00000004010000000000001400000000000000290000000000000160000000000000002900000037"], 0x100}}], 0x2, 0x0) [ 2803.041455][ T2356] bond1261: entered promiscuous mode [ 2803.047668][ T2356] 8021q: adding VLAN 0 to HW filter on device bond1261 [ 2803.286903][ T2360] bond1261: (slave bridge1182): making interface the new active one [ 2803.317120][ T2360] bridge1182: entered promiscuous mode 04:12:10 executing program 0: r0 = socket$inet6(0xa, 0x6, 0x0) sendmmsg$inet6(r0, &(0x7f0000002a80)=[{{0x0, 0x0, &(0x7f0000000340)=[{&(0x7f0000002f80)="3d59d3bf0c21ecb00e86fdcbfc4b8a3d59a5675e7708b00af43a76aa5f084f9b049eaaa3f58988747f308fe5372d232e07a1954656f69fdbd0db9b8db353d813cc9fa3e3b1d0bf51adf59515c4c19a1196dfc83701040cd20f6ecedf66d51a7c972f71b0c1e55f5b6601bbd1dfb218c41c28eeb426fb8d0529d0e1dc5cd7a62cfdb4d167892e46a7e06d9b55ff9ce3a62dc3c2be09b96e5889be264065c527c8989145687fb5d98083181be3cd4ab89d396c344bb55e2865c29602973e367b16f315567de5f7746ddbd6a788782c86c6481b11328bffce20d77a854ca7f908bc02374415ab56fc0ff1144a1de83402ba61eabe4f082b985944483272e550fc4d8dcf6ca7e59c027100e35b38fdf1d837c07e550cbbfcfda412549737bbedf62f4e74596c3c8b6a80048fbdde3f8966454c098f6ec4360a028fc13d43c6ff8c292122b02b11d78a687a3cbb4233cc3779dca2a4b6850395a9a4a8f3a3de6ee4c83f699e", 0x163}, {&(0x7f0000000240)="cf4d32a806c5dde1d7cdeafa6297e40e3c7d0016cb364d1996ddf21fa8106027bd14f68f05791694e509b7510223b9b41ac0a8d2d36f7add06c9fda739d1585d9a372e0164aa44a75e9f356e7335ac11dc5c24b509d633f5dd589a8722062c44f96ee1ac04c2b7cdffababcb67a0f87826e2473d5e7775a9", 0x78}, {&(0x7f00000002c0)="f197136ff72ea7d5781b8d690a546c1c3e1d97d5172ee6be3bab68461dead12e4f1e1817c9d2ccacbf215c1828b8d45772339b51214ef5dac93bc056819973e09fc60c2eb8c780839001770f6ecc33cbc672440ce2", 0x55}, {&(0x7f00000001c0)="887df9d949d3", 0x6}], 0x4, &(0x7f0000000380)=[@dontfrag={{0x14, 0x29, 0x3e, 0xc6}}, @hopopts={{0x158, 0x29, 0x36, {0x2f, 0x27, '\x00', [@padn={0x1, 0x2, [0x0, 0x0]}, @enc_lim={0x4, 0x1, 0x10}, @calipso={0x7, 0x48, {0x3, 0x10, 0xfe, 0x7, [0xc5c, 0x616, 0x80000001, 0x1000, 0x82c0, 0x3, 0xcc, 0x7]}}, @generic={0x9, 0xec, "c09f950ede99d3786881aeceac937f1fa5dc5defb8283eba3160d81d2d96ea222c0c88c0c7609eaf399113a60a4a1ce469d5f529fe54e28d00a8fea66c3dc8d8ae4fb775589b176a72149bd1486978b742a1a6996cc32bf260b67533d069cb74cd4911d14bc28eee0ed7af07f1c1df5bb6c49d49a6af0afad0ee28fead6e0fcbf64e0377ecae0148b2dd075ffa644a44a0f17401c4ce8c421ea19856d7a5cd38d79981b2dbc1263a200514294bf994114dc81487e7920d5c61c0fd0a0b54898637b3297cf00f3e69863521bab3d7314b6dc1f75bdcd872c4cda5d8cf84c12955f1142b0e4c1c9e7e633961cd"}]}}}, @hoplimit={{0x14, 0x29, 0x34, 0xf336}}, @hopopts_2292={{0x68, 0x29, 0x36, {0x33, 0x9, '\x00', [@ra={0x5, 0x2, 0x9c}, @hao={0xc9, 0x10, @remote}, @calipso={0x7, 0x20, {0x0, 0x6, 0x5, 0x8, [0xfff, 0x6, 0x1000]}}, @hao={0xc9, 0x10, @private2={0xfc, 0x2, '\x00', 0x1}}]}}}], 0x1f0}}, {{0x0, 0x0, &(0x7f00000008c0)=[{&(0x7f00000005c0)="4461b57a6fdf2d3de73d3b8d79351ce7124a71937abf928b75f51adad1096c8f1844ea3ee18e78d7126620f013f90334a3f22c570bfa1c926f9e3ae702f623ada8e643f7e40ba4e5495b667dbdf1e2431491b7fe93a6f6fd591ada3296467b9dcef63ea62a52b1cb90ba3bdcb3606b306ca3045dab0aa3aa44c10c556e6f99e1c18062e5894c31131c27e42f66dfde5ab43fcd7dbed825c3e0abfbbad3044bcbe5e908a4468ca35427b8d01279f0af225c5b03c645b5469cb88cefc736210dcdb12226fb5c13aaa04c082e7d09ff31ee411e32b7c82a5bfb7699c7019852fc70d864169843bf6485ac7640074d6ffddedd5111ab0bf1cde97b62db20", 0xfc}, {&(0x7f00000006c0)="0f7842ccb0fdd3effc0fa5d3d22d41e2ff8871d136a7268291dfab1b4385be22ec24d511", 0x24}, {&(0x7f0000000700)="8fd28462b5ba6fe0e7056391ad6becb03a0c56320bfc4e03b22c725e248d2f2828b27c92cb466fc2bd834c581fd40b8b15ef012ec53c73e7ef3626a2925ffd391234c631c92aea0c9f17b538685342fb9b1a4347fcdd824ccd2f5b5e04bad2dfdb0c76b88e566740bb5b6faf8e3995913df2f378dea8cdb0868d65ae10bba6bc8694d2aba24abd4af1f9cd2e9ce6aca587079e6b58c1867bb47daf9e79a45e9c8dec6af0648948e9fccff3eb432b10865ea435", 0xb3}, {&(0x7f00000007c0)="9191e7c4fb9d80f3b5d236af84fcd07b2b14420baf41ea4cbafe733bb6f8e401ffa94bd551a21fb684da758f12d153f04b4bc640adf6a50511cbf409bedf28a643edd8b8365f57def13f1171aae198600c1305a7f99813ef673842def974c2e335dc964e176390dc35b690192cd4205f4425c2b88759201ccbe91d233362178f3706bf02e501d38fb338c8fd2c3dbb3e718fce028197360c1e8fc48829e00c9b33569cb85c9f764ae1bbd2ae8063370fd78681702187b60fb5ef714babb0a0914c8aed6342b2a773453a76864d5af69c99aff39b1e1dada3619cb9dfafadde399039e8c0ef0bd5c6385d7017c2da470b22f2ab2021b3fc", 0xf7}], 0x4, &(0x7f0000000900)=[@rthdr={{0x48, 0x29, 0x39, {0x62, 0x6, 0x0, 0x6, 0x0, [@rand_addr=' \x01\x00', @initdev={0xfe, 0x88, '\x00', 0x1, 0x0}, @initdev={0xfe, 0x88, '\x00', 0x1, 0x0}]}}}], 0x48}}, {{&(0x7f0000000980)={0xa, 0x4e24, 0x2, @ipv4={'\x00', '\xff\xff', @empty}, 0xffffffff}, 0x1c, &(0x7f0000001b00)=[{&(0x7f00000009c0)="ca772d1b02988a1159197763d4f20219bc6b4112ec5d32a575dc19d7691720094c3b870a72e330b30884de0c103056aecbbeef55397c105332ba74a0280e9bcd02a00c7b07f55dd969343afa530a1defbae52dc8063444d11a064873b6377ab536caa8a9488041fd947cf3d7", 0x6c}, {&(0x7f0000000a40)="bfae6249eade9a7c7b1b5b0049090b63fa1cc829e08c5f861289b2f3cf4514e537dda1eedd77f610540a0e3abc6287b2ff5a2a78b51296698389e6193db07caadc74d0801179d8c8462f519ef9575f17c2d478b35ebeb8ca58ef21067f266ece058b0ebd8a3dcdad12b43d396dde8f675eb5338df3abc25a6d5efd524a1385c1fb526f69c2fdca74ef716fad56919d92d66165f1074b4890d9a62c7b47abf9c602f9b7e11d808ce1730d280d62a60002887760a19f68e4efc48c39d8b7585cccf33ffcc084dbee0e44f5133320f65deb55220774e8393b7bb99f080c16c0df92764391b2decef741a99fa9e1319f591ce8e4c4a40cf1ad7832a9d33ca8cb9529bac6a8758a41a98d54033c758bea8171ca028d5bbdacc2cf58126a7363dae640198b35f4fd00eba4abd5413a4def62e0cf5500d550f4072e16140cf33b8ff13bebc85757b5715108f9518b67b173bd9b39a1fcf9601e140c1fe2e8ad3d70a240d39482f81796ce8f50d960438e363523bd77d6ba6d2acf1a3700187fec1452d4a2d43b81e7e9475c0a61f91c6ff7b7e75db5311f16a0365baed7315fb92b58d6ef467827cc4f576b8385c74e2f43d1a8c849cb6557635385c33dca77633473784421f32dbb2d25bc5fb1283da1d29628a7895657403b7f6f84251e09eb06263748a7befbfcf2972d2a85473a617e264b5ded71adb57659089ac3160cb623a172d85425d6906a2452ee0ec3aaa6bfeb9e75f54eb2e86d715c95d9e14886bc41303d88f1e29539bd150213860b9b7276b443ba98b0fbf35ec40de00f3b8d570d519ab482c8dc8539d8481d5c80aa730aa672b57000a036b24f9c754463fabc057d4a9ceae7d8fb9a90287ae332b749d39c5c9bbd6284761de11902cd2f89a31b6d015f66b431643a2324463ad4e174c6512fcc4df1947e52cae6f07b36cdeca2945b2e354398ef697c8bff752ab718431aab68fad7e58419af5aa48c47759d11d633d70755a32511653900773856f39cd12d77afe7d87e87921a5c0373512794927a17f218550baf6fe1f4d2749a8998fd98ae99c647448371d457008ee55a45284515ce8f2d4157cc16c2d179b3b62629ea39f6c53f7559e8df28e77b02028cdfde71b6d28ed67bb651a90c013ee4a7965344fdda1cd0b477b40b209dcc7b05ecf7d56c767ff3340864b6c240dbe673f6365f588fba89fe6b821b5a044c3f9cc46d1a2bfb4149c30752d9ead04e7c2ecf7fd8f8a8949acbf3a293ee470c9e8aed28918fbb62701ac9b65a3c1a1feb83ebc20235b6a93892513730397663d3afb21986230362e6864c1c72572caa8a87dc3558289661acfb885f73adb39a595816af2d60116eb4136988bae71a7eaac1cf4d0a92747f723f49665180644ca8c73ea669f997d11fdb91f31379726a080ef247ac3609b98b665d4713538ba0f360884311d565159eaf07f69ca62fd1b6030e72117b26f281c61a0ae13a34ae262e6a4bdd66b2f26531b89a0865894e992a346834c9f60d0e45ffdb70b2f808651bf03e123ff9cdb3fddefcfc45811cf2c763da698221ecd8fccecb7bc378647c594c4b5b35e9f9dedcec7a45c600c438de12babbcc1368b3a558e03421718d8fac6b5771ba8de151b0ce07299e8aba8252cda7c0428baec13527a9c596ca722c3c1e9f45cb113ce210db6e94b1e4e96e7a21688a87c439153523c161d551e9af08d27820400092fc3c230343a565fd6ef6b5b26d90e5878c952c6c56e970bd9a18c32fb3bfc3d9ba96638fa12561d813f9f60af35a8e1524eac1826e49465e0c9c6a207cf0cef82308f15dbd3e0ba3e30d0bf03ee9bbe92f4d0aab82e41bcde38d37e679661e4c04e6ff1b821990274af85be2fac43f3079f84b159d7bb7b0e22ba54b5a4d1e01bd7d27f4ade44d82a062b49fd7d8863e5124d136fd8a2bdc621b0d78a9f8b06344a412919cf4762e8ba643854121755f52c7bb1462c1c8025017bc83708e5aeae469f702ed356055352600af554bd65d96c865da27c66b05bf6f4aa21436c17ee3c999de6b0672b0e622727df8d5859a3a4e92f7c899fd238532e73b9385c90dfc1c87fcac3224cd177e87aa70c8d49b5eefdfb6f28ab203875b1e474129f356a2554b87a970085b4085dbc5646a68fc6c0cce94c917893588c5127d2ef6bd749c700733d99bf4d1e6e8d2c34813517bf4ac68640b71d964fcfabd39514f5770fd99d6c4939c62c58eb3db4278ede6918b4d4d509be66538ef889fc4dc91965e0c0bec5f8e10e6bfbed47352826406aa671bc43a47388440e229a6c5753cee24ee581a6edc2d5ce2e0708d1ef4d0eb31295faef4cbd1a39d5cf5f221ecf413da910b61c479eb0f622d857ff14d679c682fe44467b5e53dc571266412ef4e188a5b75c5611fa9e35dcd234ab3253523997ab19515a1e36aef0a326eceb3e2ce637726c9d8bdc95e350297eda308a7ee7cb355fcbb1d1dbfb47a82d6f865feaa264b2cf708557c4435555403de89ccd8b48bc9ff111e73e335985e25209076a74b0d477a6345ad01f7afaa2e9d05f230d2a9bf3168092b85a3d9b861f20349a54db717b0ab0a706e475ab1ac2ac937d7e4052a062cae045afcdd51e28f013cc7a127efdcef76fdbf7932e693931f6aa7ba49cd8e5ba9d46ff35c29f5ef42505c2ab25978817852db684eee19e4ae30261d72f8039d52f7fa406f18570684121cef18297a300d5bf8ed96b8d8ece62b258dd2dc20ff3f309bf18dd3d0da3e6c6859e657d8f14c0268de2825c53ceb758d69ae2dabc2ec4a864578da9cf3abd599683d8d4331091812f5964db8afef2842212a323a2b9a86cbaf4cbe1b5e326391d2e5aa57eaf17e73f4da7b322b443c5c40efb1071ecfc08104a4a0f717154a4bbb7d25a76b179124c9570bdc1df51510e19aa03f43fc43d1872a97e6c222ce002cef31bcdd9a2df99c226163d2bbd8e4033d97cea235bc74f5a379f9ee53bc8d793951b2af9db7609b81b91747fa6a3933adf583a6060032473f4563c28adddf432a500067f79da655db2fec3d0293cb50d61abe84f251b7d8630b276b3406a023b748bd973ab4bd89f34b5ad0f46ba4a73c667a15177394e876357db47e7c8a3199be3cae0f25f05ac60025523e4516d074aa911f2ad1424a07cfea3a78d99a7bc1c99b87ef492d67b606f890c39707c2c7ec810e03adb66feb886ea4a40302d32c8cf7b417a9e8f934ccaf69c5ceccc1b62a120cd02f44e51a3f8d217a5cc6b7b6b41226c8b21e4e28ee541c4d39e4c56d123a494f82802efbc6da8f00c47987076379fc0c5b78d31f183bee69104ab382eceddafb5ae1731ff85ac8d065fe303034a0241e79f1ca49a65882ba2c7019433962d12380ecf9faa6daa7bda1b5467a2858c95308f8e2f26de268b5bab088035bb909937c407f0e3d21f087a8e1836886c5c8f5609812891036f0a20ceaa64dbd756b9b94b19bdbd6da5c0dbb0ccb961955f013d6edc7ce0ac99d1f2636d4fe2719e5837e184d84bbf068a39b6fc92aed2c0972063a913821b4996b9ff60c74c566ed7fea899c7f0d0c7da3e08620eb421d8f6afc9b4c27c37554d0c990d5b04c8b36e5b39b76cc05a6eed52e28b2188e1c5e0f380e6c0adf7b789608c675e56177ff714ef79729c4e200f2e6a12153b04444616178167cb375cab742e5a4ba428f1383d80ac5923b766d52fae3305edad6df586e4ebcee51249e052ca5ede137536fe5a07e966a41bd6dcd06e7ee3feba47a62474a4404d6c4cfbcce824fe7b96b550c248f2279d82f0537b5e79fbabf3853378c1962e0710fddc86bab1f1a3728fb8569563c8b83fe89ad308623f7ff6c1222fb303254f21ed4c646eebac11abf5cf1451e2c0c32cd537afe696024ff8e118229d4f47c71f9f3fc6eeb45e4e8711f5a2ade9d9c608ce89f0bc2ad55315ca1a23c737de7dade00e84788c38665583503ffe41182a578f3d05bfa2994ecfb9b5cd60e6277f257e433ecff9d5da9ae462f5283153cf519f4523183873f80e966489eb5ae9d2155a697faa8690ed4bf47f117d0639fea581acdba161230b1f7a7bd6cd13f9141a43c3c1ebb995a70e7a60e9c5468fb7c739ffbcbe45ca2f45e1b2203efa685be5d0ba7b1cc55017934f3fce376450adf65c59436568ab360b678b185f8af3bb5c798168169698ba79934ae18ff395b3a3ae042c69bdc53bb5b47757733a0e2e305b9b51e49c95d59d190930ef42ad92e99db757b8d69e21ae31364e9271199ee9a830d2a876cbbdd09228dc3a9e2329e61809bddf254cbaab16680f33d68add2502ff589b2106bca9cf73553d74e53c236ebbcf5131edddb566b0bf0c6b21c4b42971a072ff950aed17e1a7523921d685afb7fa2b5298a89b2214d69e3bb36f08e8593483df84da66e6851d13b35bbaa9ff8861e1a94ac2c6429448c64f827143f29e35d8fc65ebce8dc5a58e39e4437307a6b65d0d32876cef6df1c085e4876f509a7c141c9bd19f3d06a1a9a186a68be50f148d31f0545be965574faf2f980a91d087ada11cc43733a51bb4a8aeab1221993f9b1c7436187f1a2fcd5ee8a80d93f0de8b47d14a72a41e7f30d9ee1a051d3bf9d0c89b401d433d7dd52f9009dfe56a9f8cb74234d70f038f40ca1e9c9f4d239f5c9422bf43141aea7268954417b29eb47fc0d8cf4a536d0d62fc8ebab5cfe5f7bada910d7ddbef365862b407197fd32aaafd4399d3847b12a1cd2e4ea9fbd96836753f9dfbf54a8ee6a9f03165b56696be89a12fca80d4c2c07801bc3d82e9056f69329a94711a5c44e230f9aaa88400ee1a1fa04616829289ee69ea0081084de8624b25a5745b269bb5b909c1c63666a022d7dd7bdfe4862324454626a473a2b49804aa7803b6f20c8383abf0e564a9ab2e1f4c58db0ad26d899dd529a76d81e8c0ac2c3ea30a722fbdbcc8eee5f384e930b372a08958df765d06bfe1fd40bf244577172c34d44d799935c3fabf8848bf20d02e9f4d0174e235dd86662928cff827fd4d9c6217530a34756b0610a53a25e54eac528df5aaee081c48e22bdf0872450c8e83c1aad037612acf164121ca9440a03e16771cde8905f59d94d0d79c01391ee1cb19c30d93bb6c17f6dade71df0233102014c1751d09c3d6597f7335cc0b444c36328018994b5526cf2cdb74366615dc6a17d657c12a2e64685667d532dbf7c50706d58efdbbb32e1382217b3d7d7d284d971bad5fbd8c3ea979c051b948cb72ff90a682c1e6935e3e7fe9078239627b910d10daa5584e2d9b89574652f5e0e58eed13d121effad1d7b7dd368e4ebc7aba045898ac13f13fc0859c40822bc820621e7b29655ae476007c7964cbb027164432d68dff004fb1b8f9d4157aaf853c17f494411411ddd82846a0c66104fc2481417ddbfbb8df979caf4040569a2b671b6cbdc2f9c6b029dcd2fa6a50538f8f329240ebba158b3fdc424af2832f90bd7143f4f20e191fd5830f705453436df5e0ce9903c7a257fb551a6a2b15f1c186344b0f33eda24deb88e833f30694ca8e9f933ff30518060af6a524965ec76393b8494e3eba852147d44b8112d28665cb598219875c2bfa5d19ecea67e07d1b791d8815f2475ffef1f400108eeb5a689603d453670ac3a2c0d117854870b4c406ecab5965cce651e8d61f9a4560629636fc017fadbf9bd9fb52dff5c5a8c97451ef053ae45e551dc305f6fc4c6369f933f056a49e5dbaa10b2a842f2dfcd06a9f19fb24e30172ac40a72da37f3399ebde0712d6f1caa1f6f77f8", 0x1000}, {&(0x7f0000001a40)="71d313ad79bb3ad89b871ce7d902439711a1cdbf0644c242ff08692171d921eafc76abbb416dfba74ffe975ff649afa23635e300a82b8e727dbaf85feabcf688c40910a580821a3483d1d6e1c0a54801064ae141439e5ed3c09fb300767f0a63813234e504a7377643fe72c5c186fb8e6ff058a1ac527d734a012c628abd7825be17830e51c8fbd32b843f127e86ec96a239d4b9ddf0d230", 0x98}], 0x3, &(0x7f0000002c40)=[@hoplimit_2292={{0x14, 0x29, 0x8, 0xbd}}, @dontfrag={{0x14, 0x29, 0x3e, 0xa0}}, @hopopts={{0x78, 0x29, 0x36, {0x87, 0xb, '\x00', [@generic={0x1, 0x31, "a8feb4aba4f3df4cd116837ad3dc1907d5962b9ed7d744b334325e1c1e90ab28aa1da93c73cc658a5853eecf8c496e9668"}, @hao={0xc9, 0x10, @loopback}, @calipso={0x7, 0x18, {0x3, 0x4, 0x81, 0x4, [0xdac, 0x400]}}]}}}, @hoplimit_2292={{0x14, 0x29, 0x8, 0x8}}, @rthdrdstopts={{0x1a0, 0x29, 0x37, {0x87, 0x30, '\x00', [@padn={0x1, 0x6, [0x0, 0x0, 0x0, 0x0, 0x0, 0x0]}, @calipso={0x7, 0x10, {0x3, 0x2, 0x80, 0x40, [0x9]}}, @ra={0x5, 0x2, 0x9}, @enc_lim={0x4, 0x1, 0xff}, @pad1, @generic={0x8, 0x139, "054e4105e7d2ad58f0b2b7c301f911b4dbe7f3e41b2357fb37da641670558c00342bdaf9576a688fba09e8efffd134d9549094dbeb33cbe0f635dd5ddf6424a21a5cbb54717cc2dcd061cc03a8e3ae5a520621863adeebba04bab3d76556f2809913e5636b7b70996f6eef19fda744195019c6b66c43f61212168947cf8cfa3702223e9694fbdfb6c62f69ce754fc7a5ee4c089fe9d97c24fbba18dff47ea60c46690f0de993eabe1ce07987e46761c85e6fdb016f3eae7d4748bc8b206792e17af1e50b0aa5e6bd48a27565804a8b22319e5cfa65eb918c3eefc813f332c81928d5a8022ca912c51e6c97c4e2a8ce61d22ac6223713c4d830e53eb9788e1b9adc87596d6c616e75a0a9c192eed83919addf1f2791202cafb3aec2b68055598ed41b2ebfc984587fc9552360a57a59febe7cd6571d8a07a4e2"}, @pad1, @generic={0x2, 0x1e, "be4b70986db39e998ac451e9f34365db2875bcadd52acb8cbc8189dbf048"}, @padn={0x1, 0x1, [0x0]}]}}}, @hoplimit_2292={{0x14, 0x29, 0x8, 0xcaae}}, @rthdr_2292={{0x28, 0x29, 0x39, {0x6c, 0x2, 0x2, 0x20, 0x0, [@initdev={0xfe, 0x88, '\x00', 0x0, 0x0}]}}}, @rthdr_2292={{0x68, 0x29, 0x39, {0x5e, 0xa, 0x2, 0x7f, 0x0, [@ipv4={'\x00', '\xff\xff', @local}, @initdev={0xfe, 0x88, '\x00', 0x1, 0x0}, @private1={0xfc, 0x1, '\x00', 0x1}, @initdev={0xfe, 0x88, '\x00', 0x1, 0x0}, @remote]}}}], 0x308}}, {{&(0x7f0000001ec0)={0xa, 0x288, 0xffffffff, @empty, 0x1}, 0x1c, &(0x7f0000002180)=[{&(0x7f0000001f00)="21f0041d16cf4926402beac7019ec214cd5b9f26ce26f24e8de103a5c2ae80b04ffb5f227bb8e951146c19ccb58030f38f61351f0f18f066fef18c532a20f3ae6f5acdc65dd640ff495579723c3c58f14d0891005454e8d706926f63bee1b4c6fb512c08ff83c2e4b0898ad0fe972454a11fa739fb6f31cca91af535", 0x7c}, {&(0x7f0000001f80)}, {&(0x7f0000001fc0)="f7a54c5b5248ad3584ce4db0828669034d786ce16641668766089dfadfc71f91f20c9e1420752fbc029c35979ab2401fcdea5fb9c09789f15a0c2097d5f2ccd40d0e1fd67bfd1b12a4d57f52384ec6001e7d086095165d6127396471598b439e17", 0x61}, {&(0x7f0000002040)="cef5729c07dc819c5f5d947ee0479bf931fb68e7f06baf0b3653144c2fc07d302ae363e4816339b0852c31d8a9c8c9a10795e24e94d51828eec8dd24cba6c716513f426dc597aa26f0859821a9b813196fac530d7ecda89d74148aeedcd656ea7c07d6537e8d8855fbec06d9c25d8067be2ffe1b8a53c6ccd033dc80d372f7893246b56ab6127d1a9d0837771309d86050c68c9437e42205b8aa9bd786733db5220832ff0d75cff62dffa695d19015bb4b6b45557a9d1e677639f2adf95f8b826aae11f3829d0abf38755c1b70a8d344026d7ecf9e75b1eebf94fc7007d9425e6cdf55d561f279d3356e", 0xea}, {&(0x7f0000002140)="f702633c572590cf52aaaa1ec50b478afa6e501be3ebacdeeacc78d64d65e66bfdfbf77b1f04", 0x26}], 0x5, &(0x7f0000002200)=[@hopopts={{0x28, 0x29, 0x36, {0x2c, 0x1, '\x00', [@ra={0x5, 0x2, 0x9}, @ra={0x5, 0x2, 0xff}, @enc_lim={0x4, 0x1, 0x6}]}}}], 0x28}}, {{0x0, 0x0, &(0x7f0000002340)=[{&(0x7f0000002240)="5fc7264e044d38ba6c7b934bc26c7f763442739e829deac60be6d5127c26c9aa1a0eb608de20590d3dc4a76c0638fd4cb7c93d99fa5db8da1203d43179da22c448f1cd24d21ddd0af6d92dc98dfe5ae9e28bf188e09d8e7ca6a198e9d8c027b80f78c1fdcaec04969a8fc2c1b26432a661ec6552f8470cd0f6968b2e1c20fabe523796c2a49c59745b2e6755ba94c0bfc7b6d97658c34184ad4db4735a71408056407fbf8534c788236f2b9628b3f6089f9d0cd54260b3a4aa0348732b15f94b0b443a150f4fcf7435725354abe98daa5f1d022d3087c858c0f6dbd5", 0xdc}], 0x1, &(0x7f0000009c80)=ANY=[@ANYBLOB="a00000000000000029000000360000002f10000000000000c91020010000000000000000000000000000010100c2040000000501070000000000000005020009c910fc0200000000000000000000000000000001000740000000020e030002090000000000000006000000000000000400000000000000c845000000000000c4000000000000000600000000000000c1000000000000000502007e0000000000a800000000000000290000003600000001110000000000000740000000030e00ffff060000000000000009000000000000000000000000000000f7ffffffffffffff04000000000000000000000000000010020000000000000005020007c910ff010000000000000000000000000001c910fc00000000000000000000000000000004018b050207ffc2040000000705020f73c910fe80000000000000000000000000001400000024000000000000002900000032000000ff020000000000000000000000000001", @ANYRES32=0x0, @ANYBLOB="0000000040000000000000002900000036000000ff04000000000000c204000000010103000000c91000000000000000000000000000000000010800000000000000000068000000000000002900000039000000840a018000000000fe8000000000000000000000000000bbfc010000000000000000000000000001fc000000000000000000000000000000ff01000000000000000000000000000120010000000000000000000000000000a80000000000000029000000370000003c1100000000000007100000000202b8b60401000000010000000720000000000601ff01080000000000000000000000000000800100000001000000c910fc0200000000000000000000000000080730000000010a8108000400000000000000060000000000000080000000000000000800000000000000020000000000000001080000000000000000c2040000003f0103000000000000ece3a8518ec14f11b19aa4eaf46cef187cbbd6c81d01caff3c98028acdf0d1b9f43cf082405ed4d449c39e0852c3018e5dcf84afb3810334fe238fa38683e11f6ee0d68d76bfd9bcf9416d2061d1c92d5e"], 0x2c0}}, {{&(0x7f0000002600)={0xa, 0x4e21, 0x2, @rand_addr=' \x01\x00\x00\x00\x00\x00\x00\x00\x00\x00\x00\x00\x00\x00\x02', 0x200}, 0x1c, &(0x7f0000002680)=[{&(0x7f0000002640)="adb91e6a4dd0a6a089be", 0xa}], 0x1, &(0x7f00000026c0)=[@rthdrdstopts={{0xb0, 0x29, 0x37, {0xff, 0x12, '\x00', [@hao={0xc9, 0x10, @initdev={0xfe, 0x88, '\x00', 0x0, 0x0}}, @calipso={0x7, 0x28, {0x1, 0x8, 0x83, 0x1786, [0x8, 0x100000001, 0x4, 0x7]}}, @calipso={0x7, 0x58, {0x3, 0x14, 0x4, 0x573c, [0xffffffffffff744e, 0x6, 0x4, 0x1, 0x8, 0xca, 0x9, 0xfff, 0x200, 0x24000000000000]}}]}}}, @dstopts={{0x160, 0x29, 0x37, {0x29, 0x29, '\x00', [@jumbo={0xc2, 0x4, 0x9}, @jumbo, @jumbo, @pad1, @pad1, @generic={0x9, 0x78, "c77ff41b188929e2b19171eec4b6fe3f67811e680bcb9fe474ac5247d42111cbe5c14d440b928ab9bfec00d3ce2c7b577877a0812875285180f8e59eb2106647f89f1498f05c7f8c116bc11aa94be33e43b249820437be370b168d770491d98cc128eb6afd84ac6e71efe0f6d7dd8f6e5d6f85961ae6fec3"}, @jumbo={0xc2, 0x4, 0x7}, @generic={0x31, 0xa5, "1d27f16f17c32631ce821e89590a8f4f248f93e206f8d8207a93d6a2fade3e0c6b5062a514882c7135059271cf628d8ae367210993d3586f7ab7cad5ce19a40e065488187ba2f413862c88533fcfb3a42d7c1e4474547a51ab4cc7ffef6742cae94d8a7a8bfd1c649225a2e2cac29441cc9f1281bc9f88e75479e393bab79e898b477efdc3b7318fa8b82d10ade1ddbb4ea9bfe9cb89d67d217a9c7c88df424534b49a594d"}, @enc_lim={0x4, 0x1, 0x1}, @jumbo={0xc2, 0x4, 0x9}]}}}], 0x210}}, {{0x0, 0x0, &(0x7f0000002a40)=[{&(0x7f0000002900)="b91f4773fefaed659041649349ee9148e5650225f033731a90eb90ad5c59552c2fa6bd9d6a11a079db0cda5d9a151d785c34258d0d30a6167bc787bbd01684c7319d9cd49bbc2a1c16132b701852a9623c1d8a5d5e9519128f819ee3ce54c3637dffe6ed1b2e918ba05a65daff464f2ad8f5e88d85418fc1652c0eb30cd9210816ee38afd6e478f3029427c161a2a5d7d3cbd8a38aaa8f1ec920668cd730f864ee980fc2f197da7cba805629659b", 0xae}, {&(0x7f00000029c0)="cff84b8ea37ebc5980412998da40e782fdc386b6054574a1c5aba414d9c36b732e1697fc5bac7e290c70af9514d2432d0cef12894789bac734f2565c356d179c2ced51adf0b8e4c636df243822c5991cc333ecf2516c2b88b8b86c489884d8bd", 0x60}], 0x2}}], 0x7, 0x80) (async) bind$inet6(r0, &(0x7f0000000000)={0xa, 0x4e20}, 0x1c) (async) r1 = socket$isdn(0x22, 0x3, 0x25) ioctl$BTRFS_IOC_SET_FEATURES(r1, 0x40309439, &(0x7f0000001f80)={0x2, 0x2, 0x8}) (async) listen(r0, 0x6) (async) r2 = socket$inet6(0xa, 0x80002, 0x88) bind$inet6(r2, &(0x7f0000000580)={0xa, 0x10010000004e20}, 0x1c) (async) setsockopt$inet6_udp_int(r2, 0x11, 0xb, &(0x7f0000000080), 0x4) (async) ioctl$sock_SIOCGIFINDEX(r2, 0x8933, &(0x7f0000000080)={'pimreg1\x00'}) (async) ioctl$BTRFS_IOC_QUOTA_RESCAN(r2, 0x4040942c, &(0x7f0000005740)={0x0, 0x4, [0x8000000000000001, 0xffff, 0x800000000000000, 0x70d, 0x100000000, 0x1c00]}) (async) r3 = socket$inet6(0xa, 0x6, 0x0) (async) r4 = socket$inet6(0xa, 0x80002, 0x88) ioctl$sock_SIOCGIFINDEX_80211(0xffffffffffffffff, 0x8933, &(0x7f00000000c0)={'wlan1\x00', 0x0}) sendmsg$NL80211_CMD_SET_INTERFACE(0xffffffffffffffff, &(0x7f0000000100)={0x0, 0x0, &(0x7f0000000140)={&(0x7f00000002c0)={0x24, 0x0, 0x5, 0x0, 0x0, {{}, {@val={0x8, 0x3, r5}, @void}}, [@NL80211_ATTR_IFTYPE={0x8, 0x5, 0x3}]}, 0x24}, 0x1, 0x0, 0x0, 0x20000080}, 0x0) (async) r6 = syz_genetlink_get_family_id$nl80211(&(0x7f0000000c00), 0xffffffffffffffff) (async) r7 = socket$netlink(0x10, 0x3, 0x0) ioctl$sock_SIOCGIFINDEX_80211(r7, 0x8933, &(0x7f0000000040)={'wlan0\x00', 0x0}) sendmsg$NL80211_CMD_DEAUTHENTICATE(0xffffffffffffffff, &(0x7f0000000680)={0x0, 0x0, &(0x7f0000000640)={&(0x7f0000000780)={0x4c, r6, 0x2, 0x0, 0x0, {{}, {@val={0x8, 0x3, r8}, @val={0xc, 0x99, {0xff, 0x26}}}}, [@NL80211_ATTR_REASON_CODE={0x6, 0x36, 0x7ff}, @NL80211_ATTR_LOCAL_STATE_CHANGE={0x4}, @NL80211_ATTR_MAC={0xa, 0x6, @random="94f538bc85b7"}, @NL80211_ATTR_SSID={0xa, 0x34, @default_ibss_ssid}]}, 0x4c}}, 0x0) (async) sendmsg$NL80211_CMD_CONNECT(0xffffffffffffffff, &(0x7f0000000640)={&(0x7f0000000180)={0x10, 0x0, 0x0, 0x1000000}, 0xc, &(0x7f0000000600)={&(0x7f0000000400)={0x24, r6, 0xd57130b9fd170e52, 0x70bd2c, 0x25dfdafb, {{}, {@void, @val={0xc, 0x99, {0x8, 0xc}}}}, [@NL80211_ATTR_USE_RRM={0x4}]}, 0x24}}, 0x84) (async) sendmsg$NL80211_CMD_CONNECT(0xffffffffffffffff, &(0x7f0000001600)={0x0, 0x0, &(0x7f0000000300)={&(0x7f00000013c0)=ANY=[@ANYRESOCT, @ANYRES64, @ANYBLOB="f7ffffff00000000a7002e00000008000300dbfdefb8e55cf0d06fab513b7f37d1a1afe926bfe9f70ba49d8401f952227b27f3eb211ea62159130000d4aa0c0827a9b6fd21e2a15cd38034152274b81d257090fb49e7c9d05d1691d0cc05e8ae0a7d0a75ba944c3024f8f920cb5ed5cb42cf91b00c93bdcee7e679776dd67d224b746980e208cc5beb68d1290d9360e307034d31c6e10c8a60a27ff932d8b1ac038204a0080000ba65c7d88dfebc62ea4631984df100aa3a7ca380389e94e48f4ed4e807e9f03025798d059b2844d8fb7c36a8c6d57388d3cca2a0c29a248bcdb8e65b98025468b6dea1ab01dea1218b0779ee8307566fdc0fd20501606260df73ab2a1cc4befeb4beacc0c58d174895371b7da77581787819707d32114f2ae0a5c02e3e9ac7ff0bb7a49d539091fdd7f2369d180fc56272c0e91ffef200b47d338b00e19768f62debc93d85061a89e1060f52be100083dbc94de1ec70a630c8ce5d4a375f4a8cca0135a560603538b9ee2bf155ec17953c896b21ae57632b6abc25edaa0129a1afa044e6e5bb8e54288e5ffcc985a247f7bbef952568d7f628a6555cd82224c1f1bed5a5b7ec63c2d6c55fd884fbda02f017bf7c2621884aba46667372475c14d05a0e39ee2572d287fa1172a7b88efd47339c35fefca982b35af50999e1858a42767382ae47de6bc6ddeeccfc8a664693ca3cc3b0777139f2f0", @ANYRES16, @ANYRESOCT], 0x28}, 0x1, 0x0, 0x0, 0x4014}, 0x0) ioctl$BTRFS_IOC_SET_RECEIVED_SUBVOL(0xffffffffffffffff, 0xc0c89425, &(0x7f00000001c0)={"bfea8b8b73e4828708777203dc85f1c6", 0x0, 0x0, {0xbe}, {0xffffffff, 0x8}, 0x9, [0x80000000, 0x81, 0x0, 0x1000, 0x80000000, 0x8000, 0x1, 0xfbc4, 0x0, 0x13e, 0x4, 0x1, 0x7, 0x0, 0x1, 0x7]}) (async) ioctl$BTRFS_IOC_RM_DEV_V2(r4, 0x5000943a, &(0x7f0000008c80)={{}, 0x0, 0x8, @inherit={0x60, &(0x7f0000005780)={0x1, 0x3, 0x2, 0x561, {0x15, 0x0, 0x101, 0x0, 0xffffffffffffffff}, [0x4, 0xec, 0x4]}}, @devid}) (async) bind$inet6(r4, &(0x7f0000000580)={0xa, 0x10010000004e20}, 0x1c) (async) setsockopt$inet6_udp_int(r4, 0x11, 0xb, &(0x7f0000000080), 0x4) ioctl$int_in(r4, 0x5421, &(0x7f0000001b40)=0x9) (async) connect$inet6(r3, &(0x7f0000000200)={0xa, 0x0, 0x0, @initdev={0xfe, 0x88, '\x00', 0x0, 0x0}, 0x3}, 0x1c) (async) sendmmsg$alg(r1, &(0x7f0000005580)=[{0x0, 0x0, &(0x7f0000001c80)=[{&(0x7f0000001b80)="ed552c167d4a55cc3c22be739c47e47179495be7e6294e52de29a9a65784e8d149156bcf08ef631cab29d2e17859e7892363d6d3c366232b422d7a55fc5ce32e7f561cd8d80b0b6c9974a37bbf6bc9c45f956f57381c93678c890d154fcacca6df17595335c18d4fd99b942460ed0c025a06e081bf7e21bb863defc30ef7f998cb96494a60ffeec419fcddddbf71452dbb5f3600d3010a1eb73c50c3c3beacff10e09d85c9e768084f107eebb93aa64af5255826ac753b1abe7dfdd25e0dda11365d5db95746b677d710837f6bb784ab4ba16ed107269fb25f9f56bd4779258c56a7d50c5687a16417389f906e09", 0xee}], 0x1, &(0x7f0000003240)=[@iv={0x38, 0x117, 0x2, 0x24, "6edbd9fd16d13be3badcce91fca07ddb8fec1f02568f62197bff5aa1e622f580d2c70542"}, @assoc={0x18, 0x117, 0x4, 0x8}, @assoc={0x18, 0x117, 0x4, 0x4}, @iv={0xe0, 0x117, 0x2, 0xcc, "f95f34a2718300908485e95b17b4655ad21d41af33566aaa3acef7f7c9037989175e76f33273ef89bd6ce9008796d831e099f58c0ce6c1ae63e4cd375fa98b0872fc482e2598a85f05811b92a6980fc2312ebf3ce0060eb24e06fb1c26f3241037b27d0644e04084bdd311cf60849c7b079e5b8eec26c9f74437ccc1190e244546afa544abff041d3888d7c2649fee491755f9c95cfb6345de1e91610f065f30ea75bee4011eb8a45277aeb1469a41cc94ff8982bec906bd0bce2f876fff225ee54ba9645b7ace154e6fb5e3"}, @op={0x18, 0x117, 0x3, 0x1}, @assoc={0x18, 0x117, 0x4, 0x480000}, @iv={0x80, 0x117, 0x2, 0x67, "fae32821d51229f830c86660d8ff59b74b08e41910816705269ff1f914550d46eb00b278863cdc57df68a8d1fcb79e6375bf1765e93c7ee927dc31c6113bcd9adb6ba20e6350bc21d40a0c75426623082ba2f746438c45c4bed084945ab7ab593e2d9e142287ac"}, @op={0x18}, @iv={0xe8, 0x117, 0x2, 0xd4, "5e1693468fb457c5db72d7cbef2c47d7ca30fe75a1ff264ddae90fc27e6799d82a5ef3f9b68a3462f5aa8099dd9d68ec5367eb395da7877438d03168e535c108966231714640d9ad3b2e52c14eceeadf67ee1bb8d5afd7662e945d815c5052e079984190c6dd51a6fbfcd4c42cb09eea4344fa91fb41ba581885609c89780338898ebe70d323690403e5b41f3e8dc24e8096802e6d1faccd98af3adc198c19e6e38d530ad597112ce7752e9bc56db2109015a9c23c60c6f83f16a91e5db0990059e85ad2ba9e26ef3bf63526f04b2243f40a4acb"}, @iv={0xa0, 0x117, 0x2, 0x8b, "a09621c66c9fd035ec3600088f10e93c4ed2f210edc75110f71fa9a97d8732628f9b5420e90093b458927c33ad71db27caf66ca49125e7e0d329e199e91389148b674df5d85c74b13133e1e563cf731c67fd0e40796db83691cff4626d82cde8aa56271d70229b0a51ecdafdc46a3d417c0044b6f21648198e40f50cbc79b93d22db60c6fa20544f05720e"}], 0x398, 0x20004001}, {0x0, 0x0, &(0x7f0000001e80)=[{&(0x7f0000001cc0)="65a1fd90cfac3fdd1549ea83afbaa3c9960d03df16e605ac2993ceafd0faa39de2b519d30e49afe4b83d76159910e5356f778575a696f4aa9e21f7abab47f294b23ac00cd011e993e74018e460c1d7b0eb8fd05f507bfee0b0db687fc988c8e87c31a0eacf539771043c3ebad7269eaf821033362a5d878018e03bbe9ce71a282a642312347e9b5970c5c6", 0x8b}, {&(0x7f0000001d80)="288ec28e089a68b53a451afe55a1fb44922fab34adb87fa95f15758edd864448e8f9ffd34f07fdb2447ebcea98414c0d9e84f3c620303d3c071ab5bcd4dc3fbc72eca8", 0x43}, {&(0x7f0000001e00)="ad6826acb1cad947648f197812e6f2e91794", 0x12}, {&(0x7f0000001e40)="fba91c77a4d43c9b71d07e1994b32cb626682b1a357083282aa153b880b2d219d3cd8381abef9941f0d56085792dd9bbb2b1d6d617a6dc263d", 0x39}], 0x4, &(0x7f0000002380)=[@iv={0x88, 0x117, 0x2, 0x6d, "915f54d745332b14e3238fc60e3d9c70338727557d882b9d0a1f4eeba33bf37506ef73d7244388835a908c6607cce6adc097de0327a5bb68ab2c492878d5db9a6a28dfa009574639bba09ba4eeb507b7af91fe8c3b6ea7cd1da2bd00ced9976399a5f921a1293a24fec27650fa"}, @assoc={0x18}, @iv={0x50, 0x117, 0x2, 0x3b, "ddedebcb225501da7e5f59a3ca9856fae49e65e9495a9cf475de8505c0b0217b4e93b468ec5ff09e1e79b11d8c1f84774c162d94dd036f03304cf4"}, @op={0x18}, @assoc={0x18, 0x117, 0x4, 0x10001}, @assoc={0x18, 0x117, 0x4, 0x8}], 0x138, 0x8000}, {0x0, 0x0, &(0x7f0000003780)=[{&(0x7f00000024c0)="df8438deea21dc7d46593dbe4f24855ce6668a8b882ad89a5efc1bd4539c460851c6c02c3fd2095e53750d5c1c0fd87fb6ad969efb56bd03c2b1156a47637f11785c9e0a84da78e3173b3fe78e9d61b1a16f03969187d72cc88b0f561811fde58eb2d65ea8b146c0d4c4ceee2558215b1347aafbf6ab1b8ebdcb6e5175ae0c600f7d82042e38d7b09bacf529efd473a9ee9d9ac0f426a7979c20a41745a2176496c38436849050a56994beed5fbece4661e33e937737c2d998138ed8", 0xbc}, {&(0x7f0000003600)="a57cd85e0f340b4961dec3cc1b9789332aef8ce7e2c1437b884819fd15c49b22894d259f9291594abb6c92754092b7f038ed9f89bdbaa8b3950f2577297d7c767b54e947a20e02e67b766c0f6efc96a68d663e99ad6cc7a720c52f533d919e2c820f2be54675206b25a7fc800e9b73cc844558e176e814af481bc562b50c2b2186e8", 0x82}, {&(0x7f0000002580)="69d59e2192dc2927503a4c96a5c421d7062ef7f087ec77da585f1b7b45549bbabdf9b3f262380f3780b965b6664184f0327634c86cc35d10defab8a6d4412065ada881f2a563a895", 0x48}, {&(0x7f00000036c0)="244727eb136383101de3c6c0d84d7e9b82e78a0dce481c7c2520c8beae36948dbb9f8a398b", 0x25}, {&(0x7f0000003700)="75e8f218187deb24d6124791fe2b563aea7f537b3905c3c4fa7b5773c83714ac52b925fc88bc322d22e0f7439774bfd0b6f6efe71ab8819cf163e58838388b7aece5a6b674020808eae29ad6a89b337fee73fc10424bd27d50f9b342ea6da758e16014d1f8b85732", 0x68}], 0x5, &(0x7f0000003800)=[@iv={0x1018, 0x117, 0x2, 0x1000, "1ce8226f58e3bb29f08fcaa189dc03d294a5e372bd2a15eb9fe2cfb4924dc37db5dfab3181b4d5911fcad08771ec058ee807770156a412155ef62c1adda20bf8e3ab3d4ac5aac7ffdebb02e2cfe024e4969411ee7d0d62d3c7c2b67c276dc4c23bd4873e79207a1fb67a6e536bdf1cbc6aab44a3bbcdeff80e1b9d80e53170eedc66285922ae4e75af2f3df1eb219f55204944b02f1fd6c0347a1c6d76e9e23d6db93c39b7e03d719e47381dbc4288a947e5f7b5eea5f1e81124d378fd5ca1010fc447ffae4feb5d2fdc1bbf7ab838bd1b576cd79a072cbdd0e532b4e71c0dc98931e480a3941632d14e69ce212e4edf5e0d180101d58adece77fa2858c377f6b70837a42c306881ec822226ea2786e2e9a4303c6c6ea8564c7adcef401c6e5c1d501175e7febccddbd7d06f3e4f086c304befd0a8aa3252079b1fb9ce14fb0775647ee311c5a0a9bbeb5c6bb38b0ed39f03716c9d023c662b60ca2dc71656444fbb3b0e6e8ca8d6a04b080f5e9d55d87283ef4ca02f2da6413d634badb448f25a8c3e21cce21e2945a5c061a1bc389fe5cded14d0fc06ad912a75918da5de990912ab7dbf67a09025eae05df2661b587135250097c7be3829451c793e22f0312b398a5b5d8335fc0d366bfa6411c025112ca70ec0879d69f8b2c3a5cd4901543940773595493ec04a8b85bced73420532fe19d699bc120157a761b1601f59a419488c6b7d3bda4ea2c9f3529b21fa8ba20ac25da0b5ec9aaf0e621a6a819b46b2a09b202387272af33eddc1787954fad70936b82c5c01c5d5c6ba6b45e6eb3cec1fbfebe1eaf604d05794d2fcb881a36056c492b5097df8db56eea13052f92bbcc273c950592c2df5f602530acae8ebe9332df938837847b970009a5c211a0bb81db4710176c14d7011c4d438b24231fad18e94c5dff77c2072bf9e04adb51b4b2317b3bf98fbe469940c4572b189752d1a00298a223a6752db6707cbb729562cf9749f2111e675b2989987595bde5bdc6fdd46284d7b4317abb6502cd5ff9c8c3a85cabc0efd4c9935a4aec76c9f219e7ef1a7546c09723a67f9fca5cfb0afbe6f38d1b7d62db14890c51ef4e8c9a08911fe7a116f23ec727641f0ae3d220a50ea0971ac8c121c76cbceddd2a33fb79e1e7835284f44f71025f2c5dd8543de2bbb8ada4c428fd2143e95db9993c549490421c1c16bc9c57cca7fed9673ae39ac97faf1c31e8b6a7afe7cd9c8b634e2d0edffdf38f8b079a7d1e6426f7926063cfaa915f7f551541f8adf3fe9658e4a57a7ee5d1b33c8c8e88fcb93e834bab79578b59f596688b2687dde73291280541fbd90b5b5fbfe274533c1d45bc8f6d93069286fb533f8dbb35aeac5a421909cda1c4614ce4eb728b215e4463ca14d38e71b90a9284677c1e9d80d580d6474e56a7a4e56c6c3bc8f760fc73153a505a83349fbb10ef4fc2f5d1ac5eb63c02821e5e36fd098750e57f664090440e419f5da146c5cfe466b3123de70a6016f623436f8f8c9f94af50bce3a719b26d3f879f816530d7f66108f8395e2e038208a75dd459223975e44e77a75f0f5fbf238df92d66f9efc9cba6b8e6f1c1b0e90ff51436ef59ba6afb93a79dc18e75b64fde6826b0e22675f22a3b785935ebcc474fd1072b3d571f9ba261adaf52525d4a17a6a1793cf481238981f7d62c74b6ad71b0e590b6ba9d75e6408571e0ec6223d3d179287e00ee128f31d4887cc9e44b4df0c5e6261f32d7bb24bb4f547e7bf2d48b98e8d488a8f9b283977ea1cf97fd3480a75af56881fc39734a8761105c280c008812d0defc2a8262383543ae3dd1a32c3347285d3495c4f0ac42290a38d17541b5a8390e27febc1ef90ec625f1e40f664b75a95a97a8a82dcd92e9d7b010cef6ebcc2b5d8f6968f4172e83807c5f596f7e9bda87988d94af77c9288b0e8d049e4151749fac4108c774093e4dec593c2f09051810e20882de1d060cb59319a4f3c4de0a127b9cc2f0547baeb8a9b1b9b05faa03672ac5733c8a94a71e300a2a0b99b4dff1801ccce65b0b58eb8eb1eeb7e8bd8543b11d7b5639f6af83afaa383b2094bce893541d8628530e7c5fa040e2a1672bfd9be5ad0c5048bf745c54af0dab6745a5644819a3c2bf3dead750984f0a8eca97b525884b8050ad250b43768c0c01509bc11d22b15121c9f2486179f7d9837e209edeaee59e457af02d9b64b9beba328b9cc240faaf58ac6b23329f170616e51df99a79ed7d3d72904ad562711f56183dc203b735ae9323c01f0e67669bd13897895850d9e08a43bdd0d6d28dff39bcc9c76b73665bc68f549d80ab62d54e5b0bf081c07f18044d4c8f8ba6ed65dec4fc77951ab0b5276613abc66d1473516d9eb1dc49b6976e2e1bf831413b159081f9d823bdac8ab8a82926e7580e54bbfe45e5de8464355a6b181f1cbb0c1ee2ad5f0f57d652580bb7889394ce7f1a7a4f17e426fbbb57b6d8c4bc2e187793839d91ffc611155596247bcb9c25b638d8ce654106a9c74315019b34eef749c847834bbb4eb78011bcc563d27cf45c2cc2bbee01b75db823320e2c162c32f31047945c431e10836e1b1f6f3b7b8c3aafe4d9655347e70c49f3d7285df7ce380fea9f9e17bc2e115c1f0d0a8d7f6619dc14beaa39fb86f1cd1fb37fe81b1eef896cb024b21bba890f2438718f83f305c52796f15f8c5c2fb874d0b83602d2510b0e5c82faaca0c97dee3a4f208ee1e2631f296c952fb2d8539eff16346abd58e09eaefcf82ac060d52afeaa6b4190cfd6f0182a876cc32575b343b6b62b75a95fd95b9a86fe95f87d739847615b3bc40126f38f579225e35864b14a924898834d149cbab4b7275027066f827665ee7821a6eab86762f2abb9fcb45475e597cf6b08da4e720dfd04d2abf0905d2396eff74d8b345936a4e176b140751e0aa824e0e9ab5f422f3d860b31be5ddc670ee45a4c453031f02a70cc1d8f6bb3ea793791f59a3dad7ceba9bb77eeae88f646c8b6ca8254c3aff7be71f944084df40b621b3ca07fb75cbce8238701fbe16212e872f78d4f0360f158de8a24560cf44db1c007a323c39086fb8864eff3eac529e877f02c461d748a1e06e07e7524930001487b4ceb605c55617b775f1e249bd1492a2926385558f4b0fb28e8cf814dcf9ff983b9ef860ac32efbf517d147518e4af7bad54c2c576dbe0cbab027c308389192d6237b146252eb1ace886d010f6d1bd4bab35e96fe6dcece551257b3b7a10418b3a0b2ec251f8d646225db4bdde3fc7b594ec01cf662225f3db40fad082fb5c2949c9a9c2f990ced3e39e419dc89c2af3ff19f764dda953ce3f66a8be4ee9267e7de350653157016e0994b07e990a210505adb4457443ebebd13239d80300f004a164da37b20096b475d168542503824c72ab2271504e01a21b32818245523301743d06d23e45f63fde95fd1992efda6a96d066baa9cc0ef5f8d4dfbb8fd29c3178e33f24b95e62d03fafbf48aacec15f71576e6f75e22812ef74739338b5aa5da5de8440fe203d523aa6853845394b78085097be22a8b93ccad4de419b29949543ddff8cd10eb9c4709e0b6b4fa9bb0cc32b89ad74ab67a19e4bcd7ce585dba83b7ab89575f32e9bd0f8da9447777f420de282faf5a8530a818ebf6b18743c5d6f3d5beb2fedfa5f775155ad8cc30dcb613e42cc212e161ae88dc3cb26fc6e73bfab646ad9e019bac667cd06a5d003952a04597f90ecf7ccb451f30afa8fbaafaaedcd9315c16eabf4d60472460374c9b7e6f74486ca6baabcc6b213d674111d0166cd1649b1011c88f8f9f1708af6f754785c087a4a07cf359ad3a4ad0c0e340d71a7357545848ce0e06443271b3f4557c51e3e9fa07df711c48fffffcc818b676dcd393df71a9832ae8a94c2bfe42332951594558c09df979b76ca13623ac6aa448a9f2ddb2e073c098030df23a89659a5b5482db17dc0acc3e73b6b63316fda7a314f048b63dd9280948a9077276e0bc0d048e084a6bc28c80cb78e4710bbe106432f2ce467a4879603d850c7f0eab13102d2c710b9a661f482d937a5f06f03138568fad3dbfea1b89db99ce44819a9a81a4a46c6c56be041a34717e65c3ea158c6ab7866331a29680ed24ac2454f9bacbd34b63bb449ebeea6866da22cbe9361a9f0d4559d4f48fd647aaf00671136f9caabc09c7f418330b647db960974dba4de8d607bb91164e6186cafce1214e1c1b61c79613056be19b86743181fe0a931fdd03f3555f2c87918b4038219f33789de1d873f9935cbb330cfe6f6c15960386853da4ccbe87594b409b9eeccd09610ff346c3bd026db1bc410037c54217a99ca6ea69081bc706b0c920a2cc2a3c749eaa6312c9e52520514cc3f3bd5cc4c9000ce681ed7e063ae0052f1d9bf190c78c466d7e621327866d79f01e45c34c7b1bdd3e081d42a446d5fd183f2b1b49b8cc0f913627479496f34823f778c8394623e903e74629563bad6eabf97296a212353bb53cb70b5a8e9d42747871fb48cf798ec22b09a2c9b419c324a7b6e8efdd62dcb1fea7085dc0bc98a761e7427bc9b63a48bda5746a850365c18df2c337835246e7a58476e6257f013fc697ebc01af54ccdde70c252cc28cff6bb8e2d7c7a076839aaae011b97a8cee2914a01b5904a027cf895c222dc4a5a15addee15d1bc40e29ee11b008248396215c822da423e467479f1c8b1073f5eec2c42c1abe698d8690f23879280006b0c5e8bb829119482279e4cc5d001dc99d68c245e07626124565a72fe6e7bd97814b52b9f6a5d9718f14106ed5d3cf996b7c2aa7bb6da7cf9ea845d4280ed2ba9cb35b4628c2eccdcc584774006f9ddfac5fa66e5984364aa1f15e4318b5e70716432d0520b706452809fbb70f84bacc124ef3871d1e68fca5ffcdb56fc211e1bd3d83beb71a1823e289efae9123e7ac5656ebf381ecfd7e5aa6a514537884b984d43741dcd8c3fe09cd4ad7271d7712a3ad642b91969bee30a089231f814d349faa29777753971914bd938fa290f1df53cee9125d71423b061c44fcbdf52e40c80386352c9bab049a15d57f27fe79bb5e83127dc049cf95abfd6aa4786cbc0a7a2054c236bdde369179f6516500c98b56e8e777eefb3efb20a673881908432ac80cc4d7dbe3fcbb5fea2255a730330e2320382fc367a5ccc842d122dc97900099a3416cbf550a4eb9f7b061ab2ecfabeb3f8f7582ad518730df123ca920c0daa3900cf3332284432123decc595395ec2623ef1e95f8335ca2a66c1ede95e234bc3094c6c3982a363aae0d1069b6a81ef9ec68aef87fe627b528a1373328173fd04dcfa42cc800e67241ecde7a8f25f3657e46dde3ef1dffa336845631a5aab9872c1f6f7461101aa05bba2bfcd3ab00694761026744b663db0364abed735d8a76b5a63b8e17278c73b3b1c9574308941245f6497cbfedf1ec3f9a5c655d7d7afb252afe4ddb4f1c538a6e3cf744958c0135be724eb0b546704c28076cfcb9887de822f92d8f454178294c45aaf5529d3065ebb634f7b80ad53e0be26ddbc3d92f8fa255c3e74c1d14675f6c33409d8ed36ddf6c2b18e290efbc302999ab257061c3eb12169be5840125ab6f3cb6fa5e4365e5ed796944d9b8d853c5b406eb272788dde5c46f69c260b30541e29e3f4da0d6d840346b57b8ee7082251f3c6134c245329161e53db2700c9cb1a6d107b39f76bc29286312a2d7d095059451963c4692737784fa2445f9aeed93daec71ccf87aa33c7ddd5db61a7e1129fe1eb5584decf61"}], 0x1018, 0x404c000}, {0x0, 0x0, &(0x7f0000004c00)=[{&(0x7f0000004840)="c0c8d784997c32230adf7bf3716b169dc6021285f50bc94d9b97fccdcad7", 0x1e}, {&(0x7f0000004880)="935362aad25d5d5c9c9f413dbca985ecc0d5766fa806a7fca4b5dcd15fa3ac9461aa0c6ab0841dffc4210babd33e0d797225c420d72af937cf80a7f0bd7c8ff5b17af2820061a0", 0x47}, {&(0x7f0000004900)="ae60da53a53367c69594bda0d5597163a84adc4777fd1c26b183db1d1661494b3eb2dfb05f7151e576fa8914d7a3c220541d5d1e982aec8f0f24397a53eba11b09", 0x41}, {&(0x7f0000004980)="3dee85c6451cab53ec37c7e91923935e267abab539c79ff4bbcb643d6b6acee593a4dcc6025b1196fc846d6ec55cdcc39bd112e0d9d4176c498038864178e7ad0962daf34b69bfe8812b72dc1aef1e52de704fff1a106d5f0196207c4100742a467df61fbcf278bb23030512e438d48671485f74d47bec52db4e469db7708aa416944d27c972d146a8c3994dd68bfb93df5d942c30d9288a781fc3e0f1f867d3d436d5860970027c83fc1d881bb0712815e52825787304b4deeafd07268ac031c6e7b5c4d6daa6eb6df4fdfc032ed7a91e3ebb4649", 0xd5}, {&(0x7f0000004a80)="952970f3f76ca48b9ca1d122846ce62a98970f9d2cfcd8bd6c9d034adf9b57c889987408964ae192277018226c05e6eb087cfa9ddf0b87bd8895f4a104f63d15124578d943cf175fcdcb5ef114989e36880866f1c2a7c60e5445175454e1b7588d98ab799f1a7a06a690ac1f4e5dd75e22c50e208e64cadb741cafff47747fdd3c5c5272a6e5098b6259a9466bdca3f4912cc49e85", 0x95}, {&(0x7f0000004b40)="43aa416047f36ac75d6762a001bd75824d8e18e71f0743c927b6afee7fc9106b906b192cf62bf5993d335375f8f5e8f8e5da9ba957c35bb042d142f66d734c86939c0d5a1f86a22686627810fa87c70eef18f8ae1a843d525076e6cbd2fe66a478593a176fa838034809943f39c0d85c96de1331b6f1cace9139df4dc557105120fd4bdd8d34f0042a5db91a48ff3ae636dfc2c86ab7c650ff58181cf954c77088d9ea0aa56c67878000bea8fdb95e2cf0f05c8936c297136e", 0xb9}], 0x6, &(0x7f0000004c80)=[@op={0x18, 0x117, 0x3, 0x1}, @op={0x18, 0x117, 0x3, 0x1}, @iv={0xb0, 0x117, 0x2, 0x97, "d254a4a3a257267fa62b8d4fdef68efb50b9ff97fd381230931fbb4387d72a49c4e1170eb13a08c9f9e2f41cf3c86c1404456b9639432f061d0fe30e0ad05ebff5692f095377ec5d1f7044bb598a674ccbd628fcde801054606ff1bb36748d4a603df16766771ace6b5f6495b51476c4af0f6d651747c66934d9b5bac5235c47634e83f7e29d93038abcc94c9841c6298c755c48ce0d39"}], 0xe0, 0x4000000}, {0x0, 0x0, &(0x7f0000004e00)=[{&(0x7f0000005c80)="9e3c5ca37c38c681038a40efe1c7e1576aca3454b7ba2a0a440bde7b43ee551ad1462d2ae487f583554e8fb0740148e6d4cbba2850cb5f043b3b6e6ba43bfd603fa6b9858e583d9dc1e54b58d47b7efdc2d7d5637d6bac36a54e5d572c3cdf78a08cc9abb2302e0c767c5c1a4e4ca76f80a17c5fc4c383f00b28a7a67f2670594f4a7890608348b58bc12a72e0a74361c123966ec39ebe9be7eb5a08dbb18dde893bad9911fb650e237bac40596725e62bd7cf62cc327473c42572b11800f744754277e426a0007fbd4e929496535880cb563142eb377d6aa6d8159351d31722b6526ea3aa20137b45925eb9b4880c01d315509be03d45ea04bdb7609f2d8b2c771abe515bc76f35acd46dcffdce518693684cc833e751f3782f80da2d141a574f60b32d3a9256d46207678fbe15304c884493b3be81954fa3dcf1b537d67a58e05b92c81768c30d1ebf78c1bc1b22c8323e0d5efee892550d6a1d4ea0de10bd3682af31fa5b7fdeed97182be9651c2d3df79d1acbb07b0c9aaa6d6044ac06c1b77fdb6296a3864b5ac2eff8a6d83e8c53a412e918396d535af828b2f3e165ec5f7b82c35fb3fa48d1a9826d775f5aad02a8513cf6faed7584133162d82bbbca8ace2ff5e2725d01c3a21b00ad69b9f7db737a8a15bac0bcc34426bcde5b35708f7c913a0be41a728799a182a8551983b8f0959ea9c730f6894f63bc79aaf23b0b278d3df900ad89a55d481eac6a03b428559bcea6bb3760f0e249fa195d0b42edf9df67969e47b6882e43f79fdd2487bc66a879966f6c0c8f272bc4d72263b4436ae126b7483d5e9b3c1db7dfc34bd6a1af69cc1f65f5903824959062542fc0058fe28996c56996e04e59fe1682357a886bd767ec2466c3eacb135c1a3f80ed736b2281a2c134d15b2a5be1628c92db8d72a72b2df5a7c187c92d9433a3387e330d29a04155d72f39a651fd2e71de27c334dfff2e0beaa0962e01aeaa57d7fb3acb54046c240763f2e02e80af7f9a3b6b062ef4b96663692a6a489b271ec209425303da0d5e6f4b7deeb6066422bcf69425ea54d8ca7f015de1a27c343534a2be6589a7990fd653723690bc249e911e13da8a59706e7372ae180f601fa4ae79f69f39c26fa63bf4af8277e04e72ded7991f3bcb9e1409ee8f9ec195cab700dada2b09529aa9eab19eede1e04de32437c99406c263fc33421c539cbde56af0c2f15f01c4948f9ba165a6f4935fb3dac22852571671e8b8e1c56029871230da84db3b9812fa43f9c9cb515bd2714a8c0832061ee0023710d8ad4cdf93b7dce57337203abf94f103c680008beaaf556c6bbc406b2888d3abe49956f5f2db3e2f14aa7366d95bade9d738ae5aeb14c68415f3d7e61fb5e731a10e5f41722cb54152bb4768c4973fb1be2b1962ee7366ab15ad543d1b147bb1ca44a5a0aae7c90e1b7216e84907d31251abe04e1adb8ee3364a10a17fa187855b67a49869e8df691bec43d06dccce6f3714ca19bdc9d89b89ec712a4f7c2f62fe38d63f553b83e36d542a5947afb8869088302069eae33bb8ca70c9b31acaa0740102aa5a8a3f1f2f5f57d2ffc21186d9a064714928e91e19188c2ea780cbf08c84aea10ef16bf6c7bb08b9edb5b89e41100957c96536b897eb83b52f81664f88f0f71532f4d4ffcf73142cd5294096ac14cb04bf5703b38722a0ec5a2489823ceb1691ed3edf1671fa9eca02edf16a8b32f3550b7511b9b88082db53fdc271d3a24b0b742b94b4f780155989e0e9b88ccb6a40a3ffbf90cfc4637396ba3f00398859ae9692ace4de22d7f47e71ebaaeb7196a3ce1bc6f1d15051161655f273a87fad8ef871b38a1d029e150ea4a9ea773b952447e9831bd5211a3789d32f379aa15d6540db8ce2deddef4c37adec7be0edefdca477f77977933e7f32112fbae4712aac549f1f85f91e531763c2f11d12aa92bfcef91c5d3209e9b0426d368335f1b0c591edd76215b9524f9e4ebbbfa00ffb8eef80cf438fe7f7d35701bea62ddfd19e2180651192c2f4d64fd96c17ee75d2d76215180bc96479d690d25c83b2ffb088e92947c6a4e7246580e9c02ae7e2dfaaf0a018a13fc351b53beaa6d32feb5c7a683b4f9594f41994d596f78b3c24fe1e7cbfb79de6fe9ef65cc1f1c166d9abfdf87fe532c7efff220788be06813913435debfe17a11f8dff7c521cd0434266887e0c4bda9c44b62a72c43ec9ab57d998a3bef5c87df5e9becaf6d54bf8d67c0851fd2bd5d61c8adfcfc885ed62b4044cd8a686a28f321a0b60b25fb7c0868d99c46e8613ec8e7b888605238065c4c04488e38255d00ac0d3b9ea474c85b643fba7ae62374a76da567c11dd6d98ff0a6d8f6989614f86e3a1cbcdb1baf8764f51c972643d97ba403e830f884f2f5e003b8905bb1ddb028ca8f01979e018f01f2c62a134486a20367899a15835231771fab1d24860de3115cbe54bda4caf9683ee9b5c96e05d477c3f6af489e7e3384325df9ff3e68118a948163be57f7c333a301251be8710fa2b741f9d579ed631dd6a54fb326c5568db5a47910bfaf198ca00823c026f1140e990ed6ebca803369771fd8fc6035b1e1f09283435126cf7624dc33873ed48934f729fb3ae611662172007118efa9a3a9021990fc9be75bb629adf858b65c56626997c4b5d4a2856295cb58a5f57633da592de7879cbff1253b8b63b6da0586f1fce7762d5bf881adb8ae42ac9385965ad632aac49770ff6644b491b4b673f3ed21a4eb74991a7e752f3c9c1ecdbd0d6397350b3ca6cc1590b55dba88bb53c2f19f06ae2378c8aaaafed257eff675254d179e6cf248e27d3dc02e788355e193080b31aa10b7a5baa24726f6b9f26c7551bd440289bb7c7b5d765a544b085a975eef9218252d4f44e0b9c462b5c87eb8ad1bb477b200a9d86b9c7949cc2d75379cd5c4407bbf2dcd142bba19bde413362f5257a9342ec935fb1ffac4c255af774b3f59f1fc072376a15df0c7bd559ece789ab339075153bc2600fa0ed4456e99a5c97aa57154af889ffbf85dd9c2c855366a3ec991ba08c8edea7690065fb9c1a6817571d422523d9abf677ab04eb96412ee9c44cadce9505222d881000a054e4447d2caf9aab2b13423df1cfbedb9a5a4a32cdae6e0917358117614fd4d8694c5fc07873332f31819e93385dfe5bbffd02d2dbb22fbc43cea6579b88b649dd2cab4e8ac33dbc8d32416c6c71e984e4fdd24b81cb5b03f0aafceee4d3a936e108939ee4c978ca15c7f1606a61628c88361ddf129b626bc4137b6e3e7b74c7a821026a5da0b144210f806b9addfda0086945b3a81e546a895fbfde17387b3b88afa15f19047a2625d784919fa033a673a1bcc930859e3ca6719b2830cb02e654175ca8488b2e4b46182cb40bd25999bec5f4a329b55206e2c38981fc874dae6c6ee35312a8411ed724ffdb372f04c01abc23904c8c2fae1c85266f2eeb994d7fe5b4cd9d73aec430ad252a3e160a9adae405de8cf56d0f678dad22c37c358342c0f3c77e2af99760409136a1c1a737150ede3d39f13f3342853a54cfa6fe56cbdfcb7a35b071b7d71e588b8165c00faf8b90d5fad43b4cb3f27d8a87b8d928c3b7182103ca5d03ea439ec5afd1ae7a0c1b7be352ae363924201ccc08f5169fbba12688c14a27ed167401a1ec3de2c61c799be939812ed818fe1268f3cf3b17d1891d6055b398b015aed22e4ab75f4390a484e288b4bb42dcfc3c2ad64f26dc1c619ade4b835a938bb10ee1548d53ed7025d3a3246995ccd200e0c151cdaaad138c2238abe5a82ed7257e2a1cdbc9811062935ee65594d57d8924ac285c8afc34cdb1c3b00cff6718326433ea044485340890e75e1388ea8691bcc517826cf435377cb036a13f6698674d33005fa9bad6a864df609acb85253ee8fa3706bc43527e26365a7b16620ded7cb200d4929498f9c990c7e37f07bcf94ce19b810615f3f308bc84947f675ff6306f59469373820deae5cf72d475386c5aff1583a918c66bdbf45785dd68bbc2b4552a0751d69cbe74023bf1076abbbccb9d29a287c0f2c7c725ad773ebc0a596a7edadd541f14a4d59e2cb1b6895b129cb701f3ed404b244368ab017fd109cd5da68b38dbae133a40ad79aeee1fa6061f4c5cc855fb0f738f3043d5adb7dc536b7a9e4e6eb317593fca57ff50318e9703966398175231b2c1ad62617bb9ac021ac34d5c03fd9ef196e1494f49a5d8b35b94feeceb989c90b0d9728d5fce02b345c898711c2bba29daa66688a8222922e94abb4d0e1629b9fb7688a2b8a2aa251a1f293cc7d7ff750e72c28a2933e8345caa9d992111794cc7670476a7b0b4e2238bf7a51ce1faf9d195b1f2a4e7f50d18f3ec8fc9d53476d5680fc80935b6caa2fbc0bcbbbf27a5a914e696b920edf61784fc285d1cb56838a9b3aca8de8914474d02eab1364c829dc2f9e8c3f1ae88e3f236b30617d72cc5d96989b4a1ccf9d84c88b06efa5bc7089516e23ba3f22699669ca2d897503ac302f7a09b0493a8a70c8a3f8e7174aa5cc9b171adc5b6eebc5e4ad965403c34ba957df0af65df61c54d4dc2dba8f6b5954a621ea357b6e82cf01273391f5e38cfe264a3e2d456e795af2251da5bbe7034d442085931cee3759695431e30b7dc62858b0eb78047a74d052584ce1e78bd17e2cc1c78362bced2437cb438fba4b139bf7ec2f142654c6747afe21e55190a34257a8c602e4171a73b0e649a521877a90836f1583d4f42401853954a9f7e98563603da4e2f749ec4da027d000d3d4ddb038ffdf9c70d69a917588ebdf6167ef8ff8c32b465d7d5f5e2cf09ffa52a553e2156794f23e7dff83d38ca278925c111954b3effed0e26f9edcf32c9346537bb601bf44af665a77cc46ded797455c43cfef3fe0b57c15f45d1ae110080393cd00940f02db81e206d963c9ff979e411d3e64634790a777c18c7413a0227cc391d414be1f0ad25138fb829ee214c5a9fe6489053c289db7633f178774de9cf4da077aaf0bf36853a5e13e404f6368af7a6ba85db4046174b3db09e3cd6ffd13a6b7a4447d6243946921f9f7fb1d6ec3a92830ec867caf6c1e166339b7bbbe6cae4a6b2da5b63e2873a40ac2a8f547bf21d1090df148160388dfb6661ab74597a0fded8bb7df9558ebeee16c95400834cc09e8db930e1fb6a1447b179a03743dde5fc7534e26947fb60ec88eceafa9efceaf014fb614430ffd3e4cc02a05b00fa2a215e5573b5b8b022b9aa8bf2f0da0e4ba1f7d51437ed7d2b79edcb092edc8d6ef0ab9cdc946b08ac47bc60484c03d6fecaf204e2dc410213a44a42b6d0a5aaea3d70fad645044fc8a178af355eeed5ce34238684e2d4bf792b5b028572861da2dde92c5856318382bd3106e8c1932b1752e957b675244198a8ed4a44326a448c24aa344120a7dd988b7379e4ccb56684cd98e79f8c0d6637a593d89f100fbb1df727f19cbeb18b7003b80df6392b1a9f26af0c0c2a46571dcf83cb5b4631de78a81e475f5c29ad3eb37f34802331705cdbfb51f2a9c828d23aa24ec5f9b1bd0c3f607efc962e62b1c429cfcfba4277026089d56323dbef366f50233e0271480421e26d39b0877118267a17a21ef456312dfcad85e7537c8a2537ab992f3eb512db9ef09c22a097958407aac11fadd58cfb225f01b46a61eb7c50bb0672a142f855444128c9bb393f4c6e71ccdc92b73f367b6fe8e46cfd3d98d441e49a95aa35ea286d7dea8a5183e618d476da36c31b95dcf48319b6765531cf318722fd", 0x1000}, {&(0x7f0000006c80)="431610d59691d8d547e6c328443e2ad259233fb89cc0a4dac675a1b106d2a7d0ae8895c5a09d1e6093b525fab0f504ad4ae97b7835e47de87b05f6c687cc9249176e429164a46b02d47901e53dd798fe835f8e76a7675edcf05634515e9ceaa023d1752cef5a4772ba5425b13344401b07c849f2570fc9e55fa58bb584261eb53ae53845aa152fe399e078baf58f589400f6ee99e9146c7e88258bfb99033409e97aa6da76126ba0ee8cab1c7cfe549f95cfd1978cb5de46fda8e204fd93d2022f05c949b0cb4d74683bb6a56e8c320c50a58659c98e1e0b4e0fcdccd7bcd99e8cef24a0958eb7e823c632bea9e14b2dbf3da56b67b5ca71b122035368a4b3c632b988ef2262be1192177d8bf8fc2d275c25f352d7a9352217feb769d32f7b9acd3069430865621d8afbd48f87d506cb56326b698419c87a621927ab1752d3eddaa340adfbb8a2aeef63c78af19e1e7ba1cde467d3fd4d584c18ac6c29cafc4333c47f355340f2d156d523e9b8e26a8388c35e839226ffdb4b5a64346d513b670ca487fa7ec81381f255deb7d0244dcdc146107e9dc8e5b3aef426a9986e3777dc6052184727c8ecb20849bc6bf6df2f5adc236f02494d0eefb4855dda883eb51e916600fa3e790815cc62247b00f45bd79229a438e63a9e97a63d27ee72ad14798a560dcae1b59822f02ec12208e9d64c11d7781d20b4ee93bcf67074bcbcea1d5b5f4602fc39d21b7342b028f5a0d626336ef3500ddf0fcd99891fcdf151716985eeae8b16f6a494e4db49211366f70ab515d54146fd365e6e96693abdbf2f38d41cf5ba08ffe14ed9f45d6e33e8afe25dbcf62350e971df7786c5ee27f66192b2a27af8abd2953f1041f1ab539c5666727d1cf6f81ab947f78729bc60b2180c1d88fe21b18793907054142972fe8651881cbcdffcb18f87e91c6583626d643b04d8742308cbb92cb0d96480316cbbd66ad7fa80a504b843e75aa75956970201d00812c94abb07621465b9a11aa3457d6784bd0fc323d8ada182e64a9ab52e410fd8d4ba68cad37e0f6033b74707ae0b68234022375f3dbeadecc7c55712d250f2ec4d8b89c1a27b3fd9726fc85c51baeff2ecfbe042377afe783f44faf7cfdcef7d1ebe9e81474b833ef8acfffea117efdeb83daa2245b461b0b305fab1f2e62e55aa64f82e21fc902d404b86ea7e6cde3f84ab8479b103512c5c0222f91a696b4dde59e2627c885d027fc11c4f538785047a242eceea94f679142126ee6986cfb686825816a5b41c870efec32b094b78870d17e92beaadb71b6c66fc60d92a9135177481aaa1f34e740e04a2de0f321f6bab72706018d2d91d032e7d749f8056a02b3a73775a8d66d1138ac3da1fc65f549a39c9250662c649d30add82b0e9046202862345162e8b58520b98677775518e759c9f1b79c8a5b8f9a1eebee30f5e816b67ea241246086a58bda4876daf7462ee96d897b156009cb3fbad7d7be320efed2a74e85804dc23f43f3856ce620594b605f83c19ed88ab2c6a5a14782b6efca89af2424535e222d2bbefd1fd03bda773cb18e7c2def301a1fc077008dacb244281e557952546025552a2f8204ed75e9dec38f22f15127dcf70128db8d8753490901bd363cf6f3f9b27871df81f9ba5f7cd7f8e74134a85c1ccbef264dd686e421517cb79c3a1158a9f4fac4237e68859685ef28a1979136235fb0470c1d59a645745059f2813cac857cbd3db12fe729623552cbf5114908e733ddce63d4e5e754bc15db576b7db72450964d43806c141fe38b2cc004af1dcb579fca27c4dc6b17d68594829a209392b0eaa954ee65801116ba367faa1b0c4ae54a7883eb7872b3952dff9c5bcba3ece9ff3fbe267deffb04e4d95fa86c36e8ef56798de3bd94e1ecfccbb284ebc627372081fe14fae1fe0f93001f222a9e82acec221f0ef0b32fc8cbb6201b766706958592af7cdb71f22c41dd1939d8547504972d50fcbf51d3b6e7f3db2530a29cf6fb4a5a7b62e713ab26087c47f224595d9790427744609effca18b05408ad047b63c5e421dec1ac78fc9cfcbf0fe17ac996915ad84c5e13bcada7be81bf8d57e844e8bf6c64fbf9225309996a228ebee453559abc984f5ef3afae0dbd8238f519f46275ec55b765f76ab0a6d9223426fa5e28e666d1e35c2a9d9a61b7558f67e93882ec2e7f93921de802b2146a538b7a0ae2aebee11927488c530ceccb43b8888746de89352c0f55f5842c00f90c41b514fddff456a6423f6657b2152d9ce06f56cdce96db076ee3a489561e1485e513256177d62ba1609040c77095d7a678cb13b719652dea20a933d12639d6b611ee7d33c4be7bae2940b1d21aeb791c57fa718d5daa481fb1578876dc65e807584e4e7e014a45fc8647cf79669e752de3567d67dc2cee81122d6be7f7e07e286d8b7bf172a665f6eff1e24ee15ce4650baa5b9b8ed0c22b1192efaf08d65eb69e70c3c27012a56e0da6589ff09d63d50720c3d54f550c052747daf0e2a619aef4d4403248fcd07718d2bf06fac821cc3911ce81975eec68693b534f21f2fd78df183034a045d4b507e277ea93eec0ffc048a7bc0285e8535e7a14717455e5a2878646fb2bc46089d815214cffa9c81d5ce802588e3c086689cd6c366c5132f8e480298e82ab01f2edf488272376b1b736dfbc7a4fed564cb81ad6c0718b92228bc34397a281d7e18f4ea8d03ce1ddb3a0111e2b6a329e66b31e1a117a771251fbae63ff710b4f3a0cae234b0bc82c0f4e78d03c82fc2c340c2947deab813d07cacf789c76dcd01e49edd39bf37e4774bdf32e2b7e4e18681321cefb88c9b15979a2d7ac983e3295d1f996bac44988fbbd89bb2e4de2f9c90933aed4e7f2c1bd329ec1164c83c5ab04e1609b44eb532e1a68aa346417ce7ce0057fabc60aa9ac95cfb32959d3ad002516e311c0e17045612400efa10531dde76afc1144c67d7fe4239e4268aad53746fe28123f069edacc8be8bce7717bd68142d5c82611ddef17bb00f8cf601f64912767cdedb1e09ae21d0315ad023e9410c5f73a4b9dda3d2141b25b3930f68d95447dd3ad3cdf4aab131a1220d7add73774d2d4e125ee2f78b51eee2fb0d8e0a61965cc36a80402b1427c32965c02be2da1377959204e111263da5812bead8cd45cf0b8748327b5c2de8681f194c6f0c77dcef9c5da3469633b36e2e9e9ec41a8d4f3e9e6f1d60943eb302556e05de3fd6a706310ac40d139fbf7bc6893ef8000d8ad05635d10a503a0069d333b14d7c870d9688bc313c0a9965a1cdaca0eae6ec2546108aa7fb546e16b31fa2c0ece9304b5fbe82f981eb634f670f246634f744ffaecaea8ab74a0a7f8dda774bffc8f21ef0ac848a583af1cbe21b9b7e894b93059c09047de1700d63ca847bbd888afe47689b0eb3acfa31251b36b81b7467688a09b290c8daa85d8fdc4e121944507a9c5f23992ef66d0e21e35d6915185fc58c1c27ea9434b078b1d9d2a382762be024fc22d6990f882eac10edd984ee0a2a5c85af00ad784e0cede2a35790b65a7ab2d624416a5b37906762913c986464f705e0beba39dd47a8156a12578509f5fb2f66e50f953d1b7b7eaa120a61a5a9210b57c3b4cbb8f5a0a388ab894c7eba958b75886a054db748c4c7d7a03595ec5ac1cdcd6d347c1b3c7c1c85af33221d03dcddd2ac9f3e9682e9df3602cb1e0b839cf8a35712b559871533453fc0f7c732314d2df6065de0361be45616d2fb4edffcd33f61b66bfec594122498824d89a28ec1fdb39e473e6b40f968bf4bbfdaf92f6376b31209be3a3f1f2b8ad5faea17299e561967c61ce142b666fb668db332fcbf4f8c86dbbe2c9aef5a75267e2d8d220e7e45c986af61873dcfd499cb89a15c03d840df4731596007e011c2c5fb6a9d003e8ca4808c6932a175b2bf690aa258eda9f136140c683289700c1a0bdbb22453f21d9c30689896f76620e68f6b8c58f2d16650c051423ea710d335437fbc99a1e9c796dcdf334524b495611ac489a013bebbe5627ba0b1e3e12310410dc84a5d2a70b3caafcfabaaba24734daf6c417a4def0d347d49843e4e3ce10bf4b28acd03686526541902dbbd8437c00551038883ff015798e1a0ee1807644f835b75848f709dc105646661bfc4836939afe89e4ca4aff3b5541e5f9b28ab35f02658eee67034fe6d0ea24f4492becd3eeac5f489e0d1b2d09c6e1c2eb1fb2f80d122a844d9e5c1cc8a0fb68ad9fe7112322a2b12fc263994132c0bb88d5c24cad00af8118becda78577496c0fc852a2e8574509006674cdd1c4680b38ef610e4a7017c86d8e8a3232d4c90f286257ad167ce4b3e87524da91972699ae896117090897ac86a76d9effe522c8194c23f6edc4248a29640e4e9cd15e97c444cbc95765fc55428161df3e20242207a87753912a80918644efb72c539be02402e6948dfd3278ae32c28b0fbf49748afe18a2c000a4d82ce215bc7f7a1ad640902f25590c78cf2894db89f5ca9788f4e9c50a9442942f1c3ecb8fe8180ec10f2e6a06c168dba4bece345759e3455626746b7b88c1efc46f9258d858242e88f048a85ff6f37f3ea8f3db1e8ec81c53200ad4a0e7ca46109fa45e487609c334e0c268813a0f327dfb14364a5d40c94cb044c371dfa3481ac2495ab14b877a02095f670415824ca6af8cd832e1fd3ac7c1a1caa42e06751090cb3a6cfbe1da60519045207f4754d90070e01ae3057423d34b420b3cd9461bba8b2a7324f06333e445abdaa0cdd80d54716cc4e1c6e8a9b3cae575fed6309b78a3d2285256c734f3a8562275d669b27b495b69409bc9fdee8cf9af5a1d1a48edba37e6befa3046a7ef36146c6a2f31518ab008e0d1c5543c624c67cf6a38547d7ecc85850408b9ba090a919f4a253b3a58d9c88aa2b3b11843f960dae91dbaa5820608efadd829144733b6a425a2264ab1f7aa6885eff6cc6c58fd603f7bbf2ece7d69389db99cc974f30a6830133d1a6bc4a48eec15bf7d2766037fc3f3fcc8ea4f9b9c334a8bd76393c5d76e81bf1e02202b9b708bb32de63345a074cadaa3baa9e1f9c3c015e77499d7c6488ab2beece7f31de74a77f048a1685311f884ab741a215db3b10cf9a587ef0bdc5a5eda8822affa7e6c72fc46e88431e5c6e873c94f4e9949c2219924c92fe606ae9e19f2e6150d1b55ff974b457caf6a9a42379013ed00672578875f9402a7e1b9a0f05b4937dbe1b24982dbd613d890e448a55fd95cda3f699998f83b93cdcbcf8ae5df4dfab5b9360052215c9bcc2febab701a0ff62c1682d7b3fed8875e0d61bbed780d15c8391263273ce9f3c7490dcf18606e7724835c3086eb91b209e8fe9443fa487333c016be869e8f975cfc6f3f6be5fbaccc9dad24dde72e377c5b78a2cafd41a97ee199710042ef2c87c722f14176ea785fb9f97c0bd6363a492fb373de8ed938aa0626f020bc1571a21c591a670d255dfcd11c3b4a68c9eb77858ff398d1ce676217007141cd17679253e72c9825397e37647075d2534b0e068e8c66754bfc213bc41183d353c8e7265b273f69446fe1f6e78bb78fdc7bb4b40433aac317ab06e2748b08b3c95ac0dbee0d642c1ade7e384f15a68f2eb676d1b09e38ca1224cb13c2e91d548e6e2a7a0e57383ec807a0517cee39a8b853c26225d256f4cb835091e6046164bd9e329478565f76156ecc0ff87fca425d41565de68a5f59d8794f48f9d798a2d353512c424d6c13108fdd773c79d0b781301ee18227a99845be765f704eb56", 0x1000}, {&(0x7f0000004d80)="9108a062e24c0f63147570abdb833f9835229ed7da3492152fe3fa0b5e864c3a669c4dee2325be801d663a2e06523c68db1b54973862bb1b43a02909952b06c49af6b8248ede86d6aaabca2d7d4e582fec62403052ec850d4eeb327ae6c05a8fb82d16920111f2df5f50be9f9086d3b8f3fb11dd61c51477dc9416f8bf02144c", 0x80}], 0x3, &(0x7f0000004e40)=[@iv={0x108, 0x117, 0x2, 0xed, "d38f8caa38184ae4bb866ad364cd7ec4ecac8eca4b201ed77b2f86ca89912020950be14e5f472322dae5431189c713b14dd14cbb77c7574fe25032cfc6819e718b883b7b75e8bb0f8b046244e1946aabf5d3aadc2420296287129e7423f3fbf3325438ea2837090de67211cebcd234b610b4586bd3ccc02141cfc3ecd2da7c3a84bda9cd25802b4415ea6168245e39631e6be6a59ba31ea9237c0dc5ff8f5de11b045d65a7037a17a16cd9f586922dc7acf1bc1fd12398ee4bb237a10f4b55981d87e41e7377e7080e7d895281e4d362f91ec4c342da32ab4021ab5f915548c0930a08ed82f2374a23fe21b364"}, @op={0x18, 0x117, 0x3, 0x1}, @assoc={0x18, 0x117, 0x4, 0xfffffffd}], 0x138, 0x4000008}, {0x0, 0x0, &(0x7f0000005240)=[{&(0x7f0000007c80)="4ca0cbc5466a021b0ed44ab1fcb14d341b5a34aca8452e50d25bcf22d6d1a84653885c976e90f1e6e0e3a9c822669c7d60b14d4bf585805c59af6d3fc476efc51f0b2d91760973b21922f9b0842c012551bc99c2421b23561792d8589b368d1315caf52ea858af941d3fe491667cfe7ff0ef8e12f7423fbeb43016cc264e6e58455e68f121114227346f8ddc09133bcc05283e31e6ce1bc08f50de6a6ea240c9f3bf895a766bf52de4761094fbdf3131b04cb430fdcc47c336c5846484bcdd9babf039ca5a1ba8d0b180827fe90ac9045c88a0862ab46431e3c60d94356f1d5d724476444a376c8adad6c3ee03f9e0788385899b6fa18879396ec2e3bbf0c513e8bab9082e50c51909505e946ff8edf6a5113bc17e6bf8e1a1e47ff57b2a2f9b8c4a4796ecd9981ad5dfa9bb5450b769a0e24089ad009202d8bf42f987c2f7225019a4485cd2e74570d18c0155bc36f2d96bf8985a04fb6186a9aea3e89c76e5e1d7276d27109ae742c421a2a4bc55658e4450bfb83978d02a292032f0ada0215ad533a09cbe55f10b857dd143eea6bd3a969137b8a2afb3b2bdaa0952f2cf4b58098a83127b39c6b563fabb8f8ecca351055dfd7e6966e86039aefabc33e83e18225e82edcf00fed4ea1def38517cffdc7ab3259d3498e2546e294fa56a360e1947ea08aa88edc7504eacb330d88dc780ee523b44da954ea5df3329950773f81af05e7f2101cf50252a2be1b5de88b45f9af15b187aa65e3791347d07e6e0f504239a8e43633e5bdac0a84237555cdbdcaef6a728422159ade0e8973d4787018c9f03f4ee017b80a8564004b8486e49013df35dd513211b6064177b58205b8293a4303e025d1f0a35d05aaa5436d092401583cfb255606c4211e6f7744c02a8edebdcbe50a91f58b3d4f143998d2b00d2a72128251c1bafa179c4414832b3c6c9672707412c55c8d5ab6e2e3553182008f1cccd9f35d151598ebb96f11ca2e16d463405552efd842744cb1477bbb3d82ea63555c52c8eb42febb6443ea6a9c25d456cd536df5ccc51e14a3263e0363b365878d0d7c6487d172eb91a5570a180860a48c8a739077ac4b5732e08c6382444ce215b4b69c50615a98faf9e937934a498654b77c04a57deba101e87d0eb3c7ea4a9c442a7952221e854ddb155b3ce1f52f340de94ccfc2f71cfeb901ec00b87163a7296767bb5ef6a39fcf57e5b5603ba4621ae9b2d2a6bbff8867346160193cf66afa9fb3ec1eeb4c6d6d584d15ddd88ac70a94a6f364ceccd3b472cb6a38a2e748abfe2acf818f6cd780690b07b1a76f07dfb0c628c802d7a19fe7e258c31c200810c3d8ddb71834f56cbab39f9f72f5ef3dd391d14157a315f5346813026be06ec3fcefe28ad4877bcc217dc0dae6246d2141a1e9a3aacbcf932e564e358ee7d2577e6920e89fceb9c25322c8e44c0db8b4c108dfee112e948a92113f286e307c7320d2167b373b03062f0055046a3510e2b4c679af03299c7001a9256af57be02800fa39fc6070ab588bddf058468ad544d306ee570cd9b84b63f1c34bc728e371a74e272dc629a64f30416dafce5955b7465c6a0df1b4cefda7e32baf1f5155e9c55f965bca03cff6d533c1a242e5da912e856198297f7fe2c325fae0cd330d5b2cba31d1734830a83cf2e42428e23628b3b63544e59fb86c074d9b1b90dd4c5911c28c3f097a9d33a5a01b831620003de2c47c3abcc2ea48d9b3e7a0f7bff11483ea66026fb71a5ac55ef6f1b037c56955f27b2fc9f8c1049ed18aeb5cf8a6cef2fc9bf13feed052ce6ac2d4b80f6bf51cdae67baf72b20f03e533480108d7ac11e61e68c9ca8db654b038139bae95df5309249ab7f8827eb15c934f77cf8163bb9019685a677d86eaadcf09c1c2df5e2dcdbf481113a6d96bff8155dfc39841a95ae86142ae8f1f558c0a23e053b357121eea7285091a32b2c285ba91d41e1c53a6bdc4a6fcaeb8354b2d65356408d1fe7add81fccbec3d05890bbbfcef39f50a70b7701d675a953df007e01ff130e697fc8d38ee620142943dbb5597c01b9d906f052bf8855f0041ccbdd6c002f8361bf479482e81e1c19505b94879711aafe1ded367772a203285ff91b02bceca85cb775f24cf337b9d162f8799d908d3c3841573fa8ca1892a1f9693225a54d867851bb49ed8d18ec835173ddef4b0103fcdfa820eb0a4ef6cc2a374b63df036bb4de327661bdfcfb22e77c2a3f3fa7292cf4584092233b4a3fc82278963d4e0c3059543b4242a0aabed28d35df1941a4db74eb65f16cae60a0c24e3dd539888695a7938f99fd739213084a468646cd6d439389a21bd880f393317c6ffefa2c6c2e2cd55c0ee337cb0515ab622e1158c19cc28673df29824781b05e14236c2240d0bf8c7c459d5b2be97013bdf18f372e824cf1a435294b13183f1dd79aeed8751fa0a7d06e0c78306e37d6bcd2815ecbd85b250334bee88b70566b80c3e441249355366fcd6f70ce35870d05ee32919e7400d48c26cffdba88ffa53b07c1f319789b95a698a450d9077bab2240fa6bdfa3a114ace196615247840aa540f04ee02f44a254ec19afdc920aa447c663f14ba95a3f5872d362a9511156982dbc953c8853a654df1e3385385777d74821a93d3cfa18673a3047cd906a00d8b5d6d893718adca65d1c8d9f9ef3f7ce4d63b449a4f68e202c42d9d8de168a840170a3c29dd5dd3521a3006c085d6573ebf71067262fc35a0386bca575e206477e5b073a9d78b708851a9719067ac5bc115df03dd67e46a0a62e8e5183a91352c32cc4a85071f419f4c7cc763f419bc7f5dcd09db08b7cf14038f7fb84210d2a12130e7db3853344d0f382b9e400112a4eb68491cb9ecd38464219c4ad36060e050241fb1e3cd3637346e326fd64b4feb7087acd2069ee552eee9455f5d25817236f7dcc9fea359694ffb37aa3a219913c529b4cfd1c7b1ac68e1dd73b8c25a703b20b4d031e4c8b4ba98c340c4a8bf429baf8cd7edf924f1ff8b0d57fa28a9c7d95a8228ff84405f5531dcb1e0448f024b7d9d5721aaac7b1ac1359c9d63e5bae51cf6928d66defdd283efbbf5a41c9c3a899cfb6f6156f585dee40342bd5fcb13a0441c07b6b069da3307a780318e560bf00e3af89dda874337a37dc54b79bf599e386e4690ddf9e19c1a25acef09a989099073cd9130ff275ba60a8bcb099c12248feca85e78bd57704fc1030833b3c6bd4ab4039f8136e533b04694215349670f65551581a13d5a26e0234c1f4f79d3c31c4480fe0098cbe5afc661b3c4fed907e996dc73a5c61a61e028ccc8f2b99813b5528c4ee6a557d8b23d00e23dc767e762fa2348c4989365d535f489114da95517f4dd7ab725b49641534b18593b51cd3716913862bf36fd1b7502ac5ca96e8754bc4a3a4269c56a6645029a988e172980819146e97e16088a6646d1d05be9f9d79c3672edfc240d28f0d16f26e8558e521235c4bf58d8ce32698ac031d1e5b77403cd9e2a637e471c1e578dccebb3853726e8c2f3bc501d2192bd32305ffaae3877c728a3a2178f29fb9e452f20a3c5237dd190e34b2f3e2c3bd0e309c91d2896f2e8de0ea11911c7adf1677019743a67223f7417e2e056c417e2eff269287aecdd397f9c14164f3f2ab4713b72462e581353c29d5d05935993d01e2312bbd2e2f24e9b97a9558a1a3239af6a13ea8ffeeda5c09be5ed7377024e1387193fb1042ce6f6cd38ba3fa0a89665797171c906fbc213a1eb93f706279e9b03d9c2fe9f3da5869b41a87c5c78aa841e83914b51535fa3a39254fd14a0769bae6db7cbf9625952e3426ff38c432891521b8caeda0980ff37a894f75deda990a1c287e266301b6ac6aed3e6162c9224880b3c67e063880dfd8f954af90c52d690086940ec58c1854e0e8d85b9cef6a2fd156ce130c5eea6267635f73b2f9faad159ba40c092df3886692a13417ef930ae2d695134aea9a937d8e410c957df7fcd8c0972656fb019fb6b1a83d489536caf4c0750d9237e17798db02064dc09ce2ac48bd6a215a494c1c4a8bb9dd56ed9ba45474bcf8b50c758ef19955b2d3a6d6320d13533f28d36f781299dba0b33916cfe28a50619be9824276faa52ecc7d5ef4d448e0d87f9c2ac7dd133bd77eb5389dce1ac957bb14b4ca5b388a529319d4db6a7569ddda24648fb6b173a2582a46316452ac61cab2deb86a8352970cc39eefb6094a3a7843824ec4ed64a9905f2cc038fcef907787e19fd9bef5b18d8eeddf69291142aadda67c2ea0b900b96eb64a731e9c27dafab03641717ebf090d6fee71f447d7748dc925a835ef81c8ec68336e80c62f6678e815f275c8ccc3f3b1283e4ed55024cd40ca62ff02f10e5ac6b711140994e22b3d64b884c1691cb93bc99fa3fd8a05c11b9810cda1c031128eece89bf5483b697eff102c4f77eb878a2c76dde4eb8595f786200e1434c94ba76d76b48576366a519d7d7edb5acfb260a857905724f994aac4a669ec5176e0358257bac91615a971fb2012ad79d961e7645bfc4aee41748f4a22fddcadec8abb138369634439e3e3c367ed229784919cd92c74947ad137866a5b658edcfe635bb50b4779065cb3623b8d3450dd92409dc8ef22f0cae8e03cfb279f85805acf7627f31e3e473ce516cccb00eac16a745591c733b4c2bd092fd6357935b0491f2e28b95e94dfe7670753741d14c91ad30891b194e3c8e512e2d65354d891ee0b7464efbc550ed004a63a1f128fc81f775a4c4b1fd995ad360c253c1320b1d34c8b9297b2af5b7468287ff8847f9bbadf282991f5a4223bc3d010a044433f0af66960c28a28cf85191385c82da3912cc62468049d0a8177d6200346bbc7719099a7201e64238e06712c85c6adb7aced588d7d40c1db246061c5bc86daef036774053bd918f4fffab60380e478b0af54c3bdc926d6a98b17a95bf55d438ed5d09b617e647dd62f64537afc6401945a5a5c582ffeec1ff9a3867a3f4fc90bce92cbdd6be3b979abdf9caa9834b1f119992eaae09c5dfb186ba175cd7e3ea243ebb0a5b9e473957f749bc2995987114274e80ee53204990d2369b6cb41d8c7a4294793418ea1255b50c7b2bc944905651fdcebdd8e00ffffec7780d4dad8124937d6c1ce6ae847b4905e6c4d97bbd32abe433afc87df8a23e83e8dd1b4e7054788231297aa57240a9b23d9c38668100732bb1fa7ae97da4c9dc6e37585ca1fb1ebf27d357c0e32edacb43ad9880825e42acc6346c31bef284f22093c53686de9bf7c4575183509168ede60d5e8d651db8af1a17b19bd037a95f23c248ff8ad7afbfc30337dac4de0f56947d974389e5d883044b80ebba1cfc40e17206651520b3b90dfafe9c70335a214819b51fc42e42f793039672da7ba1f6ee6cd95105fdea355412d4affa134fc4f65dcbba0d627d4878ad22e78ee66b7fa051dffe7640fa5d09d1a2cac0af5c085a5430091b659e5c71083637e6c0d3b8cc3da5c1ac960349603dffbcd483ce402434bd5c5997cbd8c6888011fab04bdbf34a2c6cc6c1d49ed6bf35b1cbe718bef9202eec8b1201a90b5c32c8e1f0a1d7ef92c175ef9ff12a6200e6403a6ab514300028715933f28463fc95556a4e024b8da77eda8b9f930b55ffaf97377649178a102c0d20428c8e8a5d66bbd5845da5eef01192c155cd1f4516306418c8993a8780451328ada401f6535d1882dc5b6166c8d953ca92ff72f51621d324aa741f094dece1efd4063c80d1f71c5ec639668d0087541ffa95934b509", 0x1000}, {&(0x7f0000004f80)="a27022efca4646c0b4a1611e6804f38bb9de5844c2204975a6c23606ef4f04a852ba31f7691e9342be6cfeff9c", 0x2d}, {&(0x7f0000004fc0)="ae7c3bc4f40d0c5f4825c332f33cdd294baae624f78b4e34638edf769f3cf6e6b4dd59cfc18a7808852d63464cfc979f1fa07f598e2ad80fe10558d29ef8f529e1da41bbd3dab56dda68aed1436551e96fd1d03a8e761dc419dbfe83af0f63039222d118dbdb9a5eca15ea15a9b19c0a343e91057133c0d3783c6bb2e6da83975fa21e4955eac61a8c4bdb7f4f2d8fc8f368b322756929334718f09c00fbe18e294f761efd", 0xa5}, {&(0x7f0000005080)="76cdbd55e2430375d979f1f0f596ba20d3fb5a75157e0d5a99c94b2de7473c07bdbca93e26557bdc7b42f73db9dfc61d8a9afbb0cec655d6c294ad467ec3f3c0bbb1d2949b1451d481222532b9fabea7fdd706ee15aaf1d46c61d1525dbb719e7e4e4baee8640a8148c3fed616f050f08fc255fc9a20e49749ba90b3ba08eb2cd7b8d00cc19fde2be20342368b62609fdd0cb73574c4ebfd571aa59d58768ee7b5b1a372fe4a03eae9678085bc65910f52a2783135c2d15e24fc94aa97a9407df8afdedd7e7c579fe895661deda075161c577d7316c95ca1c8f47233cc3e9fdd1d65a85f", 0xe4}, {&(0x7f0000005180)="f4234f3520381e9be7351d019fabde4d9ae67262e039d3632846ee38ff0004a28cdf6d4f79a61362b00dbc0c86aa38e8e399925469dd9e4fb351895b556f89f571eb116a4deeba4b64d0a6617fa9e4adc0b6da31f857e6f423ee7eec5fb71053de696b0cb198b14dcd00db1608652341168a988fb3c7f5a3ed9e2d29680c47f498155df6b0217bb8b102465c6a1fae506e17b094074b58f0fa", 0x99}], 0x5, 0x0, 0x0, 0x200000c4}, {0x0, 0x0, &(0x7f00000054c0)=[{&(0x7f00000052c0)="0556d4d807611313a785796d3eb788120fbcecb94d54b3b03f794aefb166242c7592c8c4d93415d46b6d12ec1c080b218c36e5dda30cd8267e1975620b7d426528a765a8fa437f9d", 0x48}, {&(0x7f0000005340)="697c41474d015ea44d80e275384657032c7bcf640b08fbed8b764fe3ef4504fdc7bd6bb6adf15657977c034471e70e640734aeb1852b5540712dc5f8d1b63847431c222b98c69c9d71c8f0bd3c4230f32054140a5bc037b4d4d8b0ab25b1ad39aca1296afca662811a3c3bdf1aed5b989e7d2de08a0a810995c39bbe30f84c69991b907c888449821677e9f9a33ee719a28e62608adef75d83766dab8583c09a8accb00fee4de612757a954674", 0xad}, {&(0x7f0000005400)="cd101bb5b3271eedc18551a1cd724d6288adba2f6d01ff157cd5df67022851764dd9a089db70ca2d8b379c9f6199a384cb377464fb4366e3b9549c8a116f8c51aa35d699bfdc952211b55bdce5e71ef8a1440fc078f08a08048507fbb5128e5babc097e7e5a677ec8c023286e836905aed15e250e9c76d8dfddd18c7142407d65becb90060449a96d4902ce222f0646603efca15ce72cbca568e85426be88095e80b", 0xa2}], 0x3, &(0x7f0000005500)=[@iv={0x50, 0x117, 0x2, 0x3c, "7b8f2e991d4a45c9e843a3492b93ded39b0af5b14bd18ffbd393138c4a367ac5a9602a73de1bf1fd247eea7d5a301ee4b9a380a745da31344d33ade4"}, @assoc={0x18, 0x117, 0x4, 0x5}], 0x68, 0x20000040}], 0x7, 0x4004054) connect$inet6(r3, &(0x7f0000000040)={0xa, 0x4e20, 0x0, @dev={0xfe, 0x80, '\x00', 0xf}}, 0x1c) (async) sendmmsg$inet6(r3, &(0x7f0000005a00)=[{{0x0, 0x0, 0x0}}, {{0x0, 0x0, 0x0, 0x0, &(0x7f0000005bc0)=ANY=[@ANYBLOB="5000000000000000290000003600000000060000000000000401000730000000000a0000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000020000000000000002900000037000000000000000000000004010000000000001400000000000000290000000000000160000000000000002900000037"], 0x100}}], 0x2, 0x0) [ 2803.344789][ T2360] bond1261: (slave bridge1182): Enslaving as an active interface with an up link [ 2803.356664][ T2363] netlink: 'syz-executor.1': attribute type 1 has an invalid length. 04:12:10 executing program 3: r0 = socket$netlink(0x10, 0x3, 0x0) sendmsg$netlink(r0, &(0x7f00000047c0)={0x0, 0x0, &(0x7f0000000000)=[{&(0x7f0000000100)=ANY=[@ANYBLOB="18000000560001060000006400db730007"], 0x18}], 0x1}, 0x0) 04:12:10 executing program 2: r0 = socket$netlink(0x10, 0x3, 0x0) r1 = socket$netlink(0x10, 0x3, 0x0) r2 = socket(0x10, 0x803, 0x0) sendmsg$NL80211_CMD_CRIT_PROTOCOL_START(r2, &(0x7f0000000580)={0x0, 0x0, &(0x7f0000000540)={0x0, 0x1c}}, 0x0) getsockname$packet(r2, &(0x7f0000000600)={0x11, 0x0, 0x0, 0x1, 0x0, 0x6, @broadcast}, &(0x7f0000000100)=0x14) sendmsg$nl_route(r1, &(0x7f0000000040)={0x0, 0x0, &(0x7f0000000000)={&(0x7f00000000c0)=ANY=[@ANYBLOB="3c00000010008506eb9afc4cd8d06e754a0081c5", @ANYRES32=r3, @ANYBLOB="2377f292252155b21c0012000c000100626f6e64000000000c0002000800010001"], 0x3c}}, 0x0) sendmsg$nl_route(r0, &(0x7f0000000240)={0x0, 0x0, &(0x7f0000000680)={&(0x7f0000000400)=@newlink={0x3c, 0x10, 0xffffff1f, 0x0, 0x253c1f, {}, [@IFLA_LINKINFO={0x14, 0x12, 0x0, 0x1, @bridge={{0xb}, {0x4}}}, @IFLA_MASTER={0x8, 0xa, r3}]}, 0x3c}}, 0x0) 04:12:11 executing program 0: r0 = socket$inet6(0xa, 0x6, 0x0) sendmmsg$inet6(r0, &(0x7f0000002a80)=[{{0x0, 0x0, &(0x7f0000000340)=[{&(0x7f0000002f80)="3d59d3bf0c21ecb00e86fdcbfc4b8a3d59a5675e7708b00af43a76aa5f084f9b049eaaa3f58988747f308fe5372d232e07a1954656f69fdbd0db9b8db353d813cc9fa3e3b1d0bf51adf59515c4c19a1196dfc83701040cd20f6ecedf66d51a7c972f71b0c1e55f5b6601bbd1dfb218c41c28eeb426fb8d0529d0e1dc5cd7a62cfdb4d167892e46a7e06d9b55ff9ce3a62dc3c2be09b96e5889be264065c527c8989145687fb5d98083181be3cd4ab89d396c344bb55e2865c29602973e367b16f315567de5f7746ddbd6a788782c86c6481b11328bffce20d77a854ca7f908bc02374415ab56fc0ff1144a1de83402ba61eabe4f082b985944483272e550fc4d8dcf6ca7e59c027100e35b38fdf1d837c07e550cbbfcfda412549737bbedf62f4e74596c3c8b6a80048fbdde3f8966454c098f6ec4360a028fc13d43c6ff8c292122b02b11d78a687a3cbb4233cc3779dca2a4b6850395a9a4a8f3a3de6ee4c83f699e", 0x163}, {&(0x7f0000000240)="cf4d32a806c5dde1d7cdeafa6297e40e3c7d0016cb364d1996ddf21fa8106027bd14f68f05791694e509b7510223b9b41ac0a8d2d36f7add06c9fda739d1585d9a372e0164aa44a75e9f356e7335ac11dc5c24b509d633f5dd589a8722062c44f96ee1ac04c2b7cdffababcb67a0f87826e2473d5e7775a9", 0x78}, {&(0x7f00000002c0)="f197136ff72ea7d5781b8d690a546c1c3e1d97d5172ee6be3bab68461dead12e4f1e1817c9d2ccacbf215c1828b8d45772339b51214ef5dac93bc056819973e09fc60c2eb8c780839001770f6ecc33cbc672440ce2", 0x55}, {&(0x7f00000001c0)="887df9d949d3", 0x6}], 0x4, &(0x7f0000000380)=[@dontfrag={{0x14, 0x29, 0x3e, 0xc6}}, @hopopts={{0x158, 0x29, 0x36, {0x2f, 0x27, '\x00', [@padn={0x1, 0x2, [0x0, 0x0]}, @enc_lim={0x4, 0x1, 0x10}, @calipso={0x7, 0x48, {0x3, 0x10, 0xfe, 0x7, [0xc5c, 0x616, 0x80000001, 0x1000, 0x82c0, 0x3, 0xcc, 0x7]}}, @generic={0x9, 0xec, "c09f950ede99d3786881aeceac937f1fa5dc5defb8283eba3160d81d2d96ea222c0c88c0c7609eaf399113a60a4a1ce469d5f529fe54e28d00a8fea66c3dc8d8ae4fb775589b176a72149bd1486978b742a1a6996cc32bf260b67533d069cb74cd4911d14bc28eee0ed7af07f1c1df5bb6c49d49a6af0afad0ee28fead6e0fcbf64e0377ecae0148b2dd075ffa644a44a0f17401c4ce8c421ea19856d7a5cd38d79981b2dbc1263a200514294bf994114dc81487e7920d5c61c0fd0a0b54898637b3297cf00f3e69863521bab3d7314b6dc1f75bdcd872c4cda5d8cf84c12955f1142b0e4c1c9e7e633961cd"}]}}}, @hoplimit={{0x14, 0x29, 0x34, 0xf336}}, @hopopts_2292={{0x68, 0x29, 0x36, {0x33, 0x9, '\x00', [@ra={0x5, 0x2, 0x9c}, @hao={0xc9, 0x10, @remote}, @calipso={0x7, 0x20, {0x0, 0x6, 0x5, 0x8, [0xfff, 0x6, 0x1000]}}, @hao={0xc9, 0x10, @private2={0xfc, 0x2, '\x00', 0x1}}]}}}], 0x1f0}}, {{0x0, 0x0, &(0x7f00000008c0)=[{&(0x7f00000005c0)="4461b57a6fdf2d3de73d3b8d79351ce7124a71937abf928b75f51adad1096c8f1844ea3ee18e78d7126620f013f90334a3f22c570bfa1c926f9e3ae702f623ada8e643f7e40ba4e5495b667dbdf1e2431491b7fe93a6f6fd591ada3296467b9dcef63ea62a52b1cb90ba3bdcb3606b306ca3045dab0aa3aa44c10c556e6f99e1c18062e5894c31131c27e42f66dfde5ab43fcd7dbed825c3e0abfbbad3044bcbe5e908a4468ca35427b8d01279f0af225c5b03c645b5469cb88cefc736210dcdb12226fb5c13aaa04c082e7d09ff31ee411e32b7c82a5bfb7699c7019852fc70d864169843bf6485ac7640074d6ffddedd5111ab0bf1cde97b62db20", 0xfc}, {&(0x7f00000006c0)="0f7842ccb0fdd3effc0fa5d3d22d41e2ff8871d136a7268291dfab1b4385be22ec24d511", 0x24}, {&(0x7f0000000700)="8fd28462b5ba6fe0e7056391ad6becb03a0c56320bfc4e03b22c725e248d2f2828b27c92cb466fc2bd834c581fd40b8b15ef012ec53c73e7ef3626a2925ffd391234c631c92aea0c9f17b538685342fb9b1a4347fcdd824ccd2f5b5e04bad2dfdb0c76b88e566740bb5b6faf8e3995913df2f378dea8cdb0868d65ae10bba6bc8694d2aba24abd4af1f9cd2e9ce6aca587079e6b58c1867bb47daf9e79a45e9c8dec6af0648948e9fccff3eb432b10865ea435", 0xb3}, {&(0x7f00000007c0)="9191e7c4fb9d80f3b5d236af84fcd07b2b14420baf41ea4cbafe733bb6f8e401ffa94bd551a21fb684da758f12d153f04b4bc640adf6a50511cbf409bedf28a643edd8b8365f57def13f1171aae198600c1305a7f99813ef673842def974c2e335dc964e176390dc35b690192cd4205f4425c2b88759201ccbe91d233362178f3706bf02e501d38fb338c8fd2c3dbb3e718fce028197360c1e8fc48829e00c9b33569cb85c9f764ae1bbd2ae8063370fd78681702187b60fb5ef714babb0a0914c8aed6342b2a773453a76864d5af69c99aff39b1e1dada3619cb9dfafadde399039e8c0ef0bd5c6385d7017c2da470b22f2ab2021b3fc", 0xf7}], 0x4, &(0x7f0000000900)=[@rthdr={{0x48, 0x29, 0x39, {0x62, 0x6, 0x0, 0x6, 0x0, [@rand_addr=' \x01\x00', @initdev={0xfe, 0x88, '\x00', 0x1, 0x0}, @initdev={0xfe, 0x88, '\x00', 0x1, 0x0}]}}}], 0x48}}, {{&(0x7f0000000980)={0xa, 0x4e24, 0x2, @ipv4={'\x00', '\xff\xff', @empty}, 0xffffffff}, 0x1c, &(0x7f0000001b00)=[{&(0x7f00000009c0)="ca772d1b02988a1159197763d4f20219bc6b4112ec5d32a575dc19d7691720094c3b870a72e330b30884de0c103056aecbbeef55397c105332ba74a0280e9bcd02a00c7b07f55dd969343afa530a1defbae52dc8063444d11a064873b6377ab536caa8a9488041fd947cf3d7", 0x6c}, {&(0x7f0000000a40)="bfae6249eade9a7c7b1b5b0049090b63fa1cc829e08c5f861289b2f3cf4514e537dda1eedd77f610540a0e3abc6287b2ff5a2a78b51296698389e6193db07caadc74d0801179d8c8462f519ef9575f17c2d478b35ebeb8ca58ef21067f266ece058b0ebd8a3dcdad12b43d396dde8f675eb5338df3abc25a6d5efd524a1385c1fb526f69c2fdca74ef716fad56919d92d66165f1074b4890d9a62c7b47abf9c602f9b7e11d808ce1730d280d62a60002887760a19f68e4efc48c39d8b7585cccf33ffcc084dbee0e44f5133320f65deb55220774e8393b7bb99f080c16c0df92764391b2decef741a99fa9e1319f591ce8e4c4a40cf1ad7832a9d33ca8cb9529bac6a8758a41a98d54033c758bea8171ca028d5bbdacc2cf58126a7363dae640198b35f4fd00eba4abd5413a4def62e0cf5500d550f4072e16140cf33b8ff13bebc85757b5715108f9518b67b173bd9b39a1fcf9601e140c1fe2e8ad3d70a240d39482f81796ce8f50d960438e363523bd77d6ba6d2acf1a3700187fec1452d4a2d43b81e7e9475c0a61f91c6ff7b7e75db5311f16a0365baed7315fb92b58d6ef467827cc4f576b8385c74e2f43d1a8c849cb6557635385c33dca77633473784421f32dbb2d25bc5fb1283da1d29628a7895657403b7f6f84251e09eb06263748a7befbfcf2972d2a85473a617e264b5ded71adb57659089ac3160cb623a172d85425d6906a2452ee0ec3aaa6bfeb9e75f54eb2e86d715c95d9e14886bc41303d88f1e29539bd150213860b9b7276b443ba98b0fbf35ec40de00f3b8d570d519ab482c8dc8539d8481d5c80aa730aa672b57000a036b24f9c754463fabc057d4a9ceae7d8fb9a90287ae332b749d39c5c9bbd6284761de11902cd2f89a31b6d015f66b431643a2324463ad4e174c6512fcc4df1947e52cae6f07b36cdeca2945b2e354398ef697c8bff752ab718431aab68fad7e58419af5aa48c47759d11d633d70755a32511653900773856f39cd12d77afe7d87e87921a5c0373512794927a17f218550baf6fe1f4d2749a8998fd98ae99c647448371d457008ee55a45284515ce8f2d4157cc16c2d179b3b62629ea39f6c53f7559e8df28e77b02028cdfde71b6d28ed67bb651a90c013ee4a7965344fdda1cd0b477b40b209dcc7b05ecf7d56c767ff3340864b6c240dbe673f6365f588fba89fe6b821b5a044c3f9cc46d1a2bfb4149c30752d9ead04e7c2ecf7fd8f8a8949acbf3a293ee470c9e8aed28918fbb62701ac9b65a3c1a1feb83ebc20235b6a93892513730397663d3afb21986230362e6864c1c72572caa8a87dc3558289661acfb885f73adb39a595816af2d60116eb4136988bae71a7eaac1cf4d0a92747f723f49665180644ca8c73ea669f997d11fdb91f31379726a080ef247ac3609b98b665d4713538ba0f360884311d565159eaf07f69ca62fd1b6030e72117b26f281c61a0ae13a34ae262e6a4bdd66b2f26531b89a0865894e992a346834c9f60d0e45ffdb70b2f808651bf03e123ff9cdb3fddefcfc45811cf2c763da698221ecd8fccecb7bc378647c594c4b5b35e9f9dedcec7a45c600c438de12babbcc1368b3a558e03421718d8fac6b5771ba8de151b0ce07299e8aba8252cda7c0428baec13527a9c596ca722c3c1e9f45cb113ce210db6e94b1e4e96e7a21688a87c439153523c161d551e9af08d27820400092fc3c230343a565fd6ef6b5b26d90e5878c952c6c56e970bd9a18c32fb3bfc3d9ba96638fa12561d813f9f60af35a8e1524eac1826e49465e0c9c6a207cf0cef82308f15dbd3e0ba3e30d0bf03ee9bbe92f4d0aab82e41bcde38d37e679661e4c04e6ff1b821990274af85be2fac43f3079f84b159d7bb7b0e22ba54b5a4d1e01bd7d27f4ade44d82a062b49fd7d8863e5124d136fd8a2bdc621b0d78a9f8b06344a412919cf4762e8ba643854121755f52c7bb1462c1c8025017bc83708e5aeae469f702ed356055352600af554bd65d96c865da27c66b05bf6f4aa21436c17ee3c999de6b0672b0e622727df8d5859a3a4e92f7c899fd238532e73b9385c90dfc1c87fcac3224cd177e87aa70c8d49b5eefdfb6f28ab203875b1e474129f356a2554b87a970085b4085dbc5646a68fc6c0cce94c917893588c5127d2ef6bd749c700733d99bf4d1e6e8d2c34813517bf4ac68640b71d964fcfabd39514f5770fd99d6c4939c62c58eb3db4278ede6918b4d4d509be66538ef889fc4dc91965e0c0bec5f8e10e6bfbed47352826406aa671bc43a47388440e229a6c5753cee24ee581a6edc2d5ce2e0708d1ef4d0eb31295faef4cbd1a39d5cf5f221ecf413da910b61c479eb0f622d857ff14d679c682fe44467b5e53dc571266412ef4e188a5b75c5611fa9e35dcd234ab3253523997ab19515a1e36aef0a326eceb3e2ce637726c9d8bdc95e350297eda308a7ee7cb355fcbb1d1dbfb47a82d6f865feaa264b2cf708557c4435555403de89ccd8b48bc9ff111e73e335985e25209076a74b0d477a6345ad01f7afaa2e9d05f230d2a9bf3168092b85a3d9b861f20349a54db717b0ab0a706e475ab1ac2ac937d7e4052a062cae045afcdd51e28f013cc7a127efdcef76fdbf7932e693931f6aa7ba49cd8e5ba9d46ff35c29f5ef42505c2ab25978817852db684eee19e4ae30261d72f8039d52f7fa406f18570684121cef18297a300d5bf8ed96b8d8ece62b258dd2dc20ff3f309bf18dd3d0da3e6c6859e657d8f14c0268de2825c53ceb758d69ae2dabc2ec4a864578da9cf3abd599683d8d4331091812f5964db8afef2842212a323a2b9a86cbaf4cbe1b5e326391d2e5aa57eaf17e73f4da7b322b443c5c40efb1071ecfc08104a4a0f717154a4bbb7d25a76b179124c9570bdc1df51510e19aa03f43fc43d1872a97e6c222ce002cef31bcdd9a2df99c226163d2bbd8e4033d97cea235bc74f5a379f9ee53bc8d793951b2af9db7609b81b91747fa6a3933adf583a6060032473f4563c28adddf432a500067f79da655db2fec3d0293cb50d61abe84f251b7d8630b276b3406a023b748bd973ab4bd89f34b5ad0f46ba4a73c667a15177394e876357db47e7c8a3199be3cae0f25f05ac60025523e4516d074aa911f2ad1424a07cfea3a78d99a7bc1c99b87ef492d67b606f890c39707c2c7ec810e03adb66feb886ea4a40302d32c8cf7b417a9e8f934ccaf69c5ceccc1b62a120cd02f44e51a3f8d217a5cc6b7b6b41226c8b21e4e28ee541c4d39e4c56d123a494f82802efbc6da8f00c47987076379fc0c5b78d31f183bee69104ab382eceddafb5ae1731ff85ac8d065fe303034a0241e79f1ca49a65882ba2c7019433962d12380ecf9faa6daa7bda1b5467a2858c95308f8e2f26de268b5bab088035bb909937c407f0e3d21f087a8e1836886c5c8f5609812891036f0a20ceaa64dbd756b9b94b19bdbd6da5c0dbb0ccb961955f013d6edc7ce0ac99d1f2636d4fe2719e5837e184d84bbf068a39b6fc92aed2c0972063a913821b4996b9ff60c74c566ed7fea899c7f0d0c7da3e08620eb421d8f6afc9b4c27c37554d0c990d5b04c8b36e5b39b76cc05a6eed52e28b2188e1c5e0f380e6c0adf7b789608c675e56177ff714ef79729c4e200f2e6a12153b04444616178167cb375cab742e5a4ba428f1383d80ac5923b766d52fae3305edad6df586e4ebcee51249e052ca5ede137536fe5a07e966a41bd6dcd06e7ee3feba47a62474a4404d6c4cfbcce824fe7b96b550c248f2279d82f0537b5e79fbabf3853378c1962e0710fddc86bab1f1a3728fb8569563c8b83fe89ad308623f7ff6c1222fb303254f21ed4c646eebac11abf5cf1451e2c0c32cd537afe696024ff8e118229d4f47c71f9f3fc6eeb45e4e8711f5a2ade9d9c608ce89f0bc2ad55315ca1a23c737de7dade00e84788c38665583503ffe41182a578f3d05bfa2994ecfb9b5cd60e6277f257e433ecff9d5da9ae462f5283153cf519f4523183873f80e966489eb5ae9d2155a697faa8690ed4bf47f117d0639fea581acdba161230b1f7a7bd6cd13f9141a43c3c1ebb995a70e7a60e9c5468fb7c739ffbcbe45ca2f45e1b2203efa685be5d0ba7b1cc55017934f3fce376450adf65c59436568ab360b678b185f8af3bb5c798168169698ba79934ae18ff395b3a3ae042c69bdc53bb5b47757733a0e2e305b9b51e49c95d59d190930ef42ad92e99db757b8d69e21ae31364e9271199ee9a830d2a876cbbdd09228dc3a9e2329e61809bddf254cbaab16680f33d68add2502ff589b2106bca9cf73553d74e53c236ebbcf5131edddb566b0bf0c6b21c4b42971a072ff950aed17e1a7523921d685afb7fa2b5298a89b2214d69e3bb36f08e8593483df84da66e6851d13b35bbaa9ff8861e1a94ac2c6429448c64f827143f29e35d8fc65ebce8dc5a58e39e4437307a6b65d0d32876cef6df1c085e4876f509a7c141c9bd19f3d06a1a9a186a68be50f148d31f0545be965574faf2f980a91d087ada11cc43733a51bb4a8aeab1221993f9b1c7436187f1a2fcd5ee8a80d93f0de8b47d14a72a41e7f30d9ee1a051d3bf9d0c89b401d433d7dd52f9009dfe56a9f8cb74234d70f038f40ca1e9c9f4d239f5c9422bf43141aea7268954417b29eb47fc0d8cf4a536d0d62fc8ebab5cfe5f7bada910d7ddbef365862b407197fd32aaafd4399d3847b12a1cd2e4ea9fbd96836753f9dfbf54a8ee6a9f03165b56696be89a12fca80d4c2c07801bc3d82e9056f69329a94711a5c44e230f9aaa88400ee1a1fa04616829289ee69ea0081084de8624b25a5745b269bb5b909c1c63666a022d7dd7bdfe4862324454626a473a2b49804aa7803b6f20c8383abf0e564a9ab2e1f4c58db0ad26d899dd529a76d81e8c0ac2c3ea30a722fbdbcc8eee5f384e930b372a08958df765d06bfe1fd40bf244577172c34d44d799935c3fabf8848bf20d02e9f4d0174e235dd86662928cff827fd4d9c6217530a34756b0610a53a25e54eac528df5aaee081c48e22bdf0872450c8e83c1aad037612acf164121ca9440a03e16771cde8905f59d94d0d79c01391ee1cb19c30d93bb6c17f6dade71df0233102014c1751d09c3d6597f7335cc0b444c36328018994b5526cf2cdb74366615dc6a17d657c12a2e64685667d532dbf7c50706d58efdbbb32e1382217b3d7d7d284d971bad5fbd8c3ea979c051b948cb72ff90a682c1e6935e3e7fe9078239627b910d10daa5584e2d9b89574652f5e0e58eed13d121effad1d7b7dd368e4ebc7aba045898ac13f13fc0859c40822bc820621e7b29655ae476007c7964cbb027164432d68dff004fb1b8f9d4157aaf853c17f494411411ddd82846a0c66104fc2481417ddbfbb8df979caf4040569a2b671b6cbdc2f9c6b029dcd2fa6a50538f8f329240ebba158b3fdc424af2832f90bd7143f4f20e191fd5830f705453436df5e0ce9903c7a257fb551a6a2b15f1c186344b0f33eda24deb88e833f30694ca8e9f933ff30518060af6a524965ec76393b8494e3eba852147d44b8112d28665cb598219875c2bfa5d19ecea67e07d1b791d8815f2475ffef1f400108eeb5a689603d453670ac3a2c0d117854870b4c406ecab5965cce651e8d61f9a4560629636fc017fadbf9bd9fb52dff5c5a8c97451ef053ae45e551dc305f6fc4c6369f933f056a49e5dbaa10b2a842f2dfcd06a9f19fb24e30172ac40a72da37f3399ebde0712d6f1caa1f6f77f8", 0x1000}, {&(0x7f0000001a40)="71d313ad79bb3ad89b871ce7d902439711a1cdbf0644c242ff08692171d921eafc76abbb416dfba74ffe975ff649afa23635e300a82b8e727dbaf85feabcf688c40910a580821a3483d1d6e1c0a54801064ae141439e5ed3c09fb300767f0a63813234e504a7377643fe72c5c186fb8e6ff058a1ac527d734a012c628abd7825be17830e51c8fbd32b843f127e86ec96a239d4b9ddf0d230", 0x98}], 0x3, &(0x7f0000002c40)=[@hoplimit_2292={{0x14, 0x29, 0x8, 0xbd}}, @dontfrag={{0x14, 0x29, 0x3e, 0xa0}}, @hopopts={{0x78, 0x29, 0x36, {0x87, 0xb, '\x00', [@generic={0x1, 0x31, "a8feb4aba4f3df4cd116837ad3dc1907d5962b9ed7d744b334325e1c1e90ab28aa1da93c73cc658a5853eecf8c496e9668"}, @hao={0xc9, 0x10, @loopback}, @calipso={0x7, 0x18, {0x3, 0x4, 0x81, 0x4, [0xdac, 0x400]}}]}}}, @hoplimit_2292={{0x14, 0x29, 0x8, 0x8}}, @rthdrdstopts={{0x1a0, 0x29, 0x37, {0x87, 0x30, '\x00', [@padn={0x1, 0x6, [0x0, 0x0, 0x0, 0x0, 0x0, 0x0]}, @calipso={0x7, 0x10, {0x3, 0x2, 0x80, 0x40, [0x9]}}, @ra={0x5, 0x2, 0x9}, @enc_lim={0x4, 0x1, 0xff}, @pad1, @generic={0x8, 0x139, "054e4105e7d2ad58f0b2b7c301f911b4dbe7f3e41b2357fb37da641670558c00342bdaf9576a688fba09e8efffd134d9549094dbeb33cbe0f635dd5ddf6424a21a5cbb54717cc2dcd061cc03a8e3ae5a520621863adeebba04bab3d76556f2809913e5636b7b70996f6eef19fda744195019c6b66c43f61212168947cf8cfa3702223e9694fbdfb6c62f69ce754fc7a5ee4c089fe9d97c24fbba18dff47ea60c46690f0de993eabe1ce07987e46761c85e6fdb016f3eae7d4748bc8b206792e17af1e50b0aa5e6bd48a27565804a8b22319e5cfa65eb918c3eefc813f332c81928d5a8022ca912c51e6c97c4e2a8ce61d22ac6223713c4d830e53eb9788e1b9adc87596d6c616e75a0a9c192eed83919addf1f2791202cafb3aec2b68055598ed41b2ebfc984587fc9552360a57a59febe7cd6571d8a07a4e2"}, @pad1, @generic={0x2, 0x1e, "be4b70986db39e998ac451e9f34365db2875bcadd52acb8cbc8189dbf048"}, @padn={0x1, 0x1, [0x0]}]}}}, @hoplimit_2292={{0x14, 0x29, 0x8, 0xcaae}}, @rthdr_2292={{0x28, 0x29, 0x39, {0x6c, 0x2, 0x2, 0x20, 0x0, [@initdev={0xfe, 0x88, '\x00', 0x0, 0x0}]}}}, @rthdr_2292={{0x68, 0x29, 0x39, {0x5e, 0xa, 0x2, 0x7f, 0x0, [@ipv4={'\x00', '\xff\xff', @local}, @initdev={0xfe, 0x88, '\x00', 0x1, 0x0}, @private1={0xfc, 0x1, '\x00', 0x1}, @initdev={0xfe, 0x88, '\x00', 0x1, 0x0}, @remote]}}}], 0x308}}, {{&(0x7f0000001ec0)={0xa, 0x288, 0xffffffff, @empty, 0x1}, 0x1c, &(0x7f0000002180)=[{&(0x7f0000001f00)="21f0041d16cf4926402beac7019ec214cd5b9f26ce26f24e8de103a5c2ae80b04ffb5f227bb8e951146c19ccb58030f38f61351f0f18f066fef18c532a20f3ae6f5acdc65dd640ff495579723c3c58f14d0891005454e8d706926f63bee1b4c6fb512c08ff83c2e4b0898ad0fe972454a11fa739fb6f31cca91af535", 0x7c}, {&(0x7f0000001f80)}, {&(0x7f0000001fc0)="f7a54c5b5248ad3584ce4db0828669034d786ce16641668766089dfadfc71f91f20c9e1420752fbc029c35979ab2401fcdea5fb9c09789f15a0c2097d5f2ccd40d0e1fd67bfd1b12a4d57f52384ec6001e7d086095165d6127396471598b439e17", 0x61}, {&(0x7f0000002040)="cef5729c07dc819c5f5d947ee0479bf931fb68e7f06baf0b3653144c2fc07d302ae363e4816339b0852c31d8a9c8c9a10795e24e94d51828eec8dd24cba6c716513f426dc597aa26f0859821a9b813196fac530d7ecda89d74148aeedcd656ea7c07d6537e8d8855fbec06d9c25d8067be2ffe1b8a53c6ccd033dc80d372f7893246b56ab6127d1a9d0837771309d86050c68c9437e42205b8aa9bd786733db5220832ff0d75cff62dffa695d19015bb4b6b45557a9d1e677639f2adf95f8b826aae11f3829d0abf38755c1b70a8d344026d7ecf9e75b1eebf94fc7007d9425e6cdf55d561f279d3356e", 0xea}, {&(0x7f0000002140)="f702633c572590cf52aaaa1ec50b478afa6e501be3ebacdeeacc78d64d65e66bfdfbf77b1f04", 0x26}], 0x5, &(0x7f0000002200)=[@hopopts={{0x28, 0x29, 0x36, {0x2c, 0x1, '\x00', [@ra={0x5, 0x2, 0x9}, @ra={0x5, 0x2, 0xff}, @enc_lim={0x4, 0x1, 0x6}]}}}], 0x28}}, {{0x0, 0x0, &(0x7f0000002340)=[{&(0x7f0000002240)="5fc7264e044d38ba6c7b934bc26c7f763442739e829deac60be6d5127c26c9aa1a0eb608de20590d3dc4a76c0638fd4cb7c93d99fa5db8da1203d43179da22c448f1cd24d21ddd0af6d92dc98dfe5ae9e28bf188e09d8e7ca6a198e9d8c027b80f78c1fdcaec04969a8fc2c1b26432a661ec6552f8470cd0f6968b2e1c20fabe523796c2a49c59745b2e6755ba94c0bfc7b6d97658c34184ad4db4735a71408056407fbf8534c788236f2b9628b3f6089f9d0cd54260b3a4aa0348732b15f94b0b443a150f4fcf7435725354abe98daa5f1d022d3087c858c0f6dbd5", 0xdc}], 0x1, &(0x7f0000009c80)=ANY=[@ANYBLOB="a00000000000000029000000360000002f10000000000000c91020010000000000000000000000000000010100c2040000000501070000000000000005020009c910fc0200000000000000000000000000000001000740000000020e030002090000000000000006000000000000000400000000000000c845000000000000c4000000000000000600000000000000c1000000000000000502007e0000000000a800000000000000290000003600000001110000000000000740000000030e00ffff060000000000000009000000000000000000000000000000f7ffffffffffffff04000000000000000000000000000010020000000000000005020007c910ff010000000000000000000000000001c910fc00000000000000000000000000000004018b050207ffc2040000000705020f73c910fe80000000000000000000000000001400000024000000000000002900000032000000ff020000000000000000000000000001", @ANYRES32=0x0, @ANYBLOB="0000000040000000000000002900000036000000ff04000000000000c204000000010103000000c91000000000000000000000000000000000010800000000000000000068000000000000002900000039000000840a018000000000fe8000000000000000000000000000bbfc010000000000000000000000000001fc000000000000000000000000000000ff01000000000000000000000000000120010000000000000000000000000000a80000000000000029000000370000003c1100000000000007100000000202b8b60401000000010000000720000000000601ff01080000000000000000000000000000800100000001000000c910fc0200000000000000000000000000080730000000010a8108000400000000000000060000000000000080000000000000000800000000000000020000000000000001080000000000000000c2040000003f0103000000000000ece3a8518ec14f11b19aa4eaf46cef187cbbd6c81d01caff3c98028acdf0d1b9f43cf082405ed4d449c39e0852c3018e5dcf84afb3810334fe238fa38683e11f6ee0d68d76bfd9bcf9416d2061d1c92d5e"], 0x2c0}}, {{&(0x7f0000002600)={0xa, 0x4e21, 0x2, @rand_addr=' \x01\x00\x00\x00\x00\x00\x00\x00\x00\x00\x00\x00\x00\x00\x02', 0x200}, 0x1c, &(0x7f0000002680)=[{&(0x7f0000002640)="adb91e6a4dd0a6a089be", 0xa}], 0x1, &(0x7f00000026c0)=[@rthdrdstopts={{0xb0, 0x29, 0x37, {0xff, 0x12, '\x00', [@hao={0xc9, 0x10, @initdev={0xfe, 0x88, '\x00', 0x0, 0x0}}, @calipso={0x7, 0x28, {0x1, 0x8, 0x83, 0x1786, [0x8, 0x100000001, 0x4, 0x7]}}, @calipso={0x7, 0x58, {0x3, 0x14, 0x4, 0x573c, [0xffffffffffff744e, 0x6, 0x4, 0x1, 0x8, 0xca, 0x9, 0xfff, 0x200, 0x24000000000000]}}]}}}, @dstopts={{0x160, 0x29, 0x37, {0x29, 0x29, '\x00', [@jumbo={0xc2, 0x4, 0x9}, @jumbo, @jumbo, @pad1, @pad1, @generic={0x9, 0x78, "c77ff41b188929e2b19171eec4b6fe3f67811e680bcb9fe474ac5247d42111cbe5c14d440b928ab9bfec00d3ce2c7b577877a0812875285180f8e59eb2106647f89f1498f05c7f8c116bc11aa94be33e43b249820437be370b168d770491d98cc128eb6afd84ac6e71efe0f6d7dd8f6e5d6f85961ae6fec3"}, @jumbo={0xc2, 0x4, 0x7}, @generic={0x31, 0xa5, "1d27f16f17c32631ce821e89590a8f4f248f93e206f8d8207a93d6a2fade3e0c6b5062a514882c7135059271cf628d8ae367210993d3586f7ab7cad5ce19a40e065488187ba2f413862c88533fcfb3a42d7c1e4474547a51ab4cc7ffef6742cae94d8a7a8bfd1c649225a2e2cac29441cc9f1281bc9f88e75479e393bab79e898b477efdc3b7318fa8b82d10ade1ddbb4ea9bfe9cb89d67d217a9c7c88df424534b49a594d"}, @enc_lim={0x4, 0x1, 0x1}, @jumbo={0xc2, 0x4, 0x9}]}}}], 0x210}}, {{0x0, 0x0, &(0x7f0000002a40)=[{&(0x7f0000002900)="b91f4773fefaed659041649349ee9148e5650225f033731a90eb90ad5c59552c2fa6bd9d6a11a079db0cda5d9a151d785c34258d0d30a6167bc787bbd01684c7319d9cd49bbc2a1c16132b701852a9623c1d8a5d5e9519128f819ee3ce54c3637dffe6ed1b2e918ba05a65daff464f2ad8f5e88d85418fc1652c0eb30cd9210816ee38afd6e478f3029427c161a2a5d7d3cbd8a38aaa8f1ec920668cd730f864ee980fc2f197da7cba805629659b", 0xae}, {&(0x7f00000029c0)="cff84b8ea37ebc5980412998da40e782fdc386b6054574a1c5aba414d9c36b732e1697fc5bac7e290c70af9514d2432d0cef12894789bac734f2565c356d179c2ced51adf0b8e4c636df243822c5991cc333ecf2516c2b88b8b86c489884d8bd", 0x60}], 0x2}}], 0x7, 0x80) (async) bind$inet6(r0, &(0x7f0000000000)={0xa, 0x4e20}, 0x1c) (async) r1 = socket$isdn(0x22, 0x3, 0x25) ioctl$BTRFS_IOC_SET_FEATURES(r1, 0x40309439, &(0x7f0000001f80)={0x2, 0x2, 0x8}) (async) listen(r0, 0x6) (async) r2 = socket$inet6(0xa, 0x80002, 0x88) bind$inet6(r2, &(0x7f0000000580)={0xa, 0x10010000004e20}, 0x1c) (async) setsockopt$inet6_udp_int(r2, 0x11, 0xb, &(0x7f0000000080), 0x4) (async) ioctl$sock_SIOCGIFINDEX(r2, 0x8933, &(0x7f0000000080)={'pimreg1\x00'}) ioctl$BTRFS_IOC_QUOTA_RESCAN(r2, 0x4040942c, &(0x7f0000005740)={0x0, 0x4, [0x8000000000000001, 0xffff, 0x800000000000000, 0x70d, 0x100000000, 0x1c00]}) (async) r3 = socket$inet6(0xa, 0x6, 0x0) (async) r4 = socket$inet6(0xa, 0x80002, 0x88) (async) ioctl$sock_SIOCGIFINDEX_80211(0xffffffffffffffff, 0x8933, &(0x7f00000000c0)={'wlan1\x00', 0x0}) sendmsg$NL80211_CMD_SET_INTERFACE(0xffffffffffffffff, &(0x7f0000000100)={0x0, 0x0, &(0x7f0000000140)={&(0x7f00000002c0)={0x24, 0x0, 0x5, 0x0, 0x0, {{}, {@val={0x8, 0x3, r5}, @void}}, [@NL80211_ATTR_IFTYPE={0x8, 0x5, 0x3}]}, 0x24}, 0x1, 0x0, 0x0, 0x20000080}, 0x0) (async) r6 = syz_genetlink_get_family_id$nl80211(&(0x7f0000000c00), 0xffffffffffffffff) (async) r7 = socket$netlink(0x10, 0x3, 0x0) ioctl$sock_SIOCGIFINDEX_80211(r7, 0x8933, &(0x7f0000000040)={'wlan0\x00', 0x0}) sendmsg$NL80211_CMD_DEAUTHENTICATE(0xffffffffffffffff, &(0x7f0000000680)={0x0, 0x0, &(0x7f0000000640)={&(0x7f0000000780)={0x4c, r6, 0x2, 0x0, 0x0, {{}, {@val={0x8, 0x3, r8}, @val={0xc, 0x99, {0xff, 0x26}}}}, [@NL80211_ATTR_REASON_CODE={0x6, 0x36, 0x7ff}, @NL80211_ATTR_LOCAL_STATE_CHANGE={0x4}, @NL80211_ATTR_MAC={0xa, 0x6, @random="94f538bc85b7"}, @NL80211_ATTR_SSID={0xa, 0x34, @default_ibss_ssid}]}, 0x4c}}, 0x0) (async) sendmsg$NL80211_CMD_CONNECT(0xffffffffffffffff, &(0x7f0000000640)={&(0x7f0000000180)={0x10, 0x0, 0x0, 0x1000000}, 0xc, &(0x7f0000000600)={&(0x7f0000000400)={0x24, r6, 0xd57130b9fd170e52, 0x70bd2c, 0x25dfdafb, {{}, {@void, @val={0xc, 0x99, {0x8, 0xc}}}}, [@NL80211_ATTR_USE_RRM={0x4}]}, 0x24}}, 0x84) (async) sendmsg$NL80211_CMD_CONNECT(0xffffffffffffffff, &(0x7f0000001600)={0x0, 0x0, &(0x7f0000000300)={&(0x7f00000013c0)=ANY=[@ANYRESOCT, @ANYRES64, @ANYBLOB="f7ffffff00000000a7002e00000008000300dbfdefb8e55cf0d06fab513b7f37d1a1afe926bfe9f70ba49d8401f952227b27f3eb211ea62159130000d4aa0c0827a9b6fd21e2a15cd38034152274b81d257090fb49e7c9d05d1691d0cc05e8ae0a7d0a75ba944c3024f8f920cb5ed5cb42cf91b00c93bdcee7e679776dd67d224b746980e208cc5beb68d1290d9360e307034d31c6e10c8a60a27ff932d8b1ac038204a0080000ba65c7d88dfebc62ea4631984df100aa3a7ca380389e94e48f4ed4e807e9f03025798d059b2844d8fb7c36a8c6d57388d3cca2a0c29a248bcdb8e65b98025468b6dea1ab01dea1218b0779ee8307566fdc0fd20501606260df73ab2a1cc4befeb4beacc0c58d174895371b7da77581787819707d32114f2ae0a5c02e3e9ac7ff0bb7a49d539091fdd7f2369d180fc56272c0e91ffef200b47d338b00e19768f62debc93d85061a89e1060f52be100083dbc94de1ec70a630c8ce5d4a375f4a8cca0135a560603538b9ee2bf155ec17953c896b21ae57632b6abc25edaa0129a1afa044e6e5bb8e54288e5ffcc985a247f7bbef952568d7f628a6555cd82224c1f1bed5a5b7ec63c2d6c55fd884fbda02f017bf7c2621884aba46667372475c14d05a0e39ee2572d287fa1172a7b88efd47339c35fefca982b35af50999e1858a42767382ae47de6bc6ddeeccfc8a664693ca3cc3b0777139f2f0", @ANYRES16, @ANYRESOCT], 0x28}, 0x1, 0x0, 0x0, 0x4014}, 0x0) (async) ioctl$BTRFS_IOC_SET_RECEIVED_SUBVOL(0xffffffffffffffff, 0xc0c89425, &(0x7f00000001c0)={"bfea8b8b73e4828708777203dc85f1c6", 0x0, 0x0, {0xbe}, {0xffffffff, 0x8}, 0x9, [0x80000000, 0x81, 0x0, 0x1000, 0x80000000, 0x8000, 0x1, 0xfbc4, 0x0, 0x13e, 0x4, 0x1, 0x7, 0x0, 0x1, 0x7]}) ioctl$BTRFS_IOC_RM_DEV_V2(r4, 0x5000943a, &(0x7f0000008c80)={{}, 0x0, 0x8, @inherit={0x60, &(0x7f0000005780)={0x1, 0x3, 0x2, 0x561, {0x15, 0x0, 0x101, 0x0, 0xffffffffffffffff}, [0x4, 0xec, 0x4]}}, @devid}) (async) bind$inet6(r4, &(0x7f0000000580)={0xa, 0x10010000004e20}, 0x1c) (async) setsockopt$inet6_udp_int(r4, 0x11, 0xb, &(0x7f0000000080), 0x4) ioctl$int_in(r4, 0x5421, &(0x7f0000001b40)=0x9) (async) connect$inet6(r3, &(0x7f0000000200)={0xa, 0x0, 0x0, @initdev={0xfe, 0x88, '\x00', 0x0, 0x0}, 0x3}, 0x1c) sendmmsg$alg(r1, &(0x7f0000005580)=[{0x0, 0x0, &(0x7f0000001c80)=[{&(0x7f0000001b80)="ed552c167d4a55cc3c22be739c47e47179495be7e6294e52de29a9a65784e8d149156bcf08ef631cab29d2e17859e7892363d6d3c366232b422d7a55fc5ce32e7f561cd8d80b0b6c9974a37bbf6bc9c45f956f57381c93678c890d154fcacca6df17595335c18d4fd99b942460ed0c025a06e081bf7e21bb863defc30ef7f998cb96494a60ffeec419fcddddbf71452dbb5f3600d3010a1eb73c50c3c3beacff10e09d85c9e768084f107eebb93aa64af5255826ac753b1abe7dfdd25e0dda11365d5db95746b677d710837f6bb784ab4ba16ed107269fb25f9f56bd4779258c56a7d50c5687a16417389f906e09", 0xee}], 0x1, &(0x7f0000003240)=[@iv={0x38, 0x117, 0x2, 0x24, "6edbd9fd16d13be3badcce91fca07ddb8fec1f02568f62197bff5aa1e622f580d2c70542"}, @assoc={0x18, 0x117, 0x4, 0x8}, @assoc={0x18, 0x117, 0x4, 0x4}, @iv={0xe0, 0x117, 0x2, 0xcc, "f95f34a2718300908485e95b17b4655ad21d41af33566aaa3acef7f7c9037989175e76f33273ef89bd6ce9008796d831e099f58c0ce6c1ae63e4cd375fa98b0872fc482e2598a85f05811b92a6980fc2312ebf3ce0060eb24e06fb1c26f3241037b27d0644e04084bdd311cf60849c7b079e5b8eec26c9f74437ccc1190e244546afa544abff041d3888d7c2649fee491755f9c95cfb6345de1e91610f065f30ea75bee4011eb8a45277aeb1469a41cc94ff8982bec906bd0bce2f876fff225ee54ba9645b7ace154e6fb5e3"}, @op={0x18, 0x117, 0x3, 0x1}, @assoc={0x18, 0x117, 0x4, 0x480000}, @iv={0x80, 0x117, 0x2, 0x67, "fae32821d51229f830c86660d8ff59b74b08e41910816705269ff1f914550d46eb00b278863cdc57df68a8d1fcb79e6375bf1765e93c7ee927dc31c6113bcd9adb6ba20e6350bc21d40a0c75426623082ba2f746438c45c4bed084945ab7ab593e2d9e142287ac"}, @op={0x18}, @iv={0xe8, 0x117, 0x2, 0xd4, "5e1693468fb457c5db72d7cbef2c47d7ca30fe75a1ff264ddae90fc27e6799d82a5ef3f9b68a3462f5aa8099dd9d68ec5367eb395da7877438d03168e535c108966231714640d9ad3b2e52c14eceeadf67ee1bb8d5afd7662e945d815c5052e079984190c6dd51a6fbfcd4c42cb09eea4344fa91fb41ba581885609c89780338898ebe70d323690403e5b41f3e8dc24e8096802e6d1faccd98af3adc198c19e6e38d530ad597112ce7752e9bc56db2109015a9c23c60c6f83f16a91e5db0990059e85ad2ba9e26ef3bf63526f04b2243f40a4acb"}, @iv={0xa0, 0x117, 0x2, 0x8b, "a09621c66c9fd035ec3600088f10e93c4ed2f210edc75110f71fa9a97d8732628f9b5420e90093b458927c33ad71db27caf66ca49125e7e0d329e199e91389148b674df5d85c74b13133e1e563cf731c67fd0e40796db83691cff4626d82cde8aa56271d70229b0a51ecdafdc46a3d417c0044b6f21648198e40f50cbc79b93d22db60c6fa20544f05720e"}], 0x398, 0x20004001}, {0x0, 0x0, &(0x7f0000001e80)=[{&(0x7f0000001cc0)="65a1fd90cfac3fdd1549ea83afbaa3c9960d03df16e605ac2993ceafd0faa39de2b519d30e49afe4b83d76159910e5356f778575a696f4aa9e21f7abab47f294b23ac00cd011e993e74018e460c1d7b0eb8fd05f507bfee0b0db687fc988c8e87c31a0eacf539771043c3ebad7269eaf821033362a5d878018e03bbe9ce71a282a642312347e9b5970c5c6", 0x8b}, {&(0x7f0000001d80)="288ec28e089a68b53a451afe55a1fb44922fab34adb87fa95f15758edd864448e8f9ffd34f07fdb2447ebcea98414c0d9e84f3c620303d3c071ab5bcd4dc3fbc72eca8", 0x43}, {&(0x7f0000001e00)="ad6826acb1cad947648f197812e6f2e91794", 0x12}, {&(0x7f0000001e40)="fba91c77a4d43c9b71d07e1994b32cb626682b1a357083282aa153b880b2d219d3cd8381abef9941f0d56085792dd9bbb2b1d6d617a6dc263d", 0x39}], 0x4, &(0x7f0000002380)=[@iv={0x88, 0x117, 0x2, 0x6d, "915f54d745332b14e3238fc60e3d9c70338727557d882b9d0a1f4eeba33bf37506ef73d7244388835a908c6607cce6adc097de0327a5bb68ab2c492878d5db9a6a28dfa009574639bba09ba4eeb507b7af91fe8c3b6ea7cd1da2bd00ced9976399a5f921a1293a24fec27650fa"}, @assoc={0x18}, @iv={0x50, 0x117, 0x2, 0x3b, "ddedebcb225501da7e5f59a3ca9856fae49e65e9495a9cf475de8505c0b0217b4e93b468ec5ff09e1e79b11d8c1f84774c162d94dd036f03304cf4"}, @op={0x18}, @assoc={0x18, 0x117, 0x4, 0x10001}, @assoc={0x18, 0x117, 0x4, 0x8}], 0x138, 0x8000}, {0x0, 0x0, &(0x7f0000003780)=[{&(0x7f00000024c0)="df8438deea21dc7d46593dbe4f24855ce6668a8b882ad89a5efc1bd4539c460851c6c02c3fd2095e53750d5c1c0fd87fb6ad969efb56bd03c2b1156a47637f11785c9e0a84da78e3173b3fe78e9d61b1a16f03969187d72cc88b0f561811fde58eb2d65ea8b146c0d4c4ceee2558215b1347aafbf6ab1b8ebdcb6e5175ae0c600f7d82042e38d7b09bacf529efd473a9ee9d9ac0f426a7979c20a41745a2176496c38436849050a56994beed5fbece4661e33e937737c2d998138ed8", 0xbc}, {&(0x7f0000003600)="a57cd85e0f340b4961dec3cc1b9789332aef8ce7e2c1437b884819fd15c49b22894d259f9291594abb6c92754092b7f038ed9f89bdbaa8b3950f2577297d7c767b54e947a20e02e67b766c0f6efc96a68d663e99ad6cc7a720c52f533d919e2c820f2be54675206b25a7fc800e9b73cc844558e176e814af481bc562b50c2b2186e8", 0x82}, {&(0x7f0000002580)="69d59e2192dc2927503a4c96a5c421d7062ef7f087ec77da585f1b7b45549bbabdf9b3f262380f3780b965b6664184f0327634c86cc35d10defab8a6d4412065ada881f2a563a895", 0x48}, {&(0x7f00000036c0)="244727eb136383101de3c6c0d84d7e9b82e78a0dce481c7c2520c8beae36948dbb9f8a398b", 0x25}, {&(0x7f0000003700)="75e8f218187deb24d6124791fe2b563aea7f537b3905c3c4fa7b5773c83714ac52b925fc88bc322d22e0f7439774bfd0b6f6efe71ab8819cf163e58838388b7aece5a6b674020808eae29ad6a89b337fee73fc10424bd27d50f9b342ea6da758e16014d1f8b85732", 0x68}], 0x5, &(0x7f0000003800)=[@iv={0x1018, 0x117, 0x2, 0x1000, "1ce8226f58e3bb29f08fcaa189dc03d294a5e372bd2a15eb9fe2cfb4924dc37db5dfab3181b4d5911fcad08771ec058ee807770156a412155ef62c1adda20bf8e3ab3d4ac5aac7ffdebb02e2cfe024e4969411ee7d0d62d3c7c2b67c276dc4c23bd4873e79207a1fb67a6e536bdf1cbc6aab44a3bbcdeff80e1b9d80e53170eedc66285922ae4e75af2f3df1eb219f55204944b02f1fd6c0347a1c6d76e9e23d6db93c39b7e03d719e47381dbc4288a947e5f7b5eea5f1e81124d378fd5ca1010fc447ffae4feb5d2fdc1bbf7ab838bd1b576cd79a072cbdd0e532b4e71c0dc98931e480a3941632d14e69ce212e4edf5e0d180101d58adece77fa2858c377f6b70837a42c306881ec822226ea2786e2e9a4303c6c6ea8564c7adcef401c6e5c1d501175e7febccddbd7d06f3e4f086c304befd0a8aa3252079b1fb9ce14fb0775647ee311c5a0a9bbeb5c6bb38b0ed39f03716c9d023c662b60ca2dc71656444fbb3b0e6e8ca8d6a04b080f5e9d55d87283ef4ca02f2da6413d634badb448f25a8c3e21cce21e2945a5c061a1bc389fe5cded14d0fc06ad912a75918da5de990912ab7dbf67a09025eae05df2661b587135250097c7be3829451c793e22f0312b398a5b5d8335fc0d366bfa6411c025112ca70ec0879d69f8b2c3a5cd4901543940773595493ec04a8b85bced73420532fe19d699bc120157a761b1601f59a419488c6b7d3bda4ea2c9f3529b21fa8ba20ac25da0b5ec9aaf0e621a6a819b46b2a09b202387272af33eddc1787954fad70936b82c5c01c5d5c6ba6b45e6eb3cec1fbfebe1eaf604d05794d2fcb881a36056c492b5097df8db56eea13052f92bbcc273c950592c2df5f602530acae8ebe9332df938837847b970009a5c211a0bb81db4710176c14d7011c4d438b24231fad18e94c5dff77c2072bf9e04adb51b4b2317b3bf98fbe469940c4572b189752d1a00298a223a6752db6707cbb729562cf9749f2111e675b2989987595bde5bdc6fdd46284d7b4317abb6502cd5ff9c8c3a85cabc0efd4c9935a4aec76c9f219e7ef1a7546c09723a67f9fca5cfb0afbe6f38d1b7d62db14890c51ef4e8c9a08911fe7a116f23ec727641f0ae3d220a50ea0971ac8c121c76cbceddd2a33fb79e1e7835284f44f71025f2c5dd8543de2bbb8ada4c428fd2143e95db9993c549490421c1c16bc9c57cca7fed9673ae39ac97faf1c31e8b6a7afe7cd9c8b634e2d0edffdf38f8b079a7d1e6426f7926063cfaa915f7f551541f8adf3fe9658e4a57a7ee5d1b33c8c8e88fcb93e834bab79578b59f596688b2687dde73291280541fbd90b5b5fbfe274533c1d45bc8f6d93069286fb533f8dbb35aeac5a421909cda1c4614ce4eb728b215e4463ca14d38e71b90a9284677c1e9d80d580d6474e56a7a4e56c6c3bc8f760fc73153a505a83349fbb10ef4fc2f5d1ac5eb63c02821e5e36fd098750e57f664090440e419f5da146c5cfe466b3123de70a6016f623436f8f8c9f94af50bce3a719b26d3f879f816530d7f66108f8395e2e038208a75dd459223975e44e77a75f0f5fbf238df92d66f9efc9cba6b8e6f1c1b0e90ff51436ef59ba6afb93a79dc18e75b64fde6826b0e22675f22a3b785935ebcc474fd1072b3d571f9ba261adaf52525d4a17a6a1793cf481238981f7d62c74b6ad71b0e590b6ba9d75e6408571e0ec6223d3d179287e00ee128f31d4887cc9e44b4df0c5e6261f32d7bb24bb4f547e7bf2d48b98e8d488a8f9b283977ea1cf97fd3480a75af56881fc39734a8761105c280c008812d0defc2a8262383543ae3dd1a32c3347285d3495c4f0ac42290a38d17541b5a8390e27febc1ef90ec625f1e40f664b75a95a97a8a82dcd92e9d7b010cef6ebcc2b5d8f6968f4172e83807c5f596f7e9bda87988d94af77c9288b0e8d049e4151749fac4108c774093e4dec593c2f09051810e20882de1d060cb59319a4f3c4de0a127b9cc2f0547baeb8a9b1b9b05faa03672ac5733c8a94a71e300a2a0b99b4dff1801ccce65b0b58eb8eb1eeb7e8bd8543b11d7b5639f6af83afaa383b2094bce893541d8628530e7c5fa040e2a1672bfd9be5ad0c5048bf745c54af0dab6745a5644819a3c2bf3dead750984f0a8eca97b525884b8050ad250b43768c0c01509bc11d22b15121c9f2486179f7d9837e209edeaee59e457af02d9b64b9beba328b9cc240faaf58ac6b23329f170616e51df99a79ed7d3d72904ad562711f56183dc203b735ae9323c01f0e67669bd13897895850d9e08a43bdd0d6d28dff39bcc9c76b73665bc68f549d80ab62d54e5b0bf081c07f18044d4c8f8ba6ed65dec4fc77951ab0b5276613abc66d1473516d9eb1dc49b6976e2e1bf831413b159081f9d823bdac8ab8a82926e7580e54bbfe45e5de8464355a6b181f1cbb0c1ee2ad5f0f57d652580bb7889394ce7f1a7a4f17e426fbbb57b6d8c4bc2e187793839d91ffc611155596247bcb9c25b638d8ce654106a9c74315019b34eef749c847834bbb4eb78011bcc563d27cf45c2cc2bbee01b75db823320e2c162c32f31047945c431e10836e1b1f6f3b7b8c3aafe4d9655347e70c49f3d7285df7ce380fea9f9e17bc2e115c1f0d0a8d7f6619dc14beaa39fb86f1cd1fb37fe81b1eef896cb024b21bba890f2438718f83f305c52796f15f8c5c2fb874d0b83602d2510b0e5c82faaca0c97dee3a4f208ee1e2631f296c952fb2d8539eff16346abd58e09eaefcf82ac060d52afeaa6b4190cfd6f0182a876cc32575b343b6b62b75a95fd95b9a86fe95f87d739847615b3bc40126f38f579225e35864b14a924898834d149cbab4b7275027066f827665ee7821a6eab86762f2abb9fcb45475e597cf6b08da4e720dfd04d2abf0905d2396eff74d8b345936a4e176b140751e0aa824e0e9ab5f422f3d860b31be5ddc670ee45a4c453031f02a70cc1d8f6bb3ea793791f59a3dad7ceba9bb77eeae88f646c8b6ca8254c3aff7be71f944084df40b621b3ca07fb75cbce8238701fbe16212e872f78d4f0360f158de8a24560cf44db1c007a323c39086fb8864eff3eac529e877f02c461d748a1e06e07e7524930001487b4ceb605c55617b775f1e249bd1492a2926385558f4b0fb28e8cf814dcf9ff983b9ef860ac32efbf517d147518e4af7bad54c2c576dbe0cbab027c308389192d6237b146252eb1ace886d010f6d1bd4bab35e96fe6dcece551257b3b7a10418b3a0b2ec251f8d646225db4bdde3fc7b594ec01cf662225f3db40fad082fb5c2949c9a9c2f990ced3e39e419dc89c2af3ff19f764dda953ce3f66a8be4ee9267e7de350653157016e0994b07e990a210505adb4457443ebebd13239d80300f004a164da37b20096b475d168542503824c72ab2271504e01a21b32818245523301743d06d23e45f63fde95fd1992efda6a96d066baa9cc0ef5f8d4dfbb8fd29c3178e33f24b95e62d03fafbf48aacec15f71576e6f75e22812ef74739338b5aa5da5de8440fe203d523aa6853845394b78085097be22a8b93ccad4de419b29949543ddff8cd10eb9c4709e0b6b4fa9bb0cc32b89ad74ab67a19e4bcd7ce585dba83b7ab89575f32e9bd0f8da9447777f420de282faf5a8530a818ebf6b18743c5d6f3d5beb2fedfa5f775155ad8cc30dcb613e42cc212e161ae88dc3cb26fc6e73bfab646ad9e019bac667cd06a5d003952a04597f90ecf7ccb451f30afa8fbaafaaedcd9315c16eabf4d60472460374c9b7e6f74486ca6baabcc6b213d674111d0166cd1649b1011c88f8f9f1708af6f754785c087a4a07cf359ad3a4ad0c0e340d71a7357545848ce0e06443271b3f4557c51e3e9fa07df711c48fffffcc818b676dcd393df71a9832ae8a94c2bfe42332951594558c09df979b76ca13623ac6aa448a9f2ddb2e073c098030df23a89659a5b5482db17dc0acc3e73b6b63316fda7a314f048b63dd9280948a9077276e0bc0d048e084a6bc28c80cb78e4710bbe106432f2ce467a4879603d850c7f0eab13102d2c710b9a661f482d937a5f06f03138568fad3dbfea1b89db99ce44819a9a81a4a46c6c56be041a34717e65c3ea158c6ab7866331a29680ed24ac2454f9bacbd34b63bb449ebeea6866da22cbe9361a9f0d4559d4f48fd647aaf00671136f9caabc09c7f418330b647db960974dba4de8d607bb91164e6186cafce1214e1c1b61c79613056be19b86743181fe0a931fdd03f3555f2c87918b4038219f33789de1d873f9935cbb330cfe6f6c15960386853da4ccbe87594b409b9eeccd09610ff346c3bd026db1bc410037c54217a99ca6ea69081bc706b0c920a2cc2a3c749eaa6312c9e52520514cc3f3bd5cc4c9000ce681ed7e063ae0052f1d9bf190c78c466d7e621327866d79f01e45c34c7b1bdd3e081d42a446d5fd183f2b1b49b8cc0f913627479496f34823f778c8394623e903e74629563bad6eabf97296a212353bb53cb70b5a8e9d42747871fb48cf798ec22b09a2c9b419c324a7b6e8efdd62dcb1fea7085dc0bc98a761e7427bc9b63a48bda5746a850365c18df2c337835246e7a58476e6257f013fc697ebc01af54ccdde70c252cc28cff6bb8e2d7c7a076839aaae011b97a8cee2914a01b5904a027cf895c222dc4a5a15addee15d1bc40e29ee11b008248396215c822da423e467479f1c8b1073f5eec2c42c1abe698d8690f23879280006b0c5e8bb829119482279e4cc5d001dc99d68c245e07626124565a72fe6e7bd97814b52b9f6a5d9718f14106ed5d3cf996b7c2aa7bb6da7cf9ea845d4280ed2ba9cb35b4628c2eccdcc584774006f9ddfac5fa66e5984364aa1f15e4318b5e70716432d0520b706452809fbb70f84bacc124ef3871d1e68fca5ffcdb56fc211e1bd3d83beb71a1823e289efae9123e7ac5656ebf381ecfd7e5aa6a514537884b984d43741dcd8c3fe09cd4ad7271d7712a3ad642b91969bee30a089231f814d349faa29777753971914bd938fa290f1df53cee9125d71423b061c44fcbdf52e40c80386352c9bab049a15d57f27fe79bb5e83127dc049cf95abfd6aa4786cbc0a7a2054c236bdde369179f6516500c98b56e8e777eefb3efb20a673881908432ac80cc4d7dbe3fcbb5fea2255a730330e2320382fc367a5ccc842d122dc97900099a3416cbf550a4eb9f7b061ab2ecfabeb3f8f7582ad518730df123ca920c0daa3900cf3332284432123decc595395ec2623ef1e95f8335ca2a66c1ede95e234bc3094c6c3982a363aae0d1069b6a81ef9ec68aef87fe627b528a1373328173fd04dcfa42cc800e67241ecde7a8f25f3657e46dde3ef1dffa336845631a5aab9872c1f6f7461101aa05bba2bfcd3ab00694761026744b663db0364abed735d8a76b5a63b8e17278c73b3b1c9574308941245f6497cbfedf1ec3f9a5c655d7d7afb252afe4ddb4f1c538a6e3cf744958c0135be724eb0b546704c28076cfcb9887de822f92d8f454178294c45aaf5529d3065ebb634f7b80ad53e0be26ddbc3d92f8fa255c3e74c1d14675f6c33409d8ed36ddf6c2b18e290efbc302999ab257061c3eb12169be5840125ab6f3cb6fa5e4365e5ed796944d9b8d853c5b406eb272788dde5c46f69c260b30541e29e3f4da0d6d840346b57b8ee7082251f3c6134c245329161e53db2700c9cb1a6d107b39f76bc29286312a2d7d095059451963c4692737784fa2445f9aeed93daec71ccf87aa33c7ddd5db61a7e1129fe1eb5584decf61"}], 0x1018, 0x404c000}, {0x0, 0x0, &(0x7f0000004c00)=[{&(0x7f0000004840)="c0c8d784997c32230adf7bf3716b169dc6021285f50bc94d9b97fccdcad7", 0x1e}, {&(0x7f0000004880)="935362aad25d5d5c9c9f413dbca985ecc0d5766fa806a7fca4b5dcd15fa3ac9461aa0c6ab0841dffc4210babd33e0d797225c420d72af937cf80a7f0bd7c8ff5b17af2820061a0", 0x47}, {&(0x7f0000004900)="ae60da53a53367c69594bda0d5597163a84adc4777fd1c26b183db1d1661494b3eb2dfb05f7151e576fa8914d7a3c220541d5d1e982aec8f0f24397a53eba11b09", 0x41}, {&(0x7f0000004980)="3dee85c6451cab53ec37c7e91923935e267abab539c79ff4bbcb643d6b6acee593a4dcc6025b1196fc846d6ec55cdcc39bd112e0d9d4176c498038864178e7ad0962daf34b69bfe8812b72dc1aef1e52de704fff1a106d5f0196207c4100742a467df61fbcf278bb23030512e438d48671485f74d47bec52db4e469db7708aa416944d27c972d146a8c3994dd68bfb93df5d942c30d9288a781fc3e0f1f867d3d436d5860970027c83fc1d881bb0712815e52825787304b4deeafd07268ac031c6e7b5c4d6daa6eb6df4fdfc032ed7a91e3ebb4649", 0xd5}, {&(0x7f0000004a80)="952970f3f76ca48b9ca1d122846ce62a98970f9d2cfcd8bd6c9d034adf9b57c889987408964ae192277018226c05e6eb087cfa9ddf0b87bd8895f4a104f63d15124578d943cf175fcdcb5ef114989e36880866f1c2a7c60e5445175454e1b7588d98ab799f1a7a06a690ac1f4e5dd75e22c50e208e64cadb741cafff47747fdd3c5c5272a6e5098b6259a9466bdca3f4912cc49e85", 0x95}, {&(0x7f0000004b40)="43aa416047f36ac75d6762a001bd75824d8e18e71f0743c927b6afee7fc9106b906b192cf62bf5993d335375f8f5e8f8e5da9ba957c35bb042d142f66d734c86939c0d5a1f86a22686627810fa87c70eef18f8ae1a843d525076e6cbd2fe66a478593a176fa838034809943f39c0d85c96de1331b6f1cace9139df4dc557105120fd4bdd8d34f0042a5db91a48ff3ae636dfc2c86ab7c650ff58181cf954c77088d9ea0aa56c67878000bea8fdb95e2cf0f05c8936c297136e", 0xb9}], 0x6, &(0x7f0000004c80)=[@op={0x18, 0x117, 0x3, 0x1}, @op={0x18, 0x117, 0x3, 0x1}, @iv={0xb0, 0x117, 0x2, 0x97, "d254a4a3a257267fa62b8d4fdef68efb50b9ff97fd381230931fbb4387d72a49c4e1170eb13a08c9f9e2f41cf3c86c1404456b9639432f061d0fe30e0ad05ebff5692f095377ec5d1f7044bb598a674ccbd628fcde801054606ff1bb36748d4a603df16766771ace6b5f6495b51476c4af0f6d651747c66934d9b5bac5235c47634e83f7e29d93038abcc94c9841c6298c755c48ce0d39"}], 0xe0, 0x4000000}, {0x0, 0x0, &(0x7f0000004e00)=[{&(0x7f0000005c80)="9e3c5ca37c38c681038a40efe1c7e1576aca3454b7ba2a0a440bde7b43ee551ad1462d2ae487f583554e8fb0740148e6d4cbba2850cb5f043b3b6e6ba43bfd603fa6b9858e583d9dc1e54b58d47b7efdc2d7d5637d6bac36a54e5d572c3cdf78a08cc9abb2302e0c767c5c1a4e4ca76f80a17c5fc4c383f00b28a7a67f2670594f4a7890608348b58bc12a72e0a74361c123966ec39ebe9be7eb5a08dbb18dde893bad9911fb650e237bac40596725e62bd7cf62cc327473c42572b11800f744754277e426a0007fbd4e929496535880cb563142eb377d6aa6d8159351d31722b6526ea3aa20137b45925eb9b4880c01d315509be03d45ea04bdb7609f2d8b2c771abe515bc76f35acd46dcffdce518693684cc833e751f3782f80da2d141a574f60b32d3a9256d46207678fbe15304c884493b3be81954fa3dcf1b537d67a58e05b92c81768c30d1ebf78c1bc1b22c8323e0d5efee892550d6a1d4ea0de10bd3682af31fa5b7fdeed97182be9651c2d3df79d1acbb07b0c9aaa6d6044ac06c1b77fdb6296a3864b5ac2eff8a6d83e8c53a412e918396d535af828b2f3e165ec5f7b82c35fb3fa48d1a9826d775f5aad02a8513cf6faed7584133162d82bbbca8ace2ff5e2725d01c3a21b00ad69b9f7db737a8a15bac0bcc34426bcde5b35708f7c913a0be41a728799a182a8551983b8f0959ea9c730f6894f63bc79aaf23b0b278d3df900ad89a55d481eac6a03b428559bcea6bb3760f0e249fa195d0b42edf9df67969e47b6882e43f79fdd2487bc66a879966f6c0c8f272bc4d72263b4436ae126b7483d5e9b3c1db7dfc34bd6a1af69cc1f65f5903824959062542fc0058fe28996c56996e04e59fe1682357a886bd767ec2466c3eacb135c1a3f80ed736b2281a2c134d15b2a5be1628c92db8d72a72b2df5a7c187c92d9433a3387e330d29a04155d72f39a651fd2e71de27c334dfff2e0beaa0962e01aeaa57d7fb3acb54046c240763f2e02e80af7f9a3b6b062ef4b96663692a6a489b271ec209425303da0d5e6f4b7deeb6066422bcf69425ea54d8ca7f015de1a27c343534a2be6589a7990fd653723690bc249e911e13da8a59706e7372ae180f601fa4ae79f69f39c26fa63bf4af8277e04e72ded7991f3bcb9e1409ee8f9ec195cab700dada2b09529aa9eab19eede1e04de32437c99406c263fc33421c539cbde56af0c2f15f01c4948f9ba165a6f4935fb3dac22852571671e8b8e1c56029871230da84db3b9812fa43f9c9cb515bd2714a8c0832061ee0023710d8ad4cdf93b7dce57337203abf94f103c680008beaaf556c6bbc406b2888d3abe49956f5f2db3e2f14aa7366d95bade9d738ae5aeb14c68415f3d7e61fb5e731a10e5f41722cb54152bb4768c4973fb1be2b1962ee7366ab15ad543d1b147bb1ca44a5a0aae7c90e1b7216e84907d31251abe04e1adb8ee3364a10a17fa187855b67a49869e8df691bec43d06dccce6f3714ca19bdc9d89b89ec712a4f7c2f62fe38d63f553b83e36d542a5947afb8869088302069eae33bb8ca70c9b31acaa0740102aa5a8a3f1f2f5f57d2ffc21186d9a064714928e91e19188c2ea780cbf08c84aea10ef16bf6c7bb08b9edb5b89e41100957c96536b897eb83b52f81664f88f0f71532f4d4ffcf73142cd5294096ac14cb04bf5703b38722a0ec5a2489823ceb1691ed3edf1671fa9eca02edf16a8b32f3550b7511b9b88082db53fdc271d3a24b0b742b94b4f780155989e0e9b88ccb6a40a3ffbf90cfc4637396ba3f00398859ae9692ace4de22d7f47e71ebaaeb7196a3ce1bc6f1d15051161655f273a87fad8ef871b38a1d029e150ea4a9ea773b952447e9831bd5211a3789d32f379aa15d6540db8ce2deddef4c37adec7be0edefdca477f77977933e7f32112fbae4712aac549f1f85f91e531763c2f11d12aa92bfcef91c5d3209e9b0426d368335f1b0c591edd76215b9524f9e4ebbbfa00ffb8eef80cf438fe7f7d35701bea62ddfd19e2180651192c2f4d64fd96c17ee75d2d76215180bc96479d690d25c83b2ffb088e92947c6a4e7246580e9c02ae7e2dfaaf0a018a13fc351b53beaa6d32feb5c7a683b4f9594f41994d596f78b3c24fe1e7cbfb79de6fe9ef65cc1f1c166d9abfdf87fe532c7efff220788be06813913435debfe17a11f8dff7c521cd0434266887e0c4bda9c44b62a72c43ec9ab57d998a3bef5c87df5e9becaf6d54bf8d67c0851fd2bd5d61c8adfcfc885ed62b4044cd8a686a28f321a0b60b25fb7c0868d99c46e8613ec8e7b888605238065c4c04488e38255d00ac0d3b9ea474c85b643fba7ae62374a76da567c11dd6d98ff0a6d8f6989614f86e3a1cbcdb1baf8764f51c972643d97ba403e830f884f2f5e003b8905bb1ddb028ca8f01979e018f01f2c62a134486a20367899a15835231771fab1d24860de3115cbe54bda4caf9683ee9b5c96e05d477c3f6af489e7e3384325df9ff3e68118a948163be57f7c333a301251be8710fa2b741f9d579ed631dd6a54fb326c5568db5a47910bfaf198ca00823c026f1140e990ed6ebca803369771fd8fc6035b1e1f09283435126cf7624dc33873ed48934f729fb3ae611662172007118efa9a3a9021990fc9be75bb629adf858b65c56626997c4b5d4a2856295cb58a5f57633da592de7879cbff1253b8b63b6da0586f1fce7762d5bf881adb8ae42ac9385965ad632aac49770ff6644b491b4b673f3ed21a4eb74991a7e752f3c9c1ecdbd0d6397350b3ca6cc1590b55dba88bb53c2f19f06ae2378c8aaaafed257eff675254d179e6cf248e27d3dc02e788355e193080b31aa10b7a5baa24726f6b9f26c7551bd440289bb7c7b5d765a544b085a975eef9218252d4f44e0b9c462b5c87eb8ad1bb477b200a9d86b9c7949cc2d75379cd5c4407bbf2dcd142bba19bde413362f5257a9342ec935fb1ffac4c255af774b3f59f1fc072376a15df0c7bd559ece789ab339075153bc2600fa0ed4456e99a5c97aa57154af889ffbf85dd9c2c855366a3ec991ba08c8edea7690065fb9c1a6817571d422523d9abf677ab04eb96412ee9c44cadce9505222d881000a054e4447d2caf9aab2b13423df1cfbedb9a5a4a32cdae6e0917358117614fd4d8694c5fc07873332f31819e93385dfe5bbffd02d2dbb22fbc43cea6579b88b649dd2cab4e8ac33dbc8d32416c6c71e984e4fdd24b81cb5b03f0aafceee4d3a936e108939ee4c978ca15c7f1606a61628c88361ddf129b626bc4137b6e3e7b74c7a821026a5da0b144210f806b9addfda0086945b3a81e546a895fbfde17387b3b88afa15f19047a2625d784919fa033a673a1bcc930859e3ca6719b2830cb02e654175ca8488b2e4b46182cb40bd25999bec5f4a329b55206e2c38981fc874dae6c6ee35312a8411ed724ffdb372f04c01abc23904c8c2fae1c85266f2eeb994d7fe5b4cd9d73aec430ad252a3e160a9adae405de8cf56d0f678dad22c37c358342c0f3c77e2af99760409136a1c1a737150ede3d39f13f3342853a54cfa6fe56cbdfcb7a35b071b7d71e588b8165c00faf8b90d5fad43b4cb3f27d8a87b8d928c3b7182103ca5d03ea439ec5afd1ae7a0c1b7be352ae363924201ccc08f5169fbba12688c14a27ed167401a1ec3de2c61c799be939812ed818fe1268f3cf3b17d1891d6055b398b015aed22e4ab75f4390a484e288b4bb42dcfc3c2ad64f26dc1c619ade4b835a938bb10ee1548d53ed7025d3a3246995ccd200e0c151cdaaad138c2238abe5a82ed7257e2a1cdbc9811062935ee65594d57d8924ac285c8afc34cdb1c3b00cff6718326433ea044485340890e75e1388ea8691bcc517826cf435377cb036a13f6698674d33005fa9bad6a864df609acb85253ee8fa3706bc43527e26365a7b16620ded7cb200d4929498f9c990c7e37f07bcf94ce19b810615f3f308bc84947f675ff6306f59469373820deae5cf72d475386c5aff1583a918c66bdbf45785dd68bbc2b4552a0751d69cbe74023bf1076abbbccb9d29a287c0f2c7c725ad773ebc0a596a7edadd541f14a4d59e2cb1b6895b129cb701f3ed404b244368ab017fd109cd5da68b38dbae133a40ad79aeee1fa6061f4c5cc855fb0f738f3043d5adb7dc536b7a9e4e6eb317593fca57ff50318e9703966398175231b2c1ad62617bb9ac021ac34d5c03fd9ef196e1494f49a5d8b35b94feeceb989c90b0d9728d5fce02b345c898711c2bba29daa66688a8222922e94abb4d0e1629b9fb7688a2b8a2aa251a1f293cc7d7ff750e72c28a2933e8345caa9d992111794cc7670476a7b0b4e2238bf7a51ce1faf9d195b1f2a4e7f50d18f3ec8fc9d53476d5680fc80935b6caa2fbc0bcbbbf27a5a914e696b920edf61784fc285d1cb56838a9b3aca8de8914474d02eab1364c829dc2f9e8c3f1ae88e3f236b30617d72cc5d96989b4a1ccf9d84c88b06efa5bc7089516e23ba3f22699669ca2d897503ac302f7a09b0493a8a70c8a3f8e7174aa5cc9b171adc5b6eebc5e4ad965403c34ba957df0af65df61c54d4dc2dba8f6b5954a621ea357b6e82cf01273391f5e38cfe264a3e2d456e795af2251da5bbe7034d442085931cee3759695431e30b7dc62858b0eb78047a74d052584ce1e78bd17e2cc1c78362bced2437cb438fba4b139bf7ec2f142654c6747afe21e55190a34257a8c602e4171a73b0e649a521877a90836f1583d4f42401853954a9f7e98563603da4e2f749ec4da027d000d3d4ddb038ffdf9c70d69a917588ebdf6167ef8ff8c32b465d7d5f5e2cf09ffa52a553e2156794f23e7dff83d38ca278925c111954b3effed0e26f9edcf32c9346537bb601bf44af665a77cc46ded797455c43cfef3fe0b57c15f45d1ae110080393cd00940f02db81e206d963c9ff979e411d3e64634790a777c18c7413a0227cc391d414be1f0ad25138fb829ee214c5a9fe6489053c289db7633f178774de9cf4da077aaf0bf36853a5e13e404f6368af7a6ba85db4046174b3db09e3cd6ffd13a6b7a4447d6243946921f9f7fb1d6ec3a92830ec867caf6c1e166339b7bbbe6cae4a6b2da5b63e2873a40ac2a8f547bf21d1090df148160388dfb6661ab74597a0fded8bb7df9558ebeee16c95400834cc09e8db930e1fb6a1447b179a03743dde5fc7534e26947fb60ec88eceafa9efceaf014fb614430ffd3e4cc02a05b00fa2a215e5573b5b8b022b9aa8bf2f0da0e4ba1f7d51437ed7d2b79edcb092edc8d6ef0ab9cdc946b08ac47bc60484c03d6fecaf204e2dc410213a44a42b6d0a5aaea3d70fad645044fc8a178af355eeed5ce34238684e2d4bf792b5b028572861da2dde92c5856318382bd3106e8c1932b1752e957b675244198a8ed4a44326a448c24aa344120a7dd988b7379e4ccb56684cd98e79f8c0d6637a593d89f100fbb1df727f19cbeb18b7003b80df6392b1a9f26af0c0c2a46571dcf83cb5b4631de78a81e475f5c29ad3eb37f34802331705cdbfb51f2a9c828d23aa24ec5f9b1bd0c3f607efc962e62b1c429cfcfba4277026089d56323dbef366f50233e0271480421e26d39b0877118267a17a21ef456312dfcad85e7537c8a2537ab992f3eb512db9ef09c22a097958407aac11fadd58cfb225f01b46a61eb7c50bb0672a142f855444128c9bb393f4c6e71ccdc92b73f367b6fe8e46cfd3d98d441e49a95aa35ea286d7dea8a5183e618d476da36c31b95dcf48319b6765531cf318722fd", 0x1000}, {&(0x7f0000006c80)="431610d59691d8d547e6c328443e2ad259233fb89cc0a4dac675a1b106d2a7d0ae8895c5a09d1e6093b525fab0f504ad4ae97b7835e47de87b05f6c687cc9249176e429164a46b02d47901e53dd798fe835f8e76a7675edcf05634515e9ceaa023d1752cef5a4772ba5425b13344401b07c849f2570fc9e55fa58bb584261eb53ae53845aa152fe399e078baf58f589400f6ee99e9146c7e88258bfb99033409e97aa6da76126ba0ee8cab1c7cfe549f95cfd1978cb5de46fda8e204fd93d2022f05c949b0cb4d74683bb6a56e8c320c50a58659c98e1e0b4e0fcdccd7bcd99e8cef24a0958eb7e823c632bea9e14b2dbf3da56b67b5ca71b122035368a4b3c632b988ef2262be1192177d8bf8fc2d275c25f352d7a9352217feb769d32f7b9acd3069430865621d8afbd48f87d506cb56326b698419c87a621927ab1752d3eddaa340adfbb8a2aeef63c78af19e1e7ba1cde467d3fd4d584c18ac6c29cafc4333c47f355340f2d156d523e9b8e26a8388c35e839226ffdb4b5a64346d513b670ca487fa7ec81381f255deb7d0244dcdc146107e9dc8e5b3aef426a9986e3777dc6052184727c8ecb20849bc6bf6df2f5adc236f02494d0eefb4855dda883eb51e916600fa3e790815cc62247b00f45bd79229a438e63a9e97a63d27ee72ad14798a560dcae1b59822f02ec12208e9d64c11d7781d20b4ee93bcf67074bcbcea1d5b5f4602fc39d21b7342b028f5a0d626336ef3500ddf0fcd99891fcdf151716985eeae8b16f6a494e4db49211366f70ab515d54146fd365e6e96693abdbf2f38d41cf5ba08ffe14ed9f45d6e33e8afe25dbcf62350e971df7786c5ee27f66192b2a27af8abd2953f1041f1ab539c5666727d1cf6f81ab947f78729bc60b2180c1d88fe21b18793907054142972fe8651881cbcdffcb18f87e91c6583626d643b04d8742308cbb92cb0d96480316cbbd66ad7fa80a504b843e75aa75956970201d00812c94abb07621465b9a11aa3457d6784bd0fc323d8ada182e64a9ab52e410fd8d4ba68cad37e0f6033b74707ae0b68234022375f3dbeadecc7c55712d250f2ec4d8b89c1a27b3fd9726fc85c51baeff2ecfbe042377afe783f44faf7cfdcef7d1ebe9e81474b833ef8acfffea117efdeb83daa2245b461b0b305fab1f2e62e55aa64f82e21fc902d404b86ea7e6cde3f84ab8479b103512c5c0222f91a696b4dde59e2627c885d027fc11c4f538785047a242eceea94f679142126ee6986cfb686825816a5b41c870efec32b094b78870d17e92beaadb71b6c66fc60d92a9135177481aaa1f34e740e04a2de0f321f6bab72706018d2d91d032e7d749f8056a02b3a73775a8d66d1138ac3da1fc65f549a39c9250662c649d30add82b0e9046202862345162e8b58520b98677775518e759c9f1b79c8a5b8f9a1eebee30f5e816b67ea241246086a58bda4876daf7462ee96d897b156009cb3fbad7d7be320efed2a74e85804dc23f43f3856ce620594b605f83c19ed88ab2c6a5a14782b6efca89af2424535e222d2bbefd1fd03bda773cb18e7c2def301a1fc077008dacb244281e557952546025552a2f8204ed75e9dec38f22f15127dcf70128db8d8753490901bd363cf6f3f9b27871df81f9ba5f7cd7f8e74134a85c1ccbef264dd686e421517cb79c3a1158a9f4fac4237e68859685ef28a1979136235fb0470c1d59a645745059f2813cac857cbd3db12fe729623552cbf5114908e733ddce63d4e5e754bc15db576b7db72450964d43806c141fe38b2cc004af1dcb579fca27c4dc6b17d68594829a209392b0eaa954ee65801116ba367faa1b0c4ae54a7883eb7872b3952dff9c5bcba3ece9ff3fbe267deffb04e4d95fa86c36e8ef56798de3bd94e1ecfccbb284ebc627372081fe14fae1fe0f93001f222a9e82acec221f0ef0b32fc8cbb6201b766706958592af7cdb71f22c41dd1939d8547504972d50fcbf51d3b6e7f3db2530a29cf6fb4a5a7b62e713ab26087c47f224595d9790427744609effca18b05408ad047b63c5e421dec1ac78fc9cfcbf0fe17ac996915ad84c5e13bcada7be81bf8d57e844e8bf6c64fbf9225309996a228ebee453559abc984f5ef3afae0dbd8238f519f46275ec55b765f76ab0a6d9223426fa5e28e666d1e35c2a9d9a61b7558f67e93882ec2e7f93921de802b2146a538b7a0ae2aebee11927488c530ceccb43b8888746de89352c0f55f5842c00f90c41b514fddff456a6423f6657b2152d9ce06f56cdce96db076ee3a489561e1485e513256177d62ba1609040c77095d7a678cb13b719652dea20a933d12639d6b611ee7d33c4be7bae2940b1d21aeb791c57fa718d5daa481fb1578876dc65e807584e4e7e014a45fc8647cf79669e752de3567d67dc2cee81122d6be7f7e07e286d8b7bf172a665f6eff1e24ee15ce4650baa5b9b8ed0c22b1192efaf08d65eb69e70c3c27012a56e0da6589ff09d63d50720c3d54f550c052747daf0e2a619aef4d4403248fcd07718d2bf06fac821cc3911ce81975eec68693b534f21f2fd78df183034a045d4b507e277ea93eec0ffc048a7bc0285e8535e7a14717455e5a2878646fb2bc46089d815214cffa9c81d5ce802588e3c086689cd6c366c5132f8e480298e82ab01f2edf488272376b1b736dfbc7a4fed564cb81ad6c0718b92228bc34397a281d7e18f4ea8d03ce1ddb3a0111e2b6a329e66b31e1a117a771251fbae63ff710b4f3a0cae234b0bc82c0f4e78d03c82fc2c340c2947deab813d07cacf789c76dcd01e49edd39bf37e4774bdf32e2b7e4e18681321cefb88c9b15979a2d7ac983e3295d1f996bac44988fbbd89bb2e4de2f9c90933aed4e7f2c1bd329ec1164c83c5ab04e1609b44eb532e1a68aa346417ce7ce0057fabc60aa9ac95cfb32959d3ad002516e311c0e17045612400efa10531dde76afc1144c67d7fe4239e4268aad53746fe28123f069edacc8be8bce7717bd68142d5c82611ddef17bb00f8cf601f64912767cdedb1e09ae21d0315ad023e9410c5f73a4b9dda3d2141b25b3930f68d95447dd3ad3cdf4aab131a1220d7add73774d2d4e125ee2f78b51eee2fb0d8e0a61965cc36a80402b1427c32965c02be2da1377959204e111263da5812bead8cd45cf0b8748327b5c2de8681f194c6f0c77dcef9c5da3469633b36e2e9e9ec41a8d4f3e9e6f1d60943eb302556e05de3fd6a706310ac40d139fbf7bc6893ef8000d8ad05635d10a503a0069d333b14d7c870d9688bc313c0a9965a1cdaca0eae6ec2546108aa7fb546e16b31fa2c0ece9304b5fbe82f981eb634f670f246634f744ffaecaea8ab74a0a7f8dda774bffc8f21ef0ac848a583af1cbe21b9b7e894b93059c09047de1700d63ca847bbd888afe47689b0eb3acfa31251b36b81b7467688a09b290c8daa85d8fdc4e121944507a9c5f23992ef66d0e21e35d6915185fc58c1c27ea9434b078b1d9d2a382762be024fc22d6990f882eac10edd984ee0a2a5c85af00ad784e0cede2a35790b65a7ab2d624416a5b37906762913c986464f705e0beba39dd47a8156a12578509f5fb2f66e50f953d1b7b7eaa120a61a5a9210b57c3b4cbb8f5a0a388ab894c7eba958b75886a054db748c4c7d7a03595ec5ac1cdcd6d347c1b3c7c1c85af33221d03dcddd2ac9f3e9682e9df3602cb1e0b839cf8a35712b559871533453fc0f7c732314d2df6065de0361be45616d2fb4edffcd33f61b66bfec594122498824d89a28ec1fdb39e473e6b40f968bf4bbfdaf92f6376b31209be3a3f1f2b8ad5faea17299e561967c61ce142b666fb668db332fcbf4f8c86dbbe2c9aef5a75267e2d8d220e7e45c986af61873dcfd499cb89a15c03d840df4731596007e011c2c5fb6a9d003e8ca4808c6932a175b2bf690aa258eda9f136140c683289700c1a0bdbb22453f21d9c30689896f76620e68f6b8c58f2d16650c051423ea710d335437fbc99a1e9c796dcdf334524b495611ac489a013bebbe5627ba0b1e3e12310410dc84a5d2a70b3caafcfabaaba24734daf6c417a4def0d347d49843e4e3ce10bf4b28acd03686526541902dbbd8437c00551038883ff015798e1a0ee1807644f835b75848f709dc105646661bfc4836939afe89e4ca4aff3b5541e5f9b28ab35f02658eee67034fe6d0ea24f4492becd3eeac5f489e0d1b2d09c6e1c2eb1fb2f80d122a844d9e5c1cc8a0fb68ad9fe7112322a2b12fc263994132c0bb88d5c24cad00af8118becda78577496c0fc852a2e8574509006674cdd1c4680b38ef610e4a7017c86d8e8a3232d4c90f286257ad167ce4b3e87524da91972699ae896117090897ac86a76d9effe522c8194c23f6edc4248a29640e4e9cd15e97c444cbc95765fc55428161df3e20242207a87753912a80918644efb72c539be02402e6948dfd3278ae32c28b0fbf49748afe18a2c000a4d82ce215bc7f7a1ad640902f25590c78cf2894db89f5ca9788f4e9c50a9442942f1c3ecb8fe8180ec10f2e6a06c168dba4bece345759e3455626746b7b88c1efc46f9258d858242e88f048a85ff6f37f3ea8f3db1e8ec81c53200ad4a0e7ca46109fa45e487609c334e0c268813a0f327dfb14364a5d40c94cb044c371dfa3481ac2495ab14b877a02095f670415824ca6af8cd832e1fd3ac7c1a1caa42e06751090cb3a6cfbe1da60519045207f4754d90070e01ae3057423d34b420b3cd9461bba8b2a7324f06333e445abdaa0cdd80d54716cc4e1c6e8a9b3cae575fed6309b78a3d2285256c734f3a8562275d669b27b495b69409bc9fdee8cf9af5a1d1a48edba37e6befa3046a7ef36146c6a2f31518ab008e0d1c5543c624c67cf6a38547d7ecc85850408b9ba090a919f4a253b3a58d9c88aa2b3b11843f960dae91dbaa5820608efadd829144733b6a425a2264ab1f7aa6885eff6cc6c58fd603f7bbf2ece7d69389db99cc974f30a6830133d1a6bc4a48eec15bf7d2766037fc3f3fcc8ea4f9b9c334a8bd76393c5d76e81bf1e02202b9b708bb32de63345a074cadaa3baa9e1f9c3c015e77499d7c6488ab2beece7f31de74a77f048a1685311f884ab741a215db3b10cf9a587ef0bdc5a5eda8822affa7e6c72fc46e88431e5c6e873c94f4e9949c2219924c92fe606ae9e19f2e6150d1b55ff974b457caf6a9a42379013ed00672578875f9402a7e1b9a0f05b4937dbe1b24982dbd613d890e448a55fd95cda3f699998f83b93cdcbcf8ae5df4dfab5b9360052215c9bcc2febab701a0ff62c1682d7b3fed8875e0d61bbed780d15c8391263273ce9f3c7490dcf18606e7724835c3086eb91b209e8fe9443fa487333c016be869e8f975cfc6f3f6be5fbaccc9dad24dde72e377c5b78a2cafd41a97ee199710042ef2c87c722f14176ea785fb9f97c0bd6363a492fb373de8ed938aa0626f020bc1571a21c591a670d255dfcd11c3b4a68c9eb77858ff398d1ce676217007141cd17679253e72c9825397e37647075d2534b0e068e8c66754bfc213bc41183d353c8e7265b273f69446fe1f6e78bb78fdc7bb4b40433aac317ab06e2748b08b3c95ac0dbee0d642c1ade7e384f15a68f2eb676d1b09e38ca1224cb13c2e91d548e6e2a7a0e57383ec807a0517cee39a8b853c26225d256f4cb835091e6046164bd9e329478565f76156ecc0ff87fca425d41565de68a5f59d8794f48f9d798a2d353512c424d6c13108fdd773c79d0b781301ee18227a99845be765f704eb56", 0x1000}, {&(0x7f0000004d80)="9108a062e24c0f63147570abdb833f9835229ed7da3492152fe3fa0b5e864c3a669c4dee2325be801d663a2e06523c68db1b54973862bb1b43a02909952b06c49af6b8248ede86d6aaabca2d7d4e582fec62403052ec850d4eeb327ae6c05a8fb82d16920111f2df5f50be9f9086d3b8f3fb11dd61c51477dc9416f8bf02144c", 0x80}], 0x3, &(0x7f0000004e40)=[@iv={0x108, 0x117, 0x2, 0xed, "d38f8caa38184ae4bb866ad364cd7ec4ecac8eca4b201ed77b2f86ca89912020950be14e5f472322dae5431189c713b14dd14cbb77c7574fe25032cfc6819e718b883b7b75e8bb0f8b046244e1946aabf5d3aadc2420296287129e7423f3fbf3325438ea2837090de67211cebcd234b610b4586bd3ccc02141cfc3ecd2da7c3a84bda9cd25802b4415ea6168245e39631e6be6a59ba31ea9237c0dc5ff8f5de11b045d65a7037a17a16cd9f586922dc7acf1bc1fd12398ee4bb237a10f4b55981d87e41e7377e7080e7d895281e4d362f91ec4c342da32ab4021ab5f915548c0930a08ed82f2374a23fe21b364"}, @op={0x18, 0x117, 0x3, 0x1}, @assoc={0x18, 0x117, 0x4, 0xfffffffd}], 0x138, 0x4000008}, {0x0, 0x0, &(0x7f0000005240)=[{&(0x7f0000007c80)="4ca0cbc5466a021b0ed44ab1fcb14d341b5a34aca8452e50d25bcf22d6d1a84653885c976e90f1e6e0e3a9c822669c7d60b14d4bf585805c59af6d3fc476efc51f0b2d91760973b21922f9b0842c012551bc99c2421b23561792d8589b368d1315caf52ea858af941d3fe491667cfe7ff0ef8e12f7423fbeb43016cc264e6e58455e68f121114227346f8ddc09133bcc05283e31e6ce1bc08f50de6a6ea240c9f3bf895a766bf52de4761094fbdf3131b04cb430fdcc47c336c5846484bcdd9babf039ca5a1ba8d0b180827fe90ac9045c88a0862ab46431e3c60d94356f1d5d724476444a376c8adad6c3ee03f9e0788385899b6fa18879396ec2e3bbf0c513e8bab9082e50c51909505e946ff8edf6a5113bc17e6bf8e1a1e47ff57b2a2f9b8c4a4796ecd9981ad5dfa9bb5450b769a0e24089ad009202d8bf42f987c2f7225019a4485cd2e74570d18c0155bc36f2d96bf8985a04fb6186a9aea3e89c76e5e1d7276d27109ae742c421a2a4bc55658e4450bfb83978d02a292032f0ada0215ad533a09cbe55f10b857dd143eea6bd3a969137b8a2afb3b2bdaa0952f2cf4b58098a83127b39c6b563fabb8f8ecca351055dfd7e6966e86039aefabc33e83e18225e82edcf00fed4ea1def38517cffdc7ab3259d3498e2546e294fa56a360e1947ea08aa88edc7504eacb330d88dc780ee523b44da954ea5df3329950773f81af05e7f2101cf50252a2be1b5de88b45f9af15b187aa65e3791347d07e6e0f504239a8e43633e5bdac0a84237555cdbdcaef6a728422159ade0e8973d4787018c9f03f4ee017b80a8564004b8486e49013df35dd513211b6064177b58205b8293a4303e025d1f0a35d05aaa5436d092401583cfb255606c4211e6f7744c02a8edebdcbe50a91f58b3d4f143998d2b00d2a72128251c1bafa179c4414832b3c6c9672707412c55c8d5ab6e2e3553182008f1cccd9f35d151598ebb96f11ca2e16d463405552efd842744cb1477bbb3d82ea63555c52c8eb42febb6443ea6a9c25d456cd536df5ccc51e14a3263e0363b365878d0d7c6487d172eb91a5570a180860a48c8a739077ac4b5732e08c6382444ce215b4b69c50615a98faf9e937934a498654b77c04a57deba101e87d0eb3c7ea4a9c442a7952221e854ddb155b3ce1f52f340de94ccfc2f71cfeb901ec00b87163a7296767bb5ef6a39fcf57e5b5603ba4621ae9b2d2a6bbff8867346160193cf66afa9fb3ec1eeb4c6d6d584d15ddd88ac70a94a6f364ceccd3b472cb6a38a2e748abfe2acf818f6cd780690b07b1a76f07dfb0c628c802d7a19fe7e258c31c200810c3d8ddb71834f56cbab39f9f72f5ef3dd391d14157a315f5346813026be06ec3fcefe28ad4877bcc217dc0dae6246d2141a1e9a3aacbcf932e564e358ee7d2577e6920e89fceb9c25322c8e44c0db8b4c108dfee112e948a92113f286e307c7320d2167b373b03062f0055046a3510e2b4c679af03299c7001a9256af57be02800fa39fc6070ab588bddf058468ad544d306ee570cd9b84b63f1c34bc728e371a74e272dc629a64f30416dafce5955b7465c6a0df1b4cefda7e32baf1f5155e9c55f965bca03cff6d533c1a242e5da912e856198297f7fe2c325fae0cd330d5b2cba31d1734830a83cf2e42428e23628b3b63544e59fb86c074d9b1b90dd4c5911c28c3f097a9d33a5a01b831620003de2c47c3abcc2ea48d9b3e7a0f7bff11483ea66026fb71a5ac55ef6f1b037c56955f27b2fc9f8c1049ed18aeb5cf8a6cef2fc9bf13feed052ce6ac2d4b80f6bf51cdae67baf72b20f03e533480108d7ac11e61e68c9ca8db654b038139bae95df5309249ab7f8827eb15c934f77cf8163bb9019685a677d86eaadcf09c1c2df5e2dcdbf481113a6d96bff8155dfc39841a95ae86142ae8f1f558c0a23e053b357121eea7285091a32b2c285ba91d41e1c53a6bdc4a6fcaeb8354b2d65356408d1fe7add81fccbec3d05890bbbfcef39f50a70b7701d675a953df007e01ff130e697fc8d38ee620142943dbb5597c01b9d906f052bf8855f0041ccbdd6c002f8361bf479482e81e1c19505b94879711aafe1ded367772a203285ff91b02bceca85cb775f24cf337b9d162f8799d908d3c3841573fa8ca1892a1f9693225a54d867851bb49ed8d18ec835173ddef4b0103fcdfa820eb0a4ef6cc2a374b63df036bb4de327661bdfcfb22e77c2a3f3fa7292cf4584092233b4a3fc82278963d4e0c3059543b4242a0aabed28d35df1941a4db74eb65f16cae60a0c24e3dd539888695a7938f99fd739213084a468646cd6d439389a21bd880f393317c6ffefa2c6c2e2cd55c0ee337cb0515ab622e1158c19cc28673df29824781b05e14236c2240d0bf8c7c459d5b2be97013bdf18f372e824cf1a435294b13183f1dd79aeed8751fa0a7d06e0c78306e37d6bcd2815ecbd85b250334bee88b70566b80c3e441249355366fcd6f70ce35870d05ee32919e7400d48c26cffdba88ffa53b07c1f319789b95a698a450d9077bab2240fa6bdfa3a114ace196615247840aa540f04ee02f44a254ec19afdc920aa447c663f14ba95a3f5872d362a9511156982dbc953c8853a654df1e3385385777d74821a93d3cfa18673a3047cd906a00d8b5d6d893718adca65d1c8d9f9ef3f7ce4d63b449a4f68e202c42d9d8de168a840170a3c29dd5dd3521a3006c085d6573ebf71067262fc35a0386bca575e206477e5b073a9d78b708851a9719067ac5bc115df03dd67e46a0a62e8e5183a91352c32cc4a85071f419f4c7cc763f419bc7f5dcd09db08b7cf14038f7fb84210d2a12130e7db3853344d0f382b9e400112a4eb68491cb9ecd38464219c4ad36060e050241fb1e3cd3637346e326fd64b4feb7087acd2069ee552eee9455f5d25817236f7dcc9fea359694ffb37aa3a219913c529b4cfd1c7b1ac68e1dd73b8c25a703b20b4d031e4c8b4ba98c340c4a8bf429baf8cd7edf924f1ff8b0d57fa28a9c7d95a8228ff84405f5531dcb1e0448f024b7d9d5721aaac7b1ac1359c9d63e5bae51cf6928d66defdd283efbbf5a41c9c3a899cfb6f6156f585dee40342bd5fcb13a0441c07b6b069da3307a780318e560bf00e3af89dda874337a37dc54b79bf599e386e4690ddf9e19c1a25acef09a989099073cd9130ff275ba60a8bcb099c12248feca85e78bd57704fc1030833b3c6bd4ab4039f8136e533b04694215349670f65551581a13d5a26e0234c1f4f79d3c31c4480fe0098cbe5afc661b3c4fed907e996dc73a5c61a61e028ccc8f2b99813b5528c4ee6a557d8b23d00e23dc767e762fa2348c4989365d535f489114da95517f4dd7ab725b49641534b18593b51cd3716913862bf36fd1b7502ac5ca96e8754bc4a3a4269c56a6645029a988e172980819146e97e16088a6646d1d05be9f9d79c3672edfc240d28f0d16f26e8558e521235c4bf58d8ce32698ac031d1e5b77403cd9e2a637e471c1e578dccebb3853726e8c2f3bc501d2192bd32305ffaae3877c728a3a2178f29fb9e452f20a3c5237dd190e34b2f3e2c3bd0e309c91d2896f2e8de0ea11911c7adf1677019743a67223f7417e2e056c417e2eff269287aecdd397f9c14164f3f2ab4713b72462e581353c29d5d05935993d01e2312bbd2e2f24e9b97a9558a1a3239af6a13ea8ffeeda5c09be5ed7377024e1387193fb1042ce6f6cd38ba3fa0a89665797171c906fbc213a1eb93f706279e9b03d9c2fe9f3da5869b41a87c5c78aa841e83914b51535fa3a39254fd14a0769bae6db7cbf9625952e3426ff38c432891521b8caeda0980ff37a894f75deda990a1c287e266301b6ac6aed3e6162c9224880b3c67e063880dfd8f954af90c52d690086940ec58c1854e0e8d85b9cef6a2fd156ce130c5eea6267635f73b2f9faad159ba40c092df3886692a13417ef930ae2d695134aea9a937d8e410c957df7fcd8c0972656fb019fb6b1a83d489536caf4c0750d9237e17798db02064dc09ce2ac48bd6a215a494c1c4a8bb9dd56ed9ba45474bcf8b50c758ef19955b2d3a6d6320d13533f28d36f781299dba0b33916cfe28a50619be9824276faa52ecc7d5ef4d448e0d87f9c2ac7dd133bd77eb5389dce1ac957bb14b4ca5b388a529319d4db6a7569ddda24648fb6b173a2582a46316452ac61cab2deb86a8352970cc39eefb6094a3a7843824ec4ed64a9905f2cc038fcef907787e19fd9bef5b18d8eeddf69291142aadda67c2ea0b900b96eb64a731e9c27dafab03641717ebf090d6fee71f447d7748dc925a835ef81c8ec68336e80c62f6678e815f275c8ccc3f3b1283e4ed55024cd40ca62ff02f10e5ac6b711140994e22b3d64b884c1691cb93bc99fa3fd8a05c11b9810cda1c031128eece89bf5483b697eff102c4f77eb878a2c76dde4eb8595f786200e1434c94ba76d76b48576366a519d7d7edb5acfb260a857905724f994aac4a669ec5176e0358257bac91615a971fb2012ad79d961e7645bfc4aee41748f4a22fddcadec8abb138369634439e3e3c367ed229784919cd92c74947ad137866a5b658edcfe635bb50b4779065cb3623b8d3450dd92409dc8ef22f0cae8e03cfb279f85805acf7627f31e3e473ce516cccb00eac16a745591c733b4c2bd092fd6357935b0491f2e28b95e94dfe7670753741d14c91ad30891b194e3c8e512e2d65354d891ee0b7464efbc550ed004a63a1f128fc81f775a4c4b1fd995ad360c253c1320b1d34c8b9297b2af5b7468287ff8847f9bbadf282991f5a4223bc3d010a044433f0af66960c28a28cf85191385c82da3912cc62468049d0a8177d6200346bbc7719099a7201e64238e06712c85c6adb7aced588d7d40c1db246061c5bc86daef036774053bd918f4fffab60380e478b0af54c3bdc926d6a98b17a95bf55d438ed5d09b617e647dd62f64537afc6401945a5a5c582ffeec1ff9a3867a3f4fc90bce92cbdd6be3b979abdf9caa9834b1f119992eaae09c5dfb186ba175cd7e3ea243ebb0a5b9e473957f749bc2995987114274e80ee53204990d2369b6cb41d8c7a4294793418ea1255b50c7b2bc944905651fdcebdd8e00ffffec7780d4dad8124937d6c1ce6ae847b4905e6c4d97bbd32abe433afc87df8a23e83e8dd1b4e7054788231297aa57240a9b23d9c38668100732bb1fa7ae97da4c9dc6e37585ca1fb1ebf27d357c0e32edacb43ad9880825e42acc6346c31bef284f22093c53686de9bf7c4575183509168ede60d5e8d651db8af1a17b19bd037a95f23c248ff8ad7afbfc30337dac4de0f56947d974389e5d883044b80ebba1cfc40e17206651520b3b90dfafe9c70335a214819b51fc42e42f793039672da7ba1f6ee6cd95105fdea355412d4affa134fc4f65dcbba0d627d4878ad22e78ee66b7fa051dffe7640fa5d09d1a2cac0af5c085a5430091b659e5c71083637e6c0d3b8cc3da5c1ac960349603dffbcd483ce402434bd5c5997cbd8c6888011fab04bdbf34a2c6cc6c1d49ed6bf35b1cbe718bef9202eec8b1201a90b5c32c8e1f0a1d7ef92c175ef9ff12a6200e6403a6ab514300028715933f28463fc95556a4e024b8da77eda8b9f930b55ffaf97377649178a102c0d20428c8e8a5d66bbd5845da5eef01192c155cd1f4516306418c8993a8780451328ada401f6535d1882dc5b6166c8d953ca92ff72f51621d324aa741f094dece1efd4063c80d1f71c5ec639668d0087541ffa95934b509", 0x1000}, {&(0x7f0000004f80)="a27022efca4646c0b4a1611e6804f38bb9de5844c2204975a6c23606ef4f04a852ba31f7691e9342be6cfeff9c", 0x2d}, {&(0x7f0000004fc0)="ae7c3bc4f40d0c5f4825c332f33cdd294baae624f78b4e34638edf769f3cf6e6b4dd59cfc18a7808852d63464cfc979f1fa07f598e2ad80fe10558d29ef8f529e1da41bbd3dab56dda68aed1436551e96fd1d03a8e761dc419dbfe83af0f63039222d118dbdb9a5eca15ea15a9b19c0a343e91057133c0d3783c6bb2e6da83975fa21e4955eac61a8c4bdb7f4f2d8fc8f368b322756929334718f09c00fbe18e294f761efd", 0xa5}, {&(0x7f0000005080)="76cdbd55e2430375d979f1f0f596ba20d3fb5a75157e0d5a99c94b2de7473c07bdbca93e26557bdc7b42f73db9dfc61d8a9afbb0cec655d6c294ad467ec3f3c0bbb1d2949b1451d481222532b9fabea7fdd706ee15aaf1d46c61d1525dbb719e7e4e4baee8640a8148c3fed616f050f08fc255fc9a20e49749ba90b3ba08eb2cd7b8d00cc19fde2be20342368b62609fdd0cb73574c4ebfd571aa59d58768ee7b5b1a372fe4a03eae9678085bc65910f52a2783135c2d15e24fc94aa97a9407df8afdedd7e7c579fe895661deda075161c577d7316c95ca1c8f47233cc3e9fdd1d65a85f", 0xe4}, {&(0x7f0000005180)="f4234f3520381e9be7351d019fabde4d9ae67262e039d3632846ee38ff0004a28cdf6d4f79a61362b00dbc0c86aa38e8e399925469dd9e4fb351895b556f89f571eb116a4deeba4b64d0a6617fa9e4adc0b6da31f857e6f423ee7eec5fb71053de696b0cb198b14dcd00db1608652341168a988fb3c7f5a3ed9e2d29680c47f498155df6b0217bb8b102465c6a1fae506e17b094074b58f0fa", 0x99}], 0x5, 0x0, 0x0, 0x200000c4}, {0x0, 0x0, &(0x7f00000054c0)=[{&(0x7f00000052c0)="0556d4d807611313a785796d3eb788120fbcecb94d54b3b03f794aefb166242c7592c8c4d93415d46b6d12ec1c080b218c36e5dda30cd8267e1975620b7d426528a765a8fa437f9d", 0x48}, {&(0x7f0000005340)="697c41474d015ea44d80e275384657032c7bcf640b08fbed8b764fe3ef4504fdc7bd6bb6adf15657977c034471e70e640734aeb1852b5540712dc5f8d1b63847431c222b98c69c9d71c8f0bd3c4230f32054140a5bc037b4d4d8b0ab25b1ad39aca1296afca662811a3c3bdf1aed5b989e7d2de08a0a810995c39bbe30f84c69991b907c888449821677e9f9a33ee719a28e62608adef75d83766dab8583c09a8accb00fee4de612757a954674", 0xad}, {&(0x7f0000005400)="cd101bb5b3271eedc18551a1cd724d6288adba2f6d01ff157cd5df67022851764dd9a089db70ca2d8b379c9f6199a384cb377464fb4366e3b9549c8a116f8c51aa35d699bfdc952211b55bdce5e71ef8a1440fc078f08a08048507fbb5128e5babc097e7e5a677ec8c023286e836905aed15e250e9c76d8dfddd18c7142407d65becb90060449a96d4902ce222f0646603efca15ce72cbca568e85426be88095e80b", 0xa2}], 0x3, &(0x7f0000005500)=[@iv={0x50, 0x117, 0x2, 0x3c, "7b8f2e991d4a45c9e843a3492b93ded39b0af5b14bd18ffbd393138c4a367ac5a9602a73de1bf1fd247eea7d5a301ee4b9a380a745da31344d33ade4"}, @assoc={0x18, 0x117, 0x4, 0x5}], 0x68, 0x20000040}], 0x7, 0x4004054) (async) connect$inet6(r3, &(0x7f0000000040)={0xa, 0x4e20, 0x0, @dev={0xfe, 0x80, '\x00', 0xf}}, 0x1c) sendmmsg$inet6(r3, &(0x7f0000005a00)=[{{0x0, 0x0, 0x0}}, {{0x0, 0x0, 0x0, 0x0, &(0x7f0000005bc0)=ANY=[@ANYBLOB="5000000000000000290000003600000000060000000000000401000730000000000a0000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000020000000000000002900000037000000000000000000000004010000000000001400000000000000290000000000000160000000000000002900000037"], 0x100}}], 0x2, 0x0) 04:12:11 executing program 4: r0 = syz_init_net_socket$nl_generic(0x10, 0x3, 0x10) sendmsg$NLBL_CIPSOV4_C_ADD(r0, &(0x7f00000000c0)={&(0x7f0000000040)={0x10, 0x0, 0x0, 0x2000000}, 0xc, &(0x7f0000000080)={&(0x7f0000000240)={0x1c8, 0x0, 0x200, 0x70bd2b, 0x25dfdbfb, {}, [@NLBL_CIPSOV4_A_MLSCATLST={0x64, 0xc, 0x0, 0x1, [{0x34, 0xb, 0x0, 0x1, [@NLBL_CIPSOV4_A_MLSCATREM={0x8, 0xa, 0x2dfd}, @NLBL_CIPSOV4_A_MLSCATLOC={0x8, 0x9, 0x3013d865}, @NLBL_CIPSOV4_A_MLSCATREM={0x8, 0xa, 0x76cc}, @NLBL_CIPSOV4_A_MLSCATLOC={0x8, 0x9, 0x66f4c801}, @NLBL_CIPSOV4_A_MLSCATREM={0x8, 0xa, 0x74de}, @NLBL_CIPSOV4_A_MLSCATLOC={0x8, 0x9, 0xbecdb5c}]}, {0x2c, 0xb, 0x0, 0x1, [@NLBL_CIPSOV4_A_MLSCATLOC={0x8, 0x9, 0x47c3498c}, @NLBL_CIPSOV4_A_MLSCATLOC={0x8, 0x9, 0x1df202c4}, @NLBL_CIPSOV4_A_MLSCATREM={0x8, 0xa, 0xb791}, @NLBL_CIPSOV4_A_MLSCATREM={0x8, 0xa, 0x780c}, @NLBL_CIPSOV4_A_MLSCATREM={0x8, 0xa, 0x3117}]}]}, @NLBL_CIPSOV4_A_TAGLST={0x1c, 0x4, 0x0, 0x1, [{0x5, 0x3, 0x1}, {0x5, 0x3, 0x1}, {0x5}]}, @NLBL_CIPSOV4_A_MLSLVLLST={0xf4, 0x8, 0x0, 0x1, [{0xc, 0x7, 0x0, 0x1, [@NLBL_CIPSOV4_A_MLSLVLREM={0x8, 0x6, 0xe2}]}, {0x34, 0x7, 0x0, 0x1, [@NLBL_CIPSOV4_A_MLSLVLREM={0x8, 0x6, 0xca}, @NLBL_CIPSOV4_A_MLSLVLREM={0x8, 0x6, 0xe7}, @NLBL_CIPSOV4_A_MLSLVLREM={0x8, 0x6, 0x23}, @NLBL_CIPSOV4_A_MLSLVLREM={0x8, 0x6, 0xa6}, @NLBL_CIPSOV4_A_MLSLVLREM={0x8, 0x6, 0x56}, @NLBL_CIPSOV4_A_MLSLVLREM={0x8, 0x6, 0x37}]}, {0xc, 0x7, 0x0, 0x1, [@NLBL_CIPSOV4_A_MLSLVLLOC={0x8, 0x5, 0x5dca46ac}]}, {0x44, 0x7, 0x0, 0x1, [@NLBL_CIPSOV4_A_MLSLVLREM={0x8, 0x6, 0x9}, @NLBL_CIPSOV4_A_MLSLVLREM={0x8, 0x6, 0x5}, @NLBL_CIPSOV4_A_MLSLVLLOC={0x8, 0x5, 0x52f76072}, @NLBL_CIPSOV4_A_MLSLVLREM={0x8, 0x6, 0xa2}, @NLBL_CIPSOV4_A_MLSLVLREM={0x8, 0x6, 0x9d}, @NLBL_CIPSOV4_A_MLSLVLLOC={0x8, 0x5, 0x58fa5dd2}, @NLBL_CIPSOV4_A_MLSLVLLOC={0x8, 0x5, 0xd41a9b6}, @NLBL_CIPSOV4_A_MLSLVLLOC={0x8, 0x5, 0x5d506357}]}, {0x4}, {0x34, 0x7, 0x0, 0x1, [@NLBL_CIPSOV4_A_MLSLVLLOC={0x8, 0x5, 0x7bfdcbcb}, @NLBL_CIPSOV4_A_MLSLVLREM={0x8, 0x6, 0x6e}, @NLBL_CIPSOV4_A_MLSLVLREM={0x8, 0x6, 0x73}, @NLBL_CIPSOV4_A_MLSLVLREM={0x8, 0x6, 0x89}, @NLBL_CIPSOV4_A_MLSLVLREM={0x8, 0x6, 0xfe}, @NLBL_CIPSOV4_A_MLSLVLREM={0x8, 0x6, 0xde}]}, {0x1c, 0x7, 0x0, 0x1, [@NLBL_CIPSOV4_A_MLSLVLREM={0x8, 0x6, 0x5f}, @NLBL_CIPSOV4_A_MLSLVLLOC={0x8, 0x5, 0x2dd16156}, @NLBL_CIPSOV4_A_MLSLVLREM={0x8, 0x6, 0x97}]}, {0xc, 0x7, 0x0, 0x1, [@NLBL_CIPSOV4_A_MLSLVLREM={0x8, 0x6, 0x59}]}]}, @NLBL_CIPSOV4_A_MLSCATLST={0x40, 0xc, 0x0, 0x1, [{0x3c, 0xb, 0x0, 0x1, [@NLBL_CIPSOV4_A_MLSCATLOC={0x8, 0x9, 0x39ec9455}, @NLBL_CIPSOV4_A_MLSCATLOC={0x8, 0x9, 0x3fe3a589}, @NLBL_CIPSOV4_A_MLSCATLOC={0x8, 0x9, 0x370bee40}, @NLBL_CIPSOV4_A_MLSCATLOC={0x8, 0x9, 0x39ff5a54}, @NLBL_CIPSOV4_A_MLSCATLOC={0x8, 0x9, 0x56c9fd5e}, @NLBL_CIPSOV4_A_MLSCATREM={0x8, 0xa, 0xa6a2}, @NLBL_CIPSOV4_A_MLSCATREM={0x8, 0xa, 0x1d8a}]}]}]}, 0x1c8}}, 0x1) r1 = socket$inet6(0xa, 0x2, 0x0) ioctl$sock_SIOCETHTOOL(r1, 0x89f0, &(0x7f0000000000)={'bridge0\x00', &(0x7f0000000200)=@ethtool_ringparam={0x10}}) (async) ioctl$sock_SIOCETHTOOL(r1, 0x89f0, &(0x7f0000000000)={'bridge0\x00', &(0x7f0000000200)=@ethtool_ringparam={0x10}}) 04:12:11 executing program 1: r0 = socket$netlink(0x10, 0x3, 0x0) r1 = socket$netlink(0x10, 0x3, 0x0) r2 = socket(0x10, 0x803, 0x0) sendmsg$NL80211_CMD_CRIT_PROTOCOL_START(r2, &(0x7f0000000580)={0x0, 0x0, &(0x7f0000000540)={0x0, 0x1c}}, 0x0) getsockname$packet(r2, &(0x7f0000000600)={0x11, 0x0, 0x0, 0x1, 0x0, 0x6, @broadcast}, &(0x7f0000000100)=0x14) sendmsg$nl_route(r1, &(0x7f0000000040)={0x0, 0x0, &(0x7f0000000000)={&(0x7f00000000c0)=ANY=[@ANYBLOB="3c00000010008506eb9afc4cd8d06e754a0081c5", @ANYRES32=r3, @ANYBLOB="2377f292252155b21c0012000c000100626f6e64000000000c0002000800010001"], 0x3c}}, 0x0) sendmsg$nl_route(r0, &(0x7f0000000240)={0x0, 0x0, &(0x7f0000000680)={&(0x7f0000000400)=@newlink={0x3c, 0x10, 0xffffff1f, 0x0, 0x7000000, {}, [@IFLA_LINKINFO={0x14, 0x12, 0x0, 0x1, @bridge={{0xb}, {0x4}}}, @IFLA_MASTER={0x8, 0xa, r3}]}, 0x3c}}, 0x0) 04:12:11 executing program 0: r0 = socket$inet6(0xa, 0x6, 0x0) bind$inet6(r0, &(0x7f0000000000)={0xa, 0x4e20}, 0x1c) listen(r0, 0x6) r1 = socket$inet6(0xa, 0x6, 0x0) connect$inet6(r1, &(0x7f0000000200)={0xa, 0x0, 0x0, @initdev={0xfe, 0x88, '\x00', 0x0, 0x0}, 0x3}, 0x1c) connect$inet6(r1, &(0x7f0000000040)={0xa, 0x4e20, 0x0, @dev={0xfe, 0x80, '\x00', 0xf}}, 0x1c) sendmmsg$inet6(r1, &(0x7f0000005a00)=[{{0x0, 0x0, 0x0}}, {{0x0, 0x0, 0x0, 0x0, &(0x7f0000005bc0)=ANY=[@ANYBLOB="5000000000000000290000003600000000060000000000000401000730000000000a0000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000020000000000000002900000037000000000000000000000004010000000000001400000000000000290000000000000160000000000000002900000037"], 0x100}}], 0x2, 0x0) shutdown(r0, 0x1) [ 2803.409230][ T2363] workqueue: Failed to create a rescuer kthread for wq "bond1282": -EINTR [ 2803.895136][ T2377] netlink: 'syz-executor.5': attribute type 1 has an invalid length. 04:12:11 executing program 0: r0 = socket$inet6(0xa, 0x6, 0x0) bind$inet6(r0, &(0x7f0000000000)={0xa, 0x4e20}, 0x1c) (async) listen(r0, 0x6) (async) r1 = socket$inet6(0xa, 0x6, 0x0) connect$inet6(r1, &(0x7f0000000200)={0xa, 0x0, 0x0, @initdev={0xfe, 0x88, '\x00', 0x0, 0x0}, 0x3}, 0x1c) (async) connect$inet6(r1, &(0x7f0000000040)={0xa, 0x4e20, 0x0, @dev={0xfe, 0x80, '\x00', 0xf}}, 0x1c) (async) sendmmsg$inet6(r1, &(0x7f0000005a00)=[{{0x0, 0x0, 0x0}}, {{0x0, 0x0, 0x0, 0x0, &(0x7f0000005bc0)=ANY=[@ANYBLOB="5000000000000000290000003600000000060000000000000401000730000000000a0000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000020000000000000002900000037000000000000000000000004010000000000001400000000000000290000000000000160000000000000002900000037"], 0x100}}], 0x2, 0x0) shutdown(r0, 0x1) 04:12:11 executing program 5: r0 = socket$netlink(0x10, 0x3, 0x0) r1 = socket$netlink(0x10, 0x3, 0x0) r2 = socket(0x10, 0x803, 0x0) sendmsg$NL80211_CMD_CRIT_PROTOCOL_START(r2, &(0x7f0000000580)={0x0, 0x0, &(0x7f0000000540)={0x0, 0x1c}}, 0x0) getsockname$packet(r2, &(0x7f0000000600)={0x11, 0x0, 0x0, 0x1, 0x0, 0x6, @broadcast}, &(0x7f0000000100)=0x14) sendmsg$nl_route(r1, &(0x7f0000000040)={0x0, 0x0, &(0x7f0000000000)={&(0x7f00000000c0)=ANY=[@ANYBLOB="3c00000010008506eb9afc4cd8d06e754a0081c5", @ANYRES32=r3, @ANYBLOB="2377f292252155b21c0012000c000100626f6e64000000000c0002000800010001"], 0x3c}}, 0x0) sendmsg$nl_route(r0, &(0x7f0000000240)={0x0, 0x0, &(0x7f0000000680)={&(0x7f0000000400)=@newlink={0x3c, 0x10, 0xffffff1f, 0x0, 0x2535f1, {}, [@IFLA_LINKINFO={0x14, 0x12, 0x0, 0x1, @bridge={{0xb}, {0x4}}}, @IFLA_MASTER={0x8, 0xa, r3}]}, 0x3c}}, 0x0) 04:12:11 executing program 0: r0 = socket$inet6(0xa, 0x6, 0x0) bind$inet6(r0, &(0x7f0000000000)={0xa, 0x4e20}, 0x1c) (async) listen(r0, 0x6) (async, rerun: 32) r1 = socket$inet6(0xa, 0x6, 0x0) (rerun: 32) connect$inet6(r1, &(0x7f0000000200)={0xa, 0x0, 0x0, @initdev={0xfe, 0x88, '\x00', 0x0, 0x0}, 0x3}, 0x1c) (async) connect$inet6(r1, &(0x7f0000000040)={0xa, 0x4e20, 0x0, @dev={0xfe, 0x80, '\x00', 0xf}}, 0x1c) (async) sendmmsg$inet6(r1, &(0x7f0000005a00)=[{{0x0, 0x0, 0x0}}, {{0x0, 0x0, 0x0, 0x0, &(0x7f0000005bc0)=ANY=[@ANYBLOB="5000000000000000290000003600000000060000000000000401000730000000000a0000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000020000000000000002900000037000000000000000000000004010000000000001400000000000000290000000000000160000000000000002900000037"], 0x100}}], 0x2, 0x0) (async) shutdown(r0, 0x1) [ 2803.974194][ T2377] workqueue: Failed to create a rescuer kthread for wq "bond1274": -EINTR [ 2804.423141][ T2404] netlink: 'syz-executor.2': attribute type 1 has an invalid length. 04:12:12 executing program 0: r0 = socket$inet6(0xa, 0x6, 0x0) bind$inet6(r0, &(0x7f0000000000)={0xa, 0x4e20}, 0x1c) listen(r0, 0x103) r1 = socket$inet6_dccp(0xa, 0x6, 0x0) sendmsg$inet6(r1, &(0x7f0000000180)={&(0x7f0000000080)={0xa, 0x4e23, 0x4, @mcast1, 0xffff}, 0x1c, &(0x7f0000000140)=[{&(0x7f00000000c0)="9803ab9f7604f8fd7fa2e35430d755418b241378692645c45f02f476dcc332ed91af9ffcefaff0ad220916dcbb732d125f211bf6fbb45754271dd46a6daaee0281dee502c82143e4f5e3e09fee05535ed4c5e77d", 0x54}, {&(0x7f0000000240)="058e31787f7f111a35900199e61c2b8e1e3ca7edac433e5d55fc9548bed10bcecd429688577066787efc8fbb546a73629c3e6a16064ee00660face27234f5b562012c234596337b59a2adff2d8b467ed3365cb58a1596e848623af9e0727492ec85786dacb13a153fe0386d466a101e2ce4ebbb46770d9cccd9ef9eae6ba3b2256678b85e4c5c3f15733839664edfd8f61b066bb65b8d93d79e28af9d42166b19942d29980536cf8d3dc4cf2117797f44f9ac659793145ec42d77a23197f4d926fbd595fdd959ef7b613545171cdc46dd0557921501f8006575e81857f3632072952f22aee5641bfd50b67de1d2d3fc902aa242cc67eb8c7d636839847ec86704b108e3d25e7207738afd113744c3fd20da190a37963da491727683959647137491a394b0000da751943f3c497322b770a58ad178ce72e45b80762d4075dfae126d2486ca5c88c23cd6087f84946008768f79d2baf951f7ae9676159cf7b581e7336fb75fc4a4d1163f40777de70c6ab2656ccb497721c9423596197d07e6be7d2b55486a0f6fe39b7264d2f1c287a0a615edab3f5c35b5c9b1fa7251e1bf8247894833600ebd8048d2829672f1c1de71da572509f662010f4ec2de33ccfd4b5039a0b58c19d52f80283d4699bea47f3cde2ef837ee4f8668b34fb8678381273098d7e99f903cb72bf973e66fbae8f1606ee8a82f8ddd75b0a39c00ff063c8588c272532e30f3a0c60dae1230f4ce584efeef4e26efc3c62025cc2497d9a9104320f9507803c95c7dbc7c6642909541e5a53311507732eedd43b5ee9bf4b897f9f272d178437d530c696ed659d29a0ded70ab3e75aaa1bebf8924804cb746e05159a701daac1af550aeb004875fb09775a1a24a7b83ba7a30ef83dc549e93e93edbcde2425fdeb9b91b215b60143e007253ccb5bd481fd5f33864e707f787f9b1d249333172d79f5877a1280d986092b8944ec24906f705f273cc6308e794be53ea47f3d6341ba1f137bc093843eb2a304bf4e73c99a6a1f96fb6dbb24d60c9e5a5968d44842d54f9a0fcb607375c873246f48e451d24145554991128862602958d9880bd565d5ab38562bd287875817f253809f6328fc7ce3dbfdfba35cde5441484ed1780cf4f37557ae33174108625aab31e6260043ed265d4c469a5b107e704c7e1f6c493304f93c197c16ff158c2ac354f3953dffac674a2f35512399ecee6648d9cf7b22549ac9a5bd500a467ca7a53eea80840ac77c73245aade6325292936624a793b1c1e601bdd4f8756149ec9ad36b0b5342a3a2f4871d14fb2507d96ea9702461a5cac0ba7f98b67fffe793ba5043c836a8625ce9d3e2fb32a47c6184b9e86edd81d3542eb4a5a15524713c0a5917b080d6dd7975129011f02584585845ef69c7d42146e32b8ae39ee16efd340c5e98a39f706f4234e97b6b994007a7a649d029d7cf6d7eb1c12fa6497ececb292772f675ca98653a68a25c3d5cb1fdac423cb521ffffb84599e5eb4fac3731ccccb57073802871dd276c29b511efd8442beaf173028564c448f52a71176dfbcf76fd44a43d3db94b09e0df58a2b350f53703cf495f86496e3e8136a888dae249b6d50cde25b05c132258e20f788c3cacbe26164ce68e43d74e6302048da6c3a96d98d68086fc45eb7d2082f4007019b158a2e084093a90592446b4dafc128d55041b3f4c9e5c08bf4a0c2585148103d8fb04788bb34d5d3148dd04a9fe96862b712416818fcc92f4fe8ee0e9112e1a2e7e149b1b67908e2b14d4f2f74ac124b41a71662708a89554aa026239fe939168248b119926eb81c20d6f85962384f4cc37b84089fe0c33ae403a7a8c88224e098d682c15894559f8bfa655ff411b007075dca6d5b7fb6c73696ff28aaa059f06150e8ac00d90f3d1519c359dda8038bd96fe2b37c94c16774cab5653caf0f506a4dbd3f0dc94688cd29138f74af534ef5960004c2d3b8444119b5f8bead128b4c1c3ca00537c5f902b6460456924be23d3be69037243db8972a9748075da24e216b7ccda79c215c2027c350e1fc3c50d4795a18cec5469c5ab48dfa1c9635a6625803e94e66802c50f5552a12cbabfee587f0e3f082672b7de2f10573b91fcabd304d5d768918967ed73ef29d6ac028639d2573ebe5c17ab3f10f9fad1b86b95d87b1731e7c3272d0b2a014accb9e3d4305a3473f60fbf63bd401301f8a73edc3b143914dd2dc0d850955eecc8927b65b90e3e098efe58ea0aafcbd9589a61f0da3db1c81b9b26f65708516187df12fd9eb9aefb2f4fc4dc4cbc3675a58127828dc0d5fba447fdd7a7f2f5d422951a1d789b519d8842ced12e2ece94a948a1762c0e81ff4be5811d2afdea0bc21b9ef56932364a28c45a86fc551971354269e90dc5326c88d708247f9c1b8bed70a99be05a5b0917330e1fa2e1d4d7c26fb9fc1117de011335c2a846e1f0687cabac113c315e6344437b251ef264ca2d69677d146eccf5ce4bfdc01ae1339dfdff3b6a43aa855cd3ad7c06876f68f09aef5d5a281a5661794fb63cf2f4ebf60d5e0e70b118c7b58df082713930af16cbd091ded8a9e6e7c153410841c1e911987ed3af015196c9e06f21be67307f701be3c1fa7a542aa69a1d2a6dbefda6159b7c4e7b624164b971b1f0d1f9ba4250f418473689af52d4e7b111d24cd15359fa994e5e3bcd49d307d8f63d11227e7df6719ea886a4b1dae2e23a8d77832e10b0aab91653d3e339fab859eb8d1c8795116121b304f4f243db1b95ab2332f2f4f26effffe739fcfbb89f4f54cfd5efe67be2774e76b34b389e80c0e1b0dad91fcdf75e125c11630c3d977ad443a66cda3aa55f6dd40732b2dfa4d3a7897e38ceefeb8c3a3c7e723e39aad56f698e61a87721e251210f84ff1fe9707e83bd7e87b35ffda5f0caa742b5e3146184d296b45968c534f1efacae1d90b692c78bd1dbd08bbeea5ee7e9ee0d330a1f37511825a39529a6b4d729fe5b6ad60f325759551c1d69fbccd292e803b1c9697eb7f877ce071127f679724c7c98defbfbfb040f09d8ec7891bd2348e4a715ce125fd34b013876897375e22600c8b146d3700e471072d757b6c5ae40d93bcc453c425d86ab119a9354049cffd521afb1f3de7e0e74245fc06cb2f022939ff8e2685aa7987519f08dbfbc34a043b576f647c5a9117b9600f0e9c850d47a6c51ee0f7f6f83676e6e8b21de10aff1beb5a5d12eb22ff6d89bc9a00b525a77392bd0fca3f06d1058d499ff6feafb1541ce540b851a4f1920bdc2ac7d199c332d081260579541f66d5b0ebccfcbba78788b5b43743b1260057902655052f1f299b5d4852c56fe443709d3592addb73a0bf325016bee1624deff6d503a33fdacd7705e7f9a47f3457a78911522340c38b9edbb5e69ae05d6dc78c33fbe2f003cfe0a182bab11d0df9bb7c52178a4e49c25af5690fc716bcad4e877db093ec3962287187752e02a2ad16c51ace16b485e03a88a43e258f902fa8cfdd5b8be0db1c5f2a86b352e7227bb5be86fbe87d7d4a47245b4bbb4a231c4782c5b088f3b7839f38818c642a0831a3e86959ba233b6b5fac8723e901f2d2185c15e05170b674ba942c258c3c490178b9fbd924f2d760de44e02ef97e2dc1057bbd2fb6aa39a64969e2afb6b9632916f3475f7b6b71df30eb9ed3a7762385580b81f20db0951f34cce02b04f76b8bf734e0997033195c6403e53bdfa45e9d1724552cb6916620f7b5db362146638072f217cd53952e61c56f41bfd21fb524220685cbc65eaef4c1b728ef41d076d9b1069b2ec6c4d90f291d1a4fc87e25c9e8615ac0a67b217a645e7a7a8211400f37a9d4f0408992d6c31b970d0a7fd4893cc39243eca6ffa276d26423b5a26001421837747053aa58fc0eea37044384b1433eca593076b7fe97384a58707b5dba84eeba13d2f90032a80c6f8270d53135a72f54b43dc7b6e82fb360bcfc43b30c8b4d9c6fa05ec2c17fd9d20406637e8f0d7f852adb5e2cc6ff04da9aa870dfccd9e1677069cad59105bde64d30de72c328158fcc5cb31246ee21e385cbddfd1a6e0a19630cfcdc707981dfdb4df454e538913595bf27a6bc1963de2432b0b9dbe6a71b43dc67f58316487e2cdb12d86f5c9db9f96726c80ee2fb74cb03e1868ae50dd382777dd627aa85ee6c9077d0de79d76b904dd5fe645c616fdfbfacc1a6572df4089c2a1bd0acb465396bf46f30060cbd78201b4719ecbb053b7c4ec78c025e25b584b0228e8e3cb90f040300378d76141a6474139b0a8eb99b54124d402d955644a78a3902e40038116f3d189c41f29f4e000c5b28a42f718d951372480ac1ed6b78c3a7f170d3e1963c8c635bdc58a928407b6b11ebf69d24fb28ef343d847b7d98589863fb4a4085d4db9d274f9f91f074214e74c9873a4a77490026bb816f49ff981e71e37545cc87cea0056683fa7ae3b371971495a75c079705cdaf247a6a88868885010992c7926071bf5793d6d1fe061e4efcf8e824e63e3bef5293ed2564e8f608ea71cf1c313776ee96670fd1430c9d1af49aa0e3540bf3d8135a60ebc367325278419d3f840e49d8cabc4aba9a024b139687fa867fcf74d2322dda111d5e7d0c892e6ae74b2162f9829819fc1433a4e445a71c0aef8aecb14e2894daef3d833ca814ce721979ec35a18865fcbfc8f4ad516e8aca22bf3c504616e2dfc792e32e443b3c8c0d699cbf1b9d05da0268761ec3d5079748ac9db57403cd197bfad46710d63d91788fea227307edbbe440473f9c4592c10122961205f9a364e97453e6812919ab1fa88ea7460035ad01b6704e25543242004acb3b5302e39ac9a674e160e6a9adcf8616034d345f45c8a55d7789db944b48723d4b90cf14409cd39713fcd336290359015404e25049e8772fe42ade078afa2e41d4f8bfda0cde1f8d069cc465fee7a8f12c7958f020137a234bc64665dfe03dea207fd60027228f85515c04a4d5a9583ec16fec173137e55d6ff7fde97ed8d1bc57f7562a9169fa5b5f7a2413bc72b0695698b9c8e2a0dda1dfae07f47e2edb2033639580f216dba9a151b6a7e6be91ef426d039f12fb1ebccfd95966a8cad4e5af1d41529a8a24f7045b7ce41baa48e42b2b882e319d9f85dee5da7191b3c5c007b7c94c85362d533b4fd45701fe19ced8149547a37144ad0815f2b6bece91f08969236bbe70b9a7f12206fa7ede73a6a783fc3b340634c850e7fe1b94d6aeca5397c8a1e7ae511948c1da616926e5997cdb32bf9861db5860ab271dad72037fbce5643c4fd206e90359609a9f49aa33b868e32abdc04fb6a19c34206d85ed769050e3b378a69bff2445691e084051fd41c456a80e3d30ce725a44399b13e54791555d117c1a3dee47e1c71a1ab64fb042bf07276dadf7054fea1b1777c866987cb109a59c59e3301a113b37888c78a3eaffbe502af461ef30395c4d17b6236411603724ac71782b35ee4430b1da3dce7847273d66ce2f193cbb7787c3a6870a3243c0530d6f20cc3df3a45c8b1b9ffb61802dc60c5e3adaa4a9ca9e79b13a7a04ff07751f465462f133428eb4186b510c267f0d7e16f6d86bc629ae331b6c391abb5115b72eda9be937159f9424b0d13d5b75749637f590b58e549b629bbcf79f768ea7a91cc5c1d989c4d23d55f0bac5bd2b114df652451e679ba2aada14d586bb4c7df4543521178835e1fff3abf0c3d4a0397c2e041388ffe9a4568a7dad9e00e89744d1c454cb55906d3ce415494c83b5a6f9f34a120925e95204490814b", 0x1000}], 0x2}, 0x40001) r2 = socket$inet6(0xa, 0x6, 0x0) connect$inet6(r2, &(0x7f0000000200)={0xa, 0x0, 0x0, @initdev={0xfe, 0x88, '\x00', 0x0, 0x0}, 0x3}, 0x1c) connect$inet6(r2, &(0x7f0000000040)={0xa, 0x4e20, 0x0, @dev={0xfe, 0x80, '\x00', 0xf}}, 0x1c) sendmmsg$inet6(r2, &(0x7f0000005a00)=[{{0x0, 0x0, 0x0}}, {{0x0, 0x0, 0x0, 0x0, &(0x7f0000005bc0)=ANY=[@ANYBLOB="5000000000000000290000003600000000060000000000000401000730000000000a0000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000020000000000000002900000037000000000000000000000004010000000000001400000000000000290000000000000160000000000000002900000037"], 0x100}}], 0x2, 0x0) [ 2804.524360][ T2404] bond1262: entered promiscuous mode [ 2804.530175][ T2404] 8021q: adding VLAN 0 to HW filter on device bond1262 [ 2804.638482][ T2407] bond1262: (slave bridge1183): making interface the new active one [ 2804.649213][ T2407] bridge1183: entered promiscuous mode 04:12:12 executing program 0: r0 = socket$inet6(0xa, 0x6, 0x0) bind$inet6(r0, &(0x7f0000000000)={0xa, 0x4e20}, 0x1c) (async) listen(r0, 0x103) r1 = socket$inet6_dccp(0xa, 0x6, 0x0) sendmsg$inet6(r1, &(0x7f0000000180)={&(0x7f0000000080)={0xa, 0x4e23, 0x4, @mcast1, 0xffff}, 0x1c, &(0x7f0000000140)=[{&(0x7f00000000c0)="9803ab9f7604f8fd7fa2e35430d755418b241378692645c45f02f476dcc332ed91af9ffcefaff0ad220916dcbb732d125f211bf6fbb45754271dd46a6daaee0281dee502c82143e4f5e3e09fee05535ed4c5e77d", 0x54}, {&(0x7f0000000240)="058e31787f7f111a35900199e61c2b8e1e3ca7edac433e5d55fc9548bed10bcecd429688577066787efc8fbb546a73629c3e6a16064ee00660face27234f5b562012c234596337b59a2adff2d8b467ed3365cb58a1596e848623af9e0727492ec85786dacb13a153fe0386d466a101e2ce4ebbb46770d9cccd9ef9eae6ba3b2256678b85e4c5c3f15733839664edfd8f61b066bb65b8d93d79e28af9d42166b19942d29980536cf8d3dc4cf2117797f44f9ac659793145ec42d77a23197f4d926fbd595fdd959ef7b613545171cdc46dd0557921501f8006575e81857f3632072952f22aee5641bfd50b67de1d2d3fc902aa242cc67eb8c7d636839847ec86704b108e3d25e7207738afd113744c3fd20da190a37963da491727683959647137491a394b0000da751943f3c497322b770a58ad178ce72e45b80762d4075dfae126d2486ca5c88c23cd6087f84946008768f79d2baf951f7ae9676159cf7b581e7336fb75fc4a4d1163f40777de70c6ab2656ccb497721c9423596197d07e6be7d2b55486a0f6fe39b7264d2f1c287a0a615edab3f5c35b5c9b1fa7251e1bf8247894833600ebd8048d2829672f1c1de71da572509f662010f4ec2de33ccfd4b5039a0b58c19d52f80283d4699bea47f3cde2ef837ee4f8668b34fb8678381273098d7e99f903cb72bf973e66fbae8f1606ee8a82f8ddd75b0a39c00ff063c8588c272532e30f3a0c60dae1230f4ce584efeef4e26efc3c62025cc2497d9a9104320f9507803c95c7dbc7c6642909541e5a53311507732eedd43b5ee9bf4b897f9f272d178437d530c696ed659d29a0ded70ab3e75aaa1bebf8924804cb746e05159a701daac1af550aeb004875fb09775a1a24a7b83ba7a30ef83dc549e93e93edbcde2425fdeb9b91b215b60143e007253ccb5bd481fd5f33864e707f787f9b1d249333172d79f5877a1280d986092b8944ec24906f705f273cc6308e794be53ea47f3d6341ba1f137bc093843eb2a304bf4e73c99a6a1f96fb6dbb24d60c9e5a5968d44842d54f9a0fcb607375c873246f48e451d24145554991128862602958d9880bd565d5ab38562bd287875817f253809f6328fc7ce3dbfdfba35cde5441484ed1780cf4f37557ae33174108625aab31e6260043ed265d4c469a5b107e704c7e1f6c493304f93c197c16ff158c2ac354f3953dffac674a2f35512399ecee6648d9cf7b22549ac9a5bd500a467ca7a53eea80840ac77c73245aade6325292936624a793b1c1e601bdd4f8756149ec9ad36b0b5342a3a2f4871d14fb2507d96ea9702461a5cac0ba7f98b67fffe793ba5043c836a8625ce9d3e2fb32a47c6184b9e86edd81d3542eb4a5a15524713c0a5917b080d6dd7975129011f02584585845ef69c7d42146e32b8ae39ee16efd340c5e98a39f706f4234e97b6b994007a7a649d029d7cf6d7eb1c12fa6497ececb292772f675ca98653a68a25c3d5cb1fdac423cb521ffffb84599e5eb4fac3731ccccb57073802871dd276c29b511efd8442beaf173028564c448f52a71176dfbcf76fd44a43d3db94b09e0df58a2b350f53703cf495f86496e3e8136a888dae249b6d50cde25b05c132258e20f788c3cacbe26164ce68e43d74e6302048da6c3a96d98d68086fc45eb7d2082f4007019b158a2e084093a90592446b4dafc128d55041b3f4c9e5c08bf4a0c2585148103d8fb04788bb34d5d3148dd04a9fe96862b712416818fcc92f4fe8ee0e9112e1a2e7e149b1b67908e2b14d4f2f74ac124b41a71662708a89554aa026239fe939168248b119926eb81c20d6f85962384f4cc37b84089fe0c33ae403a7a8c88224e098d682c15894559f8bfa655ff411b007075dca6d5b7fb6c73696ff28aaa059f06150e8ac00d90f3d1519c359dda8038bd96fe2b37c94c16774cab5653caf0f506a4dbd3f0dc94688cd29138f74af534ef5960004c2d3b8444119b5f8bead128b4c1c3ca00537c5f902b6460456924be23d3be69037243db8972a9748075da24e216b7ccda79c215c2027c350e1fc3c50d4795a18cec5469c5ab48dfa1c9635a6625803e94e66802c50f5552a12cbabfee587f0e3f082672b7de2f10573b91fcabd304d5d768918967ed73ef29d6ac028639d2573ebe5c17ab3f10f9fad1b86b95d87b1731e7c3272d0b2a014accb9e3d4305a3473f60fbf63bd401301f8a73edc3b143914dd2dc0d850955eecc8927b65b90e3e098efe58ea0aafcbd9589a61f0da3db1c81b9b26f65708516187df12fd9eb9aefb2f4fc4dc4cbc3675a58127828dc0d5fba447fdd7a7f2f5d422951a1d789b519d8842ced12e2ece94a948a1762c0e81ff4be5811d2afdea0bc21b9ef56932364a28c45a86fc551971354269e90dc5326c88d708247f9c1b8bed70a99be05a5b0917330e1fa2e1d4d7c26fb9fc1117de011335c2a846e1f0687cabac113c315e6344437b251ef264ca2d69677d146eccf5ce4bfdc01ae1339dfdff3b6a43aa855cd3ad7c06876f68f09aef5d5a281a5661794fb63cf2f4ebf60d5e0e70b118c7b58df082713930af16cbd091ded8a9e6e7c153410841c1e911987ed3af015196c9e06f21be67307f701be3c1fa7a542aa69a1d2a6dbefda6159b7c4e7b624164b971b1f0d1f9ba4250f418473689af52d4e7b111d24cd15359fa994e5e3bcd49d307d8f63d11227e7df6719ea886a4b1dae2e23a8d77832e10b0aab91653d3e339fab859eb8d1c8795116121b304f4f243db1b95ab2332f2f4f26effffe739fcfbb89f4f54cfd5efe67be2774e76b34b389e80c0e1b0dad91fcdf75e125c11630c3d977ad443a66cda3aa55f6dd40732b2dfa4d3a7897e38ceefeb8c3a3c7e723e39aad56f698e61a87721e251210f84ff1fe9707e83bd7e87b35ffda5f0caa742b5e3146184d296b45968c534f1efacae1d90b692c78bd1dbd08bbeea5ee7e9ee0d330a1f37511825a39529a6b4d729fe5b6ad60f325759551c1d69fbccd292e803b1c9697eb7f877ce071127f679724c7c98defbfbfb040f09d8ec7891bd2348e4a715ce125fd34b013876897375e22600c8b146d3700e471072d757b6c5ae40d93bcc453c425d86ab119a9354049cffd521afb1f3de7e0e74245fc06cb2f022939ff8e2685aa7987519f08dbfbc34a043b576f647c5a9117b9600f0e9c850d47a6c51ee0f7f6f83676e6e8b21de10aff1beb5a5d12eb22ff6d89bc9a00b525a77392bd0fca3f06d1058d499ff6feafb1541ce540b851a4f1920bdc2ac7d199c332d081260579541f66d5b0ebccfcbba78788b5b43743b1260057902655052f1f299b5d4852c56fe443709d3592addb73a0bf325016bee1624deff6d503a33fdacd7705e7f9a47f3457a78911522340c38b9edbb5e69ae05d6dc78c33fbe2f003cfe0a182bab11d0df9bb7c52178a4e49c25af5690fc716bcad4e877db093ec3962287187752e02a2ad16c51ace16b485e03a88a43e258f902fa8cfdd5b8be0db1c5f2a86b352e7227bb5be86fbe87d7d4a47245b4bbb4a231c4782c5b088f3b7839f38818c642a0831a3e86959ba233b6b5fac8723e901f2d2185c15e05170b674ba942c258c3c490178b9fbd924f2d760de44e02ef97e2dc1057bbd2fb6aa39a64969e2afb6b9632916f3475f7b6b71df30eb9ed3a7762385580b81f20db0951f34cce02b04f76b8bf734e0997033195c6403e53bdfa45e9d1724552cb6916620f7b5db362146638072f217cd53952e61c56f41bfd21fb524220685cbc65eaef4c1b728ef41d076d9b1069b2ec6c4d90f291d1a4fc87e25c9e8615ac0a67b217a645e7a7a8211400f37a9d4f0408992d6c31b970d0a7fd4893cc39243eca6ffa276d26423b5a26001421837747053aa58fc0eea37044384b1433eca593076b7fe97384a58707b5dba84eeba13d2f90032a80c6f8270d53135a72f54b43dc7b6e82fb360bcfc43b30c8b4d9c6fa05ec2c17fd9d20406637e8f0d7f852adb5e2cc6ff04da9aa870dfccd9e1677069cad59105bde64d30de72c328158fcc5cb31246ee21e385cbddfd1a6e0a19630cfcdc707981dfdb4df454e538913595bf27a6bc1963de2432b0b9dbe6a71b43dc67f58316487e2cdb12d86f5c9db9f96726c80ee2fb74cb03e1868ae50dd382777dd627aa85ee6c9077d0de79d76b904dd5fe645c616fdfbfacc1a6572df4089c2a1bd0acb465396bf46f30060cbd78201b4719ecbb053b7c4ec78c025e25b584b0228e8e3cb90f040300378d76141a6474139b0a8eb99b54124d402d955644a78a3902e40038116f3d189c41f29f4e000c5b28a42f718d951372480ac1ed6b78c3a7f170d3e1963c8c635bdc58a928407b6b11ebf69d24fb28ef343d847b7d98589863fb4a4085d4db9d274f9f91f074214e74c9873a4a77490026bb816f49ff981e71e37545cc87cea0056683fa7ae3b371971495a75c079705cdaf247a6a88868885010992c7926071bf5793d6d1fe061e4efcf8e824e63e3bef5293ed2564e8f608ea71cf1c313776ee96670fd1430c9d1af49aa0e3540bf3d8135a60ebc367325278419d3f840e49d8cabc4aba9a024b139687fa867fcf74d2322dda111d5e7d0c892e6ae74b2162f9829819fc1433a4e445a71c0aef8aecb14e2894daef3d833ca814ce721979ec35a18865fcbfc8f4ad516e8aca22bf3c504616e2dfc792e32e443b3c8c0d699cbf1b9d05da0268761ec3d5079748ac9db57403cd197bfad46710d63d91788fea227307edbbe440473f9c4592c10122961205f9a364e97453e6812919ab1fa88ea7460035ad01b6704e25543242004acb3b5302e39ac9a674e160e6a9adcf8616034d345f45c8a55d7789db944b48723d4b90cf14409cd39713fcd336290359015404e25049e8772fe42ade078afa2e41d4f8bfda0cde1f8d069cc465fee7a8f12c7958f020137a234bc64665dfe03dea207fd60027228f85515c04a4d5a9583ec16fec173137e55d6ff7fde97ed8d1bc57f7562a9169fa5b5f7a2413bc72b0695698b9c8e2a0dda1dfae07f47e2edb2033639580f216dba9a151b6a7e6be91ef426d039f12fb1ebccfd95966a8cad4e5af1d41529a8a24f7045b7ce41baa48e42b2b882e319d9f85dee5da7191b3c5c007b7c94c85362d533b4fd45701fe19ced8149547a37144ad0815f2b6bece91f08969236bbe70b9a7f12206fa7ede73a6a783fc3b340634c850e7fe1b94d6aeca5397c8a1e7ae511948c1da616926e5997cdb32bf9861db5860ab271dad72037fbce5643c4fd206e90359609a9f49aa33b868e32abdc04fb6a19c34206d85ed769050e3b378a69bff2445691e084051fd41c456a80e3d30ce725a44399b13e54791555d117c1a3dee47e1c71a1ab64fb042bf07276dadf7054fea1b1777c866987cb109a59c59e3301a113b37888c78a3eaffbe502af461ef30395c4d17b6236411603724ac71782b35ee4430b1da3dce7847273d66ce2f193cbb7787c3a6870a3243c0530d6f20cc3df3a45c8b1b9ffb61802dc60c5e3adaa4a9ca9e79b13a7a04ff07751f465462f133428eb4186b510c267f0d7e16f6d86bc629ae331b6c391abb5115b72eda9be937159f9424b0d13d5b75749637f590b58e549b629bbcf79f768ea7a91cc5c1d989c4d23d55f0bac5bd2b114df652451e679ba2aada14d586bb4c7df4543521178835e1fff3abf0c3d4a0397c2e041388ffe9a4568a7dad9e00e89744d1c454cb55906d3ce415494c83b5a6f9f34a120925e95204490814b", 0x1000}], 0x2}, 0x40001) r2 = socket$inet6(0xa, 0x6, 0x0) connect$inet6(r2, &(0x7f0000000200)={0xa, 0x0, 0x0, @initdev={0xfe, 0x88, '\x00', 0x0, 0x0}, 0x3}, 0x1c) (async) connect$inet6(r2, &(0x7f0000000040)={0xa, 0x4e20, 0x0, @dev={0xfe, 0x80, '\x00', 0xf}}, 0x1c) (async) sendmmsg$inet6(r2, &(0x7f0000005a00)=[{{0x0, 0x0, 0x0}}, {{0x0, 0x0, 0x0, 0x0, &(0x7f0000005bc0)=ANY=[@ANYBLOB="5000000000000000290000003600000000060000000000000401000730000000000a0000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000020000000000000002900000037000000000000000000000004010000000000001400000000000000290000000000000160000000000000002900000037"], 0x100}}], 0x2, 0x0) [ 2804.679044][ T2407] bond1262: (slave bridge1183): Enslaving as an active interface with an up link 04:12:12 executing program 2: r0 = socket$netlink(0x10, 0x3, 0x0) r1 = socket$netlink(0x10, 0x3, 0x0) r2 = socket(0x10, 0x803, 0x0) sendmsg$NL80211_CMD_CRIT_PROTOCOL_START(r2, &(0x7f0000000580)={0x0, 0x0, &(0x7f0000000540)={0x0, 0x1c}}, 0x0) getsockname$packet(r2, &(0x7f0000000600)={0x11, 0x0, 0x0, 0x1, 0x0, 0x6, @broadcast}, &(0x7f0000000100)=0x14) sendmsg$nl_route(r1, &(0x7f0000000040)={0x0, 0x0, &(0x7f0000000000)={&(0x7f00000000c0)=ANY=[@ANYBLOB="3c00000010008506eb9afc4cd8d06e754a0081c5", @ANYRES32=r3, @ANYBLOB="2377f292252155b21c0012000c000100626f6e64000000000c0002000800010001"], 0x3c}}, 0x0) sendmsg$nl_route(r0, &(0x7f0000000240)={0x0, 0x0, &(0x7f0000000680)={&(0x7f0000000400)=@newlink={0x3c, 0x10, 0xffffff1f, 0x0, 0x253c20, {}, [@IFLA_LINKINFO={0x14, 0x12, 0x0, 0x1, @bridge={{0xb}, {0x4}}}, @IFLA_MASTER={0x8, 0xa, r3}]}, 0x3c}}, 0x0) [ 2804.893801][ T2415] netlink: 'syz-executor.1': attribute type 1 has an invalid length. 04:12:12 executing program 3: r0 = socket$netlink(0x10, 0x3, 0x0) sendmsg$netlink(r0, &(0x7f00000047c0)={0x0, 0x0, &(0x7f0000000000)=[{&(0x7f0000000100)=ANY=[@ANYBLOB="18000000560001060000006500db730007"], 0x18}], 0x1}, 0x0) 04:12:12 executing program 0: r0 = socket$inet6(0xa, 0x6, 0x0) bind$inet6(r0, &(0x7f0000000000)={0xa, 0x4e20}, 0x1c) listen(r0, 0x103) (async) listen(r0, 0x103) socket$inet6_dccp(0xa, 0x6, 0x0) (async) r1 = socket$inet6_dccp(0xa, 0x6, 0x0) sendmsg$inet6(r1, &(0x7f0000000180)={&(0x7f0000000080)={0xa, 0x4e23, 0x4, @mcast1, 0xffff}, 0x1c, &(0x7f0000000140)=[{&(0x7f00000000c0)="9803ab9f7604f8fd7fa2e35430d755418b241378692645c45f02f476dcc332ed91af9ffcefaff0ad220916dcbb732d125f211bf6fbb45754271dd46a6daaee0281dee502c82143e4f5e3e09fee05535ed4c5e77d", 0x54}, {&(0x7f0000000240)="058e31787f7f111a35900199e61c2b8e1e3ca7edac433e5d55fc9548bed10bcecd429688577066787efc8fbb546a73629c3e6a16064ee00660face27234f5b562012c234596337b59a2adff2d8b467ed3365cb58a1596e848623af9e0727492ec85786dacb13a153fe0386d466a101e2ce4ebbb46770d9cccd9ef9eae6ba3b2256678b85e4c5c3f15733839664edfd8f61b066bb65b8d93d79e28af9d42166b19942d29980536cf8d3dc4cf2117797f44f9ac659793145ec42d77a23197f4d926fbd595fdd959ef7b613545171cdc46dd0557921501f8006575e81857f3632072952f22aee5641bfd50b67de1d2d3fc902aa242cc67eb8c7d636839847ec86704b108e3d25e7207738afd113744c3fd20da190a37963da491727683959647137491a394b0000da751943f3c497322b770a58ad178ce72e45b80762d4075dfae126d2486ca5c88c23cd6087f84946008768f79d2baf951f7ae9676159cf7b581e7336fb75fc4a4d1163f40777de70c6ab2656ccb497721c9423596197d07e6be7d2b55486a0f6fe39b7264d2f1c287a0a615edab3f5c35b5c9b1fa7251e1bf8247894833600ebd8048d2829672f1c1de71da572509f662010f4ec2de33ccfd4b5039a0b58c19d52f80283d4699bea47f3cde2ef837ee4f8668b34fb8678381273098d7e99f903cb72bf973e66fbae8f1606ee8a82f8ddd75b0a39c00ff063c8588c272532e30f3a0c60dae1230f4ce584efeef4e26efc3c62025cc2497d9a9104320f9507803c95c7dbc7c6642909541e5a53311507732eedd43b5ee9bf4b897f9f272d178437d530c696ed659d29a0ded70ab3e75aaa1bebf8924804cb746e05159a701daac1af550aeb004875fb09775a1a24a7b83ba7a30ef83dc549e93e93edbcde2425fdeb9b91b215b60143e007253ccb5bd481fd5f33864e707f787f9b1d249333172d79f5877a1280d986092b8944ec24906f705f273cc6308e794be53ea47f3d6341ba1f137bc093843eb2a304bf4e73c99a6a1f96fb6dbb24d60c9e5a5968d44842d54f9a0fcb607375c873246f48e451d24145554991128862602958d9880bd565d5ab38562bd287875817f253809f6328fc7ce3dbfdfba35cde5441484ed1780cf4f37557ae33174108625aab31e6260043ed265d4c469a5b107e704c7e1f6c493304f93c197c16ff158c2ac354f3953dffac674a2f35512399ecee6648d9cf7b22549ac9a5bd500a467ca7a53eea80840ac77c73245aade6325292936624a793b1c1e601bdd4f8756149ec9ad36b0b5342a3a2f4871d14fb2507d96ea9702461a5cac0ba7f98b67fffe793ba5043c836a8625ce9d3e2fb32a47c6184b9e86edd81d3542eb4a5a15524713c0a5917b080d6dd7975129011f02584585845ef69c7d42146e32b8ae39ee16efd340c5e98a39f706f4234e97b6b994007a7a649d029d7cf6d7eb1c12fa6497ececb292772f675ca98653a68a25c3d5cb1fdac423cb521ffffb84599e5eb4fac3731ccccb57073802871dd276c29b511efd8442beaf173028564c448f52a71176dfbcf76fd44a43d3db94b09e0df58a2b350f53703cf495f86496e3e8136a888dae249b6d50cde25b05c132258e20f788c3cacbe26164ce68e43d74e6302048da6c3a96d98d68086fc45eb7d2082f4007019b158a2e084093a90592446b4dafc128d55041b3f4c9e5c08bf4a0c2585148103d8fb04788bb34d5d3148dd04a9fe96862b712416818fcc92f4fe8ee0e9112e1a2e7e149b1b67908e2b14d4f2f74ac124b41a71662708a89554aa026239fe939168248b119926eb81c20d6f85962384f4cc37b84089fe0c33ae403a7a8c88224e098d682c15894559f8bfa655ff411b007075dca6d5b7fb6c73696ff28aaa059f06150e8ac00d90f3d1519c359dda8038bd96fe2b37c94c16774cab5653caf0f506a4dbd3f0dc94688cd29138f74af534ef5960004c2d3b8444119b5f8bead128b4c1c3ca00537c5f902b6460456924be23d3be69037243db8972a9748075da24e216b7ccda79c215c2027c350e1fc3c50d4795a18cec5469c5ab48dfa1c9635a6625803e94e66802c50f5552a12cbabfee587f0e3f082672b7de2f10573b91fcabd304d5d768918967ed73ef29d6ac028639d2573ebe5c17ab3f10f9fad1b86b95d87b1731e7c3272d0b2a014accb9e3d4305a3473f60fbf63bd401301f8a73edc3b143914dd2dc0d850955eecc8927b65b90e3e098efe58ea0aafcbd9589a61f0da3db1c81b9b26f65708516187df12fd9eb9aefb2f4fc4dc4cbc3675a58127828dc0d5fba447fdd7a7f2f5d422951a1d789b519d8842ced12e2ece94a948a1762c0e81ff4be5811d2afdea0bc21b9ef56932364a28c45a86fc551971354269e90dc5326c88d708247f9c1b8bed70a99be05a5b0917330e1fa2e1d4d7c26fb9fc1117de011335c2a846e1f0687cabac113c315e6344437b251ef264ca2d69677d146eccf5ce4bfdc01ae1339dfdff3b6a43aa855cd3ad7c06876f68f09aef5d5a281a5661794fb63cf2f4ebf60d5e0e70b118c7b58df082713930af16cbd091ded8a9e6e7c153410841c1e911987ed3af015196c9e06f21be67307f701be3c1fa7a542aa69a1d2a6dbefda6159b7c4e7b624164b971b1f0d1f9ba4250f418473689af52d4e7b111d24cd15359fa994e5e3bcd49d307d8f63d11227e7df6719ea886a4b1dae2e23a8d77832e10b0aab91653d3e339fab859eb8d1c8795116121b304f4f243db1b95ab2332f2f4f26effffe739fcfbb89f4f54cfd5efe67be2774e76b34b389e80c0e1b0dad91fcdf75e125c11630c3d977ad443a66cda3aa55f6dd40732b2dfa4d3a7897e38ceefeb8c3a3c7e723e39aad56f698e61a87721e251210f84ff1fe9707e83bd7e87b35ffda5f0caa742b5e3146184d296b45968c534f1efacae1d90b692c78bd1dbd08bbeea5ee7e9ee0d330a1f37511825a39529a6b4d729fe5b6ad60f325759551c1d69fbccd292e803b1c9697eb7f877ce071127f679724c7c98defbfbfb040f09d8ec7891bd2348e4a715ce125fd34b013876897375e22600c8b146d3700e471072d757b6c5ae40d93bcc453c425d86ab119a9354049cffd521afb1f3de7e0e74245fc06cb2f022939ff8e2685aa7987519f08dbfbc34a043b576f647c5a9117b9600f0e9c850d47a6c51ee0f7f6f83676e6e8b21de10aff1beb5a5d12eb22ff6d89bc9a00b525a77392bd0fca3f06d1058d499ff6feafb1541ce540b851a4f1920bdc2ac7d199c332d081260579541f66d5b0ebccfcbba78788b5b43743b1260057902655052f1f299b5d4852c56fe443709d3592addb73a0bf325016bee1624deff6d503a33fdacd7705e7f9a47f3457a78911522340c38b9edbb5e69ae05d6dc78c33fbe2f003cfe0a182bab11d0df9bb7c52178a4e49c25af5690fc716bcad4e877db093ec3962287187752e02a2ad16c51ace16b485e03a88a43e258f902fa8cfdd5b8be0db1c5f2a86b352e7227bb5be86fbe87d7d4a47245b4bbb4a231c4782c5b088f3b7839f38818c642a0831a3e86959ba233b6b5fac8723e901f2d2185c15e05170b674ba942c258c3c490178b9fbd924f2d760de44e02ef97e2dc1057bbd2fb6aa39a64969e2afb6b9632916f3475f7b6b71df30eb9ed3a7762385580b81f20db0951f34cce02b04f76b8bf734e0997033195c6403e53bdfa45e9d1724552cb6916620f7b5db362146638072f217cd53952e61c56f41bfd21fb524220685cbc65eaef4c1b728ef41d076d9b1069b2ec6c4d90f291d1a4fc87e25c9e8615ac0a67b217a645e7a7a8211400f37a9d4f0408992d6c31b970d0a7fd4893cc39243eca6ffa276d26423b5a26001421837747053aa58fc0eea37044384b1433eca593076b7fe97384a58707b5dba84eeba13d2f90032a80c6f8270d53135a72f54b43dc7b6e82fb360bcfc43b30c8b4d9c6fa05ec2c17fd9d20406637e8f0d7f852adb5e2cc6ff04da9aa870dfccd9e1677069cad59105bde64d30de72c328158fcc5cb31246ee21e385cbddfd1a6e0a19630cfcdc707981dfdb4df454e538913595bf27a6bc1963de2432b0b9dbe6a71b43dc67f58316487e2cdb12d86f5c9db9f96726c80ee2fb74cb03e1868ae50dd382777dd627aa85ee6c9077d0de79d76b904dd5fe645c616fdfbfacc1a6572df4089c2a1bd0acb465396bf46f30060cbd78201b4719ecbb053b7c4ec78c025e25b584b0228e8e3cb90f040300378d76141a6474139b0a8eb99b54124d402d955644a78a3902e40038116f3d189c41f29f4e000c5b28a42f718d951372480ac1ed6b78c3a7f170d3e1963c8c635bdc58a928407b6b11ebf69d24fb28ef343d847b7d98589863fb4a4085d4db9d274f9f91f074214e74c9873a4a77490026bb816f49ff981e71e37545cc87cea0056683fa7ae3b371971495a75c079705cdaf247a6a88868885010992c7926071bf5793d6d1fe061e4efcf8e824e63e3bef5293ed2564e8f608ea71cf1c313776ee96670fd1430c9d1af49aa0e3540bf3d8135a60ebc367325278419d3f840e49d8cabc4aba9a024b139687fa867fcf74d2322dda111d5e7d0c892e6ae74b2162f9829819fc1433a4e445a71c0aef8aecb14e2894daef3d833ca814ce721979ec35a18865fcbfc8f4ad516e8aca22bf3c504616e2dfc792e32e443b3c8c0d699cbf1b9d05da0268761ec3d5079748ac9db57403cd197bfad46710d63d91788fea227307edbbe440473f9c4592c10122961205f9a364e97453e6812919ab1fa88ea7460035ad01b6704e25543242004acb3b5302e39ac9a674e160e6a9adcf8616034d345f45c8a55d7789db944b48723d4b90cf14409cd39713fcd336290359015404e25049e8772fe42ade078afa2e41d4f8bfda0cde1f8d069cc465fee7a8f12c7958f020137a234bc64665dfe03dea207fd60027228f85515c04a4d5a9583ec16fec173137e55d6ff7fde97ed8d1bc57f7562a9169fa5b5f7a2413bc72b0695698b9c8e2a0dda1dfae07f47e2edb2033639580f216dba9a151b6a7e6be91ef426d039f12fb1ebccfd95966a8cad4e5af1d41529a8a24f7045b7ce41baa48e42b2b882e319d9f85dee5da7191b3c5c007b7c94c85362d533b4fd45701fe19ced8149547a37144ad0815f2b6bece91f08969236bbe70b9a7f12206fa7ede73a6a783fc3b340634c850e7fe1b94d6aeca5397c8a1e7ae511948c1da616926e5997cdb32bf9861db5860ab271dad72037fbce5643c4fd206e90359609a9f49aa33b868e32abdc04fb6a19c34206d85ed769050e3b378a69bff2445691e084051fd41c456a80e3d30ce725a44399b13e54791555d117c1a3dee47e1c71a1ab64fb042bf07276dadf7054fea1b1777c866987cb109a59c59e3301a113b37888c78a3eaffbe502af461ef30395c4d17b6236411603724ac71782b35ee4430b1da3dce7847273d66ce2f193cbb7787c3a6870a3243c0530d6f20cc3df3a45c8b1b9ffb61802dc60c5e3adaa4a9ca9e79b13a7a04ff07751f465462f133428eb4186b510c267f0d7e16f6d86bc629ae331b6c391abb5115b72eda9be937159f9424b0d13d5b75749637f590b58e549b629bbcf79f768ea7a91cc5c1d989c4d23d55f0bac5bd2b114df652451e679ba2aada14d586bb4c7df4543521178835e1fff3abf0c3d4a0397c2e041388ffe9a4568a7dad9e00e89744d1c454cb55906d3ce415494c83b5a6f9f34a120925e95204490814b", 0x1000}], 0x2}, 0x40001) r2 = socket$inet6(0xa, 0x6, 0x0) connect$inet6(r2, &(0x7f0000000200)={0xa, 0x0, 0x0, @initdev={0xfe, 0x88, '\x00', 0x0, 0x0}, 0x3}, 0x1c) (async) connect$inet6(r2, &(0x7f0000000200)={0xa, 0x0, 0x0, @initdev={0xfe, 0x88, '\x00', 0x0, 0x0}, 0x3}, 0x1c) connect$inet6(r2, &(0x7f0000000040)={0xa, 0x4e20, 0x0, @dev={0xfe, 0x80, '\x00', 0xf}}, 0x1c) sendmmsg$inet6(r2, &(0x7f0000005a00)=[{{0x0, 0x0, 0x0}}, {{0x0, 0x0, 0x0, 0x0, &(0x7f0000005bc0)=ANY=[@ANYBLOB="5000000000000000290000003600000000060000000000000401000730000000000a0000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000020000000000000002900000037000000000000000000000004010000000000001400000000000000290000000000000160000000000000002900000037"], 0x100}}], 0x2, 0x0) 04:12:12 executing program 4: r0 = socket$inet6(0xa, 0x4, 0x7f) ioctl$sock_SIOCETHTOOL(r0, 0x89f0, &(0x7f0000000000)={'bridge0\x00', &(0x7f0000000040)=ANY=[@ANYBLOB="100000000000000000000000000000000000000000000000000000000000000000000000819e19bed0542c39590af27aeaa5c0d596c50ba275c7eb92582d35bc41942369561e1b67acb781287f1832fab0d5635c327f0e4b22218a6cd476e5ee8db8cae57947acb8"]}) r1 = socket$netlink(0x10, 0x3, 0x6) sendmsg$NL80211_CMD_REMAIN_ON_CHANNEL(r1, &(0x7f0000000280)={&(0x7f00000001c0)={0x10, 0x0, 0x0, 0x400}, 0xc, &(0x7f0000000240)={&(0x7f0000000200)={0x3c, 0x0, 0x200, 0x70bd2d, 0x25dfdbff, {{}, {@val={0x8}, @void}}, [@NL80211_ATTR_DURATION={0x8, 0x57, 0x9}, @chandef_params=[@NL80211_ATTR_WIPHY_CHANNEL_TYPE={0x8}, @NL80211_ATTR_WIPHY_CHANNEL_TYPE={0x8, 0x27, 0x3}, @NL80211_ATTR_WIPHY_EDMG_CHANNELS={0x5, 0x118, 0x30}]]}, 0x3c}, 0x1, 0x0, 0x0, 0x801}, 0x20000000) sendmsg$NL80211_CMD_LEAVE_MESH(0xffffffffffffffff, &(0x7f0000000180)={&(0x7f00000000c0)={0x10, 0x0, 0x0, 0x40000}, 0xc, &(0x7f0000000140)={&(0x7f0000000100)={0x14, 0x0, 0x10, 0x70bd27, 0x25dfdbfe, {{}, {@void, @void}}, ["", ""]}, 0x14}, 0x1, 0x0, 0x0, 0x8000}, 0x20000000) 04:12:12 executing program 1: r0 = socket$netlink(0x10, 0x3, 0x0) r1 = socket$netlink(0x10, 0x3, 0x0) r2 = socket(0x10, 0x803, 0x0) sendmsg$NL80211_CMD_CRIT_PROTOCOL_START(r2, &(0x7f0000000580)={0x0, 0x0, &(0x7f0000000540)={0x0, 0x1c}}, 0x0) getsockname$packet(r2, &(0x7f0000000600)={0x11, 0x0, 0x0, 0x1, 0x0, 0x6, @broadcast}, &(0x7f0000000100)=0x14) sendmsg$nl_route(r1, &(0x7f0000000040)={0x0, 0x0, &(0x7f0000000000)={&(0x7f00000000c0)=ANY=[@ANYBLOB="3c00000010008506eb9afc4cd8d06e754a0081c5", @ANYRES32=r3, @ANYBLOB="2377f292252155b21c0012000c000100626f6e64000000000c0002000800010001"], 0x3c}}, 0x0) sendmsg$nl_route(r0, &(0x7f0000000240)={0x0, 0x0, &(0x7f0000000680)={&(0x7f0000000400)=@newlink={0x3c, 0x10, 0xffffff1f, 0x0, 0x7672500, {}, [@IFLA_LINKINFO={0x14, 0x12, 0x0, 0x1, @bridge={{0xb}, {0x4}}}, @IFLA_MASTER={0x8, 0xa, r3}]}, 0x3c}}, 0x0) [ 2804.921069][ T2415] workqueue: Failed to create a rescuer kthread for wq "bond1282": -EINTR [ 2805.249868][ T2426] netlink: 'syz-executor.5': attribute type 1 has an invalid length. 04:12:13 executing program 0: r0 = socket$inet6(0xa, 0x6, 0x0) bind$inet6(r0, &(0x7f0000000000)={0xa, 0x4e20}, 0x1c) listen(r0, 0x6) r1 = socket$inet6(0xa, 0x6, 0x0) connect$inet6(r1, &(0x7f0000000200)={0xa, 0x0, 0x0, @initdev={0xfe, 0x88, '\x00', 0x0, 0x0}, 0x3}, 0x1c) connect$inet6(r1, &(0x7f0000000040)={0xa, 0x4e20, 0x0, @dev={0xfe, 0x80, '\x00', 0xf}}, 0x1c) sendmmsg$inet6(r1, &(0x7f0000005a00)=[{{0x0, 0x0, 0x0}}, {{0x0, 0x0, 0x0, 0x0, &(0x7f0000005bc0)=ANY=[@ANYBLOB="5000000000000000290000003600000000060000000000000401000730000000000a0000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000020000000000000002900000037000000000000000000000004010000000000001400000000000000290000000000000160000000000000002900000037"], 0x100}}], 0x2, 0x0) r2 = socket$inet_udplite(0x2, 0x2, 0x88) ioctl$sock_SIOCGIFINDEX(r2, 0x8933, &(0x7f0000000340)={'bridge_slave_0\x00', 0x0}) r4 = socket(0x10, 0x80002, 0x0) sendmsg$nl_route(r4, &(0x7f0000000000)={0x0, 0x0, &(0x7f0000000080)={&(0x7f00000000c0)=@bridge_setlink={0x2c, 0x13, 0xa29, 0x0, 0x0, {0x7, 0x0, 0x0, r3}, [@IFLA_AF_SPEC={0xc, 0x1a, 0x0, 0x1, [@AF_INET={0x8, 0x4, 0x0, 0x1, {0x4, 0x3}}]}]}, 0x2c}}, 0x0) sendmmsg$inet6(r0, &(0x7f0000002980)=[{{&(0x7f0000000080)={0xa, 0x4e20, 0x10000, @private1={0xfc, 0x1, '\x00', 0x21}, 0x2}, 0x1c, &(0x7f0000000400)=[{&(0x7f00000000c0)="feda2ec1cb303eb00f06a45f8228b563bd35f07111fecd807ec3c6b41b15d1ba07cd7189e2a603df3244c4aa9fb49890e1a6e4d790987930dfb23e8ba45487c0", 0x40}, {&(0x7f0000000100)="2d982b346b4b981b3a8b74359e4c88906a80abac39b20e089d566998c5a951c34ce2218edb9cde7dc3073a0df468a13f928d991f79e437715520a712e80cc5972f68084292c960905a3d", 0x4a}, {&(0x7f0000000240)="dff888c7697d7dd474cd5535142cb3c13ea78c0b07109203a350fe2634ddf4063d22eb0b0058900ec5d7b79b078abbe7cb07d9c2d46c44523dc99c644cd9fec9759e3f58919d0450591745f7e8f5b79a3b86d7a78fe02a3890c1d8f8e7935dbb7865a312ef3928a6026b9ea4df2795953e944cce5cf669e567968ae7836fc511c7c9eac81a90ce71b1b11a57d70aef3fa26c30d573c144ec7481588b86977707fc1ed1e3d4a9bd8593902808b5bd577478651957aefd7b6bc4aa6ab607debdf8631eb24f5e6e1d402f2ad261023c2e3555648b6ccb458e59e4fdaab285f27d3f65f9413a9ed787d40baf13", 0xeb}, {&(0x7f0000000180)="c1c3042b092f696f088c00a19a0b38525ed579b77ba88adbf6eb505cd2ea537ff796feb5afe011cdfcd7dfcdb39feb3de8833b360f768598dcee213cec4c4a093dffb398f59d20e0bd", 0x49}, {&(0x7f0000000340)="5071c6b52b084037bca1a119b09664b64dfb834b13bed5b2e842f0547e4589e915df3c509c91741b95611bddfd6f84451adeccb5d4b15040a1fd828f6dcac794ca33c523b717e0cebddd0eb4c919be88db8b87757a5e42e423fd63abf7a16017ba8b0a527159d2df7c2e2c9e6ae02028cf3aaaed95c7553df648689cc7ab07c6cea5394ff6c0d0d44b35e1230f1354fab75df5af1ebcd292df7d809c1f2d47d3895489b79feb6b46d69c6a", 0xab}], 0x5, &(0x7f0000000480)=[@hoplimit_2292={{0x14, 0x29, 0x8, 0xba90}}, @tclass={{0x14, 0x29, 0x43, 0x9}}, @hopopts={{0x80, 0x29, 0x36, {0x2f, 0xc, '\x00', [@enc_lim={0x4, 0x1, 0xa6}, @generic={0x1f, 0x60, "c012574bc3e513d3be4641846b965d5e881e431f8f202f8bbd06afe5e86a0909840391d4fe82f298137479cd9efff48757fda6a1f22b82363f5aa2ecd0007019d99d38ced76625a9e3ab79f7f5108aa60c36f3b65e7a52bf6d28f213de1f2a0c"}]}}}, @rthdr_2292={{0x38, 0x29, 0x39, {0x2, 0x4, 0x0, 0x2, 0x0, [@mcast1, @remote]}}}, @tclass={{0x14, 0x29, 0x43, 0x7}}, @pktinfo={{0x24, 0x29, 0x32, {@ipv4={'\x00', '\xff\xff', @dev={0xac, 0x14, 0x14, 0x22}}}}}, @dontfrag={{0x14, 0x29, 0x3e, 0x3}}, @dstopts={{0x20, 0x29, 0x37, {0x21, 0x0, '\x00', [@jumbo={0xc2, 0x4, 0xfffffffd}]}}}], 0x160}}, {{&(0x7f0000000600)={0xa, 0x4e20, 0x8, @mcast2, 0x8}, 0x1c, &(0x7f0000000ac0)=[{&(0x7f0000000640)="0e2df84fe5829143c916a56badcdcee434551ca3a72b4b0976aadee23accbbc1e685bd5af90e4e0ed1de612b615f78f742589b1b495eaf85398e7435375d3f5ea9a3ddba57cd9e264a4a42d81070b2bc52088341b197434f6492a9f5341d1db87675fd134382f1ff766c57cd989a8dbd73a8aafbeb751a16eaac909af0fb86c3688b163820c920e87f28e144af3d481b01d1aa2195f23ecc5a4486dc38c277ae17f04980f000a0cf", 0xa8}, {&(0x7f0000000700)="df630a9ea335a9663e8b1fbc4508c0c67e0b68f4744e114977ac354e9c01dcaccfb658f9c06faacad96c8113c7fcdef1f0f687e9884bbadb5b32abfae96bbc2f567073784415a2e5c42c0a19ae853ace864bf31d5461c8750ba2ea0bc65b40bdd7ed6ca3b9f94ecef5d174f05c0b5de1d931d085e5c40212fab15cf564c8ff866969b0471a04caa9bbbea505ef6fe998ed35b3d0d2ca21d5ca48e8c98fa6625eafe025f306b3e7992622f494492d1018e7115f66728d2653b822d5a757a35e4906ebe33bd98b08557b93", 0xca}, {&(0x7f0000000800)="4649f5ab56503a7928b97e4766843a2733955e576147a80bd03f0e3919c2", 0x1e}, {&(0x7f0000000840)="4f83", 0x2}, {&(0x7f0000000880)="18ea84ebd444e97992260b9971c53cf9d5cff9dbd55c54af47e153686c283eec28e1c3c4c995c948266c2af02acf878f6b86484c943b0fc0aac76aa0dca0560e559f6161d64c0facdcb6b3cdde6209cf25215723210005155774e160cdbafd7da96bfa731c37849a320cc1635af8923c06f98d62ef127d50cbac86dead0c94ae29b58e359b704be41223fd8a8db396b1dc08047e93a3440463fb4624ecaecdf0ef371686075f5727a157a815f83b6ef7a17db274a8e1d8a0bfc4329f83157de8cd0134ac367fa9b4375d60d37cc6b5e0439ac7aac9fbe46dde1e02824084d3a6a5225f49634f89e1ef73fda874560bce487cd2c4d5c8a6", 0xf7}, {&(0x7f0000000980)="b6a7cb96b4a0a552afaef96506b7e8ca254428930550b36af3072c85dc2dc8f6b8b48ea5e25ab24b45f7ee19df45aebd4daa3505c2cba99c8a6ec0b40be35f1018bb8a1bcb0e52d4115f85b50530a92af04ce48bde9f2957710c836c3f4e7836fa3aaa1725817e3d11a0a41200a5e5534aac3f30", 0x74}, {&(0x7f0000000a00)="df1461b1a75640df3bcc88e2656187b0540bd6d99b47ccedcc8932ae63648d6e", 0x20}, {&(0x7f0000000a40)="0e07f15eab0e01ef7a843b7f3a28c3e0dac54e3eeeceb8767a5c21463404d9fb4a336970d62692769b432f8c4b0345287f4acce4661a05ef77718120a1c75002a108f4991043bb41acad973ac5089ddaa47b57d9d1a61e88", 0x58}], 0x8}}, {{&(0x7f0000000b40)={0xa, 0x4e24, 0x6d6, @empty, 0x5}, 0x1c, &(0x7f0000000fc0)=[{&(0x7f0000000b80)="6fa5b0d71557abbba8a7593948d41c599b0116db3e2c39f26bc3419146c04e5b84846c5501bb7a891e333d549a7382fa8163f50743568b56fc45d25509b52639460b92a67712ec62731a22141f989f77850aab7973e08383f68aae4246516ce5d17483d06c00e610fa517718e7fdf8c1478638bbb7d47d1dec7de4a3def67bf07cac0a2bf7586af4cc0df9a19f12e95cf294576d65d20572b92dd0fa36a51eb4cd155fdcdf51845461a1924ce5e0dfda", 0xb0}, {&(0x7f0000000c40)="2e863078dacbb12f78f632cece9c7d988aec6f2e7f3586d966343bfd9a9791ecf8b2fb0b1fabd530fcec72edec06bc0ecd48f4435f7342ca87307de880bad518ca0a36b99b2452d72ff4685b4995e7797455d105a2ac585486458967be7034d74f5b2ad49389aadb0570c77ed191f861ea5704e11a35068beb5b42a4bc39a287ee96868e202e419386b7e9fd63e9f5619e7b9a615a9e7917972bd15c3b43bac13a710dc78c07c5c9d10e859f56605030efa106e6d437607db66ecb2d9aaec82d385f12e340c246966f0701dc3d7005d4", 0xd0}, {&(0x7f0000000d40)="9649c68807526393025f4f8dc5c6b612237847a11f5826551719051c85a352f1b0a49c109fdc7ea2e46d2c04109a86f67606d29eb332f0f7e184c0f9fd7c6eb6f414faaac6902a9d890faed749206c52d0b5308a91b03fd083c987641a1fa7155f998e4436880bf4644c9815befd48cc30c39a9c06aa61f5cc266f898e8e456758ffd0292cd37a28d2cc1fbb79ef10980192eeb09e7c0690554c258a39c6c6ccacf53550fbdda778bb17e2d4b3a296059b1bdf3ea3658fd79a7d8ee0", 0xbc}, {&(0x7f0000000e00)="40baaae1d82d9c8bd9dd8c354852f73426a53dfbb116bf264a82cce1f392d6911dc86f3361fe12a52ade3a57772c6e1cdbf9529778f1c476bd26b11f1b1d2ff3d99b0274ff9a30e289fb5ea1650c183a7b12938e99de10078acc486ae2333cd8a27596936ab53ef1c6cd9d24b196f884d574a62f4a53e52eeb88b5e2afd57c9a6d0771574ab69b2fe8b8290bf9b4b451ebd4e191de4ea487a58e06bcb9034178d6479dc3292b31c3292c755edfd17ff2ac8819c65e3f6c2d2aa83261f7f19e4fcc2091642ce8eb8a40d072d940b714bebe4124b100a1220649f6", 0xda}, {&(0x7f0000000f00)="46665e3f183618c716d2b7032becdd1283ebc1d2ec57cb654de7509465c1d4f0a2fc97da640f5eaac5d4e377575552a6aacb4694dcdffcb8082205e6227374aebda7c83ac852964f385c488d08f200b843ed0166b1511a0f666ed923fc7033b9618473fdf62f5822634c061a7a6054a60ab3b582faca4307865119dfede554020388b39bd6b10ca6300ae264c60f032c1da54aa02e7c5b5d3bc3c3901d2eee54a138ff6aa9ac07ac01c1b15b85377e0fe7348e7daec9b7", 0xb7}], 0x5}}, {{&(0x7f0000001040)={0xa, 0x4e22, 0x7590, @initdev={0xfe, 0x88, '\x00', 0x1, 0x0}, 0x8d}, 0x1c, &(0x7f0000002500)=[{&(0x7f0000001080)="91e3822371a251242a663bbd0f1298d43771917894c6d3f330185ed8e26cc4906bd6bc25cbb3d23ccf7791cbc08f58a54693", 0x32}, {&(0x7f00000010c0)="2ff549c0aa7855a23acdb351f7312a01216b7a9b01ed58dbb4e0fdd792eeeac9c89dffab655426c9af9ec7dee035f0339f93ac08298c824ad7b2422eee9ecd40caccb60234cc3eec9ed92e7c397e0ab9ad5ce6a115ed0de7953bde44e75bf8890f18d66dfc88f323a198084015b5e4f4702cdba271a4112af76c222da6fa38ae58d9d163236556546430a7039503afc3121062de5a6fe1eea2cb74a51ac3bd13f9bc4124f970787fd80da4529b3edb0d2a7f22c9511e9c47976bc88e3bf81620837bc43449aa383c04d270c0c7be575bc932e2f9f4217218547867e327dc07f9e649806f7d4b69a898c87a830cee2ce59c372a3de7b38ec09b95cf89208a0e119f94641b1d1c19714e95145e4ad45088ffccb872ecdb3b7ecfb02c92227723d7893b32d9da6e1b8383e741f4c9313eb6ea59012b9450df669922780795f4f4ae75c345aca94c92cc4047fcfa7454fd0c5901e47e18068442188bb640002440dca504284ceffc3da14c5710cfe6e4f3d2ee1e778032d06b6b7bf6eaafebe0fc78d04f2994aa93a9f84d6ae1d0ea4ec37277ab6f219b7fade45c57305f4682041574642a7244c94905ea12ff2e079e36ba0a28156488467792f938624a5c4ff45d07f4c32a50a3387c347cbfcb92c3abb2f735886a65538c9c0214d59dde6c08f88a468e3f35b7e503358266bc845fdc71a392502c713ffcc52dcf46896b407da99465ee18f7a54f9ae7c4c4af66c5fbf6463afffd4ccf54a3762d2996dd0328ecd88e0b9bf43e6216c101410f1546e9c628ae659814987e2d1b571eb38641c5048add8a1a8a6c334b2b2a8e4ec4bbbfa781b192bd3d740c2224e58e8fd2270d2c01a2c18ff66ea9bf0fcbdcae712a2562471f9f293947d7f282492ea3bed0adbb70edcadcd4589a1a0943d4cbe5dc149b0c7c5752202ddd81ab9099cf6eb88f8d3e2dada4eed96c4cbd8ad369de503bb1725b063f5f51b0de40c957e789002845d3e09512ec7b6a182b0279770041436489e39734ce61ce2fe956aa66c5ca71048cc898abcfbca7f0cca9186aa08d2f1b08ed24e8af16a688cfaa28807f4173d52ee4b74a9de9bfd5d3d122e1c8464ea729b62f403793ece706af52ffa77de67e710b6e48a7850b5875e7eef055355ef21eba88ffbc6756b78e29e569ec1c40cf10a81ee3e47e4016f595db2aa1b227623bb51a71cd182dcfe1fbb4aed9d2cd5e3be5cdb0d956e65bb83944d86422b00df9ba15ae6a32ed786f3b5eff269502ae9d79f87a9855cb5bd9c33ef7d4fc87971572185e2db3bf343d814b6ac0d98a895f5128009fb3f91ec8a78a0e236cc5e98e23acc5ab721b85d774cbd48d1e2ce1b69aa7f53b9cf3fc3550993ecb7dd316644a60c7e357e6e50f15b16458fcb913b3b27741c7785a1c9aca2795b5883ce9affbea342b7da00447fc095feda5a3a480325f68b2f81c909acfd82f1fb6fc7dfd44483f90157cb80d361c7f30bca3569cece17fe36f0f2063c57ae4c25230ddb4f975d39b34220e6d4b84cc5d5e2de938acfabdb7a16b8b39f4e569c9c88ee36261d58f23e1446ee63c1f476b1f5cbbc72ace11161db31493a6aa1b74d4c914993be31a7f2ef95d89e1d021db8d0e889c055509eccd48ae7ab14b73450f5c2a5d03110d131683e7b2f777c4399a0012759f8ca2ca60ffeb5ecbd00f2680800c3bf20e6675acd51e0456375d353098c78c50b2a33354322b50b4233dffb281740155c4e13d138e754664f3570f793843042e2734207d72ef29e4e197406c98c6900b05986fa5f905ac3399964d44157b165ca0d80e98c95e8f1b66e7af285715ff8671beba62d65c5af2b6622f47fccf5d925634bafa2da7bb87ae1cc074212d3a3420b0e1b19ff8770e2b40af3457bca64da544904324efb885b625598d7935d65c95a9d45d293a36627565ab3c9ebdc819ab7378e2b61840081de88eb9c68572f66a05a50304b9c3e6a552985b533a1f07838b9f8b876f4477d225563a7f2d43240412714385d7d85f1a379086df6cd37be6f68c532c797c8e22197660fc3cc78f1059bbf730e1e6137aaae2f3dfbcf208a2e8bc5c088d57482e9c80e53e9281a1fe1dd520cda167aeb349d201611eedf9d05c36086e6839845cc75497104db51c8c54e8dfb1af71744ea170cbf5f717b90ba44b68c3c1ad3741bc02122ea49fc3224b4a2a1da1d35af0f306dc0a0db0a67bad5bad5f0cadf43082df025e601abe593a8ca0e03dbd6deb06187a88d5e0b276e8293bcd39c29ea3e0c8ec1e06b89081d0c130476cecc1e4a3cafc31fa6735c547a1f0eec4090e91941ca7ba9b0211593b88f59bc44e5c874ea53110637254754e75b424677e3ecb68a5c939073494db1fbfc19af4bac84330838323b8d93ee2bbb4dccfbecbbec9e3001f6941cda90b09d7d0c213b59cd68d38d21cc36795a127776e0cc387912eef7eab0778f7d5222216b6b9916f2869092a8f6eda97d7abebeb1ecf4853b222f9582e2cbf22ff07e373b9ef2160b048956154a985ed41bf57b5e8cb107a42847421f0d731ea261f81512086c717e59576ed9bd9baf4526656a07c556e55f698efa687510b627487b9c0e923a570a14b3ffb1745628f41f37ab7fe9009219e042c73b04018e0e64ae58d22417a6d5c35f025de673463586890698a8ed78852a894cc21113080d8f37fa9784d48b6c070c1dc1877182e59461335b69c564bfc7848c2c65f4a989304570015ad9b75aa7e7a9dfc85df55dad86b62e4b0527a70ee9cbdfdb8fa146ee8b719dd160521412375209edc325b781c84ec8de9021bd2846447ff588ebad29b462b7b27d0aa2ce4c64ad2e6267f7a60c2ae9d25fd08414dbf44aabcb1fe9be91c2e70f5a8a76eeab27f379bff3a08f593379f326cb082cabff65e917015fe9b7a4fe5c9b223f4ec676800acc9398c0edcdd59aede104681155c2b2d5b5bf3703b766d1ebd1d08406af007a63e0533d1bec12f525c2acd699fd60c65fb7f1c96db5a69ad027d704074f869a858d1ec8b964d50e4eeb2ae193703538207f786a4b3ba862b62327845ae71d62ec1d7fc864e2c0984fb3de7424fc43a7991d68893d2b415f359c4360391c24a9b3ffc0e7b2d92ce4e703a198b4766bb033d70d6ab9a9caa509c135e13b898eff71f2fd9b589a807a22013b2d53f1cdb3ade43b5b3a5798751375fa32129955763cbe2302010fb4fe3a60e7595899083da1bba5b9a1f844891dfc64b71d143686295f0e1579df9fd2fc1667151ead1e97968a84612cc64fab2e808860f02921321f806b72612e95ca9504451e36b126f78a49f5cae249297ea1d067950aca4ca05c43207de5b5946b672703f0198991cc6c83e0c174478b151b20d7263cbf47f92533c3e744cc42579274502e46264d09a927ef16ddcb7d418fcd3472506d2dd0e77595ddc443010f58158ca68eb35c2e191365eb5473850aef59c88263e670a9d09d3c9024ad0451f15f8c863774203637c2546f30950d348bccc6f09b125cef06696d96714252b82d9efbcf3e0daaa1f78f823595f02f55059e08474b25894f3683789c8e95b15de9c34d8ece4f0f5a1fb5ee7872025a9ea59518def338515632ff593c7fc04349b0c50b4e82fcffb627703d898707b1002500289e67289ae1bd5e0c5cafe5a4dfdceba8dea5602eba1dccaa1fd518d112c537d28182e3ef995f81f001e5dc2b894f25f54399456607c04601c230ccc4eee5de05af280d0341c01d133a04ee4ab1d0134d732592986f0a851d7b9c4318dd79c1c32694d62119ad41f8dce4b3565b22942555c8232ff39696ac99953b24fcd1289993558f3dc01e7c5071d77768aacd8c20628fd4acd051c256aa55cb47d489d85383c5be1eb5b3568b0dd415934c2fa6ae5a60d05a8d6e8ec3a8c219b518db452346a65b6086cf7747e72856886f18d4023d339edd126caf0cb5b7dff4f6fa2d3386aaf2f8c349d09da1519ef74192472ddd456ee8953d406a12d63a755d1ae6a646d811ec61b9151b96240781d301603d8c9034ec38d5f8a946a4486ab855c72dcb780ae65b6840cebb1f9a071b73ac813b1a8b218145054b920a2a553975b3ed1267cfcb9f66de5b7b91c0afc1054ce3245c4694a9dc448e40005499793268744dac46c4b7ef65f77eaa220dbde1ced009d7cbcdbb53930ceb260e21f83b9a39201014d38c46b714825f29968c13d528a32cf75baa43ff4dc4c9528c83dba4b25d11d7ddd20425b99b669812da58d9f6a9aa12c7b7eda27d58c94157230954190ec79ea131f3046c97ebc3290d2e67eba20a027989aa2552ac6abd6177c8e1667613b438adafa94f2162cc21e64676b37afacb292085ae4fde365ed3730e6947c8ef925dfefeaad6db5528e292ec16b5e44472b3e3f13b5aa186436c877119a788a2197422cc5e9537dc55a6548ba355ec1a7754f60156c930bd7a38aca09c7b84637bd16470ae41bb624d7405fe80f2600f5785e30fa457e28342c0d4d12deb5a4cf73851f7d8913e35a5346654db4729adcb00e2748985dc878b34c2e5c44062fdb7552d4d4394f2b74187712e20edc275862ca3d24879455f24188c5a198440b020391bb7673952b304046c971917e7ffba36d857219145de8f759cac2bf08b7fdb3c6509d68f521c2f9c89b73eb9e121f79237d89fadf1c310bffa35b7fb312147d29805faa73a13d3682f03e943e64ab0949a70064394ecaa90abbe9a0a89390980ec8700bc23e3d9aaf329bcc3a8994a2fbc7dc24398f51a41cdd80f5d6c3fb4f9735b392273081eb2856b8356eb178f2be0c1c9045d17c15c9071cf29f0df034aaed6aef81adfb4bbdaf3ce69c0e35c111ea4eab117f5d79a9efca4c7ca71ebb93f081b3a7bfa0525d2350eaa7b8f39200e4aa11f3576906d8a2d82a6bf24841669333334a5768e96b87f76ff6aee1f48aa3288d59f1d0c8b0d250b434592311583723dc41b160ee51dc94d3de09754da156045a749a74c8d36d51b56b43ec462264e4ea04b67dc3d576b7c25d8474d96800806b1a5603e17d823f06ec9f60583d8feed21fd2b599cd3df5e6b18288ce2269743bf708ca05733c3b64665ce1882a8b029a1aa3d0f74e985b533b49bf0c16e9ef46975654afcb45742cb9b474a7c22b0da6a3f6c1e276035c3a1677eb93ea3a8bde58d3069867b66c1f96a9f402cdb8a10ea671939b79317253a24cbf833c04d267d2e79b327ca343b975963414b1fb97bc61640e86a76a5a2990c4d35dfba7a62bcc0cf669ca6fbec975c84f35bec5cefe8fb9b1ce4f92d836eb136898a17bffc73f21faf3b3754fafb52e6cf83c958744437dc9112899e277b429d28241527ca10476de6d3c64ba4978ad7cfa6d1ee0173dceedba2d2122c35047f3fb6d6afd42b27c6acaf131bbbef239fac6ae3d39cf978f6f7af360901c2d44012a956117832812143e5f6ed63af1b4861a8a98dd0966284811cee5cebae461ff4f0e9ca416f7039a2ac21058630b39fd855cb351e46759501d4b3e31123ed82d98254c25edc2d402c185d1820f4d2af4151449e5fe8096098988a64bab8442b33ea9cd1cdb5d701a9e5763d21be632e4ce4bac1094a60f824b65176f7e420a382fdc8d20d14bb2082a64286031ec8c136a92af39cd5c4df71ebf82a42c7aedbb0e42db5613cf0cce7bd66b857744dc4712850c6a81743b6db7a3f05cebbeebef7de28a4099c3a1f77200f3e55e1fd3e56573c3062461a5c93c918c4f62026b407cdf1730d36ea7d5c65d064b8db59989598ea19e", 0x1000}, {&(0x7f00000020c0)="66d651e8a1b16cd32e856ad6cb93c167571f4c4414c47551565a61", 0x1b}, {&(0x7f0000002100)="348801661bf341b443670227708ef351a3daf01b7f1f9a3afdd0a821b4f287a9cd75c670a07ad3b6e87f136685dd1f6c3a6a3bcd9f8a441eb29f306a28131e45bf786dcfff7bb83492d6f9167e8884dc064b9b7cc6c40bb207", 0x59}, {&(0x7f0000002180)="c9ac68377e2e686937b2b70e64a53ffab73bf2a8eafabf4d91d9432832b7b2fa5bea34c3c70c947f2dc16377a8d820fda51d8d92a162e55fd52ca23bd66ff088c6e3eafc8462a748ad49be27c3045f841758733d690f6dea5be7a9d181f2ffa0797a46a2fed1289edd5b1c7622f230e00c19efd330570f93ab4f56c42a52a0ff42ad15bf8d20762459353c37a11d312bd58e68de69930f1a4ba09857e0d9cc6a1af11b0f118b7980b82fbcc9ed48d497addd5a42cf13f53dcad63dba62bf616f4650a5302091951d3ca1e3eb82f0cb168cb13439263740b7a25b38c5ef", 0xdd}, {&(0x7f0000002280)="a3bfadb6fbaae52f2e2fa67262b4c9c6cb3f964657d4b6cc3dd5b5aae919f083af5b91eb044f5c6961947eef1407b5d97a6ec25f3a19592a7a8f16c247622b9cad091f5a023b170a8cc2ec8201604fd74681d7f698c9cd1a51b0d1b1a19ced34478b1c0e28aaeb6e0520d7267d89ad9a56e3be486d9f01feac8064a55aa7d2ab3127ec194ceb5bd202ca65ecdcba22dc2e994b925406fc8dc73d8c0902365dc38f46e74b212776cf16507cd0b2e560e2da98c1e4cc5c9c4d8fdead8113e930b737", 0xc1}, {&(0x7f0000002380)="19d74ab20ba78c5642d056aef55b8bdeb008fd7a739e8ddb9dd2ad475bb401bce8994f87ca19384025a23b4186762414d8f558f224882f98840b495ab72bf9ce558c0d91a0430da4991d68b0976e536e1e53ca327442f31cc3b90e2a7f0999cbcf593ad1c07f8af811eb406012a2ac35c6f46adc89fccf7bc0bc296acc011790d5018afaa44e45ff415ac8b179cd755ac43d09b795d1e09e2232ca77ea65c6976e0599cc9509e563879bb36bbccb682259c427b13210d2df8c48677646be1b3d76d60839b1638ec94b99a28ae624cfd7140b39ecac45f398a38eb47e", 0xdc}, {&(0x7f0000002480)="139ee9852a4e2ab434e2223402e01dd988ae06864bfbb88b5240a26e26c6c495454773a54a974e057085726d470fab7c78ec87c1a599b4790d494271d0293cfd89b0d8a7dbd59014b0cdcdb3ba614a873214f72cfbd620de6292c6a1", 0x5c}], 0x8, &(0x7f0000002580)=[@pktinfo={{0x24, 0x29, 0x32, {@loopback, r3}}}, @rthdr={{0x78, 0x29, 0x39, {0x0, 0xc, 0x2, 0x5, 0x0, [@private1={0xfc, 0x1, '\x00', 0x1}, @mcast2, @private2, @rand_addr=' \x01\x00\x00\x00\x00\x00\x00\x00\x00\x00\x00\x00\x00\x00\x01', @loopback, @local]}}}, @tclass={{0x14, 0x29, 0x43, 0x1f}}, @hoplimit={{0x14, 0x29, 0x34, 0x8}}, @dstopts={{0x30, 0x29, 0x37, {0x84, 0x2, '\x00', [@hao={0xc9, 0x10, @local}]}}}, @hopopts_2292={{0x38, 0x29, 0x36, {0x5e, 0x3, '\x00', [@hao={0xc9, 0x10, @mcast2}, @enc_lim={0x4, 0x1, 0x3a}, @ra, @pad1]}}}, @dontfrag={{0x14, 0x29, 0x3e, 0x3}}, @flowinfo={{0x14, 0x29, 0xb, 0x2}}], 0x168}}, {{0x0, 0x0, &(0x7f0000002900)=[{&(0x7f0000002700)="a0983fdced6c2d9679e1a20e6d806c5e440240cfde2971157d70bceb31650cc8fb18409a6d05e411cd7fccaf83667fe3e33c79c4a12ec30e7d5889fd9dd563ef8a7c02eede87fa4adb113b943912f060aa37d5ae8a4b9c548746df17e5d8a61cac4ba3a295f7db02cfa4dceebc8b5cfaac5b136a8450fcc7f2baa097d9e50f7d93615c94a024fc2f4635142cc1c187e5e5e32d0d2f7c6a498f88a15b0a4493cc394231a7964ac6ad785b8ef9e6eefb8ebebfd0797d112b860aba95400fa281390be8a3f6e6ca9d05ade4b96d6aaf5dbc6b07946077f6", 0xd6}, {&(0x7f0000002800)="2746e592a8faabfefe4d29eae51b6a1b2c91bac16245639b70a085f5738fe6242b0aa8676d22ba769fb956a9505a8c9b067038864abb73b72cdab8e2e983af375344754cae612fcc6a74d69d1d844b59c19090e9f19bd45082d6972c467a14888f48764f541b8f6b025318c1729c9af8a52034879a028af3b1feabeae49e4daf08f3fc1adc1141875979f56b22b7df20534b7b41686fd141727f5d94ffbafe5accb656cdf39fcd49d06d2afee19cef2e2784949fab82bb27c6689f649bedbd98950d1111b860332a438b27dd8daa5d99d3e90555492364639fdf5fe1f8c457907dcc99733fa0ad2123bb4036452ab884", 0xf0}], 0x2, &(0x7f0000002940)=[@flowinfo={{0x14, 0x29, 0xb, 0x3}}], 0x18}}], 0x5, 0x4000) [ 2805.416726][ T2426] workqueue: Failed to create a rescuer kthread for wq "bond1274": -EINTR [ 2805.626385][ T2443] netlink: 'syz-executor.2': attribute type 1 has an invalid length. 04:12:13 executing program 5: r0 = socket$netlink(0x10, 0x3, 0x0) r1 = socket$netlink(0x10, 0x3, 0x0) r2 = socket(0x10, 0x803, 0x0) sendmsg$NL80211_CMD_CRIT_PROTOCOL_START(r2, &(0x7f0000000580)={0x0, 0x0, &(0x7f0000000540)={0x0, 0x1c}}, 0x0) getsockname$packet(r2, &(0x7f0000000600)={0x11, 0x0, 0x0, 0x1, 0x0, 0x6, @broadcast}, &(0x7f0000000100)=0x14) sendmsg$nl_route(r1, &(0x7f0000000040)={0x0, 0x0, &(0x7f0000000000)={&(0x7f00000000c0)=ANY=[@ANYBLOB="3c00000010008506eb9afc4cd8d06e754a0081c5", @ANYRES32=r3, @ANYBLOB="2377f292252155b21c0012000c000100626f6e64000000000c0002000800010001"], 0x3c}}, 0x0) sendmsg$nl_route(r0, &(0x7f0000000240)={0x0, 0x0, &(0x7f0000000680)={&(0x7f0000000400)=@newlink={0x3c, 0x10, 0xffffff1f, 0x0, 0x2535f2, {}, [@IFLA_LINKINFO={0x14, 0x12, 0x0, 0x1, @bridge={{0xb}, {0x4}}}, @IFLA_MASTER={0x8, 0xa, r3}]}, 0x3c}}, 0x0) 04:12:13 executing program 4: r0 = socket$inet6(0xa, 0x4, 0x7f) ioctl$sock_SIOCETHTOOL(r0, 0x89f0, &(0x7f0000000000)={'bridge0\x00', &(0x7f0000000040)=ANY=[@ANYBLOB="100000000000000000000000000000000000000000000000000000000000000000000000819e19bed0542c39590af27aeaa5c0d596c50ba275c7eb92582d35bc41942369561e1b67acb781287f1832fab0d5635c327f0e4b22218a6cd476e5ee8db8cae57947acb8"]}) (async) ioctl$sock_SIOCETHTOOL(r0, 0x89f0, &(0x7f0000000000)={'bridge0\x00', &(0x7f0000000040)=ANY=[@ANYBLOB="100000000000000000000000000000000000000000000000000000000000000000000000819e19bed0542c39590af27aeaa5c0d596c50ba275c7eb92582d35bc41942369561e1b67acb781287f1832fab0d5635c327f0e4b22218a6cd476e5ee8db8cae57947acb8"]}) r1 = socket$netlink(0x10, 0x3, 0x6) sendmsg$NL80211_CMD_REMAIN_ON_CHANNEL(r1, &(0x7f0000000280)={&(0x7f00000001c0)={0x10, 0x0, 0x0, 0x400}, 0xc, &(0x7f0000000240)={&(0x7f0000000200)={0x3c, 0x0, 0x200, 0x70bd2d, 0x25dfdbff, {{}, {@val={0x8}, @void}}, [@NL80211_ATTR_DURATION={0x8, 0x57, 0x9}, @chandef_params=[@NL80211_ATTR_WIPHY_CHANNEL_TYPE={0x8}, @NL80211_ATTR_WIPHY_CHANNEL_TYPE={0x8, 0x27, 0x3}, @NL80211_ATTR_WIPHY_EDMG_CHANNELS={0x5, 0x118, 0x30}]]}, 0x3c}, 0x1, 0x0, 0x0, 0x801}, 0x20000000) sendmsg$NL80211_CMD_LEAVE_MESH(0xffffffffffffffff, &(0x7f0000000180)={&(0x7f00000000c0)={0x10, 0x0, 0x0, 0x40000}, 0xc, &(0x7f0000000140)={&(0x7f0000000100)={0x14, 0x0, 0x10, 0x70bd27, 0x25dfdbfe, {{}, {@void, @void}}, ["", ""]}, 0x14}, 0x1, 0x0, 0x0, 0x8000}, 0x20000000) [ 2805.793460][ T2443] bond1263: entered promiscuous mode [ 2805.800114][ T2443] 8021q: adding VLAN 0 to HW filter on device bond1263 [ 2806.125007][ T2445] bond1263: (slave bridge1184): making interface the new active one [ 2806.176647][ T2445] bridge1184: entered promiscuous mode 04:12:13 executing program 2: r0 = socket$netlink(0x10, 0x3, 0x0) r1 = socket$netlink(0x10, 0x3, 0x0) r2 = socket(0x10, 0x803, 0x0) sendmsg$NL80211_CMD_CRIT_PROTOCOL_START(r2, &(0x7f0000000580)={0x0, 0x0, &(0x7f0000000540)={0x0, 0x1c}}, 0x0) getsockname$packet(r2, &(0x7f0000000600)={0x11, 0x0, 0x0, 0x1, 0x0, 0x6, @broadcast}, &(0x7f0000000100)=0x14) sendmsg$nl_route(r1, &(0x7f0000000040)={0x0, 0x0, &(0x7f0000000000)={&(0x7f00000000c0)=ANY=[@ANYBLOB="3c00000010008506eb9afc4cd8d06e754a0081c5", @ANYRES32=r3, @ANYBLOB="2377f292252155b21c0012000c000100626f6e64000000000c0002000800010001"], 0x3c}}, 0x0) sendmsg$nl_route(r0, &(0x7f0000000240)={0x0, 0x0, &(0x7f0000000680)={&(0x7f0000000400)=@newlink={0x3c, 0x10, 0xffffff1f, 0x0, 0x253c21, {}, [@IFLA_LINKINFO={0x14, 0x12, 0x0, 0x1, @bridge={{0xb}, {0x4}}}, @IFLA_MASTER={0x8, 0xa, r3}]}, 0x3c}}, 0x0) [ 2806.229980][ T2445] bond1263: (slave bridge1184): Enslaving as an active interface with an up link 04:12:13 executing program 3: r0 = socket$netlink(0x10, 0x3, 0x0) sendmsg$netlink(r0, &(0x7f00000047c0)={0x0, 0x0, &(0x7f0000000000)=[{&(0x7f0000000100)=ANY=[@ANYBLOB="18000000560001060000006600db730007"], 0x18}], 0x1}, 0x0) 04:12:13 executing program 4: r0 = socket$inet6(0xa, 0x4, 0x7f) ioctl$sock_SIOCETHTOOL(r0, 0x89f0, &(0x7f0000000000)={'bridge0\x00', &(0x7f0000000040)=ANY=[@ANYBLOB="100000000000000000000000000000000000000000000000000000000000000000000000819e19bed0542c39590af27aeaa5c0d596c50ba275c7eb92582d35bc41942369561e1b67acb781287f1832fab0d5635c327f0e4b22218a6cd476e5ee8db8cae57947acb8"]}) (async) r1 = socket$netlink(0x10, 0x3, 0x6) sendmsg$NL80211_CMD_REMAIN_ON_CHANNEL(r1, &(0x7f0000000280)={&(0x7f00000001c0)={0x10, 0x0, 0x0, 0x400}, 0xc, &(0x7f0000000240)={&(0x7f0000000200)={0x3c, 0x0, 0x200, 0x70bd2d, 0x25dfdbff, {{}, {@val={0x8}, @void}}, [@NL80211_ATTR_DURATION={0x8, 0x57, 0x9}, @chandef_params=[@NL80211_ATTR_WIPHY_CHANNEL_TYPE={0x8}, @NL80211_ATTR_WIPHY_CHANNEL_TYPE={0x8, 0x27, 0x3}, @NL80211_ATTR_WIPHY_EDMG_CHANNELS={0x5, 0x118, 0x30}]]}, 0x3c}, 0x1, 0x0, 0x0, 0x801}, 0x20000000) (async) sendmsg$NL80211_CMD_LEAVE_MESH(0xffffffffffffffff, &(0x7f0000000180)={&(0x7f00000000c0)={0x10, 0x0, 0x0, 0x40000}, 0xc, &(0x7f0000000140)={&(0x7f0000000100)={0x14, 0x0, 0x10, 0x70bd27, 0x25dfdbfe, {{}, {@void, @void}}, ["", ""]}, 0x14}, 0x1, 0x0, 0x0, 0x8000}, 0x20000000) 04:12:14 executing program 4: r0 = socket$inet6(0xa, 0x2, 0x0) r1 = socket$inet6(0xa, 0x80002, 0x88) bind$inet6(r1, &(0x7f0000000580)={0xa, 0x10010000004e20}, 0x1c) setsockopt$inet6_udp_int(r1, 0x11, 0xb, &(0x7f0000000080), 0x4) connect$inet6(r1, &(0x7f0000000040)={0xa, 0x4e20, 0x6, @empty, 0x5}, 0x1c) ioctl$sock_SIOCETHTOOL(r0, 0x89f0, &(0x7f0000000000)={'bridge0\x00', &(0x7f0000000200)=@ethtool_ringparam={0x10}}) r2 = socket$inet6(0xa, 0x80002, 0x88) bind$inet6(r2, &(0x7f0000000580)={0xa, 0x10010000004e20}, 0x1c) setsockopt$inet6_udp_int(r2, 0x11, 0xb, &(0x7f0000000080), 0x4) poll(&(0x7f00000000c0)=[{r0, 0x4000}, {r2, 0x11100}], 0x2, 0x5) [ 2806.529867][ T2455] bond1282: entered promiscuous mode [ 2806.536685][ T2455] 8021q: adding VLAN 0 to HW filter on device bond1282 04:12:14 executing program 1: r0 = socket$netlink(0x10, 0x3, 0x0) r1 = socket$netlink(0x10, 0x3, 0x0) r2 = socket(0x10, 0x803, 0x0) sendmsg$NL80211_CMD_CRIT_PROTOCOL_START(r2, &(0x7f0000000580)={0x0, 0x0, &(0x7f0000000540)={0x0, 0x1c}}, 0x0) getsockname$packet(r2, &(0x7f0000000600)={0x11, 0x0, 0x0, 0x1, 0x0, 0x6, @broadcast}, &(0x7f0000000100)=0x14) sendmsg$nl_route(r1, &(0x7f0000000040)={0x0, 0x0, &(0x7f0000000000)={&(0x7f00000000c0)=ANY=[@ANYBLOB="3c00000010008506eb9afc4cd8d06e754a0081c5", @ANYRES32=r3, @ANYBLOB="2377f292252155b21c0012000c000100626f6e64000000000c0002000800010001"], 0x3c}}, 0x0) sendmsg$nl_route(r0, &(0x7f0000000240)={0x0, 0x0, &(0x7f0000000680)={&(0x7f0000000400)=@newlink={0x3c, 0x10, 0xffffff1f, 0x0, 0x8000000, {}, [@IFLA_LINKINFO={0x14, 0x12, 0x0, 0x1, @bridge={{0xb}, {0x4}}}, @IFLA_MASTER={0x8, 0xa, r3}]}, 0x3c}}, 0x0) [ 2806.636630][ T2459] bond1282: (slave bridge1228): making interface the new active one [ 2806.646247][ T2459] bridge1228: entered promiscuous mode [ 2806.657310][ T2459] bond1282: (slave bridge1228): Enslaving as an active interface with an up link [ 2806.680254][ T2468] validate_nla: 1 callbacks suppressed 04:12:14 executing program 0: r0 = socket$inet6(0xa, 0x6, 0x0) bind$inet6(r0, &(0x7f0000000000)={0xa, 0x4e20}, 0x1c) listen(r0, 0x6) (async) r1 = socket$inet6(0xa, 0x6, 0x0) connect$inet6(r1, &(0x7f0000000200)={0xa, 0x0, 0x0, @initdev={0xfe, 0x88, '\x00', 0x0, 0x0}, 0x3}, 0x1c) (async) connect$inet6(r1, &(0x7f0000000040)={0xa, 0x4e20, 0x0, @dev={0xfe, 0x80, '\x00', 0xf}}, 0x1c) (async, rerun: 64) sendmmsg$inet6(r1, &(0x7f0000005a00)=[{{0x0, 0x0, 0x0}}, {{0x0, 0x0, 0x0, 0x0, &(0x7f0000005bc0)=ANY=[@ANYBLOB="5000000000000000290000003600000000060000000000000401000730000000000a0000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000020000000000000002900000037000000000000000000000004010000000000001400000000000000290000000000000160000000000000002900000037"], 0x100}}], 0x2, 0x0) (async, rerun: 64) r2 = socket$inet_udplite(0x2, 0x2, 0x88) ioctl$sock_SIOCGIFINDEX(r2, 0x8933, &(0x7f0000000340)={'bridge_slave_0\x00', 0x0}) (async) r4 = socket(0x10, 0x80002, 0x0) sendmsg$nl_route(r4, &(0x7f0000000000)={0x0, 0x0, &(0x7f0000000080)={&(0x7f00000000c0)=@bridge_setlink={0x2c, 0x13, 0xa29, 0x0, 0x0, {0x7, 0x0, 0x0, r3}, [@IFLA_AF_SPEC={0xc, 0x1a, 0x0, 0x1, [@AF_INET={0x8, 0x4, 0x0, 0x1, {0x4, 0x3}}]}]}, 0x2c}}, 0x0) (async) sendmmsg$inet6(r0, &(0x7f0000002980)=[{{&(0x7f0000000080)={0xa, 0x4e20, 0x10000, @private1={0xfc, 0x1, '\x00', 0x21}, 0x2}, 0x1c, &(0x7f0000000400)=[{&(0x7f00000000c0)="feda2ec1cb303eb00f06a45f8228b563bd35f07111fecd807ec3c6b41b15d1ba07cd7189e2a603df3244c4aa9fb49890e1a6e4d790987930dfb23e8ba45487c0", 0x40}, {&(0x7f0000000100)="2d982b346b4b981b3a8b74359e4c88906a80abac39b20e089d566998c5a951c34ce2218edb9cde7dc3073a0df468a13f928d991f79e437715520a712e80cc5972f68084292c960905a3d", 0x4a}, {&(0x7f0000000240)="dff888c7697d7dd474cd5535142cb3c13ea78c0b07109203a350fe2634ddf4063d22eb0b0058900ec5d7b79b078abbe7cb07d9c2d46c44523dc99c644cd9fec9759e3f58919d0450591745f7e8f5b79a3b86d7a78fe02a3890c1d8f8e7935dbb7865a312ef3928a6026b9ea4df2795953e944cce5cf669e567968ae7836fc511c7c9eac81a90ce71b1b11a57d70aef3fa26c30d573c144ec7481588b86977707fc1ed1e3d4a9bd8593902808b5bd577478651957aefd7b6bc4aa6ab607debdf8631eb24f5e6e1d402f2ad261023c2e3555648b6ccb458e59e4fdaab285f27d3f65f9413a9ed787d40baf13", 0xeb}, {&(0x7f0000000180)="c1c3042b092f696f088c00a19a0b38525ed579b77ba88adbf6eb505cd2ea537ff796feb5afe011cdfcd7dfcdb39feb3de8833b360f768598dcee213cec4c4a093dffb398f59d20e0bd", 0x49}, {&(0x7f0000000340)="5071c6b52b084037bca1a119b09664b64dfb834b13bed5b2e842f0547e4589e915df3c509c91741b95611bddfd6f84451adeccb5d4b15040a1fd828f6dcac794ca33c523b717e0cebddd0eb4c919be88db8b87757a5e42e423fd63abf7a16017ba8b0a527159d2df7c2e2c9e6ae02028cf3aaaed95c7553df648689cc7ab07c6cea5394ff6c0d0d44b35e1230f1354fab75df5af1ebcd292df7d809c1f2d47d3895489b79feb6b46d69c6a", 0xab}], 0x5, &(0x7f0000000480)=[@hoplimit_2292={{0x14, 0x29, 0x8, 0xba90}}, @tclass={{0x14, 0x29, 0x43, 0x9}}, @hopopts={{0x80, 0x29, 0x36, {0x2f, 0xc, '\x00', [@enc_lim={0x4, 0x1, 0xa6}, @generic={0x1f, 0x60, "c012574bc3e513d3be4641846b965d5e881e431f8f202f8bbd06afe5e86a0909840391d4fe82f298137479cd9efff48757fda6a1f22b82363f5aa2ecd0007019d99d38ced76625a9e3ab79f7f5108aa60c36f3b65e7a52bf6d28f213de1f2a0c"}]}}}, @rthdr_2292={{0x38, 0x29, 0x39, {0x2, 0x4, 0x0, 0x2, 0x0, [@mcast1, @remote]}}}, @tclass={{0x14, 0x29, 0x43, 0x7}}, @pktinfo={{0x24, 0x29, 0x32, {@ipv4={'\x00', '\xff\xff', @dev={0xac, 0x14, 0x14, 0x22}}}}}, @dontfrag={{0x14, 0x29, 0x3e, 0x3}}, @dstopts={{0x20, 0x29, 0x37, {0x21, 0x0, '\x00', [@jumbo={0xc2, 0x4, 0xfffffffd}]}}}], 0x160}}, {{&(0x7f0000000600)={0xa, 0x4e20, 0x8, @mcast2, 0x8}, 0x1c, &(0x7f0000000ac0)=[{&(0x7f0000000640)="0e2df84fe5829143c916a56badcdcee434551ca3a72b4b0976aadee23accbbc1e685bd5af90e4e0ed1de612b615f78f742589b1b495eaf85398e7435375d3f5ea9a3ddba57cd9e264a4a42d81070b2bc52088341b197434f6492a9f5341d1db87675fd134382f1ff766c57cd989a8dbd73a8aafbeb751a16eaac909af0fb86c3688b163820c920e87f28e144af3d481b01d1aa2195f23ecc5a4486dc38c277ae17f04980f000a0cf", 0xa8}, {&(0x7f0000000700)="df630a9ea335a9663e8b1fbc4508c0c67e0b68f4744e114977ac354e9c01dcaccfb658f9c06faacad96c8113c7fcdef1f0f687e9884bbadb5b32abfae96bbc2f567073784415a2e5c42c0a19ae853ace864bf31d5461c8750ba2ea0bc65b40bdd7ed6ca3b9f94ecef5d174f05c0b5de1d931d085e5c40212fab15cf564c8ff866969b0471a04caa9bbbea505ef6fe998ed35b3d0d2ca21d5ca48e8c98fa6625eafe025f306b3e7992622f494492d1018e7115f66728d2653b822d5a757a35e4906ebe33bd98b08557b93", 0xca}, {&(0x7f0000000800)="4649f5ab56503a7928b97e4766843a2733955e576147a80bd03f0e3919c2", 0x1e}, {&(0x7f0000000840)="4f83", 0x2}, {&(0x7f0000000880)="18ea84ebd444e97992260b9971c53cf9d5cff9dbd55c54af47e153686c283eec28e1c3c4c995c948266c2af02acf878f6b86484c943b0fc0aac76aa0dca0560e559f6161d64c0facdcb6b3cdde6209cf25215723210005155774e160cdbafd7da96bfa731c37849a320cc1635af8923c06f98d62ef127d50cbac86dead0c94ae29b58e359b704be41223fd8a8db396b1dc08047e93a3440463fb4624ecaecdf0ef371686075f5727a157a815f83b6ef7a17db274a8e1d8a0bfc4329f83157de8cd0134ac367fa9b4375d60d37cc6b5e0439ac7aac9fbe46dde1e02824084d3a6a5225f49634f89e1ef73fda874560bce487cd2c4d5c8a6", 0xf7}, {&(0x7f0000000980)="b6a7cb96b4a0a552afaef96506b7e8ca254428930550b36af3072c85dc2dc8f6b8b48ea5e25ab24b45f7ee19df45aebd4daa3505c2cba99c8a6ec0b40be35f1018bb8a1bcb0e52d4115f85b50530a92af04ce48bde9f2957710c836c3f4e7836fa3aaa1725817e3d11a0a41200a5e5534aac3f30", 0x74}, {&(0x7f0000000a00)="df1461b1a75640df3bcc88e2656187b0540bd6d99b47ccedcc8932ae63648d6e", 0x20}, {&(0x7f0000000a40)="0e07f15eab0e01ef7a843b7f3a28c3e0dac54e3eeeceb8767a5c21463404d9fb4a336970d62692769b432f8c4b0345287f4acce4661a05ef77718120a1c75002a108f4991043bb41acad973ac5089ddaa47b57d9d1a61e88", 0x58}], 0x8}}, {{&(0x7f0000000b40)={0xa, 0x4e24, 0x6d6, @empty, 0x5}, 0x1c, &(0x7f0000000fc0)=[{&(0x7f0000000b80)="6fa5b0d71557abbba8a7593948d41c599b0116db3e2c39f26bc3419146c04e5b84846c5501bb7a891e333d549a7382fa8163f50743568b56fc45d25509b52639460b92a67712ec62731a22141f989f77850aab7973e08383f68aae4246516ce5d17483d06c00e610fa517718e7fdf8c1478638bbb7d47d1dec7de4a3def67bf07cac0a2bf7586af4cc0df9a19f12e95cf294576d65d20572b92dd0fa36a51eb4cd155fdcdf51845461a1924ce5e0dfda", 0xb0}, {&(0x7f0000000c40)="2e863078dacbb12f78f632cece9c7d988aec6f2e7f3586d966343bfd9a9791ecf8b2fb0b1fabd530fcec72edec06bc0ecd48f4435f7342ca87307de880bad518ca0a36b99b2452d72ff4685b4995e7797455d105a2ac585486458967be7034d74f5b2ad49389aadb0570c77ed191f861ea5704e11a35068beb5b42a4bc39a287ee96868e202e419386b7e9fd63e9f5619e7b9a615a9e7917972bd15c3b43bac13a710dc78c07c5c9d10e859f56605030efa106e6d437607db66ecb2d9aaec82d385f12e340c246966f0701dc3d7005d4", 0xd0}, {&(0x7f0000000d40)="9649c68807526393025f4f8dc5c6b612237847a11f5826551719051c85a352f1b0a49c109fdc7ea2e46d2c04109a86f67606d29eb332f0f7e184c0f9fd7c6eb6f414faaac6902a9d890faed749206c52d0b5308a91b03fd083c987641a1fa7155f998e4436880bf4644c9815befd48cc30c39a9c06aa61f5cc266f898e8e456758ffd0292cd37a28d2cc1fbb79ef10980192eeb09e7c0690554c258a39c6c6ccacf53550fbdda778bb17e2d4b3a296059b1bdf3ea3658fd79a7d8ee0", 0xbc}, {&(0x7f0000000e00)="40baaae1d82d9c8bd9dd8c354852f73426a53dfbb116bf264a82cce1f392d6911dc86f3361fe12a52ade3a57772c6e1cdbf9529778f1c476bd26b11f1b1d2ff3d99b0274ff9a30e289fb5ea1650c183a7b12938e99de10078acc486ae2333cd8a27596936ab53ef1c6cd9d24b196f884d574a62f4a53e52eeb88b5e2afd57c9a6d0771574ab69b2fe8b8290bf9b4b451ebd4e191de4ea487a58e06bcb9034178d6479dc3292b31c3292c755edfd17ff2ac8819c65e3f6c2d2aa83261f7f19e4fcc2091642ce8eb8a40d072d940b714bebe4124b100a1220649f6", 0xda}, {&(0x7f0000000f00)="46665e3f183618c716d2b7032becdd1283ebc1d2ec57cb654de7509465c1d4f0a2fc97da640f5eaac5d4e377575552a6aacb4694dcdffcb8082205e6227374aebda7c83ac852964f385c488d08f200b843ed0166b1511a0f666ed923fc7033b9618473fdf62f5822634c061a7a6054a60ab3b582faca4307865119dfede554020388b39bd6b10ca6300ae264c60f032c1da54aa02e7c5b5d3bc3c3901d2eee54a138ff6aa9ac07ac01c1b15b85377e0fe7348e7daec9b7", 0xb7}], 0x5}}, {{&(0x7f0000001040)={0xa, 0x4e22, 0x7590, @initdev={0xfe, 0x88, '\x00', 0x1, 0x0}, 0x8d}, 0x1c, &(0x7f0000002500)=[{&(0x7f0000001080)="91e3822371a251242a663bbd0f1298d43771917894c6d3f330185ed8e26cc4906bd6bc25cbb3d23ccf7791cbc08f58a54693", 0x32}, {&(0x7f00000010c0)="2ff549c0aa7855a23acdb351f7312a01216b7a9b01ed58dbb4e0fdd792eeeac9c89dffab655426c9af9ec7dee035f0339f93ac08298c824ad7b2422eee9ecd40caccb60234cc3eec9ed92e7c397e0ab9ad5ce6a115ed0de7953bde44e75bf8890f18d66dfc88f323a198084015b5e4f4702cdba271a4112af76c222da6fa38ae58d9d163236556546430a7039503afc3121062de5a6fe1eea2cb74a51ac3bd13f9bc4124f970787fd80da4529b3edb0d2a7f22c9511e9c47976bc88e3bf81620837bc43449aa383c04d270c0c7be575bc932e2f9f4217218547867e327dc07f9e649806f7d4b69a898c87a830cee2ce59c372a3de7b38ec09b95cf89208a0e119f94641b1d1c19714e95145e4ad45088ffccb872ecdb3b7ecfb02c92227723d7893b32d9da6e1b8383e741f4c9313eb6ea59012b9450df669922780795f4f4ae75c345aca94c92cc4047fcfa7454fd0c5901e47e18068442188bb640002440dca504284ceffc3da14c5710cfe6e4f3d2ee1e778032d06b6b7bf6eaafebe0fc78d04f2994aa93a9f84d6ae1d0ea4ec37277ab6f219b7fade45c57305f4682041574642a7244c94905ea12ff2e079e36ba0a28156488467792f938624a5c4ff45d07f4c32a50a3387c347cbfcb92c3abb2f735886a65538c9c0214d59dde6c08f88a468e3f35b7e503358266bc845fdc71a392502c713ffcc52dcf46896b407da99465ee18f7a54f9ae7c4c4af66c5fbf6463afffd4ccf54a3762d2996dd0328ecd88e0b9bf43e6216c101410f1546e9c628ae659814987e2d1b571eb38641c5048add8a1a8a6c334b2b2a8e4ec4bbbfa781b192bd3d740c2224e58e8fd2270d2c01a2c18ff66ea9bf0fcbdcae712a2562471f9f293947d7f282492ea3bed0adbb70edcadcd4589a1a0943d4cbe5dc149b0c7c5752202ddd81ab9099cf6eb88f8d3e2dada4eed96c4cbd8ad369de503bb1725b063f5f51b0de40c957e789002845d3e09512ec7b6a182b0279770041436489e39734ce61ce2fe956aa66c5ca71048cc898abcfbca7f0cca9186aa08d2f1b08ed24e8af16a688cfaa28807f4173d52ee4b74a9de9bfd5d3d122e1c8464ea729b62f403793ece706af52ffa77de67e710b6e48a7850b5875e7eef055355ef21eba88ffbc6756b78e29e569ec1c40cf10a81ee3e47e4016f595db2aa1b227623bb51a71cd182dcfe1fbb4aed9d2cd5e3be5cdb0d956e65bb83944d86422b00df9ba15ae6a32ed786f3b5eff269502ae9d79f87a9855cb5bd9c33ef7d4fc87971572185e2db3bf343d814b6ac0d98a895f5128009fb3f91ec8a78a0e236cc5e98e23acc5ab721b85d774cbd48d1e2ce1b69aa7f53b9cf3fc3550993ecb7dd316644a60c7e357e6e50f15b16458fcb913b3b27741c7785a1c9aca2795b5883ce9affbea342b7da00447fc095feda5a3a480325f68b2f81c909acfd82f1fb6fc7dfd44483f90157cb80d361c7f30bca3569cece17fe36f0f2063c57ae4c25230ddb4f975d39b34220e6d4b84cc5d5e2de938acfabdb7a16b8b39f4e569c9c88ee36261d58f23e1446ee63c1f476b1f5cbbc72ace11161db31493a6aa1b74d4c914993be31a7f2ef95d89e1d021db8d0e889c055509eccd48ae7ab14b73450f5c2a5d03110d131683e7b2f777c4399a0012759f8ca2ca60ffeb5ecbd00f2680800c3bf20e6675acd51e0456375d353098c78c50b2a33354322b50b4233dffb281740155c4e13d138e754664f3570f793843042e2734207d72ef29e4e197406c98c6900b05986fa5f905ac3399964d44157b165ca0d80e98c95e8f1b66e7af285715ff8671beba62d65c5af2b6622f47fccf5d925634bafa2da7bb87ae1cc074212d3a3420b0e1b19ff8770e2b40af3457bca64da544904324efb885b625598d7935d65c95a9d45d293a36627565ab3c9ebdc819ab7378e2b61840081de88eb9c68572f66a05a50304b9c3e6a552985b533a1f07838b9f8b876f4477d225563a7f2d43240412714385d7d85f1a379086df6cd37be6f68c532c797c8e22197660fc3cc78f1059bbf730e1e6137aaae2f3dfbcf208a2e8bc5c088d57482e9c80e53e9281a1fe1dd520cda167aeb349d201611eedf9d05c36086e6839845cc75497104db51c8c54e8dfb1af71744ea170cbf5f717b90ba44b68c3c1ad3741bc02122ea49fc3224b4a2a1da1d35af0f306dc0a0db0a67bad5bad5f0cadf43082df025e601abe593a8ca0e03dbd6deb06187a88d5e0b276e8293bcd39c29ea3e0c8ec1e06b89081d0c130476cecc1e4a3cafc31fa6735c547a1f0eec4090e91941ca7ba9b0211593b88f59bc44e5c874ea53110637254754e75b424677e3ecb68a5c939073494db1fbfc19af4bac84330838323b8d93ee2bbb4dccfbecbbec9e3001f6941cda90b09d7d0c213b59cd68d38d21cc36795a127776e0cc387912eef7eab0778f7d5222216b6b9916f2869092a8f6eda97d7abebeb1ecf4853b222f9582e2cbf22ff07e373b9ef2160b048956154a985ed41bf57b5e8cb107a42847421f0d731ea261f81512086c717e59576ed9bd9baf4526656a07c556e55f698efa687510b627487b9c0e923a570a14b3ffb1745628f41f37ab7fe9009219e042c73b04018e0e64ae58d22417a6d5c35f025de673463586890698a8ed78852a894cc21113080d8f37fa9784d48b6c070c1dc1877182e59461335b69c564bfc7848c2c65f4a989304570015ad9b75aa7e7a9dfc85df55dad86b62e4b0527a70ee9cbdfdb8fa146ee8b719dd160521412375209edc325b781c84ec8de9021bd2846447ff588ebad29b462b7b27d0aa2ce4c64ad2e6267f7a60c2ae9d25fd08414dbf44aabcb1fe9be91c2e70f5a8a76eeab27f379bff3a08f593379f326cb082cabff65e917015fe9b7a4fe5c9b223f4ec676800acc9398c0edcdd59aede104681155c2b2d5b5bf3703b766d1ebd1d08406af007a63e0533d1bec12f525c2acd699fd60c65fb7f1c96db5a69ad027d704074f869a858d1ec8b964d50e4eeb2ae193703538207f786a4b3ba862b62327845ae71d62ec1d7fc864e2c0984fb3de7424fc43a7991d68893d2b415f359c4360391c24a9b3ffc0e7b2d92ce4e703a198b4766bb033d70d6ab9a9caa509c135e13b898eff71f2fd9b589a807a22013b2d53f1cdb3ade43b5b3a5798751375fa32129955763cbe2302010fb4fe3a60e7595899083da1bba5b9a1f844891dfc64b71d143686295f0e1579df9fd2fc1667151ead1e97968a84612cc64fab2e808860f02921321f806b72612e95ca9504451e36b126f78a49f5cae249297ea1d067950aca4ca05c43207de5b5946b672703f0198991cc6c83e0c174478b151b20d7263cbf47f92533c3e744cc42579274502e46264d09a927ef16ddcb7d418fcd3472506d2dd0e77595ddc443010f58158ca68eb35c2e191365eb5473850aef59c88263e670a9d09d3c9024ad0451f15f8c863774203637c2546f30950d348bccc6f09b125cef06696d96714252b82d9efbcf3e0daaa1f78f823595f02f55059e08474b25894f3683789c8e95b15de9c34d8ece4f0f5a1fb5ee7872025a9ea59518def338515632ff593c7fc04349b0c50b4e82fcffb627703d898707b1002500289e67289ae1bd5e0c5cafe5a4dfdceba8dea5602eba1dccaa1fd518d112c537d28182e3ef995f81f001e5dc2b894f25f54399456607c04601c230ccc4eee5de05af280d0341c01d133a04ee4ab1d0134d732592986f0a851d7b9c4318dd79c1c32694d62119ad41f8dce4b3565b22942555c8232ff39696ac99953b24fcd1289993558f3dc01e7c5071d77768aacd8c20628fd4acd051c256aa55cb47d489d85383c5be1eb5b3568b0dd415934c2fa6ae5a60d05a8d6e8ec3a8c219b518db452346a65b6086cf7747e72856886f18d4023d339edd126caf0cb5b7dff4f6fa2d3386aaf2f8c349d09da1519ef74192472ddd456ee8953d406a12d63a755d1ae6a646d811ec61b9151b96240781d301603d8c9034ec38d5f8a946a4486ab855c72dcb780ae65b6840cebb1f9a071b73ac813b1a8b218145054b920a2a553975b3ed1267cfcb9f66de5b7b91c0afc1054ce3245c4694a9dc448e40005499793268744dac46c4b7ef65f77eaa220dbde1ced009d7cbcdbb53930ceb260e21f83b9a39201014d38c46b714825f29968c13d528a32cf75baa43ff4dc4c9528c83dba4b25d11d7ddd20425b99b669812da58d9f6a9aa12c7b7eda27d58c94157230954190ec79ea131f3046c97ebc3290d2e67eba20a027989aa2552ac6abd6177c8e1667613b438adafa94f2162cc21e64676b37afacb292085ae4fde365ed3730e6947c8ef925dfefeaad6db5528e292ec16b5e44472b3e3f13b5aa186436c877119a788a2197422cc5e9537dc55a6548ba355ec1a7754f60156c930bd7a38aca09c7b84637bd16470ae41bb624d7405fe80f2600f5785e30fa457e28342c0d4d12deb5a4cf73851f7d8913e35a5346654db4729adcb00e2748985dc878b34c2e5c44062fdb7552d4d4394f2b74187712e20edc275862ca3d24879455f24188c5a198440b020391bb7673952b304046c971917e7ffba36d857219145de8f759cac2bf08b7fdb3c6509d68f521c2f9c89b73eb9e121f79237d89fadf1c310bffa35b7fb312147d29805faa73a13d3682f03e943e64ab0949a70064394ecaa90abbe9a0a89390980ec8700bc23e3d9aaf329bcc3a8994a2fbc7dc24398f51a41cdd80f5d6c3fb4f9735b392273081eb2856b8356eb178f2be0c1c9045d17c15c9071cf29f0df034aaed6aef81adfb4bbdaf3ce69c0e35c111ea4eab117f5d79a9efca4c7ca71ebb93f081b3a7bfa0525d2350eaa7b8f39200e4aa11f3576906d8a2d82a6bf24841669333334a5768e96b87f76ff6aee1f48aa3288d59f1d0c8b0d250b434592311583723dc41b160ee51dc94d3de09754da156045a749a74c8d36d51b56b43ec462264e4ea04b67dc3d576b7c25d8474d96800806b1a5603e17d823f06ec9f60583d8feed21fd2b599cd3df5e6b18288ce2269743bf708ca05733c3b64665ce1882a8b029a1aa3d0f74e985b533b49bf0c16e9ef46975654afcb45742cb9b474a7c22b0da6a3f6c1e276035c3a1677eb93ea3a8bde58d3069867b66c1f96a9f402cdb8a10ea671939b79317253a24cbf833c04d267d2e79b327ca343b975963414b1fb97bc61640e86a76a5a2990c4d35dfba7a62bcc0cf669ca6fbec975c84f35bec5cefe8fb9b1ce4f92d836eb136898a17bffc73f21faf3b3754fafb52e6cf83c958744437dc9112899e277b429d28241527ca10476de6d3c64ba4978ad7cfa6d1ee0173dceedba2d2122c35047f3fb6d6afd42b27c6acaf131bbbef239fac6ae3d39cf978f6f7af360901c2d44012a956117832812143e5f6ed63af1b4861a8a98dd0966284811cee5cebae461ff4f0e9ca416f7039a2ac21058630b39fd855cb351e46759501d4b3e31123ed82d98254c25edc2d402c185d1820f4d2af4151449e5fe8096098988a64bab8442b33ea9cd1cdb5d701a9e5763d21be632e4ce4bac1094a60f824b65176f7e420a382fdc8d20d14bb2082a64286031ec8c136a92af39cd5c4df71ebf82a42c7aedbb0e42db5613cf0cce7bd66b857744dc4712850c6a81743b6db7a3f05cebbeebef7de28a4099c3a1f77200f3e55e1fd3e56573c3062461a5c93c918c4f62026b407cdf1730d36ea7d5c65d064b8db59989598ea19e", 0x1000}, {&(0x7f00000020c0)="66d651e8a1b16cd32e856ad6cb93c167571f4c4414c47551565a61", 0x1b}, {&(0x7f0000002100)="348801661bf341b443670227708ef351a3daf01b7f1f9a3afdd0a821b4f287a9cd75c670a07ad3b6e87f136685dd1f6c3a6a3bcd9f8a441eb29f306a28131e45bf786dcfff7bb83492d6f9167e8884dc064b9b7cc6c40bb207", 0x59}, {&(0x7f0000002180)="c9ac68377e2e686937b2b70e64a53ffab73bf2a8eafabf4d91d9432832b7b2fa5bea34c3c70c947f2dc16377a8d820fda51d8d92a162e55fd52ca23bd66ff088c6e3eafc8462a748ad49be27c3045f841758733d690f6dea5be7a9d181f2ffa0797a46a2fed1289edd5b1c7622f230e00c19efd330570f93ab4f56c42a52a0ff42ad15bf8d20762459353c37a11d312bd58e68de69930f1a4ba09857e0d9cc6a1af11b0f118b7980b82fbcc9ed48d497addd5a42cf13f53dcad63dba62bf616f4650a5302091951d3ca1e3eb82f0cb168cb13439263740b7a25b38c5ef", 0xdd}, {&(0x7f0000002280)="a3bfadb6fbaae52f2e2fa67262b4c9c6cb3f964657d4b6cc3dd5b5aae919f083af5b91eb044f5c6961947eef1407b5d97a6ec25f3a19592a7a8f16c247622b9cad091f5a023b170a8cc2ec8201604fd74681d7f698c9cd1a51b0d1b1a19ced34478b1c0e28aaeb6e0520d7267d89ad9a56e3be486d9f01feac8064a55aa7d2ab3127ec194ceb5bd202ca65ecdcba22dc2e994b925406fc8dc73d8c0902365dc38f46e74b212776cf16507cd0b2e560e2da98c1e4cc5c9c4d8fdead8113e930b737", 0xc1}, {&(0x7f0000002380)="19d74ab20ba78c5642d056aef55b8bdeb008fd7a739e8ddb9dd2ad475bb401bce8994f87ca19384025a23b4186762414d8f558f224882f98840b495ab72bf9ce558c0d91a0430da4991d68b0976e536e1e53ca327442f31cc3b90e2a7f0999cbcf593ad1c07f8af811eb406012a2ac35c6f46adc89fccf7bc0bc296acc011790d5018afaa44e45ff415ac8b179cd755ac43d09b795d1e09e2232ca77ea65c6976e0599cc9509e563879bb36bbccb682259c427b13210d2df8c48677646be1b3d76d60839b1638ec94b99a28ae624cfd7140b39ecac45f398a38eb47e", 0xdc}, {&(0x7f0000002480)="139ee9852a4e2ab434e2223402e01dd988ae06864bfbb88b5240a26e26c6c495454773a54a974e057085726d470fab7c78ec87c1a599b4790d494271d0293cfd89b0d8a7dbd59014b0cdcdb3ba614a873214f72cfbd620de6292c6a1", 0x5c}], 0x8, &(0x7f0000002580)=[@pktinfo={{0x24, 0x29, 0x32, {@loopback, r3}}}, @rthdr={{0x78, 0x29, 0x39, {0x0, 0xc, 0x2, 0x5, 0x0, [@private1={0xfc, 0x1, '\x00', 0x1}, @mcast2, @private2, @rand_addr=' \x01\x00\x00\x00\x00\x00\x00\x00\x00\x00\x00\x00\x00\x00\x01', @loopback, @local]}}}, @tclass={{0x14, 0x29, 0x43, 0x1f}}, @hoplimit={{0x14, 0x29, 0x34, 0x8}}, @dstopts={{0x30, 0x29, 0x37, {0x84, 0x2, '\x00', [@hao={0xc9, 0x10, @local}]}}}, @hopopts_2292={{0x38, 0x29, 0x36, {0x5e, 0x3, '\x00', [@hao={0xc9, 0x10, @mcast2}, @enc_lim={0x4, 0x1, 0x3a}, @ra, @pad1]}}}, @dontfrag={{0x14, 0x29, 0x3e, 0x3}}, @flowinfo={{0x14, 0x29, 0xb, 0x2}}], 0x168}}, {{0x0, 0x0, &(0x7f0000002900)=[{&(0x7f0000002700)="a0983fdced6c2d9679e1a20e6d806c5e440240cfde2971157d70bceb31650cc8fb18409a6d05e411cd7fccaf83667fe3e33c79c4a12ec30e7d5889fd9dd563ef8a7c02eede87fa4adb113b943912f060aa37d5ae8a4b9c548746df17e5d8a61cac4ba3a295f7db02cfa4dceebc8b5cfaac5b136a8450fcc7f2baa097d9e50f7d93615c94a024fc2f4635142cc1c187e5e5e32d0d2f7c6a498f88a15b0a4493cc394231a7964ac6ad785b8ef9e6eefb8ebebfd0797d112b860aba95400fa281390be8a3f6e6ca9d05ade4b96d6aaf5dbc6b07946077f6", 0xd6}, {&(0x7f0000002800)="2746e592a8faabfefe4d29eae51b6a1b2c91bac16245639b70a085f5738fe6242b0aa8676d22ba769fb956a9505a8c9b067038864abb73b72cdab8e2e983af375344754cae612fcc6a74d69d1d844b59c19090e9f19bd45082d6972c467a14888f48764f541b8f6b025318c1729c9af8a52034879a028af3b1feabeae49e4daf08f3fc1adc1141875979f56b22b7df20534b7b41686fd141727f5d94ffbafe5accb656cdf39fcd49d06d2afee19cef2e2784949fab82bb27c6689f649bedbd98950d1111b860332a438b27dd8daa5d99d3e90555492364639fdf5fe1f8c457907dcc99733fa0ad2123bb4036452ab884", 0xf0}], 0x2, &(0x7f0000002940)=[@flowinfo={{0x14, 0x29, 0xb, 0x3}}], 0x18}}], 0x5, 0x4000) [ 2806.680276][ T2468] netlink: 'syz-executor.5': attribute type 1 has an invalid length. 04:12:14 executing program 5: r0 = socket$netlink(0x10, 0x3, 0x0) r1 = socket$netlink(0x10, 0x3, 0x0) r2 = socket(0x10, 0x803, 0x0) sendmsg$NL80211_CMD_CRIT_PROTOCOL_START(r2, &(0x7f0000000580)={0x0, 0x0, &(0x7f0000000540)={0x0, 0x1c}}, 0x0) getsockname$packet(r2, &(0x7f0000000600)={0x11, 0x0, 0x0, 0x1, 0x0, 0x6, @broadcast}, &(0x7f0000000100)=0x14) sendmsg$nl_route(r1, &(0x7f0000000040)={0x0, 0x0, &(0x7f0000000000)={&(0x7f00000000c0)=ANY=[@ANYBLOB="3c00000010008506eb9afc4cd8d06e754a0081c5", @ANYRES32=r3, @ANYBLOB="2377f292252155b21c0012000c000100626f6e64000000000c0002000800010001"], 0x3c}}, 0x0) sendmsg$nl_route(r0, &(0x7f0000000240)={0x0, 0x0, &(0x7f0000000680)={&(0x7f0000000400)=@newlink={0x3c, 0x10, 0xffffff1f, 0x0, 0x2535f3, {}, [@IFLA_LINKINFO={0x14, 0x12, 0x0, 0x1, @bridge={{0xb}, {0x4}}}, @IFLA_MASTER={0x8, 0xa, r3}]}, 0x3c}}, 0x0) 04:12:14 executing program 0: r0 = socket$inet6(0xa, 0x6, 0x0) bind$inet6(r0, &(0x7f0000000000)={0xa, 0x4e20}, 0x1c) listen(r0, 0x6) r1 = socket$inet6(0xa, 0x6, 0x0) connect$inet6(r1, &(0x7f0000000200)={0xa, 0x0, 0x0, @initdev={0xfe, 0x88, '\x00', 0x0, 0x0}, 0x3}, 0x1c) connect$inet6(r1, &(0x7f0000000040)={0xa, 0x4e20, 0x0, @dev={0xfe, 0x80, '\x00', 0xf}}, 0x1c) sendmmsg$inet6(r1, &(0x7f0000005a00)=[{{0x0, 0x0, 0x0}}, {{0x0, 0x0, 0x0, 0x0, &(0x7f0000005bc0)=ANY=[@ANYBLOB="5000000000000000290000003600000000060000000000000401000730000000000a0000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000020000000000000002900000037000000000000000000000004010000000000001400000000000000290000000000000160000000000000002900000037"], 0x100}}], 0x2, 0x0) r2 = socket$inet_udplite(0x2, 0x2, 0x88) ioctl$sock_SIOCGIFINDEX(r2, 0x8933, &(0x7f0000000340)={'bridge_slave_0\x00', 0x0}) r4 = socket(0x10, 0x80002, 0x0) sendmsg$nl_route(r4, &(0x7f0000000000)={0x0, 0x0, &(0x7f0000000080)={&(0x7f00000000c0)=@bridge_setlink={0x2c, 0x13, 0xa29, 0x0, 0x0, {0x7, 0x0, 0x0, r3}, [@IFLA_AF_SPEC={0xc, 0x1a, 0x0, 0x1, [@AF_INET={0x8, 0x4, 0x0, 0x1, {0x4, 0x3}}]}]}, 0x2c}}, 0x0) sendmmsg$inet6(r0, &(0x7f0000002980)=[{{&(0x7f0000000080)={0xa, 0x4e20, 0x10000, @private1={0xfc, 0x1, '\x00', 0x21}, 0x2}, 0x1c, &(0x7f0000000400)=[{&(0x7f00000000c0)="feda2ec1cb303eb00f06a45f8228b563bd35f07111fecd807ec3c6b41b15d1ba07cd7189e2a603df3244c4aa9fb49890e1a6e4d790987930dfb23e8ba45487c0", 0x40}, {&(0x7f0000000100)="2d982b346b4b981b3a8b74359e4c88906a80abac39b20e089d566998c5a951c34ce2218edb9cde7dc3073a0df468a13f928d991f79e437715520a712e80cc5972f68084292c960905a3d", 0x4a}, {&(0x7f0000000240)="dff888c7697d7dd474cd5535142cb3c13ea78c0b07109203a350fe2634ddf4063d22eb0b0058900ec5d7b79b078abbe7cb07d9c2d46c44523dc99c644cd9fec9759e3f58919d0450591745f7e8f5b79a3b86d7a78fe02a3890c1d8f8e7935dbb7865a312ef3928a6026b9ea4df2795953e944cce5cf669e567968ae7836fc511c7c9eac81a90ce71b1b11a57d70aef3fa26c30d573c144ec7481588b86977707fc1ed1e3d4a9bd8593902808b5bd577478651957aefd7b6bc4aa6ab607debdf8631eb24f5e6e1d402f2ad261023c2e3555648b6ccb458e59e4fdaab285f27d3f65f9413a9ed787d40baf13", 0xeb}, {&(0x7f0000000180)="c1c3042b092f696f088c00a19a0b38525ed579b77ba88adbf6eb505cd2ea537ff796feb5afe011cdfcd7dfcdb39feb3de8833b360f768598dcee213cec4c4a093dffb398f59d20e0bd", 0x49}, {&(0x7f0000000340)="5071c6b52b084037bca1a119b09664b64dfb834b13bed5b2e842f0547e4589e915df3c509c91741b95611bddfd6f84451adeccb5d4b15040a1fd828f6dcac794ca33c523b717e0cebddd0eb4c919be88db8b87757a5e42e423fd63abf7a16017ba8b0a527159d2df7c2e2c9e6ae02028cf3aaaed95c7553df648689cc7ab07c6cea5394ff6c0d0d44b35e1230f1354fab75df5af1ebcd292df7d809c1f2d47d3895489b79feb6b46d69c6a", 0xab}], 0x5, &(0x7f0000000480)=[@hoplimit_2292={{0x14, 0x29, 0x8, 0xba90}}, @tclass={{0x14, 0x29, 0x43, 0x9}}, @hopopts={{0x80, 0x29, 0x36, {0x2f, 0xc, '\x00', [@enc_lim={0x4, 0x1, 0xa6}, @generic={0x1f, 0x60, "c012574bc3e513d3be4641846b965d5e881e431f8f202f8bbd06afe5e86a0909840391d4fe82f298137479cd9efff48757fda6a1f22b82363f5aa2ecd0007019d99d38ced76625a9e3ab79f7f5108aa60c36f3b65e7a52bf6d28f213de1f2a0c"}]}}}, @rthdr_2292={{0x38, 0x29, 0x39, {0x2, 0x4, 0x0, 0x2, 0x0, [@mcast1, @remote]}}}, @tclass={{0x14, 0x29, 0x43, 0x7}}, @pktinfo={{0x24, 0x29, 0x32, {@ipv4={'\x00', '\xff\xff', @dev={0xac, 0x14, 0x14, 0x22}}}}}, @dontfrag={{0x14, 0x29, 0x3e, 0x3}}, @dstopts={{0x20, 0x29, 0x37, {0x21, 0x0, '\x00', [@jumbo={0xc2, 0x4, 0xfffffffd}]}}}], 0x160}}, {{&(0x7f0000000600)={0xa, 0x4e20, 0x8, @mcast2, 0x8}, 0x1c, &(0x7f0000000ac0)=[{&(0x7f0000000640)="0e2df84fe5829143c916a56badcdcee434551ca3a72b4b0976aadee23accbbc1e685bd5af90e4e0ed1de612b615f78f742589b1b495eaf85398e7435375d3f5ea9a3ddba57cd9e264a4a42d81070b2bc52088341b197434f6492a9f5341d1db87675fd134382f1ff766c57cd989a8dbd73a8aafbeb751a16eaac909af0fb86c3688b163820c920e87f28e144af3d481b01d1aa2195f23ecc5a4486dc38c277ae17f04980f000a0cf", 0xa8}, {&(0x7f0000000700)="df630a9ea335a9663e8b1fbc4508c0c67e0b68f4744e114977ac354e9c01dcaccfb658f9c06faacad96c8113c7fcdef1f0f687e9884bbadb5b32abfae96bbc2f567073784415a2e5c42c0a19ae853ace864bf31d5461c8750ba2ea0bc65b40bdd7ed6ca3b9f94ecef5d174f05c0b5de1d931d085e5c40212fab15cf564c8ff866969b0471a04caa9bbbea505ef6fe998ed35b3d0d2ca21d5ca48e8c98fa6625eafe025f306b3e7992622f494492d1018e7115f66728d2653b822d5a757a35e4906ebe33bd98b08557b93", 0xca}, {&(0x7f0000000800)="4649f5ab56503a7928b97e4766843a2733955e576147a80bd03f0e3919c2", 0x1e}, {&(0x7f0000000840)="4f83", 0x2}, {&(0x7f0000000880)="18ea84ebd444e97992260b9971c53cf9d5cff9dbd55c54af47e153686c283eec28e1c3c4c995c948266c2af02acf878f6b86484c943b0fc0aac76aa0dca0560e559f6161d64c0facdcb6b3cdde6209cf25215723210005155774e160cdbafd7da96bfa731c37849a320cc1635af8923c06f98d62ef127d50cbac86dead0c94ae29b58e359b704be41223fd8a8db396b1dc08047e93a3440463fb4624ecaecdf0ef371686075f5727a157a815f83b6ef7a17db274a8e1d8a0bfc4329f83157de8cd0134ac367fa9b4375d60d37cc6b5e0439ac7aac9fbe46dde1e02824084d3a6a5225f49634f89e1ef73fda874560bce487cd2c4d5c8a6", 0xf7}, {&(0x7f0000000980)="b6a7cb96b4a0a552afaef96506b7e8ca254428930550b36af3072c85dc2dc8f6b8b48ea5e25ab24b45f7ee19df45aebd4daa3505c2cba99c8a6ec0b40be35f1018bb8a1bcb0e52d4115f85b50530a92af04ce48bde9f2957710c836c3f4e7836fa3aaa1725817e3d11a0a41200a5e5534aac3f30", 0x74}, {&(0x7f0000000a00)="df1461b1a75640df3bcc88e2656187b0540bd6d99b47ccedcc8932ae63648d6e", 0x20}, {&(0x7f0000000a40)="0e07f15eab0e01ef7a843b7f3a28c3e0dac54e3eeeceb8767a5c21463404d9fb4a336970d62692769b432f8c4b0345287f4acce4661a05ef77718120a1c75002a108f4991043bb41acad973ac5089ddaa47b57d9d1a61e88", 0x58}], 0x8}}, {{&(0x7f0000000b40)={0xa, 0x4e24, 0x6d6, @empty, 0x5}, 0x1c, &(0x7f0000000fc0)=[{&(0x7f0000000b80)="6fa5b0d71557abbba8a7593948d41c599b0116db3e2c39f26bc3419146c04e5b84846c5501bb7a891e333d549a7382fa8163f50743568b56fc45d25509b52639460b92a67712ec62731a22141f989f77850aab7973e08383f68aae4246516ce5d17483d06c00e610fa517718e7fdf8c1478638bbb7d47d1dec7de4a3def67bf07cac0a2bf7586af4cc0df9a19f12e95cf294576d65d20572b92dd0fa36a51eb4cd155fdcdf51845461a1924ce5e0dfda", 0xb0}, {&(0x7f0000000c40)="2e863078dacbb12f78f632cece9c7d988aec6f2e7f3586d966343bfd9a9791ecf8b2fb0b1fabd530fcec72edec06bc0ecd48f4435f7342ca87307de880bad518ca0a36b99b2452d72ff4685b4995e7797455d105a2ac585486458967be7034d74f5b2ad49389aadb0570c77ed191f861ea5704e11a35068beb5b42a4bc39a287ee96868e202e419386b7e9fd63e9f5619e7b9a615a9e7917972bd15c3b43bac13a710dc78c07c5c9d10e859f56605030efa106e6d437607db66ecb2d9aaec82d385f12e340c246966f0701dc3d7005d4", 0xd0}, {&(0x7f0000000d40)="9649c68807526393025f4f8dc5c6b612237847a11f5826551719051c85a352f1b0a49c109fdc7ea2e46d2c04109a86f67606d29eb332f0f7e184c0f9fd7c6eb6f414faaac6902a9d890faed749206c52d0b5308a91b03fd083c987641a1fa7155f998e4436880bf4644c9815befd48cc30c39a9c06aa61f5cc266f898e8e456758ffd0292cd37a28d2cc1fbb79ef10980192eeb09e7c0690554c258a39c6c6ccacf53550fbdda778bb17e2d4b3a296059b1bdf3ea3658fd79a7d8ee0", 0xbc}, {&(0x7f0000000e00)="40baaae1d82d9c8bd9dd8c354852f73426a53dfbb116bf264a82cce1f392d6911dc86f3361fe12a52ade3a57772c6e1cdbf9529778f1c476bd26b11f1b1d2ff3d99b0274ff9a30e289fb5ea1650c183a7b12938e99de10078acc486ae2333cd8a27596936ab53ef1c6cd9d24b196f884d574a62f4a53e52eeb88b5e2afd57c9a6d0771574ab69b2fe8b8290bf9b4b451ebd4e191de4ea487a58e06bcb9034178d6479dc3292b31c3292c755edfd17ff2ac8819c65e3f6c2d2aa83261f7f19e4fcc2091642ce8eb8a40d072d940b714bebe4124b100a1220649f6", 0xda}, {&(0x7f0000000f00)="46665e3f183618c716d2b7032becdd1283ebc1d2ec57cb654de7509465c1d4f0a2fc97da640f5eaac5d4e377575552a6aacb4694dcdffcb8082205e6227374aebda7c83ac852964f385c488d08f200b843ed0166b1511a0f666ed923fc7033b9618473fdf62f5822634c061a7a6054a60ab3b582faca4307865119dfede554020388b39bd6b10ca6300ae264c60f032c1da54aa02e7c5b5d3bc3c3901d2eee54a138ff6aa9ac07ac01c1b15b85377e0fe7348e7daec9b7", 0xb7}], 0x5}}, {{&(0x7f0000001040)={0xa, 0x4e22, 0x7590, @initdev={0xfe, 0x88, '\x00', 0x1, 0x0}, 0x8d}, 0x1c, &(0x7f0000002500)=[{&(0x7f0000001080)="91e3822371a251242a663bbd0f1298d43771917894c6d3f330185ed8e26cc4906bd6bc25cbb3d23ccf7791cbc08f58a54693", 0x32}, {&(0x7f00000010c0)="2ff549c0aa7855a23acdb351f7312a01216b7a9b01ed58dbb4e0fdd792eeeac9c89dffab655426c9af9ec7dee035f0339f93ac08298c824ad7b2422eee9ecd40caccb60234cc3eec9ed92e7c397e0ab9ad5ce6a115ed0de7953bde44e75bf8890f18d66dfc88f323a198084015b5e4f4702cdba271a4112af76c222da6fa38ae58d9d163236556546430a7039503afc3121062de5a6fe1eea2cb74a51ac3bd13f9bc4124f970787fd80da4529b3edb0d2a7f22c9511e9c47976bc88e3bf81620837bc43449aa383c04d270c0c7be575bc932e2f9f4217218547867e327dc07f9e649806f7d4b69a898c87a830cee2ce59c372a3de7b38ec09b95cf89208a0e119f94641b1d1c19714e95145e4ad45088ffccb872ecdb3b7ecfb02c92227723d7893b32d9da6e1b8383e741f4c9313eb6ea59012b9450df669922780795f4f4ae75c345aca94c92cc4047fcfa7454fd0c5901e47e18068442188bb640002440dca504284ceffc3da14c5710cfe6e4f3d2ee1e778032d06b6b7bf6eaafebe0fc78d04f2994aa93a9f84d6ae1d0ea4ec37277ab6f219b7fade45c57305f4682041574642a7244c94905ea12ff2e079e36ba0a28156488467792f938624a5c4ff45d07f4c32a50a3387c347cbfcb92c3abb2f735886a65538c9c0214d59dde6c08f88a468e3f35b7e503358266bc845fdc71a392502c713ffcc52dcf46896b407da99465ee18f7a54f9ae7c4c4af66c5fbf6463afffd4ccf54a3762d2996dd0328ecd88e0b9bf43e6216c101410f1546e9c628ae659814987e2d1b571eb38641c5048add8a1a8a6c334b2b2a8e4ec4bbbfa781b192bd3d740c2224e58e8fd2270d2c01a2c18ff66ea9bf0fcbdcae712a2562471f9f293947d7f282492ea3bed0adbb70edcadcd4589a1a0943d4cbe5dc149b0c7c5752202ddd81ab9099cf6eb88f8d3e2dada4eed96c4cbd8ad369de503bb1725b063f5f51b0de40c957e789002845d3e09512ec7b6a182b0279770041436489e39734ce61ce2fe956aa66c5ca71048cc898abcfbca7f0cca9186aa08d2f1b08ed24e8af16a688cfaa28807f4173d52ee4b74a9de9bfd5d3d122e1c8464ea729b62f403793ece706af52ffa77de67e710b6e48a7850b5875e7eef055355ef21eba88ffbc6756b78e29e569ec1c40cf10a81ee3e47e4016f595db2aa1b227623bb51a71cd182dcfe1fbb4aed9d2cd5e3be5cdb0d956e65bb83944d86422b00df9ba15ae6a32ed786f3b5eff269502ae9d79f87a9855cb5bd9c33ef7d4fc87971572185e2db3bf343d814b6ac0d98a895f5128009fb3f91ec8a78a0e236cc5e98e23acc5ab721b85d774cbd48d1e2ce1b69aa7f53b9cf3fc3550993ecb7dd316644a60c7e357e6e50f15b16458fcb913b3b27741c7785a1c9aca2795b5883ce9affbea342b7da00447fc095feda5a3a480325f68b2f81c909acfd82f1fb6fc7dfd44483f90157cb80d361c7f30bca3569cece17fe36f0f2063c57ae4c25230ddb4f975d39b34220e6d4b84cc5d5e2de938acfabdb7a16b8b39f4e569c9c88ee36261d58f23e1446ee63c1f476b1f5cbbc72ace11161db31493a6aa1b74d4c914993be31a7f2ef95d89e1d021db8d0e889c055509eccd48ae7ab14b73450f5c2a5d03110d131683e7b2f777c4399a0012759f8ca2ca60ffeb5ecbd00f2680800c3bf20e6675acd51e0456375d353098c78c50b2a33354322b50b4233dffb281740155c4e13d138e754664f3570f793843042e2734207d72ef29e4e197406c98c6900b05986fa5f905ac3399964d44157b165ca0d80e98c95e8f1b66e7af285715ff8671beba62d65c5af2b6622f47fccf5d925634bafa2da7bb87ae1cc074212d3a3420b0e1b19ff8770e2b40af3457bca64da544904324efb885b625598d7935d65c95a9d45d293a36627565ab3c9ebdc819ab7378e2b61840081de88eb9c68572f66a05a50304b9c3e6a552985b533a1f07838b9f8b876f4477d225563a7f2d43240412714385d7d85f1a379086df6cd37be6f68c532c797c8e22197660fc3cc78f1059bbf730e1e6137aaae2f3dfbcf208a2e8bc5c088d57482e9c80e53e9281a1fe1dd520cda167aeb349d201611eedf9d05c36086e6839845cc75497104db51c8c54e8dfb1af71744ea170cbf5f717b90ba44b68c3c1ad3741bc02122ea49fc3224b4a2a1da1d35af0f306dc0a0db0a67bad5bad5f0cadf43082df025e601abe593a8ca0e03dbd6deb06187a88d5e0b276e8293bcd39c29ea3e0c8ec1e06b89081d0c130476cecc1e4a3cafc31fa6735c547a1f0eec4090e91941ca7ba9b0211593b88f59bc44e5c874ea53110637254754e75b424677e3ecb68a5c939073494db1fbfc19af4bac84330838323b8d93ee2bbb4dccfbecbbec9e3001f6941cda90b09d7d0c213b59cd68d38d21cc36795a127776e0cc387912eef7eab0778f7d5222216b6b9916f2869092a8f6eda97d7abebeb1ecf4853b222f9582e2cbf22ff07e373b9ef2160b048956154a985ed41bf57b5e8cb107a42847421f0d731ea261f81512086c717e59576ed9bd9baf4526656a07c556e55f698efa687510b627487b9c0e923a570a14b3ffb1745628f41f37ab7fe9009219e042c73b04018e0e64ae58d22417a6d5c35f025de673463586890698a8ed78852a894cc21113080d8f37fa9784d48b6c070c1dc1877182e59461335b69c564bfc7848c2c65f4a989304570015ad9b75aa7e7a9dfc85df55dad86b62e4b0527a70ee9cbdfdb8fa146ee8b719dd160521412375209edc325b781c84ec8de9021bd2846447ff588ebad29b462b7b27d0aa2ce4c64ad2e6267f7a60c2ae9d25fd08414dbf44aabcb1fe9be91c2e70f5a8a76eeab27f379bff3a08f593379f326cb082cabff65e917015fe9b7a4fe5c9b223f4ec676800acc9398c0edcdd59aede104681155c2b2d5b5bf3703b766d1ebd1d08406af007a63e0533d1bec12f525c2acd699fd60c65fb7f1c96db5a69ad027d704074f869a858d1ec8b964d50e4eeb2ae193703538207f786a4b3ba862b62327845ae71d62ec1d7fc864e2c0984fb3de7424fc43a7991d68893d2b415f359c4360391c24a9b3ffc0e7b2d92ce4e703a198b4766bb033d70d6ab9a9caa509c135e13b898eff71f2fd9b589a807a22013b2d53f1cdb3ade43b5b3a5798751375fa32129955763cbe2302010fb4fe3a60e7595899083da1bba5b9a1f844891dfc64b71d143686295f0e1579df9fd2fc1667151ead1e97968a84612cc64fab2e808860f02921321f806b72612e95ca9504451e36b126f78a49f5cae249297ea1d067950aca4ca05c43207de5b5946b672703f0198991cc6c83e0c174478b151b20d7263cbf47f92533c3e744cc42579274502e46264d09a927ef16ddcb7d418fcd3472506d2dd0e77595ddc443010f58158ca68eb35c2e191365eb5473850aef59c88263e670a9d09d3c9024ad0451f15f8c863774203637c2546f30950d348bccc6f09b125cef06696d96714252b82d9efbcf3e0daaa1f78f823595f02f55059e08474b25894f3683789c8e95b15de9c34d8ece4f0f5a1fb5ee7872025a9ea59518def338515632ff593c7fc04349b0c50b4e82fcffb627703d898707b1002500289e67289ae1bd5e0c5cafe5a4dfdceba8dea5602eba1dccaa1fd518d112c537d28182e3ef995f81f001e5dc2b894f25f54399456607c04601c230ccc4eee5de05af280d0341c01d133a04ee4ab1d0134d732592986f0a851d7b9c4318dd79c1c32694d62119ad41f8dce4b3565b22942555c8232ff39696ac99953b24fcd1289993558f3dc01e7c5071d77768aacd8c20628fd4acd051c256aa55cb47d489d85383c5be1eb5b3568b0dd415934c2fa6ae5a60d05a8d6e8ec3a8c219b518db452346a65b6086cf7747e72856886f18d4023d339edd126caf0cb5b7dff4f6fa2d3386aaf2f8c349d09da1519ef74192472ddd456ee8953d406a12d63a755d1ae6a646d811ec61b9151b96240781d301603d8c9034ec38d5f8a946a4486ab855c72dcb780ae65b6840cebb1f9a071b73ac813b1a8b218145054b920a2a553975b3ed1267cfcb9f66de5b7b91c0afc1054ce3245c4694a9dc448e40005499793268744dac46c4b7ef65f77eaa220dbde1ced009d7cbcdbb53930ceb260e21f83b9a39201014d38c46b714825f29968c13d528a32cf75baa43ff4dc4c9528c83dba4b25d11d7ddd20425b99b669812da58d9f6a9aa12c7b7eda27d58c94157230954190ec79ea131f3046c97ebc3290d2e67eba20a027989aa2552ac6abd6177c8e1667613b438adafa94f2162cc21e64676b37afacb292085ae4fde365ed3730e6947c8ef925dfefeaad6db5528e292ec16b5e44472b3e3f13b5aa186436c877119a788a2197422cc5e9537dc55a6548ba355ec1a7754f60156c930bd7a38aca09c7b84637bd16470ae41bb624d7405fe80f2600f5785e30fa457e28342c0d4d12deb5a4cf73851f7d8913e35a5346654db4729adcb00e2748985dc878b34c2e5c44062fdb7552d4d4394f2b74187712e20edc275862ca3d24879455f24188c5a198440b020391bb7673952b304046c971917e7ffba36d857219145de8f759cac2bf08b7fdb3c6509d68f521c2f9c89b73eb9e121f79237d89fadf1c310bffa35b7fb312147d29805faa73a13d3682f03e943e64ab0949a70064394ecaa90abbe9a0a89390980ec8700bc23e3d9aaf329bcc3a8994a2fbc7dc24398f51a41cdd80f5d6c3fb4f9735b392273081eb2856b8356eb178f2be0c1c9045d17c15c9071cf29f0df034aaed6aef81adfb4bbdaf3ce69c0e35c111ea4eab117f5d79a9efca4c7ca71ebb93f081b3a7bfa0525d2350eaa7b8f39200e4aa11f3576906d8a2d82a6bf24841669333334a5768e96b87f76ff6aee1f48aa3288d59f1d0c8b0d250b434592311583723dc41b160ee51dc94d3de09754da156045a749a74c8d36d51b56b43ec462264e4ea04b67dc3d576b7c25d8474d96800806b1a5603e17d823f06ec9f60583d8feed21fd2b599cd3df5e6b18288ce2269743bf708ca05733c3b64665ce1882a8b029a1aa3d0f74e985b533b49bf0c16e9ef46975654afcb45742cb9b474a7c22b0da6a3f6c1e276035c3a1677eb93ea3a8bde58d3069867b66c1f96a9f402cdb8a10ea671939b79317253a24cbf833c04d267d2e79b327ca343b975963414b1fb97bc61640e86a76a5a2990c4d35dfba7a62bcc0cf669ca6fbec975c84f35bec5cefe8fb9b1ce4f92d836eb136898a17bffc73f21faf3b3754fafb52e6cf83c958744437dc9112899e277b429d28241527ca10476de6d3c64ba4978ad7cfa6d1ee0173dceedba2d2122c35047f3fb6d6afd42b27c6acaf131bbbef239fac6ae3d39cf978f6f7af360901c2d44012a956117832812143e5f6ed63af1b4861a8a98dd0966284811cee5cebae461ff4f0e9ca416f7039a2ac21058630b39fd855cb351e46759501d4b3e31123ed82d98254c25edc2d402c185d1820f4d2af4151449e5fe8096098988a64bab8442b33ea9cd1cdb5d701a9e5763d21be632e4ce4bac1094a60f824b65176f7e420a382fdc8d20d14bb2082a64286031ec8c136a92af39cd5c4df71ebf82a42c7aedbb0e42db5613cf0cce7bd66b857744dc4712850c6a81743b6db7a3f05cebbeebef7de28a4099c3a1f77200f3e55e1fd3e56573c3062461a5c93c918c4f62026b407cdf1730d36ea7d5c65d064b8db59989598ea19e", 0x1000}, {&(0x7f00000020c0)="66d651e8a1b16cd32e856ad6cb93c167571f4c4414c47551565a61", 0x1b}, {&(0x7f0000002100)="348801661bf341b443670227708ef351a3daf01b7f1f9a3afdd0a821b4f287a9cd75c670a07ad3b6e87f136685dd1f6c3a6a3bcd9f8a441eb29f306a28131e45bf786dcfff7bb83492d6f9167e8884dc064b9b7cc6c40bb207", 0x59}, {&(0x7f0000002180)="c9ac68377e2e686937b2b70e64a53ffab73bf2a8eafabf4d91d9432832b7b2fa5bea34c3c70c947f2dc16377a8d820fda51d8d92a162e55fd52ca23bd66ff088c6e3eafc8462a748ad49be27c3045f841758733d690f6dea5be7a9d181f2ffa0797a46a2fed1289edd5b1c7622f230e00c19efd330570f93ab4f56c42a52a0ff42ad15bf8d20762459353c37a11d312bd58e68de69930f1a4ba09857e0d9cc6a1af11b0f118b7980b82fbcc9ed48d497addd5a42cf13f53dcad63dba62bf616f4650a5302091951d3ca1e3eb82f0cb168cb13439263740b7a25b38c5ef", 0xdd}, {&(0x7f0000002280)="a3bfadb6fbaae52f2e2fa67262b4c9c6cb3f964657d4b6cc3dd5b5aae919f083af5b91eb044f5c6961947eef1407b5d97a6ec25f3a19592a7a8f16c247622b9cad091f5a023b170a8cc2ec8201604fd74681d7f698c9cd1a51b0d1b1a19ced34478b1c0e28aaeb6e0520d7267d89ad9a56e3be486d9f01feac8064a55aa7d2ab3127ec194ceb5bd202ca65ecdcba22dc2e994b925406fc8dc73d8c0902365dc38f46e74b212776cf16507cd0b2e560e2da98c1e4cc5c9c4d8fdead8113e930b737", 0xc1}, {&(0x7f0000002380)="19d74ab20ba78c5642d056aef55b8bdeb008fd7a739e8ddb9dd2ad475bb401bce8994f87ca19384025a23b4186762414d8f558f224882f98840b495ab72bf9ce558c0d91a0430da4991d68b0976e536e1e53ca327442f31cc3b90e2a7f0999cbcf593ad1c07f8af811eb406012a2ac35c6f46adc89fccf7bc0bc296acc011790d5018afaa44e45ff415ac8b179cd755ac43d09b795d1e09e2232ca77ea65c6976e0599cc9509e563879bb36bbccb682259c427b13210d2df8c48677646be1b3d76d60839b1638ec94b99a28ae624cfd7140b39ecac45f398a38eb47e", 0xdc}, {&(0x7f0000002480)="139ee9852a4e2ab434e2223402e01dd988ae06864bfbb88b5240a26e26c6c495454773a54a974e057085726d470fab7c78ec87c1a599b4790d494271d0293cfd89b0d8a7dbd59014b0cdcdb3ba614a873214f72cfbd620de6292c6a1", 0x5c}], 0x8, &(0x7f0000002580)=[@pktinfo={{0x24, 0x29, 0x32, {@loopback, r3}}}, @rthdr={{0x78, 0x29, 0x39, {0x0, 0xc, 0x2, 0x5, 0x0, [@private1={0xfc, 0x1, '\x00', 0x1}, @mcast2, @private2, @rand_addr=' \x01\x00\x00\x00\x00\x00\x00\x00\x00\x00\x00\x00\x00\x00\x01', @loopback, @local]}}}, @tclass={{0x14, 0x29, 0x43, 0x1f}}, @hoplimit={{0x14, 0x29, 0x34, 0x8}}, @dstopts={{0x30, 0x29, 0x37, {0x84, 0x2, '\x00', [@hao={0xc9, 0x10, @local}]}}}, @hopopts_2292={{0x38, 0x29, 0x36, {0x5e, 0x3, '\x00', [@hao={0xc9, 0x10, @mcast2}, @enc_lim={0x4, 0x1, 0x3a}, @ra, @pad1]}}}, @dontfrag={{0x14, 0x29, 0x3e, 0x3}}, @flowinfo={{0x14, 0x29, 0xb, 0x2}}], 0x168}}, {{0x0, 0x0, &(0x7f0000002900)=[{&(0x7f0000002700)="a0983fdced6c2d9679e1a20e6d806c5e440240cfde2971157d70bceb31650cc8fb18409a6d05e411cd7fccaf83667fe3e33c79c4a12ec30e7d5889fd9dd563ef8a7c02eede87fa4adb113b943912f060aa37d5ae8a4b9c548746df17e5d8a61cac4ba3a295f7db02cfa4dceebc8b5cfaac5b136a8450fcc7f2baa097d9e50f7d93615c94a024fc2f4635142cc1c187e5e5e32d0d2f7c6a498f88a15b0a4493cc394231a7964ac6ad785b8ef9e6eefb8ebebfd0797d112b860aba95400fa281390be8a3f6e6ca9d05ade4b96d6aaf5dbc6b07946077f6", 0xd6}, {&(0x7f0000002800)="2746e592a8faabfefe4d29eae51b6a1b2c91bac16245639b70a085f5738fe6242b0aa8676d22ba769fb956a9505a8c9b067038864abb73b72cdab8e2e983af375344754cae612fcc6a74d69d1d844b59c19090e9f19bd45082d6972c467a14888f48764f541b8f6b025318c1729c9af8a52034879a028af3b1feabeae49e4daf08f3fc1adc1141875979f56b22b7df20534b7b41686fd141727f5d94ffbafe5accb656cdf39fcd49d06d2afee19cef2e2784949fab82bb27c6689f649bedbd98950d1111b860332a438b27dd8daa5d99d3e90555492364639fdf5fe1f8c457907dcc99733fa0ad2123bb4036452ab884", 0xf0}], 0x2, &(0x7f0000002940)=[@flowinfo={{0x14, 0x29, 0xb, 0x3}}], 0x18}}], 0x5, 0x4000) socket$inet6(0xa, 0x6, 0x0) (async) bind$inet6(r0, &(0x7f0000000000)={0xa, 0x4e20}, 0x1c) (async) listen(r0, 0x6) (async) socket$inet6(0xa, 0x6, 0x0) (async) connect$inet6(r1, &(0x7f0000000200)={0xa, 0x0, 0x0, @initdev={0xfe, 0x88, '\x00', 0x0, 0x0}, 0x3}, 0x1c) (async) connect$inet6(r1, &(0x7f0000000040)={0xa, 0x4e20, 0x0, @dev={0xfe, 0x80, '\x00', 0xf}}, 0x1c) (async) sendmmsg$inet6(r1, &(0x7f0000005a00)=[{{0x0, 0x0, 0x0}}, {{0x0, 0x0, 0x0, 0x0, &(0x7f0000005bc0)=ANY=[@ANYBLOB="5000000000000000290000003600000000060000000000000401000730000000000a0000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000020000000000000002900000037000000000000000000000004010000000000001400000000000000290000000000000160000000000000002900000037"], 0x100}}], 0x2, 0x0) (async) socket$inet_udplite(0x2, 0x2, 0x88) (async) ioctl$sock_SIOCGIFINDEX(r2, 0x8933, &(0x7f0000000340)={'bridge_slave_0\x00'}) (async) socket(0x10, 0x80002, 0x0) (async) sendmsg$nl_route(r4, &(0x7f0000000000)={0x0, 0x0, &(0x7f0000000080)={&(0x7f00000000c0)=@bridge_setlink={0x2c, 0x13, 0xa29, 0x0, 0x0, {0x7, 0x0, 0x0, r3}, [@IFLA_AF_SPEC={0xc, 0x1a, 0x0, 0x1, [@AF_INET={0x8, 0x4, 0x0, 0x1, {0x4, 0x3}}]}]}, 0x2c}}, 0x0) (async) sendmmsg$inet6(r0, &(0x7f0000002980)=[{{&(0x7f0000000080)={0xa, 0x4e20, 0x10000, @private1={0xfc, 0x1, '\x00', 0x21}, 0x2}, 0x1c, &(0x7f0000000400)=[{&(0x7f00000000c0)="feda2ec1cb303eb00f06a45f8228b563bd35f07111fecd807ec3c6b41b15d1ba07cd7189e2a603df3244c4aa9fb49890e1a6e4d790987930dfb23e8ba45487c0", 0x40}, {&(0x7f0000000100)="2d982b346b4b981b3a8b74359e4c88906a80abac39b20e089d566998c5a951c34ce2218edb9cde7dc3073a0df468a13f928d991f79e437715520a712e80cc5972f68084292c960905a3d", 0x4a}, {&(0x7f0000000240)="dff888c7697d7dd474cd5535142cb3c13ea78c0b07109203a350fe2634ddf4063d22eb0b0058900ec5d7b79b078abbe7cb07d9c2d46c44523dc99c644cd9fec9759e3f58919d0450591745f7e8f5b79a3b86d7a78fe02a3890c1d8f8e7935dbb7865a312ef3928a6026b9ea4df2795953e944cce5cf669e567968ae7836fc511c7c9eac81a90ce71b1b11a57d70aef3fa26c30d573c144ec7481588b86977707fc1ed1e3d4a9bd8593902808b5bd577478651957aefd7b6bc4aa6ab607debdf8631eb24f5e6e1d402f2ad261023c2e3555648b6ccb458e59e4fdaab285f27d3f65f9413a9ed787d40baf13", 0xeb}, {&(0x7f0000000180)="c1c3042b092f696f088c00a19a0b38525ed579b77ba88adbf6eb505cd2ea537ff796feb5afe011cdfcd7dfcdb39feb3de8833b360f768598dcee213cec4c4a093dffb398f59d20e0bd", 0x49}, {&(0x7f0000000340)="5071c6b52b084037bca1a119b09664b64dfb834b13bed5b2e842f0547e4589e915df3c509c91741b95611bddfd6f84451adeccb5d4b15040a1fd828f6dcac794ca33c523b717e0cebddd0eb4c919be88db8b87757a5e42e423fd63abf7a16017ba8b0a527159d2df7c2e2c9e6ae02028cf3aaaed95c7553df648689cc7ab07c6cea5394ff6c0d0d44b35e1230f1354fab75df5af1ebcd292df7d809c1f2d47d3895489b79feb6b46d69c6a", 0xab}], 0x5, &(0x7f0000000480)=[@hoplimit_2292={{0x14, 0x29, 0x8, 0xba90}}, @tclass={{0x14, 0x29, 0x43, 0x9}}, @hopopts={{0x80, 0x29, 0x36, {0x2f, 0xc, '\x00', [@enc_lim={0x4, 0x1, 0xa6}, @generic={0x1f, 0x60, "c012574bc3e513d3be4641846b965d5e881e431f8f202f8bbd06afe5e86a0909840391d4fe82f298137479cd9efff48757fda6a1f22b82363f5aa2ecd0007019d99d38ced76625a9e3ab79f7f5108aa60c36f3b65e7a52bf6d28f213de1f2a0c"}]}}}, @rthdr_2292={{0x38, 0x29, 0x39, {0x2, 0x4, 0x0, 0x2, 0x0, [@mcast1, @remote]}}}, @tclass={{0x14, 0x29, 0x43, 0x7}}, @pktinfo={{0x24, 0x29, 0x32, {@ipv4={'\x00', '\xff\xff', @dev={0xac, 0x14, 0x14, 0x22}}}}}, @dontfrag={{0x14, 0x29, 0x3e, 0x3}}, @dstopts={{0x20, 0x29, 0x37, {0x21, 0x0, '\x00', [@jumbo={0xc2, 0x4, 0xfffffffd}]}}}], 0x160}}, {{&(0x7f0000000600)={0xa, 0x4e20, 0x8, @mcast2, 0x8}, 0x1c, &(0x7f0000000ac0)=[{&(0x7f0000000640)="0e2df84fe5829143c916a56badcdcee434551ca3a72b4b0976aadee23accbbc1e685bd5af90e4e0ed1de612b615f78f742589b1b495eaf85398e7435375d3f5ea9a3ddba57cd9e264a4a42d81070b2bc52088341b197434f6492a9f5341d1db87675fd134382f1ff766c57cd989a8dbd73a8aafbeb751a16eaac909af0fb86c3688b163820c920e87f28e144af3d481b01d1aa2195f23ecc5a4486dc38c277ae17f04980f000a0cf", 0xa8}, {&(0x7f0000000700)="df630a9ea335a9663e8b1fbc4508c0c67e0b68f4744e114977ac354e9c01dcaccfb658f9c06faacad96c8113c7fcdef1f0f687e9884bbadb5b32abfae96bbc2f567073784415a2e5c42c0a19ae853ace864bf31d5461c8750ba2ea0bc65b40bdd7ed6ca3b9f94ecef5d174f05c0b5de1d931d085e5c40212fab15cf564c8ff866969b0471a04caa9bbbea505ef6fe998ed35b3d0d2ca21d5ca48e8c98fa6625eafe025f306b3e7992622f494492d1018e7115f66728d2653b822d5a757a35e4906ebe33bd98b08557b93", 0xca}, {&(0x7f0000000800)="4649f5ab56503a7928b97e4766843a2733955e576147a80bd03f0e3919c2", 0x1e}, {&(0x7f0000000840)="4f83", 0x2}, {&(0x7f0000000880)="18ea84ebd444e97992260b9971c53cf9d5cff9dbd55c54af47e153686c283eec28e1c3c4c995c948266c2af02acf878f6b86484c943b0fc0aac76aa0dca0560e559f6161d64c0facdcb6b3cdde6209cf25215723210005155774e160cdbafd7da96bfa731c37849a320cc1635af8923c06f98d62ef127d50cbac86dead0c94ae29b58e359b704be41223fd8a8db396b1dc08047e93a3440463fb4624ecaecdf0ef371686075f5727a157a815f83b6ef7a17db274a8e1d8a0bfc4329f83157de8cd0134ac367fa9b4375d60d37cc6b5e0439ac7aac9fbe46dde1e02824084d3a6a5225f49634f89e1ef73fda874560bce487cd2c4d5c8a6", 0xf7}, {&(0x7f0000000980)="b6a7cb96b4a0a552afaef96506b7e8ca254428930550b36af3072c85dc2dc8f6b8b48ea5e25ab24b45f7ee19df45aebd4daa3505c2cba99c8a6ec0b40be35f1018bb8a1bcb0e52d4115f85b50530a92af04ce48bde9f2957710c836c3f4e7836fa3aaa1725817e3d11a0a41200a5e5534aac3f30", 0x74}, {&(0x7f0000000a00)="df1461b1a75640df3bcc88e2656187b0540bd6d99b47ccedcc8932ae63648d6e", 0x20}, {&(0x7f0000000a40)="0e07f15eab0e01ef7a843b7f3a28c3e0dac54e3eeeceb8767a5c21463404d9fb4a336970d62692769b432f8c4b0345287f4acce4661a05ef77718120a1c75002a108f4991043bb41acad973ac5089ddaa47b57d9d1a61e88", 0x58}], 0x8}}, {{&(0x7f0000000b40)={0xa, 0x4e24, 0x6d6, @empty, 0x5}, 0x1c, &(0x7f0000000fc0)=[{&(0x7f0000000b80)="6fa5b0d71557abbba8a7593948d41c599b0116db3e2c39f26bc3419146c04e5b84846c5501bb7a891e333d549a7382fa8163f50743568b56fc45d25509b52639460b92a67712ec62731a22141f989f77850aab7973e08383f68aae4246516ce5d17483d06c00e610fa517718e7fdf8c1478638bbb7d47d1dec7de4a3def67bf07cac0a2bf7586af4cc0df9a19f12e95cf294576d65d20572b92dd0fa36a51eb4cd155fdcdf51845461a1924ce5e0dfda", 0xb0}, {&(0x7f0000000c40)="2e863078dacbb12f78f632cece9c7d988aec6f2e7f3586d966343bfd9a9791ecf8b2fb0b1fabd530fcec72edec06bc0ecd48f4435f7342ca87307de880bad518ca0a36b99b2452d72ff4685b4995e7797455d105a2ac585486458967be7034d74f5b2ad49389aadb0570c77ed191f861ea5704e11a35068beb5b42a4bc39a287ee96868e202e419386b7e9fd63e9f5619e7b9a615a9e7917972bd15c3b43bac13a710dc78c07c5c9d10e859f56605030efa106e6d437607db66ecb2d9aaec82d385f12e340c246966f0701dc3d7005d4", 0xd0}, {&(0x7f0000000d40)="9649c68807526393025f4f8dc5c6b612237847a11f5826551719051c85a352f1b0a49c109fdc7ea2e46d2c04109a86f67606d29eb332f0f7e184c0f9fd7c6eb6f414faaac6902a9d890faed749206c52d0b5308a91b03fd083c987641a1fa7155f998e4436880bf4644c9815befd48cc30c39a9c06aa61f5cc266f898e8e456758ffd0292cd37a28d2cc1fbb79ef10980192eeb09e7c0690554c258a39c6c6ccacf53550fbdda778bb17e2d4b3a296059b1bdf3ea3658fd79a7d8ee0", 0xbc}, {&(0x7f0000000e00)="40baaae1d82d9c8bd9dd8c354852f73426a53dfbb116bf264a82cce1f392d6911dc86f3361fe12a52ade3a57772c6e1cdbf9529778f1c476bd26b11f1b1d2ff3d99b0274ff9a30e289fb5ea1650c183a7b12938e99de10078acc486ae2333cd8a27596936ab53ef1c6cd9d24b196f884d574a62f4a53e52eeb88b5e2afd57c9a6d0771574ab69b2fe8b8290bf9b4b451ebd4e191de4ea487a58e06bcb9034178d6479dc3292b31c3292c755edfd17ff2ac8819c65e3f6c2d2aa83261f7f19e4fcc2091642ce8eb8a40d072d940b714bebe4124b100a1220649f6", 0xda}, {&(0x7f0000000f00)="46665e3f183618c716d2b7032becdd1283ebc1d2ec57cb654de7509465c1d4f0a2fc97da640f5eaac5d4e377575552a6aacb4694dcdffcb8082205e6227374aebda7c83ac852964f385c488d08f200b843ed0166b1511a0f666ed923fc7033b9618473fdf62f5822634c061a7a6054a60ab3b582faca4307865119dfede554020388b39bd6b10ca6300ae264c60f032c1da54aa02e7c5b5d3bc3c3901d2eee54a138ff6aa9ac07ac01c1b15b85377e0fe7348e7daec9b7", 0xb7}], 0x5}}, {{&(0x7f0000001040)={0xa, 0x4e22, 0x7590, @initdev={0xfe, 0x88, '\x00', 0x1, 0x0}, 0x8d}, 0x1c, &(0x7f0000002500)=[{&(0x7f0000001080)="91e3822371a251242a663bbd0f1298d43771917894c6d3f330185ed8e26cc4906bd6bc25cbb3d23ccf7791cbc08f58a54693", 0x32}, {&(0x7f00000010c0)="2ff549c0aa7855a23acdb351f7312a01216b7a9b01ed58dbb4e0fdd792eeeac9c89dffab655426c9af9ec7dee035f0339f93ac08298c824ad7b2422eee9ecd40caccb60234cc3eec9ed92e7c397e0ab9ad5ce6a115ed0de7953bde44e75bf8890f18d66dfc88f323a198084015b5e4f4702cdba271a4112af76c222da6fa38ae58d9d163236556546430a7039503afc3121062de5a6fe1eea2cb74a51ac3bd13f9bc4124f970787fd80da4529b3edb0d2a7f22c9511e9c47976bc88e3bf81620837bc43449aa383c04d270c0c7be575bc932e2f9f4217218547867e327dc07f9e649806f7d4b69a898c87a830cee2ce59c372a3de7b38ec09b95cf89208a0e119f94641b1d1c19714e95145e4ad45088ffccb872ecdb3b7ecfb02c92227723d7893b32d9da6e1b8383e741f4c9313eb6ea59012b9450df669922780795f4f4ae75c345aca94c92cc4047fcfa7454fd0c5901e47e18068442188bb640002440dca504284ceffc3da14c5710cfe6e4f3d2ee1e778032d06b6b7bf6eaafebe0fc78d04f2994aa93a9f84d6ae1d0ea4ec37277ab6f219b7fade45c57305f4682041574642a7244c94905ea12ff2e079e36ba0a28156488467792f938624a5c4ff45d07f4c32a50a3387c347cbfcb92c3abb2f735886a65538c9c0214d59dde6c08f88a468e3f35b7e503358266bc845fdc71a392502c713ffcc52dcf46896b407da99465ee18f7a54f9ae7c4c4af66c5fbf6463afffd4ccf54a3762d2996dd0328ecd88e0b9bf43e6216c101410f1546e9c628ae659814987e2d1b571eb38641c5048add8a1a8a6c334b2b2a8e4ec4bbbfa781b192bd3d740c2224e58e8fd2270d2c01a2c18ff66ea9bf0fcbdcae712a2562471f9f293947d7f282492ea3bed0adbb70edcadcd4589a1a0943d4cbe5dc149b0c7c5752202ddd81ab9099cf6eb88f8d3e2dada4eed96c4cbd8ad369de503bb1725b063f5f51b0de40c957e789002845d3e09512ec7b6a182b0279770041436489e39734ce61ce2fe956aa66c5ca71048cc898abcfbca7f0cca9186aa08d2f1b08ed24e8af16a688cfaa28807f4173d52ee4b74a9de9bfd5d3d122e1c8464ea729b62f403793ece706af52ffa77de67e710b6e48a7850b5875e7eef055355ef21eba88ffbc6756b78e29e569ec1c40cf10a81ee3e47e4016f595db2aa1b227623bb51a71cd182dcfe1fbb4aed9d2cd5e3be5cdb0d956e65bb83944d86422b00df9ba15ae6a32ed786f3b5eff269502ae9d79f87a9855cb5bd9c33ef7d4fc87971572185e2db3bf343d814b6ac0d98a895f5128009fb3f91ec8a78a0e236cc5e98e23acc5ab721b85d774cbd48d1e2ce1b69aa7f53b9cf3fc3550993ecb7dd316644a60c7e357e6e50f15b16458fcb913b3b27741c7785a1c9aca2795b5883ce9affbea342b7da00447fc095feda5a3a480325f68b2f81c909acfd82f1fb6fc7dfd44483f90157cb80d361c7f30bca3569cece17fe36f0f2063c57ae4c25230ddb4f975d39b34220e6d4b84cc5d5e2de938acfabdb7a16b8b39f4e569c9c88ee36261d58f23e1446ee63c1f476b1f5cbbc72ace11161db31493a6aa1b74d4c914993be31a7f2ef95d89e1d021db8d0e889c055509eccd48ae7ab14b73450f5c2a5d03110d131683e7b2f777c4399a0012759f8ca2ca60ffeb5ecbd00f2680800c3bf20e6675acd51e0456375d353098c78c50b2a33354322b50b4233dffb281740155c4e13d138e754664f3570f793843042e2734207d72ef29e4e197406c98c6900b05986fa5f905ac3399964d44157b165ca0d80e98c95e8f1b66e7af285715ff8671beba62d65c5af2b6622f47fccf5d925634bafa2da7bb87ae1cc074212d3a3420b0e1b19ff8770e2b40af3457bca64da544904324efb885b625598d7935d65c95a9d45d293a36627565ab3c9ebdc819ab7378e2b61840081de88eb9c68572f66a05a50304b9c3e6a552985b533a1f07838b9f8b876f4477d225563a7f2d43240412714385d7d85f1a379086df6cd37be6f68c532c797c8e22197660fc3cc78f1059bbf730e1e6137aaae2f3dfbcf208a2e8bc5c088d57482e9c80e53e9281a1fe1dd520cda167aeb349d201611eedf9d05c36086e6839845cc75497104db51c8c54e8dfb1af71744ea170cbf5f717b90ba44b68c3c1ad3741bc02122ea49fc3224b4a2a1da1d35af0f306dc0a0db0a67bad5bad5f0cadf43082df025e601abe593a8ca0e03dbd6deb06187a88d5e0b276e8293bcd39c29ea3e0c8ec1e06b89081d0c130476cecc1e4a3cafc31fa6735c547a1f0eec4090e91941ca7ba9b0211593b88f59bc44e5c874ea53110637254754e75b424677e3ecb68a5c939073494db1fbfc19af4bac84330838323b8d93ee2bbb4dccfbecbbec9e3001f6941cda90b09d7d0c213b59cd68d38d21cc36795a127776e0cc387912eef7eab0778f7d5222216b6b9916f2869092a8f6eda97d7abebeb1ecf4853b222f9582e2cbf22ff07e373b9ef2160b048956154a985ed41bf57b5e8cb107a42847421f0d731ea261f81512086c717e59576ed9bd9baf4526656a07c556e55f698efa687510b627487b9c0e923a570a14b3ffb1745628f41f37ab7fe9009219e042c73b04018e0e64ae58d22417a6d5c35f025de673463586890698a8ed78852a894cc21113080d8f37fa9784d48b6c070c1dc1877182e59461335b69c564bfc7848c2c65f4a989304570015ad9b75aa7e7a9dfc85df55dad86b62e4b0527a70ee9cbdfdb8fa146ee8b719dd160521412375209edc325b781c84ec8de9021bd2846447ff588ebad29b462b7b27d0aa2ce4c64ad2e6267f7a60c2ae9d25fd08414dbf44aabcb1fe9be91c2e70f5a8a76eeab27f379bff3a08f593379f326cb082cabff65e917015fe9b7a4fe5c9b223f4ec676800acc9398c0edcdd59aede104681155c2b2d5b5bf3703b766d1ebd1d08406af007a63e0533d1bec12f525c2acd699fd60c65fb7f1c96db5a69ad027d704074f869a858d1ec8b964d50e4eeb2ae193703538207f786a4b3ba862b62327845ae71d62ec1d7fc864e2c0984fb3de7424fc43a7991d68893d2b415f359c4360391c24a9b3ffc0e7b2d92ce4e703a198b4766bb033d70d6ab9a9caa509c135e13b898eff71f2fd9b589a807a22013b2d53f1cdb3ade43b5b3a5798751375fa32129955763cbe2302010fb4fe3a60e7595899083da1bba5b9a1f844891dfc64b71d143686295f0e1579df9fd2fc1667151ead1e97968a84612cc64fab2e808860f02921321f806b72612e95ca9504451e36b126f78a49f5cae249297ea1d067950aca4ca05c43207de5b5946b672703f0198991cc6c83e0c174478b151b20d7263cbf47f92533c3e744cc42579274502e46264d09a927ef16ddcb7d418fcd3472506d2dd0e77595ddc443010f58158ca68eb35c2e191365eb5473850aef59c88263e670a9d09d3c9024ad0451f15f8c863774203637c2546f30950d348bccc6f09b125cef06696d96714252b82d9efbcf3e0daaa1f78f823595f02f55059e08474b25894f3683789c8e95b15de9c34d8ece4f0f5a1fb5ee7872025a9ea59518def338515632ff593c7fc04349b0c50b4e82fcffb627703d898707b1002500289e67289ae1bd5e0c5cafe5a4dfdceba8dea5602eba1dccaa1fd518d112c537d28182e3ef995f81f001e5dc2b894f25f54399456607c04601c230ccc4eee5de05af280d0341c01d133a04ee4ab1d0134d732592986f0a851d7b9c4318dd79c1c32694d62119ad41f8dce4b3565b22942555c8232ff39696ac99953b24fcd1289993558f3dc01e7c5071d77768aacd8c20628fd4acd051c256aa55cb47d489d85383c5be1eb5b3568b0dd415934c2fa6ae5a60d05a8d6e8ec3a8c219b518db452346a65b6086cf7747e72856886f18d4023d339edd126caf0cb5b7dff4f6fa2d3386aaf2f8c349d09da1519ef74192472ddd456ee8953d406a12d63a755d1ae6a646d811ec61b9151b96240781d301603d8c9034ec38d5f8a946a4486ab855c72dcb780ae65b6840cebb1f9a071b73ac813b1a8b218145054b920a2a553975b3ed1267cfcb9f66de5b7b91c0afc1054ce3245c4694a9dc448e40005499793268744dac46c4b7ef65f77eaa220dbde1ced009d7cbcdbb53930ceb260e21f83b9a39201014d38c46b714825f29968c13d528a32cf75baa43ff4dc4c9528c83dba4b25d11d7ddd20425b99b669812da58d9f6a9aa12c7b7eda27d58c94157230954190ec79ea131f3046c97ebc3290d2e67eba20a027989aa2552ac6abd6177c8e1667613b438adafa94f2162cc21e64676b37afacb292085ae4fde365ed3730e6947c8ef925dfefeaad6db5528e292ec16b5e44472b3e3f13b5aa186436c877119a788a2197422cc5e9537dc55a6548ba355ec1a7754f60156c930bd7a38aca09c7b84637bd16470ae41bb624d7405fe80f2600f5785e30fa457e28342c0d4d12deb5a4cf73851f7d8913e35a5346654db4729adcb00e2748985dc878b34c2e5c44062fdb7552d4d4394f2b74187712e20edc275862ca3d24879455f24188c5a198440b020391bb7673952b304046c971917e7ffba36d857219145de8f759cac2bf08b7fdb3c6509d68f521c2f9c89b73eb9e121f79237d89fadf1c310bffa35b7fb312147d29805faa73a13d3682f03e943e64ab0949a70064394ecaa90abbe9a0a89390980ec8700bc23e3d9aaf329bcc3a8994a2fbc7dc24398f51a41cdd80f5d6c3fb4f9735b392273081eb2856b8356eb178f2be0c1c9045d17c15c9071cf29f0df034aaed6aef81adfb4bbdaf3ce69c0e35c111ea4eab117f5d79a9efca4c7ca71ebb93f081b3a7bfa0525d2350eaa7b8f39200e4aa11f3576906d8a2d82a6bf24841669333334a5768e96b87f76ff6aee1f48aa3288d59f1d0c8b0d250b434592311583723dc41b160ee51dc94d3de09754da156045a749a74c8d36d51b56b43ec462264e4ea04b67dc3d576b7c25d8474d96800806b1a5603e17d823f06ec9f60583d8feed21fd2b599cd3df5e6b18288ce2269743bf708ca05733c3b64665ce1882a8b029a1aa3d0f74e985b533b49bf0c16e9ef46975654afcb45742cb9b474a7c22b0da6a3f6c1e276035c3a1677eb93ea3a8bde58d3069867b66c1f96a9f402cdb8a10ea671939b79317253a24cbf833c04d267d2e79b327ca343b975963414b1fb97bc61640e86a76a5a2990c4d35dfba7a62bcc0cf669ca6fbec975c84f35bec5cefe8fb9b1ce4f92d836eb136898a17bffc73f21faf3b3754fafb52e6cf83c958744437dc9112899e277b429d28241527ca10476de6d3c64ba4978ad7cfa6d1ee0173dceedba2d2122c35047f3fb6d6afd42b27c6acaf131bbbef239fac6ae3d39cf978f6f7af360901c2d44012a956117832812143e5f6ed63af1b4861a8a98dd0966284811cee5cebae461ff4f0e9ca416f7039a2ac21058630b39fd855cb351e46759501d4b3e31123ed82d98254c25edc2d402c185d1820f4d2af4151449e5fe8096098988a64bab8442b33ea9cd1cdb5d701a9e5763d21be632e4ce4bac1094a60f824b65176f7e420a382fdc8d20d14bb2082a64286031ec8c136a92af39cd5c4df71ebf82a42c7aedbb0e42db5613cf0cce7bd66b857744dc4712850c6a81743b6db7a3f05cebbeebef7de28a4099c3a1f77200f3e55e1fd3e56573c3062461a5c93c918c4f62026b407cdf1730d36ea7d5c65d064b8db59989598ea19e", 0x1000}, {&(0x7f00000020c0)="66d651e8a1b16cd32e856ad6cb93c167571f4c4414c47551565a61", 0x1b}, {&(0x7f0000002100)="348801661bf341b443670227708ef351a3daf01b7f1f9a3afdd0a821b4f287a9cd75c670a07ad3b6e87f136685dd1f6c3a6a3bcd9f8a441eb29f306a28131e45bf786dcfff7bb83492d6f9167e8884dc064b9b7cc6c40bb207", 0x59}, {&(0x7f0000002180)="c9ac68377e2e686937b2b70e64a53ffab73bf2a8eafabf4d91d9432832b7b2fa5bea34c3c70c947f2dc16377a8d820fda51d8d92a162e55fd52ca23bd66ff088c6e3eafc8462a748ad49be27c3045f841758733d690f6dea5be7a9d181f2ffa0797a46a2fed1289edd5b1c7622f230e00c19efd330570f93ab4f56c42a52a0ff42ad15bf8d20762459353c37a11d312bd58e68de69930f1a4ba09857e0d9cc6a1af11b0f118b7980b82fbcc9ed48d497addd5a42cf13f53dcad63dba62bf616f4650a5302091951d3ca1e3eb82f0cb168cb13439263740b7a25b38c5ef", 0xdd}, {&(0x7f0000002280)="a3bfadb6fbaae52f2e2fa67262b4c9c6cb3f964657d4b6cc3dd5b5aae919f083af5b91eb044f5c6961947eef1407b5d97a6ec25f3a19592a7a8f16c247622b9cad091f5a023b170a8cc2ec8201604fd74681d7f698c9cd1a51b0d1b1a19ced34478b1c0e28aaeb6e0520d7267d89ad9a56e3be486d9f01feac8064a55aa7d2ab3127ec194ceb5bd202ca65ecdcba22dc2e994b925406fc8dc73d8c0902365dc38f46e74b212776cf16507cd0b2e560e2da98c1e4cc5c9c4d8fdead8113e930b737", 0xc1}, {&(0x7f0000002380)="19d74ab20ba78c5642d056aef55b8bdeb008fd7a739e8ddb9dd2ad475bb401bce8994f87ca19384025a23b4186762414d8f558f224882f98840b495ab72bf9ce558c0d91a0430da4991d68b0976e536e1e53ca327442f31cc3b90e2a7f0999cbcf593ad1c07f8af811eb406012a2ac35c6f46adc89fccf7bc0bc296acc011790d5018afaa44e45ff415ac8b179cd755ac43d09b795d1e09e2232ca77ea65c6976e0599cc9509e563879bb36bbccb682259c427b13210d2df8c48677646be1b3d76d60839b1638ec94b99a28ae624cfd7140b39ecac45f398a38eb47e", 0xdc}, {&(0x7f0000002480)="139ee9852a4e2ab434e2223402e01dd988ae06864bfbb88b5240a26e26c6c495454773a54a974e057085726d470fab7c78ec87c1a599b4790d494271d0293cfd89b0d8a7dbd59014b0cdcdb3ba614a873214f72cfbd620de6292c6a1", 0x5c}], 0x8, &(0x7f0000002580)=[@pktinfo={{0x24, 0x29, 0x32, {@loopback, r3}}}, @rthdr={{0x78, 0x29, 0x39, {0x0, 0xc, 0x2, 0x5, 0x0, [@private1={0xfc, 0x1, '\x00', 0x1}, @mcast2, @private2, @rand_addr=' \x01\x00\x00\x00\x00\x00\x00\x00\x00\x00\x00\x00\x00\x00\x01', @loopback, @local]}}}, @tclass={{0x14, 0x29, 0x43, 0x1f}}, @hoplimit={{0x14, 0x29, 0x34, 0x8}}, @dstopts={{0x30, 0x29, 0x37, {0x84, 0x2, '\x00', [@hao={0xc9, 0x10, @local}]}}}, @hopopts_2292={{0x38, 0x29, 0x36, {0x5e, 0x3, '\x00', [@hao={0xc9, 0x10, @mcast2}, @enc_lim={0x4, 0x1, 0x3a}, @ra, @pad1]}}}, @dontfrag={{0x14, 0x29, 0x3e, 0x3}}, @flowinfo={{0x14, 0x29, 0xb, 0x2}}], 0x168}}, {{0x0, 0x0, &(0x7f0000002900)=[{&(0x7f0000002700)="a0983fdced6c2d9679e1a20e6d806c5e440240cfde2971157d70bceb31650cc8fb18409a6d05e411cd7fccaf83667fe3e33c79c4a12ec30e7d5889fd9dd563ef8a7c02eede87fa4adb113b943912f060aa37d5ae8a4b9c548746df17e5d8a61cac4ba3a295f7db02cfa4dceebc8b5cfaac5b136a8450fcc7f2baa097d9e50f7d93615c94a024fc2f4635142cc1c187e5e5e32d0d2f7c6a498f88a15b0a4493cc394231a7964ac6ad785b8ef9e6eefb8ebebfd0797d112b860aba95400fa281390be8a3f6e6ca9d05ade4b96d6aaf5dbc6b07946077f6", 0xd6}, {&(0x7f0000002800)="2746e592a8faabfefe4d29eae51b6a1b2c91bac16245639b70a085f5738fe6242b0aa8676d22ba769fb956a9505a8c9b067038864abb73b72cdab8e2e983af375344754cae612fcc6a74d69d1d844b59c19090e9f19bd45082d6972c467a14888f48764f541b8f6b025318c1729c9af8a52034879a028af3b1feabeae49e4daf08f3fc1adc1141875979f56b22b7df20534b7b41686fd141727f5d94ffbafe5accb656cdf39fcd49d06d2afee19cef2e2784949fab82bb27c6689f649bedbd98950d1111b860332a438b27dd8daa5d99d3e90555492364639fdf5fe1f8c457907dcc99733fa0ad2123bb4036452ab884", 0xf0}], 0x2, &(0x7f0000002940)=[@flowinfo={{0x14, 0x29, 0xb, 0x3}}], 0x18}}], 0x5, 0x4000) (async) [ 2806.770080][ T2468] workqueue: Failed to create a rescuer kthread for wq "bond1274": -EINTR [ 2807.103731][ T2479] netlink: 'syz-executor.2': attribute type 1 has an invalid length. [ 2807.200668][ T2479] bond1264: entered promiscuous mode [ 2807.209901][ T2479] 8021q: adding VLAN 0 to HW filter on device bond1264 04:12:14 executing program 2: r0 = socket$netlink(0x10, 0x3, 0x0) r1 = socket$netlink(0x10, 0x3, 0x0) r2 = socket(0x10, 0x803, 0x0) sendmsg$NL80211_CMD_CRIT_PROTOCOL_START(r2, &(0x7f0000000580)={0x0, 0x0, &(0x7f0000000540)={0x0, 0x1c}}, 0x0) getsockname$packet(r2, &(0x7f0000000600)={0x11, 0x0, 0x0, 0x1, 0x0, 0x6, @broadcast}, &(0x7f0000000100)=0x14) sendmsg$nl_route(r1, &(0x7f0000000040)={0x0, 0x0, &(0x7f0000000000)={&(0x7f00000000c0)=ANY=[@ANYBLOB="3c00000010008506eb9afc4cd8d06e754a0081c5", @ANYRES32=r3, @ANYBLOB="2377f292252155b21c0012000c000100626f6e64000000000c0002000800010001"], 0x3c}}, 0x0) sendmsg$nl_route(r0, &(0x7f0000000240)={0x0, 0x0, &(0x7f0000000680)={&(0x7f0000000400)=@newlink={0x3c, 0x10, 0xffffff1f, 0x0, 0x253c22, {}, [@IFLA_LINKINFO={0x14, 0x12, 0x0, 0x1, @bridge={{0xb}, {0x4}}}, @IFLA_MASTER={0x8, 0xa, r3}]}, 0x3c}}, 0x0) [ 2807.283279][ T2485] bond1264: (slave bridge1185): making interface the new active one [ 2807.291628][ T2485] bridge1185: entered promiscuous mode [ 2807.301344][ T2485] bond1264: (slave bridge1185): Enslaving as an active interface with an up link 04:12:14 executing program 4: r0 = socket$inet6(0xa, 0x2, 0x0) (async) r1 = socket$inet6(0xa, 0x80002, 0x88) bind$inet6(r1, &(0x7f0000000580)={0xa, 0x10010000004e20}, 0x1c) setsockopt$inet6_udp_int(r1, 0x11, 0xb, &(0x7f0000000080), 0x4) connect$inet6(r1, &(0x7f0000000040)={0xa, 0x4e20, 0x6, @empty, 0x5}, 0x1c) (async) ioctl$sock_SIOCETHTOOL(r0, 0x89f0, &(0x7f0000000000)={'bridge0\x00', &(0x7f0000000200)=@ethtool_ringparam={0x10}}) (async) r2 = socket$inet6(0xa, 0x80002, 0x88) bind$inet6(r2, &(0x7f0000000580)={0xa, 0x10010000004e20}, 0x1c) setsockopt$inet6_udp_int(r2, 0x11, 0xb, &(0x7f0000000080), 0x4) (async) poll(&(0x7f00000000c0)=[{r0, 0x4000}, {r2, 0x11100}], 0x2, 0x5) [ 2807.404615][ T2497] netlink: 'syz-executor.1': attribute type 1 has an invalid length. [ 2807.515763][ T2497] bond1283: entered promiscuous mode [ 2807.522124][ T2497] 8021q: adding VLAN 0 to HW filter on device bond1283 [ 2807.557646][ T2498] bond1283: (slave bridge1229): making interface the new active one [ 2807.567713][ T2498] bridge1229: entered promiscuous mode [ 2807.584409][ T2498] bond1283: (slave bridge1229): Enslaving as an active interface with an up link 04:12:15 executing program 1: r0 = socket$netlink(0x10, 0x3, 0x0) r1 = socket$netlink(0x10, 0x3, 0x0) r2 = socket(0x10, 0x803, 0x0) sendmsg$NL80211_CMD_CRIT_PROTOCOL_START(r2, &(0x7f0000000580)={0x0, 0x0, &(0x7f0000000540)={0x0, 0x1c}}, 0x0) getsockname$packet(r2, &(0x7f0000000600)={0x11, 0x0, 0x0, 0x1, 0x0, 0x6, @broadcast}, &(0x7f0000000100)=0x14) sendmsg$nl_route(r1, &(0x7f0000000040)={0x0, 0x0, &(0x7f0000000000)={&(0x7f00000000c0)=ANY=[@ANYBLOB="3c00000010008506eb9afc4cd8d06e754a0081c5", @ANYRES32=r3, @ANYBLOB="2377f292252155b21c0012000c000100626f6e64000000000c0002000800010001"], 0x3c}}, 0x0) sendmsg$nl_route(r0, &(0x7f0000000240)={0x0, 0x0, &(0x7f0000000680)={&(0x7f0000000400)=@newlink={0x3c, 0x10, 0xffffff1f, 0x0, 0x8672500, {}, [@IFLA_LINKINFO={0x14, 0x12, 0x0, 0x1, @bridge={{0xb}, {0x4}}}, @IFLA_MASTER={0x8, 0xa, r3}]}, 0x3c}}, 0x0) [ 2807.735623][ T2504] netlink: 'syz-executor.5': attribute type 1 has an invalid length. [ 2807.808551][ T2504] bond1274: entered promiscuous mode [ 2807.814540][ T2504] 8021q: adding VLAN 0 to HW filter on device bond1274 [ 2807.922472][ T2505] bond1274: (slave bridge1237): making interface the new active one [ 2807.930599][ T2505] bridge1237: entered promiscuous mode [ 2807.940504][ T2505] bond1274: (slave bridge1237): Enslaving as an active interface with an up link 04:12:15 executing program 5: r0 = socket$netlink(0x10, 0x3, 0x0) r1 = socket$netlink(0x10, 0x3, 0x0) r2 = socket(0x10, 0x803, 0x0) sendmsg$NL80211_CMD_CRIT_PROTOCOL_START(r2, &(0x7f0000000580)={0x0, 0x0, &(0x7f0000000540)={0x0, 0x1c}}, 0x0) getsockname$packet(r2, &(0x7f0000000600)={0x11, 0x0, 0x0, 0x1, 0x0, 0x6, @broadcast}, &(0x7f0000000100)=0x14) sendmsg$nl_route(r1, &(0x7f0000000040)={0x0, 0x0, &(0x7f0000000000)={&(0x7f00000000c0)=ANY=[@ANYBLOB="3c00000010008506eb9afc4cd8d06e754a0081c5", @ANYRES32=r3, @ANYBLOB="2377f292252155b21c0012000c000100626f6e64000000000c0002000800010001"], 0x3c}}, 0x0) sendmsg$nl_route(r0, &(0x7f0000000240)={0x0, 0x0, &(0x7f0000000680)={&(0x7f0000000400)=@newlink={0x3c, 0x10, 0xffffff1f, 0x0, 0x2535f4, {}, [@IFLA_LINKINFO={0x14, 0x12, 0x0, 0x1, @bridge={{0xb}, {0x4}}}, @IFLA_MASTER={0x8, 0xa, r3}]}, 0x3c}}, 0x0) 04:12:15 executing program 3: r0 = socket$netlink(0x10, 0x3, 0x0) sendmsg$netlink(r0, &(0x7f00000047c0)={0x0, 0x0, &(0x7f0000000000)=[{&(0x7f0000000100)=ANY=[@ANYBLOB="18000000560001060000006700db730007"], 0x18}], 0x1}, 0x0) [ 2808.069472][ T2514] netlink: 'syz-executor.2': attribute type 1 has an invalid length. 04:12:15 executing program 0: r0 = socket$inet6(0xa, 0x6, 0x0) bind$inet6(r0, &(0x7f0000000000)={0xa, 0x4e20}, 0x1c) listen(r0, 0x6) r1 = socket$inet6(0xa, 0x6, 0x0) connect$inet6(r1, &(0x7f0000000200)={0xa, 0x0, 0x0, @initdev={0xfe, 0x88, '\x00', 0x0, 0x0}, 0x3}, 0x1c) connect$inet6(r1, &(0x7f0000000040)={0xa, 0x4e20, 0x0, @dev={0xfe, 0x80, '\x00', 0xf}}, 0x1c) sendmmsg$inet6(r1, &(0x7f0000005a00)=[{{0x0, 0x0, 0x0}}, {{0x0, 0x0, 0x0, 0x0, &(0x7f0000005bc0)=ANY=[@ANYBLOB="5000000000000000290000003600000000060000000000000401000730000000000a0000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000020000000000000002900000037000000000000000000000004010000000000001400000000000000290000000000000160000000000000002900000037"], 0x100}}], 0x2, 0x0) r2 = socket$inet6(0xa, 0x80002, 0x88) bind$inet6(r2, &(0x7f0000000580)={0xa, 0x10010000004e20}, 0x1c) setsockopt$inet6_udp_int(r2, 0x11, 0xb, &(0x7f0000000080), 0x4) bpf$BPF_GET_MAP_INFO(0xf, &(0x7f0000000100)={0x1, 0x58, &(0x7f0000000080)={0x0, 0x0, 0x0, 0x0, 0x0, 0x0, ""/16, 0x0}}, 0x10) ioctl$sock_inet6_SIOCADDRT(r2, 0x890b, &(0x7f0000000140)={@private1={0xfc, 0x1, '\x00', 0x1}, @initdev={0xfe, 0x88, '\x00', 0x0, 0x0}, @ipv4={'\x00', '\xff\xff', @initdev={0xac, 0x1e, 0x1, 0x0}}, 0xf8, 0x7ff, 0x7, 0x400, 0x4, 0x200000, r3}) [ 2808.250949][ T2514] bond1265: entered promiscuous mode [ 2808.261330][ T2514] 8021q: adding VLAN 0 to HW filter on device bond1265 [ 2808.355627][ T2516] bond1265: (slave bridge1186): making interface the new active one [ 2808.368262][ T2516] bridge1186: entered promiscuous mode [ 2808.389692][ T2516] bond1265: (slave bridge1186): Enslaving as an active interface with an up link 04:12:15 executing program 4: r0 = socket$inet6(0xa, 0x2, 0x0) r1 = socket$inet6(0xa, 0x80002, 0x88) bind$inet6(r1, &(0x7f0000000580)={0xa, 0x10010000004e20}, 0x1c) setsockopt$inet6_udp_int(r1, 0x11, 0xb, &(0x7f0000000080), 0x4) connect$inet6(r1, &(0x7f0000000040)={0xa, 0x4e20, 0x6, @empty, 0x5}, 0x1c) (async) ioctl$sock_SIOCETHTOOL(r0, 0x89f0, &(0x7f0000000000)={'bridge0\x00', &(0x7f0000000200)=@ethtool_ringparam={0x10}}) (async, rerun: 64) r2 = socket$inet6(0xa, 0x80002, 0x88) (rerun: 64) bind$inet6(r2, &(0x7f0000000580)={0xa, 0x10010000004e20}, 0x1c) (async) setsockopt$inet6_udp_int(r2, 0x11, 0xb, &(0x7f0000000080), 0x4) (async) poll(&(0x7f00000000c0)=[{r0, 0x4000}, {r2, 0x11100}], 0x2, 0x5) 04:12:15 executing program 2: r0 = socket$netlink(0x10, 0x3, 0x0) r1 = socket$netlink(0x10, 0x3, 0x0) r2 = socket(0x10, 0x803, 0x0) sendmsg$NL80211_CMD_CRIT_PROTOCOL_START(r2, &(0x7f0000000580)={0x0, 0x0, &(0x7f0000000540)={0x0, 0x1c}}, 0x0) getsockname$packet(r2, &(0x7f0000000600)={0x11, 0x0, 0x0, 0x1, 0x0, 0x6, @broadcast}, &(0x7f0000000100)=0x14) sendmsg$nl_route(r1, &(0x7f0000000040)={0x0, 0x0, &(0x7f0000000000)={&(0x7f00000000c0)=ANY=[@ANYBLOB="3c00000010008506eb9afc4cd8d06e754a0081c5", @ANYRES32=r3, @ANYBLOB="2377f292252155b21c0012000c000100626f6e64000000000c0002000800010001"], 0x3c}}, 0x0) sendmsg$nl_route(r0, &(0x7f0000000240)={0x0, 0x0, &(0x7f0000000680)={&(0x7f0000000400)=@newlink={0x3c, 0x10, 0xffffff1f, 0x0, 0x253c23, {}, [@IFLA_LINKINFO={0x14, 0x12, 0x0, 0x1, @bridge={{0xb}, {0x4}}}, @IFLA_MASTER={0x8, 0xa, r3}]}, 0x3c}}, 0x0) [ 2808.439078][ T2523] netlink: 'syz-executor.1': attribute type 1 has an invalid length. [ 2808.535908][ T2523] bond1284: entered promiscuous mode [ 2808.541818][ T2523] 8021q: adding VLAN 0 to HW filter on device bond1284 [ 2808.698215][ T2524] bond1284: (slave bridge1230): making interface the new active one [ 2808.722355][ T2524] bridge1230: entered promiscuous mode [ 2808.733241][ T2524] bond1284: (slave bridge1230): Enslaving as an active interface with an up link [ 2808.761381][ T2529] netlink: 'syz-executor.5': attribute type 1 has an invalid length. 04:12:16 executing program 1: r0 = socket$netlink(0x10, 0x3, 0x0) r1 = socket$netlink(0x10, 0x3, 0x0) r2 = socket(0x10, 0x803, 0x0) sendmsg$NL80211_CMD_CRIT_PROTOCOL_START(r2, &(0x7f0000000580)={0x0, 0x0, &(0x7f0000000540)={0x0, 0x1c}}, 0x0) getsockname$packet(r2, &(0x7f0000000600)={0x11, 0x0, 0x0, 0x1, 0x0, 0x6, @broadcast}, &(0x7f0000000100)=0x14) sendmsg$nl_route(r1, &(0x7f0000000040)={0x0, 0x0, &(0x7f0000000000)={&(0x7f00000000c0)=ANY=[@ANYBLOB="3c00000010008506eb9afc4cd8d06e754a0081c5", @ANYRES32=r3, @ANYBLOB="2377f292252155b21c0012000c000100626f6e64000000000c0002000800010001"], 0x3c}}, 0x0) sendmsg$nl_route(r0, &(0x7f0000000240)={0x0, 0x0, &(0x7f0000000680)={&(0x7f0000000400)=@newlink={0x3c, 0x10, 0xffffff1f, 0x0, 0x9000000, {}, [@IFLA_LINKINFO={0x14, 0x12, 0x0, 0x1, @bridge={{0xb}, {0x4}}}, @IFLA_MASTER={0x8, 0xa, r3}]}, 0x3c}}, 0x0) 04:12:16 executing program 3: r0 = socket$netlink(0x10, 0x3, 0x0) sendmsg$netlink(r0, &(0x7f00000047c0)={0x0, 0x0, &(0x7f0000000000)=[{&(0x7f0000000100)=ANY=[@ANYBLOB="18000000560001060000006800db730007"], 0x18}], 0x1}, 0x0) 04:12:16 executing program 0: r0 = socket$inet6(0xa, 0x6, 0x0) bind$inet6(r0, &(0x7f0000000000)={0xa, 0x4e20}, 0x1c) (async) bind$inet6(r0, &(0x7f0000000000)={0xa, 0x4e20}, 0x1c) listen(r0, 0x6) r1 = socket$inet6(0xa, 0x6, 0x0) connect$inet6(r1, &(0x7f0000000200)={0xa, 0x0, 0x0, @initdev={0xfe, 0x88, '\x00', 0x0, 0x0}, 0x3}, 0x1c) (async) connect$inet6(r1, &(0x7f0000000200)={0xa, 0x0, 0x0, @initdev={0xfe, 0x88, '\x00', 0x0, 0x0}, 0x3}, 0x1c) connect$inet6(r1, &(0x7f0000000040)={0xa, 0x4e20, 0x0, @dev={0xfe, 0x80, '\x00', 0xf}}, 0x1c) (async) connect$inet6(r1, &(0x7f0000000040)={0xa, 0x4e20, 0x0, @dev={0xfe, 0x80, '\x00', 0xf}}, 0x1c) sendmmsg$inet6(r1, &(0x7f0000005a00)=[{{0x0, 0x0, 0x0}}, {{0x0, 0x0, 0x0, 0x0, &(0x7f0000005bc0)=ANY=[@ANYBLOB="5000000000000000290000003600000000060000000000000401000730000000000a0000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000020000000000000002900000037000000000000000000000004010000000000001400000000000000290000000000000160000000000000002900000037"], 0x100}}], 0x2, 0x0) socket$inet6(0xa, 0x80002, 0x88) (async) r2 = socket$inet6(0xa, 0x80002, 0x88) bind$inet6(r2, &(0x7f0000000580)={0xa, 0x10010000004e20}, 0x1c) setsockopt$inet6_udp_int(r2, 0x11, 0xb, &(0x7f0000000080), 0x4) bpf$BPF_GET_MAP_INFO(0xf, &(0x7f0000000100)={0x1, 0x58, &(0x7f0000000080)={0x0, 0x0, 0x0, 0x0, 0x0, 0x0, ""/16, 0x0}}, 0x10) ioctl$sock_inet6_SIOCADDRT(r2, 0x890b, &(0x7f0000000140)={@private1={0xfc, 0x1, '\x00', 0x1}, @initdev={0xfe, 0x88, '\x00', 0x0, 0x0}, @ipv4={'\x00', '\xff\xff', @initdev={0xac, 0x1e, 0x1, 0x0}}, 0xf8, 0x7ff, 0x7, 0x400, 0x4, 0x200000, r3}) [ 2808.822557][ T2529] bond1275: entered promiscuous mode [ 2808.828744][ T2529] 8021q: adding VLAN 0 to HW filter on device bond1275 [ 2808.940149][ T2534] bond1275: (slave bridge1238): making interface the new active one [ 2808.950211][ T2534] bridge1238: entered promiscuous mode [ 2808.968444][ T2534] bond1275: (slave bridge1238): Enslaving as an active interface with an up link 04:12:16 executing program 5: r0 = socket$netlink(0x10, 0x3, 0x0) r1 = socket$netlink(0x10, 0x3, 0x0) r2 = socket(0x10, 0x803, 0x0) sendmsg$NL80211_CMD_CRIT_PROTOCOL_START(r2, &(0x7f0000000580)={0x0, 0x0, &(0x7f0000000540)={0x0, 0x1c}}, 0x0) getsockname$packet(r2, &(0x7f0000000600)={0x11, 0x0, 0x0, 0x1, 0x0, 0x6, @broadcast}, &(0x7f0000000100)=0x14) sendmsg$nl_route(r1, &(0x7f0000000040)={0x0, 0x0, &(0x7f0000000000)={&(0x7f00000000c0)=ANY=[@ANYBLOB="3c00000010008506eb9afc4cd8d06e754a0081c5", @ANYRES32=r3, @ANYBLOB="2377f292252155b21c0012000c000100626f6e64000000000c0002000800010001"], 0x3c}}, 0x0) sendmsg$nl_route(r0, &(0x7f0000000240)={0x0, 0x0, &(0x7f0000000680)={&(0x7f0000000400)=@newlink={0x3c, 0x10, 0xffffff1f, 0x0, 0x2535f5, {}, [@IFLA_LINKINFO={0x14, 0x12, 0x0, 0x1, @bridge={{0xb}, {0x4}}}, @IFLA_MASTER={0x8, 0xa, r3}]}, 0x3c}}, 0x0) [ 2809.008786][ T2540] netlink: 'syz-executor.2': attribute type 1 has an invalid length. [ 2809.120324][ T2540] bond1266: entered promiscuous mode [ 2809.126181][ T2540] 8021q: adding VLAN 0 to HW filter on device bond1266 [ 2809.216250][ T2543] bond1266: (slave bridge1187): making interface the new active one [ 2809.225127][ T2543] bridge1187: entered promiscuous mode [ 2809.238037][ T2543] bond1266: (slave bridge1187): Enslaving as an active interface with an up link 04:12:16 executing program 2: r0 = socket$netlink(0x10, 0x3, 0x0) r1 = socket$netlink(0x10, 0x3, 0x0) r2 = socket(0x10, 0x803, 0x0) sendmsg$NL80211_CMD_CRIT_PROTOCOL_START(r2, &(0x7f0000000580)={0x0, 0x0, &(0x7f0000000540)={0x0, 0x1c}}, 0x0) getsockname$packet(r2, &(0x7f0000000600)={0x11, 0x0, 0x0, 0x1, 0x0, 0x6, @broadcast}, &(0x7f0000000100)=0x14) sendmsg$nl_route(r1, &(0x7f0000000040)={0x0, 0x0, &(0x7f0000000000)={&(0x7f00000000c0)=ANY=[@ANYBLOB="3c00000010008506eb9afc4cd8d06e754a0081c5", @ANYRES32=r3, @ANYBLOB="2377f292252155b21c0012000c000100626f6e64000000000c0002000800010001"], 0x3c}}, 0x0) sendmsg$nl_route(r0, &(0x7f0000000240)={0x0, 0x0, &(0x7f0000000680)={&(0x7f0000000400)=@newlink={0x3c, 0x10, 0xffffff1f, 0x0, 0x253c24, {}, [@IFLA_LINKINFO={0x14, 0x12, 0x0, 0x1, @bridge={{0xb}, {0x4}}}, @IFLA_MASTER={0x8, 0xa, r3}]}, 0x3c}}, 0x0) [ 2809.281337][ T2547] netlink: 'syz-executor.1': attribute type 1 has an invalid length. [ 2809.414184][ T2547] bond1285: entered promiscuous mode [ 2809.419906][ T2547] 8021q: adding VLAN 0 to HW filter on device bond1285 04:12:17 executing program 3: r0 = socket$netlink(0x10, 0x3, 0x0) sendmsg$netlink(r0, &(0x7f00000047c0)={0x0, 0x0, &(0x7f0000000000)=[{&(0x7f0000000100)=ANY=[@ANYBLOB="18000000560001060000006900db730007"], 0x18}], 0x1}, 0x0) [ 2809.516561][ T2550] bond1285: (slave bridge1231): making interface the new active one [ 2809.524779][ T2550] bridge1231: entered promiscuous mode [ 2809.536102][ T2550] bond1285: (slave bridge1231): Enslaving as an active interface with an up link 04:12:17 executing program 0: r0 = socket$inet6(0xa, 0x6, 0x0) bind$inet6(r0, &(0x7f0000000000)={0xa, 0x4e20}, 0x1c) (async) listen(r0, 0x6) (async) r1 = socket$inet6(0xa, 0x6, 0x0) connect$inet6(r1, &(0x7f0000000200)={0xa, 0x0, 0x0, @initdev={0xfe, 0x88, '\x00', 0x0, 0x0}, 0x3}, 0x1c) connect$inet6(r1, &(0x7f0000000040)={0xa, 0x4e20, 0x0, @dev={0xfe, 0x80, '\x00', 0xf}}, 0x1c) (async) sendmmsg$inet6(r1, &(0x7f0000005a00)=[{{0x0, 0x0, 0x0}}, {{0x0, 0x0, 0x0, 0x0, &(0x7f0000005bc0)=ANY=[@ANYBLOB="5000000000000000290000003600000000060000000000000401000730000000000a0000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000020000000000000002900000037000000000000000000000004010000000000001400000000000000290000000000000160000000000000002900000037"], 0x100}}], 0x2, 0x0) r2 = socket$inet6(0xa, 0x80002, 0x88) bind$inet6(r2, &(0x7f0000000580)={0xa, 0x10010000004e20}, 0x1c) (async) setsockopt$inet6_udp_int(r2, 0x11, 0xb, &(0x7f0000000080), 0x4) (async) bpf$BPF_GET_MAP_INFO(0xf, &(0x7f0000000100)={0x1, 0x58, &(0x7f0000000080)={0x0, 0x0, 0x0, 0x0, 0x0, 0x0, ""/16, 0x0}}, 0x10) ioctl$sock_inet6_SIOCADDRT(r2, 0x890b, &(0x7f0000000140)={@private1={0xfc, 0x1, '\x00', 0x1}, @initdev={0xfe, 0x88, '\x00', 0x0, 0x0}, @ipv4={'\x00', '\xff\xff', @initdev={0xac, 0x1e, 0x1, 0x0}}, 0xf8, 0x7ff, 0x7, 0x400, 0x4, 0x200000, r3}) [ 2809.579793][ T2558] netlink: 'syz-executor.5': attribute type 1 has an invalid length. 04:12:17 executing program 4: socket$inet6(0xa, 0x1, 0x900000) 04:12:17 executing program 1: r0 = socket$netlink(0x10, 0x3, 0x0) r1 = socket$netlink(0x10, 0x3, 0x0) r2 = socket(0x10, 0x803, 0x0) sendmsg$NL80211_CMD_CRIT_PROTOCOL_START(r2, &(0x7f0000000580)={0x0, 0x0, &(0x7f0000000540)={0x0, 0x1c}}, 0x0) getsockname$packet(r2, &(0x7f0000000600)={0x11, 0x0, 0x0, 0x1, 0x0, 0x6, @broadcast}, &(0x7f0000000100)=0x14) sendmsg$nl_route(r1, &(0x7f0000000040)={0x0, 0x0, &(0x7f0000000000)={&(0x7f00000000c0)=ANY=[@ANYBLOB="3c00000010008506eb9afc4cd8d06e754a0081c5", @ANYRES32=r3, @ANYBLOB="2377f292252155b21c0012000c000100626f6e64000000000c0002000800010001"], 0x3c}}, 0x0) sendmsg$nl_route(r0, &(0x7f0000000240)={0x0, 0x0, &(0x7f0000000680)={&(0x7f0000000400)=@newlink={0x3c, 0x10, 0xffffff1f, 0x0, 0x9672500, {}, [@IFLA_LINKINFO={0x14, 0x12, 0x0, 0x1, @bridge={{0xb}, {0x4}}}, @IFLA_MASTER={0x8, 0xa, r3}]}, 0x3c}}, 0x0) 04:12:17 executing program 4: socket$inet6(0xa, 0x1, 0x900000) [ 2809.799371][ T2558] bond1276: entered promiscuous mode [ 2809.814794][ T2558] 8021q: adding VLAN 0 to HW filter on device bond1276 04:12:17 executing program 4: socket$inet6(0xa, 0x1, 0x900000) socket$inet6(0xa, 0x1, 0x900000) (async) [ 2810.031160][ T2559] bond1276: (slave bridge1239): making interface the new active one [ 2810.069036][ T2559] bridge1239: entered promiscuous mode [ 2810.119837][ T2559] bond1276: (slave bridge1239): Enslaving as an active interface with an up link 04:12:17 executing program 4: r0 = socket$inet6(0xa, 0x800, 0x80) ioctl$sock_SIOCETHTOOL(r0, 0x89f0, &(0x7f0000000000)={'bridge0\x00', &(0x7f0000000200)=ANY=[@ANYBLOB="10000000008000"/36]}) connect$inet6(r0, &(0x7f0000000040)={0xa, 0x4e23, 0x2, @local, 0x3}, 0x1c) r1 = socket$inet_sctp(0x2, 0x5, 0x84) socket$inet_sctp(0x2, 0x1, 0x84) r2 = socket$inet(0x2, 0x4000000000080001, 0x0) setsockopt$IPT_SO_SET_REPLACE(r2, 0x0, 0x40, &(0x7f00000008c0)=ANY=[@ANYBLOB="7261770000000000000000000000000000000000000096dd89ad65dec810000108f80000030000003003000098020000000000000000000000000000000000009802000098020000980200009802000098020000030080000000000000000000ffffffffe00000010000000000000000e4000000010000000000bd00000000007465616d5f736c6176655f31000000000000000000002000000000000000000000000000000000061e2695eddaca41000000000000000000c001080200000000000000000000000000000000000000005001686173686c696d6974000000000000000000000000000000000000000002726f736530000000000000000000000000000000000000000000000000000000000000e4ff080000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000090000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000bcef5a18d4a047d60000000000000000000000070000000000000000000000000000000000090000000000000000000000000000000000000000000000000000000000006c0000000000000000210002000000000000000000000000000000000000000000fcffffffffffffff00000000000000000000000003000000070000000000000000000000000000004800435400000000000002000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000001000000000000000000000000000000000000000000000000000000000000000000000000000009000000000000000000000000000800200000000000000000000000000400000000000000000000000000000070009000000000094100000000000000000000000000000020004e4f545241434b0000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000007000980000000000000000000000001000000000000000002800000000000000004000000000000000000000000000000000000000000000feffffff"], 0x1) getsockopt$inet_sctp_SCTP_MAX_BURST(r2, 0x84, 0xd, &(0x7f00000001c0)=@assoc_value={0x0}, &(0x7f0000000180)=0x8) setsockopt$inet_sctp_SCTP_ASSOCINFO(r1, 0x84, 0x1, &(0x7f0000000100)={r3}, 0x14) setsockopt$inet_sctp6_SCTP_PEER_ADDR_PARAMS(r0, 0x84, 0x9, &(0x7f00000000c0)={r3, @in={{0x2, 0x4e22, @remote}}, 0x9, 0x800, 0x4fad7375, 0x6, 0x2, 0x3, 0xff}, 0x9c) r4 = socket(0x1a, 0x4, 0x3) setsockopt$inet6_tcp_TCP_CONGESTION(r4, 0x6, 0xd, &(0x7f0000000080)='nv\x00', 0x3) [ 2810.226714][ T2562] bond1267: entered promiscuous mode [ 2810.236718][ T2562] 8021q: adding VLAN 0 to HW filter on device bond1267 04:12:17 executing program 5: r0 = socket$netlink(0x10, 0x3, 0x0) r1 = socket$netlink(0x10, 0x3, 0x0) r2 = socket(0x10, 0x803, 0x0) sendmsg$NL80211_CMD_CRIT_PROTOCOL_START(r2, &(0x7f0000000580)={0x0, 0x0, &(0x7f0000000540)={0x0, 0x1c}}, 0x0) getsockname$packet(r2, &(0x7f0000000600)={0x11, 0x0, 0x0, 0x1, 0x0, 0x6, @broadcast}, &(0x7f0000000100)=0x14) sendmsg$nl_route(r1, &(0x7f0000000040)={0x0, 0x0, &(0x7f0000000000)={&(0x7f00000000c0)=ANY=[@ANYBLOB="3c00000010008506eb9afc4cd8d06e754a0081c5", @ANYRES32=r3, @ANYBLOB="2377f292252155b21c0012000c000100626f6e64000000000c0002000800010001"], 0x3c}}, 0x0) sendmsg$nl_route(r0, &(0x7f0000000240)={0x0, 0x0, &(0x7f0000000680)={&(0x7f0000000400)=@newlink={0x3c, 0x10, 0xffffff1f, 0x0, 0x2535f6, {}, [@IFLA_LINKINFO={0x14, 0x12, 0x0, 0x1, @bridge={{0xb}, {0x4}}}, @IFLA_MASTER={0x8, 0xa, r3}]}, 0x3c}}, 0x0) [ 2810.295270][ T2563] bond1267: (slave bridge1188): making interface the new active one [ 2810.303767][ T2563] bridge1188: entered promiscuous mode [ 2810.322165][ T2563] bond1267: (slave bridge1188): Enslaving as an active interface with an up link 04:12:17 executing program 2: r0 = socket$netlink(0x10, 0x3, 0x0) r1 = socket$netlink(0x10, 0x3, 0x0) r2 = socket(0x10, 0x803, 0x0) sendmsg$NL80211_CMD_CRIT_PROTOCOL_START(r2, &(0x7f0000000580)={0x0, 0x0, &(0x7f0000000540)={0x0, 0x1c}}, 0x0) getsockname$packet(r2, &(0x7f0000000600)={0x11, 0x0, 0x0, 0x1, 0x0, 0x6, @broadcast}, &(0x7f0000000100)=0x14) sendmsg$nl_route(r1, &(0x7f0000000040)={0x0, 0x0, &(0x7f0000000000)={&(0x7f00000000c0)=ANY=[@ANYBLOB="3c00000010008506eb9afc4cd8d06e754a0081c5", @ANYRES32=r3, @ANYBLOB="2377f292252155b21c0012000c000100626f6e64000000000c0002000800010001"], 0x3c}}, 0x0) sendmsg$nl_route(r0, &(0x7f0000000240)={0x0, 0x0, &(0x7f0000000680)={&(0x7f0000000400)=@newlink={0x3c, 0x10, 0xffffff1f, 0x0, 0x253c25, {}, [@IFLA_LINKINFO={0x14, 0x12, 0x0, 0x1, @bridge={{0xb}, {0x4}}}, @IFLA_MASTER={0x8, 0xa, r3}]}, 0x3c}}, 0x0) 04:12:17 executing program 3: r0 = socket$netlink(0x10, 0x3, 0x0) sendmsg$netlink(r0, &(0x7f00000047c0)={0x0, 0x0, &(0x7f0000000000)=[{&(0x7f0000000100)=ANY=[@ANYBLOB="18000000560001060000006a00db730007"], 0x18}], 0x1}, 0x0) 04:12:17 executing program 0: r0 = socket$inet6(0xa, 0x6, 0x0) bind$inet6(r0, &(0x7f0000000000)={0xa, 0x4e20}, 0x1c) listen(r0, 0x6) r1 = socket$inet6(0xa, 0x6, 0x0) connect$inet6(r1, &(0x7f0000000200)={0xa, 0x0, 0x0, @initdev={0xfe, 0x88, '\x00', 0x0, 0x0}, 0x3}, 0x1c) connect$inet6(r1, &(0x7f0000000040)={0xa, 0x4e20, 0x0, @dev={0xfe, 0x80, '\x00', 0xf}}, 0x1c) sendmmsg$inet6(r1, &(0x7f0000005a00)=[{{0x0, 0x0, 0x0}}, {{0x0, 0x0, 0x0, 0x0, &(0x7f0000000080)=ANY=[@ANYBLOB="5010000000000000290000003600000000060000000000000401000730000000000a00000000000000603df0c8ab5155860000000000000000000000000000000000000000000000000000000000000020000000000000002900000037000000000000000000000004010000000000001400000000000000290000000008000160000000000000002900000037016c76ecba90d1b11a05260d3562b5b3ba398cf4f236fd9e79d06b166824fd27402524730f75f69e080bbd263093805081251d4dfa00000000"], 0x100}}], 0x2, 0x0) r2 = socket$inet6(0xa, 0x80002, 0x88) bind$inet6(r2, &(0x7f0000000580)={0xa, 0x10010000004e20}, 0x1c) setsockopt$inet6_udp_int(r2, 0x11, 0xb, &(0x7f0000000080), 0x4) r3 = socket$inet6(0xa, 0x80002, 0x88) bind$inet6(r3, &(0x7f0000000580)={0xa, 0x10010000004e20}, 0x1c) setsockopt$inet6_udp_int(r3, 0x11, 0xb, &(0x7f0000000080), 0x4) setsockopt$inet6_int(r3, 0x29, 0x35, &(0x7f00000001c0)=0x5, 0x4) sendfile(r2, r1, &(0x7f0000000180)=0x3, 0x100000000) [ 2810.560927][ T2575] bond1286: entered promiscuous mode [ 2810.570878][ T2575] 8021q: adding VLAN 0 to HW filter on device bond1286 04:12:18 executing program 0: r0 = socket$inet6(0xa, 0x6, 0x0) bind$inet6(r0, &(0x7f0000000000)={0xa, 0x4e20}, 0x1c) listen(r0, 0x6) r1 = socket$inet6(0xa, 0x6, 0x0) connect$inet6(r1, &(0x7f0000000200)={0xa, 0x0, 0x0, @initdev={0xfe, 0x88, '\x00', 0x0, 0x0}, 0x3}, 0x1c) connect$inet6(r1, &(0x7f0000000040)={0xa, 0x4e20, 0x0, @dev={0xfe, 0x80, '\x00', 0xf}}, 0x1c) (async) sendmmsg$inet6(r1, &(0x7f0000005a00)=[{{0x0, 0x0, 0x0}}, {{0x0, 0x0, 0x0, 0x0, &(0x7f0000000080)=ANY=[@ANYBLOB="5010000000000000290000003600000000060000000000000401000730000000000a00000000000000603df0c8ab5155860000000000000000000000000000000000000000000000000000000000000020000000000000002900000037000000000000000000000004010000000000001400000000000000290000000008000160000000000000002900000037016c76ecba90d1b11a05260d3562b5b3ba398cf4f236fd9e79d06b166824fd27402524730f75f69e080bbd263093805081251d4dfa00000000"], 0x100}}], 0x2, 0x0) (async) r2 = socket$inet6(0xa, 0x80002, 0x88) bind$inet6(r2, &(0x7f0000000580)={0xa, 0x10010000004e20}, 0x1c) (async) setsockopt$inet6_udp_int(r2, 0x11, 0xb, &(0x7f0000000080), 0x4) (async) r3 = socket$inet6(0xa, 0x80002, 0x88) bind$inet6(r3, &(0x7f0000000580)={0xa, 0x10010000004e20}, 0x1c) setsockopt$inet6_udp_int(r3, 0x11, 0xb, &(0x7f0000000080), 0x4) setsockopt$inet6_int(r3, 0x29, 0x35, &(0x7f00000001c0)=0x5, 0x4) sendfile(r2, r1, &(0x7f0000000180)=0x3, 0x100000000) 04:12:18 executing program 0: r0 = socket$inet6(0xa, 0x6, 0x0) bind$inet6(r0, &(0x7f0000000000)={0xa, 0x4e20}, 0x1c) (async) bind$inet6(r0, &(0x7f0000000000)={0xa, 0x4e20}, 0x1c) listen(r0, 0x6) (async) listen(r0, 0x6) r1 = socket$inet6(0xa, 0x6, 0x0) connect$inet6(r1, &(0x7f0000000200)={0xa, 0x0, 0x0, @initdev={0xfe, 0x88, '\x00', 0x0, 0x0}, 0x3}, 0x1c) connect$inet6(r1, &(0x7f0000000040)={0xa, 0x4e20, 0x0, @dev={0xfe, 0x80, '\x00', 0xf}}, 0x1c) sendmmsg$inet6(r1, &(0x7f0000005a00)=[{{0x0, 0x0, 0x0}}, {{0x0, 0x0, 0x0, 0x0, &(0x7f0000000080)=ANY=[@ANYBLOB="5010000000000000290000003600000000060000000000000401000730000000000a00000000000000603df0c8ab5155860000000000000000000000000000000000000000000000000000000000000020000000000000002900000037000000000000000000000004010000000000001400000000000000290000000008000160000000000000002900000037016c76ecba90d1b11a05260d3562b5b3ba398cf4f236fd9e79d06b166824fd27402524730f75f69e080bbd263093805081251d4dfa00000000"], 0x100}}], 0x2, 0x0) r2 = socket$inet6(0xa, 0x80002, 0x88) bind$inet6(r2, &(0x7f0000000580)={0xa, 0x10010000004e20}, 0x1c) setsockopt$inet6_udp_int(r2, 0x11, 0xb, &(0x7f0000000080), 0x4) r3 = socket$inet6(0xa, 0x80002, 0x88) bind$inet6(r3, &(0x7f0000000580)={0xa, 0x10010000004e20}, 0x1c) setsockopt$inet6_udp_int(r3, 0x11, 0xb, &(0x7f0000000080), 0x4) (async) setsockopt$inet6_udp_int(r3, 0x11, 0xb, &(0x7f0000000080), 0x4) setsockopt$inet6_int(r3, 0x29, 0x35, &(0x7f00000001c0)=0x5, 0x4) (async) setsockopt$inet6_int(r3, 0x29, 0x35, &(0x7f00000001c0)=0x5, 0x4) sendfile(r2, r1, &(0x7f0000000180)=0x3, 0x100000000) (async) sendfile(r2, r1, &(0x7f0000000180)=0x3, 0x100000000) [ 2810.885492][ T2578] bond1286: (slave bridge1232): making interface the new active one [ 2810.935522][ T2578] bridge1232: entered promiscuous mode 04:12:18 executing program 0: r0 = socket$inet6(0xa, 0x6, 0x0) bind$inet6(r0, &(0x7f0000000000)={0xa, 0x4e20}, 0x1c) listen(r0, 0x6) r1 = socket$inet6(0xa, 0x6, 0x0) connect$inet6(r1, &(0x7f0000000200)={0xa, 0x0, 0x0, @initdev={0xfe, 0x88, '\x00', 0x0, 0x0}, 0x3}, 0x1c) connect$inet6(r1, &(0x7f0000000040)={0xa, 0x4e20, 0x0, @dev={0xfe, 0x80, '\x00', 0xf}}, 0x1c) sendmmsg$inet6(r1, &(0x7f0000005a00)=[{{0x0, 0x0, 0x0}}, {{0x0, 0x0, 0x0, 0x0, &(0x7f0000005bc0)=ANY=[@ANYBLOB="5000000000000000290000003600000000060040000000000000000730000000000a5134a31eb5b6a60191796200000000000000ea0000000000000000000000000000000000000000000000000000000000000000000000000000002000000000000000290000003700000079ca681a5a9aff30000000000000000004010000000000002900"], 0x100}}], 0x2, 0x0) [ 2811.045447][ T2578] bond1286: (slave bridge1232): Enslaving as an active interface with an up link 04:12:18 executing program 1: r0 = socket$netlink(0x10, 0x3, 0x0) r1 = socket$netlink(0x10, 0x3, 0x0) r2 = socket(0x10, 0x803, 0x0) sendmsg$NL80211_CMD_CRIT_PROTOCOL_START(r2, &(0x7f0000000580)={0x0, 0x0, &(0x7f0000000540)={0x0, 0x1c}}, 0x0) getsockname$packet(r2, &(0x7f0000000600)={0x11, 0x0, 0x0, 0x1, 0x0, 0x6, @broadcast}, &(0x7f0000000100)=0x14) sendmsg$nl_route(r1, &(0x7f0000000040)={0x0, 0x0, &(0x7f0000000000)={&(0x7f00000000c0)=ANY=[@ANYBLOB="3c00000010008506eb9afc4cd8d06e754a0081c5", @ANYRES32=r3, @ANYBLOB="2377f292252155b21c0012000c000100626f6e64000000000c0002000800010001"], 0x3c}}, 0x0) sendmsg$nl_route(r0, &(0x7f0000000240)={0x0, 0x0, &(0x7f0000000680)={&(0x7f0000000400)=@newlink={0x3c, 0x10, 0xffffff1f, 0x0, 0xa000000, {}, [@IFLA_LINKINFO={0x14, 0x12, 0x0, 0x1, @bridge={{0xb}, {0x4}}}, @IFLA_MASTER={0x8, 0xa, r3}]}, 0x3c}}, 0x0) 04:12:18 executing program 0: r0 = socket$inet6(0xa, 0x6, 0x0) bind$inet6(r0, &(0x7f0000000000)={0xa, 0x4e20}, 0x1c) listen(r0, 0x6) (async) r1 = socket$inet6(0xa, 0x6, 0x0) connect$inet6(r1, &(0x7f0000000200)={0xa, 0x0, 0x0, @initdev={0xfe, 0x88, '\x00', 0x0, 0x0}, 0x3}, 0x1c) (async) connect$inet6(r1, &(0x7f0000000040)={0xa, 0x4e20, 0x0, @dev={0xfe, 0x80, '\x00', 0xf}}, 0x1c) sendmmsg$inet6(r1, &(0x7f0000005a00)=[{{0x0, 0x0, 0x0}}, {{0x0, 0x0, 0x0, 0x0, &(0x7f0000005bc0)=ANY=[@ANYBLOB="5000000000000000290000003600000000060040000000000000000730000000000a5134a31eb5b6a60191796200000000000000ea0000000000000000000000000000000000000000000000000000000000000000000000000000002000000000000000290000003700000079ca681a5a9aff30000000000000000004010000000000002900"], 0x100}}], 0x2, 0x0) 04:12:18 executing program 0: r0 = socket$inet6(0xa, 0x6, 0x0) bind$inet6(r0, &(0x7f0000000000)={0xa, 0x4e20}, 0x1c) (async) listen(r0, 0x6) r1 = socket$inet6(0xa, 0x6, 0x0) connect$inet6(r1, &(0x7f0000000200)={0xa, 0x0, 0x0, @initdev={0xfe, 0x88, '\x00', 0x0, 0x0}, 0x3}, 0x1c) (async) connect$inet6(r1, &(0x7f0000000040)={0xa, 0x4e20, 0x0, @dev={0xfe, 0x80, '\x00', 0xf}}, 0x1c) (async) sendmmsg$inet6(r1, &(0x7f0000005a00)=[{{0x0, 0x0, 0x0}}, {{0x0, 0x0, 0x0, 0x0, &(0x7f0000005bc0)=ANY=[@ANYBLOB="5000000000000000290000003600000000060040000000000000000730000000000a5134a31eb5b6a60191796200000000000000ea0000000000000000000000000000000000000000000000000000000000000000000000000000002000000000000000290000003700000079ca681a5a9aff30000000000000000004010000000000002900"], 0x100}}], 0x2, 0x0) [ 2811.540128][ T2594] bond1277: entered promiscuous mode [ 2811.572614][ T2594] 8021q: adding VLAN 0 to HW filter on device bond1277 [ 2811.781194][ T2596] bond1277: (slave bridge1240): making interface the new active one [ 2811.842473][ T2596] bridge1240: entered promiscuous mode 04:12:19 executing program 0: r0 = socket$inet6(0xa, 0x6, 0x0) bind$inet6(r0, &(0x7f0000000000)={0xa, 0x4e20}, 0x1c) listen(r0, 0x6) r1 = socket$inet6(0xa, 0x6, 0x0) r2 = socket(0x3, 0x6, 0xdd) connect$inet6(r2, &(0x7f0000003ec0)={0xa, 0x4e21, 0x2, @private2={0xfc, 0x2, '\x00', 0x1}, 0xf5}, 0x1c) writev(r1, &(0x7f0000006040)=[{&(0x7f0000005ec0)="d1b56ae7f5a88544248e70b5dca818f965d20b97c74cfd4172e48d3ded3298ccdea9535c41534e24e1d864c34808ff776c1cd552ed78716e9bbb6bf0d27294f68eff5ecff7bb60c92faee5be484608f732b8eec6007722f7e3415acd5d1ca98631f2fe7b988fd5edcd54ce33f3703d8323ebe93cee7725955562422d5870fdb44f787a342a0faef293a775a6c012a403e715888dcee99eced99b", 0x9a}, {&(0x7f0000005f80)="83359419dec84b68e69e37b05f7e18d8acd2c6bb670b00ecb72010b2498a9220061d20b69af76f5207e108851b378b1713afc8dec2b86b2e26a883a0b5b5129b4b63be3048b5783de84a06e1eac86852959bdcbbc29deb00014e8b504b", 0x5d}, {&(0x7f0000006000)="cdc05fe9b555a26673c510b247d6c0c8837484d2c47b70affa9fcdbeff6978e197f0fc56de229dd3", 0x28}], 0x3) connect$inet6(0xffffffffffffffff, &(0x7f0000005e80)={0xa, 0x4e20, 0x10001, @rand_addr=' \x01\x00\x00\x00\x00\x00\x00\x00\x00\x00\x00\x00\x00\x00\x01', 0x7f}, 0x1c) sendto$inet6(0xffffffffffffffff, &(0x7f0000006140)="f6007c6225216fa83f80d015f9b2fafe139eee1756686ad1dfca9ae0273476e4836b5ce4ff42ba41cb389f50f075659a7402c60859cbd979aaa4d40aeae2dd2651326fb4141e974637f898923ac39e10c214ce2f9350fe84b351be7416c4c8144e0490533a0faa9e3894cdb094b6dca983e50c7a4bf10097de809490d15ec32495480b5f66ca85234600e3dc23193f4c76c10bb53a0f16e47d564b94d0681c4594f9153a82daee30500a7e4bbd877fbe17de30af9e6270df83296c736587c57a8f7675f2151b30d4ebdbc76b99795afcd32c578955758273353d178944082376f0d0ae4dc7d255d0c6b143ba8d93db9454ae74b802", 0xf5, 0x40810, 0x0, 0x0) connect$inet6(0xffffffffffffffff, &(0x7f0000000200)={0xa, 0x4e22, 0x200fff, @remote, 0x2}, 0x1c) connect$inet6(r1, &(0x7f0000000040)={0xa, 0x4e20, 0x0, @dev={0xfe, 0x80, '\x00', 0xf}}, 0x1c) sendmmsg$inet6(r1, &(0x7f0000005a00)=[{{0x0, 0x0, 0x0}}, {{0x0, 0x0, 0x0, 0x0, &(0x7f0000000080)=ANY=[@ANYBLOB="5000000000000000839a6de73600000000060000000000000401000730000000000a000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000002000000000000000810000003700000000000000000000000401000000000000140000000000000029000000000000016000000000000000290000003721c218959b09a462fa9c0e1c6feadbfda83ee9a1785d09930585374e38ebc2f37a461f"], 0x100}}], 0x2, 0x0) r3 = socket$inet6_udp(0xa, 0x2, 0x0) ioctl$ifreq_SIOCGIFINDEX_team(r0, 0x8933, &(0x7f0000003e80)) sendmmsg$inet6(r3, &(0x7f0000005c80)=[{{&(0x7f0000000140)={0xa, 0x4e22, 0x7fff, @rand_addr=' \x01\x00', 0x7ff}, 0x1c, &(0x7f0000003480)=[{&(0x7f0000000180)="c22de8de49c8bc74ed6a27990b96c7bbffde6e451106cd7db38c6cfe57439b6ccfd0b713f70ae54f02270b4776d47ca2288261af3a5cc7029af46bf8119dc3311516af33ab353064c342299bac4eeb1392f74f4ffed1ad7addabaf1fa8010cf13fadac6c44bf6d2b", 0x68}, {&(0x7f0000000240)="c97c8d45fc51ccbd284a494f7fa337de3064db9bf3406abe667335fb85ab0ed822756db072b12fc6d77789ab7c1acd7eb42208104b970485ba497cf158a0ca4f98eea848531b15b3ea19cf906671947387cf789ce725ed5b8e20db53c4ba19fe700a12e3d8569518995b863575644a44f0956259c90c98c6add1eb0ecee40d799b20c5bc7186300cb062b46915a2e5b13af446e725cc02114689646ac7290dd637dd1392c1bf5189299197b6ac5106256696f0ff163a3c739be5", 0xba}, {&(0x7f0000000300)="16bead57a423edd4b1cca1b2504dcb0a38fdc890f65b7636d8daea0c", 0x1c}, {&(0x7f0000000340)="51522000b4abc08f282fa6a69ceecb5e211d38939472ce2f2f5337ba56a8c81a5e9b51d5413e3444980182b0fb8b08782b71830fc891213a280a354245d36cac7d7d6185d124a6f1eccf9bed0c8a9423da237af98bdf4a9190affe49352eac845cb25a71680a1157516364c90f75a0a7f544e8512912b91163aeb614e3b677d318163e3fdf33e3e17f9c45f0364acd3c39ce5f8a204553ee4e6ea7", 0x9b}, {&(0x7f0000000400)="103ddb76f5c3a6822e67decd0875a1b06ef2eb440a6503c0d920b69baf9137df4c1f3dbee66104c86d6908e8e497ac4f2ec7656337294621b9113bffa9f202879a2bbc3188545e5158e3795c5a6ce1382cf68b8965d2507615a23c22580813718c8c5bd7d0cb1e8cf01865e19026a5c2c724772a543d1fccdb2952561885ef64ab3f4c8c45d19e2a39892f0dc27dcb5c7546cdeb07502d85daa11369b995a80571577a527ccb869d0dc362aa6ce57da19b6e4671e95a966fe8d707037ecf9942bbf8e61d4c16a5d9cb5e2388ff4d4707dc266207627a1969170396ff0fd00ecfe2cb616dffb2ceb658b5622f6cdf5f2d06f6764ff46ea4bbced621e95aab636b4653e848a781ff8df5c69f7d357f4145628c57a9f4aae6af9279ffe252c3be1ea2c693b693a353ab092d3ccc4e99ab076fb290d2326e317d65548ff64373b888a2f2817f62a73458059168d118424d88bc0f8f067678eac0ab22ef1123083dcd46fbda980201f2163286286d2e086ac7daa12e40880bc0647415e568f2d7af092c8f4d85008b1a335f9ce8a3df4216861b579152b1ab678e294fc46c815d638ce4925c8b985ff55072ae52df0a6be6f6d17c8f781a0dc1e114449c3d947814fe3decc167719d0556ff1007aea91abd64a1dd151eb0e1887409fab15a4355f610213bc2b3dca4b94daeb00419af0a5c16470d3ce57dc844eff080c09772b1b2ccda8373d8d4c6934aadbba6bdd76735f78f2076624132614a1f86769ec76aaeeecabf3fb21a5db4f807e4dd64ac62385ec5970b7c07bcb50cbd6cdbf2e0325022db87c9cf1212a1145ee4f0d4e67abd86c5dcc1e00982ed884654ff0f9cd4fa5dc6988685c91dfa182789eff886c4b9abc0323fe4fd05c1cc37e922d7ff66e8b7e7186015a28dd32dce7cfdca009f27ddd091f385af46add3449a9c9195e1c9b0b2d78919a60ffdc072a3ac1ed862968f83701bc25ee9d60af2afd1aae81edd72822d536044399ab2fa79252c633a3adc23b51616a4906ae17f51662033997f68a4ed695bf04e0cc1fd380873d70fb26c56ee26ca04092993eec119beaf5810ce6dc260b9b0036551b3532b2c7def534d3a21f55d326a357b22b5be248c70aaf1f0753dc379e70c2ca8420e75eb0f718a152cc7a12d34c7cbfb9de3ec35ba311b2a26050b2fd6730ec21f9c48d13f60561951d6be3eb6cd6feb7fce4b8049e504d9ad2469620910e24fe6d2358fdd195dea1e427b2fb0cd8d91276c42bb45e25b7253e8d38851b1cbfa73b67070aa6d576a78eb9283d30b38825e73413db827cad05211552ce49fb364ac6cb5d4e26a86c041dd009290cfa8e261b5cf26162fdc8a5d091dc351ff09911b1cffb352e6b814b16edc6b059e0ec7e0e218041343676c703a17a2bec3c97d68f44562491f9ac0353f1a0657dd3511c2a489d13a9b97feb37abf2af95a9c4241a12429954ee24afd3b7ac672e479954f90c0a591ee903afc8866fa38c1358b2963683e49d54bce0594eda4015b1526599d51cbe21f93c5d7e0a446e782767bffb793f691d07929670c82d373c6648e82b70ef9840b3ada8d2aab84d1017b2a913da6b3b64175245010659bc3532f69c58599a03dcce41c29d05dd8e3a8d4716e8b19a5f4a3a8f37ed0f0903e5d6e38d66596a3c1cd48df3347fda424e7691d733359bbfb3feda692331ad81a6ce60764bc0889ce40301f46c6a71d8da8f732d25a3c94dfc1ef0eca420723851e37f3b3f21b659c09cdd174c610693f326d8bc5aae2b72bf114f95ff5e95d509208641afa8a262cfad171494ee62904174fd9bde07dec67ca1f071975d8a800034d4c13ec8a420708d10f9d5d7fb46124dff3bb7cf4fae06e114550122cf764194186f24a9701bdb279383760843cb0c3594d2fcd74607e4e7681eca1e4ae6f7a703795bd66eb07603e3b8920bd8538c479b8b9af34fe4cd2465bcfea5c15ba4c2330e9b6c9165da731eb6213894da8effb72773b70a812c7e110e4cbfe91a0afa372545bc0d3d60fcdcff889ffc6034c208ed41e37d60418e69ab6f285b47294d3e9ecbfb3a2a100189acafb9523e8fcbbcdd99a4be6a32b2a4e62c1f657eb1a3a1e9f151681773b3443f79dca58d45118aaa92463c6d3f47345bc074ae37477b4cbfe06e89bdff5c2660ab422cefd03f7b4d0cfce6c90b2ba134f4ae176f9acfab18e8d06931c44557d10373929a80a95696913a0dc29c01c827d79c984526fecbfd6d480083d47db817f68b5ebbb2e97a97b37ace1215872ae775267dc1ae6d41bf2b99faffc154c2afcabcbde090cbdb2d9b9e7fc3ae9aa3c7603e72f6ab25ac9487954c6ecdd0c43e336e495d7f0d84d9452f413eca875377f4fb78ba9bd16550fc90752c8a58db8f08811e513162ba329e75152a4503834db7b5d728c68daa245e318bbe23ef274c4617756b5edbf6527927269e4162468aa3097587718668152733a88b28748541fbf9ad9ffbf8ba9a9b8ba0a251f2eb214e72d08b64389ef1040c825931bd3a78cb9ac6fcbde9528fe57c2e4dacc9b5499258c3c3ae40a5c70d8599cadc1d200d9f7686e1ff0dfcc0bcb22e6550c531fbf185abf47dffd485b85783ce6f79191498de308b7bffbb16ab88bacf952413a7669c5057dc392b1a4055ed8dbf3cabcfd5e7c6bf46abf92fc6fa5077a1b1c669fd554cce25659d09e63fcef040cf8489af6b7485449a616453d733435fd225d56102e49af1977ca972b54b2c566e1ad98048c684e020b4523842a85a13f27795ad4378160844ec06190d358b2e38bc155980095370620945279e1648e299ac25e217731a3d81da91bf0a9cf111601877c347d3f48728bddbdc31c4c5093972c1f1d513a89a332e0046a8c911af2ee273c14ad3155d6d15e0d4816cc3d7ab58999126e1e039fc882e7b1fa63e4f366664f2f487538f6b1e4f6bcbffdc7d70f34c2cf4f53885a0e7ee1224c992189504abc8a702e10608aa66dbf4b4b305a3c4d99937058806c8af9b853546ba1bbb84bef9937c308757863817272ef8b95061a07dfd57518ee1c4e200303c24bcf982d9dd9fffcd677bfc2e33e4026af218e76c2637b0ab840097ceac645b17c107fa2c50bfbe97846c3f202430c750b3815590000e0a74e231efeffc234269aed69232426b2408e1d362879498c0e425d1f02c53cc86a4892cf680ea63fce419a6c7df843b64f217b25b1fb10049678874d81d733bf27fc5ec89aa7c416e4f2dd4c88b530036551b4465cc756bc31e15327caa299863a99307916412503f90a30ee79eb1068230bc304ef145a1bdf7a095c1196e687bf8e9f780358314571840f2b2ef35ca94f131404b8379b2d71b9503c6431ea932771994f9be635e5df670302c90745023c37169489517b372f373d0f52717cd572a713d1e93cda0e1fb05e96e0116f4df465f7bb271943e958ce9fee5845f79847dcd4c01b485c7b271d40b2d704a66b32c9b3f2733576733e03bd4af60c8b1079ad0af867f8601ce869d5f512b77c6bd4110eb61b6da4caae42b75309f978f39596e7eed58c7638901a3aa6b9f97d851cb4efa915a682db74cb64f8c28e268d65689846bcdd4fc2cafc5934e686639c0e2a08a4eb1fd14ab7231956a75decb263e99f74a7ec9baf6e146c157d63a32fbe40b9c4cb8f238a9122a5f34d9b62420cfb296c20e0be292cf57c90d8461425cf32197d35c7747d95051c3c4b4c459733f245f40922c32d4764a8777878df1be8fd42f85de3fd01225a1d8d9cdc756306dbff987c3b4bde1c229ea45f81b9f20e6958f8c8a6fbd7d6f7bf9259b060c2d64f74769c4a66dd0be11ee9abe3b625a9a7677a2ddb94a8ab825a8e4b6b75ffeb9f54acabc68ae115fabc093a8f444e19ee266d7ad96a4fa87e70effcf18cf184c32d5dd71e7067157ddbe5f0e2f875ca0c56aad214cc9ce8b1a8b1681c74873d5dbe6c93dceb3458854b07bece891005ac2b8b4800eac5ce0211b234e6670bea324806a3ae54f8e4a48c3d7bcd102c408120c7eef179a9f2cc336fdab92b0e78511022d371cff1cc2c8f23eb9051b1648821db6830e454a2b343fe670c063c222efab424fadd58707352a4f17c4faae8e738a1bc60def40e9b8c731a11ff1ab41876593a82e7f60c53499778ad44986fc9dfefadd6562b59257faf3b45097980ade15c50f9eb6d5ce622ee86b55aafd5ea532cad3877d8965b1d831b9577dd9743f7026163c07410af842496916db1360133ddfd26feb918530a0f7ab664b6f62dbf10aa55e78a00061451006b3f81577c7e55771a08c835b56c332ae77b3f529ecc72f22f8ff4b9296480b50f88a751f05032a109502a3715d0411349de02be10a59284a91dd22fa3f8246ab1513c1cf9e964734878023a9cf89536f4c0a5c3286ff654642a0ac42d72ed68583f48c267fa26c8554d44bd312552b1121db719fdceddb9b7f3ae50fbe82984cba814662ad3e6d3f644947c89caf902de2eb394885ce08e580f4d087bb009d6bd41be5b6ddd41e95d0d7919b2469632dbef5194983500cf6b008320d6b657f6506bd9ac992a1723c477807379226e3acb4d8da21adf5c4e7065bf3d3fd395fbfab060dbda3b54e65f64e3ce2d3d20000bf4748c369ab01849051adcb6f7699034a872204e2588b6a52dbe81c0996d01612e15d0cbf21a9fddf714aa549e1717e10d24e7b919fa4e87d83b7613b7b0ec2a0e7446a59dfbf82beb6524f41f0ad05898f8f0e47e1f6beef9c5519f92f516f6a2d6ff4a56c083e4af8e801681e8782e698ab11ad3410a210112bc6cdedb09dddf7ca681d38e03b38aa5c90838e3203a232364458917dda37dc1b161ae1d0febd69347af30769528ee293909c645c6120a6f34b5504bd20701011a3cea8bc301cbb90f6b8fb1b3b94fd9a6a9d53227848d54d9b5749d30b3bfa23763d41e64fc8c8e07a74d4e12001de7f4c51e14ea8a763575dabc9cbc2d65745ac9b2eda7fa146579513375d7ce55713618c4b5cd44a4633d6c2647866f745e405b6632af9aae292001ab7581a11289a4a4ae1f235c11040024f89f5317b58e6fc7d7b9ac8ebea567debc7a5d0efbe8eafb374d4159db45767a829a22f7e503b82e0399c1350e8f25b0f428745379f9cc66f093615577fa527d9fcddb8cade543ec4aacd0f46ebb297e8cc6709915ba3b21ed93ea9284306cf6b3759853ddc1bb994a4b2dff3de13309d70dc0fa76ac64aa8618ce6127f38d3013a4bf0d2a29c6319ffd3987b7ce98a3aba4affc12c52e8c2953e3024a8e200992580601fd0c0acbfb8645453ca752a7932d552462109d208d58ea0c10f2b9da8d214456959a71bebe20f2193d0abed2bb4af4cc7b6def0ad288abe9b96fd256931dd2e4476f237daff691bd9539fc736cf5b286f39535f7b35c821b745ac24a19cf3af9228603c713e9eb70eab31fe0804710b9d6f2ce435dd2b45dee997fa333a3d9c2f2ca15271c385936ce687b102a9be42711c1baaf9794ce69a82fde4a39aacda111fbb965d4f2f58f6b48c544983fabd3a05839e204618dde1855754278da4d48d48ccef3d1d9bef906cbd60af2aca183b802b6daa41ba3c6f72ec6b12942fcb9c0c74cf79f9c0e7137273de4f196896f9cab3f471a182ca31268e86be5b5c5b00ece320d683330c37ef00ff498950b91a6c71442af1bc16868a0ecdbbe71d0d2e513724c3e9c44c0a058a1e95122493380cea437048532fd8871c00e1d8e52f6ce95aada96968a76a2c8db620747d622bded19f9afd6", 0x1000}, {&(0x7f0000001400)="6fee70109ca972282af8091517580ea97c02", 0x12}, {&(0x7f0000001440)="cf58e1164b21dbfd7361a6608edb7c3d74752d30bc803105cdc9c267ebb3180a712f3329e2daddea7dfe9a", 0x2b}, {&(0x7f0000001480)="57b10b5886521c21e2472f7c92c83189c0bb4ce02bb5a06885f74e06936c6cbee4ec566161101f109bb1ef4c83803331bdd5e94831bf88064113339a3e7c4658a414a386877130b9746550b2551cdfbbd99d9675b28d2d3c8b00f1866b89a4dcb1cc223754e09889f4eb0c9a5f3be77531564fda6b4aef20a152c5b5f3ebe7c26930e9e95361615e8edd1aff4a9caac3e87e6c467f4abdf8640dfe6cb42d89b48e1130745d47ff43353e0de9cb20156bb2cd8a64bbfd22d32d7998d405c3d26eba0e48d42ec8b9147ff41cb6c56416819d66c00c82c64b8662ff23690062578f31de8265f252d50a42ce16b16c64324a26700af0e31c3b8f09dcee863895be7fd1ebeeb89f57a691d06ea81337977e41970a37feedbb49cf7e6462a3d755124a58fab46e662e7db352706838b9b24ced05f4a52c542f537de34f0fdee6032367d49697db1247e7a43e9a4a0102ba8d71f4893bb5474121c84e43cff9e233671866a73fa7716cf2012e80905dce31ce3d1c5d9fc38f8ec29ceb82c924efd47d5f4f7f9f24673a3dc1fa5f4bcc81caf5d96e0ab3a2eb093c057a32a3122d104dd306e4382c0ecb425d94737e3693f1117a4b1780cd1f9d8f1a5b61e11f52b48cfa21777e9833ee0a8a065e3d17c514b76a04c69fef9b430b6830415d8557b23fd5f26e39a620f4128923a6c0c4c2a24ca8b27b110333968ad99075cc9f222d724fd87866d9858df6d1259452914f14f8c6c7a70daca6b85055684b8dba8225078fde417c6029e6c45fe7d6730c1fd965ab7628539585b25059b1c5d20e98478ec92b796ad25d10b08483b118d146d2b73c7145c6cb05f521c48fbb98e50dd75ce522ffe2c2076463030f7dbf61c113bd27d3b2886a5d12d578f97d8623c93befb22972be01d3af9f5d4edbcf4da523053a3539c8d0e4ef407ed52ef98a49c7229aed12903a141cc2b16acae5f92e0dcc27719c5aebb93b08cd5d86f9eac038cec3476e4ed5adb727830567fb405b3929be507152e9e2230facddcc133103acd7209aedb0f251fda299454b40a721017627545eb5b8f0deec7797211375f14b57dbb9e3f06f4a33dabf179d5e98c03798251d62dc22d6a760244bbe6858542183b2d40f7c3c755fcdd04a8502521e98a7a5f59e4dff26746d275e0b40746e76699789f88e45e85abc9327c7773f73cb91b43a35694def999fb960723851660452882860fe6623ab87f13337cecf7c8b18f3ef5afed07d5134e168722097b7d45e16228821ef99dac527e961b63fd61628f4a7cf9f1231262c83ad5db9b550c98adad1cd15db17cb947f192030fb44d638172329c7d87c35e510f439c3c72d952632ceb044cc57ac3d374dede3c81c8e75ea20e731de636f541205cf8ceb14d998793eef95ba2320e5090b299894fbf885462a4d909cfb8d021ccce95d6eb86ed47c0c8c8bc384333b81ea1557a66603c5a2f23ab0100cd94e3521f9bdb14c4c32ba9d9bfb1011e132678289d52e6c0c8c877974fe187f80c7fd11a80a59055d6d9a132f456df339fe90cd1fcf0216bb1f1c139bfcc1dc72334d08f467996ce61b692da00eefdae66bc54db30388665eca192f360ca41e7401ef3dc0f150f5831bf6356cba27b642b01bb8dc7e3f711c80d43e751ea7a1ea3376edbc704f8765df546e11a2fd23e1abb034c62eff880806fc2098981ad0388f233bb1053dd2a0bb5b1605ebce3cdf39f58db49b85dbeabdf9e4561ba2de2d2436b4d411ad040b419f5ddfe8058d02cda2edd51c320f1a14ceb206bf79aa649fbb3804b3e15c1e937358bb88a4b8be3e6de6b2e28c7c84964619677745ff8c009807e40017df35c60c11410f5dc7335d539d1c3918ea760b822bef85de33208d0f3395bad8db442c26752a675015211a0d49524f2b5b7b85c1f9c8e7364be0b095b2e194c04d6d96d9ff030a46f5b6324bea678d27d1050b38cb09c3e5973b0ab783b60b58a9ea119f28ac728203eb444b743f130fe99d291195836be19c24690f26f850edb5dbb409cdccbb06015e7d5b3e64fbe31a18191989d85f9196c39b0c226d189504da902f5c1efebcc5582ac35b098181c9ea94aab515d5691f71c0a267ae4b52593901e0820645b57da3540e986b14e8c32d00e86efcf1192cba857bffa78b6dfb99f5e69bc7e1d7453ee4b8db6800f7b89535b79d1f9304d99dcf33a7aa7a1022a2b664f65e33cff13cc34e86ebad1ca6551fe7f3431f72cd1c4563232a084ef283ef4972c1aaa8fe182a9838cc2ab5d1608a577406c663dd183ee6f6a6c1089a30f8924e858c6c6987619aab0987fbc413108913b013162043ec5344a818d97403bfa8d3c61ad5c6c9a28c4059e5786435718b1327c4f202cbe956f37f99186d57373c40daf05f1a9bb4372ff5fb66a9d5f7c7a5c98e987835ab8650a5fd68c9368146ba98c83a780288e9bf26f57ff4cef3c06569c21d30c64767834c5a7705e4e0c916703c8cabd08c787ee5edbd6fd90d3f56a0ade79fb11a41ca08ba9b4da3816ffa10fa14fbd76aba422ea613760295157e33e2b20fd4621f6ae1a6d39873f04cbc88eed0e313935de7432f55fbe09b9bbb110011fc167378229898b128e7f47dc6dc48ad759bbe3338405c17690651a14f313b540d9dd3639decad86d364d4e313fdaa7e221c862599d9333d8d307611f9c5ca8bf96920697eac2167308a1f22cb220edb497d978a6f7dc0034f7f85eb9858a44df554d8d622fa31c0bd915ce7b57b77d9141cd832a943aaf593f219310528103c3fcb316618c3b649d43faef7cbaf0f844eab8095edddcdcf6dd56e4d507e7d5f7582b90c9fcb7e9f69066e1c6cd8ebc6097412bca7965c44142ef5764c8252d8b015dfb270ced3a39824e3da0c11a6af451d0c59a6c5ecd5b367bd0b4859a291d896e38c285ac3abee0631f2337df0a70a3997bd28ec626fbd2c1ea66192f20e41a4b5fa0c3395898efcc01da101c77eb551240be337805cde29cf637ab7f57467fa16bddac3ae58b15eeced6931098093f155afcb9ee3661b081b7102a701841cf14f2dcf13e955df44956d702d95bbc8bcb8a6542cc8c8bc50711365eb95dbf78fe0bab478ccb2873fc8f8ad4201a36620ed11a24caf032147e8d1ed7bde55f15203ced96a4c45258b8467b691947ac4a165dd70a0bf7a5ce5c9c1a7dfc62fc83bfec43d05b741d22d7d42146b534428180ff9658544e3d5da92824430caee91f2cb063c41afc99b4c6b36503de16d74654d0ea86c357c788fec88a39baa1fe4f0a35dffa42db60b35ae78a310931b93607fa3b21ed9b87ae1c6965bf69a206e52a11b0a836368549ef535179d1f9edece1c33e0f9f005629b3c16be4fe8662fd093ee149ed2ce005f98c97474ccf04cd1b558eb81d6245a47736a214be31855f610c09071717bda0b54d8faf177e597af4f3770ff68e4357bdf507f713f92d23b35db3eedddf1afb1cd37efae980a77b5c218fb118f36b9d8e0182ae8970f403a838bfd50d38368f52bfcce98fa007278779da80da7d9769c9d7e2d801c8de4e40ae4899207ef057c0794e4b6ba10ad5fc56e42fe0807c05819169bb9b8499901ee31f49745017532c248fdbc0ade0a6a25744ae73643bbb47e715e617a2138a5cc66d25eb50abc2687b738b31ff9cb6586903acc739b8c7983fff6216f089d4d6d1f6f31bd42bee8fe326f47bca495810e48b266c778974a1442b86976d0b3be808d02b39d88294676482865fe2d833c2a39b401738635bd6ae0231705a487305dddb3af2cc42a8b39a7e811f3ed52edb439acf3e12fa8f8b2b66bed00b9894caa5b6a0688f7e0f699815146eff5e3f6471812bf49bc1a7f1447efe523ff058d739cdba9694094a5c64a4933682a0096e94f29e1b6ee8b84672da3cf1c17f2c2a546b09163861d0b090eabc23ce54a9c8a5f10525ac44076906d7f7991b2a08d1a7ae60fb18f468950c6589d5c1f698612917ea49b66495d03b0c3db11fc300e39ffde962c5bcf401c3502f2db3ec29e823c0691d911ef85e82f97ffe3e834c92ac496f20e46c809bc976a49669721ad46d3a4b315931d516a426339354e9760874b8a612f63c50380657f99d7eb3501cc29f9372f161816ce1ae4d285e31e279929dee2b87ebce4360ef0b632990506d310fb8c67a6a6f122959d8331623ff308bebd105b36ce9d2ec9538a8bde87a5808a555b0ceb931c0c57d139ac1fc576e55cb7e58d4a1cb24889bbbcd670d0bae65f1eb2cdc4870810e7962c4bbe21115779c19a5d90fae05c621f1bb5b4c741f9164619759802fd5c887fea540f218077c94cee430a750dae7f1c1784096862a5158b75456b6827a27aaef0ac9d59ed1210e25127035c745ad148bedb8474a7abe10c866fa3f86cbe49b51f80bf22cd830ccf0e81de8ef71acce5520f9047fa376dc39ff9ce9ee11f57512392ff9910df3b6be9b58ac5df9968fe6d1840b9bcd4d132c0a4eb174474dfc6f9ba0f9001679f87af42eb3ca0ce07cd04cb5049a4b8d4f6517d44f83fe32c2f5fbfd0e673264342b9ce268fa42392dc5a27c7e34e3cba258488f380e3dfe633ca5a3f9d0833823ef2e43631096535bccf67305bf0b09a5d2a6bc20886936ba1c1615da02e98990e4a77f750c1aff8f8891a9eabd54b036da2c90589ac96c74f1fd43e01719d641ccf10568f9b7faece6fe0380668998a8896aa5d898a549e653f178a96dbff81bef5b6da102e71ecad171bad34423d3bf19940b287c4a2fda064d6ce4619ee5c726f0c8dae017b6a0cf204cb5cad2d945e2ad2c1551e6499467704b01e62741d7d752e3f58efa878f5ad42fc86aba15ddf0650fee8c88f5d431069494ef2bcb3685e1c732b347976d1a2377dae133e6e3060c1317125cf6507b1ec001973694ec60497464db4d74e7695f7e07d2c47caf6574dec77fae787cb12a9be276b99c61c54e74ef601485f800ab58f4069312cba6d09fd4c0024490742b38557f5198e7c15a928a73e958322d35a6764743e942676f11ee872564d7b2aec5a92b3930813c9d998a3b876ea52dafe5ce481fcf3bf28baa4729bf250e87b8ea5aeebc4506f15e0c60fd114105b3cb0e7457d3b1e348cb93ee220160719f9d53cbf8bbcfed64f2a3a93e426c11856199d0f402f3575e8080e8b0b8afa556001254bd4767982a13448ce049e80b1ca806d13c3780093c3e09766fb6a5768a14b9bea98bc84af060aac734299d81b1f94ddf1fad4c939d70114e0ea7d2e9652615a4a311f0e7cb23f4f288bf1b4b8164233fa50e12cedbfd336286d0c2871099590f05c68b2b4ace213cc1062e5f4e452627bfc3dcb87c040ff926be00f583a1a4691a708b9dcfc59469d1a08f24dd392fc2a0fc910e0fa59a1678b6ac5e3e1a340db8ab7814ef49c81c20c0433a7a3f1cfa8ce6eb88d9cbc5a7b839faab5ac199b6517043ae5a45eecadd8d1af204facea965f85227adbd8e84da33db5d1bc995015dd164969ab8309ff92190e91339781110a8e29a1a054b4e729ac76e1ed941a0fada54c04863e98aa0ab9586a3598078ee45efe102dfebdd223a203c0c8541748124537f54a9df0c6d7e7186e844f7fdd27d4ad1defb99170153f3a3df2eddbf7df5647394665b0165b5489219b1dbe0777c67d384f73946d8d25382cbcc760966f6e5ea21cff2c26d0e601d6b3051c3c92ba47b904aacadb5b6da6f65e2da15537b0a52aa3346b330be2b2b190b7c959d016391a9d2ecb6b6c800fa339d3ab70945e90147840a7563fd9e1", 0x1000}, {&(0x7f0000002480)="216d2255aed99be1147baa94d0465161193b42958a0d74eab35bc3e21fb46a3c5a059744bea8efa1abecd221d9f2d3d504dfafee023723037f27e3275f882b24fec454723cbf5ac96c7c56df94397d8d2934a512a2f27a38f76c5d14fe50e04a799c0ba06475be2ab1349d5aa1d4f8db47f35334ce14f0f40e8fe733fe4070b09a9484317ac3ae887f11dcdf001144c00e0a9a25aafd47294c7696d758905690fd1b2615cb2f4155f0bca056a9ea23c168927df1491f6c9742980b0ed703ac2ada240e6f022182b4662791274fa3214ff0b5f917694ec3b01e0c753489c68a3b1e9b2901e97d2cb81456b13a780513ea02fa79de92eb59ec56565d213826e56dcefc51c04b50e7af691562684beaf06cb27ed5d4095abbaead24de8f88b498b3dd15780d1c4b4d747d1f3cd33a5acfb1cd020e5020bb01353ec5f5c46fac84d00014a9bfcefc91c56b0b92a80a402c6cff887bd6c0315cdc8c79818a87372a4d36f8a35e9f3eafd4b37b0b08c9aa6a7baf91a0458609d3abd88a054c7bf180e371a77975afa9edd941b2cc9c05badac1c677abcd9a2547b788122aeb7034a9c6d21999c4c74a51fb1176220db94525719faac17319516ed9fffb75f9c1df822a4f4df07c01761e2a6a663b0fcea058c372ed6946644cbf737a3e0fa996d3f3a12d1789a73459075521334cb0717bb36632fdf1713a2fe2ddce70aa3a5c1322d1f17ac1f717e25fee59e47b954ef5ed45043d6bb063685013d539fb643efef00ca945391d8d0d05858a09928585e6e13a0915c51d38c07fb9a88801ef84cb11980b9e18486bbee9cda0ab460468a9803040947239d4473d391babe47303e6024ed921dd2aaf5c137058360d51d0d8994388788e2d6b14296330775d2bbd34e7c58147d55fed75aa7f97aaabde3f4ec784c221957badd2ad6d0024d92c2530dfd8851beae88df4efb334def1bf24edc2d3fa772c3cb6418349686f8a205d998f9e0daa4b26e8d87c200e9c5617cdbabaca1ea381be91f26a2db95199394f3f50f258f71ee30912d53f84571d10bbdb9301acc73d34a47c8253cbb80ba441c70bb1825913bda17c85d3996ec754257fd48b636f6c0a16847aa28665f25398fa1fc76a12ac6ed2e51cb63b67614a08604de0c1a25ca36b50be2392003d1c4b1022ffd53316af9a6578940f59c37af498b61493847cab34c4aa4023fdcd7e2a4d164e2d9255be8abcf6a64e7e132f5df353b5cdaff6e7846a0bc09536c69606698c76590086ab79b988a7826d4a98858ce5b5c78f97f61befaae8458c6e1b8bc31ec7996be064783f98187892cde36d529dde6101f1ab19e9d17c2db56978f74f3af8cc9ed5cb7b661536c4747754cbc8153f30aef0d12fbc8354aa13e9d1084e9d2b71971830687eccacdf85ee9ef42900166a03c4b80950ca5d51af438d4465a1651b468b047ef57894caedc5c4ea6639ab5c43cd2c0e0f5684cf0e678833a7aa24d854ec6d8034aee508c275c0f3f1be92b4d35911ba915997f44a10f81dfb027af84b4991fc72a5df57dfb1ae98ff7755fdc7a450df7d424604b9941f8d176aa78f547b948d93b304249f931b7417931b8c9942b826dacfccdf4ae3512e4e00fcfa06cc83f58af096c927bb4c7de8a3958c1f1ebe6bf8723030845ccf639725f053e0d702b404bc4b948b91d6c40bf35f5b0ba5b5abb578b3c35c06d0ca70a076003f2ba0072c5a29b33a073a509092c3705acc41e0a4ca46ae219f3489b28d9686060ede47d93b1bfe56f12820cfd35f4c0613c60315ebcdd1b8b213cfd302863254eb52adf904fe3c426e677b07389cdd47824246023fe003540e235d79054656a769b5048a30642125253442ed72c8c75a137d60eb626737a2fcc40043e31689d2b762238fd0deee115accd4b1fcae42367fed0d7fc671986283f276a4b0f9af108445cc6451dbb4cc0eb99e62a3ce17fb5f921808363a6916f22c1693dffd2e3907159e42910b334194ebb27c2fc29e57a937d4b3ad83960644075d9fe879c9a2ceb923e38c818af3fe516719acb875d08b0bc75f44e049610abb24db5a21f1b0c93071a15bc732d78c63c41d256d25f626b00b13edef5f7ad00ea3dc07ecc4f70545fa694b4cb3f8ba25c06e6755261efd29cdf1c280a86d727dec3f976ccb2d6b8b069b2e645f9e9c1f666cb2638723ff053c7e7631b3a5ba65b6b6ff053bb313fda95c3803557aa9daeb73b936f54679987188dfdb770af72e83df19e35fe08b7bfc2c9841ed2540f0ee005a4d1a4e8ac1b2eeed228d319f9f31761718f4cd3bfee38513905e5d276e59af4f0c3c7c3d935522e821ea1c54416148867c220469dfc7c6ed3ffcded7e51dd42594d1874ac1a6f7d98555443ab2e6b4f7270ee8e8a6aa7a93056247b5dc06d4f4e6416df3ad37662a913b1e128b68d2c13b0df064199f7a636980dc7d7ed3b8f6e6ba3369f00b416d3b111528bab2c51a2074f17a1b3afcaab6171bc64df64772930be2b6713d6c9a7cd10fc81dbde65fbb35fe20b12e8a3d560c39443788e09ef27771e216417f4ff2c1656e668f06e3b3a2b3c4073296a7296bbbdf0bf64979705f07f4f3a273a2e625f4ae328c79093f3c0324c072ada3de32693318fe563f66ea73a78e91e8fa69f99f3910c4f0878ca105a037bbb7ec2964fc991ba66a13f45f77108200226616e56d8084e75ff6a5b4a631421e868709471051972fd3104ed63681c9cb292e24e1f8a3d72901ebaf6689e96d5177d5c897b0100e64e363395e8dd5f62670eb2b67e63a2c80b6cea3f6c5b2bf118941df5d49d7b1ee0ba0d9366c827c21d6b4cc7f72099311b3b62420c14807329c4ca17da600db37e188b8781e294fcfe87f40bd5b2224ecdec13b36da9cda55caee246fcf603037699fd58bc10ae811736b86b816f56912e3018fa341a254b79c4682c94dc93384b57408a9c863d01a1f8607a33cc983e8fd363d67654cee5deec055dfe01a6fa1f4ce305d392cf2b2342e4b4ca4e61f5795c20b9355fead2abf4c590895be9f7c76530eb0a10adcb631e8a0ce1f6150d1a66f4ea77274b9de3926bb918e476cf8a095ebec2d3fa3cafd61fe143e3ed47749b7702e62b17ef6636a1b4921e892d8844e05fdf14b9c5b3e9a63f9374f42847bfb58079a8cb87c39004a888e5493dbcfc30a1f345808e2faaa2cfbb28b38f97587c3dd9f4c49430e15b1ab56582767374788ab6a83effe0c21a2e5b759e68f2e0896a160c4d3c962ede288733828f8e1c402a92987d3407f5c2943825bc902f9a4acc5f0daa96374dceb623594c92209cabbc30d31ad9525d85897f2087bf89ed5e9fa24e49dccc4d69d226a89bfca5b9142201dd78a5bf76a9165a8fed75500b0032e7eeb5b7def8a3a8590db6ffff5873a7d766b402c7ab10dbf77fc4ea6602af6c174a83c725d5b545bad2b589a4b16cf559c66162d80204cb03fc0d4e502e0d1ee10edb1506edb936747563c2ab6afa1e4d5f04266f0cd64bb8332ad825f11f00a9d2906534867fe5f9352f6ac97705acf171c8ddbe350e5b709133cf2d27577b15176768e6903f13c4c49bb0b0ac35ef95f6918737132f6a94f2e812853e6ae1295d7d567d4d4b953f6c5bdc9f36867247657ea44b58d54c9c39fb8dfc685bbc75ce9995c825fba731a3349d4a2d2da19486dfb7f18415b5e6cc8a67cd92dea7d0b146449e2c655ff98cd3f5d80574e072473b3854c0594f9458e769e8c96c71cbcf72c55491cd960c5d83540f4d53bc43867fccee87543507bfaee5200842b7c1fca23dc72fc7a5b89aa81031b69d4bf2544c35bf864fe9024301356d0887e35a28ac304edd5002ccaaf61c8b3323d0a1b039fe57fba27384e67add2ecc242f7e2821b02b34ad6fa7a22af23d66ec1911579df373ce8361ec1e3249c1ce618cd43d1fac6019ee174fe750bdd87bd61a4f0f44174ef68182f7b20c104c37bf2c0aadd5f7cdd8d7c1892e106067658e3a05d3133130137cef0a3384a39c54b6f01b02fd32444ff16e31d567b88a2d1a5f02064cb9306ae5a6cff532dbe5c65b0796692b00a95c9edb2b69330ddc461d95400fb13716b84ac0c587f5d6d6516f8afccf83d15732f8e07326e8758638e40f0a49038abb1725f5bbb8eda2fe4a19b2b9fe2ca71aad4d878e5441cf71cbc4738adcc6049b4cb936c6305a21fd90464f8a3c211873296866919e3fb6ee712a99c0c2de5bb6e7f3aa82aa2046278b761bcf676818e70b93fb814bd66a35a273fb96322af95b3c00a4df7c8c90f4dcd2f363951ec3d929043e449197d6fb2eb3bb4335afe7ae64e4f2134d25a0ebd59ac7292a145af5083d06086b84b78b3655fc7b1ed4450494148ff4a3f646617676381c963857b7b0096614399e4e2aa1f1a871d3d0ab7742ab720f75a494dfd1a83f76697f23597f7eba7b6ba5e3881ba5e0fe27f05223d8b733ce038bb83c472736e2cd563e74524d7a2d2164950d6cd3a6f06b63cf3a3c8f852a7ea4d4dde1e569cd93eecba0a6ce5798f05637d02bfbbe4989b62e2b2053244408bc1f5250639534d1215420f0aca2c3aa2622377270c886c56485e8dc690ccf790828705af2e659338832ed748df008060e2f938b47c9923f2446904b30abab3020f672785a35e8387ff22912e1c93f778cce578228c547bfcade5a1912067155c704df83121de3ff75fec50e67a1ad0969b0e40fb64c7753186c56977bc1ec7348a383af65c424acceff79a203cc222f9633ad4d8f878fba99bc164b11e5845673500c8c2ec5e5917040c55569fb6950337d10b5fbe73eca80be29eb02766c877bf22b4737b09ec42fbe3b0b9a3051de63e32dca713d1a05fe9641e14157ada1c500e9c45ceeb9edadb80f3baab16a92373f01f830e7600a5d1f36965d8e5ccbdc3c0e2b239227649a6f7ac9fb8e8d771698657fe5b3a4f608381c6bc1623ab9710c83970d7d880aa6cbb8d1ccdc23a7fe8e51645a0d2f6cbe434a137a49b8f5a8fc79bd4e01c7ced63ee3edc2befffcff92c246fbe512ae4d0b4ed259b7f989a98a6eadc1a2f178c84e07b28fb3a4b10cb575fa1cd2d0c25b2ce6b8b8fa4bcd4f814c6a1b75a3df7c0f67ea830d0db2ed7e45f79afc065f962a2b8a56d338bec08b107030e32cae7e62e51a7966871c2d6e8ec047720b95792ea5aa577d389f011f42bb7d1501432575c656f10ae1f13ef7287133b2cf631934525f8b9ee45814c55994028dc3425db6564b5755814bffffa2db6cdec29a9914fabaadaa370f271700aa27a750ef2db69eb540c8f9f2bf5457b41c11ca17c6ce3821fdb982193f52df9cb6d8710b247c9a59e5ee83af4e8e69574dfdd8fbc9d1083856e75ac88cdc7046ce17a10e286efa087b133a1f3d7918756b878b7ded016f470206c7bc0c51f325ec30b634026e0beb851d522fbbe0fcef664878660ea12a60bf4f3cbcd98d0078e0761668f548a164b8c0f049fce31c9433bda4d8261c5aa8963e3d2eb96ff6a02d83347651bf4d1d43865ab57d180d394588a2f78cc87bf710eabc5e110ce1a0ee3b7171844e3d1d4bf476a24ff51ef8f7e47829bf7c0d3228dd7e2a94b2a69e35686f9e6282e6561bbdce7a9bc212e91b82486d0e63f17830dbbc04ac0c7bd1a1526042db6e7a64a6070a6a8c47cd0c916671d1ad47a1a12bbd4075e4ea9b9f64aae5e9c5f1c9d2625f97f5fc5d1d50c9c0de7f91665ddbb1939058b02842c64607eb76a3d91cb2f9c17982078ec4a12e3e11fe2d6ef55796c7ef165b3ba43c079", 0x1000}], 0x9}}, {{&(0x7f0000003540)={0xa, 0x4e24, 0x0, @ipv4={'\x00', '\xff\xff', @broadcast}, 0x6}, 0x1c, &(0x7f0000003740)=[{&(0x7f0000003580)="33c3f780a5cee531f31c782e74ce2f72b3b2ae15a761fd2c4bc4f5c1bbe71c6a94666905a95fb8c8b0719fa1d57862b3cc17ba6f285303171719091a6c54d10044febb5deec37016b35848b0baac5abb90397070e73cefdccae6af6f60797b1c6339c5f3a7eabc3e4392756904", 0x6d}, {&(0x7f0000003600)="ba8eaba419a4", 0x6}, {&(0x7f0000003640)="88e1d3d29446c8f05bc4243cc4830b6cfe37e7adeb7f7e9632accc343bce7e9718108c1d1ee961b79cef34a11a2b9975763ae92ac0dae8f7d1d5bdf52341dff1e1841456146a7bccb55f11fde9d132c35e7b1b53132df80485303bf9", 0x5c}, {&(0x7f00000036c0)="49bae93ba5d20c0a0991838fc588f1d4fb4cc14361090baae6aa2efb05dc1aab2e40589b91447a8c", 0x28}, {&(0x7f0000003700)="e88b15a8c4b3b1678bfbbd16f1551a596c5f825ea1cf039ffe9ce97ab73f7bb61bf72172e0b33a657d72ef880f4226edc0be22169698c6650fc242f42378ec", 0x3f}], 0x5, &(0x7f00000037c0)=[@rthdr_2292={{0x48, 0x29, 0x39, {0x3b, 0x6, 0x0, 0x8, 0x0, [@remote, @dev={0xfe, 0x80, '\x00', 0x2d}, @private1={0xfc, 0x1, '\x00', 0x1}]}}}, @tclass={{0x14, 0x29, 0x43, 0xe4}}, @dstopts_2292={{0xe8, 0x29, 0x4, {0x2, 0x19, '\x00', [@calipso={0x7, 0x30, {0x0, 0xa, 0x40, 0xf801, [0x24, 0x6, 0xd1, 0x2, 0x4]}}, @padn={0x1, 0x2, [0x0, 0x0]}, @calipso={0x7, 0x20, {0x2, 0x6, 0x7, 0xfff, [0x100000001, 0xffffffffffff0001, 0x9]}}, @enc_lim={0x4, 0x1, 0x8}, @calipso={0x7, 0x8, {0x2, 0x0, 0x0, 0x7}}, @enc_lim={0x4, 0x1, 0x8}, @generic={0x2, 0x60, "349c1679b4ce86614ca52ca60ca97566d4fe36106af890642546b33bb70b4ade7ec08ac4da51a5e64c8499fca2db188e22a434c3a1f3da46bdcdde3ac8df60cecf9ab4819b2f2cc5c9fed84155b6a28fdee9f0e0edc3fc6aae39ede4746c157c"}]}}}, @flowinfo={{0x14, 0x29, 0xb, 0x7fff}}, @dontfrag={{0x14, 0x29, 0x3e, 0x1d85}}], 0x178}}, {{&(0x7f0000003940)={0xa, 0x4e23, 0x8, @remote, 0x5}, 0x1c, &(0x7f0000003ac0)=[{&(0x7f0000003980)="76e960", 0x3}, {&(0x7f00000039c0)="422516b8e217da13815499ac67e749e7b03eaa12772f9b926c562821027f65033ba56bc609dd38849c56ce1d1ceedda75420a578565014419affd534f79ec81dcd72e94a5fc781ec4825fd33696f27e15906a04e4fe0085d12e0c5535dcffb79b0e9f31a7b16fceb6ffc1220b54e2ddc9f50e923eba78c6baa659878861f9eecceefbd7d5386773051beba0266d0c51b82e45f9194d91f073bcd4d594887922d99905f35922b06908fb0b7e43556ccbf015734e236d22cc8272315400e56068482c89c74a1a7be8a2cb3bb8b52f0ad3da277e6f1e8f59bc698f1f86f8a539912b5c4b10d57e347f4eaa97e96ed5a", 0xee}], 0x2}}, {{&(0x7f0000003b00)={0xa, 0x4e23, 0x2, @empty, 0xa00}, 0x1c, &(0x7f0000003e40)=[{&(0x7f0000003b40)="1e6d0bd3542829a7a388a86a271c396f1458ce7e7ee2ae164e6c556b53147eb66f5b56b9c3a3fb1367d60980aaf6d86809909367bfdf6d54744926cd7868a2dbfe4d7c9a1df872cb6de5badde16c804c2e2cb5a0afc6e996056d361435a4160a8e41a148f097ca0a9fff52bbb4975d9e5490e7a84a05d15dfae07b9c081bd0f7204864d9066167b8d408e2dc95d73ade3d", 0x91}, {&(0x7f0000003c00)="975652", 0x3}, {&(0x7f0000003c40)="94375d5329768661d465af9a4de0772f688e3f3d7d72a18106a4f7e5c276575cf10a49e5946a6e6661640de5b5e2fd398b1a5c8c108dff60b8f8fc56b3094be528f85f9357fc163f23c50aa55d934372981bbd9999d1f50330eafce56cdf905cc42e564ddf0ba8f0122fa828d7105c1245dd38bb8b6c5a1d464c280bcab9868c798858ea19a31d4d45b3f2eaf82b28b7287544bdaa898f20a0f659c9cf90c0053fb82a3bacacda8a373aa5ac5961c7971d46f9526b18c55ec1f189601e88844ffc95f063c01af89614ff885f0594a87d9dd2a6a2fb3f", 0xd6}, {&(0x7f0000003d40)="064c2b3d35d1a8dd8a4050ed2bbb707ec54f9f2676d550308fdda31aa62b18e72cf7d4d3510333cf64c8defd3712a6dc12e4f15cf21b90b69c946f33eecae331d0df5f8440faf245657aaf398f9357d034809636ac9a42047e76f34a2de2f585ff51dcd4726050bf024bae4bcb8f66317e147e3b129ae0be47009b60a4555fc7b5f3a335bcc8d59d9fab4c4a50b52b59fcaafd3da0c1c62ef1f67d71f54bb74619b42af17385fdb90468646e47f8ac0df48aca82b1ab6e82144ff0dc1212a7a0df", 0xc1}], 0x4, &(0x7f0000004140)=[@rthdr_2292={{0x48, 0x29, 0x39, {0x47, 0x6, 0x2, 0xc8, 0x0, [@empty, @loopback, @mcast2]}}}, @rthdr_2292={{0x38, 0x29, 0x39, {0x84, 0x4, 0x2, 0xe, 0x0, [@private1={0xfc, 0x1, '\x00', 0x1}, @rand_addr=' \x01\x00\x00\x00\x00\x00\x00\x00\x00\x00\x00\x00\x00\x00\x01']}}}], 0x80}}, {{&(0x7f0000003f40)={0xa, 0x4e24, 0x8, @initdev={0xfe, 0x88, '\x00', 0x0, 0x0}, 0x1}, 0x1c, &(0x7f00000040c0)=[{&(0x7f0000003f80)="b8f7bc4ed73ea4bcf39bc0fc1ea8f87f7f18f3a7b30424699f", 0x19}, {&(0x7f0000003fc0)="11b1d0492a2b64c8468bd8ce87616052a4ec41ab7ac2306b2c2167aa6a233f4caef9fcd74d8e1f7aff1a3a52f3318c15df0c9946ea6b7546cf8b61b895a97a432c55da868c88387a6ac9d4664b7ce82e181b9b2a12e60629c476504084aceb3164dddc29bfecd2bccf50e26c7494fdb3d177d68be8e05bac1b54ff690deeac1018169179f16da8b0a50a4b3ed4ce96d367e1a8a6ecf351356dab84233bf21bd29b580803ff5d2b4f67778fb688c7e2e45a10e79df9cebff3ff12a988f795", 0xbe}, {&(0x7f0000004080)='l', 0x1}], 0x3}}, {{&(0x7f0000004100)={0xa, 0x4e22, 0x100, @dev={0xfe, 0x80, '\x00', 0x40}, 0x8000}, 0x1c, &(0x7f0000004300)=[{&(0x7f0000006080)="aee88b3a78a34f78d800aa4c9353c32ec59381f3897debfe1eba32de4317fb64984477b5993c6543f1249032530214916423c179c7ceebde6330defd244b8e3deaa652f271794da5df3ccc1cd7ee49ea0209ff9e3f4fb960c8c982bda5bcafb4ed15845c72c660350998dd4aa7f533d7bd9f2e158d4fba3159660c9fcbad8acf19dcca37f9a4d6d5807e9f3196de8c16", 0x90}, {&(0x7f0000004200)="a2f150e619190ed70131476b21b8cad5e1ac6c30d3db6e3e6c1264c4dc4481d4d92fba8886506eaf541b687d084b415218eb604fbae004c8b97147a55353b588854c12", 0x43}, {&(0x7f0000004280)="4a8d11d6518e58924585060db33d9afc7d939e3c69d0d27904ebb4c2a726a49617fdd09f0302eadde097c39841af2c849b1ea799dd5b15b9cd6506e4c24d85fff82c65e65142e9fea763eb743a6425d011de6687e9b7", 0x56}], 0x3, &(0x7f0000004340)=[@hopopts={{0x28, 0x29, 0x36, {0x33, 0x1, '\x00', [@padn={0x1, 0x7, [0x0, 0x0, 0x0, 0x0, 0x0, 0x0, 0x0]}, @padn={0x1, 0x1, [0x0]}]}}}, @dstopts={{0xc0, 0x29, 0x37, {0x67, 0x14, '\x00', [@jumbo, @jumbo={0xc2, 0x4, 0x4}, @generic={0x5, 0x78, "1bc18d4aec17f5a4590702e1bf57d7dde40cacfc88c5a0659e3c2fcde1bd37619730ff764056eefb783e5e03226775a0651e4193b9b9b18b0d5804914a48f759614e40f8c0cb4d2a93b6d9a5d9cd42bd2ba2da4aa524b60dcb8e558d2e11a56d0e25c57c56b16bc2ecea51647cfc3ea20321d402399dc76f"}, @enc_lim={0x4, 0x1, 0x1}, @hao={0xc9, 0x10, @rand_addr=' \x01\x00'}, @ra, @enc_lim={0x4, 0x1, 0x7f}]}}}], 0xe8}}, {{0x0, 0x0, &(0x7f00000046c0)=[{&(0x7f0000004440)="53fd5f89addc7638eef2a05a0c1d4a88a98dacb4cee3c24021b591be0de5fa732cea65b247aaeedd2e15056e1c64eb1ea06dbaffa2e05b72e51dcd7d7478d64ce2dd8675cdee3bb7be9a62fd1ec35bbab54adfd3eb1bc41d95ed3105ed985afb538d", 0x62}, {&(0x7f00000044c0)="d0fad7888f66de898467003e8ea2f49c9c1fcf33bfa040c794998df539aa496887cd9562b2e154c8e22a672e6608b86102d57345965673c6db21230672a9f11009a282c49c2f448c44817172b7ddfc1a4a43aae43db970a7062e7dc4dda23b747db9e3a477", 0x65}, {&(0x7f0000004540)="8c6ba3cc62c973bd7f7c6b118a846ccd432f5e5aa15960f6c4a8f09391898d5c357a3b0fd96874fa86ac6cba8965bb6411215106fdcd0b23652ab882c963a8bc4898f83a542fa7f4ce94410ecf9e0d8be0f9b9c716a01179aa274414fbe44b47943668228005ce2f50f53030460166bdd744f0e659de", 0x76}, {&(0x7f00000045c0)="3411fb6a69dff242cf21daf3b3d7042581803cc4a8163ab8bfd84e548749a3bd95c80849afd82b912e71f9232d095e50aeda6e4b27a78eb508031fe5287f50b255ec3539a4638e4d9492b95a1ffc8b528574f84e271bb8edb33b4595edf1703ff0f487ef62096233df247450c9af1df76f796d9ba8d84aee76e6a03aa5e86d01f2c8d381b5d53134511a94b982fb43765e783a7b54f09e2388f3696dc867b0ec4d5d2848ec1b9cccb96fb4489dd29f54c7cb580832efac9ec833f4c06576cdff18f6b4b1e2c70aa509e21c532860c1fd0a0385c5868f4a96923d1cf062fa354a5cc252b33cad33505d4f6b32d88dfcca0ea440539d", 0xf5}], 0x4, &(0x7f0000004700)=[@hoplimit_2292={{0x14, 0x29, 0x8, 0xfffffffe}}], 0x18}}, {{&(0x7f0000004740)={0xa, 0x4e21, 0x676, @mcast1, 0x8}, 0x1c, &(0x7f0000005980)=[{&(0x7f0000004780)="1e3a7c60b47a5e6f72dd1fb43385416966fb09552c66fdb67d9908bcbe25045a5c5ed574b34c2eddf074ba440c79d3a870d9e5ca60ab253b0a03b719dc649a766f40d0eceac609440f5b4b46df14a868852468d80d345c85e1b8f44fa1bd2f33b279ac2f7f17cef809d379315db2eeb475422e42848cd0258874c27c69cac246543f03ea83110cc12b1cd01dd3c88bb65a7126296a856092f47fbcb3f7bef0bfe25e3aaddc41d68c67874dbf332d103bead1bd7fc010aafd346e2189cd3822cec0bf8f9cdd7be0ae7b9257fd1bb53a5fa0e664207ab2b56d4dd71f09f062afb7f1c593ce51a4bba135cae0fa1de757a64a6f5fea440d2bfad001023a4f28e2450ef41e14919ec3b623e7fbfce1f6d5de774de1712f48bc6c3e11ebab952d3f097f060c82ff54aa12b261e44f51f58e27db622c04c5a97d5e24f1da771b516aba024fdc8abf94bd20a674e3aa5bab85756a2318d4126c474609b77286fe9fedf7bd0c6d0c96a3fc65cf91911fe0cf2f2956df830a7f2ef23d43307a07e665b50d999dc27ff4da79065f2dd750c9fd5d7ba0733b8e3d0985050f5338880a8604d04b486396695a79e78256881ee7d7e2650fac0d57a54cbb43280f6097adaa14b07590ba2b80b2d97d7ee94353be1f53b78842bbd3133cf31b2acf82f520a3a9a3cd694e140dc70fdeb63af3bbd6a9d263ff8e973223197014aa68b98d99e61b4b67afe52cea5c04dfabb3c80de0590bff993654e0129ca1497ca97646937434f2cfda4817bb8b0b0f1f7bbe1efcf27daa3b5794084c0df8eb654f49516be8659411cb1f8ec404f119a5f92fe1085a6291cdf37b5095f4ee4073e60d418600cea4afe684cc4e70dbba23286ca50c400853f2c6629a72935a7afed74e5045e52a8b872b533757ae0be2b2fd758781ba0e25e77e67222b191584091646b3b56474cdbc70602655464d95615b99179cb3ef8af5b42171cbc6e05ba076a52adb170a94dfbcdc8c6f7444c417a141be499b5e89f3a4d33fb851d861f14065ad1e8065a7f68b2d3ee420fd1f37fbed56beab590415f392b8996d5a035e72edea7ac769b5c1393c7374f9a6d89bf727edbba921061d252fa53d9459384dcd3c96450327cbbe28d3315cf24b1ddbd3a691fcced3d5d833fbc2c971f959e63f85ae81aba96e70cf26e5ad08af2eb5464a6738c28eaf0371b6639cefffc11aa8347f0aeb9f7086ec070dee59ab12286ee3ee385cde46f29057b4150bdd1acb0ed301d34b49885afe4861ee1da60cdce3279f45d30d05eeb43019e0d40c2fd48c7590d8ec751074c40927b6e83abbf8ecc1a15f3ad9def1d44573929c8d564cd3e298b689aaac9763c0b714c82bafed329685fa0b40607462c65bbb62b288f592a82723fea9f827b523c7193258388ec9912fab7bfb061b311a6a4869250d726f83f7dcc27ca54c788afc131355dc59e791d4c15d2c7961ce91e1c0303dcee7fc8ed497cf5a9696d750af581711d8229120fab599bab00e7afbe66d4f0e75ed5ec9c734a572a13e2747098ad850af93d9fa5c58530ea9fc070149ff1d1c59966fd3e44d540ef55fe89ea0c0a1cfaa74e2b514d916ea39a7dacd19a990a227bc9eb476626d4eae213aa7133f201afd968a2dcff350f685e0ee07956fe6090b5c26349387ab671ed00d43f41ed1edb228046e7eeb3a7e4203f0149f1ca529ada2052e48a22edf88c575a094e46d57b07fb979eed8bedd9c72f9fd334b6e422a873b82db6b2496c88bc8991d09c64464866b192ae2312391778fd74ee12fb6614fd57d46931f48b8777b576fb196fe87680b8620aacc4c71596de956c2f46ff653fe126debcbf20b353d5b015604f5546f29bbd770f6594d3b6a17090131780319883e3e7b1b40b13b3cb9df066b6e4571dab487738ea98795d0e0e176e5dcae2280c37523c209d0e3387b7643a5f5b850be5d8e8d6f5a2b1a8e8edb351ad91d78b179fb1701e2a9627a0c290cf4ad7efb6330b045364227e11cfd71843979003c1b39ae7a9cbe665406d4b49d1e14bf51a68d6d55165a6b908771dfe41d93075a5c84796a8bab042f0bfd02f8c7b7511c85f367a48074d1181d29b69fc18822bf53ba4c3a2247379bbb5aeb51f17a1e0daedbc8af7829847dbce17af116f518f1fcfcf2165742985580b68e371ceaae4e1077dc816cfe398cbceeea0c3eec4b5a4ba33daa5578e3d41a24fe1eeea45c29ea840c782c2b666efb991663088ceae71851ded61aef80035e31303b596907ccaab7eab977225be355a701492c78e0b5f4d1914aef34f6ff20cdf4f4ccc23695b35585d2503b6a9e741a3ba68073ddb5df93f55ee806f1036b3ee6f346f2fcae98ac2e94d2b1f0245e6079781a6a9adcc42b6456b7d7804e2925874ad4e80cda367791028945ba6e564fd3d2b977ab2f248bdbc2640cba532e273180b4c708ea10715d8f55c453defc24a7939bef74f0d0fc56d4f22f960cc03069d6b9be1eacec23ae9c31a82726a250c9d25afcd645414262731797fa92c0b0ff4bdfc81e1d607be72d949a45886920d23360c98cb47d58190b60f5c3af2d51d25763ef6c04d33a7cfc87f1ebddd14d706c61182b98acc6f0310ef4b72e0719a3dda86d135600423310f9ab6610daad135ff1bbb9dc4af14b09022f99d73bc4589748b3e1ec3c157c5d97608e63e97690e864c31635a7204150c45c70a8fa7670409bd5e30670b703ecb735415b1a743f186c4a0c84edd7274ad583ec2ddd3e67e5e760db74914c9ef3003e6c512f0b9c6d26830a8d9eac9b540b5929f3eacc9279c872c8698cb2184b1c8f2648e20a5c4c5ad13fba630c885e71ff26cdea26319c401aead72081418c161910b30932b4a250bafa3ae69b9e4b5a98fc2efa57eee8e7194d7dd2af9778ce822781039f951cdfe7237ba1d9e572d898ca51866b199aebb2e512a8e4fc9a80320ee0d9e3f093e086aec0bf20759e43bf8ca637bd9f61e78026da737475685b9371730f810bb9d2c15f0f52f2598b529200e7c669e42f4aa00ac9542672b8073cbc56c525d19684930e2d8f79bfd1d67b35a888e7947f8cc1ed00b9b4986d7c2aca6c4111da79cd85bcfb584e8d1ddb3838a08d4bb099827316b73cd7909f6c306642d4df87935ec815d04622968d27f63e8c1e2300692f5c19bee3237eeb8f76f89d0d267213eda627a8fc821100510e939286c9fda33e5680d86692a05a0f6223842e649f4b06dbf6af0285d8f86396683ad8917c898b5d9431b1779d28911507537ece29a86093487208ae71fc8488d655a9352049e4e7d134ecfe683a372c77c7762dfeb48960a2b60efeb8efc9163ece1042baa952fe329be36ebca6466ae52ff70190f95ebdc5671cd6e88cff552631230ecdea82f1dec08a391fed26981a6907999c45f9d75487226d32d3adcbebe9a7987d3486f198e1670b7547c3c15807f6f853f697b85999c7aae5a7623358d285e5ee82e7e0dbbd3e8b5f87c84148134da1aba1553b5060a887e193ce0607a2072c76cf4740b452e5c3802eeeb55a2703af0b642d4591c1de25734248001ef2147e660de8d0ad975f006ec836936aae07c6e169ed4bda6eaf8edc75defc8f00c2954a86abe25087fba178d837d0e7974575fea20c96eb32dd4dcca981ba08f11f1047f11a28fcd52521f98e8babd1b86006817a1a24a13540755dda395df961eda7ca3fce7337c3384f8ace09dac94801c5e50a634c38193682b9173f4398df86e6e97713ac3b9717b44eb60f27d72ad1b92aacd1a66c7f9970acf886f0eaa1d5203388f8dfe9fc6b1f07b206bbabc4cd982b5dba34b15801dcaab907506890c4955a3b76bd93cc0e3d12403bd3a66754909fc14447f02eaf9717b219c9280f7c32d1f5425f05a4871b40e9cf8a5dcdc7d8c3fb63556d44cca59556a9245ee20e982e98251a464b757793ef6c126c6003eef73393a5cc05e6ec1e49016f33f7f9b76064dd04021645402d49a6453e63df4a29a84fc1b397aa922f9fcf755c06ba45df2eff2554ce555b395b1f795816179224359842d1f95836224a9c1f3e58815daac0d19bfc08c8e1d06c3c5de227c68cf2d11bc5233a481048ed657114a2da7d4e7068f87eea91d4aa4a2c469616aeaf8e490a5c1e444689b0d33fe57d815ff1020f94a2da597e02abe324fdb760cfd4d5bce89a95f7fa230ed0c8448abc0688417e5cedd095c6be83f024ec7f6091558fb8f83b25b0a3f6ca12966105ca76217a22b222077a67cf853d4b33322fa81978cacd97e72712b9efedb4ba740565440c49ecfc37d54f42ed5818012b54db628ec5507fb30df55b1ff522c5e4f246f171df146f620a914673e68b4a838198a459b59f2e65ab9e27d3fc4fade7458f561a71e80313995a2ac01e55299c5fb5c6aa0c7391e633f3c99e815b3e6c6c947e9f0d0545d930fac6ff59de06cd451e63c7980bab09ac97a556ae69cb5ea4bd10129eb36b4d1edcc0e97fbd4d095a18d91f5a07532a5b385d7d2c043537b627f6a1a972073244653f173c7f119ddd88e16c1c3336d81bcfdbdce5411e5fa4079913a330205dce4785f218aa928d8b81ec93a5cb950606187fd5c234bdcc06467bcca26434d33d64f2add9396d8cd783a0ce4cc81fa2a1eb13c80c2e35d40750dca170a9b7935247196cf5a95f7e429d91a2c77e9e5439b2f9fbf4a98091f0ea359d0483e34cc14a089b5ac8c92c3982e5b3b526049d12a2a8f88cafb45a00b46a9a9b15a5d3ecc7de10a386ac7bae1d53517dc65647df3eef816a42b30cabf7e35ed027badc5caa87b80a93c89f705b8cb233b2d24abf1c181001c911fec6fc7ec7a49c0ec24fecc5eefcf3a72f3deee7b8615f6702f7b10b60b8f41a53c3a0f7071f805b2180b771b89da65d4c4f484c2b3aeb43018a7a888ab45cbd691f3b32a23db7303ea1fc9f48194223462530fa6a2f69535aa6de8f735789bf346708cf5dfb9e498097dac2ca0209cad47ab6d5c4d7ff837819c615c5edbfb7efebc885b062efa27920a396f7ed0ae695a3db25988a0b3fae8e7656ea343a341c85cadf0ffeb54e341af736ea1fe129e1234579ee9640a993efb8151a32bcce94a25afd0765d82e7f05e128e644c0e2e4021a2c346732fccde5a6b6ccc45dfbb1e0f32a9d3e7afe2b97212c48996b6e07e80d86a9af4fdbc7d2b7ca7cdbcf78bae662cd6cc6a314dddb0590ff9b4b4c92976b553466d201ba6176c0c13eaadbedbe940958a16573005d227229b5e2917a09894b047bb0678f154955f214dc12cae1cde859317fc573fb517ae168ba2902888202caaee9c1dcd546524eb1e4a0c7043543981e9ab6723666c80f30a4921a4d40282a644ec9a5b8f458efcf384080fc71587e37df31c7de59dd5245fe5c9577cd35a2e398c7be48fc807b380b698b9c3defc8c9aa690df614686e95c7c7bf9d725d85da93bb544c9acdf81eddad2f279e121c52ca14e6f8e165ee549f48a517432849d0857acf6a19aa35147995ea34dfbb9dde1ec64e14ff12ac7624dc3d4f7195f21bddce38de911d20f21afd5ef21eaa8a2d929b9d12da532445a5fcb13dcc158293cb40267f84fe1cae4fa64d470105f1490ca5ee8240866348ea25890b12e5861ff7ad533f4da3dc6c68ef3c71ece81b2b358cc33a9c026086dc7c82256e801b0c03ca849013617baa79b9f0ff0649a667bcab10c99a344f425913deb6ebeda5f8859dae2342c680044bf856ba444847aac070495c29adf6952c8490c9489c926a7ae0bb4166e392ae03677c", 0x1000}, {&(0x7f0000005780)="725a14d768bae84499bd5b528fbb97a71e321c050fec8587c0f854fb2fc86ccf58232313532f0f64d002e9b128e5fc62ea1bd6432dad0ea752ad0c8f346e9df38b3873fe9b720010973ffb0f39c5597746a91dcd55c46ae947e6aa4062072e8193a76911e930147b3c011bfb20a46b", 0x6f}, {&(0x7f0000005800)="d6268a5528f585e113399f0c3ad37b5bc14a4608a86a42dd9134a96774356c2345d98b95390d4bc0439c067d36af24011e2344756b963d48da70d749f796976a07fd5f7b0419de30fe9662e4f3da46be21e0a68a2012fdc7f70c505b553ab07968770d86895c73fa1d46d272bf448f0b6b85fbd0966b7b6395c501e00549c3f7114f8693b21cc1d3a7f6655eb1a2ce122feedb0dbb484b4dd806effed2741d5b1bf0adaf0d36f67e5e2356c4ff2fc5b9e88f3c76b495b8658b6e4870f19634a6ecc8cc4888fe3308436aad", 0xcb}, {&(0x7f0000005900)="9ac94d9d22868b60ec510aa10601a340d72400314797f37b48a3a986ac76b52241f0122bc3692dd97e0eae6aded464b35b2763be5d83fd0d9a19af4a23beb7875f0cdfbd1f6cbb1649252ab851e23584a400d590b55fb2dd7fbc727b68882311148bea93ed8aa7e583709d80eb771228ca01cf5601d5928b885845", 0x7b}, {&(0x7f0000005a80)="f482c32a06a41039cf295e832cb0d32ef9756d12ff47602b893dc14d6d445451604684bc1f57f94e0ceca45d20a751e558466387b8bfc4487472d8e452d137eec82175f984d589239512a29d4f2a1cd3decdaed1ddb47fee6e37f2c506c805f64879dbda072bba47e45d784f923cda42b2a5274b5797b37e3c3599035ee1298db83940f4173625f6f57f94608cd333f636d0e514aa09df886a1d0fefb88671af71159aecfb69eb2fdc8c39a4f53842431ed4545ef6221849f0edd1d30ade334aa4bdd67f3bb6ee6de8cd0f83cbd7301f21a1a335500d6a2d130dc456b45e26efd6c0bd17ff7003b66f4a1d7acbac41c10069ba31583f", 0xf6}], 0x5, &(0x7f0000005b80)=[@hoplimit_2292={{0x14, 0x29, 0x8, 0x1}}, @hoplimit={{0x14, 0x29, 0x34, 0x3}}, @dstopts={{0xa0, 0x29, 0x37, {0x87, 0x10, '\x00', [@calipso={0x7, 0x18, {0x2, 0x4, 0x7f, 0x2, [0x304, 0x4]}}, @calipso={0x7, 0x30, {0x0, 0xa, 0x5, 0x0, [0x600000000000, 0x2, 0x240000000000, 0x2, 0x1]}}, @padn={0x1, 0x3, [0x0, 0x0, 0x0]}, @enc_lim={0x4, 0x1, 0xc3}, @jumbo={0xc2, 0x4, 0xf26}, @hao={0xc9, 0x10, @rand_addr=' \x01\x00'}, @enc_lim={0x4, 0x1, 0x20}, @hao={0xc9, 0x10, @loopback}, @padn={0x1, 0x2, [0x0, 0x0]}]}}}], 0xd0}}], 0x8, 0x1) [ 2811.967189][ T2596] bond1277: (slave bridge1240): Enslaving as an active interface with an up link [ 2812.005904][ T2603] validate_nla: 3 callbacks suppressed [ 2812.005923][ T2603] netlink: 'syz-executor.2': attribute type 1 has an invalid length. 04:12:19 executing program 4: r0 = socket$inet6(0xa, 0x800, 0x80) ioctl$sock_SIOCETHTOOL(r0, 0x89f0, &(0x7f0000000000)={'bridge0\x00', &(0x7f0000000200)=ANY=[@ANYBLOB="10000000008000"/36]}) (async) connect$inet6(r0, &(0x7f0000000040)={0xa, 0x4e23, 0x2, @local, 0x3}, 0x1c) r1 = socket$inet_sctp(0x2, 0x5, 0x84) (async) socket$inet_sctp(0x2, 0x1, 0x84) (async) r2 = socket$inet(0x2, 0x4000000000080001, 0x0) setsockopt$IPT_SO_SET_REPLACE(r2, 0x0, 0x40, &(0x7f00000008c0)=ANY=[@ANYBLOB="7261770000000000000000000000000000000000000096dd89ad65dec810000108f80000030000003003000098020000000000000000000000000000000000009802000098020000980200009802000098020000030080000000000000000000ffffffffe00000010000000000000000e4000000010000000000bd00000000007465616d5f736c6176655f31000000000000000000002000000000000000000000000000000000061e2695eddaca41000000000000000000c001080200000000000000000000000000000000000000005001686173686c696d6974000000000000000000000000000000000000000002726f736530000000000000000000000000000000000000000000000000000000000000e4ff080000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000090000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000bcef5a18d4a047d60000000000000000000000070000000000000000000000000000000000090000000000000000000000000000000000000000000000000000000000006c0000000000000000210002000000000000000000000000000000000000000000fcffffffffffffff00000000000000000000000003000000070000000000000000000000000000004800435400000000000002000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000001000000000000000000000000000000000000000000000000000000000000000000000000000009000000000000000000000000000800200000000000000000000000000400000000000000000000000000000070009000000000094100000000000000000000000000000020004e4f545241434b0000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000007000980000000000000000000000001000000000000000002800000000000000004000000000000000000000000000000000000000000000feffffff"], 0x1) (async) getsockopt$inet_sctp_SCTP_MAX_BURST(r2, 0x84, 0xd, &(0x7f00000001c0)=@assoc_value={0x0}, &(0x7f0000000180)=0x8) setsockopt$inet_sctp_SCTP_ASSOCINFO(r1, 0x84, 0x1, &(0x7f0000000100)={r3}, 0x14) (async) setsockopt$inet_sctp6_SCTP_PEER_ADDR_PARAMS(r0, 0x84, 0x9, &(0x7f00000000c0)={r3, @in={{0x2, 0x4e22, @remote}}, 0x9, 0x800, 0x4fad7375, 0x6, 0x2, 0x3, 0xff}, 0x9c) (async) r4 = socket(0x1a, 0x4, 0x3) setsockopt$inet6_tcp_TCP_CONGESTION(r4, 0x6, 0xd, &(0x7f0000000080)='nv\x00', 0x3) 04:12:19 executing program 0: socket$inet6(0xa, 0x6, 0x0) (async) r0 = socket$inet6(0xa, 0x6, 0x0) bind$inet6(r0, &(0x7f0000000000)={0xa, 0x4e20}, 0x1c) (async) bind$inet6(r0, &(0x7f0000000000)={0xa, 0x4e20}, 0x1c) listen(r0, 0x6) (async) listen(r0, 0x6) r1 = socket$inet6(0xa, 0x6, 0x0) r2 = socket(0x3, 0x6, 0xdd) connect$inet6(r2, &(0x7f0000003ec0)={0xa, 0x4e21, 0x2, @private2={0xfc, 0x2, '\x00', 0x1}, 0xf5}, 0x1c) writev(r1, &(0x7f0000006040)=[{&(0x7f0000005ec0)="d1b56ae7f5a88544248e70b5dca818f965d20b97c74cfd4172e48d3ded3298ccdea9535c41534e24e1d864c34808ff776c1cd552ed78716e9bbb6bf0d27294f68eff5ecff7bb60c92faee5be484608f732b8eec6007722f7e3415acd5d1ca98631f2fe7b988fd5edcd54ce33f3703d8323ebe93cee7725955562422d5870fdb44f787a342a0faef293a775a6c012a403e715888dcee99eced99b", 0x9a}, {&(0x7f0000005f80)="83359419dec84b68e69e37b05f7e18d8acd2c6bb670b00ecb72010b2498a9220061d20b69af76f5207e108851b378b1713afc8dec2b86b2e26a883a0b5b5129b4b63be3048b5783de84a06e1eac86852959bdcbbc29deb00014e8b504b", 0x5d}, {&(0x7f0000006000)="cdc05fe9b555a26673c510b247d6c0c8837484d2c47b70affa9fcdbeff6978e197f0fc56de229dd3", 0x28}], 0x3) connect$inet6(0xffffffffffffffff, &(0x7f0000005e80)={0xa, 0x4e20, 0x10001, @rand_addr=' \x01\x00\x00\x00\x00\x00\x00\x00\x00\x00\x00\x00\x00\x00\x01', 0x7f}, 0x1c) (async) connect$inet6(0xffffffffffffffff, &(0x7f0000005e80)={0xa, 0x4e20, 0x10001, @rand_addr=' \x01\x00\x00\x00\x00\x00\x00\x00\x00\x00\x00\x00\x00\x00\x01', 0x7f}, 0x1c) sendto$inet6(0xffffffffffffffff, &(0x7f0000006140)="f6007c6225216fa83f80d015f9b2fafe139eee1756686ad1dfca9ae0273476e4836b5ce4ff42ba41cb389f50f075659a7402c60859cbd979aaa4d40aeae2dd2651326fb4141e974637f898923ac39e10c214ce2f9350fe84b351be7416c4c8144e0490533a0faa9e3894cdb094b6dca983e50c7a4bf10097de809490d15ec32495480b5f66ca85234600e3dc23193f4c76c10bb53a0f16e47d564b94d0681c4594f9153a82daee30500a7e4bbd877fbe17de30af9e6270df83296c736587c57a8f7675f2151b30d4ebdbc76b99795afcd32c578955758273353d178944082376f0d0ae4dc7d255d0c6b143ba8d93db9454ae74b802", 0xf5, 0x40810, 0x0, 0x0) connect$inet6(0xffffffffffffffff, &(0x7f0000000200)={0xa, 0x4e22, 0x200fff, @remote, 0x2}, 0x1c) connect$inet6(r1, &(0x7f0000000040)={0xa, 0x4e20, 0x0, @dev={0xfe, 0x80, '\x00', 0xf}}, 0x1c) sendmmsg$inet6(r1, &(0x7f0000005a00)=[{{0x0, 0x0, 0x0}}, {{0x0, 0x0, 0x0, 0x0, &(0x7f0000000080)=ANY=[@ANYBLOB="5000000000000000839a6de73600000000060000000000000401000730000000000a000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000002000000000000000810000003700000000000000000000000401000000000000140000000000000029000000000000016000000000000000290000003721c218959b09a462fa9c0e1c6feadbfda83ee9a1785d09930585374e38ebc2f37a461f"], 0x100}}], 0x2, 0x0) r3 = socket$inet6_udp(0xa, 0x2, 0x0) ioctl$ifreq_SIOCGIFINDEX_team(r0, 0x8933, &(0x7f0000003e80)) sendmmsg$inet6(r3, &(0x7f0000005c80)=[{{&(0x7f0000000140)={0xa, 0x4e22, 0x7fff, @rand_addr=' \x01\x00', 0x7ff}, 0x1c, &(0x7f0000003480)=[{&(0x7f0000000180)="c22de8de49c8bc74ed6a27990b96c7bbffde6e451106cd7db38c6cfe57439b6ccfd0b713f70ae54f02270b4776d47ca2288261af3a5cc7029af46bf8119dc3311516af33ab353064c342299bac4eeb1392f74f4ffed1ad7addabaf1fa8010cf13fadac6c44bf6d2b", 0x68}, {&(0x7f0000000240)="c97c8d45fc51ccbd284a494f7fa337de3064db9bf3406abe667335fb85ab0ed822756db072b12fc6d77789ab7c1acd7eb42208104b970485ba497cf158a0ca4f98eea848531b15b3ea19cf906671947387cf789ce725ed5b8e20db53c4ba19fe700a12e3d8569518995b863575644a44f0956259c90c98c6add1eb0ecee40d799b20c5bc7186300cb062b46915a2e5b13af446e725cc02114689646ac7290dd637dd1392c1bf5189299197b6ac5106256696f0ff163a3c739be5", 0xba}, {&(0x7f0000000300)="16bead57a423edd4b1cca1b2504dcb0a38fdc890f65b7636d8daea0c", 0x1c}, {&(0x7f0000000340)="51522000b4abc08f282fa6a69ceecb5e211d38939472ce2f2f5337ba56a8c81a5e9b51d5413e3444980182b0fb8b08782b71830fc891213a280a354245d36cac7d7d6185d124a6f1eccf9bed0c8a9423da237af98bdf4a9190affe49352eac845cb25a71680a1157516364c90f75a0a7f544e8512912b91163aeb614e3b677d318163e3fdf33e3e17f9c45f0364acd3c39ce5f8a204553ee4e6ea7", 0x9b}, {&(0x7f0000000400)="103ddb76f5c3a6822e67decd0875a1b06ef2eb440a6503c0d920b69baf9137df4c1f3dbee66104c86d6908e8e497ac4f2ec7656337294621b9113bffa9f202879a2bbc3188545e5158e3795c5a6ce1382cf68b8965d2507615a23c22580813718c8c5bd7d0cb1e8cf01865e19026a5c2c724772a543d1fccdb2952561885ef64ab3f4c8c45d19e2a39892f0dc27dcb5c7546cdeb07502d85daa11369b995a80571577a527ccb869d0dc362aa6ce57da19b6e4671e95a966fe8d707037ecf9942bbf8e61d4c16a5d9cb5e2388ff4d4707dc266207627a1969170396ff0fd00ecfe2cb616dffb2ceb658b5622f6cdf5f2d06f6764ff46ea4bbced621e95aab636b4653e848a781ff8df5c69f7d357f4145628c57a9f4aae6af9279ffe252c3be1ea2c693b693a353ab092d3ccc4e99ab076fb290d2326e317d65548ff64373b888a2f2817f62a73458059168d118424d88bc0f8f067678eac0ab22ef1123083dcd46fbda980201f2163286286d2e086ac7daa12e40880bc0647415e568f2d7af092c8f4d85008b1a335f9ce8a3df4216861b579152b1ab678e294fc46c815d638ce4925c8b985ff55072ae52df0a6be6f6d17c8f781a0dc1e114449c3d947814fe3decc167719d0556ff1007aea91abd64a1dd151eb0e1887409fab15a4355f610213bc2b3dca4b94daeb00419af0a5c16470d3ce57dc844eff080c09772b1b2ccda8373d8d4c6934aadbba6bdd76735f78f2076624132614a1f86769ec76aaeeecabf3fb21a5db4f807e4dd64ac62385ec5970b7c07bcb50cbd6cdbf2e0325022db87c9cf1212a1145ee4f0d4e67abd86c5dcc1e00982ed884654ff0f9cd4fa5dc6988685c91dfa182789eff886c4b9abc0323fe4fd05c1cc37e922d7ff66e8b7e7186015a28dd32dce7cfdca009f27ddd091f385af46add3449a9c9195e1c9b0b2d78919a60ffdc072a3ac1ed862968f83701bc25ee9d60af2afd1aae81edd72822d536044399ab2fa79252c633a3adc23b51616a4906ae17f51662033997f68a4ed695bf04e0cc1fd380873d70fb26c56ee26ca04092993eec119beaf5810ce6dc260b9b0036551b3532b2c7def534d3a21f55d326a357b22b5be248c70aaf1f0753dc379e70c2ca8420e75eb0f718a152cc7a12d34c7cbfb9de3ec35ba311b2a26050b2fd6730ec21f9c48d13f60561951d6be3eb6cd6feb7fce4b8049e504d9ad2469620910e24fe6d2358fdd195dea1e427b2fb0cd8d91276c42bb45e25b7253e8d38851b1cbfa73b67070aa6d576a78eb9283d30b38825e73413db827cad05211552ce49fb364ac6cb5d4e26a86c041dd009290cfa8e261b5cf26162fdc8a5d091dc351ff09911b1cffb352e6b814b16edc6b059e0ec7e0e218041343676c703a17a2bec3c97d68f44562491f9ac0353f1a0657dd3511c2a489d13a9b97feb37abf2af95a9c4241a12429954ee24afd3b7ac672e479954f90c0a591ee903afc8866fa38c1358b2963683e49d54bce0594eda4015b1526599d51cbe21f93c5d7e0a446e782767bffb793f691d07929670c82d373c6648e82b70ef9840b3ada8d2aab84d1017b2a913da6b3b64175245010659bc3532f69c58599a03dcce41c29d05dd8e3a8d4716e8b19a5f4a3a8f37ed0f0903e5d6e38d66596a3c1cd48df3347fda424e7691d733359bbfb3feda692331ad81a6ce60764bc0889ce40301f46c6a71d8da8f732d25a3c94dfc1ef0eca420723851e37f3b3f21b659c09cdd174c610693f326d8bc5aae2b72bf114f95ff5e95d509208641afa8a262cfad171494ee62904174fd9bde07dec67ca1f071975d8a800034d4c13ec8a420708d10f9d5d7fb46124dff3bb7cf4fae06e114550122cf764194186f24a9701bdb279383760843cb0c3594d2fcd74607e4e7681eca1e4ae6f7a703795bd66eb07603e3b8920bd8538c479b8b9af34fe4cd2465bcfea5c15ba4c2330e9b6c9165da731eb6213894da8effb72773b70a812c7e110e4cbfe91a0afa372545bc0d3d60fcdcff889ffc6034c208ed41e37d60418e69ab6f285b47294d3e9ecbfb3a2a100189acafb9523e8fcbbcdd99a4be6a32b2a4e62c1f657eb1a3a1e9f151681773b3443f79dca58d45118aaa92463c6d3f47345bc074ae37477b4cbfe06e89bdff5c2660ab422cefd03f7b4d0cfce6c90b2ba134f4ae176f9acfab18e8d06931c44557d10373929a80a95696913a0dc29c01c827d79c984526fecbfd6d480083d47db817f68b5ebbb2e97a97b37ace1215872ae775267dc1ae6d41bf2b99faffc154c2afcabcbde090cbdb2d9b9e7fc3ae9aa3c7603e72f6ab25ac9487954c6ecdd0c43e336e495d7f0d84d9452f413eca875377f4fb78ba9bd16550fc90752c8a58db8f08811e513162ba329e75152a4503834db7b5d728c68daa245e318bbe23ef274c4617756b5edbf6527927269e4162468aa3097587718668152733a88b28748541fbf9ad9ffbf8ba9a9b8ba0a251f2eb214e72d08b64389ef1040c825931bd3a78cb9ac6fcbde9528fe57c2e4dacc9b5499258c3c3ae40a5c70d8599cadc1d200d9f7686e1ff0dfcc0bcb22e6550c531fbf185abf47dffd485b85783ce6f79191498de308b7bffbb16ab88bacf952413a7669c5057dc392b1a4055ed8dbf3cabcfd5e7c6bf46abf92fc6fa5077a1b1c669fd554cce25659d09e63fcef040cf8489af6b7485449a616453d733435fd225d56102e49af1977ca972b54b2c566e1ad98048c684e020b4523842a85a13f27795ad4378160844ec06190d358b2e38bc155980095370620945279e1648e299ac25e217731a3d81da91bf0a9cf111601877c347d3f48728bddbdc31c4c5093972c1f1d513a89a332e0046a8c911af2ee273c14ad3155d6d15e0d4816cc3d7ab58999126e1e039fc882e7b1fa63e4f366664f2f487538f6b1e4f6bcbffdc7d70f34c2cf4f53885a0e7ee1224c992189504abc8a702e10608aa66dbf4b4b305a3c4d99937058806c8af9b853546ba1bbb84bef9937c308757863817272ef8b95061a07dfd57518ee1c4e200303c24bcf982d9dd9fffcd677bfc2e33e4026af218e76c2637b0ab840097ceac645b17c107fa2c50bfbe97846c3f202430c750b3815590000e0a74e231efeffc234269aed69232426b2408e1d362879498c0e425d1f02c53cc86a4892cf680ea63fce419a6c7df843b64f217b25b1fb10049678874d81d733bf27fc5ec89aa7c416e4f2dd4c88b530036551b4465cc756bc31e15327caa299863a99307916412503f90a30ee79eb1068230bc304ef145a1bdf7a095c1196e687bf8e9f780358314571840f2b2ef35ca94f131404b8379b2d71b9503c6431ea932771994f9be635e5df670302c90745023c37169489517b372f373d0f52717cd572a713d1e93cda0e1fb05e96e0116f4df465f7bb271943e958ce9fee5845f79847dcd4c01b485c7b271d40b2d704a66b32c9b3f2733576733e03bd4af60c8b1079ad0af867f8601ce869d5f512b77c6bd4110eb61b6da4caae42b75309f978f39596e7eed58c7638901a3aa6b9f97d851cb4efa915a682db74cb64f8c28e268d65689846bcdd4fc2cafc5934e686639c0e2a08a4eb1fd14ab7231956a75decb263e99f74a7ec9baf6e146c157d63a32fbe40b9c4cb8f238a9122a5f34d9b62420cfb296c20e0be292cf57c90d8461425cf32197d35c7747d95051c3c4b4c459733f245f40922c32d4764a8777878df1be8fd42f85de3fd01225a1d8d9cdc756306dbff987c3b4bde1c229ea45f81b9f20e6958f8c8a6fbd7d6f7bf9259b060c2d64f74769c4a66dd0be11ee9abe3b625a9a7677a2ddb94a8ab825a8e4b6b75ffeb9f54acabc68ae115fabc093a8f444e19ee266d7ad96a4fa87e70effcf18cf184c32d5dd71e7067157ddbe5f0e2f875ca0c56aad214cc9ce8b1a8b1681c74873d5dbe6c93dceb3458854b07bece891005ac2b8b4800eac5ce0211b234e6670bea324806a3ae54f8e4a48c3d7bcd102c408120c7eef179a9f2cc336fdab92b0e78511022d371cff1cc2c8f23eb9051b1648821db6830e454a2b343fe670c063c222efab424fadd58707352a4f17c4faae8e738a1bc60def40e9b8c731a11ff1ab41876593a82e7f60c53499778ad44986fc9dfefadd6562b59257faf3b45097980ade15c50f9eb6d5ce622ee86b55aafd5ea532cad3877d8965b1d831b9577dd9743f7026163c07410af842496916db1360133ddfd26feb918530a0f7ab664b6f62dbf10aa55e78a00061451006b3f81577c7e55771a08c835b56c332ae77b3f529ecc72f22f8ff4b9296480b50f88a751f05032a109502a3715d0411349de02be10a59284a91dd22fa3f8246ab1513c1cf9e964734878023a9cf89536f4c0a5c3286ff654642a0ac42d72ed68583f48c267fa26c8554d44bd312552b1121db719fdceddb9b7f3ae50fbe82984cba814662ad3e6d3f644947c89caf902de2eb394885ce08e580f4d087bb009d6bd41be5b6ddd41e95d0d7919b2469632dbef5194983500cf6b008320d6b657f6506bd9ac992a1723c477807379226e3acb4d8da21adf5c4e7065bf3d3fd395fbfab060dbda3b54e65f64e3ce2d3d20000bf4748c369ab01849051adcb6f7699034a872204e2588b6a52dbe81c0996d01612e15d0cbf21a9fddf714aa549e1717e10d24e7b919fa4e87d83b7613b7b0ec2a0e7446a59dfbf82beb6524f41f0ad05898f8f0e47e1f6beef9c5519f92f516f6a2d6ff4a56c083e4af8e801681e8782e698ab11ad3410a210112bc6cdedb09dddf7ca681d38e03b38aa5c90838e3203a232364458917dda37dc1b161ae1d0febd69347af30769528ee293909c645c6120a6f34b5504bd20701011a3cea8bc301cbb90f6b8fb1b3b94fd9a6a9d53227848d54d9b5749d30b3bfa23763d41e64fc8c8e07a74d4e12001de7f4c51e14ea8a763575dabc9cbc2d65745ac9b2eda7fa146579513375d7ce55713618c4b5cd44a4633d6c2647866f745e405b6632af9aae292001ab7581a11289a4a4ae1f235c11040024f89f5317b58e6fc7d7b9ac8ebea567debc7a5d0efbe8eafb374d4159db45767a829a22f7e503b82e0399c1350e8f25b0f428745379f9cc66f093615577fa527d9fcddb8cade543ec4aacd0f46ebb297e8cc6709915ba3b21ed93ea9284306cf6b3759853ddc1bb994a4b2dff3de13309d70dc0fa76ac64aa8618ce6127f38d3013a4bf0d2a29c6319ffd3987b7ce98a3aba4affc12c52e8c2953e3024a8e200992580601fd0c0acbfb8645453ca752a7932d552462109d208d58ea0c10f2b9da8d214456959a71bebe20f2193d0abed2bb4af4cc7b6def0ad288abe9b96fd256931dd2e4476f237daff691bd9539fc736cf5b286f39535f7b35c821b745ac24a19cf3af9228603c713e9eb70eab31fe0804710b9d6f2ce435dd2b45dee997fa333a3d9c2f2ca15271c385936ce687b102a9be42711c1baaf9794ce69a82fde4a39aacda111fbb965d4f2f58f6b48c544983fabd3a05839e204618dde1855754278da4d48d48ccef3d1d9bef906cbd60af2aca183b802b6daa41ba3c6f72ec6b12942fcb9c0c74cf79f9c0e7137273de4f196896f9cab3f471a182ca31268e86be5b5c5b00ece320d683330c37ef00ff498950b91a6c71442af1bc16868a0ecdbbe71d0d2e513724c3e9c44c0a058a1e95122493380cea437048532fd8871c00e1d8e52f6ce95aada96968a76a2c8db620747d622bded19f9afd6", 0x1000}, {&(0x7f0000001400)="6fee70109ca972282af8091517580ea97c02", 0x12}, {&(0x7f0000001440)="cf58e1164b21dbfd7361a6608edb7c3d74752d30bc803105cdc9c267ebb3180a712f3329e2daddea7dfe9a", 0x2b}, {&(0x7f0000001480)="57b10b5886521c21e2472f7c92c83189c0bb4ce02bb5a06885f74e06936c6cbee4ec566161101f109bb1ef4c83803331bdd5e94831bf88064113339a3e7c4658a414a386877130b9746550b2551cdfbbd99d9675b28d2d3c8b00f1866b89a4dcb1cc223754e09889f4eb0c9a5f3be77531564fda6b4aef20a152c5b5f3ebe7c26930e9e95361615e8edd1aff4a9caac3e87e6c467f4abdf8640dfe6cb42d89b48e1130745d47ff43353e0de9cb20156bb2cd8a64bbfd22d32d7998d405c3d26eba0e48d42ec8b9147ff41cb6c56416819d66c00c82c64b8662ff23690062578f31de8265f252d50a42ce16b16c64324a26700af0e31c3b8f09dcee863895be7fd1ebeeb89f57a691d06ea81337977e41970a37feedbb49cf7e6462a3d755124a58fab46e662e7db352706838b9b24ced05f4a52c542f537de34f0fdee6032367d49697db1247e7a43e9a4a0102ba8d71f4893bb5474121c84e43cff9e233671866a73fa7716cf2012e80905dce31ce3d1c5d9fc38f8ec29ceb82c924efd47d5f4f7f9f24673a3dc1fa5f4bcc81caf5d96e0ab3a2eb093c057a32a3122d104dd306e4382c0ecb425d94737e3693f1117a4b1780cd1f9d8f1a5b61e11f52b48cfa21777e9833ee0a8a065e3d17c514b76a04c69fef9b430b6830415d8557b23fd5f26e39a620f4128923a6c0c4c2a24ca8b27b110333968ad99075cc9f222d724fd87866d9858df6d1259452914f14f8c6c7a70daca6b85055684b8dba8225078fde417c6029e6c45fe7d6730c1fd965ab7628539585b25059b1c5d20e98478ec92b796ad25d10b08483b118d146d2b73c7145c6cb05f521c48fbb98e50dd75ce522ffe2c2076463030f7dbf61c113bd27d3b2886a5d12d578f97d8623c93befb22972be01d3af9f5d4edbcf4da523053a3539c8d0e4ef407ed52ef98a49c7229aed12903a141cc2b16acae5f92e0dcc27719c5aebb93b08cd5d86f9eac038cec3476e4ed5adb727830567fb405b3929be507152e9e2230facddcc133103acd7209aedb0f251fda299454b40a721017627545eb5b8f0deec7797211375f14b57dbb9e3f06f4a33dabf179d5e98c03798251d62dc22d6a760244bbe6858542183b2d40f7c3c755fcdd04a8502521e98a7a5f59e4dff26746d275e0b40746e76699789f88e45e85abc9327c7773f73cb91b43a35694def999fb960723851660452882860fe6623ab87f13337cecf7c8b18f3ef5afed07d5134e168722097b7d45e16228821ef99dac527e961b63fd61628f4a7cf9f1231262c83ad5db9b550c98adad1cd15db17cb947f192030fb44d638172329c7d87c35e510f439c3c72d952632ceb044cc57ac3d374dede3c81c8e75ea20e731de636f541205cf8ceb14d998793eef95ba2320e5090b299894fbf885462a4d909cfb8d021ccce95d6eb86ed47c0c8c8bc384333b81ea1557a66603c5a2f23ab0100cd94e3521f9bdb14c4c32ba9d9bfb1011e132678289d52e6c0c8c877974fe187f80c7fd11a80a59055d6d9a132f456df339fe90cd1fcf0216bb1f1c139bfcc1dc72334d08f467996ce61b692da00eefdae66bc54db30388665eca192f360ca41e7401ef3dc0f150f5831bf6356cba27b642b01bb8dc7e3f711c80d43e751ea7a1ea3376edbc704f8765df546e11a2fd23e1abb034c62eff880806fc2098981ad0388f233bb1053dd2a0bb5b1605ebce3cdf39f58db49b85dbeabdf9e4561ba2de2d2436b4d411ad040b419f5ddfe8058d02cda2edd51c320f1a14ceb206bf79aa649fbb3804b3e15c1e937358bb88a4b8be3e6de6b2e28c7c84964619677745ff8c009807e40017df35c60c11410f5dc7335d539d1c3918ea760b822bef85de33208d0f3395bad8db442c26752a675015211a0d49524f2b5b7b85c1f9c8e7364be0b095b2e194c04d6d96d9ff030a46f5b6324bea678d27d1050b38cb09c3e5973b0ab783b60b58a9ea119f28ac728203eb444b743f130fe99d291195836be19c24690f26f850edb5dbb409cdccbb06015e7d5b3e64fbe31a18191989d85f9196c39b0c226d189504da902f5c1efebcc5582ac35b098181c9ea94aab515d5691f71c0a267ae4b52593901e0820645b57da3540e986b14e8c32d00e86efcf1192cba857bffa78b6dfb99f5e69bc7e1d7453ee4b8db6800f7b89535b79d1f9304d99dcf33a7aa7a1022a2b664f65e33cff13cc34e86ebad1ca6551fe7f3431f72cd1c4563232a084ef283ef4972c1aaa8fe182a9838cc2ab5d1608a577406c663dd183ee6f6a6c1089a30f8924e858c6c6987619aab0987fbc413108913b013162043ec5344a818d97403bfa8d3c61ad5c6c9a28c4059e5786435718b1327c4f202cbe956f37f99186d57373c40daf05f1a9bb4372ff5fb66a9d5f7c7a5c98e987835ab8650a5fd68c9368146ba98c83a780288e9bf26f57ff4cef3c06569c21d30c64767834c5a7705e4e0c916703c8cabd08c787ee5edbd6fd90d3f56a0ade79fb11a41ca08ba9b4da3816ffa10fa14fbd76aba422ea613760295157e33e2b20fd4621f6ae1a6d39873f04cbc88eed0e313935de7432f55fbe09b9bbb110011fc167378229898b128e7f47dc6dc48ad759bbe3338405c17690651a14f313b540d9dd3639decad86d364d4e313fdaa7e221c862599d9333d8d307611f9c5ca8bf96920697eac2167308a1f22cb220edb497d978a6f7dc0034f7f85eb9858a44df554d8d622fa31c0bd915ce7b57b77d9141cd832a943aaf593f219310528103c3fcb316618c3b649d43faef7cbaf0f844eab8095edddcdcf6dd56e4d507e7d5f7582b90c9fcb7e9f69066e1c6cd8ebc6097412bca7965c44142ef5764c8252d8b015dfb270ced3a39824e3da0c11a6af451d0c59a6c5ecd5b367bd0b4859a291d896e38c285ac3abee0631f2337df0a70a3997bd28ec626fbd2c1ea66192f20e41a4b5fa0c3395898efcc01da101c77eb551240be337805cde29cf637ab7f57467fa16bddac3ae58b15eeced6931098093f155afcb9ee3661b081b7102a701841cf14f2dcf13e955df44956d702d95bbc8bcb8a6542cc8c8bc50711365eb95dbf78fe0bab478ccb2873fc8f8ad4201a36620ed11a24caf032147e8d1ed7bde55f15203ced96a4c45258b8467b691947ac4a165dd70a0bf7a5ce5c9c1a7dfc62fc83bfec43d05b741d22d7d42146b534428180ff9658544e3d5da92824430caee91f2cb063c41afc99b4c6b36503de16d74654d0ea86c357c788fec88a39baa1fe4f0a35dffa42db60b35ae78a310931b93607fa3b21ed9b87ae1c6965bf69a206e52a11b0a836368549ef535179d1f9edece1c33e0f9f005629b3c16be4fe8662fd093ee149ed2ce005f98c97474ccf04cd1b558eb81d6245a47736a214be31855f610c09071717bda0b54d8faf177e597af4f3770ff68e4357bdf507f713f92d23b35db3eedddf1afb1cd37efae980a77b5c218fb118f36b9d8e0182ae8970f403a838bfd50d38368f52bfcce98fa007278779da80da7d9769c9d7e2d801c8de4e40ae4899207ef057c0794e4b6ba10ad5fc56e42fe0807c05819169bb9b8499901ee31f49745017532c248fdbc0ade0a6a25744ae73643bbb47e715e617a2138a5cc66d25eb50abc2687b738b31ff9cb6586903acc739b8c7983fff6216f089d4d6d1f6f31bd42bee8fe326f47bca495810e48b266c778974a1442b86976d0b3be808d02b39d88294676482865fe2d833c2a39b401738635bd6ae0231705a487305dddb3af2cc42a8b39a7e811f3ed52edb439acf3e12fa8f8b2b66bed00b9894caa5b6a0688f7e0f699815146eff5e3f6471812bf49bc1a7f1447efe523ff058d739cdba9694094a5c64a4933682a0096e94f29e1b6ee8b84672da3cf1c17f2c2a546b09163861d0b090eabc23ce54a9c8a5f10525ac44076906d7f7991b2a08d1a7ae60fb18f468950c6589d5c1f698612917ea49b66495d03b0c3db11fc300e39ffde962c5bcf401c3502f2db3ec29e823c0691d911ef85e82f97ffe3e834c92ac496f20e46c809bc976a49669721ad46d3a4b315931d516a426339354e9760874b8a612f63c50380657f99d7eb3501cc29f9372f161816ce1ae4d285e31e279929dee2b87ebce4360ef0b632990506d310fb8c67a6a6f122959d8331623ff308bebd105b36ce9d2ec9538a8bde87a5808a555b0ceb931c0c57d139ac1fc576e55cb7e58d4a1cb24889bbbcd670d0bae65f1eb2cdc4870810e7962c4bbe21115779c19a5d90fae05c621f1bb5b4c741f9164619759802fd5c887fea540f218077c94cee430a750dae7f1c1784096862a5158b75456b6827a27aaef0ac9d59ed1210e25127035c745ad148bedb8474a7abe10c866fa3f86cbe49b51f80bf22cd830ccf0e81de8ef71acce5520f9047fa376dc39ff9ce9ee11f57512392ff9910df3b6be9b58ac5df9968fe6d1840b9bcd4d132c0a4eb174474dfc6f9ba0f9001679f87af42eb3ca0ce07cd04cb5049a4b8d4f6517d44f83fe32c2f5fbfd0e673264342b9ce268fa42392dc5a27c7e34e3cba258488f380e3dfe633ca5a3f9d0833823ef2e43631096535bccf67305bf0b09a5d2a6bc20886936ba1c1615da02e98990e4a77f750c1aff8f8891a9eabd54b036da2c90589ac96c74f1fd43e01719d641ccf10568f9b7faece6fe0380668998a8896aa5d898a549e653f178a96dbff81bef5b6da102e71ecad171bad34423d3bf19940b287c4a2fda064d6ce4619ee5c726f0c8dae017b6a0cf204cb5cad2d945e2ad2c1551e6499467704b01e62741d7d752e3f58efa878f5ad42fc86aba15ddf0650fee8c88f5d431069494ef2bcb3685e1c732b347976d1a2377dae133e6e3060c1317125cf6507b1ec001973694ec60497464db4d74e7695f7e07d2c47caf6574dec77fae787cb12a9be276b99c61c54e74ef601485f800ab58f4069312cba6d09fd4c0024490742b38557f5198e7c15a928a73e958322d35a6764743e942676f11ee872564d7b2aec5a92b3930813c9d998a3b876ea52dafe5ce481fcf3bf28baa4729bf250e87b8ea5aeebc4506f15e0c60fd114105b3cb0e7457d3b1e348cb93ee220160719f9d53cbf8bbcfed64f2a3a93e426c11856199d0f402f3575e8080e8b0b8afa556001254bd4767982a13448ce049e80b1ca806d13c3780093c3e09766fb6a5768a14b9bea98bc84af060aac734299d81b1f94ddf1fad4c939d70114e0ea7d2e9652615a4a311f0e7cb23f4f288bf1b4b8164233fa50e12cedbfd336286d0c2871099590f05c68b2b4ace213cc1062e5f4e452627bfc3dcb87c040ff926be00f583a1a4691a708b9dcfc59469d1a08f24dd392fc2a0fc910e0fa59a1678b6ac5e3e1a340db8ab7814ef49c81c20c0433a7a3f1cfa8ce6eb88d9cbc5a7b839faab5ac199b6517043ae5a45eecadd8d1af204facea965f85227adbd8e84da33db5d1bc995015dd164969ab8309ff92190e91339781110a8e29a1a054b4e729ac76e1ed941a0fada54c04863e98aa0ab9586a3598078ee45efe102dfebdd223a203c0c8541748124537f54a9df0c6d7e7186e844f7fdd27d4ad1defb99170153f3a3df2eddbf7df5647394665b0165b5489219b1dbe0777c67d384f73946d8d25382cbcc760966f6e5ea21cff2c26d0e601d6b3051c3c92ba47b904aacadb5b6da6f65e2da15537b0a52aa3346b330be2b2b190b7c959d016391a9d2ecb6b6c800fa339d3ab70945e90147840a7563fd9e1", 0x1000}, {&(0x7f0000002480)="216d2255aed99be1147baa94d0465161193b42958a0d74eab35bc3e21fb46a3c5a059744bea8efa1abecd221d9f2d3d504dfafee023723037f27e3275f882b24fec454723cbf5ac96c7c56df94397d8d2934a512a2f27a38f76c5d14fe50e04a799c0ba06475be2ab1349d5aa1d4f8db47f35334ce14f0f40e8fe733fe4070b09a9484317ac3ae887f11dcdf001144c00e0a9a25aafd47294c7696d758905690fd1b2615cb2f4155f0bca056a9ea23c168927df1491f6c9742980b0ed703ac2ada240e6f022182b4662791274fa3214ff0b5f917694ec3b01e0c753489c68a3b1e9b2901e97d2cb81456b13a780513ea02fa79de92eb59ec56565d213826e56dcefc51c04b50e7af691562684beaf06cb27ed5d4095abbaead24de8f88b498b3dd15780d1c4b4d747d1f3cd33a5acfb1cd020e5020bb01353ec5f5c46fac84d00014a9bfcefc91c56b0b92a80a402c6cff887bd6c0315cdc8c79818a87372a4d36f8a35e9f3eafd4b37b0b08c9aa6a7baf91a0458609d3abd88a054c7bf180e371a77975afa9edd941b2cc9c05badac1c677abcd9a2547b788122aeb7034a9c6d21999c4c74a51fb1176220db94525719faac17319516ed9fffb75f9c1df822a4f4df07c01761e2a6a663b0fcea058c372ed6946644cbf737a3e0fa996d3f3a12d1789a73459075521334cb0717bb36632fdf1713a2fe2ddce70aa3a5c1322d1f17ac1f717e25fee59e47b954ef5ed45043d6bb063685013d539fb643efef00ca945391d8d0d05858a09928585e6e13a0915c51d38c07fb9a88801ef84cb11980b9e18486bbee9cda0ab460468a9803040947239d4473d391babe47303e6024ed921dd2aaf5c137058360d51d0d8994388788e2d6b14296330775d2bbd34e7c58147d55fed75aa7f97aaabde3f4ec784c221957badd2ad6d0024d92c2530dfd8851beae88df4efb334def1bf24edc2d3fa772c3cb6418349686f8a205d998f9e0daa4b26e8d87c200e9c5617cdbabaca1ea381be91f26a2db95199394f3f50f258f71ee30912d53f84571d10bbdb9301acc73d34a47c8253cbb80ba441c70bb1825913bda17c85d3996ec754257fd48b636f6c0a16847aa28665f25398fa1fc76a12ac6ed2e51cb63b67614a08604de0c1a25ca36b50be2392003d1c4b1022ffd53316af9a6578940f59c37af498b61493847cab34c4aa4023fdcd7e2a4d164e2d9255be8abcf6a64e7e132f5df353b5cdaff6e7846a0bc09536c69606698c76590086ab79b988a7826d4a98858ce5b5c78f97f61befaae8458c6e1b8bc31ec7996be064783f98187892cde36d529dde6101f1ab19e9d17c2db56978f74f3af8cc9ed5cb7b661536c4747754cbc8153f30aef0d12fbc8354aa13e9d1084e9d2b71971830687eccacdf85ee9ef42900166a03c4b80950ca5d51af438d4465a1651b468b047ef57894caedc5c4ea6639ab5c43cd2c0e0f5684cf0e678833a7aa24d854ec6d8034aee508c275c0f3f1be92b4d35911ba915997f44a10f81dfb027af84b4991fc72a5df57dfb1ae98ff7755fdc7a450df7d424604b9941f8d176aa78f547b948d93b304249f931b7417931b8c9942b826dacfccdf4ae3512e4e00fcfa06cc83f58af096c927bb4c7de8a3958c1f1ebe6bf8723030845ccf639725f053e0d702b404bc4b948b91d6c40bf35f5b0ba5b5abb578b3c35c06d0ca70a076003f2ba0072c5a29b33a073a509092c3705acc41e0a4ca46ae219f3489b28d9686060ede47d93b1bfe56f12820cfd35f4c0613c60315ebcdd1b8b213cfd302863254eb52adf904fe3c426e677b07389cdd47824246023fe003540e235d79054656a769b5048a30642125253442ed72c8c75a137d60eb626737a2fcc40043e31689d2b762238fd0deee115accd4b1fcae42367fed0d7fc671986283f276a4b0f9af108445cc6451dbb4cc0eb99e62a3ce17fb5f921808363a6916f22c1693dffd2e3907159e42910b334194ebb27c2fc29e57a937d4b3ad83960644075d9fe879c9a2ceb923e38c818af3fe516719acb875d08b0bc75f44e049610abb24db5a21f1b0c93071a15bc732d78c63c41d256d25f626b00b13edef5f7ad00ea3dc07ecc4f70545fa694b4cb3f8ba25c06e6755261efd29cdf1c280a86d727dec3f976ccb2d6b8b069b2e645f9e9c1f666cb2638723ff053c7e7631b3a5ba65b6b6ff053bb313fda95c3803557aa9daeb73b936f54679987188dfdb770af72e83df19e35fe08b7bfc2c9841ed2540f0ee005a4d1a4e8ac1b2eeed228d319f9f31761718f4cd3bfee38513905e5d276e59af4f0c3c7c3d935522e821ea1c54416148867c220469dfc7c6ed3ffcded7e51dd42594d1874ac1a6f7d98555443ab2e6b4f7270ee8e8a6aa7a93056247b5dc06d4f4e6416df3ad37662a913b1e128b68d2c13b0df064199f7a636980dc7d7ed3b8f6e6ba3369f00b416d3b111528bab2c51a2074f17a1b3afcaab6171bc64df64772930be2b6713d6c9a7cd10fc81dbde65fbb35fe20b12e8a3d560c39443788e09ef27771e216417f4ff2c1656e668f06e3b3a2b3c4073296a7296bbbdf0bf64979705f07f4f3a273a2e625f4ae328c79093f3c0324c072ada3de32693318fe563f66ea73a78e91e8fa69f99f3910c4f0878ca105a037bbb7ec2964fc991ba66a13f45f77108200226616e56d8084e75ff6a5b4a631421e868709471051972fd3104ed63681c9cb292e24e1f8a3d72901ebaf6689e96d5177d5c897b0100e64e363395e8dd5f62670eb2b67e63a2c80b6cea3f6c5b2bf118941df5d49d7b1ee0ba0d9366c827c21d6b4cc7f72099311b3b62420c14807329c4ca17da600db37e188b8781e294fcfe87f40bd5b2224ecdec13b36da9cda55caee246fcf603037699fd58bc10ae811736b86b816f56912e3018fa341a254b79c4682c94dc93384b57408a9c863d01a1f8607a33cc983e8fd363d67654cee5deec055dfe01a6fa1f4ce305d392cf2b2342e4b4ca4e61f5795c20b9355fead2abf4c590895be9f7c76530eb0a10adcb631e8a0ce1f6150d1a66f4ea77274b9de3926bb918e476cf8a095ebec2d3fa3cafd61fe143e3ed47749b7702e62b17ef6636a1b4921e892d8844e05fdf14b9c5b3e9a63f9374f42847bfb58079a8cb87c39004a888e5493dbcfc30a1f345808e2faaa2cfbb28b38f97587c3dd9f4c49430e15b1ab56582767374788ab6a83effe0c21a2e5b759e68f2e0896a160c4d3c962ede288733828f8e1c402a92987d3407f5c2943825bc902f9a4acc5f0daa96374dceb623594c92209cabbc30d31ad9525d85897f2087bf89ed5e9fa24e49dccc4d69d226a89bfca5b9142201dd78a5bf76a9165a8fed75500b0032e7eeb5b7def8a3a8590db6ffff5873a7d766b402c7ab10dbf77fc4ea6602af6c174a83c725d5b545bad2b589a4b16cf559c66162d80204cb03fc0d4e502e0d1ee10edb1506edb936747563c2ab6afa1e4d5f04266f0cd64bb8332ad825f11f00a9d2906534867fe5f9352f6ac97705acf171c8ddbe350e5b709133cf2d27577b15176768e6903f13c4c49bb0b0ac35ef95f6918737132f6a94f2e812853e6ae1295d7d567d4d4b953f6c5bdc9f36867247657ea44b58d54c9c39fb8dfc685bbc75ce9995c825fba731a3349d4a2d2da19486dfb7f18415b5e6cc8a67cd92dea7d0b146449e2c655ff98cd3f5d80574e072473b3854c0594f9458e769e8c96c71cbcf72c55491cd960c5d83540f4d53bc43867fccee87543507bfaee5200842b7c1fca23dc72fc7a5b89aa81031b69d4bf2544c35bf864fe9024301356d0887e35a28ac304edd5002ccaaf61c8b3323d0a1b039fe57fba27384e67add2ecc242f7e2821b02b34ad6fa7a22af23d66ec1911579df373ce8361ec1e3249c1ce618cd43d1fac6019ee174fe750bdd87bd61a4f0f44174ef68182f7b20c104c37bf2c0aadd5f7cdd8d7c1892e106067658e3a05d3133130137cef0a3384a39c54b6f01b02fd32444ff16e31d567b88a2d1a5f02064cb9306ae5a6cff532dbe5c65b0796692b00a95c9edb2b69330ddc461d95400fb13716b84ac0c587f5d6d6516f8afccf83d15732f8e07326e8758638e40f0a49038abb1725f5bbb8eda2fe4a19b2b9fe2ca71aad4d878e5441cf71cbc4738adcc6049b4cb936c6305a21fd90464f8a3c211873296866919e3fb6ee712a99c0c2de5bb6e7f3aa82aa2046278b761bcf676818e70b93fb814bd66a35a273fb96322af95b3c00a4df7c8c90f4dcd2f363951ec3d929043e449197d6fb2eb3bb4335afe7ae64e4f2134d25a0ebd59ac7292a145af5083d06086b84b78b3655fc7b1ed4450494148ff4a3f646617676381c963857b7b0096614399e4e2aa1f1a871d3d0ab7742ab720f75a494dfd1a83f76697f23597f7eba7b6ba5e3881ba5e0fe27f05223d8b733ce038bb83c472736e2cd563e74524d7a2d2164950d6cd3a6f06b63cf3a3c8f852a7ea4d4dde1e569cd93eecba0a6ce5798f05637d02bfbbe4989b62e2b2053244408bc1f5250639534d1215420f0aca2c3aa2622377270c886c56485e8dc690ccf790828705af2e659338832ed748df008060e2f938b47c9923f2446904b30abab3020f672785a35e8387ff22912e1c93f778cce578228c547bfcade5a1912067155c704df83121de3ff75fec50e67a1ad0969b0e40fb64c7753186c56977bc1ec7348a383af65c424acceff79a203cc222f9633ad4d8f878fba99bc164b11e5845673500c8c2ec5e5917040c55569fb6950337d10b5fbe73eca80be29eb02766c877bf22b4737b09ec42fbe3b0b9a3051de63e32dca713d1a05fe9641e14157ada1c500e9c45ceeb9edadb80f3baab16a92373f01f830e7600a5d1f36965d8e5ccbdc3c0e2b239227649a6f7ac9fb8e8d771698657fe5b3a4f608381c6bc1623ab9710c83970d7d880aa6cbb8d1ccdc23a7fe8e51645a0d2f6cbe434a137a49b8f5a8fc79bd4e01c7ced63ee3edc2befffcff92c246fbe512ae4d0b4ed259b7f989a98a6eadc1a2f178c84e07b28fb3a4b10cb575fa1cd2d0c25b2ce6b8b8fa4bcd4f814c6a1b75a3df7c0f67ea830d0db2ed7e45f79afc065f962a2b8a56d338bec08b107030e32cae7e62e51a7966871c2d6e8ec047720b95792ea5aa577d389f011f42bb7d1501432575c656f10ae1f13ef7287133b2cf631934525f8b9ee45814c55994028dc3425db6564b5755814bffffa2db6cdec29a9914fabaadaa370f271700aa27a750ef2db69eb540c8f9f2bf5457b41c11ca17c6ce3821fdb982193f52df9cb6d8710b247c9a59e5ee83af4e8e69574dfdd8fbc9d1083856e75ac88cdc7046ce17a10e286efa087b133a1f3d7918756b878b7ded016f470206c7bc0c51f325ec30b634026e0beb851d522fbbe0fcef664878660ea12a60bf4f3cbcd98d0078e0761668f548a164b8c0f049fce31c9433bda4d8261c5aa8963e3d2eb96ff6a02d83347651bf4d1d43865ab57d180d394588a2f78cc87bf710eabc5e110ce1a0ee3b7171844e3d1d4bf476a24ff51ef8f7e47829bf7c0d3228dd7e2a94b2a69e35686f9e6282e6561bbdce7a9bc212e91b82486d0e63f17830dbbc04ac0c7bd1a1526042db6e7a64a6070a6a8c47cd0c916671d1ad47a1a12bbd4075e4ea9b9f64aae5e9c5f1c9d2625f97f5fc5d1d50c9c0de7f91665ddbb1939058b02842c64607eb76a3d91cb2f9c17982078ec4a12e3e11fe2d6ef55796c7ef165b3ba43c079", 0x1000}], 0x9}}, {{&(0x7f0000003540)={0xa, 0x4e24, 0x0, @ipv4={'\x00', '\xff\xff', @broadcast}, 0x6}, 0x1c, &(0x7f0000003740)=[{&(0x7f0000003580)="33c3f780a5cee531f31c782e74ce2f72b3b2ae15a761fd2c4bc4f5c1bbe71c6a94666905a95fb8c8b0719fa1d57862b3cc17ba6f285303171719091a6c54d10044febb5deec37016b35848b0baac5abb90397070e73cefdccae6af6f60797b1c6339c5f3a7eabc3e4392756904", 0x6d}, {&(0x7f0000003600)="ba8eaba419a4", 0x6}, {&(0x7f0000003640)="88e1d3d29446c8f05bc4243cc4830b6cfe37e7adeb7f7e9632accc343bce7e9718108c1d1ee961b79cef34a11a2b9975763ae92ac0dae8f7d1d5bdf52341dff1e1841456146a7bccb55f11fde9d132c35e7b1b53132df80485303bf9", 0x5c}, {&(0x7f00000036c0)="49bae93ba5d20c0a0991838fc588f1d4fb4cc14361090baae6aa2efb05dc1aab2e40589b91447a8c", 0x28}, {&(0x7f0000003700)="e88b15a8c4b3b1678bfbbd16f1551a596c5f825ea1cf039ffe9ce97ab73f7bb61bf72172e0b33a657d72ef880f4226edc0be22169698c6650fc242f42378ec", 0x3f}], 0x5, &(0x7f00000037c0)=[@rthdr_2292={{0x48, 0x29, 0x39, {0x3b, 0x6, 0x0, 0x8, 0x0, [@remote, @dev={0xfe, 0x80, '\x00', 0x2d}, @private1={0xfc, 0x1, '\x00', 0x1}]}}}, @tclass={{0x14, 0x29, 0x43, 0xe4}}, @dstopts_2292={{0xe8, 0x29, 0x4, {0x2, 0x19, '\x00', [@calipso={0x7, 0x30, {0x0, 0xa, 0x40, 0xf801, [0x24, 0x6, 0xd1, 0x2, 0x4]}}, @padn={0x1, 0x2, [0x0, 0x0]}, @calipso={0x7, 0x20, {0x2, 0x6, 0x7, 0xfff, [0x100000001, 0xffffffffffff0001, 0x9]}}, @enc_lim={0x4, 0x1, 0x8}, @calipso={0x7, 0x8, {0x2, 0x0, 0x0, 0x7}}, @enc_lim={0x4, 0x1, 0x8}, @generic={0x2, 0x60, "349c1679b4ce86614ca52ca60ca97566d4fe36106af890642546b33bb70b4ade7ec08ac4da51a5e64c8499fca2db188e22a434c3a1f3da46bdcdde3ac8df60cecf9ab4819b2f2cc5c9fed84155b6a28fdee9f0e0edc3fc6aae39ede4746c157c"}]}}}, @flowinfo={{0x14, 0x29, 0xb, 0x7fff}}, @dontfrag={{0x14, 0x29, 0x3e, 0x1d85}}], 0x178}}, {{&(0x7f0000003940)={0xa, 0x4e23, 0x8, @remote, 0x5}, 0x1c, &(0x7f0000003ac0)=[{&(0x7f0000003980)="76e960", 0x3}, {&(0x7f00000039c0)="422516b8e217da13815499ac67e749e7b03eaa12772f9b926c562821027f65033ba56bc609dd38849c56ce1d1ceedda75420a578565014419affd534f79ec81dcd72e94a5fc781ec4825fd33696f27e15906a04e4fe0085d12e0c5535dcffb79b0e9f31a7b16fceb6ffc1220b54e2ddc9f50e923eba78c6baa659878861f9eecceefbd7d5386773051beba0266d0c51b82e45f9194d91f073bcd4d594887922d99905f35922b06908fb0b7e43556ccbf015734e236d22cc8272315400e56068482c89c74a1a7be8a2cb3bb8b52f0ad3da277e6f1e8f59bc698f1f86f8a539912b5c4b10d57e347f4eaa97e96ed5a", 0xee}], 0x2}}, {{&(0x7f0000003b00)={0xa, 0x4e23, 0x2, @empty, 0xa00}, 0x1c, &(0x7f0000003e40)=[{&(0x7f0000003b40)="1e6d0bd3542829a7a388a86a271c396f1458ce7e7ee2ae164e6c556b53147eb66f5b56b9c3a3fb1367d60980aaf6d86809909367bfdf6d54744926cd7868a2dbfe4d7c9a1df872cb6de5badde16c804c2e2cb5a0afc6e996056d361435a4160a8e41a148f097ca0a9fff52bbb4975d9e5490e7a84a05d15dfae07b9c081bd0f7204864d9066167b8d408e2dc95d73ade3d", 0x91}, {&(0x7f0000003c00)="975652", 0x3}, {&(0x7f0000003c40)="94375d5329768661d465af9a4de0772f688e3f3d7d72a18106a4f7e5c276575cf10a49e5946a6e6661640de5b5e2fd398b1a5c8c108dff60b8f8fc56b3094be528f85f9357fc163f23c50aa55d934372981bbd9999d1f50330eafce56cdf905cc42e564ddf0ba8f0122fa828d7105c1245dd38bb8b6c5a1d464c280bcab9868c798858ea19a31d4d45b3f2eaf82b28b7287544bdaa898f20a0f659c9cf90c0053fb82a3bacacda8a373aa5ac5961c7971d46f9526b18c55ec1f189601e88844ffc95f063c01af89614ff885f0594a87d9dd2a6a2fb3f", 0xd6}, {&(0x7f0000003d40)="064c2b3d35d1a8dd8a4050ed2bbb707ec54f9f2676d550308fdda31aa62b18e72cf7d4d3510333cf64c8defd3712a6dc12e4f15cf21b90b69c946f33eecae331d0df5f8440faf245657aaf398f9357d034809636ac9a42047e76f34a2de2f585ff51dcd4726050bf024bae4bcb8f66317e147e3b129ae0be47009b60a4555fc7b5f3a335bcc8d59d9fab4c4a50b52b59fcaafd3da0c1c62ef1f67d71f54bb74619b42af17385fdb90468646e47f8ac0df48aca82b1ab6e82144ff0dc1212a7a0df", 0xc1}], 0x4, &(0x7f0000004140)=[@rthdr_2292={{0x48, 0x29, 0x39, {0x47, 0x6, 0x2, 0xc8, 0x0, [@empty, @loopback, @mcast2]}}}, @rthdr_2292={{0x38, 0x29, 0x39, {0x84, 0x4, 0x2, 0xe, 0x0, [@private1={0xfc, 0x1, '\x00', 0x1}, @rand_addr=' \x01\x00\x00\x00\x00\x00\x00\x00\x00\x00\x00\x00\x00\x00\x01']}}}], 0x80}}, {{&(0x7f0000003f40)={0xa, 0x4e24, 0x8, @initdev={0xfe, 0x88, '\x00', 0x0, 0x0}, 0x1}, 0x1c, &(0x7f00000040c0)=[{&(0x7f0000003f80)="b8f7bc4ed73ea4bcf39bc0fc1ea8f87f7f18f3a7b30424699f", 0x19}, {&(0x7f0000003fc0)="11b1d0492a2b64c8468bd8ce87616052a4ec41ab7ac2306b2c2167aa6a233f4caef9fcd74d8e1f7aff1a3a52f3318c15df0c9946ea6b7546cf8b61b895a97a432c55da868c88387a6ac9d4664b7ce82e181b9b2a12e60629c476504084aceb3164dddc29bfecd2bccf50e26c7494fdb3d177d68be8e05bac1b54ff690deeac1018169179f16da8b0a50a4b3ed4ce96d367e1a8a6ecf351356dab84233bf21bd29b580803ff5d2b4f67778fb688c7e2e45a10e79df9cebff3ff12a988f795", 0xbe}, {&(0x7f0000004080)='l', 0x1}], 0x3}}, {{&(0x7f0000004100)={0xa, 0x4e22, 0x100, @dev={0xfe, 0x80, '\x00', 0x40}, 0x8000}, 0x1c, &(0x7f0000004300)=[{&(0x7f0000006080)="aee88b3a78a34f78d800aa4c9353c32ec59381f3897debfe1eba32de4317fb64984477b5993c6543f1249032530214916423c179c7ceebde6330defd244b8e3deaa652f271794da5df3ccc1cd7ee49ea0209ff9e3f4fb960c8c982bda5bcafb4ed15845c72c660350998dd4aa7f533d7bd9f2e158d4fba3159660c9fcbad8acf19dcca37f9a4d6d5807e9f3196de8c16", 0x90}, {&(0x7f0000004200)="a2f150e619190ed70131476b21b8cad5e1ac6c30d3db6e3e6c1264c4dc4481d4d92fba8886506eaf541b687d084b415218eb604fbae004c8b97147a55353b588854c12", 0x43}, {&(0x7f0000004280)="4a8d11d6518e58924585060db33d9afc7d939e3c69d0d27904ebb4c2a726a49617fdd09f0302eadde097c39841af2c849b1ea799dd5b15b9cd6506e4c24d85fff82c65e65142e9fea763eb743a6425d011de6687e9b7", 0x56}], 0x3, &(0x7f0000004340)=[@hopopts={{0x28, 0x29, 0x36, {0x33, 0x1, '\x00', [@padn={0x1, 0x7, [0x0, 0x0, 0x0, 0x0, 0x0, 0x0, 0x0]}, @padn={0x1, 0x1, [0x0]}]}}}, @dstopts={{0xc0, 0x29, 0x37, {0x67, 0x14, '\x00', [@jumbo, @jumbo={0xc2, 0x4, 0x4}, @generic={0x5, 0x78, "1bc18d4aec17f5a4590702e1bf57d7dde40cacfc88c5a0659e3c2fcde1bd37619730ff764056eefb783e5e03226775a0651e4193b9b9b18b0d5804914a48f759614e40f8c0cb4d2a93b6d9a5d9cd42bd2ba2da4aa524b60dcb8e558d2e11a56d0e25c57c56b16bc2ecea51647cfc3ea20321d402399dc76f"}, @enc_lim={0x4, 0x1, 0x1}, @hao={0xc9, 0x10, @rand_addr=' \x01\x00'}, @ra, @enc_lim={0x4, 0x1, 0x7f}]}}}], 0xe8}}, {{0x0, 0x0, &(0x7f00000046c0)=[{&(0x7f0000004440)="53fd5f89addc7638eef2a05a0c1d4a88a98dacb4cee3c24021b591be0de5fa732cea65b247aaeedd2e15056e1c64eb1ea06dbaffa2e05b72e51dcd7d7478d64ce2dd8675cdee3bb7be9a62fd1ec35bbab54adfd3eb1bc41d95ed3105ed985afb538d", 0x62}, {&(0x7f00000044c0)="d0fad7888f66de898467003e8ea2f49c9c1fcf33bfa040c794998df539aa496887cd9562b2e154c8e22a672e6608b86102d57345965673c6db21230672a9f11009a282c49c2f448c44817172b7ddfc1a4a43aae43db970a7062e7dc4dda23b747db9e3a477", 0x65}, {&(0x7f0000004540)="8c6ba3cc62c973bd7f7c6b118a846ccd432f5e5aa15960f6c4a8f09391898d5c357a3b0fd96874fa86ac6cba8965bb6411215106fdcd0b23652ab882c963a8bc4898f83a542fa7f4ce94410ecf9e0d8be0f9b9c716a01179aa274414fbe44b47943668228005ce2f50f53030460166bdd744f0e659de", 0x76}, {&(0x7f00000045c0)="3411fb6a69dff242cf21daf3b3d7042581803cc4a8163ab8bfd84e548749a3bd95c80849afd82b912e71f9232d095e50aeda6e4b27a78eb508031fe5287f50b255ec3539a4638e4d9492b95a1ffc8b528574f84e271bb8edb33b4595edf1703ff0f487ef62096233df247450c9af1df76f796d9ba8d84aee76e6a03aa5e86d01f2c8d381b5d53134511a94b982fb43765e783a7b54f09e2388f3696dc867b0ec4d5d2848ec1b9cccb96fb4489dd29f54c7cb580832efac9ec833f4c06576cdff18f6b4b1e2c70aa509e21c532860c1fd0a0385c5868f4a96923d1cf062fa354a5cc252b33cad33505d4f6b32d88dfcca0ea440539d", 0xf5}], 0x4, &(0x7f0000004700)=[@hoplimit_2292={{0x14, 0x29, 0x8, 0xfffffffe}}], 0x18}}, {{&(0x7f0000004740)={0xa, 0x4e21, 0x676, @mcast1, 0x8}, 0x1c, &(0x7f0000005980)=[{&(0x7f0000004780)="1e3a7c60b47a5e6f72dd1fb43385416966fb09552c66fdb67d9908bcbe25045a5c5ed574b34c2eddf074ba440c79d3a870d9e5ca60ab253b0a03b719dc649a766f40d0eceac609440f5b4b46df14a868852468d80d345c85e1b8f44fa1bd2f33b279ac2f7f17cef809d379315db2eeb475422e42848cd0258874c27c69cac246543f03ea83110cc12b1cd01dd3c88bb65a7126296a856092f47fbcb3f7bef0bfe25e3aaddc41d68c67874dbf332d103bead1bd7fc010aafd346e2189cd3822cec0bf8f9cdd7be0ae7b9257fd1bb53a5fa0e664207ab2b56d4dd71f09f062afb7f1c593ce51a4bba135cae0fa1de757a64a6f5fea440d2bfad001023a4f28e2450ef41e14919ec3b623e7fbfce1f6d5de774de1712f48bc6c3e11ebab952d3f097f060c82ff54aa12b261e44f51f58e27db622c04c5a97d5e24f1da771b516aba024fdc8abf94bd20a674e3aa5bab85756a2318d4126c474609b77286fe9fedf7bd0c6d0c96a3fc65cf91911fe0cf2f2956df830a7f2ef23d43307a07e665b50d999dc27ff4da79065f2dd750c9fd5d7ba0733b8e3d0985050f5338880a8604d04b486396695a79e78256881ee7d7e2650fac0d57a54cbb43280f6097adaa14b07590ba2b80b2d97d7ee94353be1f53b78842bbd3133cf31b2acf82f520a3a9a3cd694e140dc70fdeb63af3bbd6a9d263ff8e973223197014aa68b98d99e61b4b67afe52cea5c04dfabb3c80de0590bff993654e0129ca1497ca97646937434f2cfda4817bb8b0b0f1f7bbe1efcf27daa3b5794084c0df8eb654f49516be8659411cb1f8ec404f119a5f92fe1085a6291cdf37b5095f4ee4073e60d418600cea4afe684cc4e70dbba23286ca50c400853f2c6629a72935a7afed74e5045e52a8b872b533757ae0be2b2fd758781ba0e25e77e67222b191584091646b3b56474cdbc70602655464d95615b99179cb3ef8af5b42171cbc6e05ba076a52adb170a94dfbcdc8c6f7444c417a141be499b5e89f3a4d33fb851d861f14065ad1e8065a7f68b2d3ee420fd1f37fbed56beab590415f392b8996d5a035e72edea7ac769b5c1393c7374f9a6d89bf727edbba921061d252fa53d9459384dcd3c96450327cbbe28d3315cf24b1ddbd3a691fcced3d5d833fbc2c971f959e63f85ae81aba96e70cf26e5ad08af2eb5464a6738c28eaf0371b6639cefffc11aa8347f0aeb9f7086ec070dee59ab12286ee3ee385cde46f29057b4150bdd1acb0ed301d34b49885afe4861ee1da60cdce3279f45d30d05eeb43019e0d40c2fd48c7590d8ec751074c40927b6e83abbf8ecc1a15f3ad9def1d44573929c8d564cd3e298b689aaac9763c0b714c82bafed329685fa0b40607462c65bbb62b288f592a82723fea9f827b523c7193258388ec9912fab7bfb061b311a6a4869250d726f83f7dcc27ca54c788afc131355dc59e791d4c15d2c7961ce91e1c0303dcee7fc8ed497cf5a9696d750af581711d8229120fab599bab00e7afbe66d4f0e75ed5ec9c734a572a13e2747098ad850af93d9fa5c58530ea9fc070149ff1d1c59966fd3e44d540ef55fe89ea0c0a1cfaa74e2b514d916ea39a7dacd19a990a227bc9eb476626d4eae213aa7133f201afd968a2dcff350f685e0ee07956fe6090b5c26349387ab671ed00d43f41ed1edb228046e7eeb3a7e4203f0149f1ca529ada2052e48a22edf88c575a094e46d57b07fb979eed8bedd9c72f9fd334b6e422a873b82db6b2496c88bc8991d09c64464866b192ae2312391778fd74ee12fb6614fd57d46931f48b8777b576fb196fe87680b8620aacc4c71596de956c2f46ff653fe126debcbf20b353d5b015604f5546f29bbd770f6594d3b6a17090131780319883e3e7b1b40b13b3cb9df066b6e4571dab487738ea98795d0e0e176e5dcae2280c37523c209d0e3387b7643a5f5b850be5d8e8d6f5a2b1a8e8edb351ad91d78b179fb1701e2a9627a0c290cf4ad7efb6330b045364227e11cfd71843979003c1b39ae7a9cbe665406d4b49d1e14bf51a68d6d55165a6b908771dfe41d93075a5c84796a8bab042f0bfd02f8c7b7511c85f367a48074d1181d29b69fc18822bf53ba4c3a2247379bbb5aeb51f17a1e0daedbc8af7829847dbce17af116f518f1fcfcf2165742985580b68e371ceaae4e1077dc816cfe398cbceeea0c3eec4b5a4ba33daa5578e3d41a24fe1eeea45c29ea840c782c2b666efb991663088ceae71851ded61aef80035e31303b596907ccaab7eab977225be355a701492c78e0b5f4d1914aef34f6ff20cdf4f4ccc23695b35585d2503b6a9e741a3ba68073ddb5df93f55ee806f1036b3ee6f346f2fcae98ac2e94d2b1f0245e6079781a6a9adcc42b6456b7d7804e2925874ad4e80cda367791028945ba6e564fd3d2b977ab2f248bdbc2640cba532e273180b4c708ea10715d8f55c453defc24a7939bef74f0d0fc56d4f22f960cc03069d6b9be1eacec23ae9c31a82726a250c9d25afcd645414262731797fa92c0b0ff4bdfc81e1d607be72d949a45886920d23360c98cb47d58190b60f5c3af2d51d25763ef6c04d33a7cfc87f1ebddd14d706c61182b98acc6f0310ef4b72e0719a3dda86d135600423310f9ab6610daad135ff1bbb9dc4af14b09022f99d73bc4589748b3e1ec3c157c5d97608e63e97690e864c31635a7204150c45c70a8fa7670409bd5e30670b703ecb735415b1a743f186c4a0c84edd7274ad583ec2ddd3e67e5e760db74914c9ef3003e6c512f0b9c6d26830a8d9eac9b540b5929f3eacc9279c872c8698cb2184b1c8f2648e20a5c4c5ad13fba630c885e71ff26cdea26319c401aead72081418c161910b30932b4a250bafa3ae69b9e4b5a98fc2efa57eee8e7194d7dd2af9778ce822781039f951cdfe7237ba1d9e572d898ca51866b199aebb2e512a8e4fc9a80320ee0d9e3f093e086aec0bf20759e43bf8ca637bd9f61e78026da737475685b9371730f810bb9d2c15f0f52f2598b529200e7c669e42f4aa00ac9542672b8073cbc56c525d19684930e2d8f79bfd1d67b35a888e7947f8cc1ed00b9b4986d7c2aca6c4111da79cd85bcfb584e8d1ddb3838a08d4bb099827316b73cd7909f6c306642d4df87935ec815d04622968d27f63e8c1e2300692f5c19bee3237eeb8f76f89d0d267213eda627a8fc821100510e939286c9fda33e5680d86692a05a0f6223842e649f4b06dbf6af0285d8f86396683ad8917c898b5d9431b1779d28911507537ece29a86093487208ae71fc8488d655a9352049e4e7d134ecfe683a372c77c7762dfeb48960a2b60efeb8efc9163ece1042baa952fe329be36ebca6466ae52ff70190f95ebdc5671cd6e88cff552631230ecdea82f1dec08a391fed26981a6907999c45f9d75487226d32d3adcbebe9a7987d3486f198e1670b7547c3c15807f6f853f697b85999c7aae5a7623358d285e5ee82e7e0dbbd3e8b5f87c84148134da1aba1553b5060a887e193ce0607a2072c76cf4740b452e5c3802eeeb55a2703af0b642d4591c1de25734248001ef2147e660de8d0ad975f006ec836936aae07c6e169ed4bda6eaf8edc75defc8f00c2954a86abe25087fba178d837d0e7974575fea20c96eb32dd4dcca981ba08f11f1047f11a28fcd52521f98e8babd1b86006817a1a24a13540755dda395df961eda7ca3fce7337c3384f8ace09dac94801c5e50a634c38193682b9173f4398df86e6e97713ac3b9717b44eb60f27d72ad1b92aacd1a66c7f9970acf886f0eaa1d5203388f8dfe9fc6b1f07b206bbabc4cd982b5dba34b15801dcaab907506890c4955a3b76bd93cc0e3d12403bd3a66754909fc14447f02eaf9717b219c9280f7c32d1f5425f05a4871b40e9cf8a5dcdc7d8c3fb63556d44cca59556a9245ee20e982e98251a464b757793ef6c126c6003eef73393a5cc05e6ec1e49016f33f7f9b76064dd04021645402d49a6453e63df4a29a84fc1b397aa922f9fcf755c06ba45df2eff2554ce555b395b1f795816179224359842d1f95836224a9c1f3e58815daac0d19bfc08c8e1d06c3c5de227c68cf2d11bc5233a481048ed657114a2da7d4e7068f87eea91d4aa4a2c469616aeaf8e490a5c1e444689b0d33fe57d815ff1020f94a2da597e02abe324fdb760cfd4d5bce89a95f7fa230ed0c8448abc0688417e5cedd095c6be83f024ec7f6091558fb8f83b25b0a3f6ca12966105ca76217a22b222077a67cf853d4b33322fa81978cacd97e72712b9efedb4ba740565440c49ecfc37d54f42ed5818012b54db628ec5507fb30df55b1ff522c5e4f246f171df146f620a914673e68b4a838198a459b59f2e65ab9e27d3fc4fade7458f561a71e80313995a2ac01e55299c5fb5c6aa0c7391e633f3c99e815b3e6c6c947e9f0d0545d930fac6ff59de06cd451e63c7980bab09ac97a556ae69cb5ea4bd10129eb36b4d1edcc0e97fbd4d095a18d91f5a07532a5b385d7d2c043537b627f6a1a972073244653f173c7f119ddd88e16c1c3336d81bcfdbdce5411e5fa4079913a330205dce4785f218aa928d8b81ec93a5cb950606187fd5c234bdcc06467bcca26434d33d64f2add9396d8cd783a0ce4cc81fa2a1eb13c80c2e35d40750dca170a9b7935247196cf5a95f7e429d91a2c77e9e5439b2f9fbf4a98091f0ea359d0483e34cc14a089b5ac8c92c3982e5b3b526049d12a2a8f88cafb45a00b46a9a9b15a5d3ecc7de10a386ac7bae1d53517dc65647df3eef816a42b30cabf7e35ed027badc5caa87b80a93c89f705b8cb233b2d24abf1c181001c911fec6fc7ec7a49c0ec24fecc5eefcf3a72f3deee7b8615f6702f7b10b60b8f41a53c3a0f7071f805b2180b771b89da65d4c4f484c2b3aeb43018a7a888ab45cbd691f3b32a23db7303ea1fc9f48194223462530fa6a2f69535aa6de8f735789bf346708cf5dfb9e498097dac2ca0209cad47ab6d5c4d7ff837819c615c5edbfb7efebc885b062efa27920a396f7ed0ae695a3db25988a0b3fae8e7656ea343a341c85cadf0ffeb54e341af736ea1fe129e1234579ee9640a993efb8151a32bcce94a25afd0765d82e7f05e128e644c0e2e4021a2c346732fccde5a6b6ccc45dfbb1e0f32a9d3e7afe2b97212c48996b6e07e80d86a9af4fdbc7d2b7ca7cdbcf78bae662cd6cc6a314dddb0590ff9b4b4c92976b553466d201ba6176c0c13eaadbedbe940958a16573005d227229b5e2917a09894b047bb0678f154955f214dc12cae1cde859317fc573fb517ae168ba2902888202caaee9c1dcd546524eb1e4a0c7043543981e9ab6723666c80f30a4921a4d40282a644ec9a5b8f458efcf384080fc71587e37df31c7de59dd5245fe5c9577cd35a2e398c7be48fc807b380b698b9c3defc8c9aa690df614686e95c7c7bf9d725d85da93bb544c9acdf81eddad2f279e121c52ca14e6f8e165ee549f48a517432849d0857acf6a19aa35147995ea34dfbb9dde1ec64e14ff12ac7624dc3d4f7195f21bddce38de911d20f21afd5ef21eaa8a2d929b9d12da532445a5fcb13dcc158293cb40267f84fe1cae4fa64d470105f1490ca5ee8240866348ea25890b12e5861ff7ad533f4da3dc6c68ef3c71ece81b2b358cc33a9c026086dc7c82256e801b0c03ca849013617baa79b9f0ff0649a667bcab10c99a344f425913deb6ebeda5f8859dae2342c680044bf856ba444847aac070495c29adf6952c8490c9489c926a7ae0bb4166e392ae03677c", 0x1000}, {&(0x7f0000005780)="725a14d768bae84499bd5b528fbb97a71e321c050fec8587c0f854fb2fc86ccf58232313532f0f64d002e9b128e5fc62ea1bd6432dad0ea752ad0c8f346e9df38b3873fe9b720010973ffb0f39c5597746a91dcd55c46ae947e6aa4062072e8193a76911e930147b3c011bfb20a46b", 0x6f}, {&(0x7f0000005800)="d6268a5528f585e113399f0c3ad37b5bc14a4608a86a42dd9134a96774356c2345d98b95390d4bc0439c067d36af24011e2344756b963d48da70d749f796976a07fd5f7b0419de30fe9662e4f3da46be21e0a68a2012fdc7f70c505b553ab07968770d86895c73fa1d46d272bf448f0b6b85fbd0966b7b6395c501e00549c3f7114f8693b21cc1d3a7f6655eb1a2ce122feedb0dbb484b4dd806effed2741d5b1bf0adaf0d36f67e5e2356c4ff2fc5b9e88f3c76b495b8658b6e4870f19634a6ecc8cc4888fe3308436aad", 0xcb}, {&(0x7f0000005900)="9ac94d9d22868b60ec510aa10601a340d72400314797f37b48a3a986ac76b52241f0122bc3692dd97e0eae6aded464b35b2763be5d83fd0d9a19af4a23beb7875f0cdfbd1f6cbb1649252ab851e23584a400d590b55fb2dd7fbc727b68882311148bea93ed8aa7e583709d80eb771228ca01cf5601d5928b885845", 0x7b}, {&(0x7f0000005a80)="f482c32a06a41039cf295e832cb0d32ef9756d12ff47602b893dc14d6d445451604684bc1f57f94e0ceca45d20a751e558466387b8bfc4487472d8e452d137eec82175f984d589239512a29d4f2a1cd3decdaed1ddb47fee6e37f2c506c805f64879dbda072bba47e45d784f923cda42b2a5274b5797b37e3c3599035ee1298db83940f4173625f6f57f94608cd333f636d0e514aa09df886a1d0fefb88671af71159aecfb69eb2fdc8c39a4f53842431ed4545ef6221849f0edd1d30ade334aa4bdd67f3bb6ee6de8cd0f83cbd7301f21a1a335500d6a2d130dc456b45e26efd6c0bd17ff7003b66f4a1d7acbac41c10069ba31583f", 0xf6}], 0x5, &(0x7f0000005b80)=[@hoplimit_2292={{0x14, 0x29, 0x8, 0x1}}, @hoplimit={{0x14, 0x29, 0x34, 0x3}}, @dstopts={{0xa0, 0x29, 0x37, {0x87, 0x10, '\x00', [@calipso={0x7, 0x18, {0x2, 0x4, 0x7f, 0x2, [0x304, 0x4]}}, @calipso={0x7, 0x30, {0x0, 0xa, 0x5, 0x0, [0x600000000000, 0x2, 0x240000000000, 0x2, 0x1]}}, @padn={0x1, 0x3, [0x0, 0x0, 0x0]}, @enc_lim={0x4, 0x1, 0xc3}, @jumbo={0xc2, 0x4, 0xf26}, @hao={0xc9, 0x10, @rand_addr=' \x01\x00'}, @enc_lim={0x4, 0x1, 0x20}, @hao={0xc9, 0x10, @loopback}, @padn={0x1, 0x2, [0x0, 0x0]}]}}}], 0xd0}}], 0x8, 0x1) (async) sendmmsg$inet6(r3, &(0x7f0000005c80)=[{{&(0x7f0000000140)={0xa, 0x4e22, 0x7fff, @rand_addr=' \x01\x00', 0x7ff}, 0x1c, &(0x7f0000003480)=[{&(0x7f0000000180)="c22de8de49c8bc74ed6a27990b96c7bbffde6e451106cd7db38c6cfe57439b6ccfd0b713f70ae54f02270b4776d47ca2288261af3a5cc7029af46bf8119dc3311516af33ab353064c342299bac4eeb1392f74f4ffed1ad7addabaf1fa8010cf13fadac6c44bf6d2b", 0x68}, {&(0x7f0000000240)="c97c8d45fc51ccbd284a494f7fa337de3064db9bf3406abe667335fb85ab0ed822756db072b12fc6d77789ab7c1acd7eb42208104b970485ba497cf158a0ca4f98eea848531b15b3ea19cf906671947387cf789ce725ed5b8e20db53c4ba19fe700a12e3d8569518995b863575644a44f0956259c90c98c6add1eb0ecee40d799b20c5bc7186300cb062b46915a2e5b13af446e725cc02114689646ac7290dd637dd1392c1bf5189299197b6ac5106256696f0ff163a3c739be5", 0xba}, {&(0x7f0000000300)="16bead57a423edd4b1cca1b2504dcb0a38fdc890f65b7636d8daea0c", 0x1c}, {&(0x7f0000000340)="51522000b4abc08f282fa6a69ceecb5e211d38939472ce2f2f5337ba56a8c81a5e9b51d5413e3444980182b0fb8b08782b71830fc891213a280a354245d36cac7d7d6185d124a6f1eccf9bed0c8a9423da237af98bdf4a9190affe49352eac845cb25a71680a1157516364c90f75a0a7f544e8512912b91163aeb614e3b677d318163e3fdf33e3e17f9c45f0364acd3c39ce5f8a204553ee4e6ea7", 0x9b}, {&(0x7f0000000400)="103ddb76f5c3a6822e67decd0875a1b06ef2eb440a6503c0d920b69baf9137df4c1f3dbee66104c86d6908e8e497ac4f2ec7656337294621b9113bffa9f202879a2bbc3188545e5158e3795c5a6ce1382cf68b8965d2507615a23c22580813718c8c5bd7d0cb1e8cf01865e19026a5c2c724772a543d1fccdb2952561885ef64ab3f4c8c45d19e2a39892f0dc27dcb5c7546cdeb07502d85daa11369b995a80571577a527ccb869d0dc362aa6ce57da19b6e4671e95a966fe8d707037ecf9942bbf8e61d4c16a5d9cb5e2388ff4d4707dc266207627a1969170396ff0fd00ecfe2cb616dffb2ceb658b5622f6cdf5f2d06f6764ff46ea4bbced621e95aab636b4653e848a781ff8df5c69f7d357f4145628c57a9f4aae6af9279ffe252c3be1ea2c693b693a353ab092d3ccc4e99ab076fb290d2326e317d65548ff64373b888a2f2817f62a73458059168d118424d88bc0f8f067678eac0ab22ef1123083dcd46fbda980201f2163286286d2e086ac7daa12e40880bc0647415e568f2d7af092c8f4d85008b1a335f9ce8a3df4216861b579152b1ab678e294fc46c815d638ce4925c8b985ff55072ae52df0a6be6f6d17c8f781a0dc1e114449c3d947814fe3decc167719d0556ff1007aea91abd64a1dd151eb0e1887409fab15a4355f610213bc2b3dca4b94daeb00419af0a5c16470d3ce57dc844eff080c09772b1b2ccda8373d8d4c6934aadbba6bdd76735f78f2076624132614a1f86769ec76aaeeecabf3fb21a5db4f807e4dd64ac62385ec5970b7c07bcb50cbd6cdbf2e0325022db87c9cf1212a1145ee4f0d4e67abd86c5dcc1e00982ed884654ff0f9cd4fa5dc6988685c91dfa182789eff886c4b9abc0323fe4fd05c1cc37e922d7ff66e8b7e7186015a28dd32dce7cfdca009f27ddd091f385af46add3449a9c9195e1c9b0b2d78919a60ffdc072a3ac1ed862968f83701bc25ee9d60af2afd1aae81edd72822d536044399ab2fa79252c633a3adc23b51616a4906ae17f51662033997f68a4ed695bf04e0cc1fd380873d70fb26c56ee26ca04092993eec119beaf5810ce6dc260b9b0036551b3532b2c7def534d3a21f55d326a357b22b5be248c70aaf1f0753dc379e70c2ca8420e75eb0f718a152cc7a12d34c7cbfb9de3ec35ba311b2a26050b2fd6730ec21f9c48d13f60561951d6be3eb6cd6feb7fce4b8049e504d9ad2469620910e24fe6d2358fdd195dea1e427b2fb0cd8d91276c42bb45e25b7253e8d38851b1cbfa73b67070aa6d576a78eb9283d30b38825e73413db827cad05211552ce49fb364ac6cb5d4e26a86c041dd009290cfa8e261b5cf26162fdc8a5d091dc351ff09911b1cffb352e6b814b16edc6b059e0ec7e0e218041343676c703a17a2bec3c97d68f44562491f9ac0353f1a0657dd3511c2a489d13a9b97feb37abf2af95a9c4241a12429954ee24afd3b7ac672e479954f90c0a591ee903afc8866fa38c1358b2963683e49d54bce0594eda4015b1526599d51cbe21f93c5d7e0a446e782767bffb793f691d07929670c82d373c6648e82b70ef9840b3ada8d2aab84d1017b2a913da6b3b64175245010659bc3532f69c58599a03dcce41c29d05dd8e3a8d4716e8b19a5f4a3a8f37ed0f0903e5d6e38d66596a3c1cd48df3347fda424e7691d733359bbfb3feda692331ad81a6ce60764bc0889ce40301f46c6a71d8da8f732d25a3c94dfc1ef0eca420723851e37f3b3f21b659c09cdd174c610693f326d8bc5aae2b72bf114f95ff5e95d509208641afa8a262cfad171494ee62904174fd9bde07dec67ca1f071975d8a800034d4c13ec8a420708d10f9d5d7fb46124dff3bb7cf4fae06e114550122cf764194186f24a9701bdb279383760843cb0c3594d2fcd74607e4e7681eca1e4ae6f7a703795bd66eb07603e3b8920bd8538c479b8b9af34fe4cd2465bcfea5c15ba4c2330e9b6c9165da731eb6213894da8effb72773b70a812c7e110e4cbfe91a0afa372545bc0d3d60fcdcff889ffc6034c208ed41e37d60418e69ab6f285b47294d3e9ecbfb3a2a100189acafb9523e8fcbbcdd99a4be6a32b2a4e62c1f657eb1a3a1e9f151681773b3443f79dca58d45118aaa92463c6d3f47345bc074ae37477b4cbfe06e89bdff5c2660ab422cefd03f7b4d0cfce6c90b2ba134f4ae176f9acfab18e8d06931c44557d10373929a80a95696913a0dc29c01c827d79c984526fecbfd6d480083d47db817f68b5ebbb2e97a97b37ace1215872ae775267dc1ae6d41bf2b99faffc154c2afcabcbde090cbdb2d9b9e7fc3ae9aa3c7603e72f6ab25ac9487954c6ecdd0c43e336e495d7f0d84d9452f413eca875377f4fb78ba9bd16550fc90752c8a58db8f08811e513162ba329e75152a4503834db7b5d728c68daa245e318bbe23ef274c4617756b5edbf6527927269e4162468aa3097587718668152733a88b28748541fbf9ad9ffbf8ba9a9b8ba0a251f2eb214e72d08b64389ef1040c825931bd3a78cb9ac6fcbde9528fe57c2e4dacc9b5499258c3c3ae40a5c70d8599cadc1d200d9f7686e1ff0dfcc0bcb22e6550c531fbf185abf47dffd485b85783ce6f79191498de308b7bffbb16ab88bacf952413a7669c5057dc392b1a4055ed8dbf3cabcfd5e7c6bf46abf92fc6fa5077a1b1c669fd554cce25659d09e63fcef040cf8489af6b7485449a616453d733435fd225d56102e49af1977ca972b54b2c566e1ad98048c684e020b4523842a85a13f27795ad4378160844ec06190d358b2e38bc155980095370620945279e1648e299ac25e217731a3d81da91bf0a9cf111601877c347d3f48728bddbdc31c4c5093972c1f1d513a89a332e0046a8c911af2ee273c14ad3155d6d15e0d4816cc3d7ab58999126e1e039fc882e7b1fa63e4f366664f2f487538f6b1e4f6bcbffdc7d70f34c2cf4f53885a0e7ee1224c992189504abc8a702e10608aa66dbf4b4b305a3c4d99937058806c8af9b853546ba1bbb84bef9937c308757863817272ef8b95061a07dfd57518ee1c4e200303c24bcf982d9dd9fffcd677bfc2e33e4026af218e76c2637b0ab840097ceac645b17c107fa2c50bfbe97846c3f202430c750b3815590000e0a74e231efeffc234269aed69232426b2408e1d362879498c0e425d1f02c53cc86a4892cf680ea63fce419a6c7df843b64f217b25b1fb10049678874d81d733bf27fc5ec89aa7c416e4f2dd4c88b530036551b4465cc756bc31e15327caa299863a99307916412503f90a30ee79eb1068230bc304ef145a1bdf7a095c1196e687bf8e9f780358314571840f2b2ef35ca94f131404b8379b2d71b9503c6431ea932771994f9be635e5df670302c90745023c37169489517b372f373d0f52717cd572a713d1e93cda0e1fb05e96e0116f4df465f7bb271943e958ce9fee5845f79847dcd4c01b485c7b271d40b2d704a66b32c9b3f2733576733e03bd4af60c8b1079ad0af867f8601ce869d5f512b77c6bd4110eb61b6da4caae42b75309f978f39596e7eed58c7638901a3aa6b9f97d851cb4efa915a682db74cb64f8c28e268d65689846bcdd4fc2cafc5934e686639c0e2a08a4eb1fd14ab7231956a75decb263e99f74a7ec9baf6e146c157d63a32fbe40b9c4cb8f238a9122a5f34d9b62420cfb296c20e0be292cf57c90d8461425cf32197d35c7747d95051c3c4b4c459733f245f40922c32d4764a8777878df1be8fd42f85de3fd01225a1d8d9cdc756306dbff987c3b4bde1c229ea45f81b9f20e6958f8c8a6fbd7d6f7bf9259b060c2d64f74769c4a66dd0be11ee9abe3b625a9a7677a2ddb94a8ab825a8e4b6b75ffeb9f54acabc68ae115fabc093a8f444e19ee266d7ad96a4fa87e70effcf18cf184c32d5dd71e7067157ddbe5f0e2f875ca0c56aad214cc9ce8b1a8b1681c74873d5dbe6c93dceb3458854b07bece891005ac2b8b4800eac5ce0211b234e6670bea324806a3ae54f8e4a48c3d7bcd102c408120c7eef179a9f2cc336fdab92b0e78511022d371cff1cc2c8f23eb9051b1648821db6830e454a2b343fe670c063c222efab424fadd58707352a4f17c4faae8e738a1bc60def40e9b8c731a11ff1ab41876593a82e7f60c53499778ad44986fc9dfefadd6562b59257faf3b45097980ade15c50f9eb6d5ce622ee86b55aafd5ea532cad3877d8965b1d831b9577dd9743f7026163c07410af842496916db1360133ddfd26feb918530a0f7ab664b6f62dbf10aa55e78a00061451006b3f81577c7e55771a08c835b56c332ae77b3f529ecc72f22f8ff4b9296480b50f88a751f05032a109502a3715d0411349de02be10a59284a91dd22fa3f8246ab1513c1cf9e964734878023a9cf89536f4c0a5c3286ff654642a0ac42d72ed68583f48c267fa26c8554d44bd312552b1121db719fdceddb9b7f3ae50fbe82984cba814662ad3e6d3f644947c89caf902de2eb394885ce08e580f4d087bb009d6bd41be5b6ddd41e95d0d7919b2469632dbef5194983500cf6b008320d6b657f6506bd9ac992a1723c477807379226e3acb4d8da21adf5c4e7065bf3d3fd395fbfab060dbda3b54e65f64e3ce2d3d20000bf4748c369ab01849051adcb6f7699034a872204e2588b6a52dbe81c0996d01612e15d0cbf21a9fddf714aa549e1717e10d24e7b919fa4e87d83b7613b7b0ec2a0e7446a59dfbf82beb6524f41f0ad05898f8f0e47e1f6beef9c5519f92f516f6a2d6ff4a56c083e4af8e801681e8782e698ab11ad3410a210112bc6cdedb09dddf7ca681d38e03b38aa5c90838e3203a232364458917dda37dc1b161ae1d0febd69347af30769528ee293909c645c6120a6f34b5504bd20701011a3cea8bc301cbb90f6b8fb1b3b94fd9a6a9d53227848d54d9b5749d30b3bfa23763d41e64fc8c8e07a74d4e12001de7f4c51e14ea8a763575dabc9cbc2d65745ac9b2eda7fa146579513375d7ce55713618c4b5cd44a4633d6c2647866f745e405b6632af9aae292001ab7581a11289a4a4ae1f235c11040024f89f5317b58e6fc7d7b9ac8ebea567debc7a5d0efbe8eafb374d4159db45767a829a22f7e503b82e0399c1350e8f25b0f428745379f9cc66f093615577fa527d9fcddb8cade543ec4aacd0f46ebb297e8cc6709915ba3b21ed93ea9284306cf6b3759853ddc1bb994a4b2dff3de13309d70dc0fa76ac64aa8618ce6127f38d3013a4bf0d2a29c6319ffd3987b7ce98a3aba4affc12c52e8c2953e3024a8e200992580601fd0c0acbfb8645453ca752a7932d552462109d208d58ea0c10f2b9da8d214456959a71bebe20f2193d0abed2bb4af4cc7b6def0ad288abe9b96fd256931dd2e4476f237daff691bd9539fc736cf5b286f39535f7b35c821b745ac24a19cf3af9228603c713e9eb70eab31fe0804710b9d6f2ce435dd2b45dee997fa333a3d9c2f2ca15271c385936ce687b102a9be42711c1baaf9794ce69a82fde4a39aacda111fbb965d4f2f58f6b48c544983fabd3a05839e204618dde1855754278da4d48d48ccef3d1d9bef906cbd60af2aca183b802b6daa41ba3c6f72ec6b12942fcb9c0c74cf79f9c0e7137273de4f196896f9cab3f471a182ca31268e86be5b5c5b00ece320d683330c37ef00ff498950b91a6c71442af1bc16868a0ecdbbe71d0d2e513724c3e9c44c0a058a1e95122493380cea437048532fd8871c00e1d8e52f6ce95aada96968a76a2c8db620747d622bded19f9afd6", 0x1000}, {&(0x7f0000001400)="6fee70109ca972282af8091517580ea97c02", 0x12}, {&(0x7f0000001440)="cf58e1164b21dbfd7361a6608edb7c3d74752d30bc803105cdc9c267ebb3180a712f3329e2daddea7dfe9a", 0x2b}, {&(0x7f0000001480)="57b10b5886521c21e2472f7c92c83189c0bb4ce02bb5a06885f74e06936c6cbee4ec566161101f109bb1ef4c83803331bdd5e94831bf88064113339a3e7c4658a414a386877130b9746550b2551cdfbbd99d9675b28d2d3c8b00f1866b89a4dcb1cc223754e09889f4eb0c9a5f3be77531564fda6b4aef20a152c5b5f3ebe7c26930e9e95361615e8edd1aff4a9caac3e87e6c467f4abdf8640dfe6cb42d89b48e1130745d47ff43353e0de9cb20156bb2cd8a64bbfd22d32d7998d405c3d26eba0e48d42ec8b9147ff41cb6c56416819d66c00c82c64b8662ff23690062578f31de8265f252d50a42ce16b16c64324a26700af0e31c3b8f09dcee863895be7fd1ebeeb89f57a691d06ea81337977e41970a37feedbb49cf7e6462a3d755124a58fab46e662e7db352706838b9b24ced05f4a52c542f537de34f0fdee6032367d49697db1247e7a43e9a4a0102ba8d71f4893bb5474121c84e43cff9e233671866a73fa7716cf2012e80905dce31ce3d1c5d9fc38f8ec29ceb82c924efd47d5f4f7f9f24673a3dc1fa5f4bcc81caf5d96e0ab3a2eb093c057a32a3122d104dd306e4382c0ecb425d94737e3693f1117a4b1780cd1f9d8f1a5b61e11f52b48cfa21777e9833ee0a8a065e3d17c514b76a04c69fef9b430b6830415d8557b23fd5f26e39a620f4128923a6c0c4c2a24ca8b27b110333968ad99075cc9f222d724fd87866d9858df6d1259452914f14f8c6c7a70daca6b85055684b8dba8225078fde417c6029e6c45fe7d6730c1fd965ab7628539585b25059b1c5d20e98478ec92b796ad25d10b08483b118d146d2b73c7145c6cb05f521c48fbb98e50dd75ce522ffe2c2076463030f7dbf61c113bd27d3b2886a5d12d578f97d8623c93befb22972be01d3af9f5d4edbcf4da523053a3539c8d0e4ef407ed52ef98a49c7229aed12903a141cc2b16acae5f92e0dcc27719c5aebb93b08cd5d86f9eac038cec3476e4ed5adb727830567fb405b3929be507152e9e2230facddcc133103acd7209aedb0f251fda299454b40a721017627545eb5b8f0deec7797211375f14b57dbb9e3f06f4a33dabf179d5e98c03798251d62dc22d6a760244bbe6858542183b2d40f7c3c755fcdd04a8502521e98a7a5f59e4dff26746d275e0b40746e76699789f88e45e85abc9327c7773f73cb91b43a35694def999fb960723851660452882860fe6623ab87f13337cecf7c8b18f3ef5afed07d5134e168722097b7d45e16228821ef99dac527e961b63fd61628f4a7cf9f1231262c83ad5db9b550c98adad1cd15db17cb947f192030fb44d638172329c7d87c35e510f439c3c72d952632ceb044cc57ac3d374dede3c81c8e75ea20e731de636f541205cf8ceb14d998793eef95ba2320e5090b299894fbf885462a4d909cfb8d021ccce95d6eb86ed47c0c8c8bc384333b81ea1557a66603c5a2f23ab0100cd94e3521f9bdb14c4c32ba9d9bfb1011e132678289d52e6c0c8c877974fe187f80c7fd11a80a59055d6d9a132f456df339fe90cd1fcf0216bb1f1c139bfcc1dc72334d08f467996ce61b692da00eefdae66bc54db30388665eca192f360ca41e7401ef3dc0f150f5831bf6356cba27b642b01bb8dc7e3f711c80d43e751ea7a1ea3376edbc704f8765df546e11a2fd23e1abb034c62eff880806fc2098981ad0388f233bb1053dd2a0bb5b1605ebce3cdf39f58db49b85dbeabdf9e4561ba2de2d2436b4d411ad040b419f5ddfe8058d02cda2edd51c320f1a14ceb206bf79aa649fbb3804b3e15c1e937358bb88a4b8be3e6de6b2e28c7c84964619677745ff8c009807e40017df35c60c11410f5dc7335d539d1c3918ea760b822bef85de33208d0f3395bad8db442c26752a675015211a0d49524f2b5b7b85c1f9c8e7364be0b095b2e194c04d6d96d9ff030a46f5b6324bea678d27d1050b38cb09c3e5973b0ab783b60b58a9ea119f28ac728203eb444b743f130fe99d291195836be19c24690f26f850edb5dbb409cdccbb06015e7d5b3e64fbe31a18191989d85f9196c39b0c226d189504da902f5c1efebcc5582ac35b098181c9ea94aab515d5691f71c0a267ae4b52593901e0820645b57da3540e986b14e8c32d00e86efcf1192cba857bffa78b6dfb99f5e69bc7e1d7453ee4b8db6800f7b89535b79d1f9304d99dcf33a7aa7a1022a2b664f65e33cff13cc34e86ebad1ca6551fe7f3431f72cd1c4563232a084ef283ef4972c1aaa8fe182a9838cc2ab5d1608a577406c663dd183ee6f6a6c1089a30f8924e858c6c6987619aab0987fbc413108913b013162043ec5344a818d97403bfa8d3c61ad5c6c9a28c4059e5786435718b1327c4f202cbe956f37f99186d57373c40daf05f1a9bb4372ff5fb66a9d5f7c7a5c98e987835ab8650a5fd68c9368146ba98c83a780288e9bf26f57ff4cef3c06569c21d30c64767834c5a7705e4e0c916703c8cabd08c787ee5edbd6fd90d3f56a0ade79fb11a41ca08ba9b4da3816ffa10fa14fbd76aba422ea613760295157e33e2b20fd4621f6ae1a6d39873f04cbc88eed0e313935de7432f55fbe09b9bbb110011fc167378229898b128e7f47dc6dc48ad759bbe3338405c17690651a14f313b540d9dd3639decad86d364d4e313fdaa7e221c862599d9333d8d307611f9c5ca8bf96920697eac2167308a1f22cb220edb497d978a6f7dc0034f7f85eb9858a44df554d8d622fa31c0bd915ce7b57b77d9141cd832a943aaf593f219310528103c3fcb316618c3b649d43faef7cbaf0f844eab8095edddcdcf6dd56e4d507e7d5f7582b90c9fcb7e9f69066e1c6cd8ebc6097412bca7965c44142ef5764c8252d8b015dfb270ced3a39824e3da0c11a6af451d0c59a6c5ecd5b367bd0b4859a291d896e38c285ac3abee0631f2337df0a70a3997bd28ec626fbd2c1ea66192f20e41a4b5fa0c3395898efcc01da101c77eb551240be337805cde29cf637ab7f57467fa16bddac3ae58b15eeced6931098093f155afcb9ee3661b081b7102a701841cf14f2dcf13e955df44956d702d95bbc8bcb8a6542cc8c8bc50711365eb95dbf78fe0bab478ccb2873fc8f8ad4201a36620ed11a24caf032147e8d1ed7bde55f15203ced96a4c45258b8467b691947ac4a165dd70a0bf7a5ce5c9c1a7dfc62fc83bfec43d05b741d22d7d42146b534428180ff9658544e3d5da92824430caee91f2cb063c41afc99b4c6b36503de16d74654d0ea86c357c788fec88a39baa1fe4f0a35dffa42db60b35ae78a310931b93607fa3b21ed9b87ae1c6965bf69a206e52a11b0a836368549ef535179d1f9edece1c33e0f9f005629b3c16be4fe8662fd093ee149ed2ce005f98c97474ccf04cd1b558eb81d6245a47736a214be31855f610c09071717bda0b54d8faf177e597af4f3770ff68e4357bdf507f713f92d23b35db3eedddf1afb1cd37efae980a77b5c218fb118f36b9d8e0182ae8970f403a838bfd50d38368f52bfcce98fa007278779da80da7d9769c9d7e2d801c8de4e40ae4899207ef057c0794e4b6ba10ad5fc56e42fe0807c05819169bb9b8499901ee31f49745017532c248fdbc0ade0a6a25744ae73643bbb47e715e617a2138a5cc66d25eb50abc2687b738b31ff9cb6586903acc739b8c7983fff6216f089d4d6d1f6f31bd42bee8fe326f47bca495810e48b266c778974a1442b86976d0b3be808d02b39d88294676482865fe2d833c2a39b401738635bd6ae0231705a487305dddb3af2cc42a8b39a7e811f3ed52edb439acf3e12fa8f8b2b66bed00b9894caa5b6a0688f7e0f699815146eff5e3f6471812bf49bc1a7f1447efe523ff058d739cdba9694094a5c64a4933682a0096e94f29e1b6ee8b84672da3cf1c17f2c2a546b09163861d0b090eabc23ce54a9c8a5f10525ac44076906d7f7991b2a08d1a7ae60fb18f468950c6589d5c1f698612917ea49b66495d03b0c3db11fc300e39ffde962c5bcf401c3502f2db3ec29e823c0691d911ef85e82f97ffe3e834c92ac496f20e46c809bc976a49669721ad46d3a4b315931d516a426339354e9760874b8a612f63c50380657f99d7eb3501cc29f9372f161816ce1ae4d285e31e279929dee2b87ebce4360ef0b632990506d310fb8c67a6a6f122959d8331623ff308bebd105b36ce9d2ec9538a8bde87a5808a555b0ceb931c0c57d139ac1fc576e55cb7e58d4a1cb24889bbbcd670d0bae65f1eb2cdc4870810e7962c4bbe21115779c19a5d90fae05c621f1bb5b4c741f9164619759802fd5c887fea540f218077c94cee430a750dae7f1c1784096862a5158b75456b6827a27aaef0ac9d59ed1210e25127035c745ad148bedb8474a7abe10c866fa3f86cbe49b51f80bf22cd830ccf0e81de8ef71acce5520f9047fa376dc39ff9ce9ee11f57512392ff9910df3b6be9b58ac5df9968fe6d1840b9bcd4d132c0a4eb174474dfc6f9ba0f9001679f87af42eb3ca0ce07cd04cb5049a4b8d4f6517d44f83fe32c2f5fbfd0e673264342b9ce268fa42392dc5a27c7e34e3cba258488f380e3dfe633ca5a3f9d0833823ef2e43631096535bccf67305bf0b09a5d2a6bc20886936ba1c1615da02e98990e4a77f750c1aff8f8891a9eabd54b036da2c90589ac96c74f1fd43e01719d641ccf10568f9b7faece6fe0380668998a8896aa5d898a549e653f178a96dbff81bef5b6da102e71ecad171bad34423d3bf19940b287c4a2fda064d6ce4619ee5c726f0c8dae017b6a0cf204cb5cad2d945e2ad2c1551e6499467704b01e62741d7d752e3f58efa878f5ad42fc86aba15ddf0650fee8c88f5d431069494ef2bcb3685e1c732b347976d1a2377dae133e6e3060c1317125cf6507b1ec001973694ec60497464db4d74e7695f7e07d2c47caf6574dec77fae787cb12a9be276b99c61c54e74ef601485f800ab58f4069312cba6d09fd4c0024490742b38557f5198e7c15a928a73e958322d35a6764743e942676f11ee872564d7b2aec5a92b3930813c9d998a3b876ea52dafe5ce481fcf3bf28baa4729bf250e87b8ea5aeebc4506f15e0c60fd114105b3cb0e7457d3b1e348cb93ee220160719f9d53cbf8bbcfed64f2a3a93e426c11856199d0f402f3575e8080e8b0b8afa556001254bd4767982a13448ce049e80b1ca806d13c3780093c3e09766fb6a5768a14b9bea98bc84af060aac734299d81b1f94ddf1fad4c939d70114e0ea7d2e9652615a4a311f0e7cb23f4f288bf1b4b8164233fa50e12cedbfd336286d0c2871099590f05c68b2b4ace213cc1062e5f4e452627bfc3dcb87c040ff926be00f583a1a4691a708b9dcfc59469d1a08f24dd392fc2a0fc910e0fa59a1678b6ac5e3e1a340db8ab7814ef49c81c20c0433a7a3f1cfa8ce6eb88d9cbc5a7b839faab5ac199b6517043ae5a45eecadd8d1af204facea965f85227adbd8e84da33db5d1bc995015dd164969ab8309ff92190e91339781110a8e29a1a054b4e729ac76e1ed941a0fada54c04863e98aa0ab9586a3598078ee45efe102dfebdd223a203c0c8541748124537f54a9df0c6d7e7186e844f7fdd27d4ad1defb99170153f3a3df2eddbf7df5647394665b0165b5489219b1dbe0777c67d384f73946d8d25382cbcc760966f6e5ea21cff2c26d0e601d6b3051c3c92ba47b904aacadb5b6da6f65e2da15537b0a52aa3346b330be2b2b190b7c959d016391a9d2ecb6b6c800fa339d3ab70945e90147840a7563fd9e1", 0x1000}, {&(0x7f0000002480)="216d2255aed99be1147baa94d0465161193b42958a0d74eab35bc3e21fb46a3c5a059744bea8efa1abecd221d9f2d3d504dfafee023723037f27e3275f882b24fec454723cbf5ac96c7c56df94397d8d2934a512a2f27a38f76c5d14fe50e04a799c0ba06475be2ab1349d5aa1d4f8db47f35334ce14f0f40e8fe733fe4070b09a9484317ac3ae887f11dcdf001144c00e0a9a25aafd47294c7696d758905690fd1b2615cb2f4155f0bca056a9ea23c168927df1491f6c9742980b0ed703ac2ada240e6f022182b4662791274fa3214ff0b5f917694ec3b01e0c753489c68a3b1e9b2901e97d2cb81456b13a780513ea02fa79de92eb59ec56565d213826e56dcefc51c04b50e7af691562684beaf06cb27ed5d4095abbaead24de8f88b498b3dd15780d1c4b4d747d1f3cd33a5acfb1cd020e5020bb01353ec5f5c46fac84d00014a9bfcefc91c56b0b92a80a402c6cff887bd6c0315cdc8c79818a87372a4d36f8a35e9f3eafd4b37b0b08c9aa6a7baf91a0458609d3abd88a054c7bf180e371a77975afa9edd941b2cc9c05badac1c677abcd9a2547b788122aeb7034a9c6d21999c4c74a51fb1176220db94525719faac17319516ed9fffb75f9c1df822a4f4df07c01761e2a6a663b0fcea058c372ed6946644cbf737a3e0fa996d3f3a12d1789a73459075521334cb0717bb36632fdf1713a2fe2ddce70aa3a5c1322d1f17ac1f717e25fee59e47b954ef5ed45043d6bb063685013d539fb643efef00ca945391d8d0d05858a09928585e6e13a0915c51d38c07fb9a88801ef84cb11980b9e18486bbee9cda0ab460468a9803040947239d4473d391babe47303e6024ed921dd2aaf5c137058360d51d0d8994388788e2d6b14296330775d2bbd34e7c58147d55fed75aa7f97aaabde3f4ec784c221957badd2ad6d0024d92c2530dfd8851beae88df4efb334def1bf24edc2d3fa772c3cb6418349686f8a205d998f9e0daa4b26e8d87c200e9c5617cdbabaca1ea381be91f26a2db95199394f3f50f258f71ee30912d53f84571d10bbdb9301acc73d34a47c8253cbb80ba441c70bb1825913bda17c85d3996ec754257fd48b636f6c0a16847aa28665f25398fa1fc76a12ac6ed2e51cb63b67614a08604de0c1a25ca36b50be2392003d1c4b1022ffd53316af9a6578940f59c37af498b61493847cab34c4aa4023fdcd7e2a4d164e2d9255be8abcf6a64e7e132f5df353b5cdaff6e7846a0bc09536c69606698c76590086ab79b988a7826d4a98858ce5b5c78f97f61befaae8458c6e1b8bc31ec7996be064783f98187892cde36d529dde6101f1ab19e9d17c2db56978f74f3af8cc9ed5cb7b661536c4747754cbc8153f30aef0d12fbc8354aa13e9d1084e9d2b71971830687eccacdf85ee9ef42900166a03c4b80950ca5d51af438d4465a1651b468b047ef57894caedc5c4ea6639ab5c43cd2c0e0f5684cf0e678833a7aa24d854ec6d8034aee508c275c0f3f1be92b4d35911ba915997f44a10f81dfb027af84b4991fc72a5df57dfb1ae98ff7755fdc7a450df7d424604b9941f8d176aa78f547b948d93b304249f931b7417931b8c9942b826dacfccdf4ae3512e4e00fcfa06cc83f58af096c927bb4c7de8a3958c1f1ebe6bf8723030845ccf639725f053e0d702b404bc4b948b91d6c40bf35f5b0ba5b5abb578b3c35c06d0ca70a076003f2ba0072c5a29b33a073a509092c3705acc41e0a4ca46ae219f3489b28d9686060ede47d93b1bfe56f12820cfd35f4c0613c60315ebcdd1b8b213cfd302863254eb52adf904fe3c426e677b07389cdd47824246023fe003540e235d79054656a769b5048a30642125253442ed72c8c75a137d60eb626737a2fcc40043e31689d2b762238fd0deee115accd4b1fcae42367fed0d7fc671986283f276a4b0f9af108445cc6451dbb4cc0eb99e62a3ce17fb5f921808363a6916f22c1693dffd2e3907159e42910b334194ebb27c2fc29e57a937d4b3ad83960644075d9fe879c9a2ceb923e38c818af3fe516719acb875d08b0bc75f44e049610abb24db5a21f1b0c93071a15bc732d78c63c41d256d25f626b00b13edef5f7ad00ea3dc07ecc4f70545fa694b4cb3f8ba25c06e6755261efd29cdf1c280a86d727dec3f976ccb2d6b8b069b2e645f9e9c1f666cb2638723ff053c7e7631b3a5ba65b6b6ff053bb313fda95c3803557aa9daeb73b936f54679987188dfdb770af72e83df19e35fe08b7bfc2c9841ed2540f0ee005a4d1a4e8ac1b2eeed228d319f9f31761718f4cd3bfee38513905e5d276e59af4f0c3c7c3d935522e821ea1c54416148867c220469dfc7c6ed3ffcded7e51dd42594d1874ac1a6f7d98555443ab2e6b4f7270ee8e8a6aa7a93056247b5dc06d4f4e6416df3ad37662a913b1e128b68d2c13b0df064199f7a636980dc7d7ed3b8f6e6ba3369f00b416d3b111528bab2c51a2074f17a1b3afcaab6171bc64df64772930be2b6713d6c9a7cd10fc81dbde65fbb35fe20b12e8a3d560c39443788e09ef27771e216417f4ff2c1656e668f06e3b3a2b3c4073296a7296bbbdf0bf64979705f07f4f3a273a2e625f4ae328c79093f3c0324c072ada3de32693318fe563f66ea73a78e91e8fa69f99f3910c4f0878ca105a037bbb7ec2964fc991ba66a13f45f77108200226616e56d8084e75ff6a5b4a631421e868709471051972fd3104ed63681c9cb292e24e1f8a3d72901ebaf6689e96d5177d5c897b0100e64e363395e8dd5f62670eb2b67e63a2c80b6cea3f6c5b2bf118941df5d49d7b1ee0ba0d9366c827c21d6b4cc7f72099311b3b62420c14807329c4ca17da600db37e188b8781e294fcfe87f40bd5b2224ecdec13b36da9cda55caee246fcf603037699fd58bc10ae811736b86b816f56912e3018fa341a254b79c4682c94dc93384b57408a9c863d01a1f8607a33cc983e8fd363d67654cee5deec055dfe01a6fa1f4ce305d392cf2b2342e4b4ca4e61f5795c20b9355fead2abf4c590895be9f7c76530eb0a10adcb631e8a0ce1f6150d1a66f4ea77274b9de3926bb918e476cf8a095ebec2d3fa3cafd61fe143e3ed47749b7702e62b17ef6636a1b4921e892d8844e05fdf14b9c5b3e9a63f9374f42847bfb58079a8cb87c39004a888e5493dbcfc30a1f345808e2faaa2cfbb28b38f97587c3dd9f4c49430e15b1ab56582767374788ab6a83effe0c21a2e5b759e68f2e0896a160c4d3c962ede288733828f8e1c402a92987d3407f5c2943825bc902f9a4acc5f0daa96374dceb623594c92209cabbc30d31ad9525d85897f2087bf89ed5e9fa24e49dccc4d69d226a89bfca5b9142201dd78a5bf76a9165a8fed75500b0032e7eeb5b7def8a3a8590db6ffff5873a7d766b402c7ab10dbf77fc4ea6602af6c174a83c725d5b545bad2b589a4b16cf559c66162d80204cb03fc0d4e502e0d1ee10edb1506edb936747563c2ab6afa1e4d5f04266f0cd64bb8332ad825f11f00a9d2906534867fe5f9352f6ac97705acf171c8ddbe350e5b709133cf2d27577b15176768e6903f13c4c49bb0b0ac35ef95f6918737132f6a94f2e812853e6ae1295d7d567d4d4b953f6c5bdc9f36867247657ea44b58d54c9c39fb8dfc685bbc75ce9995c825fba731a3349d4a2d2da19486dfb7f18415b5e6cc8a67cd92dea7d0b146449e2c655ff98cd3f5d80574e072473b3854c0594f9458e769e8c96c71cbcf72c55491cd960c5d83540f4d53bc43867fccee87543507bfaee5200842b7c1fca23dc72fc7a5b89aa81031b69d4bf2544c35bf864fe9024301356d0887e35a28ac304edd5002ccaaf61c8b3323d0a1b039fe57fba27384e67add2ecc242f7e2821b02b34ad6fa7a22af23d66ec1911579df373ce8361ec1e3249c1ce618cd43d1fac6019ee174fe750bdd87bd61a4f0f44174ef68182f7b20c104c37bf2c0aadd5f7cdd8d7c1892e106067658e3a05d3133130137cef0a3384a39c54b6f01b02fd32444ff16e31d567b88a2d1a5f02064cb9306ae5a6cff532dbe5c65b0796692b00a95c9edb2b69330ddc461d95400fb13716b84ac0c587f5d6d6516f8afccf83d15732f8e07326e8758638e40f0a49038abb1725f5bbb8eda2fe4a19b2b9fe2ca71aad4d878e5441cf71cbc4738adcc6049b4cb936c6305a21fd90464f8a3c211873296866919e3fb6ee712a99c0c2de5bb6e7f3aa82aa2046278b761bcf676818e70b93fb814bd66a35a273fb96322af95b3c00a4df7c8c90f4dcd2f363951ec3d929043e449197d6fb2eb3bb4335afe7ae64e4f2134d25a0ebd59ac7292a145af5083d06086b84b78b3655fc7b1ed4450494148ff4a3f646617676381c963857b7b0096614399e4e2aa1f1a871d3d0ab7742ab720f75a494dfd1a83f76697f23597f7eba7b6ba5e3881ba5e0fe27f05223d8b733ce038bb83c472736e2cd563e74524d7a2d2164950d6cd3a6f06b63cf3a3c8f852a7ea4d4dde1e569cd93eecba0a6ce5798f05637d02bfbbe4989b62e2b2053244408bc1f5250639534d1215420f0aca2c3aa2622377270c886c56485e8dc690ccf790828705af2e659338832ed748df008060e2f938b47c9923f2446904b30abab3020f672785a35e8387ff22912e1c93f778cce578228c547bfcade5a1912067155c704df83121de3ff75fec50e67a1ad0969b0e40fb64c7753186c56977bc1ec7348a383af65c424acceff79a203cc222f9633ad4d8f878fba99bc164b11e5845673500c8c2ec5e5917040c55569fb6950337d10b5fbe73eca80be29eb02766c877bf22b4737b09ec42fbe3b0b9a3051de63e32dca713d1a05fe9641e14157ada1c500e9c45ceeb9edadb80f3baab16a92373f01f830e7600a5d1f36965d8e5ccbdc3c0e2b239227649a6f7ac9fb8e8d771698657fe5b3a4f608381c6bc1623ab9710c83970d7d880aa6cbb8d1ccdc23a7fe8e51645a0d2f6cbe434a137a49b8f5a8fc79bd4e01c7ced63ee3edc2befffcff92c246fbe512ae4d0b4ed259b7f989a98a6eadc1a2f178c84e07b28fb3a4b10cb575fa1cd2d0c25b2ce6b8b8fa4bcd4f814c6a1b75a3df7c0f67ea830d0db2ed7e45f79afc065f962a2b8a56d338bec08b107030e32cae7e62e51a7966871c2d6e8ec047720b95792ea5aa577d389f011f42bb7d1501432575c656f10ae1f13ef7287133b2cf631934525f8b9ee45814c55994028dc3425db6564b5755814bffffa2db6cdec29a9914fabaadaa370f271700aa27a750ef2db69eb540c8f9f2bf5457b41c11ca17c6ce3821fdb982193f52df9cb6d8710b247c9a59e5ee83af4e8e69574dfdd8fbc9d1083856e75ac88cdc7046ce17a10e286efa087b133a1f3d7918756b878b7ded016f470206c7bc0c51f325ec30b634026e0beb851d522fbbe0fcef664878660ea12a60bf4f3cbcd98d0078e0761668f548a164b8c0f049fce31c9433bda4d8261c5aa8963e3d2eb96ff6a02d83347651bf4d1d43865ab57d180d394588a2f78cc87bf710eabc5e110ce1a0ee3b7171844e3d1d4bf476a24ff51ef8f7e47829bf7c0d3228dd7e2a94b2a69e35686f9e6282e6561bbdce7a9bc212e91b82486d0e63f17830dbbc04ac0c7bd1a1526042db6e7a64a6070a6a8c47cd0c916671d1ad47a1a12bbd4075e4ea9b9f64aae5e9c5f1c9d2625f97f5fc5d1d50c9c0de7f91665ddbb1939058b02842c64607eb76a3d91cb2f9c17982078ec4a12e3e11fe2d6ef55796c7ef165b3ba43c079", 0x1000}], 0x9}}, {{&(0x7f0000003540)={0xa, 0x4e24, 0x0, @ipv4={'\x00', '\xff\xff', @broadcast}, 0x6}, 0x1c, &(0x7f0000003740)=[{&(0x7f0000003580)="33c3f780a5cee531f31c782e74ce2f72b3b2ae15a761fd2c4bc4f5c1bbe71c6a94666905a95fb8c8b0719fa1d57862b3cc17ba6f285303171719091a6c54d10044febb5deec37016b35848b0baac5abb90397070e73cefdccae6af6f60797b1c6339c5f3a7eabc3e4392756904", 0x6d}, {&(0x7f0000003600)="ba8eaba419a4", 0x6}, {&(0x7f0000003640)="88e1d3d29446c8f05bc4243cc4830b6cfe37e7adeb7f7e9632accc343bce7e9718108c1d1ee961b79cef34a11a2b9975763ae92ac0dae8f7d1d5bdf52341dff1e1841456146a7bccb55f11fde9d132c35e7b1b53132df80485303bf9", 0x5c}, {&(0x7f00000036c0)="49bae93ba5d20c0a0991838fc588f1d4fb4cc14361090baae6aa2efb05dc1aab2e40589b91447a8c", 0x28}, {&(0x7f0000003700)="e88b15a8c4b3b1678bfbbd16f1551a596c5f825ea1cf039ffe9ce97ab73f7bb61bf72172e0b33a657d72ef880f4226edc0be22169698c6650fc242f42378ec", 0x3f}], 0x5, &(0x7f00000037c0)=[@rthdr_2292={{0x48, 0x29, 0x39, {0x3b, 0x6, 0x0, 0x8, 0x0, [@remote, @dev={0xfe, 0x80, '\x00', 0x2d}, @private1={0xfc, 0x1, '\x00', 0x1}]}}}, @tclass={{0x14, 0x29, 0x43, 0xe4}}, @dstopts_2292={{0xe8, 0x29, 0x4, {0x2, 0x19, '\x00', [@calipso={0x7, 0x30, {0x0, 0xa, 0x40, 0xf801, [0x24, 0x6, 0xd1, 0x2, 0x4]}}, @padn={0x1, 0x2, [0x0, 0x0]}, @calipso={0x7, 0x20, {0x2, 0x6, 0x7, 0xfff, [0x100000001, 0xffffffffffff0001, 0x9]}}, @enc_lim={0x4, 0x1, 0x8}, @calipso={0x7, 0x8, {0x2, 0x0, 0x0, 0x7}}, @enc_lim={0x4, 0x1, 0x8}, @generic={0x2, 0x60, "349c1679b4ce86614ca52ca60ca97566d4fe36106af890642546b33bb70b4ade7ec08ac4da51a5e64c8499fca2db188e22a434c3a1f3da46bdcdde3ac8df60cecf9ab4819b2f2cc5c9fed84155b6a28fdee9f0e0edc3fc6aae39ede4746c157c"}]}}}, @flowinfo={{0x14, 0x29, 0xb, 0x7fff}}, @dontfrag={{0x14, 0x29, 0x3e, 0x1d85}}], 0x178}}, {{&(0x7f0000003940)={0xa, 0x4e23, 0x8, @remote, 0x5}, 0x1c, &(0x7f0000003ac0)=[{&(0x7f0000003980)="76e960", 0x3}, {&(0x7f00000039c0)="422516b8e217da13815499ac67e749e7b03eaa12772f9b926c562821027f65033ba56bc609dd38849c56ce1d1ceedda75420a578565014419affd534f79ec81dcd72e94a5fc781ec4825fd33696f27e15906a04e4fe0085d12e0c5535dcffb79b0e9f31a7b16fceb6ffc1220b54e2ddc9f50e923eba78c6baa659878861f9eecceefbd7d5386773051beba0266d0c51b82e45f9194d91f073bcd4d594887922d99905f35922b06908fb0b7e43556ccbf015734e236d22cc8272315400e56068482c89c74a1a7be8a2cb3bb8b52f0ad3da277e6f1e8f59bc698f1f86f8a539912b5c4b10d57e347f4eaa97e96ed5a", 0xee}], 0x2}}, {{&(0x7f0000003b00)={0xa, 0x4e23, 0x2, @empty, 0xa00}, 0x1c, &(0x7f0000003e40)=[{&(0x7f0000003b40)="1e6d0bd3542829a7a388a86a271c396f1458ce7e7ee2ae164e6c556b53147eb66f5b56b9c3a3fb1367d60980aaf6d86809909367bfdf6d54744926cd7868a2dbfe4d7c9a1df872cb6de5badde16c804c2e2cb5a0afc6e996056d361435a4160a8e41a148f097ca0a9fff52bbb4975d9e5490e7a84a05d15dfae07b9c081bd0f7204864d9066167b8d408e2dc95d73ade3d", 0x91}, {&(0x7f0000003c00)="975652", 0x3}, {&(0x7f0000003c40)="94375d5329768661d465af9a4de0772f688e3f3d7d72a18106a4f7e5c276575cf10a49e5946a6e6661640de5b5e2fd398b1a5c8c108dff60b8f8fc56b3094be528f85f9357fc163f23c50aa55d934372981bbd9999d1f50330eafce56cdf905cc42e564ddf0ba8f0122fa828d7105c1245dd38bb8b6c5a1d464c280bcab9868c798858ea19a31d4d45b3f2eaf82b28b7287544bdaa898f20a0f659c9cf90c0053fb82a3bacacda8a373aa5ac5961c7971d46f9526b18c55ec1f189601e88844ffc95f063c01af89614ff885f0594a87d9dd2a6a2fb3f", 0xd6}, {&(0x7f0000003d40)="064c2b3d35d1a8dd8a4050ed2bbb707ec54f9f2676d550308fdda31aa62b18e72cf7d4d3510333cf64c8defd3712a6dc12e4f15cf21b90b69c946f33eecae331d0df5f8440faf245657aaf398f9357d034809636ac9a42047e76f34a2de2f585ff51dcd4726050bf024bae4bcb8f66317e147e3b129ae0be47009b60a4555fc7b5f3a335bcc8d59d9fab4c4a50b52b59fcaafd3da0c1c62ef1f67d71f54bb74619b42af17385fdb90468646e47f8ac0df48aca82b1ab6e82144ff0dc1212a7a0df", 0xc1}], 0x4, &(0x7f0000004140)=[@rthdr_2292={{0x48, 0x29, 0x39, {0x47, 0x6, 0x2, 0xc8, 0x0, [@empty, @loopback, @mcast2]}}}, @rthdr_2292={{0x38, 0x29, 0x39, {0x84, 0x4, 0x2, 0xe, 0x0, [@private1={0xfc, 0x1, '\x00', 0x1}, @rand_addr=' \x01\x00\x00\x00\x00\x00\x00\x00\x00\x00\x00\x00\x00\x00\x01']}}}], 0x80}}, {{&(0x7f0000003f40)={0xa, 0x4e24, 0x8, @initdev={0xfe, 0x88, '\x00', 0x0, 0x0}, 0x1}, 0x1c, &(0x7f00000040c0)=[{&(0x7f0000003f80)="b8f7bc4ed73ea4bcf39bc0fc1ea8f87f7f18f3a7b30424699f", 0x19}, {&(0x7f0000003fc0)="11b1d0492a2b64c8468bd8ce87616052a4ec41ab7ac2306b2c2167aa6a233f4caef9fcd74d8e1f7aff1a3a52f3318c15df0c9946ea6b7546cf8b61b895a97a432c55da868c88387a6ac9d4664b7ce82e181b9b2a12e60629c476504084aceb3164dddc29bfecd2bccf50e26c7494fdb3d177d68be8e05bac1b54ff690deeac1018169179f16da8b0a50a4b3ed4ce96d367e1a8a6ecf351356dab84233bf21bd29b580803ff5d2b4f67778fb688c7e2e45a10e79df9cebff3ff12a988f795", 0xbe}, {&(0x7f0000004080)='l', 0x1}], 0x3}}, {{&(0x7f0000004100)={0xa, 0x4e22, 0x100, @dev={0xfe, 0x80, '\x00', 0x40}, 0x8000}, 0x1c, &(0x7f0000004300)=[{&(0x7f0000006080)="aee88b3a78a34f78d800aa4c9353c32ec59381f3897debfe1eba32de4317fb64984477b5993c6543f1249032530214916423c179c7ceebde6330defd244b8e3deaa652f271794da5df3ccc1cd7ee49ea0209ff9e3f4fb960c8c982bda5bcafb4ed15845c72c660350998dd4aa7f533d7bd9f2e158d4fba3159660c9fcbad8acf19dcca37f9a4d6d5807e9f3196de8c16", 0x90}, {&(0x7f0000004200)="a2f150e619190ed70131476b21b8cad5e1ac6c30d3db6e3e6c1264c4dc4481d4d92fba8886506eaf541b687d084b415218eb604fbae004c8b97147a55353b588854c12", 0x43}, {&(0x7f0000004280)="4a8d11d6518e58924585060db33d9afc7d939e3c69d0d27904ebb4c2a726a49617fdd09f0302eadde097c39841af2c849b1ea799dd5b15b9cd6506e4c24d85fff82c65e65142e9fea763eb743a6425d011de6687e9b7", 0x56}], 0x3, &(0x7f0000004340)=[@hopopts={{0x28, 0x29, 0x36, {0x33, 0x1, '\x00', [@padn={0x1, 0x7, [0x0, 0x0, 0x0, 0x0, 0x0, 0x0, 0x0]}, @padn={0x1, 0x1, [0x0]}]}}}, @dstopts={{0xc0, 0x29, 0x37, {0x67, 0x14, '\x00', [@jumbo, @jumbo={0xc2, 0x4, 0x4}, @generic={0x5, 0x78, "1bc18d4aec17f5a4590702e1bf57d7dde40cacfc88c5a0659e3c2fcde1bd37619730ff764056eefb783e5e03226775a0651e4193b9b9b18b0d5804914a48f759614e40f8c0cb4d2a93b6d9a5d9cd42bd2ba2da4aa524b60dcb8e558d2e11a56d0e25c57c56b16bc2ecea51647cfc3ea20321d402399dc76f"}, @enc_lim={0x4, 0x1, 0x1}, @hao={0xc9, 0x10, @rand_addr=' \x01\x00'}, @ra, @enc_lim={0x4, 0x1, 0x7f}]}}}], 0xe8}}, {{0x0, 0x0, &(0x7f00000046c0)=[{&(0x7f0000004440)="53fd5f89addc7638eef2a05a0c1d4a88a98dacb4cee3c24021b591be0de5fa732cea65b247aaeedd2e15056e1c64eb1ea06dbaffa2e05b72e51dcd7d7478d64ce2dd8675cdee3bb7be9a62fd1ec35bbab54adfd3eb1bc41d95ed3105ed985afb538d", 0x62}, {&(0x7f00000044c0)="d0fad7888f66de898467003e8ea2f49c9c1fcf33bfa040c794998df539aa496887cd9562b2e154c8e22a672e6608b86102d57345965673c6db21230672a9f11009a282c49c2f448c44817172b7ddfc1a4a43aae43db970a7062e7dc4dda23b747db9e3a477", 0x65}, {&(0x7f0000004540)="8c6ba3cc62c973bd7f7c6b118a846ccd432f5e5aa15960f6c4a8f09391898d5c357a3b0fd96874fa86ac6cba8965bb6411215106fdcd0b23652ab882c963a8bc4898f83a542fa7f4ce94410ecf9e0d8be0f9b9c716a01179aa274414fbe44b47943668228005ce2f50f53030460166bdd744f0e659de", 0x76}, {&(0x7f00000045c0)="3411fb6a69dff242cf21daf3b3d7042581803cc4a8163ab8bfd84e548749a3bd95c80849afd82b912e71f9232d095e50aeda6e4b27a78eb508031fe5287f50b255ec3539a4638e4d9492b95a1ffc8b528574f84e271bb8edb33b4595edf1703ff0f487ef62096233df247450c9af1df76f796d9ba8d84aee76e6a03aa5e86d01f2c8d381b5d53134511a94b982fb43765e783a7b54f09e2388f3696dc867b0ec4d5d2848ec1b9cccb96fb4489dd29f54c7cb580832efac9ec833f4c06576cdff18f6b4b1e2c70aa509e21c532860c1fd0a0385c5868f4a96923d1cf062fa354a5cc252b33cad33505d4f6b32d88dfcca0ea440539d", 0xf5}], 0x4, &(0x7f0000004700)=[@hoplimit_2292={{0x14, 0x29, 0x8, 0xfffffffe}}], 0x18}}, {{&(0x7f0000004740)={0xa, 0x4e21, 0x676, @mcast1, 0x8}, 0x1c, &(0x7f0000005980)=[{&(0x7f0000004780)="1e3a7c60b47a5e6f72dd1fb43385416966fb09552c66fdb67d9908bcbe25045a5c5ed574b34c2eddf074ba440c79d3a870d9e5ca60ab253b0a03b719dc649a766f40d0eceac609440f5b4b46df14a868852468d80d345c85e1b8f44fa1bd2f33b279ac2f7f17cef809d379315db2eeb475422e42848cd0258874c27c69cac246543f03ea83110cc12b1cd01dd3c88bb65a7126296a856092f47fbcb3f7bef0bfe25e3aaddc41d68c67874dbf332d103bead1bd7fc010aafd346e2189cd3822cec0bf8f9cdd7be0ae7b9257fd1bb53a5fa0e664207ab2b56d4dd71f09f062afb7f1c593ce51a4bba135cae0fa1de757a64a6f5fea440d2bfad001023a4f28e2450ef41e14919ec3b623e7fbfce1f6d5de774de1712f48bc6c3e11ebab952d3f097f060c82ff54aa12b261e44f51f58e27db622c04c5a97d5e24f1da771b516aba024fdc8abf94bd20a674e3aa5bab85756a2318d4126c474609b77286fe9fedf7bd0c6d0c96a3fc65cf91911fe0cf2f2956df830a7f2ef23d43307a07e665b50d999dc27ff4da79065f2dd750c9fd5d7ba0733b8e3d0985050f5338880a8604d04b486396695a79e78256881ee7d7e2650fac0d57a54cbb43280f6097adaa14b07590ba2b80b2d97d7ee94353be1f53b78842bbd3133cf31b2acf82f520a3a9a3cd694e140dc70fdeb63af3bbd6a9d263ff8e973223197014aa68b98d99e61b4b67afe52cea5c04dfabb3c80de0590bff993654e0129ca1497ca97646937434f2cfda4817bb8b0b0f1f7bbe1efcf27daa3b5794084c0df8eb654f49516be8659411cb1f8ec404f119a5f92fe1085a6291cdf37b5095f4ee4073e60d418600cea4afe684cc4e70dbba23286ca50c400853f2c6629a72935a7afed74e5045e52a8b872b533757ae0be2b2fd758781ba0e25e77e67222b191584091646b3b56474cdbc70602655464d95615b99179cb3ef8af5b42171cbc6e05ba076a52adb170a94dfbcdc8c6f7444c417a141be499b5e89f3a4d33fb851d861f14065ad1e8065a7f68b2d3ee420fd1f37fbed56beab590415f392b8996d5a035e72edea7ac769b5c1393c7374f9a6d89bf727edbba921061d252fa53d9459384dcd3c96450327cbbe28d3315cf24b1ddbd3a691fcced3d5d833fbc2c971f959e63f85ae81aba96e70cf26e5ad08af2eb5464a6738c28eaf0371b6639cefffc11aa8347f0aeb9f7086ec070dee59ab12286ee3ee385cde46f29057b4150bdd1acb0ed301d34b49885afe4861ee1da60cdce3279f45d30d05eeb43019e0d40c2fd48c7590d8ec751074c40927b6e83abbf8ecc1a15f3ad9def1d44573929c8d564cd3e298b689aaac9763c0b714c82bafed329685fa0b40607462c65bbb62b288f592a82723fea9f827b523c7193258388ec9912fab7bfb061b311a6a4869250d726f83f7dcc27ca54c788afc131355dc59e791d4c15d2c7961ce91e1c0303dcee7fc8ed497cf5a9696d750af581711d8229120fab599bab00e7afbe66d4f0e75ed5ec9c734a572a13e2747098ad850af93d9fa5c58530ea9fc070149ff1d1c59966fd3e44d540ef55fe89ea0c0a1cfaa74e2b514d916ea39a7dacd19a990a227bc9eb476626d4eae213aa7133f201afd968a2dcff350f685e0ee07956fe6090b5c26349387ab671ed00d43f41ed1edb228046e7eeb3a7e4203f0149f1ca529ada2052e48a22edf88c575a094e46d57b07fb979eed8bedd9c72f9fd334b6e422a873b82db6b2496c88bc8991d09c64464866b192ae2312391778fd74ee12fb6614fd57d46931f48b8777b576fb196fe87680b8620aacc4c71596de956c2f46ff653fe126debcbf20b353d5b015604f5546f29bbd770f6594d3b6a17090131780319883e3e7b1b40b13b3cb9df066b6e4571dab487738ea98795d0e0e176e5dcae2280c37523c209d0e3387b7643a5f5b850be5d8e8d6f5a2b1a8e8edb351ad91d78b179fb1701e2a9627a0c290cf4ad7efb6330b045364227e11cfd71843979003c1b39ae7a9cbe665406d4b49d1e14bf51a68d6d55165a6b908771dfe41d93075a5c84796a8bab042f0bfd02f8c7b7511c85f367a48074d1181d29b69fc18822bf53ba4c3a2247379bbb5aeb51f17a1e0daedbc8af7829847dbce17af116f518f1fcfcf2165742985580b68e371ceaae4e1077dc816cfe398cbceeea0c3eec4b5a4ba33daa5578e3d41a24fe1eeea45c29ea840c782c2b666efb991663088ceae71851ded61aef80035e31303b596907ccaab7eab977225be355a701492c78e0b5f4d1914aef34f6ff20cdf4f4ccc23695b35585d2503b6a9e741a3ba68073ddb5df93f55ee806f1036b3ee6f346f2fcae98ac2e94d2b1f0245e6079781a6a9adcc42b6456b7d7804e2925874ad4e80cda367791028945ba6e564fd3d2b977ab2f248bdbc2640cba532e273180b4c708ea10715d8f55c453defc24a7939bef74f0d0fc56d4f22f960cc03069d6b9be1eacec23ae9c31a82726a250c9d25afcd645414262731797fa92c0b0ff4bdfc81e1d607be72d949a45886920d23360c98cb47d58190b60f5c3af2d51d25763ef6c04d33a7cfc87f1ebddd14d706c61182b98acc6f0310ef4b72e0719a3dda86d135600423310f9ab6610daad135ff1bbb9dc4af14b09022f99d73bc4589748b3e1ec3c157c5d97608e63e97690e864c31635a7204150c45c70a8fa7670409bd5e30670b703ecb735415b1a743f186c4a0c84edd7274ad583ec2ddd3e67e5e760db74914c9ef3003e6c512f0b9c6d26830a8d9eac9b540b5929f3eacc9279c872c8698cb2184b1c8f2648e20a5c4c5ad13fba630c885e71ff26cdea26319c401aead72081418c161910b30932b4a250bafa3ae69b9e4b5a98fc2efa57eee8e7194d7dd2af9778ce822781039f951cdfe7237ba1d9e572d898ca51866b199aebb2e512a8e4fc9a80320ee0d9e3f093e086aec0bf20759e43bf8ca637bd9f61e78026da737475685b9371730f810bb9d2c15f0f52f2598b529200e7c669e42f4aa00ac9542672b8073cbc56c525d19684930e2d8f79bfd1d67b35a888e7947f8cc1ed00b9b4986d7c2aca6c4111da79cd85bcfb584e8d1ddb3838a08d4bb099827316b73cd7909f6c306642d4df87935ec815d04622968d27f63e8c1e2300692f5c19bee3237eeb8f76f89d0d267213eda627a8fc821100510e939286c9fda33e5680d86692a05a0f6223842e649f4b06dbf6af0285d8f86396683ad8917c898b5d9431b1779d28911507537ece29a86093487208ae71fc8488d655a9352049e4e7d134ecfe683a372c77c7762dfeb48960a2b60efeb8efc9163ece1042baa952fe329be36ebca6466ae52ff70190f95ebdc5671cd6e88cff552631230ecdea82f1dec08a391fed26981a6907999c45f9d75487226d32d3adcbebe9a7987d3486f198e1670b7547c3c15807f6f853f697b85999c7aae5a7623358d285e5ee82e7e0dbbd3e8b5f87c84148134da1aba1553b5060a887e193ce0607a2072c76cf4740b452e5c3802eeeb55a2703af0b642d4591c1de25734248001ef2147e660de8d0ad975f006ec836936aae07c6e169ed4bda6eaf8edc75defc8f00c2954a86abe25087fba178d837d0e7974575fea20c96eb32dd4dcca981ba08f11f1047f11a28fcd52521f98e8babd1b86006817a1a24a13540755dda395df961eda7ca3fce7337c3384f8ace09dac94801c5e50a634c38193682b9173f4398df86e6e97713ac3b9717b44eb60f27d72ad1b92aacd1a66c7f9970acf886f0eaa1d5203388f8dfe9fc6b1f07b206bbabc4cd982b5dba34b15801dcaab907506890c4955a3b76bd93cc0e3d12403bd3a66754909fc14447f02eaf9717b219c9280f7c32d1f5425f05a4871b40e9cf8a5dcdc7d8c3fb63556d44cca59556a9245ee20e982e98251a464b757793ef6c126c6003eef73393a5cc05e6ec1e49016f33f7f9b76064dd04021645402d49a6453e63df4a29a84fc1b397aa922f9fcf755c06ba45df2eff2554ce555b395b1f795816179224359842d1f95836224a9c1f3e58815daac0d19bfc08c8e1d06c3c5de227c68cf2d11bc5233a481048ed657114a2da7d4e7068f87eea91d4aa4a2c469616aeaf8e490a5c1e444689b0d33fe57d815ff1020f94a2da597e02abe324fdb760cfd4d5bce89a95f7fa230ed0c8448abc0688417e5cedd095c6be83f024ec7f6091558fb8f83b25b0a3f6ca12966105ca76217a22b222077a67cf853d4b33322fa81978cacd97e72712b9efedb4ba740565440c49ecfc37d54f42ed5818012b54db628ec5507fb30df55b1ff522c5e4f246f171df146f620a914673e68b4a838198a459b59f2e65ab9e27d3fc4fade7458f561a71e80313995a2ac01e55299c5fb5c6aa0c7391e633f3c99e815b3e6c6c947e9f0d0545d930fac6ff59de06cd451e63c7980bab09ac97a556ae69cb5ea4bd10129eb36b4d1edcc0e97fbd4d095a18d91f5a07532a5b385d7d2c043537b627f6a1a972073244653f173c7f119ddd88e16c1c3336d81bcfdbdce5411e5fa4079913a330205dce4785f218aa928d8b81ec93a5cb950606187fd5c234bdcc06467bcca26434d33d64f2add9396d8cd783a0ce4cc81fa2a1eb13c80c2e35d40750dca170a9b7935247196cf5a95f7e429d91a2c77e9e5439b2f9fbf4a98091f0ea359d0483e34cc14a089b5ac8c92c3982e5b3b526049d12a2a8f88cafb45a00b46a9a9b15a5d3ecc7de10a386ac7bae1d53517dc65647df3eef816a42b30cabf7e35ed027badc5caa87b80a93c89f705b8cb233b2d24abf1c181001c911fec6fc7ec7a49c0ec24fecc5eefcf3a72f3deee7b8615f6702f7b10b60b8f41a53c3a0f7071f805b2180b771b89da65d4c4f484c2b3aeb43018a7a888ab45cbd691f3b32a23db7303ea1fc9f48194223462530fa6a2f69535aa6de8f735789bf346708cf5dfb9e498097dac2ca0209cad47ab6d5c4d7ff837819c615c5edbfb7efebc885b062efa27920a396f7ed0ae695a3db25988a0b3fae8e7656ea343a341c85cadf0ffeb54e341af736ea1fe129e1234579ee9640a993efb8151a32bcce94a25afd0765d82e7f05e128e644c0e2e4021a2c346732fccde5a6b6ccc45dfbb1e0f32a9d3e7afe2b97212c48996b6e07e80d86a9af4fdbc7d2b7ca7cdbcf78bae662cd6cc6a314dddb0590ff9b4b4c92976b553466d201ba6176c0c13eaadbedbe940958a16573005d227229b5e2917a09894b047bb0678f154955f214dc12cae1cde859317fc573fb517ae168ba2902888202caaee9c1dcd546524eb1e4a0c7043543981e9ab6723666c80f30a4921a4d40282a644ec9a5b8f458efcf384080fc71587e37df31c7de59dd5245fe5c9577cd35a2e398c7be48fc807b380b698b9c3defc8c9aa690df614686e95c7c7bf9d725d85da93bb544c9acdf81eddad2f279e121c52ca14e6f8e165ee549f48a517432849d0857acf6a19aa35147995ea34dfbb9dde1ec64e14ff12ac7624dc3d4f7195f21bddce38de911d20f21afd5ef21eaa8a2d929b9d12da532445a5fcb13dcc158293cb40267f84fe1cae4fa64d470105f1490ca5ee8240866348ea25890b12e5861ff7ad533f4da3dc6c68ef3c71ece81b2b358cc33a9c026086dc7c82256e801b0c03ca849013617baa79b9f0ff0649a667bcab10c99a344f425913deb6ebeda5f8859dae2342c680044bf856ba444847aac070495c29adf6952c8490c9489c926a7ae0bb4166e392ae03677c", 0x1000}, {&(0x7f0000005780)="725a14d768bae84499bd5b528fbb97a71e321c050fec8587c0f854fb2fc86ccf58232313532f0f64d002e9b128e5fc62ea1bd6432dad0ea752ad0c8f346e9df38b3873fe9b720010973ffb0f39c5597746a91dcd55c46ae947e6aa4062072e8193a76911e930147b3c011bfb20a46b", 0x6f}, {&(0x7f0000005800)="d6268a5528f585e113399f0c3ad37b5bc14a4608a86a42dd9134a96774356c2345d98b95390d4bc0439c067d36af24011e2344756b963d48da70d749f796976a07fd5f7b0419de30fe9662e4f3da46be21e0a68a2012fdc7f70c505b553ab07968770d86895c73fa1d46d272bf448f0b6b85fbd0966b7b6395c501e00549c3f7114f8693b21cc1d3a7f6655eb1a2ce122feedb0dbb484b4dd806effed2741d5b1bf0adaf0d36f67e5e2356c4ff2fc5b9e88f3c76b495b8658b6e4870f19634a6ecc8cc4888fe3308436aad", 0xcb}, {&(0x7f0000005900)="9ac94d9d22868b60ec510aa10601a340d72400314797f37b48a3a986ac76b52241f0122bc3692dd97e0eae6aded464b35b2763be5d83fd0d9a19af4a23beb7875f0cdfbd1f6cbb1649252ab851e23584a400d590b55fb2dd7fbc727b68882311148bea93ed8aa7e583709d80eb771228ca01cf5601d5928b885845", 0x7b}, {&(0x7f0000005a80)="f482c32a06a41039cf295e832cb0d32ef9756d12ff47602b893dc14d6d445451604684bc1f57f94e0ceca45d20a751e558466387b8bfc4487472d8e452d137eec82175f984d589239512a29d4f2a1cd3decdaed1ddb47fee6e37f2c506c805f64879dbda072bba47e45d784f923cda42b2a5274b5797b37e3c3599035ee1298db83940f4173625f6f57f94608cd333f636d0e514aa09df886a1d0fefb88671af71159aecfb69eb2fdc8c39a4f53842431ed4545ef6221849f0edd1d30ade334aa4bdd67f3bb6ee6de8cd0f83cbd7301f21a1a335500d6a2d130dc456b45e26efd6c0bd17ff7003b66f4a1d7acbac41c10069ba31583f", 0xf6}], 0x5, &(0x7f0000005b80)=[@hoplimit_2292={{0x14, 0x29, 0x8, 0x1}}, @hoplimit={{0x14, 0x29, 0x34, 0x3}}, @dstopts={{0xa0, 0x29, 0x37, {0x87, 0x10, '\x00', [@calipso={0x7, 0x18, {0x2, 0x4, 0x7f, 0x2, [0x304, 0x4]}}, @calipso={0x7, 0x30, {0x0, 0xa, 0x5, 0x0, [0x600000000000, 0x2, 0x240000000000, 0x2, 0x1]}}, @padn={0x1, 0x3, [0x0, 0x0, 0x0]}, @enc_lim={0x4, 0x1, 0xc3}, @jumbo={0xc2, 0x4, 0xf26}, @hao={0xc9, 0x10, @rand_addr=' \x01\x00'}, @enc_lim={0x4, 0x1, 0x20}, @hao={0xc9, 0x10, @loopback}, @padn={0x1, 0x2, [0x0, 0x0]}]}}}], 0xd0}}], 0x8, 0x1) 04:12:19 executing program 2: r0 = socket$netlink(0x10, 0x3, 0x0) r1 = socket$netlink(0x10, 0x3, 0x0) r2 = socket(0x10, 0x803, 0x0) sendmsg$NL80211_CMD_CRIT_PROTOCOL_START(r2, &(0x7f0000000580)={0x0, 0x0, &(0x7f0000000540)={0x0, 0x1c}}, 0x0) getsockname$packet(r2, &(0x7f0000000600)={0x11, 0x0, 0x0, 0x1, 0x0, 0x6, @broadcast}, &(0x7f0000000100)=0x14) sendmsg$nl_route(r1, &(0x7f0000000040)={0x0, 0x0, &(0x7f0000000000)={&(0x7f00000000c0)=ANY=[@ANYBLOB="3c00000010008506eb9afc4cd8d06e754a0081c5", @ANYRES32=r3, @ANYBLOB="2377f292252155b21c0012000c000100626f6e64000000000c0002000800010001"], 0x3c}}, 0x0) sendmsg$nl_route(r0, &(0x7f0000000240)={0x0, 0x0, &(0x7f0000000680)={&(0x7f0000000400)=@newlink={0x3c, 0x10, 0xffffff1f, 0x0, 0x253c26, {}, [@IFLA_LINKINFO={0x14, 0x12, 0x0, 0x1, @bridge={{0xb}, {0x4}}}, @IFLA_MASTER={0x8, 0xa, r3}]}, 0x3c}}, 0x0) [ 2812.120381][ T2603] workqueue: Failed to create a rescuer kthread for wq "bond1268": -EINTR [ 2812.470071][ T2618] netlink: 'syz-executor.1': attribute type 1 has an invalid length. 04:12:20 executing program 3: r0 = socket$netlink(0x10, 0x3, 0x0) sendmsg$netlink(r0, &(0x7f00000047c0)={0x0, 0x0, &(0x7f0000000000)=[{&(0x7f0000000100)=ANY=[@ANYBLOB="18000000560001060000006b00db730007"], 0x18}], 0x1}, 0x0) 04:12:20 executing program 5: r0 = socket$netlink(0x10, 0x3, 0x0) r1 = socket$netlink(0x10, 0x3, 0x0) r2 = socket(0x10, 0x803, 0x0) sendmsg$NL80211_CMD_CRIT_PROTOCOL_START(r2, &(0x7f0000000580)={0x0, 0x0, &(0x7f0000000540)={0x0, 0x1c}}, 0x0) getsockname$packet(r2, &(0x7f0000000600)={0x11, 0x0, 0x0, 0x1, 0x0, 0x6, @broadcast}, &(0x7f0000000100)=0x14) sendmsg$nl_route(r1, &(0x7f0000000040)={0x0, 0x0, &(0x7f0000000000)={&(0x7f00000000c0)=ANY=[@ANYBLOB="3c00000010008506eb9afc4cd8d06e754a0081c5", @ANYRES32=r3, @ANYBLOB="2377f292252155b21c0012000c000100626f6e64000000000c0002000800010001"], 0x3c}}, 0x0) sendmsg$nl_route(r0, &(0x7f0000000240)={0x0, 0x0, &(0x7f0000000680)={&(0x7f0000000400)=@newlink={0x3c, 0x10, 0xffffff1f, 0x0, 0x2535f7, {}, [@IFLA_LINKINFO={0x14, 0x12, 0x0, 0x1, @bridge={{0xb}, {0x4}}}, @IFLA_MASTER={0x8, 0xa, r3}]}, 0x3c}}, 0x0) 04:12:20 executing program 4: r0 = socket$inet6(0xa, 0x800, 0x80) ioctl$sock_SIOCETHTOOL(r0, 0x89f0, &(0x7f0000000000)={'bridge0\x00', &(0x7f0000000200)=ANY=[@ANYBLOB="10000000008000"/36]}) (async) connect$inet6(r0, &(0x7f0000000040)={0xa, 0x4e23, 0x2, @local, 0x3}, 0x1c) r1 = socket$inet_sctp(0x2, 0x5, 0x84) (async) socket$inet_sctp(0x2, 0x1, 0x84) (async) r2 = socket$inet(0x2, 0x4000000000080001, 0x0) setsockopt$IPT_SO_SET_REPLACE(r2, 0x0, 0x40, &(0x7f00000008c0)=ANY=[@ANYBLOB="7261770000000000000000000000000000000000000096dd89ad65dec810000108f80000030000003003000098020000000000000000000000000000000000009802000098020000980200009802000098020000030080000000000000000000ffffffffe00000010000000000000000e4000000010000000000bd00000000007465616d5f736c6176655f31000000000000000000002000000000000000000000000000000000061e2695eddaca41000000000000000000c001080200000000000000000000000000000000000000005001686173686c696d6974000000000000000000000000000000000000000002726f736530000000000000000000000000000000000000000000000000000000000000e4ff080000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000090000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000bcef5a18d4a047d60000000000000000000000070000000000000000000000000000000000090000000000000000000000000000000000000000000000000000000000006c0000000000000000210002000000000000000000000000000000000000000000fcffffffffffffff00000000000000000000000003000000070000000000000000000000000000004800435400000000000002000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000001000000000000000000000000000000000000000000000000000000000000000000000000000009000000000000000000000000000800200000000000000000000000000400000000000000000000000000000070009000000000094100000000000000000000000000000020004e4f545241434b0000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000007000980000000000000000000000001000000000000000002800000000000000004000000000000000000000000000000000000000000000feffffff"], 0x1) (async) getsockopt$inet_sctp_SCTP_MAX_BURST(r2, 0x84, 0xd, &(0x7f00000001c0)=@assoc_value={0x0}, &(0x7f0000000180)=0x8) setsockopt$inet_sctp_SCTP_ASSOCINFO(r1, 0x84, 0x1, &(0x7f0000000100)={r3}, 0x14) setsockopt$inet_sctp6_SCTP_PEER_ADDR_PARAMS(r0, 0x84, 0x9, &(0x7f00000000c0)={r3, @in={{0x2, 0x4e22, @remote}}, 0x9, 0x800, 0x4fad7375, 0x6, 0x2, 0x3, 0xff}, 0x9c) (async) r4 = socket(0x1a, 0x4, 0x3) setsockopt$inet6_tcp_TCP_CONGESTION(r4, 0x6, 0xd, &(0x7f0000000080)='nv\x00', 0x3) 04:12:20 executing program 0: r0 = socket$inet6(0xa, 0x6, 0x0) bind$inet6(r0, &(0x7f0000000000)={0xa, 0x4e20}, 0x1c) (async) listen(r0, 0x6) (async) r1 = socket$inet6(0xa, 0x6, 0x0) (async) r2 = socket(0x3, 0x6, 0xdd) connect$inet6(r2, &(0x7f0000003ec0)={0xa, 0x4e21, 0x2, @private2={0xfc, 0x2, '\x00', 0x1}, 0xf5}, 0x1c) (async) writev(r1, &(0x7f0000006040)=[{&(0x7f0000005ec0)="d1b56ae7f5a88544248e70b5dca818f965d20b97c74cfd4172e48d3ded3298ccdea9535c41534e24e1d864c34808ff776c1cd552ed78716e9bbb6bf0d27294f68eff5ecff7bb60c92faee5be484608f732b8eec6007722f7e3415acd5d1ca98631f2fe7b988fd5edcd54ce33f3703d8323ebe93cee7725955562422d5870fdb44f787a342a0faef293a775a6c012a403e715888dcee99eced99b", 0x9a}, {&(0x7f0000005f80)="83359419dec84b68e69e37b05f7e18d8acd2c6bb670b00ecb72010b2498a9220061d20b69af76f5207e108851b378b1713afc8dec2b86b2e26a883a0b5b5129b4b63be3048b5783de84a06e1eac86852959bdcbbc29deb00014e8b504b", 0x5d}, {&(0x7f0000006000)="cdc05fe9b555a26673c510b247d6c0c8837484d2c47b70affa9fcdbeff6978e197f0fc56de229dd3", 0x28}], 0x3) (async) connect$inet6(0xffffffffffffffff, &(0x7f0000005e80)={0xa, 0x4e20, 0x10001, @rand_addr=' \x01\x00\x00\x00\x00\x00\x00\x00\x00\x00\x00\x00\x00\x00\x01', 0x7f}, 0x1c) (async) sendto$inet6(0xffffffffffffffff, &(0x7f0000006140)="f6007c6225216fa83f80d015f9b2fafe139eee1756686ad1dfca9ae0273476e4836b5ce4ff42ba41cb389f50f075659a7402c60859cbd979aaa4d40aeae2dd2651326fb4141e974637f898923ac39e10c214ce2f9350fe84b351be7416c4c8144e0490533a0faa9e3894cdb094b6dca983e50c7a4bf10097de809490d15ec32495480b5f66ca85234600e3dc23193f4c76c10bb53a0f16e47d564b94d0681c4594f9153a82daee30500a7e4bbd877fbe17de30af9e6270df83296c736587c57a8f7675f2151b30d4ebdbc76b99795afcd32c578955758273353d178944082376f0d0ae4dc7d255d0c6b143ba8d93db9454ae74b802", 0xf5, 0x40810, 0x0, 0x0) connect$inet6(0xffffffffffffffff, &(0x7f0000000200)={0xa, 0x4e22, 0x200fff, @remote, 0x2}, 0x1c) (async) connect$inet6(r1, &(0x7f0000000040)={0xa, 0x4e20, 0x0, @dev={0xfe, 0x80, '\x00', 0xf}}, 0x1c) (async) sendmmsg$inet6(r1, &(0x7f0000005a00)=[{{0x0, 0x0, 0x0}}, {{0x0, 0x0, 0x0, 0x0, &(0x7f0000000080)=ANY=[@ANYBLOB="5000000000000000839a6de73600000000060000000000000401000730000000000a000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000002000000000000000810000003700000000000000000000000401000000000000140000000000000029000000000000016000000000000000290000003721c218959b09a462fa9c0e1c6feadbfda83ee9a1785d09930585374e38ebc2f37a461f"], 0x100}}], 0x2, 0x0) r3 = socket$inet6_udp(0xa, 0x2, 0x0) ioctl$ifreq_SIOCGIFINDEX_team(r0, 0x8933, &(0x7f0000003e80)) sendmmsg$inet6(r3, &(0x7f0000005c80)=[{{&(0x7f0000000140)={0xa, 0x4e22, 0x7fff, @rand_addr=' \x01\x00', 0x7ff}, 0x1c, &(0x7f0000003480)=[{&(0x7f0000000180)="c22de8de49c8bc74ed6a27990b96c7bbffde6e451106cd7db38c6cfe57439b6ccfd0b713f70ae54f02270b4776d47ca2288261af3a5cc7029af46bf8119dc3311516af33ab353064c342299bac4eeb1392f74f4ffed1ad7addabaf1fa8010cf13fadac6c44bf6d2b", 0x68}, {&(0x7f0000000240)="c97c8d45fc51ccbd284a494f7fa337de3064db9bf3406abe667335fb85ab0ed822756db072b12fc6d77789ab7c1acd7eb42208104b970485ba497cf158a0ca4f98eea848531b15b3ea19cf906671947387cf789ce725ed5b8e20db53c4ba19fe700a12e3d8569518995b863575644a44f0956259c90c98c6add1eb0ecee40d799b20c5bc7186300cb062b46915a2e5b13af446e725cc02114689646ac7290dd637dd1392c1bf5189299197b6ac5106256696f0ff163a3c739be5", 0xba}, {&(0x7f0000000300)="16bead57a423edd4b1cca1b2504dcb0a38fdc890f65b7636d8daea0c", 0x1c}, {&(0x7f0000000340)="51522000b4abc08f282fa6a69ceecb5e211d38939472ce2f2f5337ba56a8c81a5e9b51d5413e3444980182b0fb8b08782b71830fc891213a280a354245d36cac7d7d6185d124a6f1eccf9bed0c8a9423da237af98bdf4a9190affe49352eac845cb25a71680a1157516364c90f75a0a7f544e8512912b91163aeb614e3b677d318163e3fdf33e3e17f9c45f0364acd3c39ce5f8a204553ee4e6ea7", 0x9b}, {&(0x7f0000000400)="103ddb76f5c3a6822e67decd0875a1b06ef2eb440a6503c0d920b69baf9137df4c1f3dbee66104c86d6908e8e497ac4f2ec7656337294621b9113bffa9f202879a2bbc3188545e5158e3795c5a6ce1382cf68b8965d2507615a23c22580813718c8c5bd7d0cb1e8cf01865e19026a5c2c724772a543d1fccdb2952561885ef64ab3f4c8c45d19e2a39892f0dc27dcb5c7546cdeb07502d85daa11369b995a80571577a527ccb869d0dc362aa6ce57da19b6e4671e95a966fe8d707037ecf9942bbf8e61d4c16a5d9cb5e2388ff4d4707dc266207627a1969170396ff0fd00ecfe2cb616dffb2ceb658b5622f6cdf5f2d06f6764ff46ea4bbced621e95aab636b4653e848a781ff8df5c69f7d357f4145628c57a9f4aae6af9279ffe252c3be1ea2c693b693a353ab092d3ccc4e99ab076fb290d2326e317d65548ff64373b888a2f2817f62a73458059168d118424d88bc0f8f067678eac0ab22ef1123083dcd46fbda980201f2163286286d2e086ac7daa12e40880bc0647415e568f2d7af092c8f4d85008b1a335f9ce8a3df4216861b579152b1ab678e294fc46c815d638ce4925c8b985ff55072ae52df0a6be6f6d17c8f781a0dc1e114449c3d947814fe3decc167719d0556ff1007aea91abd64a1dd151eb0e1887409fab15a4355f610213bc2b3dca4b94daeb00419af0a5c16470d3ce57dc844eff080c09772b1b2ccda8373d8d4c6934aadbba6bdd76735f78f2076624132614a1f86769ec76aaeeecabf3fb21a5db4f807e4dd64ac62385ec5970b7c07bcb50cbd6cdbf2e0325022db87c9cf1212a1145ee4f0d4e67abd86c5dcc1e00982ed884654ff0f9cd4fa5dc6988685c91dfa182789eff886c4b9abc0323fe4fd05c1cc37e922d7ff66e8b7e7186015a28dd32dce7cfdca009f27ddd091f385af46add3449a9c9195e1c9b0b2d78919a60ffdc072a3ac1ed862968f83701bc25ee9d60af2afd1aae81edd72822d536044399ab2fa79252c633a3adc23b51616a4906ae17f51662033997f68a4ed695bf04e0cc1fd380873d70fb26c56ee26ca04092993eec119beaf5810ce6dc260b9b0036551b3532b2c7def534d3a21f55d326a357b22b5be248c70aaf1f0753dc379e70c2ca8420e75eb0f718a152cc7a12d34c7cbfb9de3ec35ba311b2a26050b2fd6730ec21f9c48d13f60561951d6be3eb6cd6feb7fce4b8049e504d9ad2469620910e24fe6d2358fdd195dea1e427b2fb0cd8d91276c42bb45e25b7253e8d38851b1cbfa73b67070aa6d576a78eb9283d30b38825e73413db827cad05211552ce49fb364ac6cb5d4e26a86c041dd009290cfa8e261b5cf26162fdc8a5d091dc351ff09911b1cffb352e6b814b16edc6b059e0ec7e0e218041343676c703a17a2bec3c97d68f44562491f9ac0353f1a0657dd3511c2a489d13a9b97feb37abf2af95a9c4241a12429954ee24afd3b7ac672e479954f90c0a591ee903afc8866fa38c1358b2963683e49d54bce0594eda4015b1526599d51cbe21f93c5d7e0a446e782767bffb793f691d07929670c82d373c6648e82b70ef9840b3ada8d2aab84d1017b2a913da6b3b64175245010659bc3532f69c58599a03dcce41c29d05dd8e3a8d4716e8b19a5f4a3a8f37ed0f0903e5d6e38d66596a3c1cd48df3347fda424e7691d733359bbfb3feda692331ad81a6ce60764bc0889ce40301f46c6a71d8da8f732d25a3c94dfc1ef0eca420723851e37f3b3f21b659c09cdd174c610693f326d8bc5aae2b72bf114f95ff5e95d509208641afa8a262cfad171494ee62904174fd9bde07dec67ca1f071975d8a800034d4c13ec8a420708d10f9d5d7fb46124dff3bb7cf4fae06e114550122cf764194186f24a9701bdb279383760843cb0c3594d2fcd74607e4e7681eca1e4ae6f7a703795bd66eb07603e3b8920bd8538c479b8b9af34fe4cd2465bcfea5c15ba4c2330e9b6c9165da731eb6213894da8effb72773b70a812c7e110e4cbfe91a0afa372545bc0d3d60fcdcff889ffc6034c208ed41e37d60418e69ab6f285b47294d3e9ecbfb3a2a100189acafb9523e8fcbbcdd99a4be6a32b2a4e62c1f657eb1a3a1e9f151681773b3443f79dca58d45118aaa92463c6d3f47345bc074ae37477b4cbfe06e89bdff5c2660ab422cefd03f7b4d0cfce6c90b2ba134f4ae176f9acfab18e8d06931c44557d10373929a80a95696913a0dc29c01c827d79c984526fecbfd6d480083d47db817f68b5ebbb2e97a97b37ace1215872ae775267dc1ae6d41bf2b99faffc154c2afcabcbde090cbdb2d9b9e7fc3ae9aa3c7603e72f6ab25ac9487954c6ecdd0c43e336e495d7f0d84d9452f413eca875377f4fb78ba9bd16550fc90752c8a58db8f08811e513162ba329e75152a4503834db7b5d728c68daa245e318bbe23ef274c4617756b5edbf6527927269e4162468aa3097587718668152733a88b28748541fbf9ad9ffbf8ba9a9b8ba0a251f2eb214e72d08b64389ef1040c825931bd3a78cb9ac6fcbde9528fe57c2e4dacc9b5499258c3c3ae40a5c70d8599cadc1d200d9f7686e1ff0dfcc0bcb22e6550c531fbf185abf47dffd485b85783ce6f79191498de308b7bffbb16ab88bacf952413a7669c5057dc392b1a4055ed8dbf3cabcfd5e7c6bf46abf92fc6fa5077a1b1c669fd554cce25659d09e63fcef040cf8489af6b7485449a616453d733435fd225d56102e49af1977ca972b54b2c566e1ad98048c684e020b4523842a85a13f27795ad4378160844ec06190d358b2e38bc155980095370620945279e1648e299ac25e217731a3d81da91bf0a9cf111601877c347d3f48728bddbdc31c4c5093972c1f1d513a89a332e0046a8c911af2ee273c14ad3155d6d15e0d4816cc3d7ab58999126e1e039fc882e7b1fa63e4f366664f2f487538f6b1e4f6bcbffdc7d70f34c2cf4f53885a0e7ee1224c992189504abc8a702e10608aa66dbf4b4b305a3c4d99937058806c8af9b853546ba1bbb84bef9937c308757863817272ef8b95061a07dfd57518ee1c4e200303c24bcf982d9dd9fffcd677bfc2e33e4026af218e76c2637b0ab840097ceac645b17c107fa2c50bfbe97846c3f202430c750b3815590000e0a74e231efeffc234269aed69232426b2408e1d362879498c0e425d1f02c53cc86a4892cf680ea63fce419a6c7df843b64f217b25b1fb10049678874d81d733bf27fc5ec89aa7c416e4f2dd4c88b530036551b4465cc756bc31e15327caa299863a99307916412503f90a30ee79eb1068230bc304ef145a1bdf7a095c1196e687bf8e9f780358314571840f2b2ef35ca94f131404b8379b2d71b9503c6431ea932771994f9be635e5df670302c90745023c37169489517b372f373d0f52717cd572a713d1e93cda0e1fb05e96e0116f4df465f7bb271943e958ce9fee5845f79847dcd4c01b485c7b271d40b2d704a66b32c9b3f2733576733e03bd4af60c8b1079ad0af867f8601ce869d5f512b77c6bd4110eb61b6da4caae42b75309f978f39596e7eed58c7638901a3aa6b9f97d851cb4efa915a682db74cb64f8c28e268d65689846bcdd4fc2cafc5934e686639c0e2a08a4eb1fd14ab7231956a75decb263e99f74a7ec9baf6e146c157d63a32fbe40b9c4cb8f238a9122a5f34d9b62420cfb296c20e0be292cf57c90d8461425cf32197d35c7747d95051c3c4b4c459733f245f40922c32d4764a8777878df1be8fd42f85de3fd01225a1d8d9cdc756306dbff987c3b4bde1c229ea45f81b9f20e6958f8c8a6fbd7d6f7bf9259b060c2d64f74769c4a66dd0be11ee9abe3b625a9a7677a2ddb94a8ab825a8e4b6b75ffeb9f54acabc68ae115fabc093a8f444e19ee266d7ad96a4fa87e70effcf18cf184c32d5dd71e7067157ddbe5f0e2f875ca0c56aad214cc9ce8b1a8b1681c74873d5dbe6c93dceb3458854b07bece891005ac2b8b4800eac5ce0211b234e6670bea324806a3ae54f8e4a48c3d7bcd102c408120c7eef179a9f2cc336fdab92b0e78511022d371cff1cc2c8f23eb9051b1648821db6830e454a2b343fe670c063c222efab424fadd58707352a4f17c4faae8e738a1bc60def40e9b8c731a11ff1ab41876593a82e7f60c53499778ad44986fc9dfefadd6562b59257faf3b45097980ade15c50f9eb6d5ce622ee86b55aafd5ea532cad3877d8965b1d831b9577dd9743f7026163c07410af842496916db1360133ddfd26feb918530a0f7ab664b6f62dbf10aa55e78a00061451006b3f81577c7e55771a08c835b56c332ae77b3f529ecc72f22f8ff4b9296480b50f88a751f05032a109502a3715d0411349de02be10a59284a91dd22fa3f8246ab1513c1cf9e964734878023a9cf89536f4c0a5c3286ff654642a0ac42d72ed68583f48c267fa26c8554d44bd312552b1121db719fdceddb9b7f3ae50fbe82984cba814662ad3e6d3f644947c89caf902de2eb394885ce08e580f4d087bb009d6bd41be5b6ddd41e95d0d7919b2469632dbef5194983500cf6b008320d6b657f6506bd9ac992a1723c477807379226e3acb4d8da21adf5c4e7065bf3d3fd395fbfab060dbda3b54e65f64e3ce2d3d20000bf4748c369ab01849051adcb6f7699034a872204e2588b6a52dbe81c0996d01612e15d0cbf21a9fddf714aa549e1717e10d24e7b919fa4e87d83b7613b7b0ec2a0e7446a59dfbf82beb6524f41f0ad05898f8f0e47e1f6beef9c5519f92f516f6a2d6ff4a56c083e4af8e801681e8782e698ab11ad3410a210112bc6cdedb09dddf7ca681d38e03b38aa5c90838e3203a232364458917dda37dc1b161ae1d0febd69347af30769528ee293909c645c6120a6f34b5504bd20701011a3cea8bc301cbb90f6b8fb1b3b94fd9a6a9d53227848d54d9b5749d30b3bfa23763d41e64fc8c8e07a74d4e12001de7f4c51e14ea8a763575dabc9cbc2d65745ac9b2eda7fa146579513375d7ce55713618c4b5cd44a4633d6c2647866f745e405b6632af9aae292001ab7581a11289a4a4ae1f235c11040024f89f5317b58e6fc7d7b9ac8ebea567debc7a5d0efbe8eafb374d4159db45767a829a22f7e503b82e0399c1350e8f25b0f428745379f9cc66f093615577fa527d9fcddb8cade543ec4aacd0f46ebb297e8cc6709915ba3b21ed93ea9284306cf6b3759853ddc1bb994a4b2dff3de13309d70dc0fa76ac64aa8618ce6127f38d3013a4bf0d2a29c6319ffd3987b7ce98a3aba4affc12c52e8c2953e3024a8e200992580601fd0c0acbfb8645453ca752a7932d552462109d208d58ea0c10f2b9da8d214456959a71bebe20f2193d0abed2bb4af4cc7b6def0ad288abe9b96fd256931dd2e4476f237daff691bd9539fc736cf5b286f39535f7b35c821b745ac24a19cf3af9228603c713e9eb70eab31fe0804710b9d6f2ce435dd2b45dee997fa333a3d9c2f2ca15271c385936ce687b102a9be42711c1baaf9794ce69a82fde4a39aacda111fbb965d4f2f58f6b48c544983fabd3a05839e204618dde1855754278da4d48d48ccef3d1d9bef906cbd60af2aca183b802b6daa41ba3c6f72ec6b12942fcb9c0c74cf79f9c0e7137273de4f196896f9cab3f471a182ca31268e86be5b5c5b00ece320d683330c37ef00ff498950b91a6c71442af1bc16868a0ecdbbe71d0d2e513724c3e9c44c0a058a1e95122493380cea437048532fd8871c00e1d8e52f6ce95aada96968a76a2c8db620747d622bded19f9afd6", 0x1000}, {&(0x7f0000001400)="6fee70109ca972282af8091517580ea97c02", 0x12}, {&(0x7f0000001440)="cf58e1164b21dbfd7361a6608edb7c3d74752d30bc803105cdc9c267ebb3180a712f3329e2daddea7dfe9a", 0x2b}, {&(0x7f0000001480)="57b10b5886521c21e2472f7c92c83189c0bb4ce02bb5a06885f74e06936c6cbee4ec566161101f109bb1ef4c83803331bdd5e94831bf88064113339a3e7c4658a414a386877130b9746550b2551cdfbbd99d9675b28d2d3c8b00f1866b89a4dcb1cc223754e09889f4eb0c9a5f3be77531564fda6b4aef20a152c5b5f3ebe7c26930e9e95361615e8edd1aff4a9caac3e87e6c467f4abdf8640dfe6cb42d89b48e1130745d47ff43353e0de9cb20156bb2cd8a64bbfd22d32d7998d405c3d26eba0e48d42ec8b9147ff41cb6c56416819d66c00c82c64b8662ff23690062578f31de8265f252d50a42ce16b16c64324a26700af0e31c3b8f09dcee863895be7fd1ebeeb89f57a691d06ea81337977e41970a37feedbb49cf7e6462a3d755124a58fab46e662e7db352706838b9b24ced05f4a52c542f537de34f0fdee6032367d49697db1247e7a43e9a4a0102ba8d71f4893bb5474121c84e43cff9e233671866a73fa7716cf2012e80905dce31ce3d1c5d9fc38f8ec29ceb82c924efd47d5f4f7f9f24673a3dc1fa5f4bcc81caf5d96e0ab3a2eb093c057a32a3122d104dd306e4382c0ecb425d94737e3693f1117a4b1780cd1f9d8f1a5b61e11f52b48cfa21777e9833ee0a8a065e3d17c514b76a04c69fef9b430b6830415d8557b23fd5f26e39a620f4128923a6c0c4c2a24ca8b27b110333968ad99075cc9f222d724fd87866d9858df6d1259452914f14f8c6c7a70daca6b85055684b8dba8225078fde417c6029e6c45fe7d6730c1fd965ab7628539585b25059b1c5d20e98478ec92b796ad25d10b08483b118d146d2b73c7145c6cb05f521c48fbb98e50dd75ce522ffe2c2076463030f7dbf61c113bd27d3b2886a5d12d578f97d8623c93befb22972be01d3af9f5d4edbcf4da523053a3539c8d0e4ef407ed52ef98a49c7229aed12903a141cc2b16acae5f92e0dcc27719c5aebb93b08cd5d86f9eac038cec3476e4ed5adb727830567fb405b3929be507152e9e2230facddcc133103acd7209aedb0f251fda299454b40a721017627545eb5b8f0deec7797211375f14b57dbb9e3f06f4a33dabf179d5e98c03798251d62dc22d6a760244bbe6858542183b2d40f7c3c755fcdd04a8502521e98a7a5f59e4dff26746d275e0b40746e76699789f88e45e85abc9327c7773f73cb91b43a35694def999fb960723851660452882860fe6623ab87f13337cecf7c8b18f3ef5afed07d5134e168722097b7d45e16228821ef99dac527e961b63fd61628f4a7cf9f1231262c83ad5db9b550c98adad1cd15db17cb947f192030fb44d638172329c7d87c35e510f439c3c72d952632ceb044cc57ac3d374dede3c81c8e75ea20e731de636f541205cf8ceb14d998793eef95ba2320e5090b299894fbf885462a4d909cfb8d021ccce95d6eb86ed47c0c8c8bc384333b81ea1557a66603c5a2f23ab0100cd94e3521f9bdb14c4c32ba9d9bfb1011e132678289d52e6c0c8c877974fe187f80c7fd11a80a59055d6d9a132f456df339fe90cd1fcf0216bb1f1c139bfcc1dc72334d08f467996ce61b692da00eefdae66bc54db30388665eca192f360ca41e7401ef3dc0f150f5831bf6356cba27b642b01bb8dc7e3f711c80d43e751ea7a1ea3376edbc704f8765df546e11a2fd23e1abb034c62eff880806fc2098981ad0388f233bb1053dd2a0bb5b1605ebce3cdf39f58db49b85dbeabdf9e4561ba2de2d2436b4d411ad040b419f5ddfe8058d02cda2edd51c320f1a14ceb206bf79aa649fbb3804b3e15c1e937358bb88a4b8be3e6de6b2e28c7c84964619677745ff8c009807e40017df35c60c11410f5dc7335d539d1c3918ea760b822bef85de33208d0f3395bad8db442c26752a675015211a0d49524f2b5b7b85c1f9c8e7364be0b095b2e194c04d6d96d9ff030a46f5b6324bea678d27d1050b38cb09c3e5973b0ab783b60b58a9ea119f28ac728203eb444b743f130fe99d291195836be19c24690f26f850edb5dbb409cdccbb06015e7d5b3e64fbe31a18191989d85f9196c39b0c226d189504da902f5c1efebcc5582ac35b098181c9ea94aab515d5691f71c0a267ae4b52593901e0820645b57da3540e986b14e8c32d00e86efcf1192cba857bffa78b6dfb99f5e69bc7e1d7453ee4b8db6800f7b89535b79d1f9304d99dcf33a7aa7a1022a2b664f65e33cff13cc34e86ebad1ca6551fe7f3431f72cd1c4563232a084ef283ef4972c1aaa8fe182a9838cc2ab5d1608a577406c663dd183ee6f6a6c1089a30f8924e858c6c6987619aab0987fbc413108913b013162043ec5344a818d97403bfa8d3c61ad5c6c9a28c4059e5786435718b1327c4f202cbe956f37f99186d57373c40daf05f1a9bb4372ff5fb66a9d5f7c7a5c98e987835ab8650a5fd68c9368146ba98c83a780288e9bf26f57ff4cef3c06569c21d30c64767834c5a7705e4e0c916703c8cabd08c787ee5edbd6fd90d3f56a0ade79fb11a41ca08ba9b4da3816ffa10fa14fbd76aba422ea613760295157e33e2b20fd4621f6ae1a6d39873f04cbc88eed0e313935de7432f55fbe09b9bbb110011fc167378229898b128e7f47dc6dc48ad759bbe3338405c17690651a14f313b540d9dd3639decad86d364d4e313fdaa7e221c862599d9333d8d307611f9c5ca8bf96920697eac2167308a1f22cb220edb497d978a6f7dc0034f7f85eb9858a44df554d8d622fa31c0bd915ce7b57b77d9141cd832a943aaf593f219310528103c3fcb316618c3b649d43faef7cbaf0f844eab8095edddcdcf6dd56e4d507e7d5f7582b90c9fcb7e9f69066e1c6cd8ebc6097412bca7965c44142ef5764c8252d8b015dfb270ced3a39824e3da0c11a6af451d0c59a6c5ecd5b367bd0b4859a291d896e38c285ac3abee0631f2337df0a70a3997bd28ec626fbd2c1ea66192f20e41a4b5fa0c3395898efcc01da101c77eb551240be337805cde29cf637ab7f57467fa16bddac3ae58b15eeced6931098093f155afcb9ee3661b081b7102a701841cf14f2dcf13e955df44956d702d95bbc8bcb8a6542cc8c8bc50711365eb95dbf78fe0bab478ccb2873fc8f8ad4201a36620ed11a24caf032147e8d1ed7bde55f15203ced96a4c45258b8467b691947ac4a165dd70a0bf7a5ce5c9c1a7dfc62fc83bfec43d05b741d22d7d42146b534428180ff9658544e3d5da92824430caee91f2cb063c41afc99b4c6b36503de16d74654d0ea86c357c788fec88a39baa1fe4f0a35dffa42db60b35ae78a310931b93607fa3b21ed9b87ae1c6965bf69a206e52a11b0a836368549ef535179d1f9edece1c33e0f9f005629b3c16be4fe8662fd093ee149ed2ce005f98c97474ccf04cd1b558eb81d6245a47736a214be31855f610c09071717bda0b54d8faf177e597af4f3770ff68e4357bdf507f713f92d23b35db3eedddf1afb1cd37efae980a77b5c218fb118f36b9d8e0182ae8970f403a838bfd50d38368f52bfcce98fa007278779da80da7d9769c9d7e2d801c8de4e40ae4899207ef057c0794e4b6ba10ad5fc56e42fe0807c05819169bb9b8499901ee31f49745017532c248fdbc0ade0a6a25744ae73643bbb47e715e617a2138a5cc66d25eb50abc2687b738b31ff9cb6586903acc739b8c7983fff6216f089d4d6d1f6f31bd42bee8fe326f47bca495810e48b266c778974a1442b86976d0b3be808d02b39d88294676482865fe2d833c2a39b401738635bd6ae0231705a487305dddb3af2cc42a8b39a7e811f3ed52edb439acf3e12fa8f8b2b66bed00b9894caa5b6a0688f7e0f699815146eff5e3f6471812bf49bc1a7f1447efe523ff058d739cdba9694094a5c64a4933682a0096e94f29e1b6ee8b84672da3cf1c17f2c2a546b09163861d0b090eabc23ce54a9c8a5f10525ac44076906d7f7991b2a08d1a7ae60fb18f468950c6589d5c1f698612917ea49b66495d03b0c3db11fc300e39ffde962c5bcf401c3502f2db3ec29e823c0691d911ef85e82f97ffe3e834c92ac496f20e46c809bc976a49669721ad46d3a4b315931d516a426339354e9760874b8a612f63c50380657f99d7eb3501cc29f9372f161816ce1ae4d285e31e279929dee2b87ebce4360ef0b632990506d310fb8c67a6a6f122959d8331623ff308bebd105b36ce9d2ec9538a8bde87a5808a555b0ceb931c0c57d139ac1fc576e55cb7e58d4a1cb24889bbbcd670d0bae65f1eb2cdc4870810e7962c4bbe21115779c19a5d90fae05c621f1bb5b4c741f9164619759802fd5c887fea540f218077c94cee430a750dae7f1c1784096862a5158b75456b6827a27aaef0ac9d59ed1210e25127035c745ad148bedb8474a7abe10c866fa3f86cbe49b51f80bf22cd830ccf0e81de8ef71acce5520f9047fa376dc39ff9ce9ee11f57512392ff9910df3b6be9b58ac5df9968fe6d1840b9bcd4d132c0a4eb174474dfc6f9ba0f9001679f87af42eb3ca0ce07cd04cb5049a4b8d4f6517d44f83fe32c2f5fbfd0e673264342b9ce268fa42392dc5a27c7e34e3cba258488f380e3dfe633ca5a3f9d0833823ef2e43631096535bccf67305bf0b09a5d2a6bc20886936ba1c1615da02e98990e4a77f750c1aff8f8891a9eabd54b036da2c90589ac96c74f1fd43e01719d641ccf10568f9b7faece6fe0380668998a8896aa5d898a549e653f178a96dbff81bef5b6da102e71ecad171bad34423d3bf19940b287c4a2fda064d6ce4619ee5c726f0c8dae017b6a0cf204cb5cad2d945e2ad2c1551e6499467704b01e62741d7d752e3f58efa878f5ad42fc86aba15ddf0650fee8c88f5d431069494ef2bcb3685e1c732b347976d1a2377dae133e6e3060c1317125cf6507b1ec001973694ec60497464db4d74e7695f7e07d2c47caf6574dec77fae787cb12a9be276b99c61c54e74ef601485f800ab58f4069312cba6d09fd4c0024490742b38557f5198e7c15a928a73e958322d35a6764743e942676f11ee872564d7b2aec5a92b3930813c9d998a3b876ea52dafe5ce481fcf3bf28baa4729bf250e87b8ea5aeebc4506f15e0c60fd114105b3cb0e7457d3b1e348cb93ee220160719f9d53cbf8bbcfed64f2a3a93e426c11856199d0f402f3575e8080e8b0b8afa556001254bd4767982a13448ce049e80b1ca806d13c3780093c3e09766fb6a5768a14b9bea98bc84af060aac734299d81b1f94ddf1fad4c939d70114e0ea7d2e9652615a4a311f0e7cb23f4f288bf1b4b8164233fa50e12cedbfd336286d0c2871099590f05c68b2b4ace213cc1062e5f4e452627bfc3dcb87c040ff926be00f583a1a4691a708b9dcfc59469d1a08f24dd392fc2a0fc910e0fa59a1678b6ac5e3e1a340db8ab7814ef49c81c20c0433a7a3f1cfa8ce6eb88d9cbc5a7b839faab5ac199b6517043ae5a45eecadd8d1af204facea965f85227adbd8e84da33db5d1bc995015dd164969ab8309ff92190e91339781110a8e29a1a054b4e729ac76e1ed941a0fada54c04863e98aa0ab9586a3598078ee45efe102dfebdd223a203c0c8541748124537f54a9df0c6d7e7186e844f7fdd27d4ad1defb99170153f3a3df2eddbf7df5647394665b0165b5489219b1dbe0777c67d384f73946d8d25382cbcc760966f6e5ea21cff2c26d0e601d6b3051c3c92ba47b904aacadb5b6da6f65e2da15537b0a52aa3346b330be2b2b190b7c959d016391a9d2ecb6b6c800fa339d3ab70945e90147840a7563fd9e1", 0x1000}, {&(0x7f0000002480)="216d2255aed99be1147baa94d0465161193b42958a0d74eab35bc3e21fb46a3c5a059744bea8efa1abecd221d9f2d3d504dfafee023723037f27e3275f882b24fec454723cbf5ac96c7c56df94397d8d2934a512a2f27a38f76c5d14fe50e04a799c0ba06475be2ab1349d5aa1d4f8db47f35334ce14f0f40e8fe733fe4070b09a9484317ac3ae887f11dcdf001144c00e0a9a25aafd47294c7696d758905690fd1b2615cb2f4155f0bca056a9ea23c168927df1491f6c9742980b0ed703ac2ada240e6f022182b4662791274fa3214ff0b5f917694ec3b01e0c753489c68a3b1e9b2901e97d2cb81456b13a780513ea02fa79de92eb59ec56565d213826e56dcefc51c04b50e7af691562684beaf06cb27ed5d4095abbaead24de8f88b498b3dd15780d1c4b4d747d1f3cd33a5acfb1cd020e5020bb01353ec5f5c46fac84d00014a9bfcefc91c56b0b92a80a402c6cff887bd6c0315cdc8c79818a87372a4d36f8a35e9f3eafd4b37b0b08c9aa6a7baf91a0458609d3abd88a054c7bf180e371a77975afa9edd941b2cc9c05badac1c677abcd9a2547b788122aeb7034a9c6d21999c4c74a51fb1176220db94525719faac17319516ed9fffb75f9c1df822a4f4df07c01761e2a6a663b0fcea058c372ed6946644cbf737a3e0fa996d3f3a12d1789a73459075521334cb0717bb36632fdf1713a2fe2ddce70aa3a5c1322d1f17ac1f717e25fee59e47b954ef5ed45043d6bb063685013d539fb643efef00ca945391d8d0d05858a09928585e6e13a0915c51d38c07fb9a88801ef84cb11980b9e18486bbee9cda0ab460468a9803040947239d4473d391babe47303e6024ed921dd2aaf5c137058360d51d0d8994388788e2d6b14296330775d2bbd34e7c58147d55fed75aa7f97aaabde3f4ec784c221957badd2ad6d0024d92c2530dfd8851beae88df4efb334def1bf24edc2d3fa772c3cb6418349686f8a205d998f9e0daa4b26e8d87c200e9c5617cdbabaca1ea381be91f26a2db95199394f3f50f258f71ee30912d53f84571d10bbdb9301acc73d34a47c8253cbb80ba441c70bb1825913bda17c85d3996ec754257fd48b636f6c0a16847aa28665f25398fa1fc76a12ac6ed2e51cb63b67614a08604de0c1a25ca36b50be2392003d1c4b1022ffd53316af9a6578940f59c37af498b61493847cab34c4aa4023fdcd7e2a4d164e2d9255be8abcf6a64e7e132f5df353b5cdaff6e7846a0bc09536c69606698c76590086ab79b988a7826d4a98858ce5b5c78f97f61befaae8458c6e1b8bc31ec7996be064783f98187892cde36d529dde6101f1ab19e9d17c2db56978f74f3af8cc9ed5cb7b661536c4747754cbc8153f30aef0d12fbc8354aa13e9d1084e9d2b71971830687eccacdf85ee9ef42900166a03c4b80950ca5d51af438d4465a1651b468b047ef57894caedc5c4ea6639ab5c43cd2c0e0f5684cf0e678833a7aa24d854ec6d8034aee508c275c0f3f1be92b4d35911ba915997f44a10f81dfb027af84b4991fc72a5df57dfb1ae98ff7755fdc7a450df7d424604b9941f8d176aa78f547b948d93b304249f931b7417931b8c9942b826dacfccdf4ae3512e4e00fcfa06cc83f58af096c927bb4c7de8a3958c1f1ebe6bf8723030845ccf639725f053e0d702b404bc4b948b91d6c40bf35f5b0ba5b5abb578b3c35c06d0ca70a076003f2ba0072c5a29b33a073a509092c3705acc41e0a4ca46ae219f3489b28d9686060ede47d93b1bfe56f12820cfd35f4c0613c60315ebcdd1b8b213cfd302863254eb52adf904fe3c426e677b07389cdd47824246023fe003540e235d79054656a769b5048a30642125253442ed72c8c75a137d60eb626737a2fcc40043e31689d2b762238fd0deee115accd4b1fcae42367fed0d7fc671986283f276a4b0f9af108445cc6451dbb4cc0eb99e62a3ce17fb5f921808363a6916f22c1693dffd2e3907159e42910b334194ebb27c2fc29e57a937d4b3ad83960644075d9fe879c9a2ceb923e38c818af3fe516719acb875d08b0bc75f44e049610abb24db5a21f1b0c93071a15bc732d78c63c41d256d25f626b00b13edef5f7ad00ea3dc07ecc4f70545fa694b4cb3f8ba25c06e6755261efd29cdf1c280a86d727dec3f976ccb2d6b8b069b2e645f9e9c1f666cb2638723ff053c7e7631b3a5ba65b6b6ff053bb313fda95c3803557aa9daeb73b936f54679987188dfdb770af72e83df19e35fe08b7bfc2c9841ed2540f0ee005a4d1a4e8ac1b2eeed228d319f9f31761718f4cd3bfee38513905e5d276e59af4f0c3c7c3d935522e821ea1c54416148867c220469dfc7c6ed3ffcded7e51dd42594d1874ac1a6f7d98555443ab2e6b4f7270ee8e8a6aa7a93056247b5dc06d4f4e6416df3ad37662a913b1e128b68d2c13b0df064199f7a636980dc7d7ed3b8f6e6ba3369f00b416d3b111528bab2c51a2074f17a1b3afcaab6171bc64df64772930be2b6713d6c9a7cd10fc81dbde65fbb35fe20b12e8a3d560c39443788e09ef27771e216417f4ff2c1656e668f06e3b3a2b3c4073296a7296bbbdf0bf64979705f07f4f3a273a2e625f4ae328c79093f3c0324c072ada3de32693318fe563f66ea73a78e91e8fa69f99f3910c4f0878ca105a037bbb7ec2964fc991ba66a13f45f77108200226616e56d8084e75ff6a5b4a631421e868709471051972fd3104ed63681c9cb292e24e1f8a3d72901ebaf6689e96d5177d5c897b0100e64e363395e8dd5f62670eb2b67e63a2c80b6cea3f6c5b2bf118941df5d49d7b1ee0ba0d9366c827c21d6b4cc7f72099311b3b62420c14807329c4ca17da600db37e188b8781e294fcfe87f40bd5b2224ecdec13b36da9cda55caee246fcf603037699fd58bc10ae811736b86b816f56912e3018fa341a254b79c4682c94dc93384b57408a9c863d01a1f8607a33cc983e8fd363d67654cee5deec055dfe01a6fa1f4ce305d392cf2b2342e4b4ca4e61f5795c20b9355fead2abf4c590895be9f7c76530eb0a10adcb631e8a0ce1f6150d1a66f4ea77274b9de3926bb918e476cf8a095ebec2d3fa3cafd61fe143e3ed47749b7702e62b17ef6636a1b4921e892d8844e05fdf14b9c5b3e9a63f9374f42847bfb58079a8cb87c39004a888e5493dbcfc30a1f345808e2faaa2cfbb28b38f97587c3dd9f4c49430e15b1ab56582767374788ab6a83effe0c21a2e5b759e68f2e0896a160c4d3c962ede288733828f8e1c402a92987d3407f5c2943825bc902f9a4acc5f0daa96374dceb623594c92209cabbc30d31ad9525d85897f2087bf89ed5e9fa24e49dccc4d69d226a89bfca5b9142201dd78a5bf76a9165a8fed75500b0032e7eeb5b7def8a3a8590db6ffff5873a7d766b402c7ab10dbf77fc4ea6602af6c174a83c725d5b545bad2b589a4b16cf559c66162d80204cb03fc0d4e502e0d1ee10edb1506edb936747563c2ab6afa1e4d5f04266f0cd64bb8332ad825f11f00a9d2906534867fe5f9352f6ac97705acf171c8ddbe350e5b709133cf2d27577b15176768e6903f13c4c49bb0b0ac35ef95f6918737132f6a94f2e812853e6ae1295d7d567d4d4b953f6c5bdc9f36867247657ea44b58d54c9c39fb8dfc685bbc75ce9995c825fba731a3349d4a2d2da19486dfb7f18415b5e6cc8a67cd92dea7d0b146449e2c655ff98cd3f5d80574e072473b3854c0594f9458e769e8c96c71cbcf72c55491cd960c5d83540f4d53bc43867fccee87543507bfaee5200842b7c1fca23dc72fc7a5b89aa81031b69d4bf2544c35bf864fe9024301356d0887e35a28ac304edd5002ccaaf61c8b3323d0a1b039fe57fba27384e67add2ecc242f7e2821b02b34ad6fa7a22af23d66ec1911579df373ce8361ec1e3249c1ce618cd43d1fac6019ee174fe750bdd87bd61a4f0f44174ef68182f7b20c104c37bf2c0aadd5f7cdd8d7c1892e106067658e3a05d3133130137cef0a3384a39c54b6f01b02fd32444ff16e31d567b88a2d1a5f02064cb9306ae5a6cff532dbe5c65b0796692b00a95c9edb2b69330ddc461d95400fb13716b84ac0c587f5d6d6516f8afccf83d15732f8e07326e8758638e40f0a49038abb1725f5bbb8eda2fe4a19b2b9fe2ca71aad4d878e5441cf71cbc4738adcc6049b4cb936c6305a21fd90464f8a3c211873296866919e3fb6ee712a99c0c2de5bb6e7f3aa82aa2046278b761bcf676818e70b93fb814bd66a35a273fb96322af95b3c00a4df7c8c90f4dcd2f363951ec3d929043e449197d6fb2eb3bb4335afe7ae64e4f2134d25a0ebd59ac7292a145af5083d06086b84b78b3655fc7b1ed4450494148ff4a3f646617676381c963857b7b0096614399e4e2aa1f1a871d3d0ab7742ab720f75a494dfd1a83f76697f23597f7eba7b6ba5e3881ba5e0fe27f05223d8b733ce038bb83c472736e2cd563e74524d7a2d2164950d6cd3a6f06b63cf3a3c8f852a7ea4d4dde1e569cd93eecba0a6ce5798f05637d02bfbbe4989b62e2b2053244408bc1f5250639534d1215420f0aca2c3aa2622377270c886c56485e8dc690ccf790828705af2e659338832ed748df008060e2f938b47c9923f2446904b30abab3020f672785a35e8387ff22912e1c93f778cce578228c547bfcade5a1912067155c704df83121de3ff75fec50e67a1ad0969b0e40fb64c7753186c56977bc1ec7348a383af65c424acceff79a203cc222f9633ad4d8f878fba99bc164b11e5845673500c8c2ec5e5917040c55569fb6950337d10b5fbe73eca80be29eb02766c877bf22b4737b09ec42fbe3b0b9a3051de63e32dca713d1a05fe9641e14157ada1c500e9c45ceeb9edadb80f3baab16a92373f01f830e7600a5d1f36965d8e5ccbdc3c0e2b239227649a6f7ac9fb8e8d771698657fe5b3a4f608381c6bc1623ab9710c83970d7d880aa6cbb8d1ccdc23a7fe8e51645a0d2f6cbe434a137a49b8f5a8fc79bd4e01c7ced63ee3edc2befffcff92c246fbe512ae4d0b4ed259b7f989a98a6eadc1a2f178c84e07b28fb3a4b10cb575fa1cd2d0c25b2ce6b8b8fa4bcd4f814c6a1b75a3df7c0f67ea830d0db2ed7e45f79afc065f962a2b8a56d338bec08b107030e32cae7e62e51a7966871c2d6e8ec047720b95792ea5aa577d389f011f42bb7d1501432575c656f10ae1f13ef7287133b2cf631934525f8b9ee45814c55994028dc3425db6564b5755814bffffa2db6cdec29a9914fabaadaa370f271700aa27a750ef2db69eb540c8f9f2bf5457b41c11ca17c6ce3821fdb982193f52df9cb6d8710b247c9a59e5ee83af4e8e69574dfdd8fbc9d1083856e75ac88cdc7046ce17a10e286efa087b133a1f3d7918756b878b7ded016f470206c7bc0c51f325ec30b634026e0beb851d522fbbe0fcef664878660ea12a60bf4f3cbcd98d0078e0761668f548a164b8c0f049fce31c9433bda4d8261c5aa8963e3d2eb96ff6a02d83347651bf4d1d43865ab57d180d394588a2f78cc87bf710eabc5e110ce1a0ee3b7171844e3d1d4bf476a24ff51ef8f7e47829bf7c0d3228dd7e2a94b2a69e35686f9e6282e6561bbdce7a9bc212e91b82486d0e63f17830dbbc04ac0c7bd1a1526042db6e7a64a6070a6a8c47cd0c916671d1ad47a1a12bbd4075e4ea9b9f64aae5e9c5f1c9d2625f97f5fc5d1d50c9c0de7f91665ddbb1939058b02842c64607eb76a3d91cb2f9c17982078ec4a12e3e11fe2d6ef55796c7ef165b3ba43c079", 0x1000}], 0x9}}, {{&(0x7f0000003540)={0xa, 0x4e24, 0x0, @ipv4={'\x00', '\xff\xff', @broadcast}, 0x6}, 0x1c, &(0x7f0000003740)=[{&(0x7f0000003580)="33c3f780a5cee531f31c782e74ce2f72b3b2ae15a761fd2c4bc4f5c1bbe71c6a94666905a95fb8c8b0719fa1d57862b3cc17ba6f285303171719091a6c54d10044febb5deec37016b35848b0baac5abb90397070e73cefdccae6af6f60797b1c6339c5f3a7eabc3e4392756904", 0x6d}, {&(0x7f0000003600)="ba8eaba419a4", 0x6}, {&(0x7f0000003640)="88e1d3d29446c8f05bc4243cc4830b6cfe37e7adeb7f7e9632accc343bce7e9718108c1d1ee961b79cef34a11a2b9975763ae92ac0dae8f7d1d5bdf52341dff1e1841456146a7bccb55f11fde9d132c35e7b1b53132df80485303bf9", 0x5c}, {&(0x7f00000036c0)="49bae93ba5d20c0a0991838fc588f1d4fb4cc14361090baae6aa2efb05dc1aab2e40589b91447a8c", 0x28}, {&(0x7f0000003700)="e88b15a8c4b3b1678bfbbd16f1551a596c5f825ea1cf039ffe9ce97ab73f7bb61bf72172e0b33a657d72ef880f4226edc0be22169698c6650fc242f42378ec", 0x3f}], 0x5, &(0x7f00000037c0)=[@rthdr_2292={{0x48, 0x29, 0x39, {0x3b, 0x6, 0x0, 0x8, 0x0, [@remote, @dev={0xfe, 0x80, '\x00', 0x2d}, @private1={0xfc, 0x1, '\x00', 0x1}]}}}, @tclass={{0x14, 0x29, 0x43, 0xe4}}, @dstopts_2292={{0xe8, 0x29, 0x4, {0x2, 0x19, '\x00', [@calipso={0x7, 0x30, {0x0, 0xa, 0x40, 0xf801, [0x24, 0x6, 0xd1, 0x2, 0x4]}}, @padn={0x1, 0x2, [0x0, 0x0]}, @calipso={0x7, 0x20, {0x2, 0x6, 0x7, 0xfff, [0x100000001, 0xffffffffffff0001, 0x9]}}, @enc_lim={0x4, 0x1, 0x8}, @calipso={0x7, 0x8, {0x2, 0x0, 0x0, 0x7}}, @enc_lim={0x4, 0x1, 0x8}, @generic={0x2, 0x60, "349c1679b4ce86614ca52ca60ca97566d4fe36106af890642546b33bb70b4ade7ec08ac4da51a5e64c8499fca2db188e22a434c3a1f3da46bdcdde3ac8df60cecf9ab4819b2f2cc5c9fed84155b6a28fdee9f0e0edc3fc6aae39ede4746c157c"}]}}}, @flowinfo={{0x14, 0x29, 0xb, 0x7fff}}, @dontfrag={{0x14, 0x29, 0x3e, 0x1d85}}], 0x178}}, {{&(0x7f0000003940)={0xa, 0x4e23, 0x8, @remote, 0x5}, 0x1c, &(0x7f0000003ac0)=[{&(0x7f0000003980)="76e960", 0x3}, {&(0x7f00000039c0)="422516b8e217da13815499ac67e749e7b03eaa12772f9b926c562821027f65033ba56bc609dd38849c56ce1d1ceedda75420a578565014419affd534f79ec81dcd72e94a5fc781ec4825fd33696f27e15906a04e4fe0085d12e0c5535dcffb79b0e9f31a7b16fceb6ffc1220b54e2ddc9f50e923eba78c6baa659878861f9eecceefbd7d5386773051beba0266d0c51b82e45f9194d91f073bcd4d594887922d99905f35922b06908fb0b7e43556ccbf015734e236d22cc8272315400e56068482c89c74a1a7be8a2cb3bb8b52f0ad3da277e6f1e8f59bc698f1f86f8a539912b5c4b10d57e347f4eaa97e96ed5a", 0xee}], 0x2}}, {{&(0x7f0000003b00)={0xa, 0x4e23, 0x2, @empty, 0xa00}, 0x1c, &(0x7f0000003e40)=[{&(0x7f0000003b40)="1e6d0bd3542829a7a388a86a271c396f1458ce7e7ee2ae164e6c556b53147eb66f5b56b9c3a3fb1367d60980aaf6d86809909367bfdf6d54744926cd7868a2dbfe4d7c9a1df872cb6de5badde16c804c2e2cb5a0afc6e996056d361435a4160a8e41a148f097ca0a9fff52bbb4975d9e5490e7a84a05d15dfae07b9c081bd0f7204864d9066167b8d408e2dc95d73ade3d", 0x91}, {&(0x7f0000003c00)="975652", 0x3}, {&(0x7f0000003c40)="94375d5329768661d465af9a4de0772f688e3f3d7d72a18106a4f7e5c276575cf10a49e5946a6e6661640de5b5e2fd398b1a5c8c108dff60b8f8fc56b3094be528f85f9357fc163f23c50aa55d934372981bbd9999d1f50330eafce56cdf905cc42e564ddf0ba8f0122fa828d7105c1245dd38bb8b6c5a1d464c280bcab9868c798858ea19a31d4d45b3f2eaf82b28b7287544bdaa898f20a0f659c9cf90c0053fb82a3bacacda8a373aa5ac5961c7971d46f9526b18c55ec1f189601e88844ffc95f063c01af89614ff885f0594a87d9dd2a6a2fb3f", 0xd6}, {&(0x7f0000003d40)="064c2b3d35d1a8dd8a4050ed2bbb707ec54f9f2676d550308fdda31aa62b18e72cf7d4d3510333cf64c8defd3712a6dc12e4f15cf21b90b69c946f33eecae331d0df5f8440faf245657aaf398f9357d034809636ac9a42047e76f34a2de2f585ff51dcd4726050bf024bae4bcb8f66317e147e3b129ae0be47009b60a4555fc7b5f3a335bcc8d59d9fab4c4a50b52b59fcaafd3da0c1c62ef1f67d71f54bb74619b42af17385fdb90468646e47f8ac0df48aca82b1ab6e82144ff0dc1212a7a0df", 0xc1}], 0x4, &(0x7f0000004140)=[@rthdr_2292={{0x48, 0x29, 0x39, {0x47, 0x6, 0x2, 0xc8, 0x0, [@empty, @loopback, @mcast2]}}}, @rthdr_2292={{0x38, 0x29, 0x39, {0x84, 0x4, 0x2, 0xe, 0x0, [@private1={0xfc, 0x1, '\x00', 0x1}, @rand_addr=' \x01\x00\x00\x00\x00\x00\x00\x00\x00\x00\x00\x00\x00\x00\x01']}}}], 0x80}}, {{&(0x7f0000003f40)={0xa, 0x4e24, 0x8, @initdev={0xfe, 0x88, '\x00', 0x0, 0x0}, 0x1}, 0x1c, &(0x7f00000040c0)=[{&(0x7f0000003f80)="b8f7bc4ed73ea4bcf39bc0fc1ea8f87f7f18f3a7b30424699f", 0x19}, {&(0x7f0000003fc0)="11b1d0492a2b64c8468bd8ce87616052a4ec41ab7ac2306b2c2167aa6a233f4caef9fcd74d8e1f7aff1a3a52f3318c15df0c9946ea6b7546cf8b61b895a97a432c55da868c88387a6ac9d4664b7ce82e181b9b2a12e60629c476504084aceb3164dddc29bfecd2bccf50e26c7494fdb3d177d68be8e05bac1b54ff690deeac1018169179f16da8b0a50a4b3ed4ce96d367e1a8a6ecf351356dab84233bf21bd29b580803ff5d2b4f67778fb688c7e2e45a10e79df9cebff3ff12a988f795", 0xbe}, {&(0x7f0000004080)='l', 0x1}], 0x3}}, {{&(0x7f0000004100)={0xa, 0x4e22, 0x100, @dev={0xfe, 0x80, '\x00', 0x40}, 0x8000}, 0x1c, &(0x7f0000004300)=[{&(0x7f0000006080)="aee88b3a78a34f78d800aa4c9353c32ec59381f3897debfe1eba32de4317fb64984477b5993c6543f1249032530214916423c179c7ceebde6330defd244b8e3deaa652f271794da5df3ccc1cd7ee49ea0209ff9e3f4fb960c8c982bda5bcafb4ed15845c72c660350998dd4aa7f533d7bd9f2e158d4fba3159660c9fcbad8acf19dcca37f9a4d6d5807e9f3196de8c16", 0x90}, {&(0x7f0000004200)="a2f150e619190ed70131476b21b8cad5e1ac6c30d3db6e3e6c1264c4dc4481d4d92fba8886506eaf541b687d084b415218eb604fbae004c8b97147a55353b588854c12", 0x43}, {&(0x7f0000004280)="4a8d11d6518e58924585060db33d9afc7d939e3c69d0d27904ebb4c2a726a49617fdd09f0302eadde097c39841af2c849b1ea799dd5b15b9cd6506e4c24d85fff82c65e65142e9fea763eb743a6425d011de6687e9b7", 0x56}], 0x3, &(0x7f0000004340)=[@hopopts={{0x28, 0x29, 0x36, {0x33, 0x1, '\x00', [@padn={0x1, 0x7, [0x0, 0x0, 0x0, 0x0, 0x0, 0x0, 0x0]}, @padn={0x1, 0x1, [0x0]}]}}}, @dstopts={{0xc0, 0x29, 0x37, {0x67, 0x14, '\x00', [@jumbo, @jumbo={0xc2, 0x4, 0x4}, @generic={0x5, 0x78, "1bc18d4aec17f5a4590702e1bf57d7dde40cacfc88c5a0659e3c2fcde1bd37619730ff764056eefb783e5e03226775a0651e4193b9b9b18b0d5804914a48f759614e40f8c0cb4d2a93b6d9a5d9cd42bd2ba2da4aa524b60dcb8e558d2e11a56d0e25c57c56b16bc2ecea51647cfc3ea20321d402399dc76f"}, @enc_lim={0x4, 0x1, 0x1}, @hao={0xc9, 0x10, @rand_addr=' \x01\x00'}, @ra, @enc_lim={0x4, 0x1, 0x7f}]}}}], 0xe8}}, {{0x0, 0x0, &(0x7f00000046c0)=[{&(0x7f0000004440)="53fd5f89addc7638eef2a05a0c1d4a88a98dacb4cee3c24021b591be0de5fa732cea65b247aaeedd2e15056e1c64eb1ea06dbaffa2e05b72e51dcd7d7478d64ce2dd8675cdee3bb7be9a62fd1ec35bbab54adfd3eb1bc41d95ed3105ed985afb538d", 0x62}, {&(0x7f00000044c0)="d0fad7888f66de898467003e8ea2f49c9c1fcf33bfa040c794998df539aa496887cd9562b2e154c8e22a672e6608b86102d57345965673c6db21230672a9f11009a282c49c2f448c44817172b7ddfc1a4a43aae43db970a7062e7dc4dda23b747db9e3a477", 0x65}, {&(0x7f0000004540)="8c6ba3cc62c973bd7f7c6b118a846ccd432f5e5aa15960f6c4a8f09391898d5c357a3b0fd96874fa86ac6cba8965bb6411215106fdcd0b23652ab882c963a8bc4898f83a542fa7f4ce94410ecf9e0d8be0f9b9c716a01179aa274414fbe44b47943668228005ce2f50f53030460166bdd744f0e659de", 0x76}, {&(0x7f00000045c0)="3411fb6a69dff242cf21daf3b3d7042581803cc4a8163ab8bfd84e548749a3bd95c80849afd82b912e71f9232d095e50aeda6e4b27a78eb508031fe5287f50b255ec3539a4638e4d9492b95a1ffc8b528574f84e271bb8edb33b4595edf1703ff0f487ef62096233df247450c9af1df76f796d9ba8d84aee76e6a03aa5e86d01f2c8d381b5d53134511a94b982fb43765e783a7b54f09e2388f3696dc867b0ec4d5d2848ec1b9cccb96fb4489dd29f54c7cb580832efac9ec833f4c06576cdff18f6b4b1e2c70aa509e21c532860c1fd0a0385c5868f4a96923d1cf062fa354a5cc252b33cad33505d4f6b32d88dfcca0ea440539d", 0xf5}], 0x4, &(0x7f0000004700)=[@hoplimit_2292={{0x14, 0x29, 0x8, 0xfffffffe}}], 0x18}}, {{&(0x7f0000004740)={0xa, 0x4e21, 0x676, @mcast1, 0x8}, 0x1c, &(0x7f0000005980)=[{&(0x7f0000004780)="1e3a7c60b47a5e6f72dd1fb43385416966fb09552c66fdb67d9908bcbe25045a5c5ed574b34c2eddf074ba440c79d3a870d9e5ca60ab253b0a03b719dc649a766f40d0eceac609440f5b4b46df14a868852468d80d345c85e1b8f44fa1bd2f33b279ac2f7f17cef809d379315db2eeb475422e42848cd0258874c27c69cac246543f03ea83110cc12b1cd01dd3c88bb65a7126296a856092f47fbcb3f7bef0bfe25e3aaddc41d68c67874dbf332d103bead1bd7fc010aafd346e2189cd3822cec0bf8f9cdd7be0ae7b9257fd1bb53a5fa0e664207ab2b56d4dd71f09f062afb7f1c593ce51a4bba135cae0fa1de757a64a6f5fea440d2bfad001023a4f28e2450ef41e14919ec3b623e7fbfce1f6d5de774de1712f48bc6c3e11ebab952d3f097f060c82ff54aa12b261e44f51f58e27db622c04c5a97d5e24f1da771b516aba024fdc8abf94bd20a674e3aa5bab85756a2318d4126c474609b77286fe9fedf7bd0c6d0c96a3fc65cf91911fe0cf2f2956df830a7f2ef23d43307a07e665b50d999dc27ff4da79065f2dd750c9fd5d7ba0733b8e3d0985050f5338880a8604d04b486396695a79e78256881ee7d7e2650fac0d57a54cbb43280f6097adaa14b07590ba2b80b2d97d7ee94353be1f53b78842bbd3133cf31b2acf82f520a3a9a3cd694e140dc70fdeb63af3bbd6a9d263ff8e973223197014aa68b98d99e61b4b67afe52cea5c04dfabb3c80de0590bff993654e0129ca1497ca97646937434f2cfda4817bb8b0b0f1f7bbe1efcf27daa3b5794084c0df8eb654f49516be8659411cb1f8ec404f119a5f92fe1085a6291cdf37b5095f4ee4073e60d418600cea4afe684cc4e70dbba23286ca50c400853f2c6629a72935a7afed74e5045e52a8b872b533757ae0be2b2fd758781ba0e25e77e67222b191584091646b3b56474cdbc70602655464d95615b99179cb3ef8af5b42171cbc6e05ba076a52adb170a94dfbcdc8c6f7444c417a141be499b5e89f3a4d33fb851d861f14065ad1e8065a7f68b2d3ee420fd1f37fbed56beab590415f392b8996d5a035e72edea7ac769b5c1393c7374f9a6d89bf727edbba921061d252fa53d9459384dcd3c96450327cbbe28d3315cf24b1ddbd3a691fcced3d5d833fbc2c971f959e63f85ae81aba96e70cf26e5ad08af2eb5464a6738c28eaf0371b6639cefffc11aa8347f0aeb9f7086ec070dee59ab12286ee3ee385cde46f29057b4150bdd1acb0ed301d34b49885afe4861ee1da60cdce3279f45d30d05eeb43019e0d40c2fd48c7590d8ec751074c40927b6e83abbf8ecc1a15f3ad9def1d44573929c8d564cd3e298b689aaac9763c0b714c82bafed329685fa0b40607462c65bbb62b288f592a82723fea9f827b523c7193258388ec9912fab7bfb061b311a6a4869250d726f83f7dcc27ca54c788afc131355dc59e791d4c15d2c7961ce91e1c0303dcee7fc8ed497cf5a9696d750af581711d8229120fab599bab00e7afbe66d4f0e75ed5ec9c734a572a13e2747098ad850af93d9fa5c58530ea9fc070149ff1d1c59966fd3e44d540ef55fe89ea0c0a1cfaa74e2b514d916ea39a7dacd19a990a227bc9eb476626d4eae213aa7133f201afd968a2dcff350f685e0ee07956fe6090b5c26349387ab671ed00d43f41ed1edb228046e7eeb3a7e4203f0149f1ca529ada2052e48a22edf88c575a094e46d57b07fb979eed8bedd9c72f9fd334b6e422a873b82db6b2496c88bc8991d09c64464866b192ae2312391778fd74ee12fb6614fd57d46931f48b8777b576fb196fe87680b8620aacc4c71596de956c2f46ff653fe126debcbf20b353d5b015604f5546f29bbd770f6594d3b6a17090131780319883e3e7b1b40b13b3cb9df066b6e4571dab487738ea98795d0e0e176e5dcae2280c37523c209d0e3387b7643a5f5b850be5d8e8d6f5a2b1a8e8edb351ad91d78b179fb1701e2a9627a0c290cf4ad7efb6330b045364227e11cfd71843979003c1b39ae7a9cbe665406d4b49d1e14bf51a68d6d55165a6b908771dfe41d93075a5c84796a8bab042f0bfd02f8c7b7511c85f367a48074d1181d29b69fc18822bf53ba4c3a2247379bbb5aeb51f17a1e0daedbc8af7829847dbce17af116f518f1fcfcf2165742985580b68e371ceaae4e1077dc816cfe398cbceeea0c3eec4b5a4ba33daa5578e3d41a24fe1eeea45c29ea840c782c2b666efb991663088ceae71851ded61aef80035e31303b596907ccaab7eab977225be355a701492c78e0b5f4d1914aef34f6ff20cdf4f4ccc23695b35585d2503b6a9e741a3ba68073ddb5df93f55ee806f1036b3ee6f346f2fcae98ac2e94d2b1f0245e6079781a6a9adcc42b6456b7d7804e2925874ad4e80cda367791028945ba6e564fd3d2b977ab2f248bdbc2640cba532e273180b4c708ea10715d8f55c453defc24a7939bef74f0d0fc56d4f22f960cc03069d6b9be1eacec23ae9c31a82726a250c9d25afcd645414262731797fa92c0b0ff4bdfc81e1d607be72d949a45886920d23360c98cb47d58190b60f5c3af2d51d25763ef6c04d33a7cfc87f1ebddd14d706c61182b98acc6f0310ef4b72e0719a3dda86d135600423310f9ab6610daad135ff1bbb9dc4af14b09022f99d73bc4589748b3e1ec3c157c5d97608e63e97690e864c31635a7204150c45c70a8fa7670409bd5e30670b703ecb735415b1a743f186c4a0c84edd7274ad583ec2ddd3e67e5e760db74914c9ef3003e6c512f0b9c6d26830a8d9eac9b540b5929f3eacc9279c872c8698cb2184b1c8f2648e20a5c4c5ad13fba630c885e71ff26cdea26319c401aead72081418c161910b30932b4a250bafa3ae69b9e4b5a98fc2efa57eee8e7194d7dd2af9778ce822781039f951cdfe7237ba1d9e572d898ca51866b199aebb2e512a8e4fc9a80320ee0d9e3f093e086aec0bf20759e43bf8ca637bd9f61e78026da737475685b9371730f810bb9d2c15f0f52f2598b529200e7c669e42f4aa00ac9542672b8073cbc56c525d19684930e2d8f79bfd1d67b35a888e7947f8cc1ed00b9b4986d7c2aca6c4111da79cd85bcfb584e8d1ddb3838a08d4bb099827316b73cd7909f6c306642d4df87935ec815d04622968d27f63e8c1e2300692f5c19bee3237eeb8f76f89d0d267213eda627a8fc821100510e939286c9fda33e5680d86692a05a0f6223842e649f4b06dbf6af0285d8f86396683ad8917c898b5d9431b1779d28911507537ece29a86093487208ae71fc8488d655a9352049e4e7d134ecfe683a372c77c7762dfeb48960a2b60efeb8efc9163ece1042baa952fe329be36ebca6466ae52ff70190f95ebdc5671cd6e88cff552631230ecdea82f1dec08a391fed26981a6907999c45f9d75487226d32d3adcbebe9a7987d3486f198e1670b7547c3c15807f6f853f697b85999c7aae5a7623358d285e5ee82e7e0dbbd3e8b5f87c84148134da1aba1553b5060a887e193ce0607a2072c76cf4740b452e5c3802eeeb55a2703af0b642d4591c1de25734248001ef2147e660de8d0ad975f006ec836936aae07c6e169ed4bda6eaf8edc75defc8f00c2954a86abe25087fba178d837d0e7974575fea20c96eb32dd4dcca981ba08f11f1047f11a28fcd52521f98e8babd1b86006817a1a24a13540755dda395df961eda7ca3fce7337c3384f8ace09dac94801c5e50a634c38193682b9173f4398df86e6e97713ac3b9717b44eb60f27d72ad1b92aacd1a66c7f9970acf886f0eaa1d5203388f8dfe9fc6b1f07b206bbabc4cd982b5dba34b15801dcaab907506890c4955a3b76bd93cc0e3d12403bd3a66754909fc14447f02eaf9717b219c9280f7c32d1f5425f05a4871b40e9cf8a5dcdc7d8c3fb63556d44cca59556a9245ee20e982e98251a464b757793ef6c126c6003eef73393a5cc05e6ec1e49016f33f7f9b76064dd04021645402d49a6453e63df4a29a84fc1b397aa922f9fcf755c06ba45df2eff2554ce555b395b1f795816179224359842d1f95836224a9c1f3e58815daac0d19bfc08c8e1d06c3c5de227c68cf2d11bc5233a481048ed657114a2da7d4e7068f87eea91d4aa4a2c469616aeaf8e490a5c1e444689b0d33fe57d815ff1020f94a2da597e02abe324fdb760cfd4d5bce89a95f7fa230ed0c8448abc0688417e5cedd095c6be83f024ec7f6091558fb8f83b25b0a3f6ca12966105ca76217a22b222077a67cf853d4b33322fa81978cacd97e72712b9efedb4ba740565440c49ecfc37d54f42ed5818012b54db628ec5507fb30df55b1ff522c5e4f246f171df146f620a914673e68b4a838198a459b59f2e65ab9e27d3fc4fade7458f561a71e80313995a2ac01e55299c5fb5c6aa0c7391e633f3c99e815b3e6c6c947e9f0d0545d930fac6ff59de06cd451e63c7980bab09ac97a556ae69cb5ea4bd10129eb36b4d1edcc0e97fbd4d095a18d91f5a07532a5b385d7d2c043537b627f6a1a972073244653f173c7f119ddd88e16c1c3336d81bcfdbdce5411e5fa4079913a330205dce4785f218aa928d8b81ec93a5cb950606187fd5c234bdcc06467bcca26434d33d64f2add9396d8cd783a0ce4cc81fa2a1eb13c80c2e35d40750dca170a9b7935247196cf5a95f7e429d91a2c77e9e5439b2f9fbf4a98091f0ea359d0483e34cc14a089b5ac8c92c3982e5b3b526049d12a2a8f88cafb45a00b46a9a9b15a5d3ecc7de10a386ac7bae1d53517dc65647df3eef816a42b30cabf7e35ed027badc5caa87b80a93c89f705b8cb233b2d24abf1c181001c911fec6fc7ec7a49c0ec24fecc5eefcf3a72f3deee7b8615f6702f7b10b60b8f41a53c3a0f7071f805b2180b771b89da65d4c4f484c2b3aeb43018a7a888ab45cbd691f3b32a23db7303ea1fc9f48194223462530fa6a2f69535aa6de8f735789bf346708cf5dfb9e498097dac2ca0209cad47ab6d5c4d7ff837819c615c5edbfb7efebc885b062efa27920a396f7ed0ae695a3db25988a0b3fae8e7656ea343a341c85cadf0ffeb54e341af736ea1fe129e1234579ee9640a993efb8151a32bcce94a25afd0765d82e7f05e128e644c0e2e4021a2c346732fccde5a6b6ccc45dfbb1e0f32a9d3e7afe2b97212c48996b6e07e80d86a9af4fdbc7d2b7ca7cdbcf78bae662cd6cc6a314dddb0590ff9b4b4c92976b553466d201ba6176c0c13eaadbedbe940958a16573005d227229b5e2917a09894b047bb0678f154955f214dc12cae1cde859317fc573fb517ae168ba2902888202caaee9c1dcd546524eb1e4a0c7043543981e9ab6723666c80f30a4921a4d40282a644ec9a5b8f458efcf384080fc71587e37df31c7de59dd5245fe5c9577cd35a2e398c7be48fc807b380b698b9c3defc8c9aa690df614686e95c7c7bf9d725d85da93bb544c9acdf81eddad2f279e121c52ca14e6f8e165ee549f48a517432849d0857acf6a19aa35147995ea34dfbb9dde1ec64e14ff12ac7624dc3d4f7195f21bddce38de911d20f21afd5ef21eaa8a2d929b9d12da532445a5fcb13dcc158293cb40267f84fe1cae4fa64d470105f1490ca5ee8240866348ea25890b12e5861ff7ad533f4da3dc6c68ef3c71ece81b2b358cc33a9c026086dc7c82256e801b0c03ca849013617baa79b9f0ff0649a667bcab10c99a344f425913deb6ebeda5f8859dae2342c680044bf856ba444847aac070495c29adf6952c8490c9489c926a7ae0bb4166e392ae03677c", 0x1000}, {&(0x7f0000005780)="725a14d768bae84499bd5b528fbb97a71e321c050fec8587c0f854fb2fc86ccf58232313532f0f64d002e9b128e5fc62ea1bd6432dad0ea752ad0c8f346e9df38b3873fe9b720010973ffb0f39c5597746a91dcd55c46ae947e6aa4062072e8193a76911e930147b3c011bfb20a46b", 0x6f}, {&(0x7f0000005800)="d6268a5528f585e113399f0c3ad37b5bc14a4608a86a42dd9134a96774356c2345d98b95390d4bc0439c067d36af24011e2344756b963d48da70d749f796976a07fd5f7b0419de30fe9662e4f3da46be21e0a68a2012fdc7f70c505b553ab07968770d86895c73fa1d46d272bf448f0b6b85fbd0966b7b6395c501e00549c3f7114f8693b21cc1d3a7f6655eb1a2ce122feedb0dbb484b4dd806effed2741d5b1bf0adaf0d36f67e5e2356c4ff2fc5b9e88f3c76b495b8658b6e4870f19634a6ecc8cc4888fe3308436aad", 0xcb}, {&(0x7f0000005900)="9ac94d9d22868b60ec510aa10601a340d72400314797f37b48a3a986ac76b52241f0122bc3692dd97e0eae6aded464b35b2763be5d83fd0d9a19af4a23beb7875f0cdfbd1f6cbb1649252ab851e23584a400d590b55fb2dd7fbc727b68882311148bea93ed8aa7e583709d80eb771228ca01cf5601d5928b885845", 0x7b}, {&(0x7f0000005a80)="f482c32a06a41039cf295e832cb0d32ef9756d12ff47602b893dc14d6d445451604684bc1f57f94e0ceca45d20a751e558466387b8bfc4487472d8e452d137eec82175f984d589239512a29d4f2a1cd3decdaed1ddb47fee6e37f2c506c805f64879dbda072bba47e45d784f923cda42b2a5274b5797b37e3c3599035ee1298db83940f4173625f6f57f94608cd333f636d0e514aa09df886a1d0fefb88671af71159aecfb69eb2fdc8c39a4f53842431ed4545ef6221849f0edd1d30ade334aa4bdd67f3bb6ee6de8cd0f83cbd7301f21a1a335500d6a2d130dc456b45e26efd6c0bd17ff7003b66f4a1d7acbac41c10069ba31583f", 0xf6}], 0x5, &(0x7f0000005b80)=[@hoplimit_2292={{0x14, 0x29, 0x8, 0x1}}, @hoplimit={{0x14, 0x29, 0x34, 0x3}}, @dstopts={{0xa0, 0x29, 0x37, {0x87, 0x10, '\x00', [@calipso={0x7, 0x18, {0x2, 0x4, 0x7f, 0x2, [0x304, 0x4]}}, @calipso={0x7, 0x30, {0x0, 0xa, 0x5, 0x0, [0x600000000000, 0x2, 0x240000000000, 0x2, 0x1]}}, @padn={0x1, 0x3, [0x0, 0x0, 0x0]}, @enc_lim={0x4, 0x1, 0xc3}, @jumbo={0xc2, 0x4, 0xf26}, @hao={0xc9, 0x10, @rand_addr=' \x01\x00'}, @enc_lim={0x4, 0x1, 0x20}, @hao={0xc9, 0x10, @loopback}, @padn={0x1, 0x2, [0x0, 0x0]}]}}}], 0xd0}}], 0x8, 0x1) [ 2812.519072][ T2618] workqueue: Failed to create a rescuer kthread for wq "bond1287": -EINTR [ 2812.731265][ T2648] netlink: 'syz-executor.2': attribute type 1 has an invalid length. 04:12:20 executing program 1: r0 = socket$netlink(0x10, 0x3, 0x0) r1 = socket$netlink(0x10, 0x3, 0x0) r2 = socket(0x10, 0x803, 0x0) sendmsg$NL80211_CMD_CRIT_PROTOCOL_START(r2, &(0x7f0000000580)={0x0, 0x0, &(0x7f0000000540)={0x0, 0x1c}}, 0x0) getsockname$packet(r2, &(0x7f0000000600)={0x11, 0x0, 0x0, 0x1, 0x0, 0x6, @broadcast}, &(0x7f0000000100)=0x14) sendmsg$nl_route(r1, &(0x7f0000000040)={0x0, 0x0, &(0x7f0000000000)={&(0x7f00000000c0)=ANY=[@ANYBLOB="3c00000010008506eb9afc4cd8d06e754a0081c5", @ANYRES32=r3, @ANYBLOB="2377f292252155b21c0012000c000100626f6e64000000000c0002000800010001"], 0x3c}}, 0x0) sendmsg$nl_route(r0, &(0x7f0000000240)={0x0, 0x0, &(0x7f0000000680)={&(0x7f0000000400)=@newlink={0x3c, 0x10, 0xffffff1f, 0x0, 0xa030000, {}, [@IFLA_LINKINFO={0x14, 0x12, 0x0, 0x1, @bridge={{0xb}, {0x4}}}, @IFLA_MASTER={0x8, 0xa, r3}]}, 0x3c}}, 0x0) [ 2812.827830][ T2648] bond1268: entered promiscuous mode [ 2812.838333][ T2648] 8021q: adding VLAN 0 to HW filter on device bond1268 04:12:20 executing program 4: r0 = socket$inet6(0xa, 0x2, 0x0) r1 = socket$inet6(0xa, 0x80002, 0x88) bind$inet6(r1, &(0x7f0000000580)={0xa, 0x10010000004e20}, 0x1c) setsockopt$inet6_udp_int(r1, 0x11, 0xb, &(0x7f0000000080), 0x4) bind$inet6(r1, &(0x7f0000000080)={0xa, 0x4e23, 0x8825, @private1={0xfc, 0x1, '\x00', 0x1}, 0x10000}, 0x1c) ioctl$sock_SIOCETHTOOL(r0, 0x89f0, &(0x7f0000000000)={'bridge0\x00', &(0x7f0000000040)=@ethtool_rxfh_indir={0x38, 0x9, [0x10000, 0x7, 0xfff, 0x20, 0x1, 0x8, 0x200, 0x0, 0x9]}}) [ 2813.001139][ T2649] bond1268: (slave bridge1189): making interface the new active one [ 2813.039126][ T2649] bridge1189: entered promiscuous mode 04:12:20 executing program 0: r0 = socket$inet6(0xa, 0x6, 0x0) listen(r0, 0x6) r1 = socket$inet6(0xa, 0x6, 0x0) connect$inet6(r1, &(0x7f0000000200)={0xa, 0x0, 0x0, @initdev={0xfe, 0x88, '\x00', 0x0, 0x0}, 0x3}, 0x1c) connect$inet6(r1, &(0x7f0000000040)={0xa, 0x4e20, 0x0, @dev={0xfe, 0x80, '\x00', 0xf}}, 0x1c) sendmmsg$inet6(r1, &(0x7f0000005a00)=[{{0x0, 0x0, 0x0}}, {{0x0, 0x0, 0x0, 0x0, &(0x7f0000000080)=ANY=[@ANYBLOB="5000000000000000290000003600000000060000000000000401000730000000000a0000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000020000000000000002900000037000000000000000000000004010000000000001400000000000000290000000004000160000000000000002900000037377d078f6cc92e9df3d00238a93cdd737056f6857bc4de06f0510bdc0926b2a14c881dffbaff8096d219a55e2b86a677608cea03f8449dddd9222fc701eba2461a67b84f84e25e5710941f8f2fee35df5fa96b8ec9c8610e793d84bad92cd7c6bfac6b"], 0x100}}], 0x2, 0x0) [ 2813.125447][ T2649] bond1268: (slave bridge1189): Enslaving as an active interface with an up link [ 2813.177594][ T2657] netlink: 'syz-executor.5': attribute type 1 has an invalid length. 04:12:20 executing program 0: r0 = socket$inet6(0xa, 0x6, 0x0) listen(r0, 0x6) r1 = socket$inet6(0xa, 0x6, 0x0) connect$inet6(r1, &(0x7f0000000200)={0xa, 0x0, 0x0, @initdev={0xfe, 0x88, '\x00', 0x0, 0x0}, 0x3}, 0x1c) connect$inet6(r1, &(0x7f0000000040)={0xa, 0x4e20, 0x0, @dev={0xfe, 0x80, '\x00', 0xf}}, 0x1c) sendmmsg$inet6(r1, &(0x7f0000005a00)=[{{0x0, 0x0, 0x0}}, {{0x0, 0x0, 0x0, 0x0, &(0x7f0000000080)=ANY=[@ANYBLOB="5000000000000000290000003600000000060000000000000401000730000000000a0000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000020000000000000002900000037000000000000000000000004010000000000001400000000000000290000000004000160000000000000002900000037377d078f6cc92e9df3d00238a93cdd737056f6857bc4de06f0510bdc0926b2a14c881dffbaff8096d219a55e2b86a677608cea03f8449dddd9222fc701eba2461a67b84f84e25e5710941f8f2fee35df5fa96b8ec9c8610e793d84bad92cd7c6bfac6b"], 0x100}}], 0x2, 0x0) socket$inet6(0xa, 0x6, 0x0) (async) listen(r0, 0x6) (async) socket$inet6(0xa, 0x6, 0x0) (async) connect$inet6(r1, &(0x7f0000000200)={0xa, 0x0, 0x0, @initdev={0xfe, 0x88, '\x00', 0x0, 0x0}, 0x3}, 0x1c) (async) connect$inet6(r1, &(0x7f0000000040)={0xa, 0x4e20, 0x0, @dev={0xfe, 0x80, '\x00', 0xf}}, 0x1c) (async) sendmmsg$inet6(r1, &(0x7f0000005a00)=[{{0x0, 0x0, 0x0}}, {{0x0, 0x0, 0x0, 0x0, &(0x7f0000000080)=ANY=[@ANYBLOB="5000000000000000290000003600000000060000000000000401000730000000000a0000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000020000000000000002900000037000000000000000000000004010000000000001400000000000000290000000004000160000000000000002900000037377d078f6cc92e9df3d00238a93cdd737056f6857bc4de06f0510bdc0926b2a14c881dffbaff8096d219a55e2b86a677608cea03f8449dddd9222fc701eba2461a67b84f84e25e5710941f8f2fee35df5fa96b8ec9c8610e793d84bad92cd7c6bfac6b"], 0x100}}], 0x2, 0x0) (async) 04:12:20 executing program 2: r0 = socket$netlink(0x10, 0x3, 0x0) r1 = socket$netlink(0x10, 0x3, 0x0) r2 = socket(0x10, 0x803, 0x0) sendmsg$NL80211_CMD_CRIT_PROTOCOL_START(r2, &(0x7f0000000580)={0x0, 0x0, &(0x7f0000000540)={0x0, 0x1c}}, 0x0) getsockname$packet(r2, &(0x7f0000000600)={0x11, 0x0, 0x0, 0x1, 0x0, 0x6, @broadcast}, &(0x7f0000000100)=0x14) sendmsg$nl_route(r1, &(0x7f0000000040)={0x0, 0x0, &(0x7f0000000000)={&(0x7f00000000c0)=ANY=[@ANYBLOB="3c00000010008506eb9afc4cd8d06e754a0081c5", @ANYRES32=r3, @ANYBLOB="2377f292252155b21c0012000c000100626f6e64000000000c0002000800010001"], 0x3c}}, 0x0) sendmsg$nl_route(r0, &(0x7f0000000240)={0x0, 0x0, &(0x7f0000000680)={&(0x7f0000000400)=@newlink={0x3c, 0x10, 0xffffff1f, 0x0, 0x253c27, {}, [@IFLA_LINKINFO={0x14, 0x12, 0x0, 0x1, @bridge={{0xb}, {0x4}}}, @IFLA_MASTER={0x8, 0xa, r3}]}, 0x3c}}, 0x0) [ 2813.400179][ T2657] bond1278: entered promiscuous mode 04:12:20 executing program 0: r0 = socket$inet6(0xa, 0x6, 0x0) listen(r0, 0x6) r1 = socket$inet6(0xa, 0x6, 0x0) connect$inet6(r1, &(0x7f0000000200)={0xa, 0x0, 0x0, @initdev={0xfe, 0x88, '\x00', 0x0, 0x0}, 0x3}, 0x1c) connect$inet6(r1, &(0x7f0000000040)={0xa, 0x4e20, 0x0, @dev={0xfe, 0x80, '\x00', 0xf}}, 0x1c) sendmmsg$inet6(r1, &(0x7f0000005a00)=[{{0x0, 0x0, 0x0}}, {{0x0, 0x0, 0x0, 0x0, &(0x7f0000000080)=ANY=[@ANYBLOB="5000000000000000290000003600000000060000000000000401000730000000000a0000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000020000000000000002900000037000000000000000000000004010000000000001400000000000000290000000004000160000000000000002900000037377d078f6cc92e9df3d00238a93cdd737056f6857bc4de06f0510bdc0926b2a14c881dffbaff8096d219a55e2b86a677608cea03f8449dddd9222fc701eba2461a67b84f84e25e5710941f8f2fee35df5fa96b8ec9c8610e793d84bad92cd7c6bfac6b"], 0x100}}], 0x2, 0x0) socket$inet6(0xa, 0x6, 0x0) (async) listen(r0, 0x6) (async) socket$inet6(0xa, 0x6, 0x0) (async) connect$inet6(r1, &(0x7f0000000200)={0xa, 0x0, 0x0, @initdev={0xfe, 0x88, '\x00', 0x0, 0x0}, 0x3}, 0x1c) (async) connect$inet6(r1, &(0x7f0000000040)={0xa, 0x4e20, 0x0, @dev={0xfe, 0x80, '\x00', 0xf}}, 0x1c) (async) sendmmsg$inet6(r1, &(0x7f0000005a00)=[{{0x0, 0x0, 0x0}}, {{0x0, 0x0, 0x0, 0x0, &(0x7f0000000080)=ANY=[@ANYBLOB="5000000000000000290000003600000000060000000000000401000730000000000a0000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000020000000000000002900000037000000000000000000000004010000000000001400000000000000290000000004000160000000000000002900000037377d078f6cc92e9df3d00238a93cdd737056f6857bc4de06f0510bdc0926b2a14c881dffbaff8096d219a55e2b86a677608cea03f8449dddd9222fc701eba2461a67b84f84e25e5710941f8f2fee35df5fa96b8ec9c8610e793d84bad92cd7c6bfac6b"], 0x100}}], 0x2, 0x0) (async) [ 2813.438072][ T2657] 8021q: adding VLAN 0 to HW filter on device bond1278 04:12:21 executing program 0: r0 = socket$inet6(0xa, 0x6, 0x0) bind$inet6(r0, &(0x7f0000000000)={0xa, 0x4e20}, 0x1c) listen(r0, 0x6) r1 = socket$inet6(0xa, 0x6, 0x0) connect$inet6(r1, &(0x7f0000000200)={0xa, 0x0, 0x0, @initdev={0xfe, 0x88, '\x00', 0x0, 0x0}, 0x3}, 0x1c) connect$inet6(r1, &(0x7f0000000040)={0xa, 0x4e20, 0x0, @dev={0xfe, 0x80, '\x00', 0xf}}, 0x1c) sendmmsg$inet6(r1, &(0x7f0000005a00)=[{{0x0, 0x0, 0x0}}, {{0x0, 0x0, 0x0, 0x0, &(0x7f0000000080)=ANY=[@ANYBLOB="5000000000000000290000003600000000060000000000000401000730000000000a0000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000020000000000000002900000037000000000000000000000004010000000000001400000000000000290000000000000160000000000000002900000037f0789d733dae4ddd1e8a618b11df504abf89ab7e9c2ecbb4d3b8e21dffcdcccaff393a2f55c5faf8ce78373b31033f662920cad3859ed2c67fe252a032d8b010628aba6fe39a8971bfb79f6adf3533197e3299176fd298d31e1b68feea3190a394208986c26db4fb4a6713cf708ab908ea70c1f2cebfe15c9a58f392ab9c9f7ea95a758a95d479a2d39213a2e35ce2caedfadd2c2e5a72e301539101b20411d2a280d6b9a71bfb5ddb019997b93864fb4de3"], 0x100}}], 0x2, 0x0) [ 2813.641317][ T2667] bond1278: (slave bridge1241): making interface the new active one 04:12:21 executing program 0: r0 = socket$inet6(0xa, 0x6, 0x0) bind$inet6(r0, &(0x7f0000000000)={0xa, 0x4e20}, 0x1c) (async) listen(r0, 0x6) (async) r1 = socket$inet6(0xa, 0x6, 0x0) connect$inet6(r1, &(0x7f0000000200)={0xa, 0x0, 0x0, @initdev={0xfe, 0x88, '\x00', 0x0, 0x0}, 0x3}, 0x1c) connect$inet6(r1, &(0x7f0000000040)={0xa, 0x4e20, 0x0, @dev={0xfe, 0x80, '\x00', 0xf}}, 0x1c) sendmmsg$inet6(r1, &(0x7f0000005a00)=[{{0x0, 0x0, 0x0}}, {{0x0, 0x0, 0x0, 0x0, &(0x7f0000000080)=ANY=[@ANYBLOB="5000000000000000290000003600000000060000000000000401000730000000000a0000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000020000000000000002900000037000000000000000000000004010000000000001400000000000000290000000000000160000000000000002900000037f0789d733dae4ddd1e8a618b11df504abf89ab7e9c2ecbb4d3b8e21dffcdcccaff393a2f55c5faf8ce78373b31033f662920cad3859ed2c67fe252a032d8b010628aba6fe39a8971bfb79f6adf3533197e3299176fd298d31e1b68feea3190a394208986c26db4fb4a6713cf708ab908ea70c1f2cebfe15c9a58f392ab9c9f7ea95a758a95d479a2d39213a2e35ce2caedfadd2c2e5a72e301539101b20411d2a280d6b9a71bfb5ddb019997b93864fb4de3"], 0x100}}], 0x2, 0x0) [ 2813.714960][ T2667] bridge1241: entered promiscuous mode [ 2813.769632][ T2667] bond1278: (slave bridge1241): Enslaving as an active interface with an up link [ 2813.836640][ T2670] netlink: 'syz-executor.1': attribute type 1 has an invalid length. 04:12:21 executing program 3: r0 = socket$netlink(0x10, 0x3, 0x0) sendmsg$netlink(r0, &(0x7f00000047c0)={0x0, 0x0, &(0x7f0000000000)=[{&(0x7f0000000100)=ANY=[@ANYBLOB="18000000560001060000006c00db730007"], 0x18}], 0x1}, 0x0) 04:12:21 executing program 5: r0 = socket$netlink(0x10, 0x3, 0x0) r1 = socket$netlink(0x10, 0x3, 0x0) r2 = socket(0x10, 0x803, 0x0) sendmsg$NL80211_CMD_CRIT_PROTOCOL_START(r2, &(0x7f0000000580)={0x0, 0x0, &(0x7f0000000540)={0x0, 0x1c}}, 0x0) getsockname$packet(r2, &(0x7f0000000600)={0x11, 0x0, 0x0, 0x1, 0x0, 0x6, @broadcast}, &(0x7f0000000100)=0x14) sendmsg$nl_route(r1, &(0x7f0000000040)={0x0, 0x0, &(0x7f0000000000)={&(0x7f00000000c0)=ANY=[@ANYBLOB="3c00000010008506eb9afc4cd8d06e754a0081c5", @ANYRES32=r3, @ANYBLOB="2377f292252155b21c0012000c000100626f6e64000000000c0002000800010001"], 0x3c}}, 0x0) sendmsg$nl_route(r0, &(0x7f0000000240)={0x0, 0x0, &(0x7f0000000680)={&(0x7f0000000400)=@newlink={0x3c, 0x10, 0xffffff1f, 0x0, 0x2535f8, {}, [@IFLA_LINKINFO={0x14, 0x12, 0x0, 0x1, @bridge={{0xb}, {0x4}}}, @IFLA_MASTER={0x8, 0xa, r3}]}, 0x3c}}, 0x0) 04:12:21 executing program 0: r0 = socket$inet6(0xa, 0x6, 0x0) bind$inet6(r0, &(0x7f0000000000)={0xa, 0x4e20}, 0x1c) (async) listen(r0, 0x6) (async) r1 = socket$inet6(0xa, 0x6, 0x0) connect$inet6(r1, &(0x7f0000000200)={0xa, 0x0, 0x0, @initdev={0xfe, 0x88, '\x00', 0x0, 0x0}, 0x3}, 0x1c) (async) connect$inet6(r1, &(0x7f0000000040)={0xa, 0x4e20, 0x0, @dev={0xfe, 0x80, '\x00', 0xf}}, 0x1c) (async) sendmmsg$inet6(r1, &(0x7f0000005a00)=[{{0x0, 0x0, 0x0}}, {{0x0, 0x0, 0x0, 0x0, &(0x7f0000000080)=ANY=[@ANYBLOB="5000000000000000290000003600000000060000000000000401000730000000000a0000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000020000000000000002900000037000000000000000000000004010000000000001400000000000000290000000000000160000000000000002900000037f0789d733dae4ddd1e8a618b11df504abf89ab7e9c2ecbb4d3b8e21dffcdcccaff393a2f55c5faf8ce78373b31033f662920cad3859ed2c67fe252a032d8b010628aba6fe39a8971bfb79f6adf3533197e3299176fd298d31e1b68feea3190a394208986c26db4fb4a6713cf708ab908ea70c1f2cebfe15c9a58f392ab9c9f7ea95a758a95d479a2d39213a2e35ce2caedfadd2c2e5a72e301539101b20411d2a280d6b9a71bfb5ddb019997b93864fb4de3"], 0x100}}], 0x2, 0x0) 04:12:21 executing program 0: r0 = socket$inet6(0xa, 0x6, 0x0) bind$inet6(r0, &(0x7f0000000000)={0xa, 0x4e20}, 0x1c) listen(r0, 0x6) r1 = socket$inet6(0xa, 0x6, 0x0) r2 = socket$inet6(0xa, 0x80002, 0x88) bind$inet6(r2, &(0x7f0000000580)={0xa, 0x10010000004e20}, 0x1c) setsockopt$inet6_udp_int(r2, 0x11, 0xb, &(0x7f0000000080), 0x4) r3 = accept(r0, 0x0, &(0x7f0000000100)) setsockopt$inet6_tcp_TCP_MD5SIG(r3, 0x6, 0xe, &(0x7f0000000240)={@in={{0x2, 0x4e24, @empty}}, 0x0, 0x0, 0xe, 0x0, "f7e7cc8eefaf1c185361d9fd84f2f5f5da1fc0c85899c5fa73b0962b36bcf8d02916687c145ad91a568951b903f26927ecda656f85ee50ddc5788f8f42770db312d7023cc7b4fba91b852d9bf7ad6eb8"}, 0xd8) getsockopt$inet6_IPV6_FLOWLABEL_MGR(r2, 0x29, 0x20, &(0x7f0000000080)={@private1, 0x4, 0x2, 0x2, 0xc, 0x81, 0x3}, &(0x7f00000000c0)=0x20) connect$inet6(r1, &(0x7f0000000200)={0xa, 0x0, 0x0, @initdev={0xfe, 0x88, '\x00', 0x0, 0x0}, 0x3}, 0x1c) r4 = syz_init_net_socket$ax25(0x3, 0x2, 0x0) bind$ax25(r4, &(0x7f0000000000)={{0x3, @remote={0xcc, 0xcc, 0xcc, 0xcc, 0xcc, 0xcc, 0x1}, 0x1}, [@null, @netrom={0xbb, 0xbb, 0xbb, 0xbb, 0xbb, 0x0, 0x0}, @bcast, @rose={0xbb, 0xbb, 0xbb, 0x1, 0x0}, @bcast, @netrom={0xbb, 0xbb, 0xbb, 0xbb, 0xbb, 0x0, 0x0}, @null, @bcast]}, 0x48) close(r4) connect$inet6(r1, &(0x7f0000000040)={0xa, 0x4e20, 0x0, @dev={0xfe, 0x80, '\x00', 0xd}}, 0x1c) sendmmsg$inet6(r1, &(0x7f0000005a00)=[{{0x0, 0x0, 0x0}}, {{0x0, 0x0, 0x0, 0x0, &(0x7f0000005bc0)=ANY=[@ANYBLOB="5000000000000000290000003600000000060000000000000401000730000000000a0000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000020000000000000002900000037000000000000000000000004010000000000001400000000000000290000000000000160000000000000002900000037"], 0x100}}], 0x2, 0x0) 04:12:21 executing program 1: r0 = socket$netlink(0x10, 0x3, 0x0) r1 = socket$netlink(0x10, 0x3, 0x0) r2 = socket(0x10, 0x803, 0x0) sendmsg$NL80211_CMD_CRIT_PROTOCOL_START(r2, &(0x7f0000000580)={0x0, 0x0, &(0x7f0000000540)={0x0, 0x1c}}, 0x0) getsockname$packet(r2, &(0x7f0000000600)={0x11, 0x0, 0x0, 0x1, 0x0, 0x6, @broadcast}, &(0x7f0000000100)=0x14) sendmsg$nl_route(r1, &(0x7f0000000040)={0x0, 0x0, &(0x7f0000000000)={&(0x7f00000000c0)=ANY=[@ANYBLOB="3c00000010008506eb9afc4cd8d06e754a0081c5", @ANYRES32=r3, @ANYBLOB="2377f292252155b21c0012000c000100626f6e64000000000c0002000800010001"], 0x3c}}, 0x0) sendmsg$nl_route(r0, &(0x7f0000000240)={0x0, 0x0, &(0x7f0000000680)={&(0x7f0000000400)=@newlink={0x3c, 0x10, 0xffffff1f, 0x0, 0xa672500, {}, [@IFLA_LINKINFO={0x14, 0x12, 0x0, 0x1, @bridge={{0xb}, {0x4}}}, @IFLA_MASTER={0x8, 0xa, r3}]}, 0x3c}}, 0x0) 04:12:21 executing program 4: socket$inet6(0xa, 0x2, 0x0) (async) r0 = socket$inet6(0xa, 0x2, 0x0) r1 = socket$inet6(0xa, 0x80002, 0x88) bind$inet6(r1, &(0x7f0000000580)={0xa, 0x10010000004e20}, 0x1c) setsockopt$inet6_udp_int(r1, 0x11, 0xb, &(0x7f0000000080), 0x4) (async) setsockopt$inet6_udp_int(r1, 0x11, 0xb, &(0x7f0000000080), 0x4) bind$inet6(r1, &(0x7f0000000080)={0xa, 0x4e23, 0x8825, @private1={0xfc, 0x1, '\x00', 0x1}, 0x10000}, 0x1c) ioctl$sock_SIOCETHTOOL(r0, 0x89f0, &(0x7f0000000000)={'bridge0\x00', &(0x7f0000000040)=@ethtool_rxfh_indir={0x38, 0x9, [0x10000, 0x7, 0xfff, 0x20, 0x1, 0x8, 0x200, 0x0, 0x9]}}) [ 2813.865195][ T2670] workqueue: Failed to create a rescuer kthread for wq "bond1287": -EINTR [ 2814.138610][ T2683] netlink: 'syz-executor.2': attribute type 1 has an invalid length. [ 2814.317170][ T2683] bond1269: entered promiscuous mode [ 2814.338000][ T2683] 8021q: adding VLAN 0 to HW filter on device bond1269 [ 2814.528107][ T2688] bond1269: (slave bridge1190): making interface the new active one [ 2814.536660][ T2688] bridge1190: entered promiscuous mode [ 2814.550197][ T2688] bond1269: (slave bridge1190): Enslaving as an active interface with an up link 04:12:22 executing program 2: r0 = socket$netlink(0x10, 0x3, 0x0) r1 = socket$netlink(0x10, 0x3, 0x0) r2 = socket(0x10, 0x803, 0x0) sendmsg$NL80211_CMD_CRIT_PROTOCOL_START(r2, &(0x7f0000000580)={0x0, 0x0, &(0x7f0000000540)={0x0, 0x1c}}, 0x0) getsockname$packet(r2, &(0x7f0000000600)={0x11, 0x0, 0x0, 0x1, 0x0, 0x6, @broadcast}, &(0x7f0000000100)=0x14) sendmsg$nl_route(r1, &(0x7f0000000040)={0x0, 0x0, &(0x7f0000000000)={&(0x7f00000000c0)=ANY=[@ANYBLOB="3c00000010008506eb9afc4cd8d06e754a0081c5", @ANYRES32=r3, @ANYBLOB="2377f292252155b21c0012000c000100626f6e64000000000c0002000800010001"], 0x3c}}, 0x0) sendmsg$nl_route(r0, &(0x7f0000000240)={0x0, 0x0, &(0x7f0000000680)={&(0x7f0000000400)=@newlink={0x3c, 0x10, 0xffffff1f, 0x0, 0x253c28, {}, [@IFLA_LINKINFO={0x14, 0x12, 0x0, 0x1, @bridge={{0xb}, {0x4}}}, @IFLA_MASTER={0x8, 0xa, r3}]}, 0x3c}}, 0x0) [ 2814.574746][ T2700] netlink: 'syz-executor.5': attribute type 1 has an invalid length. 04:12:22 executing program 3: r0 = socket$netlink(0x10, 0x3, 0x0) sendmsg$netlink(r0, &(0x7f00000047c0)={0x0, 0x0, &(0x7f0000000000)=[{&(0x7f0000000100)=ANY=[@ANYBLOB="18000000560001060000006d00db730007"], 0x18}], 0x1}, 0x0) [ 2814.694475][ T2700] bond1279: entered promiscuous mode [ 2814.700492][ T2700] 8021q: adding VLAN 0 to HW filter on device bond1279 [ 2814.863383][ T2704] bond1279: (slave bridge1242): making interface the new active one [ 2814.871463][ T2704] bridge1242: entered promiscuous mode [ 2814.924961][ T2704] bond1279: (slave bridge1242): Enslaving as an active interface with an up link 04:12:22 executing program 5: r0 = socket$netlink(0x10, 0x3, 0x0) r1 = socket$netlink(0x10, 0x3, 0x0) r2 = socket(0x10, 0x803, 0x0) sendmsg$NL80211_CMD_CRIT_PROTOCOL_START(r2, &(0x7f0000000580)={0x0, 0x0, &(0x7f0000000540)={0x0, 0x1c}}, 0x0) getsockname$packet(r2, &(0x7f0000000600)={0x11, 0x0, 0x0, 0x1, 0x0, 0x6, @broadcast}, &(0x7f0000000100)=0x14) sendmsg$nl_route(r1, &(0x7f0000000040)={0x0, 0x0, &(0x7f0000000000)={&(0x7f00000000c0)=ANY=[@ANYBLOB="3c00000010008506eb9afc4cd8d06e754a0081c5", @ANYRES32=r3, @ANYBLOB="2377f292252155b21c0012000c000100626f6e64000000000c0002000800010001"], 0x3c}}, 0x0) sendmsg$nl_route(r0, &(0x7f0000000240)={0x0, 0x0, &(0x7f0000000680)={&(0x7f0000000400)=@newlink={0x3c, 0x10, 0xffffff1f, 0x0, 0x2535f9, {}, [@IFLA_LINKINFO={0x14, 0x12, 0x0, 0x1, @bridge={{0xb}, {0x4}}}, @IFLA_MASTER={0x8, 0xa, r3}]}, 0x3c}}, 0x0) [ 2814.965578][ T2712] netlink: 'syz-executor.1': attribute type 1 has an invalid length. 04:12:22 executing program 4: r0 = socket$inet6(0xa, 0x2, 0x0) r1 = socket$inet6(0xa, 0x80002, 0x88) bind$inet6(r1, &(0x7f0000000580)={0xa, 0x10010000004e20}, 0x1c) setsockopt$inet6_udp_int(r1, 0x11, 0xb, &(0x7f0000000080), 0x4) bind$inet6(r1, &(0x7f0000000080)={0xa, 0x4e23, 0x8825, @private1={0xfc, 0x1, '\x00', 0x1}, 0x10000}, 0x1c) ioctl$sock_SIOCETHTOOL(r0, 0x89f0, &(0x7f0000000000)={'bridge0\x00', &(0x7f0000000040)=@ethtool_rxfh_indir={0x38, 0x9, [0x10000, 0x7, 0xfff, 0x20, 0x1, 0x8, 0x200, 0x0, 0x9]}}) socket$inet6(0xa, 0x2, 0x0) (async) socket$inet6(0xa, 0x80002, 0x88) (async) bind$inet6(r1, &(0x7f0000000580)={0xa, 0x10010000004e20}, 0x1c) (async) setsockopt$inet6_udp_int(r1, 0x11, 0xb, &(0x7f0000000080), 0x4) (async) bind$inet6(r1, &(0x7f0000000080)={0xa, 0x4e23, 0x8825, @private1={0xfc, 0x1, '\x00', 0x1}, 0x10000}, 0x1c) (async) ioctl$sock_SIOCETHTOOL(r0, 0x89f0, &(0x7f0000000000)={'bridge0\x00', &(0x7f0000000040)=@ethtool_rxfh_indir={0x38, 0x9, [0x10000, 0x7, 0xfff, 0x20, 0x1, 0x8, 0x200, 0x0, 0x9]}}) (async) [ 2815.058424][ T2712] bond1287: entered promiscuous mode [ 2815.079001][ T2712] 8021q: adding VLAN 0 to HW filter on device bond1287 04:12:22 executing program 0: r0 = socket$inet6(0xa, 0x6, 0x0) bind$inet6(r0, &(0x7f0000000000)={0xa, 0x4e20}, 0x1c) (async) listen(r0, 0x6) (async) r1 = socket$inet6(0xa, 0x6, 0x0) (async) r2 = socket$inet6(0xa, 0x80002, 0x88) bind$inet6(r2, &(0x7f0000000580)={0xa, 0x10010000004e20}, 0x1c) (async) setsockopt$inet6_udp_int(r2, 0x11, 0xb, &(0x7f0000000080), 0x4) (async) r3 = accept(r0, 0x0, &(0x7f0000000100)) setsockopt$inet6_tcp_TCP_MD5SIG(r3, 0x6, 0xe, &(0x7f0000000240)={@in={{0x2, 0x4e24, @empty}}, 0x0, 0x0, 0xe, 0x0, "f7e7cc8eefaf1c185361d9fd84f2f5f5da1fc0c85899c5fa73b0962b36bcf8d02916687c145ad91a568951b903f26927ecda656f85ee50ddc5788f8f42770db312d7023cc7b4fba91b852d9bf7ad6eb8"}, 0xd8) (async, rerun: 32) getsockopt$inet6_IPV6_FLOWLABEL_MGR(r2, 0x29, 0x20, &(0x7f0000000080)={@private1, 0x4, 0x2, 0x2, 0xc, 0x81, 0x3}, &(0x7f00000000c0)=0x20) (rerun: 32) connect$inet6(r1, &(0x7f0000000200)={0xa, 0x0, 0x0, @initdev={0xfe, 0x88, '\x00', 0x0, 0x0}, 0x3}, 0x1c) (async) r4 = syz_init_net_socket$ax25(0x3, 0x2, 0x0) bind$ax25(r4, &(0x7f0000000000)={{0x3, @remote={0xcc, 0xcc, 0xcc, 0xcc, 0xcc, 0xcc, 0x1}, 0x1}, [@null, @netrom={0xbb, 0xbb, 0xbb, 0xbb, 0xbb, 0x0, 0x0}, @bcast, @rose={0xbb, 0xbb, 0xbb, 0x1, 0x0}, @bcast, @netrom={0xbb, 0xbb, 0xbb, 0xbb, 0xbb, 0x0, 0x0}, @null, @bcast]}, 0x48) (async) close(r4) (async, rerun: 64) connect$inet6(r1, &(0x7f0000000040)={0xa, 0x4e20, 0x0, @dev={0xfe, 0x80, '\x00', 0xd}}, 0x1c) (rerun: 64) sendmmsg$inet6(r1, &(0x7f0000005a00)=[{{0x0, 0x0, 0x0}}, {{0x0, 0x0, 0x0, 0x0, &(0x7f0000005bc0)=ANY=[@ANYBLOB="5000000000000000290000003600000000060000000000000401000730000000000a0000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000020000000000000002900000037000000000000000000000004010000000000001400000000000000290000000000000160000000000000002900000037"], 0x100}}], 0x2, 0x0) 04:12:22 executing program 0: r0 = socket$inet6(0xa, 0x6, 0x0) bind$inet6(r0, &(0x7f0000000000)={0xa, 0x4e20}, 0x1c) (async) listen(r0, 0x6) r1 = socket$inet6(0xa, 0x6, 0x0) (async) r2 = socket$inet6(0xa, 0x80002, 0x88) bind$inet6(r2, &(0x7f0000000580)={0xa, 0x10010000004e20}, 0x1c) setsockopt$inet6_udp_int(r2, 0x11, 0xb, &(0x7f0000000080), 0x4) r3 = accept(r0, 0x0, &(0x7f0000000100)) setsockopt$inet6_tcp_TCP_MD5SIG(r3, 0x6, 0xe, &(0x7f0000000240)={@in={{0x2, 0x4e24, @empty}}, 0x0, 0x0, 0xe, 0x0, "f7e7cc8eefaf1c185361d9fd84f2f5f5da1fc0c85899c5fa73b0962b36bcf8d02916687c145ad91a568951b903f26927ecda656f85ee50ddc5788f8f42770db312d7023cc7b4fba91b852d9bf7ad6eb8"}, 0xd8) getsockopt$inet6_IPV6_FLOWLABEL_MGR(r2, 0x29, 0x20, &(0x7f0000000080)={@private1, 0x4, 0x2, 0x2, 0xc, 0x81, 0x3}, &(0x7f00000000c0)=0x20) (async) connect$inet6(r1, &(0x7f0000000200)={0xa, 0x0, 0x0, @initdev={0xfe, 0x88, '\x00', 0x0, 0x0}, 0x3}, 0x1c) (async) r4 = syz_init_net_socket$ax25(0x3, 0x2, 0x0) bind$ax25(r4, &(0x7f0000000000)={{0x3, @remote={0xcc, 0xcc, 0xcc, 0xcc, 0xcc, 0xcc, 0x1}, 0x1}, [@null, @netrom={0xbb, 0xbb, 0xbb, 0xbb, 0xbb, 0x0, 0x0}, @bcast, @rose={0xbb, 0xbb, 0xbb, 0x1, 0x0}, @bcast, @netrom={0xbb, 0xbb, 0xbb, 0xbb, 0xbb, 0x0, 0x0}, @null, @bcast]}, 0x48) (async) close(r4) (async) connect$inet6(r1, &(0x7f0000000040)={0xa, 0x4e20, 0x0, @dev={0xfe, 0x80, '\x00', 0xd}}, 0x1c) sendmmsg$inet6(r1, &(0x7f0000005a00)=[{{0x0, 0x0, 0x0}}, {{0x0, 0x0, 0x0, 0x0, &(0x7f0000005bc0)=ANY=[@ANYBLOB="5000000000000000290000003600000000060000000000000401000730000000000a0000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000020000000000000002900000037000000000000000000000004010000000000001400000000000000290000000000000160000000000000002900000037"], 0x100}}], 0x2, 0x0) [ 2815.264790][ T2716] bond1287: (slave bridge1233): making interface the new active one [ 2815.287966][ T2716] bridge1233: entered promiscuous mode 04:12:22 executing program 1: r0 = socket$netlink(0x10, 0x3, 0x0) r1 = socket$netlink(0x10, 0x3, 0x0) r2 = socket(0x10, 0x803, 0x0) sendmsg$NL80211_CMD_CRIT_PROTOCOL_START(r2, &(0x7f0000000580)={0x0, 0x0, &(0x7f0000000540)={0x0, 0x1c}}, 0x0) getsockname$packet(r2, &(0x7f0000000600)={0x11, 0x0, 0x0, 0x1, 0x0, 0x6, @broadcast}, &(0x7f0000000100)=0x14) sendmsg$nl_route(r1, &(0x7f0000000040)={0x0, 0x0, &(0x7f0000000000)={&(0x7f00000000c0)=ANY=[@ANYBLOB="3c00000010008506eb9afc4cd8d06e754a0081c5", @ANYRES32=r3, @ANYBLOB="2377f292252155b21c0012000c000100626f6e64000000000c0002000800010001"], 0x3c}}, 0x0) sendmsg$nl_route(r0, &(0x7f0000000240)={0x0, 0x0, &(0x7f0000000680)={&(0x7f0000000400)=@newlink={0x3c, 0x10, 0xffffff1f, 0x0, 0xb672500, {}, [@IFLA_LINKINFO={0x14, 0x12, 0x0, 0x1, @bridge={{0xb}, {0x4}}}, @IFLA_MASTER={0x8, 0xa, r3}]}, 0x3c}}, 0x0) [ 2815.357718][ T2716] bond1287: (slave bridge1233): Enslaving as an active interface with an up link 04:12:22 executing program 3: r0 = socket$netlink(0x10, 0x3, 0x0) sendmsg$netlink(r0, &(0x7f00000047c0)={0x0, 0x0, &(0x7f0000000000)=[{&(0x7f0000000100)=ANY=[@ANYBLOB="18000000560001060000006e00db730007"], 0x18}], 0x1}, 0x0) [ 2815.425140][ T2727] netlink: 'syz-executor.2': attribute type 1 has an invalid length. 04:12:22 executing program 0: r0 = socket$inet6(0xa, 0x6, 0x0) bind$inet6(r0, &(0x7f0000000000)={0xa, 0x4e20}, 0x1c) listen(r0, 0x6) r1 = socket$inet6(0xa, 0x6, 0x0) connect$inet6(r1, &(0x7f0000000200)={0xa, 0x0, 0x0, @initdev={0xfe, 0x88, '\x00', 0x0, 0x0}, 0x3}, 0x1c) connect$inet6(r1, &(0x7f0000000040)={0xa, 0x4e20, 0xe3bd, @dev={0xfe, 0x80, '\x00', 0x3f}}, 0x1c) sendmmsg$inet6(r1, &(0x7f0000005a00)=[{{0x0, 0x0, 0x0}}, {{0x0, 0x0, 0x0, 0x0, &(0x7f0000005bc0)=ANY=[@ANYBLOB="5000000000000000290000003600000000060000000000000401000730000000000a0000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000020000000000000002900000037000000000000000000000004010000000000001400000000000000290000000000000160000000000000002900000037"], 0x100}}], 0x2, 0x0) r2 = socket$inet6_udp(0xa, 0x2, 0x0) connect$inet6(r2, &(0x7f0000000080)={0xa, 0x4e22, 0x0, @private0={0xfc, 0x0, '\x00', 0x1}, 0x7f}, 0x1c) ioctl$F2FS_IOC_MOVE_RANGE(r1, 0xc020f509, &(0x7f00000000c0)={r2, 0x7, 0x800, 0x8000000000000001}) setsockopt$inet6_IPV6_IPSEC_POLICY(r3, 0x29, 0x22, &(0x7f0000000100)={{{@in=@loopback, @in6=@private0, 0x4e21, 0x1, 0x4e23, 0x0, 0x2}, {0x7, 0x8, 0x7, 0x5, 0x1, 0x7, 0x6, 0x49cf}, {0xffffffffffff3271, 0x0, 0x4f6f601c, 0x3}, 0x1, 0x6e6bb2, 0x1, 0x1, 0x2}, {{@in6=@loopback, 0x4d3, 0x32}, 0xa, @in=@multicast2, 0x3504, 0x0, 0x2, 0x5f, 0xfffffffb, 0x0, 0x8}}, 0xe8) [ 2815.539012][ T2727] bond1270: entered promiscuous mode [ 2815.544984][ T2727] 8021q: adding VLAN 0 to HW filter on device bond1270 [ 2815.647733][ T2730] bond1270: (slave bridge1191): making interface the new active one [ 2815.687330][ T2730] bridge1191: entered promiscuous mode [ 2815.701185][ T2730] bond1270: (slave bridge1191): Enslaving as an active interface with an up link [ 2815.741204][ T2734] netlink: 'syz-executor.5': attribute type 1 has an invalid length. 04:12:23 executing program 2: r0 = socket$netlink(0x10, 0x3, 0x0) r1 = socket$netlink(0x10, 0x3, 0x0) r2 = socket(0x10, 0x803, 0x0) sendmsg$NL80211_CMD_CRIT_PROTOCOL_START(r2, &(0x7f0000000580)={0x0, 0x0, &(0x7f0000000540)={0x0, 0x1c}}, 0x0) getsockname$packet(r2, &(0x7f0000000600)={0x11, 0x0, 0x0, 0x1, 0x0, 0x6, @broadcast}, &(0x7f0000000100)=0x14) sendmsg$nl_route(r1, &(0x7f0000000040)={0x0, 0x0, &(0x7f0000000000)={&(0x7f00000000c0)=ANY=[@ANYBLOB="3c00000010008506eb9afc4cd8d06e754a0081c5", @ANYRES32=r3, @ANYBLOB="2377f292252155b21c0012000c000100626f6e64000000000c0002000800010001"], 0x3c}}, 0x0) sendmsg$nl_route(r0, &(0x7f0000000240)={0x0, 0x0, &(0x7f0000000680)={&(0x7f0000000400)=@newlink={0x3c, 0x10, 0xffffff1f, 0x0, 0x253c29, {}, [@IFLA_LINKINFO={0x14, 0x12, 0x0, 0x1, @bridge={{0xb}, {0x4}}}, @IFLA_MASTER={0x8, 0xa, r3}]}, 0x3c}}, 0x0) [ 2815.867742][ T2734] bond1280: entered promiscuous mode [ 2815.873725][ T2734] 8021q: adding VLAN 0 to HW filter on device bond1280 [ 2815.917370][ T2737] bond1280: (slave bridge1243): making interface the new active one [ 2815.925600][ T2737] bridge1243: entered promiscuous mode [ 2815.936109][ T2737] bond1280: (slave bridge1243): Enslaving as an active interface with an up link 04:12:23 executing program 5: r0 = socket$netlink(0x10, 0x3, 0x0) r1 = socket$netlink(0x10, 0x3, 0x0) r2 = socket(0x10, 0x803, 0x0) sendmsg$NL80211_CMD_CRIT_PROTOCOL_START(r2, &(0x7f0000000580)={0x0, 0x0, &(0x7f0000000540)={0x0, 0x1c}}, 0x0) getsockname$packet(r2, &(0x7f0000000600)={0x11, 0x0, 0x0, 0x1, 0x0, 0x6, @broadcast}, &(0x7f0000000100)=0x14) sendmsg$nl_route(r1, &(0x7f0000000040)={0x0, 0x0, &(0x7f0000000000)={&(0x7f00000000c0)=ANY=[@ANYBLOB="3c00000010008506eb9afc4cd8d06e754a0081c5", @ANYRES32=r3, @ANYBLOB="2377f292252155b21c0012000c000100626f6e64000000000c0002000800010001"], 0x3c}}, 0x0) sendmsg$nl_route(r0, &(0x7f0000000240)={0x0, 0x0, &(0x7f0000000680)={&(0x7f0000000400)=@newlink={0x3c, 0x10, 0xffffff1f, 0x0, 0x2535fa, {}, [@IFLA_LINKINFO={0x14, 0x12, 0x0, 0x1, @bridge={{0xb}, {0x4}}}, @IFLA_MASTER={0x8, 0xa, r3}]}, 0x3c}}, 0x0) 04:12:23 executing program 4: r0 = socket$inet6(0xa, 0x2, 0x0) ioctl$sock_SIOCETHTOOL(r0, 0x89f0, &(0x7f0000000040)={'bridge0\x00', &(0x7f0000000200)=@ethtool_ringparam={0x10, 0x0, 0x0, 0x0, 0x0, 0x0, 0x10}}) [ 2816.156353][ T2757] bond1288: entered promiscuous mode [ 2816.167703][ T2757] 8021q: adding VLAN 0 to HW filter on device bond1288 [ 2816.212692][ T2760] bond1288: (slave bridge1234): making interface the new active one [ 2816.220903][ T2760] bridge1234: entered promiscuous mode [ 2816.231753][ T2760] bond1288: (slave bridge1234): Enslaving as an active interface with an up link 04:12:23 executing program 1: r0 = socket$netlink(0x10, 0x3, 0x0) r1 = socket$netlink(0x10, 0x3, 0x0) r2 = socket(0x10, 0x803, 0x0) sendmsg$NL80211_CMD_CRIT_PROTOCOL_START(r2, &(0x7f0000000580)={0x0, 0x0, &(0x7f0000000540)={0x0, 0x1c}}, 0x0) getsockname$packet(r2, &(0x7f0000000600)={0x11, 0x0, 0x0, 0x1, 0x0, 0x6, @broadcast}, &(0x7f0000000100)=0x14) sendmsg$nl_route(r1, &(0x7f0000000040)={0x0, 0x0, &(0x7f0000000000)={&(0x7f00000000c0)=ANY=[@ANYBLOB="3c00000010008506eb9afc4cd8d06e754a0081c5", @ANYRES32=r3, @ANYBLOB="2377f292252155b21c0012000c000100626f6e64000000000c0002000800010001"], 0x3c}}, 0x0) sendmsg$nl_route(r0, &(0x7f0000000240)={0x0, 0x0, &(0x7f0000000680)={&(0x7f0000000400)=@newlink={0x3c, 0x10, 0xffffff1f, 0x0, 0xc000000, {}, [@IFLA_LINKINFO={0x14, 0x12, 0x0, 0x1, @bridge={{0xb}, {0x4}}}, @IFLA_MASTER={0x8, 0xa, r3}]}, 0x3c}}, 0x0) 04:12:23 executing program 3: r0 = socket$netlink(0x10, 0x3, 0x0) sendmsg$netlink(r0, &(0x7f00000047c0)={0x0, 0x0, &(0x7f0000000000)=[{&(0x7f0000000100)=ANY=[@ANYBLOB="18000000560001060000006f00db730007"], 0x18}], 0x1}, 0x0) 04:12:23 executing program 0: r0 = socket$inet6(0xa, 0x6, 0x0) bind$inet6(r0, &(0x7f0000000000)={0xa, 0x4e20}, 0x1c) listen(r0, 0x6) (async) listen(r0, 0x6) r1 = socket$inet6(0xa, 0x6, 0x0) connect$inet6(r1, &(0x7f0000000200)={0xa, 0x0, 0x0, @initdev={0xfe, 0x88, '\x00', 0x0, 0x0}, 0x3}, 0x1c) (async) connect$inet6(r1, &(0x7f0000000200)={0xa, 0x0, 0x0, @initdev={0xfe, 0x88, '\x00', 0x0, 0x0}, 0x3}, 0x1c) connect$inet6(r1, &(0x7f0000000040)={0xa, 0x4e20, 0xe3bd, @dev={0xfe, 0x80, '\x00', 0x3f}}, 0x1c) (async) connect$inet6(r1, &(0x7f0000000040)={0xa, 0x4e20, 0xe3bd, @dev={0xfe, 0x80, '\x00', 0x3f}}, 0x1c) sendmmsg$inet6(r1, &(0x7f0000005a00)=[{{0x0, 0x0, 0x0}}, {{0x0, 0x0, 0x0, 0x0, &(0x7f0000005bc0)=ANY=[@ANYBLOB="5000000000000000290000003600000000060000000000000401000730000000000a0000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000020000000000000002900000037000000000000000000000004010000000000001400000000000000290000000000000160000000000000002900000037"], 0x100}}], 0x2, 0x0) r2 = socket$inet6_udp(0xa, 0x2, 0x0) connect$inet6(r2, &(0x7f0000000080)={0xa, 0x4e22, 0x0, @private0={0xfc, 0x0, '\x00', 0x1}, 0x7f}, 0x1c) ioctl$F2FS_IOC_MOVE_RANGE(r1, 0xc020f509, &(0x7f00000000c0)={r2, 0x7, 0x800, 0x8000000000000001}) setsockopt$inet6_IPV6_IPSEC_POLICY(r3, 0x29, 0x22, &(0x7f0000000100)={{{@in=@loopback, @in6=@private0, 0x4e21, 0x1, 0x4e23, 0x0, 0x2}, {0x7, 0x8, 0x7, 0x5, 0x1, 0x7, 0x6, 0x49cf}, {0xffffffffffff3271, 0x0, 0x4f6f601c, 0x3}, 0x1, 0x6e6bb2, 0x1, 0x1, 0x2}, {{@in6=@loopback, 0x4d3, 0x32}, 0xa, @in=@multicast2, 0x3504, 0x0, 0x2, 0x5f, 0xfffffffb, 0x0, 0x8}}, 0xe8) [ 2816.438379][ T2769] bond1271: entered promiscuous mode [ 2816.459433][ T2769] 8021q: adding VLAN 0 to HW filter on device bond1271 04:12:24 executing program 2: r0 = socket$netlink(0x10, 0x3, 0x0) r1 = socket$netlink(0x10, 0x3, 0x0) r2 = socket(0x10, 0x803, 0x0) sendmsg$NL80211_CMD_CRIT_PROTOCOL_START(r2, &(0x7f0000000580)={0x0, 0x0, &(0x7f0000000540)={0x0, 0x1c}}, 0x0) getsockname$packet(r2, &(0x7f0000000600)={0x11, 0x0, 0x0, 0x1, 0x0, 0x6, @broadcast}, &(0x7f0000000100)=0x14) sendmsg$nl_route(r1, &(0x7f0000000040)={0x0, 0x0, &(0x7f0000000000)={&(0x7f00000000c0)=ANY=[@ANYBLOB="3c00000010008506eb9afc4cd8d06e754a0081c5", @ANYRES32=r3, @ANYBLOB="2377f292252155b21c0012000c000100626f6e64000000000c0002000800010001"], 0x3c}}, 0x0) sendmsg$nl_route(r0, &(0x7f0000000240)={0x0, 0x0, &(0x7f0000000680)={&(0x7f0000000400)=@newlink={0x3c, 0x10, 0xffffff1f, 0x0, 0x253c2a, {}, [@IFLA_LINKINFO={0x14, 0x12, 0x0, 0x1, @bridge={{0xb}, {0x4}}}, @IFLA_MASTER={0x8, 0xa, r3}]}, 0x3c}}, 0x0) [ 2816.650726][ T2770] bond1271: (slave bridge1192): making interface the new active one [ 2816.659304][ T2770] bridge1192: entered promiscuous mode [ 2816.673330][ T2770] bond1271: (slave bridge1192): Enslaving as an active interface with an up link 04:12:24 executing program 0: r0 = socket$inet6(0xa, 0x6, 0x0) bind$inet6(r0, &(0x7f0000000000)={0xa, 0x4e20}, 0x1c) (async) listen(r0, 0x6) (async) r1 = socket$inet6(0xa, 0x6, 0x0) connect$inet6(r1, &(0x7f0000000200)={0xa, 0x0, 0x0, @initdev={0xfe, 0x88, '\x00', 0x0, 0x0}, 0x3}, 0x1c) connect$inet6(r1, &(0x7f0000000040)={0xa, 0x4e20, 0xe3bd, @dev={0xfe, 0x80, '\x00', 0x3f}}, 0x1c) (async) sendmmsg$inet6(r1, &(0x7f0000005a00)=[{{0x0, 0x0, 0x0}}, {{0x0, 0x0, 0x0, 0x0, &(0x7f0000005bc0)=ANY=[@ANYBLOB="5000000000000000290000003600000000060000000000000401000730000000000a0000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000020000000000000002900000037000000000000000000000004010000000000001400000000000000290000000000000160000000000000002900000037"], 0x100}}], 0x2, 0x0) r2 = socket$inet6_udp(0xa, 0x2, 0x0) connect$inet6(r2, &(0x7f0000000080)={0xa, 0x4e22, 0x0, @private0={0xfc, 0x0, '\x00', 0x1}, 0x7f}, 0x1c) (async) ioctl$F2FS_IOC_MOVE_RANGE(r1, 0xc020f509, &(0x7f00000000c0)={r2, 0x7, 0x800, 0x8000000000000001}) setsockopt$inet6_IPV6_IPSEC_POLICY(r3, 0x29, 0x22, &(0x7f0000000100)={{{@in=@loopback, @in6=@private0, 0x4e21, 0x1, 0x4e23, 0x0, 0x2}, {0x7, 0x8, 0x7, 0x5, 0x1, 0x7, 0x6, 0x49cf}, {0xffffffffffff3271, 0x0, 0x4f6f601c, 0x3}, 0x1, 0x6e6bb2, 0x1, 0x1, 0x2}, {{@in6=@loopback, 0x4d3, 0x32}, 0xa, @in=@multicast2, 0x3504, 0x0, 0x2, 0x5f, 0xfffffffb, 0x0, 0x8}}, 0xe8) [ 2816.839350][ T2773] bond1281: entered promiscuous mode [ 2816.854933][ T2773] 8021q: adding VLAN 0 to HW filter on device bond1281 04:12:24 executing program 4: r0 = socket$inet6(0xa, 0x2, 0x0) ioctl$sock_SIOCETHTOOL(r0, 0x89f0, &(0x7f0000000040)={'bridge0\x00', &(0x7f0000000200)=@ethtool_ringparam={0x10, 0x0, 0x0, 0x0, 0x0, 0x0, 0x10}}) 04:12:24 executing program 0: r0 = socket$inet6(0xa, 0x6, 0x0) bind$inet6(r0, &(0x7f0000000000)={0xa, 0x4e20}, 0x1c) listen(r0, 0x6) sendto$inet6(r0, &(0x7f0000000080)="ba34844a875f6a24ad706fcf443d535f8a21821ebd95bead9dbdf8af7b503f9a58eed1f7da8db7d8a88f61486f865aa11e0c41d37ba4b96278744da8a4708309166b0e607a18147f6ae59efed7be70a1c2127bd1d20ed49f7fcd09f0b0276408abf5", 0x62, 0x20004000, &(0x7f0000000100)={0xa, 0x4e24, 0x5, @private2={0xfc, 0x2, '\x00', 0x1}, 0x8}, 0x1c) r1 = socket$inet6(0xa, 0x6, 0x0) connect$inet6(r1, &(0x7f0000000200)={0xa, 0x0, 0x0, @initdev={0xfe, 0x88, '\x00', 0x0, 0x0}, 0x3}, 0x1c) connect$inet6(r1, &(0x7f0000000040)={0xa, 0x4e20, 0x0, @dev={0xfe, 0x80, '\x00', 0xf}}, 0x1c) sendmmsg$inet6(r1, &(0x7f0000005a00)=[{{0x0, 0x0, 0x0}}, {{0x0, 0x0, 0x0, 0x0, &(0x7f0000005bc0)=ANY=[@ANYBLOB="5000000000000000290000003600000000060000000000000401000730000000000a0000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000020000000000000002900000037000000000000000000000004010000000000001400000000000000290000000000000160000000000000002900000037"], 0x100}}], 0x2, 0x0) [ 2817.049439][ T2777] bond1281: (slave bridge1244): making interface the new active one [ 2817.072115][ T2777] bridge1244: entered promiscuous mode [ 2817.090767][ T2777] bond1281: (slave bridge1244): Enslaving as an active interface with an up link 04:12:24 executing program 0: r0 = socket$inet6(0xa, 0x6, 0x0) bind$inet6(r0, &(0x7f0000000000)={0xa, 0x4e20}, 0x1c) listen(r0, 0x6) (async, rerun: 32) sendto$inet6(r0, &(0x7f0000000080)="ba34844a875f6a24ad706fcf443d535f8a21821ebd95bead9dbdf8af7b503f9a58eed1f7da8db7d8a88f61486f865aa11e0c41d37ba4b96278744da8a4708309166b0e607a18147f6ae59efed7be70a1c2127bd1d20ed49f7fcd09f0b0276408abf5", 0x62, 0x20004000, &(0x7f0000000100)={0xa, 0x4e24, 0x5, @private2={0xfc, 0x2, '\x00', 0x1}, 0x8}, 0x1c) (rerun: 32) r1 = socket$inet6(0xa, 0x6, 0x0) connect$inet6(r1, &(0x7f0000000200)={0xa, 0x0, 0x0, @initdev={0xfe, 0x88, '\x00', 0x0, 0x0}, 0x3}, 0x1c) connect$inet6(r1, &(0x7f0000000040)={0xa, 0x4e20, 0x0, @dev={0xfe, 0x80, '\x00', 0xf}}, 0x1c) (async, rerun: 32) sendmmsg$inet6(r1, &(0x7f0000005a00)=[{{0x0, 0x0, 0x0}}, {{0x0, 0x0, 0x0, 0x0, &(0x7f0000005bc0)=ANY=[@ANYBLOB="5000000000000000290000003600000000060000000000000401000730000000000a0000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000020000000000000002900000037000000000000000000000004010000000000001400000000000000290000000000000160000000000000002900000037"], 0x100}}], 0x2, 0x0) (rerun: 32) 04:12:24 executing program 5: r0 = socket$netlink(0x10, 0x3, 0x0) r1 = socket$netlink(0x10, 0x3, 0x0) r2 = socket(0x10, 0x803, 0x0) sendmsg$NL80211_CMD_CRIT_PROTOCOL_START(r2, &(0x7f0000000580)={0x0, 0x0, &(0x7f0000000540)={0x0, 0x1c}}, 0x0) getsockname$packet(r2, &(0x7f0000000600)={0x11, 0x0, 0x0, 0x1, 0x0, 0x6, @broadcast}, &(0x7f0000000100)=0x14) sendmsg$nl_route(r1, &(0x7f0000000040)={0x0, 0x0, &(0x7f0000000000)={&(0x7f00000000c0)=ANY=[@ANYBLOB="3c00000010008506eb9afc4cd8d06e754a0081c5", @ANYRES32=r3, @ANYBLOB="2377f292252155b21c0012000c000100626f6e64000000000c0002000800010001"], 0x3c}}, 0x0) sendmsg$nl_route(r0, &(0x7f0000000240)={0x0, 0x0, &(0x7f0000000680)={&(0x7f0000000400)=@newlink={0x3c, 0x10, 0xffffff1f, 0x0, 0x2535fb, {}, [@IFLA_LINKINFO={0x14, 0x12, 0x0, 0x1, @bridge={{0xb}, {0x4}}}, @IFLA_MASTER={0x8, 0xa, r3}]}, 0x3c}}, 0x0) [ 2817.196313][ T2781] validate_nla: 3 callbacks suppressed [ 2817.196333][ T2781] netlink: 'syz-executor.1': attribute type 1 has an invalid length. 04:12:24 executing program 0: r0 = socket$inet6(0xa, 0x6, 0x0) bind$inet6(r0, &(0x7f0000000000)={0xa, 0x4e20}, 0x1c) listen(r0, 0x6) (async) sendto$inet6(r0, &(0x7f0000000080)="ba34844a875f6a24ad706fcf443d535f8a21821ebd95bead9dbdf8af7b503f9a58eed1f7da8db7d8a88f61486f865aa11e0c41d37ba4b96278744da8a4708309166b0e607a18147f6ae59efed7be70a1c2127bd1d20ed49f7fcd09f0b0276408abf5", 0x62, 0x20004000, &(0x7f0000000100)={0xa, 0x4e24, 0x5, @private2={0xfc, 0x2, '\x00', 0x1}, 0x8}, 0x1c) (async) r1 = socket$inet6(0xa, 0x6, 0x0) connect$inet6(r1, &(0x7f0000000200)={0xa, 0x0, 0x0, @initdev={0xfe, 0x88, '\x00', 0x0, 0x0}, 0x3}, 0x1c) (async) connect$inet6(r1, &(0x7f0000000040)={0xa, 0x4e20, 0x0, @dev={0xfe, 0x80, '\x00', 0xf}}, 0x1c) sendmmsg$inet6(r1, &(0x7f0000005a00)=[{{0x0, 0x0, 0x0}}, {{0x0, 0x0, 0x0, 0x0, &(0x7f0000005bc0)=ANY=[@ANYBLOB="5000000000000000290000003600000000060000000000000401000730000000000a0000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000020000000000000002900000037000000000000000000000004010000000000001400000000000000290000000000000160000000000000002900000037"], 0x100}}], 0x2, 0x0) [ 2817.343507][ T2781] bond1289: entered promiscuous mode [ 2817.349108][ T2781] 8021q: adding VLAN 0 to HW filter on device bond1289 04:12:24 executing program 3: r0 = socket$netlink(0x10, 0x3, 0x0) sendmsg$netlink(r0, &(0x7f00000047c0)={0x0, 0x0, &(0x7f0000000000)=[{&(0x7f0000000100)=ANY=[@ANYBLOB="18000000560001060000007000db730007"], 0x18}], 0x1}, 0x0) 04:12:24 executing program 0: r0 = socket$inet6(0xa, 0x6, 0x0) bind$inet6(r0, &(0x7f0000000000)={0xa, 0x4e20}, 0x1c) listen(r0, 0x6) r1 = socket$inet6(0xa, 0x6, 0x0) connect$inet6(r1, &(0x7f0000000200)={0xa, 0x0, 0x0, @initdev={0xfe, 0x88, '\x00', 0x0, 0x0}, 0x3}, 0x1c) connect$inet6(r1, &(0x7f0000000040)={0xa, 0x4e20, 0x0, @dev={0xfe, 0x80, '\x00', 0xf}}, 0x1c) sendmmsg$inet6(r1, &(0x7f0000005a00)=[{{0x0, 0x0, 0x0}}, {{0x0, 0x0, 0x0, 0x0, &(0x7f0000005bc0)=ANY=[@ANYBLOB="5000000000000000290000003600000000060000000000000401000730000000000a0000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000020000000000000002900000037000000000000000000000004010000000000001400000000000000290000000000000160000000000000002900000037"], 0x100}}], 0x2, 0x0) [ 2817.647074][ T2784] bond1289: (slave bridge1235): making interface the new active one [ 2817.670135][ T2784] bridge1235: entered promiscuous mode [ 2817.680865][ T2784] bond1289: (slave bridge1235): Enslaving as an active interface with an up link 04:12:25 executing program 1: r0 = socket$netlink(0x10, 0x3, 0x0) r1 = socket$netlink(0x10, 0x3, 0x0) r2 = socket(0x10, 0x803, 0x0) sendmsg$NL80211_CMD_CRIT_PROTOCOL_START(r2, &(0x7f0000000580)={0x0, 0x0, &(0x7f0000000540)={0x0, 0x1c}}, 0x0) getsockname$packet(r2, &(0x7f0000000600)={0x11, 0x0, 0x0, 0x1, 0x0, 0x6, @broadcast}, &(0x7f0000000100)=0x14) sendmsg$nl_route(r1, &(0x7f0000000040)={0x0, 0x0, &(0x7f0000000000)={&(0x7f00000000c0)=ANY=[@ANYBLOB="3c00000010008506eb9afc4cd8d06e754a0081c5", @ANYRES32=r3, @ANYBLOB="2377f292252155b21c0012000c000100626f6e64000000000c0002000800010001"], 0x3c}}, 0x0) sendmsg$nl_route(r0, &(0x7f0000000240)={0x0, 0x0, &(0x7f0000000680)={&(0x7f0000000400)=@newlink={0x3c, 0x10, 0xffffff1f, 0x0, 0xc672500, {}, [@IFLA_LINKINFO={0x14, 0x12, 0x0, 0x1, @bridge={{0xb}, {0x4}}}, @IFLA_MASTER={0x8, 0xa, r3}]}, 0x3c}}, 0x0) 04:12:25 executing program 0: r0 = socket$inet6(0xa, 0x6, 0x0) bind$inet6(r0, &(0x7f0000000000)={0xa, 0x4e20}, 0x1c) listen(r0, 0x6) r1 = socket$inet6(0xa, 0x6, 0x0) connect$inet6(r1, &(0x7f0000000200)={0xa, 0x0, 0x0, @initdev={0xfe, 0x88, '\x00', 0x0, 0x0}, 0x3}, 0x1c) connect$inet6(r1, &(0x7f0000000040)={0xa, 0x4e20, 0x0, @dev={0xfe, 0x80, '\x00', 0xf}}, 0x1c) sendmmsg$inet6(r1, &(0x7f0000005a00)=[{{0x0, 0x0, 0x0}}, {{0x0, 0x0, 0x0, 0x0, &(0x7f0000005bc0)=ANY=[@ANYBLOB="5000000000000000290000003600000000060000000000000401000730000000000a0000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000020000000000000002900000037000000000000000000000004010000000000001400000000000000290000000000000160000000000000002900000037"], 0x100}}], 0x2, 0x0) socket$inet6(0xa, 0x6, 0x0) (async) bind$inet6(r0, &(0x7f0000000000)={0xa, 0x4e20}, 0x1c) (async) listen(r0, 0x6) (async) socket$inet6(0xa, 0x6, 0x0) (async) connect$inet6(r1, &(0x7f0000000200)={0xa, 0x0, 0x0, @initdev={0xfe, 0x88, '\x00', 0x0, 0x0}, 0x3}, 0x1c) (async) connect$inet6(r1, &(0x7f0000000040)={0xa, 0x4e20, 0x0, @dev={0xfe, 0x80, '\x00', 0xf}}, 0x1c) (async) sendmmsg$inet6(r1, &(0x7f0000005a00)=[{{0x0, 0x0, 0x0}}, {{0x0, 0x0, 0x0, 0x0, &(0x7f0000005bc0)=ANY=[@ANYBLOB="5000000000000000290000003600000000060000000000000401000730000000000a0000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000020000000000000002900000037000000000000000000000004010000000000001400000000000000290000000000000160000000000000002900000037"], 0x100}}], 0x2, 0x0) (async) [ 2817.714151][ T2793] netlink: 'syz-executor.2': attribute type 1 has an invalid length. [ 2817.782422][ T2793] bond1272: entered promiscuous mode [ 2817.788934][ T2793] 8021q: adding VLAN 0 to HW filter on device bond1272 04:12:25 executing program 0: r0 = socket$inet6(0xa, 0x6, 0x0) bind$inet6(r0, &(0x7f0000000000)={0xa, 0x4e20}, 0x1c) listen(r0, 0x6) r1 = socket$inet6(0xa, 0x6, 0x0) connect$inet6(r1, &(0x7f0000000200)={0xa, 0x0, 0x0, @initdev={0xfe, 0x88, '\x00', 0x0, 0x0}, 0x3}, 0x1c) connect$inet6(r1, &(0x7f0000000040)={0xa, 0x4e20, 0x0, @dev={0xfe, 0x80, '\x00', 0xf}}, 0x1c) sendmmsg$inet6(r1, &(0x7f0000005a00)=[{{0x0, 0x0, 0x0}}, {{0x0, 0x0, 0x0, 0x0, &(0x7f0000005bc0)=ANY=[@ANYBLOB="5000000000000000290000003600000000060000000000000401000730000000000a0000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000020000000000000002900000037000000000000000000000004010000000000001400000000000000290000000000000160000000000000002900000037"], 0x100}}], 0x2, 0x0) socket$inet6(0xa, 0x6, 0x0) (async) bind$inet6(r0, &(0x7f0000000000)={0xa, 0x4e20}, 0x1c) (async) listen(r0, 0x6) (async) socket$inet6(0xa, 0x6, 0x0) (async) connect$inet6(r1, &(0x7f0000000200)={0xa, 0x0, 0x0, @initdev={0xfe, 0x88, '\x00', 0x0, 0x0}, 0x3}, 0x1c) (async) connect$inet6(r1, &(0x7f0000000040)={0xa, 0x4e20, 0x0, @dev={0xfe, 0x80, '\x00', 0xf}}, 0x1c) (async) sendmmsg$inet6(r1, &(0x7f0000005a00)=[{{0x0, 0x0, 0x0}}, {{0x0, 0x0, 0x0, 0x0, &(0x7f0000005bc0)=ANY=[@ANYBLOB="5000000000000000290000003600000000060000000000000401000730000000000a0000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000020000000000000002900000037000000000000000000000004010000000000001400000000000000290000000000000160000000000000002900000037"], 0x100}}], 0x2, 0x0) (async) [ 2818.004236][ T2798] bond1272: (slave bridge1193): making interface the new active one [ 2818.029278][ T2798] bridge1193: entered promiscuous mode [ 2818.059562][ T2798] bond1272: (slave bridge1193): Enslaving as an active interface with an up link [ 2818.099145][ T2814] netlink: 'syz-executor.5': attribute type 1 has an invalid length. 04:12:25 executing program 2: r0 = socket$netlink(0x10, 0x3, 0x0) r1 = socket$netlink(0x10, 0x3, 0x0) r2 = socket(0x10, 0x803, 0x0) sendmsg$NL80211_CMD_CRIT_PROTOCOL_START(r2, &(0x7f0000000580)={0x0, 0x0, &(0x7f0000000540)={0x0, 0x1c}}, 0x0) getsockname$packet(r2, &(0x7f0000000600)={0x11, 0x0, 0x0, 0x1, 0x0, 0x6, @broadcast}, &(0x7f0000000100)=0x14) sendmsg$nl_route(r1, &(0x7f0000000040)={0x0, 0x0, &(0x7f0000000000)={&(0x7f00000000c0)=ANY=[@ANYBLOB="3c00000010008506eb9afc4cd8d06e754a0081c5", @ANYRES32=r3, @ANYBLOB="2377f292252155b21c0012000c000100626f6e64000000000c0002000800010001"], 0x3c}}, 0x0) sendmsg$nl_route(r0, &(0x7f0000000240)={0x0, 0x0, &(0x7f0000000680)={&(0x7f0000000400)=@newlink={0x3c, 0x10, 0xffffff1f, 0x0, 0x253c2b, {}, [@IFLA_LINKINFO={0x14, 0x12, 0x0, 0x1, @bridge={{0xb}, {0x4}}}, @IFLA_MASTER={0x8, 0xa, r3}]}, 0x3c}}, 0x0) 04:12:25 executing program 0: r0 = socket$inet6(0xa, 0x6, 0x0) bind$inet6(r0, &(0x7f0000000000)={0xa, 0x4e20}, 0x1c) socket$inet6(0xa, 0x800, 0x8d3) connect$inet6(r0, &(0x7f0000000080)={0xa, 0x4e20, 0x7fff, @private2}, 0x1c) listen(r0, 0x6) r1 = socket$inet6(0xa, 0x800, 0x0) connect$inet6(r1, &(0x7f0000000200)={0xa, 0x0, 0x0, @initdev={0xfe, 0x88, '\x00', 0x0, 0x0}, 0x3}, 0x1c) connect$inet6(r1, &(0x7f0000000040)={0xa, 0x4e20, 0x0, @dev={0xfe, 0x80, '\x00', 0xf}}, 0x1c) sendmmsg$inet6(r1, &(0x7f0000005a00)=[{{0x0, 0x0, 0x0}}, {{0x0, 0x0, 0x0, 0x0, &(0x7f0000005bc0)=ANY=[@ANYBLOB="5000000000000000290000003600000000060000000000000401000730000000000a0000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000020000000000000002900000037000000000000000000000004010000000000001400000000000000290000000000000160000000000000002900000037"], 0x100}}], 0x2, 0x0) 04:12:25 executing program 4: r0 = socket$inet6(0xa, 0x2, 0x0) ioctl$sock_SIOCETHTOOL(r0, 0x89f0, &(0x7f0000000040)={'bridge0\x00', &(0x7f0000000200)=@ethtool_ringparam={0x10, 0x0, 0x0, 0x0, 0x0, 0x0, 0x10}}) [ 2818.210223][ T2814] bond1282: entered promiscuous mode [ 2818.216858][ T2814] 8021q: adding VLAN 0 to HW filter on device bond1282 [ 2818.239672][ T2846] netlink: 'syz-executor.2': attribute type 1 has an invalid length. [ 2818.386421][ T2846] bond1273: entered promiscuous mode [ 2818.392850][ T2846] 8021q: adding VLAN 0 to HW filter on device bond1273 04:12:25 executing program 0: r0 = socket$inet6(0xa, 0x6, 0x0) bind$inet6(r0, &(0x7f0000000000)={0xa, 0x4e20}, 0x1c) socket$inet6(0xa, 0x800, 0x8d3) (async) connect$inet6(r0, &(0x7f0000000080)={0xa, 0x4e20, 0x7fff, @private2}, 0x1c) (async) listen(r0, 0x6) r1 = socket$inet6(0xa, 0x800, 0x0) connect$inet6(r1, &(0x7f0000000200)={0xa, 0x0, 0x0, @initdev={0xfe, 0x88, '\x00', 0x0, 0x0}, 0x3}, 0x1c) (async) connect$inet6(r1, &(0x7f0000000040)={0xa, 0x4e20, 0x0, @dev={0xfe, 0x80, '\x00', 0xf}}, 0x1c) sendmmsg$inet6(r1, &(0x7f0000005a00)=[{{0x0, 0x0, 0x0}}, {{0x0, 0x0, 0x0, 0x0, &(0x7f0000005bc0)=ANY=[@ANYBLOB="5000000000000000290000003600000000060000000000000401000730000000000a0000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000020000000000000002900000037000000000000000000000004010000000000001400000000000000290000000000000160000000000000002900000037"], 0x100}}], 0x2, 0x0) [ 2818.533780][ T2818] bond1282: (slave bridge1245): making interface the new active one [ 2818.564880][ T2818] bridge1245: entered promiscuous mode [ 2818.596770][ T2818] bond1282: (slave bridge1245): Enslaving as an active interface with an up link [ 2818.634234][ T2831] netlink: 'syz-executor.1': attribute type 1 has an invalid length. 04:12:26 executing program 5: r0 = socket$netlink(0x10, 0x3, 0x0) r1 = socket$netlink(0x10, 0x3, 0x0) r2 = socket(0x10, 0x803, 0x0) sendmsg$NL80211_CMD_CRIT_PROTOCOL_START(r2, &(0x7f0000000580)={0x0, 0x0, &(0x7f0000000540)={0x0, 0x1c}}, 0x0) getsockname$packet(r2, &(0x7f0000000600)={0x11, 0x0, 0x0, 0x1, 0x0, 0x6, @broadcast}, &(0x7f0000000100)=0x14) sendmsg$nl_route(r1, &(0x7f0000000040)={0x0, 0x0, &(0x7f0000000000)={&(0x7f00000000c0)=ANY=[@ANYBLOB="3c00000010008506eb9afc4cd8d06e754a0081c5", @ANYRES32=r3, @ANYBLOB="2377f292252155b21c0012000c000100626f6e64000000000c0002000800010001"], 0x3c}}, 0x0) sendmsg$nl_route(r0, &(0x7f0000000240)={0x0, 0x0, &(0x7f0000000680)={&(0x7f0000000400)=@newlink={0x3c, 0x10, 0xffffff1f, 0x0, 0x2535fc, {}, [@IFLA_LINKINFO={0x14, 0x12, 0x0, 0x1, @bridge={{0xb}, {0x4}}}, @IFLA_MASTER={0x8, 0xa, r3}]}, 0x3c}}, 0x0) 04:12:26 executing program 3: r0 = socket$netlink(0x10, 0x3, 0x0) sendmsg$netlink(r0, &(0x7f00000047c0)={0x0, 0x0, &(0x7f0000000000)=[{&(0x7f0000000100)=ANY=[@ANYBLOB="18000000560001060000007100db730007"], 0x18}], 0x1}, 0x0) 04:12:26 executing program 0: r0 = socket$inet6(0xa, 0x6, 0x0) bind$inet6(r0, &(0x7f0000000000)={0xa, 0x4e20}, 0x1c) socket$inet6(0xa, 0x800, 0x8d3) connect$inet6(r0, &(0x7f0000000080)={0xa, 0x4e20, 0x7fff, @private2}, 0x1c) (async) listen(r0, 0x6) r1 = socket$inet6(0xa, 0x800, 0x0) connect$inet6(r1, &(0x7f0000000200)={0xa, 0x0, 0x0, @initdev={0xfe, 0x88, '\x00', 0x0, 0x0}, 0x3}, 0x1c) (async) connect$inet6(r1, &(0x7f0000000040)={0xa, 0x4e20, 0x0, @dev={0xfe, 0x80, '\x00', 0xf}}, 0x1c) sendmmsg$inet6(r1, &(0x7f0000005a00)=[{{0x0, 0x0, 0x0}}, {{0x0, 0x0, 0x0, 0x0, &(0x7f0000005bc0)=ANY=[@ANYBLOB="5000000000000000290000003600000000060000000000000401000730000000000a0000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000020000000000000002900000037000000000000000000000004010000000000001400000000000000290000000000000160000000000000002900000037"], 0x100}}], 0x2, 0x0) [ 2818.928398][ T2831] bond1290: entered promiscuous mode [ 2818.960704][ T2831] 8021q: adding VLAN 0 to HW filter on device bond1290 [ 2819.205466][ T2834] bond1290: (slave bridge1236): making interface the new active one [ 2819.238258][ T2834] bridge1236: entered promiscuous mode 04:12:26 executing program 0: r0 = socket$inet6(0xa, 0x6, 0x0) r1 = socket$inet6(0xa, 0x80002, 0x88) bind$inet6(r1, &(0x7f0000000580)={0xa, 0x10010000004e20}, 0x1c) setsockopt$inet6_udp_int(r1, 0x11, 0xb, &(0x7f0000000080), 0x4) connect$inet6(r1, &(0x7f00000001c0)={0xa, 0x4e22, 0x57, @local, 0x81}, 0x1c) bind$inet6(r0, &(0x7f0000000000)={0xa, 0x4e20}, 0x1c) listen(r0, 0x6) r2 = socket$inet6(0xa, 0x6, 0x0) bind$inet6(r0, &(0x7f0000000180)={0xa, 0x4e24, 0x8, @remote, 0x200}, 0x1c) connect$inet6(r2, &(0x7f0000000200)={0xa, 0x0, 0x0, @initdev={0xfe, 0x88, '\x00', 0x0, 0x0}, 0x3}, 0x1c) connect$inet6(r2, &(0x7f0000000040)={0xa, 0x4e20, 0x0, @dev={0xfe, 0x80, '\x00', 0xf}}, 0x1c) sendmmsg$inet6(r2, &(0x7f0000000140)=[{{&(0x7f0000000080)={0xa, 0x4e24, 0x1000, @local, 0x9}, 0x1c, &(0x7f0000000100)=[{&(0x7f00000000c0)="274666e7114baa87a41ad47ef318c278a0e9150ba7ed62b80b3c7472e4f3671a7e3925f1e359ff786323165ab2", 0x2d}], 0x1, &(0x7f0000000240)=[@flowinfo={{0x14, 0x29, 0xb, 0x1}}, @flowinfo={{0x14, 0x29, 0xb, 0x7}}, @rthdr={{0x58, 0x29, 0x39, {0x2b, 0x8, 0x0, 0x8, 0x0, [@empty, @loopback, @loopback, @local]}}}, @hoplimit={{0x14, 0x29, 0x34, 0x8}}, @dstopts_2292={{0x50, 0x29, 0x4, {0x88, 0x7, '\x00', [@calipso={0x7, 0x30, {0x3, 0xa, 0x1f, 0x9, [0x80, 0x0, 0x0, 0x5, 0x101]}}, @jumbo]}}}, @dstopts_2292={{0x20, 0x29, 0x4, {0x2f, 0x0, '\x00', [@padn={0x1, 0x1, [0x0]}, @ra={0x5, 0x2, 0x2}]}}}], 0x110}}], 0x1, 0x0) sendmmsg$inet6(r2, &(0x7f0000005a00)=[{{0x0, 0x0, 0x0}}, {{0x0, 0x0, 0x0, 0x0, &(0x7f0000005bc0)=ANY=[@ANYBLOB="5000000000000000290000003600000000060000000000000401000730000000000a0000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000020000000000000002900000037000000000000000000000004010000000000001400000000000000290000000000000160000000000000002900000037"], 0x100}}], 0x2, 0x0) [ 2819.249415][ T2834] bond1290: (slave bridge1236): Enslaving as an active interface with an up link 04:12:26 executing program 1: r0 = socket$netlink(0x10, 0x3, 0x0) r1 = socket$netlink(0x10, 0x3, 0x0) r2 = socket(0x10, 0x803, 0x0) sendmsg$NL80211_CMD_CRIT_PROTOCOL_START(r2, &(0x7f0000000580)={0x0, 0x0, &(0x7f0000000540)={0x0, 0x1c}}, 0x0) getsockname$packet(r2, &(0x7f0000000600)={0x11, 0x0, 0x0, 0x1, 0x0, 0x6, @broadcast}, &(0x7f0000000100)=0x14) sendmsg$nl_route(r1, &(0x7f0000000040)={0x0, 0x0, &(0x7f0000000000)={&(0x7f00000000c0)=ANY=[@ANYBLOB="3c00000010008506eb9afc4cd8d06e754a0081c5", @ANYRES32=r3, @ANYBLOB="2377f292252155b21c0012000c000100626f6e64000000000c0002000800010001"], 0x3c}}, 0x0) sendmsg$nl_route(r0, &(0x7f0000000240)={0x0, 0x0, &(0x7f0000000680)={&(0x7f0000000400)=@newlink={0x3c, 0x10, 0xffffff1f, 0x0, 0xd672500, {}, [@IFLA_LINKINFO={0x14, 0x12, 0x0, 0x1, @bridge={{0xb}, {0x4}}}, @IFLA_MASTER={0x8, 0xa, r3}]}, 0x3c}}, 0x0) 04:12:26 executing program 4: r0 = socket$inet6(0xa, 0x4, 0x0) ioctl$sock_SIOCETHTOOL(r0, 0x89f0, &(0x7f0000000000)={'bridge0\x00', &(0x7f0000000200)=@ethtool_ringparam={0x10, 0x0, 0x0, 0x0, 0x0, 0x0, 0xfffffffd}}) [ 2819.534604][ T2851] bond1273: (slave bridge1194): making interface the new active one [ 2819.560338][ T2851] bridge1194: entered promiscuous mode 04:12:27 executing program 0: r0 = socket$inet6(0xa, 0x6, 0x0) r1 = socket$inet6(0xa, 0x80002, 0x88) bind$inet6(r1, &(0x7f0000000580)={0xa, 0x10010000004e20}, 0x1c) (async) bind$inet6(r1, &(0x7f0000000580)={0xa, 0x10010000004e20}, 0x1c) setsockopt$inet6_udp_int(r1, 0x11, 0xb, &(0x7f0000000080), 0x4) connect$inet6(r1, &(0x7f00000001c0)={0xa, 0x4e22, 0x57, @local, 0x81}, 0x1c) (async) connect$inet6(r1, &(0x7f00000001c0)={0xa, 0x4e22, 0x57, @local, 0x81}, 0x1c) bind$inet6(r0, &(0x7f0000000000)={0xa, 0x4e20}, 0x1c) listen(r0, 0x6) r2 = socket$inet6(0xa, 0x6, 0x0) bind$inet6(r0, &(0x7f0000000180)={0xa, 0x4e24, 0x8, @remote, 0x200}, 0x1c) connect$inet6(r2, &(0x7f0000000200)={0xa, 0x0, 0x0, @initdev={0xfe, 0x88, '\x00', 0x0, 0x0}, 0x3}, 0x1c) connect$inet6(r2, &(0x7f0000000040)={0xa, 0x4e20, 0x0, @dev={0xfe, 0x80, '\x00', 0xf}}, 0x1c) (async) connect$inet6(r2, &(0x7f0000000040)={0xa, 0x4e20, 0x0, @dev={0xfe, 0x80, '\x00', 0xf}}, 0x1c) sendmmsg$inet6(r2, &(0x7f0000000140)=[{{&(0x7f0000000080)={0xa, 0x4e24, 0x1000, @local, 0x9}, 0x1c, &(0x7f0000000100)=[{&(0x7f00000000c0)="274666e7114baa87a41ad47ef318c278a0e9150ba7ed62b80b3c7472e4f3671a7e3925f1e359ff786323165ab2", 0x2d}], 0x1, &(0x7f0000000240)=[@flowinfo={{0x14, 0x29, 0xb, 0x1}}, @flowinfo={{0x14, 0x29, 0xb, 0x7}}, @rthdr={{0x58, 0x29, 0x39, {0x2b, 0x8, 0x0, 0x8, 0x0, [@empty, @loopback, @loopback, @local]}}}, @hoplimit={{0x14, 0x29, 0x34, 0x8}}, @dstopts_2292={{0x50, 0x29, 0x4, {0x88, 0x7, '\x00', [@calipso={0x7, 0x30, {0x3, 0xa, 0x1f, 0x9, [0x80, 0x0, 0x0, 0x5, 0x101]}}, @jumbo]}}}, @dstopts_2292={{0x20, 0x29, 0x4, {0x2f, 0x0, '\x00', [@padn={0x1, 0x1, [0x0]}, @ra={0x5, 0x2, 0x2}]}}}], 0x110}}], 0x1, 0x0) sendmmsg$inet6(r2, &(0x7f0000005a00)=[{{0x0, 0x0, 0x0}}, {{0x0, 0x0, 0x0, 0x0, &(0x7f0000005bc0)=ANY=[@ANYBLOB="5000000000000000290000003600000000060000000000000401000730000000000a0000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000020000000000000002900000037000000000000000000000004010000000000001400000000000000290000000000000160000000000000002900000037"], 0x100}}], 0x2, 0x0) (async) sendmmsg$inet6(r2, &(0x7f0000005a00)=[{{0x0, 0x0, 0x0}}, {{0x0, 0x0, 0x0, 0x0, &(0x7f0000005bc0)=ANY=[@ANYBLOB="5000000000000000290000003600000000060000000000000401000730000000000a0000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000020000000000000002900000037000000000000000000000004010000000000001400000000000000290000000000000160000000000000002900000037"], 0x100}}], 0x2, 0x0) [ 2819.596288][ T2851] bond1273: (slave bridge1194): Enslaving as an active interface with an up link [ 2819.628776][ T2861] netlink: 'syz-executor.5': attribute type 1 has an invalid length. 04:12:27 executing program 2: r0 = socket$netlink(0x10, 0x3, 0x0) r1 = socket$netlink(0x10, 0x3, 0x0) r2 = socket(0x10, 0x803, 0x0) sendmsg$NL80211_CMD_CRIT_PROTOCOL_START(r2, &(0x7f0000000580)={0x0, 0x0, &(0x7f0000000540)={0x0, 0x1c}}, 0x0) getsockname$packet(r2, &(0x7f0000000600)={0x11, 0x0, 0x0, 0x1, 0x0, 0x6, @broadcast}, &(0x7f0000000100)=0x14) sendmsg$nl_route(r1, &(0x7f0000000040)={0x0, 0x0, &(0x7f0000000000)={&(0x7f00000000c0)=ANY=[@ANYBLOB="3c00000010008506eb9afc4cd8d06e754a0081c5", @ANYRES32=r3, @ANYBLOB="2377f292252155b21c0012000c000100626f6e64000000000c0002000800010001"], 0x3c}}, 0x0) sendmsg$nl_route(r0, &(0x7f0000000240)={0x0, 0x0, &(0x7f0000000680)={&(0x7f0000000400)=@newlink={0x3c, 0x10, 0xffffff1f, 0x0, 0x253c2c, {}, [@IFLA_LINKINFO={0x14, 0x12, 0x0, 0x1, @bridge={{0xb}, {0x4}}}, @IFLA_MASTER={0x8, 0xa, r3}]}, 0x3c}}, 0x0) 04:12:27 executing program 4: r0 = socket$inet6(0xa, 0x4, 0x0) ioctl$sock_SIOCETHTOOL(r0, 0x89f0, &(0x7f0000000000)={'bridge0\x00', &(0x7f0000000200)=@ethtool_ringparam={0x10, 0x0, 0x0, 0x0, 0x0, 0x0, 0xfffffffd}}) 04:12:27 executing program 0: r0 = socket$inet6(0xa, 0x6, 0x0) (async) r1 = socket$inet6(0xa, 0x80002, 0x88) bind$inet6(r1, &(0x7f0000000580)={0xa, 0x10010000004e20}, 0x1c) (async) setsockopt$inet6_udp_int(r1, 0x11, 0xb, &(0x7f0000000080), 0x4) (async) connect$inet6(r1, &(0x7f00000001c0)={0xa, 0x4e22, 0x57, @local, 0x81}, 0x1c) (async) bind$inet6(r0, &(0x7f0000000000)={0xa, 0x4e20}, 0x1c) listen(r0, 0x6) (async) r2 = socket$inet6(0xa, 0x6, 0x0) bind$inet6(r0, &(0x7f0000000180)={0xa, 0x4e24, 0x8, @remote, 0x200}, 0x1c) (async) connect$inet6(r2, &(0x7f0000000200)={0xa, 0x0, 0x0, @initdev={0xfe, 0x88, '\x00', 0x0, 0x0}, 0x3}, 0x1c) (async) connect$inet6(r2, &(0x7f0000000040)={0xa, 0x4e20, 0x0, @dev={0xfe, 0x80, '\x00', 0xf}}, 0x1c) (async) sendmmsg$inet6(r2, &(0x7f0000000140)=[{{&(0x7f0000000080)={0xa, 0x4e24, 0x1000, @local, 0x9}, 0x1c, &(0x7f0000000100)=[{&(0x7f00000000c0)="274666e7114baa87a41ad47ef318c278a0e9150ba7ed62b80b3c7472e4f3671a7e3925f1e359ff786323165ab2", 0x2d}], 0x1, &(0x7f0000000240)=[@flowinfo={{0x14, 0x29, 0xb, 0x1}}, @flowinfo={{0x14, 0x29, 0xb, 0x7}}, @rthdr={{0x58, 0x29, 0x39, {0x2b, 0x8, 0x0, 0x8, 0x0, [@empty, @loopback, @loopback, @local]}}}, @hoplimit={{0x14, 0x29, 0x34, 0x8}}, @dstopts_2292={{0x50, 0x29, 0x4, {0x88, 0x7, '\x00', [@calipso={0x7, 0x30, {0x3, 0xa, 0x1f, 0x9, [0x80, 0x0, 0x0, 0x5, 0x101]}}, @jumbo]}}}, @dstopts_2292={{0x20, 0x29, 0x4, {0x2f, 0x0, '\x00', [@padn={0x1, 0x1, [0x0]}, @ra={0x5, 0x2, 0x2}]}}}], 0x110}}], 0x1, 0x0) (async) sendmmsg$inet6(r2, &(0x7f0000005a00)=[{{0x0, 0x0, 0x0}}, {{0x0, 0x0, 0x0, 0x0, &(0x7f0000005bc0)=ANY=[@ANYBLOB="5000000000000000290000003600000000060000000000000401000730000000000a0000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000020000000000000002900000037000000000000000000000004010000000000001400000000000000290000000000000160000000000000002900000037"], 0x100}}], 0x2, 0x0) [ 2819.757368][ T2861] workqueue: Failed to create a rescuer kthread for wq "bond1283": -EINTR [ 2819.954155][ T2878] netlink: 'syz-executor.1': attribute type 1 has an invalid length. 04:12:27 executing program 5: r0 = socket$netlink(0x10, 0x3, 0x0) r1 = socket$netlink(0x10, 0x3, 0x0) r2 = socket(0x10, 0x803, 0x0) sendmsg$NL80211_CMD_CRIT_PROTOCOL_START(r2, &(0x7f0000000580)={0x0, 0x0, &(0x7f0000000540)={0x0, 0x1c}}, 0x0) getsockname$packet(r2, &(0x7f0000000600)={0x11, 0x0, 0x0, 0x1, 0x0, 0x6, @broadcast}, &(0x7f0000000100)=0x14) sendmsg$nl_route(r1, &(0x7f0000000040)={0x0, 0x0, &(0x7f0000000000)={&(0x7f00000000c0)=ANY=[@ANYBLOB="3c00000010008506eb9afc4cd8d06e754a0081c5", @ANYRES32=r3, @ANYBLOB="2377f292252155b21c0012000c000100626f6e64000000000c0002000800010001"], 0x3c}}, 0x0) sendmsg$nl_route(r0, &(0x7f0000000240)={0x0, 0x0, &(0x7f0000000680)={&(0x7f0000000400)=@newlink={0x3c, 0x10, 0xffffff1f, 0x0, 0x2535fd, {}, [@IFLA_LINKINFO={0x14, 0x12, 0x0, 0x1, @bridge={{0xb}, {0x4}}}, @IFLA_MASTER={0x8, 0xa, r3}]}, 0x3c}}, 0x0) [ 2820.085834][ T2878] bond1291: entered promiscuous mode [ 2820.091438][ T2878] 8021q: adding VLAN 0 to HW filter on device bond1291 04:12:27 executing program 0: r0 = socket$inet6(0xa, 0x1, 0x200) bind$inet6(r0, &(0x7f0000000000)={0xa, 0x4e20}, 0x1c) listen(r0, 0x6) r1 = socket$inet6(0xa, 0x6, 0x0) connect$inet6(r1, &(0x7f0000000200)={0xa, 0x0, 0x0, @initdev={0xfe, 0x88, '\x00', 0x0, 0x0}, 0x3}, 0x1c) connect$inet6(r1, &(0x7f0000000040)={0xa, 0x4e20, 0x5, @dev={0xfe, 0x80, '\x00', 0x3a}}, 0x1c) r2 = accept$inet6(r1, &(0x7f0000000080)={0xa, 0x0, 0x0, @local}, &(0x7f00000000c0)=0x1c) bind$inet6(r1, &(0x7f0000000280)={0xa, 0x4e23, 0x0, @private2, 0x1}, 0x1c) sendto$inet6(r2, &(0x7f0000000100)="ef5e152b036be3e4a3fd87e9b25c36ec51c97ea193d88e2733fe9d2c48a9cd33ddef7cc56d6648d062a39848732b2774b363f3d9ad36d715a3dd", 0x3a, 0x8001, &(0x7f0000000140)={0xa, 0x4e23, 0x80000000, @mcast2, 0x619e}, 0x1c) r3 = socket$inet6(0xa, 0x80002, 0x88) bind$inet6(r3, &(0x7f0000000580)={0xa, 0x10010000004e20}, 0x1c) setsockopt$inet6_udp_int(r3, 0x11, 0xb, &(0x7f0000000080), 0x4) setsockopt$inet_sctp6_SCTP_PARTIAL_DELIVERY_POINT(r0, 0x84, 0x13, &(0x7f0000000240)=0x2, 0x4) sendto$inet6(r3, &(0x7f0000000180)="1ecc13986388d5d9f99741567d95618abec38cab0618", 0x16, 0x90, &(0x7f00000001c0)={0xa, 0x4e23, 0x6, @initdev={0xfe, 0x88, '\x00', 0x1, 0x0}, 0x8}, 0x1c) 04:12:27 executing program 4: r0 = socket$inet6(0xa, 0x4, 0x0) ioctl$sock_SIOCETHTOOL(r0, 0x89f0, &(0x7f0000000000)={'bridge0\x00', &(0x7f0000000200)=@ethtool_ringparam={0x10, 0x0, 0x0, 0x0, 0x0, 0x0, 0xfffffffd}}) [ 2820.472576][ T2879] bond1291: (slave bridge1237): making interface the new active one [ 2820.496942][ T2879] bridge1237: entered promiscuous mode [ 2820.531177][ T2879] bond1291: (slave bridge1237): Enslaving as an active interface with an up link [ 2820.567440][ T2888] netlink: 'syz-executor.2': attribute type 1 has an invalid length. 04:12:28 executing program 3: r0 = socket$netlink(0x10, 0x3, 0x0) sendmsg$netlink(r0, &(0x7f00000047c0)={0x0, 0x0, &(0x7f0000000000)=[{&(0x7f0000000100)=ANY=[@ANYBLOB="18000000560001060000007200db730007"], 0x18}], 0x1}, 0x0) 04:12:28 executing program 1: r0 = socket$netlink(0x10, 0x3, 0x0) r1 = socket$netlink(0x10, 0x3, 0x0) r2 = socket(0x10, 0x803, 0x0) sendmsg$NL80211_CMD_CRIT_PROTOCOL_START(r2, &(0x7f0000000580)={0x0, 0x0, &(0x7f0000000540)={0x0, 0x1c}}, 0x0) getsockname$packet(r2, &(0x7f0000000600)={0x11, 0x0, 0x0, 0x1, 0x0, 0x6, @broadcast}, &(0x7f0000000100)=0x14) sendmsg$nl_route(r1, &(0x7f0000000040)={0x0, 0x0, &(0x7f0000000000)={&(0x7f00000000c0)=ANY=[@ANYBLOB="3c00000010008506eb9afc4cd8d06e754a0081c5", @ANYRES32=r3, @ANYBLOB="2377f292252155b21c0012000c000100626f6e64000000000c0002000800010001"], 0x3c}}, 0x0) sendmsg$nl_route(r0, &(0x7f0000000240)={0x0, 0x0, &(0x7f0000000680)={&(0x7f0000000400)=@newlink={0x3c, 0x10, 0xffffff1f, 0x0, 0xe672500, {}, [@IFLA_LINKINFO={0x14, 0x12, 0x0, 0x1, @bridge={{0xb}, {0x4}}}, @IFLA_MASTER={0x8, 0xa, r3}]}, 0x3c}}, 0x0) 04:12:28 executing program 4: r0 = socket$inet6(0xa, 0x2, 0x0) ioctl$sock_SIOCETHTOOL(r0, 0x89f0, &(0x7f0000000000)={'bridge0\x00', &(0x7f0000000200)=@ethtool_ringparam={0x10}}) r1 = socket$inet6(0xa, 0x80002, 0x88) bind$inet6(r1, &(0x7f0000000580)={0xa, 0x10010000004e20}, 0x1c) setsockopt$inet6_udp_int(r1, 0x11, 0xb, &(0x7f0000000080), 0x4) bind$inet6(r1, &(0x7f0000000040)={0xa, 0x4e24, 0x7, @private2, 0x1ff}, 0x1c) [ 2820.737202][ T2888] bond1274: entered promiscuous mode [ 2820.770400][ T2888] 8021q: adding VLAN 0 to HW filter on device bond1274 [ 2820.919599][ T2890] bond1274: (slave bridge1195): making interface the new active one 04:12:28 executing program 2: r0 = socket$netlink(0x10, 0x3, 0x0) r1 = socket$netlink(0x10, 0x3, 0x0) r2 = socket(0x10, 0x803, 0x0) sendmsg$NL80211_CMD_CRIT_PROTOCOL_START(r2, &(0x7f0000000580)={0x0, 0x0, &(0x7f0000000540)={0x0, 0x1c}}, 0x0) getsockname$packet(r2, &(0x7f0000000600)={0x11, 0x0, 0x0, 0x1, 0x0, 0x6, @broadcast}, &(0x7f0000000100)=0x14) sendmsg$nl_route(r1, &(0x7f0000000040)={0x0, 0x0, &(0x7f0000000000)={&(0x7f00000000c0)=ANY=[@ANYBLOB="3c00000010008506eb9afc4cd8d06e754a0081c5", @ANYRES32=r3, @ANYBLOB="2377f292252155b21c0012000c000100626f6e64000000000c0002000800010001"], 0x3c}}, 0x0) sendmsg$nl_route(r0, &(0x7f0000000240)={0x0, 0x0, &(0x7f0000000680)={&(0x7f0000000400)=@newlink={0x3c, 0x10, 0xffffff1f, 0x0, 0x253c2d, {}, [@IFLA_LINKINFO={0x14, 0x12, 0x0, 0x1, @bridge={{0xb}, {0x4}}}, @IFLA_MASTER={0x8, 0xa, r3}]}, 0x3c}}, 0x0) [ 2820.963550][ T2890] bridge1195: entered promiscuous mode [ 2820.976787][ T2890] bond1274: (slave bridge1195): Enslaving as an active interface with an up link [ 2820.989005][ T2905] netlink: 'syz-executor.5': attribute type 1 has an invalid length. 04:12:28 executing program 5: r0 = socket$netlink(0x10, 0x3, 0x0) r1 = socket$netlink(0x10, 0x3, 0x0) r2 = socket(0x10, 0x803, 0x0) sendmsg$NL80211_CMD_CRIT_PROTOCOL_START(r2, &(0x7f0000000580)={0x0, 0x0, &(0x7f0000000540)={0x0, 0x1c}}, 0x0) getsockname$packet(r2, &(0x7f0000000600)={0x11, 0x0, 0x0, 0x1, 0x0, 0x6, @broadcast}, &(0x7f0000000100)=0x14) sendmsg$nl_route(r1, &(0x7f0000000040)={0x0, 0x0, &(0x7f0000000000)={&(0x7f00000000c0)=ANY=[@ANYBLOB="3c00000010008506eb9afc4cd8d06e754a0081c5", @ANYRES32=r3, @ANYBLOB="2377f292252155b21c0012000c000100626f6e64000000000c0002000800010001"], 0x3c}}, 0x0) sendmsg$nl_route(r0, &(0x7f0000000240)={0x0, 0x0, &(0x7f0000000680)={&(0x7f0000000400)=@newlink={0x3c, 0x10, 0xffffff1f, 0x0, 0x400300, {}, [@IFLA_LINKINFO={0x14, 0x12, 0x0, 0x1, @bridge={{0xb}, {0x4}}}, @IFLA_MASTER={0x8, 0xa, r3}]}, 0x3c}}, 0x0) 04:12:28 executing program 0: r0 = socket$inet6(0xa, 0x1, 0x200) bind$inet6(r0, &(0x7f0000000000)={0xa, 0x4e20}, 0x1c) (async) listen(r0, 0x6) (async) r1 = socket$inet6(0xa, 0x6, 0x0) connect$inet6(r1, &(0x7f0000000200)={0xa, 0x0, 0x0, @initdev={0xfe, 0x88, '\x00', 0x0, 0x0}, 0x3}, 0x1c) (async) connect$inet6(r1, &(0x7f0000000040)={0xa, 0x4e20, 0x5, @dev={0xfe, 0x80, '\x00', 0x3a}}, 0x1c) r2 = accept$inet6(r1, &(0x7f0000000080)={0xa, 0x0, 0x0, @local}, &(0x7f00000000c0)=0x1c) bind$inet6(r1, &(0x7f0000000280)={0xa, 0x4e23, 0x0, @private2, 0x1}, 0x1c) (async) sendto$inet6(r2, &(0x7f0000000100)="ef5e152b036be3e4a3fd87e9b25c36ec51c97ea193d88e2733fe9d2c48a9cd33ddef7cc56d6648d062a39848732b2774b363f3d9ad36d715a3dd", 0x3a, 0x8001, &(0x7f0000000140)={0xa, 0x4e23, 0x80000000, @mcast2, 0x619e}, 0x1c) r3 = socket$inet6(0xa, 0x80002, 0x88) bind$inet6(r3, &(0x7f0000000580)={0xa, 0x10010000004e20}, 0x1c) (async) setsockopt$inet6_udp_int(r3, 0x11, 0xb, &(0x7f0000000080), 0x4) setsockopt$inet_sctp6_SCTP_PARTIAL_DELIVERY_POINT(r0, 0x84, 0x13, &(0x7f0000000240)=0x2, 0x4) (async) sendto$inet6(r3, &(0x7f0000000180)="1ecc13986388d5d9f99741567d95618abec38cab0618", 0x16, 0x90, &(0x7f00000001c0)={0xa, 0x4e23, 0x6, @initdev={0xfe, 0x88, '\x00', 0x1, 0x0}, 0x8}, 0x1c) [ 2821.035662][ T2905] workqueue: Failed to create a rescuer kthread for wq "bond1283": -EINTR [ 2821.222204][ T2920] netlink: 'syz-executor.1': attribute type 1 has an invalid length. 04:12:28 executing program 4: r0 = socket$inet6(0xa, 0x2, 0x0) ioctl$sock_SIOCETHTOOL(r0, 0x89f0, &(0x7f0000000000)={'bridge0\x00', &(0x7f0000000200)=@ethtool_ringparam={0x10}}) r1 = socket$inet6(0xa, 0x80002, 0x88) bind$inet6(r1, &(0x7f0000000580)={0xa, 0x10010000004e20}, 0x1c) (async) setsockopt$inet6_udp_int(r1, 0x11, 0xb, &(0x7f0000000080), 0x4) bind$inet6(r1, &(0x7f0000000040)={0xa, 0x4e24, 0x7, @private2, 0x1ff}, 0x1c) [ 2821.351133][ T2920] bond1292: entered promiscuous mode [ 2821.358671][ T2920] 8021q: adding VLAN 0 to HW filter on device bond1292 [ 2821.537168][ T2923] bond1292: (slave bridge1238): making interface the new active one [ 2821.550383][ T2923] bridge1238: entered promiscuous mode [ 2821.580189][ T2923] bond1292: (slave bridge1238): Enslaving as an active interface with an up link 04:12:29 executing program 1: r0 = socket$netlink(0x10, 0x3, 0x0) r1 = socket$netlink(0x10, 0x3, 0x0) r2 = socket(0x10, 0x803, 0x0) sendmsg$NL80211_CMD_CRIT_PROTOCOL_START(r2, &(0x7f0000000580)={0x0, 0x0, &(0x7f0000000540)={0x0, 0x1c}}, 0x0) getsockname$packet(r2, &(0x7f0000000600)={0x11, 0x0, 0x0, 0x1, 0x0, 0x6, @broadcast}, &(0x7f0000000100)=0x14) sendmsg$nl_route(r1, &(0x7f0000000040)={0x0, 0x0, &(0x7f0000000000)={&(0x7f00000000c0)=ANY=[@ANYBLOB="3c00000010008506eb9afc4cd8d06e754a0081c5", @ANYRES32=r3, @ANYBLOB="2377f292252155b21c0012000c000100626f6e64000000000c0002000800010001"], 0x3c}}, 0x0) sendmsg$nl_route(r0, &(0x7f0000000240)={0x0, 0x0, &(0x7f0000000680)={&(0x7f0000000400)=@newlink={0x3c, 0x10, 0xffffff1f, 0x0, 0xf672500, {}, [@IFLA_LINKINFO={0x14, 0x12, 0x0, 0x1, @bridge={{0xb}, {0x4}}}, @IFLA_MASTER={0x8, 0xa, r3}]}, 0x3c}}, 0x0) [ 2821.706603][ T2927] bond1283: entered promiscuous mode [ 2821.739223][ T2927] 8021q: adding VLAN 0 to HW filter on device bond1283 04:12:29 executing program 3: r0 = socket$netlink(0x10, 0x3, 0x0) sendmsg$netlink(r0, &(0x7f00000047c0)={0x0, 0x0, &(0x7f0000000000)=[{&(0x7f0000000100)=ANY=[@ANYBLOB="18000000560001060000007300db730007"], 0x18}], 0x1}, 0x0) [ 2821.938079][ T2930] bond1275: entered promiscuous mode [ 2821.943851][ T2930] 8021q: adding VLAN 0 to HW filter on device bond1275 04:12:29 executing program 0: r0 = socket$inet6(0xa, 0x1, 0x200) bind$inet6(r0, &(0x7f0000000000)={0xa, 0x4e20}, 0x1c) listen(r0, 0x6) r1 = socket$inet6(0xa, 0x6, 0x0) connect$inet6(r1, &(0x7f0000000200)={0xa, 0x0, 0x0, @initdev={0xfe, 0x88, '\x00', 0x0, 0x0}, 0x3}, 0x1c) connect$inet6(r1, &(0x7f0000000040)={0xa, 0x4e20, 0x5, @dev={0xfe, 0x80, '\x00', 0x3a}}, 0x1c) (async) connect$inet6(r1, &(0x7f0000000040)={0xa, 0x4e20, 0x5, @dev={0xfe, 0x80, '\x00', 0x3a}}, 0x1c) r2 = accept$inet6(r1, &(0x7f0000000080)={0xa, 0x0, 0x0, @local}, &(0x7f00000000c0)=0x1c) bind$inet6(r1, &(0x7f0000000280)={0xa, 0x4e23, 0x0, @private2, 0x1}, 0x1c) (async) bind$inet6(r1, &(0x7f0000000280)={0xa, 0x4e23, 0x0, @private2, 0x1}, 0x1c) sendto$inet6(r2, &(0x7f0000000100)="ef5e152b036be3e4a3fd87e9b25c36ec51c97ea193d88e2733fe9d2c48a9cd33ddef7cc56d6648d062a39848732b2774b363f3d9ad36d715a3dd", 0x3a, 0x8001, &(0x7f0000000140)={0xa, 0x4e23, 0x80000000, @mcast2, 0x619e}, 0x1c) (async) sendto$inet6(r2, &(0x7f0000000100)="ef5e152b036be3e4a3fd87e9b25c36ec51c97ea193d88e2733fe9d2c48a9cd33ddef7cc56d6648d062a39848732b2774b363f3d9ad36d715a3dd", 0x3a, 0x8001, &(0x7f0000000140)={0xa, 0x4e23, 0x80000000, @mcast2, 0x619e}, 0x1c) r3 = socket$inet6(0xa, 0x80002, 0x88) bind$inet6(r3, &(0x7f0000000580)={0xa, 0x10010000004e20}, 0x1c) setsockopt$inet6_udp_int(r3, 0x11, 0xb, &(0x7f0000000080), 0x4) (async) setsockopt$inet6_udp_int(r3, 0x11, 0xb, &(0x7f0000000080), 0x4) setsockopt$inet_sctp6_SCTP_PARTIAL_DELIVERY_POINT(r0, 0x84, 0x13, &(0x7f0000000240)=0x2, 0x4) sendto$inet6(r3, &(0x7f0000000180)="1ecc13986388d5d9f99741567d95618abec38cab0618", 0x16, 0x90, &(0x7f00000001c0)={0xa, 0x4e23, 0x6, @initdev={0xfe, 0x88, '\x00', 0x1, 0x0}, 0x8}, 0x1c) [ 2822.150710][ T2931] bond1283: (slave bridge1246): making interface the new active one [ 2822.195720][ T2931] bridge1246: entered promiscuous mode [ 2822.219645][ T2931] bond1283: (slave bridge1246): Enslaving as an active interface with an up link 04:12:29 executing program 5: r0 = socket$netlink(0x10, 0x3, 0x0) r1 = socket$netlink(0x10, 0x3, 0x0) r2 = socket(0x10, 0x803, 0x0) sendmsg$NL80211_CMD_CRIT_PROTOCOL_START(r2, &(0x7f0000000580)={0x0, 0x0, &(0x7f0000000540)={0x0, 0x1c}}, 0x0) getsockname$packet(r2, &(0x7f0000000600)={0x11, 0x0, 0x0, 0x1, 0x0, 0x6, @broadcast}, &(0x7f0000000100)=0x14) sendmsg$nl_route(r1, &(0x7f0000000040)={0x0, 0x0, &(0x7f0000000000)={&(0x7f00000000c0)=ANY=[@ANYBLOB="3c00000010008506eb9afc4cd8d06e754a0081c5", @ANYRES32=r3, @ANYBLOB="2377f292252155b21c0012000c000100626f6e64000000000c0002000800010001"], 0x3c}}, 0x0) sendmsg$nl_route(r0, &(0x7f0000000240)={0x0, 0x0, &(0x7f0000000680)={&(0x7f0000000400)=@newlink={0x3c, 0x10, 0xffffff1f, 0x0, 0xf0ffff, {}, [@IFLA_LINKINFO={0x14, 0x12, 0x0, 0x1, @bridge={{0xb}, {0x4}}}, @IFLA_MASTER={0x8, 0xa, r3}]}, 0x3c}}, 0x0) 04:12:29 executing program 0: r0 = socket$inet6(0xa, 0x6, 0x0) r1 = socket$inet6_tcp(0xa, 0x1, 0x0) getsockopt$inet6_mreq(r1, 0x29, 0x15, &(0x7f0000000080)={@empty}, &(0x7f00000000c0)=0x14) bind$inet6(r0, &(0x7f0000000000)={0xa, 0x4e20}, 0x1c) listen(r0, 0x6) r2 = socket$inet6(0xa, 0x6, 0x0) connect$inet6(r2, &(0x7f0000000200)={0xa, 0x0, 0x0, @initdev={0xfe, 0x88, '\x00', 0x0, 0x0}, 0x3}, 0x1c) connect$inet6(r2, &(0x7f0000000040)={0xa, 0x4e20, 0x0, @dev={0xfe, 0x80, '\x00', 0xf}}, 0x1c) sendmmsg$inet6(r2, &(0x7f0000005a00)=[{{0x0, 0x0, 0x0}}, {{0x0, 0x0, 0x0, 0x0, &(0x7f0000005bc0)=ANY=[@ANYBLOB="5000000000000000290000003600000000060000000000000401000730000000000a0000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000020000000000000002900000037000000000000000000000004010000000000001400000000000000290000000000000160000000000000002900000037"], 0x100}}], 0x2, 0x0) [ 2822.376978][ T2932] bond1275: (slave bridge1196): making interface the new active one [ 2822.432926][ T2932] bridge1196: entered promiscuous mode [ 2822.485225][ T2932] bond1275: (slave bridge1196): Enslaving as an active interface with an up link 04:12:30 executing program 0: r0 = socket$inet6(0xa, 0x6, 0x0) r1 = socket$inet6_tcp(0xa, 0x1, 0x0) getsockopt$inet6_mreq(r1, 0x29, 0x15, &(0x7f0000000080)={@empty}, &(0x7f00000000c0)=0x14) (async) bind$inet6(r0, &(0x7f0000000000)={0xa, 0x4e20}, 0x1c) listen(r0, 0x6) r2 = socket$inet6(0xa, 0x6, 0x0) connect$inet6(r2, &(0x7f0000000200)={0xa, 0x0, 0x0, @initdev={0xfe, 0x88, '\x00', 0x0, 0x0}, 0x3}, 0x1c) (async) connect$inet6(r2, &(0x7f0000000040)={0xa, 0x4e20, 0x0, @dev={0xfe, 0x80, '\x00', 0xf}}, 0x1c) sendmmsg$inet6(r2, &(0x7f0000005a00)=[{{0x0, 0x0, 0x0}}, {{0x0, 0x0, 0x0, 0x0, &(0x7f0000005bc0)=ANY=[@ANYBLOB="5000000000000000290000003600000000060000000000000401000730000000000a0000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000020000000000000002900000037000000000000000000000004010000000000001400000000000000290000000000000160000000000000002900000037"], 0x100}}], 0x2, 0x0) 04:12:30 executing program 4: r0 = socket$inet6(0xa, 0x2, 0x0) ioctl$sock_SIOCETHTOOL(r0, 0x89f0, &(0x7f0000000000)={'bridge0\x00', &(0x7f0000000200)=@ethtool_ringparam={0x10}}) socket$inet6(0xa, 0x80002, 0x88) (async) r1 = socket$inet6(0xa, 0x80002, 0x88) bind$inet6(r1, &(0x7f0000000580)={0xa, 0x10010000004e20}, 0x1c) setsockopt$inet6_udp_int(r1, 0x11, 0xb, &(0x7f0000000080), 0x4) (async) setsockopt$inet6_udp_int(r1, 0x11, 0xb, &(0x7f0000000080), 0x4) bind$inet6(r1, &(0x7f0000000040)={0xa, 0x4e24, 0x7, @private2, 0x1ff}, 0x1c) 04:12:30 executing program 2: r0 = socket$netlink(0x10, 0x3, 0x0) r1 = socket$netlink(0x10, 0x3, 0x0) r2 = socket(0x10, 0x803, 0x0) sendmsg$NL80211_CMD_CRIT_PROTOCOL_START(r2, &(0x7f0000000580)={0x0, 0x0, &(0x7f0000000540)={0x0, 0x1c}}, 0x0) getsockname$packet(r2, &(0x7f0000000600)={0x11, 0x0, 0x0, 0x1, 0x0, 0x6, @broadcast}, &(0x7f0000000100)=0x14) sendmsg$nl_route(r1, &(0x7f0000000040)={0x0, 0x0, &(0x7f0000000000)={&(0x7f00000000c0)=ANY=[@ANYBLOB="3c00000010008506eb9afc4cd8d06e754a0081c5", @ANYRES32=r3, @ANYBLOB="2377f292252155b21c0012000c000100626f6e64000000000c0002000800010001"], 0x3c}}, 0x0) sendmsg$nl_route(r0, &(0x7f0000000240)={0x0, 0x0, &(0x7f0000000680)={&(0x7f0000000400)=@newlink={0x3c, 0x10, 0xffffff1f, 0x0, 0x253c2e, {}, [@IFLA_LINKINFO={0x14, 0x12, 0x0, 0x1, @bridge={{0xb}, {0x4}}}, @IFLA_MASTER={0x8, 0xa, r3}]}, 0x3c}}, 0x0) [ 2822.580609][ T2944] validate_nla: 2 callbacks suppressed [ 2822.580631][ T2944] netlink: 'syz-executor.1': attribute type 1 has an invalid length. 04:12:30 executing program 0: r0 = socket$inet6(0xa, 0x6, 0x0) r1 = socket$inet6_tcp(0xa, 0x1, 0x0) getsockopt$inet6_mreq(r1, 0x29, 0x15, &(0x7f0000000080)={@empty}, &(0x7f00000000c0)=0x14) bind$inet6(r0, &(0x7f0000000000)={0xa, 0x4e20}, 0x1c) listen(r0, 0x6) r2 = socket$inet6(0xa, 0x6, 0x0) connect$inet6(r2, &(0x7f0000000200)={0xa, 0x0, 0x0, @initdev={0xfe, 0x88, '\x00', 0x0, 0x0}, 0x3}, 0x1c) (async) connect$inet6(r2, &(0x7f0000000040)={0xa, 0x4e20, 0x0, @dev={0xfe, 0x80, '\x00', 0xf}}, 0x1c) (async, rerun: 32) sendmmsg$inet6(r2, &(0x7f0000005a00)=[{{0x0, 0x0, 0x0}}, {{0x0, 0x0, 0x0, 0x0, &(0x7f0000005bc0)=ANY=[@ANYBLOB="5000000000000000290000003600000000060000000000000401000730000000000a0000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000020000000000000002900000037000000000000000000000004010000000000001400000000000000290000000000000160000000000000002900000037"], 0x100}}], 0x2, 0x0) (rerun: 32) 04:12:30 executing program 0: r0 = openat$tun(0xffffffffffffff9c, &(0x7f0000000140), 0x0, 0x0) ioctl$TUNSETQUEUE(r0, 0x400454d9, &(0x7f0000000240)={'dvmrp0\x00'}) 04:12:30 executing program 1: r0 = socket$netlink(0x10, 0x3, 0x0) r1 = socket$netlink(0x10, 0x3, 0x0) r2 = socket(0x10, 0x803, 0x0) sendmsg$NL80211_CMD_CRIT_PROTOCOL_START(r2, &(0x7f0000000580)={0x0, 0x0, &(0x7f0000000540)={0x0, 0x1c}}, 0x0) getsockname$packet(r2, &(0x7f0000000600)={0x11, 0x0, 0x0, 0x1, 0x0, 0x6, @broadcast}, &(0x7f0000000100)=0x14) sendmsg$nl_route(r1, &(0x7f0000000040)={0x0, 0x0, &(0x7f0000000000)={&(0x7f00000000c0)=ANY=[@ANYBLOB="3c00000010008506eb9afc4cd8d06e754a0081c5", @ANYRES32=r3, @ANYBLOB="2377f292252155b21c0012000c000100626f6e64000000000c0002000800010001"], 0x3c}}, 0x0) sendmsg$nl_route(r0, &(0x7f0000000240)={0x0, 0x0, &(0x7f0000000680)={&(0x7f0000000400)=@newlink={0x3c, 0x10, 0xffffff1f, 0x0, 0x10672500, {}, [@IFLA_LINKINFO={0x14, 0x12, 0x0, 0x1, @bridge={{0xb}, {0x4}}}, @IFLA_MASTER={0x8, 0xa, r3}]}, 0x3c}}, 0x0) [ 2822.679936][ T2944] workqueue: Failed to create a rescuer kthread for wq "bond1293": -EINTR [ 2822.966305][ T2958] netlink: 'syz-executor.5': attribute type 1 has an invalid length. 04:12:30 executing program 3: r0 = socket$netlink(0x10, 0x3, 0x0) sendmsg$netlink(r0, &(0x7f00000047c0)={0x0, 0x0, &(0x7f0000000000)=[{&(0x7f0000000100)=ANY=[@ANYBLOB="18000000560001060000007400db730007"], 0x18}], 0x1}, 0x0) [ 2823.034234][ T2958] bond1284: entered promiscuous mode [ 2823.040167][ T2958] 8021q: adding VLAN 0 to HW filter on device bond1284 [ 2823.179260][ T2959] bond1284: (slave bridge1247): making interface the new active one [ 2823.201426][ T2959] bridge1247: entered promiscuous mode [ 2823.217155][ T2959] bond1284: (slave bridge1247): Enslaving as an active interface with an up link [ 2823.247677][ T2966] netlink: 'syz-executor.2': attribute type 1 has an invalid length. 04:12:30 executing program 5: r0 = socket$netlink(0x10, 0x3, 0x0) r1 = socket$netlink(0x10, 0x3, 0x0) r2 = socket(0x10, 0x803, 0x0) sendmsg$NL80211_CMD_CRIT_PROTOCOL_START(r2, &(0x7f0000000580)={0x0, 0x0, &(0x7f0000000540)={0x0, 0x1c}}, 0x0) getsockname$packet(r2, &(0x7f0000000600)={0x11, 0x0, 0x0, 0x1, 0x0, 0x6, @broadcast}, &(0x7f0000000100)=0x14) sendmsg$nl_route(r1, &(0x7f0000000040)={0x0, 0x0, &(0x7f0000000000)={&(0x7f00000000c0)=ANY=[@ANYBLOB="3c00000010008506eb9afc4cd8d06e754a0081c5", @ANYRES32=r3, @ANYBLOB="2377f292252155b21c0012000c000100626f6e64000000000c0002000800010001"], 0x3c}}, 0x0) sendmsg$nl_route(r0, &(0x7f0000000240)={0x0, 0x0, &(0x7f0000000680)={&(0x7f0000000400)=@newlink={0x3c, 0x10, 0xffffff1f, 0x0, 0x1000000, {}, [@IFLA_LINKINFO={0x14, 0x12, 0x0, 0x1, @bridge={{0xb}, {0x4}}}, @IFLA_MASTER={0x8, 0xa, r3}]}, 0x3c}}, 0x0) [ 2823.409382][ T2966] bond1276: entered promiscuous mode [ 2823.415771][ T2966] 8021q: adding VLAN 0 to HW filter on device bond1276 [ 2823.573972][ T2969] bond1276: (slave bridge1197): making interface the new active one [ 2823.588025][ T2969] bridge1197: entered promiscuous mode [ 2823.599629][ T2969] bond1276: (slave bridge1197): Enslaving as an active interface with an up link 04:12:31 executing program 3: r0 = socket$netlink(0x10, 0x3, 0x0) sendmsg$netlink(r0, &(0x7f00000047c0)={0x0, 0x0, &(0x7f0000000000)=[{&(0x7f0000000100)=ANY=[@ANYBLOB="18000000560001060000007500db730007"], 0x18}], 0x1}, 0x0) 04:12:31 executing program 4: r0 = socket$inet6(0xa, 0x2, 0x0) ioctl$sock_SIOCETHTOOL(r0, 0x89f0, &(0x7f0000000000)={'bridge0\x00', &(0x7f0000000200)=ANY=[@ANYBLOB="100000000000000000000000000000002000"/36]}) connect$inet6(r0, &(0x7f0000000040)={0xa, 0x4e21, 0x24, @private2, 0x5}, 0x1c) [ 2823.637657][ T2980] netlink: 'syz-executor.1': attribute type 1 has an invalid length. 04:12:31 executing program 2: r0 = socket$netlink(0x10, 0x3, 0x0) r1 = socket$netlink(0x10, 0x3, 0x0) r2 = socket(0x10, 0x803, 0x0) sendmsg$NL80211_CMD_CRIT_PROTOCOL_START(r2, &(0x7f0000000580)={0x0, 0x0, &(0x7f0000000540)={0x0, 0x1c}}, 0x0) getsockname$packet(r2, &(0x7f0000000600)={0x11, 0x0, 0x0, 0x1, 0x0, 0x6, @broadcast}, &(0x7f0000000100)=0x14) sendmsg$nl_route(r1, &(0x7f0000000040)={0x0, 0x0, &(0x7f0000000000)={&(0x7f00000000c0)=ANY=[@ANYBLOB="3c00000010008506eb9afc4cd8d06e754a0081c5", @ANYRES32=r3, @ANYBLOB="2377f292252155b21c0012000c000100626f6e64000000000c0002000800010001"], 0x3c}}, 0x0) sendmsg$nl_route(r0, &(0x7f0000000240)={0x0, 0x0, &(0x7f0000000680)={&(0x7f0000000400)=@newlink={0x3c, 0x10, 0xffffff1f, 0x0, 0x253c2f, {}, [@IFLA_LINKINFO={0x14, 0x12, 0x0, 0x1, @bridge={{0xb}, {0x4}}}, @IFLA_MASTER={0x8, 0xa, r3}]}, 0x3c}}, 0x0) [ 2823.804926][ T2980] bond1293: entered promiscuous mode [ 2823.810718][ T2980] 8021q: adding VLAN 0 to HW filter on device bond1293 [ 2823.848742][ T2984] bond1293: (slave bridge1239): making interface the new active one [ 2823.857282][ T2984] bridge1239: entered promiscuous mode [ 2823.867217][ T2984] bond1293: (slave bridge1239): Enslaving as an active interface with an up link [ 2823.985521][ T2987] netlink: 'syz-executor.5': attribute type 1 has an invalid length. 04:12:31 executing program 1: r0 = socket$netlink(0x10, 0x3, 0x0) r1 = socket$netlink(0x10, 0x3, 0x0) r2 = socket(0x10, 0x803, 0x0) sendmsg$NL80211_CMD_CRIT_PROTOCOL_START(r2, &(0x7f0000000580)={0x0, 0x0, &(0x7f0000000540)={0x0, 0x1c}}, 0x0) getsockname$packet(r2, &(0x7f0000000600)={0x11, 0x0, 0x0, 0x1, 0x0, 0x6, @broadcast}, &(0x7f0000000100)=0x14) sendmsg$nl_route(r1, &(0x7f0000000040)={0x0, 0x0, &(0x7f0000000000)={&(0x7f00000000c0)=ANY=[@ANYBLOB="3c00000010008506eb9afc4cd8d06e754a0081c5", @ANYRES32=r3, @ANYBLOB="2377f292252155b21c0012000c000100626f6e64000000000c0002000800010001"], 0x3c}}, 0x0) sendmsg$nl_route(r0, &(0x7f0000000240)={0x0, 0x0, &(0x7f0000000680)={&(0x7f0000000400)=@newlink={0x3c, 0x10, 0xffffff1f, 0x0, 0x11672500, {}, [@IFLA_LINKINFO={0x14, 0x12, 0x0, 0x1, @bridge={{0xb}, {0x4}}}, @IFLA_MASTER={0x8, 0xa, r3}]}, 0x3c}}, 0x0) [ 2824.098409][ T2987] bond1285: entered promiscuous mode [ 2824.104368][ T2987] 8021q: adding VLAN 0 to HW filter on device bond1285 04:12:31 executing program 0: bpf$BPF_PROG_RAW_TRACEPOINT_LOAD(0x5, &(0x7f0000000240)={0x18, 0x10, &(0x7f0000000700)=ANY=[@ANYBLOB="1800000093f9397b000000008000000005000000ffffffff1830e8000300000000000000000000009500000000000000182b0000fda338cb19cc07629461496137132819ad4bbd5a297e7ceff7b1e6e5bdab4a"], &(0x7f0000000080)='syzkaller\x00', 0x9, 0x0, 0x0, 0x0, 0x0, '\x00', 0x0, 0x0, 0xffffffffffffffff, 0x8, 0x0, 0x0, 0x10, 0x0}, 0x90) [ 2824.154494][ T2988] bond1285: (slave bridge1248): making interface the new active one [ 2824.162789][ T2988] bridge1248: entered promiscuous mode [ 2824.174542][ T2988] bond1285: (slave bridge1248): Enslaving as an active interface with an up link 04:12:31 executing program 0: socketpair$unix(0x1, 0x5, 0x0, &(0x7f0000000380)={0xffffffffffffffff, 0xffffffffffffffff}) sendmsg$inet(r0, &(0x7f00000006c0)={0x0, 0x0, &(0x7f0000000580)=[{&(0x7f0000000400)="0f", 0x3ec0}], 0x1}, 0x0) 04:12:31 executing program 5: r0 = socket$netlink(0x10, 0x3, 0x0) r1 = socket$netlink(0x10, 0x3, 0x0) r2 = socket(0x10, 0x803, 0x0) sendmsg$NL80211_CMD_CRIT_PROTOCOL_START(r2, &(0x7f0000000580)={0x0, 0x0, &(0x7f0000000540)={0x0, 0x1c}}, 0x0) getsockname$packet(r2, &(0x7f0000000600)={0x11, 0x0, 0x0, 0x1, 0x0, 0x6, @broadcast}, &(0x7f0000000100)=0x14) sendmsg$nl_route(r1, &(0x7f0000000040)={0x0, 0x0, &(0x7f0000000000)={&(0x7f00000000c0)=ANY=[@ANYBLOB="3c00000010008506eb9afc4cd8d06e754a0081c5", @ANYRES32=r3, @ANYBLOB="2377f292252155b21c0012000c000100626f6e64000000000c0002000800010001"], 0x3c}}, 0x0) sendmsg$nl_route(r0, &(0x7f0000000240)={0x0, 0x0, &(0x7f0000000680)={&(0x7f0000000400)=@newlink={0x3c, 0x10, 0xffffff1f, 0x0, 0x2000000, {}, [@IFLA_LINKINFO={0x14, 0x12, 0x0, 0x1, @bridge={{0xb}, {0x4}}}, @IFLA_MASTER={0x8, 0xa, r3}]}, 0x3c}}, 0x0) 04:12:31 executing program 0: bpf$BPF_TASK_FD_QUERY(0x14, &(0x7f0000000100)={0xffffffffffffffff, 0xffffffffffffffff, 0x0, 0x0, 0x0}, 0x30) [ 2824.337701][ T2996] netlink: 'syz-executor.2': attribute type 1 has an invalid length. 04:12:31 executing program 4: r0 = socket$inet6(0xa, 0x2, 0x0) ioctl$sock_SIOCETHTOOL(r0, 0x89f0, &(0x7f0000000000)={'bridge0\x00', &(0x7f0000000200)=ANY=[@ANYBLOB="100000000000000000000000000000002000"/36]}) (async) ioctl$sock_SIOCETHTOOL(r0, 0x89f0, &(0x7f0000000000)={'bridge0\x00', &(0x7f0000000200)=ANY=[@ANYBLOB="100000000000000000000000000000002000"/36]}) connect$inet6(r0, &(0x7f0000000040)={0xa, 0x4e21, 0x24, @private2, 0x5}, 0x1c) 04:12:31 executing program 0: bpf$BPF_PROG_RAW_TRACEPOINT_LOAD(0x5, &(0x7f00000007c0)={0x18, 0x3, &(0x7f0000000480)=@framed={{0x18, 0x0, 0x0, 0x0, 0x0, 0x0, 0x0, 0x0, 0x4018}}, &(0x7f00000004c0)='syzkaller\x00', 0x0, 0x0, 0x0, 0x0, 0x0, '\x00', 0x0, 0x0, 0xffffffffffffffff, 0x8, 0x0, 0x0, 0x10, 0x0}, 0x90) [ 2824.488897][ T2996] bond1277: entered promiscuous mode [ 2824.508791][ T2996] 8021q: adding VLAN 0 to HW filter on device bond1277 04:12:32 executing program 0: socketpair$tipc(0x1e, 0x2, 0x0, &(0x7f0000000480)) 04:12:32 executing program 0: r0 = socket$inet6(0xa, 0x6, 0x0) bind$inet6(r0, &(0x7f0000000000)={0xa, 0x4e20}, 0x1c) listen(r0, 0x6) sendto$inet6(r0, &(0x7f0000000080)="ba34844a875f6a24ad706fcf443d535f8a21821ebd95bead9dbdf8af7b503f9a58eed1f7da8db7d8a88f61486f865aa11e0c41d37ba4b96278744da8a4708309166b0e607a18147f6ae59efed7be70a1c2127bd1d20ed49f7fcd09f0b0276408abf5", 0x62, 0x20004000, &(0x7f0000000100)={0xa, 0x4e24, 0x5, @private2={0xfc, 0x2, '\x00', 0x1}, 0x8}, 0x1c) r1 = socket$inet6(0xa, 0x6, 0x0) connect$inet6(r1, &(0x7f0000000200)={0xa, 0x0, 0x0, @initdev={0xfe, 0x88, '\x00', 0x0, 0x0}, 0x3}, 0x1c) connect$inet6(r1, &(0x7f0000000040)={0xa, 0x4e20, 0x0, @dev={0xfe, 0x80, '\x00', 0xf}}, 0x1c) sendmmsg$inet6(r1, &(0x7f0000005a00)=[{{0x0, 0x0, 0x0}}, {{0x0, 0x0, 0x0, 0x0, &(0x7f0000005bc0)=ANY=[@ANYBLOB="5000000000000000290000003600000000060000000000000401000730000000000a0000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000020000000000000002900000037000000000000000000000004010000000000001400000000000000290000000000000160000000000000002900000037"], 0x100}}], 0x2, 0x0) [ 2824.766127][ T2998] bond1277: (slave bridge1198): making interface the new active one [ 2824.793353][ T2998] bridge1198: entered promiscuous mode [ 2824.813822][ T2998] bond1277: (slave bridge1198): Enslaving as an active interface with an up link [ 2824.828953][ T3001] netlink: 'syz-executor.1': attribute type 1 has an invalid length. [ 2824.977799][ T3001] bond1294: entered promiscuous mode [ 2824.993714][ T3001] 8021q: adding VLAN 0 to HW filter on device bond1294 [ 2825.082880][ T3002] bond1294: (slave bridge1240): making interface the new active one [ 2825.090925][ T3002] bridge1240: entered promiscuous mode [ 2825.102980][ T3002] bond1294: (slave bridge1240): Enslaving as an active interface with an up link 04:12:32 executing program 3: r0 = socket$netlink(0x10, 0x3, 0x0) sendmsg$netlink(r0, &(0x7f00000047c0)={0x0, 0x0, &(0x7f0000000000)=[{&(0x7f0000000100)=ANY=[@ANYBLOB="18000000560001060000007600db730007"], 0x18}], 0x1}, 0x0) 04:12:32 executing program 0: r0 = socket$inet6(0xa, 0x6, 0x0) bind$inet6(r0, &(0x7f0000000000)={0xa, 0x4e20}, 0x1c) listen(r0, 0x6) r1 = socket$inet6(0xa, 0x6, 0x0) r2 = socket$inet6(0xa, 0x80002, 0x88) bind$inet6(r2, &(0x7f0000000580)={0xa, 0x10010000004e20}, 0x1c) setsockopt$inet6_udp_int(r2, 0x11, 0xb, &(0x7f0000000080), 0x4) r3 = accept(r0, 0x0, &(0x7f0000000100)) setsockopt$inet6_tcp_TCP_MD5SIG(r3, 0x6, 0xe, &(0x7f0000000240)={@in={{0x2, 0x4e24, @empty}}, 0x0, 0x0, 0xe, 0x0, "f7e7cc8eefaf1c185361d9fd84f2f5f5da1fc0c85899c5fa73b0962b36bcf8d02916687c145ad91a568951b903f26927ecda656f85ee50ddc5788f8f42770db312d7023cc7b4fba91b852d9bf7ad6eb8"}, 0xd8) getsockopt$inet6_IPV6_FLOWLABEL_MGR(r2, 0x29, 0x20, &(0x7f0000000080)={@private1, 0x4, 0x2, 0x2, 0xc, 0x81, 0x3}, &(0x7f00000000c0)=0x20) connect$inet6(r1, &(0x7f0000000200)={0xa, 0x0, 0x0, @initdev={0xfe, 0x88, '\x00', 0x0, 0x0}, 0x3}, 0x1c) r4 = syz_init_net_socket$ax25(0x3, 0x2, 0x0) bind$ax25(r4, &(0x7f0000000000)={{0x3, @remote={0xcc, 0xcc, 0xcc, 0xcc, 0xcc, 0xcc, 0x1}, 0x1}, [@null, @netrom={0xbb, 0xbb, 0xbb, 0xbb, 0xbb, 0x0, 0x0}, @bcast, @rose={0xbb, 0xbb, 0xbb, 0x1, 0x0}, @bcast, @netrom={0xbb, 0xbb, 0xbb, 0xbb, 0xbb, 0x0, 0x0}, @null, @bcast]}, 0x48) close(r4) connect$inet6(r1, &(0x7f0000000040)={0xa, 0x4e20, 0x0, @dev={0xfe, 0x80, '\x00', 0xd}}, 0x1c) sendmmsg$inet6(r1, &(0x7f0000005a00)=[{{0x0, 0x0, 0x0}}, {{0x0, 0x0, 0x0, 0x0, &(0x7f0000005bc0)=ANY=[@ANYBLOB="5000000000000000290000003600000000060000000000000401000730000000000a0000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000020000000000000002900000037000000000000000000000004010000000000001400000000000000290000000000000160000000000000002900000037"], 0x100}}], 0x2, 0x0) 04:12:32 executing program 2: r0 = socket$netlink(0x10, 0x3, 0x0) r1 = socket$netlink(0x10, 0x3, 0x0) r2 = socket(0x10, 0x803, 0x0) sendmsg$NL80211_CMD_CRIT_PROTOCOL_START(r2, &(0x7f0000000580)={0x0, 0x0, &(0x7f0000000540)={0x0, 0x1c}}, 0x0) getsockname$packet(r2, &(0x7f0000000600)={0x11, 0x0, 0x0, 0x1, 0x0, 0x6, @broadcast}, &(0x7f0000000100)=0x14) sendmsg$nl_route(r1, &(0x7f0000000040)={0x0, 0x0, &(0x7f0000000000)={&(0x7f00000000c0)=ANY=[@ANYBLOB="3c00000010008506eb9afc4cd8d06e754a0081c5", @ANYRES32=r3, @ANYBLOB="2377f292252155b21c0012000c000100626f6e64000000000c0002000800010001"], 0x3c}}, 0x0) sendmsg$nl_route(r0, &(0x7f0000000240)={0x0, 0x0, &(0x7f0000000680)={&(0x7f0000000400)=@newlink={0x3c, 0x10, 0xffffff1f, 0x0, 0x253c30, {}, [@IFLA_LINKINFO={0x14, 0x12, 0x0, 0x1, @bridge={{0xb}, {0x4}}}, @IFLA_MASTER={0x8, 0xa, r3}]}, 0x3c}}, 0x0) [ 2825.129885][ T3011] netlink: 'syz-executor.5': attribute type 1 has an invalid length. 04:12:32 executing program 1: r0 = socket$netlink(0x10, 0x3, 0x0) r1 = socket$netlink(0x10, 0x3, 0x0) r2 = socket(0x10, 0x803, 0x0) sendmsg$NL80211_CMD_CRIT_PROTOCOL_START(r2, &(0x7f0000000580)={0x0, 0x0, &(0x7f0000000540)={0x0, 0x1c}}, 0x0) getsockname$packet(r2, &(0x7f0000000600)={0x11, 0x0, 0x0, 0x1, 0x0, 0x6, @broadcast}, &(0x7f0000000100)=0x14) sendmsg$nl_route(r1, &(0x7f0000000040)={0x0, 0x0, &(0x7f0000000000)={&(0x7f00000000c0)=ANY=[@ANYBLOB="3c00000010008506eb9afc4cd8d06e754a0081c5", @ANYRES32=r3, @ANYBLOB="2377f292252155b21c0012000c000100626f6e64000000000c0002000800010001"], 0x3c}}, 0x0) sendmsg$nl_route(r0, &(0x7f0000000240)={0x0, 0x0, &(0x7f0000000680)={&(0x7f0000000400)=@newlink={0x3c, 0x10, 0xffffff1f, 0x0, 0x12000000, {}, [@IFLA_LINKINFO={0x14, 0x12, 0x0, 0x1, @bridge={{0xb}, {0x4}}}, @IFLA_MASTER={0x8, 0xa, r3}]}, 0x3c}}, 0x0) [ 2825.294874][ T3011] bond1286: entered promiscuous mode [ 2825.300751][ T3011] 8021q: adding VLAN 0 to HW filter on device bond1286 04:12:32 executing program 4: r0 = socket$inet6(0xa, 0x2, 0x0) ioctl$sock_SIOCETHTOOL(r0, 0x89f0, &(0x7f0000000000)={'bridge0\x00', &(0x7f0000000200)=ANY=[@ANYBLOB="100000000000000000000000000000002000"/36]}) (async) connect$inet6(r0, &(0x7f0000000040)={0xa, 0x4e21, 0x24, @private2, 0x5}, 0x1c) [ 2825.490028][ T3016] bond1286: (slave bridge1249): making interface the new active one [ 2825.511229][ T3016] bridge1249: entered promiscuous mode [ 2825.537333][ T3016] bond1286: (slave bridge1249): Enslaving as an active interface with an up link 04:12:33 executing program 5: r0 = socket$netlink(0x10, 0x3, 0x0) r1 = socket$netlink(0x10, 0x3, 0x0) r2 = socket(0x10, 0x803, 0x0) sendmsg$NL80211_CMD_CRIT_PROTOCOL_START(r2, &(0x7f0000000580)={0x0, 0x0, &(0x7f0000000540)={0x0, 0x1c}}, 0x0) getsockname$packet(r2, &(0x7f0000000600)={0x11, 0x0, 0x0, 0x1, 0x0, 0x6, @broadcast}, &(0x7f0000000100)=0x14) sendmsg$nl_route(r1, &(0x7f0000000040)={0x0, 0x0, &(0x7f0000000000)={&(0x7f00000000c0)=ANY=[@ANYBLOB="3c00000010008506eb9afc4cd8d06e754a0081c5", @ANYRES32=r3, @ANYBLOB="2377f292252155b21c0012000c000100626f6e64000000000c0002000800010001"], 0x3c}}, 0x0) sendmsg$nl_route(r0, &(0x7f0000000240)={0x0, 0x0, &(0x7f0000000680)={&(0x7f0000000400)=@newlink={0x3c, 0x10, 0xffffff1f, 0x0, 0x2030000, {}, [@IFLA_LINKINFO={0x14, 0x12, 0x0, 0x1, @bridge={{0xb}, {0x4}}}, @IFLA_MASTER={0x8, 0xa, r3}]}, 0x3c}}, 0x0) [ 2825.586781][ T3029] netlink: 'syz-executor.2': attribute type 1 has an invalid length. [ 2825.688992][ T3029] bond1278: entered promiscuous mode [ 2825.696543][ T3029] 8021q: adding VLAN 0 to HW filter on device bond1278 [ 2825.805422][ T3036] bond1278: (slave bridge1199): making interface the new active one [ 2825.825129][ T3036] bridge1199: entered promiscuous mode [ 2825.850373][ T3036] bond1278: (slave bridge1199): Enslaving as an active interface with an up link [ 2825.878029][ T3039] netlink: 'syz-executor.1': attribute type 1 has an invalid length. 04:12:33 executing program 2: r0 = socket$netlink(0x10, 0x3, 0x0) r1 = socket$netlink(0x10, 0x3, 0x0) r2 = socket(0x10, 0x803, 0x0) sendmsg$NL80211_CMD_CRIT_PROTOCOL_START(r2, &(0x7f0000000580)={0x0, 0x0, &(0x7f0000000540)={0x0, 0x1c}}, 0x0) getsockname$packet(r2, &(0x7f0000000600)={0x11, 0x0, 0x0, 0x1, 0x0, 0x6, @broadcast}, &(0x7f0000000100)=0x14) sendmsg$nl_route(r1, &(0x7f0000000040)={0x0, 0x0, &(0x7f0000000000)={&(0x7f00000000c0)=ANY=[@ANYBLOB="3c00000010008506eb9afc4cd8d06e754a0081c5", @ANYRES32=r3, @ANYBLOB="2377f292252155b21c0012000c000100626f6e64000000000c0002000800010001"], 0x3c}}, 0x0) sendmsg$nl_route(r0, &(0x7f0000000240)={0x0, 0x0, &(0x7f0000000680)={&(0x7f0000000400)=@newlink={0x3c, 0x10, 0xffffff1f, 0x0, 0x253c31, {}, [@IFLA_LINKINFO={0x14, 0x12, 0x0, 0x1, @bridge={{0xb}, {0x4}}}, @IFLA_MASTER={0x8, 0xa, r3}]}, 0x3c}}, 0x0) [ 2825.928903][ T3039] bond1295: entered promiscuous mode [ 2825.936755][ T3039] 8021q: adding VLAN 0 to HW filter on device bond1295 04:12:33 executing program 0: r0 = socket$inet6(0xa, 0x6, 0x0) bind$inet6(r0, &(0x7f0000000000)={0xa, 0x4e20}, 0x1c) listen(r0, 0x6) r1 = socket$inet6(0xa, 0x6, 0x0) r2 = socket$inet6(0xa, 0x80002, 0x88) bind$inet6(r2, &(0x7f0000000580)={0xa, 0x10010000004e20}, 0x1c) setsockopt$inet6_udp_int(r2, 0x11, 0xb, &(0x7f0000000080), 0x4) r3 = accept(r0, 0x0, &(0x7f0000000100)) setsockopt$inet6_tcp_TCP_MD5SIG(r3, 0x6, 0xe, &(0x7f0000000240)={@in={{0x2, 0x4e24, @empty}}, 0x0, 0x0, 0xe, 0x0, "f7e7cc8eefaf1c185361d9fd84f2f5f5da1fc0c85899c5fa73b0962b36bcf8d02916687c145ad91a568951b903f26927ecda656f85ee50ddc5788f8f42770db312d7023cc7b4fba91b852d9bf7ad6eb8"}, 0xd8) getsockopt$inet6_IPV6_FLOWLABEL_MGR(r2, 0x29, 0x20, &(0x7f0000000080)={@private1, 0x4, 0x2, 0x2, 0xc, 0x81, 0x3}, &(0x7f00000000c0)=0x20) connect$inet6(r1, &(0x7f0000000200)={0xa, 0x0, 0x0, @initdev={0xfe, 0x88, '\x00', 0x0, 0x0}, 0x3}, 0x1c) r4 = syz_init_net_socket$ax25(0x3, 0x2, 0x0) bind$ax25(r4, &(0x7f0000000000)={{0x3, @remote={0xcc, 0xcc, 0xcc, 0xcc, 0xcc, 0xcc, 0x1}, 0x1}, [@null, @netrom={0xbb, 0xbb, 0xbb, 0xbb, 0xbb, 0x0, 0x0}, @bcast, @rose={0xbb, 0xbb, 0xbb, 0x1, 0x0}, @bcast, @netrom={0xbb, 0xbb, 0xbb, 0xbb, 0xbb, 0x0, 0x0}, @null, @bcast]}, 0x48) close(r4) connect$inet6(r1, &(0x7f0000000040)={0xa, 0x4e20, 0x0, @dev={0xfe, 0x80, '\x00', 0xd}}, 0x1c) sendmmsg$inet6(r1, &(0x7f0000005a00)=[{{0x0, 0x0, 0x0}}, {{0x0, 0x0, 0x0, 0x0, &(0x7f0000005bc0)=ANY=[@ANYBLOB="5000000000000000290000003600000000060000000000000401000730000000000a0000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000020000000000000002900000037000000000000000000000004010000000000001400000000000000290000000000000160000000000000002900000037"], 0x100}}], 0x2, 0x0) [ 2826.195140][ T3040] bond1295: (slave bridge1241): making interface the new active one [ 2826.205456][ T3040] bridge1241: entered promiscuous mode [ 2826.220441][ T3040] bond1295: (slave bridge1241): Enslaving as an active interface with an up link 04:12:33 executing program 4: ioctl$F2FS_IOC_MOVE_RANGE(0xffffffffffffffff, 0xc020f509, &(0x7f0000000040)={0xffffffffffffffff, 0x2, 0x1f, 0x77}) ioctl$ifreq_SIOCGIFINDEX_batadv_mesh(0xffffffffffffffff, 0x8933, &(0x7f00000000c0)={'batadv0\x00', 0x0}) r2 = socket$inet_udplite(0x2, 0x2, 0x88) ioctl$sock_SIOCGIFINDEX(r2, 0x8933, &(0x7f0000000340)={'bridge_slave_0\x00', 0x0}) r4 = socket(0x10, 0x80002, 0x0) sendmsg$nl_route(r4, &(0x7f0000000000)={0x0, 0x0, &(0x7f0000000080)={&(0x7f00000000c0)=@bridge_setlink={0x2c, 0x13, 0xa29, 0x0, 0x0, {0x7, 0x0, 0x0, r3}, [@IFLA_AF_SPEC={0xc, 0x1a, 0x0, 0x1, [@AF_INET={0x8, 0x4, 0x0, 0x1, {0x4, 0x3}}]}]}, 0x2c}}, 0x0) r5 = syz_genetlink_get_family_id$batadv(&(0x7f00000001c0), r0) sendmsg$BATADV_CMD_GET_GATEWAYS(0xffffffffffffffff, &(0x7f00000004c0)={&(0x7f0000000180)={0x10, 0x0, 0x0, 0x10}, 0xc, &(0x7f0000000480)={&(0x7f0000000440)={0x1c, r5, 0x200, 0x70bd28, 0x25dfdbfe, {}, [@BATADV_ATTR_TPMETER_TEST_TIME={0x8, 0xb, 0x2a63}]}, 0x1c}, 0x1, 0x0, 0x0, 0x4080}, 0x80) socket$inet_udplite(0x2, 0x2, 0x88) r6 = socket(0x10, 0x80002, 0x0) sendmsg$nl_route(r6, &(0x7f0000000000)={0x0, 0x0, &(0x7f0000000080)={&(0x7f00000000c0)=@bridge_setlink={0x2c, 0x13, 0xa29, 0x0, 0x0, {}, [@IFLA_AF_SPEC={0xc, 0x1a, 0x0, 0x1, [@AF_INET={0x8, 0x4, 0x0, 0x1, {0x4, 0x3}}]}]}, 0x2c}}, 0x0) sendmsg$nl_route(r0, &(0x7f0000000140)={&(0x7f0000000080)={0x10, 0x0, 0x0, 0x20000}, 0xc, &(0x7f0000000100)={&(0x7f0000000240)=@ipv6_newroute={0x1e0, 0x18, 0x500, 0x70bd2c, 0x25dfdbff, {0xa, 0x14, 0x10, 0x0, 0xd341d5271abc844a, 0x0, 0xc8, 0x2, 0x700}, [@RTA_OIF={0x8, 0x4, r1}, @RTA_METRICS={0x88, 0x8, 0x0, 0x1, "dc4ae466bdee725324a2ed7dbc6a9825c8bbe19a9ce529fa4f55bf462317b21095af79853d356138f729d8aef20c8bea26a288f858ecf5f1854fc514debd2367c32a52319397fece2f4a693fdd33696eb7bccaaf14fe64c150b199e954c7dc367227e11b1fb7192cec9bbdd46ee96361ba299654636f60a70b45e9e6b5fa989050de0460"}, @RTA_OIF={0x8, 0x4, r3}, @RTA_GATEWAY={0x14, 0x5, @initdev={0xfe, 0x88, '\x00', 0x0, 0x0}}, @RTA_OIF={0x8}, @RTA_METRICS={0xf8, 0x8, 0x0, 0x1, "3f892218677b04daae345aaa66e2f1f71c4d348ea8427aad818baaeb00ccafbbcae7e204a4e0dbfffe5b1748662645f9dfad1c448841b64cc9b5e408d18c0aa2d9c9f3aba140670dda320d8845466bf148e8ea3a2fc13037db3b9b2768239b60ea1692e4c41df5b00ac0b2aaf0214ac6da9924042ac2632edbf04c0d2dbbfd923193e6fb575fd2a4fdda04a26047f872d6415a0fbf4e73f132dd900a7d72f5b6327cc986340dbe8039abcc307de251deec9c264872363819e8fdd38608f65aa552841e8f0494c8243b3e19f2b7ff1fa8ca3be0660db0730b5baacfc829b6400dc444742ff43e045a7e6d803ca9e8bb8e6036aec0"}, @RTA_PREF={0x5, 0x14, 0x5}, @RTA_PRIORITY={0x8, 0x6, 0x7}, @RTA_MARK={0x8, 0x10, 0xfffff123}]}, 0x1e0}, 0x1, 0x0, 0x0, 0x40011}, 0x20010080) r7 = socket$inet6(0xa, 0x2, 0x0) ioctl$sock_SIOCETHTOOL(r7, 0x89f0, &(0x7f0000000000)={'bridge0\x00', &(0x7f0000000200)=@ethtool_ringparam={0x10}}) 04:12:33 executing program 1: r0 = socket$netlink(0x10, 0x3, 0x0) r1 = socket$netlink(0x10, 0x3, 0x0) r2 = socket(0x10, 0x803, 0x0) sendmsg$NL80211_CMD_CRIT_PROTOCOL_START(r2, &(0x7f0000000580)={0x0, 0x0, &(0x7f0000000540)={0x0, 0x1c}}, 0x0) getsockname$packet(r2, &(0x7f0000000600)={0x11, 0x0, 0x0, 0x1, 0x0, 0x6, @broadcast}, &(0x7f0000000100)=0x14) sendmsg$nl_route(r1, &(0x7f0000000040)={0x0, 0x0, &(0x7f0000000000)={&(0x7f00000000c0)=ANY=[@ANYBLOB="3c00000010008506eb9afc4cd8d06e754a0081c5", @ANYRES32=r3, @ANYBLOB="2377f292252155b21c0012000c000100626f6e64000000000c0002000800010001"], 0x3c}}, 0x0) sendmsg$nl_route(r0, &(0x7f0000000240)={0x0, 0x0, &(0x7f0000000680)={&(0x7f0000000400)=@newlink={0x3c, 0x10, 0xffffff1f, 0x0, 0x12672500, {}, [@IFLA_LINKINFO={0x14, 0x12, 0x0, 0x1, @bridge={{0xb}, {0x4}}}, @IFLA_MASTER={0x8, 0xa, r3}]}, 0x3c}}, 0x0) [ 2826.421396][ T3047] bond1287: entered promiscuous mode [ 2826.427210][ T3047] 8021q: adding VLAN 0 to HW filter on device bond1287 04:12:34 executing program 3: r0 = socket$netlink(0x10, 0x3, 0x0) sendmsg$netlink(r0, &(0x7f00000047c0)={0x0, 0x0, &(0x7f0000000000)=[{&(0x7f0000000100)=ANY=[@ANYBLOB="18000000560001060000007700db730007"], 0x18}], 0x1}, 0x0) [ 2826.551801][ T3048] bond1287: (slave bridge1250): making interface the new active one [ 2826.560057][ T3048] bridge1250: entered promiscuous mode [ 2826.580638][ T3048] bond1287: (slave bridge1250): Enslaving as an active interface with an up link 04:12:34 executing program 5: r0 = socket$netlink(0x10, 0x3, 0x0) r1 = socket$netlink(0x10, 0x3, 0x0) r2 = socket(0x10, 0x803, 0x0) sendmsg$NL80211_CMD_CRIT_PROTOCOL_START(r2, &(0x7f0000000580)={0x0, 0x0, &(0x7f0000000540)={0x0, 0x1c}}, 0x0) getsockname$packet(r2, &(0x7f0000000600)={0x11, 0x0, 0x0, 0x1, 0x0, 0x6, @broadcast}, &(0x7f0000000100)=0x14) sendmsg$nl_route(r1, &(0x7f0000000040)={0x0, 0x0, &(0x7f0000000000)={&(0x7f00000000c0)=ANY=[@ANYBLOB="3c00000010008506eb9afc4cd8d06e754a0081c5", @ANYRES32=r3, @ANYBLOB="2377f292252155b21c0012000c000100626f6e64000000000c0002000800010001"], 0x3c}}, 0x0) sendmsg$nl_route(r0, &(0x7f0000000240)={0x0, 0x0, &(0x7f0000000680)={&(0x7f0000000400)=@newlink={0x3c, 0x10, 0xffffff1f, 0x0, 0x3000000, {}, [@IFLA_LINKINFO={0x14, 0x12, 0x0, 0x1, @bridge={{0xb}, {0x4}}}, @IFLA_MASTER={0x8, 0xa, r3}]}, 0x3c}}, 0x0) [ 2826.846457][ T3054] bond1279: entered promiscuous mode [ 2826.852386][ T3054] 8021q: adding VLAN 0 to HW filter on device bond1279 [ 2826.925914][ T3057] bond1279: (slave bridge1200): making interface the new active one [ 2826.934211][ T3057] bridge1200: entered promiscuous mode [ 2826.946016][ T3057] bond1279: (slave bridge1200): Enslaving as an active interface with an up link 04:12:34 executing program 0: r0 = socket$inet6(0xa, 0x6, 0x0) bind$inet6(r0, &(0x7f0000000000)={0xa, 0x4e20}, 0x1c) listen(r0, 0x6) r1 = socket$inet6(0xa, 0x6, 0x0) r2 = socket$inet6(0xa, 0x80002, 0x88) bind$inet6(r2, &(0x7f0000000580)={0xa, 0x10010000004e20}, 0x1c) setsockopt$inet6_udp_int(r2, 0x11, 0xb, &(0x7f0000000080), 0x4) r3 = accept(r0, 0x0, &(0x7f0000000100)) setsockopt$inet6_tcp_TCP_MD5SIG(r3, 0x6, 0xe, &(0x7f0000000240)={@in={{0x2, 0x4e24, @empty}}, 0x0, 0x0, 0xe, 0x0, "f7e7cc8eefaf1c185361d9fd84f2f5f5da1fc0c85899c5fa73b0962b36bcf8d02916687c145ad91a568951b903f26927ecda656f85ee50ddc5788f8f42770db312d7023cc7b4fba91b852d9bf7ad6eb8"}, 0xd8) getsockopt$inet6_IPV6_FLOWLABEL_MGR(r2, 0x29, 0x20, &(0x7f0000000080)={@private1, 0x4, 0x2, 0x2, 0xc, 0x81, 0x3}, &(0x7f00000000c0)=0x20) connect$inet6(r1, &(0x7f0000000200)={0xa, 0x0, 0x0, @initdev={0xfe, 0x88, '\x00', 0x0, 0x0}, 0x3}, 0x1c) r4 = syz_init_net_socket$ax25(0x3, 0x2, 0x0) bind$ax25(r4, &(0x7f0000000000)={{0x3, @remote={0xcc, 0xcc, 0xcc, 0xcc, 0xcc, 0xcc, 0x1}, 0x1}, [@null, @netrom={0xbb, 0xbb, 0xbb, 0xbb, 0xbb, 0x0, 0x0}, @bcast, @rose={0xbb, 0xbb, 0xbb, 0x1, 0x0}, @bcast, @netrom={0xbb, 0xbb, 0xbb, 0xbb, 0xbb, 0x0, 0x0}, @null, @bcast]}, 0x48) close(r4) connect$inet6(r1, &(0x7f0000000040)={0xa, 0x4e20, 0x0, @dev={0xfe, 0x80, '\x00', 0xd}}, 0x1c) sendmmsg$inet6(r1, &(0x7f0000005a00)=[{{0x0, 0x0, 0x0}}, {{0x0, 0x0, 0x0, 0x0, &(0x7f0000005bc0)=ANY=[@ANYBLOB="5000000000000000290000003600000000060000000000000401000730000000000a0000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000020000000000000002900000037000000000000000000000004010000000000001400000000000000290000000000000160000000000000002900000037"], 0x100}}], 0x2, 0x0) 04:12:34 executing program 2: r0 = socket$netlink(0x10, 0x3, 0x0) r1 = socket$netlink(0x10, 0x3, 0x0) r2 = socket(0x10, 0x803, 0x0) sendmsg$NL80211_CMD_CRIT_PROTOCOL_START(r2, &(0x7f0000000580)={0x0, 0x0, &(0x7f0000000540)={0x0, 0x1c}}, 0x0) getsockname$packet(r2, &(0x7f0000000600)={0x11, 0x0, 0x0, 0x1, 0x0, 0x6, @broadcast}, &(0x7f0000000100)=0x14) sendmsg$nl_route(r1, &(0x7f0000000040)={0x0, 0x0, &(0x7f0000000000)={&(0x7f00000000c0)=ANY=[@ANYBLOB="3c00000010008506eb9afc4cd8d06e754a0081c5", @ANYRES32=r3, @ANYBLOB="2377f292252155b21c0012000c000100626f6e64000000000c0002000800010001"], 0x3c}}, 0x0) sendmsg$nl_route(r0, &(0x7f0000000240)={0x0, 0x0, &(0x7f0000000680)={&(0x7f0000000400)=@newlink={0x3c, 0x10, 0xffffff1f, 0x0, 0x253c32, {}, [@IFLA_LINKINFO={0x14, 0x12, 0x0, 0x1, @bridge={{0xb}, {0x4}}}, @IFLA_MASTER={0x8, 0xa, r3}]}, 0x3c}}, 0x0) [ 2827.125064][ T3062] bond1296: entered promiscuous mode [ 2827.131403][ T3062] 8021q: adding VLAN 0 to HW filter on device bond1296 [ 2827.237160][ T3064] bond1296: (slave bridge1242): making interface the new active one [ 2827.258717][ T3064] bridge1242: entered promiscuous mode [ 2827.279480][ T3064] bond1296: (slave bridge1242): Enslaving as an active interface with an up link 04:12:34 executing program 4: ioctl$F2FS_IOC_MOVE_RANGE(0xffffffffffffffff, 0xc020f509, &(0x7f0000000040)={0xffffffffffffffff, 0x2, 0x1f, 0x77}) ioctl$ifreq_SIOCGIFINDEX_batadv_mesh(0xffffffffffffffff, 0x8933, &(0x7f00000000c0)={'batadv0\x00', 0x0}) r2 = socket$inet_udplite(0x2, 0x2, 0x88) ioctl$sock_SIOCGIFINDEX(r2, 0x8933, &(0x7f0000000340)={'bridge_slave_0\x00', 0x0}) r4 = socket(0x10, 0x80002, 0x0) sendmsg$nl_route(r4, &(0x7f0000000000)={0x0, 0x0, &(0x7f0000000080)={&(0x7f00000000c0)=@bridge_setlink={0x2c, 0x13, 0xa29, 0x0, 0x0, {0x7, 0x0, 0x0, r3}, [@IFLA_AF_SPEC={0xc, 0x1a, 0x0, 0x1, [@AF_INET={0x8, 0x4, 0x0, 0x1, {0x4, 0x3}}]}]}, 0x2c}}, 0x0) syz_genetlink_get_family_id$batadv(&(0x7f00000001c0), r0) (async) r5 = syz_genetlink_get_family_id$batadv(&(0x7f00000001c0), r0) sendmsg$BATADV_CMD_GET_GATEWAYS(0xffffffffffffffff, &(0x7f00000004c0)={&(0x7f0000000180)={0x10, 0x0, 0x0, 0x10}, 0xc, &(0x7f0000000480)={&(0x7f0000000440)={0x1c, r5, 0x200, 0x70bd28, 0x25dfdbfe, {}, [@BATADV_ATTR_TPMETER_TEST_TIME={0x8, 0xb, 0x2a63}]}, 0x1c}, 0x1, 0x0, 0x0, 0x4080}, 0x80) (async) sendmsg$BATADV_CMD_GET_GATEWAYS(0xffffffffffffffff, &(0x7f00000004c0)={&(0x7f0000000180)={0x10, 0x0, 0x0, 0x10}, 0xc, &(0x7f0000000480)={&(0x7f0000000440)={0x1c, r5, 0x200, 0x70bd28, 0x25dfdbfe, {}, [@BATADV_ATTR_TPMETER_TEST_TIME={0x8, 0xb, 0x2a63}]}, 0x1c}, 0x1, 0x0, 0x0, 0x4080}, 0x80) socket$inet_udplite(0x2, 0x2, 0x88) (async) socket$inet_udplite(0x2, 0x2, 0x88) r6 = socket(0x10, 0x80002, 0x0) sendmsg$nl_route(r6, &(0x7f0000000000)={0x0, 0x0, &(0x7f0000000080)={&(0x7f00000000c0)=@bridge_setlink={0x2c, 0x13, 0xa29, 0x0, 0x0, {}, [@IFLA_AF_SPEC={0xc, 0x1a, 0x0, 0x1, [@AF_INET={0x8, 0x4, 0x0, 0x1, {0x4, 0x3}}]}]}, 0x2c}}, 0x0) sendmsg$nl_route(r0, &(0x7f0000000140)={&(0x7f0000000080)={0x10, 0x0, 0x0, 0x20000}, 0xc, &(0x7f0000000100)={&(0x7f0000000240)=@ipv6_newroute={0x1e0, 0x18, 0x500, 0x70bd2c, 0x25dfdbff, {0xa, 0x14, 0x10, 0x0, 0xd341d5271abc844a, 0x0, 0xc8, 0x2, 0x700}, [@RTA_OIF={0x8, 0x4, r1}, @RTA_METRICS={0x88, 0x8, 0x0, 0x1, "dc4ae466bdee725324a2ed7dbc6a9825c8bbe19a9ce529fa4f55bf462317b21095af79853d356138f729d8aef20c8bea26a288f858ecf5f1854fc514debd2367c32a52319397fece2f4a693fdd33696eb7bccaaf14fe64c150b199e954c7dc367227e11b1fb7192cec9bbdd46ee96361ba299654636f60a70b45e9e6b5fa989050de0460"}, @RTA_OIF={0x8, 0x4, r3}, @RTA_GATEWAY={0x14, 0x5, @initdev={0xfe, 0x88, '\x00', 0x0, 0x0}}, @RTA_OIF={0x8}, @RTA_METRICS={0xf8, 0x8, 0x0, 0x1, "3f892218677b04daae345aaa66e2f1f71c4d348ea8427aad818baaeb00ccafbbcae7e204a4e0dbfffe5b1748662645f9dfad1c448841b64cc9b5e408d18c0aa2d9c9f3aba140670dda320d8845466bf148e8ea3a2fc13037db3b9b2768239b60ea1692e4c41df5b00ac0b2aaf0214ac6da9924042ac2632edbf04c0d2dbbfd923193e6fb575fd2a4fdda04a26047f872d6415a0fbf4e73f132dd900a7d72f5b6327cc986340dbe8039abcc307de251deec9c264872363819e8fdd38608f65aa552841e8f0494c8243b3e19f2b7ff1fa8ca3be0660db0730b5baacfc829b6400dc444742ff43e045a7e6d803ca9e8bb8e6036aec0"}, @RTA_PREF={0x5, 0x14, 0x5}, @RTA_PRIORITY={0x8, 0x6, 0x7}, @RTA_MARK={0x8, 0x10, 0xfffff123}]}, 0x1e0}, 0x1, 0x0, 0x0, 0x40011}, 0x20010080) r7 = socket$inet6(0xa, 0x2, 0x0) ioctl$sock_SIOCETHTOOL(r7, 0x89f0, &(0x7f0000000000)={'bridge0\x00', &(0x7f0000000200)=@ethtool_ringparam={0x10}}) (async) ioctl$sock_SIOCETHTOOL(r7, 0x89f0, &(0x7f0000000000)={'bridge0\x00', &(0x7f0000000200)=@ethtool_ringparam={0x10}}) 04:12:34 executing program 1: r0 = socket$netlink(0x10, 0x3, 0x0) r1 = socket$netlink(0x10, 0x3, 0x0) r2 = socket(0x10, 0x803, 0x0) sendmsg$NL80211_CMD_CRIT_PROTOCOL_START(r2, &(0x7f0000000580)={0x0, 0x0, &(0x7f0000000540)={0x0, 0x1c}}, 0x0) getsockname$packet(r2, &(0x7f0000000600)={0x11, 0x0, 0x0, 0x1, 0x0, 0x6, @broadcast}, &(0x7f0000000100)=0x14) sendmsg$nl_route(r1, &(0x7f0000000040)={0x0, 0x0, &(0x7f0000000000)={&(0x7f00000000c0)=ANY=[@ANYBLOB="3c00000010008506eb9afc4cd8d06e754a0081c5", @ANYRES32=r3, @ANYBLOB="2377f292252155b21c0012000c000100626f6e64000000000c0002000800010001"], 0x3c}}, 0x0) sendmsg$nl_route(r0, &(0x7f0000000240)={0x0, 0x0, &(0x7f0000000680)={&(0x7f0000000400)=@newlink={0x3c, 0x10, 0xffffff1f, 0x0, 0x13672500, {}, [@IFLA_LINKINFO={0x14, 0x12, 0x0, 0x1, @bridge={{0xb}, {0x4}}}, @IFLA_MASTER={0x8, 0xa, r3}]}, 0x3c}}, 0x0) 04:12:34 executing program 3: r0 = socket$netlink(0x10, 0x3, 0x0) sendmsg$netlink(r0, &(0x7f00000047c0)={0x0, 0x0, &(0x7f0000000000)=[{&(0x7f0000000100)=ANY=[@ANYBLOB="18000000560001060000007800db730007"], 0x18}], 0x1}, 0x0) [ 2827.433587][ T3069] bond1288: entered promiscuous mode [ 2827.439270][ T3069] 8021q: adding VLAN 0 to HW filter on device bond1288 04:12:35 executing program 4: ioctl$F2FS_IOC_MOVE_RANGE(0xffffffffffffffff, 0xc020f509, &(0x7f0000000040)={0xffffffffffffffff, 0x2, 0x1f, 0x77}) (async) ioctl$ifreq_SIOCGIFINDEX_batadv_mesh(0xffffffffffffffff, 0x8933, &(0x7f00000000c0)={'batadv0\x00', 0x0}) r2 = socket$inet_udplite(0x2, 0x2, 0x88) ioctl$sock_SIOCGIFINDEX(r2, 0x8933, &(0x7f0000000340)={'bridge_slave_0\x00', 0x0}) (async) r4 = socket(0x10, 0x80002, 0x0) sendmsg$nl_route(r4, &(0x7f0000000000)={0x0, 0x0, &(0x7f0000000080)={&(0x7f00000000c0)=@bridge_setlink={0x2c, 0x13, 0xa29, 0x0, 0x0, {0x7, 0x0, 0x0, r3}, [@IFLA_AF_SPEC={0xc, 0x1a, 0x0, 0x1, [@AF_INET={0x8, 0x4, 0x0, 0x1, {0x4, 0x3}}]}]}, 0x2c}}, 0x0) r5 = syz_genetlink_get_family_id$batadv(&(0x7f00000001c0), r0) sendmsg$BATADV_CMD_GET_GATEWAYS(0xffffffffffffffff, &(0x7f00000004c0)={&(0x7f0000000180)={0x10, 0x0, 0x0, 0x10}, 0xc, &(0x7f0000000480)={&(0x7f0000000440)={0x1c, r5, 0x200, 0x70bd28, 0x25dfdbfe, {}, [@BATADV_ATTR_TPMETER_TEST_TIME={0x8, 0xb, 0x2a63}]}, 0x1c}, 0x1, 0x0, 0x0, 0x4080}, 0x80) socket$inet_udplite(0x2, 0x2, 0x88) (async) r6 = socket(0x10, 0x80002, 0x0) sendmsg$nl_route(r6, &(0x7f0000000000)={0x0, 0x0, &(0x7f0000000080)={&(0x7f00000000c0)=@bridge_setlink={0x2c, 0x13, 0xa29, 0x0, 0x0, {}, [@IFLA_AF_SPEC={0xc, 0x1a, 0x0, 0x1, [@AF_INET={0x8, 0x4, 0x0, 0x1, {0x4, 0x3}}]}]}, 0x2c}}, 0x0) sendmsg$nl_route(r0, &(0x7f0000000140)={&(0x7f0000000080)={0x10, 0x0, 0x0, 0x20000}, 0xc, &(0x7f0000000100)={&(0x7f0000000240)=@ipv6_newroute={0x1e0, 0x18, 0x500, 0x70bd2c, 0x25dfdbff, {0xa, 0x14, 0x10, 0x0, 0xd341d5271abc844a, 0x0, 0xc8, 0x2, 0x700}, [@RTA_OIF={0x8, 0x4, r1}, @RTA_METRICS={0x88, 0x8, 0x0, 0x1, "dc4ae466bdee725324a2ed7dbc6a9825c8bbe19a9ce529fa4f55bf462317b21095af79853d356138f729d8aef20c8bea26a288f858ecf5f1854fc514debd2367c32a52319397fece2f4a693fdd33696eb7bccaaf14fe64c150b199e954c7dc367227e11b1fb7192cec9bbdd46ee96361ba299654636f60a70b45e9e6b5fa989050de0460"}, @RTA_OIF={0x8, 0x4, r3}, @RTA_GATEWAY={0x14, 0x5, @initdev={0xfe, 0x88, '\x00', 0x0, 0x0}}, @RTA_OIF={0x8}, @RTA_METRICS={0xf8, 0x8, 0x0, 0x1, "3f892218677b04daae345aaa66e2f1f71c4d348ea8427aad818baaeb00ccafbbcae7e204a4e0dbfffe5b1748662645f9dfad1c448841b64cc9b5e408d18c0aa2d9c9f3aba140670dda320d8845466bf148e8ea3a2fc13037db3b9b2768239b60ea1692e4c41df5b00ac0b2aaf0214ac6da9924042ac2632edbf04c0d2dbbfd923193e6fb575fd2a4fdda04a26047f872d6415a0fbf4e73f132dd900a7d72f5b6327cc986340dbe8039abcc307de251deec9c264872363819e8fdd38608f65aa552841e8f0494c8243b3e19f2b7ff1fa8ca3be0660db0730b5baacfc829b6400dc444742ff43e045a7e6d803ca9e8bb8e6036aec0"}, @RTA_PREF={0x5, 0x14, 0x5}, @RTA_PRIORITY={0x8, 0x6, 0x7}, @RTA_MARK={0x8, 0x10, 0xfffff123}]}, 0x1e0}, 0x1, 0x0, 0x0, 0x40011}, 0x20010080) (async) r7 = socket$inet6(0xa, 0x2, 0x0) ioctl$sock_SIOCETHTOOL(r7, 0x89f0, &(0x7f0000000000)={'bridge0\x00', &(0x7f0000000200)=@ethtool_ringparam={0x10}}) [ 2827.537184][ T3073] bond1288: (slave bridge1251): making interface the new active one [ 2827.569984][ T3073] bridge1251: entered promiscuous mode [ 2827.609493][ T3073] bond1288: (slave bridge1251): Enslaving as an active interface with an up link 04:12:35 executing program 5: r0 = socket$netlink(0x10, 0x3, 0x0) r1 = socket$netlink(0x10, 0x3, 0x0) r2 = socket(0x10, 0x803, 0x0) sendmsg$NL80211_CMD_CRIT_PROTOCOL_START(r2, &(0x7f0000000580)={0x0, 0x0, &(0x7f0000000540)={0x0, 0x1c}}, 0x0) getsockname$packet(r2, &(0x7f0000000600)={0x11, 0x0, 0x0, 0x1, 0x0, 0x6, @broadcast}, &(0x7f0000000100)=0x14) sendmsg$nl_route(r1, &(0x7f0000000040)={0x0, 0x0, &(0x7f0000000000)={&(0x7f00000000c0)=ANY=[@ANYBLOB="3c00000010008506eb9afc4cd8d06e754a0081c5", @ANYRES32=r3, @ANYBLOB="2377f292252155b21c0012000c000100626f6e64000000000c0002000800010001"], 0x3c}}, 0x0) sendmsg$nl_route(r0, &(0x7f0000000240)={0x0, 0x0, &(0x7f0000000680)={&(0x7f0000000400)=@newlink={0x3c, 0x10, 0xffffff1f, 0x0, 0x4000000, {}, [@IFLA_LINKINFO={0x14, 0x12, 0x0, 0x1, @bridge={{0xb}, {0x4}}}, @IFLA_MASTER={0x8, 0xa, r3}]}, 0x3c}}, 0x0) [ 2827.741123][ T3077] validate_nla: 4 callbacks suppressed [ 2827.741145][ T3077] netlink: 'syz-executor.2': attribute type 1 has an invalid length. [ 2827.811132][ T3077] bond1280: entered promiscuous mode [ 2827.816994][ T3077] 8021q: adding VLAN 0 to HW filter on device bond1280 04:12:35 executing program 0: r0 = socket$inet6(0xa, 0x6, 0x0) bind$inet6(r0, &(0x7f0000000000)={0xa, 0x4e20}, 0x1c) listen(r0, 0x6) r1 = socket$inet6(0xa, 0x6, 0x0) r2 = socket$inet6(0xa, 0x80002, 0x88) bind$inet6(r2, &(0x7f0000000580)={0xa, 0x10010000004e20}, 0x1c) setsockopt$inet6_udp_int(r2, 0x11, 0xb, &(0x7f0000000080), 0x4) r3 = accept(r0, 0x0, &(0x7f0000000100)) setsockopt$inet6_tcp_TCP_MD5SIG(r3, 0x6, 0xe, &(0x7f0000000240)={@in={{0x2, 0x4e24, @empty}}, 0x0, 0x0, 0xe, 0x0, "f7e7cc8eefaf1c185361d9fd84f2f5f5da1fc0c85899c5fa73b0962b36bcf8d02916687c145ad91a568951b903f26927ecda656f85ee50ddc5788f8f42770db312d7023cc7b4fba91b852d9bf7ad6eb8"}, 0xd8) getsockopt$inet6_IPV6_FLOWLABEL_MGR(r2, 0x29, 0x20, &(0x7f0000000080)={@private1, 0x4, 0x2, 0x2, 0xc, 0x81, 0x3}, &(0x7f00000000c0)=0x20) connect$inet6(r1, &(0x7f0000000200)={0xa, 0x0, 0x0, @initdev={0xfe, 0x88, '\x00', 0x0, 0x0}, 0x3}, 0x1c) r4 = syz_init_net_socket$ax25(0x3, 0x2, 0x0) bind$ax25(r4, &(0x7f0000000000)={{0x3, @remote={0xcc, 0xcc, 0xcc, 0xcc, 0xcc, 0xcc, 0x1}, 0x1}, [@null, @netrom={0xbb, 0xbb, 0xbb, 0xbb, 0xbb, 0x0, 0x0}, @bcast, @rose={0xbb, 0xbb, 0xbb, 0x1, 0x0}, @bcast, @netrom={0xbb, 0xbb, 0xbb, 0xbb, 0xbb, 0x0, 0x0}, @null, @bcast]}, 0x48) close(r4) connect$inet6(r1, &(0x7f0000000040)={0xa, 0x4e20, 0x0, @dev={0xfe, 0x80, '\x00', 0xd}}, 0x1c) [ 2827.942609][ T3079] bond1280: (slave bridge1201): making interface the new active one [ 2827.950939][ T3079] bridge1201: entered promiscuous mode [ 2827.969109][ T3079] bond1280: (slave bridge1201): Enslaving as an active interface with an up link 04:12:35 executing program 2: r0 = socket$netlink(0x10, 0x3, 0x0) r1 = socket$netlink(0x10, 0x3, 0x0) r2 = socket(0x10, 0x803, 0x0) sendmsg$NL80211_CMD_CRIT_PROTOCOL_START(r2, &(0x7f0000000580)={0x0, 0x0, &(0x7f0000000540)={0x0, 0x1c}}, 0x0) getsockname$packet(r2, &(0x7f0000000600)={0x11, 0x0, 0x0, 0x1, 0x0, 0x6, @broadcast}, &(0x7f0000000100)=0x14) sendmsg$nl_route(r1, &(0x7f0000000040)={0x0, 0x0, &(0x7f0000000000)={&(0x7f00000000c0)=ANY=[@ANYBLOB="3c00000010008506eb9afc4cd8d06e754a0081c5", @ANYRES32=r3, @ANYBLOB="2377f292252155b21c0012000c000100626f6e64000000000c0002000800010001"], 0x3c}}, 0x0) sendmsg$nl_route(r0, &(0x7f0000000240)={0x0, 0x0, &(0x7f0000000680)={&(0x7f0000000400)=@newlink={0x3c, 0x10, 0xffffff1f, 0x0, 0x253c33, {}, [@IFLA_LINKINFO={0x14, 0x12, 0x0, 0x1, @bridge={{0xb}, {0x4}}}, @IFLA_MASTER={0x8, 0xa, r3}]}, 0x3c}}, 0x0) [ 2828.104247][ T3085] netlink: 'syz-executor.1': attribute type 1 has an invalid length. [ 2828.235911][ T3085] bond1297: entered promiscuous mode [ 2828.247032][ T3085] 8021q: adding VLAN 0 to HW filter on device bond1297 04:12:35 executing program 3: r0 = socket$netlink(0x10, 0x3, 0x0) sendmsg$netlink(r0, &(0x7f00000047c0)={0x0, 0x0, &(0x7f0000000000)=[{&(0x7f0000000100)=ANY=[@ANYBLOB="18000000560001060000007900db730007"], 0x18}], 0x1}, 0x0) 04:12:35 executing program 1: r0 = socket$netlink(0x10, 0x3, 0x0) r1 = socket$netlink(0x10, 0x3, 0x0) r2 = socket(0x10, 0x803, 0x0) sendmsg$NL80211_CMD_CRIT_PROTOCOL_START(r2, &(0x7f0000000580)={0x0, 0x0, &(0x7f0000000540)={0x0, 0x1c}}, 0x0) getsockname$packet(r2, &(0x7f0000000600)={0x11, 0x0, 0x0, 0x1, 0x0, 0x6, @broadcast}, &(0x7f0000000100)=0x14) sendmsg$nl_route(r1, &(0x7f0000000040)={0x0, 0x0, &(0x7f0000000000)={&(0x7f00000000c0)=ANY=[@ANYBLOB="3c00000010008506eb9afc4cd8d06e754a0081c5", @ANYRES32=r3, @ANYBLOB="2377f292252155b21c0012000c000100626f6e64000000000c0002000800010001"], 0x3c}}, 0x0) sendmsg$nl_route(r0, &(0x7f0000000240)={0x0, 0x0, &(0x7f0000000680)={&(0x7f0000000400)=@newlink={0x3c, 0x10, 0xffffff1f, 0x0, 0x14672500, {}, [@IFLA_LINKINFO={0x14, 0x12, 0x0, 0x1, @bridge={{0xb}, {0x4}}}, @IFLA_MASTER={0x8, 0xa, r3}]}, 0x3c}}, 0x0) [ 2828.415897][ T3090] bond1297: (slave bridge1243): making interface the new active one [ 2828.429816][ T3090] bridge1243: entered promiscuous mode [ 2828.443719][ T3090] bond1297: (slave bridge1243): Enslaving as an active interface with an up link 04:12:36 executing program 4: r0 = socket$inet6(0xa, 0x2, 0x0) ioctl$sock_SIOCETHTOOL(r0, 0x89f0, &(0x7f0000000000)={'bridge0\x00', &(0x7f00000000c0)=ANY=[@ANYBLOB="1000000000000000000000000000000000000000000000000000000000000000000000001aeba036f1ba0e8a9787f68257d8328e6a226a8174cd51fac0abe08608753a4f7cb0bff6c41ad13822193c88b53dd85b0cf81c29c9c8a8000000000000000000000000f085767ae65bd55126e75c04dca74c7297e66bbf12bfb0a74b111a3a0c2d43673b9b840db111f78f38985e7581dbab7e1f1d53eb6dcf5413d725b24b9dfa9d3a53dc3e30d1d280fea5ef7b265877aef237576f97c88b323137f58f1ab1cfde2eff4dd93ec64a4f19900a3fe390d9e69a93b93a07cc9c0fa588b2bdee4703cb763304160853bf55785c86d05cd1f9d42e8cada37dea387719d9783611e54f"]}) [ 2828.580427][ T3099] netlink: 'syz-executor.5': attribute type 1 has an invalid length. [ 2828.629984][ T3099] workqueue: Failed to create a rescuer kthread for wq "bond1289": -EINTR [ 2828.887928][ T3108] netlink: 'syz-executor.2': attribute type 1 has an invalid length. 04:12:36 executing program 5: r0 = socket$netlink(0x10, 0x3, 0x0) r1 = socket$netlink(0x10, 0x3, 0x0) r2 = socket(0x10, 0x803, 0x0) sendmsg$NL80211_CMD_CRIT_PROTOCOL_START(r2, &(0x7f0000000580)={0x0, 0x0, &(0x7f0000000540)={0x0, 0x1c}}, 0x0) getsockname$packet(r2, &(0x7f0000000600)={0x11, 0x0, 0x0, 0x1, 0x0, 0x6, @broadcast}, &(0x7f0000000100)=0x14) sendmsg$nl_route(r1, &(0x7f0000000040)={0x0, 0x0, &(0x7f0000000000)={&(0x7f00000000c0)=ANY=[@ANYBLOB="3c00000010008506eb9afc4cd8d06e754a0081c5", @ANYRES32=r3, @ANYBLOB="2377f292252155b21c0012000c000100626f6e64000000000c0002000800010001"], 0x3c}}, 0x0) sendmsg$nl_route(r0, &(0x7f0000000240)={0x0, 0x0, &(0x7f0000000680)={&(0x7f0000000400)=@newlink={0x3c, 0x10, 0xffffff1f, 0x0, 0x5000000, {}, [@IFLA_LINKINFO={0x14, 0x12, 0x0, 0x1, @bridge={{0xb}, {0x4}}}, @IFLA_MASTER={0x8, 0xa, r3}]}, 0x3c}}, 0x0) 04:12:36 executing program 0: r0 = socket$inet6(0xa, 0x6, 0x0) bind$inet6(r0, &(0x7f0000000000)={0xa, 0x4e20}, 0x1c) listen(r0, 0x6) r1 = socket$inet6(0xa, 0x6, 0x0) r2 = socket$inet6(0xa, 0x80002, 0x88) bind$inet6(r2, &(0x7f0000000580)={0xa, 0x10010000004e20}, 0x1c) setsockopt$inet6_udp_int(r2, 0x11, 0xb, &(0x7f0000000080), 0x4) r3 = accept(r0, 0x0, &(0x7f0000000100)) setsockopt$inet6_tcp_TCP_MD5SIG(r3, 0x6, 0xe, &(0x7f0000000240)={@in={{0x2, 0x4e24, @empty}}, 0x0, 0x0, 0xe, 0x0, "f7e7cc8eefaf1c185361d9fd84f2f5f5da1fc0c85899c5fa73b0962b36bcf8d02916687c145ad91a568951b903f26927ecda656f85ee50ddc5788f8f42770db312d7023cc7b4fba91b852d9bf7ad6eb8"}, 0xd8) getsockopt$inet6_IPV6_FLOWLABEL_MGR(r2, 0x29, 0x20, &(0x7f0000000080)={@private1, 0x4, 0x2, 0x2, 0xc, 0x81, 0x3}, &(0x7f00000000c0)=0x20) connect$inet6(r1, &(0x7f0000000200)={0xa, 0x0, 0x0, @initdev={0xfe, 0x88, '\x00', 0x0, 0x0}, 0x3}, 0x1c) r4 = syz_init_net_socket$ax25(0x3, 0x2, 0x0) bind$ax25(r4, &(0x7f0000000000)={{0x3, @remote={0xcc, 0xcc, 0xcc, 0xcc, 0xcc, 0xcc, 0x1}, 0x1}, [@null, @netrom={0xbb, 0xbb, 0xbb, 0xbb, 0xbb, 0x0, 0x0}, @bcast, @rose={0xbb, 0xbb, 0xbb, 0x1, 0x0}, @bcast, @netrom={0xbb, 0xbb, 0xbb, 0xbb, 0xbb, 0x0, 0x0}, @null, @bcast]}, 0x48) close(r4) [ 2829.085390][ T3108] bond1281: entered promiscuous mode [ 2829.091070][ T3108] 8021q: adding VLAN 0 to HW filter on device bond1281 [ 2829.160068][ T3111] bond1281: (slave bridge1202): making interface the new active one [ 2829.176977][ T3111] bridge1202: entered promiscuous mode [ 2829.208437][ T3111] bond1281: (slave bridge1202): Enslaving as an active interface with an up link [ 2829.223342][ T3114] netlink: 'syz-executor.1': attribute type 1 has an invalid length. 04:12:36 executing program 2: r0 = socket$netlink(0x10, 0x3, 0x0) r1 = socket$netlink(0x10, 0x3, 0x0) r2 = socket(0x10, 0x803, 0x0) sendmsg$NL80211_CMD_CRIT_PROTOCOL_START(r2, &(0x7f0000000580)={0x0, 0x0, &(0x7f0000000540)={0x0, 0x1c}}, 0x0) getsockname$packet(r2, &(0x7f0000000600)={0x11, 0x0, 0x0, 0x1, 0x0, 0x6, @broadcast}, &(0x7f0000000100)=0x14) sendmsg$nl_route(r1, &(0x7f0000000040)={0x0, 0x0, &(0x7f0000000000)={&(0x7f00000000c0)=ANY=[@ANYBLOB="3c00000010008506eb9afc4cd8d06e754a0081c5", @ANYRES32=r3, @ANYBLOB="2377f292252155b21c0012000c000100626f6e64000000000c0002000800010001"], 0x3c}}, 0x0) sendmsg$nl_route(r0, &(0x7f0000000240)={0x0, 0x0, &(0x7f0000000680)={&(0x7f0000000400)=@newlink={0x3c, 0x10, 0xffffff1f, 0x0, 0x253c34, {}, [@IFLA_LINKINFO={0x14, 0x12, 0x0, 0x1, @bridge={{0xb}, {0x4}}}, @IFLA_MASTER={0x8, 0xa, r3}]}, 0x3c}}, 0x0) 04:12:36 executing program 3: r0 = socket$netlink(0x10, 0x3, 0x0) sendmsg$netlink(r0, &(0x7f00000047c0)={0x0, 0x0, &(0x7f0000000000)=[{&(0x7f0000000100)=ANY=[@ANYBLOB="18000000560001060000007a00db730007"], 0x18}], 0x1}, 0x0) [ 2829.347123][ T3114] bond1298: entered promiscuous mode [ 2829.364110][T25146] wlan1: No active IBSS STAs - trying to scan for other IBSS networks with same SSID (merge) [ 2829.377013][ T3114] 8021q: adding VLAN 0 to HW filter on device bond1298 [ 2829.419592][ T3115] bond1298: (slave bridge1244): making interface the new active one [ 2829.439017][ T3115] bridge1244: entered promiscuous mode [ 2829.450050][ T3115] bond1298: (slave bridge1244): Enslaving as an active interface with an up link 04:12:37 executing program 1: r0 = socket$netlink(0x10, 0x3, 0x0) r1 = socket$netlink(0x10, 0x3, 0x0) r2 = socket(0x10, 0x803, 0x0) sendmsg$NL80211_CMD_CRIT_PROTOCOL_START(r2, &(0x7f0000000580)={0x0, 0x0, &(0x7f0000000540)={0x0, 0x1c}}, 0x0) getsockname$packet(r2, &(0x7f0000000600)={0x11, 0x0, 0x0, 0x1, 0x0, 0x6, @broadcast}, &(0x7f0000000100)=0x14) sendmsg$nl_route(r1, &(0x7f0000000040)={0x0, 0x0, &(0x7f0000000000)={&(0x7f00000000c0)=ANY=[@ANYBLOB="3c00000010008506eb9afc4cd8d06e754a0081c5", @ANYRES32=r3, @ANYBLOB="2377f292252155b21c0012000c000100626f6e64000000000c0002000800010001"], 0x3c}}, 0x0) sendmsg$nl_route(r0, &(0x7f0000000240)={0x0, 0x0, &(0x7f0000000680)={&(0x7f0000000400)=@newlink={0x3c, 0x10, 0xffffff1f, 0x0, 0x15672500, {}, [@IFLA_LINKINFO={0x14, 0x12, 0x0, 0x1, @bridge={{0xb}, {0x4}}}, @IFLA_MASTER={0x8, 0xa, r3}]}, 0x3c}}, 0x0) 04:12:37 executing program 4: socket$inet6(0xa, 0x2, 0x0) (async) r0 = socket$inet6(0xa, 0x2, 0x0) ioctl$sock_SIOCETHTOOL(r0, 0x89f0, &(0x7f0000000000)={'bridge0\x00', &(0x7f00000000c0)=ANY=[@ANYBLOB="1000000000000000000000000000000000000000000000000000000000000000000000001aeba036f1ba0e8a9787f68257d8328e6a226a8174cd51fac0abe08608753a4f7cb0bff6c41ad13822193c88b53dd85b0cf81c29c9c8a8000000000000000000000000f085767ae65bd55126e75c04dca74c7297e66bbf12bfb0a74b111a3a0c2d43673b9b840db111f78f38985e7581dbab7e1f1d53eb6dcf5413d725b24b9dfa9d3a53dc3e30d1d280fea5ef7b265877aef237576f97c88b323137f58f1ab1cfde2eff4dd93ec64a4f19900a3fe390d9e69a93b93a07cc9c0fa588b2bdee4703cb763304160853bf55785c86d05cd1f9d42e8cada37dea387719d9783611e54f"]}) [ 2829.758177][ T3125] netlink: 'syz-executor.5': attribute type 1 has an invalid length. [ 2829.852613][ T3125] bond1289: entered promiscuous mode [ 2829.858326][ T3125] 8021q: adding VLAN 0 to HW filter on device bond1289 04:12:37 executing program 0: r0 = socket$inet6(0xa, 0x6, 0x0) bind$inet6(r0, &(0x7f0000000000)={0xa, 0x4e20}, 0x1c) listen(r0, 0x6) r1 = socket$inet6(0xa, 0x6, 0x0) r2 = socket$inet6(0xa, 0x80002, 0x88) bind$inet6(r2, &(0x7f0000000580)={0xa, 0x10010000004e20}, 0x1c) setsockopt$inet6_udp_int(r2, 0x11, 0xb, &(0x7f0000000080), 0x4) r3 = accept(r0, 0x0, &(0x7f0000000100)) setsockopt$inet6_tcp_TCP_MD5SIG(r3, 0x6, 0xe, &(0x7f0000000240)={@in={{0x2, 0x4e24, @empty}}, 0x0, 0x0, 0xe, 0x0, "f7e7cc8eefaf1c185361d9fd84f2f5f5da1fc0c85899c5fa73b0962b36bcf8d02916687c145ad91a568951b903f26927ecda656f85ee50ddc5788f8f42770db312d7023cc7b4fba91b852d9bf7ad6eb8"}, 0xd8) getsockopt$inet6_IPV6_FLOWLABEL_MGR(r2, 0x29, 0x20, &(0x7f0000000080)={@private1, 0x4, 0x2, 0x2, 0xc, 0x81, 0x3}, &(0x7f00000000c0)=0x20) connect$inet6(r1, &(0x7f0000000200)={0xa, 0x0, 0x0, @initdev={0xfe, 0x88, '\x00', 0x0, 0x0}, 0x3}, 0x1c) r4 = syz_init_net_socket$ax25(0x3, 0x2, 0x0) bind$ax25(r4, &(0x7f0000000000)={{0x3, @remote={0xcc, 0xcc, 0xcc, 0xcc, 0xcc, 0xcc, 0x1}, 0x1}, [@null, @netrom={0xbb, 0xbb, 0xbb, 0xbb, 0xbb, 0x0, 0x0}, @bcast, @rose={0xbb, 0xbb, 0xbb, 0x1, 0x0}, @bcast, @netrom={0xbb, 0xbb, 0xbb, 0xbb, 0xbb, 0x0, 0x0}, @null, @bcast]}, 0x48) [ 2830.189674][ T3127] bond1289: (slave bridge1252): making interface the new active one [ 2830.240257][ T3127] bridge1252: entered promiscuous mode [ 2830.252661][ T3127] bond1289: (slave bridge1252): Enslaving as an active interface with an up link 04:12:37 executing program 5: r0 = socket$netlink(0x10, 0x3, 0x0) r1 = socket$netlink(0x10, 0x3, 0x0) r2 = socket(0x10, 0x803, 0x0) sendmsg$NL80211_CMD_CRIT_PROTOCOL_START(r2, &(0x7f0000000580)={0x0, 0x0, &(0x7f0000000540)={0x0, 0x1c}}, 0x0) getsockname$packet(r2, &(0x7f0000000600)={0x11, 0x0, 0x0, 0x1, 0x0, 0x6, @broadcast}, &(0x7f0000000100)=0x14) sendmsg$nl_route(r1, &(0x7f0000000040)={0x0, 0x0, &(0x7f0000000000)={&(0x7f00000000c0)=ANY=[@ANYBLOB="3c00000010008506eb9afc4cd8d06e754a0081c5", @ANYRES32=r3, @ANYBLOB="2377f292252155b21c0012000c000100626f6e64000000000c0002000800010001"], 0x3c}}, 0x0) sendmsg$nl_route(r0, &(0x7f0000000240)={0x0, 0x0, &(0x7f0000000680)={&(0x7f0000000400)=@newlink={0x3c, 0x10, 0xffffff1f, 0x0, 0x6000000, {}, [@IFLA_LINKINFO={0x14, 0x12, 0x0, 0x1, @bridge={{0xb}, {0x4}}}, @IFLA_MASTER={0x8, 0xa, r3}]}, 0x3c}}, 0x0) 04:12:37 executing program 3: r0 = socket$netlink(0x10, 0x3, 0x0) sendmsg$netlink(r0, &(0x7f00000047c0)={0x0, 0x0, &(0x7f0000000000)=[{&(0x7f0000000100)=ANY=[@ANYBLOB="18000000560001060000007b00db730007"], 0x18}], 0x1}, 0x0) [ 2830.357067][ T3132] netlink: 'syz-executor.2': attribute type 1 has an invalid length. [ 2830.399508][ T3132] workqueue: Failed to create a rescuer kthread for wq "bond1282": -EINTR [ 2830.579265][ T3134] netlink: 'syz-executor.1': attribute type 1 has an invalid length. 04:12:38 executing program 0: r0 = socket$inet6(0xa, 0x6, 0x0) bind$inet6(r0, &(0x7f0000000000)={0xa, 0x4e20}, 0x1c) listen(r0, 0x6) r1 = socket$inet6(0xa, 0x6, 0x0) r2 = socket$inet6(0xa, 0x80002, 0x88) bind$inet6(r2, &(0x7f0000000580)={0xa, 0x10010000004e20}, 0x1c) setsockopt$inet6_udp_int(r2, 0x11, 0xb, &(0x7f0000000080), 0x4) r3 = accept(r0, 0x0, &(0x7f0000000100)) setsockopt$inet6_tcp_TCP_MD5SIG(r3, 0x6, 0xe, &(0x7f0000000240)={@in={{0x2, 0x4e24, @empty}}, 0x0, 0x0, 0xe, 0x0, "f7e7cc8eefaf1c185361d9fd84f2f5f5da1fc0c85899c5fa73b0962b36bcf8d02916687c145ad91a568951b903f26927ecda656f85ee50ddc5788f8f42770db312d7023cc7b4fba91b852d9bf7ad6eb8"}, 0xd8) getsockopt$inet6_IPV6_FLOWLABEL_MGR(r2, 0x29, 0x20, &(0x7f0000000080)={@private1, 0x4, 0x2, 0x2, 0xc, 0x81, 0x3}, &(0x7f00000000c0)=0x20) connect$inet6(r1, &(0x7f0000000200)={0xa, 0x0, 0x0, @initdev={0xfe, 0x88, '\x00', 0x0, 0x0}, 0x3}, 0x1c) bind$ax25(0xffffffffffffffff, &(0x7f0000000000)={{0x3, @remote={0xcc, 0xcc, 0xcc, 0xcc, 0xcc, 0xcc, 0x1}, 0x1}, [@null, @netrom={0xbb, 0xbb, 0xbb, 0xbb, 0xbb, 0x0, 0x0}, @bcast, @rose={0xbb, 0xbb, 0xbb, 0x1, 0x0}, @bcast, @netrom={0xbb, 0xbb, 0xbb, 0xbb, 0xbb, 0x0, 0x0}, @null, @bcast]}, 0x48) 04:12:38 executing program 2: r0 = socket$netlink(0x10, 0x3, 0x0) r1 = socket$netlink(0x10, 0x3, 0x0) r2 = socket(0x10, 0x803, 0x0) sendmsg$NL80211_CMD_CRIT_PROTOCOL_START(r2, &(0x7f0000000580)={0x0, 0x0, &(0x7f0000000540)={0x0, 0x1c}}, 0x0) getsockname$packet(r2, &(0x7f0000000600)={0x11, 0x0, 0x0, 0x1, 0x0, 0x6, @broadcast}, &(0x7f0000000100)=0x14) sendmsg$nl_route(r1, &(0x7f0000000040)={0x0, 0x0, &(0x7f0000000000)={&(0x7f00000000c0)=ANY=[@ANYBLOB="3c00000010008506eb9afc4cd8d06e754a0081c5", @ANYRES32=r3, @ANYBLOB="2377f292252155b21c0012000c000100626f6e64000000000c0002000800010001"], 0x3c}}, 0x0) sendmsg$nl_route(r0, &(0x7f0000000240)={0x0, 0x0, &(0x7f0000000680)={&(0x7f0000000400)=@newlink={0x3c, 0x10, 0xffffff1f, 0x0, 0x253c35, {}, [@IFLA_LINKINFO={0x14, 0x12, 0x0, 0x1, @bridge={{0xb}, {0x4}}}, @IFLA_MASTER={0x8, 0xa, r3}]}, 0x3c}}, 0x0) 04:12:38 executing program 4: r0 = socket$inet6(0xa, 0x2, 0x0) ioctl$sock_SIOCETHTOOL(r0, 0x89f0, &(0x7f0000000000)={'bridge0\x00', &(0x7f00000000c0)=ANY=[@ANYBLOB="1000000000000000000000000000000000000000000000000000000000000000000000001aeba036f1ba0e8a9787f68257d8328e6a226a8174cd51fac0abe08608753a4f7cb0bff6c41ad13822193c88b53dd85b0cf81c29c9c8a8000000000000000000000000f085767ae65bd55126e75c04dca74c7297e66bbf12bfb0a74b111a3a0c2d43673b9b840db111f78f38985e7581dbab7e1f1d53eb6dcf5413d725b24b9dfa9d3a53dc3e30d1d280fea5ef7b265877aef237576f97c88b323137f58f1ab1cfde2eff4dd93ec64a4f19900a3fe390d9e69a93b93a07cc9c0fa588b2bdee4703cb763304160853bf55785c86d05cd1f9d42e8cada37dea387719d9783611e54f"]}) [ 2830.632080][ T3134] workqueue: Failed to create a rescuer kthread for wq "bond1299": -EINTR [ 2831.115222][ T3148] netlink: 'syz-executor.5': attribute type 1 has an invalid length. 04:12:38 executing program 1: r0 = socket$netlink(0x10, 0x3, 0x0) r1 = socket$netlink(0x10, 0x3, 0x0) r2 = socket(0x10, 0x803, 0x0) sendmsg$NL80211_CMD_CRIT_PROTOCOL_START(r2, &(0x7f0000000580)={0x0, 0x0, &(0x7f0000000540)={0x0, 0x1c}}, 0x0) getsockname$packet(r2, &(0x7f0000000600)={0x11, 0x0, 0x0, 0x1, 0x0, 0x6, @broadcast}, &(0x7f0000000100)=0x14) sendmsg$nl_route(r1, &(0x7f0000000040)={0x0, 0x0, &(0x7f0000000000)={&(0x7f00000000c0)=ANY=[@ANYBLOB="3c00000010008506eb9afc4cd8d06e754a0081c5", @ANYRES32=r3, @ANYBLOB="2377f292252155b21c0012000c000100626f6e64000000000c0002000800010001"], 0x3c}}, 0x0) sendmsg$nl_route(r0, &(0x7f0000000240)={0x0, 0x0, &(0x7f0000000680)={&(0x7f0000000400)=@newlink={0x3c, 0x10, 0xffffff1f, 0x0, 0x16672500, {}, [@IFLA_LINKINFO={0x14, 0x12, 0x0, 0x1, @bridge={{0xb}, {0x4}}}, @IFLA_MASTER={0x8, 0xa, r3}]}, 0x3c}}, 0x0) [ 2831.337590][ T3148] bond1290: entered promiscuous mode [ 2831.360710][ T3148] 8021q: adding VLAN 0 to HW filter on device bond1290 [ 2831.479351][ T3152] bond1290: (slave bridge1253): making interface the new active one [ 2831.519991][ T3152] bridge1253: entered promiscuous mode [ 2831.570014][ T3152] bond1290: (slave bridge1253): Enslaving as an active interface with an up link [ 2831.636622][ T3159] netlink: 'syz-executor.2': attribute type 1 has an invalid length. 04:12:39 executing program 5: r0 = socket$netlink(0x10, 0x3, 0x0) r1 = socket$netlink(0x10, 0x3, 0x0) r2 = socket(0x10, 0x803, 0x0) sendmsg$NL80211_CMD_CRIT_PROTOCOL_START(r2, &(0x7f0000000580)={0x0, 0x0, &(0x7f0000000540)={0x0, 0x1c}}, 0x0) getsockname$packet(r2, &(0x7f0000000600)={0x11, 0x0, 0x0, 0x1, 0x0, 0x6, @broadcast}, &(0x7f0000000100)=0x14) sendmsg$nl_route(r1, &(0x7f0000000040)={0x0, 0x0, &(0x7f0000000000)={&(0x7f00000000c0)=ANY=[@ANYBLOB="3c00000010008506eb9afc4cd8d06e754a0081c5", @ANYRES32=r3, @ANYBLOB="2377f292252155b21c0012000c000100626f6e64000000000c0002000800010001"], 0x3c}}, 0x0) sendmsg$nl_route(r0, &(0x7f0000000240)={0x0, 0x0, &(0x7f0000000680)={&(0x7f0000000400)=@newlink={0x3c, 0x10, 0xffffff1f, 0x0, 0x7000000, {}, [@IFLA_LINKINFO={0x14, 0x12, 0x0, 0x1, @bridge={{0xb}, {0x4}}}, @IFLA_MASTER={0x8, 0xa, r3}]}, 0x3c}}, 0x0) [ 2831.691198][ T3159] bond1282: entered promiscuous mode 04:12:39 executing program 0: r0 = socket$inet6(0xa, 0x6, 0x0) bind$inet6(r0, &(0x7f0000000000)={0xa, 0x4e20}, 0x1c) listen(r0, 0x6) r1 = socket$inet6(0xa, 0x6, 0x0) r2 = socket$inet6(0xa, 0x80002, 0x88) bind$inet6(r2, &(0x7f0000000580)={0xa, 0x10010000004e20}, 0x1c) setsockopt$inet6_udp_int(r2, 0x11, 0xb, &(0x7f0000000080), 0x4) r3 = accept(r0, 0x0, &(0x7f0000000100)) setsockopt$inet6_tcp_TCP_MD5SIG(r3, 0x6, 0xe, &(0x7f0000000240)={@in={{0x2, 0x4e24, @empty}}, 0x0, 0x0, 0xe, 0x0, "f7e7cc8eefaf1c185361d9fd84f2f5f5da1fc0c85899c5fa73b0962b36bcf8d02916687c145ad91a568951b903f26927ecda656f85ee50ddc5788f8f42770db312d7023cc7b4fba91b852d9bf7ad6eb8"}, 0xd8) getsockopt$inet6_IPV6_FLOWLABEL_MGR(r2, 0x29, 0x20, &(0x7f0000000080)={@private1, 0x4, 0x2, 0x2, 0xc, 0x81, 0x3}, &(0x7f00000000c0)=0x20) connect$inet6(r1, &(0x7f0000000200)={0xa, 0x0, 0x0, @initdev={0xfe, 0x88, '\x00', 0x0, 0x0}, 0x3}, 0x1c) bind$ax25(0xffffffffffffffff, &(0x7f0000000000)={{0x3, @remote={0xcc, 0xcc, 0xcc, 0xcc, 0xcc, 0xcc, 0x1}, 0x1}, [@null, @netrom={0xbb, 0xbb, 0xbb, 0xbb, 0xbb, 0x0, 0x0}, @bcast, @rose={0xbb, 0xbb, 0xbb, 0x1, 0x0}, @bcast, @netrom={0xbb, 0xbb, 0xbb, 0xbb, 0xbb, 0x0, 0x0}, @null, @bcast]}, 0x48) [ 2831.754265][ T3159] 8021q: adding VLAN 0 to HW filter on device bond1282 [ 2831.919792][ T3160] bond1282: (slave bridge1203): making interface the new active one [ 2831.941839][ T3160] bridge1203: entered promiscuous mode [ 2831.970326][ T3160] bond1282: (slave bridge1203): Enslaving as an active interface with an up link 04:12:39 executing program 2: r0 = socket$netlink(0x10, 0x3, 0x0) r1 = socket$netlink(0x10, 0x3, 0x0) r2 = socket(0x10, 0x803, 0x0) sendmsg$NL80211_CMD_CRIT_PROTOCOL_START(r2, &(0x7f0000000580)={0x0, 0x0, &(0x7f0000000540)={0x0, 0x1c}}, 0x0) getsockname$packet(r2, &(0x7f0000000600)={0x11, 0x0, 0x0, 0x1, 0x0, 0x6, @broadcast}, &(0x7f0000000100)=0x14) sendmsg$nl_route(r1, &(0x7f0000000040)={0x0, 0x0, &(0x7f0000000000)={&(0x7f00000000c0)=ANY=[@ANYBLOB="3c00000010008506eb9afc4cd8d06e754a0081c5", @ANYRES32=r3, @ANYBLOB="2377f292252155b21c0012000c000100626f6e64000000000c0002000800010001"], 0x3c}}, 0x0) sendmsg$nl_route(r0, &(0x7f0000000240)={0x0, 0x0, &(0x7f0000000680)={&(0x7f0000000400)=@newlink={0x3c, 0x10, 0xffffff1f, 0x0, 0x253c36, {}, [@IFLA_LINKINFO={0x14, 0x12, 0x0, 0x1, @bridge={{0xb}, {0x4}}}, @IFLA_MASTER={0x8, 0xa, r3}]}, 0x3c}}, 0x0) 04:12:39 executing program 3: r0 = socket$netlink(0x10, 0x3, 0x0) sendmsg$netlink(r0, &(0x7f00000047c0)={0x0, 0x0, &(0x7f0000000000)=[{&(0x7f0000000100)=ANY=[@ANYBLOB="18000000560001060000007c00db730007"], 0x18}], 0x1}, 0x0) 04:12:39 executing program 4: r0 = socket$inet_sctp(0x2, 0x5, 0x84) setsockopt$inet_sctp_SCTP_DEFAULT_PRINFO(r0, 0x84, 0x72, &(0x7f0000000040)={0x0, 0x0, 0x2}, 0xc) r1 = accept$inet(r0, 0x0, &(0x7f0000000100)) setsockopt$inet_sctp_SCTP_PEER_ADDR_THLDS(r1, 0x84, 0x1f, &(0x7f0000000140)={0x0, @in={{0x2, 0x4e21, @empty}}, 0x711f, 0x6f}, 0x90) r2 = socket$inet6(0xa, 0x2, 0x0) setsockopt$inet_sctp_SCTP_ENABLE_STREAM_RESET(r0, 0x84, 0x76, &(0x7f0000000080)={0x0, 0x9}, 0x8) r3 = socket$inet6(0xa, 0x80002, 0x88) bind$inet6(r3, &(0x7f0000000580)={0xa, 0x10010000004e20}, 0x1c) setsockopt$inet6_udp_int(r3, 0x11, 0xb, &(0x7f0000000080), 0x4) ioctl$sock_SIOCETHTOOL(r2, 0x89f0, &(0x7f0000000000)={'bridge0\x00', &(0x7f00000000c0)=ANY=[@ANYRES64=r3]}) [ 2832.177763][ T3165] bond1299: entered promiscuous mode [ 2832.220557][ T3165] 8021q: adding VLAN 0 to HW filter on device bond1299 [ 2832.346742][ T3166] bond1299: (slave bridge1245): making interface the new active one [ 2832.397485][ T3166] bridge1245: entered promiscuous mode 04:12:39 executing program 1: r0 = socket$netlink(0x10, 0x3, 0x0) r1 = socket$netlink(0x10, 0x3, 0x0) r2 = socket(0x10, 0x803, 0x0) sendmsg$NL80211_CMD_CRIT_PROTOCOL_START(r2, &(0x7f0000000580)={0x0, 0x0, &(0x7f0000000540)={0x0, 0x1c}}, 0x0) getsockname$packet(r2, &(0x7f0000000600)={0x11, 0x0, 0x0, 0x1, 0x0, 0x6, @broadcast}, &(0x7f0000000100)=0x14) sendmsg$nl_route(r1, &(0x7f0000000040)={0x0, 0x0, &(0x7f0000000000)={&(0x7f00000000c0)=ANY=[@ANYBLOB="3c00000010008506eb9afc4cd8d06e754a0081c5", @ANYRES32=r3, @ANYBLOB="2377f292252155b21c0012000c000100626f6e64000000000c0002000800010001"], 0x3c}}, 0x0) sendmsg$nl_route(r0, &(0x7f0000000240)={0x0, 0x0, &(0x7f0000000680)={&(0x7f0000000400)=@newlink={0x3c, 0x10, 0xffffff1f, 0x0, 0x17672500, {}, [@IFLA_LINKINFO={0x14, 0x12, 0x0, 0x1, @bridge={{0xb}, {0x4}}}, @IFLA_MASTER={0x8, 0xa, r3}]}, 0x3c}}, 0x0) [ 2832.439495][ T3166] bond1299: (slave bridge1245): Enslaving as an active interface with an up link [ 2832.606208][ T3170] bond1291: entered promiscuous mode [ 2832.619757][ T3170] 8021q: adding VLAN 0 to HW filter on device bond1291 04:12:40 executing program 0: r0 = socket$inet6(0xa, 0x6, 0x0) bind$inet6(r0, &(0x7f0000000000)={0xa, 0x4e20}, 0x1c) listen(r0, 0x6) r1 = socket$inet6(0xa, 0x6, 0x0) r2 = socket$inet6(0xa, 0x80002, 0x88) bind$inet6(r2, &(0x7f0000000580)={0xa, 0x10010000004e20}, 0x1c) setsockopt$inet6_udp_int(r2, 0x11, 0xb, &(0x7f0000000080), 0x4) r3 = accept(r0, 0x0, &(0x7f0000000100)) setsockopt$inet6_tcp_TCP_MD5SIG(r3, 0x6, 0xe, &(0x7f0000000240)={@in={{0x2, 0x4e24, @empty}}, 0x0, 0x0, 0xe, 0x0, "f7e7cc8eefaf1c185361d9fd84f2f5f5da1fc0c85899c5fa73b0962b36bcf8d02916687c145ad91a568951b903f26927ecda656f85ee50ddc5788f8f42770db312d7023cc7b4fba91b852d9bf7ad6eb8"}, 0xd8) getsockopt$inet6_IPV6_FLOWLABEL_MGR(r2, 0x29, 0x20, &(0x7f0000000080)={@private1, 0x4, 0x2, 0x2, 0xc, 0x81, 0x3}, &(0x7f00000000c0)=0x20) connect$inet6(r1, &(0x7f0000000200)={0xa, 0x0, 0x0, @initdev={0xfe, 0x88, '\x00', 0x0, 0x0}, 0x3}, 0x1c) bind$ax25(0xffffffffffffffff, &(0x7f0000000000)={{0x3, @remote={0xcc, 0xcc, 0xcc, 0xcc, 0xcc, 0xcc, 0x1}, 0x1}, [@null, @netrom={0xbb, 0xbb, 0xbb, 0xbb, 0xbb, 0x0, 0x0}, @bcast, @rose={0xbb, 0xbb, 0xbb, 0x1, 0x0}, @bcast, @netrom={0xbb, 0xbb, 0xbb, 0xbb, 0xbb, 0x0, 0x0}, @null, @bcast]}, 0x48) 04:12:40 executing program 5: r0 = socket$netlink(0x10, 0x3, 0x0) r1 = socket$netlink(0x10, 0x3, 0x0) r2 = socket(0x10, 0x803, 0x0) sendmsg$NL80211_CMD_CRIT_PROTOCOL_START(r2, &(0x7f0000000580)={0x0, 0x0, &(0x7f0000000540)={0x0, 0x1c}}, 0x0) getsockname$packet(r2, &(0x7f0000000600)={0x11, 0x0, 0x0, 0x1, 0x0, 0x6, @broadcast}, &(0x7f0000000100)=0x14) sendmsg$nl_route(r1, &(0x7f0000000040)={0x0, 0x0, &(0x7f0000000000)={&(0x7f00000000c0)=ANY=[@ANYBLOB="3c00000010008506eb9afc4cd8d06e754a0081c5", @ANYRES32=r3, @ANYBLOB="2377f292252155b21c0012000c000100626f6e64000000000c0002000800010001"], 0x3c}}, 0x0) sendmsg$nl_route(r0, &(0x7f0000000240)={0x0, 0x0, &(0x7f0000000680)={&(0x7f0000000400)=@newlink={0x3c, 0x10, 0xffffff1f, 0x0, 0x8000000, {}, [@IFLA_LINKINFO={0x14, 0x12, 0x0, 0x1, @bridge={{0xb}, {0x4}}}, @IFLA_MASTER={0x8, 0xa, r3}]}, 0x3c}}, 0x0) [ 2832.658167][ T3173] bond1291: (slave bridge1254): making interface the new active one [ 2832.666448][ T3173] bridge1254: entered promiscuous mode [ 2832.676234][ T3173] bond1291: (slave bridge1254): Enslaving as an active interface with an up link [ 2832.753804][ T3180] validate_nla: 2 callbacks suppressed [ 2832.753824][ T3180] netlink: 'syz-executor.2': attribute type 1 has an invalid length. [ 2832.894297][ T3180] bond1283: entered promiscuous mode [ 2832.909098][ T3180] 8021q: adding VLAN 0 to HW filter on device bond1283 [ 2833.019539][ T3181] bond1283: (slave bridge1204): making interface the new active one [ 2833.028072][ T3181] bridge1204: entered promiscuous mode [ 2833.045036][ T3181] bond1283: (slave bridge1204): Enslaving as an active interface with an up link 04:12:40 executing program 4: r0 = socket$inet_sctp(0x2, 0x5, 0x84) setsockopt$inet_sctp_SCTP_DEFAULT_PRINFO(r0, 0x84, 0x72, &(0x7f0000000040)={0x0, 0x0, 0x2}, 0xc) r1 = accept$inet(r0, 0x0, &(0x7f0000000100)) setsockopt$inet_sctp_SCTP_PEER_ADDR_THLDS(r1, 0x84, 0x1f, &(0x7f0000000140)={0x0, @in={{0x2, 0x4e21, @empty}}, 0x711f, 0x6f}, 0x90) r2 = socket$inet6(0xa, 0x2, 0x0) setsockopt$inet_sctp_SCTP_ENABLE_STREAM_RESET(r0, 0x84, 0x76, &(0x7f0000000080)={0x0, 0x9}, 0x8) r3 = socket$inet6(0xa, 0x80002, 0x88) bind$inet6(r3, &(0x7f0000000580)={0xa, 0x10010000004e20}, 0x1c) setsockopt$inet6_udp_int(r3, 0x11, 0xb, &(0x7f0000000080), 0x4) ioctl$sock_SIOCETHTOOL(r2, 0x89f0, &(0x7f0000000000)={'bridge0\x00', &(0x7f00000000c0)=ANY=[@ANYRES64=r3]}) socket$inet_sctp(0x2, 0x5, 0x84) (async) setsockopt$inet_sctp_SCTP_DEFAULT_PRINFO(r0, 0x84, 0x72, &(0x7f0000000040)={0x0, 0x0, 0x2}, 0xc) (async) accept$inet(r0, 0x0, &(0x7f0000000100)) (async) setsockopt$inet_sctp_SCTP_PEER_ADDR_THLDS(r1, 0x84, 0x1f, &(0x7f0000000140)={0x0, @in={{0x2, 0x4e21, @empty}}, 0x711f, 0x6f}, 0x90) (async) socket$inet6(0xa, 0x2, 0x0) (async) setsockopt$inet_sctp_SCTP_ENABLE_STREAM_RESET(r0, 0x84, 0x76, &(0x7f0000000080)={0x0, 0x9}, 0x8) (async) socket$inet6(0xa, 0x80002, 0x88) (async) bind$inet6(r3, &(0x7f0000000580)={0xa, 0x10010000004e20}, 0x1c) (async) setsockopt$inet6_udp_int(r3, 0x11, 0xb, &(0x7f0000000080), 0x4) (async) ioctl$sock_SIOCETHTOOL(r2, 0x89f0, &(0x7f0000000000)={'bridge0\x00', &(0x7f00000000c0)=ANY=[@ANYRES64=r3]}) (async) 04:12:40 executing program 2: r0 = socket$netlink(0x10, 0x3, 0x0) r1 = socket$netlink(0x10, 0x3, 0x0) r2 = socket(0x10, 0x803, 0x0) sendmsg$NL80211_CMD_CRIT_PROTOCOL_START(r2, &(0x7f0000000580)={0x0, 0x0, &(0x7f0000000540)={0x0, 0x1c}}, 0x0) getsockname$packet(r2, &(0x7f0000000600)={0x11, 0x0, 0x0, 0x1, 0x0, 0x6, @broadcast}, &(0x7f0000000100)=0x14) sendmsg$nl_route(r1, &(0x7f0000000040)={0x0, 0x0, &(0x7f0000000000)={&(0x7f00000000c0)=ANY=[@ANYBLOB="3c00000010008506eb9afc4cd8d06e754a0081c5", @ANYRES32=r3, @ANYBLOB="2377f292252155b21c0012000c000100626f6e64000000000c0002000800010001"], 0x3c}}, 0x0) sendmsg$nl_route(r0, &(0x7f0000000240)={0x0, 0x0, &(0x7f0000000680)={&(0x7f0000000400)=@newlink={0x3c, 0x10, 0xffffff1f, 0x0, 0x253c37, {}, [@IFLA_LINKINFO={0x14, 0x12, 0x0, 0x1, @bridge={{0xb}, {0x4}}}, @IFLA_MASTER={0x8, 0xa, r3}]}, 0x3c}}, 0x0) [ 2833.269435][ T3186] netlink: 'syz-executor.1': attribute type 1 has an invalid length. [ 2833.338326][ T3186] bond1300: entered promiscuous mode [ 2833.346597][ T3186] 8021q: adding VLAN 0 to HW filter on device bond1300 04:12:40 executing program 3: r0 = socket$netlink(0x10, 0x3, 0x0) sendmsg$netlink(r0, &(0x7f00000047c0)={0x0, 0x0, &(0x7f0000000000)=[{&(0x7f0000000100)=ANY=[@ANYBLOB="18000000560001060000007d00db730007"], 0x18}], 0x1}, 0x0) [ 2833.500332][ T3189] bond1300: (slave bridge1246): making interface the new active one [ 2833.525858][ T3189] bridge1246: entered promiscuous mode 04:12:41 executing program 1: r0 = socket$netlink(0x10, 0x3, 0x0) r1 = socket$netlink(0x10, 0x3, 0x0) r2 = socket(0x10, 0x803, 0x0) sendmsg$NL80211_CMD_CRIT_PROTOCOL_START(r2, &(0x7f0000000580)={0x0, 0x0, &(0x7f0000000540)={0x0, 0x1c}}, 0x0) getsockname$packet(r2, &(0x7f0000000600)={0x11, 0x0, 0x0, 0x1, 0x0, 0x6, @broadcast}, &(0x7f0000000100)=0x14) sendmsg$nl_route(r1, &(0x7f0000000040)={0x0, 0x0, &(0x7f0000000000)={&(0x7f00000000c0)=ANY=[@ANYBLOB="3c00000010008506eb9afc4cd8d06e754a0081c5", @ANYRES32=r3, @ANYBLOB="2377f292252155b21c0012000c000100626f6e64000000000c0002000800010001"], 0x3c}}, 0x0) sendmsg$nl_route(r0, &(0x7f0000000240)={0x0, 0x0, &(0x7f0000000680)={&(0x7f0000000400)=@newlink={0x3c, 0x10, 0xffffff1f, 0x0, 0x18672500, {}, [@IFLA_LINKINFO={0x14, 0x12, 0x0, 0x1, @bridge={{0xb}, {0x4}}}, @IFLA_MASTER={0x8, 0xa, r3}]}, 0x3c}}, 0x0) [ 2833.549668][ T3189] bond1300: (slave bridge1246): Enslaving as an active interface with an up link [ 2833.570336][ T3194] netlink: 'syz-executor.5': attribute type 1 has an invalid length. 04:12:41 executing program 0: r0 = socket$inet6(0xa, 0x6, 0x0) bind$inet6(r0, &(0x7f0000000000)={0xa, 0x4e20}, 0x1c) listen(r0, 0x6) socket$inet6(0xa, 0x6, 0x0) r1 = socket$inet6(0xa, 0x80002, 0x88) bind$inet6(r1, &(0x7f0000000580)={0xa, 0x10010000004e20}, 0x1c) setsockopt$inet6_udp_int(r1, 0x11, 0xb, &(0x7f0000000080), 0x4) r2 = accept(r0, 0x0, &(0x7f0000000100)) setsockopt$inet6_tcp_TCP_MD5SIG(r2, 0x6, 0xe, &(0x7f0000000240)={@in={{0x2, 0x4e24, @empty}}, 0x0, 0x0, 0xe, 0x0, "f7e7cc8eefaf1c185361d9fd84f2f5f5da1fc0c85899c5fa73b0962b36bcf8d02916687c145ad91a568951b903f26927ecda656f85ee50ddc5788f8f42770db312d7023cc7b4fba91b852d9bf7ad6eb8"}, 0xd8) getsockopt$inet6_IPV6_FLOWLABEL_MGR(r1, 0x29, 0x20, &(0x7f0000000080)={@private1, 0x4, 0x2, 0x2, 0xc, 0x81, 0x3}, &(0x7f00000000c0)=0x20) r3 = syz_init_net_socket$ax25(0x3, 0x2, 0x0) bind$ax25(r3, &(0x7f0000000000)={{0x3, @remote={0xcc, 0xcc, 0xcc, 0xcc, 0xcc, 0xcc, 0x1}, 0x1}, [@null, @netrom={0xbb, 0xbb, 0xbb, 0xbb, 0xbb, 0x0, 0x0}, @bcast, @rose={0xbb, 0xbb, 0xbb, 0x1, 0x0}, @bcast, @netrom={0xbb, 0xbb, 0xbb, 0xbb, 0xbb, 0x0, 0x0}, @null, @bcast]}, 0x48) [ 2833.640589][ T3194] bond1292: entered promiscuous mode [ 2833.646818][ T3194] 8021q: adding VLAN 0 to HW filter on device bond1292 [ 2833.769815][ T3195] bond1292: (slave bridge1255): making interface the new active one [ 2833.789792][ T3195] bridge1255: entered promiscuous mode [ 2833.800818][ T3195] bond1292: (slave bridge1255): Enslaving as an active interface with an up link 04:12:41 executing program 5: r0 = socket$netlink(0x10, 0x3, 0x0) r1 = socket$netlink(0x10, 0x3, 0x0) r2 = socket(0x10, 0x803, 0x0) sendmsg$NL80211_CMD_CRIT_PROTOCOL_START(r2, &(0x7f0000000580)={0x0, 0x0, &(0x7f0000000540)={0x0, 0x1c}}, 0x0) getsockname$packet(r2, &(0x7f0000000600)={0x11, 0x0, 0x0, 0x1, 0x0, 0x6, @broadcast}, &(0x7f0000000100)=0x14) sendmsg$nl_route(r1, &(0x7f0000000040)={0x0, 0x0, &(0x7f0000000000)={&(0x7f00000000c0)=ANY=[@ANYBLOB="3c00000010008506eb9afc4cd8d06e754a0081c5", @ANYRES32=r3, @ANYBLOB="2377f292252155b21c0012000c000100626f6e64000000000c0002000800010001"], 0x3c}}, 0x0) sendmsg$nl_route(r0, &(0x7f0000000240)={0x0, 0x0, &(0x7f0000000680)={&(0x7f0000000400)=@newlink={0x3c, 0x10, 0xffffff1f, 0x0, 0x9000000, {}, [@IFLA_LINKINFO={0x14, 0x12, 0x0, 0x1, @bridge={{0xb}, {0x4}}}, @IFLA_MASTER={0x8, 0xa, r3}]}, 0x3c}}, 0x0) [ 2833.823420][ T3199] netlink: 'syz-executor.2': attribute type 1 has an invalid length. 04:12:41 executing program 4: r0 = socket$inet_sctp(0x2, 0x5, 0x84) setsockopt$inet_sctp_SCTP_DEFAULT_PRINFO(r0, 0x84, 0x72, &(0x7f0000000040)={0x0, 0x0, 0x2}, 0xc) (async) r1 = accept$inet(r0, 0x0, &(0x7f0000000100)) setsockopt$inet_sctp_SCTP_PEER_ADDR_THLDS(r1, 0x84, 0x1f, &(0x7f0000000140)={0x0, @in={{0x2, 0x4e21, @empty}}, 0x711f, 0x6f}, 0x90) (async) r2 = socket$inet6(0xa, 0x2, 0x0) setsockopt$inet_sctp_SCTP_ENABLE_STREAM_RESET(r0, 0x84, 0x76, &(0x7f0000000080)={0x0, 0x9}, 0x8) (async) r3 = socket$inet6(0xa, 0x80002, 0x88) bind$inet6(r3, &(0x7f0000000580)={0xa, 0x10010000004e20}, 0x1c) (async) setsockopt$inet6_udp_int(r3, 0x11, 0xb, &(0x7f0000000080), 0x4) (async) ioctl$sock_SIOCETHTOOL(r2, 0x89f0, &(0x7f0000000000)={'bridge0\x00', &(0x7f00000000c0)=ANY=[@ANYRES64=r3]}) [ 2833.960889][ T3199] bond1284: entered promiscuous mode [ 2833.967690][ T3199] 8021q: adding VLAN 0 to HW filter on device bond1284 [ 2834.031002][ T3204] bond1284: (slave bridge1205): making interface the new active one [ 2834.054918][ T3204] bridge1205: entered promiscuous mode 04:12:41 executing program 2: r0 = socket$netlink(0x10, 0x3, 0x0) r1 = socket$netlink(0x10, 0x3, 0x0) r2 = socket(0x10, 0x803, 0x0) sendmsg$NL80211_CMD_CRIT_PROTOCOL_START(r2, &(0x7f0000000580)={0x0, 0x0, &(0x7f0000000540)={0x0, 0x1c}}, 0x0) getsockname$packet(r2, &(0x7f0000000600)={0x11, 0x0, 0x0, 0x1, 0x0, 0x6, @broadcast}, &(0x7f0000000100)=0x14) sendmsg$nl_route(r1, &(0x7f0000000040)={0x0, 0x0, &(0x7f0000000000)={&(0x7f00000000c0)=ANY=[@ANYBLOB="3c00000010008506eb9afc4cd8d06e754a0081c5", @ANYRES32=r3, @ANYBLOB="2377f292252155b21c0012000c000100626f6e64000000000c0002000800010001"], 0x3c}}, 0x0) sendmsg$nl_route(r0, &(0x7f0000000240)={0x0, 0x0, &(0x7f0000000680)={&(0x7f0000000400)=@newlink={0x3c, 0x10, 0xffffff1f, 0x0, 0x253c38, {}, [@IFLA_LINKINFO={0x14, 0x12, 0x0, 0x1, @bridge={{0xb}, {0x4}}}, @IFLA_MASTER={0x8, 0xa, r3}]}, 0x3c}}, 0x0) [ 2834.081137][ T3204] bond1284: (slave bridge1205): Enslaving as an active interface with an up link 04:12:41 executing program 3: r0 = socket$netlink(0x10, 0x3, 0x0) sendmsg$netlink(r0, &(0x7f00000047c0)={0x0, 0x0, &(0x7f0000000000)=[{&(0x7f0000000100)=ANY=[@ANYBLOB="18000000560001060000007e00db730007"], 0x18}], 0x1}, 0x0) [ 2834.258533][ T3213] netlink: 'syz-executor.1': attribute type 1 has an invalid length. [ 2834.348551][ T3213] bond1301: entered promiscuous mode [ 2834.355360][ T3213] 8021q: adding VLAN 0 to HW filter on device bond1301 04:12:41 executing program 1: r0 = socket$netlink(0x10, 0x3, 0x0) r1 = socket$netlink(0x10, 0x3, 0x0) r2 = socket(0x10, 0x803, 0x0) sendmsg$NL80211_CMD_CRIT_PROTOCOL_START(r2, &(0x7f0000000580)={0x0, 0x0, &(0x7f0000000540)={0x0, 0x1c}}, 0x0) getsockname$packet(r2, &(0x7f0000000600)={0x11, 0x0, 0x0, 0x1, 0x0, 0x6, @broadcast}, &(0x7f0000000100)=0x14) sendmsg$nl_route(r1, &(0x7f0000000040)={0x0, 0x0, &(0x7f0000000000)={&(0x7f00000000c0)=ANY=[@ANYBLOB="3c00000010008506eb9afc4cd8d06e754a0081c5", @ANYRES32=r3, @ANYBLOB="2377f292252155b21c0012000c000100626f6e64000000000c0002000800010001"], 0x3c}}, 0x0) sendmsg$nl_route(r0, &(0x7f0000000240)={0x0, 0x0, &(0x7f0000000680)={&(0x7f0000000400)=@newlink={0x3c, 0x10, 0xffffff1f, 0x0, 0x19672500, {}, [@IFLA_LINKINFO={0x14, 0x12, 0x0, 0x1, @bridge={{0xb}, {0x4}}}, @IFLA_MASTER={0x8, 0xa, r3}]}, 0x3c}}, 0x0) [ 2834.397373][ T3215] bond1301: (slave bridge1247): making interface the new active one [ 2834.405751][ T3215] bridge1247: entered promiscuous mode [ 2834.417634][ T3215] bond1301: (slave bridge1247): Enslaving as an active interface with an up link [ 2834.427157][ T3219] netlink: 'syz-executor.5': attribute type 1 has an invalid length. 04:12:42 executing program 0: r0 = socket$inet6(0xa, 0x6, 0x0) bind$inet6(r0, &(0x7f0000000000)={0xa, 0x4e20}, 0x1c) listen(r0, 0x6) socket$inet6(0xa, 0x6, 0x0) r1 = socket$inet6(0xa, 0x80002, 0x88) bind$inet6(r1, &(0x7f0000000580)={0xa, 0x10010000004e20}, 0x1c) setsockopt$inet6_udp_int(r1, 0x11, 0xb, &(0x7f0000000080), 0x4) r2 = accept(r0, 0x0, &(0x7f0000000100)) setsockopt$inet6_tcp_TCP_MD5SIG(r2, 0x6, 0xe, &(0x7f0000000240)={@in={{0x2, 0x4e24, @empty}}, 0x0, 0x0, 0xe, 0x0, "f7e7cc8eefaf1c185361d9fd84f2f5f5da1fc0c85899c5fa73b0962b36bcf8d02916687c145ad91a568951b903f26927ecda656f85ee50ddc5788f8f42770db312d7023cc7b4fba91b852d9bf7ad6eb8"}, 0xd8) r3 = syz_init_net_socket$ax25(0x3, 0x2, 0x0) bind$ax25(r3, &(0x7f0000000000)={{0x3, @remote={0xcc, 0xcc, 0xcc, 0xcc, 0xcc, 0xcc, 0x1}, 0x1}, [@null, @netrom={0xbb, 0xbb, 0xbb, 0xbb, 0xbb, 0x0, 0x0}, @bcast, @rose={0xbb, 0xbb, 0xbb, 0x1, 0x0}, @bcast, @netrom={0xbb, 0xbb, 0xbb, 0xbb, 0xbb, 0x0, 0x0}, @null, @bcast]}, 0x48) [ 2834.551335][ T3219] bond1293: entered promiscuous mode [ 2834.583818][ T3219] 8021q: adding VLAN 0 to HW filter on device bond1293 04:12:42 executing program 4: ioctl$BTRFS_IOC_SCRUB_PROGRESS(0xffffffffffffffff, 0xc400941d, &(0x7f0000000240)={0x0, 0x7, 0x3, 0x1}) r1 = socket$inet6(0xa, 0x2, 0x0) ioctl$sock_SIOCETHTOOL(r1, 0x89f0, &(0x7f0000000000)={'bridge0\x00', &(0x7f0000000200)=@ethtool_ringparam={0x10}}) r2 = socket$inet6(0xa, 0x80002, 0x88) bind$inet6(r2, &(0x7f0000000580)={0xa, 0x10010000004e20}, 0x1c) setsockopt$inet6_udp_int(r2, 0x11, 0xb, &(0x7f0000000080), 0x4) sendmmsg$inet6(r2, &(0x7f0000007600)=[{{&(0x7f0000000040)={0xa, 0x4e22, 0x7, @dev={0xfe, 0x80, '\x00', 0x1d}, 0x3}, 0x1c, &(0x7f0000000c80)=[{&(0x7f0000000080)="7404edfae53ede7d1a09859a6b3deeac57e43d596e3c4306791801aa8f33af74978ffd7335d4b3b3322941899ef8d99ac803391fc661e08d67d096e4cdaa4a67392a3b79bf47a9417b7587a9a1d260eb17554b01d84a38548b1015cddd01b5e16ae9c0e2587d4cf47afe4ad5ab01bcffc6ec361e85f1937b18d4a97da8e562ec85858d6b16ca6d146c4132b5d6cd1cbb867e866fca85b2716e98bf62b11225c17af3dc6ffa8e646dfec27b19025c69ae34499e23b267db917dac85556d0afd1fb6f2197caaf7d6b1916a6adb88c1470eb32c278a3e2ebff4f7", 0xd9}, {&(0x7f0000000a40)="3ede83855b5c4c5cb1e3f189b0fc516401361f677c3432c5a0a53cabce5d17ef6abbf132df751a14d38d4c86fe6ba04579030e2436a835da9bdb64a18fcfbbf7183316a6aa33d215a8c18fc3a09ea7ab4576794d8f315684b0af9ff8f144b196a9b1086b3b20119349abb79e0df941576e1376e04395fa98cb668367eb0d2dc180d15091a7d91ebfee8c11922f045772799ebdea6dbe1c4f5aca9647d2a03e", 0x9f}, {&(0x7f0000000180)="f484cd0060ba8f6139a6ea0eb3f6a338a6677b7298937e8283f582027f7080e6b4a612522c23c0bf2d2fe6a6ec8a56004122ce57731cbf3fe6c946cf1ae4", 0x3e}, {&(0x7f0000000b00)="6dd61a68d3c0fc641b5eb13d48c2a6c3d097cd5a44d96549275ec0f3a236c5315a14b665a8a3e3e5c1d34ea134e7e1546cf8537945e1906259bc856d5d6466b9ac37e5b96f1722237f2b00bddb9046d96400cbe8d52d550787cbab808c5068ea832ffbe112548e4bbf253908e5a45cc4d2d8b31664f5bc9f0527bbd3b3601f328979d61e6761132c10324de1a0676c97d8f8e3964eef59c99a777116c41d90eceae5308fdfd7aae49d7170a798fe6f7da775e4328cfc31938aaa30aee7a6291254796b3252240e7ec9711429dceb10d9bd880620271c6feea17780f280a0ed9fe092", 0xe2}, {&(0x7f0000000c00)="db27c4ff6a09ab2043e1d2cbfc5e53056ef813b840fb73c98b0be1ac0c8e3acb9a0c3460c890a4c6e5a227a85a0d38eec453787cef082ee62d259248562438700f467ab5689076ca6f5381b8d89a252f2e5555535992b1", 0x57}], 0x5, &(0x7f0000000d00)=[@tclass={{0x14, 0x29, 0x43, 0x2}}, @hoplimit={{0x14, 0x29, 0x34, 0x5}}, @tclass={{0x14, 0x29, 0x43, 0x4}}, @rthdr={{0x38, 0x29, 0x39, {0x1, 0x4, 0x0, 0x6, 0x0, [@mcast2, @remote]}}}], 0x80}}, {{&(0x7f00000001c0)={0xa, 0x4e24, 0x3, @empty, 0x7}, 0x1c, &(0x7f0000000ec0)=[{&(0x7f0000000d80)="8243bf28d847c328e11f2f250fc748e7454f01", 0x13}, {&(0x7f0000000dc0)="e98086afbeab9bbf0c9ebefa6d493c4b5c589e774d12f869db3bdfc70618b78fc72cb423931a0b2bba3b4c746d2d68c0cb151628da47f9da4e8cf5342952dcf45e1f76a285d0efe301c306700d23abd38ef37b14de999a69d2d30efc9ab8d7619317cde3758c4fb8be5f8faf0c97cb2cf1937279a51a1d41917f8a42bd986eed6848f54f6b28dd9496305406010682dc04da27f4a9bac7c84ffefa628aa3fed21ed48d0ebbeeb8b7f3923bfcb7e801f7b024fa930a05bee6ee5c853bf9461c0854882e7126d4f182b4b90e4431361b84989c3aea08", 0xd5}], 0x2, &(0x7f0000000f00)=[@rthdr={{0x88, 0x29, 0x39, {0x2, 0xe, 0x2, 0x1, 0x0, [@loopback, @private1, @initdev={0xfe, 0x88, '\x00', 0x1, 0x0}, @private2, @initdev={0xfe, 0x88, '\x00', 0x0, 0x0}, @mcast2, @local]}}}], 0x88}}, {{&(0x7f0000000fc0)={0xa, 0x4e21, 0x20, @ipv4={'\x00', '\xff\xff', @initdev={0xac, 0x1e, 0x1, 0x0}}, 0xff}, 0x1c, &(0x7f00000031c0)=[{&(0x7f0000001000)="5fb906fb71ad76120c1c90e7eedbf0acef97ed911df0fd053d1391a9f079a43751f62dd438bdeec3123f26cc73f874e789c29b6d866eb07df5f18002b214056c2662a2fd0020b430ca9b6f3ae85ed7a61a73250935736c7f14167a9d6fba9d7b01a315fac51b2be36decdd98affbc9e1e3383e9f71e27d239fbfbeb41fc5a84694dfb792a606630845d883f0f6f15e0a2cde0115c6376d0fa81435353aff60a08b8ee81ad7d4d5fc0027c04594f7414d6e1de1b7fd1b8250273d63f2b42295226d6871cb94f7d0af49105038cb257a442e7f59792ee843b5a3cbcce7f5ae361d3e36550c5e3872e090524f9223949de4f2a696d14a9b5056095af7c414b98cc8b429c67cb4e1f991f29e6f505f3aa6956d733b49ba9a85bdd634feaf1613ebb8f4d7dd3d939510d2eebe9a40679b156343efb4284281f93a206bde9d1986f0dec3c5a263749f25404421a6a495cd1f302662ae176a2e8974fcdfb197153058b2f5a6695932875060dd348d1d233fa9807dc5a98507e8c6fe726167f276499633b4b7463b392abadfb7b68da12aac146004f46e679e3ed18ffad84afc95100fa09bab6944d25ea55a78f10a85a063ddf537bce2ef1bdd57b101db4eb5ee4fbf0a9d6c184342a119546efbcb71ada2db4c1500a46e837cf8c7606d0be343d0e460338a05957274054c092cfa3e0eca9fb8d2f845993d9454f7782c8c3a9b823e674d2db4a924b4946827b3b8f521c517aebfc9542dd799bbb6607290faa4743acf98934892db1bf744e7efcad821c07edc0f9f18de11036b950d92eb58d4dfb7c845730e9e236ea6c6cc56287de14f6e169da6874dd5fb442c37f6d18ac40af97fe92d5d20c4e04095918cc69f7306dc474ae33886492f8c8e746edd7d96c776fa76aa9ae243f6acd93d982ddbeaeb301ff64f25cda9eaa7afd138b840b52f183bc3a7f16f88526e9b2b6913335b6cb72913c1b59eb83eddba71f28e793f3ededf994cb6a80b47c1c59edf73feb0b831a342dc0c8a26c0c4b8132db0eb3656e9702bcd924b74e6853ea0e35d01d5a1da103f359fcbcce084ad603f61343d6f6a2b47bb131c6e1391551e38712132ddd0ac842ee4b2c4a605bba68aa56c18b6e785163194e924a84804c7b1ab5b201fb9684ff20c767a1fd19b10b3d6bf8658ddbc1364ac761784e3fce85d250fed72c4aa6fdb45593973c4b020c31b3c366d17223f64233b2ecf8c37cba64631c56aa066f2257718bb499e17ca903bddb35ebed11e2720cb1b5a027b4c0eae4519ab2bbaea2bd3dd643d30cd441b364762dbcaf3e598b2996ccf2aecd177dc4e0dd5603f9d71be5b9d11c76a7adc012963b31fadd07c1ae825099ebbeeff10b1b9e9bf5ba542d2be91d145332606d1ea4847ebe55a891746b4e09f7b6ea1cedb974bb6ed649a35318945767f17eae0f212e44176bca8797129723bfc9e6ac402981b65a1e72aab2efac6c592306dee6070f2e76d199c722753ddfb74b772fc530ff403fd85bb9f17651d70a871d395ee2b047fecf254877feb7f6bdac80af1e5892510b0b559fee90b6c7b5b287498c0922bd911d723856a0ecedceb63d8b93c014e1bd7eb9491fad5e986eda95ca1eba6ff3a39f93ab75334b38e850694685d9d323bcd70370532ce87f00f0d38678325d174712ef9c9f64b0a03fc236c991797d0f167996ce58895b1fc62e2a302a15907544a6dfa92833cdea659f6417c742f40fa699008036318b1971329cc7ca67930e33606402515b53eaa83cc54e2bcd67867e1357f56c1315b9070b327e1840bb08679955509bd83f7e1a2e92b1e3b5fc90e5537714732af072dadef2de4034077fbfeed9ed3b676bc8b6372a46c9f18a1962949262698e718950ea1685d33a9f8642eca180614feeb5b108bcfd46a8f0615be219120545f702dcb192a3f2f54f450586bc8766dfd9dc19a949a95cb9d62d85e40bcd64911138a83760845d2f3e9d33209807db5a21dee4d9e54b8c20a5f67892b06b13272054f9e39dd54efb57bee4e321cae1db7b283d86d324755f6e97be34d7352036910ba24c03fa9cf84aff230cbbca8145721e8c70929d4c70cf96d4670e5124dcddbf571e9311f30d2ee111990b1d41fbd1a2ce422d8129651a4eace61da1d8952e60cef9f029b4497a06c4c79e819c3905c4761e89076ee5a6bf023eb54b1a7ce4b615c7eea5482ee7e77ded251596d3c4ec21cacffdf9eed00d3a141bf5d6488bbefb269438eb4e60487f1b2616cdb79869304af7043f47e60e6c8c900498f59756a72fb6bc1e74fd07cd1679b0349c69b6ccc36dd2cd74bba95388216f77133c22ff20783fb06bd84cff2fbb3e32a260c8741f5725c54c6f7e659eda998e3d312c55f203343e4c68a4c5adeb2859db2dddf17a82bc2de14c52ba4a83106ba89330dac4c05f3d981d0fa5307d2de60d146fb1dfc7a7bab1d9f86dec55db77dfedd0f9224ab157ac746326781283847015c61fc06528ebd620ebc815196e4a3a6ad293a1c895c3767c61518c72bd0699355e32f5c38193215880806b85f71ae748dbe2a04ea31cff5aeb65e1e7dd8ee8575a5579dfb5f84c2e9e707e6d5b4721ed461ece928a658ced136dae166d512a3323ebc064bff652c4de6bd7108239d3709f7472c91f54ea504d97c1ed2bdcd873b17539d19a620b15d26b343afceb251f4485cf91be16d4e47b883d755cc6d2dbf25a196742214a854ad1918214bca42bc1b28a952a5054c2b26b172f4bd96ce2e22512ef59ed2eadb496debfe06e50ddfa9f941e85b17d2651d976c1e9f03574bcea247e2d0903c5a2e3b0fdd1bef735bc97126424d6f94f1d18f5e4313599bec2bade7dc6d1fda2fd05d6d325aaa4c50fa9ad95a414338515cee9e00b39c552a86966d64df5320e2a9638c1a2e3a794d79a46e05079ee108b9d4c26174c9d05d55c854d9b23dc963f5264331cba5c97be78e89bd2c3172d40277b9bbf4728c4bb884c5fbbef8fdb89f97b4ffef27f9af22ddbea2909706989960a58accb4f97f6be0d7a27f4fba9628c5d0d07a8c19e897963ef29b959f283cf8049d1c0a011310f45cd8b9cf6d22e8f94334ba54281d718fe5252808182f45607b84fdc551d6c0ca98ec47246533f4efb1bb83170b39a11dbdf0dceda366adb009ef48cd807aabdfd76ebfa7d1689ea0ad638fd2c5873dd62542e98e150c4c29eef20842fbfb9dbfe4829614667a4bbf38f727e66c614c4ce12ac034a92689ae10f3d00c872b67a3c69c39560964f50420a4ba3dcf62d6d39c603e60231b36b0356a7a0080756b761feab0253e68e67f317adf72a7dfef797469547a84fccb764e7b78cd012a77ede0177a8e402e56540e9e5e9b20b686d50d213226be12d067c28469827ddcdf7e15d8d4a6200e9e52bf17384041b8aa70ca097275fe70f44643a8beff01be4035929d765d7e863023cdb90f3bde63fff91d023187a0a3fcc96018475d8767d978dbe58a5555feecb6699e2234cf481ded5b8d3bec293faf12c16f890a52238d3e3664118d4bc919f05fd7a299c2c9d6a5fab9dc0eee8aec75f57ad0bc91eb29d34a485691e2c225654b3abd4afe162346118a87e33f8c2cbdb3f68df5148eb27b427ffdab10f29475ad17d8ddec3025a98b22b30bbc27fb6123bfa711889e051db56cf07aba5256458e5059ee5b4ae791b96c5a45170ff845c244d37a5fd28dd756e1892fa9545a38a9a925c41dfcc55147215bb2e09c1df454c23ab3c5a64cf339ed75530967765c81f0bee2d0f7804e841d76b6aed9e361b433e0089e9681fe497d473cd9f39737770f4e887ae84cc9f2becfe93180fee9dec22b9b8697e7c5f5dcfa2e3f3f3bb0c94f40400caa45971a6c7c90eeee5e924035521bb33a28f590dbddf58251047fe33bfba2f3e9b9cb31d2f9b0b688c21e8e7be2a3add2bd337eb4884931d78430aaafcff9da037b0ef972506fc12723a1ea49ff0e20bfe1bc3359afed16337bdd79370d992fb5ac6f6c1f6da011bda39d64f6e559589810fc184bc985745ac7a9c5b3926d98c6742b609efb0586077c4537969255eaf90acc134f80c1c062cf313de92a40dbac1f6e011716335632036a2c79b6adf0859f8ced80776517d6a13e990eff0a169c79e953784e986f3f18abc7fcf13f583b6e9f383543b41f955815dc93a687e88d10f5689ef7cc38a469e5b2f0dc4fb3f2d62de06c4d0f1787d50d873c5d4c510d6cf5e62702b84110a7fd3432d1bd1cec285b6fef129c9f8c411ea9797694d45dabf3ff0251db6577984728a23bfd60331f4b69a7307737e6ab90d79db6c107adccdb35af3b28ed2da1e449f10d5760b37b56f484c7a56eadde8c3f0c7dce573eac3d4a6f46d82dc6463d4c02425fbcbcf4e3d47dadb96a2bae03750690a7232f3d0bb5f45ab80f61d0bd2495702359c19abe73b730ddc8ac6cc87ff4e0916b0d401c579b33bf4eec208227cadf3fcf8e2761eefec22625eaa8b1820557a2de152f2baea18c348c1337c5fd402088a81bba2f5544f208ebc684a6fc20eb4c8a40cc270c937a5ac417db1b59b00e0659756466d7f1cda12092dd138fd8b5fa4c79431d4f3c1f2dac347c29cfc2ef23f39e01a6e35c7b0c20b76d29eef315a2b49d3aac78e5a6c07682eef5528982142a555cf571463f2acefcae1056762ab6e6adfd1109399062f836d08a5203393d58ef72c1b730dd7b6728dd6a3389321a01a54a96c27115b594cd401fc1922831097f333d182764c92a3146297948a53549758825b5cd2e63412c493c7f57aa3d79e062bcbf3d3e8b8f94917b6df48f4a27d4d09330c427f105ec4d56117f0301504cca747506894af3caca99b934640ef57bfe93805b9fd4e4c9880da8675eaf3bee1247777827018e54bdfd788faeb84e1c471ace504a7910670bbc3c8e75d546940ce0768c543bbd9348180484f0e9b279e3667ded2423d324fd098d4585a57feb05319ec7a137034e721538cecc06bf6c87541f31f4ea4b38623425df336ddd233d72e1bc9ac556cf97ca88f9116aa23b4bc493691277ce38e2600033c6dea8bcdade7e5648a8eacd4358c6b04f8b4bc5e459e9b9e91e8c0f26129c05474f062a00ba85d50db442d6ecea983726fc8fb47332406a03ca40c2796eccb545056968c1f2cde6446549ecbc21da3618b1eb46bc986bfb652bb93178db6d7b4beffda22ac417f86a959fe3788c628e4d5a3ffd8a14e3bbdeb6103907179cfb9420cbea5fd2112e1409ff92ccfd13380e82df79fac73bd23137635f12e293e8b3077ae9dabbd2fe2e13257c2f8ab11d20e1e386b0637caa4ca227dd1670e3e4c4bf78dadad16540144a0e7cd67f0ee5e864bbaf17b4d22bfb2e4a3b13705658577422799cdd991239e0b0d0de5e55112cbeadaa4a4a065c165edc31eaf17a96ff6be007a123b5600a3309ac4c7138b52d1257ecda2a1a9d84f5dd66f7121ab6d342294c6cc7b564c609d5f9381406b2a7fcec40d779a02b95843731cb3e741b4c5152eba9e4797f95d2c7eed2d0e5f66b2bb36679ca2b5e483d9c19760f24137471a05d0773c84876cd54969bab649f954d954838ed352cadfe63df02391f8bc057da723aa33c478508e85663815f4131dd965410159f3f91563549cf4b55bd7688624135b905795dc0cbf5da7fd62232dadb4b7c0b7ff9ef33ff73d3400ab41e702eadfa266905f4406cf277d0a447f3b93b766e5b2d0414f81396897ee66d942353cef68f2df92d8a8208388b000411a69169946a2ee38b4a47bc8c1b", 0x1000}, {&(0x7f0000002000)="7252ba8bc1d75a97d4f12549ef766b80b2e4c36d1f11d984de5807b5d9b7be5badfe5d1d79f1bcd48743507c2fe1e6bc93afb1a18bf8aa4f1141382e5df43a20851f2ab3b8353af0ec54cffc54378586ee8c9eaf3d4d8cf6640aa94fd203acd514e78e44f7217406320ae0723dbf749ebf3685b7f12db16702dc7aa938b1abef9c789602bec343e1f0d54fca9b7146884f0c4da57a0d4011c5cb61c52f36a5469a455746ddfbe4a5db43f26ba9826d6f61ce1ed9a0b46ba516f243d048fa2211307d06109403e98f8e16770036cf580f5d6db8c2dde148bc0579740913cc3868bea39a497f8942", 0xe7}, {&(0x7f0000002100)="e0949db7f4ef749c4d9532615e3090e995eacb74b40b8c8f6b2566685cf6b1c677473e380dabcaf9031d076b23fcbfb108952dd43a2af98cbf95d25f2524925a2f86c4be9f22175c1612f0a5ecea1fc256c0d287395bde0c4b8f3a9d328a00e12096836f20f8ce2f6974ba144c98cd1d3fc176eb37f066f52ec2659d57c00010571b878831bf2d01c2b85f", 0x8b}, {&(0x7f00000021c0)="e39901f38cf493b7c80ba34d6d55f3ba3dfc148fb0eb59a7dcdd4f0fa3ea59c81921a08e3eb6be97db3bb6321ef9a212d8b49d32aa26a2d0fa01bf22723c432241f5ba018fbcf6de10bfd692a1e92b38f9ad8def34e79070a778b6f446c1b1e8f26e19b11fc190afe9d0a75bff256751a45bddf56954fd16fe13b3afe9be1b98435e5a2ce95a2d7c52b5d1069bacbf1ffb85c414c0eda68fe7d70c77a4605890f52499b751f3567ed978f329870d606369b129d83a74c7f4656ffa1ac2d6741e3728e3ad6d3253c375e552a53fe9d2776d27fbe2c1d698dfec48a1b6bd1e297cd2bc6e3b844d56934a88ba2a4b14ac208af2672f37e26be9d429624dd33f22361807c631d346e4debd8b1e3baa1b6a2dac341531d473b126667aad2f4617a504fc752b7283ea0bdec706910745ddda451ce81df4c6c33cac2479ebf380aaf0061d8203703656fbb6b88845b743e642752c2267f7c1045bb5159676c7bf53502e157457c6a0e0d62ce83d7b74b789a2a2bd4238b4236549dfbb377522683f362c1c06200d2c0e9b985b565a91aacd64f6db36c904ae56ca8f4b718e71bb0651d5113f560c0fe3ffd70ceee7a07b8df0e7a76d24cf87a902b4eb913ad10bd0c965dddeba3ca473dc8698ffb879a405e489a917bd80656122c75fc27b85079e1a63600212c75675e2246e586d081e1a14da434903ad7392f9218c4ea6c2966f116bd292889403720c647693bf35adf963b490478e8af07268fb3ad49e0ee096058af003f5e7b80dce9032b209e7cd79a2a5e61c0caee2a4e2d90d307b4b1c9176a8625ba994dffdb4a3a1523c5091d86af016749e8bf9671629b77a40a44de950260e2e15f3ec54bb23c1b8e4e31db20e8d2b2a0307b1f988b20a374d1e5e1052bbc35dd0766be1fd241b7269931b0b510fe0ba38cb33cedf95991b78cccdedefa5eb953759c1a435db4e2104305ae73e4b839f2e8af4a053a74fc3e22d612c75e355c01e44911758ec8493ed3af672216dc1bbd42490a189da994baea138bf9d27e70444a6db92de757d31c91631c6bd15e92417d3b5fd04cafc7bcf126877aa0affdd0154a0ca10d39980782ecff7d37b1cfb6ac209ca6771d470ae820b21189bdd74f5fbbd4336dde08d7adef8b20d73b2ebe5b6be51edbbd8d368c705f6f67461555790fc38372d61c85bc65fc9b2019bca8ae1764be788cc04a0108fb30e83f8dca571d1dd45d3b4d6b5eac0b716fc5b245fe9669d7a229e6ee72bfbc2b70a02569d87f1bcd74fb99b3f93fc39566ea4e152963aa88e219e0665e93d0b116e92e32a6e222f85971edfd3bcb310631cc7528df95d1c5a4b71c5e9c384ab1554c0db65cf6bcbcd3be5ca379ab704cf33258d7542aadaf71423f9b655792d9f3e73e2c800322c39e530f4dd596f9cb6795db108f2fda6222ae60ca3851c839bce543f3127b6b5d411b8e41b4b9f585e517824ad89fda37ffc65d45284fe985cb159d7370f36e551ff718dea5387f0abc2b39af55b5ddf0b96a05b12817c12166e75c4b33c2ca1765b1346dc87e21ad13abd6402cb22f65ac47b078af23571348494235e8e5483b05e1c8c3a4e09f3679ecb866795e439d87fa852e59581c5410600500b18f7352dacd4a756b4a49d689e260f047f2bd495e58034e65b2473e2b8ca4e763294ee4b2f74db56f6c2a8fc28e3316e23d538583d125e8420a50337724bdd46794d19401c0356a906769238417ec4e217c9f37dd77232e996aa0ac09390f604e023a7ff12868d8dd161ca3d80f8acadfccf11c2d301a9c0ab465f42a10e950ac9973381714ceb3dfe6c5acf1f6d83ca25c70a5c24d49b127d78153cb31297683ea0568f038f520e49f720061d6c2dfbc8b0bfc0eb4b202f2213251792c6f741d1ee3c5489c76a7143473928ff938c97683a316c72d6e35844fe8c25d017bfb8a29532c9f444f43559e909ec507fffe66759bbf584cecb0e4570228e4513c08124d91a17b90c2230de488f38f3f2e606ca69e4f006803d6e5cab9b11c9f37e9f01b86301f7addcfa4624d6747d82552fed7d7c77183415ebe6a9d02bf40588a240001b0c0d13c74944d8bc009296abc6ad16c73bde8aeb81a5c92e9c3b971c4cd4d4a1423772b1c5e5d000796474e38c3f89b1c99e19ec76cbdd7339d064221ad02c54a9e0b3d46573bbaa7c0af49c8e68e6f17ef33735aec1a982af8ffad35235f0944541ef0c35c9824748614187d91219986bb2fd45a88877965e7aff352d2ef4eaa5f1df28799ebfca9e7544cfe93e7810663b956aef8ec6e3087d3fc4b44e1d0f422456f22ff89d5dde9ef5c448b646142f2fc8977834f245ba4528b70e1c4daabb6612b91f96361be441e0081cdf04e5b0b7151ea3acf05efeaeb8a523326be040044c5cc450777cd5cf2ef549c2f712277dff37ce30b7595b09a1e1e384702dda93a46e63f4da116ff67b6d1e8779d99ab460f95dcb0a1ac300e2a1c3bee243741086e1a0254231ccc34a9b9b77a86a86c36b590dea58b2b943b2eb15212dd4e13d796fe23973da9b1c01d346f0098591bb3f37f40b77f31cc7e73b15db81270d8249e03b2554e2c32e54ddeb1b286e6cb874d8a3e0d442dad9574668fdba19abf13f8b170952b4175626ad31fc40b7bea69a9fd4c0fe016bad0a4e988172fa81b226d3695ce89602760009e97b587a3317bf6870a70ed910f1355e03e80cd518be82d7e06f54f8632772247fbfdc0e95068741dc32bb4daeaa3821da0a1b586f78228762eae6195e913b500e5d63d75217e8a5a1c3dd9d4dec8ee798def5c0066c713591b83aa3aadb0887493c4066b2a95fb3b4d316685682b7fc92f8a19093a7d790d7a65b430eb6e7e96b2da3d349f54df43d4cc054fbc041ea9c54dbaa61349f669f32c1cd2f3c98e2d9939da4ed165264568c099fcea1a5a413470b1592a582887bcdc87785b273aee82ca3bd05986ecb860554f01e5a4aa0b1bfbc6ba5df6a3abe73e14da64fbde485f1af83b7c4ddb656f57c8c857b79417d5d51e997717a00138f6a7a0b604999b77f73f0edb9e7d59413de9825fd49533eda109ebc2324f432f2ee563b52780f701c81778816042a5f5706b273635ca05274a981ff742f01df9bfc59c3f2792260521a819404e3111529cdee969e55ef9e2c6ca6fca90a179d3f20bc3b42b1d69ddef324f0cf04ea6983fdf4e8d29217a496e6fbcdc4627e49eaeefaa9368bd3dc96de5248a8c57a3dc80459f1c65b0a1bd70de3d896521245b5c9952429b68aac94660327616b6d97568c1dd6d11e7c664ac077389185b659309e79954a0fc42b8c21a6caf29fe05486e78b786854376cca6f18cd8b790ef0c83621b469f6e6175a29432e1138782417269a90593ad205b3c691c13134c0f0cacf18a9874ebf3286e2e3a4d6a86f0c8b838642f39b5e7538a68072379a23fbe8f5a2c679da346f67487d5630fcc7d40634c8309299e00c2b352d69c8a11c34b20c1e0999108e9d922b68289a1403c1de9d6806cd4112a3d24c9c34bad74c6efebfd06497a73f6bf6a6a0bebb127abeab1931ad70a0b4bfa9b40be372a7e45abe7b611b5bf712d681a57c0254053fb104389ca5e25865ddf7ef4dca6a2f286524e263ea2c166a687faf60fbc13bf4d623521682b3104e69a1556d6f9996fc25d1fa36f8df5967b8a7b0732097453a897099f5fb4d47a1bccd1db44da1837686afc9f5415f51c0640f9eb90d1b80b75bb11d31e1f553933af3048b77ce77d7abe6e556d7293b1365a357b3a714968b4cf9699eba4f2b62073d9c7588275f243dd4020839b6a1838a19f5ac19e00220294fd914fd10198cf42cd3ee5cc16aba3bb554db17ed771ed2bbaeedf1477ff712b13f929982db8eecee7cc3d9a29b28932c59a9c94f67f89116c33a551c906c5f4d96e5ceaeeae6197b9463c3e2b0456dad025484267f11d041213c21e89d073c6170efc478fb49878003a4f03156e5daae5dfb8f1f04b8b7f8788351ae7c445133137801401cb496d426ab0a05d9d933f6df8d593389ad2d5cc6453d7384007f23c6e3098af6fefe9fa27d3783168e24776b432b872023379b20665c131cc13f8b08bbbc9d6a6d9772f37750602ec66c2447415b97800027a1f62cece33c0f2e3703bc27f5a6a25e261b90452ccaf0bceb73ffe9c91356ee28e466682708b4aa245f7b708ded0e4efa31c33793eff3e960708d7df3f731e61e6f02ff94398fd26250e43452c07dd1984c701e26086fc1e2cae819e8420d14cfd120998212a7b316b39b9c64aedc3138973571ef408b5555ef5c5ea8e35a4d62f8eb93c0aa854ca17533e71d078c3d066e8b05c205d1b9497f7988d40699dcb2e82b3f053517c4781764c9d031a88b0494ad3e29faea4310037710a98087ab3d75e4671d8098854d278dfefae74957e168a9b7e458debefb85e23e009ccc7fa31ca4d23dc2c7925d14112584edfe1a27e8ead18c3b81e621eff3e8d307d2ec47d3ee388b5404e9be3bc9f46be6e9a27c0f65c582df62e1286dfb9b805d5f490e993c278a66536155e06f4c306ad0240cdc3108ec9e8e63741aa1dbc24e521dedb43edfb84b4ba506915183d1873124eba9e7d0776ce476b8b8280e6f662520294cf484b5b155692a818437444a494ab0445a48f9a74f7bb61fec523de839f527e4b116741069da2e6aec6d7835d786205055acfe7be63ba07d8d215f3cd948ebf360d384ba1783f8db18fc6362f3875876c1ceda2dc8441d21e0b430c9c3bf4e3eeac9ce965a37d9f6d126fa8f9a0340a039c61762ed4588a2abb47e0cb88b791e8a9154eb95402eea495a3a49234c83480dd628239d76f058a3302a3ca4f693aa3a994dd18cb5f99af03404889ea713f2def3116588fac94c94fd7e19134668c24cd949835f85a70c7686917f5cfeec8bd1bf3e198bcbcc980bf66d53a26b4472394424adde196c2fd73e452fe6c51164049f0d7eb76fec48ede903d1aed21b1796ca0c869a1b7974e01db412c44c3cb991214ea56d40dc296a4e4b9b39712e17ad8250ad5274db318f7a3f4c373427ea1c440668f9c94bbce7e064aa1cf5f6224fe9e58e4d488c85b1e0e1dc0d72945b1286dbf37bfa4de1f4013e7177c259790f44e081e0def7b9bd45620dfafc5783b0da095aebaf2736547a0955dcfcfb2b34f9719daf8fbbbb89c9bb0a49bef904234d9626a5debfa8731e61d326d635af87120b40e7158d54227462d3c7a88b40e9639338149db971aa99b48c33dc1cf759fb74a9ead522574a8fe5c870f5af1179313228415b0b894a76235efe001b915e0f8206ace9dfdebd90e36b6dc9542f43c680a694ff086017aa46171c2d98ff9c90f3116876aab63d58cfec732be181dd19aa6e68c3897510617b135cd3e578e11c88a106c3204b769ab91cbf454e601f9921de11c68e7c0c52c8916924e866c9d48f510dff5b9d3fb96796cf079fddf3c755c9040601b493214706fd37bf89e0c737305da67bd219384b87bb2807b4c109eb30fc96fde40bfb2e9c0321e0089a24ea72351a1a13a4cdb0e1dbfd4c50d7327458d94f0c49dce905955235b259d7a808aeaa9eb9d8a4d781545b7f4cf77d6140d48404376e08533d269316e7760ab7faed70ed938a97a5294e8f737c56be715e61ee7378ce0bc99b9e53e9143aa3ce1dc768704655e485c047ab8248286886321f7c9d41830ee8cd886201748ec891970d5b8d99f9941c32e5be2f1f77f3aef57c0c32c33fb3bc71e333dced2e5a269c09f40fcffffba2e0", 0x1000}], 0x4}}, {{&(0x7f0000003200)={0xa, 0x4e24, 0x1, @local, 0x200}, 0x1c, &(0x7f0000005440)=[{&(0x7f0000003240)="7df665bb5181ae567a88ed05d75347c1fa484886428a1668ea1f35a7ff4df0e57e7ae2d148340c5239f050aa2e6da57039d0e62164d484f8ea3fb53712b63da4d8c88d5e4a4865d02b0cfd4da18de33c80d601f30877ccd6d1209c3abdc1c2a065d4bd29bdc51208e791c4dfb39c84a8f87fadd9e248f917b5f85c3b5ee321b27cbdf5b46b603822ea9997501311e59d9724e805a1fa064a8b9fb0502f0f0df0aa2e0197c7d3e45c4cee38883af5c2a52818b9744559e71ee031d03a73c71a350af164f871bb9531ee415df25ccaacab14fe439349a865c1db2679736757e0522e890ece9884e2d9e2ee675f8d4fdd65ff991a761f00dfd7", 0xf8}, {&(0x7f0000003340)="c7ba26e2fd4cb119bc254f5c25dc33fc4d25eadbff01de0e0f02699edbef7a257df9e9bbbe6cb085a0562ffe414a6bfc4e516eb5b46ef48a434d5c90df38b1d26d32cf746891854680da8194ca464b5f4e58f1c8043254d4e026dcb18ab991d6d1ddf39ae3971d16bb5a9f47d376ae13639087c6c8a21968df71c4c24ac771718c17e3f92a23ca98e6256904211f69ba66278ee0b7b060c8f0e207645903f334bb3a8647ffb7abc8b6b582d50497dc76e3c587e85dfcf30beb9c5c261a", 0xbd}, {&(0x7f0000003400)="e190e89f605f00bd91f69ab4666bd7ae317b474c1e6c78fab735984eab18644c15ec4a3a993e9f96e9bd9837d93d32e5ead1c84653cafd94d76a36d7ccee1d91367c55b74dab557da03766ab6120564244fadaa27f3fcfc00189aeb15d9a6afc8de532466999fb732a42e728cdfb676ed47375f5ee81ac6342123963db2f79c622ea5b63c5836ef02c24246fd93ef0eaf2a3463c48c4bdc7296f36099882cef0a39558e946fa98102661d5332c13a696ad2524b4942bdfa45946710831f3ab5613b1806c454384ebb0137d17fc3b181b9ff9bd5433fa09ebfdbba962ef61d4a1cd1c6432b8803ebf133847d88f660041d5491ec2c7d6b99fd12e8259a724331633a587284d31ba1cc9b0af650eabd386797d149097e00dda9fd9efb2695fc7120c159cfac8dc612c66186dbb3944afb55f27d1e8d2be27a538e3c3e2c278ee8215fadfb9f082cc945cab4f9324bb27a5cc2b3ffae06564c9c2f91e848c93f190b60b140fad71a3c732bd1edfa09ff85c585791dd5a68dd8fbcd3a8a804f474dad52c53c724778f7bfcc6f41345c997d95372c7c0635e9ee40f2d03fc9cb3fbb7a418854488f9de79ed78cb5ce6cc7e1c158e7157eb2409a8129000b022a738248f96e1e22145b180fa8844b2539deff9a7826735db39cec8380da35d50ae0d854e6f6e26e5b8418ac67e4f74a93ac122fabee469650d541ee6efead191203cd21ec9e30159e5ec75f9707555cf563897a7ba6b50cb679a998e8e73f92c7f1c609d5d763fb46279f4ff09dd11d1d3192e399bf554ab29a5520f6e7dc7be2d796bf0e66ed1a07d86de5bbbb9c77a9eefd2479ce23cb5e38eb731b8d169a3cb810835b4cba05ae023f6bf5dae84ef6342220cfe57337c266b5afdca0bc9f705eb0d48a2922b1cfafec64c540b81e636d95fab9e72a791210c67e5339ec101635bcd56a929146bb8db3e1b76d9651df2f63fb27fbb048358badf039b2a28f650c90c0147d4daf0e79734eafab8cdab89150c97058ccf64cda9f39234eb20e2cc6312e303e52cb8492afaec5d7c7c9c4098ff2895abb6eaa3806af9fc88e8f6772a64b852a82a64fd1d7b80bb2217e438b7eaf8437e782d398946bb6a78a7487bac4b482727b19450524bad0ca57034449abc5a5bb2aa39be4e76afa96ccc73a574e64a9ecdadd4eae8b88b56460c2ebf6e4fb40ea835a8591d580cdd1560ebaeb52378cfc52f6e70005767e11de0147751fb854a45628067bbc3cfae0f18e3b7764e38a8f92010dd52d2726b85440ea7ba4b341a8a65e9449f37a90f37cefad5b9ee598b57632a5db6681234c7c37c290827f895a3097ad0819e3739fa65f80e83ac4bc78f68752620ec9c227bcbc7efad226492252f42ce76540792a3bdc22064eb39ebccb046f448a6ba874731e979a5b784abc3488b572e3a28ef5a193097f83d18aed933b0c81d6de3c4596eb5759440f32aaa934f6da62f0b806c45bf1d5c4781d9269fbedfb983cfff1dcb44387454ceb05d02eb62188b585169f51e42cae92bc88197637b4914f5fe1e150734c86ea42f0732ba2909cfa77628d82d6b928d189810021a95ce16c98132e4b624455cdde4209196080a8fe5dbda4a5f31d7ef58fc89fac5f95bbbccadc0e0ebfde5b0ea2dfca449b5306240ef3343718812aa31a3ec2ae37022faf263d7b26de9a8008b3bc70d47b814e821e49d6b8efd313edcc435feea9dc640fbe5beeac93a3190aaa2fa0b8ee1801eb6a8eb73d0b2fcf6c5f7be41294194c2a9131489f39b7f1089ab925ed2535b2dda584849954546fb60aebf2f0067954ab00155770c26d7abbcd3b25c8c6aa820090eed50bf7e819587c7049b7ae7a1902ac44e18cfd42af34c0ea98a6d67d34cb7bb1f0098a74a5857b3dc154c155d04e8d4cd0536b741f37fd67f6b3ae3e643117de7148d329981c14d5a6c30d7a4898aca8e0d10e5568ad096113d3ca4db898141297b6063708e840595b4e1e954b79b8b4b0f8ffaed314a0ba7c3b913700a65397a338660e4304eb7dad08f326dbbb987fc6203e0d7c7b9dad8948400c82ed707c3c360220c20d0516b367e48107cc93ec4b2e2dff9b7892226b665a66096ef0afea2a13f896938a3c30be8e1b3f3893bedfa1a8bfe31114f5ca785c9783ed2e1e10e5fac8ad89be10fbcbb89bb6e62a11bc5729d6c3f9104b83aff8205cd84ab674d8690d223724b3b2883ad57454284d9b53f6ca0e628adf64669f13ba923cd81103833bf63de19bef07eb50d8dd4feec09793d5fce031bfb686c7a5085f442edfe9d2ef459a1580a2f345751c9086aeb9c8a22728cab6ce0aad9d607af0af70f4d2b6c2cd764f636cf1cf527aa899f0d6db4d4b83e3ee1078017fa74ac080580695159e418facba20a3e02286509834f81674ecf896056ea94fa98021531e1602acd956df05fc9ddf329531792b00aec7ae7a67fea04f0e3023f88d7f557a38924e3b9ce52b3121cecafc0612d43afebe99f87ef739634480dc01843f4e3bd7f5e822d362fd4fd7a3f18a7c46aa1a9934ef807ad161463ef5675d071dc52ff0caac2a3dc4903c702a2d8c6c300b342953088b8207dc66d56c90903cb0eed2ef8a9673b6365dde6ee2088fdb4e5bc385775acb843490f71779143dae3e7e61d48824ec039bf7cc34afccf02c4008749bafa2475dab12c54b5979adc785ad7f5baab1b0cb7fa53b0075b917b673174580fdf22556079e50d81a336f042de65a99fa614fe2c0ff11bf7539674fe5d81d60bb720934d63ef7937dded244539e2f098ddb301d090d6bbc21a6c9e218de0bfef700f23203d867007199704dd99d3d9864cde573503ef99a5c4eb933f7381a3b3181fd0eb195d5ad6d5ff1fcbc767bde3714983b73e2b7105d7402ed4fc7ce01f833f59f4ef09ad91ff3fe1d69b72656df47104287d677f996d9f854c3a1b17b1ef89bc5a656b2a7d6749395cb20bb1e4376f1221582189783faabdf75549da281235656d74b5398e74e9003b87c69d47b6223791df2b1e2d321aff48468902e2ba1fbbffcbc334fd497ed4cf48abee97418af7994cfc687c3844f122bbdfb1aaf918bbf2e0a74af7714dcd248bfef5684812e2b93e92a4fd0059447e16dd52b319e373cc302bed8464cee7faf29f6fd5a8e615edefbbf9a86c353eee4ea4ce8d22ff258d7bfcb7659cf0734e2b84f00d1add3cfb8434bd40230fbfaf6738e1463a20969573e5f71abcbe603f4878010d047103e9e6e4006414d82785d44ab63f4bc0190c18951505ff70424406dedc0aed19ac4953b8e6a1ba53577386a65edf549af0ee2c6e4a81d8f5767aa4dea43c95a8570e0e7887162d28ab669f62f69a83be5f86b88d6a963858eeb2db12c939ab3f6b9541713079a61cad6bdd3059ec3fbecf4da5421a75a2038bdc9b4596aca55efc040292c6d7d51041d5828a411f20f9aa755fb95331863c0fff0fe9d75b12357b21d0c7966b7f6e01b2e0c3556a1501562bc8af093641da3d2b643425d401d91d3349f375e9536692e7813c5a69f91ec0c61a4371904349ae99c6435f79a82fb720e12b0412f71758a5c28494cf2f703eb6e092b5607714519d1411d8875b394da6878fdd4bb130bdbdc11f83b92d98bf66d95f1cdd52391e6167c5e293f34c90a9e2d35830729e91cf24c5cd4d1a1039e9d0c4bf0038bf31f046602daabeaccff4dec6130f6d6b35e9fb4791bc5b6594706b694e5d154df372cd5d4567e407769db1299342d6672dea2ebb324c7805335a9fee0330b987f249dd429d4d5d43e64db525db8815274fc6f7b3f7853561b9616f2d102d210f6cf4e1346f788f321a0296e93193dc43346449ea12ea9078a3f0d709550a5480aa88787e0a0c8d4c923557e7dd3c50402383b56716e0f4bdd8e9f77ef0bf682836da9370d39bf3f1b70cc382c0a9d5346b697977ee5992642297c72b227560c43be3a4b8eddde211297d48858fb5b9bb92687a66c4f26bcf067dd174134119982ab05d30f42c9f4385ae22d208891390276c053653059319da0ca69a39f421e6ecbd5314d3e82b570bbe1e8d05e30ff76e982928a768cb544243dad5e1cfcecdab7e62fd54962e5e4f7a556bce62353dd2c8bd849d9028fe148f33c72f93b219ab33c90bc61f65254f8713868fbd8310fa79f03b22f2e84d3213bbffa7b47d8ba2f4a49cef8c2e8d6fc8d9f7433e56cc750c56d7cff337eeb8d390007f6aeccaf50f771224bd2afb0ee4610783fb2702eb87b9d26a49cb5c993877715d803fdc1cbec94b13bb3ff0c1db7bbd55c6cfaafdb965d490f3949d6617c516b446e6e4347b5351d3018e0d85f25e5d97a2e462761ca6fa2302de383df22d52bbd190f7e7f69ca03bc6cb89034f70465e67a8ac80acecb79aee4a5f09a0f293634bd2c22aeeb064b0fb354a39bb38aaad300c798baf41e8e302586a0dec50d3ee7cf7619582aa02b003a119fac9d3b72192008b8a6d9de5b6822eeec428da7249209c539f3a8ea01791ddb8a7f7628bf7a13144e8cff89659e1309f0f1f4cd286c87db50041da1de6e09eb84210d9cf0f1420b3759e3c35594af2c2fedb312754598b4a3e436a9c3fb3faae8c6afc234256d119b524cbaa5f9dc25e55add364373c821a4cc9c9571fc88cb1aa4e5d53fa72a88fa31472c0d0491cf95231098ceb76d53b9fdeedd44c1f26afa0fc00f7aa918a1a497d65653b637a32fe5cfcb468eb228eeaef9bb106b04e716d94d90a8f0c11b37a9c5eb3812f83a472face6105c78fde86131dc0538ddf026bcc22b33d38803e2ba3659ebe425f57e201e75b1f0df4485c2f682ec5c51481e75672e089a8378aca10e96ef41bd2a797cb3ee7f83b2b2a3be89d921155c5b49ce7fbe2fb5db1c4bfa470e1a5abaab7c54d919c8019871e013be3cc7f85383c5d2dad51b695c0eb4e296a8185b234e0fc1ce6ba9edde2cc31d79c90cc88a9423d5446008a8a965e12fcc7072d67b10e3dbf322cb6f46f7c4c7e3092cac5c211e669336322bb4c7d933d05ec4a0815bec5654e040877fe4b8447b1614dffd78cc12b294cca42434b47f6b6bcc7840ddf8230fd8061965c7f2ac1908a23665e881660647faf8fee0d16a4c106fbd20de8a9e506378bcfe92e7243e7726ec0204580eeb82328d4728c97ece7adb3e82b461931e3d7f9231db4b97ef538a04c50e549415e7efd036f6b64032a8b31a3fe517314a82db1199f1e8167b02535bfa5d6f707129d4ea1e7abd727a038805cb5cf548f7a5f2ecd1fdbb2700921dddbf5e472fdf83ecc8c9213ce4324c7ced7ef4628ae019bceaef6556ef82b1a8427cb2faacfd3074d9e686c39cfbd5313fb766e6a17c6902026dd3599a19cdf980a205905a04a2acde838584c45158a73fa64d231323bb92ca0714cbe6d5aebcce8c058cf9e345d19282b74eac07d0aac40c2d667f3537ea5568c828fb292f570c232994a7f543fd633b011a4f56cf027e6c4b435eac59cbb32a93bd6e0fce1ce8884b737063900a9345f8c42e404b8bc8bba7277897aa39afe10116b85c134ccda98cc4219288e68833e67311fe4319c1a7f4e4145f5c76a1f925a5b00b5313c80156ce54d1f3739c3cd19badf1a3254b86317d0c5691865dd7f6ebef4e7f3c6605e219c7742b798ae0b4b7f180fc7633d32d7e55a58174bc1d4ef433672e98de259d69c9a8f8a440538c0655866203657ce62fbe9b6cc4278c9a354d762cea9d8b0125566fac2edd1eb9dda0a83428a43c2e6a7963cca4355063a9430119410ac36c2ffdf84", 0x1000}, {&(0x7f0000004400)="8bbf7629efe09e01c9bc74b559d33afbf45d4c865149925582ff2aaef4dd083cb49f7e57b9bf5d22d070587cab0b3446b0fc90d58736b5eaf8aa1d684de75807b03d78dccc46eae447491d850b4178a87fe3d3f34652b793177c9da0d545e5de495807214188b8dab91bfd40998a24c4036829c2a90897fba295aea5c4a457592da9fe1d06b0034969be51ec84aa0622f1405da25189f274fee40689e992edd76a175aa6a29bdcf1a678527f4f3cdbb662c5ca7451af14b3227c1944d8561bbb2c38f36dc1a38d8b34cdea45c840a4e324b928167d9cdfdb2dc66a0d87056e2a24a0cbf40fe32a5ea805a2cfc528900dad9d0de42b601d802734e4a0d50956d568d47db8f8df6cfff35f71017b52a14542fab7a24374076930748c48bca672a3672d2d09da644508c3ed17f13e1404e6724e401a3652ca7e7bd611a3a679f36d9ffc41a55ffb301abc2add59227a88c9546fabd3c8982eab3c16e752e32469a31c6aa351fdf292796a7f70f11e7b9b7260d90a9a17b14beb73039b518070bdd4d6a85d95b2d1ee5cb0d880734280d701940489faffd5388700a0c58bec2ede65af3325625a0d14ade7756d802c8490fde9d59c1cda6f7574bee9917b0db7c0191d94e4aea6edca930264c20e0b4375ae158839a89efd23a2f7191b097d020372fddf994248d25443f3fe8bd425ece2ed98e295b3e8d544897f447ce5875ed69c582db51cac8f9e4f580c8f3cb820ca2ac5ab0aefa185c2c6d68ffa008bbf58d906a8521ec279cbd6cc2b5e4314ff1e3b47bc048fd8833df51b185412b32ce6f0844d58c45cc1968c65236d4d6deacef9a8332d523a6fa867ddbc09614c71661488cb989bb46699f867b0cb40251e1f5e0068f0f49b90471579942d64cf27b85011de30c45a75d31c5bcc7f465cb4b76ed1667de5cb67bbe706ce41ce5b9a87873e0de05640e9b97fd01e26d0cc64d65e4982276bb411a12c29841643c61262da56f1f7dea492492f90bb292b06327cc48147568eb73de86c29b00a86b3639ca35714a54059a1478c6149069fb48425f4b35b81ebcb3b465327f987756613c44c05f388384fc5c7ea01d2c4cb216314396123a206ccc142c1cc71516793d3dcf763b819cc932f10f693c1d5c24130a37d932ebfd0baa255f99d3639e548822a42dc13b1147203e2e5fabde8193a433a948d923183257a41bec149c34256c4c037ce25eb2532a76f9bc64ebd84b4edb7a90e3f6edb0fbccfca6df4f888e797c1878c2d71cb59e0980b0290866b78d30d2a9e51f9661724f44ff26c6f26c05022576a120cdde0fc827811e4286348e363dcf6eae1a8ef6f94f596b87009f53608f2c8edd4255d468b0bab997534be6184c2b3608b0f8fef7ebd322b0da99cc407526cdfdc095fa59a9945a9234c078a6431a7e00ea5159228482d26d4f2ce93ead428d8bddbf8dfe92dc9957ab4f9f710ae22a20f63a7c53985d4934ed283e27c32cd45d148f4965444258e7f5dab0881f17f9d61b7cab7d20710afca52c444a3a05c2d7108a20fe648f8290b5fd22b70c32572c69ae9b68c622a77a2707811df93c143aede12c3ca9acb51857b6321ee85e2d9158daaf56d452bc04e49440dc7be51615d3155da939cf6cb380d8805d3094390bd494e96d5b42456a239985fc39278f6155628adb1597360badc108e4aca8a9865464d89645f341a4674d4aff7eba78e85950e8868fca661f996281494095ab9e26fb5a55b04a0d8fd455c77b196a1e325e585630bd989b272695e5432d9b8eac5a8711b8ca06786777fe68e12a33e88374f8ece11f86653f7a157a4f133574f4371ed9a0c12b2dfd840c75bd1669f1a80b4f6f1f4ca41febf5405ffa6fe7924d48886dd037ef5301532c047acd80525ddba5912443350164e7fd6e93956fbde183c44e9620e417f9e7172967776b6aeab165fb1a18d34ee43c4d38c4a41b6634a7b9f5584960b1cb8194f3f18523efca6a9bdbbec01060083907ae78a5bec113117d4aad430d32fb08090b248b01cf5a42f4c317cd0720c70dd7dd2f35124aa7a3ad06a6b73a236a311a0d0c518f076bd98eebc242badf29ffa628b58062b166a9452de53d6f19feae02eea9b2037cc843b4f4eb05e95264476186aa250e79a4a3c0a7ab2fe46f8d932b981ab9debc4c8a6a5f3939d858fe9c0624a6b098e203026bf23c3a7b1d75decd76096ba202b74a0b778a6b7ab281497a24c1e84c72a2a10a474ae27f54ea75b190d5113de562d332e20764828197d1cbd247853b914caf24c815a715925091d12c0be622eb8b9111c2d172ba3bc34976883449a18c61252db9937752183c0c6a64fbe178788ac3f11f49aaa248b796322d03198ef9fb370608fc2fcb0d15248d41d88e7885e19d427baf2cc59de8f83ab01756e22e38f8ee6d993a647157ce9ee5d65b495d3888dfe9a0436c428b207328f56a8f20604688d5b024fecfc3be3072a49aa678847bf6fe57a49a129ac72e601f3dfb1f416d3d12502d5b53860dcd5dcfc1a47fdf072eabbd3255f00d5eac0044c3635510b162f0a52c7b8cb04be6f5a35129813b53fae28206909aa6d85006c69f174029a2a46137d4690a86d700e53e2c068ec0fb073b04e37da67b806b04be60afea3fe9e4ebbc1decea09864abd0c8b7dbf06459c05e7ac1c550c661c104f315b2712aa7ab283fca0b7cfdfbf33aeae3b5997b0844cf637873e0d8c5c54c4e373c5b10321c147390da3abe39abcea936bdcc5a0d83a4c21313ff5029310e7384c5a19a5dbd89c8344cf850011923782c1ccdf9ee408767c8948ad94f75389f1b18df4e27e382300f537d4101474ae294a498b0c1237a5d09506f8cb5ca928536bc99a01b6943164d7852aa3d252eb110d5562463044df575e12615deafc82e20c4c138c0dd562f765c021276bb861c3cacf3eafad3533500b957c5664ac860612903a8081d02fe1108f48a9b709c8da0bef93eea79f772126dd41e86dab0a4ca5ab229192a26af9530b39972875ce9d2cfc649a491e159466eb760fbae16da56d70c16247d84062b98223d9dc158bf14402612e22051e0ebf21f139f4712beb10ddfa9207df2f758c71b3cf1d317602c8d698e3f53935c9ba957567c5e2321634b88bea68edd161b391d1c63e9f070a23d7eb2b78ad5b56cf18090b8845ebe91f32e20d48777fd884b55517f9f345026a4c41e5f1c7e2d559bcce163d278e2d13358091050e6760ddaa2b761d253790e95d56dc8d65c0f702c08d5553f9a9dd4f364e50b3c98f43d8a5e0f043c157dc3f34d4f375be18df39e0e094405112f0f43b05c7856d5a321fdf53bcd909372c8ea70c217ee984ba03ab567709d647c3e95d5cc100f7ccc284aceb1f463d7caebc74b2514d5fd4c0e96059e5a76d4921f23dadd78aee4a911117276cf55e429726a8d4bc54c9dee10d505a83d0a2b4a3317db5158f69d987094609021f32ad2279e1d8f1cd0cb2ad1c08022ec3cb9215ecf8c08ddfb14712b3b42bb014c26dafd54afe17f05099edb6cff7edd67dd05221367265de0ca9fc1e8ee7c84f26f42732a6f3764a4ebf07168231b16ccee6b8d82b5f73b748f0e088f59068c317cafa09e8d658cb512fa8a0d2bcce97046d73eaf30332b519f915b687c09ef79b7c7baa2ed532593a36088c863ae853ddb3195a3417533ba8bf1e071e010518fe7a5d4a7f3717ff2b3def3918aaa066016ca2f618a8c2c5b04e6c4f2e4f38c9f4f745195ba38d15d05f1ae4e60ad61bc92f78fad000c026a0474a810a87c59e193113f130ed5b0cbe3e46369d7d9c098b6455275129569c74aaef70460a119da97cc082c55c6ca4f68b00644b3e76bd197f42dd9c837b94380605b855af2aa48268bf025183099bf545bdfe39b3d6a18267db20a6bd35e030b3537fea99d403626f24649264946afc0b592e1fb8dc496b1205052d48fb4579c44b88da414ee2e2c1af9f611f44aced02ab760e6d564b053e0fa285d34f46acc2e33a8298eb5d9e1e815bd61e7be00070b82dcf9358e4d3749bae9c2c2db9c902cabde19e34eb98d02d55ac1a191b1e7e18cad5243e2c971b9d92cb104c26669c1b99548eb7d59d9245f04cf5362886b7412ccc83a33273c2368b57ac5756952437b0d5e64ea763a55ebdcfc3867e4f8f812b71aec816b10a254ef6f862b856e05e158a4a56a540b7e2fc3382a830113ebe8ac6bd03f7d56d661ddf965d6461c90fbd5f6a74b429c82bc78ffdc0cb78f8ef76124b4697edb45afcb0290df7e68666c52596b934e4199e73f553120fcf21c0b3ca7e7a76b805a0a229b7f40e4483049f3711d80ee68d9bf6de2f15f80826e0438af9151a74f8be986ffeeef5f40fc66fa1be7218b8b5134aa0adc3afe08a89b332707f693df04fc7ce3bfc788a0c8f470587d28fd8faa7429202767c68d0294cff930b03c251a848a52fc1520bfef70e5b3df755f78115767923dbad460f4cee461325cc33118d48f38fba40119c7dd8ed6b7b3ebeef7c65508a05819cada0bdfbbe3a1c6a8d0f82d5130f2f3d6d96af3c76b7b6911e44a4861ef78de6461966b03aa05801e3ae0ba4f648588e8d7ddcda0f76f8ef1a76b1b16cbc3db37cadf5b825764dd3a2b2fbc850147112b389f9fdd7d96fc8f1a58bda0b25cc262cccf442b19034a93940ff9d4bff96aaae9206bdec0c23f3789875fe865cfa8954f75fcdaddee17e296b287a49b34dfe0def35f2b59da773340b002b699ad44ba20df0237860048083d5b9f7d69e045343fa645d1673e1d86254e22c62a6f883e32a9effd1c13d91ee0fd48b9d57f030a458f39e3e75edd571e11957f9ccbe64b774f56bb188fc93f733ab59c87a88703949cf873170c3dac46d53677f6fcac0d9ba3d6c36505eb34a9e08817e33de84576b824c30a9c2fc87d2abacee3e2dad46386bc36aa35ed23ea11fb4d2355e2a01ef7df578004dfa57e3eb00db2edf1fcfaa4a608d31facb25a5de6e08b35151fb3d73ec06dce9e3326082cf5ece63f0a3eefba8c5fe9a69fe30ac843cc38cac8442574e9956d0fc8c257482458c10c5593958c3c1d3d1915996780c5beaceb3990246ba325208b9a9e0eb163b153c36ed73f171394d8b5bde20bb392961715aa2772386ecbd442e2ef7cb1d09d65d2623f7c4e2d79d3be789048e9a2060fb7b887251b0c7e0a7cb3dd4ddf39195502710f7d3fcb358fae6840a411846efc0be2b2716b3edca31b1a731d4492d686b12f9ec8c3006aabeb41a8c2b9f53f7f9c75010cfc53eb639596a8d301b07eaacaebda45b2547447c14a4d087cb7e404a2c94549399cac09aeff7b52a43908f8b6d66e035911b34435f8790e77abb2b407b8da897a5dbbbf0da378af668639a4fff02a9393226233766e2c535a5de7e0bc187431b72608c9fa8aea2c6bb7bcaa9d2c5eaf281fef168c49a97ca7ada9e0a9e433603684acac2fc181b57d4c899ae4b619eca7c308035bb7ca5b94b3aac37efc0d771c95aa9167d0169fac65ef5879e27a6b9d51721567d6f2e95170c9080fbb83ab5b0eefb80a48fff0304f71afa4062e84fff8fb96ed126564f27b6ef936e0656ea998a5adc2ee19435baa31b2b62bc1b4a9f13f25ab5875139bad731cbaf1d6e4992a3e8831118fcb89e465c7bdce1ff10a1867c5817593f18cd62f82b4107a14299ab2519ee05ce3871982ca4f19199b04512c7548223930ee2be18ffc4b726777711df1f07bceaf856da75ec1a35bb04ddf63a7f3b68401c74ac322e975a94e569ebe91fcf499e6", 0x1000}, {&(0x7f0000005400)="c1b41f9f3168625bf5ece59ce57a8f0d9f4bbb243d07a923cfbb0a8859895a9a9bdca488d9680aa3585f891e689608819fb9978d57ad1492a63681bff65080", 0x3f}], 0x5}}, {{&(0x7f00000054c0)={0xa, 0x4e24, 0x617, @private2, 0x464}, 0x1c, &(0x7f0000005580)=[{&(0x7f0000005500)="4b08cea4af7fb5effb5ce8352b9a28251987308b2117c8a4ce32670e1f7d823d705b6aa4faccd6f1c1475ddd9df92275c681089bd427e43913457ef96e241112fefd00c159796d8ff41d8044f33bfe8823e3795ec3732aa52b32f3c38eb9df5e2960badf856d67f6158dad018b85dae0f5478fbc407ec2c4b1", 0x79}], 0x1, &(0x7f00000055c0)=[@hoplimit={{0x14, 0x29, 0x34, 0x1}}], 0x18}}, {{&(0x7f0000005600)={0xa, 0x4e23, 0x2e1, @ipv4={'\x00', '\xff\xff', @broadcast}}, 0x1c, &(0x7f0000005700)=[{&(0x7f0000005640)="5ba2cae580dc49fc169d0641e18bf968d9cd146b0af56837401f8288c475b9bf7d4fbbee58c73519ff043c0ea253b72d023724c9b1fcd73985eecfad808fe32b21790f47081462d519533760560342eea3ed0b4579db1b774695a0233be67f6e300d45494bca87a6d6e7b789daba9203889493257bc950592ccb3f8adf120221d25e3029743efbe6b5cb72674d2e75853c122ccfd82b8144a732c54a3b29d7493b4b948b2d57fa852597ae42220101a4c93201cb0845ffb189c8", 0xba}], 0x1, &(0x7f0000005740)=[@flowinfo={{0x14, 0x29, 0xb, 0x1}}, @hoplimit={{0x14, 0x29, 0x34, 0x20}}, @rthdr_2292={{0x38, 0x29, 0x39, {0x2c, 0x4, 0x1, 0x0, 0x0, [@dev={0xfe, 0x80, '\x00', 0x1e}, @mcast1]}}}, @hopopts_2292={{0x20, 0x29, 0x36, {0x1d, 0x0, '\x00', [@jumbo={0xc2, 0x4, 0x8}]}}}, @hopopts_2292={{0x78, 0x29, 0x36, {0x32, 0xb, '\x00', [@enc_lim, @padn={0x1, 0x8, [0x0, 0x0, 0x0, 0x0, 0x0, 0x0, 0x0, 0x0]}, @jumbo={0xc2, 0x4, 0xffffffc1}, @calipso={0x7, 0x28, {0x0, 0x8, 0x0, 0x23, [0x100000000, 0x1, 0x100000000, 0xfff]}}, @enc_lim={0x4, 0x1, 0x40}, @enc_lim={0x4, 0x1, 0x2}, @jumbo={0xc2, 0x4, 0x7}, @enc_lim={0x4, 0x1, 0x3}, @jumbo={0xc2, 0x4, 0x80}, @padn={0x1, 0x6, [0x0, 0x0, 0x0, 0x0, 0x0, 0x0]}]}}}, @hoplimit={{0x14, 0x29, 0x34, 0x225}}, @hopopts={{0x170, 0x29, 0x36, {0x2c, 0x2a, '\x00', [@pad1, @calipso={0x7, 0x38, {0x3, 0xc, 0x7f, 0xbb5, [0x40, 0x1000, 0x5, 0x3, 0x7, 0x3]}}, @hao={0xc9, 0x10, @ipv4={'\x00', '\xff\xff', @multicast1}}, @hao={0xc9, 0x10, @mcast2}, @hao={0xc9, 0x10, @empty}, @calipso={0x7, 0x10, {0x1, 0x2, 0x77, 0x101, [0xfffffffffffffffe]}}, @padn={0x1, 0x3, [0x0, 0x0, 0x0]}, @pad1, @generic={0x20, 0xc5, "53108c437e0999adf47f6778c43b28cd4f49f1194d80509b18f68b59bccb057032d5b475c3e0769149a1f8d3d8df06aa5668d8ab3d9a0e3c742cb3153f7b5e65f2ba90bf5fa46a10733cd35760280cee062eb242944f78223145b8bb8713bc6c2c42b7421fb5d762e7a5937a45ad7e93862a704d20401a706cad5be91e94d120231e89170c76e078121e8b6459f06f93f5076cb424940a59f2058a47c441fe5ab195d44ca4b97c8df9c4e7ff9973840fde7b426c7153f69995b820745f67b528b6880d30c8"}, @pad1]}}}], 0x288}}, {{0x0, 0x0, &(0x7f0000006e40)=[{&(0x7f0000005a00)="3f24a61dac45a3915be9e85c3d6da7bee1ce6e27a2c1d3dd2ff8ebc2cb17456e97bb6078f9bd2a94fd0b581b9b4cf5c9430c66eb6761d3e4b856e0f1ff5b1c4a58a33e33c9dca51f68f9d3ea06214ba703a44e2aef21a72142c1bf4b82cdc1fec109f97af66ee8e0449c0cf7", 0x6c}, {&(0x7f0000005a80)="be1d1409ddf392b46d366bf209c29ac46157627d5f0abbed918a67e969c2fc05b296f94cbbb7fa62ac61a54657e6e7335159e925159e63943c4e0c04ac85c8bd2c231266ed49bca18a54c1e01f4048d991c98a18679fd07f30815a91a3ddd98a96f003036347af78d8b3e802dce608606c106ddf8632958b2a69281bb5b97df407d02a6e29a44aa63cf626802532ba242286e4769d4aed787adda75d72200ebe92070f6a558527273e5e185168353e3d5ab078c4a75a24810c7e73f5990bc36ed867ac3003ec50c374962f69cf1c8d239809b2", 0xd3}, {&(0x7f0000005b80)="dd3b04f9beac32d22f8c9f9dbe23819f366edd0bedfebca2f6569885471bee84bd0997f8a2b7bc3db803265bf61414b77f267a5d45bf1257fe10a43d890e3aa39201be83a0c3783e55698de6f828d17096fe6c9b2415dee92975c9eebec69b05a913af2db2415e4d7141f01c5f5958eb53d8581cc49d4e4d7d9b960c29bcbd552268de2463fd544cf45d190d28ef0a22308fded4eb6a04e0b131ebbefdf10992faf20d282b0b3421", 0xa8}, {&(0x7f0000005c40)="7fccda0cc8ead8aae0408a025b4df83ab54fbef5531598c67d1097ecd28ea443b6580499bbd30484f1be206f90ac72178377a35f5df1db4bd26972446dc085019baed3763c2f7125c6fb00ef5ef3c3074cecbb0e7e2ebc35c88cc8fe883d0282ffbedabbfab50e048ed6f91b84b89ccbc344dbcfc642c8cac6f34a3b294ab53866f46ebeb0579b388fe36483d482279ae1bc6e0dd1e2142485b5605b5f4cc853b0a374769c7677ba156392c2d35ba0c131242bf2ad15b567666f0e1b0034d0786d3ce9b5bdacca5671ecfdc32b0f7002fa15da4fc6d4331df428bc733e203bb9aa1507af4f09909e", 0xe8}, {&(0x7f0000005d40)="bfee1056b57b8a87fac60fc21a46be552c01e563c8e40bb9a903f5a654b36e82f21c55ef1ed1a5e179705d2626b7e5ec0fca187a1b4c1b3c08b2f49fec875d30d734ac1f6ecdcfc0a4693a0386de8362fe491b42631a86daea0fd6ca8f34d545584cce2c966e7a35060bc759da4de1b0ba6c52fa9332bb156059256bb35b48b1fbb9ad5f8967765b24a87230ecaceb481a2a434aa4db2e9961f7ed8126e3175b807a67a465b02b9bc8e332641aab857d7047425b8e4712bf5b5e9557aa53ef857de882e5d106b5141d53d87b7078c19664f260fa3e7f54a6cb281cf2c155642bb11156cd56bf882b6d73c643c2f828a93a7c2f36cbf90e2991ceed4e9062b43b13724fc47942b850aa4250c8133f0c8fab15589d3caea45cf891c47b5e3e5b9c3e5677ae5186aba6be5217a4d306f5ccd0645d460b7d2bfdcdcc48b643ea54333a8d66ae864508d8922b1ad3c44f85b27ccc92c52c96bdd9e2d4f918d771498742af7a0701610c906ce401e3f2bb400a67f711b6ac624986fdf0e372391bd8e30b07648dafe3f0b06976f1ece8c4f33483895ef016e26b0f16f8acf622aacc5bd9c69d5ace1b7e8ef753d339548f555bc94000683f1384a1d8f854572a15ab74db640a4aa77a03213e8a1bcb1f361cee0dbbeb05e48fe1700fd262dc094672dd1a1c8f2cbe825b246c0df1e7b9b49fa055fbc68e334a2e47dcc50ffed3a65dab7dbda3ca1788681cfb74e212d445c30f9b603aa3751db72eed21b90add76d1cf6306b519cf61da960a0e472ce1cd4c779d7227f01b8d6f61636b1d6d16bf03b8f8dfd45f4021587ac4a1ff66aa6ba7c8725ac91f599c4db8279cbbb91cdbbb45a18d224741b71eb79ff296135f96c67689281242ec0c1bb694618b5597603e3226bde62f885144c2f65b160a291e08e8ae248180d9cb20f9635e15dd7446f4ca692fa8ccbde5a20f7951fb33d5a50ccba4f509017555f6699dcd1d626e29d5aea8986d96d7ba488aad9f512ea50bc665e12e5e9d7c6390637a64f5cf2767e717a8ba1a746c5aea061881034e89f58f8d075d673307ec1cbc1b598f18cad0101f1abda02c3f5fe6f72a0e2929a95923cef2c01832112a47c37b91cb8d65c4ca41f81d450ba92a32d482c59d2a839f0c36c1c301f3992307a5be2601088146f9fe89f33f7eb0a4b9ddd52aaee0663f683e352715a0594f4bea49b2f090c85fa7bb3ac11f436212edf7fa7864cef8296a675af92b5ef5e5131749646ef451ba3305cd00d764c997b2da08ab065e5c8c2d0e968303f2adb34f118e3f1159f47e51427614d208c5e2a4b2247d31d92df9f87197fbf5d5637bbe85e47269d3a41e2d4f7a1ffc0a6281e6624e9fb10f1eae343ca5b94f19f92041467b91128a4a7eefd26d46bb0e816d14fada6c50de931a9c6ba006aa3156d86a01fec57725f955506953a5b8364f857b130fee93e3a8202cd9601a25e03fc301beef7f66533f55cdff32fc0ac0ad154b864e975ceac381b2d8fce3232a9b8a9b41a7e4a2a671f57c275250efbf75f54773b3b00dbabc21e64ffa508173c9deec95a2e4298286c3eae6fbd9f22d28c16c475e2cc4c5d68b6e421b746595601941da19b4824244caf8cd6d1013fac0cf7b1e25909150c29e31d077f5f0f2f239f9164526191c1e604525ee6ae34314052e13b38ea6bcc59c46c49c3f332ba4129acc4edf364eb5e83f9affb943368c81325e787a85c96a6b95b6788d62e8167e377ac4d9f548705ce15afa8ce9d76b3ee238a1919b3ca852669df6efc695daa1677758ef49c42487e5059b48d26816f9e7e352faf84ccb0520f45f5ca62d8b45ef19e02d51010d7dcfd572cdb1fc9a884eae5e41ad0c5f06bbaf2e909a87233d79c6ea44668452ab985a92175890848590b4c2aa322b9b01828484ba84f6ac91b262cd42c8d88bcc41010c9a7cc1d3afd617a3cb0ed2dbaa91dda6140b04586f85e7d92ab92989a1cd676060800ebf92eec3a0b7de05e15100ba916f03ad1dc7753a4426981296bc9cbac3b773a097ab16ce41f4067f5b879e7b4d040a3c9360ef77c6a31e6e1b6d578086bc82963e910a9544b3ae2f771f5099932a18391ba0987cc3e66fd7849d7062b64ff12b80c9584026ca3751d49e6b78ddfd7f968bba507078b2eb90bafa94dcb1c9f88c777e538bf48ac976bd1df87c01911e35e3d47ecc56bcce9d6bf8559b135b6609af71a73db8923d70b041ba758fc1e7ab1ddccfc94071cbf78a4ed929f5ea9f1cf5373753323a22f99eee1639603d48ae68ef868f793ad02b81ec6ad6293df0d37c28a5c251602670f9b7a6a21b327abb84d46dc6960a1f4d28a9aac3dab1d4b919a3f6d37f36147dab3c79c94b848a1b995bd396a8435d4ffa5e78530a802f79a3653b28ea52e154347d0c960b452b3c7f8341b8b26a778d43466d9a54403233feffa74ea48a68f5fa8aaf72f5eb51e3b07f6c2449c6f5f2f970edbda9a725d80997bebe0c54373275b56de857a742e46479e09e663fe756b11ac45ae3c141939cb9ce03ff1e6fb12d01be311ce5f59c9f124aa26b2bd87e5ffdb7f73545c7aab71633f646dc85380800aecd2975447d87e7eb6c16addaaecc919bb401789952e9ff14705bdcad744af1ecd018bb7f79bd5cae90427cff478183f3b28233a55ea750d3d88972869c31a2e0c932d06a7c186c4185efa36233edc88c186fcd2f1544d382cc0744167f54e359d8547538f7f6926ae017e5ba662843960ce1a3140fa1290f9541db6ff9507de2fd0e8e77dc0a55c8096ab22a6c21767e18318706feec569be584c13d5337d52414ae60448aa383e780785eb3f196c7f10fd376f911759e420ddb27cbd744fa2a7e2dad5543a07ed103e5a467ceee127030a0cb48d98f7b4821926dbf43dd4f95125feefed4f48c3195a61d8d8a1e848244ed9f7b332eaaa88e55195a509daf964b224e7fe34b5092f6ddb6b309d57963651d1ba1c447e02711374efeaf4c380979b1810f240d39708d23476c686edece9d991048f7f710e8a31a0bf3bd6479fff543a0698933261c29955710f294873df42599d91aedc0a9f9172d131275a83d1bda56784cf1139dc33dd33ca0cd2d0c4ec2f4912e2f2ee59c1b3ade58553f7df74929517dfe28b989cf1207f1e5bb530c3afe62f404af18196a90222323c8189a4bac438ad35a1d31efa40e65fde299ffc582195d2e995666b78cdb5e2d10913c935e92936c4935eb503977c635bc4e34ff5c26f8a8dca7bb55c6f99b4bfaa97b75cc08c75dd133fb8b09fa028a4ae9e9529121fcd49bdad51f0e128814cef7772f3e822cb29bcef1544b0c54c08b6d40f24727897a0a0e9e4a15e829bf4262bb0c5fff063e277e8e511ea4b0c91093ec3bf215acdddc3c6fac93f8996a2d10ea838fade70a41b354cef38a53ba1c71343b0b4b209f671c63051a9019f34355676a69176101d92cc0a7f2dbe0a4115c3f9473a9614b0684f99feb478691b9fd250a426d4e8bdebe27191a815fdb49f1ff2bef88f8cdd20142fc56ae3893c68486dca47b7a60453508bdccc410ebb0e088677ac1875e562624750523634956baed75fe7a10ef987a6e4b71f303a2c41c72dfa86e33bd7cde76c6c033c74418b95cc37638314e70e6116d7f321a11ca13db037bcbf41ef8b6b8bcc95c32d8af5f984a961bc47fbb1c5233d200a95006ddcaee294ff7caf5a8aa192023504eea099bed488ac8074c0df343b1e3d73c3073028654398f81864911fc106282ff5e22f2da39c709380f71e28d391d171eec72fd8affc3dc91b2d0096e2943270351679483baef9671325dfa1ab923659257a973820c2e5e4229eb630fa6a12719ae8b0ec0a498b69bbc94a8400c54b66560579736523a655f7481d9330cd7e6d9db3e359f9478bc1eeafb969a0c4942c0d1bbaa20b5be683b0a3f4d35d0df39b829ccbbdee0c6b39bfb1964f7b2a6f19055a57de2062caa3ed43595a92ac949421fbc659c31256d5c7605ae88cdb119829824f35b256b3de1668a50a290a356e32a645e8746069e4e3b2aa65ccb2e591471ae241b63f4495dd622d3df0f312eb9e243712880bdbc4000e0f9e490e7580d9c1f08f7144033a4af0763b93c4d2fd1e6ace9ce8f1c9760b581d51af613eccde9487583fc359ba412e590f19fc132e51564a23c0c2b767760a324547a7d04b271915adcaa436aa17fd1a81002e3e3a8f412b5d281126c86c0b83d82b4566b38ac9ac601d5507593d72b403b39a6bee4754adb27e33a44a713c84fccbd6091024b107940eab518243b4d01b02a253c86163826133717c04268cbe64011b1834b1568612e884211ba20f4b8503f3837ab8e7bafbf4f55b888fdfc4eb02a8aa592341ee755252af23f5db80105360645e0071e0098d6a0cc82176f471d67a18f06c3b2ccd849ce65f4755fa0ab802e77b0cf5f24229574e4e2d8446e45d212d4f2d554221a97b37a381f76a7b918572b70c7f9e37f30458fa38bc075e3ec575ff9e5c1af6499833a90176d3e9e7729ec9c2c81920c9ec0b457c1641963f1491238c3df52cc8c04f66b1f0bbd4577cbcbefbc0c7fa266724bb6af72145af3da1f131bf7990a0b1094f3cfd1b44ee43656c95b9e75ffc8de888504bab8795951c4ab5672d5daeabfccf6771c367910e563e269e093343081264b813667bbd48fb8932f89ba32429659e92dcbda8ce138ca31d6232905c63eddcd4ec935c4a4f59d970325b102f9b4b52c09d7305b88116480d069be7ccee249dd3af3530e4f79c7169bb13fb5573fdcd7c164a418fda059b8d37f91ac9efd69600dda57525171bfcd554c412b4f40b2b8176eed980f1c55f4ae3bf1543f040ff13eceff746b31dac494bb5409ea6d26f62edc630686423a8a5d4348f56b38d276babdca09cdd13d5d4b509cf600e81e8b79954328952002f47b3508d2894845f64c19f04d2dd0d9680fc40494e39a17fabe122c50a823ac06bf04c52827a3714974a4ecab62d1034011d021713464839a6a95ac583b8b36d90ff9afe10660372d9a9c71b1f61598d574c3bc3d779a22a61c2a466a9eb738aca6e55b3c45ae1c1aa26a4e8fe34d7176cc9c25fe553cd1b5f98f046f6a63da13e2de4d953d3e5bc8f559a4edbb1a13af904aaaa832e606852b3ca47d338bffdd4fc2c576cd188144a2a8a4b6ca81ec6e2134ad19eeb0b78a55c62a978a891d4d49050c918008c8b0b5651e1823b4ae89139691177e32436da940d4357ab012e1e394532aaab6f03d6576facaa71649fe3c2a1c92c4d02c48d65a684b3744810dbd6f73326ae4caf6d0e7da163a6a12d8499e4f9f860481656df3d308a55941a061f24335d9723ea2bb34f518a7081d91c76f695fac623388b7b24264d44e6e2fc2a3dcc032e9bb45fe83f89ea14ec3f9b962c53be6fccc563857e1e87dc20454e2f46623b0c350d11930853a44becb53a340f0674dd9a1ebec56b933c53916cc15429bedddddc2d670b1085dc02f0e3086b7b937550934d3e52bd2b4d31ffd7d90237fda2a593188bc92f7790deb1f2f660af4070d2601af107d3586ee94b3c2924a397458cc7e06538b8bce8d28212073fb0706453634d287e5c1a62987357ff3d1125ff9c70f7e60fd4e23af41c7ae93224aab69868db2a20a8ef76b00cf1c7dced6ba8f2e313a222dacbabdf612d659fe9ac3721d9de164be80e9328f5ec1133af28fa496f5bf336c683e3a7f1dd65bb6fb2ea300686c7f1e3a7b5be1927ad51065d4977fd8ed4582bc0f3512acb9805e9a", 0x1000}, {&(0x7f0000006d40)="7534da42ab8ebea4423f2fc1354f832bf3197028a50eb9562a1319e68849bdc7bd0fe345802d5ef1c48b586968170bbe22168fd8011b05c979289d36b4d05161782c6b1c0a44e9f6081d2a3599c0978c", 0x50}, {&(0x7f0000006dc0)="56ddc9c96cf571414c61f6754cff4eb202fd95314598c614c24739b62940a41a5e0c18d4600b7f5eb7498fb12a81779558faf631673c04ffc0914544a1ff292c2bd37dc13d37c6c99e7a4446ff4d3aba057a81afaf1f1770599918c3ebca305a47c05978576ff0512be7c56b47b887", 0x6f}], 0x7}}, {{&(0x7f0000006ec0)={0xa, 0x4e21, 0x2, @rand_addr=' \x01\x00\x00\x00\x00\x00\x00\x00\x00\x00\x00\x00\x00\x00\x02', 0xa00}, 0x1c, &(0x7f0000006f40)=[{&(0x7f0000006f00)="9fff08a2", 0x4}], 0x1, &(0x7f0000006f80)=[@rthdr_2292={{0x78, 0x29, 0x39, {0x3772466af79949df, 0xc, 0x2, 0x6, 0x0, [@loopback, @mcast1, @initdev={0xfe, 0x88, '\x00', 0x1, 0x0}, @private2={0xfc, 0x2, '\x00', 0x1}, @private0={0xfc, 0x0, '\x00', 0x1}, @dev={0xfe, 0x80, '\x00', 0x12}]}}}], 0x78}}, {{&(0x7f0000007000)={0xa, 0x4e24, 0x2, @mcast1, 0x2}, 0x1c, &(0x7f0000007200)=[{&(0x7f0000007040)="28357b828bfac2c815793e46d9c9d23dd66973f83bcb7af7e6ba0da677af64bbab92adf0c17fef707fbc64e36f26513f5a8bcec1d766a76f11bac51e187e383d4253fc993b06042a941258142cfad391d252ea2c21c2faa5266c7301f229", 0x5e}, {&(0x7f00000070c0)="253f98c2bbb8f7e64e7974e455fee0a69cbebac7e94979a49d788709cdb17e886dcfde86226461a63ec59adaad90ceafee2b19023b8aa04091ea4e0b128b1b526065fbb757e36cd551000d5538082a002ad35f0c436a044b39d166659c4d2b5125b7faf422b55b4a2c7245e0d8717054e261c05c9c7bac3257d6fdeadcf5688aeb678e7967c3b704216055ab039b9e8b75569390c6d00d055249c4c4cd75826bca85d05342028bbf622467e9977d54b909c7f6579c5f97609552", 0xba}, {&(0x7f0000007180)="85e273e9152c799a6557a0362f4181b95bf2794fd5ec1f83b53ef4aab07d5498cb5ff8b1b6e2dabdeef02cd977d5025efe659c331887f2a56f1c3552356fa405fcee979bcc12c95776cbf788e5a06c0f33e39fb12d5dc2f755798f73", 0x5c}], 0x3, &(0x7f0000007240)=[@rthdrdstopts={{0x20, 0x29, 0x37, {0x32, 0x0, '\x00', [@ra={0x5, 0x2, 0x5b4d}]}}}, @rthdr_2292={{0x48, 0x29, 0x39, {0x2e, 0x6, 0x2, 0x8, 0x0, [@initdev={0xfe, 0x88, '\x00', 0x0, 0x0}, @remote, @remote]}}}, @rthdrdstopts={{0x20, 0x29, 0x37, {0x29, 0x0, '\x00', [@jumbo={0xc2, 0x4, 0x5}]}}}, @hopopts={{0x20, 0x29, 0x36, {0x3b, 0x0, '\x00', [@ra={0x5, 0x2, 0xcb49}, @enc_lim]}}}, @dstopts_2292={{0x30, 0x29, 0x4, {0x33, 0x2, '\x00', [@hao={0xc9, 0x10, @initdev={0xfe, 0x88, '\x00', 0x1, 0x0}}]}}}, @rthdr={{0x28, 0x29, 0x39, {0x0, 0x2, 0x0, 0x3, 0x0, [@loopback]}}}, @tclass={{0x14, 0x29, 0x43, 0xffff8000}}, @dontfrag={{0x14}}], 0x130}}, {{&(0x7f0000007380)={0xa, 0x4e20, 0x6, @private2, 0xb99e}, 0x1c, &(0x7f0000007540)=[{&(0x7f00000073c0)="9b493e883c245d00743a2979372b087e556d2c6a1a147d745d361241eb4bcc61ca9ee47092f93b118a97ca16c7f3a842e838d951ac2743f018dbbc269b17b6a381b46d849674990f387743d2570a9408de434f5c86042e5c4fd1cadf764ca1efddb267d436f357f56e347a8d5378d46996a2835df9b410b54472490e9bac68409bf773fdb170e4ac6388cb8cfa8d3c1b6d735fab86c5", 0x96}, {&(0x7f0000007480)="821ac7a68d525e25b36fce772aec357f3861b222a9f729837224614dc2d28fdad2e9a03caadecaf1cfaf620de3d0d25030f648b915c6558d9c28c4d201c71b708952f38d1261e71a4e4dcd30824d45eba558da6910cc9c0b5639637ba053bb8512f8c07e29b320a8f15e6ed4e29881cbbbd4d542481e463217f570df132147dd69a3b4649eb982e6182431", 0x8b}], 0x2, &(0x7f0000007580)=[@flowinfo={{0x14, 0x29, 0xb, 0x1}}, @dontfrag={{0x14, 0x29, 0x3e, 0x9}}, @tclass={{0x14, 0x29, 0x43, 0xffffffff}}, @dontfrag={{0x14, 0x29, 0x3e, 0x7}}, @dontfrag={{0x14, 0x29, 0x3e, 0x19}}], 0x78}}], 0xa, 0x40000d4) r3 = syz_init_net_socket$ax25(0x3, 0x5, 0xcd) ioctl$BTRFS_IOC_SCRUB(r3, 0xc400941b, &(0x7f0000000640)={r0, 0x8000000000000001, 0xab05}) [ 2834.795778][ T3223] bond1293: (slave bridge1256): making interface the new active one [ 2834.807987][ T3223] bridge1256: entered promiscuous mode [ 2834.826640][ T3223] bond1293: (slave bridge1256): Enslaving as an active interface with an up link [ 2834.846297][ T3225] netlink: 'syz-executor.2': attribute type 1 has an invalid length. 04:12:42 executing program 5: r0 = socket$netlink(0x10, 0x3, 0x0) r1 = socket$netlink(0x10, 0x3, 0x0) r2 = socket(0x10, 0x803, 0x0) sendmsg$NL80211_CMD_CRIT_PROTOCOL_START(r2, &(0x7f0000000580)={0x0, 0x0, &(0x7f0000000540)={0x0, 0x1c}}, 0x0) getsockname$packet(r2, &(0x7f0000000600)={0x11, 0x0, 0x0, 0x1, 0x0, 0x6, @broadcast}, &(0x7f0000000100)=0x14) sendmsg$nl_route(r1, &(0x7f0000000040)={0x0, 0x0, &(0x7f0000000000)={&(0x7f00000000c0)=ANY=[@ANYBLOB="3c00000010008506eb9afc4cd8d06e754a0081c5", @ANYRES32=r3, @ANYBLOB="2377f292252155b21c0012000c000100626f6e64000000000c0002000800010001"], 0x3c}}, 0x0) sendmsg$nl_route(r0, &(0x7f0000000240)={0x0, 0x0, &(0x7f0000000680)={&(0x7f0000000400)=@newlink={0x3c, 0x10, 0xffffff1f, 0x0, 0xa000000, {}, [@IFLA_LINKINFO={0x14, 0x12, 0x0, 0x1, @bridge={{0xb}, {0x4}}}, @IFLA_MASTER={0x8, 0xa, r3}]}, 0x3c}}, 0x0) 04:12:42 executing program 3: r0 = socket$netlink(0x10, 0x3, 0x0) sendmsg$netlink(r0, &(0x7f00000047c0)={0x0, 0x0, &(0x7f0000000000)=[{&(0x7f0000000100)=ANY=[@ANYBLOB="18000000560001060000008200db730007"], 0x18}], 0x1}, 0x0) [ 2834.998826][ T3225] bond1285: entered promiscuous mode [ 2835.011211][ T3225] 8021q: adding VLAN 0 to HW filter on device bond1285 [ 2835.079052][ T3229] bond1285: (slave bridge1206): making interface the new active one [ 2835.117480][ T3229] bridge1206: entered promiscuous mode [ 2835.154623][ T3229] bond1285: (slave bridge1206): Enslaving as an active interface with an up link 04:12:42 executing program 2: r0 = socket$netlink(0x10, 0x3, 0x0) r1 = socket$netlink(0x10, 0x3, 0x0) r2 = socket(0x10, 0x803, 0x0) sendmsg$NL80211_CMD_CRIT_PROTOCOL_START(r2, &(0x7f0000000580)={0x0, 0x0, &(0x7f0000000540)={0x0, 0x1c}}, 0x0) getsockname$packet(r2, &(0x7f0000000600)={0x11, 0x0, 0x0, 0x1, 0x0, 0x6, @broadcast}, &(0x7f0000000100)=0x14) sendmsg$nl_route(r1, &(0x7f0000000040)={0x0, 0x0, &(0x7f0000000000)={&(0x7f00000000c0)=ANY=[@ANYBLOB="3c00000010008506eb9afc4cd8d06e754a0081c5", @ANYRES32=r3, @ANYBLOB="2377f292252155b21c0012000c000100626f6e64000000000c0002000800010001"], 0x3c}}, 0x0) sendmsg$nl_route(r0, &(0x7f0000000240)={0x0, 0x0, &(0x7f0000000680)={&(0x7f0000000400)=@newlink={0x3c, 0x10, 0xffffff1f, 0x0, 0x253c39, {}, [@IFLA_LINKINFO={0x14, 0x12, 0x0, 0x1, @bridge={{0xb}, {0x4}}}, @IFLA_MASTER={0x8, 0xa, r3}]}, 0x3c}}, 0x0) [ 2835.203749][ T3234] netlink: 'syz-executor.1': attribute type 1 has an invalid length. [ 2835.283752][ T3234] bond1302: entered promiscuous mode [ 2835.290784][ T3234] 8021q: adding VLAN 0 to HW filter on device bond1302 04:12:42 executing program 1: r0 = socket$netlink(0x10, 0x3, 0x0) r1 = socket$netlink(0x10, 0x3, 0x0) r2 = socket(0x10, 0x803, 0x0) sendmsg$NL80211_CMD_CRIT_PROTOCOL_START(r2, &(0x7f0000000580)={0x0, 0x0, &(0x7f0000000540)={0x0, 0x1c}}, 0x0) getsockname$packet(r2, &(0x7f0000000600)={0x11, 0x0, 0x0, 0x1, 0x0, 0x6, @broadcast}, &(0x7f0000000100)=0x14) sendmsg$nl_route(r1, &(0x7f0000000040)={0x0, 0x0, &(0x7f0000000000)={&(0x7f00000000c0)=ANY=[@ANYBLOB="3c00000010008506eb9afc4cd8d06e754a0081c5", @ANYRES32=r3, @ANYBLOB="2377f292252155b21c0012000c000100626f6e64000000000c0002000800010001"], 0x3c}}, 0x0) sendmsg$nl_route(r0, &(0x7f0000000240)={0x0, 0x0, &(0x7f0000000680)={&(0x7f0000000400)=@newlink={0x3c, 0x10, 0xffffff1f, 0x0, 0x1a030000, {}, [@IFLA_LINKINFO={0x14, 0x12, 0x0, 0x1, @bridge={{0xb}, {0x4}}}, @IFLA_MASTER={0x8, 0xa, r3}]}, 0x3c}}, 0x0) [ 2835.338595][ T3236] bond1302: (slave bridge1248): making interface the new active one [ 2835.346984][ T3236] bridge1248: entered promiscuous mode [ 2835.356976][ T3236] bond1302: (slave bridge1248): Enslaving as an active interface with an up link 04:12:42 executing program 0: r0 = socket$inet6(0xa, 0x6, 0x0) bind$inet6(r0, &(0x7f0000000000)={0xa, 0x4e20}, 0x1c) listen(r0, 0x6) socket$inet6(0xa, 0x6, 0x0) r1 = socket$inet6(0xa, 0x80002, 0x88) bind$inet6(r1, &(0x7f0000000580)={0xa, 0x10010000004e20}, 0x1c) setsockopt$inet6_udp_int(r1, 0x11, 0xb, &(0x7f0000000080), 0x4) accept(r0, 0x0, &(0x7f0000000100)) r2 = syz_init_net_socket$ax25(0x3, 0x2, 0x0) bind$ax25(r2, &(0x7f0000000000)={{0x3, @remote={0xcc, 0xcc, 0xcc, 0xcc, 0xcc, 0xcc, 0x1}, 0x1}, [@null, @netrom={0xbb, 0xbb, 0xbb, 0xbb, 0xbb, 0x0, 0x0}, @bcast, @rose={0xbb, 0xbb, 0xbb, 0x1, 0x0}, @bcast, @netrom={0xbb, 0xbb, 0xbb, 0xbb, 0xbb, 0x0, 0x0}, @null, @bcast]}, 0x48) [ 2835.507351][ T3244] netlink: 'syz-executor.5': attribute type 1 has an invalid length. 04:12:43 executing program 4: ioctl$BTRFS_IOC_SCRUB_PROGRESS(0xffffffffffffffff, 0xc400941d, &(0x7f0000000240)={0x0, 0x7, 0x3, 0x1}) r1 = socket$inet6(0xa, 0x2, 0x0) ioctl$sock_SIOCETHTOOL(r1, 0x89f0, &(0x7f0000000000)={'bridge0\x00', &(0x7f0000000200)=@ethtool_ringparam={0x10}}) (async) r2 = socket$inet6(0xa, 0x80002, 0x88) bind$inet6(r2, &(0x7f0000000580)={0xa, 0x10010000004e20}, 0x1c) setsockopt$inet6_udp_int(r2, 0x11, 0xb, &(0x7f0000000080), 0x4) sendmmsg$inet6(r2, &(0x7f0000007600)=[{{&(0x7f0000000040)={0xa, 0x4e22, 0x7, @dev={0xfe, 0x80, '\x00', 0x1d}, 0x3}, 0x1c, &(0x7f0000000c80)=[{&(0x7f0000000080)="7404edfae53ede7d1a09859a6b3deeac57e43d596e3c4306791801aa8f33af74978ffd7335d4b3b3322941899ef8d99ac803391fc661e08d67d096e4cdaa4a67392a3b79bf47a9417b7587a9a1d260eb17554b01d84a38548b1015cddd01b5e16ae9c0e2587d4cf47afe4ad5ab01bcffc6ec361e85f1937b18d4a97da8e562ec85858d6b16ca6d146c4132b5d6cd1cbb867e866fca85b2716e98bf62b11225c17af3dc6ffa8e646dfec27b19025c69ae34499e23b267db917dac85556d0afd1fb6f2197caaf7d6b1916a6adb88c1470eb32c278a3e2ebff4f7", 0xd9}, {&(0x7f0000000a40)="3ede83855b5c4c5cb1e3f189b0fc516401361f677c3432c5a0a53cabce5d17ef6abbf132df751a14d38d4c86fe6ba04579030e2436a835da9bdb64a18fcfbbf7183316a6aa33d215a8c18fc3a09ea7ab4576794d8f315684b0af9ff8f144b196a9b1086b3b20119349abb79e0df941576e1376e04395fa98cb668367eb0d2dc180d15091a7d91ebfee8c11922f045772799ebdea6dbe1c4f5aca9647d2a03e", 0x9f}, {&(0x7f0000000180)="f484cd0060ba8f6139a6ea0eb3f6a338a6677b7298937e8283f582027f7080e6b4a612522c23c0bf2d2fe6a6ec8a56004122ce57731cbf3fe6c946cf1ae4", 0x3e}, {&(0x7f0000000b00)="6dd61a68d3c0fc641b5eb13d48c2a6c3d097cd5a44d96549275ec0f3a236c5315a14b665a8a3e3e5c1d34ea134e7e1546cf8537945e1906259bc856d5d6466b9ac37e5b96f1722237f2b00bddb9046d96400cbe8d52d550787cbab808c5068ea832ffbe112548e4bbf253908e5a45cc4d2d8b31664f5bc9f0527bbd3b3601f328979d61e6761132c10324de1a0676c97d8f8e3964eef59c99a777116c41d90eceae5308fdfd7aae49d7170a798fe6f7da775e4328cfc31938aaa30aee7a6291254796b3252240e7ec9711429dceb10d9bd880620271c6feea17780f280a0ed9fe092", 0xe2}, {&(0x7f0000000c00)="db27c4ff6a09ab2043e1d2cbfc5e53056ef813b840fb73c98b0be1ac0c8e3acb9a0c3460c890a4c6e5a227a85a0d38eec453787cef082ee62d259248562438700f467ab5689076ca6f5381b8d89a252f2e5555535992b1", 0x57}], 0x5, &(0x7f0000000d00)=[@tclass={{0x14, 0x29, 0x43, 0x2}}, @hoplimit={{0x14, 0x29, 0x34, 0x5}}, @tclass={{0x14, 0x29, 0x43, 0x4}}, @rthdr={{0x38, 0x29, 0x39, {0x1, 0x4, 0x0, 0x6, 0x0, [@mcast2, @remote]}}}], 0x80}}, {{&(0x7f00000001c0)={0xa, 0x4e24, 0x3, @empty, 0x7}, 0x1c, &(0x7f0000000ec0)=[{&(0x7f0000000d80)="8243bf28d847c328e11f2f250fc748e7454f01", 0x13}, {&(0x7f0000000dc0)="e98086afbeab9bbf0c9ebefa6d493c4b5c589e774d12f869db3bdfc70618b78fc72cb423931a0b2bba3b4c746d2d68c0cb151628da47f9da4e8cf5342952dcf45e1f76a285d0efe301c306700d23abd38ef37b14de999a69d2d30efc9ab8d7619317cde3758c4fb8be5f8faf0c97cb2cf1937279a51a1d41917f8a42bd986eed6848f54f6b28dd9496305406010682dc04da27f4a9bac7c84ffefa628aa3fed21ed48d0ebbeeb8b7f3923bfcb7e801f7b024fa930a05bee6ee5c853bf9461c0854882e7126d4f182b4b90e4431361b84989c3aea08", 0xd5}], 0x2, &(0x7f0000000f00)=[@rthdr={{0x88, 0x29, 0x39, {0x2, 0xe, 0x2, 0x1, 0x0, [@loopback, @private1, @initdev={0xfe, 0x88, '\x00', 0x1, 0x0}, @private2, @initdev={0xfe, 0x88, '\x00', 0x0, 0x0}, @mcast2, @local]}}}], 0x88}}, {{&(0x7f0000000fc0)={0xa, 0x4e21, 0x20, @ipv4={'\x00', '\xff\xff', @initdev={0xac, 0x1e, 0x1, 0x0}}, 0xff}, 0x1c, &(0x7f00000031c0)=[{&(0x7f0000001000)="5fb906fb71ad76120c1c90e7eedbf0acef97ed911df0fd053d1391a9f079a43751f62dd438bdeec3123f26cc73f874e789c29b6d866eb07df5f18002b214056c2662a2fd0020b430ca9b6f3ae85ed7a61a73250935736c7f14167a9d6fba9d7b01a315fac51b2be36decdd98affbc9e1e3383e9f71e27d239fbfbeb41fc5a84694dfb792a606630845d883f0f6f15e0a2cde0115c6376d0fa81435353aff60a08b8ee81ad7d4d5fc0027c04594f7414d6e1de1b7fd1b8250273d63f2b42295226d6871cb94f7d0af49105038cb257a442e7f59792ee843b5a3cbcce7f5ae361d3e36550c5e3872e090524f9223949de4f2a696d14a9b5056095af7c414b98cc8b429c67cb4e1f991f29e6f505f3aa6956d733b49ba9a85bdd634feaf1613ebb8f4d7dd3d939510d2eebe9a40679b156343efb4284281f93a206bde9d1986f0dec3c5a263749f25404421a6a495cd1f302662ae176a2e8974fcdfb197153058b2f5a6695932875060dd348d1d233fa9807dc5a98507e8c6fe726167f276499633b4b7463b392abadfb7b68da12aac146004f46e679e3ed18ffad84afc95100fa09bab6944d25ea55a78f10a85a063ddf537bce2ef1bdd57b101db4eb5ee4fbf0a9d6c184342a119546efbcb71ada2db4c1500a46e837cf8c7606d0be343d0e460338a05957274054c092cfa3e0eca9fb8d2f845993d9454f7782c8c3a9b823e674d2db4a924b4946827b3b8f521c517aebfc9542dd799bbb6607290faa4743acf98934892db1bf744e7efcad821c07edc0f9f18de11036b950d92eb58d4dfb7c845730e9e236ea6c6cc56287de14f6e169da6874dd5fb442c37f6d18ac40af97fe92d5d20c4e04095918cc69f7306dc474ae33886492f8c8e746edd7d96c776fa76aa9ae243f6acd93d982ddbeaeb301ff64f25cda9eaa7afd138b840b52f183bc3a7f16f88526e9b2b6913335b6cb72913c1b59eb83eddba71f28e793f3ededf994cb6a80b47c1c59edf73feb0b831a342dc0c8a26c0c4b8132db0eb3656e9702bcd924b74e6853ea0e35d01d5a1da103f359fcbcce084ad603f61343d6f6a2b47bb131c6e1391551e38712132ddd0ac842ee4b2c4a605bba68aa56c18b6e785163194e924a84804c7b1ab5b201fb9684ff20c767a1fd19b10b3d6bf8658ddbc1364ac761784e3fce85d250fed72c4aa6fdb45593973c4b020c31b3c366d17223f64233b2ecf8c37cba64631c56aa066f2257718bb499e17ca903bddb35ebed11e2720cb1b5a027b4c0eae4519ab2bbaea2bd3dd643d30cd441b364762dbcaf3e598b2996ccf2aecd177dc4e0dd5603f9d71be5b9d11c76a7adc012963b31fadd07c1ae825099ebbeeff10b1b9e9bf5ba542d2be91d145332606d1ea4847ebe55a891746b4e09f7b6ea1cedb974bb6ed649a35318945767f17eae0f212e44176bca8797129723bfc9e6ac402981b65a1e72aab2efac6c592306dee6070f2e76d199c722753ddfb74b772fc530ff403fd85bb9f17651d70a871d395ee2b047fecf254877feb7f6bdac80af1e5892510b0b559fee90b6c7b5b287498c0922bd911d723856a0ecedceb63d8b93c014e1bd7eb9491fad5e986eda95ca1eba6ff3a39f93ab75334b38e850694685d9d323bcd70370532ce87f00f0d38678325d174712ef9c9f64b0a03fc236c991797d0f167996ce58895b1fc62e2a302a15907544a6dfa92833cdea659f6417c742f40fa699008036318b1971329cc7ca67930e33606402515b53eaa83cc54e2bcd67867e1357f56c1315b9070b327e1840bb08679955509bd83f7e1a2e92b1e3b5fc90e5537714732af072dadef2de4034077fbfeed9ed3b676bc8b6372a46c9f18a1962949262698e718950ea1685d33a9f8642eca180614feeb5b108bcfd46a8f0615be219120545f702dcb192a3f2f54f450586bc8766dfd9dc19a949a95cb9d62d85e40bcd64911138a83760845d2f3e9d33209807db5a21dee4d9e54b8c20a5f67892b06b13272054f9e39dd54efb57bee4e321cae1db7b283d86d324755f6e97be34d7352036910ba24c03fa9cf84aff230cbbca8145721e8c70929d4c70cf96d4670e5124dcddbf571e9311f30d2ee111990b1d41fbd1a2ce422d8129651a4eace61da1d8952e60cef9f029b4497a06c4c79e819c3905c4761e89076ee5a6bf023eb54b1a7ce4b615c7eea5482ee7e77ded251596d3c4ec21cacffdf9eed00d3a141bf5d6488bbefb269438eb4e60487f1b2616cdb79869304af7043f47e60e6c8c900498f59756a72fb6bc1e74fd07cd1679b0349c69b6ccc36dd2cd74bba95388216f77133c22ff20783fb06bd84cff2fbb3e32a260c8741f5725c54c6f7e659eda998e3d312c55f203343e4c68a4c5adeb2859db2dddf17a82bc2de14c52ba4a83106ba89330dac4c05f3d981d0fa5307d2de60d146fb1dfc7a7bab1d9f86dec55db77dfedd0f9224ab157ac746326781283847015c61fc06528ebd620ebc815196e4a3a6ad293a1c895c3767c61518c72bd0699355e32f5c38193215880806b85f71ae748dbe2a04ea31cff5aeb65e1e7dd8ee8575a5579dfb5f84c2e9e707e6d5b4721ed461ece928a658ced136dae166d512a3323ebc064bff652c4de6bd7108239d3709f7472c91f54ea504d97c1ed2bdcd873b17539d19a620b15d26b343afceb251f4485cf91be16d4e47b883d755cc6d2dbf25a196742214a854ad1918214bca42bc1b28a952a5054c2b26b172f4bd96ce2e22512ef59ed2eadb496debfe06e50ddfa9f941e85b17d2651d976c1e9f03574bcea247e2d0903c5a2e3b0fdd1bef735bc97126424d6f94f1d18f5e4313599bec2bade7dc6d1fda2fd05d6d325aaa4c50fa9ad95a414338515cee9e00b39c552a86966d64df5320e2a9638c1a2e3a794d79a46e05079ee108b9d4c26174c9d05d55c854d9b23dc963f5264331cba5c97be78e89bd2c3172d40277b9bbf4728c4bb884c5fbbef8fdb89f97b4ffef27f9af22ddbea2909706989960a58accb4f97f6be0d7a27f4fba9628c5d0d07a8c19e897963ef29b959f283cf8049d1c0a011310f45cd8b9cf6d22e8f94334ba54281d718fe5252808182f45607b84fdc551d6c0ca98ec47246533f4efb1bb83170b39a11dbdf0dceda366adb009ef48cd807aabdfd76ebfa7d1689ea0ad638fd2c5873dd62542e98e150c4c29eef20842fbfb9dbfe4829614667a4bbf38f727e66c614c4ce12ac034a92689ae10f3d00c872b67a3c69c39560964f50420a4ba3dcf62d6d39c603e60231b36b0356a7a0080756b761feab0253e68e67f317adf72a7dfef797469547a84fccb764e7b78cd012a77ede0177a8e402e56540e9e5e9b20b686d50d213226be12d067c28469827ddcdf7e15d8d4a6200e9e52bf17384041b8aa70ca097275fe70f44643a8beff01be4035929d765d7e863023cdb90f3bde63fff91d023187a0a3fcc96018475d8767d978dbe58a5555feecb6699e2234cf481ded5b8d3bec293faf12c16f890a52238d3e3664118d4bc919f05fd7a299c2c9d6a5fab9dc0eee8aec75f57ad0bc91eb29d34a485691e2c225654b3abd4afe162346118a87e33f8c2cbdb3f68df5148eb27b427ffdab10f29475ad17d8ddec3025a98b22b30bbc27fb6123bfa711889e051db56cf07aba5256458e5059ee5b4ae791b96c5a45170ff845c244d37a5fd28dd756e1892fa9545a38a9a925c41dfcc55147215bb2e09c1df454c23ab3c5a64cf339ed75530967765c81f0bee2d0f7804e841d76b6aed9e361b433e0089e9681fe497d473cd9f39737770f4e887ae84cc9f2becfe93180fee9dec22b9b8697e7c5f5dcfa2e3f3f3bb0c94f40400caa45971a6c7c90eeee5e924035521bb33a28f590dbddf58251047fe33bfba2f3e9b9cb31d2f9b0b688c21e8e7be2a3add2bd337eb4884931d78430aaafcff9da037b0ef972506fc12723a1ea49ff0e20bfe1bc3359afed16337bdd79370d992fb5ac6f6c1f6da011bda39d64f6e559589810fc184bc985745ac7a9c5b3926d98c6742b609efb0586077c4537969255eaf90acc134f80c1c062cf313de92a40dbac1f6e011716335632036a2c79b6adf0859f8ced80776517d6a13e990eff0a169c79e953784e986f3f18abc7fcf13f583b6e9f383543b41f955815dc93a687e88d10f5689ef7cc38a469e5b2f0dc4fb3f2d62de06c4d0f1787d50d873c5d4c510d6cf5e62702b84110a7fd3432d1bd1cec285b6fef129c9f8c411ea9797694d45dabf3ff0251db6577984728a23bfd60331f4b69a7307737e6ab90d79db6c107adccdb35af3b28ed2da1e449f10d5760b37b56f484c7a56eadde8c3f0c7dce573eac3d4a6f46d82dc6463d4c02425fbcbcf4e3d47dadb96a2bae03750690a7232f3d0bb5f45ab80f61d0bd2495702359c19abe73b730ddc8ac6cc87ff4e0916b0d401c579b33bf4eec208227cadf3fcf8e2761eefec22625eaa8b1820557a2de152f2baea18c348c1337c5fd402088a81bba2f5544f208ebc684a6fc20eb4c8a40cc270c937a5ac417db1b59b00e0659756466d7f1cda12092dd138fd8b5fa4c79431d4f3c1f2dac347c29cfc2ef23f39e01a6e35c7b0c20b76d29eef315a2b49d3aac78e5a6c07682eef5528982142a555cf571463f2acefcae1056762ab6e6adfd1109399062f836d08a5203393d58ef72c1b730dd7b6728dd6a3389321a01a54a96c27115b594cd401fc1922831097f333d182764c92a3146297948a53549758825b5cd2e63412c493c7f57aa3d79e062bcbf3d3e8b8f94917b6df48f4a27d4d09330c427f105ec4d56117f0301504cca747506894af3caca99b934640ef57bfe93805b9fd4e4c9880da8675eaf3bee1247777827018e54bdfd788faeb84e1c471ace504a7910670bbc3c8e75d546940ce0768c543bbd9348180484f0e9b279e3667ded2423d324fd098d4585a57feb05319ec7a137034e721538cecc06bf6c87541f31f4ea4b38623425df336ddd233d72e1bc9ac556cf97ca88f9116aa23b4bc493691277ce38e2600033c6dea8bcdade7e5648a8eacd4358c6b04f8b4bc5e459e9b9e91e8c0f26129c05474f062a00ba85d50db442d6ecea983726fc8fb47332406a03ca40c2796eccb545056968c1f2cde6446549ecbc21da3618b1eb46bc986bfb652bb93178db6d7b4beffda22ac417f86a959fe3788c628e4d5a3ffd8a14e3bbdeb6103907179cfb9420cbea5fd2112e1409ff92ccfd13380e82df79fac73bd23137635f12e293e8b3077ae9dabbd2fe2e13257c2f8ab11d20e1e386b0637caa4ca227dd1670e3e4c4bf78dadad16540144a0e7cd67f0ee5e864bbaf17b4d22bfb2e4a3b13705658577422799cdd991239e0b0d0de5e55112cbeadaa4a4a065c165edc31eaf17a96ff6be007a123b5600a3309ac4c7138b52d1257ecda2a1a9d84f5dd66f7121ab6d342294c6cc7b564c609d5f9381406b2a7fcec40d779a02b95843731cb3e741b4c5152eba9e4797f95d2c7eed2d0e5f66b2bb36679ca2b5e483d9c19760f24137471a05d0773c84876cd54969bab649f954d954838ed352cadfe63df02391f8bc057da723aa33c478508e85663815f4131dd965410159f3f91563549cf4b55bd7688624135b905795dc0cbf5da7fd62232dadb4b7c0b7ff9ef33ff73d3400ab41e702eadfa266905f4406cf277d0a447f3b93b766e5b2d0414f81396897ee66d942353cef68f2df92d8a8208388b000411a69169946a2ee38b4a47bc8c1b", 0x1000}, {&(0x7f0000002000)="7252ba8bc1d75a97d4f12549ef766b80b2e4c36d1f11d984de5807b5d9b7be5badfe5d1d79f1bcd48743507c2fe1e6bc93afb1a18bf8aa4f1141382e5df43a20851f2ab3b8353af0ec54cffc54378586ee8c9eaf3d4d8cf6640aa94fd203acd514e78e44f7217406320ae0723dbf749ebf3685b7f12db16702dc7aa938b1abef9c789602bec343e1f0d54fca9b7146884f0c4da57a0d4011c5cb61c52f36a5469a455746ddfbe4a5db43f26ba9826d6f61ce1ed9a0b46ba516f243d048fa2211307d06109403e98f8e16770036cf580f5d6db8c2dde148bc0579740913cc3868bea39a497f8942", 0xe7}, {&(0x7f0000002100)="e0949db7f4ef749c4d9532615e3090e995eacb74b40b8c8f6b2566685cf6b1c677473e380dabcaf9031d076b23fcbfb108952dd43a2af98cbf95d25f2524925a2f86c4be9f22175c1612f0a5ecea1fc256c0d287395bde0c4b8f3a9d328a00e12096836f20f8ce2f6974ba144c98cd1d3fc176eb37f066f52ec2659d57c00010571b878831bf2d01c2b85f", 0x8b}, {&(0x7f00000021c0)="e39901f38cf493b7c80ba34d6d55f3ba3dfc148fb0eb59a7dcdd4f0fa3ea59c81921a08e3eb6be97db3bb6321ef9a212d8b49d32aa26a2d0fa01bf22723c432241f5ba018fbcf6de10bfd692a1e92b38f9ad8def34e79070a778b6f446c1b1e8f26e19b11fc190afe9d0a75bff256751a45bddf56954fd16fe13b3afe9be1b98435e5a2ce95a2d7c52b5d1069bacbf1ffb85c414c0eda68fe7d70c77a4605890f52499b751f3567ed978f329870d606369b129d83a74c7f4656ffa1ac2d6741e3728e3ad6d3253c375e552a53fe9d2776d27fbe2c1d698dfec48a1b6bd1e297cd2bc6e3b844d56934a88ba2a4b14ac208af2672f37e26be9d429624dd33f22361807c631d346e4debd8b1e3baa1b6a2dac341531d473b126667aad2f4617a504fc752b7283ea0bdec706910745ddda451ce81df4c6c33cac2479ebf380aaf0061d8203703656fbb6b88845b743e642752c2267f7c1045bb5159676c7bf53502e157457c6a0e0d62ce83d7b74b789a2a2bd4238b4236549dfbb377522683f362c1c06200d2c0e9b985b565a91aacd64f6db36c904ae56ca8f4b718e71bb0651d5113f560c0fe3ffd70ceee7a07b8df0e7a76d24cf87a902b4eb913ad10bd0c965dddeba3ca473dc8698ffb879a405e489a917bd80656122c75fc27b85079e1a63600212c75675e2246e586d081e1a14da434903ad7392f9218c4ea6c2966f116bd292889403720c647693bf35adf963b490478e8af07268fb3ad49e0ee096058af003f5e7b80dce9032b209e7cd79a2a5e61c0caee2a4e2d90d307b4b1c9176a8625ba994dffdb4a3a1523c5091d86af016749e8bf9671629b77a40a44de950260e2e15f3ec54bb23c1b8e4e31db20e8d2b2a0307b1f988b20a374d1e5e1052bbc35dd0766be1fd241b7269931b0b510fe0ba38cb33cedf95991b78cccdedefa5eb953759c1a435db4e2104305ae73e4b839f2e8af4a053a74fc3e22d612c75e355c01e44911758ec8493ed3af672216dc1bbd42490a189da994baea138bf9d27e70444a6db92de757d31c91631c6bd15e92417d3b5fd04cafc7bcf126877aa0affdd0154a0ca10d39980782ecff7d37b1cfb6ac209ca6771d470ae820b21189bdd74f5fbbd4336dde08d7adef8b20d73b2ebe5b6be51edbbd8d368c705f6f67461555790fc38372d61c85bc65fc9b2019bca8ae1764be788cc04a0108fb30e83f8dca571d1dd45d3b4d6b5eac0b716fc5b245fe9669d7a229e6ee72bfbc2b70a02569d87f1bcd74fb99b3f93fc39566ea4e152963aa88e219e0665e93d0b116e92e32a6e222f85971edfd3bcb310631cc7528df95d1c5a4b71c5e9c384ab1554c0db65cf6bcbcd3be5ca379ab704cf33258d7542aadaf71423f9b655792d9f3e73e2c800322c39e530f4dd596f9cb6795db108f2fda6222ae60ca3851c839bce543f3127b6b5d411b8e41b4b9f585e517824ad89fda37ffc65d45284fe985cb159d7370f36e551ff718dea5387f0abc2b39af55b5ddf0b96a05b12817c12166e75c4b33c2ca1765b1346dc87e21ad13abd6402cb22f65ac47b078af23571348494235e8e5483b05e1c8c3a4e09f3679ecb866795e439d87fa852e59581c5410600500b18f7352dacd4a756b4a49d689e260f047f2bd495e58034e65b2473e2b8ca4e763294ee4b2f74db56f6c2a8fc28e3316e23d538583d125e8420a50337724bdd46794d19401c0356a906769238417ec4e217c9f37dd77232e996aa0ac09390f604e023a7ff12868d8dd161ca3d80f8acadfccf11c2d301a9c0ab465f42a10e950ac9973381714ceb3dfe6c5acf1f6d83ca25c70a5c24d49b127d78153cb31297683ea0568f038f520e49f720061d6c2dfbc8b0bfc0eb4b202f2213251792c6f741d1ee3c5489c76a7143473928ff938c97683a316c72d6e35844fe8c25d017bfb8a29532c9f444f43559e909ec507fffe66759bbf584cecb0e4570228e4513c08124d91a17b90c2230de488f38f3f2e606ca69e4f006803d6e5cab9b11c9f37e9f01b86301f7addcfa4624d6747d82552fed7d7c77183415ebe6a9d02bf40588a240001b0c0d13c74944d8bc009296abc6ad16c73bde8aeb81a5c92e9c3b971c4cd4d4a1423772b1c5e5d000796474e38c3f89b1c99e19ec76cbdd7339d064221ad02c54a9e0b3d46573bbaa7c0af49c8e68e6f17ef33735aec1a982af8ffad35235f0944541ef0c35c9824748614187d91219986bb2fd45a88877965e7aff352d2ef4eaa5f1df28799ebfca9e7544cfe93e7810663b956aef8ec6e3087d3fc4b44e1d0f422456f22ff89d5dde9ef5c448b646142f2fc8977834f245ba4528b70e1c4daabb6612b91f96361be441e0081cdf04e5b0b7151ea3acf05efeaeb8a523326be040044c5cc450777cd5cf2ef549c2f712277dff37ce30b7595b09a1e1e384702dda93a46e63f4da116ff67b6d1e8779d99ab460f95dcb0a1ac300e2a1c3bee243741086e1a0254231ccc34a9b9b77a86a86c36b590dea58b2b943b2eb15212dd4e13d796fe23973da9b1c01d346f0098591bb3f37f40b77f31cc7e73b15db81270d8249e03b2554e2c32e54ddeb1b286e6cb874d8a3e0d442dad9574668fdba19abf13f8b170952b4175626ad31fc40b7bea69a9fd4c0fe016bad0a4e988172fa81b226d3695ce89602760009e97b587a3317bf6870a70ed910f1355e03e80cd518be82d7e06f54f8632772247fbfdc0e95068741dc32bb4daeaa3821da0a1b586f78228762eae6195e913b500e5d63d75217e8a5a1c3dd9d4dec8ee798def5c0066c713591b83aa3aadb0887493c4066b2a95fb3b4d316685682b7fc92f8a19093a7d790d7a65b430eb6e7e96b2da3d349f54df43d4cc054fbc041ea9c54dbaa61349f669f32c1cd2f3c98e2d9939da4ed165264568c099fcea1a5a413470b1592a582887bcdc87785b273aee82ca3bd05986ecb860554f01e5a4aa0b1bfbc6ba5df6a3abe73e14da64fbde485f1af83b7c4ddb656f57c8c857b79417d5d51e997717a00138f6a7a0b604999b77f73f0edb9e7d59413de9825fd49533eda109ebc2324f432f2ee563b52780f701c81778816042a5f5706b273635ca05274a981ff742f01df9bfc59c3f2792260521a819404e3111529cdee969e55ef9e2c6ca6fca90a179d3f20bc3b42b1d69ddef324f0cf04ea6983fdf4e8d29217a496e6fbcdc4627e49eaeefaa9368bd3dc96de5248a8c57a3dc80459f1c65b0a1bd70de3d896521245b5c9952429b68aac94660327616b6d97568c1dd6d11e7c664ac077389185b659309e79954a0fc42b8c21a6caf29fe05486e78b786854376cca6f18cd8b790ef0c83621b469f6e6175a29432e1138782417269a90593ad205b3c691c13134c0f0cacf18a9874ebf3286e2e3a4d6a86f0c8b838642f39b5e7538a68072379a23fbe8f5a2c679da346f67487d5630fcc7d40634c8309299e00c2b352d69c8a11c34b20c1e0999108e9d922b68289a1403c1de9d6806cd4112a3d24c9c34bad74c6efebfd06497a73f6bf6a6a0bebb127abeab1931ad70a0b4bfa9b40be372a7e45abe7b611b5bf712d681a57c0254053fb104389ca5e25865ddf7ef4dca6a2f286524e263ea2c166a687faf60fbc13bf4d623521682b3104e69a1556d6f9996fc25d1fa36f8df5967b8a7b0732097453a897099f5fb4d47a1bccd1db44da1837686afc9f5415f51c0640f9eb90d1b80b75bb11d31e1f553933af3048b77ce77d7abe6e556d7293b1365a357b3a714968b4cf9699eba4f2b62073d9c7588275f243dd4020839b6a1838a19f5ac19e00220294fd914fd10198cf42cd3ee5cc16aba3bb554db17ed771ed2bbaeedf1477ff712b13f929982db8eecee7cc3d9a29b28932c59a9c94f67f89116c33a551c906c5f4d96e5ceaeeae6197b9463c3e2b0456dad025484267f11d041213c21e89d073c6170efc478fb49878003a4f03156e5daae5dfb8f1f04b8b7f8788351ae7c445133137801401cb496d426ab0a05d9d933f6df8d593389ad2d5cc6453d7384007f23c6e3098af6fefe9fa27d3783168e24776b432b872023379b20665c131cc13f8b08bbbc9d6a6d9772f37750602ec66c2447415b97800027a1f62cece33c0f2e3703bc27f5a6a25e261b90452ccaf0bceb73ffe9c91356ee28e466682708b4aa245f7b708ded0e4efa31c33793eff3e960708d7df3f731e61e6f02ff94398fd26250e43452c07dd1984c701e26086fc1e2cae819e8420d14cfd120998212a7b316b39b9c64aedc3138973571ef408b5555ef5c5ea8e35a4d62f8eb93c0aa854ca17533e71d078c3d066e8b05c205d1b9497f7988d40699dcb2e82b3f053517c4781764c9d031a88b0494ad3e29faea4310037710a98087ab3d75e4671d8098854d278dfefae74957e168a9b7e458debefb85e23e009ccc7fa31ca4d23dc2c7925d14112584edfe1a27e8ead18c3b81e621eff3e8d307d2ec47d3ee388b5404e9be3bc9f46be6e9a27c0f65c582df62e1286dfb9b805d5f490e993c278a66536155e06f4c306ad0240cdc3108ec9e8e63741aa1dbc24e521dedb43edfb84b4ba506915183d1873124eba9e7d0776ce476b8b8280e6f662520294cf484b5b155692a818437444a494ab0445a48f9a74f7bb61fec523de839f527e4b116741069da2e6aec6d7835d786205055acfe7be63ba07d8d215f3cd948ebf360d384ba1783f8db18fc6362f3875876c1ceda2dc8441d21e0b430c9c3bf4e3eeac9ce965a37d9f6d126fa8f9a0340a039c61762ed4588a2abb47e0cb88b791e8a9154eb95402eea495a3a49234c83480dd628239d76f058a3302a3ca4f693aa3a994dd18cb5f99af03404889ea713f2def3116588fac94c94fd7e19134668c24cd949835f85a70c7686917f5cfeec8bd1bf3e198bcbcc980bf66d53a26b4472394424adde196c2fd73e452fe6c51164049f0d7eb76fec48ede903d1aed21b1796ca0c869a1b7974e01db412c44c3cb991214ea56d40dc296a4e4b9b39712e17ad8250ad5274db318f7a3f4c373427ea1c440668f9c94bbce7e064aa1cf5f6224fe9e58e4d488c85b1e0e1dc0d72945b1286dbf37bfa4de1f4013e7177c259790f44e081e0def7b9bd45620dfafc5783b0da095aebaf2736547a0955dcfcfb2b34f9719daf8fbbbb89c9bb0a49bef904234d9626a5debfa8731e61d326d635af87120b40e7158d54227462d3c7a88b40e9639338149db971aa99b48c33dc1cf759fb74a9ead522574a8fe5c870f5af1179313228415b0b894a76235efe001b915e0f8206ace9dfdebd90e36b6dc9542f43c680a694ff086017aa46171c2d98ff9c90f3116876aab63d58cfec732be181dd19aa6e68c3897510617b135cd3e578e11c88a106c3204b769ab91cbf454e601f9921de11c68e7c0c52c8916924e866c9d48f510dff5b9d3fb96796cf079fddf3c755c9040601b493214706fd37bf89e0c737305da67bd219384b87bb2807b4c109eb30fc96fde40bfb2e9c0321e0089a24ea72351a1a13a4cdb0e1dbfd4c50d7327458d94f0c49dce905955235b259d7a808aeaa9eb9d8a4d781545b7f4cf77d6140d48404376e08533d269316e7760ab7faed70ed938a97a5294e8f737c56be715e61ee7378ce0bc99b9e53e9143aa3ce1dc768704655e485c047ab8248286886321f7c9d41830ee8cd886201748ec891970d5b8d99f9941c32e5be2f1f77f3aef57c0c32c33fb3bc71e333dced2e5a269c09f40fcffffba2e0", 0x1000}], 0x4}}, {{&(0x7f0000003200)={0xa, 0x4e24, 0x1, @local, 0x200}, 0x1c, &(0x7f0000005440)=[{&(0x7f0000003240)="7df665bb5181ae567a88ed05d75347c1fa484886428a1668ea1f35a7ff4df0e57e7ae2d148340c5239f050aa2e6da57039d0e62164d484f8ea3fb53712b63da4d8c88d5e4a4865d02b0cfd4da18de33c80d601f30877ccd6d1209c3abdc1c2a065d4bd29bdc51208e791c4dfb39c84a8f87fadd9e248f917b5f85c3b5ee321b27cbdf5b46b603822ea9997501311e59d9724e805a1fa064a8b9fb0502f0f0df0aa2e0197c7d3e45c4cee38883af5c2a52818b9744559e71ee031d03a73c71a350af164f871bb9531ee415df25ccaacab14fe439349a865c1db2679736757e0522e890ece9884e2d9e2ee675f8d4fdd65ff991a761f00dfd7", 0xf8}, {&(0x7f0000003340)="c7ba26e2fd4cb119bc254f5c25dc33fc4d25eadbff01de0e0f02699edbef7a257df9e9bbbe6cb085a0562ffe414a6bfc4e516eb5b46ef48a434d5c90df38b1d26d32cf746891854680da8194ca464b5f4e58f1c8043254d4e026dcb18ab991d6d1ddf39ae3971d16bb5a9f47d376ae13639087c6c8a21968df71c4c24ac771718c17e3f92a23ca98e6256904211f69ba66278ee0b7b060c8f0e207645903f334bb3a8647ffb7abc8b6b582d50497dc76e3c587e85dfcf30beb9c5c261a", 0xbd}, {&(0x7f0000003400)="e190e89f605f00bd91f69ab4666bd7ae317b474c1e6c78fab735984eab18644c15ec4a3a993e9f96e9bd9837d93d32e5ead1c84653cafd94d76a36d7ccee1d91367c55b74dab557da03766ab6120564244fadaa27f3fcfc00189aeb15d9a6afc8de532466999fb732a42e728cdfb676ed47375f5ee81ac6342123963db2f79c622ea5b63c5836ef02c24246fd93ef0eaf2a3463c48c4bdc7296f36099882cef0a39558e946fa98102661d5332c13a696ad2524b4942bdfa45946710831f3ab5613b1806c454384ebb0137d17fc3b181b9ff9bd5433fa09ebfdbba962ef61d4a1cd1c6432b8803ebf133847d88f660041d5491ec2c7d6b99fd12e8259a724331633a587284d31ba1cc9b0af650eabd386797d149097e00dda9fd9efb2695fc7120c159cfac8dc612c66186dbb3944afb55f27d1e8d2be27a538e3c3e2c278ee8215fadfb9f082cc945cab4f9324bb27a5cc2b3ffae06564c9c2f91e848c93f190b60b140fad71a3c732bd1edfa09ff85c585791dd5a68dd8fbcd3a8a804f474dad52c53c724778f7bfcc6f41345c997d95372c7c0635e9ee40f2d03fc9cb3fbb7a418854488f9de79ed78cb5ce6cc7e1c158e7157eb2409a8129000b022a738248f96e1e22145b180fa8844b2539deff9a7826735db39cec8380da35d50ae0d854e6f6e26e5b8418ac67e4f74a93ac122fabee469650d541ee6efead191203cd21ec9e30159e5ec75f9707555cf563897a7ba6b50cb679a998e8e73f92c7f1c609d5d763fb46279f4ff09dd11d1d3192e399bf554ab29a5520f6e7dc7be2d796bf0e66ed1a07d86de5bbbb9c77a9eefd2479ce23cb5e38eb731b8d169a3cb810835b4cba05ae023f6bf5dae84ef6342220cfe57337c266b5afdca0bc9f705eb0d48a2922b1cfafec64c540b81e636d95fab9e72a791210c67e5339ec101635bcd56a929146bb8db3e1b76d9651df2f63fb27fbb048358badf039b2a28f650c90c0147d4daf0e79734eafab8cdab89150c97058ccf64cda9f39234eb20e2cc6312e303e52cb8492afaec5d7c7c9c4098ff2895abb6eaa3806af9fc88e8f6772a64b852a82a64fd1d7b80bb2217e438b7eaf8437e782d398946bb6a78a7487bac4b482727b19450524bad0ca57034449abc5a5bb2aa39be4e76afa96ccc73a574e64a9ecdadd4eae8b88b56460c2ebf6e4fb40ea835a8591d580cdd1560ebaeb52378cfc52f6e70005767e11de0147751fb854a45628067bbc3cfae0f18e3b7764e38a8f92010dd52d2726b85440ea7ba4b341a8a65e9449f37a90f37cefad5b9ee598b57632a5db6681234c7c37c290827f895a3097ad0819e3739fa65f80e83ac4bc78f68752620ec9c227bcbc7efad226492252f42ce76540792a3bdc22064eb39ebccb046f448a6ba874731e979a5b784abc3488b572e3a28ef5a193097f83d18aed933b0c81d6de3c4596eb5759440f32aaa934f6da62f0b806c45bf1d5c4781d9269fbedfb983cfff1dcb44387454ceb05d02eb62188b585169f51e42cae92bc88197637b4914f5fe1e150734c86ea42f0732ba2909cfa77628d82d6b928d189810021a95ce16c98132e4b624455cdde4209196080a8fe5dbda4a5f31d7ef58fc89fac5f95bbbccadc0e0ebfde5b0ea2dfca449b5306240ef3343718812aa31a3ec2ae37022faf263d7b26de9a8008b3bc70d47b814e821e49d6b8efd313edcc435feea9dc640fbe5beeac93a3190aaa2fa0b8ee1801eb6a8eb73d0b2fcf6c5f7be41294194c2a9131489f39b7f1089ab925ed2535b2dda584849954546fb60aebf2f0067954ab00155770c26d7abbcd3b25c8c6aa820090eed50bf7e819587c7049b7ae7a1902ac44e18cfd42af34c0ea98a6d67d34cb7bb1f0098a74a5857b3dc154c155d04e8d4cd0536b741f37fd67f6b3ae3e643117de7148d329981c14d5a6c30d7a4898aca8e0d10e5568ad096113d3ca4db898141297b6063708e840595b4e1e954b79b8b4b0f8ffaed314a0ba7c3b913700a65397a338660e4304eb7dad08f326dbbb987fc6203e0d7c7b9dad8948400c82ed707c3c360220c20d0516b367e48107cc93ec4b2e2dff9b7892226b665a66096ef0afea2a13f896938a3c30be8e1b3f3893bedfa1a8bfe31114f5ca785c9783ed2e1e10e5fac8ad89be10fbcbb89bb6e62a11bc5729d6c3f9104b83aff8205cd84ab674d8690d223724b3b2883ad57454284d9b53f6ca0e628adf64669f13ba923cd81103833bf63de19bef07eb50d8dd4feec09793d5fce031bfb686c7a5085f442edfe9d2ef459a1580a2f345751c9086aeb9c8a22728cab6ce0aad9d607af0af70f4d2b6c2cd764f636cf1cf527aa899f0d6db4d4b83e3ee1078017fa74ac080580695159e418facba20a3e02286509834f81674ecf896056ea94fa98021531e1602acd956df05fc9ddf329531792b00aec7ae7a67fea04f0e3023f88d7f557a38924e3b9ce52b3121cecafc0612d43afebe99f87ef739634480dc01843f4e3bd7f5e822d362fd4fd7a3f18a7c46aa1a9934ef807ad161463ef5675d071dc52ff0caac2a3dc4903c702a2d8c6c300b342953088b8207dc66d56c90903cb0eed2ef8a9673b6365dde6ee2088fdb4e5bc385775acb843490f71779143dae3e7e61d48824ec039bf7cc34afccf02c4008749bafa2475dab12c54b5979adc785ad7f5baab1b0cb7fa53b0075b917b673174580fdf22556079e50d81a336f042de65a99fa614fe2c0ff11bf7539674fe5d81d60bb720934d63ef7937dded244539e2f098ddb301d090d6bbc21a6c9e218de0bfef700f23203d867007199704dd99d3d9864cde573503ef99a5c4eb933f7381a3b3181fd0eb195d5ad6d5ff1fcbc767bde3714983b73e2b7105d7402ed4fc7ce01f833f59f4ef09ad91ff3fe1d69b72656df47104287d677f996d9f854c3a1b17b1ef89bc5a656b2a7d6749395cb20bb1e4376f1221582189783faabdf75549da281235656d74b5398e74e9003b87c69d47b6223791df2b1e2d321aff48468902e2ba1fbbffcbc334fd497ed4cf48abee97418af7994cfc687c3844f122bbdfb1aaf918bbf2e0a74af7714dcd248bfef5684812e2b93e92a4fd0059447e16dd52b319e373cc302bed8464cee7faf29f6fd5a8e615edefbbf9a86c353eee4ea4ce8d22ff258d7bfcb7659cf0734e2b84f00d1add3cfb8434bd40230fbfaf6738e1463a20969573e5f71abcbe603f4878010d047103e9e6e4006414d82785d44ab63f4bc0190c18951505ff70424406dedc0aed19ac4953b8e6a1ba53577386a65edf549af0ee2c6e4a81d8f5767aa4dea43c95a8570e0e7887162d28ab669f62f69a83be5f86b88d6a963858eeb2db12c939ab3f6b9541713079a61cad6bdd3059ec3fbecf4da5421a75a2038bdc9b4596aca55efc040292c6d7d51041d5828a411f20f9aa755fb95331863c0fff0fe9d75b12357b21d0c7966b7f6e01b2e0c3556a1501562bc8af093641da3d2b643425d401d91d3349f375e9536692e7813c5a69f91ec0c61a4371904349ae99c6435f79a82fb720e12b0412f71758a5c28494cf2f703eb6e092b5607714519d1411d8875b394da6878fdd4bb130bdbdc11f83b92d98bf66d95f1cdd52391e6167c5e293f34c90a9e2d35830729e91cf24c5cd4d1a1039e9d0c4bf0038bf31f046602daabeaccff4dec6130f6d6b35e9fb4791bc5b6594706b694e5d154df372cd5d4567e407769db1299342d6672dea2ebb324c7805335a9fee0330b987f249dd429d4d5d43e64db525db8815274fc6f7b3f7853561b9616f2d102d210f6cf4e1346f788f321a0296e93193dc43346449ea12ea9078a3f0d709550a5480aa88787e0a0c8d4c923557e7dd3c50402383b56716e0f4bdd8e9f77ef0bf682836da9370d39bf3f1b70cc382c0a9d5346b697977ee5992642297c72b227560c43be3a4b8eddde211297d48858fb5b9bb92687a66c4f26bcf067dd174134119982ab05d30f42c9f4385ae22d208891390276c053653059319da0ca69a39f421e6ecbd5314d3e82b570bbe1e8d05e30ff76e982928a768cb544243dad5e1cfcecdab7e62fd54962e5e4f7a556bce62353dd2c8bd849d9028fe148f33c72f93b219ab33c90bc61f65254f8713868fbd8310fa79f03b22f2e84d3213bbffa7b47d8ba2f4a49cef8c2e8d6fc8d9f7433e56cc750c56d7cff337eeb8d390007f6aeccaf50f771224bd2afb0ee4610783fb2702eb87b9d26a49cb5c993877715d803fdc1cbec94b13bb3ff0c1db7bbd55c6cfaafdb965d490f3949d6617c516b446e6e4347b5351d3018e0d85f25e5d97a2e462761ca6fa2302de383df22d52bbd190f7e7f69ca03bc6cb89034f70465e67a8ac80acecb79aee4a5f09a0f293634bd2c22aeeb064b0fb354a39bb38aaad300c798baf41e8e302586a0dec50d3ee7cf7619582aa02b003a119fac9d3b72192008b8a6d9de5b6822eeec428da7249209c539f3a8ea01791ddb8a7f7628bf7a13144e8cff89659e1309f0f1f4cd286c87db50041da1de6e09eb84210d9cf0f1420b3759e3c35594af2c2fedb312754598b4a3e436a9c3fb3faae8c6afc234256d119b524cbaa5f9dc25e55add364373c821a4cc9c9571fc88cb1aa4e5d53fa72a88fa31472c0d0491cf95231098ceb76d53b9fdeedd44c1f26afa0fc00f7aa918a1a497d65653b637a32fe5cfcb468eb228eeaef9bb106b04e716d94d90a8f0c11b37a9c5eb3812f83a472face6105c78fde86131dc0538ddf026bcc22b33d38803e2ba3659ebe425f57e201e75b1f0df4485c2f682ec5c51481e75672e089a8378aca10e96ef41bd2a797cb3ee7f83b2b2a3be89d921155c5b49ce7fbe2fb5db1c4bfa470e1a5abaab7c54d919c8019871e013be3cc7f85383c5d2dad51b695c0eb4e296a8185b234e0fc1ce6ba9edde2cc31d79c90cc88a9423d5446008a8a965e12fcc7072d67b10e3dbf322cb6f46f7c4c7e3092cac5c211e669336322bb4c7d933d05ec4a0815bec5654e040877fe4b8447b1614dffd78cc12b294cca42434b47f6b6bcc7840ddf8230fd8061965c7f2ac1908a23665e881660647faf8fee0d16a4c106fbd20de8a9e506378bcfe92e7243e7726ec0204580eeb82328d4728c97ece7adb3e82b461931e3d7f9231db4b97ef538a04c50e549415e7efd036f6b64032a8b31a3fe517314a82db1199f1e8167b02535bfa5d6f707129d4ea1e7abd727a038805cb5cf548f7a5f2ecd1fdbb2700921dddbf5e472fdf83ecc8c9213ce4324c7ced7ef4628ae019bceaef6556ef82b1a8427cb2faacfd3074d9e686c39cfbd5313fb766e6a17c6902026dd3599a19cdf980a205905a04a2acde838584c45158a73fa64d231323bb92ca0714cbe6d5aebcce8c058cf9e345d19282b74eac07d0aac40c2d667f3537ea5568c828fb292f570c232994a7f543fd633b011a4f56cf027e6c4b435eac59cbb32a93bd6e0fce1ce8884b737063900a9345f8c42e404b8bc8bba7277897aa39afe10116b85c134ccda98cc4219288e68833e67311fe4319c1a7f4e4145f5c76a1f925a5b00b5313c80156ce54d1f3739c3cd19badf1a3254b86317d0c5691865dd7f6ebef4e7f3c6605e219c7742b798ae0b4b7f180fc7633d32d7e55a58174bc1d4ef433672e98de259d69c9a8f8a440538c0655866203657ce62fbe9b6cc4278c9a354d762cea9d8b0125566fac2edd1eb9dda0a83428a43c2e6a7963cca4355063a9430119410ac36c2ffdf84", 0x1000}, {&(0x7f0000004400)="8bbf7629efe09e01c9bc74b559d33afbf45d4c865149925582ff2aaef4dd083cb49f7e57b9bf5d22d070587cab0b3446b0fc90d58736b5eaf8aa1d684de75807b03d78dccc46eae447491d850b4178a87fe3d3f34652b793177c9da0d545e5de495807214188b8dab91bfd40998a24c4036829c2a90897fba295aea5c4a457592da9fe1d06b0034969be51ec84aa0622f1405da25189f274fee40689e992edd76a175aa6a29bdcf1a678527f4f3cdbb662c5ca7451af14b3227c1944d8561bbb2c38f36dc1a38d8b34cdea45c840a4e324b928167d9cdfdb2dc66a0d87056e2a24a0cbf40fe32a5ea805a2cfc528900dad9d0de42b601d802734e4a0d50956d568d47db8f8df6cfff35f71017b52a14542fab7a24374076930748c48bca672a3672d2d09da644508c3ed17f13e1404e6724e401a3652ca7e7bd611a3a679f36d9ffc41a55ffb301abc2add59227a88c9546fabd3c8982eab3c16e752e32469a31c6aa351fdf292796a7f70f11e7b9b7260d90a9a17b14beb73039b518070bdd4d6a85d95b2d1ee5cb0d880734280d701940489faffd5388700a0c58bec2ede65af3325625a0d14ade7756d802c8490fde9d59c1cda6f7574bee9917b0db7c0191d94e4aea6edca930264c20e0b4375ae158839a89efd23a2f7191b097d020372fddf994248d25443f3fe8bd425ece2ed98e295b3e8d544897f447ce5875ed69c582db51cac8f9e4f580c8f3cb820ca2ac5ab0aefa185c2c6d68ffa008bbf58d906a8521ec279cbd6cc2b5e4314ff1e3b47bc048fd8833df51b185412b32ce6f0844d58c45cc1968c65236d4d6deacef9a8332d523a6fa867ddbc09614c71661488cb989bb46699f867b0cb40251e1f5e0068f0f49b90471579942d64cf27b85011de30c45a75d31c5bcc7f465cb4b76ed1667de5cb67bbe706ce41ce5b9a87873e0de05640e9b97fd01e26d0cc64d65e4982276bb411a12c29841643c61262da56f1f7dea492492f90bb292b06327cc48147568eb73de86c29b00a86b3639ca35714a54059a1478c6149069fb48425f4b35b81ebcb3b465327f987756613c44c05f388384fc5c7ea01d2c4cb216314396123a206ccc142c1cc71516793d3dcf763b819cc932f10f693c1d5c24130a37d932ebfd0baa255f99d3639e548822a42dc13b1147203e2e5fabde8193a433a948d923183257a41bec149c34256c4c037ce25eb2532a76f9bc64ebd84b4edb7a90e3f6edb0fbccfca6df4f888e797c1878c2d71cb59e0980b0290866b78d30d2a9e51f9661724f44ff26c6f26c05022576a120cdde0fc827811e4286348e363dcf6eae1a8ef6f94f596b87009f53608f2c8edd4255d468b0bab997534be6184c2b3608b0f8fef7ebd322b0da99cc407526cdfdc095fa59a9945a9234c078a6431a7e00ea5159228482d26d4f2ce93ead428d8bddbf8dfe92dc9957ab4f9f710ae22a20f63a7c53985d4934ed283e27c32cd45d148f4965444258e7f5dab0881f17f9d61b7cab7d20710afca52c444a3a05c2d7108a20fe648f8290b5fd22b70c32572c69ae9b68c622a77a2707811df93c143aede12c3ca9acb51857b6321ee85e2d9158daaf56d452bc04e49440dc7be51615d3155da939cf6cb380d8805d3094390bd494e96d5b42456a239985fc39278f6155628adb1597360badc108e4aca8a9865464d89645f341a4674d4aff7eba78e85950e8868fca661f996281494095ab9e26fb5a55b04a0d8fd455c77b196a1e325e585630bd989b272695e5432d9b8eac5a8711b8ca06786777fe68e12a33e88374f8ece11f86653f7a157a4f133574f4371ed9a0c12b2dfd840c75bd1669f1a80b4f6f1f4ca41febf5405ffa6fe7924d48886dd037ef5301532c047acd80525ddba5912443350164e7fd6e93956fbde183c44e9620e417f9e7172967776b6aeab165fb1a18d34ee43c4d38c4a41b6634a7b9f5584960b1cb8194f3f18523efca6a9bdbbec01060083907ae78a5bec113117d4aad430d32fb08090b248b01cf5a42f4c317cd0720c70dd7dd2f35124aa7a3ad06a6b73a236a311a0d0c518f076bd98eebc242badf29ffa628b58062b166a9452de53d6f19feae02eea9b2037cc843b4f4eb05e95264476186aa250e79a4a3c0a7ab2fe46f8d932b981ab9debc4c8a6a5f3939d858fe9c0624a6b098e203026bf23c3a7b1d75decd76096ba202b74a0b778a6b7ab281497a24c1e84c72a2a10a474ae27f54ea75b190d5113de562d332e20764828197d1cbd247853b914caf24c815a715925091d12c0be622eb8b9111c2d172ba3bc34976883449a18c61252db9937752183c0c6a64fbe178788ac3f11f49aaa248b796322d03198ef9fb370608fc2fcb0d15248d41d88e7885e19d427baf2cc59de8f83ab01756e22e38f8ee6d993a647157ce9ee5d65b495d3888dfe9a0436c428b207328f56a8f20604688d5b024fecfc3be3072a49aa678847bf6fe57a49a129ac72e601f3dfb1f416d3d12502d5b53860dcd5dcfc1a47fdf072eabbd3255f00d5eac0044c3635510b162f0a52c7b8cb04be6f5a35129813b53fae28206909aa6d85006c69f174029a2a46137d4690a86d700e53e2c068ec0fb073b04e37da67b806b04be60afea3fe9e4ebbc1decea09864abd0c8b7dbf06459c05e7ac1c550c661c104f315b2712aa7ab283fca0b7cfdfbf33aeae3b5997b0844cf637873e0d8c5c54c4e373c5b10321c147390da3abe39abcea936bdcc5a0d83a4c21313ff5029310e7384c5a19a5dbd89c8344cf850011923782c1ccdf9ee408767c8948ad94f75389f1b18df4e27e382300f537d4101474ae294a498b0c1237a5d09506f8cb5ca928536bc99a01b6943164d7852aa3d252eb110d5562463044df575e12615deafc82e20c4c138c0dd562f765c021276bb861c3cacf3eafad3533500b957c5664ac860612903a8081d02fe1108f48a9b709c8da0bef93eea79f772126dd41e86dab0a4ca5ab229192a26af9530b39972875ce9d2cfc649a491e159466eb760fbae16da56d70c16247d84062b98223d9dc158bf14402612e22051e0ebf21f139f4712beb10ddfa9207df2f758c71b3cf1d317602c8d698e3f53935c9ba957567c5e2321634b88bea68edd161b391d1c63e9f070a23d7eb2b78ad5b56cf18090b8845ebe91f32e20d48777fd884b55517f9f345026a4c41e5f1c7e2d559bcce163d278e2d13358091050e6760ddaa2b761d253790e95d56dc8d65c0f702c08d5553f9a9dd4f364e50b3c98f43d8a5e0f043c157dc3f34d4f375be18df39e0e094405112f0f43b05c7856d5a321fdf53bcd909372c8ea70c217ee984ba03ab567709d647c3e95d5cc100f7ccc284aceb1f463d7caebc74b2514d5fd4c0e96059e5a76d4921f23dadd78aee4a911117276cf55e429726a8d4bc54c9dee10d505a83d0a2b4a3317db5158f69d987094609021f32ad2279e1d8f1cd0cb2ad1c08022ec3cb9215ecf8c08ddfb14712b3b42bb014c26dafd54afe17f05099edb6cff7edd67dd05221367265de0ca9fc1e8ee7c84f26f42732a6f3764a4ebf07168231b16ccee6b8d82b5f73b748f0e088f59068c317cafa09e8d658cb512fa8a0d2bcce97046d73eaf30332b519f915b687c09ef79b7c7baa2ed532593a36088c863ae853ddb3195a3417533ba8bf1e071e010518fe7a5d4a7f3717ff2b3def3918aaa066016ca2f618a8c2c5b04e6c4f2e4f38c9f4f745195ba38d15d05f1ae4e60ad61bc92f78fad000c026a0474a810a87c59e193113f130ed5b0cbe3e46369d7d9c098b6455275129569c74aaef70460a119da97cc082c55c6ca4f68b00644b3e76bd197f42dd9c837b94380605b855af2aa48268bf025183099bf545bdfe39b3d6a18267db20a6bd35e030b3537fea99d403626f24649264946afc0b592e1fb8dc496b1205052d48fb4579c44b88da414ee2e2c1af9f611f44aced02ab760e6d564b053e0fa285d34f46acc2e33a8298eb5d9e1e815bd61e7be00070b82dcf9358e4d3749bae9c2c2db9c902cabde19e34eb98d02d55ac1a191b1e7e18cad5243e2c971b9d92cb104c26669c1b99548eb7d59d9245f04cf5362886b7412ccc83a33273c2368b57ac5756952437b0d5e64ea763a55ebdcfc3867e4f8f812b71aec816b10a254ef6f862b856e05e158a4a56a540b7e2fc3382a830113ebe8ac6bd03f7d56d661ddf965d6461c90fbd5f6a74b429c82bc78ffdc0cb78f8ef76124b4697edb45afcb0290df7e68666c52596b934e4199e73f553120fcf21c0b3ca7e7a76b805a0a229b7f40e4483049f3711d80ee68d9bf6de2f15f80826e0438af9151a74f8be986ffeeef5f40fc66fa1be7218b8b5134aa0adc3afe08a89b332707f693df04fc7ce3bfc788a0c8f470587d28fd8faa7429202767c68d0294cff930b03c251a848a52fc1520bfef70e5b3df755f78115767923dbad460f4cee461325cc33118d48f38fba40119c7dd8ed6b7b3ebeef7c65508a05819cada0bdfbbe3a1c6a8d0f82d5130f2f3d6d96af3c76b7b6911e44a4861ef78de6461966b03aa05801e3ae0ba4f648588e8d7ddcda0f76f8ef1a76b1b16cbc3db37cadf5b825764dd3a2b2fbc850147112b389f9fdd7d96fc8f1a58bda0b25cc262cccf442b19034a93940ff9d4bff96aaae9206bdec0c23f3789875fe865cfa8954f75fcdaddee17e296b287a49b34dfe0def35f2b59da773340b002b699ad44ba20df0237860048083d5b9f7d69e045343fa645d1673e1d86254e22c62a6f883e32a9effd1c13d91ee0fd48b9d57f030a458f39e3e75edd571e11957f9ccbe64b774f56bb188fc93f733ab59c87a88703949cf873170c3dac46d53677f6fcac0d9ba3d6c36505eb34a9e08817e33de84576b824c30a9c2fc87d2abacee3e2dad46386bc36aa35ed23ea11fb4d2355e2a01ef7df578004dfa57e3eb00db2edf1fcfaa4a608d31facb25a5de6e08b35151fb3d73ec06dce9e3326082cf5ece63f0a3eefba8c5fe9a69fe30ac843cc38cac8442574e9956d0fc8c257482458c10c5593958c3c1d3d1915996780c5beaceb3990246ba325208b9a9e0eb163b153c36ed73f171394d8b5bde20bb392961715aa2772386ecbd442e2ef7cb1d09d65d2623f7c4e2d79d3be789048e9a2060fb7b887251b0c7e0a7cb3dd4ddf39195502710f7d3fcb358fae6840a411846efc0be2b2716b3edca31b1a731d4492d686b12f9ec8c3006aabeb41a8c2b9f53f7f9c75010cfc53eb639596a8d301b07eaacaebda45b2547447c14a4d087cb7e404a2c94549399cac09aeff7b52a43908f8b6d66e035911b34435f8790e77abb2b407b8da897a5dbbbf0da378af668639a4fff02a9393226233766e2c535a5de7e0bc187431b72608c9fa8aea2c6bb7bcaa9d2c5eaf281fef168c49a97ca7ada9e0a9e433603684acac2fc181b57d4c899ae4b619eca7c308035bb7ca5b94b3aac37efc0d771c95aa9167d0169fac65ef5879e27a6b9d51721567d6f2e95170c9080fbb83ab5b0eefb80a48fff0304f71afa4062e84fff8fb96ed126564f27b6ef936e0656ea998a5adc2ee19435baa31b2b62bc1b4a9f13f25ab5875139bad731cbaf1d6e4992a3e8831118fcb89e465c7bdce1ff10a1867c5817593f18cd62f82b4107a14299ab2519ee05ce3871982ca4f19199b04512c7548223930ee2be18ffc4b726777711df1f07bceaf856da75ec1a35bb04ddf63a7f3b68401c74ac322e975a94e569ebe91fcf499e6", 0x1000}, {&(0x7f0000005400)="c1b41f9f3168625bf5ece59ce57a8f0d9f4bbb243d07a923cfbb0a8859895a9a9bdca488d9680aa3585f891e689608819fb9978d57ad1492a63681bff65080", 0x3f}], 0x5}}, {{&(0x7f00000054c0)={0xa, 0x4e24, 0x617, @private2, 0x464}, 0x1c, &(0x7f0000005580)=[{&(0x7f0000005500)="4b08cea4af7fb5effb5ce8352b9a28251987308b2117c8a4ce32670e1f7d823d705b6aa4faccd6f1c1475ddd9df92275c681089bd427e43913457ef96e241112fefd00c159796d8ff41d8044f33bfe8823e3795ec3732aa52b32f3c38eb9df5e2960badf856d67f6158dad018b85dae0f5478fbc407ec2c4b1", 0x79}], 0x1, &(0x7f00000055c0)=[@hoplimit={{0x14, 0x29, 0x34, 0x1}}], 0x18}}, {{&(0x7f0000005600)={0xa, 0x4e23, 0x2e1, @ipv4={'\x00', '\xff\xff', @broadcast}}, 0x1c, &(0x7f0000005700)=[{&(0x7f0000005640)="5ba2cae580dc49fc169d0641e18bf968d9cd146b0af56837401f8288c475b9bf7d4fbbee58c73519ff043c0ea253b72d023724c9b1fcd73985eecfad808fe32b21790f47081462d519533760560342eea3ed0b4579db1b774695a0233be67f6e300d45494bca87a6d6e7b789daba9203889493257bc950592ccb3f8adf120221d25e3029743efbe6b5cb72674d2e75853c122ccfd82b8144a732c54a3b29d7493b4b948b2d57fa852597ae42220101a4c93201cb0845ffb189c8", 0xba}], 0x1, &(0x7f0000005740)=[@flowinfo={{0x14, 0x29, 0xb, 0x1}}, @hoplimit={{0x14, 0x29, 0x34, 0x20}}, @rthdr_2292={{0x38, 0x29, 0x39, {0x2c, 0x4, 0x1, 0x0, 0x0, [@dev={0xfe, 0x80, '\x00', 0x1e}, @mcast1]}}}, @hopopts_2292={{0x20, 0x29, 0x36, {0x1d, 0x0, '\x00', [@jumbo={0xc2, 0x4, 0x8}]}}}, @hopopts_2292={{0x78, 0x29, 0x36, {0x32, 0xb, '\x00', [@enc_lim, @padn={0x1, 0x8, [0x0, 0x0, 0x0, 0x0, 0x0, 0x0, 0x0, 0x0]}, @jumbo={0xc2, 0x4, 0xffffffc1}, @calipso={0x7, 0x28, {0x0, 0x8, 0x0, 0x23, [0x100000000, 0x1, 0x100000000, 0xfff]}}, @enc_lim={0x4, 0x1, 0x40}, @enc_lim={0x4, 0x1, 0x2}, @jumbo={0xc2, 0x4, 0x7}, @enc_lim={0x4, 0x1, 0x3}, @jumbo={0xc2, 0x4, 0x80}, @padn={0x1, 0x6, [0x0, 0x0, 0x0, 0x0, 0x0, 0x0]}]}}}, @hoplimit={{0x14, 0x29, 0x34, 0x225}}, @hopopts={{0x170, 0x29, 0x36, {0x2c, 0x2a, '\x00', [@pad1, @calipso={0x7, 0x38, {0x3, 0xc, 0x7f, 0xbb5, [0x40, 0x1000, 0x5, 0x3, 0x7, 0x3]}}, @hao={0xc9, 0x10, @ipv4={'\x00', '\xff\xff', @multicast1}}, @hao={0xc9, 0x10, @mcast2}, @hao={0xc9, 0x10, @empty}, @calipso={0x7, 0x10, {0x1, 0x2, 0x77, 0x101, [0xfffffffffffffffe]}}, @padn={0x1, 0x3, [0x0, 0x0, 0x0]}, @pad1, @generic={0x20, 0xc5, "53108c437e0999adf47f6778c43b28cd4f49f1194d80509b18f68b59bccb057032d5b475c3e0769149a1f8d3d8df06aa5668d8ab3d9a0e3c742cb3153f7b5e65f2ba90bf5fa46a10733cd35760280cee062eb242944f78223145b8bb8713bc6c2c42b7421fb5d762e7a5937a45ad7e93862a704d20401a706cad5be91e94d120231e89170c76e078121e8b6459f06f93f5076cb424940a59f2058a47c441fe5ab195d44ca4b97c8df9c4e7ff9973840fde7b426c7153f69995b820745f67b528b6880d30c8"}, @pad1]}}}], 0x288}}, {{0x0, 0x0, &(0x7f0000006e40)=[{&(0x7f0000005a00)="3f24a61dac45a3915be9e85c3d6da7bee1ce6e27a2c1d3dd2ff8ebc2cb17456e97bb6078f9bd2a94fd0b581b9b4cf5c9430c66eb6761d3e4b856e0f1ff5b1c4a58a33e33c9dca51f68f9d3ea06214ba703a44e2aef21a72142c1bf4b82cdc1fec109f97af66ee8e0449c0cf7", 0x6c}, {&(0x7f0000005a80)="be1d1409ddf392b46d366bf209c29ac46157627d5f0abbed918a67e969c2fc05b296f94cbbb7fa62ac61a54657e6e7335159e925159e63943c4e0c04ac85c8bd2c231266ed49bca18a54c1e01f4048d991c98a18679fd07f30815a91a3ddd98a96f003036347af78d8b3e802dce608606c106ddf8632958b2a69281bb5b97df407d02a6e29a44aa63cf626802532ba242286e4769d4aed787adda75d72200ebe92070f6a558527273e5e185168353e3d5ab078c4a75a24810c7e73f5990bc36ed867ac3003ec50c374962f69cf1c8d239809b2", 0xd3}, {&(0x7f0000005b80)="dd3b04f9beac32d22f8c9f9dbe23819f366edd0bedfebca2f6569885471bee84bd0997f8a2b7bc3db803265bf61414b77f267a5d45bf1257fe10a43d890e3aa39201be83a0c3783e55698de6f828d17096fe6c9b2415dee92975c9eebec69b05a913af2db2415e4d7141f01c5f5958eb53d8581cc49d4e4d7d9b960c29bcbd552268de2463fd544cf45d190d28ef0a22308fded4eb6a04e0b131ebbefdf10992faf20d282b0b3421", 0xa8}, {&(0x7f0000005c40)="7fccda0cc8ead8aae0408a025b4df83ab54fbef5531598c67d1097ecd28ea443b6580499bbd30484f1be206f90ac72178377a35f5df1db4bd26972446dc085019baed3763c2f7125c6fb00ef5ef3c3074cecbb0e7e2ebc35c88cc8fe883d0282ffbedabbfab50e048ed6f91b84b89ccbc344dbcfc642c8cac6f34a3b294ab53866f46ebeb0579b388fe36483d482279ae1bc6e0dd1e2142485b5605b5f4cc853b0a374769c7677ba156392c2d35ba0c131242bf2ad15b567666f0e1b0034d0786d3ce9b5bdacca5671ecfdc32b0f7002fa15da4fc6d4331df428bc733e203bb9aa1507af4f09909e", 0xe8}, {&(0x7f0000005d40)="bfee1056b57b8a87fac60fc21a46be552c01e563c8e40bb9a903f5a654b36e82f21c55ef1ed1a5e179705d2626b7e5ec0fca187a1b4c1b3c08b2f49fec875d30d734ac1f6ecdcfc0a4693a0386de8362fe491b42631a86daea0fd6ca8f34d545584cce2c966e7a35060bc759da4de1b0ba6c52fa9332bb156059256bb35b48b1fbb9ad5f8967765b24a87230ecaceb481a2a434aa4db2e9961f7ed8126e3175b807a67a465b02b9bc8e332641aab857d7047425b8e4712bf5b5e9557aa53ef857de882e5d106b5141d53d87b7078c19664f260fa3e7f54a6cb281cf2c155642bb11156cd56bf882b6d73c643c2f828a93a7c2f36cbf90e2991ceed4e9062b43b13724fc47942b850aa4250c8133f0c8fab15589d3caea45cf891c47b5e3e5b9c3e5677ae5186aba6be5217a4d306f5ccd0645d460b7d2bfdcdcc48b643ea54333a8d66ae864508d8922b1ad3c44f85b27ccc92c52c96bdd9e2d4f918d771498742af7a0701610c906ce401e3f2bb400a67f711b6ac624986fdf0e372391bd8e30b07648dafe3f0b06976f1ece8c4f33483895ef016e26b0f16f8acf622aacc5bd9c69d5ace1b7e8ef753d339548f555bc94000683f1384a1d8f854572a15ab74db640a4aa77a03213e8a1bcb1f361cee0dbbeb05e48fe1700fd262dc094672dd1a1c8f2cbe825b246c0df1e7b9b49fa055fbc68e334a2e47dcc50ffed3a65dab7dbda3ca1788681cfb74e212d445c30f9b603aa3751db72eed21b90add76d1cf6306b519cf61da960a0e472ce1cd4c779d7227f01b8d6f61636b1d6d16bf03b8f8dfd45f4021587ac4a1ff66aa6ba7c8725ac91f599c4db8279cbbb91cdbbb45a18d224741b71eb79ff296135f96c67689281242ec0c1bb694618b5597603e3226bde62f885144c2f65b160a291e08e8ae248180d9cb20f9635e15dd7446f4ca692fa8ccbde5a20f7951fb33d5a50ccba4f509017555f6699dcd1d626e29d5aea8986d96d7ba488aad9f512ea50bc665e12e5e9d7c6390637a64f5cf2767e717a8ba1a746c5aea061881034e89f58f8d075d673307ec1cbc1b598f18cad0101f1abda02c3f5fe6f72a0e2929a95923cef2c01832112a47c37b91cb8d65c4ca41f81d450ba92a32d482c59d2a839f0c36c1c301f3992307a5be2601088146f9fe89f33f7eb0a4b9ddd52aaee0663f683e352715a0594f4bea49b2f090c85fa7bb3ac11f436212edf7fa7864cef8296a675af92b5ef5e5131749646ef451ba3305cd00d764c997b2da08ab065e5c8c2d0e968303f2adb34f118e3f1159f47e51427614d208c5e2a4b2247d31d92df9f87197fbf5d5637bbe85e47269d3a41e2d4f7a1ffc0a6281e6624e9fb10f1eae343ca5b94f19f92041467b91128a4a7eefd26d46bb0e816d14fada6c50de931a9c6ba006aa3156d86a01fec57725f955506953a5b8364f857b130fee93e3a8202cd9601a25e03fc301beef7f66533f55cdff32fc0ac0ad154b864e975ceac381b2d8fce3232a9b8a9b41a7e4a2a671f57c275250efbf75f54773b3b00dbabc21e64ffa508173c9deec95a2e4298286c3eae6fbd9f22d28c16c475e2cc4c5d68b6e421b746595601941da19b4824244caf8cd6d1013fac0cf7b1e25909150c29e31d077f5f0f2f239f9164526191c1e604525ee6ae34314052e13b38ea6bcc59c46c49c3f332ba4129acc4edf364eb5e83f9affb943368c81325e787a85c96a6b95b6788d62e8167e377ac4d9f548705ce15afa8ce9d76b3ee238a1919b3ca852669df6efc695daa1677758ef49c42487e5059b48d26816f9e7e352faf84ccb0520f45f5ca62d8b45ef19e02d51010d7dcfd572cdb1fc9a884eae5e41ad0c5f06bbaf2e909a87233d79c6ea44668452ab985a92175890848590b4c2aa322b9b01828484ba84f6ac91b262cd42c8d88bcc41010c9a7cc1d3afd617a3cb0ed2dbaa91dda6140b04586f85e7d92ab92989a1cd676060800ebf92eec3a0b7de05e15100ba916f03ad1dc7753a4426981296bc9cbac3b773a097ab16ce41f4067f5b879e7b4d040a3c9360ef77c6a31e6e1b6d578086bc82963e910a9544b3ae2f771f5099932a18391ba0987cc3e66fd7849d7062b64ff12b80c9584026ca3751d49e6b78ddfd7f968bba507078b2eb90bafa94dcb1c9f88c777e538bf48ac976bd1df87c01911e35e3d47ecc56bcce9d6bf8559b135b6609af71a73db8923d70b041ba758fc1e7ab1ddccfc94071cbf78a4ed929f5ea9f1cf5373753323a22f99eee1639603d48ae68ef868f793ad02b81ec6ad6293df0d37c28a5c251602670f9b7a6a21b327abb84d46dc6960a1f4d28a9aac3dab1d4b919a3f6d37f36147dab3c79c94b848a1b995bd396a8435d4ffa5e78530a802f79a3653b28ea52e154347d0c960b452b3c7f8341b8b26a778d43466d9a54403233feffa74ea48a68f5fa8aaf72f5eb51e3b07f6c2449c6f5f2f970edbda9a725d80997bebe0c54373275b56de857a742e46479e09e663fe756b11ac45ae3c141939cb9ce03ff1e6fb12d01be311ce5f59c9f124aa26b2bd87e5ffdb7f73545c7aab71633f646dc85380800aecd2975447d87e7eb6c16addaaecc919bb401789952e9ff14705bdcad744af1ecd018bb7f79bd5cae90427cff478183f3b28233a55ea750d3d88972869c31a2e0c932d06a7c186c4185efa36233edc88c186fcd2f1544d382cc0744167f54e359d8547538f7f6926ae017e5ba662843960ce1a3140fa1290f9541db6ff9507de2fd0e8e77dc0a55c8096ab22a6c21767e18318706feec569be584c13d5337d52414ae60448aa383e780785eb3f196c7f10fd376f911759e420ddb27cbd744fa2a7e2dad5543a07ed103e5a467ceee127030a0cb48d98f7b4821926dbf43dd4f95125feefed4f48c3195a61d8d8a1e848244ed9f7b332eaaa88e55195a509daf964b224e7fe34b5092f6ddb6b309d57963651d1ba1c447e02711374efeaf4c380979b1810f240d39708d23476c686edece9d991048f7f710e8a31a0bf3bd6479fff543a0698933261c29955710f294873df42599d91aedc0a9f9172d131275a83d1bda56784cf1139dc33dd33ca0cd2d0c4ec2f4912e2f2ee59c1b3ade58553f7df74929517dfe28b989cf1207f1e5bb530c3afe62f404af18196a90222323c8189a4bac438ad35a1d31efa40e65fde299ffc582195d2e995666b78cdb5e2d10913c935e92936c4935eb503977c635bc4e34ff5c26f8a8dca7bb55c6f99b4bfaa97b75cc08c75dd133fb8b09fa028a4ae9e9529121fcd49bdad51f0e128814cef7772f3e822cb29bcef1544b0c54c08b6d40f24727897a0a0e9e4a15e829bf4262bb0c5fff063e277e8e511ea4b0c91093ec3bf215acdddc3c6fac93f8996a2d10ea838fade70a41b354cef38a53ba1c71343b0b4b209f671c63051a9019f34355676a69176101d92cc0a7f2dbe0a4115c3f9473a9614b0684f99feb478691b9fd250a426d4e8bdebe27191a815fdb49f1ff2bef88f8cdd20142fc56ae3893c68486dca47b7a60453508bdccc410ebb0e088677ac1875e562624750523634956baed75fe7a10ef987a6e4b71f303a2c41c72dfa86e33bd7cde76c6c033c74418b95cc37638314e70e6116d7f321a11ca13db037bcbf41ef8b6b8bcc95c32d8af5f984a961bc47fbb1c5233d200a95006ddcaee294ff7caf5a8aa192023504eea099bed488ac8074c0df343b1e3d73c3073028654398f81864911fc106282ff5e22f2da39c709380f71e28d391d171eec72fd8affc3dc91b2d0096e2943270351679483baef9671325dfa1ab923659257a973820c2e5e4229eb630fa6a12719ae8b0ec0a498b69bbc94a8400c54b66560579736523a655f7481d9330cd7e6d9db3e359f9478bc1eeafb969a0c4942c0d1bbaa20b5be683b0a3f4d35d0df39b829ccbbdee0c6b39bfb1964f7b2a6f19055a57de2062caa3ed43595a92ac949421fbc659c31256d5c7605ae88cdb119829824f35b256b3de1668a50a290a356e32a645e8746069e4e3b2aa65ccb2e591471ae241b63f4495dd622d3df0f312eb9e243712880bdbc4000e0f9e490e7580d9c1f08f7144033a4af0763b93c4d2fd1e6ace9ce8f1c9760b581d51af613eccde9487583fc359ba412e590f19fc132e51564a23c0c2b767760a324547a7d04b271915adcaa436aa17fd1a81002e3e3a8f412b5d281126c86c0b83d82b4566b38ac9ac601d5507593d72b403b39a6bee4754adb27e33a44a713c84fccbd6091024b107940eab518243b4d01b02a253c86163826133717c04268cbe64011b1834b1568612e884211ba20f4b8503f3837ab8e7bafbf4f55b888fdfc4eb02a8aa592341ee755252af23f5db80105360645e0071e0098d6a0cc82176f471d67a18f06c3b2ccd849ce65f4755fa0ab802e77b0cf5f24229574e4e2d8446e45d212d4f2d554221a97b37a381f76a7b918572b70c7f9e37f30458fa38bc075e3ec575ff9e5c1af6499833a90176d3e9e7729ec9c2c81920c9ec0b457c1641963f1491238c3df52cc8c04f66b1f0bbd4577cbcbefbc0c7fa266724bb6af72145af3da1f131bf7990a0b1094f3cfd1b44ee43656c95b9e75ffc8de888504bab8795951c4ab5672d5daeabfccf6771c367910e563e269e093343081264b813667bbd48fb8932f89ba32429659e92dcbda8ce138ca31d6232905c63eddcd4ec935c4a4f59d970325b102f9b4b52c09d7305b88116480d069be7ccee249dd3af3530e4f79c7169bb13fb5573fdcd7c164a418fda059b8d37f91ac9efd69600dda57525171bfcd554c412b4f40b2b8176eed980f1c55f4ae3bf1543f040ff13eceff746b31dac494bb5409ea6d26f62edc630686423a8a5d4348f56b38d276babdca09cdd13d5d4b509cf600e81e8b79954328952002f47b3508d2894845f64c19f04d2dd0d9680fc40494e39a17fabe122c50a823ac06bf04c52827a3714974a4ecab62d1034011d021713464839a6a95ac583b8b36d90ff9afe10660372d9a9c71b1f61598d574c3bc3d779a22a61c2a466a9eb738aca6e55b3c45ae1c1aa26a4e8fe34d7176cc9c25fe553cd1b5f98f046f6a63da13e2de4d953d3e5bc8f559a4edbb1a13af904aaaa832e606852b3ca47d338bffdd4fc2c576cd188144a2a8a4b6ca81ec6e2134ad19eeb0b78a55c62a978a891d4d49050c918008c8b0b5651e1823b4ae89139691177e32436da940d4357ab012e1e394532aaab6f03d6576facaa71649fe3c2a1c92c4d02c48d65a684b3744810dbd6f73326ae4caf6d0e7da163a6a12d8499e4f9f860481656df3d308a55941a061f24335d9723ea2bb34f518a7081d91c76f695fac623388b7b24264d44e6e2fc2a3dcc032e9bb45fe83f89ea14ec3f9b962c53be6fccc563857e1e87dc20454e2f46623b0c350d11930853a44becb53a340f0674dd9a1ebec56b933c53916cc15429bedddddc2d670b1085dc02f0e3086b7b937550934d3e52bd2b4d31ffd7d90237fda2a593188bc92f7790deb1f2f660af4070d2601af107d3586ee94b3c2924a397458cc7e06538b8bce8d28212073fb0706453634d287e5c1a62987357ff3d1125ff9c70f7e60fd4e23af41c7ae93224aab69868db2a20a8ef76b00cf1c7dced6ba8f2e313a222dacbabdf612d659fe9ac3721d9de164be80e9328f5ec1133af28fa496f5bf336c683e3a7f1dd65bb6fb2ea300686c7f1e3a7b5be1927ad51065d4977fd8ed4582bc0f3512acb9805e9a", 0x1000}, {&(0x7f0000006d40)="7534da42ab8ebea4423f2fc1354f832bf3197028a50eb9562a1319e68849bdc7bd0fe345802d5ef1c48b586968170bbe22168fd8011b05c979289d36b4d05161782c6b1c0a44e9f6081d2a3599c0978c", 0x50}, {&(0x7f0000006dc0)="56ddc9c96cf571414c61f6754cff4eb202fd95314598c614c24739b62940a41a5e0c18d4600b7f5eb7498fb12a81779558faf631673c04ffc0914544a1ff292c2bd37dc13d37c6c99e7a4446ff4d3aba057a81afaf1f1770599918c3ebca305a47c05978576ff0512be7c56b47b887", 0x6f}], 0x7}}, {{&(0x7f0000006ec0)={0xa, 0x4e21, 0x2, @rand_addr=' \x01\x00\x00\x00\x00\x00\x00\x00\x00\x00\x00\x00\x00\x00\x02', 0xa00}, 0x1c, &(0x7f0000006f40)=[{&(0x7f0000006f00)="9fff08a2", 0x4}], 0x1, &(0x7f0000006f80)=[@rthdr_2292={{0x78, 0x29, 0x39, {0x3772466af79949df, 0xc, 0x2, 0x6, 0x0, [@loopback, @mcast1, @initdev={0xfe, 0x88, '\x00', 0x1, 0x0}, @private2={0xfc, 0x2, '\x00', 0x1}, @private0={0xfc, 0x0, '\x00', 0x1}, @dev={0xfe, 0x80, '\x00', 0x12}]}}}], 0x78}}, {{&(0x7f0000007000)={0xa, 0x4e24, 0x2, @mcast1, 0x2}, 0x1c, &(0x7f0000007200)=[{&(0x7f0000007040)="28357b828bfac2c815793e46d9c9d23dd66973f83bcb7af7e6ba0da677af64bbab92adf0c17fef707fbc64e36f26513f5a8bcec1d766a76f11bac51e187e383d4253fc993b06042a941258142cfad391d252ea2c21c2faa5266c7301f229", 0x5e}, {&(0x7f00000070c0)="253f98c2bbb8f7e64e7974e455fee0a69cbebac7e94979a49d788709cdb17e886dcfde86226461a63ec59adaad90ceafee2b19023b8aa04091ea4e0b128b1b526065fbb757e36cd551000d5538082a002ad35f0c436a044b39d166659c4d2b5125b7faf422b55b4a2c7245e0d8717054e261c05c9c7bac3257d6fdeadcf5688aeb678e7967c3b704216055ab039b9e8b75569390c6d00d055249c4c4cd75826bca85d05342028bbf622467e9977d54b909c7f6579c5f97609552", 0xba}, {&(0x7f0000007180)="85e273e9152c799a6557a0362f4181b95bf2794fd5ec1f83b53ef4aab07d5498cb5ff8b1b6e2dabdeef02cd977d5025efe659c331887f2a56f1c3552356fa405fcee979bcc12c95776cbf788e5a06c0f33e39fb12d5dc2f755798f73", 0x5c}], 0x3, &(0x7f0000007240)=[@rthdrdstopts={{0x20, 0x29, 0x37, {0x32, 0x0, '\x00', [@ra={0x5, 0x2, 0x5b4d}]}}}, @rthdr_2292={{0x48, 0x29, 0x39, {0x2e, 0x6, 0x2, 0x8, 0x0, [@initdev={0xfe, 0x88, '\x00', 0x0, 0x0}, @remote, @remote]}}}, @rthdrdstopts={{0x20, 0x29, 0x37, {0x29, 0x0, '\x00', [@jumbo={0xc2, 0x4, 0x5}]}}}, @hopopts={{0x20, 0x29, 0x36, {0x3b, 0x0, '\x00', [@ra={0x5, 0x2, 0xcb49}, @enc_lim]}}}, @dstopts_2292={{0x30, 0x29, 0x4, {0x33, 0x2, '\x00', [@hao={0xc9, 0x10, @initdev={0xfe, 0x88, '\x00', 0x1, 0x0}}]}}}, @rthdr={{0x28, 0x29, 0x39, {0x0, 0x2, 0x0, 0x3, 0x0, [@loopback]}}}, @tclass={{0x14, 0x29, 0x43, 0xffff8000}}, @dontfrag={{0x14}}], 0x130}}, {{&(0x7f0000007380)={0xa, 0x4e20, 0x6, @private2, 0xb99e}, 0x1c, &(0x7f0000007540)=[{&(0x7f00000073c0)="9b493e883c245d00743a2979372b087e556d2c6a1a147d745d361241eb4bcc61ca9ee47092f93b118a97ca16c7f3a842e838d951ac2743f018dbbc269b17b6a381b46d849674990f387743d2570a9408de434f5c86042e5c4fd1cadf764ca1efddb267d436f357f56e347a8d5378d46996a2835df9b410b54472490e9bac68409bf773fdb170e4ac6388cb8cfa8d3c1b6d735fab86c5", 0x96}, {&(0x7f0000007480)="821ac7a68d525e25b36fce772aec357f3861b222a9f729837224614dc2d28fdad2e9a03caadecaf1cfaf620de3d0d25030f648b915c6558d9c28c4d201c71b708952f38d1261e71a4e4dcd30824d45eba558da6910cc9c0b5639637ba053bb8512f8c07e29b320a8f15e6ed4e29881cbbbd4d542481e463217f570df132147dd69a3b4649eb982e6182431", 0x8b}], 0x2, &(0x7f0000007580)=[@flowinfo={{0x14, 0x29, 0xb, 0x1}}, @dontfrag={{0x14, 0x29, 0x3e, 0x9}}, @tclass={{0x14, 0x29, 0x43, 0xffffffff}}, @dontfrag={{0x14, 0x29, 0x3e, 0x7}}, @dontfrag={{0x14, 0x29, 0x3e, 0x19}}], 0x78}}], 0xa, 0x40000d4) (async) r3 = syz_init_net_socket$ax25(0x3, 0x5, 0xcd) ioctl$BTRFS_IOC_SCRUB(r3, 0xc400941b, &(0x7f0000000640)={r0, 0x8000000000000001, 0xab05}) [ 2835.627121][ T3244] bond1294: entered promiscuous mode [ 2835.661740][ T3244] 8021q: adding VLAN 0 to HW filter on device bond1294 [ 2835.906131][ T3248] bond1294: (slave bridge1257): making interface the new active one [ 2835.920240][ T3248] bridge1257: entered promiscuous mode [ 2835.956205][ T3248] bond1294: (slave bridge1257): Enslaving as an active interface with an up link [ 2835.991300][ T3251] netlink: 'syz-executor.2': attribute type 1 has an invalid length. 04:12:43 executing program 5: r0 = socket$netlink(0x10, 0x3, 0x0) r1 = socket$netlink(0x10, 0x3, 0x0) r2 = socket(0x10, 0x803, 0x0) sendmsg$NL80211_CMD_CRIT_PROTOCOL_START(r2, &(0x7f0000000580)={0x0, 0x0, &(0x7f0000000540)={0x0, 0x1c}}, 0x0) getsockname$packet(r2, &(0x7f0000000600)={0x11, 0x0, 0x0, 0x1, 0x0, 0x6, @broadcast}, &(0x7f0000000100)=0x14) sendmsg$nl_route(r1, &(0x7f0000000040)={0x0, 0x0, &(0x7f0000000000)={&(0x7f00000000c0)=ANY=[@ANYBLOB="3c00000010008506eb9afc4cd8d06e754a0081c5", @ANYRES32=r3, @ANYBLOB="2377f292252155b21c0012000c000100626f6e64000000000c0002000800010001"], 0x3c}}, 0x0) sendmsg$nl_route(r0, &(0x7f0000000240)={0x0, 0x0, &(0x7f0000000680)={&(0x7f0000000400)=@newlink={0x3c, 0x10, 0xffffff1f, 0x0, 0xa030000, {}, [@IFLA_LINKINFO={0x14, 0x12, 0x0, 0x1, @bridge={{0xb}, {0x4}}}, @IFLA_MASTER={0x8, 0xa, r3}]}, 0x3c}}, 0x0) [ 2836.106194][ T3251] bond1286: entered promiscuous mode [ 2836.119444][ T3251] 8021q: adding VLAN 0 to HW filter on device bond1286 [ 2836.210799][ T3253] bond1286: (slave bridge1207): making interface the new active one [ 2836.244709][ T3253] bridge1207: entered promiscuous mode [ 2836.267679][ T3253] bond1286: (slave bridge1207): Enslaving as an active interface with an up link [ 2836.319030][ T1236] ieee802154 phy0 wpan0: encryption failed: -22 [ 2836.325873][ T1236] ieee802154 phy1 wpan1: encryption failed: -22 04:12:43 executing program 2: r0 = socket$netlink(0x10, 0x3, 0x0) r1 = socket$netlink(0x10, 0x3, 0x0) r2 = socket(0x10, 0x803, 0x0) sendmsg$NL80211_CMD_CRIT_PROTOCOL_START(r2, &(0x7f0000000580)={0x0, 0x0, &(0x7f0000000540)={0x0, 0x1c}}, 0x0) getsockname$packet(r2, &(0x7f0000000600)={0x11, 0x0, 0x0, 0x1, 0x0, 0x6, @broadcast}, &(0x7f0000000100)=0x14) sendmsg$nl_route(r1, &(0x7f0000000040)={0x0, 0x0, &(0x7f0000000000)={&(0x7f00000000c0)=ANY=[@ANYBLOB="3c00000010008506eb9afc4cd8d06e754a0081c5", @ANYRES32=r3, @ANYBLOB="2377f292252155b21c0012000c000100626f6e64000000000c0002000800010001"], 0x3c}}, 0x0) sendmsg$nl_route(r0, &(0x7f0000000240)={0x0, 0x0, &(0x7f0000000680)={&(0x7f0000000400)=@newlink={0x3c, 0x10, 0xffffff1f, 0x0, 0x253c3a, {}, [@IFLA_LINKINFO={0x14, 0x12, 0x0, 0x1, @bridge={{0xb}, {0x4}}}, @IFLA_MASTER={0x8, 0xa, r3}]}, 0x3c}}, 0x0) [ 2836.390473][ T3256] bond1303: entered promiscuous mode [ 2836.404889][ T3256] 8021q: adding VLAN 0 to HW filter on device bond1303 04:12:43 executing program 0: r0 = socket$inet6(0xa, 0x6, 0x0) bind$inet6(r0, &(0x7f0000000000)={0xa, 0x4e20}, 0x1c) listen(r0, 0x6) socket$inet6(0xa, 0x6, 0x0) r1 = socket$inet6(0xa, 0x80002, 0x88) bind$inet6(r1, &(0x7f0000000580)={0xa, 0x10010000004e20}, 0x1c) setsockopt$inet6_udp_int(r1, 0x11, 0xb, &(0x7f0000000080), 0x4) r2 = syz_init_net_socket$ax25(0x3, 0x2, 0x0) bind$ax25(r2, &(0x7f0000000000)={{0x3, @remote={0xcc, 0xcc, 0xcc, 0xcc, 0xcc, 0xcc, 0x1}, 0x1}, [@null, @netrom={0xbb, 0xbb, 0xbb, 0xbb, 0xbb, 0x0, 0x0}, @bcast, @rose={0xbb, 0xbb, 0xbb, 0x1, 0x0}, @bcast, @netrom={0xbb, 0xbb, 0xbb, 0xbb, 0xbb, 0x0, 0x0}, @null, @bcast]}, 0x48) 04:12:44 executing program 4: ioctl$BTRFS_IOC_SCRUB_PROGRESS(0xffffffffffffffff, 0xc400941d, &(0x7f0000000240)={0x0, 0x7, 0x3, 0x1}) r1 = socket$inet6(0xa, 0x2, 0x0) ioctl$sock_SIOCETHTOOL(r1, 0x89f0, &(0x7f0000000000)={'bridge0\x00', &(0x7f0000000200)=@ethtool_ringparam={0x10}}) r2 = socket$inet6(0xa, 0x80002, 0x88) bind$inet6(r2, &(0x7f0000000580)={0xa, 0x10010000004e20}, 0x1c) setsockopt$inet6_udp_int(r2, 0x11, 0xb, &(0x7f0000000080), 0x4) sendmmsg$inet6(r2, &(0x7f0000007600)=[{{&(0x7f0000000040)={0xa, 0x4e22, 0x7, @dev={0xfe, 0x80, '\x00', 0x1d}, 0x3}, 0x1c, &(0x7f0000000c80)=[{&(0x7f0000000080)="7404edfae53ede7d1a09859a6b3deeac57e43d596e3c4306791801aa8f33af74978ffd7335d4b3b3322941899ef8d99ac803391fc661e08d67d096e4cdaa4a67392a3b79bf47a9417b7587a9a1d260eb17554b01d84a38548b1015cddd01b5e16ae9c0e2587d4cf47afe4ad5ab01bcffc6ec361e85f1937b18d4a97da8e562ec85858d6b16ca6d146c4132b5d6cd1cbb867e866fca85b2716e98bf62b11225c17af3dc6ffa8e646dfec27b19025c69ae34499e23b267db917dac85556d0afd1fb6f2197caaf7d6b1916a6adb88c1470eb32c278a3e2ebff4f7", 0xd9}, {&(0x7f0000000a40)="3ede83855b5c4c5cb1e3f189b0fc516401361f677c3432c5a0a53cabce5d17ef6abbf132df751a14d38d4c86fe6ba04579030e2436a835da9bdb64a18fcfbbf7183316a6aa33d215a8c18fc3a09ea7ab4576794d8f315684b0af9ff8f144b196a9b1086b3b20119349abb79e0df941576e1376e04395fa98cb668367eb0d2dc180d15091a7d91ebfee8c11922f045772799ebdea6dbe1c4f5aca9647d2a03e", 0x9f}, {&(0x7f0000000180)="f484cd0060ba8f6139a6ea0eb3f6a338a6677b7298937e8283f582027f7080e6b4a612522c23c0bf2d2fe6a6ec8a56004122ce57731cbf3fe6c946cf1ae4", 0x3e}, {&(0x7f0000000b00)="6dd61a68d3c0fc641b5eb13d48c2a6c3d097cd5a44d96549275ec0f3a236c5315a14b665a8a3e3e5c1d34ea134e7e1546cf8537945e1906259bc856d5d6466b9ac37e5b96f1722237f2b00bddb9046d96400cbe8d52d550787cbab808c5068ea832ffbe112548e4bbf253908e5a45cc4d2d8b31664f5bc9f0527bbd3b3601f328979d61e6761132c10324de1a0676c97d8f8e3964eef59c99a777116c41d90eceae5308fdfd7aae49d7170a798fe6f7da775e4328cfc31938aaa30aee7a6291254796b3252240e7ec9711429dceb10d9bd880620271c6feea17780f280a0ed9fe092", 0xe2}, {&(0x7f0000000c00)="db27c4ff6a09ab2043e1d2cbfc5e53056ef813b840fb73c98b0be1ac0c8e3acb9a0c3460c890a4c6e5a227a85a0d38eec453787cef082ee62d259248562438700f467ab5689076ca6f5381b8d89a252f2e5555535992b1", 0x57}], 0x5, &(0x7f0000000d00)=[@tclass={{0x14, 0x29, 0x43, 0x2}}, @hoplimit={{0x14, 0x29, 0x34, 0x5}}, @tclass={{0x14, 0x29, 0x43, 0x4}}, @rthdr={{0x38, 0x29, 0x39, {0x1, 0x4, 0x0, 0x6, 0x0, [@mcast2, @remote]}}}], 0x80}}, {{&(0x7f00000001c0)={0xa, 0x4e24, 0x3, @empty, 0x7}, 0x1c, &(0x7f0000000ec0)=[{&(0x7f0000000d80)="8243bf28d847c328e11f2f250fc748e7454f01", 0x13}, {&(0x7f0000000dc0)="e98086afbeab9bbf0c9ebefa6d493c4b5c589e774d12f869db3bdfc70618b78fc72cb423931a0b2bba3b4c746d2d68c0cb151628da47f9da4e8cf5342952dcf45e1f76a285d0efe301c306700d23abd38ef37b14de999a69d2d30efc9ab8d7619317cde3758c4fb8be5f8faf0c97cb2cf1937279a51a1d41917f8a42bd986eed6848f54f6b28dd9496305406010682dc04da27f4a9bac7c84ffefa628aa3fed21ed48d0ebbeeb8b7f3923bfcb7e801f7b024fa930a05bee6ee5c853bf9461c0854882e7126d4f182b4b90e4431361b84989c3aea08", 0xd5}], 0x2, &(0x7f0000000f00)=[@rthdr={{0x88, 0x29, 0x39, {0x2, 0xe, 0x2, 0x1, 0x0, [@loopback, @private1, @initdev={0xfe, 0x88, '\x00', 0x1, 0x0}, @private2, @initdev={0xfe, 0x88, '\x00', 0x0, 0x0}, @mcast2, @local]}}}], 0x88}}, {{&(0x7f0000000fc0)={0xa, 0x4e21, 0x20, @ipv4={'\x00', '\xff\xff', @initdev={0xac, 0x1e, 0x1, 0x0}}, 0xff}, 0x1c, &(0x7f00000031c0)=[{&(0x7f0000001000)="5fb906fb71ad76120c1c90e7eedbf0acef97ed911df0fd053d1391a9f079a43751f62dd438bdeec3123f26cc73f874e789c29b6d866eb07df5f18002b214056c2662a2fd0020b430ca9b6f3ae85ed7a61a73250935736c7f14167a9d6fba9d7b01a315fac51b2be36decdd98affbc9e1e3383e9f71e27d239fbfbeb41fc5a84694dfb792a606630845d883f0f6f15e0a2cde0115c6376d0fa81435353aff60a08b8ee81ad7d4d5fc0027c04594f7414d6e1de1b7fd1b8250273d63f2b42295226d6871cb94f7d0af49105038cb257a442e7f59792ee843b5a3cbcce7f5ae361d3e36550c5e3872e090524f9223949de4f2a696d14a9b5056095af7c414b98cc8b429c67cb4e1f991f29e6f505f3aa6956d733b49ba9a85bdd634feaf1613ebb8f4d7dd3d939510d2eebe9a40679b156343efb4284281f93a206bde9d1986f0dec3c5a263749f25404421a6a495cd1f302662ae176a2e8974fcdfb197153058b2f5a6695932875060dd348d1d233fa9807dc5a98507e8c6fe726167f276499633b4b7463b392abadfb7b68da12aac146004f46e679e3ed18ffad84afc95100fa09bab6944d25ea55a78f10a85a063ddf537bce2ef1bdd57b101db4eb5ee4fbf0a9d6c184342a119546efbcb71ada2db4c1500a46e837cf8c7606d0be343d0e460338a05957274054c092cfa3e0eca9fb8d2f845993d9454f7782c8c3a9b823e674d2db4a924b4946827b3b8f521c517aebfc9542dd799bbb6607290faa4743acf98934892db1bf744e7efcad821c07edc0f9f18de11036b950d92eb58d4dfb7c845730e9e236ea6c6cc56287de14f6e169da6874dd5fb442c37f6d18ac40af97fe92d5d20c4e04095918cc69f7306dc474ae33886492f8c8e746edd7d96c776fa76aa9ae243f6acd93d982ddbeaeb301ff64f25cda9eaa7afd138b840b52f183bc3a7f16f88526e9b2b6913335b6cb72913c1b59eb83eddba71f28e793f3ededf994cb6a80b47c1c59edf73feb0b831a342dc0c8a26c0c4b8132db0eb3656e9702bcd924b74e6853ea0e35d01d5a1da103f359fcbcce084ad603f61343d6f6a2b47bb131c6e1391551e38712132ddd0ac842ee4b2c4a605bba68aa56c18b6e785163194e924a84804c7b1ab5b201fb9684ff20c767a1fd19b10b3d6bf8658ddbc1364ac761784e3fce85d250fed72c4aa6fdb45593973c4b020c31b3c366d17223f64233b2ecf8c37cba64631c56aa066f2257718bb499e17ca903bddb35ebed11e2720cb1b5a027b4c0eae4519ab2bbaea2bd3dd643d30cd441b364762dbcaf3e598b2996ccf2aecd177dc4e0dd5603f9d71be5b9d11c76a7adc012963b31fadd07c1ae825099ebbeeff10b1b9e9bf5ba542d2be91d145332606d1ea4847ebe55a891746b4e09f7b6ea1cedb974bb6ed649a35318945767f17eae0f212e44176bca8797129723bfc9e6ac402981b65a1e72aab2efac6c592306dee6070f2e76d199c722753ddfb74b772fc530ff403fd85bb9f17651d70a871d395ee2b047fecf254877feb7f6bdac80af1e5892510b0b559fee90b6c7b5b287498c0922bd911d723856a0ecedceb63d8b93c014e1bd7eb9491fad5e986eda95ca1eba6ff3a39f93ab75334b38e850694685d9d323bcd70370532ce87f00f0d38678325d174712ef9c9f64b0a03fc236c991797d0f167996ce58895b1fc62e2a302a15907544a6dfa92833cdea659f6417c742f40fa699008036318b1971329cc7ca67930e33606402515b53eaa83cc54e2bcd67867e1357f56c1315b9070b327e1840bb08679955509bd83f7e1a2e92b1e3b5fc90e5537714732af072dadef2de4034077fbfeed9ed3b676bc8b6372a46c9f18a1962949262698e718950ea1685d33a9f8642eca180614feeb5b108bcfd46a8f0615be219120545f702dcb192a3f2f54f450586bc8766dfd9dc19a949a95cb9d62d85e40bcd64911138a83760845d2f3e9d33209807db5a21dee4d9e54b8c20a5f67892b06b13272054f9e39dd54efb57bee4e321cae1db7b283d86d324755f6e97be34d7352036910ba24c03fa9cf84aff230cbbca8145721e8c70929d4c70cf96d4670e5124dcddbf571e9311f30d2ee111990b1d41fbd1a2ce422d8129651a4eace61da1d8952e60cef9f029b4497a06c4c79e819c3905c4761e89076ee5a6bf023eb54b1a7ce4b615c7eea5482ee7e77ded251596d3c4ec21cacffdf9eed00d3a141bf5d6488bbefb269438eb4e60487f1b2616cdb79869304af7043f47e60e6c8c900498f59756a72fb6bc1e74fd07cd1679b0349c69b6ccc36dd2cd74bba95388216f77133c22ff20783fb06bd84cff2fbb3e32a260c8741f5725c54c6f7e659eda998e3d312c55f203343e4c68a4c5adeb2859db2dddf17a82bc2de14c52ba4a83106ba89330dac4c05f3d981d0fa5307d2de60d146fb1dfc7a7bab1d9f86dec55db77dfedd0f9224ab157ac746326781283847015c61fc06528ebd620ebc815196e4a3a6ad293a1c895c3767c61518c72bd0699355e32f5c38193215880806b85f71ae748dbe2a04ea31cff5aeb65e1e7dd8ee8575a5579dfb5f84c2e9e707e6d5b4721ed461ece928a658ced136dae166d512a3323ebc064bff652c4de6bd7108239d3709f7472c91f54ea504d97c1ed2bdcd873b17539d19a620b15d26b343afceb251f4485cf91be16d4e47b883d755cc6d2dbf25a196742214a854ad1918214bca42bc1b28a952a5054c2b26b172f4bd96ce2e22512ef59ed2eadb496debfe06e50ddfa9f941e85b17d2651d976c1e9f03574bcea247e2d0903c5a2e3b0fdd1bef735bc97126424d6f94f1d18f5e4313599bec2bade7dc6d1fda2fd05d6d325aaa4c50fa9ad95a414338515cee9e00b39c552a86966d64df5320e2a9638c1a2e3a794d79a46e05079ee108b9d4c26174c9d05d55c854d9b23dc963f5264331cba5c97be78e89bd2c3172d40277b9bbf4728c4bb884c5fbbef8fdb89f97b4ffef27f9af22ddbea2909706989960a58accb4f97f6be0d7a27f4fba9628c5d0d07a8c19e897963ef29b959f283cf8049d1c0a011310f45cd8b9cf6d22e8f94334ba54281d718fe5252808182f45607b84fdc551d6c0ca98ec47246533f4efb1bb83170b39a11dbdf0dceda366adb009ef48cd807aabdfd76ebfa7d1689ea0ad638fd2c5873dd62542e98e150c4c29eef20842fbfb9dbfe4829614667a4bbf38f727e66c614c4ce12ac034a92689ae10f3d00c872b67a3c69c39560964f50420a4ba3dcf62d6d39c603e60231b36b0356a7a0080756b761feab0253e68e67f317adf72a7dfef797469547a84fccb764e7b78cd012a77ede0177a8e402e56540e9e5e9b20b686d50d213226be12d067c28469827ddcdf7e15d8d4a6200e9e52bf17384041b8aa70ca097275fe70f44643a8beff01be4035929d765d7e863023cdb90f3bde63fff91d023187a0a3fcc96018475d8767d978dbe58a5555feecb6699e2234cf481ded5b8d3bec293faf12c16f890a52238d3e3664118d4bc919f05fd7a299c2c9d6a5fab9dc0eee8aec75f57ad0bc91eb29d34a485691e2c225654b3abd4afe162346118a87e33f8c2cbdb3f68df5148eb27b427ffdab10f29475ad17d8ddec3025a98b22b30bbc27fb6123bfa711889e051db56cf07aba5256458e5059ee5b4ae791b96c5a45170ff845c244d37a5fd28dd756e1892fa9545a38a9a925c41dfcc55147215bb2e09c1df454c23ab3c5a64cf339ed75530967765c81f0bee2d0f7804e841d76b6aed9e361b433e0089e9681fe497d473cd9f39737770f4e887ae84cc9f2becfe93180fee9dec22b9b8697e7c5f5dcfa2e3f3f3bb0c94f40400caa45971a6c7c90eeee5e924035521bb33a28f590dbddf58251047fe33bfba2f3e9b9cb31d2f9b0b688c21e8e7be2a3add2bd337eb4884931d78430aaafcff9da037b0ef972506fc12723a1ea49ff0e20bfe1bc3359afed16337bdd79370d992fb5ac6f6c1f6da011bda39d64f6e559589810fc184bc985745ac7a9c5b3926d98c6742b609efb0586077c4537969255eaf90acc134f80c1c062cf313de92a40dbac1f6e011716335632036a2c79b6adf0859f8ced80776517d6a13e990eff0a169c79e953784e986f3f18abc7fcf13f583b6e9f383543b41f955815dc93a687e88d10f5689ef7cc38a469e5b2f0dc4fb3f2d62de06c4d0f1787d50d873c5d4c510d6cf5e62702b84110a7fd3432d1bd1cec285b6fef129c9f8c411ea9797694d45dabf3ff0251db6577984728a23bfd60331f4b69a7307737e6ab90d79db6c107adccdb35af3b28ed2da1e449f10d5760b37b56f484c7a56eadde8c3f0c7dce573eac3d4a6f46d82dc6463d4c02425fbcbcf4e3d47dadb96a2bae03750690a7232f3d0bb5f45ab80f61d0bd2495702359c19abe73b730ddc8ac6cc87ff4e0916b0d401c579b33bf4eec208227cadf3fcf8e2761eefec22625eaa8b1820557a2de152f2baea18c348c1337c5fd402088a81bba2f5544f208ebc684a6fc20eb4c8a40cc270c937a5ac417db1b59b00e0659756466d7f1cda12092dd138fd8b5fa4c79431d4f3c1f2dac347c29cfc2ef23f39e01a6e35c7b0c20b76d29eef315a2b49d3aac78e5a6c07682eef5528982142a555cf571463f2acefcae1056762ab6e6adfd1109399062f836d08a5203393d58ef72c1b730dd7b6728dd6a3389321a01a54a96c27115b594cd401fc1922831097f333d182764c92a3146297948a53549758825b5cd2e63412c493c7f57aa3d79e062bcbf3d3e8b8f94917b6df48f4a27d4d09330c427f105ec4d56117f0301504cca747506894af3caca99b934640ef57bfe93805b9fd4e4c9880da8675eaf3bee1247777827018e54bdfd788faeb84e1c471ace504a7910670bbc3c8e75d546940ce0768c543bbd9348180484f0e9b279e3667ded2423d324fd098d4585a57feb05319ec7a137034e721538cecc06bf6c87541f31f4ea4b38623425df336ddd233d72e1bc9ac556cf97ca88f9116aa23b4bc493691277ce38e2600033c6dea8bcdade7e5648a8eacd4358c6b04f8b4bc5e459e9b9e91e8c0f26129c05474f062a00ba85d50db442d6ecea983726fc8fb47332406a03ca40c2796eccb545056968c1f2cde6446549ecbc21da3618b1eb46bc986bfb652bb93178db6d7b4beffda22ac417f86a959fe3788c628e4d5a3ffd8a14e3bbdeb6103907179cfb9420cbea5fd2112e1409ff92ccfd13380e82df79fac73bd23137635f12e293e8b3077ae9dabbd2fe2e13257c2f8ab11d20e1e386b0637caa4ca227dd1670e3e4c4bf78dadad16540144a0e7cd67f0ee5e864bbaf17b4d22bfb2e4a3b13705658577422799cdd991239e0b0d0de5e55112cbeadaa4a4a065c165edc31eaf17a96ff6be007a123b5600a3309ac4c7138b52d1257ecda2a1a9d84f5dd66f7121ab6d342294c6cc7b564c609d5f9381406b2a7fcec40d779a02b95843731cb3e741b4c5152eba9e4797f95d2c7eed2d0e5f66b2bb36679ca2b5e483d9c19760f24137471a05d0773c84876cd54969bab649f954d954838ed352cadfe63df02391f8bc057da723aa33c478508e85663815f4131dd965410159f3f91563549cf4b55bd7688624135b905795dc0cbf5da7fd62232dadb4b7c0b7ff9ef33ff73d3400ab41e702eadfa266905f4406cf277d0a447f3b93b766e5b2d0414f81396897ee66d942353cef68f2df92d8a8208388b000411a69169946a2ee38b4a47bc8c1b", 0x1000}, {&(0x7f0000002000)="7252ba8bc1d75a97d4f12549ef766b80b2e4c36d1f11d984de5807b5d9b7be5badfe5d1d79f1bcd48743507c2fe1e6bc93afb1a18bf8aa4f1141382e5df43a20851f2ab3b8353af0ec54cffc54378586ee8c9eaf3d4d8cf6640aa94fd203acd514e78e44f7217406320ae0723dbf749ebf3685b7f12db16702dc7aa938b1abef9c789602bec343e1f0d54fca9b7146884f0c4da57a0d4011c5cb61c52f36a5469a455746ddfbe4a5db43f26ba9826d6f61ce1ed9a0b46ba516f243d048fa2211307d06109403e98f8e16770036cf580f5d6db8c2dde148bc0579740913cc3868bea39a497f8942", 0xe7}, {&(0x7f0000002100)="e0949db7f4ef749c4d9532615e3090e995eacb74b40b8c8f6b2566685cf6b1c677473e380dabcaf9031d076b23fcbfb108952dd43a2af98cbf95d25f2524925a2f86c4be9f22175c1612f0a5ecea1fc256c0d287395bde0c4b8f3a9d328a00e12096836f20f8ce2f6974ba144c98cd1d3fc176eb37f066f52ec2659d57c00010571b878831bf2d01c2b85f", 0x8b}, {&(0x7f00000021c0)="e39901f38cf493b7c80ba34d6d55f3ba3dfc148fb0eb59a7dcdd4f0fa3ea59c81921a08e3eb6be97db3bb6321ef9a212d8b49d32aa26a2d0fa01bf22723c432241f5ba018fbcf6de10bfd692a1e92b38f9ad8def34e79070a778b6f446c1b1e8f26e19b11fc190afe9d0a75bff256751a45bddf56954fd16fe13b3afe9be1b98435e5a2ce95a2d7c52b5d1069bacbf1ffb85c414c0eda68fe7d70c77a4605890f52499b751f3567ed978f329870d606369b129d83a74c7f4656ffa1ac2d6741e3728e3ad6d3253c375e552a53fe9d2776d27fbe2c1d698dfec48a1b6bd1e297cd2bc6e3b844d56934a88ba2a4b14ac208af2672f37e26be9d429624dd33f22361807c631d346e4debd8b1e3baa1b6a2dac341531d473b126667aad2f4617a504fc752b7283ea0bdec706910745ddda451ce81df4c6c33cac2479ebf380aaf0061d8203703656fbb6b88845b743e642752c2267f7c1045bb5159676c7bf53502e157457c6a0e0d62ce83d7b74b789a2a2bd4238b4236549dfbb377522683f362c1c06200d2c0e9b985b565a91aacd64f6db36c904ae56ca8f4b718e71bb0651d5113f560c0fe3ffd70ceee7a07b8df0e7a76d24cf87a902b4eb913ad10bd0c965dddeba3ca473dc8698ffb879a405e489a917bd80656122c75fc27b85079e1a63600212c75675e2246e586d081e1a14da434903ad7392f9218c4ea6c2966f116bd292889403720c647693bf35adf963b490478e8af07268fb3ad49e0ee096058af003f5e7b80dce9032b209e7cd79a2a5e61c0caee2a4e2d90d307b4b1c9176a8625ba994dffdb4a3a1523c5091d86af016749e8bf9671629b77a40a44de950260e2e15f3ec54bb23c1b8e4e31db20e8d2b2a0307b1f988b20a374d1e5e1052bbc35dd0766be1fd241b7269931b0b510fe0ba38cb33cedf95991b78cccdedefa5eb953759c1a435db4e2104305ae73e4b839f2e8af4a053a74fc3e22d612c75e355c01e44911758ec8493ed3af672216dc1bbd42490a189da994baea138bf9d27e70444a6db92de757d31c91631c6bd15e92417d3b5fd04cafc7bcf126877aa0affdd0154a0ca10d39980782ecff7d37b1cfb6ac209ca6771d470ae820b21189bdd74f5fbbd4336dde08d7adef8b20d73b2ebe5b6be51edbbd8d368c705f6f67461555790fc38372d61c85bc65fc9b2019bca8ae1764be788cc04a0108fb30e83f8dca571d1dd45d3b4d6b5eac0b716fc5b245fe9669d7a229e6ee72bfbc2b70a02569d87f1bcd74fb99b3f93fc39566ea4e152963aa88e219e0665e93d0b116e92e32a6e222f85971edfd3bcb310631cc7528df95d1c5a4b71c5e9c384ab1554c0db65cf6bcbcd3be5ca379ab704cf33258d7542aadaf71423f9b655792d9f3e73e2c800322c39e530f4dd596f9cb6795db108f2fda6222ae60ca3851c839bce543f3127b6b5d411b8e41b4b9f585e517824ad89fda37ffc65d45284fe985cb159d7370f36e551ff718dea5387f0abc2b39af55b5ddf0b96a05b12817c12166e75c4b33c2ca1765b1346dc87e21ad13abd6402cb22f65ac47b078af23571348494235e8e5483b05e1c8c3a4e09f3679ecb866795e439d87fa852e59581c5410600500b18f7352dacd4a756b4a49d689e260f047f2bd495e58034e65b2473e2b8ca4e763294ee4b2f74db56f6c2a8fc28e3316e23d538583d125e8420a50337724bdd46794d19401c0356a906769238417ec4e217c9f37dd77232e996aa0ac09390f604e023a7ff12868d8dd161ca3d80f8acadfccf11c2d301a9c0ab465f42a10e950ac9973381714ceb3dfe6c5acf1f6d83ca25c70a5c24d49b127d78153cb31297683ea0568f038f520e49f720061d6c2dfbc8b0bfc0eb4b202f2213251792c6f741d1ee3c5489c76a7143473928ff938c97683a316c72d6e35844fe8c25d017bfb8a29532c9f444f43559e909ec507fffe66759bbf584cecb0e4570228e4513c08124d91a17b90c2230de488f38f3f2e606ca69e4f006803d6e5cab9b11c9f37e9f01b86301f7addcfa4624d6747d82552fed7d7c77183415ebe6a9d02bf40588a240001b0c0d13c74944d8bc009296abc6ad16c73bde8aeb81a5c92e9c3b971c4cd4d4a1423772b1c5e5d000796474e38c3f89b1c99e19ec76cbdd7339d064221ad02c54a9e0b3d46573bbaa7c0af49c8e68e6f17ef33735aec1a982af8ffad35235f0944541ef0c35c9824748614187d91219986bb2fd45a88877965e7aff352d2ef4eaa5f1df28799ebfca9e7544cfe93e7810663b956aef8ec6e3087d3fc4b44e1d0f422456f22ff89d5dde9ef5c448b646142f2fc8977834f245ba4528b70e1c4daabb6612b91f96361be441e0081cdf04e5b0b7151ea3acf05efeaeb8a523326be040044c5cc450777cd5cf2ef549c2f712277dff37ce30b7595b09a1e1e384702dda93a46e63f4da116ff67b6d1e8779d99ab460f95dcb0a1ac300e2a1c3bee243741086e1a0254231ccc34a9b9b77a86a86c36b590dea58b2b943b2eb15212dd4e13d796fe23973da9b1c01d346f0098591bb3f37f40b77f31cc7e73b15db81270d8249e03b2554e2c32e54ddeb1b286e6cb874d8a3e0d442dad9574668fdba19abf13f8b170952b4175626ad31fc40b7bea69a9fd4c0fe016bad0a4e988172fa81b226d3695ce89602760009e97b587a3317bf6870a70ed910f1355e03e80cd518be82d7e06f54f8632772247fbfdc0e95068741dc32bb4daeaa3821da0a1b586f78228762eae6195e913b500e5d63d75217e8a5a1c3dd9d4dec8ee798def5c0066c713591b83aa3aadb0887493c4066b2a95fb3b4d316685682b7fc92f8a19093a7d790d7a65b430eb6e7e96b2da3d349f54df43d4cc054fbc041ea9c54dbaa61349f669f32c1cd2f3c98e2d9939da4ed165264568c099fcea1a5a413470b1592a582887bcdc87785b273aee82ca3bd05986ecb860554f01e5a4aa0b1bfbc6ba5df6a3abe73e14da64fbde485f1af83b7c4ddb656f57c8c857b79417d5d51e997717a00138f6a7a0b604999b77f73f0edb9e7d59413de9825fd49533eda109ebc2324f432f2ee563b52780f701c81778816042a5f5706b273635ca05274a981ff742f01df9bfc59c3f2792260521a819404e3111529cdee969e55ef9e2c6ca6fca90a179d3f20bc3b42b1d69ddef324f0cf04ea6983fdf4e8d29217a496e6fbcdc4627e49eaeefaa9368bd3dc96de5248a8c57a3dc80459f1c65b0a1bd70de3d896521245b5c9952429b68aac94660327616b6d97568c1dd6d11e7c664ac077389185b659309e79954a0fc42b8c21a6caf29fe05486e78b786854376cca6f18cd8b790ef0c83621b469f6e6175a29432e1138782417269a90593ad205b3c691c13134c0f0cacf18a9874ebf3286e2e3a4d6a86f0c8b838642f39b5e7538a68072379a23fbe8f5a2c679da346f67487d5630fcc7d40634c8309299e00c2b352d69c8a11c34b20c1e0999108e9d922b68289a1403c1de9d6806cd4112a3d24c9c34bad74c6efebfd06497a73f6bf6a6a0bebb127abeab1931ad70a0b4bfa9b40be372a7e45abe7b611b5bf712d681a57c0254053fb104389ca5e25865ddf7ef4dca6a2f286524e263ea2c166a687faf60fbc13bf4d623521682b3104e69a1556d6f9996fc25d1fa36f8df5967b8a7b0732097453a897099f5fb4d47a1bccd1db44da1837686afc9f5415f51c0640f9eb90d1b80b75bb11d31e1f553933af3048b77ce77d7abe6e556d7293b1365a357b3a714968b4cf9699eba4f2b62073d9c7588275f243dd4020839b6a1838a19f5ac19e00220294fd914fd10198cf42cd3ee5cc16aba3bb554db17ed771ed2bbaeedf1477ff712b13f929982db8eecee7cc3d9a29b28932c59a9c94f67f89116c33a551c906c5f4d96e5ceaeeae6197b9463c3e2b0456dad025484267f11d041213c21e89d073c6170efc478fb49878003a4f03156e5daae5dfb8f1f04b8b7f8788351ae7c445133137801401cb496d426ab0a05d9d933f6df8d593389ad2d5cc6453d7384007f23c6e3098af6fefe9fa27d3783168e24776b432b872023379b20665c131cc13f8b08bbbc9d6a6d9772f37750602ec66c2447415b97800027a1f62cece33c0f2e3703bc27f5a6a25e261b90452ccaf0bceb73ffe9c91356ee28e466682708b4aa245f7b708ded0e4efa31c33793eff3e960708d7df3f731e61e6f02ff94398fd26250e43452c07dd1984c701e26086fc1e2cae819e8420d14cfd120998212a7b316b39b9c64aedc3138973571ef408b5555ef5c5ea8e35a4d62f8eb93c0aa854ca17533e71d078c3d066e8b05c205d1b9497f7988d40699dcb2e82b3f053517c4781764c9d031a88b0494ad3e29faea4310037710a98087ab3d75e4671d8098854d278dfefae74957e168a9b7e458debefb85e23e009ccc7fa31ca4d23dc2c7925d14112584edfe1a27e8ead18c3b81e621eff3e8d307d2ec47d3ee388b5404e9be3bc9f46be6e9a27c0f65c582df62e1286dfb9b805d5f490e993c278a66536155e06f4c306ad0240cdc3108ec9e8e63741aa1dbc24e521dedb43edfb84b4ba506915183d1873124eba9e7d0776ce476b8b8280e6f662520294cf484b5b155692a818437444a494ab0445a48f9a74f7bb61fec523de839f527e4b116741069da2e6aec6d7835d786205055acfe7be63ba07d8d215f3cd948ebf360d384ba1783f8db18fc6362f3875876c1ceda2dc8441d21e0b430c9c3bf4e3eeac9ce965a37d9f6d126fa8f9a0340a039c61762ed4588a2abb47e0cb88b791e8a9154eb95402eea495a3a49234c83480dd628239d76f058a3302a3ca4f693aa3a994dd18cb5f99af03404889ea713f2def3116588fac94c94fd7e19134668c24cd949835f85a70c7686917f5cfeec8bd1bf3e198bcbcc980bf66d53a26b4472394424adde196c2fd73e452fe6c51164049f0d7eb76fec48ede903d1aed21b1796ca0c869a1b7974e01db412c44c3cb991214ea56d40dc296a4e4b9b39712e17ad8250ad5274db318f7a3f4c373427ea1c440668f9c94bbce7e064aa1cf5f6224fe9e58e4d488c85b1e0e1dc0d72945b1286dbf37bfa4de1f4013e7177c259790f44e081e0def7b9bd45620dfafc5783b0da095aebaf2736547a0955dcfcfb2b34f9719daf8fbbbb89c9bb0a49bef904234d9626a5debfa8731e61d326d635af87120b40e7158d54227462d3c7a88b40e9639338149db971aa99b48c33dc1cf759fb74a9ead522574a8fe5c870f5af1179313228415b0b894a76235efe001b915e0f8206ace9dfdebd90e36b6dc9542f43c680a694ff086017aa46171c2d98ff9c90f3116876aab63d58cfec732be181dd19aa6e68c3897510617b135cd3e578e11c88a106c3204b769ab91cbf454e601f9921de11c68e7c0c52c8916924e866c9d48f510dff5b9d3fb96796cf079fddf3c755c9040601b493214706fd37bf89e0c737305da67bd219384b87bb2807b4c109eb30fc96fde40bfb2e9c0321e0089a24ea72351a1a13a4cdb0e1dbfd4c50d7327458d94f0c49dce905955235b259d7a808aeaa9eb9d8a4d781545b7f4cf77d6140d48404376e08533d269316e7760ab7faed70ed938a97a5294e8f737c56be715e61ee7378ce0bc99b9e53e9143aa3ce1dc768704655e485c047ab8248286886321f7c9d41830ee8cd886201748ec891970d5b8d99f9941c32e5be2f1f77f3aef57c0c32c33fb3bc71e333dced2e5a269c09f40fcffffba2e0", 0x1000}], 0x4}}, {{&(0x7f0000003200)={0xa, 0x4e24, 0x1, @local, 0x200}, 0x1c, &(0x7f0000005440)=[{&(0x7f0000003240)="7df665bb5181ae567a88ed05d75347c1fa484886428a1668ea1f35a7ff4df0e57e7ae2d148340c5239f050aa2e6da57039d0e62164d484f8ea3fb53712b63da4d8c88d5e4a4865d02b0cfd4da18de33c80d601f30877ccd6d1209c3abdc1c2a065d4bd29bdc51208e791c4dfb39c84a8f87fadd9e248f917b5f85c3b5ee321b27cbdf5b46b603822ea9997501311e59d9724e805a1fa064a8b9fb0502f0f0df0aa2e0197c7d3e45c4cee38883af5c2a52818b9744559e71ee031d03a73c71a350af164f871bb9531ee415df25ccaacab14fe439349a865c1db2679736757e0522e890ece9884e2d9e2ee675f8d4fdd65ff991a761f00dfd7", 0xf8}, {&(0x7f0000003340)="c7ba26e2fd4cb119bc254f5c25dc33fc4d25eadbff01de0e0f02699edbef7a257df9e9bbbe6cb085a0562ffe414a6bfc4e516eb5b46ef48a434d5c90df38b1d26d32cf746891854680da8194ca464b5f4e58f1c8043254d4e026dcb18ab991d6d1ddf39ae3971d16bb5a9f47d376ae13639087c6c8a21968df71c4c24ac771718c17e3f92a23ca98e6256904211f69ba66278ee0b7b060c8f0e207645903f334bb3a8647ffb7abc8b6b582d50497dc76e3c587e85dfcf30beb9c5c261a", 0xbd}, {&(0x7f0000003400)="e190e89f605f00bd91f69ab4666bd7ae317b474c1e6c78fab735984eab18644c15ec4a3a993e9f96e9bd9837d93d32e5ead1c84653cafd94d76a36d7ccee1d91367c55b74dab557da03766ab6120564244fadaa27f3fcfc00189aeb15d9a6afc8de532466999fb732a42e728cdfb676ed47375f5ee81ac6342123963db2f79c622ea5b63c5836ef02c24246fd93ef0eaf2a3463c48c4bdc7296f36099882cef0a39558e946fa98102661d5332c13a696ad2524b4942bdfa45946710831f3ab5613b1806c454384ebb0137d17fc3b181b9ff9bd5433fa09ebfdbba962ef61d4a1cd1c6432b8803ebf133847d88f660041d5491ec2c7d6b99fd12e8259a724331633a587284d31ba1cc9b0af650eabd386797d149097e00dda9fd9efb2695fc7120c159cfac8dc612c66186dbb3944afb55f27d1e8d2be27a538e3c3e2c278ee8215fadfb9f082cc945cab4f9324bb27a5cc2b3ffae06564c9c2f91e848c93f190b60b140fad71a3c732bd1edfa09ff85c585791dd5a68dd8fbcd3a8a804f474dad52c53c724778f7bfcc6f41345c997d95372c7c0635e9ee40f2d03fc9cb3fbb7a418854488f9de79ed78cb5ce6cc7e1c158e7157eb2409a8129000b022a738248f96e1e22145b180fa8844b2539deff9a7826735db39cec8380da35d50ae0d854e6f6e26e5b8418ac67e4f74a93ac122fabee469650d541ee6efead191203cd21ec9e30159e5ec75f9707555cf563897a7ba6b50cb679a998e8e73f92c7f1c609d5d763fb46279f4ff09dd11d1d3192e399bf554ab29a5520f6e7dc7be2d796bf0e66ed1a07d86de5bbbb9c77a9eefd2479ce23cb5e38eb731b8d169a3cb810835b4cba05ae023f6bf5dae84ef6342220cfe57337c266b5afdca0bc9f705eb0d48a2922b1cfafec64c540b81e636d95fab9e72a791210c67e5339ec101635bcd56a929146bb8db3e1b76d9651df2f63fb27fbb048358badf039b2a28f650c90c0147d4daf0e79734eafab8cdab89150c97058ccf64cda9f39234eb20e2cc6312e303e52cb8492afaec5d7c7c9c4098ff2895abb6eaa3806af9fc88e8f6772a64b852a82a64fd1d7b80bb2217e438b7eaf8437e782d398946bb6a78a7487bac4b482727b19450524bad0ca57034449abc5a5bb2aa39be4e76afa96ccc73a574e64a9ecdadd4eae8b88b56460c2ebf6e4fb40ea835a8591d580cdd1560ebaeb52378cfc52f6e70005767e11de0147751fb854a45628067bbc3cfae0f18e3b7764e38a8f92010dd52d2726b85440ea7ba4b341a8a65e9449f37a90f37cefad5b9ee598b57632a5db6681234c7c37c290827f895a3097ad0819e3739fa65f80e83ac4bc78f68752620ec9c227bcbc7efad226492252f42ce76540792a3bdc22064eb39ebccb046f448a6ba874731e979a5b784abc3488b572e3a28ef5a193097f83d18aed933b0c81d6de3c4596eb5759440f32aaa934f6da62f0b806c45bf1d5c4781d9269fbedfb983cfff1dcb44387454ceb05d02eb62188b585169f51e42cae92bc88197637b4914f5fe1e150734c86ea42f0732ba2909cfa77628d82d6b928d189810021a95ce16c98132e4b624455cdde4209196080a8fe5dbda4a5f31d7ef58fc89fac5f95bbbccadc0e0ebfde5b0ea2dfca449b5306240ef3343718812aa31a3ec2ae37022faf263d7b26de9a8008b3bc70d47b814e821e49d6b8efd313edcc435feea9dc640fbe5beeac93a3190aaa2fa0b8ee1801eb6a8eb73d0b2fcf6c5f7be41294194c2a9131489f39b7f1089ab925ed2535b2dda584849954546fb60aebf2f0067954ab00155770c26d7abbcd3b25c8c6aa820090eed50bf7e819587c7049b7ae7a1902ac44e18cfd42af34c0ea98a6d67d34cb7bb1f0098a74a5857b3dc154c155d04e8d4cd0536b741f37fd67f6b3ae3e643117de7148d329981c14d5a6c30d7a4898aca8e0d10e5568ad096113d3ca4db898141297b6063708e840595b4e1e954b79b8b4b0f8ffaed314a0ba7c3b913700a65397a338660e4304eb7dad08f326dbbb987fc6203e0d7c7b9dad8948400c82ed707c3c360220c20d0516b367e48107cc93ec4b2e2dff9b7892226b665a66096ef0afea2a13f896938a3c30be8e1b3f3893bedfa1a8bfe31114f5ca785c9783ed2e1e10e5fac8ad89be10fbcbb89bb6e62a11bc5729d6c3f9104b83aff8205cd84ab674d8690d223724b3b2883ad57454284d9b53f6ca0e628adf64669f13ba923cd81103833bf63de19bef07eb50d8dd4feec09793d5fce031bfb686c7a5085f442edfe9d2ef459a1580a2f345751c9086aeb9c8a22728cab6ce0aad9d607af0af70f4d2b6c2cd764f636cf1cf527aa899f0d6db4d4b83e3ee1078017fa74ac080580695159e418facba20a3e02286509834f81674ecf896056ea94fa98021531e1602acd956df05fc9ddf329531792b00aec7ae7a67fea04f0e3023f88d7f557a38924e3b9ce52b3121cecafc0612d43afebe99f87ef739634480dc01843f4e3bd7f5e822d362fd4fd7a3f18a7c46aa1a9934ef807ad161463ef5675d071dc52ff0caac2a3dc4903c702a2d8c6c300b342953088b8207dc66d56c90903cb0eed2ef8a9673b6365dde6ee2088fdb4e5bc385775acb843490f71779143dae3e7e61d48824ec039bf7cc34afccf02c4008749bafa2475dab12c54b5979adc785ad7f5baab1b0cb7fa53b0075b917b673174580fdf22556079e50d81a336f042de65a99fa614fe2c0ff11bf7539674fe5d81d60bb720934d63ef7937dded244539e2f098ddb301d090d6bbc21a6c9e218de0bfef700f23203d867007199704dd99d3d9864cde573503ef99a5c4eb933f7381a3b3181fd0eb195d5ad6d5ff1fcbc767bde3714983b73e2b7105d7402ed4fc7ce01f833f59f4ef09ad91ff3fe1d69b72656df47104287d677f996d9f854c3a1b17b1ef89bc5a656b2a7d6749395cb20bb1e4376f1221582189783faabdf75549da281235656d74b5398e74e9003b87c69d47b6223791df2b1e2d321aff48468902e2ba1fbbffcbc334fd497ed4cf48abee97418af7994cfc687c3844f122bbdfb1aaf918bbf2e0a74af7714dcd248bfef5684812e2b93e92a4fd0059447e16dd52b319e373cc302bed8464cee7faf29f6fd5a8e615edefbbf9a86c353eee4ea4ce8d22ff258d7bfcb7659cf0734e2b84f00d1add3cfb8434bd40230fbfaf6738e1463a20969573e5f71abcbe603f4878010d047103e9e6e4006414d82785d44ab63f4bc0190c18951505ff70424406dedc0aed19ac4953b8e6a1ba53577386a65edf549af0ee2c6e4a81d8f5767aa4dea43c95a8570e0e7887162d28ab669f62f69a83be5f86b88d6a963858eeb2db12c939ab3f6b9541713079a61cad6bdd3059ec3fbecf4da5421a75a2038bdc9b4596aca55efc040292c6d7d51041d5828a411f20f9aa755fb95331863c0fff0fe9d75b12357b21d0c7966b7f6e01b2e0c3556a1501562bc8af093641da3d2b643425d401d91d3349f375e9536692e7813c5a69f91ec0c61a4371904349ae99c6435f79a82fb720e12b0412f71758a5c28494cf2f703eb6e092b5607714519d1411d8875b394da6878fdd4bb130bdbdc11f83b92d98bf66d95f1cdd52391e6167c5e293f34c90a9e2d35830729e91cf24c5cd4d1a1039e9d0c4bf0038bf31f046602daabeaccff4dec6130f6d6b35e9fb4791bc5b6594706b694e5d154df372cd5d4567e407769db1299342d6672dea2ebb324c7805335a9fee0330b987f249dd429d4d5d43e64db525db8815274fc6f7b3f7853561b9616f2d102d210f6cf4e1346f788f321a0296e93193dc43346449ea12ea9078a3f0d709550a5480aa88787e0a0c8d4c923557e7dd3c50402383b56716e0f4bdd8e9f77ef0bf682836da9370d39bf3f1b70cc382c0a9d5346b697977ee5992642297c72b227560c43be3a4b8eddde211297d48858fb5b9bb92687a66c4f26bcf067dd174134119982ab05d30f42c9f4385ae22d208891390276c053653059319da0ca69a39f421e6ecbd5314d3e82b570bbe1e8d05e30ff76e982928a768cb544243dad5e1cfcecdab7e62fd54962e5e4f7a556bce62353dd2c8bd849d9028fe148f33c72f93b219ab33c90bc61f65254f8713868fbd8310fa79f03b22f2e84d3213bbffa7b47d8ba2f4a49cef8c2e8d6fc8d9f7433e56cc750c56d7cff337eeb8d390007f6aeccaf50f771224bd2afb0ee4610783fb2702eb87b9d26a49cb5c993877715d803fdc1cbec94b13bb3ff0c1db7bbd55c6cfaafdb965d490f3949d6617c516b446e6e4347b5351d3018e0d85f25e5d97a2e462761ca6fa2302de383df22d52bbd190f7e7f69ca03bc6cb89034f70465e67a8ac80acecb79aee4a5f09a0f293634bd2c22aeeb064b0fb354a39bb38aaad300c798baf41e8e302586a0dec50d3ee7cf7619582aa02b003a119fac9d3b72192008b8a6d9de5b6822eeec428da7249209c539f3a8ea01791ddb8a7f7628bf7a13144e8cff89659e1309f0f1f4cd286c87db50041da1de6e09eb84210d9cf0f1420b3759e3c35594af2c2fedb312754598b4a3e436a9c3fb3faae8c6afc234256d119b524cbaa5f9dc25e55add364373c821a4cc9c9571fc88cb1aa4e5d53fa72a88fa31472c0d0491cf95231098ceb76d53b9fdeedd44c1f26afa0fc00f7aa918a1a497d65653b637a32fe5cfcb468eb228eeaef9bb106b04e716d94d90a8f0c11b37a9c5eb3812f83a472face6105c78fde86131dc0538ddf026bcc22b33d38803e2ba3659ebe425f57e201e75b1f0df4485c2f682ec5c51481e75672e089a8378aca10e96ef41bd2a797cb3ee7f83b2b2a3be89d921155c5b49ce7fbe2fb5db1c4bfa470e1a5abaab7c54d919c8019871e013be3cc7f85383c5d2dad51b695c0eb4e296a8185b234e0fc1ce6ba9edde2cc31d79c90cc88a9423d5446008a8a965e12fcc7072d67b10e3dbf322cb6f46f7c4c7e3092cac5c211e669336322bb4c7d933d05ec4a0815bec5654e040877fe4b8447b1614dffd78cc12b294cca42434b47f6b6bcc7840ddf8230fd8061965c7f2ac1908a23665e881660647faf8fee0d16a4c106fbd20de8a9e506378bcfe92e7243e7726ec0204580eeb82328d4728c97ece7adb3e82b461931e3d7f9231db4b97ef538a04c50e549415e7efd036f6b64032a8b31a3fe517314a82db1199f1e8167b02535bfa5d6f707129d4ea1e7abd727a038805cb5cf548f7a5f2ecd1fdbb2700921dddbf5e472fdf83ecc8c9213ce4324c7ced7ef4628ae019bceaef6556ef82b1a8427cb2faacfd3074d9e686c39cfbd5313fb766e6a17c6902026dd3599a19cdf980a205905a04a2acde838584c45158a73fa64d231323bb92ca0714cbe6d5aebcce8c058cf9e345d19282b74eac07d0aac40c2d667f3537ea5568c828fb292f570c232994a7f543fd633b011a4f56cf027e6c4b435eac59cbb32a93bd6e0fce1ce8884b737063900a9345f8c42e404b8bc8bba7277897aa39afe10116b85c134ccda98cc4219288e68833e67311fe4319c1a7f4e4145f5c76a1f925a5b00b5313c80156ce54d1f3739c3cd19badf1a3254b86317d0c5691865dd7f6ebef4e7f3c6605e219c7742b798ae0b4b7f180fc7633d32d7e55a58174bc1d4ef433672e98de259d69c9a8f8a440538c0655866203657ce62fbe9b6cc4278c9a354d762cea9d8b0125566fac2edd1eb9dda0a83428a43c2e6a7963cca4355063a9430119410ac36c2ffdf84", 0x1000}, {&(0x7f0000004400)="8bbf7629efe09e01c9bc74b559d33afbf45d4c865149925582ff2aaef4dd083cb49f7e57b9bf5d22d070587cab0b3446b0fc90d58736b5eaf8aa1d684de75807b03d78dccc46eae447491d850b4178a87fe3d3f34652b793177c9da0d545e5de495807214188b8dab91bfd40998a24c4036829c2a90897fba295aea5c4a457592da9fe1d06b0034969be51ec84aa0622f1405da25189f274fee40689e992edd76a175aa6a29bdcf1a678527f4f3cdbb662c5ca7451af14b3227c1944d8561bbb2c38f36dc1a38d8b34cdea45c840a4e324b928167d9cdfdb2dc66a0d87056e2a24a0cbf40fe32a5ea805a2cfc528900dad9d0de42b601d802734e4a0d50956d568d47db8f8df6cfff35f71017b52a14542fab7a24374076930748c48bca672a3672d2d09da644508c3ed17f13e1404e6724e401a3652ca7e7bd611a3a679f36d9ffc41a55ffb301abc2add59227a88c9546fabd3c8982eab3c16e752e32469a31c6aa351fdf292796a7f70f11e7b9b7260d90a9a17b14beb73039b518070bdd4d6a85d95b2d1ee5cb0d880734280d701940489faffd5388700a0c58bec2ede65af3325625a0d14ade7756d802c8490fde9d59c1cda6f7574bee9917b0db7c0191d94e4aea6edca930264c20e0b4375ae158839a89efd23a2f7191b097d020372fddf994248d25443f3fe8bd425ece2ed98e295b3e8d544897f447ce5875ed69c582db51cac8f9e4f580c8f3cb820ca2ac5ab0aefa185c2c6d68ffa008bbf58d906a8521ec279cbd6cc2b5e4314ff1e3b47bc048fd8833df51b185412b32ce6f0844d58c45cc1968c65236d4d6deacef9a8332d523a6fa867ddbc09614c71661488cb989bb46699f867b0cb40251e1f5e0068f0f49b90471579942d64cf27b85011de30c45a75d31c5bcc7f465cb4b76ed1667de5cb67bbe706ce41ce5b9a87873e0de05640e9b97fd01e26d0cc64d65e4982276bb411a12c29841643c61262da56f1f7dea492492f90bb292b06327cc48147568eb73de86c29b00a86b3639ca35714a54059a1478c6149069fb48425f4b35b81ebcb3b465327f987756613c44c05f388384fc5c7ea01d2c4cb216314396123a206ccc142c1cc71516793d3dcf763b819cc932f10f693c1d5c24130a37d932ebfd0baa255f99d3639e548822a42dc13b1147203e2e5fabde8193a433a948d923183257a41bec149c34256c4c037ce25eb2532a76f9bc64ebd84b4edb7a90e3f6edb0fbccfca6df4f888e797c1878c2d71cb59e0980b0290866b78d30d2a9e51f9661724f44ff26c6f26c05022576a120cdde0fc827811e4286348e363dcf6eae1a8ef6f94f596b87009f53608f2c8edd4255d468b0bab997534be6184c2b3608b0f8fef7ebd322b0da99cc407526cdfdc095fa59a9945a9234c078a6431a7e00ea5159228482d26d4f2ce93ead428d8bddbf8dfe92dc9957ab4f9f710ae22a20f63a7c53985d4934ed283e27c32cd45d148f4965444258e7f5dab0881f17f9d61b7cab7d20710afca52c444a3a05c2d7108a20fe648f8290b5fd22b70c32572c69ae9b68c622a77a2707811df93c143aede12c3ca9acb51857b6321ee85e2d9158daaf56d452bc04e49440dc7be51615d3155da939cf6cb380d8805d3094390bd494e96d5b42456a239985fc39278f6155628adb1597360badc108e4aca8a9865464d89645f341a4674d4aff7eba78e85950e8868fca661f996281494095ab9e26fb5a55b04a0d8fd455c77b196a1e325e585630bd989b272695e5432d9b8eac5a8711b8ca06786777fe68e12a33e88374f8ece11f86653f7a157a4f133574f4371ed9a0c12b2dfd840c75bd1669f1a80b4f6f1f4ca41febf5405ffa6fe7924d48886dd037ef5301532c047acd80525ddba5912443350164e7fd6e93956fbde183c44e9620e417f9e7172967776b6aeab165fb1a18d34ee43c4d38c4a41b6634a7b9f5584960b1cb8194f3f18523efca6a9bdbbec01060083907ae78a5bec113117d4aad430d32fb08090b248b01cf5a42f4c317cd0720c70dd7dd2f35124aa7a3ad06a6b73a236a311a0d0c518f076bd98eebc242badf29ffa628b58062b166a9452de53d6f19feae02eea9b2037cc843b4f4eb05e95264476186aa250e79a4a3c0a7ab2fe46f8d932b981ab9debc4c8a6a5f3939d858fe9c0624a6b098e203026bf23c3a7b1d75decd76096ba202b74a0b778a6b7ab281497a24c1e84c72a2a10a474ae27f54ea75b190d5113de562d332e20764828197d1cbd247853b914caf24c815a715925091d12c0be622eb8b9111c2d172ba3bc34976883449a18c61252db9937752183c0c6a64fbe178788ac3f11f49aaa248b796322d03198ef9fb370608fc2fcb0d15248d41d88e7885e19d427baf2cc59de8f83ab01756e22e38f8ee6d993a647157ce9ee5d65b495d3888dfe9a0436c428b207328f56a8f20604688d5b024fecfc3be3072a49aa678847bf6fe57a49a129ac72e601f3dfb1f416d3d12502d5b53860dcd5dcfc1a47fdf072eabbd3255f00d5eac0044c3635510b162f0a52c7b8cb04be6f5a35129813b53fae28206909aa6d85006c69f174029a2a46137d4690a86d700e53e2c068ec0fb073b04e37da67b806b04be60afea3fe9e4ebbc1decea09864abd0c8b7dbf06459c05e7ac1c550c661c104f315b2712aa7ab283fca0b7cfdfbf33aeae3b5997b0844cf637873e0d8c5c54c4e373c5b10321c147390da3abe39abcea936bdcc5a0d83a4c21313ff5029310e7384c5a19a5dbd89c8344cf850011923782c1ccdf9ee408767c8948ad94f75389f1b18df4e27e382300f537d4101474ae294a498b0c1237a5d09506f8cb5ca928536bc99a01b6943164d7852aa3d252eb110d5562463044df575e12615deafc82e20c4c138c0dd562f765c021276bb861c3cacf3eafad3533500b957c5664ac860612903a8081d02fe1108f48a9b709c8da0bef93eea79f772126dd41e86dab0a4ca5ab229192a26af9530b39972875ce9d2cfc649a491e159466eb760fbae16da56d70c16247d84062b98223d9dc158bf14402612e22051e0ebf21f139f4712beb10ddfa9207df2f758c71b3cf1d317602c8d698e3f53935c9ba957567c5e2321634b88bea68edd161b391d1c63e9f070a23d7eb2b78ad5b56cf18090b8845ebe91f32e20d48777fd884b55517f9f345026a4c41e5f1c7e2d559bcce163d278e2d13358091050e6760ddaa2b761d253790e95d56dc8d65c0f702c08d5553f9a9dd4f364e50b3c98f43d8a5e0f043c157dc3f34d4f375be18df39e0e094405112f0f43b05c7856d5a321fdf53bcd909372c8ea70c217ee984ba03ab567709d647c3e95d5cc100f7ccc284aceb1f463d7caebc74b2514d5fd4c0e96059e5a76d4921f23dadd78aee4a911117276cf55e429726a8d4bc54c9dee10d505a83d0a2b4a3317db5158f69d987094609021f32ad2279e1d8f1cd0cb2ad1c08022ec3cb9215ecf8c08ddfb14712b3b42bb014c26dafd54afe17f05099edb6cff7edd67dd05221367265de0ca9fc1e8ee7c84f26f42732a6f3764a4ebf07168231b16ccee6b8d82b5f73b748f0e088f59068c317cafa09e8d658cb512fa8a0d2bcce97046d73eaf30332b519f915b687c09ef79b7c7baa2ed532593a36088c863ae853ddb3195a3417533ba8bf1e071e010518fe7a5d4a7f3717ff2b3def3918aaa066016ca2f618a8c2c5b04e6c4f2e4f38c9f4f745195ba38d15d05f1ae4e60ad61bc92f78fad000c026a0474a810a87c59e193113f130ed5b0cbe3e46369d7d9c098b6455275129569c74aaef70460a119da97cc082c55c6ca4f68b00644b3e76bd197f42dd9c837b94380605b855af2aa48268bf025183099bf545bdfe39b3d6a18267db20a6bd35e030b3537fea99d403626f24649264946afc0b592e1fb8dc496b1205052d48fb4579c44b88da414ee2e2c1af9f611f44aced02ab760e6d564b053e0fa285d34f46acc2e33a8298eb5d9e1e815bd61e7be00070b82dcf9358e4d3749bae9c2c2db9c902cabde19e34eb98d02d55ac1a191b1e7e18cad5243e2c971b9d92cb104c26669c1b99548eb7d59d9245f04cf5362886b7412ccc83a33273c2368b57ac5756952437b0d5e64ea763a55ebdcfc3867e4f8f812b71aec816b10a254ef6f862b856e05e158a4a56a540b7e2fc3382a830113ebe8ac6bd03f7d56d661ddf965d6461c90fbd5f6a74b429c82bc78ffdc0cb78f8ef76124b4697edb45afcb0290df7e68666c52596b934e4199e73f553120fcf21c0b3ca7e7a76b805a0a229b7f40e4483049f3711d80ee68d9bf6de2f15f80826e0438af9151a74f8be986ffeeef5f40fc66fa1be7218b8b5134aa0adc3afe08a89b332707f693df04fc7ce3bfc788a0c8f470587d28fd8faa7429202767c68d0294cff930b03c251a848a52fc1520bfef70e5b3df755f78115767923dbad460f4cee461325cc33118d48f38fba40119c7dd8ed6b7b3ebeef7c65508a05819cada0bdfbbe3a1c6a8d0f82d5130f2f3d6d96af3c76b7b6911e44a4861ef78de6461966b03aa05801e3ae0ba4f648588e8d7ddcda0f76f8ef1a76b1b16cbc3db37cadf5b825764dd3a2b2fbc850147112b389f9fdd7d96fc8f1a58bda0b25cc262cccf442b19034a93940ff9d4bff96aaae9206bdec0c23f3789875fe865cfa8954f75fcdaddee17e296b287a49b34dfe0def35f2b59da773340b002b699ad44ba20df0237860048083d5b9f7d69e045343fa645d1673e1d86254e22c62a6f883e32a9effd1c13d91ee0fd48b9d57f030a458f39e3e75edd571e11957f9ccbe64b774f56bb188fc93f733ab59c87a88703949cf873170c3dac46d53677f6fcac0d9ba3d6c36505eb34a9e08817e33de84576b824c30a9c2fc87d2abacee3e2dad46386bc36aa35ed23ea11fb4d2355e2a01ef7df578004dfa57e3eb00db2edf1fcfaa4a608d31facb25a5de6e08b35151fb3d73ec06dce9e3326082cf5ece63f0a3eefba8c5fe9a69fe30ac843cc38cac8442574e9956d0fc8c257482458c10c5593958c3c1d3d1915996780c5beaceb3990246ba325208b9a9e0eb163b153c36ed73f171394d8b5bde20bb392961715aa2772386ecbd442e2ef7cb1d09d65d2623f7c4e2d79d3be789048e9a2060fb7b887251b0c7e0a7cb3dd4ddf39195502710f7d3fcb358fae6840a411846efc0be2b2716b3edca31b1a731d4492d686b12f9ec8c3006aabeb41a8c2b9f53f7f9c75010cfc53eb639596a8d301b07eaacaebda45b2547447c14a4d087cb7e404a2c94549399cac09aeff7b52a43908f8b6d66e035911b34435f8790e77abb2b407b8da897a5dbbbf0da378af668639a4fff02a9393226233766e2c535a5de7e0bc187431b72608c9fa8aea2c6bb7bcaa9d2c5eaf281fef168c49a97ca7ada9e0a9e433603684acac2fc181b57d4c899ae4b619eca7c308035bb7ca5b94b3aac37efc0d771c95aa9167d0169fac65ef5879e27a6b9d51721567d6f2e95170c9080fbb83ab5b0eefb80a48fff0304f71afa4062e84fff8fb96ed126564f27b6ef936e0656ea998a5adc2ee19435baa31b2b62bc1b4a9f13f25ab5875139bad731cbaf1d6e4992a3e8831118fcb89e465c7bdce1ff10a1867c5817593f18cd62f82b4107a14299ab2519ee05ce3871982ca4f19199b04512c7548223930ee2be18ffc4b726777711df1f07bceaf856da75ec1a35bb04ddf63a7f3b68401c74ac322e975a94e569ebe91fcf499e6", 0x1000}, {&(0x7f0000005400)="c1b41f9f3168625bf5ece59ce57a8f0d9f4bbb243d07a923cfbb0a8859895a9a9bdca488d9680aa3585f891e689608819fb9978d57ad1492a63681bff65080", 0x3f}], 0x5}}, {{&(0x7f00000054c0)={0xa, 0x4e24, 0x617, @private2, 0x464}, 0x1c, &(0x7f0000005580)=[{&(0x7f0000005500)="4b08cea4af7fb5effb5ce8352b9a28251987308b2117c8a4ce32670e1f7d823d705b6aa4faccd6f1c1475ddd9df92275c681089bd427e43913457ef96e241112fefd00c159796d8ff41d8044f33bfe8823e3795ec3732aa52b32f3c38eb9df5e2960badf856d67f6158dad018b85dae0f5478fbc407ec2c4b1", 0x79}], 0x1, &(0x7f00000055c0)=[@hoplimit={{0x14, 0x29, 0x34, 0x1}}], 0x18}}, {{&(0x7f0000005600)={0xa, 0x4e23, 0x2e1, @ipv4={'\x00', '\xff\xff', @broadcast}}, 0x1c, &(0x7f0000005700)=[{&(0x7f0000005640)="5ba2cae580dc49fc169d0641e18bf968d9cd146b0af56837401f8288c475b9bf7d4fbbee58c73519ff043c0ea253b72d023724c9b1fcd73985eecfad808fe32b21790f47081462d519533760560342eea3ed0b4579db1b774695a0233be67f6e300d45494bca87a6d6e7b789daba9203889493257bc950592ccb3f8adf120221d25e3029743efbe6b5cb72674d2e75853c122ccfd82b8144a732c54a3b29d7493b4b948b2d57fa852597ae42220101a4c93201cb0845ffb189c8", 0xba}], 0x1, &(0x7f0000005740)=[@flowinfo={{0x14, 0x29, 0xb, 0x1}}, @hoplimit={{0x14, 0x29, 0x34, 0x20}}, @rthdr_2292={{0x38, 0x29, 0x39, {0x2c, 0x4, 0x1, 0x0, 0x0, [@dev={0xfe, 0x80, '\x00', 0x1e}, @mcast1]}}}, @hopopts_2292={{0x20, 0x29, 0x36, {0x1d, 0x0, '\x00', [@jumbo={0xc2, 0x4, 0x8}]}}}, @hopopts_2292={{0x78, 0x29, 0x36, {0x32, 0xb, '\x00', [@enc_lim, @padn={0x1, 0x8, [0x0, 0x0, 0x0, 0x0, 0x0, 0x0, 0x0, 0x0]}, @jumbo={0xc2, 0x4, 0xffffffc1}, @calipso={0x7, 0x28, {0x0, 0x8, 0x0, 0x23, [0x100000000, 0x1, 0x100000000, 0xfff]}}, @enc_lim={0x4, 0x1, 0x40}, @enc_lim={0x4, 0x1, 0x2}, @jumbo={0xc2, 0x4, 0x7}, @enc_lim={0x4, 0x1, 0x3}, @jumbo={0xc2, 0x4, 0x80}, @padn={0x1, 0x6, [0x0, 0x0, 0x0, 0x0, 0x0, 0x0]}]}}}, @hoplimit={{0x14, 0x29, 0x34, 0x225}}, @hopopts={{0x170, 0x29, 0x36, {0x2c, 0x2a, '\x00', [@pad1, @calipso={0x7, 0x38, {0x3, 0xc, 0x7f, 0xbb5, [0x40, 0x1000, 0x5, 0x3, 0x7, 0x3]}}, @hao={0xc9, 0x10, @ipv4={'\x00', '\xff\xff', @multicast1}}, @hao={0xc9, 0x10, @mcast2}, @hao={0xc9, 0x10, @empty}, @calipso={0x7, 0x10, {0x1, 0x2, 0x77, 0x101, [0xfffffffffffffffe]}}, @padn={0x1, 0x3, [0x0, 0x0, 0x0]}, @pad1, @generic={0x20, 0xc5, "53108c437e0999adf47f6778c43b28cd4f49f1194d80509b18f68b59bccb057032d5b475c3e0769149a1f8d3d8df06aa5668d8ab3d9a0e3c742cb3153f7b5e65f2ba90bf5fa46a10733cd35760280cee062eb242944f78223145b8bb8713bc6c2c42b7421fb5d762e7a5937a45ad7e93862a704d20401a706cad5be91e94d120231e89170c76e078121e8b6459f06f93f5076cb424940a59f2058a47c441fe5ab195d44ca4b97c8df9c4e7ff9973840fde7b426c7153f69995b820745f67b528b6880d30c8"}, @pad1]}}}], 0x288}}, {{0x0, 0x0, &(0x7f0000006e40)=[{&(0x7f0000005a00)="3f24a61dac45a3915be9e85c3d6da7bee1ce6e27a2c1d3dd2ff8ebc2cb17456e97bb6078f9bd2a94fd0b581b9b4cf5c9430c66eb6761d3e4b856e0f1ff5b1c4a58a33e33c9dca51f68f9d3ea06214ba703a44e2aef21a72142c1bf4b82cdc1fec109f97af66ee8e0449c0cf7", 0x6c}, {&(0x7f0000005a80)="be1d1409ddf392b46d366bf209c29ac46157627d5f0abbed918a67e969c2fc05b296f94cbbb7fa62ac61a54657e6e7335159e925159e63943c4e0c04ac85c8bd2c231266ed49bca18a54c1e01f4048d991c98a18679fd07f30815a91a3ddd98a96f003036347af78d8b3e802dce608606c106ddf8632958b2a69281bb5b97df407d02a6e29a44aa63cf626802532ba242286e4769d4aed787adda75d72200ebe92070f6a558527273e5e185168353e3d5ab078c4a75a24810c7e73f5990bc36ed867ac3003ec50c374962f69cf1c8d239809b2", 0xd3}, {&(0x7f0000005b80)="dd3b04f9beac32d22f8c9f9dbe23819f366edd0bedfebca2f6569885471bee84bd0997f8a2b7bc3db803265bf61414b77f267a5d45bf1257fe10a43d890e3aa39201be83a0c3783e55698de6f828d17096fe6c9b2415dee92975c9eebec69b05a913af2db2415e4d7141f01c5f5958eb53d8581cc49d4e4d7d9b960c29bcbd552268de2463fd544cf45d190d28ef0a22308fded4eb6a04e0b131ebbefdf10992faf20d282b0b3421", 0xa8}, {&(0x7f0000005c40)="7fccda0cc8ead8aae0408a025b4df83ab54fbef5531598c67d1097ecd28ea443b6580499bbd30484f1be206f90ac72178377a35f5df1db4bd26972446dc085019baed3763c2f7125c6fb00ef5ef3c3074cecbb0e7e2ebc35c88cc8fe883d0282ffbedabbfab50e048ed6f91b84b89ccbc344dbcfc642c8cac6f34a3b294ab53866f46ebeb0579b388fe36483d482279ae1bc6e0dd1e2142485b5605b5f4cc853b0a374769c7677ba156392c2d35ba0c131242bf2ad15b567666f0e1b0034d0786d3ce9b5bdacca5671ecfdc32b0f7002fa15da4fc6d4331df428bc733e203bb9aa1507af4f09909e", 0xe8}, {&(0x7f0000005d40)="bfee1056b57b8a87fac60fc21a46be552c01e563c8e40bb9a903f5a654b36e82f21c55ef1ed1a5e179705d2626b7e5ec0fca187a1b4c1b3c08b2f49fec875d30d734ac1f6ecdcfc0a4693a0386de8362fe491b42631a86daea0fd6ca8f34d545584cce2c966e7a35060bc759da4de1b0ba6c52fa9332bb156059256bb35b48b1fbb9ad5f8967765b24a87230ecaceb481a2a434aa4db2e9961f7ed8126e3175b807a67a465b02b9bc8e332641aab857d7047425b8e4712bf5b5e9557aa53ef857de882e5d106b5141d53d87b7078c19664f260fa3e7f54a6cb281cf2c155642bb11156cd56bf882b6d73c643c2f828a93a7c2f36cbf90e2991ceed4e9062b43b13724fc47942b850aa4250c8133f0c8fab15589d3caea45cf891c47b5e3e5b9c3e5677ae5186aba6be5217a4d306f5ccd0645d460b7d2bfdcdcc48b643ea54333a8d66ae864508d8922b1ad3c44f85b27ccc92c52c96bdd9e2d4f918d771498742af7a0701610c906ce401e3f2bb400a67f711b6ac624986fdf0e372391bd8e30b07648dafe3f0b06976f1ece8c4f33483895ef016e26b0f16f8acf622aacc5bd9c69d5ace1b7e8ef753d339548f555bc94000683f1384a1d8f854572a15ab74db640a4aa77a03213e8a1bcb1f361cee0dbbeb05e48fe1700fd262dc094672dd1a1c8f2cbe825b246c0df1e7b9b49fa055fbc68e334a2e47dcc50ffed3a65dab7dbda3ca1788681cfb74e212d445c30f9b603aa3751db72eed21b90add76d1cf6306b519cf61da960a0e472ce1cd4c779d7227f01b8d6f61636b1d6d16bf03b8f8dfd45f4021587ac4a1ff66aa6ba7c8725ac91f599c4db8279cbbb91cdbbb45a18d224741b71eb79ff296135f96c67689281242ec0c1bb694618b5597603e3226bde62f885144c2f65b160a291e08e8ae248180d9cb20f9635e15dd7446f4ca692fa8ccbde5a20f7951fb33d5a50ccba4f509017555f6699dcd1d626e29d5aea8986d96d7ba488aad9f512ea50bc665e12e5e9d7c6390637a64f5cf2767e717a8ba1a746c5aea061881034e89f58f8d075d673307ec1cbc1b598f18cad0101f1abda02c3f5fe6f72a0e2929a95923cef2c01832112a47c37b91cb8d65c4ca41f81d450ba92a32d482c59d2a839f0c36c1c301f3992307a5be2601088146f9fe89f33f7eb0a4b9ddd52aaee0663f683e352715a0594f4bea49b2f090c85fa7bb3ac11f436212edf7fa7864cef8296a675af92b5ef5e5131749646ef451ba3305cd00d764c997b2da08ab065e5c8c2d0e968303f2adb34f118e3f1159f47e51427614d208c5e2a4b2247d31d92df9f87197fbf5d5637bbe85e47269d3a41e2d4f7a1ffc0a6281e6624e9fb10f1eae343ca5b94f19f92041467b91128a4a7eefd26d46bb0e816d14fada6c50de931a9c6ba006aa3156d86a01fec57725f955506953a5b8364f857b130fee93e3a8202cd9601a25e03fc301beef7f66533f55cdff32fc0ac0ad154b864e975ceac381b2d8fce3232a9b8a9b41a7e4a2a671f57c275250efbf75f54773b3b00dbabc21e64ffa508173c9deec95a2e4298286c3eae6fbd9f22d28c16c475e2cc4c5d68b6e421b746595601941da19b4824244caf8cd6d1013fac0cf7b1e25909150c29e31d077f5f0f2f239f9164526191c1e604525ee6ae34314052e13b38ea6bcc59c46c49c3f332ba4129acc4edf364eb5e83f9affb943368c81325e787a85c96a6b95b6788d62e8167e377ac4d9f548705ce15afa8ce9d76b3ee238a1919b3ca852669df6efc695daa1677758ef49c42487e5059b48d26816f9e7e352faf84ccb0520f45f5ca62d8b45ef19e02d51010d7dcfd572cdb1fc9a884eae5e41ad0c5f06bbaf2e909a87233d79c6ea44668452ab985a92175890848590b4c2aa322b9b01828484ba84f6ac91b262cd42c8d88bcc41010c9a7cc1d3afd617a3cb0ed2dbaa91dda6140b04586f85e7d92ab92989a1cd676060800ebf92eec3a0b7de05e15100ba916f03ad1dc7753a4426981296bc9cbac3b773a097ab16ce41f4067f5b879e7b4d040a3c9360ef77c6a31e6e1b6d578086bc82963e910a9544b3ae2f771f5099932a18391ba0987cc3e66fd7849d7062b64ff12b80c9584026ca3751d49e6b78ddfd7f968bba507078b2eb90bafa94dcb1c9f88c777e538bf48ac976bd1df87c01911e35e3d47ecc56bcce9d6bf8559b135b6609af71a73db8923d70b041ba758fc1e7ab1ddccfc94071cbf78a4ed929f5ea9f1cf5373753323a22f99eee1639603d48ae68ef868f793ad02b81ec6ad6293df0d37c28a5c251602670f9b7a6a21b327abb84d46dc6960a1f4d28a9aac3dab1d4b919a3f6d37f36147dab3c79c94b848a1b995bd396a8435d4ffa5e78530a802f79a3653b28ea52e154347d0c960b452b3c7f8341b8b26a778d43466d9a54403233feffa74ea48a68f5fa8aaf72f5eb51e3b07f6c2449c6f5f2f970edbda9a725d80997bebe0c54373275b56de857a742e46479e09e663fe756b11ac45ae3c141939cb9ce03ff1e6fb12d01be311ce5f59c9f124aa26b2bd87e5ffdb7f73545c7aab71633f646dc85380800aecd2975447d87e7eb6c16addaaecc919bb401789952e9ff14705bdcad744af1ecd018bb7f79bd5cae90427cff478183f3b28233a55ea750d3d88972869c31a2e0c932d06a7c186c4185efa36233edc88c186fcd2f1544d382cc0744167f54e359d8547538f7f6926ae017e5ba662843960ce1a3140fa1290f9541db6ff9507de2fd0e8e77dc0a55c8096ab22a6c21767e18318706feec569be584c13d5337d52414ae60448aa383e780785eb3f196c7f10fd376f911759e420ddb27cbd744fa2a7e2dad5543a07ed103e5a467ceee127030a0cb48d98f7b4821926dbf43dd4f95125feefed4f48c3195a61d8d8a1e848244ed9f7b332eaaa88e55195a509daf964b224e7fe34b5092f6ddb6b309d57963651d1ba1c447e02711374efeaf4c380979b1810f240d39708d23476c686edece9d991048f7f710e8a31a0bf3bd6479fff543a0698933261c29955710f294873df42599d91aedc0a9f9172d131275a83d1bda56784cf1139dc33dd33ca0cd2d0c4ec2f4912e2f2ee59c1b3ade58553f7df74929517dfe28b989cf1207f1e5bb530c3afe62f404af18196a90222323c8189a4bac438ad35a1d31efa40e65fde299ffc582195d2e995666b78cdb5e2d10913c935e92936c4935eb503977c635bc4e34ff5c26f8a8dca7bb55c6f99b4bfaa97b75cc08c75dd133fb8b09fa028a4ae9e9529121fcd49bdad51f0e128814cef7772f3e822cb29bcef1544b0c54c08b6d40f24727897a0a0e9e4a15e829bf4262bb0c5fff063e277e8e511ea4b0c91093ec3bf215acdddc3c6fac93f8996a2d10ea838fade70a41b354cef38a53ba1c71343b0b4b209f671c63051a9019f34355676a69176101d92cc0a7f2dbe0a4115c3f9473a9614b0684f99feb478691b9fd250a426d4e8bdebe27191a815fdb49f1ff2bef88f8cdd20142fc56ae3893c68486dca47b7a60453508bdccc410ebb0e088677ac1875e562624750523634956baed75fe7a10ef987a6e4b71f303a2c41c72dfa86e33bd7cde76c6c033c74418b95cc37638314e70e6116d7f321a11ca13db037bcbf41ef8b6b8bcc95c32d8af5f984a961bc47fbb1c5233d200a95006ddcaee294ff7caf5a8aa192023504eea099bed488ac8074c0df343b1e3d73c3073028654398f81864911fc106282ff5e22f2da39c709380f71e28d391d171eec72fd8affc3dc91b2d0096e2943270351679483baef9671325dfa1ab923659257a973820c2e5e4229eb630fa6a12719ae8b0ec0a498b69bbc94a8400c54b66560579736523a655f7481d9330cd7e6d9db3e359f9478bc1eeafb969a0c4942c0d1bbaa20b5be683b0a3f4d35d0df39b829ccbbdee0c6b39bfb1964f7b2a6f19055a57de2062caa3ed43595a92ac949421fbc659c31256d5c7605ae88cdb119829824f35b256b3de1668a50a290a356e32a645e8746069e4e3b2aa65ccb2e591471ae241b63f4495dd622d3df0f312eb9e243712880bdbc4000e0f9e490e7580d9c1f08f7144033a4af0763b93c4d2fd1e6ace9ce8f1c9760b581d51af613eccde9487583fc359ba412e590f19fc132e51564a23c0c2b767760a324547a7d04b271915adcaa436aa17fd1a81002e3e3a8f412b5d281126c86c0b83d82b4566b38ac9ac601d5507593d72b403b39a6bee4754adb27e33a44a713c84fccbd6091024b107940eab518243b4d01b02a253c86163826133717c04268cbe64011b1834b1568612e884211ba20f4b8503f3837ab8e7bafbf4f55b888fdfc4eb02a8aa592341ee755252af23f5db80105360645e0071e0098d6a0cc82176f471d67a18f06c3b2ccd849ce65f4755fa0ab802e77b0cf5f24229574e4e2d8446e45d212d4f2d554221a97b37a381f76a7b918572b70c7f9e37f30458fa38bc075e3ec575ff9e5c1af6499833a90176d3e9e7729ec9c2c81920c9ec0b457c1641963f1491238c3df52cc8c04f66b1f0bbd4577cbcbefbc0c7fa266724bb6af72145af3da1f131bf7990a0b1094f3cfd1b44ee43656c95b9e75ffc8de888504bab8795951c4ab5672d5daeabfccf6771c367910e563e269e093343081264b813667bbd48fb8932f89ba32429659e92dcbda8ce138ca31d6232905c63eddcd4ec935c4a4f59d970325b102f9b4b52c09d7305b88116480d069be7ccee249dd3af3530e4f79c7169bb13fb5573fdcd7c164a418fda059b8d37f91ac9efd69600dda57525171bfcd554c412b4f40b2b8176eed980f1c55f4ae3bf1543f040ff13eceff746b31dac494bb5409ea6d26f62edc630686423a8a5d4348f56b38d276babdca09cdd13d5d4b509cf600e81e8b79954328952002f47b3508d2894845f64c19f04d2dd0d9680fc40494e39a17fabe122c50a823ac06bf04c52827a3714974a4ecab62d1034011d021713464839a6a95ac583b8b36d90ff9afe10660372d9a9c71b1f61598d574c3bc3d779a22a61c2a466a9eb738aca6e55b3c45ae1c1aa26a4e8fe34d7176cc9c25fe553cd1b5f98f046f6a63da13e2de4d953d3e5bc8f559a4edbb1a13af904aaaa832e606852b3ca47d338bffdd4fc2c576cd188144a2a8a4b6ca81ec6e2134ad19eeb0b78a55c62a978a891d4d49050c918008c8b0b5651e1823b4ae89139691177e32436da940d4357ab012e1e394532aaab6f03d6576facaa71649fe3c2a1c92c4d02c48d65a684b3744810dbd6f73326ae4caf6d0e7da163a6a12d8499e4f9f860481656df3d308a55941a061f24335d9723ea2bb34f518a7081d91c76f695fac623388b7b24264d44e6e2fc2a3dcc032e9bb45fe83f89ea14ec3f9b962c53be6fccc563857e1e87dc20454e2f46623b0c350d11930853a44becb53a340f0674dd9a1ebec56b933c53916cc15429bedddddc2d670b1085dc02f0e3086b7b937550934d3e52bd2b4d31ffd7d90237fda2a593188bc92f7790deb1f2f660af4070d2601af107d3586ee94b3c2924a397458cc7e06538b8bce8d28212073fb0706453634d287e5c1a62987357ff3d1125ff9c70f7e60fd4e23af41c7ae93224aab69868db2a20a8ef76b00cf1c7dced6ba8f2e313a222dacbabdf612d659fe9ac3721d9de164be80e9328f5ec1133af28fa496f5bf336c683e3a7f1dd65bb6fb2ea300686c7f1e3a7b5be1927ad51065d4977fd8ed4582bc0f3512acb9805e9a", 0x1000}, {&(0x7f0000006d40)="7534da42ab8ebea4423f2fc1354f832bf3197028a50eb9562a1319e68849bdc7bd0fe345802d5ef1c48b586968170bbe22168fd8011b05c979289d36b4d05161782c6b1c0a44e9f6081d2a3599c0978c", 0x50}, {&(0x7f0000006dc0)="56ddc9c96cf571414c61f6754cff4eb202fd95314598c614c24739b62940a41a5e0c18d4600b7f5eb7498fb12a81779558faf631673c04ffc0914544a1ff292c2bd37dc13d37c6c99e7a4446ff4d3aba057a81afaf1f1770599918c3ebca305a47c05978576ff0512be7c56b47b887", 0x6f}], 0x7}}, {{&(0x7f0000006ec0)={0xa, 0x4e21, 0x2, @rand_addr=' \x01\x00\x00\x00\x00\x00\x00\x00\x00\x00\x00\x00\x00\x00\x02', 0xa00}, 0x1c, &(0x7f0000006f40)=[{&(0x7f0000006f00)="9fff08a2", 0x4}], 0x1, &(0x7f0000006f80)=[@rthdr_2292={{0x78, 0x29, 0x39, {0x3772466af79949df, 0xc, 0x2, 0x6, 0x0, [@loopback, @mcast1, @initdev={0xfe, 0x88, '\x00', 0x1, 0x0}, @private2={0xfc, 0x2, '\x00', 0x1}, @private0={0xfc, 0x0, '\x00', 0x1}, @dev={0xfe, 0x80, '\x00', 0x12}]}}}], 0x78}}, {{&(0x7f0000007000)={0xa, 0x4e24, 0x2, @mcast1, 0x2}, 0x1c, &(0x7f0000007200)=[{&(0x7f0000007040)="28357b828bfac2c815793e46d9c9d23dd66973f83bcb7af7e6ba0da677af64bbab92adf0c17fef707fbc64e36f26513f5a8bcec1d766a76f11bac51e187e383d4253fc993b06042a941258142cfad391d252ea2c21c2faa5266c7301f229", 0x5e}, {&(0x7f00000070c0)="253f98c2bbb8f7e64e7974e455fee0a69cbebac7e94979a49d788709cdb17e886dcfde86226461a63ec59adaad90ceafee2b19023b8aa04091ea4e0b128b1b526065fbb757e36cd551000d5538082a002ad35f0c436a044b39d166659c4d2b5125b7faf422b55b4a2c7245e0d8717054e261c05c9c7bac3257d6fdeadcf5688aeb678e7967c3b704216055ab039b9e8b75569390c6d00d055249c4c4cd75826bca85d05342028bbf622467e9977d54b909c7f6579c5f97609552", 0xba}, {&(0x7f0000007180)="85e273e9152c799a6557a0362f4181b95bf2794fd5ec1f83b53ef4aab07d5498cb5ff8b1b6e2dabdeef02cd977d5025efe659c331887f2a56f1c3552356fa405fcee979bcc12c95776cbf788e5a06c0f33e39fb12d5dc2f755798f73", 0x5c}], 0x3, &(0x7f0000007240)=[@rthdrdstopts={{0x20, 0x29, 0x37, {0x32, 0x0, '\x00', [@ra={0x5, 0x2, 0x5b4d}]}}}, @rthdr_2292={{0x48, 0x29, 0x39, {0x2e, 0x6, 0x2, 0x8, 0x0, [@initdev={0xfe, 0x88, '\x00', 0x0, 0x0}, @remote, @remote]}}}, @rthdrdstopts={{0x20, 0x29, 0x37, {0x29, 0x0, '\x00', [@jumbo={0xc2, 0x4, 0x5}]}}}, @hopopts={{0x20, 0x29, 0x36, {0x3b, 0x0, '\x00', [@ra={0x5, 0x2, 0xcb49}, @enc_lim]}}}, @dstopts_2292={{0x30, 0x29, 0x4, {0x33, 0x2, '\x00', [@hao={0xc9, 0x10, @initdev={0xfe, 0x88, '\x00', 0x1, 0x0}}]}}}, @rthdr={{0x28, 0x29, 0x39, {0x0, 0x2, 0x0, 0x3, 0x0, [@loopback]}}}, @tclass={{0x14, 0x29, 0x43, 0xffff8000}}, @dontfrag={{0x14}}], 0x130}}, {{&(0x7f0000007380)={0xa, 0x4e20, 0x6, @private2, 0xb99e}, 0x1c, &(0x7f0000007540)=[{&(0x7f00000073c0)="9b493e883c245d00743a2979372b087e556d2c6a1a147d745d361241eb4bcc61ca9ee47092f93b118a97ca16c7f3a842e838d951ac2743f018dbbc269b17b6a381b46d849674990f387743d2570a9408de434f5c86042e5c4fd1cadf764ca1efddb267d436f357f56e347a8d5378d46996a2835df9b410b54472490e9bac68409bf773fdb170e4ac6388cb8cfa8d3c1b6d735fab86c5", 0x96}, {&(0x7f0000007480)="821ac7a68d525e25b36fce772aec357f3861b222a9f729837224614dc2d28fdad2e9a03caadecaf1cfaf620de3d0d25030f648b915c6558d9c28c4d201c71b708952f38d1261e71a4e4dcd30824d45eba558da6910cc9c0b5639637ba053bb8512f8c07e29b320a8f15e6ed4e29881cbbbd4d542481e463217f570df132147dd69a3b4649eb982e6182431", 0x8b}], 0x2, &(0x7f0000007580)=[@flowinfo={{0x14, 0x29, 0xb, 0x1}}, @dontfrag={{0x14, 0x29, 0x3e, 0x9}}, @tclass={{0x14, 0x29, 0x43, 0xffffffff}}, @dontfrag={{0x14, 0x29, 0x3e, 0x7}}, @dontfrag={{0x14, 0x29, 0x3e, 0x19}}], 0x78}}], 0xa, 0x40000d4) r3 = syz_init_net_socket$ax25(0x3, 0x5, 0xcd) ioctl$BTRFS_IOC_SCRUB(r3, 0xc400941b, &(0x7f0000000640)={r0, 0x8000000000000001, 0xab05}) ioctl$BTRFS_IOC_SCRUB_PROGRESS(0xffffffffffffffff, 0xc400941d, &(0x7f0000000240)={0x0, 0x7, 0x3, 0x1}) (async) socket$inet6(0xa, 0x2, 0x0) (async) ioctl$sock_SIOCETHTOOL(r1, 0x89f0, &(0x7f0000000000)={'bridge0\x00', &(0x7f0000000200)=@ethtool_ringparam={0x10}}) (async) socket$inet6(0xa, 0x80002, 0x88) (async) bind$inet6(r2, &(0x7f0000000580)={0xa, 0x10010000004e20}, 0x1c) (async) setsockopt$inet6_udp_int(r2, 0x11, 0xb, &(0x7f0000000080), 0x4) (async) sendmmsg$inet6(r2, &(0x7f0000007600)=[{{&(0x7f0000000040)={0xa, 0x4e22, 0x7, @dev={0xfe, 0x80, '\x00', 0x1d}, 0x3}, 0x1c, &(0x7f0000000c80)=[{&(0x7f0000000080)="7404edfae53ede7d1a09859a6b3deeac57e43d596e3c4306791801aa8f33af74978ffd7335d4b3b3322941899ef8d99ac803391fc661e08d67d096e4cdaa4a67392a3b79bf47a9417b7587a9a1d260eb17554b01d84a38548b1015cddd01b5e16ae9c0e2587d4cf47afe4ad5ab01bcffc6ec361e85f1937b18d4a97da8e562ec85858d6b16ca6d146c4132b5d6cd1cbb867e866fca85b2716e98bf62b11225c17af3dc6ffa8e646dfec27b19025c69ae34499e23b267db917dac85556d0afd1fb6f2197caaf7d6b1916a6adb88c1470eb32c278a3e2ebff4f7", 0xd9}, {&(0x7f0000000a40)="3ede83855b5c4c5cb1e3f189b0fc516401361f677c3432c5a0a53cabce5d17ef6abbf132df751a14d38d4c86fe6ba04579030e2436a835da9bdb64a18fcfbbf7183316a6aa33d215a8c18fc3a09ea7ab4576794d8f315684b0af9ff8f144b196a9b1086b3b20119349abb79e0df941576e1376e04395fa98cb668367eb0d2dc180d15091a7d91ebfee8c11922f045772799ebdea6dbe1c4f5aca9647d2a03e", 0x9f}, {&(0x7f0000000180)="f484cd0060ba8f6139a6ea0eb3f6a338a6677b7298937e8283f582027f7080e6b4a612522c23c0bf2d2fe6a6ec8a56004122ce57731cbf3fe6c946cf1ae4", 0x3e}, {&(0x7f0000000b00)="6dd61a68d3c0fc641b5eb13d48c2a6c3d097cd5a44d96549275ec0f3a236c5315a14b665a8a3e3e5c1d34ea134e7e1546cf8537945e1906259bc856d5d6466b9ac37e5b96f1722237f2b00bddb9046d96400cbe8d52d550787cbab808c5068ea832ffbe112548e4bbf253908e5a45cc4d2d8b31664f5bc9f0527bbd3b3601f328979d61e6761132c10324de1a0676c97d8f8e3964eef59c99a777116c41d90eceae5308fdfd7aae49d7170a798fe6f7da775e4328cfc31938aaa30aee7a6291254796b3252240e7ec9711429dceb10d9bd880620271c6feea17780f280a0ed9fe092", 0xe2}, {&(0x7f0000000c00)="db27c4ff6a09ab2043e1d2cbfc5e53056ef813b840fb73c98b0be1ac0c8e3acb9a0c3460c890a4c6e5a227a85a0d38eec453787cef082ee62d259248562438700f467ab5689076ca6f5381b8d89a252f2e5555535992b1", 0x57}], 0x5, &(0x7f0000000d00)=[@tclass={{0x14, 0x29, 0x43, 0x2}}, @hoplimit={{0x14, 0x29, 0x34, 0x5}}, @tclass={{0x14, 0x29, 0x43, 0x4}}, @rthdr={{0x38, 0x29, 0x39, {0x1, 0x4, 0x0, 0x6, 0x0, [@mcast2, @remote]}}}], 0x80}}, {{&(0x7f00000001c0)={0xa, 0x4e24, 0x3, @empty, 0x7}, 0x1c, &(0x7f0000000ec0)=[{&(0x7f0000000d80)="8243bf28d847c328e11f2f250fc748e7454f01", 0x13}, {&(0x7f0000000dc0)="e98086afbeab9bbf0c9ebefa6d493c4b5c589e774d12f869db3bdfc70618b78fc72cb423931a0b2bba3b4c746d2d68c0cb151628da47f9da4e8cf5342952dcf45e1f76a285d0efe301c306700d23abd38ef37b14de999a69d2d30efc9ab8d7619317cde3758c4fb8be5f8faf0c97cb2cf1937279a51a1d41917f8a42bd986eed6848f54f6b28dd9496305406010682dc04da27f4a9bac7c84ffefa628aa3fed21ed48d0ebbeeb8b7f3923bfcb7e801f7b024fa930a05bee6ee5c853bf9461c0854882e7126d4f182b4b90e4431361b84989c3aea08", 0xd5}], 0x2, &(0x7f0000000f00)=[@rthdr={{0x88, 0x29, 0x39, {0x2, 0xe, 0x2, 0x1, 0x0, [@loopback, @private1, @initdev={0xfe, 0x88, '\x00', 0x1, 0x0}, @private2, @initdev={0xfe, 0x88, '\x00', 0x0, 0x0}, @mcast2, @local]}}}], 0x88}}, {{&(0x7f0000000fc0)={0xa, 0x4e21, 0x20, @ipv4={'\x00', '\xff\xff', @initdev={0xac, 0x1e, 0x1, 0x0}}, 0xff}, 0x1c, &(0x7f00000031c0)=[{&(0x7f0000001000)="5fb906fb71ad76120c1c90e7eedbf0acef97ed911df0fd053d1391a9f079a43751f62dd438bdeec3123f26cc73f874e789c29b6d866eb07df5f18002b214056c2662a2fd0020b430ca9b6f3ae85ed7a61a73250935736c7f14167a9d6fba9d7b01a315fac51b2be36decdd98affbc9e1e3383e9f71e27d239fbfbeb41fc5a84694dfb792a606630845d883f0f6f15e0a2cde0115c6376d0fa81435353aff60a08b8ee81ad7d4d5fc0027c04594f7414d6e1de1b7fd1b8250273d63f2b42295226d6871cb94f7d0af49105038cb257a442e7f59792ee843b5a3cbcce7f5ae361d3e36550c5e3872e090524f9223949de4f2a696d14a9b5056095af7c414b98cc8b429c67cb4e1f991f29e6f505f3aa6956d733b49ba9a85bdd634feaf1613ebb8f4d7dd3d939510d2eebe9a40679b156343efb4284281f93a206bde9d1986f0dec3c5a263749f25404421a6a495cd1f302662ae176a2e8974fcdfb197153058b2f5a6695932875060dd348d1d233fa9807dc5a98507e8c6fe726167f276499633b4b7463b392abadfb7b68da12aac146004f46e679e3ed18ffad84afc95100fa09bab6944d25ea55a78f10a85a063ddf537bce2ef1bdd57b101db4eb5ee4fbf0a9d6c184342a119546efbcb71ada2db4c1500a46e837cf8c7606d0be343d0e460338a05957274054c092cfa3e0eca9fb8d2f845993d9454f7782c8c3a9b823e674d2db4a924b4946827b3b8f521c517aebfc9542dd799bbb6607290faa4743acf98934892db1bf744e7efcad821c07edc0f9f18de11036b950d92eb58d4dfb7c845730e9e236ea6c6cc56287de14f6e169da6874dd5fb442c37f6d18ac40af97fe92d5d20c4e04095918cc69f7306dc474ae33886492f8c8e746edd7d96c776fa76aa9ae243f6acd93d982ddbeaeb301ff64f25cda9eaa7afd138b840b52f183bc3a7f16f88526e9b2b6913335b6cb72913c1b59eb83eddba71f28e793f3ededf994cb6a80b47c1c59edf73feb0b831a342dc0c8a26c0c4b8132db0eb3656e9702bcd924b74e6853ea0e35d01d5a1da103f359fcbcce084ad603f61343d6f6a2b47bb131c6e1391551e38712132ddd0ac842ee4b2c4a605bba68aa56c18b6e785163194e924a84804c7b1ab5b201fb9684ff20c767a1fd19b10b3d6bf8658ddbc1364ac761784e3fce85d250fed72c4aa6fdb45593973c4b020c31b3c366d17223f64233b2ecf8c37cba64631c56aa066f2257718bb499e17ca903bddb35ebed11e2720cb1b5a027b4c0eae4519ab2bbaea2bd3dd643d30cd441b364762dbcaf3e598b2996ccf2aecd177dc4e0dd5603f9d71be5b9d11c76a7adc012963b31fadd07c1ae825099ebbeeff10b1b9e9bf5ba542d2be91d145332606d1ea4847ebe55a891746b4e09f7b6ea1cedb974bb6ed649a35318945767f17eae0f212e44176bca8797129723bfc9e6ac402981b65a1e72aab2efac6c592306dee6070f2e76d199c722753ddfb74b772fc530ff403fd85bb9f17651d70a871d395ee2b047fecf254877feb7f6bdac80af1e5892510b0b559fee90b6c7b5b287498c0922bd911d723856a0ecedceb63d8b93c014e1bd7eb9491fad5e986eda95ca1eba6ff3a39f93ab75334b38e850694685d9d323bcd70370532ce87f00f0d38678325d174712ef9c9f64b0a03fc236c991797d0f167996ce58895b1fc62e2a302a15907544a6dfa92833cdea659f6417c742f40fa699008036318b1971329cc7ca67930e33606402515b53eaa83cc54e2bcd67867e1357f56c1315b9070b327e1840bb08679955509bd83f7e1a2e92b1e3b5fc90e5537714732af072dadef2de4034077fbfeed9ed3b676bc8b6372a46c9f18a1962949262698e718950ea1685d33a9f8642eca180614feeb5b108bcfd46a8f0615be219120545f702dcb192a3f2f54f450586bc8766dfd9dc19a949a95cb9d62d85e40bcd64911138a83760845d2f3e9d33209807db5a21dee4d9e54b8c20a5f67892b06b13272054f9e39dd54efb57bee4e321cae1db7b283d86d324755f6e97be34d7352036910ba24c03fa9cf84aff230cbbca8145721e8c70929d4c70cf96d4670e5124dcddbf571e9311f30d2ee111990b1d41fbd1a2ce422d8129651a4eace61da1d8952e60cef9f029b4497a06c4c79e819c3905c4761e89076ee5a6bf023eb54b1a7ce4b615c7eea5482ee7e77ded251596d3c4ec21cacffdf9eed00d3a141bf5d6488bbefb269438eb4e60487f1b2616cdb79869304af7043f47e60e6c8c900498f59756a72fb6bc1e74fd07cd1679b0349c69b6ccc36dd2cd74bba95388216f77133c22ff20783fb06bd84cff2fbb3e32a260c8741f5725c54c6f7e659eda998e3d312c55f203343e4c68a4c5adeb2859db2dddf17a82bc2de14c52ba4a83106ba89330dac4c05f3d981d0fa5307d2de60d146fb1dfc7a7bab1d9f86dec55db77dfedd0f9224ab157ac746326781283847015c61fc06528ebd620ebc815196e4a3a6ad293a1c895c3767c61518c72bd0699355e32f5c38193215880806b85f71ae748dbe2a04ea31cff5aeb65e1e7dd8ee8575a5579dfb5f84c2e9e707e6d5b4721ed461ece928a658ced136dae166d512a3323ebc064bff652c4de6bd7108239d3709f7472c91f54ea504d97c1ed2bdcd873b17539d19a620b15d26b343afceb251f4485cf91be16d4e47b883d755cc6d2dbf25a196742214a854ad1918214bca42bc1b28a952a5054c2b26b172f4bd96ce2e22512ef59ed2eadb496debfe06e50ddfa9f941e85b17d2651d976c1e9f03574bcea247e2d0903c5a2e3b0fdd1bef735bc97126424d6f94f1d18f5e4313599bec2bade7dc6d1fda2fd05d6d325aaa4c50fa9ad95a414338515cee9e00b39c552a86966d64df5320e2a9638c1a2e3a794d79a46e05079ee108b9d4c26174c9d05d55c854d9b23dc963f5264331cba5c97be78e89bd2c3172d40277b9bbf4728c4bb884c5fbbef8fdb89f97b4ffef27f9af22ddbea2909706989960a58accb4f97f6be0d7a27f4fba9628c5d0d07a8c19e897963ef29b959f283cf8049d1c0a011310f45cd8b9cf6d22e8f94334ba54281d718fe5252808182f45607b84fdc551d6c0ca98ec47246533f4efb1bb83170b39a11dbdf0dceda366adb009ef48cd807aabdfd76ebfa7d1689ea0ad638fd2c5873dd62542e98e150c4c29eef20842fbfb9dbfe4829614667a4bbf38f727e66c614c4ce12ac034a92689ae10f3d00c872b67a3c69c39560964f50420a4ba3dcf62d6d39c603e60231b36b0356a7a0080756b761feab0253e68e67f317adf72a7dfef797469547a84fccb764e7b78cd012a77ede0177a8e402e56540e9e5e9b20b686d50d213226be12d067c28469827ddcdf7e15d8d4a6200e9e52bf17384041b8aa70ca097275fe70f44643a8beff01be4035929d765d7e863023cdb90f3bde63fff91d023187a0a3fcc96018475d8767d978dbe58a5555feecb6699e2234cf481ded5b8d3bec293faf12c16f890a52238d3e3664118d4bc919f05fd7a299c2c9d6a5fab9dc0eee8aec75f57ad0bc91eb29d34a485691e2c225654b3abd4afe162346118a87e33f8c2cbdb3f68df5148eb27b427ffdab10f29475ad17d8ddec3025a98b22b30bbc27fb6123bfa711889e051db56cf07aba5256458e5059ee5b4ae791b96c5a45170ff845c244d37a5fd28dd756e1892fa9545a38a9a925c41dfcc55147215bb2e09c1df454c23ab3c5a64cf339ed75530967765c81f0bee2d0f7804e841d76b6aed9e361b433e0089e9681fe497d473cd9f39737770f4e887ae84cc9f2becfe93180fee9dec22b9b8697e7c5f5dcfa2e3f3f3bb0c94f40400caa45971a6c7c90eeee5e924035521bb33a28f590dbddf58251047fe33bfba2f3e9b9cb31d2f9b0b688c21e8e7be2a3add2bd337eb4884931d78430aaafcff9da037b0ef972506fc12723a1ea49ff0e20bfe1bc3359afed16337bdd79370d992fb5ac6f6c1f6da011bda39d64f6e559589810fc184bc985745ac7a9c5b3926d98c6742b609efb0586077c4537969255eaf90acc134f80c1c062cf313de92a40dbac1f6e011716335632036a2c79b6adf0859f8ced80776517d6a13e990eff0a169c79e953784e986f3f18abc7fcf13f583b6e9f383543b41f955815dc93a687e88d10f5689ef7cc38a469e5b2f0dc4fb3f2d62de06c4d0f1787d50d873c5d4c510d6cf5e62702b84110a7fd3432d1bd1cec285b6fef129c9f8c411ea9797694d45dabf3ff0251db6577984728a23bfd60331f4b69a7307737e6ab90d79db6c107adccdb35af3b28ed2da1e449f10d5760b37b56f484c7a56eadde8c3f0c7dce573eac3d4a6f46d82dc6463d4c02425fbcbcf4e3d47dadb96a2bae03750690a7232f3d0bb5f45ab80f61d0bd2495702359c19abe73b730ddc8ac6cc87ff4e0916b0d401c579b33bf4eec208227cadf3fcf8e2761eefec22625eaa8b1820557a2de152f2baea18c348c1337c5fd402088a81bba2f5544f208ebc684a6fc20eb4c8a40cc270c937a5ac417db1b59b00e0659756466d7f1cda12092dd138fd8b5fa4c79431d4f3c1f2dac347c29cfc2ef23f39e01a6e35c7b0c20b76d29eef315a2b49d3aac78e5a6c07682eef5528982142a555cf571463f2acefcae1056762ab6e6adfd1109399062f836d08a5203393d58ef72c1b730dd7b6728dd6a3389321a01a54a96c27115b594cd401fc1922831097f333d182764c92a3146297948a53549758825b5cd2e63412c493c7f57aa3d79e062bcbf3d3e8b8f94917b6df48f4a27d4d09330c427f105ec4d56117f0301504cca747506894af3caca99b934640ef57bfe93805b9fd4e4c9880da8675eaf3bee1247777827018e54bdfd788faeb84e1c471ace504a7910670bbc3c8e75d546940ce0768c543bbd9348180484f0e9b279e3667ded2423d324fd098d4585a57feb05319ec7a137034e721538cecc06bf6c87541f31f4ea4b38623425df336ddd233d72e1bc9ac556cf97ca88f9116aa23b4bc493691277ce38e2600033c6dea8bcdade7e5648a8eacd4358c6b04f8b4bc5e459e9b9e91e8c0f26129c05474f062a00ba85d50db442d6ecea983726fc8fb47332406a03ca40c2796eccb545056968c1f2cde6446549ecbc21da3618b1eb46bc986bfb652bb93178db6d7b4beffda22ac417f86a959fe3788c628e4d5a3ffd8a14e3bbdeb6103907179cfb9420cbea5fd2112e1409ff92ccfd13380e82df79fac73bd23137635f12e293e8b3077ae9dabbd2fe2e13257c2f8ab11d20e1e386b0637caa4ca227dd1670e3e4c4bf78dadad16540144a0e7cd67f0ee5e864bbaf17b4d22bfb2e4a3b13705658577422799cdd991239e0b0d0de5e55112cbeadaa4a4a065c165edc31eaf17a96ff6be007a123b5600a3309ac4c7138b52d1257ecda2a1a9d84f5dd66f7121ab6d342294c6cc7b564c609d5f9381406b2a7fcec40d779a02b95843731cb3e741b4c5152eba9e4797f95d2c7eed2d0e5f66b2bb36679ca2b5e483d9c19760f24137471a05d0773c84876cd54969bab649f954d954838ed352cadfe63df02391f8bc057da723aa33c478508e85663815f4131dd965410159f3f91563549cf4b55bd7688624135b905795dc0cbf5da7fd62232dadb4b7c0b7ff9ef33ff73d3400ab41e702eadfa266905f4406cf277d0a447f3b93b766e5b2d0414f81396897ee66d942353cef68f2df92d8a8208388b000411a69169946a2ee38b4a47bc8c1b", 0x1000}, {&(0x7f0000002000)="7252ba8bc1d75a97d4f12549ef766b80b2e4c36d1f11d984de5807b5d9b7be5badfe5d1d79f1bcd48743507c2fe1e6bc93afb1a18bf8aa4f1141382e5df43a20851f2ab3b8353af0ec54cffc54378586ee8c9eaf3d4d8cf6640aa94fd203acd514e78e44f7217406320ae0723dbf749ebf3685b7f12db16702dc7aa938b1abef9c789602bec343e1f0d54fca9b7146884f0c4da57a0d4011c5cb61c52f36a5469a455746ddfbe4a5db43f26ba9826d6f61ce1ed9a0b46ba516f243d048fa2211307d06109403e98f8e16770036cf580f5d6db8c2dde148bc0579740913cc3868bea39a497f8942", 0xe7}, {&(0x7f0000002100)="e0949db7f4ef749c4d9532615e3090e995eacb74b40b8c8f6b2566685cf6b1c677473e380dabcaf9031d076b23fcbfb108952dd43a2af98cbf95d25f2524925a2f86c4be9f22175c1612f0a5ecea1fc256c0d287395bde0c4b8f3a9d328a00e12096836f20f8ce2f6974ba144c98cd1d3fc176eb37f066f52ec2659d57c00010571b878831bf2d01c2b85f", 0x8b}, {&(0x7f00000021c0)="e39901f38cf493b7c80ba34d6d55f3ba3dfc148fb0eb59a7dcdd4f0fa3ea59c81921a08e3eb6be97db3bb6321ef9a212d8b49d32aa26a2d0fa01bf22723c432241f5ba018fbcf6de10bfd692a1e92b38f9ad8def34e79070a778b6f446c1b1e8f26e19b11fc190afe9d0a75bff256751a45bddf56954fd16fe13b3afe9be1b98435e5a2ce95a2d7c52b5d1069bacbf1ffb85c414c0eda68fe7d70c77a4605890f52499b751f3567ed978f329870d606369b129d83a74c7f4656ffa1ac2d6741e3728e3ad6d3253c375e552a53fe9d2776d27fbe2c1d698dfec48a1b6bd1e297cd2bc6e3b844d56934a88ba2a4b14ac208af2672f37e26be9d429624dd33f22361807c631d346e4debd8b1e3baa1b6a2dac341531d473b126667aad2f4617a504fc752b7283ea0bdec706910745ddda451ce81df4c6c33cac2479ebf380aaf0061d8203703656fbb6b88845b743e642752c2267f7c1045bb5159676c7bf53502e157457c6a0e0d62ce83d7b74b789a2a2bd4238b4236549dfbb377522683f362c1c06200d2c0e9b985b565a91aacd64f6db36c904ae56ca8f4b718e71bb0651d5113f560c0fe3ffd70ceee7a07b8df0e7a76d24cf87a902b4eb913ad10bd0c965dddeba3ca473dc8698ffb879a405e489a917bd80656122c75fc27b85079e1a63600212c75675e2246e586d081e1a14da434903ad7392f9218c4ea6c2966f116bd292889403720c647693bf35adf963b490478e8af07268fb3ad49e0ee096058af003f5e7b80dce9032b209e7cd79a2a5e61c0caee2a4e2d90d307b4b1c9176a8625ba994dffdb4a3a1523c5091d86af016749e8bf9671629b77a40a44de950260e2e15f3ec54bb23c1b8e4e31db20e8d2b2a0307b1f988b20a374d1e5e1052bbc35dd0766be1fd241b7269931b0b510fe0ba38cb33cedf95991b78cccdedefa5eb953759c1a435db4e2104305ae73e4b839f2e8af4a053a74fc3e22d612c75e355c01e44911758ec8493ed3af672216dc1bbd42490a189da994baea138bf9d27e70444a6db92de757d31c91631c6bd15e92417d3b5fd04cafc7bcf126877aa0affdd0154a0ca10d39980782ecff7d37b1cfb6ac209ca6771d470ae820b21189bdd74f5fbbd4336dde08d7adef8b20d73b2ebe5b6be51edbbd8d368c705f6f67461555790fc38372d61c85bc65fc9b2019bca8ae1764be788cc04a0108fb30e83f8dca571d1dd45d3b4d6b5eac0b716fc5b245fe9669d7a229e6ee72bfbc2b70a02569d87f1bcd74fb99b3f93fc39566ea4e152963aa88e219e0665e93d0b116e92e32a6e222f85971edfd3bcb310631cc7528df95d1c5a4b71c5e9c384ab1554c0db65cf6bcbcd3be5ca379ab704cf33258d7542aadaf71423f9b655792d9f3e73e2c800322c39e530f4dd596f9cb6795db108f2fda6222ae60ca3851c839bce543f3127b6b5d411b8e41b4b9f585e517824ad89fda37ffc65d45284fe985cb159d7370f36e551ff718dea5387f0abc2b39af55b5ddf0b96a05b12817c12166e75c4b33c2ca1765b1346dc87e21ad13abd6402cb22f65ac47b078af23571348494235e8e5483b05e1c8c3a4e09f3679ecb866795e439d87fa852e59581c5410600500b18f7352dacd4a756b4a49d689e260f047f2bd495e58034e65b2473e2b8ca4e763294ee4b2f74db56f6c2a8fc28e3316e23d538583d125e8420a50337724bdd46794d19401c0356a906769238417ec4e217c9f37dd77232e996aa0ac09390f604e023a7ff12868d8dd161ca3d80f8acadfccf11c2d301a9c0ab465f42a10e950ac9973381714ceb3dfe6c5acf1f6d83ca25c70a5c24d49b127d78153cb31297683ea0568f038f520e49f720061d6c2dfbc8b0bfc0eb4b202f2213251792c6f741d1ee3c5489c76a7143473928ff938c97683a316c72d6e35844fe8c25d017bfb8a29532c9f444f43559e909ec507fffe66759bbf584cecb0e4570228e4513c08124d91a17b90c2230de488f38f3f2e606ca69e4f006803d6e5cab9b11c9f37e9f01b86301f7addcfa4624d6747d82552fed7d7c77183415ebe6a9d02bf40588a240001b0c0d13c74944d8bc009296abc6ad16c73bde8aeb81a5c92e9c3b971c4cd4d4a1423772b1c5e5d000796474e38c3f89b1c99e19ec76cbdd7339d064221ad02c54a9e0b3d46573bbaa7c0af49c8e68e6f17ef33735aec1a982af8ffad35235f0944541ef0c35c9824748614187d91219986bb2fd45a88877965e7aff352d2ef4eaa5f1df28799ebfca9e7544cfe93e7810663b956aef8ec6e3087d3fc4b44e1d0f422456f22ff89d5dde9ef5c448b646142f2fc8977834f245ba4528b70e1c4daabb6612b91f96361be441e0081cdf04e5b0b7151ea3acf05efeaeb8a523326be040044c5cc450777cd5cf2ef549c2f712277dff37ce30b7595b09a1e1e384702dda93a46e63f4da116ff67b6d1e8779d99ab460f95dcb0a1ac300e2a1c3bee243741086e1a0254231ccc34a9b9b77a86a86c36b590dea58b2b943b2eb15212dd4e13d796fe23973da9b1c01d346f0098591bb3f37f40b77f31cc7e73b15db81270d8249e03b2554e2c32e54ddeb1b286e6cb874d8a3e0d442dad9574668fdba19abf13f8b170952b4175626ad31fc40b7bea69a9fd4c0fe016bad0a4e988172fa81b226d3695ce89602760009e97b587a3317bf6870a70ed910f1355e03e80cd518be82d7e06f54f8632772247fbfdc0e95068741dc32bb4daeaa3821da0a1b586f78228762eae6195e913b500e5d63d75217e8a5a1c3dd9d4dec8ee798def5c0066c713591b83aa3aadb0887493c4066b2a95fb3b4d316685682b7fc92f8a19093a7d790d7a65b430eb6e7e96b2da3d349f54df43d4cc054fbc041ea9c54dbaa61349f669f32c1cd2f3c98e2d9939da4ed165264568c099fcea1a5a413470b1592a582887bcdc87785b273aee82ca3bd05986ecb860554f01e5a4aa0b1bfbc6ba5df6a3abe73e14da64fbde485f1af83b7c4ddb656f57c8c857b79417d5d51e997717a00138f6a7a0b604999b77f73f0edb9e7d59413de9825fd49533eda109ebc2324f432f2ee563b52780f701c81778816042a5f5706b273635ca05274a981ff742f01df9bfc59c3f2792260521a819404e3111529cdee969e55ef9e2c6ca6fca90a179d3f20bc3b42b1d69ddef324f0cf04ea6983fdf4e8d29217a496e6fbcdc4627e49eaeefaa9368bd3dc96de5248a8c57a3dc80459f1c65b0a1bd70de3d896521245b5c9952429b68aac94660327616b6d97568c1dd6d11e7c664ac077389185b659309e79954a0fc42b8c21a6caf29fe05486e78b786854376cca6f18cd8b790ef0c83621b469f6e6175a29432e1138782417269a90593ad205b3c691c13134c0f0cacf18a9874ebf3286e2e3a4d6a86f0c8b838642f39b5e7538a68072379a23fbe8f5a2c679da346f67487d5630fcc7d40634c8309299e00c2b352d69c8a11c34b20c1e0999108e9d922b68289a1403c1de9d6806cd4112a3d24c9c34bad74c6efebfd06497a73f6bf6a6a0bebb127abeab1931ad70a0b4bfa9b40be372a7e45abe7b611b5bf712d681a57c0254053fb104389ca5e25865ddf7ef4dca6a2f286524e263ea2c166a687faf60fbc13bf4d623521682b3104e69a1556d6f9996fc25d1fa36f8df5967b8a7b0732097453a897099f5fb4d47a1bccd1db44da1837686afc9f5415f51c0640f9eb90d1b80b75bb11d31e1f553933af3048b77ce77d7abe6e556d7293b1365a357b3a714968b4cf9699eba4f2b62073d9c7588275f243dd4020839b6a1838a19f5ac19e00220294fd914fd10198cf42cd3ee5cc16aba3bb554db17ed771ed2bbaeedf1477ff712b13f929982db8eecee7cc3d9a29b28932c59a9c94f67f89116c33a551c906c5f4d96e5ceaeeae6197b9463c3e2b0456dad025484267f11d041213c21e89d073c6170efc478fb49878003a4f03156e5daae5dfb8f1f04b8b7f8788351ae7c445133137801401cb496d426ab0a05d9d933f6df8d593389ad2d5cc6453d7384007f23c6e3098af6fefe9fa27d3783168e24776b432b872023379b20665c131cc13f8b08bbbc9d6a6d9772f37750602ec66c2447415b97800027a1f62cece33c0f2e3703bc27f5a6a25e261b90452ccaf0bceb73ffe9c91356ee28e466682708b4aa245f7b708ded0e4efa31c33793eff3e960708d7df3f731e61e6f02ff94398fd26250e43452c07dd1984c701e26086fc1e2cae819e8420d14cfd120998212a7b316b39b9c64aedc3138973571ef408b5555ef5c5ea8e35a4d62f8eb93c0aa854ca17533e71d078c3d066e8b05c205d1b9497f7988d40699dcb2e82b3f053517c4781764c9d031a88b0494ad3e29faea4310037710a98087ab3d75e4671d8098854d278dfefae74957e168a9b7e458debefb85e23e009ccc7fa31ca4d23dc2c7925d14112584edfe1a27e8ead18c3b81e621eff3e8d307d2ec47d3ee388b5404e9be3bc9f46be6e9a27c0f65c582df62e1286dfb9b805d5f490e993c278a66536155e06f4c306ad0240cdc3108ec9e8e63741aa1dbc24e521dedb43edfb84b4ba506915183d1873124eba9e7d0776ce476b8b8280e6f662520294cf484b5b155692a818437444a494ab0445a48f9a74f7bb61fec523de839f527e4b116741069da2e6aec6d7835d786205055acfe7be63ba07d8d215f3cd948ebf360d384ba1783f8db18fc6362f3875876c1ceda2dc8441d21e0b430c9c3bf4e3eeac9ce965a37d9f6d126fa8f9a0340a039c61762ed4588a2abb47e0cb88b791e8a9154eb95402eea495a3a49234c83480dd628239d76f058a3302a3ca4f693aa3a994dd18cb5f99af03404889ea713f2def3116588fac94c94fd7e19134668c24cd949835f85a70c7686917f5cfeec8bd1bf3e198bcbcc980bf66d53a26b4472394424adde196c2fd73e452fe6c51164049f0d7eb76fec48ede903d1aed21b1796ca0c869a1b7974e01db412c44c3cb991214ea56d40dc296a4e4b9b39712e17ad8250ad5274db318f7a3f4c373427ea1c440668f9c94bbce7e064aa1cf5f6224fe9e58e4d488c85b1e0e1dc0d72945b1286dbf37bfa4de1f4013e7177c259790f44e081e0def7b9bd45620dfafc5783b0da095aebaf2736547a0955dcfcfb2b34f9719daf8fbbbb89c9bb0a49bef904234d9626a5debfa8731e61d326d635af87120b40e7158d54227462d3c7a88b40e9639338149db971aa99b48c33dc1cf759fb74a9ead522574a8fe5c870f5af1179313228415b0b894a76235efe001b915e0f8206ace9dfdebd90e36b6dc9542f43c680a694ff086017aa46171c2d98ff9c90f3116876aab63d58cfec732be181dd19aa6e68c3897510617b135cd3e578e11c88a106c3204b769ab91cbf454e601f9921de11c68e7c0c52c8916924e866c9d48f510dff5b9d3fb96796cf079fddf3c755c9040601b493214706fd37bf89e0c737305da67bd219384b87bb2807b4c109eb30fc96fde40bfb2e9c0321e0089a24ea72351a1a13a4cdb0e1dbfd4c50d7327458d94f0c49dce905955235b259d7a808aeaa9eb9d8a4d781545b7f4cf77d6140d48404376e08533d269316e7760ab7faed70ed938a97a5294e8f737c56be715e61ee7378ce0bc99b9e53e9143aa3ce1dc768704655e485c047ab8248286886321f7c9d41830ee8cd886201748ec891970d5b8d99f9941c32e5be2f1f77f3aef57c0c32c33fb3bc71e333dced2e5a269c09f40fcffffba2e0", 0x1000}], 0x4}}, {{&(0x7f0000003200)={0xa, 0x4e24, 0x1, @local, 0x200}, 0x1c, &(0x7f0000005440)=[{&(0x7f0000003240)="7df665bb5181ae567a88ed05d75347c1fa484886428a1668ea1f35a7ff4df0e57e7ae2d148340c5239f050aa2e6da57039d0e62164d484f8ea3fb53712b63da4d8c88d5e4a4865d02b0cfd4da18de33c80d601f30877ccd6d1209c3abdc1c2a065d4bd29bdc51208e791c4dfb39c84a8f87fadd9e248f917b5f85c3b5ee321b27cbdf5b46b603822ea9997501311e59d9724e805a1fa064a8b9fb0502f0f0df0aa2e0197c7d3e45c4cee38883af5c2a52818b9744559e71ee031d03a73c71a350af164f871bb9531ee415df25ccaacab14fe439349a865c1db2679736757e0522e890ece9884e2d9e2ee675f8d4fdd65ff991a761f00dfd7", 0xf8}, {&(0x7f0000003340)="c7ba26e2fd4cb119bc254f5c25dc33fc4d25eadbff01de0e0f02699edbef7a257df9e9bbbe6cb085a0562ffe414a6bfc4e516eb5b46ef48a434d5c90df38b1d26d32cf746891854680da8194ca464b5f4e58f1c8043254d4e026dcb18ab991d6d1ddf39ae3971d16bb5a9f47d376ae13639087c6c8a21968df71c4c24ac771718c17e3f92a23ca98e6256904211f69ba66278ee0b7b060c8f0e207645903f334bb3a8647ffb7abc8b6b582d50497dc76e3c587e85dfcf30beb9c5c261a", 0xbd}, {&(0x7f0000003400)="e190e89f605f00bd91f69ab4666bd7ae317b474c1e6c78fab735984eab18644c15ec4a3a993e9f96e9bd9837d93d32e5ead1c84653cafd94d76a36d7ccee1d91367c55b74dab557da03766ab6120564244fadaa27f3fcfc00189aeb15d9a6afc8de532466999fb732a42e728cdfb676ed47375f5ee81ac6342123963db2f79c622ea5b63c5836ef02c24246fd93ef0eaf2a3463c48c4bdc7296f36099882cef0a39558e946fa98102661d5332c13a696ad2524b4942bdfa45946710831f3ab5613b1806c454384ebb0137d17fc3b181b9ff9bd5433fa09ebfdbba962ef61d4a1cd1c6432b8803ebf133847d88f660041d5491ec2c7d6b99fd12e8259a724331633a587284d31ba1cc9b0af650eabd386797d149097e00dda9fd9efb2695fc7120c159cfac8dc612c66186dbb3944afb55f27d1e8d2be27a538e3c3e2c278ee8215fadfb9f082cc945cab4f9324bb27a5cc2b3ffae06564c9c2f91e848c93f190b60b140fad71a3c732bd1edfa09ff85c585791dd5a68dd8fbcd3a8a804f474dad52c53c724778f7bfcc6f41345c997d95372c7c0635e9ee40f2d03fc9cb3fbb7a418854488f9de79ed78cb5ce6cc7e1c158e7157eb2409a8129000b022a738248f96e1e22145b180fa8844b2539deff9a7826735db39cec8380da35d50ae0d854e6f6e26e5b8418ac67e4f74a93ac122fabee469650d541ee6efead191203cd21ec9e30159e5ec75f9707555cf563897a7ba6b50cb679a998e8e73f92c7f1c609d5d763fb46279f4ff09dd11d1d3192e399bf554ab29a5520f6e7dc7be2d796bf0e66ed1a07d86de5bbbb9c77a9eefd2479ce23cb5e38eb731b8d169a3cb810835b4cba05ae023f6bf5dae84ef6342220cfe57337c266b5afdca0bc9f705eb0d48a2922b1cfafec64c540b81e636d95fab9e72a791210c67e5339ec101635bcd56a929146bb8db3e1b76d9651df2f63fb27fbb048358badf039b2a28f650c90c0147d4daf0e79734eafab8cdab89150c97058ccf64cda9f39234eb20e2cc6312e303e52cb8492afaec5d7c7c9c4098ff2895abb6eaa3806af9fc88e8f6772a64b852a82a64fd1d7b80bb2217e438b7eaf8437e782d398946bb6a78a7487bac4b482727b19450524bad0ca57034449abc5a5bb2aa39be4e76afa96ccc73a574e64a9ecdadd4eae8b88b56460c2ebf6e4fb40ea835a8591d580cdd1560ebaeb52378cfc52f6e70005767e11de0147751fb854a45628067bbc3cfae0f18e3b7764e38a8f92010dd52d2726b85440ea7ba4b341a8a65e9449f37a90f37cefad5b9ee598b57632a5db6681234c7c37c290827f895a3097ad0819e3739fa65f80e83ac4bc78f68752620ec9c227bcbc7efad226492252f42ce76540792a3bdc22064eb39ebccb046f448a6ba874731e979a5b784abc3488b572e3a28ef5a193097f83d18aed933b0c81d6de3c4596eb5759440f32aaa934f6da62f0b806c45bf1d5c4781d9269fbedfb983cfff1dcb44387454ceb05d02eb62188b585169f51e42cae92bc88197637b4914f5fe1e150734c86ea42f0732ba2909cfa77628d82d6b928d189810021a95ce16c98132e4b624455cdde4209196080a8fe5dbda4a5f31d7ef58fc89fac5f95bbbccadc0e0ebfde5b0ea2dfca449b5306240ef3343718812aa31a3ec2ae37022faf263d7b26de9a8008b3bc70d47b814e821e49d6b8efd313edcc435feea9dc640fbe5beeac93a3190aaa2fa0b8ee1801eb6a8eb73d0b2fcf6c5f7be41294194c2a9131489f39b7f1089ab925ed2535b2dda584849954546fb60aebf2f0067954ab00155770c26d7abbcd3b25c8c6aa820090eed50bf7e819587c7049b7ae7a1902ac44e18cfd42af34c0ea98a6d67d34cb7bb1f0098a74a5857b3dc154c155d04e8d4cd0536b741f37fd67f6b3ae3e643117de7148d329981c14d5a6c30d7a4898aca8e0d10e5568ad096113d3ca4db898141297b6063708e840595b4e1e954b79b8b4b0f8ffaed314a0ba7c3b913700a65397a338660e4304eb7dad08f326dbbb987fc6203e0d7c7b9dad8948400c82ed707c3c360220c20d0516b367e48107cc93ec4b2e2dff9b7892226b665a66096ef0afea2a13f896938a3c30be8e1b3f3893bedfa1a8bfe31114f5ca785c9783ed2e1e10e5fac8ad89be10fbcbb89bb6e62a11bc5729d6c3f9104b83aff8205cd84ab674d8690d223724b3b2883ad57454284d9b53f6ca0e628adf64669f13ba923cd81103833bf63de19bef07eb50d8dd4feec09793d5fce031bfb686c7a5085f442edfe9d2ef459a1580a2f345751c9086aeb9c8a22728cab6ce0aad9d607af0af70f4d2b6c2cd764f636cf1cf527aa899f0d6db4d4b83e3ee1078017fa74ac080580695159e418facba20a3e02286509834f81674ecf896056ea94fa98021531e1602acd956df05fc9ddf329531792b00aec7ae7a67fea04f0e3023f88d7f557a38924e3b9ce52b3121cecafc0612d43afebe99f87ef739634480dc01843f4e3bd7f5e822d362fd4fd7a3f18a7c46aa1a9934ef807ad161463ef5675d071dc52ff0caac2a3dc4903c702a2d8c6c300b342953088b8207dc66d56c90903cb0eed2ef8a9673b6365dde6ee2088fdb4e5bc385775acb843490f71779143dae3e7e61d48824ec039bf7cc34afccf02c4008749bafa2475dab12c54b5979adc785ad7f5baab1b0cb7fa53b0075b917b673174580fdf22556079e50d81a336f042de65a99fa614fe2c0ff11bf7539674fe5d81d60bb720934d63ef7937dded244539e2f098ddb301d090d6bbc21a6c9e218de0bfef700f23203d867007199704dd99d3d9864cde573503ef99a5c4eb933f7381a3b3181fd0eb195d5ad6d5ff1fcbc767bde3714983b73e2b7105d7402ed4fc7ce01f833f59f4ef09ad91ff3fe1d69b72656df47104287d677f996d9f854c3a1b17b1ef89bc5a656b2a7d6749395cb20bb1e4376f1221582189783faabdf75549da281235656d74b5398e74e9003b87c69d47b6223791df2b1e2d321aff48468902e2ba1fbbffcbc334fd497ed4cf48abee97418af7994cfc687c3844f122bbdfb1aaf918bbf2e0a74af7714dcd248bfef5684812e2b93e92a4fd0059447e16dd52b319e373cc302bed8464cee7faf29f6fd5a8e615edefbbf9a86c353eee4ea4ce8d22ff258d7bfcb7659cf0734e2b84f00d1add3cfb8434bd40230fbfaf6738e1463a20969573e5f71abcbe603f4878010d047103e9e6e4006414d82785d44ab63f4bc0190c18951505ff70424406dedc0aed19ac4953b8e6a1ba53577386a65edf549af0ee2c6e4a81d8f5767aa4dea43c95a8570e0e7887162d28ab669f62f69a83be5f86b88d6a963858eeb2db12c939ab3f6b9541713079a61cad6bdd3059ec3fbecf4da5421a75a2038bdc9b4596aca55efc040292c6d7d51041d5828a411f20f9aa755fb95331863c0fff0fe9d75b12357b21d0c7966b7f6e01b2e0c3556a1501562bc8af093641da3d2b643425d401d91d3349f375e9536692e7813c5a69f91ec0c61a4371904349ae99c6435f79a82fb720e12b0412f71758a5c28494cf2f703eb6e092b5607714519d1411d8875b394da6878fdd4bb130bdbdc11f83b92d98bf66d95f1cdd52391e6167c5e293f34c90a9e2d35830729e91cf24c5cd4d1a1039e9d0c4bf0038bf31f046602daabeaccff4dec6130f6d6b35e9fb4791bc5b6594706b694e5d154df372cd5d4567e407769db1299342d6672dea2ebb324c7805335a9fee0330b987f249dd429d4d5d43e64db525db8815274fc6f7b3f7853561b9616f2d102d210f6cf4e1346f788f321a0296e93193dc43346449ea12ea9078a3f0d709550a5480aa88787e0a0c8d4c923557e7dd3c50402383b56716e0f4bdd8e9f77ef0bf682836da9370d39bf3f1b70cc382c0a9d5346b697977ee5992642297c72b227560c43be3a4b8eddde211297d48858fb5b9bb92687a66c4f26bcf067dd174134119982ab05d30f42c9f4385ae22d208891390276c053653059319da0ca69a39f421e6ecbd5314d3e82b570bbe1e8d05e30ff76e982928a768cb544243dad5e1cfcecdab7e62fd54962e5e4f7a556bce62353dd2c8bd849d9028fe148f33c72f93b219ab33c90bc61f65254f8713868fbd8310fa79f03b22f2e84d3213bbffa7b47d8ba2f4a49cef8c2e8d6fc8d9f7433e56cc750c56d7cff337eeb8d390007f6aeccaf50f771224bd2afb0ee4610783fb2702eb87b9d26a49cb5c993877715d803fdc1cbec94b13bb3ff0c1db7bbd55c6cfaafdb965d490f3949d6617c516b446e6e4347b5351d3018e0d85f25e5d97a2e462761ca6fa2302de383df22d52bbd190f7e7f69ca03bc6cb89034f70465e67a8ac80acecb79aee4a5f09a0f293634bd2c22aeeb064b0fb354a39bb38aaad300c798baf41e8e302586a0dec50d3ee7cf7619582aa02b003a119fac9d3b72192008b8a6d9de5b6822eeec428da7249209c539f3a8ea01791ddb8a7f7628bf7a13144e8cff89659e1309f0f1f4cd286c87db50041da1de6e09eb84210d9cf0f1420b3759e3c35594af2c2fedb312754598b4a3e436a9c3fb3faae8c6afc234256d119b524cbaa5f9dc25e55add364373c821a4cc9c9571fc88cb1aa4e5d53fa72a88fa31472c0d0491cf95231098ceb76d53b9fdeedd44c1f26afa0fc00f7aa918a1a497d65653b637a32fe5cfcb468eb228eeaef9bb106b04e716d94d90a8f0c11b37a9c5eb3812f83a472face6105c78fde86131dc0538ddf026bcc22b33d38803e2ba3659ebe425f57e201e75b1f0df4485c2f682ec5c51481e75672e089a8378aca10e96ef41bd2a797cb3ee7f83b2b2a3be89d921155c5b49ce7fbe2fb5db1c4bfa470e1a5abaab7c54d919c8019871e013be3cc7f85383c5d2dad51b695c0eb4e296a8185b234e0fc1ce6ba9edde2cc31d79c90cc88a9423d5446008a8a965e12fcc7072d67b10e3dbf322cb6f46f7c4c7e3092cac5c211e669336322bb4c7d933d05ec4a0815bec5654e040877fe4b8447b1614dffd78cc12b294cca42434b47f6b6bcc7840ddf8230fd8061965c7f2ac1908a23665e881660647faf8fee0d16a4c106fbd20de8a9e506378bcfe92e7243e7726ec0204580eeb82328d4728c97ece7adb3e82b461931e3d7f9231db4b97ef538a04c50e549415e7efd036f6b64032a8b31a3fe517314a82db1199f1e8167b02535bfa5d6f707129d4ea1e7abd727a038805cb5cf548f7a5f2ecd1fdbb2700921dddbf5e472fdf83ecc8c9213ce4324c7ced7ef4628ae019bceaef6556ef82b1a8427cb2faacfd3074d9e686c39cfbd5313fb766e6a17c6902026dd3599a19cdf980a205905a04a2acde838584c45158a73fa64d231323bb92ca0714cbe6d5aebcce8c058cf9e345d19282b74eac07d0aac40c2d667f3537ea5568c828fb292f570c232994a7f543fd633b011a4f56cf027e6c4b435eac59cbb32a93bd6e0fce1ce8884b737063900a9345f8c42e404b8bc8bba7277897aa39afe10116b85c134ccda98cc4219288e68833e67311fe4319c1a7f4e4145f5c76a1f925a5b00b5313c80156ce54d1f3739c3cd19badf1a3254b86317d0c5691865dd7f6ebef4e7f3c6605e219c7742b798ae0b4b7f180fc7633d32d7e55a58174bc1d4ef433672e98de259d69c9a8f8a440538c0655866203657ce62fbe9b6cc4278c9a354d762cea9d8b0125566fac2edd1eb9dda0a83428a43c2e6a7963cca4355063a9430119410ac36c2ffdf84", 0x1000}, {&(0x7f0000004400)="8bbf7629efe09e01c9bc74b559d33afbf45d4c865149925582ff2aaef4dd083cb49f7e57b9bf5d22d070587cab0b3446b0fc90d58736b5eaf8aa1d684de75807b03d78dccc46eae447491d850b4178a87fe3d3f34652b793177c9da0d545e5de495807214188b8dab91bfd40998a24c4036829c2a90897fba295aea5c4a457592da9fe1d06b0034969be51ec84aa0622f1405da25189f274fee40689e992edd76a175aa6a29bdcf1a678527f4f3cdbb662c5ca7451af14b3227c1944d8561bbb2c38f36dc1a38d8b34cdea45c840a4e324b928167d9cdfdb2dc66a0d87056e2a24a0cbf40fe32a5ea805a2cfc528900dad9d0de42b601d802734e4a0d50956d568d47db8f8df6cfff35f71017b52a14542fab7a24374076930748c48bca672a3672d2d09da644508c3ed17f13e1404e6724e401a3652ca7e7bd611a3a679f36d9ffc41a55ffb301abc2add59227a88c9546fabd3c8982eab3c16e752e32469a31c6aa351fdf292796a7f70f11e7b9b7260d90a9a17b14beb73039b518070bdd4d6a85d95b2d1ee5cb0d880734280d701940489faffd5388700a0c58bec2ede65af3325625a0d14ade7756d802c8490fde9d59c1cda6f7574bee9917b0db7c0191d94e4aea6edca930264c20e0b4375ae158839a89efd23a2f7191b097d020372fddf994248d25443f3fe8bd425ece2ed98e295b3e8d544897f447ce5875ed69c582db51cac8f9e4f580c8f3cb820ca2ac5ab0aefa185c2c6d68ffa008bbf58d906a8521ec279cbd6cc2b5e4314ff1e3b47bc048fd8833df51b185412b32ce6f0844d58c45cc1968c65236d4d6deacef9a8332d523a6fa867ddbc09614c71661488cb989bb46699f867b0cb40251e1f5e0068f0f49b90471579942d64cf27b85011de30c45a75d31c5bcc7f465cb4b76ed1667de5cb67bbe706ce41ce5b9a87873e0de05640e9b97fd01e26d0cc64d65e4982276bb411a12c29841643c61262da56f1f7dea492492f90bb292b06327cc48147568eb73de86c29b00a86b3639ca35714a54059a1478c6149069fb48425f4b35b81ebcb3b465327f987756613c44c05f388384fc5c7ea01d2c4cb216314396123a206ccc142c1cc71516793d3dcf763b819cc932f10f693c1d5c24130a37d932ebfd0baa255f99d3639e548822a42dc13b1147203e2e5fabde8193a433a948d923183257a41bec149c34256c4c037ce25eb2532a76f9bc64ebd84b4edb7a90e3f6edb0fbccfca6df4f888e797c1878c2d71cb59e0980b0290866b78d30d2a9e51f9661724f44ff26c6f26c05022576a120cdde0fc827811e4286348e363dcf6eae1a8ef6f94f596b87009f53608f2c8edd4255d468b0bab997534be6184c2b3608b0f8fef7ebd322b0da99cc407526cdfdc095fa59a9945a9234c078a6431a7e00ea5159228482d26d4f2ce93ead428d8bddbf8dfe92dc9957ab4f9f710ae22a20f63a7c53985d4934ed283e27c32cd45d148f4965444258e7f5dab0881f17f9d61b7cab7d20710afca52c444a3a05c2d7108a20fe648f8290b5fd22b70c32572c69ae9b68c622a77a2707811df93c143aede12c3ca9acb51857b6321ee85e2d9158daaf56d452bc04e49440dc7be51615d3155da939cf6cb380d8805d3094390bd494e96d5b42456a239985fc39278f6155628adb1597360badc108e4aca8a9865464d89645f341a4674d4aff7eba78e85950e8868fca661f996281494095ab9e26fb5a55b04a0d8fd455c77b196a1e325e585630bd989b272695e5432d9b8eac5a8711b8ca06786777fe68e12a33e88374f8ece11f86653f7a157a4f133574f4371ed9a0c12b2dfd840c75bd1669f1a80b4f6f1f4ca41febf5405ffa6fe7924d48886dd037ef5301532c047acd80525ddba5912443350164e7fd6e93956fbde183c44e9620e417f9e7172967776b6aeab165fb1a18d34ee43c4d38c4a41b6634a7b9f5584960b1cb8194f3f18523efca6a9bdbbec01060083907ae78a5bec113117d4aad430d32fb08090b248b01cf5a42f4c317cd0720c70dd7dd2f35124aa7a3ad06a6b73a236a311a0d0c518f076bd98eebc242badf29ffa628b58062b166a9452de53d6f19feae02eea9b2037cc843b4f4eb05e95264476186aa250e79a4a3c0a7ab2fe46f8d932b981ab9debc4c8a6a5f3939d858fe9c0624a6b098e203026bf23c3a7b1d75decd76096ba202b74a0b778a6b7ab281497a24c1e84c72a2a10a474ae27f54ea75b190d5113de562d332e20764828197d1cbd247853b914caf24c815a715925091d12c0be622eb8b9111c2d172ba3bc34976883449a18c61252db9937752183c0c6a64fbe178788ac3f11f49aaa248b796322d03198ef9fb370608fc2fcb0d15248d41d88e7885e19d427baf2cc59de8f83ab01756e22e38f8ee6d993a647157ce9ee5d65b495d3888dfe9a0436c428b207328f56a8f20604688d5b024fecfc3be3072a49aa678847bf6fe57a49a129ac72e601f3dfb1f416d3d12502d5b53860dcd5dcfc1a47fdf072eabbd3255f00d5eac0044c3635510b162f0a52c7b8cb04be6f5a35129813b53fae28206909aa6d85006c69f174029a2a46137d4690a86d700e53e2c068ec0fb073b04e37da67b806b04be60afea3fe9e4ebbc1decea09864abd0c8b7dbf06459c05e7ac1c550c661c104f315b2712aa7ab283fca0b7cfdfbf33aeae3b5997b0844cf637873e0d8c5c54c4e373c5b10321c147390da3abe39abcea936bdcc5a0d83a4c21313ff5029310e7384c5a19a5dbd89c8344cf850011923782c1ccdf9ee408767c8948ad94f75389f1b18df4e27e382300f537d4101474ae294a498b0c1237a5d09506f8cb5ca928536bc99a01b6943164d7852aa3d252eb110d5562463044df575e12615deafc82e20c4c138c0dd562f765c021276bb861c3cacf3eafad3533500b957c5664ac860612903a8081d02fe1108f48a9b709c8da0bef93eea79f772126dd41e86dab0a4ca5ab229192a26af9530b39972875ce9d2cfc649a491e159466eb760fbae16da56d70c16247d84062b98223d9dc158bf14402612e22051e0ebf21f139f4712beb10ddfa9207df2f758c71b3cf1d317602c8d698e3f53935c9ba957567c5e2321634b88bea68edd161b391d1c63e9f070a23d7eb2b78ad5b56cf18090b8845ebe91f32e20d48777fd884b55517f9f345026a4c41e5f1c7e2d559bcce163d278e2d13358091050e6760ddaa2b761d253790e95d56dc8d65c0f702c08d5553f9a9dd4f364e50b3c98f43d8a5e0f043c157dc3f34d4f375be18df39e0e094405112f0f43b05c7856d5a321fdf53bcd909372c8ea70c217ee984ba03ab567709d647c3e95d5cc100f7ccc284aceb1f463d7caebc74b2514d5fd4c0e96059e5a76d4921f23dadd78aee4a911117276cf55e429726a8d4bc54c9dee10d505a83d0a2b4a3317db5158f69d987094609021f32ad2279e1d8f1cd0cb2ad1c08022ec3cb9215ecf8c08ddfb14712b3b42bb014c26dafd54afe17f05099edb6cff7edd67dd05221367265de0ca9fc1e8ee7c84f26f42732a6f3764a4ebf07168231b16ccee6b8d82b5f73b748f0e088f59068c317cafa09e8d658cb512fa8a0d2bcce97046d73eaf30332b519f915b687c09ef79b7c7baa2ed532593a36088c863ae853ddb3195a3417533ba8bf1e071e010518fe7a5d4a7f3717ff2b3def3918aaa066016ca2f618a8c2c5b04e6c4f2e4f38c9f4f745195ba38d15d05f1ae4e60ad61bc92f78fad000c026a0474a810a87c59e193113f130ed5b0cbe3e46369d7d9c098b6455275129569c74aaef70460a119da97cc082c55c6ca4f68b00644b3e76bd197f42dd9c837b94380605b855af2aa48268bf025183099bf545bdfe39b3d6a18267db20a6bd35e030b3537fea99d403626f24649264946afc0b592e1fb8dc496b1205052d48fb4579c44b88da414ee2e2c1af9f611f44aced02ab760e6d564b053e0fa285d34f46acc2e33a8298eb5d9e1e815bd61e7be00070b82dcf9358e4d3749bae9c2c2db9c902cabde19e34eb98d02d55ac1a191b1e7e18cad5243e2c971b9d92cb104c26669c1b99548eb7d59d9245f04cf5362886b7412ccc83a33273c2368b57ac5756952437b0d5e64ea763a55ebdcfc3867e4f8f812b71aec816b10a254ef6f862b856e05e158a4a56a540b7e2fc3382a830113ebe8ac6bd03f7d56d661ddf965d6461c90fbd5f6a74b429c82bc78ffdc0cb78f8ef76124b4697edb45afcb0290df7e68666c52596b934e4199e73f553120fcf21c0b3ca7e7a76b805a0a229b7f40e4483049f3711d80ee68d9bf6de2f15f80826e0438af9151a74f8be986ffeeef5f40fc66fa1be7218b8b5134aa0adc3afe08a89b332707f693df04fc7ce3bfc788a0c8f470587d28fd8faa7429202767c68d0294cff930b03c251a848a52fc1520bfef70e5b3df755f78115767923dbad460f4cee461325cc33118d48f38fba40119c7dd8ed6b7b3ebeef7c65508a05819cada0bdfbbe3a1c6a8d0f82d5130f2f3d6d96af3c76b7b6911e44a4861ef78de6461966b03aa05801e3ae0ba4f648588e8d7ddcda0f76f8ef1a76b1b16cbc3db37cadf5b825764dd3a2b2fbc850147112b389f9fdd7d96fc8f1a58bda0b25cc262cccf442b19034a93940ff9d4bff96aaae9206bdec0c23f3789875fe865cfa8954f75fcdaddee17e296b287a49b34dfe0def35f2b59da773340b002b699ad44ba20df0237860048083d5b9f7d69e045343fa645d1673e1d86254e22c62a6f883e32a9effd1c13d91ee0fd48b9d57f030a458f39e3e75edd571e11957f9ccbe64b774f56bb188fc93f733ab59c87a88703949cf873170c3dac46d53677f6fcac0d9ba3d6c36505eb34a9e08817e33de84576b824c30a9c2fc87d2abacee3e2dad46386bc36aa35ed23ea11fb4d2355e2a01ef7df578004dfa57e3eb00db2edf1fcfaa4a608d31facb25a5de6e08b35151fb3d73ec06dce9e3326082cf5ece63f0a3eefba8c5fe9a69fe30ac843cc38cac8442574e9956d0fc8c257482458c10c5593958c3c1d3d1915996780c5beaceb3990246ba325208b9a9e0eb163b153c36ed73f171394d8b5bde20bb392961715aa2772386ecbd442e2ef7cb1d09d65d2623f7c4e2d79d3be789048e9a2060fb7b887251b0c7e0a7cb3dd4ddf39195502710f7d3fcb358fae6840a411846efc0be2b2716b3edca31b1a731d4492d686b12f9ec8c3006aabeb41a8c2b9f53f7f9c75010cfc53eb639596a8d301b07eaacaebda45b2547447c14a4d087cb7e404a2c94549399cac09aeff7b52a43908f8b6d66e035911b34435f8790e77abb2b407b8da897a5dbbbf0da378af668639a4fff02a9393226233766e2c535a5de7e0bc187431b72608c9fa8aea2c6bb7bcaa9d2c5eaf281fef168c49a97ca7ada9e0a9e433603684acac2fc181b57d4c899ae4b619eca7c308035bb7ca5b94b3aac37efc0d771c95aa9167d0169fac65ef5879e27a6b9d51721567d6f2e95170c9080fbb83ab5b0eefb80a48fff0304f71afa4062e84fff8fb96ed126564f27b6ef936e0656ea998a5adc2ee19435baa31b2b62bc1b4a9f13f25ab5875139bad731cbaf1d6e4992a3e8831118fcb89e465c7bdce1ff10a1867c5817593f18cd62f82b4107a14299ab2519ee05ce3871982ca4f19199b04512c7548223930ee2be18ffc4b726777711df1f07bceaf856da75ec1a35bb04ddf63a7f3b68401c74ac322e975a94e569ebe91fcf499e6", 0x1000}, {&(0x7f0000005400)="c1b41f9f3168625bf5ece59ce57a8f0d9f4bbb243d07a923cfbb0a8859895a9a9bdca488d9680aa3585f891e689608819fb9978d57ad1492a63681bff65080", 0x3f}], 0x5}}, {{&(0x7f00000054c0)={0xa, 0x4e24, 0x617, @private2, 0x464}, 0x1c, &(0x7f0000005580)=[{&(0x7f0000005500)="4b08cea4af7fb5effb5ce8352b9a28251987308b2117c8a4ce32670e1f7d823d705b6aa4faccd6f1c1475ddd9df92275c681089bd427e43913457ef96e241112fefd00c159796d8ff41d8044f33bfe8823e3795ec3732aa52b32f3c38eb9df5e2960badf856d67f6158dad018b85dae0f5478fbc407ec2c4b1", 0x79}], 0x1, &(0x7f00000055c0)=[@hoplimit={{0x14, 0x29, 0x34, 0x1}}], 0x18}}, {{&(0x7f0000005600)={0xa, 0x4e23, 0x2e1, @ipv4={'\x00', '\xff\xff', @broadcast}}, 0x1c, &(0x7f0000005700)=[{&(0x7f0000005640)="5ba2cae580dc49fc169d0641e18bf968d9cd146b0af56837401f8288c475b9bf7d4fbbee58c73519ff043c0ea253b72d023724c9b1fcd73985eecfad808fe32b21790f47081462d519533760560342eea3ed0b4579db1b774695a0233be67f6e300d45494bca87a6d6e7b789daba9203889493257bc950592ccb3f8adf120221d25e3029743efbe6b5cb72674d2e75853c122ccfd82b8144a732c54a3b29d7493b4b948b2d57fa852597ae42220101a4c93201cb0845ffb189c8", 0xba}], 0x1, &(0x7f0000005740)=[@flowinfo={{0x14, 0x29, 0xb, 0x1}}, @hoplimit={{0x14, 0x29, 0x34, 0x20}}, @rthdr_2292={{0x38, 0x29, 0x39, {0x2c, 0x4, 0x1, 0x0, 0x0, [@dev={0xfe, 0x80, '\x00', 0x1e}, @mcast1]}}}, @hopopts_2292={{0x20, 0x29, 0x36, {0x1d, 0x0, '\x00', [@jumbo={0xc2, 0x4, 0x8}]}}}, @hopopts_2292={{0x78, 0x29, 0x36, {0x32, 0xb, '\x00', [@enc_lim, @padn={0x1, 0x8, [0x0, 0x0, 0x0, 0x0, 0x0, 0x0, 0x0, 0x0]}, @jumbo={0xc2, 0x4, 0xffffffc1}, @calipso={0x7, 0x28, {0x0, 0x8, 0x0, 0x23, [0x100000000, 0x1, 0x100000000, 0xfff]}}, @enc_lim={0x4, 0x1, 0x40}, @enc_lim={0x4, 0x1, 0x2}, @jumbo={0xc2, 0x4, 0x7}, @enc_lim={0x4, 0x1, 0x3}, @jumbo={0xc2, 0x4, 0x80}, @padn={0x1, 0x6, [0x0, 0x0, 0x0, 0x0, 0x0, 0x0]}]}}}, @hoplimit={{0x14, 0x29, 0x34, 0x225}}, @hopopts={{0x170, 0x29, 0x36, {0x2c, 0x2a, '\x00', [@pad1, @calipso={0x7, 0x38, {0x3, 0xc, 0x7f, 0xbb5, [0x40, 0x1000, 0x5, 0x3, 0x7, 0x3]}}, @hao={0xc9, 0x10, @ipv4={'\x00', '\xff\xff', @multicast1}}, @hao={0xc9, 0x10, @mcast2}, @hao={0xc9, 0x10, @empty}, @calipso={0x7, 0x10, {0x1, 0x2, 0x77, 0x101, [0xfffffffffffffffe]}}, @padn={0x1, 0x3, [0x0, 0x0, 0x0]}, @pad1, @generic={0x20, 0xc5, "53108c437e0999adf47f6778c43b28cd4f49f1194d80509b18f68b59bccb057032d5b475c3e0769149a1f8d3d8df06aa5668d8ab3d9a0e3c742cb3153f7b5e65f2ba90bf5fa46a10733cd35760280cee062eb242944f78223145b8bb8713bc6c2c42b7421fb5d762e7a5937a45ad7e93862a704d20401a706cad5be91e94d120231e89170c76e078121e8b6459f06f93f5076cb424940a59f2058a47c441fe5ab195d44ca4b97c8df9c4e7ff9973840fde7b426c7153f69995b820745f67b528b6880d30c8"}, @pad1]}}}], 0x288}}, {{0x0, 0x0, &(0x7f0000006e40)=[{&(0x7f0000005a00)="3f24a61dac45a3915be9e85c3d6da7bee1ce6e27a2c1d3dd2ff8ebc2cb17456e97bb6078f9bd2a94fd0b581b9b4cf5c9430c66eb6761d3e4b856e0f1ff5b1c4a58a33e33c9dca51f68f9d3ea06214ba703a44e2aef21a72142c1bf4b82cdc1fec109f97af66ee8e0449c0cf7", 0x6c}, {&(0x7f0000005a80)="be1d1409ddf392b46d366bf209c29ac46157627d5f0abbed918a67e969c2fc05b296f94cbbb7fa62ac61a54657e6e7335159e925159e63943c4e0c04ac85c8bd2c231266ed49bca18a54c1e01f4048d991c98a18679fd07f30815a91a3ddd98a96f003036347af78d8b3e802dce608606c106ddf8632958b2a69281bb5b97df407d02a6e29a44aa63cf626802532ba242286e4769d4aed787adda75d72200ebe92070f6a558527273e5e185168353e3d5ab078c4a75a24810c7e73f5990bc36ed867ac3003ec50c374962f69cf1c8d239809b2", 0xd3}, {&(0x7f0000005b80)="dd3b04f9beac32d22f8c9f9dbe23819f366edd0bedfebca2f6569885471bee84bd0997f8a2b7bc3db803265bf61414b77f267a5d45bf1257fe10a43d890e3aa39201be83a0c3783e55698de6f828d17096fe6c9b2415dee92975c9eebec69b05a913af2db2415e4d7141f01c5f5958eb53d8581cc49d4e4d7d9b960c29bcbd552268de2463fd544cf45d190d28ef0a22308fded4eb6a04e0b131ebbefdf10992faf20d282b0b3421", 0xa8}, {&(0x7f0000005c40)="7fccda0cc8ead8aae0408a025b4df83ab54fbef5531598c67d1097ecd28ea443b6580499bbd30484f1be206f90ac72178377a35f5df1db4bd26972446dc085019baed3763c2f7125c6fb00ef5ef3c3074cecbb0e7e2ebc35c88cc8fe883d0282ffbedabbfab50e048ed6f91b84b89ccbc344dbcfc642c8cac6f34a3b294ab53866f46ebeb0579b388fe36483d482279ae1bc6e0dd1e2142485b5605b5f4cc853b0a374769c7677ba156392c2d35ba0c131242bf2ad15b567666f0e1b0034d0786d3ce9b5bdacca5671ecfdc32b0f7002fa15da4fc6d4331df428bc733e203bb9aa1507af4f09909e", 0xe8}, {&(0x7f0000005d40)="bfee1056b57b8a87fac60fc21a46be552c01e563c8e40bb9a903f5a654b36e82f21c55ef1ed1a5e179705d2626b7e5ec0fca187a1b4c1b3c08b2f49fec875d30d734ac1f6ecdcfc0a4693a0386de8362fe491b42631a86daea0fd6ca8f34d545584cce2c966e7a35060bc759da4de1b0ba6c52fa9332bb156059256bb35b48b1fbb9ad5f8967765b24a87230ecaceb481a2a434aa4db2e9961f7ed8126e3175b807a67a465b02b9bc8e332641aab857d7047425b8e4712bf5b5e9557aa53ef857de882e5d106b5141d53d87b7078c19664f260fa3e7f54a6cb281cf2c155642bb11156cd56bf882b6d73c643c2f828a93a7c2f36cbf90e2991ceed4e9062b43b13724fc47942b850aa4250c8133f0c8fab15589d3caea45cf891c47b5e3e5b9c3e5677ae5186aba6be5217a4d306f5ccd0645d460b7d2bfdcdcc48b643ea54333a8d66ae864508d8922b1ad3c44f85b27ccc92c52c96bdd9e2d4f918d771498742af7a0701610c906ce401e3f2bb400a67f711b6ac624986fdf0e372391bd8e30b07648dafe3f0b06976f1ece8c4f33483895ef016e26b0f16f8acf622aacc5bd9c69d5ace1b7e8ef753d339548f555bc94000683f1384a1d8f854572a15ab74db640a4aa77a03213e8a1bcb1f361cee0dbbeb05e48fe1700fd262dc094672dd1a1c8f2cbe825b246c0df1e7b9b49fa055fbc68e334a2e47dcc50ffed3a65dab7dbda3ca1788681cfb74e212d445c30f9b603aa3751db72eed21b90add76d1cf6306b519cf61da960a0e472ce1cd4c779d7227f01b8d6f61636b1d6d16bf03b8f8dfd45f4021587ac4a1ff66aa6ba7c8725ac91f599c4db8279cbbb91cdbbb45a18d224741b71eb79ff296135f96c67689281242ec0c1bb694618b5597603e3226bde62f885144c2f65b160a291e08e8ae248180d9cb20f9635e15dd7446f4ca692fa8ccbde5a20f7951fb33d5a50ccba4f509017555f6699dcd1d626e29d5aea8986d96d7ba488aad9f512ea50bc665e12e5e9d7c6390637a64f5cf2767e717a8ba1a746c5aea061881034e89f58f8d075d673307ec1cbc1b598f18cad0101f1abda02c3f5fe6f72a0e2929a95923cef2c01832112a47c37b91cb8d65c4ca41f81d450ba92a32d482c59d2a839f0c36c1c301f3992307a5be2601088146f9fe89f33f7eb0a4b9ddd52aaee0663f683e352715a0594f4bea49b2f090c85fa7bb3ac11f436212edf7fa7864cef8296a675af92b5ef5e5131749646ef451ba3305cd00d764c997b2da08ab065e5c8c2d0e968303f2adb34f118e3f1159f47e51427614d208c5e2a4b2247d31d92df9f87197fbf5d5637bbe85e47269d3a41e2d4f7a1ffc0a6281e6624e9fb10f1eae343ca5b94f19f92041467b91128a4a7eefd26d46bb0e816d14fada6c50de931a9c6ba006aa3156d86a01fec57725f955506953a5b8364f857b130fee93e3a8202cd9601a25e03fc301beef7f66533f55cdff32fc0ac0ad154b864e975ceac381b2d8fce3232a9b8a9b41a7e4a2a671f57c275250efbf75f54773b3b00dbabc21e64ffa508173c9deec95a2e4298286c3eae6fbd9f22d28c16c475e2cc4c5d68b6e421b746595601941da19b4824244caf8cd6d1013fac0cf7b1e25909150c29e31d077f5f0f2f239f9164526191c1e604525ee6ae34314052e13b38ea6bcc59c46c49c3f332ba4129acc4edf364eb5e83f9affb943368c81325e787a85c96a6b95b6788d62e8167e377ac4d9f548705ce15afa8ce9d76b3ee238a1919b3ca852669df6efc695daa1677758ef49c42487e5059b48d26816f9e7e352faf84ccb0520f45f5ca62d8b45ef19e02d51010d7dcfd572cdb1fc9a884eae5e41ad0c5f06bbaf2e909a87233d79c6ea44668452ab985a92175890848590b4c2aa322b9b01828484ba84f6ac91b262cd42c8d88bcc41010c9a7cc1d3afd617a3cb0ed2dbaa91dda6140b04586f85e7d92ab92989a1cd676060800ebf92eec3a0b7de05e15100ba916f03ad1dc7753a4426981296bc9cbac3b773a097ab16ce41f4067f5b879e7b4d040a3c9360ef77c6a31e6e1b6d578086bc82963e910a9544b3ae2f771f5099932a18391ba0987cc3e66fd7849d7062b64ff12b80c9584026ca3751d49e6b78ddfd7f968bba507078b2eb90bafa94dcb1c9f88c777e538bf48ac976bd1df87c01911e35e3d47ecc56bcce9d6bf8559b135b6609af71a73db8923d70b041ba758fc1e7ab1ddccfc94071cbf78a4ed929f5ea9f1cf5373753323a22f99eee1639603d48ae68ef868f793ad02b81ec6ad6293df0d37c28a5c251602670f9b7a6a21b327abb84d46dc6960a1f4d28a9aac3dab1d4b919a3f6d37f36147dab3c79c94b848a1b995bd396a8435d4ffa5e78530a802f79a3653b28ea52e154347d0c960b452b3c7f8341b8b26a778d43466d9a54403233feffa74ea48a68f5fa8aaf72f5eb51e3b07f6c2449c6f5f2f970edbda9a725d80997bebe0c54373275b56de857a742e46479e09e663fe756b11ac45ae3c141939cb9ce03ff1e6fb12d01be311ce5f59c9f124aa26b2bd87e5ffdb7f73545c7aab71633f646dc85380800aecd2975447d87e7eb6c16addaaecc919bb401789952e9ff14705bdcad744af1ecd018bb7f79bd5cae90427cff478183f3b28233a55ea750d3d88972869c31a2e0c932d06a7c186c4185efa36233edc88c186fcd2f1544d382cc0744167f54e359d8547538f7f6926ae017e5ba662843960ce1a3140fa1290f9541db6ff9507de2fd0e8e77dc0a55c8096ab22a6c21767e18318706feec569be584c13d5337d52414ae60448aa383e780785eb3f196c7f10fd376f911759e420ddb27cbd744fa2a7e2dad5543a07ed103e5a467ceee127030a0cb48d98f7b4821926dbf43dd4f95125feefed4f48c3195a61d8d8a1e848244ed9f7b332eaaa88e55195a509daf964b224e7fe34b5092f6ddb6b309d57963651d1ba1c447e02711374efeaf4c380979b1810f240d39708d23476c686edece9d991048f7f710e8a31a0bf3bd6479fff543a0698933261c29955710f294873df42599d91aedc0a9f9172d131275a83d1bda56784cf1139dc33dd33ca0cd2d0c4ec2f4912e2f2ee59c1b3ade58553f7df74929517dfe28b989cf1207f1e5bb530c3afe62f404af18196a90222323c8189a4bac438ad35a1d31efa40e65fde299ffc582195d2e995666b78cdb5e2d10913c935e92936c4935eb503977c635bc4e34ff5c26f8a8dca7bb55c6f99b4bfaa97b75cc08c75dd133fb8b09fa028a4ae9e9529121fcd49bdad51f0e128814cef7772f3e822cb29bcef1544b0c54c08b6d40f24727897a0a0e9e4a15e829bf4262bb0c5fff063e277e8e511ea4b0c91093ec3bf215acdddc3c6fac93f8996a2d10ea838fade70a41b354cef38a53ba1c71343b0b4b209f671c63051a9019f34355676a69176101d92cc0a7f2dbe0a4115c3f9473a9614b0684f99feb478691b9fd250a426d4e8bdebe27191a815fdb49f1ff2bef88f8cdd20142fc56ae3893c68486dca47b7a60453508bdccc410ebb0e088677ac1875e562624750523634956baed75fe7a10ef987a6e4b71f303a2c41c72dfa86e33bd7cde76c6c033c74418b95cc37638314e70e6116d7f321a11ca13db037bcbf41ef8b6b8bcc95c32d8af5f984a961bc47fbb1c5233d200a95006ddcaee294ff7caf5a8aa192023504eea099bed488ac8074c0df343b1e3d73c3073028654398f81864911fc106282ff5e22f2da39c709380f71e28d391d171eec72fd8affc3dc91b2d0096e2943270351679483baef9671325dfa1ab923659257a973820c2e5e4229eb630fa6a12719ae8b0ec0a498b69bbc94a8400c54b66560579736523a655f7481d9330cd7e6d9db3e359f9478bc1eeafb969a0c4942c0d1bbaa20b5be683b0a3f4d35d0df39b829ccbbdee0c6b39bfb1964f7b2a6f19055a57de2062caa3ed43595a92ac949421fbc659c31256d5c7605ae88cdb119829824f35b256b3de1668a50a290a356e32a645e8746069e4e3b2aa65ccb2e591471ae241b63f4495dd622d3df0f312eb9e243712880bdbc4000e0f9e490e7580d9c1f08f7144033a4af0763b93c4d2fd1e6ace9ce8f1c9760b581d51af613eccde9487583fc359ba412e590f19fc132e51564a23c0c2b767760a324547a7d04b271915adcaa436aa17fd1a81002e3e3a8f412b5d281126c86c0b83d82b4566b38ac9ac601d5507593d72b403b39a6bee4754adb27e33a44a713c84fccbd6091024b107940eab518243b4d01b02a253c86163826133717c04268cbe64011b1834b1568612e884211ba20f4b8503f3837ab8e7bafbf4f55b888fdfc4eb02a8aa592341ee755252af23f5db80105360645e0071e0098d6a0cc82176f471d67a18f06c3b2ccd849ce65f4755fa0ab802e77b0cf5f24229574e4e2d8446e45d212d4f2d554221a97b37a381f76a7b918572b70c7f9e37f30458fa38bc075e3ec575ff9e5c1af6499833a90176d3e9e7729ec9c2c81920c9ec0b457c1641963f1491238c3df52cc8c04f66b1f0bbd4577cbcbefbc0c7fa266724bb6af72145af3da1f131bf7990a0b1094f3cfd1b44ee43656c95b9e75ffc8de888504bab8795951c4ab5672d5daeabfccf6771c367910e563e269e093343081264b813667bbd48fb8932f89ba32429659e92dcbda8ce138ca31d6232905c63eddcd4ec935c4a4f59d970325b102f9b4b52c09d7305b88116480d069be7ccee249dd3af3530e4f79c7169bb13fb5573fdcd7c164a418fda059b8d37f91ac9efd69600dda57525171bfcd554c412b4f40b2b8176eed980f1c55f4ae3bf1543f040ff13eceff746b31dac494bb5409ea6d26f62edc630686423a8a5d4348f56b38d276babdca09cdd13d5d4b509cf600e81e8b79954328952002f47b3508d2894845f64c19f04d2dd0d9680fc40494e39a17fabe122c50a823ac06bf04c52827a3714974a4ecab62d1034011d021713464839a6a95ac583b8b36d90ff9afe10660372d9a9c71b1f61598d574c3bc3d779a22a61c2a466a9eb738aca6e55b3c45ae1c1aa26a4e8fe34d7176cc9c25fe553cd1b5f98f046f6a63da13e2de4d953d3e5bc8f559a4edbb1a13af904aaaa832e606852b3ca47d338bffdd4fc2c576cd188144a2a8a4b6ca81ec6e2134ad19eeb0b78a55c62a978a891d4d49050c918008c8b0b5651e1823b4ae89139691177e32436da940d4357ab012e1e394532aaab6f03d6576facaa71649fe3c2a1c92c4d02c48d65a684b3744810dbd6f73326ae4caf6d0e7da163a6a12d8499e4f9f860481656df3d308a55941a061f24335d9723ea2bb34f518a7081d91c76f695fac623388b7b24264d44e6e2fc2a3dcc032e9bb45fe83f89ea14ec3f9b962c53be6fccc563857e1e87dc20454e2f46623b0c350d11930853a44becb53a340f0674dd9a1ebec56b933c53916cc15429bedddddc2d670b1085dc02f0e3086b7b937550934d3e52bd2b4d31ffd7d90237fda2a593188bc92f7790deb1f2f660af4070d2601af107d3586ee94b3c2924a397458cc7e06538b8bce8d28212073fb0706453634d287e5c1a62987357ff3d1125ff9c70f7e60fd4e23af41c7ae93224aab69868db2a20a8ef76b00cf1c7dced6ba8f2e313a222dacbabdf612d659fe9ac3721d9de164be80e9328f5ec1133af28fa496f5bf336c683e3a7f1dd65bb6fb2ea300686c7f1e3a7b5be1927ad51065d4977fd8ed4582bc0f3512acb9805e9a", 0x1000}, {&(0x7f0000006d40)="7534da42ab8ebea4423f2fc1354f832bf3197028a50eb9562a1319e68849bdc7bd0fe345802d5ef1c48b586968170bbe22168fd8011b05c979289d36b4d05161782c6b1c0a44e9f6081d2a3599c0978c", 0x50}, {&(0x7f0000006dc0)="56ddc9c96cf571414c61f6754cff4eb202fd95314598c614c24739b62940a41a5e0c18d4600b7f5eb7498fb12a81779558faf631673c04ffc0914544a1ff292c2bd37dc13d37c6c99e7a4446ff4d3aba057a81afaf1f1770599918c3ebca305a47c05978576ff0512be7c56b47b887", 0x6f}], 0x7}}, {{&(0x7f0000006ec0)={0xa, 0x4e21, 0x2, @rand_addr=' \x01\x00\x00\x00\x00\x00\x00\x00\x00\x00\x00\x00\x00\x00\x02', 0xa00}, 0x1c, &(0x7f0000006f40)=[{&(0x7f0000006f00)="9fff08a2", 0x4}], 0x1, &(0x7f0000006f80)=[@rthdr_2292={{0x78, 0x29, 0x39, {0x3772466af79949df, 0xc, 0x2, 0x6, 0x0, [@loopback, @mcast1, @initdev={0xfe, 0x88, '\x00', 0x1, 0x0}, @private2={0xfc, 0x2, '\x00', 0x1}, @private0={0xfc, 0x0, '\x00', 0x1}, @dev={0xfe, 0x80, '\x00', 0x12}]}}}], 0x78}}, {{&(0x7f0000007000)={0xa, 0x4e24, 0x2, @mcast1, 0x2}, 0x1c, &(0x7f0000007200)=[{&(0x7f0000007040)="28357b828bfac2c815793e46d9c9d23dd66973f83bcb7af7e6ba0da677af64bbab92adf0c17fef707fbc64e36f26513f5a8bcec1d766a76f11bac51e187e383d4253fc993b06042a941258142cfad391d252ea2c21c2faa5266c7301f229", 0x5e}, {&(0x7f00000070c0)="253f98c2bbb8f7e64e7974e455fee0a69cbebac7e94979a49d788709cdb17e886dcfde86226461a63ec59adaad90ceafee2b19023b8aa04091ea4e0b128b1b526065fbb757e36cd551000d5538082a002ad35f0c436a044b39d166659c4d2b5125b7faf422b55b4a2c7245e0d8717054e261c05c9c7bac3257d6fdeadcf5688aeb678e7967c3b704216055ab039b9e8b75569390c6d00d055249c4c4cd75826bca85d05342028bbf622467e9977d54b909c7f6579c5f97609552", 0xba}, {&(0x7f0000007180)="85e273e9152c799a6557a0362f4181b95bf2794fd5ec1f83b53ef4aab07d5498cb5ff8b1b6e2dabdeef02cd977d5025efe659c331887f2a56f1c3552356fa405fcee979bcc12c95776cbf788e5a06c0f33e39fb12d5dc2f755798f73", 0x5c}], 0x3, &(0x7f0000007240)=[@rthdrdstopts={{0x20, 0x29, 0x37, {0x32, 0x0, '\x00', [@ra={0x5, 0x2, 0x5b4d}]}}}, @rthdr_2292={{0x48, 0x29, 0x39, {0x2e, 0x6, 0x2, 0x8, 0x0, [@initdev={0xfe, 0x88, '\x00', 0x0, 0x0}, @remote, @remote]}}}, @rthdrdstopts={{0x20, 0x29, 0x37, {0x29, 0x0, '\x00', [@jumbo={0xc2, 0x4, 0x5}]}}}, @hopopts={{0x20, 0x29, 0x36, {0x3b, 0x0, '\x00', [@ra={0x5, 0x2, 0xcb49}, @enc_lim]}}}, @dstopts_2292={{0x30, 0x29, 0x4, {0x33, 0x2, '\x00', [@hao={0xc9, 0x10, @initdev={0xfe, 0x88, '\x00', 0x1, 0x0}}]}}}, @rthdr={{0x28, 0x29, 0x39, {0x0, 0x2, 0x0, 0x3, 0x0, [@loopback]}}}, @tclass={{0x14, 0x29, 0x43, 0xffff8000}}, @dontfrag={{0x14}}], 0x130}}, {{&(0x7f0000007380)={0xa, 0x4e20, 0x6, @private2, 0xb99e}, 0x1c, &(0x7f0000007540)=[{&(0x7f00000073c0)="9b493e883c245d00743a2979372b087e556d2c6a1a147d745d361241eb4bcc61ca9ee47092f93b118a97ca16c7f3a842e838d951ac2743f018dbbc269b17b6a381b46d849674990f387743d2570a9408de434f5c86042e5c4fd1cadf764ca1efddb267d436f357f56e347a8d5378d46996a2835df9b410b54472490e9bac68409bf773fdb170e4ac6388cb8cfa8d3c1b6d735fab86c5", 0x96}, {&(0x7f0000007480)="821ac7a68d525e25b36fce772aec357f3861b222a9f729837224614dc2d28fdad2e9a03caadecaf1cfaf620de3d0d25030f648b915c6558d9c28c4d201c71b708952f38d1261e71a4e4dcd30824d45eba558da6910cc9c0b5639637ba053bb8512f8c07e29b320a8f15e6ed4e29881cbbbd4d542481e463217f570df132147dd69a3b4649eb982e6182431", 0x8b}], 0x2, &(0x7f0000007580)=[@flowinfo={{0x14, 0x29, 0xb, 0x1}}, @dontfrag={{0x14, 0x29, 0x3e, 0x9}}, @tclass={{0x14, 0x29, 0x43, 0xffffffff}}, @dontfrag={{0x14, 0x29, 0x3e, 0x7}}, @dontfrag={{0x14, 0x29, 0x3e, 0x19}}], 0x78}}], 0xa, 0x40000d4) (async) syz_init_net_socket$ax25(0x3, 0x5, 0xcd) (async) ioctl$BTRFS_IOC_SCRUB(r3, 0xc400941b, &(0x7f0000000640)={r0, 0x8000000000000001, 0xab05}) (async) [ 2836.576112][ T3259] bond1303: (slave bridge1249): making interface the new active one [ 2836.597550][ T3259] bridge1249: entered promiscuous mode [ 2836.620843][ T3259] bond1303: (slave bridge1249): Enslaving as an active interface with an up link 04:12:44 executing program 3: r0 = socket$netlink(0x10, 0x3, 0x0) sendmsg$netlink(r0, &(0x7f00000047c0)={0x0, 0x0, &(0x7f0000000000)=[{&(0x7f0000000100)=ANY=[@ANYBLOB="18000000560001060000008300db730007"], 0x18}], 0x1}, 0x0) 04:12:44 executing program 1: r0 = socket$netlink(0x10, 0x3, 0x0) r1 = socket$netlink(0x10, 0x3, 0x0) r2 = socket(0x10, 0x803, 0x0) sendmsg$NL80211_CMD_CRIT_PROTOCOL_START(r2, &(0x7f0000000580)={0x0, 0x0, &(0x7f0000000540)={0x0, 0x1c}}, 0x0) getsockname$packet(r2, &(0x7f0000000600)={0x11, 0x0, 0x0, 0x1, 0x0, 0x6, @broadcast}, &(0x7f0000000100)=0x14) sendmsg$nl_route(r1, &(0x7f0000000040)={0x0, 0x0, &(0x7f0000000000)={&(0x7f00000000c0)=ANY=[@ANYBLOB="3c00000010008506eb9afc4cd8d06e754a0081c5", @ANYRES32=r3, @ANYBLOB="2377f292252155b21c0012000c000100626f6e64000000000c0002000800010001"], 0x3c}}, 0x0) sendmsg$nl_route(r0, &(0x7f0000000240)={0x0, 0x0, &(0x7f0000000680)={&(0x7f0000000400)=@newlink={0x3c, 0x10, 0xffffff1f, 0x0, 0x1a672500, {}, [@IFLA_LINKINFO={0x14, 0x12, 0x0, 0x1, @bridge={{0xb}, {0x4}}}, @IFLA_MASTER={0x8, 0xa, r3}]}, 0x3c}}, 0x0) 04:12:44 executing program 0: r0 = socket$inet6(0xa, 0x6, 0x0) bind$inet6(r0, &(0x7f0000000000)={0xa, 0x4e20}, 0x1c) listen(r0, 0x6) socket$inet6(0xa, 0x6, 0x0) r1 = socket$inet6(0xa, 0x80002, 0x88) bind$inet6(r1, &(0x7f0000000580)={0xa, 0x10010000004e20}, 0x1c) r2 = syz_init_net_socket$ax25(0x3, 0x2, 0x0) bind$ax25(r2, &(0x7f0000000000)={{0x3, @remote={0xcc, 0xcc, 0xcc, 0xcc, 0xcc, 0xcc, 0x1}, 0x1}, [@null, @netrom={0xbb, 0xbb, 0xbb, 0xbb, 0xbb, 0x0, 0x0}, @bcast, @rose={0xbb, 0xbb, 0xbb, 0x1, 0x0}, @bcast, @netrom={0xbb, 0xbb, 0xbb, 0xbb, 0xbb, 0x0, 0x0}, @null, @bcast]}, 0x48) 04:12:44 executing program 0: r0 = socket$inet6(0xa, 0x6, 0x0) bind$inet6(r0, &(0x7f0000000000)={0xa, 0x4e20}, 0x1c) listen(r0, 0x6) socket$inet6(0xa, 0x6, 0x0) socket$inet6(0xa, 0x80002, 0x88) r1 = syz_init_net_socket$ax25(0x3, 0x2, 0x0) bind$ax25(r1, &(0x7f0000000000)={{0x3, @remote={0xcc, 0xcc, 0xcc, 0xcc, 0xcc, 0xcc, 0x1}, 0x1}, [@null, @netrom={0xbb, 0xbb, 0xbb, 0xbb, 0xbb, 0x0, 0x0}, @bcast, @rose={0xbb, 0xbb, 0xbb, 0x1, 0x0}, @bcast, @netrom={0xbb, 0xbb, 0xbb, 0xbb, 0xbb, 0x0, 0x0}, @null, @bcast]}, 0x48) [ 2836.996509][ T3268] bond1295: entered promiscuous mode [ 2837.002584][ T3268] 8021q: adding VLAN 0 to HW filter on device bond1295 04:12:44 executing program 0: r0 = socket$inet6(0xa, 0x6, 0x0) bind$inet6(r0, &(0x7f0000000000)={0xa, 0x4e20}, 0x1c) listen(r0, 0x6) socket$inet6(0xa, 0x6, 0x0) r1 = syz_init_net_socket$ax25(0x3, 0x2, 0x0) bind$ax25(r1, &(0x7f0000000000)={{0x3, @remote={0xcc, 0xcc, 0xcc, 0xcc, 0xcc, 0xcc, 0x1}, 0x1}, [@null, @netrom={0xbb, 0xbb, 0xbb, 0xbb, 0xbb, 0x0, 0x0}, @bcast, @rose={0xbb, 0xbb, 0xbb, 0x1, 0x0}, @bcast, @netrom={0xbb, 0xbb, 0xbb, 0xbb, 0xbb, 0x0, 0x0}, @null, @bcast]}, 0x48) 04:12:44 executing program 0: r0 = socket$inet6(0xa, 0x6, 0x0) bind$inet6(r0, &(0x7f0000000000)={0xa, 0x4e20}, 0x1c) listen(r0, 0x6) r1 = syz_init_net_socket$ax25(0x3, 0x2, 0x0) bind$ax25(r1, &(0x7f0000000000)={{0x3, @remote={0xcc, 0xcc, 0xcc, 0xcc, 0xcc, 0xcc, 0x1}, 0x1}, [@null, @netrom={0xbb, 0xbb, 0xbb, 0xbb, 0xbb, 0x0, 0x0}, @bcast, @rose={0xbb, 0xbb, 0xbb, 0x1, 0x0}, @bcast, @netrom={0xbb, 0xbb, 0xbb, 0xbb, 0xbb, 0x0, 0x0}, @null, @bcast]}, 0x48) 04:12:44 executing program 0: r0 = socket$inet6(0xa, 0x6, 0x0) bind$inet6(r0, &(0x7f0000000000)={0xa, 0x4e20}, 0x1c) r1 = syz_init_net_socket$ax25(0x3, 0x2, 0x0) bind$ax25(r1, &(0x7f0000000000)={{0x3, @remote={0xcc, 0xcc, 0xcc, 0xcc, 0xcc, 0xcc, 0x1}, 0x1}, [@null, @netrom={0xbb, 0xbb, 0xbb, 0xbb, 0xbb, 0x0, 0x0}, @bcast, @rose={0xbb, 0xbb, 0xbb, 0x1, 0x0}, @bcast, @netrom={0xbb, 0xbb, 0xbb, 0xbb, 0xbb, 0x0, 0x0}, @null, @bcast]}, 0x48) [ 2837.220370][ T3269] bond1295: (slave bridge1258): making interface the new active one [ 2837.244483][ T3269] bridge1258: entered promiscuous mode 04:12:44 executing program 0: socket$inet6(0xa, 0x6, 0x0) r0 = syz_init_net_socket$ax25(0x3, 0x2, 0x0) bind$ax25(r0, &(0x7f0000000000)={{0x3, @remote={0xcc, 0xcc, 0xcc, 0xcc, 0xcc, 0xcc, 0x1}, 0x1}, [@null, @netrom={0xbb, 0xbb, 0xbb, 0xbb, 0xbb, 0x0, 0x0}, @bcast, @rose={0xbb, 0xbb, 0xbb, 0x1, 0x0}, @bcast, @netrom={0xbb, 0xbb, 0xbb, 0xbb, 0xbb, 0x0, 0x0}, @null, @bcast]}, 0x48) [ 2837.287907][ T3269] bond1295: (slave bridge1258): Enslaving as an active interface with an up link 04:12:44 executing program 5: r0 = socket$netlink(0x10, 0x3, 0x0) r1 = socket$netlink(0x10, 0x3, 0x0) r2 = socket(0x10, 0x803, 0x0) sendmsg$NL80211_CMD_CRIT_PROTOCOL_START(r2, &(0x7f0000000580)={0x0, 0x0, &(0x7f0000000540)={0x0, 0x1c}}, 0x0) getsockname$packet(r2, &(0x7f0000000600)={0x11, 0x0, 0x0, 0x1, 0x0, 0x6, @broadcast}, &(0x7f0000000100)=0x14) sendmsg$nl_route(r1, &(0x7f0000000040)={0x0, 0x0, &(0x7f0000000000)={&(0x7f00000000c0)=ANY=[@ANYBLOB="3c00000010008506eb9afc4cd8d06e754a0081c5", @ANYRES32=r3, @ANYBLOB="2377f292252155b21c0012000c000100626f6e64000000000c0002000800010001"], 0x3c}}, 0x0) sendmsg$nl_route(r0, &(0x7f0000000240)={0x0, 0x0, &(0x7f0000000680)={&(0x7f0000000400)=@newlink={0x3c, 0x10, 0xffffff1f, 0x0, 0xc000000, {}, [@IFLA_LINKINFO={0x14, 0x12, 0x0, 0x1, @bridge={{0xb}, {0x4}}}, @IFLA_MASTER={0x8, 0xa, r3}]}, 0x3c}}, 0x0) 04:12:45 executing program 2: r0 = socket$netlink(0x10, 0x3, 0x0) r1 = socket$netlink(0x10, 0x3, 0x0) r2 = socket(0x10, 0x803, 0x0) sendmsg$NL80211_CMD_CRIT_PROTOCOL_START(r2, &(0x7f0000000580)={0x0, 0x0, &(0x7f0000000540)={0x0, 0x1c}}, 0x0) getsockname$packet(r2, &(0x7f0000000600)={0x11, 0x0, 0x0, 0x1, 0x0, 0x6, @broadcast}, &(0x7f0000000100)=0x14) sendmsg$nl_route(r1, &(0x7f0000000040)={0x0, 0x0, &(0x7f0000000000)={&(0x7f00000000c0)=ANY=[@ANYBLOB="3c00000010008506eb9afc4cd8d06e754a0081c5", @ANYRES32=r3, @ANYBLOB="2377f292252155b21c0012000c000100626f6e64000000000c0002000800010001"], 0x3c}}, 0x0) sendmsg$nl_route(r0, &(0x7f0000000240)={0x0, 0x0, &(0x7f0000000680)={&(0x7f0000000400)=@newlink={0x3c, 0x10, 0xffffff1f, 0x0, 0x253c3b, {}, [@IFLA_LINKINFO={0x14, 0x12, 0x0, 0x1, @bridge={{0xb}, {0x4}}}, @IFLA_MASTER={0x8, 0xa, r3}]}, 0x3c}}, 0x0) 04:12:45 executing program 0: r0 = syz_init_net_socket$ax25(0x3, 0x2, 0x0) bind$ax25(r0, &(0x7f0000000000)={{0x3, @remote={0xcc, 0xcc, 0xcc, 0xcc, 0xcc, 0xcc, 0x1}, 0x1}, [@null, @netrom={0xbb, 0xbb, 0xbb, 0xbb, 0xbb, 0x0, 0x0}, @bcast, @rose={0xbb, 0xbb, 0xbb, 0x1, 0x0}, @bcast, @netrom={0xbb, 0xbb, 0xbb, 0xbb, 0xbb, 0x0, 0x0}, @null, @bcast]}, 0x48) 04:12:45 executing program 4: r0 = socket$inet6(0xa, 0x2, 0x0) r1 = socket$inet6(0xa, 0x80002, 0x88) bind$inet6(r1, &(0x7f0000000580)={0xa, 0x10010000004e20}, 0x1c) setsockopt$inet6_udp_int(r1, 0x11, 0xb, &(0x7f0000000080), 0x4) r2 = socket$inet_udplite(0x2, 0x2, 0x88) ioctl$sock_SIOCGIFINDEX(r2, 0x8933, &(0x7f0000000340)={'bridge_slave_0\x00', 0x0}) r4 = socket(0x10, 0x80002, 0x0) sendmsg$nl_route(r4, &(0x7f0000000000)={0x0, 0x0, &(0x7f0000000080)={&(0x7f00000000c0)=@bridge_setlink={0x2c, 0x13, 0xa29, 0x0, 0x0, {0x7, 0x0, 0x0, r3}, [@IFLA_AF_SPEC={0xc, 0x1a, 0x0, 0x1, [@AF_INET={0x8, 0x4, 0x0, 0x1, {0x4, 0x3}}]}]}, 0x2c}}, 0x0) sendmmsg$inet6(r1, &(0x7f0000001dc0)=[{{&(0x7f0000000040)={0xa, 0x4e20, 0x9, @remote, 0x5}, 0x1c, &(0x7f0000000080)=[{&(0x7f0000000240)="dd1ce36b160f3507faaa3ce9d79c64d3afdb297ae040fe7e97eb2d48bd5575e16940c44459dbf315bf6df71a4c309a954eb91b9469a8280aaa31aa1117b6013d2561e4b019ac80f40c31177d73762d1ddaaf0cebd1aa3385d4500b271a1dd0ea0a0d89f0ad7c7fd515fba4c3496ab0c806bd5615975c4da653d6835d176874ed1a455a2efcc819007c84cbb0e235c6fb168077e2190cded56f5f1e1c49c2c303af1dc122ed60b8eca662d413293b9f20f6892372c699fc1b1dc4cfbd21d6a40ad06a68f364a96c80c65aa460374fed63bdf621cdb4f5232887273011ff7af41a9adc68e70623cd90ced64a908f9b9a55013c62194b4cb76494534cbd9762f351bdcf56874a8654d55b26375082fe116548aa72e4bc38452c6ac93aa54b8834a6d9992e9b38037204e66a325d4dffe5a9100d0e845984d3bc542876680f458c6df9046f992f437553e7fd90a0c911a02f6efcc3591ece9427e9784dac0a8e802649846804ef8331dbbd4afd106d4057034f32a008887a252907244edb2fd33c970989b7f76b9ce096d045aada4109111134e2f8dcb48926e122ed706d0b96d9970f9349f2b2d9a7a9b98a08a48651bc65161b1c67c3324eaee83c7d8ff0f99be6182770381e9151899a2858ab89bb5ae606e7188cd29feb2eee4cb41e93cad00d2a3cfad5d0abdc7cebc69d4414205bb3f2ebf6d88b3cee8e8a46fc4523d13763c63866e9b5a89926b1ced6cd3c0a24638b1a4658f168cb45bb0765c18058596a6eb3669dc6eb243c4325ea3b4532cf7ca8e8446144821d54120d36812f4d6b4d63537b04b3ddeaf583a71a090c2fd6c37879d234b43abcc77543b6e2df98ee7e4c5268d96954d68ed1256224486e61300272fe4944ee01c9c7ea92ba66e321c0441377fae47c8fe596ff3f9b0cc9698a51251e219f93a78ac77b57c2904346367219931189357e32133a6ae4e439b3b4471c7f1b9ce55e7cf9c5b123c2573cc6739f0ccae3430763aa8cffc7faa1c2a1a36f3d07946b38a4601e856f13fea9742c60161455ec9f68b3b9fc965359c6e68da11f13e2ae814cdb09e2e706d8230bd1b40ee439442995f05f5a2580c3f12dede9ab2c82b2c4b7eb808bf7a3e114f5cdf30eab015ee991a05f679c106ebb176090801a92e913ab647436e9b6394c695811d04b851212eb2b5d9dc51cc9a26d89c56cc7506a427e49d065b16c00819493cf011dbb090f17eeac9339e0f1fd3728be362411ae8040580bcaadac7d143bc9310e364db7f2e734f948dfba4b8c32cd5398d17fb739e40d0c674ee09936638fc0bf4d6515e68e87581b20d33172b5333673b1e01d5c67ab60e77472655ffc38b0216f9ecec11cc6f6d61f63195b8bf914a89444096c0e88f8682e9a3739cb2b5b66a8a3750209ea0af8c778250fde57243dcb9e7ccec2552327082d2df1d67ad793d1ed89573bd6e3ae6938c33d3a26ee1cd11687f211d26eee4a1e1ca90a555d22cabed33c5edf432c15b3fdc71bdde680e14aed8cd7dd557dfa67e17be6535a6651c66f65fe7201b84b0912ed3f554fe95c8774a19c3227a3c196532d68a6efae2270ea91f34c1f9da2bc91630caed927f386864eca1e536e14913a5b6ceb8e5bfd0f35eac465434710e07bf9bd230e9253b96a0b04f56a7a772d2698f08549584baf69adbc8c2816747eb75cafdc61f5738821dd5cd23e3afbb9c2afbe079c0fdc05b6ac2d8d0e8430e3d496616f36f9665da097880658b13c0285daa09f710f8c11d4ef020dc79374f803c28af1497d063b4b00ca8441cc890abf5c82f7ea36c6001caee1fcc0080b41e0e238f2faf330b64b055101880770c0758ab1da32b8ff151deb9c9dd135e9ebe0be4d799265b735f6789c7c65b340135c71f37cf4fad6cdea553f91daecd1542dc80988726e48ff47bb410854e8bf232d1dbff4633322a0b00991fb8c4c248488811b52521c31d534a17c74d8174083d1863100bc8529516f18ee2a85ef7f824c52bbfd6ca9eee723e14e389c2dfe492de552cff45a1865cb563762922f9741cf9422b5f211a9b5f1e8f68cf2da8274004bef7ede7fb07c9838cb0965771becb6fad77d3dc0218b86f7053957bb31adf80f60937d6d471d4f88cc814e386c888fb7a8f337438e434ab9fd0e7516740d2f66514bb097e21b22da8822a061a0ba7c758fb86b9c16e35c835a41e3357b71a0dc1be8146992ff7768cc48d7a86b8a1c0084c582c7fd09812c74940500d02024ace9a9df3b092c38ff62a7752c2c65233ac918b9bf98529bacaa63b8e219ff460342e11c445bd257839225b6bd2182616843ad103549aa92c2db00ead76a402e92d679741cda04129380c0a1edcb3d4aec6417819ecf16dd47f643a8805660fe33a61663bd6f5d4d215350f0e73710376d3ba1b850e18c0768885e0654c5eb6e383a116414819346e534162a949713d80df9c9a5d10536cb63571e777908564823e566d93bd057764b5d4880679cada7b6336e25d026b129547d36fcc66693207f3df14b83f2f8f106452524e9942b6ade43b340241f04f87eacc65d9b1a62693b7c2cc848a2cd908a1068c614184f07024b52e27ac73297006e25fa14aaf064bca0f6d40ed8d2f884393b287af8d3e2af9d3226e676542c0283a474ff36ecc1a532b34e7fbfb4017ca96b6bc423f9e17ff63cd44cfba7d8a8841b2c277e960c12427666104b67962e5fad35d4c34bfe7c0b69de2b52d07daf10e573e10128af3e695812576bba6b1dd8d02f79d4842808bb1885abbdb7198e2aa839af0a642df6619fd5dbf09296996415aa2ad94e6d8a8a8c2e95eab634c3590d2e6a17689a0a0eb64ecf22744fb98ddc0ef6fc55a90420510fc722f339c583f9767eb1c993f98527c83d36f20860b710669407352742352f02f439e3e1e7933700a75d95765a5d2132f1a1b73cc487113d7988d15c41b8baecead428c495c36824f8c1115f6795c8bf2a9817ba6382bd5d669c3aa80cafd8396352228bf693af529d88273a34fb4ec90901992eb425230614b9b361515e9e4f65631c385bbe8da0e735c4d6fc680aa6664b730fa4f15bc397b19a9791d8c53e631acd930bc365ad05f5e8f7e825eba4e7588acfe0568611a046246278562217aaaa0766168091b348791b876bead75de07d153babb3e0665a326bcd71a472d51e86835471994fec156b07e04303ffa8089afc49b484fbfaf9a86255fe218f541ee500bb45fc355d6557fb884f1a840f7f256f158d1673e1ad8fe837ce9b3e77aaca3d5b90bfbd642c7f766e61549944848780c815febdcb97ba35a2702303f338286c1a94e2ff8b46d0d0f541a891f5bd424a7321de44d0c1a568c792d6bb99e6dbc16b389a7e83c3fabcb9e5fefb5b2c5c4329d01bb898e6c8f2067ee73c73e90dfa18da7a82bf217723efdd7d27e6f132f3a648c485a1b9363dc7224e4b846e99e3737ec9eb6f6d8816563887acd317361251cd46ece0a6d9d797951b4cd21b3b848d5384cb85488c08ca952ca5a1d65fab840f81ec56d6425a9a7da8be24588856bdbfc7df08e5f13f7e10e80a8220851db1b41fcf23b0616a83ecace12bfbd917ac5a9af8dcfc2258a1be5634e7c0d8511d94162caf32e9800343689f30cc373ffd3f354f23ec11bcdf986e711a683d60d6bfa66ae83d7aec00f9ee822e2e3abc4322232c27e3ff543fb12772d216ac7726a770f1f6879c79d41693b85917b17b10f71b54c3828b38624dba9ff9cfd9e322234800128bf921ffee1501f0940f3d2bbc081c5ff67c3f4c63a01a2d14a04abea0ecb8fc47f2d365a7986a12bb1af7896ab1e97d62ca96a55deb0e751c855446f53c27520ba831230610c1102639c109e47c988968ce140c92e4e555d3b46a38b2778ae03efbed5a28337b0a04f39707439d22d38b9b185702874f2570cb98faaf1b8a77ba7f88d056d5266e9ced1828751accd418985952a9d851c98485a9f703e0ac98232084d1f7e16ede91c73f679e25c974284d71ecee1bc580ade172ebb1209caa5558df2d830b4390a0eb92d26f0f937f53455a3003f9978bc92b8471c71d1e26d046555361417b91b81768bd23ddce564e8224c8ef9aa2387f87c709266bfd01526ac207fe8b324ac30c3474e1e1ad2c92016883548859867814d8522979d4b31a2c1fb0706693525581badea1a077fecc18e8150d398d2854068e1868af81ab23cb92287b1984f550def18ce074c0eb4cc8611721016ced6c696292259d147f525ad493f3ce57ce63cecd44b33536ab5762eacbd7501a6d3df899e0ccc83edb0ac9633aa64d0bbdbf533f16c49a8bf8ffd0d6c05803a93da4af3ee56f61fa72aef02fee6df5aa6c040812fc7cce65f4242cea1afd9d82443e3d4b4fe98a7ba0b783e7ee94ad3f1d08039f32b29029819b94d6ef158e8d6316be9d47f13e3243e1bdc946dde5249264e0d7df0bbffa1a4828e714a996283c6a0935fcd34f8d0040ce8668308eca6f6fbe85063f768ee9130e4a5b7570a69a8723e2f41a4bc9eba67f016171d6b787c13458fb7655a1b7b26f6cd0d7e339aaffcda7624f9396ff9b45edfea9c593a5d9d0168534bd7203365c2edfe21b1253ce340114d4e1874be56dc8502c23f201fda9951f9a41fc0bfc92a738330efc6c6eae8d6d316b317e9e2bc792634162d58734820492b646a0a6a06ec126b5c2284313993452ebed96a031f3dcd11913cf7766b8236208ea6240c3fcc2d3524b7d216f41c47307aea4f01286134e78a88738d6448eec0b5e073bcfdaea76e3d2afe9a0fe4f4d3c08e30124652dc43cfd2a502fe61261bd134bdb3a10721f2eaea8170dd8dabfb7581d3acdfadff51774ba03a971d3229cc4b54db1bb7583f60f7851e2a93627db21e442f33ccd66b90047861af55203adcb1c3001af055a096180b346a62019947c71fe08c2ce115374f1c1678aca401deae482bf0edc994cfd122da0d93df031bfc56ef9642d6b1465d66003a9e2f5820d194c4df01cd885a96a16d5fcbe986ef416371b42179fc76052ee298d2f0ea96afd968740d4ccf16ab7db154deab314616a6c38c9e9daaaaa1306349f7aa913d7395f024815de3d067f39d9f9ddc077873dbd1556f0912a20ae4cefb3f4faaab30506dade5d6369e6e580d2b610572ae768d09f0a813782a1920228b7c2e9d83508b3d7e5bd306d79a2da8f1455745b30c23c0b297e67c99bb3b12a0954a924ac7d4925e7f77c785cb3f74ff85a5e9e3c7b38359f4f22c314177cf1155cc7505cbc7919e2525cf41c16052461e69036f1acdafa488dea7f9bc6608a25b1df13aec49ae51c6b643a8fdffaf6645d1acd1e2ab309283f6b64bc6dd1b2a718b37e1ad2d4b854d99d12afedfa42f7c42ebf6db4994eaf62f28f0d6b9f6ed172143926a785a3eb19fad8715f1f958cddc2a4a805438485ea9a5f974387f29a81a7261898c69b18ac04edfe79480849745b305ebca727537438d13f0328cef8931fbd8284cf644d89ff8d978f5dda5de5602b285230cd6d419fc63f6d1a02d649c69ecb8d093e603036faaae4a59b5d0aaed76fd539a6b4b7d7433df00893b3775d5fbe718539cdfcd0b247bc22f2ef7d5c58c99c8fd039367024ff94893bdee566bda6784a131f1633f572fdb87b406299cf345ee1de46a78e591f6d45eb37c2a4a481cce6eb6c67bb9928b3d33965e4693831e99194c41283d8f493886cca86fade836bdb6e7eac357adc0d507be9cff4d4635dc61086ba939d593607f094fb72fadc9eb648540a6b54", 0x1000}], 0x1, &(0x7f0000001240)=[@hopopts={{0xa0, 0x29, 0x36, {0x3c, 0x10, '\x00', [@enc_lim={0x4, 0x1, 0x4}, @enc_lim={0x4, 0x1, 0x2}, @generic={0x81, 0x60, "1f369872c4e4612f8bb266eb86912ac884c06ac35e7cf1eb3ff6684aab0a98432cf9e23c55d9248c910213d4b99f117bed505fe085b9ea096b8eca29bfd971a23933097d1b8268ec8a5d280837c5d2283adbe00eb1184416071d8510ced881bb"}, @ra={0x5, 0x2, 0x7245}, @hao={0xc9, 0x10, @mcast1}, @ra={0x5, 0x2, 0x6}, @enc_lim={0x4, 0x1, 0x1}]}}}, @hoplimit={{0x14, 0x29, 0x34, 0xffffffff}}, @rthdr={{0x98, 0x29, 0x39, {0x21, 0x10, 0x0, 0x4, 0x0, [@loopback, @rand_addr=' \x01\x00', @private0, @remote, @ipv4={'\x00', '\xff\xff', @dev={0xac, 0x14, 0x14, 0x2b}}, @dev={0xfe, 0x80, '\x00', 0x40}, @private0={0xfc, 0x0, '\x00', 0x1}, @loopback]}}}, @dontfrag={{0x14, 0x29, 0x3e, 0x6}}, @rthdr_2292={{0x48, 0x29, 0x39, {0x1d, 0x6, 0x2, 0x0, 0x0, [@initdev={0xfe, 0x88, '\x00', 0x1, 0x0}, @remote, @private0]}}}], 0x1b0}}, {{&(0x7f00000000c0)={0xa, 0x4e23, 0x1, @private2={0xfc, 0x2, '\x00', 0x1}, 0x7ff}, 0x1c, &(0x7f0000001540)=[{&(0x7f0000000100)="b4d537fc4ec5292b087a9a19d73d33ee91cb83f20643778c7ae309cd39bf3fe60466967057bea074b619fe631de3fab9fa36bc064ddbc3ab97a1cb29ba5a5d8b2b76c6c2bf001b45c7cc0f36d06c435f5c3d99cb4d050d1de994b075c0cfd3de7e2d27d95956b962b871377c42b8f62bf5f46433e1eb44c395903270696cdfd46528edbe80be6d09bb471450618e9de4b0cd63dec325823d8913d5dadb2fb57786ca5142b92c10847aec01aaa4118eea7e73b951f8048db1d19e0bdc46ebf6a50c3a8677146fcdcd5bea67d2", 0xcc}, {&(0x7f0000001400)="48affc6589751527c1662de5caaea4cfbf91f195e2474ad43fc8708054af73bbbc4ff13cce1df94a922b7bc9b22f8da2046691d8b2baeeaab1ba61a62f9ee5dbcede38ae2b8fd731f3b102", 0x4b}, {&(0x7f0000001480)="b2abb19bc646600056bf615c6eb86d2affd0c6a19703efa75fae356f2814da02d32c4ec3668accc6fe2152959d31764d69a2cf8897e590f587aa7c12c279996195bfa37e93d60b9032673da89cfa9cfc52f99b623f97eb38d648a4a1afda82274d503957efd6a089bb10d15498a3062a097d5ac1b688cd4439a877744a123e0f3b003f1e", 0x84}], 0x3, &(0x7f0000001580)=[@hopopts={{0x38, 0x29, 0x36, {0x62, 0x3, '\x00', [@calipso={0x7, 0x18, {0x2, 0x4, 0x7f, 0x7, [0x3, 0xffff]}}, @ra={0x5, 0x2, 0x8001}]}}}, @pktinfo={{0x24, 0x29, 0x32, {@mcast2, r3}}}, @hoplimit_2292={{0x14, 0x29, 0x8, 0x20}}, @rthdrdstopts={{0x160, 0x29, 0x37, {0x5e, 0x28, '\x00', [@jumbo={0xc2, 0x4, 0x5}, @ra={0x5, 0x2, 0x9}, @calipso={0x7, 0x50, {0x2, 0x12, 0x2, 0x0, [0x7, 0x7, 0x8, 0xe6, 0x80000000, 0x4, 0x5a3, 0x5, 0x100000000]}}, @generic={0xf2, 0xe9, "6c24ee61412a83545ce26853175fc8054839e39e56d7c7807054dec0596a3f3ead7d12fbdf75e0eac3319c2d8ee18023c547511173ddb68acd3264b3a118c8a78c9c03f65bc4ec62e9e7879110e0aed4b80d752fd329a17e51bf207baf6229afc8416cb69b00acd5bc250abc1abb41005824e38b7fde7b7c4711eba3f4c6e0b02bac902f496cc9d75928e1fe598f2e4485b88c68f5378144d52269f43fb9ca90ba4fc295aa4022402efd99fdc37c54ddf3818d7b5c36d0b81b8247df839a365cfbe0f7116b7c21a5c1e2014710ea432eff5d3a65fdaa36c317f1ec7c0b97843afae20c9d7fd7d4bf70"}]}}}], 0x1d8}}, {{0x0, 0x0, &(0x7f0000001b00)=[{&(0x7f0000001780)="5feea81a24ebddeeac", 0x9}, {&(0x7f00000017c0)="95d527e3bd5b310da8bf16bd6fe8b85ef0371db049f969ec59147745ff0ef5fbb65f6a3094408c14f69f15a07f274db20b255251da84967f0ffa9fe50be037072bcc4411d9551a3150c0c5b10926966423d765455d65a17b", 0x58}, {&(0x7f0000001840)="bb6c6bf4a9151ae18e87006298fd33f59e346e44892d5ed19ad6157bacd297f597f9c7e3ab972069154bb109d8262c82b79a0de44798afc5947611828a6ac8160bdec7a40c49150226a16427f8210246cb3482d2f84bfe8a1d36b5ef5a4c2fbfb615a48ef5d3074f01c14145b42a2e7285e4dbc1a150f9eb8a38775652c6583fe8aaf0fb5040569d90bd6682120b947b31dc", 0x92}, {&(0x7f0000001900)="9b37a484f3851e1fe1cd8bdcc339374a22197cd54ba150aa1d45526bb63108f22c75aa24a93540934912923150b647163341d2aa123de3e9ac167cd2d0d71a19091513c3374818ea981fb781df772b4baec81e912e9c2148512eacdaa7e7fbe0d7f61aad6aa19de932a7157eca8794090a25ee4eb1825403bdb698e6184ed348c175e2e7d6c0e300eed5fa750f894c05e3979e0f485327768b4777e05b2a63113a79a08dda92031b4c03e30c85cc7298e8b0776d20131478e62fb578f4c54e351c29c3d7dfafd956cc4ba3c41cf4a24ea7ae273ae57722a5f64cdb94723bae4b108e", 0xe2}, {&(0x7f0000001a00)="e524d03409aaefcdf6c6659616d99bb6e508859c33221742d2d1f9aacf84e11676f3eb074215c007feabb1225502343f4247b4c922f0cc0ac0d3350d9f6d5ac9cc2ecb0754150a873879e776c6ac06208a2c52ffc63338f333", 0x59}, {&(0x7f0000001a80)="1e8401aa707748e6574c40e9c2c49625a1994dbd1a55b56a70f376f1a281e17c84d17e80d7b6", 0x26}, {&(0x7f0000001ac0)="1c594fe43f90b6ffd9d8099baf46a84865846782e5844128d638481b6ead1734fb98f40d498aa82ba5772863d4da", 0x2e}], 0x7, &(0x7f0000001b80)=[@dstopts_2292={{0x28, 0x29, 0x4, {0x32, 0x2, '\x00', [@pad1, @pad1, @jumbo={0xc2, 0x4, 0x3}, @ra={0x5, 0x2, 0x9}]}}}, @dstopts_2292={{0xf8, 0x29, 0x4, {0x2b, 0x1b, '\x00', [@jumbo={0xc2, 0x4, 0x4}, @calipso={0x7, 0x40, {0x1, 0xe, 0x2, 0xf463, [0xfffffffffffffb46, 0xfff, 0x8, 0x3ff, 0x0, 0x6, 0x8]}}, @padn={0x1, 0x5, [0x0, 0x0, 0x0, 0x0, 0x0]}, @hao={0xc9, 0x10, @empty}, @hao={0xc9, 0x10, @dev={0xfe, 0x80, '\x00', 0xd}}, @enc_lim={0x4, 0x1, 0x2}, @pad1, @hao={0xc9, 0x10, @initdev={0xfe, 0x88, '\x00', 0x1, 0x0}}, @calipso={0x7, 0x48, {0x0, 0x10, 0x1, 0x9, [0xc3, 0x1, 0x1, 0xa0f0, 0x7, 0x8, 0x5, 0x0]}}, @ra={0x5, 0x2, 0xfffe}]}}}, @hopopts_2292={{0x20, 0x29, 0x36, {0x2e, 0x0, '\x00', [@pad1]}}}, @dontfrag={{0x14, 0x29, 0x3e, 0x1}}, @pktinfo={{0x24, 0x29, 0x32, {@rand_addr=' \x01\x00\x00\x00\x00\x00\x00\x00\x00\x00\x00\x00\x00\x00\x02'}}}, @dstopts={{0xb0, 0x29, 0x37, {0x4a, 0x12, '\x00', [@jumbo={0xc2, 0x4, 0xfedd}, @calipso={0x7, 0x58, {0x1, 0x14, 0x6, 0x3, [0x9, 0x2, 0x1, 0x4, 0x401, 0x8, 0x9, 0x2, 0xb4, 0x8]}}, @generic={0xa5, 0x25, "3cbb5e9ca85727c64c56265f036fb83230d933e58f0bbb4135ac8436f962806661674c8891"}, @pad1, @enc_lim={0x4, 0x1, 0x7}, @ra={0x5, 0x2, 0x4}]}}}], 0x230}}], 0x3, 0x10) ioctl$sock_SIOCETHTOOL(r0, 0x89f0, &(0x7f0000000000)={'bridge0\x00', &(0x7f0000000200)=ANY=[@ANYBLOB="1000000000000000000000000000001000"/36]}) 04:12:45 executing program 3: r0 = socket$netlink(0x10, 0x3, 0x0) sendmsg$netlink(r0, &(0x7f00000047c0)={0x0, 0x0, &(0x7f0000000000)=[{&(0x7f0000000100)=ANY=[@ANYBLOB="18000000560001060000008400db730007"], 0x18}], 0x1}, 0x0) [ 2837.390814][ T3273] workqueue: Failed to create a rescuer kthread for wq "bond1287": -EINTR [ 2837.947560][ T3287] validate_nla: 3 callbacks suppressed [ 2837.947580][ T3287] netlink: 'syz-executor.1': attribute type 1 has an invalid length. 04:12:45 executing program 1: r0 = socket$netlink(0x10, 0x3, 0x0) r1 = socket$netlink(0x10, 0x3, 0x0) r2 = socket(0x10, 0x803, 0x0) sendmsg$NL80211_CMD_CRIT_PROTOCOL_START(r2, &(0x7f0000000580)={0x0, 0x0, &(0x7f0000000540)={0x0, 0x1c}}, 0x0) getsockname$packet(r2, &(0x7f0000000600)={0x11, 0x0, 0x0, 0x1, 0x0, 0x6, @broadcast}, &(0x7f0000000100)=0x14) sendmsg$nl_route(r1, &(0x7f0000000040)={0x0, 0x0, &(0x7f0000000000)={&(0x7f00000000c0)=ANY=[@ANYBLOB="3c00000010008506eb9afc4cd8d06e754a0081c5", @ANYRES32=r3, @ANYBLOB="2377f292252155b21c0012000c000100626f6e64000000000c0002000800010001"], 0x3c}}, 0x0) sendmsg$nl_route(r0, &(0x7f0000000240)={0x0, 0x0, &(0x7f0000000680)={&(0x7f0000000400)=@newlink={0x3c, 0x10, 0xffffff1f, 0x0, 0x1b672500, {}, [@IFLA_LINKINFO={0x14, 0x12, 0x0, 0x1, @bridge={{0xb}, {0x4}}}, @IFLA_MASTER={0x8, 0xa, r3}]}, 0x3c}}, 0x0) 04:12:45 executing program 0: r0 = syz_init_net_socket$ax25(0x3, 0x0, 0x0) bind$ax25(r0, &(0x7f0000000000)={{0x3, @remote={0xcc, 0xcc, 0xcc, 0xcc, 0xcc, 0xcc, 0x1}, 0x1}, [@null, @netrom={0xbb, 0xbb, 0xbb, 0xbb, 0xbb, 0x0, 0x0}, @bcast, @rose={0xbb, 0xbb, 0xbb, 0x1, 0x0}, @bcast, @netrom={0xbb, 0xbb, 0xbb, 0xbb, 0xbb, 0x0, 0x0}, @null, @bcast]}, 0x48) [ 2837.997081][ T3287] workqueue: Failed to create a rescuer kthread for wq "bond1304": -EINTR [ 2838.170356][ T3306] netlink: 'syz-executor.5': attribute type 1 has an invalid length. 04:12:45 executing program 0: r0 = syz_init_net_socket$ax25(0x3, 0x0, 0x0) bind$ax25(r0, &(0x7f0000000000)={{0x3, @remote={0xcc, 0xcc, 0xcc, 0xcc, 0xcc, 0xcc, 0x1}, 0x1}, [@null, @netrom={0xbb, 0xbb, 0xbb, 0xbb, 0xbb, 0x0, 0x0}, @bcast, @rose={0xbb, 0xbb, 0xbb, 0x1, 0x0}, @bcast, @netrom={0xbb, 0xbb, 0xbb, 0xbb, 0xbb, 0x0, 0x0}, @null, @bcast]}, 0x48) 04:12:45 executing program 0: r0 = syz_init_net_socket$ax25(0x3, 0x0, 0x0) bind$ax25(r0, &(0x7f0000000000)={{0x3, @remote={0xcc, 0xcc, 0xcc, 0xcc, 0xcc, 0xcc, 0x1}, 0x1}, [@null, @netrom={0xbb, 0xbb, 0xbb, 0xbb, 0xbb, 0x0, 0x0}, @bcast, @rose={0xbb, 0xbb, 0xbb, 0x1, 0x0}, @bcast, @netrom={0xbb, 0xbb, 0xbb, 0xbb, 0xbb, 0x0, 0x0}, @null, @bcast]}, 0x48) 04:12:45 executing program 0: syz_init_net_socket$ax25(0x3, 0x2, 0x0) bind$ax25(0xffffffffffffffff, &(0x7f0000000000)={{0x3, @remote={0xcc, 0xcc, 0xcc, 0xcc, 0xcc, 0xcc, 0x1}, 0x1}, [@null, @netrom={0xbb, 0xbb, 0xbb, 0xbb, 0xbb, 0x0, 0x0}, @bcast, @rose={0xbb, 0xbb, 0xbb, 0x1, 0x0}, @bcast, @netrom={0xbb, 0xbb, 0xbb, 0xbb, 0xbb, 0x0, 0x0}, @null, @bcast]}, 0x48) 04:12:45 executing program 0: syz_init_net_socket$ax25(0x3, 0x2, 0x0) bind$ax25(0xffffffffffffffff, &(0x7f0000000000)={{0x3, @remote={0xcc, 0xcc, 0xcc, 0xcc, 0xcc, 0xcc, 0x1}, 0x1}, [@null, @netrom={0xbb, 0xbb, 0xbb, 0xbb, 0xbb, 0x0, 0x0}, @bcast, @rose={0xbb, 0xbb, 0xbb, 0x1, 0x0}, @bcast, @netrom={0xbb, 0xbb, 0xbb, 0xbb, 0xbb, 0x0, 0x0}, @null, @bcast]}, 0x48) [ 2838.435585][ T3306] bond1296: entered promiscuous mode [ 2838.455945][ T3306] 8021q: adding VLAN 0 to HW filter on device bond1296 [ 2838.558356][ T3307] bond1296: (slave bridge1259): making interface the new active one [ 2838.584868][ T3307] bridge1259: entered promiscuous mode 04:12:46 executing program 0: syz_init_net_socket$ax25(0x3, 0x2, 0x0) bind$ax25(0xffffffffffffffff, &(0x7f0000000000)={{0x3, @remote={0xcc, 0xcc, 0xcc, 0xcc, 0xcc, 0xcc, 0x1}, 0x1}, [@null, @netrom={0xbb, 0xbb, 0xbb, 0xbb, 0xbb, 0x0, 0x0}, @bcast, @rose={0xbb, 0xbb, 0xbb, 0x1, 0x0}, @bcast, @netrom={0xbb, 0xbb, 0xbb, 0xbb, 0xbb, 0x0, 0x0}, @null, @bcast]}, 0x48) [ 2838.614845][ T3307] bond1296: (slave bridge1259): Enslaving as an active interface with an up link [ 2838.648002][ T3315] netlink: 'syz-executor.2': attribute type 1 has an invalid length. 04:12:46 executing program 5: r0 = socket$netlink(0x10, 0x3, 0x0) r1 = socket$netlink(0x10, 0x3, 0x0) r2 = socket(0x10, 0x803, 0x0) sendmsg$NL80211_CMD_CRIT_PROTOCOL_START(r2, &(0x7f0000000580)={0x0, 0x0, &(0x7f0000000540)={0x0, 0x1c}}, 0x0) getsockname$packet(r2, &(0x7f0000000600)={0x11, 0x0, 0x0, 0x1, 0x0, 0x6, @broadcast}, &(0x7f0000000100)=0x14) sendmsg$nl_route(r1, &(0x7f0000000040)={0x0, 0x0, &(0x7f0000000000)={&(0x7f00000000c0)=ANY=[@ANYBLOB="3c00000010008506eb9afc4cd8d06e754a0081c5", @ANYRES32=r3, @ANYBLOB="2377f292252155b21c0012000c000100626f6e64000000000c0002000800010001"], 0x3c}}, 0x0) sendmsg$nl_route(r0, &(0x7f0000000240)={0x0, 0x0, &(0x7f0000000680)={&(0x7f0000000400)=@newlink={0x3c, 0x10, 0xffffff1f, 0x0, 0x12000000, {}, [@IFLA_LINKINFO={0x14, 0x12, 0x0, 0x1, @bridge={{0xb}, {0x4}}}, @IFLA_MASTER={0x8, 0xa, r3}]}, 0x3c}}, 0x0) [ 2838.776353][ T3315] bond1287: entered promiscuous mode [ 2838.791177][ T3315] 8021q: adding VLAN 0 to HW filter on device bond1287 [ 2838.975247][ T3318] bond1287: (slave bridge1208): making interface the new active one [ 2838.992073][ T3318] bridge1208: entered promiscuous mode 04:12:46 executing program 4: r0 = socket$inet6(0xa, 0x2, 0x0) r1 = socket$inet6(0xa, 0x80002, 0x88) bind$inet6(r1, &(0x7f0000000580)={0xa, 0x10010000004e20}, 0x1c) setsockopt$inet6_udp_int(r1, 0x11, 0xb, &(0x7f0000000080), 0x4) r2 = socket$inet_udplite(0x2, 0x2, 0x88) ioctl$sock_SIOCGIFINDEX(r2, 0x8933, &(0x7f0000000340)={'bridge_slave_0\x00', 0x0}) r4 = socket(0x10, 0x80002, 0x0) sendmsg$nl_route(r4, &(0x7f0000000000)={0x0, 0x0, &(0x7f0000000080)={&(0x7f00000000c0)=@bridge_setlink={0x2c, 0x13, 0xa29, 0x0, 0x0, {0x7, 0x0, 0x0, r3}, [@IFLA_AF_SPEC={0xc, 0x1a, 0x0, 0x1, [@AF_INET={0x8, 0x4, 0x0, 0x1, {0x4, 0x3}}]}]}, 0x2c}}, 0x0) sendmmsg$inet6(r1, &(0x7f0000001dc0)=[{{&(0x7f0000000040)={0xa, 0x4e20, 0x9, @remote, 0x5}, 0x1c, &(0x7f0000000080)=[{&(0x7f0000000240)="dd1ce36b160f3507faaa3ce9d79c64d3afdb297ae040fe7e97eb2d48bd5575e16940c44459dbf315bf6df71a4c309a954eb91b9469a8280aaa31aa1117b6013d2561e4b019ac80f40c31177d73762d1ddaaf0cebd1aa3385d4500b271a1dd0ea0a0d89f0ad7c7fd515fba4c3496ab0c806bd5615975c4da653d6835d176874ed1a455a2efcc819007c84cbb0e235c6fb168077e2190cded56f5f1e1c49c2c303af1dc122ed60b8eca662d413293b9f20f6892372c699fc1b1dc4cfbd21d6a40ad06a68f364a96c80c65aa460374fed63bdf621cdb4f5232887273011ff7af41a9adc68e70623cd90ced64a908f9b9a55013c62194b4cb76494534cbd9762f351bdcf56874a8654d55b26375082fe116548aa72e4bc38452c6ac93aa54b8834a6d9992e9b38037204e66a325d4dffe5a9100d0e845984d3bc542876680f458c6df9046f992f437553e7fd90a0c911a02f6efcc3591ece9427e9784dac0a8e802649846804ef8331dbbd4afd106d4057034f32a008887a252907244edb2fd33c970989b7f76b9ce096d045aada4109111134e2f8dcb48926e122ed706d0b96d9970f9349f2b2d9a7a9b98a08a48651bc65161b1c67c3324eaee83c7d8ff0f99be6182770381e9151899a2858ab89bb5ae606e7188cd29feb2eee4cb41e93cad00d2a3cfad5d0abdc7cebc69d4414205bb3f2ebf6d88b3cee8e8a46fc4523d13763c63866e9b5a89926b1ced6cd3c0a24638b1a4658f168cb45bb0765c18058596a6eb3669dc6eb243c4325ea3b4532cf7ca8e8446144821d54120d36812f4d6b4d63537b04b3ddeaf583a71a090c2fd6c37879d234b43abcc77543b6e2df98ee7e4c5268d96954d68ed1256224486e61300272fe4944ee01c9c7ea92ba66e321c0441377fae47c8fe596ff3f9b0cc9698a51251e219f93a78ac77b57c2904346367219931189357e32133a6ae4e439b3b4471c7f1b9ce55e7cf9c5b123c2573cc6739f0ccae3430763aa8cffc7faa1c2a1a36f3d07946b38a4601e856f13fea9742c60161455ec9f68b3b9fc965359c6e68da11f13e2ae814cdb09e2e706d8230bd1b40ee439442995f05f5a2580c3f12dede9ab2c82b2c4b7eb808bf7a3e114f5cdf30eab015ee991a05f679c106ebb176090801a92e913ab647436e9b6394c695811d04b851212eb2b5d9dc51cc9a26d89c56cc7506a427e49d065b16c00819493cf011dbb090f17eeac9339e0f1fd3728be362411ae8040580bcaadac7d143bc9310e364db7f2e734f948dfba4b8c32cd5398d17fb739e40d0c674ee09936638fc0bf4d6515e68e87581b20d33172b5333673b1e01d5c67ab60e77472655ffc38b0216f9ecec11cc6f6d61f63195b8bf914a89444096c0e88f8682e9a3739cb2b5b66a8a3750209ea0af8c778250fde57243dcb9e7ccec2552327082d2df1d67ad793d1ed89573bd6e3ae6938c33d3a26ee1cd11687f211d26eee4a1e1ca90a555d22cabed33c5edf432c15b3fdc71bdde680e14aed8cd7dd557dfa67e17be6535a6651c66f65fe7201b84b0912ed3f554fe95c8774a19c3227a3c196532d68a6efae2270ea91f34c1f9da2bc91630caed927f386864eca1e536e14913a5b6ceb8e5bfd0f35eac465434710e07bf9bd230e9253b96a0b04f56a7a772d2698f08549584baf69adbc8c2816747eb75cafdc61f5738821dd5cd23e3afbb9c2afbe079c0fdc05b6ac2d8d0e8430e3d496616f36f9665da097880658b13c0285daa09f710f8c11d4ef020dc79374f803c28af1497d063b4b00ca8441cc890abf5c82f7ea36c6001caee1fcc0080b41e0e238f2faf330b64b055101880770c0758ab1da32b8ff151deb9c9dd135e9ebe0be4d799265b735f6789c7c65b340135c71f37cf4fad6cdea553f91daecd1542dc80988726e48ff47bb410854e8bf232d1dbff4633322a0b00991fb8c4c248488811b52521c31d534a17c74d8174083d1863100bc8529516f18ee2a85ef7f824c52bbfd6ca9eee723e14e389c2dfe492de552cff45a1865cb563762922f9741cf9422b5f211a9b5f1e8f68cf2da8274004bef7ede7fb07c9838cb0965771becb6fad77d3dc0218b86f7053957bb31adf80f60937d6d471d4f88cc814e386c888fb7a8f337438e434ab9fd0e7516740d2f66514bb097e21b22da8822a061a0ba7c758fb86b9c16e35c835a41e3357b71a0dc1be8146992ff7768cc48d7a86b8a1c0084c582c7fd09812c74940500d02024ace9a9df3b092c38ff62a7752c2c65233ac918b9bf98529bacaa63b8e219ff460342e11c445bd257839225b6bd2182616843ad103549aa92c2db00ead76a402e92d679741cda04129380c0a1edcb3d4aec6417819ecf16dd47f643a8805660fe33a61663bd6f5d4d215350f0e73710376d3ba1b850e18c0768885e0654c5eb6e383a116414819346e534162a949713d80df9c9a5d10536cb63571e777908564823e566d93bd057764b5d4880679cada7b6336e25d026b129547d36fcc66693207f3df14b83f2f8f106452524e9942b6ade43b340241f04f87eacc65d9b1a62693b7c2cc848a2cd908a1068c614184f07024b52e27ac73297006e25fa14aaf064bca0f6d40ed8d2f884393b287af8d3e2af9d3226e676542c0283a474ff36ecc1a532b34e7fbfb4017ca96b6bc423f9e17ff63cd44cfba7d8a8841b2c277e960c12427666104b67962e5fad35d4c34bfe7c0b69de2b52d07daf10e573e10128af3e695812576bba6b1dd8d02f79d4842808bb1885abbdb7198e2aa839af0a642df6619fd5dbf09296996415aa2ad94e6d8a8a8c2e95eab634c3590d2e6a17689a0a0eb64ecf22744fb98ddc0ef6fc55a90420510fc722f339c583f9767eb1c993f98527c83d36f20860b710669407352742352f02f439e3e1e7933700a75d95765a5d2132f1a1b73cc487113d7988d15c41b8baecead428c495c36824f8c1115f6795c8bf2a9817ba6382bd5d669c3aa80cafd8396352228bf693af529d88273a34fb4ec90901992eb425230614b9b361515e9e4f65631c385bbe8da0e735c4d6fc680aa6664b730fa4f15bc397b19a9791d8c53e631acd930bc365ad05f5e8f7e825eba4e7588acfe0568611a046246278562217aaaa0766168091b348791b876bead75de07d153babb3e0665a326bcd71a472d51e86835471994fec156b07e04303ffa8089afc49b484fbfaf9a86255fe218f541ee500bb45fc355d6557fb884f1a840f7f256f158d1673e1ad8fe837ce9b3e77aaca3d5b90bfbd642c7f766e61549944848780c815febdcb97ba35a2702303f338286c1a94e2ff8b46d0d0f541a891f5bd424a7321de44d0c1a568c792d6bb99e6dbc16b389a7e83c3fabcb9e5fefb5b2c5c4329d01bb898e6c8f2067ee73c73e90dfa18da7a82bf217723efdd7d27e6f132f3a648c485a1b9363dc7224e4b846e99e3737ec9eb6f6d8816563887acd317361251cd46ece0a6d9d797951b4cd21b3b848d5384cb85488c08ca952ca5a1d65fab840f81ec56d6425a9a7da8be24588856bdbfc7df08e5f13f7e10e80a8220851db1b41fcf23b0616a83ecace12bfbd917ac5a9af8dcfc2258a1be5634e7c0d8511d94162caf32e9800343689f30cc373ffd3f354f23ec11bcdf986e711a683d60d6bfa66ae83d7aec00f9ee822e2e3abc4322232c27e3ff543fb12772d216ac7726a770f1f6879c79d41693b85917b17b10f71b54c3828b38624dba9ff9cfd9e322234800128bf921ffee1501f0940f3d2bbc081c5ff67c3f4c63a01a2d14a04abea0ecb8fc47f2d365a7986a12bb1af7896ab1e97d62ca96a55deb0e751c855446f53c27520ba831230610c1102639c109e47c988968ce140c92e4e555d3b46a38b2778ae03efbed5a28337b0a04f39707439d22d38b9b185702874f2570cb98faaf1b8a77ba7f88d056d5266e9ced1828751accd418985952a9d851c98485a9f703e0ac98232084d1f7e16ede91c73f679e25c974284d71ecee1bc580ade172ebb1209caa5558df2d830b4390a0eb92d26f0f937f53455a3003f9978bc92b8471c71d1e26d046555361417b91b81768bd23ddce564e8224c8ef9aa2387f87c709266bfd01526ac207fe8b324ac30c3474e1e1ad2c92016883548859867814d8522979d4b31a2c1fb0706693525581badea1a077fecc18e8150d398d2854068e1868af81ab23cb92287b1984f550def18ce074c0eb4cc8611721016ced6c696292259d147f525ad493f3ce57ce63cecd44b33536ab5762eacbd7501a6d3df899e0ccc83edb0ac9633aa64d0bbdbf533f16c49a8bf8ffd0d6c05803a93da4af3ee56f61fa72aef02fee6df5aa6c040812fc7cce65f4242cea1afd9d82443e3d4b4fe98a7ba0b783e7ee94ad3f1d08039f32b29029819b94d6ef158e8d6316be9d47f13e3243e1bdc946dde5249264e0d7df0bbffa1a4828e714a996283c6a0935fcd34f8d0040ce8668308eca6f6fbe85063f768ee9130e4a5b7570a69a8723e2f41a4bc9eba67f016171d6b787c13458fb7655a1b7b26f6cd0d7e339aaffcda7624f9396ff9b45edfea9c593a5d9d0168534bd7203365c2edfe21b1253ce340114d4e1874be56dc8502c23f201fda9951f9a41fc0bfc92a738330efc6c6eae8d6d316b317e9e2bc792634162d58734820492b646a0a6a06ec126b5c2284313993452ebed96a031f3dcd11913cf7766b8236208ea6240c3fcc2d3524b7d216f41c47307aea4f01286134e78a88738d6448eec0b5e073bcfdaea76e3d2afe9a0fe4f4d3c08e30124652dc43cfd2a502fe61261bd134bdb3a10721f2eaea8170dd8dabfb7581d3acdfadff51774ba03a971d3229cc4b54db1bb7583f60f7851e2a93627db21e442f33ccd66b90047861af55203adcb1c3001af055a096180b346a62019947c71fe08c2ce115374f1c1678aca401deae482bf0edc994cfd122da0d93df031bfc56ef9642d6b1465d66003a9e2f5820d194c4df01cd885a96a16d5fcbe986ef416371b42179fc76052ee298d2f0ea96afd968740d4ccf16ab7db154deab314616a6c38c9e9daaaaa1306349f7aa913d7395f024815de3d067f39d9f9ddc077873dbd1556f0912a20ae4cefb3f4faaab30506dade5d6369e6e580d2b610572ae768d09f0a813782a1920228b7c2e9d83508b3d7e5bd306d79a2da8f1455745b30c23c0b297e67c99bb3b12a0954a924ac7d4925e7f77c785cb3f74ff85a5e9e3c7b38359f4f22c314177cf1155cc7505cbc7919e2525cf41c16052461e69036f1acdafa488dea7f9bc6608a25b1df13aec49ae51c6b643a8fdffaf6645d1acd1e2ab309283f6b64bc6dd1b2a718b37e1ad2d4b854d99d12afedfa42f7c42ebf6db4994eaf62f28f0d6b9f6ed172143926a785a3eb19fad8715f1f958cddc2a4a805438485ea9a5f974387f29a81a7261898c69b18ac04edfe79480849745b305ebca727537438d13f0328cef8931fbd8284cf644d89ff8d978f5dda5de5602b285230cd6d419fc63f6d1a02d649c69ecb8d093e603036faaae4a59b5d0aaed76fd539a6b4b7d7433df00893b3775d5fbe718539cdfcd0b247bc22f2ef7d5c58c99c8fd039367024ff94893bdee566bda6784a131f1633f572fdb87b406299cf345ee1de46a78e591f6d45eb37c2a4a481cce6eb6c67bb9928b3d33965e4693831e99194c41283d8f493886cca86fade836bdb6e7eac357adc0d507be9cff4d4635dc61086ba939d593607f094fb72fadc9eb648540a6b54", 0x1000}], 0x1, &(0x7f0000001240)=[@hopopts={{0xa0, 0x29, 0x36, {0x3c, 0x10, '\x00', [@enc_lim={0x4, 0x1, 0x4}, @enc_lim={0x4, 0x1, 0x2}, @generic={0x81, 0x60, "1f369872c4e4612f8bb266eb86912ac884c06ac35e7cf1eb3ff6684aab0a98432cf9e23c55d9248c910213d4b99f117bed505fe085b9ea096b8eca29bfd971a23933097d1b8268ec8a5d280837c5d2283adbe00eb1184416071d8510ced881bb"}, @ra={0x5, 0x2, 0x7245}, @hao={0xc9, 0x10, @mcast1}, @ra={0x5, 0x2, 0x6}, @enc_lim={0x4, 0x1, 0x1}]}}}, @hoplimit={{0x14, 0x29, 0x34, 0xffffffff}}, @rthdr={{0x98, 0x29, 0x39, {0x21, 0x10, 0x0, 0x4, 0x0, [@loopback, @rand_addr=' \x01\x00', @private0, @remote, @ipv4={'\x00', '\xff\xff', @dev={0xac, 0x14, 0x14, 0x2b}}, @dev={0xfe, 0x80, '\x00', 0x40}, @private0={0xfc, 0x0, '\x00', 0x1}, @loopback]}}}, @dontfrag={{0x14, 0x29, 0x3e, 0x6}}, @rthdr_2292={{0x48, 0x29, 0x39, {0x1d, 0x6, 0x2, 0x0, 0x0, [@initdev={0xfe, 0x88, '\x00', 0x1, 0x0}, @remote, @private0]}}}], 0x1b0}}, {{&(0x7f00000000c0)={0xa, 0x4e23, 0x1, @private2={0xfc, 0x2, '\x00', 0x1}, 0x7ff}, 0x1c, &(0x7f0000001540)=[{&(0x7f0000000100)="b4d537fc4ec5292b087a9a19d73d33ee91cb83f20643778c7ae309cd39bf3fe60466967057bea074b619fe631de3fab9fa36bc064ddbc3ab97a1cb29ba5a5d8b2b76c6c2bf001b45c7cc0f36d06c435f5c3d99cb4d050d1de994b075c0cfd3de7e2d27d95956b962b871377c42b8f62bf5f46433e1eb44c395903270696cdfd46528edbe80be6d09bb471450618e9de4b0cd63dec325823d8913d5dadb2fb57786ca5142b92c10847aec01aaa4118eea7e73b951f8048db1d19e0bdc46ebf6a50c3a8677146fcdcd5bea67d2", 0xcc}, {&(0x7f0000001400)="48affc6589751527c1662de5caaea4cfbf91f195e2474ad43fc8708054af73bbbc4ff13cce1df94a922b7bc9b22f8da2046691d8b2baeeaab1ba61a62f9ee5dbcede38ae2b8fd731f3b102", 0x4b}, {&(0x7f0000001480)="b2abb19bc646600056bf615c6eb86d2affd0c6a19703efa75fae356f2814da02d32c4ec3668accc6fe2152959d31764d69a2cf8897e590f587aa7c12c279996195bfa37e93d60b9032673da89cfa9cfc52f99b623f97eb38d648a4a1afda82274d503957efd6a089bb10d15498a3062a097d5ac1b688cd4439a877744a123e0f3b003f1e", 0x84}], 0x3, &(0x7f0000001580)=[@hopopts={{0x38, 0x29, 0x36, {0x62, 0x3, '\x00', [@calipso={0x7, 0x18, {0x2, 0x4, 0x7f, 0x7, [0x3, 0xffff]}}, @ra={0x5, 0x2, 0x8001}]}}}, @pktinfo={{0x24, 0x29, 0x32, {@mcast2, r3}}}, @hoplimit_2292={{0x14, 0x29, 0x8, 0x20}}, @rthdrdstopts={{0x160, 0x29, 0x37, {0x5e, 0x28, '\x00', [@jumbo={0xc2, 0x4, 0x5}, @ra={0x5, 0x2, 0x9}, @calipso={0x7, 0x50, {0x2, 0x12, 0x2, 0x0, [0x7, 0x7, 0x8, 0xe6, 0x80000000, 0x4, 0x5a3, 0x5, 0x100000000]}}, @generic={0xf2, 0xe9, "6c24ee61412a83545ce26853175fc8054839e39e56d7c7807054dec0596a3f3ead7d12fbdf75e0eac3319c2d8ee18023c547511173ddb68acd3264b3a118c8a78c9c03f65bc4ec62e9e7879110e0aed4b80d752fd329a17e51bf207baf6229afc8416cb69b00acd5bc250abc1abb41005824e38b7fde7b7c4711eba3f4c6e0b02bac902f496cc9d75928e1fe598f2e4485b88c68f5378144d52269f43fb9ca90ba4fc295aa4022402efd99fdc37c54ddf3818d7b5c36d0b81b8247df839a365cfbe0f7116b7c21a5c1e2014710ea432eff5d3a65fdaa36c317f1ec7c0b97843afae20c9d7fd7d4bf70"}]}}}], 0x1d8}}, {{0x0, 0x0, &(0x7f0000001b00)=[{&(0x7f0000001780)="5feea81a24ebddeeac", 0x9}, {&(0x7f00000017c0)="95d527e3bd5b310da8bf16bd6fe8b85ef0371db049f969ec59147745ff0ef5fbb65f6a3094408c14f69f15a07f274db20b255251da84967f0ffa9fe50be037072bcc4411d9551a3150c0c5b10926966423d765455d65a17b", 0x58}, {&(0x7f0000001840)="bb6c6bf4a9151ae18e87006298fd33f59e346e44892d5ed19ad6157bacd297f597f9c7e3ab972069154bb109d8262c82b79a0de44798afc5947611828a6ac8160bdec7a40c49150226a16427f8210246cb3482d2f84bfe8a1d36b5ef5a4c2fbfb615a48ef5d3074f01c14145b42a2e7285e4dbc1a150f9eb8a38775652c6583fe8aaf0fb5040569d90bd6682120b947b31dc", 0x92}, {&(0x7f0000001900)="9b37a484f3851e1fe1cd8bdcc339374a22197cd54ba150aa1d45526bb63108f22c75aa24a93540934912923150b647163341d2aa123de3e9ac167cd2d0d71a19091513c3374818ea981fb781df772b4baec81e912e9c2148512eacdaa7e7fbe0d7f61aad6aa19de932a7157eca8794090a25ee4eb1825403bdb698e6184ed348c175e2e7d6c0e300eed5fa750f894c05e3979e0f485327768b4777e05b2a63113a79a08dda92031b4c03e30c85cc7298e8b0776d20131478e62fb578f4c54e351c29c3d7dfafd956cc4ba3c41cf4a24ea7ae273ae57722a5f64cdb94723bae4b108e", 0xe2}, {&(0x7f0000001a00)="e524d03409aaefcdf6c6659616d99bb6e508859c33221742d2d1f9aacf84e11676f3eb074215c007feabb1225502343f4247b4c922f0cc0ac0d3350d9f6d5ac9cc2ecb0754150a873879e776c6ac06208a2c52ffc63338f333", 0x59}, {&(0x7f0000001a80)="1e8401aa707748e6574c40e9c2c49625a1994dbd1a55b56a70f376f1a281e17c84d17e80d7b6", 0x26}, {&(0x7f0000001ac0)="1c594fe43f90b6ffd9d8099baf46a84865846782e5844128d638481b6ead1734fb98f40d498aa82ba5772863d4da", 0x2e}], 0x7, &(0x7f0000001b80)=[@dstopts_2292={{0x28, 0x29, 0x4, {0x32, 0x2, '\x00', [@pad1, @pad1, @jumbo={0xc2, 0x4, 0x3}, @ra={0x5, 0x2, 0x9}]}}}, @dstopts_2292={{0xf8, 0x29, 0x4, {0x2b, 0x1b, '\x00', [@jumbo={0xc2, 0x4, 0x4}, @calipso={0x7, 0x40, {0x1, 0xe, 0x2, 0xf463, [0xfffffffffffffb46, 0xfff, 0x8, 0x3ff, 0x0, 0x6, 0x8]}}, @padn={0x1, 0x5, [0x0, 0x0, 0x0, 0x0, 0x0]}, @hao={0xc9, 0x10, @empty}, @hao={0xc9, 0x10, @dev={0xfe, 0x80, '\x00', 0xd}}, @enc_lim={0x4, 0x1, 0x2}, @pad1, @hao={0xc9, 0x10, @initdev={0xfe, 0x88, '\x00', 0x1, 0x0}}, @calipso={0x7, 0x48, {0x0, 0x10, 0x1, 0x9, [0xc3, 0x1, 0x1, 0xa0f0, 0x7, 0x8, 0x5, 0x0]}}, @ra={0x5, 0x2, 0xfffe}]}}}, @hopopts_2292={{0x20, 0x29, 0x36, {0x2e, 0x0, '\x00', [@pad1]}}}, @dontfrag={{0x14, 0x29, 0x3e, 0x1}}, @pktinfo={{0x24, 0x29, 0x32, {@rand_addr=' \x01\x00\x00\x00\x00\x00\x00\x00\x00\x00\x00\x00\x00\x00\x02'}}}, @dstopts={{0xb0, 0x29, 0x37, {0x4a, 0x12, '\x00', [@jumbo={0xc2, 0x4, 0xfedd}, @calipso={0x7, 0x58, {0x1, 0x14, 0x6, 0x3, [0x9, 0x2, 0x1, 0x4, 0x401, 0x8, 0x9, 0x2, 0xb4, 0x8]}}, @generic={0xa5, 0x25, "3cbb5e9ca85727c64c56265f036fb83230d933e58f0bbb4135ac8436f962806661674c8891"}, @pad1, @enc_lim={0x4, 0x1, 0x7}, @ra={0x5, 0x2, 0x4}]}}}], 0x230}}], 0x3, 0x10) ioctl$sock_SIOCETHTOOL(r0, 0x89f0, &(0x7f0000000000)={'bridge0\x00', &(0x7f0000000200)=ANY=[@ANYBLOB="1000000000000000000000000000001000"/36]}) socket$inet6(0xa, 0x2, 0x0) (async) socket$inet6(0xa, 0x80002, 0x88) (async) bind$inet6(r1, &(0x7f0000000580)={0xa, 0x10010000004e20}, 0x1c) (async) setsockopt$inet6_udp_int(r1, 0x11, 0xb, &(0x7f0000000080), 0x4) (async) socket$inet_udplite(0x2, 0x2, 0x88) (async) ioctl$sock_SIOCGIFINDEX(r2, 0x8933, &(0x7f0000000340)={'bridge_slave_0\x00'}) (async) socket(0x10, 0x80002, 0x0) (async) sendmsg$nl_route(r4, &(0x7f0000000000)={0x0, 0x0, &(0x7f0000000080)={&(0x7f00000000c0)=@bridge_setlink={0x2c, 0x13, 0xa29, 0x0, 0x0, {0x7, 0x0, 0x0, r3}, [@IFLA_AF_SPEC={0xc, 0x1a, 0x0, 0x1, [@AF_INET={0x8, 0x4, 0x0, 0x1, {0x4, 0x3}}]}]}, 0x2c}}, 0x0) (async) sendmmsg$inet6(r1, &(0x7f0000001dc0)=[{{&(0x7f0000000040)={0xa, 0x4e20, 0x9, @remote, 0x5}, 0x1c, &(0x7f0000000080)=[{&(0x7f0000000240)="dd1ce36b160f3507faaa3ce9d79c64d3afdb297ae040fe7e97eb2d48bd5575e16940c44459dbf315bf6df71a4c309a954eb91b9469a8280aaa31aa1117b6013d2561e4b019ac80f40c31177d73762d1ddaaf0cebd1aa3385d4500b271a1dd0ea0a0d89f0ad7c7fd515fba4c3496ab0c806bd5615975c4da653d6835d176874ed1a455a2efcc819007c84cbb0e235c6fb168077e2190cded56f5f1e1c49c2c303af1dc122ed60b8eca662d413293b9f20f6892372c699fc1b1dc4cfbd21d6a40ad06a68f364a96c80c65aa460374fed63bdf621cdb4f5232887273011ff7af41a9adc68e70623cd90ced64a908f9b9a55013c62194b4cb76494534cbd9762f351bdcf56874a8654d55b26375082fe116548aa72e4bc38452c6ac93aa54b8834a6d9992e9b38037204e66a325d4dffe5a9100d0e845984d3bc542876680f458c6df9046f992f437553e7fd90a0c911a02f6efcc3591ece9427e9784dac0a8e802649846804ef8331dbbd4afd106d4057034f32a008887a252907244edb2fd33c970989b7f76b9ce096d045aada4109111134e2f8dcb48926e122ed706d0b96d9970f9349f2b2d9a7a9b98a08a48651bc65161b1c67c3324eaee83c7d8ff0f99be6182770381e9151899a2858ab89bb5ae606e7188cd29feb2eee4cb41e93cad00d2a3cfad5d0abdc7cebc69d4414205bb3f2ebf6d88b3cee8e8a46fc4523d13763c63866e9b5a89926b1ced6cd3c0a24638b1a4658f168cb45bb0765c18058596a6eb3669dc6eb243c4325ea3b4532cf7ca8e8446144821d54120d36812f4d6b4d63537b04b3ddeaf583a71a090c2fd6c37879d234b43abcc77543b6e2df98ee7e4c5268d96954d68ed1256224486e61300272fe4944ee01c9c7ea92ba66e321c0441377fae47c8fe596ff3f9b0cc9698a51251e219f93a78ac77b57c2904346367219931189357e32133a6ae4e439b3b4471c7f1b9ce55e7cf9c5b123c2573cc6739f0ccae3430763aa8cffc7faa1c2a1a36f3d07946b38a4601e856f13fea9742c60161455ec9f68b3b9fc965359c6e68da11f13e2ae814cdb09e2e706d8230bd1b40ee439442995f05f5a2580c3f12dede9ab2c82b2c4b7eb808bf7a3e114f5cdf30eab015ee991a05f679c106ebb176090801a92e913ab647436e9b6394c695811d04b851212eb2b5d9dc51cc9a26d89c56cc7506a427e49d065b16c00819493cf011dbb090f17eeac9339e0f1fd3728be362411ae8040580bcaadac7d143bc9310e364db7f2e734f948dfba4b8c32cd5398d17fb739e40d0c674ee09936638fc0bf4d6515e68e87581b20d33172b5333673b1e01d5c67ab60e77472655ffc38b0216f9ecec11cc6f6d61f63195b8bf914a89444096c0e88f8682e9a3739cb2b5b66a8a3750209ea0af8c778250fde57243dcb9e7ccec2552327082d2df1d67ad793d1ed89573bd6e3ae6938c33d3a26ee1cd11687f211d26eee4a1e1ca90a555d22cabed33c5edf432c15b3fdc71bdde680e14aed8cd7dd557dfa67e17be6535a6651c66f65fe7201b84b0912ed3f554fe95c8774a19c3227a3c196532d68a6efae2270ea91f34c1f9da2bc91630caed927f386864eca1e536e14913a5b6ceb8e5bfd0f35eac465434710e07bf9bd230e9253b96a0b04f56a7a772d2698f08549584baf69adbc8c2816747eb75cafdc61f5738821dd5cd23e3afbb9c2afbe079c0fdc05b6ac2d8d0e8430e3d496616f36f9665da097880658b13c0285daa09f710f8c11d4ef020dc79374f803c28af1497d063b4b00ca8441cc890abf5c82f7ea36c6001caee1fcc0080b41e0e238f2faf330b64b055101880770c0758ab1da32b8ff151deb9c9dd135e9ebe0be4d799265b735f6789c7c65b340135c71f37cf4fad6cdea553f91daecd1542dc80988726e48ff47bb410854e8bf232d1dbff4633322a0b00991fb8c4c248488811b52521c31d534a17c74d8174083d1863100bc8529516f18ee2a85ef7f824c52bbfd6ca9eee723e14e389c2dfe492de552cff45a1865cb563762922f9741cf9422b5f211a9b5f1e8f68cf2da8274004bef7ede7fb07c9838cb0965771becb6fad77d3dc0218b86f7053957bb31adf80f60937d6d471d4f88cc814e386c888fb7a8f337438e434ab9fd0e7516740d2f66514bb097e21b22da8822a061a0ba7c758fb86b9c16e35c835a41e3357b71a0dc1be8146992ff7768cc48d7a86b8a1c0084c582c7fd09812c74940500d02024ace9a9df3b092c38ff62a7752c2c65233ac918b9bf98529bacaa63b8e219ff460342e11c445bd257839225b6bd2182616843ad103549aa92c2db00ead76a402e92d679741cda04129380c0a1edcb3d4aec6417819ecf16dd47f643a8805660fe33a61663bd6f5d4d215350f0e73710376d3ba1b850e18c0768885e0654c5eb6e383a116414819346e534162a949713d80df9c9a5d10536cb63571e777908564823e566d93bd057764b5d4880679cada7b6336e25d026b129547d36fcc66693207f3df14b83f2f8f106452524e9942b6ade43b340241f04f87eacc65d9b1a62693b7c2cc848a2cd908a1068c614184f07024b52e27ac73297006e25fa14aaf064bca0f6d40ed8d2f884393b287af8d3e2af9d3226e676542c0283a474ff36ecc1a532b34e7fbfb4017ca96b6bc423f9e17ff63cd44cfba7d8a8841b2c277e960c12427666104b67962e5fad35d4c34bfe7c0b69de2b52d07daf10e573e10128af3e695812576bba6b1dd8d02f79d4842808bb1885abbdb7198e2aa839af0a642df6619fd5dbf09296996415aa2ad94e6d8a8a8c2e95eab634c3590d2e6a17689a0a0eb64ecf22744fb98ddc0ef6fc55a90420510fc722f339c583f9767eb1c993f98527c83d36f20860b710669407352742352f02f439e3e1e7933700a75d95765a5d2132f1a1b73cc487113d7988d15c41b8baecead428c495c36824f8c1115f6795c8bf2a9817ba6382bd5d669c3aa80cafd8396352228bf693af529d88273a34fb4ec90901992eb425230614b9b361515e9e4f65631c385bbe8da0e735c4d6fc680aa6664b730fa4f15bc397b19a9791d8c53e631acd930bc365ad05f5e8f7e825eba4e7588acfe0568611a046246278562217aaaa0766168091b348791b876bead75de07d153babb3e0665a326bcd71a472d51e86835471994fec156b07e04303ffa8089afc49b484fbfaf9a86255fe218f541ee500bb45fc355d6557fb884f1a840f7f256f158d1673e1ad8fe837ce9b3e77aaca3d5b90bfbd642c7f766e61549944848780c815febdcb97ba35a2702303f338286c1a94e2ff8b46d0d0f541a891f5bd424a7321de44d0c1a568c792d6bb99e6dbc16b389a7e83c3fabcb9e5fefb5b2c5c4329d01bb898e6c8f2067ee73c73e90dfa18da7a82bf217723efdd7d27e6f132f3a648c485a1b9363dc7224e4b846e99e3737ec9eb6f6d8816563887acd317361251cd46ece0a6d9d797951b4cd21b3b848d5384cb85488c08ca952ca5a1d65fab840f81ec56d6425a9a7da8be24588856bdbfc7df08e5f13f7e10e80a8220851db1b41fcf23b0616a83ecace12bfbd917ac5a9af8dcfc2258a1be5634e7c0d8511d94162caf32e9800343689f30cc373ffd3f354f23ec11bcdf986e711a683d60d6bfa66ae83d7aec00f9ee822e2e3abc4322232c27e3ff543fb12772d216ac7726a770f1f6879c79d41693b85917b17b10f71b54c3828b38624dba9ff9cfd9e322234800128bf921ffee1501f0940f3d2bbc081c5ff67c3f4c63a01a2d14a04abea0ecb8fc47f2d365a7986a12bb1af7896ab1e97d62ca96a55deb0e751c855446f53c27520ba831230610c1102639c109e47c988968ce140c92e4e555d3b46a38b2778ae03efbed5a28337b0a04f39707439d22d38b9b185702874f2570cb98faaf1b8a77ba7f88d056d5266e9ced1828751accd418985952a9d851c98485a9f703e0ac98232084d1f7e16ede91c73f679e25c974284d71ecee1bc580ade172ebb1209caa5558df2d830b4390a0eb92d26f0f937f53455a3003f9978bc92b8471c71d1e26d046555361417b91b81768bd23ddce564e8224c8ef9aa2387f87c709266bfd01526ac207fe8b324ac30c3474e1e1ad2c92016883548859867814d8522979d4b31a2c1fb0706693525581badea1a077fecc18e8150d398d2854068e1868af81ab23cb92287b1984f550def18ce074c0eb4cc8611721016ced6c696292259d147f525ad493f3ce57ce63cecd44b33536ab5762eacbd7501a6d3df899e0ccc83edb0ac9633aa64d0bbdbf533f16c49a8bf8ffd0d6c05803a93da4af3ee56f61fa72aef02fee6df5aa6c040812fc7cce65f4242cea1afd9d82443e3d4b4fe98a7ba0b783e7ee94ad3f1d08039f32b29029819b94d6ef158e8d6316be9d47f13e3243e1bdc946dde5249264e0d7df0bbffa1a4828e714a996283c6a0935fcd34f8d0040ce8668308eca6f6fbe85063f768ee9130e4a5b7570a69a8723e2f41a4bc9eba67f016171d6b787c13458fb7655a1b7b26f6cd0d7e339aaffcda7624f9396ff9b45edfea9c593a5d9d0168534bd7203365c2edfe21b1253ce340114d4e1874be56dc8502c23f201fda9951f9a41fc0bfc92a738330efc6c6eae8d6d316b317e9e2bc792634162d58734820492b646a0a6a06ec126b5c2284313993452ebed96a031f3dcd11913cf7766b8236208ea6240c3fcc2d3524b7d216f41c47307aea4f01286134e78a88738d6448eec0b5e073bcfdaea76e3d2afe9a0fe4f4d3c08e30124652dc43cfd2a502fe61261bd134bdb3a10721f2eaea8170dd8dabfb7581d3acdfadff51774ba03a971d3229cc4b54db1bb7583f60f7851e2a93627db21e442f33ccd66b90047861af55203adcb1c3001af055a096180b346a62019947c71fe08c2ce115374f1c1678aca401deae482bf0edc994cfd122da0d93df031bfc56ef9642d6b1465d66003a9e2f5820d194c4df01cd885a96a16d5fcbe986ef416371b42179fc76052ee298d2f0ea96afd968740d4ccf16ab7db154deab314616a6c38c9e9daaaaa1306349f7aa913d7395f024815de3d067f39d9f9ddc077873dbd1556f0912a20ae4cefb3f4faaab30506dade5d6369e6e580d2b610572ae768d09f0a813782a1920228b7c2e9d83508b3d7e5bd306d79a2da8f1455745b30c23c0b297e67c99bb3b12a0954a924ac7d4925e7f77c785cb3f74ff85a5e9e3c7b38359f4f22c314177cf1155cc7505cbc7919e2525cf41c16052461e69036f1acdafa488dea7f9bc6608a25b1df13aec49ae51c6b643a8fdffaf6645d1acd1e2ab309283f6b64bc6dd1b2a718b37e1ad2d4b854d99d12afedfa42f7c42ebf6db4994eaf62f28f0d6b9f6ed172143926a785a3eb19fad8715f1f958cddc2a4a805438485ea9a5f974387f29a81a7261898c69b18ac04edfe79480849745b305ebca727537438d13f0328cef8931fbd8284cf644d89ff8d978f5dda5de5602b285230cd6d419fc63f6d1a02d649c69ecb8d093e603036faaae4a59b5d0aaed76fd539a6b4b7d7433df00893b3775d5fbe718539cdfcd0b247bc22f2ef7d5c58c99c8fd039367024ff94893bdee566bda6784a131f1633f572fdb87b406299cf345ee1de46a78e591f6d45eb37c2a4a481cce6eb6c67bb9928b3d33965e4693831e99194c41283d8f493886cca86fade836bdb6e7eac357adc0d507be9cff4d4635dc61086ba939d593607f094fb72fadc9eb648540a6b54", 0x1000}], 0x1, &(0x7f0000001240)=[@hopopts={{0xa0, 0x29, 0x36, {0x3c, 0x10, '\x00', [@enc_lim={0x4, 0x1, 0x4}, @enc_lim={0x4, 0x1, 0x2}, @generic={0x81, 0x60, "1f369872c4e4612f8bb266eb86912ac884c06ac35e7cf1eb3ff6684aab0a98432cf9e23c55d9248c910213d4b99f117bed505fe085b9ea096b8eca29bfd971a23933097d1b8268ec8a5d280837c5d2283adbe00eb1184416071d8510ced881bb"}, @ra={0x5, 0x2, 0x7245}, @hao={0xc9, 0x10, @mcast1}, @ra={0x5, 0x2, 0x6}, @enc_lim={0x4, 0x1, 0x1}]}}}, @hoplimit={{0x14, 0x29, 0x34, 0xffffffff}}, @rthdr={{0x98, 0x29, 0x39, {0x21, 0x10, 0x0, 0x4, 0x0, [@loopback, @rand_addr=' \x01\x00', @private0, @remote, @ipv4={'\x00', '\xff\xff', @dev={0xac, 0x14, 0x14, 0x2b}}, @dev={0xfe, 0x80, '\x00', 0x40}, @private0={0xfc, 0x0, '\x00', 0x1}, @loopback]}}}, @dontfrag={{0x14, 0x29, 0x3e, 0x6}}, @rthdr_2292={{0x48, 0x29, 0x39, {0x1d, 0x6, 0x2, 0x0, 0x0, [@initdev={0xfe, 0x88, '\x00', 0x1, 0x0}, @remote, @private0]}}}], 0x1b0}}, {{&(0x7f00000000c0)={0xa, 0x4e23, 0x1, @private2={0xfc, 0x2, '\x00', 0x1}, 0x7ff}, 0x1c, &(0x7f0000001540)=[{&(0x7f0000000100)="b4d537fc4ec5292b087a9a19d73d33ee91cb83f20643778c7ae309cd39bf3fe60466967057bea074b619fe631de3fab9fa36bc064ddbc3ab97a1cb29ba5a5d8b2b76c6c2bf001b45c7cc0f36d06c435f5c3d99cb4d050d1de994b075c0cfd3de7e2d27d95956b962b871377c42b8f62bf5f46433e1eb44c395903270696cdfd46528edbe80be6d09bb471450618e9de4b0cd63dec325823d8913d5dadb2fb57786ca5142b92c10847aec01aaa4118eea7e73b951f8048db1d19e0bdc46ebf6a50c3a8677146fcdcd5bea67d2", 0xcc}, {&(0x7f0000001400)="48affc6589751527c1662de5caaea4cfbf91f195e2474ad43fc8708054af73bbbc4ff13cce1df94a922b7bc9b22f8da2046691d8b2baeeaab1ba61a62f9ee5dbcede38ae2b8fd731f3b102", 0x4b}, {&(0x7f0000001480)="b2abb19bc646600056bf615c6eb86d2affd0c6a19703efa75fae356f2814da02d32c4ec3668accc6fe2152959d31764d69a2cf8897e590f587aa7c12c279996195bfa37e93d60b9032673da89cfa9cfc52f99b623f97eb38d648a4a1afda82274d503957efd6a089bb10d15498a3062a097d5ac1b688cd4439a877744a123e0f3b003f1e", 0x84}], 0x3, &(0x7f0000001580)=[@hopopts={{0x38, 0x29, 0x36, {0x62, 0x3, '\x00', [@calipso={0x7, 0x18, {0x2, 0x4, 0x7f, 0x7, [0x3, 0xffff]}}, @ra={0x5, 0x2, 0x8001}]}}}, @pktinfo={{0x24, 0x29, 0x32, {@mcast2, r3}}}, @hoplimit_2292={{0x14, 0x29, 0x8, 0x20}}, @rthdrdstopts={{0x160, 0x29, 0x37, {0x5e, 0x28, '\x00', [@jumbo={0xc2, 0x4, 0x5}, @ra={0x5, 0x2, 0x9}, @calipso={0x7, 0x50, {0x2, 0x12, 0x2, 0x0, [0x7, 0x7, 0x8, 0xe6, 0x80000000, 0x4, 0x5a3, 0x5, 0x100000000]}}, @generic={0xf2, 0xe9, "6c24ee61412a83545ce26853175fc8054839e39e56d7c7807054dec0596a3f3ead7d12fbdf75e0eac3319c2d8ee18023c547511173ddb68acd3264b3a118c8a78c9c03f65bc4ec62e9e7879110e0aed4b80d752fd329a17e51bf207baf6229afc8416cb69b00acd5bc250abc1abb41005824e38b7fde7b7c4711eba3f4c6e0b02bac902f496cc9d75928e1fe598f2e4485b88c68f5378144d52269f43fb9ca90ba4fc295aa4022402efd99fdc37c54ddf3818d7b5c36d0b81b8247df839a365cfbe0f7116b7c21a5c1e2014710ea432eff5d3a65fdaa36c317f1ec7c0b97843afae20c9d7fd7d4bf70"}]}}}], 0x1d8}}, {{0x0, 0x0, &(0x7f0000001b00)=[{&(0x7f0000001780)="5feea81a24ebddeeac", 0x9}, {&(0x7f00000017c0)="95d527e3bd5b310da8bf16bd6fe8b85ef0371db049f969ec59147745ff0ef5fbb65f6a3094408c14f69f15a07f274db20b255251da84967f0ffa9fe50be037072bcc4411d9551a3150c0c5b10926966423d765455d65a17b", 0x58}, {&(0x7f0000001840)="bb6c6bf4a9151ae18e87006298fd33f59e346e44892d5ed19ad6157bacd297f597f9c7e3ab972069154bb109d8262c82b79a0de44798afc5947611828a6ac8160bdec7a40c49150226a16427f8210246cb3482d2f84bfe8a1d36b5ef5a4c2fbfb615a48ef5d3074f01c14145b42a2e7285e4dbc1a150f9eb8a38775652c6583fe8aaf0fb5040569d90bd6682120b947b31dc", 0x92}, {&(0x7f0000001900)="9b37a484f3851e1fe1cd8bdcc339374a22197cd54ba150aa1d45526bb63108f22c75aa24a93540934912923150b647163341d2aa123de3e9ac167cd2d0d71a19091513c3374818ea981fb781df772b4baec81e912e9c2148512eacdaa7e7fbe0d7f61aad6aa19de932a7157eca8794090a25ee4eb1825403bdb698e6184ed348c175e2e7d6c0e300eed5fa750f894c05e3979e0f485327768b4777e05b2a63113a79a08dda92031b4c03e30c85cc7298e8b0776d20131478e62fb578f4c54e351c29c3d7dfafd956cc4ba3c41cf4a24ea7ae273ae57722a5f64cdb94723bae4b108e", 0xe2}, {&(0x7f0000001a00)="e524d03409aaefcdf6c6659616d99bb6e508859c33221742d2d1f9aacf84e11676f3eb074215c007feabb1225502343f4247b4c922f0cc0ac0d3350d9f6d5ac9cc2ecb0754150a873879e776c6ac06208a2c52ffc63338f333", 0x59}, {&(0x7f0000001a80)="1e8401aa707748e6574c40e9c2c49625a1994dbd1a55b56a70f376f1a281e17c84d17e80d7b6", 0x26}, {&(0x7f0000001ac0)="1c594fe43f90b6ffd9d8099baf46a84865846782e5844128d638481b6ead1734fb98f40d498aa82ba5772863d4da", 0x2e}], 0x7, &(0x7f0000001b80)=[@dstopts_2292={{0x28, 0x29, 0x4, {0x32, 0x2, '\x00', [@pad1, @pad1, @jumbo={0xc2, 0x4, 0x3}, @ra={0x5, 0x2, 0x9}]}}}, @dstopts_2292={{0xf8, 0x29, 0x4, {0x2b, 0x1b, '\x00', [@jumbo={0xc2, 0x4, 0x4}, @calipso={0x7, 0x40, {0x1, 0xe, 0x2, 0xf463, [0xfffffffffffffb46, 0xfff, 0x8, 0x3ff, 0x0, 0x6, 0x8]}}, @padn={0x1, 0x5, [0x0, 0x0, 0x0, 0x0, 0x0]}, @hao={0xc9, 0x10, @empty}, @hao={0xc9, 0x10, @dev={0xfe, 0x80, '\x00', 0xd}}, @enc_lim={0x4, 0x1, 0x2}, @pad1, @hao={0xc9, 0x10, @initdev={0xfe, 0x88, '\x00', 0x1, 0x0}}, @calipso={0x7, 0x48, {0x0, 0x10, 0x1, 0x9, [0xc3, 0x1, 0x1, 0xa0f0, 0x7, 0x8, 0x5, 0x0]}}, @ra={0x5, 0x2, 0xfffe}]}}}, @hopopts_2292={{0x20, 0x29, 0x36, {0x2e, 0x0, '\x00', [@pad1]}}}, @dontfrag={{0x14, 0x29, 0x3e, 0x1}}, @pktinfo={{0x24, 0x29, 0x32, {@rand_addr=' \x01\x00\x00\x00\x00\x00\x00\x00\x00\x00\x00\x00\x00\x00\x02'}}}, @dstopts={{0xb0, 0x29, 0x37, {0x4a, 0x12, '\x00', [@jumbo={0xc2, 0x4, 0xfedd}, @calipso={0x7, 0x58, {0x1, 0x14, 0x6, 0x3, [0x9, 0x2, 0x1, 0x4, 0x401, 0x8, 0x9, 0x2, 0xb4, 0x8]}}, @generic={0xa5, 0x25, "3cbb5e9ca85727c64c56265f036fb83230d933e58f0bbb4135ac8436f962806661674c8891"}, @pad1, @enc_lim={0x4, 0x1, 0x7}, @ra={0x5, 0x2, 0x4}]}}}], 0x230}}], 0x3, 0x10) (async) ioctl$sock_SIOCETHTOOL(r0, 0x89f0, &(0x7f0000000000)={'bridge0\x00', &(0x7f0000000200)=ANY=[@ANYBLOB="1000000000000000000000000000001000"/36]}) (async) 04:12:46 executing program 0: r0 = syz_init_net_socket$ax25(0x3, 0x2, 0x0) bind$ax25(r0, 0x0, 0x0) 04:12:46 executing program 3: r0 = socket$netlink(0x10, 0x3, 0x0) sendmsg$netlink(r0, &(0x7f00000047c0)={0x0, 0x0, &(0x7f0000000000)=[{&(0x7f0000000100)=ANY=[@ANYBLOB="18000000560001060000008500db730007"], 0x18}], 0x1}, 0x0) [ 2839.018861][ T3318] bond1287: (slave bridge1208): Enslaving as an active interface with an up link [ 2839.047757][ T3329] netlink: 'syz-executor.1': attribute type 1 has an invalid length. 04:12:46 executing program 2: r0 = socket$netlink(0x10, 0x3, 0x0) r1 = socket$netlink(0x10, 0x3, 0x0) r2 = socket(0x10, 0x803, 0x0) sendmsg$NL80211_CMD_CRIT_PROTOCOL_START(r2, &(0x7f0000000580)={0x0, 0x0, &(0x7f0000000540)={0x0, 0x1c}}, 0x0) getsockname$packet(r2, &(0x7f0000000600)={0x11, 0x0, 0x0, 0x1, 0x0, 0x6, @broadcast}, &(0x7f0000000100)=0x14) sendmsg$nl_route(r1, &(0x7f0000000040)={0x0, 0x0, &(0x7f0000000000)={&(0x7f00000000c0)=ANY=[@ANYBLOB="3c00000010008506eb9afc4cd8d06e754a0081c5", @ANYRES32=r3, @ANYBLOB="2377f292252155b21c0012000c000100626f6e64000000000c0002000800010001"], 0x3c}}, 0x0) sendmsg$nl_route(r0, &(0x7f0000000240)={0x0, 0x0, &(0x7f0000000680)={&(0x7f0000000400)=@newlink={0x3c, 0x10, 0xffffff1f, 0x0, 0x253c3c, {}, [@IFLA_LINKINFO={0x14, 0x12, 0x0, 0x1, @bridge={{0xb}, {0x4}}}, @IFLA_MASTER={0x8, 0xa, r3}]}, 0x3c}}, 0x0) [ 2839.233352][ T3329] bond1304: entered promiscuous mode [ 2839.239139][ T3329] 8021q: adding VLAN 0 to HW filter on device bond1304 [ 2839.286047][ T3331] bond1304: (slave bridge1250): making interface the new active one [ 2839.295274][ T3331] bridge1250: entered promiscuous mode [ 2839.318533][ T3331] bond1304: (slave bridge1250): Enslaving as an active interface with an up link 04:12:46 executing program 1: r0 = socket$netlink(0x10, 0x3, 0x0) r1 = socket$netlink(0x10, 0x3, 0x0) r2 = socket(0x10, 0x803, 0x0) sendmsg$NL80211_CMD_CRIT_PROTOCOL_START(r2, &(0x7f0000000580)={0x0, 0x0, &(0x7f0000000540)={0x0, 0x1c}}, 0x0) getsockname$packet(r2, &(0x7f0000000600)={0x11, 0x0, 0x0, 0x1, 0x0, 0x6, @broadcast}, &(0x7f0000000100)=0x14) sendmsg$nl_route(r1, &(0x7f0000000040)={0x0, 0x0, &(0x7f0000000000)={&(0x7f00000000c0)=ANY=[@ANYBLOB="3c00000010008506eb9afc4cd8d06e754a0081c5", @ANYRES32=r3, @ANYBLOB="2377f292252155b21c0012000c000100626f6e64000000000c0002000800010001"], 0x3c}}, 0x0) sendmsg$nl_route(r0, &(0x7f0000000240)={0x0, 0x0, &(0x7f0000000680)={&(0x7f0000000400)=@newlink={0x3c, 0x10, 0xffffff1f, 0x0, 0x1c672500, {}, [@IFLA_LINKINFO={0x14, 0x12, 0x0, 0x1, @bridge={{0xb}, {0x4}}}, @IFLA_MASTER={0x8, 0xa, r3}]}, 0x3c}}, 0x0) 04:12:46 executing program 0: r0 = syz_init_net_socket$ax25(0x3, 0x2, 0x0) bind$ax25(r0, 0x0, 0x0) [ 2839.430191][ T3339] netlink: 'syz-executor.5': attribute type 1 has an invalid length. 04:12:47 executing program 0: r0 = syz_init_net_socket$ax25(0x3, 0x2, 0x0) bind$ax25(r0, 0x0, 0x0) [ 2839.576206][ T3339] bond1297: entered promiscuous mode [ 2839.603144][ T3339] 8021q: adding VLAN 0 to HW filter on device bond1297 04:12:47 executing program 0: r0 = syz_init_net_socket$ax25(0x3, 0x2, 0x0) bind$ax25(r0, &(0x7f0000000000)={{0x3, @remote={0xcc, 0xcc, 0xcc, 0xcc, 0xcc, 0xcc, 0x1}}, [@null, @netrom={0xbb, 0xbb, 0xbb, 0xbb, 0xbb, 0x0, 0x0}, @bcast, @rose={0xbb, 0xbb, 0xbb, 0x1, 0x0}, @bcast, @netrom={0xbb, 0xbb, 0xbb, 0xbb, 0xbb, 0x0, 0x0}, @null, @bcast]}, 0x48) [ 2839.740038][ T3343] bond1297: (slave bridge1260): making interface the new active one [ 2839.765889][ T3343] bridge1260: entered promiscuous mode 04:12:47 executing program 3: r0 = socket$netlink(0x10, 0x3, 0x0) sendmsg$netlink(r0, &(0x7f00000047c0)={0x0, 0x0, &(0x7f0000000000)=[{&(0x7f0000000100)=ANY=[@ANYBLOB="18000000560001060000008600db730007"], 0x18}], 0x1}, 0x0) [ 2839.790191][ T3343] bond1297: (slave bridge1260): Enslaving as an active interface with an up link [ 2839.820557][ T3352] netlink: 'syz-executor.2': attribute type 1 has an invalid length. 04:12:47 executing program 5: r0 = socket$netlink(0x10, 0x3, 0x0) r1 = socket$netlink(0x10, 0x3, 0x0) r2 = socket(0x10, 0x803, 0x0) sendmsg$NL80211_CMD_CRIT_PROTOCOL_START(r2, &(0x7f0000000580)={0x0, 0x0, &(0x7f0000000540)={0x0, 0x1c}}, 0x0) getsockname$packet(r2, &(0x7f0000000600)={0x11, 0x0, 0x0, 0x1, 0x0, 0x6, @broadcast}, &(0x7f0000000100)=0x14) sendmsg$nl_route(r1, &(0x7f0000000040)={0x0, 0x0, &(0x7f0000000000)={&(0x7f00000000c0)=ANY=[@ANYBLOB="3c00000010008506eb9afc4cd8d06e754a0081c5", @ANYRES32=r3, @ANYBLOB="2377f292252155b21c0012000c000100626f6e64000000000c0002000800010001"], 0x3c}}, 0x0) sendmsg$nl_route(r0, &(0x7f0000000240)={0x0, 0x0, &(0x7f0000000680)={&(0x7f0000000400)=@newlink={0x3c, 0x10, 0xffffff1f, 0x0, 0x1a030000, {}, [@IFLA_LINKINFO={0x14, 0x12, 0x0, 0x1, @bridge={{0xb}, {0x4}}}, @IFLA_MASTER={0x8, 0xa, r3}]}, 0x3c}}, 0x0) [ 2839.956883][ T3352] bond1288: entered promiscuous mode [ 2839.965985][ T3352] 8021q: adding VLAN 0 to HW filter on device bond1288 04:12:47 executing program 0: r0 = syz_init_net_socket$ax25(0x3, 0x2, 0x0) bind$ax25(r0, &(0x7f0000000000)={{0x3, @remote={0xcc, 0xcc, 0xcc, 0xcc, 0xcc, 0xcc, 0x1}}, [@null, @netrom={0xbb, 0xbb, 0xbb, 0xbb, 0xbb, 0x0, 0x0}, @bcast, @rose={0xbb, 0xbb, 0xbb, 0x1, 0x0}, @bcast, @netrom={0xbb, 0xbb, 0xbb, 0xbb, 0xbb, 0x0, 0x0}, @null, @bcast]}, 0x48) 04:12:47 executing program 0: r0 = syz_init_net_socket$ax25(0x3, 0x2, 0x0) bind$ax25(r0, &(0x7f0000000000)={{0x3, @remote={0xcc, 0xcc, 0xcc, 0xcc, 0xcc, 0xcc, 0x1}}, [@null, @netrom={0xbb, 0xbb, 0xbb, 0xbb, 0xbb, 0x0, 0x0}, @bcast, @rose={0xbb, 0xbb, 0xbb, 0x1, 0x0}, @bcast, @netrom={0xbb, 0xbb, 0xbb, 0xbb, 0xbb, 0x0, 0x0}, @null, @bcast]}, 0x48) 04:12:47 executing program 4: r0 = socket$inet6(0xa, 0x2, 0x0) (async, rerun: 32) r1 = socket$inet6(0xa, 0x80002, 0x88) (rerun: 32) bind$inet6(r1, &(0x7f0000000580)={0xa, 0x10010000004e20}, 0x1c) setsockopt$inet6_udp_int(r1, 0x11, 0xb, &(0x7f0000000080), 0x4) (async) r2 = socket$inet_udplite(0x2, 0x2, 0x88) ioctl$sock_SIOCGIFINDEX(r2, 0x8933, &(0x7f0000000340)={'bridge_slave_0\x00', 0x0}) (async) r4 = socket(0x10, 0x80002, 0x0) sendmsg$nl_route(r4, &(0x7f0000000000)={0x0, 0x0, &(0x7f0000000080)={&(0x7f00000000c0)=@bridge_setlink={0x2c, 0x13, 0xa29, 0x0, 0x0, {0x7, 0x0, 0x0, r3}, [@IFLA_AF_SPEC={0xc, 0x1a, 0x0, 0x1, [@AF_INET={0x8, 0x4, 0x0, 0x1, {0x4, 0x3}}]}]}, 0x2c}}, 0x0) (async) sendmmsg$inet6(r1, &(0x7f0000001dc0)=[{{&(0x7f0000000040)={0xa, 0x4e20, 0x9, @remote, 0x5}, 0x1c, &(0x7f0000000080)=[{&(0x7f0000000240)="dd1ce36b160f3507faaa3ce9d79c64d3afdb297ae040fe7e97eb2d48bd5575e16940c44459dbf315bf6df71a4c309a954eb91b9469a8280aaa31aa1117b6013d2561e4b019ac80f40c31177d73762d1ddaaf0cebd1aa3385d4500b271a1dd0ea0a0d89f0ad7c7fd515fba4c3496ab0c806bd5615975c4da653d6835d176874ed1a455a2efcc819007c84cbb0e235c6fb168077e2190cded56f5f1e1c49c2c303af1dc122ed60b8eca662d413293b9f20f6892372c699fc1b1dc4cfbd21d6a40ad06a68f364a96c80c65aa460374fed63bdf621cdb4f5232887273011ff7af41a9adc68e70623cd90ced64a908f9b9a55013c62194b4cb76494534cbd9762f351bdcf56874a8654d55b26375082fe116548aa72e4bc38452c6ac93aa54b8834a6d9992e9b38037204e66a325d4dffe5a9100d0e845984d3bc542876680f458c6df9046f992f437553e7fd90a0c911a02f6efcc3591ece9427e9784dac0a8e802649846804ef8331dbbd4afd106d4057034f32a008887a252907244edb2fd33c970989b7f76b9ce096d045aada4109111134e2f8dcb48926e122ed706d0b96d9970f9349f2b2d9a7a9b98a08a48651bc65161b1c67c3324eaee83c7d8ff0f99be6182770381e9151899a2858ab89bb5ae606e7188cd29feb2eee4cb41e93cad00d2a3cfad5d0abdc7cebc69d4414205bb3f2ebf6d88b3cee8e8a46fc4523d13763c63866e9b5a89926b1ced6cd3c0a24638b1a4658f168cb45bb0765c18058596a6eb3669dc6eb243c4325ea3b4532cf7ca8e8446144821d54120d36812f4d6b4d63537b04b3ddeaf583a71a090c2fd6c37879d234b43abcc77543b6e2df98ee7e4c5268d96954d68ed1256224486e61300272fe4944ee01c9c7ea92ba66e321c0441377fae47c8fe596ff3f9b0cc9698a51251e219f93a78ac77b57c2904346367219931189357e32133a6ae4e439b3b4471c7f1b9ce55e7cf9c5b123c2573cc6739f0ccae3430763aa8cffc7faa1c2a1a36f3d07946b38a4601e856f13fea9742c60161455ec9f68b3b9fc965359c6e68da11f13e2ae814cdb09e2e706d8230bd1b40ee439442995f05f5a2580c3f12dede9ab2c82b2c4b7eb808bf7a3e114f5cdf30eab015ee991a05f679c106ebb176090801a92e913ab647436e9b6394c695811d04b851212eb2b5d9dc51cc9a26d89c56cc7506a427e49d065b16c00819493cf011dbb090f17eeac9339e0f1fd3728be362411ae8040580bcaadac7d143bc9310e364db7f2e734f948dfba4b8c32cd5398d17fb739e40d0c674ee09936638fc0bf4d6515e68e87581b20d33172b5333673b1e01d5c67ab60e77472655ffc38b0216f9ecec11cc6f6d61f63195b8bf914a89444096c0e88f8682e9a3739cb2b5b66a8a3750209ea0af8c778250fde57243dcb9e7ccec2552327082d2df1d67ad793d1ed89573bd6e3ae6938c33d3a26ee1cd11687f211d26eee4a1e1ca90a555d22cabed33c5edf432c15b3fdc71bdde680e14aed8cd7dd557dfa67e17be6535a6651c66f65fe7201b84b0912ed3f554fe95c8774a19c3227a3c196532d68a6efae2270ea91f34c1f9da2bc91630caed927f386864eca1e536e14913a5b6ceb8e5bfd0f35eac465434710e07bf9bd230e9253b96a0b04f56a7a772d2698f08549584baf69adbc8c2816747eb75cafdc61f5738821dd5cd23e3afbb9c2afbe079c0fdc05b6ac2d8d0e8430e3d496616f36f9665da097880658b13c0285daa09f710f8c11d4ef020dc79374f803c28af1497d063b4b00ca8441cc890abf5c82f7ea36c6001caee1fcc0080b41e0e238f2faf330b64b055101880770c0758ab1da32b8ff151deb9c9dd135e9ebe0be4d799265b735f6789c7c65b340135c71f37cf4fad6cdea553f91daecd1542dc80988726e48ff47bb410854e8bf232d1dbff4633322a0b00991fb8c4c248488811b52521c31d534a17c74d8174083d1863100bc8529516f18ee2a85ef7f824c52bbfd6ca9eee723e14e389c2dfe492de552cff45a1865cb563762922f9741cf9422b5f211a9b5f1e8f68cf2da8274004bef7ede7fb07c9838cb0965771becb6fad77d3dc0218b86f7053957bb31adf80f60937d6d471d4f88cc814e386c888fb7a8f337438e434ab9fd0e7516740d2f66514bb097e21b22da8822a061a0ba7c758fb86b9c16e35c835a41e3357b71a0dc1be8146992ff7768cc48d7a86b8a1c0084c582c7fd09812c74940500d02024ace9a9df3b092c38ff62a7752c2c65233ac918b9bf98529bacaa63b8e219ff460342e11c445bd257839225b6bd2182616843ad103549aa92c2db00ead76a402e92d679741cda04129380c0a1edcb3d4aec6417819ecf16dd47f643a8805660fe33a61663bd6f5d4d215350f0e73710376d3ba1b850e18c0768885e0654c5eb6e383a116414819346e534162a949713d80df9c9a5d10536cb63571e777908564823e566d93bd057764b5d4880679cada7b6336e25d026b129547d36fcc66693207f3df14b83f2f8f106452524e9942b6ade43b340241f04f87eacc65d9b1a62693b7c2cc848a2cd908a1068c614184f07024b52e27ac73297006e25fa14aaf064bca0f6d40ed8d2f884393b287af8d3e2af9d3226e676542c0283a474ff36ecc1a532b34e7fbfb4017ca96b6bc423f9e17ff63cd44cfba7d8a8841b2c277e960c12427666104b67962e5fad35d4c34bfe7c0b69de2b52d07daf10e573e10128af3e695812576bba6b1dd8d02f79d4842808bb1885abbdb7198e2aa839af0a642df6619fd5dbf09296996415aa2ad94e6d8a8a8c2e95eab634c3590d2e6a17689a0a0eb64ecf22744fb98ddc0ef6fc55a90420510fc722f339c583f9767eb1c993f98527c83d36f20860b710669407352742352f02f439e3e1e7933700a75d95765a5d2132f1a1b73cc487113d7988d15c41b8baecead428c495c36824f8c1115f6795c8bf2a9817ba6382bd5d669c3aa80cafd8396352228bf693af529d88273a34fb4ec90901992eb425230614b9b361515e9e4f65631c385bbe8da0e735c4d6fc680aa6664b730fa4f15bc397b19a9791d8c53e631acd930bc365ad05f5e8f7e825eba4e7588acfe0568611a046246278562217aaaa0766168091b348791b876bead75de07d153babb3e0665a326bcd71a472d51e86835471994fec156b07e04303ffa8089afc49b484fbfaf9a86255fe218f541ee500bb45fc355d6557fb884f1a840f7f256f158d1673e1ad8fe837ce9b3e77aaca3d5b90bfbd642c7f766e61549944848780c815febdcb97ba35a2702303f338286c1a94e2ff8b46d0d0f541a891f5bd424a7321de44d0c1a568c792d6bb99e6dbc16b389a7e83c3fabcb9e5fefb5b2c5c4329d01bb898e6c8f2067ee73c73e90dfa18da7a82bf217723efdd7d27e6f132f3a648c485a1b9363dc7224e4b846e99e3737ec9eb6f6d8816563887acd317361251cd46ece0a6d9d797951b4cd21b3b848d5384cb85488c08ca952ca5a1d65fab840f81ec56d6425a9a7da8be24588856bdbfc7df08e5f13f7e10e80a8220851db1b41fcf23b0616a83ecace12bfbd917ac5a9af8dcfc2258a1be5634e7c0d8511d94162caf32e9800343689f30cc373ffd3f354f23ec11bcdf986e711a683d60d6bfa66ae83d7aec00f9ee822e2e3abc4322232c27e3ff543fb12772d216ac7726a770f1f6879c79d41693b85917b17b10f71b54c3828b38624dba9ff9cfd9e322234800128bf921ffee1501f0940f3d2bbc081c5ff67c3f4c63a01a2d14a04abea0ecb8fc47f2d365a7986a12bb1af7896ab1e97d62ca96a55deb0e751c855446f53c27520ba831230610c1102639c109e47c988968ce140c92e4e555d3b46a38b2778ae03efbed5a28337b0a04f39707439d22d38b9b185702874f2570cb98faaf1b8a77ba7f88d056d5266e9ced1828751accd418985952a9d851c98485a9f703e0ac98232084d1f7e16ede91c73f679e25c974284d71ecee1bc580ade172ebb1209caa5558df2d830b4390a0eb92d26f0f937f53455a3003f9978bc92b8471c71d1e26d046555361417b91b81768bd23ddce564e8224c8ef9aa2387f87c709266bfd01526ac207fe8b324ac30c3474e1e1ad2c92016883548859867814d8522979d4b31a2c1fb0706693525581badea1a077fecc18e8150d398d2854068e1868af81ab23cb92287b1984f550def18ce074c0eb4cc8611721016ced6c696292259d147f525ad493f3ce57ce63cecd44b33536ab5762eacbd7501a6d3df899e0ccc83edb0ac9633aa64d0bbdbf533f16c49a8bf8ffd0d6c05803a93da4af3ee56f61fa72aef02fee6df5aa6c040812fc7cce65f4242cea1afd9d82443e3d4b4fe98a7ba0b783e7ee94ad3f1d08039f32b29029819b94d6ef158e8d6316be9d47f13e3243e1bdc946dde5249264e0d7df0bbffa1a4828e714a996283c6a0935fcd34f8d0040ce8668308eca6f6fbe85063f768ee9130e4a5b7570a69a8723e2f41a4bc9eba67f016171d6b787c13458fb7655a1b7b26f6cd0d7e339aaffcda7624f9396ff9b45edfea9c593a5d9d0168534bd7203365c2edfe21b1253ce340114d4e1874be56dc8502c23f201fda9951f9a41fc0bfc92a738330efc6c6eae8d6d316b317e9e2bc792634162d58734820492b646a0a6a06ec126b5c2284313993452ebed96a031f3dcd11913cf7766b8236208ea6240c3fcc2d3524b7d216f41c47307aea4f01286134e78a88738d6448eec0b5e073bcfdaea76e3d2afe9a0fe4f4d3c08e30124652dc43cfd2a502fe61261bd134bdb3a10721f2eaea8170dd8dabfb7581d3acdfadff51774ba03a971d3229cc4b54db1bb7583f60f7851e2a93627db21e442f33ccd66b90047861af55203adcb1c3001af055a096180b346a62019947c71fe08c2ce115374f1c1678aca401deae482bf0edc994cfd122da0d93df031bfc56ef9642d6b1465d66003a9e2f5820d194c4df01cd885a96a16d5fcbe986ef416371b42179fc76052ee298d2f0ea96afd968740d4ccf16ab7db154deab314616a6c38c9e9daaaaa1306349f7aa913d7395f024815de3d067f39d9f9ddc077873dbd1556f0912a20ae4cefb3f4faaab30506dade5d6369e6e580d2b610572ae768d09f0a813782a1920228b7c2e9d83508b3d7e5bd306d79a2da8f1455745b30c23c0b297e67c99bb3b12a0954a924ac7d4925e7f77c785cb3f74ff85a5e9e3c7b38359f4f22c314177cf1155cc7505cbc7919e2525cf41c16052461e69036f1acdafa488dea7f9bc6608a25b1df13aec49ae51c6b643a8fdffaf6645d1acd1e2ab309283f6b64bc6dd1b2a718b37e1ad2d4b854d99d12afedfa42f7c42ebf6db4994eaf62f28f0d6b9f6ed172143926a785a3eb19fad8715f1f958cddc2a4a805438485ea9a5f974387f29a81a7261898c69b18ac04edfe79480849745b305ebca727537438d13f0328cef8931fbd8284cf644d89ff8d978f5dda5de5602b285230cd6d419fc63f6d1a02d649c69ecb8d093e603036faaae4a59b5d0aaed76fd539a6b4b7d7433df00893b3775d5fbe718539cdfcd0b247bc22f2ef7d5c58c99c8fd039367024ff94893bdee566bda6784a131f1633f572fdb87b406299cf345ee1de46a78e591f6d45eb37c2a4a481cce6eb6c67bb9928b3d33965e4693831e99194c41283d8f493886cca86fade836bdb6e7eac357adc0d507be9cff4d4635dc61086ba939d593607f094fb72fadc9eb648540a6b54", 0x1000}], 0x1, &(0x7f0000001240)=[@hopopts={{0xa0, 0x29, 0x36, {0x3c, 0x10, '\x00', [@enc_lim={0x4, 0x1, 0x4}, @enc_lim={0x4, 0x1, 0x2}, @generic={0x81, 0x60, "1f369872c4e4612f8bb266eb86912ac884c06ac35e7cf1eb3ff6684aab0a98432cf9e23c55d9248c910213d4b99f117bed505fe085b9ea096b8eca29bfd971a23933097d1b8268ec8a5d280837c5d2283adbe00eb1184416071d8510ced881bb"}, @ra={0x5, 0x2, 0x7245}, @hao={0xc9, 0x10, @mcast1}, @ra={0x5, 0x2, 0x6}, @enc_lim={0x4, 0x1, 0x1}]}}}, @hoplimit={{0x14, 0x29, 0x34, 0xffffffff}}, @rthdr={{0x98, 0x29, 0x39, {0x21, 0x10, 0x0, 0x4, 0x0, [@loopback, @rand_addr=' \x01\x00', @private0, @remote, @ipv4={'\x00', '\xff\xff', @dev={0xac, 0x14, 0x14, 0x2b}}, @dev={0xfe, 0x80, '\x00', 0x40}, @private0={0xfc, 0x0, '\x00', 0x1}, @loopback]}}}, @dontfrag={{0x14, 0x29, 0x3e, 0x6}}, @rthdr_2292={{0x48, 0x29, 0x39, {0x1d, 0x6, 0x2, 0x0, 0x0, [@initdev={0xfe, 0x88, '\x00', 0x1, 0x0}, @remote, @private0]}}}], 0x1b0}}, {{&(0x7f00000000c0)={0xa, 0x4e23, 0x1, @private2={0xfc, 0x2, '\x00', 0x1}, 0x7ff}, 0x1c, &(0x7f0000001540)=[{&(0x7f0000000100)="b4d537fc4ec5292b087a9a19d73d33ee91cb83f20643778c7ae309cd39bf3fe60466967057bea074b619fe631de3fab9fa36bc064ddbc3ab97a1cb29ba5a5d8b2b76c6c2bf001b45c7cc0f36d06c435f5c3d99cb4d050d1de994b075c0cfd3de7e2d27d95956b962b871377c42b8f62bf5f46433e1eb44c395903270696cdfd46528edbe80be6d09bb471450618e9de4b0cd63dec325823d8913d5dadb2fb57786ca5142b92c10847aec01aaa4118eea7e73b951f8048db1d19e0bdc46ebf6a50c3a8677146fcdcd5bea67d2", 0xcc}, {&(0x7f0000001400)="48affc6589751527c1662de5caaea4cfbf91f195e2474ad43fc8708054af73bbbc4ff13cce1df94a922b7bc9b22f8da2046691d8b2baeeaab1ba61a62f9ee5dbcede38ae2b8fd731f3b102", 0x4b}, {&(0x7f0000001480)="b2abb19bc646600056bf615c6eb86d2affd0c6a19703efa75fae356f2814da02d32c4ec3668accc6fe2152959d31764d69a2cf8897e590f587aa7c12c279996195bfa37e93d60b9032673da89cfa9cfc52f99b623f97eb38d648a4a1afda82274d503957efd6a089bb10d15498a3062a097d5ac1b688cd4439a877744a123e0f3b003f1e", 0x84}], 0x3, &(0x7f0000001580)=[@hopopts={{0x38, 0x29, 0x36, {0x62, 0x3, '\x00', [@calipso={0x7, 0x18, {0x2, 0x4, 0x7f, 0x7, [0x3, 0xffff]}}, @ra={0x5, 0x2, 0x8001}]}}}, @pktinfo={{0x24, 0x29, 0x32, {@mcast2, r3}}}, @hoplimit_2292={{0x14, 0x29, 0x8, 0x20}}, @rthdrdstopts={{0x160, 0x29, 0x37, {0x5e, 0x28, '\x00', [@jumbo={0xc2, 0x4, 0x5}, @ra={0x5, 0x2, 0x9}, @calipso={0x7, 0x50, {0x2, 0x12, 0x2, 0x0, [0x7, 0x7, 0x8, 0xe6, 0x80000000, 0x4, 0x5a3, 0x5, 0x100000000]}}, @generic={0xf2, 0xe9, "6c24ee61412a83545ce26853175fc8054839e39e56d7c7807054dec0596a3f3ead7d12fbdf75e0eac3319c2d8ee18023c547511173ddb68acd3264b3a118c8a78c9c03f65bc4ec62e9e7879110e0aed4b80d752fd329a17e51bf207baf6229afc8416cb69b00acd5bc250abc1abb41005824e38b7fde7b7c4711eba3f4c6e0b02bac902f496cc9d75928e1fe598f2e4485b88c68f5378144d52269f43fb9ca90ba4fc295aa4022402efd99fdc37c54ddf3818d7b5c36d0b81b8247df839a365cfbe0f7116b7c21a5c1e2014710ea432eff5d3a65fdaa36c317f1ec7c0b97843afae20c9d7fd7d4bf70"}]}}}], 0x1d8}}, {{0x0, 0x0, &(0x7f0000001b00)=[{&(0x7f0000001780)="5feea81a24ebddeeac", 0x9}, {&(0x7f00000017c0)="95d527e3bd5b310da8bf16bd6fe8b85ef0371db049f969ec59147745ff0ef5fbb65f6a3094408c14f69f15a07f274db20b255251da84967f0ffa9fe50be037072bcc4411d9551a3150c0c5b10926966423d765455d65a17b", 0x58}, {&(0x7f0000001840)="bb6c6bf4a9151ae18e87006298fd33f59e346e44892d5ed19ad6157bacd297f597f9c7e3ab972069154bb109d8262c82b79a0de44798afc5947611828a6ac8160bdec7a40c49150226a16427f8210246cb3482d2f84bfe8a1d36b5ef5a4c2fbfb615a48ef5d3074f01c14145b42a2e7285e4dbc1a150f9eb8a38775652c6583fe8aaf0fb5040569d90bd6682120b947b31dc", 0x92}, {&(0x7f0000001900)="9b37a484f3851e1fe1cd8bdcc339374a22197cd54ba150aa1d45526bb63108f22c75aa24a93540934912923150b647163341d2aa123de3e9ac167cd2d0d71a19091513c3374818ea981fb781df772b4baec81e912e9c2148512eacdaa7e7fbe0d7f61aad6aa19de932a7157eca8794090a25ee4eb1825403bdb698e6184ed348c175e2e7d6c0e300eed5fa750f894c05e3979e0f485327768b4777e05b2a63113a79a08dda92031b4c03e30c85cc7298e8b0776d20131478e62fb578f4c54e351c29c3d7dfafd956cc4ba3c41cf4a24ea7ae273ae57722a5f64cdb94723bae4b108e", 0xe2}, {&(0x7f0000001a00)="e524d03409aaefcdf6c6659616d99bb6e508859c33221742d2d1f9aacf84e11676f3eb074215c007feabb1225502343f4247b4c922f0cc0ac0d3350d9f6d5ac9cc2ecb0754150a873879e776c6ac06208a2c52ffc63338f333", 0x59}, {&(0x7f0000001a80)="1e8401aa707748e6574c40e9c2c49625a1994dbd1a55b56a70f376f1a281e17c84d17e80d7b6", 0x26}, {&(0x7f0000001ac0)="1c594fe43f90b6ffd9d8099baf46a84865846782e5844128d638481b6ead1734fb98f40d498aa82ba5772863d4da", 0x2e}], 0x7, &(0x7f0000001b80)=[@dstopts_2292={{0x28, 0x29, 0x4, {0x32, 0x2, '\x00', [@pad1, @pad1, @jumbo={0xc2, 0x4, 0x3}, @ra={0x5, 0x2, 0x9}]}}}, @dstopts_2292={{0xf8, 0x29, 0x4, {0x2b, 0x1b, '\x00', [@jumbo={0xc2, 0x4, 0x4}, @calipso={0x7, 0x40, {0x1, 0xe, 0x2, 0xf463, [0xfffffffffffffb46, 0xfff, 0x8, 0x3ff, 0x0, 0x6, 0x8]}}, @padn={0x1, 0x5, [0x0, 0x0, 0x0, 0x0, 0x0]}, @hao={0xc9, 0x10, @empty}, @hao={0xc9, 0x10, @dev={0xfe, 0x80, '\x00', 0xd}}, @enc_lim={0x4, 0x1, 0x2}, @pad1, @hao={0xc9, 0x10, @initdev={0xfe, 0x88, '\x00', 0x1, 0x0}}, @calipso={0x7, 0x48, {0x0, 0x10, 0x1, 0x9, [0xc3, 0x1, 0x1, 0xa0f0, 0x7, 0x8, 0x5, 0x0]}}, @ra={0x5, 0x2, 0xfffe}]}}}, @hopopts_2292={{0x20, 0x29, 0x36, {0x2e, 0x0, '\x00', [@pad1]}}}, @dontfrag={{0x14, 0x29, 0x3e, 0x1}}, @pktinfo={{0x24, 0x29, 0x32, {@rand_addr=' \x01\x00\x00\x00\x00\x00\x00\x00\x00\x00\x00\x00\x00\x00\x02'}}}, @dstopts={{0xb0, 0x29, 0x37, {0x4a, 0x12, '\x00', [@jumbo={0xc2, 0x4, 0xfedd}, @calipso={0x7, 0x58, {0x1, 0x14, 0x6, 0x3, [0x9, 0x2, 0x1, 0x4, 0x401, 0x8, 0x9, 0x2, 0xb4, 0x8]}}, @generic={0xa5, 0x25, "3cbb5e9ca85727c64c56265f036fb83230d933e58f0bbb4135ac8436f962806661674c8891"}, @pad1, @enc_lim={0x4, 0x1, 0x7}, @ra={0x5, 0x2, 0x4}]}}}], 0x230}}], 0x3, 0x10) ioctl$sock_SIOCETHTOOL(r0, 0x89f0, &(0x7f0000000000)={'bridge0\x00', &(0x7f0000000200)=ANY=[@ANYBLOB="1000000000000000000000000000001000"/36]}) 04:12:47 executing program 0: r0 = socket$netlink(0x10, 0x3, 0x0) r1 = socket$netlink(0x10, 0x3, 0x0) r2 = socket(0x10, 0x803, 0x0) sendmsg$NL80211_CMD_CRIT_PROTOCOL_START(r2, &(0x7f0000000580)={0x0, 0x0, &(0x7f0000000540)={0x0, 0x1c}}, 0x0) getsockname$packet(r2, &(0x7f0000000600)={0x11, 0x0, 0x0, 0x1, 0x0, 0x6, @broadcast}, &(0x7f0000000100)=0x14) sendmsg$nl_route(r1, &(0x7f0000000040)={0x0, 0x0, &(0x7f0000000000)={&(0x7f00000000c0)=ANY=[@ANYBLOB="3c00000010008506eb9afc4cd8d06e754a0081c5", @ANYRES32=r3, @ANYBLOB="2377f292252155b21c0012000c000100626f6e64000000000c0002000800010001"], 0x3c}}, 0x0) sendmsg$nl_route(r0, &(0x7f0000000240)={0x0, 0x0, &(0x7f0000000680)={&(0x7f0000000400)=@newlink={0x3c, 0x10, 0xffffff1f, 0x0, 0x253c35, {}, [@IFLA_LINKINFO={0x14, 0x12, 0x0, 0x1, @bridge={{0xb}, {0x4}}}, @IFLA_MASTER={0x8, 0xa, r3}]}, 0x3c}}, 0x0) [ 2840.439754][ T3354] bond1288: (slave bridge1209): making interface the new active one [ 2840.465991][ T3354] bridge1209: entered promiscuous mode [ 2840.488293][ T3354] bond1288: (slave bridge1209): Enslaving as an active interface with an up link [ 2840.517907][ T3361] netlink: 'syz-executor.1': attribute type 1 has an invalid length. 04:12:48 executing program 2: r0 = socket$netlink(0x10, 0x3, 0x0) r1 = socket$netlink(0x10, 0x3, 0x0) r2 = socket(0x10, 0x803, 0x0) sendmsg$NL80211_CMD_CRIT_PROTOCOL_START(r2, &(0x7f0000000580)={0x0, 0x0, &(0x7f0000000540)={0x0, 0x1c}}, 0x0) getsockname$packet(r2, &(0x7f0000000600)={0x11, 0x0, 0x0, 0x1, 0x0, 0x6, @broadcast}, &(0x7f0000000100)=0x14) sendmsg$nl_route(r1, &(0x7f0000000040)={0x0, 0x0, &(0x7f0000000000)={&(0x7f00000000c0)=ANY=[@ANYBLOB="3c00000010008506eb9afc4cd8d06e754a0081c5", @ANYRES32=r3, @ANYBLOB="2377f292252155b21c0012000c000100626f6e64000000000c0002000800010001"], 0x3c}}, 0x0) sendmsg$nl_route(r0, &(0x7f0000000240)={0x0, 0x0, &(0x7f0000000680)={&(0x7f0000000400)=@newlink={0x3c, 0x10, 0xffffff1f, 0x0, 0x253c3d, {}, [@IFLA_LINKINFO={0x14, 0x12, 0x0, 0x1, @bridge={{0xb}, {0x4}}}, @IFLA_MASTER={0x8, 0xa, r3}]}, 0x3c}}, 0x0) [ 2840.627791][ T3361] bond1305: entered promiscuous mode [ 2840.636485][ T3361] 8021q: adding VLAN 0 to HW filter on device bond1305 [ 2840.696337][ T3363] bond1305: (slave bridge1251): making interface the new active one [ 2840.710486][ T3363] bridge1251: entered promiscuous mode [ 2840.731162][ T3363] bond1305: (slave bridge1251): Enslaving as an active interface with an up link 04:12:48 executing program 1: r0 = socket$netlink(0x10, 0x3, 0x0) r1 = socket$netlink(0x10, 0x3, 0x0) r2 = socket(0x10, 0x803, 0x0) sendmsg$NL80211_CMD_CRIT_PROTOCOL_START(r2, &(0x7f0000000580)={0x0, 0x0, &(0x7f0000000540)={0x0, 0x1c}}, 0x0) getsockname$packet(r2, &(0x7f0000000600)={0x11, 0x0, 0x0, 0x1, 0x0, 0x6, @broadcast}, &(0x7f0000000100)=0x14) sendmsg$nl_route(r1, &(0x7f0000000040)={0x0, 0x0, &(0x7f0000000000)={&(0x7f00000000c0)=ANY=[@ANYBLOB="3c00000010008506eb9afc4cd8d06e754a0081c5", @ANYRES32=r3, @ANYBLOB="2377f292252155b21c0012000c000100626f6e64000000000c0002000800010001"], 0x3c}}, 0x0) sendmsg$nl_route(r0, &(0x7f0000000240)={0x0, 0x0, &(0x7f0000000680)={&(0x7f0000000400)=@newlink={0x3c, 0x10, 0xffffff1f, 0x0, 0x1d672500, {}, [@IFLA_LINKINFO={0x14, 0x12, 0x0, 0x1, @bridge={{0xb}, {0x4}}}, @IFLA_MASTER={0x8, 0xa, r3}]}, 0x3c}}, 0x0) [ 2840.880224][ T3374] netlink: 'syz-executor.5': attribute type 1 has an invalid length. 04:12:48 executing program 3: r0 = socket$netlink(0x10, 0x3, 0x0) sendmsg$netlink(r0, &(0x7f00000047c0)={0x0, 0x0, &(0x7f0000000000)=[{&(0x7f0000000100)=ANY=[@ANYBLOB="18000000560001060000008700db730007"], 0x18}], 0x1}, 0x0) [ 2840.915765][ T3374] workqueue: Failed to create a rescuer kthread for wq "bond1298": -EINTR [ 2841.078265][ T3382] netlink: 'syz-executor.0': attribute type 1 has an invalid length. 04:12:48 executing program 5: r0 = socket$netlink(0x10, 0x3, 0x0) r1 = socket$netlink(0x10, 0x3, 0x0) r2 = socket(0x10, 0x803, 0x0) sendmsg$NL80211_CMD_CRIT_PROTOCOL_START(r2, &(0x7f0000000580)={0x0, 0x0, &(0x7f0000000540)={0x0, 0x1c}}, 0x0) getsockname$packet(r2, &(0x7f0000000600)={0x11, 0x0, 0x0, 0x1, 0x0, 0x6, @broadcast}, &(0x7f0000000100)=0x14) sendmsg$nl_route(r1, &(0x7f0000000040)={0x0, 0x0, &(0x7f0000000000)={&(0x7f00000000c0)=ANY=[@ANYBLOB="3c00000010008506eb9afc4cd8d06e754a0081c5", @ANYRES32=r3, @ANYBLOB="2377f292252155b21c0012000c000100626f6e64000000000c0002000800010001"], 0x3c}}, 0x0) sendmsg$nl_route(r0, &(0x7f0000000240)={0x0, 0x0, &(0x7f0000000680)={&(0x7f0000000400)=@newlink={0x3c, 0x10, 0xffffff1f, 0x0, 0x1f000000, {}, [@IFLA_LINKINFO={0x14, 0x12, 0x0, 0x1, @bridge={{0xb}, {0x4}}}, @IFLA_MASTER={0x8, 0xa, r3}]}, 0x3c}}, 0x0) [ 2841.145685][ T3382] bond10: entered promiscuous mode [ 2841.151386][ T3382] 8021q: adding VLAN 0 to HW filter on device bond10 04:12:48 executing program 4: r0 = socket$inet6(0xa, 0x2, 0x0) ioctl$sock_SIOCETHTOOL(r0, 0x89f0, &(0x7f0000000000)={'bridge0\x00', &(0x7f0000000200)=@ethtool_ringparam={0x10}}) write$binfmt_misc(r0, &(0x7f0000000040)=ANY=[@ANYBLOB="73797a3191132a3ee5279ea5d7e758b00dd3e0bf1e4a411fe4c63586aefa14c0a4693a9c2ebe8ca6efadae70e0bafc07c986fb7827cd55f1b0fae82ec76aa5c6b85b178ac5f2218f282d581a91045874c25777845ffb43695324e94e9c892700000000"], 0x64) [ 2841.288137][ T3387] bond10: (slave bridge18): making interface the new active one [ 2841.317893][ T3387] bridge18: entered promiscuous mode [ 2841.368855][ T3387] bond10: (slave bridge18): Enslaving as an active interface with an up link [ 2841.391958][ T3390] netlink: 'syz-executor.2': attribute type 1 has an invalid length. 04:12:48 executing program 0: r0 = socket$inet6(0xa, 0x6, 0x0) bind$inet6(r0, &(0x7f0000000000)={0xa, 0x4e20}, 0x1c) listen(r0, 0x6) r1 = socket$inet6(0xa, 0x6, 0x0) connect$inet6(r1, &(0x7f0000000200)={0xa, 0x0, 0x0, @initdev={0xfe, 0x88, '\x00', 0x0, 0x0}, 0x3}, 0x1c) r2 = accept4$inet6(r0, &(0x7f0000000080)={0xa, 0x0, 0x0, @private2}, &(0x7f00000000c0)=0x1c, 0x80000) setsockopt$inet6_group_source_req(r2, 0x29, 0x2c, &(0x7f0000000240)={0x7fffffff, {{0xa, 0x4e24, 0x3, @mcast1, 0x3}}, {{0xa, 0x4e20, 0x3f, @mcast2, 0xd52}}}, 0x108) connect$inet6(r1, &(0x7f0000000040)={0xa, 0x4e20, 0x0, @dev={0xfe, 0x80, '\x00', 0xf}}, 0x1c) sendmmsg$inet6(r1, &(0x7f0000005a00)=[{{0x0, 0x0, 0x0}}, {{0x0, 0x0, 0x0, 0x0, &(0x7f0000005bc0)=ANY=[@ANYBLOB="5000000000000000290000003600000000060000000000000401000730000000000a0000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000020000000000000002900000037000000000000000000000004010000000000001400000000000000290000000000000160000000000000002900000037"], 0x100}}], 0x2, 0x0) [ 2841.550185][ T3390] bond1289: entered promiscuous mode [ 2841.651968][ T3390] 8021q: adding VLAN 0 to HW filter on device bond1289 [ 2841.829199][ T3391] bond1289: (slave bridge1210): making interface the new active one [ 2841.849615][ T3391] bridge1210: entered promiscuous mode 04:12:49 executing program 0: r0 = socket$inet6(0xa, 0x6, 0x0) bind$inet6(r0, &(0x7f0000000000)={0xa, 0x4e20}, 0x1c) listen(r0, 0x6) bind$inet6(r0, &(0x7f0000000080)={0xa, 0x4e21, 0x7fff, @mcast1, 0x9}, 0x1c) r1 = socket$inet6(0xa, 0x6, 0x0) connect$inet6(r1, &(0x7f0000000200)={0xa, 0x0, 0x0, @initdev={0xfe, 0x88, '\x00', 0x0, 0x0}, 0x3}, 0x1c) connect$inet6(r1, &(0x7f0000000040)={0xa, 0x4e20, 0x0, @dev={0xfe, 0x80, '\x00', 0xf}}, 0x1c) sendmmsg$inet6(r1, &(0x7f0000005a00)=[{{0x0, 0x0, 0x0}}, {{0x0, 0x0, 0x0, 0x0, &(0x7f0000005bc0)=ANY=[@ANYBLOB="5000000000000000290000003600000000060000000000000401000730000000000a0000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000020000000000000002900000037000000000000000000000004010000000000001400000000000000290000000000000160000000000000002900000037"], 0x100}}], 0x2, 0x0) [ 2841.875850][ T3391] bond1289: (slave bridge1210): Enslaving as an active interface with an up link 04:12:49 executing program 2: r0 = socket$netlink(0x10, 0x3, 0x0) r1 = socket$netlink(0x10, 0x3, 0x0) r2 = socket(0x10, 0x803, 0x0) sendmsg$NL80211_CMD_CRIT_PROTOCOL_START(r2, &(0x7f0000000580)={0x0, 0x0, &(0x7f0000000540)={0x0, 0x1c}}, 0x0) getsockname$packet(r2, &(0x7f0000000600)={0x11, 0x0, 0x0, 0x1, 0x0, 0x6, @broadcast}, &(0x7f0000000100)=0x14) sendmsg$nl_route(r1, &(0x7f0000000040)={0x0, 0x0, &(0x7f0000000000)={&(0x7f00000000c0)=ANY=[@ANYBLOB="3c00000010008506eb9afc4cd8d06e754a0081c5", @ANYRES32=r3, @ANYBLOB="2377f292252155b21c0012000c000100626f6e64000000000c0002000800010001"], 0x3c}}, 0x0) sendmsg$nl_route(r0, &(0x7f0000000240)={0x0, 0x0, &(0x7f0000000680)={&(0x7f0000000400)=@newlink={0x3c, 0x10, 0xffffff1f, 0x0, 0x253c3e, {}, [@IFLA_LINKINFO={0x14, 0x12, 0x0, 0x1, @bridge={{0xb}, {0x4}}}, @IFLA_MASTER={0x8, 0xa, r3}]}, 0x3c}}, 0x0) 04:12:49 executing program 0: r0 = socket$inet6(0xa, 0x6, 0x0) bind$inet6(r0, &(0x7f0000000000)={0xa, 0x4e20}, 0x1c) listen(r0, 0x6) r1 = socket$inet6(0xa, 0x6, 0x0) connect$inet6(r1, &(0x7f0000000200)={0xa, 0x0, 0x0, @initdev={0xfe, 0x88, '\x00', 0x0, 0x0}, 0x3}, 0x1c) connect$inet6(r1, &(0x7f0000000040)={0xa, 0x4e20, 0x0, @dev={0xfe, 0x80, '\x00', 0xf}}, 0x1c) sendmmsg$inet6(r1, &(0x7f0000005a00)=[{{0x0, 0x0, 0x0}}, {{0x0, 0x0, 0x0, 0x0, &(0x7f0000000080)=ANY=[@ANYBLOB="5000000000000000290000080000000000000000000000000401000730000000000a0000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000020000000000000002900000037000000f3ff00000000000004010000000000001400000000000000290000000000000160000000006aa70d8a7248e76dd8acfb29ef872325f19e6982aa494e20ad6244060c5872c9ff5c8ea2eb5e5c33e41b5c83003bc1ed83fe6d32680a6b1f05a0b2ce472bcf30b8a78594fd208094c4cc767810d83ff1b1f3c91349a3665f2ab35f7a41ef5400"], 0x100}}], 0x2, 0x20000010) 04:12:49 executing program 0: r0 = socket$netlink(0x10, 0x3, 0x0) r1 = socket$netlink(0x10, 0x3, 0x0) r2 = socket(0x10, 0x803, 0x0) sendmsg$NL80211_CMD_CRIT_PROTOCOL_START(r2, &(0x7f0000000580)={0x0, 0x0, &(0x7f0000000540)={0x0, 0x1c}}, 0x0) getsockname$packet(r2, &(0x7f0000000600)={0x11, 0x0, 0x0, 0x1, 0x0, 0x6, @broadcast}, &(0x7f0000000100)=0x14) sendmsg$nl_route(r1, &(0x7f0000000040)={0x0, 0x0, &(0x7f0000000000)={&(0x7f00000000c0)=ANY=[@ANYBLOB="3c00000010008506eb9afc4cd8d06e754a0081c5", @ANYRES32=r3, @ANYBLOB="2377f292252155b21c0012000c000100626f6e64000000000c0002000800010001"], 0x3c}}, 0x0) sendmsg$nl_route(r0, &(0x7f0000000240)={0x0, 0x0, &(0x7f0000000680)={&(0x7f0000000400)=@newlink={0x3c, 0x10, 0xffffff1f, 0x0, 0x25676e, {}, [@IFLA_LINKINFO={0x14, 0x12, 0x0, 0x1, @bridge={{0xb}, {0x4}}}, @IFLA_MASTER={0x8, 0xa, r3}]}, 0x3c}}, 0x0) 04:12:49 executing program 3: r0 = socket$netlink(0x10, 0x3, 0x0) sendmsg$netlink(r0, &(0x7f00000047c0)={0x0, 0x0, &(0x7f0000000000)=[{&(0x7f0000000100)=ANY=[@ANYBLOB="18000000560001060000008800db730007"], 0x18}], 0x1}, 0x0) [ 2842.170561][ T3393] workqueue: Failed to create a rescuer kthread for wq "bond1306": -EINTR 04:12:50 executing program 1: r0 = socket$netlink(0x10, 0x3, 0x0) r1 = socket$netlink(0x10, 0x3, 0x0) r2 = socket(0x10, 0x803, 0x0) sendmsg$NL80211_CMD_CRIT_PROTOCOL_START(r2, &(0x7f0000000580)={0x0, 0x0, &(0x7f0000000540)={0x0, 0x1c}}, 0x0) getsockname$packet(r2, &(0x7f0000000600)={0x11, 0x0, 0x0, 0x1, 0x0, 0x6, @broadcast}, &(0x7f0000000100)=0x14) sendmsg$nl_route(r1, &(0x7f0000000040)={0x0, 0x0, &(0x7f0000000000)={&(0x7f00000000c0)=ANY=[@ANYBLOB="3c00000010008506eb9afc4cd8d06e754a0081c5", @ANYRES32=r3, @ANYBLOB="2377f292252155b21c0012000c000100626f6e64000000000c0002000800010001"], 0x3c}}, 0x0) sendmsg$nl_route(r0, &(0x7f0000000240)={0x0, 0x0, &(0x7f0000000680)={&(0x7f0000000400)=@newlink={0x3c, 0x10, 0xffffff1f, 0x0, 0x1e672500, {}, [@IFLA_LINKINFO={0x14, 0x12, 0x0, 0x1, @bridge={{0xb}, {0x4}}}, @IFLA_MASTER={0x8, 0xa, r3}]}, 0x3c}}, 0x0) 04:12:50 executing program 4: r0 = socket$inet6(0xa, 0x2, 0x0) ioctl$sock_SIOCETHTOOL(r0, 0x89f0, &(0x7f0000000000)={'bridge0\x00', &(0x7f0000000200)=@ethtool_ringparam={0x10}}) write$binfmt_misc(r0, &(0x7f0000000040)=ANY=[@ANYBLOB="73797a3191132a3ee5279ea5d7e758b00dd3e0bf1e4a411fe4c63586aefa14c0a4693a9c2ebe8ca6efadae70e0bafc07c986fb7827cd55f1b0fae82ec76aa5c6b85b178ac5f2218f282d581a91045874c25777845ffb43695324e94e9c892700000000"], 0x64) socket$inet6(0xa, 0x2, 0x0) (async) ioctl$sock_SIOCETHTOOL(r0, 0x89f0, &(0x7f0000000000)={'bridge0\x00', &(0x7f0000000200)=@ethtool_ringparam={0x10}}) (async) write$binfmt_misc(r0, &(0x7f0000000040)=ANY=[@ANYBLOB="73797a3191132a3ee5279ea5d7e758b00dd3e0bf1e4a411fe4c63586aefa14c0a4693a9c2ebe8ca6efadae70e0bafc07c986fb7827cd55f1b0fae82ec76aa5c6b85b178ac5f2218f282d581a91045874c25777845ffb43695324e94e9c892700000000"], 0x64) (async) 04:12:50 executing program 5: r0 = socket$netlink(0x10, 0x3, 0x0) r1 = socket$netlink(0x10, 0x3, 0x0) r2 = socket(0x10, 0x803, 0x0) sendmsg$NL80211_CMD_CRIT_PROTOCOL_START(r2, &(0x7f0000000580)={0x0, 0x0, &(0x7f0000000540)={0x0, 0x1c}}, 0x0) getsockname$packet(r2, &(0x7f0000000600)={0x11, 0x0, 0x0, 0x1, 0x0, 0x6, @broadcast}, &(0x7f0000000100)=0x14) sendmsg$nl_route(r1, &(0x7f0000000040)={0x0, 0x0, &(0x7f0000000000)={&(0x7f00000000c0)=ANY=[@ANYBLOB="3c00000010008506eb9afc4cd8d06e754a0081c5", @ANYRES32=r3, @ANYBLOB="2377f292252155b21c0012000c000100626f6e64000000000c0002000800010001"], 0x3c}}, 0x0) sendmsg$nl_route(r0, &(0x7f0000000240)={0x0, 0x0, &(0x7f0000000680)={&(0x7f0000000400)=@newlink={0x3c, 0x10, 0xffffff1f, 0x0, 0x20000000, {}, [@IFLA_LINKINFO={0x14, 0x12, 0x0, 0x1, @bridge={{0xb}, {0x4}}}, @IFLA_MASTER={0x8, 0xa, r3}]}, 0x3c}}, 0x0) [ 2842.446082][ T3401] workqueue: Failed to create a rescuer kthread for wq "bond1298": -EINTR [ 2842.949420][ T3418] bond1290: entered promiscuous mode [ 2842.980688][ T3418] 8021q: adding VLAN 0 to HW filter on device bond1290 [ 2843.132956][ T3419] bond1290: (slave bridge1211): making interface the new active one [ 2843.141408][ T3419] bridge1211: entered promiscuous mode [ 2843.158889][ T3419] bond1290: (slave bridge1211): Enslaving as an active interface with an up link [ 2843.175427][ T3422] validate_nla: 3 callbacks suppressed 04:12:50 executing program 2: r0 = socket$netlink(0x10, 0x3, 0x0) r1 = socket$netlink(0x10, 0x3, 0x0) r2 = socket(0x10, 0x803, 0x0) sendmsg$NL80211_CMD_CRIT_PROTOCOL_START(r2, &(0x7f0000000580)={0x0, 0x0, &(0x7f0000000540)={0x0, 0x1c}}, 0x0) getsockname$packet(r2, &(0x7f0000000600)={0x11, 0x0, 0x0, 0x1, 0x0, 0x6, @broadcast}, &(0x7f0000000100)=0x14) sendmsg$nl_route(r1, &(0x7f0000000040)={0x0, 0x0, &(0x7f0000000000)={&(0x7f00000000c0)=ANY=[@ANYBLOB="3c00000010008506eb9afc4cd8d06e754a0081c5", @ANYRES32=r3, @ANYBLOB="2377f292252155b21c0012000c000100626f6e64000000000c0002000800010001"], 0x3c}}, 0x0) sendmsg$nl_route(r0, &(0x7f0000000240)={0x0, 0x0, &(0x7f0000000680)={&(0x7f0000000400)=@newlink={0x3c, 0x10, 0xffffff1f, 0x0, 0x253c3f, {}, [@IFLA_LINKINFO={0x14, 0x12, 0x0, 0x1, @bridge={{0xb}, {0x4}}}, @IFLA_MASTER={0x8, 0xa, r3}]}, 0x3c}}, 0x0) [ 2843.175450][ T3422] netlink: 'syz-executor.0': attribute type 1 has an invalid length. [ 2843.263236][ T3422] bond11: entered promiscuous mode [ 2843.268793][ T3422] 8021q: adding VLAN 0 to HW filter on device bond11 [ 2843.300051][ T3427] bond11: (slave bridge19): making interface the new active one 04:12:50 executing program 3: r0 = socket$netlink(0x10, 0x3, 0x0) sendmsg$netlink(r0, &(0x7f00000047c0)={0x0, 0x0, &(0x7f0000000000)=[{&(0x7f0000000100)=ANY=[@ANYBLOB="18000000560001060000008900db730007"], 0x18}], 0x1}, 0x0) [ 2843.310756][ T3427] bridge19: entered promiscuous mode [ 2843.321210][ T3427] bond11: (slave bridge19): Enslaving as an active interface with an up link 04:12:50 executing program 0: r0 = socket$inet6(0xa, 0x6, 0x0) bind$inet6(r0, &(0x7f0000000000)={0xa, 0x4e20}, 0x1c) listen(r0, 0x6) r1 = socket$inet6(0xa, 0x6, 0x0) connect$inet6(r1, &(0x7f0000000200)={0xa, 0x4, 0xbffffffe, @remote, 0x3}, 0x1c) connect$inet6(r1, &(0x7f0000000040)={0xa, 0x4e20, 0x0, @dev={0xfe, 0x80, '\x00', 0xf}}, 0x1c) sendmmsg$inet6(r1, &(0x7f0000005a00)=[{{0x0, 0x0, 0x0}}, {{0x0, 0x0, 0x0, 0x0, &(0x7f0000005bc0)=ANY=[@ANYBLOB="5000000000000000290000003600000000060000000000000401000730000000000a0000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000020000000000000002900000037000000000000000000000004010000000000001400000000000000290000000000000160000000000000002900000037"], 0x100}}], 0x2, 0x0) [ 2843.545521][ T3428] netlink: 'syz-executor.1': attribute type 1 has an invalid length. [ 2843.603378][ T3428] bond1306: entered promiscuous mode [ 2843.609336][ T3428] 8021q: adding VLAN 0 to HW filter on device bond1306 [ 2843.757031][ T3430] bond1306: (slave bridge1252): making interface the new active one [ 2843.767304][ T3430] bridge1252: entered promiscuous mode [ 2843.788048][ T3430] bond1306: (slave bridge1252): Enslaving as an active interface with an up link 04:12:51 executing program 1: r0 = socket$netlink(0x10, 0x3, 0x0) r1 = socket$netlink(0x10, 0x3, 0x0) r2 = socket(0x10, 0x803, 0x0) sendmsg$NL80211_CMD_CRIT_PROTOCOL_START(r2, &(0x7f0000000580)={0x0, 0x0, &(0x7f0000000540)={0x0, 0x1c}}, 0x0) getsockname$packet(r2, &(0x7f0000000600)={0x11, 0x0, 0x0, 0x1, 0x0, 0x6, @broadcast}, &(0x7f0000000100)=0x14) sendmsg$nl_route(r1, &(0x7f0000000040)={0x0, 0x0, &(0x7f0000000000)={&(0x7f00000000c0)=ANY=[@ANYBLOB="3c00000010008506eb9afc4cd8d06e754a0081c5", @ANYRES32=r3, @ANYBLOB="2377f292252155b21c0012000c000100626f6e64000000000c0002000800010001"], 0x3c}}, 0x0) sendmsg$nl_route(r0, &(0x7f0000000240)={0x0, 0x0, &(0x7f0000000680)={&(0x7f0000000400)=@newlink={0x3c, 0x10, 0xffffff1f, 0x0, 0x1f000000, {}, [@IFLA_LINKINFO={0x14, 0x12, 0x0, 0x1, @bridge={{0xb}, {0x4}}}, @IFLA_MASTER={0x8, 0xa, r3}]}, 0x3c}}, 0x0) [ 2843.819156][ T3433] netlink: 'syz-executor.5': attribute type 1 has an invalid length. [ 2843.943163][ T3433] bond1298: entered promiscuous mode [ 2843.949174][ T3433] 8021q: adding VLAN 0 to HW filter on device bond1298 [ 2844.094905][ T3435] bond1298: (slave bridge1261): making interface the new active one [ 2844.115425][ T3435] bridge1261: entered promiscuous mode [ 2844.158393][ T3435] bond1298: (slave bridge1261): Enslaving as an active interface with an up link 04:12:51 executing program 5: r0 = socket$netlink(0x10, 0x3, 0x0) r1 = socket$netlink(0x10, 0x3, 0x0) r2 = socket(0x10, 0x803, 0x0) sendmsg$NL80211_CMD_CRIT_PROTOCOL_START(r2, &(0x7f0000000580)={0x0, 0x0, &(0x7f0000000540)={0x0, 0x1c}}, 0x0) getsockname$packet(r2, &(0x7f0000000600)={0x11, 0x0, 0x0, 0x1, 0x0, 0x6, @broadcast}, &(0x7f0000000100)=0x14) sendmsg$nl_route(r1, &(0x7f0000000040)={0x0, 0x0, &(0x7f0000000000)={&(0x7f00000000c0)=ANY=[@ANYBLOB="3c00000010008506eb9afc4cd8d06e754a0081c5", @ANYRES32=r3, @ANYBLOB="2377f292252155b21c0012000c000100626f6e64000000000c0002000800010001"], 0x3c}}, 0x0) sendmsg$nl_route(r0, &(0x7f0000000240)={0x0, 0x0, &(0x7f0000000680)={&(0x7f0000000400)=@newlink={0x3c, 0x10, 0xffffff1f, 0x0, 0x20030000, {}, [@IFLA_LINKINFO={0x14, 0x12, 0x0, 0x1, @bridge={{0xb}, {0x4}}}, @IFLA_MASTER={0x8, 0xa, r3}]}, 0x3c}}, 0x0) 04:12:51 executing program 4: r0 = socket$inet6(0xa, 0x2, 0x0) ioctl$sock_SIOCETHTOOL(r0, 0x89f0, &(0x7f0000000000)={'bridge0\x00', &(0x7f0000000200)=@ethtool_ringparam={0x10}}) write$binfmt_misc(r0, &(0x7f0000000040)=ANY=[@ANYBLOB="73797a3191132a3ee5279ea5d7e758b00dd3e0bf1e4a411fe4c63586aefa14c0a4693a9c2ebe8ca6efadae70e0bafc07c986fb7827cd55f1b0fae82ec76aa5c6b85b178ac5f2218f282d581a91045874c25777845ffb43695324e94e9c892700000000"], 0x64) (async) write$binfmt_misc(r0, &(0x7f0000000040)=ANY=[@ANYBLOB="73797a3191132a3ee5279ea5d7e758b00dd3e0bf1e4a411fe4c63586aefa14c0a4693a9c2ebe8ca6efadae70e0bafc07c986fb7827cd55f1b0fae82ec76aa5c6b85b178ac5f2218f282d581a91045874c25777845ffb43695324e94e9c892700000000"], 0x64) [ 2844.319773][ T3441] netlink: 'syz-executor.2': attribute type 1 has an invalid length. 04:12:51 executing program 0: r0 = socket$inet6(0xa, 0x6, 0x0) bind$inet6(r0, &(0x7f0000000000)={0xa, 0x4e20}, 0x1c) listen(r0, 0x6) r1 = socket$inet6(0xa, 0x6, 0x0) connect$inet6(r1, &(0x7f0000000200)={0xa, 0x0, 0x0, @initdev={0xfe, 0x88, '\x00', 0x0, 0x0}, 0x3}, 0x1c) connect$inet6(r1, &(0x7f0000000040)={0xa, 0x4e20, 0x0, @dev={0xfe, 0x80, '\x00', 0xf}}, 0x1c) sendmmsg$inet6(r1, &(0x7f0000005a00)=[{{0x0, 0x0, 0x0}}, {{0x0, 0x0, 0x0, 0x0, &(0x7f0000005bc0)=ANY=[@ANYBLOB="5000000000000000290000003600000000060000000000000401000730000000000a0000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000020000000000000002900000037000000000000000000000004010000000000001400000000000000290000000000000160000000000000002900000037"], 0x100}}], 0x2, 0x62000000) [ 2844.383944][ T3441] bond1291: entered promiscuous mode [ 2844.389713][ T3441] 8021q: adding VLAN 0 to HW filter on device bond1291 04:12:51 executing program 0: r0 = socket$inet6(0xa, 0x6, 0x0) bind$inet6(r0, &(0x7f0000000000)={0xa, 0x4e20}, 0x1c) listen(r0, 0x6) r1 = socket$inet6(0xa, 0x6, 0x0) connect$inet6(r1, &(0x7f0000000200)={0xa, 0x0, 0x0, @initdev={0xfe, 0x88, '\x00', 0x0, 0x0}, 0x3}, 0x1c) connect$inet6(r1, &(0x7f0000000040)={0xa, 0x4e20, 0x0, @dev={0xfe, 0x80, '\x00', 0xf}}, 0x1c) sendmmsg$inet6(r1, &(0x7f0000005a00)=[{{0x0, 0x0, 0x0}}, {{0x0, 0x0, 0x0, 0x0, &(0x7f0000005bc0)=ANY=[@ANYBLOB="5000000000000000290000003600000000060000000000000401000730000000000a0000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000020000000000000002900000037000000000000000000000004010000000000001400000000000000290000000000000160000000000000002900000037"], 0x100}}], 0x2, 0x1f4) 04:12:52 executing program 2: r0 = socket$netlink(0x10, 0x3, 0x0) r1 = socket$netlink(0x10, 0x3, 0x0) r2 = socket(0x10, 0x803, 0x0) sendmsg$NL80211_CMD_CRIT_PROTOCOL_START(r2, &(0x7f0000000580)={0x0, 0x0, &(0x7f0000000540)={0x0, 0x1c}}, 0x0) getsockname$packet(r2, &(0x7f0000000600)={0x11, 0x0, 0x0, 0x1, 0x0, 0x6, @broadcast}, &(0x7f0000000100)=0x14) sendmsg$nl_route(r1, &(0x7f0000000040)={0x0, 0x0, &(0x7f0000000000)={&(0x7f00000000c0)=ANY=[@ANYBLOB="3c00000010008506eb9afc4cd8d06e754a0081c5", @ANYRES32=r3, @ANYBLOB="2377f292252155b21c0012000c000100626f6e64000000000c0002000800010001"], 0x3c}}, 0x0) sendmsg$nl_route(r0, &(0x7f0000000240)={0x0, 0x0, &(0x7f0000000680)={&(0x7f0000000400)=@newlink={0x3c, 0x10, 0xffffff1f, 0x0, 0x253c40, {}, [@IFLA_LINKINFO={0x14, 0x12, 0x0, 0x1, @bridge={{0xb}, {0x4}}}, @IFLA_MASTER={0x8, 0xa, r3}]}, 0x3c}}, 0x0) [ 2844.478907][ T3444] bond1291: (slave bridge1212): making interface the new active one [ 2844.487102][ T3444] bridge1212: entered promiscuous mode [ 2844.496991][ T3444] bond1291: (slave bridge1212): Enslaving as an active interface with an up link [ 2844.509317][ T3453] netlink: 'syz-executor.1': attribute type 1 has an invalid length. [ 2844.648948][ T3453] bond1307: entered promiscuous mode [ 2844.655206][ T3453] 8021q: adding VLAN 0 to HW filter on device bond1307 04:12:52 executing program 0: r0 = socket$inet6(0xa, 0x6, 0x0) bind$inet6(r0, &(0x7f0000000000)={0xa, 0x4e20}, 0x1c) listen(r0, 0x6) r1 = socket$inet6(0xa, 0x6, 0x0) connect$inet6(r1, &(0x7f0000000200)={0xa, 0x0, 0x0, @initdev={0xfe, 0x88, '\x00', 0x0, 0x0}, 0x3}, 0x1c) connect$inet6(r1, &(0x7f0000000040)={0xa, 0x4e20, 0x0, @dev={0xfe, 0x80, '\x00', 0xf}}, 0x1c) sendmmsg$inet6(r1, &(0x7f0000005a00)=[{{0x0, 0x0, 0x0}}, {{0x0, 0x0, 0x0, 0x0, &(0x7f0000005bc0)=ANY=[@ANYBLOB="5000000000000000290000003600000000060000000000000401000730000000000a0000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000020000000000000002900000037000000000000000000000004010000000000001400000000000000290000000000000160000000000000002900000037"], 0x100}}], 0x1a000, 0x0) 04:12:52 executing program 0: r0 = socket$inet6(0xa, 0x6, 0x0) bind$inet6(r0, &(0x7f0000000000)={0xa, 0x4e20}, 0x1c) listen(r0, 0x6) r1 = socket$inet6(0xa, 0x6, 0x0) connect$inet6(r1, &(0x7f0000000200)={0xa, 0x0, 0x0, @initdev={0xfe, 0x88, '\x00', 0x0, 0x0}, 0x3}, 0x1c) connect$inet6(r1, &(0x7f0000000040)={0xa, 0x4e20, 0x0, @dev={0xfe, 0x80, '\x00', 0xf}}, 0x1c) sendmmsg$inet6(r1, &(0x7f0000005a00)=[{{0x0, 0x0, 0x0}}, {{0x0, 0x0, 0x0, 0x0, &(0x7f0000005bc0)=ANY=[@ANYBLOB="5000000000000000290000003600000000060000000000000401000730000000000a0000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000020000000000000002900000037000000000000000000000004010000000000001400000000000000290000000000000160000000000000002900000037"], 0x100}, 0x11000}], 0x2, 0x0) [ 2844.898611][ T3454] bond1307: (slave bridge1253): making interface the new active one [ 2844.928110][ T3454] bridge1253: entered promiscuous mode [ 2844.968741][ T3454] bond1307: (slave bridge1253): Enslaving as an active interface with an up link [ 2845.008847][ T3457] netlink: 'syz-executor.5': attribute type 1 has an invalid length. 04:12:52 executing program 0: r0 = socket$inet6(0xa, 0x6, 0x0) bind$inet6(r0, &(0x7f0000000000)={0xa, 0x4e20}, 0x1c) listen(r0, 0x6) r1 = socket$inet6(0xa, 0x6, 0x0) connect$inet6(r1, &(0x7f0000000200)={0xa, 0x0, 0x0, @initdev={0xfe, 0x88, '\x00', 0x0, 0x0}, 0x3}, 0x1c) connect$inet6(r1, &(0x7f0000000040)={0xa, 0x4e20, 0x0, @dev={0xfe, 0x80, '\x00', 0xf}}, 0x1c) sendmmsg$inet6(r1, &(0x7f0000005a00)=[{{0x0, 0x0, 0x0}}, {{0x0, 0x0, 0x0, 0x0, &(0x7f0000005bc0)=ANY=[@ANYBLOB="5000000000000000290000003600000000060000000000000401000730000000000a0000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000020000000000000002900000037000000000000000000000004010000000000001400000000000000290000000000000160000000000000002900000037"], 0x100}, 0x11000}], 0x2, 0x0) [ 2845.101158][ T3457] bond1299: entered promiscuous mode [ 2845.107139][ T3457] 8021q: adding VLAN 0 to HW filter on device bond1299 04:12:52 executing program 1: r0 = socket$netlink(0x10, 0x3, 0x0) r1 = socket$netlink(0x10, 0x3, 0x0) r2 = socket(0x10, 0x803, 0x0) sendmsg$NL80211_CMD_CRIT_PROTOCOL_START(r2, &(0x7f0000000580)={0x0, 0x0, &(0x7f0000000540)={0x0, 0x1c}}, 0x0) getsockname$packet(r2, &(0x7f0000000600)={0x11, 0x0, 0x0, 0x1, 0x0, 0x6, @broadcast}, &(0x7f0000000100)=0x14) sendmsg$nl_route(r1, &(0x7f0000000040)={0x0, 0x0, &(0x7f0000000000)={&(0x7f00000000c0)=ANY=[@ANYBLOB="3c00000010008506eb9afc4cd8d06e754a0081c5", @ANYRES32=r3, @ANYBLOB="2377f292252155b21c0012000c000100626f6e64000000000c0002000800010001"], 0x3c}}, 0x0) sendmsg$nl_route(r0, &(0x7f0000000240)={0x0, 0x0, &(0x7f0000000680)={&(0x7f0000000400)=@newlink={0x3c, 0x10, 0xffffff1f, 0x0, 0x1f672500, {}, [@IFLA_LINKINFO={0x14, 0x12, 0x0, 0x1, @bridge={{0xb}, {0x4}}}, @IFLA_MASTER={0x8, 0xa, r3}]}, 0x3c}}, 0x0) [ 2845.309426][ T3459] bond1299: (slave bridge1262): making interface the new active one [ 2845.373290][ T3459] bridge1262: entered promiscuous mode [ 2845.408764][ T3459] bond1299: (slave bridge1262): Enslaving as an active interface with an up link 04:12:53 executing program 3: r0 = socket$netlink(0x10, 0x3, 0x0) sendmsg$netlink(r0, &(0x7f00000047c0)={0x0, 0x0, &(0x7f0000000000)=[{&(0x7f0000000100)=ANY=[@ANYBLOB="18000000560001060000008a00db730007"], 0x18}], 0x1}, 0x0) 04:12:53 executing program 4: r0 = socket$inet6(0xa, 0x2, 0x0) ioctl$sock_SIOCETHTOOL(r0, 0x89f0, &(0x7f0000000000)={'bridge0\x00', &(0x7f0000000200)=ANY=[@ANYBLOB="10000000000000000000000000000100"/36]}) setsockopt$packet_tx_ring(0xffffffffffffffff, 0x107, 0xd, &(0x7f0000000040)=@req3={0x9, 0x80000000, 0x22b5992c, 0x21, 0x1, 0x10f, 0x7fffffff}, 0x1c) 04:12:53 executing program 0: r0 = socket$inet6(0xa, 0x6, 0x0) bind$inet6(r0, &(0x7f0000000000)={0xa, 0x4e20}, 0x1c) listen(r0, 0x6) r1 = socket$inet6(0xa, 0x6, 0x0) connect$inet6(r1, &(0x7f0000000200)={0xa, 0x0, 0x0, @initdev={0xfe, 0x88, '\x00', 0x0, 0x0}, 0x3}, 0x1c) connect$inet6(r1, &(0x7f0000000040)={0xa, 0x4e20, 0x0, @dev={0xfe, 0x80, '\x00', 0xf}}, 0x1c) sendmmsg$inet6(r1, &(0x7f0000005a00)=[{{0x0, 0x0, 0x0}}, {{0x0, 0x0, 0x0, 0x0, &(0x7f0000005bc0)=ANY=[@ANYBLOB="5000000000000000290000003600000000060000000000000401000730000000000a0000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000020000000000000002900000037000000000000000000000004010000000000001400000000000000290000000000000160000000000000002900000037"], 0x100, 0xff30}}], 0x2, 0x0) 04:12:53 executing program 5: r0 = socket$netlink(0x10, 0x3, 0x0) r1 = socket$netlink(0x10, 0x3, 0x0) r2 = socket(0x10, 0x803, 0x0) sendmsg$NL80211_CMD_CRIT_PROTOCOL_START(r2, &(0x7f0000000580)={0x0, 0x0, &(0x7f0000000540)={0x0, 0x1c}}, 0x0) getsockname$packet(r2, &(0x7f0000000600)={0x11, 0x0, 0x0, 0x1, 0x0, 0x6, @broadcast}, &(0x7f0000000100)=0x14) sendmsg$nl_route(r1, &(0x7f0000000040)={0x0, 0x0, &(0x7f0000000000)={&(0x7f00000000c0)=ANY=[@ANYBLOB="3c00000010008506eb9afc4cd8d06e754a0081c5", @ANYRES32=r3, @ANYBLOB="2377f292252155b21c0012000c000100626f6e64000000000c0002000800010001"], 0x3c}}, 0x0) sendmsg$nl_route(r0, &(0x7f0000000240)={0x0, 0x0, &(0x7f0000000680)={&(0x7f0000000400)=@newlink={0x3c, 0x10, 0xffffff1f, 0x0, 0x24000000, {}, [@IFLA_LINKINFO={0x14, 0x12, 0x0, 0x1, @bridge={{0xb}, {0x4}}}, @IFLA_MASTER={0x8, 0xa, r3}]}, 0x3c}}, 0x0) [ 2845.675348][ T3471] netlink: 'syz-executor.2': attribute type 1 has an invalid length. 04:12:53 executing program 0: r0 = socket$netlink(0x10, 0x3, 0x0) r1 = socket$netlink(0x10, 0x3, 0x0) r2 = socket(0x10, 0x803, 0x0) sendmsg$NL80211_CMD_CRIT_PROTOCOL_START(r2, &(0x7f0000000580)={0x0, 0x0, &(0x7f0000000540)={0x0, 0x1c}}, 0x0) getsockname$packet(r2, &(0x7f0000000600)={0x11, 0x0, 0x0, 0x1, 0x0, 0x6, @broadcast}, &(0x7f0000000100)=0x14) sendmsg$nl_route(r1, &(0x7f0000000040)={0x0, 0x0, &(0x7f0000000000)={&(0x7f00000000c0)=ANY=[@ANYBLOB="3c00000010008506eb9afc4cd8d06e754a0081c5", @ANYRES32=r3, @ANYBLOB="2377f292252155b21c0012000c000100626f6e64000000000c0002000800010001"], 0x3c}}, 0x0) sendmsg$nl_route(r0, &(0x7f0000000240)={0x0, 0x0, &(0x7f0000000680)={&(0x7f0000000400)=@newlink={0x3c, 0x10, 0xffffff1f, 0x0, 0x9a03, {}, [@IFLA_LINKINFO={0x14, 0x12, 0x0, 0x1, @bridge={{0xb}, {0x4}}}, @IFLA_MASTER={0x8, 0xa, r3}]}, 0x3c}}, 0x0) 04:12:53 executing program 2: r0 = socket$netlink(0x10, 0x3, 0x0) r1 = socket$netlink(0x10, 0x3, 0x0) r2 = socket(0x10, 0x803, 0x0) sendmsg$NL80211_CMD_CRIT_PROTOCOL_START(r2, &(0x7f0000000580)={0x0, 0x0, &(0x7f0000000540)={0x0, 0x1c}}, 0x0) getsockname$packet(r2, &(0x7f0000000600)={0x11, 0x0, 0x0, 0x1, 0x0, 0x6, @broadcast}, &(0x7f0000000100)=0x14) sendmsg$nl_route(r1, &(0x7f0000000040)={0x0, 0x0, &(0x7f0000000000)={&(0x7f00000000c0)=ANY=[@ANYBLOB="3c00000010008506eb9afc4cd8d06e754a0081c5", @ANYRES32=r3, @ANYBLOB="2377f292252155b21c0012000c000100626f6e64000000000c0002000800010001"], 0x3c}}, 0x0) sendmsg$nl_route(r0, &(0x7f0000000240)={0x0, 0x0, &(0x7f0000000680)={&(0x7f0000000400)=@newlink={0x3c, 0x10, 0xffffff1f, 0x0, 0x253c41, {}, [@IFLA_LINKINFO={0x14, 0x12, 0x0, 0x1, @bridge={{0xb}, {0x4}}}, @IFLA_MASTER={0x8, 0xa, r3}]}, 0x3c}}, 0x0) [ 2845.730461][ T3471] workqueue: Failed to create a rescuer kthread for wq "bond1292": -EINTR [ 2845.987635][ T3480] netlink: 'syz-executor.1': attribute type 1 has an invalid length. [ 2846.074796][ T3480] bond1308: entered promiscuous mode [ 2846.080397][ T3480] 8021q: adding VLAN 0 to HW filter on device bond1308 04:12:53 executing program 1: r0 = socket$netlink(0x10, 0x3, 0x0) r1 = socket$netlink(0x10, 0x3, 0x0) r2 = socket(0x10, 0x803, 0x0) sendmsg$NL80211_CMD_CRIT_PROTOCOL_START(r2, &(0x7f0000000580)={0x0, 0x0, &(0x7f0000000540)={0x0, 0x1c}}, 0x0) getsockname$packet(r2, &(0x7f0000000600)={0x11, 0x0, 0x0, 0x1, 0x0, 0x6, @broadcast}, &(0x7f0000000100)=0x14) sendmsg$nl_route(r1, &(0x7f0000000040)={0x0, 0x0, &(0x7f0000000000)={&(0x7f00000000c0)=ANY=[@ANYBLOB="3c00000010008506eb9afc4cd8d06e754a0081c5", @ANYRES32=r3, @ANYBLOB="2377f292252155b21c0012000c000100626f6e64000000000c0002000800010001"], 0x3c}}, 0x0) sendmsg$nl_route(r0, &(0x7f0000000240)={0x0, 0x0, &(0x7f0000000680)={&(0x7f0000000400)=@newlink={0x3c, 0x10, 0xffffff1f, 0x0, 0x20000000, {}, [@IFLA_LINKINFO={0x14, 0x12, 0x0, 0x1, @bridge={{0xb}, {0x4}}}, @IFLA_MASTER={0x8, 0xa, r3}]}, 0x3c}}, 0x0) [ 2846.290822][ T3483] bond1308: (slave bridge1254): making interface the new active one [ 2846.299319][ T3483] bridge1254: entered promiscuous mode [ 2846.314402][ T3483] bond1308: (slave bridge1254): Enslaving as an active interface with an up link [ 2846.387233][ T3487] netlink: 'syz-executor.5': attribute type 1 has an invalid length. [ 2846.501275][ T3487] bond1300: entered promiscuous mode [ 2846.511168][ T3487] 8021q: adding VLAN 0 to HW filter on device bond1300 04:12:54 executing program 4: r0 = socket$inet6(0xa, 0x2, 0x0) ioctl$sock_SIOCETHTOOL(r0, 0x89f0, &(0x7f0000000000)={'bridge0\x00', &(0x7f0000000200)=ANY=[@ANYBLOB="10000000000000000000000000000100"/36]}) (async) setsockopt$packet_tx_ring(0xffffffffffffffff, 0x107, 0xd, &(0x7f0000000040)=@req3={0x9, 0x80000000, 0x22b5992c, 0x21, 0x1, 0x10f, 0x7fffffff}, 0x1c) [ 2846.600772][ T3491] bond1300: (slave bridge1263): making interface the new active one [ 2846.615925][ T3491] bridge1263: entered promiscuous mode [ 2846.628522][ T3491] bond1300: (slave bridge1263): Enslaving as an active interface with an up link [ 2846.667462][ T3497] netlink: 'syz-executor.0': attribute type 1 has an invalid length. 04:12:54 executing program 3: r0 = socket$netlink(0x10, 0x3, 0x0) sendmsg$netlink(r0, &(0x7f00000047c0)={0x0, 0x0, &(0x7f0000000000)=[{&(0x7f0000000100)=ANY=[@ANYBLOB="18000000560001060000008b00db730007"], 0x18}], 0x1}, 0x0) 04:12:54 executing program 5: r0 = socket$netlink(0x10, 0x3, 0x0) r1 = socket$netlink(0x10, 0x3, 0x0) r2 = socket(0x10, 0x803, 0x0) sendmsg$NL80211_CMD_CRIT_PROTOCOL_START(r2, &(0x7f0000000580)={0x0, 0x0, &(0x7f0000000540)={0x0, 0x1c}}, 0x0) getsockname$packet(r2, &(0x7f0000000600)={0x11, 0x0, 0x0, 0x1, 0x0, 0x6, @broadcast}, &(0x7f0000000100)=0x14) sendmsg$nl_route(r1, &(0x7f0000000040)={0x0, 0x0, &(0x7f0000000000)={&(0x7f00000000c0)=ANY=[@ANYBLOB="3c00000010008506eb9afc4cd8d06e754a0081c5", @ANYRES32=r3, @ANYBLOB="2377f292252155b21c0012000c000100626f6e64000000000c0002000800010001"], 0x3c}}, 0x0) sendmsg$nl_route(r0, &(0x7f0000000240)={0x0, 0x0, &(0x7f0000000680)={&(0x7f0000000400)=@newlink={0x3c, 0x10, 0xffffff1f, 0x0, 0x32010000, {}, [@IFLA_LINKINFO={0x14, 0x12, 0x0, 0x1, @bridge={{0xb}, {0x4}}}, @IFLA_MASTER={0x8, 0xa, r3}]}, 0x3c}}, 0x0) [ 2846.811252][ T3497] bond12: entered promiscuous mode [ 2846.818044][ T3497] 8021q: adding VLAN 0 to HW filter on device bond12 [ 2846.915879][ T3498] bond12: (slave bridge20): making interface the new active one [ 2846.945734][ T3498] bridge20: entered promiscuous mode [ 2846.978732][ T3498] bond12: (slave bridge20): Enslaving as an active interface with an up link 04:12:54 executing program 0: r0 = socket$netlink(0x10, 0x3, 0x0) r1 = socket$netlink(0x10, 0x3, 0x0) r2 = socket(0x10, 0x803, 0x0) sendmsg$NL80211_CMD_CRIT_PROTOCOL_START(r2, &(0x7f0000000580)={0x0, 0x0, &(0x7f0000000540)={0x0, 0x1c}}, 0x0) getsockname$packet(r2, &(0x7f0000000600)={0x11, 0x0, 0x0, 0x1, 0x0, 0x6, @broadcast}, &(0x7f0000000100)=0x14) sendmsg$nl_route(r1, &(0x7f0000000040)={0x0, 0x0, &(0x7f0000000000)={&(0x7f00000000c0)=ANY=[@ANYBLOB="3c00000010008506eb9afc4cd8d06e754a0081c5", @ANYRES32=r3, @ANYBLOB="2377f292252155b21c0012000c000100626f6e64000000000c0002000800010001"], 0x3c}}, 0x0) sendmsg$nl_route(r0, &(0x7f0000000240)={0x0, 0x0, &(0x7f0000000680)={&(0x7f0000000400)=@newlink={0x3c, 0x10, 0xffffff1f, 0x0, 0x9a03, {}, [@IFLA_LINKINFO={0x14, 0x12, 0x0, 0x1, @bridge={{0xb}, {0x4}}}, @IFLA_MASTER={0x8, 0xa, r3}]}, 0x3c}}, 0x0) 04:12:54 executing program 2: r0 = socket$netlink(0x10, 0x3, 0x0) r1 = socket$netlink(0x10, 0x3, 0x0) r2 = socket(0x10, 0x803, 0x0) sendmsg$NL80211_CMD_CRIT_PROTOCOL_START(r2, &(0x7f0000000580)={0x0, 0x0, &(0x7f0000000540)={0x0, 0x1c}}, 0x0) getsockname$packet(r2, &(0x7f0000000600)={0x11, 0x0, 0x0, 0x1, 0x0, 0x6, @broadcast}, &(0x7f0000000100)=0x14) sendmsg$nl_route(r1, &(0x7f0000000040)={0x0, 0x0, &(0x7f0000000000)={&(0x7f00000000c0)=ANY=[@ANYBLOB="3c00000010008506eb9afc4cd8d06e754a0081c5", @ANYRES32=r3, @ANYBLOB="2377f292252155b21c0012000c000100626f6e64000000000c0002000800010001"], 0x3c}}, 0x0) sendmsg$nl_route(r0, &(0x7f0000000240)={0x0, 0x0, &(0x7f0000000680)={&(0x7f0000000400)=@newlink={0x3c, 0x10, 0xffffff1f, 0x0, 0x253c42, {}, [@IFLA_LINKINFO={0x14, 0x12, 0x0, 0x1, @bridge={{0xb}, {0x4}}}, @IFLA_MASTER={0x8, 0xa, r3}]}, 0x3c}}, 0x0) [ 2847.015106][ T3502] workqueue: Failed to create a rescuer kthread for wq "bond1292": -EINTR 04:12:55 executing program 4: r0 = socket$inet6(0xa, 0x2, 0x0) ioctl$sock_SIOCETHTOOL(r0, 0x89f0, &(0x7f0000000000)={'bridge0\x00', &(0x7f0000000200)=ANY=[@ANYBLOB="10000000000000000000000000000100"/36]}) (async) setsockopt$packet_tx_ring(0xffffffffffffffff, 0x107, 0xd, &(0x7f0000000040)=@req3={0x9, 0x80000000, 0x22b5992c, 0x21, 0x1, 0x10f, 0x7fffffff}, 0x1c) 04:12:55 executing program 3: r0 = socket$netlink(0x10, 0x3, 0x0) sendmsg$netlink(r0, &(0x7f00000047c0)={0x0, 0x0, &(0x7f0000000000)=[{&(0x7f0000000100)=ANY=[@ANYBLOB="18000000560001060000008c00db730007"], 0x18}], 0x1}, 0x0) 04:12:55 executing program 1: r0 = socket$netlink(0x10, 0x3, 0x0) r1 = socket$netlink(0x10, 0x3, 0x0) r2 = socket(0x10, 0x803, 0x0) sendmsg$NL80211_CMD_CRIT_PROTOCOL_START(r2, &(0x7f0000000580)={0x0, 0x0, &(0x7f0000000540)={0x0, 0x1c}}, 0x0) getsockname$packet(r2, &(0x7f0000000600)={0x11, 0x0, 0x0, 0x1, 0x0, 0x6, @broadcast}, &(0x7f0000000100)=0x14) sendmsg$nl_route(r1, &(0x7f0000000040)={0x0, 0x0, &(0x7f0000000000)={&(0x7f00000000c0)=ANY=[@ANYBLOB="3c00000010008506eb9afc4cd8d06e754a0081c5", @ANYRES32=r3, @ANYBLOB="2377f292252155b21c0012000c000100626f6e64000000000c0002000800010001"], 0x3c}}, 0x0) sendmsg$nl_route(r0, &(0x7f0000000240)={0x0, 0x0, &(0x7f0000000680)={&(0x7f0000000400)=@newlink={0x3c, 0x10, 0xffffff1f, 0x0, 0x20672500, {}, [@IFLA_LINKINFO={0x14, 0x12, 0x0, 0x1, @bridge={{0xb}, {0x4}}}, @IFLA_MASTER={0x8, 0xa, r3}]}, 0x3c}}, 0x0) [ 2847.498966][ T3505] workqueue: Failed to create a rescuer kthread for wq "bond1309": -EINTR [ 2847.859759][ T3517] workqueue: Failed to create a rescuer kthread for wq "bond1301": -EINTR 04:12:55 executing program 5: r0 = socket$netlink(0x10, 0x3, 0x0) r1 = socket$netlink(0x10, 0x3, 0x0) r2 = socket(0x10, 0x803, 0x0) sendmsg$NL80211_CMD_CRIT_PROTOCOL_START(r2, &(0x7f0000000580)={0x0, 0x0, &(0x7f0000000540)={0x0, 0x1c}}, 0x0) getsockname$packet(r2, &(0x7f0000000600)={0x11, 0x0, 0x0, 0x1, 0x0, 0x6, @broadcast}, &(0x7f0000000100)=0x14) sendmsg$nl_route(r1, &(0x7f0000000040)={0x0, 0x0, &(0x7f0000000000)={&(0x7f00000000c0)=ANY=[@ANYBLOB="3c00000010008506eb9afc4cd8d06e754a0081c5", @ANYRES32=r3, @ANYBLOB="2377f292252155b21c0012000c000100626f6e64000000000c0002000800010001"], 0x3c}}, 0x0) sendmsg$nl_route(r0, &(0x7f0000000240)={0x0, 0x0, &(0x7f0000000680)={&(0x7f0000000400)=@newlink={0x3c, 0x10, 0xffffff1f, 0x0, 0x3a020000, {}, [@IFLA_LINKINFO={0x14, 0x12, 0x0, 0x1, @bridge={{0xb}, {0x4}}}, @IFLA_MASTER={0x8, 0xa, r3}]}, 0x3c}}, 0x0) [ 2848.071068][ T3522] workqueue: Failed to create a rescuer kthread for wq "bond13": -EINTR [ 2848.259389][ T3526] validate_nla: 4 callbacks suppressed [ 2848.259410][ T3526] netlink: 'syz-executor.2': attribute type 1 has an invalid length. [ 2848.339286][ T3526] bond1292: entered promiscuous mode [ 2848.345108][ T3526] 8021q: adding VLAN 0 to HW filter on device bond1292 04:12:55 executing program 2: r0 = socket$netlink(0x10, 0x3, 0x0) r1 = socket$netlink(0x10, 0x3, 0x0) r2 = socket(0x10, 0x803, 0x0) sendmsg$NL80211_CMD_CRIT_PROTOCOL_START(r2, &(0x7f0000000580)={0x0, 0x0, &(0x7f0000000540)={0x0, 0x1c}}, 0x0) getsockname$packet(r2, &(0x7f0000000600)={0x11, 0x0, 0x0, 0x1, 0x0, 0x6, @broadcast}, &(0x7f0000000100)=0x14) sendmsg$nl_route(r1, &(0x7f0000000040)={0x0, 0x0, &(0x7f0000000000)={&(0x7f00000000c0)=ANY=[@ANYBLOB="3c00000010008506eb9afc4cd8d06e754a0081c5", @ANYRES32=r3, @ANYBLOB="2377f292252155b21c0012000c000100626f6e64000000000c0002000800010001"], 0x3c}}, 0x0) sendmsg$nl_route(r0, &(0x7f0000000240)={0x0, 0x0, &(0x7f0000000680)={&(0x7f0000000400)=@newlink={0x3c, 0x10, 0xffffff1f, 0x0, 0x253c43, {}, [@IFLA_LINKINFO={0x14, 0x12, 0x0, 0x1, @bridge={{0xb}, {0x4}}}, @IFLA_MASTER={0x8, 0xa, r3}]}, 0x3c}}, 0x0) 04:12:55 executing program 0: r0 = socket$netlink(0x10, 0x3, 0x0) r1 = socket$netlink(0x10, 0x3, 0x0) r2 = socket(0x10, 0x803, 0x0) sendmsg$NL80211_CMD_CRIT_PROTOCOL_START(r2, &(0x7f0000000580)={0x0, 0x0, &(0x7f0000000540)={0x0, 0x1c}}, 0x0) getsockname$packet(r2, &(0x7f0000000600)={0x11, 0x0, 0x0, 0x1, 0x0, 0x6, @broadcast}, &(0x7f0000000100)=0x14) sendmsg$nl_route(r1, &(0x7f0000000040)={0x0, 0x0, &(0x7f0000000000)={&(0x7f00000000c0)=ANY=[@ANYBLOB="3c00000010008506eb9afc4cd8d06e754a0081c5", @ANYRES32=r3, @ANYBLOB="2377f292252155b21c0012000c000100626f6e64000000000c0002000800010001"], 0x3c}}, 0x0) sendmsg$nl_route(r0, &(0x7f0000000240)={0x0, 0x0, &(0x7f0000000680)={&(0x7f0000000400)=@newlink={0x3c, 0x10, 0xffffff1f, 0x0, 0x9a03, {}, [@IFLA_LINKINFO={0x14, 0x12, 0x0, 0x1, @bridge={{0xb}, {0x4}}}, @IFLA_MASTER={0x8, 0xa, r3}]}, 0x3c}}, 0x0) [ 2848.384007][ T3527] bond1292: (slave bridge1213): making interface the new active one [ 2848.393979][ T3527] bridge1213: entered promiscuous mode [ 2848.406164][ T3527] bond1292: (slave bridge1213): Enslaving as an active interface with an up link 04:12:55 executing program 4: ioctl$sock_SIOCETHTOOL(0xffffffffffffffff, 0x89f0, &(0x7f0000000000)={'bridge0\x00', &(0x7f0000000200)=ANY=[@ANYBLOB="1000000000000000000000000000000000000000000000007bc7821b311318c713000000"]}) r0 = accept4(0xffffffffffffffff, &(0x7f0000000040)=@pppol2tpv3in6={0x18, 0x1, {0x0, 0xffffffffffffffff, 0x0, 0x0, 0x0, 0x0, {0xa, 0x0, 0x0, @dev}}}, &(0x7f00000000c0)=0x80, 0x80800) ioctl$SIOCX25GFACILITIES(r0, 0x89e2, &(0x7f0000000100)) r1 = socket$can_bcm(0x1d, 0x2, 0x2) read(r1, &(0x7f0000000140)=""/74, 0x4a) 04:12:56 executing program 3: r0 = socket$netlink(0x10, 0x3, 0x0) sendmsg$netlink(r0, &(0x7f00000047c0)={0x0, 0x0, &(0x7f0000000000)=[{&(0x7f0000000100)=ANY=[@ANYBLOB="18000000560001060000008d00db730007"], 0x18}], 0x1}, 0x0) [ 2848.655761][ T3536] netlink: 'syz-executor.1': attribute type 1 has an invalid length. [ 2848.755919][ T3536] bond1309: entered promiscuous mode [ 2848.768445][ T3536] 8021q: adding VLAN 0 to HW filter on device bond1309 04:12:56 executing program 1: r0 = socket$netlink(0x10, 0x3, 0x0) r1 = socket$netlink(0x10, 0x3, 0x0) r2 = socket(0x10, 0x803, 0x0) sendmsg$NL80211_CMD_CRIT_PROTOCOL_START(r2, &(0x7f0000000580)={0x0, 0x0, &(0x7f0000000540)={0x0, 0x1c}}, 0x0) getsockname$packet(r2, &(0x7f0000000600)={0x11, 0x0, 0x0, 0x1, 0x0, 0x6, @broadcast}, &(0x7f0000000100)=0x14) sendmsg$nl_route(r1, &(0x7f0000000040)={0x0, 0x0, &(0x7f0000000000)={&(0x7f00000000c0)=ANY=[@ANYBLOB="3c00000010008506eb9afc4cd8d06e754a0081c5", @ANYRES32=r3, @ANYBLOB="2377f292252155b21c0012000c000100626f6e64000000000c0002000800010001"], 0x3c}}, 0x0) sendmsg$nl_route(r0, &(0x7f0000000240)={0x0, 0x0, &(0x7f0000000680)={&(0x7f0000000400)=@newlink={0x3c, 0x10, 0xffffff1f, 0x0, 0x21672500, {}, [@IFLA_LINKINFO={0x14, 0x12, 0x0, 0x1, @bridge={{0xb}, {0x4}}}, @IFLA_MASTER={0x8, 0xa, r3}]}, 0x3c}}, 0x0) [ 2848.834662][ T3538] bond1309: (slave bridge1255): making interface the new active one [ 2848.844180][ T3538] bridge1255: entered promiscuous mode [ 2848.863060][ T3538] bond1309: (slave bridge1255): Enslaving as an active interface with an up link [ 2849.008240][ T3541] netlink: 'syz-executor.5': attribute type 1 has an invalid length. [ 2849.076552][ T3541] bond1301: entered promiscuous mode [ 2849.086327][ T3541] 8021q: adding VLAN 0 to HW filter on device bond1301 [ 2849.246441][ T3544] bond1301: (slave bridge1264): making interface the new active one [ 2849.269130][ T3544] bridge1264: entered promiscuous mode [ 2849.297767][ T3544] bond1301: (slave bridge1264): Enslaving as an active interface with an up link 04:12:56 executing program 5: r0 = socket$netlink(0x10, 0x3, 0x0) r1 = socket$netlink(0x10, 0x3, 0x0) r2 = socket(0x10, 0x803, 0x0) sendmsg$NL80211_CMD_CRIT_PROTOCOL_START(r2, &(0x7f0000000580)={0x0, 0x0, &(0x7f0000000540)={0x0, 0x1c}}, 0x0) getsockname$packet(r2, &(0x7f0000000600)={0x11, 0x0, 0x0, 0x1, 0x0, 0x6, @broadcast}, &(0x7f0000000100)=0x14) sendmsg$nl_route(r1, &(0x7f0000000040)={0x0, 0x0, &(0x7f0000000000)={&(0x7f00000000c0)=ANY=[@ANYBLOB="3c00000010008506eb9afc4cd8d06e754a0081c5", @ANYRES32=r3, @ANYBLOB="2377f292252155b21c0012000c000100626f6e64000000000c0002000800010001"], 0x3c}}, 0x0) sendmsg$nl_route(r0, &(0x7f0000000240)={0x0, 0x0, &(0x7f0000000680)={&(0x7f0000000400)=@newlink={0x3c, 0x10, 0xffffff1f, 0x0, 0x3c000000, {}, [@IFLA_LINKINFO={0x14, 0x12, 0x0, 0x1, @bridge={{0xb}, {0x4}}}, @IFLA_MASTER={0x8, 0xa, r3}]}, 0x3c}}, 0x0) 04:12:56 executing program 4: ioctl$sock_SIOCETHTOOL(0xffffffffffffffff, 0x89f0, &(0x7f0000000000)={'bridge0\x00', &(0x7f0000000200)=ANY=[@ANYBLOB="1000000000000000000000000000000000000000000000007bc7821b311318c713000000"]}) r0 = accept4(0xffffffffffffffff, &(0x7f0000000040)=@pppol2tpv3in6={0x18, 0x1, {0x0, 0xffffffffffffffff, 0x0, 0x0, 0x0, 0x0, {0xa, 0x0, 0x0, @dev}}}, &(0x7f00000000c0)=0x80, 0x80800) ioctl$SIOCX25GFACILITIES(r0, 0x89e2, &(0x7f0000000100)) (async) ioctl$SIOCX25GFACILITIES(r0, 0x89e2, &(0x7f0000000100)) r1 = socket$can_bcm(0x1d, 0x2, 0x2) read(r1, &(0x7f0000000140)=""/74, 0x4a) [ 2849.500033][ T3546] netlink: 'syz-executor.0': attribute type 1 has an invalid length. [ 2849.576976][ T3546] bond13: entered promiscuous mode [ 2849.583483][ T3546] 8021q: adding VLAN 0 to HW filter on device bond13 [ 2849.599659][ T3551] netlink: 'syz-executor.2': attribute type 1 has an invalid length. [ 2849.620878][ T3551] workqueue: Failed to create a rescuer kthread for wq "bond1293": -EINTR [ 2849.705245][ T3552] bond13: (slave bridge21): making interface the new active one [ 2849.728040][ T3552] bridge21: entered promiscuous mode [ 2849.757172][ T3552] bond13: (slave bridge21): Enslaving as an active interface with an up link 04:12:57 executing program 0: r0 = socket$netlink(0x10, 0x3, 0x0) r1 = socket$netlink(0x10, 0x3, 0x0) r2 = socket(0x10, 0x803, 0x0) sendmsg$NL80211_CMD_CRIT_PROTOCOL_START(r2, &(0x7f0000000580)={0x0, 0x0, &(0x7f0000000540)={0x0, 0x1c}}, 0x0) getsockname$packet(r2, &(0x7f0000000600)={0x11, 0x0, 0x0, 0x1, 0x0, 0x6, @broadcast}, &(0x7f0000000100)=0x14) sendmsg$nl_route(r1, &(0x7f0000000040)={0x0, 0x0, &(0x7f0000000000)={&(0x7f00000000c0)=ANY=[@ANYBLOB="3c00000010008506eb9afc4cd8d06e754a0081c5", @ANYRES32=r3, @ANYBLOB="2377f292252155b21c0012000c000100626f6e64000000000c0002000800010001"], 0x3c}}, 0x0) sendmsg$nl_route(r0, &(0x7f0000000240)={0x0, 0x0, &(0x7f0000000680)={&(0x7f0000000400)=@newlink={0x3c, 0x10, 0xffffff1f, 0x0, 0x8611, {}, [@IFLA_LINKINFO={0x14, 0x12, 0x0, 0x1, @bridge={{0xb}, {0x4}}}, @IFLA_MASTER={0x8, 0xa, r3}]}, 0x3c}}, 0x0) 04:12:57 executing program 3: r0 = socket$netlink(0x10, 0x3, 0x0) sendmsg$netlink(r0, &(0x7f00000047c0)={0x0, 0x0, &(0x7f0000000000)=[{&(0x7f0000000100)=ANY=[@ANYBLOB="18000000560001060000008e00db730007"], 0x18}], 0x1}, 0x0) 04:12:57 executing program 2: r0 = socket$netlink(0x10, 0x3, 0x0) r1 = socket$netlink(0x10, 0x3, 0x0) r2 = socket(0x10, 0x803, 0x0) sendmsg$NL80211_CMD_CRIT_PROTOCOL_START(r2, &(0x7f0000000580)={0x0, 0x0, &(0x7f0000000540)={0x0, 0x1c}}, 0x0) getsockname$packet(r2, &(0x7f0000000600)={0x11, 0x0, 0x0, 0x1, 0x0, 0x6, @broadcast}, &(0x7f0000000100)=0x14) sendmsg$nl_route(r1, &(0x7f0000000040)={0x0, 0x0, &(0x7f0000000000)={&(0x7f00000000c0)=ANY=[@ANYBLOB="3c00000010008506eb9afc4cd8d06e754a0081c5", @ANYRES32=r3, @ANYBLOB="2377f292252155b21c0012000c000100626f6e64000000000c0002000800010001"], 0x3c}}, 0x0) sendmsg$nl_route(r0, &(0x7f0000000240)={0x0, 0x0, &(0x7f0000000680)={&(0x7f0000000400)=@newlink={0x3c, 0x10, 0xffffff1f, 0x0, 0x253c44, {}, [@IFLA_LINKINFO={0x14, 0x12, 0x0, 0x1, @bridge={{0xb}, {0x4}}}, @IFLA_MASTER={0x8, 0xa, r3}]}, 0x3c}}, 0x0) [ 2850.028017][ T3559] netlink: 'syz-executor.1': attribute type 1 has an invalid length. [ 2850.091022][ T3559] workqueue: Failed to create a rescuer kthread for wq "bond1310": -EINTR [ 2850.159465][ T3577] netlink: 'syz-executor.2': attribute type 1 has an invalid length. [ 2850.252751][ T3577] bond1293: entered promiscuous mode [ 2850.258764][ T3577] 8021q: adding VLAN 0 to HW filter on device bond1293 04:12:57 executing program 1: r0 = socket$netlink(0x10, 0x3, 0x0) r1 = socket$netlink(0x10, 0x3, 0x0) r2 = socket(0x10, 0x803, 0x0) sendmsg$NL80211_CMD_CRIT_PROTOCOL_START(r2, &(0x7f0000000580)={0x0, 0x0, &(0x7f0000000540)={0x0, 0x1c}}, 0x0) getsockname$packet(r2, &(0x7f0000000600)={0x11, 0x0, 0x0, 0x1, 0x0, 0x6, @broadcast}, &(0x7f0000000100)=0x14) sendmsg$nl_route(r1, &(0x7f0000000040)={0x0, 0x0, &(0x7f0000000000)={&(0x7f00000000c0)=ANY=[@ANYBLOB="3c00000010008506eb9afc4cd8d06e754a0081c5", @ANYRES32=r3, @ANYBLOB="2377f292252155b21c0012000c000100626f6e64000000000c0002000800010001"], 0x3c}}, 0x0) sendmsg$nl_route(r0, &(0x7f0000000240)={0x0, 0x0, &(0x7f0000000680)={&(0x7f0000000400)=@newlink={0x3c, 0x10, 0xffffff1f, 0x0, 0x22672500, {}, [@IFLA_LINKINFO={0x14, 0x12, 0x0, 0x1, @bridge={{0xb}, {0x4}}}, @IFLA_MASTER={0x8, 0xa, r3}]}, 0x3c}}, 0x0) [ 2850.368081][ T3563] netlink: 'syz-executor.5': attribute type 1 has an invalid length. 04:12:57 executing program 4: ioctl$sock_SIOCETHTOOL(0xffffffffffffffff, 0x89f0, &(0x7f0000000000)={'bridge0\x00', &(0x7f0000000200)=ANY=[@ANYBLOB="1000000000000000000000000000000000000000000000007bc7821b311318c713000000"]}) r0 = accept4(0xffffffffffffffff, &(0x7f0000000040)=@pppol2tpv3in6={0x18, 0x1, {0x0, 0xffffffffffffffff, 0x0, 0x0, 0x0, 0x0, {0xa, 0x0, 0x0, @dev}}}, &(0x7f00000000c0)=0x80, 0x80800) ioctl$SIOCX25GFACILITIES(r0, 0x89e2, &(0x7f0000000100)) r1 = socket$can_bcm(0x1d, 0x2, 0x2) read(r1, &(0x7f0000000140)=""/74, 0x4a) ioctl$sock_SIOCETHTOOL(0xffffffffffffffff, 0x89f0, &(0x7f0000000000)={'bridge0\x00', &(0x7f0000000200)=ANY=[@ANYBLOB="1000000000000000000000000000000000000000000000007bc7821b311318c713000000"]}) (async) accept4(0xffffffffffffffff, &(0x7f0000000040)=@pppol2tpv3in6={0x18, 0x1, {0x0, 0xffffffffffffffff, 0x0, 0x0, 0x0, 0x0, {0xa, 0x0, 0x0, @dev}}}, &(0x7f00000000c0)=0x80, 0x80800) (async) ioctl$SIOCX25GFACILITIES(r0, 0x89e2, &(0x7f0000000100)) (async) socket$can_bcm(0x1d, 0x2, 0x2) (async) read(r1, &(0x7f0000000140)=""/74, 0x4a) (async) [ 2850.554713][ T3563] bond1302: entered promiscuous mode [ 2850.561408][ T3563] 8021q: adding VLAN 0 to HW filter on device bond1302 04:12:58 executing program 4: r0 = socket$inet6(0xa, 0x2, 0x0) r1 = socket$nl_netfilter(0x10, 0x3, 0xc) r2 = openat$cgroup_ro(0xffffffffffffff9c, &(0x7f0000000300)='cgroup.controllers\x00', 0x275a, 0x0) ioctl$FS_IOC_SETFLAGS(r2, 0x40086602, &(0x7f00000002c0)=0x2000000) close(r2) r3 = socket$inet_udp(0x2, 0x2, 0x0) write$binfmt_misc(r3, &(0x7f0000000040)={'syz1', "eab2321ffc785019f6ad25ad5c64b4fa036be4dfcc35fbb17bb0852869a3aeb05ae0bd5b77827adb14525c5bfa5a"}, 0x32) sendmsg$NFT_BATCH(r2, &(0x7f0000000140)={&(0x7f00000000c0)={0x10, 0x0, 0x0, 0x8000}, 0xc, &(0x7f0000000100)={&(0x7f0000000900)=ANY=[@ANYBLOB="140000001000010000000000000000000000000a40010000090a010100000000000000000300000a040112803c0001800c0001007061796c6f6164002c00028008000640000000000800074000000007080008400000000008000340fffffffd0800084000000000400001800e000100636f6e6e6c696d69740000002c00028008000140000000000800014000000020080002400000000008000240000000000800014000000007100001800a00010071756f7461000000100001800a0001006d61746368000000540001800b000100736f636b6574000044000280080003000000002d080002400000000b08000140000000000800014000000001080002400000001608000300000000c00800014000000001080003000000000010000180090001007866726d0000000008000f400000001808000c40000000070800034000000010080003400000005908000640ffffff0064000000180a0102000000000000000003000003080007400000000108000740000000010c00054000000000000000050c000380080002400000003d0c0005400000000000000004040003800c00054000000000000000040c000380080002400000000010010000050a01040000000000000000050000041c0008800c00014000000000000000020c000240000000000000007008000b4000000100cf000c00b559aa8f5980897e8c586506ac7dc9717475a7904a650fe1323f7b97c0628b18ad03f7010148be0211e53292d872813bc6c5a88fb1c7b7e0c4434b8bc6fc5f8d3cbb87f1eb4fd03cff16186ab4d538578c6a5e4c1b1f19ed714cdf45f8668f654861ba30ea6ef76c30af9bb55e6aa9caf69939e3cf9fce1c2193f3c6adbfc8cf4bdb9cfb4d7e1b405c1804de040db80939b5cfff1b26b5953b0d313a2feb9bfc4021c601032125865b7697fe77c319d65447226e38eab271916cef80a7e486009a2976dfea058a66ebf87700080007006e617400140000001100010000000000000000000000000a63ed8b5be0"], 0x2dc}, 0x1, 0x0, 0x0, 0x24000895}, 0x20000080) sendmsg$IPCTNL_MSG_CT_GET_CTRZERO(r1, &(0x7f0000000300)={0x0, 0x0, &(0x7f0000000080)={&(0x7f00000005c0)={0x24, 0x3, 0x1, 0x201, 0x0, 0x0, {0x2}, [@CTA_TUPLE_REPLY={0x4}, @CTA_FILTER={0xc, 0x19, 0x0, 0x1, [@CTA_FILTER_REPLY_FLAGS={0x8, 0x2, 0x718}]}]}, 0x24}}, 0x0) ioctl$sock_SIOCETHTOOL(r0, 0x89f0, &(0x7f0000000000)={'veth0_macvtap\x00', &(0x7f0000000340)=ANY=[@ANYBLOB="130000005000050007000000ff070000b2a5a6d8f40bfb057282796d074309305b029df3289934a0c3eece604b5c146b2d6bcfc7a8b37e6748d9876b93319331a2f97ab24371c7dbc9931a2faea4f03c980581d787bf86144cc2a38bc29aadda570486371e4f7020f2a8646eb5127e943048ab8714a61ff7a07f0f13eaabcf5ed5bccb079b68d29f32a982ed7534660da0d85d81009bb87c54e0fb8dd005c8061183b80f0193487eb999a7234dbb0c9a61c41f4fbd692fec19ec002963ddee2c804966a7944f5af4ad71dcb157d8afc3613f6f7de34f2665b1942cca9e87e8cff11c95c8cbda27c0742a717244875da0bd3ad44bb00727ceb2eaf0a1488a4d89d432ccef53332e9ac7ad781664b14095cc3c96a1b854c85964ee4db25ade4f05887fdd518205253b3de0374e4459d8d85c4ecc86e43bec0bd60f4a390b1a5daeb7576ba5d1b64a1337c2ae8bf278fe2578a119882ac83326f4f35f6b1067cbb4a27ef8ff782d544e058ef78603349aa7efee8c86507f0cb337e92bac39081c44b426af8f178198a1953e3e030a75a44f8d9b814bf924f680"]}) [ 2850.790597][ T3567] bond1302: (slave bridge1265): making interface the new active one [ 2850.811129][ T3567] bridge1265: entered promiscuous mode 04:12:58 executing program 5: r0 = socket$netlink(0x10, 0x3, 0x0) r1 = socket$netlink(0x10, 0x3, 0x0) r2 = socket(0x10, 0x803, 0x0) sendmsg$NL80211_CMD_CRIT_PROTOCOL_START(r2, &(0x7f0000000580)={0x0, 0x0, &(0x7f0000000540)={0x0, 0x1c}}, 0x0) getsockname$packet(r2, &(0x7f0000000600)={0x11, 0x0, 0x0, 0x1, 0x0, 0x6, @broadcast}, &(0x7f0000000100)=0x14) sendmsg$nl_route(r1, &(0x7f0000000040)={0x0, 0x0, &(0x7f0000000000)={&(0x7f00000000c0)=ANY=[@ANYBLOB="3c00000010008506eb9afc4cd8d06e754a0081c5", @ANYRES32=r3, @ANYBLOB="2377f292252155b21c0012000c000100626f6e64000000000c0002000800010001"], 0x3c}}, 0x0) sendmsg$nl_route(r0, &(0x7f0000000240)={0x0, 0x0, &(0x7f0000000680)={&(0x7f0000000400)=@newlink={0x3c, 0x10, 0xffffff1f, 0x0, 0x3f000000, {}, [@IFLA_LINKINFO={0x14, 0x12, 0x0, 0x1, @bridge={{0xb}, {0x4}}}, @IFLA_MASTER={0x8, 0xa, r3}]}, 0x3c}}, 0x0) [ 2850.838160][ T3567] bond1302: (slave bridge1265): Enslaving as an active interface with an up link [ 2850.864762][ T3571] netlink: 'syz-executor.0': attribute type 1 has an invalid length. [ 2850.938581][ T3571] bond14: entered promiscuous mode [ 2850.949726][ T3571] 8021q: adding VLAN 0 to HW filter on device bond14 04:12:58 executing program 0: r0 = socket$netlink(0x10, 0x3, 0x0) r1 = socket$netlink(0x10, 0x3, 0x0) r2 = socket(0x10, 0x803, 0x0) sendmsg$NL80211_CMD_CRIT_PROTOCOL_START(r2, &(0x7f0000000580)={0x0, 0x0, &(0x7f0000000540)={0x0, 0x1c}}, 0x0) getsockname$packet(r2, &(0x7f0000000600)={0x11, 0x0, 0x0, 0x1, 0x0, 0x6, @broadcast}, &(0x7f0000000100)=0x14) sendmsg$nl_route(r1, &(0x7f0000000040)={0x0, 0x0, &(0x7f0000000000)={&(0x7f00000000c0)=ANY=[@ANYBLOB="3c00000010008506eb9afc4cd8d06e754a0081c5", @ANYRES32=r3, @ANYBLOB="2377f292252155b21c0012000c000100626f6e64000000000c0002000800010001"], 0x3c}}, 0x0) sendmsg$nl_route(r0, &(0x7f0000000240)={0x0, 0x0, &(0x7f0000000680)={&(0x7f0000000400)=@newlink={0x3c, 0x10, 0xffffff1f, 0x0, 0x7a00, {}, [@IFLA_LINKINFO={0x14, 0x12, 0x0, 0x1, @bridge={{0xb}, {0x4}}}, @IFLA_MASTER={0x8, 0xa, r3}]}, 0x3c}}, 0x0) [ 2851.016588][ T3572] bond14: (slave bridge22): making interface the new active one [ 2851.026048][ T3572] bridge22: entered promiscuous mode [ 2851.035652][ T3572] bond14: (slave bridge22): Enslaving as an active interface with an up link 04:12:58 executing program 3: r0 = socket$netlink(0x10, 0x3, 0x0) sendmsg$netlink(r0, &(0x7f00000047c0)={0x0, 0x0, &(0x7f0000000000)=[{&(0x7f0000000100)=ANY=[@ANYBLOB="18000000560001060000008f00db730007"], 0x18}], 0x1}, 0x0) [ 2851.433955][ T3580] bond1293: (slave bridge1214): making interface the new active one [ 2851.444008][ T3580] bridge1214: entered promiscuous mode [ 2851.465298][ T3580] bond1293: (slave bridge1214): Enslaving as an active interface with an up link [ 2851.475135][ T3583] netlink: 'syz-executor.1': attribute type 1 has an invalid length. 04:12:58 executing program 2: r0 = socket$netlink(0x10, 0x3, 0x0) r1 = socket$netlink(0x10, 0x3, 0x0) r2 = socket(0x10, 0x803, 0x0) sendmsg$NL80211_CMD_CRIT_PROTOCOL_START(r2, &(0x7f0000000580)={0x0, 0x0, &(0x7f0000000540)={0x0, 0x1c}}, 0x0) getsockname$packet(r2, &(0x7f0000000600)={0x11, 0x0, 0x0, 0x1, 0x0, 0x6, @broadcast}, &(0x7f0000000100)=0x14) sendmsg$nl_route(r1, &(0x7f0000000040)={0x0, 0x0, &(0x7f0000000000)={&(0x7f00000000c0)=ANY=[@ANYBLOB="3c00000010008506eb9afc4cd8d06e754a0081c5", @ANYRES32=r3, @ANYBLOB="2377f292252155b21c0012000c000100626f6e64000000000c0002000800010001"], 0x3c}}, 0x0) sendmsg$nl_route(r0, &(0x7f0000000240)={0x0, 0x0, &(0x7f0000000680)={&(0x7f0000000400)=@newlink={0x3c, 0x10, 0xffffff1f, 0x0, 0x253c45, {}, [@IFLA_LINKINFO={0x14, 0x12, 0x0, 0x1, @bridge={{0xb}, {0x4}}}, @IFLA_MASTER={0x8, 0xa, r3}]}, 0x3c}}, 0x0) [ 2851.610276][ T3583] bond1310: entered promiscuous mode [ 2851.626753][ T3583] 8021q: adding VLAN 0 to HW filter on device bond1310 [ 2851.777927][ T3587] bond1310: (slave bridge1256): making interface the new active one [ 2851.788070][ T3587] bridge1256: entered promiscuous mode [ 2851.800306][ T3587] bond1310: (slave bridge1256): Enslaving as an active interface with an up link 04:12:59 executing program 4: r0 = socket$inet6(0xa, 0x2, 0x0) r1 = socket$nl_netfilter(0x10, 0x3, 0xc) r2 = openat$cgroup_ro(0xffffffffffffff9c, &(0x7f0000000300)='cgroup.controllers\x00', 0x275a, 0x0) ioctl$FS_IOC_SETFLAGS(r2, 0x40086602, &(0x7f00000002c0)=0x2000000) close(r2) r3 = socket$inet_udp(0x2, 0x2, 0x0) write$binfmt_misc(r3, &(0x7f0000000040)={'syz1', "eab2321ffc785019f6ad25ad5c64b4fa036be4dfcc35fbb17bb0852869a3aeb05ae0bd5b77827adb14525c5bfa5a"}, 0x32) sendmsg$NFT_BATCH(r2, &(0x7f0000000140)={&(0x7f00000000c0)={0x10, 0x0, 0x0, 0x8000}, 0xc, &(0x7f0000000100)={&(0x7f0000000900)=ANY=[@ANYBLOB="140000001000010000000000000000000000000a40010000090a010100000000000000000300000a040112803c0001800c0001007061796c6f6164002c00028008000640000000000800074000000007080008400000000008000340fffffffd0800084000000000400001800e000100636f6e6e6c696d69740000002c00028008000140000000000800014000000020080002400000000008000240000000000800014000000007100001800a00010071756f7461000000100001800a0001006d61746368000000540001800b000100736f636b6574000044000280080003000000002d080002400000000b08000140000000000800014000000001080002400000001608000300000000c00800014000000001080003000000000010000180090001007866726d0000000008000f400000001808000c40000000070800034000000010080003400000005908000640ffffff0064000000180a0102000000000000000003000003080007400000000108000740000000010c00054000000000000000050c000380080002400000003d0c0005400000000000000004040003800c00054000000000000000040c000380080002400000000010010000050a01040000000000000000050000041c0008800c00014000000000000000020c000240000000000000007008000b4000000100cf000c00b559aa8f5980897e8c586506ac7dc9717475a7904a650fe1323f7b97c0628b18ad03f7010148be0211e53292d872813bc6c5a88fb1c7b7e0c4434b8bc6fc5f8d3cbb87f1eb4fd03cff16186ab4d538578c6a5e4c1b1f19ed714cdf45f8668f654861ba30ea6ef76c30af9bb55e6aa9caf69939e3cf9fce1c2193f3c6adbfc8cf4bdb9cfb4d7e1b405c1804de040db80939b5cfff1b26b5953b0d313a2feb9bfc4021c601032125865b7697fe77c319d65447226e38eab271916cef80a7e486009a2976dfea058a66ebf87700080007006e617400140000001100010000000000000000000000000a63ed8b5be0"], 0x2dc}, 0x1, 0x0, 0x0, 0x24000895}, 0x20000080) sendmsg$IPCTNL_MSG_CT_GET_CTRZERO(r1, &(0x7f0000000300)={0x0, 0x0, &(0x7f0000000080)={&(0x7f00000005c0)={0x24, 0x3, 0x1, 0x201, 0x0, 0x0, {0x2}, [@CTA_TUPLE_REPLY={0x4}, @CTA_FILTER={0xc, 0x19, 0x0, 0x1, [@CTA_FILTER_REPLY_FLAGS={0x8, 0x2, 0x718}]}]}, 0x24}}, 0x0) ioctl$sock_SIOCETHTOOL(r0, 0x89f0, &(0x7f0000000000)={'veth0_macvtap\x00', &(0x7f0000000340)=ANY=[@ANYBLOB="130000005000050007000000ff070000b2a5a6d8f40bfb057282796d074309305b029df3289934a0c3eece604b5c146b2d6bcfc7a8b37e6748d9876b93319331a2f97ab24371c7dbc9931a2faea4f03c980581d787bf86144cc2a38bc29aadda570486371e4f7020f2a8646eb5127e943048ab8714a61ff7a07f0f13eaabcf5ed5bccb079b68d29f32a982ed7534660da0d85d81009bb87c54e0fb8dd005c8061183b80f0193487eb999a7234dbb0c9a61c41f4fbd692fec19ec002963ddee2c804966a7944f5af4ad71dcb157d8afc3613f6f7de34f2665b1942cca9e87e8cff11c95c8cbda27c0742a717244875da0bd3ad44bb00727ceb2eaf0a1488a4d89d432ccef53332e9ac7ad781664b14095cc3c96a1b854c85964ee4db25ade4f05887fdd518205253b3de0374e4459d8d85c4ecc86e43bec0bd60f4a390b1a5daeb7576ba5d1b64a1337c2ae8bf278fe2578a119882ac83326f4f35f6b1067cbb4a27ef8ff782d544e058ef78603349aa7efee8c86507f0cb337e92bac39081c44b426af8f178198a1953e3e030a75a44f8d9b814bf924f680"]}) socket$inet6(0xa, 0x2, 0x0) (async) socket$nl_netfilter(0x10, 0x3, 0xc) (async) openat$cgroup_ro(0xffffffffffffff9c, &(0x7f0000000300)='cgroup.controllers\x00', 0x275a, 0x0) (async) ioctl$FS_IOC_SETFLAGS(r2, 0x40086602, &(0x7f00000002c0)=0x2000000) (async) close(r2) (async) socket$inet_udp(0x2, 0x2, 0x0) (async) write$binfmt_misc(r3, &(0x7f0000000040)={'syz1', "eab2321ffc785019f6ad25ad5c64b4fa036be4dfcc35fbb17bb0852869a3aeb05ae0bd5b77827adb14525c5bfa5a"}, 0x32) (async) sendmsg$NFT_BATCH(r2, &(0x7f0000000140)={&(0x7f00000000c0)={0x10, 0x0, 0x0, 0x8000}, 0xc, &(0x7f0000000100)={&(0x7f0000000900)=ANY=[@ANYBLOB="140000001000010000000000000000000000000a40010000090a010100000000000000000300000a040112803c0001800c0001007061796c6f6164002c00028008000640000000000800074000000007080008400000000008000340fffffffd0800084000000000400001800e000100636f6e6e6c696d69740000002c00028008000140000000000800014000000020080002400000000008000240000000000800014000000007100001800a00010071756f7461000000100001800a0001006d61746368000000540001800b000100736f636b6574000044000280080003000000002d080002400000000b08000140000000000800014000000001080002400000001608000300000000c00800014000000001080003000000000010000180090001007866726d0000000008000f400000001808000c40000000070800034000000010080003400000005908000640ffffff0064000000180a0102000000000000000003000003080007400000000108000740000000010c00054000000000000000050c000380080002400000003d0c0005400000000000000004040003800c00054000000000000000040c000380080002400000000010010000050a01040000000000000000050000041c0008800c00014000000000000000020c000240000000000000007008000b4000000100cf000c00b559aa8f5980897e8c586506ac7dc9717475a7904a650fe1323f7b97c0628b18ad03f7010148be0211e53292d872813bc6c5a88fb1c7b7e0c4434b8bc6fc5f8d3cbb87f1eb4fd03cff16186ab4d538578c6a5e4c1b1f19ed714cdf45f8668f654861ba30ea6ef76c30af9bb55e6aa9caf69939e3cf9fce1c2193f3c6adbfc8cf4bdb9cfb4d7e1b405c1804de040db80939b5cfff1b26b5953b0d313a2feb9bfc4021c601032125865b7697fe77c319d65447226e38eab271916cef80a7e486009a2976dfea058a66ebf87700080007006e617400140000001100010000000000000000000000000a63ed8b5be0"], 0x2dc}, 0x1, 0x0, 0x0, 0x24000895}, 0x20000080) (async) sendmsg$IPCTNL_MSG_CT_GET_CTRZERO(r1, &(0x7f0000000300)={0x0, 0x0, &(0x7f0000000080)={&(0x7f00000005c0)={0x24, 0x3, 0x1, 0x201, 0x0, 0x0, {0x2}, [@CTA_TUPLE_REPLY={0x4}, @CTA_FILTER={0xc, 0x19, 0x0, 0x1, [@CTA_FILTER_REPLY_FLAGS={0x8, 0x2, 0x718}]}]}, 0x24}}, 0x0) (async) ioctl$sock_SIOCETHTOOL(r0, 0x89f0, &(0x7f0000000000)={'veth0_macvtap\x00', &(0x7f0000000340)=ANY=[@ANYBLOB="130000005000050007000000ff070000b2a5a6d8f40bfb057282796d074309305b029df3289934a0c3eece604b5c146b2d6bcfc7a8b37e6748d9876b93319331a2f97ab24371c7dbc9931a2faea4f03c980581d787bf86144cc2a38bc29aadda570486371e4f7020f2a8646eb5127e943048ab8714a61ff7a07f0f13eaabcf5ed5bccb079b68d29f32a982ed7534660da0d85d81009bb87c54e0fb8dd005c8061183b80f0193487eb999a7234dbb0c9a61c41f4fbd692fec19ec002963ddee2c804966a7944f5af4ad71dcb157d8afc3613f6f7de34f2665b1942cca9e87e8cff11c95c8cbda27c0742a717244875da0bd3ad44bb00727ceb2eaf0a1488a4d89d432ccef53332e9ac7ad781664b14095cc3c96a1b854c85964ee4db25ade4f05887fdd518205253b3de0374e4459d8d85c4ecc86e43bec0bd60f4a390b1a5daeb7576ba5d1b64a1337c2ae8bf278fe2578a119882ac83326f4f35f6b1067cbb4a27ef8ff782d544e058ef78603349aa7efee8c86507f0cb337e92bac39081c44b426af8f178198a1953e3e030a75a44f8d9b814bf924f680"]}) (async) 04:12:59 executing program 1: r0 = socket$netlink(0x10, 0x3, 0x0) r1 = socket$netlink(0x10, 0x3, 0x0) r2 = socket(0x10, 0x803, 0x0) sendmsg$NL80211_CMD_CRIT_PROTOCOL_START(r2, &(0x7f0000000580)={0x0, 0x0, &(0x7f0000000540)={0x0, 0x1c}}, 0x0) getsockname$packet(r2, &(0x7f0000000600)={0x11, 0x0, 0x0, 0x1, 0x0, 0x6, @broadcast}, &(0x7f0000000100)=0x14) sendmsg$nl_route(r1, &(0x7f0000000040)={0x0, 0x0, &(0x7f0000000000)={&(0x7f00000000c0)=ANY=[@ANYBLOB="3c00000010008506eb9afc4cd8d06e754a0081c5", @ANYRES32=r3, @ANYBLOB="2377f292252155b21c0012000c000100626f6e64000000000c0002000800010001"], 0x3c}}, 0x0) sendmsg$nl_route(r0, &(0x7f0000000240)={0x0, 0x0, &(0x7f0000000680)={&(0x7f0000000400)=@newlink={0x3c, 0x10, 0xffffff1f, 0x0, 0x23672500, {}, [@IFLA_LINKINFO={0x14, 0x12, 0x0, 0x1, @bridge={{0xb}, {0x4}}}, @IFLA_MASTER={0x8, 0xa, r3}]}, 0x3c}}, 0x0) [ 2851.981112][ T3598] workqueue: Failed to create a rescuer kthread for wq "bond15": -EINTR 04:12:59 executing program 0: r0 = socket$netlink(0x10, 0x3, 0x0) r1 = socket$netlink(0x10, 0x3, 0x0) r2 = socket(0x10, 0x803, 0x0) sendmsg$NL80211_CMD_CRIT_PROTOCOL_START(r2, &(0x7f0000000580)={0x0, 0x0, &(0x7f0000000540)={0x0, 0x1c}}, 0x0) getsockname$packet(r2, &(0x7f0000000600)={0x11, 0x0, 0x0, 0x1, 0x0, 0x6, @broadcast}, &(0x7f0000000100)=0x14) sendmsg$nl_route(r1, &(0x7f0000000040)={0x0, 0x0, &(0x7f0000000000)={&(0x7f00000000c0)=ANY=[@ANYBLOB="3c00000010008506eb9afc4cd8d06e754a0081c5", @ANYRES32=r3, @ANYBLOB="2377f292252155b21c0012000c000100626f6e64000000000c0002000800010001"], 0x3c}}, 0x0) sendmsg$nl_route(r0, &(0x7f0000000240)={0x0, 0x0, &(0x7f0000000680)={&(0x7f0000000400)=@newlink={0x3c, 0x10, 0xffffff1f, 0x0, 0x7a00, {}, [@IFLA_LINKINFO={0x14, 0x12, 0x0, 0x1, @bridge={{0xb}, {0x4}}}, @IFLA_MASTER={0x8, 0xa, r3}]}, 0x3c}}, 0x0) 04:12:59 executing program 3: r0 = socket$netlink(0x10, 0x3, 0x0) sendmsg$netlink(r0, &(0x7f00000047c0)={0x0, 0x0, &(0x7f0000000000)=[{&(0x7f0000000100)=ANY=[@ANYBLOB="18000000560001060000009000db730007"], 0x18}], 0x1}, 0x0) 04:12:59 executing program 5: r0 = socket$netlink(0x10, 0x3, 0x0) r1 = socket$netlink(0x10, 0x3, 0x0) r2 = socket(0x10, 0x803, 0x0) sendmsg$NL80211_CMD_CRIT_PROTOCOL_START(r2, &(0x7f0000000580)={0x0, 0x0, &(0x7f0000000540)={0x0, 0x1c}}, 0x0) getsockname$packet(r2, &(0x7f0000000600)={0x11, 0x0, 0x0, 0x1, 0x0, 0x6, @broadcast}, &(0x7f0000000100)=0x14) sendmsg$nl_route(r1, &(0x7f0000000040)={0x0, 0x0, &(0x7f0000000000)={&(0x7f00000000c0)=ANY=[@ANYBLOB="3c00000010008506eb9afc4cd8d06e754a0081c5", @ANYRES32=r3, @ANYBLOB="2377f292252155b21c0012000c000100626f6e64000000000c0002000800010001"], 0x3c}}, 0x0) sendmsg$nl_route(r0, &(0x7f0000000240)={0x0, 0x0, &(0x7f0000000680)={&(0x7f0000000400)=@newlink={0x3c, 0x10, 0xffffff1f, 0x0, 0x40000000, {}, [@IFLA_LINKINFO={0x14, 0x12, 0x0, 0x1, @bridge={{0xb}, {0x4}}}, @IFLA_MASTER={0x8, 0xa, r3}]}, 0x3c}}, 0x0) [ 2852.089903][ T3599] workqueue: Failed to create a rescuer kthread for wq "bond1303": -EINTR [ 2852.486188][ T3606] bond1294: entered promiscuous mode [ 2852.507051][ T3606] 8021q: adding VLAN 0 to HW filter on device bond1294 [ 2852.624502][ T3607] bond1294: (slave bridge1215): making interface the new active one [ 2852.633159][ T3607] bridge1215: entered promiscuous mode [ 2852.656436][ T3607] bond1294: (slave bridge1215): Enslaving as an active interface with an up link 04:13:00 executing program 2: r0 = socket$netlink(0x10, 0x3, 0x0) r1 = socket$netlink(0x10, 0x3, 0x0) r2 = socket(0x10, 0x803, 0x0) sendmsg$NL80211_CMD_CRIT_PROTOCOL_START(r2, &(0x7f0000000580)={0x0, 0x0, &(0x7f0000000540)={0x0, 0x1c}}, 0x0) getsockname$packet(r2, &(0x7f0000000600)={0x11, 0x0, 0x0, 0x1, 0x0, 0x6, @broadcast}, &(0x7f0000000100)=0x14) sendmsg$nl_route(r1, &(0x7f0000000040)={0x0, 0x0, &(0x7f0000000000)={&(0x7f00000000c0)=ANY=[@ANYBLOB="3c00000010008506eb9afc4cd8d06e754a0081c5", @ANYRES32=r3, @ANYBLOB="2377f292252155b21c0012000c000100626f6e64000000000c0002000800010001"], 0x3c}}, 0x0) sendmsg$nl_route(r0, &(0x7f0000000240)={0x0, 0x0, &(0x7f0000000680)={&(0x7f0000000400)=@newlink={0x3c, 0x10, 0xffffff1f, 0x0, 0x253c46, {}, [@IFLA_LINKINFO={0x14, 0x12, 0x0, 0x1, @bridge={{0xb}, {0x4}}}, @IFLA_MASTER={0x8, 0xa, r3}]}, 0x3c}}, 0x0) 04:13:00 executing program 4: r0 = socket$inet6(0xa, 0x2, 0x0) r1 = socket$nl_netfilter(0x10, 0x3, 0xc) r2 = openat$cgroup_ro(0xffffffffffffff9c, &(0x7f0000000300)='cgroup.controllers\x00', 0x275a, 0x0) ioctl$FS_IOC_SETFLAGS(r2, 0x40086602, &(0x7f00000002c0)=0x2000000) close(r2) r3 = socket$inet_udp(0x2, 0x2, 0x0) write$binfmt_misc(r3, &(0x7f0000000040)={'syz1', "eab2321ffc785019f6ad25ad5c64b4fa036be4dfcc35fbb17bb0852869a3aeb05ae0bd5b77827adb14525c5bfa5a"}, 0x32) (async) write$binfmt_misc(r3, &(0x7f0000000040)={'syz1', "eab2321ffc785019f6ad25ad5c64b4fa036be4dfcc35fbb17bb0852869a3aeb05ae0bd5b77827adb14525c5bfa5a"}, 0x32) sendmsg$NFT_BATCH(r2, &(0x7f0000000140)={&(0x7f00000000c0)={0x10, 0x0, 0x0, 0x8000}, 0xc, &(0x7f0000000100)={&(0x7f0000000900)=ANY=[@ANYBLOB="140000001000010000000000000000000000000a40010000090a010100000000000000000300000a040112803c0001800c0001007061796c6f6164002c00028008000640000000000800074000000007080008400000000008000340fffffffd0800084000000000400001800e000100636f6e6e6c696d69740000002c00028008000140000000000800014000000020080002400000000008000240000000000800014000000007100001800a00010071756f7461000000100001800a0001006d61746368000000540001800b000100736f636b6574000044000280080003000000002d080002400000000b08000140000000000800014000000001080002400000001608000300000000c00800014000000001080003000000000010000180090001007866726d0000000008000f400000001808000c40000000070800034000000010080003400000005908000640ffffff0064000000180a0102000000000000000003000003080007400000000108000740000000010c00054000000000000000050c000380080002400000003d0c0005400000000000000004040003800c00054000000000000000040c000380080002400000000010010000050a01040000000000000000050000041c0008800c00014000000000000000020c000240000000000000007008000b4000000100cf000c00b559aa8f5980897e8c586506ac7dc9717475a7904a650fe1323f7b97c0628b18ad03f7010148be0211e53292d872813bc6c5a88fb1c7b7e0c4434b8bc6fc5f8d3cbb87f1eb4fd03cff16186ab4d538578c6a5e4c1b1f19ed714cdf45f8668f654861ba30ea6ef76c30af9bb55e6aa9caf69939e3cf9fce1c2193f3c6adbfc8cf4bdb9cfb4d7e1b405c1804de040db80939b5cfff1b26b5953b0d313a2feb9bfc4021c601032125865b7697fe77c319d65447226e38eab271916cef80a7e486009a2976dfea058a66ebf87700080007006e617400140000001100010000000000000000000000000a63ed8b5be0"], 0x2dc}, 0x1, 0x0, 0x0, 0x24000895}, 0x20000080) (async) sendmsg$NFT_BATCH(r2, &(0x7f0000000140)={&(0x7f00000000c0)={0x10, 0x0, 0x0, 0x8000}, 0xc, &(0x7f0000000100)={&(0x7f0000000900)=ANY=[@ANYBLOB="140000001000010000000000000000000000000a40010000090a010100000000000000000300000a040112803c0001800c0001007061796c6f6164002c00028008000640000000000800074000000007080008400000000008000340fffffffd0800084000000000400001800e000100636f6e6e6c696d69740000002c00028008000140000000000800014000000020080002400000000008000240000000000800014000000007100001800a00010071756f7461000000100001800a0001006d61746368000000540001800b000100736f636b6574000044000280080003000000002d080002400000000b08000140000000000800014000000001080002400000001608000300000000c00800014000000001080003000000000010000180090001007866726d0000000008000f400000001808000c40000000070800034000000010080003400000005908000640ffffff0064000000180a0102000000000000000003000003080007400000000108000740000000010c00054000000000000000050c000380080002400000003d0c0005400000000000000004040003800c00054000000000000000040c000380080002400000000010010000050a01040000000000000000050000041c0008800c00014000000000000000020c000240000000000000007008000b4000000100cf000c00b559aa8f5980897e8c586506ac7dc9717475a7904a650fe1323f7b97c0628b18ad03f7010148be0211e53292d872813bc6c5a88fb1c7b7e0c4434b8bc6fc5f8d3cbb87f1eb4fd03cff16186ab4d538578c6a5e4c1b1f19ed714cdf45f8668f654861ba30ea6ef76c30af9bb55e6aa9caf69939e3cf9fce1c2193f3c6adbfc8cf4bdb9cfb4d7e1b405c1804de040db80939b5cfff1b26b5953b0d313a2feb9bfc4021c601032125865b7697fe77c319d65447226e38eab271916cef80a7e486009a2976dfea058a66ebf87700080007006e617400140000001100010000000000000000000000000a63ed8b5be0"], 0x2dc}, 0x1, 0x0, 0x0, 0x24000895}, 0x20000080) sendmsg$IPCTNL_MSG_CT_GET_CTRZERO(r1, &(0x7f0000000300)={0x0, 0x0, &(0x7f0000000080)={&(0x7f00000005c0)={0x24, 0x3, 0x1, 0x201, 0x0, 0x0, {0x2}, [@CTA_TUPLE_REPLY={0x4}, @CTA_FILTER={0xc, 0x19, 0x0, 0x1, [@CTA_FILTER_REPLY_FLAGS={0x8, 0x2, 0x718}]}]}, 0x24}}, 0x0) (async) sendmsg$IPCTNL_MSG_CT_GET_CTRZERO(r1, &(0x7f0000000300)={0x0, 0x0, &(0x7f0000000080)={&(0x7f00000005c0)={0x24, 0x3, 0x1, 0x201, 0x0, 0x0, {0x2}, [@CTA_TUPLE_REPLY={0x4}, @CTA_FILTER={0xc, 0x19, 0x0, 0x1, [@CTA_FILTER_REPLY_FLAGS={0x8, 0x2, 0x718}]}]}, 0x24}}, 0x0) ioctl$sock_SIOCETHTOOL(r0, 0x89f0, &(0x7f0000000000)={'veth0_macvtap\x00', &(0x7f0000000340)=ANY=[@ANYBLOB="130000005000050007000000ff070000b2a5a6d8f40bfb057282796d074309305b029df3289934a0c3eece604b5c146b2d6bcfc7a8b37e6748d9876b93319331a2f97ab24371c7dbc9931a2faea4f03c980581d787bf86144cc2a38bc29aadda570486371e4f7020f2a8646eb5127e943048ab8714a61ff7a07f0f13eaabcf5ed5bccb079b68d29f32a982ed7534660da0d85d81009bb87c54e0fb8dd005c8061183b80f0193487eb999a7234dbb0c9a61c41f4fbd692fec19ec002963ddee2c804966a7944f5af4ad71dcb157d8afc3613f6f7de34f2665b1942cca9e87e8cff11c95c8cbda27c0742a717244875da0bd3ad44bb00727ceb2eaf0a1488a4d89d432ccef53332e9ac7ad781664b14095cc3c96a1b854c85964ee4db25ade4f05887fdd518205253b3de0374e4459d8d85c4ecc86e43bec0bd60f4a390b1a5daeb7576ba5d1b64a1337c2ae8bf278fe2578a119882ac83326f4f35f6b1067cbb4a27ef8ff782d544e058ef78603349aa7efee8c86507f0cb337e92bac39081c44b426af8f178198a1953e3e030a75a44f8d9b814bf924f680"]}) [ 2852.834962][ T3612] bond1311: entered promiscuous mode [ 2852.845575][ T3612] 8021q: adding VLAN 0 to HW filter on device bond1311 [ 2852.987242][ T3620] bond1311: (slave bridge1257): making interface the new active one [ 2853.010364][ T3620] bridge1257: entered promiscuous mode 04:13:00 executing program 1: r0 = socket$netlink(0x10, 0x3, 0x0) r1 = socket$netlink(0x10, 0x3, 0x0) r2 = socket(0x10, 0x803, 0x0) sendmsg$NL80211_CMD_CRIT_PROTOCOL_START(r2, &(0x7f0000000580)={0x0, 0x0, &(0x7f0000000540)={0x0, 0x1c}}, 0x0) getsockname$packet(r2, &(0x7f0000000600)={0x11, 0x0, 0x0, 0x1, 0x0, 0x6, @broadcast}, &(0x7f0000000100)=0x14) sendmsg$nl_route(r1, &(0x7f0000000040)={0x0, 0x0, &(0x7f0000000000)={&(0x7f00000000c0)=ANY=[@ANYBLOB="3c00000010008506eb9afc4cd8d06e754a0081c5", @ANYRES32=r3, @ANYBLOB="2377f292252155b21c0012000c000100626f6e64000000000c0002000800010001"], 0x3c}}, 0x0) sendmsg$nl_route(r0, &(0x7f0000000240)={0x0, 0x0, &(0x7f0000000680)={&(0x7f0000000400)=@newlink={0x3c, 0x10, 0xffffff1f, 0x0, 0x24000000, {}, [@IFLA_LINKINFO={0x14, 0x12, 0x0, 0x1, @bridge={{0xb}, {0x4}}}, @IFLA_MASTER={0x8, 0xa, r3}]}, 0x3c}}, 0x0) [ 2853.026627][ T3620] bond1311: (slave bridge1257): Enslaving as an active interface with an up link [ 2853.130599][ T3625] bond15: entered promiscuous mode [ 2853.149599][ T3625] 8021q: adding VLAN 0 to HW filter on device bond15 04:13:00 executing program 3: r0 = socket$netlink(0x10, 0x3, 0x0) sendmsg$netlink(r0, &(0x7f00000047c0)={0x0, 0x0, &(0x7f0000000000)=[{&(0x7f0000000100)=ANY=[@ANYBLOB="18000000560001060000009100db730007"], 0x18}], 0x1}, 0x0) [ 2853.347012][ T3628] bond15: (slave bridge23): making interface the new active one [ 2853.370366][ T3628] bridge23: entered promiscuous mode [ 2853.381306][ T3628] bond15: (slave bridge23): Enslaving as an active interface with an up link [ 2853.413945][ T3630] validate_nla: 5 callbacks suppressed [ 2853.413964][ T3630] netlink: 'syz-executor.5': attribute type 1 has an invalid length. 04:13:00 executing program 0: r0 = socket$inet6(0xa, 0x6, 0x0) bind$inet6(r0, &(0x7f0000000000)={0xa, 0x4e20}, 0x1c) listen(r0, 0x6) r1 = socket$inet6(0xa, 0x6, 0x0) connect$inet6(r1, &(0x7f0000000200)={0xa, 0x0, 0x0, @initdev={0xfe, 0x88, '\x00', 0x0, 0x0}, 0x3}, 0x1c) connect$inet6(r1, &(0x7f0000000040)={0xa, 0x4e20, 0x0, @dev={0xfe, 0x80, '\x00', 0xf}}, 0x1c) sendmmsg$inet6(r1, &(0x7f0000005a00)=[{{0x0, 0x0, 0x0}}, {{0x0, 0x0, 0x0, 0x0, &(0x7f0000005bc0)=ANY=[@ANYBLOB="5000000000000000290000003600000000060000000000000401000730000000000a0000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000020000000000000002900000037000000000000000000000004010000000000001400000000000000290000000000000160000000000000002900000037"], 0x100, 0x3}}], 0x2, 0x0) [ 2853.498173][ T3630] workqueue: Failed to create a rescuer kthread for wq "bond1303": -EINTR [ 2853.697653][ T3639] netlink: 'syz-executor.2': attribute type 1 has an invalid length. 04:13:01 executing program 4: r0 = socket$inet6(0xa, 0x801, 0x800) r1 = socket$inet6_icmp_raw(0xa, 0x3, 0x3a) accept4$inet6(r1, 0x0, &(0x7f0000000040), 0x0) setsockopt$IP6T_SO_SET_REPLACE(r1, 0x29, 0x40, &(0x7f0000000240)=@raw={'raw\x00', 0x9, 0x3, 0x2a0, 0x0, 0xffffffff, 0xffffffff, 0xd8, 0xffffffff, 0x1d0, 0xffffffff, 0xffffffff, 0x1d0, 0xffffffff, 0x3, &(0x7f0000000080), {[{{@uncond, 0x0, 0xa8, 0xd8}, @common=@inet=@SET2={0x30, 'SET\x00', 0x2, {{0x3, 0x0, 0x2}, {0x3, 0x6, 0x2}, 0x8, 0x40}}}, {{@uncond, 0x0, 0xd8, 0xf8, 0x0, {}, [@common=@srh={{0x30}, {0x2c, 0x7, 0x7d, 0x0, 0xffff, 0x820, 0x400}}]}, @unspec=@NOTRACK={0x20}}], {{'\x00', 0x0, 0xa8, 0xd0}, {0x28}}}}, 0x300) ioctl$sock_SIOCETHTOOL(r0, 0x89f0, &(0x7f0000000000)={'bridge0\x00', &(0x7f0000000200)=@ethtool_ringparam={0x10}}) 04:13:01 executing program 0: r0 = socket$inet6(0xa, 0x2, 0x0) ioctl$sock_inet_SIOCSIFPFLAGS(0xffffffffffffffff, 0x8934, &(0x7f0000000040)={'wlan0\x00', 0x100000}) r1 = syz_init_net_socket$bt_hci(0x1f, 0x3, 0x5) ioctl(r1, 0x8b0f, &(0x7f0000000040)) ioctl$sock_SIOCETHTOOL(r0, 0x89f0, &(0x7f0000000000)={'bridge0\x00', &(0x7f0000000200)=@ethtool_ringparam={0x10}}) bind$inet6(r0, &(0x7f0000000040)={0xa, 0x4e24, 0x1, @dev={0xfe, 0x80, '\x00', 0x22}, 0x6}, 0x1c) r2 = openat$cgroup_ro(0xffffffffffffff9c, &(0x7f0000000300)='cgroup.controllers\x00', 0x275a, 0x0) ioctl$FS_IOC_SETFLAGS(r2, 0x40086602, &(0x7f00000002c0)=0x2000000) close(r2) sendmsg$nl_route_sched(r2, &(0x7f0000000180)={&(0x7f0000000080)={0x10, 0x0, 0x0, 0x40}, 0xc, &(0x7f0000000140)={&(0x7f00000000c0)=@newqdisc={0x6c, 0x24, 0x800, 0x70bd29, 0x25dfdbff, {0x0, 0x0, 0x0, 0x0, {0xfff1, 0xffe0}, {0x2}, {0x4, 0xffff}}, [@qdisc_kind_options=@q_plug={{0x9}, {0xc, 0x2, {0x0, 0xa1}}}, @TCA_RATE={0x6, 0x5, {0x6, 0x7}}, @qdisc_kind_options=@q_plug={{0x9}, {0xc, 0x2, {0x3, 0x7fffffff}}}, @TCA_RATE={0x6, 0x5, {0x2, 0x4}}, @TCA_EGRESS_BLOCK={0x8}]}, 0x6c}, 0x1, 0x0, 0x0, 0x20000050}, 0x41) 04:13:01 executing program 5: r0 = socket$netlink(0x10, 0x3, 0x0) r1 = socket$netlink(0x10, 0x3, 0x0) r2 = socket(0x10, 0x803, 0x0) sendmsg$NL80211_CMD_CRIT_PROTOCOL_START(r2, &(0x7f0000000580)={0x0, 0x0, &(0x7f0000000540)={0x0, 0x1c}}, 0x0) getsockname$packet(r2, &(0x7f0000000600)={0x11, 0x0, 0x0, 0x1, 0x0, 0x6, @broadcast}, &(0x7f0000000100)=0x14) sendmsg$nl_route(r1, &(0x7f0000000040)={0x0, 0x0, &(0x7f0000000000)={&(0x7f00000000c0)=ANY=[@ANYBLOB="3c00000010008506eb9afc4cd8d06e754a0081c5", @ANYRES32=r3, @ANYBLOB="2377f292252155b21c0012000c000100626f6e64000000000c0002000800010001"], 0x3c}}, 0x0) sendmsg$nl_route(r0, &(0x7f0000000240)={0x0, 0x0, &(0x7f0000000680)={&(0x7f0000000400)=@newlink={0x3c, 0x10, 0xffffff1f, 0x0, 0x48000000, {}, [@IFLA_LINKINFO={0x14, 0x12, 0x0, 0x1, @bridge={{0xb}, {0x4}}}, @IFLA_MASTER={0x8, 0xa, r3}]}, 0x3c}}, 0x0) [ 2853.828038][ T3639] workqueue: Failed to create a rescuer kthread for wq "bond1295": -EINTR [ 2853.928801][ T3660] x_tables: duplicate underflow at hook 3 04:13:01 executing program 4: r0 = socket$inet6(0xa, 0x801, 0x800) socket$inet6_icmp_raw(0xa, 0x3, 0x3a) (async) r1 = socket$inet6_icmp_raw(0xa, 0x3, 0x3a) accept4$inet6(r1, 0x0, &(0x7f0000000040), 0x0) setsockopt$IP6T_SO_SET_REPLACE(r1, 0x29, 0x40, &(0x7f0000000240)=@raw={'raw\x00', 0x9, 0x3, 0x2a0, 0x0, 0xffffffff, 0xffffffff, 0xd8, 0xffffffff, 0x1d0, 0xffffffff, 0xffffffff, 0x1d0, 0xffffffff, 0x3, &(0x7f0000000080), {[{{@uncond, 0x0, 0xa8, 0xd8}, @common=@inet=@SET2={0x30, 'SET\x00', 0x2, {{0x3, 0x0, 0x2}, {0x3, 0x6, 0x2}, 0x8, 0x40}}}, {{@uncond, 0x0, 0xd8, 0xf8, 0x0, {}, [@common=@srh={{0x30}, {0x2c, 0x7, 0x7d, 0x0, 0xffff, 0x820, 0x400}}]}, @unspec=@NOTRACK={0x20}}], {{'\x00', 0x0, 0xa8, 0xd0}, {0x28}}}}, 0x300) ioctl$sock_SIOCETHTOOL(r0, 0x89f0, &(0x7f0000000000)={'bridge0\x00', &(0x7f0000000200)=@ethtool_ringparam={0x10}}) 04:13:01 executing program 4: r0 = socket$inet6(0xa, 0x801, 0x800) (async) r1 = socket$inet6_icmp_raw(0xa, 0x3, 0x3a) accept4$inet6(r1, 0x0, &(0x7f0000000040), 0x0) (async) setsockopt$IP6T_SO_SET_REPLACE(r1, 0x29, 0x40, &(0x7f0000000240)=@raw={'raw\x00', 0x9, 0x3, 0x2a0, 0x0, 0xffffffff, 0xffffffff, 0xd8, 0xffffffff, 0x1d0, 0xffffffff, 0xffffffff, 0x1d0, 0xffffffff, 0x3, &(0x7f0000000080), {[{{@uncond, 0x0, 0xa8, 0xd8}, @common=@inet=@SET2={0x30, 'SET\x00', 0x2, {{0x3, 0x0, 0x2}, {0x3, 0x6, 0x2}, 0x8, 0x40}}}, {{@uncond, 0x0, 0xd8, 0xf8, 0x0, {}, [@common=@srh={{0x30}, {0x2c, 0x7, 0x7d, 0x0, 0xffff, 0x820, 0x400}}]}, @unspec=@NOTRACK={0x20}}], {{'\x00', 0x0, 0xa8, 0xd0}, {0x28}}}}, 0x300) ioctl$sock_SIOCETHTOOL(r0, 0x89f0, &(0x7f0000000000)={'bridge0\x00', &(0x7f0000000200)=@ethtool_ringparam={0x10}}) [ 2854.097442][ T3665] x_tables: duplicate underflow at hook 3 [ 2854.107080][ T3644] netlink: 'syz-executor.1': attribute type 1 has an invalid length. 04:13:01 executing program 2: r0 = socket$netlink(0x10, 0x3, 0x0) r1 = socket$netlink(0x10, 0x3, 0x0) r2 = socket(0x10, 0x803, 0x0) sendmsg$NL80211_CMD_CRIT_PROTOCOL_START(r2, &(0x7f0000000580)={0x0, 0x0, &(0x7f0000000540)={0x0, 0x1c}}, 0x0) getsockname$packet(r2, &(0x7f0000000600)={0x11, 0x0, 0x0, 0x1, 0x0, 0x6, @broadcast}, &(0x7f0000000100)=0x14) sendmsg$nl_route(r1, &(0x7f0000000040)={0x0, 0x0, &(0x7f0000000000)={&(0x7f00000000c0)=ANY=[@ANYBLOB="3c00000010008506eb9afc4cd8d06e754a0081c5", @ANYRES32=r3, @ANYBLOB="2377f292252155b21c0012000c000100626f6e64000000000c0002000800010001"], 0x3c}}, 0x0) sendmsg$nl_route(r0, &(0x7f0000000240)={0x0, 0x0, &(0x7f0000000680)={&(0x7f0000000400)=@newlink={0x3c, 0x10, 0xffffff1f, 0x0, 0x253c47, {}, [@IFLA_LINKINFO={0x14, 0x12, 0x0, 0x1, @bridge={{0xb}, {0x4}}}, @IFLA_MASTER={0x8, 0xa, r3}]}, 0x3c}}, 0x0) [ 2854.142512][ T3644] workqueue: Failed to create a rescuer kthread for wq "bond1312": -EINTR [ 2854.231720][ T3669] x_tables: duplicate underflow at hook 3 04:13:01 executing program 4: r0 = socket$inet6(0xa, 0x2, 0x0) ioctl$sock_SIOCETHTOOL(r0, 0x89f0, &(0x7f0000000000)={'bridge0\x00', &(0x7f0000000200)=ANY=[@ANYBLOB="10000000000000000000839d46d1ad3642fb17121a7682b043701e175ccfa33b6651"]}) r1 = socket$l2tp6(0xa, 0x2, 0x73) ioctl$F2FS_IOC_MOVE_RANGE(r0, 0xc020f509, &(0x7f0000000040)={r0, 0x1, 0xff, 0x6}) bind$l2tp6(r2, &(0x7f0000000080)={0xa, 0x0, 0xe84c, @mcast2, 0x10001, 0x3}, 0x20) setsockopt$inet6_MCAST_MSFILTER(r1, 0x29, 0x30, &(0x7f0000000240)={0xd2, {{0xa, 0x4e23, 0x8001, @ipv4={'\x00', '\xff\xff', @rand_addr=0x64010100}, 0xfffffffd}}, 0x1, 0x3, [{{0xa, 0x4e23, 0x9, @remote, 0xffffebe2}}, {{0xa, 0x4e21, 0x200, @ipv4={'\x00', '\xff\xff', @dev={0xac, 0x14, 0x14, 0x22}}, 0x5}}, {{0xa, 0x4e20, 0x23a, @local, 0x3}}]}, 0x210) r3 = socket$inet6(0xa, 0x80002, 0x88) bind$inet6(r3, &(0x7f0000000580)={0xa, 0x10010000004e20}, 0x1c) setsockopt$inet6_udp_int(r3, 0x11, 0xb, &(0x7f0000000080), 0x4) bind$l2tp6(r3, &(0x7f00000000c0)={0xa, 0x0, 0x2b5, @mcast1, 0x10001, 0x4}, 0x20) r4 = socket$inet6(0xa, 0x80002, 0x88) bind$inet6(r4, &(0x7f0000000580)={0xa, 0x10010000004e20}, 0x1c) setsockopt$inet6_udp_int(r4, 0x11, 0xb, &(0x7f0000000080), 0x4) setsockopt$sock_int(r4, 0x1, 0x1, &(0x7f0000000100)=0x9, 0x4) 04:13:01 executing program 3: r0 = socket$netlink(0x10, 0x3, 0x0) sendmsg$netlink(r0, &(0x7f00000047c0)={0x0, 0x0, &(0x7f0000000000)=[{&(0x7f0000000100)=ANY=[@ANYBLOB="18000000560001060000009200db730007"], 0x18}], 0x1}, 0x0) [ 2854.410926][ T3657] netlink: 'syz-executor.5': attribute type 1 has an invalid length. 04:13:01 executing program 1: r0 = socket$netlink(0x10, 0x3, 0x0) r1 = socket$netlink(0x10, 0x3, 0x0) r2 = socket(0x10, 0x803, 0x0) sendmsg$NL80211_CMD_CRIT_PROTOCOL_START(r2, &(0x7f0000000580)={0x0, 0x0, &(0x7f0000000540)={0x0, 0x1c}}, 0x0) getsockname$packet(r2, &(0x7f0000000600)={0x11, 0x0, 0x0, 0x1, 0x0, 0x6, @broadcast}, &(0x7f0000000100)=0x14) sendmsg$nl_route(r1, &(0x7f0000000040)={0x0, 0x0, &(0x7f0000000000)={&(0x7f00000000c0)=ANY=[@ANYBLOB="3c00000010008506eb9afc4cd8d06e754a0081c5", @ANYRES32=r3, @ANYBLOB="2377f292252155b21c0012000c000100626f6e64000000000c0002000800010001"], 0x3c}}, 0x0) sendmsg$nl_route(r0, &(0x7f0000000240)={0x0, 0x0, &(0x7f0000000680)={&(0x7f0000000400)=@newlink={0x3c, 0x10, 0xffffff1f, 0x0, 0x24672500, {}, [@IFLA_LINKINFO={0x14, 0x12, 0x0, 0x1, @bridge={{0xb}, {0x4}}}, @IFLA_MASTER={0x8, 0xa, r3}]}, 0x3c}}, 0x0) [ 2854.558072][ T3657] bond1303: entered promiscuous mode [ 2854.563751][ T3657] 8021q: adding VLAN 0 to HW filter on device bond1303 [ 2854.652822][ T3661] bond1303: (slave bridge1266): making interface the new active one [ 2854.664484][ T3661] bridge1266: entered promiscuous mode [ 2854.677681][ T3661] bond1303: (slave bridge1266): Enslaving as an active interface with an up link 04:13:02 executing program 5: r0 = socket$netlink(0x10, 0x3, 0x0) r1 = socket$netlink(0x10, 0x3, 0x0) r2 = socket(0x10, 0x803, 0x0) sendmsg$NL80211_CMD_CRIT_PROTOCOL_START(r2, &(0x7f0000000580)={0x0, 0x0, &(0x7f0000000540)={0x0, 0x1c}}, 0x0) getsockname$packet(r2, &(0x7f0000000600)={0x11, 0x0, 0x0, 0x1, 0x0, 0x6, @broadcast}, &(0x7f0000000100)=0x14) sendmsg$nl_route(r1, &(0x7f0000000040)={0x0, 0x0, &(0x7f0000000000)={&(0x7f00000000c0)=ANY=[@ANYBLOB="3c00000010008506eb9afc4cd8d06e754a0081c5", @ANYRES32=r3, @ANYBLOB="2377f292252155b21c0012000c000100626f6e64000000000c0002000800010001"], 0x3c}}, 0x0) sendmsg$nl_route(r0, &(0x7f0000000240)={0x0, 0x0, &(0x7f0000000680)={&(0x7f0000000400)=@newlink={0x3c, 0x10, 0xffffff1f, 0x0, 0x4a000000, {}, [@IFLA_LINKINFO={0x14, 0x12, 0x0, 0x1, @bridge={{0xb}, {0x4}}}, @IFLA_MASTER={0x8, 0xa, r3}]}, 0x3c}}, 0x0) 04:13:02 executing program 0: r0 = socket$inet6(0xa, 0x6, 0x0) bind$inet6(r0, &(0x7f0000000000)={0xa, 0x4e20}, 0x1c) listen(r0, 0x6) r1 = socket$inet6(0xa, 0x6, 0x0) connect$inet6(r1, &(0x7f0000000200)={0xa, 0x0, 0x0, @initdev={0xfe, 0x88, '\x00', 0x0, 0x0}, 0x3}, 0x1c) connect$inet6(r1, &(0x7f0000000040)={0xa, 0x4e20, 0x0, @dev={0xfe, 0x80, '\x00', 0xf}}, 0x1c) sendmmsg$inet6(r1, &(0x7f0000005a00)=[{{0x0, 0x0, 0x0}}, {{0x0, 0x0, 0x0, 0x0, &(0x7f0000005bc0)=ANY=[@ANYBLOB="5000000000000000290000003600000000060000000000000401000730000000000a00000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000200000000000000029000000370000000000000000001f0004010000000000001400000000000000290000000000000160000000000000002900000037"], 0x100}}], 0x2, 0x0) [ 2854.766037][ T3673] netlink: 'syz-executor.2': attribute type 1 has an invalid length. [ 2854.828424][ T3673] bond1295: entered promiscuous mode [ 2854.836174][ T3673] 8021q: adding VLAN 0 to HW filter on device bond1295 04:13:02 executing program 0: r0 = socket$inet6(0xa, 0x6, 0x0) bind$inet6(r0, &(0x7f0000000000)={0xa, 0x4e20}, 0x1c) listen(r0, 0x6) r1 = socket$inet6(0xa, 0x6, 0x0) connect$inet6(r1, &(0x7f0000000200)={0xa, 0x0, 0x0, @initdev={0xfe, 0x88, '\x00', 0x0, 0x0}, 0x3}, 0x1c) connect$inet6(r1, &(0x7f0000000040)={0xa, 0x4e20, 0x0, @dev={0xfe, 0x80, '\x00', 0xf}}, 0x1c) sendmmsg$inet6(r1, &(0x7f0000005a00)=[{{0x0, 0x0, 0x0}}, {{0x0, 0x0, 0x0, 0x0, &(0x7f0000005bc0)=ANY=[@ANYBLOB="5000000000000000290000003600000000060000000000000401000730000000000a0000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000020000000000000002900000037000000000000000000030004010000000000001400000000000000290000000000000160000000000000002900000037"], 0x100}}], 0x2, 0x0) 04:13:02 executing program 0: r0 = socket$inet6(0xa, 0x6, 0x0) bind$inet6(r0, &(0x7f0000000000)={0xa, 0x4e20}, 0x1c) listen(r0, 0x6) r1 = socket$inet6(0xa, 0x6, 0x0) connect$inet6(r1, &(0x7f0000000200)={0xa, 0x0, 0x0, @initdev={0xfe, 0x88, '\x00', 0x0, 0x0}, 0x3}, 0x1c) connect$inet6(r1, &(0x7f0000000040)={0xa, 0x4e20, 0x0, @dev={0xfe, 0x80, '\x00', 0xf}}, 0x1c) sendmmsg$inet6(r1, &(0x7f0000005a00)=[{{0x0, 0x0, 0x0}}, {{0x0, 0x0, 0x0, 0x0, &(0x7f0000005bc0)=ANY=[@ANYBLOB="5000000000000000290000003600000000060000000000000401000730000000000a0000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000020000000000000002900000037000000000000000000020004010000000000001400000000000000290000000000000160000000000000002900000037"], 0x100}}], 0x2, 0x0) 04:13:02 executing program 0: r0 = socket$inet6(0xa, 0x6, 0x0) bind$inet6(r0, &(0x7f0000000000)={0xa, 0x4e20}, 0x1c) listen(r0, 0x6) r1 = socket$inet6(0xa, 0x6, 0x0) connect$inet6(r1, &(0x7f0000000200)={0xa, 0x0, 0x0, @initdev={0xfe, 0x88, '\x00', 0x0, 0x0}, 0x3}, 0x1c) connect$inet6(r1, &(0x7f0000000040)={0xa, 0x4e20, 0x0, @dev={0xfe, 0x80, '\x00', 0xf}}, 0x1c) sendmmsg$inet6(r1, &(0x7f0000005a00)=[{{0x0, 0x0, 0x0}}, {{0x0, 0x0, 0x0, 0x0, &(0x7f0000005bc0)=ANY=[@ANYBLOB="5000000000000000290000003600000000060000000000000401000730000000000a0000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000020000000000000002900000037000000000000000000020004010000000000001400000000000000290000000000000160000000000000002900000037"], 0x100}}], 0x2, 0x0) [ 2855.316809][ T3675] bond1295: (slave bridge1216): making interface the new active one [ 2855.344611][ T3675] bridge1216: entered promiscuous mode 04:13:02 executing program 3: r0 = socket$netlink(0x10, 0x3, 0x0) sendmsg$netlink(r0, &(0x7f00000047c0)={0x0, 0x0, &(0x7f0000000000)=[{&(0x7f0000000100)=ANY=[@ANYBLOB="18000000560001060000009300db730007"], 0x18}], 0x1}, 0x0) 04:13:02 executing program 0: r0 = socket$inet6(0xa, 0x6, 0x0) bind$inet6(r0, &(0x7f0000000000)={0xa, 0x4e20}, 0x1c) listen(r0, 0x6) r1 = socket$inet6(0xa, 0x6, 0x0) connect$inet6(r1, &(0x7f0000000200)={0xa, 0x0, 0x0, @initdev={0xfe, 0x88, '\x00', 0x0, 0x0}, 0x3}, 0x1c) connect$inet6(r1, &(0x7f0000000040)={0xa, 0x4e20, 0x0, @dev={0xfe, 0x80, '\x00', 0xf}}, 0x1c) sendmmsg$inet6(r1, &(0x7f0000005a00)=[{{0x0, 0x0, 0x0}}, {{0x0, 0x0, 0x0, 0x0, &(0x7f0000005bc0)=ANY=[@ANYBLOB="5000000000000000290000003600000000060000000000000401000730000000000a0000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000020000000000000002900000037000000000000001f00000004010000000000001400000000000000290000000000000160000000000000002900000037"], 0x100}}], 0x2, 0x0) 04:13:02 executing program 4: r0 = socket$inet6(0xa, 0x2, 0x0) ioctl$sock_SIOCETHTOOL(r0, 0x89f0, &(0x7f0000000000)={'bridge0\x00', &(0x7f0000000200)=ANY=[@ANYBLOB="10000000000000000000839d46d1ad3642fb17121a7682b043701e175ccfa33b6651"]}) r1 = socket$l2tp6(0xa, 0x2, 0x73) ioctl$F2FS_IOC_MOVE_RANGE(r0, 0xc020f509, &(0x7f0000000040)={r0, 0x1, 0xff, 0x6}) (async) ioctl$F2FS_IOC_MOVE_RANGE(r0, 0xc020f509, &(0x7f0000000040)={r0, 0x1, 0xff, 0x6}) bind$l2tp6(r2, &(0x7f0000000080)={0xa, 0x0, 0xe84c, @mcast2, 0x10001, 0x3}, 0x20) setsockopt$inet6_MCAST_MSFILTER(r1, 0x29, 0x30, &(0x7f0000000240)={0xd2, {{0xa, 0x4e23, 0x8001, @ipv4={'\x00', '\xff\xff', @rand_addr=0x64010100}, 0xfffffffd}}, 0x1, 0x3, [{{0xa, 0x4e23, 0x9, @remote, 0xffffebe2}}, {{0xa, 0x4e21, 0x200, @ipv4={'\x00', '\xff\xff', @dev={0xac, 0x14, 0x14, 0x22}}, 0x5}}, {{0xa, 0x4e20, 0x23a, @local, 0x3}}]}, 0x210) r3 = socket$inet6(0xa, 0x80002, 0x88) bind$inet6(r3, &(0x7f0000000580)={0xa, 0x10010000004e20}, 0x1c) (async) bind$inet6(r3, &(0x7f0000000580)={0xa, 0x10010000004e20}, 0x1c) setsockopt$inet6_udp_int(r3, 0x11, 0xb, &(0x7f0000000080), 0x4) (async) setsockopt$inet6_udp_int(r3, 0x11, 0xb, &(0x7f0000000080), 0x4) bind$l2tp6(r3, &(0x7f00000000c0)={0xa, 0x0, 0x2b5, @mcast1, 0x10001, 0x4}, 0x20) r4 = socket$inet6(0xa, 0x80002, 0x88) bind$inet6(r4, &(0x7f0000000580)={0xa, 0x10010000004e20}, 0x1c) setsockopt$inet6_udp_int(r4, 0x11, 0xb, &(0x7f0000000080), 0x4) setsockopt$sock_int(r4, 0x1, 0x1, &(0x7f0000000100)=0x9, 0x4) (async) setsockopt$sock_int(r4, 0x1, 0x1, &(0x7f0000000100)=0x9, 0x4) [ 2855.370772][ T3675] bond1295: (slave bridge1216): Enslaving as an active interface with an up link [ 2855.408708][ T3682] netlink: 'syz-executor.1': attribute type 1 has an invalid length. 04:13:03 executing program 2: r0 = socket$netlink(0x10, 0x3, 0x0) r1 = socket$netlink(0x10, 0x3, 0x0) r2 = socket(0x10, 0x803, 0x0) sendmsg$NL80211_CMD_CRIT_PROTOCOL_START(r2, &(0x7f0000000580)={0x0, 0x0, &(0x7f0000000540)={0x0, 0x1c}}, 0x0) getsockname$packet(r2, &(0x7f0000000600)={0x11, 0x0, 0x0, 0x1, 0x0, 0x6, @broadcast}, &(0x7f0000000100)=0x14) sendmsg$nl_route(r1, &(0x7f0000000040)={0x0, 0x0, &(0x7f0000000000)={&(0x7f00000000c0)=ANY=[@ANYBLOB="3c00000010008506eb9afc4cd8d06e754a0081c5", @ANYRES32=r3, @ANYBLOB="2377f292252155b21c0012000c000100626f6e64000000000c0002000800010001"], 0x3c}}, 0x0) sendmsg$nl_route(r0, &(0x7f0000000240)={0x0, 0x0, &(0x7f0000000680)={&(0x7f0000000400)=@newlink={0x3c, 0x10, 0xffffff1f, 0x0, 0x253c48, {}, [@IFLA_LINKINFO={0x14, 0x12, 0x0, 0x1, @bridge={{0xb}, {0x4}}}, @IFLA_MASTER={0x8, 0xa, r3}]}, 0x3c}}, 0x0) 04:13:03 executing program 0: r0 = socket$inet6(0xa, 0x6, 0x0) bind$inet6(r0, &(0x7f0000000000)={0xa, 0x4e20}, 0x1c) listen(r0, 0x6) r1 = socket$inet6(0xa, 0x6, 0x0) connect$inet6(r1, &(0x7f0000000200)={0xa, 0x0, 0x0, @initdev={0xfe, 0x88, '\x00', 0x0, 0x0}, 0x3}, 0x1c) connect$inet6(r1, &(0x7f0000000040)={0xa, 0x4e20, 0x0, @dev={0xfe, 0x80, '\x00', 0xf}}, 0x1c) sendmmsg$inet6(r1, &(0x7f0000005a00)=[{{0x0, 0x0, 0x0}}, {{0x0, 0x0, 0x0, 0x0, &(0x7f0000005bc0)=ANY=[@ANYBLOB="5000000000000000290000003600000000060000000000000401000730000000000a0000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000020000000000000002900000037000000000000001f00000004010000000000001400000000000000290000000000000160000000000000002900000037"], 0x100}}], 0x2, 0x0) 04:13:03 executing program 1: r0 = socket$netlink(0x10, 0x3, 0x0) r1 = socket$netlink(0x10, 0x3, 0x0) r2 = socket(0x10, 0x803, 0x0) sendmsg$NL80211_CMD_CRIT_PROTOCOL_START(r2, &(0x7f0000000580)={0x0, 0x0, &(0x7f0000000540)={0x0, 0x1c}}, 0x0) getsockname$packet(r2, &(0x7f0000000600)={0x11, 0x0, 0x0, 0x1, 0x0, 0x6, @broadcast}, &(0x7f0000000100)=0x14) sendmsg$nl_route(r1, &(0x7f0000000040)={0x0, 0x0, &(0x7f0000000000)={&(0x7f00000000c0)=ANY=[@ANYBLOB="3c00000010008506eb9afc4cd8d06e754a0081c5", @ANYRES32=r3, @ANYBLOB="2377f292252155b21c0012000c000100626f6e64000000000c0002000800010001"], 0x3c}}, 0x0) sendmsg$nl_route(r0, &(0x7f0000000240)={0x0, 0x0, &(0x7f0000000680)={&(0x7f0000000400)=@newlink={0x3c, 0x10, 0xffffff1f, 0x0, 0x25672500, {}, [@IFLA_LINKINFO={0x14, 0x12, 0x0, 0x1, @bridge={{0xb}, {0x4}}}, @IFLA_MASTER={0x8, 0xa, r3}]}, 0x3c}}, 0x0) 04:13:03 executing program 0: r0 = socket$inet6(0xa, 0x6, 0x0) bind$inet6(r0, &(0x7f0000000000)={0xa, 0x4e20}, 0x1c) listen(r0, 0x6) r1 = socket$inet6(0xa, 0x6, 0x0) connect$inet6(r1, &(0x7f0000000200)={0xa, 0x0, 0x0, @initdev={0xfe, 0x88, '\x00', 0x0, 0x0}, 0x3}, 0x1c) connect$inet6(r1, &(0x7f0000000040)={0xa, 0x4e20, 0x0, @dev={0xfe, 0x80, '\x00', 0xf}}, 0x1c) sendmmsg$inet6(r1, &(0x7f0000005a00)=[{{0x0, 0x0, 0x0}}, {{0x0, 0x0, 0x0, 0x0, &(0x7f0000005bc0)=ANY=[@ANYBLOB="5000000000000000290000003600000000060000000000000401000730000000000a0000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000020000000000000002900000037000000000000000060000004010000000000001400000000000000290000000000000160000000000000002900000037"], 0x100}}], 0x2, 0x0) [ 2855.438826][ T3682] workqueue: Failed to create a rescuer kthread for wq "bond1312": -EINTR [ 2855.894422][ T3688] netlink: 'syz-executor.5': attribute type 1 has an invalid length. 04:13:03 executing program 5: r0 = socket$netlink(0x10, 0x3, 0x0) r1 = socket$netlink(0x10, 0x3, 0x0) r2 = socket(0x10, 0x803, 0x0) sendmsg$NL80211_CMD_CRIT_PROTOCOL_START(r2, &(0x7f0000000580)={0x0, 0x0, &(0x7f0000000540)={0x0, 0x1c}}, 0x0) getsockname$packet(r2, &(0x7f0000000600)={0x11, 0x0, 0x0, 0x1, 0x0, 0x6, @broadcast}, &(0x7f0000000100)=0x14) sendmsg$nl_route(r1, &(0x7f0000000040)={0x0, 0x0, &(0x7f0000000000)={&(0x7f00000000c0)=ANY=[@ANYBLOB="3c00000010008506eb9afc4cd8d06e754a0081c5", @ANYRES32=r3, @ANYBLOB="2377f292252155b21c0012000c000100626f6e64000000000c0002000800010001"], 0x3c}}, 0x0) sendmsg$nl_route(r0, &(0x7f0000000240)={0x0, 0x0, &(0x7f0000000680)={&(0x7f0000000400)=@newlink={0x3c, 0x10, 0xffffff1f, 0x0, 0x4c000000, {}, [@IFLA_LINKINFO={0x14, 0x12, 0x0, 0x1, @bridge={{0xb}, {0x4}}}, @IFLA_MASTER={0x8, 0xa, r3}]}, 0x3c}}, 0x0) 04:13:03 executing program 0: r0 = socket$inet6(0xa, 0x6, 0x0) bind$inet6(r0, &(0x7f0000000000)={0xa, 0x4e20}, 0x1c) listen(r0, 0x6) r1 = socket$inet6(0xa, 0x6, 0x0) connect$inet6(r1, &(0x7f0000000200)={0xa, 0x0, 0x0, @initdev={0xfe, 0x88, '\x00', 0x0, 0x0}, 0x3}, 0x1c) connect$inet6(r1, &(0x7f0000000040)={0xa, 0x4e20, 0x0, @dev={0xfe, 0x80, '\x00', 0xf}}, 0x1c) sendmmsg$inet6(r1, &(0x7f0000005a00)=[{{0x0, 0x0, 0x0}}, {{0x0, 0x0, 0x0, 0x0, &(0x7f0000005bc0)=ANY=[@ANYBLOB="5000000000000000290000003600000000060000000000000401000730000000000a0000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000020000000000000002900000037000000000000011000000004010000000000001400000000000000290000000000000160000000000000002900000037"], 0x100}}], 0x2, 0x0) [ 2855.967580][ T3688] workqueue: Failed to create a rescuer kthread for wq "bond1304": -EINTR [ 2856.130323][ T3710] netlink: 'syz-executor.2': attribute type 1 has an invalid length. 04:13:03 executing program 3: r0 = socket$netlink(0x10, 0x3, 0x0) sendmsg$netlink(r0, &(0x7f00000047c0)={0x0, 0x0, &(0x7f0000000000)=[{&(0x7f0000000100)=ANY=[@ANYBLOB="18000000560001060000009400db730007"], 0x18}], 0x1}, 0x0) 04:13:03 executing program 4: r0 = socket$inet6(0xa, 0x2, 0x0) ioctl$sock_SIOCETHTOOL(r0, 0x89f0, &(0x7f0000000000)={'bridge0\x00', &(0x7f0000000200)=ANY=[@ANYBLOB="10000000000000000000839d46d1ad3642fb17121a7682b043701e175ccfa33b6651"]}) (async) r1 = socket$l2tp6(0xa, 0x2, 0x73) (async) ioctl$F2FS_IOC_MOVE_RANGE(r0, 0xc020f509, &(0x7f0000000040)={r0, 0x1, 0xff, 0x6}) bind$l2tp6(r2, &(0x7f0000000080)={0xa, 0x0, 0xe84c, @mcast2, 0x10001, 0x3}, 0x20) setsockopt$inet6_MCAST_MSFILTER(r1, 0x29, 0x30, &(0x7f0000000240)={0xd2, {{0xa, 0x4e23, 0x8001, @ipv4={'\x00', '\xff\xff', @rand_addr=0x64010100}, 0xfffffffd}}, 0x1, 0x3, [{{0xa, 0x4e23, 0x9, @remote, 0xffffebe2}}, {{0xa, 0x4e21, 0x200, @ipv4={'\x00', '\xff\xff', @dev={0xac, 0x14, 0x14, 0x22}}, 0x5}}, {{0xa, 0x4e20, 0x23a, @local, 0x3}}]}, 0x210) (async) r3 = socket$inet6(0xa, 0x80002, 0x88) bind$inet6(r3, &(0x7f0000000580)={0xa, 0x10010000004e20}, 0x1c) (async) setsockopt$inet6_udp_int(r3, 0x11, 0xb, &(0x7f0000000080), 0x4) bind$l2tp6(r3, &(0x7f00000000c0)={0xa, 0x0, 0x2b5, @mcast1, 0x10001, 0x4}, 0x20) r4 = socket$inet6(0xa, 0x80002, 0x88) bind$inet6(r4, &(0x7f0000000580)={0xa, 0x10010000004e20}, 0x1c) (async) setsockopt$inet6_udp_int(r4, 0x11, 0xb, &(0x7f0000000080), 0x4) setsockopt$sock_int(r4, 0x1, 0x1, &(0x7f0000000100)=0x9, 0x4) 04:13:03 executing program 0: r0 = socket$inet6(0xa, 0x6, 0x0) bind$inet6(r0, &(0x7f0000000000)={0xa, 0x4e20}, 0x1c) listen(r0, 0x6) r1 = socket$inet6(0xa, 0x6, 0x0) connect$inet6(r1, &(0x7f0000000200)={0xa, 0x0, 0x0, @initdev={0xfe, 0x88, '\x00', 0x0, 0x0}, 0x3}, 0x1c) connect$inet6(r1, &(0x7f0000000040)={0xa, 0x4e20, 0x0, @dev={0xfe, 0x80, '\x00', 0xf}}, 0x1c) sendmmsg$inet6(r1, &(0x7f0000005a00)=[{{0x0, 0x0, 0x0}}, {{0x0, 0x0, 0x0, 0x0, &(0x7f0000005bc0)=ANY=[@ANYBLOB="5000000000000000290000003600000000060000000000000401000730000000000a00000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000200000000000000029000000370000000000b0ff0000000004010000000000001400000000000000290000000000000160000000000000002900000037"], 0x100}}], 0x2, 0x0) [ 2856.341445][ T3710] bond1296: entered promiscuous mode [ 2856.348449][ T3710] 8021q: adding VLAN 0 to HW filter on device bond1296 [ 2856.416170][ T3714] bond1296: (slave bridge1217): making interface the new active one [ 2856.459887][ T3714] bridge1217: entered promiscuous mode 04:13:04 executing program 0: r0 = socket$inet6(0xa, 0x6, 0x0) bind$inet6(r0, &(0x7f0000000000)={0xa, 0x4e20}, 0x1c) listen(r0, 0x6) r1 = socket$inet6(0xa, 0x6, 0x0) connect$inet6(r1, &(0x7f0000000200)={0xa, 0x0, 0x0, @initdev={0xfe, 0x88, '\x00', 0x0, 0x0}, 0x3}, 0x1c) connect$inet6(r1, &(0x7f0000000040)={0xa, 0x4e20, 0x0, @dev={0xfe, 0x80, '\x00', 0xf}}, 0x1c) sendmmsg$inet6(r1, &(0x7f0000005a00)=[{{0x0, 0x0, 0x0}}, {{0x0, 0x0, 0x0, 0x0, &(0x7f0000005bc0)=ANY=[@ANYBLOB="5000000000000000290000003600000000060000000000000401000730000000000a00000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000200000000000000029000000370000000000f4010000000004010000000000001400000000000000290000000000000160000000000000002900000037"], 0x100}}], 0x2, 0x0) 04:13:04 executing program 2: r0 = socket$netlink(0x10, 0x3, 0x0) r1 = socket$netlink(0x10, 0x3, 0x0) r2 = socket(0x10, 0x803, 0x0) sendmsg$NL80211_CMD_CRIT_PROTOCOL_START(r2, &(0x7f0000000580)={0x0, 0x0, &(0x7f0000000540)={0x0, 0x1c}}, 0x0) getsockname$packet(r2, &(0x7f0000000600)={0x11, 0x0, 0x0, 0x1, 0x0, 0x6, @broadcast}, &(0x7f0000000100)=0x14) sendmsg$nl_route(r1, &(0x7f0000000040)={0x0, 0x0, &(0x7f0000000000)={&(0x7f00000000c0)=ANY=[@ANYBLOB="3c00000010008506eb9afc4cd8d06e754a0081c5", @ANYRES32=r3, @ANYBLOB="2377f292252155b21c0012000c000100626f6e64000000000c0002000800010001"], 0x3c}}, 0x0) sendmsg$nl_route(r0, &(0x7f0000000240)={0x0, 0x0, &(0x7f0000000680)={&(0x7f0000000400)=@newlink={0x3c, 0x10, 0xffffff1f, 0x0, 0x253c49, {}, [@IFLA_LINKINFO={0x14, 0x12, 0x0, 0x1, @bridge={{0xb}, {0x4}}}, @IFLA_MASTER={0x8, 0xa, r3}]}, 0x3c}}, 0x0) [ 2856.515907][ T3714] bond1296: (slave bridge1217): Enslaving as an active interface with an up link [ 2856.555331][ T3721] netlink: 'syz-executor.1': attribute type 1 has an invalid length. 04:13:04 executing program 0: r0 = socket$netlink(0x10, 0x3, 0x0) r1 = socket$netlink(0x10, 0x3, 0x0) r2 = socket(0x10, 0x803, 0x0) sendmsg$NL80211_CMD_CRIT_PROTOCOL_START(r2, &(0x7f0000000580)={0x0, 0x0, &(0x7f0000000540)={0x0, 0x1c}}, 0x0) getsockname$packet(r2, &(0x7f0000000600)={0x11, 0x0, 0x0, 0x1, 0x0, 0x6, @broadcast}, &(0x7f0000000100)=0x14) sendmsg$nl_route(r1, &(0x7f0000000040)={0x0, 0x0, &(0x7f0000000000)={&(0x7f00000000c0)=ANY=[@ANYBLOB="3c00000010008506eb9afc4cd8d06e754a0081c5", @ANYRES32=r3, @ANYBLOB="2377f292252155b21c0012000c000100626f6e64000000000c0002000800010001"], 0x3c}}, 0x0) sendmsg$nl_route(r0, &(0x7f0000000240)={0x0, 0x0, &(0x7f0000000680)={&(0x7f0000000400)=@newlink={0x3c, 0x10, 0xffffff1f, 0x0, 0x1186, {}, [@IFLA_LINKINFO={0x14, 0x12, 0x0, 0x1, @bridge={{0xb}, {0x4}}}, @IFLA_MASTER={0x8, 0xa, r3}]}, 0x3c}}, 0x0) [ 2856.709079][ T3721] bond1312: entered promiscuous mode [ 2856.715000][ T3721] 8021q: adding VLAN 0 to HW filter on device bond1312 [ 2856.884719][ T3724] bond1312: (slave bridge1258): making interface the new active one [ 2856.893573][ T3724] bridge1258: entered promiscuous mode [ 2856.907818][ T3724] bond1312: (slave bridge1258): Enslaving as an active interface with an up link [ 2856.917511][ T3725] netlink: 'syz-executor.5': attribute type 1 has an invalid length. 04:13:04 executing program 1: r0 = socket$netlink(0x10, 0x3, 0x0) r1 = socket$netlink(0x10, 0x3, 0x0) r2 = socket(0x10, 0x803, 0x0) sendmsg$NL80211_CMD_CRIT_PROTOCOL_START(r2, &(0x7f0000000580)={0x0, 0x0, &(0x7f0000000540)={0x0, 0x1c}}, 0x0) getsockname$packet(r2, &(0x7f0000000600)={0x11, 0x0, 0x0, 0x1, 0x0, 0x6, @broadcast}, &(0x7f0000000100)=0x14) sendmsg$nl_route(r1, &(0x7f0000000040)={0x0, 0x0, &(0x7f0000000000)={&(0x7f00000000c0)=ANY=[@ANYBLOB="3c00000010008506eb9afc4cd8d06e754a0081c5", @ANYRES32=r3, @ANYBLOB="2377f292252155b21c0012000c000100626f6e64000000000c0002000800010001"], 0x3c}}, 0x0) sendmsg$nl_route(r0, &(0x7f0000000240)={0x0, 0x0, &(0x7f0000000680)={&(0x7f0000000400)=@newlink={0x3c, 0x10, 0xffffff1f, 0x0, 0x26672500, {}, [@IFLA_LINKINFO={0x14, 0x12, 0x0, 0x1, @bridge={{0xb}, {0x4}}}, @IFLA_MASTER={0x8, 0xa, r3}]}, 0x3c}}, 0x0) [ 2857.062476][ T3725] bond1304: entered promiscuous mode [ 2857.068062][ T3725] 8021q: adding VLAN 0 to HW filter on device bond1304 [ 2857.263824][T32023] BUG: MAX_LOCKDEP_CHAINS too low! [ 2857.268972][T32023] turning off the locking correctness validator. [ 2857.275304][T32023] CPU: 1 PID: 32023 Comm: kworker/u4:3 Not tainted 6.5.0-syzkaller-03967-gbd6c11bc43c4 #0 [ 2857.285210][T32023] Hardware name: Google Google Compute Engine/Google Compute Engine, BIOS Google 07/26/2023 [ 2857.285565][ T3730] bond1304: (slave bridge1267): making interface the new active one [ 2857.295250][T32023] Workqueue: bond1304 bond_netdev_notify_work [ 2857.303333][ T3730] bridge1267: entered promiscuous mode [ 2857.309347][T32023] Call Trace: [ 2857.309358][T32023] [ 2857.309369][T32023] dump_stack_lvl+0xd9/0x1b0 [ 2857.319776][ T3730] bond1304: (slave bridge1267): Enslaving as an active interface with an up link [ 2857.320999][T32023] __lock_acquire+0x4286/0x5de0 [ 2857.340029][T32023] ? lock_acquire+0x1ae/0x510 [ 2857.344744][T32023] ? lockdep_hardirqs_on_prepare+0x410/0x410 [ 2857.350762][T32023] ? debug_object_assert_init+0x220/0x370 [ 2857.356519][T32023] ? reacquire_held_locks+0x4b0/0x4b0 04:13:04 executing program 5: r0 = socket$netlink(0x10, 0x3, 0x0) r1 = socket$netlink(0x10, 0x3, 0x0) r2 = socket(0x10, 0x803, 0x0) sendmsg$NL80211_CMD_CRIT_PROTOCOL_START(r2, &(0x7f0000000580)={0x0, 0x0, &(0x7f0000000540)={0x0, 0x1c}}, 0x0) getsockname$packet(r2, &(0x7f0000000600)={0x11, 0x0, 0x0, 0x1, 0x0, 0x6, @broadcast}, &(0x7f0000000100)=0x14) sendmsg$nl_route(r1, &(0x7f0000000040)={0x0, 0x0, &(0x7f0000000000)={&(0x7f00000000c0)=ANY=[@ANYBLOB="3c00000010008506eb9afc4cd8d06e754a0081c5", @ANYRES32=r3, @ANYBLOB="2377f292252155b21c0012000c000100626f6e64000000000c0002000800010001"], 0x3c}}, 0x0) sendmsg$nl_route(r0, &(0x7f0000000240)={0x0, 0x0, &(0x7f0000000680)={&(0x7f0000000400)=@newlink={0x3c, 0x10, 0xffffff1f, 0x0, 0x52010000, {}, [@IFLA_LINKINFO={0x14, 0x12, 0x0, 0x1, @bridge={{0xb}, {0x4}}}, @IFLA_MASTER={0x8, 0xa, r3}]}, 0x3c}}, 0x0) [ 2857.361927][T32023] lock_acquire+0x1ae/0x510 [ 2857.366464][T32023] ? lock_timer_base+0x5d/0x200 [ 2857.371353][T32023] ? lock_sync+0x190/0x190 [ 2857.375805][T32023] ? _raw_spin_unlock_irqrestore+0x3b/0x70 [ 2857.381635][T32023] ? debug_object_assert_init+0x220/0x370 [ 2857.387370][T32023] ? stack_trace_save+0x96/0xd0 [ 2857.392258][T32023] ? debug_object_free+0x360/0x360 [ 2857.397399][T32023] _raw_spin_lock_irqsave+0x3a/0x50 [ 2857.402623][T32023] ? lock_timer_base+0x5d/0x200 [ 2857.407510][T32023] lock_timer_base+0x5d/0x200 [ 2857.412212][T32023] __mod_timer+0x420/0xea0 [ 2857.416658][T32023] ? timer_fixup_activate+0x2b0/0x2b0 [ 2857.422059][T32023] ? lockdep_hardirqs_on_prepare+0x410/0x410 [ 2857.428069][T32023] ? lockdep_hardirqs_on+0x7d/0x100 [ 2857.433317][T32023] add_timer+0x62/0x90 [ 2857.437417][T32023] __queue_delayed_work+0x19d/0x260 [ 2857.442647][T32023] queue_delayed_work_on+0x106/0x130 [ 2857.447974][T32023] bond_netdev_notify_work+0x26d/0x2c0 [ 2857.453459][T32023] ? bond_xmit_activebackup_slave_get+0xd0/0xd0 [ 2857.459732][T32023] ? reacquire_held_locks+0x4b0/0x4b0 [ 2857.465150][T32023] ? do_raw_spin_lock+0x12e/0x2b0 [ 2857.470209][T32023] ? spin_bug+0x1d0/0x1d0 [ 2857.474583][T32023] process_one_work+0xaa2/0x16f0 [ 2857.479556][T32023] ? bond_xmit_activebackup_slave_get+0xd0/0xd0 [ 2857.485826][T32023] ? pwq_dec_nr_in_flight+0x2a0/0x2a0 [ 2857.491232][T32023] ? spin_bug+0x1d0/0x1d0 [ 2857.495600][T32023] ? kthread_data+0x53/0xc0 [ 2857.500139][T32023] worker_thread+0x687/0x1110 [ 2857.504835][T32023] ? __kthread_parkme+0x152/0x220 [ 2857.508617][ T3757] bond1305: entered promiscuous mode [ 2857.509857][T32023] ? process_one_work+0x16f0/0x16f0 [ 2857.509895][T32023] kthread+0x33a/0x430 [ 2857.515535][ T3757] 8021q: adding VLAN 0 to HW filter on device bond1305 [ 2857.520331][T32023] ? kthread_complete_and_exit+0x40/0x40 [ 2857.520368][T32023] ret_from_fork+0x2c/0x70 [ 2857.520398][T32023] ? kthread_complete_and_exit+0x40/0x40 [ 2857.520427][T32023] ret_from_fork_asm+0x11/0x20 [ 2857.520498][T32023] 04:13:05 executing program 3: r0 = socket$netlink(0x10, 0x3, 0x0) sendmsg$netlink(r0, &(0x7f00000047c0)={0x0, 0x0, &(0x7f0000000000)=[{&(0x7f0000000100)=ANY=[@ANYBLOB="18000000560001060000009500db730007"], 0x18}], 0x1}, 0x0) 04:13:05 executing program 4: r0 = socket$inet6(0xa, 0x2, 0x0) ioctl$sock_SIOCETHTOOL(r0, 0x89f0, &(0x7f0000000000)={'bridge0\x00', &(0x7f0000000200)=@ethtool_ringparam={0x10}}) connect$inet6(r0, &(0x7f0000000040)={0xa, 0x4e20, 0x2, @rand_addr=' \x01\x00\x00\x00\x00\x00\x00\x00\x00\x00\x00\x00\x00\x00\x02', 0x2}, 0x1c) [ 2857.690473][ T3744] bond1297: entered promiscuous mode [ 2857.702592][ T3744] 8021q: adding VLAN 0 to HW filter on device bond1297 04:13:05 executing program 2: r0 = socket$netlink(0x10, 0x3, 0x0) r1 = socket$netlink(0x10, 0x3, 0x0) r2 = socket(0x10, 0x803, 0x0) sendmsg$NL80211_CMD_CRIT_PROTOCOL_START(r2, &(0x7f0000000580)={0x0, 0x0, &(0x7f0000000540)={0x0, 0x1c}}, 0x0) getsockname$packet(r2, &(0x7f0000000600)={0x11, 0x0, 0x0, 0x1, 0x0, 0x6, @broadcast}, &(0x7f0000000100)=0x14) sendmsg$nl_route(r1, &(0x7f0000000040)={0x0, 0x0, &(0x7f0000000000)={&(0x7f00000000c0)=ANY=[@ANYBLOB="3c00000010008506eb9afc4cd8d06e754a0081c5", @ANYRES32=r3, @ANYBLOB="2377f292252155b21c0012000c000100626f6e64000000000c0002000800010001"], 0x3c}}, 0x0) sendmsg$nl_route(r0, &(0x7f0000000240)={0x0, 0x0, &(0x7f0000000680)={&(0x7f0000000400)=@newlink={0x3c, 0x10, 0xffffff1f, 0x0, 0x253c4a, {}, [@IFLA_LINKINFO={0x14, 0x12, 0x0, 0x1, @bridge={{0xb}, {0x4}}}, @IFLA_MASTER={0x8, 0xa, r3}]}, 0x3c}}, 0x0) [ 2857.776904][ T3746] bond1297: (slave bridge1218): making interface the new active one [ 2857.788942][ T3746] bridge1218: entered promiscuous mode [ 2857.803806][ T3746] bond1297: (slave bridge1218): Enslaving as an active interface with an up link 04:13:05 executing program 0: r0 = socket$netlink(0x10, 0x3, 0x0) r1 = socket$netlink(0x10, 0x3, 0x0) r2 = socket(0x10, 0x803, 0x0) sendmsg$NL80211_CMD_CRIT_PROTOCOL_START(r2, &(0x7f0000000580)={0x0, 0x0, &(0x7f0000000540)={0x0, 0x1c}}, 0x0) getsockname$packet(r2, &(0x7f0000000600)={0x11, 0x0, 0x0, 0x1, 0x0, 0x6, @broadcast}, &(0x7f0000000100)=0x14) sendmsg$nl_route(r1, &(0x7f0000000040)={0x0, 0x0, &(0x7f0000000000)={&(0x7f00000000c0)=ANY=[@ANYBLOB="3c00000010008506eb9afc4cd8d06e754a0081c5", @ANYRES32=r3, @ANYBLOB="2377f292252155b21c0012000c000100626f6e64000000000c0002000800010001"], 0x3c}}, 0x0) sendmsg$nl_route(r0, &(0x7f0000000240)={0x0, 0x0, &(0x7f0000000680)={&(0x7f0000000400)=@newlink={0x3c, 0x10, 0xffffff1f, 0x0, 0x1186, {}, [@IFLA_LINKINFO={0x14, 0x12, 0x0, 0x1, @bridge={{0xb}, {0x4}}}, @IFLA_MASTER={0x8, 0xa, r3}]}, 0x3c}}, 0x0) [ 2857.819208][ T3748] workqueue: Failed to create a rescuer kthread for wq "bond16": -EINTR [ 2857.929732][ T3753] bond1313: entered promiscuous mode [ 2857.948202][ T3753] 8021q: adding VLAN 0 to HW filter on device bond1313 04:13:05 executing program 1: r0 = socket$netlink(0x10, 0x3, 0x0) r1 = socket$netlink(0x10, 0x3, 0x0) r2 = socket(0x10, 0x803, 0x0) sendmsg$NL80211_CMD_CRIT_PROTOCOL_START(r2, &(0x7f0000000580)={0x0, 0x0, &(0x7f0000000540)={0x0, 0x1c}}, 0x0) getsockname$packet(r2, &(0x7f0000000600)={0x11, 0x0, 0x0, 0x1, 0x0, 0x6, @broadcast}, &(0x7f0000000100)=0x14) sendmsg$nl_route(r1, &(0x7f0000000040)={0x0, 0x0, &(0x7f0000000000)={&(0x7f00000000c0)=ANY=[@ANYBLOB="3c00000010008506eb9afc4cd8d06e754a0081c5", @ANYRES32=r3, @ANYBLOB="2377f292252155b21c0012000c000100626f6e64000000000c0002000800010001"], 0x3c}}, 0x0) sendmsg$nl_route(r0, &(0x7f0000000240)={0x0, 0x0, &(0x7f0000000680)={&(0x7f0000000400)=@newlink={0x3c, 0x10, 0xffffff1f, 0x0, 0x27672500, {}, [@IFLA_LINKINFO={0x14, 0x12, 0x0, 0x1, @bridge={{0xb}, {0x4}}}, @IFLA_MASTER={0x8, 0xa, r3}]}, 0x3c}}, 0x0) [ 2857.988786][ T3754] bond1313: (slave bridge1259): making interface the new active one [ 2857.997112][ T3754] bridge1259: entered promiscuous mode [ 2858.011005][ T3754] bond1313: (slave bridge1259): Enslaving as an active interface with an up link 04:13:05 executing program 5: r0 = socket$netlink(0x10, 0x3, 0x0) r1 = socket$netlink(0x10, 0x3, 0x0) r2 = socket(0x10, 0x803, 0x0) sendmsg$NL80211_CMD_CRIT_PROTOCOL_START(r2, &(0x7f0000000580)={0x0, 0x0, &(0x7f0000000540)={0x0, 0x1c}}, 0x0) getsockname$packet(r2, &(0x7f0000000600)={0x11, 0x0, 0x0, 0x1, 0x0, 0x6, @broadcast}, &(0x7f0000000100)=0x14) sendmsg$nl_route(r1, &(0x7f0000000040)={0x0, 0x0, &(0x7f0000000000)={&(0x7f00000000c0)=ANY=[@ANYBLOB="3c00000010008506eb9afc4cd8d06e754a0081c5", @ANYRES32=r3, @ANYBLOB="2377f292252155b21c0012000c000100626f6e64000000000c0002000800010001"], 0x3c}}, 0x0) sendmsg$nl_route(r0, &(0x7f0000000240)={0x0, 0x0, &(0x7f0000000680)={&(0x7f0000000400)=@newlink={0x3c, 0x10, 0xffffff1f, 0x0, 0x52020000, {}, [@IFLA_LINKINFO={0x14, 0x12, 0x0, 0x1, @bridge={{0xb}, {0x4}}}, @IFLA_MASTER={0x8, 0xa, r3}]}, 0x3c}}, 0x0) [ 2858.053676][ T3760] bond1305: (slave bridge1268): making interface the new active one [ 2858.062152][ T3760] bridge1268: entered promiscuous mode [ 2858.074492][ T3760] bond1305: (slave bridge1268): Enslaving as an active interface with an up link 04:13:05 executing program 3: r0 = socket$netlink(0x10, 0x3, 0x0) sendmsg$netlink(r0, &(0x7f00000047c0)={0x0, 0x0, &(0x7f0000000000)=[{&(0x7f0000000100)=ANY=[@ANYBLOB="18000000560001060000009600db730007"], 0x18}], 0x1}, 0x0) 04:13:05 executing program 4: r0 = socket$inet6(0xa, 0x2, 0x0) ioctl$sock_SIOCETHTOOL(r0, 0x89f0, &(0x7f0000000000)={'bridge0\x00', &(0x7f0000000200)=@ethtool_ringparam={0x10}}) (async) connect$inet6(r0, &(0x7f0000000040)={0xa, 0x4e20, 0x2, @rand_addr=' \x01\x00\x00\x00\x00\x00\x00\x00\x00\x00\x00\x00\x00\x00\x02', 0x2}, 0x1c) [ 2858.237624][ T3777] bond1314: entered promiscuous mode [ 2858.245542][ T3777] 8021q: adding VLAN 0 to HW filter on device bond1314 [ 2858.304764][ T3770] bond1298: entered promiscuous mode [ 2858.322948][ T3770] 8021q: adding VLAN 0 to HW filter on device bond1298 [ 2858.398752][ T3772] bond1298: (slave bridge1219): making interface the new active one [ 2858.407603][ T3772] bridge1219: entered promiscuous mode [ 2858.421654][ T3772] bond1298: (slave bridge1219): Enslaving as an active interface with an up link [ 2858.430978][ T3774] validate_nla: 6 callbacks suppressed 04:13:05 executing program 2: r0 = socket$netlink(0x10, 0x3, 0x0) r1 = socket$netlink(0x10, 0x3, 0x0) r2 = socket(0x10, 0x803, 0x0) sendmsg$NL80211_CMD_CRIT_PROTOCOL_START(r2, &(0x7f0000000580)={0x0, 0x0, &(0x7f0000000540)={0x0, 0x1c}}, 0x0) getsockname$packet(r2, &(0x7f0000000600)={0x11, 0x0, 0x0, 0x1, 0x0, 0x6, @broadcast}, &(0x7f0000000100)=0x14) sendmsg$nl_route(r1, &(0x7f0000000040)={0x0, 0x0, &(0x7f0000000000)={&(0x7f00000000c0)=ANY=[@ANYBLOB="3c00000010008506eb9afc4cd8d06e754a0081c5", @ANYRES32=r3, @ANYBLOB="2377f292252155b21c0012000c000100626f6e64000000000c0002000800010001"], 0x3c}}, 0x0) sendmsg$nl_route(r0, &(0x7f0000000240)={0x0, 0x0, &(0x7f0000000680)={&(0x7f0000000400)=@newlink={0x3c, 0x10, 0xffffff1f, 0x0, 0x253c4b, {}, [@IFLA_LINKINFO={0x14, 0x12, 0x0, 0x1, @bridge={{0xb}, {0x4}}}, @IFLA_MASTER={0x8, 0xa, r3}]}, 0x3c}}, 0x0) [ 2858.430995][ T3774] netlink: 'syz-executor.0': attribute type 1 has an invalid length. [ 2858.545297][ T3774] bond16: entered promiscuous mode [ 2858.560660][ T3774] 8021q: adding VLAN 0 to HW filter on device bond16 04:13:06 executing program 4: socket$inet6(0xa, 0x2, 0x0) (async) r0 = socket$inet6(0xa, 0x2, 0x0) ioctl$sock_SIOCETHTOOL(r0, 0x89f0, &(0x7f0000000000)={'bridge0\x00', &(0x7f0000000200)=@ethtool_ringparam={0x10}}) connect$inet6(r0, &(0x7f0000000040)={0xa, 0x4e20, 0x2, @rand_addr=' \x01\x00\x00\x00\x00\x00\x00\x00\x00\x00\x00\x00\x00\x00\x02', 0x2}, 0x1c) [ 2858.612687][ T3775] bond16: (slave bridge24): making interface the new active one [ 2858.620688][ T3775] bridge24: entered promiscuous mode [ 2858.630587][ T3775] bond16: (slave bridge24): Enslaving as an active interface with an up link [ 2858.651664][ T3784] netlink: 'syz-executor.5': attribute type 1 has an invalid length. [ 2858.699421][ T3784] bond1306: entered promiscuous mode [ 2858.710147][ T3784] 8021q: adding VLAN 0 to HW filter on device bond1306 04:13:06 executing program 0: r0 = socket$netlink(0x10, 0x3, 0x0) sendmsg$netlink(r0, &(0x7f00000047c0)={0x0, 0x0, &(0x7f0000000000)=[{&(0x7f0000000100)=ANY=[@ANYBLOB="18000000560001060000009500db730007"], 0x18}], 0x1}, 0x0) [ 2858.811198][ T3785] bond1314: (slave bridge1260): making interface the new active one [ 2858.830969][ T3785] bridge1260: entered promiscuous mode 04:13:06 executing program 3: r0 = socket$netlink(0x10, 0x3, 0x0) sendmsg$netlink(r0, &(0x7f00000047c0)={0x0, 0x0, &(0x7f0000000000)=[{&(0x7f0000000100)=ANY=[@ANYBLOB="18000000560001060000009700db730007"], 0x18}], 0x1}, 0x0) [ 2858.880826][ T3785] bond1314: (slave bridge1260): Enslaving as an active interface with an up link 04:13:06 executing program 1: r0 = socket$netlink(0x10, 0x3, 0x0) r1 = socket$netlink(0x10, 0x3, 0x0) r2 = socket(0x10, 0x803, 0x0) sendmsg$NL80211_CMD_CRIT_PROTOCOL_START(r2, &(0x7f0000000580)={0x0, 0x0, &(0x7f0000000540)={0x0, 0x1c}}, 0x0) getsockname$packet(r2, &(0x7f0000000600)={0x11, 0x0, 0x0, 0x1, 0x0, 0x6, @broadcast}, &(0x7f0000000100)=0x14) sendmsg$nl_route(r1, &(0x7f0000000040)={0x0, 0x0, &(0x7f0000000000)={&(0x7f00000000c0)=ANY=[@ANYBLOB="3c00000010008506eb9afc4cd8d06e754a0081c5", @ANYRES32=r3, @ANYBLOB="2377f292252155b21c0012000c000100626f6e64000000000c0002000800010001"], 0x3c}}, 0x0) sendmsg$nl_route(r0, &(0x7f0000000240)={0x0, 0x0, &(0x7f0000000680)={&(0x7f0000000400)=@newlink={0x3c, 0x10, 0xffffff1f, 0x0, 0x28672500, {}, [@IFLA_LINKINFO={0x14, 0x12, 0x0, 0x1, @bridge={{0xb}, {0x4}}}, @IFLA_MASTER={0x8, 0xa, r3}]}, 0x3c}}, 0x0) [ 2859.009043][ T3790] bond1306: (slave bridge1269): making interface the new active one [ 2859.030536][ T3790] bridge1269: entered promiscuous mode [ 2859.048923][ T3790] bond1306: (slave bridge1269): Enslaving as an active interface with an up link 04:13:06 executing program 5: r0 = socket$netlink(0x10, 0x3, 0x0) r1 = socket$netlink(0x10, 0x3, 0x0) r2 = socket(0x10, 0x803, 0x0) sendmsg$NL80211_CMD_CRIT_PROTOCOL_START(r2, &(0x7f0000000580)={0x0, 0x0, &(0x7f0000000540)={0x0, 0x1c}}, 0x0) getsockname$packet(r2, &(0x7f0000000600)={0x11, 0x0, 0x0, 0x1, 0x0, 0x6, @broadcast}, &(0x7f0000000100)=0x14) sendmsg$nl_route(r1, &(0x7f0000000040)={0x0, 0x0, &(0x7f0000000000)={&(0x7f00000000c0)=ANY=[@ANYBLOB="3c00000010008506eb9afc4cd8d06e754a0081c5", @ANYRES32=r3, @ANYBLOB="2377f292252155b21c0012000c000100626f6e64000000000c0002000800010001"], 0x3c}}, 0x0) sendmsg$nl_route(r0, &(0x7f0000000240)={0x0, 0x0, &(0x7f0000000680)={&(0x7f0000000400)=@newlink={0x3c, 0x10, 0xffffff1f, 0x0, 0x5c120000, {}, [@IFLA_LINKINFO={0x14, 0x12, 0x0, 0x1, @bridge={{0xb}, {0x4}}}, @IFLA_MASTER={0x8, 0xa, r3}]}, 0x3c}}, 0x0) [ 2859.106494][ T3795] netlink: 'syz-executor.2': attribute type 1 has an invalid length. [ 2859.173596][ T3795] bond1299: entered promiscuous mode [ 2859.179330][ T3795] 8021q: adding VLAN 0 to HW filter on device bond1299 [ 2859.229155][ T3796] bond1299: (slave bridge1220): making interface the new active one [ 2859.248999][ T3796] bridge1220: entered promiscuous mode [ 2859.261023][ T3796] bond1299: (slave bridge1220): Enslaving as an active interface with an up link 04:13:06 executing program 4: r0 = socket$inet6(0xa, 0x2, 0x0) ioctl$sock_SIOCETHTOOL(r0, 0x89f0, &(0x7f0000000000)={'ip6_vti0\x00', &(0x7f0000000040)=@ethtool_stats={0x1d, 0x6, [0xc50b, 0x3471, 0x3ff, 0x6, 0x101, 0x2000004e]}}) 04:13:06 executing program 2: r0 = socket$netlink(0x10, 0x3, 0x0) r1 = socket$netlink(0x10, 0x3, 0x0) r2 = socket(0x10, 0x803, 0x0) sendmsg$NL80211_CMD_CRIT_PROTOCOL_START(r2, &(0x7f0000000580)={0x0, 0x0, &(0x7f0000000540)={0x0, 0x1c}}, 0x0) getsockname$packet(r2, &(0x7f0000000600)={0x11, 0x0, 0x0, 0x1, 0x0, 0x6, @broadcast}, &(0x7f0000000100)=0x14) sendmsg$nl_route(r1, &(0x7f0000000040)={0x0, 0x0, &(0x7f0000000000)={&(0x7f00000000c0)=ANY=[@ANYBLOB="3c00000010008506eb9afc4cd8d06e754a0081c5", @ANYRES32=r3, @ANYBLOB="2377f292252155b21c0012000c000100626f6e64000000000c0002000800010001"], 0x3c}}, 0x0) sendmsg$nl_route(r0, &(0x7f0000000240)={0x0, 0x0, &(0x7f0000000680)={&(0x7f0000000400)=@newlink={0x3c, 0x10, 0xffffff1f, 0x0, 0x3c2500, {}, [@IFLA_LINKINFO={0x14, 0x12, 0x0, 0x1, @bridge={{0xb}, {0x4}}}, @IFLA_MASTER={0x8, 0xa, r3}]}, 0x3c}}, 0x0) 04:13:06 executing program 3: r0 = socket$netlink(0x10, 0x3, 0x0) sendmsg$netlink(r0, &(0x7f00000047c0)={0x0, 0x0, &(0x7f0000000000)=[{&(0x7f0000000100)=ANY=[@ANYBLOB="18000000560001060000009800db730007"], 0x18}], 0x1}, 0x0) 04:13:06 executing program 0: r0 = socket$netlink(0x10, 0x3, 0x0) sendmsg$netlink(r0, &(0x7f00000047c0)={0x0, 0x0, &(0x7f0000000000)=[{&(0x7f0000000100)=ANY=[@ANYBLOB="18000000560001060000009500db730007"], 0x18}], 0x1}, 0x0) [ 2859.278830][ T3808] netlink: 'syz-executor.1': attribute type 1 has an invalid length. [ 2859.384353][ T3808] bond1315: entered promiscuous mode [ 2859.390075][ T3808] 8021q: adding VLAN 0 to HW filter on device bond1315 [ 2859.435432][ T3809] bond1315: (slave bridge1261): making interface the new active one [ 2859.453703][ T3809] bridge1261: entered promiscuous mode [ 2859.494292][ T3809] bond1315: (slave bridge1261): Enslaving as an active interface with an up link [ 2859.509677][ T3811] netlink: 'syz-executor.5': attribute type 1 has an invalid length. 04:13:07 executing program 1: r0 = socket$netlink(0x10, 0x3, 0x0) r1 = socket$netlink(0x10, 0x3, 0x0) r2 = socket(0x10, 0x803, 0x0) sendmsg$NL80211_CMD_CRIT_PROTOCOL_START(r2, &(0x7f0000000580)={0x0, 0x0, &(0x7f0000000540)={0x0, 0x1c}}, 0x0) getsockname$packet(r2, &(0x7f0000000600)={0x11, 0x0, 0x0, 0x1, 0x0, 0x6, @broadcast}, &(0x7f0000000100)=0x14) sendmsg$nl_route(r1, &(0x7f0000000040)={0x0, 0x0, &(0x7f0000000000)={&(0x7f00000000c0)=ANY=[@ANYBLOB="3c00000010008506eb9afc4cd8d06e754a0081c5", @ANYRES32=r3, @ANYBLOB="2377f292252155b21c0012000c000100626f6e64000000000c0002000800010001"], 0x3c}}, 0x0) sendmsg$nl_route(r0, &(0x7f0000000240)={0x0, 0x0, &(0x7f0000000680)={&(0x7f0000000400)=@newlink={0x3c, 0x10, 0xffffff1f, 0x0, 0x29672500, {}, [@IFLA_LINKINFO={0x14, 0x12, 0x0, 0x1, @bridge={{0xb}, {0x4}}}, @IFLA_MASTER={0x8, 0xa, r3}]}, 0x3c}}, 0x0) [ 2859.618809][ T3811] bond1307: entered promiscuous mode [ 2859.628784][ T3811] 8021q: adding VLAN 0 to HW filter on device bond1307 [ 2859.754131][ T3813] bond1307: (slave bridge1270): making interface the new active one [ 2859.779616][ T3813] bridge1270: entered promiscuous mode 04:13:07 executing program 4: r0 = socket$inet6(0xa, 0x2, 0x0) ioctl$sock_SIOCETHTOOL(r0, 0x89f0, &(0x7f0000000000)={'ip6_vti0\x00', &(0x7f0000000040)=@ethtool_stats={0x1d, 0x6, [0xc50b, 0x3471, 0x3ff, 0x6, 0x101, 0x2000004e]}}) socket$inet6(0xa, 0x2, 0x0) (async) ioctl$sock_SIOCETHTOOL(r0, 0x89f0, &(0x7f0000000000)={'ip6_vti0\x00', &(0x7f0000000040)=@ethtool_stats={0x1d, 0x6, [0xc50b, 0x3471, 0x3ff, 0x6, 0x101, 0x2000004e]}}) (async) 04:13:07 executing program 5: r0 = socket$netlink(0x10, 0x3, 0x0) r1 = socket$netlink(0x10, 0x3, 0x0) r2 = socket(0x10, 0x803, 0x0) sendmsg$NL80211_CMD_CRIT_PROTOCOL_START(r2, &(0x7f0000000580)={0x0, 0x0, &(0x7f0000000540)={0x0, 0x1c}}, 0x0) getsockname$packet(r2, &(0x7f0000000600)={0x11, 0x0, 0x0, 0x1, 0x0, 0x6, @broadcast}, &(0x7f0000000100)=0x14) sendmsg$nl_route(r1, &(0x7f0000000040)={0x0, 0x0, &(0x7f0000000000)={&(0x7f00000000c0)=ANY=[@ANYBLOB="3c00000010008506eb9afc4cd8d06e754a0081c5", @ANYRES32=r3, @ANYBLOB="2377f292252155b21c0012000c000100626f6e64000000000c0002000800010001"], 0x3c}}, 0x0) sendmsg$nl_route(r0, &(0x7f0000000240)={0x0, 0x0, &(0x7f0000000680)={&(0x7f0000000400)=@newlink={0x3c, 0x10, 0xffffff1f, 0x0, 0x5d120000, {}, [@IFLA_LINKINFO={0x14, 0x12, 0x0, 0x1, @bridge={{0xb}, {0x4}}}, @IFLA_MASTER={0x8, 0xa, r3}]}, 0x3c}}, 0x0) [ 2859.800910][ T3813] bond1307: (slave bridge1270): Enslaving as an active interface with an up link [ 2859.837365][ T3822] netlink: 'syz-executor.2': attribute type 1 has an invalid length. 04:13:07 executing program 0: r0 = socket$netlink(0x10, 0x3, 0x0) sendmsg$netlink(r0, &(0x7f00000047c0)={0x0, 0x0, &(0x7f0000000000)=[{&(0x7f0000000100)=ANY=[@ANYBLOB="18000000560001060000009500db730007"], 0x18}], 0x1}, 0x0) [ 2859.904748][ T3822] bond1300: entered promiscuous mode [ 2859.910489][ T3822] 8021q: adding VLAN 0 to HW filter on device bond1300 04:13:07 executing program 2: r0 = socket$netlink(0x10, 0x3, 0x0) r1 = socket$netlink(0x10, 0x3, 0x0) r2 = socket(0x10, 0x803, 0x0) sendmsg$NL80211_CMD_CRIT_PROTOCOL_START(r2, &(0x7f0000000580)={0x0, 0x0, &(0x7f0000000540)={0x0, 0x1c}}, 0x0) getsockname$packet(r2, &(0x7f0000000600)={0x11, 0x0, 0x0, 0x1, 0x0, 0x6, @broadcast}, &(0x7f0000000100)=0x14) sendmsg$nl_route(r1, &(0x7f0000000040)={0x0, 0x0, &(0x7f0000000000)={&(0x7f00000000c0)=ANY=[@ANYBLOB="3c00000010008506eb9afc4cd8d06e754a0081c5", @ANYRES32=r3, @ANYBLOB="2377f292252155b21c0012000c000100626f6e64000000000c0002000800010001"], 0x3c}}, 0x0) sendmsg$nl_route(r0, &(0x7f0000000240)={0x0, 0x0, &(0x7f0000000680)={&(0x7f0000000400)=@newlink={0x3c, 0x10, 0xffffff1f, 0x0, 0x400300, {}, [@IFLA_LINKINFO={0x14, 0x12, 0x0, 0x1, @bridge={{0xb}, {0x4}}}, @IFLA_MASTER={0x8, 0xa, r3}]}, 0x3c}}, 0x0) 04:13:07 executing program 3: r0 = socket$netlink(0x10, 0x3, 0x0) sendmsg$netlink(r0, &(0x7f00000047c0)={0x0, 0x0, &(0x7f0000000000)=[{&(0x7f0000000100)=ANY=[@ANYBLOB="18000000560001060000009900db730007"], 0x18}], 0x1}, 0x0) [ 2859.963700][ T3824] bond1300: (slave bridge1221): making interface the new active one [ 2859.973615][ T3824] bridge1221: entered promiscuous mode [ 2859.985544][ T3824] bond1300: (slave bridge1221): Enslaving as an active interface with an up link [ 2859.998375][ T3828] netlink: 'syz-executor.1': attribute type 1 has an invalid length. [ 2860.088128][ T3828] bond1316: entered promiscuous mode [ 2860.113084][ T3828] 8021q: adding VLAN 0 to HW filter on device bond1316 [ 2860.343477][ T3829] bond1316: (slave bridge1262): making interface the new active one [ 2860.361969][ T3829] bridge1262: entered promiscuous mode [ 2860.389360][T32481] wlan1: No active IBSS STAs - trying to scan for other IBSS networks with same SSID (merge) [ 2860.391846][ T3829] bond1316: (slave bridge1262): Enslaving as an active interface with an up link 04:13:07 executing program 0: sendmsg$netlink(0xffffffffffffffff, &(0x7f00000047c0)={0x0, 0x0, &(0x7f0000000000)=[{&(0x7f0000000100)=ANY=[@ANYBLOB="18000000560001060000009500db730007"], 0x18}], 0x1}, 0x0) 04:13:07 executing program 1: r0 = socket$netlink(0x10, 0x3, 0x0) r1 = socket$netlink(0x10, 0x3, 0x0) r2 = socket(0x10, 0x803, 0x0) sendmsg$NL80211_CMD_CRIT_PROTOCOL_START(r2, &(0x7f0000000580)={0x0, 0x0, &(0x7f0000000540)={0x0, 0x1c}}, 0x0) getsockname$packet(r2, &(0x7f0000000600)={0x11, 0x0, 0x0, 0x1, 0x0, 0x6, @broadcast}, &(0x7f0000000100)=0x14) sendmsg$nl_route(r1, &(0x7f0000000040)={0x0, 0x0, &(0x7f0000000000)={&(0x7f00000000c0)=ANY=[@ANYBLOB="3c00000010008506eb9afc4cd8d06e754a0081c5", @ANYRES32=r3, @ANYBLOB="2377f292252155b21c0012000c000100626f6e64000000000c0002000800010001"], 0x3c}}, 0x0) sendmsg$nl_route(r0, &(0x7f0000000240)={0x0, 0x0, &(0x7f0000000680)={&(0x7f0000000400)=@newlink={0x3c, 0x10, 0xffffff1f, 0x0, 0x2a672500, {}, [@IFLA_LINKINFO={0x14, 0x12, 0x0, 0x1, @bridge={{0xb}, {0x4}}}, @IFLA_MASTER={0x8, 0xa, r3}]}, 0x3c}}, 0x0) [ 2860.464035][ T3834] netlink: 'syz-executor.5': attribute type 1 has an invalid length. 04:13:08 executing program 0: sendmsg$netlink(0xffffffffffffffff, &(0x7f00000047c0)={0x0, 0x0, &(0x7f0000000000)=[{&(0x7f0000000100)=ANY=[@ANYBLOB="18000000560001060000009500db730007"], 0x18}], 0x1}, 0x0) 04:13:08 executing program 0: sendmsg$netlink(0xffffffffffffffff, &(0x7f00000047c0)={0x0, 0x0, &(0x7f0000000000)=[{&(0x7f0000000100)=ANY=[@ANYBLOB="18000000560001060000009500db730007"], 0x18}], 0x1}, 0x0) 04:13:08 executing program 0: socket$netlink(0x10, 0x3, 0x0) sendmsg$netlink(0xffffffffffffffff, &(0x7f00000047c0)={0x0, 0x0, &(0x7f0000000000)=[{&(0x7f0000000100)=ANY=[@ANYBLOB="18000000560001060000009500db730007"], 0x18}], 0x1}, 0x0) [ 2860.629763][ T3834] bond1308: entered promiscuous mode [ 2860.661758][ T3834] 8021q: adding VLAN 0 to HW filter on device bond1308 04:13:08 executing program 0: socket$netlink(0x10, 0x3, 0x0) sendmsg$netlink(0xffffffffffffffff, &(0x7f00000047c0)={0x0, 0x0, &(0x7f0000000000)=[{&(0x7f0000000100)=ANY=[@ANYBLOB="18000000560001060000009500db730007"], 0x18}], 0x1}, 0x0) 04:13:08 executing program 3: r0 = socket$netlink(0x10, 0x3, 0x0) sendmsg$netlink(r0, &(0x7f00000047c0)={0x0, 0x0, &(0x7f0000000000)=[{&(0x7f0000000100)=ANY=[@ANYBLOB="18000000560001060000009a00db730007"], 0x18}], 0x1}, 0x0) [ 2860.682162][ T3841] netlink: 'syz-executor.2': attribute type 1 has an invalid length. [ 2860.812573][ T3841] bond1301: entered promiscuous mode [ 2860.822504][ T3841] 8021q: adding VLAN 0 to HW filter on device bond1301 04:13:08 executing program 4: socket$inet6(0xa, 0x2, 0x0) (async) r0 = socket$inet6(0xa, 0x2, 0x0) ioctl$sock_SIOCETHTOOL(r0, 0x89f0, &(0x7f0000000000)={'ip6_vti0\x00', &(0x7f0000000040)=@ethtool_stats={0x1d, 0x6, [0xc50b, 0x3471, 0x3ff, 0x6, 0x101, 0x2000004e]}}) 04:13:08 executing program 0: socket$netlink(0x10, 0x3, 0x0) sendmsg$netlink(0xffffffffffffffff, &(0x7f00000047c0)={0x0, 0x0, &(0x7f0000000000)=[{&(0x7f0000000100)=ANY=[@ANYBLOB="18000000560001060000009500db730007"], 0x18}], 0x1}, 0x0) 04:13:08 executing program 5: r0 = socket$netlink(0x10, 0x3, 0x0) r1 = socket$netlink(0x10, 0x3, 0x0) r2 = socket(0x10, 0x803, 0x0) sendmsg$NL80211_CMD_CRIT_PROTOCOL_START(r2, &(0x7f0000000580)={0x0, 0x0, &(0x7f0000000540)={0x0, 0x1c}}, 0x0) getsockname$packet(r2, &(0x7f0000000600)={0x11, 0x0, 0x0, 0x1, 0x0, 0x6, @broadcast}, &(0x7f0000000100)=0x14) sendmsg$nl_route(r1, &(0x7f0000000040)={0x0, 0x0, &(0x7f0000000000)={&(0x7f00000000c0)=ANY=[@ANYBLOB="3c00000010008506eb9afc4cd8d06e754a0081c5", @ANYRES32=r3, @ANYBLOB="2377f292252155b21c0012000c000100626f6e64000000000c0002000800010001"], 0x3c}}, 0x0) sendmsg$nl_route(r0, &(0x7f0000000240)={0x0, 0x0, &(0x7f0000000680)={&(0x7f0000000400)=@newlink={0x3c, 0x10, 0xffffff1f, 0x0, 0x62020000, {}, [@IFLA_LINKINFO={0x14, 0x12, 0x0, 0x1, @bridge={{0xb}, {0x4}}}, @IFLA_MASTER={0x8, 0xa, r3}]}, 0x3c}}, 0x0) 04:13:08 executing program 2: r0 = socket$netlink(0x10, 0x3, 0x0) r1 = socket$netlink(0x10, 0x3, 0x0) r2 = socket(0x10, 0x803, 0x0) sendmsg$NL80211_CMD_CRIT_PROTOCOL_START(r2, &(0x7f0000000580)={0x0, 0x0, &(0x7f0000000540)={0x0, 0x1c}}, 0x0) getsockname$packet(r2, &(0x7f0000000600)={0x11, 0x0, 0x0, 0x1, 0x0, 0x6, @broadcast}, &(0x7f0000000100)=0x14) sendmsg$nl_route(r1, &(0x7f0000000040)={0x0, 0x0, &(0x7f0000000000)={&(0x7f00000000c0)=ANY=[@ANYBLOB="3c00000010008506eb9afc4cd8d06e754a0081c5", @ANYRES32=r3, @ANYBLOB="2377f292252155b21c0012000c000100626f6e64000000000c0002000800010001"], 0x3c}}, 0x0) sendmsg$nl_route(r0, &(0x7f0000000240)={0x0, 0x0, &(0x7f0000000680)={&(0x7f0000000400)=@newlink={0x3c, 0x10, 0xffffff1f, 0x0, 0xf0ffff, {}, [@IFLA_LINKINFO={0x14, 0x12, 0x0, 0x1, @bridge={{0xb}, {0x4}}}, @IFLA_MASTER={0x8, 0xa, r3}]}, 0x3c}}, 0x0) [ 2860.935723][ T3842] bond1308: (slave bridge1271): making interface the new active one [ 2860.944087][ T3842] bridge1271: entered promiscuous mode [ 2860.954821][ T3842] bond1308: (slave bridge1271): Enslaving as an active interface with an up link [ 2860.969243][ T3852] netlink: 'syz-executor.1': attribute type 1 has an invalid length. 04:13:08 executing program 0: r0 = socket$netlink(0x10, 0x3, 0x0) sendmsg$netlink(r0, 0x0, 0x0) [ 2861.056219][ T3852] bond1317: entered promiscuous mode [ 2861.062252][ T3852] 8021q: adding VLAN 0 to HW filter on device bond1317 04:13:08 executing program 0: r0 = socket$netlink(0x10, 0x3, 0x0) sendmsg$netlink(r0, 0x0, 0x0) [ 2861.145233][ T3853] bond1317: (slave bridge1263): making interface the new active one [ 2861.218198][ T3853] bridge1263: entered promiscuous mode [ 2861.269458][ T3853] bond1317: (slave bridge1263): Enslaving as an active interface with an up link 04:13:08 executing program 1: r0 = socket$netlink(0x10, 0x3, 0x0) r1 = socket$netlink(0x10, 0x3, 0x0) r2 = socket(0x10, 0x803, 0x0) sendmsg$NL80211_CMD_CRIT_PROTOCOL_START(r2, &(0x7f0000000580)={0x0, 0x0, &(0x7f0000000540)={0x0, 0x1c}}, 0x0) getsockname$packet(r2, &(0x7f0000000600)={0x11, 0x0, 0x0, 0x1, 0x0, 0x6, @broadcast}, &(0x7f0000000100)=0x14) sendmsg$nl_route(r1, &(0x7f0000000040)={0x0, 0x0, &(0x7f0000000000)={&(0x7f00000000c0)=ANY=[@ANYBLOB="3c00000010008506eb9afc4cd8d06e754a0081c5", @ANYRES32=r3, @ANYBLOB="2377f292252155b21c0012000c000100626f6e64000000000c0002000800010001"], 0x3c}}, 0x0) sendmsg$nl_route(r0, &(0x7f0000000240)={0x0, 0x0, &(0x7f0000000680)={&(0x7f0000000400)=@newlink={0x3c, 0x10, 0xffffff1f, 0x0, 0x2b672500, {}, [@IFLA_LINKINFO={0x14, 0x12, 0x0, 0x1, @bridge={{0xb}, {0x4}}}, @IFLA_MASTER={0x8, 0xa, r3}]}, 0x3c}}, 0x0) 04:13:08 executing program 0: r0 = socket$netlink(0x10, 0x3, 0x0) sendmsg$netlink(r0, 0x0, 0x0) 04:13:08 executing program 4: r0 = socket$inet6(0xa, 0x2, 0x0) ioctl$sock_SIOCETHTOOL(r0, 0x89f0, &(0x7f0000000000)={'bridge0\x00', &(0x7f0000000200)=@ethtool_ringparam={0x10}}) r1 = syz_init_net_socket$nl_rdma(0x10, 0x3, 0x10) sendmsg$RDMA_NLDEV_CMD_RES_QP_GET(r1, &(0x7f0000000100)={&(0x7f0000000040)={0x10, 0x0, 0x0, 0x20000}, 0xc, &(0x7f00000000c0)={&(0x7f0000000140)=ANY=[@ANYBLOB="400000000a142214fc07ce982fa941e4b3b0040025bd7000fdbadf250800150004000000084da7000500000008001500017ed5a4920015000500000008000300010038d51406b102000800150005000000311a8d4a031744f2438c0a5824376364c41a0020f4db1b00000000"], 0x40}, 0x1, 0x0, 0x0, 0x20040010}, 0x24000000) 04:13:08 executing program 3: r0 = socket$netlink(0x10, 0x3, 0x0) sendmsg$netlink(r0, &(0x7f00000047c0)={0x0, 0x0, &(0x7f0000000000)=[{&(0x7f0000000100)=ANY=[@ANYBLOB="18000000560001060000009b00db730007"], 0x18}], 0x1}, 0x0) 04:13:08 executing program 0: r0 = socket$netlink(0x10, 0x3, 0x0) sendmsg$netlink(r0, &(0x7f00000047c0)={0x0, 0x0, 0x0}, 0x0) [ 2861.465604][ T3873] bond1302: entered promiscuous mode [ 2861.474108][ T3873] 8021q: adding VLAN 0 to HW filter on device bond1302 [ 2861.517850][ T3877] bond1302: (slave bridge1222): making interface the new active one [ 2861.526871][ T3877] bridge1222: entered promiscuous mode 04:13:09 executing program 0: r0 = socket$netlink(0x10, 0x3, 0x0) sendmsg$netlink(r0, &(0x7f00000047c0)={0x0, 0x0, 0x0}, 0x0) [ 2861.560195][ T3877] bond1302: (slave bridge1222): Enslaving as an active interface with an up link 04:13:09 executing program 2: r0 = socket$netlink(0x10, 0x3, 0x0) r1 = socket$netlink(0x10, 0x3, 0x0) r2 = socket(0x10, 0x803, 0x0) sendmsg$NL80211_CMD_CRIT_PROTOCOL_START(r2, &(0x7f0000000580)={0x0, 0x0, &(0x7f0000000540)={0x0, 0x1c}}, 0x0) getsockname$packet(r2, &(0x7f0000000600)={0x11, 0x0, 0x0, 0x1, 0x0, 0x6, @broadcast}, &(0x7f0000000100)=0x14) sendmsg$nl_route(r1, &(0x7f0000000040)={0x0, 0x0, &(0x7f0000000000)={&(0x7f00000000c0)=ANY=[@ANYBLOB="3c00000010008506eb9afc4cd8d06e754a0081c5", @ANYRES32=r3, @ANYBLOB="2377f292252155b21c0012000c000100626f6e64000000000c0002000800010001"], 0x3c}}, 0x0) sendmsg$nl_route(r0, &(0x7f0000000240)={0x0, 0x0, &(0x7f0000000680)={&(0x7f0000000400)=@newlink={0x3c, 0x10, 0xffffff1f, 0x0, 0x1000000, {}, [@IFLA_LINKINFO={0x14, 0x12, 0x0, 0x1, @bridge={{0xb}, {0x4}}}, @IFLA_MASTER={0x8, 0xa, r3}]}, 0x3c}}, 0x0) 04:13:09 executing program 0: r0 = socket$netlink(0x10, 0x3, 0x0) sendmsg$netlink(r0, &(0x7f00000047c0)={0x0, 0x0, 0x0}, 0x0) [ 2861.768698][ T3876] bond1309: entered promiscuous mode [ 2861.787904][ T3876] 8021q: adding VLAN 0 to HW filter on device bond1309 [ 2861.829839][ T3883] bond1309: (slave bridge1272): making interface the new active one [ 2861.838448][ T3883] bridge1272: entered promiscuous mode [ 2861.848198][ T3883] bond1309: (slave bridge1272): Enslaving as an active interface with an up link 04:13:09 executing program 5: r0 = socket$netlink(0x10, 0x3, 0x0) r1 = socket$netlink(0x10, 0x3, 0x0) r2 = socket(0x10, 0x803, 0x0) sendmsg$NL80211_CMD_CRIT_PROTOCOL_START(r2, &(0x7f0000000580)={0x0, 0x0, &(0x7f0000000540)={0x0, 0x1c}}, 0x0) getsockname$packet(r2, &(0x7f0000000600)={0x11, 0x0, 0x0, 0x1, 0x0, 0x6, @broadcast}, &(0x7f0000000100)=0x14) sendmsg$nl_route(r1, &(0x7f0000000040)={0x0, 0x0, &(0x7f0000000000)={&(0x7f00000000c0)=ANY=[@ANYBLOB="3c00000010008506eb9afc4cd8d06e754a0081c5", @ANYRES32=r3, @ANYBLOB="2377f292252155b21c0012000c000100626f6e64000000000c0002000800010001"], 0x3c}}, 0x0) sendmsg$nl_route(r0, &(0x7f0000000240)={0x0, 0x0, &(0x7f0000000680)={&(0x7f0000000400)=@newlink={0x3c, 0x10, 0xffffff1f, 0x0, 0x62030000, {}, [@IFLA_LINKINFO={0x14, 0x12, 0x0, 0x1, @bridge={{0xb}, {0x4}}}, @IFLA_MASTER={0x8, 0xa, r3}]}, 0x3c}}, 0x0) 04:13:09 executing program 0: r0 = socket$netlink(0x10, 0x3, 0x0) sendmsg$netlink(r0, &(0x7f00000047c0), 0x0) 04:13:09 executing program 0: r0 = socket$netlink(0x10, 0x3, 0x0) sendmsg$netlink(r0, &(0x7f00000047c0), 0x0) [ 2861.967018][ T3892] bond1318: entered promiscuous mode [ 2861.987900][ T3892] 8021q: adding VLAN 0 to HW filter on device bond1318 04:13:09 executing program 0: r0 = socket$netlink(0x10, 0x3, 0x0) sendmsg$netlink(r0, &(0x7f00000047c0), 0x0) 04:13:09 executing program 0: r0 = socket$netlink(0x10, 0x3, 0x0) sendmsg$netlink(r0, &(0x7f00000047c0)={0x0, 0x0, &(0x7f0000000000)=[{0x0, 0x18}], 0x1}, 0x0) 04:13:09 executing program 4: r0 = socket$inet6(0xa, 0x2, 0x0) ioctl$sock_SIOCETHTOOL(r0, 0x89f0, &(0x7f0000000000)={'bridge0\x00', &(0x7f0000000200)=@ethtool_ringparam={0x10}}) (async) ioctl$sock_SIOCETHTOOL(r0, 0x89f0, &(0x7f0000000000)={'bridge0\x00', &(0x7f0000000200)=@ethtool_ringparam={0x10}}) r1 = syz_init_net_socket$nl_rdma(0x10, 0x3, 0x10) sendmsg$RDMA_NLDEV_CMD_RES_QP_GET(r1, &(0x7f0000000100)={&(0x7f0000000040)={0x10, 0x0, 0x0, 0x20000}, 0xc, &(0x7f00000000c0)={&(0x7f0000000140)=ANY=[@ANYBLOB="400000000a142214fc07ce982fa941e4b3b0040025bd7000fdbadf250800150004000000084da7000500000008001500017ed5a4920015000500000008000300010038d51406b102000800150005000000311a8d4a031744f2438c0a5824376364c41a0020f4db1b00000000"], 0x40}, 0x1, 0x0, 0x0, 0x20040010}, 0x24000000) [ 2862.283406][ T3894] bond1318: (slave bridge1264): making interface the new active one [ 2862.295233][ T3894] bridge1264: entered promiscuous mode [ 2862.317931][ T3894] bond1318: (slave bridge1264): Enslaving as an active interface with an up link 04:13:09 executing program 1: r0 = socket$netlink(0x10, 0x3, 0x0) r1 = socket$netlink(0x10, 0x3, 0x0) r2 = socket(0x10, 0x803, 0x0) sendmsg$NL80211_CMD_CRIT_PROTOCOL_START(r2, &(0x7f0000000580)={0x0, 0x0, &(0x7f0000000540)={0x0, 0x1c}}, 0x0) getsockname$packet(r2, &(0x7f0000000600)={0x11, 0x0, 0x0, 0x1, 0x0, 0x6, @broadcast}, &(0x7f0000000100)=0x14) sendmsg$nl_route(r1, &(0x7f0000000040)={0x0, 0x0, &(0x7f0000000000)={&(0x7f00000000c0)=ANY=[@ANYBLOB="3c00000010008506eb9afc4cd8d06e754a0081c5", @ANYRES32=r3, @ANYBLOB="2377f292252155b21c0012000c000100626f6e64000000000c0002000800010001"], 0x3c}}, 0x0) sendmsg$nl_route(r0, &(0x7f0000000240)={0x0, 0x0, &(0x7f0000000680)={&(0x7f0000000400)=@newlink={0x3c, 0x10, 0xffffff1f, 0x0, 0x2c672500, {}, [@IFLA_LINKINFO={0x14, 0x12, 0x0, 0x1, @bridge={{0xb}, {0x4}}}, @IFLA_MASTER={0x8, 0xa, r3}]}, 0x3c}}, 0x0) [ 2862.483409][ T3907] bond1303: entered promiscuous mode [ 2862.498514][ T3907] 8021q: adding VLAN 0 to HW filter on device bond1303 [ 2862.592491][ T3908] bond1303: (slave bridge1223): making interface the new active one [ 2862.604108][ T3908] bridge1223: entered promiscuous mode [ 2862.626687][ T3908] bond1303: (slave bridge1223): Enslaving as an active interface with an up link 04:13:10 executing program 3: r0 = socket$netlink(0x10, 0x3, 0x0) sendmsg$netlink(r0, &(0x7f00000047c0)={0x0, 0x0, &(0x7f0000000000)=[{&(0x7f0000000100)=ANY=[@ANYBLOB="18000000560001060000009c00db730007"], 0x18}], 0x1}, 0x0) 04:13:10 executing program 0: r0 = socket$netlink(0x10, 0x3, 0x0) sendmsg$netlink(r0, &(0x7f00000047c0)={0x0, 0x0, &(0x7f0000000000)=[{0x0, 0x18}], 0x1}, 0x0) 04:13:10 executing program 0: r0 = socket$netlink(0x10, 0x3, 0x0) sendmsg$netlink(r0, &(0x7f00000047c0)={0x0, 0x0, &(0x7f0000000000)=[{0x0, 0x18}], 0x1}, 0x0) [ 2862.734700][ T3911] bond1310: entered promiscuous mode [ 2862.740632][ T3911] 8021q: adding VLAN 0 to HW filter on device bond1310 04:13:10 executing program 2: r0 = socket$netlink(0x10, 0x3, 0x0) r1 = socket$netlink(0x10, 0x3, 0x0) r2 = socket(0x10, 0x803, 0x0) sendmsg$NL80211_CMD_CRIT_PROTOCOL_START(r2, &(0x7f0000000580)={0x0, 0x0, &(0x7f0000000540)={0x0, 0x1c}}, 0x0) getsockname$packet(r2, &(0x7f0000000600)={0x11, 0x0, 0x0, 0x1, 0x0, 0x6, @broadcast}, &(0x7f0000000100)=0x14) sendmsg$nl_route(r1, &(0x7f0000000040)={0x0, 0x0, &(0x7f0000000000)={&(0x7f00000000c0)=ANY=[@ANYBLOB="3c00000010008506eb9afc4cd8d06e754a0081c5", @ANYRES32=r3, @ANYBLOB="2377f292252155b21c0012000c000100626f6e64000000000c0002000800010001"], 0x3c}}, 0x0) sendmsg$nl_route(r0, &(0x7f0000000240)={0x0, 0x0, &(0x7f0000000680)={&(0x7f0000000400)=@newlink={0x3c, 0x10, 0xffffff1f, 0x0, 0x13c2500, {}, [@IFLA_LINKINFO={0x14, 0x12, 0x0, 0x1, @bridge={{0xb}, {0x4}}}, @IFLA_MASTER={0x8, 0xa, r3}]}, 0x3c}}, 0x0) 04:13:10 executing program 4: r0 = socket$inet6(0xa, 0x2, 0x0) ioctl$sock_SIOCETHTOOL(r0, 0x89f0, &(0x7f0000000000)={'bridge0\x00', &(0x7f0000000200)=@ethtool_ringparam={0x10}}) r1 = syz_init_net_socket$nl_rdma(0x10, 0x3, 0x10) sendmsg$RDMA_NLDEV_CMD_RES_QP_GET(r1, &(0x7f0000000100)={&(0x7f0000000040)={0x10, 0x0, 0x0, 0x20000}, 0xc, &(0x7f00000000c0)={&(0x7f0000000140)=ANY=[@ANYBLOB="400000000a142214fc07ce982fa941e4b3b0040025bd7000fdbadf250800150004000000084da7000500000008001500017ed5a4920015000500000008000300010038d51406b102000800150005000000311a8d4a031744f2438c0a5824376364c41a0020f4db1b00000000"], 0x40}, 0x1, 0x0, 0x0, 0x20040010}, 0x24000000) 04:13:10 executing program 5: r0 = socket$netlink(0x10, 0x3, 0x0) r1 = socket$netlink(0x10, 0x3, 0x0) r2 = socket(0x10, 0x803, 0x0) sendmsg$NL80211_CMD_CRIT_PROTOCOL_START(r2, &(0x7f0000000580)={0x0, 0x0, &(0x7f0000000540)={0x0, 0x1c}}, 0x0) getsockname$packet(r2, &(0x7f0000000600)={0x11, 0x0, 0x0, 0x1, 0x0, 0x6, @broadcast}, &(0x7f0000000100)=0x14) sendmsg$nl_route(r1, &(0x7f0000000040)={0x0, 0x0, &(0x7f0000000000)={&(0x7f00000000c0)=ANY=[@ANYBLOB="3c00000010008506eb9afc4cd8d06e754a0081c5", @ANYRES32=r3, @ANYBLOB="2377f292252155b21c0012000c000100626f6e64000000000c0002000800010001"], 0x3c}}, 0x0) sendmsg$nl_route(r0, &(0x7f0000000240)={0x0, 0x0, &(0x7f0000000680)={&(0x7f0000000400)=@newlink={0x3c, 0x10, 0xffffff1f, 0x0, 0x68000000, {}, [@IFLA_LINKINFO={0x14, 0x12, 0x0, 0x1, @bridge={{0xb}, {0x4}}}, @IFLA_MASTER={0x8, 0xa, r3}]}, 0x3c}}, 0x0) 04:13:10 executing program 0: r0 = socket$netlink(0x10, 0x3, 0x0) sendmsg$netlink(r0, &(0x7f00000047c0)={0x0, 0x0, &(0x7f0000000000)=[{&(0x7f0000000100)=ANY=[], 0x18}], 0x1}, 0x0) [ 2862.885127][ T3926] bond1319: entered promiscuous mode [ 2862.890839][ T3926] 8021q: adding VLAN 0 to HW filter on device bond1319 [ 2862.926911][ T3929] bond1319: (slave bridge1265): making interface the new active one [ 2862.935110][ T3929] bridge1265: entered promiscuous mode [ 2862.951440][ T3929] bond1319: (slave bridge1265): Enslaving as an active interface with an up link 04:13:10 executing program 0: r0 = socket$netlink(0x10, 0x3, 0x0) sendmsg$netlink(r0, &(0x7f00000047c0)={0x0, 0x0, &(0x7f0000000000)=[{&(0x7f0000000100)=ANY=[], 0x18}], 0x1}, 0x0) 04:13:10 executing program 3: r0 = socket$netlink(0x10, 0x3, 0x0) sendmsg$netlink(r0, &(0x7f00000047c0)={0x0, 0x0, &(0x7f0000000000)=[{&(0x7f0000000100)=ANY=[@ANYBLOB="18000000560001060000009d00db730007"], 0x18}], 0x1}, 0x0) [ 2863.020802][ T3935] bond1304: entered promiscuous mode [ 2863.043744][ T3935] 8021q: adding VLAN 0 to HW filter on device bond1304 04:13:10 executing program 4: r0 = socket$inet6(0xa, 0x2, 0x0) r1 = socket(0x28, 0x5, 0x0) getsockopt$sock_buf(r1, 0x1, 0x1c, 0x0, &(0x7f0000000100)) ioctl$sock_SIOCETHTOOL(r0, 0x89f0, &(0x7f0000000000)={'bridge0\x00', &(0x7f0000000040)=ANY=[@ANYBLOB="100000000000256a00000000000020000000000000000000000009c6921b9c92fd7814950000002000"]}) 04:13:10 executing program 0: r0 = socket$netlink(0x10, 0x3, 0x0) sendmsg$netlink(r0, &(0x7f00000047c0)={0x0, 0x0, &(0x7f0000000000)=[{&(0x7f0000000100)=ANY=[], 0x18}], 0x1}, 0x0) 04:13:10 executing program 1: r0 = socket$netlink(0x10, 0x3, 0x0) r1 = socket$netlink(0x10, 0x3, 0x0) r2 = socket(0x10, 0x803, 0x0) sendmsg$NL80211_CMD_CRIT_PROTOCOL_START(r2, &(0x7f0000000580)={0x0, 0x0, &(0x7f0000000540)={0x0, 0x1c}}, 0x0) getsockname$packet(r2, &(0x7f0000000600)={0x11, 0x0, 0x0, 0x1, 0x0, 0x6, @broadcast}, &(0x7f0000000100)=0x14) sendmsg$nl_route(r1, &(0x7f0000000040)={0x0, 0x0, &(0x7f0000000000)={&(0x7f00000000c0)=ANY=[@ANYBLOB="3c00000010008506eb9afc4cd8d06e754a0081c5", @ANYRES32=r3, @ANYBLOB="2377f292252155b21c0012000c000100626f6e64000000000c0002000800010001"], 0x3c}}, 0x0) sendmsg$nl_route(r0, &(0x7f0000000240)={0x0, 0x0, &(0x7f0000000680)={&(0x7f0000000400)=@newlink={0x3c, 0x10, 0xffffff1f, 0x0, 0x2d672500, {}, [@IFLA_LINKINFO={0x14, 0x12, 0x0, 0x1, @bridge={{0xb}, {0x4}}}, @IFLA_MASTER={0x8, 0xa, r3}]}, 0x3c}}, 0x0) [ 2863.178581][ T3943] bond1304: (slave bridge1224): making interface the new active one [ 2863.186956][ T3943] bridge1224: entered promiscuous mode [ 2863.203546][ T3943] bond1304: (slave bridge1224): Enslaving as an active interface with an up link 04:13:10 executing program 0: r0 = socket$netlink(0x10, 0x3, 0x0) sendmsg$netlink(r0, &(0x7f00000047c0)={0x0, 0x0, &(0x7f0000000000)=[{&(0x7f0000000100)=ANY=[@ANYBLOB], 0x18}], 0x1}, 0x0) 04:13:10 executing program 0: r0 = socket$netlink(0x10, 0x3, 0x0) sendmsg$netlink(r0, &(0x7f00000047c0)={0x0, 0x0, &(0x7f0000000000)=[{&(0x7f0000000100)=ANY=[@ANYBLOB], 0x18}], 0x1}, 0x0) 04:13:10 executing program 2: r0 = socket$netlink(0x10, 0x3, 0x0) r1 = socket$netlink(0x10, 0x3, 0x0) r2 = socket(0x10, 0x803, 0x0) sendmsg$NL80211_CMD_CRIT_PROTOCOL_START(r2, &(0x7f0000000580)={0x0, 0x0, &(0x7f0000000540)={0x0, 0x1c}}, 0x0) getsockname$packet(r2, &(0x7f0000000600)={0x11, 0x0, 0x0, 0x1, 0x0, 0x6, @broadcast}, &(0x7f0000000100)=0x14) sendmsg$nl_route(r1, &(0x7f0000000040)={0x0, 0x0, &(0x7f0000000000)={&(0x7f00000000c0)=ANY=[@ANYBLOB="3c00000010008506eb9afc4cd8d06e754a0081c5", @ANYRES32=r3, @ANYBLOB="2377f292252155b21c0012000c000100626f6e64000000000c0002000800010001"], 0x3c}}, 0x0) sendmsg$nl_route(r0, &(0x7f0000000240)={0x0, 0x0, &(0x7f0000000680)={&(0x7f0000000400)=@newlink={0x3c, 0x10, 0xffffff1f, 0x0, 0x2000000, {}, [@IFLA_LINKINFO={0x14, 0x12, 0x0, 0x1, @bridge={{0xb}, {0x4}}}, @IFLA_MASTER={0x8, 0xa, r3}]}, 0x3c}}, 0x0) [ 2863.311794][ T3950] bond1311: entered promiscuous mode [ 2863.317400][ T3950] 8021q: adding VLAN 0 to HW filter on device bond1311 04:13:10 executing program 0: r0 = socket$netlink(0x10, 0x3, 0x0) sendmsg$netlink(r0, &(0x7f00000047c0)={0x0, 0x0, &(0x7f0000000000)=[{&(0x7f0000000100)=ANY=[@ANYBLOB], 0x18}], 0x1}, 0x0) 04:13:10 executing program 3: r0 = socket$netlink(0x10, 0x3, 0x0) sendmsg$netlink(r0, &(0x7f00000047c0)={0x0, 0x0, &(0x7f0000000000)=[{&(0x7f0000000100)=ANY=[@ANYBLOB="18000000560001060000009e00db730007"], 0x18}], 0x1}, 0x0) [ 2863.361326][ T3951] bond1311: (slave bridge1273): making interface the new active one [ 2863.369500][ T3951] bridge1273: entered promiscuous mode [ 2863.385567][ T3951] bond1311: (slave bridge1273): Enslaving as an active interface with an up link 04:13:10 executing program 5: r0 = socket$netlink(0x10, 0x3, 0x0) r1 = socket$netlink(0x10, 0x3, 0x0) r2 = socket(0x10, 0x803, 0x0) sendmsg$NL80211_CMD_CRIT_PROTOCOL_START(r2, &(0x7f0000000580)={0x0, 0x0, &(0x7f0000000540)={0x0, 0x1c}}, 0x0) getsockname$packet(r2, &(0x7f0000000600)={0x11, 0x0, 0x0, 0x1, 0x0, 0x6, @broadcast}, &(0x7f0000000100)=0x14) sendmsg$nl_route(r1, &(0x7f0000000040)={0x0, 0x0, &(0x7f0000000000)={&(0x7f00000000c0)=ANY=[@ANYBLOB="3c00000010008506eb9afc4cd8d06e754a0081c5", @ANYRES32=r3, @ANYBLOB="2377f292252155b21c0012000c000100626f6e64000000000c0002000800010001"], 0x3c}}, 0x0) sendmsg$nl_route(r0, &(0x7f0000000240)={0x0, 0x0, &(0x7f0000000680)={&(0x7f0000000400)=@newlink={0x3c, 0x10, 0xffffff1f, 0x0, 0x6a020000, {}, [@IFLA_LINKINFO={0x14, 0x12, 0x0, 0x1, @bridge={{0xb}, {0x4}}}, @IFLA_MASTER={0x8, 0xa, r3}]}, 0x3c}}, 0x0) 04:13:10 executing program 4: r0 = socket$inet6(0xa, 0x2, 0x0) (async) r1 = socket(0x28, 0x5, 0x0) getsockopt$sock_buf(r1, 0x1, 0x1c, 0x0, &(0x7f0000000100)) ioctl$sock_SIOCETHTOOL(r0, 0x89f0, &(0x7f0000000000)={'bridge0\x00', &(0x7f0000000040)=ANY=[@ANYBLOB="100000000000256a00000000000020000000000000000000000009c6921b9c92fd7814950000002000"]}) 04:13:10 executing program 0: r0 = socket$netlink(0x10, 0x3, 0x0) sendmsg$netlink(r0, &(0x7f00000047c0)={0x0, 0x0, &(0x7f0000000000)=[{&(0x7f0000000100)=ANY=[@ANYBLOB="180000005600010600"], 0x18}], 0x1}, 0x0) 04:13:11 executing program 0: r0 = socket$netlink(0x10, 0x3, 0x0) sendmsg$netlink(r0, &(0x7f00000047c0)={0x0, 0x0, &(0x7f0000000000)=[{&(0x7f0000000100)=ANY=[@ANYBLOB="180000005600010600"], 0x18}], 0x1}, 0x0) [ 2863.597448][ T3963] bond1320: entered promiscuous mode 04:13:11 executing program 4: r0 = socket$inet6(0xa, 0x2, 0x0) (async) r1 = socket(0x28, 0x5, 0x0) getsockopt$sock_buf(r1, 0x1, 0x1c, 0x0, &(0x7f0000000100)) (async) ioctl$sock_SIOCETHTOOL(r0, 0x89f0, &(0x7f0000000000)={'bridge0\x00', &(0x7f0000000040)=ANY=[@ANYBLOB="100000000000256a00000000000020000000000000000000000009c6921b9c92fd7814950000002000"]}) 04:13:11 executing program 0: r0 = socket$netlink(0x10, 0x3, 0x0) sendmsg$netlink(r0, &(0x7f00000047c0)={0x0, 0x0, &(0x7f0000000000)=[{&(0x7f0000000100)=ANY=[@ANYBLOB="180000005600010600"], 0x18}], 0x1}, 0x0) [ 2863.631287][ T3963] 8021q: adding VLAN 0 to HW filter on device bond1320 [ 2863.773863][ T3967] bond1320: (slave bridge1266): making interface the new active one [ 2863.800291][ T3967] bridge1266: entered promiscuous mode 04:13:11 executing program 1: r0 = socket$netlink(0x10, 0x3, 0x0) r1 = socket$netlink(0x10, 0x3, 0x0) r2 = socket(0x10, 0x803, 0x0) sendmsg$NL80211_CMD_CRIT_PROTOCOL_START(r2, &(0x7f0000000580)={0x0, 0x0, &(0x7f0000000540)={0x0, 0x1c}}, 0x0) getsockname$packet(r2, &(0x7f0000000600)={0x11, 0x0, 0x0, 0x1, 0x0, 0x6, @broadcast}, &(0x7f0000000100)=0x14) sendmsg$nl_route(r1, &(0x7f0000000040)={0x0, 0x0, &(0x7f0000000000)={&(0x7f00000000c0)=ANY=[@ANYBLOB="3c00000010008506eb9afc4cd8d06e754a0081c5", @ANYRES32=r3, @ANYBLOB="2377f292252155b21c0012000c000100626f6e64000000000c0002000800010001"], 0x3c}}, 0x0) sendmsg$nl_route(r0, &(0x7f0000000240)={0x0, 0x0, &(0x7f0000000680)={&(0x7f0000000400)=@newlink={0x3c, 0x10, 0xffffff1f, 0x0, 0x2e672500, {}, [@IFLA_LINKINFO={0x14, 0x12, 0x0, 0x1, @bridge={{0xb}, {0x4}}}, @IFLA_MASTER={0x8, 0xa, r3}]}, 0x3c}}, 0x0) 04:13:11 executing program 0: r0 = socket$netlink(0x10, 0x3, 0x0) sendmsg$netlink(r0, &(0x7f00000047c0)={0x0, 0x0, &(0x7f0000000000)=[{&(0x7f0000000100)=ANY=[@ANYBLOB="18000000560001060000009500"], 0x18}], 0x1}, 0x0) [ 2863.820242][ T3967] bond1320: (slave bridge1266): Enslaving as an active interface with an up link [ 2863.859263][ T3974] validate_nla: 9 callbacks suppressed [ 2863.859284][ T3974] netlink: 'syz-executor.2': attribute type 1 has an invalid length. 04:13:11 executing program 0: r0 = socket$netlink(0x10, 0x3, 0x0) sendmsg$netlink(r0, &(0x7f00000047c0)={0x0, 0x0, &(0x7f0000000000)=[{&(0x7f0000000100)=ANY=[@ANYBLOB="18000000560001060000009500"], 0x18}], 0x1}, 0x0) [ 2863.967494][ T3974] bond1305: entered promiscuous mode [ 2863.974255][ T3974] 8021q: adding VLAN 0 to HW filter on device bond1305 [ 2864.016699][ T3977] bond1305: (slave bridge1225): making interface the new active one [ 2864.024986][ T3977] bridge1225: entered promiscuous mode [ 2864.036195][ T3977] bond1305: (slave bridge1225): Enslaving as an active interface with an up link 04:13:11 executing program 2: r0 = socket$netlink(0x10, 0x3, 0x0) r1 = socket$netlink(0x10, 0x3, 0x0) r2 = socket(0x10, 0x803, 0x0) sendmsg$NL80211_CMD_CRIT_PROTOCOL_START(r2, &(0x7f0000000580)={0x0, 0x0, &(0x7f0000000540)={0x0, 0x1c}}, 0x0) getsockname$packet(r2, &(0x7f0000000600)={0x11, 0x0, 0x0, 0x1, 0x0, 0x6, @broadcast}, &(0x7f0000000100)=0x14) sendmsg$nl_route(r1, &(0x7f0000000040)={0x0, 0x0, &(0x7f0000000000)={&(0x7f00000000c0)=ANY=[@ANYBLOB="3c00000010008506eb9afc4cd8d06e754a0081c5", @ANYRES32=r3, @ANYBLOB="2377f292252155b21c0012000c000100626f6e64000000000c0002000800010001"], 0x3c}}, 0x0) sendmsg$nl_route(r0, &(0x7f0000000240)={0x0, 0x0, &(0x7f0000000680)={&(0x7f0000000400)=@newlink={0x3c, 0x10, 0xffffff1f, 0x0, 0x2030000, {}, [@IFLA_LINKINFO={0x14, 0x12, 0x0, 0x1, @bridge={{0xb}, {0x4}}}, @IFLA_MASTER={0x8, 0xa, r3}]}, 0x3c}}, 0x0) 04:13:11 executing program 0: r0 = socket$netlink(0x10, 0x3, 0x0) sendmsg$netlink(r0, &(0x7f00000047c0)={0x0, 0x0, &(0x7f0000000000)=[{&(0x7f0000000100)=ANY=[@ANYBLOB="18000000560001060000009500"], 0x18}], 0x1}, 0x0) [ 2864.069205][ T3982] netlink: 'syz-executor.5': attribute type 1 has an invalid length. 04:13:11 executing program 3: r0 = socket$netlink(0x10, 0x3, 0x0) sendmsg$netlink(r0, &(0x7f00000047c0)={0x0, 0x0, &(0x7f0000000000)=[{&(0x7f0000000100)=ANY=[@ANYBLOB="18000000560001060000009f00db730007"], 0x18}], 0x1}, 0x0) [ 2864.174264][ T3982] bond1312: entered promiscuous mode [ 2864.201343][ T3982] 8021q: adding VLAN 0 to HW filter on device bond1312 04:13:11 executing program 5: r0 = socket$netlink(0x10, 0x3, 0x0) r1 = socket$netlink(0x10, 0x3, 0x0) r2 = socket(0x10, 0x803, 0x0) sendmsg$NL80211_CMD_CRIT_PROTOCOL_START(r2, &(0x7f0000000580)={0x0, 0x0, &(0x7f0000000540)={0x0, 0x1c}}, 0x0) getsockname$packet(r2, &(0x7f0000000600)={0x11, 0x0, 0x0, 0x1, 0x0, 0x6, @broadcast}, &(0x7f0000000100)=0x14) sendmsg$nl_route(r1, &(0x7f0000000040)={0x0, 0x0, &(0x7f0000000000)={&(0x7f00000000c0)=ANY=[@ANYBLOB="3c00000010008506eb9afc4cd8d06e754a0081c5", @ANYRES32=r3, @ANYBLOB="2377f292252155b21c0012000c000100626f6e64000000000c0002000800010001"], 0x3c}}, 0x0) sendmsg$nl_route(r0, &(0x7f0000000240)={0x0, 0x0, &(0x7f0000000680)={&(0x7f0000000400)=@newlink={0x3c, 0x10, 0xffffff1f, 0x0, 0x6a030000, {}, [@IFLA_LINKINFO={0x14, 0x12, 0x0, 0x1, @bridge={{0xb}, {0x4}}}, @IFLA_MASTER={0x8, 0xa, r3}]}, 0x3c}}, 0x0) 04:13:11 executing program 0: r0 = socket$netlink(0x10, 0x3, 0x0) sendmsg$netlink(r0, &(0x7f00000047c0)={0x0, 0x0, &(0x7f0000000000)=[{&(0x7f0000000100)=ANY=[@ANYBLOB="18000000560001060000009500db73"], 0x18}], 0x1}, 0x0) 04:13:11 executing program 4: r0 = socket$inet6(0xa, 0x2, 0x0) ioctl$sock_SIOCGIFVLAN_GET_VLAN_REALDEV_NAME_CMD(r0, 0x8982, &(0x7f0000000040)={0x8, 'ip_vti0\x00', {'ip_vti0\x00'}, 0x8001}) r1 = socket$inet6(0xa, 0x80002, 0x88) bind$inet6(r1, &(0x7f0000000100)={0xa, 0x4e20, 0x2, @initdev={0xfe, 0x88, '\x00', 0x1, 0x0}}, 0x1c) setsockopt$inet6_udp_int(r1, 0x11, 0xb, &(0x7f0000000080), 0x4) ioctl$ifreq_SIOCGIFINDEX_team(r0, 0x8933, &(0x7f00000000c0)={'team0\x00', 0x0}) ioctl$sock_SIOCETHTOOL(r1, 0x89f0, &(0x7f0000000000)={'bridge0\x00', &(0x7f00000001c0)=ANY=[@ANYBLOB="1000000061ee24eb2fb78228909bab660388080000001000000077000b000000000000000000010000a7cf55f3ae0c11f341ca13b955d099c858155548dbbc4c9d2ce7f6711eff0d3c883758654c1a8a5fc322f8fa4d34eae2d92920e147bb0892e5b27e43f1e578c66472b473432b9a5411e8d5b54aa090ad8caecf116951b688219d25ff081502714492a90b743dad8b1a02c911b2f3a3c1e760ad7ca8152171492adfe74d3a054c74809dc0", @ANYRESDEC=r2, @ANYRESHEX=r1]}) [ 2864.247691][ T3992] bond1312: (slave bridge1274): making interface the new active one [ 2864.257299][ T3992] bridge1274: entered promiscuous mode [ 2864.273423][ T3992] bond1312: (slave bridge1274): Enslaving as an active interface with an up link [ 2864.283799][ T4004] netlink: 'syz-executor.1': attribute type 1 has an invalid length. 04:13:11 executing program 0: r0 = socket$netlink(0x10, 0x3, 0x0) sendmsg$netlink(r0, &(0x7f00000047c0)={0x0, 0x0, &(0x7f0000000000)=[{&(0x7f0000000100)=ANY=[@ANYBLOB="18000000560001060000009500db73"], 0x18}], 0x1}, 0x0) [ 2864.431236][ T4004] bond1321: entered promiscuous mode [ 2864.449281][ T4004] 8021q: adding VLAN 0 to HW filter on device bond1321 04:13:11 executing program 0: r0 = socket$netlink(0x10, 0x3, 0x0) sendmsg$netlink(r0, &(0x7f00000047c0)={0x0, 0x0, &(0x7f0000000000)=[{&(0x7f0000000100)=ANY=[@ANYBLOB="18000000560001060000009500db73"], 0x18}], 0x1}, 0x0) 04:13:12 executing program 0: r0 = socket$netlink(0x10, 0x3, 0x0) sendmsg$netlink(r0, &(0x7f00000047c0)={0x0, 0x0, &(0x7f0000000000)=[{&(0x7f0000000100)=ANY=[@ANYBLOB="18000000560001060000009500db7300"], 0x18}], 0x1}, 0x0) [ 2864.635814][ T4008] bond1321: (slave bridge1267): making interface the new active one [ 2864.665756][ T4008] bridge1267: entered promiscuous mode 04:13:12 executing program 1: r0 = socket$netlink(0x10, 0x3, 0x0) r1 = socket$netlink(0x10, 0x3, 0x0) r2 = socket(0x10, 0x803, 0x0) sendmsg$NL80211_CMD_CRIT_PROTOCOL_START(r2, &(0x7f0000000580)={0x0, 0x0, &(0x7f0000000540)={0x0, 0x1c}}, 0x0) getsockname$packet(r2, &(0x7f0000000600)={0x11, 0x0, 0x0, 0x1, 0x0, 0x6, @broadcast}, &(0x7f0000000100)=0x14) sendmsg$nl_route(r1, &(0x7f0000000040)={0x0, 0x0, &(0x7f0000000000)={&(0x7f00000000c0)=ANY=[@ANYBLOB="3c00000010008506eb9afc4cd8d06e754a0081c5", @ANYRES32=r3, @ANYBLOB="2377f292252155b21c0012000c000100626f6e64000000000c0002000800010001"], 0x3c}}, 0x0) sendmsg$nl_route(r0, &(0x7f0000000240)={0x0, 0x0, &(0x7f0000000680)={&(0x7f0000000400)=@newlink={0x3c, 0x10, 0xffffff1f, 0x0, 0x2f672500, {}, [@IFLA_LINKINFO={0x14, 0x12, 0x0, 0x1, @bridge={{0xb}, {0x4}}}, @IFLA_MASTER={0x8, 0xa, r3}]}, 0x3c}}, 0x0) 04:13:12 executing program 0: r0 = socket$netlink(0x10, 0x3, 0x0) sendmsg$netlink(r0, &(0x7f00000047c0)={0x0, 0x0, &(0x7f0000000000)=[{&(0x7f0000000100)=ANY=[@ANYBLOB="18000000560001060000009500db7300"], 0x18}], 0x1}, 0x0) [ 2864.720835][ T4008] bond1321: (slave bridge1267): Enslaving as an active interface with an up link [ 2864.746454][ T4015] netlink: 'syz-executor.2': attribute type 1 has an invalid length. [ 2864.808884][ T4015] bond1306: entered promiscuous mode [ 2864.814641][ T4015] 8021q: adding VLAN 0 to HW filter on device bond1306 04:13:12 executing program 0: r0 = socket$netlink(0x10, 0x3, 0x0) sendmsg$netlink(r0, &(0x7f00000047c0)={0x0, 0x0, &(0x7f0000000000)=[{&(0x7f0000000100)=ANY=[@ANYBLOB="18000000560001060000009500db7300"], 0x18}], 0x1}, 0x0) [ 2865.025971][ T4017] bond1306: (slave bridge1226): making interface the new active one [ 2865.060419][ T4017] bridge1226: entered promiscuous mode [ 2865.079056][ T4017] bond1306: (slave bridge1226): Enslaving as an active interface with an up link [ 2865.114229][ T4025] netlink: 'syz-executor.5': attribute type 1 has an invalid length. 04:13:12 executing program 0: r0 = socket$netlink(0x10, 0x3, 0x0) sendmsg$netlink(r0, &(0x7f00000047c0)={0x0, 0x0, &(0x7f0000000000)=[{&(0x7f0000000100)=ANY=[@ANYBLOB="180000005600010600"], 0x18}], 0x1}, 0x0) 04:13:12 executing program 3: r0 = socket$netlink(0x10, 0x3, 0x0) sendmsg$netlink(r0, &(0x7f00000047c0)={0x0, 0x0, &(0x7f0000000000)=[{&(0x7f0000000100)=ANY=[@ANYBLOB="1800000056000106000000a000db730007"], 0x18}], 0x1}, 0x0) 04:13:12 executing program 2: r0 = socket$netlink(0x10, 0x3, 0x0) r1 = socket$netlink(0x10, 0x3, 0x0) r2 = socket(0x10, 0x803, 0x0) sendmsg$NL80211_CMD_CRIT_PROTOCOL_START(r2, &(0x7f0000000580)={0x0, 0x0, &(0x7f0000000540)={0x0, 0x1c}}, 0x0) getsockname$packet(r2, &(0x7f0000000600)={0x11, 0x0, 0x0, 0x1, 0x0, 0x6, @broadcast}, &(0x7f0000000100)=0x14) sendmsg$nl_route(r1, &(0x7f0000000040)={0x0, 0x0, &(0x7f0000000000)={&(0x7f00000000c0)=ANY=[@ANYBLOB="3c00000010008506eb9afc4cd8d06e754a0081c5", @ANYRES32=r3, @ANYBLOB="2377f292252155b21c0012000c000100626f6e64000000000c0002000800010001"], 0x3c}}, 0x0) sendmsg$nl_route(r0, &(0x7f0000000240)={0x0, 0x0, &(0x7f0000000680)={&(0x7f0000000400)=@newlink={0x3c, 0x10, 0xffffff1f, 0x0, 0x23c2500, {}, [@IFLA_LINKINFO={0x14, 0x12, 0x0, 0x1, @bridge={{0xb}, {0x4}}}, @IFLA_MASTER={0x8, 0xa, r3}]}, 0x3c}}, 0x0) [ 2865.226235][ T4025] bond1313 (unregistering): Released all slaves 04:13:12 executing program 5: r0 = socket$netlink(0x10, 0x3, 0x0) r1 = socket$netlink(0x10, 0x3, 0x0) r2 = socket(0x10, 0x803, 0x0) sendmsg$NL80211_CMD_CRIT_PROTOCOL_START(r2, &(0x7f0000000580)={0x0, 0x0, &(0x7f0000000540)={0x0, 0x1c}}, 0x0) getsockname$packet(r2, &(0x7f0000000600)={0x11, 0x0, 0x0, 0x1, 0x0, 0x6, @broadcast}, &(0x7f0000000100)=0x14) sendmsg$nl_route(r1, &(0x7f0000000040)={0x0, 0x0, &(0x7f0000000000)={&(0x7f00000000c0)=ANY=[@ANYBLOB="3c00000010008506eb9afc4cd8d06e754a0081c5", @ANYRES32=r3, @ANYBLOB="2377f292252155b21c0012000c000100626f6e64000000000c0002000800010001"], 0x3c}}, 0x0) sendmsg$nl_route(r0, &(0x7f0000000240)={0x0, 0x0, &(0x7f0000000680)={&(0x7f0000000400)=@newlink={0x3c, 0x10, 0xffffff1f, 0x0, 0x6c000000, {}, [@IFLA_LINKINFO={0x14, 0x12, 0x0, 0x1, @bridge={{0xb}, {0x4}}}, @IFLA_MASTER={0x8, 0xa, r3}]}, 0x3c}}, 0x0) 04:13:12 executing program 4: r0 = socket$inet6(0xa, 0x2, 0x0) ioctl$sock_SIOCGIFVLAN_GET_VLAN_REALDEV_NAME_CMD(r0, 0x8982, &(0x7f0000000040)={0x8, 'ip_vti0\x00', {'ip_vti0\x00'}, 0x8001}) (async) r1 = socket$inet6(0xa, 0x80002, 0x88) bind$inet6(r1, &(0x7f0000000100)={0xa, 0x4e20, 0x2, @initdev={0xfe, 0x88, '\x00', 0x1, 0x0}}, 0x1c) (async) setsockopt$inet6_udp_int(r1, 0x11, 0xb, &(0x7f0000000080), 0x4) (async) ioctl$ifreq_SIOCGIFINDEX_team(r0, 0x8933, &(0x7f00000000c0)={'team0\x00', 0x0}) ioctl$sock_SIOCETHTOOL(r1, 0x89f0, &(0x7f0000000000)={'bridge0\x00', &(0x7f00000001c0)=ANY=[@ANYBLOB="1000000061ee24eb2fb78228909bab660388080000001000000077000b000000000000000000010000a7cf55f3ae0c11f341ca13b955d099c858155548dbbc4c9d2ce7f6711eff0d3c883758654c1a8a5fc322f8fa4d34eae2d92920e147bb0892e5b27e43f1e578c66472b473432b9a5411e8d5b54aa090ad8caecf116951b688219d25ff081502714492a90b743dad8b1a02c911b2f3a3c1e760ad7ca8152171492adfe74d3a054c74809dc0", @ANYRESDEC=r2, @ANYRESHEX=r1]}) 04:13:12 executing program 0: r0 = socket$netlink(0x10, 0x3, 0x0) sendmsg$netlink(r0, &(0x7f00000047c0)={0x0, 0x0, &(0x7f0000000000)=[{&(0x7f0000000100)=ANY=[@ANYBLOB="180000005600010600"], 0x18}], 0x1}, 0x0) [ 2865.417529][ T4041] netlink: 'syz-executor.1': attribute type 1 has an invalid length. 04:13:13 executing program 0: r0 = socket$netlink(0x10, 0x3, 0x0) sendmsg$netlink(r0, &(0x7f00000047c0)={0x0, 0x0, &(0x7f0000000000)=[{&(0x7f0000000100)=ANY=[@ANYBLOB="180000005600010600"], 0x18}], 0x1}, 0x0) [ 2865.556582][ T4041] bond1322: entered promiscuous mode [ 2865.572152][ T4041] 8021q: adding VLAN 0 to HW filter on device bond1322 04:13:13 executing program 0: sendmsg$netlink(0xffffffffffffffff, &(0x7f00000047c0)={0x0, 0x0, &(0x7f0000000000)=[{&(0x7f0000000100)=ANY=[@ANYBLOB="180000005600010600"], 0x18}], 0x1}, 0x0) [ 2865.656589][ T4043] bond1322: (slave bridge1268): making interface the new active one [ 2865.673935][ T4043] bridge1268: entered promiscuous mode [ 2865.720186][ T4043] bond1322: (slave bridge1268): Enslaving as an active interface with an up link 04:13:13 executing program 0: sendmsg$netlink(0xffffffffffffffff, &(0x7f00000047c0)={0x0, 0x0, &(0x7f0000000000)=[{&(0x7f0000000100)=ANY=[@ANYBLOB="180000005600010600"], 0x18}], 0x1}, 0x0) [ 2865.766908][ T4054] netlink: 'syz-executor.2': attribute type 1 has an invalid length. 04:13:13 executing program 1: r0 = socket$netlink(0x10, 0x3, 0x0) r1 = socket$netlink(0x10, 0x3, 0x0) r2 = socket(0x10, 0x803, 0x0) sendmsg$NL80211_CMD_CRIT_PROTOCOL_START(r2, &(0x7f0000000580)={0x0, 0x0, &(0x7f0000000540)={0x0, 0x1c}}, 0x0) getsockname$packet(r2, &(0x7f0000000600)={0x11, 0x0, 0x0, 0x1, 0x0, 0x6, @broadcast}, &(0x7f0000000100)=0x14) sendmsg$nl_route(r1, &(0x7f0000000040)={0x0, 0x0, &(0x7f0000000000)={&(0x7f00000000c0)=ANY=[@ANYBLOB="3c00000010008506eb9afc4cd8d06e754a0081c5", @ANYRES32=r3, @ANYBLOB="2377f292252155b21c0012000c000100626f6e64000000000c0002000800010001"], 0x3c}}, 0x0) sendmsg$nl_route(r0, &(0x7f0000000240)={0x0, 0x0, &(0x7f0000000680)={&(0x7f0000000400)=@newlink={0x3c, 0x10, 0xffffff1f, 0x0, 0x30672500, {}, [@IFLA_LINKINFO={0x14, 0x12, 0x0, 0x1, @bridge={{0xb}, {0x4}}}, @IFLA_MASTER={0x8, 0xa, r3}]}, 0x3c}}, 0x0) 04:13:13 executing program 3: r0 = socket$netlink(0x10, 0x3, 0x0) sendmsg$netlink(r0, &(0x7f00000047c0)={0x0, 0x0, &(0x7f0000000000)=[{&(0x7f0000000100)=ANY=[@ANYBLOB="1800000056000106000000a100db730007"], 0x18}], 0x1}, 0x0) [ 2865.897266][ T4054] bond1307: entered promiscuous mode [ 2865.907355][ T4054] 8021q: adding VLAN 0 to HW filter on device bond1307 04:13:13 executing program 0: sendmsg$netlink(0xffffffffffffffff, &(0x7f00000047c0)={0x0, 0x0, &(0x7f0000000000)=[{&(0x7f0000000100)=ANY=[@ANYBLOB="180000005600010600"], 0x18}], 0x1}, 0x0) [ 2865.955111][ T4055] bond1307: (slave bridge1227): making interface the new active one [ 2865.972059][ T4055] bridge1227: entered promiscuous mode [ 2865.981276][ T4055] bond1307: (slave bridge1227): Enslaving as an active interface with an up link 04:13:13 executing program 4: r0 = socket$inet6(0xa, 0x2, 0x0) ioctl$sock_SIOCGIFVLAN_GET_VLAN_REALDEV_NAME_CMD(r0, 0x8982, &(0x7f0000000040)={0x8, 'ip_vti0\x00', {'ip_vti0\x00'}, 0x8001}) r1 = socket$inet6(0xa, 0x80002, 0x88) bind$inet6(r1, &(0x7f0000000100)={0xa, 0x4e20, 0x2, @initdev={0xfe, 0x88, '\x00', 0x1, 0x0}}, 0x1c) setsockopt$inet6_udp_int(r1, 0x11, 0xb, &(0x7f0000000080), 0x4) (async) ioctl$ifreq_SIOCGIFINDEX_team(r0, 0x8933, &(0x7f00000000c0)={'team0\x00', 0x0}) ioctl$sock_SIOCETHTOOL(r1, 0x89f0, &(0x7f0000000000)={'bridge0\x00', &(0x7f00000001c0)=ANY=[@ANYBLOB="1000000061ee24eb2fb78228909bab660388080000001000000077000b000000000000000000010000a7cf55f3ae0c11f341ca13b955d099c858155548dbbc4c9d2ce7f6711eff0d3c883758654c1a8a5fc322f8fa4d34eae2d92920e147bb0892e5b27e43f1e578c66472b473432b9a5411e8d5b54aa090ad8caecf116951b688219d25ff081502714492a90b743dad8b1a02c911b2f3a3c1e760ad7ca8152171492adfe74d3a054c74809dc0", @ANYRESDEC=r2, @ANYRESHEX=r1]}) [ 2866.001566][ T4064] netlink: 'syz-executor.5': attribute type 1 has an invalid length. 04:13:13 executing program 0: socket$netlink(0x10, 0x3, 0x0) sendmsg$netlink(0xffffffffffffffff, &(0x7f00000047c0)={0x0, 0x0, &(0x7f0000000000)=[{&(0x7f0000000100)=ANY=[@ANYBLOB="180000005600010600"], 0x18}], 0x1}, 0x0) [ 2866.032876][ T4064] bond1313: entered promiscuous mode [ 2866.042074][ T4064] 8021q: adding VLAN 0 to HW filter on device bond1313 04:13:13 executing program 2: r0 = socket$netlink(0x10, 0x3, 0x0) r1 = socket$netlink(0x10, 0x3, 0x0) r2 = socket(0x10, 0x803, 0x0) sendmsg$NL80211_CMD_CRIT_PROTOCOL_START(r2, &(0x7f0000000580)={0x0, 0x0, &(0x7f0000000540)={0x0, 0x1c}}, 0x0) getsockname$packet(r2, &(0x7f0000000600)={0x11, 0x0, 0x0, 0x1, 0x0, 0x6, @broadcast}, &(0x7f0000000100)=0x14) sendmsg$nl_route(r1, &(0x7f0000000040)={0x0, 0x0, &(0x7f0000000000)={&(0x7f00000000c0)=ANY=[@ANYBLOB="3c00000010008506eb9afc4cd8d06e754a0081c5", @ANYRES32=r3, @ANYBLOB="2377f292252155b21c0012000c000100626f6e64000000000c0002000800010001"], 0x3c}}, 0x0) sendmsg$nl_route(r0, &(0x7f0000000240)={0x0, 0x0, &(0x7f0000000680)={&(0x7f0000000400)=@newlink={0x3c, 0x10, 0xffffff1f, 0x0, 0x3000000, {}, [@IFLA_LINKINFO={0x14, 0x12, 0x0, 0x1, @bridge={{0xb}, {0x4}}}, @IFLA_MASTER={0x8, 0xa, r3}]}, 0x3c}}, 0x0) [ 2866.283172][ T4067] bond1313: (slave bridge1275): making interface the new active one [ 2866.302595][ T4067] bridge1275: entered promiscuous mode 04:13:13 executing program 5: r0 = socket$netlink(0x10, 0x3, 0x0) r1 = socket$netlink(0x10, 0x3, 0x0) r2 = socket(0x10, 0x803, 0x0) sendmsg$NL80211_CMD_CRIT_PROTOCOL_START(r2, &(0x7f0000000580)={0x0, 0x0, &(0x7f0000000540)={0x0, 0x1c}}, 0x0) getsockname$packet(r2, &(0x7f0000000600)={0x11, 0x0, 0x0, 0x1, 0x0, 0x6, @broadcast}, &(0x7f0000000100)=0x14) sendmsg$nl_route(r1, &(0x7f0000000040)={0x0, 0x0, &(0x7f0000000000)={&(0x7f00000000c0)=ANY=[@ANYBLOB="3c00000010008506eb9afc4cd8d06e754a0081c5", @ANYRES32=r3, @ANYBLOB="2377f292252155b21c0012000c000100626f6e64000000000c0002000800010001"], 0x3c}}, 0x0) sendmsg$nl_route(r0, &(0x7f0000000240)={0x0, 0x0, &(0x7f0000000680)={&(0x7f0000000400)=@newlink={0x3c, 0x10, 0xffffff1f, 0x0, 0x74000000, {}, [@IFLA_LINKINFO={0x14, 0x12, 0x0, 0x1, @bridge={{0xb}, {0x4}}}, @IFLA_MASTER={0x8, 0xa, r3}]}, 0x3c}}, 0x0) 04:13:13 executing program 0: socket$netlink(0x10, 0x3, 0x0) sendmsg$netlink(0xffffffffffffffff, &(0x7f00000047c0)={0x0, 0x0, &(0x7f0000000000)=[{&(0x7f0000000100)=ANY=[@ANYBLOB="180000005600010600"], 0x18}], 0x1}, 0x0) [ 2866.328059][ T4067] bond1313: (slave bridge1275): Enslaving as an active interface with an up link [ 2866.357804][ T4077] netlink: 'syz-executor.1': attribute type 1 has an invalid length. 04:13:13 executing program 0: socket$netlink(0x10, 0x3, 0x0) sendmsg$netlink(0xffffffffffffffff, &(0x7f00000047c0)={0x0, 0x0, &(0x7f0000000000)=[{&(0x7f0000000100)=ANY=[@ANYBLOB="180000005600010600"], 0x18}], 0x1}, 0x0) [ 2866.468986][ T4077] bond1323: entered promiscuous mode [ 2866.487255][ T4077] 8021q: adding VLAN 0 to HW filter on device bond1323 04:13:14 executing program 0: r0 = socket$netlink(0x10, 0x3, 0x0) sendmsg$netlink(r0, 0x0, 0x0) 04:13:14 executing program 4: r0 = socket$kcm(0x29, 0x5, 0x0) ioctl$sock_SIOCOUTQNSD(r0, 0x894b, &(0x7f0000000040)) r1 = socket$inet6(0xa, 0x2, 0x0) ioctl$sock_SIOCETHTOOL(r1, 0x89f0, &(0x7f0000000080)={'bridge0\x00', &(0x7f0000000200)=ANY=[@ANYBLOB="110000000000000000000000080000000000f3ff0b000000000000000020000000000000"]}) [ 2866.555309][ T4079] bond1323: (slave bridge1269): making interface the new active one [ 2866.563545][ T4079] bridge1269: entered promiscuous mode [ 2866.574251][ T4079] bond1323: (slave bridge1269): Enslaving as an active interface with an up link 04:13:14 executing program 1: r0 = socket$netlink(0x10, 0x3, 0x0) r1 = socket$netlink(0x10, 0x3, 0x0) r2 = socket(0x10, 0x803, 0x0) sendmsg$NL80211_CMD_CRIT_PROTOCOL_START(r2, &(0x7f0000000580)={0x0, 0x0, &(0x7f0000000540)={0x0, 0x1c}}, 0x0) getsockname$packet(r2, &(0x7f0000000600)={0x11, 0x0, 0x0, 0x1, 0x0, 0x6, @broadcast}, &(0x7f0000000100)=0x14) sendmsg$nl_route(r1, &(0x7f0000000040)={0x0, 0x0, &(0x7f0000000000)={&(0x7f00000000c0)=ANY=[@ANYBLOB="3c00000010008506eb9afc4cd8d06e754a0081c5", @ANYRES32=r3, @ANYBLOB="2377f292252155b21c0012000c000100626f6e64000000000c0002000800010001"], 0x3c}}, 0x0) sendmsg$nl_route(r0, &(0x7f0000000240)={0x0, 0x0, &(0x7f0000000680)={&(0x7f0000000400)=@newlink={0x3c, 0x10, 0xffffff1f, 0x0, 0x31672500, {}, [@IFLA_LINKINFO={0x14, 0x12, 0x0, 0x1, @bridge={{0xb}, {0x4}}}, @IFLA_MASTER={0x8, 0xa, r3}]}, 0x3c}}, 0x0) 04:13:14 executing program 3: r0 = socket$netlink(0x10, 0x3, 0x0) sendmsg$netlink(r0, &(0x7f00000047c0)={0x0, 0x0, &(0x7f0000000000)=[{&(0x7f0000000100)=ANY=[@ANYBLOB="1800000056000106000000a200db730007"], 0x18}], 0x1}, 0x0) [ 2866.610610][ T4092] netlink: 'syz-executor.2': attribute type 1 has an invalid length. 04:13:14 executing program 0: r0 = socket$netlink(0x10, 0x3, 0x0) sendmsg$netlink(r0, 0x0, 0x0) [ 2866.793625][ T4092] bond1308: entered promiscuous mode [ 2866.810688][ T4092] 8021q: adding VLAN 0 to HW filter on device bond1308 04:13:14 executing program 0: r0 = socket$netlink(0x10, 0x3, 0x0) sendmsg$netlink(r0, 0x0, 0x0) 04:13:14 executing program 0: r0 = socket$netlink(0x10, 0x3, 0x0) sendmsg$netlink(r0, &(0x7f00000047c0)={0x0, 0x0, 0x0}, 0x0) [ 2867.013893][ T4093] bond1308: (slave bridge1228): making interface the new active one [ 2867.037673][ T4093] bridge1228: entered promiscuous mode [ 2867.049692][ T4093] bond1308: (slave bridge1228): Enslaving as an active interface with an up link [ 2867.128531][ T4098] bond1314: entered promiscuous mode [ 2867.134261][ T4098] 8021q: adding VLAN 0 to HW filter on device bond1314 04:13:14 executing program 0: r0 = socket$netlink(0x10, 0x3, 0x0) sendmsg$netlink(r0, &(0x7f00000047c0)={0x0, 0x0, 0x0}, 0x0) 04:13:14 executing program 2: r0 = socket$netlink(0x10, 0x3, 0x0) r1 = socket$netlink(0x10, 0x3, 0x0) r2 = socket(0x10, 0x803, 0x0) sendmsg$NL80211_CMD_CRIT_PROTOCOL_START(r2, &(0x7f0000000580)={0x0, 0x0, &(0x7f0000000540)={0x0, 0x1c}}, 0x0) getsockname$packet(r2, &(0x7f0000000600)={0x11, 0x0, 0x0, 0x1, 0x0, 0x6, @broadcast}, &(0x7f0000000100)=0x14) sendmsg$nl_route(r1, &(0x7f0000000040)={0x0, 0x0, &(0x7f0000000000)={&(0x7f00000000c0)=ANY=[@ANYBLOB="3c00000010008506eb9afc4cd8d06e754a0081c5", @ANYRES32=r3, @ANYBLOB="2377f292252155b21c0012000c000100626f6e64000000000c0002000800010001"], 0x3c}}, 0x0) sendmsg$nl_route(r0, &(0x7f0000000240)={0x0, 0x0, &(0x7f0000000680)={&(0x7f0000000400)=@newlink={0x3c, 0x10, 0xffffff1f, 0x0, 0x33c2500, {}, [@IFLA_LINKINFO={0x14, 0x12, 0x0, 0x1, @bridge={{0xb}, {0x4}}}, @IFLA_MASTER={0x8, 0xa, r3}]}, 0x3c}}, 0x0) [ 2867.358975][ T4101] bond1314: (slave bridge1276): making interface the new active one [ 2867.369205][ T4101] bridge1276: entered promiscuous mode [ 2867.385186][ T4101] bond1314: (slave bridge1276): Enslaving as an active interface with an up link 04:13:14 executing program 0: r0 = socket$netlink(0x10, 0x3, 0x0) sendmsg$netlink(r0, &(0x7f00000047c0)={0x0, 0x0, 0x0}, 0x0) 04:13:14 executing program 4: r0 = socket$kcm(0x29, 0x5, 0x0) ioctl$sock_SIOCOUTQNSD(r0, 0x894b, &(0x7f0000000040)) (async) ioctl$sock_SIOCOUTQNSD(r0, 0x894b, &(0x7f0000000040)) r1 = socket$inet6(0xa, 0x2, 0x0) ioctl$sock_SIOCETHTOOL(r1, 0x89f0, &(0x7f0000000080)={'bridge0\x00', &(0x7f0000000200)=ANY=[@ANYBLOB="110000000000000000000000080000000000f3ff0b000000000000000020000000000000"]}) 04:13:14 executing program 5: r0 = socket$netlink(0x10, 0x3, 0x0) r1 = socket$netlink(0x10, 0x3, 0x0) r2 = socket(0x10, 0x803, 0x0) sendmsg$NL80211_CMD_CRIT_PROTOCOL_START(r2, &(0x7f0000000580)={0x0, 0x0, &(0x7f0000000540)={0x0, 0x1c}}, 0x0) getsockname$packet(r2, &(0x7f0000000600)={0x11, 0x0, 0x0, 0x1, 0x0, 0x6, @broadcast}, &(0x7f0000000100)=0x14) sendmsg$nl_route(r1, &(0x7f0000000040)={0x0, 0x0, &(0x7f0000000000)={&(0x7f00000000c0)=ANY=[@ANYBLOB="3c00000010008506eb9afc4cd8d06e754a0081c5", @ANYRES32=r3, @ANYBLOB="2377f292252155b21c0012000c000100626f6e64000000000c0002000800010001"], 0x3c}}, 0x0) sendmsg$nl_route(r0, &(0x7f0000000240)={0x0, 0x0, &(0x7f0000000680)={&(0x7f0000000400)=@newlink={0x3c, 0x10, 0xffffff1f, 0x0, 0x7a000000, {}, [@IFLA_LINKINFO={0x14, 0x12, 0x0, 0x1, @bridge={{0xb}, {0x4}}}, @IFLA_MASTER={0x8, 0xa, r3}]}, 0x3c}}, 0x0) 04:13:15 executing program 3: r0 = socket$netlink(0x10, 0x3, 0x0) sendmsg$netlink(r0, &(0x7f00000047c0)={0x0, 0x0, &(0x7f0000000000)=[{&(0x7f0000000100)=ANY=[@ANYBLOB="1800000056000106000000a300db730007"], 0x18}], 0x1}, 0x0) [ 2867.496828][ T4110] bond1324: entered promiscuous mode [ 2867.502625][ T4110] 8021q: adding VLAN 0 to HW filter on device bond1324 04:13:15 executing program 0: r0 = socket$netlink(0x10, 0x3, 0x0) sendmsg$netlink(r0, &(0x7f00000047c0), 0x0) [ 2867.725915][ T4113] bond1324: (slave bridge1270): making interface the new active one [ 2867.745168][ T4113] bridge1270: entered promiscuous mode [ 2867.776785][ T4113] bond1324: (slave bridge1270): Enslaving as an active interface with an up link 04:13:15 executing program 1: r0 = socket$netlink(0x10, 0x3, 0x0) r1 = socket$netlink(0x10, 0x3, 0x0) r2 = socket(0x10, 0x803, 0x0) sendmsg$NL80211_CMD_CRIT_PROTOCOL_START(r2, &(0x7f0000000580)={0x0, 0x0, &(0x7f0000000540)={0x0, 0x1c}}, 0x0) getsockname$packet(r2, &(0x7f0000000600)={0x11, 0x0, 0x0, 0x1, 0x0, 0x6, @broadcast}, &(0x7f0000000100)=0x14) sendmsg$nl_route(r1, &(0x7f0000000040)={0x0, 0x0, &(0x7f0000000000)={&(0x7f00000000c0)=ANY=[@ANYBLOB="3c00000010008506eb9afc4cd8d06e754a0081c5", @ANYRES32=r3, @ANYBLOB="2377f292252155b21c0012000c000100626f6e64000000000c0002000800010001"], 0x3c}}, 0x0) sendmsg$nl_route(r0, &(0x7f0000000240)={0x0, 0x0, &(0x7f0000000680)={&(0x7f0000000400)=@newlink={0x3c, 0x10, 0xffffff1f, 0x0, 0x32672500, {}, [@IFLA_LINKINFO={0x14, 0x12, 0x0, 0x1, @bridge={{0xb}, {0x4}}}, @IFLA_MASTER={0x8, 0xa, r3}]}, 0x3c}}, 0x0) 04:13:15 executing program 0: r0 = socket$netlink(0x10, 0x3, 0x0) sendmsg$netlink(r0, &(0x7f00000047c0), 0x0) 04:13:15 executing program 0: r0 = socket$netlink(0x10, 0x3, 0x0) sendmsg$netlink(r0, &(0x7f00000047c0), 0x0) [ 2867.929810][ T4123] bond1309: entered promiscuous mode [ 2867.959454][ T4123] 8021q: adding VLAN 0 to HW filter on device bond1309 04:13:15 executing program 0: r0 = socket$netlink(0x10, 0x3, 0x0) sendmsg$netlink(r0, &(0x7f00000047c0)={0x0, 0x0, &(0x7f0000000000)=[{0x0, 0x18}], 0x1}, 0x0) 04:13:15 executing program 0: r0 = socket$netlink(0x10, 0x3, 0x0) sendmsg$netlink(r0, &(0x7f00000047c0)={0x0, 0x0, &(0x7f0000000000)=[{0x0, 0x18}], 0x1}, 0x0) [ 2868.164866][ T4126] bond1309: (slave bridge1229): making interface the new active one [ 2868.202778][ T4126] bridge1229: entered promiscuous mode 04:13:15 executing program 3: r0 = socket$netlink(0x10, 0x3, 0x0) sendmsg$netlink(r0, &(0x7f00000047c0)={0x0, 0x0, &(0x7f0000000000)=[{&(0x7f0000000100)=ANY=[@ANYBLOB="1800000056000106000000a400db730007"], 0x18}], 0x1}, 0x0) [ 2868.237984][ T4126] bond1309: (slave bridge1229): Enslaving as an active interface with an up link 04:13:15 executing program 2: r0 = socket$netlink(0x10, 0x3, 0x0) r1 = socket$netlink(0x10, 0x3, 0x0) r2 = socket(0x10, 0x803, 0x0) sendmsg$NL80211_CMD_CRIT_PROTOCOL_START(r2, &(0x7f0000000580)={0x0, 0x0, &(0x7f0000000540)={0x0, 0x1c}}, 0x0) getsockname$packet(r2, &(0x7f0000000600)={0x11, 0x0, 0x0, 0x1, 0x0, 0x6, @broadcast}, &(0x7f0000000100)=0x14) sendmsg$nl_route(r1, &(0x7f0000000040)={0x0, 0x0, &(0x7f0000000000)={&(0x7f00000000c0)=ANY=[@ANYBLOB="3c00000010008506eb9afc4cd8d06e754a0081c5", @ANYRES32=r3, @ANYBLOB="2377f292252155b21c0012000c000100626f6e64000000000c0002000800010001"], 0x3c}}, 0x0) sendmsg$nl_route(r0, &(0x7f0000000240)={0x0, 0x0, &(0x7f0000000680)={&(0x7f0000000400)=@newlink={0x3c, 0x10, 0xffffff1f, 0x0, 0x4000000, {}, [@IFLA_LINKINFO={0x14, 0x12, 0x0, 0x1, @bridge={{0xb}, {0x4}}}, @IFLA_MASTER={0x8, 0xa, r3}]}, 0x3c}}, 0x0) 04:13:15 executing program 4: r0 = socket$kcm(0x29, 0x5, 0x0) ioctl$sock_SIOCOUTQNSD(r0, 0x894b, &(0x7f0000000040)) (async) r1 = socket$inet6(0xa, 0x2, 0x0) ioctl$sock_SIOCETHTOOL(r1, 0x89f0, &(0x7f0000000080)={'bridge0\x00', &(0x7f0000000200)=ANY=[@ANYBLOB="110000000000000000000000080000000000f3ff0b000000000000000020000000000000"]}) 04:13:15 executing program 0: r0 = socket$netlink(0x10, 0x3, 0x0) sendmsg$netlink(r0, &(0x7f00000047c0)={0x0, 0x0, &(0x7f0000000000)=[{0x0, 0x18}], 0x1}, 0x0) [ 2868.359527][ T4131] bond1315: entered promiscuous mode [ 2868.369165][ T4131] 8021q: adding VLAN 0 to HW filter on device bond1315 [ 2868.407945][ T4141] bond1315: (slave bridge1277): making interface the new active one [ 2868.420099][ T4141] bridge1277: entered promiscuous mode [ 2868.430586][ T4141] bond1315: (slave bridge1277): Enslaving as an active interface with an up link 04:13:16 executing program 5: r0 = socket$netlink(0x10, 0x3, 0x0) r1 = socket$netlink(0x10, 0x3, 0x0) r2 = socket(0x10, 0x803, 0x0) sendmsg$NL80211_CMD_CRIT_PROTOCOL_START(r2, &(0x7f0000000580)={0x0, 0x0, &(0x7f0000000540)={0x0, 0x1c}}, 0x0) getsockname$packet(r2, &(0x7f0000000600)={0x11, 0x0, 0x0, 0x1, 0x0, 0x6, @broadcast}, &(0x7f0000000100)=0x14) sendmsg$nl_route(r1, &(0x7f0000000040)={0x0, 0x0, &(0x7f0000000000)={&(0x7f00000000c0)=ANY=[@ANYBLOB="3c00000010008506eb9afc4cd8d06e754a0081c5", @ANYRES32=r3, @ANYBLOB="2377f292252155b21c0012000c000100626f6e64000000000c0002000800010001"], 0x3c}}, 0x0) sendmsg$nl_route(r0, &(0x7f0000000240)={0x0, 0x0, &(0x7f0000000680)={&(0x7f0000000400)=@newlink={0x3c, 0x10, 0xffffff1f, 0x0, 0x7b352500, {}, [@IFLA_LINKINFO={0x14, 0x12, 0x0, 0x1, @bridge={{0xb}, {0x4}}}, @IFLA_MASTER={0x8, 0xa, r3}]}, 0x3c}}, 0x0) 04:13:16 executing program 0: r0 = socket$netlink(0x10, 0x3, 0x0) sendmsg$netlink(r0, &(0x7f00000047c0)={0x0, 0x0, &(0x7f0000000000)=[{&(0x7f0000000100)=ANY=[], 0x18}], 0x1}, 0x0) 04:13:16 executing program 0: r0 = socket$netlink(0x10, 0x3, 0x0) sendmsg$netlink(r0, &(0x7f00000047c0)={0x0, 0x0, &(0x7f0000000000)=[{&(0x7f0000000100)=ANY=[], 0x18}], 0x1}, 0x0) [ 2868.590992][ T4147] bond1325: entered promiscuous mode [ 2868.606092][ T4147] 8021q: adding VLAN 0 to HW filter on device bond1325 [ 2868.809673][ T4149] bond1325: (slave bridge1271): making interface the new active one [ 2868.840640][ T4149] bridge1271: entered promiscuous mode [ 2868.877435][ T4149] bond1325: (slave bridge1271): Enslaving as an active interface with an up link 04:13:16 executing program 1: r0 = socket$netlink(0x10, 0x3, 0x0) r1 = socket$netlink(0x10, 0x3, 0x0) r2 = socket(0x10, 0x803, 0x0) sendmsg$NL80211_CMD_CRIT_PROTOCOL_START(r2, &(0x7f0000000580)={0x0, 0x0, &(0x7f0000000540)={0x0, 0x1c}}, 0x0) getsockname$packet(r2, &(0x7f0000000600)={0x11, 0x0, 0x0, 0x1, 0x0, 0x6, @broadcast}, &(0x7f0000000100)=0x14) sendmsg$nl_route(r1, &(0x7f0000000040)={0x0, 0x0, &(0x7f0000000000)={&(0x7f00000000c0)=ANY=[@ANYBLOB="3c00000010008506eb9afc4cd8d06e754a0081c5", @ANYRES32=r3, @ANYBLOB="2377f292252155b21c0012000c000100626f6e64000000000c0002000800010001"], 0x3c}}, 0x0) sendmsg$nl_route(r0, &(0x7f0000000240)={0x0, 0x0, &(0x7f0000000680)={&(0x7f0000000400)=@newlink={0x3c, 0x10, 0xffffff1f, 0x0, 0x33672500, {}, [@IFLA_LINKINFO={0x14, 0x12, 0x0, 0x1, @bridge={{0xb}, {0x4}}}, @IFLA_MASTER={0x8, 0xa, r3}]}, 0x3c}}, 0x0) 04:13:16 executing program 0: r0 = socket$netlink(0x10, 0x3, 0x0) sendmsg$netlink(r0, &(0x7f00000047c0)={0x0, 0x0, &(0x7f0000000000)=[{&(0x7f0000000100)=ANY=[], 0x18}], 0x1}, 0x0) 04:13:16 executing program 4: ioctl$SIOCSIFHWADDR(0xffffffffffffffff, 0x8924, &(0x7f0000000040)={'geneve0\x00', @random="3cfeaa4162d8"}) r0 = socket$nl_generic(0x10, 0x3, 0x10) r1 = accept(0xffffffffffffffff, &(0x7f0000000080)=@nl, &(0x7f0000000100)=0x80) ioctl$BTRFS_IOC_SUBVOL_CREATE(r0, 0x5000940e, &(0x7f0000000240)={{r1}, "8e812815f4f4dae1b7a62bfb3f7189ee5d6cbf31e4e2c5c26ae83af35969714ce7f9e7f01c6c7096f9fd93d6c1bd23b03dc90892f91bb5abfcbe4548c0a94478c9101706e0dd214888a6ddf4b23181c45d24cc16ab2d6240bb84f4e8627ee8f34867f72a963fb13a9a3db73b87cb332cf9a14f9f480f4b7e9e78994fd768867b79104408895b0a27ad8e65e8014dfe2e0ef5b84490d58cd661b1ee297f0ab6e126cec3adc1ee292f2f231efda46a9e43c0c8cde6488039e406ccd15236ccd35636b723fd50de7703c5beea79f2edd5df96e04d060a9b387cb202ef677983c4088ba2b7b367e354eee1656fac128fa212fecc99c59422ebe8dd062d24b43e5b88a9b8f6163728b8dc1c40b03dcf15fec041a80c88d817ecd42feb910dd569916a3518745116976008b450600058cc1ee49012933b01054126e0998d4062708c34198e7f90b49740b4b312c1b78c535350ec06bda321e2a633fb7b499edb13dffe6c72463d60d50c4a8ef58a405ce31ca869a4d027e1dbf1ecfb0de201e7f087aa2e5d093f8840c0071c97e72b7946712130d8ba2540fe37674e86325cbb1bbc2e9770448feb81a28d7fb6a50abadacac9c255c63f838d352f7eaf32ff957e550ba3a0a4f844aa6207310384460ca2eaef6fa03e16cfc4af9cfd4f21a35c617bace68fb4a8d080979f34a3eea661f670ad43bb2095c7a4e4d9749cdfe7a146362414105eb656aad3c402bc36698bc84c1d1786bd64b2405487188922aad1e49825b9761adf683f73a7e84d56ed77fde3617bc55d0a57244148893c77c0ebe8a15ae33a54f5141f196c0fc66736a7926388de9399c0521f958158b4589e5de747e2d2141bb197fbc35dda2677881065013c6a2ca42b2025987733c8e9e4aace3003a15094ec935171eae98ee3a4be0ba1c47436ef035a3debbdc662dfa8d5bdbe3a3148df9a026246c00d2b7be48d3cdf19536386efbd4dc15d12739e0789a2ce1330b97f5c7f35c7cf6c36eabacbc9aa4fadf23fed088f280180cb525cae921d6300b3a9c771daeb5c2b633722f05a8f89a19ed4447e5c57f95122cd6db018d8967b0669e183f0d314fc92998352408af7ece1f7a637908eda91d30f6c365f1e83d53a840e422944194503021497a6c42f06077c257aa0bc25aad8b5cf713c066e043379c4eeac1322adf8e88a4f8dfaa0277cffe918ff08f7723deb1c74021b0d40d5a9d0e90bfcb1c209f227b45ba8888ade9cf4c984d0f3496b344cf1a1cf3654de7ebbc3812d4cc794c942b0c8b2c016358c96f1d578528476cf46144b3fbb53f3c7a49a9bf29ecf871a6bfb290eab40efeb007ca995413b57ca56f76d3a4b8c99a1e6c4ca9850539db4faeaa54b38b800e160377021c2dd4f37d2aa5373d8acb732c82dc897459867f7cba21fb34e3bbad9e9c5a94139e35dbff23dac8926fae52ebfc520ead6aff62c87487be29aec4146ec6eb91f52c32e912f02a74c2d9791da3b519a98c0de151a1b0e204e8d23145054d2f2c6416966a65f751353aabebaaf4da5123ff0c9987f17c5e1ea492a24c94068d28587bf8ce08ad086b20eb085335693ecc30bb3a8a677e4c1203293fd395b113e92143ae4b3c30b4d8162fea5be04343c2bb8419f59ac49d5034bfcb2e9d5df04c9b7883e54b42a522395fa6a2ba0b60b418718ebad76d978d31e4eecb419c32f1c4a5227267753420be0bf5ca183fb40076f68b3e0dabbf8a9a57730c4c10c5d41cb183d825ec81ba9ec884ded75b2a85b0a51c1f3f6dd4c18e80f9cba8c0518855ea87710cd896159e5971b54db4b0c694f655e424d2f1cc6dc8980d5ec0d877dfd4a4efeb529168f12546b9ec52136888c88ade85df8acf6f91fd77ee3cb799b188c3171c8ad2c834e41ccdbc22ebe4ef1a5a08997b9059cf44893fe2da9fc0b911296062f2f9d78655093064d303ce68119581032ca5a7d9ff0a72c6ae622c654dec214dcd60244f66d180c7c855558d7173df94b6acf7c87697bcb39a5f062d591f66ecae84f44077e7576d782f22777af70c8a34224f34a12fa54439fbe03b3c403082278f48513f4f8bd9e25bfe1df53454edad3fccb96bf5c506a07cc3450db9cffd27c8b2ffe83f6eb88423aa9095fe17f33277bdbc02215e3df4a2afbdf81200808421d076bdcd6405427f8b654e5c2e30bf6970d9ad4821bd185e14ea5a630b53aba6a6852752a679ab6a5d7d09bfe48e95b744e471336ccd53643f866dbc08093b06c85a3b0529fabab63bf348083fd14e8617bbf091ab6a16c70f8324f6493c644432b96da7752c44e94d3bb873f1d4eab123864a7968075064f8eaab0ab8812d916871530cbe37d057ee3a56af8de23fc2590680abd6cdf6e4c97c13ebe9c6e5ebd7a3b4119c8704be5e02b6e100d34abd2986e2c48b5dfaa0779e6048ffa9929f433a4cf4551e82f63c9b65c6a7aece3992719112d53cedcb7d41bc307085a27555eca3c7d6085a7338475f0344948a77fd62cb06dcbcb8c0cf13147b396a9864a11d62b4db69aea8bb213cb2f97e280153b565890f003844443097c0b39374e0278ccf6750617128c51b658860a7133640fe0833aec74f7f871f00d1d5d11040561d79501ab7f7e924e4a28cfb65de0c954943f6adfa5c331c167ab90508776414712da5ef502aed5bf5d9092b1c2e30a1fb517af63604ccb921d0985f546653a258dd0465ba2ab8fdc1657990880bfd010a7ecafaeaff7a0976fd6dbc18b507f74d134e5809a9dbd3d3c61c7555673fbb2c3ce4047726d5776f3c175403dd22438e8023395e61b9089933f14c2c5f3a794244cfcb0370af6b8a879714751ef288f94311eaad924348eec58bc52732aa4ddc84507b606750909166b5ffe3197eb8755a127444ebeb484a1a21057d9ce799424d07f905659ed3f7716d35210d605292f4e152e15c36523964ecae60d331999372947e00ee346dd2c53575886fc636293d66fc32faea001d0d28d51848de0c583ed79e0f5aa7971d06b31d7f87ae734306b7fbf069d7986a7298f9b471e83db1b95f52f7ae3fc7b658d45cc6c2c4b5d95e07f491a06249c2d1b526f51bbb48238c39c1c91e5a4f6b7d7d72db9ed4474c0f98c5a5a12b6d417f3d3f0a9edcb3e7debeb87ec0a95ef8da1f48e99999f928c3a210595074012fe32dfc3ec4f0b88e07e82a89b36aa3b3a3ec7279cdb7bbcc4f02f79b5b6ee9b23a1579443767f9099529efd1825bbf6cdaaf52bb130a4e01fe10e7702eb563478277993aa042f5ab39d5707ee55379a88309ca980ce021d64692267c0b923a49ce191c036a9fd1f3b5da2e40bcc9781e79e57250e6292e3d5b489cc874168bbcd609ca00006d38f1300a23a4f3b9fd249836937d7cf9fb6f0d1a313ca815c8d3dff0213570eb63e8dc15c83434be08302b263c6e94a974fd2919b0c410734b3d4ec98b6ffd03b2be1d0ccf341127715aa4b796373561c8c17396e1b9a2334ef1e878d350d8241d9b32bd8311e8cd41e80a7435a66eb34dd18dbe1ed35c2caf3801f53b771a4b854aa3c9e815261995e1c842825292674995826da25542a12c092d44748e76deb23a9739edf71564d5d72ff62e3e5c3ea51c57fd2389bd663f85794004b9face367af5f2981e5596149f6ed67b766ce3c185de7333d237fdfcad99df3629d28622769611a5a72b672cc1e0d28e2b05e39915159087c083f5b83daa4f1ca4fb3405a29dc6b8f41904a7e272c5fce37ec8e29313a44f0f804bf417459da42065063c3877062f96cb8c1d3a4115113e95d38e68ca2f07246a96f77d1cd9971fa03e3ce781971a9f088e460072c451fb36232cf015093fba2e11347e7389cfdf1531ca079901a64c51495045a4f5e86aff5b7724533eadfce98127b4810438b79e9c750720e0b5cd0323d3081c5ebbfe3bc024e342986b88776fec77a6ef379e1dda9ad5cf5426c0af821ca986a0100352be1ac1d35cef9f2282a655046ef0cd1b002683db007bf4330c5f1994c239ae43e3f70a2d065be88266c5f7cd5b1530672571c97e08fce74cdba1908a1c48f3d19e5261799dacf106052966d76a26c5830eed3e3ef5620a1c435282ecfcfa3fbf4935f3c37546d7b97c5c2a5a0c009d2674741c788413345309e487111035b10086ec82da9d722b3905d5db535d287bde77d73442e87fb3ede2a561528f2a7d48ce6b98202e5f8eb5843b1c1a4cc452587f145cfa38ae33742b35ccb516f73cc90090c721ca4e4e522a2d3045d63b69566334434a1b62e2baf1b142eabd62dd6113999e10e1b4d6305e5121c1a7ceb901fe8e31e011dc7cbfe8ca913f3d267d0fe6ac1e7cb10dd903ee83f51259f0f83770594d7f667ed85c646ab0c94a156f8068e1b441e9e3c810d244510ec04943e9d93e9ac4152b720228489e4a4fbda7a7b0baf437105e4f9868e4ded50e9abfc293c8f9ab6784c255ebec495d05a872f4381b452b628f1f705bc8dbd2e1a6f4fde5c42bf65b93ac6dce1bbb00d5726d80e7fa429fac664b225fb78e56573a184732fd72947034ad8265c7b6bde1782e6f679cfcd4cab903b9a6104fa07f3425a14e60f96ed92278936d76add20720331b3b02380149483e7f68328b6a69375c43053917cd416119da7f546e456714d14cd4dbfe72413fcd79a26ede2c71cd130b9dfb2b353b80bd54766ed3e6239304dc2eeea33b1fcbfa9de4673e1aff1c3baec9ee352012fd6f22643428fb0ac4f043aea4ac7c2d6368597428ebb50201b3fbf5601294f7a6f6fbf3da60bfcace75140ecc81233f9da2e1eeafef7d71edcb8242968a485de10dd1a5026b7ef501ed7ddd559c569af6a7450c30e0a85dd2b3831acbe655a756b22bcf655dae0c9ffae8f4efe1324244260333c5d9161605e992f2b1784b9c93a326c3f0319a9f4b1ba6a994771041614aa2abcc9f10d4a70b2d5c95dbe246d4d5470c390d4caaaee36ee732347ea5b0dbe4ef97d0b9d2a446b78f577acc247dab3863ef539b902a47e32f9813a54f6c9a1fa04ab7bb7dd2e34924091ee6e1f52da02f13c16d46ff15d177361b5984ca36418f928910aa75fb73f08be8ebe9533015d3b11c45023bd54c4f91100f6e615f894f947a0f26741b6377f1eec7ef001e5a1ac713a927df62c7959380f449ddd7455d15ac088e73c0739240ac930fac4d9bac7fdf1ca3eeeda23e51033b5636e4fc55aa24c57fed495dd72ad502c563431bafbd0a3acdbd532c64df0dd0396f67947020742de75079d4178fb3f66dc4464e4bc20028c3cdbd9a4df6e898a9dc2c1c0b92d540255f0daf7f934fde9503c7268ef519a4c957ab433891c13e716645210fe80bfef39facd92313289002421ea9f0aa98a62c40353eb1d1e717327f9ffaefe245aac5a8617ae0a22c83d9ecfeffd756f899339c425ae610ddd9afe5aed3c07ec440916ca054427094f47bfd5d93b83d648030ed912686289d222417c2a8485f1b758c6e7eea85576e5160783bdeedade9d113f7b0f03f76642ff7ac89c0c8e8e395763aeb00b2b0c8cc7507225db4210bcfab834fef62457bcc5efa4a31f2ea8ace28ef9db79063712638b55949208ea52f8a7020f8028396f2662dccc7449b720c3c71a9488c87640577306b6b7ee0e03eec6c748890a8b2e39400b634ee2b4f82069f5ff030c83f4679a7add11e6711ca4212dceb051be9eac54c48c3677fad77778ceac60f2ee76895339214966d96188eee7891b904457ad42855527fd9b0bebc05d327e02a2d73aad59663eb217cd"}) r2 = socket$inet6(0xa, 0x2, 0x0) ioctl$sock_SIOCETHTOOL(r2, 0x89f0, &(0x7f0000000000)={'bridge0\x00', &(0x7f0000000200)=@ethtool_ringparam={0x10}}) 04:13:16 executing program 3: r0 = socket$netlink(0x10, 0x3, 0x0) sendmsg$netlink(r0, &(0x7f00000047c0)={0x0, 0x0, &(0x7f0000000000)=[{&(0x7f0000000100)=ANY=[@ANYBLOB="1800000056000106000000a500db730007"], 0x18}], 0x1}, 0x0) [ 2868.939084][ T4166] validate_nla: 5 callbacks suppressed [ 2868.939106][ T4166] netlink: 'syz-executor.2': attribute type 1 has an invalid length. 04:13:16 executing program 0: r0 = socket$netlink(0x10, 0x3, 0x0) sendmsg$netlink(r0, &(0x7f00000047c0)={0x0, 0x0, &(0x7f0000000000)=[{&(0x7f0000000100)=ANY=[@ANYBLOB], 0x18}], 0x1}, 0x0) [ 2869.111763][ T4166] bond1310: entered promiscuous mode [ 2869.127816][ T4166] 8021q: adding VLAN 0 to HW filter on device bond1310 04:13:16 executing program 0: r0 = socket$netlink(0x10, 0x3, 0x0) sendmsg$netlink(r0, &(0x7f00000047c0)={0x0, 0x0, &(0x7f0000000000)=[{&(0x7f0000000100)=ANY=[@ANYBLOB], 0x18}], 0x1}, 0x0) [ 2869.306373][ T4167] bond1310: (slave bridge1230): making interface the new active one [ 2869.335883][ T4167] bridge1230: entered promiscuous mode 04:13:16 executing program 2: r0 = socket$netlink(0x10, 0x3, 0x0) r1 = socket$netlink(0x10, 0x3, 0x0) r2 = socket(0x10, 0x803, 0x0) sendmsg$NL80211_CMD_CRIT_PROTOCOL_START(r2, &(0x7f0000000580)={0x0, 0x0, &(0x7f0000000540)={0x0, 0x1c}}, 0x0) getsockname$packet(r2, &(0x7f0000000600)={0x11, 0x0, 0x0, 0x1, 0x0, 0x6, @broadcast}, &(0x7f0000000100)=0x14) sendmsg$nl_route(r1, &(0x7f0000000040)={0x0, 0x0, &(0x7f0000000000)={&(0x7f00000000c0)=ANY=[@ANYBLOB="3c00000010008506eb9afc4cd8d06e754a0081c5", @ANYRES32=r3, @ANYBLOB="2377f292252155b21c0012000c000100626f6e64000000000c0002000800010001"], 0x3c}}, 0x0) sendmsg$nl_route(r0, &(0x7f0000000240)={0x0, 0x0, &(0x7f0000000680)={&(0x7f0000000400)=@newlink={0x3c, 0x10, 0xffffff1f, 0x0, 0x43c2500, {}, [@IFLA_LINKINFO={0x14, 0x12, 0x0, 0x1, @bridge={{0xb}, {0x4}}}, @IFLA_MASTER={0x8, 0xa, r3}]}, 0x3c}}, 0x0) 04:13:16 executing program 0: r0 = socket$netlink(0x10, 0x3, 0x0) sendmsg$netlink(r0, &(0x7f00000047c0)={0x0, 0x0, &(0x7f0000000000)=[{&(0x7f0000000100)=ANY=[@ANYBLOB], 0x18}], 0x1}, 0x0) [ 2869.355619][ T4167] bond1310: (slave bridge1230): Enslaving as an active interface with an up link [ 2869.380869][ T4171] netlink: 'syz-executor.5': attribute type 1 has an invalid length. 04:13:16 executing program 0: r0 = socket$netlink(0x10, 0x3, 0x0) sendmsg$netlink(r0, &(0x7f00000047c0)={0x0, 0x0, &(0x7f0000000000)=[{&(0x7f0000000100)=ANY=[@ANYBLOB="1800000056"], 0x18}], 0x1}, 0x0) [ 2869.464135][ T4171] bond1316: entered promiscuous mode [ 2869.470347][ T4171] 8021q: adding VLAN 0 to HW filter on device bond1316 [ 2869.624737][ T4173] bond1316: (slave bridge1278): making interface the new active one [ 2869.654017][ T4173] bridge1278: entered promiscuous mode [ 2869.693838][ T4173] bond1316: (slave bridge1278): Enslaving as an active interface with an up link [ 2869.723291][ T4185] netlink: 'syz-executor.1': attribute type 1 has an invalid length. 04:13:17 executing program 5: r0 = socket$netlink(0x10, 0x3, 0x0) r1 = socket$netlink(0x10, 0x3, 0x0) r2 = socket(0x10, 0x803, 0x0) sendmsg$NL80211_CMD_CRIT_PROTOCOL_START(r2, &(0x7f0000000580)={0x0, 0x0, &(0x7f0000000540)={0x0, 0x1c}}, 0x0) getsockname$packet(r2, &(0x7f0000000600)={0x11, 0x0, 0x0, 0x1, 0x0, 0x6, @broadcast}, &(0x7f0000000100)=0x14) sendmsg$nl_route(r1, &(0x7f0000000040)={0x0, 0x0, &(0x7f0000000000)={&(0x7f00000000c0)=ANY=[@ANYBLOB="3c00000010008506eb9afc4cd8d06e754a0081c5", @ANYRES32=r3, @ANYBLOB="2377f292252155b21c0012000c000100626f6e64000000000c0002000800010001"], 0x3c}}, 0x0) sendmsg$nl_route(r0, &(0x7f0000000240)={0x0, 0x0, &(0x7f0000000680)={&(0x7f0000000400)=@newlink={0x3c, 0x10, 0xffffff1f, 0x0, 0x7c030000, {}, [@IFLA_LINKINFO={0x14, 0x12, 0x0, 0x1, @bridge={{0xb}, {0x4}}}, @IFLA_MASTER={0x8, 0xa, r3}]}, 0x3c}}, 0x0) 04:13:17 executing program 0: r0 = socket$netlink(0x10, 0x3, 0x0) sendmsg$netlink(r0, &(0x7f00000047c0)={0x0, 0x0, &(0x7f0000000000)=[{&(0x7f0000000100)=ANY=[@ANYBLOB="1800000056"], 0x18}], 0x1}, 0x0) 04:13:17 executing program 4: ioctl$SIOCSIFHWADDR(0xffffffffffffffff, 0x8924, &(0x7f0000000040)={'geneve0\x00', @random="3cfeaa4162d8"}) r0 = socket$nl_generic(0x10, 0x3, 0x10) (async) r1 = accept(0xffffffffffffffff, &(0x7f0000000080)=@nl, &(0x7f0000000100)=0x80) ioctl$BTRFS_IOC_SUBVOL_CREATE(r0, 0x5000940e, &(0x7f0000000240)={{r1}, "8e812815f4f4dae1b7a62bfb3f7189ee5d6cbf31e4e2c5c26ae83af35969714ce7f9e7f01c6c7096f9fd93d6c1bd23b03dc90892f91bb5abfcbe4548c0a94478c9101706e0dd214888a6ddf4b23181c45d24cc16ab2d6240bb84f4e8627ee8f34867f72a963fb13a9a3db73b87cb332cf9a14f9f480f4b7e9e78994fd768867b79104408895b0a27ad8e65e8014dfe2e0ef5b84490d58cd661b1ee297f0ab6e126cec3adc1ee292f2f231efda46a9e43c0c8cde6488039e406ccd15236ccd35636b723fd50de7703c5beea79f2edd5df96e04d060a9b387cb202ef677983c4088ba2b7b367e354eee1656fac128fa212fecc99c59422ebe8dd062d24b43e5b88a9b8f6163728b8dc1c40b03dcf15fec041a80c88d817ecd42feb910dd569916a3518745116976008b450600058cc1ee49012933b01054126e0998d4062708c34198e7f90b49740b4b312c1b78c535350ec06bda321e2a633fb7b499edb13dffe6c72463d60d50c4a8ef58a405ce31ca869a4d027e1dbf1ecfb0de201e7f087aa2e5d093f8840c0071c97e72b7946712130d8ba2540fe37674e86325cbb1bbc2e9770448feb81a28d7fb6a50abadacac9c255c63f838d352f7eaf32ff957e550ba3a0a4f844aa6207310384460ca2eaef6fa03e16cfc4af9cfd4f21a35c617bace68fb4a8d080979f34a3eea661f670ad43bb2095c7a4e4d9749cdfe7a146362414105eb656aad3c402bc36698bc84c1d1786bd64b2405487188922aad1e49825b9761adf683f73a7e84d56ed77fde3617bc55d0a57244148893c77c0ebe8a15ae33a54f5141f196c0fc66736a7926388de9399c0521f958158b4589e5de747e2d2141bb197fbc35dda2677881065013c6a2ca42b2025987733c8e9e4aace3003a15094ec935171eae98ee3a4be0ba1c47436ef035a3debbdc662dfa8d5bdbe3a3148df9a026246c00d2b7be48d3cdf19536386efbd4dc15d12739e0789a2ce1330b97f5c7f35c7cf6c36eabacbc9aa4fadf23fed088f280180cb525cae921d6300b3a9c771daeb5c2b633722f05a8f89a19ed4447e5c57f95122cd6db018d8967b0669e183f0d314fc92998352408af7ece1f7a637908eda91d30f6c365f1e83d53a840e422944194503021497a6c42f06077c257aa0bc25aad8b5cf713c066e043379c4eeac1322adf8e88a4f8dfaa0277cffe918ff08f7723deb1c74021b0d40d5a9d0e90bfcb1c209f227b45ba8888ade9cf4c984d0f3496b344cf1a1cf3654de7ebbc3812d4cc794c942b0c8b2c016358c96f1d578528476cf46144b3fbb53f3c7a49a9bf29ecf871a6bfb290eab40efeb007ca995413b57ca56f76d3a4b8c99a1e6c4ca9850539db4faeaa54b38b800e160377021c2dd4f37d2aa5373d8acb732c82dc897459867f7cba21fb34e3bbad9e9c5a94139e35dbff23dac8926fae52ebfc520ead6aff62c87487be29aec4146ec6eb91f52c32e912f02a74c2d9791da3b519a98c0de151a1b0e204e8d23145054d2f2c6416966a65f751353aabebaaf4da5123ff0c9987f17c5e1ea492a24c94068d28587bf8ce08ad086b20eb085335693ecc30bb3a8a677e4c1203293fd395b113e92143ae4b3c30b4d8162fea5be04343c2bb8419f59ac49d5034bfcb2e9d5df04c9b7883e54b42a522395fa6a2ba0b60b418718ebad76d978d31e4eecb419c32f1c4a5227267753420be0bf5ca183fb40076f68b3e0dabbf8a9a57730c4c10c5d41cb183d825ec81ba9ec884ded75b2a85b0a51c1f3f6dd4c18e80f9cba8c0518855ea87710cd896159e5971b54db4b0c694f655e424d2f1cc6dc8980d5ec0d877dfd4a4efeb529168f12546b9ec52136888c88ade85df8acf6f91fd77ee3cb799b188c3171c8ad2c834e41ccdbc22ebe4ef1a5a08997b9059cf44893fe2da9fc0b911296062f2f9d78655093064d303ce68119581032ca5a7d9ff0a72c6ae622c654dec214dcd60244f66d180c7c855558d7173df94b6acf7c87697bcb39a5f062d591f66ecae84f44077e7576d782f22777af70c8a34224f34a12fa54439fbe03b3c403082278f48513f4f8bd9e25bfe1df53454edad3fccb96bf5c506a07cc3450db9cffd27c8b2ffe83f6eb88423aa9095fe17f33277bdbc02215e3df4a2afbdf81200808421d076bdcd6405427f8b654e5c2e30bf6970d9ad4821bd185e14ea5a630b53aba6a6852752a679ab6a5d7d09bfe48e95b744e471336ccd53643f866dbc08093b06c85a3b0529fabab63bf348083fd14e8617bbf091ab6a16c70f8324f6493c644432b96da7752c44e94d3bb873f1d4eab123864a7968075064f8eaab0ab8812d916871530cbe37d057ee3a56af8de23fc2590680abd6cdf6e4c97c13ebe9c6e5ebd7a3b4119c8704be5e02b6e100d34abd2986e2c48b5dfaa0779e6048ffa9929f433a4cf4551e82f63c9b65c6a7aece3992719112d53cedcb7d41bc307085a27555eca3c7d6085a7338475f0344948a77fd62cb06dcbcb8c0cf13147b396a9864a11d62b4db69aea8bb213cb2f97e280153b565890f003844443097c0b39374e0278ccf6750617128c51b658860a7133640fe0833aec74f7f871f00d1d5d11040561d79501ab7f7e924e4a28cfb65de0c954943f6adfa5c331c167ab90508776414712da5ef502aed5bf5d9092b1c2e30a1fb517af63604ccb921d0985f546653a258dd0465ba2ab8fdc1657990880bfd010a7ecafaeaff7a0976fd6dbc18b507f74d134e5809a9dbd3d3c61c7555673fbb2c3ce4047726d5776f3c175403dd22438e8023395e61b9089933f14c2c5f3a794244cfcb0370af6b8a879714751ef288f94311eaad924348eec58bc52732aa4ddc84507b606750909166b5ffe3197eb8755a127444ebeb484a1a21057d9ce799424d07f905659ed3f7716d35210d605292f4e152e15c36523964ecae60d331999372947e00ee346dd2c53575886fc636293d66fc32faea001d0d28d51848de0c583ed79e0f5aa7971d06b31d7f87ae734306b7fbf069d7986a7298f9b471e83db1b95f52f7ae3fc7b658d45cc6c2c4b5d95e07f491a06249c2d1b526f51bbb48238c39c1c91e5a4f6b7d7d72db9ed4474c0f98c5a5a12b6d417f3d3f0a9edcb3e7debeb87ec0a95ef8da1f48e99999f928c3a210595074012fe32dfc3ec4f0b88e07e82a89b36aa3b3a3ec7279cdb7bbcc4f02f79b5b6ee9b23a1579443767f9099529efd1825bbf6cdaaf52bb130a4e01fe10e7702eb563478277993aa042f5ab39d5707ee55379a88309ca980ce021d64692267c0b923a49ce191c036a9fd1f3b5da2e40bcc9781e79e57250e6292e3d5b489cc874168bbcd609ca00006d38f1300a23a4f3b9fd249836937d7cf9fb6f0d1a313ca815c8d3dff0213570eb63e8dc15c83434be08302b263c6e94a974fd2919b0c410734b3d4ec98b6ffd03b2be1d0ccf341127715aa4b796373561c8c17396e1b9a2334ef1e878d350d8241d9b32bd8311e8cd41e80a7435a66eb34dd18dbe1ed35c2caf3801f53b771a4b854aa3c9e815261995e1c842825292674995826da25542a12c092d44748e76deb23a9739edf71564d5d72ff62e3e5c3ea51c57fd2389bd663f85794004b9face367af5f2981e5596149f6ed67b766ce3c185de7333d237fdfcad99df3629d28622769611a5a72b672cc1e0d28e2b05e39915159087c083f5b83daa4f1ca4fb3405a29dc6b8f41904a7e272c5fce37ec8e29313a44f0f804bf417459da42065063c3877062f96cb8c1d3a4115113e95d38e68ca2f07246a96f77d1cd9971fa03e3ce781971a9f088e460072c451fb36232cf015093fba2e11347e7389cfdf1531ca079901a64c51495045a4f5e86aff5b7724533eadfce98127b4810438b79e9c750720e0b5cd0323d3081c5ebbfe3bc024e342986b88776fec77a6ef379e1dda9ad5cf5426c0af821ca986a0100352be1ac1d35cef9f2282a655046ef0cd1b002683db007bf4330c5f1994c239ae43e3f70a2d065be88266c5f7cd5b1530672571c97e08fce74cdba1908a1c48f3d19e5261799dacf106052966d76a26c5830eed3e3ef5620a1c435282ecfcfa3fbf4935f3c37546d7b97c5c2a5a0c009d2674741c788413345309e487111035b10086ec82da9d722b3905d5db535d287bde77d73442e87fb3ede2a561528f2a7d48ce6b98202e5f8eb5843b1c1a4cc452587f145cfa38ae33742b35ccb516f73cc90090c721ca4e4e522a2d3045d63b69566334434a1b62e2baf1b142eabd62dd6113999e10e1b4d6305e5121c1a7ceb901fe8e31e011dc7cbfe8ca913f3d267d0fe6ac1e7cb10dd903ee83f51259f0f83770594d7f667ed85c646ab0c94a156f8068e1b441e9e3c810d244510ec04943e9d93e9ac4152b720228489e4a4fbda7a7b0baf437105e4f9868e4ded50e9abfc293c8f9ab6784c255ebec495d05a872f4381b452b628f1f705bc8dbd2e1a6f4fde5c42bf65b93ac6dce1bbb00d5726d80e7fa429fac664b225fb78e56573a184732fd72947034ad8265c7b6bde1782e6f679cfcd4cab903b9a6104fa07f3425a14e60f96ed92278936d76add20720331b3b02380149483e7f68328b6a69375c43053917cd416119da7f546e456714d14cd4dbfe72413fcd79a26ede2c71cd130b9dfb2b353b80bd54766ed3e6239304dc2eeea33b1fcbfa9de4673e1aff1c3baec9ee352012fd6f22643428fb0ac4f043aea4ac7c2d6368597428ebb50201b3fbf5601294f7a6f6fbf3da60bfcace75140ecc81233f9da2e1eeafef7d71edcb8242968a485de10dd1a5026b7ef501ed7ddd559c569af6a7450c30e0a85dd2b3831acbe655a756b22bcf655dae0c9ffae8f4efe1324244260333c5d9161605e992f2b1784b9c93a326c3f0319a9f4b1ba6a994771041614aa2abcc9f10d4a70b2d5c95dbe246d4d5470c390d4caaaee36ee732347ea5b0dbe4ef97d0b9d2a446b78f577acc247dab3863ef539b902a47e32f9813a54f6c9a1fa04ab7bb7dd2e34924091ee6e1f52da02f13c16d46ff15d177361b5984ca36418f928910aa75fb73f08be8ebe9533015d3b11c45023bd54c4f91100f6e615f894f947a0f26741b6377f1eec7ef001e5a1ac713a927df62c7959380f449ddd7455d15ac088e73c0739240ac930fac4d9bac7fdf1ca3eeeda23e51033b5636e4fc55aa24c57fed495dd72ad502c563431bafbd0a3acdbd532c64df0dd0396f67947020742de75079d4178fb3f66dc4464e4bc20028c3cdbd9a4df6e898a9dc2c1c0b92d540255f0daf7f934fde9503c7268ef519a4c957ab433891c13e716645210fe80bfef39facd92313289002421ea9f0aa98a62c40353eb1d1e717327f9ffaefe245aac5a8617ae0a22c83d9ecfeffd756f899339c425ae610ddd9afe5aed3c07ec440916ca054427094f47bfd5d93b83d648030ed912686289d222417c2a8485f1b758c6e7eea85576e5160783bdeedade9d113f7b0f03f76642ff7ac89c0c8e8e395763aeb00b2b0c8cc7507225db4210bcfab834fef62457bcc5efa4a31f2ea8ace28ef9db79063712638b55949208ea52f8a7020f8028396f2662dccc7449b720c3c71a9488c87640577306b6b7ee0e03eec6c748890a8b2e39400b634ee2b4f82069f5ff030c83f4679a7add11e6711ca4212dceb051be9eac54c48c3677fad77778ceac60f2ee76895339214966d96188eee7891b904457ad42855527fd9b0bebc05d327e02a2d73aad59663eb217cd"}) r2 = socket$inet6(0xa, 0x2, 0x0) ioctl$sock_SIOCETHTOOL(r2, 0x89f0, &(0x7f0000000000)={'bridge0\x00', &(0x7f0000000200)=@ethtool_ringparam={0x10}}) [ 2869.815897][ T4185] bond1326: entered promiscuous mode [ 2869.833879][ T4185] 8021q: adding VLAN 0 to HW filter on device bond1326 04:13:17 executing program 1: r0 = socket$netlink(0x10, 0x3, 0x0) r1 = socket$netlink(0x10, 0x3, 0x0) r2 = socket(0x10, 0x803, 0x0) sendmsg$NL80211_CMD_CRIT_PROTOCOL_START(r2, &(0x7f0000000580)={0x0, 0x0, &(0x7f0000000540)={0x0, 0x1c}}, 0x0) getsockname$packet(r2, &(0x7f0000000600)={0x11, 0x0, 0x0, 0x1, 0x0, 0x6, @broadcast}, &(0x7f0000000100)=0x14) sendmsg$nl_route(r1, &(0x7f0000000040)={0x0, 0x0, &(0x7f0000000000)={&(0x7f00000000c0)=ANY=[@ANYBLOB="3c00000010008506eb9afc4cd8d06e754a0081c5", @ANYRES32=r3, @ANYBLOB="2377f292252155b21c0012000c000100626f6e64000000000c0002000800010001"], 0x3c}}, 0x0) sendmsg$nl_route(r0, &(0x7f0000000240)={0x0, 0x0, &(0x7f0000000680)={&(0x7f0000000400)=@newlink={0x3c, 0x10, 0xffffff1f, 0x0, 0x34010000, {}, [@IFLA_LINKINFO={0x14, 0x12, 0x0, 0x1, @bridge={{0xb}, {0x4}}}, @IFLA_MASTER={0x8, 0xa, r3}]}, 0x3c}}, 0x0) 04:13:17 executing program 0: r0 = socket$netlink(0x10, 0x3, 0x0) sendmsg$netlink(r0, &(0x7f00000047c0)={0x0, 0x0, &(0x7f0000000000)=[{&(0x7f0000000100)=ANY=[@ANYBLOB="1800000056"], 0x18}], 0x1}, 0x0) 04:13:17 executing program 3: r0 = socket$netlink(0x10, 0x3, 0x0) sendmsg$netlink(r0, &(0x7f00000047c0)={0x0, 0x0, &(0x7f0000000000)=[{&(0x7f0000000100)=ANY=[@ANYBLOB="1800000056000106000000a600db730007"], 0x18}], 0x1}, 0x0) [ 2869.954847][ T4188] bond1326: (slave bridge1272): making interface the new active one [ 2869.963876][ T4188] bridge1272: entered promiscuous mode [ 2869.976506][ T4188] bond1326: (slave bridge1272): Enslaving as an active interface with an up link 04:13:17 executing program 0: r0 = socket$netlink(0x10, 0x3, 0x0) sendmsg$netlink(r0, &(0x7f00000047c0)={0x0, 0x0, &(0x7f0000000000)=[{&(0x7f0000000100)=ANY=[@ANYBLOB="18000000560001"], 0x18}], 0x1}, 0x0) [ 2870.041423][ T4196] netlink: 'syz-executor.2': attribute type 1 has an invalid length. 04:13:17 executing program 0: r0 = socket$netlink(0x10, 0x3, 0x0) sendmsg$netlink(r0, &(0x7f00000047c0)={0x0, 0x0, &(0x7f0000000000)=[{&(0x7f0000000100)=ANY=[@ANYBLOB="18000000560001"], 0x18}], 0x1}, 0x0) [ 2870.140202][ T4196] bond1311: entered promiscuous mode [ 2870.148439][ T4196] 8021q: adding VLAN 0 to HW filter on device bond1311 [ 2870.237270][ T4199] bond1311: (slave bridge1231): making interface the new active one [ 2870.250363][ T4199] bridge1231: entered promiscuous mode [ 2870.261357][ T4199] bond1311: (slave bridge1231): Enslaving as an active interface with an up link 04:13:17 executing program 2: r0 = socket$netlink(0x10, 0x3, 0x0) r1 = socket$netlink(0x10, 0x3, 0x0) r2 = socket(0x10, 0x803, 0x0) sendmsg$NL80211_CMD_CRIT_PROTOCOL_START(r2, &(0x7f0000000580)={0x0, 0x0, &(0x7f0000000540)={0x0, 0x1c}}, 0x0) getsockname$packet(r2, &(0x7f0000000600)={0x11, 0x0, 0x0, 0x1, 0x0, 0x6, @broadcast}, &(0x7f0000000100)=0x14) sendmsg$nl_route(r1, &(0x7f0000000040)={0x0, 0x0, &(0x7f0000000000)={&(0x7f00000000c0)=ANY=[@ANYBLOB="3c00000010008506eb9afc4cd8d06e754a0081c5", @ANYRES32=r3, @ANYBLOB="2377f292252155b21c0012000c000100626f6e64000000000c0002000800010001"], 0x3c}}, 0x0) sendmsg$nl_route(r0, &(0x7f0000000240)={0x0, 0x0, &(0x7f0000000680)={&(0x7f0000000400)=@newlink={0x3c, 0x10, 0xffffff1f, 0x0, 0x5000000, {}, [@IFLA_LINKINFO={0x14, 0x12, 0x0, 0x1, @bridge={{0xb}, {0x4}}}, @IFLA_MASTER={0x8, 0xa, r3}]}, 0x3c}}, 0x0) [ 2870.281163][ T4209] netlink: 'syz-executor.5': attribute type 1 has an invalid length. 04:13:17 executing program 0: r0 = socket$netlink(0x10, 0x3, 0x0) sendmsg$netlink(r0, &(0x7f00000047c0)={0x0, 0x0, &(0x7f0000000000)=[{&(0x7f0000000100)=ANY=[@ANYBLOB="18000000560001"], 0x18}], 0x1}, 0x0) 04:13:17 executing program 0: r0 = socket$netlink(0x10, 0x3, 0x0) sendmsg$netlink(r0, &(0x7f00000047c0)={0x0, 0x0, &(0x7f0000000000)=[{&(0x7f0000000100)=ANY=[@ANYBLOB="1800000056000106"], 0x18}], 0x1}, 0x0) 04:13:17 executing program 0: r0 = socket$netlink(0x10, 0x3, 0x0) sendmsg$netlink(r0, &(0x7f00000047c0)={0x0, 0x0, &(0x7f0000000000)=[{&(0x7f0000000100)=ANY=[@ANYBLOB="18000000560001"], 0x18}], 0x1}, 0x0) [ 2870.586280][ T4209] bond1317: entered promiscuous mode [ 2870.600814][ T4209] 8021q: adding VLAN 0 to HW filter on device bond1317 [ 2870.710246][ T4212] bond1317: (slave bridge1279): making interface the new active one [ 2870.721606][ T4212] bridge1279: entered promiscuous mode [ 2870.747446][ T4212] bond1317: (slave bridge1279): Enslaving as an active interface with an up link 04:13:18 executing program 5: r0 = socket$netlink(0x10, 0x3, 0x0) r1 = socket$netlink(0x10, 0x3, 0x0) r2 = socket(0x10, 0x803, 0x0) sendmsg$NL80211_CMD_CRIT_PROTOCOL_START(r2, &(0x7f0000000580)={0x0, 0x0, &(0x7f0000000540)={0x0, 0x1c}}, 0x0) getsockname$packet(r2, &(0x7f0000000600)={0x11, 0x0, 0x0, 0x1, 0x0, 0x6, @broadcast}, &(0x7f0000000100)=0x14) sendmsg$nl_route(r1, &(0x7f0000000040)={0x0, 0x0, &(0x7f0000000000)={&(0x7f00000000c0)=ANY=[@ANYBLOB="3c00000010008506eb9afc4cd8d06e754a0081c5", @ANYRES32=r3, @ANYBLOB="2377f292252155b21c0012000c000100626f6e64000000000c0002000800010001"], 0x3c}}, 0x0) sendmsg$nl_route(r0, &(0x7f0000000240)={0x0, 0x0, &(0x7f0000000680)={&(0x7f0000000400)=@newlink={0x3c, 0x10, 0xffffff1f, 0x0, 0x7c352500, {}, [@IFLA_LINKINFO={0x14, 0x12, 0x0, 0x1, @bridge={{0xb}, {0x4}}}, @IFLA_MASTER={0x8, 0xa, r3}]}, 0x3c}}, 0x0) 04:13:18 executing program 0: r0 = socket$netlink(0x10, 0x3, 0x0) sendmsg$netlink(r0, &(0x7f00000047c0)={0x0, 0x0, &(0x7f0000000000)=[{&(0x7f0000000100)=ANY=[@ANYBLOB="18000000560001"], 0x18}], 0x1}, 0x0) 04:13:18 executing program 4: ioctl$SIOCSIFHWADDR(0xffffffffffffffff, 0x8924, &(0x7f0000000040)={'geneve0\x00', @random="3cfeaa4162d8"}) (async) r0 = socket$nl_generic(0x10, 0x3, 0x10) (async) r1 = accept(0xffffffffffffffff, &(0x7f0000000080)=@nl, &(0x7f0000000100)=0x80) ioctl$BTRFS_IOC_SUBVOL_CREATE(r0, 0x5000940e, &(0x7f0000000240)={{r1}, "8e812815f4f4dae1b7a62bfb3f7189ee5d6cbf31e4e2c5c26ae83af35969714ce7f9e7f01c6c7096f9fd93d6c1bd23b03dc90892f91bb5abfcbe4548c0a94478c9101706e0dd214888a6ddf4b23181c45d24cc16ab2d6240bb84f4e8627ee8f34867f72a963fb13a9a3db73b87cb332cf9a14f9f480f4b7e9e78994fd768867b79104408895b0a27ad8e65e8014dfe2e0ef5b84490d58cd661b1ee297f0ab6e126cec3adc1ee292f2f231efda46a9e43c0c8cde6488039e406ccd15236ccd35636b723fd50de7703c5beea79f2edd5df96e04d060a9b387cb202ef677983c4088ba2b7b367e354eee1656fac128fa212fecc99c59422ebe8dd062d24b43e5b88a9b8f6163728b8dc1c40b03dcf15fec041a80c88d817ecd42feb910dd569916a3518745116976008b450600058cc1ee49012933b01054126e0998d4062708c34198e7f90b49740b4b312c1b78c535350ec06bda321e2a633fb7b499edb13dffe6c72463d60d50c4a8ef58a405ce31ca869a4d027e1dbf1ecfb0de201e7f087aa2e5d093f8840c0071c97e72b7946712130d8ba2540fe37674e86325cbb1bbc2e9770448feb81a28d7fb6a50abadacac9c255c63f838d352f7eaf32ff957e550ba3a0a4f844aa6207310384460ca2eaef6fa03e16cfc4af9cfd4f21a35c617bace68fb4a8d080979f34a3eea661f670ad43bb2095c7a4e4d9749cdfe7a146362414105eb656aad3c402bc36698bc84c1d1786bd64b2405487188922aad1e49825b9761adf683f73a7e84d56ed77fde3617bc55d0a57244148893c77c0ebe8a15ae33a54f5141f196c0fc66736a7926388de9399c0521f958158b4589e5de747e2d2141bb197fbc35dda2677881065013c6a2ca42b2025987733c8e9e4aace3003a15094ec935171eae98ee3a4be0ba1c47436ef035a3debbdc662dfa8d5bdbe3a3148df9a026246c00d2b7be48d3cdf19536386efbd4dc15d12739e0789a2ce1330b97f5c7f35c7cf6c36eabacbc9aa4fadf23fed088f280180cb525cae921d6300b3a9c771daeb5c2b633722f05a8f89a19ed4447e5c57f95122cd6db018d8967b0669e183f0d314fc92998352408af7ece1f7a637908eda91d30f6c365f1e83d53a840e422944194503021497a6c42f06077c257aa0bc25aad8b5cf713c066e043379c4eeac1322adf8e88a4f8dfaa0277cffe918ff08f7723deb1c74021b0d40d5a9d0e90bfcb1c209f227b45ba8888ade9cf4c984d0f3496b344cf1a1cf3654de7ebbc3812d4cc794c942b0c8b2c016358c96f1d578528476cf46144b3fbb53f3c7a49a9bf29ecf871a6bfb290eab40efeb007ca995413b57ca56f76d3a4b8c99a1e6c4ca9850539db4faeaa54b38b800e160377021c2dd4f37d2aa5373d8acb732c82dc897459867f7cba21fb34e3bbad9e9c5a94139e35dbff23dac8926fae52ebfc520ead6aff62c87487be29aec4146ec6eb91f52c32e912f02a74c2d9791da3b519a98c0de151a1b0e204e8d23145054d2f2c6416966a65f751353aabebaaf4da5123ff0c9987f17c5e1ea492a24c94068d28587bf8ce08ad086b20eb085335693ecc30bb3a8a677e4c1203293fd395b113e92143ae4b3c30b4d8162fea5be04343c2bb8419f59ac49d5034bfcb2e9d5df04c9b7883e54b42a522395fa6a2ba0b60b418718ebad76d978d31e4eecb419c32f1c4a5227267753420be0bf5ca183fb40076f68b3e0dabbf8a9a57730c4c10c5d41cb183d825ec81ba9ec884ded75b2a85b0a51c1f3f6dd4c18e80f9cba8c0518855ea87710cd896159e5971b54db4b0c694f655e424d2f1cc6dc8980d5ec0d877dfd4a4efeb529168f12546b9ec52136888c88ade85df8acf6f91fd77ee3cb799b188c3171c8ad2c834e41ccdbc22ebe4ef1a5a08997b9059cf44893fe2da9fc0b911296062f2f9d78655093064d303ce68119581032ca5a7d9ff0a72c6ae622c654dec214dcd60244f66d180c7c855558d7173df94b6acf7c87697bcb39a5f062d591f66ecae84f44077e7576d782f22777af70c8a34224f34a12fa54439fbe03b3c403082278f48513f4f8bd9e25bfe1df53454edad3fccb96bf5c506a07cc3450db9cffd27c8b2ffe83f6eb88423aa9095fe17f33277bdbc02215e3df4a2afbdf81200808421d076bdcd6405427f8b654e5c2e30bf6970d9ad4821bd185e14ea5a630b53aba6a6852752a679ab6a5d7d09bfe48e95b744e471336ccd53643f866dbc08093b06c85a3b0529fabab63bf348083fd14e8617bbf091ab6a16c70f8324f6493c644432b96da7752c44e94d3bb873f1d4eab123864a7968075064f8eaab0ab8812d916871530cbe37d057ee3a56af8de23fc2590680abd6cdf6e4c97c13ebe9c6e5ebd7a3b4119c8704be5e02b6e100d34abd2986e2c48b5dfaa0779e6048ffa9929f433a4cf4551e82f63c9b65c6a7aece3992719112d53cedcb7d41bc307085a27555eca3c7d6085a7338475f0344948a77fd62cb06dcbcb8c0cf13147b396a9864a11d62b4db69aea8bb213cb2f97e280153b565890f003844443097c0b39374e0278ccf6750617128c51b658860a7133640fe0833aec74f7f871f00d1d5d11040561d79501ab7f7e924e4a28cfb65de0c954943f6adfa5c331c167ab90508776414712da5ef502aed5bf5d9092b1c2e30a1fb517af63604ccb921d0985f546653a258dd0465ba2ab8fdc1657990880bfd010a7ecafaeaff7a0976fd6dbc18b507f74d134e5809a9dbd3d3c61c7555673fbb2c3ce4047726d5776f3c175403dd22438e8023395e61b9089933f14c2c5f3a794244cfcb0370af6b8a879714751ef288f94311eaad924348eec58bc52732aa4ddc84507b606750909166b5ffe3197eb8755a127444ebeb484a1a21057d9ce799424d07f905659ed3f7716d35210d605292f4e152e15c36523964ecae60d331999372947e00ee346dd2c53575886fc636293d66fc32faea001d0d28d51848de0c583ed79e0f5aa7971d06b31d7f87ae734306b7fbf069d7986a7298f9b471e83db1b95f52f7ae3fc7b658d45cc6c2c4b5d95e07f491a06249c2d1b526f51bbb48238c39c1c91e5a4f6b7d7d72db9ed4474c0f98c5a5a12b6d417f3d3f0a9edcb3e7debeb87ec0a95ef8da1f48e99999f928c3a210595074012fe32dfc3ec4f0b88e07e82a89b36aa3b3a3ec7279cdb7bbcc4f02f79b5b6ee9b23a1579443767f9099529efd1825bbf6cdaaf52bb130a4e01fe10e7702eb563478277993aa042f5ab39d5707ee55379a88309ca980ce021d64692267c0b923a49ce191c036a9fd1f3b5da2e40bcc9781e79e57250e6292e3d5b489cc874168bbcd609ca00006d38f1300a23a4f3b9fd249836937d7cf9fb6f0d1a313ca815c8d3dff0213570eb63e8dc15c83434be08302b263c6e94a974fd2919b0c410734b3d4ec98b6ffd03b2be1d0ccf341127715aa4b796373561c8c17396e1b9a2334ef1e878d350d8241d9b32bd8311e8cd41e80a7435a66eb34dd18dbe1ed35c2caf3801f53b771a4b854aa3c9e815261995e1c842825292674995826da25542a12c092d44748e76deb23a9739edf71564d5d72ff62e3e5c3ea51c57fd2389bd663f85794004b9face367af5f2981e5596149f6ed67b766ce3c185de7333d237fdfcad99df3629d28622769611a5a72b672cc1e0d28e2b05e39915159087c083f5b83daa4f1ca4fb3405a29dc6b8f41904a7e272c5fce37ec8e29313a44f0f804bf417459da42065063c3877062f96cb8c1d3a4115113e95d38e68ca2f07246a96f77d1cd9971fa03e3ce781971a9f088e460072c451fb36232cf015093fba2e11347e7389cfdf1531ca079901a64c51495045a4f5e86aff5b7724533eadfce98127b4810438b79e9c750720e0b5cd0323d3081c5ebbfe3bc024e342986b88776fec77a6ef379e1dda9ad5cf5426c0af821ca986a0100352be1ac1d35cef9f2282a655046ef0cd1b002683db007bf4330c5f1994c239ae43e3f70a2d065be88266c5f7cd5b1530672571c97e08fce74cdba1908a1c48f3d19e5261799dacf106052966d76a26c5830eed3e3ef5620a1c435282ecfcfa3fbf4935f3c37546d7b97c5c2a5a0c009d2674741c788413345309e487111035b10086ec82da9d722b3905d5db535d287bde77d73442e87fb3ede2a561528f2a7d48ce6b98202e5f8eb5843b1c1a4cc452587f145cfa38ae33742b35ccb516f73cc90090c721ca4e4e522a2d3045d63b69566334434a1b62e2baf1b142eabd62dd6113999e10e1b4d6305e5121c1a7ceb901fe8e31e011dc7cbfe8ca913f3d267d0fe6ac1e7cb10dd903ee83f51259f0f83770594d7f667ed85c646ab0c94a156f8068e1b441e9e3c810d244510ec04943e9d93e9ac4152b720228489e4a4fbda7a7b0baf437105e4f9868e4ded50e9abfc293c8f9ab6784c255ebec495d05a872f4381b452b628f1f705bc8dbd2e1a6f4fde5c42bf65b93ac6dce1bbb00d5726d80e7fa429fac664b225fb78e56573a184732fd72947034ad8265c7b6bde1782e6f679cfcd4cab903b9a6104fa07f3425a14e60f96ed92278936d76add20720331b3b02380149483e7f68328b6a69375c43053917cd416119da7f546e456714d14cd4dbfe72413fcd79a26ede2c71cd130b9dfb2b353b80bd54766ed3e6239304dc2eeea33b1fcbfa9de4673e1aff1c3baec9ee352012fd6f22643428fb0ac4f043aea4ac7c2d6368597428ebb50201b3fbf5601294f7a6f6fbf3da60bfcace75140ecc81233f9da2e1eeafef7d71edcb8242968a485de10dd1a5026b7ef501ed7ddd559c569af6a7450c30e0a85dd2b3831acbe655a756b22bcf655dae0c9ffae8f4efe1324244260333c5d9161605e992f2b1784b9c93a326c3f0319a9f4b1ba6a994771041614aa2abcc9f10d4a70b2d5c95dbe246d4d5470c390d4caaaee36ee732347ea5b0dbe4ef97d0b9d2a446b78f577acc247dab3863ef539b902a47e32f9813a54f6c9a1fa04ab7bb7dd2e34924091ee6e1f52da02f13c16d46ff15d177361b5984ca36418f928910aa75fb73f08be8ebe9533015d3b11c45023bd54c4f91100f6e615f894f947a0f26741b6377f1eec7ef001e5a1ac713a927df62c7959380f449ddd7455d15ac088e73c0739240ac930fac4d9bac7fdf1ca3eeeda23e51033b5636e4fc55aa24c57fed495dd72ad502c563431bafbd0a3acdbd532c64df0dd0396f67947020742de75079d4178fb3f66dc4464e4bc20028c3cdbd9a4df6e898a9dc2c1c0b92d540255f0daf7f934fde9503c7268ef519a4c957ab433891c13e716645210fe80bfef39facd92313289002421ea9f0aa98a62c40353eb1d1e717327f9ffaefe245aac5a8617ae0a22c83d9ecfeffd756f899339c425ae610ddd9afe5aed3c07ec440916ca054427094f47bfd5d93b83d648030ed912686289d222417c2a8485f1b758c6e7eea85576e5160783bdeedade9d113f7b0f03f76642ff7ac89c0c8e8e395763aeb00b2b0c8cc7507225db4210bcfab834fef62457bcc5efa4a31f2ea8ace28ef9db79063712638b55949208ea52f8a7020f8028396f2662dccc7449b720c3c71a9488c87640577306b6b7ee0e03eec6c748890a8b2e39400b634ee2b4f82069f5ff030c83f4679a7add11e6711ca4212dceb051be9eac54c48c3677fad77778ceac60f2ee76895339214966d96188eee7891b904457ad42855527fd9b0bebc05d327e02a2d73aad59663eb217cd"}) r2 = socket$inet6(0xa, 0x2, 0x0) ioctl$sock_SIOCETHTOOL(r2, 0x89f0, &(0x7f0000000000)={'bridge0\x00', &(0x7f0000000200)=@ethtool_ringparam={0x10}}) [ 2870.766441][ T4219] netlink: 'syz-executor.1': attribute type 1 has an invalid length. [ 2870.925419][ T4219] bond1327: entered promiscuous mode [ 2870.944018][ T4219] 8021q: adding VLAN 0 to HW filter on device bond1327 [ 2871.046543][ T4221] bond1327: (slave bridge1273): making interface the new active one [ 2871.060834][ T4221] bridge1273: entered promiscuous mode [ 2871.123359][ T4221] bond1327: (slave bridge1273): Enslaving as an active interface with an up link 04:13:18 executing program 1: r0 = socket$netlink(0x10, 0x3, 0x0) r1 = socket$netlink(0x10, 0x3, 0x0) r2 = socket(0x10, 0x803, 0x0) sendmsg$NL80211_CMD_CRIT_PROTOCOL_START(r2, &(0x7f0000000580)={0x0, 0x0, &(0x7f0000000540)={0x0, 0x1c}}, 0x0) getsockname$packet(r2, &(0x7f0000000600)={0x11, 0x0, 0x0, 0x1, 0x0, 0x6, @broadcast}, &(0x7f0000000100)=0x14) sendmsg$nl_route(r1, &(0x7f0000000040)={0x0, 0x0, &(0x7f0000000000)={&(0x7f00000000c0)=ANY=[@ANYBLOB="3c00000010008506eb9afc4cd8d06e754a0081c5", @ANYRES32=r3, @ANYBLOB="2377f292252155b21c0012000c000100626f6e64000000000c0002000800010001"], 0x3c}}, 0x0) sendmsg$nl_route(r0, &(0x7f0000000240)={0x0, 0x0, &(0x7f0000000680)={&(0x7f0000000400)=@newlink={0x3c, 0x10, 0xffffff1f, 0x0, 0x34672500, {}, [@IFLA_LINKINFO={0x14, 0x12, 0x0, 0x1, @bridge={{0xb}, {0x4}}}, @IFLA_MASTER={0x8, 0xa, r3}]}, 0x3c}}, 0x0) 04:13:18 executing program 3: r0 = socket$netlink(0x10, 0x3, 0x0) sendmsg$netlink(r0, &(0x7f00000047c0)={0x0, 0x0, &(0x7f0000000000)=[{&(0x7f0000000100)=ANY=[@ANYBLOB="1800000056000106000000a700db730007"], 0x18}], 0x1}, 0x0) 04:13:18 executing program 0: r0 = socket$netlink(0x10, 0x3, 0x0) sendmsg$netlink(r0, &(0x7f00000047c0)={0x0, 0x0, &(0x7f0000000000)=[{&(0x7f0000000100)=ANY=[@ANYBLOB="18000000560001"], 0x18}], 0x1}, 0x0) [ 2871.220109][ T4228] netlink: 'syz-executor.2': attribute type 1 has an invalid length. 04:13:18 executing program 0: r0 = socket$netlink(0x10, 0x3, 0x0) sendmsg$netlink(r0, &(0x7f00000047c0)={0x0, 0x0, &(0x7f0000000000)=[{&(0x7f0000000100)=ANY=[@ANYBLOB="18000000560001"], 0x18}], 0x1}, 0x0) [ 2871.299311][ T4228] bond1312: entered promiscuous mode [ 2871.305013][ T4228] 8021q: adding VLAN 0 to HW filter on device bond1312 04:13:18 executing program 0: r0 = socket$netlink(0x10, 0x3, 0x0) sendmsg$netlink(r0, &(0x7f00000047c0)={0x0, 0x0, &(0x7f0000000000)=[{&(0x7f0000000100)=ANY=[@ANYBLOB="18000000560001"], 0x18}], 0x1}, 0x0) [ 2871.388270][ T4231] bond1312: (slave bridge1232): making interface the new active one [ 2871.408765][ T4231] bridge1232: entered promiscuous mode [ 2871.419154][ T4231] bond1312: (slave bridge1232): Enslaving as an active interface with an up link 04:13:18 executing program 2: r0 = socket$netlink(0x10, 0x3, 0x0) r1 = socket$netlink(0x10, 0x3, 0x0) r2 = socket(0x10, 0x803, 0x0) sendmsg$NL80211_CMD_CRIT_PROTOCOL_START(r2, &(0x7f0000000580)={0x0, 0x0, &(0x7f0000000540)={0x0, 0x1c}}, 0x0) getsockname$packet(r2, &(0x7f0000000600)={0x11, 0x0, 0x0, 0x1, 0x0, 0x6, @broadcast}, &(0x7f0000000100)=0x14) sendmsg$nl_route(r1, &(0x7f0000000040)={0x0, 0x0, &(0x7f0000000000)={&(0x7f00000000c0)=ANY=[@ANYBLOB="3c00000010008506eb9afc4cd8d06e754a0081c5", @ANYRES32=r3, @ANYBLOB="2377f292252155b21c0012000c000100626f6e64000000000c0002000800010001"], 0x3c}}, 0x0) sendmsg$nl_route(r0, &(0x7f0000000240)={0x0, 0x0, &(0x7f0000000680)={&(0x7f0000000400)=@newlink={0x3c, 0x10, 0xffffff1f, 0x0, 0x53c2500, {}, [@IFLA_LINKINFO={0x14, 0x12, 0x0, 0x1, @bridge={{0xb}, {0x4}}}, @IFLA_MASTER={0x8, 0xa, r3}]}, 0x3c}}, 0x0) 04:13:18 executing program 0: r0 = socket$netlink(0x10, 0x3, 0x0) sendmsg$netlink(r0, &(0x7f00000047c0)={0x0, 0x0, &(0x7f0000000000)=[{&(0x7f0000000100)=ANY=[@ANYBLOB="18000000560001"], 0x18}], 0x1}, 0x0) 04:13:18 executing program 4: r0 = socket$inet6(0xa, 0x6, 0x7de) ioctl$sock_SIOCETHTOOL(r0, 0x89f0, &(0x7f0000000000)={'veth0_to_bridge\x00', &(0x7f0000000040)=ANY=[@ANYBLOB="0500000002000000450400000000000800000000"]}) [ 2871.441676][ T4246] netlink: 'syz-executor.5': attribute type 1 has an invalid length. 04:13:19 executing program 0: sendmsg$netlink(0xffffffffffffffff, &(0x7f00000047c0)={0x0, 0x0, &(0x7f0000000000)=[{&(0x7f0000000100)=ANY=[@ANYBLOB="18000000560001"], 0x18}], 0x1}, 0x0) [ 2871.576192][ T4246] bond1318: entered promiscuous mode [ 2871.596307][ T4246] 8021q: adding VLAN 0 to HW filter on device bond1318 [ 2871.719185][ T4247] bond1318: (slave bridge1280): making interface the new active one [ 2871.738818][ T4247] bridge1280: entered promiscuous mode 04:13:19 executing program 5: r0 = socket$netlink(0x10, 0x3, 0x0) r1 = socket$netlink(0x10, 0x3, 0x0) r2 = socket(0x10, 0x803, 0x0) sendmsg$NL80211_CMD_CRIT_PROTOCOL_START(r2, &(0x7f0000000580)={0x0, 0x0, &(0x7f0000000540)={0x0, 0x1c}}, 0x0) getsockname$packet(r2, &(0x7f0000000600)={0x11, 0x0, 0x0, 0x1, 0x0, 0x6, @broadcast}, &(0x7f0000000100)=0x14) sendmsg$nl_route(r1, &(0x7f0000000040)={0x0, 0x0, &(0x7f0000000000)={&(0x7f00000000c0)=ANY=[@ANYBLOB="3c00000010008506eb9afc4cd8d06e754a0081c5", @ANYRES32=r3, @ANYBLOB="2377f292252155b21c0012000c000100626f6e64000000000c0002000800010001"], 0x3c}}, 0x0) sendmsg$nl_route(r0, &(0x7f0000000240)={0x0, 0x0, &(0x7f0000000680)={&(0x7f0000000400)=@newlink={0x3c, 0x10, 0xffffff1f, 0x0, 0x7d352500, {}, [@IFLA_LINKINFO={0x14, 0x12, 0x0, 0x1, @bridge={{0xb}, {0x4}}}, @IFLA_MASTER={0x8, 0xa, r3}]}, 0x3c}}, 0x0) 04:13:19 executing program 4: r0 = socket$inet6(0xa, 0x6, 0x7de) ioctl$sock_SIOCETHTOOL(r0, 0x89f0, &(0x7f0000000000)={'veth0_to_bridge\x00', &(0x7f0000000040)=ANY=[@ANYBLOB="0500000002000000450400000000000800000000"]}) 04:13:19 executing program 0: sendmsg$netlink(0xffffffffffffffff, &(0x7f00000047c0)={0x0, 0x0, &(0x7f0000000000)=[{&(0x7f0000000100)=ANY=[@ANYBLOB="18000000560001"], 0x18}], 0x1}, 0x0) [ 2871.795597][ T4247] bond1318: (slave bridge1280): Enslaving as an active interface with an up link [ 2871.831740][ T4254] netlink: 'syz-executor.1': attribute type 1 has an invalid length. [ 2872.000371][ T4254] bond1328: entered promiscuous mode [ 2872.022979][ T4254] 8021q: adding VLAN 0 to HW filter on device bond1328 [ 2872.130525][ T4258] bond1328: (slave bridge1274): making interface the new active one [ 2872.169956][ T4258] bridge1274: entered promiscuous mode 04:13:19 executing program 1: r0 = socket$netlink(0x10, 0x3, 0x0) r1 = socket$netlink(0x10, 0x3, 0x0) r2 = socket(0x10, 0x803, 0x0) sendmsg$NL80211_CMD_CRIT_PROTOCOL_START(r2, &(0x7f0000000580)={0x0, 0x0, &(0x7f0000000540)={0x0, 0x1c}}, 0x0) getsockname$packet(r2, &(0x7f0000000600)={0x11, 0x0, 0x0, 0x1, 0x0, 0x6, @broadcast}, &(0x7f0000000100)=0x14) sendmsg$nl_route(r1, &(0x7f0000000040)={0x0, 0x0, &(0x7f0000000000)={&(0x7f00000000c0)=ANY=[@ANYBLOB="3c00000010008506eb9afc4cd8d06e754a0081c5", @ANYRES32=r3, @ANYBLOB="2377f292252155b21c0012000c000100626f6e64000000000c0002000800010001"], 0x3c}}, 0x0) sendmsg$nl_route(r0, &(0x7f0000000240)={0x0, 0x0, &(0x7f0000000680)={&(0x7f0000000400)=@newlink={0x3c, 0x10, 0xffffff1f, 0x0, 0x35672500, {}, [@IFLA_LINKINFO={0x14, 0x12, 0x0, 0x1, @bridge={{0xb}, {0x4}}}, @IFLA_MASTER={0x8, 0xa, r3}]}, 0x3c}}, 0x0) 04:13:19 executing program 3: r0 = socket$netlink(0x10, 0x3, 0x0) sendmsg$netlink(r0, &(0x7f00000047c0)={0x0, 0x0, &(0x7f0000000000)=[{&(0x7f0000000100)=ANY=[@ANYBLOB="1800000056000106000000a800db730007"], 0x18}], 0x1}, 0x0) 04:13:19 executing program 0: sendmsg$netlink(0xffffffffffffffff, &(0x7f00000047c0)={0x0, 0x0, &(0x7f0000000000)=[{&(0x7f0000000100)=ANY=[@ANYBLOB="18000000560001"], 0x18}], 0x1}, 0x0) 04:13:19 executing program 4: r0 = socket$inet6(0xa, 0x6, 0x7de) ioctl$sock_SIOCETHTOOL(r0, 0x89f0, &(0x7f0000000000)={'veth0_to_bridge\x00', &(0x7f0000000040)=ANY=[@ANYBLOB="0500000002000000450400000000000800000000"]}) [ 2872.180842][ T4258] bond1328: (slave bridge1274): Enslaving as an active interface with an up link 04:13:19 executing program 4: getsockopt$inet_sctp6_SCTP_DEFAULT_SEND_PARAM(0xffffffffffffffff, 0x84, 0xa, &(0x7f0000000180)={0x6, 0x101, 0x1, 0xffff6eeb, 0x8000, 0x8, 0x4, 0x5748, 0x0}, &(0x7f00000001c0)=0x20) setsockopt$inet_sctp_SCTP_AUTH_DELETE_KEY(0xffffffffffffffff, 0x84, 0x19, &(0x7f0000000240)={r0, 0x7f}, 0x8) r1 = socket$inet6(0xa, 0x2, 0x0) ioctl$ifreq_SIOCGIFINDEX_vcan(r1, 0x8933, &(0x7f0000000040)={'vxcan0\x00', 0x0}) setsockopt$inet6_IPV6_XFRM_POLICY(r1, 0x29, 0x23, &(0x7f0000000080)={{{@in6=@mcast2, @in=@private=0xa010100, 0x4e24, 0x3, 0x4e24, 0x9, 0x8, 0x20, 0xa0, 0x87, r2}, {0x6, 0x8, 0x6, 0x9, 0x1, 0x2, 0x2, 0xe6a}, {0x7, 0x0, 0x4, 0x2999}, 0x7fffffff, 0x6e6bb0, 0x1, 0x0, 0x2, 0x2}, {{@in=@private=0xa010101, 0x4d3, 0x33}, 0xa, @in=@dev={0xac, 0x14, 0x14, 0xf}, 0x3504, 0x2, 0x2, 0xf7, 0x101, 0x6, 0x8}}, 0xe8) ioctl$sock_SIOCETHTOOL(r1, 0x89f0, &(0x7f0000000000)={'bridge0\x00', &(0x7f0000000200)=@ethtool_ringparam={0x10}}) [ 2872.275862][ T4269] netlink: 'syz-executor.2': attribute type 1 has an invalid length. [ 2872.486584][ T4269] bond1313: entered promiscuous mode [ 2872.495636][ T4269] 8021q: adding VLAN 0 to HW filter on device bond1313 [ 2872.580759][ T4272] bond1313: (slave bridge1233): making interface the new active one [ 2872.619265][ T4272] bridge1233: entered promiscuous mode 04:13:20 executing program 2: r0 = socket$netlink(0x10, 0x3, 0x0) r1 = socket$netlink(0x10, 0x3, 0x0) r2 = socket(0x10, 0x803, 0x0) sendmsg$NL80211_CMD_CRIT_PROTOCOL_START(r2, &(0x7f0000000580)={0x0, 0x0, &(0x7f0000000540)={0x0, 0x1c}}, 0x0) getsockname$packet(r2, &(0x7f0000000600)={0x11, 0x0, 0x0, 0x1, 0x0, 0x6, @broadcast}, &(0x7f0000000100)=0x14) sendmsg$nl_route(r1, &(0x7f0000000040)={0x0, 0x0, &(0x7f0000000000)={&(0x7f00000000c0)=ANY=[@ANYBLOB="3c00000010008506eb9afc4cd8d06e754a0081c5", @ANYRES32=r3, @ANYBLOB="2377f292252155b21c0012000c000100626f6e64000000000c0002000800010001"], 0x3c}}, 0x0) sendmsg$nl_route(r0, &(0x7f0000000240)={0x0, 0x0, &(0x7f0000000680)={&(0x7f0000000400)=@newlink={0x3c, 0x10, 0xffffff1f, 0x0, 0x6000000, {}, [@IFLA_LINKINFO={0x14, 0x12, 0x0, 0x1, @bridge={{0xb}, {0x4}}}, @IFLA_MASTER={0x8, 0xa, r3}]}, 0x3c}}, 0x0) 04:13:20 executing program 0: socket$netlink(0x10, 0x3, 0x0) sendmsg$netlink(0xffffffffffffffff, &(0x7f00000047c0)={0x0, 0x0, &(0x7f0000000000)=[{&(0x7f0000000100)=ANY=[@ANYBLOB="18000000560001"], 0x18}], 0x1}, 0x0) [ 2872.640826][ T4272] bond1313: (slave bridge1233): Enslaving as an active interface with an up link 04:13:20 executing program 0: socket$netlink(0x10, 0x3, 0x0) sendmsg$netlink(0xffffffffffffffff, &(0x7f00000047c0)={0x0, 0x0, &(0x7f0000000000)=[{&(0x7f0000000100)=ANY=[@ANYBLOB="18000000560001"], 0x18}], 0x1}, 0x0) [ 2872.779032][ T4280] bond1319: entered promiscuous mode [ 2872.804385][ T4280] 8021q: adding VLAN 0 to HW filter on device bond1319 04:13:20 executing program 0: socket$netlink(0x10, 0x3, 0x0) sendmsg$netlink(0xffffffffffffffff, &(0x7f00000047c0)={0x0, 0x0, &(0x7f0000000000)=[{&(0x7f0000000100)=ANY=[@ANYBLOB="18000000560001"], 0x18}], 0x1}, 0x0) [ 2873.035070][ T4281] bond1319: (slave bridge1281): making interface the new active one [ 2873.067572][ T4281] bridge1281: entered promiscuous mode 04:13:20 executing program 5: r0 = socket$netlink(0x10, 0x3, 0x0) r1 = socket$netlink(0x10, 0x3, 0x0) r2 = socket(0x10, 0x803, 0x0) sendmsg$NL80211_CMD_CRIT_PROTOCOL_START(r2, &(0x7f0000000580)={0x0, 0x0, &(0x7f0000000540)={0x0, 0x1c}}, 0x0) getsockname$packet(r2, &(0x7f0000000600)={0x11, 0x0, 0x0, 0x1, 0x0, 0x6, @broadcast}, &(0x7f0000000100)=0x14) sendmsg$nl_route(r1, &(0x7f0000000040)={0x0, 0x0, &(0x7f0000000000)={&(0x7f00000000c0)=ANY=[@ANYBLOB="3c00000010008506eb9afc4cd8d06e754a0081c5", @ANYRES32=r3, @ANYBLOB="2377f292252155b21c0012000c000100626f6e64000000000c0002000800010001"], 0x3c}}, 0x0) sendmsg$nl_route(r0, &(0x7f0000000240)={0x0, 0x0, &(0x7f0000000680)={&(0x7f0000000400)=@newlink={0x3c, 0x10, 0xffffff1f, 0x0, 0x7e352500, {}, [@IFLA_LINKINFO={0x14, 0x12, 0x0, 0x1, @bridge={{0xb}, {0x4}}}, @IFLA_MASTER={0x8, 0xa, r3}]}, 0x3c}}, 0x0) 04:13:20 executing program 3: r0 = socket$netlink(0x10, 0x3, 0x0) sendmsg$netlink(r0, &(0x7f00000047c0)={0x0, 0x0, &(0x7f0000000000)=[{&(0x7f0000000100)=ANY=[@ANYBLOB="1800000056000106000000a900db730007"], 0x18}], 0x1}, 0x0) [ 2873.095573][ T4281] bond1319: (slave bridge1281): Enslaving as an active interface with an up link 04:13:20 executing program 0: r0 = socket$netlink(0x10, 0x3, 0x0) sendmsg$netlink(r0, 0x0, 0x0) 04:13:20 executing program 4: getsockopt$inet_sctp6_SCTP_DEFAULT_SEND_PARAM(0xffffffffffffffff, 0x84, 0xa, &(0x7f0000000180)={0x6, 0x101, 0x1, 0xffff6eeb, 0x8000, 0x8, 0x4, 0x5748, 0x0}, &(0x7f00000001c0)=0x20) setsockopt$inet_sctp_SCTP_AUTH_DELETE_KEY(0xffffffffffffffff, 0x84, 0x19, &(0x7f0000000240)={r0, 0x7f}, 0x8) (async) r1 = socket$inet6(0xa, 0x2, 0x0) ioctl$ifreq_SIOCGIFINDEX_vcan(r1, 0x8933, &(0x7f0000000040)={'vxcan0\x00', 0x0}) setsockopt$inet6_IPV6_XFRM_POLICY(r1, 0x29, 0x23, &(0x7f0000000080)={{{@in6=@mcast2, @in=@private=0xa010100, 0x4e24, 0x3, 0x4e24, 0x9, 0x8, 0x20, 0xa0, 0x87, r2}, {0x6, 0x8, 0x6, 0x9, 0x1, 0x2, 0x2, 0xe6a}, {0x7, 0x0, 0x4, 0x2999}, 0x7fffffff, 0x6e6bb0, 0x1, 0x0, 0x2, 0x2}, {{@in=@private=0xa010101, 0x4d3, 0x33}, 0xa, @in=@dev={0xac, 0x14, 0x14, 0xf}, 0x3504, 0x2, 0x2, 0xf7, 0x101, 0x6, 0x8}}, 0xe8) ioctl$sock_SIOCETHTOOL(r1, 0x89f0, &(0x7f0000000000)={'bridge0\x00', &(0x7f0000000200)=@ethtool_ringparam={0x10}}) 04:13:20 executing program 1: r0 = socket$netlink(0x10, 0x3, 0x0) r1 = socket$netlink(0x10, 0x3, 0x0) r2 = socket(0x10, 0x803, 0x0) sendmsg$NL80211_CMD_CRIT_PROTOCOL_START(r2, &(0x7f0000000580)={0x0, 0x0, &(0x7f0000000540)={0x0, 0x1c}}, 0x0) getsockname$packet(r2, &(0x7f0000000600)={0x11, 0x0, 0x0, 0x1, 0x0, 0x6, @broadcast}, &(0x7f0000000100)=0x14) sendmsg$nl_route(r1, &(0x7f0000000040)={0x0, 0x0, &(0x7f0000000000)={&(0x7f00000000c0)=ANY=[@ANYBLOB="3c00000010008506eb9afc4cd8d06e754a0081c5", @ANYRES32=r3, @ANYBLOB="2377f292252155b21c0012000c000100626f6e64000000000c0002000800010001"], 0x3c}}, 0x0) sendmsg$nl_route(r0, &(0x7f0000000240)={0x0, 0x0, &(0x7f0000000680)={&(0x7f0000000400)=@newlink={0x3c, 0x10, 0xffffff1f, 0x0, 0x36672500, {}, [@IFLA_LINKINFO={0x14, 0x12, 0x0, 0x1, @bridge={{0xb}, {0x4}}}, @IFLA_MASTER={0x8, 0xa, r3}]}, 0x3c}}, 0x0) 04:13:20 executing program 0: r0 = socket$netlink(0x10, 0x3, 0x0) sendmsg$netlink(r0, 0x0, 0x0) 04:13:20 executing program 0: r0 = socket$netlink(0x10, 0x3, 0x0) sendmsg$netlink(r0, 0x0, 0x0) 04:13:20 executing program 0: r0 = socket$netlink(0x10, 0x3, 0x0) sendmsg$netlink(r0, &(0x7f00000047c0)={0x0, 0x0, 0x0}, 0x0) [ 2873.137484][ T4290] workqueue: Failed to create a rescuer kthread for wq "bond1329": -EINTR [ 2873.431980][ T4300] bond1314: entered promiscuous mode [ 2873.464092][ T4300] 8021q: adding VLAN 0 to HW filter on device bond1314 04:13:21 executing program 0: r0 = socket$netlink(0x10, 0x3, 0x0) sendmsg$netlink(r0, &(0x7f00000047c0)={0x0, 0x0, 0x0}, 0x0) 04:13:21 executing program 3: r0 = socket$netlink(0x10, 0x3, 0x0) sendmsg$netlink(r0, &(0x7f00000047c0)={0x0, 0x0, &(0x7f0000000000)=[{&(0x7f0000000100)=ANY=[@ANYBLOB="1800000056000106000000aa00db730007"], 0x18}], 0x1}, 0x0) 04:13:21 executing program 2: r0 = socket$netlink(0x10, 0x3, 0x0) r1 = socket$netlink(0x10, 0x3, 0x0) r2 = socket(0x10, 0x803, 0x0) sendmsg$NL80211_CMD_CRIT_PROTOCOL_START(r2, &(0x7f0000000580)={0x0, 0x0, &(0x7f0000000540)={0x0, 0x1c}}, 0x0) getsockname$packet(r2, &(0x7f0000000600)={0x11, 0x0, 0x0, 0x1, 0x0, 0x6, @broadcast}, &(0x7f0000000100)=0x14) sendmsg$nl_route(r1, &(0x7f0000000040)={0x0, 0x0, &(0x7f0000000000)={&(0x7f00000000c0)=ANY=[@ANYBLOB="3c00000010008506eb9afc4cd8d06e754a0081c5", @ANYRES32=r3, @ANYBLOB="2377f292252155b21c0012000c000100626f6e64000000000c0002000800010001"], 0x3c}}, 0x0) sendmsg$nl_route(r0, &(0x7f0000000240)={0x0, 0x0, &(0x7f0000000680)={&(0x7f0000000400)=@newlink={0x3c, 0x10, 0xffffff1f, 0x0, 0x63c2500, {}, [@IFLA_LINKINFO={0x14, 0x12, 0x0, 0x1, @bridge={{0xb}, {0x4}}}, @IFLA_MASTER={0x8, 0xa, r3}]}, 0x3c}}, 0x0) 04:13:21 executing program 0: r0 = socket$netlink(0x10, 0x3, 0x0) sendmsg$netlink(r0, &(0x7f00000047c0)={0x0, 0x0, 0x0}, 0x0) [ 2873.588781][ T4313] bond1320: entered promiscuous mode [ 2873.613577][ T4313] 8021q: adding VLAN 0 to HW filter on device bond1320 [ 2873.817642][ T4315] bond1320: (slave bridge1282): making interface the new active one [ 2873.828190][ T4315] bridge1282: entered promiscuous mode 04:13:21 executing program 5: r0 = socket$netlink(0x10, 0x3, 0x0) r1 = socket$netlink(0x10, 0x3, 0x0) r2 = socket(0x10, 0x803, 0x0) sendmsg$NL80211_CMD_CRIT_PROTOCOL_START(r2, &(0x7f0000000580)={0x0, 0x0, &(0x7f0000000540)={0x0, 0x1c}}, 0x0) getsockname$packet(r2, &(0x7f0000000600)={0x11, 0x0, 0x0, 0x1, 0x0, 0x6, @broadcast}, &(0x7f0000000100)=0x14) sendmsg$nl_route(r1, &(0x7f0000000040)={0x0, 0x0, &(0x7f0000000000)={&(0x7f00000000c0)=ANY=[@ANYBLOB="3c00000010008506eb9afc4cd8d06e754a0081c5", @ANYRES32=r3, @ANYBLOB="2377f292252155b21c0012000c000100626f6e64000000000c0002000800010001"], 0x3c}}, 0x0) sendmsg$nl_route(r0, &(0x7f0000000240)={0x0, 0x0, &(0x7f0000000680)={&(0x7f0000000400)=@newlink={0x3c, 0x10, 0xffffff1f, 0x0, 0x7f352500, {}, [@IFLA_LINKINFO={0x14, 0x12, 0x0, 0x1, @bridge={{0xb}, {0x4}}}, @IFLA_MASTER={0x8, 0xa, r3}]}, 0x3c}}, 0x0) 04:13:21 executing program 0: r0 = socket$netlink(0x10, 0x3, 0x0) sendmsg$netlink(r0, &(0x7f00000047c0), 0x0) [ 2873.870999][ T4315] bond1320: (slave bridge1282): Enslaving as an active interface with an up link 04:13:21 executing program 0: r0 = socket$netlink(0x10, 0x3, 0x0) sendmsg$netlink(r0, &(0x7f00000047c0), 0x0) [ 2873.964022][ T4323] bond1329: entered promiscuous mode [ 2873.969660][ T4323] 8021q: adding VLAN 0 to HW filter on device bond1329 04:13:21 executing program 4: getsockopt$inet_sctp6_SCTP_DEFAULT_SEND_PARAM(0xffffffffffffffff, 0x84, 0xa, &(0x7f0000000180)={0x6, 0x101, 0x1, 0xffff6eeb, 0x8000, 0x8, 0x4, 0x5748, 0x0}, &(0x7f00000001c0)=0x20) setsockopt$inet_sctp_SCTP_AUTH_DELETE_KEY(0xffffffffffffffff, 0x84, 0x19, &(0x7f0000000240)={r0, 0x7f}, 0x8) r1 = socket$inet6(0xa, 0x2, 0 VM DIAGNOSIS: Warning: Permanently added '10.128.1.122' (ED25519) to the list of known hosts. lock-classes: 6586 [max: 8192] direct dependencies: 52410 [max: 131072] indirect dependencies: 1092242 all direct dependencies: 2556282 dependency chains: 262144 [max: 262144] dependency chain hlocks used: 1303608 [max: 1310720] dependency chain hlocks lost: 0 in-hardirq chains: 103 in-softirq chains: 3417 in-process chains: 258624 stack-trace entries: 332862 [max: 1048576] number of stack traces: 17008 number of stack hash chains: 10550 combined max dependencies:hardirq-safe locks: 57 hardirq-unsafe locks: 5796 softirq-safe locks: 367 softirq-unsafe locks: 5376 irq-safe locks: 375 irq-unsafe locks: 5796 hardirq-read-safe locks: 4 hardirq-read-unsafe locks: 227 softirq-read-safe locks: 25 softirq-read-unsafe locks: 203 irq-read-safe locks: 25 irq-read-unsafe locks: 227 uncategorized locks: 431 unused locks: 0 max locking depth: 18 max bfs queue depth: 732 max lock class index: 6585 debug_locks: 0 zapped classes: 1505 zapped lock chains: 2818 large chain blocks: 1 all lock classes: FD: 34 BD: 1 +.+.: fill_pool_map-wait-type-override ->&zone->lock ->&____s->seqcount ->pool_lock#2 ->&c->lock ->pool_lock ->&rq->__lock ->rcu_node_0 ->&cfs_rq->removed.lock ->&obj_hash[i].lock ->&n->list_lock ->batched_entropy_u8.lock ->kfence_freelist_lock ->&____s->seqcount#2 ->pgd_lock ->stock_lock ->key ->pcpu_lock ->percpu_counters_lock ->&rcu_state.expedited_wq ->init_task.mems_allowed_seq.seqcount FD: 2 BD: 5278 -.-.: &obj_hash[i].lock ->pool_lock FD: 1 BD: 5279 -.-.: pool_lock FD: 899 BD: 17 +.+.: cgroup_mutex ->pcpu_alloc_mutex ->pool_lock#2 ->lock ->&root->kernfs_rwsem ->&obj_hash[i].lock ->cgroup_file_kn_lock ->css_set_lock ->&c->lock ->&____s->seqcount ->blkcg_pol_mutex ->&n->list_lock ->&zone->lock ->percpu_counters_lock ->shrinker_rwsem ->&base->lock ->batched_entropy_u8.lock ->&pgdat->memcg_lru.lock ->devcgroup_mutex ->cpu_hotplug_lock ->fs_reclaim ->rcu_node_0 ->&pool->lock ->&rq->__lock ->cgroup_rstat_lock ->cgroup_mutex.wait_lock ->&rcu_state.expedited_wq ->cpuset_mutex ->&dom->lock ->batched_entropy_u32.lock ->cgroup_idr_lock ->task_group_lock ->(wq_completion)cpuset_migrate_mm ->&wq->mutex ->&____s->seqcount#2 ->kfence_freelist_lock ->rtnl_mutex ->rtnl_mutex.wait_lock ->&p->pi_lock ->remove_cache_srcu ->stock_lock ->per_cpu_ptr(&cgroup_rstat_cpu_lock, cpu) ->&cfs_rq->removed.lock ->krc.lock ->&mm->mmap_lock FD: 1 BD: 4145 -.-.: (console_sem).lock FD: 222 BD: 13 +.+.: console_lock ->console_owner_lock ->resource_lock ->pool_lock#2 ->&obj_hash[i].lock ->&zone->lock ->&____s->seqcount ->&c->lock ->kbd_event_lock ->vga_lock ->(console_sem).lock ->fs_reclaim ->&x->wait#9 ->&k->list_lock ->gdp_mutex ->lock ->&root->kernfs_rwsem ->bus_type_sem ->sysfs_symlink_target_lock ->&dev->power.lock ->dpm_list_mtx ->uevent_sock_mutex ->running_helpers_waitq.lock ->subsys mutex#11 ->&fb_info->lock ->&base->lock ->subsys mutex#5 ->&helper->lock ->&helper->damage_lock ->&lock->wait_lock ->&p->pi_lock ->&rq->__lock ->vt_event_lock FD: 1 BD: 11 ....: console_srcu FD: 286 BD: 141 ++++: cpu_hotplug_lock ->jump_label_mutex ->static_call_mutex ->cpuhp_state_mutex ->wq_pool_mutex ->freezer_mutex ->rcu_tasks_trace__percpu.cbs_pcpu_lock ->&ACCESS_PRIVATE(rtpcp, lock) ->smpboot_threads_lock ->&obj_hash[i].lock ->&pool->lock ->&x->wait#4 ->&rq->__lock ->mem_hotplug_lock ->mem_hotplug_lock.waiters.lock ->mem_hotplug_lock.rss.gp_wait.lock ->cpu_hotplug_lock.rss.gp_wait.lock ->rcu_node_0 ->&swhash->hlist_mutex ->pmus_lock ->pcp_batch_high_lock ->&xa->xa_lock ->fs_reclaim ->pool_lock#2 ->kthread_create_lock ->&p->pi_lock ->&x->wait ->wq_pool_attach_mutex ->pcpu_alloc_mutex ->relay_channels_mutex ->&c->lock ->&n->list_lock ->&zone->lock ->&____s->seqcount ->text_mutex ->free_vmap_area_lock ->vmap_area_lock ->init_mm.page_table_lock ->(console_sem).lock ->rtc_lock ->sparse_irq_lock ->&x->wait#6 ->cpuhp_state-up ->stop_cpus_mutex ->&wq->mutex ->flush_lock ->&md->mutex ->&irq_desc_lock_class ->xps_map_mutex ->css_set_lock ->cpuset_mutex ->cgroup_threadgroup_rwsem ->cgroup_threadgroup_rwsem.waiters.lock ->cgroup_threadgroup_rwsem.rss.gp_wait.lock ->&list->lock#5 ->jump_label_mutex.wait_lock ->(work_completion)(flush) ->&x->wait#10 ->&____s->seqcount#2 ->&p->alloc_lock FD: 57 BD: 148 +.+.: jump_label_mutex ->text_mutex ->&rq->__lock ->jump_label_mutex.wait_lock ->text_mutex.wait_lock ->&p->pi_lock ->&cfs_rq->removed.lock ->&obj_hash[i].lock ->pool_lock#2 ->rcu_node_0 FD: 56 BD: 142 +.+.: static_call_mutex ->text_mutex ->text_mutex.wait_lock ->&p->pi_lock ->&rq->__lock FD: 55 BD: 154 +.+.: text_mutex ->ptlock_ptr(page)#2 ->&rq->__lock ->text_mutex.wait_lock ->&cfs_rq->removed.lock ->&obj_hash[i].lock ->&pool->lock ->rcu_node_0 ->&rcu_state.expedited_wq ->pool_lock#2 FD: 228 BD: 10 +.+.: console_mutex ->syslog_lock ->(console_sem).lock ->&port_lock_key ->console_lock ->console_srcu_srcu_usage.lock ->&ACCESS_PRIVATE(sdp, lock) ->console_srcu ->&root->kernfs_rwsem ->kernfs_notify_lock ->&rq->__lock FD: 23 BD: 11 +.+.: syslog_lock ->&rq->__lock FD: 1 BD: 4144 -.-.: console_owner_lock FD: 36 BD: 4143 -.-.: console_owner ->console_owner_lock ->&port_lock_key FD: 1 BD: 166 ..-.: input_pool.lock FD: 220 BD: 142 +.+.: cpuhp_state_mutex ->cpuhp_state-down ->cpuhp_state-up ->&p->pi_lock ->&x->wait#6 ->&rq->__lock ->&cfs_rq->removed.lock ->&obj_hash[i].lock ->pool_lock#2 ->fs_reclaim ->lock ->&root->kernfs_rwsem ->&c->lock ->&____s->seqcount ->&zone->lock ->&n->list_lock ->crypto_alg_sem ->scomp_lock FD: 288 BD: 1 +.+.: clocksource_mutex ->watchdog_lock ->cpu_hotplug_lock ->(console_sem).lock ->&rq->__lock FD: 1 BD: 2 ....: watchdog_lock FD: 10 BD: 166 ++++: resource_lock ->&c->lock ->&____s->seqcount ->pool_lock#2 ->&obj_hash[i].lock FD: 1 BD: 1 ....: cache_disable_lock FD: 1 BD: 4418 +.+.: pgd_lock FD: 26 BD: 309 +.+.: init_mm.page_table_lock ->pgd_lock ->&obj_hash[i].lock FD: 1 BD: 1 ....: early_pfn_lock FD: 173 BD: 1 +.+.: acpi_ioapic_lock ->ioapic_lock ->(console_sem).lock ->ioapic_mutex FD: 2 BD: 162 ....: ioapic_lock ->i8259A_lock FD: 1 BD: 1 +.+.: syscore_ops_lock FD: 1 BD: 1 ....: map_entries_lock FD: 1 BD: 7 ....: devtree_lock FD: 3 BD: 4453 ..-.: pcpu_lock ->stock_lock FD: 128 BD: 79 +.+.: param_lock ->rate_ctrl_mutex ->disk_events_mutex FD: 1 BD: 5263 ..-.: base_crng.lock FD: 1 BD: 1 ....: rcu_read_lock FD: 1 BD: 1 ....: crng_init_wait.lock FD: 2 BD: 1 ....: zonelist_update_seq ->zonelist_update_seq.seqcount FD: 1 BD: 2 ....: zonelist_update_seq.seqcount FD: 1 BD: 1 +.+.: dmar_global_lock FD: 2 BD: 4845 -.-.: &zone->lock ->&____s->seqcount FD: 1 BD: 5237 .-.-: &____s->seqcount FD: 3 BD: 4543 +.+.: &pcp->lock ->&zone->lock FD: 1 BD: 5304 -.-.: pool_lock#2 FD: 130 BD: 220 +.+.: pcpu_alloc_mutex ->pcpu_lock ->fs_reclaim ->pool_lock#2 ->free_vmap_area_lock ->vmap_area_lock ->&zone->lock ->&____s->seqcount ->init_mm.page_table_lock ->&rq->__lock ->pcpu_alloc_mutex.wait_lock ->&c->lock ->&obj_hash[i].lock ->&cfs_rq->removed.lock ->&pool->lock ->rcu_node_0 ->&rcu_state.expedited_wq ->pgd_lock ->key ->percpu_counters_lock ->&n->list_lock FD: 6 BD: 5215 -.-.: &n->list_lock ->&c->lock FD: 5 BD: 5252 -.-.: &c->lock ->batched_entropy_u8.lock ->kfence_freelist_lock FD: 153 BD: 92 +.+.: slab_mutex ->pool_lock#2 ->&c->lock ->&n->list_lock ->pcpu_alloc_mutex ->&zone->lock ->&____s->seqcount ->&obj_hash[i].lock ->fs_reclaim ->&rq->__lock ->batched_entropy_u8.lock ->kfence_freelist_lock ->lock ->&root->kernfs_rwsem ->&k->list_lock FD: 3 BD: 3931 ....: batched_entropy_u64.lock ->crngs.lock FD: 2 BD: 5262 ..-.: crngs.lock ->base_crng.lock FD: 4 BD: 1 ....: espfix_init_mutex ->&zone->lock ->&____s->seqcount ->pool_lock#2 FD: 1 BD: 4420 ..-.: percpu_counters_lock FD: 8 BD: 4496 +.+.: &mm->page_table_lock ->&obj_hash[i].lock ->pool_lock#2 ->quarantine_lock ->stock_lock FD: 11 BD: 4502 +.+.: ptlock_ptr(page) ->lock#4 FD: 51 BD: 4528 +.+.: ptlock_ptr(page)#2 ->lock#4 ->ptlock_ptr(page)#2/1 ->key ->&____s->seqcount ->pool_lock#2 ->lock#5 ->&zone->lock ->&lruvec->lru_lock ->&mapping->private_lock ->&obj_hash[i].lock ->&folio_wait_table[i] ->&pgdat->kswapd_wait FD: 130 BD: 3 +.+.: trace_types_lock ->fs_reclaim ->pool_lock#2 ->pin_fs_lock ->&sb->s_type->i_mutex_key#5 ->&obj_hash[i].lock FD: 1 BD: 1 ....: panic_notifier_list.lock FD: 1 BD: 1 ....: die_chain.lock FD: 132 BD: 4 +.+.: trace_event_sem ->trace_event_ida.xa_lock ->&rq->__lock ->trace_event_sem.wait_lock ->&cfs_rq->removed.lock ->&obj_hash[i].lock ->pool_lock#2 ->fs_reclaim ->batched_entropy_u8.lock ->kfence_freelist_lock ->pin_fs_lock ->&sb->s_type->i_mutex_key#5 ->&c->lock ->&zone->lock ->&____s->seqcount FD: 3 BD: 4138 ..-.: batched_entropy_u32.lock ->crngs.lock FD: 22 BD: 5120 -.-.: &rq->__lock ->&per_cpu_ptr(group->pcpu, cpu)->seq ->&obj_hash[i].lock ->&base->lock ->&rq->__lock/1 ->&cfs_rq->removed.lock ->&rt_b->rt_runtime_lock ->&cp->lock ->pool_lock#2 ->per_cpu_ptr(&cgroup_rstat_cpu_lock, cpu) ->cid_lock FD: 1 BD: 5121 ....: &cfs_b->lock FD: 23 BD: 147 ....: init_task.pi_lock ->&rq->__lock FD: 1 BD: 1 ....: init_task.vtime_seqcount FD: 142 BD: 146 +.+.: wq_pool_mutex ->&zone->lock ->&____s->seqcount ->pool_lock#2 ->&c->lock ->&wq->mutex ->&obj_hash[i].lock ->&pool->lock/1 ->fs_reclaim ->kthread_create_lock ->&p->pi_lock ->&rq->__lock ->&x->wait ->wq_pool_attach_mutex ->(console_sem).lock ->&xa->xa_lock ->&n->list_lock ->&____s->seqcount#2 ->batched_entropy_u8.lock ->kfence_freelist_lock ->&meta->lock ->rcu_node_0 ->&cfs_rq->removed.lock ->quarantine_lock FD: 32 BD: 156 +.+.: &wq->mutex ->&pool->lock ->&pool->lock/1 ->&x->wait#10 ->&rq->__lock ->rcu_node_0 ->&rcu_state.expedited_wq FD: 24 BD: 4787 -.-.: &pool->lock ->&obj_hash[i].lock ->&p->pi_lock ->pool_lock#2 ->&base->lock FD: 29 BD: 4640 ..-.: &pool->lock/1 ->&obj_hash[i].lock ->&p->pi_lock ->pool_lock#2 ->&base->lock ->wq_mayday_lock ->&x->wait#10 FD: 127 BD: 64 ++++: shrinker_rwsem ->pool_lock#2 ->&c->lock ->&n->list_lock ->&____s->seqcount ->fs_reclaim ->&rq->__lock ->shrinker_rwsem.wait_lock ->&pool->lock ->&obj_hash[i].lock ->krc.lock ->f2fs_list_lock ->tk_core.seq.seqcount ->&sbi->s_es_lock ->&journal->j_list_lock FD: 1 BD: 4604 -.-.: rcu_node_0 FD: 5 BD: 88 -.-.: rcu_state.barrier_lock ->rcu_node_0 ->&obj_hash[i].lock FD: 25 BD: 3 ....: &rnp->exp_poll_lock FD: 9 BD: 5 ....: trace_event_ida.xa_lock ->&zone->lock ->&____s->seqcount ->pool_lock#2 ->&c->lock FD: 1 BD: 1 ....: trigger_cmd_mutex FD: 1 BD: 163 ....: i8259A_lock FD: 126 BD: 143 +.+.: irq_domain_mutex ->fs_reclaim ->pool_lock#2 ->&obj_hash[i].lock FD: 29 BD: 308 +.+.: free_vmap_area_lock ->&obj_hash[i].lock ->pool_lock#2 ->init_mm.page_table_lock ->quarantine_lock ->&meta->lock ->kfence_freelist_lock FD: 1 BD: 308 +.+.: vmap_area_lock FD: 7 BD: 159 -.-.: &irq_desc_lock_class ->i8259A_lock ->vector_lock ->ioapic_lock ->mask_lock ->tmp_mask_lock ->irq_resend_lock FD: 33 BD: 89 +.+.: vmap_purge_lock ->purge_vmap_area_lock ->free_vmap_area_lock ->&rq->__lock ->&cfs_rq->removed.lock ->&obj_hash[i].lock ->pool_lock#2 ->rcu_node_0 ->&rcu_state.expedited_wq FD: 26 BD: 106 +.+.: purge_vmap_area_lock ->&obj_hash[i].lock ->pool_lock#2 ->quarantine_lock ->&meta->lock ->kfence_freelist_lock ->&____s->seqcount ->&base->lock FD: 2 BD: 88 +.+.: cpa_lock ->pgd_lock FD: 5 BD: 2 -.-.: timekeeper_lock ->tk_core.seq.seqcount ->pvclock_gtod_data FD: 4 BD: 4802 ----: tk_core.seq.seqcount ->&obj_hash[i].lock ->pvclock_gtod_data FD: 12 BD: 5172 -.-.: &base->lock ->&obj_hash[i].lock FD: 176 BD: 144 +.+.: pmus_lock ->pcpu_alloc_mutex ->pool_lock#2 ->&obj_hash[i].lock ->&cpuctx_mutex ->fs_reclaim ->&k->list_lock ->lock ->&root->kernfs_rwsem ->uevent_sock_mutex ->running_helpers_waitq.lock ->&zone->lock ->&____s->seqcount ->&x->wait#9 ->bus_type_sem ->&c->lock ->sysfs_symlink_target_lock ->&k->k_lock ->&dev->power.lock ->dpm_list_mtx ->subsys mutex#29 FD: 1 BD: 144 +.+.: &swhash->hlist_mutex FD: 1 BD: 145 +.+.: &cpuctx_mutex FD: 1 BD: 2 ....: tty_ldiscs_lock FD: 2 BD: 14 ....: kbd_event_lock ->led_lock FD: 1 BD: 15 ..-.: led_lock FD: 1 BD: 14 ....: vga_lock FD: 34 BD: 4148 -.-.: &port_lock_key ->&dev->power.lock ->&port->lock ->&tty->write_wait FD: 3 BD: 11 ....: console_srcu_srcu_usage.lock ->&obj_hash[i].lock FD: 1 BD: 43 ..-.: &ACCESS_PRIVATE(sdp, lock) FD: 42 BD: 4 +.+.: init_task.alloc_lock ->init_fs.lock FD: 33 BD: 1 +.+.: acpi_ioremap_lock ->pool_lock#2 ->resource_lock ->memtype_lock ->free_vmap_area_lock ->vmap_area_lock FD: 1 BD: 2 +.+.: memtype_lock FD: 1 BD: 17 ....: semaphore->lock FD: 1 BD: 13 ....: *(&acpi_gbl_reference_count_lock) FD: 9 BD: 1 ....: clockevents_lock ->tk_core.seq.seqcount ->tick_broadcast_lock ->i8253_lock FD: 3 BD: 2 -...: tick_broadcast_lock ->jiffies_lock FD: 1 BD: 2 ....: i8253_lock FD: 22 BD: 12 +.+.: &desc->request_mutex ->&irq_desc_lock_class ->proc_subdir_lock ->&ent->pde_unload_lock ->proc_inum_ida.xa_lock ->&obj_hash[i].lock ->pool_lock#2 FD: 163 BD: 2 +.+.: ioapic_mutex ->&domain->mutex FD: 162 BD: 145 +.+.: &domain->mutex ->pool_lock#2 ->vector_lock ->&irq_desc_lock_class ->i8259A_lock ->&c->lock ->&zone->lock ->&____s->seqcount ->sparse_irq_lock ->fs_reclaim ->&obj_hash[i].lock FD: 1 BD: 162 -...: vector_lock FD: 1 BD: 1 +.+.: &pool->lock#2 FD: 2 BD: 3 -.-.: jiffies_lock ->jiffies_seq.seqcount FD: 1 BD: 4 -.-.: jiffies_seq.seqcount FD: 15 BD: 4744 -.-.: hrtimer_bases.lock ->tk_core.seq.seqcount ->&obj_hash[i].lock FD: 24 BD: 1 -.-.: log_wait.lock ->&p->pi_lock FD: 287 BD: 2 +.+.: spec_ctrl_mutex ->cpu_hotplug_lock ->(console_sem).lock ->&rq->__lock FD: 2 BD: 1 +.-.: drivers/char/random.c:1010 ->input_pool.lock FD: 18 BD: 4599 +.+.: sysctl_lock ->&obj_hash[i].lock ->pool_lock#2 ->krc.lock FD: 34 BD: 2 +.+.: tomoyo_policy_lock ->pool_lock#2 ->mmu_notifier_invalidate_range_start ->&c->lock ->&zone->lock ->&____s->seqcount ->&n->list_lock ->&rq->__lock FD: 2 BD: 1 ....: aa_secids.xa_lock ->pool_lock#2 FD: 1 BD: 2 +.+.: aa_buffers_lock FD: 1036 BD: 4 ++++: pernet_ops_rwsem ->stack_depot_init_mutex ->crngs.lock ->net_rwsem ->proc_inum_ida.xa_lock ->pool_lock#2 ->proc_subdir_lock ->fs_reclaim ->&zone->lock ->&____s->seqcount ->&c->lock ->sysctl_lock ->pcpu_alloc_mutex ->net_generic_ids.xa_lock ->mmu_notifier_invalidate_range_start ->&sb->s_type->i_lock_key#8 ->&dir->lock ->&obj_hash[i].lock ->k-sk_lock-AF_NETLINK ->k-slock-AF_NETLINK ->nl_table_lock ->nl_table_wait.lock ->rtnl_mutex ->uevent_sock_mutex ->&net->rules_mod_lock ->slab_mutex ->batched_entropy_u32.lock ->percpu_counters_lock ->&rq->__lock ->cache_list_lock ->rcu_node_0 ->pool_lock ->tk_core.seq.seqcount ->&k->list_lock ->lock ->&root->kernfs_rwsem ->&pool->lock/1 ->running_helpers_waitq.lock ->&sn->pipefs_sb_lock ->krc.lock ->&s->s_inode_list_lock ->nf_hook_mutex ->cpu_hotplug_lock ->hwsim_netgroup_ida.xa_lock ->nf_connlabels_lock ->nf_ct_ecache_mutex ->nf_log_mutex ->batched_entropy_u8.lock ->kfence_freelist_lock ->ipvs->est_mutex ->&base->lock ->__ip_vs_app_mutex ->&cfs_rq->removed.lock ->&hashinfo->lock#2 ->&net->ipv6.ip6addrlbl_table.lock ->(console_sem).lock ->k-sk_lock-AF_INET6 ->k-slock-AF_INET6 ->k-clock-AF_INET6 ->wq_pool_mutex ->pcpu_lock ->&list->lock#4 ->&dir->lock#2 ->ptype_lock ->k-clock-AF_TIPC ->k-sk_lock-AF_TIPC ->k-slock-AF_TIPC ->&this->receive_lock ->once_lock ->nf_ct_proto_mutex ->k-sk_lock-AF_RXRPC ->k-slock-AF_RXRPC ->&rxnet->conn_lock ->&call->waitq ->&rx->call_lock ->&rxnet->call_lock ->&n->list_lock ->rdma_nets.xa_lock ->devices_rwsem ->rtnl_mutex.wait_lock ->&p->pi_lock ->remove_cache_srcu ->stock_lock ->&____s->seqcount#2 ->ebt_mutex ->&xt[i].mutex ->&nft_net->commit_mutex ->netns_bpf_mutex ->&rnp->exp_wq[1] ->rcu_state.barrier_mutex ->rcu_state.barrier_mutex.wait_lock ->dev_base_lock ->lweventlist_lock ->netdev_unregistering_wq.lock ->(work_completion)(&ht->run_work) ->&ht->mutex ->napi_hash_lock ->&fn->fou_lock ->ipvs->sync_mutex ->hwsim_radio_lock ->&ent->pde_unload_lock ->rdma_nets_rwsem ->k-clock-AF_NETLINK ->&nlk->wait ->wlock-AF_NETLINK ->&xa->xa_lock#9 ->&fsnotify_mark_srcu ->&hn->hn_lock ->&pnettable->lock ->&pnetids_ndev->lock ->k-sk_lock-AF_INET6/1 ->&net->sctp.addr_wq_lock ->k-sk_lock-AF_INET ->k-slock-AF_INET ->per_cpu_ptr(&cgroup_rstat_cpu_lock, cpu) ->&sn->gssp_lock ->&cd->hash_lock ->(&net->can.stattimer) ->xfrm_state_gc_work ->&net->xfrm.xfrm_state_lock ->(work_completion)(&(&net->ipv6.addr_chk_work)->work) ->ip6_fl_lock ->(&net->ipv6.ip6_fib_timer) ->__ip_vs_mutex ->(&ipvs->dest_trash_timer) ->(work_completion)(&(&ipvs->expire_nodest_conn_work)->work) ->(work_completion)(&(&ipvs->defense_work)->work) ->(work_completion)(&(&ipvs->est_reload_work)->work) ->nfnl_subsys_ipset ->recent_lock ->hashlimit_mutex ->trans_gc_work ->&rnp->exp_wq[0] ->(work_completion)(&(&cnet->ecache.dwork)->work) ->sysfs_symlink_target_lock ->kernfs_idr_lock ->tcp_metrics_lock ->k-clock-AF_INET ->(work_completion)(&net->xfrm.policy_hash_work) ->&net->xfrm.xfrm_policy_lock ->(work_completion)(&net->xfrm.state_hash_work) ->&list->lock#2 ->&xa->xa_lock#3 ->genl_sk_destructing_waitq.lock ->&rnp->exp_wq[2] ->&net->nsid_lock ->&rnp->exp_lock ->rcu_state.exp_mutex ->(&net->fs_probe_timer) ->&net->cells_lock ->(&net->cells_timer) ->bit_wait_table + i ->(&net->fs_timer) ->(wq_completion)kafsd ->&wq->mutex ->k-clock-AF_RXRPC ->&local->services_lock ->(wq_completion)krxrpcd ->rlock-AF_RXRPC ->&x->wait ->ovs_mutex ->(work_completion)(&(&ovs_net->masks_rebalance)->work) ->(work_completion)(&ovs_net->dp_notify_work) ->&srv->idr_lock ->rcu_state.exp_mutex.wait_lock ->(work_completion)(&tn->work) ->&tn->nametbl_lock ->(work_completion)(&(&c->work)->work) ->(wq_completion)krdsd ->(work_completion)(&rtn->rds_tcp_accept_w) ->rds_tcp_conn_lock ->loop_conns_lock ->(wq_completion)l2tp ->(&rxnet->peer_keepalive_timer) ->(work_completion)(&rxnet->peer_keepalive_work) ->(&rxnet->service_conn_reap_timer) ->&x->wait#10 ->quarantine_lock ->&rcu_state.expedited_wq ->&meta->lock ->&rnp->exp_wq[3] ->&sem->wait_lock ->pcpu_alloc_mutex.wait_lock ->&device->compat_devs_mutex ->dev_pm_qos_sysfs_mtx ->subsys mutex#84 ->&x->wait#9 ->dpm_list_mtx ->&dev->power.lock ->deferred_probe_mutex ->device_links_lock ->gdp_mutex ->&device->unregistration_lock ->uevent_sock_mutex.wait_lock ->pgd_lock ->key ->rdma_nets_rwsem.wait_lock ->key#24 ->&lock->wait_lock ->nf_conntrack_mutex ->key#25 ->__ip_vs_app_mutex.wait_lock FD: 26 BD: 81 +.+.: stack_depot_init_mutex ->&rq->__lock ->rcu_node_0 ->&rcu_state.expedited_wq FD: 156 BD: 3897 ++++: net_rwsem ->&list->lock#2 ->&rq->__lock ->&c->lock ->pool_lock#2 ->&obj_hash[i].lock ->nl_table_lock ->nl_table_wait.lock ->&n->list_lock ->&table->lock#4 ->&ndev->lock ->rcu_node_0 ->&rcu_state.expedited_wq ->&data->lock ->&cfs_rq->removed.lock FD: 12 BD: 100 ....: proc_inum_ida.xa_lock ->&____s->seqcount ->pool_lock#2 ->&c->lock ->&obj_hash[i].lock ->&n->list_lock ->&____s->seqcount#2 FD: 889 BD: 77 +.+.: rtnl_mutex ->&c->lock ->&zone->lock ->&____s->seqcount ->pool_lock#2 ->fs_reclaim ->pcpu_alloc_mutex ->&xa->xa_lock#3 ->&x->wait#9 ->&obj_hash[i].lock ->&k->list_lock ->gdp_mutex ->lock ->&root->kernfs_rwsem ->bus_type_sem ->sysfs_symlink_target_lock ->&dev->power.lock ->dpm_list_mtx ->uevent_sock_mutex ->running_helpers_waitq.lock ->subsys mutex#17 ->&dir->lock#2 ->dev_hotplug_mutex ->dev_base_lock ->input_pool.lock ->nl_table_lock ->nl_table_wait.lock ->net_rwsem ->batched_entropy_u32.lock ->&tbl->lock ->sysctl_lock ->&rq->__lock ->batched_entropy_u8.lock ->kfence_freelist_lock ->krc.lock ->stack_depot_init_mutex ->cpu_hotplug_lock ->kthread_create_lock ->&p->pi_lock ->&x->wait ->wq_pool_mutex ->crngs.lock ->&pool->lock/1 ->lweventlist_lock ->&pool->lock ->rtnl_mutex.wait_lock ->proc_subdir_lock ->proc_inum_ida.xa_lock ->&cfs_rq->removed.lock ->&k->k_lock ->param_lock ->(console_sem).lock ->&rdev->wiphy.mtx ->&base->lock ->subsys mutex#55 ->&sdata->sec_mtx ->&local->iflist_mtx#2 ->lock#7 ->failover_lock ->&tn->lock ->&idev->mc_lock ->&ndev->lock ->pool_lock ->&pnettable->lock ->smc_ib_devices.mutex ->&(&net->nexthop.notifier_chain)->rwsem ->reg_requests_lock ->reg_pending_beacons_lock ->rlock-AF_NETLINK ->(inetaddr_validator_chain).rwsem ->(inetaddr_chain).rwsem ->_xmit_LOOPBACK ->netpoll_srcu ->&in_dev->mc_tomb_lock ->&im->lock ->fib_info_lock ->cbs_list_lock ->(inet6addr_validator_chain).rwsem ->&net->ipv6.addrconf_hash_lock ->&ifa->lock ->&tb->tb6_lock ->&n->list_lock ->&dev_addr_list_lock_key ->napi_hash_lock ->lapb_list_lock ->x25_neigh_list_lock ->console_owner_lock ->console_owner ->_xmit_ETHER ->_xmit_SLIP ->&sem->wait_lock ->&vi->refill_lock ->noop_qdisc.q.lock ->&rfkill->lock ->&local->chanctx_mtx ->&dev->tx_global_lock ->rcu_node_0 ->&rnp->exp_wq[3] ->&sch->q.lock ->class ->(&tbl->proxy_timer) ->_xmit_VOID ->_xmit_X25 ->&lapbeth->up_lock ->&lapb->lock ->pgd_lock ->key ->pcpu_lock ->percpu_counters_lock ->&rnp->exp_wq[1] ->&dir->lock ->&ul->lock#2 ->&n->lock ->dev_addr_sem ->_xmit_IEEE802154 ->reg_indoor_lock ->remove_cache_srcu ->&nr_netdev_addr_lock_key ->listen_lock ->pcpu_alloc_mutex.wait_lock ->quarantine_lock ->&meta->lock ->&r->consumer_lock ->&mm->mmap_lock ->(switchdev_blocking_notif_chain).rwsem ->&br->hash_lock ->nf_hook_mutex ->j1939_netdev_lock ->&bat_priv->tvlv.handler_list_lock ->&bat_priv->tvlv.container_list_lock ->&bat_priv->softif_vlan_list_lock ->key#16 ->&bat_priv->tt.changes_list_lock ->kernfs_idr_lock ->&rnp->exp_wq[0] ->&rnp->exp_wq[2] ->tk_core.seq.seqcount ->&wq->mutex ->init_lock ->deferred_lock ->target_list_lock ->&br->lock ->&pn->hash_lock ->team->team_lock_key ->team->team_lock_key#2 ->&rcu_state.expedited_wq ->&hard_iface->bat_iv.ogm_buff_mutex ->ptype_lock ->team->team_lock_key#3 ->_xmit_NONE ->lock#9 ->team->team_lock_key#4 ->&hsr->list_lock ->pin_fs_lock ->&sb->s_type->i_mutex_key#3 ->team->team_lock_key#5 ->team->team_lock_key#6 ->mount_lock ->&xa->xa_lock#15 ->&dev_addr_list_lock_key#3/1 ->req_lock ->&x->wait#11 ->subsys mutex#82 ->bpf_devs_lock ->(work_completion)(&(&devlink_port->type_warn_dw)->work) ->&devlink_port->type_lock ->&vn->sock_lock ->devnet_rename_sem ->&nft_net->commit_mutex ->&ent->pde_unload_lock ->&wg->device_update_lock ->_xmit_SIT ->&bridge_netdev_addr_lock_key/1 ->_xmit_TUNNEL ->_xmit_IPGRE ->_xmit_TUNNEL6 ->&dev_addr_list_lock_key/1 ->&dev_addr_list_lock_key#2/1 ->_xmit_ETHER/1 ->&nn->netlink_tap_lock ->&batadv_netdev_addr_lock_key/1 ->&vlan_netdev_addr_lock_key/1 ->&macvlan_netdev_addr_lock_key/1 ->&ipvlan->addrs_lock ->&macsec_netdev_addr_lock_key/1 ->key#20 ->&bat_priv->tt.commit_lock ->mmu_notifier_invalidate_range_start ->&sb->s_type->i_lock_key#8 ->k-sk_lock-AF_INET ->k-slock-AF_INET ->k-sk_lock-AF_INET6 ->k-slock-AF_INET6 ->&____s->seqcount#2 ->&ul->lock ->__ip_vs_mutex ->flowtable_lock ->raw_notifier_lock ->bcm_notifier_lock ->isotp_notifier_lock ->stock_lock ->qdisc_mod_lock ->&block->lock ->&block->cb_lock ->rcu_state.exp_mutex.wait_lock ->&idev->mc_query_lock ->(work_completion)(&(&idev->mc_report_work)->work) ->&rnp->exp_lock ->rcu_state.exp_mutex ->&hwstats->hwsdev_list_lock ->&net->xdp.lock ->mirred_list_lock ->&idev->mc_report_lock ->(&pmctx->ip6_mc_router_timer) ->(&pmctx->ip4_mc_router_timer) ->(work_completion)(&ht->run_work) ->&ht->mutex ->&br->multicast_lock ->&pnn->pndevs.lock ->&pnn->routes.lock ->dev_pm_qos_sysfs_mtx ->deferred_probe_mutex ->device_links_lock ->&r->consumer_lock#2 ->&wg->socket_update_lock ->&table->hash[i].lock ->k-clock-AF_INET ->&xa->xa_lock#9 ->&fsnotify_mark_srcu ->k-clock-AF_INET6 ->&caifn->caifdevs.lock ->&net->rules_mod_lock ->(&mrt->ipmr_expire_timer) ->nf_connlabels_lock ->key#19 ->&bat_priv->forw_bcast_list_lock ->&bat_priv->forw_bat_list_lock ->(work_completion)(&(&forw_packet_aggr->delayed_work)->work) ->per_cpu_ptr(&cgroup_rstat_cpu_lock, cpu) ->&chain->filter_chain_lock ->cls_mod_lock ->&block->proto_destroy_lock ->&pn->all_ppp_mutex ->&ppp->rlock ->&ppp->wlock ->&dev_addr_list_lock_key#4 ->&pf->rwait ->&bridge_netdev_addr_lock_key ->&batadv_netdev_addr_lock_key ->&bond->mode_lock ->&net->xfrm.xfrm_state_lock ->(work_completion)(&(&slave->notify_work)->work) ->(work_completion)(&(&br->gc_work)->work) ->(&brmctx->ip4_mc_router_timer) ->(&brmctx->ip4_other_query.timer) ->(&brmctx->ip4_own_query.timer) ->(&brmctx->ip6_mc_router_timer) ->(&brmctx->ip6_other_query.timer) ->(&brmctx->ip6_own_query.timer) ->(work_completion)(&br->mcast_gc_work) ->rcu_state.barrier_mutex ->_xmit_PHONET_PIPE ->uevent_sock_mutex.wait_lock ->sk_lock-AF_INET6 ->slock-AF_INET6 ->acaddr_hash_lock ->sk_lock-AF_INET ->slock-AF_INET ->&sb->s_type->i_lock_key#23 ->&dentry->d_lock ->rename_lock.seqcount ->&s->s_inode_list_lock ->&app->lock ->(&app->join_timer) ->(&app->periodic_timer) ->&list->lock#11 ->(&app->join_timer)#2 ->&app->lock#2 ->&list->lock#12 ->(&br->hello_timer) ->(&br->topology_change_timer) ->(&br->tcn_timer) ->&tun->lock ->wlock-AF_UNSPEC ->elock-AF_UNSPEC ->dev->qdisc_tx_busylock ?: &qdisc_tx_busylock ->&newf->file_lock ->&sb->s_type->i_lock_key#15 ->bpf_dispatcher_xdp.mutex ->dev_pm_qos_sysfs_mtx.wait_lock ->&pmc->lock ->sk_lock-AF_UNSPEC ->slock-AF_UNSPEC ->sk_lock-AF_CAN ->slock-AF_CAN ->&xs->mutex ->&dev_addr_list_lock_key#3/2 ->(work_completion)(&port->wq) ->&p->alloc_lock ->&list->lock#2 ->(work_completion)(&port->bc_work) ->(work_completion)(&(&bond->mii_work)->work) ->(work_completion)(&(&bond->arp_work)->work) ->(work_completion)(&(&bond->alb_work)->work) ->(work_completion)(&(&bond->ad_work)->work) ->(work_completion)(&(&bond->mcast_work)->work) ->(work_completion)(&(&bond->slave_arr_work)->work) ->(work_completion)(&wdev->disconnect_wk) ->(work_completion)(&wdev->pmsr_free_wk) ->(work_completion)(&sdata->activate_links_work) ->&rdev->dev_wait ->_xmit_PIMREG ->mrt_lock ->_xmit_NETROM#2 ->dev->qdisc_tx_busylock ?: &qdisc_tx_busylock#2 ->&net->xfrm.xfrm_policy_lock ->&vlan_netdev_addr_lock_key/2 ->&lock->wait_lock ->&tn->nametbl_lock ->&ht->lock ->sk_lock-AF_TIPC ->slock-AF_TIPC ->k-sk_lock-AF_TIPC ->k-slock-AF_TIPC ->&tipc_net(net)->bclock ->&wdev->mtx ->&fq->lock ->ematch_mod_lock ->&dev_addr_list_lock_key#2 ->prog_idr_lock ->bpf_lock ->(&mp->timer) ->&this->info_list_lock ->(&hsr->prune_timer) ->(&hsr->announce_timer) ->&sb->s_type->i_lock_key#7 ->act_mod_lock ->&tn->idrinfo->lock ->&p->tcfa_lock ->mfc_unres_lock ->team->team_lock_key#7 ->team->team_lock_key#8 ->team->team_lock_key#9 ->&tn->idrinfo->lock#2 ->&net->ipv4.ra_mutex ->mfc_unres_lock#2 ->mrt_lock#2 ->&tn->node_list_lock ->&dev_addr_list_lock_key/2 ->ifalias_mutex ->&macsec_netdev_addr_lock_key#2/2 ->&tn->idrinfo->lock#3 ->team->team_lock_key#10 ->team->team_lock_key#11 ->&data->lock ->&dev_addr_list_lock_key#2/3 ->_xmit_ETHER/4 ->&macvlan_netdev_addr_lock_key/2 ->&tn->idrinfo->lock#4 ->free_vmap_area_lock ->vmap_area_lock ->purge_vmap_area_lock FD: 45 BD: 247 +.+.: lock ->kernfs_idr_lock ->cgroup_idr_lock ->pidmap_lock ->drm_minor_lock ->&file_private->table_lock ->&q->queue_lock ->sg_index_lock ->map_idr_lock ->prog_idr_lock ->btf_idr_lock ->&group->inotify_data.idr_lock ->link_idr_lock ->sctp_assocs_id_lock FD: 13 BD: 4521 +.+.: kernfs_idr_lock ->pool_lock#2 ->&c->lock ->&____s->seqcount ->&zone->lock ->&obj_hash[i].lock ->batched_entropy_u8.lock ->kfence_freelist_lock ->&n->list_lock ->&____s->seqcount#2 FD: 129 BD: 250 ++++: &root->kernfs_rwsem ->&root->kernfs_iattr_rwsem ->kernfs_idr_lock ->&obj_hash[i].lock ->pool_lock#2 ->&rq->__lock ->&sem->wait_lock ->&cfs_rq->removed.lock ->rcu_node_0 ->quarantine_lock ->inode_hash_lock ->fs_reclaim ->mmu_notifier_invalidate_range_start ->tk_core.seq.seqcount ->&sb->s_type->i_lock_key#24 ->&c->lock ->&____s->seqcount ->remove_cache_srcu ->&n->list_lock ->&p->pi_lock ->&sb->s_type->i_lock_key#30 ->&sb->s_type->i_lock_key#31 ->&meta->lock ->kfence_freelist_lock ->kernfs_rename_lock ->&xa->xa_lock#4 ->stock_lock ->&____s->seqcount#2 ->pgd_lock ->key ->pcpu_lock ->percpu_counters_lock ->&rcu_state.expedited_wq ->&base->lock ->batched_entropy_u8.lock FD: 1 BD: 5 ++++: file_systems_lock FD: 127 BD: 255 ++++: &root->kernfs_iattr_rwsem ->&rq->__lock ->&cfs_rq->removed.lock ->&obj_hash[i].lock ->pool_lock#2 ->iattr_mutex ->&sem->wait_lock ->tk_core.seq.seqcount ->rcu_node_0 ->&rcu_state.expedited_wq FD: 5 BD: 48 +.+.: sb_lock ->unnamed_dev_ida.xa_lock ->&obj_hash[i].lock ->pool_lock#2 FD: 138 BD: 1 +.+.: &type->s_umount_key/1 ->pool_lock#2 ->pcpu_alloc_mutex ->shrinker_rwsem ->list_lrus_mutex ->sb_lock ->&obj_hash[i].lock ->percpu_counters_lock ->crngs.lock ->&sbinfo->stat_lock ->&zone->lock ->&____s->seqcount ->&c->lock ->&sb->s_type->i_lock_key ->&s->s_inode_list_lock ->tk_core.seq.seqcount ->batched_entropy_u32.lock ->&dentry->d_lock ->fs_reclaim ->mmu_notifier_invalidate_range_start FD: 23 BD: 40 +.+.: list_lrus_mutex ->&rq->__lock FD: 2 BD: 49 ....: unnamed_dev_ida.xa_lock ->pool_lock#2 FD: 1 BD: 11 +.+.: &sbinfo->stat_lock FD: 55 BD: 4540 +.+.: &sb->s_type->i_lock_key ->&dentry->d_lock ->&xa->xa_lock#9 FD: 1 BD: 4533 +.+.: &s->s_inode_list_lock FD: 39 BD: 4598 +.+.: &dentry->d_lock ->&wq ->&dentry->d_lock/1 ->&wq#2 ->&lru->node[i].lock ->sysctl_lock ->&wq#3 ->&dentry->d_lock/2 ->&p->pi_lock FD: 2 BD: 29 ....: mnt_id_ida.xa_lock ->pool_lock#2 FD: 44 BD: 201 +.+.: mount_lock ->mount_lock.seqcount ->&dentry->d_lock ->&obj_hash[i].lock ->pool_lock#2 FD: 42 BD: 206 +.+.: mount_lock.seqcount ->&new_ns->poll ->&dentry->d_lock ->&obj_hash[i].lock ->pool_lock#2 ->quarantine_lock ->&p->pi_lock FD: 136 BD: 1 +.+.: &type->s_umount_key#2/1 ->pool_lock#2 ->pcpu_alloc_mutex ->shrinker_rwsem ->list_lrus_mutex ->sb_lock ->&zone->lock ->&____s->seqcount ->&c->lock ->&sb->s_type->i_lock_key#2 ->&s->s_inode_list_lock ->tk_core.seq.seqcount ->&dentry->d_lock FD: 40 BD: 4540 +.+.: &sb->s_type->i_lock_key#2 ->&dentry->d_lock FD: 1 BD: 2 ..-.: ucounts_lock FD: 41 BD: 220 +.+.: init_fs.lock ->init_fs.seq.seqcount ->&dentry->d_lock FD: 1 BD: 218 +.+.: init_fs.seq.seqcount FD: 136 BD: 1 +.+.: &type->s_umount_key#3/1 ->pool_lock#2 ->pcpu_alloc_mutex ->shrinker_rwsem ->list_lrus_mutex ->sb_lock ->&zone->lock ->&____s->seqcount ->&c->lock ->&sb->s_type->i_lock_key#3 ->&s->s_inode_list_lock ->tk_core.seq.seqcount ->&dentry->d_lock FD: 55 BD: 4544 +.+.: &sb->s_type->i_lock_key#3 ->&dentry->d_lock ->&xa->xa_lock#9 FD: 1 BD: 143 +.+.: cpuhp_state-down FD: 215 BD: 143 +.+.: cpuhp_state-up ->smpboot_threads_lock ->sparse_irq_lock ->&swhash->hlist_mutex ->pmus_lock ->&x->wait#5 ->&obj_hash[i].lock ->hrtimer_bases.lock ->wq_pool_mutex ->rcu_node_0 ->&rq->__lock ->jump_label_mutex ->fs_reclaim ->&c->lock ->&n->list_lock ->&zone->lock ->&____s->seqcount ->pool_lock#2 ->&x->wait#9 ->&k->list_lock ->gdp_mutex ->lock ->&root->kernfs_rwsem ->bus_type_sem ->sysfs_symlink_target_lock ->&dev->power.lock ->dpm_list_mtx ->req_lock ->&p->pi_lock ->&x->wait#11 ->uevent_sock_mutex ->running_helpers_waitq.lock ->subsys mutex#24 ->subsys mutex#25 ->&cfs_rq->removed.lock ->&k->k_lock ->subsys mutex#79 ->&base->lock ->swap_slots_cache_mutex FD: 1 BD: 102 ++++: proc_subdir_lock FD: 137 BD: 1 +.+.: &type->s_umount_key#4/1 ->pool_lock#2 ->pcpu_alloc_mutex ->shrinker_rwsem ->list_lrus_mutex ->sb_lock ->&sb->s_type->i_lock_key#4 ->&s->s_inode_list_lock ->tk_core.seq.seqcount ->&dentry->d_lock FD: 40 BD: 2 +.+.: &sb->s_type->i_lock_key#4 ->&dentry->d_lock FD: 26 BD: 151 ....: cgroup_file_kn_lock ->kernfs_notify_lock FD: 33 BD: 150 ..-.: css_set_lock ->cgroup_file_kn_lock ->&p->pi_lock ->&obj_hash[i].lock ->pool_lock#2 ->krc.lock ->kernfs_rename_lock FD: 9 BD: 248 +...: cgroup_idr_lock ->pool_lock#2 ->&c->lock ->&zone->lock ->&____s->seqcount FD: 145 BD: 144 +.+.: cpuset_mutex ->callback_lock ->jump_label_mutex ->&p->pi_lock ->&p->alloc_lock ->cpuset_attach_wq.lock ->fs_reclaim ->pool_lock#2 ->css_set_lock ->&obj_hash[i].lock FD: 1 BD: 145 ....: callback_lock FD: 131 BD: 18 +.+.: blkcg_pol_mutex ->pcpu_alloc_mutex ->fs_reclaim ->pool_lock#2 ->&c->lock ->&n->list_lock ->&rq->__lock FD: 3 BD: 5252 ..-.: batched_entropy_u8.lock ->crngs.lock FD: 1 BD: 4515 ....: &pgdat->memcg_lru.lock FD: 23 BD: 19 +.+.: devcgroup_mutex ->&rq->__lock FD: 47 BD: 144 +.+.: freezer_mutex ->freezer_lock ->&rq->__lock ->rcu_node_0 ->freezer_mutex.wait_lock FD: 42 BD: 196 +.+.: rcu_state.exp_mutex ->rcu_node_0 ->rcu_state.exp_wake_mutex ->&obj_hash[i].lock ->&pool->lock ->&rnp->exp_wq[2] ->&rq->__lock ->&rnp->exp_wq[3] ->&rnp->exp_wq[0] ->rcu_state.exp_mutex.wait_lock ->&rcu_state.expedited_wq ->&rnp->exp_wq[1] ->pool_lock#2 ->&cfs_rq->removed.lock ->pgd_lock ->stock_lock ->key ->pcpu_lock ->percpu_counters_lock FD: 33 BD: 242 +.+.: rcu_state.exp_wake_mutex ->rcu_node_0 ->&rnp->exp_lock ->&rnp->exp_wq[0] ->&rnp->exp_wq[1] ->&rnp->exp_wq[2] ->&rnp->exp_wq[3] ->&rq->__lock ->&cfs_rq->removed.lock ->&obj_hash[i].lock ->rcu_state.exp_wake_mutex.wait_lock ->&pool->lock ->pool_lock#2 ->&rcu_state.expedited_wq FD: 1 BD: 252 +.+.: &rnp->exp_lock FD: 24 BD: 256 ....: &rnp->exp_wq[0] ->&p->pi_lock FD: 24 BD: 247 ....: &rnp->exp_wq[1] ->&p->pi_lock FD: 1 BD: 147 ....: init_sighand.siglock FD: 1 BD: 3 +.+.: init_files.file_lock FD: 13 BD: 259 ....: pidmap_lock ->pool_lock#2 ->&obj_hash[i].lock ->&c->lock ->&zone->lock ->&____s->seqcount ->&____s->seqcount#2 ->&n->list_lock ->batched_entropy_u8.lock ->kfence_freelist_lock FD: 173 BD: 143 ++++: cgroup_threadgroup_rwsem ->css_set_lock ->&p->pi_lock ->tk_core.seq.seqcount ->tasklist_lock ->&rq->__lock ->&cfs_rq->removed.lock ->&obj_hash[i].lock ->pool_lock#2 ->&c->lock ->pgd_lock ->key ->pcpu_lock ->percpu_counters_lock ->&sighand->siglock ->cgroup_threadgroup_rwsem.rss.gp_wait.lock ->rcu_node_0 ->inode_hash_lock ->fs_reclaim ->mmu_notifier_invalidate_range_start ->&sb->s_type->i_lock_key#30 ->&root->kernfs_iattr_rwsem ->&s->s_inode_list_lock ->&xa->xa_lock#9 ->&fsnotify_mark_srcu ->cpuset_mutex ->&p->alloc_lock ->freezer_mutex ->cgroup_threadgroup_rwsem.waiters.lock ->&rcu_state.expedited_wq ->freezer_mutex.wait_lock ->stock_lock FD: 23 BD: 5016 -.-.: &p->pi_lock ->&rq->__lock ->&cfs_rq->removed.lock FD: 68 BD: 146 .+.+: tasklist_lock ->init_task.pi_lock ->init_sighand.siglock ->&p->pi_lock ->&sighand->siglock ->&pid->wait_pidfd ->&obj_hash[i].lock ->&c->lock ->&n->list_lock ->stock_lock ->&p->alloc_lock ->per_cpu_ptr(&cgroup_rstat_cpu_lock, cpu) FD: 1 BD: 5121 -.-.: &per_cpu_ptr(group->pcpu, cpu)->seq FD: 1 BD: 1 ....: (kthreadd_done).wait.lock FD: 42 BD: 155 ....: &sighand->siglock ->&sig->wait_chldexit ->input_pool.lock ->&(&sig->stats_lock)->lock ->&____s->seqcount ->pool_lock#2 ->&c->lock ->hrtimer_bases.lock ->&p->pi_lock ->&obj_hash[i].lock ->&sighand->signalfd_wqh ->batched_entropy_u8.lock ->kfence_freelist_lock ->&meta->lock ->&tty->ctrl.lock ->&prev->lock ->&rq->__lock ->stock_lock ->per_cpu_ptr(&cgroup_rstat_cpu_lock, cpu) ->&n->list_lock ->&____s->seqcount#2 ->quarantine_lock FD: 49 BD: 218 +.+.: &p->alloc_lock ->&____s->seqcount#2 ->init_fs.lock ->&fs->lock ->&x->wait ->&memcg->mm_list.lock ->&x->wait#25 ->&newf->file_lock ->&p->pi_lock FD: 1 BD: 5204 .-.-: &____s->seqcount#2 FD: 125 BD: 4514 +.+.: fs_reclaim ->mmu_notifier_invalidate_range_start ->&mapping->i_mmap_rwsem ->&rq->__lock ->&cfs_rq->removed.lock ->&obj_hash[i].lock ->pgd_lock ->stock_lock ->key ->pcpu_lock ->percpu_counters_lock ->rcu_node_0 ->&rcu_state.expedited_wq ->pool_lock#2 ->lock#4 ->lock#5 ->batched_entropy_u8.lock ->&lruvec->lru_lock ->&sb->s_type->i_lock_key#22 ->&zone->lock ->&mapping->private_lock ->&sb->s_type->i_lock_key#3 ->&meta->lock ->kfence_freelist_lock ->quarantine_lock ->&sem->wait_lock ->&p->pi_lock ->&c->lock ->&____s->seqcount#2 ->&____s->seqcount ->&pgdat->kcompactd_wait ->&pgdat->memcg_lru.lock ->&memcg->mm_list.lock ->&n->list_lock ->swap_slots_cache_mutex ->&cache->alloc_lock ->shmem_swaplist_mutex ->&p->lock#2 ->&tree->lock ->&xa->xa_lock#21 ->remove_cache_srcu ->&base->lock FD: 32 BD: 4534 +.+.: mmu_notifier_invalidate_range_start ->dma_fence_map ->&rq->__lock ->&cfs_rq->removed.lock ->&obj_hash[i].lock ->pgd_lock ->pool_lock#2 ->key ->pcpu_lock ->percpu_counters_lock ->rcu_node_0 ->&rcu_state.expedited_wq ->stock_lock FD: 1 BD: 168 +.+.: kthread_create_lock FD: 24 BD: 245 ....: &x->wait ->&p->pi_lock FD: 32 BD: 148 +.+.: wq_pool_attach_mutex ->&p->pi_lock ->&x->wait#7 ->&pool->lock ->&pool->lock/1 ->&rq->__lock ->wq_pool_attach_mutex.wait_lock ->&cfs_rq->removed.lock ->&obj_hash[i].lock ->pool_lock#2 FD: 24 BD: 4641 ..-.: wq_mayday_lock ->&p->pi_lock FD: 1 BD: 147 ....: &xa->xa_lock FD: 30 BD: 1 +.-.: (&pool->mayday_timer) ->&pool->lock/1 ->&obj_hash[i].lock ->&base->lock FD: 50 BD: 1 +.+.: (wq_completion)rcu_gp ->(work_completion)(&rnp->exp_poll_wq) ->(work_completion)(&(&ssp->srcu_sup->work)->work) ->(work_completion)(&sdp->work) ->(work_completion)(&rew->rew_work) ->&rq->__lock ->&cfs_rq->removed.lock ->&obj_hash[i].lock FD: 26 BD: 2 +.+.: (work_completion)(&rnp->exp_poll_wq) ->&rnp->exp_poll_lock FD: 1000 BD: 1 +.+.: (wq_completion)events_unbound ->(work_completion)(&(&kfence_timer)->work) ->(work_completion)(&entry->work) ->(next_reseed).work ->(work_completion)(&sub_info->work) ->(stats_flush_dwork).work ->deferred_probe_work ->(work_completion)(&barr->work) ->(work_completion)(&map->work) ->connector_reaper_work ->(reaper_work).work ->(work_completion)(&port->bc_work) ->&rq->__lock ->(work_completion)(&pool->idle_cull_work) FD: 288 BD: 2 +.+.: (work_completion)(&(&kfence_timer)->work) ->cpu_hotplug_lock ->allocation_wait.lock ->&rq->__lock ->&obj_hash[i].lock ->&base->lock ->&cfs_rq->removed.lock ->pool_lock#2 ->rcu_node_0 ->&rcu_state.expedited_wq FD: 24 BD: 3 -.-.: allocation_wait.lock ->&p->pi_lock FD: 13 BD: 1 +.-.: (&wq_watchdog_timer) ->&obj_hash[i].lock ->&base->lock FD: 1 BD: 5249 ..-.: kfence_freelist_lock FD: 1 BD: 4593 ..-.: &meta->lock FD: 56 BD: 1 +.+.: rcu_tasks.tasks_gp_mutex ->&obj_hash[i].lock ->&base->lock ->tasks_rcu_exit_srcu_srcu_usage.lock ->&ACCESS_PRIVATE(sdp, lock) ->tasks_rcu_exit_srcu ->&x->wait#2 ->&rq->__lock ->kernel/rcu/tasks.h:152 ->(&timer.timer) ->rcu_tasks__percpu.cbs_pcpu_lock ->&x->wait#3 ->rcu_state.exp_mutex.wait_lock ->&p->pi_lock ->(console_sem).lock FD: 1 BD: 1 ....: rcu_tasks.cbs_gbl_lock FD: 289 BD: 1 +.+.: rcu_tasks_trace.tasks_gp_mutex ->cpu_hotplug_lock ->rcu_tasks_trace__percpu.cbs_pcpu_lock ->&x->wait#3 ->&obj_hash[i].lock ->&base->lock ->&rq->__lock ->(&timer.timer) ->(console_sem).lock FD: 1 BD: 1 ....: rcu_tasks_trace.cbs_gbl_lock FD: 13 BD: 3 ..-.: rcu_tasks__percpu.cbs_pcpu_lock ->&obj_hash[i].lock ->&base->lock FD: 24 BD: 257 ....: &rnp->exp_wq[2] ->&p->pi_lock FD: 26 BD: 8 ....: tasks_rcu_exit_srcu_srcu_usage.lock ->&obj_hash[i].lock ->&ACCESS_PRIVATE(sdp, lock) FD: 1 BD: 2 ....: tasks_rcu_exit_srcu FD: 24 BD: 25 ....: &x->wait#2 ->&p->pi_lock FD: 35 BD: 4 +.+.: (work_completion)(&(&ssp->srcu_sup->work)->work) ->&ssp->srcu_sup->srcu_gp_mutex ->&ssp->srcu_sup->srcu_cb_mutex ->tasks_rcu_exit_srcu_srcu_usage.lock ->&obj_hash[i].lock ->&base->lock ->remove_cache_srcu_srcu_usage.lock ->&rq->__lock ->wakeup_srcu_srcu_usage.lock ->&ACCESS_PRIVATE(ssp->srcu_sup, lock) ->tracepoint_srcu_srcu_usage.lock ->&cfs_rq->removed.lock FD: 34 BD: 5 +.+.: &ssp->srcu_sup->srcu_gp_mutex ->tasks_rcu_exit_srcu_srcu_usage.lock ->&ssp->srcu_sup->srcu_cb_mutex ->remove_cache_srcu_srcu_usage.lock ->&rq->__lock ->wakeup_srcu_srcu_usage.lock ->&ACCESS_PRIVATE(ssp->srcu_sup, lock) ->tracepoint_srcu_srcu_usage.lock ->&cfs_rq->removed.lock ->&obj_hash[i].lock ->pool_lock#2 ->rcu_node_0 ->&rcu_state.expedited_wq FD: 31 BD: 6 +.+.: &ssp->srcu_sup->srcu_cb_mutex ->tasks_rcu_exit_srcu_srcu_usage.lock ->remove_cache_srcu_srcu_usage.lock ->wakeup_srcu_srcu_usage.lock ->&ACCESS_PRIVATE(ssp->srcu_sup, lock) ->&rq->__lock ->tracepoint_srcu_srcu_usage.lock ->&obj_hash[i].lock ->&base->lock FD: 29 BD: 4 +.+.: (work_completion)(&sdp->work) ->&ACCESS_PRIVATE(sdp, lock) ->&obj_hash[i].lock ->&x->wait#2 ->&rq->__lock ->rcu_node_0 ->&rcu_state.expedited_wq ->pool_lock#2 ->&cfs_rq->removed.lock ->&meta->lock ->kfence_freelist_lock FD: 1 BD: 2 ....: kernel/rcu/tasks.h:152 FD: 24 BD: 3 ....: &x->wait#3 ->&p->pi_lock FD: 5 BD: 1 -.-.: (null) ->tk_core.seq.seqcount FD: 24 BD: 218 +.-.: (&timer.timer) ->&p->pi_lock FD: 30 BD: 1 ..-.: &(&kfence_timer)->timer FD: 24 BD: 255 ....: &rnp->exp_wq[3] ->&p->pi_lock FD: 13 BD: 143 ..-.: rcu_tasks_trace__percpu.cbs_pcpu_lock ->&obj_hash[i].lock ->&base->lock FD: 1 BD: 142 ....: &ACCESS_PRIVATE(rtpcp, lock) FD: 1 BD: 1 ....: &nmi_desc[0].lock FD: 128 BD: 144 +.+.: smpboot_threads_lock ->fs_reclaim ->&rq->__lock ->batched_entropy_u8.lock ->kfence_freelist_lock ->pool_lock#2 ->kthread_create_lock ->&p->pi_lock ->&x->wait ->&obj_hash[i].lock ->&c->lock ->&n->list_lock FD: 24 BD: 4495 -.-.: &rcu_state.gp_wq ->&p->pi_lock FD: 23 BD: 250 -.-.: &stop_pi_lock ->&rq->__lock FD: 1 BD: 250 -.-.: &stopper->lock FD: 1 BD: 2 +.+.: (module_notify_list).rwsem FD: 1 BD: 1 +.+.: ddebug_lock FD: 1 BD: 1 .+.+: &pmus_srcu FD: 1 BD: 1 ....: rcu_callback FD: 287 BD: 1 +.+.: watchdog_mutex ->cpu_hotplug_lock FD: 24 BD: 143 ....: &x->wait#4 ->&p->pi_lock FD: 1082 BD: 1 +.+.: (wq_completion)events ->(work_completion)(&sscs.work) ->pcpu_balance_work ->(work_completion)(&pwq->unbound_release_work) ->(shepherd).work ->(work_completion)(&rfkill_global_led_trigger_work) ->timer_update_work ->(work_completion)(&p->wq) ->(work_completion)(&(&group->avgs_work)->work) ->(work_completion)(&(&krcp->monitor_work)->work) ->(work_completion)(&helper->damage_work) ->(work_completion)(&rfkill->sync_work) ->(linkwatch_work).work ->(work_completion)(&w->work) ->(work_completion)(&vi->config_work) ->(debug_obj_work).work ->(work_completion)(&gadget->work) ->kernfs_notify_work ->(work_completion)(&blkg->free_work) ->async_lookup_work ->autoload_work ->(work_completion)(&barr->work) ->drain_vmap_work ->netstamp_work ->reg_work ->(work_completion)(&fw_work->work) ->(delayed_fput_work).work ->(work_completion)(&s->destroy_work) ->(work_completion)(&(&ovs_net->masks_rebalance)->work) ->(work_completion)(&aux->work) ->(work_completion)(&ht->run_work) ->(work_completion)(&w->w) ->(work_completion)(&(&krcp->krw_arr[i].rcu_work)->work) ->(deferred_probe_timeout_work).work ->(work_completion)(&w->work)#2 ->(regulator_init_complete_work).work ->(work_completion)(&cgrp->bpf.release_work) ->deferred_process_work ->(work_completion)(&data->fib_event_work) ->(work_completion)(&(&nsim_dev->trap_data->trap_report_dw)->work) ->(work_completion)(&(&hwstats->traffic_dw)->work) ->(work_completion)(&(&conn->info_timer)->work) ->(work_completion)(&rdev->wiphy_work) ->wireless_nlevent_work ->fqdir_free_work ->free_ipc_work ->(work_completion)(&msk->work) ->&rq->__lock ->(work_completion)(&smcibdev->port_event_work) ->(work_completion)(&(&chan->chan_timer)->work) ->(work_completion)(&nlk->work) ->(ima_keys_delayed_work).work ->trans_destroy_work ->((tcp_md5_needed).work).work ->(work_completion)(&rdev->destroy_work) ->(work_completion)(&umem->work) ->(work_completion)(&work->work)#3 ->(work_completion)(&(&cnet->ecache.dwork)->work) ->(work_completion)(&data->dm_alert_work) ->(work_completion)(&aux->work)#2 ->(work_completion)(&(&krcp->page_cache_work)->work) FD: 30 BD: 2 +.+.: (work_completion)(&sscs.work) ->&x->wait#5 ->&obj_hash[i].lock ->hrtimer_bases.lock ->&x->wait#4 FD: 1 BD: 145 -.-.: &x->wait#5 FD: 2 BD: 221 +.+.: &newf->file_lock ->&newf->resize_wait FD: 1 BD: 1 ....: &p->vtime.seqcount FD: 35 BD: 142 +.+.: mem_hotplug_lock ->mem_hotplug_lock.rss.gp_wait.lock FD: 3 BD: 143 ..-.: mem_hotplug_lock.rss.gp_wait.lock ->&obj_hash[i].lock FD: 1 BD: 142 ....: mem_hotplug_lock.waiters.lock FD: 290 BD: 1 +.+.: cpu_add_remove_lock ->cpu_hotplug_lock ->cpu_hotplug_lock.waiters.lock ->cpu_hotplug_lock.rss.gp_wait.lock ->spec_ctrl_mutex ->cpuset_hotplug_work FD: 3 BD: 142 ..-.: cpu_hotplug_lock.rss.gp_wait.lock ->&obj_hash[i].lock FD: 1 BD: 2 ....: cpu_hotplug_lock.waiters.lock FD: 1 BD: 2 +.+.: cpuset_hotplug_work FD: 1 BD: 143 +.+.: pcp_batch_high_lock FD: 1 BD: 142 +.+.: relay_channels_mutex FD: 1 BD: 150 ....: rtc_lock FD: 161 BD: 148 +.+.: sparse_irq_lock ->tk_core.seq.seqcount ->rtc_lock ->&x->wait#6 ->&rq->__lock ->&p->pi_lock ->&irq_desc_lock_class ->fs_reclaim ->pool_lock#2 ->lock ->&root->kernfs_rwsem ->&c->lock ->&____s->seqcount ->&zone->lock ->pcpu_alloc_mutex ->&obj_hash[i].lock ->&n->list_lock FD: 24 BD: 149 ....: &x->wait#6 ->&p->pi_lock FD: 1 BD: 5121 ....: &rq->__lock/1 FD: 1 BD: 5121 -.-.: &cfs_rq->removed.lock FD: 1 BD: 149 ....: &x->wait#7 FD: 2 BD: 5121 ....: &rt_b->rt_runtime_lock ->&rt_rq->rt_runtime_lock FD: 1 BD: 5122 ....: &rt_rq->rt_runtime_lock FD: 27 BD: 142 +.+.: stop_cpus_mutex ->&stopper->lock ->&stop_pi_lock ->&rq->__lock ->&x->wait#8 FD: 24 BD: 144 ....: &x->wait#8 ->&p->pi_lock FD: 131 BD: 1 +.+.: sched_domains_mutex ->fs_reclaim ->pool_lock#2 ->&obj_hash[i].lock ->pcpu_alloc_mutex ->&zone->lock ->&____s->seqcount ->&c->lock ->pcpu_lock FD: 1 BD: 5121 ....: &cp->lock FD: 1 BD: 1 +.+.: (memory_chain).rwsem FD: 138 BD: 1 +.+.: &type->s_umount_key#5/1 ->fs_reclaim ->pool_lock#2 ->pcpu_alloc_mutex ->shrinker_rwsem ->list_lrus_mutex ->sb_lock ->&obj_hash[i].lock ->percpu_counters_lock ->crngs.lock ->&sbinfo->stat_lock ->mmu_notifier_invalidate_range_start ->&sb->s_type->i_lock_key#5 ->&s->s_inode_list_lock ->tk_core.seq.seqcount ->batched_entropy_u32.lock ->&dentry->d_lock FD: 55 BD: 4540 +.+.: &sb->s_type->i_lock_key#5 ->&dentry->d_lock ->&xa->xa_lock#9 FD: 24 BD: 1 ....: (setup_done).wait.lock ->&p->pi_lock FD: 139 BD: 26 ++++: namespace_sem ->fs_reclaim ->&zone->lock ->&____s->seqcount ->pool_lock#2 ->&c->lock ->mnt_id_ida.xa_lock ->pcpu_alloc_mutex ->&dentry->d_lock ->mount_lock ->rename_lock ->&obj_hash[i].lock ->&new_ns->ns_lock ->stock_lock ->&____s->seqcount#2 ->rcu_node_0 ->&rq->__lock ->&rcu_state.expedited_wq ->&n->list_lock ->remove_cache_srcu ->namespace_sem.wait_lock ->pcpu_alloc_mutex.wait_lock ->&p->pi_lock ->batched_entropy_u8.lock ->kfence_freelist_lock ->&cfs_rq->removed.lock FD: 1 BD: 211 +.+.: &____s->seqcount#3 FD: 127 BD: 1 +.+.: &type->s_umount_key#6 ->fs_reclaim ->pool_lock#2 ->&dentry->d_lock ->&zone->lock ->&____s->seqcount ->&c->lock ->&lru->node[i].lock ->&sbinfo->stat_lock ->&obj_hash[i].lock FD: 24 BD: 4600 +.+.: &lru->node[i].lock FD: 141 BD: 8 ++++: &sb->s_type->i_mutex_key ->namespace_sem ->fs_reclaim ->pool_lock#2 ->&dentry->d_lock ->rename_lock.seqcount ->tomoyo_ss ->tk_core.seq.seqcount ->&sb->s_type->i_lock_key#2 ->&wb->list_lock ->&c->lock ->&zone->lock ->&____s->seqcount ->&obj_hash[i].lock ->&rq->__lock ->&cfs_rq->removed.lock ->&dentry->d_lock/1 FD: 41 BD: 93 +.+.: rename_lock ->rename_lock.seqcount FD: 40 BD: 227 +.+.: rename_lock.seqcount ->&dentry->d_lock ->&dentry->d_lock/2 FD: 1 BD: 207 ....: &new_ns->poll FD: 2 BD: 4601 +.+.: &____s->seqcount#4 ->&____s->seqcount#4/1 FD: 41 BD: 219 +.+.: &fs->lock ->&____s->seqcount#3 ->&dentry->d_lock FD: 1 BD: 175 +.+.: req_lock FD: 147 BD: 1 +.+.: of_mutex ->fs_reclaim ->pool_lock#2 ->lock ->&root->kernfs_rwsem FD: 1 BD: 210 ....: &x->wait#9 FD: 1 BD: 236 +.+.: &k->list_lock FD: 23 BD: 216 ++++: bus_type_sem ->&rq->__lock FD: 31 BD: 4185 -...: &dev->power.lock ->&dev->power.lock/1 ->&dev->power.wait_queue ->hrtimer_bases.lock FD: 23 BD: 212 +.+.: dpm_list_mtx ->&rq->__lock FD: 136 BD: 226 +.+.: uevent_sock_mutex ->fs_reclaim ->pool_lock#2 ->nl_table_lock ->&obj_hash[i].lock ->nl_table_wait.lock ->&c->lock ->&zone->lock ->&rq->__lock ->&____s->seqcount ->&n->list_lock ->&cfs_rq->removed.lock ->rlock-AF_NETLINK ->batched_entropy_u8.lock ->kfence_freelist_lock ->rcu_node_0 ->remove_cache_srcu ->&meta->lock ->quarantine_lock ->mmu_notifier_invalidate_range_start ->uevent_sock_mutex.wait_lock ->&____s->seqcount#2 ->&rcu_state.expedited_wq ->&base->lock ->&data->lock ->pgd_lock ->stock_lock ->key ->pcpu_lock ->percpu_counters_lock FD: 1 BD: 202 ....: running_helpers_waitq.lock FD: 1 BD: 226 +.+.: sysfs_symlink_target_lock FD: 2 BD: 286 +.+.: &k->k_lock ->klist_remove_lock FD: 1 BD: 1 ....: &dev->mutex FD: 1 BD: 1 +.+.: subsys mutex FD: 2 BD: 1 +.+.: memory_blocks.xa_lock ->pool_lock#2 FD: 1 BD: 1 +.+.: subsys mutex#2 FD: 128 BD: 12 +.+.: register_lock ->proc_subdir_lock ->fs_reclaim ->pool_lock#2 ->proc_inum_ida.xa_lock ->&c->lock ->&zone->lock ->&____s->seqcount ->&obj_hash[i].lock FD: 1 BD: 2 +.+.: (pm_chain_head).rwsem FD: 1 BD: 1 +.+.: cpufreq_governor_mutex FD: 35 BD: 2 +.+.: (work_completion)(&rew->rew_work) ->rcu_node_0 ->rcu_state.exp_wake_mutex ->&rcu_state.expedited_wq ->&obj_hash[i].lock ->&base->lock ->&rq->__lock ->(&timer.timer) ->&pool->lock ->pool_lock#2 ->pool_lock ->rcu_state.exp_wake_mutex.wait_lock ->&p->pi_lock ->&cfs_rq->removed.lock FD: 1 BD: 1 +.+.: dyn_event_ops_mutex FD: 1 BD: 2 ++++: binfmt_lock FD: 1 BD: 114 +.+.: pin_fs_lock FD: 137 BD: 1 +.+.: &type->s_umount_key#7/1 ->fs_reclaim ->pool_lock#2 ->pcpu_alloc_mutex ->shrinker_rwsem ->list_lrus_mutex ->sb_lock ->&zone->lock ->&____s->seqcount ->&c->lock ->mmu_notifier_invalidate_range_start ->&sb->s_type->i_lock_key#6 ->&s->s_inode_list_lock ->tk_core.seq.seqcount ->&dentry->d_lock FD: 40 BD: 3 +.+.: &sb->s_type->i_lock_key#6 ->&dentry->d_lock FD: 128 BD: 1 +.+.: &sb->s_type->i_mutex_key#2 ->&sb->s_type->i_lock_key#6 ->rename_lock.seqcount ->fs_reclaim ->pool_lock#2 ->&dentry->d_lock ->mmu_notifier_invalidate_range_start ->&s->s_inode_list_lock ->tk_core.seq.seqcount ->&c->lock ->&zone->lock ->&____s->seqcount FD: 24 BD: 4601 ....: &wq ->&p->pi_lock FD: 1 BD: 37 +.+.: chrdevs_lock FD: 958 BD: 1 ++++: cb_lock ->genl_mutex ->fs_reclaim ->pool_lock#2 ->rlock-AF_NETLINK ->&c->lock ->&n->list_lock ->rtnl_mutex ->&obj_hash[i].lock ->&rdev->wiphy.mtx ->nlk_cb_mutex-GENERIC ->&____s->seqcount ->&rq->__lock ->genl_mutex.wait_lock ->&p->pi_lock ->rtnl_mutex.wait_lock ->&lock->wait_lock ->&____s->seqcount#2 ->(console_sem).lock ->console_owner_lock ->console_owner ->rcu_node_0 ->remove_cache_srcu ->tk_core.seq.seqcount ->&dir->lock#2 ->quarantine_lock ->batched_entropy_u8.lock ->kfence_freelist_lock ->&rcu_state.expedited_wq ->&meta->lock ->&fw_cache.lock ->stock_lock ->async_lock ->init_task.alloc_lock ->&dentry->d_lock ->&type->i_mutex_dir_key#3 ->&sb->s_type->i_lock_key#22 ->umhelper_sem ->fw_lock ->&x->wait#23 ->&cfs_rq->removed.lock ->pgd_lock ->key ->pcpu_lock ->percpu_counters_lock ->&data->lock ->reg_requests_lock FD: 930 BD: 5 +.+.: genl_mutex ->fs_reclaim ->pool_lock#2 ->nl_table_lock ->nl_table_wait.lock ->&c->lock ->&____s->seqcount ->&obj_hash[i].lock ->rlock-AF_NETLINK ->&n->list_lock ->rtnl_mutex ->rtnl_mutex.wait_lock ->&p->pi_lock ->genl_mutex.wait_lock ->&rq->__lock ->rcu_node_0 ->hwsim_radio_lock ->&x->wait#9 ->&zone->lock ->batched_entropy_u32.lock ->&k->list_lock ->gdp_mutex ->lock ->&root->kernfs_rwsem ->bus_type_sem ->sysfs_symlink_target_lock ->&dev->power.lock ->dpm_list_mtx ->uevent_sock_mutex ->subsys mutex#53 ->device_links_lock ->&k->k_lock ->deferred_probe_mutex ->cpu_hotplug_lock ->wq_pool_mutex ->crngs.lock ->triggers_list_lock ->leds_list_lock ->rfkill_global_mutex ->rfkill_global_mutex.wait_lock ->pin_fs_lock ->&sb->s_type->i_mutex_key#3 ->(inetaddr_chain).rwsem ->inet6addr_chain.lock ->&____s->seqcount#2 ->&cfs_rq->removed.lock ->(console_sem).lock ->console_owner_lock ->console_owner ->tcp_metrics_lock ->&hash->list_locks[i] ->&ht->lock ->stock_lock ->&pernet->lock ->sk_lock-AF_INET ->slock-AF_INET ->calipso_doi_list_lock ->__ip_vs_mutex ->&pn->l2tp_tunnel_idr_lock ->mmu_notifier_invalidate_range_start ->&sb->s_type->i_lock_key#8 ->&dir->lock ->l2tp_ip6_lock ->k-sk_lock-AF_INET6 ->k-slock-AF_INET6 ->k-clock-AF_INET6 ->&xa->xa_lock#9 ->&fsnotify_mark_srcu ->nbd_index_mutex ->&nbd->config_lock ->quarantine_lock ->remove_cache_srcu ->&lock->wait_lock ->k-sk_lock-AF_INET ->k-slock-AF_INET ->&table->hash[i].lock ->k-clock-AF_INET ->&rcu_state.expedited_wq ->&sdata->sec_mtx ->smcd_dev_list.mutex ->batched_entropy_u8.lock ->kfence_freelist_lock ->&meta->lock ->&sem->wait_lock ->&bat_priv->gw.list_lock ->&fn->fou_lock ->sk_lock-AF_NETLINK ->slock-AF_NETLINK ->rcu_state.barrier_mutex ->dev_base_lock ->lweventlist_lock ->krc.lock ->&dir->lock#2 ->netdev_unregistering_wq.lock ->nbd_index_mutex.wait_lock ->net_dm_mutex ->&data->lock ->key#32 FD: 137 BD: 1 +.+.: &type->s_umount_key#8/1 ->fs_reclaim ->pool_lock#2 ->pcpu_alloc_mutex ->shrinker_rwsem ->list_lrus_mutex ->sb_lock ->mmu_notifier_invalidate_range_start ->&sb->s_type->i_lock_key#7 ->&s->s_inode_list_lock ->tk_core.seq.seqcount ->&dentry->d_lock FD: 40 BD: 107 +.+.: &sb->s_type->i_lock_key#7 ->&dentry->d_lock FD: 141 BD: 105 +.+.: &sb->s_type->i_mutex_key#3 ->&sb->s_type->i_lock_key#7 ->rename_lock.seqcount ->fs_reclaim ->pool_lock#2 ->&dentry->d_lock ->mmu_notifier_invalidate_range_start ->&s->s_inode_list_lock ->tk_core.seq.seqcount ->&c->lock ->&____s->seqcount ->&zone->lock ->&obj_hash[i].lock ->batched_entropy_u8.lock ->kfence_freelist_lock ->&rq->__lock ->&cfs_rq->removed.lock ->&base->lock ->&n->list_lock ->(console_sem).lock ->rcu_node_0 ->remove_cache_srcu ->&rcu_state.expedited_wq ->pin_fs_lock ->mount_lock ->&fsnotify_mark_srcu ->&xa->xa_lock#9 ->&____s->seqcount#2 ->&xa->xa_lock#4 ->stock_lock ->per_cpu_ptr(&cgroup_rstat_cpu_lock, cpu) FD: 1 BD: 4 +.+.: subsys mutex#3 FD: 4 BD: 7 ....: async_lock ->&obj_hash[i].lock ->pool_lock#2 FD: 395 BD: 2 +.+.: (work_completion)(&entry->work) ->tk_core.seq.seqcount ->&dev->power.lock ->&k->list_lock ->sysfs_symlink_target_lock ->fs_reclaim ->pool_lock#2 ->lock ->&root->kernfs_rwsem ->&x->wait#9 ->&obj_hash[i].lock ->&c->lock ->&zone->lock ->&____s->seqcount ->uevent_sock_mutex ->running_helpers_waitq.lock ->&k->k_lock ->async_lock ->async_done.lock ->&dentry->d_lock ->&sb->s_type->i_mutex_key ->sb_writers#2 ->&pool->lock/1 ->cpu_hotplug_lock ->wq_pool_mutex ->pcpu_alloc_mutex ->batched_entropy_u32.lock ->&n->list_lock ->mmu_notifier_invalidate_range_start ->blk_queue_ida.xa_lock ->&q->sysfs_lock ->&set->tag_list_lock ->bio_slab_lock ->percpu_counters_lock ->&sb->s_type->i_lock_key#3 ->&s->s_inode_list_lock ->&xa->xa_lock#11 ->&q->mq_freeze_lock ->percpu_ref_switch_lock ->&q->queue_lock ->major_names_lock ->floppy_lock ->rtc_lock ->&wq->mutex ->&desc->request_mutex ->register_lock ->&irq_desc_lock_class ->proc_subdir_lock ->proc_inum_ida.xa_lock ->resource_lock ->&base->lock ->&rq->__lock ->(&timer.timer) ->command_done.lock ->&cfs_rq->removed.lock ->&shost->scan_mutex ->(console_sem).lock ->console_owner_lock ->console_owner ->async_scan_lock ->&q->debugfs_mutex ->klist_remove_lock ->kernfs_idr_lock ->(&motor_off_timer[drive]) ->&xa->xa_lock#10 ->&q->unused_hctx_lock ->(&sq->pending_timer) ->(work_completion)(&td->dispatch_work) ->&q->blkcg_mutex ->pcpu_lock ->&xa->xa_lock#9 ->&fsnotify_mark_srcu FD: 1 BD: 22 .+.+: device_links_srcu FD: 3 BD: 21 +.+.: fwnode_link_lock ->&k->k_lock FD: 26 BD: 123 +.+.: device_links_lock ->&k->list_lock ->&k->k_lock ->&rq->__lock FD: 1 BD: 3 +.+.: regulator_list_mutex FD: 1 BD: 4 ....: &dev->devres_lock FD: 1 BD: 4 +.+.: regulator_nesting_mutex FD: 2 BD: 1 +.+.: regulator_ww_class_mutex ->regulator_nesting_mutex FD: 150 BD: 184 +.+.: gdp_mutex ->&k->list_lock ->fs_reclaim ->pool_lock#2 ->lock ->&root->kernfs_rwsem ->kobj_ns_type_lock ->&c->lock ->&zone->lock ->&____s->seqcount ->&obj_hash[i].lock ->sysfs_symlink_target_lock ->&____s->seqcount#2 ->&rq->__lock ->&sem->wait_lock ->&p->pi_lock ->kernfs_idr_lock ->&n->list_lock ->remove_cache_srcu FD: 3 BD: 3 +.+.: subsys mutex#4 ->&k->k_lock FD: 23 BD: 123 +.+.: deferred_probe_mutex ->&rq->__lock FD: 1 BD: 20 ....: probe_waitqueue.lock FD: 1 BD: 3 ....: async_done.lock FD: 136 BD: 1 +.+.: &type->s_umount_key#9/1 ->fs_reclaim ->pool_lock#2 ->pcpu_alloc_mutex ->shrinker_rwsem ->list_lrus_mutex ->sb_lock ->&zone->lock ->&____s->seqcount ->&c->lock ->mmu_notifier_invalidate_range_start ->&sb->s_type->i_lock_key#8 ->&s->s_inode_list_lock ->tk_core.seq.seqcount ->&dentry->d_lock FD: 41 BD: 4542 +.+.: &sb->s_type->i_lock_key#8 ->&dentry->d_lock ->bit_wait_table + i FD: 135 BD: 87 +.+.: pack_mutex ->fs_reclaim ->&zone->lock ->&____s->seqcount ->pool_lock#2 ->free_vmap_area_lock ->vmap_area_lock ->init_mm.page_table_lock ->rcu_node_0 ->&rq->__lock ->vmap_purge_lock ->cpa_lock ->text_mutex ->pack_mutex.wait_lock FD: 23 BD: 86 +.+.: &fp->aux->used_maps_mutex ->&rq->__lock FD: 1 BD: 1 +.+.: proto_list_mutex FD: 1 BD: 1 +.+.: targets_mutex FD: 26 BD: 4167 ...-: nl_table_lock ->pool_lock#2 ->&c->lock ->nl_table_wait.lock ->&obj_hash[i].lock ->&____s->seqcount ->&meta->lock ->kfence_freelist_lock ->&n->list_lock FD: 24 BD: 4168 ..-.: nl_table_wait.lock ->&p->pi_lock FD: 1 BD: 1 +.+.: net_family_lock FD: 2 BD: 5 ....: net_generic_ids.xa_lock ->pool_lock#2 FD: 4 BD: 158 ..-.: &dir->lock ->&obj_hash[i].lock ->pool_lock#2 FD: 31 BD: 5 +.+.: k-sk_lock-AF_NETLINK ->k-slock-AF_NETLINK ->&rq->__lock FD: 1 BD: 6 +...: k-slock-AF_NETLINK FD: 2 BD: 3942 ..-.: rhashtable_bucket ->rhashtable_bucket/1 FD: 1 BD: 10 ....: &list->lock FD: 24 BD: 10 ....: kauditd_wait.lock ->&p->pi_lock FD: 3 BD: 2 +.+.: lock#2 ->&zone->lock FD: 130 BD: 1 +.+.: khugepaged_mutex ->fs_reclaim ->pool_lock#2 ->kthread_create_lock ->&p->pi_lock ->&x->wait ->&rq->__lock ->&obj_hash[i].lock ->lock#2 ->pcp_batch_high_lock FD: 3 BD: 14 +.+.: subsys mutex#5 ->&k->k_lock FD: 4 BD: 1 +.+.: subsys mutex#6 ->&k->list_lock ->&k->k_lock FD: 1 BD: 1 +.+.: regmap_debugfs_early_lock FD: 1 BD: 1 +.+.: (acpi_reconfig_chain).rwsem FD: 1 BD: 1 +.+.: __i2c_board_lock FD: 129 BD: 1 +.+.: core_lock ->&k->list_lock ->&k->k_lock ->&rq->__lock ->&cfs_rq->removed.lock ->&obj_hash[i].lock ->pool_lock#2 ->fs_reclaim FD: 2 BD: 1 +.+.: thermal_governor_lock ->thermal_list_lock FD: 1 BD: 2 +.+.: thermal_list_lock FD: 167 BD: 1 +.+.: cpuidle_lock ->&rnp->exp_lock ->rcu_state.exp_mutex ->&obj_hash[i].lock ->(console_sem).lock ->fs_reclaim ->pool_lock#2 ->lock ->&root->kernfs_rwsem ->&c->lock ->&zone->lock ->&____s->seqcount FD: 1 BD: 198 +.+.: rcu_state.exp_mutex.wait_lock FD: 24 BD: 4552 -.-.: &rcu_state.expedited_wq ->&p->pi_lock FD: 43 BD: 145 ....: freezer_lock ->&sighand->siglock ->&p->pi_lock FD: 1 BD: 1 ....: printk_ratelimit_state.lock FD: 1 BD: 1 ....: audit_backlog_wait.lock FD: 129 BD: 1 +.+.: k-sk_lock-AF_QIPCRTR ->k-slock-AF_QIPCRTR ->fs_reclaim ->qrtr_ports.xa_lock ->pool_lock#2 ->qrtr_node_lock ->&obj_hash[i].lock FD: 1 BD: 2 +...: k-slock-AF_QIPCRTR FD: 1 BD: 2 +.+.: qrtr_ports.xa_lock FD: 1 BD: 2 +.+.: qrtr_node_lock FD: 128 BD: 157 ++++: (crypto_chain).rwsem ->fs_reclaim ->&c->lock ->&____s->seqcount ->pool_lock#2 ->kthread_create_lock ->&p->pi_lock ->&x->wait ->&rq->__lock ->&obj_hash[i].lock ->&zone->lock ->&cfs_rq->removed.lock ->&____s->seqcount#2 ->&n->list_lock FD: 288 BD: 1 +.+.: iova_cache_mutex ->cpu_hotplug_lock ->slab_mutex FD: 3 BD: 1 +.+.: subsys mutex#7 ->&k->k_lock FD: 1 BD: 148 ....: pci_config_lock FD: 1 BD: 1 +.+.: subsys mutex#8 FD: 130 BD: 116 +.+.: dev_pm_qos_mtx ->fs_reclaim ->pool_lock#2 ->&dev->power.lock ->pm_qos_lock ->&c->lock ->&zone->lock ->&____s->seqcount ->&rq->__lock FD: 1 BD: 117 ....: pm_qos_lock FD: 153 BD: 115 +.+.: dev_pm_qos_sysfs_mtx ->dev_pm_qos_mtx ->&root->kernfs_rwsem ->fs_reclaim ->pool_lock#2 ->lock ->&c->lock ->&____s->seqcount ->&zone->lock ->&rq->__lock ->&sem->wait_lock ->&p->pi_lock ->dev_pm_qos_sysfs_mtx.wait_lock FD: 126 BD: 1 +.+.: mtrr_mutex ->fs_reclaim ->pool_lock#2 FD: 1 BD: 1 ..-.: uidhash_lock FD: 12 BD: 5 +.+.: (work_completion)(work) ->lock#4 ->lock#5 FD: 133 BD: 1 +.+.: (work_completion)(&eval_map_work) ->trace_event_sem ->trace_event_sem.wait_lock ->&p->pi_lock ->&rq->__lock FD: 1 BD: 1 ....: oom_reaper_wait.lock FD: 1 BD: 1 +.+.: subsys mutex#9 FD: 25 BD: 3 ..-.: &(&ssp->srcu_sup->work)->timer FD: 24 BD: 4515 ....: &pgdat->kcompactd_wait ->&p->pi_lock FD: 131 BD: 2 +.+.: pcpu_balance_work ->pcpu_alloc_mutex ->pcpu_alloc_mutex.wait_lock ->&p->pi_lock ->&rq->__lock FD: 1 BD: 221 +.+.: pcpu_alloc_mutex.wait_lock FD: 171 BD: 1 +.+.: memory_tier_lock ->fs_reclaim ->pool_lock#2 ->&x->wait#9 ->&obj_hash[i].lock ->&k->list_lock ->lock ->&root->kernfs_rwsem ->bus_type_sem ->sysfs_symlink_target_lock ->&k->k_lock ->&c->lock ->&____s->seqcount ->&dev->power.lock ->dpm_list_mtx ->uevent_sock_mutex ->running_helpers_waitq.lock ->&zone->lock ->subsys mutex#10 FD: 1 BD: 2 +.+.: subsys mutex#10 FD: 1 BD: 1 +.+.: ksm_thread_mutex FD: 1 BD: 1 ....: ksm_thread_wait.lock FD: 1 BD: 2 +.+.: damon_ops_lock FD: 129 BD: 156 ++++: crypto_alg_sem ->(crypto_chain).rwsem ->&rq->__lock FD: 37 BD: 3 +.+.: lock#3 ->&rq->__lock ->&obj_hash[i].lock ->(work_completion)(work) ->&x->wait#10 ->&cfs_rq->removed.lock ->pool_lock#2 ->pool_lock FD: 1 BD: 4569 ..-.: quarantine_lock FD: 35 BD: 4498 .+.+: remove_cache_srcu ->quarantine_lock ->&c->lock ->&n->list_lock ->pool_lock#2 ->&obj_hash[i].lock ->rcu_node_0 ->&rq->__lock ->&____s->seqcount ->&rcu_state.expedited_wq ->&cfs_rq->removed.lock ->pgd_lock ->key ->pcpu_lock ->percpu_counters_lock ->&base->lock ->stock_lock ->&meta->lock ->kfence_freelist_lock FD: 154 BD: 2 +.+.: (work_completion)(&pwq->unbound_release_work) ->&wq->mutex ->wq_pool_mutex ->&obj_hash[i].lock ->pool_lock#2 ->rcu_node_0 ->&pool->lock ->&rnp->exp_wq[0] ->&rq->__lock ->&rnp->exp_wq[1] ->&rnp->exp_lock ->rcu_state.exp_mutex ->&rnp->exp_wq[2] ->rcu_state.exp_mutex.wait_lock ->&p->pi_lock ->&rnp->exp_wq[3] ->&cfs_rq->removed.lock ->&rcu_state.expedited_wq FD: 1 BD: 153 +.+.: khugepaged_mm_lock FD: 24 BD: 153 ....: khugepaged_wait.lock ->&p->pi_lock FD: 155 BD: 7 +.+.: bio_slab_lock ->fs_reclaim ->pool_lock#2 ->slab_mutex ->bio_slabs.xa_lock FD: 2 BD: 8 +.+.: bio_slabs.xa_lock ->pool_lock#2 FD: 127 BD: 3 +.+.: major_names_lock ->fs_reclaim ->pool_lock#2 ->major_names_spinlock ->&c->lock ->&____s->seqcount FD: 1 BD: 4 +.+.: major_names_spinlock FD: 3 BD: 14 +.+.: subsys mutex#11 ->&k->k_lock FD: 26 BD: 1 +.-.: (&rtpcp->lazy_timer) ->rcu_tasks_trace__percpu.cbs_pcpu_lock ->rcu_tasks__percpu.cbs_pcpu_lock FD: 1 BD: 1 ....: *(&acpi_gbl_hardware_lock) FD: 38 BD: 1 ....: *(&acpi_gbl_gpe_lock) ->(console_sem).lock FD: 5 BD: 160 ....: mask_lock ->tmp_mask_lock FD: 4 BD: 161 -...: tmp_mask_lock ->vector_lock ->ioapic_lock FD: 1 BD: 1 -...: shrink_qlist.lock FD: 26 BD: 7 ....: remove_cache_srcu_srcu_usage.lock ->&obj_hash[i].lock ->&ACCESS_PRIVATE(sdp, lock) FD: 29 BD: 142 +.+.: flush_lock ->&obj_hash[i].lock ->(work_completion)(&sfw->work) ->&x->wait#10 ->&rq->__lock FD: 10 BD: 144 +.+.: (work_completion)(&sfw->work) ->&c->lock ->&obj_hash[i].lock ->&n->list_lock FD: 27 BD: 143 +.+.: (wq_completion)slub_flushwq ->(work_completion)(&sfw->work) ->(work_completion)(&barr->work) FD: 24 BD: 4646 ....: &x->wait#10 ->&p->pi_lock FD: 25 BD: 150 +.+.: (work_completion)(&barr->work) ->&x->wait#10 ->&rq->__lock ->&cfs_rq->removed.lock ->&obj_hash[i].lock ->pool_lock#2 FD: 1 BD: 1 +.+.: system_transition_mutex FD: 1 BD: 1 +.+.: (power_off_prep_handler_list).rwsem FD: 1 BD: 1 ....: power_off_handler_list.lock FD: 1 BD: 1 +.+.: (restart_prep_handler_list).rwsem FD: 1 BD: 1 +.+.: (reboot_notifier_list).rwsem FD: 205 BD: 1 +.+.: acpi_scan_lock ->semaphore->lock ->fs_reclaim ->pool_lock#2 ->&obj_hash[i].lock ->&c->lock ->&zone->lock ->&____s->seqcount ->&x->wait#9 ->acpi_device_lock ->&k->list_lock ->lock ->&root->kernfs_rwsem ->bus_type_sem ->sysfs_symlink_target_lock ->&k->k_lock ->&dev->power.lock ->dpm_list_mtx ->subsys mutex#12 ->uevent_sock_mutex ->running_helpers_waitq.lock ->*(&acpi_gbl_reference_count_lock) ->&n->list_lock ->pci_config_lock ->batched_entropy_u8.lock ->kfence_freelist_lock ->&meta->lock ->quarantine_lock ->(console_sem).lock ->pci_bus_sem ->pci_mmcfg_lock ->resource_lock ->&device->physical_node_lock ->fwnode_link_lock ->devtree_lock ->gdp_mutex ->subsys mutex#13 ->pci_acpi_companion_lookup_sem ->pci_slot_mutex ->tk_core.seq.seqcount ->resource_alignment_lock ->device_links_srcu ->subsys mutex#14 ->acpi_pm_notifier_install_lock ->&rq->__lock ->pci_rescan_remove_lock ->subsys mutex#3 ->acpi_link_lock ->acpi_dep_list_lock ->wakeup_ida.xa_lock ->subsys mutex#15 ->events_lock ->power_resource_list_lock FD: 128 BD: 2 +.+.: acpi_device_lock ->fs_reclaim ->pool_lock#2 ->&xa->xa_lock#2 ->semaphore->lock ->&obj_hash[i].lock ->&c->lock ->&zone->lock ->&____s->seqcount FD: 1 BD: 3 ....: &xa->xa_lock#2 FD: 1 BD: 2 +.+.: subsys mutex#12 FD: 1 BD: 2 ++++: pci_bus_sem FD: 1 BD: 2 +.+.: pci_mmcfg_lock FD: 148 BD: 14 +.+.: &device->physical_node_lock ->sysfs_symlink_target_lock ->fs_reclaim ->pool_lock#2 ->lock ->&root->kernfs_rwsem ->&c->lock ->&zone->lock ->&____s->seqcount ->&obj_hash[i].lock FD: 3 BD: 2 +.+.: subsys mutex#13 ->&k->k_lock FD: 1 BD: 2 .+.+: pci_acpi_companion_lookup_sem FD: 1 BD: 2 +.+.: pci_slot_mutex FD: 1 BD: 2 +.+.: resource_alignment_lock FD: 1 BD: 4186 ....: &dev->power.lock/1 FD: 1 BD: 2 +.+.: subsys mutex#14 FD: 173 BD: 2 +.+.: acpi_pm_notifier_install_lock ->semaphore->lock ->fs_reclaim ->pool_lock#2 ->*(&acpi_gbl_reference_count_lock) ->acpi_pm_notifier_lock FD: 170 BD: 3 +.+.: acpi_pm_notifier_lock ->fs_reclaim ->pool_lock#2 ->wakeup_ida.xa_lock ->&x->wait#9 ->&obj_hash[i].lock ->&k->list_lock ->gdp_mutex ->lock ->&root->kernfs_rwsem ->bus_type_sem ->sysfs_symlink_target_lock ->&c->lock ->&____s->seqcount ->uevent_sock_mutex ->running_helpers_waitq.lock ->&k->k_lock ->subsys mutex#15 ->events_lock ->&zone->lock FD: 1 BD: 7 ....: wakeup_ida.xa_lock FD: 25 BD: 7 +.+.: subsys mutex#15 ->&k->k_lock ->&rq->__lock FD: 1 BD: 7 ....: events_lock FD: 128 BD: 1 +.+.: &pgdat->kswapd_lock ->fs_reclaim ->pool_lock#2 ->kthread_create_lock ->&p->pi_lock ->&rq->__lock ->&x->wait ->&obj_hash[i].lock FD: 30 BD: 1 ..-.: drivers/char/random.c:251 FD: 24 BD: 2 +.+.: (next_reseed).work ->&obj_hash[i].lock ->&base->lock ->input_pool.lock ->base_crng.lock ->&rq->__lock FD: 35 BD: 2 +.+.: pci_rescan_remove_lock FD: 138 BD: 2 +.+.: acpi_link_lock ->fs_reclaim ->pool_lock#2 ->semaphore->lock ->&obj_hash[i].lock ->*(&acpi_gbl_reference_count_lock) ->pci_config_lock ->&zone->lock ->&____s->seqcount ->(console_sem).lock ->&c->lock ->&rq->__lock FD: 25 BD: 1 ..-.: mm/vmstat.c:2018 FD: 287 BD: 2 +.+.: (shepherd).work ->cpu_hotplug_lock ->&obj_hash[i].lock ->&base->lock ->&rq->__lock FD: 34 BD: 4 +.+.: (wq_completion)mm_percpu_wq ->(work_completion)(&(({ do { const void *__vpp_verify = (typeof((&vmstat_work) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __asm__ ("" : "=r"(__ptr) : "0"((typeof(*((&vmstat_work))) *)((&vmstat_work)))); (typeof((typeof(*((&vmstat_work))) *)((&vmstat_work)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); }))->work) ->(work_completion)(work) ->(work_completion)(&barr->work) ->&rq->__lock FD: 25 BD: 5 +.+.: (work_completion)(&(({ do { const void *__vpp_verify = (typeof((&vmstat_work) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __asm__ ("" : "=r"(__ptr) : "0"((typeof(*((&vmstat_work))) *)((&vmstat_work)))); (typeof((typeof(*((&vmstat_work))) *)((&vmstat_work)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); }))->work) ->&obj_hash[i].lock ->&base->lock ->&pcp->lock ->&rq->__lock FD: 1 BD: 2 +.+.: acpi_dep_list_lock FD: 1 BD: 2 +.+.: power_resource_list_lock FD: 176 BD: 7 ++++: &(&priv->bus_notifier)->rwsem ->fs_reclaim ->&c->lock ->&____s->seqcount ->pool_lock#2 ->lock ->&root->kernfs_rwsem ->&zone->lock ->i2c_dev_list_lock ->&x->wait#9 ->&obj_hash[i].lock ->chrdevs_lock ->&k->list_lock ->gdp_mutex ->bus_type_sem ->sysfs_symlink_target_lock ->&dev->power.lock ->dpm_list_mtx ->req_lock ->&p->pi_lock ->&rq->__lock ->&x->wait#11 ->uevent_sock_mutex ->running_helpers_waitq.lock ->&k->k_lock ->subsys mutex#66 FD: 137 BD: 1 +.+.: &type->s_umount_key#10/1 ->fs_reclaim ->pool_lock#2 ->pcpu_alloc_mutex ->shrinker_rwsem ->list_lrus_mutex ->sb_lock ->&zone->lock ->&____s->seqcount ->&c->lock ->mmu_notifier_invalidate_range_start ->&sb->s_type->i_lock_key#9 ->&s->s_inode_list_lock ->tk_core.seq.seqcount ->&dentry->d_lock FD: 40 BD: 2 +.+.: &sb->s_type->i_lock_key#9 ->&dentry->d_lock FD: 137 BD: 1 +.+.: &type->s_umount_key#11/1 ->fs_reclaim ->pool_lock#2 ->pcpu_alloc_mutex ->shrinker_rwsem ->&c->lock ->&zone->lock ->&____s->seqcount ->list_lrus_mutex ->sb_lock ->mmu_notifier_invalidate_range_start ->&sb->s_type->i_lock_key#10 ->&s->s_inode_list_lock ->tk_core.seq.seqcount ->&dentry->d_lock FD: 40 BD: 2 +.+.: &sb->s_type->i_lock_key#10 ->&dentry->d_lock FD: 229 BD: 152 ++++: &mm->mmap_lock ->reservation_ww_class_acquire ->fs_reclaim ->&zone->lock ->&____s->seqcount ->pool_lock#2 ->&c->lock ->&mm->page_table_lock ->ptlock_ptr(page) ->&anon_vma->rwsem ->per_cpu_ptr(&cgroup_rstat_cpu_lock, cpu) ->ptlock_ptr(page)#2 ->lock#4 ->lock#5 ->mmu_notifier_invalidate_range_start ->&vma->vm_lock->lock ->&obj_hash[i].lock ->&lruvec->lru_lock ->rcu_node_0 ->&rq->__lock ->&cfs_rq->removed.lock ->&meta->lock ->kfence_freelist_lock ->quarantine_lock ->batched_entropy_u8.lock ->&mapping->i_mmap_rwsem ->resource_lock ->&p->alloc_lock ->tk_core.seq.seqcount ->&mm->mmap_lock/1 ->&n->list_lock ->&sem->wait_lock ->&p->pi_lock ->remove_cache_srcu ->&folio_wait_table[i] ->pgd_lock ->key ->pcpu_lock ->percpu_counters_lock ->pool_lock ->&rcu_state.expedited_wq ->khugepaged_mm_lock ->khugepaged_wait.lock ->&sb->s_type->i_lock_key ->&xa->xa_lock#9 ->&info->lock ->mount_lock ->&wb->list_lock ->&kcov->lock ->stock_lock ->&____s->seqcount#2 ->sb_pagefaults ->&mapping->private_lock ->&xa->xa_lock#4 ->&s->s_inode_list_lock ->batched_entropy_u32.lock ->&base->lock ->&po->pg_vec_lock ->&dd->lock ->key#22 ->&sb->s_type->i_mutex_key#21 ->&hugetlbfs_i_mmap_rwsem_key ->&vma_lock->rw_sema ->&hugetlb_fault_mutex_table[i] ->mapping.invalidate_lock ->lock#10 ->key#8 ->&pgdat->kswapd_wait FD: 132 BD: 169 +.+.: reservation_ww_class_acquire ->reservation_ww_class_mutex FD: 131 BD: 170 +.+.: reservation_ww_class_mutex ->fs_reclaim ->&shmem->vmap_lock FD: 66 BD: 4515 ++++: &mapping->i_mmap_rwsem ->&obj_hash[i].lock ->pool_lock#2 ->&zone->lock ->&____s->seqcount ->&anon_vma->rwsem ->quarantine_lock ->&meta->lock ->kfence_freelist_lock ->&rq->__lock ->&sem->wait_lock ->mmu_notifier_invalidate_range_start ->ptlock_ptr(page) ->rcu_node_0 ->pool_lock ->&p->pi_lock ->&base->lock ->&cfs_rq->removed.lock ->pgd_lock ->key ->pcpu_lock ->percpu_counters_lock ->ptlock_ptr(page)#2 ->lock#4 ->lock#5 ->&rcu_state.expedited_wq ->&lruvec->lru_lock ->stock_lock FD: 1 BD: 4535 +.+.: dma_fence_map FD: 24 BD: 3 +.+.: delayed_uprobe_lock ->&rq->__lock ->delayed_uprobe_lock.wait_lock ->&cfs_rq->removed.lock ->&obj_hash[i].lock ->pool_lock#2 FD: 1 BD: 4415 ....: key FD: 1 BD: 4 +.+.: attribute_container_mutex FD: 141 BD: 19 ++++: triggers_list_lock ->&led_cdev->trigger_lock ->&rq->__lock FD: 141 BD: 19 ++++: leds_list_lock ->&led_cdev->trigger_lock FD: 188 BD: 2 ++++: (usb_notifier_list).rwsem ->fs_reclaim ->pool_lock#2 ->pin_fs_lock ->&sb->s_type->i_mutex_key#3 ->&x->wait#9 ->&obj_hash[i].lock ->&k->list_lock ->gdp_mutex ->lock ->&root->kernfs_rwsem ->bus_type_sem ->sysfs_symlink_target_lock ->&c->lock ->&zone->lock ->&____s->seqcount ->&dev->power.lock ->dpm_list_mtx ->req_lock ->&p->pi_lock ->&x->wait#11 ->&rq->__lock ->uevent_sock_mutex ->running_helpers_waitq.lock ->&k->k_lock ->subsys mutex#60 ->mon_lock ->&cfs_rq->removed.lock FD: 1 BD: 1 +.+.: rc_map_lock FD: 1 BD: 1 +.+.: subsys mutex#16 FD: 1 BD: 2 +.+.: &entry->access FD: 128 BD: 2 +.+.: info_mutex ->proc_subdir_lock ->fs_reclaim ->pool_lock#2 ->proc_inum_ida.xa_lock ->&c->lock ->&____s->seqcount ->&zone->lock FD: 1 BD: 185 +.+.: kobj_ns_type_lock FD: 13 BD: 87 +.+.: &xa->xa_lock#3 ->pool_lock#2 ->&c->lock ->&n->list_lock ->&obj_hash[i].lock ->&____s->seqcount#2 ->&____s->seqcount ->batched_entropy_u8.lock ->kfence_freelist_lock FD: 25 BD: 81 +.+.: subsys mutex#17 ->&k->k_lock ->&rq->__lock FD: 7 BD: 4035 ..-.: &dir->lock#2 ->&obj_hash[i].lock ->pool_lock#2 ->quarantine_lock ->&meta->lock ->kfence_freelist_lock FD: 34 BD: 86 +.+.: dev_hotplug_mutex ->&dev->power.lock ->&rq->__lock ->&k->k_lock FD: 14 BD: 86 ++++: dev_base_lock ->&xa->xa_lock#3 FD: 1 BD: 78 ++++: qdisc_mod_lock FD: 22 BD: 2 ++++: bt_proto_lock ->pool_lock#2 ->&dir->lock ->&obj_hash[i].lock ->&c->lock ->&zone->lock ->&____s->seqcount ->chan_list_lock ->l2cap_sk_list.lock ->&sk->sk_peer_lock ->hci_sk_list.lock ->&n->list_lock ->&____s->seqcount#2 ->sco_sk_list.lock ->rfcomm_sk_list.lock ->cmtp_sk_list.lock FD: 157 BD: 23 +.+.: hci_cb_list_lock ->fs_reclaim ->pool_lock#2 ->&obj_hash[i].lock ->chan_list_lock ->&conn->ident_lock ->&base->lock ->&list->lock#9 ->&conn->chan_lock ->&c->lock ->&rq->__lock ->&list->lock#10 ->(work_completion)(&(&conn->id_addr_timer)->work) ->rcu_node_0 ->hci_cb_list_lock.wait_lock ->(work_completion)(&(&conn->info_timer)->work) ->&rnp->exp_lock ->rcu_state.exp_mutex ->rcu_state.exp_mutex.wait_lock ->&p->pi_lock ->&rnp->exp_wq[1] ->&rnp->exp_wq[2] ->&conn->lock#2 ->(work_completion)(&(&conn->timeout_work)->work) FD: 310 BD: 4 +.+.: mgmt_chan_list_lock ->pool_lock#2 ->tk_core.seq.seqcount ->hci_sk_list.lock ->&obj_hash[i].lock ->hci_dev_list_lock ->(console_sem).lock ->&rq->__lock ->&hdev->lock ->&c->lock ->&data->lock ->fs_reclaim ->rlock-AF_BLUETOOTH ->&____s->seqcount#2 ->&____s->seqcount ->&lock->wait_lock ->&p->pi_lock FD: 1 BD: 3903 ....: &list->lock#2 FD: 126 BD: 80 +.+.: rate_ctrl_mutex ->fs_reclaim ->pool_lock#2 FD: 2 BD: 6 +.+.: netlbl_domhsh_lock ->pool_lock#2 FD: 1 BD: 6 +.+.: netlbl_unlhsh_lock FD: 186 BD: 1 +.+.: misc_mtx ->fs_reclaim ->pool_lock#2 ->&x->wait#9 ->&obj_hash[i].lock ->&k->list_lock ->gdp_mutex ->lock ->&root->kernfs_rwsem ->bus_type_sem ->&c->lock ->&____s->seqcount ->sysfs_symlink_target_lock ->&dev->power.lock ->dpm_list_mtx ->req_lock ->&p->pi_lock ->&x->wait#11 ->&rq->__lock ->uevent_sock_mutex ->running_helpers_waitq.lock ->subsys mutex#18 ->misc_minors_ida.xa_lock ->&zone->lock ->&cfs_rq->removed.lock ->batched_entropy_u8.lock ->kfence_freelist_lock ->&base->lock ->&dir->lock ->rfkill_global_mutex ->remove_cache_srcu ->&n->list_lock ->&____s->seqcount#2 ->rcu_node_0 FD: 24 BD: 175 ....: &x->wait#11 ->&p->pi_lock FD: 150 BD: 2 .+.+: sb_writers ->mount_lock ->&type->i_mutex_dir_key/1 ->&sb->s_type->i_mutex_key#4 ->tk_core.seq.seqcount ->&sb->s_type->i_lock_key#5 ->&wb->list_lock ->&type->i_mutex_dir_key#2 ->&sem->wait_lock ->&p->pi_lock ->&rq->__lock ->&dentry->d_lock ->tomoyo_ss ->&s->s_inode_list_lock ->&sbinfo->stat_lock ->&xa->xa_lock#9 ->&obj_hash[i].lock ->&fsnotify_mark_srcu FD: 139 BD: 3 +.+.: &type->i_mutex_dir_key/1 ->rename_lock.seqcount ->fs_reclaim ->pool_lock#2 ->&dentry->d_lock ->&obj_hash[i].lock ->&sbinfo->stat_lock ->&zone->lock ->&____s->seqcount ->&c->lock ->mmu_notifier_invalidate_range_start ->&sb->s_type->i_lock_key#5 ->&s->s_inode_list_lock ->tk_core.seq.seqcount ->batched_entropy_u32.lock ->&simple_offset_xa_lock ->&sb->s_type->i_mutex_key#4 ->quarantine_lock ->&meta->lock ->kfence_freelist_lock ->&rq->__lock ->tomoyo_ss ->&u->bindlock ->&fsnotify_mark_srcu ->&n->list_lock ->&sem->wait_lock ->krc.lock ->&xa->xa_lock#9 ->&sb->s_type->i_mutex_key#4/4 ->&cfs_rq->removed.lock ->remove_cache_srcu FD: 13 BD: 12 +.+.: &simple_offset_xa_lock ->pool_lock#2 ->&c->lock ->&zone->lock ->&____s->seqcount ->&obj_hash[i].lock ->&n->list_lock ->&____s->seqcount#2 ->batched_entropy_u8.lock ->kfence_freelist_lock FD: 115 BD: 4 +.+.: &sb->s_type->i_mutex_key#4 ->tk_core.seq.seqcount ->&rq->__lock ->&cfs_rq->removed.lock ->&obj_hash[i].lock ->tomoyo_ss ->&xattrs->lock ->&dentry->d_lock ->&fsnotify_mark_srcu ->&sb->s_type->i_mutex_key#4/4 ->&sb->s_type->i_lock_key#5 ->fs_reclaim ->&____s->seqcount ->pool_lock#2 ->stock_lock ->&xa->xa_lock#9 ->lock#4 ->&info->lock ->key#9 ->rcu_node_0 ->&sem->wait_lock ->&wb->list_lock ->lock#5 ->&lruvec->lru_lock ->&rcu_state.expedited_wq ->&simple_offset_xa_lock FD: 3 BD: 2 +.+.: subsys mutex#18 ->&k->k_lock FD: 190 BD: 6 +.+.: input_mutex ->input_devices_poll_wait.lock ->fs_reclaim ->pool_lock#2 ->&dev->mutex#2 ->input_ida.xa_lock ->&c->lock ->&zone->lock ->&____s->seqcount ->&x->wait#9 ->&obj_hash[i].lock ->chrdevs_lock ->&k->list_lock ->lock ->&root->kernfs_rwsem ->bus_type_sem ->sysfs_symlink_target_lock ->&dev->power.lock ->dpm_list_mtx ->req_lock ->&p->pi_lock ->&rq->__lock ->&x->wait#11 ->uevent_sock_mutex ->running_helpers_waitq.lock ->&k->k_lock ->subsys mutex#30 ->&led_cdev->led_access ->&cfs_rq->removed.lock ->&pool->lock ->&mousedev->mutex/1 FD: 1 BD: 7 ....: input_devices_poll_wait.lock FD: 318 BD: 3 ++++: (netlink_chain).rwsem ->pool_lock#2 ->&obj_hash[i].lock ->reg_indoor_lock ->hwsim_radio_lock ->&rq->__lock ->&c->lock ->&____s->seqcount ->&____s->seqcount#2 ->&q->instances_lock ->&log->instances_lock ->&nft_net->commit_mutex ->rcu_node_0 ->batched_entropy_u8.lock ->kfence_freelist_lock ->&n->list_lock ->&cfs_rq->removed.lock ->pgd_lock ->stock_lock ->key ->pcpu_lock ->percpu_counters_lock FD: 12 BD: 1 ++++: proto_tab_lock ->pool_lock#2 ->&dir->lock ->&obj_hash[i].lock ->raw_sk_list.lock ->&c->lock ->&n->list_lock FD: 3 BD: 1 ....: random_ready_notifier.lock ->crngs.lock FD: 1 BD: 2 ....: misc_minors_ida.xa_lock FD: 178 BD: 2 +.+.: (work_completion)(&rfkill_global_led_trigger_work) ->rfkill_global_mutex FD: 177 BD: 10 +.+.: rfkill_global_mutex ->fs_reclaim ->pool_lock#2 ->&k->list_lock ->lock ->&root->kernfs_rwsem ->bus_type_sem ->sysfs_symlink_target_lock ->&c->lock ->&____s->seqcount ->&dev->power.lock ->dpm_list_mtx ->&rfkill->lock ->uevent_sock_mutex ->&n->list_lock ->&obj_hash[i].lock ->running_helpers_waitq.lock ->&k->k_lock ->subsys mutex#41 ->triggers_list_lock ->leds_list_lock ->&pool->lock ->&rq->__lock ->rfkill_global_mutex.wait_lock ->&zone->lock ->&cfs_rq->removed.lock ->uevent_sock_mutex.wait_lock ->&p->pi_lock ->&data->mtx ->&____s->seqcount#2 ->&sem->wait_lock FD: 39 BD: 1 ....: vga_lock#2 ->pci_config_lock ->(console_sem).lock FD: 178 BD: 1 +.+.: disable_lock ->fs_reclaim ->pool_lock#2 ->&x->wait#9 ->&obj_hash[i].lock ->&k->list_lock ->lock ->&root->kernfs_rwsem ->bus_type_sem ->sysfs_symlink_target_lock ->&k->k_lock ->&c->lock ->&____s->seqcount ->&dev->power.lock ->dpm_list_mtx ->&(&priv->bus_notifier)->rwsem ->uevent_sock_mutex ->running_helpers_waitq.lock ->subsys mutex#3 FD: 137 BD: 1 +.+.: &type->s_umount_key#12/1 ->fs_reclaim ->pool_lock#2 ->pcpu_alloc_mutex ->shrinker_rwsem ->list_lrus_mutex ->sb_lock ->mmu_notifier_invalidate_range_start ->&sb->s_type->i_lock_key#11 ->&s->s_inode_list_lock ->tk_core.seq.seqcount ->&dentry->d_lock FD: 40 BD: 2 +.+.: &sb->s_type->i_lock_key#11 ->&dentry->d_lock FD: 288 BD: 2 +.+.: timer_update_work ->timer_keys_mutex FD: 287 BD: 3 +.+.: timer_keys_mutex ->cpu_hotplug_lock FD: 320 BD: 1 +.+.: (work_completion)(&tracerfs_init_work) ->pin_fs_lock ->fs_reclaim ->pool_lock#2 ->&zone->lock ->&____s->seqcount ->&c->lock ->sb_lock ->&type->s_umount_key#13/1 ->&type->s_umount_key#14 ->mnt_id_ida.xa_lock ->pcpu_alloc_mutex ->&dentry->d_lock ->mount_lock ->&obj_hash[i].lock ->&fsnotify_mark_srcu ->&sb->s_type->i_mutex_key#5 ->event_mutex ->(module_notify_list).rwsem ->trace_types_lock FD: 139 BD: 2 +.+.: &type->s_umount_key#13/1 ->fs_reclaim ->pool_lock#2 ->pcpu_alloc_mutex ->shrinker_rwsem ->list_lrus_mutex ->sb_lock ->mmu_notifier_invalidate_range_start ->&sb->s_type->i_lock_key#12 ->&s->s_inode_list_lock ->tk_core.seq.seqcount ->&dentry->d_lock ->&type->s_umount_key#14 FD: 40 BD: 8 +.+.: &sb->s_type->i_lock_key#12 ->&dentry->d_lock FD: 132 BD: 3 +.+.: &type->s_umount_key#14 ->sb_lock ->list_lrus_mutex ->&xa->xa_lock#4 ->&obj_hash[i].lock ->pool_lock#2 ->shrinker_rwsem ->&rsp->gp_wait ->pcpu_lock ->fs_reclaim ->&dentry->d_lock ->&lru->node[i].lock FD: 128 BD: 6 +.+.: &sb->s_type->i_mutex_key#5 ->&sb->s_type->i_lock_key#12 ->rename_lock.seqcount ->fs_reclaim ->pool_lock#2 ->&dentry->d_lock ->mmu_notifier_invalidate_range_start ->&s->s_inode_list_lock ->tk_core.seq.seqcount ->&c->lock ->&zone->lock ->&____s->seqcount ->&obj_hash[i].lock ->&rq->__lock ->batched_entropy_u8.lock ->kfence_freelist_lock FD: 13 BD: 4562 ....: &xa->xa_lock#4 ->pool_lock#2 ->&obj_hash[i].lock ->&____s->seqcount ->&c->lock ->&____s->seqcount#2 ->&n->list_lock FD: 24 BD: 169 ..-.: &rsp->gp_wait ->&obj_hash[i].lock ->pool_lock#2 ->&p->pi_lock FD: 131 BD: 4525 .+.+: &fsnotify_mark_srcu ->&conn->lock ->fs_reclaim ->pool_lock#2 ->&group->notification_lock ->&group->notification_waitq ->&obj_hash[i].lock ->&rq->__lock ->remove_cache_srcu ->&c->lock ->&____s->seqcount#2 ->&____s->seqcount ->&cfs_rq->removed.lock ->&n->list_lock ->pgd_lock ->stock_lock ->key ->pcpu_lock ->percpu_counters_lock ->batched_entropy_u8.lock ->kfence_freelist_lock ->rcu_node_0 FD: 137 BD: 1 +.+.: &type->s_umount_key#15/1 ->fs_reclaim ->pool_lock#2 ->pcpu_alloc_mutex ->shrinker_rwsem ->list_lrus_mutex ->sb_lock ->mmu_notifier_invalidate_range_start ->&sb->s_type->i_lock_key#13 ->&s->s_inode_list_lock ->tk_core.seq.seqcount ->&dentry->d_lock FD: 306 BD: 2 +.+.: event_mutex ->pin_fs_lock ->&sb->s_type->i_mutex_key#5 ->trace_event_sem ->trace_event_sem.wait_lock ->&p->pi_lock ->trace_types_lock ->sched_register_mutex ->tracepoints_mutex ->&rq->__lock ->event_mutex.wait_lock FD: 40 BD: 2 +.+.: &sb->s_type->i_lock_key#13 ->&dentry->d_lock FD: 137 BD: 1 +.+.: &type->s_umount_key#16/1 ->fs_reclaim ->pool_lock#2 ->pcpu_alloc_mutex ->shrinker_rwsem ->list_lrus_mutex ->sb_lock ->mmu_notifier_invalidate_range_start ->&sb->s_type->i_lock_key#14 ->&s->s_inode_list_lock ->tk_core.seq.seqcount ->&dentry->d_lock FD: 1 BD: 5 ....: trace_event_sem.wait_lock FD: 40 BD: 2 +.+.: &sb->s_type->i_lock_key#14 ->&dentry->d_lock FD: 137 BD: 1 +.+.: &type->s_umount_key#17/1 ->fs_reclaim ->pool_lock#2 ->pcpu_alloc_mutex ->shrinker_rwsem ->list_lrus_mutex ->sb_lock ->mmu_notifier_invalidate_range_start ->&sb->s_type->i_lock_key#15 ->&s->s_inode_list_lock ->tk_core.seq.seqcount ->&dentry->d_lock FD: 40 BD: 80 +.+.: &sb->s_type->i_lock_key#15 ->&dentry->d_lock FD: 127 BD: 1 +.+.: kclist_lock ->resource_lock ->fs_reclaim ->pool_lock#2 FD: 137 BD: 1 +.+.: &type->s_umount_key#18/1 ->fs_reclaim ->pool_lock#2 ->pcpu_alloc_mutex ->shrinker_rwsem ->&c->lock ->&____s->seqcount ->list_lrus_mutex ->sb_lock ->mmu_notifier_invalidate_range_start ->&sb->s_type->i_lock_key#16 ->&s->s_inode_list_lock ->tk_core.seq.seqcount ->&dentry->d_lock FD: 55 BD: 156 +.+.: &sb->s_type->i_lock_key#16 ->&dentry->d_lock ->&xa->xa_lock#9 FD: 244 BD: 34 .+.+: tomoyo_ss ->mmu_notifier_invalidate_range_start ->pool_lock#2 ->tomoyo_policy_lock ->(console_sem).lock ->console_owner_lock ->console_owner ->&c->lock ->&zone->lock ->&____s->seqcount ->&obj_hash[i].lock ->&dentry->d_lock ->tomoyo_log_lock ->tomoyo_log_wait.lock ->quarantine_lock ->file_systems_lock ->fs_reclaim ->&mm->mmap_lock ->&rq->__lock ->batched_entropy_u8.lock ->kfence_freelist_lock ->&meta->lock ->&n->list_lock ->remove_cache_srcu ->rcu_node_0 ->&rcu_state.expedited_wq ->&base->lock ->pgd_lock ->key ->pcpu_lock ->percpu_counters_lock ->&cfs_rq->removed.lock ->&____s->seqcount#2 ->mount_lock ->stock_lock FD: 137 BD: 1 +.+.: &type->s_umount_key#19/1 ->fs_reclaim ->pool_lock#2 ->pcpu_alloc_mutex ->shrinker_rwsem ->list_lrus_mutex ->sb_lock ->mmu_notifier_invalidate_range_start ->&sb->s_type->i_lock_key#17 ->&s->s_inode_list_lock ->tk_core.seq.seqcount ->&dentry->d_lock FD: 40 BD: 4 +.+.: &sb->s_type->i_lock_key#17 ->&dentry->d_lock FD: 130 BD: 1 +.+.: &ns->lock ->&dentry->d_lock ->pin_fs_lock ->&sb->s_type->i_mutex_key#6 FD: 128 BD: 2 +.+.: &sb->s_type->i_mutex_key#6 ->&sb->s_type->i_lock_key#17 ->rename_lock.seqcount ->fs_reclaim ->&____s->seqcount ->&zone->lock ->pool_lock#2 ->&obj_hash[i].lock ->&dentry->d_lock ->mmu_notifier_invalidate_range_start ->&s->s_inode_list_lock ->tk_core.seq.seqcount ->&c->lock FD: 42 BD: 1 +.+.: &type->s_umount_key#20 ->sb_lock ->&dentry->d_lock FD: 126 BD: 1 +.+.: pnp_lock ->fs_reclaim ->pool_lock#2 FD: 1 BD: 1 +.+.: subsys mutex#19 FD: 3 BD: 1 +.+.: subsys mutex#20 ->&k->k_lock FD: 3 BD: 10 +.+.: subsys mutex#21 ->&k->k_lock FD: 3 BD: 1 +.+.: subsys mutex#22 ->&k->k_lock FD: 371 BD: 1 +.+.: tty_mutex ->(console_sem).lock ->console_lock ->fs_reclaim ->pool_lock#2 ->tty_ldiscs_lock ->&obj_hash[i].lock ->&k->list_lock ->&k->k_lock ->&tty->legacy_mutex FD: 4 BD: 1 +.+.: subsys mutex#23 ->&k->list_lock ->&k->k_lock FD: 1 BD: 1 ....: netevent_notif_chain.lock FD: 323 BD: 12 ++++: clients_rwsem ->fs_reclaim ->clients.xa_lock ->&device->client_data_rwsem FD: 2 BD: 13 +.+.: clients.xa_lock ->pool_lock#2 FD: 923 BD: 11 ++++: devices_rwsem ->rcu_node_0 ->&rq->__lock ->fs_reclaim ->pool_lock#2 ->devices.xa_lock ->&obj_hash[i].lock ->(console_sem).lock ->clients_rwsem ->rdma_nets_rwsem ->&pdata->netdev_lock ->&table->lock#4 ->rdma_nets_rwsem.wait_lock ->&p->pi_lock ->&rcu_state.expedited_wq ->&cfs_rq->removed.lock FD: 1 BD: 1 +.+.: (blocking_lsm_notifier_chain).rwsem FD: 229 BD: 78 ++++: (inetaddr_chain).rwsem ->fs_reclaim ->pool_lock#2 ->pcpu_alloc_mutex ->fib_info_lock ->&c->lock ->&dir->lock#2 ->&____s->seqcount ->nl_table_lock ->&obj_hash[i].lock ->nl_table_wait.lock ->&net->sctp.local_addr_lock ->rlock-AF_NETLINK ->&rq->__lock ->&n->list_lock ->remove_cache_srcu ->quarantine_lock ->&ipvlan->addrs_lock ->&____s->seqcount#2 ->&tbl->lock ->class ->(&tbl->proxy_timer) ->&base->lock ->krc.lock ->stock_lock ->&dir->lock FD: 1 BD: 7 ....: inet6addr_chain.lock FD: 1 BD: 1 +.+.: buses_mutex FD: 1 BD: 1 +.+.: offload_lock FD: 1 BD: 1 +...: inetsw_lock FD: 913 BD: 1 +.+.: (wq_completion)events_power_efficient ->(work_completion)(&(&tbl->managed_work)->work) ->(check_lifetime_work).work ->(work_completion)(&(&cache_cleaner)->work) ->(work_completion)(&(&ops->cursor_work)->work) ->(work_completion)(&(&hub->init_work)->work) ->(work_completion)(&(&gc_work->dwork)->work) ->(work_completion)(&(&tbl->gc_work)->work) ->(reg_check_chans).work ->(crda_timeout).work ->(gc_work).work ->(work_completion)(&(&hinfo->gc_work)->work) ->(work_completion)(&barr->work) ->&rq->__lock FD: 49 BD: 2 +.+.: (work_completion)(&(&tbl->managed_work)->work) ->&tbl->lock ->&rq->__lock FD: 48 BD: 3913 ++-.: &tbl->lock ->&obj_hash[i].lock ->&base->lock ->pool_lock#2 ->batched_entropy_u32.lock ->&n->lock ->&c->lock ->&n->list_lock ->nl_table_lock ->nl_table_wait.lock ->&dir->lock#2 ->krc.lock ->&____s->seqcount ->rlock-AF_NETLINK ->&____s->seqcount#2 ->init_task.mems_allowed_seq.seqcount ->batched_entropy_u8.lock ->kfence_freelist_lock ->&meta->lock ->quarantine_lock ->&zone->lock FD: 1 BD: 82 +.+.: ptype_lock FD: 24 BD: 2 +.+.: (check_lifetime_work).work ->&obj_hash[i].lock ->&base->lock ->&rq->__lock ->rcu_node_0 FD: 1 BD: 78 +.+.: &net->rules_mod_lock FD: 1 BD: 1 +.+.: tcp_ulp_list_lock FD: 1 BD: 1 +...: xfrm_state_afinfo_lock FD: 1 BD: 1 +.+.: xfrm_policy_afinfo_lock FD: 1 BD: 1 +...: xfrm_input_afinfo_lock FD: 17 BD: 4696 ..-.: krc.lock ->&obj_hash[i].lock ->hrtimer_bases.lock ->&base->lock FD: 130 BD: 1 +.+.: (wq_completion)events_highpri ->(work_completion)(&(&krcp->page_cache_work)->work) ->(work_completion)(flush) ->(work_completion)(&barr->work) FD: 126 BD: 3 +.+.: (work_completion)(&(&krcp->page_cache_work)->work) ->fs_reclaim ->&zone->lock ->&____s->seqcount ->pool_lock#2 ->krc.lock ->&rq->__lock FD: 1 BD: 3 +.+.: &hashinfo->lock FD: 1 BD: 1 +.+.: tcp_cong_list_lock FD: 1 BD: 1 +.+.: mptcp_sched_list_lock FD: 2 BD: 7 +.+.: cache_list_lock ->&cd->hash_lock FD: 25 BD: 2 +.+.: (work_completion)(&(&cache_cleaner)->work) ->cache_list_lock ->&obj_hash[i].lock ->&base->lock ->&rq->__lock FD: 1 BD: 1 +.+.: (rpc_pipefs_notifier_list).rwsem FD: 1 BD: 1 +.+.: svc_xprt_class_lock FD: 38 BD: 1 +.+.: xprt_list_lock ->(console_sem).lock FD: 1 BD: 1 ....: pcibios_fwaddrmap_lock FD: 142 BD: 3 .+.+: sb_writers#2 ->mount_lock ->&sb->s_type->i_mutex_key/1 ->&sb->s_type->i_mutex_key FD: 138 BD: 4 +.+.: &sb->s_type->i_mutex_key/1 ->rename_lock.seqcount ->fs_reclaim ->pool_lock#2 ->&dentry->d_lock ->&obj_hash[i].lock ->tomoyo_ss ->&c->lock ->&zone->lock ->&____s->seqcount ->mmu_notifier_invalidate_range_start ->&sb->s_type->i_lock_key#2 ->&s->s_inode_list_lock ->tk_core.seq.seqcount ->&sb->s_type->i_mutex_key FD: 1 BD: 2 +.+.: tomoyo_log_lock FD: 1 BD: 2 ....: tomoyo_log_wait.lock FD: 66 BD: 4539 +.+.: &wb->list_lock ->&sb->s_type->i_lock_key#2 ->&sb->s_type->i_lock_key#23 ->&sb->s_type->i_lock_key#22 ->&sb->s_type->i_lock_key ->&sb->s_type->i_lock_key#5 ->&sb->s_type->i_lock_key#8 ->&sb->s_type->i_lock_key#24 ->&sb->s_type->i_lock_key#3 ->&sb->s_type->i_lock_key#27 ->&sb->s_type->i_lock_key#31 FD: 183 BD: 4 ++++: umhelper_sem ->usermodehelper_disabled_waitq.lock ->fs_reclaim ->&c->lock ->&zone->lock ->&____s->seqcount ->pool_lock#2 ->&x->wait#9 ->&obj_hash[i].lock ->&k->list_lock ->gdp_mutex ->lock ->&root->kernfs_rwsem ->bus_type_sem ->sysfs_symlink_target_lock ->&dev->power.lock ->dpm_list_mtx ->&k->k_lock ->subsys mutex#80 ->fw_lock ->uevent_sock_mutex ->running_helpers_waitq.lock ->&rq->__lock ->&cfs_rq->removed.lock ->&x->wait#23 ->&base->lock ->(&timer.timer) ->dev_pm_qos_sysfs_mtx ->kernfs_idr_lock ->deferred_probe_mutex ->device_links_lock ->mmu_notifier_invalidate_range_start ->&n->list_lock ->&____s->seqcount#2 ->fw_lock.wait_lock ->&p->pi_lock ->batched_entropy_u8.lock ->kfence_freelist_lock FD: 1 BD: 5 ....: usermodehelper_disabled_waitq.lock FD: 1 BD: 1 +.+.: &drv->dynids.lock FD: 196 BD: 2 +.+.: (work_completion)(&sub_info->work) ->&sighand->siglock ->fs_reclaim ->pool_lock#2 ->free_vmap_area_lock ->vmap_area_lock ->&zone->lock ->&____s->seqcount ->init_mm.page_table_lock ->batched_entropy_u64.lock ->&obj_hash[i].lock ->&c->lock ->init_files.file_lock ->init_fs.lock ->&p->alloc_lock ->lock ->pidmap_lock ->cgroup_threadgroup_rwsem ->input_pool.lock ->&p->pi_lock ->rcu_node_0 ->&rq->__lock ->&cfs_rq->removed.lock ->batched_entropy_u8.lock ->kfence_freelist_lock ->&n->list_lock ->&sig->wait_chldexit ->tasklist_lock ->&prev->lock ->css_set_lock ->&x->wait#17 ->&____s->seqcount#2 ->remove_cache_srcu ->quarantine_lock ->&meta->lock FD: 1 BD: 1 +.+.: umh_sysctl_lock FD: 65 BD: 4496 ++++: &anon_vma->rwsem ->&mm->page_table_lock ->&obj_hash[i].lock ->pool_lock#2 ->quarantine_lock ->&rq->__lock ->&meta->lock ->kfence_freelist_lock ->&c->lock ->&____s->seqcount ->&sem->wait_lock ->&n->list_lock ->rcu_node_0 ->pgd_lock ->key ->pcpu_lock ->percpu_counters_lock ->&cfs_rq->removed.lock ->mmu_notifier_invalidate_range_start ->ptlock_ptr(page) ->ptlock_ptr(page)#2 ->stock_lock ->&rcu_state.gp_wq ->&rcu_state.expedited_wq ->&____s->seqcount#2 ->&base->lock ->batched_entropy_u8.lock ->per_cpu_ptr(&cgroup_rstat_cpu_lock, cpu) FD: 1 BD: 5139 -.-.: per_cpu_ptr(&cgroup_rstat_cpu_lock, cpu) FD: 10 BD: 4540 +.+.: lock#4 ->&lruvec->lru_lock ->&obj_hash[i].lock ->&pcp->lock ->lock#11 FD: 272 BD: 1 +.+.: &sig->cred_guard_mutex ->fs_reclaim ->pool_lock#2 ->&fs->lock ->&zone->lock ->&____s->seqcount ->&c->lock ->&dentry->d_lock ->&sb->s_type->i_mutex_key ->&obj_hash[i].lock ->delayed_uprobe_lock ->&mm->mmap_lock ->pgd_lock ->key ->pcpu_lock ->percpu_counters_lock ->&rq->__lock ->pool_lock ->rcu_node_0 ->&n->list_lock ->&cfs_rq->removed.lock ->quarantine_lock ->&dentry->d_lock/1 ->init_fs.lock ->&type->i_mutex_dir_key#3 ->&sb->s_type->i_lock_key#22 ->&sb->s_type->i_mutex_key#8 ->&p->pi_lock ->aa_buffers_lock ->mapping.invalidate_lock ->&folio_wait_table[i] ->tomoyo_ss ->&iint->mutex ->binfmt_lock ->entries_lock ->&ei->xattr_sem ->&tsk->futex_exit_mutex ->&sig->exec_update_lock ->&p->alloc_lock ->tk_core.seq.seqcount ->&stopper->lock ->&stop_pi_lock ->&x->wait#8 ->&lock->wait_lock ->batched_entropy_u8.lock ->kfence_freelist_lock ->key#8 ->remove_cache_srcu ->&____s->seqcount#2 FD: 2 BD: 4544 ..-.: &lruvec->lru_lock ->per_cpu_ptr(&cgroup_rstat_cpu_lock, cpu) FD: 6 BD: 4535 +.+.: lock#5 ->&lruvec->lru_lock ->&obj_hash[i].lock FD: 126 BD: 155 ++++: &vma->vm_lock->lock ->&rq->__lock ->&cfs_rq->removed.lock ->&obj_hash[i].lock ->fs_reclaim ->&____s->seqcount ->pool_lock#2 ->ptlock_ptr(page)#2 ->mmu_notifier_invalidate_range_start ->&lruvec->lru_lock ->rcu_node_0 ->per_cpu_ptr(&cgroup_rstat_cpu_lock, cpu) ->ptlock_ptr(page) ->pgd_lock ->key ->pcpu_lock ->percpu_counters_lock ->&rcu_state.gp_wq ->&c->lock ->batched_entropy_u8.lock ->kfence_freelist_lock ->&p->pi_lock ->remove_cache_srcu ->&rcu_state.expedited_wq ->stock_lock ->&____s->seqcount#2 ->&n->list_lock ->lock#4 ->lock#5 ->&sem->wait_lock FD: 230 BD: 2 +.+.: &tsk->futex_exit_mutex ->&p->pi_lock ->&rq->__lock ->&mm->mmap_lock ->&cfs_rq->removed.lock ->&obj_hash[i].lock ->pool_lock#2 FD: 23 BD: 1 +.+.: &child->perf_event_mutex ->&rq->__lock ->&cfs_rq->removed.lock ->&obj_hash[i].lock ->pool_lock#2 FD: 1 BD: 147 ....: &pid->wait_pidfd FD: 24 BD: 156 ....: &sig->wait_chldexit ->&p->pi_lock FD: 15 BD: 156 ....: &(&sig->stats_lock)->lock ->&____s->seqcount#5 FD: 14 BD: 157 ....: &____s->seqcount#5 ->pidmap_lock ->&obj_hash[i].lock ->pool_lock#2 FD: 38 BD: 1 +.+.: low_water_lock ->(console_sem).lock ->console_owner_lock ->console_owner FD: 291 BD: 1 +.+.: vendor_module_lock ->slab_mutex ->pcpu_alloc_mutex ->&obj_hash[i].lock ->percpu_counters_lock ->fs_reclaim ->pool_lock#2 ->shrinker_rwsem ->&zone->lock ->&____s->seqcount ->cpu_hotplug_lock ->timekeeper_lock FD: 13 BD: 1 +.-.: (&tcp_orphan_timer) ->&obj_hash[i].lock ->&base->lock FD: 25 BD: 1 ..-.: &(&cache_cleaner)->timer FD: 1 BD: 155 +.+.: text_mutex.wait_lock FD: 1 BD: 4803 -.-.: pvclock_gtod_data FD: 144 BD: 3 ++++: &type->i_mutex_dir_key#2 ->fs_reclaim ->pool_lock#2 ->&dentry->d_lock ->rename_lock.seqcount ->&c->lock ->&zone->lock ->&____s->seqcount ->namespace_sem ->tk_core.seq.seqcount ->&sb->s_type->i_lock_key#5 ->&n->list_lock ->&sem->wait_lock ->&rq->__lock ->pgd_lock ->&obj_hash[i].lock ->key ->pcpu_lock ->percpu_counters_lock ->&cfs_rq->removed.lock ->&xa->xa_lock#4 ->stock_lock ->tomoyo_ss ->&sbinfo->stat_lock ->mmu_notifier_invalidate_range_start ->&s->s_inode_list_lock ->batched_entropy_u32.lock ->&simple_offset_xa_lock FD: 3 BD: 144 +.+.: subsys mutex#24 ->&k->k_lock FD: 3 BD: 144 +.+.: subsys mutex#25 ->&k->k_lock FD: 37 BD: 1 +.+.: put_task_map-wait-type-override ->&obj_hash[i].lock ->pool_lock#2 ->quarantine_lock ->&base->lock ->stock_lock ->&meta->lock ->kfence_freelist_lock ->per_cpu_ptr(&cgroup_rstat_cpu_lock, cpu) ->css_set_lock FD: 1 BD: 1 +.+.: subsys mutex#26 FD: 188 BD: 1 +.+.: subsys mutex#27 ->&k->list_lock ->&k->k_lock ->fs_reclaim ->pool_lock#2 ->&x->wait#9 ->&obj_hash[i].lock ->platform_devid_ida.xa_lock ->lock ->&root->kernfs_rwsem ->bus_type_sem ->sysfs_symlink_target_lock ->&dev->power.lock ->dpm_list_mtx ->&(&priv->bus_notifier)->rwsem ->uevent_sock_mutex ->running_helpers_waitq.lock ->&c->lock ->&____s->seqcount ->subsys mutex#3 ->&rq->__lock ->wakeup_ida.xa_lock ->gdp_mutex ->&zone->lock ->subsys mutex#15 ->events_lock ->rtcdev_lock FD: 1 BD: 1 +.+.: subsys mutex#28 FD: 33 BD: 2 +.+.: (work_completion)(&p->wq) ->vmap_area_lock ->&obj_hash[i].lock ->purge_vmap_area_lock ->pool_lock#2 ->rcu_node_0 ->&rq->__lock ->&meta->lock ->kfence_freelist_lock ->&cfs_rq->removed.lock ->&base->lock ->&rcu_state.expedited_wq ->quarantine_lock FD: 25 BD: 1 ..-.: &(&group->avgs_work)->timer FD: 25 BD: 1 ..-.: &(({ do { const void *__vpp_verify = (typeof((&vmstat_work) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __asm__ ("" : "=r"(__ptr) : "0"((typeof(*((&vmstat_work))) *)((&vmstat_work)))); (typeof((typeof(*((&vmstat_work))) *)((&vmstat_work)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); }))->timer FD: 30 BD: 1 ..-.: mm/memcontrol.c:589 FD: 24 BD: 2 +.+.: (work_completion)(&(&group->avgs_work)->work) ->&group->avgs_lock ->&rq->__lock FD: 23 BD: 3 +.+.: &group->avgs_lock ->&per_cpu_ptr(group->pcpu, cpu)->seq ->&obj_hash[i].lock ->&base->lock ->&rq->__lock FD: 27 BD: 2 +.+.: (stats_flush_dwork).work ->cgroup_rstat_lock ->&obj_hash[i].lock ->&base->lock ->&rq->__lock ->rcu_node_0 ->&rcu_state.expedited_wq ->&cfs_rq->removed.lock ->pool_lock#2 FD: 2 BD: 19 ....: cgroup_rstat_lock ->per_cpu_ptr(&cgroup_rstat_cpu_lock, cpu) FD: 1 BD: 145 +.+.: subsys mutex#29 FD: 1 BD: 4 +.+.: key_user_lock FD: 1 BD: 4 +.+.: key_serial_lock FD: 5 BD: 5 +.+.: key_construction_mutex ->&obj_hash[i].lock ->pool_lock#2 ->keyring_name_lock FD: 133 BD: 3 +.+.: &type->lock_class ->keyring_serialise_link_lock ->fs_reclaim ->pool_lock#2 ->key_user_lock ->crngs.lock ->key_serial_lock ->key_construction_mutex ->ima_keys_lock ->&c->lock ->&zone->lock ->&____s->seqcount FD: 129 BD: 4 +.+.: keyring_serialise_link_lock ->fs_reclaim ->pool_lock#2 ->&obj_hash[i].lock ->root_key_user.lock ->key_construction_mutex FD: 24 BD: 4531 ....: &pgdat->kswapd_wait ->&p->pi_lock FD: 1 BD: 1 +.+.: drivers_lock FD: 143 BD: 1 +.+.: damon_dbgfs_lock ->fs_reclaim ->pool_lock#2 ->tk_core.seq.seqcount ->damon_ops_lock ->pin_fs_lock ->&sb->s_type->i_mutex_key#3 FD: 137 BD: 1 +.+.: &type->s_umount_key#21/1 ->fs_reclaim ->pool_lock#2 ->pcpu_alloc_mutex ->shrinker_rwsem ->list_lrus_mutex ->sb_lock ->&zone->lock ->&____s->seqcount ->&obj_hash[i].lock ->mmu_notifier_invalidate_range_start ->&sb->s_type->i_lock_key#18 ->&s->s_inode_list_lock ->tk_core.seq.seqcount ->&dentry->d_lock FD: 40 BD: 2 +.+.: &sb->s_type->i_lock_key#18 ->&dentry->d_lock FD: 1 BD: 1 +.+.: dq_list_lock FD: 137 BD: 1 +.+.: &type->s_umount_key#22/1 ->fs_reclaim ->pool_lock#2 ->pcpu_alloc_mutex ->shrinker_rwsem ->&c->lock ->&____s->seqcount ->list_lrus_mutex ->sb_lock ->mmu_notifier_invalidate_range_start ->&sb->s_type->i_lock_key#19 ->&s->s_inode_list_lock ->tk_core.seq.seqcount ->&dentry->d_lock FD: 40 BD: 7 +.+.: &sb->s_type->i_lock_key#19 ->&dentry->d_lock FD: 1 BD: 1 +.+.: configfs_subsystem_mutex FD: 135 BD: 1 +.+.: &sb->s_type->i_mutex_key#7/1 ->fs_reclaim ->pool_lock#2 ->&dentry->d_lock ->&____s->seqcount ->&c->lock ->configfs_dirent_lock ->mmu_notifier_invalidate_range_start ->&sb->s_type->i_lock_key#19 ->&s->s_inode_list_lock ->tk_core.seq.seqcount ->&default_group_class[depth - 1]/2 ->&default_group_class[depth - 1]#2 FD: 1 BD: 8 +.+.: configfs_dirent_lock FD: 133 BD: 2 +.+.: &default_group_class[depth - 1]/2 ->fs_reclaim ->pool_lock#2 ->&dentry->d_lock ->configfs_dirent_lock ->mmu_notifier_invalidate_range_start ->&sb->s_type->i_lock_key#19 ->&s->s_inode_list_lock ->tk_core.seq.seqcount ->&default_group_class[depth - 1]#3/2 FD: 126 BD: 1 +.+.: ecryptfs_daemon_hash_mux ->fs_reclaim ->&c->lock ->&zone->lock ->&____s->seqcount ->pool_lock#2 FD: 2 BD: 1 +.+.: ecryptfs_msg_ctx_lists_mux ->&ecryptfs_msg_ctx_arr[i].mux FD: 1 BD: 2 +.+.: &ecryptfs_msg_ctx_arr[i].mux FD: 1 BD: 1 ....: &ecryptfs_kthread_ctl.wait FD: 1 BD: 1 +.+.: nfs_version_lock FD: 150 BD: 1 ++++: key_types_sem ->(console_sem).lock ->&rq->__lock ->asymmetric_key_parsers_sem ->&type->lock_class ->&obj_hash[i].lock ->pool_lock#2 FD: 1 BD: 1 +.+.: pnfs_spinlock FD: 1 BD: 5 +.+.: &sn->pipefs_sb_lock FD: 1 BD: 1 +.+.: nls_lock FD: 1 BD: 1 +.+.: jffs2_compressor_list_lock FD: 1 BD: 1 +.+.: next_tag_value_lock FD: 1 BD: 1 ....: log_redrive_lock FD: 2 BD: 1 ....: &TxAnchor.LazyLock ->jfs_commit_thread_wait.lock FD: 1 BD: 2 ....: jfs_commit_thread_wait.lock FD: 1 BD: 1 +.+.: jfsTxnLock FD: 38 BD: 1 +.+.: ocfs2_stack_lock ->(console_sem).lock FD: 1 BD: 1 +.+.: o2hb_callback_sem FD: 1 BD: 1 +.+.: o2net_handler_lock FD: 137 BD: 1 +.+.: &type->s_umount_key#23/1 ->fs_reclaim ->pool_lock#2 ->pcpu_alloc_mutex ->shrinker_rwsem ->list_lrus_mutex ->sb_lock ->&____s->seqcount ->&c->lock ->mmu_notifier_invalidate_range_start ->&sb->s_type->i_lock_key#20 ->&s->s_inode_list_lock ->tk_core.seq.seqcount ->&dentry->d_lock ->&n->list_lock ->&xa->xa_lock#4 ->&obj_hash[i].lock ->stock_lock ->&rq->__lock ->&____s->seqcount#2 ->batched_entropy_u8.lock ->kfence_freelist_lock ->rcu_node_0 ->shrinker_rwsem.wait_lock ->&p->pi_lock ->&cfs_rq->removed.lock ->remove_cache_srcu FD: 40 BD: 5 +.+.: &sb->s_type->i_lock_key#20 ->&dentry->d_lock FD: 287 BD: 90 +.+.: nf_hook_mutex ->fs_reclaim ->&____s->seqcount ->pool_lock#2 ->&c->lock ->&zone->lock ->stock_lock ->&____s->seqcount#2 ->per_cpu_ptr(&cgroup_rstat_cpu_lock, cpu) ->&rq->__lock ->remove_cache_srcu ->cpu_hotplug_lock ->rcu_node_0 ->batched_entropy_u8.lock ->kfence_freelist_lock ->&n->list_lock FD: 126 BD: 1 ++++: alg_types_sem ->fs_reclaim ->pool_lock#2 FD: 1 BD: 1 +.+.: dma_list_mutex FD: 141 BD: 2 ++++: asymmetric_key_parsers_sem ->(console_sem).lock ->fs_reclaim ->pool_lock#2 ->&c->lock ->&____s->seqcount ->&zone->lock ->&obj_hash[i].lock ->crypto_alg_sem ->(crypto_chain).rwsem ->&x->wait#21 ->&base->lock ->&rq->__lock ->&cfs_rq->removed.lock ->(&timer.timer) FD: 900 BD: 1 +.+.: blkcg_pol_register_mutex ->blkcg_pol_mutex ->cgroup_mutex FD: 1 BD: 4 +.+.: elv_list_lock FD: 130 BD: 1 +.+.: crc_t10dif_mutex ->crypto_alg_sem ->fs_reclaim ->pool_lock#2 FD: 130 BD: 1 +.+.: crc64_rocksoft_mutex ->crypto_alg_sem ->fs_reclaim ->pool_lock#2 FD: 1 BD: 1 +.+.: ts_mod_lock FD: 3 BD: 7 +.+.: subsys mutex#30 ->&k->k_lock FD: 31 BD: 10 +.+.: &dev->mutex#2 ->&obj_hash[i].lock ->&rnp->exp_wq[3] ->&rq->__lock ->&rnp->exp_lock ->&lock->wait_lock FD: 26 BD: 7 ....: wakeup_srcu_srcu_usage.lock ->&obj_hash[i].lock ->&ACCESS_PRIVATE(sdp, lock) FD: 1 BD: 3 ....: wakeup_srcu FD: 1 BD: 3 ....: (&ws->timer) FD: 1 BD: 287 +.+.: klist_remove_lock FD: 5 BD: 4193 ....: &ws->lock ->tk_core.seq.seqcount ->&obj_hash[i].lock FD: 1 BD: 3 ....: deleted_ws.lock FD: 162 BD: 1 +.+.: register_count_mutex ->&k->list_lock ->fs_reclaim ->pool_lock#2 ->&c->lock ->&zone->lock ->&____s->seqcount ->lock ->&root->kernfs_rwsem ->&k->k_lock ->&obj_hash[i].lock ->uevent_sock_mutex ->running_helpers_waitq.lock FD: 1 BD: 1 +.+.: (cpufreq_policy_notifier_list).rwsem FD: 1 BD: 1 +.+.: cpuidle_driver_lock FD: 1 BD: 1 ....: thermal_cdev_ida.xa_lock FD: 1 BD: 4 ....: cpufreq_driver_lock FD: 25 BD: 1 +.+.: subsys mutex#31 ->&rq->__lock ->&k->k_lock FD: 1 BD: 1 +.+.: (x86_mce_decoder_chain).rwsem FD: 1 BD: 1 ....: virtio_index_ida.xa_lock FD: 1 BD: 1 +.+.: subsys mutex#32 FD: 166 BD: 142 +.+.: &md->mutex ->fs_reclaim ->pool_lock#2 ->irq_domain_mutex ->pci_config_lock ->&xa->xa_lock#5 ->&domain->mutex ->&irq_desc_lock_class ->vector_lock ->&root->kernfs_rwsem ->lock ->&c->lock ->&____s->seqcount ->&zone->lock ->&rq->__lock FD: 2 BD: 143 +.+.: &xa->xa_lock#5 ->pool_lock#2 FD: 1 BD: 1 +.+.: &dev->vqs_list_lock FD: 1 BD: 1 ....: &vp_dev->lock FD: 1 BD: 1 +.+.: (oom_notify_list).rwsem FD: 1 BD: 1 ....: &dev->config_lock FD: 23 BD: 1 +.+.: vdpa_dev_lock ->&rq->__lock ->&cfs_rq->removed.lock ->&obj_hash[i].lock FD: 3 BD: 1 +.+.: subsys mutex#33 ->&k->k_lock FD: 25 BD: 1 -.-.: &vb->stop_update_lock FD: 288 BD: 1 +.+.: (wq_completion)events_freezable ->(work_completion)(&vb->update_balloon_stats_work) ->&rq->__lock FD: 287 BD: 2 +.+.: (work_completion)(&vb->update_balloon_stats_work) ->cpu_hotplug_lock ->&s->s_inode_list_lock ->&rq->__lock ->rcu_node_0 ->&rcu_state.expedited_wq FD: 25 BD: 1 ..-.: &(&krcp->monitor_work)->timer FD: 25 BD: 1 ..-.: &(&tbl->managed_work)->timer FD: 34 BD: 2 +.+.: (work_completion)(&(&krcp->monitor_work)->work) ->krc.lock ->&obj_hash[i].lock ->&rq->__lock ->rcu_node_0 ->&rcu_state.expedited_wq FD: 265 BD: 1 +.+.: serial_mutex ->gpio_lookup_lock ->port_mutex FD: 1 BD: 2 +.+.: gpio_lookup_lock FD: 263 BD: 2 +.+.: port_mutex ->fs_reclaim ->pool_lock#2 ->&x->wait#9 ->&obj_hash[i].lock ->&k->list_lock ->&____s->seqcount ->lock ->&root->kernfs_rwsem ->bus_type_sem ->sysfs_symlink_target_lock ->&k->k_lock ->&c->lock ->&dev->power.lock ->dpm_list_mtx ->uevent_sock_mutex ->running_helpers_waitq.lock ->subsys mutex#34 ->&xa->xa_lock#6 ->&rq->__lock ->&port->mutex ->&zone->lock FD: 1 BD: 3 +.+.: subsys mutex#34 FD: 1 BD: 3 ....: &xa->xa_lock#6 FD: 1 BD: 4186 ....: &dev->power.wait_queue FD: 146 BD: 1 +.+.: (wq_completion)pm ->(work_completion)(&dev->power.work) FD: 145 BD: 2 +.+.: (work_completion)(&dev->power.work) ->&dev->power.lock ->&hub->irq_urb_lock ->(&hub->irq_urb_retry) ->&obj_hash[i].lock ->&base->lock ->hcd_urb_unlink_lock ->hcd_root_hub_lock ->&rq->__lock ->(work_completion)(&hub->tt.clear_work) ->&dum_hcd->dum->lock ->device_state_lock ->hcd_urb_list_lock ->usb_kill_urb_queue.lock ->&pool->lock ->&vhci_hcd->vhci->lock ->mmu_notifier_invalidate_range_start ->pool_lock#2 ->fs_reclaim ->&x->wait#19 ->(&timer.timer) ->&c->lock ->&zone->lock ->&____s->seqcount ->batched_entropy_u8.lock ->kfence_freelist_lock ->&meta->lock ->&cfs_rq->removed.lock ->&port_lock_key FD: 257 BD: 9 +.+.: &port->mutex ->fs_reclaim ->pool_lock#2 ->console_mutex ->resource_lock ->&port_lock_key ->(console_sem).lock ->&rq->__lock ->&cfs_rq->removed.lock ->&obj_hash[i].lock ->ctrl_ida.xa_lock ->&x->wait#9 ->&dev->power.lock ->&k->list_lock ->&c->lock ->&____s->seqcount ->lock ->&root->kernfs_rwsem ->bus_type_sem ->sysfs_symlink_target_lock ->&k->k_lock ->dpm_list_mtx ->&zone->lock ->uevent_sock_mutex ->running_helpers_waitq.lock ->subsys mutex#35 ->semaphore->lock ->*(&acpi_gbl_reference_count_lock) ->dev_pm_qos_sysfs_mtx ->kernfs_idr_lock ->deferred_probe_mutex ->device_links_lock ->mmu_notifier_invalidate_range_start ->gdp_mutex ->req_lock ->&p->pi_lock ->&x->wait#11 ->subsys mutex#21 ->chrdevs_lock ->batched_entropy_u8.lock ->kfence_freelist_lock ->&meta->lock ->hash_mutex ->&i->lock ->&desc->request_mutex ->register_lock ->&irq_desc_lock_class ->proc_subdir_lock ->proc_inum_ida.xa_lock FD: 1 BD: 10 ....: ctrl_ida.xa_lock FD: 1 BD: 10 +.+.: subsys mutex#35 FD: 1 BD: 1 ....: rng_index_ida.xa_lock FD: 129 BD: 1 +.+.: rng_mutex ->&x->wait#13 ->fs_reclaim ->pool_lock#2 ->kthread_create_lock ->&p->pi_lock ->&rq->__lock ->&x->wait ->&obj_hash[i].lock FD: 24 BD: 2 -.-.: &x->wait#12 ->&p->pi_lock FD: 1 BD: 2 ....: &x->wait#13 FD: 26 BD: 1 +.+.: reading_mutex ->reading_mutex.wait_lock ->&rq->__lock ->&x->wait#12 FD: 1 BD: 2 +.+.: reading_mutex.wait_lock FD: 1 BD: 1 ....: &dev->managed.lock FD: 137 BD: 1 +.+.: &type->s_umount_key#24/1 ->fs_reclaim ->pool_lock#2 ->pcpu_alloc_mutex ->shrinker_rwsem ->&c->lock ->&zone->lock ->&____s->seqcount ->list_lrus_mutex ->sb_lock ->mmu_notifier_invalidate_range_start ->&sb->s_type->i_lock_key#21 ->&s->s_inode_list_lock ->tk_core.seq.seqcount ->&dentry->d_lock FD: 40 BD: 2 +.+.: &sb->s_type->i_lock_key#21 ->&dentry->d_lock FD: 9 BD: 248 ....: drm_minor_lock ->&c->lock ->&zone->lock ->&____s->seqcount ->pool_lock#2 FD: 1 BD: 1 +.+.: &dev->debugfs_mutex FD: 3 BD: 3 +.+.: subsys mutex#36 ->&k->k_lock FD: 1 BD: 1 ....: (worker)->lock FD: 126 BD: 24 +.+.: &dev->mode_config.idr_mutex ->fs_reclaim ->pool_lock#2 FD: 149 BD: 20 +.+.: crtc_ww_class_acquire ->crtc_ww_class_mutex ->fs_reclaim ->pool_lock#2 FD: 148 BD: 21 +.+.: crtc_ww_class_mutex ->reservation_ww_class_acquire ->fs_reclaim ->pool_lock#2 ->&c->lock ->&zone->lock ->&____s->seqcount ->&rq->__lock ->&obj_hash[i].lock ->&dev->mode_config.idr_mutex ->&dev->mode_config.blob_lock ->&crtc->commit_lock ->reservation_ww_class_mutex ->tk_core.seq.seqcount ->&vkms_out->lock ->&dev->vbl_lock ->&x->wait#14 ->(work_completion)(&vkms_state->composer_work) ->&base->lock ->(&timer.timer) ->(work_completion)(&vkms_state->composer_work)#2 FD: 1 BD: 22 +.+.: &dev->mode_config.blob_lock FD: 1 BD: 1 ....: &xa->xa_lock#7 FD: 1 BD: 1 ....: &xa->xa_lock#8 FD: 1 BD: 23 ....: &dev->mode_config.connector_list_lock FD: 19 BD: 25 ..-.: &dev->vbl_lock ->&dev->vblank_time_lock FD: 187 BD: 1 .+.+: drm_connector_list_iter ->&dev->mode_config.connector_list_lock ->fs_reclaim ->pool_lock#2 ->&connector->mutex FD: 185 BD: 2 +.+.: &connector->mutex ->fs_reclaim ->pool_lock#2 ->&x->wait#9 ->&obj_hash[i].lock ->&k->list_lock ->lock ->&root->kernfs_rwsem ->bus_type_sem ->sysfs_symlink_target_lock ->&c->lock ->&____s->seqcount ->&dev->power.lock ->dpm_list_mtx ->&zone->lock ->uevent_sock_mutex ->running_helpers_waitq.lock ->&rq->__lock ->&k->k_lock ->subsys mutex#36 ->pin_fs_lock ->&sb->s_type->i_mutex_key#3 ->&dev->mode_config.idr_mutex ->connector_list_lock FD: 1 BD: 3 +.+.: connector_list_lock FD: 1 BD: 1 +.+.: &dev->filelist_mutex FD: 228 BD: 1 +.+.: &dev->clientlist_mutex ->&helper->lock ->registration_lock ->(console_sem).lock ->kernel_fb_helper_lock FD: 178 BD: 16 +.+.: &helper->lock ->fs_reclaim ->pool_lock#2 ->&client->modeset_mutex ->&obj_hash[i].lock ->mmu_notifier_invalidate_range_start ->&sb->s_type->i_lock_key ->&s->s_inode_list_lock ->tk_core.seq.seqcount ->batched_entropy_u32.lock ->&mgr->vm_lock ->&dev->object_name_lock ->&node->vm_lock ->&file_private->table_lock ->&dev->mode_config.idr_mutex ->&dev->mode_config.fb_lock ->&file->fbs_lock ->&prime_fpriv->lock ->free_vmap_area_lock ->vmap_area_lock ->&____s->seqcount ->init_mm.page_table_lock ->&zone->lock ->&rq->__lock ->&dev->master_mutex ->&lock->wait_lock ->&pool->lock ->reservation_ww_class_mutex FD: 151 BD: 18 +.+.: &client->modeset_mutex ->&dev->mode_config.mutex ->fs_reclaim ->pool_lock#2 ->crtc_ww_class_acquire FD: 150 BD: 19 +.+.: &dev->mode_config.mutex ->crtc_ww_class_acquire ->fs_reclaim ->pool_lock#2 ->&obj_hash[i].lock FD: 6 BD: 17 +.+.: &mgr->vm_lock ->&____s->seqcount ->&zone->lock ->pool_lock#2 ->&obj_hash[i].lock FD: 46 BD: 17 +.+.: &dev->object_name_lock ->lock FD: 4 BD: 248 +.+.: &file_private->table_lock ->pool_lock#2 ->&obj_hash[i].lock FD: 4 BD: 17 +.+.: &node->vm_lock ->&obj_hash[i].lock ->pool_lock#2 FD: 1 BD: 17 +.+.: &dev->mode_config.fb_lock FD: 1 BD: 17 +.+.: &file->fbs_lock FD: 1 BD: 17 +.+.: &prime_fpriv->lock FD: 226 BD: 2 +.+.: registration_lock ->fs_reclaim ->pool_lock#2 ->&x->wait#9 ->&obj_hash[i].lock ->&k->list_lock ->gdp_mutex ->lock ->&root->kernfs_rwsem ->bus_type_sem ->sysfs_symlink_target_lock ->&dev->power.lock ->dpm_list_mtx ->req_lock ->&p->pi_lock ->&rq->__lock ->&x->wait#11 ->&c->lock ->&zone->lock ->&____s->seqcount ->uevent_sock_mutex ->running_helpers_waitq.lock ->&k->k_lock ->subsys mutex#11 ->vt_switch_mutex ->(console_sem).lock ->console_lock FD: 126 BD: 3 +.+.: vt_switch_mutex ->fs_reclaim ->pool_lock#2 FD: 1 BD: 14 +.+.: &fb_info->lock FD: 152 BD: 17 +.+.: &dev->master_mutex ->&client->modeset_mutex FD: 1 BD: 22 +.+.: &crtc->commit_lock FD: 130 BD: 171 +.+.: &shmem->vmap_lock ->&shmem->pages_lock ->fs_reclaim ->pool_lock#2 ->free_vmap_area_lock ->vmap_area_lock ->&zone->lock ->&____s->seqcount ->init_mm.page_table_lock FD: 126 BD: 172 +.+.: &shmem->pages_lock ->fs_reclaim ->pool_lock#2 ->&sb->s_type->i_lock_key ->&zone->lock ->&____s->seqcount ->&xa->xa_lock#9 ->lock#4 ->&info->lock ->&rq->__lock FD: 39 BD: 4561 ..-.: &xa->xa_lock#9 ->pool_lock#2 ->batched_entropy_u8.lock ->kfence_freelist_lock ->&obj_hash[i].lock ->&c->lock ->&zone->lock ->&____s->seqcount ->&n->list_lock ->key#10 ->&s->s_inode_wblist_lock ->&base->lock ->key#12 ->&wb->work_lock ->&pl->lock ->key#13 ->stock_lock ->&____s->seqcount#2 ->&xa->xa_lock#4 ->key#30 FD: 1 BD: 4501 +.+.: &info->lock FD: 35 BD: 22 -.-.: &vkms_out->lock ->&dev->event_lock FD: 34 BD: 23 -.-.: &dev->event_lock ->&dev->vbl_lock ->&____s->seqcount#6 ->&x->wait#14 ->&obj_hash[i].lock ->pool_lock#2 ->&dev->vblank_time_lock ->&vblank->queue ->&base->lock FD: 1 BD: 28 ----: &____s->seqcount#6 FD: 24 BD: 24 -...: &x->wait#14 ->&p->pi_lock FD: 18 BD: 26 -.-.: &dev->vblank_time_lock ->tk_core.seq.seqcount ->&(&vblank->seqlock)->lock ->&obj_hash[i].lock ->hrtimer_bases.lock FD: 2 BD: 27 -.-.: &(&vblank->seqlock)->lock ->&____s->seqcount#6 FD: 1 BD: 22 +.+.: (work_completion)(&vkms_state->composer_work) FD: 1 BD: 18 ....: &helper->damage_lock FD: 180 BD: 2 +.+.: (work_completion)(&helper->damage_work) ->&helper->damage_lock ->&helper->lock FD: 1 BD: 4048 +.+.: &lock->wait_lock FD: 1 BD: 24 -.-.: &vblank->queue FD: 1 BD: 22 +.+.: (work_completion)(&vkms_state->composer_work)#2 FD: 1 BD: 14 ....: vt_event_lock FD: 1 BD: 2 +.+.: kernel_fb_helper_lock FD: 1 BD: 1 +...: &dev->queue_lock FD: 1 BD: 8 ....: blk_queue_ida.xa_lock FD: 306 BD: 10 +.+.: &q->sysfs_lock ->&q->unused_hctx_lock ->mmu_notifier_invalidate_range_start ->pool_lock#2 ->&obj_hash[i].lock ->&c->lock ->&n->list_lock ->&zone->lock ->&____s->seqcount ->cpu_hotplug_lock ->fs_reclaim ->&xa->xa_lock#10 ->&q->debugfs_mutex ->pcpu_alloc_mutex ->&q->rq_qos_mutex ->&stats->lock ->&rq->__lock ->lock ->&root->kernfs_rwsem ->sysfs_symlink_target_lock ->kernfs_idr_lock ->&q->queue_lock ->&tags->lock ->&eq->sysfs_lock FD: 1 BD: 11 +.+.: &q->unused_hctx_lock FD: 2 BD: 13 +.+.: &xa->xa_lock#10 ->pool_lock#2 FD: 27 BD: 7 +.+.: &set->tag_list_lock ->&q->mq_freeze_lock ->percpu_ref_switch_lock ->&rq->__lock FD: 2 BD: 8 +.+.: &xa->xa_lock#11 ->pool_lock#2 FD: 32 BD: 306 ....: &q->queue_lock ->&blkcg->lock ->pool_lock#2 ->pcpu_lock ->&obj_hash[i].lock ->percpu_counters_lock ->&c->lock ->&____s->seqcount ->&zone->lock FD: 30 BD: 307 ....: &blkcg->lock ->pool_lock#2 ->percpu_ref_switch_lock ->(&sq->pending_timer) ->&obj_hash[i].lock ->&base->lock ->percpu_counters_lock ->pcpu_lock ->pool_lock FD: 26 BD: 13 +.+.: &q->mq_freeze_lock ->percpu_ref_switch_lock ->&q->mq_freeze_wq ->&rq->__lock FD: 25 BD: 313 ..-.: percpu_ref_switch_lock ->&obj_hash[i].lock ->pool_lock#2 FD: 24 BD: 314 ..-.: &q->mq_freeze_wq ->&p->pi_lock FD: 1 BD: 15 +.+.: &bdev->bd_size_lock FD: 3 BD: 8 +.+.: subsys mutex#37 ->&k->k_lock FD: 307 BD: 8 +.+.: &q->sysfs_dir_lock ->fs_reclaim ->pool_lock#2 ->lock ->&root->kernfs_rwsem ->&c->lock ->&____s->seqcount ->&q->sysfs_lock ->&zone->lock ->&obj_hash[i].lock ->sysfs_symlink_target_lock ->kernfs_idr_lock ->&sem->wait_lock ->&p->pi_lock ->&rq->__lock FD: 138 BD: 13 +.+.: &q->debugfs_mutex ->pin_fs_lock ->&sb->s_type->i_mutex_key#3 ->&dentry->d_lock ->&fsnotify_mark_srcu ->&sb->s_type->i_lock_key#7 ->&s->s_inode_list_lock ->&xa->xa_lock#9 ->&rq->__lock ->&obj_hash[i].lock ->pool_lock#2 ->mount_lock FD: 3 BD: 7 +.+.: subsys mutex#38 ->&k->k_lock FD: 1 BD: 8 ....: cgwb_lock FD: 1 BD: 7 +...: bdi_lock FD: 62 BD: 4531 +.+.: inode_hash_lock ->&sb->s_type->i_lock_key#3 ->&sb->s_type->i_lock_key#22 ->&s->s_inode_list_lock ->&sb->s_type->i_lock_key#24 ->&sb->s_type->i_lock_key#30 ->&sb->s_type->i_lock_key#31 FD: 2 BD: 6 +.+.: bdev_lock ->&bdev->bd_holder_lock FD: 328 BD: 5 +.+.: &disk->open_mutex ->fs_reclaim ->pool_lock#2 ->&c->lock ->&____s->seqcount ->free_vmap_area_lock ->vmap_area_lock ->&zone->lock ->init_mm.page_table_lock ->&xa->xa_lock#9 ->lock#4 ->mmu_notifier_invalidate_range_start ->&mapping->private_lock ->tk_core.seq.seqcount ->&ret->b_uptodate_lock ->&obj_hash[i].lock ->purge_vmap_area_lock ->&sb->s_type->i_lock_key#3 ->lock#5 ->&lruvec->lru_lock ->&rq->__lock ->&base->lock ->&hctx->lock ->&x->wait#16 ->(&timer.timer) ->&q->sysfs_dir_lock ->&bdev->bd_size_lock ->&dd->lock ->&folio_wait_table[i] ->(console_sem).lock ->console_owner_lock ->console_owner ->&s->s_inode_list_lock ->pcpu_alloc_mutex ->&x->wait#9 ->&k->list_lock ->lock ->&root->kernfs_rwsem ->bus_type_sem ->sysfs_symlink_target_lock ->&dev->power.lock ->dpm_list_mtx ->req_lock ->&p->pi_lock ->&x->wait#11 ->&cfs_rq->removed.lock ->&k->k_lock ->subsys mutex#37 ->&xa->xa_lock#11 ->inode_hash_lock ->bdev_lock ->&lo->lo_mutex ->nbd_index_mutex ->&nbd->config_lock ->&new->lock ->rcu_node_0 ->nbd_index_mutex.wait_lock ->&lock->wait_lock FD: 41 BD: 4535 +.+.: &mapping->private_lock ->&xa->xa_lock#9 FD: 25 BD: 8 ..-.: &ret->b_uptodate_lock ->bit_wait_table + i FD: 13 BD: 5 ....: floppy_lock ->&obj_hash[i].lock ->&base->lock FD: 36 BD: 2 +.+.: (work_completion)(&blkg->free_work) ->&q->blkcg_mutex ->&obj_hash[i].lock ->pool_lock#2 ->&xa->xa_lock#10 ->pcpu_lock ->blk_queue_ida.xa_lock ->percpu_ref_switch_lock FD: 24 BD: 4 ....: command_done.lock ->&p->pi_lock FD: 16 BD: 2 +.+.: floppy_work ->dma_spin_lock ->floppy_lock ->&obj_hash[i].lock ->fdc_wait.lock FD: 1 BD: 3 ....: dma_spin_lock FD: 126 BD: 1 +.+.: loop_ctl_mutex ->fs_reclaim ->pool_lock#2 FD: 146 BD: 11 +.+.: &q->rq_qos_mutex ->&q->mq_freeze_lock ->percpu_ref_switch_lock ->&q->debugfs_mutex ->set->srcu ->&stats->lock ->(&cb->timer) ->&obj_hash[i].lock ->&base->lock ->pool_lock#2 FD: 1 BD: 12 ....: &stats->lock FD: 25 BD: 1 ..-.: &(&ops->cursor_work)->timer FD: 27 BD: 2 +.+.: (work_completion)(&(&ops->cursor_work)->work) ->(console_sem).lock ->&obj_hash[i].lock ->&base->lock FD: 158 BD: 11 +.+.: nbd_index_mutex ->fs_reclaim ->pool_lock#2 ->&nbd->config_lock ->nbd_index_mutex.wait_lock ->&rq->__lock FD: 23 BD: 19 .+.+: set->srcu ->&rq->__lock FD: 30 BD: 8 +.+.: (work_completion)(&(&q->requeue_work)->work) ->&q->requeue_lock ->&hctx->lock ->&dd->lock ->rcu_node_0 ->&rq->__lock FD: 28 BD: 8 +.+.: (work_completion)(&(&hctx->run_work)->work) FD: 319 BD: 1 +.+.: zram_index_mutex ->fs_reclaim ->pool_lock#2 ->blk_queue_ida.xa_lock ->&obj_hash[i].lock ->pcpu_alloc_mutex ->bio_slab_lock ->&c->lock ->&____s->seqcount ->percpu_counters_lock ->mmu_notifier_invalidate_range_start ->&sb->s_type->i_lock_key#3 ->&s->s_inode_list_lock ->&xa->xa_lock#11 ->lock ->&q->queue_lock ->&x->wait#9 ->&bdev->bd_size_lock ->&n->list_lock ->&k->list_lock ->gdp_mutex ->&zone->lock ->&root->kernfs_rwsem ->bus_type_sem ->sysfs_symlink_target_lock ->&dev->power.lock ->dpm_list_mtx ->req_lock ->&p->pi_lock ->&x->wait#11 ->&rq->__lock ->subsys mutex#37 ->dev_hotplug_mutex ->&q->sysfs_dir_lock ->percpu_ref_switch_lock ->uevent_sock_mutex ->running_helpers_waitq.lock ->subsys mutex#38 ->cgwb_lock ->pin_fs_lock ->&sb->s_type->i_mutex_key#3 ->bdi_lock ->inode_hash_lock ->(console_sem).lock FD: 3 BD: 1 +.+.: subsys mutex#39 ->&k->k_lock FD: 127 BD: 2 +.+.: &default_group_class[depth - 1]#2 ->fs_reclaim ->pool_lock#2 ->configfs_dirent_lock FD: 2 BD: 1 +.+.: &lock ->nullb_indexes.xa_lock FD: 1 BD: 2 ....: nullb_indexes.xa_lock FD: 1 BD: 1 +.+.: ctx_list.lock FD: 1 BD: 1 ....: nfc_index_ida.xa_lock FD: 172 BD: 3 +.+.: nfc_devlist_mutex ->fs_reclaim ->&c->lock ->&____s->seqcount ->pool_lock#2 ->&k->list_lock ->gdp_mutex ->lock ->&root->kernfs_rwsem ->bus_type_sem ->sysfs_symlink_target_lock ->&dev->power.lock ->dpm_list_mtx ->uevent_sock_mutex ->&obj_hash[i].lock ->running_helpers_waitq.lock ->subsys mutex#40 ->&zone->lock ->&k->k_lock ->&genl_data->genl_data_mutex ->&rq->__lock FD: 3 BD: 4 +.+.: subsys mutex#40 ->&k->k_lock FD: 1 BD: 87 ....: &rfkill->lock FD: 3 BD: 11 +.+.: subsys mutex#41 ->&k->k_lock FD: 178 BD: 2 +.+.: (work_completion)(&rfkill->sync_work) ->rfkill_global_mutex ->rfkill_global_mutex.wait_lock ->&p->pi_lock ->&rq->__lock FD: 1 BD: 11 +.+.: rfkill_global_mutex.wait_lock FD: 1 BD: 1 +.+.: dma_heap_minors.xa_lock FD: 3 BD: 1 +.+.: subsys mutex#42 ->&k->k_lock FD: 1 BD: 1 +.+.: heap_list_lock FD: 1 BD: 1 ....: host_index_ida.xa_lock FD: 154 BD: 1 +.+.: scsi_sense_cache_mutex ->slab_mutex FD: 1 BD: 4 +.+.: subsys mutex#43 FD: 3 BD: 1 +.+.: subsys mutex#44 ->&k->k_lock FD: 1 BD: 182 -.-.: &virtscsi_vq->vq_lock FD: 337 BD: 3 +.+.: &shost->scan_mutex ->fs_reclaim ->pool_lock#2 ->shost->host_lock ->&dev->power.lock ->&x->wait#9 ->&obj_hash[i].lock ->attribute_container_mutex ->blk_queue_ida.xa_lock ->pcpu_alloc_mutex ->&q->sysfs_lock ->&set->tag_list_lock ->batched_entropy_u32.lock ->&c->lock ->&zone->lock ->&____s->seqcount ->tk_core.seq.seqcount ->mmu_notifier_invalidate_range_start ->&hctx->lock ->&base->lock ->&x->wait#16 ->&rq->__lock ->&cfs_rq->removed.lock ->(&timer.timer) ->&sdev->state_mutex ->&q->mq_freeze_lock ->&q->mq_freeze_wq ->percpu_ref_switch_lock ->(&q->timeout) ->(work_completion)(&q->timeout_work) ->(work_completion)(&(&q->requeue_work)->work) ->(work_completion)(&(&hctx->run_work)->work) ->cpu_hotplug_lock ->&xa->xa_lock#10 ->&q->unused_hctx_lock ->(work_completion)(&sdev->requeue_work) ->(work_completion)(&sdev->event_work) ->pcpu_lock ->&sdev->inquiry_mutex ->(console_sem).lock ->batched_entropy_u8.lock ->kfence_freelist_lock ->&tags->lock ->pool_lock ->quarantine_lock ->&meta->lock ->&x->wait#15 ->&k->list_lock ->lock ->&root->kernfs_rwsem ->bus_type_sem ->sysfs_symlink_target_lock ->&k->k_lock ->dpm_list_mtx ->uevent_sock_mutex ->subsys mutex#43 ->device_links_srcu ->running_helpers_waitq.lock ->async_lock ->gdp_mutex ->subsys mutex#45 ->bsg_minor_ida.xa_lock ->chrdevs_lock ->req_lock ->&p->pi_lock ->&x->wait#11 ->subsys mutex#58 FD: 1 BD: 4 ....: shost->host_lock FD: 2 BD: 3 +.+.: async_scan_lock ->&x->wait#15 FD: 1 BD: 5 ....: &x->wait#15 FD: 181 BD: 4 +.+.: subsys mutex#45 ->&k->list_lock ->&k->k_lock ->fs_reclaim ->pool_lock#2 ->lock ->chrdevs_lock ->&x->wait#9 ->&obj_hash[i].lock ->gdp_mutex ->&c->lock ->&zone->lock ->&____s->seqcount ->&root->kernfs_rwsem ->bus_type_sem ->sysfs_symlink_target_lock ->&dev->power.lock ->dpm_list_mtx ->req_lock ->&p->pi_lock ->&x->wait#11 ->&rq->__lock ->uevent_sock_mutex ->running_helpers_waitq.lock ->subsys mutex#57 ->(console_sem).lock ->console_owner_lock ->console_owner FD: 1 BD: 171 +.+.: &hctx->lock FD: 24 BD: 7 ..-.: &x->wait#16 ->&p->pi_lock FD: 1 BD: 4 +.+.: &sdev->state_mutex FD: 25 BD: 6 +.-.: (&q->timeout) FD: 24 BD: 7 +.+.: (work_completion)(&q->timeout_work) ->&tags->lock ->&obj_hash[i].lock ->&base->lock FD: 1 BD: 4 +.+.: (work_completion)(&sdev->requeue_work) FD: 1 BD: 4 +.+.: (work_completion)(&sdev->event_work) FD: 1 BD: 4 +.+.: &sdev->inquiry_mutex FD: 1 BD: 1 +.+.: nvmf_hosts_mutex FD: 3 BD: 1 +.+.: subsys mutex#46 ->&k->k_lock FD: 1 BD: 1 +.+.: nvmf_transports_rwsem FD: 3 BD: 1 +.+.: subsys mutex#47 ->&k->k_lock FD: 132 BD: 3 +.+.: &default_group_class[depth - 1]#3/2 ->fs_reclaim ->pool_lock#2 ->&dentry->d_lock ->configfs_dirent_lock ->mmu_notifier_invalidate_range_start ->&sb->s_type->i_lock_key#19 ->&s->s_inode_list_lock ->tk_core.seq.seqcount ->&default_group_class[depth - 1]#4/2 FD: 1 BD: 1 +.+.: nvmet_config_sem FD: 3 BD: 1 +.+.: subsys mutex#48 ->&k->k_lock FD: 1 BD: 13 ....: &tags->lock FD: 131 BD: 4 +.+.: &default_group_class[depth - 1]#4/2 ->fs_reclaim ->pool_lock#2 ->&dentry->d_lock ->configfs_dirent_lock ->mmu_notifier_invalidate_range_start ->&sb->s_type->i_lock_key#19 ->&s->s_inode_list_lock ->tk_core.seq.seqcount ->&default_group_class[depth - 1]#5/2 FD: 130 BD: 5 +.+.: &default_group_class[depth - 1]#5/2 ->fs_reclaim ->pool_lock#2 ->&dentry->d_lock ->configfs_dirent_lock ->mmu_notifier_invalidate_range_start ->&sb->s_type->i_lock_key#19 ->&s->s_inode_list_lock ->tk_core.seq.seqcount ->&default_group_class[depth - 1]#6 ->&default_group_class[depth - 1]#6/2 FD: 127 BD: 6 +.+.: &default_group_class[depth - 1]#6 ->fs_reclaim ->pool_lock#2 ->configfs_dirent_lock FD: 1 BD: 6 +.+.: &default_group_class[depth - 1]#6/2 FD: 1 BD: 1 +.+.: backend_mutex FD: 1 BD: 1 +.+.: scsi_mib_index_lock FD: 1 BD: 1 +.+.: hba_lock FD: 126 BD: 1 +.+.: device_mutex ->fs_reclaim ->pool_lock#2 FD: 1 BD: 1 +.+.: &hba->device_lock FD: 332 BD: 1 +.+.: mtd_table_mutex ->fs_reclaim ->pool_lock#2 ->&x->wait#9 ->&obj_hash[i].lock ->&k->list_lock ->gdp_mutex ->lock ->&root->kernfs_rwsem ->bus_type_sem ->sysfs_symlink_target_lock ->&zone->lock ->&____s->seqcount ->&c->lock ->&dev->power.lock ->dpm_list_mtx ->req_lock ->&p->pi_lock ->&x->wait#11 ->&rq->__lock ->uevent_sock_mutex ->running_helpers_waitq.lock ->subsys mutex#49 ->devtree_lock ->nvmem_ida.xa_lock ->nvmem_cell_mutex ->&k->k_lock ->subsys mutex#50 ->pin_fs_lock ->&sb->s_type->i_mutex_key#3 ->&cfs_rq->removed.lock ->(console_sem).lock ->pcpu_alloc_mutex ->cpu_hotplug_lock ->batched_entropy_u32.lock ->mmu_notifier_invalidate_range_start ->blk_queue_ida.xa_lock ->&q->sysfs_lock ->&set->tag_list_lock ->bio_slab_lock ->percpu_counters_lock ->&sb->s_type->i_lock_key#3 ->&s->s_inode_list_lock ->&xa->xa_lock#11 ->&q->mq_freeze_lock ->set->srcu ->percpu_ref_switch_lock ->&q->queue_lock ->&bdev->bd_size_lock ->elv_list_lock ->(work_completion)(&(&q->requeue_work)->work) ->(work_completion)(&(&hctx->run_work)->work) ->&q->debugfs_mutex ->subsys mutex#37 ->dev_hotplug_mutex ->&q->sysfs_dir_lock ->subsys mutex#38 ->cgwb_lock ->bdi_lock ->inode_hash_lock FD: 1 BD: 1 +.+.: part_parser_lock FD: 1 BD: 86 ....: (kmod_concurrent_max).lock FD: 24 BD: 87 ....: &x->wait#17 ->&p->pi_lock FD: 1 BD: 156 ....: &prev->lock FD: 3 BD: 2 +.+.: subsys mutex#49 ->&k->k_lock FD: 1 BD: 2 ....: nvmem_ida.xa_lock FD: 1 BD: 2 +.+.: nvmem_cell_mutex FD: 1 BD: 2 +.+.: subsys mutex#50 FD: 1 BD: 3826 +.+.: &bond->stats_lock FD: 17 BD: 108 ....: lweventlist_lock ->pool_lock#2 ->&dir->lock#2 ->&c->lock ->&n->list_lock ->&____s->seqcount ->&obj_hash[i].lock ->&base->lock ->&____s->seqcount#2 ->batched_entropy_u8.lock ->kfence_freelist_lock FD: 890 BD: 2 +.+.: (linkwatch_work).work ->rtnl_mutex ->rtnl_mutex.wait_lock ->&p->pi_lock ->&rq->__lock ->&cfs_rq->removed.lock ->&obj_hash[i].lock ->pool_lock#2 FD: 1 BD: 3889 +.+.: rtnl_mutex.wait_lock FD: 926 BD: 1 +.+.: (wq_completion)gid-cache-wq ->(work_completion)(&ndev_work->work) ->(work_completion)(&work->work) ->&rq->__lock FD: 924 BD: 2 +.+.: (work_completion)(&ndev_work->work) ->devices_rwsem ->&obj_hash[i].lock ->pool_lock#2 ->&rq->__lock ->&meta->lock ->kfence_freelist_lock ->&base->lock ->quarantine_lock ->&cfs_rq->removed.lock FD: 3 BD: 85 ..-.: once_lock ->crngs.lock FD: 287 BD: 2 +.+.: (work_completion)(&w->work) ->cpu_hotplug_lock ->&obj_hash[i].lock FD: 23 BD: 81 ++++: (inet6addr_validator_chain).rwsem ->&rq->__lock FD: 23 BD: 78 ++++: (inetaddr_validator_chain).rwsem ->&rq->__lock FD: 3 BD: 1 +.+.: subsys mutex#51 ->&k->k_lock FD: 1 BD: 1 +.+.: mdio_board_lock FD: 1 BD: 80 +.+.: mode_list_lock FD: 1 BD: 79 +.+.: napi_hash_lock FD: 130 BD: 142 +.+.: xps_map_mutex ->fs_reclaim ->pool_lock#2 ->&c->lock ->&n->list_lock ->&zone->lock ->&____s->seqcount ->jump_label_mutex ->&obj_hash[i].lock ->krc.lock ->&rq->__lock ->&cfs_rq->removed.lock FD: 1 BD: 1 +.+.: l3mdev_lock FD: 1 BD: 2 +.+.: (work_completion)(&vi->config_work) FD: 3 BD: 1 +.+.: subsys mutex#52 ->&k->k_lock FD: 2 BD: 2 +.+.: compressor_list_lock ->pool_lock#2 FD: 1 BD: 5 ....: hwsim_netgroup_ida.xa_lock FD: 30 BD: 3881 +.-.: hwsim_radio_lock ->pool_lock#2 ->&list->lock#19 ->&c->lock ->&____s->seqcount#2 ->&zone->lock ->&____s->seqcount ->&n->list_lock ->batched_entropy_u8.lock ->kfence_freelist_lock ->init_task.mems_allowed_seq.seqcount FD: 3 BD: 6 +.+.: subsys mutex#53 ->&k->k_lock FD: 542 BD: 80 +.+.: &rdev->wiphy.mtx ->fs_reclaim ->pool_lock#2 ->&k->list_lock ->gdp_mutex ->lock ->&root->kernfs_rwsem ->bus_type_sem ->sysfs_symlink_target_lock ->&____s->seqcount ->&c->lock ->&zone->lock ->&dev->power.lock ->dpm_list_mtx ->uevent_sock_mutex ->&obj_hash[i].lock ->running_helpers_waitq.lock ->&k->k_lock ->subsys mutex#54 ->pin_fs_lock ->&sb->s_type->i_mutex_key#3 ->nl_table_lock ->nl_table_wait.lock ->reg_requests_lock ->stack_depot_init_mutex ->pcpu_alloc_mutex ->&local->iflist_mtx ->&xa->xa_lock#3 ->net_rwsem ->&x->wait#9 ->subsys mutex#17 ->&dir->lock#2 ->dev_hotplug_mutex ->dev_base_lock ->input_pool.lock ->batched_entropy_u32.lock ->&tbl->lock ->sysctl_lock ->&wdev->mtx ->&fq->lock ->quarantine_lock ->rlock-AF_NETLINK ->lweventlist_lock ->&pool->lock ->&data->mutex ->&base->lock ->&rq->__lock ->&tn->lock ->remove_cache_srcu ->failover_lock ->proc_subdir_lock ->proc_inum_ida.xa_lock ->&idev->mc_lock ->&pnettable->lock ->smc_ib_devices.mutex ->&ndev->lock ->&wdev->event_lock ->&rdev->mgmt_registrations_lock ->(work_completion)(&(&sdata->dec_tailroom_needed_wk)->work) ->&local->key_mtx ->&dentry->d_lock ->&fsnotify_mark_srcu ->&sb->s_type->i_lock_key#7 ->&s->s_inode_list_lock ->&xa->xa_lock#9 ->mount_lock ->&rdev->wiphy_work_lock ->(&dwork->timer) ->(work_completion)(&(&link->color_collision_detect_work)->work) ->&local->chanctx_mtx ->rtnl_mutex.wait_lock ->&p->pi_lock ->&list->lock#18 ->&lock->wait_lock ->lock#6 ->&____s->seqcount#2 ->&n->list_lock ->&local->mtx ->batched_entropy_u8.lock ->kfence_freelist_lock ->&meta->lock ->&list->lock#2 ->rcu_node_0 ->&rcu_state.expedited_wq ->&cfs_rq->removed.lock ->&local->sta_mtx ->_xmit_ETHER ->(&local->dynamic_ps_timer) ->(work_completion)(&local->dynamic_ps_enable_work) ->(work_completion)(&sdata->recalc_smps) ->(work_completion)(&link->csa_finalize_work) ->(work_completion)(&link->color_change_finalize_work) ->(work_completion)(&(&link->dfs_cac_timer_work)->work) ->&local->queue_stop_reason_lock ->&list->lock#19 ->(work_completion)(&local->reconfig_filter) ->(wq_completion)phy8 ->&wq->mutex ->(wq_completion)phy4 ->(wq_completion)phy7 ->stock_lock ->&sem->wait_lock ->(wq_completion)phy11 ->(wq_completion)phy14 ->kernfs_rename_lock ->&sb->s_type->i_mutex_key#3/1 ->cpu_hotplug_lock ->bpf_devs_lock ->&hwstats->hwsdev_list_lock ->&in_dev->mc_tomb_lock ->class ->(&tbl->proxy_timer) ->&ul->lock ->&net->xdp.lock ->krc.lock ->mirred_list_lock ->&nft_net->commit_mutex ->&ent->pde_unload_lock ->&net->ipv6.addrconf_hash_lock ->&idev->mc_query_lock ->(work_completion)(&(&idev->mc_report_work)->work) ->&idev->mc_report_lock ->&pnn->pndevs.lock ->&pnn->routes.lock ->target_list_lock ->kernfs_idr_lock ->dev_pm_qos_sysfs_mtx ->deferred_probe_mutex ->device_links_lock ->mmu_notifier_invalidate_range_start ->&local->filter_lock ->&data->lock ->(console_sem).lock FD: 3 BD: 81 +.+.: subsys mutex#54 ->&k->k_lock FD: 1 BD: 81 +.+.: reg_requests_lock FD: 27 BD: 91 +.+.: &local->iflist_mtx ->hrtimer_bases.lock ->tk_core.seq.seqcount ->&rq->__lock FD: 186 BD: 85 +.+.: &wdev->mtx ->&rdev->bss_lock ->&rq->__lock ->&local->chanctx_mtx ->&rdev->wiphy_work_lock ->&ifibss->incomplete_lock ->(console_sem).lock ->console_owner_lock ->console_owner ->&local->mtx ->fs_reclaim ->pool_lock#2 ->tk_core.seq.seqcount ->hrtimer_bases.lock ->&obj_hash[i].lock ->&base->lock ->&____s->seqcount ->&wdev->event_lock ->&c->lock ->nl_table_lock ->nl_table_wait.lock ->&list->lock#2 ->&n->list_lock ->&sta->lock ->&local->sta_mtx ->&____s->seqcount#2 ->rcu_node_0 ->quarantine_lock ->batched_entropy_u8.lock ->kfence_freelist_lock ->&meta->lock ->lweventlist_lock ->krc.lock ->rcu_state.exp_mutex.wait_lock ->&p->pi_lock ->&list->lock#18 ->remove_cache_srcu ->&rcu_state.expedited_wq ->&cfs_rq->removed.lock ->(&ifibss->timer) ->&wdev->pmsr_lock ->&lock->wait_lock ->crngs.lock ->&local->key_mtx FD: 5 BD: 3880 +.-.: &fq->lock ->tk_core.seq.seqcount FD: 3 BD: 78 +.+.: subsys mutex#55 ->&k->k_lock FD: 2 BD: 79 +.+.: &sdata->sec_mtx ->&sec->lock FD: 1 BD: 80 ++..: &sec->lock FD: 1 BD: 78 +.+.: &local->iflist_mtx#2 FD: 126 BD: 1 +.+.: hwsim_phys_lock ->fs_reclaim ->pool_lock#2 FD: 2 BD: 248 ....: sg_index_lock ->pool_lock#2 FD: 1 BD: 3 ....: sd_index_ida.xa_lock FD: 3 BD: 3 +.+.: subsys mutex#56 ->&k->k_lock FD: 3 BD: 5 +.+.: subsys mutex#57 ->&k->k_lock FD: 1 BD: 4 ....: bsg_minor_ida.xa_lock FD: 3 BD: 4 +.+.: subsys mutex#58 ->&k->k_lock FD: 1 BD: 190 +.+.: &dd->lock FD: 24 BD: 4513 ..-.: &folio_wait_table[i] ->&p->pi_lock FD: 36 BD: 1 +.+.: (wq_completion)kblockd ->(work_completion)(&(&hctx->run_work)->work) ->(work_completion)(&q->timeout_work) ->(work_completion)(&(&q->requeue_work)->work) ->&rq->__lock ->rcu_node_0 ->&rcu_state.expedited_wq FD: 126 BD: 1 +.+.: xdomain_lock ->fs_reclaim ->pool_lock#2 FD: 23 BD: 1 +.+.: ioctl_mutex ->&rq->__lock ->&cfs_rq->removed.lock ->&obj_hash[i].lock ->pool_lock#2 FD: 1 BD: 1 +.+.: address_handler_list_lock FD: 1 BD: 1 +.+.: card_mutex FD: 3 BD: 1 +.+.: subsys mutex#59 ->&k->k_lock FD: 24 BD: 1 ....: &x->wait#18 ->&p->pi_lock FD: 26 BD: 2 ..-.: &txlock ->&list->lock#3 ->&txwq FD: 1 BD: 3 ..-.: &list->lock#3 FD: 24 BD: 3 ..-.: &txwq ->&p->pi_lock FD: 2 BD: 1 ....: &iocq[i].lock ->&ktiowq[i] FD: 1 BD: 2 ....: &ktiowq[i] FD: 1 BD: 1 ....: rcu_read_lock_bh FD: 4 BD: 3908 +.-.: noop_qdisc.q.lock ->batched_entropy_u64.lock FD: 3 BD: 3 +.+.: subsys mutex#60 ->&k->k_lock FD: 225 BD: 1 +.+.: usb_bus_idr_lock ->(usb_notifier_list).rwsem ->fs_reclaim ->pool_lock#2 ->mmu_notifier_invalidate_range_start ->hcd_root_hub_lock ->&obj_hash[i].lock ->&rq->__lock ->&x->wait#19 ->&____s->seqcount ->&zone->lock ->&dev->power.lock ->device_links_srcu ->&base->lock ->(&timer.timer) ->(console_sem).lock ->input_pool.lock ->&k->list_lock ->lock ->&root->kernfs_rwsem ->bus_type_sem ->&c->lock ->sysfs_symlink_target_lock ->&k->k_lock ->dpm_list_mtx ->req_lock ->&p->pi_lock ->&x->wait#11 ->&(&priv->bus_notifier)->rwsem ->uevent_sock_mutex ->running_helpers_waitq.lock ->device_state_lock ->&dum_hcd->dum->lock ->batched_entropy_u8.lock ->kfence_freelist_lock ->subsys mutex#61 ->&x->wait#9 ->&lock->wait_lock ->&hub->irq_urb_lock ->(&hub->irq_urb_retry) ->hcd_urb_unlink_lock ->(work_completion)(&hub->tt.clear_work) ->hcd_urb_list_lock ->&vhci_hcd->vhci->lock ->quarantine_lock ->&cfs_rq->removed.lock ->usb_kill_urb_queue.lock FD: 171 BD: 1 +.+.: table_lock ->&k->list_lock ->fs_reclaim ->pool_lock#2 ->lock ->&root->kernfs_rwsem ->&k->k_lock ->&c->lock ->&____s->seqcount ->uevent_sock_mutex ->&obj_hash[i].lock ->running_helpers_waitq.lock ->(console_sem).lock ->&zone->lock ->&rq->__lock ->&cfs_rq->removed.lock FD: 1 BD: 3 +.+.: mon_lock FD: 163 BD: 2 +.+.: usb_port_peer_mutex ->fs_reclaim ->pool_lock#2 ->&x->wait#9 ->&obj_hash[i].lock ->&k->list_lock ->lock ->&root->kernfs_rwsem ->bus_type_sem ->&c->lock ->&zone->lock ->&____s->seqcount ->&dev->power.lock ->dpm_list_mtx ->&k->k_lock ->dev_pm_qos_mtx ->component_mutex ->device_links_srcu ->dev_pm_qos_sysfs_mtx ->&rq->__lock ->sysfs_symlink_target_lock FD: 1 BD: 4 ....: device_state_lock FD: 26 BD: 8 ....: hcd_root_hub_lock ->hcd_urb_list_lock ->&bh->lock ->&p->pi_lock FD: 1 BD: 9 ....: hcd_urb_list_lock FD: 1 BD: 9 ..-.: &bh->lock FD: 9 BD: 99 ..-.: lock#6 ->kcov_remote_lock ->&kcov->lock FD: 7 BD: 159 ..-.: kcov_remote_lock ->pool_lock#2 ->&c->lock FD: 24 BD: 8 ..-.: &x->wait#19 ->&p->pi_lock FD: 1 BD: 2 +.+.: set_config_lock FD: 136 BD: 2 +.+.: hcd->bandwidth_mutex ->devtree_lock ->&obj_hash[i].lock ->&x->wait#9 ->&dev->power.lock ->fs_reclaim ->pool_lock#2 ->mmu_notifier_invalidate_range_start ->hcd_root_hub_lock ->&rq->__lock ->&x->wait#19 ->&base->lock ->(&timer.timer) ->&c->lock ->&zone->lock ->&____s->seqcount FD: 1 BD: 2 +.+.: &new_driver->dynids.lock FD: 1 BD: 7 ....: &dum_hcd->dum->lock FD: 133 BD: 4 +.+.: &hub->status_mutex ->mmu_notifier_invalidate_range_start ->pool_lock#2 ->hcd_root_hub_lock ->fs_reclaim ->&dum_hcd->dum->lock ->&obj_hash[i].lock ->&rq->__lock ->&x->wait#19 ->&c->lock ->&zone->lock ->&____s->seqcount ->&base->lock ->(&timer.timer) ->&pool->lock ->&vhci_hcd->vhci->lock FD: 1 BD: 3 +.+.: component_mutex FD: 1 BD: 2 +.+.: subsys mutex#61 FD: 135 BD: 2 +.+.: (work_completion)(&(&hub->init_work)->work) FD: 36 BD: 1 +.+.: (wq_completion)usb_hub_wq ->(work_completion)(&hub->events) FD: 35 BD: 2 +.+.: (work_completion)(&hub->events) ->lock#6 ->&dev->power.lock FD: 1 BD: 4 ....: &hub->irq_urb_lock FD: 1 BD: 4 ....: (&hub->irq_urb_retry) FD: 1 BD: 4 ....: hcd_urb_unlink_lock FD: 24 BD: 4 ..-.: usb_kill_urb_queue.lock ->&p->pi_lock FD: 1 BD: 4 +.+.: (work_completion)(&hub->tt.clear_work) FD: 25 BD: 1 ..-.: lib/debugobjects.c:101 FD: 25 BD: 2 +.+.: (debug_obj_work).work ->pool_lock#2 ->&rq->__lock ->&meta->lock ->kfence_freelist_lock ->quarantine_lock FD: 1 BD: 10 +.+.: udc_lock FD: 3 BD: 1 +.+.: subsys mutex#62 ->&k->k_lock FD: 1 BD: 1 ....: gadget_id_numbers.xa_lock FD: 131 BD: 2 +.+.: (work_completion)(&gadget->work) ->&root->kernfs_rwsem ->kernfs_notify_lock FD: 25 BD: 164 ....: kernfs_notify_lock FD: 65 BD: 2 +.+.: kernfs_notify_work ->kernfs_notify_lock ->&root->kernfs_supers_rwsem FD: 63 BD: 7 ++++: &root->kernfs_supers_rwsem ->inode_hash_lock FD: 1 BD: 1 +.+.: subsys mutex#63 FD: 1 BD: 1 +.+.: func_lock FD: 1 BD: 1 +.+.: g_tf_lock FD: 25 BD: 4599 +.+.: &dentry->d_lock/1 ->&lru->node[i].lock FD: 1 BD: 7 ....: &vhci_hcd->vhci->lock FD: 25 BD: 1 ..-.: net/core/link_watch.c:31 FD: 39 BD: 7 -.-.: i8042_lock ->(console_sem).lock ->&x->wait#20 FD: 24 BD: 8 -...: &x->wait#20 ->&p->pi_lock FD: 1 BD: 160 ....: irq_resend_lock FD: 1 BD: 95 +.+.: &ent->pde_unload_lock FD: 25 BD: 4 ....: serio_event_lock ->pool_lock#2 FD: 246 BD: 1 +.+.: (wq_completion)events_long ->serio_event_work ->(work_completion)(&(&ipvs->defense_work)->work) ->(work_completion)(&(&br->gc_work)->work) ->(work_completion)(&br->mcast_gc_work) ->&rq->__lock FD: 220 BD: 2 +.+.: serio_event_work ->serio_mutex FD: 219 BD: 3 +.+.: serio_mutex ->serio_event_lock ->i8042_lock ->fs_reclaim ->pool_lock#2 ->&k->list_lock ->lock ->&root->kernfs_rwsem ->&device->physical_node_lock ->&c->lock ->&zone->lock ->&____s->seqcount ->semaphore->lock ->&obj_hash[i].lock ->sysfs_symlink_target_lock ->&k->k_lock ->&dev->power.lock ->dpm_list_mtx ->&(&priv->bus_notifier)->rwsem ->uevent_sock_mutex ->subsys mutex#64 ->bus_type_sem ->running_helpers_waitq.lock FD: 2 BD: 7 ....: input_ida.xa_lock ->pool_lock#2 FD: 1 BD: 4 +.+.: subsys mutex#64 FD: 33 BD: 7 +.+.: &mousedev->mutex/1 ->&mousedev->mutex#2 FD: 1 BD: 4527 ....: &sem->wait_lock FD: 206 BD: 4 +.+.: &serio->drv_mutex ->fs_reclaim ->pool_lock#2 ->&x->wait#9 ->&obj_hash[i].lock ->&serio->lock ->i8042_mutex ->&k->list_lock ->gdp_mutex ->lock ->&root->kernfs_rwsem ->bus_type_sem ->sysfs_symlink_target_lock ->&c->lock ->&zone->lock ->&____s->seqcount ->&dev->power.lock ->dpm_list_mtx ->uevent_sock_mutex ->running_helpers_waitq.lock ->&k->k_lock ->subsys mutex#30 ->(console_sem).lock ->input_mutex ->i8042_lock ->psmouse_mutex FD: 34 BD: 7 -.-.: &serio->lock ->&ps2dev->wait ->&dev->power.lock ->&dev->event_lock#2 FD: 44 BD: 6 +.+.: i8042_mutex ->&serio->lock ->i8042_lock ->&ps2dev->wait ->&obj_hash[i].lock ->pool_lock#2 ->&base->lock ->&pool->lock ->&rq->__lock ->&cfs_rq->removed.lock ->(&timer.timer) FD: 24 BD: 8 -.-.: &ps2dev->wait ->&p->pi_lock FD: 1 BD: 1 ....: rtc_ida.xa_lock FD: 2 BD: 1 +.+.: &rtc->ops_lock ->rtc_lock FD: 1 BD: 2 ....: platform_devid_ida.xa_lock FD: 1 BD: 2 ....: rtcdev_lock FD: 126 BD: 1 +.+.: g_smscore_deviceslock ->fs_reclaim ->pool_lock#2 FD: 175 BD: 7 +.+.: &led_cdev->led_access ->fs_reclaim ->pool_lock#2 ->&x->wait#9 ->&obj_hash[i].lock ->&k->list_lock ->lock ->&root->kernfs_rwsem ->bus_type_sem ->sysfs_symlink_target_lock ->&c->lock ->&zone->lock ->&____s->seqcount ->&dev->power.lock ->dpm_list_mtx ->uevent_sock_mutex ->running_helpers_waitq.lock ->&k->k_lock ->subsys mutex#65 ->leds_list_lock ->triggers_list_lock FD: 3 BD: 8 +.+.: subsys mutex#65 ->&k->k_lock FD: 140 BD: 21 +.+.: &led_cdev->trigger_lock ->fs_reclaim ->pool_lock#2 ->&trig->leddev_list_lock ->&c->lock ->&zone->lock ->&____s->seqcount ->uevent_sock_mutex ->&obj_hash[i].lock ->running_helpers_waitq.lock FD: 1 BD: 22 +.+.: &trig->leddev_list_lock FD: 1 BD: 24 -...: &dev->event_lock#2 FD: 1 BD: 1 +.+.: cx231xx_devlist_mutex FD: 1 BD: 1 +.+.: em28xx_devlist_mutex FD: 205 BD: 5 +.+.: psmouse_mutex ->fs_reclaim ->pool_lock#2 ->&x->wait#9 ->&obj_hash[i].lock ->&serio->lock ->i8042_mutex ->&rq->__lock ->&k->list_lock ->gdp_mutex ->lock ->&root->kernfs_rwsem ->bus_type_sem ->sysfs_symlink_target_lock ->&c->lock ->&zone->lock ->&____s->seqcount ->&dev->power.lock ->dpm_list_mtx ->uevent_sock_mutex ->running_helpers_waitq.lock ->&k->k_lock ->subsys mutex#30 ->(console_sem).lock ->console_owner_lock ->console_owner ->&cfs_rq->removed.lock ->input_mutex FD: 1 BD: 1 ....: pvr2_context_sync_data.lock FD: 1 BD: 15 +.+.: i2c_dev_list_lock FD: 3 BD: 8 +.+.: subsys mutex#66 ->&k->k_lock FD: 1 BD: 1 +.+.: subsys mutex#67 FD: 183 BD: 2 +.+.: dvbdev_register_lock ->(console_sem).lock ->fs_reclaim ->pool_lock#2 ->minor_rwsem ->&xa->xa_lock#12 ->&mdev->graph_mutex ->&x->wait#9 ->&obj_hash[i].lock ->&k->list_lock ->gdp_mutex ->lock ->&root->kernfs_rwsem ->bus_type_sem ->sysfs_symlink_target_lock ->&c->lock ->&zone->lock ->&____s->seqcount ->&dev->power.lock ->dpm_list_mtx ->req_lock ->&p->pi_lock ->&x->wait#11 ->&rq->__lock ->uevent_sock_mutex ->running_helpers_waitq.lock ->&k->k_lock ->subsys mutex#68 FD: 184 BD: 1 +.+.: frontend_mutex ->fs_reclaim ->pool_lock#2 ->(console_sem).lock ->dvbdev_register_lock FD: 1 BD: 3 +.+.: minor_rwsem FD: 8 BD: 3 ....: &xa->xa_lock#12 ->&c->lock ->&____s->seqcount ->pool_lock#2 FD: 126 BD: 4 +.+.: &mdev->graph_mutex ->fs_reclaim ->pool_lock#2 ->&obj_hash[i].lock FD: 3 BD: 3 +.+.: subsys mutex#68 ->&k->k_lock FD: 1 BD: 1 ....: &dmxdev->lock FD: 1 BD: 1 +.+.: &dvbdemux->mutex FD: 1 BD: 1 +.+.: media_devnode_lock FD: 1 BD: 1 +.+.: subsys mutex#69 FD: 1 BD: 1 +.+.: videodev_lock FD: 3 BD: 1 +.+.: subsys mutex#70 ->&k->k_lock FD: 1 BD: 1 +.+.: vimc_sensor:393:(&vsensor->hdl)->_lock FD: 1 BD: 1 +.+.: &v4l2_dev->lock FD: 1 BD: 1 +.+.: vimc_debayer:578:(&vdebayer->hdl)->_lock FD: 1 BD: 1 +.+.: vimc_lens:61:(&vlens->hdl)->_lock FD: 136 BD: 1 +.+.: vivid_ctrls:1606:(hdl_user_gen)->_lock ->vivid_ctrls:1620:(hdl_vid_cap)->_lock ->fs_reclaim ->pool_lock#2 ->&c->lock ->&____s->seqcount ->vivid_ctrls:1622:(hdl_vid_out)->_lock ->vivid_ctrls:1625:(hdl_vbi_cap)->_lock ->vivid_ctrls:1627:(hdl_vbi_out)->_lock ->vivid_ctrls:1630:(hdl_radio_rx)->_lock ->vivid_ctrls:1632:(hdl_radio_tx)->_lock ->&zone->lock ->vivid_ctrls:1634:(hdl_sdr_cap)->_lock ->vivid_ctrls:1636:(hdl_meta_cap)->_lock ->vivid_ctrls:1638:(hdl_meta_out)->_lock ->vivid_ctrls:1640:(hdl_tch_cap)->_lock ->&obj_hash[i].lock ->batched_entropy_u8.lock ->kfence_freelist_lock FD: 127 BD: 1 +.+.: vivid_ctrls:1608:(hdl_user_vid)->_lock ->vivid_ctrls:1620:(hdl_vid_cap)->_lock ->fs_reclaim ->pool_lock#2 ->&c->lock ->&zone->lock ->&____s->seqcount FD: 130 BD: 1 +.+.: vivid_ctrls:1610:(hdl_user_aud)->_lock ->vivid_ctrls:1620:(hdl_vid_cap)->_lock ->fs_reclaim ->pool_lock#2 ->vivid_ctrls:1622:(hdl_vid_out)->_lock ->vivid_ctrls:1630:(hdl_radio_rx)->_lock ->vivid_ctrls:1632:(hdl_radio_tx)->_lock ->&c->lock ->&____s->seqcount ->&zone->lock FD: 134 BD: 1 +.+.: vivid_ctrls:1612:(hdl_streaming)->_lock ->vivid_ctrls:1620:(hdl_vid_cap)->_lock ->fs_reclaim ->pool_lock#2 ->&c->lock ->&____s->seqcount ->vivid_ctrls:1622:(hdl_vid_out)->_lock ->vivid_ctrls:1625:(hdl_vbi_cap)->_lock ->&zone->lock ->vivid_ctrls:1627:(hdl_vbi_out)->_lock ->vivid_ctrls:1634:(hdl_sdr_cap)->_lock ->vivid_ctrls:1636:(hdl_meta_cap)->_lock ->vivid_ctrls:1638:(hdl_meta_out)->_lock ->vivid_ctrls:1640:(hdl_tch_cap)->_lock ->&obj_hash[i].lock FD: 128 BD: 1 +.+.: vivid_ctrls:1614:(hdl_sdtv_cap)->_lock ->vivid_ctrls:1620:(hdl_vid_cap)->_lock ->fs_reclaim ->pool_lock#2 ->vivid_ctrls:1625:(hdl_vbi_cap)->_lock ->&c->lock ->&zone->lock ->&____s->seqcount FD: 128 BD: 1 +.+.: vivid_ctrls:1616:(hdl_loop_cap)->_lock ->&rq->__lock ->vivid_ctrls:1620:(hdl_vid_cap)->_lock ->fs_reclaim ->pool_lock#2 ->vivid_ctrls:1625:(hdl_vbi_cap)->_lock ->&c->lock ->&zone->lock ->&____s->seqcount FD: 1 BD: 1 +.+.: vivid_ctrls:1618:(hdl_fb)->_lock FD: 1 BD: 7 +.+.: vivid_ctrls:1620:(hdl_vid_cap)->_lock FD: 1 BD: 4 +.+.: vivid_ctrls:1622:(hdl_vid_out)->_lock FD: 1 BD: 5 +.+.: vivid_ctrls:1625:(hdl_vbi_cap)->_lock FD: 1 BD: 3 +.+.: vivid_ctrls:1627:(hdl_vbi_out)->_lock FD: 1 BD: 3 +.+.: vivid_ctrls:1630:(hdl_radio_rx)->_lock FD: 1 BD: 3 +.+.: vivid_ctrls:1632:(hdl_radio_tx)->_lock FD: 1 BD: 3 +.+.: vivid_ctrls:1634:(hdl_sdr_cap)->_lock FD: 1 BD: 3 +.+.: vivid_ctrls:1636:(hdl_meta_cap)->_lock FD: 1 BD: 3 +.+.: vivid_ctrls:1638:(hdl_meta_out)->_lock FD: 1 BD: 3 +.+.: vivid_ctrls:1640:(hdl_tch_cap)->_lock FD: 1 BD: 1 ....: &adap->kthread_waitq FD: 1 BD: 1 +.+.: &dev->cec_xfers_slock FD: 1 BD: 1 +.+.: cec_devnode_lock FD: 1 BD: 1 ....: &dev->kthread_waitq_cec FD: 1 BD: 1 +.+.: subsys mutex#71 FD: 6 BD: 1 +.+.: &adap->lock ->tk_core.seq.seqcount ->&adap->devnode.lock_fhs FD: 1 BD: 2 +.+.: &adap->devnode.lock_fhs FD: 30 BD: 1 ..-.: drivers/block/floppy.c:640 FD: 35 BD: 1 +.+.: (fd_timeout).work ->&obj_hash[i].lock ->floppy_work ->dma_spin_lock ->floppy_lock ->command_done.lock FD: 1 BD: 3 ....: fdc_wait.lock FD: 1 BD: 3 ....: (&motor_off_timer[drive]) FD: 1 BD: 308 ....: (&sq->pending_timer) FD: 1 BD: 5 +.+.: (work_completion)(&td->dispatch_work) FD: 33 BD: 7 +.+.: &q->blkcg_mutex ->&q->queue_lock ->&obj_hash[i].lock ->pool_lock#2 FD: 1 BD: 1 ..-.: percpu_ref_switch_waitq.lock FD: 1 BD: 1 ....: ptp_clocks_map.xa_lock FD: 3 BD: 1 +.+.: subsys mutex#72 ->&k->k_lock FD: 1 BD: 1 +.+.: pers_lock FD: 1 BD: 1 +.+.: _lock FD: 23 BD: 3 +.+.: dm_bufio_clients_lock ->&rq->__lock FD: 1 BD: 1 +.+.: _ps_lock FD: 1 BD: 1 +.+.: _lock#2 FD: 1 BD: 1 +.+.: _lock#3 FD: 1 BD: 1 +.+.: register_lock#2 FD: 3 BD: 1 +.+.: subsys mutex#73 ->&k->k_lock FD: 1 BD: 1 .+.+: bp_lock FD: 3 BD: 1 +.+.: subsys mutex#74 ->&k->k_lock FD: 15 BD: 1 +.-.: (&dsp_spl_tl) ->dsp_lock FD: 14 BD: 2 ..-.: dsp_lock ->iclock_lock ->&obj_hash[i].lock ->&base->lock FD: 5 BD: 3 ...-: iclock_lock ->tk_core.seq.seqcount FD: 133 BD: 80 +.+.: lock#7 ->fs_reclaim ->pool_lock#2 ->&xa->xa_lock#14 ->crngs.lock ->&xa->xa_lock#19 ->&id_priv->qp_mutex ->&id_priv->lock ->&xa->xa_lock#20 ->&cm_id_priv->lock ->&c->lock ->&rq->__lock FD: 1 BD: 1 ....: iscsi_transport_lock FD: 3 BD: 1 +.+.: subsys mutex#75 ->&k->k_lock FD: 926 BD: 2 ++++: link_ops_rwsem ->fs_reclaim ->&c->lock ->pool_lock#2 ->&x->wait#9 ->&obj_hash[i].lock ->(console_sem).lock ->&pdata->netdev_lock ->ndev_hash_lock ->crypto_alg_sem ->devices_rwsem ->&rxe->usdev_lock ->rtnl_mutex ->&device->cache_lock ->rdmacg_mutex ->&k->list_lock ->gdp_mutex ->lock ->&root->kernfs_rwsem ->bus_type_sem ->sysfs_symlink_target_lock ->&dev->power.lock ->dpm_list_mtx ->subsys mutex#84 ->&zone->lock ->&____s->seqcount ->&n->list_lock ->&rq->__lock ->rcu_node_0 ->&____s->seqcount#2 ->&sem->wait_lock ->&p->pi_lock ->batched_entropy_u8.lock ->kfence_freelist_lock ->uevent_sock_mutex ->&xa->xa_lock#19 ->krc.lock ->&xa->xa_lock#17 ->remove_cache_srcu FD: 1 BD: 1 ....: &tx_task->waiting FD: 3 BD: 1 +.+.: subsys mutex#76 ->&k->k_lock FD: 1 BD: 1 +.+.: service_lock FD: 1 BD: 1 +.+.: vsock_register_mutex FD: 1 BD: 1 +.+.: comedi_drivers_list_lock FD: 3 BD: 6 +.+.: subsys mutex#77 ->&k->k_lock FD: 158 BD: 2 ++++: snd_ctl_layer_rwsem ->snd_ctl_led_mutex ->fs_reclaim ->&c->lock ->&____s->seqcount ->pool_lock#2 ->&x->wait#9 ->&obj_hash[i].lock ->&k->list_lock ->lock ->&root->kernfs_rwsem ->bus_type_sem ->&dev->power.lock ->dpm_list_mtx ->&k->k_lock ->sysfs_symlink_target_lock FD: 1 BD: 3 +.+.: snd_card_mutex FD: 1 BD: 1 +.+.: snd_ioctl_rwsem FD: 126 BD: 2 +.+.: strings ->fs_reclaim ->pool_lock#2 FD: 1 BD: 2 +.+.: register_mutex FD: 173 BD: 3 +.+.: sound_mutex ->fs_reclaim ->pool_lock#2 ->&k->list_lock ->gdp_mutex ->lock ->&root->kernfs_rwsem ->bus_type_sem ->sysfs_symlink_target_lock ->&c->lock ->&____s->seqcount ->&dev->power.lock ->dpm_list_mtx ->req_lock ->&p->pi_lock ->&x->wait#11 ->&rq->__lock ->&obj_hash[i].lock ->uevent_sock_mutex ->running_helpers_waitq.lock ->subsys mutex#77 ->&cfs_rq->removed.lock ->&zone->lock ->&k->k_lock FD: 183 BD: 1 +.+.: register_mutex#2 ->fs_reclaim ->pool_lock#2 ->sound_mutex ->&obj_hash[i].lock ->register_mutex ->&c->lock ->&zone->lock ->&____s->seqcount ->sound_oss_mutex ->strings ->&entry->access ->info_mutex FD: 175 BD: 1 +.+.: register_mutex#3 ->fs_reclaim ->pool_lock#2 ->sound_mutex ->clients_lock FD: 1 BD: 5 ....: clients_lock FD: 2 BD: 1 +.+.: &client->ports_mutex ->&client->ports_lock FD: 1 BD: 5 .+.+: &client->ports_lock FD: 176 BD: 1 +.+.: register_mutex#4 ->fs_reclaim ->pool_lock#2 ->sound_oss_mutex FD: 175 BD: 3 +.+.: sound_oss_mutex ->fs_reclaim ->pool_lock#2 ->sound_loader_lock ->&x->wait#9 ->&obj_hash[i].lock ->&zone->lock ->&____s->seqcount ->&k->list_lock ->gdp_mutex ->lock ->&root->kernfs_rwsem ->bus_type_sem ->&c->lock ->sysfs_symlink_target_lock ->&dev->power.lock ->dpm_list_mtx ->req_lock ->&p->pi_lock ->&x->wait#11 ->&rq->__lock ->uevent_sock_mutex ->running_helpers_waitq.lock ->subsys mutex#77 ->&k->k_lock FD: 1 BD: 4 +.+.: sound_loader_lock FD: 129 BD: 1 .+.+: &grp->list_mutex/1 ->clients_lock ->&client->ports_lock ->register_lock#3 ->fs_reclaim ->pool_lock#2 FD: 2 BD: 1 +.+.: &grp->list_mutex#2 ->&grp->list_lock FD: 1 BD: 2 ....: &grp->list_lock FD: 136 BD: 2 +.+.: async_lookup_work ->fs_reclaim ->pool_lock#2 ->clients_lock ->&client->ports_lock ->snd_card_mutex ->(kmod_concurrent_max).lock ->&obj_hash[i].lock ->&x->wait#17 ->&pool->lock ->&rq->__lock ->running_helpers_waitq.lock ->autoload_work ->&x->wait#10 FD: 1 BD: 2 ....: register_lock#3 FD: 4 BD: 3 +.+.: autoload_work ->&k->list_lock ->&k->k_lock FD: 161 BD: 1 ++++: &card->controls_rwsem ->&xa->xa_lock#13 ->fs_reclaim ->&card->ctl_files_rwlock ->snd_ctl_layer_rwsem ->&rq->__lock ->pool_lock#2 ->&c->lock ->&zone->lock ->&____s->seqcount FD: 9 BD: 2 +.+.: &xa->xa_lock#13 ->pool_lock#2 ->&c->lock ->&zone->lock ->&____s->seqcount FD: 1 BD: 2 ....: &card->ctl_files_rwlock FD: 1 BD: 3 +.+.: snd_ctl_led_mutex FD: 1 BD: 1 +.+.: register_mutex#5 FD: 1 BD: 81 +.+.: failover_lock FD: 8 BD: 5 +...: llc_sap_list_lock ->pool_lock#2 ->&c->lock ->&n->list_lock ->batched_entropy_u8.lock ->kfence_freelist_lock FD: 126 BD: 1 +.+.: act_id_mutex ->fs_reclaim ->pool_lock#2 FD: 1 BD: 78 ++++: act_mod_lock FD: 1 BD: 1 +.+.: ife_mod_lock FD: 1 BD: 80 +.+.: nf_connlabels_lock FD: 1 BD: 78 ++++: cls_mod_lock FD: 1 BD: 78 ++++: ematch_mod_lock FD: 897 BD: 2 +.+.: sock_diag_table_mutex ->nlk_cb_mutex-SOCK_DIAG ->&c->lock ->pool_lock#2 ->&obj_hash[i].lock ->rlock-AF_NETLINK ->&lock->wait_lock ->&p->pi_lock ->&rq->__lock ->inet_diag_table_mutex ->(kmod_concurrent_max).lock ->fs_reclaim ->&x->wait#17 ->running_helpers_waitq.lock FD: 1 BD: 1 +.+.: nfnl_subsys_acct FD: 26 BD: 1 +.+.: nfnl_subsys_queue FD: 1 BD: 1 +.+.: nfnl_subsys_ulog FD: 23 BD: 5 +.+.: nf_log_mutex ->&rq->__lock FD: 1 BD: 1 +.+.: nfnl_subsys_osf FD: 27 BD: 7 +.+.: nf_sockopt_mutex ->&rq->__lock ->nf_sockopt_mutex.wait_lock ->rcu_node_0 ->&cfs_rq->removed.lock ->&obj_hash[i].lock ->pool_lock#2 ->&rcu_state.expedited_wq FD: 149 BD: 4 +.+.: nfnl_subsys_ctnetlink ->pool_lock#2 ->&obj_hash[i].lock ->rcu_node_0 ->&rq->__lock ->&c->lock ->&n->list_lock ->batched_entropy_u8.lock ->kfence_freelist_lock ->&____s->seqcount#2 ->&____s->seqcount ->nf_conntrack_mutex ->(console_sem).lock ->nlk_cb_mutex-NETFILTER FD: 1 BD: 1 +.+.: nfnl_subsys_ctnetlink_exp FD: 1 BD: 5 +.+.: nf_ct_ecache_mutex FD: 132 BD: 1 +.+.: nfnl_subsys_cttimeout ->fs_reclaim ->pool_lock#2 ->&obj_hash[i].lock ->krc.lock ->nf_conntrack_mutex FD: 39 BD: 1 +.+.: nfnl_subsys_cthelper ->(console_sem).lock ->&rq->__lock ->&lock->wait_lock FD: 1 BD: 1 +.+.: nf_ct_helper_mutex FD: 1 BD: 1 +...: nf_conntrack_expect_lock FD: 32 BD: 12 +.+.: nf_conntrack_mutex ->&nf_conntrack_locks[i] ->&rq->__lock ->&cfs_rq->removed.lock ->&obj_hash[i].lock ->nf_conntrack_mutex.wait_lock ->&pool->lock ->rcu_node_0 ->&rcu_state.expedited_wq ->&____s->seqcount#7 ->pool_lock#2 ->&c->lock ->&n->list_lock ->&nf_conntrack_locks[i]/1 FD: 1 BD: 1 +.+.: nf_ct_nat_helpers_mutex FD: 310 BD: 1 +.+.: nfnl_subsys_nftables ->&nft_net->commit_mutex ->&rq->__lock ->&lock->wait_lock FD: 1 BD: 1 +.+.: nfnl_subsys_nftcompat FD: 1038 BD: 1 +.+.: masq_mutex ->pernet_ops_rwsem ->(inetaddr_chain).rwsem ->inet6addr_chain.lock FD: 232 BD: 11 +.+.: &xt[i].mutex ->fs_reclaim ->pool_lock#2 ->&c->lock ->&n->list_lock ->&mm->mmap_lock ->free_vmap_area_lock ->vmap_area_lock ->&____s->seqcount ->&per_cpu(xt_recseq, i) ->&obj_hash[i].lock ->purge_vmap_area_lock ->&rq->__lock ->&____s->seqcount#2 ->init_mm.page_table_lock ->rcu_node_0 ->&lock->wait_lock ->&cfs_rq->removed.lock ->remove_cache_srcu ->&zone->lock ->batched_entropy_u8.lock ->kfence_freelist_lock ->&meta->lock ->&rcu_state.expedited_wq ->pgd_lock ->stock_lock ->key ->pcpu_lock ->percpu_counters_lock ->quarantine_lock ->&base->lock ->&pgdat->kswapd_wait FD: 26 BD: 3894 +.+.: &tn->lock ->&rq->__lock ->rcu_node_0 ->&cfs_rq->removed.lock ->&obj_hash[i].lock ->pool_lock#2 ->&rcu_state.expedited_wq FD: 3 BD: 1 +.+.: subsys mutex#78 ->&k->k_lock FD: 150 BD: 5 +.+.: nfnl_subsys_ipset ->fs_reclaim ->pool_lock#2 ->&obj_hash[i].lock ->&c->lock ->rlock-AF_NETLINK ->&n->list_lock ->&rq->__lock ->free_vmap_area_lock ->vmap_area_lock ->&____s->seqcount ->&zone->lock ->init_mm.page_table_lock ->rcu_node_0 ->&lock->wait_lock ->stock_lock ->crngs.lock ->ip_set_ref_lock ->rcu_state.barrier_mutex ->&cfs_rq->removed.lock ->purge_vmap_area_lock ->(console_sem).lock ->console_owner_lock ->console_owner ->&rcu_state.expedited_wq FD: 1 BD: 1 +.+.: ip_set_type_mutex FD: 131 BD: 5 +.+.: ipvs->est_mutex ->fs_reclaim ->pool_lock#2 ->pcpu_alloc_mutex ->&c->lock ->pcpu_lock ->&obj_hash[i].lock ->&n->list_lock ->&____s->seqcount#2 ->pcpu_alloc_mutex.wait_lock ->&p->pi_lock ->&____s->seqcount ->&rq->__lock ->rcu_node_0 ->quarantine_lock FD: 1 BD: 1 +.+.: ip_vs_sched_mutex FD: 127 BD: 5 +.+.: __ip_vs_app_mutex ->fs_reclaim ->pool_lock#2 ->&c->lock ->&n->list_lock ->&obj_hash[i].lock ->&____s->seqcount ->&rq->__lock ->&____s->seqcount#2 ->__ip_vs_app_mutex.wait_lock ->&cfs_rq->removed.lock FD: 1 BD: 1 +.+.: ip_vs_pe_mutex FD: 1 BD: 1 +.+.: tunnel4_mutex FD: 1 BD: 1 +.+.: xfrm4_protocol_mutex FD: 205 BD: 4 +.+.: inet_diag_table_mutex ->&h->lhash2[i].lock ->&rq->__lock ->&hashinfo->ehash_locks[i] ->k-clock-AF_INET ->k-sk_lock-AF_INET ->k-slock-AF_INET ->&ht->lock FD: 1 BD: 1 +...: xfrm_km_lock FD: 1 BD: 1 +...: xfrm_translator_lock FD: 1 BD: 1 +.+.: xfrm6_protocol_mutex FD: 1 BD: 1 +.+.: tunnel6_mutex FD: 1 BD: 1 +.+.: xfrm_if_cb_lock FD: 1 BD: 1 +...: inetsw6_lock FD: 1 BD: 7 +.+.: &hashinfo->lock#2 FD: 18 BD: 5 +.+.: &net->ipv6.ip6addrlbl_table.lock ->&obj_hash[i].lock ->pool_lock#2 ->krc.lock FD: 220 BD: 3835 +.+.: &idev->mc_lock ->fs_reclaim ->pool_lock#2 ->&obj_hash[i].lock ->&c->lock ->&____s->seqcount ->&dev_addr_list_lock_key ->_xmit_ETHER ->&zone->lock ->batched_entropy_u32.lock ->&base->lock ->&rq->__lock ->krc.lock ->batched_entropy_u8.lock ->kfence_freelist_lock ->rcu_node_0 ->&n->list_lock ->&bridge_netdev_addr_lock_key ->&dev_addr_list_lock_key#2 ->&batadv_netdev_addr_lock_key ->&vlan_netdev_addr_lock_key ->&macvlan_netdev_addr_lock_key ->&dev_addr_list_lock_key#3 ->&bridge_netdev_addr_lock_key/1 ->&dev_addr_list_lock_key/1 ->&dev_addr_list_lock_key#2/1 ->_xmit_ETHER/1 ->&pool->lock ->remove_cache_srcu ->&batadv_netdev_addr_lock_key/1 ->&vlan_netdev_addr_lock_key/1 ->&macvlan_netdev_addr_lock_key/1 ->&dev_addr_list_lock_key#3/1 ->&macsec_netdev_addr_lock_key/1 ->&____s->seqcount#2 ->&rcu_state.expedited_wq ->&cfs_rq->removed.lock ->&dev_addr_list_lock_key#3/2 ->_xmit_IPGRE ->&vlan_netdev_addr_lock_key/2 ->&macsec_netdev_addr_lock_key#2/2 ->_xmit_ETHER/4 ->&macvlan_netdev_addr_lock_key/2 ->&lock->wait_lock FD: 18 BD: 3844 +...: &dev_addr_list_lock_key ->pool_lock#2 ->&c->lock ->&n->list_lock ->&obj_hash[i].lock ->krc.lock ->&____s->seqcount#2 ->&____s->seqcount ->batched_entropy_u8.lock ->kfence_freelist_lock ->&zone->lock FD: 44 BD: 3864 +...: _xmit_ETHER ->&c->lock ->&zone->lock ->&____s->seqcount ->pool_lock#2 ->&local->filter_lock ->&n->list_lock ->&obj_hash[i].lock ->batched_entropy_u8.lock ->kfence_freelist_lock ->(console_sem).lock ->console_owner_lock ->console_owner ->&____s->seqcount#2 ->krc.lock FD: 892 BD: 1 +.+.: (wq_completion)ipv6_addrconf ->(work_completion)(&(&net->ipv6.addr_chk_work)->work) ->(work_completion)(&(&ifa->dad_work)->work) ->&rq->__lock FD: 890 BD: 6 +.+.: (work_completion)(&(&net->ipv6.addr_chk_work)->work) ->rtnl_mutex ->rtnl_mutex.wait_lock ->&p->pi_lock ->&rq->__lock FD: 24 BD: 85 ....: &x->wait#21 ->&p->pi_lock FD: 57 BD: 3940 ++--: &ndev->lock ->&ifa->lock ->&____s->seqcount ->pool_lock#2 ->&obj_hash[i].lock ->&dir->lock#2 ->pcpu_lock ->&tb->tb6_lock ->&c->lock ->&n->list_lock ->batched_entropy_u32.lock ->&base->lock ->&____s->seqcount#2 FD: 9 BD: 1 +.+.: stp_proto_mutex ->llc_sap_list_lock FD: 1 BD: 1 ....: switchdev_notif_chain.lock FD: 23 BD: 78 ++++: (switchdev_blocking_notif_chain).rwsem ->&rq->__lock FD: 892 BD: 1 +.+.: br_ioctl_mutex ->rtnl_mutex ->rtnl_mutex.wait_lock ->&p->pi_lock ->&rq->__lock ->rcu_state.barrier_mutex ->dev_base_lock ->lweventlist_lock ->stock_lock ->&obj_hash[i].lock ->pool_lock#2 ->krc.lock ->&dir->lock#2 ->netdev_unregistering_wq.lock ->br_ioctl_mutex.wait_lock FD: 290 BD: 86 +.+.: nf_ct_proto_mutex ->defrag4_mutex ->nf_hook_mutex ->cpu_hotplug_lock ->&obj_hash[i].lock ->pool_lock#2 ->defrag6_mutex ->&rq->__lock ->rcu_node_0 ->&rcu_state.expedited_wq FD: 231 BD: 5 +.+.: ebt_mutex ->fs_reclaim ->pool_lock#2 ->&mm->mmap_lock ->&c->lock ->&____s->seqcount#2 ->&____s->seqcount ->&rq->__lock ->ebt_mutex.wait_lock ->&cfs_rq->removed.lock ->&obj_hash[i].lock ->pgd_lock ->stock_lock ->key ->pcpu_lock ->percpu_counters_lock FD: 1 BD: 1 +.+.: dsa_tag_drivers_lock FD: 1 BD: 1 +...: protocol_list_lock FD: 1 BD: 1 +...: linkfail_lock FD: 1 BD: 5 +...: rose_neigh_list_lock FD: 1 BD: 1 +.+.: proto_tab_lock#2 FD: 1 BD: 31 ++++: chan_list_lock FD: 1 BD: 5 +.+.: l2cap_sk_list.lock FD: 237 BD: 4 +.+.: sk_lock-AF_BLUETOOTH-BTPROTO_L2CAP ->slock-AF_BLUETOOTH-BTPROTO_L2CAP ->chan_list_lock ->&mm->mmap_lock ->&rq->__lock ->&ei->socket.wq.wait FD: 1 BD: 31 +...: slock-AF_BLUETOOTH-BTPROTO_L2CAP FD: 24 BD: 31 ..-.: rfcomm_wq.lock ->&p->pi_lock FD: 318 BD: 1 +.+.: rfcomm_mutex ->fs_reclaim ->stock_lock ->pool_lock#2 ->mmu_notifier_invalidate_range_start ->&sb->s_type->i_lock_key#8 ->bt_proto_lock ->sk_lock-AF_BLUETOOTH-BTPROTO_L2CAP ->slock-AF_BLUETOOTH-BTPROTO_L2CAP ->&c->lock ->&obj_hash[i].lock ->hci_dev_list_lock ->&hdev->lock ->(&s->timer) ->&rq->__lock ->&base->lock ->&d->lock ->&list->lock#33 ->rlock-AF_BLUETOOTH ->l2cap_sk_list.lock ->&chan->lock/1 ->chan_list_lock ->&xa->xa_lock#9 ->&fsnotify_mark_srcu FD: 1 BD: 1 +.+.: auth_domain_lock FD: 1 BD: 1 +.+.: registered_mechs_lock FD: 1 BD: 1 ....: atm_dev_notify_chain.lock FD: 1 BD: 1 +.+.: proto_tab_lock#3 FD: 892 BD: 1 +.+.: vlan_ioctl_mutex ->&mm->mmap_lock ->rtnl_mutex ->rtnl_mutex.wait_lock ->&p->pi_lock ->&rq->__lock ->vlan_ioctl_mutex.wait_lock ->rcu_state.barrier_mutex ->dev_base_lock ->lweventlist_lock ->pcpu_lock ->pool_lock#2 ->&dir->lock#2 ->&obj_hash[i].lock ->krc.lock ->netdev_unregistering_wq.lock ->rcu_state.barrier_mutex.wait_lock ->stock_lock FD: 1 BD: 1 +.+.: rds_info_lock FD: 139 BD: 8 ++++: rds_trans_sem ->(console_sem).lock ->&rq->__lock ->fs_reclaim ->&c->lock ->pool_lock#2 ->crngs.lock ->&id_priv->handler_mutex ->id_table_lock ->&x->wait#29 ->&obj_hash[i].lock ->&id_priv->lock FD: 1 BD: 86 ....: &id_priv->lock FD: 2 BD: 81 +.+.: &xa->xa_lock#14 ->pool_lock#2 FD: 204 BD: 90 +.+.: k-sk_lock-AF_INET6 ->k-slock-AF_INET6 ->&tcp_hashinfo.bhash[i].lock ->&h->lhash2[i].lock ->&table->hash[i].lock ->&obj_hash[i].lock ->k-clock-AF_INET6 ->&queue->rskq_lock ->&rq->__lock ->&____s->seqcount#8 ->batched_entropy_u32.lock ->clock-AF_INET6 ->&base->lock ->pool_lock#2 ->&dir->lock ->fs_reclaim ->&____s->seqcount ->tk_core.seq.seqcount ->rcu_node_0 ->&hashinfo->ehash_locks[i] ->&c->lock ->slock-AF_INET6 ->&rcu_state.expedited_wq FD: 85 BD: 93 +.-.: k-slock-AF_INET6 ->pool_lock#2 ->tk_core.seq.seqcount ->&c->lock ->&n->list_lock ->&____s->seqcount#2 ->&____s->seqcount ->kfence_freelist_lock ->&obj_hash[i].lock ->elock-AF_INET6 ->&tcp_hashinfo.bhash[i].lock ->krc.lock ->crngs.lock ->&hashinfo->ehash_locks[i] ->(&req->rsk_timer) ->&base->lock ->&queue->rskq_lock ->k-clock-AF_INET6 ->&meta->lock ->&(&bp->lock)->lock FD: 30 BD: 129 ++.-: k-clock-AF_INET6 FD: 25 BD: 125 +.-.: &tcp_hashinfo.bhash[i].lock ->&zone->lock ->&____s->seqcount ->pool_lock#2 ->&c->lock ->&tcp_hashinfo.bhash2[i].lock ->k-clock-AF_INET6 ->clock-AF_INET ->clock-AF_INET6 ->stock_lock ->&obj_hash[i].lock ->k-clock-AF_INET ->per_cpu_ptr(&cgroup_rstat_cpu_lock, cpu) ->&____s->seqcount#2 FD: 24 BD: 126 +.-.: &tcp_hashinfo.bhash2[i].lock ->&zone->lock ->&____s->seqcount ->pool_lock#2 ->&c->lock ->k-clock-AF_INET6 ->clock-AF_INET ->clock-AF_INET6 ->&obj_hash[i].lock ->batched_entropy_u8.lock ->&hashinfo->ehash_locks[i] ->stock_lock ->k-clock-AF_INET ->quarantine_lock ->per_cpu_ptr(&cgroup_rstat_cpu_lock, cpu) ->&____s->seqcount#2 FD: 2 BD: 97 +.+.: &h->lhash2[i].lock ->k-clock-AF_INET6 FD: 1 BD: 5 +...: &list->lock#4 FD: 53 BD: 93 ++..: k-clock-AF_TIPC ->&con->sub_lock ->pool_lock#2 ->&tn->nametbl_lock ->&obj_hash[i].lock ->&base->lock ->&c->lock ->&n->list_lock ->(console_sem).lock ->batched_entropy_u8.lock ->kfence_freelist_lock ->&____s->seqcount#2 ->&____s->seqcount FD: 159 BD: 89 +.+.: k-sk_lock-AF_TIPC ->k-slock-AF_TIPC ->&tn->nametbl_lock ->&obj_hash[i].lock ->k-clock-AF_TIPC ->&rq->__lock ->fs_reclaim ->pool_lock#2 ->&dir->lock ->batched_entropy_u32.lock ->k-sk_lock-AF_TIPC/1 ->quarantine_lock ->&base->lock ->&c->lock ->&list->lock#31 ->rcu_node_0 ->&zone->lock ->&____s->seqcount#2 ->&____s->seqcount ->&n->list_lock ->remove_cache_srcu ->&meta->lock ->kfence_freelist_lock ->&rcu_state.expedited_wq FD: 31 BD: 91 +...: k-slock-AF_TIPC ->&list->lock#31 ->k-clock-AF_TIPC ->pool_lock#2 ->&c->lock ->&____s->seqcount#2 ->&____s->seqcount ->&list->lock#40 ->&obj_hash[i].lock ->quarantine_lock ->&sk->sk_lock.wq#2 ->&meta->lock ->kfence_freelist_lock ->&data->lock FD: 40 BD: 93 +...: &tn->nametbl_lock ->pool_lock#2 ->&service->lock ->&c->lock ->&nt->cluster_scope_lock ->&obj_hash[i].lock ->krc.lock ->&n->list_lock ->&____s->seqcount#2 ->&____s->seqcount FD: 38 BD: 94 +...: &service->lock ->pool_lock#2 ->&obj_hash[i].lock ->krc.lock ->&c->lock ->&n->list_lock ->&sub->lock ->&____s->seqcount#2 ->&____s->seqcount FD: 23 BD: 83 +.+.: &pnettable->lock ->&rq->__lock FD: 23 BD: 83 +.+.: smc_ib_devices.mutex ->&rq->__lock ->&cfs_rq->removed.lock FD: 1 BD: 1 +.+.: smc_wr_rx_hash_lock FD: 1 BD: 1 +.+.: v9fs_trans_lock FD: 1 BD: 5 +...: &this->receive_lock FD: 1 BD: 1 +...: lowpan_nhc_lock FD: 20 BD: 1 +.-.: (&vblank->disable_timer) ->&dev->vbl_lock FD: 306 BD: 7 +.+.: ovs_mutex ->(work_completion)(&data->gc_work) ->nf_ct_proto_mutex ->&obj_hash[i].lock ->pool_lock#2 ->nf_connlabels_lock ->net_rwsem ->quarantine_lock ->&rq->__lock FD: 288 BD: 87 +.+.: defrag4_mutex ->nf_hook_mutex ->cpu_hotplug_lock ->&obj_hash[i].lock ->pool_lock#2 ->&rq->__lock FD: 288 BD: 87 +.+.: defrag6_mutex ->nf_hook_mutex ->cpu_hotplug_lock ->&obj_hash[i].lock ->pool_lock#2 ->&rq->__lock FD: 1 BD: 144 +.+.: subsys mutex#79 FD: 25 BD: 1 ..-.: &(&gc_work->dwork)->timer FD: 33 BD: 2 +.+.: (work_completion)(&(&gc_work->dwork)->work) ->rcu_node_0 ->&rq->__lock ->&cfs_rq->removed.lock ->&obj_hash[i].lock ->pool_lock#2 ->&base->lock FD: 1 BD: 3922 ...-: &____s->seqcount#7 FD: 25 BD: 1 ..-.: &(&ipvs->defense_work)->timer FD: 28 BD: 6 +.+.: (work_completion)(&(&ipvs->defense_work)->work) ->&s->s_inode_list_lock ->&ipvs->dropentry_lock ->&ipvs->droppacket_lock ->&ipvs->securetcp_lock ->&obj_hash[i].lock ->&base->lock ->&rq->__lock ->rcu_node_0 FD: 1 BD: 7 +...: &ipvs->dropentry_lock FD: 1 BD: 7 +...: &ipvs->droppacket_lock FD: 1 BD: 7 +...: &ipvs->securetcp_lock FD: 34 BD: 2 +.+.: drain_vmap_work ->vmap_purge_lock ->&rq->__lock FD: 13 BD: 5 +.-.: (&net->can.stattimer) ->&obj_hash[i].lock ->&base->lock FD: 10 BD: 249 +...: map_idr_lock ->pool_lock#2 ->&obj_hash[i].lock ->&c->lock ->&n->list_lock FD: 1 BD: 1 ....: rcu_read_lock_sched FD: 12 BD: 248 +.-.: prog_idr_lock ->pool_lock#2 ->&obj_hash[i].lock ->&c->lock ->&n->list_lock ->&____s->seqcount#2 ->&____s->seqcount ->batched_entropy_u8.lock ->kfence_freelist_lock FD: 1 BD: 84 +.-.: bpf_lock FD: 1 BD: 1 ....: rcu_read_lock_trace FD: 10 BD: 248 +...: btf_idr_lock ->pool_lock#2 ->&obj_hash[i].lock ->&c->lock ->&n->list_lock FD: 1 BD: 1 +.+.: &map->freeze_mutex FD: 1 BD: 6 +.+.: ima_keys_lock FD: 129 BD: 143 +.+.: scomp_lock ->fs_reclaim ->pool_lock#2 ->free_vmap_area_lock ->vmap_area_lock ->&zone->lock ->&____s->seqcount ->init_mm.page_table_lock ->&c->lock FD: 25 BD: 1 +.+.: pcpu_drain_mutex ->&pcp->lock ->&rq->__lock FD: 386 BD: 5 +.+.: k-sk_lock-AF_RXRPC ->k-slock-AF_RXRPC ->&rxnet->local_mutex ->&local->services_lock ->fs_reclaim ->&c->lock ->&zone->lock ->&____s->seqcount ->pool_lock#2 ->&rx->incoming_lock ->&obj_hash[i].lock ->&rq->__lock ->&rxnet->conn_lock ->&call->waitq ->(rxrpc_call_limiter).lock ->&rx->recvmsg_lock ->&rx->call_lock ->&rxnet->call_lock ->(&call->timer) ->&base->lock ->&list->lock#22 ->&meta->lock ->kfence_freelist_lock ->rcu_node_0 ->&n->list_lock ->quarantine_lock ->remove_cache_srcu ->batched_entropy_u8.lock ->&rcu_state.expedited_wq FD: 1 BD: 6 +...: k-slock-AF_RXRPC FD: 373 BD: 7 +.+.: &rxnet->local_mutex ->fs_reclaim ->pool_lock#2 ->&obj_hash[i].lock ->crngs.lock ->mmu_notifier_invalidate_range_start ->&sb->s_type->i_lock_key#8 ->&zone->lock ->&____s->seqcount ->&c->lock ->&dir->lock ->k-sk_lock-AF_INET6 ->k-slock-AF_INET6 ->cpu_hotplug_lock ->&rq->__lock ->kthread_create_lock ->&p->pi_lock ->&x->wait ->&x->wait#22 ->stock_lock ->&____s->seqcount#2 ->&n->list_lock ->&cfs_rq->removed.lock ->rcu_node_0 ->&table->hash[i].lock ->k-clock-AF_INET6 ->&xa->xa_lock#9 ->&fsnotify_mark_srcu ->quarantine_lock ->batched_entropy_u8.lock ->kfence_freelist_lock ->k-sk_lock-AF_INET ->k-slock-AF_INET ->remove_cache_srcu ->&lock->wait_lock ->k-clock-AF_INET FD: 18 BD: 98 +...: &table->hash[i].lock ->k-clock-AF_INET6 ->&table->hash2[i].lock ->k-clock-AF_INET ->clock-AF_INET ->clock-AF_INET6 FD: 1 BD: 99 +...: &table->hash2[i].lock FD: 287 BD: 2 +.+.: netstamp_work ->cpu_hotplug_lock FD: 24 BD: 8 ....: &x->wait#22 ->&p->pi_lock FD: 1 BD: 9 +.+.: &local->services_lock FD: 1 BD: 10 +.+.: &rxnet->conn_lock FD: 1 BD: 6 ....: &call->waitq FD: 1 BD: 6 +.+.: &rx->call_lock FD: 1 BD: 6 +.+.: &rxnet->call_lock FD: 30 BD: 5 +.-.: (&rxnet->peer_keepalive_timer) FD: 134 BD: 1 +.+.: init_user_ns.keyring_sem ->key_user_lock ->root_key_user.lock ->fs_reclaim ->pool_lock#2 ->crngs.lock ->key_serial_lock ->key_construction_mutex ->&type->lock_class ->keyring_serialise_link_lock FD: 1 BD: 5 +.+.: root_key_user.lock FD: 27 BD: 7 +.+.: (wq_completion)krxrpcd ->(work_completion)(&rxnet->peer_keepalive_work) ->(work_completion)(&rxnet->service_conn_reaper) ->&rq->__lock FD: 24 BD: 8 +.+.: (work_completion)(&rxnet->peer_keepalive_work) ->&rxnet->peer_hash_lock ->&obj_hash[i].lock ->&base->lock ->&rq->__lock FD: 1 BD: 9 +.+.: &rxnet->peer_hash_lock FD: 1 BD: 6 +.+.: keyring_name_lock FD: 1 BD: 1 +.+.: template_list FD: 1 BD: 1 +.+.: idr_lock FD: 127 BD: 9 +.+.: ima_extend_list_mutex ->fs_reclaim ->&c->lock ->&zone->lock ->&____s->seqcount ->pool_lock#2 ->&n->list_lock ->&rq->__lock ->rcu_node_0 ->ima_extend_list_mutex.wait_lock ->remove_cache_srcu ->&____s->seqcount#2 ->&cfs_rq->removed.lock ->&obj_hash[i].lock FD: 1 BD: 1 +.+.: clk_debug_lock FD: 24 BD: 4 +.+.: deferred_probe_work ->deferred_probe_mutex FD: 129 BD: 78 ++++: &(&net->nexthop.notifier_chain)->rwsem ->&data->nh_lock FD: 210 BD: 89 +.+.: k-sk_lock-AF_INET ->k-slock-AF_INET ->&table->hash[i].lock ->&obj_hash[i].lock ->k-clock-AF_INET ->&tcp_hashinfo.bhash[i].lock ->&____s->seqcount ->pool_lock#2 ->&c->lock ->stock_lock ->&rq->__lock ->per_cpu_ptr(&cgroup_rstat_cpu_lock, cpu) ->&hashinfo->ehash_locks[i] ->batched_entropy_u32.lock ->tk_core.seq.seqcount ->batched_entropy_u16.lock ->fs_reclaim ->&base->lock ->slock-AF_INET ->batched_entropy_u8.lock ->kfence_freelist_lock ->&meta->lock ->&____s->seqcount#2 ->(&tw->tw_timer) ->&n->list_lock ->&____s->seqcount#8 ->&ei->socket.wq.wait ->crngs.lock ->&token_hash[i].lock ->&in_dev->mc_tomb_lock ->&im->lock ->_xmit_ETHER ->&data->lock FD: 73 BD: 91 +.-.: k-slock-AF_INET ->pool_lock#2 ->&c->lock ->batched_entropy_u32.lock ->&obj_hash[i].lock ->&____s->seqcount#2 ->&____s->seqcount ->elock-AF_INET ->krc.lock ->&tcp_hashinfo.bhash[i].lock ->&n->list_lock ->key#24 ->&base->lock FD: 31 BD: 128 ++..: k-clock-AF_INET FD: 890 BD: 2 +.+.: reg_work ->rtnl_mutex ->rtnl_mutex.wait_lock ->&p->pi_lock ->&rq->__lock FD: 1 BD: 1 +.+.: detector_work FD: 1 BD: 78 +...: reg_pending_beacons_lock FD: 1 BD: 1 +.+.: acpi_gpio_deferred_req_irqs_lock FD: 904 BD: 2 +.+.: (work_completion)(&fw_work->work) ->fs_reclaim ->pool_lock#2 ->&fw_cache.lock ->&c->lock ->&zone->lock ->&____s->seqcount ->tk_core.seq.seqcount ->async_lock ->init_task.alloc_lock ->&obj_hash[i].lock ->&dentry->d_lock ->&sb->s_type->i_mutex_key ->&base->lock ->(console_sem).lock ->console_owner_lock ->console_owner ->umhelper_sem ->fw_lock ->&rq->__lock ->rtnl_mutex ->rtnl_mutex.wait_lock ->&p->pi_lock FD: 9 BD: 4 +.+.: &fw_cache.lock ->&c->lock ->&zone->lock ->&____s->seqcount ->pool_lock#2 FD: 1 BD: 1 +.+.: prepare_lock FD: 3 BD: 5 +.+.: subsys mutex#80 ->&k->k_lock FD: 26 BD: 14 +.+.: fw_lock ->&x->wait#23 ->&rq->__lock ->fw_lock.wait_lock FD: 24 BD: 15 ....: &x->wait#23 ->&p->pi_lock FD: 25 BD: 1 ..-.: fs/file_table.c:431 FD: 4 BD: 2 +.+.: (delayed_fput_work).work ->&obj_hash[i].lock ->pool_lock#2 FD: 1 BD: 1 +.+.: cdev_lock FD: 369 BD: 2 +.+.: &tty->legacy_mutex ->&tty->read_wait ->&tty->write_wait ->&tty->ldisc_sem ->&tty->files_lock ->&port->lock ->&port->mutex ->&port_lock_key ->tasklist_lock ->&tty->ctrl.lock ->&f->f_lock ->&obj_hash[i].lock ->pool_lock#2 FD: 1 BD: 7 ....: &tty->read_wait FD: 24 BD: 4149 -.-.: &tty->write_wait ->&p->pi_lock FD: 353 BD: 3 ++++: &tty->ldisc_sem ->fs_reclaim ->pool_lock#2 ->free_vmap_area_lock ->vmap_area_lock ->&____s->seqcount ->&zone->lock ->&tty->write_wait ->&tty->read_wait ->&tty->termios_rwsem ->&mm->mmap_lock ->&port_lock_key ->&port->lock ->&tty->flow.lock ->&ldata->atomic_read_lock FD: 261 BD: 6 ++++: &tty->termios_rwsem ->&port->mutex ->&tty->write_wait ->&tty->read_wait ->&ldata->output_lock ->&port_lock_key FD: 1 BD: 5 +.+.: &tty->files_lock FD: 1 BD: 4149 -.-.: &port->lock FD: 126 BD: 10 +.+.: hash_mutex ->fs_reclaim ->pool_lock#2 FD: 35 BD: 10 -.-.: &i->lock ->&port_lock_key FD: 1 BD: 1 +.+.: detected_devices_mutex FD: 24 BD: 4599 ....: &wq#2 ->&p->pi_lock FD: 136 BD: 1 +.+.: &type->s_umount_key#25/1 ->fs_reclaim ->pool_lock#2 ->pcpu_alloc_mutex ->shrinker_rwsem ->list_lrus_mutex ->sb_lock FD: 1 BD: 7 +.+.: &bdev->bd_holder_lock FD: 6 BD: 1 +.+.: &bdev->bd_fsfreeze_mutex ->sb_lock FD: 154 BD: 1 +.+.: &type->s_umount_key#26 ->fs_reclaim ->&c->lock ->&zone->lock ->&____s->seqcount ->pool_lock#2 ->&obj_hash[i].lock ->pcpu_alloc_mutex ->&wq->mutex ->kthread_create_lock ->&p->pi_lock ->&rq->__lock ->&x->wait ->wq_pool_mutex ->mmu_notifier_invalidate_range_start ->&xa->xa_lock#9 ->lock#4 ->&mapping->private_lock ->tk_core.seq.seqcount ->&dd->lock ->bit_wait_table + i ->wq_mayday_lock ->&sbi->old_work_lock ->(work_completion)(&(&sbi->old_work)->work) ->shrinker_rwsem FD: 24 BD: 4556 ..-.: bit_wait_table + i ->&p->pi_lock FD: 28 BD: 2 +.+.: (work_completion)(&s->destroy_work) ->&rsp->gp_wait ->pcpu_lock ->&obj_hash[i].lock ->quarantine_lock ->&base->lock FD: 1 BD: 2 +.+.: &sbi->old_work_lock FD: 1 BD: 2 +.+.: (work_completion)(&(&sbi->old_work)->work) FD: 136 BD: 1 +.+.: &type->s_umount_key#27/1 ->fs_reclaim ->pool_lock#2 ->pcpu_alloc_mutex ->shrinker_rwsem ->list_lrus_mutex ->sb_lock FD: 138 BD: 1 +.+.: &type->s_umount_key#28 ->fs_reclaim ->pool_lock#2 ->mmu_notifier_invalidate_range_start ->&____s->seqcount ->&xa->xa_lock#9 ->lock#4 ->&mapping->private_lock ->tk_core.seq.seqcount ->&dd->lock ->&obj_hash[i].lock ->bit_wait_table + i ->&rq->__lock ->&sb->s_type->i_lock_key#3 ->lock#5 ->&lruvec->lru_lock ->&zone->lock ->crypto_alg_sem ->lock#3 ->shrinker_rwsem FD: 136 BD: 1 +.+.: &type->s_umount_key#29/1 ->fs_reclaim ->pcpu_alloc_mutex ->shrinker_rwsem ->list_lrus_mutex ->sb_lock FD: 138 BD: 1 +.+.: &type->s_umount_key#30 ->fs_reclaim ->mmu_notifier_invalidate_range_start ->&____s->seqcount ->&xa->xa_lock#9 ->lock#4 ->&mapping->private_lock ->tk_core.seq.seqcount ->&dd->lock ->&obj_hash[i].lock ->bit_wait_table + i ->&rq->__lock ->&sb->s_type->i_lock_key#3 ->lock#5 ->&lruvec->lru_lock ->&zone->lock ->crypto_alg_sem ->lock#3 ->shrinker_rwsem FD: 136 BD: 1 +.+.: &type->s_umount_key#31/1 ->fs_reclaim ->pcpu_alloc_mutex ->shrinker_rwsem ->list_lrus_mutex ->sb_lock FD: 386 BD: 1 ++++: &type->s_umount_key#32 ->fs_reclaim ->mmu_notifier_invalidate_range_start ->&____s->seqcount ->&xa->xa_lock#9 ->lock#4 ->&mapping->private_lock ->tk_core.seq.seqcount ->&dd->lock ->&obj_hash[i].lock ->bit_wait_table + i ->&rq->__lock ->&sb->s_type->i_lock_key#3 ->lock#5 ->&lruvec->lru_lock ->&zone->lock ->&c->lock ->pool_lock#2 ->crypto_alg_sem ->pcpu_alloc_mutex ->percpu_counters_lock ->shrinker_rwsem ->inode_hash_lock ->&sb->s_type->i_lock_key#22 ->&sb->s_type->i_mutex_key#8 ->proc_subdir_lock ->proc_inum_ida.xa_lock ->&journal->j_state_lock ->kthread_create_lock ->&p->pi_lock ->&x->wait ->&journal->j_wait_done_commit ->&p->alloc_lock ->cpu_hotplug_lock ->wq_pool_mutex ->&ei->i_es_lock ->ext4_grpinfo_slab_create_mutex ->&s->s_inode_list_lock ->ext4_li_mtx ->lock ->&root->kernfs_rwsem ->(console_sem).lock ->&dentry->d_lock ->&lru->node[i].lock ->(work_completion)(&sbi->s_error_work) ->key#3 ->key#4 ->&sbi->s_error_lock ->&base->lock ->&fq->mq_flush_lock ->mount_lock ->&eli->li_list_mtx ->&wb->list_lock ->&sbi->s_writepages_rwsem ->rcu_node_0 ->&bdi->wb_waitq ->&fsnotify_mark_srcu ->&cfs_rq->removed.lock ->&ei->i_prealloc_lock ->integrity_iint_lock ->&journal->j_list_lock ->&rcu_state.expedited_wq FD: 31 BD: 172 +.+.: &bgl->locks[i].lock ->&sbi->s_md_lock ->&obj_hash[i].lock ->pool_lock#2 ->&ei->i_prealloc_lock ->&meta->lock ->kfence_freelist_lock ->&pa->pa_lock#2 ->&pa->pa_lock ->&lg->lg_prealloc_lock ->quarantine_lock ->&____s->seqcount FD: 56 BD: 4544 +.+.: &sb->s_type->i_lock_key#22 ->&dentry->d_lock ->&lru->node[i].lock ->&xa->xa_lock#9 ->bit_wait_table + i FD: 261 BD: 7 ++++: &sb->s_type->i_mutex_key#8 ->&ei->i_es_lock ->&ei->i_data_sem ->mmu_notifier_invalidate_range_start ->&zone->lock ->&____s->seqcount ->pool_lock#2 ->&c->lock ->integrity_iint_lock ->remove_cache_srcu ->&ei->xattr_sem ->tk_core.seq.seqcount ->fs_reclaim ->&xa->xa_lock#9 ->lock#4 ->&mapping->private_lock ->&sb->s_type->i_lock_key#22 ->&wb->list_lock ->&journal->j_state_lock ->jbd2_handle ->&obj_hash[i].lock ->per_cpu_ptr(&cgroup_rstat_cpu_lock, cpu) ->&rq->__lock ->batched_entropy_u8.lock ->kfence_freelist_lock ->&meta->lock ->rcu_node_0 ->&rcu_state.expedited_wq ->mapping.invalidate_lock ->free_vmap_area_lock ->vmap_area_lock ->init_mm.page_table_lock ->swap_cgroup_mutex ->&base->lock ->&fq->mq_flush_lock ->&x->wait#26 ->(&timer.timer) ->swapon_mutex ->proc_poll_wait.lock ->&dentry->d_lock ->&sbi->s_writepages_rwsem ->&sem->waiters ->&rsp->gp_wait ->stock_lock ->&____s->seqcount#2 ->&sem->wait_lock ->&p->pi_lock ->&mm->mmap_lock ->ima_extend_list_mutex ->&p->alloc_lock ->&list->lock ->kauditd_wait.lock ->&n->list_lock ->&folio_wait_table[i] ->lock#5 ->&dd->lock ->&cfs_rq->removed.lock ->quarantine_lock ->pool_lock ->&journal->j_wait_commit ->&journal->j_wait_done_commit ->&lruvec->lru_lock ->pgd_lock ->key ->pcpu_lock ->percpu_counters_lock ->key#3 ->key#14 ->&journal->j_wait_transaction_locked ->&sb->s_type->i_mutex_key#8/4 ->bit_wait_table + i ->&mapping->i_mmap_rwsem ->&journal->j_list_lock ->lock#10 ->batched_entropy_u32.lock FD: 32 BD: 4530 ++++: &ei->i_es_lock ->&zone->lock ->&____s->seqcount ->pool_lock#2 ->&c->lock ->&sbi->s_es_lock ->&obj_hash[i].lock ->key#2 ->key#5 ->key#6 ->key#7 ->&____s->seqcount#2 ->batched_entropy_u8.lock ->kfence_freelist_lock ->&meta->lock ->&n->list_lock ->&base->lock ->quarantine_lock FD: 117 BD: 172 ++++: &ei->i_data_sem ->mmu_notifier_invalidate_range_start ->pool_lock#2 ->&ei->i_es_lock ->&obj_hash[i].lock ->&c->lock ->&____s->seqcount ->&ei->i_prealloc_lock ->&rq->__lock ->&n->list_lock ->&sb->s_type->i_lock_key#22 ->&(ei->i_block_reservation_lock) ->&ei->i_raw_lock ->&wb->list_lock ->&mapping->private_lock ->&ret->b_state_lock ->&journal->j_revoke_lock ->key#14 ->&sbi->s_md_lock ->key#3 ->&lg->lg_mutex ->bit_wait_table + i ->&dd->lock ->batched_entropy_u8.lock ->kfence_freelist_lock ->&meta->lock ->&wb->work_lock ->rcu_node_0 ->&____s->seqcount#2 ->&journal->j_wait_updates ->&pa->pa_lock#2 ->remove_cache_srcu ->&bgl->locks[i].lock ->&rcu_state.expedited_wq ->stock_lock ->&xa->xa_lock#9 ->lock#4 ->&cfs_rq->removed.lock ->&sem->wait_lock ->quarantine_lock ->pgd_lock ->key ->pcpu_lock ->percpu_counters_lock ->&journal->j_state_lock ->&ei->i_data_sem/1 ->&base->lock FD: 1 BD: 4531 +.+.: &sbi->s_es_lock FD: 76 BD: 173 ++++: &journal->j_state_lock ->&journal->j_wait_done_commit ->&journal->j_wait_commit ->tk_core.seq.seqcount ->&obj_hash[i].lock ->&base->lock ->&journal->j_wait_updates ->&journal->j_wait_transaction_locked ->&journal->j_list_lock ->&journal->j_wait_reserved FD: 24 BD: 174 ....: &journal->j_wait_done_commit ->&p->pi_lock FD: 24 BD: 174 ....: &journal->j_wait_commit ->&p->pi_lock FD: 154 BD: 2 +.+.: ext4_grpinfo_slab_create_mutex ->slab_mutex FD: 130 BD: 2 +.+.: ext4_li_mtx ->fs_reclaim ->pool_lock#2 ->batched_entropy_u16.lock ->&eli->li_list_mtx ->kthread_create_lock ->&p->pi_lock ->&x->wait ->&rq->__lock ->&obj_hash[i].lock FD: 1 BD: 1 ....: &rs->lock FD: 182 BD: 6 ++++: &type->i_mutex_dir_key#3 ->fs_reclaim ->pool_lock#2 ->&dentry->d_lock ->rename_lock.seqcount ->&ei->i_es_lock ->&ei->i_data_sem ->mmu_notifier_invalidate_range_start ->&____s->seqcount ->&xa->xa_lock#9 ->lock#4 ->&mapping->private_lock ->tk_core.seq.seqcount ->&dd->lock ->&obj_hash[i].lock ->bit_wait_table + i ->&rq->__lock ->inode_hash_lock ->&journal->j_state_lock ->&sb->s_type->i_lock_key#22 ->namespace_sem ->&zone->lock ->&c->lock ->pgd_lock ->key ->pcpu_lock ->percpu_counters_lock ->tomoyo_ss ->&s->s_inode_list_lock ->&ei->xattr_sem ->jbd2_handle ->batched_entropy_u8.lock ->kfence_freelist_lock ->&n->list_lock ->remove_cache_srcu ->&____s->seqcount#2 ->stock_lock ->rcu_node_0 ->&cfs_rq->removed.lock ->&meta->lock ->&sem->wait_lock ->&rcu_state.gp_wq ->per_cpu_ptr(&cgroup_rstat_cpu_lock, cpu) ->&journal->j_wait_transaction_locked ->&rcu_state.expedited_wq FD: 37 BD: 87 +.+.: rcu_state.barrier_mutex ->rcu_state.barrier_lock ->&x->wait#24 ->&rq->__lock ->rcu_state.barrier_mutex.wait_lock ->&pool->lock ->&cfs_rq->removed.lock ->&obj_hash[i].lock ->pool_lock#2 ->rcu_node_0 ->&rcu_state.expedited_wq ->pgd_lock ->stock_lock ->key ->pcpu_lock ->percpu_counters_lock ->pool_lock FD: 24 BD: 88 ..-.: &x->wait#24 ->&p->pi_lock FD: 23 BD: 1 +.+.: (init_mm).mmap_lock ->&rq->__lock FD: 155 BD: 1 +.+.: &type->s_umount_key#33/1 ->fs_reclaim ->pool_lock#2 ->pcpu_alloc_mutex ->shrinker_rwsem ->list_lrus_mutex ->sb_lock ->&zone->lock ->&____s->seqcount ->&c->lock ->mmu_notifier_invalidate_range_start ->&sb->s_type->i_lock_key#23 ->&s->s_inode_list_lock ->tk_core.seq.seqcount ->&sb->s_type->i_mutex_key#9 ->&dentry->d_lock FD: 41 BD: 4540 +.+.: &sb->s_type->i_lock_key#23 ->&dentry->d_lock ->bit_wait_table + i ->&dentry->d_lock/1 FD: 149 BD: 4 ++++: &sb->s_type->i_mutex_key#9 ->fs_reclaim ->pool_lock#2 ->&dentry->d_lock ->mmu_notifier_invalidate_range_start ->&zone->lock ->&____s->seqcount ->&obj_hash[i].lock ->&sb->s_type->i_lock_key#23 ->&s->s_inode_list_lock ->tk_core.seq.seqcount ->rename_lock.seqcount ->proc_subdir_lock ->sysctl_lock ->&c->lock ->&p->alloc_lock ->&pid->lock ->namespace_sem ->tomoyo_ss ->&n->list_lock ->&rq->__lock ->&xa->xa_lock#4 ->stock_lock ->&____s->seqcount#2 ->rcu_node_0 ->remove_cache_srcu ->&rcu_state.gp_wq ->batched_entropy_u8.lock ->kfence_freelist_lock ->&cfs_rq->removed.lock ->&rcu_state.expedited_wq ->per_cpu_ptr(&cgroup_rstat_cpu_lock, cpu) ->pgd_lock ->key ->pcpu_lock ->percpu_counters_lock FD: 252 BD: 2 .+.+: sb_writers#3 ->mount_lock ->&sb->s_type->i_mutex_key#9 ->sysctl_lock ->fs_reclaim ->pool_lock#2 ->&obj_hash[i].lock ->&h->resize_lock ->hugetlb_lock ->tk_core.seq.seqcount ->&sb->s_type->i_lock_key#23 ->&wb->list_lock ->&dentry->d_lock ->tomoyo_ss ->&mm->mmap_lock ->oom_adj_mutex ->&p->pi_lock ->&rq->__lock ->&c->lock ->&____s->seqcount#11 ->&(&net->ipv4.ping_group_range.lock)->lock ->rcu_node_0 ->&rcu_state.expedited_wq ->oom_adj_mutex.wait_lock ->&cfs_rq->removed.lock ->(console_sem).lock FD: 128 BD: 3 +.+.: &h->resize_lock ->free_hpage_work ->hugetlb_lock ->fs_reclaim ->&____s->seqcount ->pool_lock#2 FD: 1 BD: 4 +.+.: free_hpage_work FD: 2 BD: 157 ....: hugetlb_lock ->&____s->seqcount#2 FD: 25 BD: 1 ..-.: &(&ovs_net->masks_rebalance)->timer FD: 307 BD: 6 +.+.: (work_completion)(&(&ovs_net->masks_rebalance)->work) ->ovs_mutex ->&obj_hash[i].lock ->&base->lock ->&rq->__lock FD: 189 BD: 159 ++++: mapping.invalidate_lock ->mmu_notifier_invalidate_range_start ->&zone->lock ->&____s->seqcount ->pool_lock#2 ->&xa->xa_lock#9 ->lock#4 ->&ei->i_es_lock ->&ei->i_data_sem ->tk_core.seq.seqcount ->&dd->lock ->&obj_hash[i].lock ->&c->lock ->batched_entropy_u8.lock ->kfence_freelist_lock ->&base->lock ->&rq->__lock ->rcu_node_0 ->&mapping->i_mmap_rwsem ->&journal->j_state_lock ->jbd2_handle ->&mapping->private_lock ->stock_lock ->&sb->s_type->i_lock_key#22 ->lock#5 ->&lruvec->lru_lock ->&cfs_rq->removed.lock ->&rcu_state.expedited_wq ->remove_cache_srcu ->fs_reclaim ->&____s->seqcount#2 ->&folio_wait_table[i] ->per_cpu_ptr(&cgroup_rstat_cpu_lock, cpu) ->&sem->wait_lock ->pgd_lock ->key ->pcpu_lock ->percpu_counters_lock ->quarantine_lock ->&meta->lock ->&journal->j_wait_transaction_locked ->&n->list_lock ->&journal->j_wait_commit ->&journal->j_wait_done_commit ->&p->pi_lock ->&sbi->s_writepages_rwsem ->&sem->waiters ->&rsp->gp_wait ->batched_entropy_u32.lock FD: 1 BD: 4518 ++++: integrity_iint_lock FD: 204 BD: 4 +.+.: &iint->mutex ->&ei->xattr_sem ->fs_reclaim ->pool_lock#2 ->&obj_hash[i].lock ->mmu_notifier_invalidate_range_start ->ima_extend_list_mutex ->mapping.invalidate_lock ->&folio_wait_table[i] ->&rq->__lock ->batched_entropy_u8.lock ->kfence_freelist_lock ->&c->lock ->&____s->seqcount ->tk_core.seq.seqcount ->quarantine_lock ->&lock->wait_lock ->&n->list_lock ->remove_cache_srcu ->pgd_lock ->key ->pcpu_lock ->percpu_counters_lock ->&____s->seqcount#2 ->rcu_node_0 ->&p->alloc_lock ->&list->lock ->kauditd_wait.lock ->&cfs_rq->removed.lock ->ima_extend_list_mutex.wait_lock ->&p->pi_lock ->&lruvec->lru_lock ->&sem->wait_lock ->&rcu_state.expedited_wq ->&meta->lock FD: 59 BD: 11 .+.+: &ei->xattr_sem ->&mapping->private_lock ->&rq->__lock ->&cfs_rq->removed.lock ->&obj_hash[i].lock ->mmu_notifier_invalidate_range_start ->&____s->seqcount ->pool_lock#2 ->&xa->xa_lock#9 ->lock#4 ->tk_core.seq.seqcount ->&dd->lock ->&c->lock ->&n->list_lock ->batched_entropy_u8.lock ->kfence_freelist_lock ->bit_wait_table + i FD: 1 BD: 4 ++++: entries_lock FD: 245 BD: 2 +.+.: &sig->exec_update_lock ->&p->alloc_lock ->&sighand->siglock ->&newf->file_lock ->batched_entropy_u64.lock ->&mm->mmap_lock ->delayed_uprobe_lock ->&memcg->mm_list.lock ->pgd_lock ->pool_lock#2 ->&obj_hash[i].lock ->key ->pcpu_lock ->percpu_counters_lock ->quarantine_lock ->&rq->__lock ->pool_lock ->&sem->wait_lock ->&p->pi_lock ->stock_lock FD: 1 BD: 4518 +.+.: &memcg->mm_list.lock FD: 3 BD: 3925 ..-.: batched_entropy_u16.lock ->crngs.lock FD: 24 BD: 4529 +.+.: ptlock_ptr(page)#2/1 FD: 124 BD: 1 ++++: &type->s_umount_key#34 ->shrinker_rwsem ->&dentry->d_lock ->rename_lock.seqcount ->&dentry->d_lock/1 ->&sb->s_type->i_lock_key#23 ->&s->s_inode_list_lock ->&xa->xa_lock#9 ->sysctl_lock ->&obj_hash[i].lock ->pool_lock#2 ->&fsnotify_mark_srcu ->&lru->node[i].lock ->&rq->__lock ->&pid->lock FD: 903 BD: 2 +.+.: (work_completion)(&map->work) ->&obj_hash[i].lock ->pool_lock#2 ->cgroup_mutex ->stock_lock ->vmap_area_lock ->purge_vmap_area_lock ->per_cpu_ptr(&cgroup_rstat_cpu_lock, cpu) ->pcpu_lock ->rcu_node_0 ->&rq->__lock ->&rnp->exp_wq[0] ->&rnp->exp_lock ->rcu_state.exp_mutex ->rcu_state.exp_mutex.wait_lock ->&p->pi_lock ->&rnp->exp_wq[1] ->callchain_mutex ->&htab->buckets[i].lock ->&cfs_rq->removed.lock ->percpu_counters_lock ->map_idr_lock ->dev_map_lock ->rcu_state.barrier_mutex ->quarantine_lock ->&rcu_state.expedited_wq FD: 1 BD: 4531 ....: key#2 FD: 890 BD: 2 +.+.: (work_completion)(&aux->work) ->map_idr_lock ->&obj_hash[i].lock ->pool_lock#2 ->pack_mutex ->pcpu_lock ->vmap_area_lock ->purge_vmap_area_lock ->stock_lock ->rtnl_mutex ->rtnl_mutex.wait_lock ->&p->pi_lock ->&base->lock ->per_cpu_ptr(&cgroup_rstat_cpu_lock, cpu) ->&rq->__lock ->rcu_node_0 ->&rcu_state.expedited_wq ->quarantine_lock ->&meta->lock ->kfence_freelist_lock ->pack_mutex.wait_lock FD: 937 BD: 3 +.+.: &p->lock ->fs_reclaim ->pool_lock#2 ->&mm->mmap_lock ->file_systems_lock ->namespace_sem ->&c->lock ->&____s->seqcount ->&of->mutex ->&n->list_lock ->remove_cache_srcu ->rcu_node_0 ->&rq->__lock ->cpufreq_driver_lock ->module_mutex ->&____s->seqcount#2 ->&cfs_rq->removed.lock ->&obj_hash[i].lock ->batched_entropy_u8.lock ->kfence_freelist_lock ->&rcu_state.expedited_wq ->stock_lock ->per_cpu_ptr(&cgroup_rstat_cpu_lock, cpu) ->cgroup_mutex FD: 141 BD: 1 +.+.: &type->s_umount_key#35/1 ->fs_reclaim ->pool_lock#2 ->pcpu_alloc_mutex ->shrinker_rwsem ->list_lrus_mutex ->&c->lock ->&____s->seqcount ->sb_lock ->&root->kernfs_rwsem ->&sb->s_type->i_lock_key#24 ->&root->kernfs_supers_rwsem ->&dentry->d_lock FD: 40 BD: 4544 +.+.: &sb->s_type->i_lock_key#24 ->&dentry->d_lock FD: 144 BD: 3 ++++: &type->i_mutex_dir_key#4 ->fs_reclaim ->pool_lock#2 ->&dentry->d_lock ->rename_lock.seqcount ->&root->kernfs_rwsem ->&sb->s_type->i_lock_key#24 ->&____s->seqcount ->&obj_hash[i].lock ->namespace_sem ->tk_core.seq.seqcount ->&c->lock ->remove_cache_srcu ->&n->list_lock ->&rq->__lock ->rcu_node_0 ->&sem->wait_lock ->&p->pi_lock ->&____s->seqcount#2 ->&cfs_rq->removed.lock ->rename_lock FD: 24 BD: 219 ....: &x->wait#25 ->&p->pi_lock FD: 41 BD: 11 +.+.: &net->unx.table.locks[i] ->&net->unx.table.locks[i]/1 FD: 1067 BD: 2 +.+.: &sb->s_type->i_mutex_key#10 ->&net->unx.table.locks[i] ->&u->lock ->&u->peer_wait ->rlock-AF_UNIX ->pool_lock#2 ->&dir->lock ->&obj_hash[i].lock ->quarantine_lock ->nl_table_lock ->nl_table_wait.lock ->clock-AF_NETLINK ->genl_sk_destructing_waitq.lock ->&nlk->wait ->wlock-AF_NETLINK ->(netlink_chain).rwsem ->tomoyo_ss ->tk_core.seq.seqcount ->&sb->s_type->i_lock_key#8 ->&wb->list_lock ->&dentry->d_lock ->rcu_node_0 ->&rq->__lock ->sk_lock-AF_INET ->slock-AF_INET ->clock-AF_INET ->sk_lock-AF_INET6 ->slock-AF_INET6 ->clock-AF_INET6 ->&table->hash[i].lock ->&net->packet.sklist_lock ->&po->bind_lock ->sk_lock-AF_PACKET ->slock-AF_PACKET ->fanout_mutex ->&rnp->exp_wq[0] ->clock-AF_PACKET ->rlock-AF_PACKET ->pcpu_lock ->elock-AF_PACKET ->&rnp->exp_wq[2] ->&rnp->exp_wq[1] ->&rnp->exp_wq[3] ->sk_lock-AF_BLUETOOTH-BTPROTO_HCI ->slock-AF_BLUETOOTH-BTPROTO_HCI ->hci_dev_list_lock ->rlock-AF_BLUETOOTH ->wlock-AF_BLUETOOTH ->&rcu_state.expedited_wq ->stock_lock ->&pnsocks.lock ->resource_mutex ->clock-AF_PHONET ->rlock-AF_PHONET ->pfkey_mutex ->clock-AF_KEY ->wlock-AF_KEY ->rlock-AF_KEY ->&net->ipv4.ra_mutex ->&hashinfo->lock ->(console_sem).lock ->base_sockets.lock ->clock-AF_ISDN ->clock-AF_ROSE ->sk_lock-AF_ROSE ->slock-AF_ROSE ->wlock-AF_ROSE ->&list->lock#20 ->per_cpu_ptr(&cgroup_rstat_cpu_lock, cpu) ->l2tp_ip_lock ->sk_lock-AF_INET6/1 ->&net->sctp.addr_wq_lock ->sk_lock-AF_INET/1 ->&hashinfo->lock#2 ->rlock-AF_CAIF ->sk_lock-AF_CAIF ->slock-AF_CAIF ->elock-AF_CAIF ->&rnp->exp_lock ->rcu_state.exp_mutex ->raw_sk_list.lock ->clock-AF_NFC ->rlock-AF_NFC ->pool_lock ->clock-AF_NETROM ->sk_lock-AF_NETROM ->slock-AF_NETROM ->(work_completion)(&msk->work) ->clock-AF_RDS ->&rs->rs_recv_lock ->rds_cong_monitor_lock ->rds_cong_lock ->&rs->rs_lock ->&rs->rs_rdma_lock ->&q->lock ->rds_sock_lock ->raw_lock ->clock-AF_IEEE802154 ->rlock-AF_IEEE802154 ->sk_lock-AF_X25 ->slock-AF_X25 ->clock-AF_RXRPC ->(wq_completion)krxrpcd ->&wq->mutex ->rlock-AF_RXRPC ->&c->lock ->sk_lock-AF_PHONET ->slock-AF_PHONET ->&list->lock#29 ->&ping_table.lock ->dgram_lock ->l2cap_sk_list.lock ->sk_lock-AF_BLUETOOTH-BTPROTO_L2CAP ->slock-AF_BLUETOOTH-BTPROTO_L2CAP ->&chan->lock/1 ->chan_list_lock ->&____s->seqcount ->rcu_state.exp_mutex.wait_lock ->&p->pi_lock ->&match->lock ->(work_completion)(&smc->connect_work) ->sk_lock-AF_SMC ->slock-AF_SMC ->&smc->clcsock_release_lock ->rtnl_mutex ->l2tp_ip6_lock ->&cfs_rq->removed.lock ->sk_lock-AF_KCM ->slock-AF_KCM ->&mux->lock ->(work_completion)(&kcm->tx_work) ->&mux->rx_lock ->&knet->mutex ->sk_lock-AF_TIPC ->slock-AF_TIPC ->sk_lock-AF_VSOCK ->slock-AF_VSOCK ->bcm_notifier_lock ->sk_lock-AF_CAN ->slock-AF_CAN ->rlock-AF_CAN ->elock-AF_CAN ->sk_lock-AF_PPPOX ->slock-AF_PPPOX ->&net->xdp.lock ->&xs->map_list_lock ->&xs->mutex ->clock-AF_XDP ->&rng->jent_lock ->&net->xfrm.xfrm_policy_lock ->&policy->lock ->&list->lock#32 ->sk_lock-AF_QIPCRTR ->slock-AF_QIPCRTR ->sk_lock-AF_BLUETOOTH-BTPROTO_SCO ->slock-AF_BLUETOOTH-BTPROTO_SCO ->clock-AF_BLUETOOTH ->sco_sk_list.lock ->sk_lock-AF_BLUETOOTH-BTPROTO_RFCOMM ->slock-AF_BLUETOOTH-BTPROTO_RFCOMM ->rfcomm_sk_list.lock ->&d->lock ->&list->lock#33 ->&bsd_socket_locks[i] ->&n->list_lock ->&zone->lock ->sk_lock-AF_AX25 ->slock-AF_AX25 ->krc.lock ->vmap_area_lock ->purge_vmap_area_lock ->kfence_freelist_lock ->raw_notifier_lock ->&meta->lock ->rlock-AF_PPPOX ->wlock-AF_PPPOX ->sk_lock-AF_LLC ->slock-AF_LLC ->(&llc->pf_cycle_timer.timer) ->&base->lock ->(&llc->ack_timer.timer) ->(&llc->rej_sent_timer.timer) ->(&llc->busy_state_timer.timer) ->rlock-AF_LLC ->wlock-AF_LLC ->&list->lock#36 ->&list->lock#37 ->&dir->lock#2 ->(work_completion)(&(&sw_ctx_tx->tx_work.work)->work) ->rtnl_mutex.wait_lock ->(work_completion)(&strp->work) ->&x->wait#10 ->cpu_hotplug_lock ->unix_gc_lock ->crypto_default_null_skcipher_lock ->pgd_lock ->key ->percpu_counters_lock ->nfnl_grp_active_lock ->&data->lock ->&x->wait ->isotp_notifier_lock ->(work_completion)(&(&strp->msg_timer_work)->work) ->(work_completion)(&strp->work)#2 ->&local->services_lock ->cmtp_sk_list.lock FD: 52 BD: 10 +.+.: &u->lock ->clock-AF_UNIX ->&u->lock/1 ->&sk->sk_peer_lock ->rlock-AF_UNIX ->&u->peer_wait ->&ei->socket.wq.wait ->&f->f_owner.lock FD: 1 BD: 11 +...: clock-AF_UNIX FD: 31 BD: 11 +.+.: &u->peer_wait ->&p->pi_lock ->&ei->socket.wq.wait FD: 1 BD: 12 +.+.: rlock-AF_UNIX FD: 934 BD: 2 .+.+: sb_writers#4 ->mount_lock ->tk_core.seq.seqcount ->mmu_notifier_invalidate_range_start ->&____s->seqcount ->pool_lock#2 ->&c->lock ->&journal->j_state_lock ->jbd2_handle ->&obj_hash[i].lock ->&sb->s_type->i_lock_key#22 ->&wb->list_lock ->&wb->work_lock ->&type->i_mutex_dir_key#3 ->&type->i_mutex_dir_key#3/1 ->&xa->xa_lock#9 ->lock#4 ->&mapping->private_lock ->&dd->lock ->bit_wait_table + i ->&rq->__lock ->batched_entropy_u8.lock ->kfence_freelist_lock ->&meta->lock ->&sb->s_type->i_mutex_key#8 ->tomoyo_ss ->&sem->wait_lock ->&p->pi_lock ->&s->s_inode_list_lock ->sb_internal ->inode_hash_lock ->&fsnotify_mark_srcu ->&dentry->d_lock ->rcu_node_0 ->&iint->mutex ->fs_reclaim ->&ei->xattr_sem ->integrity_iint_lock ->stock_lock ->lock#5 ->&lruvec->lru_lock ->&____s->seqcount#2 ->&cfs_rq->removed.lock ->quarantine_lock ->&rcu_state.expedited_wq ->mapping.invalidate_lock ->&folio_wait_table[i] ->&n->list_lock ->remove_cache_srcu ->&sbi->s_writepages_rwsem ->&journal->j_wait_commit ->&journal->j_wait_done_commit ->&journal->j_list_lock ->&journal->j_wait_transaction_locked ->&sb->s_type->i_mutex_key#8/4 ->(console_sem).lock ->&journal->j_barrier ->pgd_lock ->key ->pcpu_lock ->percpu_counters_lock ->per_cpu_ptr(&cgroup_rstat_cpu_lock, cpu) ->&base->lock ->&lock->wait_lock ->&fq->mq_flush_lock ->&x->wait#26 ->(&timer.timer) ->&pipe->mutex/1 ->&pipe->rd_wait FD: 1 BD: 4518 +.+.: &pid->lock FD: 1 BD: 27 +.+.: &new_ns->ns_lock FD: 1 BD: 2 +.+.: (work_completion)(&sbi->s_error_work) FD: 1 BD: 176 ....: key#3 FD: 1 BD: 170 ....: key#4 FD: 1 BD: 2 +.+.: &sbi->s_error_lock FD: 30 BD: 172 ..-.: &fq->mq_flush_lock ->tk_core.seq.seqcount ->&q->requeue_lock ->&obj_hash[i].lock ->bit_wait_table + i ->&x->wait#26 FD: 1 BD: 178 ..-.: &q->requeue_lock FD: 4 BD: 3 +.+.: &eli->li_list_mtx ->&obj_hash[i].lock ->pool_lock#2 FD: 143 BD: 169 ++++: jbd2_handle ->mmu_notifier_invalidate_range_start ->&____s->seqcount ->pool_lock#2 ->&c->lock ->&ret->b_state_lock ->&journal->j_revoke_lock ->&ei->i_raw_lock ->&journal->j_wait_updates ->&mapping->private_lock ->&meta_group_info[i]->alloc_sem ->tk_core.seq.seqcount ->inode_hash_lock ->batched_entropy_u32.lock ->&ei->i_es_lock ->&sb->s_type->i_lock_key#22 ->&obj_hash[i].lock ->&rq->__lock ->&journal->j_state_lock ->rcu_node_0 ->&sbi->s_orphan_lock ->&ei->i_data_sem ->&journal->j_list_lock ->&xa->xa_lock#9 ->lock#4 ->lock#5 ->&base->lock ->&dd->lock ->&rq_wait->wait ->stock_lock ->&____s->seqcount#2 ->key#4 ->&ei->i_prealloc_lock ->&(ei->i_block_reservation_lock) ->bit_wait_table + i ->&bgl->locks[i].lock ->&cfs_rq->removed.lock ->&journal->j_wait_reserved ->&folio_wait_table[i] ->per_cpu_ptr(&cgroup_rstat_cpu_lock, cpu) ->&n->list_lock ->&lock->wait_lock ->&p->pi_lock ->&rcu_state.expedited_wq ->&sem->wait_lock ->pgd_lock ->key ->pcpu_lock ->percpu_counters_lock ->remove_cache_srcu ->batched_entropy_u8.lock ->kfence_freelist_lock ->&lruvec->lru_lock ->&ei->i_data_sem/1 ->hrtimer_bases.lock ->&meta->lock ->&s->s_inode_list_lock ->key#31 FD: 71 BD: 174 +.+.: &ret->b_state_lock ->&journal->j_list_lock ->&obj_hash[i].lock ->bit_wait_table + i FD: 70 BD: 4534 +.+.: &journal->j_list_lock ->&sb->s_type->i_lock_key#3 ->&wb->list_lock ->&obj_hash[i].lock ->&c->lock ->pool_lock#2 ->key#15 ->&meta->lock ->kfence_freelist_lock ->bit_wait_table + i FD: 1 BD: 173 +.+.: &journal->j_revoke_lock FD: 1 BD: 173 +.+.: &ei->i_raw_lock FD: 24 BD: 174 ....: &journal->j_wait_updates ->&p->pi_lock FD: 30 BD: 4562 ..-.: &wb->work_lock ->&obj_hash[i].lock ->&base->lock FD: 51 BD: 170 ++++: &meta_group_info[i]->alloc_sem ->mmu_notifier_invalidate_range_start ->pool_lock#2 ->tk_core.seq.seqcount ->&dd->lock ->&x->wait#26 ->&obj_hash[i].lock ->rcu_node_0 ->&rq->__lock ->&base->lock ->(&timer.timer) ->&fq->mq_flush_lock ->&bgl->locks[i].lock FD: 148 BD: 165 .+.+: sb_internal ->mmu_notifier_invalidate_range_start ->pool_lock#2 ->&journal->j_state_lock ->jbd2_handle ->&obj_hash[i].lock ->&c->lock ->&rq->__lock ->&____s->seqcount#2 ->&____s->seqcount ->rcu_node_0 ->remove_cache_srcu ->&n->list_lock ->batched_entropy_u8.lock ->kfence_freelist_lock ->&meta->lock ->&cfs_rq->removed.lock ->quarantine_lock ->&journal->j_wait_transaction_locked ->&journal->j_wait_commit ->&journal->j_wait_done_commit ->pgd_lock ->key ->pcpu_lock ->percpu_counters_lock ->&rcu_state.expedited_wq ->&base->lock FD: 2 BD: 4532 ++++: &ei->i_prealloc_lock ->&pa->pa_lock#2 FD: 26 BD: 1 .+.+: file_rwsem ->&ctx->flc_lock ->&rq->__lock ->rcu_node_0 FD: 2 BD: 2 +.+.: &ctx->flc_lock ->&fll->lock FD: 1 BD: 3 +.+.: &fll->lock FD: 234 BD: 3 +.+.: &type->i_mutex_dir_key#3/1 ->rename_lock.seqcount ->&dentry->d_lock ->fs_reclaim ->&ei->i_es_lock ->&ei->i_data_sem ->mmu_notifier_invalidate_range_start ->&____s->seqcount ->pool_lock#2 ->&xa->xa_lock#9 ->lock#4 ->&mapping->private_lock ->tk_core.seq.seqcount ->&dd->lock ->&obj_hash[i].lock ->bit_wait_table + i ->&rq->__lock ->inode_hash_lock ->&c->lock ->&n->list_lock ->&journal->j_state_lock ->&sb->s_type->i_lock_key#22 ->tomoyo_ss ->&s->s_inode_list_lock ->&ei->xattr_sem ->jbd2_handle ->&sb->s_type->i_mutex_key#8 ->&sem->wait_lock ->remove_cache_srcu ->&xa->xa_lock#4 ->stock_lock ->&____s->seqcount#2 ->&fsnotify_mark_srcu ->&type->i_mutex_dir_key#3 ->&wb->list_lock ->sb_internal ->rcu_node_0 ->per_cpu_ptr(&cgroup_rstat_cpu_lock, cpu) ->&rcu_state.gp_wq ->&rcu_state.expedited_wq ->&cfs_rq->removed.lock ->batched_entropy_u8.lock ->kfence_freelist_lock ->&meta->lock ->&journal->j_wait_commit ->&journal->j_wait_done_commit ->&u->bindlock ->&journal->j_wait_transaction_locked ->quarantine_lock ->&base->lock ->pgd_lock ->key ->pcpu_lock ->percpu_counters_lock FD: 138 BD: 1 +.+.: &type->s_umount_key#36/1 ->fs_reclaim ->pool_lock#2 ->pcpu_alloc_mutex ->shrinker_rwsem ->list_lrus_mutex ->sb_lock ->mmu_notifier_invalidate_range_start ->&sb->s_type->i_lock_key#25 ->&s->s_inode_list_lock ->tk_core.seq.seqcount ->&sb->s_type->i_mutex_key#11 ->&dentry->d_lock FD: 40 BD: 3 +.+.: &sb->s_type->i_lock_key#25 ->&dentry->d_lock FD: 127 BD: 2 +.+.: &sb->s_type->i_mutex_key#11 ->fs_reclaim ->pool_lock#2 ->&dentry->d_lock ->mmu_notifier_invalidate_range_start ->&sb->s_type->i_lock_key#25 ->&s->s_inode_list_lock ->tk_core.seq.seqcount FD: 128 BD: 1 +.+.: &type->s_umount_key#37 ->sb_lock ->fs_reclaim ->pool_lock#2 ->&dentry->d_lock ->&lru->node[i].lock ->&obj_hash[i].lock FD: 42 BD: 1 +.+.: &type->s_umount_key#38 ->sb_lock ->&dentry->d_lock FD: 137 BD: 1 +.+.: &type->s_umount_key#39/1 ->fs_reclaim ->pool_lock#2 ->pcpu_alloc_mutex ->shrinker_rwsem ->list_lrus_mutex ->sb_lock ->mmu_notifier_invalidate_range_start ->&c->lock ->&____s->seqcount ->&sb->s_type->i_lock_key#26 ->&s->s_inode_list_lock ->tk_core.seq.seqcount ->&dentry->d_lock FD: 40 BD: 4 +.+.: &sb->s_type->i_lock_key#26 ->&dentry->d_lock FD: 1 BD: 1 +.+.: redirect_lock FD: 350 BD: 1 +.+.: &tty->atomic_write_lock ->fs_reclaim ->pool_lock#2 ->&mm->mmap_lock ->&tty->termios_rwsem ->&tty->files_lock FD: 35 BD: 7 +.+.: &ldata->output_lock ->&port_lock_key ->&rq->__lock FD: 137 BD: 1 +.+.: &type->s_umount_key#40/1 ->fs_reclaim ->pcpu_alloc_mutex ->shrinker_rwsem ->list_lrus_mutex ->&c->lock ->&____s->seqcount ->pool_lock#2 ->sb_lock ->mmu_notifier_invalidate_range_start ->&sb->s_type->i_lock_key#27 ->&s->s_inode_list_lock ->tk_core.seq.seqcount ->fuse_mutex ->&dentry->d_lock FD: 40 BD: 4540 +.+.: &sb->s_type->i_lock_key#27 ->&dentry->d_lock FD: 1 BD: 2 +.+.: fuse_mutex FD: 138 BD: 1 +.+.: &type->s_umount_key#41/1 ->fs_reclaim ->pool_lock#2 ->pcpu_alloc_mutex ->shrinker_rwsem ->list_lrus_mutex ->sb_lock ->mmu_notifier_invalidate_range_start ->&sb->s_type->i_lock_key#28 ->&s->s_inode_list_lock ->tk_core.seq.seqcount ->pstore_sb_lock ->&dentry->d_lock FD: 40 BD: 2 +.+.: &sb->s_type->i_lock_key#28 ->&dentry->d_lock FD: 1 BD: 2 +.+.: pstore_sb_lock FD: 141 BD: 1 +.+.: &type->s_umount_key#42/1 ->fs_reclaim ->pcpu_alloc_mutex ->shrinker_rwsem ->list_lrus_mutex ->sb_lock ->pool_lock#2 ->mmu_notifier_invalidate_range_start ->&sb->s_type->i_lock_key#29 ->&s->s_inode_list_lock ->tk_core.seq.seqcount ->bpf_preload_lock ->&dentry->d_lock FD: 40 BD: 2 +.+.: &sb->s_type->i_lock_key#29 ->&dentry->d_lock FD: 129 BD: 2 +.+.: bpf_preload_lock ->(kmod_concurrent_max).lock ->fs_reclaim ->pool_lock#2 ->&obj_hash[i].lock ->&x->wait#17 ->&rq->__lock ->running_helpers_waitq.lock FD: 27 BD: 12 +.-.: (&cb->timer) ->&obj_hash[i].lock ->&base->lock ->tk_core.seq.seqcount ->&rq_wait->wait FD: 24 BD: 1 ++++: uts_sem ->hostname_poll.wait.lock ->&rq->__lock FD: 132 BD: 3 ++++: &type->i_mutex_dir_key#5 ->fs_reclaim ->&dentry->d_lock ->rename_lock.seqcount ->tomoyo_ss ->&sbinfo->stat_lock ->pool_lock#2 ->mmu_notifier_invalidate_range_start ->&sb->s_type->i_lock_key ->&s->s_inode_list_lock ->tk_core.seq.seqcount ->batched_entropy_u32.lock ->&simple_offset_xa_lock ->&c->lock ->&____s->seqcount ->&n->list_lock ->&sem->wait_lock ->&obj_hash[i].lock ->&rq->__lock ->rcu_node_0 ->&dentry->d_lock/1 ->remove_cache_srcu ->&p->pi_lock ->batched_entropy_u8.lock ->kfence_freelist_lock ->&____s->seqcount#2 ->&cfs_rq->removed.lock ->pgd_lock ->stock_lock ->key ->pcpu_lock ->percpu_counters_lock FD: 143 BD: 2 .+.+: sb_writers#5 ->mount_lock ->&type->i_mutex_dir_key#5 ->&type->i_mutex_dir_key#5/1 ->tk_core.seq.seqcount ->&sb->s_type->i_lock_key ->&wb->list_lock ->&sb->s_type->i_mutex_key#12 ->&rq->__lock ->&sem->wait_lock ->&p->pi_lock ->&s->s_inode_list_lock ->&info->lock ->&sbinfo->stat_lock ->&xa->xa_lock#9 ->&obj_hash[i].lock ->pool_lock#2 ->&fsnotify_mark_srcu ->tomoyo_ss ->&xattrs->lock ->fs_reclaim ->&c->lock ->lock#4 ->lock#5 ->&lruvec->lru_lock ->&dentry->d_lock ->rcu_node_0 ->&cfs_rq->removed.lock ->&rcu_state.expedited_wq FD: 133 BD: 4 +.+.: &sb->s_type->i_mutex_key#12 ->&xattrs->lock ->tk_core.seq.seqcount ->&sb->s_type->i_lock_key ->&wb->list_lock ->fs_reclaim ->&____s->seqcount ->pool_lock#2 ->&xa->xa_lock#9 ->lock#4 ->&info->lock ->key#9 ->&dentry->d_lock ->&simple_offset_xa_lock ->rename_lock ->&rq->__lock ->&sb->s_type->i_mutex_key#12/4 ->tomoyo_ss ->&mapping->i_mmap_rwsem ->lock#5 ->&lruvec->lru_lock ->&obj_hash[i].lock ->per_cpu_ptr(&cgroup_rstat_cpu_lock, cpu) ->rcu_node_0 ->&cfs_rq->removed.lock ->&rcu_state.expedited_wq FD: 139 BD: 3 +.+.: &type->i_mutex_dir_key#5/1 ->rename_lock.seqcount ->fs_reclaim ->&c->lock ->&____s->seqcount ->&dentry->d_lock ->tomoyo_ss ->&sbinfo->stat_lock ->mmu_notifier_invalidate_range_start ->&sb->s_type->i_lock_key ->&s->s_inode_list_lock ->tk_core.seq.seqcount ->batched_entropy_u32.lock ->&simple_offset_xa_lock ->&obj_hash[i].lock ->pool_lock#2 ->&u->bindlock ->&sb->s_type->i_mutex_key#12 ->&fsnotify_mark_srcu ->&sem->wait_lock ->&rq->__lock ->&p->pi_lock ->&n->list_lock ->&dentry->d_lock/1 ->lock#4 ->lock#5 ->&lruvec->lru_lock ->&info->lock ->&xa->xa_lock#9 ->&sb->s_type->i_mutex_key#12/4 ->remove_cache_srcu ->&cfs_rq->removed.lock ->&____s->seqcount#2 ->rcu_node_0 ->&rcu_state.gp_wq ->batched_entropy_u8.lock ->kfence_freelist_lock ->&rcu_state.expedited_wq ->pgd_lock ->stock_lock ->key ->pcpu_lock ->percpu_counters_lock ->key#9 FD: 7 BD: 88 +.+.: &f->f_lock ->fasync_lock FD: 1 BD: 2 ....: hostname_poll.wait.lock FD: 1013 BD: 1 +.+.: &f->f_pos_lock ->&type->i_mutex_dir_key#3 ->&mm->mmap_lock ->&sb->s_type->i_mutex_key#9 ->&type->i_mutex_dir_key#2 ->&type->i_mutex_dir_key#4 ->&type->i_mutex_dir_key#5 ->sb_writers#5 ->&p->lock ->&rq->__lock ->sb_writers#4 ->rcu_node_0 ->&lock->wait_lock ->&cfs_rq->removed.lock ->&obj_hash[i].lock ->pool_lock#2 ->&sb->s_type->i_mutex_key#18 ->sb_writers#10 ->sb_writers#3 ->(console_sem).lock ->console_owner_lock ->console_owner ->sb_writers#11 ->&rcu_state.expedited_wq ->sb_writers ->&sem->wait_lock ->&p->pi_lock ->sb_writers#15 FD: 230 BD: 1 .+.+: dup_mmap_sem ->&mm->mmap_lock ->&rq->__lock FD: 131 BD: 153 +.+.: &mm->mmap_lock/1 ->pool_lock#2 ->&c->lock ->&____s->seqcount ->&vma->vm_lock->lock ->fs_reclaim ->&mapping->i_mmap_rwsem ->&anon_vma->rwsem ->mmu_notifier_invalidate_range_start ->&mm->page_table_lock ->ptlock_ptr(page) ->ptlock_ptr(page)#2 ->&mm->context.lock ->&obj_hash[i].lock ->rcu_node_0 ->&rcu_state.expedited_wq ->&rq->__lock ->&zone->lock ->&n->list_lock ->remove_cache_srcu ->&sem->wait_lock ->&p->pi_lock ->batched_entropy_u8.lock ->kfence_freelist_lock ->quarantine_lock ->pgd_lock ->key ->pcpu_lock ->percpu_counters_lock ->&cfs_rq->removed.lock ->stock_lock ->&____s->seqcount#2 ->per_cpu_ptr(&cgroup_rstat_cpu_lock, cpu) ->key#22 ->&base->lock ->&meta->lock ->pool_lock ->lock#10 FD: 26 BD: 154 +.+.: &mm->context.lock ->&rq->__lock ->rcu_node_0 ->&rcu_state.expedited_wq FD: 1 BD: 10 .+.+: &xattrs->lock FD: 130 BD: 8 +.+.: &u->bindlock ->&net->unx.table.locks[i] ->&net->unx.table.locks[i]/1 ->&bsd_socket_locks[i] ->fs_reclaim ->pool_lock#2 ->batched_entropy_u32.lock FD: 40 BD: 12 +.+.: &net->unx.table.locks[i]/1 ->&dentry->d_lock FD: 1 BD: 11 +.+.: &bsd_socket_locks[i] FD: 246 BD: 5 +.+.: &u->iolock ->rlock-AF_UNIX ->&mm->mmap_lock ->&obj_hash[i].lock ->pool_lock#2 ->&u->peer_wait ->quarantine_lock ->&rq->__lock ->&meta->lock ->kfence_freelist_lock ->fs_reclaim ->&____s->seqcount ->&u->lock ->&dir->lock ->rcu_node_0 ->&cfs_rq->removed.lock ->stock_lock ->&rcu_state.expedited_wq ->&lock->wait_lock ->&sem->wait_lock ->&p->pi_lock ->unix_gc_lock ->&c->lock ->&____s->seqcount#2 ->&n->list_lock ->batched_entropy_u8.lock ->remove_cache_srcu ->&base->lock FD: 30 BD: 4179 ..-.: &ei->socket.wq.wait ->&p->pi_lock ->&ep->lock ->&ep->poll_wait/1 FD: 1 BD: 4531 ....: key#5 FD: 1 BD: 4531 ....: key#6 FD: 1 BD: 4531 ....: key#7 FD: 1 BD: 4599 ....: &wq#3 FD: 42 BD: 11 +.+.: &u->lock/1 ->&sk->sk_peer_lock ->&dentry->d_lock ->&sk->sk_peer_lock/1 FD: 141 BD: 1 +.+.: &group->mark_mutex ->&fsnotify_mark_srcu ->fs_reclaim ->&____s->seqcount ->pool_lock#2 ->&c->lock ->lock ->ucounts_lock ->&mark->lock ->&conn->lock ->&sb->s_type->i_lock_key#5 ->&sb->s_type->i_lock_key#22 ->&sb->s_type->i_lock_key ->&rq->__lock ->&lock->wait_lock ->batched_entropy_u8.lock ->kfence_freelist_lock ->&____s->seqcount#2 ->remove_cache_srcu ->&cfs_rq->removed.lock ->&obj_hash[i].lock ->&n->list_lock FD: 12 BD: 248 +.+.: &group->inotify_data.idr_lock ->&c->lock ->&____s->seqcount ->pool_lock#2 ->&obj_hash[i].lock ->&n->list_lock ->&____s->seqcount#2 FD: 3 BD: 2 +.+.: &mark->lock ->&fsnotify_mark_srcu ->&conn->lock FD: 1 BD: 7 +.+.: &conn->lock FD: 1 BD: 1 +.+.: &evdev->client_lock FD: 231 BD: 1 +.+.: &evdev->mutex ->&dev->mutex#2 ->&mm->mmap_lock FD: 245 BD: 9 +.+.: sk_lock-AF_NETLINK ->slock-AF_NETLINK ->&mm->mmap_lock ->fs_reclaim ->&c->lock ->pool_lock#2 ->free_vmap_area_lock ->vmap_area_lock ->&____s->seqcount ->pcpu_alloc_mutex ->&obj_hash[i].lock ->pack_mutex ->batched_entropy_u32.lock ->text_mutex ->&fp->aux->used_maps_mutex ->rcu_node_0 ->&rq->__lock ->clock-AF_NETLINK FD: 25 BD: 10 +...: slock-AF_NETLINK ->&sk->sk_lock.wq FD: 1 BD: 4147 ..-.: rlock-AF_NETLINK FD: 1 BD: 10 ....: &nlk->wait FD: 1 BD: 15 ++..: clock-AF_NETLINK FD: 1 BD: 7 ....: genl_sk_destructing_waitq.lock FD: 1 BD: 7 ....: wlock-AF_NETLINK FD: 1 BD: 5 +.-.: &rdev->beacon_registrations_lock FD: 1 BD: 83 +.-.: &rdev->mgmt_registrations_lock FD: 1 BD: 87 +...: &wdev->pmsr_lock FD: 1 BD: 79 +.+.: reg_indoor_lock FD: 1072 BD: 1 .+.+: sb_writers#6 ->mount_lock ->&sb->s_type->i_mutex_key#10 FD: 131 BD: 88 +.+.: (work_completion)(&ht->run_work) ->&ht->mutex ->&rq->__lock FD: 130 BD: 89 +.+.: &ht->mutex ->fs_reclaim ->pool_lock#2 ->batched_entropy_u32.lock ->rhashtable_bucket ->&ht->lock ->remove_cache_srcu ->&c->lock ->&____s->seqcount#2 ->&____s->seqcount ->&rq->__lock ->&obj_hash[i].lock ->batched_entropy_u8.lock ->kfence_freelist_lock ->&n->list_lock ->&base->lock ->quarantine_lock ->rcu_node_0 ->&rcu_state.expedited_wq ->&meta->lock FD: 1 BD: 3943 ....: rhashtable_bucket/1 FD: 11 BD: 92 +.+.: &ht->lock ->&obj_hash[i].lock ->pool_lock#2 FD: 173 BD: 2 +.+.: (work_completion)(&w->w) ->nfc_devlist_mutex ->&obj_hash[i].lock ->pool_lock#2 ->&meta->lock ->kfence_freelist_lock ->&rq->__lock ->&base->lock FD: 23 BD: 4 +.+.: &genl_data->genl_data_mutex ->&rq->__lock FD: 2 BD: 14 +.+.: &sk->sk_peer_lock ->&sk->sk_peer_lock/1 FD: 29 BD: 7 ....: &group->notification_waitq ->&p->pi_lock ->&ep->lock FD: 1 BD: 7 +.+.: &group->notification_lock FD: 1 BD: 1 ....: &client->wait FD: 1 BD: 154 ....: key#8 FD: 924 BD: 3 +.+.: &pipe->mutex/1 ->&pipe->rd_wait ->&pipe->wr_wait ->&rq->__lock ->&lock->wait_lock ->fs_reclaim ->&____s->seqcount ->pool_lock#2 ->&mm->mmap_lock ->&cfs_rq->removed.lock ->&obj_hash[i].lock ->rcu_node_0 ->&rcu_state.expedited_wq ->stock_lock ->per_cpu_ptr(&cgroup_rstat_cpu_lock, cpu) ->sk_lock-AF_NETLINK ->slock-AF_NETLINK ->free_vmap_area_lock ->vmap_area_lock ->init_mm.page_table_lock ->&c->lock ->nfnl_subsys_ctnetlink ->purge_vmap_area_lock ->&sighand->siglock ->&n->list_lock ->sk_lock-AF_INET6 ->slock-AF_INET6 ->sk_lock-AF_INET ->slock-AF_INET ->&list->lock#34 ->&ei->socket.wq.wait ->&pipe->mutex#2/2 ->remove_cache_srcu ->rtnl_mutex ->rtnl_mutex.wait_lock ->&p->pi_lock ->rlock-AF_NETLINK ->&nlk->wait ->&sb->s_type->i_mutex_key#8 ->&____s->seqcount#2 ->batched_entropy_u8.lock ->kfence_freelist_lock ->&meta->lock ->lock#10 ->pgd_lock ->key ->pcpu_lock ->percpu_counters_lock ->batched_entropy_u32.lock ->&sem->wait_lock ->&f->f_lock ->&u->iolock ->&u->lock FD: 29 BD: 6 ....: &pipe->rd_wait ->&p->pi_lock ->&ep->lock FD: 24 BD: 173 ..-.: &x->wait#26 ->&p->pi_lock FD: 1 BD: 8 ....: key#9 FD: 28 BD: 4599 +.+.: &dentry->d_lock/2 ->&dentry->d_lock/3 FD: 27 BD: 4600 +.+.: &dentry->d_lock/3 ->&____s->seqcount#4 ->&wq FD: 1 BD: 4602 +.+.: &____s->seqcount#4/1 FD: 24 BD: 6 ....: &pipe->wr_wait ->&p->pi_lock FD: 47 BD: 1 .+.+: sb_writers#7 ->tk_core.seq.seqcount ->mount_lock ->&rq->__lock ->rcu_node_0 ->&rcu_state.expedited_wq FD: 248 BD: 1 +.+.: sk_lock-AF_UNIX ->slock-AF_UNIX ->&mm->mmap_lock ->fs_reclaim ->&obj_hash[i].lock ->pool_lock#2 ->&u->iolock FD: 1 BD: 2 +...: slock-AF_UNIX FD: 1 BD: 1 ....: &rs->lock#2 FD: 54 BD: 3 +.+.: oom_adj_mutex ->&p->alloc_lock ->&rq->__lock ->rcu_node_0 ->oom_adj_mutex.wait_lock ->&cfs_rq->removed.lock ->&obj_hash[i].lock ->pool_lock#2 FD: 321 BD: 2 +.+.: &ep->mtx ->fs_reclaim ->&____s->seqcount ->pool_lock#2 ->&c->lock ->&f->f_lock ->&ei->socket.wq.wait ->&ep->lock ->&group->notification_waitq ->&group->notification_lock ->&sighand->signalfd_wqh ->&sighand->siglock ->&rq->__lock ->&pipe->rd_wait ->&obj_hash[i].lock ->batched_entropy_u8.lock ->kfence_freelist_lock ->&meta->lock ->key#11 ->remove_cache_srcu ->pgd_lock ->key ->pcpu_lock ->percpu_counters_lock ->&cfs_rq->removed.lock ->&lock->wait_lock ->&pipe->wr_wait ->&p->pi_lock ->rcu_node_0 ->&rcu_state.expedited_wq ->stock_lock ->&n->list_lock ->&ep->mtx/1 ->&ep->poll_wait ->wakeup_ida.xa_lock ->&x->wait#9 ->&k->list_lock ->gdp_mutex ->lock ->&root->kernfs_rwsem ->bus_type_sem ->sysfs_symlink_target_lock ->uevent_sock_mutex ->subsys mutex#15 ->events_lock ->&dentry->d_lock ->&u->lock ->&ws->lock ->&ACCESS_PRIVATE(sdp, lock) ->wakeup_srcu ->&x->wait#2 ->(&ws->timer) ->&base->lock ->deferred_probe_mutex ->device_links_lock ->mmu_notifier_invalidate_range_start ->deleted_ws.lock ->&____s->seqcount#2 ->&sem->wait_lock ->uevent_sock_mutex.wait_lock ->kernfs_idr_lock ->kn->active#5 ->quarantine_lock ->sk_lock-AF_VSOCK ->slock-AF_VSOCK ->rlock-AF_PACKET ->wlock-AF_PACKET FD: 322 BD: 1 +.+.: epnested_mutex ->&ep->mtx ->&ep->mtx/1 FD: 28 BD: 4192 ...-: &ep->lock ->&ep->wq ->&ws->lock FD: 29 BD: 156 ....: &sighand->signalfd_wqh ->&ep->lock ->&p->pi_lock FD: 931 BD: 1 .+.+: sb_writers#8 ->mount_lock ->tk_core.seq.seqcount ->&sb->s_type->i_lock_key#24 ->&wb->list_lock ->&type->i_mutex_dir_key#4 ->fs_reclaim ->pool_lock#2 ->&mm->mmap_lock ->&of->mutex ->&obj_hash[i].lock ->&c->lock ->&n->list_lock ->&rq->__lock ->remove_cache_srcu ->rcu_node_0 ->&root->kernfs_iattr_rwsem ->&dentry->d_lock ->tomoyo_ss ->&sb->s_type->i_mutex_key#13 ->iattr_mutex ->&xattrs->lock ->&____s->seqcount#2 ->&____s->seqcount ->&cfs_rq->removed.lock FD: 3 BD: 10 +.+.: swap_lock ->&p->lock#2 FD: 140 BD: 1 .+.+: kn->active ->fs_reclaim ->pool_lock#2 ->&kernfs_locks->open_file_mutex[count] ->&k->list_lock ->uevent_sock_mutex ->&obj_hash[i].lock ->&c->lock ->&____s->seqcount ->quarantine_lock ->&n->list_lock ->&rq->__lock FD: 127 BD: 80 +.+.: &kernfs_locks->open_file_mutex[count] ->fs_reclaim ->pool_lock#2 ->&obj_hash[i].lock ->krc.lock ->&c->lock ->&____s->seqcount ->&n->list_lock ->&rq->__lock ->remove_cache_srcu ->rcu_node_0 ->&lock->wait_lock ->&____s->seqcount#2 ->&rcu_state.expedited_wq ->quarantine_lock ->&cfs_rq->removed.lock FD: 128 BD: 2 +.+.: &sb->s_type->i_mutex_key#13 ->tk_core.seq.seqcount ->&root->kernfs_iattr_rwsem ->&sem->wait_lock ->&p->pi_lock ->&rq->__lock FD: 929 BD: 6 +.+.: &of->mutex ->&rq->__lock ->cgroup_mutex ->&p->pi_lock ->cgroup_mutex.wait_lock ->&root->deactivate_waitq FD: 24 BD: 4193 ..-.: &ep->wq ->&p->pi_lock FD: 139 BD: 1 .+.+: kn->active#2 ->fs_reclaim ->&kernfs_locks->open_file_mutex[count] ->&c->lock ->&n->list_lock ->pool_lock#2 ->uevent_sock_mutex ->&obj_hash[i].lock ->&____s->seqcount ->remove_cache_srcu ->quarantine_lock ->&rq->__lock FD: 35 BD: 2 +.+.: (work_completion)(&(&krcp->krw_arr[i].rcu_work)->work) ->krc.lock ->&obj_hash[i].lock ->&rq->__lock ->rcu_node_0 ->&rcu_state.expedited_wq FD: 139 BD: 1 .+.+: kn->active#3 ->fs_reclaim ->&kernfs_locks->open_file_mutex[count] ->pool_lock#2 ->uevent_sock_mutex ->&obj_hash[i].lock ->&c->lock ->&____s->seqcount ->&rq->__lock ->&n->list_lock ->remove_cache_srcu ->quarantine_lock FD: 132 BD: 1 .+.+: kn->active#4 ->fs_reclaim ->&kernfs_locks->open_file_mutex[count] ->param_lock ->pool_lock#2 ->&on->poll ->&c->lock ->&rq->__lock ->&n->list_lock ->remove_cache_srcu FD: 126 BD: 256 +.+.: iattr_mutex ->fs_reclaim ->&____s->seqcount ->pool_lock#2 ->&rq->__lock ->&c->lock ->tk_core.seq.seqcount ->&cfs_rq->removed.lock FD: 1 BD: 80 +.+.: disk_events_mutex FD: 168 BD: 3 ++++: kn->active#5 ->fs_reclaim ->&kernfs_locks->open_file_mutex[count] ->pool_lock#2 ->uevent_sock_mutex ->&obj_hash[i].lock ->&c->lock ->remove_cache_srcu ->&n->list_lock ->&____s->seqcount ->&device->physical_node_lock ->&rq->__lock ->udc_lock ->quarantine_lock ->fw_lock ->batched_entropy_u8.lock ->kfence_freelist_lock ->&meta->lock ->rcu_node_0 ->&rfkill->lock ->&____s->seqcount#2 ->uevent_sock_mutex.wait_lock ->&p->pi_lock ->&rcu_state.expedited_wq ->&cfs_rq->removed.lock ->&root->deactivate_waitq FD: 128 BD: 1 .+.+: kn->active#6 ->fs_reclaim ->&kernfs_locks->open_file_mutex[count] ->&c->lock ->&____s->seqcount ->&n->list_lock FD: 128 BD: 1 .+.+: kn->active#7 ->fs_reclaim ->&c->lock ->&n->list_lock ->&kernfs_locks->open_file_mutex[count] FD: 128 BD: 1 .+.+: kn->active#8 ->fs_reclaim ->&kernfs_locks->open_file_mutex[count] ->&c->lock FD: 128 BD: 1 .+.+: kn->active#9 ->fs_reclaim ->&kernfs_locks->open_file_mutex[count] ->&c->lock ->&____s->seqcount ->&n->list_lock FD: 128 BD: 1 .+.+: kn->active#10 ->fs_reclaim ->&kernfs_locks->open_file_mutex[count] ->&c->lock ->&____s->seqcount ->remove_cache_srcu FD: 128 BD: 1 .+.+: kn->active#11 ->fs_reclaim ->&kernfs_locks->open_file_mutex[count] ->&c->lock ->&____s->seqcount FD: 128 BD: 1 .+.+: kn->active#12 ->fs_reclaim ->&kernfs_locks->open_file_mutex[count] ->&c->lock ->&n->list_lock FD: 128 BD: 1 .+.+: kn->active#13 ->fs_reclaim ->&c->lock ->&____s->seqcount ->&kernfs_locks->open_file_mutex[count] FD: 128 BD: 1 .+.+: kn->active#14 ->fs_reclaim ->&kernfs_locks->open_file_mutex[count] ->pool_lock#2 ->&obj_hash[i].lock FD: 130 BD: 1 .+.+: kn->active#15 ->fs_reclaim ->&c->lock ->&n->list_lock ->&kernfs_locks->open_file_mutex[count] ->dev_base_lock ->&____s->seqcount ->remove_cache_srcu ->&____s->seqcount#2 FD: 130 BD: 1 .+.+: kn->active#16 ->fs_reclaim ->&kernfs_locks->open_file_mutex[count] ->dev_base_lock ->&c->lock ->&____s->seqcount ->&n->list_lock FD: 128 BD: 1 .+.+: kn->active#17 ->fs_reclaim ->&kernfs_locks->open_file_mutex[count] ->&c->lock ->&rq->__lock ->&____s->seqcount#2 ->&____s->seqcount FD: 130 BD: 1 .+.+: kn->active#18 ->fs_reclaim ->&c->lock ->&kernfs_locks->open_file_mutex[count] ->dev_base_lock ->&n->list_lock FD: 128 BD: 1 .+.+: kn->active#19 ->fs_reclaim ->pool_lock#2 ->&kernfs_locks->open_file_mutex[count] ->&c->lock FD: 130 BD: 1 .+.+: kn->active#20 ->fs_reclaim ->&kernfs_locks->open_file_mutex[count] ->dev_base_lock ->&c->lock ->&rq->__lock FD: 128 BD: 1 .+.+: kn->active#21 ->fs_reclaim ->&kernfs_locks->open_file_mutex[count] ->&c->lock ->&____s->seqcount ->&n->list_lock FD: 128 BD: 1 .+.+: kn->active#22 ->fs_reclaim ->&kernfs_locks->open_file_mutex[count] ->&c->lock ->&n->list_lock FD: 130 BD: 1 .+.+: kn->active#23 ->fs_reclaim ->&kernfs_locks->open_file_mutex[count] ->dev_base_lock FD: 128 BD: 1 .+.+: kn->active#24 ->fs_reclaim ->&kernfs_locks->open_file_mutex[count] ->&c->lock ->&n->list_lock FD: 133 BD: 1 .+.+: kn->active#25 ->fs_reclaim ->&kernfs_locks->open_file_mutex[count] ->&dev->power.lock ->pci_lock FD: 2 BD: 8 ....: pci_lock ->pci_config_lock FD: 128 BD: 1 .+.+: kn->active#26 ->fs_reclaim ->&kernfs_locks->open_file_mutex[count] ->&c->lock ->remove_cache_srcu ->&n->list_lock FD: 128 BD: 1 .+.+: kn->active#27 ->fs_reclaim ->&kernfs_locks->open_file_mutex[count] ->&c->lock ->&n->list_lock ->&____s->seqcount FD: 128 BD: 1 .+.+: kn->active#28 ->fs_reclaim ->&kernfs_locks->open_file_mutex[count] FD: 128 BD: 1 .+.+: kn->active#29 ->&rq->__lock ->fs_reclaim ->&kernfs_locks->open_file_mutex[count] ->&c->lock ->&n->list_lock FD: 128 BD: 1 .+.+: kn->active#30 ->fs_reclaim ->&kernfs_locks->open_file_mutex[count] FD: 128 BD: 1 .+.+: kn->active#31 ->fs_reclaim ->&kernfs_locks->open_file_mutex[count] FD: 128 BD: 1 .+.+: kn->active#32 ->fs_reclaim ->&kernfs_locks->open_file_mutex[count] ->&c->lock FD: 128 BD: 1 .+.+: kn->active#33 ->fs_reclaim ->&kernfs_locks->open_file_mutex[count] FD: 1 BD: 1 +.+.: &mousedev->client_lock FD: 32 BD: 8 +.+.: &mousedev->mutex#2 ->&dev->mutex#2 ->&lock->wait_lock ->&p->pi_lock FD: 1 BD: 1 +.+.: &sb->s_type->i_mutex_key#14 FD: 56 BD: 1 .+.+: mapping.invalidate_lock#2 ->mmu_notifier_invalidate_range_start ->&____s->seqcount ->pool_lock#2 ->&xa->xa_lock#9 ->lock#4 ->tk_core.seq.seqcount ->&dd->lock ->&c->lock ->&rq->__lock ->per_cpu_ptr(&cgroup_rstat_cpu_lock, cpu) FD: 128 BD: 1 .+.+: kn->active#34 ->fs_reclaim ->&kernfs_locks->open_file_mutex[count] ->&c->lock ->&n->list_lock ->&____s->seqcount ->remove_cache_srcu ->batched_entropy_u8.lock ->kfence_freelist_lock FD: 129 BD: 5 +.+.: &sb->s_type->i_mutex_key#12/4 ->&dentry->d_lock ->&simple_offset_xa_lock ->fs_reclaim ->tk_core.seq.seqcount ->rename_lock ->&rq->__lock ->&cfs_rq->removed.lock ->&obj_hash[i].lock ->pool_lock#2 FD: 128 BD: 1 .+.+: kn->active#35 ->fs_reclaim ->&kernfs_locks->open_file_mutex[count] FD: 129 BD: 1 .+.+: kn->active#36 ->fs_reclaim ->&kernfs_locks->open_file_mutex[count] ->i2c_dev_list_lock FD: 128 BD: 1 .+.+: kn->active#37 ->fs_reclaim ->&kernfs_locks->open_file_mutex[count] FD: 128 BD: 1 .+.+: kn->active#38 ->fs_reclaim ->&c->lock ->&kernfs_locks->open_file_mutex[count] FD: 131 BD: 2 +.+.: &dev_instance->mutex ->fs_reclaim ->&c->lock ->pool_lock#2 ->vicodec_core:1844:(hdl)->_lock ->&vdev->fh_lock ->&m2m_dev->job_spinlock ->&q->done_wq ->&q->mmap_lock ->&obj_hash[i].lock ->&rq->__lock FD: 4 BD: 3 +.+.: vicodec_core:1844:(hdl)->_lock ->&obj_hash[i].lock ->pool_lock#2 FD: 1 BD: 5 ....: &vdev->fh_lock FD: 137 BD: 1 +.+.: &mdev->req_queue_mutex ->&dev_instance->mutex ->&lock->wait_lock ->&rq->__lock ->&vdev->fh_lock ->&mdev->graph_mutex ->vicodec_core:1844:(hdl)->_lock ->&obj_hash[i].lock ->pool_lock#2 ->vim2m:1183:(hdl)->_lock ->&dev->dev_mutex ->&dev->mutex#3 FD: 1 BD: 4 ....: &m2m_dev->job_spinlock FD: 1 BD: 4 ....: &q->done_wq FD: 1 BD: 4 +.+.: &q->mmap_lock FD: 128 BD: 1 .+.+: kn->active#39 ->fs_reclaim ->&kernfs_locks->open_file_mutex[count] ->&c->lock ->&n->list_lock ->&____s->seqcount ->remove_cache_srcu ->&rq->__lock FD: 1 BD: 1 +.+.: fh->state->lock FD: 30 BD: 3 ..-.: &(&wb->dwork)->timer FD: 172 BD: 3 +.+.: (wq_completion)writeback ->(work_completion)(&(&wb->dwork)->work) ->(work_completion)(&(&wb->bw_dwork)->work) ->&rq->__lock ->(work_completion)(&barr->work) FD: 169 BD: 4 +.+.: (work_completion)(&(&wb->dwork)->work) ->&wb->work_lock ->&wb->list_lock ->&p->sequence ->key#10 ->&sb->s_type->i_lock_key#22 ->&sbi->s_writepages_rwsem ->pool_lock#2 ->&dd->lock ->&obj_hash[i].lock ->&pl->lock ->&rq->__lock ->&bdi->wb_waitq FD: 2 BD: 6 +.-.: &p->sequence ->key#13 FD: 1 BD: 4562 ..-.: key#10 FD: 131 BD: 2 +.+.: &dev->dev_mutex ->fs_reclaim ->pool_lock#2 ->vim2m:1183:(hdl)->_lock ->&c->lock ->&obj_hash[i].lock ->&vdev->fh_lock ->&m2m_dev->job_spinlock ->&q->done_wq ->&q->mmap_lock FD: 4 BD: 3 +.+.: vim2m:1183:(hdl)->_lock ->&obj_hash[i].lock ->pool_lock#2 FD: 1 BD: 1 +.+.: &vcapture->lock FD: 2 BD: 2 +.+.: &dev->mutex#3 ->&vdev->fh_lock FD: 24 BD: 1 +.-.: (&journal->j_commit_timer) ->&p->pi_lock FD: 106 BD: 166 +.+.: &journal->j_checkpoint_mutex ->mmu_notifier_invalidate_range_start ->pool_lock#2 ->tk_core.seq.seqcount ->&dd->lock ->&obj_hash[i].lock ->&base->lock ->bit_wait_table + i ->&rq->__lock ->&journal->j_state_lock ->&fq->mq_flush_lock ->&x->wait#26 ->&journal->j_list_lock ->&c->lock ->rcu_node_0 ->(&timer.timer) ->&ei->i_es_lock ->&mapping->private_lock ->&meta->lock ->kfence_freelist_lock ->&sb->s_type->i_lock_key#3 ->lock#4 ->lock#5 ->&lruvec->lru_lock ->&n->list_lock ->&____s->seqcount#2 ->&____s->seqcount ->batched_entropy_u8.lock ->&cfs_rq->removed.lock ->key#27 ->quarantine_lock FD: 24 BD: 176 ....: &journal->j_wait_transaction_locked ->&p->pi_lock FD: 1 BD: 4544 ..-.: &memcg->move_lock FD: 1 BD: 174 +.+.: &sbi->s_md_lock FD: 1 BD: 1 ....: &journal->j_fc_wait FD: 1 BD: 1 +.+.: &journal->j_history_lock FD: 1 BD: 15 +.+.: &sk->sk_peer_lock/1 FD: 1 BD: 3 ....: key#11 FD: 129 BD: 5 +.+.: &sb->s_type->i_mutex_key#4/4 ->&dentry->d_lock ->&simple_offset_xa_lock ->fs_reclaim ->tk_core.seq.seqcount ->rename_lock FD: 25 BD: 1 ..-.: drivers/base/dd.c:321 FD: 36 BD: 2 +.+.: (deferred_probe_timeout_work).work ->device_links_lock ->deferred_probe_mutex ->deferred_probe_work ->&x->wait#10 ->&pool->lock ->&rq->__lock ->&obj_hash[i].lock FD: 166 BD: 164 ++++: &sbi->s_writepages_rwsem ->&xa->xa_lock#9 ->mmu_notifier_invalidate_range_start ->&____s->seqcount ->pool_lock#2 ->&c->lock ->lock#4 ->lock#5 ->&obj_hash[i].lock ->&journal->j_state_lock ->jbd2_handle ->tk_core.seq.seqcount ->&dd->lock ->&base->lock ->&rq->__lock ->&rq_wait->wait ->rcu_node_0 ->&rsp->gp_wait ->&rnp->exp_wq[2] ->&ei->i_data_sem ->&____s->seqcount#2 ->remove_cache_srcu ->&rnp->exp_wq[0] ->&rnp->exp_lock ->rcu_state.exp_mutex ->rcu_state.exp_mutex.wait_lock ->&p->pi_lock ->&cfs_rq->removed.lock ->&rcu_state.expedited_wq ->pool_lock ->&mapping->private_lock ->&folio_wait_table[i] ->&n->list_lock ->quarantine_lock ->batched_entropy_u8.lock ->kfence_freelist_lock ->&meta->lock ->&ext4__ioend_wq[i] ->&sem->waiters ->&journal->j_wait_commit ->&journal->j_wait_done_commit ->&journal->j_barrier ->&journal->j_wait_transaction_locked ->&rnp->exp_wq[1] ->&sb->s_type->i_lock_key#22 ->&s->s_inode_list_lock ->sb_internal ->inode_hash_lock ->&fsnotify_mark_srcu FD: 84 BD: 1 .+.+: &type->s_umount_key#43 ->&sb->s_type->i_lock_key#3 ->mmu_notifier_invalidate_range_start ->pool_lock#2 ->tk_core.seq.seqcount ->&dd->lock ->&obj_hash[i].lock ->&base->lock ->&c->lock ->lock#4 ->lock#5 ->&wb->list_lock ->&____s->seqcount ->&rq_wait->wait ->rcu_node_0 ->&rq->__lock ->&n->list_lock ->&____s->seqcount#2 ->lock#11 FD: 1 BD: 4562 ..-.: &s->s_inode_wblist_lock FD: 1 BD: 4563 ..-.: key#12 FD: 30 BD: 3 ..-.: &(&wb->bw_dwork)->timer FD: 67 BD: 4 +.+.: (work_completion)(&(&wb->bw_dwork)->work) ->&wb->list_lock ->&rq->__lock FD: 128 BD: 1 .+.+: kn->active#40 ->fs_reclaim ->&kernfs_locks->open_file_mutex[count] FD: 128 BD: 1 .+.+: kn->active#41 ->fs_reclaim ->&kernfs_locks->open_file_mutex[count] FD: 128 BD: 1 .+.+: kn->active#42 ->fs_reclaim ->&kernfs_locks->open_file_mutex[count] FD: 128 BD: 1 .+.+: kn->active#43 ->fs_reclaim ->&kernfs_locks->open_file_mutex[count] FD: 23 BD: 7 +.+.: &lo->lo_mutex ->&rq->__lock ->&cfs_rq->removed.lock ->&obj_hash[i].lock ->pool_lock#2 FD: 156 BD: 12 +.+.: &nbd->config_lock ->mmu_notifier_invalidate_range_start ->pool_lock#2 ->&bdev->bd_size_lock ->&q->queue_lock ->&ACCESS_PRIVATE(sdp, lock) ->set->srcu ->&obj_hash[i].lock ->&rq->__lock ->&x->wait#2 ->&c->lock ->(console_sem).lock ->fs_reclaim ->uevent_sock_mutex ->&lock->wait_lock ->&n->list_lock ->rcu_node_0 ->&cfs_rq->removed.lock ->&base->lock ->&____s->seqcount#2 ->&____s->seqcount ->remove_cache_srcu ->pgd_lock ->stock_lock ->key ->pcpu_lock ->percpu_counters_lock FD: 26 BD: 10 ....: &ACCESS_PRIVATE(ssp->srcu_sup, lock) ->&ACCESS_PRIVATE(sdp, lock) FD: 15 BD: 1 +.-.: (&dom->period_timer) ->key#13 ->&p->sequence ->&obj_hash[i].lock ->&base->lock FD: 1 BD: 4564 ..-.: key#13 FD: 128 BD: 1 .+.+: kn->active#44 ->fs_reclaim ->&kernfs_locks->open_file_mutex[count] FD: 2 BD: 6 +.+.: &new->lock ->&mtdblk->cache_mutex FD: 1 BD: 7 +.+.: &mtdblk->cache_mutex FD: 128 BD: 1 .+.+: kn->active#45 ->fs_reclaim ->&kernfs_locks->open_file_mutex[count] FD: 230 BD: 1 +.+.: &mtd->master.chrdev_lock ->&mm->mmap_lock FD: 1 BD: 4 +.+.: destroy_lock FD: 30 BD: 1 ..-.: fs/notify/mark.c:89 FD: 136 BD: 2 +.+.: (reaper_work).work ->destroy_lock ->&ACCESS_PRIVATE(sdp, lock) ->&fsnotify_mark_srcu ->&obj_hash[i].lock ->&rq->__lock ->&x->wait#2 ->pool_lock#2 ->&meta->lock ->kfence_freelist_lock ->rcu_node_0 ->&rcu_state.expedited_wq ->&cfs_rq->removed.lock ->pool_lock ->&ACCESS_PRIVATE(ssp->srcu_sup, lock) FD: 136 BD: 2 +.+.: connector_reaper_work ->destroy_lock ->&ACCESS_PRIVATE(sdp, lock) ->&fsnotify_mark_srcu ->&obj_hash[i].lock ->&x->wait#2 ->&rq->__lock ->pool_lock#2 ->pool_lock ->&cfs_rq->removed.lock ->&ACCESS_PRIVATE(ssp->srcu_sup, lock) ->rcu_node_0 ->&rcu_state.expedited_wq FD: 1 BD: 1 +.+.: userns_state_mutex FD: 4 BD: 79 +...: fib_info_lock ->&obj_hash[i].lock ->pool_lock#2 FD: 130 BD: 82 +...: &net->sctp.local_addr_lock ->&net->sctp.addr_wq_lock FD: 129 BD: 87 +.-.: &net->sctp.addr_wq_lock ->pool_lock#2 ->&obj_hash[i].lock ->&base->lock ->&c->lock ->&n->list_lock ->&____s->seqcount ->&____s->seqcount#2 ->slock-AF_INET6/1 ->slock-AF_INET/1 ->k-slock-AF_INET6/1 ->batched_entropy_u8.lock ->kfence_freelist_lock ->&meta->lock FD: 1 BD: 78 +...: _xmit_LOOPBACK FD: 23 BD: 85 .+.+: netpoll_srcu ->&rq->__lock FD: 14 BD: 101 +.-.: &in_dev->mc_tomb_lock ->pool_lock#2 ->&obj_hash[i].lock ->&c->lock ->&n->list_lock FD: 22 BD: 98 +.-.: &im->lock ->pool_lock#2 ->&c->lock ->&____s->seqcount#2 ->&____s->seqcount ->&n->list_lock ->&obj_hash[i].lock ->&zone->lock ->batched_entropy_u8.lock ->kfence_freelist_lock ->batched_entropy_u32.lock ->&base->lock ->init_task.mems_allowed_seq.seqcount ->&data->lock FD: 1 BD: 85 +.+.: cbs_list_lock FD: 25 BD: 84 +...: &net->ipv6.addrconf_hash_lock ->&obj_hash[i].lock ->&base->lock FD: 26 BD: 3941 +...: &ifa->lock ->batched_entropy_u32.lock ->crngs.lock ->&obj_hash[i].lock ->&base->lock FD: 52 BD: 3945 +.-.: &tb->tb6_lock ->&net->ipv6.fib6_walker_lock ->&____s->seqcount ->pool_lock#2 ->&c->lock ->nl_table_lock ->&obj_hash[i].lock ->nl_table_wait.lock ->rlock-AF_NETLINK ->rt6_exception_lock ->&n->list_lock ->&data->fib_event_queue_lock ->&zone->lock ->quarantine_lock ->&____s->seqcount#2 ->batched_entropy_u8.lock ->kfence_freelist_lock ->&meta->lock ->stock_lock ->(console_sem).lock ->&base->lock FD: 1 BD: 3946 ++.-: &net->ipv6.fib6_walker_lock FD: 457 BD: 82 +.+.: sk_lock-AF_INET ->slock-AF_INET ->&table->hash[i].lock ->&tcp_hashinfo.bhash[i].lock ->&h->lhash2[i].lock ->&queue->rskq_lock ->clock-AF_INET ->&obj_hash[i].lock ->&base->lock ->fs_reclaim ->&____s->seqcount ->pool_lock#2 ->&c->lock ->&mm->mmap_lock ->tk_core.seq.seqcount ->&sd->defer_lock ->rcu_node_0 ->&rq->__lock ->&hashinfo->ehash_locks[i] ->elock-AF_INET ->mmu_notifier_invalidate_range_start ->remove_cache_srcu ->&n->list_lock ->batched_entropy_u8.lock ->kfence_freelist_lock ->&meta->lock ->&____s->seqcount#8 ->once_mutex ->&pool->lock ->batched_entropy_u32.lock ->batched_entropy_u16.lock ->&ei->socket.wq.wait ->quarantine_lock ->&rcu_state.expedited_wq ->stock_lock ->&sb->s_type->i_lock_key#8 ->&dir->lock ->k-sk_lock-AF_INET/1 ->k-slock-AF_INET ->k-clock-AF_INET ->&xa->xa_lock#9 ->&fsnotify_mark_srcu ->k-sk_lock-AF_INET ->crngs.lock ->&token_hash[i].lock ->&msk->pm.lock ->&sighand->siglock ->&cfs_rq->removed.lock ->&dccp_hashinfo.bhash[i].lock ->&____s->seqcount#2 ->free_vmap_area_lock ->vmap_area_lock ->pcpu_alloc_mutex ->pack_mutex ->text_mutex ->&fp->aux->used_maps_mutex ->hrtimer_bases.lock ->&f->f_owner.lock ->&net->xfrm.xfrm_policy_lock ->&in_dev->mc_tomb_lock ->&im->lock ->krc.lock ->sctp_assocs_id_lock ->wlock-AF_INET ->&p->pi_lock ->(console_sem).lock ->l2tp_ip_lock ->&sctp_port_hashtable[i].lock ->lock ->&asoc->wait ->cpu_hotplug_lock ->&ping_table.lock ->&list->lock#24 ->_xmit_ETHER ->(&tw->tw_timer) ->&data->lock ->&mux->lock ->prog_idr_lock ->bpf_lock ->pcpu_alloc_mutex.wait_lock ->key#26 FD: 100 BD: 96 +.-.: slock-AF_INET ->&obj_hash[i].lock ->batched_entropy_u16.lock ->&tcp_hashinfo.bhash[i].lock ->&hashinfo->ehash_locks[i] ->tk_core.seq.seqcount ->(&req->rsk_timer) ->&base->lock ->&queue->rskq_lock ->pool_lock#2 ->&c->lock ->batched_entropy_u8.lock ->kfence_freelist_lock ->&____s->seqcount ->elock-AF_INET ->&sk->sk_lock.wq ->hrtimer_bases.lock ->batched_entropy_u32.lock ->&n->list_lock ->key#24 ->&____s->seqcount#2 ->quarantine_lock ->&meta->lock ->krc.lock ->&data->lock ->&dccp_hashinfo.bhash[i].lock ->stock_lock FD: 5 BD: 130 ++..: clock-AF_INET ->&obj_hash[i].lock ->&mux->rx_lock FD: 511 BD: 80 +.+.: sk_lock-AF_INET6 ->slock-AF_INET6 ->&table->hash[i].lock ->&____s->seqcount#8 ->batched_entropy_u32.lock ->&____s->seqcount ->pool_lock#2 ->&c->lock ->&obj_hash[i].lock ->batched_entropy_u16.lock ->&tcp_hashinfo.bhash[i].lock ->&h->lhash2[i].lock ->fs_reclaim ->&mm->mmap_lock ->once_lock ->&pool->lock ->rcu_node_0 ->&rq->__lock ->pgd_lock ->key ->pcpu_lock ->percpu_counters_lock ->&zone->lock ->tk_core.seq.seqcount ->&n->list_lock ->clock-AF_INET6 ->&list->lock#5 ->&rcu_state.expedited_wq ->remove_cache_srcu ->&____s->seqcount#2 ->&cfs_rq->removed.lock ->batched_entropy_u8.lock ->kfence_freelist_lock ->&sctp_port_hashtable[i].lock ->crngs.lock ->&base->lock ->&asoc->wait ->stock_lock ->krc.lock ->sctp_assocs_id_lock ->&list->lock#24 ->per_cpu_ptr(&cgroup_rstat_cpu_lock, cpu) ->&f->f_lock ->&hashinfo->ehash_locks[i] ->&f->f_owner.lock ->&ei->socket.wq.wait ->&sighand->siglock ->crypto_alg_sem ->&dccp_hashinfo.bhash[i].lock ->&queue->rskq_lock ->&ping_table.lock ->&ndev->lock ->acaddr_hash_lock ->&tb->tb6_lock ->&idev->mc_lock ->mmu_notifier_invalidate_range_start ->&sem->wait_lock ->&p->pi_lock ->&sd->defer_lock ->(console_sem).lock ->console_owner_lock ->console_owner ->&meta->lock ->&sb->s_type->i_lock_key#8 ->&dir->lock ->k-sk_lock-AF_INET6/1 ->k-slock-AF_INET6 ->k-clock-AF_INET6 ->&xa->xa_lock#9 ->&fsnotify_mark_srcu ->&msk->pm.lock ->elock-AF_INET6 ->quarantine_lock ->free_vmap_area_lock ->vmap_area_lock ->init_mm.page_table_lock ->pcpu_alloc_mutex ->cpu_hotplug_lock ->&net->sctp.addr_wq_lock ->tcp_md5sig_mutex ->wlock-AF_INET6 ->lock ->tcpv6_prot_mutex ->device_spinlock ->(kmod_concurrent_max).lock ->&x->wait#17 ->k-sk_lock-AF_INET6 ->running_helpers_waitq.lock ->(crypto_chain).rwsem ->&x->wait#21 ->(&timer.timer) ->&sw_ctx_tx->encrypt_compl_lock ->&list->lock#39 ->&token_hash[i].lock ->rlock-AF_INET6 ->key#26 ->sk_lock-AF_INET6/1 ->&newf->file_lock ->&sctp_ep_hashtable[i].lock ->&net->xfrm.xfrm_policy_lock ->&data->lock ->&policy->lock ->&list->lock#32 FD: 101 BD: 106 +.-.: slock-AF_INET6 ->&obj_hash[i].lock ->elock-AF_INET6 ->&sk->sk_lock.wq ->&tcp_hashinfo.bhash[i].lock ->pool_lock#2 ->&c->lock ->tk_core.seq.seqcount ->stock_lock ->&base->lock ->batched_entropy_u32.lock ->&____s->seqcount ->&n->list_lock ->key#24 ->&zone->lock ->&hashinfo->ehash_locks[i] ->batched_entropy_u16.lock ->&dccp_hashinfo.bhash[i].lock ->(&req->rsk_timer) ->&queue->rskq_lock ->clock-AF_INET6 ->&____s->seqcount#2 ->krc.lock ->wlock-AF_INET6 ->&list->lock#39 ->&list->lock#24 ->quarantine_lock ->&data->lock ->batched_entropy_u8.lock ->kfence_freelist_lock ->&meta->lock ->crngs.lock ->sctp_assocs_id_lock ->&asoc->wait FD: 33 BD: 131 ++--: clock-AF_INET6 ->pool_lock#2 ->&c->lock ->rds_tcp_tc_list_lock ->&cp->cp_lock ->&rm->m_rs_lock ->&obj_hash[i].lock ->&n->list_lock FD: 128 BD: 1 .+.+: kn->active#46 ->fs_reclaim ->&c->lock ->&kernfs_locks->open_file_mutex[count] FD: 445 BD: 78 ++++: dev_addr_sem ->net_rwsem ->&tn->lock ->&sdata->sec_mtx ->fs_reclaim ->pool_lock#2 ->nl_table_lock ->rlock-AF_NETLINK ->nl_table_wait.lock ->&tbl->lock ->&pn->hash_lock ->&obj_hash[i].lock ->input_pool.lock ->&rq->__lock ->&c->lock ->&n->list_lock ->rcu_node_0 ->&____s->seqcount ->&br->lock ->team->team_lock_key ->team->team_lock_key#3 ->team->team_lock_key#4 ->batched_entropy_u8.lock ->kfence_freelist_lock ->team->team_lock_key#6 ->team->team_lock_key#2 ->team->team_lock_key#5 ->_xmit_ETHER ->quarantine_lock ->remove_cache_srcu ->&hard_iface->bat_iv.ogm_buff_mutex ->&____s->seqcount#2 ->&rcu_state.expedited_wq ->&cfs_rq->removed.lock FD: 942 BD: 2 +.+.: nlk_cb_mutex-GENERIC ->fs_reclaim ->pool_lock#2 ->&____s->seqcount ->rtnl_mutex ->&rdev->wiphy.mtx ->rlock-AF_NETLINK ->&obj_hash[i].lock ->&c->lock ->&devlink->lock_key ->&devlink->lock_key#2 ->&devlink->lock_key#3 ->&devlink->lock_key#4 ->&devlink->lock_key#5 ->&devlink->lock_key#6 ->genl_mutex ->&n->list_lock ->&rq->__lock ->rcu_node_0 ->&____s->seqcount#2 ->genl_mutex.wait_lock ->&p->pi_lock ->&cfs_rq->removed.lock ->&lock->wait_lock ->rtnl_mutex.wait_lock ->remove_cache_srcu ->&rcu_state.expedited_wq ->&dir->lock#2 FD: 20 BD: 93 +.-.: &rdev->bss_lock ->pool_lock#2 ->&obj_hash[i].lock ->krc.lock ->&c->lock ->&____s->seqcount#2 ->&____s->seqcount ->&zone->lock ->&n->list_lock ->quarantine_lock ->&base->lock ->batched_entropy_u8.lock ->kfence_freelist_lock ->&meta->lock FD: 130 BD: 1 +.-.: (&net->sctp.addr_wq_timer) ->&net->sctp.addr_wq_lock FD: 13 BD: 78 ++..: lapb_list_lock ->pool_lock#2 ->&obj_hash[i].lock ->&base->lock ->&c->lock ->&n->list_lock FD: 1 BD: 78 ++.-: x25_neigh_list_lock FD: 1 BD: 78 +...: _xmit_SLIP FD: 14 BD: 1 +.-.: (&eql->timer) ->&eql->queue.lock ->&obj_hash[i].lock ->&base->lock FD: 5 BD: 81 +.-.: &eql->queue.lock ->&obj_hash[i].lock ->&____s->seqcount ->pool_lock#2 FD: 1 BD: 78 +...: &vi->refill_lock FD: 69 BD: 3866 +.-.: _xmit_ETHER#2 ->&obj_hash[i].lock ->pool_lock#2 ->&____s->seqcount ->quarantine_lock ->&meta->lock ->kfence_freelist_lock ->&base->lock ->&data->lock FD: 137 BD: 92 +.+.: &local->chanctx_mtx ->fs_reclaim ->pool_lock#2 ->&data->mutex ->&rq->__lock ->&c->lock ->&____s->seqcount#2 ->&n->list_lock ->&local->queue_stop_reason_lock ->&obj_hash[i].lock ->krc.lock ->nl_table_lock ->nl_table_wait.lock ->&rdev->bss_lock ->&____s->seqcount FD: 23 BD: 93 +.+.: &data->mutex ->&rq->__lock FD: 18 BD: 3896 +...: &local->filter_lock ->pool_lock#2 ->&obj_hash[i].lock ->krc.lock ->&c->lock ->&____s->seqcount#2 ->&____s->seqcount FD: 30 BD: 1 +.+.: (wq_completion)phy0 ->(work_completion)(&local->reconfig_filter) FD: 29 BD: 106 +.+.: (work_completion)(&local->reconfig_filter) ->&local->filter_lock ->&rq->__lock FD: 99 BD: 82 +.-.: &dev->tx_global_lock ->_xmit_ETHER#2 ->&obj_hash[i].lock ->&base->lock ->_xmit_NONE#2 ->&batadv_netdev_xmit_lock_key ->&vlan_netdev_xmit_lock_key ->&qdisc_xmit_lock_key ->_xmit_LOOPBACK#2 ->_xmit_NETROM ->&qdisc_xmit_lock_key#2 ->&qdisc_xmit_lock_key#3 ->_xmit_PIMREG#2 ->_xmit_IPGRE#2 ->_xmit_TUNNEL#2 ->_xmit_TUNNEL6#2 ->_xmit_SIT#2 ->&qdisc_xmit_lock_key#4 ->&qdisc_xmit_lock_key#5 FD: 35 BD: 3921 +.-.: &sch->q.lock ->tk_core.seq.seqcount ->batched_entropy_u64.lock ->hrtimer_bases.lock ->pool_lock#2 ->&obj_hash[i].lock ->crngs.lock ->batched_entropy_u16.lock ->&c->lock ->&n->list_lock ->&____s->seqcount#2 ->&____s->seqcount ->batched_entropy_u8.lock ->kfence_freelist_lock FD: 1 BD: 85 ....: class FD: 1 BD: 85 ....: (&tbl->proxy_timer) FD: 30 BD: 1 +.+.: (wq_completion)phy1 ->(work_completion)(&local->reconfig_filter) FD: 1 BD: 78 +...: _xmit_VOID FD: 1 BD: 98 ....: &____s->seqcount#8 FD: 8 BD: 3919 +.-.: &ul->lock ->pool_lock#2 ->&dir->lock#2 FD: 1 BD: 78 +...: _xmit_X25 FD: 14 BD: 79 +...: &lapbeth->up_lock ->&obj_hash[i].lock ->pool_lock#2 ->&data->lock FD: 74 BD: 79 +.-.: &lapb->lock ->&c->lock ->pool_lock#2 ->&obj_hash[i].lock ->&base->lock ->batched_entropy_u8.lock ->kfence_freelist_lock ->&n->list_lock ->&____s->seqcount ->&list->lock#13 ->&list->lock#14 ->&____s->seqcount#2 FD: 1 BD: 157 ....: &tty->ctrl.lock FD: 6 BD: 89 +.+.: fasync_lock ->&new->fa_lock ->&obj_hash[i].lock ->pool_lock#2 FD: 1 BD: 1 +.+.: &buf->lock FD: 1 BD: 7 ....: &tty->flow.lock FD: 3 BD: 171 +.+.: &(ei->i_block_reservation_lock) ->key#14 ->key#3 FD: 924 BD: 2 +.+.: (work_completion)(&work->work) ->devices_rwsem ->&obj_hash[i].lock ->&rq->__lock ->&meta->lock ->kfence_freelist_lock ->quarantine_lock FD: 890 BD: 2 +.+.: (work_completion)(&(&ifa->dad_work)->work) ->rtnl_mutex ->rtnl_mutex.wait_lock ->&p->pi_lock ->&rq->__lock ->rcu_node_0 ->&rcu_state.expedited_wq ->&cfs_rq->removed.lock ->&obj_hash[i].lock ->pool_lock#2 FD: 1 BD: 3946 +.-.: rt6_exception_lock FD: 263 BD: 4 +.+.: &ldata->atomic_read_lock ->&tty->termios_rwsem ->(work_completion)(&buf->work) ->&rq->__lock FD: 1 BD: 5 +.+.: (work_completion)(&buf->work) FD: 25 BD: 1 ..-.: &(&idev->mc_dad_work)->timer FD: 225 BD: 1 +.+.: (wq_completion)mld ->(work_completion)(&(&idev->mc_dad_work)->work) ->(work_completion)(&(&idev->mc_ifc_work)->work) ->&rq->__lock ->(work_completion)(&(&idev->mc_query_work)->work) FD: 221 BD: 2 +.+.: (work_completion)(&(&idev->mc_dad_work)->work) ->&idev->mc_lock ->&rq->__lock FD: 81 BD: 78 +...: dev->qdisc_tx_busylock ?: &qdisc_tx_busylock ->_xmit_ETHER#2 ->_xmit_SLIP#2 ->_xmit_NETROM ->&obj_hash[i].lock ->pool_lock#2 ->&____s->seqcount ->&c->lock ->&sch->q.lock ->quarantine_lock ->&meta->lock ->kfence_freelist_lock ->&data->lock FD: 894 BD: 6 +.+.: &net->packet.sklist_lock ->&rq->__lock ->clock-AF_PACKET ->rtnl_mutex ->&po->pg_vec_lock ->fanout_mutex FD: 245 BD: 3 +.+.: sk_lock-AF_PACKET ->slock-AF_PACKET ->&po->bind_lock ->rcu_node_0 ->&obj_hash[i].lock ->&rnp->exp_wq[0] ->&rq->__lock ->&mm->mmap_lock ->fs_reclaim ->pool_lock#2 ->free_vmap_area_lock ->vmap_area_lock ->&____s->seqcount ->&c->lock ->pcpu_alloc_mutex ->pack_mutex ->batched_entropy_u32.lock ->text_mutex ->&fp->aux->used_maps_mutex ->&rnp->exp_wq[2] ->&rnp->exp_wq[3] ->&n->list_lock ->&zone->lock ->&po->pg_vec_lock ->rcu_state.exp_mutex.wait_lock ->&p->pi_lock ->&rnp->exp_lock ->rcu_state.exp_mutex ->stock_lock ->&cfs_rq->removed.lock ->pgd_lock ->key ->pcpu_lock ->percpu_counters_lock FD: 1 BD: 4 +...: slock-AF_PACKET FD: 17 BD: 9 +.+.: &po->bind_lock ->ptype_lock ->pool_lock#2 ->&dir->lock#2 ->&match->lock ->&obj_hash[i].lock FD: 1 BD: 4001 +.-.: rlock-AF_PACKET FD: 1 BD: 156 +...: wlock-AF_PACKET FD: 2 BD: 4562 ..-.: &pl->lock ->key#12 FD: 25 BD: 1 ..-.: &(&idev->mc_ifc_work)->timer FD: 221 BD: 2 +.+.: (work_completion)(&(&idev->mc_ifc_work)->work) ->&idev->mc_lock ->&rq->__lock FD: 15 BD: 3878 +.-.: &ul->lock#2 ->pool_lock#2 ->&dir->lock#2 ->&____s->seqcount ->&obj_hash[i].lock ->&c->lock ->&n->list_lock ->&____s->seqcount#2 FD: 18 BD: 3917 ++--: &n->lock ->&obj_hash[i].lock ->&base->lock ->&c->lock ->pool_lock#2 ->&(&n->ha_lock)->lock ->&____s->seqcount#9 ->&____s->seqcount#2 ->&____s->seqcount ->&n->list_lock ->batched_entropy_u8.lock ->kfence_freelist_lock ->quarantine_lock ->&meta->lock ->&data->lock FD: 1 BD: 3919 +.--: &____s->seqcount#9 FD: 35 BD: 2 +.+.: (work_completion)(&w->work)#2 ->pool_lock#2 ->&dir->lock ->&obj_hash[i].lock ->nf_conntrack_mutex ->&meta->lock ->kfence_freelist_lock ->&rq->__lock ->nf_conntrack_mutex.wait_lock ->&p->pi_lock ->&cfs_rq->removed.lock FD: 1 BD: 3921 ...-: &____s->seqcount#10 FD: 25 BD: 1 ..-.: &(&ifa->dad_work)->timer FD: 130 BD: 7 +.+.: fanout_mutex ->fs_reclaim ->pool_lock#2 ->&po->bind_lock ->&c->lock ->&n->list_lock ->&rq->__lock FD: 1 BD: 7 ++..: clock-AF_PACKET FD: 1 BD: 3 ..-.: elock-AF_PACKET FD: 38 BD: 125 +.-.: &ct->lock ->(console_sem).lock FD: 25 BD: 1 ..-.: &(&tbl->gc_work)->timer FD: 51 BD: 2 +.+.: (work_completion)(&(&tbl->gc_work)->work) ->&tbl->lock ->&rq->__lock ->rcu_node_0 ->&rcu_state.expedited_wq FD: 75 BD: 1 +.-.: (&lapb->t1timer) ->&lapb->lock FD: 100 BD: 1 +.-.: (&dev->watchdog_timer) ->&dev->tx_global_lock FD: 25 BD: 1 ..-.: &(&dm_bufio_cleanup_old_work)->timer FD: 25 BD: 1 +.+.: (wq_completion)dm_bufio_cache ->(work_completion)(&(&dm_bufio_cleanup_old_work)->work) FD: 24 BD: 2 +.+.: (work_completion)(&(&dm_bufio_cleanup_old_work)->work) ->dm_bufio_clients_lock ->&obj_hash[i].lock ->&base->lock FD: 5 BD: 3914 +.-.: &nf_conntrack_locks[i] ->&nf_conntrack_locks[i]/1 ->batched_entropy_u8.lock FD: 4 BD: 3915 +.-.: &nf_conntrack_locks[i]/1 ->batched_entropy_u8.lock FD: 1 BD: 129 +.-.: &hashinfo->ehash_locks[i] FD: 2 BD: 3918 +.-.: &(&n->ha_lock)->lock ->&____s->seqcount#9 FD: 1 BD: 3914 +.-.: lock#8 FD: 1 BD: 3916 ..-.: id_table_lock FD: 1 BD: 121 ..-.: (&req->rsk_timer) FD: 1 BD: 121 +.-.: &queue->rskq_lock FD: 8 BD: 115 +.-.: tcp_metrics_lock ->pool_lock#2 ->&c->lock ->&n->list_lock FD: 98 BD: 88 +.-.: slock-AF_INET/1 ->tk_core.seq.seqcount ->pool_lock#2 ->&obj_hash[i].lock ->&base->lock ->&c->lock ->&n->list_lock ->&____s->seqcount ->&meta->lock ->kfence_freelist_lock ->&hashinfo->ehash_locks[i] ->&tcp_hashinfo.bhash[i].lock ->elock-AF_INET ->&zone->lock ->quarantine_lock ->batched_entropy_u8.lock ->&sctp_ep_hashtable[i].lock ->clock-AF_INET ->&____s->seqcount#2 ->&f->f_owner.lock ->hrtimer_bases.lock ->key#25 ->krc.lock ->&sctp_port_hashtable[i].lock ->&data->lock ->batched_entropy_u32.lock FD: 1 BD: 85 +.-.: &sd->defer_lock FD: 125 BD: 1 +.-.: (&icsk->icsk_delack_timer) ->slock-AF_INET ->slock-AF_INET6 ->k-slock-AF_INET6 FD: 123 BD: 1 +.-.: (&icsk->icsk_retransmit_timer) ->slock-AF_INET ->slock-AF_INET6 ->&obj_hash[i].lock ->pool_lock#2 ->&dir->lock ->&____s->seqcount ->stock_lock ->&n->list_lock ->&c->lock ->per_cpu_ptr(&cgroup_rstat_cpu_lock, cpu) FD: 1 BD: 108 ..-.: elock-AF_INET FD: 1 BD: 173 ....: key#14 FD: 25 BD: 1 ..-.: drivers/regulator/core.c:6262 FD: 4 BD: 2 +.+.: (regulator_init_complete_work).work ->&k->list_lock ->&k->k_lock FD: 84 BD: 170 +.+.: &sbi->s_orphan_lock ->&rq->__lock ->&ei->i_raw_lock ->&ret->b_state_lock ->&lock->wait_lock ->rcu_node_0 ->&mapping->private_lock ->&cfs_rq->removed.lock ->&obj_hash[i].lock ->&rcu_state.expedited_wq ->mmu_notifier_invalidate_range_start ->&____s->seqcount ->pool_lock#2 ->bit_wait_table + i ->pgd_lock ->stock_lock ->key ->pcpu_lock ->percpu_counters_lock FD: 128 BD: 1 .+.+: kn->active#47 ->fs_reclaim ->&kernfs_locks->open_file_mutex[count] FD: 1 BD: 1 +.+.: &futex_queues[i].lock FD: 1 BD: 4 ....: &on->poll FD: 1 BD: 4 +.+.: module_mutex FD: 3 BD: 83 +.+.: once_mutex ->crngs.lock FD: 231 BD: 1 .+.+: sb_writers#9 ->&attr->mutex ->&mm->mmap_lock FD: 230 BD: 2 +.+.: &attr->mutex ->&mm->mmap_lock FD: 141 BD: 1 +.+.: &type->s_umount_key#44/1 ->fs_reclaim ->pool_lock#2 ->pcpu_alloc_mutex ->shrinker_rwsem ->list_lrus_mutex ->&c->lock ->sb_lock ->&root->kernfs_rwsem ->&sb->s_type->i_lock_key#30 ->&root->kernfs_supers_rwsem ->&dentry->d_lock FD: 40 BD: 4532 +.+.: &sb->s_type->i_lock_key#30 ->&dentry->d_lock FD: 933 BD: 2 .+.+: sb_writers#10 ->mount_lock ->&type->i_mutex_dir_key#6 ->fs_reclaim ->&mm->mmap_lock ->&of->mutex ->&obj_hash[i].lock ->&type->i_mutex_dir_key#6/1 ->&rq->__lock ->&root->kernfs_iattr_rwsem ->&dentry->d_lock ->tomoyo_ss ->&sb->s_type->i_mutex_key#15 ->iattr_mutex ->rcu_node_0 ->&rcu_state.expedited_wq ->&c->lock ->&n->list_lock ->&xattrs->lock ->remove_cache_srcu ->(console_sem).lock ->batched_entropy_u8.lock ->kfence_freelist_lock ->&meta->lock FD: 112 BD: 3 ++++: &type->i_mutex_dir_key#6 ->tomoyo_ss ->tk_core.seq.seqcount ->&root->kernfs_iattr_rwsem ->rename_lock.seqcount ->fs_reclaim ->&dentry->d_lock ->&root->kernfs_rwsem ->&sb->s_type->i_lock_key#30 ->pool_lock#2 ->&xa->xa_lock#4 ->&obj_hash[i].lock ->stock_lock ->&c->lock ->&____s->seqcount#2 ->&____s->seqcount ->&n->list_lock ->&rq->__lock FD: 128 BD: 1 ++++: kn->active#48 ->fs_reclaim ->&kernfs_locks->open_file_mutex[count] ->pool_lock#2 ->stock_lock ->&c->lock ->&lock->wait_lock ->&rq->__lock ->&p->pi_lock ->&n->list_lock ->rcu_node_0 ->&____s->seqcount#2 ->&____s->seqcount ->&rcu_state.expedited_wq ->remove_cache_srcu FD: 128 BD: 3 +.+.: &sb->s_type->i_mutex_key#15 ->tk_core.seq.seqcount ->&root->kernfs_iattr_rwsem FD: 141 BD: 1 +.+.: &type->s_umount_key#45/1 ->fs_reclaim ->pool_lock#2 ->pcpu_alloc_mutex ->shrinker_rwsem ->&____s->seqcount ->&obj_hash[i].lock ->list_lrus_mutex ->sb_lock ->&root->kernfs_rwsem ->&sb->s_type->i_lock_key#31 ->&root->kernfs_supers_rwsem ->&dentry->d_lock ->&c->lock ->&n->list_lock FD: 40 BD: 4544 +.+.: &sb->s_type->i_lock_key#31 ->&dentry->d_lock FD: 127 BD: 1 ++++: &type->s_umount_key#46 ->shrinker_rwsem ->percpu_ref_switch_lock ->&root->kernfs_supers_rwsem ->rename_lock.seqcount ->&dentry->d_lock ->&sb->s_type->i_lock_key#31 ->&s->s_inode_list_lock ->&xa->xa_lock#9 ->inode_hash_lock ->&obj_hash[i].lock ->pool_lock#2 ->&fsnotify_mark_srcu ->&lru->node[i].lock ->&rq->__lock FD: 900 BD: 2 +.+.: (work_completion)(&cgrp->bpf.release_work) ->cgroup_mutex ->percpu_ref_switch_lock ->&obj_hash[i].lock ->pool_lock#2 FD: 906 BD: 1 +.+.: (wq_completion)cgroup_destroy ->(work_completion)(&css->destroy_work) ->(work_completion)(&(&css->destroy_rwork)->work) FD: 900 BD: 2 +.+.: (work_completion)(&css->destroy_work) ->cgroup_mutex ->&obj_hash[i].lock ->pool_lock#2 FD: 904 BD: 2 +.+.: (work_completion)(&(&css->destroy_rwork)->work) ->percpu_ref_switch_lock ->&obj_hash[i].lock ->pool_lock#2 ->&cgrp->pidlist_mutex ->(wq_completion)cgroup_pidlist_destroy ->&wq->mutex ->(work_completion)(&cgrp->release_agent_work) ->cgroup_mutex ->cgroup_rstat_lock ->pcpu_lock ->&root->kernfs_rwsem ->kernfs_idr_lock ->cgroup_mutex.wait_lock ->&p->pi_lock FD: 130 BD: 12 +.+.: &cgrp->pidlist_mutex ->&rq->__lock ->css_set_lock ->fs_reclaim ->pool_lock#2 ->&obj_hash[i].lock ->&base->lock ->&c->lock FD: 132 BD: 3 +.+.: (wq_completion)cgroup_pidlist_destroy ->(work_completion)(&(&l->destroy_dwork)->work) FD: 1 BD: 3 +.+.: (work_completion)(&cgrp->release_agent_work) FD: 1 BD: 18 +.+.: cgroup_mutex.wait_lock FD: 932 BD: 2 .+.+: sb_writers#11 ->mount_lock ->&type->i_mutex_dir_key#7 ->fs_reclaim ->&mm->mmap_lock ->&of->mutex ->&obj_hash[i].lock ->&type->i_mutex_dir_key#7/1 ->&c->lock ->&rq->__lock ->&____s->seqcount ->stock_lock ->&n->list_lock ->rcu_node_0 ->&rcu_state.expedited_wq ->pool_lock#2 ->&p->lock ->tk_core.seq.seqcount ->&sb->s_type->i_lock_key#31 ->&wb->list_lock ->&____s->seqcount#2 ->batched_entropy_u8.lock ->kfence_freelist_lock ->&meta->lock ->remove_cache_srcu FD: 112 BD: 3 ++++: &type->i_mutex_dir_key#7 ->tomoyo_ss ->tk_core.seq.seqcount ->&root->kernfs_iattr_rwsem ->rename_lock.seqcount ->fs_reclaim ->&dentry->d_lock ->&root->kernfs_rwsem ->&sb->s_type->i_lock_key#31 ->pool_lock#2 ->&xa->xa_lock#4 ->&obj_hash[i].lock ->stock_lock ->&c->lock ->&n->list_lock ->&____s->seqcount ->&____s->seqcount#2 ->&rq->__lock FD: 1 BD: 18 +.+.: &dom->lock FD: 128 BD: 1 .+.+: kn->active#49 ->fs_reclaim ->&kernfs_locks->open_file_mutex[count] FD: 23 BD: 1 +.+.: &sb->s_type->i_mutex_key#16 ->&rq->__lock FD: 289 BD: 1 .+.+: kn->active#50 ->fs_reclaim ->&kernfs_locks->open_file_mutex[count] ->cpu_hotplug_lock FD: 42 BD: 3 +.+.: &type->s_umount_key#47 ->sb_lock ->&dentry->d_lock FD: 146 BD: 2 +.+.: &sb->s_type->i_mutex_key#17 ->namespace_sem ->rename_lock.seqcount ->fs_reclaim ->pool_lock#2 ->&dentry->d_lock ->mmu_notifier_invalidate_range_start ->&sb->s_type->i_lock_key#26 ->&s->s_inode_list_lock ->tk_core.seq.seqcount ->pin_fs_lock ->sb_lock ->&type->s_umount_key#47 ->mnt_id_ida.xa_lock ->pcpu_alloc_mutex ->mount_lock ->&obj_hash[i].lock ->entries_lock ->&c->lock ->&____s->seqcount FD: 244 BD: 1 .+.+: sb_writers#12 ->fs_reclaim ->&c->lock ->pool_lock#2 ->&mm->mmap_lock ->&sb->s_type->i_mutex_key#17 FD: 1 BD: 83 ++..: &pn->hash_lock FD: 53 BD: 6 +.-.: &net->ipv6.fib6_gc_lock ->&obj_hash[i].lock FD: 1 BD: 78 +...: _xmit_IEEE802154 FD: 30 BD: 3 ..-.: &ei->i_completed_io_lock FD: 148 BD: 1 +.+.: (wq_completion)ext4-rsv-conversion ->(work_completion)(&ei->i_rsv_conversion_work) ->&rq->__lock FD: 147 BD: 2 +.+.: (work_completion)(&ei->i_rsv_conversion_work) ->&ei->i_completed_io_lock ->&journal->j_state_lock ->jbd2_handle ->&obj_hash[i].lock ->pool_lock#2 ->&ext4__ioend_wq[i] ->&ret->b_uptodate_lock ->&folio_wait_table[i] ->rcu_node_0 ->&rq->__lock ->&cfs_rq->removed.lock ->quarantine_lock ->&rcu_state.expedited_wq ->mmu_notifier_invalidate_range_start ->remove_cache_srcu ->&meta->lock ->kfence_freelist_lock ->&c->lock ->&base->lock ->batched_entropy_u8.lock ->&n->list_lock ->&____s->seqcount#2 ->&____s->seqcount FD: 1 BD: 174 ....: &journal->j_wait_reserved FD: 1 BD: 167 ....: &ext4__ioend_wq[i] FD: 1 BD: 4535 ....: key#15 FD: 126 BD: 8 +.+.: swap_cgroup_mutex ->fs_reclaim ->&____s->seqcount ->pool_lock#2 FD: 138 BD: 8 +.+.: swapon_mutex ->fs_reclaim ->&c->lock ->pool_lock#2 ->swap_lock ->percpu_ref_switch_lock ->(console_sem).lock FD: 2 BD: 4518 +.+.: &p->lock#2 ->swap_avail_lock FD: 1 BD: 4519 +.+.: swap_avail_lock FD: 1 BD: 8 ....: proc_poll_wait.lock FD: 288 BD: 1 +.+.: swap_slots_cache_enable_mutex ->cpu_hotplug_lock ->swap_lock FD: 23 BD: 4515 +.+.: swap_slots_cache_mutex ->&rq->__lock FD: 88 BD: 173 +.+.: &lg->lg_mutex ->&ei->i_prealloc_lock ->mmu_notifier_invalidate_range_start ->pool_lock#2 ->&mapping->private_lock ->&ret->b_state_lock ->&journal->j_revoke_lock ->&pa->pa_lock ->&lg->lg_prealloc_lock ->&____s->seqcount ->&c->lock ->&rq->__lock ->bit_wait_table + i ->key#3 ->&obj_hash[i].lock FD: 1 BD: 174 +.+.: &pa->pa_lock FD: 1 BD: 174 +.+.: &lg->lg_prealloc_lock FD: 28 BD: 1 +.-.: (&timer) ->&obj_hash[i].lock ->&base->lock ->&txlock ->&txwq FD: 1 BD: 4007 ..-.: &list->lock#5 FD: 6 BD: 79 +...: _xmit_SLIP#2 ->&eql->queue.lock FD: 42 BD: 84 +...: _xmit_NETROM ->(console_sem).lock ->console_owner_lock ->console_owner ->&obj_hash[i].lock ->pool_lock#2 ->&rdev->wpan_phy.queue_lock ->&rdev->wpan_phy.sync_txq ->&data->lock FD: 15 BD: 1 +...: _xmit_X25#2 ->&lapbeth->up_lock FD: 24 BD: 180 ..-.: &rq_wait->wait ->&p->pi_lock FD: 79 BD: 1 +.-.: (&n->timer) ->&n->lock ->pool_lock#2 ->&dir->lock#2 ->&ul->lock#2 ->&obj_hash[i].lock ->&c->lock ->icmp_global.lock ->&n->list_lock ->nl_table_lock ->nl_table_wait.lock ->&dir->lock ->stock_lock ->&meta->lock ->kfence_freelist_lock ->quarantine_lock ->batched_entropy_u8.lock ->&____s->seqcount#2 ->&____s->seqcount ->&data->lock FD: 25 BD: 1 ..-.: net/wireless/reg.c:236 FD: 890 BD: 2 +.+.: (reg_check_chans).work ->rtnl_mutex ->rtnl_mutex.wait_lock ->&p->pi_lock ->&rq->__lock FD: 25 BD: 1 ..-.: &(&hctx->run_work)->timer FD: 25 BD: 1 ..-.: net/wireless/reg.c:533 FD: 890 BD: 2 +.+.: (crda_timeout).work ->rtnl_mutex ->rtnl_mutex.wait_lock ->&p->pi_lock ->&rq->__lock FD: 101 BD: 1 +.-.: (&sk->sk_timer) ->slock-AF_INET FD: 1 BD: 222 ....: &newf->resize_wait FD: 8 BD: 156 ..-.: &kcov->lock ->kcov_remote_lock FD: 154 BD: 1 +.+.: pid_caches_mutex ->slab_mutex FD: 42 BD: 1 +.+.: &type->s_umount_key#48 ->sb_lock ->&dentry->d_lock FD: 140 BD: 2 ++++: &sb->s_type->i_mutex_key#18 ->namespace_sem ->&dentry->d_lock ->tk_core.seq.seqcount ->&rq->__lock FD: 1 BD: 25 ++++: hci_sk_list.lock FD: 1 BD: 1 +.+.: (work_completion)(&(&data->open_timeout)->work) FD: 313 BD: 1 +.+.: &data->open_mutex ->fs_reclaim ->pool_lock#2 ->&____s->seqcount ->&obj_hash[i].lock ->&x->wait#9 ->hci_index_ida.xa_lock ->cpu_hotplug_lock ->wq_pool_mutex ->pin_fs_lock ->&sb->s_type->i_mutex_key#3 ->&k->list_lock ->gdp_mutex ->lock ->&root->kernfs_rwsem ->bus_type_sem ->sysfs_symlink_target_lock ->&c->lock ->&dev->power.lock ->dpm_list_mtx ->uevent_sock_mutex ->&n->list_lock ->uevent_sock_mutex.wait_lock ->&p->pi_lock ->subsys mutex#81 ->&dev->devres_lock ->triggers_list_lock ->leds_list_lock ->rfkill_global_mutex ->&rfkill->lock ->hci_dev_list_lock ->tk_core.seq.seqcount ->hci_sk_list.lock ->(pm_chain_head).rwsem ->&list->lock#6 ->&data->read_wait ->rfkill_global_mutex.wait_lock ->&rq->__lock FD: 1 BD: 2 ....: hci_index_ida.xa_lock FD: 1 BD: 227 +.+.: uevent_sock_mutex.wait_lock FD: 3 BD: 24 +.+.: subsys mutex#81 ->&k->k_lock FD: 14 BD: 16 ++++: hci_dev_list_lock ->pool_lock#2 ->tk_core.seq.seqcount ->rlock-AF_BLUETOOTH ->&____s->seqcount ->&obj_hash[i].lock ->&c->lock ->&n->list_lock FD: 1 BD: 19 ....: &list->lock#6 FD: 24 BD: 19 ....: &data->read_wait ->&p->pi_lock FD: 328 BD: 1 +.+.: (wq_completion)hci0 ->(work_completion)(&hdev->power_on) ->(work_completion)(&hdev->cmd_sync_work) FD: 326 BD: 7 +.+.: (work_completion)(&hdev->power_on) ->&hdev->req_lock ->fs_reclaim ->pool_lock#2 ->tk_core.seq.seqcount ->hci_sk_list.lock ->&obj_hash[i].lock FD: 325 BD: 9 +.+.: &hdev->req_lock ->&obj_hash[i].lock ->&list->lock#7 ->&c->lock ->pool_lock#2 ->&list->lock#8 ->&hdev->req_wait_q ->&base->lock ->&rq->__lock ->(&timer.timer) ->batched_entropy_u8.lock ->kfence_freelist_lock ->&____s->seqcount ->tk_core.seq.seqcount ->hci_sk_list.lock ->&hdev->lock ->&n->list_lock ->&lock->wait_lock ->&p->pi_lock ->(work_completion)(&(&hdev->interleave_scan)->work) ->hci_dev_list_lock ->(work_completion)(&hdev->tx_work) ->(work_completion)(&hdev->rx_work) ->(work_completion)(&(&hdev->rpa_expired)->work) ->&wq->mutex ->(wq_completion)hci0#2 ->&data->lock ->&cfs_rq->removed.lock ->&msft->filter_lock ->&list->lock#6 ->(work_completion)(&hdev->cmd_work) ->(work_completion)(&(&hdev->cmd_timer)->work) FD: 1 BD: 25 ....: &list->lock#7 FD: 1 BD: 10 ....: &list->lock#8 FD: 24 BD: 17 ....: &hdev->req_wait_q ->&p->pi_lock FD: 313 BD: 3 +.+.: sk_lock-AF_BLUETOOTH-BTPROTO_HCI ->slock-AF_BLUETOOTH-BTPROTO_HCI ->sock_cookie_ida.xa_lock ->&p->alloc_lock ->pool_lock#2 ->tk_core.seq.seqcount ->hci_sk_list.lock ->&obj_hash[i].lock ->&c->lock ->&____s->seqcount ->clock-AF_BLUETOOTH ->&mm->mmap_lock ->mgmt_chan_list_lock ->&rq->__lock ->&n->list_lock ->quarantine_lock ->hci_dev_list_lock ->rlock-AF_BLUETOOTH ->&data->lock FD: 1 BD: 4 +...: slock-AF_BLUETOOTH-BTPROTO_HCI FD: 1 BD: 4 ....: sock_cookie_ida.xa_lock FD: 318 BD: 10 +.+.: (wq_completion)hci0#2 ->(work_completion)(&hdev->cmd_work) ->(work_completion)(&hdev->rx_work) ->(work_completion)(&hdev->tx_work) ->(work_completion)(&conn->pending_rx_work) ->(work_completion)(&(&hdev->cmd_timer)->work) ->(work_completion)(&(&conn->disc_work)->work) FD: 129 BD: 16 +.+.: (work_completion)(&hdev->cmd_work) ->&list->lock#7 ->fs_reclaim ->pool_lock#2 ->tk_core.seq.seqcount ->&list->lock#6 ->&data->read_wait ->&obj_hash[i].lock ->&c->lock ->&____s->seqcount ->&rq->__lock ->batched_entropy_u8.lock ->kfence_freelist_lock ->&meta->lock FD: 328 BD: 1 +.+.: (wq_completion)hci1 ->(work_completion)(&hdev->power_on) ->(work_completion)(&hdev->cmd_sync_work) FD: 318 BD: 1 +.+.: (wq_completion)hci1#2 ->(work_completion)(&hdev->cmd_work) ->(work_completion)(&hdev->rx_work) ->(work_completion)(&hdev->tx_work) ->(work_completion)(&conn->pending_rx_work) ->(work_completion)(&(&hdev->cmd_timer)->work) ->(work_completion)(&(&conn->disc_work)->work) FD: 311 BD: 16 +.+.: (work_completion)(&hdev->rx_work) ->&list->lock#7 ->lock#6 ->fs_reclaim ->pool_lock#2 ->free_vmap_area_lock ->vmap_area_lock ->&____s->seqcount ->init_mm.page_table_lock ->&c->lock ->&hdev->lock ->(console_sem).lock ->console_owner_lock ->console_owner ->&rq->__lock ->&obj_hash[i].lock ->&hdev->req_wait_q ->&base->lock ->chan_list_lock FD: 328 BD: 1 +.+.: (wq_completion)hci2 ->(work_completion)(&hdev->power_on) ->(work_completion)(&hdev->cmd_sync_work) FD: 302 BD: 22 +.+.: &hdev->lock ->fs_reclaim ->pool_lock#2 ->&obj_hash[i].lock ->&x->wait#9 ->&k->list_lock ->lock ->&root->kernfs_rwsem ->bus_type_sem ->sysfs_symlink_target_lock ->&____s->seqcount ->&c->lock ->&dev->power.lock ->dpm_list_mtx ->uevent_sock_mutex ->&k->k_lock ->subsys mutex#81 ->&list->lock#7 ->&hdev->unregister_lock ->hci_cb_list_lock ->&base->lock ->tk_core.seq.seqcount ->hci_sk_list.lock ->&n->list_lock ->&rq->__lock ->&conn->chan_lock ->(work_completion)(&(&conn->disc_work)->work) ->(work_completion)(&(&conn->auto_accept_work)->work) ->(work_completion)(&(&conn->idle_work)->work) ->&list->lock#9 ->rcu_node_0 ->&rnp->exp_wq[1] ->hci_cb_list_lock.wait_lock ->&p->pi_lock ->rcu_state.exp_mutex.wait_lock ->dev_pm_qos_sysfs_mtx ->kernfs_idr_lock ->deferred_probe_mutex ->device_links_lock ->mmu_notifier_invalidate_range_start ->&rnp->exp_lock ->rcu_state.exp_mutex ->&rnp->exp_wq[2] ->uevent_sock_mutex.wait_lock ->sk_lock-AF_BLUETOOTH-BTPROTO_SCO ->slock-AF_BLUETOOTH-BTPROTO_SCO ->&rnp->exp_wq[3] ->&lock->wait_lock ->rlock-AF_BLUETOOTH ->&data->lock FD: 328 BD: 1 +.+.: (wq_completion)hci3 ->(work_completion)(&hdev->power_on) ->(work_completion)(&hdev->cmd_sync_work) FD: 318 BD: 1 +.+.: (wq_completion)hci2#2 ->(work_completion)(&hdev->cmd_work) ->(work_completion)(&hdev->rx_work) ->(work_completion)(&hdev->tx_work) ->(work_completion)(&conn->pending_rx_work) ->(work_completion)(&(&hdev->cmd_timer)->work) ->(work_completion)(&(&conn->disc_work)->work) FD: 328 BD: 1 +.+.: (wq_completion)hci4 ->(work_completion)(&hdev->power_on) ->(work_completion)(&hdev->cmd_sync_work) FD: 318 BD: 1 +.+.: (wq_completion)hci3#2 ->(work_completion)(&hdev->cmd_work) ->(work_completion)(&hdev->rx_work) ->(work_completion)(&hdev->tx_work) ->(work_completion)(&conn->pending_rx_work) ->(work_completion)(&(&hdev->cmd_timer)->work) ->(work_completion)(&(&conn->disc_work)->work) FD: 318 BD: 1 +.+.: (wq_completion)hci4#2 ->(work_completion)(&hdev->cmd_work) ->(work_completion)(&hdev->rx_work) ->(work_completion)(&hdev->tx_work) ->(work_completion)(&conn->pending_rx_work) ->(work_completion)(&(&hdev->cmd_timer)->work) ->(work_completion)(&(&conn->disc_work)->work) FD: 328 BD: 1 +.+.: (wq_completion)hci5 ->(work_completion)(&hdev->power_on) ->(work_completion)(&hdev->cmd_sync_work) FD: 318 BD: 1 +.+.: (wq_completion)hci5#2 ->(work_completion)(&hdev->cmd_work) ->(work_completion)(&hdev->rx_work) ->(work_completion)(&hdev->tx_work) ->(work_completion)(&conn->pending_rx_work) ->(work_completion)(&(&hdev->cmd_timer)->work) ->(work_completion)(&(&conn->disc_work)->work) FD: 1 BD: 29 +...: clock-AF_BLUETOOTH FD: 1 BD: 33 ....: rlock-AF_BLUETOOTH FD: 1 BD: 29 ....: wlock-AF_BLUETOOTH FD: 127 BD: 24 +.+.: &hdev->unregister_lock ->fs_reclaim ->pool_lock#2 ->&hdev->cmd_sync_work_lock ->&c->lock ->remove_cache_srcu ->&rq->__lock ->rcu_node_0 ->&rcu_state.expedited_wq FD: 1 BD: 25 +.+.: &hdev->cmd_sync_work_lock FD: 1 BD: 24 +.+.: &conn->ident_lock FD: 1 BD: 25 ....: &list->lock#9 FD: 41 BD: 27 +.+.: &conn->chan_lock ->&chan->lock/1 ->&lock->wait_lock ->&rq->__lock FD: 326 BD: 7 +.+.: (work_completion)(&hdev->cmd_sync_work) ->&hdev->cmd_sync_work_lock ->&hdev->req_lock ->&obj_hash[i].lock ->pool_lock#2 FD: 30 BD: 16 +.+.: (work_completion)(&hdev->tx_work) ->&list->lock#9 ->tk_core.seq.seqcount ->&list->lock#6 ->&data->read_wait ->&list->lock#7 FD: 2 BD: 16 +.+.: (work_completion)(&conn->pending_rx_work) ->&list->lock#10 FD: 1 BD: 25 ....: &list->lock#10 FD: 23 BD: 1 +.+.: &sb->s_type->i_mutex_key#19 ->&rq->__lock FD: 1 BD: 1 +.+.: &undo_list->lock FD: 1 BD: 78 +...: &nr_netdev_addr_lock_key FD: 1 BD: 78 +...: listen_lock FD: 2 BD: 13 +.+.: rdma_nets.xa_lock ->pool_lock#2 FD: 1 BD: 4 +.+.: &____s->seqcount#11 FD: 2 BD: 3 +.+.: &(&net->ipv4.ping_group_range.lock)->lock ->&____s->seqcount#11 FD: 2 BD: 78 +.+.: &r->consumer_lock ->&r->producer_lock FD: 1 BD: 3839 +.-.: &r->producer_lock FD: 18 BD: 3844 +...: &bridge_netdev_addr_lock_key ->pool_lock#2 ->&c->lock ->&obj_hash[i].lock ->krc.lock ->&n->list_lock ->&____s->seqcount#2 ->&____s->seqcount ->batched_entropy_u8.lock ->kfence_freelist_lock ->&zone->lock FD: 64 BD: 83 +.-.: &br->hash_lock ->&____s->seqcount ->pool_lock#2 ->&c->lock ->nl_table_lock ->&obj_hash[i].lock ->nl_table_wait.lock ->&n->list_lock ->batched_entropy_u8.lock ->kfence_freelist_lock ->&meta->lock ->&____s->seqcount#2 ->&zone->lock ->rlock-AF_NETLINK ->&dev_addr_list_lock_key#2/3 ->&dev_addr_list_lock_key#2/1 ->quarantine_lock FD: 128 BD: 79 +.+.: j1939_netdev_lock ->fs_reclaim ->pool_lock#2 ->&net->can.rcvlists_lock ->&obj_hash[i].lock ->&priv->lock FD: 9 BD: 3836 +...: &dev_addr_list_lock_key#2 ->pool_lock#2 ->&c->lock ->&____s->seqcount ->&n->list_lock FD: 10 BD: 78 +...: &bat_priv->tvlv.handler_list_lock ->pool_lock#2 ->&c->lock ->&n->list_lock ->&____s->seqcount#2 ->&____s->seqcount FD: 13 BD: 85 +...: &bat_priv->tvlv.container_list_lock ->pool_lock#2 ->&obj_hash[i].lock ->&c->lock ->&____s->seqcount#2 ->&n->list_lock ->&____s->seqcount ->quarantine_lock FD: 18 BD: 3837 +...: &batadv_netdev_addr_lock_key ->pool_lock#2 ->&obj_hash[i].lock ->krc.lock ->&c->lock ->&n->list_lock FD: 10 BD: 103 +...: &bat_priv->softif_vlan_list_lock ->pool_lock#2 ->&c->lock ->&n->list_lock ->batched_entropy_u8.lock ->kfence_freelist_lock ->&____s->seqcount ->&____s->seqcount#2 FD: 21 BD: 102 +...: key#16 ->&bat_priv->softif_vlan_list_lock ->&obj_hash[i].lock ->pool_lock#2 ->krc.lock ->&bat_priv->tt.changes_list_lock FD: 6 BD: 103 +...: &bat_priv->tt.changes_list_lock ->&obj_hash[i].lock ->pool_lock#2 ->&meta->lock ->kfence_freelist_lock FD: 30 BD: 1 ..-.: &(&bat_priv->nc.work)->timer FD: 79 BD: 1 +.+.: (wq_completion)bat_events ->(work_completion)(&(&bat_priv->nc.work)->work) ->(work_completion)(&(&bat_priv->mcast.work)->work) ->(work_completion)(&(&bat_priv->orig_work)->work) ->(work_completion)(&(&forw_packet_aggr->delayed_work)->work) ->(work_completion)(&(&bat_priv->tt.work)->work) ->(work_completion)(&(&bat_priv->dat.work)->work) ->(work_completion)(&(&bat_priv->bla.work)->work) ->&rq->__lock ->&cfs_rq->removed.lock ->&obj_hash[i].lock FD: 30 BD: 2 +.+.: (work_completion)(&(&bat_priv->nc.work)->work) ->rcu_node_0 ->&rq->__lock ->key#17 ->key#18 ->&obj_hash[i].lock ->&base->lock ->&rcu_state.expedited_wq ->pool_lock#2 ->&cfs_rq->removed.lock FD: 1 BD: 3 +...: key#17 FD: 1 BD: 3 +...: key#18 FD: 154 BD: 79 +.+.: init_lock ->slab_mutex ->fs_reclaim ->&zone->lock ->&____s->seqcount ->pool_lock#2 ->&obj_hash[i].lock ->&base->lock ->crngs.lock FD: 1 BD: 3847 +.-.: deferred_lock FD: 890 BD: 2 +.+.: deferred_process_work ->rtnl_mutex ->rtnl_mutex.wait_lock ->&p->pi_lock ->&rq->__lock FD: 1 BD: 81 ....: target_list_lock FD: 70 BD: 80 +.-.: &br->lock ->&br->hash_lock ->lweventlist_lock ->&obj_hash[i].lock ->&base->lock ->pool_lock#2 ->&dir->lock#2 ->deferred_lock ->(console_sem).lock ->&c->lock ->&____s->seqcount ->nl_table_lock ->nl_table_wait.lock ->&br->multicast_lock ->&n->list_lock FD: 30 BD: 1 ..-.: &(&bat_priv->mcast.work)->timer FD: 49 BD: 2 +.+.: (work_completion)(&(&bat_priv->mcast.work)->work) ->pool_lock#2 ->&bat_priv->mcast.mla_lock ->&obj_hash[i].lock ->&base->lock ->kfence_freelist_lock ->&meta->lock ->rcu_node_0 ->&rq->__lock ->&cfs_rq->removed.lock ->&rcu_state.expedited_wq ->quarantine_lock FD: 45 BD: 3 +.+.: &bat_priv->mcast.mla_lock ->pool_lock#2 ->key#16 ->&bat_priv->tt.changes_list_lock ->&bat_priv->tvlv.container_list_lock ->&obj_hash[i].lock ->&c->lock ->&n->list_lock ->&____s->seqcount#2 ->&____s->seqcount ->(console_sem).lock FD: 2 BD: 3823 +.+.: &bond->stats_lock/1 FD: 160 BD: 1 +.+.: (wq_completion)bond0 ->(work_completion)(&(&slave->notify_work)->work) FD: 159 BD: 3880 +.+.: (work_completion)(&(&slave->notify_work)->work) ->&obj_hash[i].lock ->&base->lock ->rtnl_mutex.wait_lock ->&p->pi_lock ->&rq->__lock ->rcu_node_0 ->pool_lock#2 ->&cfs_rq->removed.lock ->&rcu_state.expedited_wq FD: 30 BD: 1 ..-.: &(&slave->notify_work)->timer FD: 160 BD: 1 +.+.: (wq_completion)bond0#2 ->(work_completion)(&(&slave->notify_work)->work) FD: 160 BD: 1 +.+.: (wq_completion)bond0#3 ->(work_completion)(&(&slave->notify_work)->work) FD: 190 BD: 79 +.+.: team->team_lock_key ->fs_reclaim ->pool_lock#2 ->netpoll_srcu ->net_rwsem ->&tn->lock ->_xmit_ETHER ->&dir->lock#2 ->input_pool.lock ->&ndev->lock ->&obj_hash[i].lock ->nl_table_lock ->nl_table_wait.lock ->&in_dev->mc_tomb_lock ->&im->lock ->cbs_list_lock ->sysfs_symlink_target_lock ->lock ->&root->kernfs_rwsem ->&c->lock ->&n->list_lock ->&____s->seqcount ->lweventlist_lock ->(console_sem).lock ->&____s->seqcount#2 FD: 160 BD: 1 +.+.: (wq_completion)bond0#4 ->(work_completion)(&(&slave->notify_work)->work) FD: 207 BD: 79 +.+.: team->team_lock_key#2 ->fs_reclaim ->netpoll_srcu ->net_rwsem ->&tn->lock ->_xmit_ETHER ->&dir->lock#2 ->input_pool.lock ->&ndev->lock ->&obj_hash[i].lock ->nl_table_lock ->nl_table_wait.lock ->&c->lock ->&in_dev->mc_tomb_lock ->&im->lock ->cbs_list_lock ->&rq->__lock ->sysfs_symlink_target_lock ->lock ->&root->kernfs_rwsem ->&____s->seqcount ->quarantine_lock ->remove_cache_srcu ->lweventlist_lock ->(console_sem).lock ->console_owner_lock ->console_owner ->&base->lock ->pool_lock#2 ->krc.lock ->&dev_addr_list_lock_key#2/1 ->&____s->seqcount#2 ->&tbl->lock ->&pn->hash_lock ->&bond->mode_lock ->&dev_addr_list_lock_key ->&n->list_lock FD: 41 BD: 81 +.+.: &hard_iface->bat_iv.ogm_buff_mutex ->crngs.lock ->pool_lock#2 ->batched_entropy_u8.lock ->&bat_priv->forw_bat_list_lock ->&c->lock ->&obj_hash[i].lock ->rcu_node_0 ->&rq->__lock ->&n->list_lock ->&rcu_state.expedited_wq ->&____s->seqcount ->&bat_priv->tt.commit_lock ->&bat_priv->tvlv.container_list_lock ->kfence_freelist_lock ->&____s->seqcount#2 ->&cfs_rq->removed.lock FD: 30 BD: 1 ..-.: &(&bat_priv->orig_work)->timer FD: 27 BD: 2 +.+.: (work_completion)(&(&bat_priv->orig_work)->work) ->key#19 ->&obj_hash[i].lock ->&base->lock ->&rq->__lock ->pool_lock#2 ->rcu_node_0 ->&rcu_state.expedited_wq ->&cfs_rq->removed.lock FD: 1 BD: 80 +...: key#19 FD: 13 BD: 82 +...: &bat_priv->forw_bat_list_lock ->&obj_hash[i].lock ->&base->lock FD: 424 BD: 79 +.+.: team->team_lock_key#3 ->fs_reclaim ->netpoll_srcu ->net_rwsem ->&tn->lock ->_xmit_ETHER ->&dir->lock#2 ->input_pool.lock ->&ndev->lock ->&obj_hash[i].lock ->nl_table_lock ->nl_table_wait.lock ->&c->lock ->&in_dev->mc_tomb_lock ->&im->lock ->cbs_list_lock ->sysfs_symlink_target_lock ->lock ->&root->kernfs_rwsem ->&rq->__lock ->quarantine_lock ->remove_cache_srcu ->&____s->seqcount ->(console_sem).lock ->pool_lock#2 ->lweventlist_lock ->&macvlan_netdev_addr_lock_key/1 ->pcpu_alloc_mutex ->&____s->seqcount#2 ->&idev->mc_lock ->&tb->tb6_lock ->(inet6addr_validator_chain).rwsem ->stock_lock ->&net->ipv6.addrconf_hash_lock ->pcpu_lock ->&ifa->lock ->&n->list_lock ->&pn->hash_lock ->&dev->tx_global_lock ->dev->qdisc_tx_busylock ?: &qdisc_tx_busylock#2 ->&sch->q.lock ->__ip_vs_mutex ->krc.lock ->&tbl->lock ->class ->(&tbl->proxy_timer) ->&base->lock ->flowtable_lock ->&dir->lock ->&idev->mc_query_lock ->(work_completion)(&(&idev->mc_report_work)->work) ->&batadv_netdev_addr_lock_key ->&bat_priv->softif_vlan_list_lock ->key#16 ->&bat_priv->tt.changes_list_lock ->&cfs_rq->removed.lock FD: 1 BD: 78 +...: _xmit_NONE FD: 1 BD: 78 +...: lock#9 FD: 25 BD: 1 ..-.: drivers/net/wireguard/ratelimiter.c:20 FD: 29 BD: 2 +.+.: (gc_work).work ->tk_core.seq.seqcount ->"ratelimiter_table_lock" ->&rq->__lock ->&obj_hash[i].lock ->&base->lock ->rcu_node_0 ->&rcu_state.expedited_wq ->&cfs_rq->removed.lock ->pool_lock#2 FD: 1 BD: 3 +.+.: "ratelimiter_table_lock" FD: 429 BD: 79 +.+.: team->team_lock_key#4 ->fs_reclaim ->netpoll_srcu ->net_rwsem ->&tn->lock ->_xmit_ETHER ->&dir->lock#2 ->input_pool.lock ->&ndev->lock ->&obj_hash[i].lock ->nl_table_lock ->nl_table_wait.lock ->&in_dev->mc_tomb_lock ->&im->lock ->cbs_list_lock ->&rq->__lock ->&c->lock ->sysfs_symlink_target_lock ->lock ->&root->kernfs_rwsem ->&____s->seqcount ->lweventlist_lock ->(console_sem).lock ->&n->list_lock ->&macvlan_netdev_addr_lock_key/1 ->pcpu_alloc_mutex ->&idev->mc_lock ->&tb->tb6_lock ->(inet6addr_validator_chain).rwsem ->stock_lock ->&net->ipv6.addrconf_hash_lock ->pcpu_lock ->&ifa->lock ->&pn->hash_lock ->&dev->tx_global_lock ->dev->qdisc_tx_busylock ?: &qdisc_tx_busylock#2 ->&sch->q.lock ->__ip_vs_mutex ->krc.lock ->&tbl->lock ->class ->(&tbl->proxy_timer) ->&base->lock ->flowtable_lock ->&dir->lock ->&idev->mc_query_lock ->(work_completion)(&(&idev->mc_report_work)->work) ->pool_lock#2 ->rlock-AF_NETLINK ->&bond->mode_lock ->&dev_addr_list_lock_key ->&dev_addr_list_lock_key#2/1 ->&cfs_rq->removed.lock ->&dev_addr_list_lock_key#2/3 ->&vlan_netdev_addr_lock_key/1 FD: 160 BD: 1 +.+.: (wq_completion)bond0#5 ->(work_completion)(&(&slave->notify_work)->work) FD: 1 BD: 3906 ...-: &____s->seqcount#12 FD: 30 BD: 1 ..-.: &(&hdev->cmd_timer)->timer FD: 44 BD: 16 +.+.: (work_completion)(&(&hdev->cmd_timer)->work) ->(console_sem).lock ->console_owner_lock ->console_owner ->&rq->__lock ->rcu_node_0 ->&rcu_state.expedited_wq FD: 1 BD: 3840 +.-.: &hsr->list_lock FD: 160 BD: 1 +.+.: (wq_completion)bond0#6 ->(work_completion)(&(&slave->notify_work)->work) FD: 1 BD: 243 +.+.: rcu_state.exp_wake_mutex.wait_lock FD: 190 BD: 79 +.+.: team->team_lock_key#5 ->fs_reclaim ->netpoll_srcu ->net_rwsem ->&tn->lock ->_xmit_ETHER ->&dir->lock#2 ->input_pool.lock ->&c->lock ->&____s->seqcount ->&ndev->lock ->&obj_hash[i].lock ->nl_table_lock ->nl_table_wait.lock ->&rq->__lock ->&in_dev->mc_tomb_lock ->&im->lock ->cbs_list_lock ->sysfs_symlink_target_lock ->lock ->&root->kernfs_rwsem ->lweventlist_lock ->(console_sem).lock ->quarantine_lock ->remove_cache_srcu FD: 190 BD: 79 +.+.: team->team_lock_key#6 ->fs_reclaim ->netpoll_srcu ->net_rwsem ->&tn->lock ->_xmit_ETHER ->&dir->lock#2 ->input_pool.lock ->&ndev->lock ->&obj_hash[i].lock ->nl_table_lock ->nl_table_wait.lock ->&in_dev->mc_tomb_lock ->&im->lock ->cbs_list_lock ->&c->lock ->sysfs_symlink_target_lock ->lock ->&root->kernfs_rwsem ->&n->list_lock ->&rq->__lock ->&____s->seqcount ->lweventlist_lock ->(console_sem).lock ->remove_cache_srcu FD: 9 BD: 3836 +...: &vlan_netdev_addr_lock_key ->pool_lock#2 ->&c->lock ->&____s->seqcount ->&n->list_lock FD: 24 BD: 78 +.-.: (&app->join_timer) ->&app->lock ->&list->lock#11 FD: 16 BD: 80 +.-.: &app->lock ->batched_entropy_u32.lock ->&obj_hash[i].lock ->&base->lock ->pool_lock#2 ->&zone->lock ->&____s->seqcount ->&list->lock#11 ->&c->lock ->&n->list_lock ->&____s->seqcount#2 ->batched_entropy_u8.lock ->kfence_freelist_lock FD: 1 BD: 81 ..-.: &list->lock#11 FD: 20 BD: 78 +.-.: (&app->join_timer)#2 ->&app->lock#2 ->&list->lock#12 ->batched_entropy_u32.lock ->&obj_hash[i].lock ->&base->lock FD: 10 BD: 79 +.-.: &app->lock#2 ->pool_lock#2 ->&c->lock ->&list->lock#12 ->&obj_hash[i].lock FD: 1 BD: 80 ..-.: &list->lock#12 FD: 30 BD: 1 ..-.: &(&forw_packet_aggr->delayed_work)->timer FD: 50 BD: 79 +.+.: (work_completion)(&(&forw_packet_aggr->delayed_work)->work) ->&hard_iface->bat_iv.ogm_buff_mutex ->&bat_priv->forw_bat_list_lock ->&obj_hash[i].lock ->rcu_node_0 ->&rcu_state.expedited_wq ->&rq->__lock ->pool_lock#2 ->&meta->lock ->kfence_freelist_lock ->&c->lock ->&____s->seqcount#2 ->&____s->seqcount ->&n->list_lock ->batched_entropy_u8.lock ->quarantine_lock ->&cfs_rq->removed.lock ->&base->lock FD: 9 BD: 3836 +...: &macvlan_netdev_addr_lock_key ->pool_lock#2 ->&c->lock ->&____s->seqcount ->&n->list_lock FD: 8 BD: 3836 +...: &dev_addr_list_lock_key#3 ->pool_lock#2 ->&c->lock ->&____s->seqcount FD: 1 BD: 78 ....: &xa->xa_lock#15 FD: 18 BD: 3836 +...: &dev_addr_list_lock_key#3/1 ->&obj_hash[i].lock ->pool_lock#2 ->krc.lock FD: 2 BD: 78 +.+.: &tap_major->minor_lock ->pool_lock#2 FD: 3 BD: 78 +.+.: subsys mutex#82 ->&k->k_lock FD: 909 BD: 1 .+.+: kn->active#51 ->fs_reclaim ->&kernfs_locks->open_file_mutex[count] ->nsim_bus_dev_list_lock ->&c->lock ->nsim_bus_dev_list_lock.wait_lock ->&p->pi_lock FD: 907 BD: 9 +.+.: nsim_bus_dev_list_lock ->fs_reclaim ->pool_lock#2 ->nsim_bus_dev_ids.xa_lock ->&x->wait#9 ->&obj_hash[i].lock ->&k->list_lock ->lock ->&root->kernfs_rwsem ->bus_type_sem ->sysfs_symlink_target_lock ->&k->k_lock ->&c->lock ->&____s->seqcount ->&dev->power.lock ->dpm_list_mtx ->uevent_sock_mutex ->nsim_bus_dev_list_lock.wait_lock ->&rq->__lock ->subsys mutex#83 FD: 909 BD: 1 .+.+: kn->active#52 ->fs_reclaim ->&kernfs_locks->open_file_mutex[count] ->nsim_bus_dev_list_lock ->nsim_bus_dev_list_lock.wait_lock ->&p->pi_lock ->&c->lock ->&rq->__lock FD: 1 BD: 10 ....: nsim_bus_dev_ids.xa_lock FD: 8 BD: 18 +.+.: devlinks.xa_lock ->&c->lock ->&____s->seqcount ->pool_lock#2 FD: 895 BD: 12 +.+.: &devlink->lock_key ->crngs.lock ->fs_reclaim ->pool_lock#2 ->devlinks.xa_lock ->&obj_hash[i].lock ->nl_table_lock ->nl_table_wait.lock ->&c->lock ->&____s->seqcount ->&xa->xa_lock#16 ->pcpu_alloc_mutex ->quarantine_lock ->&n->list_lock ->&base->lock ->pin_fs_lock ->&sb->s_type->i_mutex_key#3 ->batched_entropy_u32.lock ->rtnl_mutex ->rtnl_mutex.wait_lock ->&p->pi_lock ->rcu_node_0 ->&rq->__lock ->&(&fn_net->fib_chain)->lock ->&devlink_port->type_lock ->stack_depot_init_mutex ->&rcu_state.expedited_wq ->&nsim_trap_data->trap_lock ->&cfs_rq->removed.lock FD: 8 BD: 18 +.+.: &xa->xa_lock#16 ->pool_lock#2 ->&c->lock ->&____s->seqcount FD: 1 BD: 3947 +...: &data->fib_event_queue_lock FD: 131 BD: 2 +.+.: (work_completion)(&data->fib_event_work) ->&data->fib_event_queue_lock ->&data->fib_lock ->&rq->__lock FD: 129 BD: 3 +.+.: &data->fib_lock ->fs_reclaim ->pool_lock#2 ->&obj_hash[i].lock ->&base->lock ->&pool->lock ->&rq->__lock ->(&timer.timer) ->&c->lock ->batched_entropy_u8.lock ->kfence_freelist_lock ->&n->list_lock ->remove_cache_srcu ->&____s->seqcount ->&cfs_rq->removed.lock ->pool_lock ->&____s->seqcount#2 ->rcu_node_0 ->&meta->lock ->&rcu_state.expedited_wq FD: 1 BD: 18 ....: &(&fn_net->fib_chain)->lock FD: 1 BD: 10 +.+.: nsim_bus_dev_list_lock.wait_lock FD: 25 BD: 1 ..-.: &(&nsim_dev->trap_data->trap_report_dw)->timer FD: 33 BD: 2 +.+.: (work_completion)(&(&nsim_dev->trap_data->trap_report_dw)->work) ->&rq->__lock ->&obj_hash[i].lock ->&base->lock ->rcu_node_0 FD: 1 BD: 78 +...: &devlink_port->type_lock FD: 128 BD: 81 +.+.: bpf_devs_lock ->fs_reclaim ->pool_lock#2 ->&c->lock ->&n->list_lock ->&rq->__lock ->stock_lock ->&obj_hash[i].lock FD: 1 BD: 78 +.+.: (work_completion)(&(&devlink_port->type_warn_dw)->work) FD: 1 BD: 78 +.+.: &vn->sock_lock FD: 1 BD: 80 ..-.: &list->lock#13 FD: 1 BD: 80 ..-.: &list->lock#14 FD: 1 BD: 1 ..-.: &list->lock#15 FD: 1 BD: 4 +.-.: x25_list_lock FD: 1 BD: 1 +.-.: x25_forward_list_lock FD: 1 BD: 10 +.+.: subsys mutex#83 FD: 895 BD: 12 +.+.: &devlink->lock_key#2 ->crngs.lock ->fs_reclaim ->devlinks.xa_lock ->&c->lock ->&obj_hash[i].lock ->nl_table_lock ->nl_table_wait.lock ->&____s->seqcount ->pool_lock#2 ->&xa->xa_lock#16 ->pcpu_alloc_mutex ->&n->list_lock ->&rq->__lock ->&base->lock ->pin_fs_lock ->&sb->s_type->i_mutex_key#3 ->batched_entropy_u32.lock ->rtnl_mutex ->rtnl_mutex.wait_lock ->&p->pi_lock ->rcu_node_0 ->&(&fn_net->fib_chain)->lock ->&devlink_port->type_lock ->stack_depot_init_mutex ->&nsim_trap_data->trap_lock ->&rcu_state.expedited_wq ->&cfs_rq->removed.lock FD: 151 BD: 78 ++++: devnet_rename_sem ->(console_sem).lock ->fs_reclaim ->pool_lock#2 ->&k->list_lock ->&root->kernfs_rwsem ->&c->lock ->&____s->seqcount ->kernfs_rename_lock ->uevent_sock_mutex ->&obj_hash[i].lock ->&rq->__lock ->batched_entropy_u8.lock ->kfence_freelist_lock ->remove_cache_srcu ->&meta->lock FD: 1 BD: 305 ....: kernfs_rename_lock FD: 309 BD: 83 +.+.: &nft_net->commit_mutex ->fs_reclaim ->pool_lock#2 ->stock_lock ->batched_entropy_u32.lock ->&obj_hash[i].lock ->rcu_node_0 ->&rnp->exp_wq[1] ->&rq->__lock ->(work_completion)(&ht->run_work) ->&ht->mutex ->&c->lock ->&____s->seqcount#2 ->&n->list_lock ->&rnp->exp_wq[2] ->&lock->wait_lock ->&____s->seqcount ->&rnp->exp_wq[3] ->nf_ct_proto_mutex ->nl_table_lock ->nl_table_wait.lock ->rlock-AF_NETLINK ->&p->alloc_lock ->&rnp->exp_wq[0] ->(console_sem).lock ->&base->lock ->(work_completion)(&(&priv->gc_work)->work) ->rcu_state.exp_mutex.wait_lock ->&p->pi_lock ->&rnp->exp_lock ->rcu_state.exp_mutex ->nf_hook_mutex ->cpu_hotplug_lock ->nf_tables_destroy_list_lock ->&cfs_rq->removed.lock ->remove_cache_srcu ->defrag4_mutex ->(work_completion)(&(&priv->gc_work)->work)#2 ->rcu_state.barrier_mutex ->batched_entropy_u8.lock ->kfence_freelist_lock ->flowtable_lock ->krc.lock ->(work_completion)(&(&flowtable->gc_work)->work) ->&ht->lock ->&meta->lock FD: 25 BD: 1 ..-.: &(&hwstats->traffic_dw)->timer FD: 25 BD: 2 +.+.: (work_completion)(&(&hwstats->traffic_dw)->work) ->&hwstats->hwsdev_list_lock ->&obj_hash[i].lock ->&base->lock ->&rq->__lock ->&cfs_rq->removed.lock FD: 24 BD: 82 +.+.: &hwstats->hwsdev_list_lock ->&rq->__lock ->rcu_node_0 FD: 397 BD: 78 +.+.: &wg->device_update_lock ->&wg->static_identity.lock ->fs_reclaim ->&____s->seqcount ->pool_lock#2 ->&c->lock ->pcpu_alloc_mutex ->&handshake->lock ->&obj_hash[i].lock ->tk_core.seq.seqcount ->&table->lock ->&peer->endpoint_lock ->&zone->lock ->&n->list_lock ->mmu_notifier_invalidate_range_start ->&sb->s_type->i_lock_key#8 ->&dir->lock ->k-sk_lock-AF_INET ->k-slock-AF_INET ->cpu_hotplug_lock ->k-sk_lock-AF_INET6 ->k-slock-AF_INET6 ->&wg->socket_update_lock ->&rq->__lock ->&list->lock#17 ->&pool->lock/1 ->&rnp->exp_wq[0] ->&____s->seqcount#2 ->(&peer->timer_retransmit_handshake) ->&base->lock ->(&peer->timer_send_keepalive) ->(&peer->timer_new_handshake) ->(&peer->timer_zero_key_material) ->(&peer->timer_persistent_keepalive) ->(work_completion)(&peer->clear_peer_work) ->&peer->keypairs.keypair_update_lock ->&wq->mutex ->napi_hash_lock ->&table->lock#2 ->wq_pool_mutex ->wq_mayday_lock ->&p->pi_lock ->&cfs_rq->removed.lock ->&x->wait ->rcu_state.exp_mutex.wait_lock ->pcpu_lock ->&r->consumer_lock#2 ->rcu_state.barrier_mutex ->rcu_state.barrier_mutex.wait_lock ->init_lock ->stock_lock ->&rnp->exp_lock ->rcu_state.exp_mutex FD: 130 BD: 132 ++++: &wg->static_identity.lock ->&handshake->lock ->&rq->__lock ->&sem->wait_lock ->&p->pi_lock FD: 129 BD: 134 ++++: &handshake->lock ->crngs.lock ->tk_core.seq.seqcount ->&table->lock#2 ->fs_reclaim ->pool_lock#2 ->&c->lock ->&____s->seqcount ->&____s->seqcount#2 ->&rq->__lock ->&sem->wait_lock ->remove_cache_srcu ->&n->list_lock ->rcu_node_0 ->&rcu_state.expedited_wq FD: 1 BD: 79 +.+.: &table->lock FD: 64 BD: 135 ++-.: &peer->endpoint_lock ->pool_lock#2 ->&obj_hash[i].lock FD: 29 BD: 20 +.+.: &nsim_trap_data->trap_lock ->pool_lock#2 ->crngs.lock ->&nsim_dev->fa_cookie_lock ->&obj_hash[i].lock ->&c->lock ->&zone->lock ->&____s->seqcount ->&n->list_lock ->quarantine_lock ->batched_entropy_u8.lock ->kfence_freelist_lock ->&meta->lock ->&____s->seqcount#2 ->&base->lock ->&pgdat->kswapd_wait FD: 1 BD: 21 +...: &nsim_dev->fa_cookie_lock FD: 1 BD: 78 +...: _xmit_SIT FD: 895 BD: 12 +.+.: &devlink->lock_key#3 ->crngs.lock ->fs_reclaim ->devlinks.xa_lock ->&obj_hash[i].lock ->nl_table_lock ->nl_table_wait.lock ->&c->lock ->&____s->seqcount ->pool_lock#2 ->&xa->xa_lock#16 ->pcpu_alloc_mutex ->&rq->__lock ->&base->lock ->pin_fs_lock ->&sb->s_type->i_mutex_key#3 ->batched_entropy_u32.lock ->rtnl_mutex ->rtnl_mutex.wait_lock ->&p->pi_lock ->&(&fn_net->fib_chain)->lock ->&devlink_port->type_lock ->stack_depot_init_mutex ->&nsim_trap_data->trap_lock ->rcu_node_0 ->&rcu_state.expedited_wq ->&cfs_rq->removed.lock FD: 18 BD: 3836 +...: &bridge_netdev_addr_lock_key/1 ->pool_lock#2 ->&c->lock ->&____s->seqcount ->&obj_hash[i].lock ->krc.lock FD: 38 BD: 78 +.-.: (&brmctx->ip6_own_query.timer) ->&br->multicast_lock FD: 37 BD: 3846 +.-.: &br->multicast_lock ->&obj_hash[i].lock ->&base->lock ->pool_lock#2 ->&dir->lock#2 ->deferred_lock ->nl_table_lock ->nl_table_wait.lock ->&c->lock ->&____s->seqcount ->&n->list_lock ->&____s->seqcount#2 ->quarantine_lock ->batched_entropy_u8.lock ->kfence_freelist_lock ->&meta->lock ->tk_core.seq.seqcount ->&list->lock#5 FD: 38 BD: 78 +.-.: (&brmctx->ip4_own_query.timer) ->&br->multicast_lock FD: 92 BD: 1 +.-.: (&in_dev->mr_ifc_timer) ->&obj_hash[i].lock ->batched_entropy_u32.lock ->&base->lock FD: 895 BD: 12 +.+.: &devlink->lock_key#4 ->crngs.lock ->fs_reclaim ->devlinks.xa_lock ->&obj_hash[i].lock ->nl_table_lock ->nl_table_wait.lock ->&xa->xa_lock#16 ->&c->lock ->&____s->seqcount ->pcpu_alloc_mutex ->remove_cache_srcu ->&base->lock ->pin_fs_lock ->&sb->s_type->i_mutex_key#3 ->pool_lock#2 ->batched_entropy_u32.lock ->rtnl_mutex ->rtnl_mutex.wait_lock ->&p->pi_lock ->&n->list_lock ->&(&fn_net->fib_chain)->lock ->&devlink_port->type_lock ->stack_depot_init_mutex ->&rq->__lock ->&nsim_trap_data->trap_lock ->&cfs_rq->removed.lock ->rcu_node_0 ->&rcu_state.expedited_wq FD: 1 BD: 78 +...: _xmit_TUNNEL FD: 39 BD: 3836 +...: _xmit_IPGRE ->pool_lock#2 ->(console_sem).lock ->console_owner_lock ->console_owner ->&obj_hash[i].lock ->krc.lock ->&c->lock FD: 18 BD: 78 +...: _xmit_TUNNEL6 ->pool_lock#2 ->&obj_hash[i].lock ->krc.lock FD: 25 BD: 1 ..-.: &(&br->gc_work)->timer FD: 66 BD: 79 +.+.: (work_completion)(&(&br->gc_work)->work) ->&obj_hash[i].lock ->&base->lock ->rcu_node_0 ->&rq->__lock FD: 70 BD: 3839 +.-.: _xmit_TUNNEL6#2 ->&obj_hash[i].lock ->pool_lock#2 ->&meta->lock ->kfence_freelist_lock ->icmp_global.lock ->&data->lock FD: 17 BD: 78 +.-.: (&app->periodic_timer) ->&app->lock FD: 47 BD: 3836 +...: &dev_addr_list_lock_key/1 ->_xmit_ETHER ->&c->lock ->&____s->seqcount ->&obj_hash[i].lock ->pool_lock#2 ->krc.lock ->&batadv_netdev_addr_lock_key ->&bridge_netdev_addr_lock_key ->&n->list_lock ->&____s->seqcount#2 FD: 895 BD: 12 +.+.: &devlink->lock_key#5 ->crngs.lock ->fs_reclaim ->devlinks.xa_lock ->&c->lock ->&____s->seqcount ->pool_lock#2 ->&obj_hash[i].lock ->nl_table_lock ->nl_table_wait.lock ->&xa->xa_lock#16 ->pcpu_alloc_mutex ->&base->lock ->pin_fs_lock ->&sb->s_type->i_mutex_key#3 ->batched_entropy_u32.lock ->rtnl_mutex ->rtnl_mutex.wait_lock ->&p->pi_lock ->&(&fn_net->fib_chain)->lock ->&devlink_port->type_lock ->stack_depot_init_mutex ->&rq->__lock ->&n->list_lock ->&nsim_trap_data->trap_lock ->rcu_node_0 ->&rcu_state.expedited_wq ->&cfs_rq->removed.lock FD: 1 BD: 6 +.+.: genl_mutex.wait_lock FD: 46 BD: 3842 +...: &dev_addr_list_lock_key#2/1 ->_xmit_ETHER ->&obj_hash[i].lock ->krc.lock ->&dev_addr_list_lock_key ->&c->lock FD: 895 BD: 12 +.+.: &devlink->lock_key#6 ->crngs.lock ->fs_reclaim ->devlinks.xa_lock ->&c->lock ->&n->list_lock ->&obj_hash[i].lock ->nl_table_lock ->nl_table_wait.lock ->&xa->xa_lock#16 ->&____s->seqcount ->pcpu_alloc_mutex ->&base->lock ->&rq->__lock ->pin_fs_lock ->&sb->s_type->i_mutex_key#3 ->batched_entropy_u32.lock ->rtnl_mutex ->rtnl_mutex.wait_lock ->&p->pi_lock ->rcu_node_0 ->&(&fn_net->fib_chain)->lock ->&devlink_port->type_lock ->stack_depot_init_mutex ->&nsim_trap_data->trap_lock ->&rcu_state.expedited_wq ->&cfs_rq->removed.lock FD: 2 BD: 1 +.-.: (&tun->flow_gc_timer) ->&tun->lock FD: 1 BD: 79 +.-.: &tun->lock FD: 38 BD: 1 +.-.: (&pmctx->ip6_own_query.timer) ->&br->multicast_lock FD: 38 BD: 1 +.-.: (&pmctx->ip4_own_query.timer) ->&br->multicast_lock FD: 25 BD: 1 ..-.: &(&conn->info_timer)->timer FD: 42 BD: 25 +.+.: (work_completion)(&(&conn->info_timer)->work) ->&conn->chan_lock FD: 18 BD: 3836 +...: _xmit_ETHER/1 ->&c->lock ->&____s->seqcount ->&obj_hash[i].lock ->krc.lock ->pool_lock#2 ->&____s->seqcount#2 ->&n->list_lock FD: 44 BD: 3838 +.-.: &hsr->seqnr_lock ->pool_lock#2 ->&obj_hash[i].lock ->&____s->seqcount ->quarantine_lock ->&meta->lock ->kfence_freelist_lock ->&data->lock FD: 1 BD: 3839 +.-.: &new_node->seq_out_lock FD: 45 BD: 78 +.-.: (&hsr->announce_timer) FD: 1 BD: 78 +.+.: &nn->netlink_tap_lock FD: 18 BD: 3836 +...: &batadv_netdev_addr_lock_key/1 ->&c->lock ->&____s->seqcount ->&obj_hash[i].lock ->krc.lock FD: 46 BD: 3842 +...: &vlan_netdev_addr_lock_key/1 ->_xmit_ETHER ->&c->lock ->&____s->seqcount ->pool_lock#2 ->&____s->seqcount#2 ->&obj_hash[i].lock ->krc.lock ->&bridge_netdev_addr_lock_key FD: 14 BD: 78 +.-.: (&hsr->prune_timer) ->&hsr->list_lock ->&obj_hash[i].lock ->&base->lock FD: 45 BD: 3844 +...: &macvlan_netdev_addr_lock_key/1 ->_xmit_ETHER ->&c->lock ->&____s->seqcount#2 ->&____s->seqcount ->pool_lock#2 ->&obj_hash[i].lock ->krc.lock ->&n->list_lock FD: 1 BD: 79 +.-.: &list->lock#16 FD: 27 BD: 78 +.+.: (work_completion)(&port->bc_work) ->&list->lock#16 ->&obj_hash[i].lock ->pool_lock#2 ->quarantine_lock ->&rq->__lock ->&meta->lock ->kfence_freelist_lock FD: 18 BD: 79 +...: &ipvlan->addrs_lock ->pool_lock#2 ->&c->lock ->&____s->seqcount#2 ->&____s->seqcount ->&obj_hash[i].lock ->krc.lock FD: 45 BD: 3836 +...: &macsec_netdev_addr_lock_key/1 ->_xmit_ETHER ->&obj_hash[i].lock ->pool_lock#2 ->krc.lock ->&c->lock FD: 37 BD: 81 +...: dev->qdisc_tx_busylock ?: &qdisc_tx_busylock#2 ->&sch->q.lock FD: 18 BD: 3839 +.-.: key#20 ->&obj_hash[i].lock ->pool_lock#2 ->krc.lock FD: 25 BD: 82 +...: &bat_priv->tt.commit_lock ->key#16 ->&bat_priv->softif_vlan_list_lock ->&bat_priv->tt.changes_list_lock ->&bat_priv->tt.last_changeset_lock ->pool_lock#2 ->&bat_priv->tvlv.container_list_lock ->&obj_hash[i].lock ->&c->lock ->&n->list_lock ->&____s->seqcount#2 ->&____s->seqcount ->&meta->lock ->kfence_freelist_lock FD: 30 BD: 1 ..-.: &(&bat_priv->tt.work)->timer FD: 38 BD: 2 +.+.: (work_completion)(&(&bat_priv->tt.work)->work) ->key#16 ->key#21 ->&rq->__lock ->&bat_priv->tt.req_list_lock ->&bat_priv->tt.roam_list_lock ->&obj_hash[i].lock ->&base->lock ->rcu_node_0 ->&rcu_state.expedited_wq ->&cfs_rq->removed.lock ->pool_lock#2 FD: 1 BD: 3 +...: key#21 FD: 1 BD: 3 +...: &bat_priv->tt.req_list_lock FD: 1 BD: 3 +...: &bat_priv->tt.roam_list_lock FD: 1 BD: 79 +.+.: &wg->socket_update_lock FD: 14 BD: 121 +.-.: &list->lock#17 ->&obj_hash[i].lock ->pool_lock#2 ->&data->lock FD: 126 BD: 1 +.+.: (wq_completion)wg-kex-wg0 ->(work_completion)(&peer->transmit_handshake_work) FD: 125 BD: 18 +.+.: (work_completion)(&peer->transmit_handshake_work) ->tk_core.seq.seqcount ->&wg->static_identity.lock ->&cookie->lock ->&obj_hash[i].lock ->&base->lock ->pool_lock#2 ->&peer->endpoint_lock ->batched_entropy_u8.lock ->&c->lock ->&rq->__lock ->rcu_node_0 ->&rcu_state.expedited_wq ->&____s->seqcount#2 ->&____s->seqcount ->kfence_freelist_lock ->&n->list_lock ->&cfs_rq->removed.lock FD: 1 BD: 136 +...: &table->lock#2 FD: 23 BD: 54 ++++: &cookie->lock ->&rq->__lock FD: 1 BD: 117 +.-.: &r->producer_lock#2 FD: 168 BD: 1 +.+.: (wq_completion)wg-kex-wg1 ->(work_completion)(&({ do { const void *__vpp_verify = (typeof((worker) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __asm__ ("" : "=r"(__ptr) : "0"((typeof(*((worker))) *)((worker)))); (typeof((typeof(*((worker))) *)((worker)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); })->work) FD: 167 BD: 35 +.+.: (work_completion)(&({ do { const void *__vpp_verify = (typeof((worker) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __asm__ ("" : "=r"(__ptr) : "0"((typeof(*((worker))) *)((worker)))); (typeof((typeof(*((worker))) *)((worker)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); })->work) ->&r->consumer_lock#2 ->&wg->static_identity.lock ->&peer->endpoint_lock ->tk_core.seq.seqcount ->&cookie->lock ->&handshake->lock ->&obj_hash[i].lock ->&base->lock ->pool_lock#2 ->&rq->__lock ->&list->lock#17 ->&c->lock ->rcu_node_0 ->&____s->seqcount#2 ->&____s->seqcount ->&n->list_lock ->&rcu_state.expedited_wq ->&cfs_rq->removed.lock ->batched_entropy_u8.lock ->kfence_freelist_lock ->&meta->lock FD: 1 BD: 114 +.+.: &r->consumer_lock#2 FD: 10 BD: 135 +.-.: &peer->keypairs.keypair_update_lock ->&table->lock#2 ->&obj_hash[i].lock ->pool_lock#2 FD: 168 BD: 1 +.+.: (wq_completion)wg-kex-wg0#2 ->(work_completion)(&({ do { const void *__vpp_verify = (typeof((worker) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __asm__ ("" : "=r"(__ptr) : "0"((typeof(*((worker))) *)((worker)))); (typeof((typeof(*((worker))) *)((worker)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); })->work) FD: 169 BD: 1 +.+.: (wq_completion)wg-crypt-wg0 ->(work_completion)(&({ do { const void *__vpp_verify = (typeof((worker) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __asm__ ("" : "=r"(__ptr) : "0"((typeof(*((worker))) *)((worker)))); (typeof((typeof(*((worker))) *)((worker)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); })->work) ->(work_completion)(&peer->transmit_packet_work) FD: 29 BD: 18 +.+.: (work_completion)(&peer->transmit_packet_work) ->&obj_hash[i].lock ->&peer->endpoint_lock ->&base->lock ->batched_entropy_u8.lock ->&rq->__lock ->rcu_node_0 ->&cfs_rq->removed.lock ->pool_lock#2 ->&rcu_state.expedited_wq FD: 169 BD: 1 +.+.: (wq_completion)wg-crypt-wg1 ->(work_completion)(&({ do { const void *__vpp_verify = (typeof((worker) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __asm__ ("" : "=r"(__ptr) : "0"((typeof(*((worker))) *)((worker)))); (typeof((typeof(*((worker))) *)((worker)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); })->work) ->(work_completion)(&peer->transmit_packet_work) ->&rq->__lock FD: 1 BD: 1 +.-.: &keypair->receiving_counter.lock FD: 126 BD: 1 +.+.: (wq_completion)wg-kex-wg1#2 ->(work_completion)(&peer->transmit_handshake_work) FD: 105 BD: 1 +.-.: (&ndev->rs_timer) ->&ndev->lock ->pool_lock#2 ->&c->lock ->&____s->seqcount ->&dir->lock#2 ->&ul->lock#2 ->&obj_hash[i].lock ->&n->list_lock ->&____s->seqcount#2 ->batched_entropy_u8.lock ->kfence_freelist_lock ->init_task.mems_allowed_seq.seqcount ->&data->lock ->quarantine_lock ->key#29 FD: 127 BD: 11 +.+.: &data->mtx ->fs_reclaim ->pool_lock#2 ->&rfkill->lock ->&c->lock ->&____s->seqcount ->&____s->seqcount#2 FD: 126 BD: 1 +.+.: (wq_completion)wg-kex-wg2 ->(work_completion)(&peer->transmit_handshake_work) FD: 168 BD: 1 +.+.: (wq_completion)wg-kex-wg2#2 ->(work_completion)(&({ do { const void *__vpp_verify = (typeof((worker) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __asm__ ("" : "=r"(__ptr) : "0"((typeof(*((worker))) *)((worker)))); (typeof((typeof(*((worker))) *)((worker)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); })->work) FD: 169 BD: 1 +.+.: (wq_completion)wg-crypt-wg2 ->(work_completion)(&({ do { const void *__vpp_verify = (typeof((worker) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __asm__ ("" : "=r"(__ptr) : "0"((typeof(*((worker))) *)((worker)))); (typeof((typeof(*((worker))) *)((worker)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); })->work) ->(work_completion)(&peer->transmit_packet_work) FD: 1 BD: 3839 +.-.: &entry->crc_lock FD: 9 BD: 83 +...: &bat_priv->tt.last_changeset_lock ->pool_lock#2 ->&obj_hash[i].lock ->&c->lock FD: 1 BD: 86 ....: &wdev->event_lock FD: 1 BD: 81 +.+.: (work_completion)(&(&sdata->dec_tailroom_needed_wk)->work) FD: 29 BD: 87 +.+.: &local->key_mtx ->&rq->__lock ->&obj_hash[i].lock FD: 25 BD: 93 ..-.: &rdev->wiphy_work_lock FD: 1 BD: 81 ....: (&dwork->timer) FD: 1 BD: 81 +.+.: (work_completion)(&(&link->color_collision_detect_work)->work) FD: 189 BD: 1 +.+.: (wq_completion)phy3 ->(work_completion)(&local->reconfig_filter) ->(work_completion)(&link->csa_finalize_work) FD: 543 BD: 2 +.+.: (work_completion)(&rdev->wiphy_work) ->&rq->__lock ->&rdev->wiphy.mtx ->&lock->wait_lock ->&p->pi_lock FD: 1 BD: 87 ..-.: &list->lock#18 FD: 1 BD: 86 +.-.: &ifibss->incomplete_lock FD: 154 BD: 90 +.+.: &local->mtx ->&local->chanctx_mtx ->&data->mutex ->&local->queue_stop_reason_lock ->&local->iflist_mtx ->&local->filter_lock ->&obj_hash[i].lock ->&base->lock ->&____s->seqcount ->pool_lock#2 ->fs_reclaim ->&local->ack_status_lock ->rcu_node_0 ->&rq->__lock ->&c->lock ->&n->list_lock ->&rdev->wiphy_work_lock ->&____s->seqcount#2 ->remove_cache_srcu ->&cfs_rq->removed.lock ->nl_table_lock ->nl_table_wait.lock ->&lock->wait_lock ->rcu_state.exp_mutex.wait_lock ->&p->pi_lock ->batched_entropy_u8.lock ->kfence_freelist_lock ->&rnp->exp_lock ->rcu_state.exp_mutex FD: 892 BD: 1 +.+.: (wq_completion)cfg80211 ->(work_completion)(&rdev->event_work) ->(work_completion)(&(&rdev->dfs_update_channels_wk)->work) FD: 543 BD: 2 +.+.: (work_completion)(&rdev->event_work) ->&rdev->wiphy.mtx ->&lock->wait_lock ->&p->pi_lock FD: 157 BD: 2 +.+.: wireless_nlevent_work ->net_rwsem FD: 157 BD: 81 +.+.: (wq_completion)phy4 ->(work_completion)(&local->reconfig_filter) ->(work_completion)(&(&local->scan_work)->work) FD: 142 BD: 1 +.+.: &type->s_umount_key#49/1 ->fs_reclaim ->pcpu_alloc_mutex ->shrinker_rwsem ->list_lrus_mutex ->sb_lock ->pool_lock#2 ->mmu_notifier_invalidate_range_start ->&sb->s_type->i_lock_key#32 ->&rq->__lock ->&s->s_inode_list_lock ->tk_core.seq.seqcount ->binderfs_minors_mutex ->&dentry->d_lock ->&sb->s_type->i_mutex_key#20 ->&c->lock ->&____s->seqcount#2 ->&____s->seqcount FD: 40 BD: 3 +.+.: &sb->s_type->i_lock_key#32 ->&dentry->d_lock FD: 2 BD: 2 +.+.: binderfs_minors_mutex ->binderfs_minors.xa_lock FD: 1 BD: 3 ....: binderfs_minors.xa_lock FD: 129 BD: 2 +.+.: &sb->s_type->i_mutex_key#20 ->&sb->s_type->i_lock_key#32 ->rename_lock.seqcount ->fs_reclaim ->&c->lock ->pool_lock#2 ->&dentry->d_lock ->mmu_notifier_invalidate_range_start ->&s->s_inode_list_lock ->tk_core.seq.seqcount ->&____s->seqcount ->&____s->seqcount#2 FD: 1 BD: 3 +.+.: iunique_lock FD: 866 BD: 3 +.+.: &type->i_mutex_dir_key#6/1 ->rename_lock.seqcount ->fs_reclaim ->&dentry->d_lock ->&root->kernfs_rwsem ->tomoyo_ss ->&root->kernfs_iattr_rwsem ->cgroup_mutex ->&xa->xa_lock#4 ->&obj_hash[i].lock ->stock_lock ->&c->lock ->&sb->s_type->i_lock_key#30 FD: 1 BD: 18 ....: task_group_lock FD: 128 BD: 1 .+.+: kn->active#53 ->fs_reclaim ->&kernfs_locks->open_file_mutex[count] ->&c->lock ->&____s->seqcount#2 ->&____s->seqcount FD: 128 BD: 1 ++++: kn->active#54 ->fs_reclaim ->&kernfs_locks->open_file_mutex[count] ->&c->lock ->&____s->seqcount#2 ->&____s->seqcount ->stock_lock ->&rq->__lock ->&n->list_lock ->remove_cache_srcu FD: 3 BD: 144 ..-.: cgroup_threadgroup_rwsem.rss.gp_wait.lock ->&obj_hash[i].lock FD: 1 BD: 3882 ..-.: &list->lock#19 FD: 30 BD: 1 +.-.: &local->rx_path_lock ->&list->lock#18 ->&rdev->wiphy_work_lock ->&obj_hash[i].lock ->pool_lock#2 ->&rdev->mgmt_registrations_lock ->&rdev->beacon_registrations_lock ->&data->lock FD: 18 BD: 94 +...: &sta->lock ->pool_lock#2 ->&obj_hash[i].lock ->krc.lock ->batched_entropy_u8.lock ->kfence_freelist_lock ->&c->lock ->&____s->seqcount#2 ->&____s->seqcount FD: 18 BD: 86 +.-.: &sta->rate_ctrl_lock ->pool_lock#2 ->&obj_hash[i].lock ->krc.lock ->&c->lock ->&n->list_lock ->batched_entropy_u8.lock ->kfence_freelist_lock FD: 170 BD: 86 +.+.: &local->sta_mtx ->fs_reclaim ->pool_lock#2 ->&local->chanctx_mtx ->pin_fs_lock ->&sb->s_type->i_mutex_key#3 ->&obj_hash[i].lock ->nl_table_lock ->nl_table_wait.lock ->&c->lock ->&____s->seqcount#2 ->&____s->seqcount ->&sta->ampdu_mlme.mtx ->(work_completion)(&sta->ampdu_mlme.work) ->&rq->__lock ->&sta->lock ->krc.lock ->&local->key_mtx ->&fq->lock ->&dentry->d_lock ->&cfs_rq->removed.lock ->&fsnotify_mark_srcu ->&sb->s_type->i_lock_key#7 ->&s->s_inode_list_lock ->&xa->xa_lock#9 ->mount_lock ->&local->active_txq_lock[i] ->(work_completion)(&sta->drv_deliver_wk) ->batched_entropy_u8.lock ->kfence_freelist_lock ->remove_cache_srcu ->&rnp->exp_lock ->rcu_state.exp_mutex ->&meta->lock ->&n->list_lock FD: 24 BD: 144 ....: cgroup_threadgroup_rwsem.waiters.lock ->&p->pi_lock FD: 1 BD: 18 +.+.: (wq_completion)cpuset_migrate_mm FD: 866 BD: 3 +.+.: &type->i_mutex_dir_key#7/1 ->&rq->__lock ->rename_lock.seqcount ->fs_reclaim ->&dentry->d_lock ->&root->kernfs_rwsem ->tomoyo_ss ->&root->kernfs_iattr_rwsem ->cgroup_mutex ->pool_lock#2 ->&xa->xa_lock#4 ->&obj_hash[i].lock ->stock_lock ->cgroup_mutex.wait_lock ->&p->pi_lock ->rcu_node_0 ->&sb->s_type->i_lock_key#31 ->&c->lock ->&n->list_lock ->&____s->seqcount#2 ->&____s->seqcount FD: 128 BD: 1 ++++: kn->active#55 ->fs_reclaim ->&kernfs_locks->open_file_mutex[count] ->stock_lock ->&c->lock ->&____s->seqcount#2 ->&____s->seqcount ->&n->list_lock FD: 1 BD: 145 ....: cpuset_attach_wq.lock FD: 2 BD: 4633 ..-.: stock_lock ->per_cpu_ptr(&cgroup_rstat_cpu_lock, cpu) FD: 128 BD: 1 .+.+: kn->active#56 ->fs_reclaim ->stock_lock ->pool_lock#2 ->&kernfs_locks->open_file_mutex[count] ->&c->lock ->&____s->seqcount#2 ->&____s->seqcount ->&n->list_lock FD: 129 BD: 1 .+.+: kn->active#57 ->fs_reclaim ->stock_lock ->&kernfs_locks->open_file_mutex[count] ->memcg_max_mutex ->&c->lock ->&____s->seqcount#2 ->&n->list_lock FD: 1 BD: 8 +.+.: memcg_max_mutex FD: 1 BD: 3874 +.-.: &local->active_txq_lock[i] FD: 36 BD: 3867 +.-.: &local->handle_wake_tx_queue_lock ->&local->active_txq_lock[i] ->&local->queue_stop_reason_lock ->&fq->lock ->tk_core.seq.seqcount ->hwsim_radio_lock ->&list->lock#19 FD: 1 BD: 3880 ..-.: &local->queue_stop_reason_lock FD: 1 BD: 12 ....: &per_cpu(xt_recseq, i) FD: 288 BD: 1 +.+.: nf_nat_proto_mutex ->fs_reclaim ->pool_lock#2 ->nf_hook_mutex ->cpu_hotplug_lock ->&obj_hash[i].lock ->stock_lock ->&c->lock ->&____s->seqcount#2 ->&____s->seqcount FD: 1 BD: 125 ..-.: elock-AF_INET6 FD: 25 BD: 1 +.+.: loop_validate_mutex ->&lo->lo_mutex ->&rq->__lock ->loop_validate_mutex.wait_lock ->&cfs_rq->removed.lock ->&obj_hash[i].lock ->pool_lock#2 FD: 1 BD: 5 +.+.: &pnsocks.lock FD: 1 BD: 4 +.+.: resource_mutex FD: 1 BD: 3 +...: clock-AF_PHONET FD: 1 BD: 3 ....: rlock-AF_PHONET FD: 1 BD: 3911 +.-.: &nf_nat_locks[i] FD: 35 BD: 79 +.-.: (&peer->timer_persistent_keepalive) ->pool_lock#2 ->&list->lock#17 ->tk_core.seq.seqcount ->batched_entropy_u8.lock ->kfence_freelist_lock ->&c->lock ->&n->list_lock ->&____s->seqcount#2 ->&____s->seqcount ->&zone->lock ->init_task.mems_allowed_seq.seqcount FD: 126 BD: 1 +.+.: (wq_completion)wg-kex-wg0#3 ->(work_completion)(&peer->transmit_handshake_work) FD: 126 BD: 1 +.+.: (wq_completion)wg-kex-wg1#3 ->(work_completion)(&peer->transmit_handshake_work) FD: 168 BD: 1 +.+.: (wq_completion)wg-kex-wg0#4 ->(work_completion)(&({ do { const void *__vpp_verify = (typeof((worker) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __asm__ ("" : "=r"(__ptr) : "0"((typeof(*((worker))) *)((worker)))); (typeof((typeof(*((worker))) *)((worker)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); })->work) FD: 168 BD: 1 +.+.: (wq_completion)wg-kex-wg1#4 ->(work_completion)(&({ do { const void *__vpp_verify = (typeof((worker) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __asm__ ("" : "=r"(__ptr) : "0"((typeof(*((worker))) *)((worker)))); (typeof((typeof(*((worker))) *)((worker)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); })->work) FD: 169 BD: 1 +.+.: (wq_completion)wg-crypt-wg1#2 ->(work_completion)(&({ do { const void *__vpp_verify = (typeof((worker) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __asm__ ("" : "=r"(__ptr) : "0"((typeof(*((worker))) *)((worker)))); (typeof((typeof(*((worker))) *)((worker)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); })->work) ->(work_completion)(&peer->transmit_packet_work) FD: 169 BD: 1 +.+.: (wq_completion)wg-crypt-wg0#2 ->(work_completion)(&({ do { const void *__vpp_verify = (typeof((worker) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __asm__ ("" : "=r"(__ptr) : "0"((typeof(*((worker))) *)((worker)))); (typeof((typeof(*((worker))) *)((worker)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); })->work) ->(work_completion)(&peer->transmit_packet_work) ->&rq->__lock FD: 126 BD: 1 +.+.: (wq_completion)wg-kex-wg2#3 ->(work_completion)(&peer->transmit_handshake_work) FD: 168 BD: 1 +.+.: (wq_completion)wg-kex-wg2#4 ->(work_completion)(&({ do { const void *__vpp_verify = (typeof((worker) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __asm__ ("" : "=r"(__ptr) : "0"((typeof(*((worker))) *)((worker)))); (typeof((typeof(*((worker))) *)((worker)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); })->work) FD: 169 BD: 1 +.+.: (wq_completion)wg-crypt-wg2#2 ->&rq->__lock ->(work_completion)(&({ do { const void *__vpp_verify = (typeof((worker) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __asm__ ("" : "=r"(__ptr) : "0"((typeof(*((worker))) *)((worker)))); (typeof((typeof(*((worker))) *)((worker)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); })->work) ->(work_completion)(&peer->transmit_packet_work) FD: 30 BD: 1 +.+.: (wq_completion)phy5 ->(work_completion)(&local->reconfig_filter) FD: 126 BD: 1 +.+.: (wq_completion)wg-kex-wg0#5 ->(work_completion)(&peer->transmit_handshake_work) FD: 126 BD: 1 +.+.: (wq_completion)wg-kex-wg1#5 ->(work_completion)(&peer->transmit_handshake_work) FD: 168 BD: 1 +.+.: (wq_completion)wg-kex-wg0#6 ->(work_completion)(&({ do { const void *__vpp_verify = (typeof((worker) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __asm__ ("" : "=r"(__ptr) : "0"((typeof(*((worker))) *)((worker)))); (typeof((typeof(*((worker))) *)((worker)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); })->work) FD: 168 BD: 1 +.+.: (wq_completion)wg-kex-wg1#6 ->(work_completion)(&({ do { const void *__vpp_verify = (typeof((worker) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __asm__ ("" : "=r"(__ptr) : "0"((typeof(*((worker))) *)((worker)))); (typeof((typeof(*((worker))) *)((worker)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); })->work) FD: 169 BD: 1 +.+.: (wq_completion)wg-crypt-wg1#3 ->(work_completion)(&({ do { const void *__vpp_verify = (typeof((worker) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __asm__ ("" : "=r"(__ptr) : "0"((typeof(*((worker))) *)((worker)))); (typeof((typeof(*((worker))) *)((worker)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); })->work) ->(work_completion)(&peer->transmit_packet_work) FD: 169 BD: 1 +.+.: (wq_completion)wg-crypt-wg0#3 ->(work_completion)(&({ do { const void *__vpp_verify = (typeof((worker) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __asm__ ("" : "=r"(__ptr) : "0"((typeof(*((worker))) *)((worker)))); (typeof((typeof(*((worker))) *)((worker)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); })->work) ->(work_completion)(&peer->transmit_packet_work) FD: 30 BD: 1 +.+.: (wq_completion)phy6 ->(work_completion)(&local->reconfig_filter) FD: 127 BD: 1 +.+.: (wq_completion)wg-kex-wg0#7 ->(work_completion)(&peer->transmit_handshake_work) ->(work_completion)(&peer->clear_peer_work) FD: 126 BD: 1 +.+.: (wq_completion)wg-kex-wg0#8 ->(work_completion)(&peer->transmit_handshake_work) FD: 126 BD: 1 +.+.: (wq_completion)wg-kex-wg0#9 ->(work_completion)(&peer->transmit_handshake_work) FD: 23 BD: 10 +.+.: &fn->fou_lock ->&rq->__lock FD: 168 BD: 1 +.+.: (wq_completion)wg-kex-wg0#10 ->(work_completion)(&({ do { const void *__vpp_verify = (typeof((worker) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __asm__ ("" : "=r"(__ptr) : "0"((typeof(*((worker))) *)((worker)))); (typeof((typeof(*((worker))) *)((worker)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); })->work) FD: 898 BD: 12 ++++: rdma_nets_rwsem ->rdma_nets.xa_lock ->&device->compat_devs_mutex ->&rq->__lock ->&lock->wait_lock ->&p->pi_lock ->rdma_nets_rwsem.wait_lock FD: 317 BD: 11 +.+.: ipvs->sync_mutex ->fs_reclaim ->pool_lock#2 ->&c->lock ->stock_lock ->mmu_notifier_invalidate_range_start ->&sb->s_type->i_lock_key#8 ->&dir->lock ->&obj_hash[i].lock ->k-sk_lock-AF_INET ->k-slock-AF_INET ->kthread_create_lock ->&p->pi_lock ->&x->wait ->&rq->__lock ->&ipvs->sync_buff_lock ->(console_sem).lock ->console_owner_lock ->console_owner ->&cfs_rq->removed.lock ->&n->list_lock ->rtnl_mutex.wait_lock ->&mm->mmap_lock FD: 126 BD: 1 +.+.: (wq_completion)wg-kex-wg2#5 ->(work_completion)(&peer->transmit_handshake_work) FD: 169 BD: 1 +.+.: (wq_completion)wg-crypt-wg0#4 ->&rq->__lock ->(work_completion)(&({ do { const void *__vpp_verify = (typeof((worker) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __asm__ ("" : "=r"(__ptr) : "0"((typeof(*((worker))) *)((worker)))); (typeof((typeof(*((worker))) *)((worker)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); })->work) ->(work_completion)(&peer->transmit_packet_work) FD: 168 BD: 1 +.+.: (wq_completion)wg-kex-wg2#6 ->(work_completion)(&({ do { const void *__vpp_verify = (typeof((worker) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __asm__ ("" : "=r"(__ptr) : "0"((typeof(*((worker))) *)((worker)))); (typeof((typeof(*((worker))) *)((worker)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); })->work) FD: 169 BD: 1 +.+.: (wq_completion)wg-crypt-wg2#3 ->(work_completion)(&({ do { const void *__vpp_verify = (typeof((worker) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __asm__ ("" : "=r"(__ptr) : "0"((typeof(*((worker))) *)((worker)))); (typeof((typeof(*((worker))) *)((worker)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); })->work) ->(work_completion)(&peer->transmit_packet_work) ->&rq->__lock FD: 126 BD: 1 +.+.: (wq_completion)wg-kex-wg1#9 ->(work_completion)(&peer->transmit_handshake_work) FD: 168 BD: 1 +.+.: (wq_completion)wg-kex-wg0#11 ->(work_completion)(&({ do { const void *__vpp_verify = (typeof((worker) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __asm__ ("" : "=r"(__ptr) : "0"((typeof(*((worker))) *)((worker)))); (typeof((typeof(*((worker))) *)((worker)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); })->work) FD: 126 BD: 1 +.+.: (wq_completion)wg-kex-wg1#10 ->(work_completion)(&peer->transmit_handshake_work) FD: 168 BD: 1 +.+.: (wq_completion)wg-kex-wg1#11 ->(work_completion)(&({ do { const void *__vpp_verify = (typeof((worker) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __asm__ ("" : "=r"(__ptr) : "0"((typeof(*((worker))) *)((worker)))); (typeof((typeof(*((worker))) *)((worker)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); })->work) FD: 168 BD: 1 +.+.: (wq_completion)wg-kex-wg0#12 ->(work_completion)(&({ do { const void *__vpp_verify = (typeof((worker) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __asm__ ("" : "=r"(__ptr) : "0"((typeof(*((worker))) *)((worker)))); (typeof((typeof(*((worker))) *)((worker)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); })->work) FD: 168 BD: 1 +.+.: (wq_completion)wg-kex-wg1#12 ->(work_completion)(&({ do { const void *__vpp_verify = (typeof((worker) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __asm__ ("" : "=r"(__ptr) : "0"((typeof(*((worker))) *)((worker)))); (typeof((typeof(*((worker))) *)((worker)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); })->work) FD: 126 BD: 1 +.+.: (wq_completion)wg-kex-wg2#7 ->(work_completion)(&peer->transmit_handshake_work) FD: 169 BD: 1 +.+.: (wq_completion)wg-crypt-wg1#5 ->(work_completion)(&({ do { const void *__vpp_verify = (typeof((worker) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __asm__ ("" : "=r"(__ptr) : "0"((typeof(*((worker))) *)((worker)))); (typeof((typeof(*((worker))) *)((worker)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); })->work) ->(work_completion)(&peer->transmit_packet_work) ->&rq->__lock FD: 169 BD: 1 +.+.: (wq_completion)wg-crypt-wg0#5 ->(work_completion)(&({ do { const void *__vpp_verify = (typeof((worker) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __asm__ ("" : "=r"(__ptr) : "0"((typeof(*((worker))) *)((worker)))); (typeof((typeof(*((worker))) *)((worker)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); })->work) ->(work_completion)(&peer->transmit_packet_work) ->&rq->__lock FD: 168 BD: 1 +.+.: (wq_completion)wg-kex-wg2#8 ->(work_completion)(&({ do { const void *__vpp_verify = (typeof((worker) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __asm__ ("" : "=r"(__ptr) : "0"((typeof(*((worker))) *)((worker)))); (typeof((typeof(*((worker))) *)((worker)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); })->work) FD: 169 BD: 1 +.+.: (wq_completion)wg-crypt-wg1#6 ->&rq->__lock ->(work_completion)(&({ do { const void *__vpp_verify = (typeof((worker) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __asm__ ("" : "=r"(__ptr) : "0"((typeof(*((worker))) *)((worker)))); (typeof((typeof(*((worker))) *)((worker)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); })->work) ->(work_completion)(&peer->transmit_packet_work) FD: 169 BD: 1 +.+.: (wq_completion)wg-crypt-wg2#4 ->(work_completion)(&({ do { const void *__vpp_verify = (typeof((worker) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __asm__ ("" : "=r"(__ptr) : "0"((typeof(*((worker))) *)((worker)))); (typeof((typeof(*((worker))) *)((worker)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); })->work) ->(work_completion)(&peer->transmit_packet_work) FD: 169 BD: 1 +.+.: (wq_completion)wg-crypt-wg0#6 ->(work_completion)(&({ do { const void *__vpp_verify = (typeof((worker) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __asm__ ("" : "=r"(__ptr) : "0"((typeof(*((worker))) *)((worker)))); (typeof((typeof(*((worker))) *)((worker)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); })->work) ->(work_completion)(&peer->transmit_packet_work) ->&rq->__lock FD: 127 BD: 1 +.+.: (wq_completion)wg-kex-wg2#9 ->(work_completion)(&peer->transmit_handshake_work) ->&rq->__lock ->(work_completion)(&peer->clear_peer_work) FD: 168 BD: 1 +.+.: (wq_completion)wg-kex-wg2#10 ->(work_completion)(&({ do { const void *__vpp_verify = (typeof((worker) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __asm__ ("" : "=r"(__ptr) : "0"((typeof(*((worker))) *)((worker)))); (typeof((typeof(*((worker))) *)((worker)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); })->work) FD: 169 BD: 1 +.+.: (wq_completion)wg-crypt-wg2#5 ->(work_completion)(&({ do { const void *__vpp_verify = (typeof((worker) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __asm__ ("" : "=r"(__ptr) : "0"((typeof(*((worker))) *)((worker)))); (typeof((typeof(*((worker))) *)((worker)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); })->work) ->(work_completion)(&peer->transmit_packet_work) FD: 126 BD: 1 +.+.: (wq_completion)wg-kex-wg2#11 ->(work_completion)(&peer->transmit_handshake_work) FD: 168 BD: 1 +.+.: (wq_completion)wg-kex-wg2#12 ->(work_completion)(&({ do { const void *__vpp_verify = (typeof((worker) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __asm__ ("" : "=r"(__ptr) : "0"((typeof(*((worker))) *)((worker)))); (typeof((typeof(*((worker))) *)((worker)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); })->work) FD: 169 BD: 1 +.+.: (wq_completion)wg-crypt-wg2#6 ->(work_completion)(&({ do { const void *__vpp_verify = (typeof((worker) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __asm__ ("" : "=r"(__ptr) : "0"((typeof(*((worker))) *)((worker)))); (typeof((typeof(*((worker))) *)((worker)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); })->work) ->(work_completion)(&peer->transmit_packet_work) FD: 30 BD: 81 +.+.: (wq_completion)phy7 ->(work_completion)(&local->reconfig_filter) ->&rq->__lock FD: 135 BD: 4 +.+.: pfkey_mutex ->crypto_alg_sem ->(kmod_concurrent_max).lock ->fs_reclaim ->pool_lock#2 ->&c->lock ->&obj_hash[i].lock ->&x->wait#17 ->&rq->__lock ->running_helpers_waitq.lock ->(crypto_chain).rwsem ->&x->wait#21 ->&base->lock ->(&timer.timer) ->remove_cache_srcu ->&n->list_lock ->quarantine_lock ->&cfs_rq->removed.lock ->&____s->seqcount#2 ->&____s->seqcount ->pgd_lock ->key ->pcpu_lock ->percpu_counters_lock FD: 30 BD: 81 +.+.: (wq_completion)phy8 ->(work_completion)(&local->reconfig_filter) FD: 30 BD: 1 +.+.: (wq_completion)phy9 ->(work_completion)(&local->reconfig_filter) FD: 23 BD: 1 +.+.: &type->s_umount_key#50 ->&rq->__lock FD: 190 BD: 1 +.+.: (wq_completion)phy10 ->(work_completion)(&local->reconfig_filter) ->(work_completion)(&link->csa_finalize_work) ->(work_completion)(&(&local->roc_work)->work) FD: 1 BD: 3 +...: base_sockets.lock FD: 30 BD: 81 +.+.: (wq_completion)phy11 ->(work_completion)(&local->reconfig_filter) FD: 1 BD: 3848 .+.-: &table->lock#3 FD: 1 BD: 3 +...: clock-AF_KEY FD: 1 BD: 3 ....: wlock-AF_KEY FD: 1 BD: 4 ....: rlock-AF_KEY FD: 1 BD: 78 +.+.: &net->ipv4.ra_mutex FD: 158 BD: 1 +.+.: (wq_completion)phy12 ->(work_completion)(&local->reconfig_filter) ->(work_completion)(&(&local->roc_work)->work) ->(work_completion)(&(&local->scan_work)->work) FD: 190 BD: 1 +.+.: (wq_completion)phy13 ->(work_completion)(&local->reconfig_filter) ->(work_completion)(&link->csa_finalize_work) ->(work_completion)(&(&local->scan_work)->work) FD: 30 BD: 81 +.+.: (wq_completion)phy14 ->(work_completion)(&local->reconfig_filter) FD: 1 BD: 4 +...: clock-AF_ISDN FD: 1 BD: 3 +...: clock-AF_ROSE FD: 241 BD: 3 +.+.: sk_lock-AF_ROSE ->slock-AF_ROSE ->rose_list_lock ->&obj_hash[i].lock ->wlock-AF_ROSE ->&list->lock#20 ->rlock-AF_ROSE ->rose_node_list_lock ->&rq->__lock ->fs_reclaim ->stock_lock ->&f->f_lock ->&mm->mmap_lock FD: 1 BD: 4 +...: slock-AF_ROSE FD: 1 BD: 4 ....: wlock-AF_ROSE FD: 1 BD: 4 ....: &list->lock#20 FD: 1 BD: 4 +...: rose_list_lock FD: 1 BD: 4 ....: rlock-AF_ROSE FD: 1 BD: 83 +...: l2tp_ip_lock FD: 4 BD: 5 +.+.: &q->instances_lock ->pool_lock#2 ->&obj_hash[i].lock FD: 1 BD: 4 +...: &log->instances_lock FD: 24 BD: 165 ....: &sem->waiters ->&p->pi_lock FD: 16 BD: 104 +.-.: _xmit_NONE#2 ->&data->lock ->&obj_hash[i].lock ->pool_lock#2 ->elock-AF_INET6 ->quarantine_lock FD: 241 BD: 81 +.+.: __ip_vs_mutex ->&ipvs->dest_trash_lock ->(console_sem).lock ->&rq->__lock ->&s->lock ->&mm->mmap_lock ->&cfs_rq->removed.lock ->&obj_hash[i].lock ->pool_lock#2 FD: 1 BD: 82 +...: &ipvs->dest_trash_lock FD: 1 BD: 87 +.+.: flowtable_lock FD: 1 BD: 78 +.+.: raw_notifier_lock FD: 1 BD: 78 +.+.: bcm_notifier_lock FD: 1 BD: 78 +.+.: isotp_notifier_lock FD: 190 BD: 155 .+.+: sb_pagefaults ->tk_core.seq.seqcount ->mmu_notifier_invalidate_range_start ->pool_lock#2 ->&journal->j_state_lock ->jbd2_handle ->&obj_hash[i].lock ->mapping.invalidate_lock ->&rq->__lock ->remove_cache_srcu ->&c->lock ->&____s->seqcount#2 ->&____s->seqcount ->batched_entropy_u8.lock ->kfence_freelist_lock ->&meta->lock ->&n->list_lock ->&journal->j_wait_transaction_locked ->&sb->s_type->i_lock_key#22 ->&wb->list_lock ->&sem->wait_lock ->&p->pi_lock ->quarantine_lock FD: 1 BD: 4533 +.+.: &pa->pa_lock#2 FD: 1 BD: 1 ....: _rs.lock FD: 30 BD: 1 ..-.: &(&bat_priv->dat.work)->timer FD: 30 BD: 1 ..-.: &(&bat_priv->bla.work)->timer FD: 253 BD: 1 +.+.: sk_lock-AF_ALG ->slock-AF_ALG ->fs_reclaim ->pool_lock#2 ->&mm->mmap_lock ->&obj_hash[i].lock ->&dir->lock ->&ei->socket.wq.wait ->&c->lock ->&____s->seqcount#2 ->&____s->seqcount ->rcu_node_0 ->&rq->__lock ->&n->list_lock ->&sem->wait_lock ->&p->pi_lock ->remove_cache_srcu ->sk_lock-AF_ALG/1 ->&drbg->drbg_mutex ->batched_entropy_u8.lock ->kfence_freelist_lock ->&meta->lock ->quarantine_lock ->&cfs_rq->removed.lock ->(console_sem).lock FD: 25 BD: 3 +...: slock-AF_ALG ->&sk->sk_lock.wq FD: 27 BD: 2 +.+.: (work_completion)(&(&bat_priv->dat.work)->work) ->&hash->list_locks[i] ->&rq->__lock ->&obj_hash[i].lock ->&base->lock ->&cfs_rq->removed.lock ->pool_lock#2 ->rcu_node_0 ->&rcu_state.expedited_wq FD: 1 BD: 8 +...: &hash->list_locks[i] FD: 40 BD: 2 +.+.: (work_completion)(&(&bat_priv->bla.work)->work) ->key#20 ->&obj_hash[i].lock ->&base->lock ->&rq->__lock ->rcu_node_0 ->crngs.lock ->&cfs_rq->removed.lock FD: 193 BD: 81 +.+.: sk_lock-AF_INET6/1 ->slock-AF_INET6 ->rlock-AF_INET6 ->&list->lock#21 ->pool_lock#2 ->&c->lock ->&obj_hash[i].lock ->&base->lock ->&rq->__lock ->fs_reclaim ->tk_core.seq.seqcount ->&list->lock#24 ->&____s->seqcount ->&n->list_lock ->krc.lock ->&____s->seqcount#2 ->rcu_node_0 ->&rcu_state.expedited_wq ->sctp_assocs_id_lock ->quarantine_lock ->&zone->lock ->&cfs_rq->removed.lock ->remove_cache_srcu ->&meta->lock ->kfence_freelist_lock ->&asoc->wait FD: 1 BD: 84 +.-.: rlock-AF_INET6 FD: 1 BD: 84 ....: &list->lock#21 FD: 83 BD: 88 +.-.: slock-AF_INET6/1 ->&sctp_ep_hashtable[i].lock ->&obj_hash[i].lock ->pool_lock#2 ->clock-AF_INET6 ->krc.lock ->&sctp_port_hashtable[i].lock ->tk_core.seq.seqcount ->&base->lock ->&c->lock ->&hashinfo->ehash_locks[i] ->&f->f_owner.lock ->&tcp_hashinfo.bhash[i].lock ->elock-AF_INET6 ->key#25 ->&data->lock ->&zone->lock FD: 1 BD: 91 ++.-: &sctp_ep_hashtable[i].lock FD: 104 BD: 3 +.+.: sk_lock-AF_INET/1 ->slock-AF_INET ->rlock-AF_INET ->&list->lock#21 ->&rq->__lock ->&obj_hash[i].lock ->&base->lock ->pool_lock#2 ->krc.lock ->sctp_assocs_id_lock FD: 1 BD: 5 +.-.: rlock-AF_INET FD: 139 BD: 1 +.+.: crypto_cfg_mutex ->(console_sem).lock ->&rq->__lock ->crypto_alg_sem FD: 128 BD: 79 +.+.: &block->lock ->fs_reclaim ->pool_lock#2 ->&obj_hash[i].lock ->&rq->__lock ->nl_table_lock ->nl_table_wait.lock ->&c->lock ->&____s->seqcount#2 ->&____s->seqcount ->&n->list_lock FD: 127 BD: 78 ++++: &block->cb_lock ->flow_indr_block_lock ->&rq->__lock FD: 126 BD: 79 +.+.: flow_indr_block_lock ->fs_reclaim ->pool_lock#2 ->&c->lock ->&obj_hash[i].lock ->&rq->__lock ->&cfs_rq->removed.lock ->&____s->seqcount#2 ->&____s->seqcount FD: 13 BD: 248 +...: link_idr_lock ->pool_lock#2 ->&obj_hash[i].lock ->&c->lock ->&n->list_lock ->&____s->seqcount#2 ->&____s->seqcount FD: 295 BD: 10 +.+.: tracepoints_mutex ->fs_reclaim ->pool_lock#2 ->cpu_hotplug_lock ->tracepoint_srcu_srcu_usage.lock ->&ACCESS_PRIVATE(sdp, lock) ->&rq->__lock ->&obj_hash[i].lock ->rcu_node_0 ->&c->lock ->tasklist_lock ->rcu_state.exp_mutex.wait_lock ->&p->pi_lock ->tracepoint_srcu ->&x->wait#2 ->&n->list_lock ->tracepoints_mutex.wait_lock ->&cfs_rq->removed.lock ->&rnp->exp_lock ->rcu_state.exp_mutex ->&____s->seqcount#2 ->batched_entropy_u8.lock ->kfence_freelist_lock ->reg_lock ->remove_cache_srcu ->&____s->seqcount FD: 25 BD: 86 +.-.: &idev->mc_query_lock ->&obj_hash[i].lock FD: 23 BD: 84 +.+.: (work_completion)(&(&idev->mc_report_work)->work) ->&rq->__lock FD: 38 BD: 80 +.-.: (&mp->timer) ->&br->multicast_lock FD: 43 BD: 79 +.+.: (work_completion)(&br->mcast_gc_work) ->&br->multicast_lock ->(&mp->timer) ->&obj_hash[i].lock ->&base->lock ->pool_lock#2 ->krc.lock ->(&p->rexmit_timer) ->(&p->timer) ->&rq->__lock FD: 1 BD: 80 ....: (&p->rexmit_timer) FD: 38 BD: 80 +.-.: (&p->timer) ->&br->multicast_lock FD: 243 BD: 81 +.+.: &net->xdp.lock ->&rq->__lock ->&xs->mutex ->&lock->wait_lock ->&p->pi_lock FD: 1 BD: 81 +.+.: mirred_list_lock FD: 1 BD: 81 +...: &idev->mc_report_lock FD: 1 BD: 78 ....: (&pmctx->ip6_mc_router_timer) FD: 1 BD: 78 ....: (&pmctx->ip4_mc_router_timer) FD: 24 BD: 81 +.+.: &pnn->pndevs.lock ->&rq->__lock FD: 23 BD: 81 +.+.: &pnn->routes.lock ->&rq->__lock FD: 1 BD: 13 ....: netdev_unregistering_wq.lock FD: 14 BD: 17 ....: tracepoint_srcu_srcu_usage.lock ->&obj_hash[i].lock ->&ACCESS_PRIVATE(sdp, lock) ->&base->lock FD: 25 BD: 3 +.-.: (&sdp->delay_work) FD: 1 BD: 78 +...: &tn->node_list_lock FD: 1 BD: 5 +.+.: netns_bpf_mutex FD: 70 BD: 79 +.-.: (&peer->timer_retransmit_handshake) ->&peer->endpoint_lock ->&obj_hash[i].lock ->&list->lock#17 FD: 30 BD: 79 +.-.: (&peer->timer_send_keepalive) ->pool_lock#2 ->&list->lock#17 ->tk_core.seq.seqcount ->&c->lock ->&____s->seqcount#2 ->&____s->seqcount ->&n->list_lock ->batched_entropy_u8.lock ->kfence_freelist_lock FD: 69 BD: 79 +.-.: (&peer->timer_new_handshake) ->&peer->endpoint_lock FD: 30 BD: 79 +.-.: (&peer->timer_zero_key_material) FD: 130 BD: 81 +.+.: (work_completion)(&peer->clear_peer_work) ->&handshake->lock ->&peer->keypairs.keypair_update_lock FD: 1 BD: 88 +.+.: rcu_state.barrier_mutex.wait_lock FD: 1 BD: 8 ++..: k-clock-AF_NETLINK FD: 1 BD: 5 +.+.: &hn->hn_lock FD: 133 BD: 1 +.+.: (wq_completion)inet_frag_wq ->(work_completion)(&fqdir->destroy_work) FD: 132 BD: 2 +.+.: (work_completion)(&fqdir->destroy_work) ->(work_completion)(&ht->run_work) ->&ht->mutex ->&rq->__lock FD: 40 BD: 2 +.+.: fqdir_free_work ->rcu_state.barrier_mutex ->&obj_hash[i].lock ->pool_lock#2 ->rcu_state.barrier_mutex.wait_lock ->&p->pi_lock ->&rq->__lock ->quarantine_lock ->rcu_node_0 ->&meta->lock ->kfence_freelist_lock ->&cfs_rq->removed.lock FD: 45 BD: 78 +.+.: &caifn->caifdevs.lock ->&obj_hash[i].lock ->&rq->__lock ->pool_lock#2 ->&this->info_list_lock ->&rnp->exp_wq[2] ->&rnp->exp_lock ->&rnp->exp_wq[0] ->quarantine_lock ->rcu_state.exp_mutex FD: 1 BD: 80 +...: &this->info_list_lock FD: 1 BD: 5 +.+.: &pnetids_ndev->lock FD: 182 BD: 81 +.+.: k-sk_lock-AF_INET6/1 ->k-slock-AF_INET6 ->rlock-AF_INET6 ->&list->lock#21 ->&dir->lock ->fs_reclaim ->k-clock-AF_INET6 ->&c->lock ->&rq->__lock ->&h->lhash2[i].lock ->&tcp_hashinfo.bhash[i].lock ->&queue->rskq_lock ->pool_lock#2 ->&n->list_lock ->&obj_hash[i].lock FD: 102 BD: 88 +.-.: k-slock-AF_INET6/1 ->&sctp_ep_hashtable[i].lock ->&obj_hash[i].lock ->pool_lock#2 ->k-clock-AF_INET6 ->key#25 ->tk_core.seq.seqcount ->clock-AF_INET6 ->&base->lock ->&hashinfo->ehash_locks[i] ->&tcp_hashinfo.bhash[i].lock ->slock-AF_INET6 ->elock-AF_INET6 FD: 23 BD: 5 +.+.: &sn->gssp_lock ->&rq->__lock FD: 1 BD: 8 +.+.: &cd->hash_lock FD: 1 BD: 5 +.+.: xfrm_state_gc_work FD: 17 BD: 79 +...: &net->xfrm.xfrm_state_lock ->hrtimer_bases.lock ->&obj_hash[i].lock ->&base->lock FD: 1 BD: 5 +...: ip6_fl_lock FD: 54 BD: 5 +.-.: (&net->ipv6.ip6_fib_timer) ->&net->ipv6.fib6_gc_lock FD: 1 BD: 78 ....: (&mrt->ipmr_expire_timer) FD: 1 BD: 5 ....: (&ipvs->dest_trash_timer) FD: 1 BD: 5 +.+.: (work_completion)(&(&ipvs->expire_nodest_conn_work)->work) FD: 1 BD: 5 +.+.: (work_completion)(&(&ipvs->est_reload_work)->work) FD: 1 BD: 6 +...: recent_lock FD: 133 BD: 5 +.+.: hashlimit_mutex ->fs_reclaim ->pool_lock#2 ->free_vmap_area_lock ->vmap_area_lock ->&____s->seqcount ->init_mm.page_table_lock ->&c->lock ->proc_subdir_lock ->proc_inum_ida.xa_lock ->&obj_hash[i].lock ->&base->lock ->&rq->__lock ->&ent->pde_unload_lock ->&____s->seqcount#2 ->&n->list_lock ->purge_vmap_area_lock ->remove_cache_srcu FD: 1 BD: 5 +.+.: trans_gc_work FD: 31 BD: 6 +.+.: (work_completion)(&(&cnet->ecache.dwork)->work) ->&cnet->ecache.dying_lock ->&obj_hash[i].lock ->pool_lock#2 ->&c->lock ->&rq->__lock ->&base->lock ->&n->list_lock ->quarantine_lock FD: 1 BD: 5 +.+.: (work_completion)(&net->xfrm.policy_hash_work) FD: 56 BD: 86 +...: &net->xfrm.xfrm_policy_lock ->pool_lock#2 ->&obj_hash[i].lock ->&base->lock ->&____s->seqcount#16 FD: 1 BD: 5 +.+.: (work_completion)(&net->xfrm.state_hash_work) FD: 1 BD: 1 +.+.: &mq_lock FD: 125 BD: 2 +.+.: free_ipc_work ->&rnp->exp_lock ->rcu_state.exp_mutex ->&obj_hash[i].lock ->mount_lock ->&fsnotify_mark_srcu ->&type->s_umount_key#51 ->unnamed_dev_ida.xa_lock ->list_lrus_mutex ->&xa->xa_lock#4 ->pool_lock#2 ->sb_lock ->mnt_id_ida.xa_lock ->&ids->rwsem ->(work_completion)(&ht->run_work) ->&ht->mutex ->percpu_counters_lock ->pcpu_lock ->sysctl_lock ->proc_inum_ida.xa_lock ->stock_lock ->&rnp->exp_wq[3] ->rcu_state.exp_mutex.wait_lock ->&p->pi_lock ->&rnp->exp_wq[1] ->&rq->__lock ->rcu_node_0 ->&rcu_state.expedited_wq ->quarantine_lock ->&pool->lock ->&cfs_rq->removed.lock ->&rnp->exp_wq[2] ->&rnp->exp_wq[0] ->per_cpu_ptr(&cgroup_rstat_cpu_lock, cpu) FD: 125 BD: 3 +.+.: &type->s_umount_key#51 ->shrinker_rwsem ->rename_lock.seqcount ->&dentry->d_lock ->&sb->s_type->i_lock_key#20 ->&s->s_inode_list_lock ->&xa->xa_lock#9 ->&obj_hash[i].lock ->pool_lock#2 ->&fsnotify_mark_srcu ->&rq->__lock FD: 1 BD: 3 +.+.: &ids->rwsem FD: 1 BD: 78 +...: &bat_priv->forw_bcast_list_lock FD: 1 BD: 83 +...: &batadv_netdev_xmit_lock_key FD: 1 BD: 6 +...: &bat_priv->gw.list_lock FD: 1 BD: 1 +.+.: (work_completion)(&(&bat_priv->bat_v.ogm_wq)->work) FD: 23 BD: 1 +.+.: &bat_priv->bat_v.ogm_buff_mutex ->&obj_hash[i].lock ->pool_lock#2 ->&rq->__lock FD: 289 BD: 1 +.+.: bpf_stats_enabled_mutex ->&newf->file_lock ->fs_reclaim ->pool_lock#2 ->&xa->xa_lock#4 ->&rq->__lock ->&obj_hash[i].lock ->stock_lock ->&sb->s_type->i_lock_key#15 ->cpu_hotplug_lock ->&c->lock ->&n->list_lock ->key#8 ->batched_entropy_u8.lock ->kfence_freelist_lock ->&____s->seqcount#2 ->&____s->seqcount FD: 1038 BD: 1 +.+.: (wq_completion)netns ->net_cleanup_work FD: 1037 BD: 2 +.+.: net_cleanup_work ->pernet_ops_rwsem ->rcu_state.barrier_mutex ->&obj_hash[i].lock ->pool_lock#2 ->krc.lock ->&dir->lock ->stock_lock ->rcu_state.barrier_mutex.wait_lock ->&p->pi_lock ->&rq->__lock ->quarantine_lock ->per_cpu_ptr(&cgroup_rstat_cpu_lock, cpu) ->&base->lock FD: 1 BD: 5 +...: &net->nsid_lock FD: 1 BD: 5 ....: (&net->fs_probe_timer) FD: 1 BD: 7 ++++: &net->cells_lock FD: 1 BD: 5 ....: (&net->cells_timer) FD: 29 BD: 1 +.+.: (wq_completion)afs ->(work_completion)(&net->cells_manager) ->(work_completion)(&net->fs_manager) FD: 26 BD: 2 +.+.: (work_completion)(&net->cells_manager) ->&net->cells_lock ->bit_wait_table + i ->&rq->__lock FD: 1 BD: 5 ....: (&net->fs_timer) FD: 26 BD: 2 +.+.: (work_completion)(&net->fs_manager) ->&(&net->fs_lock)->lock ->bit_wait_table + i ->&rq->__lock ->&cfs_rq->removed.lock ->&obj_hash[i].lock ->pool_lock#2 FD: 1 BD: 3 +.+.: &(&net->fs_lock)->lock FD: 1 BD: 6 +.+.: &rx->incoming_lock FD: 1 BD: 6 +.+.: &call->notify_lock FD: 1 BD: 6 ....: (rxrpc_call_limiter).lock FD: 1 BD: 6 +.+.: &rx->recvmsg_lock FD: 1 BD: 6 ....: (&call->timer) FD: 1 BD: 6 ....: &list->lock#22 FD: 23 BD: 5 +.+.: (wq_completion)kafsd ->&rq->__lock FD: 1 BD: 5 +...: k-clock-AF_RXRPC FD: 1 BD: 7 ..-.: rlock-AF_RXRPC FD: 1 BD: 1 ....: (&local->client_conn_reap_timer) FD: 1 BD: 1 ....: &list->lock#23 FD: 1 BD: 8 +.+.: (work_completion)(&data->gc_work) FD: 1 BD: 5 +.+.: (work_completion)(&ovs_net->dp_notify_work) FD: 10 BD: 97 +.-.: &srv->idr_lock ->pool_lock#2 ->&obj_hash[i].lock ->&c->lock ->&n->list_lock FD: 1 BD: 5 +.+.: (work_completion)(&tn->work) FD: 1 BD: 5 ....: (&rxnet->service_conn_reap_timer) FD: 1 BD: 94 +...: &nt->cluster_scope_lock FD: 1 BD: 5 +.+.: (work_completion)(&(&c->work)->work) FD: 1 BD: 154 ....: key#22 FD: 221 BD: 5 +.+.: (wq_completion)krdsd ->(work_completion)(&rtn->rds_tcp_accept_w) ->(work_completion)(&(&cp->cp_send_w)->work) ->(work_completion)(&(&cp->cp_recv_w)->work) ->(work_completion)(&cp->cp_down_w) FD: 214 BD: 6 +.+.: (work_completion)(&rtn->rds_tcp_accept_w) ->fs_reclaim ->pool_lock#2 ->mmu_notifier_invalidate_range_start ->&sb->s_type->i_lock_key#8 ->k-sk_lock-AF_INET6 ->k-slock-AF_INET6 ->&obj_hash[i].lock ->once_lock ->&____s->seqcount#2 ->&____s->seqcount ->&c->lock ->rds_cong_lock ->&n->list_lock ->rds_trans_sem ->&tc->t_conn_path_lock ->&xa->xa_lock#9 ->&fsnotify_mark_srcu FD: 1 BD: 7 ....: rds_tcp_conn_lock FD: 1 BD: 5 ....: loop_conns_lock FD: 47 BD: 5 +.+.: (wq_completion)l2tp ->(work_completion)(&tunnel->del_work) FD: 1 BD: 149 +.+.: jump_label_mutex.wait_lock FD: 2 BD: 8 +.+.: (work_completion)(&rxnet->service_conn_reaper) ->&rxnet->conn_lock FD: 12 BD: 94 +.-.: &sctp_port_hashtable[i].lock ->&____s->seqcount#2 ->&____s->seqcount ->pool_lock#2 ->&c->lock ->&obj_hash[i].lock ->batched_entropy_u8.lock ->kfence_freelist_lock ->&meta->lock FD: 24 BD: 111 ..-.: &asoc->wait ->&p->pi_lock FD: 24 BD: 137 ....: &sk->sk_lock.wq ->&p->pi_lock FD: 1 BD: 111 ..-.: key#23 FD: 12 BD: 268 +.-.: sctp_assocs_id_lock ->pool_lock#2 ->&obj_hash[i].lock ->&c->lock ->&____s->seqcount#2 ->&____s->seqcount ->&n->list_lock FD: 1 BD: 111 ..-.: &list->lock#24 FD: 71 BD: 1 +.-.: (&p->forward_delay_timer) ->&br->lock FD: 248 BD: 3 +.+.: sk_lock-AF_CAIF ->&rq->__lock ->slock-AF_CAIF ->&obj_hash[i].lock ->&this->info_list_lock ->(console_sem).lock ->&ei->socket.wq.wait ->clock-AF_CAIF ->elock-AF_CAIF ->&rnp->exp_lock ->rcu_state.exp_mutex ->&rnp->exp_wq[3] ->rcu_state.exp_mutex.wait_lock ->&p->pi_lock ->&mm->mmap_lock ->console_owner_lock ->console_owner FD: 1 BD: 4 +...: slock-AF_CAIF FD: 1 BD: 3 +...: rlock-AF_CAIF FD: 1 BD: 4 +...: clock-AF_CAIF FD: 1 BD: 4 ....: elock-AF_CAIF FD: 130 BD: 78 +.+.: &chain->filter_chain_lock ->&block->lock ->&block->proto_destroy_lock FD: 1 BD: 79 +.+.: &block->proto_destroy_lock FD: 2 BD: 98 ...-: &new->fa_lock ->&f->f_owner.lock FD: 1 BD: 107 ...-: &f->f_owner.lock FD: 25 BD: 1 ..-.: &(&net->ipv6.addr_chk_work)->timer FD: 1 BD: 4 +.+.: raw_sk_list.lock FD: 155 BD: 84 +.+.: (work_completion)(&(&local->scan_work)->work) ->&local->mtx ->&rq->__lock FD: 13 BD: 91 ..-.: &local->ack_status_lock ->pool_lock#2 ->&obj_hash[i].lock ->&c->lock ->&n->list_lock ->batched_entropy_u8.lock ->kfence_freelist_lock ->&____s->seqcount#2 ->&____s->seqcount FD: 30 BD: 1 ..-.: &(&local->scan_work)->timer FD: 1 BD: 83 +...: &vlan_netdev_xmit_lock_key FD: 29 BD: 87 +.+.: &sta->ampdu_mlme.mtx ->&rq->__lock ->&sta->lock FD: 1 BD: 87 +.+.: (work_completion)(&sta->ampdu_mlme.work) FD: 23 BD: 87 +.+.: (work_completion)(&sta->drv_deliver_wk) ->&rq->__lock FD: 1 BD: 3 +...: clock-AF_NFC FD: 1 BD: 3 ....: rlock-AF_NFC FD: 1 BD: 83 +...: &qdisc_xmit_lock_key FD: 1 BD: 92 ....: fastopen_seqlock.seqcount FD: 892 BD: 1 +.+.: ppp_mutex ->&mm->mmap_lock ->fs_reclaim ->stock_lock ->pool_lock#2 ->stack_depot_init_mutex ->rtnl_mutex ->rtnl_mutex.wait_lock ->&p->pi_lock ->&rq->__lock ->&c->lock ->ppp_mutex.wait_lock ->&cfs_rq->removed.lock ->&obj_hash[i].lock ->remove_cache_srcu ->&n->list_lock ->&ppp->wlock ->&____s->seqcount#2 ->&____s->seqcount ->compressor_list_lock ->&pn->all_ppp_mutex FD: 126 BD: 78 +.+.: &pn->all_ppp_mutex ->&rq->__lock ->fs_reclaim ->pool_lock#2 ->&obj_hash[i].lock ->&c->lock ->&n->list_lock ->remove_cache_srcu FD: 4 BD: 79 +...: &ppp->rlock ->&obj_hash[i].lock ->pool_lock#2 FD: 5 BD: 78 +...: &ppp->wlock ->&ppp->rlock FD: 194 BD: 83 +.+.: k-sk_lock-AF_INET/1 ->k-slock-AF_INET ->pool_lock#2 ->&dir->lock ->fs_reclaim ->&c->lock ->k-clock-AF_INET ->&obj_hash[i].lock ->slock-AF_INET ->&hashinfo->ehash_locks[i] ->&rq->__lock ->tk_core.seq.seqcount ->&base->lock ->&tcp_hashinfo.bhash[i].lock ->&n->list_lock FD: 1 BD: 92 +...: &token_hash[i].lock FD: 101 BD: 1 +.-.: k-slock-AF_INET/1 ->tk_core.seq.seqcount ->slock-AF_INET ->&obj_hash[i].lock ->pool_lock#2 ->&hashinfo->ehash_locks[i] ->&tcp_hashinfo.bhash[i].lock ->&base->lock FD: 1 BD: 78 +...: &dev_addr_list_lock_key#4 FD: 1 BD: 79 ....: &pf->rwait FD: 1 BD: 3 +...: rds_sock_lock FD: 1 BD: 3 +...: clock-AF_NETROM FD: 8 BD: 3 +.+.: sk_lock-AF_NETROM ->slock-AF_NETROM ->&obj_hash[i].lock ->wlock-AF_NETROM ->&list->lock#25 ->nr_list_lock ->rlock-AF_NETROM FD: 1 BD: 4 +...: slock-AF_NETROM FD: 1 BD: 4 ....: wlock-AF_NETROM FD: 1 BD: 4 ....: &list->lock#25 FD: 1 BD: 4 +...: nr_list_lock FD: 1 BD: 4 ....: rlock-AF_NETROM FD: 1 BD: 85 +...: &msk->pm.lock FD: 531 BD: 4 +.+.: (work_completion)(&msk->work) ->sk_lock-AF_INET ->slock-AF_INET ->&rq->__lock ->sk_lock-AF_INET6 ->slock-AF_INET6 FD: 1 BD: 3 +...: clock-AF_RDS FD: 1 BD: 3 ....: &rs->rs_recv_lock FD: 1 BD: 3 ....: rds_cong_monitor_lock FD: 1 BD: 10 ....: rds_cong_lock FD: 1 BD: 3 ....: &rs->rs_lock FD: 1 BD: 3 ....: &rs->rs_rdma_lock FD: 1 BD: 3 ....: &q->lock FD: 1 BD: 4 ....: &list->lock#26 FD: 29 BD: 90 +.-.: (&tw->tw_timer) ->&hashinfo->ehash_locks[i] ->&tcp_hashinfo.bhash[i].lock ->stock_lock ->&obj_hash[i].lock ->&dccp_hashinfo.bhash[i].lock ->per_cpu_ptr(&cgroup_rstat_cpu_lock, cpu) ->&c->lock ->&n->list_lock FD: 1 BD: 2 +.+.: ppp_mutex.wait_lock FD: 1 BD: 3 ++.-: raw_lock FD: 1 BD: 3 +...: clock-AF_IEEE802154 FD: 1 BD: 3 ....: rlock-AF_IEEE802154 FD: 27 BD: 2 +.+.: (work_completion)(&(&hinfo->gc_work)->work) ->&hinfo->lock ->rcu_node_0 ->&rcu_state.expedited_wq ->&rq->__lock ->&cfs_rq->removed.lock ->&obj_hash[i].lock ->pool_lock#2 ->&base->lock FD: 1 BD: 3 +...: &hinfo->lock FD: 24 BD: 122 +.-.: &dccp_hashinfo.bhash[i].lock ->&dccp_hashinfo.bhash2[i].lock ->stock_lock ->pool_lock#2 ->&obj_hash[i].lock ->&____s->seqcount#2 ->&____s->seqcount ->&c->lock ->clock-AF_INET6 ->per_cpu_ptr(&cgroup_rstat_cpu_lock, cpu) ->&meta->lock ->kfence_freelist_lock ->batched_entropy_u8.lock ->&n->list_lock ->&hashinfo->ehash_locks[i] ->quarantine_lock FD: 1 BD: 4 +...: slock-AF_X25 FD: 23 BD: 123 +.-.: &dccp_hashinfo.bhash2[i].lock ->stock_lock ->&____s->seqcount#2 ->&____s->seqcount ->pool_lock#2 ->&c->lock ->clock-AF_INET6 ->batched_entropy_u8.lock ->&hashinfo->ehash_locks[i] ->&obj_hash[i].lock ->kfence_freelist_lock ->&n->list_lock ->per_cpu_ptr(&cgroup_rstat_cpu_lock, cpu) ->&meta->lock ->quarantine_lock ->clock-AF_INET FD: 28 BD: 3 +.+.: sk_lock-AF_X25 ->slock-AF_X25 ->wlock-AF_X25 ->&list->lock#27 ->&rq->__lock ->&obj_hash[i].lock ->x25_list_lock ->rlock-AF_X25 FD: 1 BD: 4 ....: wlock-AF_X25 FD: 1 BD: 4 ....: &list->lock#27 FD: 1 BD: 4 ....: rlock-AF_X25 FD: 1 BD: 1 ....: _rs.lock#2 FD: 49 BD: 93 +...: &bond->mode_lock ->pool_lock#2 ->&obj_hash[i].lock ->&c->lock ->(console_sem).lock ->&____s->seqcount#2 ->&____s->seqcount FD: 160 BD: 1 +.+.: (wq_completion)bond1 ->(work_completion)(&(&slave->notify_work)->work) FD: 2 BD: 1 +...: &list->lock#28 ->rlock-AF_INET FD: 4 BD: 3841 +.-.: icmp_global.lock ->batched_entropy_u8.lock FD: 1 BD: 3 ....: &list->lock#29 FD: 1 BD: 3 +...: clock-AF_RXRPC FD: 48 BD: 3 +.+.: sk_lock-AF_PHONET ->slock-AF_PHONET ->port_mutex#2 ->&pnsocks.lock ->resource_mutex ->&obj_hash[i].lock ->&rq->__lock ->&rnp->exp_wq[0] ->&rnp->exp_wq[1] ->rcu_state.exp_mutex.wait_lock ->&p->pi_lock ->&rnp->exp_lock ->rcu_state.exp_mutex FD: 1 BD: 4 +...: slock-AF_PHONET FD: 25 BD: 4 +.+.: port_mutex#2 ->local_port_range_lock.seqcount ->&rq->__lock ->&pnsocks.lock FD: 1 BD: 5 ....: local_port_range_lock.seqcount FD: 160 BD: 1 +.+.: (wq_completion)bond2 ->(work_completion)(&(&slave->notify_work)->work) FD: 1 BD: 78 ....: (&brmctx->ip4_other_query.timer) FD: 1 BD: 78 ....: (&brmctx->ip6_mc_router_timer) FD: 1 BD: 78 ....: (&brmctx->ip6_other_query.timer) FD: 1 BD: 4652 ...-: init_task.mems_allowed_seq.seqcount FD: 72 BD: 1 .+.+: sb_writers#13 ->mount_lock ->tk_core.seq.seqcount ->&sb->s_type->i_lock_key#27 ->&wb->list_lock FD: 160 BD: 1 +.+.: (wq_completion)bond1#2 ->(work_completion)(&(&slave->notify_work)->work) FD: 1 BD: 78 ....: (&brmctx->ip4_mc_router_timer) FD: 26 BD: 2 +.+.: sk_lock-AF_ALG/1 ->slock-AF_ALG FD: 1 BD: 85 +.+.: &ping_table.lock FD: 1 BD: 78 +...: _xmit_PHONET_PIPE FD: 39 BD: 28 +.+.: &chan->lock/1 ->sk_lock-AF_BLUETOOTH-BTPROTO_L2CAP/1 ->slock-AF_BLUETOOTH-BTPROTO_L2CAP ->clock-AF_BLUETOOTH ->rlock-AF_BLUETOOTH ->wlock-AF_BLUETOOTH ->pool_lock#2 ->&dir->lock ->&obj_hash[i].lock ->&rq->__lock ->&base->lock ->chan_list_lock ->k-clock-AF_BLUETOOTH FD: 1 BD: 3 +...: dgram_lock FD: 26 BD: 29 +.+.: sk_lock-AF_BLUETOOTH-BTPROTO_L2CAP/1 ->slock-AF_BLUETOOTH-BTPROTO_L2CAP ->rfcomm_wq.lock FD: 1 BD: 85 ....: &rdev->wpan_phy.queue_lock FD: 2 BD: 10 +.+.: &match->lock ->ptype_lock FD: 1 BD: 5121 ....: cid_lock FD: 1 BD: 6 +...: &pernet->lock FD: 1 BD: 6 +...: smc_v4_hashinfo.lock FD: 896 BD: 5 +.+.: sk_lock-AF_SMC ->slock-AF_SMC ->&smc->clcsock_release_lock ->k-sk_lock-AF_INET ->k-slock-AF_INET ->k-clock-AF_INET ->smc_v4_hashinfo.lock ->clock-AF_SMC ->&rq->__lock FD: 1 BD: 6 +...: slock-AF_SMC FD: 892 BD: 6 +.+.: &smc->clcsock_release_lock ->&net->smc.mutex_fback_rsn ->k-clock-AF_INET ->k-sk_lock-AF_INET ->k-slock-AF_INET ->pool_lock#2 ->&dir->lock ->&obj_hash[i].lock ->stock_lock ->&sb->s_type->i_lock_key#8 ->&xa->xa_lock#9 ->&fsnotify_mark_srcu ->&rq->__lock ->&c->lock ->&n->list_lock ->nf_sockopt_mutex ->&mm->mmap_lock ->rtnl_mutex ->rtnl_mutex.wait_lock ->&p->pi_lock ->pgd_lock ->key ->pcpu_lock ->percpu_counters_lock ->ipvs->sync_mutex ->&table->hash[i].lock ->&lock->wait_lock ->&xt[i].mutex ->&meta->lock ->kfence_freelist_lock FD: 1 BD: 85 ..-.: &list->lock#30 FD: 1 BD: 85 ....: &rdev->wpan_phy.sync_txq FD: 1 BD: 7 +.+.: &net->smc.mutex_fback_rsn FD: 899 BD: 4 +.+.: (work_completion)(&smc->connect_work) ->k-sk_lock-AF_INET ->k-slock-AF_INET ->sk_lock-AF_SMC ->slock-AF_SMC FD: 1 BD: 6 +...: clock-AF_SMC FD: 900 BD: 1 +.+.: (wq_completion)smc_hs_wq ->(work_completion)(&smc->connect_work) FD: 1 BD: 81 +.+.: acaddr_hash_lock FD: 1 BD: 8 +...: l2tp_ip6_lock FD: 23 BD: 3 +.+.: &knet->mutex ->&rq->__lock FD: 1 BD: 83 +...: &mux->lock FD: 2 BD: 131 +...: &mux->rx_lock ->rlock-AF_KCM FD: 460 BD: 3 +.+.: sk_lock-AF_KCM ->slock-AF_KCM ->clock-AF_KCM ->fs_reclaim ->pool_lock#2 ->&____s->seqcount ->&obj_hash[i].lock ->&mm->mmap_lock ->&mux->lock ->&rq->__lock ->&c->lock ->rcu_node_0 ->remove_cache_srcu ->&sem->wait_lock ->&p->pi_lock ->&n->list_lock ->&cfs_rq->removed.lock ->batched_entropy_u8.lock ->kfence_freelist_lock ->&meta->lock ->&____s->seqcount#2 ->sk_lock-AF_INET ->slock-AF_INET ->&data->lock FD: 25 BD: 4 +...: slock-AF_KCM ->&sk->sk_lock.wq FD: 1 BD: 4 +...: clock-AF_KCM FD: 1 BD: 3 +.+.: (work_completion)(&kcm->tx_work) FD: 1 BD: 132 ....: rlock-AF_KCM FD: 250 BD: 78 +.+.: sk_lock-AF_TIPC ->slock-AF_TIPC ->&rq->__lock ->fs_reclaim ->pool_lock#2 ->&mm->mmap_lock ->&list->lock#31 ->&obj_hash[i].lock ->&base->lock ->clock-AF_TIPC ->&c->lock ->&n->list_lock ->&ei->socket.wq.wait ->&____s->seqcount#2 ->&____s->seqcount ->rcu_node_0 ->&zone->lock ->remove_cache_srcu ->&rcu_state.expedited_wq ->&srv->idr_lock ->&tn->nametbl_lock ->&con->sub_lock ->&tipc_net(net)->bclock ->&con->outqueue_lock ->&list->lock#40 ->quarantine_lock ->batched_entropy_u8.lock ->kfence_freelist_lock ->&meta->lock ->&cfs_rq->removed.lock ->tk_core.seq.seqcount ->&list->lock#5 ->pcpu_lock ->&sem->wait_lock ->&p->pi_lock ->&data->lock FD: 34 BD: 79 +...: slock-AF_TIPC ->&list->lock#31 ->&obj_hash[i].lock ->&base->lock ->pool_lock#2 ->&list->lock#38 ->&c->lock ->&____s->seqcount ->&____s->seqcount#2 ->&n->list_lock ->&meta->lock ->kfence_freelist_lock FD: 1 BD: 94 +...: &list->lock#31 FD: 1 BD: 79 ++..: clock-AF_TIPC FD: 928 BD: 1 .+.+: &rdma_nl_types[idx].sem ->nlk_cb_mutex-RDMA ->link_ops_rwsem ->&rq->__lock FD: 924 BD: 2 +.+.: nlk_cb_mutex-RDMA ->fs_reclaim ->pool_lock#2 ->devices_rwsem ->rlock-AF_NETLINK FD: 129 BD: 155 +.+.: &po->pg_vec_lock ->wlock-AF_PACKET ->&vma->vm_lock->lock ->fs_reclaim ->&____s->seqcount ->pool_lock#2 ->stock_lock ->ptlock_ptr(page) ->ptlock_ptr(page)#2 ->rlock-AF_PACKET ->&rq->__lock ->&c->lock FD: 25 BD: 1 ..-.: &(&hinfo->gc_work)->timer FD: 4 BD: 9 +...: vsock_table_lock ->batched_entropy_u32.lock FD: 246 BD: 7 +.+.: sk_lock-AF_VSOCK ->&rq->__lock ->slock-AF_VSOCK ->vsock_table_lock ->clock-AF_VSOCK ->rlock-AF_VSOCK ->fs_reclaim ->&c->lock ->&n->list_lock ->pool_lock#2 ->&vvs->rx_lock ->&obj_hash[i].lock ->&mm->mmap_lock ->&____s->seqcount#2 ->&____s->seqcount ->&list->lock#42 ->&pool->lock ->&dir->lock ->sk_lock-AF_VSOCK/1 ->&ei->socket.wq.wait ->&vvs->tx_lock ->&zone->lock ->&base->lock ->&data->lock FD: 25 BD: 9 +...: slock-AF_VSOCK ->&sk->sk_lock.wq FD: 1 BD: 9 +...: clock-AF_VSOCK FD: 1 BD: 9 ....: rlock-AF_VSOCK FD: 1 BD: 3 ..-.: rlock-AF_CAN FD: 1 BD: 84 +.+.: (work_completion)(&(&priv->gc_work)->work) FD: 244 BD: 78 +.+.: sk_lock-AF_CAN ->slock-AF_CAN ->&obj_hash[i].lock ->&rq->__lock ->clock-AF_CAN ->&rnp->exp_lock ->rcu_state.exp_mutex ->pcpu_lock ->proc_subdir_lock ->fs_reclaim ->pool_lock#2 ->proc_inum_ida.xa_lock ->&ent->pde_unload_lock ->pool_lock ->&mm->mmap_lock ->rcu_state.exp_mutex.wait_lock ->&p->pi_lock ->&____s->seqcount#2 ->&____s->seqcount ->&c->lock ->&net->can.rcvlists_lock ->&n->list_lock ->batched_entropy_u8.lock ->kfence_freelist_lock ->&meta->lock ->j1939_netdev_lock ->&priv->lock ->&priv->j1939_socks_lock ->&jsk->sk_session_queue_lock ->stock_lock ->&f->f_lock FD: 1 BD: 79 +...: slock-AF_CAN FD: 1 BD: 79 ++..: clock-AF_CAN FD: 1 BD: 3 ..-.: elock-AF_CAN FD: 1 BD: 1 ....: &qs->lock FD: 1 BD: 6 +.+.: ebt_mutex.wait_lock FD: 245 BD: 3 +.+.: sk_lock-AF_PPPOX ->slock-AF_PPPOX ->&pn->hash_lock ->clock-AF_PPPOX ->rlock-AF_PPPOX ->chan_lock ->&obj_hash[i].lock ->&rq->__lock ->rcu_state.exp_mutex.wait_lock ->&p->pi_lock ->&rnp->exp_lock ->rcu_state.exp_mutex ->fs_reclaim ->pool_lock#2 ->&ps->sk_lock ->&c->lock ->&tunnel->hlist_lock ->&list->lock#41 ->&dir->lock ->&pn->all_channels_lock ->&mm->mmap_lock ->&pch->chan_sem ->&pch->upl ->&pf->rwait ->&list->lock#26 FD: 1 BD: 4 +...: slock-AF_PPPOX FD: 1 BD: 4 +...: clock-AF_PPPOX FD: 1 BD: 4 ..-.: rlock-AF_PPPOX FD: 33 BD: 3 +.+.: &ep->mtx/1 ->&f->f_lock ->&rq->__lock ->&ep->lock FD: 1 BD: 3 ....: &ep->poll_wait FD: 1 BD: 3 +...: &xs->map_list_lock FD: 242 BD: 82 +.+.: &xs->mutex ->fs_reclaim ->&c->lock ->pool_lock#2 ->free_vmap_area_lock ->&rq->__lock ->vmap_area_lock ->&____s->seqcount ->init_mm.page_table_lock ->rcu_node_0 ->&zone->lock ->&lock->wait_lock ->&obj_hash[i].lock ->purge_vmap_area_lock ->&rcu_state.expedited_wq ->(console_sem).lock ->console_owner_lock ->console_owner ->&cfs_rq->removed.lock ->&n->list_lock ->remove_cache_srcu ->&____s->seqcount#2 ->umem_ida.xa_lock ->&mm->mmap_lock ->&base->lock ->batched_entropy_u8.lock ->kfence_freelist_lock ->&sem->wait_lock ->&p->pi_lock FD: 1 BD: 3 +...: clock-AF_XDP FD: 1 BD: 8 +.+.: nf_sockopt_mutex.wait_lock FD: 131 BD: 2 +.+.: &drbg->drbg_mutex ->crypto_alg_sem ->fs_reclaim ->pool_lock#2 ->&c->lock ->&rq->__lock ->crngs.lock ->&rng->jent_lock ->&n->list_lock ->&cfs_rq->removed.lock ->&obj_hash[i].lock ->rcu_node_0 FD: 7 BD: 5 +.+.: &rng->jent_lock ->&obj_hash[i].lock ->pool_lock#2 ->&____s->seqcount ->quarantine_lock FD: 17 BD: 1 +...: &nr_netdev_xmit_lock_key ->nr_node_list_lock ->&obj_hash[i].lock ->pool_lock#2 ->&meta->lock ->kfence_freelist_lock ->&data->lock ->quarantine_lock FD: 1 BD: 2 +...: nr_node_list_lock FD: 129 BD: 1 +.+.: &audit_cmd_mutex.lock ->fs_reclaim ->pool_lock#2 ->rlock-AF_NETLINK ->&c->lock ->tk_core.seq.seqcount ->&obj_hash[i].lock ->&list->lock ->kauditd_wait.lock ->&rq->__lock ->&n->list_lock FD: 1 BD: 83 ++.-: &policy->lock FD: 1 BD: 82 ....: &list->lock#32 FD: 1 BD: 1 ....: _rs.lock#3 FD: 899 BD: 1 +.+.: sock_diag_mutex ->fs_reclaim ->pool_lock#2 ->&c->lock ->&obj_hash[i].lock ->&rq->__lock ->rlock-AF_NETLINK ->sock_diag_table_mutex ->&n->list_lock ->&____s->seqcount#2 ->&____s->seqcount ->sock_diag_mutex.wait_lock ->&cfs_rq->removed.lock FD: 1 BD: 6 +.+.: calipso_doi_list_lock FD: 249 BD: 1 +.+.: sk_lock-AF_RDS ->slock-AF_RDS ->batched_entropy_u16.lock ->once_lock ->&rq->__lock ->pool_lock#2 ->&obj_hash[i].lock ->&mm->mmap_lock ->rds_trans_sem FD: 1 BD: 2 +...: slock-AF_RDS FD: 27 BD: 3 +.+.: sk_lock-AF_QIPCRTR ->slock-AF_QIPCRTR ->clock-AF_QIPCRTR ->rlock-AF_QIPCRTR ->&rq->__lock FD: 1 BD: 4 +...: slock-AF_QIPCRTR FD: 1 BD: 4 +...: clock-AF_QIPCRTR FD: 1 BD: 4 ....: rlock-AF_QIPCRTR FD: 1 BD: 5 +.+.: sco_sk_list.lock FD: 237 BD: 23 +.+.: sk_lock-AF_BLUETOOTH-BTPROTO_SCO ->&rq->__lock ->slock-AF_BLUETOOTH-BTPROTO_SCO ->&mm->mmap_lock ->&conn->lock#2 ->&obj_hash[i].lock ->&base->lock ->&ei->socket.wq.wait ->rcu_node_0 FD: 1 BD: 24 +...: slock-AF_BLUETOOTH-BTPROTO_SCO FD: 168 BD: 1 +.+.: &net->xfrm.xfrm_cfg_mutex ->fs_reclaim ->&c->lock ->pool_lock#2 ->&obj_hash[i].lock ->rlock-AF_NETLINK ->&rq->__lock ->&n->list_lock ->pfkey_mutex ->&lock->wait_lock ->rlock-AF_KEY ->&____s->seqcount ->&____s->seqcount#2 ->quarantine_lock ->&data->lock ->crypto_alg_sem ->(kmod_concurrent_max).lock ->&x->wait#17 ->running_helpers_waitq.lock ->(crypto_chain).rwsem ->&x->wait#21 ->&base->lock ->&cfs_rq->removed.lock ->(&timer.timer) ->crypto_default_null_skcipher_lock ->&net->xfrm.xfrm_state_lock ->&net->xfrm.xfrm_policy_lock ->&policy->lock ->&list->lock#32 ->(console_sem).lock FD: 2 BD: 4 +...: rose_node_list_lock ->rose_neigh_list_lock FD: 1 BD: 5 +.+.: rfcomm_sk_list.lock FD: 238 BD: 5 +.+.: sk_lock-AF_BLUETOOTH-BTPROTO_RFCOMM ->&rq->__lock ->slock-AF_BLUETOOTH-BTPROTO_RFCOMM ->&ei->socket.wq.wait ->&mm->mmap_lock ->fs_reclaim ->batched_entropy_u8.lock ->kfence_freelist_lock ->rfcomm_dev_lock ->&obj_hash[i].lock ->&meta->lock ->&c->lock ->&n->list_lock ->pool_lock#2 ->rlock-AF_BLUETOOTH ->rcu_node_0 ->&rcu_state.expedited_wq ->&cfs_rq->removed.lock FD: 1 BD: 6 +...: slock-AF_BLUETOOTH-BTPROTO_RFCOMM FD: 25 BD: 2 +.-.: (&s->timer) ->rfcomm_wq.lock FD: 239 BD: 4 +.+.: &d->lock ->sk_lock-AF_BLUETOOTH-BTPROTO_RFCOMM ->slock-AF_BLUETOOTH-BTPROTO_RFCOMM FD: 1 BD: 4 ....: &list->lock#33 FD: 1 BD: 11 ....: tracepoint_srcu FD: 1 BD: 78 ....: (&br->tcn_timer) FD: 1 BD: 78 ....: (&br->hello_timer) FD: 1 BD: 78 ....: (&br->topology_change_timer) FD: 23 BD: 1 +.+.: bpf_module_mutex ->&rq->__lock FD: 1 BD: 1 ....: nopage_rs.lock FD: 3 BD: 84 ....: kernfs_pr_cont_lock ->kernfs_rename_lock ->(console_sem).lock FD: 24 BD: 164 ....: &bdi->wb_waitq ->&p->pi_lock FD: 1 BD: 83 +...: _xmit_LOOPBACK#2 FD: 1 BD: 78 +...: &bond->ipsec_lock FD: 1 BD: 20 ....: ndev_hash_lock FD: 13 BD: 19 +.+.: &pdata->netdev_lock ->pool_lock#2 ->&dir->lock#2 ->ndev_hash_lock ->&c->lock FD: 1 BD: 1 ....: _rs.lock#4 FD: 2 BD: 12 +.+.: devices.xa_lock ->pool_lock#2 FD: 892 BD: 18 +.+.: &rxe->usdev_lock ->&rq->__lock ->&pdata->netdev_lock ->rtnl_mutex ->rtnl_mutex.wait_lock ->&p->pi_lock ->&cfs_rq->removed.lock ->&obj_hash[i].lock ->pool_lock#2 ->(console_sem).lock ->console_owner_lock ->console_owner ->rcu_node_0 ->&rcu_state.expedited_wq ->pgd_lock ->stock_lock ->key ->pcpu_lock ->percpu_counters_lock FD: 128 BD: 3898 +.+.: &table->lock#4 ->fs_reclaim ->pool_lock#2 ->&obj_hash[i].lock ->&table->rwlock ->&device->event_handler_rwsem ->&rq->__lock ->&c->lock FD: 1 BD: 3899 ....: &table->rwlock FD: 1 BD: 3901 ++++: &device->event_handler_rwsem FD: 1 BD: 5 ....: &device->cache_lock FD: 1 BD: 3 +.+.: rdmacg_mutex FD: 25 BD: 14 +.+.: subsys mutex#84 ->&k->k_lock ->&rq->__lock FD: 895 BD: 1 +.+.: (wq_completion)infiniband ->(work_completion)(&work->work)#2 FD: 894 BD: 2 +.+.: (work_completion)(&work->work)#2 ->fs_reclaim ->pool_lock#2 ->&rxe->usdev_lock ->&device->cache_lock ->&obj_hash[i].lock ->&device->event_handler_rwsem ->batched_entropy_u8.lock ->kfence_freelist_lock ->&meta->lock FD: 321 BD: 13 ++++: &device->client_data_rwsem ->&xa->xa_lock#17 ->fs_reclaim ->pool_lock#2 ->&c->lock ->rcu_node_0 ->&rq->__lock ->&cfs_rq->removed.lock ->&obj_hash[i].lock ->&xa->xa_lock#18 ->&xa->xa_lock#19 ->crngs.lock ->free_vmap_area_lock ->vmap_area_lock ->&____s->seqcount ->init_mm.page_table_lock ->purge_vmap_area_lock ->&x->wait#27 ->(console_sem).lock ->&x->wait#28 ->krc.lock ->lock ->&root->kernfs_rwsem ->&____s->seqcount#2 ->ib_mad_port_list_lock ->kernfs_idr_lock ->lock#7 ->umad_ida.xa_lock ->&x->wait#9 ->&n->list_lock ->chrdevs_lock ->&k->list_lock ->gdp_mutex ->bus_type_sem ->sysfs_symlink_target_lock ->&dev->power.lock ->dpm_list_mtx ->req_lock ->&p->pi_lock ->&x->wait#11 ->uevent_sock_mutex ->subsys mutex#85 ->pcpu_alloc_mutex ->uverbs_ida.xa_lock ->subsys mutex#86 ->subsys mutex#87 ->rds_ib_devices_lock ->ib_nodev_conns_lock ->smc_ib_devices.mutex ->&device->event_handler_rwsem ->&pnettable->lock ->&cq->cq_lock ->stock_lock ->mmu_notifier_invalidate_range_start ->&sb->s_type->i_lock_key#8 ->&dir->lock ->&qp->state_lock ->cpu_hotplug_lock ->kthread_create_lock ->&x->wait ->wq_pool_mutex ->&mad_queue->lock ->&qp->rq.producer_lock ->ib_mad_clients.xa_lock ->&port_priv->reg_lock ->ib_agent_port_list_lock ->&cm.device_lock FD: 4 BD: 16 +.+.: &xa->xa_lock#17 ->pool_lock#2 ->&obj_hash[i].lock FD: 10 BD: 14 +.+.: &xa->xa_lock#18 ->pool_lock#2 ->&c->lock ->&n->list_lock ->&obj_hash[i].lock FD: 7 BD: 81 +.+.: &xa->xa_lock#19 ->pool_lock#2 ->&c->lock FD: 1 BD: 14 ....: &x->wait#27 FD: 1 BD: 1 ....: _rs.lock#5 FD: 1 BD: 14 ....: &x->wait#28 FD: 1 BD: 14 ....: ib_mad_port_list_lock FD: 1 BD: 81 +.+.: &id_priv->qp_mutex FD: 2 BD: 81 +.+.: &xa->xa_lock#20 ->pool_lock#2 FD: 2 BD: 81 ....: &cm_id_priv->lock ->&cm.lock FD: 1 BD: 82 ....: &cm.lock FD: 1 BD: 14 ....: umad_ida.xa_lock FD: 3 BD: 14 +.+.: subsys mutex#85 ->&k->k_lock FD: 1 BD: 14 ....: uverbs_ida.xa_lock FD: 3 BD: 14 +.+.: subsys mutex#86 ->&k->k_lock FD: 3 BD: 14 +.+.: subsys mutex#87 ->&k->k_lock FD: 1 BD: 14 +.+.: rds_ib_devices_lock FD: 1 BD: 14 +.+.: ib_nodev_conns_lock FD: 1 BD: 1 ....: _rs.lock#6 FD: 1 BD: 1 ....: _rs.lock#7 FD: 895 BD: 13 +.+.: &device->compat_devs_mutex ->fs_reclaim ->&xa->xa_lock#17 ->pool_lock#2 ->&x->wait#9 ->&obj_hash[i].lock ->&k->list_lock ->gdp_mutex ->&c->lock ->lock ->&root->kernfs_rwsem ->bus_type_sem ->sysfs_symlink_target_lock ->&____s->seqcount#2 ->&____s->seqcount ->&dev->power.lock ->dpm_list_mtx ->uevent_sock_mutex ->subsys mutex#84 ->&n->list_lock ->&rxe->usdev_lock ->&zone->lock ->&rq->__lock ->&sem->wait_lock ->&p->pi_lock ->remove_cache_srcu ->rcu_node_0 ->batched_entropy_u8.lock ->kfence_freelist_lock ->&lock->wait_lock ->&rcu_state.expedited_wq ->&cfs_rq->removed.lock ->uevent_sock_mutex.wait_lock ->pgd_lock ->key ->pcpu_lock ->percpu_counters_lock ->stock_lock ->quarantine_lock ->&meta->lock ->per_cpu_ptr(&cgroup_rstat_cpu_lock, cpu) FD: 894 BD: 2 +.+.: (work_completion)(&smcibdev->port_event_work) ->&rxe->usdev_lock ->&table->rwlock ->smc_lgr_list.lock FD: 1 BD: 3 +...: smc_lgr_list.lock FD: 1 BD: 1 ....: _rs.lock#8 FD: 1 BD: 78 ....: wlock-AF_UNSPEC FD: 1 BD: 78 ....: elock-AF_UNSPEC FD: 1 BD: 1 ....: _rs.lock#9 FD: 1 BD: 4 +.-.: &sap->sk_lock FD: 295 BD: 78 +.+.: bpf_dispatcher_xdp.mutex ->pack_mutex ->fs_reclaim ->pool_lock#2 ->free_vmap_area_lock ->vmap_area_lock ->&____s->seqcount ->init_mm.page_table_lock ->bpf_lock ->text_mutex ->cpu_hotplug_lock ->&rnp->exp_lock ->rcu_state.exp_mutex ->&obj_hash[i].lock ->rcu_state.exp_mutex.wait_lock ->&p->pi_lock FD: 1 BD: 2 +.+.: br_ioctl_mutex.wait_lock FD: 1 BD: 1 ....: net_ratelimit_state.lock FD: 1 BD: 82 +.+.: &s->lock FD: 1 BD: 145 +.+.: freezer_mutex.wait_lock FD: 1 BD: 5 +.+.: &device->unregistration_lock FD: 28 BD: 82 +.-.: &q->lock#2 ->&obj_hash[i].lock ->&base->lock ->pool_lock#2 ->&____s->seqcount ->quarantine_lock FD: 2 BD: 4 +...: &list->lock#34 ->rlock-AF_INET6 FD: 9 BD: 12 +...: &pn->l2tp_tunnel_idr_lock ->pool_lock#2 ->&obj_hash[i].lock ->&c->lock FD: 1 BD: 116 +.+.: dev_pm_qos_sysfs_mtx.wait_lock FD: 1 BD: 1 ....: _rs.lock#10 FD: 1 BD: 13 ....: rdma_nets_rwsem.wait_lock FD: 1 BD: 78 +...: &pmc->lock FD: 25 BD: 1 ..-.: &(&chan->chan_timer)->timer FD: 42 BD: 2 +.+.: (work_completion)(&(&chan->chan_timer)->work) ->&conn->chan_lock ->&lock->wait_lock ->&p->pi_lock ->&rq->__lock FD: 128 BD: 23 +.+.: (work_completion)(&(&conn->disc_work)->work) ->&hdev->unregister_lock ->&rq->__lock FD: 1 BD: 24 +.+.: (work_completion)(&(&conn->id_addr_timer)->work) FD: 1 BD: 23 +.+.: (work_completion)(&(&conn->auto_accept_work)->work) FD: 1 BD: 29 +...: k-clock-AF_BLUETOOTH FD: 1 BD: 23 +.+.: (work_completion)(&(&conn->idle_work)->work) FD: 931 BD: 2 +.+.: (work_completion)(&nlk->work) ->&obj_hash[i].lock ->pool_lock#2 ->rlock-AF_NETLINK ->&dir->lock ->&rq->__lock ->&base->lock ->genl_mutex ->vmap_area_lock ->purge_vmap_area_lock ->&ht->lock ->&____s->seqcount ->quarantine_lock ->&data->lock FD: 1 BD: 2 +.+.: vlan_ioctl_mutex.wait_lock FD: 241 BD: 78 +.+.: sk_lock-AF_UNSPEC ->slock-AF_UNSPEC ->fs_reclaim ->pool_lock#2 ->free_vmap_area_lock ->vmap_area_lock ->&____s->seqcount ->stock_lock ->&c->lock ->pcpu_alloc_mutex ->&mm->mmap_lock ->&obj_hash[i].lock ->pack_mutex ->batched_entropy_u32.lock ->text_mutex ->&fp->aux->used_maps_mutex ->&n->list_lock ->&rq->__lock ->&____s->seqcount#2 FD: 1 BD: 79 +...: slock-AF_UNSPEC FD: 30 BD: 1 ..-.: &(&conn->disc_work)->timer FD: 1 BD: 24 +.+.: hci_cb_list_lock.wait_lock FD: 238 BD: 3 +.+.: sk_lock-AF_AX25 ->slock-AF_AX25 ->clock-AF_AX25 ->ax25_list_lock ->&rq->__lock ->&obj_hash[i].lock ->&list->lock#35 ->rlock-AF_AX25 ->wlock-AF_AX25 ->&mm->mmap_lock ->ax25_uid_lock ->ax25_dev_lock ->fs_reclaim ->&c->lock ->pool_lock#2 FD: 1 BD: 4 +...: slock-AF_AX25 FD: 1 BD: 4 +...: clock-AF_AX25 FD: 1 BD: 4 +...: ax25_list_lock FD: 1 BD: 4 ....: &list->lock#35 FD: 1 BD: 4 ....: rlock-AF_AX25 FD: 1 BD: 4 ....: wlock-AF_AX25 FD: 128 BD: 1 ++++: kn->active#58 ->&rq->__lock ->fs_reclaim ->stock_lock ->&kernfs_locks->open_file_mutex[count] ->&c->lock ->&n->list_lock FD: 1 BD: 1 ....: _rs.lock#11 FD: 146 BD: 165 +.+.: &journal->j_barrier ->&journal->j_state_lock ->&rq->__lock ->&journal->j_list_lock ->&journal->j_checkpoint_mutex ->jbd2_handle ->&journal->j_wait_commit ->&journal->j_wait_done_commit ->&cfs_rq->removed.lock ->&obj_hash[i].lock ->pool_lock#2 ->&lock->wait_lock FD: 137 BD: 90 +.+.: k-sk_lock-AF_TIPC/1 ->k-slock-AF_TIPC ->&obj_hash[i].lock ->&base->lock ->fs_reclaim ->pool_lock#2 ->&list->lock#31 ->&c->lock ->&rq->__lock FD: 1 BD: 120 ....: key#24 FD: 162 BD: 1 +.+.: (wq_completion)tipc_rcv ->(work_completion)(&srv->awork) ->(work_completion)(&con->rwork) FD: 156 BD: 4 +.+.: (work_completion)(&con->rwork) ->&rq->__lock ->k-sk_lock-AF_TIPC ->k-slock-AF_TIPC ->k-clock-AF_TIPC ->&srv->idr_lock ->&obj_hash[i].lock ->pool_lock#2 ->&sb->s_type->i_lock_key#8 ->&xa->xa_lock#9 ->&fsnotify_mark_srcu ->&con->outqueue_lock FD: 43 BD: 94 +...: &con->sub_lock ->&tn->nametbl_lock ->&obj_hash[i].lock ->pool_lock#2 ->(&sub->timer) ->&base->lock ->&meta->lock ->kfence_freelist_lock FD: 160 BD: 4 +.+.: (work_completion)(&srv->awork) ->&srv->idr_lock ->fs_reclaim ->pool_lock#2 ->mmu_notifier_invalidate_range_start ->&sb->s_type->i_lock_key#8 ->k-sk_lock-AF_TIPC ->k-slock-AF_TIPC ->&obj_hash[i].lock ->k-clock-AF_TIPC ->&rq->__lock ->&xa->xa_lock#9 ->&fsnotify_mark_srcu ->&c->lock FD: 5 BD: 97 +.-.: &con->outqueue_lock ->&obj_hash[i].lock ->pool_lock#2 ->quarantine_lock FD: 14 BD: 1 +.-.: (t) ->cmci_poll_lock ->&obj_hash[i].lock ->&base->lock FD: 20 BD: 1 ....: &trie->lock ->stock_lock ->&c->lock ->pool_lock#2 ->&obj_hash[i].lock ->krc.lock ->&____s->seqcount#2 ->&____s->seqcount ->&n->list_lock FD: 162 BD: 1 +.+.: (wq_completion)tipc_rcv#2 ->(work_completion)(&srv->awork) ->(work_completion)(&con->rwork) ->&rq->__lock FD: 30 BD: 1 +.-.: (&pool->idle_timer) ->&pool->lock/1 ->&pool->lock FD: 1 BD: 2 +.-.: cmci_poll_lock FD: 144 BD: 8 +.+.: &sb->s_type->i_mutex_key#8/4 ->&sem->wait_lock ->&rq->__lock ->&ei->i_data_sem ->&ei->i_data_sem/1 ->mmu_notifier_invalidate_range_start ->pool_lock#2 ->&journal->j_state_lock ->jbd2_handle ->&obj_hash[i].lock ->&cfs_rq->removed.lock ->&journal->j_wait_commit ->&journal->j_wait_done_commit ->remove_cache_srcu ->key#3 ->key#14 ->&c->lock ->&____s->seqcount#2 ->&____s->seqcount ->&n->list_lock ->&p->pi_lock ->pgd_lock ->stock_lock ->key ->pcpu_lock ->percpu_counters_lock FD: 29 BD: 1 +.-.: (&q->timer) ->&obj_hash[i].lock ->pool_lock#2 FD: 102 BD: 171 +.+.: &ei->i_data_sem/1 ->mmu_notifier_invalidate_range_start ->pool_lock#2 ->&ei->i_es_lock ->&rq->__lock ->&c->lock ->&ei->i_raw_lock ->&obj_hash[i].lock ->&ei->i_prealloc_lock ->&ret->b_state_lock ->key#3 ->key#14 ->&mapping->private_lock ->&____s->seqcount#2 ->&n->list_lock ->rcu_node_0 ->batched_entropy_u8.lock ->kfence_freelist_lock ->&meta->lock ->&cfs_rq->removed.lock ->remove_cache_srcu ->&____s->seqcount ->&sb->s_type->i_lock_key#22 ->&journal->j_state_lock ->&journal->j_revoke_lock FD: 25 BD: 1 ..-.: security/integrity/ima/ima_queue_keys.c:35 FD: 5 BD: 2 +.+.: (ima_keys_delayed_work).work ->ima_keys_lock ->&obj_hash[i].lock ->pool_lock#2 FD: 128 BD: 1 .+.+: kn->active#59 ->&rq->__lock ->fs_reclaim ->stock_lock ->&kernfs_locks->open_file_mutex[count] ->&c->lock ->&____s->seqcount#2 ->&____s->seqcount ->&n->list_lock ->remove_cache_srcu ->rcu_node_0 FD: 1 BD: 10 +.+.: ima_extend_list_mutex.wait_lock FD: 1 BD: 91 ....: key#25 FD: 25 BD: 1 ..-.: net/ipv4/devinet.c:474 FD: 898 BD: 3 +.+.: nlk_cb_mutex-SOCK_DIAG ->fs_reclaim ->pool_lock#2 ->&c->lock ->&n->list_lock ->&net->packet.sklist_lock ->&ht->lock ->nl_table_lock ->&rq->__lock ->inet_diag_table_mutex ->&obj_hash[i].lock ->rlock-AF_NETLINK ->k-sk_lock-AF_TIPC ->k-slock-AF_TIPC ->sk_lock-AF_TIPC ->slock-AF_TIPC ->&lock->wait_lock FD: 1 BD: 4 +.+.: oom_adj_mutex.wait_lock FD: 1 BD: 1 ....: _rs.lock#12 FD: 103 BD: 1 +.-.: &dreq->dreq_lock ->pool_lock#2 ->&dir->lock ->slock-AF_INET6 ->&c->lock ->&____s->seqcount#2 ->&____s->seqcount ->&n->list_lock ->batched_entropy_u8.lock ->kfence_freelist_lock FD: 1 BD: 83 ....: wlock-AF_INET FD: 1 BD: 4 +.+.: delayed_uprobe_lock.wait_lock FD: 1 BD: 4 +.+.: &pipe->mutex#2/2 FD: 128 BD: 1 .+.+: kn->active#60 ->fs_reclaim ->stock_lock ->&kernfs_locks->open_file_mutex[count] ->&c->lock ->&rq->__lock ->&n->list_lock ->&____s->seqcount#2 ->&____s->seqcount ->remove_cache_srcu ->pool_lock#2 FD: 2 BD: 143 +.+.: (work_completion)(flush) ->&list->lock#5 FD: 1 BD: 1 ....: _rs.lock#13 FD: 1 BD: 1 ....: elock-AF_RDS FD: 1 BD: 3 ..-.: wlock-AF_PPPOX FD: 1 BD: 1 ....: _rs.lock#14 FD: 33 BD: 2 +.+.: (work_completion)(&pool->idle_cull_work) ->wq_pool_attach_mutex ->wq_pool_attach_mutex.wait_lock ->&p->pi_lock ->&rq->__lock ->&cfs_rq->removed.lock ->&obj_hash[i].lock ->pool_lock#2 FD: 1 BD: 1 +...: clock-AF_LLC FD: 139 BD: 3 +.+.: sk_lock-AF_LLC ->slock-AF_LLC ->&rq->__lock ->llc_sap_list_lock ->fs_reclaim ->&c->lock ->pool_lock#2 ->&dir->lock#2 ->&sap->sk_lock ->wlock-AF_LLC ->&obj_hash[i].lock ->&base->lock ->&ei->socket.wq.wait ->rcu_node_0 ->quarantine_lock ->&cfs_rq->removed.lock ->&meta->lock ->kfence_freelist_lock ->&rcu_state.expedited_wq ->&n->list_lock FD: 35 BD: 5 +.-.: slock-AF_LLC ->&sk->sk_lock.wq ->pool_lock#2 ->wlock-AF_LLC ->&obj_hash[i].lock ->&base->lock FD: 1 BD: 3 ....: (&llc->pf_cycle_timer.timer) FD: 36 BD: 3 +.-.: (&llc->ack_timer.timer) ->pool_lock#2 ->slock-AF_LLC FD: 1 BD: 3 ....: (&llc->rej_sent_timer.timer) FD: 1 BD: 3 ....: (&llc->busy_state_timer.timer) FD: 1 BD: 3 ....: rlock-AF_LLC FD: 1 BD: 6 ..-.: wlock-AF_LLC FD: 1 BD: 3 ....: &list->lock#36 FD: 1 BD: 2 +...: slock-AF_IEEE802154 FD: 1 BD: 85 +.+.: nf_tables_destroy_list_lock FD: 2 BD: 1 +.+.: sk_lock-AF_IEEE802154 ->slock-AF_IEEE802154 FD: 1 BD: 84 ....: umem_ida.xa_lock FD: 134 BD: 2 +.+.: trans_destroy_work ->nf_tables_destroy_list_lock ->&obj_hash[i].lock ->stock_lock ->pool_lock#2 ->(work_completion)(&ht->run_work) ->&ht->mutex FD: 1 BD: 1 ....: &head->lock FD: 1 BD: 15 +.+.: fw_lock.wait_lock FD: 160 BD: 1 +.+.: (wq_completion)bond1#3 ->(work_completion)(&(&slave->notify_work)->work) FD: 46 BD: 3836 +...: &dev_addr_list_lock_key#3/2 ->&macvlan_netdev_addr_lock_key/1 ->&obj_hash[i].lock ->krc.lock ->pool_lock#2 FD: 1 BD: 83 +...: &qdisc_xmit_lock_key#2 FD: 1 BD: 78 +.+.: (work_completion)(&port->wq) FD: 55 BD: 1 +.+.: (wq_completion)bond3 ->(work_completion)(&(&bond->alb_work)->work) ->(work_completion)(&(&bond->mii_work)->work) ->&rq->__lock FD: 50 BD: 83 +.+.: (work_completion)(&(&bond->alb_work)->work) ->&obj_hash[i].lock ->&base->lock ->&rq->__lock FD: 27 BD: 87 +.+.: (work_completion)(&(&bond->mii_work)->work) ->&obj_hash[i].lock ->&base->lock ->&rq->__lock ->rcu_node_0 ->&cfs_rq->removed.lock FD: 1 BD: 83 +...: &qdisc_xmit_lock_key#3 FD: 30 BD: 1 ..-.: &(&bond->alb_work)->timer FD: 30 BD: 1 ..-.: &(&bond->mii_work)->timer FD: 55 BD: 1 +.+.: (wq_completion)bond4 ->(work_completion)(&(&bond->alb_work)->work) ->(work_completion)(&(&bond->mii_work)->work) ->&rq->__lock FD: 55 BD: 1 +.+.: (wq_completion)bond5 ->(work_completion)(&(&bond->alb_work)->work) ->(work_completion)(&(&bond->mii_work)->work) ->&rq->__lock FD: 130 BD: 81 +.+.: tcp_md5sig_mutex ->crypto_alg_sem ->fs_reclaim ->pool_lock#2 ->&c->lock FD: 1 BD: 107 ..-.: wlock-AF_INET6 FD: 103 BD: 1 +.-.: (&hc->tx_rtotimer) ->slock-AF_INET6 ->&obj_hash[i].lock ->pool_lock#2 ->&dir->lock ->stock_lock ->per_cpu_ptr(&cgroup_rstat_cpu_lock, cpu) ->&zone->lock ->quarantine_lock ->&meta->lock ->kfence_freelist_lock ->&c->lock ->&n->list_lock FD: 25 BD: 1 ..-.: net/ipv4/tcp_ipv4.c:1063 FD: 287 BD: 2 +.+.: ((tcp_md5_needed).work).work ->cpu_hotplug_lock FD: 26 BD: 86 +.-.: (&ifibss->timer) ->&rdev->wiphy_work_lock FD: 1 BD: 149 +.+.: wq_pool_attach_mutex.wait_lock FD: 1 BD: 3 ....: &list->lock#37 FD: 891 BD: 2 +.+.: (work_completion)(&rdev->destroy_work) ->rtnl_mutex ->rtnl_mutex.wait_lock ->&p->pi_lock ->&rq->__lock ->rcu_state.barrier_mutex ->rcu_state.barrier_mutex.wait_lock ->dev_base_lock ->lweventlist_lock ->pcpu_lock ->stock_lock ->&obj_hash[i].lock ->pool_lock#2 ->krc.lock ->&dir->lock#2 ->netdev_unregistering_wq.lock FD: 890 BD: 2 +.+.: (work_completion)(&(&rdev->dfs_update_channels_wk)->work) ->rtnl_mutex ->rtnl_mutex.wait_lock ->&p->pi_lock ->&rq->__lock FD: 1 BD: 79 +...: &ipvs->sync_buff_lock FD: 144 BD: 1 +.+.: xt_led_mutex ->fs_reclaim ->pool_lock#2 ->triggers_list_lock ->leds_list_lock ->&obj_hash[i].lock ->(&ledinternal->timer) ->&base->lock ->&c->lock ->&rq->__lock FD: 1 BD: 1 ....: _rs.lock#16 FD: 1 BD: 65 ....: shrinker_rwsem.wait_lock FD: 1 BD: 1 ....: _rs.lock#15 FD: 1 BD: 2 ....: (&ledinternal->timer) FD: 23 BD: 78 +.+.: (work_completion)(&(&bond->arp_work)->work) ->&rq->__lock FD: 166 BD: 82 +.+.: (work_completion)(&(&bond->ad_work)->work) ->&bond->mode_lock ->&rq->__lock ->&obj_hash[i].lock ->rtnl_mutex.wait_lock ->&p->pi_lock ->&base->lock ->rcu_node_0 ->&rcu_state.expedited_wq ->&cfs_rq->removed.lock FD: 230 BD: 3821 +.+.: (work_completion)(&(&bond->mcast_work)->work) ->&obj_hash[i].lock ->&base->lock ->pool_lock#2 ->&rq->__lock ->rtnl_mutex.wait_lock ->&p->pi_lock ->rcu_node_0 ->&cfs_rq->removed.lock ->&rcu_state.expedited_wq FD: 145 BD: 79 +.+.: (work_completion)(&(&bond->slave_arr_work)->work) ->&rq->__lock ->&obj_hash[i].lock ->rtnl_mutex.wait_lock ->&p->pi_lock ->&base->lock FD: 1 BD: 81 +.+.: tcpv6_prot_mutex FD: 1 BD: 81 +...: device_spinlock FD: 512 BD: 1 +.+.: &ctx->tx_lock ->sk_lock-AF_INET6 ->&lock->wait_lock ->&rq->__lock ->slock-AF_INET6 ->&cfs_rq->removed.lock ->&obj_hash[i].lock ->pool_lock#2 FD: 1 BD: 3 +.+.: (work_completion)(&(&sw_ctx_tx->tx_work.work)->work) FD: 1 BD: 78 +.+.: (work_completion)(&wdev->pmsr_free_wk) FD: 1 BD: 81 +...: &sw_ctx_tx->encrypt_compl_lock FD: 1 BD: 11 +.+.: tracepoints_mutex.wait_lock FD: 1 BD: 78 +.+.: (work_completion)(&wdev->disconnect_wk) FD: 1 BD: 78 +.+.: (work_completion)(&sdata->activate_links_work) FD: 1 BD: 81 ....: (&local->dynamic_ps_timer) FD: 1 BD: 81 +.+.: (work_completion)(&local->dynamic_ps_enable_work) FD: 1 BD: 81 +.+.: (work_completion)(&sdata->recalc_smps) FD: 187 BD: 84 +.+.: (work_completion)(&link->csa_finalize_work) ->&wdev->mtx FD: 1 BD: 81 +.+.: (work_completion)(&link->color_change_finalize_work) FD: 1 BD: 81 +.+.: (work_completion)(&(&link->dfs_cac_timer_work)->work) FD: 1 BD: 78 ....: &rdev->dev_wait FD: 1 BD: 3 +.+.: (work_completion)(&strp->work) FD: 1 BD: 4 +.+.: chan_lock FD: 12 BD: 78 +.+.: mrt_lock ->&c->lock ->pool_lock#2 ->&dir->lock#2 FD: 1 BD: 78 +...: _xmit_PIMREG FD: 1 BD: 83 +...: _xmit_PIMREG#2 FD: 128 BD: 153 +.+.: &sb->s_type->i_mutex_key#21 ->&rq->__lock ->tk_core.seq.seqcount ->fs_reclaim ->&c->lock ->pool_lock#2 ->&resv_map->lock ->hugetlb_lock FD: 4 BD: 156 +.+.: &resv_map->lock ->&obj_hash[i].lock ->pool_lock#2 FD: 29 BD: 155 +.+.: &hugetlbfs_i_mmap_rwsem_key ->&obj_hash[i].lock ->&rq->__lock ->pool_lock#2 ->ptlock_ptr(page) FD: 130 BD: 154 ++++: &vma_lock->rw_sema ->ptlock_ptr(page) ->&rq->__lock ->&resv_map->lock ->fs_reclaim ->pool_lock#2 ->hugetlb_lock ->&xa->xa_lock#9 ->&sb->s_type->i_lock_key#16 ->&hugetlbfs_i_mmap_rwsem_key ->&c->lock FD: 131 BD: 153 +.+.: &hugetlb_fault_mutex_table[i] ->&vma_lock->rw_sema ->&sb->s_type->i_lock_key#16 FD: 1 BD: 1 ....: _rs.lock#17 FD: 1 BD: 83 +...: _xmit_IPGRE#2 FD: 1 BD: 6 +.+.: rfcomm_dev_lock FD: 26 BD: 1 +.+.: sk_lock-AF_XDP ->slock-AF_XDP ->&rq->__lock FD: 25 BD: 2 +...: slock-AF_XDP ->&sk->sk_lock.wq FD: 1 BD: 1 ....: _rs.lock#18 FD: 1 BD: 83 +...: _xmit_TUNNEL#2 FD: 1 BD: 83 +...: _xmit_SIT#2 FD: 1 BD: 13 +.+.: nf_conntrack_mutex.wait_lock FD: 1 BD: 83 +...: &qdisc_xmit_lock_key#4 FD: 1 BD: 78 +...: _xmit_NETROM#2 FD: 1 BD: 83 +...: &qdisc_xmit_lock_key#5 FD: 18 BD: 3836 +...: &vlan_netdev_addr_lock_key/2 ->&obj_hash[i].lock ->krc.lock FD: 126 BD: 3 +.+.: callchain_mutex ->fs_reclaim ->pool_lock#2 ->&c->lock ->&obj_hash[i].lock ->&n->list_lock ->&rq->__lock ->&____s->seqcount FD: 33 BD: 96 +.-.: &sub->lock ->&srv->idr_lock ->pool_lock#2 ->&con->outqueue_lock ->&c->lock ->&n->list_lock FD: 1 BD: 6 +.+.: smcd_dev_list.mutex FD: 162 BD: 1 +.+.: (wq_completion)tipc_send ->(work_completion)(&con->swork) FD: 13 BD: 1 +.-.: (&tsc_sync_check_timer) ->&obj_hash[i].lock ->&base->lock FD: 161 BD: 6 +.+.: (work_completion)(&con->swork) ->&con->outqueue_lock ->pool_lock#2 ->&list->lock#38 ->&c->lock ->&list->lock#31 ->&n->list_lock ->batched_entropy_u8.lock ->kfence_freelist_lock ->&____s->seqcount#2 ->&____s->seqcount ->&con->sub_lock ->&rq->__lock ->k-sk_lock-AF_TIPC ->k-slock-AF_TIPC FD: 1 BD: 86 +...: &list->lock#38 FD: 1 BD: 79 +...: &tipc_net(net)->bclock FD: 4 BD: 1 +.+.: sk_lock-AF_ISDN ->slock-AF_ISDN ->clock-AF_ISDN ->rlock-AF_ISDN FD: 162 BD: 1 +.+.: (wq_completion)tipc_send#2 ->(work_completion)(&con->swork) FD: 1 BD: 1 +...: data_sockets.lock FD: 1 BD: 9 +...: &vvs->rx_lock FD: 1 BD: 2 +...: slock-AF_ISDN FD: 1 BD: 2 ....: rlock-AF_ISDN FD: 34 BD: 2 +.+.: (work_completion)(&umem->work) ->umem_ida.xa_lock ->vmap_area_lock ->&obj_hash[i].lock ->purge_vmap_area_lock ->pool_lock#2 ->&rq->__lock ->&lruvec->lru_lock ->rcu_node_0 ->&cfs_rq->removed.lock FD: 1 BD: 132 +...: rds_tcp_tc_list_lock FD: 34 BD: 7 +.+.: &tc->t_conn_path_lock ->clock-AF_INET6 ->&cp->cp_lock ->pool_lock#2 FD: 2 BD: 7 ....: rds_conn_lock ->rds_cong_lock FD: 2 BD: 9 +.+.: &id_priv->handler_mutex ->&id_priv->lock FD: 1 BD: 9 ....: &x->wait#29 FD: 1 BD: 109 ..-.: &cp->cp_lock FD: 205 BD: 6 +.+.: (work_completion)(&(&cp->cp_send_w)->work) ->k-sk_lock-AF_INET6 ->k-slock-AF_INET6 ->pool_lock#2 ->&obj_hash[i].lock ->&cp->cp_lock FD: 205 BD: 6 +.+.: (work_completion)(&(&cp->cp_recv_w)->work) ->k-sk_lock-AF_INET6 ->k-slock-AF_INET6 FD: 202 BD: 6 +.+.: (work_completion)(&cp->cp_down_w) ->&cp->cp_cm_lock ->k-sk_lock-AF_INET6 ->k-slock-AF_INET6 ->pool_lock#2 ->&dir->lock ->&obj_hash[i].lock ->&sb->s_type->i_lock_key#8 ->&xa->xa_lock#9 ->&fsnotify_mark_srcu ->&rm->m_flush_wait ->&cp->cp_lock ->(work_completion)(&(&cp->cp_conn_w)->work) FD: 1 BD: 7 +.+.: &cp->cp_cm_lock FD: 1 BD: 7 ....: &rm->m_flush_wait FD: 1 BD: 7 +.+.: (work_completion)(&(&cp->cp_conn_w)->work) FD: 1 BD: 109 ..-.: &rm->m_rs_lock FD: 1 BD: 107 ..-.: &list->lock#39 FD: 102 BD: 1 +.-.: (&asoc->timers[i]) ->slock-AF_INET6 FD: 1 BD: 93 +...: &list->lock#40 FD: 24 BD: 4 +.+.: &ps->sk_lock ->&rq->__lock ->&tunnel->hlist_lock FD: 1 BD: 11 +...: &tunnel->hlist_lock FD: 46 BD: 6 +.+.: (work_completion)(&tunnel->del_work) ->&tunnel->hlist_lock ->&obj_hash[i].lock ->&list->lock#41 ->&pn->l2tp_tunnel_idr_lock ->&rnp->exp_lock ->rcu_state.exp_mutex ->rcu_state.exp_mutex.wait_lock ->&p->pi_lock FD: 1 BD: 10 ....: &list->lock#41 FD: 162 BD: 1 +.+.: (wq_completion)tipc_rcv#3 ->(work_completion)(&srv->awork) ->(work_completion)(&con->rwork) FD: 24 BD: 4180 ....: &ep->poll_wait/1 ->&p->pi_lock FD: 1 BD: 24 +.+.: (work_completion)(&(&conn->timeout_work)->work) FD: 1 BD: 25 +.+.: &conn->lock#2 FD: 1 BD: 27 ....: namespace_sem.wait_lock FD: 162 BD: 1 +.+.: (wq_completion)tipc_send#3 ->&rq->__lock ->(work_completion)(&con->swork) FD: 1 BD: 3 +...: &htab->buckets[i].lock FD: 1 BD: 78 ....: &____s->seqcount#13 FD: 24 BD: 92 ....: &sk->sk_lock.wq#2 ->&p->pi_lock FD: 232 BD: 1 +.+.: &tfile->napi_mutex ->&mm->mmap_lock ->pcpu_lock ->&obj_hash[i].lock ->pool_lock#2 ->&rq->__lock ->&lock->wait_lock ->&____s->seqcount ->&c->lock ->&n->list_lock FD: 34 BD: 95 +.-.: (&sub->timer) ->&sub->lock FD: 1 BD: 78 ....: &____s->seqcount#14 FD: 289 BD: 1 .+.+: kn->active#61 ->fs_reclaim ->stock_lock ->&kernfs_locks->open_file_mutex[count] ->cpu_hotplug_lock ->&c->lock ->&rq->__lock ->&n->list_lock ->&____s->seqcount#2 ->&____s->seqcount FD: 133 BD: 1 ++++: kn->active#62 ->&rq->__lock ->fs_reclaim ->&c->lock ->stock_lock ->&kernfs_locks->open_file_mutex[count] ->&cgrp->pidlist_mutex ->&n->list_lock FD: 6 BD: 1 +...: fnhe_lock ->&____s->seqcount ->pool_lock#2 ->batched_entropy_u8.lock FD: 296 BD: 3 +.+.: sched_register_mutex ->tracepoints_mutex FD: 25 BD: 1 ..-.: &(&l->destroy_dwork)->timer FD: 131 BD: 4 +.+.: (work_completion)(&(&l->destroy_dwork)->work) ->&cgrp->pidlist_mutex ->&obj_hash[i].lock FD: 126 BD: 78 +.+.: &tn->idrinfo->lock ->fs_reclaim ->pool_lock#2 FD: 13 BD: 78 +...: &p->tcfa_lock ->&c->lock ->&n->list_lock ->pool_lock#2 ->&(to_police(*a)->tcfp_lock) FD: 73 BD: 2 +.+.: (work_completion)(&work->work)#3 ->&rq->__lock ->pool_lock#2 ->&c->lock ->&dir->lock#2 ->&ul->lock#2 ->&obj_hash[i].lock ->&data->lock FD: 1 BD: 2 +.+.: sock_diag_mutex.wait_lock FD: 160 BD: 1 +.+.: (wq_completion)bond8 ->(work_completion)(&(&slave->notify_work)->work) FD: 160 BD: 1 +.+.: (wq_completion)bond9 ->(work_completion)(&(&slave->notify_work)->work) FD: 55 BD: 1 +.+.: (wq_completion)bond6 ->(work_completion)(&(&bond->alb_work)->work) ->(work_completion)(&(&bond->mii_work)->work) ->&rq->__lock FD: 55 BD: 1 +.+.: (wq_completion)bond7 ->(work_completion)(&(&bond->alb_work)->work) ->(work_completion)(&(&bond->mii_work)->work) ->&rq->__lock FD: 128 BD: 81 +.+.: &sb->s_type->i_mutex_key#3/1 ->rename_lock.seqcount ->fs_reclaim ->stock_lock ->&c->lock ->pool_lock#2 ->&dentry->d_lock ->&rq->__lock ->tk_core.seq.seqcount ->rename_lock FD: 128 BD: 1 .+.+: kn->active#63 ->fs_reclaim ->stock_lock ->&kernfs_locks->open_file_mutex[count] ->&c->lock FD: 155 BD: 3 +.+.: (work_completion)(&(&local->roc_work)->work) ->&local->mtx FD: 1 BD: 1 ....: _rs.lock#19 FD: 30 BD: 1 ..-.: &(&local->roc_work)->timer FD: 1 BD: 85 ....: key#26 FD: 1 BD: 9 ....: unix_gc_wait.lock FD: 128 BD: 1 .+.+: kn->active#64 ->&rq->__lock ->fs_reclaim ->stock_lock ->&kernfs_locks->open_file_mutex[count] FD: 3 BD: 8 +.+.: unix_gc_lock ->unix_gc_wait.lock ->rlock-AF_UNIX FD: 3 BD: 78 +...: mfc_unres_lock ->&obj_hash[i].lock FD: 1 BD: 1 ....: _rs.lock#20 FD: 130 BD: 4 +.+.: crypto_default_null_skcipher_lock ->crypto_alg_sem ->fs_reclaim ->pool_lock#2 ->&obj_hash[i].lock FD: 1 BD: 1 +.+.: nfnl_subsys_none FD: 160 BD: 1 +.+.: (wq_completion)bond3#2 ->(work_completion)(&(&slave->notify_work)->work) FD: 30 BD: 1 ..-.: &(&bond->mcast_work)->timer FD: 1 BD: 78 +.+.: team->team_lock_key#7 FD: 1 BD: 78 +.+.: team->team_lock_key#8 FD: 1 BD: 78 +.+.: team->team_lock_key#9 FD: 128 BD: 1 .+.+: kn->active#65 ->fs_reclaim ->&c->lock ->&n->list_lock ->&kernfs_locks->open_file_mutex[count] FD: 126 BD: 78 +.+.: &tn->idrinfo->lock#2 ->fs_reclaim ->pool_lock#2 FD: 1 BD: 3 +.+.: nfnl_grp_active_lock FD: 2 BD: 1 ....: &loc_l->lock ->&l->lock FD: 1 BD: 2 ....: &l->lock FD: 3 BD: 78 +...: mfc_unres_lock#2 ->&obj_hash[i].lock FD: 162 BD: 1 +.+.: (wq_completion)tipc_send#4 ->(work_completion)(&con->swork) FD: 24 BD: 1 +.+.: put_task_map-wait-type-override#2 ->&obj_hash[i].lock ->pool_lock#2 ->stock_lock ->per_cpu_ptr(&cgroup_rstat_cpu_lock, cpu) ->&rq->__lock FD: 30 BD: 1 ..-.: &(&bond->ad_work)->timer FD: 1 BD: 84 +.+.: (work_completion)(&(&priv->gc_work)->work)#2 FD: 1 BD: 3 +.+.: event_mutex.wait_lock FD: 222 BD: 2 +.+.: (work_completion)(&(&idev->mc_query_work)->work) ->&idev->mc_query_lock ->&idev->mc_lock FD: 12 BD: 78 +.+.: mrt_lock#2 ->pool_lock#2 ->&dir->lock#2 ->&c->lock FD: 248 BD: 1 +.+.: (wq_completion)vsock-loopback ->(work_completion)(&vsock->pkt_work) FD: 1 BD: 9 +...: &list->lock#42 FD: 169 BD: 1 +.+.: (wq_completion)bond8#2 ->(work_completion)(&(&bond->mii_work)->work) ->(work_completion)(&(&bond->ad_work)->work) ->(work_completion)(&(&slave->notify_work)->work) ->&rq->__lock FD: 131 BD: 1 +.+.: mem_id_lock ->&rq->__lock ->fs_reclaim ->&c->lock ->&____s->seqcount#2 ->&____s->seqcount ->pool_lock#2 ->batched_entropy_u32.lock ->&obj_hash[i].lock ->mem_id_pool.xa_lock ->&ht->lock FD: 162 BD: 1 +.+.: (wq_completion)tipc_send#5 ->(work_completion)(&con->swork) FD: 1 BD: 1 ....: _rs.lock#21 FD: 1 BD: 3 +.+.: dev_map_lock FD: 1 BD: 2 ..-.: mem_id_pool.xa_lock FD: 1 BD: 1 +...: &r->producer_lock#3 FD: 1 BD: 81 +.+.: &bond->stats_lock/2 FD: 46 BD: 86 +...: &dev_addr_list_lock_key/2 ->&macvlan_netdev_addr_lock_key/1 ->&obj_hash[i].lock ->krc.lock ->&c->lock FD: 1 BD: 1 +...: &r->consumer_lock#3 FD: 1 BD: 167 ....: key#27 FD: 247 BD: 2 +.+.: (work_completion)(&vsock->pkt_work) ->&list->lock#42 ->vsock_table_lock ->sk_lock-AF_VSOCK ->slock-AF_VSOCK ->&obj_hash[i].lock ->pool_lock#2 ->&dir->lock ->stock_lock FD: 135 BD: 8 +.+.: sk_lock-AF_VSOCK/1 ->slock-AF_VSOCK ->fs_reclaim ->pool_lock#2 ->&vvs->tx_lock ->vsock_table_lock ->&vvs->rx_lock ->&list->lock#42 ->&c->lock ->&obj_hash[i].lock ->&base->lock ->clock-AF_VSOCK ->rlock-AF_VSOCK FD: 1 BD: 9 +...: &vvs->tx_lock FD: 128 BD: 1 .+.+: kn->active#66 ->fs_reclaim ->stock_lock ->&kernfs_locks->open_file_mutex[count] ->&c->lock ->&n->list_lock ->&rq->__lock ->&____s->seqcount#2 ->&____s->seqcount FD: 1 BD: 78 +.+.: ifalias_mutex FD: 18 BD: 3836 +...: &macsec_netdev_addr_lock_key#2/2 ->&obj_hash[i].lock ->krc.lock FD: 160 BD: 1 +.+.: (wq_completion)bond5#2 ->(work_completion)(&(&slave->notify_work)->work) FD: 160 BD: 1 +.+.: (wq_completion)bond7#2 ->(work_completion)(&(&slave->notify_work)->work) FD: 45 BD: 1 .+.+: sb_writers#14 ->mount_lock FD: 1 BD: 14 ....: &cq->cq_lock FD: 1 BD: 14 ....: &qp->state_lock FD: 1 BD: 14 ....: &mad_queue->lock FD: 1 BD: 14 ....: &qp->rq.producer_lock FD: 2 BD: 14 +.+.: ib_mad_clients.xa_lock ->pool_lock#2 FD: 2 BD: 14 ....: &port_priv->reg_lock ->pool_lock#2 FD: 1 BD: 14 ....: ib_agent_port_list_lock FD: 1 BD: 14 ....: &cm.device_lock FD: 409 BD: 1 +.+.: (wq_completion)nbd-del ->(work_completion)(&nbd->remove_work) FD: 408 BD: 2 +.+.: (work_completion)(&nbd->remove_work) ->&disk->open_mutex ->&bdev->bd_holder_lock ->&bdev->bd_size_lock ->&q->mq_freeze_lock ->set->srcu ->&q->mq_freeze_wq ->&root->kernfs_rwsem ->(&bdi->laptop_mode_wb_timer) ->&obj_hash[i].lock ->&base->lock ->bdi_lock ->&rq->__lock ->rcu_state.exp_mutex.wait_lock ->&p->pi_lock ->&wb->work_lock ->cgwb_lock ->&pool->lock/1 ->&(&wb->dwork)->timer ->(work_completion)(&(&wb->dwork)->work) ->&(&wb->bw_dwork)->timer ->(work_completion)(&(&wb->bw_dwork)->work) ->&bdi->cgwb_release_mutex ->pin_fs_lock ->&dentry->d_lock ->&sb->s_type->i_mutex_key#3 ->&fsnotify_mark_srcu ->&sb->s_type->i_lock_key#7 ->&s->s_inode_list_lock ->&xa->xa_lock#9 ->pool_lock#2 ->mount_lock ->dev_pm_qos_sysfs_mtx ->kernfs_idr_lock ->&k->list_lock ->sysfs_symlink_target_lock ->subsys mutex#38 ->&x->wait#9 ->dpm_list_mtx ->&dev->power.lock ->deferred_probe_mutex ->device_links_lock ->mmu_notifier_invalidate_range_start ->uevent_sock_mutex ->uevent_sock_mutex.wait_lock ->gdp_mutex ->&q->sysfs_lock ->&q->sysfs_dir_lock ->&q->debugfs_mutex ->dev_hotplug_mutex ->req_lock ->&x->wait#11 ->subsys mutex#37 ->percpu_ref_switch_lock ->&q->queue_lock ->(&q->timeout) ->(work_completion)(&q->timeout_work) ->(wq_completion)kintegrityd ->&wq->mutex ->(work_completion)(&(&q->requeue_work)->work) ->(work_completion)(&(&hctx->run_work)->work) ->&ACCESS_PRIVATE(sdp, lock) ->&x->wait#2 ->&q->rq_qos_mutex ->&tags->lock ->cpu_hotplug_lock ->&xa->xa_lock#10 ->&q->unused_hctx_lock ->&set->tag_list_lock ->(&sq->pending_timer) ->(work_completion)(&td->dispatch_work) ->&q->blkcg_mutex ->pcpu_lock ->&c->lock ->bio_slab_lock ->&xa->xa_lock#11 ->&sb->s_type->i_lock_key#3 ->&zone->lock ->&(&ssp->srcu_sup->work)->timer ->(work_completion)(&(&ssp->srcu_sup->work)->work) ->(&sdp->delay_work) ->(work_completion)(&sdp->work) ->nbd_index_mutex ->wq_mayday_lock ->&x->wait ->wq_pool_mutex ->&x->wait#10 ->&cfs_rq->removed.lock ->&meta->lock ->kfence_freelist_lock FD: 1 BD: 3 ....: (&bdi->laptop_mode_wb_timer) FD: 2 BD: 3 +.+.: &bdi->cgwb_release_mutex ->cgwb_lock FD: 1 BD: 3 +.+.: (wq_completion)kintegrityd FD: 141 BD: 11 +.+.: &eq->sysfs_lock ->&q->debugfs_mutex ->&dd->lock ->&obj_hash[i].lock ->pool_lock#2 ->pcpu_lock FD: 1 BD: 12 +.+.: nbd_index_mutex.wait_lock FD: 2 BD: 156 +.+.: lock#10 FD: 1 BD: 2 +.-.: &inst->lock FD: 128 BD: 11 +.+.: reg_lock ->&rq->__lock ->fs_reclaim ->pool_lock#2 ->&obj_hash[i].lock ->&c->lock ->&n->list_lock FD: 128 BD: 1 .+.+: kn->active#67 ->fs_reclaim ->stock_lock ->&kernfs_locks->open_file_mutex[count] ->&c->lock ->&rq->__lock ->&n->list_lock FD: 1 BD: 84 +.+.: (work_completion)(&(&flowtable->gc_work)->work) FD: 129 BD: 1 .+.+: kn->active#68 ->fs_reclaim ->stock_lock ->&kernfs_locks->open_file_mutex[count] ->devcgroup_mutex ->&c->lock FD: 30 BD: 1 +.+.: (wq_completion)phy19 ->(work_completion)(&local->reconfig_filter) FD: 30 BD: 1 +.+.: (wq_completion)phy20 ->(work_completion)(&local->reconfig_filter) FD: 30 BD: 1 +.+.: (wq_completion)phy21 ->(work_completion)(&local->reconfig_filter) FD: 128 BD: 1 .+.+: kn->active#69 ->fs_reclaim ->&c->lock ->stock_lock ->&kernfs_locks->open_file_mutex[count] FD: 28 BD: 11 +...: &cnet->ecache.dying_lock FD: 25 BD: 1 ..-.: &(&cnet->ecache.dwork)->timer FD: 30 BD: 1 +.+.: (wq_completion)phy22 ->(work_completion)(&local->reconfig_filter) FD: 30 BD: 1 +.+.: (wq_completion)phy26 ->(work_completion)(&local->reconfig_filter) FD: 30 BD: 1 +.+.: (wq_completion)phy24 ->(work_completion)(&local->reconfig_filter) FD: 30 BD: 1 +.+.: (wq_completion)phy23 ->&rq->__lock ->(work_completion)(&local->reconfig_filter) FD: 30 BD: 1 +.+.: (wq_completion)phy25 ->(work_completion)(&local->reconfig_filter) FD: 6 BD: 78 ....: &dtab->index_lock ->stock_lock ->pool_lock#2 ->&obj_hash[i].lock FD: 30 BD: 1 +.+.: (wq_completion)phy27 ->(work_completion)(&local->reconfig_filter) FD: 126 BD: 78 +.+.: &tn->idrinfo->lock#3 ->fs_reclaim ->pool_lock#2 FD: 5 BD: 79 +...: &(to_police(*a)->tcfp_lock) ->tk_core.seq.seqcount FD: 160 BD: 1 +.+.: (wq_completion)bond8#3 ->(work_completion)(&(&slave->notify_work)->work) FD: 128 BD: 1 .+.+: kn->active#70 ->&rq->__lock ->fs_reclaim ->stock_lock ->&c->lock ->&kernfs_locks->open_file_mutex[count] ->&n->list_lock FD: 129 BD: 1 .+.+: kn->active#71 ->fs_reclaim ->stock_lock ->&kernfs_locks->open_file_mutex[count] ->&prev->lock#2 ->&rq->__lock ->&c->lock ->&____s->seqcount#2 ->&____s->seqcount FD: 128 BD: 1 .+.+: kn->active#72 ->fs_reclaim ->stock_lock ->&kernfs_locks->open_file_mutex[count] ->&c->lock ->&rq->__lock ->&n->list_lock FD: 1 BD: 8 ....: &prev->lock#2 FD: 160 BD: 1 +.+.: (wq_completion)bond9#2 ->(work_completion)(&(&slave->notify_work)->work) FD: 102 BD: 1 +.-.: (&dp->dccps_xmit_timer) ->slock-AF_INET6 FD: 160 BD: 1 +.+.: (wq_completion)bond9#3 ->(work_completion)(&(&slave->notify_work)->work) FD: 144 BD: 78 +.+.: team->team_lock_key#10 ->&rq->__lock ->mode_list_lock ->fs_reclaim ->&____s->seqcount ->pool_lock#2 ->pcpu_alloc_mutex ->&obj_hash[i].lock ->&c->lock ->&____s->seqcount#2 ->&n->list_lock ->nl_table_lock ->rlock-AF_NETLINK ->nl_table_wait.lock ->(console_sem).lock FD: 144 BD: 78 +.+.: team->team_lock_key#11 ->mode_list_lock ->fs_reclaim ->&____s->seqcount ->pcpu_alloc_mutex ->&obj_hash[i].lock ->&c->lock ->&n->list_lock ->&rq->__lock ->rcu_node_0 ->&____s->seqcount#2 ->nl_table_lock ->rlock-AF_NETLINK ->nl_table_wait.lock ->(console_sem).lock FD: 25 BD: 1 +.-.: (&data->send_timer) FD: 1 BD: 1 ....: _rs.lock#22 FD: 296 BD: 6 +.+.: net_dm_mutex ->&rq->__lock ->&obj_hash[i].lock ->fs_reclaim ->&c->lock ->pool_lock#2 ->&data->lock ->tracepoints_mutex FD: 13 BD: 4178 ..-.: &data->lock ->&obj_hash[i].lock ->&base->lock FD: 129 BD: 2 +.+.: (work_completion)(&data->dm_alert_work) ->fs_reclaim ->batched_entropy_u8.lock ->kfence_freelist_lock ->pool_lock#2 ->&data->lock ->&obj_hash[i].lock ->nl_table_lock ->nl_table_wait.lock ->&c->lock ->&n->list_lock ->rcu_node_0 ->&rcu_state.expedited_wq ->&rq->__lock ->&meta->lock ->remove_cache_srcu ->&____s->seqcount#2 ->&____s->seqcount ->quarantine_lock FD: 30 BD: 1 +.+.: (wq_completion)phy28 ->(work_completion)(&local->reconfig_filter) FD: 30 BD: 1 +.+.: (wq_completion)phy29 ->(work_completion)(&local->reconfig_filter) FD: 1 BD: 2 +.-.: &x->lock FD: 18 BD: 3836 +...: _xmit_ETHER/4 ->&obj_hash[i].lock ->krc.lock FD: 24 BD: 10 ....: &root->deactivate_waitq ->&p->pi_lock FD: 50 BD: 85 +...: &dev_addr_list_lock_key#2/3 ->&dev_addr_list_lock_key ->pool_lock#2 ->&c->lock ->&obj_hash[i].lock ->krc.lock ->&vlan_netdev_addr_lock_key/1 FD: 2 BD: 1 +.-.: (&x->rtimer) ->&x->lock FD: 168 BD: 1 +.+.: (wq_completion)bond11 ->&rq->__lock ->(work_completion)(&(&bond->mii_work)->work) ->(work_completion)(&(&bond->ad_work)->work) ->&cfs_rq->removed.lock ->&obj_hash[i].lock FD: 9 BD: 95 +...: &____s->seqcount#15 ->&____s->seqcount#2 ->&____s->seqcount ->pool_lock#2 ->&c->lock FD: 10 BD: 94 +...: &(&bp->lock)->lock ->&____s->seqcount#15 FD: 24 BD: 3837 ....: &tfile->socket.wq.wait ->&p->pi_lock FD: 18 BD: 3836 +...: &macvlan_netdev_addr_lock_key/2 ->&obj_hash[i].lock ->krc.lock FD: 32 BD: 2 +.+.: (work_completion)(&aux->work)#2 ->&aux->poke_mutex ->&rq->__lock ->map_idr_lock ->&obj_hash[i].lock FD: 1 BD: 88 +.+.: pack_mutex.wait_lock FD: 170 BD: 1 +.+.: (wq_completion)bond10 ->(work_completion)(&(&bond->mii_work)->work) ->(work_completion)(&(&bond->ad_work)->work) ->(work_completion)(&(&slave->notify_work)->work) ->(work_completion)(&(&bond->slave_arr_work)->work) ->&rq->__lock FD: 23 BD: 3 +.+.: &aux->poke_mutex ->&rq->__lock ->&cfs_rq->removed.lock ->&obj_hash[i].lock ->pool_lock#2 FD: 169 BD: 1 +.+.: (wq_completion)bond13 ->(work_completion)(&(&bond->mii_work)->work) ->(work_completion)(&(&bond->ad_work)->work) ->(work_completion)(&(&slave->notify_work)->work) ->&rq->__lock FD: 1 BD: 1 ....: _rs.lock#23 FD: 126 BD: 78 +.+.: &tn->idrinfo->lock#4 ->fs_reclaim ->pool_lock#2 ->&obj_hash[i].lock ->&c->lock ->&n->list_lock ->&rq->__lock FD: 1 BD: 6 +.+.: __ip_vs_app_mutex.wait_lock FD: 376 BD: 1 +.+.: sk_lock-AF_RXRPC ->slock-AF_RXRPC ->&rxnet->local_mutex ->&lock->wait_lock ->&p->pi_lock ->&rq->__lock ->&obj_hash[i].lock ->pool_lock#2 ->&local->services_lock FD: 1 BD: 2 +...: slock-AF_RXRPC FD: 1 BD: 80 +...: &net->can.rcvlists_lock FD: 1 BD: 10 +.+.: &msft->filter_lock FD: 1 BD: 10 +.+.: (work_completion)(&(&hdev->interleave_scan)->work) FD: 1 BD: 2 +.+.: loop_validate_mutex.wait_lock FD: 1 BD: 1 ....: _rs.lock#24 FD: 23 BD: 10 +.+.: (work_completion)(&(&hdev->rpa_expired)->work) ->&rq->__lock FD: 2 BD: 1 +.-.: (&policy->timer) ->&policy->lock FD: 1 BD: 1 ....: &so->wait FD: 1 BD: 3892 +..-: &____s->seqcount#16 FD: 459 BD: 1 +.+.: (wq_completion)kstrp ->(work_completion)(&strp->work)#2 FD: 458 BD: 4 +.+.: (work_completion)(&strp->work)#2 ->sk_lock-AF_INET ->slock-AF_INET FD: 1 BD: 3 +.+.: (work_completion)(&(&strp->msg_timer_work)->work) FD: 1 BD: 2 .+.+: sb_writers#15 FD: 130 BD: 1 +.+.: recent_mutex ->&rq->__lock ->fs_reclaim ->pool_lock#2 ->proc_subdir_lock ->proc_inum_ida.xa_lock ->recent_lock ->&ent->pde_unload_lock ->&obj_hash[i].lock FD: 1 BD: 1 ....: _rs.lock#25 FD: 128 BD: 1 .+.+: kn->active#73 ->&rq->__lock ->fs_reclaim ->stock_lock ->&kernfs_locks->open_file_mutex[count] ->&c->lock FD: 1 BD: 4 ++++: ax25_uid_lock FD: 1 BD: 4 +...: ax25_dev_lock FD: 232 BD: 1 +.+.: (wq_completion)bond14 ->(work_completion)(&(&slave->notify_work)->work) ->(work_completion)(&(&bond->mcast_work)->work) FD: 232 BD: 1 +.+.: (wq_completion)bond24 ->(work_completion)(&(&slave->notify_work)->work) ->(work_completion)(&(&bond->mcast_work)->work) FD: 232 BD: 1 +.+.: (wq_completion)bond15 ->(work_completion)(&(&slave->notify_work)->work) ->(work_completion)(&(&bond->mcast_work)->work) FD: 232 BD: 1 +.+.: (wq_completion)bond15#2 ->(work_completion)(&(&slave->notify_work)->work) ->(work_completion)(&(&bond->mcast_work)->work) FD: 232 BD: 1 +.+.: (wq_completion)bond16 ->(work_completion)(&(&slave->notify_work)->work) ->(work_completion)(&(&bond->mcast_work)->work) FD: 232 BD: 1 +.+.: (wq_completion)bond17 ->(work_completion)(&(&slave->notify_work)->work) ->(work_completion)(&(&bond->mcast_work)->work) FD: 232 BD: 1 +.+.: (wq_completion)bond18 ->(work_completion)(&(&slave->notify_work)->work) ->(work_completion)(&(&bond->mcast_work)->work) FD: 232 BD: 1 +.+.: (wq_completion)bond19 ->(work_completion)(&(&slave->notify_work)->work) ->(work_completion)(&(&bond->mcast_work)->work) FD: 232 BD: 1 +.+.: (wq_completion)bond21 ->(work_completion)(&(&slave->notify_work)->work) ->(work_completion)(&(&bond->mcast_work)->work) FD: 232 BD: 1 +.+.: (wq_completion)bond20 ->(work_completion)(&(&slave->notify_work)->work) ->(work_completion)(&(&bond->mcast_work)->work) FD: 232 BD: 1 +.+.: (wq_completion)bond22 ->(work_completion)(&(&slave->notify_work)->work) ->(work_completion)(&(&bond->mcast_work)->work) FD: 232 BD: 1 +.+.: (wq_completion)bond23 ->(work_completion)(&(&slave->notify_work)->work) ->(work_completion)(&(&bond->mcast_work)->work) FD: 232 BD: 1 +.+.: (wq_completion)bond26 ->(work_completion)(&(&slave->notify_work)->work) ->(work_completion)(&(&bond->mcast_work)->work) FD: 232 BD: 1 +.+.: (wq_completion)bond25 ->(work_completion)(&(&slave->notify_work)->work) ->(work_completion)(&(&bond->mcast_work)->work) FD: 232 BD: 1 +.+.: (wq_completion)bond16#2 ->(work_completion)(&(&slave->notify_work)->work) ->(work_completion)(&(&bond->mcast_work)->work) FD: 232 BD: 1 +.+.: (wq_completion)bond28 ->(work_completion)(&(&slave->notify_work)->work) ->(work_completion)(&(&bond->mcast_work)->work) FD: 232 BD: 1 +.+.: (wq_completion)bond27 ->(work_completion)(&(&slave->notify_work)->work) ->(work_completion)(&(&bond->mcast_work)->work) FD: 232 BD: 1 +.+.: (wq_completion)bond29 ->(work_completion)(&(&slave->notify_work)->work) ->(work_completion)(&(&bond->mcast_work)->work) FD: 232 BD: 1 +.+.: (wq_completion)bond30 ->(work_completion)(&(&slave->notify_work)->work) ->(work_completion)(&(&bond->mcast_work)->work) FD: 232 BD: 1 +.+.: (wq_completion)bond31 ->(work_completion)(&(&slave->notify_work)->work) ->(work_completion)(&(&bond->mcast_work)->work) FD: 232 BD: 1 +.+.: (wq_completion)bond17#2 ->(work_completion)(&(&slave->notify_work)->work) ->(work_completion)(&(&bond->mcast_work)->work) FD: 232 BD: 1 +.+.: (wq_completion)bond32 ->(work_completion)(&(&slave->notify_work)->work) ->(work_completion)(&(&bond->mcast_work)->work) FD: 232 BD: 1 +.+.: (wq_completion)bond18#2 ->(work_completion)(&(&slave->notify_work)->work) ->(work_completion)(&(&bond->mcast_work)->work) FD: 232 BD: 1 +.+.: (wq_completion)bond19#2 ->(work_completion)(&(&slave->notify_work)->work) ->(work_completion)(&(&bond->mcast_work)->work) ->&rq->__lock FD: 232 BD: 1 +.+.: (wq_completion)bond20#2 ->(work_completion)(&(&slave->notify_work)->work) ->(work_completion)(&(&bond->mcast_work)->work) FD: 232 BD: 1 +.+.: (wq_completion)bond21#2 ->(work_completion)(&(&slave->notify_work)->work) ->(work_completion)(&(&bond->mcast_work)->work) FD: 232 BD: 1 +.+.: (wq_completion)bond22#2 ->(work_completion)(&(&slave->notify_work)->work) ->(work_completion)(&(&bond->mcast_work)->work) FD: 232 BD: 1 +.+.: (wq_completion)bond36 ->(work_completion)(&(&slave->notify_work)->work) ->(work_completion)(&(&bond->mcast_work)->work) FD: 232 BD: 1 +.+.: (wq_completion)bond37 ->(work_completion)(&(&slave->notify_work)->work) ->(work_completion)(&(&bond->mcast_work)->work) FD: 232 BD: 1 +.+.: (wq_completion)bond38 ->(work_completion)(&(&slave->notify_work)->work) ->(work_completion)(&(&bond->mcast_work)->work) FD: 232 BD: 1 +.+.: (wq_completion)bond2#2 ->(work_completion)(&(&slave->notify_work)->work) ->(work_completion)(&(&bond->mcast_work)->work) FD: 232 BD: 1 +.+.: (wq_completion)bond3#3 ->(work_completion)(&(&slave->notify_work)->work) ->(work_completion)(&(&bond->mcast_work)->work) FD: 232 BD: 1 +.+.: (wq_completion)bond26#2 ->(work_completion)(&(&slave->notify_work)->work) ->(work_completion)(&(&bond->mcast_work)->work) FD: 232 BD: 1 +.+.: (wq_completion)bond4#2 ->(work_completion)(&(&slave->notify_work)->work) ->(work_completion)(&(&bond->mcast_work)->work) FD: 232 BD: 1 +.+.: (wq_completion)bond27#2 ->(work_completion)(&(&slave->notify_work)->work) ->(work_completion)(&(&bond->mcast_work)->work) FD: 232 BD: 1 +.+.: (wq_completion)bond28#2 ->(work_completion)(&(&slave->notify_work)->work) ->(work_completion)(&(&bond->mcast_work)->work) FD: 232 BD: 1 +.+.: (wq_completion)bond29#2 ->(work_completion)(&(&slave->notify_work)->work) ->(work_completion)(&(&bond->mcast_work)->work) FD: 232 BD: 1 +.+.: (wq_completion)bond30#2 ->&rq->__lock ->(work_completion)(&(&slave->notify_work)->work) ->(work_completion)(&(&bond->mcast_work)->work) FD: 2 BD: 1 +.+.: sk_lock-AF_NFC ->slock-AF_NFC FD: 1 BD: 2 +...: slock-AF_NFC FD: 232 BD: 1 +.+.: (wq_completion)bond8#4 ->(work_completion)(&(&slave->notify_work)->work) ->(work_completion)(&(&bond->mcast_work)->work) FD: 232 BD: 1 +.+.: (wq_completion)bond9#4 ->(work_completion)(&(&slave->notify_work)->work) ->(work_completion)(&(&bond->mcast_work)->work) FD: 160 BD: 1 +.+.: (wq_completion)bond51 ->(work_completion)(&(&slave->notify_work)->work) FD: 160 BD: 1 +.+.: (wq_completion)bond52 ->(work_completion)(&(&slave->notify_work)->work) FD: 160 BD: 1 +.+.: (wq_completion)bond53 ->(work_completion)(&(&slave->notify_work)->work) FD: 160 BD: 1 +.+.: (wq_completion)bond54 ->(work_completion)(&(&slave->notify_work)->work) FD: 160 BD: 1 +.+.: (wq_completion)bond55 ->(work_completion)(&(&slave->notify_work)->work) FD: 160 BD: 1 +.+.: (wq_completion)bond56 ->(work_completion)(&(&slave->notify_work)->work) FD: 160 BD: 1 +.+.: (wq_completion)bond57 ->(work_completion)(&(&slave->notify_work)->work) FD: 160 BD: 1 +.+.: (wq_completion)bond58 ->(work_completion)(&(&slave->notify_work)->work) FD: 160 BD: 1 +.+.: (wq_completion)bond59 ->(work_completion)(&(&slave->notify_work)->work) FD: 160 BD: 1 +.+.: (wq_completion)bond60 ->(work_completion)(&(&slave->notify_work)->work) FD: 160 BD: 1 +.+.: (wq_completion)bond43 ->(work_completion)(&(&slave->notify_work)->work) FD: 160 BD: 1 +.+.: (wq_completion)bond44 ->(work_completion)(&(&slave->notify_work)->work) FD: 160 BD: 1 +.+.: (wq_completion)bond62 ->(work_completion)(&(&slave->notify_work)->work) FD: 160 BD: 1 +.+.: (wq_completion)bond45 ->(work_completion)(&(&slave->notify_work)->work) FD: 160 BD: 1 +.+.: (wq_completion)bond46 ->(work_completion)(&(&slave->notify_work)->work) FD: 160 BD: 1 +.+.: (wq_completion)bond47 ->(work_completion)(&(&slave->notify_work)->work) FD: 160 BD: 1 +.+.: (wq_completion)bond48 ->(work_completion)(&(&slave->notify_work)->work) FD: 160 BD: 1 +.+.: (wq_completion)bond49 ->(work_completion)(&(&slave->notify_work)->work) FD: 160 BD: 1 +.+.: (wq_completion)bond22#3 ->(work_completion)(&(&slave->notify_work)->work) FD: 160 BD: 1 +.+.: (wq_completion)bond50 ->(work_completion)(&(&slave->notify_work)->work) FD: 160 BD: 1 +.+.: (wq_completion)bond23#2 ->(work_completion)(&(&slave->notify_work)->work) FD: 160 BD: 1 +.+.: (wq_completion)bond24#2 ->(work_completion)(&(&slave->notify_work)->work) FD: 160 BD: 1 +.+.: (wq_completion)bond51#2 ->(work_completion)(&(&slave->notify_work)->work) FD: 160 BD: 1 +.+.: (wq_completion)bond25#2 ->(work_completion)(&(&slave->notify_work)->work) FD: 160 BD: 1 +.+.: (wq_completion)bond52#2 ->&rq->__lock ->(work_completion)(&(&slave->notify_work)->work) FD: 160 BD: 1 +.+.: (wq_completion)bond26#3 ->(work_completion)(&(&slave->notify_work)->work) FD: 160 BD: 1 +.+.: (wq_completion)bond53#2 ->(work_completion)(&(&slave->notify_work)->work) FD: 160 BD: 1 +.+.: (wq_completion)bond54#2 ->(work_completion)(&(&slave->notify_work)->work) FD: 160 BD: 1 +.+.: (wq_completion)bond27#3 ->(work_completion)(&(&slave->notify_work)->work) FD: 160 BD: 1 +.+.: (wq_completion)bond28#3 ->(work_completion)(&(&slave->notify_work)->work) FD: 160 BD: 1 +.+.: (wq_completion)bond29#3 ->(work_completion)(&(&slave->notify_work)->work) FD: 160 BD: 1 +.+.: (wq_completion)bond30#3 ->(work_completion)(&(&slave->notify_work)->work) FD: 160 BD: 1 +.+.: (wq_completion)bond31#2 ->(work_completion)(&(&slave->notify_work)->work) FD: 160 BD: 1 +.+.: (wq_completion)bond32#2 ->(work_completion)(&(&slave->notify_work)->work) FD: 160 BD: 1 +.+.: (wq_completion)bond33 ->(work_completion)(&(&slave->notify_work)->work) FD: 160 BD: 1 +.+.: (wq_completion)bond79 ->(work_completion)(&(&slave->notify_work)->work) FD: 160 BD: 1 +.+.: (wq_completion)bond80 ->(work_completion)(&(&slave->notify_work)->work) FD: 232 BD: 1 +.+.: (wq_completion)bond81 ->(work_completion)(&(&slave->notify_work)->work) ->(work_completion)(&(&bond->mcast_work)->work) FD: 232 BD: 1 +.+.: (wq_completion)bond62#2 ->&rq->__lock ->(work_completion)(&(&slave->notify_work)->work) ->(work_completion)(&(&bond->mcast_work)->work) FD: 232 BD: 1 +.+.: (wq_completion)bond96 ->(work_completion)(&(&slave->notify_work)->work) ->(work_completion)(&(&bond->mcast_work)->work) FD: 232 BD: 1 +.+.: (wq_completion)bond67 ->(work_completion)(&(&slave->notify_work)->work) ->(work_completion)(&(&bond->mcast_work)->work) FD: 232 BD: 1 +.+.: (wq_completion)bond97 ->(work_completion)(&(&slave->notify_work)->work) ->(work_completion)(&(&bond->mcast_work)->work) FD: 232 BD: 1 +.+.: (wq_completion)bond85 ->(work_completion)(&(&slave->notify_work)->work) ->(work_completion)(&(&bond->mcast_work)->work) FD: 232 BD: 1 +.+.: (wq_completion)bond68 ->(work_completion)(&(&slave->notify_work)->work) ->(work_completion)(&(&bond->mcast_work)->work) FD: 232 BD: 1 +.+.: (wq_completion)bond98 ->(work_completion)(&(&slave->notify_work)->work) ->(work_completion)(&(&bond->mcast_work)->work) FD: 232 BD: 1 +.+.: (wq_completion)bond69 ->(work_completion)(&(&slave->notify_work)->work) ->(work_completion)(&(&bond->mcast_work)->work) FD: 232 BD: 1 +.+.: (wq_completion)bond99 ->(work_completion)(&(&slave->notify_work)->work) ->(work_completion)(&(&bond->mcast_work)->work) FD: 232 BD: 1 +.+.: (wq_completion)bond70 ->(work_completion)(&(&slave->notify_work)->work) ->(work_completion)(&(&bond->mcast_work)->work) FD: 232 BD: 1 +.+.: (wq_completion)bond100 ->(work_completion)(&(&slave->notify_work)->work) ->(work_completion)(&(&bond->mcast_work)->work) FD: 232 BD: 1 +.+.: (wq_completion)bond88 ->(work_completion)(&(&slave->notify_work)->work) ->(work_completion)(&(&bond->mcast_work)->work) FD: 232 BD: 1 +.+.: (wq_completion)bond71 ->(work_completion)(&(&slave->notify_work)->work) ->(work_completion)(&(&bond->mcast_work)->work) FD: 232 BD: 1 +.+.: (wq_completion)bond101 ->(work_completion)(&(&slave->notify_work)->work) ->(work_completion)(&(&bond->mcast_work)->work) FD: 232 BD: 1 +.+.: (wq_completion)bond89 ->(work_completion)(&(&slave->notify_work)->work) ->(work_completion)(&(&bond->mcast_work)->work) FD: 232 BD: 1 +.+.: (wq_completion)bond72 ->&rq->__lock ->(work_completion)(&(&slave->notify_work)->work) ->(work_completion)(&(&bond->mcast_work)->work) FD: 232 BD: 1 +.+.: (wq_completion)bond102 ->&rq->__lock ->(work_completion)(&(&slave->notify_work)->work) ->(work_completion)(&(&bond->mcast_work)->work) FD: 232 BD: 1 +.+.: (wq_completion)bond73 ->(work_completion)(&(&slave->notify_work)->work) ->(work_completion)(&(&bond->mcast_work)->work) FD: 232 BD: 1 +.+.: (wq_completion)bond103 ->&rq->__lock ->(work_completion)(&(&slave->notify_work)->work) ->(work_completion)(&(&bond->mcast_work)->work) FD: 232 BD: 1 +.+.: (wq_completion)bond74 ->(work_completion)(&(&slave->notify_work)->work) ->(work_completion)(&(&bond->mcast_work)->work) FD: 232 BD: 1 +.+.: (wq_completion)bond104 ->(work_completion)(&(&slave->notify_work)->work) ->(work_completion)(&(&bond->mcast_work)->work) FD: 232 BD: 1 +.+.: (wq_completion)bond92 ->(work_completion)(&(&slave->notify_work)->work) ->(work_completion)(&(&bond->mcast_work)->work) FD: 232 BD: 1 +.+.: (wq_completion)bond75 ->(work_completion)(&(&slave->notify_work)->work) ->(work_completion)(&(&bond->mcast_work)->work) FD: 232 BD: 1 +.+.: (wq_completion)bond105 ->(work_completion)(&(&slave->notify_work)->work) ->(work_completion)(&(&bond->mcast_work)->work) FD: 232 BD: 1 +.+.: (wq_completion)bond93 ->(work_completion)(&(&slave->notify_work)->work) ->(work_completion)(&(&bond->mcast_work)->work) FD: 232 BD: 1 +.+.: (wq_completion)bond76 ->&rq->__lock ->(work_completion)(&(&slave->notify_work)->work) ->(work_completion)(&(&bond->mcast_work)->work) FD: 232 BD: 1 +.+.: (wq_completion)bond106 ->(work_completion)(&(&slave->notify_work)->work) ->(work_completion)(&(&bond->mcast_work)->work) FD: 232 BD: 1 +.+.: (wq_completion)bond77 ->(work_completion)(&(&slave->notify_work)->work) ->(work_completion)(&(&bond->mcast_work)->work) FD: 232 BD: 1 +.+.: (wq_completion)bond94 ->(work_completion)(&(&slave->notify_work)->work) ->(work_completion)(&(&bond->mcast_work)->work) FD: 232 BD: 1 +.+.: (wq_completion)bond107 ->(work_completion)(&(&slave->notify_work)->work) ->(work_completion)(&(&bond->mcast_work)->work) FD: 232 BD: 1 +.+.: (wq_completion)bond78 ->(work_completion)(&(&slave->notify_work)->work) ->(work_completion)(&(&bond->mcast_work)->work) ->&rq->__lock FD: 232 BD: 1 +.+.: (wq_completion)bond95 ->&rq->__lock ->(work_completion)(&(&slave->notify_work)->work) ->(work_completion)(&(&bond->mcast_work)->work) FD: 232 BD: 1 +.+.: (wq_completion)bond108 ->(work_completion)(&(&slave->notify_work)->work) ->(work_completion)(&(&bond->mcast_work)->work) FD: 232 BD: 1 +.+.: (wq_completion)bond79#2 ->(work_completion)(&(&slave->notify_work)->work) ->(work_completion)(&(&bond->mcast_work)->work) FD: 232 BD: 1 +.+.: (wq_completion)bond96#2 ->(work_completion)(&(&slave->notify_work)->work) ->(work_completion)(&(&bond->mcast_work)->work) FD: 232 BD: 1 +.+.: (wq_completion)bond109 ->(work_completion)(&(&slave->notify_work)->work) ->(work_completion)(&(&bond->mcast_work)->work) ->&rq->__lock FD: 232 BD: 1 +.+.: (wq_completion)bond80#2 ->(work_completion)(&(&slave->notify_work)->work) ->(work_completion)(&(&bond->mcast_work)->work) ->&rq->__lock FD: 232 BD: 1 +.+.: (wq_completion)bond97#2 ->(work_completion)(&(&slave->notify_work)->work) ->(work_completion)(&(&bond->mcast_work)->work) FD: 232 BD: 1 +.+.: (wq_completion)bond110 ->(work_completion)(&(&slave->notify_work)->work) ->(work_completion)(&(&bond->mcast_work)->work) FD: 1 BD: 81 ..-.: key#28 FD: 232 BD: 1 +.+.: (wq_completion)bond81#2 ->(work_completion)(&(&slave->notify_work)->work) ->(work_completion)(&(&bond->mcast_work)->work) ->&rq->__lock FD: 232 BD: 1 +.+.: (wq_completion)bond98#2 ->&rq->__lock ->(work_completion)(&(&slave->notify_work)->work) ->(work_completion)(&(&bond->mcast_work)->work) FD: 232 BD: 1 +.+.: (wq_completion)bond111 ->&rq->__lock ->(work_completion)(&(&slave->notify_work)->work) ->(work_completion)(&(&bond->mcast_work)->work) FD: 232 BD: 1 +.+.: (wq_completion)bond82 ->(work_completion)(&(&slave->notify_work)->work) ->(work_completion)(&(&bond->mcast_work)->work) FD: 232 BD: 1 +.+.: (wq_completion)bond99#2 ->(work_completion)(&(&slave->notify_work)->work) ->(work_completion)(&(&bond->mcast_work)->work) ->&rq->__lock FD: 232 BD: 1 +.+.: (wq_completion)bond13#2 ->(work_completion)(&(&slave->notify_work)->work) ->(work_completion)(&(&bond->mcast_work)->work) FD: 232 BD: 1 +.+.: (wq_completion)bond112 ->(work_completion)(&(&slave->notify_work)->work) ->(work_completion)(&(&bond->mcast_work)->work) FD: 232 BD: 1 +.+.: (wq_completion)bond83 ->(work_completion)(&(&slave->notify_work)->work) ->(work_completion)(&(&bond->mcast_work)->work) FD: 232 BD: 1 +.+.: (wq_completion)bond100#2 ->(work_completion)(&(&slave->notify_work)->work) ->(work_completion)(&(&bond->mcast_work)->work) FD: 232 BD: 1 +.+.: (wq_completion)bond113 ->(work_completion)(&(&slave->notify_work)->work) ->(work_completion)(&(&bond->mcast_work)->work) FD: 232 BD: 1 +.+.: (wq_completion)bond84 ->(work_completion)(&(&slave->notify_work)->work) ->(work_completion)(&(&bond->mcast_work)->work) FD: 232 BD: 1 +.+.: (wq_completion)bond101#2 ->(work_completion)(&(&slave->notify_work)->work) ->(work_completion)(&(&bond->mcast_work)->work) ->&rq->__lock FD: 232 BD: 1 +.+.: (wq_completion)bond14#2 ->(work_completion)(&(&slave->notify_work)->work) ->(work_completion)(&(&bond->mcast_work)->work) FD: 232 BD: 1 +.+.: (wq_completion)bond114 ->(work_completion)(&(&slave->notify_work)->work) ->(work_completion)(&(&bond->mcast_work)->work) FD: 232 BD: 1 +.+.: (wq_completion)bond85#2 ->(work_completion)(&(&slave->notify_work)->work) ->(work_completion)(&(&bond->mcast_work)->work) FD: 232 BD: 1 +.+.: (wq_completion)bond102#2 ->(work_completion)(&(&slave->notify_work)->work) ->(work_completion)(&(&bond->mcast_work)->work) ->&rq->__lock FD: 232 BD: 1 +.+.: (wq_completion)bond15#3 ->&rq->__lock ->(work_completion)(&(&slave->notify_work)->work) ->(work_completion)(&(&bond->mcast_work)->work) FD: 232 BD: 1 +.+.: (wq_completion)bond115 ->(work_completion)(&(&slave->notify_work)->work) ->(work_completion)(&(&bond->mcast_work)->work) ->&rq->__lock FD: 232 BD: 1 +.+.: (wq_completion)bond86 ->(work_completion)(&(&slave->notify_work)->work) ->(work_completion)(&(&bond->mcast_work)->work) FD: 232 BD: 1 +.+.: (wq_completion)bond103#2 ->(work_completion)(&(&slave->notify_work)->work) ->(work_completion)(&(&bond->mcast_work)->work) FD: 232 BD: 1 +.+.: (wq_completion)bond116 ->(work_completion)(&(&slave->notify_work)->work) ->(work_completion)(&(&bond->mcast_work)->work) FD: 232 BD: 1 +.+.: (wq_completion)bond87 ->(work_completion)(&(&slave->notify_work)->work) ->(work_completion)(&(&bond->mcast_work)->work) FD: 232 BD: 1 +.+.: (wq_completion)bond104#2 ->&rq->__lock ->(work_completion)(&(&slave->notify_work)->work) ->(work_completion)(&(&bond->mcast_work)->work) FD: 232 BD: 1 +.+.: (wq_completion)bond117 ->(work_completion)(&(&slave->notify_work)->work) ->(work_completion)(&(&bond->mcast_work)->work) FD: 232 BD: 1 +.+.: (wq_completion)bond88#2 ->(work_completion)(&(&slave->notify_work)->work) ->(work_completion)(&(&bond->mcast_work)->work) FD: 232 BD: 1 +.+.: (wq_completion)bond105#2 ->(work_completion)(&(&slave->notify_work)->work) ->(work_completion)(&(&bond->mcast_work)->work) FD: 232 BD: 1 +.+.: (wq_completion)bond17#3 ->(work_completion)(&(&slave->notify_work)->work) ->(work_completion)(&(&bond->mcast_work)->work) FD: 232 BD: 1 +.+.: (wq_completion)bond118 ->(work_completion)(&(&slave->notify_work)->work) ->(work_completion)(&(&bond->mcast_work)->work) FD: 232 BD: 1 +.+.: (wq_completion)bond89#2 ->(work_completion)(&(&slave->notify_work)->work) ->(work_completion)(&(&bond->mcast_work)->work) FD: 232 BD: 1 +.+.: (wq_completion)bond106#2 ->(work_completion)(&(&slave->notify_work)->work) ->(work_completion)(&(&bond->mcast_work)->work) ->&rq->__lock FD: 232 BD: 1 +.+.: (wq_completion)bond119 ->(work_completion)(&(&slave->notify_work)->work) ->(work_completion)(&(&bond->mcast_work)->work) FD: 232 BD: 1 +.+.: (wq_completion)bond90 ->(work_completion)(&(&slave->notify_work)->work) ->(work_completion)(&(&bond->mcast_work)->work) FD: 232 BD: 1 +.+.: (wq_completion)bond107#2 ->(work_completion)(&(&slave->notify_work)->work) ->(work_completion)(&(&bond->mcast_work)->work) ->&rq->__lock FD: 232 BD: 1 +.+.: (wq_completion)bond120 ->(work_completion)(&(&slave->notify_work)->work) ->(work_completion)(&(&bond->mcast_work)->work) FD: 232 BD: 1 +.+.: (wq_completion)bond91 ->(work_completion)(&(&slave->notify_work)->work) ->(work_completion)(&(&bond->mcast_work)->work) FD: 232 BD: 1 +.+.: (wq_completion)bond18#3 ->(work_completion)(&(&slave->notify_work)->work) ->(work_completion)(&(&bond->mcast_work)->work) FD: 232 BD: 1 +.+.: (wq_completion)bond108#2 ->(work_completion)(&(&slave->notify_work)->work) ->(work_completion)(&(&bond->mcast_work)->work) ->&rq->__lock FD: 232 BD: 1 +.+.: (wq_completion)bond121 ->(work_completion)(&(&slave->notify_work)->work) ->(work_completion)(&(&bond->mcast_work)->work) FD: 232 BD: 1 +.+.: (wq_completion)bond92#2 ->(work_completion)(&(&slave->notify_work)->work) ->(work_completion)(&(&bond->mcast_work)->work) FD: 232 BD: 1 +.+.: (wq_completion)bond109#2 ->&rq->__lock ->(work_completion)(&(&slave->notify_work)->work) ->(work_completion)(&(&bond->mcast_work)->work) FD: 232 BD: 1 +.+.: (wq_completion)bond122 ->(work_completion)(&(&slave->notify_work)->work) ->(work_completion)(&(&bond->mcast_work)->work) ->&rq->__lock FD: 232 BD: 1 +.+.: (wq_completion)bond93#2 ->(work_completion)(&(&slave->notify_work)->work) ->(work_completion)(&(&bond->mcast_work)->work) FD: 232 BD: 1 +.+.: (wq_completion)bond110#2 ->(work_completion)(&(&slave->notify_work)->work) ->(work_completion)(&(&bond->mcast_work)->work) FD: 232 BD: 1 +.+.: (wq_completion)bond123 ->(work_completion)(&(&slave->notify_work)->work) ->(work_completion)(&(&bond->mcast_work)->work) ->&rq->__lock ->&cfs_rq->removed.lock ->&obj_hash[i].lock FD: 232 BD: 1 +.+.: (wq_completion)bond94#2 ->(work_completion)(&(&slave->notify_work)->work) ->(work_completion)(&(&bond->mcast_work)->work) ->&rq->__lock FD: 232 BD: 1 +.+.: (wq_completion)bond111#2 ->(work_completion)(&(&slave->notify_work)->work) ->(work_completion)(&(&bond->mcast_work)->work) ->&rq->__lock FD: 232 BD: 1 +.+.: (wq_completion)bond124 ->(work_completion)(&(&slave->notify_work)->work) ->(work_completion)(&(&bond->mcast_work)->work) FD: 232 BD: 1 +.+.: (wq_completion)bond95#2 ->(work_completion)(&(&slave->notify_work)->work) ->(work_completion)(&(&bond->mcast_work)->work) FD: 232 BD: 1 +.+.: (wq_completion)bond20#3 ->(work_completion)(&(&slave->notify_work)->work) ->(work_completion)(&(&bond->mcast_work)->work) ->&rq->__lock FD: 232 BD: 1 +.+.: (wq_completion)bond112#2 ->(work_completion)(&(&slave->notify_work)->work) ->(work_completion)(&(&bond->mcast_work)->work) FD: 232 BD: 1 +.+.: (wq_completion)bond125 ->(work_completion)(&(&slave->notify_work)->work) ->(work_completion)(&(&bond->mcast_work)->work) FD: 232 BD: 1 +.+.: (wq_completion)bond96#3 ->(work_completion)(&(&slave->notify_work)->work) ->(work_completion)(&(&bond->mcast_work)->work) FD: 232 BD: 1 +.+.: (wq_completion)bond113#2 ->(work_completion)(&(&slave->notify_work)->work) ->(work_completion)(&(&bond->mcast_work)->work) FD: 232 BD: 1 +.+.: (wq_completion)bond21#3 ->(work_completion)(&(&slave->notify_work)->work) ->(work_completion)(&(&bond->mcast_work)->work) FD: 232 BD: 1 +.+.: (wq_completion)bond126 ->&rq->__lock ->(work_completion)(&(&slave->notify_work)->work) ->(work_completion)(&(&bond->mcast_work)->work) FD: 232 BD: 1 +.+.: (wq_completion)bond97#3 ->(work_completion)(&(&slave->notify_work)->work) ->(work_completion)(&(&bond->mcast_work)->work) ->&rq->__lock FD: 232 BD: 1 +.+.: (wq_completion)bond114#2 ->(work_completion)(&(&slave->notify_work)->work) ->(work_completion)(&(&bond->mcast_work)->work) ->&rq->__lock FD: 232 BD: 1 +.+.: (wq_completion)bond22#4 ->(work_completion)(&(&slave->notify_work)->work) ->(work_completion)(&(&bond->mcast_work)->work) ->&rq->__lock FD: 232 BD: 1 +.+.: (wq_completion)bond127 ->(work_completion)(&(&slave->notify_work)->work) ->(work_completion)(&(&bond->mcast_work)->work) ->&rq->__lock FD: 232 BD: 1 +.+.: (wq_completion)bond98#3 ->(work_completion)(&(&slave->notify_work)->work) ->(work_completion)(&(&bond->mcast_work)->work) ->&rq->__lock FD: 232 BD: 1 +.+.: (wq_completion)bond115#2 ->(work_completion)(&(&slave->notify_work)->work) ->(work_completion)(&(&bond->mcast_work)->work) FD: 232 BD: 1 +.+.: (wq_completion)bond128 ->(work_completion)(&(&slave->notify_work)->work) ->(work_completion)(&(&bond->mcast_work)->work) ->&rq->__lock FD: 232 BD: 1 +.+.: (wq_completion)bond99#3 ->&rq->__lock ->(work_completion)(&(&slave->notify_work)->work) ->(work_completion)(&(&bond->mcast_work)->work) FD: 232 BD: 1 +.+.: (wq_completion)bond23#3 ->(work_completion)(&(&slave->notify_work)->work) ->(work_completion)(&(&bond->mcast_work)->work) ->&rq->__lock FD: 232 BD: 1 +.+.: (wq_completion)bond116#2 ->&rq->__lock ->(work_completion)(&(&slave->notify_work)->work) ->(work_completion)(&(&bond->mcast_work)->work) FD: 232 BD: 1 +.+.: (wq_completion)bond129 ->(work_completion)(&(&slave->notify_work)->work) ->(work_completion)(&(&bond->mcast_work)->work) ->&rq->__lock FD: 232 BD: 1 +.+.: (wq_completion)bond100#3 ->(work_completion)(&(&slave->notify_work)->work) ->(work_completion)(&(&bond->mcast_work)->work) ->&rq->__lock FD: 232 BD: 1 +.+.: (wq_completion)bond117#2 ->&rq->__lock ->(work_completion)(&(&slave->notify_work)->work) ->(work_completion)(&(&bond->mcast_work)->work) FD: 232 BD: 1 +.+.: (wq_completion)bond130 ->(work_completion)(&(&slave->notify_work)->work) ->(work_completion)(&(&bond->mcast_work)->work) FD: 232 BD: 1 +.+.: (wq_completion)bond24#3 ->&rq->__lock ->(work_completion)(&(&slave->notify_work)->work) ->(work_completion)(&(&bond->mcast_work)->work) FD: 232 BD: 1 +.+.: (wq_completion)bond101#3 ->(work_completion)(&(&slave->notify_work)->work) ->(work_completion)(&(&bond->mcast_work)->work) FD: 232 BD: 1 +.+.: (wq_completion)bond118#2 ->(work_completion)(&(&slave->notify_work)->work) ->(work_completion)(&(&bond->mcast_work)->work) ->&rq->__lock FD: 232 BD: 1 +.+.: (wq_completion)bond131 ->(work_completion)(&(&slave->notify_work)->work) ->(work_completion)(&(&bond->mcast_work)->work) ->&rq->__lock FD: 232 BD: 1 +.+.: (wq_completion)bond102#3 ->(work_completion)(&(&slave->notify_work)->work) ->(work_completion)(&(&bond->mcast_work)->work) ->&rq->__lock FD: 232 BD: 1 +.+.: (wq_completion)bond119#2 ->(work_completion)(&(&slave->notify_work)->work) ->(work_completion)(&(&bond->mcast_work)->work) FD: 232 BD: 1 +.+.: (wq_completion)bond132 ->(work_completion)(&(&slave->notify_work)->work) ->(work_completion)(&(&bond->mcast_work)->work) ->&rq->__lock FD: 232 BD: 1 +.+.: (wq_completion)bond103#3 ->(work_completion)(&(&slave->notify_work)->work) ->(work_completion)(&(&bond->mcast_work)->work) FD: 232 BD: 1 +.+.: (wq_completion)bond120#2 ->(work_completion)(&(&slave->notify_work)->work) ->(work_completion)(&(&bond->mcast_work)->work) FD: 232 BD: 1 +.+.: (wq_completion)bond25#3 ->(work_completion)(&(&slave->notify_work)->work) ->(work_completion)(&(&bond->mcast_work)->work) FD: 232 BD: 1 +.+.: (wq_completion)bond133 ->&rq->__lock ->(work_completion)(&(&slave->notify_work)->work) ->(work_completion)(&(&bond->mcast_work)->work) FD: 232 BD: 1 +.+.: (wq_completion)bond104#3 ->(work_completion)(&(&slave->notify_work)->work) ->(work_completion)(&(&bond->mcast_work)->work) ->&rq->__lock FD: 232 BD: 1 +.+.: (wq_completion)bond121#2 ->(work_completion)(&(&slave->notify_work)->work) ->(work_completion)(&(&bond->mcast_work)->work) FD: 232 BD: 1 +.+.: (wq_completion)bond134 ->(work_completion)(&(&slave->notify_work)->work) ->(work_completion)(&(&bond->mcast_work)->work) ->&rq->__lock FD: 232 BD: 1 +.+.: (wq_completion)bond105#3 ->(work_completion)(&(&slave->notify_work)->work) ->(work_completion)(&(&bond->mcast_work)->work) ->&rq->__lock FD: 232 BD: 1 +.+.: (wq_completion)bond26#4 ->(work_completion)(&(&slave->notify_work)->work) ->(work_completion)(&(&bond->mcast_work)->work) ->&rq->__lock FD: 232 BD: 1 +.+.: (wq_completion)bond122#2 ->(work_completion)(&(&slave->notify_work)->work) ->(work_completion)(&(&bond->mcast_work)->work) ->&rq->__lock FD: 232 BD: 1 +.+.: (wq_completion)bond135 ->(work_completion)(&(&slave->notify_work)->work) ->(work_completion)(&(&bond->mcast_work)->work) FD: 232 BD: 1 +.+.: (wq_completion)bond106#3 ->(work_completion)(&(&slave->notify_work)->work) ->(work_completion)(&(&bond->mcast_work)->work) FD: 232 BD: 1 +.+.: (wq_completion)bond123#2 ->(work_completion)(&(&slave->notify_work)->work) ->(work_completion)(&(&bond->mcast_work)->work) ->&rq->__lock FD: 232 BD: 1 +.+.: (wq_completion)bond27#4 ->(work_completion)(&(&slave->notify_work)->work) ->(work_completion)(&(&bond->mcast_work)->work) ->&rq->__lock FD: 232 BD: 1 +.+.: (wq_completion)bond136 ->(work_completion)(&(&slave->notify_work)->work) ->(work_completion)(&(&bond->mcast_work)->work) ->&rq->__lock FD: 232 BD: 1 +.+.: (wq_completion)bond107#3 ->(work_completion)(&(&slave->notify_work)->work) ->(work_completion)(&(&bond->mcast_work)->work) ->&rq->__lock FD: 232 BD: 1 +.+.: (wq_completion)bond124#2 ->(work_completion)(&(&slave->notify_work)->work) ->(work_completion)(&(&bond->mcast_work)->work) FD: 232 BD: 1 +.+.: (wq_completion)bond28#4 ->(work_completion)(&(&slave->notify_work)->work) ->(work_completion)(&(&bond->mcast_work)->work) FD: 232 BD: 1 +.+.: (wq_completion)bond137 ->(work_completion)(&(&slave->notify_work)->work) ->(work_completion)(&(&bond->mcast_work)->work) FD: 232 BD: 1 +.+.: (wq_completion)bond108#3 ->&rq->__lock ->(work_completion)(&(&slave->notify_work)->work) ->(work_completion)(&(&bond->mcast_work)->work) FD: 232 BD: 1 +.+.: (wq_completion)bond125#2 ->&rq->__lock ->(work_completion)(&(&slave->notify_work)->work) ->(work_completion)(&(&bond->mcast_work)->work) FD: 232 BD: 1 +.+.: (wq_completion)bond109#3 ->(work_completion)(&(&slave->notify_work)->work) ->(work_completion)(&(&bond->mcast_work)->work) FD: 232 BD: 1 +.+.: (wq_completion)bond126#2 ->&rq->__lock ->(work_completion)(&(&slave->notify_work)->work) ->(work_completion)(&(&bond->mcast_work)->work) FD: 232 BD: 1 +.+.: (wq_completion)bond139 ->(work_completion)(&(&slave->notify_work)->work) ->(work_completion)(&(&bond->mcast_work)->work) FD: 232 BD: 1 +.+.: (wq_completion)bond110#3 ->(work_completion)(&(&slave->notify_work)->work) ->(work_completion)(&(&bond->mcast_work)->work) FD: 232 BD: 1 +.+.: (wq_completion)bond29#4 ->(work_completion)(&(&slave->notify_work)->work) ->(work_completion)(&(&bond->mcast_work)->work) FD: 232 BD: 1 +.+.: (wq_completion)bond111#3 ->(work_completion)(&(&slave->notify_work)->work) ->(work_completion)(&(&bond->mcast_work)->work) FD: 232 BD: 1 +.+.: (wq_completion)bond127#2 ->(work_completion)(&(&slave->notify_work)->work) ->(work_completion)(&(&bond->mcast_work)->work) FD: 232 BD: 1 +.+.: (wq_completion)bond140 ->(work_completion)(&(&slave->notify_work)->work) ->(work_completion)(&(&bond->mcast_work)->work) FD: 232 BD: 1 +.+.: (wq_completion)bond112#3 ->&rq->__lock ->(work_completion)(&(&slave->notify_work)->work) ->(work_completion)(&(&bond->mcast_work)->work) FD: 232 BD: 1 +.+.: (wq_completion)bond128#2 ->(work_completion)(&(&slave->notify_work)->work) ->(work_completion)(&(&bond->mcast_work)->work) ->&rq->__lock FD: 232 BD: 1 +.+.: (wq_completion)bond30#4 ->(work_completion)(&(&slave->notify_work)->work) ->(work_completion)(&(&bond->mcast_work)->work) ->&rq->__lock FD: 232 BD: 1 +.+.: (wq_completion)bond141 ->(work_completion)(&(&slave->notify_work)->work) ->(work_completion)(&(&bond->mcast_work)->work) FD: 232 BD: 1 +.+.: (wq_completion)bond113#3 ->(work_completion)(&(&slave->notify_work)->work) ->(work_completion)(&(&bond->mcast_work)->work) FD: 232 BD: 1 +.+.: (wq_completion)bond129#2 ->(work_completion)(&(&slave->notify_work)->work) ->(work_completion)(&(&bond->mcast_work)->work) ->&rq->__lock FD: 232 BD: 1 +.+.: (wq_completion)bond31#3 ->&rq->__lock ->(work_completion)(&(&slave->notify_work)->work) ->(work_completion)(&(&bond->mcast_work)->work) FD: 232 BD: 1 +.+.: (wq_completion)bond142 ->(work_completion)(&(&slave->notify_work)->work) ->(work_completion)(&(&bond->mcast_work)->work) FD: 232 BD: 1 +.+.: (wq_completion)bond114#3 ->(work_completion)(&(&slave->notify_work)->work) ->(work_completion)(&(&bond->mcast_work)->work) FD: 232 BD: 1 +.+.: (wq_completion)bond130#2 ->(work_completion)(&(&slave->notify_work)->work) ->(work_completion)(&(&bond->mcast_work)->work) FD: 232 BD: 1 +.+.: (wq_completion)bond143 ->(work_completion)(&(&slave->notify_work)->work) ->(work_completion)(&(&bond->mcast_work)->work) ->&rq->__lock FD: 232 BD: 1 +.+.: (wq_completion)bond115#3 ->(work_completion)(&(&slave->notify_work)->work) ->(work_completion)(&(&bond->mcast_work)->work) ->&rq->__lock FD: 232 BD: 1 +.+.: (wq_completion)bond32#3 ->(work_completion)(&(&slave->notify_work)->work) ->(work_completion)(&(&bond->mcast_work)->work) ->&rq->__lock FD: 232 BD: 1 +.+.: (wq_completion)bond144 ->(work_completion)(&(&slave->notify_work)->work) ->(work_completion)(&(&bond->mcast_work)->work) ->&rq->__lock FD: 232 BD: 1 +.+.: (wq_completion)bond116#3 ->(work_completion)(&(&slave->notify_work)->work) ->(work_completion)(&(&bond->mcast_work)->work) ->&rq->__lock FD: 232 BD: 1 +.+.: (wq_completion)bond132#2 ->(work_completion)(&(&slave->notify_work)->work) ->(work_completion)(&(&bond->mcast_work)->work) ->&rq->__lock FD: 232 BD: 1 +.+.: (wq_completion)bond145 ->(work_completion)(&(&slave->notify_work)->work) ->(work_completion)(&(&bond->mcast_work)->work) FD: 232 BD: 1 +.+.: (wq_completion)bond117#3 ->(work_completion)(&(&slave->notify_work)->work) ->(work_completion)(&(&bond->mcast_work)->work) ->&rq->__lock FD: 232 BD: 1 +.+.: (wq_completion)bond33#2 ->&rq->__lock ->(work_completion)(&(&slave->notify_work)->work) ->(work_completion)(&(&bond->mcast_work)->work) FD: 232 BD: 1 +.+.: (wq_completion)bond133#2 ->(work_completion)(&(&slave->notify_work)->work) ->(work_completion)(&(&bond->mcast_work)->work) ->&rq->__lock FD: 232 BD: 1 +.+.: (wq_completion)bond146 ->&rq->__lock ->(work_completion)(&(&slave->notify_work)->work) ->(work_completion)(&(&bond->mcast_work)->work) FD: 232 BD: 1 +.+.: (wq_completion)bond118#3 ->(work_completion)(&(&slave->notify_work)->work) ->(work_completion)(&(&bond->mcast_work)->work) FD: 232 BD: 1 +.+.: (wq_completion)bond134#2 ->(work_completion)(&(&slave->notify_work)->work) ->(work_completion)(&(&bond->mcast_work)->work) FD: 232 BD: 1 +.+.: (wq_completion)bond119#3 ->(work_completion)(&(&slave->notify_work)->work) ->(work_completion)(&(&bond->mcast_work)->work) FD: 232 BD: 1 +.+.: (wq_completion)bond147 ->(work_completion)(&(&slave->notify_work)->work) ->(work_completion)(&(&bond->mcast_work)->work) FD: 232 BD: 1 +.+.: (wq_completion)bond135#2 ->(work_completion)(&(&slave->notify_work)->work) ->(work_completion)(&(&bond->mcast_work)->work) ->&rq->__lock FD: 232 BD: 1 +.+.: (wq_completion)bond34 ->&rq->__lock ->(work_completion)(&(&slave->notify_work)->work) ->(work_completion)(&(&bond->mcast_work)->work) FD: 232 BD: 1 +.+.: (wq_completion)bond120#3 ->(work_completion)(&(&slave->notify_work)->work) ->(work_completion)(&(&bond->mcast_work)->work) ->&rq->__lock FD: 232 BD: 1 +.+.: (wq_completion)bond148 ->(work_completion)(&(&slave->notify_work)->work) ->(work_completion)(&(&bond->mcast_work)->work) ->&rq->__lock FD: 232 BD: 1 +.+.: (wq_completion)bond136#2 ->(work_completion)(&(&slave->notify_work)->work) ->(work_completion)(&(&bond->mcast_work)->work) FD: 232 BD: 1 +.+.: (wq_completion)bond149 ->(work_completion)(&(&slave->notify_work)->work) ->(work_completion)(&(&bond->mcast_work)->work) ->&rq->__lock FD: 232 BD: 1 +.+.: (wq_completion)bond121#3 ->(work_completion)(&(&slave->notify_work)->work) ->(work_completion)(&(&bond->mcast_work)->work) FD: 232 BD: 1 +.+.: (wq_completion)bond137#2 ->(work_completion)(&(&slave->notify_work)->work) ->(work_completion)(&(&bond->mcast_work)->work) FD: 232 BD: 1 +.+.: (wq_completion)bond150 ->(work_completion)(&(&slave->notify_work)->work) ->(work_completion)(&(&bond->mcast_work)->work) ->&rq->__lock FD: 232 BD: 1 +.+.: (wq_completion)bond122#3 ->(work_completion)(&(&slave->notify_work)->work) ->(work_completion)(&(&bond->mcast_work)->work) ->&rq->__lock FD: 232 BD: 1 +.+.: (wq_completion)bond138 ->(work_completion)(&(&slave->notify_work)->work) ->(work_completion)(&(&bond->mcast_work)->work) FD: 232 BD: 1 +.+.: (wq_completion)bond123#3 ->(work_completion)(&(&slave->notify_work)->work) ->(work_completion)(&(&bond->mcast_work)->work) ->&rq->__lock FD: 232 BD: 1 +.+.: (wq_completion)bond139#2 ->(work_completion)(&(&slave->notify_work)->work) ->(work_completion)(&(&bond->mcast_work)->work) FD: 232 BD: 1 +.+.: (wq_completion)bond152 ->(work_completion)(&(&slave->notify_work)->work) ->(work_completion)(&(&bond->mcast_work)->work) ->&rq->__lock FD: 232 BD: 1 +.+.: (wq_completion)bond124#3 ->(work_completion)(&(&slave->notify_work)->work) ->(work_completion)(&(&bond->mcast_work)->work) ->&rq->__lock FD: 232 BD: 1 +.+.: (wq_completion)bond35 ->(work_completion)(&(&slave->notify_work)->work) ->(work_completion)(&(&bond->mcast_work)->work) ->&rq->__lock FD: 232 BD: 1 +.+.: (wq_completion)bond140#2 ->(work_completion)(&(&slave->notify_work)->work) ->(work_completion)(&(&bond->mcast_work)->work) ->&rq->__lock FD: 232 BD: 1 +.+.: (wq_completion)bond153 ->(work_completion)(&(&slave->notify_work)->work) ->(work_completion)(&(&bond->mcast_work)->work) FD: 232 BD: 1 +.+.: (wq_completion)bond125#3 ->(work_completion)(&(&slave->notify_work)->work) ->(work_completion)(&(&bond->mcast_work)->work) ->&rq->__lock FD: 232 BD: 1 +.+.: (wq_completion)bond141#2 ->&rq->__lock ->(work_completion)(&(&slave->notify_work)->work) ->(work_completion)(&(&bond->mcast_work)->work) FD: 232 BD: 1 +.+.: (wq_completion)bond154 ->(work_completion)(&(&slave->notify_work)->work) ->(work_completion)(&(&bond->mcast_work)->work) ->&rq->__lock FD: 232 BD: 1 +.+.: (wq_completion)bond126#3 ->&rq->__lock ->(work_completion)(&(&slave->notify_work)->work) ->(work_completion)(&(&bond->mcast_work)->work) FD: 232 BD: 1 +.+.: (wq_completion)bond36#2 ->(work_completion)(&(&slave->notify_work)->work) ->(work_completion)(&(&bond->mcast_work)->work) ->&rq->__lock FD: 232 BD: 1 +.+.: (wq_completion)bond142#2 ->(work_completion)(&(&slave->notify_work)->work) ->(work_completion)(&(&bond->mcast_work)->work) ->&rq->__lock FD: 232 BD: 1 +.+.: (wq_completion)bond155 ->(work_completion)(&(&slave->notify_work)->work) ->(work_completion)(&(&bond->mcast_work)->work) FD: 232 BD: 1 +.+.: (wq_completion)bond127#3 ->&rq->__lock ->(work_completion)(&(&slave->notify_work)->work) ->(work_completion)(&(&bond->mcast_work)->work) FD: 232 BD: 1 +.+.: (wq_completion)bond37#2 ->&rq->__lock ->(work_completion)(&(&slave->notify_work)->work) ->(work_completion)(&(&bond->mcast_work)->work) FD: 232 BD: 1 +.+.: (wq_completion)bond143#2 ->(work_completion)(&(&slave->notify_work)->work) ->(work_completion)(&(&bond->mcast_work)->work) ->&rq->__lock FD: 232 BD: 1 +.+.: (wq_completion)bond156 ->(work_completion)(&(&slave->notify_work)->work) ->(work_completion)(&(&bond->mcast_work)->work) ->&rq->__lock FD: 232 BD: 1 +.+.: (wq_completion)bond128#3 ->&rq->__lock ->(work_completion)(&(&slave->notify_work)->work) ->(work_completion)(&(&bond->mcast_work)->work) FD: 232 BD: 1 +.+.: (wq_completion)bond144#2 ->(work_completion)(&(&slave->notify_work)->work) ->(work_completion)(&(&bond->mcast_work)->work) ->&rq->__lock FD: 232 BD: 1 +.+.: (wq_completion)bond157 ->(work_completion)(&(&slave->notify_work)->work) ->(work_completion)(&(&bond->mcast_work)->work) ->&rq->__lock FD: 232 BD: 1 +.+.: (wq_completion)bond38#2 ->&rq->__lock ->(work_completion)(&(&slave->notify_work)->work) ->(work_completion)(&(&bond->mcast_work)->work) FD: 232 BD: 1 +.+.: (wq_completion)bond129#3 ->(work_completion)(&(&slave->notify_work)->work) ->(work_completion)(&(&bond->mcast_work)->work) FD: 232 BD: 1 +.+.: (wq_completion)bond145#2 ->(work_completion)(&(&slave->notify_work)->work) ->(work_completion)(&(&bond->mcast_work)->work) ->&rq->__lock FD: 232 BD: 1 +.+.: (wq_completion)bond158 ->&rq->__lock ->(work_completion)(&(&slave->notify_work)->work) ->(work_completion)(&(&bond->mcast_work)->work) FD: 232 BD: 1 +.+.: (wq_completion)bond39 ->(work_completion)(&(&slave->notify_work)->work) ->(work_completion)(&(&bond->mcast_work)->work) FD: 232 BD: 1 +.+.: (wq_completion)bond130#3 ->(work_completion)(&(&slave->notify_work)->work) ->(work_completion)(&(&bond->mcast_work)->work) FD: 232 BD: 1 +.+.: (wq_completion)bond146#2 ->&rq->__lock ->(work_completion)(&(&slave->notify_work)->work) ->(work_completion)(&(&bond->mcast_work)->work) FD: 232 BD: 1 +.+.: (wq_completion)bond159 ->&rq->__lock ->(work_completion)(&(&slave->notify_work)->work) ->(work_completion)(&(&bond->mcast_work)->work) FD: 232 BD: 1 +.+.: (wq_completion)bond131#2 ->(work_completion)(&(&slave->notify_work)->work) ->(work_completion)(&(&bond->mcast_work)->work) FD: 232 BD: 1 +.+.: (wq_completion)bond147#2 ->(work_completion)(&(&slave->notify_work)->work) ->(work_completion)(&(&bond->mcast_work)->work) ->&rq->__lock FD: 232 BD: 1 +.+.: (wq_completion)bond160 ->(work_completion)(&(&slave->notify_work)->work) ->(work_completion)(&(&bond->mcast_work)->work) ->&rq->__lock FD: 232 BD: 1 +.+.: (wq_completion)bond132#3 ->&rq->__lock ->(work_completion)(&(&slave->notify_work)->work) ->(work_completion)(&(&bond->mcast_work)->work) FD: 232 BD: 1 +.+.: (wq_completion)bond148#2 ->(work_completion)(&(&slave->notify_work)->work) ->(work_completion)(&(&bond->mcast_work)->work) ->&rq->__lock FD: 232 BD: 1 +.+.: (wq_completion)bond40 ->(work_completion)(&(&slave->notify_work)->work) ->(work_completion)(&(&bond->mcast_work)->work) ->&rq->__lock FD: 232 BD: 1 +.+.: (wq_completion)bond161 ->(work_completion)(&(&slave->notify_work)->work) ->(work_completion)(&(&bond->mcast_work)->work) FD: 232 BD: 1 +.+.: (wq_completion)bond133#3 ->(work_completion)(&(&slave->notify_work)->work) ->(work_completion)(&(&bond->mcast_work)->work) ->&rq->__lock FD: 232 BD: 1 +.+.: (wq_completion)bond149#2 ->(work_completion)(&(&slave->notify_work)->work) ->(work_completion)(&(&bond->mcast_work)->work) FD: 232 BD: 1 +.+.: (wq_completion)bond162 ->(work_completion)(&(&slave->notify_work)->work) ->(work_completion)(&(&bond->mcast_work)->work) FD: 232 BD: 1 +.+.: (wq_completion)bond134#3 ->(work_completion)(&(&slave->notify_work)->work) ->(work_completion)(&(&bond->mcast_work)->work) ->&rq->__lock FD: 232 BD: 1 +.+.: (wq_completion)bond150#2 ->(work_completion)(&(&slave->notify_work)->work) ->(work_completion)(&(&bond->mcast_work)->work) ->&rq->__lock FD: 232 BD: 1 +.+.: (wq_completion)bond163 ->(work_completion)(&(&slave->notify_work)->work) ->(work_completion)(&(&bond->mcast_work)->work) FD: 232 BD: 1 +.+.: (wq_completion)bond41 ->(work_completion)(&(&slave->notify_work)->work) ->(work_completion)(&(&bond->mcast_work)->work) ->&rq->__lock FD: 232 BD: 1 +.+.: (wq_completion)bond151 ->(work_completion)(&(&slave->notify_work)->work) ->(work_completion)(&(&bond->mcast_work)->work) FD: 232 BD: 1 +.+.: (wq_completion)bond135#3 ->(work_completion)(&(&slave->notify_work)->work) ->(work_completion)(&(&bond->mcast_work)->work) ->&rq->__lock FD: 232 BD: 1 +.+.: (wq_completion)bond164 ->(work_completion)(&(&slave->notify_work)->work) ->(work_completion)(&(&bond->mcast_work)->work) FD: 232 BD: 1 +.+.: (wq_completion)bond42 ->&rq->__lock ->(work_completion)(&(&slave->notify_work)->work) ->(work_completion)(&(&bond->mcast_work)->work) FD: 232 BD: 1 +.+.: (wq_completion)bond152#2 ->&rq->__lock ->(work_completion)(&(&slave->notify_work)->work) ->(work_completion)(&(&bond->mcast_work)->work) FD: 232 BD: 1 +.+.: (wq_completion)bond136#3 ->(work_completion)(&(&slave->notify_work)->work) ->(work_completion)(&(&bond->mcast_work)->work) FD: 232 BD: 1 +.+.: (wq_completion)bond165 ->(work_completion)(&(&slave->notify_work)->work) ->(work_completion)(&(&bond->mcast_work)->work) FD: 232 BD: 1 +.+.: (wq_completion)bond153#2 ->(work_completion)(&(&slave->notify_work)->work) ->(work_completion)(&(&bond->mcast_work)->work) ->&rq->__lock FD: 232 BD: 1 +.+.: (wq_completion)bond137#3 ->(work_completion)(&(&slave->notify_work)->work) ->(work_completion)(&(&bond->mcast_work)->work) ->&rq->__lock FD: 232 BD: 1 +.+.: (wq_completion)bond43#2 ->&rq->__lock ->(work_completion)(&(&slave->notify_work)->work) ->(work_completion)(&(&bond->mcast_work)->work) FD: 232 BD: 1 +.+.: (wq_completion)bond166 ->(work_completion)(&(&slave->notify_work)->work) ->(work_completion)(&(&bond->mcast_work)->work) FD: 232 BD: 1 +.+.: (wq_completion)bond154#2 ->(work_completion)(&(&slave->notify_work)->work) ->(work_completion)(&(&bond->mcast_work)->work) ->&rq->__lock FD: 232 BD: 1 +.+.: (wq_completion)bond138#2 ->(work_completion)(&(&slave->notify_work)->work) ->(work_completion)(&(&bond->mcast_work)->work) ->&rq->__lock FD: 232 BD: 1 +.+.: (wq_completion)bond167 ->(work_completion)(&(&slave->notify_work)->work) ->(work_completion)(&(&bond->mcast_work)->work) ->&rq->__lock FD: 232 BD: 1 +.+.: (wq_completion)bond155#2 ->(work_completion)(&(&slave->notify_work)->work) ->(work_completion)(&(&bond->mcast_work)->work) ->&rq->__lock FD: 232 BD: 1 +.+.: (wq_completion)bond139#3 ->(work_completion)(&(&slave->notify_work)->work) ->(work_completion)(&(&bond->mcast_work)->work) FD: 232 BD: 1 +.+.: (wq_completion)bond44#2 ->&rq->__lock ->(work_completion)(&(&slave->notify_work)->work) ->(work_completion)(&(&bond->mcast_work)->work) FD: 232 BD: 1 +.+.: (wq_completion)bond168 ->&rq->__lock ->(work_completion)(&(&slave->notify_work)->work) ->(work_completion)(&(&bond->mcast_work)->work) FD: 232 BD: 1 +.+.: (wq_completion)bond156#2 ->(work_completion)(&(&slave->notify_work)->work) ->(work_completion)(&(&bond->mcast_work)->work) ->&rq->__lock FD: 232 BD: 1 +.+.: (wq_completion)bond140#3 ->(work_completion)(&(&slave->notify_work)->work) ->(work_completion)(&(&bond->mcast_work)->work) ->&rq->__lock FD: 232 BD: 1 +.+.: (wq_completion)bond169 ->&rq->__lock ->(work_completion)(&(&slave->notify_work)->work) ->(work_completion)(&(&bond->mcast_work)->work) FD: 232 BD: 1 +.+.: (wq_completion)bond157#2 ->(work_completion)(&(&slave->notify_work)->work) ->(work_completion)(&(&bond->mcast_work)->work) ->&rq->__lock FD: 232 BD: 1 +.+.: (wq_completion)bond141#3 ->(work_completion)(&(&slave->notify_work)->work) ->(work_completion)(&(&bond->mcast_work)->work) FD: 232 BD: 1 +.+.: (wq_completion)bond170 ->(work_completion)(&(&slave->notify_work)->work) ->(work_completion)(&(&bond->mcast_work)->work) ->&rq->__lock ->&cfs_rq->removed.lock FD: 232 BD: 1 +.+.: (wq_completion)bond45#2 ->(work_completion)(&(&slave->notify_work)->work) ->(work_completion)(&(&bond->mcast_work)->work) ->&rq->__lock FD: 232 BD: 1 +.+.: (wq_completion)bond158#2 ->(work_completion)(&(&slave->notify_work)->work) ->(work_completion)(&(&bond->mcast_work)->work) ->&rq->__lock FD: 232 BD: 1 +.+.: (wq_completion)bond142#3 ->(work_completion)(&(&slave->notify_work)->work) ->(work_completion)(&(&bond->mcast_work)->work) ->&rq->__lock FD: 232 BD: 1 +.+.: (wq_completion)bond171 ->(work_completion)(&(&slave->notify_work)->work) ->(work_completion)(&(&bond->mcast_work)->work) FD: 232 BD: 1 +.+.: (wq_completion)bond46#2 ->(work_completion)(&(&slave->notify_work)->work) ->(work_completion)(&(&bond->mcast_work)->work) FD: 232 BD: 1 +.+.: (wq_completion)bond159#2 ->(work_completion)(&(&slave->notify_work)->work) ->(work_completion)(&(&bond->mcast_work)->work) ->&rq->__lock FD: 232 BD: 1 +.+.: (wq_completion)bond143#3 ->(work_completion)(&(&slave->notify_work)->work) ->(work_completion)(&(&bond->mcast_work)->work) ->&rq->__lock FD: 232 BD: 1 +.+.: (wq_completion)bond172 ->&rq->__lock ->(work_completion)(&(&slave->notify_work)->work) ->(work_completion)(&(&bond->mcast_work)->work) FD: 232 BD: 1 +.+.: (wq_completion)bond160#2 ->(work_completion)(&(&slave->notify_work)->work) ->(work_completion)(&(&bond->mcast_work)->work) ->&rq->__lock FD: 232 BD: 1 +.+.: (wq_completion)bond144#3 ->(work_completion)(&(&slave->notify_work)->work) ->(work_completion)(&(&bond->mcast_work)->work) ->&rq->__lock FD: 232 BD: 1 +.+.: (wq_completion)bond173 ->&rq->__lock ->(work_completion)(&(&slave->notify_work)->work) ->(work_completion)(&(&bond->mcast_work)->work) FD: 232 BD: 1 +.+.: (wq_completion)bond47#2 ->(work_completion)(&(&slave->notify_work)->work) ->(work_completion)(&(&bond->mcast_work)->work) FD: 232 BD: 1 +.+.: (wq_completion)bond161#2 ->(work_completion)(&(&slave->notify_work)->work) ->(work_completion)(&(&bond->mcast_work)->work) FD: 232 BD: 1 +.+.: (wq_completion)bond145#3 ->(work_completion)(&(&slave->notify_work)->work) ->(work_completion)(&(&bond->mcast_work)->work) FD: 232 BD: 1 +.+.: (wq_completion)bond174 ->(work_completion)(&(&slave->notify_work)->work) ->(work_completion)(&(&bond->mcast_work)->work) ->&rq->__lock FD: 232 BD: 1 +.+.: (wq_completion)bond162#2 ->&rq->__lock ->(work_completion)(&(&slave->notify_work)->work) ->(work_completion)(&(&bond->mcast_work)->work) FD: 232 BD: 1 +.+.: (wq_completion)bond146#3 ->&rq->__lock ->(work_completion)(&(&slave->notify_work)->work) ->(work_completion)(&(&bond->mcast_work)->work) FD: 232 BD: 1 +.+.: (wq_completion)bond48#2 ->(work_completion)(&(&slave->notify_work)->work) ->(work_completion)(&(&bond->mcast_work)->work) ->&rq->__lock FD: 232 BD: 1 +.+.: (wq_completion)bond175 ->&rq->__lock ->(work_completion)(&(&slave->notify_work)->work) ->(work_completion)(&(&bond->mcast_work)->work) FD: 232 BD: 1 +.+.: (wq_completion)bond163#2 ->(work_completion)(&(&slave->notify_work)->work) ->(work_completion)(&(&bond->mcast_work)->work) ->&rq->__lock FD: 232 BD: 1 +.+.: (wq_completion)bond147#3 ->(work_completion)(&(&slave->notify_work)->work) ->(work_completion)(&(&bond->mcast_work)->work) ->&rq->__lock FD: 232 BD: 1 +.+.: (wq_completion)bond49#2 ->(work_completion)(&(&slave->notify_work)->work) ->(work_completion)(&(&bond->mcast_work)->work) FD: 232 BD: 1 +.+.: (wq_completion)bond176 ->(work_completion)(&(&slave->notify_work)->work) ->(work_completion)(&(&bond->mcast_work)->work) FD: 232 BD: 1 +.+.: (wq_completion)bond164#2 ->(work_completion)(&(&slave->notify_work)->work) ->(work_completion)(&(&bond->mcast_work)->work) FD: 232 BD: 1 +.+.: (wq_completion)bond148#3 ->&rq->__lock ->(work_completion)(&(&slave->notify_work)->work) ->(work_completion)(&(&bond->mcast_work)->work) FD: 232 BD: 1 +.+.: (wq_completion)bond177 ->(work_completion)(&(&slave->notify_work)->work) ->(work_completion)(&(&bond->mcast_work)->work) FD: 232 BD: 1 +.+.: (wq_completion)bond50#2 ->(work_completion)(&(&slave->notify_work)->work) ->(work_completion)(&(&bond->mcast_work)->work) FD: 232 BD: 1 +.+.: (wq_completion)bond165#2 ->(work_completion)(&(&slave->notify_work)->work) ->(work_completion)(&(&bond->mcast_work)->work) ->&rq->__lock FD: 232 BD: 1 +.+.: (wq_completion)bond149#3 ->&rq->__lock ->(work_completion)(&(&slave->notify_work)->work) ->(work_completion)(&(&bond->mcast_work)->work) FD: 232 BD: 1 +.+.: (wq_completion)bond178 ->&rq->__lock ->(work_completion)(&(&slave->notify_work)->work) ->(work_completion)(&(&bond->mcast_work)->work) FD: 232 BD: 1 +.+.: (wq_completion)bond51#3 ->&rq->__lock ->(work_completion)(&(&slave->notify_work)->work) ->(work_completion)(&(&bond->mcast_work)->work) FD: 232 BD: 1 +.+.: (wq_completion)bond166#2 ->&rq->__lock ->(work_completion)(&(&slave->notify_work)->work) ->(work_completion)(&(&bond->mcast_work)->work) FD: 232 BD: 1 +.+.: (wq_completion)bond150#3 ->&rq->__lock ->(work_completion)(&(&slave->notify_work)->work) ->(work_completion)(&(&bond->mcast_work)->work) FD: 232 BD: 1 +.+.: (wq_completion)bond52#3 ->(work_completion)(&(&slave->notify_work)->work) ->(work_completion)(&(&bond->mcast_work)->work) ->&rq->__lock FD: 232 BD: 1 +.+.: (wq_completion)bond179 ->(work_completion)(&(&slave->notify_work)->work) ->(work_completion)(&(&bond->mcast_work)->work) FD: 232 BD: 1 +.+.: (wq_completion)bond167#2 ->(work_completion)(&(&slave->notify_work)->work) ->(work_completion)(&(&bond->mcast_work)->work) ->&rq->__lock FD: 232 BD: 1 +.+.: (wq_completion)bond151#2 ->&rq->__lock ->(work_completion)(&(&slave->notify_work)->work) ->(work_completion)(&(&bond->mcast_work)->work) FD: 232 BD: 1 +.+.: (wq_completion)bond180 ->&rq->__lock ->(work_completion)(&(&slave->notify_work)->work) ->(work_completion)(&(&bond->mcast_work)->work) FD: 232 BD: 1 +.+.: (wq_completion)bond53#3 ->(work_completion)(&(&slave->notify_work)->work) ->(work_completion)(&(&bond->mcast_work)->work) ->&rq->__lock FD: 232 BD: 1 +.+.: (wq_completion)bond168#2 ->(work_completion)(&(&slave->notify_work)->work) ->(work_completion)(&(&bond->mcast_work)->work) ->&rq->__lock FD: 232 BD: 1 +.+.: (wq_completion)bond152#3 ->(work_completion)(&(&slave->notify_work)->work) ->(work_completion)(&(&bond->mcast_work)->work) FD: 232 BD: 1 +.+.: (wq_completion)bond181 ->(work_completion)(&(&slave->notify_work)->work) ->(work_completion)(&(&bond->mcast_work)->work) FD: 232 BD: 1 +.+.: (wq_completion)bond169#2 ->&rq->__lock ->(work_completion)(&(&slave->notify_work)->work) ->(work_completion)(&(&bond->mcast_work)->work) FD: 232 BD: 1 +.+.: (wq_completion)bond153#3 ->&rq->__lock ->(work_completion)(&(&slave->notify_work)->work) ->(work_completion)(&(&bond->mcast_work)->work) FD: 232 BD: 1 +.+.: (wq_completion)bond54#3 ->(work_completion)(&(&slave->notify_work)->work) ->(work_completion)(&(&bond->mcast_work)->work) ->&rq->__lock FD: 232 BD: 1 +.+.: (wq_completion)bond182 ->(work_completion)(&(&slave->notify_work)->work) ->(work_completion)(&(&bond->mcast_work)->work) FD: 232 BD: 1 +.+.: (wq_completion)bond170#2 ->&rq->__lock ->(work_completion)(&(&slave->notify_work)->work) ->(work_completion)(&(&bond->mcast_work)->work) FD: 232 BD: 1 +.+.: (wq_completion)bond154#3 ->(work_completion)(&(&slave->notify_work)->work) ->(work_completion)(&(&bond->mcast_work)->work) FD: 232 BD: 1 +.+.: (wq_completion)bond183 ->(work_completion)(&(&slave->notify_work)->work) ->(work_completion)(&(&bond->mcast_work)->work) FD: 232 BD: 1 +.+.: (wq_completion)bond171#2 ->(work_completion)(&(&slave->notify_work)->work) ->(work_completion)(&(&bond->mcast_work)->work) ->&rq->__lock FD: 232 BD: 1 +.+.: (wq_completion)bond155#3 ->(work_completion)(&(&slave->notify_work)->work) ->(work_completion)(&(&bond->mcast_work)->work) ->&rq->__lock FD: 232 BD: 1 +.+.: (wq_completion)bond184 ->(work_completion)(&(&slave->notify_work)->work) ->(work_completion)(&(&bond->mcast_work)->work) ->&rq->__lock FD: 232 BD: 1 +.+.: (wq_completion)bond172#2 ->(work_completion)(&(&slave->notify_work)->work) ->(work_completion)(&(&bond->mcast_work)->work) FD: 232 BD: 1 +.+.: (wq_completion)bond156#3 ->(work_completion)(&(&slave->notify_work)->work) ->(work_completion)(&(&bond->mcast_work)->work) FD: 232 BD: 1 +.+.: (wq_completion)bond185 ->(work_completion)(&(&slave->notify_work)->work) ->(work_completion)(&(&bond->mcast_work)->work) ->&rq->__lock FD: 232 BD: 1 +.+.: (wq_completion)bond55#2 ->(work_completion)(&(&slave->notify_work)->work) ->(work_completion)(&(&bond->mcast_work)->work) FD: 232 BD: 1 +.+.: (wq_completion)bond173#2 ->(work_completion)(&(&slave->notify_work)->work) ->(work_completion)(&(&bond->mcast_work)->work) FD: 232 BD: 1 +.+.: (wq_completion)bond157#3 ->(work_completion)(&(&slave->notify_work)->work) ->(work_completion)(&(&bond->mcast_work)->work) ->&rq->__lock FD: 232 BD: 1 +.+.: (wq_completion)bond186 ->(work_completion)(&(&slave->notify_work)->work) ->(work_completion)(&(&bond->mcast_work)->work) FD: 232 BD: 1 +.+.: (wq_completion)bond174#2 ->(work_completion)(&(&slave->notify_work)->work) ->(work_completion)(&(&bond->mcast_work)->work) ->&rq->__lock FD: 232 BD: 1 +.+.: (wq_completion)bond158#3 ->(work_completion)(&(&slave->notify_work)->work) ->(work_completion)(&(&bond->mcast_work)->work) FD: 232 BD: 1 +.+.: (wq_completion)bond187 ->(work_completion)(&(&slave->notify_work)->work) ->(work_completion)(&(&bond->mcast_work)->work) FD: 232 BD: 1 +.+.: (wq_completion)bond159#3 ->(work_completion)(&(&slave->notify_work)->work) ->(work_completion)(&(&bond->mcast_work)->work) FD: 232 BD: 1 +.+.: (wq_completion)bond188 ->(work_completion)(&(&slave->notify_work)->work) ->(work_completion)(&(&bond->mcast_work)->work) ->&rq->__lock FD: 232 BD: 1 +.+.: (wq_completion)bond176#2 ->(work_completion)(&(&slave->notify_work)->work) ->(work_completion)(&(&bond->mcast_work)->work) ->&rq->__lock FD: 232 BD: 1 +.+.: (wq_completion)bond160#3 ->(work_completion)(&(&slave->notify_work)->work) ->(work_completion)(&(&bond->mcast_work)->work) ->&rq->__lock FD: 232 BD: 1 +.+.: (wq_completion)bond189 ->(work_completion)(&(&slave->notify_work)->work) ->(work_completion)(&(&bond->mcast_work)->work) ->&rq->__lock FD: 232 BD: 1 +.+.: (wq_completion)bond56#2 ->(work_completion)(&(&slave->notify_work)->work) ->(work_completion)(&(&bond->mcast_work)->work) FD: 232 BD: 1 +.+.: (wq_completion)bond177#2 ->(work_completion)(&(&slave->notify_work)->work) ->(work_completion)(&(&bond->mcast_work)->work) FD: 232 BD: 1 +.+.: (wq_completion)bond161#3 ->(work_completion)(&(&slave->notify_work)->work) ->(work_completion)(&(&bond->mcast_work)->work) ->&rq->__lock FD: 232 BD: 1 +.+.: (wq_completion)bond190 ->(work_completion)(&(&slave->notify_work)->work) ->(work_completion)(&(&bond->mcast_work)->work) ->&rq->__lock FD: 232 BD: 1 +.+.: (wq_completion)bond57#2 ->&rq->__lock ->(work_completion)(&(&slave->notify_work)->work) ->(work_completion)(&(&bond->mcast_work)->work) FD: 232 BD: 1 +.+.: (wq_completion)bond178#2 ->(work_completion)(&(&slave->notify_work)->work) ->(work_completion)(&(&bond->mcast_work)->work) ->&rq->__lock FD: 232 BD: 1 +.+.: (wq_completion)bond162#3 ->(work_completion)(&(&slave->notify_work)->work) ->(work_completion)(&(&bond->mcast_work)->work) FD: 232 BD: 1 +.+.: (wq_completion)bond191 ->(work_completion)(&(&slave->notify_work)->work) ->(work_completion)(&(&bond->mcast_work)->work) ->&rq->__lock FD: 232 BD: 1 +.+.: (wq_completion)bond179#2 ->&rq->__lock ->(work_completion)(&(&slave->notify_work)->work) ->(work_completion)(&(&bond->mcast_work)->work) FD: 232 BD: 1 +.+.: (wq_completion)bond163#3 ->(work_completion)(&(&slave->notify_work)->work) ->(work_completion)(&(&bond->mcast_work)->work) ->&rq->__lock FD: 232 BD: 1 +.+.: (wq_completion)bond192 ->(work_completion)(&(&slave->notify_work)->work) ->(work_completion)(&(&bond->mcast_work)->work) FD: 232 BD: 1 +.+.: (wq_completion)bond58#2 ->(work_completion)(&(&slave->notify_work)->work) ->(work_completion)(&(&bond->mcast_work)->work) FD: 232 BD: 1 +.+.: (wq_completion)bond180#2 ->(work_completion)(&(&slave->notify_work)->work) ->(work_completion)(&(&bond->mcast_work)->work) FD: 232 BD: 1 +.+.: (wq_completion)bond164#3 ->&rq->__lock ->(work_completion)(&(&slave->notify_work)->work) ->(work_completion)(&(&bond->mcast_work)->work) FD: 232 BD: 1 +.+.: (wq_completion)bond193 ->(work_completion)(&(&slave->notify_work)->work) ->(work_completion)(&(&bond->mcast_work)->work) FD: 232 BD: 1 +.+.: (wq_completion)bond181#2 ->(work_completion)(&(&slave->notify_work)->work) ->(work_completion)(&(&bond->mcast_work)->work) FD: 232 BD: 1 +.+.: (wq_completion)bond165#3 ->(work_completion)(&(&slave->notify_work)->work) ->(work_completion)(&(&bond->mcast_work)->work) FD: 232 BD: 1 +.+.: (wq_completion)bond194 ->(work_completion)(&(&slave->notify_work)->work) ->(work_completion)(&(&bond->mcast_work)->work) ->&rq->__lock FD: 232 BD: 1 +.+.: (wq_completion)bond182#2 ->(work_completion)(&(&slave->notify_work)->work) ->(work_completion)(&(&bond->mcast_work)->work) FD: 232 BD: 1 +.+.: (wq_completion)bond166#3 ->(work_completion)(&(&slave->notify_work)->work) ->(work_completion)(&(&bond->mcast_work)->work) FD: 232 BD: 1 +.+.: (wq_completion)bond183#2 ->&rq->__lock ->(work_completion)(&(&slave->notify_work)->work) ->(work_completion)(&(&bond->mcast_work)->work) FD: 232 BD: 1 +.+.: (wq_completion)bond195 ->(work_completion)(&(&slave->notify_work)->work) ->(work_completion)(&(&bond->mcast_work)->work) ->&rq->__lock FD: 232 BD: 1 +.+.: (wq_completion)bond167#3 ->(work_completion)(&(&slave->notify_work)->work) ->(work_completion)(&(&bond->mcast_work)->work) FD: 232 BD: 1 +.+.: (wq_completion)bond196 ->(work_completion)(&(&slave->notify_work)->work) ->(work_completion)(&(&bond->mcast_work)->work) FD: 232 BD: 1 +.+.: (wq_completion)bond184#2 ->(work_completion)(&(&slave->notify_work)->work) ->(work_completion)(&(&bond->mcast_work)->work) ->&rq->__lock FD: 232 BD: 1 +.+.: (wq_completion)bond168#3 ->(work_completion)(&(&slave->notify_work)->work) ->(work_completion)(&(&bond->mcast_work)->work) FD: 232 BD: 1 +.+.: (wq_completion)bond197 ->&rq->__lock ->(work_completion)(&(&slave->notify_work)->work) ->(work_completion)(&(&bond->mcast_work)->work) FD: 232 BD: 1 +.+.: (wq_completion)bond185#2 ->(work_completion)(&(&slave->notify_work)->work) ->(work_completion)(&(&bond->mcast_work)->work) FD: 232 BD: 1 +.+.: (wq_completion)bond169#3 ->(work_completion)(&(&slave->notify_work)->work) ->(work_completion)(&(&bond->mcast_work)->work) ->&rq->__lock FD: 232 BD: 1 +.+.: (wq_completion)bond198 ->(work_completion)(&(&slave->notify_work)->work) ->(work_completion)(&(&bond->mcast_work)->work) FD: 232 BD: 1 +.+.: (wq_completion)bond186#2 ->(work_completion)(&(&slave->notify_work)->work) ->(work_completion)(&(&bond->mcast_work)->work) ->&rq->__lock FD: 232 BD: 1 +.+.: (wq_completion)bond170#3 ->(work_completion)(&(&slave->notify_work)->work) ->(work_completion)(&(&bond->mcast_work)->work) FD: 232 BD: 1 +.+.: (wq_completion)bond199 ->&rq->__lock ->(work_completion)(&(&slave->notify_work)->work) ->(work_completion)(&(&bond->mcast_work)->work) FD: 232 BD: 1 +.+.: (wq_completion)bond59#2 ->(work_completion)(&(&slave->notify_work)->work) ->(work_completion)(&(&bond->mcast_work)->work) ->&rq->__lock FD: 232 BD: 1 +.+.: (wq_completion)bond187#2 ->(work_completion)(&(&slave->notify_work)->work) ->(work_completion)(&(&bond->mcast_work)->work) ->&rq->__lock FD: 232 BD: 1 +.+.: (wq_completion)bond171#3 ->(work_completion)(&(&slave->notify_work)->work) ->(work_completion)(&(&bond->mcast_work)->work) FD: 232 BD: 1 +.+.: (wq_completion)bond200 ->&rq->__lock ->(work_completion)(&(&slave->notify_work)->work) ->(work_completion)(&(&bond->mcast_work)->work) FD: 232 BD: 1 +.+.: (wq_completion)bond188#2 ->(work_completion)(&(&slave->notify_work)->work) ->(work_completion)(&(&bond->mcast_work)->work) ->&rq->__lock FD: 232 BD: 1 +.+.: (wq_completion)bond172#3 ->(work_completion)(&(&slave->notify_work)->work) ->(work_completion)(&(&bond->mcast_work)->work) FD: 232 BD: 1 +.+.: (wq_completion)bond201 ->(work_completion)(&(&slave->notify_work)->work) ->(work_completion)(&(&bond->mcast_work)->work) ->&rq->__lock FD: 232 BD: 1 +.+.: (wq_completion)bond189#2 ->&rq->__lock ->(work_completion)(&(&slave->notify_work)->work) ->(work_completion)(&(&bond->mcast_work)->work) FD: 232 BD: 1 +.+.: (wq_completion)bond173#3 ->(work_completion)(&(&slave->notify_work)->work) ->(work_completion)(&(&bond->mcast_work)->work) FD: 232 BD: 1 +.+.: (wq_completion)bond202 ->(work_completion)(&(&slave->notify_work)->work) ->(work_completion)(&(&bond->mcast_work)->work) ->&rq->__lock FD: 232 BD: 1 +.+.: (wq_completion)bond190#2 ->&rq->__lock ->(work_completion)(&(&slave->notify_work)->work) ->(work_completion)(&(&bond->mcast_work)->work) FD: 232 BD: 1 +.+.: (wq_completion)bond174#3 ->(work_completion)(&(&slave->notify_work)->work) ->(work_completion)(&(&bond->mcast_work)->work) ->&rq->__lock FD: 232 BD: 1 +.+.: (wq_completion)bond203 ->(work_completion)(&(&slave->notify_work)->work) ->(work_completion)(&(&bond->mcast_work)->work) ->&rq->__lock FD: 232 BD: 1 +.+.: (wq_completion)bond191#2 ->(work_completion)(&(&slave->notify_work)->work) ->(work_completion)(&(&bond->mcast_work)->work) ->&rq->__lock FD: 232 BD: 1 +.+.: (wq_completion)bond175#2 ->&rq->__lock ->(work_completion)(&(&slave->notify_work)->work) ->(work_completion)(&(&bond->mcast_work)->work) FD: 232 BD: 1 +.+.: (wq_completion)bond204 ->(work_completion)(&(&slave->notify_work)->work) ->(work_completion)(&(&bond->mcast_work)->work) ->&rq->__lock FD: 232 BD: 1 +.+.: (wq_completion)bond192#2 ->(work_completion)(&(&slave->notify_work)->work) ->(work_completion)(&(&bond->mcast_work)->work) FD: 232 BD: 1 +.+.: (wq_completion)bond176#3 ->&rq->__lock ->(work_completion)(&(&slave->notify_work)->work) ->(work_completion)(&(&bond->mcast_work)->work) FD: 232 BD: 1 +.+.: (wq_completion)bond177#3 ->&rq->__lock ->(work_completion)(&(&slave->notify_work)->work) ->(work_completion)(&(&bond->mcast_work)->work) FD: 232 BD: 1 +.+.: (wq_completion)bond193#2 ->(work_completion)(&(&slave->notify_work)->work) ->(work_completion)(&(&bond->mcast_work)->work) FD: 232 BD: 1 +.+.: (wq_completion)bond205 ->(work_completion)(&(&slave->notify_work)->work) ->(work_completion)(&(&bond->mcast_work)->work) FD: 232 BD: 1 +.+.: (wq_completion)bond194#2 ->&rq->__lock ->(work_completion)(&(&slave->notify_work)->work) ->(work_completion)(&(&bond->mcast_work)->work) FD: 232 BD: 1 +.+.: (wq_completion)bond178#3 ->(work_completion)(&(&slave->notify_work)->work) ->(work_completion)(&(&bond->mcast_work)->work) FD: 232 BD: 1 +.+.: (wq_completion)bond206 ->(work_completion)(&(&slave->notify_work)->work) ->(work_completion)(&(&bond->mcast_work)->work) ->&rq->__lock FD: 232 BD: 1 +.+.: (wq_completion)bond195#2 ->&rq->__lock ->(work_completion)(&(&slave->notify_work)->work) ->(work_completion)(&(&bond->mcast_work)->work) FD: 232 BD: 1 +.+.: (wq_completion)bond179#3 ->&rq->__lock ->(work_completion)(&(&slave->notify_work)->work) ->(work_completion)(&(&bond->mcast_work)->work) FD: 232 BD: 1 +.+.: (wq_completion)bond207 ->(work_completion)(&(&slave->notify_work)->work) ->(work_completion)(&(&bond->mcast_work)->work) ->&rq->__lock FD: 232 BD: 1 +.+.: (wq_completion)bond180#3 ->&rq->__lock ->(work_completion)(&(&slave->notify_work)->work) ->(work_completion)(&(&bond->mcast_work)->work) FD: 232 BD: 1 +.+.: (wq_completion)bond208 ->&rq->__lock ->(work_completion)(&(&slave->notify_work)->work) ->(work_completion)(&(&bond->mcast_work)->work) FD: 232 BD: 1 +.+.: (wq_completion)bond197#2 ->&rq->__lock ->(work_completion)(&(&slave->notify_work)->work) ->(work_completion)(&(&bond->mcast_work)->work) FD: 232 BD: 1 +.+.: (wq_completion)bond209 ->&rq->__lock ->(work_completion)(&(&slave->notify_work)->work) ->(work_completion)(&(&bond->mcast_work)->work) FD: 232 BD: 1 +.+.: (wq_completion)bond198#2 ->(work_completion)(&(&slave->notify_work)->work) ->(work_completion)(&(&bond->mcast_work)->work) ->&rq->__lock FD: 232 BD: 1 +.+.: (wq_completion)bond182#3 ->(work_completion)(&(&slave->notify_work)->work) ->(work_completion)(&(&bond->mcast_work)->work) FD: 232 BD: 1 +.+.: (wq_completion)bond210 ->(work_completion)(&(&slave->notify_work)->work) ->(work_completion)(&(&bond->mcast_work)->work) FD: 232 BD: 1 +.+.: (wq_completion)bond199#2 ->(work_completion)(&(&slave->notify_work)->work) ->(work_completion)(&(&bond->mcast_work)->work) FD: 232 BD: 1 +.+.: (wq_completion)bond211 ->(work_completion)(&(&slave->notify_work)->work) ->(work_completion)(&(&bond->mcast_work)->work) FD: 232 BD: 1 +.+.: (wq_completion)bond183#3 ->(work_completion)(&(&slave->notify_work)->work) ->(work_completion)(&(&bond->mcast_work)->work) FD: 232 BD: 1 +.+.: (wq_completion)bond200#2 ->&rq->__lock ->(work_completion)(&(&slave->notify_work)->work) ->(work_completion)(&(&bond->mcast_work)->work) FD: 232 BD: 1 +.+.: (wq_completion)bond212 ->(work_completion)(&(&slave->notify_work)->work) ->(work_completion)(&(&bond->mcast_work)->work) FD: 232 BD: 1 +.+.: (wq_completion)bond60#2 ->&rq->__lock ->(work_completion)(&(&slave->notify_work)->work) ->(work_completion)(&(&bond->mcast_work)->work) FD: 232 BD: 1 +.+.: (wq_completion)bond184#3 ->(work_completion)(&(&slave->notify_work)->work) ->(work_completion)(&(&bond->mcast_work)->work) ->&rq->__lock FD: 232 BD: 1 +.+.: (wq_completion)bond201#2 ->(work_completion)(&(&slave->notify_work)->work) ->(work_completion)(&(&bond->mcast_work)->work) FD: 232 BD: 1 +.+.: (wq_completion)bond213 ->(work_completion)(&(&slave->notify_work)->work) ->(work_completion)(&(&bond->mcast_work)->work) FD: 232 BD: 1 +.+.: (wq_completion)bond185#3 ->(work_completion)(&(&slave->notify_work)->work) ->(work_completion)(&(&bond->mcast_work)->work) FD: 232 BD: 1 +.+.: (wq_completion)bond202#2 ->(work_completion)(&(&slave->notify_work)->work) ->(work_completion)(&(&bond->mcast_work)->work) FD: 232 BD: 1 +.+.: (wq_completion)bond214 ->(work_completion)(&(&slave->notify_work)->work) ->(work_completion)(&(&bond->mcast_work)->work) ->&rq->__lock FD: 232 BD: 1 +.+.: (wq_completion)bond203#2 ->(work_completion)(&(&slave->notify_work)->work) ->(work_completion)(&(&bond->mcast_work)->work) FD: 232 BD: 1 +.+.: (wq_completion)bond187#3 ->(work_completion)(&(&slave->notify_work)->work) ->(work_completion)(&(&bond->mcast_work)->work) FD: 232 BD: 1 +.+.: (wq_completion)bond204#2 ->(work_completion)(&(&slave->notify_work)->work) ->(work_completion)(&(&bond->mcast_work)->work) FD: 232 BD: 1 +.+.: (wq_completion)bond215 ->&rq->__lock ->(work_completion)(&(&slave->notify_work)->work) ->(work_completion)(&(&bond->mcast_work)->work) FD: 232 BD: 1 +.+.: (wq_completion)bond188#3 ->(work_completion)(&(&slave->notify_work)->work) ->(work_completion)(&(&bond->mcast_work)->work) FD: 232 BD: 1 +.+.: (wq_completion)bond205#2 ->(work_completion)(&(&slave->notify_work)->work) ->(work_completion)(&(&bond->mcast_work)->work) FD: 232 BD: 1 +.+.: (wq_completion)bond216 ->&rq->__lock ->(work_completion)(&(&slave->notify_work)->work) ->(work_completion)(&(&bond->mcast_work)->work) FD: 232 BD: 1 +.+.: (wq_completion)bond61 ->(work_completion)(&(&slave->notify_work)->work) ->(work_completion)(&(&bond->mcast_work)->work) FD: 232 BD: 1 +.+.: (wq_completion)bond189#3 ->(work_completion)(&(&slave->notify_work)->work) ->(work_completion)(&(&bond->mcast_work)->work) FD: 232 BD: 1 +.+.: (wq_completion)bond206#2 ->(work_completion)(&(&slave->notify_work)->work) ->(work_completion)(&(&bond->mcast_work)->work) FD: 232 BD: 1 +.+.: (wq_completion)bond217 ->(work_completion)(&(&slave->notify_work)->work) ->(work_completion)(&(&bond->mcast_work)->work) ->&rq->__lock FD: 232 BD: 1 +.+.: (wq_completion)bond62#3 ->(work_completion)(&(&slave->notify_work)->work) ->(work_completion)(&(&bond->mcast_work)->work) FD: 232 BD: 1 +.+.: (wq_completion)bond190#3 ->(work_completion)(&(&slave->notify_work)->work) ->(work_completion)(&(&bond->mcast_work)->work) FD: 232 BD: 1 +.+.: (wq_completion)bond207#2 ->&rq->__lock ->(work_completion)(&(&slave->notify_work)->work) ->(work_completion)(&(&bond->mcast_work)->work) FD: 232 BD: 1 +.+.: (wq_completion)bond218 ->(work_completion)(&(&slave->notify_work)->work) ->(work_completion)(&(&bond->mcast_work)->work) ->&rq->__lock FD: 232 BD: 1 +.+.: (wq_completion)bond191#3 ->(work_completion)(&(&slave->notify_work)->work) ->(work_completion)(&(&bond->mcast_work)->work) FD: 232 BD: 1 +.+.: (wq_completion)bond219 ->(work_completion)(&(&slave->notify_work)->work) ->(work_completion)(&(&bond->mcast_work)->work) FD: 232 BD: 1 +.+.: (wq_completion)bond63 ->(work_completion)(&(&slave->notify_work)->work) ->(work_completion)(&(&bond->mcast_work)->work) FD: 232 BD: 1 +.+.: (wq_completion)bond208#2 ->(work_completion)(&(&slave->notify_work)->work) ->(work_completion)(&(&bond->mcast_work)->work) ->&rq->__lock FD: 232 BD: 1 +.+.: (wq_completion)bond192#3 ->(work_completion)(&(&slave->notify_work)->work) ->(work_completion)(&(&bond->mcast_work)->work) ->&rq->__lock FD: 232 BD: 1 +.+.: (wq_completion)bond220 ->(work_completion)(&(&slave->notify_work)->work) ->(work_completion)(&(&bond->mcast_work)->work) FD: 232 BD: 1 +.+.: (wq_completion)bond209#2 ->(work_completion)(&(&slave->notify_work)->work) ->(work_completion)(&(&bond->mcast_work)->work) ->&rq->__lock FD: 232 BD: 1 +.+.: (wq_completion)bond64 ->&rq->__lock ->(work_completion)(&(&slave->notify_work)->work) ->(work_completion)(&(&bond->mcast_work)->work) FD: 232 BD: 1 +.+.: (wq_completion)bond193#3 ->(work_completion)(&(&slave->notify_work)->work) ->(work_completion)(&(&bond->mcast_work)->work) ->&rq->__lock FD: 232 BD: 1 +.+.: (wq_completion)bond221 ->&rq->__lock ->(work_completion)(&(&slave->notify_work)->work) ->(work_completion)(&(&bond->mcast_work)->work) FD: 232 BD: 1 +.+.: (wq_completion)bond210#2 ->(work_completion)(&(&slave->notify_work)->work) ->(work_completion)(&(&bond->mcast_work)->work) ->&rq->__lock FD: 232 BD: 1 +.+.: (wq_completion)bond65 ->(work_completion)(&(&slave->notify_work)->work) ->(work_completion)(&(&bond->mcast_work)->work) ->&rq->__lock FD: 232 BD: 1 +.+.: (wq_completion)bond194#3 ->(work_completion)(&(&slave->notify_work)->work) ->(work_completion)(&(&bond->mcast_work)->work) FD: 232 BD: 1 +.+.: (wq_completion)bond222 ->(work_completion)(&(&slave->notify_work)->work) ->(work_completion)(&(&bond->mcast_work)->work) FD: 232 BD: 1 +.+.: (wq_completion)bond211#2 ->&rq->__lock ->(work_completion)(&(&slave->notify_work)->work) ->(work_completion)(&(&bond->mcast_work)->work) FD: 232 BD: 1 +.+.: (wq_completion)bond66 ->&rq->__lock ->(work_completion)(&(&slave->notify_work)->work) ->(work_completion)(&(&bond->mcast_work)->work) FD: 232 BD: 1 +.+.: (wq_completion)bond195#3 ->&rq->__lock ->(work_completion)(&(&slave->notify_work)->work) ->(work_completion)(&(&bond->mcast_work)->work) FD: 232 BD: 1 +.+.: (wq_completion)bond223 ->(work_completion)(&(&slave->notify_work)->work) ->(work_completion)(&(&bond->mcast_work)->work) FD: 232 BD: 1 +.+.: (wq_completion)bond212#2 ->&rq->__lock ->(work_completion)(&(&slave->notify_work)->work) ->(work_completion)(&(&bond->mcast_work)->work) FD: 232 BD: 1 +.+.: (wq_completion)bond196#2 ->(work_completion)(&(&slave->notify_work)->work) ->(work_completion)(&(&bond->mcast_work)->work) FD: 232 BD: 1 +.+.: (wq_completion)bond224 ->&rq->__lock ->(work_completion)(&(&slave->notify_work)->work) ->(work_completion)(&(&bond->mcast_work)->work) FD: 232 BD: 1 +.+.: (wq_completion)bond67#2 ->(work_completion)(&(&slave->notify_work)->work) ->(work_completion)(&(&bond->mcast_work)->work) ->&rq->__lock FD: 232 BD: 1 +.+.: (wq_completion)bond197#3 ->(work_completion)(&(&slave->notify_work)->work) ->(work_completion)(&(&bond->mcast_work)->work) FD: 232 BD: 1 +.+.: (wq_completion)bond214#2 ->(work_completion)(&(&slave->notify_work)->work) ->(work_completion)(&(&bond->mcast_work)->work) FD: 232 BD: 1 +.+.: (wq_completion)bond225 ->(work_completion)(&(&slave->notify_work)->work) ->(work_completion)(&(&bond->mcast_work)->work) FD: 232 BD: 1 +.+.: (wq_completion)bond198#3 ->&rq->__lock ->(work_completion)(&(&slave->notify_work)->work) ->(work_completion)(&(&bond->mcast_work)->work) FD: 232 BD: 1 +.+.: (wq_completion)bond226 ->(work_completion)(&(&slave->notify_work)->work) ->(work_completion)(&(&bond->mcast_work)->work) FD: 232 BD: 1 +.+.: (wq_completion)bond215#2 ->(work_completion)(&(&slave->notify_work)->work) ->(work_completion)(&(&bond->mcast_work)->work) FD: 232 BD: 1 +.+.: (wq_completion)bond199#3 ->(work_completion)(&(&slave->notify_work)->work) ->(work_completion)(&(&bond->mcast_work)->work) FD: 232 BD: 1 +.+.: (wq_completion)bond227 ->(work_completion)(&(&slave->notify_work)->work) ->(work_completion)(&(&bond->mcast_work)->work) FD: 232 BD: 1 +.+.: (wq_completion)bond216#2 ->&rq->__lock ->(work_completion)(&(&slave->notify_work)->work) ->(work_completion)(&(&bond->mcast_work)->work) FD: 232 BD: 1 +.+.: (wq_completion)bond200#3 ->(work_completion)(&(&slave->notify_work)->work) ->(work_completion)(&(&bond->mcast_work)->work) FD: 232 BD: 1 +.+.: (wq_completion)bond228 ->(work_completion)(&(&slave->notify_work)->work) ->(work_completion)(&(&bond->mcast_work)->work) FD: 232 BD: 1 +.+.: (wq_completion)bond217#2 ->&rq->__lock ->(work_completion)(&(&slave->notify_work)->work) ->(work_completion)(&(&bond->mcast_work)->work) FD: 232 BD: 1 +.+.: (wq_completion)bond68#2 ->(work_completion)(&(&slave->notify_work)->work) ->(work_completion)(&(&bond->mcast_work)->work) ->&rq->__lock FD: 232 BD: 1 +.+.: (wq_completion)bond201#3 ->(work_completion)(&(&slave->notify_work)->work) ->(work_completion)(&(&bond->mcast_work)->work) ->&rq->__lock FD: 232 BD: 1 +.+.: (wq_completion)bond229 ->(work_completion)(&(&slave->notify_work)->work) ->(work_completion)(&(&bond->mcast_work)->work) FD: 232 BD: 1 +.+.: (wq_completion)bond218#2 ->&rq->__lock ->(work_completion)(&(&slave->notify_work)->work) ->(work_completion)(&(&bond->mcast_work)->work) FD: 232 BD: 1 +.+.: (wq_completion)bond202#3 ->(work_completion)(&(&slave->notify_work)->work) ->(work_completion)(&(&bond->mcast_work)->work) FD: 232 BD: 1 +.+.: (wq_completion)bond230 ->(work_completion)(&(&slave->notify_work)->work) ->(work_completion)(&(&bond->mcast_work)->work) ->&rq->__lock FD: 232 BD: 1 +.+.: (wq_completion)bond69#2 ->(work_completion)(&(&slave->notify_work)->work) ->(work_completion)(&(&bond->mcast_work)->work) FD: 232 BD: 1 +.+.: (wq_completion)bond219#2 ->(work_completion)(&(&slave->notify_work)->work) ->(work_completion)(&(&bond->mcast_work)->work) FD: 232 BD: 1 +.+.: (wq_completion)bond203#3 ->(work_completion)(&(&slave->notify_work)->work) ->(work_completion)(&(&bond->mcast_work)->work) FD: 232 BD: 1 +.+.: (wq_completion)bond231 ->&rq->__lock ->(work_completion)(&(&slave->notify_work)->work) ->(work_completion)(&(&bond->mcast_work)->work) FD: 232 BD: 1 +.+.: (wq_completion)bond220#2 ->&rq->__lock ->(work_completion)(&(&slave->notify_work)->work) ->(work_completion)(&(&bond->mcast_work)->work) FD: 232 BD: 1 +.+.: (wq_completion)bond204#3 ->(work_completion)(&(&slave->notify_work)->work) ->(work_completion)(&(&bond->mcast_work)->work) FD: 232 BD: 1 +.+.: (wq_completion)bond232 ->(work_completion)(&(&slave->notify_work)->work) ->(work_completion)(&(&bond->mcast_work)->work) FD: 232 BD: 1 +.+.: (wq_completion)bond221#2 ->&rq->__lock ->(work_completion)(&(&slave->notify_work)->work) ->(work_completion)(&(&bond->mcast_work)->work) FD: 232 BD: 1 +.+.: (wq_completion)bond205#3 ->(work_completion)(&(&slave->notify_work)->work) ->(work_completion)(&(&bond->mcast_work)->work) ->&rq->__lock FD: 232 BD: 1 +.+.: (wq_completion)bond70#2 ->&rq->__lock ->(work_completion)(&(&slave->notify_work)->work) ->(work_completion)(&(&bond->mcast_work)->work) FD: 232 BD: 1 +.+.: (wq_completion)bond233 ->(work_completion)(&(&slave->notify_work)->work) ->(work_completion)(&(&bond->mcast_work)->work) FD: 232 BD: 1 +.+.: (wq_completion)bond222#2 ->&rq->__lock ->(work_completion)(&(&slave->notify_work)->work) ->(work_completion)(&(&bond->mcast_work)->work) FD: 232 BD: 1 +.+.: (wq_completion)bond206#3 ->(work_completion)(&(&slave->notify_work)->work) ->(work_completion)(&(&bond->mcast_work)->work) FD: 232 BD: 1 +.+.: (wq_completion)bond71#2 ->&rq->__lock ->(work_completion)(&(&slave->notify_work)->work) ->(work_completion)(&(&bond->mcast_work)->work) FD: 232 BD: 1 +.+.: (wq_completion)bond234 ->(work_completion)(&(&slave->notify_work)->work) ->(work_completion)(&(&bond->mcast_work)->work) ->&rq->__lock FD: 232 BD: 1 +.+.: (wq_completion)bond223#2 ->(work_completion)(&(&slave->notify_work)->work) ->(work_completion)(&(&bond->mcast_work)->work) FD: 232 BD: 1 +.+.: (wq_completion)bond207#3 ->(work_completion)(&(&slave->notify_work)->work) ->(work_completion)(&(&bond->mcast_work)->work) FD: 232 BD: 1 +.+.: (wq_completion)bond72#2 ->&rq->__lock ->(work_completion)(&(&slave->notify_work)->work) ->(work_completion)(&(&bond->mcast_work)->work) FD: 232 BD: 1 +.+.: (wq_completion)bond235 ->(work_completion)(&(&slave->notify_work)->work) ->(work_completion)(&(&bond->mcast_work)->work) FD: 232 BD: 1 +.+.: (wq_completion)bond224#2 ->(work_completion)(&(&slave->notify_work)->work) ->(work_completion)(&(&bond->mcast_work)->work) ->&rq->__lock FD: 232 BD: 1 +.+.: (wq_completion)bond208#3 ->(work_completion)(&(&slave->notify_work)->work) ->(work_completion)(&(&bond->mcast_work)->work) FD: 232 BD: 1 +.+.: (wq_completion)bond236 ->(work_completion)(&(&slave->notify_work)->work) ->(work_completion)(&(&bond->mcast_work)->work) FD: 232 BD: 1 +.+.: (wq_completion)bond73#2 ->(work_completion)(&(&slave->notify_work)->work) ->(work_completion)(&(&bond->mcast_work)->work) FD: 232 BD: 1 +.+.: (wq_completion)bond225#2 ->(work_completion)(&(&slave->notify_work)->work) ->(work_completion)(&(&bond->mcast_work)->work) FD: 232 BD: 1 +.+.: (wq_completion)bond209#3 ->&rq->__lock ->(work_completion)(&(&slave->notify_work)->work) ->(work_completion)(&(&bond->mcast_work)->work) FD: 232 BD: 1 +.+.: (wq_completion)bond237 ->(work_completion)(&(&slave->notify_work)->work) ->(work_completion)(&(&bond->mcast_work)->work) FD: 232 BD: 1 +.+.: (wq_completion)bond226#2 ->&rq->__lock ->(work_completion)(&(&slave->notify_work)->work) ->(work_completion)(&(&bond->mcast_work)->work) FD: 232 BD: 1 +.+.: (wq_completion)bond210#3 ->(work_completion)(&(&slave->notify_work)->work) ->(work_completion)(&(&bond->mcast_work)->work) FD: 232 BD: 1 +.+.: (wq_completion)bond238 ->(work_completion)(&(&slave->notify_work)->work) ->(work_completion)(&(&bond->mcast_work)->work) FD: 232 BD: 1 +.+.: (wq_completion)bond227#2 ->(work_completion)(&(&slave->notify_work)->work) ->(work_completion)(&(&bond->mcast_work)->work) FD: 232 BD: 1 +.+.: (wq_completion)bond211#3 ->&rq->__lock ->(work_completion)(&(&slave->notify_work)->work) ->(work_completion)(&(&bond->mcast_work)->work) FD: 232 BD: 1 +.+.: (wq_completion)bond239 ->(work_completion)(&(&slave->notify_work)->work) ->(work_completion)(&(&bond->mcast_work)->work) FD: 232 BD: 1 +.+.: (wq_completion)bond74#2 ->(work_completion)(&(&slave->notify_work)->work) ->(work_completion)(&(&bond->mcast_work)->work) FD: 232 BD: 1 +.+.: (wq_completion)bond228#2 ->(work_completion)(&(&slave->notify_work)->work) ->(work_completion)(&(&bond->mcast_work)->work) ->&rq->__lock FD: 232 BD: 1 +.+.: (wq_completion)bond212#3 ->(work_completion)(&(&slave->notify_work)->work) ->(work_completion)(&(&bond->mcast_work)->work) FD: 232 BD: 1 +.+.: (wq_completion)bond240 ->(work_completion)(&(&slave->notify_work)->work) ->(work_completion)(&(&bond->mcast_work)->work) FD: 232 BD: 1 +.+.: (wq_completion)bond229#2 ->&rq->__lock ->(work_completion)(&(&slave->notify_work)->work) ->(work_completion)(&(&bond->mcast_work)->work) FD: 232 BD: 1 +.+.: (wq_completion)bond213#2 ->(work_completion)(&(&slave->notify_work)->work) ->(work_completion)(&(&bond->mcast_work)->work) FD: 232 BD: 1 +.+.: (wq_completion)bond241 ->(work_completion)(&(&slave->notify_work)->work) ->(work_completion)(&(&bond->mcast_work)->work) FD: 232 BD: 1 +.+.: (wq_completion)bond214#3 ->(work_completion)(&(&slave->notify_work)->work) ->(work_completion)(&(&bond->mcast_work)->work) FD: 232 BD: 1 +.+.: (wq_completion)bond230#2 ->(work_completion)(&(&slave->notify_work)->work) ->(work_completion)(&(&bond->mcast_work)->work) FD: 232 BD: 1 +.+.: (wq_completion)bond242 ->(work_completion)(&(&slave->notify_work)->work) ->(work_completion)(&(&bond->mcast_work)->work) FD: 232 BD: 1 +.+.: (wq_completion)bond215#3 ->(work_completion)(&(&slave->notify_work)->work) ->(work_completion)(&(&bond->mcast_work)->work) FD: 232 BD: 1 +.+.: (wq_completion)bond231#2 ->&rq->__lock ->(work_completion)(&(&slave->notify_work)->work) ->(work_completion)(&(&bond->mcast_work)->work) FD: 232 BD: 1 +.+.: (wq_completion)bond232#2 ->(work_completion)(&(&slave->notify_work)->work) ->(work_completion)(&(&bond->mcast_work)->work) FD: 232 BD: 1 +.+.: (wq_completion)bond216#3 ->(work_completion)(&(&slave->notify_work)->work) ->(work_completion)(&(&bond->mcast_work)->work) FD: 232 BD: 1 +.+.: (wq_completion)bond243 ->(work_completion)(&(&slave->notify_work)->work) ->(work_completion)(&(&bond->mcast_work)->work) FD: 232 BD: 1 +.+.: (wq_completion)bond217#3 ->(work_completion)(&(&slave->notify_work)->work) ->(work_completion)(&(&bond->mcast_work)->work) FD: 232 BD: 1 +.+.: (wq_completion)bond244 ->(work_completion)(&(&slave->notify_work)->work) ->(work_completion)(&(&bond->mcast_work)->work) FD: 232 BD: 1 +.+.: (wq_completion)bond233#2 ->(work_completion)(&(&slave->notify_work)->work) ->(work_completion)(&(&bond->mcast_work)->work) FD: 232 BD: 1 +.+.: (wq_completion)bond218#3 ->(work_completion)(&(&slave->notify_work)->work) ->(work_completion)(&(&bond->mcast_work)->work) FD: 232 BD: 1 +.+.: (wq_completion)bond75#2 ->(work_completion)(&(&slave->notify_work)->work) ->(work_completion)(&(&bond->mcast_work)->work) FD: 232 BD: 1 +.+.: (wq_completion)bond245 ->(work_completion)(&(&slave->notify_work)->work) ->(work_completion)(&(&bond->mcast_work)->work) FD: 232 BD: 1 +.+.: (wq_completion)bond234#2 ->(work_completion)(&(&slave->notify_work)->work) ->(work_completion)(&(&bond->mcast_work)->work) FD: 232 BD: 1 +.+.: (wq_completion)bond219#3 ->(work_completion)(&(&slave->notify_work)->work) ->(work_completion)(&(&bond->mcast_work)->work) FD: 232 BD: 1 +.+.: (wq_completion)bond235#2 ->(work_completion)(&(&slave->notify_work)->work) ->(work_completion)(&(&bond->mcast_work)->work) FD: 232 BD: 1 +.+.: (wq_completion)bond246 ->(work_completion)(&(&slave->notify_work)->work) ->(work_completion)(&(&bond->mcast_work)->work) FD: 232 BD: 1 +.+.: (wq_completion)bond220#3 ->(work_completion)(&(&slave->notify_work)->work) ->(work_completion)(&(&bond->mcast_work)->work) FD: 232 BD: 1 +.+.: (wq_completion)bond236#2 ->(work_completion)(&(&slave->notify_work)->work) ->(work_completion)(&(&bond->mcast_work)->work) FD: 232 BD: 1 +.+.: (wq_completion)bond247 ->(work_completion)(&(&slave->notify_work)->work) ->(work_completion)(&(&bond->mcast_work)->work) FD: 232 BD: 1 +.+.: (wq_completion)bond221#3 ->(work_completion)(&(&slave->notify_work)->work) ->(work_completion)(&(&bond->mcast_work)->work) ->&rq->__lock FD: 232 BD: 1 +.+.: (wq_completion)bond237#2 ->(work_completion)(&(&slave->notify_work)->work) ->(work_completion)(&(&bond->mcast_work)->work) FD: 232 BD: 1 +.+.: (wq_completion)bond248 ->&rq->__lock ->(work_completion)(&(&slave->notify_work)->work) ->(work_completion)(&(&bond->mcast_work)->work) FD: 232 BD: 1 +.+.: (wq_completion)bond222#3 ->(work_completion)(&(&slave->notify_work)->work) ->(work_completion)(&(&bond->mcast_work)->work) ->&rq->__lock FD: 232 BD: 1 +.+.: (wq_completion)bond238#2 ->(work_completion)(&(&slave->notify_work)->work) ->(work_completion)(&(&bond->mcast_work)->work) FD: 232 BD: 1 +.+.: (wq_completion)bond249 ->(work_completion)(&(&slave->notify_work)->work) ->(work_completion)(&(&bond->mcast_work)->work) FD: 232 BD: 1 +.+.: (wq_completion)bond223#3 ->(work_completion)(&(&slave->notify_work)->work) ->(work_completion)(&(&bond->mcast_work)->work) FD: 232 BD: 1 +.+.: (wq_completion)bond239#2 ->(work_completion)(&(&slave->notify_work)->work) ->(work_completion)(&(&bond->mcast_work)->work) FD: 232 BD: 1 +.+.: (wq_completion)bond76#2 ->(work_completion)(&(&slave->notify_work)->work) ->(work_completion)(&(&bond->mcast_work)->work) FD: 232 BD: 1 +.+.: (wq_completion)bond250 ->(work_completion)(&(&slave->notify_work)->work) ->(work_completion)(&(&bond->mcast_work)->work) FD: 232 BD: 1 +.+.: (wq_completion)bond224#3 ->&rq->__lock ->(work_completion)(&(&slave->notify_work)->work) ->(work_completion)(&(&bond->mcast_work)->work) FD: 232 BD: 1 +.+.: (wq_completion)bond240#2 ->(work_completion)(&(&slave->notify_work)->work) ->(work_completion)(&(&bond->mcast_work)->work) FD: 232 BD: 1 +.+.: (wq_completion)bond251 ->&rq->__lock ->(work_completion)(&(&slave->notify_work)->work) ->(work_completion)(&(&bond->mcast_work)->work) FD: 232 BD: 1 +.+.: (wq_completion)bond225#3 ->(work_completion)(&(&slave->notify_work)->work) ->(work_completion)(&(&bond->mcast_work)->work) FD: 232 BD: 1 +.+.: (wq_completion)bond241#2 ->(work_completion)(&(&slave->notify_work)->work) ->(work_completion)(&(&bond->mcast_work)->work) FD: 232 BD: 1 +.+.: (wq_completion)bond252 ->&rq->__lock ->(work_completion)(&(&slave->notify_work)->work) ->(work_completion)(&(&bond->mcast_work)->work) FD: 232 BD: 1 +.+.: (wq_completion)bond77#2 ->(work_completion)(&(&slave->notify_work)->work) ->(work_completion)(&(&bond->mcast_work)->work) FD: 232 BD: 1 +.+.: (wq_completion)bond226#3 ->(work_completion)(&(&slave->notify_work)->work) ->(work_completion)(&(&bond->mcast_work)->work) ->&rq->__lock FD: 232 BD: 1 +.+.: (wq_completion)bond242#2 ->(work_completion)(&(&slave->notify_work)->work) ->(work_completion)(&(&bond->mcast_work)->work) FD: 232 BD: 1 +.+.: (wq_completion)bond253 ->(work_completion)(&(&slave->notify_work)->work) ->(work_completion)(&(&bond->mcast_work)->work) FD: 232 BD: 1 +.+.: (wq_completion)bond227#3 ->(work_completion)(&(&slave->notify_work)->work) ->(work_completion)(&(&bond->mcast_work)->work) FD: 232 BD: 1 +.+.: (wq_completion)bond254 ->&rq->__lock ->(work_completion)(&(&slave->notify_work)->work) ->(work_completion)(&(&bond->mcast_work)->work) FD: 232 BD: 1 +.+.: (wq_completion)bond243#2 ->(work_completion)(&(&slave->notify_work)->work) ->(work_completion)(&(&bond->mcast_work)->work) FD: 232 BD: 1 +.+.: (wq_completion)bond228#3 ->&rq->__lock ->(work_completion)(&(&slave->notify_work)->work) ->(work_completion)(&(&bond->mcast_work)->work) FD: 232 BD: 1 +.+.: (wq_completion)bond255 ->(work_completion)(&(&slave->notify_work)->work) ->(work_completion)(&(&bond->mcast_work)->work) FD: 232 BD: 1 +.+.: (wq_completion)bond244#2 ->&rq->__lock ->(work_completion)(&(&slave->notify_work)->work) ->(work_completion)(&(&bond->mcast_work)->work) FD: 232 BD: 1 +.+.: (wq_completion)bond78#2 ->&rq->__lock ->(work_completion)(&(&slave->notify_work)->work) ->(work_completion)(&(&bond->mcast_work)->work) FD: 232 BD: 1 +.+.: (wq_completion)bond229#3 ->(work_completion)(&(&slave->notify_work)->work) ->(work_completion)(&(&bond->mcast_work)->work) FD: 232 BD: 1 +.+.: (wq_completion)bond256 ->(work_completion)(&(&slave->notify_work)->work) ->(work_completion)(&(&bond->mcast_work)->work) FD: 232 BD: 1 +.+.: (wq_completion)bond245#2 ->(work_completion)(&(&slave->notify_work)->work) ->(work_completion)(&(&bond->mcast_work)->work) FD: 232 BD: 1 +.+.: (wq_completion)bond230#3 ->(work_completion)(&(&slave->notify_work)->work) ->(work_completion)(&(&bond->mcast_work)->work) FD: 232 BD: 1 +.+.: (wq_completion)bond79#3 ->(work_completion)(&(&slave->notify_work)->work) ->(work_completion)(&(&bond->mcast_work)->work) FD: 232 BD: 1 +.+.: (wq_completion)bond257 ->(work_completion)(&(&slave->notify_work)->work) ->(work_completion)(&(&bond->mcast_work)->work) FD: 232 BD: 1 +.+.: (wq_completion)bond246#2 ->(work_completion)(&(&slave->notify_work)->work) ->(work_completion)(&(&bond->mcast_work)->work) ->&rq->__lock FD: 232 BD: 1 +.+.: (wq_completion)bond231#3 ->(work_completion)(&(&slave->notify_work)->work) ->(work_completion)(&(&bond->mcast_work)->work) FD: 232 BD: 1 +.+.: (wq_completion)bond258 ->&rq->__lock ->(work_completion)(&(&slave->notify_work)->work) ->(work_completion)(&(&bond->mcast_work)->work) FD: 232 BD: 1 +.+.: (wq_completion)bond247#2 ->(work_completion)(&(&slave->notify_work)->work) ->(work_completion)(&(&bond->mcast_work)->work) FD: 232 BD: 1 +.+.: (wq_completion)bond232#3 ->(work_completion)(&(&slave->notify_work)->work) ->(work_completion)(&(&bond->mcast_work)->work) ->&rq->__lock FD: 232 BD: 1 +.+.: (wq_completion)bond80#3 ->(work_completion)(&(&slave->notify_work)->work) ->(work_completion)(&(&bond->mcast_work)->work) FD: 232 BD: 1 +.+.: (wq_completion)bond259 ->(work_completion)(&(&slave->notify_work)->work) ->(work_completion)(&(&bond->mcast_work)->work) ->&rq->__lock FD: 232 BD: 1 +.+.: (wq_completion)bond248#2 ->(work_completion)(&(&slave->notify_work)->work) ->(work_completion)(&(&bond->mcast_work)->work) ->&rq->__lock FD: 232 BD: 1 +.+.: (wq_completion)bond260 ->(work_completion)(&(&slave->notify_work)->work) ->(work_completion)(&(&bond->mcast_work)->work) ->&rq->__lock FD: 232 BD: 1 +.+.: (wq_completion)bond249#2 ->&rq->__lock ->(work_completion)(&(&slave->notify_work)->work) ->(work_completion)(&(&bond->mcast_work)->work) FD: 232 BD: 1 +.+.: (wq_completion)bond234#3 ->(work_completion)(&(&slave->notify_work)->work) ->(work_completion)(&(&bond->mcast_work)->work) FD: 232 BD: 1 +.+.: (wq_completion)bond250#2 ->(work_completion)(&(&slave->notify_work)->work) ->(work_completion)(&(&bond->mcast_work)->work) FD: 232 BD: 1 +.+.: (wq_completion)bond235#3 ->(work_completion)(&(&slave->notify_work)->work) ->(work_completion)(&(&bond->mcast_work)->work) ->&rq->__lock FD: 232 BD: 1 +.+.: (wq_completion)bond262 ->(work_completion)(&(&slave->notify_work)->work) ->(work_completion)(&(&bond->mcast_work)->work) FD: 232 BD: 1 +.+.: (wq_completion)bond251#2 ->&rq->__lock ->(work_completion)(&(&slave->notify_work)->work) ->(work_completion)(&(&bond->mcast_work)->work) FD: 232 BD: 1 +.+.: (wq_completion)bond263 ->(work_completion)(&(&slave->notify_work)->work) ->(work_completion)(&(&bond->mcast_work)->work) FD: 232 BD: 1 +.+.: (wq_completion)bond252#2 ->(work_completion)(&(&slave->notify_work)->work) ->(work_completion)(&(&bond->mcast_work)->work) ->&rq->__lock FD: 232 BD: 1 +.+.: (wq_completion)bond237#3 ->(work_completion)(&(&slave->notify_work)->work) ->(work_completion)(&(&bond->mcast_work)->work) FD: 232 BD: 1 +.+.: (wq_completion)bond264 ->(work_completion)(&(&slave->notify_work)->work) ->(work_completion)(&(&bond->mcast_work)->work) FD: 232 BD: 1 +.+.: (wq_completion)bond253#2 ->(work_completion)(&(&slave->notify_work)->work) ->(work_completion)(&(&bond->mcast_work)->work) FD: 232 BD: 1 +.+.: (wq_completion)bond81#3 ->(work_completion)(&(&slave->notify_work)->work) ->(work_completion)(&(&bond->mcast_work)->work) FD: 232 BD: 1 +.+.: (wq_completion)bond238#3 ->(work_completion)(&(&slave->notify_work)->work) ->(work_completion)(&(&bond->mcast_work)->work) FD: 232 BD: 1 +.+.: (wq_completion)bond265 ->(work_completion)(&(&slave->notify_work)->work) ->(work_completion)(&(&bond->mcast_work)->work) ->&rq->__lock FD: 232 BD: 1 +.+.: (wq_completion)bond254#2 ->(work_completion)(&(&slave->notify_work)->work) ->(work_completion)(&(&bond->mcast_work)->work) ->&rq->__lock FD: 1 BD: 1 ....: &____s->seqcount#17 FD: 232 BD: 1 +.+.: (wq_completion)bond239#3 ->(work_completion)(&(&slave->notify_work)->work) ->(work_completion)(&(&bond->mcast_work)->work) ->&rq->__lock FD: 232 BD: 1 +.+.: (wq_completion)bond266 ->(work_completion)(&(&slave->notify_work)->work) ->(work_completion)(&(&bond->mcast_work)->work) FD: 232 BD: 1 +.+.: (wq_completion)bond255#2 ->(work_completion)(&(&slave->notify_work)->work) ->(work_completion)(&(&bond->mcast_work)->work) FD: 232 BD: 1 +.+.: (wq_completion)bond240#3 ->(work_completion)(&(&slave->notify_work)->work) ->(work_completion)(&(&bond->mcast_work)->work) ->&rq->__lock FD: 232 BD: 1 +.+.: (wq_completion)bond267 ->(work_completion)(&(&slave->notify_work)->work) ->(work_completion)(&(&bond->mcast_work)->work) FD: 232 BD: 1 +.+.: (wq_completion)bond256#2 ->&rq->__lock ->(work_completion)(&(&slave->notify_work)->work) ->(work_completion)(&(&bond->mcast_work)->work) FD: 232 BD: 1 +.+.: (wq_completion)bond241#3 ->&rq->__lock ->(work_completion)(&(&slave->notify_work)->work) ->(work_completion)(&(&bond->mcast_work)->work) FD: 232 BD: 1 +.+.: (wq_completion)bond268 ->(work_completion)(&(&slave->notify_work)->work) ->(work_completion)(&(&bond->mcast_work)->work) FD: 232 BD: 1 +.+.: (wq_completion)bond257#2 ->&rq->__lock ->(work_completion)(&(&slave->notify_work)->work) ->(work_completion)(&(&bond->mcast_work)->work) FD: 232 BD: 1 +.+.: (wq_completion)bond242#3 ->(work_completion)(&(&slave->notify_work)->work) ->(work_completion)(&(&bond->mcast_work)->work) ->&rq->__lock FD: 232 BD: 1 +.+.: (wq_completion)bond269 ->(work_completion)(&(&slave->notify_work)->work) ->(work_completion)(&(&bond->mcast_work)->work) FD: 232 BD: 1 +.+.: (wq_completion)bond258#2 ->(work_completion)(&(&slave->notify_work)->work) ->(work_completion)(&(&bond->mcast_work)->work) ->&rq->__lock FD: 232 BD: 1 +.+.: (wq_completion)bond243#3 ->(work_completion)(&(&slave->notify_work)->work) ->(work_completion)(&(&bond->mcast_work)->work) FD: 232 BD: 1 +.+.: (wq_completion)bond270 ->(work_completion)(&(&slave->notify_work)->work) ->(work_completion)(&(&bond->mcast_work)->work) ->&rq->__lock FD: 232 BD: 1 +.+.: (wq_completion)bond259#2 ->(work_completion)(&(&slave->notify_work)->work) ->(work_completion)(&(&bond->mcast_work)->work) ->&rq->__lock FD: 232 BD: 1 +.+.: (wq_completion)bond244#3 ->(work_completion)(&(&slave->notify_work)->work) ->(work_completion)(&(&bond->mcast_work)->work) ->&rq->__lock FD: 232 BD: 1 +.+.: (wq_completion)bond82#2 ->&rq->__lock ->(work_completion)(&(&slave->notify_work)->work) ->(work_completion)(&(&bond->mcast_work)->work) FD: 232 BD: 1 +.+.: (wq_completion)bond271 ->(work_completion)(&(&slave->notify_work)->work) ->(work_completion)(&(&bond->mcast_work)->work) FD: 232 BD: 1 +.+.: (wq_completion)bond260#2 ->(work_completion)(&(&slave->notify_work)->work) ->(work_completion)(&(&bond->mcast_work)->work) FD: 232 BD: 1 +.+.: (wq_completion)bond245#3 ->(work_completion)(&(&slave->notify_work)->work) ->(work_completion)(&(&bond->mcast_work)->work) ->&rq->__lock FD: 232 BD: 1 +.+.: (wq_completion)bond272 ->(work_completion)(&(&slave->notify_work)->work) ->(work_completion)(&(&bond->mcast_work)->work) ->&rq->__lock FD: 232 BD: 1 +.+.: (wq_completion)bond83#2 ->(work_completion)(&(&slave->notify_work)->work) ->(work_completion)(&(&bond->mcast_work)->work) ->&rq->__lock FD: 232 BD: 1 +.+.: (wq_completion)bond261 ->(work_completion)(&(&slave->notify_work)->work) ->(work_completion)(&(&bond->mcast_work)->work) FD: 232 BD: 1 +.+.: (wq_completion)bond273 ->(work_completion)(&(&slave->notify_work)->work) ->(work_completion)(&(&bond->mcast_work)->work) FD: 232 BD: 1 +.+.: (wq_completion)bond84#2 ->(work_completion)(&(&slave->notify_work)->work) ->(work_completion)(&(&bond->mcast_work)->work) FD: 232 BD: 1 +.+.: (wq_completion)bond247#3 ->(work_completion)(&(&slave->notify_work)->work) ->(work_completion)(&(&bond->mcast_work)->work) FD: 232 BD: 1 +.+.: (wq_completion)bond262#2 ->(work_completion)(&(&slave->notify_work)->work) ->(work_completion)(&(&bond->mcast_work)->work) FD: 232 BD: 1 +.+.: (wq_completion)bond274 ->&rq->__lock ->(work_completion)(&(&slave->notify_work)->work) ->(work_completion)(&(&bond->mcast_work)->work) FD: 232 BD: 1 +.+.: (wq_completion)bond248#3 ->(work_completion)(&(&slave->notify_work)->work) ->(work_completion)(&(&bond->mcast_work)->work) ->&rq->__lock FD: 232 BD: 1 +.+.: (wq_completion)bond263#2 ->(work_completion)(&(&slave->notify_work)->work) ->(work_completion)(&(&bond->mcast_work)->work) ->&rq->__lock FD: 232 BD: 1 +.+.: (wq_completion)bond275 ->(work_completion)(&(&slave->notify_work)->work) ->(work_completion)(&(&bond->mcast_work)->work) FD: 232 BD: 1 +.+.: (wq_completion)bond249#3 ->&rq->__lock ->(work_completion)(&(&slave->notify_work)->work) ->(work_completion)(&(&bond->mcast_work)->work) FD: 232 BD: 1 +.+.: (wq_completion)bond85#3 ->&rq->__lock ->(work_completion)(&(&slave->notify_work)->work) ->(work_completion)(&(&bond->mcast_work)->work) FD: 232 BD: 1 +.+.: (wq_completion)bond264#2 ->(work_completion)(&(&slave->notify_work)->work) ->(work_completion)(&(&bond->mcast_work)->work) FD: 232 BD: 1 +.+.: (wq_completion)bond276 ->(work_completion)(&(&slave->notify_work)->work) ->(work_completion)(&(&bond->mcast_work)->work) FD: 232 BD: 1 +.+.: (wq_completion)bond250#3 ->&rq->__lock ->(work_completion)(&(&slave->notify_work)->work) ->(work_completion)(&(&bond->mcast_work)->work) FD: 232 BD: 1 +.+.: (wq_completion)bond86#2 ->&rq->__lock ->(work_completion)(&(&slave->notify_work)->work) ->(work_completion)(&(&bond->mcast_work)->work) FD: 232 BD: 1 +.+.: (wq_completion)bond265#2 ->(work_completion)(&(&slave->notify_work)->work) ->(work_completion)(&(&bond->mcast_work)->work) FD: 232 BD: 1 +.+.: (wq_completion)bond277 ->(work_completion)(&(&slave->notify_work)->work) ->(work_completion)(&(&bond->mcast_work)->work) FD: 232 BD: 1 +.+.: (wq_completion)bond251#3 ->(work_completion)(&(&slave->notify_work)->work) ->(work_completion)(&(&bond->mcast_work)->work) FD: 232 BD: 1 +.+.: (wq_completion)bond87#2 ->&rq->__lock ->(work_completion)(&(&slave->notify_work)->work) ->(work_completion)(&(&bond->mcast_work)->work) FD: 232 BD: 1 +.+.: (wq_completion)bond266#2 ->(work_completion)(&(&slave->notify_work)->work) ->(work_completion)(&(&bond->mcast_work)->work) FD: 232 BD: 1 +.+.: (wq_completion)bond278 ->(work_completion)(&(&slave->notify_work)->work) ->(work_completion)(&(&bond->mcast_work)->work) FD: 232 BD: 1 +.+.: (wq_completion)bond252#3 ->&rq->__lock ->(work_completion)(&(&slave->notify_work)->work) ->(work_completion)(&(&bond->mcast_work)->work) FD: 232 BD: 1 +.+.: (wq_completion)bond88#3 ->(work_completion)(&(&slave->notify_work)->work) ->(work_completion)(&(&bond->mcast_work)->work) FD: 232 BD: 1 +.+.: (wq_completion)bond267#2 ->&rq->__lock ->(work_completion)(&(&slave->notify_work)->work) ->(work_completion)(&(&bond->mcast_work)->work) FD: 232 BD: 1 +.+.: (wq_completion)bond279 ->(work_completion)(&(&slave->notify_work)->work) ->(work_completion)(&(&bond->mcast_work)->work) FD: 232 BD: 1 +.+.: (wq_completion)bond253#3 ->(work_completion)(&(&slave->notify_work)->work) ->(work_completion)(&(&bond->mcast_work)->work) FD: 232 BD: 1 +.+.: (wq_completion)bond268#2 ->(work_completion)(&(&slave->notify_work)->work) ->(work_completion)(&(&bond->mcast_work)->work) FD: 232 BD: 1 +.+.: (wq_completion)bond280 ->(work_completion)(&(&slave->notify_work)->work) ->(work_completion)(&(&bond->mcast_work)->work) ->&rq->__lock FD: 232 BD: 1 +.+.: (wq_completion)bond254#3 ->(work_completion)(&(&slave->notify_work)->work) ->(work_completion)(&(&bond->mcast_work)->work) ->&rq->__lock FD: 232 BD: 1 +.+.: (wq_completion)bond269#2 ->(work_completion)(&(&slave->notify_work)->work) ->(work_completion)(&(&bond->mcast_work)->work) FD: 232 BD: 1 +.+.: (wq_completion)bond281 ->&rq->__lock ->(work_completion)(&(&slave->notify_work)->work) ->(work_completion)(&(&bond->mcast_work)->work) FD: 232 BD: 1 +.+.: (wq_completion)bond89#3 ->(work_completion)(&(&slave->notify_work)->work) ->(work_completion)(&(&bond->mcast_work)->work) FD: 232 BD: 1 +.+.: (wq_completion)bond255#3 ->(work_completion)(&(&slave->notify_work)->work) ->(work_completion)(&(&bond->mcast_work)->work) FD: 232 BD: 1 +.+.: (wq_completion)bond270#2 ->(work_completion)(&(&slave->notify_work)->work) ->(work_completion)(&(&bond->mcast_work)->work) ->&rq->__lock FD: 232 BD: 1 +.+.: (wq_completion)bond282 ->(work_completion)(&(&slave->notify_work)->work) ->(work_completion)(&(&bond->mcast_work)->work) ->&rq->__lock FD: 232 BD: 1 +.+.: (wq_completion)bond256#3 ->(work_completion)(&(&slave->notify_work)->work) ->(work_completion)(&(&bond->mcast_work)->work) FD: 232 BD: 1 +.+.: (wq_completion)bond271#2 ->(work_completion)(&(&slave->notify_work)->work) ->(work_completion)(&(&bond->mcast_work)->work) FD: 232 BD: 1 +.+.: (wq_completion)bond283 ->(work_completion)(&(&slave->notify_work)->work) ->(work_completion)(&(&bond->mcast_work)->work) FD: 232 BD: 1 +.+.: (wq_completion)bond257#3 ->(work_completion)(&(&slave->notify_work)->work) ->(work_completion)(&(&bond->mcast_work)->work) ->&rq->__lock FD: 232 BD: 1 +.+.: (wq_completion)bond272#2 ->&rq->__lock ->(work_completion)(&(&slave->notify_work)->work) ->(work_completion)(&(&bond->mcast_work)->work) FD: 232 BD: 1 +.+.: (wq_completion)bond284 ->(work_completion)(&(&slave->notify_work)->work) ->(work_completion)(&(&bond->mcast_work)->work) FD: 232 BD: 1 +.+.: (wq_completion)bond273#2 ->(work_completion)(&(&slave->notify_work)->work) ->(work_completion)(&(&bond->mcast_work)->work) FD: 232 BD: 1 +.+.: (wq_completion)bond258#3 ->&rq->__lock ->(work_completion)(&(&slave->notify_work)->work) ->(work_completion)(&(&bond->mcast_work)->work) FD: 232 BD: 1 +.+.: (wq_completion)bond259#3 ->&rq->__lock ->(work_completion)(&(&slave->notify_work)->work) ->(work_completion)(&(&bond->mcast_work)->work) FD: 232 BD: 1 +.+.: (wq_completion)bond274#2 ->(work_completion)(&(&slave->notify_work)->work) ->(work_completion)(&(&bond->mcast_work)->work) ->&rq->__lock FD: 232 BD: 1 +.+.: (wq_completion)bond286 ->(work_completion)(&(&slave->notify_work)->work) ->(work_completion)(&(&bond->mcast_work)->work) FD: 232 BD: 1 +.+.: (wq_completion)bond260#3 ->&rq->__lock ->(work_completion)(&(&slave->notify_work)->work) ->(work_completion)(&(&bond->mcast_work)->work) FD: 232 BD: 1 +.+.: (wq_completion)bond275#2 ->&rq->__lock ->(work_completion)(&(&slave->notify_work)->work) ->(work_completion)(&(&bond->mcast_work)->work) FD: 232 BD: 1 +.+.: (wq_completion)bond90#2 ->&rq->__lock ->(work_completion)(&(&slave->notify_work)->work) ->(work_completion)(&(&bond->mcast_work)->work) FD: 232 BD: 1 +.+.: (wq_completion)bond287 ->&rq->__lock ->(work_completion)(&(&slave->notify_work)->work) ->(work_completion)(&(&bond->mcast_work)->work) FD: 232 BD: 1 +.+.: (wq_completion)bond261#2 ->(work_completion)(&(&slave->notify_work)->work) ->(work_completion)(&(&bond->mcast_work)->work) ->&rq->__lock FD: 232 BD: 1 +.+.: (wq_completion)bond276#2 ->&rq->__lock ->(work_completion)(&(&slave->notify_work)->work) ->(work_completion)(&(&bond->mcast_work)->work) FD: 232 BD: 1 +.+.: (wq_completion)bond288 ->&rq->__lock ->(work_completion)(&(&slave->notify_work)->work) ->(work_completion)(&(&bond->mcast_work)->work) FD: 232 BD: 1 +.+.: (wq_completion)bond262#3 ->(work_completion)(&(&slave->notify_work)->work) ->(work_completion)(&(&bond->mcast_work)->work) ->&rq->__lock FD: 232 BD: 1 +.+.: (wq_completion)bond277#2 ->(work_completion)(&(&slave->notify_work)->work) ->(work_completion)(&(&bond->mcast_work)->work) FD: 232 BD: 1 +.+.: (wq_completion)bond289 ->(work_completion)(&(&slave->notify_work)->work) ->(work_completion)(&(&bond->mcast_work)->work) ->&rq->__lock FD: 232 BD: 1 +.+.: (wq_completion)bond263#3 ->(work_completion)(&(&slave->notify_work)->work) ->(work_completion)(&(&bond->mcast_work)->work) ->&rq->__lock FD: 232 BD: 1 +.+.: (wq_completion)bond278#2 ->(work_completion)(&(&slave->notify_work)->work) ->(work_completion)(&(&bond->mcast_work)->work) ->&rq->__lock FD: 232 BD: 1 +.+.: (wq_completion)bond290 ->&rq->__lock ->(work_completion)(&(&slave->notify_work)->work) ->(work_completion)(&(&bond->mcast_work)->work) FD: 232 BD: 1 +.+.: (wq_completion)bond264#3 ->(work_completion)(&(&slave->notify_work)->work) ->(work_completion)(&(&bond->mcast_work)->work) FD: 232 BD: 1 +.+.: (wq_completion)bond279#2 ->(work_completion)(&(&slave->notify_work)->work) ->(work_completion)(&(&bond->mcast_work)->work) ->&rq->__lock FD: 232 BD: 1 +.+.: (wq_completion)bond291 ->(work_completion)(&(&slave->notify_work)->work) ->(work_completion)(&(&bond->mcast_work)->work) ->&rq->__lock FD: 232 BD: 1 +.+.: (wq_completion)bond265#3 ->&rq->__lock ->(work_completion)(&(&slave->notify_work)->work) ->(work_completion)(&(&bond->mcast_work)->work) FD: 232 BD: 1 +.+.: (wq_completion)bond280#2 ->(work_completion)(&(&slave->notify_work)->work) ->(work_completion)(&(&bond->mcast_work)->work) ->&rq->__lock FD: 232 BD: 1 +.+.: (wq_completion)bond292 ->(work_completion)(&(&slave->notify_work)->work) ->(work_completion)(&(&bond->mcast_work)->work) ->&rq->__lock FD: 232 BD: 1 +.+.: (wq_completion)bond266#3 ->(work_completion)(&(&slave->notify_work)->work) ->(work_completion)(&(&bond->mcast_work)->work) FD: 232 BD: 1 +.+.: (wq_completion)bond281#2 ->&rq->__lock ->(work_completion)(&(&slave->notify_work)->work) ->(work_completion)(&(&bond->mcast_work)->work) FD: 232 BD: 1 +.+.: (wq_completion)bond91#2 ->(work_completion)(&(&slave->notify_work)->work) ->(work_completion)(&(&bond->mcast_work)->work) ->&rq->__lock FD: 232 BD: 1 +.+.: (wq_completion)bond293 ->(work_completion)(&(&slave->notify_work)->work) ->(work_completion)(&(&bond->mcast_work)->work) FD: 232 BD: 1 +.+.: (wq_completion)bond267#3 ->&rq->__lock ->(work_completion)(&(&slave->notify_work)->work) ->(work_completion)(&(&bond->mcast_work)->work) FD: 232 BD: 1 +.+.: (wq_completion)bond282#2 ->&rq->__lock ->(work_completion)(&(&slave->notify_work)->work) ->(work_completion)(&(&bond->mcast_work)->work) FD: 232 BD: 1 +.+.: (wq_completion)bond92#3 ->&rq->__lock ->(work_completion)(&(&slave->notify_work)->work) ->(work_completion)(&(&bond->mcast_work)->work) FD: 232 BD: 1 +.+.: (wq_completion)bond268#3 ->(work_completion)(&(&slave->notify_work)->work) ->(work_completion)(&(&bond->mcast_work)->work) FD: 232 BD: 1 +.+.: (wq_completion)bond294 ->(work_completion)(&(&slave->notify_work)->work) ->(work_completion)(&(&bond->mcast_work)->work) ->&rq->__lock FD: 232 BD: 1 +.+.: (wq_completion)bond283#2 ->&rq->__lock ->(work_completion)(&(&slave->notify_work)->work) ->(work_completion)(&(&bond->mcast_work)->work) FD: 232 BD: 1 +.+.: (wq_completion)bond269#3 ->(work_completion)(&(&slave->notify_work)->work) ->(work_completion)(&(&bond->mcast_work)->work) ->&rq->__lock FD: 232 BD: 1 +.+.: (wq_completion)bond295 ->(work_completion)(&(&slave->notify_work)->work) ->(work_completion)(&(&bond->mcast_work)->work) FD: 232 BD: 1 +.+.: (wq_completion)bond284#2 ->(work_completion)(&(&slave->notify_work)->work) ->(work_completion)(&(&bond->mcast_work)->work) ->&rq->__lock FD: 232 BD: 1 +.+.: (wq_completion)bond270#3 ->(work_completion)(&(&slave->notify_work)->work) ->(work_completion)(&(&bond->mcast_work)->work) ->&rq->__lock FD: 232 BD: 1 +.+.: (wq_completion)bond93#3 ->(work_completion)(&(&slave->notify_work)->work) ->(work_completion)(&(&bond->mcast_work)->work) FD: 232 BD: 1 +.+.: (wq_completion)bond296 ->&rq->__lock ->(work_completion)(&(&slave->notify_work)->work) ->(work_completion)(&(&bond->mcast_work)->work) FD: 232 BD: 1 +.+.: (wq_completion)bond285 ->(work_completion)(&(&slave->notify_work)->work) ->(work_completion)(&(&bond->mcast_work)->work) FD: 232 BD: 1 +.+.: (wq_completion)bond271#3 ->(work_completion)(&(&slave->notify_work)->work) ->(work_completion)(&(&bond->mcast_work)->work) FD: 232 BD: 1 +.+.: (wq_completion)bond297 ->(work_completion)(&(&slave->notify_work)->work) ->(work_completion)(&(&bond->mcast_work)->work) ->&rq->__lock FD: 232 BD: 1 +.+.: (wq_completion)bond94#3 ->&rq->__lock ->(work_completion)(&(&slave->notify_work)->work) ->(work_completion)(&(&bond->mcast_work)->work) FD: 232 BD: 1 +.+.: (wq_completion)bond286#2 ->(work_completion)(&(&slave->notify_work)->work) ->(work_completion)(&(&bond->mcast_work)->work) ->&rq->__lock FD: 232 BD: 1 +.+.: (wq_completion)bond272#3 ->&rq->__lock ->(work_completion)(&(&slave->notify_work)->work) ->(work_completion)(&(&bond->mcast_work)->work) FD: 232 BD: 1 +.+.: (wq_completion)bond298 ->&rq->__lock ->(work_completion)(&(&slave->notify_work)->work) ->(work_completion)(&(&bond->mcast_work)->work) FD: 232 BD: 1 +.+.: (wq_completion)bond95#3 ->(work_completion)(&(&slave->notify_work)->work) ->(work_completion)(&(&bond->mcast_work)->work) FD: 232 BD: 1 +.+.: (wq_completion)bond287#2 ->(work_completion)(&(&slave->notify_work)->work) ->(work_completion)(&(&bond->mcast_work)->work) ->&rq->__lock FD: 232 BD: 1 +.+.: (wq_completion)bond273#3 ->(work_completion)(&(&slave->notify_work)->work) ->(work_completion)(&(&bond->mcast_work)->work) FD: 232 BD: 1 +.+.: (wq_completion)bond299 ->(work_completion)(&(&slave->notify_work)->work) ->(work_completion)(&(&bond->mcast_work)->work) FD: 232 BD: 1 +.+.: (wq_completion)bond288#2 ->(work_completion)(&(&slave->notify_work)->work) ->(work_completion)(&(&bond->mcast_work)->work) ->&rq->__lock FD: 232 BD: 1 +.+.: (wq_completion)bond96#4 ->(work_completion)(&(&slave->notify_work)->work) ->(work_completion)(&(&bond->mcast_work)->work) FD: 232 BD: 1 +.+.: (wq_completion)bond274#3 ->(work_completion)(&(&slave->notify_work)->work) ->(work_completion)(&(&bond->mcast_work)->work) ->&rq->__lock FD: 232 BD: 1 +.+.: (wq_completion)bond300 ->(work_completion)(&(&slave->notify_work)->work) ->(work_completion)(&(&bond->mcast_work)->work) FD: 232 BD: 1 +.+.: (wq_completion)bond289#2 ->&rq->__lock ->(work_completion)(&(&slave->notify_work)->work) ->(work_completion)(&(&bond->mcast_work)->work) FD: 232 BD: 1 +.+.: (wq_completion)bond301 ->&rq->__lock ->&cfs_rq->removed.lock ->&obj_hash[i].lock ->(work_completion)(&(&slave->notify_work)->work) ->(work_completion)(&(&bond->mcast_work)->work) FD: 232 BD: 1 +.+.: (wq_completion)bond97#4 ->(work_completion)(&(&slave->notify_work)->work) ->(work_completion)(&(&bond->mcast_work)->work) ->&rq->__lock FD: 232 BD: 1 +.+.: (wq_completion)bond290#2 ->&rq->__lock ->(work_completion)(&(&slave->notify_work)->work) ->(work_completion)(&(&bond->mcast_work)->work) FD: 232 BD: 1 +.+.: (wq_completion)bond276#3 ->(work_completion)(&(&slave->notify_work)->work) ->(work_completion)(&(&bond->mcast_work)->work) FD: 232 BD: 1 +.+.: (wq_completion)bond302 ->(work_completion)(&(&slave->notify_work)->work) ->(work_completion)(&(&bond->mcast_work)->work) FD: 232 BD: 1 +.+.: (wq_completion)bond98#4 ->(work_completion)(&(&slave->notify_work)->work) ->(work_completion)(&(&bond->mcast_work)->work) ->&rq->__lock FD: 232 BD: 1 +.+.: (wq_completion)bond291#2 ->&rq->__lock ->(work_completion)(&(&slave->notify_work)->work) ->(work_completion)(&(&bond->mcast_work)->work) FD: 232 BD: 1 +.+.: (wq_completion)bond277#3 ->(work_completion)(&(&slave->notify_work)->work) ->(work_completion)(&(&bond->mcast_work)->work) ->&rq->__lock FD: 232 BD: 1 +.+.: (wq_completion)bond303 ->&rq->__lock ->(work_completion)(&(&slave->notify_work)->work) ->(work_completion)(&(&bond->mcast_work)->work) FD: 232 BD: 1 +.+.: (wq_completion)bond292#2 ->(work_completion)(&(&slave->notify_work)->work) ->(work_completion)(&(&bond->mcast_work)->work) FD: 232 BD: 1 +.+.: (wq_completion)bond278#3 ->&rq->__lock ->(work_completion)(&(&slave->notify_work)->work) ->(work_completion)(&(&bond->mcast_work)->work) FD: 232 BD: 1 +.+.: (wq_completion)bond304 ->(work_completion)(&(&slave->notify_work)->work) ->(work_completion)(&(&bond->mcast_work)->work) FD: 232 BD: 1 +.+.: (wq_completion)bond99#4 ->&rq->__lock ->(work_completion)(&(&slave->notify_work)->work) ->(work_completion)(&(&bond->mcast_work)->work) FD: 232 BD: 1 +.+.: (wq_completion)bond293#2 ->&rq->__lock ->(work_completion)(&(&slave->notify_work)->work) ->(work_completion)(&(&bond->mcast_work)->work) FD: 232 BD: 1 +.+.: (wq_completion)bond279#3 ->&rq->__lock ->(work_completion)(&(&slave->notify_work)->work) ->(work_completion)(&(&bond->mcast_work)->work) FD: 232 BD: 1 +.+.: (wq_completion)bond305 ->(work_completion)(&(&slave->notify_work)->work) ->(work_completion)(&(&bond->mcast_work)->work) ->&rq->__lock FD: 232 BD: 1 +.+.: (wq_completion)bond100#4 ->(work_completion)(&(&slave->notify_work)->work) ->(work_completion)(&(&bond->mcast_work)->work) FD: 232 BD: 1 +.+.: (wq_completion)bond294#2 ->(work_completion)(&(&slave->notify_work)->work) ->(work_completion)(&(&bond->mcast_work)->work) FD: 232 BD: 1 +.+.: (wq_completion)bond280#3 ->(work_completion)(&(&slave->notify_work)->work) ->(work_completion)(&(&bond->mcast_work)->work) ->&rq->__lock FD: 232 BD: 1 +.+.: (wq_completion)bond306 ->(work_completion)(&(&slave->notify_work)->work) ->(work_completion)(&(&bond->mcast_work)->work) FD: 232 BD: 1 +.+.: (wq_completion)bond295#2 ->&rq->__lock ->(work_completion)(&(&slave->notify_work)->work) ->(work_completion)(&(&bond->mcast_work)->work) FD: 232 BD: 1 +.+.: (wq_completion)bond101#4 ->(work_completion)(&(&slave->notify_work)->work) ->(work_completion)(&(&bond->mcast_work)->work) ->&rq->__lock FD: 232 BD: 1 +.+.: (wq_completion)bond307 ->&rq->__lock ->(work_completion)(&(&slave->notify_work)->work) ->(work_completion)(&(&bond->mcast_work)->work) FD: 232 BD: 1 +.+.: (wq_completion)bond296#2 ->(work_completion)(&(&slave->notify_work)->work) ->(work_completion)(&(&bond->mcast_work)->work) FD: 232 BD: 1 +.+.: (wq_completion)bond282#3 ->&rq->__lock ->(work_completion)(&(&slave->notify_work)->work) ->(work_completion)(&(&bond->mcast_work)->work) FD: 232 BD: 1 +.+.: (wq_completion)bond297#2 ->(work_completion)(&(&slave->notify_work)->work) ->(work_completion)(&(&bond->mcast_work)->work) FD: 232 BD: 1 +.+.: (wq_completion)bond308 ->&rq->__lock ->(work_completion)(&(&slave->notify_work)->work) ->(work_completion)(&(&bond->mcast_work)->work) FD: 232 BD: 1 +.+.: (wq_completion)bond102#4 ->(work_completion)(&(&slave->notify_work)->work) ->(work_completion)(&(&bond->mcast_work)->work) FD: 232 BD: 1 +.+.: (wq_completion)bond283#3 ->(work_completion)(&(&slave->notify_work)->work) ->(work_completion)(&(&bond->mcast_work)->work) FD: 232 BD: 1 +.+.: (wq_completion)bond298#2 ->&rq->__lock ->(work_completion)(&(&slave->notify_work)->work) ->(work_completion)(&(&bond->mcast_work)->work) FD: 232 BD: 1 +.+.: (wq_completion)bond309 ->&rq->__lock ->(work_completion)(&(&slave->notify_work)->work) ->(work_completion)(&(&bond->mcast_work)->work) FD: 232 BD: 1 +.+.: (wq_completion)bond103#4 ->&rq->__lock ->(work_completion)(&(&slave->notify_work)->work) ->(work_completion)(&(&bond->mcast_work)->work) FD: 232 BD: 1 +.+.: (wq_completion)bond284#3 ->(work_completion)(&(&slave->notify_work)->work) ->(work_completion)(&(&bond->mcast_work)->work) FD: 232 BD: 1 +.+.: (wq_completion)bond299#2 ->&rq->__lock ->(work_completion)(&(&slave->notify_work)->work) ->(work_completion)(&(&bond->mcast_work)->work) FD: 232 BD: 1 +.+.: (wq_completion)bond310 ->&rq->__lock ->(work_completion)(&(&slave->notify_work)->work) ->(work_completion)(&(&bond->mcast_work)->work) FD: 232 BD: 1 +.+.: (wq_completion)bond104#4 ->(work_completion)(&(&slave->notify_work)->work) ->(work_completion)(&(&bond->mcast_work)->work) FD: 232 BD: 1 +.+.: (wq_completion)bond285#2 ->(work_completion)(&(&slave->notify_work)->work) ->(work_completion)(&(&bond->mcast_work)->work) FD: 232 BD: 1 +.+.: (wq_completion)bond300#2 ->&rq->__lock ->(work_completion)(&(&slave->notify_work)->work) ->(work_completion)(&(&bond->mcast_work)->work) FD: 232 BD: 1 +.+.: (wq_completion)bond311 ->(work_completion)(&(&slave->notify_work)->work) ->(work_completion)(&(&bond->mcast_work)->work) FD: 232 BD: 1 +.+.: (wq_completion)bond286#3 ->&rq->__lock ->(work_completion)(&(&slave->notify_work)->work) ->(work_completion)(&(&bond->mcast_work)->work) FD: 232 BD: 1 +.+.: (wq_completion)bond301#2 ->(work_completion)(&(&slave->notify_work)->work) ->(work_completion)(&(&bond->mcast_work)->work) FD: 232 BD: 1 +.+.: (wq_completion)bond105#4 ->(work_completion)(&(&slave->notify_work)->work) ->(work_completion)(&(&bond->mcast_work)->work) ->&rq->__lock FD: 232 BD: 1 +.+.: (wq_completion)bond312 ->(work_completion)(&(&slave->notify_work)->work) ->(work_completion)(&(&bond->mcast_work)->work) FD: 232 BD: 1 +.+.: (wq_completion)bond287#3 ->(work_completion)(&(&slave->notify_work)->work) ->(work_completion)(&(&bond->mcast_work)->work) FD: 232 BD: 1 +.+.: (wq_completion)bond302#2 ->&rq->__lock ->(work_completion)(&(&slave->notify_work)->work) ->(work_completion)(&(&bond->mcast_work)->work) FD: 232 BD: 1 +.+.: (wq_completion)bond313 ->(work_completion)(&(&slave->notify_work)->work) ->(work_completion)(&(&bond->mcast_work)->work) FD: 232 BD: 1 +.+.: (wq_completion)bond106#4 ->&rq->__lock ->(work_completion)(&(&slave->notify_work)->work) ->(work_completion)(&(&bond->mcast_work)->work) FD: 232 BD: 1 +.+.: (wq_completion)bond288#3 ->(work_completion)(&(&slave->notify_work)->work) ->(work_completion)(&(&bond->mcast_work)->work) FD: 232 BD: 1 +.+.: (wq_completion)bond303#2 ->(work_completion)(&(&slave->notify_work)->work) ->(work_completion)(&(&bond->mcast_work)->work) FD: 232 BD: 1 +.+.: (wq_completion)bond314 ->(work_completion)(&(&slave->notify_work)->work) ->(work_completion)(&(&bond->mcast_work)->work) FD: 232 BD: 1 +.+.: (wq_completion)bond289#3 ->(work_completion)(&(&slave->notify_work)->work) ->(work_completion)(&(&bond->mcast_work)->work) ->&rq->__lock FD: 232 BD: 1 +.+.: (wq_completion)bond304#2 ->&rq->__lock ->(work_completion)(&(&slave->notify_work)->work) ->(work_completion)(&(&bond->mcast_work)->work) FD: 232 BD: 1 +.+.: (wq_completion)bond107#4 ->(work_completion)(&(&slave->notify_work)->work) ->(work_completion)(&(&bond->mcast_work)->work) ->&rq->__lock FD: 232 BD: 1 +.+.: (wq_completion)bond315 ->&rq->__lock ->(work_completion)(&(&slave->notify_work)->work) ->(work_completion)(&(&bond->mcast_work)->work) FD: 232 BD: 1 +.+.: (wq_completion)bond290#3 ->(work_completion)(&(&slave->notify_work)->work) ->(work_completion)(&(&bond->mcast_work)->work) FD: 232 BD: 1 +.+.: (wq_completion)bond305#2 ->(work_completion)(&(&slave->notify_work)->work) ->(work_completion)(&(&bond->mcast_work)->work) FD: 232 BD: 1 +.+.: (wq_completion)bond108#4 ->&rq->__lock ->(work_completion)(&(&slave->notify_work)->work) ->(work_completion)(&(&bond->mcast_work)->work) FD: 232 BD: 1 +.+.: (wq_completion)bond316 ->(work_completion)(&(&slave->notify_work)->work) ->(work_completion)(&(&bond->mcast_work)->work) FD: 232 BD: 1 +.+.: (wq_completion)bond291#3 ->&rq->__lock ->(work_completion)(&(&slave->notify_work)->work) ->(work_completion)(&(&bond->mcast_work)->work) FD: 232 BD: 1 +.+.: (wq_completion)bond306#2 ->(work_completion)(&(&slave->notify_work)->work) ->(work_completion)(&(&bond->mcast_work)->work) FD: 232 BD: 1 +.+.: (wq_completion)bond109#4 ->&rq->__lock ->(work_completion)(&(&slave->notify_work)->work) ->(work_completion)(&(&bond->mcast_work)->work) FD: 232 BD: 1 +.+.: (wq_completion)bond317 ->(work_completion)(&(&slave->notify_work)->work) ->(work_completion)(&(&bond->mcast_work)->work) FD: 232 BD: 1 +.+.: (wq_completion)bond292#3 ->(work_completion)(&(&slave->notify_work)->work) ->(work_completion)(&(&bond->mcast_work)->work) FD: 232 BD: 1 +.+.: (wq_completion)bond307#2 ->(work_completion)(&(&slave->notify_work)->work) ->(work_completion)(&(&bond->mcast_work)->work) FD: 232 BD: 1 +.+.: (wq_completion)bond110#4 ->&rq->__lock ->(work_completion)(&(&slave->notify_work)->work) ->(work_completion)(&(&bond->mcast_work)->work) FD: 232 BD: 1 +.+.: (wq_completion)bond318 ->&rq->__lock ->(work_completion)(&(&slave->notify_work)->work) ->(work_completion)(&(&bond->mcast_work)->work) FD: 232 BD: 1 +.+.: (wq_completion)bond293#3 ->&rq->__lock ->(work_completion)(&(&slave->notify_work)->work) ->(work_completion)(&(&bond->mcast_work)->work) FD: 232 BD: 1 +.+.: (wq_completion)bond308#2 ->(work_completion)(&(&slave->notify_work)->work) ->(work_completion)(&(&bond->mcast_work)->work) FD: 232 BD: 1 +.+.: (wq_completion)bond319 ->(work_completion)(&(&slave->notify_work)->work) ->(work_completion)(&(&bond->mcast_work)->work) FD: 232 BD: 1 +.+.: (wq_completion)bond111#4 ->&rq->__lock ->(work_completion)(&(&slave->notify_work)->work) ->(work_completion)(&(&bond->mcast_work)->work) FD: 232 BD: 1 +.+.: (wq_completion)bond294#3 ->&rq->__lock ->(work_completion)(&(&slave->notify_work)->work) ->(work_completion)(&(&bond->mcast_work)->work) FD: 232 BD: 1 +.+.: (wq_completion)bond309#2 ->(work_completion)(&(&slave->notify_work)->work) ->(work_completion)(&(&bond->mcast_work)->work) FD: 232 BD: 1 +.+.: (wq_completion)bond320 ->(work_completion)(&(&slave->notify_work)->work) ->(work_completion)(&(&bond->mcast_work)->work) FD: 232 BD: 1 +.+.: (wq_completion)bond112#4 ->&rq->__lock ->(work_completion)(&(&slave->notify_work)->work) ->(work_completion)(&(&bond->mcast_work)->work) FD: 232 BD: 1 +.+.: (wq_completion)bond295#3 ->(work_completion)(&(&slave->notify_work)->work) ->(work_completion)(&(&bond->mcast_work)->work) FD: 232 BD: 1 +.+.: (wq_completion)bond310#2 ->(work_completion)(&(&slave->notify_work)->work) ->(work_completion)(&(&bond->mcast_work)->work) FD: 232 BD: 1 +.+.: (wq_completion)bond321 ->(work_completion)(&(&slave->notify_work)->work) ->(work_completion)(&(&bond->mcast_work)->work) FD: 232 BD: 1 +.+.: (wq_completion)bond296#3 ->(work_completion)(&(&slave->notify_work)->work) ->(work_completion)(&(&bond->mcast_work)->work) FD: 232 BD: 1 +.+.: (wq_completion)bond311#2 ->(work_completion)(&(&slave->notify_work)->work) ->(work_completion)(&(&bond->mcast_work)->work) FD: 232 BD: 1 +.+.: (wq_completion)bond322 ->(work_completion)(&(&slave->notify_work)->work) ->(work_completion)(&(&bond->mcast_work)->work) FD: 232 BD: 1 +.+.: (wq_completion)bond297#3 ->&rq->__lock ->(work_completion)(&(&slave->notify_work)->work) ->(work_completion)(&(&bond->mcast_work)->work) FD: 232 BD: 1 +.+.: (wq_completion)bond312#2 ->&rq->__lock ->(work_completion)(&(&slave->notify_work)->work) ->(work_completion)(&(&bond->mcast_work)->work) FD: 232 BD: 1 +.+.: (wq_completion)bond323 ->(work_completion)(&(&slave->notify_work)->work) ->(work_completion)(&(&bond->mcast_work)->work) FD: 232 BD: 1 +.+.: (wq_completion)bond298#3 ->(work_completion)(&(&slave->notify_work)->work) ->(work_completion)(&(&bond->mcast_work)->work) FD: 232 BD: 1 +.+.: (wq_completion)bond324 ->(work_completion)(&(&slave->notify_work)->work) ->(work_completion)(&(&bond->mcast_work)->work) FD: 232 BD: 1 +.+.: (wq_completion)bond299#3 ->&rq->__lock ->(work_completion)(&(&slave->notify_work)->work) ->(work_completion)(&(&bond->mcast_work)->work) FD: 232 BD: 1 +.+.: (wq_completion)bond313#2 ->(work_completion)(&(&slave->notify_work)->work) ->(work_completion)(&(&bond->mcast_work)->work) FD: 232 BD: 1 +.+.: (wq_completion)bond300#3 ->(work_completion)(&(&slave->notify_work)->work) ->(work_completion)(&(&bond->mcast_work)->work) FD: 232 BD: 1 +.+.: (wq_completion)bond325 ->&rq->__lock ->(work_completion)(&(&slave->notify_work)->work) ->(work_completion)(&(&bond->mcast_work)->work) FD: 232 BD: 1 +.+.: (wq_completion)bond314#2 ->(work_completion)(&(&slave->notify_work)->work) ->(work_completion)(&(&bond->mcast_work)->work) FD: 232 BD: 1 +.+.: (wq_completion)bond301#3 ->(work_completion)(&(&slave->notify_work)->work) ->(work_completion)(&(&bond->mcast_work)->work) FD: 232 BD: 1 +.+.: (wq_completion)bond302#3 ->&rq->__lock ->(work_completion)(&(&slave->notify_work)->work) ->(work_completion)(&(&bond->mcast_work)->work) FD: 232 BD: 1 +.+.: (wq_completion)bond326 ->(work_completion)(&(&slave->notify_work)->work) ->(work_completion)(&(&bond->mcast_work)->work) ->&rq->__lock FD: 232 BD: 1 +.+.: (wq_completion)bond315#2 ->(work_completion)(&(&slave->notify_work)->work) ->(work_completion)(&(&bond->mcast_work)->work) FD: 232 BD: 1 +.+.: (wq_completion)bond113#4 ->(work_completion)(&(&slave->notify_work)->work) ->(work_completion)(&(&bond->mcast_work)->work) FD: 232 BD: 1 +.+.: (wq_completion)bond316#2 ->(work_completion)(&(&slave->notify_work)->work) ->(work_completion)(&(&bond->mcast_work)->work) FD: 232 BD: 1 +.+.: (wq_completion)bond327 ->(work_completion)(&(&slave->notify_work)->work) ->(work_completion)(&(&bond->mcast_work)->work) FD: 232 BD: 1 +.+.: (wq_completion)bond303#3 ->(work_completion)(&(&slave->notify_work)->work) ->(work_completion)(&(&bond->mcast_work)->work) FD: 232 BD: 1 +.+.: (wq_completion)bond114#4 ->(work_completion)(&(&slave->notify_work)->work) ->(work_completion)(&(&bond->mcast_work)->work) FD: 232 BD: 1 +.+.: (wq_completion)bond317#2 ->(work_completion)(&(&slave->notify_work)->work) ->(work_completion)(&(&bond->mcast_work)->work) FD: 232 BD: 1 +.+.: (wq_completion)bond328 ->(work_completion)(&(&slave->notify_work)->work) ->(work_completion)(&(&bond->mcast_work)->work) FD: 232 BD: 1 +.+.: (wq_completion)bond304#3 ->(work_completion)(&(&slave->notify_work)->work) ->(work_completion)(&(&bond->mcast_work)->work) ->&rq->__lock FD: 232 BD: 1 +.+.: (wq_completion)bond318#2 ->(work_completion)(&(&slave->notify_work)->work) ->(work_completion)(&(&bond->mcast_work)->work) FD: 232 BD: 1 +.+.: (wq_completion)bond329 ->(work_completion)(&(&slave->notify_work)->work) ->(work_completion)(&(&bond->mcast_work)->work) FD: 232 BD: 1 +.+.: (wq_completion)bond305#3 ->(work_completion)(&(&slave->notify_work)->work) ->(work_completion)(&(&bond->mcast_work)->work) FD: 232 BD: 1 +.+.: (wq_completion)bond319#2 ->&rq->__lock ->(work_completion)(&(&slave->notify_work)->work) ->(work_completion)(&(&bond->mcast_work)->work) FD: 232 BD: 1 +.+.: (wq_completion)bond330 ->(work_completion)(&(&slave->notify_work)->work) ->(work_completion)(&(&bond->mcast_work)->work) FD: 232 BD: 1 +.+.: (wq_completion)bond306#3 ->(work_completion)(&(&slave->notify_work)->work) ->(work_completion)(&(&bond->mcast_work)->work) FD: 232 BD: 1 +.+.: (wq_completion)bond320#2 ->(work_completion)(&(&slave->notify_work)->work) ->(work_completion)(&(&bond->mcast_work)->work) ->&rq->__lock FD: 232 BD: 1 +.+.: (wq_completion)bond307#3 ->(work_completion)(&(&slave->notify_work)->work) ->(work_completion)(&(&bond->mcast_work)->work) FD: 232 BD: 1 +.+.: (wq_completion)bond331 ->(work_completion)(&(&slave->notify_work)->work) ->(work_completion)(&(&bond->mcast_work)->work) FD: 232 BD: 1 +.+.: (wq_completion)bond321#2 ->&rq->__lock ->(work_completion)(&(&slave->notify_work)->work) ->(work_completion)(&(&bond->mcast_work)->work) FD: 232 BD: 1 +.+.: (wq_completion)bond308#3 ->(work_completion)(&(&slave->notify_work)->work) ->(work_completion)(&(&bond->mcast_work)->work) FD: 232 BD: 1 +.+.: (wq_completion)bond332 ->(work_completion)(&(&slave->notify_work)->work) ->(work_completion)(&(&bond->mcast_work)->work) FD: 232 BD: 1 +.+.: (wq_completion)bond322#2 ->(work_completion)(&(&slave->notify_work)->work) ->(work_completion)(&(&bond->mcast_work)->work) FD: 232 BD: 1 +.+.: (wq_completion)bond309#3 ->(work_completion)(&(&slave->notify_work)->work) ->(work_completion)(&(&bond->mcast_work)->work) FD: 232 BD: 1 +.+.: (wq_completion)bond115#4 ->(work_completion)(&(&slave->notify_work)->work) ->(work_completion)(&(&bond->mcast_work)->work) FD: 232 BD: 1 +.+.: (wq_completion)bond323#2 ->(work_completion)(&(&slave->notify_work)->work) ->(work_completion)(&(&bond->mcast_work)->work) ->&rq->__lock FD: 232 BD: 1 +.+.: (wq_completion)bond116#4 ->(work_completion)(&(&slave->notify_work)->work) ->(work_completion)(&(&bond->mcast_work)->work) FD: 232 BD: 1 +.+.: (wq_completion)bond333 ->(work_completion)(&(&slave->notify_work)->work) ->(work_completion)(&(&bond->mcast_work)->work) FD: 232 BD: 1 +.+.: (wq_completion)bond311#3 ->(work_completion)(&(&slave->notify_work)->work) ->(work_completion)(&(&bond->mcast_work)->work) FD: 232 BD: 1 +.+.: (wq_completion)bond324#2 ->&rq->__lock ->(work_completion)(&(&slave->notify_work)->work) ->(work_completion)(&(&bond->mcast_work)->work) FD: 232 BD: 1 +.+.: (wq_completion)bond334 ->&rq->__lock ->(work_completion)(&(&slave->notify_work)->work) ->(work_completion)(&(&bond->mcast_work)->work) FD: 232 BD: 1 +.+.: (wq_completion)bond117#4 ->(work_completion)(&(&slave->notify_work)->work) ->(work_completion)(&(&bond->mcast_work)->work) FD: 232 BD: 1 +.+.: (wq_completion)bond312#3 ->&rq->__lock ->(work_completion)(&(&slave->notify_work)->work) ->(work_completion)(&(&bond->mcast_work)->work) FD: 232 BD: 1 +.+.: (wq_completion)bond335 ->(work_completion)(&(&slave->notify_work)->work) ->(work_completion)(&(&bond->mcast_work)->work) FD: 232 BD: 1 +.+.: (wq_completion)bond325#2 ->&rq->__lock ->(work_completion)(&(&slave->notify_work)->work) ->(work_completion)(&(&bond->mcast_work)->work) FD: 232 BD: 1 +.+.: (wq_completion)bond118#4 ->(work_completion)(&(&slave->notify_work)->work) ->(work_completion)(&(&bond->mcast_work)->work) FD: 232 BD: 1 +.+.: (wq_completion)bond313#3 ->(work_completion)(&(&slave->notify_work)->work) ->(work_completion)(&(&bond->mcast_work)->work) FD: 232 BD: 1 +.+.: (wq_completion)bond336 ->(work_completion)(&(&slave->notify_work)->work) ->(work_completion)(&(&bond->mcast_work)->work) FD: 232 BD: 1 +.+.: (wq_completion)bond326#2 ->(work_completion)(&(&slave->notify_work)->work) ->(work_completion)(&(&bond->mcast_work)->work) FD: 232 BD: 1 +.+.: (wq_completion)bond119#4 ->(work_completion)(&(&slave->notify_work)->work) ->(work_completion)(&(&bond->mcast_work)->work) FD: 232 BD: 1 +.+.: (wq_completion)bond314#3 ->(work_completion)(&(&slave->notify_work)->work) ->(work_completion)(&(&bond->mcast_work)->work) FD: 232 BD: 1 +.+.: (wq_completion)bond337 ->(work_completion)(&(&slave->notify_work)->work) ->(work_completion)(&(&bond->mcast_work)->work) ->&rq->__lock FD: 232 BD: 1 +.+.: (wq_completion)bond327#2 ->&rq->__lock ->(work_completion)(&(&slave->notify_work)->work) ->(work_completion)(&(&bond->mcast_work)->work) FD: 232 BD: 1 +.+.: (wq_completion)bond315#3 ->(work_completion)(&(&slave->notify_work)->work) ->(work_completion)(&(&bond->mcast_work)->work) FD: 232 BD: 1 +.+.: (wq_completion)bond338 ->&rq->__lock ->(work_completion)(&(&slave->notify_work)->work) ->(work_completion)(&(&bond->mcast_work)->work) FD: 232 BD: 1 +.+.: (wq_completion)bond120#4 ->(work_completion)(&(&slave->notify_work)->work) ->(work_completion)(&(&bond->mcast_work)->work) FD: 232 BD: 1 +.+.: (wq_completion)bond328#2 ->(work_completion)(&(&slave->notify_work)->work) ->(work_completion)(&(&bond->mcast_work)->work) FD: 232 BD: 1 +.+.: (wq_completion)bond316#3 ->(work_completion)(&(&slave->notify_work)->work) ->(work_completion)(&(&bond->mcast_work)->work) FD: 232 BD: 1 +.+.: (wq_completion)bond339 ->(work_completion)(&(&slave->notify_work)->work) ->(work_completion)(&(&bond->mcast_work)->work) FD: 232 BD: 1 +.+.: (wq_completion)bond329#2 ->(work_completion)(&(&slave->notify_work)->work) ->(work_completion)(&(&bond->mcast_work)->work) ->&rq->__lock FD: 232 BD: 1 +.+.: (wq_completion)bond340 ->(work_completion)(&(&slave->notify_work)->work) ->(work_completion)(&(&bond->mcast_work)->work) FD: 232 BD: 1 +.+.: (wq_completion)bond317#3 ->(work_completion)(&(&slave->notify_work)->work) ->(work_completion)(&(&bond->mcast_work)->work) FD: 232 BD: 1 +.+.: (wq_completion)bond330#2 ->(work_completion)(&(&slave->notify_work)->work) ->(work_completion)(&(&bond->mcast_work)->work) FD: 232 BD: 1 +.+.: (wq_completion)bond318#3 ->(work_completion)(&(&slave->notify_work)->work) ->(work_completion)(&(&bond->mcast_work)->work) FD: 232 BD: 1 +.+.: (wq_completion)bond341 ->(work_completion)(&(&slave->notify_work)->work) ->(work_completion)(&(&bond->mcast_work)->work) FD: 232 BD: 1 +.+.: (wq_completion)bond121#4 ->(work_completion)(&(&slave->notify_work)->work) ->(work_completion)(&(&bond->mcast_work)->work) FD: 232 BD: 1 +.+.: (wq_completion)bond331#2 ->&rq->__lock ->(work_completion)(&(&slave->notify_work)->work) ->(work_completion)(&(&bond->mcast_work)->work) FD: 232 BD: 1 +.+.: (wq_completion)bond319#3 ->&rq->__lock ->(work_completion)(&(&slave->notify_work)->work) ->(work_completion)(&(&bond->mcast_work)->work) FD: 232 BD: 1 +.+.: (wq_completion)bond342 ->(work_completion)(&(&slave->notify_work)->work) ->(work_completion)(&(&bond->mcast_work)->work) FD: 232 BD: 1 +.+.: (wq_completion)bond332#2 ->(work_completion)(&(&slave->notify_work)->work) ->(work_completion)(&(&bond->mcast_work)->work) FD: 232 BD: 1 +.+.: (wq_completion)bond320#3 ->(work_completion)(&(&slave->notify_work)->work) ->&rq->__lock ->(work_completion)(&(&bond->mcast_work)->work) FD: 232 BD: 1 +.+.: (wq_completion)bond343 ->(work_completion)(&(&slave->notify_work)->work) ->(work_completion)(&(&bond->mcast_work)->work) FD: 232 BD: 1 +.+.: (wq_completion)bond333#2 ->&rq->__lock ->(work_completion)(&(&slave->notify_work)->work) ->(work_completion)(&(&bond->mcast_work)->work) FD: 232 BD: 1 +.+.: (wq_completion)bond321#3 ->&rq->__lock ->(work_completion)(&(&slave->notify_work)->work) ->(work_completion)(&(&bond->mcast_work)->work) FD: 232 BD: 1 +.+.: (wq_completion)bond344 ->(work_completion)(&(&slave->notify_work)->work) ->(work_completion)(&(&bond->mcast_work)->work) FD: 232 BD: 1 +.+.: (wq_completion)bond335#2 ->(work_completion)(&(&slave->notify_work)->work) ->(work_completion)(&(&bond->mcast_work)->work) FD: 232 BD: 1 +.+.: (wq_completion)bond322#3 ->(work_completion)(&(&slave->notify_work)->work) ->(work_completion)(&(&bond->mcast_work)->work) FD: 232 BD: 1 +.+.: (wq_completion)bond345 ->(work_completion)(&(&slave->notify_work)->work) ->(work_completion)(&(&bond->mcast_work)->work) FD: 232 BD: 1 +.+.: (wq_completion)bond336#2 ->(work_completion)(&(&slave->notify_work)->work) ->(work_completion)(&(&bond->mcast_work)->work) FD: 232 BD: 1 +.+.: (wq_completion)bond323#3 ->(work_completion)(&(&slave->notify_work)->work) ->(work_completion)(&(&bond->mcast_work)->work) ->&rq->__lock FD: 232 BD: 1 +.+.: (wq_completion)bond346 ->(work_completion)(&(&slave->notify_work)->work) ->(work_completion)(&(&bond->mcast_work)->work) FD: 232 BD: 1 +.+.: (wq_completion)bond337#2 ->(work_completion)(&(&slave->notify_work)->work) ->(work_completion)(&(&bond->mcast_work)->work) FD: 232 BD: 1 +.+.: (wq_completion)bond324#3 ->(work_completion)(&(&slave->notify_work)->work) ->(work_completion)(&(&bond->mcast_work)->work) FD: 232 BD: 1 +.+.: (wq_completion)bond338#2 ->(work_completion)(&(&slave->notify_work)->work) ->(work_completion)(&(&bond->mcast_work)->work) FD: 232 BD: 1 +.+.: (wq_completion)bond348 ->&rq->__lock ->(work_completion)(&(&slave->notify_work)->work) ->(work_completion)(&(&bond->mcast_work)->work) FD: 232 BD: 1 +.+.: (wq_completion)bond325#3 ->(work_completion)(&(&slave->notify_work)->work) ->(work_completion)(&(&bond->mcast_work)->work) FD: 232 BD: 1 +.+.: (wq_completion)bond122#4 ->(work_completion)(&(&slave->notify_work)->work) ->(work_completion)(&(&bond->mcast_work)->work) FD: 232 BD: 1 +.+.: (wq_completion)bond339#2 ->(work_completion)(&(&slave->notify_work)->work) ->(work_completion)(&(&bond->mcast_work)->work) FD: 232 BD: 1 +.+.: (wq_completion)bond349 ->(work_completion)(&(&slave->notify_work)->work) ->(work_completion)(&(&bond->mcast_work)->work) ->&rq->__lock FD: 232 BD: 1 +.+.: (wq_completion)bond326#3 ->(work_completion)(&(&slave->notify_work)->work) ->(work_completion)(&(&bond->mcast_work)->work) ->&rq->__lock FD: 232 BD: 1 +.+.: (wq_completion)bond340#2 ->(work_completion)(&(&slave->notify_work)->work) ->(work_completion)(&(&bond->mcast_work)->work) ->&rq->__lock FD: 232 BD: 1 +.+.: (wq_completion)bond350 ->(work_completion)(&(&slave->notify_work)->work) ->(work_completion)(&(&bond->mcast_work)->work) FD: 232 BD: 1 +.+.: (wq_completion)bond327#3 ->(work_completion)(&(&slave->notify_work)->work) ->(work_completion)(&(&bond->mcast_work)->work) ->&rq->__lock FD: 232 BD: 1 +.+.: (wq_completion)bond341#2 ->(work_completion)(&(&slave->notify_work)->work) ->(work_completion)(&(&bond->mcast_work)->work) ->&rq->__lock FD: 232 BD: 1 +.+.: (wq_completion)bond351 ->(work_completion)(&(&slave->notify_work)->work) ->(work_completion)(&(&bond->mcast_work)->work) FD: 232 BD: 1 +.+.: (wq_completion)bond328#3 ->(work_completion)(&(&slave->notify_work)->work) ->(work_completion)(&(&bond->mcast_work)->work) FD: 232 BD: 1 +.+.: (wq_completion)bond342#2 ->(work_completion)(&(&slave->notify_work)->work) ->(work_completion)(&(&bond->mcast_work)->work) FD: 232 BD: 1 +.+.: (wq_completion)bond352 ->(work_completion)(&(&slave->notify_work)->work) ->(work_completion)(&(&bond->mcast_work)->work) ->&rq->__lock FD: 232 BD: 1 +.+.: (wq_completion)bond329#3 ->&rq->__lock ->(work_completion)(&(&slave->notify_work)->work) ->(work_completion)(&(&bond->mcast_work)->work) FD: 232 BD: 1 +.+.: (wq_completion)bond343#2 ->(work_completion)(&(&slave->notify_work)->work) ->(work_completion)(&(&bond->mcast_work)->work) ->&rq->__lock FD: 232 BD: 1 +.+.: (wq_completion)bond123#4 ->(work_completion)(&(&slave->notify_work)->work) ->(work_completion)(&(&bond->mcast_work)->work) FD: 232 BD: 1 +.+.: (wq_completion)bond330#3 ->(work_completion)(&(&slave->notify_work)->work) ->(work_completion)(&(&bond->mcast_work)->work) ->&rq->__lock FD: 232 BD: 1 +.+.: (wq_completion)bond344#2 ->(work_completion)(&(&slave->notify_work)->work) ->(work_completion)(&(&bond->mcast_work)->work) ->&rq->__lock FD: 232 BD: 1 +.+.: (wq_completion)bond354 ->&rq->__lock ->(work_completion)(&(&slave->notify_work)->work) ->(work_completion)(&(&bond->mcast_work)->work) FD: 232 BD: 1 +.+.: (wq_completion)bond124#4 ->(work_completion)(&(&slave->notify_work)->work) ->(work_completion)(&(&bond->mcast_work)->work) ->&rq->__lock FD: 232 BD: 1 +.+.: (wq_completion)bond331#3 ->&rq->__lock ->(work_completion)(&(&slave->notify_work)->work) ->(work_completion)(&(&bond->mcast_work)->work) FD: 232 BD: 1 +.+.: (wq_completion)bond345#2 ->(work_completion)(&(&slave->notify_work)->work) ->(work_completion)(&(&bond->mcast_work)->work) ->&rq->__lock FD: 232 BD: 1 +.+.: (wq_completion)bond355 ->(work_completion)(&(&slave->notify_work)->work) ->(work_completion)(&(&bond->mcast_work)->work) FD: 232 BD: 1 +.+.: (wq_completion)bond125#4 ->&rq->__lock ->(work_completion)(&(&slave->notify_work)->work) ->(work_completion)(&(&bond->mcast_work)->work) FD: 232 BD: 1 +.+.: (wq_completion)bond332#3 ->&rq->__lock ->(work_completion)(&(&slave->notify_work)->work) ->(work_completion)(&(&bond->mcast_work)->work) FD: 232 BD: 1 +.+.: (wq_completion)bond346#2 ->&rq->__lock ->(work_completion)(&(&slave->notify_work)->work) ->(work_completion)(&(&bond->mcast_work)->work) FD: 232 BD: 1 +.+.: (wq_completion)bond356 ->(work_completion)(&(&slave->notify_work)->work) ->(work_completion)(&(&bond->mcast_work)->work) ->&rq->__lock FD: 232 BD: 1 +.+.: (wq_completion)bond126#4 ->&rq->__lock ->(work_completion)(&(&slave->notify_work)->work) ->(work_completion)(&(&bond->mcast_work)->work) FD: 232 BD: 1 +.+.: (wq_completion)bond333#3 ->(work_completion)(&(&slave->notify_work)->work) ->(work_completion)(&(&bond->mcast_work)->work) ->&rq->__lock FD: 232 BD: 1 +.+.: (wq_completion)bond347 ->(work_completion)(&(&slave->notify_work)->work) ->(work_completion)(&(&bond->mcast_work)->work) ->&rq->__lock FD: 232 BD: 1 +.+.: (wq_completion)bond357 ->&rq->__lock ->(work_completion)(&(&slave->notify_work)->work) ->(work_completion)(&(&bond->mcast_work)->work) FD: 232 BD: 1 +.+.: (wq_completion)bond334#2 ->(work_completion)(&(&slave->notify_work)->work) ->(work_completion)(&(&bond->mcast_work)->work) FD: 232 BD: 1 +.+.: (wq_completion)bond348#2 ->&rq->__lock ->(work_completion)(&(&slave->notify_work)->work) ->(work_completion)(&(&bond->mcast_work)->work) FD: 232 BD: 1 +.+.: (wq_completion)bond358 ->(work_completion)(&(&slave->notify_work)->work) ->(work_completion)(&(&bond->mcast_work)->work) FD: 232 BD: 1 +.+.: (wq_completion)bond335#3 ->(work_completion)(&(&slave->notify_work)->work) ->(work_completion)(&(&bond->mcast_work)->work) FD: 232 BD: 1 +.+.: (wq_completion)bond349#2 ->(work_completion)(&(&slave->notify_work)->work) ->(work_completion)(&(&bond->mcast_work)->work) ->&rq->__lock FD: 232 BD: 1 +.+.: (wq_completion)bond359 ->(work_completion)(&(&slave->notify_work)->work) ->(work_completion)(&(&bond->mcast_work)->work) FD: 232 BD: 1 +.+.: (wq_completion)bond350#2 ->(work_completion)(&(&slave->notify_work)->work) ->(work_completion)(&(&bond->mcast_work)->work) FD: 232 BD: 1 +.+.: (wq_completion)bond336#3 ->&rq->__lock ->(work_completion)(&(&slave->notify_work)->work) ->(work_completion)(&(&bond->mcast_work)->work) FD: 232 BD: 1 +.+.: (wq_completion)bond360 ->&rq->__lock ->(work_completion)(&(&slave->notify_work)->work) ->(work_completion)(&(&bond->mcast_work)->work) FD: 232 BD: 1 +.+.: (wq_completion)bond351#2 ->(work_completion)(&(&slave->notify_work)->work) ->(work_completion)(&(&bond->mcast_work)->work) ->&rq->__lock FD: 232 BD: 1 +.+.: (wq_completion)bond337#3 ->(work_completion)(&(&slave->notify_work)->work) ->(work_completion)(&(&bond->mcast_work)->work) ->&rq->__lock FD: 232 BD: 1 +.+.: (wq_completion)bond361 ->(work_completion)(&(&slave->notify_work)->work) ->(work_completion)(&(&bond->mcast_work)->work) FD: 232 BD: 1 +.+.: (wq_completion)bond352#2 ->&rq->__lock ->(work_completion)(&(&slave->notify_work)->work) ->(work_completion)(&(&bond->mcast_work)->work) FD: 232 BD: 1 +.+.: (wq_completion)bond338#3 ->(work_completion)(&(&slave->notify_work)->work) ->(work_completion)(&(&bond->mcast_work)->work) ->&rq->__lock FD: 232 BD: 1 +.+.: (wq_completion)bond362 ->(work_completion)(&(&slave->notify_work)->work) ->(work_completion)(&(&bond->mcast_work)->work) ->&rq->__lock FD: 232 BD: 1 +.+.: (wq_completion)bond353 ->(work_completion)(&(&slave->notify_work)->work) ->(work_completion)(&(&bond->mcast_work)->work) ->&rq->__lock FD: 232 BD: 1 +.+.: (wq_completion)bond339#3 ->(work_completion)(&(&slave->notify_work)->work) ->(work_completion)(&(&bond->mcast_work)->work) ->&rq->__lock FD: 232 BD: 1 +.+.: (wq_completion)bond127#4 ->&rq->__lock ->(work_completion)(&(&slave->notify_work)->work) ->(work_completion)(&(&bond->mcast_work)->work) FD: 232 BD: 1 +.+.: (wq_completion)bond363 ->(work_completion)(&(&slave->notify_work)->work) ->(work_completion)(&(&bond->mcast_work)->work) FD: 232 BD: 1 +.+.: (wq_completion)bond354#2 ->(work_completion)(&(&slave->notify_work)->work) ->(work_completion)(&(&bond->mcast_work)->work) ->&rq->__lock FD: 232 BD: 1 +.+.: (wq_completion)bond340#3 ->(work_completion)(&(&slave->notify_work)->work) ->(work_completion)(&(&bond->mcast_work)->work) FD: 232 BD: 1 +.+.: (wq_completion)bond364 ->&rq->__lock ->(work_completion)(&(&slave->notify_work)->work) ->(work_completion)(&(&bond->mcast_work)->work) FD: 232 BD: 1 +.+.: (wq_completion)bond355#2 ->&rq->__lock ->(work_completion)(&(&slave->notify_work)->work) ->(work_completion)(&(&bond->mcast_work)->work) FD: 232 BD: 1 +.+.: (wq_completion)bond341#3 ->&rq->__lock ->(work_completion)(&(&slave->notify_work)->work) ->(work_completion)(&(&bond->mcast_work)->work) FD: 232 BD: 1 +.+.: (wq_completion)bond128#4 ->(work_completion)(&(&slave->notify_work)->work) ->(work_completion)(&(&bond->mcast_work)->work) ->&rq->__lock FD: 232 BD: 1 +.+.: (wq_completion)bond365 ->(work_completion)(&(&slave->notify_work)->work) ->(work_completion)(&(&bond->mcast_work)->work) FD: 232 BD: 1 +.+.: (wq_completion)bond356#2 ->&rq->__lock ->(work_completion)(&(&slave->notify_work)->work) ->(work_completion)(&(&bond->mcast_work)->work) FD: 232 BD: 1 +.+.: (wq_completion)bond342#3 ->&rq->__lock ->(work_completion)(&(&slave->notify_work)->work) ->(work_completion)(&(&bond->mcast_work)->work) FD: 232 BD: 1 +.+.: (wq_completion)bond366 ->(work_completion)(&(&slave->notify_work)->work) ->(work_completion)(&(&bond->mcast_work)->work) ->&rq->__lock FD: 232 BD: 1 +.+.: (wq_completion)bond129#4 ->&rq->__lock ->(work_completion)(&(&slave->notify_work)->work) ->(work_completion)(&(&bond->mcast_work)->work) FD: 232 BD: 1 +.+.: (wq_completion)bond357#2 ->&rq->__lock ->(work_completion)(&(&slave->notify_work)->work) ->(work_completion)(&(&bond->mcast_work)->work) FD: 232 BD: 1 +.+.: (wq_completion)bond343#3 ->&rq->__lock ->(work_completion)(&(&slave->notify_work)->work) ->(work_completion)(&(&bond->mcast_work)->work) FD: 232 BD: 1 +.+.: (wq_completion)bond367 ->&rq->__lock ->(work_completion)(&(&slave->notify_work)->work) ->(work_completion)(&(&bond->mcast_work)->work) FD: 232 BD: 1 +.+.: (wq_completion)bond344#3 ->(work_completion)(&(&slave->notify_work)->work) ->(work_completion)(&(&bond->mcast_work)->work) FD: 232 BD: 1 +.+.: (wq_completion)bond358#2 ->&rq->__lock ->(work_completion)(&(&slave->notify_work)->work) ->(work_completion)(&(&bond->mcast_work)->work) FD: 232 BD: 1 +.+.: (wq_completion)bond368 ->(work_completion)(&(&slave->notify_work)->work) ->(work_completion)(&(&bond->mcast_work)->work) ->&rq->__lock FD: 232 BD: 1 +.+.: (wq_completion)bond345#3 ->(work_completion)(&(&slave->notify_work)->work) ->(work_completion)(&(&bond->mcast_work)->work) FD: 232 BD: 1 +.+.: (wq_completion)bond359#2 ->(work_completion)(&(&slave->notify_work)->work) ->(work_completion)(&(&bond->mcast_work)->work) ->&rq->__lock FD: 232 BD: 1 +.+.: (wq_completion)bond369 ->&rq->__lock ->(work_completion)(&(&slave->notify_work)->work) ->(work_completion)(&(&bond->mcast_work)->work) FD: 232 BD: 1 +.+.: (wq_completion)bond346#3 ->&rq->__lock ->(work_completion)(&(&slave->notify_work)->work) ->(work_completion)(&(&bond->mcast_work)->work) FD: 232 BD: 1 +.+.: (wq_completion)bond360#2 ->(work_completion)(&(&slave->notify_work)->work) ->(work_completion)(&(&bond->mcast_work)->work) ->&rq->__lock FD: 232 BD: 1 +.+.: (wq_completion)bond130#4 ->&rq->__lock ->(work_completion)(&(&slave->notify_work)->work) ->(work_completion)(&(&bond->mcast_work)->work) FD: 232 BD: 1 +.+.: (wq_completion)bond370 ->(work_completion)(&(&slave->notify_work)->work) ->(work_completion)(&(&bond->mcast_work)->work) ->&rq->__lock FD: 232 BD: 1 +.+.: (wq_completion)bond347#2 ->(work_completion)(&(&slave->notify_work)->work) ->(work_completion)(&(&bond->mcast_work)->work) FD: 232 BD: 1 +.+.: (wq_completion)bond361#2 ->(work_completion)(&(&slave->notify_work)->work) ->(work_completion)(&(&bond->mcast_work)->work) FD: 232 BD: 1 +.+.: (wq_completion)bond371 ->(work_completion)(&(&slave->notify_work)->work) ->(work_completion)(&(&bond->mcast_work)->work) FD: 232 BD: 1 +.+.: (wq_completion)bond348#3 ->&rq->__lock ->(work_completion)(&(&slave->notify_work)->work) ->(work_completion)(&(&bond->mcast_work)->work) FD: 232 BD: 1 +.+.: (wq_completion)bond362#2 ->(work_completion)(&(&slave->notify_work)->work) ->(work_completion)(&(&bond->mcast_work)->work) ->&rq->__lock FD: 232 BD: 1 +.+.: (wq_completion)bond372 ->&rq->__lock ->(work_completion)(&(&slave->notify_work)->work) ->(work_completion)(&(&bond->mcast_work)->work) FD: 232 BD: 1 +.+.: (wq_completion)bond349#3 ->&rq->__lock ->(work_completion)(&(&slave->notify_work)->work) ->(work_completion)(&(&bond->mcast_work)->work) FD: 232 BD: 1 +.+.: (wq_completion)bond132#4 ->&rq->__lock ->(work_completion)(&(&slave->notify_work)->work) ->(work_completion)(&(&bond->mcast_work)->work) FD: 232 BD: 1 +.+.: (wq_completion)bond363#2 ->(work_completion)(&(&slave->notify_work)->work) ->(work_completion)(&(&bond->mcast_work)->work) ->&rq->__lock FD: 232 BD: 1 +.+.: (wq_completion)bond373 ->&rq->__lock ->(work_completion)(&(&slave->notify_work)->work) ->(work_completion)(&(&bond->mcast_work)->work) FD: 232 BD: 1 +.+.: (wq_completion)bond350#3 ->(work_completion)(&(&slave->notify_work)->work) ->(work_completion)(&(&bond->mcast_work)->work) ->&rq->__lock FD: 232 BD: 1 +.+.: (wq_completion)bond364#2 ->&rq->__lock ->(work_completion)(&(&slave->notify_work)->work) ->(work_completion)(&(&bond->mcast_work)->work) FD: 232 BD: 1 +.+.: (wq_completion)bond374 ->&rq->__lock ->(work_completion)(&(&slave->notify_work)->work) ->(work_completion)(&(&bond->mcast_work)->work) FD: 232 BD: 1 +.+.: (wq_completion)bond351#3 ->(work_completion)(&(&slave->notify_work)->work) ->(work_completion)(&(&bond->mcast_work)->work) FD: 232 BD: 1 +.+.: (wq_completion)bond365#2 ->(work_completion)(&(&slave->notify_work)->work) ->(work_completion)(&(&bond->mcast_work)->work) ->&rq->__lock FD: 232 BD: 1 +.+.: (wq_completion)bond375 ->(work_completion)(&(&slave->notify_work)->work) ->(work_completion)(&(&bond->mcast_work)->work) ->&rq->__lock FD: 232 BD: 1 +.+.: (wq_completion)bond352#3 ->(work_completion)(&(&slave->notify_work)->work) ->(work_completion)(&(&bond->mcast_work)->work) FD: 232 BD: 1 +.+.: (wq_completion)bond366#2 ->&rq->__lock ->(work_completion)(&(&slave->notify_work)->work) ->(work_completion)(&(&bond->mcast_work)->work) FD: 232 BD: 1 +.+.: (wq_completion)bond376 ->(work_completion)(&(&slave->notify_work)->work) ->(work_completion)(&(&bond->mcast_work)->work) ->&rq->__lock FD: 232 BD: 1 +.+.: (wq_completion)bond353#2 ->&rq->__lock ->(work_completion)(&(&slave->notify_work)->work) ->(work_completion)(&(&bond->mcast_work)->work) FD: 232 BD: 1 +.+.: (wq_completion)bond367#2 ->(work_completion)(&(&slave->notify_work)->work) ->(work_completion)(&(&bond->mcast_work)->work) ->&rq->__lock FD: 232 BD: 1 +.+.: (wq_completion)bond377 ->(work_completion)(&(&slave->notify_work)->work) ->(work_completion)(&(&bond->mcast_work)->work) ->&rq->__lock FD: 232 BD: 1 +.+.: (wq_completion)bond354#3 ->&rq->__lock ->(work_completion)(&(&slave->notify_work)->work) ->(work_completion)(&(&bond->mcast_work)->work) FD: 232 BD: 1 +.+.: (wq_completion)bond368#2 ->(work_completion)(&(&slave->notify_work)->work) ->(work_completion)(&(&bond->mcast_work)->work) FD: 232 BD: 1 +.+.: (wq_completion)bond378 ->&rq->__lock ->(work_completion)(&(&slave->notify_work)->work) ->(work_completion)(&(&bond->mcast_work)->work) FD: 232 BD: 1 +.+.: (wq_completion)bond355#3 ->&rq->__lock ->(work_completion)(&(&slave->notify_work)->work) ->(work_completion)(&(&bond->mcast_work)->work) FD: 232 BD: 1 +.+.: (wq_completion)bond369#2 ->(work_completion)(&(&slave->notify_work)->work) ->(work_completion)(&(&bond->mcast_work)->work) FD: 232 BD: 1 +.+.: (wq_completion)bond370#2 ->(work_completion)(&(&slave->notify_work)->work) ->(work_completion)(&(&bond->mcast_work)->work) ->&rq->__lock FD: 232 BD: 1 +.+.: (wq_completion)bond379 ->&rq->__lock ->(work_completion)(&(&slave->notify_work)->work) ->(work_completion)(&(&bond->mcast_work)->work) FD: 232 BD: 1 +.+.: (wq_completion)bond356#3 ->&rq->__lock ->(work_completion)(&(&slave->notify_work)->work) ->(work_completion)(&(&bond->mcast_work)->work) FD: 232 BD: 1 +.+.: (wq_completion)bond371#2 ->(work_completion)(&(&slave->notify_work)->work) ->(work_completion)(&(&bond->mcast_work)->work) ->&rq->__lock FD: 232 BD: 1 +.+.: (wq_completion)bond380 ->(work_completion)(&(&slave->notify_work)->work) ->(work_completion)(&(&bond->mcast_work)->work) FD: 232 BD: 1 +.+.: (wq_completion)bond357#3 ->(work_completion)(&(&slave->notify_work)->work) ->(work_completion)(&(&bond->mcast_work)->work) ->&rq->__lock FD: 232 BD: 1 +.+.: (wq_completion)bond372#2 ->(work_completion)(&(&slave->notify_work)->work) ->(work_completion)(&(&bond->mcast_work)->work) ->&rq->__lock FD: 232 BD: 1 +.+.: (wq_completion)bond358#3 ->&rq->__lock ->(work_completion)(&(&slave->notify_work)->work) ->(work_completion)(&(&bond->mcast_work)->work) FD: 232 BD: 1 +.+.: (wq_completion)bond381 ->(work_completion)(&(&slave->notify_work)->work) ->(work_completion)(&(&bond->mcast_work)->work) FD: 232 BD: 1 +.+.: (wq_completion)bond373#2 ->&rq->__lock ->(work_completion)(&(&slave->notify_work)->work) ->(work_completion)(&(&bond->mcast_work)->work) FD: 232 BD: 1 +.+.: (wq_completion)bond359#3 ->(work_completion)(&(&slave->notify_work)->work) ->(work_completion)(&(&bond->mcast_work)->work) FD: 232 BD: 1 +.+.: (wq_completion)bond382 ->(work_completion)(&(&slave->notify_work)->work) ->(work_completion)(&(&bond->mcast_work)->work) FD: 232 BD: 1 +.+.: (wq_completion)bond374#2 ->&rq->__lock ->(work_completion)(&(&slave->notify_work)->work) ->(work_completion)(&(&bond->mcast_work)->work) FD: 232 BD: 1 +.+.: (wq_completion)bond133#4 ->(work_completion)(&(&slave->notify_work)->work) ->(work_completion)(&(&bond->mcast_work)->work) ->&rq->__lock FD: 232 BD: 1 +.+.: (wq_completion)bond360#3 ->(work_completion)(&(&slave->notify_work)->work) ->(work_completion)(&(&bond->mcast_work)->work) FD: 232 BD: 1 +.+.: (wq_completion)bond383 ->&rq->__lock ->(work_completion)(&(&slave->notify_work)->work) ->(work_completion)(&(&bond->mcast_work)->work) FD: 127 BD: 5 +.+.: nlk_cb_mutex-NETFILTER ->&rq->__lock ->pool_lock#2 ->fs_reclaim ->&obj_hash[i].lock ->rlock-AF_NETLINK ->&c->lock ->&n->list_lock ->&____s->seqcount#2 ->&____s->seqcount FD: 1 BD: 6 ++..: ip_set_ref_lock FD: 232 BD: 1 +.+.: (wq_completion)bond375#2 ->(work_completion)(&(&slave->notify_work)->work) ->(work_completion)(&(&bond->mcast_work)->work) ->&rq->__lock FD: 232 BD: 1 +.+.: (wq_completion)bond361#3 ->(work_completion)(&(&slave->notify_work)->work) ->(work_completion)(&(&bond->mcast_work)->work) ->&rq->__lock FD: 232 BD: 1 +.+.: (wq_completion)bond384 ->(work_completion)(&(&slave->notify_work)->work) ->(work_completion)(&(&bond->mcast_work)->work) ->&rq->__lock FD: 232 BD: 1 +.+.: (wq_completion)bond376#2 ->(work_completion)(&(&slave->notify_work)->work) ->(work_completion)(&(&bond->mcast_work)->work) ->&rq->__lock FD: 232 BD: 1 +.+.: (wq_completion)bond362#3 ->(work_completion)(&(&slave->notify_work)->work) ->(work_completion)(&(&bond->mcast_work)->work) ->&rq->__lock FD: 232 BD: 1 +.+.: (wq_completion)bond385 ->(work_completion)(&(&slave->notify_work)->work) ->(work_completion)(&(&bond->mcast_work)->work) ->&rq->__lock FD: 232 BD: 1 +.+.: (wq_completion)bond377#2 ->(work_completion)(&(&slave->notify_work)->work) ->(work_completion)(&(&bond->mcast_work)->work) ->&rq->__lock FD: 232 BD: 1 +.+.: (wq_completion)bond363#3 ->(work_completion)(&(&slave->notify_work)->work) ->(work_completion)(&(&bond->mcast_work)->work) ->&rq->__lock FD: 232 BD: 1 +.+.: (wq_completion)bond386 ->(work_completion)(&(&slave->notify_work)->work) ->(work_completion)(&(&bond->mcast_work)->work) ->&rq->__lock FD: 232 BD: 1 +.+.: (wq_completion)bond378#2 ->(work_completion)(&(&slave->notify_work)->work) ->(work_completion)(&(&bond->mcast_work)->work) ->&rq->__lock FD: 232 BD: 1 +.+.: (wq_completion)bond364#3 ->(work_completion)(&(&slave->notify_work)->work) ->(work_completion)(&(&bond->mcast_work)->work) FD: 232 BD: 1 +.+.: (wq_completion)bond387 ->(work_completion)(&(&slave->notify_work)->work) ->(work_completion)(&(&bond->mcast_work)->work) ->&rq->__lock FD: 232 BD: 1 +.+.: (wq_completion)bond379#2 ->(work_completion)(&(&slave->notify_work)->work) ->(work_completion)(&(&bond->mcast_work)->work) FD: 232 BD: 1 +.+.: (wq_completion)bond365#3 ->&rq->__lock ->(work_completion)(&(&slave->notify_work)->work) ->(work_completion)(&(&bond->mcast_work)->work) FD: 232 BD: 1 +.+.: (wq_completion)bond388 ->(work_completion)(&(&slave->notify_work)->work) ->(work_completion)(&(&bond->mcast_work)->work) ->&rq->__lock FD: 232 BD: 1 +.+.: (wq_completion)bond380#2 ->&rq->__lock ->(work_completion)(&(&slave->notify_work)->work) ->(work_completion)(&(&bond->mcast_work)->work) FD: 232 BD: 1 +.+.: (wq_completion)bond366#3 ->(work_completion)(&(&slave->notify_work)->work) ->(work_completion)(&(&bond->mcast_work)->work) FD: 232 BD: 1 +.+.: (wq_completion)bond389 ->&rq->__lock ->(work_completion)(&(&slave->notify_work)->work) ->(work_completion)(&(&bond->mcast_work)->work) FD: 232 BD: 1 +.+.: (wq_completion)bond381#2 ->&rq->__lock ->(work_completion)(&(&slave->notify_work)->work) ->(work_completion)(&(&bond->mcast_work)->work) FD: 232 BD: 1 +.+.: (wq_completion)bond367#3 ->(work_completion)(&(&slave->notify_work)->work) ->(work_completion)(&(&bond->mcast_work)->work) ->&rq->__lock FD: 232 BD: 1 +.+.: (wq_completion)bond390 ->(work_completion)(&(&slave->notify_work)->work) ->(work_completion)(&(&bond->mcast_work)->work) ->&rq->__lock FD: 232 BD: 1 +.+.: (wq_completion)bond382#2 ->(work_completion)(&(&slave->notify_work)->work) ->(work_completion)(&(&bond->mcast_work)->work) FD: 232 BD: 1 +.+.: (wq_completion)bond368#3 ->&rq->__lock ->(work_completion)(&(&slave->notify_work)->work) ->(work_completion)(&(&bond->mcast_work)->work) FD: 232 BD: 1 +.+.: (wq_completion)bond391 ->(work_completion)(&(&slave->notify_work)->work) ->(work_completion)(&(&bond->mcast_work)->work) FD: 232 BD: 1 +.+.: (wq_completion)bond383#2 ->(work_completion)(&(&slave->notify_work)->work) ->(work_completion)(&(&bond->mcast_work)->work) ->&rq->__lock FD: 232 BD: 1 +.+.: (wq_completion)bond369#3 ->(work_completion)(&(&slave->notify_work)->work) ->(work_completion)(&(&bond->mcast_work)->work) ->&rq->__lock FD: 232 BD: 1 +.+.: (wq_completion)bond392 ->&rq->__lock ->(work_completion)(&(&slave->notify_work)->work) ->(work_completion)(&(&bond->mcast_work)->work) FD: 232 BD: 1 +.+.: (wq_completion)bond384#2 ->(work_completion)(&(&slave->notify_work)->work) ->(work_completion)(&(&bond->mcast_work)->work) ->&rq->__lock FD: 232 BD: 1 +.+.: (wq_completion)bond134#4 ->(work_completion)(&(&slave->notify_work)->work) ->(work_completion)(&(&bond->mcast_work)->work) ->&rq->__lock FD: 232 BD: 1 +.+.: (wq_completion)bond370#3 ->(work_completion)(&(&slave->notify_work)->work) ->(work_completion)(&(&bond->mcast_work)->work) FD: 232 BD: 1 +.+.: (wq_completion)bond393 ->(work_completion)(&(&slave->notify_work)->work) ->(work_completion)(&(&bond->mcast_work)->work) FD: 232 BD: 1 +.+.: (wq_completion)bond385#2 ->(work_completion)(&(&slave->notify_work)->work) ->(work_completion)(&(&bond->mcast_work)->work) ->&rq->__lock FD: 232 BD: 1 +.+.: (wq_completion)bond371#3 ->(work_completion)(&(&slave->notify_work)->work) ->(work_completion)(&(&bond->mcast_work)->work) FD: 232 BD: 1 +.+.: (wq_completion)bond394 ->(work_completion)(&(&slave->notify_work)->work) ->(work_completion)(&(&bond->mcast_work)->work) ->&rq->__lock FD: 232 BD: 1 +.+.: (wq_completion)bond386#2 ->(work_completion)(&(&slave->notify_work)->work) ->(work_completion)(&(&bond->mcast_work)->work) ->&rq->__lock FD: 232 BD: 1 +.+.: (wq_completion)bond372#3 ->(work_completion)(&(&slave->notify_work)->work) ->(work_completion)(&(&bond->mcast_work)->work) ->&rq->__lock FD: 232 BD: 1 +.+.: (wq_completion)bond395 ->&rq->__lock ->(work_completion)(&(&slave->notify_work)->work) ->(work_completion)(&(&bond->mcast_work)->work) FD: 232 BD: 1 +.+.: (wq_completion)bond387#2 ->&rq->__lock ->(work_completion)(&(&slave->notify_work)->work) ->(work_completion)(&(&bond->mcast_work)->work) FD: 232 BD: 1 +.+.: (wq_completion)bond373#3 ->&rq->__lock ->(work_completion)(&(&slave->notify_work)->work) ->(work_completion)(&(&bond->mcast_work)->work) FD: 232 BD: 1 +.+.: (wq_completion)bond396 ->&rq->__lock ->(work_completion)(&(&slave->notify_work)->work) ->(work_completion)(&(&bond->mcast_work)->work) FD: 232 BD: 1 +.+.: (wq_completion)bond388#2 ->(work_completion)(&(&slave->notify_work)->work) ->(work_completion)(&(&bond->mcast_work)->work) FD: 232 BD: 1 +.+.: (wq_completion)bond374#3 ->&rq->__lock ->(work_completion)(&(&slave->notify_work)->work) ->(work_completion)(&(&bond->mcast_work)->work) FD: 232 BD: 1 +.+.: (wq_completion)bond397 ->&rq->__lock ->(work_completion)(&(&slave->notify_work)->work) ->(work_completion)(&(&bond->mcast_work)->work) FD: 232 BD: 1 +.+.: (wq_completion)bond389#2 ->&rq->__lock ->(work_completion)(&(&slave->notify_work)->work) ->(work_completion)(&(&bond->mcast_work)->work) FD: 232 BD: 1 +.+.: (wq_completion)bond375#3 ->(work_completion)(&(&slave->notify_work)->work) ->(work_completion)(&(&bond->mcast_work)->work) FD: 232 BD: 1 +.+.: (wq_completion)bond398 ->(work_completion)(&(&slave->notify_work)->work) ->(work_completion)(&(&bond->mcast_work)->work) FD: 232 BD: 1 +.+.: (wq_completion)bond390#2 ->(work_completion)(&(&slave->notify_work)->work) ->(work_completion)(&(&bond->mcast_work)->work) ->&rq->__lock FD: 232 BD: 1 +.+.: (wq_completion)bond399 ->(work_completion)(&(&slave->notify_work)->work) ->(work_completion)(&(&bond->mcast_work)->work) ->&rq->__lock FD: 232 BD: 1 +.+.: (wq_completion)bond376#3 ->(work_completion)(&(&slave->notify_work)->work) ->(work_completion)(&(&bond->mcast_work)->work) FD: 232 BD: 1 +.+.: (wq_completion)bond135#4 ->(work_completion)(&(&slave->notify_work)->work) ->(work_completion)(&(&bond->mcast_work)->work) ->&rq->__lock FD: 232 BD: 1 +.+.: (wq_completion)bond391#2 ->&rq->__lock ->(work_completion)(&(&slave->notify_work)->work) ->(work_completion)(&(&bond->mcast_work)->work) FD: 232 BD: 1 +.+.: (wq_completion)bond377#3 ->(work_completion)(&(&slave->notify_work)->work) ->(work_completion)(&(&bond->mcast_work)->work) FD: 232 BD: 1 +.+.: (wq_completion)bond400 ->(work_completion)(&(&slave->notify_work)->work) ->(work_completion)(&(&bond->mcast_work)->work) FD: 232 BD: 1 +.+.: (wq_completion)bond136#4 ->&rq->__lock ->(work_completion)(&(&slave->notify_work)->work) ->(work_completion)(&(&bond->mcast_work)->work) FD: 232 BD: 1 +.+.: (wq_completion)bond392#2 ->(work_completion)(&(&slave->notify_work)->work) ->(work_completion)(&(&bond->mcast_work)->work) ->&rq->__lock FD: 232 BD: 1 +.+.: (wq_completion)bond378#3 ->(work_completion)(&(&slave->notify_work)->work) ->(work_completion)(&(&bond->mcast_work)->work) ->&rq->__lock FD: 232 BD: 1 +.+.: (wq_completion)bond401 ->(work_completion)(&(&slave->notify_work)->work) ->(work_completion)(&(&bond->mcast_work)->work) ->&rq->__lock FD: 232 BD: 1 +.+.: (wq_completion)bond393#2 ->(work_completion)(&(&slave->notify_work)->work) ->(work_completion)(&(&bond->mcast_work)->work) ->&rq->__lock FD: 232 BD: 1 +.+.: (wq_completion)bond379#3 ->(work_completion)(&(&slave->notify_work)->work) ->(work_completion)(&(&bond->mcast_work)->work) ->&rq->__lock FD: 232 BD: 1 +.+.: (wq_completion)bond402 ->(work_completion)(&(&slave->notify_work)->work) ->(work_completion)(&(&bond->mcast_work)->work) ->&rq->__lock FD: 232 BD: 1 +.+.: (wq_completion)bond394#2 ->(work_completion)(&(&slave->notify_work)->work) ->(work_completion)(&(&bond->mcast_work)->work) ->&rq->__lock FD: 232 BD: 1 +.+.: (wq_completion)bond380#3 ->(work_completion)(&(&slave->notify_work)->work) ->(work_completion)(&(&bond->mcast_work)->work) FD: 232 BD: 1 +.+.: (wq_completion)bond137#4 ->(work_completion)(&(&slave->notify_work)->work) ->(work_completion)(&(&bond->mcast_work)->work) ->&rq->__lock FD: 232 BD: 1 +.+.: (wq_completion)bond403 ->(work_completion)(&(&slave->notify_work)->work) ->(work_completion)(&(&bond->mcast_work)->work) FD: 232 BD: 1 +.+.: (wq_completion)bond395#2 ->(work_completion)(&(&slave->notify_work)->work) ->(work_completion)(&(&bond->mcast_work)->work) ->&rq->__lock FD: 232 BD: 1 +.+.: (wq_completion)bond381#3 ->&rq->__lock ->(work_completion)(&(&slave->notify_work)->work) ->(work_completion)(&(&bond->mcast_work)->work) FD: 232 BD: 1 +.+.: (wq_completion)bond404 ->&rq->__lock ->(work_completion)(&(&slave->notify_work)->work) ->(work_completion)(&(&bond->mcast_work)->work) FD: 232 BD: 1 +.+.: (wq_completion)bond396#2 ->(work_completion)(&(&slave->notify_work)->work) ->(work_completion)(&(&bond->mcast_work)->work) FD: 232 BD: 1 +.+.: (wq_completion)bond382#3 ->(work_completion)(&(&slave->notify_work)->work) ->(work_completion)(&(&bond->mcast_work)->work) ->&rq->__lock FD: 232 BD: 1 +.+.: (wq_completion)bond405 ->(work_completion)(&(&slave->notify_work)->work) ->(work_completion)(&(&bond->mcast_work)->work) ->&rq->__lock FD: 232 BD: 1 +.+.: (wq_completion)bond397#2 ->&rq->__lock ->(work_completion)(&(&slave->notify_work)->work) ->(work_completion)(&(&bond->mcast_work)->work) FD: 232 BD: 1 +.+.: (wq_completion)bond383#3 ->&rq->__lock ->(work_completion)(&(&slave->notify_work)->work) ->(work_completion)(&(&bond->mcast_work)->work) FD: 232 BD: 1 +.+.: (wq_completion)bond406 ->(work_completion)(&(&slave->notify_work)->work) ->(work_completion)(&(&bond->mcast_work)->work) ->&rq->__lock FD: 232 BD: 1 +.+.: (wq_completion)bond398#2 ->(work_completion)(&(&slave->notify_work)->work) ->(work_completion)(&(&bond->mcast_work)->work) FD: 232 BD: 1 +.+.: (wq_completion)bond384#3 ->(work_completion)(&(&slave->notify_work)->work) ->(work_completion)(&(&bond->mcast_work)->work) ->&rq->__lock FD: 232 BD: 1 +.+.: (wq_completion)bond407 ->(work_completion)(&(&slave->notify_work)->work) ->(work_completion)(&(&bond->mcast_work)->work) ->&rq->__lock FD: 232 BD: 1 +.+.: (wq_completion)bond399#2 ->(work_completion)(&(&slave->notify_work)->work) ->(work_completion)(&(&bond->mcast_work)->work) ->&rq->__lock FD: 232 BD: 1 +.+.: (wq_completion)bond385#3 ->(work_completion)(&(&slave->notify_work)->work) ->(work_completion)(&(&bond->mcast_work)->work) ->&rq->__lock FD: 232 BD: 1 +.+.: (wq_completion)bond408 ->&rq->__lock ->(work_completion)(&(&slave->notify_work)->work) ->(work_completion)(&(&bond->mcast_work)->work) FD: 232 BD: 1 +.+.: (wq_completion)bond400#2 ->(work_completion)(&(&slave->notify_work)->work) ->(work_completion)(&(&bond->mcast_work)->work) ->&rq->__lock FD: 232 BD: 1 +.+.: (wq_completion)bond409 ->&rq->__lock ->(work_completion)(&(&slave->notify_work)->work) ->(work_completion)(&(&bond->mcast_work)->work) FD: 232 BD: 1 +.+.: (wq_completion)bond401#2 ->&rq->__lock ->(work_completion)(&(&slave->notify_work)->work) ->(work_completion)(&(&bond->mcast_work)->work) FD: 232 BD: 1 +.+.: (wq_completion)bond387#3 ->(work_completion)(&(&slave->notify_work)->work) ->(work_completion)(&(&bond->mcast_work)->work) ->&rq->__lock FD: 232 BD: 1 +.+.: (wq_completion)bond138#3 ->(work_completion)(&(&slave->notify_work)->work) ->(work_completion)(&(&bond->mcast_work)->work) ->&rq->__lock FD: 232 BD: 1 +.+.: (wq_completion)bond410 ->&rq->__lock ->(work_completion)(&(&slave->notify_work)->work) ->(work_completion)(&(&bond->mcast_work)->work) FD: 232 BD: 1 +.+.: (wq_completion)bond402#2 ->(work_completion)(&(&slave->notify_work)->work) ->(work_completion)(&(&bond->mcast_work)->work) FD: 232 BD: 1 +.+.: (wq_completion)bond139#4 ->&rq->__lock ->(work_completion)(&(&slave->notify_work)->work) ->(work_completion)(&(&bond->mcast_work)->work) FD: 232 BD: 1 +.+.: (wq_completion)bond388#3 ->&rq->__lock ->(work_completion)(&(&slave->notify_work)->work) ->(work_completion)(&(&bond->mcast_work)->work) FD: 232 BD: 1 +.+.: (wq_completion)bond411 ->&rq->__lock ->(work_completion)(&(&slave->notify_work)->work) ->(work_completion)(&(&bond->mcast_work)->work) FD: 232 BD: 1 +.+.: (wq_completion)bond403#2 ->(work_completion)(&(&slave->notify_work)->work) ->(work_completion)(&(&bond->mcast_work)->work) ->&rq->__lock FD: 232 BD: 1 +.+.: (wq_completion)bond389#3 ->&rq->__lock ->(work_completion)(&(&slave->notify_work)->work) ->(work_completion)(&(&bond->mcast_work)->work) FD: 232 BD: 1 +.+.: (wq_completion)bond412 ->&rq->__lock ->(work_completion)(&(&slave->notify_work)->work) ->(work_completion)(&(&bond->mcast_work)->work) FD: 232 BD: 1 +.+.: (wq_completion)bond140#4 ->&rq->__lock ->(work_completion)(&(&slave->notify_work)->work) ->(work_completion)(&(&bond->mcast_work)->work) FD: 232 BD: 1 +.+.: (wq_completion)bond404#2 ->(work_completion)(&(&slave->notify_work)->work) ->(work_completion)(&(&bond->mcast_work)->work) ->&rq->__lock FD: 232 BD: 1 +.+.: (wq_completion)bond390#3 ->(work_completion)(&(&slave->notify_work)->work) ->(work_completion)(&(&bond->mcast_work)->work) ->&rq->__lock FD: 232 BD: 1 +.+.: (wq_completion)bond413 ->(work_completion)(&(&slave->notify_work)->work) ->(work_completion)(&(&bond->mcast_work)->work) FD: 232 BD: 1 +.+.: (wq_completion)bond405#2 ->(work_completion)(&(&slave->notify_work)->work) ->(work_completion)(&(&bond->mcast_work)->work) FD: 232 BD: 1 +.+.: (wq_completion)bond414 ->(work_completion)(&(&slave->notify_work)->work) ->(work_completion)(&(&bond->mcast_work)->work) FD: 232 BD: 1 +.+.: (wq_completion)bond406#2 ->(work_completion)(&(&slave->notify_work)->work) ->(work_completion)(&(&bond->mcast_work)->work) FD: 232 BD: 1 +.+.: (wq_completion)bond392#3 ->(work_completion)(&(&slave->notify_work)->work) ->(work_completion)(&(&bond->mcast_work)->work) ->&rq->__lock FD: 232 BD: 1 +.+.: (wq_completion)bond141#4 ->(work_completion)(&(&slave->notify_work)->work) ->(work_completion)(&(&bond->mcast_work)->work) FD: 232 BD: 1 +.+.: (wq_completion)bond415 ->(work_completion)(&(&slave->notify_work)->work) ->(work_completion)(&(&bond->mcast_work)->work) ->&rq->__lock FD: 232 BD: 1 +.+.: (wq_completion)bond407#2 ->(work_completion)(&(&slave->notify_work)->work) ->(work_completion)(&(&bond->mcast_work)->work) ->&rq->__lock FD: 232 BD: 1 +.+.: (wq_completion)bond393#3 ->(work_completion)(&(&slave->notify_work)->work) ->(work_completion)(&(&bond->mcast_work)->work) ->&rq->__lock FD: 232 BD: 1 +.+.: (wq_completion)bond416 ->(work_completion)(&(&slave->notify_work)->work) ->(work_completion)(&(&bond->mcast_work)->work) FD: 232 BD: 1 +.+.: (wq_completion)bond408#2 ->(work_completion)(&(&slave->notify_work)->work) ->(work_completion)(&(&bond->mcast_work)->work) FD: 232 BD: 1 +.+.: (wq_completion)bond394#3 ->(work_completion)(&(&slave->notify_work)->work) ->(work_completion)(&(&bond->mcast_work)->work) ->&rq->__lock FD: 232 BD: 1 +.+.: (wq_completion)bond417 ->(work_completion)(&(&slave->notify_work)->work) ->(work_completion)(&(&bond->mcast_work)->work) FD: 232 BD: 1 +.+.: (wq_completion)bond409#2 ->(work_completion)(&(&slave->notify_work)->work) ->(work_completion)(&(&bond->mcast_work)->work) FD: 232 BD: 1 +.+.: (wq_completion)bond395#3 ->(work_completion)(&(&slave->notify_work)->work) ->(work_completion)(&(&bond->mcast_work)->work) ->&rq->__lock FD: 232 BD: 1 +.+.: (wq_completion)bond418 ->&rq->__lock ->(work_completion)(&(&slave->notify_work)->work) ->(work_completion)(&(&bond->mcast_work)->work) FD: 232 BD: 1 +.+.: (wq_completion)bond410#2 ->(work_completion)(&(&slave->notify_work)->work) ->(work_completion)(&(&bond->mcast_work)->work) ->&rq->__lock FD: 232 BD: 1 +.+.: (wq_completion)bond396#3 ->(work_completion)(&(&slave->notify_work)->work) ->(work_completion)(&(&bond->mcast_work)->work) ->&rq->__lock FD: 232 BD: 1 +.+.: (wq_completion)bond419 ->(work_completion)(&(&slave->notify_work)->work) ->(work_completion)(&(&bond->mcast_work)->work) ->&rq->__lock FD: 232 BD: 1 +.+.: (wq_completion)bond411#2 ->(work_completion)(&(&slave->notify_work)->work) ->(work_completion)(&(&bond->mcast_work)->work) ->&rq->__lock FD: 232 BD: 1 +.+.: (wq_completion)bond397#3 ->&rq->__lock ->(work_completion)(&(&slave->notify_work)->work) ->(work_completion)(&(&bond->mcast_work)->work) FD: 232 BD: 1 +.+.: (wq_completion)bond420 ->(work_completion)(&(&slave->notify_work)->work) ->(work_completion)(&(&bond->mcast_work)->work) FD: 232 BD: 1 +.+.: (wq_completion)bond412#2 ->(work_completion)(&(&slave->notify_work)->work) ->(work_completion)(&(&bond->mcast_work)->work) FD: 232 BD: 1 +.+.: (wq_completion)bond398#3 ->(work_completion)(&(&slave->notify_work)->work) ->(work_completion)(&(&bond->mcast_work)->work) FD: 232 BD: 1 +.+.: (wq_completion)bond421 ->(work_completion)(&(&slave->notify_work)->work) ->(work_completion)(&(&bond->mcast_work)->work) FD: 232 BD: 1 +.+.: (wq_completion)bond142#4 ->(work_completion)(&(&slave->notify_work)->work) ->(work_completion)(&(&bond->mcast_work)->work) ->&rq->__lock FD: 232 BD: 1 +.+.: (wq_completion)bond413#2 ->(work_completion)(&(&slave->notify_work)->work) ->(work_completion)(&(&bond->mcast_work)->work) ->&rq->__lock FD: 232 BD: 1 +.+.: (wq_completion)bond399#3 ->(work_completion)(&(&slave->notify_work)->work) ->(work_completion)(&(&bond->mcast_work)->work) ->&rq->__lock FD: 232 BD: 1 +.+.: (wq_completion)bond422 ->(work_completion)(&(&slave->notify_work)->work) ->(work_completion)(&(&bond->mcast_work)->work) ->&rq->__lock FD: 232 BD: 1 +.+.: (wq_completion)bond414#2 ->&rq->__lock ->(work_completion)(&(&slave->notify_work)->work) ->(work_completion)(&(&bond->mcast_work)->work) FD: 232 BD: 1 +.+.: (wq_completion)bond400#3 ->&rq->__lock ->(work_completion)(&(&slave->notify_work)->work) ->(work_completion)(&(&bond->mcast_work)->work) FD: 232 BD: 1 +.+.: (wq_completion)bond423 ->(work_completion)(&(&slave->notify_work)->work) ->(work_completion)(&(&bond->mcast_work)->work) FD: 232 BD: 1 +.+.: (wq_completion)bond401#3 ->(work_completion)(&(&slave->notify_work)->work) ->(work_completion)(&(&bond->mcast_work)->work) FD: 232 BD: 1 +.+.: (wq_completion)bond415#2 ->&rq->__lock ->(work_completion)(&(&slave->notify_work)->work) ->(work_completion)(&(&bond->mcast_work)->work) FD: 232 BD: 1 +.+.: (wq_completion)bond143#4 ->(work_completion)(&(&slave->notify_work)->work) ->(work_completion)(&(&bond->mcast_work)->work) FD: 232 BD: 1 +.+.: (wq_completion)bond424 ->(work_completion)(&(&slave->notify_work)->work) ->(work_completion)(&(&bond->mcast_work)->work) FD: 232 BD: 1 +.+.: (wq_completion)bond402#3 ->(work_completion)(&(&slave->notify_work)->work) ->(work_completion)(&(&bond->mcast_work)->work) ->&rq->__lock FD: 232 BD: 1 +.+.: (wq_completion)bond416#2 ->(work_completion)(&(&slave->notify_work)->work) ->(work_completion)(&(&bond->mcast_work)->work) ->&rq->__lock FD: 232 BD: 1 +.+.: (wq_completion)bond425 ->&rq->__lock ->(work_completion)(&(&slave->notify_work)->work) ->(work_completion)(&(&bond->mcast_work)->work) FD: 232 BD: 1 +.+.: (wq_completion)bond403#3 ->&rq->__lock ->(work_completion)(&(&slave->notify_work)->work) ->(work_completion)(&(&bond->mcast_work)->work) FD: 232 BD: 1 +.+.: (wq_completion)bond417#2 ->&rq->__lock ->(work_completion)(&(&slave->notify_work)->work) ->(work_completion)(&(&bond->mcast_work)->work) FD: 232 BD: 1 +.+.: (wq_completion)bond426 ->(work_completion)(&(&slave->notify_work)->work) ->(work_completion)(&(&bond->mcast_work)->work) ->&rq->__lock FD: 232 BD: 1 +.+.: (wq_completion)bond144#4 ->&rq->__lock ->(work_completion)(&(&slave->notify_work)->work) ->(work_completion)(&(&bond->mcast_work)->work) FD: 232 BD: 1 +.+.: (wq_completion)bond418#2 ->(work_completion)(&(&slave->notify_work)->work) ->(work_completion)(&(&bond->mcast_work)->work) ->&rq->__lock FD: 232 BD: 1 +.+.: (wq_completion)bond427 ->(work_completion)(&(&slave->notify_work)->work) ->(work_completion)(&(&bond->mcast_work)->work) ->&rq->__lock FD: 232 BD: 1 +.+.: (wq_completion)bond405#3 ->(work_completion)(&(&slave->notify_work)->work) ->(work_completion)(&(&bond->mcast_work)->work) ->&rq->__lock FD: 232 BD: 1 +.+.: (wq_completion)bond419#2 ->&rq->__lock ->(work_completion)(&(&slave->notify_work)->work) ->(work_completion)(&(&bond->mcast_work)->work) FD: 232 BD: 1 +.+.: (wq_completion)bond428 ->(work_completion)(&(&slave->notify_work)->work) ->(work_completion)(&(&bond->mcast_work)->work) ->&rq->__lock FD: 232 BD: 1 +.+.: (wq_completion)bond406#3 ->&rq->__lock ->(work_completion)(&(&slave->notify_work)->work) ->(work_completion)(&(&bond->mcast_work)->work) FD: 232 BD: 1 +.+.: (wq_completion)bond420#2 ->&rq->__lock ->(work_completion)(&(&slave->notify_work)->work) ->(work_completion)(&(&bond->mcast_work)->work) FD: 232 BD: 1 +.+.: (wq_completion)bond429 ->(work_completion)(&(&slave->notify_work)->work) ->(work_completion)(&(&bond->mcast_work)->work) FD: 232 BD: 1 +.+.: (wq_completion)bond145#4 ->(work_completion)(&(&slave->notify_work)->work) ->(work_completion)(&(&bond->mcast_work)->work) FD: 232 BD: 1 +.+.: (wq_completion)bond407#3 ->(work_completion)(&(&slave->notify_work)->work) ->(work_completion)(&(&bond->mcast_work)->work) ->&rq->__lock FD: 232 BD: 1 +.+.: (wq_completion)bond421#2 ->&rq->__lock ->(work_completion)(&(&slave->notify_work)->work) ->(work_completion)(&(&bond->mcast_work)->work) FD: 232 BD: 1 +.+.: (wq_completion)bond430 ->(work_completion)(&(&slave->notify_work)->work) ->(work_completion)(&(&bond->mcast_work)->work) FD: 232 BD: 1 +.+.: (wq_completion)bond408#3 ->(work_completion)(&(&slave->notify_work)->work) ->(work_completion)(&(&bond->mcast_work)->work) ->&rq->__lock FD: 232 BD: 1 +.+.: (wq_completion)bond146#4 ->(work_completion)(&(&slave->notify_work)->work) ->(work_completion)(&(&bond->mcast_work)->work) ->&rq->__lock FD: 232 BD: 1 +.+.: (wq_completion)bond422#2 ->(work_completion)(&(&slave->notify_work)->work) ->(work_completion)(&(&bond->mcast_work)->work) ->&rq->__lock FD: 232 BD: 1 +.+.: (wq_completion)bond431 ->&rq->__lock ->(work_completion)(&(&slave->notify_work)->work) ->(work_completion)(&(&bond->mcast_work)->work) FD: 232 BD: 1 +.+.: (wq_completion)bond409#3 ->(work_completion)(&(&slave->notify_work)->work) ->(work_completion)(&(&bond->mcast_work)->work) FD: 232 BD: 1 +.+.: (wq_completion)bond147#4 ->&rq->__lock ->(work_completion)(&(&slave->notify_work)->work) ->(work_completion)(&(&bond->mcast_work)->work) FD: 232 BD: 1 +.+.: (wq_completion)bond423#2 ->(work_completion)(&(&slave->notify_work)->work) ->(work_completion)(&(&bond->mcast_work)->work) FD: 232 BD: 1 +.+.: (wq_completion)bond432 ->&rq->__lock ->(work_completion)(&(&slave->notify_work)->work) ->(work_completion)(&(&bond->mcast_work)->work) FD: 232 BD: 1 +.+.: (wq_completion)bond410#3 ->(work_completion)(&(&slave->notify_work)->work) ->(work_completion)(&(&bond->mcast_work)->work) ->&rq->__lock FD: 232 BD: 1 +.+.: (wq_completion)bond424#2 ->&rq->__lock ->(work_completion)(&(&slave->notify_work)->work) ->(work_completion)(&(&bond->mcast_work)->work) FD: 232 BD: 1 +.+.: (wq_completion)bond433 ->&rq->__lock ->(work_completion)(&(&slave->notify_work)->work) ->(work_completion)(&(&bond->mcast_work)->work) FD: 232 BD: 1 +.+.: (wq_completion)bond425#2 ->(work_completion)(&(&slave->notify_work)->work) ->(work_completion)(&(&bond->mcast_work)->work) ->&rq->__lock FD: 232 BD: 1 +.+.: (wq_completion)bond434 ->(work_completion)(&(&slave->notify_work)->work) ->(work_completion)(&(&bond->mcast_work)->work) FD: 232 BD: 1 +.+.: (wq_completion)bond435 ->(work_completion)(&(&slave->notify_work)->work) ->(work_completion)(&(&bond->mcast_work)->work) ->&rq->__lock FD: 232 BD: 1 +.+.: (wq_completion)bond412#3 ->(work_completion)(&(&slave->notify_work)->work) ->(work_completion)(&(&bond->mcast_work)->work) ->&rq->__lock FD: 232 BD: 1 +.+.: (wq_completion)bond426#2 ->(work_completion)(&(&slave->notify_work)->work) ->(work_completion)(&(&bond->mcast_work)->work) ->&rq->__lock FD: 232 BD: 1 +.+.: (wq_completion)bond413#3 ->&rq->__lock ->(work_completion)(&(&slave->notify_work)->work) ->(work_completion)(&(&bond->mcast_work)->work) FD: 232 BD: 1 +.+.: (wq_completion)bond436 ->(work_completion)(&(&slave->notify_work)->work) ->(work_completion)(&(&bond->mcast_work)->work) ->&rq->__lock FD: 232 BD: 1 +.+.: (wq_completion)bond427#2 ->(work_completion)(&(&slave->notify_work)->work) ->(work_completion)(&(&bond->mcast_work)->work) ->&rq->__lock FD: 232 BD: 1 +.+.: (wq_completion)bond414#3 ->(work_completion)(&(&slave->notify_work)->work) ->(work_completion)(&(&bond->mcast_work)->work) FD: 232 BD: 1 +.+.: (wq_completion)bond437 ->&rq->__lock ->(work_completion)(&(&slave->notify_work)->work) ->(work_completion)(&(&bond->mcast_work)->work) FD: 232 BD: 1 +.+.: (wq_completion)bond428#2 ->(work_completion)(&(&slave->notify_work)->work) ->(work_completion)(&(&bond->mcast_work)->work) FD: 232 BD: 1 +.+.: (wq_completion)bond415#3 ->(work_completion)(&(&slave->notify_work)->work) ->(work_completion)(&(&bond->mcast_work)->work) ->&rq->__lock FD: 232 BD: 1 +.+.: (wq_completion)bond438 ->(work_completion)(&(&slave->notify_work)->work) ->(work_completion)(&(&bond->mcast_work)->work) FD: 232 BD: 1 +.+.: (wq_completion)bond429#2 ->(work_completion)(&(&slave->notify_work)->work) ->(work_completion)(&(&bond->mcast_work)->work) ->&rq->__lock FD: 232 BD: 1 +.+.: (wq_completion)bond416#3 ->(work_completion)(&(&slave->notify_work)->work) ->(work_completion)(&(&bond->mcast_work)->work) FD: 232 BD: 1 +.+.: (wq_completion)bond439 ->(work_completion)(&(&slave->notify_work)->work) ->(work_completion)(&(&bond->mcast_work)->work) ->&rq->__lock FD: 232 BD: 1 +.+.: (wq_completion)bond430#2 ->&rq->__lock ->(work_completion)(&(&slave->notify_work)->work) ->(work_completion)(&(&bond->mcast_work)->work) FD: 232 BD: 1 +.+.: (wq_completion)bond417#3 ->(work_completion)(&(&slave->notify_work)->work) ->(work_completion)(&(&bond->mcast_work)->work) FD: 232 BD: 1 +.+.: (wq_completion)bond431#2 ->(work_completion)(&(&slave->notify_work)->work) ->(work_completion)(&(&bond->mcast_work)->work) ->&rq->__lock FD: 232 BD: 1 +.+.: (wq_completion)bond440 ->(work_completion)(&(&slave->notify_work)->work) ->(work_completion)(&(&bond->mcast_work)->work) FD: 232 BD: 1 +.+.: (wq_completion)bond418#3 ->(work_completion)(&(&slave->notify_work)->work) ->(work_completion)(&(&bond->mcast_work)->work) ->&rq->__lock FD: 232 BD: 1 +.+.: (wq_completion)bond432#2 ->&rq->__lock ->(work_completion)(&(&slave->notify_work)->work) ->(work_completion)(&(&bond->mcast_work)->work) FD: 232 BD: 1 +.+.: (wq_completion)bond441 ->&rq->__lock ->(work_completion)(&(&slave->notify_work)->work) ->(work_completion)(&(&bond->mcast_work)->work) FD: 232 BD: 1 +.+.: (wq_completion)bond419#3 ->&rq->__lock ->(work_completion)(&(&slave->notify_work)->work) ->(work_completion)(&(&bond->mcast_work)->work) FD: 232 BD: 1 +.+.: (wq_completion)bond433#2 ->&rq->__lock ->(work_completion)(&(&slave->notify_work)->work) ->(work_completion)(&(&bond->mcast_work)->work) FD: 232 BD: 1 +.+.: (wq_completion)bond442 ->(work_completion)(&(&slave->notify_work)->work) ->(work_completion)(&(&bond->mcast_work)->work) ->&rq->__lock FD: 232 BD: 1 +.+.: (wq_completion)bond420#3 ->&rq->__lock ->(work_completion)(&(&slave->notify_work)->work) ->(work_completion)(&(&bond->mcast_work)->work) FD: 232 BD: 1 +.+.: (wq_completion)bond434#2 ->(work_completion)(&(&slave->notify_work)->work) ->(work_completion)(&(&bond->mcast_work)->work) FD: 232 BD: 1 +.+.: (wq_completion)bond443 ->(work_completion)(&(&slave->notify_work)->work) ->(work_completion)(&(&bond->mcast_work)->work) FD: 232 BD: 1 +.+.: (wq_completion)bond421#3 ->&rq->__lock ->(work_completion)(&(&slave->notify_work)->work) ->(work_completion)(&(&bond->mcast_work)->work) FD: 232 BD: 1 +.+.: (wq_completion)bond435#2 ->(work_completion)(&(&slave->notify_work)->work) ->(work_completion)(&(&bond->mcast_work)->work) ->&rq->__lock FD: 232 BD: 1 +.+.: (wq_completion)bond444 ->(work_completion)(&(&slave->notify_work)->work) ->(work_completion)(&(&bond->mcast_work)->work) FD: 232 BD: 1 +.+.: (wq_completion)bond422#3 ->&rq->__lock ->(work_completion)(&(&slave->notify_work)->work) ->(work_completion)(&(&bond->mcast_work)->work) FD: 232 BD: 1 +.+.: (wq_completion)bond436#2 ->&rq->__lock ->(work_completion)(&(&slave->notify_work)->work) ->(work_completion)(&(&bond->mcast_work)->work) FD: 232 BD: 1 +.+.: (wq_completion)bond445 ->(work_completion)(&(&slave->notify_work)->work) ->(work_completion)(&(&bond->mcast_work)->work) ->&rq->__lock FD: 232 BD: 1 +.+.: (wq_completion)bond423#3 ->&rq->__lock ->(work_completion)(&(&slave->notify_work)->work) ->(work_completion)(&(&bond->mcast_work)->work) FD: 232 BD: 1 +.+.: (wq_completion)bond437#2 ->(work_completion)(&(&slave->notify_work)->work) ->(work_completion)(&(&bond->mcast_work)->work) FD: 232 BD: 1 +.+.: (wq_completion)bond446 ->(work_completion)(&(&slave->notify_work)->work) ->(work_completion)(&(&bond->mcast_work)->work) FD: 232 BD: 1 +.+.: (wq_completion)bond438#2 ->(work_completion)(&(&slave->notify_work)->work) ->(work_completion)(&(&bond->mcast_work)->work) FD: 232 BD: 1 +.+.: (wq_completion)bond447 ->(work_completion)(&(&slave->notify_work)->work) ->(work_completion)(&(&bond->mcast_work)->work) FD: 232 BD: 1 +.+.: (wq_completion)bond424#3 ->(work_completion)(&(&slave->notify_work)->work) ->(work_completion)(&(&bond->mcast_work)->work) FD: 232 BD: 1 +.+.: (wq_completion)bond448 ->(work_completion)(&(&slave->notify_work)->work) ->(work_completion)(&(&bond->mcast_work)->work) ->&rq->__lock FD: 232 BD: 1 +.+.: (wq_completion)bond439#2 ->(work_completion)(&(&slave->notify_work)->work) ->(work_completion)(&(&bond->mcast_work)->work) FD: 232 BD: 1 +.+.: (wq_completion)bond425#3 ->&rq->__lock ->(work_completion)(&(&slave->notify_work)->work) ->(work_completion)(&(&bond->mcast_work)->work) FD: 232 BD: 1 +.+.: (wq_completion)bond449 ->(work_completion)(&(&slave->notify_work)->work) ->(work_completion)(&(&bond->mcast_work)->work) ->&rq->__lock FD: 232 BD: 1 +.+.: (wq_completion)bond440#2 ->(work_completion)(&(&slave->notify_work)->work) ->(work_completion)(&(&bond->mcast_work)->work) ->&rq->__lock FD: 232 BD: 1 +.+.: (wq_completion)bond426#3 ->&rq->__lock ->(work_completion)(&(&slave->notify_work)->work) ->(work_completion)(&(&bond->mcast_work)->work) FD: 232 BD: 1 +.+.: (wq_completion)bond450 ->(work_completion)(&(&slave->notify_work)->work) ->(work_completion)(&(&bond->mcast_work)->work) FD: 232 BD: 1 +.+.: (wq_completion)bond441#2 ->&rq->__lock ->(work_completion)(&(&slave->notify_work)->work) ->(work_completion)(&(&bond->mcast_work)->work) FD: 232 BD: 1 +.+.: (wq_completion)bond427#3 ->&rq->__lock ->(work_completion)(&(&slave->notify_work)->work) ->(work_completion)(&(&bond->mcast_work)->work) FD: 232 BD: 1 +.+.: (wq_completion)bond451 ->(work_completion)(&(&slave->notify_work)->work) ->(work_completion)(&(&bond->mcast_work)->work) FD: 232 BD: 1 +.+.: (wq_completion)bond442#2 ->(work_completion)(&(&slave->notify_work)->work) ->(work_completion)(&(&bond->mcast_work)->work) FD: 232 BD: 1 +.+.: (wq_completion)bond428#3 ->(work_completion)(&(&slave->notify_work)->work) ->(work_completion)(&(&bond->mcast_work)->work) FD: 232 BD: 1 +.+.: (wq_completion)bond452 ->(work_completion)(&(&slave->notify_work)->work) ->(work_completion)(&(&bond->mcast_work)->work) FD: 232 BD: 1 +.+.: (wq_completion)bond443#2 ->(work_completion)(&(&slave->notify_work)->work) ->(work_completion)(&(&bond->mcast_work)->work) ->&rq->__lock FD: 232 BD: 1 +.+.: (wq_completion)bond429#3 ->(work_completion)(&(&slave->notify_work)->work) ->(work_completion)(&(&bond->mcast_work)->work) FD: 232 BD: 1 +.+.: (wq_completion)bond453 ->&rq->__lock ->(work_completion)(&(&slave->notify_work)->work) ->(work_completion)(&(&bond->mcast_work)->work) FD: 232 BD: 1 +.+.: (wq_completion)bond444#2 ->&rq->__lock ->(work_completion)(&(&slave->notify_work)->work) ->(work_completion)(&(&bond->mcast_work)->work) FD: 232 BD: 1 +.+.: (wq_completion)bond430#3 ->(work_completion)(&(&slave->notify_work)->work) ->(work_completion)(&(&bond->mcast_work)->work) ->&rq->__lock FD: 232 BD: 1 +.+.: (wq_completion)bond148#4 ->(work_completion)(&(&slave->notify_work)->work) ->(work_completion)(&(&bond->mcast_work)->work) ->&rq->__lock FD: 232 BD: 1 +.+.: (wq_completion)bond454 ->&rq->__lock ->(work_completion)(&(&slave->notify_work)->work) ->(work_completion)(&(&bond->mcast_work)->work) FD: 232 BD: 1 +.+.: (wq_completion)bond445#2 ->&rq->__lock ->(work_completion)(&(&slave->notify_work)->work) ->(work_completion)(&(&bond->mcast_work)->work) FD: 232 BD: 1 +.+.: (wq_completion)bond431#3 ->(work_completion)(&(&slave->notify_work)->work) ->(work_completion)(&(&bond->mcast_work)->work) FD: 232 BD: 1 +.+.: (wq_completion)bond149#4 ->&rq->__lock ->(work_completion)(&(&slave->notify_work)->work) ->(work_completion)(&(&bond->mcast_work)->work) FD: 232 BD: 1 +.+.: (wq_completion)bond455 ->(work_completion)(&(&slave->notify_work)->work) ->(work_completion)(&(&bond->mcast_work)->work) FD: 232 BD: 1 +.+.: (wq_completion)bond446#2 ->&rq->__lock ->(work_completion)(&(&slave->notify_work)->work) ->(work_completion)(&(&bond->mcast_work)->work) FD: 232 BD: 1 +.+.: (wq_completion)bond432#3 ->&rq->__lock ->(work_completion)(&(&slave->notify_work)->work) ->(work_completion)(&(&bond->mcast_work)->work) FD: 232 BD: 1 +.+.: (wq_completion)bond150#4 ->&rq->__lock ->(work_completion)(&(&slave->notify_work)->work) ->(work_completion)(&(&bond->mcast_work)->work) FD: 232 BD: 1 +.+.: (wq_completion)bond456 ->&rq->__lock ->(work_completion)(&(&slave->notify_work)->work) ->(work_completion)(&(&bond->mcast_work)->work) FD: 232 BD: 1 +.+.: (wq_completion)bond447#2 ->&rq->__lock ->(work_completion)(&(&slave->notify_work)->work) ->(work_completion)(&(&bond->mcast_work)->work) FD: 232 BD: 1 +.+.: (wq_completion)bond433#3 ->&rq->__lock ->(work_completion)(&(&slave->notify_work)->work) ->(work_completion)(&(&bond->mcast_work)->work) FD: 232 BD: 1 +.+.: (wq_completion)bond151#3 ->&rq->__lock ->(work_completion)(&(&slave->notify_work)->work) ->(work_completion)(&(&bond->mcast_work)->work) FD: 232 BD: 1 +.+.: (wq_completion)bond457 ->(work_completion)(&(&slave->notify_work)->work) ->(work_completion)(&(&bond->mcast_work)->work) ->&rq->__lock FD: 232 BD: 1 +.+.: (wq_completion)bond448#2 ->&rq->__lock ->(work_completion)(&(&slave->notify_work)->work) ->(work_completion)(&(&bond->mcast_work)->work) FD: 232 BD: 1 +.+.: (wq_completion)bond434#3 ->(work_completion)(&(&slave->notify_work)->work) ->(work_completion)(&(&bond->mcast_work)->work) FD: 232 BD: 1 +.+.: (wq_completion)bond458 ->&rq->__lock ->(work_completion)(&(&slave->notify_work)->work) ->(work_completion)(&(&bond->mcast_work)->work) FD: 232 BD: 1 +.+.: (wq_completion)bond449#2 ->(work_completion)(&(&slave->notify_work)->work) ->(work_completion)(&(&bond->mcast_work)->work) ->&rq->__lock FD: 232 BD: 1 +.+.: (wq_completion)bond435#3 ->(work_completion)(&(&slave->notify_work)->work) ->(work_completion)(&(&bond->mcast_work)->work) ->&rq->__lock FD: 232 BD: 1 +.+.: (wq_completion)bond459 ->(work_completion)(&(&slave->notify_work)->work) ->(work_completion)(&(&bond->mcast_work)->work) FD: 232 BD: 1 +.+.: (wq_completion)bond450#2 ->(work_completion)(&(&slave->notify_work)->work) ->(work_completion)(&(&bond->mcast_work)->work) FD: 232 BD: 1 +.+.: (wq_completion)bond436#3 ->&rq->__lock ->(work_completion)(&(&slave->notify_work)->work) ->(work_completion)(&(&bond->mcast_work)->work) FD: 232 BD: 1 +.+.: (wq_completion)bond460 ->&rq->__lock ->(work_completion)(&(&slave->notify_work)->work) ->(work_completion)(&(&bond->mcast_work)->work) FD: 232 BD: 1 +.+.: (wq_completion)bond451#2 ->(work_completion)(&(&slave->notify_work)->work) ->(work_completion)(&(&bond->mcast_work)->work) ->&rq->__lock FD: 232 BD: 1 +.+.: (wq_completion)bond437#3 ->(work_completion)(&(&slave->notify_work)->work) ->(work_completion)(&(&bond->mcast_work)->work) FD: 232 BD: 1 +.+.: (wq_completion)bond461 ->(work_completion)(&(&slave->notify_work)->work) ->(work_completion)(&(&bond->mcast_work)->work) ->&rq->__lock FD: 232 BD: 1 +.+.: (wq_completion)bond438#3 ->&rq->__lock ->(work_completion)(&(&slave->notify_work)->work) ->(work_completion)(&(&bond->mcast_work)->work) FD: 232 BD: 1 +.+.: (wq_completion)bond452#2 ->(work_completion)(&(&slave->notify_work)->work) ->(work_completion)(&(&bond->mcast_work)->work) FD: 232 BD: 1 +.+.: (wq_completion)bond462 ->(work_completion)(&(&slave->notify_work)->work) ->(work_completion)(&(&bond->mcast_work)->work) FD: 232 BD: 1 +.+.: (wq_completion)bond439#3 ->(work_completion)(&(&slave->notify_work)->work) ->(work_completion)(&(&bond->mcast_work)->work) FD: 232 BD: 1 +.+.: (wq_completion)bond453#2 ->(work_completion)(&(&slave->notify_work)->work) ->(work_completion)(&(&bond->mcast_work)->work) ->&rq->__lock FD: 232 BD: 1 +.+.: (wq_completion)bond463 ->(work_completion)(&(&slave->notify_work)->work) ->(work_completion)(&(&bond->mcast_work)->work) ->&rq->__lock FD: 232 BD: 1 +.+.: (wq_completion)bond152#4 ->&rq->__lock ->(work_completion)(&(&slave->notify_work)->work) ->(work_completion)(&(&bond->mcast_work)->work) FD: 232 BD: 1 +.+.: (wq_completion)bond440#3 ->&rq->__lock ->(work_completion)(&(&slave->notify_work)->work) ->(work_completion)(&(&bond->mcast_work)->work) FD: 232 BD: 1 +.+.: (wq_completion)bond454#2 ->(work_completion)(&(&slave->notify_work)->work) ->(work_completion)(&(&bond->mcast_work)->work) ->&rq->__lock FD: 232 BD: 1 +.+.: (wq_completion)bond464 ->&rq->__lock ->(work_completion)(&(&slave->notify_work)->work) ->(work_completion)(&(&bond->mcast_work)->work) FD: 232 BD: 1 +.+.: (wq_completion)bond441#3 ->&rq->__lock ->(work_completion)(&(&slave->notify_work)->work) ->(work_completion)(&(&bond->mcast_work)->work) FD: 232 BD: 1 +.+.: (wq_completion)bond455#2 ->&rq->__lock ->(work_completion)(&(&slave->notify_work)->work) ->(work_completion)(&(&bond->mcast_work)->work) FD: 232 BD: 1 +.+.: (wq_completion)bond465 ->&rq->__lock ->(work_completion)(&(&slave->notify_work)->work) ->(work_completion)(&(&bond->mcast_work)->work) FD: 232 BD: 1 +.+.: (wq_completion)bond442#3 ->&rq->__lock ->(work_completion)(&(&slave->notify_work)->work) ->(work_completion)(&(&bond->mcast_work)->work) FD: 232 BD: 1 +.+.: (wq_completion)bond153#4 ->(work_completion)(&(&slave->notify_work)->work) ->(work_completion)(&(&bond->mcast_work)->work) ->&rq->__lock FD: 232 BD: 1 +.+.: (wq_completion)bond456#2 ->(work_completion)(&(&slave->notify_work)->work) ->(work_completion)(&(&bond->mcast_work)->work) FD: 232 BD: 1 +.+.: (wq_completion)bond466 ->&rq->__lock ->(work_completion)(&(&slave->notify_work)->work) ->(work_completion)(&(&bond->mcast_work)->work) FD: 232 BD: 1 +.+.: (wq_completion)bond443#3 ->(work_completion)(&(&slave->notify_work)->work) ->(work_completion)(&(&bond->mcast_work)->work) FD: 232 BD: 1 +.+.: (wq_completion)bond457#2 ->(work_completion)(&(&slave->notify_work)->work) ->(work_completion)(&(&bond->mcast_work)->work) FD: 232 BD: 1 +.+.: (wq_completion)bond467 ->(work_completion)(&(&slave->notify_work)->work) ->(work_completion)(&(&bond->mcast_work)->work) FD: 232 BD: 1 +.+.: (wq_completion)bond444#3 ->&rq->__lock ->(work_completion)(&(&slave->notify_work)->work) ->(work_completion)(&(&bond->mcast_work)->work) FD: 232 BD: 1 +.+.: (wq_completion)bond458#2 ->(work_completion)(&(&slave->notify_work)->work) ->(work_completion)(&(&bond->mcast_work)->work) FD: 232 BD: 1 +.+.: (wq_completion)bond468 ->&rq->__lock ->(work_completion)(&(&slave->notify_work)->work) ->(work_completion)(&(&bond->mcast_work)->work) FD: 232 BD: 1 +.+.: (wq_completion)bond445#3 ->(work_completion)(&(&slave->notify_work)->work) ->(work_completion)(&(&bond->mcast_work)->work) FD: 232 BD: 1 +.+.: (wq_completion)bond459#2 ->&rq->__lock ->(work_completion)(&(&slave->notify_work)->work) ->(work_completion)(&(&bond->mcast_work)->work) FD: 232 BD: 1 +.+.: (wq_completion)bond469 ->(work_completion)(&(&slave->notify_work)->work) ->(work_completion)(&(&bond->mcast_work)->work) FD: 232 BD: 1 +.+.: (wq_completion)bond154#4 ->(work_completion)(&(&slave->notify_work)->work) ->(work_completion)(&(&bond->mcast_work)->work) FD: 232 BD: 1 +.+.: (wq_completion)bond446#3 ->&rq->__lock ->(work_completion)(&(&slave->notify_work)->work) ->(work_completion)(&(&bond->mcast_work)->work) FD: 232 BD: 1 +.+.: (wq_completion)bond460#2 ->&rq->__lock ->(work_completion)(&(&slave->notify_work)->work) ->(work_completion)(&(&bond->mcast_work)->work) FD: 232 BD: 1 +.+.: (wq_completion)bond470 ->&rq->__lock ->(work_completion)(&(&slave->notify_work)->work) ->(work_completion)(&(&bond->mcast_work)->work) FD: 232 BD: 1 +.+.: (wq_completion)bond447#3 ->(work_completion)(&(&slave->notify_work)->work) ->(work_completion)(&(&bond->mcast_work)->work) FD: 232 BD: 1 +.+.: (wq_completion)bond461#2 ->&rq->__lock ->(work_completion)(&(&slave->notify_work)->work) ->(work_completion)(&(&bond->mcast_work)->work) FD: 232 BD: 1 +.+.: (wq_completion)bond471 ->&rq->__lock ->(work_completion)(&(&slave->notify_work)->work) ->(work_completion)(&(&bond->mcast_work)->work) FD: 232 BD: 1 +.+.: (wq_completion)bond448#3 ->(work_completion)(&(&slave->notify_work)->work) ->(work_completion)(&(&bond->mcast_work)->work) FD: 232 BD: 1 +.+.: (wq_completion)bond462#2 ->(work_completion)(&(&slave->notify_work)->work) ->(work_completion)(&(&bond->mcast_work)->work) FD: 232 BD: 1 +.+.: (wq_completion)bond472 ->(work_completion)(&(&slave->notify_work)->work) ->(work_completion)(&(&bond->mcast_work)->work) ->&rq->__lock FD: 232 BD: 1 +.+.: (wq_completion)bond449#3 ->(work_completion)(&(&slave->notify_work)->work) ->(work_completion)(&(&bond->mcast_work)->work) ->&rq->__lock FD: 232 BD: 1 +.+.: (wq_completion)bond463#2 ->(work_completion)(&(&slave->notify_work)->work) ->(work_completion)(&(&bond->mcast_work)->work) FD: 232 BD: 1 +.+.: (wq_completion)bond473 ->(work_completion)(&(&slave->notify_work)->work) ->(work_completion)(&(&bond->mcast_work)->work) FD: 232 BD: 1 +.+.: (wq_completion)bond450#3 ->(work_completion)(&(&slave->notify_work)->work) ->(work_completion)(&(&bond->mcast_work)->work) FD: 232 BD: 1 +.+.: (wq_completion)bond464#2 ->&rq->__lock ->(work_completion)(&(&slave->notify_work)->work) ->(work_completion)(&(&bond->mcast_work)->work) FD: 232 BD: 1 +.+.: (wq_completion)bond474 ->&rq->__lock ->(work_completion)(&(&slave->notify_work)->work) ->(work_completion)(&(&bond->mcast_work)->work) FD: 232 BD: 1 +.+.: (wq_completion)bond451#3 ->&rq->__lock ->(work_completion)(&(&slave->notify_work)->work) ->(work_completion)(&(&bond->mcast_work)->work) FD: 232 BD: 1 +.+.: (wq_completion)bond465#2 ->&rq->__lock ->(work_completion)(&(&slave->notify_work)->work) ->(work_completion)(&(&bond->mcast_work)->work) FD: 232 BD: 1 +.+.: (wq_completion)bond475 ->(work_completion)(&(&slave->notify_work)->work) ->(work_completion)(&(&bond->mcast_work)->work) FD: 232 BD: 1 +.+.: (wq_completion)bond452#3 ->(work_completion)(&(&slave->notify_work)->work) ->(work_completion)(&(&bond->mcast_work)->work) FD: 232 BD: 1 +.+.: (wq_completion)bond466#2 ->&rq->__lock ->(work_completion)(&(&slave->notify_work)->work) ->(work_completion)(&(&bond->mcast_work)->work) FD: 232 BD: 1 +.+.: (wq_completion)bond476 ->(work_completion)(&(&slave->notify_work)->work) ->(work_completion)(&(&bond->mcast_work)->work) FD: 232 BD: 1 +.+.: (wq_completion)bond155#4 ->(work_completion)(&(&slave->notify_work)->work) ->(work_completion)(&(&bond->mcast_work)->work) ->&rq->__lock FD: 232 BD: 1 +.+.: (wq_completion)bond453#3 ->&rq->__lock ->(work_completion)(&(&slave->notify_work)->work) ->(work_completion)(&(&bond->mcast_work)->work) FD: 232 BD: 1 +.+.: (wq_completion)bond467#2 ->(work_completion)(&(&slave->notify_work)->work) ->(work_completion)(&(&bond->mcast_work)->work) FD: 232 BD: 1 +.+.: (wq_completion)bond477 ->&rq->__lock ->(work_completion)(&(&slave->notify_work)->work) ->(work_completion)(&(&bond->mcast_work)->work) FD: 232 BD: 1 +.+.: (wq_completion)bond454#3 ->(work_completion)(&(&slave->notify_work)->work) ->(work_completion)(&(&bond->mcast_work)->work) ->&rq->__lock FD: 232 BD: 1 +.+.: (wq_completion)bond468#2 ->(work_completion)(&(&slave->notify_work)->work) ->(work_completion)(&(&bond->mcast_work)->work) FD: 232 BD: 1 +.+.: (wq_completion)bond478 ->(work_completion)(&(&slave->notify_work)->work) ->(work_completion)(&(&bond->mcast_work)->work) FD: 232 BD: 1 +.+.: (wq_completion)bond455#3 ->(work_completion)(&(&slave->notify_work)->work) ->(work_completion)(&(&bond->mcast_work)->work) FD: 232 BD: 1 +.+.: (wq_completion)bond469#2 ->(work_completion)(&(&slave->notify_work)->work) ->(work_completion)(&(&bond->mcast_work)->work) FD: 232 BD: 1 +.+.: (wq_completion)bond479 ->(work_completion)(&(&slave->notify_work)->work) ->(work_completion)(&(&bond->mcast_work)->work) FD: 232 BD: 1 +.+.: (wq_completion)bond456#3 ->(work_completion)(&(&slave->notify_work)->work) ->(work_completion)(&(&bond->mcast_work)->work) FD: 232 BD: 1 +.+.: (wq_completion)bond470#2 ->(work_completion)(&(&slave->notify_work)->work) ->(work_completion)(&(&bond->mcast_work)->work) FD: 232 BD: 1 +.+.: (wq_completion)bond480 ->&rq->__lock ->(work_completion)(&(&slave->notify_work)->work) ->(work_completion)(&(&bond->mcast_work)->work) FD: 232 BD: 1 +.+.: (wq_completion)bond471#2 ->&rq->__lock ->(work_completion)(&(&slave->notify_work)->work) ->(work_completion)(&(&bond->mcast_work)->work) FD: 232 BD: 1 +.+.: (wq_completion)bond481 ->(work_completion)(&(&slave->notify_work)->work) ->(work_completion)(&(&bond->mcast_work)->work) FD: 232 BD: 1 +.+.: (wq_completion)bond457#3 ->(work_completion)(&(&slave->notify_work)->work) ->(work_completion)(&(&bond->mcast_work)->work) FD: 232 BD: 1 +.+.: (wq_completion)bond482 ->&rq->__lock ->(work_completion)(&(&slave->notify_work)->work) ->(work_completion)(&(&bond->mcast_work)->work) FD: 232 BD: 1 +.+.: (wq_completion)bond472#2 ->&rq->__lock ->(work_completion)(&(&slave->notify_work)->work) ->(work_completion)(&(&bond->mcast_work)->work) FD: 232 BD: 1 +.+.: (wq_completion)bond473#2 ->&rq->__lock ->(work_completion)(&(&slave->notify_work)->work) ->(work_completion)(&(&bond->mcast_work)->work) FD: 232 BD: 1 +.+.: (wq_completion)bond483 ->(work_completion)(&(&slave->notify_work)->work) ->(work_completion)(&(&bond->mcast_work)->work) FD: 232 BD: 1 +.+.: (wq_completion)bond459#3 ->(work_completion)(&(&slave->notify_work)->work) ->(work_completion)(&(&bond->mcast_work)->work) FD: 232 BD: 1 +.+.: (wq_completion)bond460#3 ->&rq->__lock ->(work_completion)(&(&slave->notify_work)->work) ->(work_completion)(&(&bond->mcast_work)->work) FD: 232 BD: 1 +.+.: (wq_completion)bond461#3 ->(work_completion)(&(&slave->notify_work)->work) ->(work_completion)(&(&bond->mcast_work)->work) FD: 232 BD: 1 +.+.: (wq_completion)bond474#2 ->(work_completion)(&(&slave->notify_work)->work) ->(work_completion)(&(&bond->mcast_work)->work) FD: 232 BD: 1 +.+.: (wq_completion)bond484 ->(work_completion)(&(&slave->notify_work)->work) ->(work_completion)(&(&bond->mcast_work)->work) FD: 232 BD: 1 +.+.: (wq_completion)bond156#4 ->(work_completion)(&(&slave->notify_work)->work) ->(work_completion)(&(&bond->mcast_work)->work) ->&rq->__lock FD: 232 BD: 1 +.+.: (wq_completion)bond462#3 ->(work_completion)(&(&slave->notify_work)->work) ->(work_completion)(&(&bond->mcast_work)->work) FD: 232 BD: 1 +.+.: (wq_completion)bond485 ->(work_completion)(&(&slave->notify_work)->work) ->(work_completion)(&(&bond->mcast_work)->work) FD: 232 BD: 1 +.+.: (wq_completion)bond475#2 ->&rq->__lock ->(work_completion)(&(&slave->notify_work)->work) ->(work_completion)(&(&bond->mcast_work)->work) FD: 232 BD: 1 +.+.: (wq_completion)bond463#3 ->(work_completion)(&(&slave->notify_work)->work) ->(work_completion)(&(&bond->mcast_work)->work) ->&rq->__lock FD: 232 BD: 1 +.+.: (wq_completion)bond486 ->&rq->__lock ->(work_completion)(&(&slave->notify_work)->work) ->(work_completion)(&(&bond->mcast_work)->work) FD: 232 BD: 1 +.+.: (wq_completion)bond476#2 ->&rq->__lock ->(work_completion)(&(&slave->notify_work)->work) ->(work_completion)(&(&bond->mcast_work)->work) FD: 232 BD: 1 +.+.: (wq_completion)bond464#3 ->(work_completion)(&(&slave->notify_work)->work) ->(work_completion)(&(&bond->mcast_work)->work) FD: 232 BD: 1 +.+.: (wq_completion)bond487 ->(work_completion)(&(&slave->notify_work)->work) ->(work_completion)(&(&bond->mcast_work)->work) ->&rq->__lock FD: 232 BD: 1 +.+.: (wq_completion)bond477#2 ->(work_completion)(&(&slave->notify_work)->work) ->(work_completion)(&(&bond->mcast_work)->work) FD: 232 BD: 1 +.+.: (wq_completion)bond465#3 ->(work_completion)(&(&slave->notify_work)->work) ->(work_completion)(&(&bond->mcast_work)->work) ->&rq->__lock FD: 232 BD: 1 +.+.: (wq_completion)bond488 ->(work_completion)(&(&slave->notify_work)->work) ->(work_completion)(&(&bond->mcast_work)->work) FD: 232 BD: 1 +.+.: (wq_completion)bond478#2 ->(work_completion)(&(&slave->notify_work)->work) ->(work_completion)(&(&bond->mcast_work)->work) FD: 232 BD: 1 +.+.: (wq_completion)bond466#3 ->(work_completion)(&(&slave->notify_work)->work) ->(work_completion)(&(&bond->mcast_work)->work) FD: 232 BD: 1 +.+.: (wq_completion)bond489 ->(work_completion)(&(&slave->notify_work)->work) ->(work_completion)(&(&bond->mcast_work)->work) FD: 232 BD: 1 +.+.: (wq_completion)bond479#2 ->(work_completion)(&(&slave->notify_work)->work) ->(work_completion)(&(&bond->mcast_work)->work) FD: 232 BD: 1 +.+.: (wq_completion)bond467#3 ->(work_completion)(&(&slave->notify_work)->work) ->(work_completion)(&(&bond->mcast_work)->work) FD: 232 BD: 1 +.+.: (wq_completion)bond490 ->(work_completion)(&(&slave->notify_work)->work) ->(work_completion)(&(&bond->mcast_work)->work) FD: 232 BD: 1 +.+.: (wq_completion)bond480#2 ->&rq->__lock ->(work_completion)(&(&slave->notify_work)->work) ->(work_completion)(&(&bond->mcast_work)->work) FD: 232 BD: 1 +.+.: (wq_completion)bond468#3 ->(work_completion)(&(&slave->notify_work)->work) ->(work_completion)(&(&bond->mcast_work)->work) FD: 232 BD: 1 +.+.: (wq_completion)bond491 ->(work_completion)(&(&slave->notify_work)->work) ->(work_completion)(&(&bond->mcast_work)->work) FD: 232 BD: 1 +.+.: (wq_completion)bond481#2 ->(work_completion)(&(&slave->notify_work)->work) ->(work_completion)(&(&bond->mcast_work)->work) FD: 232 BD: 1 +.+.: (wq_completion)bond469#3 ->&rq->__lock ->(work_completion)(&(&slave->notify_work)->work) ->(work_completion)(&(&bond->mcast_work)->work) FD: 232 BD: 1 +.+.: (wq_completion)bond492 ->(work_completion)(&(&slave->notify_work)->work) ->(work_completion)(&(&bond->mcast_work)->work) ->&rq->__lock FD: 232 BD: 1 +.+.: (wq_completion)bond482#2 ->(work_completion)(&(&slave->notify_work)->work) ->(work_completion)(&(&bond->mcast_work)->work) FD: 232 BD: 1 +.+.: (wq_completion)bond470#3 ->&rq->__lock ->(work_completion)(&(&slave->notify_work)->work) ->(work_completion)(&(&bond->mcast_work)->work) FD: 232 BD: 1 +.+.: (wq_completion)bond493 ->&rq->__lock ->(work_completion)(&(&slave->notify_work)->work) ->(work_completion)(&(&bond->mcast_work)->work) FD: 232 BD: 1 +.+.: (wq_completion)bond471#3 ->(work_completion)(&(&slave->notify_work)->work) ->(work_completion)(&(&bond->mcast_work)->work) FD: 232 BD: 1 +.+.: (wq_completion)bond483#2 ->(work_completion)(&(&slave->notify_work)->work) ->(work_completion)(&(&bond->mcast_work)->work) FD: 232 BD: 1 +.+.: (wq_completion)bond494 ->(work_completion)(&(&slave->notify_work)->work) ->(work_completion)(&(&bond->mcast_work)->work) FD: 232 BD: 1 +.+.: (wq_completion)bond472#3 ->(work_completion)(&(&slave->notify_work)->work) ->(work_completion)(&(&bond->mcast_work)->work) FD: 232 BD: 1 +.+.: (wq_completion)bond495 ->&rq->__lock ->(work_completion)(&(&slave->notify_work)->work) ->(work_completion)(&(&bond->mcast_work)->work) FD: 232 BD: 1 +.+.: (wq_completion)bond473#3 ->(work_completion)(&(&slave->notify_work)->work) ->(work_completion)(&(&bond->mcast_work)->work) FD: 232 BD: 1 +.+.: (wq_completion)bond485#2 ->(work_completion)(&(&slave->notify_work)->work) ->(work_completion)(&(&bond->mcast_work)->work) FD: 232 BD: 1 +.+.: (wq_completion)bond474#3 ->(work_completion)(&(&slave->notify_work)->work) ->(work_completion)(&(&bond->mcast_work)->work) FD: 232 BD: 1 +.+.: (wq_completion)bond486#2 ->(work_completion)(&(&slave->notify_work)->work) ->(work_completion)(&(&bond->mcast_work)->work) FD: 232 BD: 1 +.+.: (wq_completion)bond496 ->&rq->__lock ->(work_completion)(&(&slave->notify_work)->work) ->(work_completion)(&(&bond->mcast_work)->work) FD: 232 BD: 1 +.+.: (wq_completion)bond487#2 ->(work_completion)(&(&slave->notify_work)->work) ->(work_completion)(&(&bond->mcast_work)->work) FD: 232 BD: 1 +.+.: (wq_completion)bond475#3 ->(work_completion)(&(&slave->notify_work)->work) ->(work_completion)(&(&bond->mcast_work)->work) ->&rq->__lock FD: 232 BD: 1 +.+.: (wq_completion)bond497 ->(work_completion)(&(&slave->notify_work)->work) ->(work_completion)(&(&bond->mcast_work)->work) FD: 232 BD: 1 +.+.: (wq_completion)bond488#2 ->(work_completion)(&(&slave->notify_work)->work) ->(work_completion)(&(&bond->mcast_work)->work) FD: 232 BD: 1 +.+.: (wq_completion)bond476#3 ->(work_completion)(&(&slave->notify_work)->work) ->(work_completion)(&(&bond->mcast_work)->work) ->&rq->__lock FD: 232 BD: 1 +.+.: (wq_completion)bond498 ->(work_completion)(&(&slave->notify_work)->work) ->(work_completion)(&(&bond->mcast_work)->work) FD: 232 BD: 1 +.+.: (wq_completion)bond489#2 ->(work_completion)(&(&slave->notify_work)->work) ->(work_completion)(&(&bond->mcast_work)->work) FD: 232 BD: 1 +.+.: (wq_completion)bond477#3 ->(work_completion)(&(&slave->notify_work)->work) ->(work_completion)(&(&bond->mcast_work)->work) FD: 232 BD: 1 +.+.: (wq_completion)bond499 ->(work_completion)(&(&slave->notify_work)->work) ->(work_completion)(&(&bond->mcast_work)->work) ->&rq->__lock FD: 232 BD: 1 +.+.: (wq_completion)bond490#2 ->&rq->__lock ->(work_completion)(&(&slave->notify_work)->work) ->(work_completion)(&(&bond->mcast_work)->work) FD: 232 BD: 1 +.+.: (wq_completion)bond478#3 ->&rq->__lock ->(work_completion)(&(&slave->notify_work)->work) ->(work_completion)(&(&bond->mcast_work)->work) FD: 232 BD: 1 +.+.: (wq_completion)bond500 ->&rq->__lock ->(work_completion)(&(&slave->notify_work)->work) ->(work_completion)(&(&bond->mcast_work)->work) FD: 232 BD: 1 +.+.: (wq_completion)bond491#2 ->(work_completion)(&(&slave->notify_work)->work) ->(work_completion)(&(&bond->mcast_work)->work) FD: 232 BD: 1 +.+.: (wq_completion)bond479#3 ->(work_completion)(&(&slave->notify_work)->work) ->(work_completion)(&(&bond->mcast_work)->work) ->&rq->__lock FD: 232 BD: 1 +.+.: (wq_completion)bond501 ->(work_completion)(&(&slave->notify_work)->work) ->(work_completion)(&(&bond->mcast_work)->work) FD: 232 BD: 1 +.+.: (wq_completion)bond492#2 ->(work_completion)(&(&slave->notify_work)->work) ->(work_completion)(&(&bond->mcast_work)->work) FD: 232 BD: 1 +.+.: (wq_completion)bond480#3 ->(work_completion)(&(&slave->notify_work)->work) ->(work_completion)(&(&bond->mcast_work)->work) FD: 232 BD: 1 +.+.: (wq_completion)bond157#4 ->(work_completion)(&(&slave->notify_work)->work) ->(work_completion)(&(&bond->mcast_work)->work) ->&rq->__lock FD: 232 BD: 1 +.+.: (wq_completion)bond502 ->&rq->__lock ->(work_completion)(&(&slave->notify_work)->work) ->(work_completion)(&(&bond->mcast_work)->work) FD: 232 BD: 1 +.+.: (wq_completion)bond493#2 ->&rq->__lock ->(work_completion)(&(&slave->notify_work)->work) ->(work_completion)(&(&bond->mcast_work)->work) FD: 232 BD: 1 +.+.: (wq_completion)bond481#3 ->(work_completion)(&(&slave->notify_work)->work) ->(work_completion)(&(&bond->mcast_work)->work) FD: 232 BD: 1 +.+.: (wq_completion)bond158#4 ->&rq->__lock ->(work_completion)(&(&slave->notify_work)->work) ->(work_completion)(&(&bond->mcast_work)->work) FD: 232 BD: 1 +.+.: (wq_completion)bond503 ->(work_completion)(&(&slave->notify_work)->work) ->(work_completion)(&(&bond->mcast_work)->work) ->&rq->__lock FD: 232 BD: 1 +.+.: (wq_completion)bond494#2 ->(work_completion)(&(&slave->notify_work)->work) ->(work_completion)(&(&bond->mcast_work)->work) FD: 232 BD: 1 +.+.: (wq_completion)bond482#3 ->(work_completion)(&(&slave->notify_work)->work) ->(work_completion)(&(&bond->mcast_work)->work) FD: 232 BD: 1 +.+.: (wq_completion)bond504 ->(work_completion)(&(&slave->notify_work)->work) ->(work_completion)(&(&bond->mcast_work)->work) FD: 232 BD: 1 +.+.: (wq_completion)bond495#2 ->&rq->__lock ->(work_completion)(&(&slave->notify_work)->work) ->(work_completion)(&(&bond->mcast_work)->work) FD: 232 BD: 1 +.+.: (wq_completion)bond483#3 ->(work_completion)(&(&slave->notify_work)->work) ->(work_completion)(&(&bond->mcast_work)->work) FD: 232 BD: 1 +.+.: (wq_completion)bond505 ->(work_completion)(&(&slave->notify_work)->work) ->(work_completion)(&(&bond->mcast_work)->work) FD: 232 BD: 1 +.+.: (wq_completion)bond496#2 ->(work_completion)(&(&slave->notify_work)->work) ->(work_completion)(&(&bond->mcast_work)->work) ->&rq->__lock FD: 232 BD: 1 +.+.: (wq_completion)bond484#2 ->(work_completion)(&(&slave->notify_work)->work) ->(work_completion)(&(&bond->mcast_work)->work) ->&rq->__lock FD: 232 BD: 1 +.+.: (wq_completion)bond497#2 ->(work_completion)(&(&slave->notify_work)->work) ->(work_completion)(&(&bond->mcast_work)->work) ->&rq->__lock FD: 232 BD: 1 +.+.: (wq_completion)bond506 ->(work_completion)(&(&slave->notify_work)->work) ->(work_completion)(&(&bond->mcast_work)->work) ->&rq->__lock FD: 232 BD: 1 +.+.: (wq_completion)bond485#3 ->(work_completion)(&(&slave->notify_work)->work) ->(work_completion)(&(&bond->mcast_work)->work) FD: 232 BD: 1 +.+.: (wq_completion)bond498#2 ->(work_completion)(&(&slave->notify_work)->work) ->(work_completion)(&(&bond->mcast_work)->work) FD: 232 BD: 1 +.+.: (wq_completion)bond507 ->&rq->__lock ->(work_completion)(&(&slave->notify_work)->work) ->(work_completion)(&(&bond->mcast_work)->work) FD: 232 BD: 1 +.+.: (wq_completion)bond486#3 ->(work_completion)(&(&slave->notify_work)->work) ->(work_completion)(&(&bond->mcast_work)->work) FD: 232 BD: 1 +.+.: (wq_completion)bond159#4 ->(work_completion)(&(&slave->notify_work)->work) ->(work_completion)(&(&bond->mcast_work)->work) ->&rq->__lock FD: 232 BD: 1 +.+.: (wq_completion)bond499#2 ->&rq->__lock ->(work_completion)(&(&slave->notify_work)->work) ->(work_completion)(&(&bond->mcast_work)->work) FD: 232 BD: 1 +.+.: (wq_completion)bond508 ->&rq->__lock ->(work_completion)(&(&slave->notify_work)->work) ->(work_completion)(&(&bond->mcast_work)->work) FD: 232 BD: 1 +.+.: (wq_completion)bond487#3 ->&rq->__lock ->(work_completion)(&(&slave->notify_work)->work) ->(work_completion)(&(&bond->mcast_work)->work) FD: 232 BD: 1 +.+.: (wq_completion)bond500#2 ->(work_completion)(&(&slave->notify_work)->work) ->(work_completion)(&(&bond->mcast_work)->work) ->&rq->__lock FD: 232 BD: 1 +.+.: (wq_completion)bond509 ->(work_completion)(&(&slave->notify_work)->work) ->(work_completion)(&(&bond->mcast_work)->work) ->&rq->__lock FD: 232 BD: 1 +.+.: (wq_completion)bond160#4 ->&rq->__lock ->(work_completion)(&(&slave->notify_work)->work) ->(work_completion)(&(&bond->mcast_work)->work) FD: 232 BD: 1 +.+.: (wq_completion)bond488#3 ->(work_completion)(&(&slave->notify_work)->work) ->(work_completion)(&(&bond->mcast_work)->work) FD: 232 BD: 1 +.+.: (wq_completion)bond501#2 ->(work_completion)(&(&slave->notify_work)->work) ->(work_completion)(&(&bond->mcast_work)->work) FD: 232 BD: 1 +.+.: (wq_completion)bond510 ->(work_completion)(&(&slave->notify_work)->work) ->(work_completion)(&(&bond->mcast_work)->work) ->&rq->__lock FD: 232 BD: 1 +.+.: (wq_completion)bond489#3 ->(work_completion)(&(&slave->notify_work)->work) ->(work_completion)(&(&bond->mcast_work)->work) ->&rq->__lock FD: 232 BD: 1 +.+.: (wq_completion)bond502#2 ->(work_completion)(&(&slave->notify_work)->work) ->(work_completion)(&(&bond->mcast_work)->work) FD: 232 BD: 1 +.+.: (wq_completion)bond511 ->(work_completion)(&(&slave->notify_work)->work) ->(work_completion)(&(&bond->mcast_work)->work) ->&rq->__lock FD: 232 BD: 1 +.+.: (wq_completion)bond490#3 ->&rq->__lock ->(work_completion)(&(&slave->notify_work)->work) ->(work_completion)(&(&bond->mcast_work)->work) FD: 232 BD: 1 +.+.: (wq_completion)bond503#2 ->&rq->__lock ->(work_completion)(&(&slave->notify_work)->work) ->(work_completion)(&(&bond->mcast_work)->work) FD: 232 BD: 1 +.+.: (wq_completion)bond512 ->&rq->__lock ->(work_completion)(&(&slave->notify_work)->work) ->(work_completion)(&(&bond->mcast_work)->work) FD: 232 BD: 1 +.+.: (wq_completion)bond161#4 ->&rq->__lock ->(work_completion)(&(&slave->notify_work)->work) ->(work_completion)(&(&bond->mcast_work)->work) FD: 232 BD: 1 +.+.: (wq_completion)bond491#3 ->&rq->__lock ->(work_completion)(&(&slave->notify_work)->work) ->(work_completion)(&(&bond->mcast_work)->work) FD: 232 BD: 1 +.+.: (wq_completion)bond504#2 ->&rq->__lock ->(work_completion)(&(&slave->notify_work)->work) ->(work_completion)(&(&bond->mcast_work)->work) FD: 232 BD: 1 +.+.: (wq_completion)bond513 ->(work_completion)(&(&slave->notify_work)->work) ->(work_completion)(&(&bond->mcast_work)->work) FD: 232 BD: 1 +.+.: (wq_completion)bond492#3 ->(work_completion)(&(&slave->notify_work)->work) ->(work_completion)(&(&bond->mcast_work)->work) ->&rq->__lock FD: 232 BD: 1 +.+.: (wq_completion)bond505#2 ->&rq->__lock ->(work_completion)(&(&slave->notify_work)->work) ->(work_completion)(&(&bond->mcast_work)->work) FD: 232 BD: 1 +.+.: (wq_completion)bond493#3 ->(work_completion)(&(&slave->notify_work)->work) ->(work_completion)(&(&bond->mcast_work)->work) FD: 232 BD: 1 +.+.: (wq_completion)bond514 ->&rq->__lock ->(work_completion)(&(&slave->notify_work)->work) ->(work_completion)(&(&bond->mcast_work)->work) FD: 232 BD: 1 +.+.: (wq_completion)bond506#2 ->(work_completion)(&(&slave->notify_work)->work) ->(work_completion)(&(&bond->mcast_work)->work) ->&rq->__lock FD: 232 BD: 1 +.+.: (wq_completion)bond515 ->(work_completion)(&(&slave->notify_work)->work) ->(work_completion)(&(&bond->mcast_work)->work) ->&rq->__lock FD: 232 BD: 1 +.+.: (wq_completion)bond162#4 ->(work_completion)(&(&slave->notify_work)->work) ->(work_completion)(&(&bond->mcast_work)->work) ->&rq->__lock FD: 232 BD: 1 +.+.: (wq_completion)bond494#3 ->(work_completion)(&(&slave->notify_work)->work) ->(work_completion)(&(&bond->mcast_work)->work) ->&rq->__lock FD: 232 BD: 1 +.+.: (wq_completion)bond507#2 ->(work_completion)(&(&slave->notify_work)->work) ->(work_completion)(&(&bond->mcast_work)->work) FD: 232 BD: 1 +.+.: (wq_completion)bond516 ->&rq->__lock ->(work_completion)(&(&slave->notify_work)->work) ->(work_completion)(&(&bond->mcast_work)->work) FD: 232 BD: 1 +.+.: (wq_completion)bond495#3 ->&rq->__lock ->(work_completion)(&(&slave->notify_work)->work) ->(work_completion)(&(&bond->mcast_work)->work) FD: 232 BD: 1 +.+.: (wq_completion)bond508#2 ->&rq->__lock ->(work_completion)(&(&slave->notify_work)->work) ->(work_completion)(&(&bond->mcast_work)->work) FD: 232 BD: 1 +.+.: (wq_completion)bond517 ->&rq->__lock ->(work_completion)(&(&slave->notify_work)->work) ->(work_completion)(&(&bond->mcast_work)->work) FD: 232 BD: 1 +.+.: (wq_completion)bond496#3 ->(work_completion)(&(&slave->notify_work)->work) ->(work_completion)(&(&bond->mcast_work)->work) FD: 232 BD: 1 +.+.: (wq_completion)bond163#4 ->(work_completion)(&(&slave->notify_work)->work) ->(work_completion)(&(&bond->mcast_work)->work) FD: 232 BD: 1 +.+.: (wq_completion)bond509#2 ->(work_completion)(&(&slave->notify_work)->work) ->(work_completion)(&(&bond->mcast_work)->work) FD: 232 BD: 1 +.+.: (wq_completion)bond518 ->&rq->__lock ->(work_completion)(&(&slave->notify_work)->work) ->(work_completion)(&(&bond->mcast_work)->work) FD: 232 BD: 1 +.+.: (wq_completion)bond497#3 ->(work_completion)(&(&slave->notify_work)->work) ->(work_completion)(&(&bond->mcast_work)->work) FD: 232 BD: 1 +.+.: (wq_completion)bond164#4 ->(work_completion)(&(&slave->notify_work)->work) ->(work_completion)(&(&bond->mcast_work)->work) FD: 232 BD: 1 +.+.: (wq_completion)bond510#2 ->(work_completion)(&(&slave->notify_work)->work) ->(work_completion)(&(&bond->mcast_work)->work) FD: 232 BD: 1 +.+.: (wq_completion)bond519 ->&rq->__lock ->(work_completion)(&(&slave->notify_work)->work) ->(work_completion)(&(&bond->mcast_work)->work) FD: 232 BD: 1 +.+.: (wq_completion)bond498#3 ->(work_completion)(&(&slave->notify_work)->work) ->(work_completion)(&(&bond->mcast_work)->work) FD: 232 BD: 1 +.+.: (wq_completion)bond520 ->(work_completion)(&(&slave->notify_work)->work) ->(work_completion)(&(&bond->mcast_work)->work) ->&rq->__lock FD: 232 BD: 1 +.+.: (wq_completion)bond512#2 ->(work_completion)(&(&slave->notify_work)->work) ->(work_completion)(&(&bond->mcast_work)->work) ->&rq->__lock FD: 232 BD: 1 +.+.: (wq_completion)bond499#3 ->&rq->__lock ->(work_completion)(&(&slave->notify_work)->work) ->(work_completion)(&(&bond->mcast_work)->work) FD: 232 BD: 1 +.+.: (wq_completion)bond521 ->(work_completion)(&(&slave->notify_work)->work) ->(work_completion)(&(&bond->mcast_work)->work) FD: 232 BD: 1 +.+.: (wq_completion)bond513#2 ->(work_completion)(&(&slave->notify_work)->work) ->(work_completion)(&(&bond->mcast_work)->work) ->&rq->__lock FD: 232 BD: 1 +.+.: (wq_completion)bond500#3 ->&rq->__lock ->(work_completion)(&(&slave->notify_work)->work) ->(work_completion)(&(&bond->mcast_work)->work) FD: 232 BD: 1 +.+.: (wq_completion)bond522 ->&rq->__lock ->(work_completion)(&(&slave->notify_work)->work) ->(work_completion)(&(&bond->mcast_work)->work) FD: 232 BD: 1 +.+.: (wq_completion)bond514#2 ->(work_completion)(&(&slave->notify_work)->work) ->(work_completion)(&(&bond->mcast_work)->work) ->&rq->__lock FD: 232 BD: 1 +.+.: (wq_completion)bond501#3 ->(work_completion)(&(&slave->notify_work)->work) ->(work_completion)(&(&bond->mcast_work)->work) FD: 232 BD: 1 +.+.: (wq_completion)bond523 ->(work_completion)(&(&slave->notify_work)->work) ->(work_completion)(&(&bond->mcast_work)->work) FD: 232 BD: 1 +.+.: (wq_completion)bond515#2 ->(work_completion)(&(&slave->notify_work)->work) ->(work_completion)(&(&bond->mcast_work)->work) ->&rq->__lock FD: 232 BD: 1 +.+.: (wq_completion)bond502#3 ->&rq->__lock ->(work_completion)(&(&slave->notify_work)->work) ->(work_completion)(&(&bond->mcast_work)->work) FD: 232 BD: 1 +.+.: (wq_completion)bond524 ->(work_completion)(&(&slave->notify_work)->work) ->(work_completion)(&(&bond->mcast_work)->work) ->&rq->__lock FD: 232 BD: 1 +.+.: (wq_completion)bond516#2 ->(work_completion)(&(&slave->notify_work)->work) ->(work_completion)(&(&bond->mcast_work)->work) FD: 232 BD: 1 +.+.: (wq_completion)bond503#3 ->(work_completion)(&(&slave->notify_work)->work) ->(work_completion)(&(&bond->mcast_work)->work) ->&rq->__lock FD: 232 BD: 1 +.+.: (wq_completion)bond525 ->(work_completion)(&(&slave->notify_work)->work) ->(work_completion)(&(&bond->mcast_work)->work) ->&rq->__lock FD: 232 BD: 1 +.+.: (wq_completion)bond165#4 ->(work_completion)(&(&slave->notify_work)->work) ->(work_completion)(&(&bond->mcast_work)->work) ->&rq->__lock FD: 232 BD: 1 +.+.: (wq_completion)bond517#2 ->&rq->__lock ->(work_completion)(&(&slave->notify_work)->work) ->(work_completion)(&(&bond->mcast_work)->work) FD: 232 BD: 1 +.+.: (wq_completion)bond504#3 ->(work_completion)(&(&slave->notify_work)->work) ->(work_completion)(&(&bond->mcast_work)->work) ->&rq->__lock FD: 232 BD: 1 +.+.: (wq_completion)bond526 ->&rq->__lock ->(work_completion)(&(&slave->notify_work)->work) ->(work_completion)(&(&bond->mcast_work)->work) FD: 232 BD: 1 +.+.: (wq_completion)bond518#2 ->&rq->__lock ->(work_completion)(&(&slave->notify_work)->work) ->(work_completion)(&(&bond->mcast_work)->work) FD: 232 BD: 1 +.+.: (wq_completion)bond505#3 ->&rq->__lock ->(work_completion)(&(&slave->notify_work)->work) ->(work_completion)(&(&bond->mcast_work)->work) FD: 232 BD: 1 +.+.: (wq_completion)bond527 ->(work_completion)(&(&slave->notify_work)->work) ->(work_completion)(&(&bond->mcast_work)->work) ->&rq->__lock FD: 232 BD: 1 +.+.: (wq_completion)bond519#2 ->(work_completion)(&(&slave->notify_work)->work) ->(work_completion)(&(&bond->mcast_work)->work) ->&rq->__lock FD: 232 BD: 1 +.+.: (wq_completion)bond506#3 ->&rq->__lock ->(work_completion)(&(&slave->notify_work)->work) ->(work_completion)(&(&bond->mcast_work)->work) FD: 232 BD: 1 +.+.: (wq_completion)bond528 ->&rq->__lock ->(work_completion)(&(&slave->notify_work)->work) ->(work_completion)(&(&bond->mcast_work)->work) FD: 232 BD: 1 +.+.: (wq_completion)bond520#2 ->&rq->__lock ->(work_completion)(&(&slave->notify_work)->work) ->(work_completion)(&(&bond->mcast_work)->work) FD: 232 BD: 1 +.+.: (wq_completion)bond507#3 ->(work_completion)(&(&slave->notify_work)->work) ->(work_completion)(&(&bond->mcast_work)->work) ->&rq->__lock FD: 232 BD: 1 +.+.: (wq_completion)bond529 ->&rq->__lock ->(work_completion)(&(&slave->notify_work)->work) ->(work_completion)(&(&bond->mcast_work)->work) FD: 232 BD: 1 +.+.: (wq_completion)bond521#2 ->(work_completion)(&(&slave->notify_work)->work) ->(work_completion)(&(&bond->mcast_work)->work) FD: 232 BD: 1 +.+.: (wq_completion)bond166#4 ->(work_completion)(&(&slave->notify_work)->work) ->(work_completion)(&(&bond->mcast_work)->work) FD: 232 BD: 1 +.+.: (wq_completion)bond508#3 ->&rq->__lock ->(work_completion)(&(&slave->notify_work)->work) ->(work_completion)(&(&bond->mcast_work)->work) FD: 232 BD: 1 +.+.: (wq_completion)bond530 ->(work_completion)(&(&slave->notify_work)->work) ->(work_completion)(&(&bond->mcast_work)->work) ->&rq->__lock FD: 232 BD: 1 +.+.: (wq_completion)bond522#2 ->&rq->__lock ->(work_completion)(&(&slave->notify_work)->work) ->(work_completion)(&(&bond->mcast_work)->work) FD: 232 BD: 1 +.+.: (wq_completion)bond167#4 ->(work_completion)(&(&slave->notify_work)->work) ->(work_completion)(&(&bond->mcast_work)->work) ->&rq->__lock FD: 232 BD: 1 +.+.: (wq_completion)bond509#3 ->&rq->__lock ->(work_completion)(&(&slave->notify_work)->work) ->(work_completion)(&(&bond->mcast_work)->work) FD: 232 BD: 1 +.+.: (wq_completion)bond531 ->&rq->__lock ->(work_completion)(&(&slave->notify_work)->work) ->(work_completion)(&(&bond->mcast_work)->work) FD: 232 BD: 1 +.+.: (wq_completion)bond523#2 ->(work_completion)(&(&slave->notify_work)->work) ->(work_completion)(&(&bond->mcast_work)->work) ->&rq->__lock FD: 232 BD: 1 +.+.: (wq_completion)bond168#4 ->(work_completion)(&(&slave->notify_work)->work) ->(work_completion)(&(&bond->mcast_work)->work) FD: 232 BD: 1 +.+.: (wq_completion)bond510#3 ->(work_completion)(&(&slave->notify_work)->work) ->(work_completion)(&(&bond->mcast_work)->work) FD: 232 BD: 1 +.+.: (wq_completion)bond532 ->&rq->__lock ->(work_completion)(&(&slave->notify_work)->work) ->(work_completion)(&(&bond->mcast_work)->work) FD: 232 BD: 1 +.+.: (wq_completion)bond524#2 ->(work_completion)(&(&slave->notify_work)->work) ->(work_completion)(&(&bond->mcast_work)->work) ->&rq->__lock FD: 232 BD: 1 +.+.: (wq_completion)bond511#2 ->&rq->__lock ->(work_completion)(&(&slave->notify_work)->work) ->(work_completion)(&(&bond->mcast_work)->work) FD: 232 BD: 1 +.+.: (wq_completion)bond533 ->(work_completion)(&(&slave->notify_work)->work) ->(work_completion)(&(&bond->mcast_work)->work) FD: 232 BD: 1 +.+.: (wq_completion)bond525#2 ->(work_completion)(&(&slave->notify_work)->work) ->(work_completion)(&(&bond->mcast_work)->work) ->&rq->__lock FD: 232 BD: 1 +.+.: (wq_completion)bond534 ->&rq->__lock ->(work_completion)(&(&slave->notify_work)->work) ->(work_completion)(&(&bond->mcast_work)->work) FD: 232 BD: 1 +.+.: (wq_completion)bond526#2 ->(work_completion)(&(&slave->notify_work)->work) ->(work_completion)(&(&bond->mcast_work)->work) FD: 232 BD: 1 +.+.: (wq_completion)bond169#4 ->(work_completion)(&(&slave->notify_work)->work) ->(work_completion)(&(&bond->mcast_work)->work) ->&rq->__lock FD: 232 BD: 1 +.+.: (wq_completion)bond535 ->(work_completion)(&(&slave->notify_work)->work) ->(work_completion)(&(&bond->mcast_work)->work) FD: 232 BD: 1 +.+.: (wq_completion)bond527#2 ->&rq->__lock ->(work_completion)(&(&slave->notify_work)->work) ->(work_completion)(&(&bond->mcast_work)->work) FD: 232 BD: 1 +.+.: (wq_completion)bond514#3 ->&rq->__lock ->(work_completion)(&(&slave->notify_work)->work) ->(work_completion)(&(&bond->mcast_work)->work) FD: 232 BD: 1 +.+.: (wq_completion)bond170#4 ->&rq->__lock ->&cfs_rq->removed.lock ->&obj_hash[i].lock ->(work_completion)(&(&slave->notify_work)->work) ->(work_completion)(&(&bond->mcast_work)->work) FD: 232 BD: 1 +.+.: (wq_completion)bond536 ->&rq->__lock ->(work_completion)(&(&slave->notify_work)->work) ->(work_completion)(&(&bond->mcast_work)->work) FD: 232 BD: 1 +.+.: (wq_completion)bond528#2 ->(work_completion)(&(&slave->notify_work)->work) ->(work_completion)(&(&bond->mcast_work)->work) ->&rq->__lock FD: 232 BD: 1 +.+.: (wq_completion)bond515#3 ->(work_completion)(&(&slave->notify_work)->work) ->(work_completion)(&(&bond->mcast_work)->work) ->&rq->__lock FD: 232 BD: 1 +.+.: (wq_completion)bond171#4 ->&rq->__lock ->(work_completion)(&(&slave->notify_work)->work) ->(work_completion)(&(&bond->mcast_work)->work) FD: 232 BD: 1 +.+.: (wq_completion)bond537 ->(work_completion)(&(&slave->notify_work)->work) ->(work_completion)(&(&bond->mcast_work)->work) FD: 232 BD: 1 +.+.: (wq_completion)bond529#2 ->&rq->__lock ->(work_completion)(&(&slave->notify_work)->work) ->(work_completion)(&(&bond->mcast_work)->work) FD: 232 BD: 1 +.+.: (wq_completion)bond516#3 ->(work_completion)(&(&slave->notify_work)->work) ->(work_completion)(&(&bond->mcast_work)->work) FD: 232 BD: 1 +.+.: (wq_completion)bond538 ->&rq->__lock ->(work_completion)(&(&slave->notify_work)->work) ->(work_completion)(&(&bond->mcast_work)->work) FD: 232 BD: 1 +.+.: (wq_completion)bond172#4 ->&rq->__lock ->(work_completion)(&(&slave->notify_work)->work) ->(work_completion)(&(&bond->mcast_work)->work) FD: 232 BD: 1 +.+.: (wq_completion)bond530#2 ->&rq->__lock ->(work_completion)(&(&slave->notify_work)->work) ->(work_completion)(&(&bond->mcast_work)->work) FD: 232 BD: 1 +.+.: (wq_completion)bond517#3 ->&rq->__lock ->(work_completion)(&(&slave->notify_work)->work) ->(work_completion)(&(&bond->mcast_work)->work) FD: 232 BD: 1 +.+.: (wq_completion)bond539 ->(work_completion)(&(&slave->notify_work)->work) ->(work_completion)(&(&bond->mcast_work)->work) FD: 232 BD: 1 +.+.: (wq_completion)bond531#2 ->(work_completion)(&(&slave->notify_work)->work) ->(work_completion)(&(&bond->mcast_work)->work) ->&rq->__lock FD: 232 BD: 1 +.+.: (wq_completion)bond540 ->&rq->__lock ->(work_completion)(&(&slave->notify_work)->work) ->(work_completion)(&(&bond->mcast_work)->work) FD: 232 BD: 1 +.+.: (wq_completion)bond519#3 ->(work_completion)(&(&slave->notify_work)->work) ->(work_completion)(&(&bond->mcast_work)->work) FD: 232 BD: 1 +.+.: (wq_completion)bond532#2 ->(work_completion)(&(&slave->notify_work)->work) ->(work_completion)(&(&bond->mcast_work)->work) ->&rq->__lock FD: 232 BD: 1 +.+.: (wq_completion)bond173#4 ->(work_completion)(&(&slave->notify_work)->work) ->(work_completion)(&(&bond->mcast_work)->work) FD: 232 BD: 1 +.+.: (wq_completion)bond541 ->(work_completion)(&(&slave->notify_work)->work) ->(work_completion)(&(&bond->mcast_work)->work) FD: 232 BD: 1 +.+.: (wq_completion)bond520#3 ->(work_completion)(&(&slave->notify_work)->work) ->(work_completion)(&(&bond->mcast_work)->work) FD: 232 BD: 1 +.+.: (wq_completion)bond533#2 ->&rq->__lock ->(work_completion)(&(&slave->notify_work)->work) ->(work_completion)(&(&bond->mcast_work)->work) FD: 232 BD: 1 +.+.: (wq_completion)bond521#3 ->&rq->__lock ->(work_completion)(&(&slave->notify_work)->work) ->(work_completion)(&(&bond->mcast_work)->work) FD: 232 BD: 1 +.+.: (wq_completion)bond534#2 ->(work_completion)(&(&slave->notify_work)->work) ->(work_completion)(&(&bond->mcast_work)->work) ->&rq->__lock FD: 232 BD: 1 +.+.: (wq_completion)bond543 ->&rq->__lock ->(work_completion)(&(&slave->notify_work)->work) ->(work_completion)(&(&bond->mcast_work)->work) FD: 232 BD: 1 +.+.: (wq_completion)bond522#3 ->(work_completion)(&(&slave->notify_work)->work) ->(work_completion)(&(&bond->mcast_work)->work) FD: 232 BD: 1 +.+.: (wq_completion)bond535#2 ->(work_completion)(&(&slave->notify_work)->work) ->(work_completion)(&(&bond->mcast_work)->work) FD: 232 BD: 1 +.+.: (wq_completion)bond544 ->(work_completion)(&(&slave->notify_work)->work) ->(work_completion)(&(&bond->mcast_work)->work) FD: 232 BD: 1 +.+.: (wq_completion)bond523#3 ->&rq->__lock ->(work_completion)(&(&slave->notify_work)->work) ->(work_completion)(&(&bond->mcast_work)->work) FD: 232 BD: 1 +.+.: (wq_completion)bond536#2 ->&rq->__lock ->(work_completion)(&(&slave->notify_work)->work) ->(work_completion)(&(&bond->mcast_work)->work) FD: 232 BD: 1 +.+.: (wq_completion)bond545 ->(work_completion)(&(&slave->notify_work)->work) ->(work_completion)(&(&bond->mcast_work)->work) FD: 232 BD: 1 +.+.: (wq_completion)bond524#3 ->(work_completion)(&(&slave->notify_work)->work) ->(work_completion)(&(&bond->mcast_work)->work) ->&rq->__lock FD: 232 BD: 1 +.+.: (wq_completion)bond537#2 ->(work_completion)(&(&slave->notify_work)->work) ->(work_completion)(&(&bond->mcast_work)->work) FD: 232 BD: 1 +.+.: (wq_completion)bond546 ->&rq->__lock ->(work_completion)(&(&slave->notify_work)->work) ->(work_completion)(&(&bond->mcast_work)->work) FD: 232 BD: 1 +.+.: (wq_completion)bond525#3 ->(work_completion)(&(&slave->notify_work)->work) ->(work_completion)(&(&bond->mcast_work)->work) FD: 232 BD: 1 +.+.: (wq_completion)bond538#2 ->&rq->__lock ->(work_completion)(&(&slave->notify_work)->work) ->(work_completion)(&(&bond->mcast_work)->work) FD: 232 BD: 1 +.+.: (wq_completion)bond547 ->&rq->__lock ->(work_completion)(&(&slave->notify_work)->work) ->(work_completion)(&(&bond->mcast_work)->work) FD: 232 BD: 1 +.+.: (wq_completion)bond526#3 ->&rq->__lock ->(work_completion)(&(&slave->notify_work)->work) ->(work_completion)(&(&bond->mcast_work)->work) FD: 232 BD: 1 +.+.: (wq_completion)bond539#2 ->(work_completion)(&(&slave->notify_work)->work) ->(work_completion)(&(&bond->mcast_work)->work) ->&rq->__lock FD: 232 BD: 1 +.+.: (wq_completion)bond548 ->(work_completion)(&(&slave->notify_work)->work) ->(work_completion)(&(&bond->mcast_work)->work) ->&rq->__lock FD: 232 BD: 1 +.+.: (wq_completion)bond527#3 ->&rq->__lock ->(work_completion)(&(&slave->notify_work)->work) ->(work_completion)(&(&bond->mcast_work)->work) FD: 232 BD: 1 +.+.: (wq_completion)bond540#2 ->(work_completion)(&(&slave->notify_work)->work) ->(work_completion)(&(&bond->mcast_work)->work) FD: 232 BD: 1 +.+.: (wq_completion)bond549 ->(work_completion)(&(&slave->notify_work)->work) ->(work_completion)(&(&bond->mcast_work)->work) FD: 232 BD: 1 +.+.: (wq_completion)bond528#3 ->(work_completion)(&(&slave->notify_work)->work) ->(work_completion)(&(&bond->mcast_work)->work) ->&rq->__lock FD: 232 BD: 1 +.+.: (wq_completion)bond541#2 ->&rq->__lock ->(work_completion)(&(&slave->notify_work)->work) ->(work_completion)(&(&bond->mcast_work)->work) FD: 232 BD: 1 +.+.: (wq_completion)bond550 ->(work_completion)(&(&slave->notify_work)->work) ->(work_completion)(&(&bond->mcast_work)->work) ->&rq->__lock FD: 232 BD: 1 +.+.: (wq_completion)bond542 ->&rq->__lock ->(work_completion)(&(&slave->notify_work)->work) ->(work_completion)(&(&bond->mcast_work)->work) FD: 232 BD: 1 +.+.: (wq_completion)bond529#3 ->(work_completion)(&(&slave->notify_work)->work) ->(work_completion)(&(&bond->mcast_work)->work) FD: 232 BD: 1 +.+.: (wq_completion)bond551 ->&rq->__lock ->(work_completion)(&(&slave->notify_work)->work) ->(work_completion)(&(&bond->mcast_work)->work) FD: 232 BD: 1 +.+.: (wq_completion)bond174#4 ->&rq->__lock ->(work_completion)(&(&slave->notify_work)->work) ->(work_completion)(&(&bond->mcast_work)->work) FD: 232 BD: 1 +.+.: (wq_completion)bond530#3 ->&rq->__lock ->(work_completion)(&(&slave->notify_work)->work) ->(work_completion)(&(&bond->mcast_work)->work) FD: 232 BD: 1 +.+.: (wq_completion)bond543#2 ->(work_completion)(&(&slave->notify_work)->work) ->(work_completion)(&(&bond->mcast_work)->work) ->&rq->__lock FD: 232 BD: 1 +.+.: (wq_completion)bond552 ->(work_completion)(&(&slave->notify_work)->work) ->(work_completion)(&(&bond->mcast_work)->work) FD: 232 BD: 1 +.+.: (wq_completion)bond531#3 ->(work_completion)(&(&slave->notify_work)->work) ->(work_completion)(&(&bond->mcast_work)->work) FD: 232 BD: 1 +.+.: (wq_completion)bond544#2 ->&rq->__lock ->(work_completion)(&(&slave->notify_work)->work) ->(work_completion)(&(&bond->mcast_work)->work) FD: 232 BD: 1 +.+.: (wq_completion)bond553 ->(work_completion)(&(&slave->notify_work)->work) ->(work_completion)(&(&bond->mcast_work)->work) FD: 232 BD: 1 +.+.: (wq_completion)bond175#3 ->&rq->__lock ->(work_completion)(&(&slave->notify_work)->work) ->(work_completion)(&(&bond->mcast_work)->work) FD: 232 BD: 1 +.+.: (wq_completion)bond532#3 ->(work_completion)(&(&slave->notify_work)->work) ->(work_completion)(&(&bond->mcast_work)->work) FD: 232 BD: 1 +.+.: (wq_completion)bond545#2 ->&rq->__lock ->(work_completion)(&(&slave->notify_work)->work) ->(work_completion)(&(&bond->mcast_work)->work) FD: 232 BD: 1 +.+.: (wq_completion)bond554 ->&rq->__lock ->(work_completion)(&(&slave->notify_work)->work) ->(work_completion)(&(&bond->mcast_work)->work) FD: 232 BD: 1 +.+.: (wq_completion)bond546#2 ->&rq->__lock ->(work_completion)(&(&slave->notify_work)->work) ->(work_completion)(&(&bond->mcast_work)->work) FD: 232 BD: 1 +.+.: (wq_completion)bond533#3 ->&rq->__lock ->(work_completion)(&(&slave->notify_work)->work) ->(work_completion)(&(&bond->mcast_work)->work) FD: 232 BD: 1 +.+.: (wq_completion)bond555 ->&rq->__lock ->(work_completion)(&(&slave->notify_work)->work) ->(work_completion)(&(&bond->mcast_work)->work) FD: 232 BD: 1 +.+.: (wq_completion)bond547#2 ->&rq->__lock ->(work_completion)(&(&slave->notify_work)->work) ->(work_completion)(&(&bond->mcast_work)->work) FD: 232 BD: 1 +.+.: (wq_completion)bond534#3 ->&rq->__lock ->(work_completion)(&(&slave->notify_work)->work) ->(work_completion)(&(&bond->mcast_work)->work) FD: 232 BD: 1 +.+.: (wq_completion)bond176#4 ->(work_completion)(&(&slave->notify_work)->work) ->(work_completion)(&(&bond->mcast_work)->work) ->&rq->__lock FD: 232 BD: 1 +.+.: (wq_completion)bond548#2 ->(work_completion)(&(&slave->notify_work)->work) ->(work_completion)(&(&bond->mcast_work)->work) FD: 232 BD: 1 +.+.: (wq_completion)bond535#3 ->(work_completion)(&(&slave->notify_work)->work) ->(work_completion)(&(&bond->mcast_work)->work) FD: 232 BD: 1 +.+.: (wq_completion)bond557 ->(work_completion)(&(&slave->notify_work)->work) ->(work_completion)(&(&bond->mcast_work)->work) ->&rq->__lock FD: 232 BD: 1 +.+.: (wq_completion)bond549#2 ->&rq->__lock ->(work_completion)(&(&slave->notify_work)->work) ->(work_completion)(&(&bond->mcast_work)->work) FD: 232 BD: 1 +.+.: (wq_completion)bond536#3 ->&rq->__lock ->(work_completion)(&(&slave->notify_work)->work) ->(work_completion)(&(&bond->mcast_work)->work) FD: 232 BD: 1 +.+.: (wq_completion)bond558 ->(work_completion)(&(&slave->notify_work)->work) ->(work_completion)(&(&bond->mcast_work)->work) FD: 232 BD: 1 +.+.: (wq_completion)bond550#2 ->&rq->__lock ->(work_completion)(&(&slave->notify_work)->work) ->(work_completion)(&(&bond->mcast_work)->work) FD: 232 BD: 1 +.+.: (wq_completion)bond537#3 ->&rq->__lock ->(work_completion)(&(&slave->notify_work)->work) ->(work_completion)(&(&bond->mcast_work)->work) FD: 232 BD: 1 +.+.: (wq_completion)bond559 ->&rq->__lock ->(work_completion)(&(&slave->notify_work)->work) ->(work_completion)(&(&bond->mcast_work)->work) FD: 232 BD: 1 +.+.: (wq_completion)bond551#2 ->&rq->__lock ->(work_completion)(&(&slave->notify_work)->work) ->(work_completion)(&(&bond->mcast_work)->work) FD: 232 BD: 1 +.+.: (wq_completion)bond538#3 ->&rq->__lock ->(work_completion)(&(&slave->notify_work)->work) ->(work_completion)(&(&bond->mcast_work)->work) FD: 232 BD: 1 +.+.: (wq_completion)bond177#4 ->(work_completion)(&(&slave->notify_work)->work) ->(work_completion)(&(&bond->mcast_work)->work) FD: 232 BD: 1 +.+.: (wq_completion)bond560 ->(work_completion)(&(&slave->notify_work)->work) ->(work_completion)(&(&bond->mcast_work)->work) FD: 232 BD: 1 +.+.: (wq_completion)bond552#2 ->(work_completion)(&(&slave->notify_work)->work) ->(work_completion)(&(&bond->mcast_work)->work) FD: 232 BD: 1 +.+.: (wq_completion)bond539#3 ->&rq->__lock ->(work_completion)(&(&slave->notify_work)->work) ->(work_completion)(&(&bond->mcast_work)->work) FD: 232 BD: 1 +.+.: (wq_completion)bond561 ->&rq->__lock ->(work_completion)(&(&slave->notify_work)->work) ->(work_completion)(&(&bond->mcast_work)->work) FD: 232 BD: 1 +.+.: (wq_completion)bond553#2 ->&rq->__lock ->(work_completion)(&(&slave->notify_work)->work) ->(work_completion)(&(&bond->mcast_work)->work) FD: 232 BD: 1 +.+.: (wq_completion)bond540#3 ->&rq->__lock ->(work_completion)(&(&slave->notify_work)->work) ->(work_completion)(&(&bond->mcast_work)->work) FD: 232 BD: 1 +.+.: (wq_completion)bond178#4 ->&rq->__lock ->(work_completion)(&(&slave->notify_work)->work) ->(work_completion)(&(&bond->mcast_work)->work) FD: 232 BD: 1 +.+.: (wq_completion)bond562 ->&rq->__lock ->(work_completion)(&(&slave->notify_work)->work) ->(work_completion)(&(&bond->mcast_work)->work) FD: 232 BD: 1 +.+.: (wq_completion)bond554#2 ->(work_completion)(&(&slave->notify_work)->work) ->(work_completion)(&(&bond->mcast_work)->work) FD: 232 BD: 1 +.+.: (wq_completion)bond541#3 ->(work_completion)(&(&slave->notify_work)->work) ->(work_completion)(&(&bond->mcast_work)->work) FD: 232 BD: 1 +.+.: (wq_completion)bond563 ->&rq->__lock ->(work_completion)(&(&slave->notify_work)->work) ->(work_completion)(&(&bond->mcast_work)->work) FD: 232 BD: 1 +.+.: (wq_completion)bond555#2 ->&rq->__lock ->(work_completion)(&(&slave->notify_work)->work) ->(work_completion)(&(&bond->mcast_work)->work) FD: 232 BD: 1 +.+.: (wq_completion)bond542#2 ->(work_completion)(&(&slave->notify_work)->work) ->(work_completion)(&(&bond->mcast_work)->work) FD: 232 BD: 1 +.+.: (wq_completion)bond564 ->(work_completion)(&(&slave->notify_work)->work) ->(work_completion)(&(&bond->mcast_work)->work) FD: 232 BD: 1 +.+.: (wq_completion)bond556 ->(work_completion)(&(&slave->notify_work)->work) ->(work_completion)(&(&bond->mcast_work)->work) FD: 232 BD: 1 +.+.: (wq_completion)bond543#3 ->&rq->__lock ->(work_completion)(&(&slave->notify_work)->work) ->(work_completion)(&(&bond->mcast_work)->work) FD: 232 BD: 1 +.+.: (wq_completion)bond565 ->(work_completion)(&(&slave->notify_work)->work) ->(work_completion)(&(&bond->mcast_work)->work) FD: 232 BD: 1 +.+.: (wq_completion)bond557#2 ->(work_completion)(&(&slave->notify_work)->work) ->(work_completion)(&(&bond->mcast_work)->work) FD: 232 BD: 1 +.+.: (wq_completion)bond544#3 ->(work_completion)(&(&slave->notify_work)->work) ->(work_completion)(&(&bond->mcast_work)->work) FD: 232 BD: 1 +.+.: (wq_completion)bond179#4 ->(work_completion)(&(&slave->notify_work)->work) ->(work_completion)(&(&bond->mcast_work)->work) FD: 232 BD: 1 +.+.: (wq_completion)bond566 ->(work_completion)(&(&slave->notify_work)->work) ->(work_completion)(&(&bond->mcast_work)->work) FD: 232 BD: 1 +.+.: (wq_completion)bond558#2 ->&rq->__lock ->(work_completion)(&(&slave->notify_work)->work) ->(work_completion)(&(&bond->mcast_work)->work) FD: 232 BD: 1 +.+.: (wq_completion)bond545#3 ->&rq->__lock ->(work_completion)(&(&slave->notify_work)->work) ->(work_completion)(&(&bond->mcast_work)->work) FD: 232 BD: 1 +.+.: (wq_completion)bond567 ->&rq->__lock ->(work_completion)(&(&slave->notify_work)->work) ->(work_completion)(&(&bond->mcast_work)->work) FD: 232 BD: 1 +.+.: (wq_completion)bond559#2 ->&rq->__lock ->(work_completion)(&(&slave->notify_work)->work) ->(work_completion)(&(&bond->mcast_work)->work) FD: 232 BD: 1 +.+.: (wq_completion)bond180#4 ->&rq->__lock ->(work_completion)(&(&slave->notify_work)->work) ->(work_completion)(&(&bond->mcast_work)->work) FD: 232 BD: 1 +.+.: (wq_completion)bond546#3 ->&rq->__lock ->(work_completion)(&(&slave->notify_work)->work) ->(work_completion)(&(&bond->mcast_work)->work) FD: 232 BD: 1 +.+.: (wq_completion)bond568 ->(work_completion)(&(&slave->notify_work)->work) ->(work_completion)(&(&bond->mcast_work)->work) ->&rq->__lock FD: 232 BD: 1 +.+.: (wq_completion)bond560#2 ->(work_completion)(&(&slave->notify_work)->work) ->(work_completion)(&(&bond->mcast_work)->work) FD: 232 BD: 1 +.+.: (wq_completion)bond547#3 ->&rq->__lock ->(work_completion)(&(&slave->notify_work)->work) ->(work_completion)(&(&bond->mcast_work)->work) FD: 232 BD: 1 +.+.: (wq_completion)bond569 ->&rq->__lock ->(work_completion)(&(&slave->notify_work)->work) ->(work_completion)(&(&bond->mcast_work)->work) FD: 232 BD: 1 +.+.: (wq_completion)bond561#2 ->(work_completion)(&(&slave->notify_work)->work) ->(work_completion)(&(&bond->mcast_work)->work) FD: 232 BD: 1 +.+.: (wq_completion)bond548#3 ->(work_completion)(&(&slave->notify_work)->work) ->(work_completion)(&(&bond->mcast_work)->work) FD: 232 BD: 1 +.+.: (wq_completion)bond570 ->&rq->__lock ->(work_completion)(&(&slave->notify_work)->work) ->(work_completion)(&(&bond->mcast_work)->work) FD: 232 BD: 1 +.+.: (wq_completion)bond562#2 ->&rq->__lock ->(work_completion)(&(&slave->notify_work)->work) ->(work_completion)(&(&bond->mcast_work)->work) FD: 232 BD: 1 +.+.: (wq_completion)bond549#3 ->&rq->__lock ->(work_completion)(&(&slave->notify_work)->work) ->(work_completion)(&(&bond->mcast_work)->work) FD: 232 BD: 1 +.+.: (wq_completion)bond571 ->(work_completion)(&(&slave->notify_work)->work) ->(work_completion)(&(&bond->mcast_work)->work) FD: 232 BD: 1 +.+.: (wq_completion)bond563#2 ->&rq->__lock ->(work_completion)(&(&slave->notify_work)->work) ->(work_completion)(&(&bond->mcast_work)->work) FD: 232 BD: 1 +.+.: (wq_completion)bond550#3 ->(work_completion)(&(&slave->notify_work)->work) ->(work_completion)(&(&bond->mcast_work)->work) FD: 232 BD: 1 +.+.: (wq_completion)bond572 ->(work_completion)(&(&slave->notify_work)->work) ->(work_completion)(&(&bond->mcast_work)->work) FD: 232 BD: 1 +.+.: (wq_completion)bond564#2 ->(work_completion)(&(&slave->notify_work)->work) ->(work_completion)(&(&bond->mcast_work)->work) FD: 232 BD: 1 +.+.: (wq_completion)bond551#3 ->&rq->__lock ->(work_completion)(&(&slave->notify_work)->work) ->(work_completion)(&(&bond->mcast_work)->work) FD: 232 BD: 1 +.+.: (wq_completion)bond573 ->(work_completion)(&(&slave->notify_work)->work) ->(work_completion)(&(&bond->mcast_work)->work) FD: 232 BD: 1 +.+.: (wq_completion)bond565#2 ->&rq->__lock ->(work_completion)(&(&slave->notify_work)->work) ->(work_completion)(&(&bond->mcast_work)->work) FD: 232 BD: 1 +.+.: (wq_completion)bond552#3 ->&rq->__lock ->(work_completion)(&(&slave->notify_work)->work) ->(work_completion)(&(&bond->mcast_work)->work) FD: 232 BD: 1 +.+.: (wq_completion)bond574 ->(work_completion)(&(&slave->notify_work)->work) ->(work_completion)(&(&bond->mcast_work)->work) FD: 232 BD: 1 +.+.: (wq_completion)bond553#3 ->&rq->__lock ->(work_completion)(&(&slave->notify_work)->work) ->(work_completion)(&(&bond->mcast_work)->work) FD: 232 BD: 1 +.+.: (wq_completion)bond566#2 ->&rq->__lock ->(work_completion)(&(&slave->notify_work)->work) ->(work_completion)(&(&bond->mcast_work)->work) FD: 232 BD: 1 +.+.: (wq_completion)bond575 ->&rq->__lock ->(work_completion)(&(&slave->notify_work)->work) ->(work_completion)(&(&bond->mcast_work)->work) FD: 232 BD: 1 +.+.: (wq_completion)bond554#3 ->(work_completion)(&(&slave->notify_work)->work) ->(work_completion)(&(&bond->mcast_work)->work) FD: 232 BD: 1 +.+.: (wq_completion)bond555#3 ->&rq->__lock ->(work_completion)(&(&slave->notify_work)->work) ->(work_completion)(&(&bond->mcast_work)->work) FD: 232 BD: 1 +.+.: (wq_completion)bond576 ->(work_completion)(&(&slave->notify_work)->work) ->(work_completion)(&(&bond->mcast_work)->work) FD: 232 BD: 1 +.+.: (wq_completion)bond567#2 ->&rq->__lock ->(work_completion)(&(&slave->notify_work)->work) ->(work_completion)(&(&bond->mcast_work)->work) FD: 232 BD: 1 +.+.: (wq_completion)bond556#2 ->(work_completion)(&(&slave->notify_work)->work) ->(work_completion)(&(&bond->mcast_work)->work) FD: 232 BD: 1 +.+.: (wq_completion)bond577 ->(work_completion)(&(&slave->notify_work)->work) ->(work_completion)(&(&bond->mcast_work)->work) FD: 232 BD: 1 +.+.: (wq_completion)bond557#3 ->&rq->__lock ->(work_completion)(&(&slave->notify_work)->work) ->(work_completion)(&(&bond->mcast_work)->work) FD: 232 BD: 1 +.+.: (wq_completion)bond568#2 ->(work_completion)(&(&slave->notify_work)->work) ->(work_completion)(&(&bond->mcast_work)->work) FD: 232 BD: 1 +.+.: (wq_completion)bond578 ->&rq->__lock ->(work_completion)(&(&slave->notify_work)->work) ->(work_completion)(&(&bond->mcast_work)->work) FD: 232 BD: 1 +.+.: (wq_completion)bond181#3 ->(work_completion)(&(&slave->notify_work)->work) ->(work_completion)(&(&bond->mcast_work)->work) FD: 232 BD: 1 +.+.: (wq_completion)bond558#3 ->(work_completion)(&(&slave->notify_work)->work) ->(work_completion)(&(&bond->mcast_work)->work) FD: 232 BD: 1 +.+.: (wq_completion)bond579 ->(work_completion)(&(&slave->notify_work)->work) ->(work_completion)(&(&bond->mcast_work)->work) FD: 232 BD: 1 +.+.: (wq_completion)bond559#3 ->(work_completion)(&(&slave->notify_work)->work) ->(work_completion)(&(&bond->mcast_work)->work) FD: 232 BD: 1 +.+.: (wq_completion)bond580 ->&rq->__lock ->(work_completion)(&(&slave->notify_work)->work) ->(work_completion)(&(&bond->mcast_work)->work) FD: 232 BD: 1 +.+.: (wq_completion)bond560#3 ->&rq->__lock ->(work_completion)(&(&slave->notify_work)->work) ->(work_completion)(&(&bond->mcast_work)->work) FD: 232 BD: 1 +.+.: (wq_completion)bond581 ->(work_completion)(&(&slave->notify_work)->work) ->(work_completion)(&(&bond->mcast_work)->work) FD: 232 BD: 1 +.+.: (wq_completion)bond582 ->&rq->__lock ->(work_completion)(&(&slave->notify_work)->work) ->(work_completion)(&(&bond->mcast_work)->work) FD: 232 BD: 1 +.+.: (wq_completion)bond561#3 ->(work_completion)(&(&slave->notify_work)->work) ->(work_completion)(&(&bond->mcast_work)->work) FD: 232 BD: 1 +.+.: (wq_completion)bond583 ->(work_completion)(&(&slave->notify_work)->work) ->(work_completion)(&(&bond->mcast_work)->work) FD: 232 BD: 1 +.+.: (wq_completion)bond562#3 ->&rq->__lock ->(work_completion)(&(&slave->notify_work)->work) ->(work_completion)(&(&bond->mcast_work)->work) FD: 232 BD: 1 +.+.: (wq_completion)bond182#4 ->(work_completion)(&(&slave->notify_work)->work) ->(work_completion)(&(&bond->mcast_work)->work) FD: 232 BD: 1 +.+.: (wq_completion)bond563#3 ->&rq->__lock ->(work_completion)(&(&slave->notify_work)->work) ->(work_completion)(&(&bond->mcast_work)->work) FD: 232 BD: 1 +.+.: (wq_completion)bond564#3 ->(work_completion)(&(&slave->notify_work)->work) ->(work_completion)(&(&bond->mcast_work)->work) FD: 232 BD: 1 +.+.: (wq_completion)bond565#3 ->(work_completion)(&(&slave->notify_work)->work) ->(work_completion)(&(&bond->mcast_work)->work) FD: 232 BD: 1 +.+.: (wq_completion)bond566#3 ->(work_completion)(&(&slave->notify_work)->work) ->(work_completion)(&(&bond->mcast_work)->work) FD: 232 BD: 1 +.+.: (wq_completion)bond567#3 ->&rq->__lock ->(work_completion)(&(&slave->notify_work)->work) ->(work_completion)(&(&bond->mcast_work)->work) FD: 232 BD: 1 +.+.: (wq_completion)bond568#3 ->&rq->__lock ->(work_completion)(&(&slave->notify_work)->work) ->(work_completion)(&(&bond->mcast_work)->work) FD: 232 BD: 1 +.+.: (wq_completion)bond569#2 ->&rq->__lock ->(work_completion)(&(&slave->notify_work)->work) ->(work_completion)(&(&bond->mcast_work)->work) FD: 232 BD: 1 +.+.: (wq_completion)bond183#4 ->&rq->__lock ->(work_completion)(&(&slave->notify_work)->work) ->(work_completion)(&(&bond->mcast_work)->work) FD: 232 BD: 1 +.+.: (wq_completion)bond570#2 ->&rq->__lock ->(work_completion)(&(&slave->notify_work)->work) ->(work_completion)(&(&bond->mcast_work)->work) FD: 232 BD: 1 +.+.: (wq_completion)bond184#4 ->(work_completion)(&(&slave->notify_work)->work) ->(work_completion)(&(&bond->mcast_work)->work) FD: 232 BD: 1 +.+.: (wq_completion)bond614 ->&rq->__lock ->(work_completion)(&(&slave->notify_work)->work) ->(work_completion)(&(&bond->mcast_work)->work) FD: 232 BD: 1 +.+.: (wq_completion)bond618 ->&rq->__lock ->(work_completion)(&(&slave->notify_work)->work) ->(work_completion)(&(&bond->mcast_work)->work) FD: 128 BD: 79 +.+.: &data->nh_lock ->fs_reclaim ->pool_lock#2 FD: 232 BD: 1 +.+.: (wq_completion)bond617 ->(work_completion)(&(&slave->notify_work)->work) ->(work_completion)(&(&bond->mcast_work)->work) FD: 232 BD: 1 +.+.: (wq_completion)bond626 ->(work_completion)(&(&slave->notify_work)->work) ->(work_completion)(&(&bond->mcast_work)->work) FD: 232 BD: 1 +.+.: (wq_completion)bond619 ->(work_completion)(&(&slave->notify_work)->work) ->(work_completion)(&(&bond->mcast_work)->work) FD: 232 BD: 1 +.+.: (wq_completion)bond620 ->(work_completion)(&(&slave->notify_work)->work) ->(work_completion)(&(&bond->mcast_work)->work) FD: 232 BD: 1 +.+.: (wq_completion)bond621 ->(work_completion)(&(&slave->notify_work)->work) ->(work_completion)(&(&bond->mcast_work)->work) FD: 232 BD: 1 +.+.: (wq_completion)bond629 ->&rq->__lock ->(work_completion)(&(&slave->notify_work)->work) ->(work_completion)(&(&bond->mcast_work)->work) FD: 232 BD: 1 +.+.: (wq_completion)bond623 ->(work_completion)(&(&slave->notify_work)->work) ->(work_completion)(&(&bond->mcast_work)->work) FD: 232 BD: 1 +.+.: (wq_completion)bond622 ->&rq->__lock ->(work_completion)(&(&slave->notify_work)->work) ->(work_completion)(&(&bond->mcast_work)->work) FD: 232 BD: 1 +.+.: (wq_completion)bond630 ->&rq->__lock ->(work_completion)(&(&slave->notify_work)->work) ->(work_completion)(&(&bond->mcast_work)->work) FD: 232 BD: 1 +.+.: (wq_completion)bond631 ->(work_completion)(&(&slave->notify_work)->work) ->(work_completion)(&(&bond->mcast_work)->work) FD: 232 BD: 1 +.+.: (wq_completion)bond633 ->(work_completion)(&(&slave->notify_work)->work) ->(work_completion)(&(&bond->mcast_work)->work) FD: 232 BD: 1 +.+.: (wq_completion)bond632 ->(work_completion)(&(&slave->notify_work)->work) ->(work_completion)(&(&bond->mcast_work)->work) FD: 232 BD: 1 +.+.: (wq_completion)bond624 ->(work_completion)(&(&slave->notify_work)->work) ->(work_completion)(&(&bond->mcast_work)->work) FD: 232 BD: 1 +.+.: (wq_completion)bond625 ->(work_completion)(&(&slave->notify_work)->work) ->(work_completion)(&(&bond->mcast_work)->work) FD: 232 BD: 1 +.+.: (wq_completion)bond627 ->&rq->__lock ->(work_completion)(&(&slave->notify_work)->work) ->(work_completion)(&(&bond->mcast_work)->work) FD: 232 BD: 1 +.+.: (wq_completion)bond634 ->(work_completion)(&(&slave->notify_work)->work) ->(work_completion)(&(&bond->mcast_work)->work) FD: 232 BD: 1 +.+.: (wq_completion)bond626#2 ->(work_completion)(&(&slave->notify_work)->work) ->(work_completion)(&(&bond->mcast_work)->work) FD: 232 BD: 1 +.+.: (wq_completion)bond635 ->(work_completion)(&(&slave->notify_work)->work) ->(work_completion)(&(&bond->mcast_work)->work) FD: 232 BD: 1 +.+.: (wq_completion)bond615 ->(work_completion)(&(&slave->notify_work)->work) ->(work_completion)(&(&bond->mcast_work)->work) FD: 232 BD: 1 +.+.: (wq_completion)bond628 ->(work_completion)(&(&slave->notify_work)->work) ->(work_completion)(&(&bond->mcast_work)->work) FD: 232 BD: 1 +.+.: (wq_completion)bond636 ->(work_completion)(&(&slave->notify_work)->work) ->(work_completion)(&(&bond->mcast_work)->work) FD: 232 BD: 1 +.+.: (wq_completion)bond629#2 ->&rq->__lock ->(work_completion)(&(&slave->notify_work)->work) ->(work_completion)(&(&bond->mcast_work)->work) FD: 232 BD: 1 +.+.: (wq_completion)bond637 ->&rq->__lock ->(work_completion)(&(&slave->notify_work)->work) ->(work_completion)(&(&bond->mcast_work)->work) FD: 1 BD: 80 +...: &priv->lock FD: 1 BD: 79 +...: &priv->j1939_socks_lock FD: 1 BD: 79 +...: &jsk->sk_session_queue_lock FD: 232 BD: 1 +.+.: (wq_completion)bond630#2 ->&rq->__lock ->(work_completion)(&(&slave->notify_work)->work) ->(work_completion)(&(&bond->mcast_work)->work) FD: 232 BD: 1 +.+.: (wq_completion)bond638 ->(work_completion)(&(&slave->notify_work)->work) ->(work_completion)(&(&bond->mcast_work)->work) FD: 232 BD: 1 +.+.: (wq_completion)bond618#2 ->(work_completion)(&(&slave->notify_work)->work) ->(work_completion)(&(&bond->mcast_work)->work) FD: 232 BD: 1 +.+.: (wq_completion)bond631#2 ->&rq->__lock ->(work_completion)(&(&slave->notify_work)->work) ->(work_completion)(&(&bond->mcast_work)->work) FD: 232 BD: 1 +.+.: (wq_completion)bond639 ->(work_completion)(&(&slave->notify_work)->work) ->(work_completion)(&(&bond->mcast_work)->work) FD: 232 BD: 1 +.+.: (wq_completion)bond619#2 ->(work_completion)(&(&slave->notify_work)->work) ->(work_completion)(&(&bond->mcast_work)->work) FD: 232 BD: 1 +.+.: (wq_completion)bond640 ->(work_completion)(&(&slave->notify_work)->work) ->(work_completion)(&(&bond->mcast_work)->work) FD: 232 BD: 1 +.+.: (wq_completion)bond632#2 ->(work_completion)(&(&slave->notify_work)->work) ->(work_completion)(&(&bond->mcast_work)->work) FD: 232 BD: 1 +.+.: (wq_completion)bond620#2 ->(work_completion)(&(&slave->notify_work)->work) ->(work_completion)(&(&bond->mcast_work)->work) FD: 232 BD: 1 +.+.: (wq_completion)bond641 ->(work_completion)(&(&slave->notify_work)->work) ->(work_completion)(&(&bond->mcast_work)->work) FD: 232 BD: 1 +.+.: (wq_completion)bond633#2 ->(work_completion)(&(&slave->notify_work)->work) ->(work_completion)(&(&bond->mcast_work)->work) FD: 232 BD: 1 +.+.: (wq_completion)bond621#2 ->&rq->__lock ->(work_completion)(&(&slave->notify_work)->work) ->(work_completion)(&(&bond->mcast_work)->work) FD: 232 BD: 1 +.+.: (wq_completion)bond642 ->(work_completion)(&(&slave->notify_work)->work) ->(work_completion)(&(&bond->mcast_work)->work) FD: 232 BD: 1 +.+.: (wq_completion)bond634#2 ->(work_completion)(&(&slave->notify_work)->work) ->(work_completion)(&(&bond->mcast_work)->work) FD: 232 BD: 1 +.+.: (wq_completion)bond622#2 ->&rq->__lock ->(work_completion)(&(&slave->notify_work)->work) ->(work_completion)(&(&bond->mcast_work)->work) FD: 232 BD: 1 +.+.: (wq_completion)bond643 ->&rq->__lock ->(work_completion)(&(&slave->notify_work)->work) ->(work_completion)(&(&bond->mcast_work)->work) FD: 232 BD: 1 +.+.: (wq_completion)bond635#2 ->(work_completion)(&(&slave->notify_work)->work) ->(work_completion)(&(&bond->mcast_work)->work) ->&rq->__lock FD: 232 BD: 1 +.+.: (wq_completion)bond623#2 ->(work_completion)(&(&slave->notify_work)->work) ->(work_completion)(&(&bond->mcast_work)->work) FD: 232 BD: 1 +.+.: (wq_completion)bond636#2 ->(work_completion)(&(&slave->notify_work)->work) ->(work_completion)(&(&bond->mcast_work)->work) ->&rq->__lock FD: 232 BD: 1 +.+.: (wq_completion)bond644 ->&rq->__lock ->(work_completion)(&(&slave->notify_work)->work) ->(work_completion)(&(&bond->mcast_work)->work) FD: 232 BD: 1 +.+.: (wq_completion)bond624#2 ->(work_completion)(&(&slave->notify_work)->work) ->(work_completion)(&(&bond->mcast_work)->work) FD: 232 BD: 1 +.+.: (wq_completion)bond637#2 ->(work_completion)(&(&slave->notify_work)->work) ->(work_completion)(&(&bond->mcast_work)->work) FD: 232 BD: 1 +.+.: (wq_completion)bond645 ->(work_completion)(&(&slave->notify_work)->work) ->(work_completion)(&(&bond->mcast_work)->work) ->&rq->__lock FD: 232 BD: 1 +.+.: (wq_completion)bond625#2 ->&rq->__lock ->(work_completion)(&(&slave->notify_work)->work) ->(work_completion)(&(&bond->mcast_work)->work) FD: 232 BD: 1 +.+.: (wq_completion)bond638#2 ->(work_completion)(&(&slave->notify_work)->work) ->(work_completion)(&(&bond->mcast_work)->work) FD: 232 BD: 1 +.+.: (wq_completion)bond646 ->(work_completion)(&(&slave->notify_work)->work) ->(work_completion)(&(&bond->mcast_work)->work) ->&rq->__lock FD: 232 BD: 1 +.+.: (wq_completion)bond626#3 ->&rq->__lock ->(work_completion)(&(&slave->notify_work)->work) ->(work_completion)(&(&bond->mcast_work)->work) FD: 232 BD: 1 +.+.: (wq_completion)bond639#2 ->&rq->__lock ->(work_completion)(&(&slave->notify_work)->work) ->(work_completion)(&(&bond->mcast_work)->work) FD: 232 BD: 1 +.+.: (wq_completion)bond647 ->(work_completion)(&(&slave->notify_work)->work) ->(work_completion)(&(&bond->mcast_work)->work) FD: 232 BD: 1 +.+.: (wq_completion)bond627#2 ->(work_completion)(&(&slave->notify_work)->work) ->(work_completion)(&(&bond->mcast_work)->work) FD: 232 BD: 1 +.+.: (wq_completion)bond640#2 ->(work_completion)(&(&slave->notify_work)->work) ->(work_completion)(&(&bond->mcast_work)->work) FD: 232 BD: 1 +.+.: (wq_completion)bond648 ->&rq->__lock ->(work_completion)(&(&slave->notify_work)->work) ->(work_completion)(&(&bond->mcast_work)->work) FD: 232 BD: 1 +.+.: (wq_completion)bond628#2 ->(work_completion)(&(&slave->notify_work)->work) ->(work_completion)(&(&bond->mcast_work)->work) ->&rq->__lock FD: 232 BD: 1 +.+.: (wq_completion)bond641#2 ->&rq->__lock ->(work_completion)(&(&slave->notify_work)->work) ->(work_completion)(&(&bond->mcast_work)->work) FD: 232 BD: 1 +.+.: (wq_completion)bond649 ->(work_completion)(&(&slave->notify_work)->work) ->(work_completion)(&(&bond->mcast_work)->work) ->&rq->__lock FD: 232 BD: 1 +.+.: (wq_completion)bond629#3 ->(work_completion)(&(&slave->notify_work)->work) ->(work_completion)(&(&bond->mcast_work)->work) FD: 232 BD: 1 +.+.: (wq_completion)bond642#2 ->&rq->__lock ->(work_completion)(&(&slave->notify_work)->work) ->(work_completion)(&(&bond->mcast_work)->work) FD: 232 BD: 1 +.+.: (wq_completion)bond650 ->(work_completion)(&(&slave->notify_work)->work) ->(work_completion)(&(&bond->mcast_work)->work) ->&rq->__lock FD: 232 BD: 1 +.+.: (wq_completion)bond630#3 ->&rq->__lock ->(work_completion)(&(&slave->notify_work)->work) ->(work_completion)(&(&bond->mcast_work)->work) FD: 232 BD: 1 +.+.: (wq_completion)bond643#2 ->(work_completion)(&(&slave->notify_work)->work) ->(work_completion)(&(&bond->mcast_work)->work) FD: 232 BD: 1 +.+.: (wq_completion)bond651 ->(work_completion)(&(&slave->notify_work)->work) ->(work_completion)(&(&bond->mcast_work)->work) FD: 232 BD: 1 +.+.: (wq_completion)bond631#3 ->(work_completion)(&(&slave->notify_work)->work) ->(work_completion)(&(&bond->mcast_work)->work) ->&rq->__lock FD: 232 BD: 1 +.+.: (wq_completion)bond652 ->(work_completion)(&(&slave->notify_work)->work) ->(work_completion)(&(&bond->mcast_work)->work) FD: 232 BD: 1 +.+.: (wq_completion)bond632#3 ->&rq->__lock ->(work_completion)(&(&slave->notify_work)->work) ->(work_completion)(&(&bond->mcast_work)->work) FD: 232 BD: 1 +.+.: (wq_completion)bond645#2 ->&rq->__lock ->(work_completion)(&(&slave->notify_work)->work) ->(work_completion)(&(&bond->mcast_work)->work) FD: 232 BD: 1 +.+.: (wq_completion)bond653 ->&rq->__lock ->(work_completion)(&(&slave->notify_work)->work) ->(work_completion)(&(&bond->mcast_work)->work) FD: 232 BD: 1 +.+.: (wq_completion)bond633#3 ->&rq->__lock ->(work_completion)(&(&slave->notify_work)->work) ->(work_completion)(&(&bond->mcast_work)->work) FD: 232 BD: 1 +.+.: (wq_completion)bond646#2 ->(work_completion)(&(&slave->notify_work)->work) ->(work_completion)(&(&bond->mcast_work)->work) FD: 232 BD: 1 +.+.: (wq_completion)bond654 ->(work_completion)(&(&slave->notify_work)->work) ->(work_completion)(&(&bond->mcast_work)->work) FD: 1 BD: 1 ....: _rs.lock#26 FD: 232 BD: 1 +.+.: (wq_completion)bond647#2 ->(work_completion)(&(&slave->notify_work)->work) ->(work_completion)(&(&bond->mcast_work)->work) ->&rq->__lock FD: 232 BD: 1 +.+.: (wq_completion)bond634#3 ->(work_completion)(&(&slave->notify_work)->work) ->(work_completion)(&(&bond->mcast_work)->work) FD: 232 BD: 1 +.+.: (wq_completion)bond655 ->&rq->__lock ->(work_completion)(&(&slave->notify_work)->work) ->(work_completion)(&(&bond->mcast_work)->work) FD: 232 BD: 1 +.+.: (wq_completion)bond635#3 ->(work_completion)(&(&slave->notify_work)->work) ->(work_completion)(&(&bond->mcast_work)->work) ->&rq->__lock FD: 232 BD: 1 +.+.: (wq_completion)bond648#2 ->&rq->__lock ->(work_completion)(&(&slave->notify_work)->work) ->(work_completion)(&(&bond->mcast_work)->work) FD: 232 BD: 1 +.+.: (wq_completion)bond656 ->(work_completion)(&(&slave->notify_work)->work) ->(work_completion)(&(&bond->mcast_work)->work) FD: 232 BD: 1 +.+.: (wq_completion)bond636#3 ->(work_completion)(&(&slave->notify_work)->work) ->(work_completion)(&(&bond->mcast_work)->work) FD: 232 BD: 1 +.+.: (wq_completion)bond649#2 ->(work_completion)(&(&slave->notify_work)->work) ->(work_completion)(&(&bond->mcast_work)->work) FD: 232 BD: 1 +.+.: (wq_completion)bond657 ->&rq->__lock ->(work_completion)(&(&slave->notify_work)->work) ->(work_completion)(&(&bond->mcast_work)->work) FD: 232 BD: 1 +.+.: (wq_completion)bond637#3 ->&rq->__lock ->(work_completion)(&(&slave->notify_work)->work) ->(work_completion)(&(&bond->mcast_work)->work) FD: 232 BD: 1 +.+.: (wq_completion)bond650#2 ->(work_completion)(&(&slave->notify_work)->work) ->(work_completion)(&(&bond->mcast_work)->work) FD: 232 BD: 1 +.+.: (wq_completion)bond658 ->(work_completion)(&(&slave->notify_work)->work) ->(work_completion)(&(&bond->mcast_work)->work) FD: 232 BD: 1 +.+.: (wq_completion)bond638#3 ->(work_completion)(&(&slave->notify_work)->work) ->(work_completion)(&(&bond->mcast_work)->work) FD: 232 BD: 1 +.+.: (wq_completion)bond651#2 ->(work_completion)(&(&slave->notify_work)->work) ->&rq->__lock ->(work_completion)(&(&bond->mcast_work)->work) FD: 232 BD: 1 +.+.: (wq_completion)bond659 ->(work_completion)(&(&slave->notify_work)->work) ->(work_completion)(&(&bond->mcast_work)->work) ->&rq->__lock ->&cfs_rq->removed.lock ->&obj_hash[i].lock FD: 232 BD: 1 +.+.: (wq_completion)bond639#3 ->&rq->__lock ->(work_completion)(&(&slave->notify_work)->work) ->(work_completion)(&(&bond->mcast_work)->work) FD: 232 BD: 1 +.+.: (wq_completion)bond652#2 ->(work_completion)(&(&slave->notify_work)->work) ->(work_completion)(&(&bond->mcast_work)->work) ->&rq->__lock FD: 232 BD: 1 +.+.: (wq_completion)bond660 ->(work_completion)(&(&slave->notify_work)->work) ->(work_completion)(&(&bond->mcast_work)->work) FD: 232 BD: 1 +.+.: (wq_completion)bond640#3 ->(work_completion)(&(&slave->notify_work)->work) ->(work_completion)(&(&bond->mcast_work)->work) FD: 232 BD: 1 +.+.: (wq_completion)bond653#2 ->(work_completion)(&(&slave->notify_work)->work) ->(work_completion)(&(&bond->mcast_work)->work) ->&rq->__lock FD: 232 BD: 1 +.+.: (wq_completion)bond661 ->&rq->__lock ->(work_completion)(&(&slave->notify_work)->work) ->(work_completion)(&(&bond->mcast_work)->work) FD: 232 BD: 1 +.+.: (wq_completion)bond641#3 ->(work_completion)(&(&slave->notify_work)->work) ->(work_completion)(&(&bond->mcast_work)->work) ->&rq->__lock FD: 232 BD: 1 +.+.: (wq_completion)bond654#2 ->&rq->__lock ->(work_completion)(&(&slave->notify_work)->work) ->(work_completion)(&(&bond->mcast_work)->work) FD: 232 BD: 1 +.+.: (wq_completion)bond662 ->(work_completion)(&(&slave->notify_work)->work) ->(work_completion)(&(&bond->mcast_work)->work) ->&rq->__lock FD: 232 BD: 1 +.+.: (wq_completion)bond642#3 ->&rq->__lock ->(work_completion)(&(&slave->notify_work)->work) ->(work_completion)(&(&bond->mcast_work)->work) FD: 232 BD: 1 +.+.: (wq_completion)bond655#2 ->(work_completion)(&(&slave->notify_work)->work) ->(work_completion)(&(&bond->mcast_work)->work) FD: 232 BD: 1 +.+.: (wq_completion)bond663 ->(work_completion)(&(&slave->notify_work)->work) ->(work_completion)(&(&bond->mcast_work)->work) FD: 232 BD: 1 +.+.: (wq_completion)bond643#3 ->(work_completion)(&(&slave->notify_work)->work) ->(work_completion)(&(&bond->mcast_work)->work) ->&rq->__lock FD: 232 BD: 1 +.+.: (wq_completion)bond656#2 ->(work_completion)(&(&slave->notify_work)->work) ->(work_completion)(&(&bond->mcast_work)->work) ->&rq->__lock FD: 232 BD: 1 +.+.: (wq_completion)bond664 ->(work_completion)(&(&slave->notify_work)->work) ->(work_completion)(&(&bond->mcast_work)->work) FD: 232 BD: 1 +.+.: (wq_completion)bond644#2 ->&rq->__lock ->(work_completion)(&(&slave->notify_work)->work) ->(work_completion)(&(&bond->mcast_work)->work) FD: 232 BD: 1 +.+.: (wq_completion)bond657#2 ->&rq->__lock ->(work_completion)(&(&slave->notify_work)->work) ->(work_completion)(&(&bond->mcast_work)->work) FD: 232 BD: 1 +.+.: (wq_completion)bond665 ->&rq->__lock ->(work_completion)(&(&slave->notify_work)->work) ->(work_completion)(&(&bond->mcast_work)->work) FD: 232 BD: 1 +.+.: (wq_completion)bond645#3 ->&rq->__lock ->(work_completion)(&(&slave->notify_work)->work) ->(work_completion)(&(&bond->mcast_work)->work) FD: 232 BD: 1 +.+.: (wq_completion)bond658#2 ->&rq->__lock ->(work_completion)(&(&slave->notify_work)->work) ->(work_completion)(&(&bond->mcast_work)->work) FD: 232 BD: 1 +.+.: (wq_completion)bond666 ->(work_completion)(&(&slave->notify_work)->work) ->(work_completion)(&(&bond->mcast_work)->work) FD: 232 BD: 1 +.+.: (wq_completion)bond646#3 ->&rq->__lock ->(work_completion)(&(&slave->notify_work)->work) ->(work_completion)(&(&bond->mcast_work)->work) FD: 232 BD: 1 +.+.: (wq_completion)bond667 ->&rq->__lock ->(work_completion)(&(&slave->notify_work)->work) ->(work_completion)(&(&bond->mcast_work)->work) FD: 232 BD: 1 +.+.: (wq_completion)bond659#2 ->&rq->__lock ->(work_completion)(&(&slave->notify_work)->work) ->(work_completion)(&(&bond->mcast_work)->work) FD: 232 BD: 1 +.+.: (wq_completion)bond647#3 ->(work_completion)(&(&slave->notify_work)->work) ->(work_completion)(&(&bond->mcast_work)->work) FD: 232 BD: 1 +.+.: (wq_completion)bond660#2 ->(work_completion)(&(&slave->notify_work)->work) ->(work_completion)(&(&bond->mcast_work)->work) FD: 232 BD: 1 +.+.: (wq_completion)bond668 ->(work_completion)(&(&slave->notify_work)->work) ->(work_completion)(&(&bond->mcast_work)->work) FD: 232 BD: 1 +.+.: (wq_completion)bond648#3 ->&rq->__lock ->(work_completion)(&(&slave->notify_work)->work) ->(work_completion)(&(&bond->mcast_work)->work) FD: 232 BD: 1 +.+.: (wq_completion)bond661#2 ->&rq->__lock ->(work_completion)(&(&slave->notify_work)->work) ->(work_completion)(&(&bond->mcast_work)->work) FD: 232 BD: 1 +.+.: (wq_completion)bond669 ->(work_completion)(&(&slave->notify_work)->work) ->(work_completion)(&(&bond->mcast_work)->work) FD: 232 BD: 1 +.+.: (wq_completion)bond649#3 ->&rq->__lock ->(work_completion)(&(&slave->notify_work)->work) ->(work_completion)(&(&bond->mcast_work)->work) FD: 232 BD: 1 +.+.: (wq_completion)bond662#2 ->(work_completion)(&(&slave->notify_work)->work) ->(work_completion)(&(&bond->mcast_work)->work) ->&rq->__lock FD: 232 BD: 1 +.+.: (wq_completion)bond670 ->(work_completion)(&(&slave->notify_work)->work) ->(work_completion)(&(&bond->mcast_work)->work) FD: 232 BD: 1 +.+.: (wq_completion)bond650#3 ->(work_completion)(&(&slave->notify_work)->work) ->(work_completion)(&(&bond->mcast_work)->work) FD: 232 BD: 1 +.+.: (wq_completion)bond671 ->&rq->__lock ->(work_completion)(&(&slave->notify_work)->work) ->(work_completion)(&(&bond->mcast_work)->work) FD: 232 BD: 1 +.+.: (wq_completion)bond651#3 ->(work_completion)(&(&slave->notify_work)->work) ->(work_completion)(&(&bond->mcast_work)->work) ->&rq->__lock FD: 232 BD: 1 +.+.: (wq_completion)bond664#2 ->&rq->__lock ->(work_completion)(&(&slave->notify_work)->work) ->(work_completion)(&(&bond->mcast_work)->work) FD: 232 BD: 1 +.+.: (wq_completion)bond672 ->(work_completion)(&(&slave->notify_work)->work) ->(work_completion)(&(&bond->mcast_work)->work) FD: 232 BD: 1 +.+.: (wq_completion)bond652#3 ->(work_completion)(&(&slave->notify_work)->work) ->(work_completion)(&(&bond->mcast_work)->work) ->&rq->__lock FD: 232 BD: 1 +.+.: (wq_completion)bond665#2 ->&rq->__lock ->(work_completion)(&(&slave->notify_work)->work) ->(work_completion)(&(&bond->mcast_work)->work) FD: 232 BD: 1 +.+.: (wq_completion)bond673 ->&rq->__lock ->(work_completion)(&(&slave->notify_work)->work) ->(work_completion)(&(&bond->mcast_work)->work) FD: 232 BD: 1 +.+.: (wq_completion)bond653#3 ->(work_completion)(&(&slave->notify_work)->work) ->(work_completion)(&(&bond->mcast_work)->work) FD: 232 BD: 1 +.+.: (wq_completion)bond666#2 ->&rq->__lock ->(work_completion)(&(&slave->notify_work)->work) ->(work_completion)(&(&bond->mcast_work)->work) FD: 232 BD: 1 +.+.: (wq_completion)bond674 ->&rq->__lock ->(work_completion)(&(&slave->notify_work)->work) ->(work_completion)(&(&bond->mcast_work)->work) FD: 232 BD: 1 +.+.: (wq_completion)bond654#3 ->(work_completion)(&(&slave->notify_work)->work) ->(work_completion)(&(&bond->mcast_work)->work) FD: 232 BD: 1 +.+.: (wq_completion)bond667#2 ->&rq->__lock ->(work_completion)(&(&slave->notify_work)->work) ->(work_completion)(&(&bond->mcast_work)->work) FD: 232 BD: 1 +.+.: (wq_completion)bond675 ->(work_completion)(&(&slave->notify_work)->work) ->(work_completion)(&(&bond->mcast_work)->work) FD: 232 BD: 1 +.+.: (wq_completion)bond655#3 ->&rq->__lock ->(work_completion)(&(&slave->notify_work)->work) ->(work_completion)(&(&bond->mcast_work)->work) FD: 232 BD: 1 +.+.: (wq_completion)bond668#2 ->&rq->__lock ->(work_completion)(&(&slave->notify_work)->work) ->(work_completion)(&(&bond->mcast_work)->work) FD: 232 BD: 1 +.+.: (wq_completion)bond676 ->(work_completion)(&(&slave->notify_work)->work) ->(work_completion)(&(&bond->mcast_work)->work) ->&rq->__lock FD: 232 BD: 1 +.+.: (wq_completion)bond656#3 ->(work_completion)(&(&slave->notify_work)->work) ->(work_completion)(&(&bond->mcast_work)->work) FD: 232 BD: 1 +.+.: (wq_completion)bond669#2 ->(work_completion)(&(&slave->notify_work)->work) ->(work_completion)(&(&bond->mcast_work)->work) ->&rq->__lock FD: 232 BD: 1 +.+.: (wq_completion)bond677 ->(work_completion)(&(&slave->notify_work)->work) ->(work_completion)(&(&bond->mcast_work)->work) ->&rq->__lock FD: 232 BD: 1 +.+.: (wq_completion)bond657#3 ->(work_completion)(&(&slave->notify_work)->work) ->(work_completion)(&(&bond->mcast_work)->work) FD: 232 BD: 1 +.+.: (wq_completion)bond670#2 ->(work_completion)(&(&slave->notify_work)->work) ->(work_completion)(&(&bond->mcast_work)->work) ->&rq->__lock FD: 232 BD: 1 +.+.: (wq_completion)bond678 ->(work_completion)(&(&slave->notify_work)->work) ->(work_completion)(&(&bond->mcast_work)->work) FD: 232 BD: 1 +.+.: (wq_completion)bond658#3 ->&rq->__lock ->(work_completion)(&(&slave->notify_work)->work) ->(work_completion)(&(&bond->mcast_work)->work) FD: 232 BD: 1 +.+.: (wq_completion)bond671#2 ->(work_completion)(&(&slave->notify_work)->work) ->(work_completion)(&(&bond->mcast_work)->work) FD: 232 BD: 1 +.+.: (wq_completion)bond679 ->&rq->__lock ->(work_completion)(&(&slave->notify_work)->work) ->(work_completion)(&(&bond->mcast_work)->work) FD: 232 BD: 1 +.+.: (wq_completion)bond659#3 ->&rq->__lock ->(work_completion)(&(&slave->notify_work)->work) ->(work_completion)(&(&bond->mcast_work)->work) FD: 232 BD: 1 +.+.: (wq_completion)bond672#2 ->&rq->__lock ->(work_completion)(&(&slave->notify_work)->work) ->(work_completion)(&(&bond->mcast_work)->work) FD: 232 BD: 1 +.+.: (wq_completion)bond680 ->(work_completion)(&(&slave->notify_work)->work) ->(work_completion)(&(&bond->mcast_work)->work) ->&rq->__lock FD: 232 BD: 1 +.+.: (wq_completion)bond660#3 ->(work_completion)(&(&slave->notify_work)->work) ->(work_completion)(&(&bond->mcast_work)->work) FD: 232 BD: 1 +.+.: (wq_completion)bond673#2 ->(work_completion)(&(&slave->notify_work)->work) ->(work_completion)(&(&bond->mcast_work)->work) FD: 232 BD: 1 +.+.: (wq_completion)bond681 ->&rq->__lock ->(work_completion)(&(&slave->notify_work)->work) ->(work_completion)(&(&bond->mcast_work)->work) FD: 232 BD: 1 +.+.: (wq_completion)bond661#3 ->&rq->__lock ->(work_completion)(&(&slave->notify_work)->work) ->(work_completion)(&(&bond->mcast_work)->work) FD: 232 BD: 1 +.+.: (wq_completion)bond674#2 ->&rq->__lock ->(work_completion)(&(&slave->notify_work)->work) ->(work_completion)(&(&bond->mcast_work)->work) FD: 232 BD: 1 +.+.: (wq_completion)bond682 ->(work_completion)(&(&slave->notify_work)->work) ->(work_completion)(&(&bond->mcast_work)->work) FD: 232 BD: 1 +.+.: (wq_completion)bond662#3 ->&rq->__lock ->(work_completion)(&(&slave->notify_work)->work) ->(work_completion)(&(&bond->mcast_work)->work) FD: 232 BD: 1 +.+.: (wq_completion)bond675#2 ->(work_completion)(&(&slave->notify_work)->work) ->(work_completion)(&(&bond->mcast_work)->work) ->&rq->__lock FD: 232 BD: 1 +.+.: (wq_completion)bond683 ->&rq->__lock ->(work_completion)(&(&slave->notify_work)->work) ->(work_completion)(&(&bond->mcast_work)->work) FD: 232 BD: 1 +.+.: (wq_completion)bond663#2 ->(work_completion)(&(&slave->notify_work)->work) ->(work_completion)(&(&bond->mcast_work)->work) ->&rq->__lock FD: 232 BD: 1 +.+.: (wq_completion)bond676#2 ->(work_completion)(&(&slave->notify_work)->work) ->(work_completion)(&(&bond->mcast_work)->work) ->&rq->__lock FD: 232 BD: 1 +.+.: (wq_completion)bond684 ->(work_completion)(&(&slave->notify_work)->work) ->(work_completion)(&(&bond->mcast_work)->work) ->&rq->__lock FD: 232 BD: 1 +.+.: (wq_completion)bond664#3 ->&rq->__lock ->(work_completion)(&(&slave->notify_work)->work) ->(work_completion)(&(&bond->mcast_work)->work) FD: 232 BD: 1 +.+.: (wq_completion)bond677#2 ->&rq->__lock ->(work_completion)(&(&slave->notify_work)->work) ->(work_completion)(&(&bond->mcast_work)->work) FD: 232 BD: 1 +.+.: (wq_completion)bond685 ->&rq->__lock ->(work_completion)(&(&slave->notify_work)->work) ->(work_completion)(&(&bond->mcast_work)->work) FD: 232 BD: 1 +.+.: (wq_completion)bond665#3 ->(work_completion)(&(&slave->notify_work)->work) ->(work_completion)(&(&bond->mcast_work)->work) ->&rq->__lock FD: 232 BD: 1 +.+.: (wq_completion)bond678#2 ->&rq->__lock ->(work_completion)(&(&slave->notify_work)->work) ->(work_completion)(&(&bond->mcast_work)->work) FD: 232 BD: 1 +.+.: (wq_completion)bond686 ->&rq->__lock ->(work_completion)(&(&slave->notify_work)->work) ->(work_completion)(&(&bond->mcast_work)->work) FD: 232 BD: 1 +.+.: (wq_completion)bond666#3 ->&rq->__lock ->(work_completion)(&(&slave->notify_work)->work) ->(work_completion)(&(&bond->mcast_work)->work) FD: 232 BD: 1 +.+.: (wq_completion)bond679#2 ->(work_completion)(&(&slave->notify_work)->work) ->(work_completion)(&(&bond->mcast_work)->work) ->&rq->__lock FD: 232 BD: 1 +.+.: (wq_completion)bond687 ->(work_completion)(&(&slave->notify_work)->work) ->(work_completion)(&(&bond->mcast_work)->work) FD: 232 BD: 1 +.+.: (wq_completion)bond667#3 ->&rq->__lock ->(work_completion)(&(&slave->notify_work)->work) ->(work_completion)(&(&bond->mcast_work)->work) FD: 232 BD: 1 +.+.: (wq_completion)bond680#2 ->&rq->__lock ->(work_completion)(&(&slave->notify_work)->work) ->(work_completion)(&(&bond->mcast_work)->work) FD: 232 BD: 1 +.+.: (wq_completion)bond688 ->(work_completion)(&(&slave->notify_work)->work) ->(work_completion)(&(&bond->mcast_work)->work) ->&rq->__lock FD: 232 BD: 1 +.+.: (wq_completion)bond668#3 ->(work_completion)(&(&slave->notify_work)->work) ->(work_completion)(&(&bond->mcast_work)->work) FD: 232 BD: 1 +.+.: (wq_completion)bond681#2 ->&rq->__lock ->(work_completion)(&(&slave->notify_work)->work) ->(work_completion)(&(&bond->mcast_work)->work) FD: 232 BD: 1 +.+.: (wq_completion)bond689 ->&rq->__lock ->(work_completion)(&(&slave->notify_work)->work) ->(work_completion)(&(&bond->mcast_work)->work) FD: 232 BD: 1 +.+.: (wq_completion)bond669#3 ->&rq->__lock ->(work_completion)(&(&slave->notify_work)->work) ->(work_completion)(&(&bond->mcast_work)->work) FD: 232 BD: 1 +.+.: (wq_completion)bond682#2 ->&rq->__lock ->(work_completion)(&(&slave->notify_work)->work) ->(work_completion)(&(&bond->mcast_work)->work) FD: 232 BD: 1 +.+.: (wq_completion)bond690 ->(work_completion)(&(&slave->notify_work)->work) ->(work_completion)(&(&bond->mcast_work)->work) ->&rq->__lock FD: 232 BD: 1 +.+.: (wq_completion)bond670#3 ->&rq->__lock ->(work_completion)(&(&slave->notify_work)->work) ->(work_completion)(&(&bond->mcast_work)->work) FD: 232 BD: 1 +.+.: (wq_completion)bond683#2 ->(work_completion)(&(&slave->notify_work)->work) ->(work_completion)(&(&bond->mcast_work)->work) FD: 232 BD: 1 +.+.: (wq_completion)bond691 ->(work_completion)(&(&slave->notify_work)->work) ->(work_completion)(&(&bond->mcast_work)->work) ->&rq->__lock FD: 232 BD: 1 +.+.: (wq_completion)bond671#3 ->&rq->__lock ->(work_completion)(&(&slave->notify_work)->work) ->(work_completion)(&(&bond->mcast_work)->work) FD: 232 BD: 1 +.+.: (wq_completion)bond684#2 ->(work_completion)(&(&slave->notify_work)->work) ->(work_completion)(&(&bond->mcast_work)->work) ->&rq->__lock FD: 232 BD: 1 +.+.: (wq_completion)bond692 ->(work_completion)(&(&slave->notify_work)->work) ->(work_completion)(&(&bond->mcast_work)->work) FD: 232 BD: 1 +.+.: (wq_completion)bond672#3 ->&rq->__lock ->(work_completion)(&(&slave->notify_work)->work) ->(work_completion)(&(&bond->mcast_work)->work) FD: 232 BD: 1 +.+.: (wq_completion)bond693 ->&rq->__lock ->(work_completion)(&(&slave->notify_work)->work) ->(work_completion)(&(&bond->mcast_work)->work) FD: 232 BD: 1 +.+.: (wq_completion)bond673#3 ->(work_completion)(&(&slave->notify_work)->work) ->(work_completion)(&(&bond->mcast_work)->work) FD: 232 BD: 1 +.+.: (wq_completion)bond694 ->(work_completion)(&(&slave->notify_work)->work) ->(work_completion)(&(&bond->mcast_work)->work) ->&rq->__lock FD: 232 BD: 1 +.+.: (wq_completion)bond674#3 ->&rq->__lock ->(work_completion)(&(&slave->notify_work)->work) ->(work_completion)(&(&bond->mcast_work)->work) FD: 232 BD: 1 +.+.: (wq_completion)bond687#2 ->(work_completion)(&(&slave->notify_work)->work) ->(work_completion)(&(&bond->mcast_work)->work) FD: 232 BD: 1 +.+.: (wq_completion)bond695 ->(work_completion)(&(&slave->notify_work)->work) ->(work_completion)(&(&bond->mcast_work)->work) FD: 232 BD: 1 +.+.: (wq_completion)bond675#3 ->&rq->__lock ->&cfs_rq->removed.lock ->(work_completion)(&(&slave->notify_work)->work) ->(work_completion)(&(&bond->mcast_work)->work) FD: 232 BD: 1 +.+.: (wq_completion)bond688#2 ->&rq->__lock ->(work_completion)(&(&slave->notify_work)->work) ->(work_completion)(&(&bond->mcast_work)->work) FD: 232 BD: 1 +.+.: (wq_completion)bond696 ->&rq->__lock ->(work_completion)(&(&slave->notify_work)->work) ->(work_completion)(&(&bond->mcast_work)->work) FD: 232 BD: 1 +.+.: (wq_completion)bond676#3 ->(work_completion)(&(&slave->notify_work)->work) ->(work_completion)(&(&bond->mcast_work)->work) FD: 232 BD: 1 +.+.: (wq_completion)bond689#2 ->(work_completion)(&(&slave->notify_work)->work) ->(work_completion)(&(&bond->mcast_work)->work) FD: 232 BD: 1 +.+.: (wq_completion)bond697 ->&rq->__lock ->(work_completion)(&(&slave->notify_work)->work) ->(work_completion)(&(&bond->mcast_work)->work) FD: 232 BD: 1 +.+.: (wq_completion)bond677#3 ->&rq->__lock ->(work_completion)(&(&slave->notify_work)->work) ->(work_completion)(&(&bond->mcast_work)->work) FD: 232 BD: 1 +.+.: (wq_completion)bond690#2 ->&rq->__lock ->(work_completion)(&(&slave->notify_work)->work) ->(work_completion)(&(&bond->mcast_work)->work) FD: 232 BD: 1 +.+.: (wq_completion)bond698 ->&rq->__lock ->(work_completion)(&(&slave->notify_work)->work) ->(work_completion)(&(&bond->mcast_work)->work) FD: 232 BD: 1 +.+.: (wq_completion)bond678#3 ->(work_completion)(&(&slave->notify_work)->work) ->(work_completion)(&(&bond->mcast_work)->work) ->&rq->__lock FD: 232 BD: 1 +.+.: (wq_completion)bond691#2 ->&rq->__lock ->(work_completion)(&(&slave->notify_work)->work) ->(work_completion)(&(&bond->mcast_work)->work) FD: 232 BD: 1 +.+.: (wq_completion)bond699 ->&rq->__lock ->(work_completion)(&(&slave->notify_work)->work) ->(work_completion)(&(&bond->mcast_work)->work) FD: 232 BD: 1 +.+.: (wq_completion)bond679#3 ->(work_completion)(&(&slave->notify_work)->work) ->(work_completion)(&(&bond->mcast_work)->work) FD: 232 BD: 1 +.+.: (wq_completion)bond692#2 ->(work_completion)(&(&slave->notify_work)->work) ->(work_completion)(&(&bond->mcast_work)->work) ->&rq->__lock FD: 232 BD: 1 +.+.: (wq_completion)bond700 ->&rq->__lock ->(work_completion)(&(&slave->notify_work)->work) ->(work_completion)(&(&bond->mcast_work)->work) FD: 232 BD: 1 +.+.: (wq_completion)bond680#3 ->(work_completion)(&(&slave->notify_work)->work) ->(work_completion)(&(&bond->mcast_work)->work) FD: 232 BD: 1 +.+.: (wq_completion)bond693#2 ->(work_completion)(&(&slave->notify_work)->work) ->(work_completion)(&(&bond->mcast_work)->work) ->&rq->__lock FD: 232 BD: 1 +.+.: (wq_completion)bond701 ->(work_completion)(&(&slave->notify_work)->work) ->(work_completion)(&(&bond->mcast_work)->work) FD: 232 BD: 1 +.+.: (wq_completion)bond694#2 ->(work_completion)(&(&slave->notify_work)->work) ->(work_completion)(&(&bond->mcast_work)->work) FD: 232 BD: 1 +.+.: (wq_completion)bond702 ->(work_completion)(&(&slave->notify_work)->work) ->(work_completion)(&(&bond->mcast_work)->work) ->&rq->__lock FD: 232 BD: 1 +.+.: (wq_completion)bond682#3 ->(work_completion)(&(&slave->notify_work)->work) ->(work_completion)(&(&bond->mcast_work)->work) FD: 232 BD: 1 +.+.: (wq_completion)bond695#2 ->&rq->__lock ->(work_completion)(&(&slave->notify_work)->work) ->(work_completion)(&(&bond->mcast_work)->work) FD: 232 BD: 1 +.+.: (wq_completion)bond703 ->(work_completion)(&(&slave->notify_work)->work) ->(work_completion)(&(&bond->mcast_work)->work) FD: 232 BD: 1 +.+.: (wq_completion)bond696#2 ->(work_completion)(&(&slave->notify_work)->work) ->(work_completion)(&(&bond->mcast_work)->work) FD: 232 BD: 1 +.+.: (wq_completion)bond683#3 ->(work_completion)(&(&slave->notify_work)->work) ->(work_completion)(&(&bond->mcast_work)->work) FD: 232 BD: 1 +.+.: (wq_completion)bond704 ->&rq->__lock ->(work_completion)(&(&slave->notify_work)->work) ->(work_completion)(&(&bond->mcast_work)->work) FD: 232 BD: 1 +.+.: (wq_completion)bond697#2 ->(work_completion)(&(&slave->notify_work)->work) ->(work_completion)(&(&bond->mcast_work)->work) FD: 232 BD: 1 +.+.: (wq_completion)bond684#3 ->(work_completion)(&(&slave->notify_work)->work) ->(work_completion)(&(&bond->mcast_work)->work) FD: 232 BD: 1 +.+.: (wq_completion)bond705 ->&rq->__lock ->(work_completion)(&(&slave->notify_work)->work) ->(work_completion)(&(&bond->mcast_work)->work) FD: 232 BD: 1 +.+.: (wq_completion)bond698#2 ->(work_completion)(&(&slave->notify_work)->work) ->(work_completion)(&(&bond->mcast_work)->work) FD: 1 BD: 4 +...: &pn->all_channels_lock FD: 24 BD: 4 +.+.: &pch->chan_sem ->&rq->__lock ->&pch->downl FD: 1 BD: 5 +...: &pch->downl FD: 1 BD: 4 +...: &pch->upl FD: 232 BD: 1 +.+.: (wq_completion)bond685#2 ->(work_completion)(&(&slave->notify_work)->work) ->(work_completion)(&(&bond->mcast_work)->work) FD: 232 BD: 1 +.+.: (wq_completion)bond706 ->&rq->__lock ->(work_completion)(&(&slave->notify_work)->work) ->(work_completion)(&(&bond->mcast_work)->work) FD: 232 BD: 1 +.+.: (wq_completion)bond699#2 ->&rq->__lock ->(work_completion)(&(&slave->notify_work)->work) ->(work_completion)(&(&bond->mcast_work)->work) FD: 232 BD: 1 +.+.: (wq_completion)bond686#2 ->&rq->__lock ->(work_completion)(&(&slave->notify_work)->work) ->(work_completion)(&(&bond->mcast_work)->work) FD: 232 BD: 1 +.+.: (wq_completion)bond700#2 ->&rq->__lock ->(work_completion)(&(&slave->notify_work)->work) ->(work_completion)(&(&bond->mcast_work)->work) FD: 232 BD: 1 +.+.: (wq_completion)bond687#3 ->&rq->__lock ->(work_completion)(&(&slave->notify_work)->work) ->(work_completion)(&(&bond->mcast_work)->work) FD: 232 BD: 1 +.+.: (wq_completion)bond708 ->(work_completion)(&(&slave->notify_work)->work) ->(work_completion)(&(&bond->mcast_work)->work) FD: 232 BD: 1 +.+.: (wq_completion)bond701#2 ->&rq->__lock ->(work_completion)(&(&slave->notify_work)->work) ->(work_completion)(&(&bond->mcast_work)->work) FD: 232 BD: 1 +.+.: (wq_completion)bond688#3 ->&rq->__lock ->(work_completion)(&(&slave->notify_work)->work) ->(work_completion)(&(&bond->mcast_work)->work) FD: 232 BD: 1 +.+.: (wq_completion)bond709 ->&rq->__lock ->(work_completion)(&(&slave->notify_work)->work) ->(work_completion)(&(&bond->mcast_work)->work) FD: 232 BD: 1 +.+.: (wq_completion)bond702#2 ->(work_completion)(&(&slave->notify_work)->work) ->(work_completion)(&(&bond->mcast_work)->work) FD: 232 BD: 1 +.+.: (wq_completion)bond689#3 ->&rq->__lock ->(work_completion)(&(&slave->notify_work)->work) ->(work_completion)(&(&bond->mcast_work)->work) FD: 232 BD: 1 +.+.: (wq_completion)bond710 ->(work_completion)(&(&slave->notify_work)->work) ->(work_completion)(&(&bond->mcast_work)->work) FD: 232 BD: 1 +.+.: (wq_completion)bond703#2 ->(work_completion)(&(&slave->notify_work)->work) ->(work_completion)(&(&bond->mcast_work)->work) FD: 232 BD: 1 +.+.: (wq_completion)bond690#3 ->&rq->__lock ->(work_completion)(&(&slave->notify_work)->work) ->(work_completion)(&(&bond->mcast_work)->work) FD: 232 BD: 1 +.+.: (wq_completion)bond711 ->&rq->__lock ->(work_completion)(&(&slave->notify_work)->work) ->(work_completion)(&(&bond->mcast_work)->work) FD: 232 BD: 1 +.+.: (wq_completion)bond704#2 ->(work_completion)(&(&slave->notify_work)->work) ->(work_completion)(&(&bond->mcast_work)->work) FD: 232 BD: 1 +.+.: (wq_completion)bond691#3 ->(work_completion)(&(&slave->notify_work)->work) ->(work_completion)(&(&bond->mcast_work)->work) FD: 232 BD: 1 +.+.: (wq_completion)bond712 ->&rq->__lock ->(work_completion)(&(&slave->notify_work)->work) ->(work_completion)(&(&bond->mcast_work)->work) FD: 232 BD: 1 +.+.: (wq_completion)bond705#2 ->&rq->__lock ->(work_completion)(&(&slave->notify_work)->work) ->(work_completion)(&(&bond->mcast_work)->work) FD: 232 BD: 1 +.+.: (wq_completion)bond692#3 ->(work_completion)(&(&slave->notify_work)->work) ->(work_completion)(&(&bond->mcast_work)->work) FD: 232 BD: 1 +.+.: (wq_completion)bond713 ->&rq->__lock ->(work_completion)(&(&slave->notify_work)->work) ->(work_completion)(&(&bond->mcast_work)->work) FD: 232 BD: 1 +.+.: (wq_completion)bond706#2 ->&rq->__lock ->(work_completion)(&(&slave->notify_work)->work) ->(work_completion)(&(&bond->mcast_work)->work) FD: 232 BD: 1 +.+.: (wq_completion)bond693#3 ->(work_completion)(&(&slave->notify_work)->work) ->(work_completion)(&(&bond->mcast_work)->work) FD: 232 BD: 1 +.+.: (wq_completion)bond714 ->&rq->__lock ->(work_completion)(&(&slave->notify_work)->work) ->(work_completion)(&(&bond->mcast_work)->work) FD: 232 BD: 1 +.+.: (wq_completion)bond694#3 ->(work_completion)(&(&slave->notify_work)->work) ->(work_completion)(&(&bond->mcast_work)->work) FD: 232 BD: 1 +.+.: (wq_completion)bond708#2 ->&rq->__lock ->(work_completion)(&(&slave->notify_work)->work) ->(work_completion)(&(&bond->mcast_work)->work) FD: 232 BD: 1 +.+.: (wq_completion)bond715 ->&rq->__lock ->(work_completion)(&(&slave->notify_work)->work) ->(work_completion)(&(&bond->mcast_work)->work) FD: 232 BD: 1 +.+.: (wq_completion)bond695#3 ->(work_completion)(&(&slave->notify_work)->work) ->(work_completion)(&(&bond->mcast_work)->work) FD: 232 BD: 1 +.+.: (wq_completion)bond709#2 ->&rq->__lock ->(work_completion)(&(&slave->notify_work)->work) ->(work_completion)(&(&bond->mcast_work)->work) FD: 232 BD: 1 +.+.: (wq_completion)bond716 ->(work_completion)(&(&slave->notify_work)->work) ->(work_completion)(&(&bond->mcast_work)->work) FD: 232 BD: 1 +.+.: (wq_completion)bond696#3 ->&rq->__lock ->(work_completion)(&(&slave->notify_work)->work) ->(work_completion)(&(&bond->mcast_work)->work) FD: 232 BD: 1 +.+.: (wq_completion)bond710#2 ->&rq->__lock ->(work_completion)(&(&slave->notify_work)->work) ->(work_completion)(&(&bond->mcast_work)->work) FD: 232 BD: 1 +.+.: (wq_completion)bond717 ->(work_completion)(&(&slave->notify_work)->work) ->(work_completion)(&(&bond->mcast_work)->work) FD: 232 BD: 1 +.+.: (wq_completion)bond697#3 ->&rq->__lock ->(work_completion)(&(&slave->notify_work)->work) ->(work_completion)(&(&bond->mcast_work)->work) FD: 232 BD: 1 +.+.: (wq_completion)bond711#2 ->(work_completion)(&(&slave->notify_work)->work) ->(work_completion)(&(&bond->mcast_work)->work) FD: 232 BD: 1 +.+.: (wq_completion)bond718 ->&rq->__lock ->(work_completion)(&(&slave->notify_work)->work) ->(work_completion)(&(&bond->mcast_work)->work) FD: 232 BD: 1 +.+.: (wq_completion)bond698#3 ->(work_completion)(&(&slave->notify_work)->work) ->(work_completion)(&(&bond->mcast_work)->work) FD: 232 BD: 1 +.+.: (wq_completion)bond712#2 ->(work_completion)(&(&slave->notify_work)->work) ->(work_completion)(&(&bond->mcast_work)->work) FD: 232 BD: 1 +.+.: (wq_completion)bond699#3 ->&rq->__lock ->(work_completion)(&(&slave->notify_work)->work) ->(work_completion)(&(&bond->mcast_work)->work) FD: 232 BD: 1 +.+.: (wq_completion)bond719 ->(work_completion)(&(&slave->notify_work)->work) ->(work_completion)(&(&bond->mcast_work)->work) ->&rq->__lock FD: 232 BD: 1 +.+.: (wq_completion)bond713#2 ->&rq->__lock ->(work_completion)(&(&slave->notify_work)->work) ->(work_completion)(&(&bond->mcast_work)->work) FD: 232 BD: 1 +.+.: (wq_completion)bond700#3 ->&rq->__lock ->(work_completion)(&(&slave->notify_work)->work) ->(work_completion)(&(&bond->mcast_work)->work) FD: 232 BD: 1 +.+.: (wq_completion)bond720 ->(work_completion)(&(&slave->notify_work)->work) ->(work_completion)(&(&bond->mcast_work)->work) FD: 232 BD: 1 +.+.: (wq_completion)bond721 ->&rq->__lock ->(work_completion)(&(&slave->notify_work)->work) ->(work_completion)(&(&bond->mcast_work)->work) FD: 232 BD: 1 +.+.: (wq_completion)bond701#3 ->&rq->__lock ->(work_completion)(&(&slave->notify_work)->work) ->(work_completion)(&(&bond->mcast_work)->work) FD: 232 BD: 1 +.+.: (wq_completion)bond714#2 ->&rq->__lock ->(work_completion)(&(&slave->notify_work)->work) ->(work_completion)(&(&bond->mcast_work)->work) FD: 232 BD: 1 +.+.: (wq_completion)bond702#3 ->(work_completion)(&(&slave->notify_work)->work) ->(work_completion)(&(&bond->mcast_work)->work) ->&rq->__lock FD: 232 BD: 1 +.+.: (wq_completion)bond722 ->(work_completion)(&(&slave->notify_work)->work) ->(work_completion)(&(&bond->mcast_work)->work) ->&rq->__lock FD: 232 BD: 1 +.+.: (wq_completion)bond715#2 ->(work_completion)(&(&slave->notify_work)->work) ->(work_completion)(&(&bond->mcast_work)->work) FD: 232 BD: 1 +.+.: (wq_completion)bond703#3 ->(work_completion)(&(&slave->notify_work)->work) ->(work_completion)(&(&bond->mcast_work)->work) FD: 232 BD: 1 +.+.: (wq_completion)bond723 ->(work_completion)(&(&slave->notify_work)->work) ->(work_completion)(&(&bond->mcast_work)->work) ->&rq->__lock FD: 232 BD: 1 +.+.: (wq_completion)bond716#2 ->(work_completion)(&(&slave->notify_work)->work) ->(work_completion)(&(&bond->mcast_work)->work) ->&rq->__lock FD: 232 BD: 1 +.+.: (wq_completion)bond704#3 ->(work_completion)(&(&slave->notify_work)->work) ->(work_completion)(&(&bond->mcast_work)->work) FD: 232 BD: 1 +.+.: (wq_completion)bond724 ->(work_completion)(&(&slave->notify_work)->work) ->(work_completion)(&(&bond->mcast_work)->work) FD: 232 BD: 1 +.+.: (wq_completion)bond717#2 ->&rq->__lock ->(work_completion)(&(&slave->notify_work)->work) ->(work_completion)(&(&bond->mcast_work)->work) FD: 232 BD: 1 +.+.: (wq_completion)bond705#3 ->(work_completion)(&(&slave->notify_work)->work) ->(work_completion)(&(&bond->mcast_work)->work) ->&rq->__lock FD: 232 BD: 1 +.+.: (wq_completion)bond725 ->(work_completion)(&(&slave->notify_work)->work) ->(work_completion)(&(&bond->mcast_work)->work) ->&rq->__lock ->&cfs_rq->removed.lock ->&obj_hash[i].lock FD: 232 BD: 1 +.+.: (wq_completion)bond718#2 ->(work_completion)(&(&slave->notify_work)->work) ->(work_completion)(&(&bond->mcast_work)->work) ->&rq->__lock FD: 232 BD: 1 +.+.: (wq_completion)bond706#3 ->(work_completion)(&(&slave->notify_work)->work) ->(work_completion)(&(&bond->mcast_work)->work) ->&rq->__lock FD: 232 BD: 1 +.+.: (wq_completion)bond726 ->&rq->__lock ->(work_completion)(&(&slave->notify_work)->work) ->(work_completion)(&(&bond->mcast_work)->work) FD: 232 BD: 1 +.+.: (wq_completion)bond719#2 ->(work_completion)(&(&slave->notify_work)->work) ->(work_completion)(&(&bond->mcast_work)->work) FD: 232 BD: 1 +.+.: (wq_completion)bond707 ->(work_completion)(&(&slave->notify_work)->work) ->(work_completion)(&(&bond->mcast_work)->work) ->&rq->__lock FD: 232 BD: 1 +.+.: (wq_completion)bond727 ->&rq->__lock ->(work_completion)(&(&slave->notify_work)->work) ->(work_completion)(&(&bond->mcast_work)->work) FD: 232 BD: 1 +.+.: (wq_completion)bond720#2 ->(work_completion)(&(&slave->notify_work)->work) ->(work_completion)(&(&bond->mcast_work)->work) ->&rq->__lock FD: 232 BD: 1 +.+.: (wq_completion)bond708#3 ->(work_completion)(&(&slave->notify_work)->work) ->(work_completion)(&(&bond->mcast_work)->work) FD: 232 BD: 1 +.+.: (wq_completion)bond728 ->(work_completion)(&(&slave->notify_work)->work) ->(work_completion)(&(&bond->mcast_work)->work) FD: 232 BD: 1 +.+.: (wq_completion)bond721#2 ->(work_completion)(&(&slave->notify_work)->work) ->(work_completion)(&(&bond->mcast_work)->work) ->&rq->__lock FD: 232 BD: 1 +.+.: (wq_completion)bond709#3 ->(work_completion)(&(&slave->notify_work)->work) ->(work_completion)(&(&bond->mcast_work)->work) ->&rq->__lock FD: 232 BD: 1 +.+.: (wq_completion)bond729 ->(work_completion)(&(&slave->notify_work)->work) ->(work_completion)(&(&bond->mcast_work)->work) FD: 232 BD: 1 +.+.: (wq_completion)bond722#2 ->&rq->__lock ->(work_completion)(&(&slave->notify_work)->work) ->(work_completion)(&(&bond->mcast_work)->work) FD: 232 BD: 1 +.+.: (wq_completion)bond730 ->&rq->__lock ->(work_completion)(&(&slave->notify_work)->work) ->(work_completion)(&(&bond->mcast_work)->work) FD: 232 BD: 1 +.+.: (wq_completion)bond710#3 ->&rq->__lock ->(work_completion)(&(&slave->notify_work)->work) ->(work_completion)(&(&bond->mcast_work)->work) FD: 232 BD: 1 +.+.: (wq_completion)bond723#2 ->&rq->__lock ->(work_completion)(&(&slave->notify_work)->work) ->(work_completion)(&(&bond->mcast_work)->work) FD: 232 BD: 1 +.+.: (wq_completion)bond731 ->(work_completion)(&(&slave->notify_work)->work) ->(work_completion)(&(&bond->mcast_work)->work) FD: 232 BD: 1 +.+.: (wq_completion)bond711#3 ->(work_completion)(&(&slave->notify_work)->work) ->(work_completion)(&(&bond->mcast_work)->work) ->&rq->__lock FD: 232 BD: 1 +.+.: (wq_completion)bond724#2 ->(work_completion)(&(&slave->notify_work)->work) ->(work_completion)(&(&bond->mcast_work)->work) ->&rq->__lock FD: 232 BD: 1 +.+.: (wq_completion)bond732 ->(work_completion)(&(&slave->notify_work)->work) ->(work_completion)(&(&bond->mcast_work)->work) ->&rq->__lock FD: 232 BD: 1 +.+.: (wq_completion)bond712#3 ->(work_completion)(&(&slave->notify_work)->work) ->(work_completion)(&(&bond->mcast_work)->work) ->&rq->__lock FD: 232 BD: 1 +.+.: (wq_completion)bond725#2 ->(work_completion)(&(&slave->notify_work)->work) ->(work_completion)(&(&bond->mcast_work)->work) FD: 232 BD: 1 +.+.: (wq_completion)bond733 ->&rq->__lock ->(work_completion)(&(&slave->notify_work)->work) ->(work_completion)(&(&bond->mcast_work)->work) FD: 232 BD: 1 +.+.: (wq_completion)bond713#3 ->&rq->__lock ->(work_completion)(&(&slave->notify_work)->work) ->(work_completion)(&(&bond->mcast_work)->work) FD: 232 BD: 1 +.+.: (wq_completion)bond726#2 ->(work_completion)(&(&slave->notify_work)->work) ->(work_completion)(&(&bond->mcast_work)->work) FD: 232 BD: 1 +.+.: (wq_completion)bond734 ->&rq->__lock ->(work_completion)(&(&slave->notify_work)->work) ->(work_completion)(&(&bond->mcast_work)->work) FD: 232 BD: 1 +.+.: (wq_completion)bond714#3 ->&rq->__lock ->(work_completion)(&(&slave->notify_work)->work) ->(work_completion)(&(&bond->mcast_work)->work) FD: 232 BD: 1 +.+.: (wq_completion)bond727#2 ->(work_completion)(&(&slave->notify_work)->work) ->(work_completion)(&(&bond->mcast_work)->work) FD: 232 BD: 1 +.+.: (wq_completion)bond735 ->&rq->__lock ->(work_completion)(&(&slave->notify_work)->work) ->(work_completion)(&(&bond->mcast_work)->work) FD: 232 BD: 1 +.+.: (wq_completion)bond715#3 ->(work_completion)(&(&slave->notify_work)->work) ->(work_completion)(&(&bond->mcast_work)->work) FD: 232 BD: 1 +.+.: (wq_completion)bond728#2 ->&rq->__lock ->(work_completion)(&(&slave->notify_work)->work) ->(work_completion)(&(&bond->mcast_work)->work) FD: 232 BD: 1 +.+.: (wq_completion)bond736 ->(work_completion)(&(&slave->notify_work)->work) ->(work_completion)(&(&bond->mcast_work)->work) FD: 232 BD: 1 +.+.: (wq_completion)bond716#3 ->(work_completion)(&(&slave->notify_work)->work) ->(work_completion)(&(&bond->mcast_work)->work) ->&rq->__lock FD: 232 BD: 1 +.+.: (wq_completion)bond729#2 ->(work_completion)(&(&slave->notify_work)->work) ->(work_completion)(&(&bond->mcast_work)->work) ->&rq->__lock FD: 232 BD: 1 +.+.: (wq_completion)bond737 ->&rq->__lock ->(work_completion)(&(&slave->notify_work)->work) ->(work_completion)(&(&bond->mcast_work)->work) FD: 232 BD: 1 +.+.: (wq_completion)bond717#3 ->(work_completion)(&(&slave->notify_work)->work) ->(work_completion)(&(&bond->mcast_work)->work) ->&rq->__lock FD: 232 BD: 1 +.+.: (wq_completion)bond730#2 ->&rq->__lock ->(work_completion)(&(&slave->notify_work)->work) ->(work_completion)(&(&bond->mcast_work)->work) FD: 232 BD: 1 +.+.: (wq_completion)bond738 ->&rq->__lock ->(work_completion)(&(&slave->notify_work)->work) ->(work_completion)(&(&bond->mcast_work)->work) FD: 232 BD: 1 +.+.: (wq_completion)bond718#3 ->&rq->__lock ->(work_completion)(&(&slave->notify_work)->work) ->(work_completion)(&(&bond->mcast_work)->work) FD: 232 BD: 1 +.+.: (wq_completion)bond731#2 ->&rq->__lock ->(work_completion)(&(&slave->notify_work)->work) ->(work_completion)(&(&bond->mcast_work)->work) FD: 232 BD: 1 +.+.: (wq_completion)bond739 ->(work_completion)(&(&slave->notify_work)->work) ->(work_completion)(&(&bond->mcast_work)->work) ->&rq->__lock FD: 232 BD: 1 +.+.: (wq_completion)bond719#3 ->(work_completion)(&(&slave->notify_work)->work) ->(work_completion)(&(&bond->mcast_work)->work) ->&rq->__lock FD: 232 BD: 1 +.+.: (wq_completion)bond732#2 ->(work_completion)(&(&slave->notify_work)->work) ->(work_completion)(&(&bond->mcast_work)->work) ->&rq->__lock FD: 232 BD: 1 +.+.: (wq_completion)bond740 ->(work_completion)(&(&slave->notify_work)->work) ->(work_completion)(&(&bond->mcast_work)->work) FD: 232 BD: 1 +.+.: (wq_completion)bond720#3 ->&rq->__lock ->(work_completion)(&(&slave->notify_work)->work) ->(work_completion)(&(&bond->mcast_work)->work) FD: 232 BD: 1 +.+.: (wq_completion)bond733#2 ->&rq->__lock ->(work_completion)(&(&slave->notify_work)->work) ->(work_completion)(&(&bond->mcast_work)->work) FD: 232 BD: 1 +.+.: (wq_completion)bond741 ->&rq->__lock ->(work_completion)(&(&slave->notify_work)->work) ->(work_completion)(&(&bond->mcast_work)->work) FD: 232 BD: 1 +.+.: (wq_completion)bond721#3 ->(work_completion)(&(&slave->notify_work)->work) ->(work_completion)(&(&bond->mcast_work)->work) ->&rq->__lock FD: 232 BD: 1 +.+.: (wq_completion)bond734#2 ->&rq->__lock ->(work_completion)(&(&slave->notify_work)->work) ->(work_completion)(&(&bond->mcast_work)->work) FD: 232 BD: 1 +.+.: (wq_completion)bond742 ->(work_completion)(&(&slave->notify_work)->work) ->(work_completion)(&(&bond->mcast_work)->work) ->&rq->__lock FD: 232 BD: 1 +.+.: (wq_completion)bond722#3 ->&rq->__lock ->(work_completion)(&(&slave->notify_work)->work) ->(work_completion)(&(&bond->mcast_work)->work) ->&cfs_rq->removed.lock ->&obj_hash[i].lock FD: 232 BD: 1 +.+.: (wq_completion)bond735#2 ->(work_completion)(&(&slave->notify_work)->work) ->(work_completion)(&(&bond->mcast_work)->work) ->&rq->__lock FD: 232 BD: 1 +.+.: (wq_completion)bond743 ->(work_completion)(&(&slave->notify_work)->work) ->(work_completion)(&(&bond->mcast_work)->work) FD: 232 BD: 1 +.+.: (wq_completion)bond723#3 ->(work_completion)(&(&slave->notify_work)->work) ->(work_completion)(&(&bond->mcast_work)->work) ->&rq->__lock FD: 232 BD: 1 +.+.: (wq_completion)bond736#2 ->&rq->__lock ->(work_completion)(&(&slave->notify_work)->work) ->(work_completion)(&(&bond->mcast_work)->work) FD: 232 BD: 1 +.+.: (wq_completion)bond744 ->(work_completion)(&(&slave->notify_work)->work) ->(work_completion)(&(&bond->mcast_work)->work) ->&rq->__lock FD: 232 BD: 1 +.+.: (wq_completion)bond724#3 ->&rq->__lock ->(work_completion)(&(&slave->notify_work)->work) ->(work_completion)(&(&bond->mcast_work)->work) FD: 232 BD: 1 +.+.: (wq_completion)bond737#2 ->(work_completion)(&(&slave->notify_work)->work) ->(work_completion)(&(&bond->mcast_work)->work) FD: 232 BD: 1 +.+.: (wq_completion)bond745 ->(work_completion)(&(&slave->notify_work)->work) ->(work_completion)(&(&bond->mcast_work)->work) ->&rq->__lock FD: 232 BD: 1 +.+.: (wq_completion)bond725#3 ->(work_completion)(&(&slave->notify_work)->work) ->(work_completion)(&(&bond->mcast_work)->work) ->&rq->__lock FD: 232 BD: 1 +.+.: (wq_completion)bond738#2 ->(work_completion)(&(&slave->notify_work)->work) ->(work_completion)(&(&bond->mcast_work)->work) FD: 232 BD: 1 +.+.: (wq_completion)bond746 ->(work_completion)(&(&slave->notify_work)->work) ->(work_completion)(&(&bond->mcast_work)->work) FD: 232 BD: 1 +.+.: (wq_completion)bond726#3 ->(work_completion)(&(&slave->notify_work)->work) ->(work_completion)(&(&bond->mcast_work)->work) FD: 232 BD: 1 +.+.: (wq_completion)bond739#2 ->(work_completion)(&(&slave->notify_work)->work) ->(work_completion)(&(&bond->mcast_work)->work) ->&rq->__lock FD: 232 BD: 1 +.+.: (wq_completion)bond747 ->&rq->__lock ->(work_completion)(&(&slave->notify_work)->work) ->(work_completion)(&(&bond->mcast_work)->work) FD: 232 BD: 1 +.+.: (wq_completion)bond727#3 ->&rq->__lock ->(work_completion)(&(&slave->notify_work)->work) ->(work_completion)(&(&bond->mcast_work)->work) FD: 232 BD: 1 +.+.: (wq_completion)bond740#2 ->&rq->__lock ->(work_completion)(&(&slave->notify_work)->work) ->(work_completion)(&(&bond->mcast_work)->work) FD: 232 BD: 1 +.+.: (wq_completion)bond748 ->(work_completion)(&(&slave->notify_work)->work) ->(work_completion)(&(&bond->mcast_work)->work) ->&rq->__lock FD: 232 BD: 1 +.+.: (wq_completion)bond728#3 ->(work_completion)(&(&slave->notify_work)->work) ->(work_completion)(&(&bond->mcast_work)->work) ->&rq->__lock FD: 232 BD: 1 +.+.: (wq_completion)bond741#2 ->&rq->__lock ->(work_completion)(&(&slave->notify_work)->work) ->(work_completion)(&(&bond->mcast_work)->work) FD: 232 BD: 1 +.+.: (wq_completion)bond749 ->&rq->__lock ->(work_completion)(&(&slave->notify_work)->work) ->(work_completion)(&(&bond->mcast_work)->work) FD: 232 BD: 1 +.+.: (wq_completion)bond742#2 ->&rq->__lock ->(work_completion)(&(&slave->notify_work)->work) ->(work_completion)(&(&bond->mcast_work)->work) FD: 232 BD: 1 +.+.: (wq_completion)bond750 ->&rq->__lock ->(work_completion)(&(&slave->notify_work)->work) ->(work_completion)(&(&bond->mcast_work)->work) FD: 232 BD: 1 +.+.: (wq_completion)bond730#3 ->(work_completion)(&(&slave->notify_work)->work) ->(work_completion)(&(&bond->mcast_work)->work) ->&rq->__lock FD: 232 BD: 1 +.+.: (wq_completion)bond743#2 ->&rq->__lock ->(work_completion)(&(&slave->notify_work)->work) ->(work_completion)(&(&bond->mcast_work)->work) FD: 232 BD: 1 +.+.: (wq_completion)bond751 ->&rq->__lock ->(work_completion)(&(&slave->notify_work)->work) ->(work_completion)(&(&bond->mcast_work)->work) FD: 232 BD: 1 +.+.: (wq_completion)bond731#3 ->&rq->__lock ->(work_completion)(&(&slave->notify_work)->work) ->(work_completion)(&(&bond->mcast_work)->work) FD: 232 BD: 1 +.+.: (wq_completion)bond744#2 ->(work_completion)(&(&slave->notify_work)->work) ->(work_completion)(&(&bond->mcast_work)->work) FD: 232 BD: 1 +.+.: (wq_completion)bond752 ->(work_completion)(&(&slave->notify_work)->work) ->(work_completion)(&(&bond->mcast_work)->work) ->&rq->__lock FD: 232 BD: 1 +.+.: (wq_completion)bond732#3 ->(work_completion)(&(&slave->notify_work)->work) ->(work_completion)(&(&bond->mcast_work)->work) FD: 232 BD: 1 +.+.: (wq_completion)bond745#2 ->(work_completion)(&(&slave->notify_work)->work) ->(work_completion)(&(&bond->mcast_work)->work) ->&rq->__lock FD: 232 BD: 1 +.+.: (wq_completion)bond753 ->&rq->__lock ->(work_completion)(&(&slave->notify_work)->work) ->(work_completion)(&(&bond->mcast_work)->work) ->&cfs_rq->removed.lock ->&obj_hash[i].lock FD: 232 BD: 1 +.+.: (wq_completion)bond733#3 ->(work_completion)(&(&slave->notify_work)->work) ->(work_completion)(&(&bond->mcast_work)->work) ->&rq->__lock ->rcu_node_0 ->&rcu_state.expedited_wq FD: 232 BD: 1 +.+.: (wq_completion)bond746#2 ->&rq->__lock ->(work_completion)(&(&slave->notify_work)->work) ->(work_completion)(&(&bond->mcast_work)->work) FD: 232 BD: 1 +.+.: (wq_completion)bond754 ->&rq->__lock ->(work_completion)(&(&slave->notify_work)->work) ->(work_completion)(&(&bond->mcast_work)->work) FD: 232 BD: 1 +.+.: (wq_completion)bond734#3 ->(work_completion)(&(&slave->notify_work)->work) ->(work_completion)(&(&bond->mcast_work)->work) ->&rq->__lock FD: 232 BD: 1 +.+.: (wq_completion)bond747#2 ->&rq->__lock ->(work_completion)(&(&slave->notify_work)->work) ->(work_completion)(&(&bond->mcast_work)->work) FD: 232 BD: 1 +.+.: (wq_completion)bond755 ->&rq->__lock ->(work_completion)(&(&slave->notify_work)->work) ->(work_completion)(&(&bond->mcast_work)->work) FD: 232 BD: 1 +.+.: (wq_completion)bond735#3 ->&rq->__lock ->(work_completion)(&(&slave->notify_work)->work) ->(work_completion)(&(&bond->mcast_work)->work) FD: 232 BD: 1 +.+.: (wq_completion)bond748#2 ->(work_completion)(&(&slave->notify_work)->work) ->(work_completion)(&(&bond->mcast_work)->work) ->&rq->__lock FD: 232 BD: 1 +.+.: (wq_completion)bond756 ->(work_completion)(&(&slave->notify_work)->work) ->(work_completion)(&(&bond->mcast_work)->work) ->&rq->__lock FD: 232 BD: 1 +.+.: (wq_completion)bond736#3 ->(work_completion)(&(&slave->notify_work)->work) ->(work_completion)(&(&bond->mcast_work)->work) ->&rq->__lock FD: 232 BD: 1 +.+.: (wq_completion)bond749#2 ->(work_completion)(&(&slave->notify_work)->work) ->(work_completion)(&(&bond->mcast_work)->work) ->&rq->__lock FD: 232 BD: 1 +.+.: (wq_completion)bond757 ->(work_completion)(&(&slave->notify_work)->work) ->(work_completion)(&(&bond->mcast_work)->work) FD: 232 BD: 1 +.+.: (wq_completion)bond737#3 ->&rq->__lock ->(work_completion)(&(&slave->notify_work)->work) ->(work_completion)(&(&bond->mcast_work)->work) FD: 232 BD: 1 +.+.: (wq_completion)bond750#2 ->&rq->__lock ->(work_completion)(&(&slave->notify_work)->work) ->(work_completion)(&(&bond->mcast_work)->work) FD: 232 BD: 1 +.+.: (wq_completion)bond758 ->&rq->__lock ->(work_completion)(&(&slave->notify_work)->work) ->(work_completion)(&(&bond->mcast_work)->work) FD: 232 BD: 1 +.+.: (wq_completion)bond738#3 ->(work_completion)(&(&slave->notify_work)->work) ->(work_completion)(&(&bond->mcast_work)->work) ->&rq->__lock FD: 232 BD: 1 +.+.: (wq_completion)bond751#2 ->&rq->__lock ->(work_completion)(&(&slave->notify_work)->work) ->(work_completion)(&(&bond->mcast_work)->work) FD: 232 BD: 1 +.+.: (wq_completion)bond759 ->&rq->__lock ->(work_completion)(&(&slave->notify_work)->work) ->(work_completion)(&(&bond->mcast_work)->work) FD: 232 BD: 1 +.+.: (wq_completion)bond739#3 ->(work_completion)(&(&slave->notify_work)->work) ->(work_completion)(&(&bond->mcast_work)->work) ->&rq->__lock FD: 232 BD: 1 +.+.: (wq_completion)bond752#2 ->&rq->__lock ->(work_completion)(&(&slave->notify_work)->work) ->(work_completion)(&(&bond->mcast_work)->work) FD: 232 BD: 1 +.+.: (wq_completion)bond760 ->&rq->__lock ->(work_completion)(&(&slave->notify_work)->work) ->(work_completion)(&(&bond->mcast_work)->work) FD: 232 BD: 1 +.+.: (wq_completion)bond740#3 ->(work_completion)(&(&slave->notify_work)->work) ->(work_completion)(&(&bond->mcast_work)->work) ->&rq->__lock FD: 232 BD: 1 +.+.: (wq_completion)bond753#2 ->&rq->__lock ->(work_completion)(&(&slave->notify_work)->work) ->(work_completion)(&(&bond->mcast_work)->work) FD: 232 BD: 1 +.+.: (wq_completion)bond761 ->(work_completion)(&(&slave->notify_work)->work) ->(work_completion)(&(&bond->mcast_work)->work) FD: 232 BD: 1 +.+.: (wq_completion)bond741#3 ->&rq->__lock ->(work_completion)(&(&slave->notify_work)->work) ->(work_completion)(&(&bond->mcast_work)->work) FD: 232 BD: 1 +.+.: (wq_completion)bond754#2 ->&rq->__lock ->(work_completion)(&(&slave->notify_work)->work) ->(work_completion)(&(&bond->mcast_work)->work) FD: 232 BD: 1 +.+.: (wq_completion)bond762 ->&rq->__lock ->(work_completion)(&(&slave->notify_work)->work) ->(work_completion)(&(&bond->mcast_work)->work) FD: 232 BD: 1 +.+.: (wq_completion)bond742#3 ->&rq->__lock ->(work_completion)(&(&slave->notify_work)->work) ->(work_completion)(&(&bond->mcast_work)->work) FD: 232 BD: 1 +.+.: (wq_completion)bond755#2 ->&rq->__lock ->(work_completion)(&(&slave->notify_work)->work) ->(work_completion)(&(&bond->mcast_work)->work) FD: 232 BD: 1 +.+.: (wq_completion)bond763 ->&rq->__lock ->(work_completion)(&(&slave->notify_work)->work) ->(work_completion)(&(&bond->mcast_work)->work) FD: 232 BD: 1 +.+.: (wq_completion)bond743#3 ->&rq->__lock ->(work_completion)(&(&slave->notify_work)->work) ->(work_completion)(&(&bond->mcast_work)->work) FD: 232 BD: 1 +.+.: (wq_completion)bond756#2 ->&rq->__lock ->(work_completion)(&(&slave->notify_work)->work) ->(work_completion)(&(&bond->mcast_work)->work) FD: 232 BD: 1 +.+.: (wq_completion)bond764 ->&rq->__lock ->(work_completion)(&(&slave->notify_work)->work) ->(work_completion)(&(&bond->mcast_work)->work) FD: 232 BD: 1 +.+.: (wq_completion)bond744#3 ->(work_completion)(&(&slave->notify_work)->work) ->(work_completion)(&(&bond->mcast_work)->work) FD: 232 BD: 1 +.+.: (wq_completion)bond757#2 ->(work_completion)(&(&slave->notify_work)->work) ->(work_completion)(&(&bond->mcast_work)->work) ->&rq->__lock FD: 232 BD: 1 +.+.: (wq_completion)bond765 ->&rq->__lock ->(work_completion)(&(&slave->notify_work)->work) ->(work_completion)(&(&bond->mcast_work)->work) FD: 232 BD: 1 +.+.: (wq_completion)bond745#3 ->&rq->__lock ->(work_completion)(&(&slave->notify_work)->work) ->(work_completion)(&(&bond->mcast_work)->work) FD: 232 BD: 1 +.+.: (wq_completion)bond758#2 ->(work_completion)(&(&slave->notify_work)->work) ->(work_completion)(&(&bond->mcast_work)->work) FD: 232 BD: 1 +.+.: (wq_completion)bond766 ->(work_completion)(&(&slave->notify_work)->work) ->(work_completion)(&(&bond->mcast_work)->work) ->&rq->__lock FD: 232 BD: 1 +.+.: (wq_completion)bond746#3 ->(work_completion)(&(&slave->notify_work)->work) ->(work_completion)(&(&bond->mcast_work)->work) ->&rq->__lock FD: 232 BD: 1 +.+.: (wq_completion)bond767 ->(work_completion)(&(&slave->notify_work)->work) ->(work_completion)(&(&bond->mcast_work)->work) ->&rq->__lock ->&cfs_rq->removed.lock ->&obj_hash[i].lock FD: 232 BD: 1 +.+.: (wq_completion)bond747#3 ->(work_completion)(&(&slave->notify_work)->work) ->(work_completion)(&(&bond->mcast_work)->work) ->&rq->__lock FD: 232 BD: 1 +.+.: (wq_completion)bond760#2 ->(work_completion)(&(&slave->notify_work)->work) ->(work_completion)(&(&bond->mcast_work)->work) ->&rq->__lock FD: 232 BD: 1 +.+.: (wq_completion)bond768 ->(work_completion)(&(&slave->notify_work)->work) ->(work_completion)(&(&bond->mcast_work)->work) ->&rq->__lock FD: 232 BD: 1 +.+.: (wq_completion)bond748#3 ->&rq->__lock ->(work_completion)(&(&slave->notify_work)->work) ->(work_completion)(&(&bond->mcast_work)->work) FD: 232 BD: 1 +.+.: (wq_completion)bond761#2 ->&rq->__lock ->(work_completion)(&(&slave->notify_work)->work) ->(work_completion)(&(&bond->mcast_work)->work) FD: 232 BD: 1 +.+.: (wq_completion)bond769 ->(work_completion)(&(&slave->notify_work)->work) ->(work_completion)(&(&bond->mcast_work)->work) ->&rq->__lock FD: 232 BD: 1 +.+.: (wq_completion)bond749#3 ->&rq->__lock ->(work_completion)(&(&slave->notify_work)->work) ->(work_completion)(&(&bond->mcast_work)->work) FD: 232 BD: 1 +.+.: (wq_completion)bond762#2 ->(work_completion)(&(&slave->notify_work)->work) ->(work_completion)(&(&bond->mcast_work)->work) ->&rq->__lock FD: 232 BD: 1 +.+.: (wq_completion)bond770 ->&rq->__lock ->(work_completion)(&(&slave->notify_work)->work) ->(work_completion)(&(&bond->mcast_work)->work) FD: 232 BD: 1 +.+.: (wq_completion)bond750#3 ->&rq->__lock ->(work_completion)(&(&slave->notify_work)->work) ->(work_completion)(&(&bond->mcast_work)->work) FD: 232 BD: 1 +.+.: (wq_completion)bond763#2 ->(work_completion)(&(&slave->notify_work)->work) ->(work_completion)(&(&bond->mcast_work)->work) ->&rq->__lock FD: 232 BD: 1 +.+.: (wq_completion)bond771 ->&rq->__lock ->(work_completion)(&(&slave->notify_work)->work) ->(work_completion)(&(&bond->mcast_work)->work) FD: 232 BD: 1 +.+.: (wq_completion)bond751#3 ->(work_completion)(&(&slave->notify_work)->work) ->(work_completion)(&(&bond->mcast_work)->work) FD: 232 BD: 1 +.+.: (wq_completion)bond764#2 ->&rq->__lock ->(work_completion)(&(&slave->notify_work)->work) ->(work_completion)(&(&bond->mcast_work)->work) FD: 232 BD: 1 +.+.: (wq_completion)bond772 ->&rq->__lock ->(work_completion)(&(&slave->notify_work)->work) ->(work_completion)(&(&bond->mcast_work)->work) FD: 232 BD: 1 +.+.: (wq_completion)bond752#3 ->&rq->__lock ->(work_completion)(&(&slave->notify_work)->work) ->(work_completion)(&(&bond->mcast_work)->work) FD: 232 BD: 1 +.+.: (wq_completion)bond765#2 ->&rq->__lock ->(work_completion)(&(&slave->notify_work)->work) ->(work_completion)(&(&bond->mcast_work)->work) FD: 232 BD: 1 +.+.: (wq_completion)bond773 ->(work_completion)(&(&slave->notify_work)->work) ->(work_completion)(&(&bond->mcast_work)->work) ->&rq->__lock FD: 232 BD: 1 +.+.: (wq_completion)bond753#3 ->&rq->__lock ->(work_completion)(&(&slave->notify_work)->work) ->(work_completion)(&(&bond->mcast_work)->work) FD: 232 BD: 1 +.+.: (wq_completion)bond766#2 ->(work_completion)(&(&slave->notify_work)->work) ->(work_completion)(&(&bond->mcast_work)->work) ->&rq->__lock FD: 232 BD: 1 +.+.: (wq_completion)bond774 ->&rq->__lock ->(work_completion)(&(&slave->notify_work)->work) ->(work_completion)(&(&bond->mcast_work)->work) FD: 232 BD: 1 +.+.: (wq_completion)bond754#3 ->&rq->__lock ->(work_completion)(&(&slave->notify_work)->work) ->(work_completion)(&(&bond->mcast_work)->work) FD: 232 BD: 1 +.+.: (wq_completion)bond767#2 ->(work_completion)(&(&slave->notify_work)->work) ->(work_completion)(&(&bond->mcast_work)->work) ->&rq->__lock FD: 232 BD: 1 +.+.: (wq_completion)bond775 ->&rq->__lock ->(work_completion)(&(&slave->notify_work)->work) ->(work_completion)(&(&bond->mcast_work)->work) FD: 232 BD: 1 +.+.: (wq_completion)bond755#3 ->(work_completion)(&(&slave->notify_work)->work) ->(work_completion)(&(&bond->mcast_work)->work) ->&rq->__lock FD: 232 BD: 1 +.+.: (wq_completion)bond768#2 ->&rq->__lock ->(work_completion)(&(&slave->notify_work)->work) ->(work_completion)(&(&bond->mcast_work)->work) FD: 232 BD: 1 +.+.: (wq_completion)bond776 ->&rq->__lock ->(work_completion)(&(&slave->notify_work)->work) ->(work_completion)(&(&bond->mcast_work)->work) FD: 232 BD: 1 +.+.: (wq_completion)bond756#3 ->(work_completion)(&(&slave->notify_work)->work) ->(work_completion)(&(&bond->mcast_work)->work) ->&rq->__lock FD: 232 BD: 1 +.+.: (wq_completion)bond769#2 ->(work_completion)(&(&slave->notify_work)->work) ->(work_completion)(&(&bond->mcast_work)->work) ->&rq->__lock FD: 232 BD: 1 +.+.: (wq_completion)bond777 ->(work_completion)(&(&slave->notify_work)->work) ->(work_completion)(&(&bond->mcast_work)->work) ->&rq->__lock ->&cfs_rq->removed.lock ->&obj_hash[i].lock FD: 232 BD: 1 +.+.: (wq_completion)bond757#3 ->(work_completion)(&(&slave->notify_work)->work) ->(work_completion)(&(&bond->mcast_work)->work) ->&rq->__lock FD: 232 BD: 1 +.+.: (wq_completion)bond770#2 ->&rq->__lock ->(work_completion)(&(&slave->notify_work)->work) ->(work_completion)(&(&bond->mcast_work)->work) FD: 232 BD: 1 +.+.: (wq_completion)bond778 ->(work_completion)(&(&slave->notify_work)->work) ->(work_completion)(&(&bond->mcast_work)->work) ->&rq->__lock FD: 232 BD: 1 +.+.: (wq_completion)bond758#3 ->(work_completion)(&(&slave->notify_work)->work) ->(work_completion)(&(&bond->mcast_work)->work) FD: 232 BD: 1 +.+.: (wq_completion)bond771#2 ->&rq->__lock ->(work_completion)(&(&slave->notify_work)->work) ->(work_completion)(&(&bond->mcast_work)->work) FD: 232 BD: 1 +.+.: (wq_completion)bond779 ->(work_completion)(&(&slave->notify_work)->work) ->(work_completion)(&(&bond->mcast_work)->work) ->&rq->__lock FD: 232 BD: 1 +.+.: (wq_completion)bond759#2 ->(work_completion)(&(&slave->notify_work)->work) ->(work_completion)(&(&bond->mcast_work)->work) ->&rq->__lock FD: 232 BD: 1 +.+.: (wq_completion)bond772#2 ->(work_completion)(&(&slave->notify_work)->work) ->(work_completion)(&(&bond->mcast_work)->work) FD: 232 BD: 1 +.+.: (wq_completion)bond780 ->(work_completion)(&(&slave->notify_work)->work) ->(work_completion)(&(&bond->mcast_work)->work) ->&rq->__lock FD: 232 BD: 1 +.+.: (wq_completion)bond760#3 ->&rq->__lock ->(work_completion)(&(&slave->notify_work)->work) ->(work_completion)(&(&bond->mcast_work)->work) FD: 232 BD: 1 +.+.: (wq_completion)bond773#2 ->&rq->__lock ->(work_completion)(&(&slave->notify_work)->work) ->(work_completion)(&(&bond->mcast_work)->work) FD: 232 BD: 1 +.+.: (wq_completion)bond781 ->&rq->__lock ->(work_completion)(&(&slave->notify_work)->work) ->(work_completion)(&(&bond->mcast_work)->work) FD: 232 BD: 1 +.+.: (wq_completion)bond774#2 ->&rq->__lock ->(work_completion)(&(&slave->notify_work)->work) ->(work_completion)(&(&bond->mcast_work)->work) FD: 232 BD: 1 +.+.: (wq_completion)bond761#3 ->&rq->__lock ->(work_completion)(&(&slave->notify_work)->work) ->(work_completion)(&(&bond->mcast_work)->work) FD: 232 BD: 1 +.+.: (wq_completion)bond782 ->(work_completion)(&(&slave->notify_work)->work) ->(work_completion)(&(&bond->mcast_work)->work) ->&rq->__lock FD: 232 BD: 1 +.+.: (wq_completion)bond775#2 ->(work_completion)(&(&slave->notify_work)->work) ->(work_completion)(&(&bond->mcast_work)->work) ->&rq->__lock FD: 232 BD: 1 +.+.: (wq_completion)bond762#3 ->(work_completion)(&(&slave->notify_work)->work) ->(work_completion)(&(&bond->mcast_work)->work) FD: 232 BD: 1 +.+.: (wq_completion)bond776#2 ->&rq->__lock ->(work_completion)(&(&slave->notify_work)->work) ->(work_completion)(&(&bond->mcast_work)->work) FD: 232 BD: 1 +.+.: (wq_completion)bond783 ->(work_completion)(&(&slave->notify_work)->work) ->(work_completion)(&(&bond->mcast_work)->work) ->&rq->__lock FD: 232 BD: 1 +.+.: (wq_completion)bond763#3 ->(work_completion)(&(&slave->notify_work)->work) ->(work_completion)(&(&bond->mcast_work)->work) ->&rq->__lock FD: 232 BD: 1 +.+.: (wq_completion)bond784 ->&rq->__lock ->(work_completion)(&(&slave->notify_work)->work) ->(work_completion)(&(&bond->mcast_work)->work) FD: 232 BD: 1 +.+.: (wq_completion)bond777#2 ->&rq->__lock ->(work_completion)(&(&slave->notify_work)->work) ->(work_completion)(&(&bond->mcast_work)->work) FD: 232 BD: 1 +.+.: (wq_completion)bond764#3 ->&rq->__lock ->(work_completion)(&(&slave->notify_work)->work) ->(work_completion)(&(&bond->mcast_work)->work) FD: 232 BD: 1 +.+.: (wq_completion)bond785 ->(work_completion)(&(&slave->notify_work)->work) ->(work_completion)(&(&bond->mcast_work)->work) ->&rq->__lock FD: 232 BD: 1 +.+.: (wq_completion)bond778#2 ->&rq->__lock ->(work_completion)(&(&slave->notify_work)->work) ->(work_completion)(&(&bond->mcast_work)->work) FD: 232 BD: 1 +.+.: (wq_completion)bond765#3 ->(work_completion)(&(&slave->notify_work)->work) ->(work_completion)(&(&bond->mcast_work)->work) ->&rq->__lock FD: 232 BD: 1 +.+.: (wq_completion)bond786 ->(work_completion)(&(&slave->notify_work)->work) ->(work_completion)(&(&bond->mcast_work)->work) ->&rq->__lock FD: 232 BD: 1 +.+.: (wq_completion)bond766#3 ->(work_completion)(&(&slave->notify_work)->work) ->(work_completion)(&(&bond->mcast_work)->work) ->&rq->__lock FD: 232 BD: 1 +.+.: (wq_completion)bond779#2 ->&rq->__lock ->(work_completion)(&(&slave->notify_work)->work) ->(work_completion)(&(&bond->mcast_work)->work) FD: 232 BD: 1 +.+.: (wq_completion)bond787 ->(work_completion)(&(&slave->notify_work)->work) ->(work_completion)(&(&bond->mcast_work)->work) FD: 232 BD: 1 +.+.: (wq_completion)bond767#3 ->&rq->__lock ->(work_completion)(&(&slave->notify_work)->work) ->(work_completion)(&(&bond->mcast_work)->work) FD: 232 BD: 1 +.+.: (wq_completion)bond788 ->(work_completion)(&(&slave->notify_work)->work) ->(work_completion)(&(&bond->mcast_work)->work) FD: 232 BD: 1 +.+.: (wq_completion)bond781#2 ->(work_completion)(&(&slave->notify_work)->work) ->(work_completion)(&(&bond->mcast_work)->work) ->&rq->__lock FD: 232 BD: 1 +.+.: (wq_completion)bond768#3 ->&rq->__lock ->(work_completion)(&(&slave->notify_work)->work) ->(work_completion)(&(&bond->mcast_work)->work) FD: 232 BD: 1 +.+.: (wq_completion)bond789 ->(work_completion)(&(&slave->notify_work)->work) ->(work_completion)(&(&bond->mcast_work)->work) ->&rq->__lock FD: 232 BD: 1 +.+.: (wq_completion)bond782#2 ->(work_completion)(&(&slave->notify_work)->work) ->(work_completion)(&(&bond->mcast_work)->work) ->&rq->__lock FD: 232 BD: 1 +.+.: (wq_completion)bond769#3 ->(work_completion)(&(&slave->notify_work)->work) ->(work_completion)(&(&bond->mcast_work)->work) ->&rq->__lock FD: 232 BD: 1 +.+.: (wq_completion)bond790 ->&rq->__lock ->(work_completion)(&(&slave->notify_work)->work) ->(work_completion)(&(&bond->mcast_work)->work) FD: 232 BD: 1 +.+.: (wq_completion)bond783#2 ->(work_completion)(&(&slave->notify_work)->work) ->(work_completion)(&(&bond->mcast_work)->work) ->&rq->__lock FD: 232 BD: 1 +.+.: (wq_completion)bond770#3 ->&rq->__lock ->(work_completion)(&(&slave->notify_work)->work) ->(work_completion)(&(&bond->mcast_work)->work) FD: 232 BD: 1 +.+.: (wq_completion)bond791 ->&rq->__lock ->(work_completion)(&(&slave->notify_work)->work) ->(work_completion)(&(&bond->mcast_work)->work) FD: 232 BD: 1 +.+.: (wq_completion)bond784#2 ->&rq->__lock ->(work_completion)(&(&slave->notify_work)->work) ->(work_completion)(&(&bond->mcast_work)->work) FD: 232 BD: 1 +.+.: (wq_completion)bond771#3 ->&rq->__lock ->(work_completion)(&(&slave->notify_work)->work) ->(work_completion)(&(&bond->mcast_work)->work) FD: 232 BD: 1 +.+.: (wq_completion)bond792 ->&rq->__lock ->(work_completion)(&(&slave->notify_work)->work) ->(work_completion)(&(&bond->mcast_work)->work) FD: 232 BD: 1 +.+.: (wq_completion)bond785#2 ->(work_completion)(&(&slave->notify_work)->work) ->(work_completion)(&(&bond->mcast_work)->work) ->&rq->__lock FD: 232 BD: 1 +.+.: (wq_completion)bond772#3 ->&rq->__lock ->(work_completion)(&(&slave->notify_work)->work) ->(work_completion)(&(&bond->mcast_work)->work) FD: 232 BD: 1 +.+.: (wq_completion)bond793 ->&rq->__lock ->(work_completion)(&(&slave->notify_work)->work) ->(work_completion)(&(&bond->mcast_work)->work) FD: 232 BD: 1 +.+.: (wq_completion)bond786#2 ->&rq->__lock ->(work_completion)(&(&slave->notify_work)->work) ->(work_completion)(&(&bond->mcast_work)->work) FD: 232 BD: 1 +.+.: (wq_completion)bond773#3 ->&rq->__lock ->(work_completion)(&(&slave->notify_work)->work) ->(work_completion)(&(&bond->mcast_work)->work) FD: 232 BD: 1 +.+.: (wq_completion)bond794 ->&rq->__lock ->(work_completion)(&(&slave->notify_work)->work) ->(work_completion)(&(&bond->mcast_work)->work) FD: 232 BD: 1 +.+.: (wq_completion)bond787#2 ->(work_completion)(&(&slave->notify_work)->work) ->(work_completion)(&(&bond->mcast_work)->work) ->&rq->__lock FD: 232 BD: 1 +.+.: (wq_completion)bond774#3 ->(work_completion)(&(&slave->notify_work)->work) ->(work_completion)(&(&bond->mcast_work)->work) ->&rq->__lock FD: 232 BD: 1 +.+.: (wq_completion)bond795 ->&rq->__lock ->(work_completion)(&(&slave->notify_work)->work) ->(work_completion)(&(&bond->mcast_work)->work) FD: 232 BD: 1 +.+.: (wq_completion)bond788#2 ->&rq->__lock ->(work_completion)(&(&slave->notify_work)->work) ->(work_completion)(&(&bond->mcast_work)->work) FD: 232 BD: 1 +.+.: (wq_completion)bond796 ->(work_completion)(&(&slave->notify_work)->work) ->(work_completion)(&(&bond->mcast_work)->work) ->&rq->__lock FD: 232 BD: 1 +.+.: (wq_completion)bond776#3 ->&rq->__lock ->(work_completion)(&(&slave->notify_work)->work) ->(work_completion)(&(&bond->mcast_work)->work) FD: 232 BD: 1 +.+.: (wq_completion)bond789#2 ->(work_completion)(&(&slave->notify_work)->work) ->(work_completion)(&(&bond->mcast_work)->work) FD: 232 BD: 1 +.+.: (wq_completion)bond797 ->&rq->__lock ->(work_completion)(&(&slave->notify_work)->work) ->(work_completion)(&(&bond->mcast_work)->work) FD: 232 BD: 1 +.+.: (wq_completion)bond790#2 ->&rq->__lock ->(work_completion)(&(&slave->notify_work)->work) ->(work_completion)(&(&bond->mcast_work)->work) FD: 232 BD: 1 +.+.: (wq_completion)bond777#3 ->(work_completion)(&(&slave->notify_work)->work) ->(work_completion)(&(&bond->mcast_work)->work) ->&rq->__lock FD: 232 BD: 1 +.+.: (wq_completion)bond798 ->&rq->__lock ->(work_completion)(&(&slave->notify_work)->work) ->(work_completion)(&(&bond->mcast_work)->work) FD: 232 BD: 1 +.+.: (wq_completion)bond791#2 ->&rq->__lock ->(work_completion)(&(&slave->notify_work)->work) ->(work_completion)(&(&bond->mcast_work)->work) FD: 232 BD: 1 +.+.: (wq_completion)bond778#3 ->(work_completion)(&(&slave->notify_work)->work) ->(work_completion)(&(&bond->mcast_work)->work) ->&rq->__lock FD: 232 BD: 1 +.+.: (wq_completion)bond799 ->&rq->__lock ->(work_completion)(&(&slave->notify_work)->work) ->(work_completion)(&(&bond->mcast_work)->work) FD: 232 BD: 1 +.+.: (wq_completion)bond792#2 ->&rq->__lock ->(work_completion)(&(&slave->notify_work)->work) ->(work_completion)(&(&bond->mcast_work)->work) FD: 232 BD: 1 +.+.: (wq_completion)bond779#3 ->&rq->__lock ->(work_completion)(&(&slave->notify_work)->work) ->(work_completion)(&(&bond->mcast_work)->work) FD: 232 BD: 1 +.+.: (wq_completion)bond800 ->&rq->__lock ->(work_completion)(&(&slave->notify_work)->work) ->(work_completion)(&(&bond->mcast_work)->work) FD: 232 BD: 1 +.+.: (wq_completion)bond793#2 ->&rq->__lock ->(work_completion)(&(&slave->notify_work)->work) ->(work_completion)(&(&bond->mcast_work)->work) FD: 232 BD: 1 +.+.: (wq_completion)bond780#2 ->&rq->__lock ->(work_completion)(&(&slave->notify_work)->work) ->(work_completion)(&(&bond->mcast_work)->work) ->&cfs_rq->removed.lock ->&obj_hash[i].lock FD: 232 BD: 1 +.+.: (wq_completion)bond801 ->(work_completion)(&(&slave->notify_work)->work) ->(work_completion)(&(&bond->mcast_work)->work) ->&rq->__lock FD: 1 BD: 3837 ..-.: key#29 FD: 232 BD: 1 +.+.: (wq_completion)bond794#2 ->&rq->__lock ->(work_completion)(&(&slave->notify_work)->work) ->(work_completion)(&(&bond->mcast_work)->work) FD: 232 BD: 1 +.+.: (wq_completion)bond781#3 ->(work_completion)(&(&slave->notify_work)->work) ->(work_completion)(&(&bond->mcast_work)->work) ->&rq->__lock FD: 232 BD: 1 +.+.: (wq_completion)bond802 ->(work_completion)(&(&slave->notify_work)->work) ->(work_completion)(&(&bond->mcast_work)->work) ->&rq->__lock FD: 232 BD: 1 +.+.: (wq_completion)bond795#2 ->&rq->__lock ->(work_completion)(&(&slave->notify_work)->work) ->(work_completion)(&(&bond->mcast_work)->work) FD: 232 BD: 1 +.+.: (wq_completion)bond782#3 ->&rq->__lock ->(work_completion)(&(&slave->notify_work)->work) ->(work_completion)(&(&bond->mcast_work)->work) FD: 232 BD: 1 +.+.: (wq_completion)bond803 ->(work_completion)(&(&slave->notify_work)->work) ->(work_completion)(&(&bond->mcast_work)->work) ->&rq->__lock FD: 232 BD: 1 +.+.: (wq_completion)bond796#2 ->(work_completion)(&(&slave->notify_work)->work) ->(work_completion)(&(&bond->mcast_work)->work) ->&rq->__lock FD: 232 BD: 1 +.+.: (wq_completion)bond783#3 ->(work_completion)(&(&slave->notify_work)->work) ->(work_completion)(&(&bond->mcast_work)->work) ->&rq->__lock FD: 232 BD: 1 +.+.: (wq_completion)bond804 ->&rq->__lock ->(work_completion)(&(&slave->notify_work)->work) ->(work_completion)(&(&bond->mcast_work)->work) ->&cfs_rq->removed.lock ->&obj_hash[i].lock FD: 232 BD: 1 +.+.: (wq_completion)bond797#2 ->&rq->__lock ->(work_completion)(&(&slave->notify_work)->work) ->(work_completion)(&(&bond->mcast_work)->work) FD: 232 BD: 1 +.+.: (wq_completion)bond784#3 ->(work_completion)(&(&slave->notify_work)->work) ->(work_completion)(&(&bond->mcast_work)->work) ->&rq->__lock FD: 232 BD: 1 +.+.: (wq_completion)bond805 ->&rq->__lock ->(work_completion)(&(&slave->notify_work)->work) ->(work_completion)(&(&bond->mcast_work)->work) FD: 232 BD: 1 +.+.: (wq_completion)bond798#2 ->&rq->__lock ->(work_completion)(&(&slave->notify_work)->work) ->(work_completion)(&(&bond->mcast_work)->work) FD: 232 BD: 1 +.+.: (wq_completion)bond785#3 ->&rq->__lock ->(work_completion)(&(&slave->notify_work)->work) ->(work_completion)(&(&bond->mcast_work)->work) FD: 232 BD: 1 +.+.: (wq_completion)bond806 ->(work_completion)(&(&slave->notify_work)->work) ->(work_completion)(&(&bond->mcast_work)->work) ->&rq->__lock FD: 232 BD: 1 +.+.: (wq_completion)bond786#3 ->(work_completion)(&(&slave->notify_work)->work) ->(work_completion)(&(&bond->mcast_work)->work) ->&rq->__lock FD: 232 BD: 1 +.+.: (wq_completion)bond799#2 ->(work_completion)(&(&slave->notify_work)->work) ->(work_completion)(&(&bond->mcast_work)->work) ->&rq->__lock FD: 232 BD: 1 +.+.: (wq_completion)bond807 ->&rq->__lock ->(work_completion)(&(&slave->notify_work)->work) ->(work_completion)(&(&bond->mcast_work)->work) FD: 232 BD: 1 +.+.: (wq_completion)bond787#3 ->&rq->__lock ->(work_completion)(&(&slave->notify_work)->work) ->(work_completion)(&(&bond->mcast_work)->work) FD: 232 BD: 1 +.+.: (wq_completion)bond800#2 ->&rq->__lock ->(work_completion)(&(&slave->notify_work)->work) ->(work_completion)(&(&bond->mcast_work)->work) FD: 232 BD: 1 +.+.: (wq_completion)bond808 ->(work_completion)(&(&slave->notify_work)->work) ->(work_completion)(&(&bond->mcast_work)->work) ->&rq->__lock FD: 232 BD: 1 +.+.: (wq_completion)bond788#3 ->(work_completion)(&(&slave->notify_work)->work) ->(work_completion)(&(&bond->mcast_work)->work) ->&rq->__lock FD: 232 BD: 1 +.+.: (wq_completion)bond801#2 ->&rq->__lock ->(work_completion)(&(&slave->notify_work)->work) ->(work_completion)(&(&bond->mcast_work)->work) FD: 232 BD: 1 +.+.: (wq_completion)bond809 ->&rq->__lock ->(work_completion)(&(&slave->notify_work)->work) ->(work_completion)(&(&bond->mcast_work)->work) FD: 232 BD: 1 +.+.: (wq_completion)bond789#3 ->(work_completion)(&(&slave->notify_work)->work) ->(work_completion)(&(&bond->mcast_work)->work) ->&rq->__lock FD: 1 BD: 4563 ....: key#30 FD: 232 BD: 1 +.+.: (wq_completion)bond802#2 ->(work_completion)(&(&slave->notify_work)->work) ->(work_completion)(&(&bond->mcast_work)->work) ->&rq->__lock FD: 232 BD: 1 +.+.: (wq_completion)bond810 ->&rq->__lock ->(work_completion)(&(&slave->notify_work)->work) ->(work_completion)(&(&bond->mcast_work)->work) FD: 232 BD: 1 +.+.: (wq_completion)bond790#3 ->&rq->__lock ->(work_completion)(&(&slave->notify_work)->work) ->(work_completion)(&(&bond->mcast_work)->work) FD: 1 BD: 4518 +.+.: f2fs_list_lock FD: 232 BD: 1 +.+.: (wq_completion)bond803#2 ->&rq->__lock ->(work_completion)(&(&slave->notify_work)->work) ->(work_completion)(&(&bond->mcast_work)->work) FD: 74 BD: 1 .+.+: &type->s_umount_key#52 ->&lru->node[i].lock ->&rq->__lock ->&dentry->d_lock ->&sb->s_type->i_lock_key#24 ->&s->s_inode_list_lock ->&xa->xa_lock#9 ->inode_hash_lock ->&obj_hash[i].lock ->pool_lock#2 ->&fsnotify_mark_srcu ->&wb->list_lock ->kernfs_idr_lock ->rcu_node_0 ->&rcu_state.expedited_wq FD: 232 BD: 1 +.+.: (wq_completion)bond811 ->&rq->__lock ->(work_completion)(&(&slave->notify_work)->work) ->(work_completion)(&(&bond->mcast_work)->work) FD: 232 BD: 1 +.+.: (wq_completion)bond791#3 ->(work_completion)(&(&slave->notify_work)->work) ->(work_completion)(&(&bond->mcast_work)->work) ->&rq->__lock FD: 232 BD: 1 +.+.: (wq_completion)bond804#2 ->(work_completion)(&(&slave->notify_work)->work) ->(work_completion)(&(&bond->mcast_work)->work) ->&rq->__lock FD: 232 BD: 1 +.+.: (wq_completion)bond812 ->&rq->__lock ->(work_completion)(&(&slave->notify_work)->work) ->(work_completion)(&(&bond->mcast_work)->work) FD: 232 BD: 1 +.+.: (wq_completion)bond792#3 ->(work_completion)(&(&slave->notify_work)->work) ->(work_completion)(&(&bond->mcast_work)->work) ->&rq->__lock FD: 232 BD: 1 +.+.: (wq_completion)bond805#2 ->&rq->__lock ->(work_completion)(&(&slave->notify_work)->work) ->(work_completion)(&(&bond->mcast_work)->work) FD: 232 BD: 1 +.+.: (wq_completion)bond813 ->&rq->__lock ->(work_completion)(&(&slave->notify_work)->work) ->(work_completion)(&(&bond->mcast_work)->work) FD: 232 BD: 1 +.+.: (wq_completion)bond793#3 ->(work_completion)(&(&slave->notify_work)->work) ->(work_completion)(&(&bond->mcast_work)->work) ->&rq->__lock FD: 232 BD: 1 +.+.: (wq_completion)bond806#2 ->&rq->__lock ->(work_completion)(&(&slave->notify_work)->work) ->(work_completion)(&(&bond->mcast_work)->work) FD: 232 BD: 1 +.+.: (wq_completion)bond814 ->(work_completion)(&(&slave->notify_work)->work) ->(work_completion)(&(&bond->mcast_work)->work) ->&rq->__lock FD: 232 BD: 1 +.+.: (wq_completion)bond794#3 ->&rq->__lock ->(work_completion)(&(&slave->notify_work)->work) ->(work_completion)(&(&bond->mcast_work)->work) FD: 232 BD: 1 +.+.: (wq_completion)bond807#2 ->(work_completion)(&(&slave->notify_work)->work) ->(work_completion)(&(&bond->mcast_work)->work) ->&rq->__lock FD: 232 BD: 1 +.+.: (wq_completion)bond815 ->(work_completion)(&(&slave->notify_work)->work) ->(work_completion)(&(&bond->mcast_work)->work) ->&rq->__lock FD: 232 BD: 1 +.+.: (wq_completion)bond795#3 ->&rq->__lock ->(work_completion)(&(&slave->notify_work)->work) ->(work_completion)(&(&bond->mcast_work)->work) FD: 232 BD: 1 +.+.: (wq_completion)bond808#2 ->(work_completion)(&(&slave->notify_work)->work) ->(work_completion)(&(&bond->mcast_work)->work) ->&rq->__lock FD: 232 BD: 1 +.+.: (wq_completion)bond816 ->&rq->__lock ->(work_completion)(&(&slave->notify_work)->work) ->(work_completion)(&(&bond->mcast_work)->work) FD: 232 BD: 1 +.+.: (wq_completion)bond796#3 ->(work_completion)(&(&slave->notify_work)->work) ->(work_completion)(&(&bond->mcast_work)->work) ->&rq->__lock FD: 232 BD: 1 +.+.: (wq_completion)bond809#2 ->&rq->__lock ->(work_completion)(&(&slave->notify_work)->work) ->(work_completion)(&(&bond->mcast_work)->work) FD: 232 BD: 1 +.+.: (wq_completion)bond817 ->&rq->__lock ->(work_completion)(&(&slave->notify_work)->work) ->(work_completion)(&(&bond->mcast_work)->work) FD: 232 BD: 1 +.+.: (wq_completion)bond797#3 ->(work_completion)(&(&slave->notify_work)->work) ->(work_completion)(&(&bond->mcast_work)->work) ->&rq->__lock FD: 232 BD: 1 +.+.: (wq_completion)bond810#2 ->&rq->__lock ->(work_completion)(&(&slave->notify_work)->work) ->(work_completion)(&(&bond->mcast_work)->work) FD: 232 BD: 1 +.+.: (wq_completion)bond818 ->&rq->__lock ->(work_completion)(&(&slave->notify_work)->work) ->(work_completion)(&(&bond->mcast_work)->work) FD: 232 BD: 1 +.+.: (wq_completion)bond798#3 ->(work_completion)(&(&slave->notify_work)->work) ->(work_completion)(&(&bond->mcast_work)->work) ->&rq->__lock FD: 232 BD: 1 +.+.: (wq_completion)bond811#2 ->&rq->__lock ->(work_completion)(&(&slave->notify_work)->work) ->(work_completion)(&(&bond->mcast_work)->work) FD: 232 BD: 1 +.+.: (wq_completion)bond819 ->(work_completion)(&(&slave->notify_work)->work) ->(work_completion)(&(&bond->mcast_work)->work) ->&rq->__lock ->rcu_node_0 FD: 232 BD: 1 +.+.: (wq_completion)bond799#3 ->&rq->__lock ->(work_completion)(&(&slave->notify_work)->work) ->(work_completion)(&(&bond->mcast_work)->work) FD: 232 BD: 1 +.+.: (wq_completion)bond812#2 ->&rq->__lock ->(work_completion)(&(&slave->notify_work)->work) ->(work_completion)(&(&bond->mcast_work)->work) FD: 232 BD: 1 +.+.: (wq_completion)bond820 ->&rq->__lock ->(work_completion)(&(&slave->notify_work)->work) ->(work_completion)(&(&bond->mcast_work)->work) FD: 232 BD: 1 +.+.: (wq_completion)bond800#3 ->&rq->__lock ->(work_completion)(&(&slave->notify_work)->work) ->(work_completion)(&(&bond->mcast_work)->work) FD: 232 BD: 1 +.+.: (wq_completion)bond813#2 ->&rq->__lock ->(work_completion)(&(&slave->notify_work)->work) ->(work_completion)(&(&bond->mcast_work)->work) FD: 232 BD: 1 +.+.: (wq_completion)bond821 ->(work_completion)(&(&slave->notify_work)->work) ->(work_completion)(&(&bond->mcast_work)->work) FD: 232 BD: 1 +.+.: (wq_completion)bond801#3 ->&rq->__lock ->(work_completion)(&(&slave->notify_work)->work) ->(work_completion)(&(&bond->mcast_work)->work) FD: 232 BD: 1 +.+.: (wq_completion)bond814#2 ->(work_completion)(&(&slave->notify_work)->work) ->(work_completion)(&(&bond->mcast_work)->work) ->&rq->__lock FD: 232 BD: 1 +.+.: (wq_completion)bond822 ->&rq->__lock ->(work_completion)(&(&slave->notify_work)->work) ->(work_completion)(&(&bond->mcast_work)->work) FD: 232 BD: 1 +.+.: (wq_completion)bond802#3 ->(work_completion)(&(&slave->notify_work)->work) ->(work_completion)(&(&bond->mcast_work)->work) ->&rq->__lock FD: 232 BD: 1 +.+.: (wq_completion)bond815#2 ->&rq->__lock ->(work_completion)(&(&slave->notify_work)->work) ->(work_completion)(&(&bond->mcast_work)->work) FD: 232 BD: 1 +.+.: (wq_completion)bond823 ->(work_completion)(&(&slave->notify_work)->work) ->(work_completion)(&(&bond->mcast_work)->work) FD: 232 BD: 1 +.+.: (wq_completion)bond803#3 ->&rq->__lock ->(work_completion)(&(&slave->notify_work)->work) ->(work_completion)(&(&bond->mcast_work)->work) FD: 232 BD: 1 +.+.: (wq_completion)bond816#2 ->(work_completion)(&(&slave->notify_work)->work) ->(work_completion)(&(&bond->mcast_work)->work) ->&rq->__lock FD: 232 BD: 1 +.+.: (wq_completion)bond824 ->(work_completion)(&(&slave->notify_work)->work) ->(work_completion)(&(&bond->mcast_work)->work) ->&rq->__lock FD: 232 BD: 1 +.+.: (wq_completion)bond804#3 ->(work_completion)(&(&slave->notify_work)->work) ->(work_completion)(&(&bond->mcast_work)->work) FD: 232 BD: 1 +.+.: (wq_completion)bond817#2 ->(work_completion)(&(&slave->notify_work)->work) ->(work_completion)(&(&bond->mcast_work)->work) ->&rq->__lock FD: 232 BD: 1 +.+.: (wq_completion)bond825 ->&rq->__lock ->(work_completion)(&(&slave->notify_work)->work) ->(work_completion)(&(&bond->mcast_work)->work) FD: 232 BD: 1 +.+.: (wq_completion)bond805#3 ->(work_completion)(&(&slave->notify_work)->work) ->(work_completion)(&(&bond->mcast_work)->work) ->&rq->__lock FD: 232 BD: 1 +.+.: (wq_completion)bond818#2 ->&rq->__lock ->(work_completion)(&(&slave->notify_work)->work) ->(work_completion)(&(&bond->mcast_work)->work) FD: 232 BD: 1 +.+.: (wq_completion)bond826 ->&rq->__lock ->(work_completion)(&(&slave->notify_work)->work) ->(work_completion)(&(&bond->mcast_work)->work) FD: 232 BD: 1 +.+.: (wq_completion)bond806#3 ->(work_completion)(&(&slave->notify_work)->work) ->(work_completion)(&(&bond->mcast_work)->work) ->&rq->__lock FD: 232 BD: 1 +.+.: (wq_completion)bond819#2 ->(work_completion)(&(&slave->notify_work)->work) ->(work_completion)(&(&bond->mcast_work)->work) FD: 232 BD: 1 +.+.: (wq_completion)bond827 ->&rq->__lock ->(work_completion)(&(&slave->notify_work)->work) ->(work_completion)(&(&bond->mcast_work)->work) FD: 232 BD: 1 +.+.: (wq_completion)bond807#3 ->&rq->__lock ->(work_completion)(&(&slave->notify_work)->work) ->(work_completion)(&(&bond->mcast_work)->work) FD: 232 BD: 1 +.+.: (wq_completion)bond828 ->&rq->__lock ->(work_completion)(&(&slave->notify_work)->work) ->(work_completion)(&(&bond->mcast_work)->work) FD: 232 BD: 1 +.+.: (wq_completion)bond808#3 ->&rq->__lock ->(work_completion)(&(&slave->notify_work)->work) ->(work_completion)(&(&bond->mcast_work)->work) FD: 1 BD: 170 ....: key#31 FD: 232 BD: 1 +.+.: (wq_completion)bond820#2 ->&rq->__lock ->(work_completion)(&(&slave->notify_work)->work) ->(work_completion)(&(&bond->mcast_work)->work) FD: 232 BD: 1 +.+.: (wq_completion)bond829 ->&rq->__lock ->(work_completion)(&(&slave->notify_work)->work) ->(work_completion)(&(&bond->mcast_work)->work) FD: 232 BD: 1 +.+.: (wq_completion)bond809#3 ->(work_completion)(&(&slave->notify_work)->work) ->(work_completion)(&(&bond->mcast_work)->work) FD: 232 BD: 1 +.+.: (wq_completion)bond821#2 ->(work_completion)(&(&slave->notify_work)->work) ->(work_completion)(&(&bond->mcast_work)->work) ->&rq->__lock FD: 232 BD: 1 +.+.: (wq_completion)bond830 ->(work_completion)(&(&slave->notify_work)->work) ->(work_completion)(&(&bond->mcast_work)->work) ->&rq->__lock FD: 232 BD: 1 +.+.: (wq_completion)bond810#3 ->&rq->__lock ->(work_completion)(&(&slave->notify_work)->work) ->(work_completion)(&(&bond->mcast_work)->work) FD: 232 BD: 1 +.+.: (wq_completion)bond822#2 ->(work_completion)(&(&slave->notify_work)->work) ->(work_completion)(&(&bond->mcast_work)->work) ->&rq->__lock FD: 232 BD: 1 +.+.: (wq_completion)bond831 ->&rq->__lock ->(work_completion)(&(&slave->notify_work)->work) ->(work_completion)(&(&bond->mcast_work)->work) FD: 232 BD: 1 +.+.: (wq_completion)bond811#3 ->&rq->__lock ->(work_completion)(&(&slave->notify_work)->work) ->(work_completion)(&(&bond->mcast_work)->work) FD: 232 BD: 1 +.+.: (wq_completion)bond823#2 ->(work_completion)(&(&slave->notify_work)->work) ->(work_completion)(&(&bond->mcast_work)->work) ->&rq->__lock FD: 232 BD: 1 +.+.: (wq_completion)bond832 ->&rq->__lock ->(work_completion)(&(&slave->notify_work)->work) ->(work_completion)(&(&bond->mcast_work)->work) FD: 232 BD: 1 +.+.: (wq_completion)bond812#3 ->&rq->__lock ->(work_completion)(&(&slave->notify_work)->work) ->(work_completion)(&(&bond->mcast_work)->work) FD: 232 BD: 1 +.+.: (wq_completion)bond824#2 ->(work_completion)(&(&slave->notify_work)->work) ->(work_completion)(&(&bond->mcast_work)->work) ->&rq->__lock FD: 232 BD: 1 +.+.: (wq_completion)bond833 ->&rq->__lock ->(work_completion)(&(&slave->notify_work)->work) ->(work_completion)(&(&bond->mcast_work)->work) FD: 232 BD: 1 +.+.: (wq_completion)bond813#3 ->(work_completion)(&(&slave->notify_work)->work) ->(work_completion)(&(&bond->mcast_work)->work) ->&rq->__lock FD: 232 BD: 1 +.+.: (wq_completion)bond825#2 ->&rq->__lock ->(work_completion)(&(&slave->notify_work)->work) ->(work_completion)(&(&bond->mcast_work)->work) FD: 232 BD: 1 +.+.: (wq_completion)bond834 ->(work_completion)(&(&slave->notify_work)->work) ->&rq->__lock ->(work_completion)(&(&bond->mcast_work)->work) FD: 232 BD: 1 +.+.: (wq_completion)bond826#2 ->&rq->__lock ->(work_completion)(&(&slave->notify_work)->work) ->(work_completion)(&(&bond->mcast_work)->work) FD: 232 BD: 1 +.+.: (wq_completion)bond814#3 ->&rq->__lock ->(work_completion)(&(&slave->notify_work)->work) ->(work_completion)(&(&bond->mcast_work)->work) FD: 232 BD: 1 +.+.: (wq_completion)bond827#2 ->(work_completion)(&(&slave->notify_work)->work) ->(work_completion)(&(&bond->mcast_work)->work) ->&rq->__lock FD: 232 BD: 1 +.+.: (wq_completion)bond835 ->(work_completion)(&(&slave->notify_work)->work) ->(work_completion)(&(&bond->mcast_work)->work) ->&rq->__lock FD: 232 BD: 1 +.+.: (wq_completion)bond815#3 ->(work_completion)(&(&slave->notify_work)->work) ->(work_completion)(&(&bond->mcast_work)->work) ->&rq->__lock FD: 232 BD: 1 +.+.: (wq_completion)bond828#2 ->(work_completion)(&(&slave->notify_work)->work) ->(work_completion)(&(&bond->mcast_work)->work) ->&rq->__lock FD: 232 BD: 1 +.+.: (wq_completion)bond836 ->(work_completion)(&(&slave->notify_work)->work) ->(work_completion)(&(&bond->mcast_work)->work) ->&rq->__lock FD: 232 BD: 1 +.+.: (wq_completion)bond816#3 ->(work_completion)(&(&slave->notify_work)->work) ->(work_completion)(&(&bond->mcast_work)->work) ->&rq->__lock FD: 232 BD: 1 +.+.: (wq_completion)bond829#2 ->&rq->__lock ->(work_completion)(&(&slave->notify_work)->work) ->(work_completion)(&(&bond->mcast_work)->work) FD: 232 BD: 1 +.+.: (wq_completion)bond837 ->(work_completion)(&(&slave->notify_work)->work) ->(work_completion)(&(&bond->mcast_work)->work) ->&rq->__lock FD: 232 BD: 1 +.+.: (wq_completion)bond817#3 ->(work_completion)(&(&slave->notify_work)->work) ->(work_completion)(&(&bond->mcast_work)->work) ->&rq->__lock FD: 232 BD: 1 +.+.: (wq_completion)bond830#2 ->&rq->__lock ->(work_completion)(&(&slave->notify_work)->work) ->(work_completion)(&(&bond->mcast_work)->work) FD: 232 BD: 1 +.+.: (wq_completion)bond838 ->&rq->__lock ->(work_completion)(&(&slave->notify_work)->work) ->(work_completion)(&(&bond->mcast_work)->work) FD: 232 BD: 1 +.+.: (wq_completion)bond818#3 ->(work_completion)(&(&slave->notify_work)->work) ->(work_completion)(&(&bond->mcast_work)->work) FD: 232 BD: 1 +.+.: (wq_completion)bond831#2 ->&rq->__lock ->(work_completion)(&(&slave->notify_work)->work) ->(work_completion)(&(&bond->mcast_work)->work) FD: 232 BD: 1 +.+.: (wq_completion)bond839 ->&rq->__lock ->(work_completion)(&(&slave->notify_work)->work) ->(work_completion)(&(&bond->mcast_work)->work) ->&cfs_rq->removed.lock ->&obj_hash[i].lock FD: 232 BD: 1 +.+.: (wq_completion)bond819#3 ->(work_completion)(&(&slave->notify_work)->work) ->(work_completion)(&(&bond->mcast_work)->work) ->&rq->__lock FD: 232 BD: 1 +.+.: (wq_completion)bond832#2 ->&rq->__lock ->(work_completion)(&(&slave->notify_work)->work) ->(work_completion)(&(&bond->mcast_work)->work) FD: 232 BD: 1 +.+.: (wq_completion)bond820#3 ->&rq->__lock ->(work_completion)(&(&slave->notify_work)->work) ->(work_completion)(&(&bond->mcast_work)->work) FD: 232 BD: 1 +.+.: (wq_completion)bond833#2 ->&rq->__lock ->(work_completion)(&(&slave->notify_work)->work) ->(work_completion)(&(&bond->mcast_work)->work) FD: 232 BD: 1 +.+.: (wq_completion)bond841 ->&rq->__lock ->(work_completion)(&(&slave->notify_work)->work) ->(work_completion)(&(&bond->mcast_work)->work) FD: 232 BD: 1 +.+.: (wq_completion)bond821#3 ->(work_completion)(&(&slave->notify_work)->work) ->(work_completion)(&(&bond->mcast_work)->work) ->&rq->__lock FD: 232 BD: 1 +.+.: (wq_completion)bond834#2 ->&rq->__lock ->(work_completion)(&(&slave->notify_work)->work) ->(work_completion)(&(&bond->mcast_work)->work) FD: 232 BD: 1 +.+.: (wq_completion)bond842 ->&rq->__lock ->(work_completion)(&(&slave->notify_work)->work) ->(work_completion)(&(&bond->mcast_work)->work) FD: 232 BD: 1 +.+.: (wq_completion)bond822#3 ->&rq->__lock ->(work_completion)(&(&slave->notify_work)->work) ->(work_completion)(&(&bond->mcast_work)->work) FD: 232 BD: 1 +.+.: (wq_completion)bond835#2 ->(work_completion)(&(&slave->notify_work)->work) ->(work_completion)(&(&bond->mcast_work)->work) ->&rq->__lock FD: 232 BD: 1 +.+.: (wq_completion)bond843 ->(work_completion)(&(&slave->notify_work)->work) ->(work_completion)(&(&bond->mcast_work)->work) ->&rq->__lock FD: 232 BD: 1 +.+.: (wq_completion)bond823#3 ->(work_completion)(&(&slave->notify_work)->work) ->(work_completion)(&(&bond->mcast_work)->work) ->&rq->__lock FD: 232 BD: 1 +.+.: (wq_completion)bond836#2 ->(work_completion)(&(&slave->notify_work)->work) ->(work_completion)(&(&bond->mcast_work)->work) ->&rq->__lock FD: 232 BD: 1 +.+.: (wq_completion)bond844 ->(work_completion)(&(&slave->notify_work)->work) ->(work_completion)(&(&bond->mcast_work)->work) ->&rq->__lock FD: 232 BD: 1 +.+.: (wq_completion)bond824#3 ->&rq->__lock ->(work_completion)(&(&slave->notify_work)->work) ->(work_completion)(&(&bond->mcast_work)->work) FD: 232 BD: 1 +.+.: (wq_completion)bond837#2 ->&rq->__lock ->(work_completion)(&(&slave->notify_work)->work) ->(work_completion)(&(&bond->mcast_work)->work) FD: 232 BD: 1 +.+.: (wq_completion)bond845 ->&rq->__lock ->(work_completion)(&(&slave->notify_work)->work) ->(work_completion)(&(&bond->mcast_work)->work) FD: 232 BD: 1 +.+.: (wq_completion)bond825#3 ->&rq->__lock ->(work_completion)(&(&slave->notify_work)->work) ->(work_completion)(&(&bond->mcast_work)->work) ->&cfs_rq->removed.lock FD: 232 BD: 1 +.+.: (wq_completion)bond838#2 ->(work_completion)(&(&slave->notify_work)->work) ->(work_completion)(&(&bond->mcast_work)->work) ->&rq->__lock FD: 232 BD: 1 +.+.: (wq_completion)bond846 ->&rq->__lock ->(work_completion)(&(&slave->notify_work)->work) ->(work_completion)(&(&bond->mcast_work)->work) FD: 232 BD: 1 +.+.: (wq_completion)bond826#3 ->&rq->__lock ->(work_completion)(&(&slave->notify_work)->work) ->(work_completion)(&(&bond->mcast_work)->work) FD: 232 BD: 1 +.+.: (wq_completion)bond839#2 ->(work_completion)(&(&slave->notify_work)->work) ->(work_completion)(&(&bond->mcast_work)->work) ->&rq->__lock FD: 232 BD: 1 +.+.: (wq_completion)bond847 ->&rq->__lock ->(work_completion)(&(&slave->notify_work)->work) ->(work_completion)(&(&bond->mcast_work)->work) FD: 232 BD: 1 +.+.: (wq_completion)bond827#3 ->&rq->__lock ->(work_completion)(&(&slave->notify_work)->work) ->(work_completion)(&(&bond->mcast_work)->work) FD: 232 BD: 1 +.+.: (wq_completion)bond840 ->&rq->__lock ->(work_completion)(&(&slave->notify_work)->work) ->(work_completion)(&(&bond->mcast_work)->work) FD: 232 BD: 1 +.+.: (wq_completion)bond848 ->&rq->__lock ->(work_completion)(&(&slave->notify_work)->work) ->(work_completion)(&(&bond->mcast_work)->work) FD: 232 BD: 1 +.+.: (wq_completion)bond828#3 ->&rq->__lock ->(work_completion)(&(&slave->notify_work)->work) ->(work_completion)(&(&bond->mcast_work)->work) FD: 232 BD: 1 +.+.: (wq_completion)bond841#2 ->&rq->__lock ->(work_completion)(&(&slave->notify_work)->work) ->(work_completion)(&(&bond->mcast_work)->work) FD: 232 BD: 1 +.+.: (wq_completion)bond849 ->&rq->__lock ->(work_completion)(&(&slave->notify_work)->work) ->(work_completion)(&(&bond->mcast_work)->work) FD: 232 BD: 1 +.+.: (wq_completion)bond829#3 ->(work_completion)(&(&slave->notify_work)->work) ->(work_completion)(&(&bond->mcast_work)->work) ->&rq->__lock FD: 232 BD: 1 +.+.: (wq_completion)bond842#2 ->(work_completion)(&(&slave->notify_work)->work) ->(work_completion)(&(&bond->mcast_work)->work) ->&rq->__lock FD: 232 BD: 1 +.+.: (wq_completion)bond850 ->&rq->__lock ->(work_completion)(&(&slave->notify_work)->work) ->(work_completion)(&(&bond->mcast_work)->work) FD: 232 BD: 1 +.+.: (wq_completion)bond830#3 ->&rq->__lock ->(work_completion)(&(&slave->notify_work)->work) ->(work_completion)(&(&bond->mcast_work)->work) FD: 232 BD: 1 +.+.: (wq_completion)bond843#2 ->(work_completion)(&(&slave->notify_work)->work) ->(work_completion)(&(&bond->mcast_work)->work) ->&rq->__lock FD: 232 BD: 1 +.+.: (wq_completion)bond851 ->&rq->__lock ->(work_completion)(&(&slave->notify_work)->work) ->(work_completion)(&(&bond->mcast_work)->work) FD: 232 BD: 1 +.+.: (wq_completion)bond831#3 ->&rq->__lock ->(work_completion)(&(&slave->notify_work)->work) ->(work_completion)(&(&bond->mcast_work)->work) FD: 232 BD: 1 +.+.: (wq_completion)bond844#2 ->&rq->__lock ->(work_completion)(&(&slave->notify_work)->work) ->(work_completion)(&(&bond->mcast_work)->work) FD: 232 BD: 1 +.+.: (wq_completion)bond852 ->&rq->__lock ->(work_completion)(&(&slave->notify_work)->work) ->(work_completion)(&(&bond->mcast_work)->work) FD: 232 BD: 1 +.+.: (wq_completion)bond832#3 ->(work_completion)(&(&slave->notify_work)->work) ->(work_completion)(&(&bond->mcast_work)->work) ->&rq->__lock FD: 232 BD: 1 +.+.: (wq_completion)bond845#2 ->(work_completion)(&(&slave->notify_work)->work) ->(work_completion)(&(&bond->mcast_work)->work) ->&rq->__lock FD: 25 BD: 1 ..-.: &(&krcp->page_cache_work)->timer FD: 232 BD: 1 +.+.: (wq_completion)bond853 ->(work_completion)(&(&slave->notify_work)->work) ->(work_completion)(&(&bond->mcast_work)->work) ->&rq->__lock FD: 232 BD: 1 +.+.: (wq_completion)bond846#2 ->&rq->__lock ->(work_completion)(&(&slave->notify_work)->work) ->(work_completion)(&(&bond->mcast_work)->work) FD: 232 BD: 1 +.+.: (wq_completion)bond854 ->&rq->__lock ->(work_completion)(&(&slave->notify_work)->work) ->(work_completion)(&(&bond->mcast_work)->work) FD: 232 BD: 1 +.+.: (wq_completion)bond834#3 ->(work_completion)(&(&slave->notify_work)->work) ->(work_completion)(&(&bond->mcast_work)->work) ->&rq->__lock FD: 232 BD: 1 +.+.: (wq_completion)bond847#2 ->&rq->__lock ->(work_completion)(&(&slave->notify_work)->work) ->(work_completion)(&(&bond->mcast_work)->work) FD: 232 BD: 1 +.+.: (wq_completion)bond855 ->(work_completion)(&(&slave->notify_work)->work) ->(work_completion)(&(&bond->mcast_work)->work) ->&rq->__lock FD: 232 BD: 1 +.+.: (wq_completion)bond835#3 ->&rq->__lock ->(work_completion)(&(&slave->notify_work)->work) ->(work_completion)(&(&bond->mcast_work)->work) FD: 232 BD: 1 +.+.: (wq_completion)bond848#2 ->(work_completion)(&(&slave->notify_work)->work) ->(work_completion)(&(&bond->mcast_work)->work) ->&rq->__lock FD: 232 BD: 1 +.+.: (wq_completion)bond856 ->&rq->__lock ->(work_completion)(&(&slave->notify_work)->work) ->(work_completion)(&(&bond->mcast_work)->work) FD: 232 BD: 1 +.+.: (wq_completion)bond836#3 ->&rq->__lock ->(work_completion)(&(&slave->notify_work)->work) ->(work_completion)(&(&bond->mcast_work)->work) FD: 232 BD: 1 +.+.: (wq_completion)bond849#2 ->&rq->__lock ->(work_completion)(&(&slave->notify_work)->work) ->(work_completion)(&(&bond->mcast_work)->work) FD: 232 BD: 1 +.+.: (wq_completion)bond857 ->(work_completion)(&(&slave->notify_work)->work) ->(work_completion)(&(&bond->mcast_work)->work) ->&rq->__lock FD: 232 BD: 1 +.+.: (wq_completion)bond837#3 ->(work_completion)(&(&slave->notify_work)->work) ->(work_completion)(&(&bond->mcast_work)->work) ->&rq->__lock FD: 232 BD: 1 +.+.: (wq_completion)bond850#2 ->&rq->__lock ->(work_completion)(&(&slave->notify_work)->work) ->(work_completion)(&(&bond->mcast_work)->work) FD: 232 BD: 1 +.+.: (wq_completion)bond838#3 ->&rq->__lock ->(work_completion)(&(&slave->notify_work)->work) ->(work_completion)(&(&bond->mcast_work)->work) FD: 3 BD: 4541 ..-.: lock#11 ->&lruvec->lru_lock FD: 232 BD: 1 +.+.: (wq_completion)bond858 ->&rq->__lock ->(work_completion)(&(&slave->notify_work)->work) ->(work_completion)(&(&bond->mcast_work)->work) FD: 25 BD: 4515 +.+.: &cache->alloc_lock ->swap_avail_lock ->&rq->__lock ->&p->lock#2 FD: 45 BD: 4515 +.+.: shmem_swaplist_mutex ->&rq->__lock ->&xa->xa_lock#21 ->&info->lock ->&p->lock#2 ->&xa->xa_lock#9 FD: 9 BD: 4516 ....: &xa->xa_lock#21 ->pool_lock#2 ->key#30 ->&ctrl->lock ->&c->lock FD: 1 BD: 4515 +.+.: &tree->lock FD: 1 BD: 4517 ....: &ctrl->lock FD: 232 BD: 1 +.+.: (wq_completion)bond851#2 ->(work_completion)(&(&slave->notify_work)->work) ->(work_completion)(&(&bond->mcast_work)->work) ->&rq->__lock FD: 232 BD: 1 +.+.: (wq_completion)bond839#3 ->(work_completion)(&(&slave->notify_work)->work) ->(work_completion)(&(&bond->mcast_work)->work) ->&rq->__lock FD: 232 BD: 1 +.+.: (wq_completion)bond859 ->(work_completion)(&(&slave->notify_work)->work) ->(work_completion)(&(&bond->mcast_work)->work) ->&rq->__lock FD: 232 BD: 1 +.+.: (wq_completion)bond852#2 ->(work_completion)(&(&slave->notify_work)->work) ->(work_completion)(&(&bond->mcast_work)->work) ->&rq->__lock FD: 232 BD: 1 +.+.: (wq_completion)bond860 ->(work_completion)(&(&slave->notify_work)->work) ->(work_completion)(&(&bond->mcast_work)->work) ->&rq->__lock FD: 232 BD: 1 +.+.: (wq_completion)bond840#2 ->&rq->__lock ->(work_completion)(&(&slave->notify_work)->work) ->(work_completion)(&(&bond->mcast_work)->work) FD: 232 BD: 1 +.+.: (wq_completion)bond854#2 ->(work_completion)(&(&slave->notify_work)->work) ->(work_completion)(&(&bond->mcast_work)->work) ->&rq->__lock FD: 232 BD: 1 +.+.: (wq_completion)bond861 ->(work_completion)(&(&slave->notify_work)->work) ->(work_completion)(&(&bond->mcast_work)->work) ->&rq->__lock FD: 232 BD: 1 +.+.: (wq_completion)bond841#3 ->&rq->__lock ->(work_completion)(&(&slave->notify_work)->work) ->(work_completion)(&(&bond->mcast_work)->work) FD: 232 BD: 1 +.+.: (wq_completion)bond856#2 ->&rq->__lock ->(work_completion)(&(&slave->notify_work)->work) ->(work_completion)(&(&bond->mcast_work)->work) FD: 232 BD: 1 +.+.: (wq_completion)bond855#2 ->(work_completion)(&(&slave->notify_work)->work) ->(work_completion)(&(&bond->mcast_work)->work) ->&rq->__lock FD: 232 BD: 1 +.+.: (wq_completion)bond862 ->&rq->__lock ->(work_completion)(&(&slave->notify_work)->work) ->(work_completion)(&(&bond->mcast_work)->work) FD: 232 BD: 1 +.+.: (wq_completion)bond842#3 ->&rq->__lock ->(work_completion)(&(&slave->notify_work)->work) ->(work_completion)(&(&bond->mcast_work)->work) FD: 232 BD: 1 +.+.: (wq_completion)bond857#2 ->&rq->__lock ->(work_completion)(&(&slave->notify_work)->work) ->(work_completion)(&(&bond->mcast_work)->work) FD: 232 BD: 1 +.+.: (wq_completion)bond863 ->&rq->__lock ->(work_completion)(&(&slave->notify_work)->work) ->(work_completion)(&(&bond->mcast_work)->work) FD: 1 BD: 6 +...: key#32 FD: 232 BD: 1 +.+.: (wq_completion)bond843#3 ->&rq->__lock ->(work_completion)(&(&slave->notify_work)->work) ->(work_completion)(&(&bond->mcast_work)->work) FD: 232 BD: 1 +.+.: (wq_completion)bond858#2 ->&rq->__lock ->(work_completion)(&(&slave->notify_work)->work) ->(work_completion)(&(&bond->mcast_work)->work) FD: 232 BD: 1 +.+.: (wq_completion)bond844#3 ->&rq->__lock ->(work_completion)(&(&slave->notify_work)->work) ->(work_completion)(&(&bond->mcast_work)->work) FD: 232 BD: 1 +.+.: (wq_completion)bond859#2 ->(work_completion)(&(&slave->notify_work)->work) ->(work_completion)(&(&bond->mcast_work)->work) ->&rq->__lock FD: 232 BD: 1 +.+.: (wq_completion)bond864 ->&rq->__lock ->(work_completion)(&(&slave->notify_work)->work) ->(work_completion)(&(&bond->mcast_work)->work) FD: 232 BD: 1 +.+.: (wq_completion)bond845#3 ->&rq->__lock ->(work_completion)(&(&slave->notify_work)->work) ->(work_completion)(&(&bond->mcast_work)->work) FD: 232 BD: 1 +.+.: (wq_completion)bond860#2 ->&rq->__lock ->(work_completion)(&(&slave->notify_work)->work) ->(work_completion)(&(&bond->mcast_work)->work) FD: 232 BD: 1 +.+.: (wq_completion)bond865 ->(work_completion)(&(&slave->notify_work)->work) ->(work_completion)(&(&bond->mcast_work)->work) ->&rq->__lock FD: 232 BD: 1 +.+.: (wq_completion)bond846#3 ->&rq->__lock ->(work_completion)(&(&slave->notify_work)->work) ->(work_completion)(&(&bond->mcast_work)->work) FD: 128 BD: 1 .+.+: kn->active#74 ->fs_reclaim ->stock_lock ->&kernfs_locks->open_file_mutex[count] FD: 232 BD: 1 +.+.: (wq_completion)bond861#2 ->&rq->__lock ->(work_completion)(&(&slave->notify_work)->work) ->(work_completion)(&(&bond->mcast_work)->work) FD: 232 BD: 1 +.+.: (wq_completion)bond866 ->(work_completion)(&(&slave->notify_work)->work) ->(work_completion)(&(&bond->mcast_work)->work) ->&rq->__lock FD: 232 BD: 1 +.+.: (wq_completion)bond847#3 ->&rq->__lock ->(work_completion)(&(&slave->notify_work)->work) ->(work_completion)(&(&bond->mcast_work)->work) FD: 232 BD: 1 +.+.: (wq_completion)bond862#2 ->(work_completion)(&(&slave->notify_work)->work) ->(work_completion)(&(&bond->mcast_work)->work) ->&rq->__lock FD: 232 BD: 1 +.+.: (wq_completion)bond867 ->&rq->__lock ->(work_completion)(&(&slave->notify_work)->work) ->(work_completion)(&(&bond->mcast_work)->work) FD: 232 BD: 1 +.+.: (wq_completion)bond848#3 ->(work_completion)(&(&slave->notify_work)->work) ->(work_completion)(&(&bond->mcast_work)->work) ->&rq->__lock FD: 232 BD: 1 +.+.: (wq_completion)bond863#2 ->(work_completion)(&(&slave->notify_work)->work) ->(work_completion)(&(&bond->mcast_work)->work) ->&rq->__lock FD: 232 BD: 1 +.+.: (wq_completion)bond868 ->&rq->__lock ->(work_completion)(&(&slave->notify_work)->work) ->(work_completion)(&(&bond->mcast_work)->work) FD: 232 BD: 1 +.+.: (wq_completion)bond849#3 ->&rq->__lock ->(work_completion)(&(&slave->notify_work)->work) ->(work_completion)(&(&bond->mcast_work)->work) FD: 232 BD: 1 +.+.: (wq_completion)bond864#2 ->(work_completion)(&(&slave->notify_work)->work) ->(work_completion)(&(&bond->mcast_work)->work) ->&rq->__lock FD: 232 BD: 1 +.+.: (wq_completion)bond869 ->&rq->__lock ->(work_completion)(&(&slave->notify_work)->work) ->(work_completion)(&(&bond->mcast_work)->work) FD: 232 BD: 1 +.+.: (wq_completion)bond850#3 ->(work_completion)(&(&slave->notify_work)->work) ->(work_completion)(&(&bond->mcast_work)->work) ->&rq->__lock FD: 232 BD: 1 +.+.: (wq_completion)bond865#2 ->&rq->__lock ->(work_completion)(&(&slave->notify_work)->work) ->(work_completion)(&(&bond->mcast_work)->work) FD: 232 BD: 1 +.+.: (wq_completion)bond870 ->(work_completion)(&(&slave->notify_work)->work) ->(work_completion)(&(&bond->mcast_work)->work) ->&rq->__lock FD: 232 BD: 1 +.+.: (wq_completion)bond866#2 ->&rq->__lock ->(work_completion)(&(&slave->notify_work)->work) ->(work_completion)(&(&bond->mcast_work)->work) FD: 232 BD: 1 +.+.: (wq_completion)bond871 ->&rq->__lock ->(work_completion)(&(&slave->notify_work)->work) ->(work_completion)(&(&bond->mcast_work)->work) FD: 232 BD: 1 +.+.: (wq_completion)bond852#3 ->&rq->__lock ->(work_completion)(&(&slave->notify_work)->work) ->(work_completion)(&(&bond->mcast_work)->work) FD: 232 BD: 1 +.+.: (wq_completion)bond867#2 ->(work_completion)(&(&slave->notify_work)->work) ->(work_completion)(&(&bond->mcast_work)->work) ->&rq->__lock FD: 232 BD: 1 +.+.: (wq_completion)bond872 ->(work_completion)(&(&slave->notify_work)->work) ->(work_completion)(&(&bond->mcast_work)->work) ->&rq->__lock FD: 232 BD: 1 +.+.: (wq_completion)bond853#2 ->&rq->__lock ->(work_completion)(&(&slave->notify_work)->work) ->(work_completion)(&(&bond->mcast_work)->work) FD: 232 BD: 1 +.+.: (wq_completion)bond868#2 ->(work_completion)(&(&slave->notify_work)->work) ->(work_completion)(&(&bond->mcast_work)->work) ->&rq->__lock FD: 232 BD: 1 +.+.: (wq_completion)bond873 ->&rq->__lock ->(work_completion)(&(&slave->notify_work)->work) ->(work_completion)(&(&bond->mcast_work)->work) FD: 232 BD: 1 +.+.: (wq_completion)bond854#3 ->(work_completion)(&(&slave->notify_work)->work) ->(work_completion)(&(&bond->mcast_work)->work) ->&rq->__lock FD: 232 BD: 1 +.+.: (wq_completion)bond869#2 ->(work_completion)(&(&slave->notify_work)->work) ->(work_completion)(&(&bond->mcast_work)->work) ->&rq->__lock FD: 232 BD: 1 +.+.: (wq_completion)bond874 ->(work_completion)(&(&slave->notify_work)->work) ->(work_completion)(&(&bond->mcast_work)->work) ->&rq->__lock FD: 232 BD: 1 +.+.: (wq_completion)bond855#3 ->(work_completion)(&(&slave->notify_work)->work) ->(work_completion)(&(&bond->mcast_work)->work) ->&rq->__lock FD: 232 BD: 1 +.+.: (wq_completion)bond870#2 ->&rq->__lock ->(work_completion)(&(&slave->notify_work)->work) ->(work_completion)(&(&bond->mcast_work)->work) FD: 232 BD: 1 +.+.: (wq_completion)bond875 ->&rq->__lock ->(work_completion)(&(&slave->notify_work)->work) ->(work_completion)(&(&bond->mcast_work)->work) FD: 232 BD: 1 +.+.: (wq_completion)bond856#3 ->&rq->__lock ->(work_completion)(&(&slave->notify_work)->work) ->(work_completion)(&(&bond->mcast_work)->work) FD: 232 BD: 1 +.+.: (wq_completion)bond871#2 ->&rq->__lock ->(work_completion)(&(&slave->notify_work)->work) ->(work_completion)(&(&bond->mcast_work)->work) FD: 232 BD: 1 +.+.: (wq_completion)bond876 ->(work_completion)(&(&slave->notify_work)->work) ->(work_completion)(&(&bond->mcast_work)->work) ->&rq->__lock FD: 232 BD: 1 +.+.: (wq_completion)bond857#3 ->&rq->__lock ->(work_completion)(&(&slave->notify_work)->work) ->(work_completion)(&(&bond->mcast_work)->work) FD: 232 BD: 1 +.+.: (wq_completion)bond872#2 ->(work_completion)(&(&slave->notify_work)->work) ->(work_completion)(&(&bond->mcast_work)->work) ->&rq->__lock ->&cfs_rq->removed.lock ->&obj_hash[i].lock FD: 232 BD: 1 +.+.: (wq_completion)bond877 ->&rq->__lock ->(work_completion)(&(&slave->notify_work)->work) ->(work_completion)(&(&bond->mcast_work)->work) FD: 232 BD: 1 +.+.: (wq_completion)bond858#3 ->&rq->__lock ->(work_completion)(&(&slave->notify_work)->work) ->(work_completion)(&(&bond->mcast_work)->work) FD: 232 BD: 1 +.+.: (wq_completion)bond873#2 ->&rq->__lock ->(work_completion)(&(&slave->notify_work)->work) ->(work_completion)(&(&bond->mcast_work)->work) ->&cfs_rq->removed.lock FD: 232 BD: 1 +.+.: (wq_completion)bond878 ->(work_completion)(&(&slave->notify_work)->work) ->(work_completion)(&(&bond->mcast_work)->work) ->&rq->__lock FD: 232 BD: 1 +.+.: (wq_completion)bond859#3 ->&rq->__lock ->(work_completion)(&(&slave->notify_work)->work) ->(work_completion)(&(&bond->mcast_work)->work) FD: 232 BD: 1 +.+.: (wq_completion)bond874#2 ->&rq->__lock ->(work_completion)(&(&slave->notify_work)->work) ->(work_completion)(&(&bond->mcast_work)->work) FD: 232 BD: 1 +.+.: (wq_completion)bond879 ->(work_completion)(&(&slave->notify_work)->work) ->(work_completion)(&(&bond->mcast_work)->work) ->&rq->__lock FD: 232 BD: 1 +.+.: (wq_completion)bond860#3 ->(work_completion)(&(&slave->notify_work)->work) ->(work_completion)(&(&bond->mcast_work)->work) FD: 232 BD: 1 +.+.: (wq_completion)bond875#2 ->&rq->__lock ->(work_completion)(&(&slave->notify_work)->work) ->(work_completion)(&(&bond->mcast_work)->work) FD: 232 BD: 1 +.+.: (wq_completion)bond880 ->&rq->__lock ->(work_completion)(&(&slave->notify_work)->work) ->(work_completion)(&(&bond->mcast_work)->work) FD: 232 BD: 1 +.+.: (wq_completion)bond861#3 ->&rq->__lock ->(work_completion)(&(&slave->notify_work)->work) ->(work_completion)(&(&bond->mcast_work)->work) FD: 232 BD: 1 +.+.: (wq_completion)bond876#2 ->&rq->__lock ->(work_completion)(&(&slave->notify_work)->work) ->(work_completion)(&(&bond->mcast_work)->work) FD: 232 BD: 1 +.+.: (wq_completion)bond881 ->(work_completion)(&(&slave->notify_work)->work) ->(work_completion)(&(&bond->mcast_work)->work) ->&rq->__lock FD: 232 BD: 1 +.+.: (wq_completion)bond862#3 ->(work_completion)(&(&slave->notify_work)->work) ->(work_completion)(&(&bond->mcast_work)->work) ->&rq->__lock FD: 232 BD: 1 +.+.: (wq_completion)bond877#2 ->(work_completion)(&(&slave->notify_work)->work) ->(work_completion)(&(&bond->mcast_work)->work) ->&rq->__lock FD: 232 BD: 1 +.+.: (wq_completion)bond882 ->&rq->__lock ->(work_completion)(&(&slave->notify_work)->work) ->(work_completion)(&(&bond->mcast_work)->work) FD: 232 BD: 1 +.+.: (wq_completion)bond863#3 ->&rq->__lock ->(work_completion)(&(&slave->notify_work)->work) ->(work_completion)(&(&bond->mcast_work)->work) FD: 232 BD: 1 +.+.: (wq_completion)bond878#2 ->(work_completion)(&(&slave->notify_work)->work) ->(work_completion)(&(&bond->mcast_work)->work) ->&rq->__lock FD: 232 BD: 1 +.+.: (wq_completion)bond883 ->&rq->__lock ->(work_completion)(&(&slave->notify_work)->work) ->(work_completion)(&(&bond->mcast_work)->work) FD: 232 BD: 1 +.+.: (wq_completion)bond864#3 ->&rq->__lock ->(work_completion)(&(&slave->notify_work)->work) ->(work_completion)(&(&bond->mcast_work)->work) FD: 232 BD: 1 +.+.: (wq_completion)bond879#2 ->(work_completion)(&(&slave->notify_work)->work) ->(work_completion)(&(&bond->mcast_work)->work) ->&rq->__lock FD: 232 BD: 1 +.+.: (wq_completion)bond884 ->(work_completion)(&(&slave->notify_work)->work) ->(work_completion)(&(&bond->mcast_work)->work) ->&rq->__lock FD: 232 BD: 1 +.+.: (wq_completion)bond865#3 ->&rq->__lock ->(work_completion)(&(&slave->notify_work)->work) ->(work_completion)(&(&bond->mcast_work)->work) FD: 232 BD: 1 +.+.: (wq_completion)bond880#2 ->(work_completion)(&(&slave->notify_work)->work) ->(work_completion)(&(&bond->mcast_work)->work) ->&rq->__lock FD: 232 BD: 1 +.+.: (wq_completion)bond866#3 ->(work_completion)(&(&slave->notify_work)->work) ->(work_completion)(&(&bond->mcast_work)->work) ->&rq->__lock FD: 232 BD: 1 +.+.: (wq_completion)bond885 ->&rq->__lock ->(work_completion)(&(&slave->notify_work)->work) ->(work_completion)(&(&bond->mcast_work)->work) FD: 232 BD: 1 +.+.: (wq_completion)bond881#2 ->(work_completion)(&(&slave->notify_work)->work) ->(work_completion)(&(&bond->mcast_work)->work) ->&rq->__lock FD: 232 BD: 1 +.+.: (wq_completion)bond867#3 ->(work_completion)(&(&slave->notify_work)->work) ->(work_completion)(&(&bond->mcast_work)->work) ->&rq->__lock FD: 232 BD: 1 +.+.: (wq_completion)bond886 ->&rq->__lock ->(work_completion)(&(&slave->notify_work)->work) ->(work_completion)(&(&bond->mcast_work)->work) FD: 232 BD: 1 +.+.: (wq_completion)bond882#2 ->&rq->__lock ->(work_completion)(&(&slave->notify_work)->work) ->(work_completion)(&(&bond->mcast_work)->work) FD: 232 BD: 1 +.+.: (wq_completion)bond868#3 ->(work_completion)(&(&slave->notify_work)->work) ->(work_completion)(&(&bond->mcast_work)->work) ->&rq->__lock FD: 232 BD: 1 +.+.: (wq_completion)bond887 ->&rq->__lock ->(work_completion)(&(&slave->notify_work)->work) ->(work_completion)(&(&bond->mcast_work)->work) FD: 232 BD: 1 +.+.: (wq_completion)bond883#2 ->&rq->__lock ->(work_completion)(&(&slave->notify_work)->work) ->(work_completion)(&(&bond->mcast_work)->work) FD: 232 BD: 1 +.+.: (wq_completion)bond869#3 ->&rq->__lock ->(work_completion)(&(&slave->notify_work)->work) ->(work_completion)(&(&bond->mcast_work)->work) FD: 232 BD: 1 +.+.: (wq_completion)bond888 ->(work_completion)(&(&slave->notify_work)->work) ->(work_completion)(&(&bond->mcast_work)->work) ->&rq->__lock FD: 232 BD: 1 +.+.: (wq_completion)bond884#2 ->&rq->__lock ->(work_completion)(&(&slave->notify_work)->work) ->(work_completion)(&(&bond->mcast_work)->work) FD: 232 BD: 1 +.+.: (wq_completion)bond870#3 ->(work_completion)(&(&slave->notify_work)->work) ->(work_completion)(&(&bond->mcast_work)->work) ->&rq->__lock FD: 232 BD: 1 +.+.: (wq_completion)bond889 ->&rq->__lock ->(work_completion)(&(&slave->notify_work)->work) ->(work_completion)(&(&bond->mcast_work)->work) FD: 232 BD: 1 +.+.: (wq_completion)bond885#2 ->(work_completion)(&(&slave->notify_work)->work) ->(work_completion)(&(&bond->mcast_work)->work) ->&rq->__lock FD: 232 BD: 1 +.+.: (wq_completion)bond871#3 ->&rq->__lock ->(work_completion)(&(&slave->notify_work)->work) ->(work_completion)(&(&bond->mcast_work)->work) FD: 232 BD: 1 +.+.: (wq_completion)bond890 ->&rq->__lock ->(work_completion)(&(&slave->notify_work)->work) ->(work_completion)(&(&bond->mcast_work)->work) FD: 232 BD: 1 +.+.: (wq_completion)bond187#4 ->&rq->__lock ->(work_completion)(&(&slave->notify_work)->work) ->(work_completion)(&(&bond->mcast_work)->work) FD: 232 BD: 1 +.+.: (wq_completion)bond891 ->&rq->__lock ->(work_completion)(&(&slave->notify_work)->work) ->(work_completion)(&(&bond->mcast_work)->work) FD: 232 BD: 1 +.+.: (wq_completion)bond886#2 ->&rq->__lock ->(work_completion)(&(&slave->notify_work)->work) ->(work_completion)(&(&bond->mcast_work)->work) ->&cfs_rq->removed.lock ->&obj_hash[i].lock FD: 232 BD: 1 +.+.: (wq_completion)bond872#3 ->&rq->__lock ->(work_completion)(&(&slave->notify_work)->work) ->(work_completion)(&(&bond->mcast_work)->work) FD: 232 BD: 1 +.+.: (wq_completion)bond892 ->&rq->__lock ->(work_completion)(&(&slave->notify_work)->work) ->(work_completion)(&(&bond->mcast_work)->work) FD: 232 BD: 1 +.+.: (wq_completion)bond188#4 ->&rq->__lock ->(work_completion)(&(&slave->notify_work)->work) ->(work_completion)(&(&bond->mcast_work)->work) FD: 232 BD: 1 +.+.: (wq_completion)bond887#2 ->&rq->__lock ->(work_completion)(&(&slave->notify_work)->work) ->(work_completion)(&(&bond->mcast_work)->work) FD: 232 BD: 1 +.+.: (wq_completion)bond893 ->&rq->__lock ->(work_completion)(&(&slave->notify_work)->work) ->(work_completion)(&(&bond->mcast_work)->work) FD: 232 BD: 1 +.+.: (wq_completion)bond888#2 ->(work_completion)(&(&slave->notify_work)->work) ->(work_completion)(&(&bond->mcast_work)->work) ->&rq->__lock FD: 232 BD: 1 +.+.: (wq_completion)bond873#3 ->&rq->__lock ->(work_completion)(&(&slave->notify_work)->work) ->(work_completion)(&(&bond->mcast_work)->work) FD: 232 BD: 1 +.+.: (wq_completion)bond894 ->(work_completion)(&(&slave->notify_work)->work) ->(work_completion)(&(&bond->mcast_work)->work) FD: 232 BD: 1 +.+.: (wq_completion)bond889#2 ->&rq->__lock ->(work_completion)(&(&slave->notify_work)->work) ->(work_completion)(&(&bond->mcast_work)->work) FD: 232 BD: 1 +.+.: (wq_completion)bond874#3 ->&rq->__lock ->(work_completion)(&(&slave->notify_work)->work) ->(work_completion)(&(&bond->mcast_work)->work) ->&cfs_rq->removed.lock ->&obj_hash[i].lock FD: 232 BD: 1 +.+.: (wq_completion)bond895 ->(work_completion)(&(&slave->notify_work)->work) ->(work_completion)(&(&bond->mcast_work)->work) ->&rq->__lock FD: 232 BD: 1 +.+.: (wq_completion)bond890#2 ->(work_completion)(&(&slave->notify_work)->work) ->(work_completion)(&(&bond->mcast_work)->work) ->&rq->__lock FD: 232 BD: 1 +.+.: (wq_completion)bond875#3 ->(work_completion)(&(&slave->notify_work)->work) ->(work_completion)(&(&bond->mcast_work)->work) ->&rq->__lock ->&cfs_rq->removed.lock FD: 232 BD: 1 +.+.: (wq_completion)bond896 ->&rq->__lock ->(work_completion)(&(&slave->notify_work)->work) ->(work_completion)(&(&bond->mcast_work)->work) FD: 232 BD: 1 +.+.: (wq_completion)bond891#2 ->(work_completion)(&(&slave->notify_work)->work) ->(work_completion)(&(&bond->mcast_work)->work) ->&rq->__lock FD: 232 BD: 1 +.+.: (wq_completion)bond876#3 ->(work_completion)(&(&slave->notify_work)->work) ->(work_completion)(&(&bond->mcast_work)->work) ->&rq->__lock FD: 232 BD: 1 +.+.: (wq_completion)bond897 ->&rq->__lock ->(work_completion)(&(&slave->notify_work)->work) ->(work_completion)(&(&bond->mcast_work)->work) FD: 232 BD: 1 +.+.: (wq_completion)bond892#2 ->&rq->__lock ->(work_completion)(&(&slave->notify_work)->work) ->(work_completion)(&(&bond->mcast_work)->work) FD: 232 BD: 1 +.+.: (wq_completion)bond189#4 ->&rq->__lock ->(work_completion)(&(&slave->notify_work)->work) ->(work_completion)(&(&bond->mcast_work)->work) FD: 232 BD: 1 +.+.: (wq_completion)bond877#3 ->&rq->__lock ->(work_completion)(&(&slave->notify_work)->work) ->(work_completion)(&(&bond->mcast_work)->work) FD: 232 BD: 1 +.+.: (wq_completion)bond898 ->&rq->__lock ->(work_completion)(&(&slave->notify_work)->work) ->(work_completion)(&(&bond->mcast_work)->work) FD: 232 BD: 1 +.+.: (wq_completion)bond893#2 ->&rq->__lock ->(work_completion)(&(&slave->notify_work)->work) ->(work_completion)(&(&bond->mcast_work)->work) FD: 232 BD: 1 +.+.: (wq_completion)bond878#3 ->&rq->__lock ->(work_completion)(&(&slave->notify_work)->work) ->(work_completion)(&(&bond->mcast_work)->work) FD: 232 BD: 1 +.+.: (wq_completion)bond899 ->(work_completion)(&(&slave->notify_work)->work) ->(work_completion)(&(&bond->mcast_work)->work) ->&rq->__lock FD: 232 BD: 1 +.+.: (wq_completion)bond894#2 ->&rq->__lock ->(work_completion)(&(&slave->notify_work)->work) ->(work_completion)(&(&bond->mcast_work)->work) FD: 232 BD: 1 +.+.: (wq_completion)bond879#3 ->(work_completion)(&(&slave->notify_work)->work) ->(work_completion)(&(&bond->mcast_work)->work) ->&rq->__lock FD: 232 BD: 1 +.+.: (wq_completion)bond895#2 ->&rq->__lock ->(work_completion)(&(&slave->notify_work)->work) ->(work_completion)(&(&bond->mcast_work)->work) FD: 232 BD: 1 +.+.: (wq_completion)bond880#3 ->&rq->__lock ->(work_completion)(&(&slave->notify_work)->work) ->(work_completion)(&(&bond->mcast_work)->work) FD: 232 BD: 1 +.+.: (wq_completion)bond900 ->(work_completion)(&(&slave->notify_work)->work) ->(work_completion)(&(&bond->mcast_work)->work) ->&rq->__lock FD: 232 BD: 1 +.+.: (wq_completion)bond896#2 ->&rq->__lock ->(work_completion)(&(&slave->notify_work)->work) ->(work_completion)(&(&bond->mcast_work)->work) FD: 232 BD: 1 +.+.: (wq_completion)bond190#4 ->(work_completion)(&(&slave->notify_work)->work) ->(work_completion)(&(&bond->mcast_work)->work) ->&rq->__lock FD: 232 BD: 1 +.+.: (wq_completion)bond881#3 ->&rq->__lock ->(work_completion)(&(&slave->notify_work)->work) ->(work_completion)(&(&bond->mcast_work)->work) FD: 232 BD: 1 +.+.: (wq_completion)bond901 ->(work_completion)(&(&slave->notify_work)->work) ->(work_completion)(&(&bond->mcast_work)->work) ->&rq->__lock FD: 232 BD: 1 +.+.: (wq_completion)bond882#3 ->(work_completion)(&(&slave->notify_work)->work) ->(work_completion)(&(&bond->mcast_work)->work) ->&rq->__lock FD: 232 BD: 1 +.+.: (wq_completion)bond902 ->&rq->__lock ->(work_completion)(&(&slave->notify_work)->work) ->(work_completion)(&(&bond->mcast_work)->work) FD: 232 BD: 1 +.+.: (wq_completion)bond898#2 ->(work_completion)(&(&slave->notify_work)->work) ->(work_completion)(&(&bond->mcast_work)->work) ->&rq->__lock FD: 232 BD: 1 +.+.: (wq_completion)bond883#3 ->(work_completion)(&(&slave->notify_work)->work) ->(work_completion)(&(&bond->mcast_work)->work) ->&rq->__lock FD: 232 BD: 1 +.+.: (wq_completion)bond899#2 ->&rq->__lock ->(work_completion)(&(&slave->notify_work)->work) ->(work_completion)(&(&bond->mcast_work)->work) FD: 232 BD: 1 +.+.: (wq_completion)bond884#3 ->(work_completion)(&(&slave->notify_work)->work) ->(work_completion)(&(&bond->mcast_work)->work) ->&rq->__lock ->&cfs_rq->removed.lock FD: 232 BD: 1 +.+.: (wq_completion)bond191#4 ->&rq->__lock ->(work_completion)(&(&slave->notify_work)->work) ->(work_completion)(&(&bond->mcast_work)->work) ->&cfs_rq->removed.lock FD: 232 BD: 1 +.+.: (wq_completion)bond903 ->&rq->__lock ->(work_completion)(&(&slave->notify_work)->work) ->(work_completion)(&(&bond->mcast_work)->work) FD: 232 BD: 1 +.+.: (wq_completion)bond904 ->(work_completion)(&(&slave->notify_work)->work) ->(work_completion)(&(&bond->mcast_work)->work) ->&rq->__lock FD: 232 BD: 1 +.+.: (wq_completion)bond900#2 ->&rq->__lock ->(work_completion)(&(&slave->notify_work)->work) ->(work_completion)(&(&bond->mcast_work)->work) FD: 232 BD: 1 +.+.: (wq_completion)bond885#3 ->&rq->__lock ->(work_completion)(&(&slave->notify_work)->work) ->(work_completion)(&(&bond->mcast_work)->work) FD: 232 BD: 1 +.+.: (wq_completion)bond905 ->(work_completion)(&(&slave->notify_work)->work) ->(work_completion)(&(&bond->mcast_work)->work) ->&rq->__lock FD: 232 BD: 1 +.+.: (wq_completion)bond192#4 ->(work_completion)(&(&slave->notify_work)->work) ->(work_completion)(&(&bond->mcast_work)->work) ->&rq->__lock FD: 232 BD: 1 +.+.: (wq_completion)bond901#2 ->&rq->__lock ->(work_completion)(&(&slave->notify_work)->work) ->(work_completion)(&(&bond->mcast_work)->work) FD: 232 BD: 1 +.+.: (wq_completion)bond886#3 ->&rq->__lock ->(work_completion)(&(&slave->notify_work)->work) ->(work_completion)(&(&bond->mcast_work)->work) FD: 232 BD: 1 +.+.: (wq_completion)bond906 ->(work_completion)(&(&slave->notify_work)->work) ->(work_completion)(&(&bond->mcast_work)->work) ->&rq->__lock FD: 232 BD: 1 +.+.: (wq_completion)bond902#2 ->&rq->__lock ->(work_completion)(&(&slave->notify_work)->work) ->(work_completion)(&(&bond->mcast_work)->work) FD: 232 BD: 1 +.+.: (wq_completion)bond887#3 ->&rq->__lock ->(work_completion)(&(&slave->notify_work)->work) ->(work_completion)(&(&bond->mcast_work)->work) FD: 232 BD: 1 +.+.: (wq_completion)bond903#2 ->(work_completion)(&(&slave->notify_work)->work) ->(work_completion)(&(&bond->mcast_work)->work) FD: 232 BD: 1 +.+.: (wq_completion)bond888#3 ->(work_completion)(&(&slave->notify_work)->work) ->(work_completion)(&(&bond->mcast_work)->work) ->&rq->__lock FD: 232 BD: 1 +.+.: (wq_completion)bond908 ->(work_completion)(&(&slave->notify_work)->work) ->(work_completion)(&(&bond->mcast_work)->work) ->&rq->__lock FD: 232 BD: 1 +.+.: (wq_completion)bond904#2 ->&rq->__lock ->(work_completion)(&(&slave->notify_work)->work) ->(work_completion)(&(&bond->mcast_work)->work) FD: 232 BD: 1 +.+.: (wq_completion)bond889#3 ->(work_completion)(&(&slave->notify_work)->work) ->(work_completion)(&(&bond->mcast_work)->work) ->&rq->__lock FD: 232 BD: 1 +.+.: (wq_completion)bond905#2 ->&rq->__lock ->(work_completion)(&(&slave->notify_work)->work) ->(work_completion)(&(&bond->mcast_work)->work) FD: 232 BD: 1 +.+.: (wq_completion)bond193#4 ->(work_completion)(&(&slave->notify_work)->work) ->(work_completion)(&(&bond->mcast_work)->work) ->&rq->__lock FD: 232 BD: 1 +.+.: (wq_completion)bond910 ->&rq->__lock ->(work_completion)(&(&slave->notify_work)->work) ->(work_completion)(&(&bond->mcast_work)->work) FD: 232 BD: 1 +.+.: (wq_completion)bond906#2 ->(work_completion)(&(&slave->notify_work)->work) ->(work_completion)(&(&bond->mcast_work)->work) FD: 232 BD: 1 +.+.: (wq_completion)bond890#3 ->&rq->__lock ->(work_completion)(&(&slave->notify_work)->work) ->(work_completion)(&(&bond->mcast_work)->work) FD: 232 BD: 1 +.+.: (wq_completion)bond911 ->&rq->__lock ->(work_completion)(&(&slave->notify_work)->work) ->(work_completion)(&(&bond->mcast_work)->work) FD: 232 BD: 1 +.+.: (wq_completion)bond907 ->(work_completion)(&(&slave->notify_work)->work) ->(work_completion)(&(&bond->mcast_work)->work) ->&rq->__lock FD: 232 BD: 1 +.+.: (wq_completion)bond891#3 ->&rq->__lock ->(work_completion)(&(&slave->notify_work)->work) ->(work_completion)(&(&bond->mcast_work)->work) FD: 232 BD: 1 +.+.: (wq_completion)bond912 ->&rq->__lock ->(work_completion)(&(&slave->notify_work)->work) ->(work_completion)(&(&bond->mcast_work)->work) FD: 232 BD: 1 +.+.: (wq_completion)bond908#2 ->(work_completion)(&(&slave->notify_work)->work) ->(work_completion)(&(&bond->mcast_work)->work) ->&rq->__lock FD: 232 BD: 1 +.+.: (wq_completion)bond892#3 ->&rq->__lock ->(work_completion)(&(&slave->notify_work)->work) ->(work_completion)(&(&bond->mcast_work)->work) FD: 232 BD: 1 +.+.: (wq_completion)bond909 ->&rq->__lock ->(work_completion)(&(&slave->notify_work)->work) ->(work_completion)(&(&bond->mcast_work)->work) FD: 232 BD: 1 +.+.: (wq_completion)bond893#3 ->&rq->__lock ->(work_completion)(&(&slave->notify_work)->work) ->(work_completion)(&(&bond->mcast_work)->work) FD: 63 BD: 1 .+.+: &type->s_umount_key#53 ->&lru->node[i].lock ->&dentry->d_lock ->&sb->s_type->i_lock_key#30 ->&s->s_inode_list_lock ->&xa->xa_lock#9 ->inode_hash_lock ->&obj_hash[i].lock ->&fsnotify_mark_srcu ->&rq->__lock FD: 232 BD: 1 +.+.: (wq_completion)bond913 ->&rq->__lock ->(work_completion)(&(&slave->notify_work)->work) ->(work_completion)(&(&bond->mcast_work)->work) FD: 232 BD: 1 +.+.: (wq_completion)bond910#2 ->(work_completion)(&(&slave->notify_work)->work) ->(work_completion)(&(&bond->mcast_work)->work) ->&rq->__lock FD: 232 BD: 1 +.+.: (wq_completion)bond894#3 ->&rq->__lock ->(work_completion)(&(&slave->notify_work)->work) ->(work_completion)(&(&bond->mcast_work)->work) FD: 232 BD: 1 +.+.: (wq_completion)bond911#2 ->(work_completion)(&(&slave->notify_work)->work) ->(work_completion)(&(&bond->mcast_work)->work) ->&rq->__lock FD: 232 BD: 1 +.+.: (wq_completion)bond895#3 ->(work_completion)(&(&slave->notify_work)->work) ->(work_completion)(&(&bond->mcast_work)->work) ->&rq->__lock FD: 232 BD: 1 +.+.: (wq_completion)bond914 ->&rq->__lock ->(work_completion)(&(&slave->notify_work)->work) ->(work_completion)(&(&bond->mcast_work)->work) FD: 232 BD: 1 +.+.: (wq_completion)bond194#4 ->&rq->__lock ->(work_completion)(&(&slave->notify_work)->work) ->(work_completion)(&(&bond->mcast_work)->work) FD: 232 BD: 1 +.+.: (wq_completion)bond896#3 ->(work_completion)(&(&slave->notify_work)->work) ->(work_completion)(&(&bond->mcast_work)->work) ->&rq->__lock FD: 232 BD: 1 +.+.: (wq_completion)bond915 ->&rq->__lock ->(work_completion)(&(&slave->notify_work)->work) ->(work_completion)(&(&bond->mcast_work)->work) FD: 232 BD: 1 +.+.: (wq_completion)bond195#4 ->&rq->__lock ->(work_completion)(&(&slave->notify_work)->work) ->(work_completion)(&(&bond->mcast_work)->work) ->&cfs_rq->removed.lock FD: 232 BD: 1 +.+.: (wq_completion)bond913#2 ->&rq->__lock ->(work_completion)(&(&slave->notify_work)->work) ->(work_completion)(&(&bond->mcast_work)->work) FD: 232 BD: 1 +.+.: (wq_completion)bond897#2 ->&rq->__lock ->(work_completion)(&(&slave->notify_work)->work) ->(work_completion)(&(&bond->mcast_work)->work) FD: 232 BD: 1 +.+.: (wq_completion)bond916 ->&rq->__lock ->(work_completion)(&(&slave->notify_work)->work) ->(work_completion)(&(&bond->mcast_work)->work) FD: 232 BD: 1 +.+.: (wq_completion)bond196#3 ->&rq->__lock ->(work_completion)(&(&slave->notify_work)->work) ->(work_completion)(&(&bond->mcast_work)->work) FD: 232 BD: 1 +.+.: (wq_completion)bond914#2 ->(work_completion)(&(&slave->notify_work)->work) ->(work_completion)(&(&bond->mcast_work)->work) ->&rq->__lock FD: 232 BD: 1 +.+.: (wq_completion)bond898#3 ->(work_completion)(&(&slave->notify_work)->work) ->(work_completion)(&(&bond->mcast_work)->work) ->&rq->__lock FD: 232 BD: 1 +.+.: (wq_completion)bond917 ->(work_completion)(&(&slave->notify_work)->work) ->(work_completion)(&(&bond->mcast_work)->work) ->&rq->__lock FD: 232 BD: 1 +.+.: (wq_completion)bond899#3 ->&rq->__lock ->(work_completion)(&(&slave->notify_work)->work) ->(work_completion)(&(&bond->mcast_work)->work) FD: 232 BD: 1 +.+.: (wq_completion)bond916#2 ->&rq->__lock ->(work_completion)(&(&slave->notify_work)->work) ->(work_completion)(&(&bond->mcast_work)->work) FD: 232 BD: 1 +.+.: (wq_completion)bond918 ->(work_completion)(&(&slave->notify_work)->work) ->(work_completion)(&(&bond->mcast_work)->work) FD: 232 BD: 1 +.+.: (wq_completion)bond900#3 ->&rq->__lock ->(work_completion)(&(&slave->notify_work)->work) ->(work_completion)(&(&bond->mcast_work)->work) FD: 232 BD: 1 +.+.: (wq_completion)bond919 ->(work_completion)(&(&slave->notify_work)->work) ->(work_completion)(&(&bond->mcast_work)->work) ->&rq->__lock FD: 232 BD: 1 +.+.: (wq_completion)bond197#4 ->(work_completion)(&(&slave->notify_work)->work) ->(work_completion)(&(&bond->mcast_work)->work) ->&rq->__lock FD: 232 BD: 1 +.+.: (wq_completion)bond917#2 ->&rq->__lock ->(work_completion)(&(&slave->notify_work)->work) ->(work_completion)(&(&bond->mcast_work)->work) FD: 232 BD: 1 +.+.: (wq_completion)bond901#3 ->&rq->__lock ->(work_completion)(&(&slave->notify_work)->work) ->(work_completion)(&(&bond->mcast_work)->work) FD: 232 BD: 1 +.+.: (wq_completion)bond920 ->&rq->__lock ->(work_completion)(&(&slave->notify_work)->work) ->(work_completion)(&(&bond->mcast_work)->work) FD: 232 BD: 1 +.+.: (wq_completion)bond198#4 ->(work_completion)(&(&slave->notify_work)->work) ->(work_completion)(&(&bond->mcast_work)->work) FD: 232 BD: 1 +.+.: (wq_completion)bond918#2 ->&rq->__lock ->(work_completion)(&(&slave->notify_work)->work) ->(work_completion)(&(&bond->mcast_work)->work) FD: 232 BD: 1 +.+.: (wq_completion)bond199#4 ->&rq->__lock ->(work_completion)(&(&slave->notify_work)->work) ->(work_completion)(&(&bond->mcast_work)->work) FD: 232 BD: 1 +.+.: (wq_completion)bond902#3 ->&rq->__lock ->(work_completion)(&(&slave->notify_work)->work) ->(work_completion)(&(&bond->mcast_work)->work) FD: 232 BD: 1 +.+.: (wq_completion)bond921 ->(work_completion)(&(&slave->notify_work)->work) ->(work_completion)(&(&bond->mcast_work)->work) ->&rq->__lock FD: 232 BD: 1 +.+.: (wq_completion)bond200#4 ->(work_completion)(&(&slave->notify_work)->work) ->(work_completion)(&(&bond->mcast_work)->work) ->&rq->__lock FD: 232 BD: 1 +.+.: (wq_completion)bond919#2 ->&rq->__lock ->(work_completion)(&(&slave->notify_work)->work) ->(work_completion)(&(&bond->mcast_work)->work) FD: 232 BD: 1 +.+.: (wq_completion)bond903#3 ->&rq->__lock ->(work_completion)(&(&slave->notify_work)->work) ->(work_completion)(&(&bond->mcast_work)->work) FD: 232 BD: 1 +.+.: (wq_completion)bond922 ->&rq->__lock ->(work_completion)(&(&slave->notify_work)->work) ->(work_completion)(&(&bond->mcast_work)->work) FD: 232 BD: 1 +.+.: (wq_completion)bond920#2 ->&rq->__lock ->(work_completion)(&(&slave->notify_work)->work) ->(work_completion)(&(&bond->mcast_work)->work) FD: 232 BD: 1 +.+.: (wq_completion)bond904#3 ->&rq->__lock ->(work_completion)(&(&slave->notify_work)->work) ->(work_completion)(&(&bond->mcast_work)->work) FD: 232 BD: 1 +.+.: (wq_completion)bond923 ->(work_completion)(&(&slave->notify_work)->work) ->(work_completion)(&(&bond->mcast_work)->work) ->&rq->__lock FD: 232 BD: 1 +.+.: (wq_completion)bond921#2 ->&rq->__lock ->(work_completion)(&(&slave->notify_work)->work) ->(work_completion)(&(&bond->mcast_work)->work) FD: 232 BD: 1 +.+.: (wq_completion)bond905#3 ->&rq->__lock ->(work_completion)(&(&slave->notify_work)->work) ->(work_completion)(&(&bond->mcast_work)->work) FD: 232 BD: 1 +.+.: (wq_completion)bond924 ->(work_completion)(&(&slave->notify_work)->work) ->(work_completion)(&(&bond->mcast_work)->work) FD: 232 BD: 1 +.+.: (wq_completion)bond922#2 ->(work_completion)(&(&slave->notify_work)->work) ->(work_completion)(&(&bond->mcast_work)->work) ->&rq->__lock FD: 232 BD: 1 +.+.: (wq_completion)bond906#3 ->(work_completion)(&(&slave->notify_work)->work) ->(work_completion)(&(&bond->mcast_work)->work) ->&rq->__lock FD: 232 BD: 1 +.+.: (wq_completion)bond201#4 ->(work_completion)(&(&slave->notify_work)->work) ->(work_completion)(&(&bond->mcast_work)->work) ->&rq->__lock FD: 232 BD: 1 +.+.: (wq_completion)bond925 ->&rq->__lock ->(work_completion)(&(&slave->notify_work)->work) ->(work_completion)(&(&bond->mcast_work)->work) ->&cfs_rq->removed.lock ->&obj_hash[i].lock FD: 232 BD: 1 +.+.: (wq_completion)bond923#2 ->(work_completion)(&(&slave->notify_work)->work) ->(work_completion)(&(&bond->mcast_work)->work) ->&rq->__lock FD: 232 BD: 1 +.+.: (wq_completion)bond907#2 ->(work_completion)(&(&slave->notify_work)->work) ->(work_completion)(&(&bond->mcast_work)->work) ->&rq->__lock FD: 232 BD: 1 +.+.: (wq_completion)bond202#4 ->&rq->__lock ->(work_completion)(&(&slave->notify_work)->work) ->(work_completion)(&(&bond->mcast_work)->work) FD: 232 BD: 1 +.+.: (wq_completion)bond924#2 ->&rq->__lock ->(work_completion)(&(&slave->notify_work)->work) ->(work_completion)(&(&bond->mcast_work)->work) FD: 232 BD: 1 +.+.: (wq_completion)bond908#3 ->&rq->__lock ->(work_completion)(&(&slave->notify_work)->work) ->(work_completion)(&(&bond->mcast_work)->work) FD: 232 BD: 1 +.+.: (wq_completion)bond926 ->&rq->__lock ->(work_completion)(&(&slave->notify_work)->work) ->(work_completion)(&(&bond->mcast_work)->work) FD: 232 BD: 1 +.+.: (wq_completion)bond925#2 ->&rq->__lock ->(work_completion)(&(&slave->notify_work)->work) ->(work_completion)(&(&bond->mcast_work)->work) FD: 232 BD: 1 +.+.: (wq_completion)bond909#2 ->&rq->__lock ->(work_completion)(&(&slave->notify_work)->work) ->(work_completion)(&(&bond->mcast_work)->work) FD: 232 BD: 1 +.+.: (wq_completion)bond927 ->&rq->__lock ->(work_completion)(&(&slave->notify_work)->work) ->(work_completion)(&(&bond->mcast_work)->work) FD: 232 BD: 1 +.+.: (wq_completion)bond926#2 ->&rq->__lock ->(work_completion)(&(&slave->notify_work)->work) ->(work_completion)(&(&bond->mcast_work)->work) FD: 232 BD: 1 +.+.: (wq_completion)bond910#3 ->(work_completion)(&(&slave->notify_work)->work) ->(work_completion)(&(&bond->mcast_work)->work) ->&rq->__lock FD: 232 BD: 1 +.+.: (wq_completion)bond928 ->(work_completion)(&(&slave->notify_work)->work) ->(work_completion)(&(&bond->mcast_work)->work) ->&rq->__lock FD: 232 BD: 1 +.+.: (wq_completion)bond927#2 ->&rq->__lock ->(work_completion)(&(&slave->notify_work)->work) ->(work_completion)(&(&bond->mcast_work)->work) FD: 232 BD: 1 +.+.: (wq_completion)bond911#3 ->(work_completion)(&(&slave->notify_work)->work) ->(work_completion)(&(&bond->mcast_work)->work) ->&rq->__lock FD: 232 BD: 1 +.+.: (wq_completion)bond203#4 ->(work_completion)(&(&slave->notify_work)->work) ->(work_completion)(&(&bond->mcast_work)->work) ->&rq->__lock FD: 232 BD: 1 +.+.: (wq_completion)bond929 ->&rq->__lock ->(work_completion)(&(&slave->notify_work)->work) ->(work_completion)(&(&bond->mcast_work)->work) FD: 232 BD: 1 +.+.: (wq_completion)bond928#2 ->&rq->__lock ->(work_completion)(&(&slave->notify_work)->work) ->(work_completion)(&(&bond->mcast_work)->work) FD: 232 BD: 1 +.+.: (wq_completion)bond912#2 ->&rq->__lock ->(work_completion)(&(&slave->notify_work)->work) ->(work_completion)(&(&bond->mcast_work)->work) ->&cfs_rq->removed.lock ->&obj_hash[i].lock FD: 232 BD: 1 +.+.: (wq_completion)bond930 ->&rq->__lock ->(work_completion)(&(&slave->notify_work)->work) ->(work_completion)(&(&bond->mcast_work)->work) FD: 232 BD: 1 +.+.: (wq_completion)bond204#4 ->&rq->__lock ->(work_completion)(&(&slave->notify_work)->work) ->(work_completion)(&(&bond->mcast_work)->work) FD: 232 BD: 1 +.+.: (wq_completion)bond929#2 ->(work_completion)(&(&slave->notify_work)->work) ->(work_completion)(&(&bond->mcast_work)->work) ->&rq->__lock FD: 232 BD: 1 +.+.: (wq_completion)bond913#3 ->(work_completion)(&(&slave->notify_work)->work) ->(work_completion)(&(&bond->mcast_work)->work) ->&rq->__lock ->&cfs_rq->removed.lock ->&obj_hash[i].lock FD: 232 BD: 1 +.+.: (wq_completion)bond931 ->&rq->__lock ->(work_completion)(&(&slave->notify_work)->work) ->(work_completion)(&(&bond->mcast_work)->work) FD: 232 BD: 1 +.+.: (wq_completion)bond205#4 ->&rq->__lock ->(work_completion)(&(&slave->notify_work)->work) ->(work_completion)(&(&bond->mcast_work)->work) FD: 232 BD: 1 +.+.: (wq_completion)bond930#2 ->(work_completion)(&(&slave->notify_work)->work) ->(work_completion)(&(&bond->mcast_work)->work) ->&rq->__lock FD: 232 BD: 1 +.+.: (wq_completion)bond932 ->&rq->__lock ->(work_completion)(&(&slave->notify_work)->work) ->(work_completion)(&(&bond->mcast_work)->work) FD: 232 BD: 1 +.+.: (wq_completion)bond206#4 ->&rq->__lock ->(work_completion)(&(&slave->notify_work)->work) ->(work_completion)(&(&bond->mcast_work)->work) FD: 232 BD: 1 +.+.: (wq_completion)bond933 ->(work_completion)(&(&slave->notify_work)->work) ->(work_completion)(&(&bond->mcast_work)->work) ->&rq->__lock FD: 232 BD: 1 +.+.: (wq_completion)bond207#4 ->&rq->__lock ->(work_completion)(&(&slave->notify_work)->work) ->(work_completion)(&(&bond->mcast_work)->work) FD: 232 BD: 1 +.+.: (wq_completion)bond931#2 ->&rq->__lock ->(work_completion)(&(&slave->notify_work)->work) ->(work_completion)(&(&bond->mcast_work)->work) FD: 232 BD: 1 +.+.: (wq_completion)bond914#3 ->&rq->__lock ->(work_completion)(&(&slave->notify_work)->work) ->(work_completion)(&(&bond->mcast_work)->work) FD: 232 BD: 1 +.+.: (wq_completion)bond934 ->&rq->__lock ->(work_completion)(&(&slave->notify_work)->work) ->(work_completion)(&(&bond->mcast_work)->work) FD: 232 BD: 1 +.+.: (wq_completion)bond932#2 ->&rq->__lock ->(work_completion)(&(&slave->notify_work)->work) ->(work_completion)(&(&bond->mcast_work)->work) FD: 232 BD: 1 +.+.: (wq_completion)bond915#2 ->(work_completion)(&(&slave->notify_work)->work) ->(work_completion)(&(&bond->mcast_work)->work) ->&rq->__lock ->&cfs_rq->removed.lock ->&obj_hash[i].lock FD: 232 BD: 1 +.+.: (wq_completion)bond933#2 ->&rq->__lock ->(work_completion)(&(&slave->notify_work)->work) ->(work_completion)(&(&bond->mcast_work)->work) FD: 232 BD: 1 +.+.: (wq_completion)bond208#4 ->&rq->__lock ->(work_completion)(&(&slave->notify_work)->work) ->(work_completion)(&(&bond->mcast_work)->work) FD: 232 BD: 1 +.+.: (wq_completion)bond916#3 ->&rq->__lock ->(work_completion)(&(&slave->notify_work)->work) ->(work_completion)(&(&bond->mcast_work)->work) FD: 232 BD: 1 +.+.: (wq_completion)bond935 ->&rq->__lock ->(work_completion)(&(&slave->notify_work)->work) ->(work_completion)(&(&bond->mcast_work)->work) FD: 232 BD: 1 +.+.: (wq_completion)bond934#2 ->&rq->__lock ->(work_completion)(&(&slave->notify_work)->work) ->(work_completion)(&(&bond->mcast_work)->work) FD: 232 BD: 1 +.+.: (wq_completion)bond209#4 ->&rq->__lock ->(work_completion)(&(&slave->notify_work)->work) ->(work_completion)(&(&bond->mcast_work)->work) FD: 232 BD: 1 +.+.: (wq_completion)bond917#3 ->&rq->__lock ->(work_completion)(&(&slave->notify_work)->work) ->(work_completion)(&(&bond->mcast_work)->work) ->rcu_node_0 ->&rcu_state.expedited_wq FD: 232 BD: 1 +.+.: (wq_completion)bond935#2 ->(work_completion)(&(&slave->notify_work)->work) ->(work_completion)(&(&bond->mcast_work)->work) ->&rq->__lock FD: 232 BD: 1 +.+.: (wq_completion)bond210#4 ->&rq->__lock ->(work_completion)(&(&slave->notify_work)->work) ->(work_completion)(&(&bond->mcast_work)->work) FD: 232 BD: 1 +.+.: (wq_completion)bond918#3 ->(work_completion)(&(&slave->notify_work)->work) ->(work_completion)(&(&bond->mcast_work)->work) ->&rq->__lock FD: 232 BD: 1 +.+.: (wq_completion)bond936 ->(work_completion)(&(&slave->notify_work)->work) ->(work_completion)(&(&bond->mcast_work)->work) ->&rq->__lock FD: 232 BD: 1 +.+.: (wq_completion)bond936#2 ->&rq->__lock ->(work_completion)(&(&slave->notify_work)->work) ->(work_completion)(&(&bond->mcast_work)->work) FD: 232 BD: 1 +.+.: (wq_completion)bond211#4 ->&rq->__lock ->(work_completion)(&(&slave->notify_work)->work) ->(work_completion)(&(&bond->mcast_work)->work) FD: 232 BD: 1 +.+.: (wq_completion)bond919#3 ->&rq->__lock ->(work_completion)(&(&slave->notify_work)->work) ->(work_completion)(&(&bond->mcast_work)->work) FD: 232 BD: 1 +.+.: (wq_completion)bond937 ->&rq->__lock ->(work_completion)(&(&slave->notify_work)->work) ->(work_completion)(&(&bond->mcast_work)->work) FD: 232 BD: 1 +.+.: (wq_completion)bond937#2 ->&rq->__lock ->(work_completion)(&(&slave->notify_work)->work) ->(work_completion)(&(&bond->mcast_work)->work) FD: 232 BD: 1 +.+.: (wq_completion)bond212#4 ->&rq->__lock ->(work_completion)(&(&slave->notify_work)->work) ->(work_completion)(&(&bond->mcast_work)->work) FD: 232 BD: 1 +.+.: (wq_completion)bond213#3 ->&rq->__lock ->(work_completion)(&(&slave->notify_work)->work) ->(work_completion)(&(&bond->mcast_work)->work) FD: 232 BD: 1 +.+.: (wq_completion)bond938 ->(work_completion)(&(&slave->notify_work)->work) ->(work_completion)(&(&bond->mcast_work)->work) ->&rq->__lock ->&cfs_rq->removed.lock ->&obj_hash[i].lock FD: 232 BD: 1 +.+.: (wq_completion)bond938#2 ->&rq->__lock ->(work_completion)(&(&slave->notify_work)->work) ->(work_completion)(&(&bond->mcast_work)->work) FD: 232 BD: 1 +.+.: (wq_completion)bond920#3 ->&rq->__lock ->(work_completion)(&(&slave->notify_work)->work) ->(work_completion)(&(&bond->mcast_work)->work) FD: 232 BD: 1 +.+.: (wq_completion)bond939 ->&rq->__lock ->(work_completion)(&(&slave->notify_work)->work) ->(work_completion)(&(&bond->mcast_work)->work) FD: 232 BD: 1 +.+.: (wq_completion)bond214#4 ->&rq->__lock ->(work_completion)(&(&slave->notify_work)->work) ->(work_completion)(&(&bond->mcast_work)->work) FD: 232 BD: 1 +.+.: (wq_completion)bond939#2 ->(work_completion)(&(&slave->notify_work)->work) ->(work_completion)(&(&bond->mcast_work)->work) ->&rq->__lock FD: 232 BD: 1 +.+.: (wq_completion)bond921#3 ->&rq->__lock ->(work_completion)(&(&slave->notify_work)->work) ->(work_completion)(&(&bond->mcast_work)->work) FD: 232 BD: 1 +.+.: (wq_completion)bond940 ->(work_completion)(&(&slave->notify_work)->work) ->(work_completion)(&(&bond->mcast_work)->work) ->&rq->__lock FD: 232 BD: 1 +.+.: (wq_completion)bond940#2 ->&rq->__lock ->(work_completion)(&(&slave->notify_work)->work) ->(work_completion)(&(&bond->mcast_work)->work) FD: 232 BD: 1 +.+.: (wq_completion)bond215#4 ->&rq->__lock ->(work_completion)(&(&slave->notify_work)->work) ->(work_completion)(&(&bond->mcast_work)->work) FD: 232 BD: 1 +.+.: (wq_completion)bond922#3 ->&rq->__lock ->(work_completion)(&(&slave->notify_work)->work) ->(work_completion)(&(&bond->mcast_work)->work) FD: 232 BD: 1 +.+.: (wq_completion)bond941 ->(work_completion)(&(&slave->notify_work)->work) ->(work_completion)(&(&bond->mcast_work)->work) ->&rq->__lock FD: 232 BD: 1 +.+.: (wq_completion)bond941#2 ->&rq->__lock ->(work_completion)(&(&slave->notify_work)->work) ->(work_completion)(&(&bond->mcast_work)->work) FD: 232 BD: 1 +.+.: (wq_completion)bond923#3 ->&rq->__lock ->(work_completion)(&(&slave->notify_work)->work) ->(work_completion)(&(&bond->mcast_work)->work) FD: 232 BD: 1 +.+.: (wq_completion)bond942 ->&rq->__lock ->(work_completion)(&(&slave->notify_work)->work) ->(work_completion)(&(&bond->mcast_work)->work) FD: 232 BD: 1 +.+.: (wq_completion)bond942#2 ->&rq->__lock ->(work_completion)(&(&slave->notify_work)->work) ->(work_completion)(&(&bond->mcast_work)->work) FD: 232 BD: 1 +.+.: (wq_completion)bond924#3 ->(work_completion)(&(&slave->notify_work)->work) ->(work_completion)(&(&bond->mcast_work)->work) ->&rq->__lock FD: 232 BD: 1 +.+.: (wq_completion)bond943 ->&rq->__lock ->(work_completion)(&(&slave->notify_work)->work) ->(work_completion)(&(&bond->mcast_work)->work) FD: 232 BD: 1 +.+.: (wq_completion)bond943#2 ->(work_completion)(&(&slave->notify_work)->work) ->(work_completion)(&(&bond->mcast_work)->work) FD: 232 BD: 1 +.+.: (wq_completion)bond925#3 ->&rq->__lock ->(work_completion)(&(&slave->notify_work)->work) ->(work_completion)(&(&bond->mcast_work)->work) FD: 232 BD: 1 +.+.: (wq_completion)bond944 ->&rq->__lock ->(work_completion)(&(&slave->notify_work)->work) ->(work_completion)(&(&bond->mcast_work)->work) FD: 232 BD: 1 +.+.: (wq_completion)bond216#4 ->&rq->__lock ->(work_completion)(&(&slave->notify_work)->work) ->(work_completion)(&(&bond->mcast_work)->work) FD: 232 BD: 1 +.+.: (wq_completion)bond926#3 ->&rq->__lock ->(work_completion)(&(&slave->notify_work)->work) ->(work_completion)(&(&bond->mcast_work)->work) FD: 232 BD: 1 +.+.: (wq_completion)bond945 ->&rq->__lock ->(work_completion)(&(&slave->notify_work)->work) ->(work_completion)(&(&bond->mcast_work)->work) FD: 232 BD: 1 +.+.: (wq_completion)bond946 ->&rq->__lock ->(work_completion)(&(&slave->notify_work)->work) ->(work_completion)(&(&bond->mcast_work)->work) FD: 232 BD: 1 +.+.: (wq_completion)bond217#4 ->(work_completion)(&(&slave->notify_work)->work) ->(work_completion)(&(&bond->mcast_work)->work) ->&rq->__lock FD: 232 BD: 1 +.+.: (wq_completion)bond944#2 ->&rq->__lock ->(work_completion)(&(&slave->notify_work)->work) ->(work_completion)(&(&bond->mcast_work)->work) FD: 232 BD: 1 +.+.: (wq_completion)bond928#3 ->&rq->__lock ->(work_completion)(&(&slave->notify_work)->work) ->(work_completion)(&(&bond->mcast_work)->work) FD: 232 BD: 1 +.+.: (wq_completion)bond947 ->&rq->__lock ->(work_completion)(&(&slave->notify_work)->work) ->(work_completion)(&(&bond->mcast_work)->work) FD: 232 BD: 1 +.+.: (wq_completion)bond945#2 ->&rq->__lock ->(work_completion)(&(&slave->notify_work)->work) ->(work_completion)(&(&bond->mcast_work)->work) FD: 232 BD: 1 +.+.: (wq_completion)bond929#3 ->(work_completion)(&(&slave->notify_work)->work) ->(work_completion)(&(&bond->mcast_work)->work) ->&rq->__lock FD: 232 BD: 1 +.+.: (wq_completion)bond948 ->&rq->__lock ->(work_completion)(&(&slave->notify_work)->work) ->(work_completion)(&(&bond->mcast_work)->work) FD: 232 BD: 1 +.+.: (wq_completion)bond946#2 ->&rq->__lock ->(work_completion)(&(&slave->notify_work)->work) ->(work_completion)(&(&bond->mcast_work)->work) FD: 232 BD: 1 +.+.: (wq_completion)bond218#4 ->&rq->__lock ->(work_completion)(&(&slave->notify_work)->work) ->(work_completion)(&(&bond->mcast_work)->work) FD: 232 BD: 1 +.+.: (wq_completion)bond949 ->&rq->__lock ->(work_completion)(&(&slave->notify_work)->work) ->(work_completion)(&(&bond->mcast_work)->work) FD: 232 BD: 1 +.+.: (wq_completion)bond930#3 ->&rq->__lock ->(work_completion)(&(&slave->notify_work)->work) ->(work_completion)(&(&bond->mcast_work)->work) FD: 232 BD: 1 +.+.: (wq_completion)bond950 ->&rq->__lock ->(work_completion)(&(&slave->notify_work)->work) ->(work_completion)(&(&bond->mcast_work)->work) FD: 232 BD: 1 +.+.: (wq_completion)bond931#3 ->&rq->__lock ->(work_completion)(&(&slave->notify_work)->work) ->(work_completion)(&(&bond->mcast_work)->work) FD: 232 BD: 1 +.+.: (wq_completion)bond947#2 ->&rq->__lock ->(work_completion)(&(&slave->notify_work)->work) ->(work_completion)(&(&bond->mcast_work)->work) FD: 232 BD: 1 +.+.: (wq_completion)bond951 ->(work_completion)(&(&slave->notify_work)->work) ->(work_completion)(&(&bond->mcast_work)->work) ->&rq->__lock FD: 232 BD: 1 +.+.: (wq_completion)bond932#3 ->(work_completion)(&(&slave->notify_work)->work) ->(work_completion)(&(&bond->mcast_work)->work) ->&rq->__lock FD: 232 BD: 1 +.+.: (wq_completion)bond219#4 ->(work_completion)(&(&slave->notify_work)->work) ->(work_completion)(&(&bond->mcast_work)->work) ->&rq->__lock FD: 232 BD: 1 +.+.: (wq_completion)bond948#2 ->&rq->__lock ->(work_completion)(&(&slave->notify_work)->work) ->(work_completion)(&(&bond->mcast_work)->work) FD: 232 BD: 1 +.+.: (wq_completion)bond952 ->(work_completion)(&(&slave->notify_work)->work) ->(work_completion)(&(&bond->mcast_work)->work) ->&rq->__lock FD: 232 BD: 1 +.+.: (wq_completion)bond933#3 ->&rq->__lock ->(work_completion)(&(&slave->notify_work)->work) ->(work_completion)(&(&bond->mcast_work)->work) FD: 232 BD: 1 +.+.: (wq_completion)bond949#2 ->&rq->__lock ->(work_completion)(&(&slave->notify_work)->work) ->(work_completion)(&(&bond->mcast_work)->work) ->&cfs_rq->removed.lock FD: 232 BD: 1 +.+.: (wq_completion)bond953 ->&rq->__lock ->(work_completion)(&(&slave->notify_work)->work) ->(work_completion)(&(&bond->mcast_work)->work) FD: 232 BD: 1 +.+.: (wq_completion)bond220#4 ->&rq->__lock ->(work_completion)(&(&slave->notify_work)->work) ->(work_completion)(&(&bond->mcast_work)->work) FD: 232 BD: 1 +.+.: (wq_completion)bond950#2 ->&rq->__lock ->(work_completion)(&(&slave->notify_work)->work) ->(work_completion)(&(&bond->mcast_work)->work) FD: 232 BD: 1 +.+.: (wq_completion)bond954 ->(work_completion)(&(&slave->notify_work)->work) ->(work_completion)(&(&bond->mcast_work)->work) ->&rq->__lock FD: 232 BD: 1 +.+.: (wq_completion)bond221#4 ->&rq->__lock ->(work_completion)(&(&slave->notify_work)->work) ->(work_completion)(&(&bond->mcast_work)->work) FD: 232 BD: 1 +.+.: (wq_completion)bond934#3 ->&rq->__lock ->(work_completion)(&(&slave->notify_work)->work) ->(work_completion)(&(&bond->mcast_work)->work) FD: 232 BD: 1 +.+.: (wq_completion)bond951#2 ->&rq->__lock ->(work_completion)(&(&slave->notify_work)->work) ->(work_completion)(&(&bond->mcast_work)->work) FD: 232 BD: 1 +.+.: (wq_completion)bond222#4 ->(work_completion)(&(&slave->notify_work)->work) ->(work_completion)(&(&bond->mcast_work)->work) ->&rq->__lock FD: 232 BD: 1 +.+.: (wq_completion)bond935#3 ->&rq->__lock ->(work_completion)(&(&slave->notify_work)->work) ->(work_completion)(&(&bond->mcast_work)->work) FD: 232 BD: 1 +.+.: (wq_completion)bond952#2 ->(work_completion)(&(&slave->notify_work)->work) ->(work_completion)(&(&bond->mcast_work)->work) ->&rq->__lock FD: 232 BD: 1 +.+.: (wq_completion)bond223#4 ->&rq->__lock ->(work_completion)(&(&slave->notify_work)->work) ->(work_completion)(&(&bond->mcast_work)->work) FD: 232 BD: 1 +.+.: (wq_completion)bond953#2 ->(work_completion)(&(&slave->notify_work)->work) ->(work_completion)(&(&bond->mcast_work)->work) ->&rq->__lock FD: 232 BD: 1 +.+.: (wq_completion)bond936#3 ->&rq->__lock ->(work_completion)(&(&slave->notify_work)->work) ->(work_completion)(&(&bond->mcast_work)->work) FD: 232 BD: 1 +.+.: (wq_completion)bond224#4 ->(work_completion)(&(&slave->notify_work)->work) ->(work_completion)(&(&bond->mcast_work)->work) ->&rq->__lock FD: 232 BD: 1 +.+.: (wq_completion)bond954#2 ->(work_completion)(&(&slave->notify_work)->work) ->(work_completion)(&(&bond->mcast_work)->work) FD: 232 BD: 1 +.+.: (wq_completion)bond937#3 ->&rq->__lock ->(work_completion)(&(&slave->notify_work)->work) ->(work_completion)(&(&bond->mcast_work)->work) FD: 232 BD: 1 +.+.: (wq_completion)bond955 ->&rq->__lock ->(work_completion)(&(&slave->notify_work)->work) ->(work_completion)(&(&bond->mcast_work)->work) FD: 232 BD: 1 +.+.: (wq_completion)bond955#2 ->&rq->__lock ->(work_completion)(&(&slave->notify_work)->work) ->(work_completion)(&(&bond->mcast_work)->work) FD: 232 BD: 1 +.+.: (wq_completion)bond938#3 ->(work_completion)(&(&slave->notify_work)->work) ->(work_completion)(&(&bond->mcast_work)->work) ->&rq->__lock FD: 232 BD: 1 +.+.: (wq_completion)bond956 ->&rq->__lock ->(work_completion)(&(&slave->notify_work)->work) ->(work_completion)(&(&bond->mcast_work)->work) FD: 232 BD: 1 +.+.: (wq_completion)bond225#4 ->&rq->__lock ->(work_completion)(&(&slave->notify_work)->work) ->(work_completion)(&(&bond->mcast_work)->work) FD: 232 BD: 1 +.+.: (wq_completion)bond939#3 ->(work_completion)(&(&slave->notify_work)->work) ->(work_completion)(&(&bond->mcast_work)->work) ->&rq->__lock FD: 232 BD: 1 +.+.: (wq_completion)bond956#2 ->&rq->__lock ->(work_completion)(&(&slave->notify_work)->work) ->(work_completion)(&(&bond->mcast_work)->work) FD: 232 BD: 1 +.+.: (wq_completion)bond957 ->&rq->__lock ->(work_completion)(&(&slave->notify_work)->work) ->(work_completion)(&(&bond->mcast_work)->work) FD: 232 BD: 1 +.+.: (wq_completion)bond226#4 ->(work_completion)(&(&slave->notify_work)->work) ->(work_completion)(&(&bond->mcast_work)->work) ->&rq->__lock FD: 232 BD: 1 +.+.: (wq_completion)bond958 ->(work_completion)(&(&slave->notify_work)->work) ->(work_completion)(&(&bond->mcast_work)->work) ->&rq->__lock FD: 232 BD: 1 +.+.: (wq_completion)bond227#4 ->&rq->__lock ->(work_completion)(&(&slave->notify_work)->work) ->(work_completion)(&(&bond->mcast_work)->work) FD: 232 BD: 1 +.+.: (wq_completion)bond940#3 ->&rq->__lock ->(work_completion)(&(&slave->notify_work)->work) ->(work_completion)(&(&bond->mcast_work)->work) FD: 232 BD: 1 +.+.: (wq_completion)bond957#2 ->(work_completion)(&(&slave->notify_work)->work) ->(work_completion)(&(&bond->mcast_work)->work) ->&rq->__lock FD: 232 BD: 1 +.+.: (wq_completion)bond959 ->(work_completion)(&(&slave->notify_work)->work) ->(work_completion)(&(&bond->mcast_work)->work) ->&rq->__lock FD: 232 BD: 1 +.+.: (wq_completion)bond941#3 ->(work_completion)(&(&slave->notify_work)->work) ->(work_completion)(&(&bond->mcast_work)->work) ->&rq->__lock FD: 232 BD: 1 +.+.: (wq_completion)bond958#2 ->&rq->__lock ->(work_completion)(&(&slave->notify_work)->work) ->(work_completion)(&(&bond->mcast_work)->work) FD: 232 BD: 1 +.+.: (wq_completion)bond960 ->(work_completion)(&(&slave->notify_work)->work) ->(work_completion)(&(&bond->mcast_work)->work) ->&rq->__lock FD: 232 BD: 1 +.+.: (wq_completion)bond228#4 ->(work_completion)(&(&slave->notify_work)->work) ->(work_completion)(&(&bond->mcast_work)->work) ->&rq->__lock FD: 232 BD: 1 +.+.: (wq_completion)bond942#3 ->&rq->__lock ->(work_completion)(&(&slave->notify_work)->work) ->(work_completion)(&(&bond->mcast_work)->work) FD: 232 BD: 1 +.+.: (wq_completion)bond961 ->&rq->__lock ->(work_completion)(&(&slave->notify_work)->work) ->(work_completion)(&(&bond->mcast_work)->work) FD: 232 BD: 1 +.+.: (wq_completion)bond229#4 ->&rq->__lock ->(work_completion)(&(&slave->notify_work)->work) ->(work_completion)(&(&bond->mcast_work)->work) FD: 232 BD: 1 +.+.: (wq_completion)bond943#3 ->(work_completion)(&(&slave->notify_work)->work) ->(work_completion)(&(&bond->mcast_work)->work) ->&rq->__lock FD: 232 BD: 1 +.+.: (wq_completion)bond959#2 ->&rq->__lock ->(work_completion)(&(&slave->notify_work)->work) ->(work_completion)(&(&bond->mcast_work)->work) FD: 232 BD: 1 +.+.: (wq_completion)bond962 ->&rq->__lock ->(work_completion)(&(&slave->notify_work)->work) ->(work_completion)(&(&bond->mcast_work)->work) FD: 232 BD: 1 +.+.: (wq_completion)bond944#3 ->&rq->__lock ->(work_completion)(&(&slave->notify_work)->work) ->(work_completion)(&(&bond->mcast_work)->work) FD: 232 BD: 1 +.+.: (wq_completion)bond960#2 ->(work_completion)(&(&slave->notify_work)->work) ->(work_completion)(&(&bond->mcast_work)->work) ->&rq->__lock FD: 232 BD: 1 +.+.: (wq_completion)bond963 ->&rq->__lock ->(work_completion)(&(&slave->notify_work)->work) ->(work_completion)(&(&bond->mcast_work)->work) FD: 232 BD: 1 +.+.: (wq_completion)bond945#3 ->&rq->__lock ->(work_completion)(&(&slave->notify_work)->work) ->(work_completion)(&(&bond->mcast_work)->work) FD: 232 BD: 1 +.+.: (wq_completion)bond961#2 ->&rq->__lock ->(work_completion)(&(&slave->notify_work)->work) ->(work_completion)(&(&bond->mcast_work)->work) FD: 232 BD: 1 +.+.: (wq_completion)bond964 ->&rq->__lock ->(work_completion)(&(&slave->notify_work)->work) ->(work_completion)(&(&bond->mcast_work)->work) FD: 232 BD: 1 +.+.: (wq_completion)bond230#4 ->(work_completion)(&(&slave->notify_work)->work) ->(work_completion)(&(&bond->mcast_work)->work) ->&rq->__lock FD: 232 BD: 1 +.+.: (wq_completion)bond946#3 ->(work_completion)(&(&slave->notify_work)->work) ->(work_completion)(&(&bond->mcast_work)->work) ->&rq->__lock FD: 232 BD: 1 +.+.: (wq_completion)bond962#2 ->&rq->__lock ->(work_completion)(&(&slave->notify_work)->work) ->(work_completion)(&(&bond->mcast_work)->work) FD: 232 BD: 1 +.+.: (wq_completion)bond963#2 ->(work_completion)(&(&slave->notify_work)->work) ->(work_completion)(&(&bond->mcast_work)->work) ->&rq->__lock FD: 232 BD: 1 +.+.: (wq_completion)bond965 ->&rq->__lock ->(work_completion)(&(&slave->notify_work)->work) ->(work_completion)(&(&bond->mcast_work)->work) FD: 232 BD: 1 +.+.: (wq_completion)bond231#4 ->(work_completion)(&(&slave->notify_work)->work) ->(work_completion)(&(&bond->mcast_work)->work) ->&rq->__lock FD: 232 BD: 1 +.+.: (wq_completion)bond947#3 ->(work_completion)(&(&slave->notify_work)->work) ->(work_completion)(&(&bond->mcast_work)->work) ->&rq->__lock FD: 232 BD: 1 +.+.: (wq_completion)bond964#2 ->(work_completion)(&(&slave->notify_work)->work) ->(work_completion)(&(&bond->mcast_work)->work) ->&rq->__lock FD: 232 BD: 1 +.+.: (wq_completion)bond966 ->&rq->__lock ->(work_completion)(&(&slave->notify_work)->work) ->(work_completion)(&(&bond->mcast_work)->work) FD: 232 BD: 1 +.+.: (wq_completion)bond948#3 ->&rq->__lock ->(work_completion)(&(&slave->notify_work)->work) ->(work_completion)(&(&bond->mcast_work)->work) FD: 232 BD: 1 +.+.: (wq_completion)bond965#2 ->&rq->__lock ->(work_completion)(&(&slave->notify_work)->work) ->(work_completion)(&(&bond->mcast_work)->work) FD: 232 BD: 1 +.+.: (wq_completion)bond967 ->&rq->__lock ->(work_completion)(&(&slave->notify_work)->work) ->(work_completion)(&(&bond->mcast_work)->work) FD: 232 BD: 1 +.+.: (wq_completion)bond949#3 ->&rq->__lock ->(work_completion)(&(&slave->notify_work)->work) ->(work_completion)(&(&bond->mcast_work)->work) FD: 232 BD: 1 +.+.: (wq_completion)bond966#2 ->&rq->__lock ->(work_completion)(&(&slave->notify_work)->work) ->(work_completion)(&(&bond->mcast_work)->work) FD: 232 BD: 1 +.+.: (wq_completion)bond968 ->&rq->__lock ->(work_completion)(&(&slave->notify_work)->work) ->(work_completion)(&(&bond->mcast_work)->work) FD: 232 BD: 1 +.+.: (wq_completion)bond950#3 ->&rq->__lock ->(work_completion)(&(&slave->notify_work)->work) ->(work_completion)(&(&bond->mcast_work)->work) FD: 232 BD: 1 +.+.: (wq_completion)bond967#2 ->(work_completion)(&(&slave->notify_work)->work) ->(work_completion)(&(&bond->mcast_work)->work) ->&rq->__lock FD: 232 BD: 1 +.+.: (wq_completion)bond969 ->&rq->__lock ->(work_completion)(&(&slave->notify_work)->work) ->(work_completion)(&(&bond->mcast_work)->work) FD: 232 BD: 1 +.+.: (wq_completion)bond232#4 ->(work_completion)(&(&slave->notify_work)->work) ->(work_completion)(&(&bond->mcast_work)->work) ->&rq->__lock FD: 232 BD: 1 +.+.: (wq_completion)bond968#2 ->&rq->__lock ->(work_completion)(&(&slave->notify_work)->work) ->(work_completion)(&(&bond->mcast_work)->work) FD: 232 BD: 1 +.+.: (wq_completion)bond951#3 ->&rq->__lock ->(work_completion)(&(&slave->notify_work)->work) ->(work_completion)(&(&bond->mcast_work)->work) FD: 232 BD: 1 +.+.: (wq_completion)bond970 ->(work_completion)(&(&slave->notify_work)->work) ->(work_completion)(&(&bond->mcast_work)->work) ->&rq->__lock FD: 232 BD: 1 +.+.: (wq_completion)bond233#3 ->&rq->__lock ->(work_completion)(&(&slave->notify_work)->work) ->(work_completion)(&(&bond->mcast_work)->work) FD: 232 BD: 1 +.+.: (wq_completion)bond971 ->(work_completion)(&(&slave->notify_work)->work) ->(work_completion)(&(&bond->mcast_work)->work) FD: 232 BD: 1 +.+.: (wq_completion)bond952#3 ->(work_completion)(&(&slave->notify_work)->work) ->(work_completion)(&(&bond->mcast_work)->work) ->&rq->__lock FD: 232 BD: 1 +.+.: (wq_completion)bond969#2 ->&rq->__lock ->(work_completion)(&(&slave->notify_work)->work) ->(work_completion)(&(&bond->mcast_work)->work) FD: 232 BD: 1 +.+.: (wq_completion)bond972 ->&rq->__lock ->(work_completion)(&(&slave->notify_work)->work) ->(work_completion)(&(&bond->mcast_work)->work) FD: 232 BD: 1 +.+.: (wq_completion)bond953#3 ->&rq->__lock ->(work_completion)(&(&slave->notify_work)->work) ->(work_completion)(&(&bond->mcast_work)->work) FD: 232 BD: 1 +.+.: (wq_completion)bond970#2 ->&rq->__lock ->(work_completion)(&(&slave->notify_work)->work) ->(work_completion)(&(&bond->mcast_work)->work) FD: 232 BD: 1 +.+.: (wq_completion)bond973 ->&rq->__lock ->(work_completion)(&(&slave->notify_work)->work) ->(work_completion)(&(&bond->mcast_work)->work) FD: 232 BD: 1 +.+.: (wq_completion)bond954#3 ->&rq->__lock ->(work_completion)(&(&slave->notify_work)->work) ->(work_completion)(&(&bond->mcast_work)->work) FD: 232 BD: 1 +.+.: (wq_completion)bond971#2 ->&rq->__lock ->(work_completion)(&(&slave->notify_work)->work) ->(work_completion)(&(&bond->mcast_work)->work) FD: 232 BD: 1 +.+.: (wq_completion)bond234#4 ->&rq->__lock ->(work_completion)(&(&slave->notify_work)->work) ->(work_completion)(&(&bond->mcast_work)->work) FD: 232 BD: 1 +.+.: (wq_completion)bond974 ->&rq->__lock ->(work_completion)(&(&slave->notify_work)->work) ->(work_completion)(&(&bond->mcast_work)->work) FD: 232 BD: 1 +.+.: (wq_completion)bond955#3 ->&rq->__lock ->(work_completion)(&(&slave->notify_work)->work) ->(work_completion)(&(&bond->mcast_work)->work) FD: 232 BD: 1 +.+.: (wq_completion)bond972#2 ->(work_completion)(&(&slave->notify_work)->work) ->(work_completion)(&(&bond->mcast_work)->work) ->&rq->__lock FD: 232 BD: 1 +.+.: (wq_completion)bond975 ->(work_completion)(&(&slave->notify_work)->work) ->(work_completion)(&(&bond->mcast_work)->work) ->&rq->__lock FD: 232 BD: 1 +.+.: (wq_completion)bond956#3 ->(work_completion)(&(&slave->notify_work)->work) ->(work_completion)(&(&bond->mcast_work)->work) ->&rq->__lock FD: 232 BD: 1 +.+.: (wq_completion)bond976 ->&rq->__lock ->(work_completion)(&(&slave->notify_work)->work) ->(work_completion)(&(&bond->mcast_work)->work) ->&cfs_rq->removed.lock ->&obj_hash[i].lock FD: 232 BD: 1 +.+.: (wq_completion)bond957#3 ->&rq->__lock ->(work_completion)(&(&slave->notify_work)->work) ->(work_completion)(&(&bond->mcast_work)->work) FD: 232 BD: 1 +.+.: (wq_completion)bond974#2 ->&rq->__lock ->(work_completion)(&(&slave->notify_work)->work) ->(work_completion)(&(&bond->mcast_work)->work) FD: 232 BD: 1 +.+.: (wq_completion)bond958#3 ->&rq->__lock ->(work_completion)(&(&slave->notify_work)->work) ->(work_completion)(&(&bond->mcast_work)->work) FD: 232 BD: 1 +.+.: (wq_completion)bond975#2 ->&rq->__lock ->(work_completion)(&(&slave->notify_work)->work) ->(work_completion)(&(&bond->mcast_work)->work) FD: 232 BD: 1 +.+.: (wq_completion)bond235#4 ->(work_completion)(&(&slave->notify_work)->work) ->(work_completion)(&(&bond->mcast_work)->work) ->&rq->__lock FD: 232 BD: 1 +.+.: (wq_completion)bond976#2 ->(work_completion)(&(&slave->notify_work)->work) ->(work_completion)(&(&bond->mcast_work)->work) ->&rq->__lock FD: 232 BD: 1 +.+.: (wq_completion)bond236#3 ->&rq->__lock ->(work_completion)(&(&slave->notify_work)->work) ->(work_completion)(&(&bond->mcast_work)->work) FD: 232 BD: 1 +.+.: (wq_completion)bond959#3 ->&rq->__lock ->(work_completion)(&(&slave->notify_work)->work) ->(work_completion)(&(&bond->mcast_work)->work) FD: 232 BD: 1 +.+.: (wq_completion)bond977 ->&rq->__lock ->(work_completion)(&(&slave->notify_work)->work) ->(work_completion)(&(&bond->mcast_work)->work) FD: 232 BD: 1 +.+.: (wq_completion)bond977#2 ->(work_completion)(&(&slave->notify_work)->work) ->(work_completion)(&(&bond->mcast_work)->work) ->&rq->__lock FD: 232 BD: 1 +.+.: (wq_completion)bond960#3 ->(work_completion)(&(&slave->notify_work)->work) ->(work_completion)(&(&bond->mcast_work)->work) ->&rq->__lock FD: 232 BD: 1 +.+.: (wq_completion)bond978 ->&rq->__lock ->(work_completion)(&(&slave->notify_work)->work) ->(work_completion)(&(&bond->mcast_work)->work) FD: 232 BD: 1 +.+.: (wq_completion)bond978#2 ->&rq->__lock ->(work_completion)(&(&slave->notify_work)->work) ->(work_completion)(&(&bond->mcast_work)->work) FD: 232 BD: 1 +.+.: (wq_completion)bond961#3 ->&rq->__lock ->(work_completion)(&(&slave->notify_work)->work) ->(work_completion)(&(&bond->mcast_work)->work) FD: 232 BD: 1 +.+.: (wq_completion)bond237#4 ->(work_completion)(&(&slave->notify_work)->work) ->(work_completion)(&(&bond->mcast_work)->work) ->&rq->__lock FD: 232 BD: 1 +.+.: (wq_completion)bond962#3 ->&rq->__lock ->(work_completion)(&(&slave->notify_work)->work) ->(work_completion)(&(&bond->mcast_work)->work) FD: 232 BD: 1 +.+.: (wq_completion)bond238#4 ->&rq->__lock ->(work_completion)(&(&slave->notify_work)->work) ->(work_completion)(&(&bond->mcast_work)->work) FD: 232 BD: 1 +.+.: (wq_completion)bond979 ->&rq->__lock ->(work_completion)(&(&slave->notify_work)->work) ->(work_completion)(&(&bond->mcast_work)->work) FD: 232 BD: 1 +.+.: (wq_completion)bond979#2 ->(work_completion)(&(&slave->notify_work)->work) ->(work_completion)(&(&bond->mcast_work)->work) ->&rq->__lock FD: 232 BD: 1 +.+.: (wq_completion)bond980 ->&rq->__lock ->(work_completion)(&(&slave->notify_work)->work) ->(work_completion)(&(&bond->mcast_work)->work) FD: 232 BD: 1 +.+.: (wq_completion)bond980#2 ->(work_completion)(&(&slave->notify_work)->work) ->(work_completion)(&(&bond->mcast_work)->work) ->&rq->__lock FD: 232 BD: 1 +.+.: (wq_completion)bond963#3 ->&rq->__lock ->(work_completion)(&(&slave->notify_work)->work) ->(work_completion)(&(&bond->mcast_work)->work) FD: 232 BD: 1 +.+.: (wq_completion)bond981 ->&rq->__lock ->(work_completion)(&(&slave->notify_work)->work) ->(work_completion)(&(&bond->mcast_work)->work) FD: 232 BD: 1 +.+.: (wq_completion)bond981#2 ->&rq->__lock ->(work_completion)(&(&slave->notify_work)->work) ->(work_completion)(&(&bond->mcast_work)->work) FD: 232 BD: 1 +.+.: (wq_completion)bond239#4 ->&rq->__lock ->(work_completion)(&(&slave->notify_work)->work) ->(work_completion)(&(&bond->mcast_work)->work) FD: 232 BD: 1 +.+.: (wq_completion)bond982 ->&rq->__lock ->(work_completion)(&(&slave->notify_work)->work) ->(work_completion)(&(&bond->mcast_work)->work) FD: 232 BD: 1 +.+.: (wq_completion)bond983 ->&rq->__lock ->(work_completion)(&(&slave->notify_work)->work) ->(work_completion)(&(&bond->mcast_work)->work) FD: 232 BD: 1 +.+.: (wq_completion)bond964#3 ->(work_completion)(&(&slave->notify_work)->work) ->(work_completion)(&(&bond->mcast_work)->work) ->&rq->__lock FD: 232 BD: 1 +.+.: (wq_completion)bond240#4 ->&rq->__lock ->(work_completion)(&(&slave->notify_work)->work) ->(work_completion)(&(&bond->mcast_work)->work) FD: 232 BD: 1 +.+.: (wq_completion)bond982#2 ->&rq->__lock ->(work_completion)(&(&slave->notify_work)->work) ->(work_completion)(&(&bond->mcast_work)->work) FD: 232 BD: 1 +.+.: (wq_completion)bond984 ->&rq->__lock ->(work_completion)(&(&slave->notify_work)->work) ->(work_completion)(&(&bond->mcast_work)->work) FD: 232 BD: 1 +.+.: (wq_completion)bond241#4 ->&rq->__lock ->(work_completion)(&(&slave->notify_work)->work) ->(work_completion)(&(&bond->mcast_work)->work) FD: 232 BD: 1 +.+.: (wq_completion)bond983#2 ->&rq->__lock ->(work_completion)(&(&slave->notify_work)->work) ->(work_completion)(&(&bond->mcast_work)->work) FD: 232 BD: 1 +.+.: (wq_completion)bond985 ->&rq->__lock ->(work_completion)(&(&slave->notify_work)->work) ->(work_completion)(&(&bond->mcast_work)->work) FD: 232 BD: 1 +.+.: (wq_completion)bond242#4 ->&rq->__lock ->(work_completion)(&(&slave->notify_work)->work) ->(work_completion)(&(&bond->mcast_work)->work) FD: 232 BD: 1 +.+.: (wq_completion)bond986 ->&rq->__lock ->(work_completion)(&(&slave->notify_work)->work) ->(work_completion)(&(&bond->mcast_work)->work) ->&cfs_rq->removed.lock ->&obj_hash[i].lock FD: 232 BD: 1 +.+.: (wq_completion)bond984#2 ->&rq->__lock ->(work_completion)(&(&slave->notify_work)->work) ->(work_completion)(&(&bond->mcast_work)->work) FD: 232 BD: 1 +.+.: (wq_completion)bond243#4 ->&rq->__lock ->(work_completion)(&(&slave->notify_work)->work) ->(work_completion)(&(&bond->mcast_work)->work) FD: 232 BD: 1 +.+.: (wq_completion)bond987 ->&rq->__lock ->(work_completion)(&(&slave->notify_work)->work) ->(work_completion)(&(&bond->mcast_work)->work) FD: 232 BD: 1 +.+.: (wq_completion)bond965#3 ->&rq->__lock ->(work_completion)(&(&slave->notify_work)->work) ->(work_completion)(&(&bond->mcast_work)->work) FD: 232 BD: 1 +.+.: (wq_completion)bond988 ->&rq->__lock ->(work_completion)(&(&slave->notify_work)->work) ->(work_completion)(&(&bond->mcast_work)->work) ->&cfs_rq->removed.lock ->&obj_hash[i].lock FD: 232 BD: 1 +.+.: (wq_completion)bond966#3 ->&rq->__lock ->(work_completion)(&(&slave->notify_work)->work) ->(work_completion)(&(&bond->mcast_work)->work) FD: 232 BD: 1 +.+.: (wq_completion)bond989 ->(work_completion)(&(&slave->notify_work)->work) ->(work_completion)(&(&bond->mcast_work)->work) ->&rq->__lock FD: 232 BD: 1 +.+.: (wq_completion)bond985#2 ->(work_completion)(&(&slave->notify_work)->work) ->(work_completion)(&(&bond->mcast_work)->work) ->&rq->__lock FD: 232 BD: 1 +.+.: (wq_completion)bond967#3 ->&rq->__lock ->(work_completion)(&(&slave->notify_work)->work) ->(work_completion)(&(&bond->mcast_work)->work) FD: 232 BD: 1 +.+.: (wq_completion)bond990 ->&rq->__lock ->(work_completion)(&(&slave->notify_work)->work) ->(work_completion)(&(&bond->mcast_work)->work) FD: 232 BD: 1 +.+.: (wq_completion)bond986#2 ->&rq->__lock ->(work_completion)(&(&slave->notify_work)->work) ->(work_completion)(&(&bond->mcast_work)->work) FD: 232 BD: 1 +.+.: (wq_completion)bond968#3 ->(work_completion)(&(&slave->notify_work)->work) ->(work_completion)(&(&bond->mcast_work)->work) ->&rq->__lock ->&cfs_rq->removed.lock ->&obj_hash[i].lock FD: 232 BD: 1 +.+.: (wq_completion)bond991 ->&rq->__lock ->(work_completion)(&(&slave->notify_work)->work) ->(work_completion)(&(&bond->mcast_work)->work) FD: 232 BD: 1 +.+.: (wq_completion)bond987#2 ->(work_completion)(&(&slave->notify_work)->work) ->(work_completion)(&(&bond->mcast_work)->work) ->&rq->__lock FD: 232 BD: 1 +.+.: (wq_completion)bond969#3 ->&rq->__lock ->(work_completion)(&(&slave->notify_work)->work) ->(work_completion)(&(&bond->mcast_work)->work) FD: 232 BD: 1 +.+.: (wq_completion)bond244#4 ->&rq->__lock ->(work_completion)(&(&slave->notify_work)->work) ->(work_completion)(&(&bond->mcast_work)->work) FD: 232 BD: 1 +.+.: (wq_completion)bond992 ->&rq->__lock ->(work_completion)(&(&slave->notify_work)->work) ->(work_completion)(&(&bond->mcast_work)->work) FD: 232 BD: 1 +.+.: (wq_completion)bond988#2 ->&rq->__lock ->(work_completion)(&(&slave->notify_work)->work) ->(work_completion)(&(&bond->mcast_work)->work) FD: 232 BD: 1 +.+.: (wq_completion)bond245#4 ->&rq->__lock ->(work_completion)(&(&slave->notify_work)->work) ->(work_completion)(&(&bond->mcast_work)->work) FD: 232 BD: 1 +.+.: (wq_completion)bond970#3 ->(work_completion)(&(&slave->notify_work)->work) ->(work_completion)(&(&bond->mcast_work)->work) ->&rq->__lock FD: 232 BD: 1 +.+.: (wq_completion)bond993 ->(work_completion)(&(&slave->notify_work)->work) ->(work_completion)(&(&bond->mcast_work)->work) ->&rq->__lock FD: 232 BD: 1 +.+.: (wq_completion)bond989#2 ->&rq->__lock ->(work_completion)(&(&slave->notify_work)->work) ->(work_completion)(&(&bond->mcast_work)->work) FD: 232 BD: 1 +.+.: (wq_completion)bond971#3 ->&rq->__lock ->(work_completion)(&(&slave->notify_work)->work) ->(work_completion)(&(&bond->mcast_work)->work) FD: 232 BD: 1 +.+.: (wq_completion)bond994 ->&rq->__lock ->(work_completion)(&(&slave->notify_work)->work) ->(work_completion)(&(&bond->mcast_work)->work) FD: 232 BD: 1 +.+.: (wq_completion)bond990#2 ->&rq->__lock ->(work_completion)(&(&slave->notify_work)->work) ->(work_completion)(&(&bond->mcast_work)->work) FD: 232 BD: 1 +.+.: (wq_completion)bond972#3 ->&rq->__lock ->(work_completion)(&(&slave->notify_work)->work) ->(work_completion)(&(&bond->mcast_work)->work) FD: 232 BD: 1 +.+.: (wq_completion)bond995 ->&rq->__lock ->(work_completion)(&(&slave->notify_work)->work) ->(work_completion)(&(&bond->mcast_work)->work) FD: 232 BD: 1 +.+.: (wq_completion)bond991#2 ->(work_completion)(&(&slave->notify_work)->work) ->(work_completion)(&(&bond->mcast_work)->work) ->&rq->__lock FD: 232 BD: 1 +.+.: (wq_completion)bond246#3 ->&rq->__lock ->(work_completion)(&(&slave->notify_work)->work) ->(work_completion)(&(&bond->mcast_work)->work) FD: 232 BD: 1 +.+.: (wq_completion)bond973#2 ->&rq->__lock ->(work_completion)(&(&slave->notify_work)->work) ->(work_completion)(&(&bond->mcast_work)->work) FD: 232 BD: 1 +.+.: (wq_completion)bond996 ->(work_completion)(&(&slave->notify_work)->work) ->(work_completion)(&(&bond->mcast_work)->work) ->&rq->__lock FD: 232 BD: 1 +.+.: (wq_completion)bond247#4 ->&rq->__lock ->(work_completion)(&(&slave->notify_work)->work) ->(work_completion)(&(&bond->mcast_work)->work) FD: 232 BD: 1 +.+.: (wq_completion)bond992#2 ->&rq->__lock ->(work_completion)(&(&slave->notify_work)->work) ->(work_completion)(&(&bond->mcast_work)->work) FD: 232 BD: 1 +.+.: (wq_completion)bond997 ->(work_completion)(&(&slave->notify_work)->work) ->(work_completion)(&(&bond->mcast_work)->work) ->&rq->__lock FD: 232 BD: 1 +.+.: (wq_completion)bond998 ->&rq->__lock ->(work_completion)(&(&slave->notify_work)->work) ->(work_completion)(&(&bond->mcast_work)->work) FD: 232 BD: 1 +.+.: (wq_completion)bond993#2 ->(work_completion)(&(&slave->notify_work)->work) ->(work_completion)(&(&bond->mcast_work)->work) ->&rq->__lock ->&cfs_rq->removed.lock ->&obj_hash[i].lock FD: 232 BD: 1 +.+.: (wq_completion)bond974#3 ->(work_completion)(&(&slave->notify_work)->work) ->(work_completion)(&(&bond->mcast_work)->work) ->&rq->__lock FD: 232 BD: 1 +.+.: (wq_completion)bond999 ->(work_completion)(&(&slave->notify_work)->work) ->(work_completion)(&(&bond->mcast_work)->work) ->&rq->__lock FD: 232 BD: 1 +.+.: (wq_completion)bond994#2 ->(work_completion)(&(&slave->notify_work)->work) ->(work_completion)(&(&bond->mcast_work)->work) ->&rq->__lock FD: 232 BD: 1 +.+.: (wq_completion)bond975#3 ->&rq->__lock ->(work_completion)(&(&slave->notify_work)->work) ->(work_completion)(&(&bond->mcast_work)->work) FD: 232 BD: 1 +.+.: (wq_completion)bond1000 ->(work_completion)(&(&slave->notify_work)->work) ->(work_completion)(&(&bond->mcast_work)->work) ->&rq->__lock FD: 232 BD: 1 +.+.: (wq_completion)bond976#3 ->(work_completion)(&(&slave->notify_work)->work) ->(work_completion)(&(&bond->mcast_work)->work) ->&rq->__lock FD: 232 BD: 1 +.+.: (wq_completion)bond996#2 ->(work_completion)(&(&slave->notify_work)->work) ->(work_completion)(&(&bond->mcast_work)->work) ->&rq->__lock FD: 232 BD: 1 +.+.: (wq_completion)bond977#3 ->&rq->__lock ->(work_completion)(&(&slave->notify_work)->work) ->(work_completion)(&(&bond->mcast_work)->work) FD: 232 BD: 1 +.+.: (wq_completion)bond1001 ->&rq->__lock ->(work_completion)(&(&slave->notify_work)->work) ->(work_completion)(&(&bond->mcast_work)->work) FD: 232 BD: 1 +.+.: (wq_completion)bond997#2 ->(work_completion)(&(&slave->notify_work)->work) ->(work_completion)(&(&bond->mcast_work)->work) FD: 232 BD: 1 +.+.: (wq_completion)bond978#3 ->&rq->__lock ->(work_completion)(&(&slave->notify_work)->work) ->(work_completion)(&(&bond->mcast_work)->work) FD: 232 BD: 1 +.+.: (wq_completion)bond1002 ->&rq->__lock ->(work_completion)(&(&slave->notify_work)->work) ->(work_completion)(&(&bond->mcast_work)->work) ->&cfs_rq->removed.lock ->&obj_hash[i].lock FD: 232 BD: 1 +.+.: (wq_completion)bond248#4 ->(work_completion)(&(&slave->notify_work)->work) ->(work_completion)(&(&bond->mcast_work)->work) ->&rq->__lock FD: 232 BD: 1 +.+.: (wq_completion)bond979#3 ->&rq->__lock ->(work_completion)(&(&slave->notify_work)->work) ->(work_completion)(&(&bond->mcast_work)->work) FD: 232 BD: 1 +.+.: (wq_completion)bond1003 ->&rq->__lock ->(work_completion)(&(&slave->notify_work)->work) ->(work_completion)(&(&bond->mcast_work)->work) ->&cfs_rq->removed.lock ->&obj_hash[i].lock FD: 232 BD: 1 +.+.: (wq_completion)bond249#4 ->&rq->__lock ->(work_completion)(&(&slave->notify_work)->work) ->(work_completion)(&(&bond->mcast_work)->work) FD: 232 BD: 1 +.+.: (wq_completion)bond998#2 ->&rq->__lock ->(work_completion)(&(&slave->notify_work)->work) ->(work_completion)(&(&bond->mcast_work)->work) FD: 232 BD: 1 +.+.: (wq_completion)bond980#3 ->&rq->__lock ->(work_completion)(&(&slave->notify_work)->work) ->(work_completion)(&(&bond->mcast_work)->work) FD: 232 BD: 1 +.+.: (wq_completion)bond1004 ->(work_completion)(&(&slave->notify_work)->work) ->(work_completion)(&(&bond->mcast_work)->work) ->&rq->__lock FD: 232 BD: 1 +.+.: (wq_completion)bond999#2 ->(work_completion)(&(&slave->notify_work)->work) ->(work_completion)(&(&bond->mcast_work)->work) ->&rq->__lock FD: 232 BD: 1 +.+.: (wq_completion)bond981#3 ->&rq->__lock ->(work_completion)(&(&slave->notify_work)->work) ->(work_completion)(&(&bond->mcast_work)->work) ->&cfs_rq->removed.lock ->&obj_hash[i].lock FD: 232 BD: 1 +.+.: (wq_completion)bond1005 ->&rq->__lock ->(work_completion)(&(&slave->notify_work)->work) ->(work_completion)(&(&bond->mcast_work)->work) FD: 232 BD: 1 +.+.: (wq_completion)bond1000#2 ->(work_completion)(&(&slave->notify_work)->work) ->(work_completion)(&(&bond->mcast_work)->work) ->&rq->__lock FD: 232 BD: 1 +.+.: (wq_completion)bond982#3 ->&rq->__lock ->(work_completion)(&(&slave->notify_work)->work) ->(work_completion)(&(&bond->mcast_work)->work) FD: 232 BD: 1 +.+.: (wq_completion)bond1006 ->(work_completion)(&(&slave->notify_work)->work) ->(work_completion)(&(&bond->mcast_work)->work) ->&rq->__lock FD: 232 BD: 1 +.+.: (wq_completion)bond1001#2 ->&rq->__lock ->(work_completion)(&(&slave->notify_work)->work) ->(work_completion)(&(&bond->mcast_work)->work) FD: 232 BD: 1 +.+.: (wq_completion)bond983#3 ->&rq->__lock ->(work_completion)(&(&slave->notify_work)->work) ->(work_completion)(&(&bond->mcast_work)->work) FD: 232 BD: 1 +.+.: (wq_completion)bond1007 ->&rq->__lock ->(work_completion)(&(&slave->notify_work)->work) ->(work_completion)(&(&bond->mcast_work)->work) FD: 232 BD: 1 +.+.: (wq_completion)bond984#3 ->&rq->__lock ->(work_completion)(&(&slave->notify_work)->work) ->(work_completion)(&(&bond->mcast_work)->work) FD: 232 BD: 1 +.+.: (wq_completion)bond1008 ->(work_completion)(&(&slave->notify_work)->work) ->(work_completion)(&(&bond->mcast_work)->work) ->&rq->__lock ->&cfs_rq->removed.lock ->&obj_hash[i].lock FD: 232 BD: 1 +.+.: (wq_completion)bond1002#2 ->(work_completion)(&(&slave->notify_work)->work) ->(work_completion)(&(&bond->mcast_work)->work) ->&rq->__lock FD: 232 BD: 1 +.+.: (wq_completion)bond985#3 ->(work_completion)(&(&slave->notify_work)->work) ->(work_completion)(&(&bond->mcast_work)->work) ->&rq->__lock FD: 232 BD: 1 +.+.: (wq_completion)bond1009 ->&rq->__lock ->(work_completion)(&(&slave->notify_work)->work) ->(work_completion)(&(&bond->mcast_work)->work) FD: 232 BD: 1 +.+.: (wq_completion)bond1003#2 ->&rq->__lock ->(work_completion)(&(&slave->notify_work)->work) ->(work_completion)(&(&bond->mcast_work)->work) FD: 232 BD: 1 +.+.: (wq_completion)bond986#3 ->&rq->__lock ->(work_completion)(&(&slave->notify_work)->work) ->(work_completion)(&(&bond->mcast_work)->work) FD: 232 BD: 1 +.+.: (wq_completion)bond1010 ->&rq->__lock ->(work_completion)(&(&slave->notify_work)->work) ->(work_completion)(&(&bond->mcast_work)->work) FD: 232 BD: 1 +.+.: (wq_completion)bond1004#2 ->&rq->__lock ->(work_completion)(&(&slave->notify_work)->work) ->(work_completion)(&(&bond->mcast_work)->work) FD: 232 BD: 1 +.+.: (wq_completion)bond250#4 ->&rq->__lock ->(work_completion)(&(&slave->notify_work)->work) ->(work_completion)(&(&bond->mcast_work)->work) FD: 232 BD: 1 +.+.: (wq_completion)bond987#3 ->(work_completion)(&(&slave->notify_work)->work) ->(work_completion)(&(&bond->mcast_work)->work) ->&rq->__lock FD: 232 BD: 1 +.+.: (wq_completion)bond1011 ->&rq->__lock ->(work_completion)(&(&slave->notify_work)->work) ->(work_completion)(&(&bond->mcast_work)->work) FD: 232 BD: 1 +.+.: (wq_completion)bond1012 ->(work_completion)(&(&slave->notify_work)->work) ->(work_completion)(&(&bond->mcast_work)->work) ->&rq->__lock FD: 232 BD: 1 +.+.: (wq_completion)bond1005#2 ->&rq->__lock ->(work_completion)(&(&slave->notify_work)->work) ->(work_completion)(&(&bond->mcast_work)->work) FD: 232 BD: 1 +.+.: (wq_completion)bond988#3 ->(work_completion)(&(&slave->notify_work)->work) ->(work_completion)(&(&bond->mcast_work)->work) ->&rq->__lock FD: 232 BD: 1 +.+.: (wq_completion)bond1013 ->(work_completion)(&(&slave->notify_work)->work) ->(work_completion)(&(&bond->mcast_work)->work) FD: 232 BD: 1 +.+.: (wq_completion)bond1006#2 ->&rq->__lock ->(work_completion)(&(&slave->notify_work)->work) ->(work_completion)(&(&bond->mcast_work)->work) FD: 232 BD: 1 +.+.: (wq_completion)bond989#3 ->&rq->__lock ->(work_completion)(&(&slave->notify_work)->work) ->(work_completion)(&(&bond->mcast_work)->work) FD: 232 BD: 1 +.+.: (wq_completion)bond1014 ->&rq->__lock ->(work_completion)(&(&slave->notify_work)->work) ->(work_completion)(&(&bond->mcast_work)->work) FD: 232 BD: 1 +.+.: (wq_completion)bond1007#2 ->&rq->__lock ->(work_completion)(&(&slave->notify_work)->work) ->(work_completion)(&(&bond->mcast_work)->work) FD: 232 BD: 1 +.+.: (wq_completion)bond251#4 ->(work_completion)(&(&slave->notify_work)->work) ->(work_completion)(&(&bond->mcast_work)->work) ->&rq->__lock FD: 232 BD: 1 +.+.: (wq_completion)bond1015 ->&rq->__lock ->(work_completion)(&(&slave->notify_work)->work) ->(work_completion)(&(&bond->mcast_work)->work) FD: 232 BD: 1 +.+.: (wq_completion)bond990#3 ->&rq->__lock ->(work_completion)(&(&slave->notify_work)->work) ->(work_completion)(&(&bond->mcast_work)->work) FD: 232 BD: 1 +.+.: (wq_completion)bond1008#2 ->&rq->__lock ->(work_completion)(&(&slave->notify_work)->work) ->(work_completion)(&(&bond->mcast_work)->work) FD: 232 BD: 1 +.+.: (wq_completion)bond1016 ->(work_completion)(&(&slave->notify_work)->work) ->(work_completion)(&(&bond->mcast_work)->work) ->&rq->__lock FD: 232 BD: 1 +.+.: (wq_completion)bond991#3 ->&rq->__lock ->(work_completion)(&(&slave->notify_work)->work) ->(work_completion)(&(&bond->mcast_work)->work) FD: 232 BD: 1 +.+.: (wq_completion)bond1009#2 ->(work_completion)(&(&slave->notify_work)->work) ->(work_completion)(&(&bond->mcast_work)->work) ->&rq->__lock FD: 232 BD: 1 +.+.: (wq_completion)bond1017 ->&rq->__lock ->(work_completion)(&(&slave->notify_work)->work) ->(work_completion)(&(&bond->mcast_work)->work) FD: 232 BD: 1 +.+.: (wq_completion)bond992#3 ->&rq->__lock ->(work_completion)(&(&slave->notify_work)->work) ->(work_completion)(&(&bond->mcast_work)->work) FD: 232 BD: 1 +.+.: (wq_completion)bond1010#2 ->&rq->__lock ->(work_completion)(&(&slave->notify_work)->work) ->(work_completion)(&(&bond->mcast_work)->work) FD: 232 BD: 1 +.+.: (wq_completion)bond1018 ->&rq->__lock ->(work_completion)(&(&slave->notify_work)->work) ->(work_completion)(&(&bond->mcast_work)->work) FD: 232 BD: 1 +.+.: (wq_completion)bond993#3 ->(work_completion)(&(&slave->notify_work)->work) ->(work_completion)(&(&bond->mcast_work)->work) ->&rq->__lock FD: 232 BD: 1 +.+.: (wq_completion)bond1011#2 ->&rq->__lock ->(work_completion)(&(&slave->notify_work)->work) ->(work_completion)(&(&bond->mcast_work)->work) ->&cfs_rq->removed.lock ->&obj_hash[i].lock FD: 232 BD: 1 +.+.: (wq_completion)bond252#4 ->(work_completion)(&(&slave->notify_work)->work) ->(work_completion)(&(&bond->mcast_work)->work) ->&rq->__lock FD: 232 BD: 1 +.+.: (wq_completion)bond1019 ->&rq->__lock ->(work_completion)(&(&slave->notify_work)->work) ->(work_completion)(&(&bond->mcast_work)->work) FD: 232 BD: 1 +.+.: (wq_completion)bond994#3 ->&rq->__lock ->(work_completion)(&(&slave->notify_work)->work) ->(work_completion)(&(&bond->mcast_work)->work) FD: 232 BD: 1 +.+.: (wq_completion)bond1012#2 ->&rq->__lock ->(work_completion)(&(&slave->notify_work)->work) ->(work_completion)(&(&bond->mcast_work)->work) FD: 232 BD: 1 +.+.: (wq_completion)bond1020 ->&rq->__lock ->(work_completion)(&(&slave->notify_work)->work) ->(work_completion)(&(&bond->mcast_work)->work) ->&cfs_rq->removed.lock ->&obj_hash[i].lock FD: 232 BD: 1 +.+.: (wq_completion)bond995#2 ->&rq->__lock ->(work_completion)(&(&slave->notify_work)->work) ->(work_completion)(&(&bond->mcast_work)->work) ->&cfs_rq->removed.lock FD: 232 BD: 1 +.+.: (wq_completion)bond1021 ->&rq->__lock ->(work_completion)(&(&slave->notify_work)->work) ->(work_completion)(&(&bond->mcast_work)->work) ->&cfs_rq->removed.lock ->&obj_hash[i].lock FD: 232 BD: 1 +.+.: (wq_completion)bond253#4 ->&rq->__lock ->(work_completion)(&(&slave->notify_work)->work) ->(work_completion)(&(&bond->mcast_work)->work) FD: 232 BD: 1 +.+.: (wq_completion)bond254#4 ->&rq->__lock ->(work_completion)(&(&slave->notify_work)->work) ->(work_completion)(&(&bond->mcast_work)->work) FD: 232 BD: 1 +.+.: (wq_completion)bond996#3 ->&rq->__lock ->(work_completion)(&(&slave->notify_work)->work) ->(work_completion)(&(&bond->mcast_work)->work) FD: 232 BD: 1 +.+.: (wq_completion)bond1014#2 ->&rq->__lock ->(work_completion)(&(&slave->notify_work)->work) ->(work_completion)(&(&bond->mcast_work)->work) FD: 232 BD: 1 +.+.: (wq_completion)bond1022 ->(work_completion)(&(&slave->notify_work)->work) ->(work_completion)(&(&bond->mcast_work)->work) ->&rq->__lock FD: 232 BD: 1 +.+.: (wq_completion)bond255#4 ->&rq->__lock ->(work_completion)(&(&slave->notify_work)->work) ->(work_completion)(&(&bond->mcast_work)->work) FD: 232 BD: 1 +.+.: (wq_completion)bond997#3 ->&rq->__lock ->(work_completion)(&(&slave->notify_work)->work) ->(work_completion)(&(&bond->mcast_work)->work) FD: 232 BD: 1 +.+.: (wq_completion)bond1015#2 ->&rq->__lock ->(work_completion)(&(&slave->notify_work)->work) ->(work_completion)(&(&bond->mcast_work)->work) FD: 232 BD: 1 +.+.: (wq_completion)bond998#3 ->(work_completion)(&(&slave->notify_work)->work) ->(work_completion)(&(&bond->mcast_work)->work) ->&rq->__lock FD: 232 BD: 1 +.+.: (wq_completion)bond1016#2 ->&rq->__lock ->(work_completion)(&(&slave->notify_work)->work) ->(work_completion)(&(&bond->mcast_work)->work) FD: 232 BD: 1 +.+.: (wq_completion)bond1023 ->&rq->__lock ->(work_completion)(&(&slave->notify_work)->work) ->(work_completion)(&(&bond->mcast_work)->work) FD: 232 BD: 1 +.+.: (wq_completion)bond256#4 ->(work_completion)(&(&slave->notify_work)->work) ->(work_completion)(&(&bond->mcast_work)->work) ->&rq->__lock FD: 232 BD: 1 +.+.: (wq_completion)bond1017#2 ->&rq->__lock ->(work_completion)(&(&slave->notify_work)->work) ->(work_completion)(&(&bond->mcast_work)->work) FD: 232 BD: 1 +.+.: (wq_completion)bond257#4 ->(work_completion)(&(&slave->notify_work)->work) ->(work_completion)(&(&bond->mcast_work)->work) ->&rq->__lock FD: 232 BD: 1 +.+.: (wq_completion)bond1024 ->(work_completion)(&(&slave->notify_work)->work) ->(work_completion)(&(&bond->mcast_work)->work) ->&rq->__lock FD: 232 BD: 1 +.+.: (wq_completion)bond999#3 ->(work_completion)(&(&slave->notify_work)->work) ->(work_completion)(&(&bond->mcast_work)->work) ->&rq->__lock FD: 232 BD: 1 +.+.: (wq_completion)bond258#4 ->(work_completion)(&(&slave->notify_work)->work) ->(work_completion)(&(&bond->mcast_work)->work) FD: 232 BD: 1 +.+.: (wq_completion)bond1025 ->(work_completion)(&(&slave->notify_work)->work) ->(work_completion)(&(&bond->mcast_work)->work) ->&rq->__lock FD: 232 BD: 1 +.+.: (wq_completion)bond1000#3 ->(work_completion)(&(&slave->notify_work)->work) ->(work_completion)(&(&bond->mcast_work)->work) ->&rq->__lock FD: 232 BD: 1 +.+.: (wq_completion)bond1018#2 ->&rq->__lock ->(work_completion)(&(&slave->notify_work)->work) ->(work_completion)(&(&bond->mcast_work)->work) FD: 232 BD: 1 +.+.: (wq_completion)bond259#4 ->&rq->__lock ->(work_completion)(&(&slave->notify_work)->work) ->(work_completion)(&(&bond->mcast_work)->work) FD: 232 BD: 1 +.+.: (wq_completion)bond1019#2 ->&rq->__lock ->(work_completion)(&(&slave->notify_work)->work) ->(work_completion)(&(&bond->mcast_work)->work) FD: 232 BD: 1 +.+.: (wq_completion)bond1026 ->&rq->__lock ->(work_completion)(&(&slave->notify_work)->work) ->(work_completion)(&(&bond->mcast_work)->work) FD: 232 BD: 1 +.+.: (wq_completion)bond1001#3 ->&rq->__lock ->(work_completion)(&(&slave->notify_work)->work) ->(work_completion)(&(&bond->mcast_work)->work) FD: 232 BD: 1 +.+.: (wq_completion)bond1020#2 ->(work_completion)(&(&slave->notify_work)->work) ->(work_completion)(&(&bond->mcast_work)->work) ->&rq->__lock FD: 232 BD: 1 +.+.: (wq_completion)bond1027 ->&rq->__lock ->(work_completion)(&(&slave->notify_work)->work) ->(work_completion)(&(&bond->mcast_work)->work) FD: 232 BD: 1 +.+.: (wq_completion)bond1002#3 ->&rq->__lock ->(work_completion)(&(&slave->notify_work)->work) ->(work_completion)(&(&bond->mcast_work)->work) FD: 232 BD: 1 +.+.: (wq_completion)bond1021#2 ->&rq->__lock ->(work_completion)(&(&slave->notify_work)->work) ->(work_completion)(&(&bond->mcast_work)->work) FD: 232 BD: 1 +.+.: (wq_completion)bond1028 ->&rq->__lock ->(work_completion)(&(&slave->notify_work)->work) ->(work_completion)(&(&bond->mcast_work)->work) FD: 232 BD: 1 +.+.: (wq_completion)bond1003#3 ->(work_completion)(&(&slave->notify_work)->work) ->(work_completion)(&(&bond->mcast_work)->work) FD: 232 BD: 1 +.+.: (wq_completion)bond260#4 ->&rq->__lock ->(work_completion)(&(&slave->notify_work)->work) ->(work_completion)(&(&bond->mcast_work)->work) FD: 232 BD: 1 +.+.: (wq_completion)bond1022#2 ->&rq->__lock ->(work_completion)(&(&slave->notify_work)->work) ->(work_completion)(&(&bond->mcast_work)->work) FD: 232 BD: 1 +.+.: (wq_completion)bond261#3 ->&rq->__lock ->(work_completion)(&(&slave->notify_work)->work) ->(work_completion)(&(&bond->mcast_work)->work) FD: 232 BD: 1 +.+.: (wq_completion)bond1029 ->&rq->__lock ->(work_completion)(&(&slave->notify_work)->work) ->(work_completion)(&(&bond->mcast_work)->work) FD: 232 BD: 1 +.+.: (wq_completion)bond1023#2 ->&rq->__lock ->(work_completion)(&(&slave->notify_work)->work) ->(work_completion)(&(&bond->mcast_work)->work) FD: 232 BD: 1 +.+.: (wq_completion)bond262#4 ->&rq->__lock ->(work_completion)(&(&slave->notify_work)->work) ->(work_completion)(&(&bond->mcast_work)->work) FD: 232 BD: 1 +.+.: (wq_completion)bond1030 ->&rq->__lock ->(work_completion)(&(&slave->notify_work)->work) ->(work_completion)(&(&bond->mcast_work)->work) FD: 232 BD: 1 +.+.: (wq_completion)bond1024#2 ->&rq->__lock ->(work_completion)(&(&slave->notify_work)->work) ->(work_completion)(&(&bond->mcast_work)->work) FD: 232 BD: 1 +.+.: (wq_completion)bond1031 ->&rq->__lock ->(work_completion)(&(&slave->notify_work)->work) ->(work_completion)(&(&bond->mcast_work)->work) FD: 232 BD: 1 +.+.: (wq_completion)bond1025#2 ->&rq->__lock ->(work_completion)(&(&slave->notify_work)->work) ->(work_completion)(&(&bond->mcast_work)->work) FD: 232 BD: 1 +.+.: (wq_completion)bond1004#3 ->&rq->__lock ->(work_completion)(&(&slave->notify_work)->work) ->(work_completion)(&(&bond->mcast_work)->work) FD: 232 BD: 1 +.+.: (wq_completion)bond1032 ->&rq->__lock ->(work_completion)(&(&slave->notify_work)->work) ->(work_completion)(&(&bond->mcast_work)->work) FD: 232 BD: 1 +.+.: (wq_completion)bond1026#2 ->&rq->__lock ->(work_completion)(&(&slave->notify_work)->work) ->(work_completion)(&(&bond->mcast_work)->work) FD: 232 BD: 1 +.+.: (wq_completion)bond1005#3 ->(work_completion)(&(&slave->notify_work)->work) ->(work_completion)(&(&bond->mcast_work)->work) ->&rq->__lock FD: 232 BD: 1 +.+.: (wq_completion)bond1033 ->&rq->__lock ->(work_completion)(&(&slave->notify_work)->work) ->(work_completion)(&(&bond->mcast_work)->work) ->&cfs_rq->removed.lock ->&obj_hash[i].lock FD: 232 BD: 1 +.+.: (wq_completion)bond1027#2 ->&rq->__lock ->(work_completion)(&(&slave->notify_work)->work) ->(work_completion)(&(&bond->mcast_work)->work) FD: 232 BD: 1 +.+.: (wq_completion)bond1006#3 ->&rq->__lock ->(work_completion)(&(&slave->notify_work)->work) ->(work_completion)(&(&bond->mcast_work)->work) FD: 232 BD: 1 +.+.: (wq_completion)bond1034 ->&rq->__lock ->(work_completion)(&(&slave->notify_work)->work) ->(work_completion)(&(&bond->mcast_work)->work) FD: 232 BD: 1 +.+.: (wq_completion)bond1028#2 ->&rq->__lock ->(work_completion)(&(&slave->notify_work)->work) ->(work_completion)(&(&bond->mcast_work)->work) FD: 232 BD: 1 +.+.: (wq_completion)bond1007#3 ->&rq->__lock ->(work_completion)(&(&slave->notify_work)->work) ->(work_completion)(&(&bond->mcast_work)->work) FD: 232 BD: 1 +.+.: (wq_completion)bond1035 ->&rq->__lock ->(work_completion)(&(&slave->notify_work)->work) ->(work_completion)(&(&bond->mcast_work)->work) FD: 232 BD: 1 +.+.: (wq_completion)bond1029#2 ->(work_completion)(&(&slave->notify_work)->work) ->(work_completion)(&(&bond->mcast_work)->work) FD: 232 BD: 1 +.+.: (wq_completion)bond1008#3 ->&rq->__lock ->(work_completion)(&(&slave->notify_work)->work) ->(work_completion)(&(&bond->mcast_work)->work) FD: 232 BD: 1 +.+.: (wq_completion)bond1036 ->&rq->__lock ->(work_completion)(&(&slave->notify_work)->work) ->(work_completion)(&(&bond->mcast_work)->work) FD: 232 BD: 1 +.+.: (wq_completion)bond1030#2 ->&rq->__lock ->(work_completion)(&(&slave->notify_work)->work) ->(work_completion)(&(&bond->mcast_work)->work) FD: 232 BD: 1 +.+.: (wq_completion)bond1009#3 ->&rq->__lock ->(work_completion)(&(&slave->notify_work)->work) ->(work_completion)(&(&bond->mcast_work)->work) FD: 232 BD: 1 +.+.: (wq_completion)bond1037 ->(work_completion)(&(&slave->notify_work)->work) ->(work_completion)(&(&bond->mcast_work)->work) ->&rq->__lock FD: 232 BD: 1 +.+.: (wq_completion)bond1031#2 ->&rq->__lock ->(work_completion)(&(&slave->notify_work)->work) ->(work_completion)(&(&bond->mcast_work)->work) FD: 232 BD: 1 +.+.: (wq_completion)bond1010#3 ->&rq->__lock ->(work_completion)(&(&slave->notify_work)->work) ->(work_completion)(&(&bond->mcast_work)->work) FD: 232 BD: 1 +.+.: (wq_completion)bond1038 ->&rq->__lock ->(work_completion)(&(&slave->notify_work)->work) ->(work_completion)(&(&bond->mcast_work)->work) FD: 232 BD: 1 +.+.: (wq_completion)bond1032#2 ->(work_completion)(&(&slave->notify_work)->work) ->(work_completion)(&(&bond->mcast_work)->work) ->&rq->__lock FD: 232 BD: 1 +.+.: (wq_completion)bond1039 ->&rq->__lock ->(work_completion)(&(&slave->notify_work)->work) ->(work_completion)(&(&bond->mcast_work)->work) FD: 232 BD: 1 +.+.: (wq_completion)bond1011#3 ->(work_completion)(&(&slave->notify_work)->work) ->(work_completion)(&(&bond->mcast_work)->work) ->&rq->__lock FD: 232 BD: 1 +.+.: (wq_completion)bond1040 ->&rq->__lock ->(work_completion)(&(&slave->notify_work)->work) ->(work_completion)(&(&bond->mcast_work)->work) FD: 232 BD: 1 +.+.: (wq_completion)bond1033#2 ->&rq->__lock ->(work_completion)(&(&slave->notify_work)->work) ->(work_completion)(&(&bond->mcast_work)->work) FD: 232 BD: 1 +.+.: (wq_completion)bond1012#3 ->&rq->__lock ->(work_completion)(&(&slave->notify_work)->work) ->(work_completion)(&(&bond->mcast_work)->work) FD: 232 BD: 1 +.+.: (wq_completion)bond1034#2 ->&rq->__lock ->(work_completion)(&(&slave->notify_work)->work) ->(work_completion)(&(&bond->mcast_work)->work) FD: 232 BD: 1 +.+.: (wq_completion)bond1041 ->&rq->__lock ->(work_completion)(&(&slave->notify_work)->work) ->(work_completion)(&(&bond->mcast_work)->work) FD: 232 BD: 1 +.+.: (wq_completion)bond1035#2 ->&rq->__lock ->(work_completion)(&(&slave->notify_work)->work) ->(work_completion)(&(&bond->mcast_work)->work) FD: 232 BD: 1 +.+.: (wq_completion)bond1013#2 ->&rq->__lock ->(work_completion)(&(&slave->notify_work)->work) ->(work_completion)(&(&bond->mcast_work)->work) FD: 232 BD: 1 +.+.: (wq_completion)bond1042 ->&rq->__lock ->(work_completion)(&(&slave->notify_work)->work) ->(work_completion)(&(&bond->mcast_work)->work) FD: 232 BD: 1 +.+.: (wq_completion)bond1036#2 ->&rq->__lock ->(work_completion)(&(&slave->notify_work)->work) ->(work_completion)(&(&bond->mcast_work)->work) FD: 232 BD: 1 +.+.: (wq_completion)bond1014#3 ->&rq->__lock ->(work_completion)(&(&slave->notify_work)->work) ->(work_completion)(&(&bond->mcast_work)->work) FD: 232 BD: 1 +.+.: (wq_completion)bond1037#2 ->&rq->__lock ->(work_completion)(&(&slave->notify_work)->work) ->(work_completion)(&(&bond->mcast_work)->work) FD: 232 BD: 1 +.+.: (wq_completion)bond1015#3 ->&rq->__lock ->(work_completion)(&(&slave->notify_work)->work) ->(work_completion)(&(&bond->mcast_work)->work) FD: 232 BD: 1 +.+.: (wq_completion)bond1043 ->(work_completion)(&(&slave->notify_work)->work) ->(work_completion)(&(&bond->mcast_work)->work) FD: 232 BD: 1 +.+.: (wq_completion)bond1038#2 ->(work_completion)(&(&slave->notify_work)->work) ->(work_completion)(&(&bond->mcast_work)->work) ->&rq->__lock FD: 232 BD: 1 +.+.: (wq_completion)bond1016#3 ->&rq->__lock ->(work_completion)(&(&slave->notify_work)->work) ->(work_completion)(&(&bond->mcast_work)->work) FD: 232 BD: 1 +.+.: (wq_completion)bond1044 ->&rq->__lock ->(work_completion)(&(&slave->notify_work)->work) ->(work_completion)(&(&bond->mcast_work)->work) FD: 232 BD: 1 +.+.: (wq_completion)bond1039#2 ->&rq->__lock ->&cfs_rq->removed.lock ->&obj_hash[i].lock ->(work_completion)(&(&slave->notify_work)->work) ->(work_completion)(&(&bond->mcast_work)->work) FD: 232 BD: 1 +.+.: (wq_completion)bond1017#3 ->&rq->__lock ->(work_completion)(&(&slave->notify_work)->work) ->(work_completion)(&(&bond->mcast_work)->work) FD: 232 BD: 1 +.+.: (wq_completion)bond1045 ->&rq->__lock ->(work_completion)(&(&slave->notify_work)->work) ->(work_completion)(&(&bond->mcast_work)->work) FD: 232 BD: 1 +.+.: (wq_completion)bond1040#2 ->(work_completion)(&(&slave->notify_work)->work) ->(work_completion)(&(&bond->mcast_work)->work) FD: 232 BD: 1 +.+.: (wq_completion)bond1046 ->(work_completion)(&(&slave->notify_work)->work) ->(work_completion)(&(&bond->mcast_work)->work) ->&rq->__lock FD: 232 BD: 1 +.+.: (wq_completion)bond1041#2 ->&rq->__lock ->(work_completion)(&(&slave->notify_work)->work) ->(work_completion)(&(&bond->mcast_work)->work) FD: 232 BD: 1 +.+.: (wq_completion)bond1042#2 ->&rq->__lock ->(work_completion)(&(&slave->notify_work)->work) ->(work_completion)(&(&bond->mcast_work)->work) FD: 232 BD: 1 +.+.: (wq_completion)bond1019#3 ->(work_completion)(&(&slave->notify_work)->work) ->(work_completion)(&(&bond->mcast_work)->work) ->&rq->__lock FD: 232 BD: 1 +.+.: (wq_completion)bond1047 ->(work_completion)(&(&slave->notify_work)->work) ->(work_completion)(&(&bond->mcast_work)->work) ->&rq->__lock FD: 232 BD: 1 +.+.: (wq_completion)bond1020#3 ->&rq->__lock ->(work_completion)(&(&slave->notify_work)->work) ->(work_completion)(&(&bond->mcast_work)->work) FD: 232 BD: 1 +.+.: (wq_completion)bond1048 ->(work_completion)(&(&slave->notify_work)->work) ->(work_completion)(&(&bond->mcast_work)->work) ->&rq->__lock FD: 232 BD: 1 +.+.: (wq_completion)bond1021#3 ->&rq->__lock ->(work_completion)(&(&slave->notify_work)->work) ->(work_completion)(&(&bond->mcast_work)->work) FD: 232 BD: 1 +.+.: (wq_completion)bond1043#2 ->&rq->__lock ->(work_completion)(&(&slave->notify_work)->work) ->(work_completion)(&(&bond->mcast_work)->work) FD: 232 BD: 1 +.+.: (wq_completion)bond1049 ->&rq->__lock ->(work_completion)(&(&slave->notify_work)->work) ->(work_completion)(&(&bond->mcast_work)->work) FD: 232 BD: 1 +.+.: (wq_completion)bond1022#3 ->&rq->__lock ->(work_completion)(&(&slave->notify_work)->work) ->(work_completion)(&(&bond->mcast_work)->work) FD: 232 BD: 1 +.+.: (wq_completion)bond1050 ->(work_completion)(&(&slave->notify_work)->work) ->(work_completion)(&(&bond->mcast_work)->work) ->&rq->__lock FD: 232 BD: 1 +.+.: (wq_completion)bond1044#2 ->(work_completion)(&(&slave->notify_work)->work) ->(work_completion)(&(&bond->mcast_work)->work) ->&rq->__lock FD: 232 BD: 1 +.+.: (wq_completion)bond1023#3 ->&rq->__lock ->(work_completion)(&(&slave->notify_work)->work) ->(work_completion)(&(&bond->mcast_work)->work) FD: 232 BD: 1 +.+.: (wq_completion)bond1051 ->&rq->__lock ->(work_completion)(&(&slave->notify_work)->work) ->(work_completion)(&(&bond->mcast_work)->work) FD: 232 BD: 1 +.+.: (wq_completion)bond1024#3 ->&rq->__lock ->(work_completion)(&(&slave->notify_work)->work) ->(work_completion)(&(&bond->mcast_work)->work) FD: 232 BD: 1 +.+.: (wq_completion)bond1052 ->&rq->__lock ->(work_completion)(&(&slave->notify_work)->work) ->(work_completion)(&(&bond->mcast_work)->work) FD: 232 BD: 1 +.+.: (wq_completion)bond1045#2 ->(work_completion)(&(&slave->notify_work)->work) ->(work_completion)(&(&bond->mcast_work)->work) ->&rq->__lock FD: 232 BD: 1 +.+.: (wq_completion)bond1025#3 ->(work_completion)(&(&slave->notify_work)->work) ->(work_completion)(&(&bond->mcast_work)->work) ->&rq->__lock FD: 232 BD: 1 +.+.: (wq_completion)bond1053 ->&rq->__lock ->(work_completion)(&(&slave->notify_work)->work) ->(work_completion)(&(&bond->mcast_work)->work) FD: 232 BD: 1 +.+.: (wq_completion)bond1046#2 ->&rq->__lock ->(work_completion)(&(&slave->notify_work)->work) ->(work_completion)(&(&bond->mcast_work)->work) FD: 232 BD: 1 +.+.: (wq_completion)bond1026#3 ->&rq->__lock ->(work_completion)(&(&slave->notify_work)->work) ->(work_completion)(&(&bond->mcast_work)->work) FD: 232 BD: 1 +.+.: (wq_completion)bond1047#2 ->&rq->__lock ->(work_completion)(&(&slave->notify_work)->work) ->(work_completion)(&(&bond->mcast_work)->work) FD: 232 BD: 1 +.+.: (wq_completion)bond1054 ->(work_completion)(&(&slave->notify_work)->work) ->(work_completion)(&(&bond->mcast_work)->work) ->&rq->__lock FD: 232 BD: 1 +.+.: (wq_completion)bond1027#3 ->(work_completion)(&(&slave->notify_work)->work) ->(work_completion)(&(&bond->mcast_work)->work) ->&rq->__lock FD: 232 BD: 1 +.+.: (wq_completion)bond1048#2 ->&rq->__lock ->(work_completion)(&(&slave->notify_work)->work) ->(work_completion)(&(&bond->mcast_work)->work) FD: 232 BD: 1 +.+.: (wq_completion)bond1055 ->&rq->__lock ->(work_completion)(&(&slave->notify_work)->work) ->(work_completion)(&(&bond->mcast_work)->work) FD: 232 BD: 1 +.+.: (wq_completion)bond1028#3 ->&rq->__lock ->(work_completion)(&(&slave->notify_work)->work) ->(work_completion)(&(&bond->mcast_work)->work) FD: 232 BD: 1 +.+.: (wq_completion)bond1049#2 ->&rq->__lock ->(work_completion)(&(&slave->notify_work)->work) ->(work_completion)(&(&bond->mcast_work)->work) FD: 232 BD: 1 +.+.: (wq_completion)bond1029#3 ->(work_completion)(&(&slave->notify_work)->work) ->(work_completion)(&(&bond->mcast_work)->work) ->&rq->__lock FD: 232 BD: 1 +.+.: (wq_completion)bond1050#2 ->(work_completion)(&(&slave->notify_work)->work) ->(work_completion)(&(&bond->mcast_work)->work) ->&rq->__lock FD: 232 BD: 1 +.+.: (wq_completion)bond1056 ->(work_completion)(&(&slave->notify_work)->work) ->(work_completion)(&(&bond->mcast_work)->work) ->&rq->__lock FD: 232 BD: 1 +.+.: (wq_completion)bond1030#3 ->&rq->__lock ->(work_completion)(&(&slave->notify_work)->work) ->(work_completion)(&(&bond->mcast_work)->work) FD: 232 BD: 1 +.+.: (wq_completion)bond1051#2 ->(work_completion)(&(&slave->notify_work)->work) ->(work_completion)(&(&bond->mcast_work)->work) ->&rq->__lock FD: 232 BD: 1 +.+.: (wq_completion)bond1057 ->&rq->__lock ->(work_completion)(&(&slave->notify_work)->work) ->(work_completion)(&(&bond->mcast_work)->work) FD: 232 BD: 1 +.+.: (wq_completion)bond1031#3 ->(work_completion)(&(&slave->notify_work)->work) ->(work_completion)(&(&bond->mcast_work)->work) ->&rq->__lock FD: 232 BD: 1 +.+.: (wq_completion)bond1052#2 ->&rq->__lock ->(work_completion)(&(&slave->notify_work)->work) ->(work_completion)(&(&bond->mcast_work)->work) FD: 232 BD: 1 +.+.: (wq_completion)bond1058 ->&rq->__lock ->(work_completion)(&(&slave->notify_work)->work) ->(work_completion)(&(&bond->mcast_work)->work) FD: 232 BD: 1 +.+.: (wq_completion)bond1032#3 ->(work_completion)(&(&slave->notify_work)->work) ->(work_completion)(&(&bond->mcast_work)->work) ->&rq->__lock FD: 232 BD: 1 +.+.: (wq_completion)bond1053#2 ->&rq->__lock ->(work_completion)(&(&slave->notify_work)->work) ->(work_completion)(&(&bond->mcast_work)->work) FD: 232 BD: 1 +.+.: (wq_completion)bond1059 ->(work_completion)(&(&slave->notify_work)->work) ->(work_completion)(&(&bond->mcast_work)->work) ->&rq->__lock FD: 232 BD: 1 +.+.: (wq_completion)bond1033#3 ->&rq->__lock ->(work_completion)(&(&slave->notify_work)->work) ->(work_completion)(&(&bond->mcast_work)->work) FD: 232 BD: 1 +.+.: (wq_completion)bond1054#2 ->&rq->__lock ->(work_completion)(&(&slave->notify_work)->work) ->(work_completion)(&(&bond->mcast_work)->work) FD: 232 BD: 1 +.+.: (wq_completion)bond1060 ->&rq->__lock ->(work_completion)(&(&slave->notify_work)->work) ->(work_completion)(&(&bond->mcast_work)->work) ->&cfs_rq->removed.lock ->&obj_hash[i].lock FD: 232 BD: 1 +.+.: (wq_completion)bond1034#3 ->(work_completion)(&(&slave->notify_work)->work) ->(work_completion)(&(&bond->mcast_work)->work) ->&rq->__lock FD: 232 BD: 1 +.+.: (wq_completion)bond1055#2 ->&rq->__lock ->(work_completion)(&(&slave->notify_work)->work) ->(work_completion)(&(&bond->mcast_work)->work) FD: 232 BD: 1 +.+.: (wq_completion)bond1061 ->(work_completion)(&(&slave->notify_work)->work) ->(work_completion)(&(&bond->mcast_work)->work) ->&rq->__lock FD: 232 BD: 1 +.+.: (wq_completion)bond1035#3 ->&rq->__lock ->(work_completion)(&(&slave->notify_work)->work) ->(work_completion)(&(&bond->mcast_work)->work) FD: 232 BD: 1 +.+.: (wq_completion)bond1056#2 ->(work_completion)(&(&slave->notify_work)->work) ->(work_completion)(&(&bond->mcast_work)->work) ->&rq->__lock FD: 232 BD: 1 +.+.: (wq_completion)bond1062 ->&rq->__lock ->(work_completion)(&(&slave->notify_work)->work) ->(work_completion)(&(&bond->mcast_work)->work) FD: 232 BD: 1 +.+.: (wq_completion)bond1036#3 ->&rq->__lock ->(work_completion)(&(&slave->notify_work)->work) ->(work_completion)(&(&bond->mcast_work)->work) FD: 232 BD: 1 +.+.: (wq_completion)bond1057#2 ->&rq->__lock ->(work_completion)(&(&slave->notify_work)->work) ->(work_completion)(&(&bond->mcast_work)->work) FD: 232 BD: 1 +.+.: (wq_completion)bond1063 ->&rq->__lock ->(work_completion)(&(&slave->notify_work)->work) ->(work_completion)(&(&bond->mcast_work)->work) FD: 232 BD: 1 +.+.: (wq_completion)bond1058#2 ->&rq->__lock ->(work_completion)(&(&slave->notify_work)->work) ->(work_completion)(&(&bond->mcast_work)->work) FD: 232 BD: 1 +.+.: (wq_completion)bond1064 ->&rq->__lock ->(work_completion)(&(&slave->notify_work)->work) ->(work_completion)(&(&bond->mcast_work)->work) FD: 232 BD: 1 +.+.: (wq_completion)bond1037#3 ->&rq->__lock ->(work_completion)(&(&slave->notify_work)->work) ->(work_completion)(&(&bond->mcast_work)->work) FD: 232 BD: 1 +.+.: (wq_completion)bond1059#2 ->(work_completion)(&(&slave->notify_work)->work) ->(work_completion)(&(&bond->mcast_work)->work) ->&rq->__lock FD: 232 BD: 1 +.+.: (wq_completion)bond1065 ->(work_completion)(&(&slave->notify_work)->work) ->(work_completion)(&(&bond->mcast_work)->work) ->&rq->__lock FD: 232 BD: 1 +.+.: (wq_completion)bond1038#3 ->&rq->__lock ->(work_completion)(&(&slave->notify_work)->work) ->(work_completion)(&(&bond->mcast_work)->work) FD: 232 BD: 1 +.+.: (wq_completion)bond1060#2 ->&rq->__lock ->(work_completion)(&(&slave->notify_work)->work) ->(work_completion)(&(&bond->mcast_work)->work) ->&cfs_rq->removed.lock ->&obj_hash[i].lock FD: 232 BD: 1 +.+.: (wq_completion)bond1039#3 ->&rq->__lock ->(work_completion)(&(&slave->notify_work)->work) ->(work_completion)(&(&bond->mcast_work)->work) FD: 232 BD: 1 +.+.: (wq_completion)bond1061#2 ->&rq->__lock ->(work_completion)(&(&slave->notify_work)->work) ->(work_completion)(&(&bond->mcast_work)->work) FD: 232 BD: 1 +.+.: (wq_completion)bond1066 ->(work_completion)(&(&slave->notify_work)->work) ->(work_completion)(&(&bond->mcast_work)->work) ->&rq->__lock FD: 232 BD: 1 +.+.: (wq_completion)bond1040#3 ->&rq->__lock ->(work_completion)(&(&slave->notify_work)->work) ->(work_completion)(&(&bond->mcast_work)->work) FD: 232 BD: 1 +.+.: (wq_completion)bond1067 ->&rq->__lock ->(work_completion)(&(&slave->notify_work)->work) ->(work_completion)(&(&bond->mcast_work)->work) FD: 232 BD: 1 +.+.: (wq_completion)bond1041#3 ->(work_completion)(&(&slave->notify_work)->work) ->(work_completion)(&(&bond->mcast_work)->work) ->&rq->__lock FD: 232 BD: 1 +.+.: (wq_completion)bond1062#2 ->&rq->__lock ->(work_completion)(&(&slave->notify_work)->work) ->(work_completion)(&(&bond->mcast_work)->work) FD: 232 BD: 1 +.+.: (wq_completion)bond1068 ->&rq->__lock ->(work_completion)(&(&slave->notify_work)->work) ->(work_completion)(&(&bond->mcast_work)->work) FD: 232 BD: 1 +.+.: (wq_completion)bond1042#3 ->&rq->__lock ->(work_completion)(&(&slave->notify_work)->work) ->(work_completion)(&(&bond->mcast_work)->work) FD: 232 BD: 1 +.+.: (wq_completion)bond1063#2 ->(work_completion)(&(&slave->notify_work)->work) ->(work_completion)(&(&bond->mcast_work)->work) ->&rq->__lock FD: 232 BD: 1 +.+.: (wq_completion)bond1064#2 ->&rq->__lock ->(work_completion)(&(&slave->notify_work)->work) ->(work_completion)(&(&bond->mcast_work)->work) FD: 232 BD: 1 +.+.: (wq_completion)bond1069 ->&rq->__lock ->(work_completion)(&(&slave->notify_work)->work) ->(work_completion)(&(&bond->mcast_work)->work) FD: 232 BD: 1 +.+.: (wq_completion)bond1065#2 ->&rq->__lock ->(work_completion)(&(&slave->notify_work)->work) ->(work_completion)(&(&bond->mcast_work)->work) FD: 232 BD: 1 +.+.: (wq_completion)bond1070 ->&rq->__lock ->(work_completion)(&(&slave->notify_work)->work) ->(work_completion)(&(&bond->mcast_work)->work) FD: 232 BD: 1 +.+.: (wq_completion)bond1066#2 ->&rq->__lock ->(work_completion)(&(&slave->notify_work)->work) ->(work_completion)(&(&bond->mcast_work)->work) FD: 232 BD: 1 +.+.: (wq_completion)bond1071 ->(work_completion)(&(&slave->notify_work)->work) ->(work_completion)(&(&bond->mcast_work)->work) ->&rq->__lock FD: 232 BD: 1 +.+.: (wq_completion)bond1044#3 ->&rq->__lock ->(work_completion)(&(&slave->notify_work)->work) ->(work_completion)(&(&bond->mcast_work)->work) FD: 232 BD: 1 +.+.: (wq_completion)bond1067#2 ->&rq->__lock ->(work_completion)(&(&slave->notify_work)->work) ->(work_completion)(&(&bond->mcast_work)->work) FD: 232 BD: 1 +.+.: (wq_completion)bond1072 ->&rq->__lock ->(work_completion)(&(&slave->notify_work)->work) ->(work_completion)(&(&bond->mcast_work)->work) FD: 232 BD: 1 +.+.: (wq_completion)bond1045#3 ->(work_completion)(&(&slave->notify_work)->work) ->(work_completion)(&(&bond->mcast_work)->work) ->&rq->__lock FD: 232 BD: 1 +.+.: (wq_completion)bond1068#2 ->&rq->__lock ->(work_completion)(&(&slave->notify_work)->work) ->(work_completion)(&(&bond->mcast_work)->work) FD: 232 BD: 1 +.+.: (wq_completion)bond1046#3 ->&rq->__lock ->(work_completion)(&(&slave->notify_work)->work) ->(work_completion)(&(&bond->mcast_work)->work) FD: 232 BD: 1 +.+.: (wq_completion)bond1069#2 ->&rq->__lock ->(work_completion)(&(&slave->notify_work)->work) ->(work_completion)(&(&bond->mcast_work)->work) FD: 232 BD: 1 +.+.: (wq_completion)bond1073 ->&rq->__lock ->(work_completion)(&(&slave->notify_work)->work) ->(work_completion)(&(&bond->mcast_work)->work) FD: 232 BD: 1 +.+.: (wq_completion)bond1070#2 ->&rq->__lock ->(work_completion)(&(&slave->notify_work)->work) ->(work_completion)(&(&bond->mcast_work)->work) FD: 232 BD: 1 +.+.: (wq_completion)bond1047#3 ->(work_completion)(&(&slave->notify_work)->work) ->(work_completion)(&(&bond->mcast_work)->work) ->&rq->__lock FD: 232 BD: 1 +.+.: (wq_completion)bond1071#2 ->&rq->__lock ->(work_completion)(&(&slave->notify_work)->work) ->(work_completion)(&(&bond->mcast_work)->work) FD: 232 BD: 1 +.+.: (wq_completion)bond1074 ->(work_completion)(&(&slave->notify_work)->work) ->(work_completion)(&(&bond->mcast_work)->work) ->&rq->__lock FD: 232 BD: 1 +.+.: (wq_completion)bond1048#3 ->&rq->__lock ->(work_completion)(&(&slave->notify_work)->work) ->(work_completion)(&(&bond->mcast_work)->work) FD: 232 BD: 1 +.+.: (wq_completion)bond1072#2 ->&rq->__lock ->(work_completion)(&(&slave->notify_work)->work) ->(work_completion)(&(&bond->mcast_work)->work) FD: 232 BD: 1 +.+.: (wq_completion)bond1049#3 ->&rq->__lock ->(work_completion)(&(&slave->notify_work)->work) ->(work_completion)(&(&bond->mcast_work)->work) FD: 232 BD: 1 +.+.: (wq_completion)bond1073#2 ->&rq->__lock ->(work_completion)(&(&slave->notify_work)->work) ->(work_completion)(&(&bond->mcast_work)->work) FD: 232 BD: 1 +.+.: (wq_completion)bond1050#3 ->&rq->__lock ->(work_completion)(&(&slave->notify_work)->work) ->(work_completion)(&(&bond->mcast_work)->work) FD: 232 BD: 1 +.+.: (wq_completion)bond1074#2 ->&rq->__lock ->(work_completion)(&(&slave->notify_work)->work) ->(work_completion)(&(&bond->mcast_work)->work) FD: 232 BD: 1 +.+.: (wq_completion)bond1075 ->(work_completion)(&(&slave->notify_work)->work) ->(work_completion)(&(&bond->mcast_work)->work) ->&rq->__lock FD: 232 BD: 1 +.+.: (wq_completion)bond1051#3 ->&rq->__lock ->(work_completion)(&(&slave->notify_work)->work) ->(work_completion)(&(&bond->mcast_work)->work) FD: 232 BD: 1 +.+.: (wq_completion)bond1075#2 ->&rq->__lock ->(work_completion)(&(&slave->notify_work)->work) ->(work_completion)(&(&bond->mcast_work)->work) FD: 232 BD: 1 +.+.: (wq_completion)bond1076 ->&rq->__lock ->(work_completion)(&(&slave->notify_work)->work) ->(work_completion)(&(&bond->mcast_work)->work) FD: 232 BD: 1 +.+.: (wq_completion)bond1052#3 ->&rq->__lock ->(work_completion)(&(&slave->notify_work)->work) ->(work_completion)(&(&bond->mcast_work)->work) FD: 232 BD: 1 +.+.: (wq_completion)bond1076#2 ->&rq->__lock ->(work_completion)(&(&slave->notify_work)->work) ->(work_completion)(&(&bond->mcast_work)->work) FD: 232 BD: 1 +.+.: (wq_completion)bond1077 ->&rq->__lock ->(work_completion)(&(&slave->notify_work)->work) ->(work_completion)(&(&bond->mcast_work)->work) FD: 232 BD: 1 +.+.: (wq_completion)bond1053#3 ->(work_completion)(&(&slave->notify_work)->work) ->(work_completion)(&(&bond->mcast_work)->work) ->&rq->__lock FD: 232 BD: 1 +.+.: (wq_completion)bond1078 ->(work_completion)(&(&slave->notify_work)->work) ->(work_completion)(&(&bond->mcast_work)->work) ->&rq->__lock FD: 232 BD: 1 +.+.: (wq_completion)bond1054#3 ->&rq->__lock ->(work_completion)(&(&slave->notify_work)->work) ->(work_completion)(&(&bond->mcast_work)->work) FD: 232 BD: 1 +.+.: (wq_completion)bond1077#2 ->&rq->__lock ->(work_completion)(&(&slave->notify_work)->work) ->(work_completion)(&(&bond->mcast_work)->work) FD: 232 BD: 1 +.+.: (wq_completion)bond1079 ->&rq->__lock ->(work_completion)(&(&slave->notify_work)->work) ->(work_completion)(&(&bond->mcast_work)->work) FD: 232 BD: 1 +.+.: (wq_completion)bond1078#2 ->&rq->__lock ->(work_completion)(&(&slave->notify_work)->work) ->(work_completion)(&(&bond->mcast_work)->work) FD: 232 BD: 1 +.+.: (wq_completion)bond1080 ->(work_completion)(&(&slave->notify_work)->work) ->(work_completion)(&(&bond->mcast_work)->work) ->&rq->__lock FD: 232 BD: 1 +.+.: (wq_completion)bond1056#3 ->(work_completion)(&(&slave->notify_work)->work) ->(work_completion)(&(&bond->mcast_work)->work) ->&rq->__lock FD: 232 BD: 1 +.+.: (wq_completion)bond1079#2 ->(work_completion)(&(&slave->notify_work)->work) ->(work_completion)(&(&bond->mcast_work)->work) ->&rq->__lock FD: 232 BD: 1 +.+.: (wq_completion)bond1081 ->&rq->__lock ->(work_completion)(&(&slave->notify_work)->work) ->(work_completion)(&(&bond->mcast_work)->work) FD: 232 BD: 1 +.+.: (wq_completion)bond1057#3 ->&rq->__lock ->(work_completion)(&(&slave->notify_work)->work) ->(work_completion)(&(&bond->mcast_work)->work) FD: 232 BD: 1 +.+.: (wq_completion)bond1080#2 ->&rq->__lock ->(work_completion)(&(&slave->notify_work)->work) ->(work_completion)(&(&bond->mcast_work)->work) FD: 232 BD: 1 +.+.: (wq_completion)bond1082 ->&rq->__lock ->(work_completion)(&(&slave->notify_work)->work) ->(work_completion)(&(&bond->mcast_work)->work) FD: 232 BD: 1 +.+.: (wq_completion)bond1058#3 ->&rq->__lock ->(work_completion)(&(&slave->notify_work)->work) ->(work_completion)(&(&bond->mcast_work)->work) FD: 232 BD: 1 +.+.: (wq_completion)bond1083 ->&rq->__lock ->(work_completion)(&(&slave->notify_work)->work) ->(work_completion)(&(&bond->mcast_work)->work) FD: 232 BD: 1 +.+.: (wq_completion)bond1059#3 ->&rq->__lock ->(work_completion)(&(&slave->notify_work)->work) ->(work_completion)(&(&bond->mcast_work)->work) FD: 232 BD: 1 +.+.: (wq_completion)bond1081#2 ->&rq->__lock ->(work_completion)(&(&slave->notify_work)->work) ->(work_completion)(&(&bond->mcast_work)->work) FD: 232 BD: 1 +.+.: (wq_completion)bond1084 ->&rq->__lock ->(work_completion)(&(&slave->notify_work)->work) ->(work_completion)(&(&bond->mcast_work)->work) FD: 232 BD: 1 +.+.: (wq_completion)bond1082#2 ->&rq->__lock ->(work_completion)(&(&slave->notify_work)->work) ->(work_completion)(&(&bond->mcast_work)->work) FD: 232 BD: 1 +.+.: (wq_completion)bond1085 ->&rq->__lock ->(work_completion)(&(&slave->notify_work)->work) ->(work_completion)(&(&bond->mcast_work)->work) FD: 232 BD: 1 +.+.: (wq_completion)bond1060#3 ->&rq->__lock ->(work_completion)(&(&slave->notify_work)->work) ->(work_completion)(&(&bond->mcast_work)->work) FD: 232 BD: 1 +.+.: (wq_completion)bond1083#2 ->&rq->__lock ->(work_completion)(&(&slave->notify_work)->work) ->(work_completion)(&(&bond->mcast_work)->work) FD: 232 BD: 1 +.+.: (wq_completion)bond1086 ->(work_completion)(&(&slave->notify_work)->work) ->(work_completion)(&(&bond->mcast_work)->work) ->&rq->__lock FD: 232 BD: 1 +.+.: (wq_completion)bond1061#3 ->&rq->__lock ->(work_completion)(&(&slave->notify_work)->work) ->(work_completion)(&(&bond->mcast_work)->work) FD: 232 BD: 1 +.+.: (wq_completion)bond1084#2 ->(work_completion)(&(&slave->notify_work)->work) ->(work_completion)(&(&bond->mcast_work)->work) ->&rq->__lock FD: 232 BD: 1 +.+.: (wq_completion)bond1062#3 ->&rq->__lock ->(work_completion)(&(&slave->notify_work)->work) ->(work_completion)(&(&bond->mcast_work)->work) FD: 232 BD: 1 +.+.: (wq_completion)bond1087 ->(work_completion)(&(&slave->notify_work)->work) ->(work_completion)(&(&bond->mcast_work)->work) ->&rq->__lock FD: 232 BD: 1 +.+.: (wq_completion)bond1063#3 ->&rq->__lock ->(work_completion)(&(&slave->notify_work)->work) ->(work_completion)(&(&bond->mcast_work)->work) FD: 232 BD: 1 +.+.: (wq_completion)bond1085#2 ->&rq->__lock ->(work_completion)(&(&slave->notify_work)->work) ->(work_completion)(&(&bond->mcast_work)->work) FD: 232 BD: 1 +.+.: (wq_completion)bond1064#3 ->&rq->__lock ->(work_completion)(&(&slave->notify_work)->work) ->(work_completion)(&(&bond->mcast_work)->work) FD: 232 BD: 1 +.+.: (wq_completion)bond1086#2 ->&rq->__lock ->(work_completion)(&(&slave->notify_work)->work) ->(work_completion)(&(&bond->mcast_work)->work) FD: 232 BD: 1 +.+.: (wq_completion)bond1088 ->&rq->__lock ->(work_completion)(&(&slave->notify_work)->work) ->(work_completion)(&(&bond->mcast_work)->work) FD: 232 BD: 1 +.+.: (wq_completion)bond1065#3 ->&rq->__lock ->(work_completion)(&(&slave->notify_work)->work) ->(work_completion)(&(&bond->mcast_work)->work) FD: 232 BD: 1 +.+.: (wq_completion)bond1087#2 ->(work_completion)(&(&slave->notify_work)->work) ->(work_completion)(&(&bond->mcast_work)->work) ->&rq->__lock FD: 232 BD: 1 +.+.: (wq_completion)bond1089 ->(work_completion)(&(&slave->notify_work)->work) ->(work_completion)(&(&bond->mcast_work)->work) ->&rq->__lock FD: 232 BD: 1 +.+.: (wq_completion)bond1066#3 ->&rq->__lock ->(work_completion)(&(&slave->notify_work)->work) ->(work_completion)(&(&bond->mcast_work)->work) FD: 232 BD: 1 +.+.: (wq_completion)bond1088#2 ->&rq->__lock ->(work_completion)(&(&slave->notify_work)->work) ->(work_completion)(&(&bond->mcast_work)->work) FD: 232 BD: 1 +.+.: (wq_completion)bond1067#3 ->&rq->__lock ->(work_completion)(&(&slave->notify_work)->work) ->(work_completion)(&(&bond->mcast_work)->work) FD: 232 BD: 1 +.+.: (wq_completion)bond1089#2 ->&rq->__lock ->(work_completion)(&(&slave->notify_work)->work) ->(work_completion)(&(&bond->mcast_work)->work) FD: 232 BD: 1 +.+.: (wq_completion)bond1090 ->(work_completion)(&(&slave->notify_work)->work) ->(work_completion)(&(&bond->mcast_work)->work) FD: 232 BD: 1 +.+.: (wq_completion)bond1068#3 ->(work_completion)(&(&slave->notify_work)->work) ->(work_completion)(&(&bond->mcast_work)->work) ->&rq->__lock FD: 232 BD: 1 +.+.: (wq_completion)bond1090#2 ->&rq->__lock ->(work_completion)(&(&slave->notify_work)->work) ->(work_completion)(&(&bond->mcast_work)->work) FD: 232 BD: 1 +.+.: (wq_completion)bond1091 ->(work_completion)(&(&slave->notify_work)->work) ->(work_completion)(&(&bond->mcast_work)->work) ->&rq->__lock FD: 232 BD: 1 +.+.: (wq_completion)bond1069#3 ->(work_completion)(&(&slave->notify_work)->work) ->(work_completion)(&(&bond->mcast_work)->work) ->&rq->__lock FD: 232 BD: 1 +.+.: (wq_completion)bond1091#2 ->(work_completion)(&(&slave->notify_work)->work) ->(work_completion)(&(&bond->mcast_work)->work) ->&rq->__lock FD: 232 BD: 1 +.+.: (wq_completion)bond1092 ->&rq->__lock ->(work_completion)(&(&slave->notify_work)->work) ->(work_completion)(&(&bond->mcast_work)->work) FD: 232 BD: 1 +.+.: (wq_completion)bond1070#3 ->&rq->__lock ->(work_completion)(&(&slave->notify_work)->work) ->(work_completion)(&(&bond->mcast_work)->work) FD: 232 BD: 1 +.+.: (wq_completion)bond1093 ->&rq->__lock ->(work_completion)(&(&slave->notify_work)->work) ->(work_completion)(&(&bond->mcast_work)->work) FD: 232 BD: 1 +.+.: (wq_completion)bond1071#3 ->(work_completion)(&(&slave->notify_work)->work) ->(work_completion)(&(&bond->mcast_work)->work) ->&rq->__lock FD: 232 BD: 1 +.+.: (wq_completion)bond1092#2 ->&rq->__lock ->(work_completion)(&(&slave->notify_work)->work) ->(work_completion)(&(&bond->mcast_work)->work) FD: 232 BD: 1 +.+.: (wq_completion)bond1094 ->&rq->__lock ->(work_completion)(&(&slave->notify_work)->work) ->(work_completion)(&(&bond->mcast_work)->work) FD: 232 BD: 1 +.+.: (wq_completion)bond1072#3 ->&rq->__lock ->(work_completion)(&(&slave->notify_work)->work) ->(work_completion)(&(&bond->mcast_work)->work) FD: 232 BD: 1 +.+.: (wq_completion)bond1093#2 ->&rq->__lock ->(work_completion)(&(&slave->notify_work)->work) ->(work_completion)(&(&bond->mcast_work)->work) ->rcu_node_0 FD: 232 BD: 1 +.+.: (wq_completion)bond1073#3 ->&rq->__lock ->(work_completion)(&(&slave->notify_work)->work) ->(work_completion)(&(&bond->mcast_work)->work) FD: 232 BD: 1 +.+.: (wq_completion)bond1094#2 ->&rq->__lock ->(work_completion)(&(&slave->notify_work)->work) ->(work_completion)(&(&bond->mcast_work)->work) FD: 232 BD: 1 +.+.: (wq_completion)bond1095 ->&rq->__lock ->(work_completion)(&(&slave->notify_work)->work) ->(work_completion)(&(&bond->mcast_work)->work) FD: 232 BD: 1 +.+.: (wq_completion)bond1074#3 ->&rq->__lock ->(work_completion)(&(&slave->notify_work)->work) ->(work_completion)(&(&bond->mcast_work)->work) FD: 232 BD: 1 +.+.: (wq_completion)bond1095#2 ->&rq->__lock ->(work_completion)(&(&slave->notify_work)->work) ->(work_completion)(&(&bond->mcast_work)->work) FD: 232 BD: 1 +.+.: (wq_completion)bond1096 ->(work_completion)(&(&slave->notify_work)->work) ->(work_completion)(&(&bond->mcast_work)->work) ->&rq->__lock FD: 232 BD: 1 +.+.: (wq_completion)bond1075#3 ->(work_completion)(&(&slave->notify_work)->work) ->(work_completion)(&(&bond->mcast_work)->work) ->&rq->__lock FD: 232 BD: 1 +.+.: (wq_completion)bond1096#2 ->(work_completion)(&(&slave->notify_work)->work) ->(work_completion)(&(&bond->mcast_work)->work) ->&rq->__lock FD: 232 BD: 1 +.+.: (wq_completion)bond1097 ->&rq->__lock ->(work_completion)(&(&slave->notify_work)->work) ->(work_completion)(&(&bond->mcast_work)->work) FD: 232 BD: 1 +.+.: (wq_completion)bond1076#3 ->&rq->__lock ->(work_completion)(&(&slave->notify_work)->work) ->(work_completion)(&(&bond->mcast_work)->work) FD: 232 BD: 1 +.+.: (wq_completion)bond1097#2 ->&rq->__lock ->(work_completion)(&(&slave->notify_work)->work) ->(work_completion)(&(&bond->mcast_work)->work) FD: 232 BD: 1 +.+.: (wq_completion)bond1098 ->&rq->__lock ->(work_completion)(&(&slave->notify_work)->work) ->(work_completion)(&(&bond->mcast_work)->work) FD: 232 BD: 1 +.+.: (wq_completion)bond1098#2 ->(work_completion)(&(&slave->notify_work)->work) ->(work_completion)(&(&bond->mcast_work)->work) ->&rq->__lock FD: 232 BD: 1 +.+.: (wq_completion)bond1099 ->&rq->__lock ->(work_completion)(&(&slave->notify_work)->work) ->(work_completion)(&(&bond->mcast_work)->work) FD: 232 BD: 1 +.+.: (wq_completion)bond1077#3 ->&rq->__lock ->(work_completion)(&(&slave->notify_work)->work) ->(work_completion)(&(&bond->mcast_work)->work) FD: 232 BD: 1 +.+.: (wq_completion)bond1099#2 ->&rq->__lock ->(work_completion)(&(&slave->notify_work)->work) ->(work_completion)(&(&bond->mcast_work)->work) FD: 232 BD: 1 +.+.: (wq_completion)bond1100 ->&rq->__lock ->(work_completion)(&(&slave->notify_work)->work) ->(work_completion)(&(&bond->mcast_work)->work) FD: 232 BD: 1 +.+.: (wq_completion)bond1078#3 ->(work_completion)(&(&slave->notify_work)->work) ->(work_completion)(&(&bond->mcast_work)->work) ->&rq->__lock FD: 232 BD: 1 +.+.: (wq_completion)bond1100#2 ->&rq->__lock ->(work_completion)(&(&slave->notify_work)->work) ->(work_completion)(&(&bond->mcast_work)->work) FD: 232 BD: 1 +.+.: (wq_completion)bond1079#3 ->&rq->__lock ->(work_completion)(&(&slave->notify_work)->work) ->(work_completion)(&(&bond->mcast_work)->work) FD: 232 BD: 1 +.+.: (wq_completion)bond1101 ->&rq->__lock ->(work_completion)(&(&slave->notify_work)->work) ->(work_completion)(&(&bond->mcast_work)->work) FD: 232 BD: 1 +.+.: (wq_completion)bond1101#2 ->(work_completion)(&(&slave->notify_work)->work) ->(work_completion)(&(&bond->mcast_work)->work) ->&rq->__lock FD: 232 BD: 1 +.+.: (wq_completion)bond1080#3 ->&rq->__lock ->(work_completion)(&(&slave->notify_work)->work) ->(work_completion)(&(&bond->mcast_work)->work) FD: 232 BD: 1 +.+.: (wq_completion)bond1102 ->(work_completion)(&(&slave->notify_work)->work) ->(work_completion)(&(&bond->mcast_work)->work) ->&rq->__lock FD: 232 BD: 1 +.+.: (wq_completion)bond1081#3 ->(work_completion)(&(&slave->notify_work)->work) ->(work_completion)(&(&bond->mcast_work)->work) ->&rq->__lock FD: 232 BD: 1 +.+.: (wq_completion)bond1103 ->&rq->__lock ->(work_completion)(&(&slave->notify_work)->work) ->(work_completion)(&(&bond->mcast_work)->work) FD: 232 BD: 1 +.+.: (wq_completion)bond1103#2 ->&rq->__lock ->(work_completion)(&(&slave->notify_work)->work) ->(work_completion)(&(&bond->mcast_work)->work) FD: 232 BD: 1 +.+.: (wq_completion)bond1082#3 ->&rq->__lock ->(work_completion)(&(&slave->notify_work)->work) ->(work_completion)(&(&bond->mcast_work)->work) FD: 232 BD: 1 +.+.: (wq_completion)bond1104 ->&rq->__lock ->(work_completion)(&(&slave->notify_work)->work) ->(work_completion)(&(&bond->mcast_work)->work) FD: 232 BD: 1 +.+.: (wq_completion)bond1104#2 ->(work_completion)(&(&slave->notify_work)->work) ->(work_completion)(&(&bond->mcast_work)->work) ->&rq->__lock FD: 232 BD: 1 +.+.: (wq_completion)bond1105 ->&rq->__lock ->(work_completion)(&(&slave->notify_work)->work) ->(work_completion)(&(&bond->mcast_work)->work) FD: 232 BD: 1 +.+.: (wq_completion)bond1106 ->&rq->__lock ->(work_completion)(&(&slave->notify_work)->work) ->(work_completion)(&(&bond->mcast_work)->work) FD: 232 BD: 1 +.+.: (wq_completion)bond1105#2 ->&rq->__lock ->(work_completion)(&(&slave->notify_work)->work) ->(work_completion)(&(&bond->mcast_work)->work) FD: 232 BD: 1 +.+.: (wq_completion)bond1085#3 ->&rq->__lock ->(work_completion)(&(&slave->notify_work)->work) ->(work_completion)(&(&bond->mcast_work)->work) FD: 232 BD: 1 +.+.: (wq_completion)bond1106#2 ->&rq->__lock ->(work_completion)(&(&slave->notify_work)->work) ->(work_completion)(&(&bond->mcast_work)->work) FD: 232 BD: 1 +.+.: (wq_completion)bond1086#3 ->&rq->__lock ->(work_completion)(&(&slave->notify_work)->work) ->(work_completion)(&(&bond->mcast_work)->work) FD: 232 BD: 1 +.+.: (wq_completion)bond1107 ->&rq->__lock ->(work_completion)(&(&slave->notify_work)->work) ->(work_completion)(&(&bond->mcast_work)->work) FD: 232 BD: 1 +.+.: (wq_completion)bond1107#2 ->&rq->__lock ->(work_completion)(&(&slave->notify_work)->work) ->(work_completion)(&(&bond->mcast_work)->work) FD: 232 BD: 1 +.+.: (wq_completion)bond1087#3 ->(work_completion)(&(&slave->notify_work)->work) ->(work_completion)(&(&bond->mcast_work)->work) ->&rq->__lock FD: 232 BD: 1 +.+.: (wq_completion)bond1108 ->(work_completion)(&(&slave->notify_work)->work) ->(work_completion)(&(&bond->mcast_work)->work) ->&rq->__lock FD: 232 BD: 1 +.+.: (wq_completion)bond1108#2 ->(work_completion)(&(&slave->notify_work)->work) ->(work_completion)(&(&bond->mcast_work)->work) ->&rq->__lock FD: 232 BD: 1 +.+.: (wq_completion)bond1088#3 ->&rq->__lock ->(work_completion)(&(&slave->notify_work)->work) ->(work_completion)(&(&bond->mcast_work)->work) FD: 232 BD: 1 +.+.: (wq_completion)bond1089#3 ->&rq->__lock ->(work_completion)(&(&slave->notify_work)->work) ->(work_completion)(&(&bond->mcast_work)->work) FD: 232 BD: 1 +.+.: (wq_completion)bond1109 ->&rq->__lock ->(work_completion)(&(&slave->notify_work)->work) ->(work_completion)(&(&bond->mcast_work)->work) FD: 232 BD: 1 +.+.: (wq_completion)bond1109#2 ->&rq->__lock ->(work_completion)(&(&slave->notify_work)->work) ->(work_completion)(&(&bond->mcast_work)->work) FD: 232 BD: 1 +.+.: (wq_completion)bond1110 ->(work_completion)(&(&slave->notify_work)->work) ->(work_completion)(&(&bond->mcast_work)->work) ->&rq->__lock FD: 232 BD: 1 +.+.: (wq_completion)bond1110#2 ->&rq->__lock ->(work_completion)(&(&slave->notify_work)->work) ->(work_completion)(&(&bond->mcast_work)->work) FD: 232 BD: 1 +.+.: (wq_completion)bond1111 ->&rq->__lock ->(work_completion)(&(&slave->notify_work)->work) ->(work_completion)(&(&bond->mcast_work)->work) FD: 232 BD: 1 +.+.: (wq_completion)bond1111#2 ->&rq->__lock ->(work_completion)(&(&slave->notify_work)->work) ->(work_completion)(&(&bond->mcast_work)->work) FD: 232 BD: 1 +.+.: (wq_completion)bond1091#3 ->&rq->__lock ->(work_completion)(&(&slave->notify_work)->work) ->(work_completion)(&(&bond->mcast_work)->work) FD: 232 BD: 1 +.+.: (wq_completion)bond1112 ->(work_completion)(&(&slave->notify_work)->work) ->(work_completion)(&(&bond->mcast_work)->work) ->&rq->__lock FD: 232 BD: 1 +.+.: (wq_completion)bond1092#3 ->&rq->__lock ->(work_completion)(&(&slave->notify_work)->work) ->(work_completion)(&(&bond->mcast_work)->work) FD: 232 BD: 1 +.+.: (wq_completion)bond1112#2 ->&rq->__lock ->(work_completion)(&(&slave->notify_work)->work) ->(work_completion)(&(&bond->mcast_work)->work) FD: 232 BD: 1 +.+.: (wq_completion)bond1113 ->&rq->__lock ->(work_completion)(&(&slave->notify_work)->work) ->(work_completion)(&(&bond->mcast_work)->work) FD: 232 BD: 1 +.+.: (wq_completion)bond1093#3 ->(work_completion)(&(&slave->notify_work)->work) ->(work_completion)(&(&bond->mcast_work)->work) ->&rq->__lock FD: 232 BD: 1 +.+.: (wq_completion)bond1113#2 ->&rq->__lock ->(work_completion)(&(&slave->notify_work)->work) ->(work_completion)(&(&bond->mcast_work)->work) FD: 232 BD: 1 +.+.: (wq_completion)bond1114 ->&rq->__lock ->(work_completion)(&(&slave->notify_work)->work) ->(work_completion)(&(&bond->mcast_work)->work) FD: 232 BD: 1 +.+.: (wq_completion)bond1094#3 ->&rq->__lock ->(work_completion)(&(&slave->notify_work)->work) ->(work_completion)(&(&bond->mcast_work)->work) FD: 232 BD: 1 +.+.: (wq_completion)bond1114#2 ->&rq->__lock ->&cfs_rq->removed.lock ->&obj_hash[i].lock ->(work_completion)(&(&slave->notify_work)->work) ->(work_completion)(&(&bond->mcast_work)->work) FD: 232 BD: 1 +.+.: (wq_completion)bond1115 ->&rq->__lock ->(work_completion)(&(&slave->notify_work)->work) ->(work_completion)(&(&bond->mcast_work)->work) FD: 232 BD: 1 +.+.: (wq_completion)bond1115#2 ->(work_completion)(&(&slave->notify_work)->work) ->(work_completion)(&(&bond->mcast_work)->work) ->&rq->__lock FD: 232 BD: 1 +.+.: (wq_completion)bond1116 ->&rq->__lock ->(work_completion)(&(&slave->notify_work)->work) ->(work_completion)(&(&bond->mcast_work)->work) FD: 232 BD: 1 +.+.: (wq_completion)bond1095#3 ->(work_completion)(&(&slave->notify_work)->work) ->(work_completion)(&(&bond->mcast_work)->work) ->&rq->__lock FD: 232 BD: 1 +.+.: (wq_completion)bond1116#2 ->&rq->__lock ->(work_completion)(&(&slave->notify_work)->work) ->(work_completion)(&(&bond->mcast_work)->work) FD: 232 BD: 1 +.+.: (wq_completion)bond1117 ->&rq->__lock ->(work_completion)(&(&slave->notify_work)->work) ->(work_completion)(&(&bond->mcast_work)->work) FD: 232 BD: 1 +.+.: (wq_completion)bond1096#3 ->&rq->__lock ->(work_completion)(&(&slave->notify_work)->work) ->(work_completion)(&(&bond->mcast_work)->work) ->&cfs_rq->removed.lock ->&obj_hash[i].lock FD: 232 BD: 1 +.+.: (wq_completion)bond1118 ->&rq->__lock ->(work_completion)(&(&slave->notify_work)->work) ->(work_completion)(&(&bond->mcast_work)->work) FD: 232 BD: 1 +.+.: (wq_completion)bond1117#2 ->&rq->__lock ->(work_completion)(&(&slave->notify_work)->work) ->(work_completion)(&(&bond->mcast_work)->work) FD: 232 BD: 1 +.+.: (wq_completion)bond1119 ->&rq->__lock ->(work_completion)(&(&slave->notify_work)->work) ->(work_completion)(&(&bond->mcast_work)->work) FD: 232 BD: 1 +.+.: (wq_completion)bond1118#2 ->&rq->__lock ->(work_completion)(&(&slave->notify_work)->work) ->(work_completion)(&(&bond->mcast_work)->work) FD: 232 BD: 1 +.+.: (wq_completion)bond1120 ->&rq->__lock ->(work_completion)(&(&slave->notify_work)->work) ->(work_completion)(&(&bond->mcast_work)->work) FD: 232 BD: 1 +.+.: (wq_completion)bond1097#3 ->(work_completion)(&(&slave->notify_work)->work) ->(work_completion)(&(&bond->mcast_work)->work) ->&rq->__lock FD: 232 BD: 1 +.+.: (wq_completion)bond1121 ->(work_completion)(&(&slave->notify_work)->work) ->(work_completion)(&(&bond->mcast_work)->work) ->&rq->__lock FD: 232 BD: 1 +.+.: (wq_completion)bond1098#3 ->&rq->__lock ->(work_completion)(&(&slave->notify_work)->work) ->(work_completion)(&(&bond->mcast_work)->work) FD: 232 BD: 1 +.+.: (wq_completion)bond1119#2 ->&rq->__lock ->(work_completion)(&(&slave->notify_work)->work) ->(work_completion)(&(&bond->mcast_work)->work) FD: 232 BD: 1 +.+.: (wq_completion)bond1099#3 ->(work_completion)(&(&slave->notify_work)->work) ->(work_completion)(&(&bond->mcast_work)->work) ->&rq->__lock ->&cfs_rq->removed.lock ->&obj_hash[i].lock FD: 232 BD: 1 +.+.: (wq_completion)bond1120#2 ->&rq->__lock ->(work_completion)(&(&slave->notify_work)->work) ->(work_completion)(&(&bond->mcast_work)->work) FD: 232 BD: 1 +.+.: (wq_completion)bond1122 ->(work_completion)(&(&slave->notify_work)->work) ->(work_completion)(&(&bond->mcast_work)->work) ->&rq->__lock FD: 232 BD: 1 +.+.: (wq_completion)bond1100#3 ->&rq->__lock ->(work_completion)(&(&slave->notify_work)->work) ->(work_completion)(&(&bond->mcast_work)->work) FD: 232 BD: 1 +.+.: (wq_completion)bond1121#2 ->&rq->__lock ->(work_completion)(&(&slave->notify_work)->work) ->(work_completion)(&(&bond->mcast_work)->work) FD: 232 BD: 1 +.+.: (wq_completion)bond1123 ->&rq->__lock ->(work_completion)(&(&slave->notify_work)->work) ->(work_completion)(&(&bond->mcast_work)->work) FD: 232 BD: 1 +.+.: (wq_completion)bond1101#3 ->&rq->__lock ->(work_completion)(&(&slave->notify_work)->work) ->(work_completion)(&(&bond->mcast_work)->work) FD: 232 BD: 1 +.+.: (wq_completion)bond1124 ->&rq->__lock ->(work_completion)(&(&slave->notify_work)->work) ->(work_completion)(&(&bond->mcast_work)->work) FD: 232 BD: 1 +.+.: (wq_completion)bond1102#2 ->(work_completion)(&(&slave->notify_work)->work) ->(work_completion)(&(&bond->mcast_work)->work) ->&rq->__lock FD: 232 BD: 1 +.+.: (wq_completion)bond1122#2 ->&rq->__lock ->(work_completion)(&(&slave->notify_work)->work) ->(work_completion)(&(&bond->mcast_work)->work) FD: 232 BD: 1 +.+.: (wq_completion)bond1125 ->&rq->__lock ->(work_completion)(&(&slave->notify_work)->work) ->(work_completion)(&(&bond->mcast_work)->work) FD: 232 BD: 1 +.+.: (wq_completion)bond1123#2 ->&rq->__lock ->(work_completion)(&(&slave->notify_work)->work) ->(work_completion)(&(&bond->mcast_work)->work) FD: 232 BD: 1 +.+.: (wq_completion)bond1126 ->&rq->__lock ->(work_completion)(&(&slave->notify_work)->work) ->(work_completion)(&(&bond->mcast_work)->work) FD: 232 BD: 1 +.+.: (wq_completion)bond1103#3 ->&rq->__lock ->(work_completion)(&(&slave->notify_work)->work) ->(work_completion)(&(&bond->mcast_work)->work) FD: 232 BD: 1 +.+.: (wq_completion)bond1124#2 ->&rq->__lock ->(work_completion)(&(&slave->notify_work)->work) ->(work_completion)(&(&bond->mcast_work)->work) FD: 232 BD: 1 +.+.: (wq_completion)bond1127 ->(work_completion)(&(&slave->notify_work)->work) ->(work_completion)(&(&bond->mcast_work)->work) ->&rq->__lock FD: 232 BD: 1 +.+.: (wq_completion)bond1128 ->&rq->__lock ->(work_completion)(&(&slave->notify_work)->work) ->(work_completion)(&(&bond->mcast_work)->work) FD: 232 BD: 1 +.+.: (wq_completion)bond1104#3 ->&rq->__lock ->(work_completion)(&(&slave->notify_work)->work) ->(work_completion)(&(&bond->mcast_work)->work) FD: 232 BD: 1 +.+.: (wq_completion)bond1125#2 ->&rq->__lock ->(work_completion)(&(&slave->notify_work)->work) ->(work_completion)(&(&bond->mcast_work)->work) FD: 232 BD: 1 +.+.: (wq_completion)bond1129 ->&rq->__lock ->(work_completion)(&(&slave->notify_work)->work) ->(work_completion)(&(&bond->mcast_work)->work) ->&cfs_rq->removed.lock ->&obj_hash[i].lock FD: 232 BD: 1 +.+.: (wq_completion)bond1105#3 ->(work_completion)(&(&slave->notify_work)->work) ->(work_completion)(&(&bond->mcast_work)->work) ->&rq->__lock FD: 232 BD: 1 +.+.: (wq_completion)bond1126#2 ->&rq->__lock ->(work_completion)(&(&slave->notify_work)->work) ->(work_completion)(&(&bond->mcast_work)->work) FD: 232 BD: 1 +.+.: (wq_completion)bond1130 ->&rq->__lock ->(work_completion)(&(&slave->notify_work)->work) ->(work_completion)(&(&bond->mcast_work)->work) FD: 232 BD: 1 +.+.: (wq_completion)bond1106#3 ->&rq->__lock ->(work_completion)(&(&slave->notify_work)->work) ->(work_completion)(&(&bond->mcast_work)->work) FD: 232 BD: 1 +.+.: (wq_completion)bond1131 ->&rq->__lock ->(work_completion)(&(&slave->notify_work)->work) ->(work_completion)(&(&bond->mcast_work)->work) FD: 232 BD: 1 +.+.: (wq_completion)bond1107#3 ->&rq->__lock ->(work_completion)(&(&slave->notify_work)->work) ->(work_completion)(&(&bond->mcast_work)->work) FD: 232 BD: 1 +.+.: (wq_completion)bond1127#2 ->&rq->__lock ->(work_completion)(&(&slave->notify_work)->work) ->(work_completion)(&(&bond->mcast_work)->work) FD: 232 BD: 1 +.+.: (wq_completion)bond1132 ->&rq->__lock ->(work_completion)(&(&slave->notify_work)->work) ->(work_completion)(&(&bond->mcast_work)->work) FD: 232 BD: 1 +.+.: (wq_completion)bond1128#2 ->&rq->__lock ->(work_completion)(&(&slave->notify_work)->work) ->(work_completion)(&(&bond->mcast_work)->work) FD: 232 BD: 1 +.+.: (wq_completion)bond1133 ->(work_completion)(&(&slave->notify_work)->work) ->(work_completion)(&(&bond->mcast_work)->work) ->&rq->__lock FD: 232 BD: 1 +.+.: (wq_completion)bond1108#3 ->&rq->__lock ->(work_completion)(&(&slave->notify_work)->work) ->(work_completion)(&(&bond->mcast_work)->work) FD: 232 BD: 1 +.+.: (wq_completion)bond1134 ->&rq->__lock ->(work_completion)(&(&slave->notify_work)->work) ->(work_completion)(&(&bond->mcast_work)->work) FD: 232 BD: 1 +.+.: (wq_completion)bond1135 ->&rq->__lock ->(work_completion)(&(&slave->notify_work)->work) ->(work_completion)(&(&bond->mcast_work)->work) FD: 232 BD: 1 +.+.: (wq_completion)bond1130#2 ->&rq->__lock ->(work_completion)(&(&slave->notify_work)->work) ->(work_completion)(&(&bond->mcast_work)->work) FD: 232 BD: 1 +.+.: (wq_completion)bond1109#3 ->&rq->__lock ->(work_completion)(&(&slave->notify_work)->work) ->(work_completion)(&(&bond->mcast_work)->work) FD: 232 BD: 1 +.+.: (wq_completion)bond1131#2 ->&rq->__lock ->(work_completion)(&(&slave->notify_work)->work) ->(work_completion)(&(&bond->mcast_work)->work) FD: 232 BD: 1 +.+.: (wq_completion)bond1136 ->&rq->__lock ->(work_completion)(&(&slave->notify_work)->work) ->(work_completion)(&(&bond->mcast_work)->work) FD: 232 BD: 1 +.+.: (wq_completion)bond1110#3 ->&rq->__lock ->(work_completion)(&(&slave->notify_work)->work) ->(work_completion)(&(&bond->mcast_work)->work) FD: 232 BD: 1 +.+.: (wq_completion)bond1111#3 ->&rq->__lock ->(work_completion)(&(&slave->notify_work)->work) ->(work_completion)(&(&bond->mcast_work)->work) FD: 232 BD: 1 +.+.: (wq_completion)bond1132#2 ->&rq->__lock ->(work_completion)(&(&slave->notify_work)->work) ->(work_completion)(&(&bond->mcast_work)->work) FD: 232 BD: 1 +.+.: (wq_completion)bond1137 ->&rq->__lock ->(work_completion)(&(&slave->notify_work)->work) ->(work_completion)(&(&bond->mcast_work)->work) FD: 232 BD: 1 +.+.: (wq_completion)bond1112#3 ->(work_completion)(&(&slave->notify_work)->work) ->(work_completion)(&(&bond->mcast_work)->work) ->&rq->__lock FD: 232 BD: 1 +.+.: (wq_completion)bond1133#2 ->&rq->__lock ->(work_completion)(&(&slave->notify_work)->work) ->(work_completion)(&(&bond->mcast_work)->work) FD: 232 BD: 1 +.+.: (wq_completion)bond1138 ->&rq->__lock ->(work_completion)(&(&slave->notify_work)->work) ->(work_completion)(&(&bond->mcast_work)->work) FD: 232 BD: 1 +.+.: (wq_completion)bond1113#3 ->(work_completion)(&(&slave->notify_work)->work) ->(work_completion)(&(&bond->mcast_work)->work) ->&rq->__lock FD: 232 BD: 1 +.+.: (wq_completion)bond1134#2 ->&rq->__lock ->(work_completion)(&(&slave->notify_work)->work) ->(work_completion)(&(&bond->mcast_work)->work) FD: 232 BD: 1 +.+.: (wq_completion)bond1139 ->(work_completion)(&(&slave->notify_work)->work) ->(work_completion)(&(&bond->mcast_work)->work) ->&rq->__lock FD: 232 BD: 1 +.+.: (wq_completion)bond1114#3 ->&rq->__lock ->(work_completion)(&(&slave->notify_work)->work) ->(work_completion)(&(&bond->mcast_work)->work) FD: 232 BD: 1 +.+.: (wq_completion)bond1135#2 ->&rq->__lock ->(work_completion)(&(&slave->notify_work)->work) ->(work_completion)(&(&bond->mcast_work)->work) FD: 232 BD: 1 +.+.: (wq_completion)bond1140 ->&rq->__lock ->(work_completion)(&(&slave->notify_work)->work) ->(work_completion)(&(&bond->mcast_work)->work) FD: 232 BD: 1 +.+.: (wq_completion)bond1115#3 ->&rq->__lock ->(work_completion)(&(&slave->notify_work)->work) ->(work_completion)(&(&bond->mcast_work)->work) ->&cfs_rq->removed.lock ->&obj_hash[i].lock FD: 232 BD: 1 +.+.: (wq_completion)bond1136#2 ->&rq->__lock ->(work_completion)(&(&slave->notify_work)->work) ->(work_completion)(&(&bond->mcast_work)->work) FD: 232 BD: 1 +.+.: (wq_completion)bond1141 ->&rq->__lock ->(work_completion)(&(&slave->notify_work)->work) ->(work_completion)(&(&bond->mcast_work)->work) FD: 232 BD: 1 +.+.: (wq_completion)bond1116#3 ->&rq->__lock ->(work_completion)(&(&slave->notify_work)->work) ->(work_completion)(&(&bond->mcast_work)->work) FD: 232 BD: 1 +.+.: (wq_completion)bond1137#2 ->&rq->__lock ->(work_completion)(&(&slave->notify_work)->work) ->(work_completion)(&(&bond->mcast_work)->work) FD: 232 BD: 1 +.+.: (wq_completion)bond1142 ->&rq->__lock ->(work_completion)(&(&slave->notify_work)->work) ->(work_completion)(&(&bond->mcast_work)->work) FD: 232 BD: 1 +.+.: (wq_completion)bond1117#3 ->&rq->__lock ->(work_completion)(&(&slave->notify_work)->work) ->(work_completion)(&(&bond->mcast_work)->work) FD: 232 BD: 1 +.+.: (wq_completion)bond1143 ->&rq->__lock ->(work_completion)(&(&slave->notify_work)->work) ->(work_completion)(&(&bond->mcast_work)->work) FD: 232 BD: 1 +.+.: (wq_completion)bond1118#3 ->&rq->__lock ->(work_completion)(&(&slave->notify_work)->work) ->(work_completion)(&(&bond->mcast_work)->work) FD: 232 BD: 1 +.+.: (wq_completion)bond1119#3 ->&rq->__lock ->(work_completion)(&(&slave->notify_work)->work) ->(work_completion)(&(&bond->mcast_work)->work) FD: 232 BD: 1 +.+.: (wq_completion)bond1138#2 ->&rq->__lock ->(work_completion)(&(&slave->notify_work)->work) ->(work_completion)(&(&bond->mcast_work)->work) FD: 232 BD: 1 +.+.: (wq_completion)bond1144 ->&rq->__lock ->(work_completion)(&(&slave->notify_work)->work) ->(work_completion)(&(&bond->mcast_work)->work) FD: 232 BD: 1 +.+.: (wq_completion)bond1139#2 ->&rq->__lock ->(work_completion)(&(&slave->notify_work)->work) ->(work_completion)(&(&bond->mcast_work)->work) FD: 232 BD: 1 +.+.: (wq_completion)bond1145 ->&rq->__lock ->(work_completion)(&(&slave->notify_work)->work) ->(work_completion)(&(&bond->mcast_work)->work) FD: 232 BD: 1 +.+.: (wq_completion)bond1120#3 ->&rq->__lock ->(work_completion)(&(&slave->notify_work)->work) ->(work_completion)(&(&bond->mcast_work)->work) FD: 232 BD: 1 +.+.: (wq_completion)bond1140#2 ->&rq->__lock ->(work_completion)(&(&slave->notify_work)->work) ->(work_completion)(&(&bond->mcast_work)->work) FD: 232 BD: 1 +.+.: (wq_completion)bond1146 ->&rq->__lock ->(work_completion)(&(&slave->notify_work)->work) ->(work_completion)(&(&bond->mcast_work)->work) FD: 232 BD: 1 +.+.: (wq_completion)bond1121#3 ->(work_completion)(&(&slave->notify_work)->work) ->(work_completion)(&(&bond->mcast_work)->work) ->&rq->__lock FD: 232 BD: 1 +.+.: (wq_completion)bond1147 ->(work_completion)(&(&slave->notify_work)->work) ->(work_completion)(&(&bond->mcast_work)->work) ->&rq->__lock FD: 232 BD: 1 +.+.: (wq_completion)bond1141#2 ->&rq->__lock ->(work_completion)(&(&slave->notify_work)->work) ->(work_completion)(&(&bond->mcast_work)->work) FD: 232 BD: 1 +.+.: (wq_completion)bond1122#3 ->&rq->__lock ->(work_completion)(&(&slave->notify_work)->work) ->(work_completion)(&(&bond->mcast_work)->work) FD: 232 BD: 1 +.+.: (wq_completion)bond1142#2 ->&rq->__lock ->(work_completion)(&(&slave->notify_work)->work) ->(work_completion)(&(&bond->mcast_work)->work) FD: 232 BD: 1 +.+.: (wq_completion)bond1148 ->&rq->__lock ->(work_completion)(&(&slave->notify_work)->work) ->(work_completion)(&(&bond->mcast_work)->work) FD: 232 BD: 1 +.+.: (wq_completion)bond1123#3 ->(work_completion)(&(&slave->notify_work)->work) ->(work_completion)(&(&bond->mcast_work)->work) ->&rq->__lock FD: 232 BD: 1 +.+.: (wq_completion)bond1143#2 ->&rq->__lock ->(work_completion)(&(&slave->notify_work)->work) ->(work_completion)(&(&bond->mcast_work)->work) FD: 232 BD: 1 +.+.: (wq_completion)bond1149 ->&rq->__lock ->(work_completion)(&(&slave->notify_work)->work) ->(work_completion)(&(&bond->mcast_work)->work) FD: 232 BD: 1 +.+.: (wq_completion)bond1144#2 ->&rq->__lock ->(work_completion)(&(&slave->notify_work)->work) ->(work_completion)(&(&bond->mcast_work)->work) ->&cfs_rq->removed.lock ->&obj_hash[i].lock FD: 232 BD: 1 +.+.: (wq_completion)bond1145#2 ->(work_completion)(&(&slave->notify_work)->work) ->(work_completion)(&(&bond->mcast_work)->work) ->&rq->__lock FD: 232 BD: 1 +.+.: (wq_completion)bond1150 ->&rq->__lock ->(work_completion)(&(&slave->notify_work)->work) ->(work_completion)(&(&bond->mcast_work)->work) FD: 232 BD: 1 +.+.: (wq_completion)bond1124#3 ->(work_completion)(&(&slave->notify_work)->work) ->(work_completion)(&(&bond->mcast_work)->work) ->&rq->__lock FD: 232 BD: 1 +.+.: (wq_completion)bond1146#2 ->&rq->__lock ->(work_completion)(&(&slave->notify_work)->work) ->(work_completion)(&(&bond->mcast_work)->work) ->&cfs_rq->removed.lock ->&obj_hash[i].lock FD: 232 BD: 1 +.+.: (wq_completion)bond1151 ->&rq->__lock ->(work_completion)(&(&slave->notify_work)->work) ->(work_completion)(&(&bond->mcast_work)->work) FD: 232 BD: 1 +.+.: (wq_completion)bond1125#3 ->&rq->__lock ->(work_completion)(&(&slave->notify_work)->work) ->(work_completion)(&(&bond->mcast_work)->work) FD: 232 BD: 1 +.+.: (wq_completion)bond1152 ->(work_completion)(&(&slave->notify_work)->work) ->(work_completion)(&(&bond->mcast_work)->work) ->&rq->__lock ->&cfs_rq->removed.lock FD: 232 BD: 1 +.+.: (wq_completion)bond1126#3 ->&rq->__lock ->(work_completion)(&(&slave->notify_work)->work) ->(work_completion)(&(&bond->mcast_work)->work) FD: 232 BD: 1 +.+.: (wq_completion)bond1153 ->&rq->__lock ->(work_completion)(&(&slave->notify_work)->work) ->(work_completion)(&(&bond->mcast_work)->work) FD: 232 BD: 1 +.+.: (wq_completion)bond1127#3 ->(work_completion)(&(&slave->notify_work)->work) ->(work_completion)(&(&bond->mcast_work)->work) ->&rq->__lock FD: 232 BD: 1 +.+.: (wq_completion)bond1154 ->(work_completion)(&(&slave->notify_work)->work) ->(work_completion)(&(&bond->mcast_work)->work) ->&rq->__lock FD: 232 BD: 1 +.+.: (wq_completion)bond1147#2 ->&rq->__lock ->(work_completion)(&(&slave->notify_work)->work) ->(work_completion)(&(&bond->mcast_work)->work) FD: 232 BD: 1 +.+.: (wq_completion)bond1128#3 ->&rq->__lock ->(work_completion)(&(&slave->notify_work)->work) ->(work_completion)(&(&bond->mcast_work)->work) FD: 232 BD: 1 +.+.: (wq_completion)bond1148#2 ->&rq->__lock ->(work_completion)(&(&slave->notify_work)->work) ->(work_completion)(&(&bond->mcast_work)->work) FD: 232 BD: 1 +.+.: (wq_completion)bond1155 ->(work_completion)(&(&slave->notify_work)->work) ->(work_completion)(&(&bond->mcast_work)->work) ->&rq->__lock FD: 232 BD: 1 +.+.: (wq_completion)bond1129#2 ->&rq->__lock ->(work_completion)(&(&slave->notify_work)->work) ->(work_completion)(&(&bond->mcast_work)->work) FD: 232 BD: 1 +.+.: (wq_completion)bond1149#2 ->&rq->__lock ->(work_completion)(&(&slave->notify_work)->work) ->(work_completion)(&(&bond->mcast_work)->work) FD: 232 BD: 1 +.+.: (wq_completion)bond1156 ->&rq->__lock ->(work_completion)(&(&slave->notify_work)->work) ->(work_completion)(&(&bond->mcast_work)->work) FD: 232 BD: 1 +.+.: (wq_completion)bond1130#3 ->&rq->__lock ->(work_completion)(&(&slave->notify_work)->work) ->(work_completion)(&(&bond->mcast_work)->work) FD: 232 BD: 1 +.+.: (wq_completion)bond1150#2 ->&rq->__lock ->(work_completion)(&(&slave->notify_work)->work) ->(work_completion)(&(&bond->mcast_work)->work) ->&cfs_rq->removed.lock ->&obj_hash[i].lock FD: 232 BD: 1 +.+.: (wq_completion)bond1157 ->&rq->__lock ->(work_completion)(&(&slave->notify_work)->work) ->(work_completion)(&(&bond->mcast_work)->work) FD: 232 BD: 1 +.+.: (wq_completion)bond1131#3 ->&rq->__lock ->(work_completion)(&(&slave->notify_work)->work) ->(work_completion)(&(&bond->mcast_work)->work) FD: 232 BD: 1 +.+.: (wq_completion)bond1151#2 ->&rq->__lock ->(work_completion)(&(&slave->notify_work)->work) ->(work_completion)(&(&bond->mcast_work)->work) FD: 232 BD: 1 +.+.: (wq_completion)bond1158 ->&rq->__lock ->(work_completion)(&(&slave->notify_work)->work) ->(work_completion)(&(&bond->mcast_work)->work) FD: 232 BD: 1 +.+.: (wq_completion)bond1132#3 ->&rq->__lock ->(work_completion)(&(&slave->notify_work)->work) ->(work_completion)(&(&bond->mcast_work)->work) FD: 232 BD: 1 +.+.: (wq_completion)bond1152#2 ->&rq->__lock ->(work_completion)(&(&slave->notify_work)->work) ->(work_completion)(&(&bond->mcast_work)->work) FD: 232 BD: 1 +.+.: (wq_completion)bond1133#3 ->&rq->__lock ->(work_completion)(&(&slave->notify_work)->work) ->(work_completion)(&(&bond->mcast_work)->work) FD: 232 BD: 1 +.+.: (wq_completion)bond1153#2 ->&rq->__lock ->(work_completion)(&(&slave->notify_work)->work) ->(work_completion)(&(&bond->mcast_work)->work) FD: 232 BD: 1 +.+.: (wq_completion)bond1159 ->&rq->__lock ->(work_completion)(&(&slave->notify_work)->work) ->(work_completion)(&(&bond->mcast_work)->work) FD: 232 BD: 1 +.+.: (wq_completion)bond1134#3 ->&rq->__lock ->(work_completion)(&(&slave->notify_work)->work) ->(work_completion)(&(&bond->mcast_work)->work) FD: 232 BD: 1 +.+.: (wq_completion)bond1135#3 ->&rq->__lock ->(work_completion)(&(&slave->notify_work)->work) ->(work_completion)(&(&bond->mcast_work)->work) ->&cfs_rq->removed.lock ->&obj_hash[i].lock FD: 232 BD: 1 +.+.: (wq_completion)bond1154#2 ->&rq->__lock ->(work_completion)(&(&slave->notify_work)->work) ->(work_completion)(&(&bond->mcast_work)->work) FD: 232 BD: 1 +.+.: (wq_completion)bond1136#3 ->&rq->__lock ->(work_completion)(&(&slave->notify_work)->work) ->(work_completion)(&(&bond->mcast_work)->work) FD: 232 BD: 1 +.+.: (wq_completion)bond1160 ->&rq->__lock ->(work_completion)(&(&slave->notify_work)->work) ->(work_completion)(&(&bond->mcast_work)->work) FD: 232 BD: 1 +.+.: (wq_completion)bond1137#3 ->&rq->__lock ->(work_completion)(&(&slave->notify_work)->work) ->(work_completion)(&(&bond->mcast_work)->work) FD: 232 BD: 1 +.+.: (wq_completion)bond1138#3 ->(work_completion)(&(&slave->notify_work)->work) ->(work_completion)(&(&bond->mcast_work)->work) ->&rq->__lock FD: 232 BD: 1 +.+.: (wq_completion)bond1155#2 ->&rq->__lock ->(work_completion)(&(&slave->notify_work)->work) ->(work_completion)(&(&bond->mcast_work)->work) FD: 232 BD: 1 +.+.: (wq_completion)bond1161 ->&rq->__lock ->(work_completion)(&(&slave->notify_work)->work) ->(work_completion)(&(&bond->mcast_work)->work) FD: 232 BD: 1 +.+.: (wq_completion)bond1139#3 ->&rq->__lock ->(work_completion)(&(&slave->notify_work)->work) ->(work_completion)(&(&bond->mcast_work)->work) FD: 232 BD: 1 +.+.: (wq_completion)bond1156#2 ->&rq->__lock ->(work_completion)(&(&slave->notify_work)->work) ->(work_completion)(&(&bond->mcast_work)->work) FD: 232 BD: 1 +.+.: (wq_completion)bond1140#3 ->&rq->__lock ->(work_completion)(&(&slave->notify_work)->work) ->(work_completion)(&(&bond->mcast_work)->work) FD: 232 BD: 1 +.+.: (wq_completion)bond1162 ->&rq->__lock ->(work_completion)(&(&slave->notify_work)->work) ->(work_completion)(&(&bond->mcast_work)->work) FD: 232 BD: 1 +.+.: (wq_completion)bond1157#2 ->(work_completion)(&(&slave->notify_work)->work) ->(work_completion)(&(&bond->mcast_work)->work) ->&rq->__lock FD: 232 BD: 1 +.+.: (wq_completion)bond1141#3 ->&rq->__lock ->(work_completion)(&(&slave->notify_work)->work) ->(work_completion)(&(&bond->mcast_work)->work) FD: 232 BD: 1 +.+.: (wq_completion)bond1158#2 ->&rq->__lock ->(work_completion)(&(&slave->notify_work)->work) ->(work_completion)(&(&bond->mcast_work)->work) FD: 232 BD: 1 +.+.: (wq_completion)bond1142#3 ->(work_completion)(&(&slave->notify_work)->work) ->(work_completion)(&(&bond->mcast_work)->work) ->&rq->__lock FD: 232 BD: 1 +.+.: (wq_completion)bond1163 ->&rq->__lock ->(work_completion)(&(&slave->notify_work)->work) ->(work_completion)(&(&bond->mcast_work)->work) ->&cfs_rq->removed.lock ->&obj_hash[i].lock FD: 232 BD: 1 +.+.: (wq_completion)bond1159#2 ->&rq->__lock ->(work_completion)(&(&slave->notify_work)->work) ->(work_completion)(&(&bond->mcast_work)->work) FD: 232 BD: 1 +.+.: (wq_completion)bond1143#3 ->&rq->__lock ->(work_completion)(&(&slave->notify_work)->work) ->(work_completion)(&(&bond->mcast_work)->work) FD: 232 BD: 1 +.+.: (wq_completion)bond1164 ->&rq->__lock ->(work_completion)(&(&slave->notify_work)->work) ->(work_completion)(&(&bond->mcast_work)->work) FD: 232 BD: 1 +.+.: (wq_completion)bond1160#2 ->&rq->__lock ->(work_completion)(&(&slave->notify_work)->work) ->(work_completion)(&(&bond->mcast_work)->work) FD: 232 BD: 1 +.+.: (wq_completion)bond1144#3 ->(work_completion)(&(&slave->notify_work)->work) ->(work_completion)(&(&bond->mcast_work)->work) ->&rq->__lock FD: 232 BD: 1 +.+.: (wq_completion)bond1165 ->(work_completion)(&(&slave->notify_work)->work) ->(work_completion)(&(&bond->mcast_work)->work) ->&rq->__lock FD: 232 BD: 1 +.+.: (wq_completion)bond1161#2 ->&rq->__lock ->(work_completion)(&(&slave->notify_work)->work) ->(work_completion)(&(&bond->mcast_work)->work) FD: 232 BD: 1 +.+.: (wq_completion)bond1162#2 ->&rq->__lock ->(work_completion)(&(&slave->notify_work)->work) ->(work_completion)(&(&bond->mcast_work)->work) FD: 232 BD: 1 +.+.: (wq_completion)bond1145#3 ->&rq->__lock ->&cfs_rq->removed.lock ->&obj_hash[i].lock ->(work_completion)(&(&slave->notify_work)->work) ->(work_completion)(&(&bond->mcast_work)->work) FD: 232 BD: 1 +.+.: (wq_completion)bond1166 ->(work_completion)(&(&slave->notify_work)->work) ->(work_completion)(&(&bond->mcast_work)->work) ->&rq->__lock FD: 232 BD: 1 +.+.: (wq_completion)bond1163#2 ->(work_completion)(&(&slave->notify_work)->work) ->(work_completion)(&(&bond->mcast_work)->work) ->&rq->__lock FD: 232 BD: 1 +.+.: (wq_completion)bond1146#3 ->&rq->__lock ->(work_completion)(&(&slave->notify_work)->work) ->(work_completion)(&(&bond->mcast_work)->work) FD: 232 BD: 1 +.+.: (wq_completion)bond1164#2 ->&rq->__lock ->(work_completion)(&(&slave->notify_work)->work) ->(work_completion)(&(&bond->mcast_work)->work) FD: 232 BD: 1 +.+.: (wq_completion)bond1167 ->&rq->__lock ->(work_completion)(&(&slave->notify_work)->work) ->(work_completion)(&(&bond->mcast_work)->work) FD: 232 BD: 1 +.+.: (wq_completion)bond1147#3 ->&rq->__lock ->(work_completion)(&(&slave->notify_work)->work) ->(work_completion)(&(&bond->mcast_work)->work) FD: 232 BD: 1 +.+.: (wq_completion)bond1165#2 ->&rq->__lock ->&cfs_rq->removed.lock ->&obj_hash[i].lock ->(work_completion)(&(&slave->notify_work)->work) ->(work_completion)(&(&bond->mcast_work)->work) FD: 232 BD: 1 +.+.: (wq_completion)bond1168 ->&rq->__lock ->(work_completion)(&(&slave->notify_work)->work) ->(work_completion)(&(&bond->mcast_work)->work) FD: 232 BD: 1 +.+.: (wq_completion)bond1148#3 ->(work_completion)(&(&slave->notify_work)->work) ->(work_completion)(&(&bond->mcast_work)->work) ->&rq->__lock FD: 232 BD: 1 +.+.: (wq_completion)bond1166#2 ->&rq->__lock ->(work_completion)(&(&slave->notify_work)->work) ->(work_completion)(&(&bond->mcast_work)->work) FD: 232 BD: 1 +.+.: (wq_completion)bond1169 ->&rq->__lock ->(work_completion)(&(&slave->notify_work)->work) ->(work_completion)(&(&bond->mcast_work)->work) FD: 232 BD: 1 +.+.: (wq_completion)bond1167#2 ->&rq->__lock ->(work_completion)(&(&slave->notify_work)->work) ->(work_completion)(&(&bond->mcast_work)->work) FD: 232 BD: 1 +.+.: (wq_completion)bond1170 ->(work_completion)(&(&slave->notify_work)->work) ->(work_completion)(&(&bond->mcast_work)->work) ->&rq->__lock FD: 232 BD: 1 +.+.: (wq_completion)bond1168#2 ->&rq->__lock ->(work_completion)(&(&slave->notify_work)->work) ->(work_completion)(&(&bond->mcast_work)->work) FD: 232 BD: 1 +.+.: (wq_completion)bond1149#3 ->&rq->__lock ->(work_completion)(&(&slave->notify_work)->work) ->(work_completion)(&(&bond->mcast_work)->work) FD: 232 BD: 1 +.+.: (wq_completion)bond1169#2 ->&rq->__lock ->(work_completion)(&(&slave->notify_work)->work) ->(work_completion)(&(&bond->mcast_work)->work) FD: 232 BD: 1 +.+.: (wq_completion)bond1172 ->(work_completion)(&(&slave->notify_work)->work) ->(work_completion)(&(&bond->mcast_work)->work) ->&rq->__lock ->&cfs_rq->removed.lock ->&obj_hash[i].lock FD: 232 BD: 1 +.+.: (wq_completion)bond1150#3 ->(work_completion)(&(&slave->notify_work)->work) ->(work_completion)(&(&bond->mcast_work)->work) ->&rq->__lock FD: 232 BD: 1 +.+.: (wq_completion)bond1173 ->&rq->__lock ->(work_completion)(&(&slave->notify_work)->work) ->(work_completion)(&(&bond->mcast_work)->work) FD: 232 BD: 1 +.+.: (wq_completion)bond1151#3 ->&rq->__lock ->(work_completion)(&(&slave->notify_work)->work) ->(work_completion)(&(&bond->mcast_work)->work) FD: 232 BD: 1 +.+.: (wq_completion)bond1170#2 ->&rq->__lock ->(work_completion)(&(&slave->notify_work)->work) ->(work_completion)(&(&bond->mcast_work)->work) FD: 232 BD: 1 +.+.: (wq_completion)bond1174 ->&rq->__lock ->(work_completion)(&(&slave->notify_work)->work) ->(work_completion)(&(&bond->mcast_work)->work) FD: 232 BD: 1 +.+.: (wq_completion)bond1152#3 ->(work_completion)(&(&slave->notify_work)->work) ->(work_completion)(&(&bond->mcast_work)->work) ->&rq->__lock FD: 232 BD: 1 +.+.: (wq_completion)bond1171 ->&rq->__lock ->(work_completion)(&(&slave->notify_work)->work) ->(work_completion)(&(&bond->mcast_work)->work) FD: 232 BD: 1 +.+.: (wq_completion)bond1175 ->&rq->__lock ->(work_completion)(&(&slave->notify_work)->work) ->(work_completion)(&(&bond->mcast_work)->work) FD: 232 BD: 1 +.+.: (wq_completion)bond1153#3 ->(work_completion)(&(&slave->notify_work)->work) ->(work_completion)(&(&bond->mcast_work)->work) FD: 232 BD: 1 +.+.: (wq_completion)bond1176 ->&rq->__lock ->(work_completion)(&(&slave->notify_work)->work) ->(work_completion)(&(&bond->mcast_work)->work) FD: 232 BD: 1 +.+.: (wq_completion)bond1172#2 ->&rq->__lock ->(work_completion)(&(&slave->notify_work)->work) ->(work_completion)(&(&bond->mcast_work)->work) FD: 232 BD: 1 +.+.: (wq_completion)bond1154#3 ->&rq->__lock ->(work_completion)(&(&slave->notify_work)->work) ->(work_completion)(&(&bond->mcast_work)->work) FD: 232 BD: 1 +.+.: (wq_completion)bond1177 ->&rq->__lock ->(work_completion)(&(&slave->notify_work)->work) ->(work_completion)(&(&bond->mcast_work)->work) FD: 232 BD: 1 +.+.: (wq_completion)bond1155#3 ->&rq->__lock ->(work_completion)(&(&slave->notify_work)->work) ->(work_completion)(&(&bond->mcast_work)->work) FD: 232 BD: 1 +.+.: (wq_completion)bond1178 ->&rq->__lock ->(work_completion)(&(&slave->notify_work)->work) ->(work_completion)(&(&bond->mcast_work)->work) FD: 232 BD: 1 +.+.: (wq_completion)bond1174#2 ->&rq->__lock ->(work_completion)(&(&slave->notify_work)->work) ->(work_completion)(&(&bond->mcast_work)->work) FD: 232 BD: 1 +.+.: (wq_completion)bond1156#3 ->(work_completion)(&(&slave->notify_work)->work) ->(work_completion)(&(&bond->mcast_work)->work) ->&rq->__lock FD: 232 BD: 1 +.+.: (wq_completion)bond1175#2 ->&rq->__lock ->(work_completion)(&(&slave->notify_work)->work) ->(work_completion)(&(&bond->mcast_work)->work) FD: 232 BD: 1 +.+.: (wq_completion)bond1157#3 ->(work_completion)(&(&slave->notify_work)->work) ->(work_completion)(&(&bond->mcast_work)->work) ->&rq->__lock FD: 232 BD: 1 +.+.: (wq_completion)bond1179 ->&rq->__lock ->(work_completion)(&(&slave->notify_work)->work) ->(work_completion)(&(&bond->mcast_work)->work) FD: 232 BD: 1 +.+.: (wq_completion)bond1176#2 ->&rq->__lock ->(work_completion)(&(&slave->notify_work)->work) ->(work_completion)(&(&bond->mcast_work)->work) FD: 232 BD: 1 +.+.: (wq_completion)bond1158#3 ->&rq->__lock ->(work_completion)(&(&slave->notify_work)->work) ->(work_completion)(&(&bond->mcast_work)->work) FD: 232 BD: 1 +.+.: (wq_completion)bond1180 ->(work_completion)(&(&slave->notify_work)->work) ->(work_completion)(&(&bond->mcast_work)->work) ->&rq->__lock FD: 232 BD: 1 +.+.: (wq_completion)bond1159#3 ->&rq->__lock ->(work_completion)(&(&slave->notify_work)->work) ->(work_completion)(&(&bond->mcast_work)->work) FD: 232 BD: 1 +.+.: (wq_completion)bond1181 ->(work_completion)(&(&slave->notify_work)->work) ->(work_completion)(&(&bond->mcast_work)->work) ->&rq->__lock FD: 232 BD: 1 +.+.: (wq_completion)bond1178#2 ->&rq->__lock ->(work_completion)(&(&slave->notify_work)->work) ->(work_completion)(&(&bond->mcast_work)->work) FD: 232 BD: 1 +.+.: (wq_completion)bond1160#3 ->(work_completion)(&(&slave->notify_work)->work) ->(work_completion)(&(&bond->mcast_work)->work) ->&rq->__lock FD: 232 BD: 1 +.+.: (wq_completion)bond1182 ->&rq->__lock ->(work_completion)(&(&slave->notify_work)->work) ->(work_completion)(&(&bond->mcast_work)->work) FD: 232 BD: 1 +.+.: (wq_completion)bond1179#2 ->&rq->__lock ->(work_completion)(&(&slave->notify_work)->work) ->(work_completion)(&(&bond->mcast_work)->work) FD: 232 BD: 1 +.+.: (wq_completion)bond1161#3 ->(work_completion)(&(&slave->notify_work)->work) ->(work_completion)(&(&bond->mcast_work)->work) ->&rq->__lock FD: 232 BD: 1 +.+.: (wq_completion)bond1180#2 ->(work_completion)(&(&slave->notify_work)->work) ->(work_completion)(&(&bond->mcast_work)->work) ->&rq->__lock FD: 232 BD: 1 +.+.: (wq_completion)bond1162#3 ->(work_completion)(&(&slave->notify_work)->work) ->(work_completion)(&(&bond->mcast_work)->work) ->&rq->__lock FD: 232 BD: 1 +.+.: (wq_completion)bond1183 ->(work_completion)(&(&slave->notify_work)->work) ->(work_completion)(&(&bond->mcast_work)->work) ->&rq->__lock FD: 232 BD: 1 +.+.: (wq_completion)bond1181#2 ->&rq->__lock ->(work_completion)(&(&slave->notify_work)->work) ->(work_completion)(&(&bond->mcast_work)->work) FD: 232 BD: 1 +.+.: (wq_completion)bond1163#3 ->&rq->__lock ->(work_completion)(&(&slave->notify_work)->work) ->(work_completion)(&(&bond->mcast_work)->work) FD: 232 BD: 1 +.+.: (wq_completion)bond1164#3 ->&rq->__lock ->(work_completion)(&(&slave->notify_work)->work) ->(work_completion)(&(&bond->mcast_work)->work) FD: 232 BD: 1 +.+.: (wq_completion)bond1184 ->&rq->__lock ->(work_completion)(&(&slave->notify_work)->work) ->(work_completion)(&(&bond->mcast_work)->work) ->&cfs_rq->removed.lock ->&obj_hash[i].lock FD: 232 BD: 1 +.+.: (wq_completion)bond1165#3 ->(work_completion)(&(&slave->notify_work)->work) ->(work_completion)(&(&bond->mcast_work)->work) FD: 232 BD: 1 +.+.: (wq_completion)bond1185 ->(work_completion)(&(&slave->notify_work)->work) ->(work_completion)(&(&bond->mcast_work)->work) ->&rq->__lock ->&cfs_rq->removed.lock ->&obj_hash[i].lock FD: 1 BD: 5 +.+.: cmtp_sk_list.lock FD: 232 BD: 1 +.+.: (wq_completion)bond1186 ->&rq->__lock ->(work_completion)(&(&slave->notify_work)->work) ->(work_completion)(&(&bond->mcast_work)->work) FD: 232 BD: 1 +.+.: (wq_completion)bond1166#3 ->&rq->__lock ->(work_completion)(&(&slave->notify_work)->work) ->(work_completion)(&(&bond->mcast_work)->work) FD: 232 BD: 1 +.+.: (wq_completion)bond1182#2 ->&rq->__lock ->(work_completion)(&(&slave->notify_work)->work) ->(work_completion)(&(&bond->mcast_work)->work) FD: 232 BD: 1 +.+.: (wq_completion)bond1187 ->(work_completion)(&(&slave->notify_work)->work) ->(work_completion)(&(&bond->mcast_work)->work) ->&rq->__lock FD: 232 BD: 1 +.+.: (wq_completion)bond1183#2 ->&rq->__lock ->(work_completion)(&(&slave->notify_work)->work) ->(work_completion)(&(&bond->mcast_work)->work) FD: 232 BD: 1 +.+.: (wq_completion)bond1167#3 ->&rq->__lock ->(work_completion)(&(&slave->notify_work)->work) ->(work_completion)(&(&bond->mcast_work)->work) FD: 232 BD: 1 +.+.: (wq_completion)bond1184#2 ->&rq->__lock ->(work_completion)(&(&slave->notify_work)->work) ->(work_completion)(&(&bond->mcast_work)->work) FD: 232 BD: 1 +.+.: (wq_completion)bond1188 ->&rq->__lock ->(work_completion)(&(&slave->notify_work)->work) ->(work_completion)(&(&bond->mcast_work)->work) FD: 232 BD: 1 +.+.: (wq_completion)bond1185#2 ->(work_completion)(&(&slave->notify_work)->work) ->(work_completion)(&(&bond->mcast_work)->work) ->&rq->__lock FD: 232 BD: 1 +.+.: (wq_completion)bond1189 ->&rq->__lock ->(work_completion)(&(&slave->notify_work)->work) ->(work_completion)(&(&bond->mcast_work)->work) FD: 232 BD: 1 +.+.: (wq_completion)bond1186#2 ->(work_completion)(&(&slave->notify_work)->work) ->(work_completion)(&(&bond->mcast_work)->work) ->&rq->__lock FD: 232 BD: 1 +.+.: (wq_completion)bond1190 ->(work_completion)(&(&slave->notify_work)->work) ->(work_completion)(&(&bond->mcast_work)->work) ->&rq->__lock FD: 232 BD: 1 +.+.: (wq_completion)bond1168#3 ->&rq->__lock ->(work_completion)(&(&slave->notify_work)->work) ->(work_completion)(&(&bond->mcast_work)->work) FD: 232 BD: 1 +.+.: (wq_completion)bond1191 ->&rq->__lock ->(work_completion)(&(&slave->notify_work)->work) ->(work_completion)(&(&bond->mcast_work)->work) FD: 232 BD: 1 +.+.: (wq_completion)bond1169#3 ->&rq->__lock ->(work_completion)(&(&slave->notify_work)->work) ->(work_completion)(&(&bond->mcast_work)->work) FD: 232 BD: 1 +.+.: (wq_completion)bond1192 ->(work_completion)(&(&slave->notify_work)->work) ->(work_completion)(&(&bond->mcast_work)->work) ->&rq->__lock FD: 232 BD: 1 +.+.: (wq_completion)bond1187#2 ->(work_completion)(&(&slave->notify_work)->work) ->(work_completion)(&(&bond->mcast_work)->work) ->&rq->__lock FD: 232 BD: 1 +.+.: (wq_completion)bond1170#3 ->(work_completion)(&(&slave->notify_work)->work) ->(work_completion)(&(&bond->mcast_work)->work) ->&rq->__lock FD: 232 BD: 1 +.+.: (wq_completion)bond1193 ->&rq->__lock ->(work_completion)(&(&slave->notify_work)->work) ->(work_completion)(&(&bond->mcast_work)->work) FD: 232 BD: 1 +.+.: (wq_completion)bond1188#2 ->&rq->__lock ->(work_completion)(&(&slave->notify_work)->work) ->(work_completion)(&(&bond->mcast_work)->work) FD: 232 BD: 1 +.+.: (wq_completion)bond1171#2 ->&rq->__lock ->(work_completion)(&(&slave->notify_work)->work) ->(work_completion)(&(&bond->mcast_work)->work) FD: 232 BD: 1 +.+.: (wq_completion)bond1194 ->&rq->__lock ->(work_completion)(&(&slave->notify_work)->work) ->(work_completion)(&(&bond->mcast_work)->work) FD: 232 BD: 1 +.+.: (wq_completion)bond1189#2 ->&rq->__lock ->(work_completion)(&(&slave->notify_work)->work) ->(work_completion)(&(&bond->mcast_work)->work) FD: 232 BD: 1 +.+.: (wq_completion)bond1172#3 ->&rq->__lock ->(work_completion)(&(&slave->notify_work)->work) ->(work_completion)(&(&bond->mcast_work)->work) FD: 232 BD: 1 +.+.: (wq_completion)bond1190#2 ->(work_completion)(&(&slave->notify_work)->work) ->(work_completion)(&(&bond->mcast_work)->work) ->&rq->__lock FD: 232 BD: 1 +.+.: (wq_completion)bond1173#2 ->&rq->__lock ->(work_completion)(&(&slave->notify_work)->work) ->(work_completion)(&(&bond->mcast_work)->work) FD: 232 BD: 1 +.+.: (wq_completion)bond1195 ->&rq->__lock ->(work_completion)(&(&slave->notify_work)->work) ->(work_completion)(&(&bond->mcast_work)->work) FD: 232 BD: 1 +.+.: (wq_completion)bond1191#2 ->&rq->__lock ->(work_completion)(&(&slave->notify_work)->work) ->(work_completion)(&(&bond->mcast_work)->work) FD: 232 BD: 1 +.+.: (wq_completion)bond1174#3 ->&rq->__lock ->(work_completion)(&(&slave->notify_work)->work) ->(work_completion)(&(&bond->mcast_work)->work) FD: 232 BD: 1 +.+.: (wq_completion)bond1192#2 ->(work_completion)(&(&slave->notify_work)->work) ->(work_completion)(&(&bond->mcast_work)->work) ->&rq->__lock FD: 232 BD: 1 +.+.: (wq_completion)bond1175#3 ->&rq->__lock ->(work_completion)(&(&slave->notify_work)->work) ->(work_completion)(&(&bond->mcast_work)->work) FD: 232 BD: 1 +.+.: (wq_completion)bond1196 ->(work_completion)(&(&slave->notify_work)->work) ->(work_completion)(&(&bond->mcast_work)->work) ->&rq->__lock FD: 232 BD: 1 +.+.: (wq_completion)bond1176#3 ->&rq->__lock ->(work_completion)(&(&slave->notify_work)->work) ->(work_completion)(&(&bond->mcast_work)->work) FD: 232 BD: 1 +.+.: (wq_completion)bond1197 ->&rq->__lock ->(work_completion)(&(&slave->notify_work)->work) ->(work_completion)(&(&bond->mcast_work)->work) FD: 232 BD: 1 +.+.: (wq_completion)bond1177#2 ->(work_completion)(&(&slave->notify_work)->work) ->(work_completion)(&(&bond->mcast_work)->work) ->&rq->__lock FD: 232 BD: 1 +.+.: (wq_completion)bond1193#2 ->&rq->__lock ->(work_completion)(&(&slave->notify_work)->work) ->(work_completion)(&(&bond->mcast_work)->work) FD: 232 BD: 1 +.+.: (wq_completion)bond1178#3 ->&rq->__lock ->(work_completion)(&(&slave->notify_work)->work) ->(work_completion)(&(&bond->mcast_work)->work) FD: 232 BD: 1 +.+.: (wq_completion)bond1198 ->&rq->__lock ->(work_completion)(&(&slave->notify_work)->work) ->(work_completion)(&(&bond->mcast_work)->work) FD: 232 BD: 1 +.+.: (wq_completion)bond1194#2 ->(work_completion)(&(&slave->notify_work)->work) ->(work_completion)(&(&bond->mcast_work)->work) ->&rq->__lock FD: 232 BD: 1 +.+.: (wq_completion)bond1179#3 ->(work_completion)(&(&slave->notify_work)->work) ->(work_completion)(&(&bond->mcast_work)->work) ->&rq->__lock FD: 232 BD: 1 +.+.: (wq_completion)bond1199 ->&rq->__lock ->(work_completion)(&(&slave->notify_work)->work) ->(work_completion)(&(&bond->mcast_work)->work) FD: 232 BD: 1 +.+.: (wq_completion)bond1180#3 ->(work_completion)(&(&slave->notify_work)->work) ->(work_completion)(&(&bond->mcast_work)->work) ->&rq->__lock FD: 232 BD: 1 +.+.: (wq_completion)bond1181#3 ->(work_completion)(&(&slave->notify_work)->work) ->(work_completion)(&(&bond->mcast_work)->work) ->&rq->__lock FD: 232 BD: 1 +.+.: (wq_completion)bond1200 ->(work_completion)(&(&slave->notify_work)->work) ->(work_completion)(&(&bond->mcast_work)->work) ->&rq->__lock FD: 232 BD: 1 +.+.: (wq_completion)bond1195#2 ->&rq->__lock ->(work_completion)(&(&slave->notify_work)->work) ->(work_completion)(&(&bond->mcast_work)->work) FD: 232 BD: 1 +.+.: (wq_completion)bond1182#3 ->&rq->__lock ->(work_completion)(&(&slave->notify_work)->work) ->(work_completion)(&(&bond->mcast_work)->work) FD: 232 BD: 1 +.+.: (wq_completion)bond1201 ->&rq->__lock ->(work_completion)(&(&slave->notify_work)->work) ->(work_completion)(&(&bond->mcast_work)->work) FD: 232 BD: 1 +.+.: (wq_completion)bond1196#2 ->&rq->__lock ->(work_completion)(&(&slave->notify_work)->work) ->(work_completion)(&(&bond->mcast_work)->work) FD: 232 BD: 1 +.+.: (wq_completion)bond1202 ->&rq->__lock ->(work_completion)(&(&slave->notify_work)->work) ->(work_completion)(&(&bond->mcast_work)->work) FD: 232 BD: 1 +.+.: (wq_completion)bond1197#2 ->(work_completion)(&(&slave->notify_work)->work) ->(work_completion)(&(&bond->mcast_work)->work) ->&rq->__lock FD: 232 BD: 1 +.+.: (wq_completion)bond1184#3 ->&rq->__lock ->(work_completion)(&(&slave->notify_work)->work) ->(work_completion)(&(&bond->mcast_work)->work) FD: 232 BD: 1 +.+.: (wq_completion)bond1203 ->(work_completion)(&(&slave->notify_work)->work) ->(work_completion)(&(&bond->mcast_work)->work) ->&rq->__lock FD: 232 BD: 1 +.+.: (wq_completion)bond1185#3 ->&rq->__lock ->(work_completion)(&(&slave->notify_work)->work) ->(work_completion)(&(&bond->mcast_work)->work) FD: 232 BD: 1 +.+.: (wq_completion)bond1198#2 ->&rq->__lock ->(work_completion)(&(&slave->notify_work)->work) ->(work_completion)(&(&bond->mcast_work)->work) FD: 232 BD: 1 +.+.: (wq_completion)bond1204 ->&rq->__lock ->(work_completion)(&(&slave->notify_work)->work) ->(work_completion)(&(&bond->mcast_work)->work) FD: 232 BD: 1 +.+.: (wq_completion)bond1186#3 ->&rq->__lock ->(work_completion)(&(&slave->notify_work)->work) ->(work_completion)(&(&bond->mcast_work)->work) FD: 232 BD: 1 +.+.: (wq_completion)bond1199#2 ->(work_completion)(&(&slave->notify_work)->work) ->(work_completion)(&(&bond->mcast_work)->work) ->&rq->__lock FD: 232 BD: 1 +.+.: (wq_completion)bond1187#3 ->&rq->__lock ->(work_completion)(&(&slave->notify_work)->work) ->(work_completion)(&(&bond->mcast_work)->work) FD: 232 BD: 1 +.+.: (wq_completion)bond1205 ->&rq->__lock ->(work_completion)(&(&slave->notify_work)->work) ->(work_completion)(&(&bond->mcast_work)->work) FD: 232 BD: 1 +.+.: (wq_completion)bond1200#2 ->&rq->__lock ->(work_completion)(&(&slave->notify_work)->work) ->(work_completion)(&(&bond->mcast_work)->work) FD: 232 BD: 1 +.+.: (wq_completion)bond1188#3 ->&rq->__lock ->(work_completion)(&(&slave->notify_work)->work) ->(work_completion)(&(&bond->mcast_work)->work) FD: 232 BD: 1 +.+.: (wq_completion)bond1206 ->&rq->__lock ->(work_completion)(&(&slave->notify_work)->work) ->(work_completion)(&(&bond->mcast_work)->work) ->&cfs_rq->removed.lock ->&obj_hash[i].lock FD: 232 BD: 1 +.+.: (wq_completion)bond1201#2 ->&rq->__lock ->(work_completion)(&(&slave->notify_work)->work) ->(work_completion)(&(&bond->mcast_work)->work) FD: 232 BD: 1 +.+.: (wq_completion)bond1189#3 ->(work_completion)(&(&slave->notify_work)->work) ->(work_completion)(&(&bond->mcast_work)->work) ->&rq->__lock FD: 232 BD: 1 +.+.: (wq_completion)bond1207 ->&rq->__lock ->(work_completion)(&(&slave->notify_work)->work) ->(work_completion)(&(&bond->mcast_work)->work) FD: 232 BD: 1 +.+.: (wq_completion)bond1208 ->&rq->__lock ->(work_completion)(&(&slave->notify_work)->work) ->(work_completion)(&(&bond->mcast_work)->work) FD: 232 BD: 1 +.+.: (wq_completion)bond1190#3 ->&rq->__lock ->(work_completion)(&(&slave->notify_work)->work) ->(work_completion)(&(&bond->mcast_work)->work) FD: 232 BD: 1 +.+.: (wq_completion)bond1209 ->&rq->__lock ->(work_completion)(&(&slave->notify_work)->work) ->(work_completion)(&(&bond->mcast_work)->work) FD: 232 BD: 1 +.+.: (wq_completion)bond1210 ->&rq->__lock ->(work_completion)(&(&slave->notify_work)->work) ->(work_completion)(&(&bond->mcast_work)->work) FD: 232 BD: 1 +.+.: (wq_completion)bond1202#2 ->&rq->__lock ->(work_completion)(&(&slave->notify_work)->work) ->(work_completion)(&(&bond->mcast_work)->work) FD: 232 BD: 1 +.+.: (wq_completion)bond1191#3 ->(work_completion)(&(&slave->notify_work)->work) ->(work_completion)(&(&bond->mcast_work)->work) ->&rq->__lock FD: 232 BD: 1 +.+.: (wq_completion)bond1211 ->(work_completion)(&(&slave->notify_work)->work) ->(work_completion)(&(&bond->mcast_work)->work) ->&rq->__lock ->&cfs_rq->removed.lock ->&obj_hash[i].lock FD: 232 BD: 1 +.+.: (wq_completion)bond1203#2 ->(work_completion)(&(&slave->notify_work)->work) ->(work_completion)(&(&bond->mcast_work)->work) ->&rq->__lock FD: 232 BD: 1 +.+.: (wq_completion)bond1192#3 ->(work_completion)(&(&slave->notify_work)->work) ->(work_completion)(&(&bond->mcast_work)->work) ->&rq->__lock FD: 232 BD: 1 +.+.: (wq_completion)bond1212 ->&rq->__lock ->(work_completion)(&(&slave->notify_work)->work) ->(work_completion)(&(&bond->mcast_work)->work) FD: 232 BD: 1 +.+.: (wq_completion)bond1213 ->&rq->__lock ->(work_completion)(&(&slave->notify_work)->work) ->(work_completion)(&(&bond->mcast_work)->work) FD: 232 BD: 1 +.+.: (wq_completion)bond1214 ->&rq->__lock ->(work_completion)(&(&slave->notify_work)->work) ->(work_completion)(&(&bond->mcast_work)->work) FD: 232 BD: 1 +.+.: (wq_completion)bond1193#3 ->&rq->__lock ->(work_completion)(&(&slave->notify_work)->work) ->(work_completion)(&(&bond->mcast_work)->work) ->&cfs_rq->removed.lock ->&obj_hash[i].lock FD: 232 BD: 1 +.+.: (wq_completion)bond1215 ->&rq->__lock ->(work_completion)(&(&slave->notify_work)->work) ->(work_completion)(&(&bond->mcast_work)->work) FD: 232 BD: 1 +.+.: (wq_completion)bond1194#3 ->&rq->__lock ->(work_completion)(&(&slave->notify_work)->work) ->(work_completion)(&(&bond->mcast_work)->work) FD: 232 BD: 1 +.+.: (wq_completion)bond1216 ->&rq->__lock ->(work_completion)(&(&slave->notify_work)->work) ->(work_completion)(&(&bond->mcast_work)->work) FD: 232 BD: 1 +.+.: (wq_completion)bond1195#3 ->&rq->__lock ->(work_completion)(&(&slave->notify_work)->work) ->(work_completion)(&(&bond->mcast_work)->work) FD: 232 BD: 1 +.+.: (wq_completion)bond1204#2 ->(work_completion)(&(&slave->notify_work)->work) ->(work_completion)(&(&bond->mcast_work)->work) ->&rq->__lock FD: 232 BD: 1 +.+.: (wq_completion)bond1217 ->(work_completion)(&(&slave->notify_work)->work) ->(work_completion)(&(&bond->mcast_work)->work) ->&rq->__lock FD: 232 BD: 1 +.+.: (wq_completion)bond1196#3 ->&rq->__lock ->(work_completion)(&(&slave->notify_work)->work) ->(work_completion)(&(&bond->mcast_work)->work) FD: 232 BD: 1 +.+.: (wq_completion)bond1205#2 ->&rq->__lock ->(work_completion)(&(&slave->notify_work)->work) ->(work_completion)(&(&bond->mcast_work)->work) FD: 232 BD: 1 +.+.: (wq_completion)bond1218 ->(work_completion)(&(&slave->notify_work)->work) ->(work_completion)(&(&bond->mcast_work)->work) ->&rq->__lock FD: 232 BD: 1 +.+.: (wq_completion)bond1206#2 ->&rq->__lock ->(work_completion)(&(&slave->notify_work)->work) ->(work_completion)(&(&bond->mcast_work)->work) FD: 232 BD: 1 +.+.: (wq_completion)bond1207#2 ->&rq->__lock ->(work_completion)(&(&slave->notify_work)->work) ->(work_completion)(&(&bond->mcast_work)->work) FD: 232 BD: 1 +.+.: (wq_completion)bond1198#3 ->&rq->__lock ->(work_completion)(&(&slave->notify_work)->work) ->(work_completion)(&(&bond->mcast_work)->work) FD: 232 BD: 1 +.+.: (wq_completion)bond1208#2 ->&rq->__lock ->(work_completion)(&(&slave->notify_work)->work) ->(work_completion)(&(&bond->mcast_work)->work) FD: 232 BD: 1 +.+.: (wq_completion)bond1199#3 ->&rq->__lock ->(work_completion)(&(&slave->notify_work)->work) ->(work_completion)(&(&bond->mcast_work)->work) FD: 232 BD: 1 +.+.: (wq_completion)bond1209#2 ->(work_completion)(&(&slave->notify_work)->work) ->(work_completion)(&(&bond->mcast_work)->work) ->&rq->__lock FD: 232 BD: 1 +.+.: (wq_completion)bond1220 ->&rq->__lock ->(work_completion)(&(&slave->notify_work)->work) ->(work_completion)(&(&bond->mcast_work)->work) FD: 232 BD: 1 +.+.: (wq_completion)bond1200#3 ->(work_completion)(&(&slave->notify_work)->work) ->(work_completion)(&(&bond->mcast_work)->work) ->&rq->__lock FD: 232 BD: 1 +.+.: (wq_completion)bond1210#2 ->&rq->__lock ->(work_completion)(&(&slave->notify_work)->work) ->(work_completion)(&(&bond->mcast_work)->work) FD: 232 BD: 1 +.+.: (wq_completion)bond1221 ->&rq->__lock ->(work_completion)(&(&slave->notify_work)->work) ->(work_completion)(&(&bond->mcast_work)->work) ->rcu_node_0 FD: 232 BD: 1 +.+.: (wq_completion)bond1211#2 ->&rq->__lock ->(work_completion)(&(&slave->notify_work)->work) ->(work_completion)(&(&bond->mcast_work)->work) FD: 232 BD: 1 +.+.: (wq_completion)bond1222 ->&rq->__lock ->(work_completion)(&(&slave->notify_work)->work) ->(work_completion)(&(&bond->mcast_work)->work) FD: 232 BD: 1 +.+.: (wq_completion)bond1201#3 ->&rq->__lock ->(work_completion)(&(&slave->notify_work)->work) ->(work_completion)(&(&bond->mcast_work)->work) FD: 232 BD: 1 +.+.: (wq_completion)bond1212#2 ->&rq->__lock ->(work_completion)(&(&slave->notify_work)->work) ->(work_completion)(&(&bond->mcast_work)->work) FD: 232 BD: 1 +.+.: (wq_completion)bond1223 ->&rq->__lock ->(work_completion)(&(&slave->notify_work)->work) ->(work_completion)(&(&bond->mcast_work)->work) FD: 232 BD: 1 +.+.: (wq_completion)bond1224 ->(work_completion)(&(&slave->notify_work)->work) ->(work_completion)(&(&bond->mcast_work)->work) ->&rq->__lock FD: 232 BD: 1 +.+.: (wq_completion)bond1202#3 ->&rq->__lock ->(work_completion)(&(&slave->notify_work)->work) ->(work_completion)(&(&bond->mcast_work)->work) FD: 232 BD: 1 +.+.: (wq_completion)bond1213#2 ->(work_completion)(&(&slave->notify_work)->work) ->(work_completion)(&(&bond->mcast_work)->work) ->&rq->__lock FD: 232 BD: 1 +.+.: (wq_completion)bond1225 ->&rq->__lock ->(work_completion)(&(&slave->notify_work)->work) ->(work_completion)(&(&bond->mcast_work)->work) FD: 232 BD: 1 +.+.: (wq_completion)bond1214#2 ->&rq->__lock ->(work_completion)(&(&slave->notify_work)->work) ->(work_completion)(&(&bond->mcast_work)->work) FD: 232 BD: 1 +.+.: (wq_completion)bond1226 ->&rq->__lock ->(work_completion)(&(&slave->notify_work)->work) ->(work_completion)(&(&bond->mcast_work)->work) FD: 232 BD: 1 +.+.: (wq_completion)bond1215#2 ->&rq->__lock ->(work_completion)(&(&slave->notify_work)->work) ->(work_completion)(&(&bond->mcast_work)->work) FD: 232 BD: 1 +.+.: (wq_completion)bond1203#3 ->&rq->__lock ->(work_completion)(&(&slave->notify_work)->work) ->(work_completion)(&(&bond->mcast_work)->work) FD: 232 BD: 1 +.+.: (wq_completion)bond1216#2 ->(work_completion)(&(&slave->notify_work)->work) ->(work_completion)(&(&bond->mcast_work)->work) FD: 232 BD: 1 +.+.: (wq_completion)bond1227 ->&rq->__lock ->(work_completion)(&(&slave->notify_work)->work) ->(work_completion)(&(&bond->mcast_work)->work) FD: 232 BD: 1 +.+.: (wq_completion)bond1217#2 ->&rq->__lock ->(work_completion)(&(&slave->notify_work)->work) ->(work_completion)(&(&bond->mcast_work)->work) FD: 232 BD: 1 +.+.: (wq_completion)bond1228 ->(work_completion)(&(&slave->notify_work)->work) ->(work_completion)(&(&bond->mcast_work)->work) ->&rq->__lock FD: 232 BD: 1 +.+.: (wq_completion)bond1205#3 ->&rq->__lock ->(work_completion)(&(&slave->notify_work)->work) ->(work_completion)(&(&bond->mcast_work)->work) FD: 232 BD: 1 +.+.: (wq_completion)bond1218#2 ->&rq->__lock ->(work_completion)(&(&slave->notify_work)->work) ->(work_completion)(&(&bond->mcast_work)->work) FD: 232 BD: 1 +.+.: (wq_completion)bond1229 ->(work_completion)(&(&slave->notify_work)->work) ->(work_completion)(&(&bond->mcast_work)->work) ->&rq->__lock FD: 232 BD: 1 +.+.: (wq_completion)bond1219 ->&rq->__lock ->(work_completion)(&(&slave->notify_work)->work) ->(work_completion)(&(&bond->mcast_work)->work) FD: 232 BD: 1 +.+.: (wq_completion)bond1230 ->&rq->__lock ->(work_completion)(&(&slave->notify_work)->work) ->(work_completion)(&(&bond->mcast_work)->work) FD: 232 BD: 1 +.+.: (wq_completion)bond1231 ->&rq->__lock ->(work_completion)(&(&slave->notify_work)->work) ->(work_completion)(&(&bond->mcast_work)->work) FD: 232 BD: 1 +.+.: (wq_completion)bond1206#3 ->&rq->__lock ->(work_completion)(&(&slave->notify_work)->work) ->(work_completion)(&(&bond->mcast_work)->work) ->&cfs_rq->removed.lock ->&obj_hash[i].lock FD: 232 BD: 1 +.+.: (wq_completion)bond1207#3 ->&rq->__lock ->(work_completion)(&(&slave->notify_work)->work) ->(work_completion)(&(&bond->mcast_work)->work) FD: 232 BD: 1 +.+.: (wq_completion)bond1220#2 ->&rq->__lock ->(work_completion)(&(&slave->notify_work)->work) ->(work_completion)(&(&bond->mcast_work)->work) FD: 232 BD: 1 +.+.: (wq_completion)bond1221#2 ->&rq->__lock ->(work_completion)(&(&slave->notify_work)->work) ->(work_completion)(&(&bond->mcast_work)->work) FD: 232 BD: 1 +.+.: (wq_completion)bond1232 ->&rq->__lock ->(work_completion)(&(&slave->notify_work)->work) ->(work_completion)(&(&bond->mcast_work)->work) FD: 232 BD: 1 +.+.: (wq_completion)bond1208#3 ->&rq->__lock ->(work_completion)(&(&slave->notify_work)->work) ->(work_completion)(&(&bond->mcast_work)->work) FD: 232 BD: 1 +.+.: (wq_completion)bond1222#2 ->&rq->__lock ->(work_completion)(&(&slave->notify_work)->work) ->(work_completion)(&(&bond->mcast_work)->work) FD: 232 BD: 1 +.+.: (wq_completion)bond1233 ->&rq->__lock ->(work_completion)(&(&slave->notify_work)->work) ->(work_completion)(&(&bond->mcast_work)->work) FD: 232 BD: 1 +.+.: (wq_completion)bond1209#3 ->&rq->__lock ->(work_completion)(&(&slave->notify_work)->work) ->(work_completion)(&(&bond->mcast_work)->work) FD: 232 BD: 1 +.+.: (wq_completion)bond1223#2 ->(work_completion)(&(&slave->notify_work)->work) ->(work_completion)(&(&bond->mcast_work)->work) ->&rq->__lock FD: 232 BD: 1 +.+.: (wq_completion)bond1234 ->(work_completion)(&(&slave->notify_work)->work) ->(work_completion)(&(&bond->mcast_work)->work) ->&rq->__lock FD: 232 BD: 1 +.+.: (wq_completion)bond1210#3 ->&rq->__lock ->(work_completion)(&(&slave->notify_work)->work) ->(work_completion)(&(&bond->mcast_work)->work) FD: 232 BD: 1 +.+.: (wq_completion)bond1211#3 ->&rq->__lock ->(work_completion)(&(&slave->notify_work)->work) ->(work_completion)(&(&bond->mcast_work)->work) FD: 232 BD: 1 +.+.: (wq_completion)bond1224#2 ->&rq->__lock ->(work_completion)(&(&slave->notify_work)->work) ->(work_completion)(&(&bond->mcast_work)->work) FD: 232 BD: 1 +.+.: (wq_completion)bond1235 ->&rq->__lock ->(work_completion)(&(&slave->notify_work)->work) ->(work_completion)(&(&bond->mcast_work)->work) FD: 232 BD: 1 +.+.: (wq_completion)bond1212#3 ->&rq->__lock ->(work_completion)(&(&slave->notify_work)->work) ->(work_completion)(&(&bond->mcast_work)->work) FD: 232 BD: 1 +.+.: (wq_completion)bond1236 ->&rq->__lock ->(work_completion)(&(&slave->notify_work)->work) ->(work_completion)(&(&bond->mcast_work)->work) FD: 232 BD: 1 +.+.: (wq_completion)bond1213#3 ->&rq->__lock ->(work_completion)(&(&slave->notify_work)->work) ->(work_completion)(&(&bond->mcast_work)->work) FD: 232 BD: 1 +.+.: (wq_completion)bond1225#2 ->(work_completion)(&(&slave->notify_work)->work) ->(work_completion)(&(&bond->mcast_work)->work) ->&rq->__lock FD: 232 BD: 1 +.+.: (wq_completion)bond1214#3 ->&rq->__lock ->(work_completion)(&(&slave->notify_work)->work) ->(work_completion)(&(&bond->mcast_work)->work) FD: 232 BD: 1 +.+.: (wq_completion)bond1226#2 ->&rq->__lock ->(work_completion)(&(&slave->notify_work)->work) ->(work_completion)(&(&bond->mcast_work)->work) FD: 232 BD: 1 +.+.: (wq_completion)bond1227#2 ->&rq->__lock ->(work_completion)(&(&slave->notify_work)->work) ->(work_completion)(&(&bond->mcast_work)->work) FD: 232 BD: 1 +.+.: (wq_completion)bond1228#2 ->&rq->__lock ->(work_completion)(&(&slave->notify_work)->work) ->(work_completion)(&(&bond->mcast_work)->work) FD: 232 BD: 1 +.+.: (wq_completion)bond1229#2 ->&rq->__lock ->(work_completion)(&(&slave->notify_work)->work) ->(work_completion)(&(&bond->mcast_work)->work) FD: 232 BD: 1 +.+.: (wq_completion)bond1237 ->&rq->__lock ->(work_completion)(&(&slave->notify_work)->work) ->(work_completion)(&(&bond->mcast_work)->work) FD: 232 BD: 1 +.+.: (wq_completion)bond1215#3 ->(work_completion)(&(&slave->notify_work)->work) ->(work_completion)(&(&bond->mcast_work)->work) ->&rq->__lock FD: 232 BD: 1 +.+.: (wq_completion)bond1238 ->&rq->__lock ->(work_completion)(&(&slave->notify_work)->work) ->(work_completion)(&(&bond->mcast_work)->work) FD: 232 BD: 1 +.+.: (wq_completion)bond1216#3 ->&rq->__lock ->(work_completion)(&(&slave->notify_work)->work) ->(work_completion)(&(&bond->mcast_work)->work) FD: 232 BD: 1 +.+.: (wq_completion)bond1231#2 ->&rq->__lock ->(work_completion)(&(&slave->notify_work)->work) ->(work_completion)(&(&bond->mcast_work)->work) ->&cfs_rq->removed.lock ->&obj_hash[i].lock FD: 232 BD: 1 +.+.: (wq_completion)bond1239 ->&rq->__lock ->(work_completion)(&(&slave->notify_work)->work) ->(work_completion)(&(&bond->mcast_work)->work) FD: 232 BD: 1 +.+.: (wq_completion)bond1217#3 ->(work_completion)(&(&slave->notify_work)->work) ->(work_completion)(&(&bond->mcast_work)->work) FD: 232 BD: 1 +.+.: (wq_completion)bond1218#3 ->&rq->__lock ->(work_completion)(&(&slave->notify_work)->work) ->(work_completion)(&(&bond->mcast_work)->work) FD: 232 BD: 1 +.+.: (wq_completion)bond1232#2 ->(work_completion)(&(&slave->notify_work)->work) ->(work_completion)(&(&bond->mcast_work)->work) ->&rq->__lock FD: 232 BD: 1 +.+.: (wq_completion)bond1241 ->(work_completion)(&(&slave->notify_work)->work) ->(work_completion)(&(&bond->mcast_work)->work) ->&rq->__lock FD: 232 BD: 1 +.+.: (wq_completion)bond1219#2 ->&rq->__lock ->(work_completion)(&(&slave->notify_work)->work) ->(work_completion)(&(&bond->mcast_work)->work) FD: 232 BD: 1 +.+.: (wq_completion)bond1233#2 ->&rq->__lock ->(work_completion)(&(&slave->notify_work)->work) ->(work_completion)(&(&bond->mcast_work)->work) FD: 232 BD: 1 +.+.: (wq_completion)bond1234#2 ->&rq->__lock ->(work_completion)(&(&slave->notify_work)->work) ->(work_completion)(&(&bond->mcast_work)->work) FD: 232 BD: 1 +.+.: (wq_completion)bond1242 ->&rq->__lock ->(work_completion)(&(&slave->notify_work)->work) ->(work_completion)(&(&bond->mcast_work)->work) FD: 232 BD: 1 +.+.: (wq_completion)bond1220#3 ->(work_completion)(&(&slave->notify_work)->work) ->(work_completion)(&(&bond->mcast_work)->work) FD: 232 BD: 1 +.+.: (wq_completion)bond1221#3 ->(work_completion)(&(&slave->notify_work)->work) ->(work_completion)(&(&bond->mcast_work)->work) ->&rq->__lock FD: 232 BD: 1 +.+.: (wq_completion)bond1243 ->&rq->__lock ->(work_completion)(&(&slave->notify_work)->work) ->(work_completion)(&(&bond->mcast_work)->work) FD: 232 BD: 1 +.+.: (wq_completion)bond1222#3 ->(work_completion)(&(&slave->notify_work)->work) ->(work_completion)(&(&bond->mcast_work)->work) ->&rq->__lock FD: 232 BD: 1 +.+.: (wq_completion)bond1235#2 ->&rq->__lock ->(work_completion)(&(&slave->notify_work)->work) ->(work_completion)(&(&bond->mcast_work)->work) FD: 232 BD: 1 +.+.: (wq_completion)bond1244 ->&rq->__lock ->(work_completion)(&(&slave->notify_work)->work) ->(work_completion)(&(&bond->mcast_work)->work) FD: 232 BD: 1 +.+.: (wq_completion)bond1236#2 ->&rq->__lock ->(work_completion)(&(&slave->notify_work)->work) ->(work_completion)(&(&bond->mcast_work)->work) FD: 232 BD: 1 +.+.: (wq_completion)bond1245 ->(work_completion)(&(&slave->notify_work)->work) ->(work_completion)(&(&bond->mcast_work)->work) ->&rq->__lock ->&cfs_rq->removed.lock ->&obj_hash[i].lock FD: 232 BD: 1 +.+.: (wq_completion)bond1223#3 ->&rq->__lock ->(work_completion)(&(&slave->notify_work)->work) ->(work_completion)(&(&bond->mcast_work)->work) FD: 232 BD: 1 +.+.: (wq_completion)bond1237#2 ->&rq->__lock ->(work_completion)(&(&slave->notify_work)->work) ->(work_completion)(&(&bond->mcast_work)->work) FD: 232 BD: 1 +.+.: (wq_completion)bond1246 ->&rq->__lock ->(work_completion)(&(&slave->notify_work)->work) ->(work_completion)(&(&bond->mcast_work)->work) FD: 232 BD: 1 +.+.: (wq_completion)bond1224#3 ->&rq->__lock ->(work_completion)(&(&slave->notify_work)->work) ->(work_completion)(&(&bond->mcast_work)->work) FD: 232 BD: 1 +.+.: (wq_completion)bond1238#2 ->(work_completion)(&(&slave->notify_work)->work) ->(work_completion)(&(&bond->mcast_work)->work) FD: 232 BD: 1 +.+.: (wq_completion)bond1247 ->&rq->__lock ->(work_completion)(&(&slave->notify_work)->work) ->(work_completion)(&(&bond->mcast_work)->work) FD: 232 BD: 1 +.+.: (wq_completion)bond1225#3 ->&rq->__lock ->(work_completion)(&(&slave->notify_work)->work) ->(work_completion)(&(&bond->mcast_work)->work) FD: 232 BD: 1 +.+.: (wq_completion)bond1239#2 ->&rq->__lock ->(work_completion)(&(&slave->notify_work)->work) ->(work_completion)(&(&bond->mcast_work)->work) FD: 232 BD: 1 +.+.: (wq_completion)bond1226#3 ->&rq->__lock ->(work_completion)(&(&slave->notify_work)->work) ->(work_completion)(&(&bond->mcast_work)->work) FD: 232 BD: 1 +.+.: (wq_completion)bond1240 ->&rq->__lock ->(work_completion)(&(&slave->notify_work)->work) ->(work_completion)(&(&bond->mcast_work)->work) FD: 232 BD: 1 +.+.: (wq_completion)bond1241#2 ->&rq->__lock ->(work_completion)(&(&slave->notify_work)->work) ->(work_completion)(&(&bond->mcast_work)->work) FD: 232 BD: 1 +.+.: (wq_completion)bond1249 ->&rq->__lock ->(work_completion)(&(&slave->notify_work)->work) ->(work_completion)(&(&bond->mcast_work)->work) FD: 232 BD: 1 +.+.: (wq_completion)bond1227#3 ->&rq->__lock ->(work_completion)(&(&slave->notify_work)->work) ->(work_completion)(&(&bond->mcast_work)->work) FD: 232 BD: 1 +.+.: (wq_completion)bond1242#2 ->(work_completion)(&(&slave->notify_work)->work) ->(work_completion)(&(&bond->mcast_work)->work) ->&rq->__lock FD: 232 BD: 1 +.+.: (wq_completion)bond1250 ->&rq->__lock ->(work_completion)(&(&slave->notify_work)->work) ->(work_completion)(&(&bond->mcast_work)->work) FD: 232 BD: 1 +.+.: (wq_completion)bond1243#2 ->&rq->__lock ->(work_completion)(&(&slave->notify_work)->work) ->(work_completion)(&(&bond->mcast_work)->work) FD: 232 BD: 1 +.+.: (wq_completion)bond1251 ->&rq->__lock ->(work_completion)(&(&slave->notify_work)->work) ->(work_completion)(&(&bond->mcast_work)->work) FD: 232 BD: 1 +.+.: (wq_completion)bond1229#3 ->&rq->__lock ->(work_completion)(&(&slave->notify_work)->work) ->(work_completion)(&(&bond->mcast_work)->work) FD: 232 BD: 1 +.+.: (wq_completion)bond1244#2 ->&rq->__lock ->(work_completion)(&(&slave->notify_work)->work) ->(work_completion)(&(&bond->mcast_work)->work) FD: 232 BD: 1 +.+.: (wq_completion)bond1252 ->(work_completion)(&(&slave->notify_work)->work) ->(work_completion)(&(&bond->mcast_work)->work) ->&rq->__lock FD: 232 BD: 1 +.+.: (wq_completion)bond1230#2 ->&rq->__lock ->(work_completion)(&(&slave->notify_work)->work) ->(work_completion)(&(&bond->mcast_work)->work) FD: 232 BD: 1 +.+.: (wq_completion)bond1245#2 ->&rq->__lock ->(work_completion)(&(&slave->notify_work)->work) ->(work_completion)(&(&bond->mcast_work)->work) FD: 232 BD: 1 +.+.: (wq_completion)bond1253 ->&rq->__lock ->(work_completion)(&(&slave->notify_work)->work) ->(work_completion)(&(&bond->mcast_work)->work) FD: 232 BD: 1 +.+.: (wq_completion)bond1231#3 ->&rq->__lock ->(work_completion)(&(&slave->notify_work)->work) ->(work_completion)(&(&bond->mcast_work)->work) FD: 232 BD: 1 +.+.: (wq_completion)bond1254 ->&rq->__lock ->&cfs_rq->removed.lock ->&obj_hash[i].lock ->(work_completion)(&(&slave->notify_work)->work) ->(work_completion)(&(&bond->mcast_work)->work) FD: 232 BD: 1 +.+.: (wq_completion)bond1232#3 ->&rq->__lock ->(work_completion)(&(&slave->notify_work)->work) ->(work_completion)(&(&bond->mcast_work)->work) FD: 232 BD: 1 +.+.: (wq_completion)bond1255 ->&rq->__lock ->(work_completion)(&(&slave->notify_work)->work) ->(work_completion)(&(&bond->mcast_work)->work) FD: 232 BD: 1 +.+.: (wq_completion)bond1233#3 ->(work_completion)(&(&slave->notify_work)->work) ->(work_completion)(&(&bond->mcast_work)->work) ->&rq->__lock FD: 232 BD: 1 +.+.: (wq_completion)bond1256 ->&rq->__lock ->(work_completion)(&(&slave->notify_work)->work) ->(work_completion)(&(&bond->mcast_work)->work) FD: 232 BD: 1 +.+.: (wq_completion)bond1234#3 ->&rq->__lock ->(work_completion)(&(&slave->notify_work)->work) ->(work_completion)(&(&bond->mcast_work)->work) FD: 232 BD: 1 +.+.: (wq_completion)bond1246#2 ->&rq->__lock ->(work_completion)(&(&slave->notify_work)->work) ->(work_completion)(&(&bond->mcast_work)->work) FD: 232 BD: 1 +.+.: (wq_completion)bond1257 ->(work_completion)(&(&slave->notify_work)->work) ->(work_completion)(&(&bond->mcast_work)->work) ->&rq->__lock FD: 232 BD: 1 +.+.: (wq_completion)bond1235#3 ->&rq->__lock ->(work_completion)(&(&slave->notify_work)->work) ->(work_completion)(&(&bond->mcast_work)->work) FD: 232 BD: 1 +.+.: (wq_completion)bond1247#2 ->(work_completion)(&(&slave->notify_work)->work) ->(work_completion)(&(&bond->mcast_work)->work) ->&rq->__lock FD: 232 BD: 1 +.+.: (wq_completion)bond1258 ->&rq->__lock ->(work_completion)(&(&slave->notify_work)->work) ->(work_completion)(&(&bond->mcast_work)->work) FD: 232 BD: 1 +.+.: (wq_completion)bond1236#3 ->&rq->__lock ->(work_completion)(&(&slave->notify_work)->work) ->(work_completion)(&(&bond->mcast_work)->work) FD: 232 BD: 1 +.+.: (wq_completion)bond1237#3 ->(work_completion)(&(&slave->notify_work)->work) ->(work_completion)(&(&bond->mcast_work)->work) ->&rq->__lock FD: 232 BD: 1 +.+.: (wq_completion)bond1248 ->(work_completion)(&(&slave->notify_work)->work) ->(work_completion)(&(&bond->mcast_work)->work) ->&rq->__lock FD: 232 BD: 1 +.+.: (wq_completion)bond1259 ->(work_completion)(&(&slave->notify_work)->work) ->(work_completion)(&(&bond->mcast_work)->work) ->&rq->__lock FD: 232 BD: 1 +.+.: (wq_completion)bond1238#3 ->&rq->__lock ->(work_completion)(&(&slave->notify_work)->work) ->(work_completion)(&(&bond->mcast_work)->work) FD: 232 BD: 1 +.+.: (wq_completion)bond1260 ->&rq->__lock ->(work_completion)(&(&slave->notify_work)->work) ->(work_completion)(&(&bond->mcast_work)->work) FD: 232 BD: 1 +.+.: (wq_completion)bond1249#2 ->&rq->__lock ->(work_completion)(&(&slave->notify_work)->work) ->(work_completion)(&(&bond->mcast_work)->work) FD: 232 BD: 1 +.+.: (wq_completion)bond1261 ->&rq->__lock ->(work_completion)(&(&slave->notify_work)->work) ->(work_completion)(&(&bond->mcast_work)->work) FD: 232 BD: 1 +.+.: (wq_completion)bond1239#3 ->(work_completion)(&(&slave->notify_work)->work) ->(work_completion)(&(&bond->mcast_work)->work) FD: 232 BD: 1 +.+.: (wq_completion)bond1240#2 ->&rq->__lock ->(work_completion)(&(&slave->notify_work)->work) ->(work_completion)(&(&bond->mcast_work)->work) FD: 232 BD: 1 +.+.: (wq_completion)bond1262 ->&rq->__lock ->(work_completion)(&(&slave->notify_work)->work) ->(work_completion)(&(&bond->mcast_work)->work) FD: 232 BD: 1 +.+.: (wq_completion)bond1263 ->&rq->__lock ->(work_completion)(&(&slave->notify_work)->work) ->(work_completion)(&(&bond->mcast_work)->work) FD: 232 BD: 1 +.+.: (wq_completion)bond1264 ->&rq->__lock ->(work_completion)(&(&slave->notify_work)->work) ->(work_completion)(&(&bond->mcast_work)->work) FD: 232 BD: 1 +.+.: (wq_completion)bond1241#3 ->&rq->__lock ->(work_completion)(&(&slave->notify_work)->work) ->(work_completion)(&(&bond->mcast_work)->work) FD: 232 BD: 1 +.+.: (wq_completion)bond1250#2 ->&rq->__lock ->(work_completion)(&(&slave->notify_work)->work) ->(work_completion)(&(&bond->mcast_work)->work) FD: 232 BD: 1 +.+.: (wq_completion)bond1265 ->(work_completion)(&(&slave->notify_work)->work) ->(work_completion)(&(&bond->mcast_work)->work) ->&rq->__lock FD: 232 BD: 1 +.+.: (wq_completion)bond1242#3 ->&rq->__lock ->(work_completion)(&(&slave->notify_work)->work) ->(work_completion)(&(&bond->mcast_work)->work) FD: 232 BD: 1 +.+.: (wq_completion)bond1251#2 ->&rq->__lock ->(work_completion)(&(&slave->notify_work)->work) ->(work_completion)(&(&bond->mcast_work)->work) FD: 232 BD: 1 +.+.: (wq_completion)bond1252#2 ->&rq->__lock ->(work_completion)(&(&slave->notify_work)->work) ->(work_completion)(&(&bond->mcast_work)->work) FD: 232 BD: 1 +.+.: (wq_completion)bond1266 ->&rq->__lock ->(work_completion)(&(&slave->notify_work)->work) ->(work_completion)(&(&bond->mcast_work)->work) FD: 232 BD: 1 +.+.: (wq_completion)bond1243#3 ->&rq->__lock ->(work_completion)(&(&slave->notify_work)->work) ->(work_completion)(&(&bond->mcast_work)->work) FD: 232 BD: 1 +.+.: (wq_completion)bond1253#2 ->&rq->__lock ->(work_completion)(&(&slave->notify_work)->work) ->(work_completion)(&(&bond->mcast_work)->work) FD: 232 BD: 1 +.+.: (wq_completion)bond1267 ->&rq->__lock ->(work_completion)(&(&slave->notify_work)->work) ->(work_completion)(&(&bond->mcast_work)->work) FD: 232 BD: 1 +.+.: (wq_completion)bond1244#3 ->&rq->__lock ->(work_completion)(&(&slave->notify_work)->work) ->(work_completion)(&(&bond->mcast_work)->work) FD: 232 BD: 1 +.+.: (wq_completion)bond1254#2 ->&rq->__lock ->(work_completion)(&(&slave->notify_work)->work) ->(work_completion)(&(&bond->mcast_work)->work) FD: 232 BD: 1 +.+.: (wq_completion)bond1245#3 ->(work_completion)(&(&slave->notify_work)->work) ->(work_completion)(&(&bond->mcast_work)->work) ->&rq->__lock FD: 232 BD: 1 +.+.: (wq_completion)bond1255#2 ->&rq->__lock ->(work_completion)(&(&slave->notify_work)->work) ->(work_completion)(&(&bond->mcast_work)->work) FD: 232 BD: 1 +.+.: (wq_completion)bond1268 ->(work_completion)(&(&slave->notify_work)->work) ->(work_completion)(&(&bond->mcast_work)->work) ->&rq->__lock FD: 232 BD: 1 +.+.: (wq_completion)bond1246#3 ->&rq->__lock ->(work_completion)(&(&slave->notify_work)->work) ->(work_completion)(&(&bond->mcast_work)->work) FD: 232 BD: 1 +.+.: (wq_completion)bond1256#2 ->&rq->__lock ->(work_completion)(&(&slave->notify_work)->work) ->(work_completion)(&(&bond->mcast_work)->work) FD: 232 BD: 1 +.+.: (wq_completion)bond1269 ->&rq->__lock ->(work_completion)(&(&slave->notify_work)->work) ->(work_completion)(&(&bond->mcast_work)->work) FD: 232 BD: 1 +.+.: (wq_completion)bond1247#3 ->&rq->__lock ->(work_completion)(&(&slave->notify_work)->work) ->(work_completion)(&(&bond->mcast_work)->work) FD: 232 BD: 1 +.+.: (wq_completion)bond1257#2 ->&rq->__lock ->(work_completion)(&(&slave->notify_work)->work) ->(work_completion)(&(&bond->mcast_work)->work) FD: 232 BD: 1 +.+.: (wq_completion)bond1270 ->(work_completion)(&(&slave->notify_work)->work) ->(work_completion)(&(&bond->mcast_work)->work) ->&rq->__lock FD: 232 BD: 1 +.+.: (wq_completion)bond1248#2 ->&rq->__lock ->(work_completion)(&(&slave->notify_work)->work) ->(work_completion)(&(&bond->mcast_work)->work) FD: 232 BD: 1 +.+.: (wq_completion)bond1271 ->(work_completion)(&(&slave->notify_work)->work) ->(work_completion)(&(&bond->mcast_work)->work) ->&rq->__lock FD: 232 BD: 1 +.+.: (wq_completion)bond1249#3 ->(work_completion)(&(&slave->notify_work)->work) ->(work_completion)(&(&bond->mcast_work)->work) ->&rq->__lock FD: 232 BD: 1 +.+.: (wq_completion)bond1259#2 ->&rq->__lock ->(work_completion)(&(&slave->notify_work)->work) ->(work_completion)(&(&bond->mcast_work)->work) FD: 232 BD: 1 +.+.: (wq_completion)bond1250#3 ->(work_completion)(&(&slave->notify_work)->work) ->(work_completion)(&(&bond->mcast_work)->work) ->&rq->__lock ->&cfs_rq->removed.lock FD: 232 BD: 1 +.+.: (wq_completion)bond1251#3 ->&rq->__lock ->(work_completion)(&(&slave->notify_work)->work) ->(work_completion)(&(&bond->mcast_work)->work) FD: 232 BD: 1 +.+.: (wq_completion)bond1260#2 ->&rq->__lock ->(work_completion)(&(&slave->notify_work)->work) ->(work_completion)(&(&bond->mcast_work)->work) FD: 232 BD: 1 +.+.: (wq_completion)bond1272 ->&rq->__lock ->(work_completion)(&(&slave->notify_work)->work) ->(work_completion)(&(&bond->mcast_work)->work) FD: 232 BD: 1 +.+.: (wq_completion)bond1261#2 ->&rq->__lock ->(work_completion)(&(&slave->notify_work)->work) ->(work_completion)(&(&bond->mcast_work)->work) FD: 232 BD: 1 +.+.: (wq_completion)bond1273 ->&rq->__lock ->(work_completion)(&(&slave->notify_work)->work) ->(work_completion)(&(&bond->mcast_work)->work) FD: 232 BD: 1 +.+.: (wq_completion)bond1262#2 ->&rq->__lock ->(work_completion)(&(&slave->notify_work)->work) ->(work_completion)(&(&bond->mcast_work)->work) FD: 232 BD: 1 +.+.: (wq_completion)bond1252#3 ->&rq->__lock ->(work_completion)(&(&slave->notify_work)->work) ->(work_completion)(&(&bond->mcast_work)->work) FD: 232 BD: 1 +.+.: (wq_completion)bond1263#2 ->(work_completion)(&(&slave->notify_work)->work) ->(work_completion)(&(&bond->mcast_work)->work) ->&rq->__lock FD: 232 BD: 1 +.+.: (wq_completion)bond1274 ->&rq->__lock ->(work_completion)(&(&slave->notify_work)->work) ->(work_completion)(&(&bond->mcast_work)->work) FD: 232 BD: 1 +.+.: (wq_completion)bond1253#3 ->&rq->__lock ->(work_completion)(&(&slave->notify_work)->work) ->(work_completion)(&(&bond->mcast_work)->work) FD: 232 BD: 1 +.+.: (wq_completion)bond1264#2 ->&rq->__lock ->(work_completion)(&(&slave->notify_work)->work) ->(work_completion)(&(&bond->mcast_work)->work) FD: 232 BD: 1 +.+.: (wq_completion)bond1275 ->(work_completion)(&(&slave->notify_work)->work) ->(work_completion)(&(&bond->mcast_work)->work) ->&rq->__lock FD: 232 BD: 1 +.+.: (wq_completion)bond1254#3 ->&rq->__lock ->(work_completion)(&(&slave->notify_work)->work) ->(work_completion)(&(&bond->mcast_work)->work) FD: 232 BD: 1 +.+.: (wq_completion)bond1265#2 ->&rq->__lock ->(work_completion)(&(&slave->notify_work)->work) ->(work_completion)(&(&bond->mcast_work)->work) FD: 232 BD: 1 +.+.: (wq_completion)bond1276 ->(work_completion)(&(&slave->notify_work)->work) ->(work_completion)(&(&bond->mcast_work)->work) ->&rq->__lock FD: 232 BD: 1 +.+.: (wq_completion)bond1255#3 ->&rq->__lock ->(work_completion)(&(&slave->notify_work)->work) ->(work_completion)(&(&bond->mcast_work)->work) FD: 232 BD: 1 +.+.: (wq_completion)bond1266#2 ->(work_completion)(&(&slave->notify_work)->work) ->(work_completion)(&(&bond->mcast_work)->work) ->&rq->__lock FD: 232 BD: 1 +.+.: (wq_completion)bond1267#2 ->(work_completion)(&(&slave->notify_work)->work) ->(work_completion)(&(&bond->mcast_work)->work) ->&rq->__lock FD: 232 BD: 1 +.+.: (wq_completion)bond1277 ->(work_completion)(&(&slave->notify_work)->work) ->(work_completion)(&(&bond->mcast_work)->work) ->&rq->__lock FD: 232 BD: 1 +.+.: (wq_completion)bond1256#3 ->(work_completion)(&(&slave->notify_work)->work) ->(work_completion)(&(&bond->mcast_work)->work) ->&rq->__lock FD: 232 BD: 1 +.+.: (wq_completion)bond1268#2 ->&rq->__lock ->(work_completion)(&(&slave->notify_work)->work) ->(work_completion)(&(&bond->mcast_work)->work) FD: 232 BD: 1 +.+.: (wq_completion)bond1278 ->&rq->__lock ->(work_completion)(&(&slave->notify_work)->work) ->(work_completion)(&(&bond->mcast_work)->work) FD: 232 BD: 1 +.+.: (wq_completion)bond1257#3 ->&rq->__lock ->(work_completion)(&(&slave->notify_work)->work) ->(work_completion)(&(&bond->mcast_work)->work) ->&cfs_rq->removed.lock ->&obj_hash[i].lock FD: 232 BD: 1 +.+.: (wq_completion)bond1269#2 ->(work_completion)(&(&slave->notify_work)->work) ->(work_completion)(&(&bond->mcast_work)->work) ->&rq->__lock FD: 232 BD: 1 +.+.: (wq_completion)bond1279 ->&rq->__lock ->(work_completion)(&(&slave->notify_work)->work) ->(work_completion)(&(&bond->mcast_work)->work) FD: 232 BD: 1 +.+.: (wq_completion)bond1258#2 ->&rq->__lock ->(work_completion)(&(&slave->notify_work)->work) ->(work_completion)(&(&bond->mcast_work)->work) FD: 232 BD: 1 +.+.: (wq_completion)bond1280 ->&rq->__lock ->(work_completion)(&(&slave->notify_work)->work) ->(work_completion)(&(&bond->mcast_work)->work) FD: 232 BD: 1 +.+.: (wq_completion)bond1259#3 ->&rq->__lock ->(work_completion)(&(&slave->notify_work)->work) ->(work_completion)(&(&bond->mcast_work)->work) FD: 232 BD: 1 +.+.: (wq_completion)bond1270#2 ->&rq->__lock ->(work_completion)(&(&slave->notify_work)->work) ->(work_completion)(&(&bond->mcast_work)->work) FD: 232 BD: 1 +.+.: (wq_completion)bond1281 ->&rq->__lock ->(work_completion)(&(&slave->notify_work)->work) ->(work_completion)(&(&bond->mcast_work)->work) FD: 232 BD: 1 +.+.: (wq_completion)bond1271#2 ->&rq->__lock ->(work_completion)(&(&slave->notify_work)->work) ->(work_completion)(&(&bond->mcast_work)->work) FD: 232 BD: 1 +.+.: (wq_completion)bond1272#2 ->&rq->__lock ->(work_completion)(&(&slave->notify_work)->work) ->(work_completion)(&(&bond->mcast_work)->work) FD: 232 BD: 1 +.+.: (wq_completion)bond1260#3 ->&rq->__lock ->(work_completion)(&(&slave->notify_work)->work) ->(work_completion)(&(&bond->mcast_work)->work) FD: 232 BD: 1 +.+.: (wq_completion)bond1273#2 ->&rq->__lock ->(work_completion)(&(&slave->notify_work)->work) ->(work_completion)(&(&bond->mcast_work)->work) FD: 232 BD: 1 +.+.: (wq_completion)bond1261#3 ->&rq->__lock ->(work_completion)(&(&slave->notify_work)->work) ->(work_completion)(&(&bond->mcast_work)->work) FD: 232 BD: 1 +.+.: (wq_completion)bond1262#3 ->(work_completion)(&(&slave->notify_work)->work) ->(work_completion)(&(&bond->mcast_work)->work) ->&rq->__lock FD: 232 BD: 1 +.+.: (wq_completion)bond1263#3 ->&rq->__lock ->(work_completion)(&(&slave->notify_work)->work) ->(work_completion)(&(&bond->mcast_work)->work) FD: 232 BD: 1 +.+.: (wq_completion)bond1282 ->&rq->__lock ->(work_completion)(&(&slave->notify_work)->work) ->(work_completion)(&(&bond->mcast_work)->work) FD: 232 BD: 1 +.+.: (wq_completion)bond1264#3 ->(work_completion)(&(&slave->notify_work)->work) ->(work_completion)(&(&bond->mcast_work)->work) FD: 232 BD: 1 +.+.: (wq_completion)bond1283 ->&rq->__lock ->(work_completion)(&(&slave->notify_work)->work) ->(work_completion)(&(&bond->mcast_work)->work) FD: 232 BD: 1 +.+.: (wq_completion)bond1274#2 ->&rq->__lock ->(work_completion)(&(&slave->notify_work)->work) ->(work_completion)(&(&bond->mcast_work)->work) FD: 232 BD: 1 +.+.: (wq_completion)bond1265#3 ->&rq->__lock ->(work_completion)(&(&slave->notify_work)->work) ->(work_completion)(&(&bond->mcast_work)->work) FD: 232 BD: 1 +.+.: (wq_completion)bond1284 ->&rq->__lock ->(work_completion)(&(&slave->notify_work)->work) ->(work_completion)(&(&bond->mcast_work)->work) FD: 232 BD: 1 +.+.: (wq_completion)bond1275#2 ->&rq->__lock ->(work_completion)(&(&slave->notify_work)->work) ->(work_completion)(&(&bond->mcast_work)->work) FD: 232 BD: 1 +.+.: (wq_completion)bond1266#3 ->(work_completion)(&(&slave->notify_work)->work) ->(work_completion)(&(&bond->mcast_work)->work) ->&rq->__lock FD: 232 BD: 1 +.+.: (wq_completion)bond1285 ->(work_completion)(&(&slave->notify_work)->work) ->(work_completion)(&(&bond->mcast_work)->work) ->&rq->__lock FD: 232 BD: 1 +.+.: (wq_completion)bond1276#2 ->&rq->__lock ->(work_completion)(&(&slave->notify_work)->work) ->(work_completion)(&(&bond->mcast_work)->work) FD: 232 BD: 1 +.+.: (wq_completion)bond1267#3 ->(work_completion)(&(&slave->notify_work)->work) ->(work_completion)(&(&bond->mcast_work)->work) ->&rq->__lock FD: 232 BD: 1 +.+.: (wq_completion)bond1286 ->(work_completion)(&(&slave->notify_work)->work) ->(work_completion)(&(&bond->mcast_work)->work) FD: 232 BD: 1 +.+.: (wq_completion)bond1277#2 ->(work_completion)(&(&slave->notify_work)->work) ->(work_completion)(&(&bond->mcast_work)->work) ->&rq->__lock FD: 232 BD: 1 +.+.: (wq_completion)bond1268#3 ->&rq->__lock ->(work_completion)(&(&slave->notify_work)->work) ->(work_completion)(&(&bond->mcast_work)->work) FD: 232 BD: 1 +.+.: (wq_completion)bond1278#2 ->&rq->__lock ->(work_completion)(&(&slave->notify_work)->work) ->(work_completion)(&(&bond->mcast_work)->work) FD: 232 BD: 1 +.+.: (wq_completion)bond1269#3 ->(work_completion)(&(&slave->notify_work)->work) ->(work_completion)(&(&bond->mcast_work)->work) ->&rq->__lock FD: 232 BD: 1 +.+.: (wq_completion)bond1279#2 ->&rq->__lock ->(work_completion)(&(&slave->notify_work)->work) ->(work_completion)(&(&bond->mcast_work)->work) FD: 232 BD: 1 +.+.: (wq_completion)bond1287 ->&rq->__lock ->(work_completion)(&(&slave->notify_work)->work) ->(work_completion)(&(&bond->mcast_work)->work) FD: 232 BD: 1 +.+.: (wq_completion)bond1270#3 ->(work_completion)(&(&slave->notify_work)->work) ->(work_completion)(&(&bond->mcast_work)->work) ->&rq->__lock FD: 232 BD: 1 +.+.: (wq_completion)bond1280#2 ->(work_completion)(&(&slave->notify_work)->work) ->(work_completion)(&(&bond->mcast_work)->work) ->&rq->__lock FD: 232 BD: 1 +.+.: (wq_completion)bond1288 ->(work_completion)(&(&slave->notify_work)->work) ->(work_completion)(&(&bond->mcast_work)->work) ->&rq->__lock FD: 232 BD: 1 +.+.: (wq_completion)bond1271#3 ->(work_completion)(&(&slave->notify_work)->work) ->(work_completion)(&(&bond->mcast_work)->work) FD: 232 BD: 1 +.+.: (wq_completion)bond1281#2 ->(work_completion)(&(&slave->notify_work)->work) ->(work_completion)(&(&bond->mcast_work)->work) ->&rq->__lock FD: 232 BD: 1 +.+.: (wq_completion)bond1289 ->&rq->__lock ->(work_completion)(&(&slave->notify_work)->work) ->(work_completion)(&(&bond->mcast_work)->work) FD: 232 BD: 1 +.+.: (wq_completion)bond1272#3 ->(work_completion)(&(&slave->notify_work)->work) ->(work_completion)(&(&bond->mcast_work)->work) FD: 232 BD: 1 +.+.: (wq_completion)bond1282#2 ->(work_completion)(&(&slave->notify_work)->work) ->(work_completion)(&(&bond->mcast_work)->work) ->&rq->__lock FD: 232 BD: 1 +.+.: (wq_completion)bond1290 ->&rq->__lock ->(work_completion)(&(&slave->notify_work)->work) ->(work_completion)(&(&bond->mcast_work)->work) FD: 232 BD: 1 +.+.: (wq_completion)bond1273#3 ->&rq->__lock ->(work_completion)(&(&slave->notify_work)->work) ->(work_completion)(&(&bond->mcast_work)->work) FD: 232 BD: 1 +.+.: (wq_completion)bond1291 ->&rq->__lock ->(work_completion)(&(&slave->notify_work)->work) ->(work_completion)(&(&bond->mcast_work)->work) FD: 232 BD: 1 +.+.: (wq_completion)bond1274#3 ->&rq->__lock ->(work_completion)(&(&slave->notify_work)->work) ->(work_completion)(&(&bond->mcast_work)->work) FD: 232 BD: 1 +.+.: (wq_completion)bond1292 ->(work_completion)(&(&slave->notify_work)->work) ->(work_completion)(&(&bond->mcast_work)->work) ->&rq->__lock FD: 232 BD: 1 +.+.: (wq_completion)bond1283#2 ->&rq->__lock ->(work_completion)(&(&slave->notify_work)->work) ->(work_completion)(&(&bond->mcast_work)->work) FD: 232 BD: 1 +.+.: (wq_completion)bond1275#3 ->&rq->__lock ->(work_completion)(&(&slave->notify_work)->work) ->(work_completion)(&(&bond->mcast_work)->work) FD: 232 BD: 1 +.+.: (wq_completion)bond1284#2 ->(work_completion)(&(&slave->notify_work)->work) ->(work_completion)(&(&bond->mcast_work)->work) FD: 232 BD: 1 +.+.: (wq_completion)bond1276#3 ->&rq->__lock ->(work_completion)(&(&slave->notify_work)->work) ->(work_completion)(&(&bond->mcast_work)->work) FD: 232 BD: 1 +.+.: (wq_completion)bond1293 ->&rq->__lock ->(work_completion)(&(&slave->notify_work)->work) ->(work_completion)(&(&bond->mcast_work)->work) FD: 232 BD: 1 +.+.: (wq_completion)bond1285#2 ->&rq->__lock ->(work_completion)(&(&slave->notify_work)->work) ->(work_completion)(&(&bond->mcast_work)->work) FD: 232 BD: 1 +.+.: (wq_completion)bond1277#3 ->&rq->__lock ->(work_completion)(&(&slave->notify_work)->work) ->(work_completion)(&(&bond->mcast_work)->work) FD: 232 BD: 1 +.+.: (wq_completion)bond1294 ->&rq->__lock ->(work_completion)(&(&slave->notify_work)->work) ->(work_completion)(&(&bond->mcast_work)->work) FD: 232 BD: 1 +.+.: (wq_completion)bond1286#2 ->&rq->__lock ->(work_completion)(&(&slave->notify_work)->work) ->(work_completion)(&(&bond->mcast_work)->work) FD: 232 BD: 1 +.+.: (wq_completion)bond1278#3 ->&rq->__lock ->(work_completion)(&(&slave->notify_work)->work) ->(work_completion)(&(&bond->mcast_work)->work) FD: 232 BD: 1 +.+.: (wq_completion)bond1295 ->&rq->__lock ->(work_completion)(&(&slave->notify_work)->work) ->(work_completion)(&(&bond->mcast_work)->work) FD: 232 BD: 1 +.+.: (wq_completion)bond1287#2 ->(work_completion)(&(&slave->notify_work)->work) ->(work_completion)(&(&bond->mcast_work)->work) ->&rq->__lock FD: 232 BD: 1 +.+.: (wq_completion)bond1279#3 ->&rq->__lock ->(work_completion)(&(&slave->notify_work)->work) ->(work_completion)(&(&bond->mcast_work)->work) FD: 232 BD: 1 +.+.: (wq_completion)bond1296 ->&rq->__lock ->(work_completion)(&(&slave->notify_work)->work) ->(work_completion)(&(&bond->mcast_work)->work) FD: 232 BD: 1 +.+.: (wq_completion)bond1288#2 ->&rq->__lock ->(work_completion)(&(&slave->notify_work)->work) ->(work_completion)(&(&bond->mcast_work)->work) FD: 232 BD: 1 +.+.: (wq_completion)bond1280#3 ->&rq->__lock ->(work_completion)(&(&slave->notify_work)->work) ->(work_completion)(&(&bond->mcast_work)->work) FD: 232 BD: 1 +.+.: (wq_completion)bond1297 ->&rq->__lock ->(work_completion)(&(&slave->notify_work)->work) ->(work_completion)(&(&bond->mcast_work)->work) FD: 232 BD: 1 +.+.: (wq_completion)bond1281#3 ->&rq->__lock ->(work_completion)(&(&slave->notify_work)->work) ->(work_completion)(&(&bond->mcast_work)->work) FD: 232 BD: 1 +.+.: (wq_completion)bond1298 ->&rq->__lock ->(work_completion)(&(&slave->notify_work)->work) ->(work_completion)(&(&bond->mcast_work)->work) FD: 232 BD: 1 +.+.: (wq_completion)bond1289#2 ->&rq->__lock ->(work_completion)(&(&slave->notify_work)->work) ->(work_completion)(&(&bond->mcast_work)->work) FD: 232 BD: 1 +.+.: (wq_completion)bond1290#2 ->(work_completion)(&(&slave->notify_work)->work) ->(work_completion)(&(&bond->mcast_work)->work) ->&rq->__lock FD: 232 BD: 1 +.+.: (wq_completion)bond1282#3 ->&rq->__lock ->(work_completion)(&(&slave->notify_work)->work) ->(work_completion)(&(&bond->mcast_work)->work) FD: 232 BD: 1 +.+.: (wq_completion)bond1299 ->&rq->__lock ->(work_completion)(&(&slave->notify_work)->work) ->(work_completion)(&(&bond->mcast_work)->work) FD: 232 BD: 1 +.+.: (wq_completion)bond1291#2 ->&rq->__lock ->(work_completion)(&(&slave->notify_work)->work) ->(work_completion)(&(&bond->mcast_work)->work) FD: 232 BD: 1 +.+.: (wq_completion)bond1283#3 ->(work_completion)(&(&slave->notify_work)->work) ->(work_completion)(&(&bond->mcast_work)->work) ->&rq->__lock FD: 232 BD: 1 +.+.: (wq_completion)bond1300 ->&rq->__lock ->(work_completion)(&(&slave->notify_work)->work) ->(work_completion)(&(&bond->mcast_work)->work) FD: 232 BD: 1 +.+.: (wq_completion)bond1292#2 ->&rq->__lock ->(work_completion)(&(&slave->notify_work)->work) ->(work_completion)(&(&bond->mcast_work)->work) FD: 232 BD: 1 +.+.: (wq_completion)bond1284#3 ->&rq->__lock ->(work_completion)(&(&slave->notify_work)->work) ->(work_completion)(&(&bond->mcast_work)->work) FD: 232 BD: 1 +.+.: (wq_completion)bond1301 ->(work_completion)(&(&slave->notify_work)->work) ->(work_completion)(&(&bond->mcast_work)->work) FD: 232 BD: 1 +.+.: (wq_completion)bond1293#2 ->&rq->__lock ->(work_completion)(&(&slave->notify_work)->work) ->(work_completion)(&(&bond->mcast_work)->work) FD: 232 BD: 1 +.+.: (wq_completion)bond1285#3 ->(work_completion)(&(&slave->notify_work)->work) ->(work_completion)(&(&bond->mcast_work)->work) ->&rq->__lock FD: 232 BD: 1 +.+.: (wq_completion)bond1302 ->(work_completion)(&(&slave->notify_work)->work) ->(work_completion)(&(&bond->mcast_work)->work) FD: 232 BD: 1 +.+.: (wq_completion)bond1294#2 ->&rq->__lock ->(work_completion)(&(&slave->notify_work)->work) ->(work_completion)(&(&bond->mcast_work)->work) FD: 232 BD: 1 +.+.: (wq_completion)bond1286#3 ->(work_completion)(&(&slave->notify_work)->work) ->(work_completion)(&(&bond->mcast_work)->work) ->&rq->__lock FD: 232 BD: 1 +.+.: (wq_completion)bond1303 ->&rq->__lock ->(work_completion)(&(&slave->notify_work)->work) ->(work_completion)(&(&bond->mcast_work)->work) FD: 232 BD: 1 +.+.: (wq_completion)bond1295#2 ->&rq->__lock ->(work_completion)(&(&slave->notify_work)->work) ->(work_completion)(&(&bond->mcast_work)->work) FD: 232 BD: 1 +.+.: (wq_completion)bond1296#2 ->&rq->__lock ->(work_completion)(&(&slave->notify_work)->work) ->(work_completion)(&(&bond->mcast_work)->work) FD: 232 BD: 1 +.+.: (wq_completion)bond1287#3 ->(work_completion)(&(&slave->notify_work)->work) ->(work_completion)(&(&bond->mcast_work)->work) ->&rq->__lock FD: 232 BD: 1 +.+.: (wq_completion)bond1304 ->(work_completion)(&(&slave->notify_work)->work) ->(work_completion)(&(&bond->mcast_work)->work) FD: 232 BD: 1 +.+.: (wq_completion)bond1297#2 ->&rq->__lock ->(work_completion)(&(&slave->notify_work)->work) ->(work_completion)(&(&bond->mcast_work)->work) FD: 232 BD: 1 +.+.: (wq_completion)bond1288#3 ->&rq->__lock ->(work_completion)(&(&slave->notify_work)->work) ->(work_completion)(&(&bond->mcast_work)->work) FD: 232 BD: 1 +.+.: (wq_completion)bond1305 ->&rq->__lock ->(work_completion)(&(&slave->notify_work)->work) ->(work_completion)(&(&bond->mcast_work)->work) FD: 232 BD: 1 +.+.: (wq_completion)bond10#2 ->&rq->__lock ->(work_completion)(&(&slave->notify_work)->work) ->(work_completion)(&(&bond->mcast_work)->work) FD: 232 BD: 1 +.+.: (wq_completion)bond1289#3 ->&rq->__lock ->(work_completion)(&(&slave->notify_work)->work) ->(work_completion)(&(&bond->mcast_work)->work) FD: 232 BD: 1 +.+.: (wq_completion)bond1290#3 ->(work_completion)(&(&slave->notify_work)->work) ->(work_completion)(&(&bond->mcast_work)->work) FD: 232 BD: 1 +.+.: (wq_completion)bond11#2 ->&rq->__lock ->(work_completion)(&(&slave->notify_work)->work) ->(work_completion)(&(&bond->mcast_work)->work) FD: 232 BD: 1 +.+.: (wq_completion)bond1306 ->(work_completion)(&(&slave->notify_work)->work) ->(work_completion)(&(&bond->mcast_work)->work) FD: 232 BD: 1 +.+.: (wq_completion)bond1298#2 ->(work_completion)(&(&slave->notify_work)->work) ->(work_completion)(&(&bond->mcast_work)->work) ->&rq->__lock FD: 232 BD: 1 +.+.: (wq_completion)bond1291#3 ->(work_completion)(&(&slave->notify_work)->work) ->(work_completion)(&(&bond->mcast_work)->work) ->&rq->__lock FD: 232 BD: 1 +.+.: (wq_completion)bond1307 ->&rq->__lock ->(work_completion)(&(&slave->notify_work)->work) ->(work_completion)(&(&bond->mcast_work)->work) FD: 232 BD: 1 +.+.: (wq_completion)bond1299#2 ->&rq->__lock ->(work_completion)(&(&slave->notify_work)->work) ->(work_completion)(&(&bond->mcast_work)->work) FD: 232 BD: 1 +.+.: (wq_completion)bond1308 ->&rq->__lock ->(work_completion)(&(&slave->notify_work)->work) ->(work_completion)(&(&bond->mcast_work)->work) FD: 232 BD: 1 +.+.: (wq_completion)bond1300#2 ->(work_completion)(&(&slave->notify_work)->work) ->(work_completion)(&(&bond->mcast_work)->work) ->&rq->__lock FD: 232 BD: 1 +.+.: (wq_completion)bond12 ->&rq->__lock ->(work_completion)(&(&slave->notify_work)->work) ->(work_completion)(&(&bond->mcast_work)->work) FD: 232 BD: 1 +.+.: (wq_completion)bond1292#3 ->&rq->__lock ->(work_completion)(&(&slave->notify_work)->work) ->(work_completion)(&(&bond->mcast_work)->work) FD: 232 BD: 1 +.+.: (wq_completion)bond1309 ->&rq->__lock ->(work_completion)(&(&slave->notify_work)->work) ->(work_completion)(&(&bond->mcast_work)->work) FD: 232 BD: 1 +.+.: (wq_completion)bond1301#2 ->&rq->__lock ->(work_completion)(&(&slave->notify_work)->work) ->(work_completion)(&(&bond->mcast_work)->work) FD: 232 BD: 1 +.+.: (wq_completion)bond1302#2 ->(work_completion)(&(&slave->notify_work)->work) ->(work_completion)(&(&bond->mcast_work)->work) ->&rq->__lock FD: 232 BD: 1 +.+.: (wq_completion)bond13#3 ->&rq->__lock ->(work_completion)(&(&slave->notify_work)->work) ->(work_completion)(&(&bond->mcast_work)->work) FD: 232 BD: 1 +.+.: (wq_completion)bond14#3 ->&rq->__lock ->(work_completion)(&(&slave->notify_work)->work) ->(work_completion)(&(&bond->mcast_work)->work) FD: 232 BD: 1 +.+.: (wq_completion)bond1293#3 ->(work_completion)(&(&slave->notify_work)->work) ->(work_completion)(&(&bond->mcast_work)->work) FD: 232 BD: 1 +.+.: (wq_completion)bond1310 ->&rq->__lock ->(work_completion)(&(&slave->notify_work)->work) ->(work_completion)(&(&bond->mcast_work)->work) FD: 232 BD: 1 +.+.: (wq_completion)bond1294#3 ->&rq->__lock ->(work_completion)(&(&slave->notify_work)->work) ->(work_completion)(&(&bond->mcast_work)->work) FD: 232 BD: 1 +.+.: (wq_completion)bond1311 ->&rq->__lock ->(work_completion)(&(&slave->notify_work)->work) ->(work_completion)(&(&bond->mcast_work)->work) FD: 232 BD: 1 +.+.: (wq_completion)bond15#4 ->&rq->__lock ->(work_completion)(&(&slave->notify_work)->work) ->(work_completion)(&(&bond->mcast_work)->work) FD: 232 BD: 1 +.+.: (wq_completion)bond1303#2 ->(work_completion)(&(&slave->notify_work)->work) ->(work_completion)(&(&bond->mcast_work)->work) FD: 232 BD: 1 +.+.: (wq_completion)bond1295#3 ->&rq->__lock ->(work_completion)(&(&slave->notify_work)->work) ->(work_completion)(&(&bond->mcast_work)->work) FD: 232 BD: 1 +.+.: (wq_completion)bond1296#3 ->&rq->__lock ->(work_completion)(&(&slave->notify_work)->work) ->(work_completion)(&(&bond->mcast_work)->work) FD: 232 BD: 1 +.+.: (wq_completion)bond1312 ->(work_completion)(&(&slave->notify_work)->work) ->(work_completion)(&(&bond->mcast_work)->work) FD: 160 BD: 1 +.+.: (wq_completion)bond1304#2 ->&rq->__lock ->(work_completion)(&(&slave->notify_work)->work) all lock chains: irq_context: 0 &obj_hash[i].lock irq_context: 0 &obj_hash[i].lock pool_lock irq_context: 0 cgroup_mutex irq_context: 0 (console_sem).lock irq_context: 0 cpu_hotplug_lock irq_context: 0 cpu_hotplug_lock jump_label_mutex irq_context: 0 cpu_hotplug_lock static_call_mutex irq_context: 0 cpu_hotplug_lock static_call_mutex text_mutex irq_context: 0 console_mutex irq_context: 0 console_mutex syslog_lock irq_context: 0 console_mutex (console_sem).lock irq_context: 0 console_mutex console_lock console_srcu console_owner_lock irq_context: 0 console_mutex console_lock console_srcu console_owner irq_context: 0 console_mutex console_lock console_srcu console_owner console_owner_lock irq_context: 0 console_lock console_srcu console_owner_lock irq_context: 0 console_lock console_srcu console_owner irq_context: 0 console_lock console_srcu console_owner console_owner_lock irq_context: 0 input_pool.lock irq_context: 0 cpu_hotplug_lock cpuhp_state_mutex irq_context: 0 clocksource_mutex irq_context: 0 clocksource_mutex watchdog_lock irq_context: 0 cpu_hotplug_lock jump_label_mutex text_mutex irq_context: 0 resource_lock irq_context: 0 cache_disable_lock irq_context: 0 pgd_lock irq_context: 0 init_mm.page_table_lock irq_context: 0 init_mm.page_table_lock pgd_lock irq_context: 0 early_pfn_lock irq_context: 0 acpi_ioapic_lock irq_context: 0 acpi_ioapic_lock ioapic_lock irq_context: 0 acpi_ioapic_lock (console_sem).lock irq_context: 0 acpi_ioapic_lock console_lock console_srcu console_owner_lock irq_context: 0 acpi_ioapic_lock console_lock console_srcu console_owner irq_context: 0 acpi_ioapic_lock console_lock console_srcu console_owner console_owner_lock irq_context: 0 syscore_ops_lock irq_context: 0 map_entries_lock irq_context: 0 devtree_lock irq_context: 0 pcpu_lock irq_context: 0 param_lock irq_context: 0 base_crng.lock irq_context: 0 crng_init_wait.lock irq_context: 0 zonelist_update_seq irq_context: 0 zonelist_update_seq zonelist_update_seq.seqcount irq_context: 0 dmar_global_lock irq_context: 0 &zone->lock irq_context: 0 &zone->lock &____s->seqcount irq_context: 0 &pcp->lock &zone->lock irq_context: 0 &pcp->lock &zone->lock &____s->seqcount irq_context: 0 &____s->seqcount irq_context: 0 pool_lock#2 irq_context: 0 pcpu_alloc_mutex irq_context: 0 pcpu_alloc_mutex pcpu_lock irq_context: 0 &n->list_lock irq_context: 0 &c->lock irq_context: 0 slab_mutex irq_context: 0 slab_mutex pool_lock#2 irq_context: 0 slab_mutex &c->lock irq_context: 0 slab_mutex &n->list_lock irq_context: 0 slab_mutex pcpu_alloc_mutex irq_context: 0 slab_mutex pcpu_alloc_mutex pcpu_lock irq_context: 0 batched_entropy_u64.lock irq_context: 0 batched_entropy_u64.lock crngs.lock irq_context: 0 batched_entropy_u64.lock crngs.lock base_crng.lock irq_context: 0 espfix_init_mutex irq_context: 0 espfix_init_mutex &pcp->lock &zone->lock irq_context: 0 espfix_init_mutex &zone->lock irq_context: 0 espfix_init_mutex &pcp->lock &zone->lock &____s->seqcount irq_context: 0 espfix_init_mutex &____s->seqcount irq_context: 0 espfix_init_mutex pool_lock#2 irq_context: 0 percpu_counters_lock irq_context: 0 &mm->page_table_lock irq_context: 0 ptlock_ptr(page) irq_context: 0 ptlock_ptr(page)#2 irq_context: 0 trace_types_lock irq_context: 0 panic_notifier_list.lock irq_context: 0 die_chain.lock irq_context: 0 trace_event_sem irq_context: 0 batched_entropy_u32.lock irq_context: 0 batched_entropy_u32.lock crngs.lock irq_context: 0 &rq->__lock irq_context: 0 &rq->__lock rcu_read_lock &cfs_b->lock irq_context: 0 init_task.pi_lock irq_context: 0 init_task.pi_lock &rq->__lock irq_context: 0 init_task.vtime_seqcount irq_context: 0 slab_mutex &pcp->lock &zone->lock irq_context: 0 slab_mutex &zone->lock irq_context: 0 slab_mutex &____s->seqcount irq_context: 0 wq_pool_mutex irq_context: 0 wq_pool_mutex &pcp->lock &zone->lock irq_context: 0 wq_pool_mutex &zone->lock irq_context: 0 wq_pool_mutex &pcp->lock &zone->lock &____s->seqcount irq_context: 0 wq_pool_mutex &____s->seqcount irq_context: 0 wq_pool_mutex pool_lock#2 irq_context: 0 wq_pool_mutex &c->lock irq_context: 0 &wq->mutex irq_context: 0 &wq->mutex &pool->lock irq_context: 0 wq_pool_mutex &wq->mutex irq_context: 0 cpu_hotplug_lock wq_pool_mutex irq_context: 0 cpu_hotplug_lock wq_pool_mutex pool_lock#2 irq_context: 0 cpu_hotplug_lock wq_pool_mutex &obj_hash[i].lock irq_context: 0 cpu_hotplug_lock wq_pool_mutex &pcp->lock &zone->lock irq_context: 0 cpu_hotplug_lock wq_pool_mutex &zone->lock irq_context: 0 cpu_hotplug_lock wq_pool_mutex &pcp->lock &zone->lock &____s->seqcount irq_context: 0 cpu_hotplug_lock wq_pool_mutex &____s->seqcount irq_context: 0 cpu_hotplug_lock wq_pool_mutex &c->lock irq_context: 0 cpu_hotplug_lock wq_pool_mutex &wq->mutex irq_context: 0 cpu_hotplug_lock wq_pool_mutex &wq->mutex &pool->lock/1 irq_context: 0 wq_pool_mutex &wq->mutex &pool->lock irq_context: 0 shrinker_rwsem irq_context: 0 rcu_node_0 irq_context: 0 rcu_state.barrier_lock irq_context: 0 rcu_state.barrier_lock rcu_node_0 irq_context: 0 &rnp->exp_poll_lock irq_context: 0 &rnp->exp_poll_lock rcu_read_lock &pool->lock irq_context: 0 &rnp->exp_poll_lock rcu_read_lock &pool->lock &obj_hash[i].lock irq_context: 0 slab_mutex &pcp->lock &zone->lock &____s->seqcount irq_context: 0 trace_event_sem trace_event_ida.xa_lock irq_context: 0 trace_event_sem trace_event_ida.xa_lock &pcp->lock &zone->lock irq_context: 0 trace_event_sem trace_event_ida.xa_lock &zone->lock irq_context: 0 trace_event_sem trace_event_ida.xa_lock &____s->seqcount irq_context: 0 trace_event_sem trace_event_ida.xa_lock pool_lock#2 irq_context: 0 trace_event_sem trace_event_ida.xa_lock &c->lock irq_context: 0 trigger_cmd_mutex irq_context: 0 i8259A_lock irq_context: 0 irq_domain_mutex irq_context: 0 free_vmap_area_lock irq_context: 0 vmap_area_lock irq_context: 0 &irq_desc_lock_class irq_context: 0 vmap_purge_lock irq_context: 0 vmap_purge_lock purge_vmap_area_lock irq_context: 0 cpa_lock irq_context: 0 cpa_lock pgd_lock irq_context: 0 timekeeper_lock irq_context: 0 timekeeper_lock tk_core.seq.seqcount irq_context: 0 timekeeper_lock tk_core.seq.seqcount &obj_hash[i].lock irq_context: 0 tk_core.seq.seqcount irq_context: 0 &base->lock irq_context: 0 &base->lock &obj_hash[i].lock irq_context: 0 batched_entropy_u32.lock crngs.lock base_crng.lock irq_context: 0 rcu_read_lock &pool->lock/1 irq_context: 0 rcu_read_lock &pool->lock/1 &obj_hash[i].lock irq_context: 0 pmus_lock irq_context: 0 pmus_lock pcpu_alloc_mutex irq_context: 0 pmus_lock pcpu_alloc_mutex pcpu_lock irq_context: 0 pmus_lock pool_lock#2 irq_context: 0 pmus_lock &obj_hash[i].lock irq_context: 0 &swhash->hlist_mutex irq_context: 0 pmus_lock &cpuctx_mutex irq_context: 0 tty_ldiscs_lock irq_context: 0 console_lock irq_context: 0 console_lock resource_lock irq_context: 0 console_lock pool_lock#2 irq_context: 0 console_lock &obj_hash[i].lock irq_context: 0 console_lock &pcp->lock &zone->lock irq_context: 0 console_lock &zone->lock irq_context: 0 console_lock &____s->seqcount irq_context: 0 console_lock &c->lock irq_context: 0 console_lock kbd_event_lock irq_context: 0 console_lock kbd_event_lock led_lock irq_context: 0 console_lock vga_lock irq_context: 0 console_lock (console_sem).lock irq_context: 0 console_lock console_owner_lock irq_context: 0 console_mutex &port_lock_key irq_context: 0 console_mutex console_lock irq_context: 0 console_mutex console_lock console_srcu console_owner &port_lock_key irq_context: 0 console_mutex console_srcu_srcu_usage.lock irq_context: 0 console_mutex console_srcu_srcu_usage.lock &obj_hash[i].lock irq_context: 0 console_mutex &ACCESS_PRIVATE(sdp, lock) irq_context: 0 console_mutex console_srcu irq_context: 0 console_lock console_srcu console_owner &port_lock_key irq_context: 0 init_task.alloc_lock irq_context: 0 acpi_ioremap_lock irq_context: 0 acpi_ioremap_lock pool_lock#2 irq_context: 0 acpi_ioremap_lock resource_lock irq_context: 0 acpi_ioremap_lock memtype_lock irq_context: 0 acpi_ioremap_lock free_vmap_area_lock irq_context: 0 acpi_ioremap_lock vmap_area_lock irq_context: 0 semaphore->lock irq_context: 0 *(&acpi_gbl_reference_count_lock) irq_context: 0 clockevents_lock irq_context: 0 clockevents_lock tk_core.seq.seqcount irq_context: 0 clockevents_lock tick_broadcast_lock irq_context: 0 clockevents_lock i8253_lock irq_context: 0 &desc->request_mutex irq_context: 0 &desc->request_mutex &irq_desc_lock_class irq_context: 0 &desc->request_mutex &irq_desc_lock_class i8259A_lock irq_context: 0 ioapic_lock irq_context: 0 ioapic_mutex irq_context: 0 ioapic_mutex &domain->mutex irq_context: 0 ioapic_mutex &domain->mutex pool_lock#2 irq_context: 0 ioapic_mutex &domain->mutex vector_lock irq_context: 0 ioapic_mutex &domain->mutex &irq_desc_lock_class irq_context: 0 ioapic_mutex &domain->mutex i8259A_lock irq_context: 0 ioapic_mutex &domain->mutex &c->lock irq_context: 0 ioapic_mutex &domain->mutex &pcp->lock &zone->lock irq_context: 0 ioapic_mutex &domain->mutex &zone->lock irq_context: 0 ioapic_mutex &domain->mutex &____s->seqcount irq_context: 0 vector_lock irq_context: 0 &pool->lock#2 irq_context: hardirq jiffies_lock irq_context: hardirq jiffies_lock jiffies_seq.seqcount irq_context: hardirq hrtimer_bases.lock irq_context: hardirq hrtimer_bases.lock tk_core.seq.seqcount irq_context: hardirq log_wait.lock irq_context: 0 spec_ctrl_mutex irq_context: 0 spec_ctrl_mutex cpu_hotplug_lock irq_context: softirq drivers/char/random.c:1010 irq_context: softirq drivers/char/random.c:1010 input_pool.lock irq_context: 0 rcu_read_lock pool_lock#2 irq_context: 0 slab_mutex rcu_read_lock pool_lock#2 irq_context: 0 slab_mutex &obj_hash[i].lock irq_context: 0 sysctl_lock irq_context: 0 tomoyo_policy_lock irq_context: 0 tomoyo_policy_lock pool_lock#2 irq_context: 0 aa_secids.xa_lock irq_context: 0 aa_secids.xa_lock pool_lock#2 irq_context: 0 aa_buffers_lock irq_context: 0 pernet_ops_rwsem irq_context: 0 pernet_ops_rwsem stack_depot_init_mutex irq_context: 0 pernet_ops_rwsem crngs.lock irq_context: 0 pernet_ops_rwsem net_rwsem irq_context: 0 pernet_ops_rwsem proc_inum_ida.xa_lock irq_context: 0 rtnl_mutex irq_context: 0 rtnl_mutex &c->lock irq_context: 0 rtnl_mutex &pcp->lock &zone->lock irq_context: 0 rtnl_mutex &zone->lock irq_context: 0 rtnl_mutex &____s->seqcount irq_context: 0 rtnl_mutex pool_lock#2 irq_context: 0 lock irq_context: 0 lock kernfs_idr_lock irq_context: 0 lock kernfs_idr_lock pool_lock#2 irq_context: 0 &root->kernfs_rwsem irq_context: 0 file_systems_lock irq_context: 0 &root->kernfs_rwsem &root->kernfs_iattr_rwsem irq_context: 0 sb_lock irq_context: 0 &type->s_umount_key/1 irq_context: 0 &type->s_umount_key/1 pool_lock#2 irq_context: 0 &type->s_umount_key/1 pcpu_alloc_mutex irq_context: 0 &type->s_umount_key/1 pcpu_alloc_mutex pcpu_lock irq_context: 0 &type->s_umount_key/1 shrinker_rwsem irq_context: 0 &type->s_umount_key/1 shrinker_rwsem pool_lock#2 irq_context: 0 &type->s_umount_key/1 list_lrus_mutex irq_context: 0 &type->s_umount_key/1 sb_lock irq_context: 0 &type->s_umount_key/1 sb_lock unnamed_dev_ida.xa_lock irq_context: 0 &type->s_umount_key/1 &obj_hash[i].lock irq_context: 0 &type->s_umount_key/1 percpu_counters_lock irq_context: 0 &type->s_umount_key/1 crngs.lock irq_context: 0 &type->s_umount_key/1 &sbinfo->stat_lock irq_context: 0 &type->s_umount_key/1 &pcp->lock &zone->lock irq_context: 0 &type->s_umount_key/1 &zone->lock irq_context: 0 &type->s_umount_key/1 &____s->seqcount irq_context: 0 &type->s_umount_key/1 &c->lock irq_context: 0 &type->s_umount_key/1 &sb->s_type->i_lock_key irq_context: 0 &type->s_umount_key/1 &s->s_inode_list_lock irq_context: 0 &type->s_umount_key/1 tk_core.seq.seqcount irq_context: 0 &type->s_umount_key/1 batched_entropy_u32.lock irq_context: 0 &type->s_umount_key/1 &pcp->lock &zone->lock &____s->seqcount irq_context: 0 &type->s_umount_key/1 &sb->s_type->i_lock_key &dentry->d_lock irq_context: 0 &type->s_umount_key/1 &dentry->d_lock irq_context: 0 mnt_id_ida.xa_lock irq_context: 0 &dentry->d_lock irq_context: 0 mount_lock irq_context: 0 mount_lock mount_lock.seqcount irq_context: 0 rcu_read_lock &dentry->d_lock irq_context: 0 &type->s_umount_key#2/1 irq_context: 0 &type->s_umount_key#2/1 pool_lock#2 irq_context: 0 &type->s_umount_key#2/1 pcpu_alloc_mutex irq_context: 0 &type->s_umount_key#2/1 pcpu_alloc_mutex pcpu_lock irq_context: 0 &type->s_umount_key#2/1 shrinker_rwsem irq_context: 0 &type->s_umount_key#2/1 list_lrus_mutex irq_context: 0 &type->s_umount_key#2/1 sb_lock irq_context: 0 &type->s_umount_key#2/1 sb_lock unnamed_dev_ida.xa_lock irq_context: 0 &type->s_umount_key#2/1 &pcp->lock &zone->lock irq_context: 0 &type->s_umount_key#2/1 &zone->lock irq_context: 0 &type->s_umount_key#2/1 &pcp->lock &zone->lock &____s->seqcount irq_context: 0 &type->s_umount_key#2/1 &____s->seqcount irq_context: 0 &type->s_umount_key#2/1 &c->lock irq_context: 0 &type->s_umount_key#2/1 &sb->s_type->i_lock_key#2 irq_context: 0 &type->s_umount_key#2/1 &s->s_inode_list_lock irq_context: 0 &type->s_umount_key#2/1 tk_core.seq.seqcount irq_context: 0 &type->s_umount_key#2/1 &sb->s_type->i_lock_key#2 &dentry->d_lock irq_context: 0 &type->s_umount_key#2/1 &dentry->d_lock irq_context: 0 ucounts_lock irq_context: 0 proc_inum_ida.xa_lock irq_context: 0 init_fs.lock irq_context: 0 init_fs.lock init_fs.seq.seqcount irq_context: 0 &type->s_umount_key#3/1 irq_context: 0 &type->s_umount_key#3/1 pool_lock#2 irq_context: 0 &type->s_umount_key#3/1 pcpu_alloc_mutex irq_context: 0 &type->s_umount_key#3/1 pcpu_alloc_mutex pcpu_lock irq_context: 0 &type->s_umount_key#3/1 shrinker_rwsem irq_context: 0 &type->s_umount_key#3/1 list_lrus_mutex irq_context: 0 &type->s_umount_key#3/1 sb_lock irq_context: 0 &type->s_umount_key#3/1 sb_lock unnamed_dev_ida.xa_lock irq_context: 0 &type->s_umount_key#3/1 &pcp->lock &zone->lock irq_context: 0 &type->s_umount_key#3/1 &zone->lock irq_context: 0 &type->s_umount_key#3/1 &pcp->lock &zone->lock &____s->seqcount irq_context: 0 &type->s_umount_key#3/1 &____s->seqcount irq_context: 0 &type->s_umount_key#3/1 &c->lock irq_context: 0 &type->s_umount_key#3/1 &sb->s_type->i_lock_key#3 irq_context: 0 &type->s_umount_key#3/1 &s->s_inode_list_lock irq_context: 0 &type->s_umount_key#3/1 tk_core.seq.seqcount irq_context: 0 &type->s_umount_key#3/1 &sb->s_type->i_lock_key#3 &dentry->d_lock irq_context: 0 &type->s_umount_key#3/1 &dentry->d_lock irq_context: 0 cpu_hotplug_lock cpuhp_state_mutex cpuhp_state-down irq_context: 0 cpu_hotplug_lock cpuhp_state_mutex cpuhp_state-up irq_context: 0 proc_subdir_lock irq_context: 0 proc_subdir_lock irq_context: 0 pernet_ops_rwsem pool_lock#2 irq_context: 0 pernet_ops_rwsem proc_subdir_lock irq_context: 0 pernet_ops_rwsem proc_subdir_lock irq_context: 0 &type->s_umount_key#4/1 irq_context: 0 &type->s_umount_key#4/1 pool_lock#2 irq_context: 0 &type->s_umount_key#4/1 pcpu_alloc_mutex irq_context: 0 &type->s_umount_key#4/1 pcpu_alloc_mutex pcpu_lock irq_context: 0 &type->s_umount_key#4/1 shrinker_rwsem irq_context: 0 &type->s_umount_key#4/1 list_lrus_mutex irq_context: 0 &type->s_umount_key#4/1 sb_lock irq_context: 0 &type->s_umount_key#4/1 sb_lock unnamed_dev_ida.xa_lock irq_context: 0 &type->s_umount_key#4/1 &sb->s_type->i_lock_key#4 irq_context: 0 &type->s_umount_key#4/1 &s->s_inode_list_lock irq_context: 0 &type->s_umount_key#4/1 tk_core.seq.seqcount irq_context: 0 &type->s_umount_key#4/1 &sb->s_type->i_lock_key#4 &dentry->d_lock irq_context: 0 &type->s_umount_key#4/1 &dentry->d_lock irq_context: 0 cgroup_mutex pcpu_alloc_mutex irq_context: 0 cgroup_mutex pcpu_alloc_mutex pcpu_lock irq_context: 0 cgroup_mutex pool_lock#2 irq_context: 0 cgroup_mutex lock irq_context: 0 cgroup_mutex lock kernfs_idr_lock irq_context: 0 cgroup_mutex lock kernfs_idr_lock pool_lock#2 irq_context: 0 cgroup_mutex &root->kernfs_rwsem irq_context: 0 cgroup_mutex &root->kernfs_rwsem &root->kernfs_iattr_rwsem irq_context: 0 cgroup_mutex &obj_hash[i].lock irq_context: 0 cgroup_mutex cgroup_file_kn_lock irq_context: 0 cgroup_mutex css_set_lock irq_context: 0 lock cgroup_idr_lock irq_context: 0 lock cgroup_idr_lock pool_lock#2 irq_context: 0 cpuset_mutex irq_context: 0 cpuset_mutex callback_lock irq_context: 0 cgroup_mutex &c->lock irq_context: 0 cgroup_mutex &____s->seqcount irq_context: 0 cgroup_mutex blkcg_pol_mutex irq_context: 0 cgroup_mutex blkcg_pol_mutex pcpu_alloc_mutex irq_context: 0 cgroup_mutex blkcg_pol_mutex pcpu_alloc_mutex pcpu_lock irq_context: 0 cgroup_mutex lock cgroup_idr_lock irq_context: 0 cgroup_mutex lock cgroup_idr_lock &c->lock irq_context: 0 cgroup_mutex lock cgroup_idr_lock &pcp->lock &zone->lock irq_context: 0 cgroup_mutex lock cgroup_idr_lock &zone->lock irq_context: 0 cgroup_mutex lock cgroup_idr_lock &____s->seqcount irq_context: 0 cgroup_mutex lock cgroup_idr_lock pool_lock#2 irq_context: 0 cgroup_mutex &n->list_lock irq_context: 0 cgroup_mutex &pcp->lock &zone->lock irq_context: 0 cgroup_mutex &pcp->lock &zone->lock &____s->seqcount irq_context: 0 cgroup_mutex percpu_counters_lock irq_context: 0 cgroup_mutex shrinker_rwsem irq_context: 0 cgroup_mutex shrinker_rwsem pool_lock#2 irq_context: 0 cgroup_mutex shrinker_rwsem &c->lock irq_context: 0 cgroup_mutex shrinker_rwsem &n->list_lock irq_context: 0 cgroup_mutex shrinker_rwsem &____s->seqcount irq_context: 0 cgroup_mutex &base->lock irq_context: 0 cgroup_mutex &base->lock &obj_hash[i].lock irq_context: 0 cgroup_mutex batched_entropy_u8.lock irq_context: 0 cgroup_mutex batched_entropy_u8.lock crngs.lock irq_context: 0 cgroup_mutex &pgdat->memcg_lru.lock irq_context: 0 cgroup_mutex devcgroup_mutex irq_context: 0 cgroup_mutex &zone->lock irq_context: 0 cgroup_mutex cpu_hotplug_lock irq_context: 0 cgroup_mutex cpu_hotplug_lock freezer_mutex irq_context: 0 rcu_state.exp_mutex rcu_node_0 irq_context: 0 rcu_state.exp_mutex rcu_state.exp_wake_mutex irq_context: 0 rcu_state.exp_mutex rcu_state.exp_wake_mutex rcu_node_0 irq_context: 0 rcu_state.exp_mutex rcu_state.exp_wake_mutex &rnp->exp_lock irq_context: 0 rcu_state.exp_mutex rcu_state.exp_wake_mutex &rnp->exp_wq[0] irq_context: 0 rcu_state.exp_mutex rcu_state.exp_wake_mutex &rnp->exp_wq[1] irq_context: 0 init_sighand.siglock irq_context: 0 init_files.file_lock irq_context: 0 lock pidmap_lock irq_context: 0 lock pidmap_lock pool_lock#2 irq_context: 0 pidmap_lock irq_context: 0 cgroup_threadgroup_rwsem irq_context: 0 cgroup_threadgroup_rwsem css_set_lock irq_context: 0 cgroup_threadgroup_rwsem &p->pi_lock irq_context: 0 cgroup_threadgroup_rwsem &p->pi_lock &rq->__lock irq_context: 0 cgroup_threadgroup_rwsem tk_core.seq.seqcount irq_context: 0 cgroup_threadgroup_rwsem tasklist_lock irq_context: 0 cgroup_threadgroup_rwsem tasklist_lock init_task.pi_lock irq_context: 0 cgroup_threadgroup_rwsem tasklist_lock init_sighand.siglock irq_context: 0 cgroup_threadgroup_rwsem css_set_lock cgroup_file_kn_lock irq_context: 0 cgroup_threadgroup_rwsem css_set_lock &p->pi_lock irq_context: 0 cgroup_threadgroup_rwsem css_set_lock &p->pi_lock &rq->__lock irq_context: 0 &p->pi_lock irq_context: 0 &p->pi_lock &rq->__lock irq_context: 0 &p->pi_lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 &p->pi_lock &rq->__lock &obj_hash[i].lock irq_context: 0 &p->pi_lock &rq->__lock &base->lock irq_context: 0 &p->pi_lock &rq->__lock &base->lock &obj_hash[i].lock irq_context: 0 rcu_read_lock &p->pi_lock irq_context: 0 rcu_read_lock &p->pi_lock &rq->__lock irq_context: 0 (kthreadd_done).wait.lock irq_context: 0 &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 &sighand->siglock irq_context: 0 &p->alloc_lock irq_context: 0 &p->alloc_lock &____s->seqcount#2 irq_context: 0 fs_reclaim irq_context: 0 fs_reclaim mmu_notifier_invalidate_range_start irq_context: 0 wq_pool_mutex &wq->mutex &pool->lock/1 irq_context: 0 wq_pool_mutex &pool->lock/1 irq_context: 0 wq_pool_mutex fs_reclaim irq_context: 0 wq_pool_mutex fs_reclaim mmu_notifier_invalidate_range_start irq_context: 0 wq_pool_mutex kthread_create_lock irq_context: 0 wq_pool_mutex &p->pi_lock irq_context: 0 wq_pool_mutex &p->pi_lock &rq->__lock irq_context: 0 wq_pool_mutex &p->pi_lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 wq_pool_mutex &rq->__lock irq_context: 0 kthread_create_lock irq_context: 0 cgroup_threadgroup_rwsem &p->pi_lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 cgroup_threadgroup_rwsem tasklist_lock &p->pi_lock irq_context: 0 cgroup_threadgroup_rwsem tasklist_lock &sighand->siglock irq_context: 0 &x->wait irq_context: 0 wq_pool_mutex &x->wait irq_context: 0 wq_pool_mutex &obj_hash[i].lock irq_context: 0 wq_pool_attach_mutex irq_context: 0 wq_mayday_lock irq_context: 0 &xa->xa_lock irq_context: 0 &pool->lock irq_context: 0 &pool->lock &p->pi_lock irq_context: 0 &pool->lock &p->pi_lock &rq->__lock irq_context: 0 &pool->lock &p->pi_lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 &x->wait &p->pi_lock irq_context: 0 &x->wait &p->pi_lock &rq->__lock irq_context: 0 &x->wait &p->pi_lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (&pool->mayday_timer) irq_context: 0 &pool->lock &obj_hash[i].lock irq_context: 0 (wq_completion)rcu_gp irq_context: 0 (wq_completion)rcu_gp (work_completion)(&rnp->exp_poll_wq) irq_context: 0 (wq_completion)rcu_gp (work_completion)(&rnp->exp_poll_wq) &rnp->exp_poll_lock irq_context: 0 &pool->lock/1 irq_context: 0 &pool->lock/1 &p->pi_lock irq_context: 0 &pool->lock/1 &p->pi_lock &rq->__lock irq_context: 0 &pool->lock/1 &p->pi_lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 &pool->lock/1 &obj_hash[i].lock irq_context: 0 (wq_completion)events_unbound irq_context: 0 (wq_completion)events_unbound (work_completion)(&(&kfence_timer)->work) irq_context: 0 (wq_completion)events_unbound (work_completion)(&(&kfence_timer)->work) cpu_hotplug_lock irq_context: 0 (wq_completion)events_unbound (work_completion)(&(&kfence_timer)->work) cpu_hotplug_lock jump_label_mutex irq_context: 0 (wq_completion)events_unbound (work_completion)(&(&kfence_timer)->work) cpu_hotplug_lock jump_label_mutex text_mutex irq_context: 0 (wq_completion)events_unbound (work_completion)(&(&kfence_timer)->work) cpu_hotplug_lock jump_label_mutex text_mutex ptlock_ptr(page)#2 irq_context: 0 (wq_completion)events_unbound (work_completion)(&(&kfence_timer)->work) allocation_wait.lock irq_context: 0 (wq_completion)events_unbound (work_completion)(&(&kfence_timer)->work) &rq->__lock irq_context: 0 (wq_completion)events_unbound (work_completion)(&(&kfence_timer)->work) &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (&wq_watchdog_timer) irq_context: hardirq allocation_wait.lock irq_context: hardirq allocation_wait.lock &p->pi_lock irq_context: hardirq allocation_wait.lock &p->pi_lock &rq->__lock irq_context: hardirq allocation_wait.lock &p->pi_lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)events_unbound (work_completion)(&(&kfence_timer)->work) &obj_hash[i].lock irq_context: 0 (wq_completion)events_unbound (work_completion)(&(&kfence_timer)->work) &base->lock irq_context: 0 (wq_completion)events_unbound (work_completion)(&(&kfence_timer)->work) &base->lock &obj_hash[i].lock irq_context: 0 batched_entropy_u8.lock irq_context: 0 kfence_freelist_lock irq_context: 0 rcu_tasks.tasks_gp_mutex irq_context: 0 rcu_tasks.cbs_gbl_lock irq_context: 0 rcu_tasks_trace.tasks_gp_mutex irq_context: 0 rcu_tasks_trace.cbs_gbl_lock irq_context: hardirq rcu_read_lock &p->pi_lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 rcu_tasks__percpu.cbs_pcpu_lock irq_context: 0 rcu_tasks.tasks_gp_mutex rcu_state.exp_mutex rcu_node_0 irq_context: 0 rcu_tasks.tasks_gp_mutex rcu_state.exp_mutex rcu_state.exp_wake_mutex irq_context: 0 rcu_tasks.tasks_gp_mutex rcu_state.exp_mutex rcu_state.exp_wake_mutex rcu_node_0 irq_context: 0 rcu_tasks.tasks_gp_mutex rcu_state.exp_mutex rcu_state.exp_wake_mutex &rnp->exp_lock irq_context: 0 rcu_tasks.tasks_gp_mutex rcu_state.exp_mutex rcu_state.exp_wake_mutex &rnp->exp_wq[2] irq_context: 0 rcu_tasks.tasks_gp_mutex &obj_hash[i].lock irq_context: 0 rcu_tasks.tasks_gp_mutex &base->lock irq_context: 0 rcu_tasks.tasks_gp_mutex &base->lock &obj_hash[i].lock irq_context: 0 rcu_tasks.tasks_gp_mutex tasks_rcu_exit_srcu_srcu_usage.lock irq_context: 0 rcu_tasks.tasks_gp_mutex tasks_rcu_exit_srcu_srcu_usage.lock &obj_hash[i].lock irq_context: 0 rcu_tasks.tasks_gp_mutex &ACCESS_PRIVATE(sdp, lock) irq_context: 0 rcu_tasks.tasks_gp_mutex tasks_rcu_exit_srcu irq_context: 0 rcu_tasks.tasks_gp_mutex tasks_rcu_exit_srcu_srcu_usage.lock &ACCESS_PRIVATE(sdp, lock) irq_context: 0 rcu_tasks.tasks_gp_mutex tasks_rcu_exit_srcu_srcu_usage.lock rcu_read_lock &pool->lock irq_context: 0 rcu_tasks.tasks_gp_mutex tasks_rcu_exit_srcu_srcu_usage.lock rcu_read_lock &pool->lock &obj_hash[i].lock irq_context: 0 rcu_tasks.tasks_gp_mutex tasks_rcu_exit_srcu_srcu_usage.lock rcu_read_lock &pool->lock &p->pi_lock irq_context: 0 rcu_tasks.tasks_gp_mutex tasks_rcu_exit_srcu_srcu_usage.lock rcu_read_lock &pool->lock &p->pi_lock &rq->__lock irq_context: 0 rcu_tasks.tasks_gp_mutex tasks_rcu_exit_srcu_srcu_usage.lock rcu_read_lock &pool->lock &p->pi_lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 rcu_tasks.tasks_gp_mutex &x->wait#2 irq_context: 0 rcu_tasks.tasks_gp_mutex &rq->__lock irq_context: 0 rcu_tasks.tasks_gp_mutex &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)rcu_gp (work_completion)(&(&ssp->srcu_sup->work)->work) irq_context: 0 (wq_completion)rcu_gp (work_completion)(&(&ssp->srcu_sup->work)->work) &ssp->srcu_sup->srcu_gp_mutex irq_context: 0 (wq_completion)rcu_gp (work_completion)(&(&ssp->srcu_sup->work)->work) &ssp->srcu_sup->srcu_gp_mutex tasks_rcu_exit_srcu_srcu_usage.lock irq_context: 0 (wq_completion)rcu_gp (work_completion)(&(&ssp->srcu_sup->work)->work) &ssp->srcu_sup->srcu_gp_mutex &ssp->srcu_sup->srcu_cb_mutex irq_context: 0 (wq_completion)rcu_gp (work_completion)(&(&ssp->srcu_sup->work)->work) &ssp->srcu_sup->srcu_gp_mutex &ssp->srcu_sup->srcu_cb_mutex tasks_rcu_exit_srcu_srcu_usage.lock irq_context: 0 (wq_completion)rcu_gp (work_completion)(&(&ssp->srcu_sup->work)->work) &ssp->srcu_sup->srcu_cb_mutex irq_context: 0 (wq_completion)rcu_gp (work_completion)(&(&ssp->srcu_sup->work)->work) &ssp->srcu_sup->srcu_cb_mutex rcu_read_lock &pool->lock irq_context: 0 (wq_completion)rcu_gp (work_completion)(&(&ssp->srcu_sup->work)->work) &ssp->srcu_sup->srcu_cb_mutex rcu_read_lock &pool->lock &obj_hash[i].lock irq_context: 0 (wq_completion)rcu_gp (work_completion)(&(&ssp->srcu_sup->work)->work) tasks_rcu_exit_srcu_srcu_usage.lock irq_context: 0 (wq_completion)rcu_gp (work_completion)(&sdp->work) irq_context: 0 (wq_completion)rcu_gp (work_completion)(&sdp->work) &ACCESS_PRIVATE(sdp, lock) irq_context: 0 (wq_completion)rcu_gp (work_completion)(&sdp->work) &obj_hash[i].lock irq_context: 0 (wq_completion)rcu_gp (work_completion)(&sdp->work) &x->wait#2 irq_context: 0 (wq_completion)rcu_gp (work_completion)(&sdp->work) &x->wait#2 &p->pi_lock irq_context: 0 (wq_completion)rcu_gp (work_completion)(&sdp->work) &x->wait#2 &p->pi_lock &rq->__lock irq_context: 0 (wq_completion)rcu_gp (work_completion)(&sdp->work) &x->wait#2 &p->pi_lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 rcu_tasks.tasks_gp_mutex kernel/rcu/tasks.h:152 irq_context: 0 &x->wait#3 irq_context: 0 (null) irq_context: 0 (null) tk_core.seq.seqcount irq_context: softirq (&timer.timer) irq_context: softirq (&timer.timer) &p->pi_lock irq_context: softirq (&timer.timer) &p->pi_lock &rq->__lock irq_context: softirq (&timer.timer) &p->pi_lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: softirq &(&kfence_timer)->timer irq_context: softirq &(&kfence_timer)->timer rcu_read_lock &pool->lock/1 irq_context: softirq &(&kfence_timer)->timer rcu_read_lock &pool->lock/1 &obj_hash[i].lock irq_context: softirq &(&kfence_timer)->timer rcu_read_lock &pool->lock/1 &p->pi_lock irq_context: softirq &(&kfence_timer)->timer rcu_read_lock &pool->lock/1 &p->pi_lock &rq->__lock irq_context: softirq &(&kfence_timer)->timer rcu_read_lock &pool->lock/1 &p->pi_lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 rcu_tasks.tasks_gp_mutex (&timer.timer) irq_context: 0 rcu_tasks.tasks_gp_mutex rcu_state.exp_mutex rcu_state.exp_wake_mutex &rnp->exp_wq[3] irq_context: 0 rcu_tasks.tasks_gp_mutex rcu_tasks__percpu.cbs_pcpu_lock irq_context: 0 rcu_tasks.tasks_gp_mutex &x->wait#3 irq_context: 0 rcu_tasks.tasks_gp_mutex &x->wait#3 &p->pi_lock irq_context: 0 rcu_tasks.tasks_gp_mutex &x->wait#3 &p->pi_lock &rq->__lock irq_context: 0 rcu_tasks.tasks_gp_mutex &x->wait#3 &p->pi_lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 rcu_read_lock rcu_tasks__percpu.cbs_pcpu_lock &obj_hash[i].lock irq_context: 0 rcu_read_lock rcu_tasks__percpu.cbs_pcpu_lock &base->lock irq_context: 0 rcu_read_lock rcu_tasks__percpu.cbs_pcpu_lock &base->lock &obj_hash[i].lock irq_context: 0 rcu_tasks_trace__percpu.cbs_pcpu_lock irq_context: 0 rcu_tasks_trace.tasks_gp_mutex cpu_hotplug_lock irq_context: 0 rcu_tasks_trace.tasks_gp_mutex cpu_hotplug_lock rcu_tasks_trace__percpu.cbs_pcpu_lock irq_context: 0 rcu_tasks_trace.tasks_gp_mutex cpu_hotplug_lock &ACCESS_PRIVATE(rtpcp, lock) irq_context: 0 rcu_tasks_trace.tasks_gp_mutex rcu_state.exp_mutex rcu_node_0 irq_context: 0 rcu_tasks_trace.tasks_gp_mutex rcu_state.exp_mutex rcu_state.exp_wake_mutex irq_context: 0 rcu_tasks_trace.tasks_gp_mutex rcu_state.exp_mutex rcu_state.exp_wake_mutex rcu_node_0 irq_context: 0 rcu_tasks_trace.tasks_gp_mutex rcu_state.exp_mutex rcu_state.exp_wake_mutex &rnp->exp_lock irq_context: 0 rcu_tasks_trace.tasks_gp_mutex rcu_state.exp_mutex rcu_state.exp_wake_mutex &rnp->exp_wq[0] irq_context: 0 rcu_tasks_trace.tasks_gp_mutex rcu_tasks_trace__percpu.cbs_pcpu_lock irq_context: 0 rcu_tasks_trace.tasks_gp_mutex &x->wait#3 irq_context: 0 rcu_read_lock rcu_tasks_trace__percpu.cbs_pcpu_lock &obj_hash[i].lock irq_context: 0 rcu_read_lock rcu_tasks_trace__percpu.cbs_pcpu_lock &base->lock irq_context: 0 rcu_read_lock rcu_tasks_trace__percpu.cbs_pcpu_lock &base->lock &obj_hash[i].lock irq_context: 0 &nmi_desc[0].lock irq_context: 0 cpu_hotplug_lock smpboot_threads_lock irq_context: 0 cpu_hotplug_lock smpboot_threads_lock fs_reclaim irq_context: 0 cpu_hotplug_lock smpboot_threads_lock fs_reclaim mmu_notifier_invalidate_range_start irq_context: 0 cpu_hotplug_lock smpboot_threads_lock &rq->__lock irq_context: 0 cpu_hotplug_lock smpboot_threads_lock batched_entropy_u8.lock irq_context: 0 cpu_hotplug_lock smpboot_threads_lock kfence_freelist_lock irq_context: 0 cpu_hotplug_lock smpboot_threads_lock pool_lock#2 irq_context: 0 cpu_hotplug_lock smpboot_threads_lock kthread_create_lock irq_context: 0 cpu_hotplug_lock smpboot_threads_lock &p->pi_lock irq_context: 0 cpu_hotplug_lock smpboot_threads_lock &p->pi_lock &rq->__lock irq_context: 0 cpu_hotplug_lock smpboot_threads_lock &p->pi_lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 cpu_hotplug_lock smpboot_threads_lock &x->wait irq_context: 0 cpu_hotplug_lock smpboot_threads_lock &obj_hash[i].lock irq_context: 0 &rcu_state.gp_wq irq_context: 0 cpu_hotplug_lock smpboot_threads_lock &c->lock irq_context: 0 cpu_hotplug_lock smpboot_threads_lock &n->list_lock irq_context: 0 &stop_pi_lock irq_context: 0 &stop_pi_lock &rq->__lock irq_context: 0 &stop_pi_lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 &stopper->lock irq_context: 0 (module_notify_list).rwsem irq_context: 0 ddebug_lock irq_context: 0 &pmus_srcu irq_context: softirq &rcu_state.gp_wq &p->pi_lock irq_context: softirq &rcu_state.gp_wq &p->pi_lock &rq->__lock irq_context: softirq &rcu_state.gp_wq &p->pi_lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: softirq rcu_callback &obj_hash[i].lock irq_context: softirq rcu_callback pool_lock#2 irq_context: 0 watchdog_mutex irq_context: 0 watchdog_mutex cpu_hotplug_lock irq_context: 0 watchdog_mutex cpu_hotplug_lock &obj_hash[i].lock irq_context: 0 watchdog_mutex cpu_hotplug_lock rcu_read_lock &pool->lock irq_context: 0 watchdog_mutex cpu_hotplug_lock rcu_read_lock &pool->lock &obj_hash[i].lock irq_context: 0 watchdog_mutex cpu_hotplug_lock rcu_read_lock &pool->lock &p->pi_lock irq_context: 0 watchdog_mutex cpu_hotplug_lock rcu_read_lock &pool->lock &p->pi_lock &rq->__lock irq_context: 0 watchdog_mutex cpu_hotplug_lock rcu_read_lock &pool->lock &p->pi_lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 watchdog_mutex cpu_hotplug_lock &x->wait#4 irq_context: 0 watchdog_mutex cpu_hotplug_lock &rq->__lock irq_context: 0 (wq_completion)events irq_context: 0 (wq_completion)events (work_completion)(&sscs.work) irq_context: 0 (wq_completion)events (work_completion)(&sscs.work) &x->wait#5 irq_context: 0 (wq_completion)events (work_completion)(&sscs.work) &obj_hash[i].lock irq_context: 0 (wq_completion)events (work_completion)(&sscs.work) hrtimer_bases.lock irq_context: 0 (wq_completion)events (work_completion)(&sscs.work) hrtimer_bases.lock tk_core.seq.seqcount irq_context: 0 (wq_completion)events (work_completion)(&sscs.work) hrtimer_bases.lock &obj_hash[i].lock irq_context: 0 (wq_completion)events (work_completion)(&sscs.work) &x->wait#4 irq_context: 0 (wq_completion)events (work_completion)(&sscs.work) &x->wait#4 &p->pi_lock irq_context: 0 (wq_completion)events (work_completion)(&sscs.work) &x->wait#4 &p->pi_lock &rq->__lock irq_context: 0 (wq_completion)events (work_completion)(&sscs.work) &x->wait#4 &p->pi_lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 watchdog_mutex cpu_hotplug_lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 &newf->file_lock irq_context: 0 init_fs.lock &dentry->d_lock irq_context: 0 &p->vtime.seqcount irq_context: 0 cpu_hotplug_lock mem_hotplug_lock irq_context: 0 cpu_hotplug_lock mem_hotplug_lock mem_hotplug_lock.rss.gp_wait.lock irq_context: 0 cpu_hotplug_lock mem_hotplug_lock rcu_state.exp_mutex rcu_node_0 irq_context: 0 cpu_hotplug_lock mem_hotplug_lock rcu_state.exp_mutex rcu_state.exp_wake_mutex irq_context: 0 cpu_hotplug_lock mem_hotplug_lock rcu_state.exp_mutex rcu_state.exp_wake_mutex rcu_node_0 irq_context: 0 cpu_hotplug_lock mem_hotplug_lock rcu_state.exp_mutex rcu_state.exp_wake_mutex &rnp->exp_lock irq_context: 0 cpu_hotplug_lock mem_hotplug_lock rcu_state.exp_mutex rcu_state.exp_wake_mutex &rnp->exp_wq[1] irq_context: 0 cpu_hotplug_lock mem_hotplug_lock.waiters.lock irq_context: 0 cpu_hotplug_lock mem_hotplug_lock.rss.gp_wait.lock irq_context: 0 cpu_hotplug_lock mem_hotplug_lock.rss.gp_wait.lock &obj_hash[i].lock irq_context: 0 cpu_add_remove_lock irq_context: 0 cpu_add_remove_lock cpu_hotplug_lock irq_context: 0 cpu_add_remove_lock cpu_hotplug_lock cpu_hotplug_lock.rss.gp_wait.lock irq_context: 0 cpu_add_remove_lock cpu_hotplug_lock rcu_state.exp_mutex rcu_node_0 irq_context: 0 cpu_add_remove_lock cpu_hotplug_lock rcu_state.exp_mutex rcu_state.exp_wake_mutex irq_context: 0 cpu_add_remove_lock cpu_hotplug_lock rcu_state.exp_mutex rcu_state.exp_wake_mutex rcu_node_0 irq_context: 0 cpu_add_remove_lock cpu_hotplug_lock rcu_state.exp_mutex rcu_state.exp_wake_mutex &rnp->exp_lock irq_context: 0 cpu_add_remove_lock cpu_hotplug_lock rcu_state.exp_mutex rcu_state.exp_wake_mutex &rnp->exp_wq[2] irq_context: 0 cpu_add_remove_lock cpu_hotplug_lock.waiters.lock irq_context: 0 cpu_add_remove_lock cpu_hotplug_lock.rss.gp_wait.lock irq_context: 0 cpu_add_remove_lock cpu_hotplug_lock.rss.gp_wait.lock &obj_hash[i].lock irq_context: 0 cpu_add_remove_lock spec_ctrl_mutex irq_context: 0 cpu_add_remove_lock spec_ctrl_mutex cpu_hotplug_lock irq_context: 0 cpu_add_remove_lock cpu_hotplug_lock irq_context: 0 cpu_add_remove_lock cpuset_hotplug_work irq_context: 0 cpu_add_remove_lock cpu_hotplug_lock smpboot_threads_lock irq_context: 0 cpu_add_remove_lock cpu_hotplug_lock smpboot_threads_lock fs_reclaim irq_context: 0 cpu_add_remove_lock cpu_hotplug_lock smpboot_threads_lock fs_reclaim mmu_notifier_invalidate_range_start irq_context: 0 cpu_add_remove_lock cpu_hotplug_lock smpboot_threads_lock pool_lock#2 irq_context: 0 cpu_add_remove_lock cpu_hotplug_lock smpboot_threads_lock kthread_create_lock irq_context: 0 cpu_add_remove_lock cpu_hotplug_lock smpboot_threads_lock &p->pi_lock irq_context: 0 cpu_add_remove_lock cpu_hotplug_lock smpboot_threads_lock &p->pi_lock &rq->__lock irq_context: 0 cpu_add_remove_lock cpu_hotplug_lock smpboot_threads_lock &p->pi_lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 cpu_add_remove_lock cpu_hotplug_lock smpboot_threads_lock &rq->__lock irq_context: 0 cpu_add_remove_lock cpu_hotplug_lock smpboot_threads_lock &x->wait irq_context: 0 cpu_add_remove_lock cpu_hotplug_lock smpboot_threads_lock &obj_hash[i].lock irq_context: 0 cpu_add_remove_lock cpu_hotplug_lock &swhash->hlist_mutex irq_context: 0 cpu_add_remove_lock cpu_hotplug_lock pmus_lock irq_context: 0 cpu_add_remove_lock cpu_hotplug_lock pmus_lock &cpuctx_mutex irq_context: 0 cpu_add_remove_lock cpu_hotplug_lock pcp_batch_high_lock irq_context: 0 cpu_add_remove_lock cpu_hotplug_lock &xa->xa_lock irq_context: 0 cpu_add_remove_lock cpu_hotplug_lock fs_reclaim irq_context: 0 cpu_add_remove_lock cpu_hotplug_lock fs_reclaim mmu_notifier_invalidate_range_start irq_context: 0 cpu_add_remove_lock cpu_hotplug_lock pool_lock#2 irq_context: 0 cpu_add_remove_lock cpu_hotplug_lock kthread_create_lock irq_context: 0 cpu_add_remove_lock cpu_hotplug_lock &p->pi_lock irq_context: 0 cpu_add_remove_lock cpu_hotplug_lock &p->pi_lock &rq->__lock irq_context: 0 cpu_add_remove_lock cpu_hotplug_lock &p->pi_lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 cpu_add_remove_lock cpu_hotplug_lock &rq->__lock irq_context: 0 cpu_add_remove_lock cpu_hotplug_lock &x->wait irq_context: 0 cpu_add_remove_lock cpu_hotplug_lock &obj_hash[i].lock irq_context: 0 cpu_add_remove_lock cpu_hotplug_lock wq_pool_attach_mutex irq_context: 0 cpu_add_remove_lock cpu_hotplug_lock &pool->lock irq_context: 0 cpu_add_remove_lock cpu_hotplug_lock &pool->lock &p->pi_lock irq_context: 0 cpu_add_remove_lock cpu_hotplug_lock &pool->lock &p->pi_lock &rq->__lock irq_context: 0 cpu_add_remove_lock cpu_hotplug_lock &pool->lock &p->pi_lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 cpu_add_remove_lock cpu_hotplug_lock pcpu_alloc_mutex irq_context: 0 cpu_add_remove_lock cpu_hotplug_lock pcpu_alloc_mutex pcpu_lock irq_context: 0 cpu_add_remove_lock cpu_hotplug_lock relay_channels_mutex irq_context: 0 cpu_add_remove_lock cpu_hotplug_lock rcu_node_0 irq_context: 0 cpu_add_remove_lock cpu_hotplug_lock &c->lock irq_context: 0 cpu_add_remove_lock cpu_hotplug_lock &n->list_lock irq_context: 0 cpu_add_remove_lock cpu_hotplug_lock &pcp->lock &zone->lock irq_context: 0 cpu_add_remove_lock cpu_hotplug_lock &zone->lock irq_context: 0 cpu_add_remove_lock cpu_hotplug_lock &____s->seqcount irq_context: 0 cpu_add_remove_lock cpu_hotplug_lock text_mutex irq_context: 0 cpu_add_remove_lock cpu_hotplug_lock free_vmap_area_lock irq_context: 0 cpu_add_remove_lock cpu_hotplug_lock vmap_area_lock irq_context: 0 cpu_add_remove_lock cpu_hotplug_lock init_mm.page_table_lock irq_context: 0 cpu_add_remove_lock cpu_hotplug_lock (console_sem).lock irq_context: 0 cpu_add_remove_lock cpu_hotplug_lock console_lock console_srcu console_owner_lock irq_context: 0 cpu_add_remove_lock cpu_hotplug_lock console_lock console_srcu console_owner irq_context: 0 cpu_add_remove_lock cpu_hotplug_lock console_lock console_srcu console_owner &port_lock_key irq_context: 0 cpu_add_remove_lock cpu_hotplug_lock console_lock console_srcu console_owner console_owner_lock irq_context: 0 cpu_add_remove_lock cpu_hotplug_lock rtc_lock irq_context: 0 cpu_add_remove_lock cpu_hotplug_lock sparse_irq_lock irq_context: 0 cpu_add_remove_lock cpu_hotplug_lock sparse_irq_lock tk_core.seq.seqcount irq_context: 0 cpu_add_remove_lock cpu_hotplug_lock sparse_irq_lock rtc_lock irq_context: 0 cpu_add_remove_lock cpu_hotplug_lock sparse_irq_lock &x->wait#6 irq_context: 0 cpu_add_remove_lock cpu_hotplug_lock sparse_irq_lock &rq->__lock irq_context: 0 cpu_add_remove_lock cpu_hotplug_lock sparse_irq_lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 &rq->__lock &rq->__lock/1 irq_context: 0 &rq->__lock/1 irq_context: 0 &x->wait#6 irq_context: 0 &x->wait#6 &p->pi_lock irq_context: 0 cpu_add_remove_lock cpu_hotplug_lock sparse_irq_lock &p->pi_lock irq_context: 0 cpu_add_remove_lock cpu_hotplug_lock sparse_irq_lock &p->pi_lock &rq->__lock irq_context: 0 cpu_add_remove_lock cpu_hotplug_lock sparse_irq_lock &p->pi_lock &cfs_rq->removed.lock irq_context: 0 cpu_add_remove_lock cpu_hotplug_lock sparse_irq_lock &p->pi_lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 cpu_add_remove_lock cpu_hotplug_lock &x->wait#6 irq_context: 0 cpu_add_remove_lock cpu_hotplug_lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 cpu_add_remove_lock cpu_hotplug_lock &rq->__lock &cfs_rq->removed.lock irq_context: 0 cpu_hotplug_lock irq_context: 0 cpu_hotplug_lock cpuhp_state-up irq_context: 0 cpu_hotplug_lock cpuhp_state-up smpboot_threads_lock irq_context: 0 cpu_hotplug_lock cpuhp_state-up smpboot_threads_lock &p->pi_lock irq_context: 0 cpu_hotplug_lock cpuhp_state-up smpboot_threads_lock &p->pi_lock &rq->__lock irq_context: 0 cpu_hotplug_lock cpuhp_state-up smpboot_threads_lock &p->pi_lock &cfs_rq->removed.lock irq_context: 0 cpu_hotplug_lock cpuhp_state-up smpboot_threads_lock &p->pi_lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 cpu_hotplug_lock cpuhp_state-up smpboot_threads_lock &rq->__lock irq_context: 0 cpu_hotplug_lock cpuhp_state-up sparse_irq_lock irq_context: 0 cpu_hotplug_lock cpuhp_state-up sparse_irq_lock &irq_desc_lock_class irq_context: 0 cpu_hotplug_lock cpuhp_state-up &swhash->hlist_mutex irq_context: 0 cpu_hotplug_lock cpuhp_state-up pmus_lock irq_context: 0 cpu_hotplug_lock cpuhp_state-up pmus_lock &cpuctx_mutex irq_context: 0 cpu_hotplug_lock cpuhp_state-up &x->wait#5 irq_context: 0 cpu_hotplug_lock cpuhp_state-up &obj_hash[i].lock irq_context: 0 cpu_hotplug_lock cpuhp_state-up hrtimer_bases.lock irq_context: 0 cpu_hotplug_lock cpuhp_state-up hrtimer_bases.lock tk_core.seq.seqcount irq_context: 0 cpu_hotplug_lock cpuhp_state-up hrtimer_bases.lock &obj_hash[i].lock irq_context: 0 cpu_hotplug_lock cpuhp_state-up wq_pool_mutex irq_context: 0 cpu_hotplug_lock cpuhp_state-up wq_pool_mutex wq_pool_attach_mutex irq_context: 0 cpu_hotplug_lock cpuhp_state-up wq_pool_mutex wq_pool_attach_mutex &p->pi_lock irq_context: 0 cpu_hotplug_lock cpuhp_state-up wq_pool_mutex wq_pool_attach_mutex &p->pi_lock &rq->__lock irq_context: 0 cpu_hotplug_lock cpuhp_state-up wq_pool_mutex wq_pool_attach_mutex &x->wait#7 irq_context: 0 cpu_hotplug_lock cpuhp_state-up wq_pool_mutex wq_pool_attach_mutex &pool->lock irq_context: 0 cpu_hotplug_lock cpuhp_state-up wq_pool_mutex &wq->mutex irq_context: 0 cpu_hotplug_lock cpuhp_state-up wq_pool_mutex &wq->mutex &pool->lock/1 irq_context: 0 cpu_hotplug_lock cpuhp_state-up wq_pool_mutex &pool->lock/1 irq_context: 0 cpu_hotplug_lock cpuhp_state-up rcu_node_0 irq_context: 0 cpu_hotplug_lock cpuhp_state-up &rq->__lock irq_context: 0 cpu_hotplug_lock cpuhp_state-up jump_label_mutex irq_context: 0 cpu_hotplug_lock cpuhp_state-up jump_label_mutex text_mutex irq_context: 0 cpu_hotplug_lock cpuhp_state-up jump_label_mutex text_mutex ptlock_ptr(page)#2 irq_context: 0 cpu_hotplug_lock cpuhp_state-up &rq->__lock &rt_b->rt_runtime_lock irq_context: 0 cpu_hotplug_lock cpuhp_state-up &rq->__lock &rt_b->rt_runtime_lock &rt_rq->rt_runtime_lock irq_context: 0 cpu_hotplug_lock cpuhp_state-up &rq->__lock rcu_read_lock &cfs_b->lock irq_context: hardirq &rq->__lock &cfs_rq->removed.lock irq_context: 0 cpu_add_remove_lock spec_ctrl_mutex cpu_hotplug_lock jump_label_mutex irq_context: 0 cpu_add_remove_lock spec_ctrl_mutex cpu_hotplug_lock jump_label_mutex text_mutex irq_context: 0 cpu_add_remove_lock spec_ctrl_mutex cpu_hotplug_lock jump_label_mutex text_mutex ptlock_ptr(page)#2 irq_context: 0 cpu_add_remove_lock spec_ctrl_mutex (console_sem).lock irq_context: 0 cpu_add_remove_lock spec_ctrl_mutex console_lock console_srcu console_owner_lock irq_context: 0 cpu_add_remove_lock spec_ctrl_mutex console_lock console_srcu console_owner irq_context: 0 cpu_add_remove_lock spec_ctrl_mutex console_lock console_srcu console_owner &port_lock_key irq_context: 0 cpu_add_remove_lock spec_ctrl_mutex console_lock console_srcu console_owner console_owner_lock irq_context: 0 cpu_add_remove_lock spec_ctrl_mutex &rq->__lock irq_context: 0 rcu_tasks_trace.tasks_gp_mutex rcu_state.exp_mutex rcu_state.exp_wake_mutex &rnp->exp_wq[3] irq_context: 0 rcu_tasks_trace.tasks_gp_mutex &obj_hash[i].lock irq_context: 0 rcu_tasks_trace.tasks_gp_mutex &base->lock irq_context: 0 rcu_tasks_trace.tasks_gp_mutex &base->lock &obj_hash[i].lock irq_context: 0 rcu_tasks_trace.tasks_gp_mutex &rq->__lock irq_context: 0 rcu_tasks_trace.tasks_gp_mutex &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 cpu_add_remove_lock cpu_hotplug_lock jump_label_mutex irq_context: 0 cpu_add_remove_lock cpu_hotplug_lock jump_label_mutex text_mutex irq_context: 0 cpu_add_remove_lock cpu_hotplug_lock jump_label_mutex text_mutex ptlock_ptr(page)#2 irq_context: 0 cpu_hotplug_lock stop_cpus_mutex irq_context: 0 cpu_hotplug_lock stop_cpus_mutex &stopper->lock irq_context: 0 cpu_hotplug_lock stop_cpus_mutex &stop_pi_lock irq_context: 0 cpu_hotplug_lock stop_cpus_mutex &stop_pi_lock &rq->__lock irq_context: 0 cpu_hotplug_lock stop_cpus_mutex &stop_pi_lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 cpu_hotplug_lock stop_cpus_mutex &rq->__lock irq_context: 0 &x->wait#8 irq_context: 0 cpu_hotplug_lock stop_cpus_mutex &x->wait#8 irq_context: 0 sched_domains_mutex irq_context: 0 sched_domains_mutex fs_reclaim irq_context: 0 sched_domains_mutex fs_reclaim mmu_notifier_invalidate_range_start irq_context: 0 sched_domains_mutex pool_lock#2 irq_context: 0 sched_domains_mutex &obj_hash[i].lock irq_context: 0 sched_domains_mutex pcpu_alloc_mutex irq_context: 0 sched_domains_mutex pcpu_alloc_mutex pcpu_lock irq_context: 0 sched_domains_mutex &pcp->lock &zone->lock irq_context: 0 sched_domains_mutex &zone->lock irq_context: 0 sched_domains_mutex &____s->seqcount irq_context: 0 sched_domains_mutex rcu_read_lock pool_lock#2 irq_context: 0 sched_domains_mutex &pcp->lock &zone->lock &____s->seqcount irq_context: 0 sched_domains_mutex &c->lock irq_context: 0 sched_domains_mutex rcu_read_lock &rq->__lock irq_context: 0 sched_domains_mutex rcu_read_lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 sched_domains_mutex rcu_read_lock &rq->__lock &cp->lock irq_context: 0 sched_domains_mutex rcu_read_lock &rq->__lock &rt_b->rt_runtime_lock irq_context: 0 sched_domains_mutex rcu_read_lock &rq->__lock &rt_b->rt_runtime_lock &rt_rq->rt_runtime_lock irq_context: 0 sched_domains_mutex rcu_read_lock &rq->__lock rcu_read_lock &cfs_b->lock irq_context: 0 sched_domains_mutex pcpu_lock irq_context: 0 slab_mutex fs_reclaim irq_context: 0 slab_mutex fs_reclaim mmu_notifier_invalidate_range_start irq_context: 0 (memory_chain).rwsem irq_context: 0 cpu_hotplug_lock wq_pool_mutex fs_reclaim irq_context: 0 cpu_hotplug_lock wq_pool_mutex fs_reclaim mmu_notifier_invalidate_range_start irq_context: 0 &type->s_umount_key#5/1 irq_context: 0 &type->s_umount_key#5/1 fs_reclaim irq_context: 0 &type->s_umount_key#5/1 fs_reclaim mmu_notifier_invalidate_range_start irq_context: 0 &type->s_umount_key#5/1 pool_lock#2 irq_context: 0 &type->s_umount_key#5/1 pcpu_alloc_mutex irq_context: 0 &type->s_umount_key#5/1 pcpu_alloc_mutex pcpu_lock irq_context: 0 &type->s_umount_key#5/1 shrinker_rwsem irq_context: 0 &type->s_umount_key#5/1 list_lrus_mutex irq_context: 0 &type->s_umount_key#5/1 sb_lock irq_context: 0 &type->s_umount_key#5/1 sb_lock unnamed_dev_ida.xa_lock irq_context: 0 &type->s_umount_key#5/1 &obj_hash[i].lock irq_context: 0 &type->s_umount_key#5/1 percpu_counters_lock irq_context: 0 &type->s_umount_key#5/1 crngs.lock irq_context: 0 &type->s_umount_key#5/1 &sbinfo->stat_lock irq_context: 0 &type->s_umount_key#5/1 mmu_notifier_invalidate_range_start irq_context: 0 &type->s_umount_key#5/1 &sb->s_type->i_lock_key#5 irq_context: 0 &type->s_umount_key#5/1 &s->s_inode_list_lock irq_context: 0 &type->s_umount_key#5/1 tk_core.seq.seqcount irq_context: 0 &type->s_umount_key#5/1 batched_entropy_u32.lock irq_context: 0 &type->s_umount_key#5/1 &sb->s_type->i_lock_key#5 &dentry->d_lock irq_context: 0 &type->s_umount_key#5/1 &dentry->d_lock irq_context: 0 (setup_done).wait.lock irq_context: 0 namespace_sem irq_context: 0 namespace_sem fs_reclaim irq_context: 0 namespace_sem fs_reclaim mmu_notifier_invalidate_range_start irq_context: 0 namespace_sem &pcp->lock &zone->lock irq_context: 0 namespace_sem &zone->lock irq_context: 0 namespace_sem &____s->seqcount irq_context: 0 namespace_sem pool_lock#2 irq_context: 0 namespace_sem &c->lock irq_context: 0 namespace_sem mnt_id_ida.xa_lock irq_context: 0 namespace_sem pcpu_alloc_mutex irq_context: 0 namespace_sem pcpu_alloc_mutex pcpu_lock irq_context: 0 namespace_sem &dentry->d_lock irq_context: 0 namespace_sem mount_lock irq_context: 0 namespace_sem mount_lock mount_lock.seqcount irq_context: 0 &p->alloc_lock init_fs.lock irq_context: 0 rcu_read_lock &____s->seqcount#3 irq_context: 0 file_systems_lock irq_context: 0 &type->s_umount_key#6 irq_context: 0 &type->s_umount_key#6 fs_reclaim irq_context: 0 &type->s_umount_key#6 fs_reclaim mmu_notifier_invalidate_range_start irq_context: 0 &type->s_umount_key#6 pool_lock#2 irq_context: 0 &type->s_umount_key#6 &dentry->d_lock irq_context: 0 &type->s_umount_key#6 &pcp->lock &zone->lock irq_context: 0 &type->s_umount_key#6 &zone->lock irq_context: 0 &type->s_umount_key#6 &____s->seqcount irq_context: 0 &type->s_umount_key#6 &c->lock irq_context: 0 &type->s_umount_key#6 &lru->node[i].lock irq_context: 0 &type->s_umount_key#6 &sbinfo->stat_lock irq_context: 0 &type->s_umount_key#6 rcu_read_lock &dentry->d_lock irq_context: 0 &type->s_umount_key#6 &obj_hash[i].lock irq_context: 0 &sb->s_type->i_mutex_key irq_context: 0 &sb->s_type->i_mutex_key namespace_sem irq_context: 0 &sb->s_type->i_mutex_key namespace_sem rcu_read_lock mount_lock.seqcount irq_context: 0 &sb->s_type->i_mutex_key namespace_sem fs_reclaim irq_context: 0 &sb->s_type->i_mutex_key namespace_sem fs_reclaim mmu_notifier_invalidate_range_start irq_context: 0 &sb->s_type->i_mutex_key namespace_sem &pcp->lock &zone->lock irq_context: 0 &sb->s_type->i_mutex_key namespace_sem &zone->lock irq_context: 0 &sb->s_type->i_mutex_key namespace_sem &____s->seqcount irq_context: 0 &sb->s_type->i_mutex_key namespace_sem pool_lock#2 irq_context: 0 &sb->s_type->i_mutex_key namespace_sem &c->lock irq_context: 0 &sb->s_type->i_mutex_key namespace_sem rename_lock irq_context: 0 &sb->s_type->i_mutex_key namespace_sem rename_lock rename_lock.seqcount irq_context: 0 &sb->s_type->i_mutex_key namespace_sem rename_lock rename_lock.seqcount &dentry->d_lock irq_context: 0 &sb->s_type->i_mutex_key namespace_sem mount_lock irq_context: 0 &sb->s_type->i_mutex_key namespace_sem mount_lock &dentry->d_lock irq_context: 0 &sb->s_type->i_mutex_key namespace_sem mount_lock mount_lock.seqcount irq_context: 0 &sb->s_type->i_mutex_key namespace_sem mount_lock mount_lock.seqcount &new_ns->poll irq_context: 0 &sb->s_type->i_mutex_key namespace_sem mount_lock mount_lock.seqcount &dentry->d_lock irq_context: 0 &sb->s_type->i_mutex_key namespace_sem mount_lock mount_lock.seqcount rcu_read_lock &dentry->d_lock irq_context: 0 &sb->s_type->i_mutex_key namespace_sem mount_lock mount_lock.seqcount &obj_hash[i].lock irq_context: 0 &sb->s_type->i_mutex_key namespace_sem mount_lock mount_lock.seqcount pool_lock#2 irq_context: 0 rcu_read_lock &sb->s_type->i_lock_key#2 irq_context: 0 rcu_read_lock &____s->seqcount#4 irq_context: 0 &sb->s_type->i_lock_key#5 irq_context: 0 &fs->lock irq_context: 0 &fs->lock &____s->seqcount#3 irq_context: 0 (setup_done).wait.lock &p->pi_lock irq_context: 0 (setup_done).wait.lock &p->pi_lock &cfs_rq->removed.lock irq_context: 0 (setup_done).wait.lock &p->pi_lock &rq->__lock irq_context: 0 (setup_done).wait.lock &p->pi_lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 req_lock irq_context: 0 of_mutex irq_context: 0 of_mutex fs_reclaim irq_context: 0 of_mutex fs_reclaim mmu_notifier_invalidate_range_start irq_context: 0 of_mutex pool_lock#2 irq_context: 0 of_mutex lock irq_context: 0 of_mutex lock kernfs_idr_lock irq_context: 0 of_mutex &root->kernfs_rwsem irq_context: 0 of_mutex &root->kernfs_rwsem &root->kernfs_iattr_rwsem irq_context: 0 &x->wait#9 irq_context: 0 &k->list_lock irq_context: 0 bus_type_sem irq_context: 0 &root->kernfs_rwsem irq_context: 0 &dev->power.lock irq_context: 0 dpm_list_mtx irq_context: 0 uevent_sock_mutex irq_context: 0 running_helpers_waitq.lock irq_context: 0 sysfs_symlink_target_lock irq_context: 0 &k->k_lock irq_context: 0 &dev->mutex &k->list_lock irq_context: 0 &dev->mutex &k->k_lock irq_context: 0 &dev->mutex &dev->power.lock irq_context: 0 subsys mutex irq_context: 0 memory_blocks.xa_lock irq_context: 0 memory_blocks.xa_lock pool_lock#2 irq_context: softirq &rcu_state.gp_wq &p->pi_lock &cfs_rq->removed.lock irq_context: 0 rcu_read_lock rcu_node_0 irq_context: 0 rcu_read_lock &rq->__lock irq_context: softirq (&timer.timer) &p->pi_lock &rq->__lock &cfs_rq->removed.lock irq_context: softirq rcu_callback mem_hotplug_lock.rss.gp_wait.lock irq_context: softirq rcu_callback mem_hotplug_lock.rss.gp_wait.lock &obj_hash[i].lock irq_context: softirq rcu_callback cpu_hotplug_lock.rss.gp_wait.lock irq_context: softirq rcu_callback cpu_hotplug_lock.rss.gp_wait.lock &obj_hash[i].lock irq_context: 0 rcu_tasks_trace.tasks_gp_mutex (&timer.timer) irq_context: 0 rcu_tasks_trace.tasks_gp_mutex (console_sem).lock irq_context: 0 rcu_tasks_trace.tasks_gp_mutex console_lock console_srcu console_owner_lock irq_context: 0 rcu_tasks_trace.tasks_gp_mutex console_lock console_srcu console_owner irq_context: 0 rcu_tasks_trace.tasks_gp_mutex console_lock console_srcu console_owner &port_lock_key irq_context: 0 rcu_tasks_trace.tasks_gp_mutex console_lock console_srcu console_owner console_owner_lock irq_context: 0 lock kernfs_idr_lock &c->lock irq_context: 0 lock kernfs_idr_lock &____s->seqcount irq_context: 0 subsys mutex#2 irq_context: 0 register_lock irq_context: 0 register_lock proc_subdir_lock irq_context: 0 register_lock fs_reclaim irq_context: 0 register_lock fs_reclaim mmu_notifier_invalidate_range_start irq_context: 0 register_lock pool_lock#2 irq_context: 0 register_lock proc_inum_ida.xa_lock irq_context: 0 register_lock proc_subdir_lock irq_context: 0 register_lock &c->lock irq_context: 0 register_lock &pcp->lock &zone->lock irq_context: 0 register_lock &zone->lock irq_context: 0 register_lock &____s->seqcount irq_context: 0 register_lock proc_inum_ida.xa_lock &____s->seqcount irq_context: 0 register_lock proc_inum_ida.xa_lock pool_lock#2 irq_context: 0 register_lock proc_inum_ida.xa_lock &c->lock irq_context: 0 register_lock &pcp->lock &zone->lock &____s->seqcount irq_context: 0 &p->pi_lock &cfs_rq->removed.lock irq_context: 0 &p->pi_lock &rq->__lock &cfs_rq->removed.lock irq_context: softirq rcu_read_lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (pm_chain_head).rwsem irq_context: 0 cpufreq_governor_mutex irq_context: 0 rcu_state.exp_mutex &obj_hash[i].lock irq_context: 0 rcu_state.exp_mutex rcu_read_lock &pool->lock irq_context: 0 rcu_state.exp_mutex rcu_read_lock &pool->lock &obj_hash[i].lock irq_context: 0 rcu_state.exp_mutex rcu_read_lock &pool->lock &p->pi_lock irq_context: 0 rcu_state.exp_mutex rcu_read_lock &pool->lock &p->pi_lock &cfs_rq->removed.lock irq_context: 0 rcu_state.exp_mutex rcu_read_lock &pool->lock &p->pi_lock &rq->__lock irq_context: 0 rcu_state.exp_mutex rcu_read_lock &pool->lock &p->pi_lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 rcu_state.exp_mutex &rnp->exp_wq[2] irq_context: 0 rcu_state.exp_mutex &rq->__lock irq_context: 0 rcu_state.exp_mutex &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 rcu_state.exp_mutex rcu_read_lock &rq->__lock irq_context: 0 &pool->lock &p->pi_lock &cfs_rq->removed.lock irq_context: 0 (wq_completion)rcu_gp (work_completion)(&rew->rew_work) irq_context: 0 (wq_completion)rcu_gp (work_completion)(&rew->rew_work) rcu_node_0 irq_context: 0 (wq_completion)rcu_gp (work_completion)(&rew->rew_work) rcu_state.exp_wake_mutex irq_context: 0 (wq_completion)rcu_gp (work_completion)(&rew->rew_work) rcu_state.exp_wake_mutex rcu_node_0 irq_context: 0 (wq_completion)rcu_gp (work_completion)(&rew->rew_work) rcu_state.exp_wake_mutex &rnp->exp_lock irq_context: 0 (wq_completion)rcu_gp (work_completion)(&rew->rew_work) rcu_state.exp_wake_mutex &rnp->exp_wq[2] irq_context: 0 (wq_completion)rcu_gp (work_completion)(&rew->rew_work) rcu_state.exp_wake_mutex &rnp->exp_wq[2] &p->pi_lock irq_context: 0 (wq_completion)rcu_gp (work_completion)(&rew->rew_work) rcu_state.exp_wake_mutex &rnp->exp_wq[2] &p->pi_lock &rq->__lock irq_context: 0 (wq_completion)rcu_gp (work_completion)(&rew->rew_work) rcu_state.exp_wake_mutex &rnp->exp_wq[2] &p->pi_lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 rcu_state.exp_mutex &rnp->exp_wq[3] irq_context: 0 (wq_completion)rcu_gp (work_completion)(&rew->rew_work) rcu_state.exp_wake_mutex &rnp->exp_wq[3] irq_context: 0 (wq_completion)rcu_gp (work_completion)(&rew->rew_work) rcu_state.exp_wake_mutex &rnp->exp_wq[3] &p->pi_lock irq_context: 0 (wq_completion)rcu_gp (work_completion)(&rew->rew_work) rcu_state.exp_wake_mutex &rnp->exp_wq[3] &p->pi_lock &rq->__lock irq_context: 0 (wq_completion)rcu_gp (work_completion)(&rew->rew_work) rcu_state.exp_wake_mutex &rnp->exp_wq[3] &p->pi_lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)rcu_gp (work_completion)(&rew->rew_work) rcu_state.exp_wake_mutex &rq->__lock irq_context: 0 dyn_event_ops_mutex irq_context: 0 binfmt_lock irq_context: 0 pin_fs_lock irq_context: 0 &type->s_umount_key#7/1 irq_context: 0 &type->s_umount_key#7/1 fs_reclaim irq_context: 0 &type->s_umount_key#7/1 fs_reclaim mmu_notifier_invalidate_range_start irq_context: 0 &type->s_umount_key#7/1 pool_lock#2 irq_context: 0 &type->s_umount_key#7/1 pcpu_alloc_mutex irq_context: 0 &type->s_umount_key#7/1 pcpu_alloc_mutex pcpu_lock irq_context: 0 &type->s_umount_key#7/1 shrinker_rwsem irq_context: 0 &type->s_umount_key#7/1 list_lrus_mutex irq_context: 0 &type->s_umount_key#7/1 sb_lock irq_context: 0 &type->s_umount_key#7/1 sb_lock unnamed_dev_ida.xa_lock irq_context: 0 &type->s_umount_key#7/1 &pcp->lock &zone->lock irq_context: 0 &type->s_umount_key#7/1 &zone->lock irq_context: 0 &type->s_umount_key#7/1 &pcp->lock &zone->lock &____s->seqcount irq_context: 0 &type->s_umount_key#7/1 &____s->seqcount irq_context: 0 &type->s_umount_key#7/1 &c->lock irq_context: 0 &type->s_umount_key#7/1 mmu_notifier_invalidate_range_start irq_context: 0 &type->s_umount_key#7/1 &sb->s_type->i_lock_key#6 irq_context: 0 &type->s_umount_key#7/1 &s->s_inode_list_lock irq_context: 0 &type->s_umount_key#7/1 tk_core.seq.seqcount irq_context: 0 &type->s_umount_key#7/1 &sb->s_type->i_lock_key#6 &dentry->d_lock irq_context: 0 &type->s_umount_key#7/1 &dentry->d_lock irq_context: 0 rcu_read_lock mount_lock irq_context: 0 rcu_read_lock mount_lock mount_lock.seqcount irq_context: 0 &sb->s_type->i_mutex_key#2 irq_context: 0 &sb->s_type->i_mutex_key#2 &sb->s_type->i_lock_key#6 irq_context: 0 &sb->s_type->i_mutex_key#2 rename_lock.seqcount irq_context: 0 &sb->s_type->i_mutex_key#2 fs_reclaim irq_context: 0 &sb->s_type->i_mutex_key#2 fs_reclaim mmu_notifier_invalidate_range_start irq_context: 0 &sb->s_type->i_mutex_key#2 pool_lock#2 irq_context: 0 &sb->s_type->i_mutex_key#2 &dentry->d_lock irq_context: 0 &sb->s_type->i_mutex_key#2 rcu_read_lock rename_lock.seqcount irq_context: 0 &sb->s_type->i_mutex_key#2 &dentry->d_lock &wq irq_context: 0 &sb->s_type->i_mutex_key#2 mmu_notifier_invalidate_range_start irq_context: 0 &sb->s_type->i_mutex_key#2 &s->s_inode_list_lock irq_context: 0 &sb->s_type->i_mutex_key#2 tk_core.seq.seqcount irq_context: 0 &sb->s_type->i_mutex_key#2 &sb->s_type->i_lock_key#6 &dentry->d_lock irq_context: 0 chrdevs_lock irq_context: 0 cb_lock irq_context: 0 cb_lock genl_mutex irq_context: 0 cb_lock genl_mutex fs_reclaim irq_context: 0 cb_lock genl_mutex fs_reclaim mmu_notifier_invalidate_range_start irq_context: 0 cb_lock genl_mutex pool_lock#2 irq_context: 0 &type->s_umount_key#8/1 irq_context: 0 &type->s_umount_key#8/1 fs_reclaim irq_context: 0 &type->s_umount_key#8/1 fs_reclaim mmu_notifier_invalidate_range_start irq_context: 0 &type->s_umount_key#8/1 pool_lock#2 irq_context: 0 &type->s_umount_key#8/1 pcpu_alloc_mutex irq_context: 0 &type->s_umount_key#8/1 pcpu_alloc_mutex pcpu_lock irq_context: 0 &type->s_umount_key#8/1 shrinker_rwsem irq_context: 0 &type->s_umount_key#8/1 list_lrus_mutex irq_context: 0 &type->s_umount_key#8/1 sb_lock irq_context: 0 &type->s_umount_key#8/1 sb_lock unnamed_dev_ida.xa_lock irq_context: 0 &type->s_umount_key#8/1 mmu_notifier_invalidate_range_start irq_context: 0 &type->s_umount_key#8/1 &sb->s_type->i_lock_key#7 irq_context: 0 &type->s_umount_key#8/1 &s->s_inode_list_lock irq_context: 0 &type->s_umount_key#8/1 tk_core.seq.seqcount irq_context: 0 &type->s_umount_key#8/1 &sb->s_type->i_lock_key#7 &dentry->d_lock irq_context: 0 &type->s_umount_key#8/1 &dentry->d_lock irq_context: 0 &sb->s_type->i_mutex_key#3 irq_context: 0 &sb->s_type->i_mutex_key#3 &sb->s_type->i_lock_key#7 irq_context: 0 &sb->s_type->i_mutex_key#3 rename_lock.seqcount irq_context: 0 &sb->s_type->i_mutex_key#3 fs_reclaim irq_context: 0 &sb->s_type->i_mutex_key#3 fs_reclaim mmu_notifier_invalidate_range_start irq_context: 0 &sb->s_type->i_mutex_key#3 pool_lock#2 irq_context: 0 &sb->s_type->i_mutex_key#3 &dentry->d_lock irq_context: 0 &sb->s_type->i_mutex_key#3 rcu_read_lock rename_lock.seqcount irq_context: 0 &sb->s_type->i_mutex_key#3 &dentry->d_lock &wq irq_context: 0 &sb->s_type->i_mutex_key#3 mmu_notifier_invalidate_range_start irq_context: 0 &sb->s_type->i_mutex_key#3 &s->s_inode_list_lock irq_context: 0 &sb->s_type->i_mutex_key#3 tk_core.seq.seqcount irq_context: 0 &sb->s_type->i_mutex_key#3 &sb->s_type->i_lock_key#7 &dentry->d_lock irq_context: 0 subsys mutex#3 irq_context: 0 async_lock irq_context: 0 rcu_read_lock &pool->lock/1 pool_lock#2 irq_context: 0 rcu_read_lock &pool->lock/1 &p->pi_lock irq_context: 0 (wq_completion)events_unbound (work_completion)(&entry->work) irq_context: 0 (wq_completion)events_unbound (work_completion)(&entry->work) tk_core.seq.seqcount irq_context: 0 (wq_completion)events_unbound (work_completion)(&entry->work) &dev->mutex device_links_srcu irq_context: 0 (wq_completion)events_unbound (work_completion)(&entry->work) &dev->mutex &dev->power.lock irq_context: 0 (wq_completion)events_unbound (work_completion)(&entry->work) &dev->mutex fwnode_link_lock irq_context: 0 (wq_completion)events_unbound (work_completion)(&entry->work) &dev->mutex device_links_lock irq_context: 0 regulator_list_mutex irq_context: 0 (wq_completion)events_unbound (work_completion)(&entry->work) &dev->mutex &k->list_lock irq_context: 0 (wq_completion)events_unbound (work_completion)(&entry->work) &dev->mutex sysfs_symlink_target_lock irq_context: 0 (wq_completion)events_unbound (work_completion)(&entry->work) &dev->mutex fs_reclaim irq_context: 0 (wq_completion)events_unbound (work_completion)(&entry->work) &dev->mutex fs_reclaim mmu_notifier_invalidate_range_start irq_context: 0 (wq_completion)events_unbound (work_completion)(&entry->work) &dev->mutex pool_lock#2 irq_context: 0 (wq_completion)events_unbound (work_completion)(&entry->work) &dev->mutex lock irq_context: 0 (wq_completion)events_unbound (work_completion)(&entry->work) &dev->mutex lock kernfs_idr_lock irq_context: 0 (wq_completion)events_unbound (work_completion)(&entry->work) &dev->mutex &root->kernfs_rwsem irq_context: 0 rtc_lock irq_context: 0 (wq_completion)events_unbound (work_completion)(&entry->work) &dev->mutex &root->kernfs_rwsem &root->kernfs_iattr_rwsem irq_context: 0 (wq_completion)events_unbound (work_completion)(&entry->work) &dev->mutex &x->wait#9 irq_context: 0 (wq_completion)events_unbound (work_completion)(&entry->work) &dev->mutex &obj_hash[i].lock irq_context: 0 (wq_completion)events_unbound (work_completion)(&entry->work) &dev->mutex &dev->devres_lock irq_context: 0 (wq_completion)events_unbound (work_completion)(&entry->work) &dev->mutex regulator_nesting_mutex irq_context: 0 (wq_completion)events_unbound (work_completion)(&entry->work) &dev->mutex regulator_ww_class_mutex regulator_nesting_mutex irq_context: 0 (wq_completion)events_unbound (work_completion)(&entry->work) &dev->mutex devtree_lock irq_context: 0 (wq_completion)events_unbound (work_completion)(&entry->work) &dev->mutex gdp_mutex irq_context: 0 (wq_completion)events_unbound (work_completion)(&entry->work) &dev->mutex gdp_mutex &k->list_lock irq_context: 0 (wq_completion)events_unbound (work_completion)(&entry->work) &dev->mutex gdp_mutex fs_reclaim irq_context: 0 (wq_completion)events_unbound (work_completion)(&entry->work) &dev->mutex gdp_mutex fs_reclaim mmu_notifier_invalidate_range_start irq_context: 0 (wq_completion)events_unbound (work_completion)(&entry->work) &dev->mutex gdp_mutex pool_lock#2 irq_context: 0 (wq_completion)events_unbound (work_completion)(&entry->work) &dev->mutex gdp_mutex lock irq_context: 0 (wq_completion)events_unbound (work_completion)(&entry->work) &dev->mutex gdp_mutex lock kernfs_idr_lock irq_context: 0 (wq_completion)events_unbound (work_completion)(&entry->work) &dev->mutex gdp_mutex &root->kernfs_rwsem irq_context: 0 (wq_completion)events_unbound (work_completion)(&entry->work) &dev->mutex gdp_mutex &root->kernfs_rwsem &root->kernfs_iattr_rwsem irq_context: 0 (wq_completion)events_unbound (work_completion)(&entry->work) &dev->mutex bus_type_sem irq_context: 0 (wq_completion)events_unbound (work_completion)(&entry->work) &dev->mutex &c->lock irq_context: 0 (wq_completion)events_unbound (work_completion)(&entry->work) &dev->mutex &pcp->lock &zone->lock irq_context: 0 (wq_completion)events_unbound (work_completion)(&entry->work) &dev->mutex &zone->lock irq_context: 0 (wq_completion)events_unbound (work_completion)(&entry->work) &dev->mutex &____s->seqcount irq_context: 0 (wq_completion)events_unbound (work_completion)(&entry->work) &dev->mutex &root->kernfs_rwsem irq_context: 0 (wq_completion)events_unbound (work_completion)(&entry->work) &dev->mutex dpm_list_mtx irq_context: 0 (wq_completion)events_unbound (work_completion)(&entry->work) &dev->mutex uevent_sock_mutex irq_context: 0 (wq_completion)events_unbound (work_completion)(&entry->work) &dev->mutex running_helpers_waitq.lock irq_context: 0 (wq_completion)events_unbound (work_completion)(&entry->work) &dev->mutex &k->k_lock irq_context: 0 (wq_completion)events_unbound (work_completion)(&entry->work) &dev->mutex subsys mutex#4 irq_context: 0 (wq_completion)events_unbound (work_completion)(&entry->work) &dev->mutex subsys mutex#4 &k->k_lock irq_context: 0 (wq_completion)events_unbound (work_completion)(&entry->work) &dev->mutex pin_fs_lock irq_context: 0 (wq_completion)events_unbound (work_completion)(&entry->work) &dev->mutex &sb->s_type->i_mutex_key#3 irq_context: 0 (wq_completion)events_unbound (work_completion)(&entry->work) &dev->mutex &sb->s_type->i_mutex_key#3 rename_lock.seqcount irq_context: 0 (wq_completion)events_unbound (work_completion)(&entry->work) &dev->mutex &sb->s_type->i_mutex_key#3 fs_reclaim irq_context: 0 (wq_completion)events_unbound (work_completion)(&entry->work) &dev->mutex &sb->s_type->i_mutex_key#3 fs_reclaim mmu_notifier_invalidate_range_start irq_context: 0 (wq_completion)events_unbound (work_completion)(&entry->work) &dev->mutex &sb->s_type->i_mutex_key#3 pool_lock#2 irq_context: 0 (wq_completion)events_unbound (work_completion)(&entry->work) &dev->mutex &sb->s_type->i_mutex_key#3 &dentry->d_lock irq_context: 0 (wq_completion)events_unbound (work_completion)(&entry->work) &dev->mutex &sb->s_type->i_mutex_key#3 rcu_read_lock rename_lock.seqcount irq_context: 0 (wq_completion)events_unbound (work_completion)(&entry->work) &dev->mutex &sb->s_type->i_mutex_key#3 &dentry->d_lock &wq irq_context: 0 (wq_completion)events_unbound (work_completion)(&entry->work) &dev->mutex &sb->s_type->i_mutex_key#3 mmu_notifier_invalidate_range_start irq_context: 0 (wq_completion)events_unbound (work_completion)(&entry->work) &dev->mutex &sb->s_type->i_mutex_key#3 &sb->s_type->i_lock_key#7 irq_context: 0 (wq_completion)events_unbound (work_completion)(&entry->work) &dev->mutex &sb->s_type->i_mutex_key#3 &s->s_inode_list_lock irq_context: 0 (wq_completion)events_unbound (work_completion)(&entry->work) &dev->mutex &sb->s_type->i_mutex_key#3 tk_core.seq.seqcount irq_context: 0 (wq_completion)events_unbound (work_completion)(&entry->work) &dev->mutex &sb->s_type->i_mutex_key#3 &sb->s_type->i_lock_key#7 &dentry->d_lock irq_context: 0 (wq_completion)events_unbound (work_completion)(&entry->work) &dev->mutex regulator_list_mutex irq_context: 0 (wq_completion)events_unbound (work_completion)(&entry->work) &dev->mutex deferred_probe_mutex irq_context: 0 (wq_completion)events_unbound (work_completion)(&entry->work) &dev->mutex probe_waitqueue.lock irq_context: 0 (wq_completion)events_unbound (work_completion)(&entry->work) async_lock irq_context: 0 (wq_completion)events_unbound (work_completion)(&entry->work) async_lock &obj_hash[i].lock irq_context: 0 (wq_completion)events_unbound (work_completion)(&entry->work) async_lock pool_lock#2 irq_context: 0 (wq_completion)events_unbound (work_completion)(&entry->work) async_done.lock irq_context: 0 &type->s_umount_key#9/1 irq_context: 0 &type->s_umount_key#9/1 fs_reclaim irq_context: 0 &type->s_umount_key#9/1 fs_reclaim mmu_notifier_invalidate_range_start irq_context: 0 &type->s_umount_key#9/1 pool_lock#2 irq_context: 0 &type->s_umount_key#9/1 pcpu_alloc_mutex irq_context: 0 &type->s_umount_key#9/1 pcpu_alloc_mutex pcpu_lock irq_context: 0 &type->s_umount_key#9/1 shrinker_rwsem irq_context: 0 &type->s_umount_key#9/1 list_lrus_mutex irq_context: 0 &type->s_umount_key#9/1 sb_lock irq_context: 0 &type->s_umount_key#9/1 sb_lock unnamed_dev_ida.xa_lock irq_context: 0 &type->s_umount_key#9/1 &pcp->lock &zone->lock irq_context: 0 &type->s_umount_key#9/1 &zone->lock irq_context: 0 &type->s_umount_key#9/1 &pcp->lock &zone->lock &____s->seqcount irq_context: 0 &type->s_umount_key#9/1 &____s->seqcount irq_context: 0 &type->s_umount_key#9/1 &c->lock irq_context: 0 &type->s_umount_key#9/1 mmu_notifier_invalidate_range_start irq_context: 0 &type->s_umount_key#9/1 &sb->s_type->i_lock_key#8 irq_context: 0 &type->s_umount_key#9/1 &s->s_inode_list_lock irq_context: 0 &type->s_umount_key#9/1 tk_core.seq.seqcount irq_context: 0 &type->s_umount_key#9/1 &sb->s_type->i_lock_key#8 &dentry->d_lock irq_context: 0 &type->s_umount_key#9/1 &dentry->d_lock irq_context: 0 pernet_ops_rwsem fs_reclaim irq_context: 0 pernet_ops_rwsem fs_reclaim mmu_notifier_invalidate_range_start irq_context: 0 pernet_ops_rwsem &pcp->lock &zone->lock irq_context: 0 pernet_ops_rwsem &zone->lock irq_context: 0 pernet_ops_rwsem &____s->seqcount irq_context: 0 pernet_ops_rwsem &c->lock irq_context: 0 pernet_ops_rwsem sysctl_lock irq_context: 0 pack_mutex irq_context: 0 pack_mutex fs_reclaim irq_context: 0 pack_mutex fs_reclaim mmu_notifier_invalidate_range_start irq_context: 0 pack_mutex &pcp->lock &zone->lock irq_context: 0 pack_mutex &zone->lock irq_context: 0 pack_mutex &____s->seqcount irq_context: 0 pack_mutex pool_lock#2 irq_context: 0 pack_mutex free_vmap_area_lock irq_context: 0 pack_mutex vmap_area_lock irq_context: 0 pack_mutex init_mm.page_table_lock irq_context: 0 pack_mutex &pcp->lock &zone->lock &____s->seqcount irq_context: 0 pack_mutex rcu_node_0 irq_context: 0 pack_mutex &rq->__lock irq_context: 0 pack_mutex &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 pack_mutex rcu_read_lock rcu_node_0 irq_context: 0 pack_mutex vmap_purge_lock irq_context: 0 pack_mutex vmap_purge_lock purge_vmap_area_lock irq_context: 0 pack_mutex cpa_lock irq_context: 0 pack_mutex cpa_lock pgd_lock irq_context: 0 text_mutex irq_context: 0 text_mutex ptlock_ptr(page)#2 irq_context: 0 &fp->aux->used_maps_mutex irq_context: 0 pernet_ops_rwsem pcpu_alloc_mutex irq_context: 0 pernet_ops_rwsem pcpu_alloc_mutex pcpu_lock irq_context: 0 proto_list_mutex irq_context: 0 targets_mutex irq_context: 0 nl_table_lock irq_context: 0 nl_table_wait.lock irq_context: 0 net_family_lock irq_context: 0 pernet_ops_rwsem net_generic_ids.xa_lock irq_context: 0 pernet_ops_rwsem mmu_notifier_invalidate_range_start irq_context: 0 pernet_ops_rwsem &sb->s_type->i_lock_key#8 irq_context: 0 pernet_ops_rwsem &dir->lock irq_context: 0 pernet_ops_rwsem &obj_hash[i].lock irq_context: 0 pernet_ops_rwsem k-sk_lock-AF_NETLINK irq_context: 0 pernet_ops_rwsem k-sk_lock-AF_NETLINK k-slock-AF_NETLINK irq_context: 0 pernet_ops_rwsem k-sk_lock-AF_NETLINK rcu_read_lock rhashtable_bucket irq_context: 0 pernet_ops_rwsem k-slock-AF_NETLINK irq_context: 0 pernet_ops_rwsem nl_table_lock irq_context: 0 pernet_ops_rwsem nl_table_wait.lock irq_context: 0 pernet_ops_rwsem rtnl_mutex irq_context: 0 rtnl_mutex fs_reclaim irq_context: 0 rtnl_mutex fs_reclaim mmu_notifier_invalidate_range_start irq_context: 0 sparse_irq_lock irq_context: 0 sparse_irq_lock fs_reclaim irq_context: 0 sparse_irq_lock fs_reclaim mmu_notifier_invalidate_range_start irq_context: 0 sparse_irq_lock pool_lock#2 irq_context: 0 sparse_irq_lock lock irq_context: 0 sparse_irq_lock lock kernfs_idr_lock irq_context: 0 sparse_irq_lock &root->kernfs_rwsem irq_context: 0 sparse_irq_lock &root->kernfs_rwsem &root->kernfs_iattr_rwsem irq_context: 0 sparse_irq_lock &c->lock irq_context: 0 sparse_irq_lock &____s->seqcount irq_context: 0 sparse_irq_lock lock kernfs_idr_lock pool_lock#2 irq_context: 0 sparse_irq_lock &pcp->lock &zone->lock irq_context: 0 sparse_irq_lock &zone->lock irq_context: 0 &list->lock irq_context: 0 kauditd_wait.lock irq_context: 0 lock#2 irq_context: 0 lock#2 &zone->lock irq_context: 0 pcp_batch_high_lock irq_context: 0 khugepaged_mutex irq_context: 0 gdp_mutex irq_context: 0 gdp_mutex &k->list_lock irq_context: 0 gdp_mutex fs_reclaim irq_context: 0 gdp_mutex fs_reclaim mmu_notifier_invalidate_range_start irq_context: 0 gdp_mutex pool_lock#2 irq_context: 0 gdp_mutex lock irq_context: 0 gdp_mutex lock kernfs_idr_lock irq_context: 0 gdp_mutex &root->kernfs_rwsem irq_context: 0 gdp_mutex &root->kernfs_rwsem &root->kernfs_iattr_rwsem irq_context: 0 subsys mutex#5 irq_context: 0 subsys mutex#5 &k->k_lock irq_context: 0 subsys mutex#6 irq_context: 0 subsys mutex#6 &k->list_lock irq_context: 0 subsys mutex#6 &k->k_lock irq_context: 0 regmap_debugfs_early_lock irq_context: 0 (acpi_reconfig_chain).rwsem irq_context: 0 __i2c_board_lock irq_context: 0 core_lock irq_context: 0 core_lock &k->list_lock irq_context: 0 core_lock &k->k_lock irq_context: 0 cb_lock genl_mutex nl_table_lock irq_context: 0 cb_lock genl_mutex nl_table_wait.lock irq_context: 0 nl_table_lock irq_context: 0 thermal_governor_lock irq_context: 0 thermal_governor_lock thermal_list_lock irq_context: softirq &(&kfence_timer)->timer rcu_read_lock &pool->lock/1 &p->pi_lock &rq->__lock &cfs_rq->removed.lock irq_context: 0 batched_entropy_u8.lock crngs.lock irq_context: 0 rcu_tasks.tasks_gp_mutex rcu_state.exp_mutex &obj_hash[i].lock irq_context: 0 cpuidle_lock irq_context: 0 rcu_tasks.tasks_gp_mutex rcu_state.exp_mutex rcu_read_lock &pool->lock irq_context: 0 cpuidle_lock &rnp->exp_lock irq_context: 0 rcu_tasks.tasks_gp_mutex rcu_state.exp_mutex rcu_read_lock &pool->lock &obj_hash[i].lock irq_context: 0 cpuidle_lock rcu_state.exp_mutex irq_context: 0 rcu_tasks.tasks_gp_mutex rcu_state.exp_mutex rcu_read_lock &pool->lock &p->pi_lock irq_context: 0 rcu_tasks.tasks_gp_mutex rcu_state.exp_mutex rcu_read_lock &pool->lock &p->pi_lock &rq->__lock irq_context: 0 rcu_tasks.tasks_gp_mutex rcu_state.exp_mutex rcu_read_lock &pool->lock &p->pi_lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 rcu_tasks.tasks_gp_mutex rcu_state.exp_mutex &rnp->exp_wq[0] irq_context: 0 rcu_tasks.tasks_gp_mutex rcu_state.exp_mutex &rq->__lock irq_context: 0 cpuidle_lock rcu_state.exp_mutex rcu_state.exp_mutex.wait_lock irq_context: 0 (wq_completion)rcu_gp (work_completion)(&rew->rew_work) &rcu_state.expedited_wq irq_context: 0 cpuidle_lock rcu_state.exp_mutex rcu_node_0 irq_context: 0 cpuidle_lock rcu_state.exp_mutex &rcu_state.expedited_wq irq_context: 0 (wq_completion)rcu_gp (work_completion)(&rew->rew_work) &obj_hash[i].lock irq_context: 0 cpuidle_lock rcu_state.exp_mutex &rcu_state.expedited_wq &p->pi_lock irq_context: 0 (wq_completion)rcu_gp (work_completion)(&rew->rew_work) &base->lock irq_context: 0 cpuidle_lock rcu_state.exp_mutex &rcu_state.expedited_wq &p->pi_lock &rq->__lock irq_context: 0 (wq_completion)rcu_gp (work_completion)(&rew->rew_work) &base->lock &obj_hash[i].lock irq_context: 0 cpuidle_lock rcu_state.exp_mutex &rq->__lock irq_context: 0 (wq_completion)rcu_gp (work_completion)(&rew->rew_work) &rq->__lock irq_context: 0 cpuidle_lock rcu_state.exp_mutex &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)rcu_gp (work_completion)(&rew->rew_work) &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)rcu_gp (work_completion)(&rew->rew_work) (&timer.timer) irq_context: 0 freezer_lock irq_context: 0 (wq_completion)rcu_gp (work_completion)(&rew->rew_work) rcu_state.exp_wake_mutex &rnp->exp_wq[0] irq_context: 0 (wq_completion)rcu_gp (work_completion)(&rew->rew_work) rcu_state.exp_wake_mutex &rnp->exp_wq[0] &p->pi_lock irq_context: 0 (wq_completion)rcu_gp (work_completion)(&rew->rew_work) rcu_state.exp_wake_mutex &rnp->exp_wq[0] &p->pi_lock &rq->__lock irq_context: 0 (wq_completion)rcu_gp (work_completion)(&rew->rew_work) rcu_state.exp_wake_mutex &rnp->exp_wq[0] &p->pi_lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 rcu_tasks.tasks_gp_mutex rcu_state.exp_mutex.wait_lock irq_context: 0 rcu_tasks.tasks_gp_mutex &p->pi_lock irq_context: 0 rcu_tasks.tasks_gp_mutex &p->pi_lock &rq->__lock irq_context: 0 rcu_tasks.tasks_gp_mutex &p->pi_lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)rcu_gp (work_completion)(&(&ssp->srcu_sup->work)->work) rcu_read_lock &pool->lock irq_context: 0 (wq_completion)rcu_gp (work_completion)(&(&ssp->srcu_sup->work)->work) rcu_read_lock &pool->lock &obj_hash[i].lock irq_context: 0 audit_backlog_wait.lock irq_context: 0 cpuidle_lock rcu_state.exp_mutex &obj_hash[i].lock irq_context: 0 cpuidle_lock rcu_state.exp_mutex rcu_read_lock &pool->lock irq_context: 0 cpuidle_lock rcu_state.exp_mutex rcu_read_lock &pool->lock &obj_hash[i].lock irq_context: 0 cpuidle_lock rcu_state.exp_mutex rcu_read_lock &pool->lock &p->pi_lock irq_context: 0 cpuidle_lock rcu_state.exp_mutex rcu_read_lock &pool->lock &p->pi_lock &rq->__lock irq_context: 0 cpuidle_lock rcu_state.exp_mutex rcu_read_lock &pool->lock &p->pi_lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 cpuidle_lock rcu_state.exp_mutex &rnp->exp_wq[1] irq_context: 0 (wq_completion)rcu_gp (work_completion)(&rew->rew_work) &pool->lock irq_context: hardirq &rcu_state.expedited_wq irq_context: hardirq &rcu_state.expedited_wq &p->pi_lock irq_context: hardirq &rcu_state.expedited_wq &p->pi_lock &rq->__lock irq_context: hardirq &rcu_state.expedited_wq &p->pi_lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 cpuidle_lock rcu_state.exp_mutex rcu_read_lock &rq->__lock irq_context: 0 cpuidle_lock rcu_state.exp_mutex rcu_read_lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)rcu_gp (work_completion)(&rew->rew_work) rcu_state.exp_wake_mutex &rnp->exp_wq[1] irq_context: 0 (wq_completion)rcu_gp (work_completion)(&rew->rew_work) rcu_state.exp_wake_mutex &rnp->exp_wq[1] &p->pi_lock irq_context: 0 (wq_completion)rcu_gp (work_completion)(&rew->rew_work) rcu_state.exp_wake_mutex &rnp->exp_wq[1] &p->pi_lock &rq->__lock irq_context: 0 (wq_completion)rcu_gp (work_completion)(&rew->rew_work) rcu_state.exp_wake_mutex &rnp->exp_wq[1] &p->pi_lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 cpuidle_lock &obj_hash[i].lock irq_context: 0 cpuidle_lock (console_sem).lock irq_context: 0 cpuidle_lock console_lock console_srcu console_owner_lock irq_context: 0 cpuidle_lock console_lock console_srcu console_owner irq_context: 0 cpuidle_lock console_lock console_srcu console_owner &port_lock_key irq_context: 0 cpuidle_lock console_lock console_srcu console_owner console_owner_lock irq_context: 0 mmu_notifier_invalidate_range_start irq_context: 0 &sb->s_type->i_lock_key#8 irq_context: 0 &dir->lock irq_context: 0 k-sk_lock-AF_QIPCRTR irq_context: 0 k-sk_lock-AF_QIPCRTR k-slock-AF_QIPCRTR irq_context: 0 k-slock-AF_QIPCRTR irq_context: 0 k-sk_lock-AF_QIPCRTR fs_reclaim irq_context: 0 k-sk_lock-AF_QIPCRTR fs_reclaim mmu_notifier_invalidate_range_start irq_context: 0 k-sk_lock-AF_QIPCRTR qrtr_ports.xa_lock irq_context: 0 k-sk_lock-AF_QIPCRTR pool_lock#2 irq_context: 0 k-sk_lock-AF_QIPCRTR qrtr_node_lock irq_context: 0 k-sk_lock-AF_QIPCRTR &obj_hash[i].lock irq_context: 0 pernet_ops_rwsem uevent_sock_mutex irq_context: 0 (wq_completion)rcu_gp (work_completion)(&(&ssp->srcu_sup->work)->work) &obj_hash[i].lock irq_context: 0 (wq_completion)rcu_gp (work_completion)(&(&ssp->srcu_sup->work)->work) &base->lock irq_context: 0 (wq_completion)rcu_gp (work_completion)(&(&ssp->srcu_sup->work)->work) &base->lock &obj_hash[i].lock irq_context: 0 &sb->s_type->i_mutex_key#3 &c->lock irq_context: 0 &sb->s_type->i_mutex_key#3 &____s->seqcount irq_context: 0 cpu_hotplug_lock jump_label_mutex text_mutex ptlock_ptr(page)#2 irq_context: 0 crngs.lock irq_context: 0 (crypto_chain).rwsem irq_context: 0 iova_cache_mutex irq_context: 0 iova_cache_mutex cpu_hotplug_lock irq_context: 0 iova_cache_mutex cpu_hotplug_lock cpuhp_state_mutex irq_context: 0 iova_cache_mutex slab_mutex irq_context: 0 iova_cache_mutex slab_mutex fs_reclaim irq_context: 0 iova_cache_mutex slab_mutex fs_reclaim mmu_notifier_invalidate_range_start irq_context: 0 iova_cache_mutex slab_mutex pool_lock#2 irq_context: 0 iova_cache_mutex slab_mutex &c->lock irq_context: 0 iova_cache_mutex slab_mutex &n->list_lock irq_context: 0 iova_cache_mutex slab_mutex pcpu_alloc_mutex irq_context: 0 iova_cache_mutex slab_mutex pcpu_alloc_mutex pcpu_lock irq_context: 0 subsys mutex#7 irq_context: 0 subsys mutex#7 &k->k_lock irq_context: 0 pci_config_lock irq_context: 0 device_links_lock irq_context: 0 subsys mutex#8 irq_context: 0 dev_pm_qos_mtx irq_context: 0 dev_pm_qos_mtx fs_reclaim irq_context: 0 dev_pm_qos_mtx fs_reclaim mmu_notifier_invalidate_range_start irq_context: 0 dev_pm_qos_mtx pool_lock#2 irq_context: 0 dev_pm_qos_mtx &dev->power.lock irq_context: 0 dev_pm_qos_mtx pm_qos_lock irq_context: 0 dev_pm_qos_sysfs_mtx irq_context: 0 dev_pm_qos_sysfs_mtx dev_pm_qos_mtx irq_context: 0 dev_pm_qos_sysfs_mtx &root->kernfs_rwsem irq_context: 0 dev_pm_qos_sysfs_mtx fs_reclaim irq_context: 0 dev_pm_qos_sysfs_mtx fs_reclaim mmu_notifier_invalidate_range_start irq_context: 0 dev_pm_qos_sysfs_mtx pool_lock#2 irq_context: 0 dev_pm_qos_sysfs_mtx lock irq_context: 0 dev_pm_qos_sysfs_mtx lock kernfs_idr_lock irq_context: 0 dev_pm_qos_sysfs_mtx &root->kernfs_rwsem irq_context: 0 dev_pm_qos_sysfs_mtx &root->kernfs_rwsem &root->kernfs_iattr_rwsem irq_context: 0 dev_pm_qos_sysfs_mtx &c->lock irq_context: 0 dev_pm_qos_sysfs_mtx &____s->seqcount irq_context: 0 mtrr_mutex irq_context: 0 mtrr_mutex fs_reclaim irq_context: 0 mtrr_mutex fs_reclaim mmu_notifier_invalidate_range_start irq_context: 0 mtrr_mutex pool_lock#2 irq_context: 0 uidhash_lock irq_context: 0 cpu_hotplug_lock wq_pool_mutex (console_sem).lock irq_context: 0 cpu_hotplug_lock wq_pool_mutex console_lock console_srcu console_owner_lock irq_context: 0 cpu_hotplug_lock wq_pool_mutex console_lock console_srcu console_owner irq_context: 0 cpu_hotplug_lock wq_pool_mutex console_lock console_srcu console_owner &port_lock_key irq_context: 0 cpu_hotplug_lock wq_pool_mutex console_lock console_srcu console_owner console_owner_lock irq_context: 0 &type->s_umount_key#28 &sb->s_type->i_lock_key#3 irq_context: 0 &type->s_umount_key#28 &sb->s_type->i_lock_key#3 &xa->xa_lock#9 irq_context: 0 &type->s_umount_key#28 lock#4 &lruvec->lru_lock irq_context: 0 oom_reaper_wait.lock irq_context: 0 subsys mutex#9 irq_context: softirq &(&ssp->srcu_sup->work)->timer irq_context: softirq &(&ssp->srcu_sup->work)->timer rcu_read_lock &pool->lock irq_context: softirq &(&ssp->srcu_sup->work)->timer rcu_read_lock &pool->lock &obj_hash[i].lock irq_context: softirq &(&ssp->srcu_sup->work)->timer rcu_read_lock &pool->lock &p->pi_lock irq_context: softirq &(&ssp->srcu_sup->work)->timer rcu_read_lock &pool->lock &p->pi_lock &rq->__lock irq_context: softirq &(&ssp->srcu_sup->work)->timer rcu_read_lock &pool->lock &p->pi_lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 &type->s_umount_key#28 lock#5 irq_context: 0 &x->wait &p->pi_lock &cfs_rq->removed.lock irq_context: 0 &pgdat->kcompactd_wait irq_context: 0 (wq_completion)rcu_gp (work_completion)(&sdp->work) &x->wait#2 &p->pi_lock &rq->__lock &cfs_rq->removed.lock irq_context: 0 slab_mutex rcu_read_lock &pool->lock irq_context: 0 slab_mutex rcu_read_lock &pool->lock &obj_hash[i].lock irq_context: 0 slab_mutex rcu_read_lock &pool->lock &p->pi_lock irq_context: 0 slab_mutex rcu_read_lock &pool->lock &p->pi_lock &rq->__lock irq_context: 0 slab_mutex rcu_read_lock &pool->lock &p->pi_lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 slab_mutex &rq->__lock irq_context: 0 (wq_completion)events pcpu_balance_work irq_context: 0 (wq_completion)events pcpu_balance_work pcpu_alloc_mutex irq_context: 0 (wq_completion)events pcpu_balance_work pcpu_alloc_mutex pcpu_lock irq_context: 0 (wq_completion)events pcpu_balance_work pcpu_alloc_mutex fs_reclaim irq_context: 0 (wq_completion)events pcpu_balance_work pcpu_alloc_mutex fs_reclaim mmu_notifier_invalidate_range_start irq_context: 0 (wq_completion)events pcpu_balance_work pcpu_alloc_mutex pool_lock#2 irq_context: 0 (wq_completion)events pcpu_balance_work pcpu_alloc_mutex free_vmap_area_lock irq_context: 0 (wq_completion)events pcpu_balance_work pcpu_alloc_mutex vmap_area_lock irq_context: 0 (wq_completion)events pcpu_balance_work pcpu_alloc_mutex &pcp->lock &zone->lock irq_context: 0 (wq_completion)events pcpu_balance_work pcpu_alloc_mutex &zone->lock irq_context: 0 (wq_completion)events pcpu_balance_work pcpu_alloc_mutex &____s->seqcount irq_context: 0 (wq_completion)events pcpu_balance_work pcpu_alloc_mutex init_mm.page_table_lock irq_context: 0 (wq_completion)events pcpu_balance_work pcpu_alloc_mutex &pcp->lock &zone->lock &____s->seqcount irq_context: 0 (wq_completion)events pcpu_balance_work pcpu_alloc_mutex &rq->__lock irq_context: 0 (wq_completion)events pcpu_balance_work pcpu_alloc_mutex &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 slab_mutex pcpu_alloc_mutex pcpu_alloc_mutex.wait_lock irq_context: 0 slab_mutex pcpu_alloc_mutex &rq->__lock irq_context: 0 slab_mutex pcpu_alloc_mutex &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)events pcpu_balance_work pcpu_alloc_mutex &c->lock irq_context: 0 (wq_completion)events pcpu_balance_work pcpu_alloc_mutex &obj_hash[i].lock irq_context: 0 (wq_completion)events pcpu_balance_work pcpu_alloc_mutex.wait_lock irq_context: 0 (wq_completion)events pcpu_balance_work &p->pi_lock irq_context: 0 (wq_completion)events pcpu_balance_work &p->pi_lock &rq->__lock irq_context: 0 (wq_completion)events pcpu_balance_work &p->pi_lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)events pcpu_balance_work &rq->__lock irq_context: 0 memory_tier_lock irq_context: 0 memory_tier_lock fs_reclaim irq_context: 0 memory_tier_lock fs_reclaim mmu_notifier_invalidate_range_start irq_context: 0 memory_tier_lock pool_lock#2 irq_context: 0 memory_tier_lock &x->wait#9 irq_context: 0 memory_tier_lock &obj_hash[i].lock irq_context: 0 memory_tier_lock &obj_hash[i].lock pool_lock irq_context: 0 memory_tier_lock &k->list_lock irq_context: 0 memory_tier_lock lock irq_context: 0 memory_tier_lock lock kernfs_idr_lock irq_context: 0 memory_tier_lock &root->kernfs_rwsem irq_context: 0 memory_tier_lock &root->kernfs_rwsem &root->kernfs_iattr_rwsem irq_context: 0 memory_tier_lock bus_type_sem irq_context: 0 memory_tier_lock sysfs_symlink_target_lock irq_context: 0 memory_tier_lock &k->k_lock irq_context: 0 memory_tier_lock &root->kernfs_rwsem irq_context: 0 memory_tier_lock &c->lock irq_context: 0 memory_tier_lock &____s->seqcount irq_context: 0 memory_tier_lock &dev->power.lock irq_context: 0 memory_tier_lock dpm_list_mtx irq_context: 0 memory_tier_lock uevent_sock_mutex irq_context: 0 memory_tier_lock running_helpers_waitq.lock irq_context: 0 memory_tier_lock &pcp->lock &zone->lock irq_context: 0 memory_tier_lock &zone->lock irq_context: 0 memory_tier_lock &pcp->lock &zone->lock &____s->seqcount irq_context: 0 memory_tier_lock rcu_read_lock pool_lock#2 irq_context: 0 memory_tier_lock &dev->mutex &k->list_lock irq_context: 0 memory_tier_lock &dev->mutex &k->k_lock irq_context: 0 memory_tier_lock &dev->mutex &dev->power.lock irq_context: 0 memory_tier_lock subsys mutex#10 irq_context: 0 memory_tier_lock rcu_state.exp_mutex rcu_node_0 irq_context: 0 memory_tier_lock rcu_state.exp_mutex &obj_hash[i].lock irq_context: 0 memory_tier_lock rcu_state.exp_mutex rcu_read_lock &pool->lock irq_context: 0 memory_tier_lock rcu_state.exp_mutex rcu_read_lock &pool->lock &obj_hash[i].lock irq_context: 0 memory_tier_lock rcu_state.exp_mutex rcu_read_lock &pool->lock &p->pi_lock irq_context: 0 memory_tier_lock rcu_state.exp_mutex rcu_read_lock &pool->lock &p->pi_lock &rq->__lock irq_context: 0 memory_tier_lock rcu_state.exp_mutex rcu_read_lock &pool->lock &p->pi_lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 memory_tier_lock rcu_state.exp_mutex &rnp->exp_wq[2] irq_context: 0 memory_tier_lock rcu_state.exp_mutex &rq->__lock irq_context: 0 memory_tier_lock rcu_state.exp_mutex &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 khugepaged_mutex fs_reclaim irq_context: 0 khugepaged_mutex fs_reclaim mmu_notifier_invalidate_range_start irq_context: 0 khugepaged_mutex pool_lock#2 irq_context: 0 khugepaged_mutex kthread_create_lock irq_context: 0 khugepaged_mutex &p->pi_lock irq_context: 0 khugepaged_mutex &p->pi_lock &rq->__lock irq_context: 0 khugepaged_mutex &p->pi_lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 khugepaged_mutex &x->wait irq_context: 0 khugepaged_mutex &rq->__lock irq_context: 0 khugepaged_mutex &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 ksm_thread_mutex irq_context: 0 ksm_thread_wait.lock irq_context: 0 khugepaged_mutex &obj_hash[i].lock irq_context: 0 khugepaged_mutex lock#2 irq_context: 0 khugepaged_mutex lock#2 &zone->lock irq_context: 0 khugepaged_mutex pcp_batch_high_lock irq_context: 0 damon_ops_lock irq_context: 0 crypto_alg_sem irq_context: 0 crypto_alg_sem (crypto_chain).rwsem irq_context: 0 cpu_hotplug_lock fs_reclaim irq_context: 0 cpu_hotplug_lock fs_reclaim mmu_notifier_invalidate_range_start irq_context: 0 cpu_hotplug_lock pool_lock#2 irq_context: 0 cpu_hotplug_lock pcpu_alloc_mutex irq_context: 0 cpu_hotplug_lock pcpu_alloc_mutex pcpu_lock irq_context: 0 cpu_hotplug_lock &obj_hash[i].lock irq_context: 0 cpu_hotplug_lock &wq->mutex irq_context: 0 cpu_hotplug_lock &wq->mutex &pool->lock irq_context: 0 cpu_hotplug_lock kthread_create_lock irq_context: 0 cpu_hotplug_lock &p->pi_lock irq_context: 0 cpu_hotplug_lock &p->pi_lock &rq->__lock irq_context: 0 cpu_hotplug_lock &p->pi_lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 cpu_hotplug_lock &x->wait irq_context: 0 cpu_hotplug_lock &rq->__lock irq_context: 0 cpu_hotplug_lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 lock#3 irq_context: 0 lock#3 &rq->__lock irq_context: 0 cpu_hotplug_lock &p->pi_lock &cfs_rq->removed.lock irq_context: 0 cpu_hotplug_lock wq_pool_mutex &pool->lock/1 irq_context: 0 cpu_hotplug_lock wq_pool_mutex &pool->lock/1 rcu_read_lock &pool->lock irq_context: 0 cpu_hotplug_lock wq_pool_mutex &pool->lock/1 rcu_read_lock &pool->lock &obj_hash[i].lock irq_context: 0 cpu_hotplug_lock wq_pool_mutex &pool->lock/1 rcu_read_lock &pool->lock pool_lock#2 irq_context: 0 cpu_hotplug_lock wq_pool_mutex &pool->lock/1 rcu_read_lock &pool->lock &p->pi_lock irq_context: 0 cpu_hotplug_lock wq_pool_mutex &pool->lock/1 rcu_read_lock &pool->lock &p->pi_lock &rq->__lock irq_context: 0 cpu_hotplug_lock wq_pool_mutex &pool->lock/1 rcu_read_lock &pool->lock &p->pi_lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 quarantine_lock irq_context: 0 remove_cache_srcu irq_context: 0 remove_cache_srcu quarantine_lock irq_context: 0 (wq_completion)events (work_completion)(&pwq->unbound_release_work) irq_context: 0 (wq_completion)events (work_completion)(&pwq->unbound_release_work) &wq->mutex irq_context: 0 (wq_completion)events (work_completion)(&pwq->unbound_release_work) wq_pool_mutex irq_context: 0 (wq_completion)events (work_completion)(&pwq->unbound_release_work) &obj_hash[i].lock irq_context: 0 (wq_completion)events (work_completion)(&pwq->unbound_release_work) pool_lock#2 irq_context: 0 khugepaged_mm_lock irq_context: 0 khugepaged_wait.lock irq_context: 0 &type->s_umount_key#28 &lruvec->lru_lock irq_context: 0 bio_slab_lock irq_context: 0 bio_slab_lock fs_reclaim irq_context: 0 bio_slab_lock fs_reclaim mmu_notifier_invalidate_range_start irq_context: 0 bio_slab_lock pool_lock#2 irq_context: 0 bio_slab_lock slab_mutex irq_context: 0 bio_slab_lock slab_mutex fs_reclaim irq_context: 0 bio_slab_lock slab_mutex fs_reclaim mmu_notifier_invalidate_range_start irq_context: 0 bio_slab_lock slab_mutex pool_lock#2 irq_context: 0 bio_slab_lock slab_mutex &c->lock irq_context: 0 bio_slab_lock slab_mutex &n->list_lock irq_context: 0 bio_slab_lock slab_mutex pcpu_alloc_mutex irq_context: 0 bio_slab_lock slab_mutex pcpu_alloc_mutex pcpu_lock irq_context: 0 bio_slab_lock bio_slabs.xa_lock irq_context: 0 bio_slab_lock bio_slabs.xa_lock pool_lock#2 irq_context: 0 major_names_lock irq_context: 0 major_names_lock fs_reclaim irq_context: 0 major_names_lock fs_reclaim mmu_notifier_invalidate_range_start irq_context: 0 major_names_lock pool_lock#2 irq_context: 0 major_names_lock major_names_spinlock irq_context: 0 console_lock fs_reclaim irq_context: 0 console_lock fs_reclaim mmu_notifier_invalidate_range_start irq_context: 0 console_lock &x->wait#9 irq_context: 0 console_lock &k->list_lock irq_context: 0 console_lock gdp_mutex irq_context: 0 console_lock gdp_mutex &k->list_lock irq_context: 0 console_lock gdp_mutex fs_reclaim irq_context: 0 console_lock gdp_mutex fs_reclaim mmu_notifier_invalidate_range_start irq_context: 0 console_lock gdp_mutex pool_lock#2 irq_context: 0 console_lock gdp_mutex lock irq_context: 0 console_lock gdp_mutex lock kernfs_idr_lock irq_context: 0 console_lock gdp_mutex &root->kernfs_rwsem irq_context: 0 console_lock gdp_mutex &root->kernfs_rwsem &root->kernfs_iattr_rwsem irq_context: 0 console_lock lock irq_context: 0 console_lock lock kernfs_idr_lock irq_context: 0 console_lock &root->kernfs_rwsem irq_context: 0 console_lock &root->kernfs_rwsem &root->kernfs_iattr_rwsem irq_context: 0 console_lock bus_type_sem irq_context: 0 console_lock sysfs_symlink_target_lock irq_context: 0 console_lock &root->kernfs_rwsem irq_context: 0 console_lock &dev->power.lock irq_context: 0 console_lock dpm_list_mtx irq_context: 0 console_lock uevent_sock_mutex irq_context: 0 console_lock running_helpers_waitq.lock irq_context: 0 console_lock subsys mutex#11 irq_context: 0 console_lock subsys mutex#11 &k->k_lock irq_context: 0 &pool->lock/1 &p->pi_lock &cfs_rq->removed.lock irq_context: 0 &meta->lock irq_context: 0 rcu_tasks.tasks_gp_mutex rcu_state.exp_mutex &rnp->exp_wq[3] irq_context: 0 rcu_tasks.tasks_gp_mutex (console_sem).lock irq_context: 0 rcu_tasks.tasks_gp_mutex console_lock console_srcu console_owner_lock irq_context: 0 rcu_tasks.tasks_gp_mutex console_lock console_srcu console_owner irq_context: 0 rcu_tasks.tasks_gp_mutex console_lock console_srcu console_owner &port_lock_key irq_context: 0 rcu_tasks.tasks_gp_mutex console_lock console_srcu console_owner console_owner_lock irq_context: softirq (&rtpcp->lazy_timer) irq_context: softirq (&rtpcp->lazy_timer) rcu_tasks_trace__percpu.cbs_pcpu_lock irq_context: softirq (&rtpcp->lazy_timer) rcu_tasks__percpu.cbs_pcpu_lock irq_context: 0 *(&acpi_gbl_hardware_lock) irq_context: 0 *(&acpi_gbl_gpe_lock) irq_context: 0 acpi_ioapic_lock ioapic_mutex irq_context: 0 &desc->request_mutex &irq_desc_lock_class vector_lock irq_context: 0 &desc->request_mutex &irq_desc_lock_class ioapic_lock irq_context: 0 &desc->request_mutex &irq_desc_lock_class mask_lock irq_context: 0 &desc->request_mutex &irq_desc_lock_class mask_lock tmp_mask_lock irq_context: 0 &desc->request_mutex &irq_desc_lock_class mask_lock tmp_mask_lock vector_lock irq_context: 0 &desc->request_mutex &irq_desc_lock_class mask_lock tmp_mask_lock ioapic_lock irq_context: 0 &desc->request_mutex &irq_desc_lock_class ioapic_lock i8259A_lock irq_context: 0 shrink_qlist.lock irq_context: 0 remove_cache_srcu_srcu_usage.lock irq_context: 0 remove_cache_srcu_srcu_usage.lock &obj_hash[i].lock irq_context: 0 &ACCESS_PRIVATE(sdp, lock) irq_context: 0 remove_cache_srcu irq_context: 0 remove_cache_srcu_srcu_usage.lock &ACCESS_PRIVATE(sdp, lock) irq_context: 0 remove_cache_srcu_srcu_usage.lock rcu_read_lock &pool->lock irq_context: 0 remove_cache_srcu_srcu_usage.lock rcu_read_lock &pool->lock &obj_hash[i].lock irq_context: 0 remove_cache_srcu_srcu_usage.lock rcu_read_lock &pool->lock &p->pi_lock irq_context: 0 remove_cache_srcu_srcu_usage.lock rcu_read_lock &pool->lock &p->pi_lock &rq->__lock irq_context: 0 remove_cache_srcu_srcu_usage.lock rcu_read_lock &pool->lock &p->pi_lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)rcu_gp (work_completion)(&(&ssp->srcu_sup->work)->work) &ssp->srcu_sup->srcu_gp_mutex remove_cache_srcu_srcu_usage.lock irq_context: 0 (wq_completion)rcu_gp (work_completion)(&(&ssp->srcu_sup->work)->work) remove_cache_srcu_srcu_usage.lock irq_context: 0 (wq_completion)rcu_gp (work_completion)(&(&ssp->srcu_sup->work)->work) &rq->__lock irq_context: 0 (wq_completion)rcu_gp (work_completion)(&(&ssp->srcu_sup->work)->work) &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)rcu_gp (work_completion)(&(&ssp->srcu_sup->work)->work) &ssp->srcu_sup->srcu_gp_mutex &rq->__lock irq_context: 0 (wq_completion)rcu_gp (work_completion)(&(&ssp->srcu_sup->work)->work) &ssp->srcu_sup->srcu_gp_mutex &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 &x->wait#2 irq_context: 0 (wq_completion)rcu_gp (work_completion)(&(&ssp->srcu_sup->work)->work) &ssp->srcu_sup->srcu_gp_mutex &ssp->srcu_sup->srcu_cb_mutex remove_cache_srcu_srcu_usage.lock irq_context: 0 (wq_completion)rcu_gp (work_completion)(&(&ssp->srcu_sup->work)->work) &ssp->srcu_sup->srcu_cb_mutex rcu_read_lock &pool->lock &p->pi_lock irq_context: 0 (wq_completion)rcu_gp (work_completion)(&(&ssp->srcu_sup->work)->work) &ssp->srcu_sup->srcu_cb_mutex rcu_read_lock &pool->lock &p->pi_lock &rq->__lock irq_context: 0 (wq_completion)rcu_gp (work_completion)(&(&ssp->srcu_sup->work)->work) &ssp->srcu_sup->srcu_cb_mutex rcu_read_lock &pool->lock &p->pi_lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)rcu_gp (work_completion)(&sdp->work) &x->wait#2 &p->pi_lock &cfs_rq->removed.lock irq_context: 0 cpu_hotplug_lock flush_lock irq_context: 0 cpu_hotplug_lock flush_lock &obj_hash[i].lock irq_context: 0 cpu_hotplug_lock flush_lock rcu_read_lock &pool->lock irq_context: 0 cpu_hotplug_lock flush_lock rcu_read_lock &pool->lock &obj_hash[i].lock irq_context: 0 cpu_hotplug_lock flush_lock rcu_read_lock &pool->lock &p->pi_lock irq_context: 0 cpu_hotplug_lock flush_lock rcu_read_lock &pool->lock &p->pi_lock &rq->__lock irq_context: 0 cpu_hotplug_lock flush_lock rcu_read_lock &pool->lock &p->pi_lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 cpu_hotplug_lock flush_lock (work_completion)(&sfw->work) irq_context: 0 cpu_hotplug_lock flush_lock rcu_read_lock (wq_completion)slub_flushwq irq_context: 0 cpu_hotplug_lock flush_lock &x->wait#10 irq_context: 0 cpu_hotplug_lock flush_lock &rq->__lock irq_context: 0 cpu_hotplug_lock flush_lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)slub_flushwq irq_context: 0 (wq_completion)slub_flushwq (work_completion)(&sfw->work) irq_context: 0 (wq_completion)slub_flushwq (work_completion)(&sfw->work) &c->lock irq_context: 0 (wq_completion)slub_flushwq (work_completion)(&sfw->work) rcu_read_lock pool_lock#2 irq_context: 0 (wq_completion)slub_flushwq (work_completion)(&sfw->work) &obj_hash[i].lock irq_context: 0 (wq_completion)slub_flushwq (work_completion)(&sfw->work) &n->list_lock irq_context: 0 (wq_completion)slub_flushwq (work_completion)(&barr->work) irq_context: 0 (wq_completion)slub_flushwq (work_completion)(&barr->work) &x->wait#10 irq_context: 0 (wq_completion)slub_flushwq (work_completion)(&barr->work) &x->wait#10 &p->pi_lock irq_context: 0 (wq_completion)slub_flushwq (work_completion)(&barr->work) &x->wait#10 &p->pi_lock &rq->__lock irq_context: 0 (wq_completion)slub_flushwq (work_completion)(&barr->work) &x->wait#10 &p->pi_lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 &n->list_lock &c->lock irq_context: 0 system_transition_mutex irq_context: 0 (power_off_prep_handler_list).rwsem irq_context: 0 power_off_handler_list.lock irq_context: 0 (restart_prep_handler_list).rwsem irq_context: 0 (reboot_notifier_list).rwsem irq_context: 0 *(&acpi_gbl_gpe_lock) (console_sem).lock irq_context: 0 *(&acpi_gbl_gpe_lock) console_lock console_srcu console_owner_lock irq_context: 0 *(&acpi_gbl_gpe_lock) console_lock console_srcu console_owner irq_context: 0 *(&acpi_gbl_gpe_lock) console_lock console_srcu console_owner &port_lock_key irq_context: 0 *(&acpi_gbl_gpe_lock) console_lock console_srcu console_owner console_owner_lock irq_context: 0 acpi_scan_lock irq_context: 0 acpi_scan_lock semaphore->lock irq_context: 0 acpi_scan_lock fs_reclaim irq_context: 0 acpi_scan_lock fs_reclaim mmu_notifier_invalidate_range_start irq_context: 0 acpi_scan_lock pool_lock#2 irq_context: 0 acpi_scan_lock &obj_hash[i].lock irq_context: 0 acpi_scan_lock &c->lock irq_context: 0 acpi_scan_lock &pcp->lock &zone->lock irq_context: 0 acpi_scan_lock &zone->lock irq_context: 0 acpi_scan_lock &____s->seqcount irq_context: 0 acpi_scan_lock &x->wait#9 irq_context: 0 acpi_scan_lock rcu_read_lock pool_lock#2 irq_context: 0 acpi_scan_lock acpi_device_lock irq_context: 0 acpi_scan_lock acpi_device_lock fs_reclaim irq_context: 0 acpi_scan_lock acpi_device_lock fs_reclaim mmu_notifier_invalidate_range_start irq_context: 0 acpi_scan_lock acpi_device_lock pool_lock#2 irq_context: 0 acpi_scan_lock acpi_device_lock &xa->xa_lock#2 irq_context: 0 acpi_scan_lock acpi_device_lock semaphore->lock irq_context: 0 acpi_scan_lock acpi_device_lock &obj_hash[i].lock irq_context: 0 acpi_scan_lock &k->list_lock irq_context: 0 acpi_scan_lock lock irq_context: 0 acpi_scan_lock lock kernfs_idr_lock irq_context: 0 acpi_scan_lock &root->kernfs_rwsem irq_context: 0 acpi_scan_lock &root->kernfs_rwsem &root->kernfs_iattr_rwsem irq_context: 0 acpi_scan_lock bus_type_sem irq_context: 0 acpi_scan_lock sysfs_symlink_target_lock irq_context: 0 acpi_scan_lock &k->k_lock irq_context: 0 acpi_scan_lock &root->kernfs_rwsem irq_context: 0 acpi_scan_lock &dev->power.lock irq_context: 0 acpi_scan_lock dpm_list_mtx irq_context: 0 acpi_scan_lock &dev->mutex &k->list_lock irq_context: 0 acpi_scan_lock &dev->mutex &k->k_lock irq_context: 0 acpi_scan_lock &dev->mutex &dev->power.lock irq_context: 0 acpi_scan_lock subsys mutex#12 irq_context: 0 acpi_scan_lock uevent_sock_mutex irq_context: 0 acpi_scan_lock running_helpers_waitq.lock irq_context: 0 acpi_scan_lock &pcp->lock &zone->lock &____s->seqcount irq_context: 0 acpi_scan_lock *(&acpi_gbl_reference_count_lock) irq_context: 0 acpi_scan_lock &n->list_lock irq_context: 0 acpi_scan_lock &n->list_lock &c->lock irq_context: 0 acpi_scan_lock acpi_device_lock &c->lock irq_context: 0 acpi_scan_lock acpi_device_lock &pcp->lock &zone->lock irq_context: 0 acpi_scan_lock acpi_device_lock &zone->lock irq_context: 0 acpi_scan_lock acpi_device_lock &____s->seqcount irq_context: 0 acpi_scan_lock pci_config_lock irq_context: 0 acpi_scan_lock lock kernfs_idr_lock &c->lock irq_context: 0 acpi_scan_lock lock kernfs_idr_lock &pcp->lock &zone->lock irq_context: 0 acpi_scan_lock lock kernfs_idr_lock &zone->lock irq_context: 0 acpi_scan_lock lock kernfs_idr_lock &____s->seqcount irq_context: 0 acpi_scan_lock lock kernfs_idr_lock pool_lock#2 irq_context: 0 acpi_scan_lock &obj_hash[i].lock pool_lock irq_context: 0 acpi_scan_lock batched_entropy_u8.lock irq_context: 0 acpi_scan_lock kfence_freelist_lock irq_context: 0 acpi_scan_lock &meta->lock irq_context: 0 acpi_scan_lock quarantine_lock irq_context: 0 acpi_scan_lock (console_sem).lock irq_context: 0 acpi_scan_lock console_lock console_srcu console_owner_lock irq_context: 0 acpi_scan_lock console_lock console_srcu console_owner irq_context: 0 acpi_scan_lock console_lock console_srcu console_owner &port_lock_key irq_context: 0 acpi_scan_lock console_lock console_srcu console_owner console_owner_lock irq_context: 0 acpi_scan_lock pci_bus_sem irq_context: 0 acpi_scan_lock pci_mmcfg_lock irq_context: 0 acpi_scan_lock resource_lock irq_context: 0 acpi_scan_lock &device->physical_node_lock irq_context: 0 acpi_scan_lock &device->physical_node_lock sysfs_symlink_target_lock irq_context: 0 acpi_scan_lock &device->physical_node_lock fs_reclaim irq_context: 0 acpi_scan_lock &device->physical_node_lock fs_reclaim mmu_notifier_invalidate_range_start irq_context: 0 acpi_scan_lock &device->physical_node_lock pool_lock#2 irq_context: 0 acpi_scan_lock &device->physical_node_lock lock irq_context: 0 acpi_scan_lock &device->physical_node_lock lock kernfs_idr_lock irq_context: 0 acpi_scan_lock &device->physical_node_lock &root->kernfs_rwsem irq_context: 0 acpi_scan_lock &device->physical_node_lock &root->kernfs_rwsem &root->kernfs_iattr_rwsem irq_context: 0 acpi_scan_lock fwnode_link_lock irq_context: 0 acpi_scan_lock fwnode_link_lock &k->k_lock irq_context: 0 acpi_scan_lock devtree_lock irq_context: 0 acpi_scan_lock gdp_mutex irq_context: 0 acpi_scan_lock gdp_mutex &k->list_lock irq_context: 0 acpi_scan_lock gdp_mutex fs_reclaim irq_context: 0 acpi_scan_lock gdp_mutex fs_reclaim mmu_notifier_invalidate_range_start irq_context: 0 acpi_scan_lock gdp_mutex pool_lock#2 irq_context: 0 acpi_scan_lock gdp_mutex lock irq_context: 0 acpi_scan_lock gdp_mutex lock kernfs_idr_lock irq_context: 0 acpi_scan_lock gdp_mutex &root->kernfs_rwsem irq_context: 0 acpi_scan_lock gdp_mutex &root->kernfs_rwsem &root->kernfs_iattr_rwsem irq_context: 0 acpi_scan_lock subsys mutex#13 irq_context: 0 acpi_scan_lock subsys mutex#13 &k->k_lock irq_context: 0 acpi_scan_lock pci_bus_sem irq_context: 0 acpi_scan_lock pci_acpi_companion_lookup_sem irq_context: 0 acpi_scan_lock pci_slot_mutex irq_context: 0 acpi_scan_lock tk_core.seq.seqcount irq_context: 0 acpi_scan_lock resource_alignment_lock irq_context: 0 acpi_scan_lock device_links_srcu irq_context: 0 acpi_scan_lock &dev->power.lock &dev->power.lock/1 irq_context: 0 acpi_scan_lock subsys mutex#14 irq_context: 0 acpi_scan_lock acpi_pm_notifier_install_lock irq_context: 0 acpi_scan_lock acpi_pm_notifier_install_lock semaphore->lock irq_context: 0 acpi_scan_lock acpi_pm_notifier_install_lock fs_reclaim irq_context: 0 acpi_scan_lock acpi_pm_notifier_install_lock fs_reclaim mmu_notifier_invalidate_range_start irq_context: 0 acpi_scan_lock acpi_pm_notifier_install_lock pool_lock#2 irq_context: 0 acpi_scan_lock acpi_pm_notifier_install_lock *(&acpi_gbl_reference_count_lock) irq_context: 0 acpi_scan_lock acpi_pm_notifier_install_lock acpi_pm_notifier_lock irq_context: 0 acpi_scan_lock acpi_pm_notifier_install_lock acpi_pm_notifier_lock fs_reclaim irq_context: 0 acpi_scan_lock acpi_pm_notifier_install_lock acpi_pm_notifier_lock fs_reclaim mmu_notifier_invalidate_range_start irq_context: 0 acpi_scan_lock acpi_pm_notifier_install_lock acpi_pm_notifier_lock pool_lock#2 irq_context: 0 acpi_scan_lock acpi_pm_notifier_install_lock acpi_pm_notifier_lock wakeup_ida.xa_lock irq_context: 0 acpi_scan_lock acpi_pm_notifier_install_lock acpi_pm_notifier_lock &x->wait#9 irq_context: 0 acpi_scan_lock acpi_pm_notifier_install_lock acpi_pm_notifier_lock &obj_hash[i].lock irq_context: 0 acpi_scan_lock acpi_pm_notifier_install_lock acpi_pm_notifier_lock &k->list_lock irq_context: 0 acpi_scan_lock acpi_pm_notifier_install_lock acpi_pm_notifier_lock gdp_mutex irq_context: 0 acpi_scan_lock acpi_pm_notifier_install_lock acpi_pm_notifier_lock gdp_mutex &k->list_lock irq_context: 0 acpi_scan_lock acpi_pm_notifier_install_lock acpi_pm_notifier_lock gdp_mutex fs_reclaim irq_context: 0 acpi_scan_lock acpi_pm_notifier_install_lock acpi_pm_notifier_lock gdp_mutex fs_reclaim mmu_notifier_invalidate_range_start irq_context: 0 acpi_scan_lock acpi_pm_notifier_install_lock acpi_pm_notifier_lock gdp_mutex pool_lock#2 irq_context: 0 acpi_scan_lock acpi_pm_notifier_install_lock acpi_pm_notifier_lock gdp_mutex lock irq_context: 0 acpi_scan_lock acpi_pm_notifier_install_lock acpi_pm_notifier_lock gdp_mutex lock kernfs_idr_lock irq_context: 0 acpi_scan_lock acpi_pm_notifier_install_lock acpi_pm_notifier_lock gdp_mutex &root->kernfs_rwsem irq_context: 0 acpi_scan_lock acpi_pm_notifier_install_lock acpi_pm_notifier_lock gdp_mutex &root->kernfs_rwsem &root->kernfs_iattr_rwsem irq_context: 0 acpi_scan_lock acpi_pm_notifier_install_lock acpi_pm_notifier_lock lock irq_context: 0 acpi_scan_lock acpi_pm_notifier_install_lock acpi_pm_notifier_lock lock kernfs_idr_lock irq_context: 0 acpi_scan_lock acpi_pm_notifier_install_lock acpi_pm_notifier_lock &root->kernfs_rwsem irq_context: 0 acpi_scan_lock acpi_pm_notifier_install_lock acpi_pm_notifier_lock &root->kernfs_rwsem &root->kernfs_iattr_rwsem irq_context: 0 acpi_scan_lock acpi_pm_notifier_install_lock acpi_pm_notifier_lock bus_type_sem irq_context: 0 acpi_scan_lock acpi_pm_notifier_install_lock acpi_pm_notifier_lock sysfs_symlink_target_lock irq_context: 0 acpi_scan_lock acpi_pm_notifier_install_lock acpi_pm_notifier_lock &c->lock irq_context: 0 acpi_scan_lock acpi_pm_notifier_install_lock acpi_pm_notifier_lock &____s->seqcount irq_context: 0 acpi_scan_lock acpi_pm_notifier_install_lock acpi_pm_notifier_lock uevent_sock_mutex irq_context: 0 acpi_scan_lock acpi_pm_notifier_install_lock acpi_pm_notifier_lock running_helpers_waitq.lock irq_context: 0 acpi_scan_lock acpi_pm_notifier_install_lock acpi_pm_notifier_lock &k->k_lock irq_context: 0 acpi_scan_lock acpi_pm_notifier_install_lock acpi_pm_notifier_lock subsys mutex#15 irq_context: 0 acpi_scan_lock acpi_pm_notifier_install_lock acpi_pm_notifier_lock subsys mutex#15 &k->k_lock irq_context: 0 acpi_scan_lock acpi_pm_notifier_install_lock acpi_pm_notifier_lock events_lock irq_context: 0 acpi_scan_lock acpi_pm_notifier_install_lock acpi_pm_notifier_lock gdp_mutex lock kernfs_idr_lock pool_lock#2 irq_context: 0 acpi_scan_lock acpi_pm_notifier_install_lock acpi_pm_notifier_lock &obj_hash[i].lock pool_lock irq_context: 0 acpi_scan_lock &device->physical_node_lock &c->lock irq_context: 0 acpi_scan_lock &device->physical_node_lock &pcp->lock &zone->lock irq_context: 0 acpi_scan_lock &device->physical_node_lock &zone->lock irq_context: 0 acpi_scan_lock &device->physical_node_lock &____s->seqcount irq_context: 0 acpi_scan_lock acpi_pm_notifier_install_lock acpi_pm_notifier_lock lock kernfs_idr_lock pool_lock#2 irq_context: 0 acpi_scan_lock acpi_pm_notifier_install_lock acpi_pm_notifier_lock &pcp->lock &zone->lock irq_context: 0 acpi_scan_lock acpi_pm_notifier_install_lock acpi_pm_notifier_lock &zone->lock irq_context: 0 &pgdat->kswapd_lock irq_context: 0 acpi_scan_lock &rq->__lock irq_context: 0 acpi_scan_lock acpi_pm_notifier_install_lock acpi_pm_notifier_lock fill_pool_map-wait-type-override &pcp->lock &zone->lock irq_context: 0 acpi_scan_lock acpi_pm_notifier_install_lock acpi_pm_notifier_lock fill_pool_map-wait-type-override &zone->lock irq_context: 0 acpi_scan_lock acpi_pm_notifier_install_lock acpi_pm_notifier_lock fill_pool_map-wait-type-override &____s->seqcount irq_context: 0 acpi_scan_lock acpi_pm_notifier_install_lock acpi_pm_notifier_lock fill_pool_map-wait-type-override pool_lock#2 irq_context: 0 acpi_scan_lock acpi_pm_notifier_install_lock acpi_pm_notifier_lock fill_pool_map-wait-type-override &c->lock irq_context: 0 acpi_scan_lock acpi_pm_notifier_install_lock acpi_pm_notifier_lock fill_pool_map-wait-type-override pool_lock irq_context: softirq drivers/char/random.c:251 irq_context: softirq drivers/char/random.c:251 rcu_read_lock &pool->lock/1 irq_context: softirq drivers/char/random.c:251 rcu_read_lock &pool->lock/1 &obj_hash[i].lock irq_context: softirq drivers/char/random.c:251 rcu_read_lock &pool->lock/1 &p->pi_lock irq_context: softirq drivers/char/random.c:251 rcu_read_lock &pool->lock/1 &p->pi_lock &rq->__lock irq_context: softirq drivers/char/random.c:251 rcu_read_lock &pool->lock/1 &p->pi_lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)events_unbound (next_reseed).work irq_context: 0 (wq_completion)events_unbound (next_reseed).work &obj_hash[i].lock irq_context: 0 (wq_completion)events_unbound (next_reseed).work &base->lock irq_context: 0 (wq_completion)events_unbound (next_reseed).work &base->lock &obj_hash[i].lock irq_context: 0 (wq_completion)events_unbound (next_reseed).work input_pool.lock irq_context: 0 (wq_completion)events_unbound (next_reseed).work base_crng.lock irq_context: 0 acpi_scan_lock pci_rescan_remove_lock irq_context: 0 acpi_scan_lock pci_rescan_remove_lock &dev->mutex &dev->power.lock irq_context: 0 acpi_scan_lock pci_rescan_remove_lock &dev->mutex &k->list_lock irq_context: 0 acpi_scan_lock pci_rescan_remove_lock &dev->mutex &k->k_lock irq_context: 0 acpi_scan_lock subsys mutex#3 irq_context: 0 acpi_scan_lock acpi_link_lock irq_context: 0 acpi_scan_lock acpi_link_lock fs_reclaim irq_context: 0 acpi_scan_lock acpi_link_lock fs_reclaim mmu_notifier_invalidate_range_start irq_context: 0 acpi_scan_lock acpi_link_lock pool_lock#2 irq_context: 0 acpi_scan_lock acpi_link_lock semaphore->lock irq_context: 0 acpi_scan_lock acpi_link_lock &obj_hash[i].lock irq_context: 0 acpi_scan_lock acpi_link_lock *(&acpi_gbl_reference_count_lock) irq_context: 0 acpi_scan_lock acpi_link_lock pci_config_lock irq_context: 0 acpi_scan_lock acpi_link_lock &pcp->lock &zone->lock irq_context: 0 acpi_scan_lock acpi_link_lock &zone->lock irq_context: 0 acpi_scan_lock acpi_link_lock &____s->seqcount irq_context: 0 acpi_scan_lock acpi_link_lock rcu_read_lock pool_lock#2 irq_context: 0 acpi_scan_lock acpi_link_lock (console_sem).lock irq_context: 0 acpi_scan_lock acpi_link_lock console_lock console_srcu console_owner_lock irq_context: 0 acpi_scan_lock acpi_link_lock console_lock console_srcu console_owner irq_context: 0 acpi_scan_lock acpi_link_lock console_lock console_srcu console_owner &port_lock_key irq_context: 0 acpi_scan_lock acpi_link_lock console_lock console_srcu console_owner console_owner_lock irq_context: 0 acpi_scan_lock acpi_link_lock &c->lock irq_context: 0 acpi_scan_lock batched_entropy_u8.lock crngs.lock irq_context: 0 acpi_scan_lock batched_entropy_u8.lock crngs.lock base_crng.lock irq_context: 0 acpi_scan_lock acpi_link_lock &pcp->lock &zone->lock &____s->seqcount irq_context: softirq mm/vmstat.c:2018 irq_context: softirq mm/vmstat.c:2018 rcu_read_lock &pool->lock irq_context: softirq mm/vmstat.c:2018 rcu_read_lock &pool->lock &obj_hash[i].lock irq_context: softirq mm/vmstat.c:2018 rcu_read_lock &pool->lock &p->pi_lock irq_context: softirq mm/vmstat.c:2018 rcu_read_lock &pool->lock &p->pi_lock &rq->__lock irq_context: softirq mm/vmstat.c:2018 rcu_read_lock &pool->lock &p->pi_lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)events (shepherd).work irq_context: 0 (wq_completion)events (shepherd).work cpu_hotplug_lock irq_context: 0 (wq_completion)events (shepherd).work cpu_hotplug_lock rcu_read_lock &pool->lock irq_context: 0 (wq_completion)events (shepherd).work cpu_hotplug_lock rcu_read_lock &pool->lock &obj_hash[i].lock irq_context: 0 (wq_completion)events (shepherd).work cpu_hotplug_lock rcu_read_lock &pool->lock &p->pi_lock irq_context: 0 (wq_completion)events (shepherd).work cpu_hotplug_lock rcu_read_lock &pool->lock &p->pi_lock &rq->__lock irq_context: 0 (wq_completion)events (shepherd).work cpu_hotplug_lock rcu_read_lock &pool->lock &p->pi_lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)events (shepherd).work &obj_hash[i].lock irq_context: 0 (wq_completion)events (shepherd).work &base->lock irq_context: 0 (wq_completion)events (shepherd).work &base->lock &obj_hash[i].lock irq_context: 0 (wq_completion)mm_percpu_wq irq_context: 0 (wq_completion)mm_percpu_wq (work_completion)(&(({ do { const void *__vpp_verify = (typeof((&vmstat_work) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __asm__ ("" : "=r"(__ptr) : "0"((typeof(*((&vmstat_work))) *)((&vmstat_work)))); (typeof((typeof(*((&vmstat_work))) *)((&vmstat_work)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); }))->work) irq_context: 0 (wq_completion)mm_percpu_wq (work_completion)(&(({ do { const void *__vpp_verify = (typeof((&vmstat_work) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __asm__ ("" : "=r"(__ptr) : "0"((typeof(*((&vmstat_work))) *)((&vmstat_work)))); (typeof((typeof(*((&vmstat_work))) *)((&vmstat_work)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); }))->work) &obj_hash[i].lock irq_context: 0 (wq_completion)mm_percpu_wq (work_completion)(&(({ do { const void *__vpp_verify = (typeof((&vmstat_work) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __asm__ ("" : "=r"(__ptr) : "0"((typeof(*((&vmstat_work))) *)((&vmstat_work)))); (typeof((typeof(*((&vmstat_work))) *)((&vmstat_work)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); }))->work) &base->lock irq_context: 0 (wq_completion)mm_percpu_wq (work_completion)(&(({ do { const void *__vpp_verify = (typeof((&vmstat_work) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __asm__ ("" : "=r"(__ptr) : "0"((typeof(*((&vmstat_work))) *)((&vmstat_work)))); (typeof((typeof(*((&vmstat_work))) *)((&vmstat_work)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); }))->work) &base->lock &obj_hash[i].lock irq_context: 0 acpi_scan_lock acpi_dep_list_lock irq_context: 0 acpi_scan_lock wakeup_ida.xa_lock irq_context: 0 acpi_scan_lock fill_pool_map-wait-type-override pool_lock#2 irq_context: 0 acpi_scan_lock fill_pool_map-wait-type-override pool_lock irq_context: 0 acpi_scan_lock subsys mutex#15 irq_context: 0 acpi_scan_lock subsys mutex#15 &k->k_lock irq_context: 0 acpi_scan_lock events_lock irq_context: 0 acpi_scan_lock power_resource_list_lock irq_context: 0 acpi_device_lock irq_context: 0 &(&priv->bus_notifier)->rwsem irq_context: 0 k-sk_lock-AF_NETLINK irq_context: 0 k-sk_lock-AF_NETLINK k-slock-AF_NETLINK irq_context: 0 k-sk_lock-AF_NETLINK rcu_read_lock rhashtable_bucket irq_context: 0 k-slock-AF_NETLINK irq_context: 0 &type->s_umount_key#10/1 irq_context: 0 &type->s_umount_key#10/1 fs_reclaim irq_context: 0 &type->s_umount_key#10/1 fs_reclaim mmu_notifier_invalidate_range_start irq_context: 0 &type->s_umount_key#10/1 pool_lock#2 irq_context: 0 &type->s_umount_key#10/1 pcpu_alloc_mutex irq_context: 0 &type->s_umount_key#10/1 pcpu_alloc_mutex pcpu_lock irq_context: 0 &type->s_umount_key#10/1 shrinker_rwsem irq_context: 0 &type->s_umount_key#10/1 list_lrus_mutex irq_context: 0 &type->s_umount_key#10/1 sb_lock irq_context: 0 &type->s_umount_key#10/1 sb_lock unnamed_dev_ida.xa_lock irq_context: 0 &type->s_umount_key#10/1 &pcp->lock &zone->lock irq_context: 0 &type->s_umount_key#10/1 &zone->lock irq_context: 0 &type->s_umount_key#10/1 &____s->seqcount irq_context: 0 &type->s_umount_key#10/1 &c->lock irq_context: 0 &type->s_umount_key#10/1 mmu_notifier_invalidate_range_start irq_context: 0 &type->s_umount_key#10/1 &sb->s_type->i_lock_key#9 irq_context: 0 &type->s_umount_key#10/1 &s->s_inode_list_lock irq_context: 0 &type->s_umount_key#10/1 tk_core.seq.seqcount irq_context: 0 &type->s_umount_key#10/1 &pcp->lock &zone->lock &____s->seqcount irq_context: 0 &type->s_umount_key#10/1 &sb->s_type->i_lock_key#9 &dentry->d_lock irq_context: 0 &type->s_umount_key#10/1 &dentry->d_lock irq_context: 0 &type->s_umount_key#11/1 irq_context: 0 &type->s_umount_key#11/1 fs_reclaim irq_context: 0 &type->s_umount_key#11/1 fs_reclaim mmu_notifier_invalidate_range_start irq_context: 0 &type->s_umount_key#11/1 pool_lock#2 irq_context: 0 &type->s_umount_key#11/1 pcpu_alloc_mutex irq_context: 0 &type->s_umount_key#11/1 pcpu_alloc_mutex pcpu_lock irq_context: 0 &type->s_umount_key#11/1 shrinker_rwsem irq_context: 0 &type->s_umount_key#11/1 &c->lock irq_context: 0 &type->s_umount_key#11/1 &pcp->lock &zone->lock irq_context: 0 &type->s_umount_key#11/1 &zone->lock irq_context: 0 &type->s_umount_key#11/1 &____s->seqcount irq_context: 0 &type->s_umount_key#11/1 list_lrus_mutex irq_context: 0 &type->s_umount_key#11/1 sb_lock irq_context: 0 &type->s_umount_key#11/1 sb_lock unnamed_dev_ida.xa_lock irq_context: 0 &type->s_umount_key#11/1 mmu_notifier_invalidate_range_start irq_context: 0 &type->s_umount_key#11/1 &sb->s_type->i_lock_key#10 irq_context: 0 &type->s_umount_key#11/1 &s->s_inode_list_lock irq_context: 0 &type->s_umount_key#11/1 tk_core.seq.seqcount irq_context: 0 &type->s_umount_key#11/1 &sb->s_type->i_lock_key#10 &dentry->d_lock irq_context: 0 &type->s_umount_key#11/1 &dentry->d_lock irq_context: 0 &mm->mmap_lock irq_context: 0 &mm->mmap_lock reservation_ww_class_acquire irq_context: 0 &mm->mmap_lock reservation_ww_class_acquire reservation_ww_class_mutex irq_context: 0 &mm->mmap_lock reservation_ww_class_acquire reservation_ww_class_mutex fs_reclaim irq_context: 0 &mm->mmap_lock reservation_ww_class_acquire reservation_ww_class_mutex fs_reclaim mmu_notifier_invalidate_range_start irq_context: 0 &mm->mmap_lock reservation_ww_class_acquire reservation_ww_class_mutex fs_reclaim &mapping->i_mmap_rwsem irq_context: 0 &mm->mmap_lock reservation_ww_class_acquire reservation_ww_class_mutex fs_reclaim mmu_notifier_invalidate_range_start dma_fence_map irq_context: 0 delayed_uprobe_lock irq_context: 0 key irq_context: 0 fill_pool_map-wait-type-override pool_lock#2 irq_context: 0 fill_pool_map-wait-type-override pool_lock irq_context: 0 attribute_container_mutex irq_context: 0 triggers_list_lock irq_context: 0 leds_list_lock irq_context: 0 &sb->s_type->i_mutex_key#3 &pcp->lock &zone->lock irq_context: 0 &sb->s_type->i_mutex_key#3 &zone->lock irq_context: 0 &sb->s_type->i_mutex_key#3 rcu_read_lock pool_lock#2 irq_context: 0 &sb->s_type->i_mutex_key#3 &obj_hash[i].lock irq_context: 0 bus_type_sem irq_context: 0 (usb_notifier_list).rwsem irq_context: 0 &device->physical_node_lock irq_context: 0 rc_map_lock irq_context: 0 subsys mutex#16 irq_context: 0 fill_pool_map-wait-type-override &c->lock irq_context: 0 fill_pool_map-wait-type-override &pcp->lock &zone->lock irq_context: 0 fill_pool_map-wait-type-override &zone->lock irq_context: 0 fill_pool_map-wait-type-override &____s->seqcount irq_context: 0 resource_lock irq_context: 0 free_vmap_area_lock &obj_hash[i].lock irq_context: 0 free_vmap_area_lock pool_lock#2 irq_context: 0 lock kernfs_idr_lock &pcp->lock &zone->lock irq_context: 0 lock kernfs_idr_lock &zone->lock irq_context: 0 &entry->access irq_context: 0 info_mutex irq_context: 0 info_mutex proc_subdir_lock irq_context: 0 info_mutex fs_reclaim irq_context: 0 info_mutex fs_reclaim mmu_notifier_invalidate_range_start irq_context: 0 info_mutex pool_lock#2 irq_context: 0 info_mutex proc_inum_ida.xa_lock irq_context: 0 info_mutex proc_subdir_lock irq_context: 0 kobj_ns_type_lock irq_context: 0 pernet_ops_rwsem &pcp->lock &zone->lock &____s->seqcount irq_context: 0 pernet_ops_rwsem rtnl_mutex fs_reclaim irq_context: 0 pernet_ops_rwsem rtnl_mutex fs_reclaim mmu_notifier_invalidate_range_start irq_context: 0 pernet_ops_rwsem rtnl_mutex pool_lock#2 irq_context: 0 pernet_ops_rwsem rtnl_mutex pcpu_alloc_mutex irq_context: 0 pernet_ops_rwsem rtnl_mutex pcpu_alloc_mutex pcpu_lock irq_context: 0 pernet_ops_rwsem rtnl_mutex &xa->xa_lock#3 irq_context: 0 pernet_ops_rwsem rtnl_mutex &xa->xa_lock#3 pool_lock#2 irq_context: 0 pernet_ops_rwsem rtnl_mutex &x->wait#9 irq_context: 0 pernet_ops_rwsem rtnl_mutex &obj_hash[i].lock irq_context: 0 pernet_ops_rwsem rtnl_mutex &k->list_lock irq_context: 0 pernet_ops_rwsem rtnl_mutex gdp_mutex irq_context: 0 pernet_ops_rwsem rtnl_mutex gdp_mutex &k->list_lock irq_context: 0 pernet_ops_rwsem rtnl_mutex gdp_mutex fs_reclaim irq_context: 0 pernet_ops_rwsem rtnl_mutex gdp_mutex fs_reclaim mmu_notifier_invalidate_range_start irq_context: 0 pernet_ops_rwsem rtnl_mutex gdp_mutex pool_lock#2 irq_context: 0 pernet_ops_rwsem rtnl_mutex gdp_mutex lock irq_context: 0 pernet_ops_rwsem rtnl_mutex gdp_mutex lock kernfs_idr_lock irq_context: 0 pernet_ops_rwsem rtnl_mutex gdp_mutex &root->kernfs_rwsem irq_context: 0 pernet_ops_rwsem rtnl_mutex gdp_mutex &root->kernfs_rwsem &root->kernfs_iattr_rwsem irq_context: 0 pernet_ops_rwsem rtnl_mutex gdp_mutex kobj_ns_type_lock irq_context: 0 pernet_ops_rwsem rtnl_mutex lock irq_context: 0 pernet_ops_rwsem rtnl_mutex lock kernfs_idr_lock irq_context: 0 pernet_ops_rwsem rtnl_mutex &root->kernfs_rwsem irq_context: 0 pernet_ops_rwsem rtnl_mutex &root->kernfs_rwsem &root->kernfs_iattr_rwsem irq_context: 0 pernet_ops_rwsem rtnl_mutex bus_type_sem irq_context: 0 pernet_ops_rwsem rtnl_mutex sysfs_symlink_target_lock irq_context: 0 pernet_ops_rwsem rtnl_mutex &c->lock irq_context: 0 pernet_ops_rwsem rtnl_mutex &____s->seqcount irq_context: 0 pernet_ops_rwsem rtnl_mutex lock kernfs_idr_lock pool_lock#2 irq_context: 0 pernet_ops_rwsem rtnl_mutex &root->kernfs_rwsem irq_context: 0 pernet_ops_rwsem rtnl_mutex &dev->power.lock irq_context: 0 pernet_ops_rwsem rtnl_mutex dpm_list_mtx irq_context: 0 pernet_ops_rwsem rtnl_mutex uevent_sock_mutex irq_context: 0 pernet_ops_rwsem rtnl_mutex uevent_sock_mutex fs_reclaim irq_context: 0 pernet_ops_rwsem rtnl_mutex uevent_sock_mutex fs_reclaim mmu_notifier_invalidate_range_start irq_context: 0 pernet_ops_rwsem rtnl_mutex uevent_sock_mutex pool_lock#2 irq_context: 0 pernet_ops_rwsem rtnl_mutex uevent_sock_mutex nl_table_lock irq_context: 0 pernet_ops_rwsem rtnl_mutex uevent_sock_mutex &obj_hash[i].lock irq_context: 0 pernet_ops_rwsem rtnl_mutex uevent_sock_mutex nl_table_wait.lock irq_context: 0 pernet_ops_rwsem rtnl_mutex running_helpers_waitq.lock irq_context: 0 pernet_ops_rwsem rtnl_mutex subsys mutex#17 irq_context: 0 pernet_ops_rwsem rtnl_mutex subsys mutex#17 &k->k_lock irq_context: 0 pernet_ops_rwsem rtnl_mutex &dir->lock#2 irq_context: 0 pernet_ops_rwsem rtnl_mutex &pcp->lock &zone->lock irq_context: 0 pernet_ops_rwsem rtnl_mutex &zone->lock irq_context: 0 pernet_ops_rwsem rtnl_mutex &pcp->lock &zone->lock &____s->seqcount irq_context: 0 pernet_ops_rwsem rtnl_mutex rcu_read_lock pool_lock#2 irq_context: 0 pernet_ops_rwsem rtnl_mutex dev_hotplug_mutex irq_context: 0 pernet_ops_rwsem rtnl_mutex dev_hotplug_mutex &dev->power.lock irq_context: 0 pernet_ops_rwsem rtnl_mutex dev_base_lock irq_context: 0 pernet_ops_rwsem rtnl_mutex input_pool.lock irq_context: 0 pernet_ops_rwsem rtnl_mutex nl_table_lock irq_context: 0 pernet_ops_rwsem rtnl_mutex nl_table_wait.lock irq_context: 0 rcu_read_lock &obj_hash[i].lock irq_context: 0 rcu_read_lock nl_table_lock irq_context: 0 rcu_read_lock nl_table_wait.lock irq_context: 0 qdisc_mod_lock irq_context: 0 bt_proto_lock irq_context: 0 hci_cb_list_lock irq_context: 0 mgmt_chan_list_lock irq_context: 0 pernet_ops_rwsem rtnl_mutex net_rwsem irq_context: 0 pernet_ops_rwsem rtnl_mutex net_rwsem &list->lock#2 irq_context: 0 rate_ctrl_mutex irq_context: 0 rate_ctrl_mutex fs_reclaim irq_context: 0 rate_ctrl_mutex fs_reclaim mmu_notifier_invalidate_range_start irq_context: 0 rate_ctrl_mutex pool_lock#2 irq_context: 0 netlbl_domhsh_lock irq_context: 0 netlbl_unlhsh_lock irq_context: 0 rcu_read_lock netlbl_domhsh_lock irq_context: 0 rcu_read_lock netlbl_domhsh_lock pool_lock#2 irq_context: 0 misc_mtx irq_context: 0 misc_mtx fs_reclaim irq_context: 0 misc_mtx fs_reclaim mmu_notifier_invalidate_range_start irq_context: 0 misc_mtx pool_lock#2 irq_context: 0 misc_mtx &x->wait#9 irq_context: 0 misc_mtx &obj_hash[i].lock irq_context: 0 misc_mtx &k->list_lock irq_context: 0 misc_mtx gdp_mutex irq_context: 0 misc_mtx gdp_mutex &k->list_lock irq_context: 0 misc_mtx gdp_mutex fs_reclaim irq_context: 0 misc_mtx gdp_mutex fs_reclaim mmu_notifier_invalidate_range_start irq_context: 0 misc_mtx gdp_mutex pool_lock#2 irq_context: 0 misc_mtx gdp_mutex lock irq_context: 0 misc_mtx gdp_mutex lock kernfs_idr_lock irq_context: 0 misc_mtx gdp_mutex &root->kernfs_rwsem irq_context: 0 misc_mtx gdp_mutex &root->kernfs_rwsem &root->kernfs_iattr_rwsem irq_context: 0 misc_mtx lock irq_context: 0 misc_mtx lock kernfs_idr_lock irq_context: 0 misc_mtx &root->kernfs_rwsem irq_context: 0 misc_mtx &root->kernfs_rwsem &root->kernfs_iattr_rwsem irq_context: 0 misc_mtx bus_type_sem irq_context: 0 misc_mtx &c->lock irq_context: 0 misc_mtx &____s->seqcount irq_context: 0 misc_mtx sysfs_symlink_target_lock irq_context: 0 misc_mtx &root->kernfs_rwsem irq_context: 0 misc_mtx &dev->power.lock irq_context: 0 misc_mtx dpm_list_mtx irq_context: 0 misc_mtx req_lock irq_context: 0 misc_mtx &p->pi_lock irq_context: 0 misc_mtx &p->pi_lock &rq->__lock irq_context: 0 misc_mtx &p->pi_lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 misc_mtx &x->wait#11 irq_context: 0 misc_mtx &rq->__lock irq_context: 0 misc_mtx &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 sb_writers irq_context: 0 sb_writers mount_lock irq_context: 0 sb_writers &type->i_mutex_dir_key/1 irq_context: 0 sb_writers &type->i_mutex_dir_key/1 rename_lock.seqcount irq_context: 0 sb_writers &type->i_mutex_dir_key/1 fs_reclaim irq_context: 0 sb_writers &type->i_mutex_dir_key/1 fs_reclaim mmu_notifier_invalidate_range_start irq_context: 0 sb_writers &type->i_mutex_dir_key/1 pool_lock#2 irq_context: 0 sb_writers &type->i_mutex_dir_key/1 &dentry->d_lock irq_context: 0 sb_writers &type->i_mutex_dir_key/1 &obj_hash[i].lock irq_context: 0 sb_writers &type->i_mutex_dir_key/1 &sbinfo->stat_lock irq_context: 0 sb_writers &type->i_mutex_dir_key/1 &pcp->lock &zone->lock irq_context: 0 sb_writers &type->i_mutex_dir_key/1 &zone->lock irq_context: 0 sb_writers &type->i_mutex_dir_key/1 &____s->seqcount irq_context: 0 sb_writers &type->i_mutex_dir_key/1 &c->lock irq_context: 0 sb_writers &type->i_mutex_dir_key/1 mmu_notifier_invalidate_range_start irq_context: 0 sb_writers &type->i_mutex_dir_key/1 &sb->s_type->i_lock_key#5 irq_context: 0 sb_writers &type->i_mutex_dir_key/1 &s->s_inode_list_lock irq_context: 0 sb_writers &type->i_mutex_dir_key/1 tk_core.seq.seqcount irq_context: 0 sb_writers &type->i_mutex_dir_key/1 batched_entropy_u32.lock irq_context: 0 sb_writers &type->i_mutex_dir_key/1 batched_entropy_u32.lock crngs.lock irq_context: 0 sb_writers &type->i_mutex_dir_key/1 &simple_offset_xa_lock irq_context: 0 sb_writers &type->i_mutex_dir_key/1 &simple_offset_xa_lock pool_lock#2 irq_context: 0 sb_writers &type->i_mutex_dir_key/1 &sb->s_type->i_lock_key#5 &dentry->d_lock irq_context: 0 sb_writers &type->i_mutex_dir_key/1 &sb->s_type->i_mutex_key#4 irq_context: 0 sb_writers &type->i_mutex_dir_key/1 &sb->s_type->i_mutex_key#4 tk_core.seq.seqcount irq_context: 0 sb_writers &type->i_mutex_dir_key/1 rcu_read_lock &dentry->d_lock irq_context: 0 &x->wait#11 irq_context: 0 &x->wait#11 &p->pi_lock irq_context: 0 &x->wait#11 &p->pi_lock &rq->__lock irq_context: 0 &x->wait#11 &p->pi_lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 misc_mtx uevent_sock_mutex irq_context: 0 misc_mtx running_helpers_waitq.lock irq_context: 0 misc_mtx subsys mutex#18 irq_context: 0 misc_mtx subsys mutex#18 &k->k_lock irq_context: 0 rcu_read_lock &pool->lock irq_context: 0 rcu_read_lock &pool->lock &obj_hash[i].lock irq_context: 0 rcu_read_lock &pool->lock &p->pi_lock irq_context: 0 rcu_read_lock &pool->lock &p->pi_lock &rq->__lock irq_context: 0 rcu_read_lock &pool->lock &p->pi_lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 input_mutex irq_context: 0 input_mutex input_devices_poll_wait.lock irq_context: 0 (netlink_chain).rwsem irq_context: 0 proto_tab_lock irq_context: 0 resource_lock &c->lock irq_context: 0 resource_lock &____s->seqcount irq_context: 0 resource_lock pool_lock#2 irq_context: 0 resource_lock &obj_hash[i].lock irq_context: 0 random_ready_notifier.lock irq_context: 0 random_ready_notifier.lock crngs.lock irq_context: 0 misc_mtx misc_minors_ida.xa_lock irq_context: 0 (wq_completion)events (work_completion)(&rfkill_global_led_trigger_work) irq_context: 0 (wq_completion)events (work_completion)(&rfkill_global_led_trigger_work) rfkill_global_mutex irq_context: 0 vga_lock#2 irq_context: 0 vga_lock#2 pci_config_lock irq_context: 0 vga_lock#2 (console_sem).lock irq_context: 0 vga_lock#2 console_lock console_srcu console_owner_lock irq_context: 0 vga_lock#2 console_lock console_srcu console_owner irq_context: 0 vga_lock#2 console_lock console_srcu console_owner &port_lock_key irq_context: 0 vga_lock#2 console_lock console_srcu console_owner console_owner_lock irq_context: 0 disable_lock irq_context: 0 disable_lock fs_reclaim irq_context: 0 disable_lock fs_reclaim mmu_notifier_invalidate_range_start irq_context: 0 disable_lock pool_lock#2 irq_context: 0 disable_lock &x->wait#9 irq_context: 0 disable_lock &obj_hash[i].lock irq_context: 0 disable_lock &k->list_lock irq_context: 0 disable_lock lock irq_context: 0 disable_lock lock kernfs_idr_lock irq_context: 0 disable_lock &root->kernfs_rwsem irq_context: 0 disable_lock &root->kernfs_rwsem &root->kernfs_iattr_rwsem irq_context: 0 disable_lock bus_type_sem irq_context: 0 disable_lock sysfs_symlink_target_lock irq_context: 0 disable_lock &k->k_lock irq_context: 0 disable_lock &root->kernfs_rwsem irq_context: 0 disable_lock &c->lock irq_context: 0 disable_lock &____s->seqcount irq_context: 0 disable_lock &dev->power.lock irq_context: 0 disable_lock dpm_list_mtx irq_context: 0 disable_lock &(&priv->bus_notifier)->rwsem irq_context: 0 disable_lock uevent_sock_mutex irq_context: 0 disable_lock running_helpers_waitq.lock irq_context: 0 disable_lock &dev->mutex &dev->power.lock irq_context: 0 disable_lock &dev->mutex &k->list_lock irq_context: 0 disable_lock &dev->mutex &k->k_lock irq_context: 0 disable_lock subsys mutex#3 irq_context: 0 &sb->s_type->i_mutex_key#3 batched_entropy_u8.lock irq_context: 0 &sb->s_type->i_mutex_key#3 kfence_freelist_lock irq_context: 0 &type->s_umount_key#12/1 irq_context: 0 &type->s_umount_key#12/1 fs_reclaim irq_context: 0 &type->s_umount_key#12/1 fs_reclaim mmu_notifier_invalidate_range_start irq_context: 0 &type->s_umount_key#12/1 pool_lock#2 irq_context: 0 &type->s_umount_key#12/1 pcpu_alloc_mutex irq_context: 0 &type->s_umount_key#12/1 pcpu_alloc_mutex pcpu_lock irq_context: 0 &type->s_umount_key#12/1 shrinker_rwsem irq_context: 0 &type->s_umount_key#12/1 list_lrus_mutex irq_context: 0 &type->s_umount_key#12/1 sb_lock irq_context: 0 &type->s_umount_key#12/1 sb_lock unnamed_dev_ida.xa_lock irq_context: 0 &type->s_umount_key#12/1 mmu_notifier_invalidate_range_start irq_context: 0 &type->s_umount_key#12/1 &sb->s_type->i_lock_key#11 irq_context: 0 &type->s_umount_key#12/1 &s->s_inode_list_lock irq_context: 0 &type->s_umount_key#12/1 tk_core.seq.seqcount irq_context: 0 &type->s_umount_key#12/1 &sb->s_type->i_lock_key#11 &dentry->d_lock irq_context: 0 &type->s_umount_key#12/1 &dentry->d_lock irq_context: 0 &sb->s_type->i_lock_key#11 irq_context: 0 clocksource_mutex cpu_hotplug_lock irq_context: 0 clocksource_mutex cpu_hotplug_lock stop_cpus_mutex irq_context: 0 clocksource_mutex cpu_hotplug_lock stop_cpus_mutex &stopper->lock irq_context: 0 clocksource_mutex cpu_hotplug_lock stop_cpus_mutex &stop_pi_lock irq_context: 0 clocksource_mutex cpu_hotplug_lock stop_cpus_mutex &stop_pi_lock &rq->__lock irq_context: 0 clocksource_mutex cpu_hotplug_lock stop_cpus_mutex &stop_pi_lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 clocksource_mutex cpu_hotplug_lock stop_cpus_mutex &rq->__lock irq_context: 0 clocksource_mutex cpu_hotplug_lock stop_cpus_mutex &x->wait#8 irq_context: 0 clocksource_mutex (console_sem).lock irq_context: 0 clocksource_mutex console_lock console_srcu console_owner_lock irq_context: 0 clocksource_mutex console_lock console_srcu console_owner irq_context: 0 clocksource_mutex console_lock console_srcu console_owner &port_lock_key irq_context: 0 clocksource_mutex console_lock console_srcu console_owner console_owner_lock irq_context: hardirq tick_broadcast_lock irq_context: hardirq tick_broadcast_lock jiffies_lock irq_context: hardirq hrtimer_bases.lock &obj_hash[i].lock irq_context: 0 clocksource_mutex &rq->__lock irq_context: 0 (wq_completion)events timer_update_work irq_context: 0 (wq_completion)events timer_update_work timer_keys_mutex irq_context: 0 (wq_completion)events timer_update_work timer_keys_mutex cpu_hotplug_lock irq_context: 0 (wq_completion)events timer_update_work timer_keys_mutex cpu_hotplug_lock jump_label_mutex irq_context: 0 (wq_completion)events timer_update_work timer_keys_mutex cpu_hotplug_lock jump_label_mutex text_mutex irq_context: 0 (wq_completion)events timer_update_work timer_keys_mutex cpu_hotplug_lock jump_label_mutex text_mutex ptlock_ptr(page)#2 irq_context: 0 rcu_read_lock &pool->lock/1 &p->pi_lock &rq->__lock irq_context: 0 rcu_read_lock &pool->lock/1 &p->pi_lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 &type->s_umount_key#28 rcu_read_lock pool_lock#2 irq_context: 0 &type->s_umount_key#28 &pcp->lock &zone->lock irq_context: 0 &type->s_umount_key#28 &zone->lock irq_context: 0 &type->s_umount_key#28 crypto_alg_sem irq_context: 0 &type->s_umount_key#28 lock#3 irq_context: 0 &type->s_umount_key#13/1 irq_context: 0 &type->s_umount_key#13/1 fs_reclaim irq_context: 0 &type->s_umount_key#28 lock#3 &obj_hash[i].lock irq_context: 0 &type->s_umount_key#13/1 fs_reclaim mmu_notifier_invalidate_range_start irq_context: 0 &type->s_umount_key#28 lock#3 rcu_read_lock &pool->lock irq_context: 0 &type->s_umount_key#13/1 pool_lock#2 irq_context: 0 &type->s_umount_key#28 lock#3 rcu_read_lock &pool->lock &obj_hash[i].lock irq_context: 0 &type->s_umount_key#13/1 pcpu_alloc_mutex irq_context: 0 &type->s_umount_key#13/1 pcpu_alloc_mutex pcpu_lock irq_context: 0 &type->s_umount_key#28 lock#3 rcu_read_lock &pool->lock &p->pi_lock irq_context: 0 &type->s_umount_key#13/1 shrinker_rwsem irq_context: 0 &type->s_umount_key#28 lock#3 rcu_read_lock &pool->lock &p->pi_lock &rq->__lock irq_context: 0 &type->s_umount_key#13/1 list_lrus_mutex irq_context: 0 &type->s_umount_key#28 lock#3 rcu_read_lock &pool->lock &p->pi_lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 &type->s_umount_key#13/1 sb_lock irq_context: 0 &type->s_umount_key#13/1 sb_lock unnamed_dev_ida.xa_lock irq_context: 0 &type->s_umount_key#28 lock#3 &rq->__lock irq_context: 0 &type->s_umount_key#13/1 mmu_notifier_invalidate_range_start irq_context: 0 (wq_completion)mm_percpu_wq (work_completion)(work) irq_context: 0 (wq_completion)mm_percpu_wq (work_completion)(work) lock#4 irq_context: 0 (wq_completion)mm_percpu_wq (work_completion)(work) lock#4 &lruvec->lru_lock irq_context: 0 (wq_completion)mm_percpu_wq (work_completion)(work) lock#5 irq_context: 0 &type->s_umount_key#28 lock#3 (work_completion)(work) irq_context: 0 &type->s_umount_key#13/1 &sb->s_type->i_lock_key#12 irq_context: 0 &type->s_umount_key#28 shrinker_rwsem irq_context: 0 &type->s_umount_key#13/1 &s->s_inode_list_lock irq_context: 0 &type->s_umount_key#13/1 tk_core.seq.seqcount irq_context: 0 &type->s_umount_key#29/1 irq_context: 0 &type->s_umount_key#13/1 &sb->s_type->i_lock_key#12 &dentry->d_lock irq_context: 0 &type->s_umount_key#29/1 fs_reclaim irq_context: 0 &type->s_umount_key#13/1 &dentry->d_lock irq_context: 0 &type->s_umount_key#29/1 fs_reclaim mmu_notifier_invalidate_range_start irq_context: 0 &type->s_umount_key#29/1 pcpu_alloc_mutex irq_context: 0 &type->s_umount_key#29/1 pcpu_alloc_mutex pcpu_lock irq_context: 0 &type->s_umount_key#29/1 shrinker_rwsem irq_context: 0 &type->s_umount_key#29/1 list_lrus_mutex irq_context: 0 &sb->s_type->i_mutex_key#5 irq_context: 0 &type->s_umount_key#29/1 sb_lock irq_context: 0 &sb->s_type->i_mutex_key#5 &sb->s_type->i_lock_key#12 irq_context: 0 &type->s_umount_key#30 irq_context: 0 &sb->s_type->i_mutex_key#5 rename_lock.seqcount irq_context: 0 &sb->s_type->i_mutex_key#5 fs_reclaim irq_context: 0 &sb->s_type->i_mutex_key#5 fs_reclaim mmu_notifier_invalidate_range_start irq_context: 0 &type->s_umount_key#30 fs_reclaim irq_context: 0 &sb->s_type->i_mutex_key#5 pool_lock#2 irq_context: 0 &sb->s_type->i_mutex_key#5 &dentry->d_lock irq_context: 0 &sb->s_type->i_mutex_key#5 rcu_read_lock rename_lock.seqcount irq_context: 0 &type->s_umount_key#30 fs_reclaim mmu_notifier_invalidate_range_start irq_context: 0 &sb->s_type->i_mutex_key#5 &dentry->d_lock &wq irq_context: 0 &type->s_umount_key#30 mmu_notifier_invalidate_range_start irq_context: 0 &sb->s_type->i_mutex_key#5 mmu_notifier_invalidate_range_start irq_context: 0 &sb->s_type->i_mutex_key#5 &s->s_inode_list_lock irq_context: 0 &type->s_umount_key#30 &____s->seqcount irq_context: 0 &sb->s_type->i_mutex_key#5 tk_core.seq.seqcount irq_context: 0 &type->s_umount_key#30 &xa->xa_lock#9 irq_context: 0 &sb->s_type->i_mutex_key#5 &sb->s_type->i_lock_key#12 &dentry->d_lock irq_context: 0 &type->s_umount_key#30 lock#4 irq_context: 0 &sb->s_type->i_mutex_key#5 &c->lock irq_context: 0 &type->s_umount_key#30 &mapping->private_lock irq_context: 0 &sb->s_type->i_mutex_key#5 &pcp->lock &zone->lock irq_context: 0 &type->s_umount_key#30 tk_core.seq.seqcount irq_context: 0 &sb->s_type->i_mutex_key#5 &zone->lock irq_context: 0 &sb->s_type->i_mutex_key#5 &____s->seqcount irq_context: 0 &type->s_umount_key#30 &dd->lock irq_context: 0 &type->s_umount_key#30 &obj_hash[i].lock irq_context: 0 &type->s_umount_key#30 rcu_read_lock &pool->lock irq_context: 0 &type->s_umount_key#30 rcu_read_lock &pool->lock &obj_hash[i].lock irq_context: 0 &type->s_umount_key#30 rcu_read_lock &pool->lock &p->pi_lock irq_context: 0 &type->s_umount_key#30 rcu_read_lock &pool->lock &p->pi_lock &rq->__lock irq_context: 0 &type->s_umount_key#30 rcu_read_lock &pool->lock &p->pi_lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 &type->s_umount_key#30 bit_wait_table + i irq_context: 0 &sb->s_type->i_mutex_key#5 rcu_read_lock pool_lock#2 irq_context: 0 &sb->s_type->i_mutex_key#5 &obj_hash[i].lock irq_context: 0 &type->s_umount_key#30 &rq->__lock irq_context: 0 &type->s_umount_key#30 &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 &type->s_umount_key#30 &sb->s_type->i_lock_key#3 irq_context: 0 &type->s_umount_key#30 &sb->s_type->i_lock_key#3 &xa->xa_lock#9 irq_context: 0 &type->s_umount_key#30 lock#4 &lruvec->lru_lock irq_context: 0 &type->s_umount_key#30 lock#5 irq_context: 0 &type->s_umount_key#30 &lruvec->lru_lock irq_context: 0 &type->s_umount_key#30 &pcp->lock &zone->lock irq_context: 0 &type->s_umount_key#30 &zone->lock irq_context: 0 &type->s_umount_key#30 crypto_alg_sem irq_context: 0 &type->s_umount_key#30 lock#3 irq_context: 0 &type->s_umount_key#30 lock#3 &obj_hash[i].lock irq_context: 0 &type->s_umount_key#30 lock#3 rcu_read_lock &pool->lock irq_context: 0 &type->s_umount_key#30 lock#3 rcu_read_lock &pool->lock &obj_hash[i].lock irq_context: 0 &type->s_umount_key#30 lock#3 rcu_read_lock &pool->lock &p->pi_lock irq_context: 0 &type->s_umount_key#30 lock#3 rcu_read_lock &pool->lock &p->pi_lock &rq->__lock irq_context: 0 &type->s_umount_key#15/1 irq_context: 0 &type->s_umount_key#15/1 fs_reclaim irq_context: 0 &type->s_umount_key#15/1 fs_reclaim mmu_notifier_invalidate_range_start irq_context: 0 &type->s_umount_key#30 lock#3 rcu_read_lock &pool->lock &p->pi_lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 &type->s_umount_key#15/1 pool_lock#2 irq_context: 0 &type->s_umount_key#30 lock#3 &rq->__lock irq_context: 0 &type->s_umount_key#15/1 pcpu_alloc_mutex irq_context: 0 &type->s_umount_key#30 lock#3 (work_completion)(work) irq_context: 0 &type->s_umount_key#15/1 pcpu_alloc_mutex pcpu_lock irq_context: 0 &type->s_umount_key#30 shrinker_rwsem irq_context: 0 &type->s_umount_key#31/1 irq_context: 0 &type->s_umount_key#31/1 fs_reclaim irq_context: 0 &type->s_umount_key#15/1 shrinker_rwsem irq_context: 0 &type->s_umount_key#31/1 fs_reclaim mmu_notifier_invalidate_range_start irq_context: 0 &type->s_umount_key#31/1 pcpu_alloc_mutex irq_context: 0 &type->s_umount_key#15/1 list_lrus_mutex irq_context: 0 &type->s_umount_key#31/1 pcpu_alloc_mutex pcpu_lock irq_context: 0 &type->s_umount_key#31/1 shrinker_rwsem irq_context: 0 &type->s_umount_key#15/1 sb_lock irq_context: 0 &type->s_umount_key#15/1 sb_lock unnamed_dev_ida.xa_lock irq_context: 0 &type->s_umount_key#31/1 list_lrus_mutex irq_context: 0 &type->s_umount_key#31/1 sb_lock irq_context: 0 &type->s_umount_key#15/1 mmu_notifier_invalidate_range_start irq_context: 0 &type->s_umount_key#32 irq_context: 0 &type->s_umount_key#32 fs_reclaim irq_context: 0 &type->s_umount_key#32 fs_reclaim mmu_notifier_invalidate_range_start irq_context: 0 &type->s_umount_key#15/1 &sb->s_type->i_lock_key#13 irq_context: 0 &type->s_umount_key#15/1 &s->s_inode_list_lock irq_context: 0 &type->s_umount_key#15/1 tk_core.seq.seqcount irq_context: 0 &type->s_umount_key#15/1 &sb->s_type->i_lock_key#13 &dentry->d_lock irq_context: 0 &type->s_umount_key#15/1 &dentry->d_lock irq_context: 0 &type->s_umount_key#32 mmu_notifier_invalidate_range_start irq_context: 0 &type->s_umount_key#32 &____s->seqcount irq_context: 0 &type->s_umount_key#32 &xa->xa_lock#9 irq_context: 0 &type->s_umount_key#32 lock#4 irq_context: 0 &type->s_umount_key#16/1 irq_context: 0 &type->s_umount_key#16/1 fs_reclaim irq_context: 0 &type->s_umount_key#16/1 fs_reclaim mmu_notifier_invalidate_range_start irq_context: 0 &type->s_umount_key#16/1 pool_lock#2 irq_context: 0 &type->s_umount_key#16/1 pcpu_alloc_mutex irq_context: 0 &type->s_umount_key#16/1 pcpu_alloc_mutex pcpu_lock irq_context: 0 &type->s_umount_key#16/1 shrinker_rwsem irq_context: 0 &type->s_umount_key#16/1 list_lrus_mutex irq_context: 0 &type->s_umount_key#32 &mapping->private_lock irq_context: 0 &type->s_umount_key#16/1 sb_lock irq_context: 0 &type->s_umount_key#16/1 sb_lock unnamed_dev_ida.xa_lock irq_context: 0 &type->s_umount_key#32 tk_core.seq.seqcount irq_context: 0 &type->s_umount_key#16/1 mmu_notifier_invalidate_range_start irq_context: 0 &type->s_umount_key#32 &dd->lock irq_context: 0 &type->s_umount_key#32 &obj_hash[i].lock irq_context: 0 &type->s_umount_key#16/1 &sb->s_type->i_lock_key#14 irq_context: 0 &type->s_umount_key#16/1 &s->s_inode_list_lock irq_context: 0 &type->s_umount_key#16/1 tk_core.seq.seqcount irq_context: 0 &type->s_umount_key#16/1 &sb->s_type->i_lock_key#14 &dentry->d_lock irq_context: 0 &type->s_umount_key#16/1 &dentry->d_lock irq_context: 0 &s->s_inode_list_lock irq_context: 0 &type->s_umount_key#17/1 irq_context: 0 &type->s_umount_key#17/1 fs_reclaim irq_context: 0 &type->s_umount_key#17/1 fs_reclaim mmu_notifier_invalidate_range_start irq_context: 0 &type->s_umount_key#17/1 pool_lock#2 irq_context: 0 &type->s_umount_key#17/1 pcpu_alloc_mutex irq_context: 0 &type->s_umount_key#17/1 pcpu_alloc_mutex pcpu_lock irq_context: 0 &type->s_umount_key#17/1 shrinker_rwsem irq_context: 0 &type->s_umount_key#17/1 list_lrus_mutex irq_context: 0 &type->s_umount_key#17/1 sb_lock irq_context: 0 &type->s_umount_key#17/1 sb_lock unnamed_dev_ida.xa_lock irq_context: 0 &type->s_umount_key#17/1 mmu_notifier_invalidate_range_start irq_context: 0 &type->s_umount_key#17/1 &sb->s_type->i_lock_key#15 irq_context: 0 &type->s_umount_key#17/1 &s->s_inode_list_lock irq_context: 0 &type->s_umount_key#17/1 tk_core.seq.seqcount irq_context: 0 &type->s_umount_key#17/1 &sb->s_type->i_lock_key#15 &dentry->d_lock irq_context: 0 &type->s_umount_key#17/1 &dentry->d_lock irq_context: 0 &sb->s_type->i_lock_key#15 irq_context: 0 kclist_lock irq_context: 0 kclist_lock resource_lock irq_context: 0 kclist_lock fs_reclaim irq_context: 0 kclist_lock fs_reclaim mmu_notifier_invalidate_range_start irq_context: 0 kclist_lock pool_lock#2 irq_context: 0 &type->s_umount_key#18/1 irq_context: 0 &type->s_umount_key#18/1 fs_reclaim irq_context: 0 &type->s_umount_key#18/1 fs_reclaim mmu_notifier_invalidate_range_start irq_context: 0 &type->s_umount_key#18/1 pool_lock#2 irq_context: 0 &type->s_umount_key#18/1 pcpu_alloc_mutex irq_context: 0 &type->s_umount_key#18/1 pcpu_alloc_mutex pcpu_lock irq_context: 0 &type->s_umount_key#18/1 shrinker_rwsem irq_context: 0 &type->s_umount_key#18/1 &c->lock irq_context: 0 &type->s_umount_key#18/1 &____s->seqcount irq_context: 0 &type->s_umount_key#18/1 list_lrus_mutex irq_context: 0 &type->s_umount_key#18/1 sb_lock irq_context: 0 &type->s_umount_key#18/1 sb_lock unnamed_dev_ida.xa_lock irq_context: 0 &type->s_umount_key#18/1 mmu_notifier_invalidate_range_start irq_context: 0 &type->s_umount_key#18/1 &sb->s_type->i_lock_key#16 irq_context: 0 &type->s_umount_key#18/1 &s->s_inode_list_lock irq_context: 0 &type->s_umount_key#18/1 tk_core.seq.seqcount irq_context: 0 &type->s_umount_key#18/1 &sb->s_type->i_lock_key#16 &dentry->d_lock irq_context: 0 &type->s_umount_key#18/1 &dentry->d_lock irq_context: 0 misc_mtx &pcp->lock &zone->lock irq_context: 0 misc_mtx &zone->lock irq_context: 0 &sb->s_type->i_mutex_key#2 &c->lock irq_context: 0 &sb->s_type->i_mutex_key#2 &pcp->lock &zone->lock irq_context: 0 &sb->s_type->i_mutex_key#2 &zone->lock irq_context: 0 &sb->s_type->i_mutex_key#2 &____s->seqcount irq_context: 0 tomoyo_ss irq_context: 0 tomoyo_ss mmu_notifier_invalidate_range_start irq_context: 0 tomoyo_ss pool_lock#2 irq_context: 0 tomoyo_ss tomoyo_policy_lock irq_context: 0 tomoyo_ss tomoyo_policy_lock mmu_notifier_invalidate_range_start irq_context: 0 tomoyo_ss tomoyo_policy_lock pool_lock#2 irq_context: 0 tomoyo_ss (console_sem).lock irq_context: 0 tomoyo_ss console_lock console_srcu console_owner_lock irq_context: 0 tomoyo_ss console_lock console_srcu console_owner irq_context: 0 tomoyo_ss console_lock console_srcu console_owner &port_lock_key irq_context: 0 tomoyo_ss console_lock console_srcu console_owner console_owner_lock irq_context: 0 &type->s_umount_key#19/1 irq_context: 0 &type->s_umount_key#19/1 fs_reclaim irq_context: 0 &type->s_umount_key#19/1 fs_reclaim mmu_notifier_invalidate_range_start irq_context: 0 &type->s_umount_key#19/1 pool_lock#2 irq_context: 0 &type->s_umount_key#19/1 pcpu_alloc_mutex irq_context: 0 &type->s_umount_key#19/1 pcpu_alloc_mutex pcpu_lock irq_context: 0 &type->s_umount_key#19/1 shrinker_rwsem irq_context: 0 &type->s_umount_key#19/1 list_lrus_mutex irq_context: 0 &type->s_umount_key#19/1 sb_lock irq_context: 0 &type->s_umount_key#19/1 sb_lock unnamed_dev_ida.xa_lock irq_context: 0 &type->s_umount_key#19/1 mmu_notifier_invalidate_range_start irq_context: 0 &type->s_umount_key#19/1 &sb->s_type->i_lock_key#17 irq_context: 0 &type->s_umount_key#19/1 &s->s_inode_list_lock irq_context: 0 &type->s_umount_key#19/1 tk_core.seq.seqcount irq_context: 0 &type->s_umount_key#19/1 &sb->s_type->i_lock_key#17 &dentry->d_lock irq_context: 0 &type->s_umount_key#19/1 &dentry->d_lock irq_context: 0 &ns->lock irq_context: 0 &ns->lock &dentry->d_lock irq_context: 0 &ns->lock pin_fs_lock irq_context: 0 &ns->lock &sb->s_type->i_mutex_key#6 irq_context: 0 &ns->lock &sb->s_type->i_mutex_key#6 &sb->s_type->i_lock_key#17 irq_context: 0 &ns->lock &sb->s_type->i_mutex_key#6 rename_lock.seqcount irq_context: 0 &ns->lock &sb->s_type->i_mutex_key#6 fs_reclaim irq_context: 0 &ns->lock &sb->s_type->i_mutex_key#6 fs_reclaim mmu_notifier_invalidate_range_start irq_context: 0 &ns->lock &sb->s_type->i_mutex_key#6 &____s->seqcount irq_context: 0 &ns->lock &sb->s_type->i_mutex_key#6 &pcp->lock &zone->lock irq_context: 0 &ns->lock &sb->s_type->i_mutex_key#6 &zone->lock irq_context: 0 &ns->lock &sb->s_type->i_mutex_key#6 pool_lock#2 irq_context: 0 &ns->lock &sb->s_type->i_mutex_key#6 rcu_read_lock pool_lock#2 irq_context: 0 &ns->lock &sb->s_type->i_mutex_key#6 &obj_hash[i].lock irq_context: 0 &ns->lock &sb->s_type->i_mutex_key#6 &dentry->d_lock irq_context: 0 &ns->lock &sb->s_type->i_mutex_key#6 rcu_read_lock rename_lock.seqcount irq_context: 0 &ns->lock &sb->s_type->i_mutex_key#6 &dentry->d_lock &wq irq_context: 0 &ns->lock &sb->s_type->i_mutex_key#6 mmu_notifier_invalidate_range_start irq_context: 0 &ns->lock &sb->s_type->i_mutex_key#6 &s->s_inode_list_lock irq_context: 0 &ns->lock &sb->s_type->i_mutex_key#6 tk_core.seq.seqcount irq_context: 0 &ns->lock &sb->s_type->i_mutex_key#6 &sb->s_type->i_lock_key#17 &dentry->d_lock irq_context: 0 &ns->lock &sb->s_type->i_mutex_key#6 &c->lock irq_context: 0 &type->s_umount_key#20 irq_context: 0 &type->s_umount_key#20 sb_lock irq_context: 0 &type->s_umount_key#20 &dentry->d_lock irq_context: 0 &sb->s_type->i_mutex_key#2 rcu_read_lock &dentry->d_lock irq_context: 0 pnp_lock irq_context: 0 pnp_lock fs_reclaim irq_context: 0 pnp_lock fs_reclaim mmu_notifier_invalidate_range_start irq_context: 0 pnp_lock pool_lock#2 irq_context: 0 &device->physical_node_lock sysfs_symlink_target_lock irq_context: 0 &device->physical_node_lock fs_reclaim irq_context: 0 &device->physical_node_lock fs_reclaim mmu_notifier_invalidate_range_start irq_context: 0 &device->physical_node_lock &____s->seqcount irq_context: 0 &device->physical_node_lock &pcp->lock &zone->lock irq_context: 0 &device->physical_node_lock &zone->lock irq_context: 0 &device->physical_node_lock pool_lock#2 irq_context: 0 &device->physical_node_lock rcu_read_lock pool_lock#2 irq_context: 0 &device->physical_node_lock &obj_hash[i].lock irq_context: 0 &device->physical_node_lock lock irq_context: 0 &device->physical_node_lock lock kernfs_idr_lock irq_context: 0 &device->physical_node_lock &root->kernfs_rwsem irq_context: 0 &device->physical_node_lock &root->kernfs_rwsem &root->kernfs_iattr_rwsem irq_context: 0 &device->physical_node_lock lock kernfs_idr_lock pool_lock#2 irq_context: 0 fwnode_link_lock irq_context: 0 fwnode_link_lock &k->k_lock irq_context: 0 subsys mutex#19 irq_context: softirq rcu_callback pcpu_lock irq_context: 0 &device->physical_node_lock &c->lock irq_context: 0 subsys mutex#20 irq_context: 0 subsys mutex#20 &k->k_lock irq_context: 0 subsys mutex#21 irq_context: 0 subsys mutex#21 &k->k_lock irq_context: 0 subsys mutex#22 irq_context: 0 subsys mutex#22 &k->k_lock irq_context: 0 tty_mutex irq_context: 0 sb_writers &type->i_mutex_dir_key/1 &simple_offset_xa_lock &c->lock irq_context: 0 sb_writers &type->i_mutex_dir_key/1 &simple_offset_xa_lock &pcp->lock &zone->lock irq_context: 0 sb_writers &type->i_mutex_dir_key/1 &simple_offset_xa_lock &zone->lock irq_context: 0 sb_writers &type->i_mutex_dir_key/1 &simple_offset_xa_lock &____s->seqcount irq_context: softirq led_lock irq_context: 0 misc_mtx &obj_hash[i].lock pool_lock irq_context: 0 misc_mtx fill_pool_map-wait-type-override pool_lock#2 irq_context: 0 misc_mtx fill_pool_map-wait-type-override pool_lock irq_context: 0 subsys mutex#23 irq_context: 0 subsys mutex#23 &k->list_lock irq_context: 0 subsys mutex#23 &k->k_lock irq_context: 0 cpu_hotplug_lock wq_pool_mutex &xa->xa_lock irq_context: 0 cpu_hotplug_lock wq_pool_mutex &n->list_lock irq_context: 0 cpu_hotplug_lock wq_pool_mutex kthread_create_lock irq_context: 0 cpu_hotplug_lock wq_pool_mutex &p->pi_lock irq_context: 0 cpu_hotplug_lock wq_pool_mutex &p->pi_lock &rq->__lock irq_context: 0 cpu_hotplug_lock wq_pool_mutex &p->pi_lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 cpu_hotplug_lock wq_pool_mutex &x->wait irq_context: 0 cpu_hotplug_lock wq_pool_mutex &rq->__lock irq_context: 0 cpu_hotplug_lock wq_pool_mutex &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 cpu_hotplug_lock wq_pool_mutex wq_pool_attach_mutex irq_context: 0 cpu_hotplug_lock wq_pool_mutex &pool->lock/1 &p->pi_lock irq_context: 0 cpu_hotplug_lock wq_pool_mutex &pool->lock/1 &p->pi_lock &rq->__lock irq_context: 0 cpu_hotplug_lock wq_pool_mutex &pool->lock/1 &p->pi_lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 netevent_notif_chain.lock irq_context: 0 clients_rwsem irq_context: 0 clients_rwsem fs_reclaim irq_context: 0 clients_rwsem fs_reclaim mmu_notifier_invalidate_range_start irq_context: 0 clients_rwsem clients.xa_lock irq_context: 0 devices_rwsem irq_context: 0 clients_rwsem clients.xa_lock pool_lock#2 irq_context: 0 (blocking_lsm_notifier_chain).rwsem irq_context: 0 (inetaddr_chain).rwsem irq_context: 0 inet6addr_chain.lock irq_context: 0 jiffies_seq.seqcount irq_context: 0 buses_mutex irq_context: 0 offload_lock irq_context: 0 inetsw_lock irq_context: 0 rcu_read_lock &pool->lock fill_pool_map-wait-type-override pool_lock#2 irq_context: 0 rcu_read_lock &pool->lock fill_pool_map-wait-type-override pool_lock irq_context: 0 (wq_completion)events_power_efficient irq_context: 0 (wq_completion)events_power_efficient (work_completion)(&(&tbl->managed_work)->work) irq_context: 0 (wq_completion)events_power_efficient (work_completion)(&(&tbl->managed_work)->work) &tbl->lock irq_context: 0 (wq_completion)events_power_efficient (work_completion)(&(&tbl->managed_work)->work) &tbl->lock &obj_hash[i].lock irq_context: 0 (wq_completion)events_power_efficient (work_completion)(&(&tbl->managed_work)->work) &tbl->lock &base->lock irq_context: 0 (wq_completion)events_power_efficient (work_completion)(&(&tbl->managed_work)->work) &tbl->lock &base->lock &obj_hash[i].lock irq_context: 0 ptype_lock irq_context: 0 pernet_ops_rwsem nl_table_lock irq_context: 0 pernet_ops_rwsem rtnl_mutex batched_entropy_u32.lock irq_context: 0 pernet_ops_rwsem rtnl_mutex &tbl->lock irq_context: 0 pernet_ops_rwsem rtnl_mutex sysctl_lock irq_context: 0 rtnl_mutex &rq->__lock irq_context: 0 (wq_completion)events_power_efficient (check_lifetime_work).work irq_context: 0 (wq_completion)events_power_efficient (check_lifetime_work).work &obj_hash[i].lock irq_context: 0 (wq_completion)events_power_efficient (check_lifetime_work).work &base->lock irq_context: 0 (wq_completion)events_power_efficient (check_lifetime_work).work &base->lock &obj_hash[i].lock irq_context: 0 rtnl_mutex batched_entropy_u8.lock irq_context: 0 rtnl_mutex kfence_freelist_lock irq_context: 0 pernet_ops_rwsem &net->rules_mod_lock irq_context: 0 pernet_ops_rwsem slab_mutex irq_context: 0 pernet_ops_rwsem slab_mutex fs_reclaim irq_context: 0 pernet_ops_rwsem slab_mutex fs_reclaim mmu_notifier_invalidate_range_start irq_context: 0 pernet_ops_rwsem slab_mutex pool_lock#2 irq_context: 0 pernet_ops_rwsem slab_mutex &c->lock irq_context: 0 pernet_ops_rwsem slab_mutex &n->list_lock irq_context: 0 pernet_ops_rwsem slab_mutex &____s->seqcount irq_context: 0 pernet_ops_rwsem slab_mutex pcpu_alloc_mutex irq_context: 0 pernet_ops_rwsem slab_mutex pcpu_alloc_mutex pcpu_lock irq_context: 0 pernet_ops_rwsem batched_entropy_u32.lock irq_context: 0 tcp_ulp_list_lock irq_context: 0 xfrm_state_afinfo_lock irq_context: 0 xfrm_policy_afinfo_lock irq_context: 0 xfrm_input_afinfo_lock irq_context: 0 pernet_ops_rwsem percpu_counters_lock irq_context: 0 rtnl_mutex &obj_hash[i].lock irq_context: 0 rtnl_mutex krc.lock irq_context: 0 rtnl_mutex krc.lock &obj_hash[i].lock irq_context: 0 rtnl_mutex krc.lock hrtimer_bases.lock irq_context: 0 rtnl_mutex krc.lock hrtimer_bases.lock tk_core.seq.seqcount irq_context: 0 rtnl_mutex krc.lock hrtimer_bases.lock &obj_hash[i].lock irq_context: 0 rtnl_mutex krc.lock &base->lock irq_context: 0 rtnl_mutex krc.lock &base->lock &obj_hash[i].lock irq_context: hardirq rcu_read_lock &pool->lock &p->pi_lock &cfs_rq->removed.lock irq_context: 0 pernet_ops_rwsem &rq->__lock irq_context: 0 (wq_completion)events_highpri irq_context: 0 (wq_completion)events_highpri (work_completion)(&(&krcp->page_cache_work)->work) irq_context: 0 (wq_completion)events_highpri (work_completion)(&(&krcp->page_cache_work)->work) fs_reclaim irq_context: 0 (wq_completion)events_highpri (work_completion)(&(&krcp->page_cache_work)->work) fs_reclaim mmu_notifier_invalidate_range_start irq_context: 0 (wq_completion)events_highpri (work_completion)(&(&krcp->page_cache_work)->work) &pcp->lock &zone->lock irq_context: 0 (wq_completion)events_highpri (work_completion)(&(&krcp->page_cache_work)->work) &zone->lock irq_context: 0 (wq_completion)events_highpri (work_completion)(&(&krcp->page_cache_work)->work) &pcp->lock &zone->lock &____s->seqcount irq_context: 0 (wq_completion)events_highpri (work_completion)(&(&krcp->page_cache_work)->work) &____s->seqcount irq_context: 0 (wq_completion)events_highpri (work_completion)(&(&krcp->page_cache_work)->work) pool_lock#2 irq_context: 0 (wq_completion)events_highpri (work_completion)(&(&krcp->page_cache_work)->work) krc.lock irq_context: 0 &hashinfo->lock irq_context: 0 tcp_cong_list_lock irq_context: 0 mptcp_sched_list_lock irq_context: 0 pernet_ops_rwsem rcu_read_lock pool_lock#2 irq_context: 0 pernet_ops_rwsem cache_list_lock irq_context: 0 pernet_ops_rwsem rcu_read_lock &pool->lock irq_context: 0 pernet_ops_rwsem rcu_read_lock &pool->lock &obj_hash[i].lock irq_context: 0 pernet_ops_rwsem rcu_read_lock &pool->lock &p->pi_lock irq_context: 0 pernet_ops_rwsem rcu_read_lock &pool->lock &p->pi_lock &rq->__lock irq_context: 0 pernet_ops_rwsem rcu_read_lock &pool->lock &p->pi_lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 pernet_ops_rwsem rcu_node_0 irq_context: 0 (wq_completion)events_power_efficient (work_completion)(&(&cache_cleaner)->work) irq_context: 0 (wq_completion)events_power_efficient (work_completion)(&(&cache_cleaner)->work) cache_list_lock irq_context: 0 (wq_completion)events_power_efficient (work_completion)(&(&cache_cleaner)->work) &obj_hash[i].lock irq_context: 0 (wq_completion)events_power_efficient (work_completion)(&(&cache_cleaner)->work) &base->lock irq_context: 0 (wq_completion)events_power_efficient (work_completion)(&(&cache_cleaner)->work) &base->lock &obj_hash[i].lock irq_context: 0 pernet_ops_rwsem rcu_read_lock rcu_node_0 irq_context: 0 pernet_ops_rwsem &obj_hash[i].lock pool_lock irq_context: 0 pernet_ops_rwsem fill_pool_map-wait-type-override pool_lock#2 irq_context: 0 pernet_ops_rwsem fill_pool_map-wait-type-override &c->lock irq_context: 0 pernet_ops_rwsem fill_pool_map-wait-type-override &____s->seqcount irq_context: 0 pernet_ops_rwsem fill_pool_map-wait-type-override pool_lock irq_context: 0 (rpc_pipefs_notifier_list).rwsem irq_context: 0 svc_xprt_class_lock irq_context: 0 xprt_list_lock irq_context: 0 xprt_list_lock (console_sem).lock irq_context: 0 xprt_list_lock console_lock console_srcu console_owner_lock irq_context: 0 xprt_list_lock console_lock console_srcu console_owner irq_context: 0 xprt_list_lock console_lock console_srcu console_owner &port_lock_key irq_context: 0 xprt_list_lock console_lock console_srcu console_owner console_owner_lock irq_context: 0 &(&priv->bus_notifier)->rwsem irq_context: 0 pcibios_fwaddrmap_lock irq_context: 0 (wq_completion)events_unbound (work_completion)(&entry->work) fs_reclaim irq_context: 0 (wq_completion)events_unbound (work_completion)(&entry->work) fs_reclaim mmu_notifier_invalidate_range_start irq_context: 0 (wq_completion)events_unbound (work_completion)(&entry->work) pool_lock#2 irq_context: 0 (wq_completion)events_unbound (work_completion)(&entry->work) &pcp->lock &zone->lock irq_context: 0 (wq_completion)events_unbound (work_completion)(&entry->work) &zone->lock irq_context: 0 (wq_completion)events_unbound (work_completion)(&entry->work) &____s->seqcount irq_context: 0 (wq_completion)events_unbound (work_completion)(&entry->work) rcu_read_lock pool_lock#2 irq_context: 0 (wq_completion)events_unbound (work_completion)(&entry->work) &obj_hash[i].lock irq_context: 0 (wq_completion)events_unbound (work_completion)(&entry->work) rcu_read_lock init_fs.seq.seqcount irq_context: 0 (wq_completion)events_unbound (work_completion)(&entry->work) rcu_read_lock &dentry->d_lock irq_context: 0 (wq_completion)events_unbound (work_completion)(&entry->work) &sb->s_type->i_mutex_key irq_context: 0 (wq_completion)events_unbound (work_completion)(&entry->work) &sb->s_type->i_mutex_key fs_reclaim irq_context: 0 (wq_completion)events_unbound (work_completion)(&entry->work) &sb->s_type->i_mutex_key fs_reclaim mmu_notifier_invalidate_range_start irq_context: 0 (wq_completion)events_unbound (work_completion)(&entry->work) &sb->s_type->i_mutex_key pool_lock#2 irq_context: 0 (wq_completion)events_unbound (work_completion)(&entry->work) &sb->s_type->i_mutex_key &dentry->d_lock irq_context: 0 (wq_completion)events_unbound (work_completion)(&entry->work) &sb->s_type->i_mutex_key rcu_read_lock rename_lock.seqcount irq_context: 0 (wq_completion)events_unbound (work_completion)(&entry->work) &sb->s_type->i_mutex_key &dentry->d_lock &wq irq_context: 0 (wq_completion)events_unbound (work_completion)(&entry->work) &dentry->d_lock irq_context: 0 (wq_completion)events_unbound (work_completion)(&entry->work) sb_writers#2 irq_context: 0 (wq_completion)events_unbound (work_completion)(&entry->work) sb_writers#2 mount_lock irq_context: 0 (wq_completion)events_unbound (work_completion)(&entry->work) sb_writers#2 &sb->s_type->i_mutex_key/1 irq_context: 0 (wq_completion)events_unbound (work_completion)(&entry->work) sb_writers#2 &sb->s_type->i_mutex_key/1 rename_lock.seqcount irq_context: 0 (wq_completion)events_unbound (work_completion)(&entry->work) sb_writers#2 &sb->s_type->i_mutex_key/1 fs_reclaim irq_context: 0 (wq_completion)events_unbound (work_completion)(&entry->work) sb_writers#2 &sb->s_type->i_mutex_key/1 fs_reclaim mmu_notifier_invalidate_range_start irq_context: 0 (wq_completion)events_unbound (work_completion)(&entry->work) sb_writers#2 &sb->s_type->i_mutex_key/1 pool_lock#2 irq_context: 0 (wq_completion)events_unbound (work_completion)(&entry->work) sb_writers#2 &sb->s_type->i_mutex_key/1 &dentry->d_lock irq_context: 0 (wq_completion)events_unbound (work_completion)(&entry->work) sb_writers#2 &sb->s_type->i_mutex_key/1 &obj_hash[i].lock irq_context: 0 (wq_completion)events_unbound (work_completion)(&entry->work) sb_writers#2 &sb->s_type->i_mutex_key/1 tomoyo_ss irq_context: 0 (wq_completion)events_unbound (work_completion)(&entry->work) sb_writers#2 &sb->s_type->i_mutex_key/1 tomoyo_ss mmu_notifier_invalidate_range_start irq_context: 0 (wq_completion)events_unbound (work_completion)(&entry->work) sb_writers#2 &sb->s_type->i_mutex_key/1 tomoyo_ss &c->lock irq_context: 0 (wq_completion)events_unbound (work_completion)(&entry->work) sb_writers#2 &sb->s_type->i_mutex_key/1 tomoyo_ss &pcp->lock &zone->lock irq_context: 0 (wq_completion)events_unbound (work_completion)(&entry->work) sb_writers#2 &sb->s_type->i_mutex_key/1 tomoyo_ss &zone->lock irq_context: 0 (wq_completion)events_unbound (work_completion)(&entry->work) sb_writers#2 &sb->s_type->i_mutex_key/1 tomoyo_ss &pcp->lock &zone->lock &____s->seqcount irq_context: 0 (wq_completion)events_unbound (work_completion)(&entry->work) sb_writers#2 &sb->s_type->i_mutex_key/1 tomoyo_ss &____s->seqcount irq_context: 0 (wq_completion)events_unbound (work_completion)(&entry->work) sb_writers#2 &sb->s_type->i_mutex_key/1 tomoyo_ss pool_lock#2 irq_context: 0 (wq_completion)events_unbound (work_completion)(&entry->work) sb_writers#2 &sb->s_type->i_mutex_key/1 tomoyo_ss rcu_read_lock mount_lock.seqcount irq_context: 0 (wq_completion)events_unbound (work_completion)(&entry->work) sb_writers#2 &sb->s_type->i_mutex_key/1 tomoyo_ss rcu_read_lock rcu_read_lock rename_lock.seqcount irq_context: 0 (wq_completion)events_unbound (work_completion)(&entry->work) sb_writers#2 &sb->s_type->i_mutex_key/1 tomoyo_ss &obj_hash[i].lock irq_context: 0 (wq_completion)events_unbound (work_completion)(&entry->work) sb_writers#2 &sb->s_type->i_mutex_key/1 tomoyo_ss rcu_read_lock &dentry->d_lock irq_context: 0 (wq_completion)events_unbound (work_completion)(&entry->work) sb_writers#2 &sb->s_type->i_mutex_key/1 tomoyo_ss tomoyo_log_lock irq_context: 0 (wq_completion)events_unbound (work_completion)(&entry->work) sb_writers#2 &sb->s_type->i_mutex_key/1 tomoyo_ss tomoyo_log_wait.lock irq_context: 0 (wq_completion)events_unbound (work_completion)(&entry->work) sb_writers#2 &sb->s_type->i_mutex_key/1 tomoyo_ss quarantine_lock irq_context: 0 (wq_completion)events_unbound (work_completion)(&entry->work) sb_writers#2 &sb->s_type->i_mutex_key/1 tomoyo_ss tomoyo_policy_lock irq_context: 0 (wq_completion)events_unbound (work_completion)(&entry->work) sb_writers#2 &sb->s_type->i_mutex_key/1 tomoyo_ss tomoyo_policy_lock mmu_notifier_invalidate_range_start irq_context: 0 (wq_completion)events_unbound (work_completion)(&entry->work) sb_writers#2 &sb->s_type->i_mutex_key/1 tomoyo_ss tomoyo_policy_lock pool_lock#2 irq_context: 0 (wq_completion)events_unbound (work_completion)(&entry->work) sb_writers#2 &sb->s_type->i_mutex_key/1 &c->lock irq_context: 0 (wq_completion)events_unbound (work_completion)(&entry->work) sb_writers#2 &sb->s_type->i_mutex_key/1 &pcp->lock &zone->lock irq_context: 0 (wq_completion)events_unbound (work_completion)(&entry->work) sb_writers#2 &sb->s_type->i_mutex_key/1 &zone->lock irq_context: 0 (wq_completion)events_unbound (work_completion)(&entry->work) sb_writers#2 &sb->s_type->i_mutex_key/1 &____s->seqcount irq_context: 0 (wq_completion)events_unbound (work_completion)(&entry->work) sb_writers#2 &sb->s_type->i_mutex_key/1 mmu_notifier_invalidate_range_start irq_context: 0 (wq_completion)events_unbound (work_completion)(&entry->work) sb_writers#2 &sb->s_type->i_mutex_key/1 &sb->s_type->i_lock_key#2 irq_context: 0 (wq_completion)events_unbound (work_completion)(&entry->work) sb_writers#2 &sb->s_type->i_mutex_key/1 &s->s_inode_list_lock irq_context: 0 (wq_completion)events_unbound (work_completion)(&entry->work) sb_writers#2 &sb->s_type->i_mutex_key/1 tk_core.seq.seqcount irq_context: 0 (wq_completion)events_unbound (work_completion)(&entry->work) sb_writers#2 &sb->s_type->i_mutex_key/1 &sb->s_type->i_lock_key#2 &dentry->d_lock irq_context: 0 (wq_completion)events_unbound (work_completion)(&entry->work) sb_writers#2 &sb->s_type->i_mutex_key/1 rcu_read_lock &dentry->d_lock irq_context: 0 (wq_completion)events_unbound (work_completion)(&entry->work) sb_writers#2 &sb->s_type->i_mutex_key irq_context: 0 (wq_completion)events_unbound (work_completion)(&entry->work) sb_writers#2 &sb->s_type->i_mutex_key tomoyo_ss irq_context: 0 (wq_completion)events_unbound (work_completion)(&entry->work) sb_writers#2 &sb->s_type->i_mutex_key tomoyo_ss mmu_notifier_invalidate_range_start irq_context: 0 (wq_completion)events_unbound (work_completion)(&entry->work) sb_writers#2 &sb->s_type->i_mutex_key tomoyo_ss pool_lock#2 irq_context: 0 (wq_completion)events_unbound (work_completion)(&entry->work) sb_writers#2 &sb->s_type->i_mutex_key tomoyo_ss rcu_read_lock mount_lock.seqcount irq_context: 0 (wq_completion)events_unbound (work_completion)(&entry->work) sb_writers#2 &sb->s_type->i_mutex_key tomoyo_ss rcu_read_lock rcu_read_lock rename_lock.seqcount irq_context: 0 (wq_completion)events_unbound (work_completion)(&entry->work) sb_writers#2 &sb->s_type->i_mutex_key tomoyo_ss &obj_hash[i].lock irq_context: 0 (wq_completion)events_unbound (work_completion)(&entry->work) sb_writers#2 &sb->s_type->i_mutex_key tomoyo_ss &c->lock irq_context: 0 (wq_completion)events_unbound (work_completion)(&entry->work) sb_writers#2 &sb->s_type->i_mutex_key tomoyo_ss &pcp->lock &zone->lock irq_context: 0 (wq_completion)events_unbound (work_completion)(&entry->work) sb_writers#2 &sb->s_type->i_mutex_key tomoyo_ss &zone->lock irq_context: 0 (wq_completion)events_unbound (work_completion)(&entry->work) sb_writers#2 &sb->s_type->i_mutex_key tomoyo_ss &____s->seqcount irq_context: 0 (wq_completion)events_unbound (work_completion)(&entry->work) sb_writers#2 &sb->s_type->i_mutex_key tomoyo_ss rcu_read_lock &dentry->d_lock irq_context: 0 (wq_completion)events_unbound (work_completion)(&entry->work) sb_writers#2 &sb->s_type->i_mutex_key tomoyo_ss tomoyo_log_lock irq_context: 0 (wq_completion)events_unbound (work_completion)(&entry->work) sb_writers#2 &sb->s_type->i_mutex_key tomoyo_ss tomoyo_log_wait.lock irq_context: 0 (wq_completion)events_unbound (work_completion)(&entry->work) sb_writers#2 &sb->s_type->i_mutex_key tomoyo_ss tomoyo_policy_lock irq_context: 0 (wq_completion)events_unbound (work_completion)(&entry->work) sb_writers#2 &sb->s_type->i_mutex_key tomoyo_ss tomoyo_policy_lock mmu_notifier_invalidate_range_start irq_context: 0 (wq_completion)events_unbound (work_completion)(&entry->work) sb_writers#2 &sb->s_type->i_mutex_key tomoyo_ss tomoyo_policy_lock pool_lock#2 irq_context: 0 (wq_completion)events_unbound (work_completion)(&entry->work) sb_writers#2 &sb->s_type->i_mutex_key tk_core.seq.seqcount irq_context: 0 (wq_completion)events_unbound (work_completion)(&entry->work) sb_writers#2 &sb->s_type->i_mutex_key &sb->s_type->i_lock_key#2 irq_context: 0 (wq_completion)events_unbound (work_completion)(&entry->work) sb_writers#2 &sb->s_type->i_mutex_key &wb->list_lock irq_context: 0 (wq_completion)events_unbound (work_completion)(&entry->work) sb_writers#2 &sb->s_type->i_mutex_key &wb->list_lock &sb->s_type->i_lock_key#2 irq_context: 0 (wq_completion)events_unbound (work_completion)(&entry->work) rcu_read_lock &sb->s_type->i_lock_key#2 irq_context: 0 (wq_completion)events_unbound (work_completion)(&entry->work) sb_writers#2 &sb->s_type->i_mutex_key/1 tomoyo_ss rcu_read_lock pool_lock#2 irq_context: 0 (wq_completion)events_unbound (work_completion)(&entry->work) &c->lock irq_context: 0 umhelper_sem irq_context: 0 umhelper_sem usermodehelper_disabled_waitq.lock irq_context: 0 &drv->dynids.lock irq_context: 0 cpu_hotplug_lock cpuhp_state_mutex &p->pi_lock irq_context: 0 cpu_hotplug_lock cpuhp_state_mutex &p->pi_lock &rq->__lock irq_context: 0 cpu_hotplug_lock cpuhp_state_mutex &p->pi_lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 cpu_hotplug_lock cpuhp_state_mutex &x->wait#6 irq_context: 0 cpu_hotplug_lock cpuhp_state_mutex &rq->__lock irq_context: 0 cpu_hotplug_lock cpuhp_state-up fs_reclaim irq_context: 0 cpu_hotplug_lock cpuhp_state_mutex &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 cpu_hotplug_lock cpuhp_state-up fs_reclaim mmu_notifier_invalidate_range_start irq_context: 0 cpu_hotplug_lock cpuhp_state-up &c->lock irq_context: 0 cpu_hotplug_lock cpuhp_state-up &n->list_lock irq_context: 0 cpu_hotplug_lock cpuhp_state-up &pcp->lock &zone->lock irq_context: 0 cpu_hotplug_lock cpuhp_state-up &zone->lock irq_context: 0 cpu_hotplug_lock cpuhp_state-up &____s->seqcount irq_context: 0 cpu_hotplug_lock cpuhp_state-up pool_lock#2 irq_context: 0 &x->wait#6 &p->pi_lock &cfs_rq->removed.lock irq_context: 0 &x->wait#6 &p->pi_lock &rq->__lock irq_context: 0 &x->wait#6 &p->pi_lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)events_unbound (work_completion)(&sub_info->work) irq_context: 0 (wq_completion)events_unbound (work_completion)(&sub_info->work) &sighand->siglock irq_context: 0 (wq_completion)events_unbound (work_completion)(&sub_info->work) fs_reclaim irq_context: 0 (wq_completion)events_unbound (work_completion)(&sub_info->work) fs_reclaim mmu_notifier_invalidate_range_start irq_context: 0 (wq_completion)events_unbound (work_completion)(&sub_info->work) pool_lock#2 irq_context: 0 (wq_completion)events_unbound (work_completion)(&sub_info->work) free_vmap_area_lock irq_context: 0 (wq_completion)events_unbound (work_completion)(&sub_info->work) vmap_area_lock irq_context: 0 (wq_completion)events_unbound (work_completion)(&sub_info->work) &pcp->lock &zone->lock irq_context: 0 (wq_completion)events_unbound (work_completion)(&sub_info->work) &zone->lock irq_context: 0 (wq_completion)events_unbound (work_completion)(&sub_info->work) &____s->seqcount irq_context: 0 (wq_completion)events_unbound (work_completion)(&sub_info->work) init_mm.page_table_lock irq_context: 0 (wq_completion)events_unbound (work_completion)(&sub_info->work) batched_entropy_u64.lock irq_context: 0 (wq_completion)events_unbound (work_completion)(&sub_info->work) &obj_hash[i].lock irq_context: 0 (wq_completion)events_unbound (work_completion)(&sub_info->work) &obj_hash[i].lock pool_lock irq_context: 0 (wq_completion)events_unbound (work_completion)(&sub_info->work) &c->lock irq_context: 0 (wq_completion)events_unbound (work_completion)(&sub_info->work) init_files.file_lock irq_context: 0 (wq_completion)events_unbound (work_completion)(&sub_info->work) init_fs.lock irq_context: 0 (wq_completion)events_unbound (work_completion)(&sub_info->work) init_fs.lock &dentry->d_lock irq_context: 0 (wq_completion)events_unbound (work_completion)(&sub_info->work) fill_pool_map-wait-type-override pool_lock#2 irq_context: 0 (wq_completion)events_unbound (work_completion)(&sub_info->work) fill_pool_map-wait-type-override pool_lock irq_context: 0 (wq_completion)events_unbound (work_completion)(&sub_info->work) &p->alloc_lock irq_context: 0 (wq_completion)events_unbound (work_completion)(&sub_info->work) lock irq_context: 0 (wq_completion)events_unbound (work_completion)(&sub_info->work) lock pidmap_lock irq_context: 0 (wq_completion)events_unbound (work_completion)(&sub_info->work) pidmap_lock irq_context: 0 (wq_completion)events_unbound (work_completion)(&sub_info->work) cgroup_threadgroup_rwsem irq_context: 0 (wq_completion)events_unbound (work_completion)(&sub_info->work) cgroup_threadgroup_rwsem css_set_lock irq_context: 0 (wq_completion)events_unbound (work_completion)(&sub_info->work) cgroup_threadgroup_rwsem &p->pi_lock irq_context: 0 (wq_completion)events_unbound (work_completion)(&sub_info->work) cgroup_threadgroup_rwsem &p->pi_lock &rq->__lock irq_context: 0 (wq_completion)events_unbound (work_completion)(&sub_info->work) cgroup_threadgroup_rwsem &p->pi_lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)events_unbound (work_completion)(&sub_info->work) cgroup_threadgroup_rwsem tk_core.seq.seqcount irq_context: 0 (wq_completion)events_unbound (work_completion)(&sub_info->work) cgroup_threadgroup_rwsem tasklist_lock irq_context: 0 (wq_completion)events_unbound (work_completion)(&sub_info->work) cgroup_threadgroup_rwsem tasklist_lock &p->pi_lock irq_context: 0 (wq_completion)events_unbound (work_completion)(&sub_info->work) cgroup_threadgroup_rwsem tasklist_lock &sighand->siglock irq_context: 0 (wq_completion)events_unbound (work_completion)(&sub_info->work) cgroup_threadgroup_rwsem css_set_lock &p->pi_lock irq_context: 0 (wq_completion)events_unbound (work_completion)(&sub_info->work) cgroup_threadgroup_rwsem css_set_lock &p->pi_lock &rq->__lock irq_context: 0 (wq_completion)events_unbound (work_completion)(&sub_info->work) input_pool.lock irq_context: 0 (wq_completion)events_unbound (work_completion)(&sub_info->work) &p->pi_lock irq_context: 0 (wq_completion)events_unbound (work_completion)(&sub_info->work) &p->pi_lock &rq->__lock irq_context: 0 (wq_completion)events_unbound (work_completion)(&sub_info->work) &p->pi_lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 umh_sysctl_lock irq_context: 0 &mm->mmap_lock irq_context: 0 &mm->mmap_lock fs_reclaim irq_context: 0 &mm->mmap_lock fs_reclaim mmu_notifier_invalidate_range_start irq_context: 0 &mm->mmap_lock &pcp->lock &zone->lock irq_context: 0 &mm->mmap_lock &zone->lock irq_context: 0 &mm->mmap_lock &____s->seqcount irq_context: 0 &mm->mmap_lock pool_lock#2 irq_context: 0 &mm->mmap_lock &c->lock irq_context: 0 &mm->mmap_lock fs_reclaim irq_context: 0 &mm->mmap_lock fs_reclaim mmu_notifier_invalidate_range_start irq_context: 0 &mm->mmap_lock &pcp->lock &zone->lock irq_context: 0 &mm->mmap_lock &zone->lock irq_context: 0 &mm->mmap_lock &____s->seqcount irq_context: 0 &mm->mmap_lock pool_lock#2 irq_context: 0 &mm->mmap_lock &mm->page_table_lock irq_context: 0 &mm->mmap_lock ptlock_ptr(page) irq_context: 0 &mm->mmap_lock &c->lock irq_context: 0 &mm->mmap_lock &anon_vma->rwsem irq_context: 0 &mm->mmap_lock &anon_vma->rwsem &mm->page_table_lock irq_context: 0 &mm->mmap_lock &pcp->lock &zone->lock &____s->seqcount irq_context: 0 &mm->mmap_lock per_cpu_ptr(&cgroup_rstat_cpu_lock, cpu) irq_context: 0 &mm->mmap_lock ptlock_ptr(page)#2 irq_context: 0 &mm->mmap_lock ptlock_ptr(page)#2 lock#4 irq_context: 0 &sig->cred_guard_mutex irq_context: 0 &sig->cred_guard_mutex fs_reclaim irq_context: 0 &sig->cred_guard_mutex fs_reclaim mmu_notifier_invalidate_range_start irq_context: 0 &sig->cred_guard_mutex pool_lock#2 irq_context: 0 &sig->cred_guard_mutex &fs->lock irq_context: 0 &sig->cred_guard_mutex &pcp->lock &zone->lock irq_context: 0 &sig->cred_guard_mutex &zone->lock irq_context: 0 &sig->cred_guard_mutex &____s->seqcount irq_context: 0 &sig->cred_guard_mutex &c->lock irq_context: 0 &sig->cred_guard_mutex rcu_read_lock &____s->seqcount#3 irq_context: 0 &sig->cred_guard_mutex rcu_read_lock &dentry->d_lock irq_context: 0 &sig->cred_guard_mutex &sb->s_type->i_mutex_key irq_context: 0 &sig->cred_guard_mutex &sb->s_type->i_mutex_key fs_reclaim irq_context: 0 &sig->cred_guard_mutex &sb->s_type->i_mutex_key fs_reclaim mmu_notifier_invalidate_range_start irq_context: 0 &sig->cred_guard_mutex &sb->s_type->i_mutex_key pool_lock#2 irq_context: 0 &sig->cred_guard_mutex &sb->s_type->i_mutex_key &dentry->d_lock irq_context: 0 &sig->cred_guard_mutex &sb->s_type->i_mutex_key rcu_read_lock rename_lock.seqcount irq_context: 0 &sig->cred_guard_mutex &sb->s_type->i_mutex_key &dentry->d_lock &wq irq_context: 0 &sig->cred_guard_mutex &dentry->d_lock irq_context: 0 &sig->cred_guard_mutex &obj_hash[i].lock irq_context: 0 &sig->cred_guard_mutex delayed_uprobe_lock irq_context: 0 &sig->cred_guard_mutex &mm->mmap_lock irq_context: 0 &sig->cred_guard_mutex &mm->mmap_lock lock#4 irq_context: 0 &sig->cred_guard_mutex &mm->mmap_lock lock#4 &lruvec->lru_lock irq_context: 0 &sig->cred_guard_mutex &mm->mmap_lock lock#5 irq_context: 0 &sig->cred_guard_mutex &mm->mmap_lock mmu_notifier_invalidate_range_start irq_context: 0 &sig->cred_guard_mutex &mm->mmap_lock ptlock_ptr(page)#2 irq_context: 0 &sig->cred_guard_mutex &mm->mmap_lock irq_context: 0 &sig->cred_guard_mutex &mm->mmap_lock &vma->vm_lock->lock irq_context: 0 &sig->cred_guard_mutex &mm->mmap_lock &anon_vma->rwsem irq_context: 0 &sig->cred_guard_mutex &mm->mmap_lock &obj_hash[i].lock irq_context: 0 &sig->cred_guard_mutex &mm->mmap_lock pool_lock#2 irq_context: 0 &sig->cred_guard_mutex &mm->mmap_lock lock#4 irq_context: 0 &sig->cred_guard_mutex &mm->mmap_lock lock#5 irq_context: 0 &sig->cred_guard_mutex &mm->mmap_lock &lruvec->lru_lock irq_context: 0 &sig->cred_guard_mutex &mm->mmap_lock rcu_read_lock pool_lock#2 irq_context: 0 &sig->cred_guard_mutex pgd_lock irq_context: 0 &sig->cred_guard_mutex rcu_read_lock pool_lock#2 irq_context: 0 &sig->cred_guard_mutex key irq_context: 0 &sig->cred_guard_mutex pcpu_lock irq_context: 0 &sig->cred_guard_mutex percpu_counters_lock irq_context: 0 &tsk->futex_exit_mutex irq_context: 0 &tsk->futex_exit_mutex &p->pi_lock irq_context: 0 &p->alloc_lock &fs->lock irq_context: 0 &child->perf_event_mutex irq_context: 0 css_set_lock irq_context: 0 tasklist_lock irq_context: 0 tasklist_lock &pid->wait_pidfd irq_context: 0 tasklist_lock &sighand->siglock irq_context: 0 tasklist_lock &sighand->siglock &sig->wait_chldexit irq_context: 0 tasklist_lock &sighand->siglock input_pool.lock irq_context: 0 tasklist_lock &sighand->siglock &(&sig->stats_lock)->lock irq_context: 0 tasklist_lock &sighand->siglock &(&sig->stats_lock)->lock &____s->seqcount#5 irq_context: 0 tasklist_lock &sighand->siglock &(&sig->stats_lock)->lock &____s->seqcount#5 pidmap_lock irq_context: 0 tasklist_lock &sighand->siglock &(&sig->stats_lock)->lock &____s->seqcount#5 &obj_hash[i].lock irq_context: 0 tasklist_lock &sighand->siglock &(&sig->stats_lock)->lock &____s->seqcount#5 pool_lock#2 irq_context: 0 tasklist_lock &obj_hash[i].lock irq_context: 0 low_water_lock irq_context: 0 low_water_lock (console_sem).lock irq_context: 0 low_water_lock console_lock console_srcu console_owner_lock irq_context: 0 low_water_lock console_lock console_srcu console_owner irq_context: 0 low_water_lock console_lock console_srcu console_owner &port_lock_key irq_context: 0 low_water_lock console_lock console_srcu console_owner console_owner_lock irq_context: 0 &cfs_rq->removed.lock irq_context: 0 &pool->lock/1 &base->lock irq_context: 0 &pool->lock/1 &base->lock &obj_hash[i].lock irq_context: 0 cpu_hotplug_lock cpuhp_state_mutex &cfs_rq->removed.lock irq_context: 0 cpu_hotplug_lock cpuhp_state_mutex &obj_hash[i].lock irq_context: 0 cpu_hotplug_lock cpuhp_state_mutex pool_lock#2 irq_context: 0 cpu_hotplug_lock cpuhp_state_mutex &p->pi_lock &rq->__lock &cfs_rq->removed.lock irq_context: 0 pmus_lock &obj_hash[i].lock pool_lock irq_context: 0 pmus_lock fill_pool_map-wait-type-override pool_lock#2 irq_context: 0 pmus_lock fill_pool_map-wait-type-override pool_lock irq_context: 0 vendor_module_lock irq_context: 0 vendor_module_lock slab_mutex irq_context: 0 vendor_module_lock slab_mutex fs_reclaim irq_context: 0 vendor_module_lock slab_mutex fs_reclaim mmu_notifier_invalidate_range_start irq_context: 0 vendor_module_lock slab_mutex pool_lock#2 irq_context: 0 vendor_module_lock slab_mutex &c->lock irq_context: 0 vendor_module_lock slab_mutex &n->list_lock irq_context: 0 vendor_module_lock slab_mutex pcpu_alloc_mutex irq_context: 0 vendor_module_lock slab_mutex pcpu_alloc_mutex pcpu_lock irq_context: 0 vendor_module_lock pcpu_alloc_mutex irq_context: 0 vendor_module_lock pcpu_alloc_mutex pcpu_lock irq_context: 0 vendor_module_lock slab_mutex &pcp->lock &zone->lock irq_context: 0 vendor_module_lock slab_mutex &zone->lock irq_context: 0 vendor_module_lock slab_mutex &____s->seqcount irq_context: 0 vendor_module_lock &obj_hash[i].lock irq_context: 0 vendor_module_lock percpu_counters_lock irq_context: 0 vendor_module_lock fs_reclaim irq_context: 0 vendor_module_lock fs_reclaim mmu_notifier_invalidate_range_start irq_context: 0 vendor_module_lock pool_lock#2 irq_context: 0 vendor_module_lock shrinker_rwsem irq_context: 0 vendor_module_lock &pcp->lock &zone->lock irq_context: 0 vendor_module_lock &pcp->lock &zone->lock &____s->seqcount irq_context: 0 vendor_module_lock &____s->seqcount irq_context: 0 vendor_module_lock &zone->lock irq_context: 0 vendor_module_lock cpu_hotplug_lock irq_context: 0 vendor_module_lock cpu_hotplug_lock static_call_mutex irq_context: 0 vendor_module_lock cpu_hotplug_lock static_call_mutex text_mutex irq_context: 0 vendor_module_lock cpu_hotplug_lock static_call_mutex text_mutex ptlock_ptr(page)#2 irq_context: softirq (&tcp_orphan_timer) irq_context: softirq (&tcp_orphan_timer) &obj_hash[i].lock irq_context: softirq (&tcp_orphan_timer) &base->lock irq_context: softirq (&tcp_orphan_timer) &base->lock &obj_hash[i].lock irq_context: softirq &(&cache_cleaner)->timer irq_context: softirq &(&cache_cleaner)->timer rcu_read_lock &pool->lock irq_context: softirq &(&cache_cleaner)->timer rcu_read_lock &pool->lock &obj_hash[i].lock irq_context: softirq &(&cache_cleaner)->timer rcu_read_lock &pool->lock &p->pi_lock irq_context: softirq &(&cache_cleaner)->timer rcu_read_lock &pool->lock &p->pi_lock &rq->__lock irq_context: softirq &(&cache_cleaner)->timer rcu_read_lock &pool->lock &p->pi_lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 vendor_module_lock cpu_hotplug_lock static_call_mutex text_mutex &rq->__lock irq_context: 0 (wq_completion)events_unbound (work_completion)(&(&kfence_timer)->work) cpu_hotplug_lock jump_label_mutex text_mutex text_mutex.wait_lock irq_context: 0 (wq_completion)events_unbound (work_completion)(&(&kfence_timer)->work) cpu_hotplug_lock jump_label_mutex text_mutex &rq->__lock irq_context: 0 (wq_completion)events_unbound (work_completion)(&(&kfence_timer)->work) cpu_hotplug_lock jump_label_mutex text_mutex &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 vendor_module_lock cpu_hotplug_lock static_call_mutex text_mutex.wait_lock irq_context: 0 vendor_module_lock cpu_hotplug_lock static_call_mutex &p->pi_lock irq_context: 0 vendor_module_lock cpu_hotplug_lock static_call_mutex &p->pi_lock &rq->__lock irq_context: 0 vendor_module_lock cpu_hotplug_lock static_call_mutex &p->pi_lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 vendor_module_lock cpu_hotplug_lock static_call_mutex &rq->__lock irq_context: 0 vendor_module_lock timekeeper_lock irq_context: 0 vendor_module_lock timekeeper_lock pvclock_gtod_data irq_context: 0 slab_mutex batched_entropy_u8.lock irq_context: 0 slab_mutex kfence_freelist_lock irq_context: 0 &sb->s_type->i_mutex_key#3 &pcp->lock &zone->lock &____s->seqcount irq_context: 0 misc_mtx lock kernfs_idr_lock pool_lock#2 irq_context: 0 misc_mtx rcu_read_lock &pool->lock/1 irq_context: 0 misc_mtx rcu_read_lock &pool->lock/1 &obj_hash[i].lock irq_context: 0 misc_mtx rcu_read_lock &pool->lock/1 pool_lock#2 irq_context: 0 misc_mtx rcu_read_lock &pool->lock/1 &p->pi_lock irq_context: 0 misc_mtx rcu_read_lock &pool->lock/1 &p->pi_lock &rq->__lock irq_context: 0 misc_mtx rcu_read_lock &pool->lock/1 &p->pi_lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 &type->s_umount_key#32 rcu_read_lock &pool->lock irq_context: 0 &type->s_umount_key#32 rcu_read_lock &pool->lock &obj_hash[i].lock irq_context: 0 &type->s_umount_key#32 rcu_read_lock &pool->lock &p->pi_lock irq_context: hardirq timekeeper_lock tk_core.seq.seqcount pvclock_gtod_data irq_context: 0 &type->s_umount_key#32 rcu_read_lock &pool->lock &p->pi_lock &rq->__lock irq_context: 0 cpu_hotplug_lock cpuhp_state-up &x->wait#9 irq_context: 0 cpu_hotplug_lock cpuhp_state-up &k->list_lock irq_context: 0 cpu_hotplug_lock cpuhp_state-up gdp_mutex irq_context: 0 cpu_hotplug_lock cpuhp_state-up gdp_mutex &k->list_lock irq_context: 0 cpu_hotplug_lock cpuhp_state-up gdp_mutex fs_reclaim irq_context: 0 cpu_hotplug_lock cpuhp_state-up gdp_mutex fs_reclaim mmu_notifier_invalidate_range_start irq_context: 0 cpu_hotplug_lock cpuhp_state-up gdp_mutex pool_lock#2 irq_context: 0 cpu_hotplug_lock cpuhp_state-up gdp_mutex lock irq_context: 0 cpu_hotplug_lock cpuhp_state-up gdp_mutex lock kernfs_idr_lock irq_context: 0 cpu_hotplug_lock cpuhp_state-up gdp_mutex &root->kernfs_rwsem irq_context: 0 cpu_hotplug_lock cpuhp_state-up gdp_mutex &root->kernfs_rwsem &root->kernfs_iattr_rwsem irq_context: 0 cpu_hotplug_lock cpuhp_state-up lock irq_context: 0 cpu_hotplug_lock cpuhp_state-up lock kernfs_idr_lock irq_context: 0 cpu_hotplug_lock cpuhp_state-up &root->kernfs_rwsem irq_context: 0 cpu_hotplug_lock cpuhp_state-up &root->kernfs_rwsem &root->kernfs_iattr_rwsem irq_context: 0 cpu_hotplug_lock cpuhp_state-up bus_type_sem irq_context: 0 cpu_hotplug_lock cpuhp_state-up sysfs_symlink_target_lock irq_context: 0 cpu_hotplug_lock cpuhp_state-up &root->kernfs_rwsem irq_context: 0 cpu_hotplug_lock cpuhp_state-up &dev->power.lock irq_context: 0 cpu_hotplug_lock cpuhp_state-up dpm_list_mtx irq_context: 0 cpu_hotplug_lock cpuhp_state-up req_lock irq_context: 0 cpu_hotplug_lock cpuhp_state-up &p->pi_lock irq_context: 0 cpu_hotplug_lock cpuhp_state-up &p->pi_lock &rq->__lock irq_context: 0 cpu_hotplug_lock cpuhp_state-up &p->pi_lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 &mm->mmap_lock rcu_read_lock rcu_node_0 irq_context: 0 cpu_hotplug_lock cpuhp_state-up &x->wait#11 irq_context: 0 &mm->mmap_lock rcu_read_lock &rq->__lock irq_context: 0 cpu_hotplug_lock cpuhp_state-up &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 &type->i_mutex_dir_key#2 irq_context: 0 &type->i_mutex_dir_key#2 fs_reclaim irq_context: 0 &type->i_mutex_dir_key#2 fs_reclaim mmu_notifier_invalidate_range_start irq_context: 0 &type->i_mutex_dir_key#2 pool_lock#2 irq_context: 0 &type->i_mutex_dir_key#2 &dentry->d_lock irq_context: 0 &type->i_mutex_dir_key#2 rcu_read_lock rename_lock.seqcount irq_context: 0 &type->i_mutex_dir_key#2 &dentry->d_lock &wq irq_context: 0 rcu_read_lock &sb->s_type->i_lock_key#5 irq_context: 0 cpu_hotplug_lock cpuhp_state-up uevent_sock_mutex irq_context: 0 cpu_hotplug_lock cpuhp_state-up rcu_read_lock &pool->lock/1 irq_context: 0 cpu_hotplug_lock cpuhp_state-up rcu_read_lock &pool->lock/1 &obj_hash[i].lock irq_context: 0 cpu_hotplug_lock cpuhp_state-up rcu_read_lock &pool->lock/1 pool_lock#2 irq_context: 0 cpu_hotplug_lock cpuhp_state-up rcu_read_lock &pool->lock/1 &p->pi_lock irq_context: 0 cpu_hotplug_lock cpuhp_state-up rcu_read_lock &pool->lock/1 &p->pi_lock &rq->__lock irq_context: 0 cpu_hotplug_lock cpuhp_state-up rcu_read_lock &pool->lock/1 &p->pi_lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 cpu_hotplug_lock cpuhp_state-up running_helpers_waitq.lock irq_context: 0 &sig->cred_guard_mutex &obj_hash[i].lock pool_lock irq_context: 0 cpu_hotplug_lock cpuhp_state-up subsys mutex#24 irq_context: 0 cpu_hotplug_lock cpuhp_state-up subsys mutex#24 &k->k_lock irq_context: 0 &sig->cred_guard_mutex &mm->mmap_lock &rq->__lock irq_context: 0 cpu_hotplug_lock cpuhp_state-up fill_pool_map-wait-type-override pool_lock#2 irq_context: 0 cpu_hotplug_lock cpuhp_state-up fill_pool_map-wait-type-override pool_lock irq_context: 0 cpu_hotplug_lock cpuhp_state-up rcu_read_lock pool_lock#2 irq_context: 0 &sig->cred_guard_mutex &rq->__lock irq_context: 0 fill_pool_map-wait-type-override &rq->__lock irq_context: 0 (wq_completion)events_unbound (work_completion)(&sub_info->work) &pcp->lock &zone->lock &____s->seqcount irq_context: 0 (wq_completion)events_unbound (work_completion)(&sub_info->work) rcu_read_lock rcu_node_0 irq_context: 0 (wq_completion)events_unbound (work_completion)(&sub_info->work) rcu_read_lock &rq->__lock irq_context: 0 cpu_hotplug_lock cpuhp_state-up subsys mutex#25 irq_context: 0 cpu_hotplug_lock cpuhp_state-up subsys mutex#25 &k->k_lock irq_context: 0 (wq_completion)events_unbound (work_completion)(&sub_info->work) lock pidmap_lock pool_lock#2 irq_context: 0 tasklist_lock &sighand->siglock &(&sig->stats_lock)->lock &____s->seqcount#5 pidmap_lock &obj_hash[i].lock irq_context: 0 tasklist_lock &sighand->siglock &(&sig->stats_lock)->lock &____s->seqcount#5 pidmap_lock pool_lock#2 irq_context: 0 (wq_completion)events_unbound (work_completion)(&sub_info->work) rcu_read_lock &cfs_rq->removed.lock irq_context: 0 (wq_completion)events_unbound (work_completion)(&sub_info->work) rcu_read_lock &obj_hash[i].lock irq_context: 0 (wq_completion)events_unbound (work_completion)(&sub_info->work) &p->pi_lock &rq->__lock &cfs_rq->removed.lock irq_context: 0 cpu_hotplug_lock cpuhp_state-up &cfs_rq->removed.lock irq_context: 0 cpu_hotplug_lock cpuhp_state-up rcu_read_lock &pool->lock/1 &p->pi_lock &rq->__lock &cfs_rq->removed.lock irq_context: softirq rcu_callback put_task_map-wait-type-override &obj_hash[i].lock irq_context: softirq rcu_callback put_task_map-wait-type-override pool_lock#2 irq_context: 0 &x->wait#6 &p->pi_lock &rq->__lock &cfs_rq->removed.lock irq_context: 0 crypto_alg_sem irq_context: 0 pm_qos_lock irq_context: 0 (wq_completion)events_unbound (work_completion)(&sub_info->work) batched_entropy_u64.lock crngs.lock irq_context: 0 &sig->cred_guard_mutex &sb->s_type->i_mutex_key &c->lock irq_context: 0 &sig->cred_guard_mutex &sb->s_type->i_mutex_key &pcp->lock &zone->lock irq_context: 0 &sig->cred_guard_mutex &sb->s_type->i_mutex_key &zone->lock irq_context: 0 &sig->cred_guard_mutex &sb->s_type->i_mutex_key &____s->seqcount irq_context: 0 misc_mtx &cfs_rq->removed.lock irq_context: 0 misc_mtx rcu_read_lock &pool->lock/1 &p->pi_lock &rq->__lock &cfs_rq->removed.lock irq_context: 0 rcu_read_lock &pool->lock/1 &p->pi_lock &rq->__lock &cfs_rq->removed.lock irq_context: 0 subsys mutex#26 irq_context: 0 subsys mutex#27 irq_context: 0 subsys mutex#27 &k->list_lock irq_context: 0 subsys mutex#27 &k->k_lock irq_context: 0 subsys mutex#28 irq_context: 0 trace_event_sem trace_event_sem.wait_lock irq_context: 0 trace_event_sem &rq->__lock irq_context: 0 trace_event_sem &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: softirq rcu_callback rcu_read_lock &pool->lock irq_context: softirq rcu_callback rcu_read_lock &pool->lock &obj_hash[i].lock irq_context: softirq rcu_callback rcu_read_lock &pool->lock &p->pi_lock irq_context: softirq rcu_callback rcu_read_lock &pool->lock &p->pi_lock &rq->__lock irq_context: softirq rcu_callback rcu_read_lock &pool->lock &p->pi_lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)events (work_completion)(&p->wq) irq_context: 0 (wq_completion)events (work_completion)(&p->wq) vmap_area_lock irq_context: softirq rcu_callback put_task_map-wait-type-override &obj_hash[i].lock pool_lock irq_context: 0 (wq_completion)events (work_completion)(&p->wq) &obj_hash[i].lock irq_context: 0 (wq_completion)events (work_completion)(&p->wq) purge_vmap_area_lock irq_context: 0 (wq_completion)events (work_completion)(&p->wq) rcu_read_lock pool_lock#2 irq_context: 0 (wq_completion)events (work_completion)(&p->wq) pool_lock#2 irq_context: softirq rcu_callback &obj_hash[i].lock pool_lock irq_context: softirq (&timer.timer) &p->pi_lock &cfs_rq->removed.lock irq_context: softirq &(&group->avgs_work)->timer irq_context: softirq &(&group->avgs_work)->timer rcu_read_lock &pool->lock irq_context: softirq &(&group->avgs_work)->timer rcu_read_lock &pool->lock &obj_hash[i].lock irq_context: softirq &(&group->avgs_work)->timer rcu_read_lock &pool->lock &p->pi_lock irq_context: softirq &(&group->avgs_work)->timer rcu_read_lock &pool->lock &p->pi_lock &rq->__lock irq_context: softirq &(&group->avgs_work)->timer rcu_read_lock &pool->lock &p->pi_lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: softirq &(({ do { const void *__vpp_verify = (typeof((&vmstat_work) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __asm__ ("" : "=r"(__ptr) : "0"((typeof(*((&vmstat_work))) *)((&vmstat_work)))); (typeof((typeof(*((&vmstat_work))) *)((&vmstat_work)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); }))->timer irq_context: softirq &(({ do { const void *__vpp_verify = (typeof((&vmstat_work) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __asm__ ("" : "=r"(__ptr) : "0"((typeof(*((&vmstat_work))) *)((&vmstat_work)))); (typeof((typeof(*((&vmstat_work))) *)((&vmstat_work)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); }))->timer rcu_read_lock &pool->lock irq_context: softirq &(({ do { const void *__vpp_verify = (typeof((&vmstat_work) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __asm__ ("" : "=r"(__ptr) : "0"((typeof(*((&vmstat_work))) *)((&vmstat_work)))); (typeof((typeof(*((&vmstat_work))) *)((&vmstat_work)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); }))->timer rcu_read_lock &pool->lock &obj_hash[i].lock irq_context: softirq &(({ do { const void *__vpp_verify = (typeof((&vmstat_work) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __asm__ ("" : "=r"(__ptr) : "0"((typeof(*((&vmstat_work))) *)((&vmstat_work)))); (typeof((typeof(*((&vmstat_work))) *)((&vmstat_work)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); }))->timer rcu_read_lock &pool->lock &p->pi_lock irq_context: softirq mm/memcontrol.c:589 irq_context: softirq mm/memcontrol.c:589 rcu_read_lock &pool->lock/1 irq_context: softirq mm/memcontrol.c:589 rcu_read_lock &pool->lock/1 &obj_hash[i].lock irq_context: softirq mm/memcontrol.c:589 rcu_read_lock &pool->lock/1 &p->pi_lock irq_context: hardirq &rq->__lock &obj_hash[i].lock irq_context: hardirq &rq->__lock &base->lock irq_context: hardirq &rq->__lock &base->lock &obj_hash[i].lock irq_context: 0 (wq_completion)events (work_completion)(&(&group->avgs_work)->work) irq_context: 0 (wq_completion)events (work_completion)(&(&group->avgs_work)->work) &group->avgs_lock irq_context: 0 (wq_completion)events_unbound (stats_flush_dwork).work irq_context: 0 (wq_completion)events (work_completion)(&(&group->avgs_work)->work) &group->avgs_lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)events_unbound (stats_flush_dwork).work cgroup_rstat_lock irq_context: 0 (wq_completion)events_unbound (stats_flush_dwork).work cgroup_rstat_lock per_cpu_ptr(&cgroup_rstat_cpu_lock, cpu) irq_context: 0 (wq_completion)events_unbound (stats_flush_dwork).work &obj_hash[i].lock irq_context: 0 (wq_completion)events_unbound (stats_flush_dwork).work &base->lock irq_context: 0 (wq_completion)events_unbound (stats_flush_dwork).work &base->lock &obj_hash[i].lock irq_context: softirq &(({ do { const void *__vpp_verify = (typeof((&vmstat_work) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __asm__ ("" : "=r"(__ptr) : "0"((typeof(*((&vmstat_work))) *)((&vmstat_work)))); (typeof((typeof(*((&vmstat_work))) *)((&vmstat_work)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); }))->timer rcu_read_lock &pool->lock &p->pi_lock &rq->__lock irq_context: softirq &(({ do { const void *__vpp_verify = (typeof((&vmstat_work) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __asm__ ("" : "=r"(__ptr) : "0"((typeof(*((&vmstat_work))) *)((&vmstat_work)))); (typeof((typeof(*((&vmstat_work))) *)((&vmstat_work)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); }))->timer rcu_read_lock &pool->lock &p->pi_lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: softirq mm/memcontrol.c:589 rcu_read_lock &pool->lock/1 &p->pi_lock &rq->__lock irq_context: softirq mm/memcontrol.c:589 rcu_read_lock &pool->lock/1 &p->pi_lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)events (work_completion)(&(&group->avgs_work)->work) &group->avgs_lock &obj_hash[i].lock irq_context: 0 (wq_completion)events (work_completion)(&(&group->avgs_work)->work) &group->avgs_lock &base->lock irq_context: 0 (wq_completion)events (work_completion)(&(&group->avgs_work)->work) &group->avgs_lock &base->lock &obj_hash[i].lock irq_context: 0 &type->s_umount_key#32 rcu_read_lock &pool->lock &p->pi_lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 &type->s_umount_key#32 bit_wait_table + i irq_context: 0 &type->s_umount_key#32 &rq->__lock irq_context: 0 &type->s_umount_key#32 &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 &type->s_umount_key#32 &sb->s_type->i_lock_key#3 irq_context: 0 &type->s_umount_key#32 &sb->s_type->i_lock_key#3 &xa->xa_lock#9 irq_context: 0 &type->s_umount_key#32 lock#4 &lruvec->lru_lock irq_context: 0 &type->s_umount_key#32 lock#5 irq_context: 0 &type->s_umount_key#32 &lruvec->lru_lock irq_context: 0 &type->s_umount_key#32 &pcp->lock &zone->lock irq_context: 0 &type->s_umount_key#32 &zone->lock irq_context: 0 &type->s_umount_key#32 &c->lock irq_context: 0 &type->s_umount_key#32 &pcp->lock &zone->lock &____s->seqcount irq_context: 0 &type->s_umount_key#32 pool_lock#2 irq_context: 0 &type->s_umount_key#32 crypto_alg_sem irq_context: 0 &type->s_umount_key#32 &xa->xa_lock#9 pool_lock#2 irq_context: 0 &type->s_umount_key#32 rcu_read_lock &pool->lock pool_lock#2 irq_context: 0 &type->s_umount_key#32 pcpu_alloc_mutex irq_context: 0 &type->s_umount_key#32 pcpu_alloc_mutex pcpu_lock irq_context: 0 &type->s_umount_key#32 percpu_counters_lock irq_context: 0 &type->s_umount_key#32 shrinker_rwsem irq_context: 0 &type->s_umount_key#32 inode_hash_lock irq_context: 0 &type->s_umount_key#32 inode_hash_lock &sb->s_type->i_lock_key#22 irq_context: 0 &type->s_umount_key#32 inode_hash_lock &s->s_inode_list_lock irq_context: 0 &type->s_umount_key#32 rcu_read_lock &dd->lock irq_context: 0 &type->s_umount_key#32 rcu_read_lock &obj_hash[i].lock irq_context: 0 &type->s_umount_key#32 rcu_read_lock &obj_hash[i].lock pool_lock irq_context: 0 &type->s_umount_key#32 rcu_read_lock pool_lock#2 irq_context: 0 &type->s_umount_key#32 rcu_read_lock tk_core.seq.seqcount irq_context: 0 &type->s_umount_key#32 rcu_read_lock &virtscsi_vq->vq_lock irq_context: 0 &type->s_umount_key#32 &sb->s_type->i_lock_key#22 irq_context: 0 &type->s_umount_key#32 &sb->s_type->i_mutex_key#8 irq_context: 0 &type->s_umount_key#32 &sb->s_type->i_mutex_key#8 &ei->i_es_lock irq_context: 0 &type->s_umount_key#32 &sb->s_type->i_mutex_key#8 &ei->i_data_sem irq_context: 0 &type->s_umount_key#32 &sb->s_type->i_mutex_key#8 &ei->i_data_sem mmu_notifier_invalidate_range_start irq_context: 0 &type->s_umount_key#32 &sb->s_type->i_mutex_key#8 &ei->i_data_sem pool_lock#2 irq_context: 0 &type->s_umount_key#32 &sb->s_type->i_mutex_key#8 &ei->i_data_sem &ei->i_es_lock irq_context: 0 &type->s_umount_key#32 &sb->s_type->i_mutex_key#8 &ei->i_data_sem &ei->i_es_lock &pcp->lock &zone->lock irq_context: 0 &type->s_umount_key#32 &sb->s_type->i_mutex_key#8 &ei->i_data_sem &ei->i_es_lock &zone->lock irq_context: 0 &type->s_umount_key#32 &sb->s_type->i_mutex_key#8 &ei->i_data_sem &ei->i_es_lock &____s->seqcount irq_context: 0 &type->s_umount_key#32 &sb->s_type->i_mutex_key#8 &ei->i_data_sem &ei->i_es_lock pool_lock#2 irq_context: 0 &type->s_umount_key#32 &sb->s_type->i_mutex_key#8 &ei->i_data_sem &ei->i_es_lock &c->lock irq_context: 0 &type->s_umount_key#32 &sb->s_type->i_mutex_key#8 &ei->i_data_sem &ei->i_es_lock &sbi->s_es_lock irq_context: 0 &type->s_umount_key#32 &sb->s_type->i_mutex_key#8 &ei->i_data_sem &obj_hash[i].lock irq_context: 0 &type->s_umount_key#32 proc_subdir_lock irq_context: softirq &(&kfence_timer)->timer rcu_read_lock &pool->lock/1 &p->pi_lock &cfs_rq->removed.lock irq_context: 0 &type->s_umount_key#32 proc_inum_ida.xa_lock irq_context: 0 &type->s_umount_key#32 proc_subdir_lock irq_context: 0 &type->s_umount_key#32 &journal->j_state_lock irq_context: 0 pmus_lock fs_reclaim irq_context: 0 pmus_lock fs_reclaim mmu_notifier_invalidate_range_start irq_context: 0 pmus_lock &k->list_lock irq_context: 0 pmus_lock lock irq_context: 0 pmus_lock lock kernfs_idr_lock irq_context: 0 pmus_lock &root->kernfs_rwsem irq_context: 0 pmus_lock &root->kernfs_rwsem &root->kernfs_iattr_rwsem irq_context: 0 pmus_lock uevent_sock_mutex irq_context: 0 pmus_lock rcu_read_lock &pool->lock/1 irq_context: 0 pmus_lock rcu_read_lock &pool->lock/1 &obj_hash[i].lock irq_context: 0 pmus_lock rcu_read_lock &pool->lock/1 pool_lock#2 irq_context: 0 pmus_lock rcu_read_lock &pool->lock/1 &p->pi_lock irq_context: 0 pmus_lock rcu_read_lock &pool->lock/1 &p->pi_lock &rq->__lock irq_context: 0 pmus_lock rcu_read_lock &pool->lock/1 &p->pi_lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 pmus_lock running_helpers_waitq.lock irq_context: 0 (wq_completion)events (work_completion)(&s->destroy_work) irq_context: 0 (wq_completion)events (work_completion)(&s->destroy_work) &rsp->gp_wait irq_context: 0 (wq_completion)events (work_completion)(&s->destroy_work) pcpu_lock irq_context: 0 (wq_completion)events (work_completion)(&s->destroy_work) &obj_hash[i].lock irq_context: 0 pmus_lock &pcp->lock &zone->lock irq_context: 0 pmus_lock &zone->lock irq_context: 0 pmus_lock &____s->seqcount irq_context: 0 pmus_lock rcu_read_lock pool_lock#2 irq_context: 0 pmus_lock &x->wait#9 irq_context: 0 pmus_lock bus_type_sem irq_context: 0 pmus_lock &c->lock irq_context: 0 pmus_lock lock kernfs_idr_lock pool_lock#2 irq_context: 0 pmus_lock sysfs_symlink_target_lock irq_context: 0 pmus_lock &k->k_lock irq_context: 0 pmus_lock &root->kernfs_rwsem irq_context: 0 pmus_lock &dev->power.lock irq_context: 0 pmus_lock dpm_list_mtx irq_context: 0 pmus_lock &dev->mutex &k->list_lock irq_context: 0 pmus_lock &dev->mutex &k->k_lock irq_context: 0 pmus_lock &dev->mutex &dev->power.lock irq_context: 0 pmus_lock subsys mutex#29 irq_context: 0 pmus_lock &pcp->lock &zone->lock &____s->seqcount irq_context: 0 &type->s_umount_key#32 kthread_create_lock irq_context: 0 &type->s_umount_key#32 &p->pi_lock irq_context: 0 &type->s_umount_key#32 &p->pi_lock &rq->__lock irq_context: 0 &type->s_umount_key#32 &p->pi_lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 &type->s_umount_key#32 &x->wait irq_context: 0 &type->s_umount_key#32 &journal->j_wait_done_commit irq_context: 0 &journal->j_wait_done_commit irq_context: 0 &journal->j_state_lock irq_context: 0 &journal->j_state_lock &journal->j_wait_done_commit irq_context: 0 &journal->j_state_lock &journal->j_wait_commit irq_context: 0 &type->s_umount_key#32 &journal->j_state_lock irq_context: 0 &type->s_umount_key#32 &p->alloc_lock irq_context: 0 &type->s_umount_key#32 cpu_hotplug_lock irq_context: 0 &type->s_umount_key#32 cpu_hotplug_lock wq_pool_mutex irq_context: 0 &type->s_umount_key#32 cpu_hotplug_lock wq_pool_mutex fs_reclaim irq_context: 0 &type->s_umount_key#32 cpu_hotplug_lock wq_pool_mutex fs_reclaim mmu_notifier_invalidate_range_start irq_context: 0 &type->s_umount_key#32 cpu_hotplug_lock wq_pool_mutex pool_lock#2 irq_context: 0 &type->s_umount_key#32 cpu_hotplug_lock wq_pool_mutex &obj_hash[i].lock irq_context: 0 &type->s_umount_key#32 cpu_hotplug_lock wq_pool_mutex &____s->seqcount irq_context: 0 &type->s_umount_key#32 cpu_hotplug_lock wq_pool_mutex &pcp->lock &zone->lock irq_context: 0 &sig->cred_guard_mutex &sb->s_type->i_mutex_key rcu_read_lock &dentry->d_lock irq_context: 0 &sig->cred_guard_mutex &sb->s_type->i_mutex_key &obj_hash[i].lock irq_context: 0 key_user_lock irq_context: 0 key_serial_lock irq_context: 0 key_construction_mutex irq_context: 0 &type->lock_class irq_context: 0 &type->lock_class keyring_serialise_link_lock irq_context: 0 &type->lock_class keyring_serialise_link_lock fs_reclaim irq_context: 0 &type->lock_class keyring_serialise_link_lock fs_reclaim mmu_notifier_invalidate_range_start irq_context: 0 &type->lock_class keyring_serialise_link_lock pool_lock#2 irq_context: 0 &type->lock_class keyring_serialise_link_lock &obj_hash[i].lock irq_context: 0 keyring_serialise_link_lock irq_context: 0 &pgdat->kswapd_lock fs_reclaim irq_context: 0 &pgdat->kswapd_lock fs_reclaim mmu_notifier_invalidate_range_start irq_context: 0 &pgdat->kswapd_lock pool_lock#2 irq_context: 0 &pgdat->kswapd_lock kthread_create_lock irq_context: 0 &pgdat->kswapd_lock &p->pi_lock irq_context: 0 &pgdat->kswapd_lock &p->pi_lock &cfs_rq->removed.lock irq_context: 0 &pgdat->kswapd_lock &p->pi_lock &rq->__lock irq_context: 0 &pgdat->kswapd_lock &p->pi_lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 &pgdat->kswapd_lock &p->pi_lock &rq->__lock &cfs_rq->removed.lock irq_context: 0 &pgdat->kswapd_lock &rq->__lock irq_context: 0 rcu_read_lock &rq->__lock &cfs_rq->removed.lock irq_context: 0 &pgdat->kswapd_lock &x->wait irq_context: 0 &pgdat->kswapd_lock &obj_hash[i].lock irq_context: 0 &pgdat->kswapd_lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 &pgdat->kswapd_wait irq_context: 0 list_lrus_mutex irq_context: 0 drivers_lock irq_context: 0 cpu_hotplug_lock cpuhp_state_mutex fs_reclaim irq_context: 0 cpu_hotplug_lock cpuhp_state_mutex fs_reclaim mmu_notifier_invalidate_range_start irq_context: 0 damon_dbgfs_lock irq_context: 0 damon_dbgfs_lock fs_reclaim irq_context: 0 damon_dbgfs_lock fs_reclaim mmu_notifier_invalidate_range_start irq_context: 0 damon_dbgfs_lock pool_lock#2 irq_context: 0 damon_dbgfs_lock tk_core.seq.seqcount irq_context: 0 damon_dbgfs_lock damon_ops_lock irq_context: 0 damon_dbgfs_lock pin_fs_lock irq_context: 0 damon_dbgfs_lock &sb->s_type->i_mutex_key#3 irq_context: 0 damon_dbgfs_lock &sb->s_type->i_mutex_key#3 rename_lock.seqcount irq_context: 0 damon_dbgfs_lock &sb->s_type->i_mutex_key#3 fs_reclaim irq_context: 0 damon_dbgfs_lock &sb->s_type->i_mutex_key#3 fs_reclaim mmu_notifier_invalidate_range_start irq_context: 0 damon_dbgfs_lock &sb->s_type->i_mutex_key#3 pool_lock#2 irq_context: 0 damon_dbgfs_lock &sb->s_type->i_mutex_key#3 &dentry->d_lock irq_context: 0 damon_dbgfs_lock &sb->s_type->i_mutex_key#3 rcu_read_lock rename_lock.seqcount irq_context: 0 damon_dbgfs_lock &sb->s_type->i_mutex_key#3 &dentry->d_lock &wq irq_context: 0 damon_dbgfs_lock &sb->s_type->i_mutex_key#3 mmu_notifier_invalidate_range_start irq_context: 0 damon_dbgfs_lock &sb->s_type->i_mutex_key#3 &sb->s_type->i_lock_key#7 irq_context: 0 damon_dbgfs_lock &sb->s_type->i_mutex_key#3 &s->s_inode_list_lock irq_context: 0 damon_dbgfs_lock &sb->s_type->i_mutex_key#3 tk_core.seq.seqcount irq_context: 0 damon_dbgfs_lock &sb->s_type->i_mutex_key#3 &sb->s_type->i_lock_key#7 &dentry->d_lock irq_context: 0 damon_dbgfs_lock &sb->s_type->i_mutex_key#3 &c->lock irq_context: 0 damon_dbgfs_lock &sb->s_type->i_mutex_key#3 &____s->seqcount irq_context: 0 misc_mtx &p->pi_lock &cfs_rq->removed.lock irq_context: 0 misc_mtx rcu_read_lock &pool->lock/1 &p->pi_lock &cfs_rq->removed.lock irq_context: 0 &type->s_umount_key#21/1 irq_context: 0 &type->s_umount_key#21/1 fs_reclaim irq_context: 0 &type->s_umount_key#21/1 fs_reclaim mmu_notifier_invalidate_range_start irq_context: 0 &type->s_umount_key#21/1 pool_lock#2 irq_context: 0 &type->s_umount_key#21/1 pcpu_alloc_mutex irq_context: 0 &type->s_umount_key#21/1 pcpu_alloc_mutex pcpu_lock irq_context: 0 &type->s_umount_key#21/1 shrinker_rwsem irq_context: 0 &type->s_umount_key#21/1 list_lrus_mutex irq_context: 0 &type->s_umount_key#21/1 sb_lock irq_context: 0 &type->s_umount_key#21/1 sb_lock unnamed_dev_ida.xa_lock irq_context: 0 &type->s_umount_key#21/1 &pcp->lock &zone->lock irq_context: 0 &type->s_umount_key#21/1 &zone->lock irq_context: 0 &type->s_umount_key#21/1 &____s->seqcount irq_context: 0 &type->s_umount_key#21/1 rcu_read_lock pool_lock#2 irq_context: 0 &type->s_umount_key#21/1 &obj_hash[i].lock irq_context: 0 &type->s_umount_key#21/1 mmu_notifier_invalidate_range_start irq_context: 0 &type->s_umount_key#21/1 &sb->s_type->i_lock_key#18 irq_context: 0 &type->s_umount_key#21/1 &s->s_inode_list_lock irq_context: 0 &type->s_umount_key#21/1 tk_core.seq.seqcount irq_context: 0 &type->s_umount_key#21/1 &sb->s_type->i_lock_key#18 &dentry->d_lock irq_context: 0 &type->s_umount_key#21/1 &dentry->d_lock irq_context: 0 dq_list_lock irq_context: 0 &type->s_umount_key#22/1 irq_context: 0 &type->s_umount_key#22/1 fs_reclaim irq_context: 0 &type->s_umount_key#22/1 fs_reclaim mmu_notifier_invalidate_range_start irq_context: 0 &type->s_umount_key#22/1 pool_lock#2 irq_context: 0 &type->s_umount_key#22/1 pcpu_alloc_mutex irq_context: 0 &type->s_umount_key#22/1 pcpu_alloc_mutex pcpu_lock irq_context: 0 &type->s_umount_key#22/1 shrinker_rwsem irq_context: 0 &type->s_umount_key#22/1 &c->lock irq_context: 0 &type->s_umount_key#22/1 &pcp->lock &zone->lock irq_context: 0 &type->s_umount_key#22/1 &pcp->lock &zone->lock &____s->seqcount irq_context: 0 &type->s_umount_key#22/1 &____s->seqcount irq_context: 0 &type->s_umount_key#22/1 list_lrus_mutex irq_context: 0 &type->s_umount_key#22/1 sb_lock irq_context: 0 &type->s_umount_key#22/1 sb_lock unnamed_dev_ida.xa_lock irq_context: 0 &type->s_umount_key#22/1 mmu_notifier_invalidate_range_start irq_context: 0 &type->s_umount_key#22/1 &sb->s_type->i_lock_key#19 irq_context: 0 &type->s_umount_key#22/1 &s->s_inode_list_lock irq_context: 0 &type->s_umount_key#22/1 tk_core.seq.seqcount irq_context: 0 &type->s_umount_key#22/1 &sb->s_type->i_lock_key#19 &dentry->d_lock irq_context: 0 &type->s_umount_key#22/1 &dentry->d_lock irq_context: 0 configfs_subsystem_mutex irq_context: 0 &sb->s_type->i_mutex_key#7/1 irq_context: 0 &sb->s_type->i_mutex_key#7/1 fs_reclaim irq_context: 0 &sb->s_type->i_mutex_key#7/1 fs_reclaim mmu_notifier_invalidate_range_start irq_context: 0 &sb->s_type->i_mutex_key#7/1 pool_lock#2 irq_context: 0 &sb->s_type->i_mutex_key#7/1 &dentry->d_lock irq_context: 0 &sb->s_type->i_mutex_key#7/1 &____s->seqcount irq_context: 0 &sb->s_type->i_mutex_key#7/1 &c->lock irq_context: 0 &sb->s_type->i_mutex_key#7/1 configfs_dirent_lock irq_context: 0 &sb->s_type->i_mutex_key#7/1 mmu_notifier_invalidate_range_start irq_context: 0 &sb->s_type->i_mutex_key#7/1 &sb->s_type->i_lock_key#19 irq_context: 0 &sb->s_type->i_mutex_key#7/1 &s->s_inode_list_lock irq_context: 0 &sb->s_type->i_mutex_key#7/1 tk_core.seq.seqcount irq_context: 0 &sb->s_type->i_mutex_key#7/1 &sb->s_type->i_lock_key#19 &dentry->d_lock irq_context: 0 &sb->s_type->i_mutex_key#7/1 &default_group_class[depth - 1]/2 irq_context: 0 sb_writers &type->i_mutex_dir_key/1 &pcp->lock &zone->lock &____s->seqcount irq_context: 0 misc_mtx rcu_read_lock pool_lock#2 irq_context: 0 ecryptfs_daemon_hash_mux irq_context: 0 ecryptfs_daemon_hash_mux fs_reclaim irq_context: 0 ecryptfs_daemon_hash_mux fs_reclaim mmu_notifier_invalidate_range_start irq_context: 0 ecryptfs_daemon_hash_mux &c->lock irq_context: 0 ecryptfs_daemon_hash_mux &pcp->lock &zone->lock irq_context: 0 ecryptfs_daemon_hash_mux &zone->lock irq_context: 0 ecryptfs_daemon_hash_mux &____s->seqcount irq_context: 0 ecryptfs_daemon_hash_mux pool_lock#2 irq_context: 0 ecryptfs_msg_ctx_lists_mux irq_context: 0 ecryptfs_msg_ctx_lists_mux &ecryptfs_msg_ctx_arr[i].mux irq_context: 0 &ecryptfs_kthread_ctl.wait irq_context: 0 pernet_ops_rwsem tk_core.seq.seqcount irq_context: 0 pernet_ops_rwsem &k->list_lock irq_context: 0 pernet_ops_rwsem lock irq_context: 0 pernet_ops_rwsem lock kernfs_idr_lock irq_context: 0 pernet_ops_rwsem &root->kernfs_rwsem irq_context: 0 pernet_ops_rwsem &root->kernfs_rwsem &root->kernfs_iattr_rwsem irq_context: 0 pernet_ops_rwsem rcu_read_lock &pool->lock/1 irq_context: 0 pernet_ops_rwsem rcu_read_lock &pool->lock/1 &obj_hash[i].lock irq_context: 0 pernet_ops_rwsem rcu_read_lock &pool->lock/1 pool_lock#2 irq_context: 0 pernet_ops_rwsem rcu_read_lock &pool->lock/1 &p->pi_lock irq_context: 0 pernet_ops_rwsem rcu_read_lock &pool->lock/1 &p->pi_lock &rq->__lock irq_context: 0 pernet_ops_rwsem rcu_read_lock &pool->lock/1 &p->pi_lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 pernet_ops_rwsem running_helpers_waitq.lock irq_context: 0 nfs_version_lock irq_context: 0 key_types_sem irq_context: 0 key_types_sem (console_sem).lock irq_context: 0 key_types_sem console_lock console_srcu console_owner_lock irq_context: 0 key_types_sem console_lock console_srcu console_owner irq_context: 0 key_types_sem console_lock console_srcu console_owner &port_lock_key irq_context: 0 key_types_sem console_lock console_srcu console_owner console_owner_lock irq_context: 0 pnfs_spinlock irq_context: 0 pernet_ops_rwsem &sn->pipefs_sb_lock irq_context: 0 pernet_ops_rwsem krc.lock irq_context: 0 pernet_ops_rwsem krc.lock &obj_hash[i].lock irq_context: 0 pernet_ops_rwsem krc.lock hrtimer_bases.lock irq_context: 0 pernet_ops_rwsem krc.lock hrtimer_bases.lock tk_core.seq.seqcount irq_context: 0 pernet_ops_rwsem krc.lock hrtimer_bases.lock &obj_hash[i].lock irq_context: 0 pernet_ops_rwsem krc.lock &base->lock irq_context: 0 pernet_ops_rwsem krc.lock &base->lock &obj_hash[i].lock irq_context: 0 pernet_ops_rwsem batched_entropy_u32.lock crngs.lock irq_context: 0 pernet_ops_rwsem &s->s_inode_list_lock irq_context: 0 nls_lock irq_context: hardirq hrtimer_bases.lock fill_pool_map-wait-type-override pool_lock#2 irq_context: hardirq hrtimer_bases.lock fill_pool_map-wait-type-override &c->lock irq_context: hardirq hrtimer_bases.lock fill_pool_map-wait-type-override &pcp->lock &zone->lock irq_context: hardirq hrtimer_bases.lock fill_pool_map-wait-type-override &zone->lock irq_context: hardirq hrtimer_bases.lock fill_pool_map-wait-type-override &____s->seqcount irq_context: hardirq hrtimer_bases.lock fill_pool_map-wait-type-override pool_lock irq_context: 0 jffs2_compressor_list_lock irq_context: 0 misc_mtx &pcp->lock &zone->lock &____s->seqcount irq_context: 0 next_tag_value_lock irq_context: 0 &sb->s_type->i_mutex_key#3 &rq->__lock irq_context: 0 log_redrive_lock irq_context: 0 &TxAnchor.LazyLock irq_context: 0 &TxAnchor.LazyLock jfs_commit_thread_wait.lock irq_context: 0 jfsTxnLock irq_context: 0 ocfs2_stack_lock irq_context: 0 ocfs2_stack_lock (console_sem).lock irq_context: 0 ocfs2_stack_lock console_lock console_srcu console_owner_lock irq_context: 0 ocfs2_stack_lock console_lock console_srcu console_owner irq_context: 0 ocfs2_stack_lock console_lock console_srcu console_owner &port_lock_key irq_context: 0 ocfs2_stack_lock console_lock console_srcu console_owner console_owner_lock irq_context: 0 &sig->cred_guard_mutex &mm->mmap_lock lock#4 rcu_read_lock pool_lock#2 irq_context: 0 &sig->cred_guard_mutex &mm->mmap_lock lock#4 &obj_hash[i].lock irq_context: 0 (wq_completion)events (work_completion)(&p->wq) rcu_read_lock rcu_node_0 irq_context: 0 (wq_completion)events (work_completion)(&p->wq) rcu_read_lock &rq->__lock irq_context: 0 o2hb_callback_sem irq_context: 0 o2net_handler_lock irq_context: 0 cpu_hotplug_lock wq_pool_mutex &obj_hash[i].lock pool_lock irq_context: 0 &type->s_umount_key#23/1 irq_context: 0 &type->s_umount_key#23/1 fs_reclaim irq_context: 0 &type->s_umount_key#23/1 fs_reclaim mmu_notifier_invalidate_range_start irq_context: 0 &type->s_umount_key#23/1 pool_lock#2 irq_context: 0 &type->s_umount_key#23/1 pcpu_alloc_mutex irq_context: 0 &type->s_umount_key#23/1 pcpu_alloc_mutex pcpu_lock irq_context: 0 &type->s_umount_key#23/1 shrinker_rwsem irq_context: 0 &type->s_umount_key#23/1 list_lrus_mutex irq_context: 0 &type->s_umount_key#23/1 sb_lock irq_context: 0 &type->s_umount_key#23/1 sb_lock unnamed_dev_ida.xa_lock irq_context: 0 &type->s_umount_key#23/1 &____s->seqcount irq_context: 0 &type->s_umount_key#23/1 &c->lock irq_context: 0 &type->s_umount_key#23/1 mmu_notifier_invalidate_range_start irq_context: 0 &type->s_umount_key#23/1 &sb->s_type->i_lock_key#20 irq_context: 0 &type->s_umount_key#23/1 &s->s_inode_list_lock irq_context: 0 &type->s_umount_key#23/1 tk_core.seq.seqcount irq_context: 0 &type->s_umount_key#23/1 &sb->s_type->i_lock_key#20 &dentry->d_lock irq_context: 0 &type->s_umount_key#23/1 &dentry->d_lock irq_context: 0 pernet_ops_rwsem nf_hook_mutex irq_context: 0 pernet_ops_rwsem nf_hook_mutex fs_reclaim irq_context: 0 pernet_ops_rwsem nf_hook_mutex fs_reclaim mmu_notifier_invalidate_range_start irq_context: 0 pernet_ops_rwsem nf_hook_mutex &____s->seqcount irq_context: 0 pernet_ops_rwsem nf_hook_mutex pool_lock#2 irq_context: 0 pernet_ops_rwsem nf_hook_mutex &c->lock irq_context: 0 pernet_ops_rwsem cpu_hotplug_lock irq_context: 0 pernet_ops_rwsem cpu_hotplug_lock jump_label_mutex irq_context: 0 pernet_ops_rwsem cpu_hotplug_lock jump_label_mutex text_mutex irq_context: 0 pernet_ops_rwsem cpu_hotplug_lock jump_label_mutex text_mutex ptlock_ptr(page)#2 irq_context: 0 alg_types_sem irq_context: 0 alg_types_sem fs_reclaim irq_context: 0 alg_types_sem fs_reclaim mmu_notifier_invalidate_range_start irq_context: 0 alg_types_sem pool_lock#2 irq_context: 0 dma_list_mutex irq_context: 0 asymmetric_key_parsers_sem irq_context: 0 asymmetric_key_parsers_sem (console_sem).lock irq_context: 0 asymmetric_key_parsers_sem console_lock console_srcu console_owner_lock irq_context: 0 asymmetric_key_parsers_sem console_lock console_srcu console_owner irq_context: 0 asymmetric_key_parsers_sem console_lock console_srcu console_owner &port_lock_key irq_context: 0 asymmetric_key_parsers_sem console_lock console_srcu console_owner console_owner_lock irq_context: 0 blkcg_pol_register_mutex irq_context: 0 blkcg_pol_register_mutex blkcg_pol_mutex irq_context: 0 blkcg_pol_register_mutex cgroup_mutex irq_context: 0 blkcg_pol_register_mutex cgroup_mutex &root->kernfs_rwsem irq_context: 0 blkcg_pol_register_mutex blkcg_pol_mutex fs_reclaim irq_context: 0 blkcg_pol_register_mutex blkcg_pol_mutex fs_reclaim mmu_notifier_invalidate_range_start irq_context: 0 blkcg_pol_register_mutex blkcg_pol_mutex pool_lock#2 irq_context: 0 blkcg_pol_register_mutex cgroup_mutex fs_reclaim irq_context: 0 blkcg_pol_register_mutex cgroup_mutex fs_reclaim mmu_notifier_invalidate_range_start irq_context: 0 blkcg_pol_register_mutex cgroup_mutex pool_lock#2 irq_context: 0 blkcg_pol_register_mutex cgroup_mutex lock irq_context: 0 blkcg_pol_register_mutex cgroup_mutex lock kernfs_idr_lock irq_context: 0 blkcg_pol_register_mutex cgroup_mutex &root->kernfs_rwsem &root->kernfs_iattr_rwsem irq_context: 0 elv_list_lock irq_context: softirq &rcu_state.gp_wq &p->pi_lock &rq->__lock &cfs_rq->removed.lock irq_context: 0 crc_t10dif_mutex irq_context: 0 crc_t10dif_mutex crypto_alg_sem irq_context: 0 crc_t10dif_mutex fs_reclaim irq_context: 0 crc_t10dif_mutex fs_reclaim mmu_notifier_invalidate_range_start irq_context: 0 crc_t10dif_mutex pool_lock#2 irq_context: 0 crc64_rocksoft_mutex irq_context: 0 crc64_rocksoft_mutex crypto_alg_sem irq_context: 0 crc64_rocksoft_mutex fs_reclaim irq_context: 0 crc64_rocksoft_mutex fs_reclaim mmu_notifier_invalidate_range_start irq_context: 0 crc64_rocksoft_mutex pool_lock#2 irq_context: 0 ts_mod_lock irq_context: 0 &sig->cred_guard_mutex fill_pool_map-wait-type-override pool_lock#2 irq_context: 0 &sig->cred_guard_mutex fill_pool_map-wait-type-override &c->lock irq_context: 0 &sig->cred_guard_mutex fill_pool_map-wait-type-override &pcp->lock &zone->lock irq_context: 0 &sig->cred_guard_mutex fill_pool_map-wait-type-override &zone->lock irq_context: 0 &sig->cred_guard_mutex fill_pool_map-wait-type-override &____s->seqcount irq_context: 0 &sig->cred_guard_mutex fill_pool_map-wait-type-override pool_lock irq_context: 0 &dev->mutex device_links_srcu irq_context: 0 &dev->mutex fwnode_link_lock irq_context: 0 &dev->mutex device_links_lock irq_context: 0 &dev->mutex sysfs_symlink_target_lock irq_context: 0 &dev->mutex fs_reclaim irq_context: 0 &dev->mutex fs_reclaim mmu_notifier_invalidate_range_start irq_context: 0 &dev->mutex pool_lock#2 irq_context: 0 &dev->mutex lock irq_context: 0 &dev->mutex lock kernfs_idr_lock irq_context: 0 &dev->mutex &root->kernfs_rwsem irq_context: 0 &dev->mutex &root->kernfs_rwsem &root->kernfs_iattr_rwsem irq_context: 0 &dev->mutex &c->lock irq_context: 0 &dev->mutex &pcp->lock &zone->lock irq_context: 0 &dev->mutex &zone->lock irq_context: 0 &dev->mutex &____s->seqcount irq_context: 0 &dev->mutex &x->wait#9 irq_context: 0 &dev->mutex &obj_hash[i].lock irq_context: 0 &dev->mutex gdp_mutex irq_context: 0 &dev->mutex gdp_mutex &k->list_lock irq_context: 0 &dev->mutex gdp_mutex fs_reclaim irq_context: 0 &dev->mutex gdp_mutex fs_reclaim mmu_notifier_invalidate_range_start irq_context: 0 &dev->mutex gdp_mutex pool_lock#2 irq_context: 0 &dev->mutex gdp_mutex lock irq_context: 0 &dev->mutex gdp_mutex lock kernfs_idr_lock irq_context: 0 &dev->mutex gdp_mutex &root->kernfs_rwsem irq_context: 0 &dev->mutex gdp_mutex &root->kernfs_rwsem &root->kernfs_iattr_rwsem irq_context: 0 &dev->mutex bus_type_sem irq_context: 0 &dev->mutex &root->kernfs_rwsem irq_context: 0 &dev->mutex dpm_list_mtx irq_context: 0 &dev->mutex uevent_sock_mutex irq_context: 0 &dev->mutex rcu_read_lock &pool->lock/1 irq_context: 0 &dev->mutex rcu_read_lock &pool->lock/1 &obj_hash[i].lock irq_context: 0 &dev->mutex rcu_read_lock &pool->lock/1 pool_lock#2 irq_context: 0 &dev->mutex rcu_read_lock &pool->lock/1 &p->pi_lock irq_context: 0 &dev->mutex running_helpers_waitq.lock irq_context: 0 &dev->mutex subsys mutex#30 irq_context: 0 &dev->mutex subsys mutex#30 &k->k_lock irq_context: 0 &dev->mutex (console_sem).lock irq_context: 0 &dev->mutex console_lock console_srcu console_owner_lock irq_context: 0 &dev->mutex console_lock console_srcu console_owner irq_context: 0 &dev->mutex console_lock console_srcu console_owner &port_lock_key irq_context: 0 &dev->mutex console_lock console_srcu console_owner console_owner_lock irq_context: 0 &dev->mutex input_mutex irq_context: 0 &dev->mutex input_mutex fs_reclaim irq_context: 0 &dev->mutex input_mutex fs_reclaim mmu_notifier_invalidate_range_start irq_context: 0 &dev->mutex input_mutex pool_lock#2 irq_context: 0 &dev->mutex input_mutex &dev->mutex#2 irq_context: 0 &dev->mutex input_mutex input_devices_poll_wait.lock irq_context: 0 &dev->mutex semaphore->lock irq_context: 0 &dev->mutex *(&acpi_gbl_hardware_lock) irq_context: 0 &dev->mutex &rq->__lock irq_context: 0 &dev->mutex &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)events_unbound (work_completion)(&sub_info->work) cgroup_threadgroup_rwsem &rq->__lock irq_context: 0 (wq_completion)events_unbound (work_completion)(&sub_info->work) cgroup_threadgroup_rwsem &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 &dev->mutex rcu_read_lock pool_lock#2 irq_context: 0 &dev->mutex wakeup_ida.xa_lock irq_context: 0 &dev->mutex lock kernfs_idr_lock pool_lock#2 irq_context: 0 &dev->mutex rcu_read_lock &pool->lock/1 &p->pi_lock &rq->__lock irq_context: 0 &dev->mutex rcu_read_lock &pool->lock/1 &p->pi_lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 &dev->mutex subsys mutex#15 irq_context: 0 &dev->mutex subsys mutex#15 &k->k_lock irq_context: 0 &dev->mutex events_lock irq_context: 0 &dev->mutex wakeup_srcu_srcu_usage.lock irq_context: 0 &dev->mutex wakeup_srcu_srcu_usage.lock &obj_hash[i].lock irq_context: 0 &dev->mutex &ACCESS_PRIVATE(sdp, lock) irq_context: 0 &dev->mutex wakeup_srcu irq_context: 0 &dev->mutex wakeup_srcu_srcu_usage.lock &ACCESS_PRIVATE(sdp, lock) irq_context: 0 &dev->mutex wakeup_srcu_srcu_usage.lock rcu_read_lock &pool->lock irq_context: 0 &dev->mutex wakeup_srcu_srcu_usage.lock rcu_read_lock &pool->lock &obj_hash[i].lock irq_context: 0 tasklist_lock &c->lock irq_context: 0 &dev->mutex wakeup_srcu_srcu_usage.lock rcu_read_lock &pool->lock &p->pi_lock irq_context: 0 &dev->mutex wakeup_srcu_srcu_usage.lock rcu_read_lock &pool->lock &p->pi_lock &rq->__lock irq_context: 0 &dev->mutex wakeup_srcu_srcu_usage.lock rcu_read_lock &pool->lock &p->pi_lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)rcu_gp (work_completion)(&(&ssp->srcu_sup->work)->work) &ssp->srcu_sup->srcu_gp_mutex wakeup_srcu_srcu_usage.lock irq_context: 0 (wq_completion)rcu_gp (work_completion)(&(&ssp->srcu_sup->work)->work) wakeup_srcu_srcu_usage.lock irq_context: 0 (wq_completion)rcu_gp (work_completion)(&(&ssp->srcu_sup->work)->work) rcu_read_lock &pool->lock fill_pool_map-wait-type-override pool_lock#2 irq_context: 0 (wq_completion)rcu_gp (work_completion)(&(&ssp->srcu_sup->work)->work) rcu_read_lock &pool->lock fill_pool_map-wait-type-override pool_lock irq_context: 0 (wq_completion)events_unbound (work_completion)(&sub_info->work) cgroup_threadgroup_rwsem &cfs_rq->removed.lock irq_context: 0 (wq_completion)events_unbound (work_completion)(&sub_info->work) cgroup_threadgroup_rwsem &obj_hash[i].lock irq_context: 0 (wq_completion)events_unbound (work_completion)(&sub_info->work) cgroup_threadgroup_rwsem pool_lock#2 irq_context: 0 (wq_completion)rcu_gp &rq->__lock irq_context: 0 (wq_completion)rcu_gp &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 &dev->mutex &x->wait#2 irq_context: 0 (wq_completion)rcu_gp &cfs_rq->removed.lock irq_context: 0 (wq_completion)rcu_gp &obj_hash[i].lock irq_context: 0 (wq_completion)rcu_gp (work_completion)(&(&ssp->srcu_sup->work)->work) &ssp->srcu_sup->srcu_gp_mutex &ssp->srcu_sup->srcu_cb_mutex wakeup_srcu_srcu_usage.lock irq_context: 0 (wq_completion)rcu_gp (work_completion)(&sdp->work) &rq->__lock irq_context: 0 &dev->mutex (&ws->timer) irq_context: 0 &dev->mutex &base->lock irq_context: 0 &dev->mutex &k->k_lock klist_remove_lock irq_context: 0 &dev->mutex &root->kernfs_rwsem kernfs_idr_lock irq_context: 0 &dev->mutex &root->kernfs_rwsem &obj_hash[i].lock irq_context: 0 &dev->mutex &root->kernfs_rwsem pool_lock#2 irq_context: 0 &dev->mutex subsys mutex#15 &k->k_lock klist_remove_lock irq_context: 0 &dev->mutex &root->kernfs_rwsem kernfs_idr_lock &obj_hash[i].lock irq_context: 0 &dev->mutex &root->kernfs_rwsem kernfs_idr_lock pool_lock#2 irq_context: 0 &dev->mutex deferred_probe_mutex irq_context: 0 &dev->mutex mmu_notifier_invalidate_range_start irq_context: 0 &dev->mutex kernfs_idr_lock irq_context: 0 &dev->mutex &ws->lock irq_context: 0 &dev->mutex deleted_ws.lock irq_context: 0 &dev->mutex probe_waitqueue.lock irq_context: 0 &dev->mutex &pcp->lock &zone->lock &____s->seqcount irq_context: 0 register_count_mutex irq_context: 0 register_count_mutex &k->list_lock irq_context: 0 &pool->lock/1 &p->pi_lock &rq->__lock &cfs_rq->removed.lock irq_context: 0 register_count_mutex fs_reclaim irq_context: 0 register_count_mutex fs_reclaim mmu_notifier_invalidate_range_start irq_context: 0 register_count_mutex pool_lock#2 irq_context: 0 register_count_mutex &c->lock irq_context: 0 register_count_mutex &pcp->lock &zone->lock irq_context: 0 register_count_mutex &zone->lock irq_context: 0 register_count_mutex &____s->seqcount irq_context: 0 register_count_mutex lock irq_context: 0 register_count_mutex lock kernfs_idr_lock irq_context: 0 register_count_mutex &root->kernfs_rwsem irq_context: 0 register_count_mutex &root->kernfs_rwsem &root->kernfs_iattr_rwsem irq_context: 0 register_count_mutex &k->k_lock irq_context: 0 register_count_mutex &pcp->lock &zone->lock &____s->seqcount irq_context: 0 register_count_mutex rcu_read_lock pool_lock#2 irq_context: 0 register_count_mutex &obj_hash[i].lock irq_context: 0 register_count_mutex uevent_sock_mutex irq_context: 0 register_count_mutex rcu_read_lock &pool->lock/1 irq_context: 0 register_count_mutex rcu_read_lock &pool->lock/1 &obj_hash[i].lock irq_context: 0 register_count_mutex rcu_read_lock &pool->lock/1 pool_lock#2 irq_context: 0 register_count_mutex rcu_read_lock &pool->lock/1 &p->pi_lock irq_context: 0 register_count_mutex rcu_read_lock &pool->lock/1 &p->pi_lock &rq->__lock irq_context: 0 register_count_mutex rcu_read_lock &pool->lock/1 &p->pi_lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 &mm->mmap_lock &rq->__lock irq_context: 0 register_count_mutex running_helpers_waitq.lock irq_context: 0 (cpufreq_policy_notifier_list).rwsem irq_context: 0 &dev->mutex cpu_add_remove_lock irq_context: 0 &dev->mutex tick_broadcast_lock irq_context: 0 &dev->mutex cpuidle_driver_lock irq_context: 0 &dev->mutex cpuidle_lock irq_context: 0 &dev->mutex cpuidle_lock fs_reclaim irq_context: 0 &dev->mutex cpuidle_lock fs_reclaim mmu_notifier_invalidate_range_start irq_context: 0 &dev->mutex cpuidle_lock pool_lock#2 irq_context: 0 &dev->mutex cpuidle_lock lock irq_context: 0 &dev->mutex cpuidle_lock lock kernfs_idr_lock irq_context: 0 &dev->mutex cpuidle_lock &root->kernfs_rwsem irq_context: 0 &dev->mutex cpuidle_lock &root->kernfs_rwsem &root->kernfs_iattr_rwsem irq_context: 0 &dev->mutex cpuidle_lock &c->lock irq_context: 0 &dev->mutex cpuidle_lock &pcp->lock &zone->lock irq_context: 0 &dev->mutex cpuidle_lock &zone->lock irq_context: 0 &dev->mutex cpuidle_lock &____s->seqcount irq_context: 0 &dev->mutex thermal_cdev_ida.xa_lock irq_context: 0 &dev->mutex cpufreq_driver_lock irq_context: 0 &mm->mmap_lock &cfs_rq->removed.lock irq_context: 0 &dev->mutex &obj_hash[i].lock pool_lock irq_context: 0 &mm->mmap_lock &obj_hash[i].lock irq_context: 0 &dev->mutex gdp_mutex &c->lock irq_context: 0 &dev->mutex gdp_mutex &pcp->lock &zone->lock irq_context: 0 &dev->mutex gdp_mutex &zone->lock irq_context: 0 &dev->mutex gdp_mutex &____s->seqcount irq_context: 0 &dev->mutex rcu_read_lock &pool->lock/1 &p->pi_lock &rq->__lock &cfs_rq->removed.lock irq_context: 0 &dev->mutex subsys mutex#31 irq_context: 0 &dev->mutex subsys mutex#31 &rq->__lock irq_context: 0 &dev->mutex subsys mutex#31 &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 &dev->mutex subsys mutex#31 &k->k_lock irq_context: 0 &dev->mutex thermal_list_lock irq_context: 0 &dev->mutex cpuidle_lock lock kernfs_idr_lock pool_lock#2 irq_context: 0 (x86_mce_decoder_chain).rwsem irq_context: 0 (wq_completion)events_unbound (work_completion)(&entry->work) &k->list_lock irq_context: 0 (wq_completion)events_unbound (work_completion)(&entry->work) &k->k_lock irq_context: 0 (wq_completion)events_unbound (work_completion)(&entry->work) lock irq_context: 0 (wq_completion)events_unbound (work_completion)(&entry->work) lock kernfs_idr_lock irq_context: 0 (wq_completion)events_unbound (work_completion)(&entry->work) &root->kernfs_rwsem irq_context: 0 (wq_completion)events_unbound (work_completion)(&entry->work) &root->kernfs_rwsem &root->kernfs_iattr_rwsem irq_context: 0 (wq_completion)events_unbound (work_completion)(&entry->work) lock kernfs_idr_lock pool_lock#2 irq_context: 0 (wq_completion)events_unbound (work_completion)(&entry->work) &pcp->lock &zone->lock &____s->seqcount irq_context: 0 (wq_completion)events_unbound (work_completion)(&entry->work) uevent_sock_mutex irq_context: 0 (wq_completion)events_unbound (work_completion)(&entry->work) rcu_read_lock &pool->lock/1 irq_context: 0 (wq_completion)events_unbound (work_completion)(&entry->work) rcu_read_lock &pool->lock/1 &obj_hash[i].lock irq_context: 0 (wq_completion)events_unbound (work_completion)(&entry->work) rcu_read_lock &pool->lock/1 pool_lock#2 irq_context: 0 (wq_completion)events_unbound (work_completion)(&entry->work) rcu_read_lock &pool->lock/1 &p->pi_lock irq_context: 0 (wq_completion)events_unbound (work_completion)(&entry->work) running_helpers_waitq.lock irq_context: 0 &dev->mutex &(&priv->bus_notifier)->rwsem irq_context: 0 &dev->mutex &drv->dynids.lock irq_context: 0 &dev->mutex pci_config_lock irq_context: 0 &dev->mutex batched_entropy_u8.lock irq_context: 0 &dev->mutex kfence_freelist_lock irq_context: 0 &dev->mutex *(&acpi_gbl_reference_count_lock) irq_context: 0 &dev->mutex &meta->lock irq_context: 0 &dev->mutex acpi_link_lock irq_context: 0 &dev->mutex acpi_link_lock fs_reclaim irq_context: 0 &dev->mutex acpi_link_lock fs_reclaim mmu_notifier_invalidate_range_start irq_context: 0 &dev->mutex acpi_link_lock pool_lock#2 irq_context: 0 &dev->mutex acpi_link_lock semaphore->lock irq_context: 0 &dev->mutex acpi_link_lock &obj_hash[i].lock irq_context: 0 &dev->mutex acpi_link_lock *(&acpi_gbl_reference_count_lock) irq_context: 0 &dev->mutex acpi_link_lock &____s->seqcount irq_context: 0 &dev->mutex acpi_link_lock &c->lock irq_context: 0 &dev->mutex acpi_link_lock pci_config_lock irq_context: 0 &dev->mutex acpi_link_lock &rq->__lock irq_context: 0 &dev->mutex acpi_link_lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 &dev->mutex acpi_link_lock &pcp->lock &zone->lock irq_context: 0 &dev->mutex acpi_link_lock &zone->lock irq_context: 0 &dev->mutex acpi_link_lock rcu_read_lock pool_lock#2 irq_context: 0 &dev->mutex acpi_link_lock (console_sem).lock irq_context: 0 &dev->mutex acpi_link_lock console_lock console_srcu console_owner_lock irq_context: 0 &dev->mutex acpi_link_lock console_lock console_srcu console_owner irq_context: 0 &dev->mutex acpi_link_lock console_lock console_srcu console_owner &port_lock_key irq_context: 0 &dev->mutex acpi_link_lock console_lock console_srcu console_owner console_owner_lock irq_context: 0 &dev->mutex acpi_ioapic_lock irq_context: 0 &dev->mutex acpi_ioapic_lock ioapic_mutex irq_context: 0 &dev->mutex resource_lock irq_context: 0 &dev->mutex virtio_index_ida.xa_lock irq_context: 0 &dev->mutex &dev->mutex &dev->power.lock irq_context: 0 &dev->mutex &dev->mutex &k->list_lock irq_context: 0 &dev->mutex &dev->mutex &k->k_lock irq_context: 0 &dev->mutex subsys mutex#32 irq_context: 0 &dev->mutex fwnode_link_lock &k->k_lock irq_context: 0 (wq_completion)events (work_completion)(&p->wq) &obj_hash[i].lock pool_lock irq_context: 0 &dev->mutex acpi_link_lock &pcp->lock &zone->lock &____s->seqcount irq_context: 0 &dev->mutex quarantine_lock irq_context: 0 &dev->mutex &dev->devres_lock irq_context: 0 &dev->mutex &md->mutex irq_context: 0 &dev->mutex &md->mutex fs_reclaim irq_context: 0 &dev->mutex &md->mutex fs_reclaim mmu_notifier_invalidate_range_start irq_context: 0 &dev->mutex &md->mutex pool_lock#2 irq_context: 0 &dev->mutex &md->mutex irq_domain_mutex irq_context: 0 &dev->mutex &md->mutex irq_domain_mutex fs_reclaim irq_context: 0 &dev->mutex &md->mutex irq_domain_mutex fs_reclaim mmu_notifier_invalidate_range_start irq_context: 0 &dev->mutex &md->mutex irq_domain_mutex pool_lock#2 irq_context: 0 &dev->mutex &md->mutex irq_domain_mutex &obj_hash[i].lock irq_context: 0 &dev->mutex resource_lock irq_context: 0 &dev->mutex memtype_lock irq_context: 0 &dev->mutex free_vmap_area_lock irq_context: 0 &dev->mutex free_vmap_area_lock &obj_hash[i].lock irq_context: 0 &dev->mutex free_vmap_area_lock pool_lock#2 irq_context: 0 &dev->mutex vmap_area_lock irq_context: 0 &dev->mutex &md->mutex pci_config_lock irq_context: 0 &dev->mutex &md->mutex &xa->xa_lock#5 irq_context: 0 &dev->mutex &md->mutex &xa->xa_lock#5 pool_lock#2 irq_context: 0 &dev->mutex &md->mutex &domain->mutex irq_context: 0 &dev->mutex &md->mutex &domain->mutex sparse_irq_lock irq_context: 0 &dev->mutex &md->mutex &domain->mutex sparse_irq_lock fs_reclaim irq_context: 0 &dev->mutex &md->mutex &domain->mutex sparse_irq_lock fs_reclaim mmu_notifier_invalidate_range_start irq_context: 0 &dev->mutex &md->mutex &domain->mutex sparse_irq_lock pool_lock#2 irq_context: 0 &dev->mutex &md->mutex &domain->mutex sparse_irq_lock pcpu_alloc_mutex irq_context: 0 &dev->mutex &md->mutex &domain->mutex sparse_irq_lock pcpu_alloc_mutex pcpu_lock irq_context: 0 &dev->mutex &md->mutex &domain->mutex sparse_irq_lock &obj_hash[i].lock irq_context: 0 &dev->mutex &md->mutex &domain->mutex sparse_irq_lock lock irq_context: 0 &dev->mutex &md->mutex &domain->mutex sparse_irq_lock lock kernfs_idr_lock irq_context: 0 &dev->mutex &md->mutex &domain->mutex sparse_irq_lock &root->kernfs_rwsem irq_context: 0 &dev->mutex &md->mutex &domain->mutex sparse_irq_lock &root->kernfs_rwsem &root->kernfs_iattr_rwsem irq_context: 0 &dev->mutex &md->mutex &domain->mutex fs_reclaim irq_context: 0 &dev->mutex &md->mutex &domain->mutex fs_reclaim mmu_notifier_invalidate_range_start irq_context: 0 &dev->mutex &md->mutex &domain->mutex pool_lock#2 irq_context: 0 &dev->mutex &md->mutex &domain->mutex &pcp->lock &zone->lock irq_context: 0 &dev->mutex &md->mutex &domain->mutex &zone->lock irq_context: 0 &dev->mutex &md->mutex &domain->mutex &____s->seqcount irq_context: 0 &dev->mutex &md->mutex &domain->mutex rcu_read_lock pool_lock#2 irq_context: 0 &dev->mutex &md->mutex &domain->mutex &obj_hash[i].lock irq_context: 0 &dev->mutex &md->mutex &domain->mutex vector_lock irq_context: 0 &dev->mutex &md->mutex &domain->mutex &irq_desc_lock_class irq_context: 0 &dev->mutex &md->mutex &irq_desc_lock_class irq_context: 0 &dev->mutex &md->mutex vector_lock irq_context: 0 &dev->mutex &md->mutex &root->kernfs_rwsem irq_context: 0 &dev->mutex &md->mutex lock irq_context: 0 &dev->mutex &md->mutex lock kernfs_idr_lock irq_context: 0 &dev->mutex &md->mutex &root->kernfs_rwsem irq_context: 0 &dev->mutex &md->mutex &root->kernfs_rwsem &root->kernfs_iattr_rwsem irq_context: 0 &dev->mutex &md->mutex &domain->mutex sparse_irq_lock &c->lock irq_context: 0 &dev->mutex &md->mutex &domain->mutex sparse_irq_lock &____s->seqcount irq_context: 0 &dev->mutex &md->mutex &c->lock irq_context: 0 &dev->mutex &md->mutex &____s->seqcount irq_context: 0 &dev->mutex &md->mutex &domain->mutex sparse_irq_lock &rq->__lock irq_context: 0 &dev->mutex &md->mutex &domain->mutex sparse_irq_lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 &dev->mutex &desc->request_mutex irq_context: 0 &dev->mutex &desc->request_mutex &irq_desc_lock_class irq_context: 0 &dev->mutex &desc->request_mutex &irq_desc_lock_class vector_lock irq_context: 0 &dev->mutex &desc->request_mutex &irq_desc_lock_class mask_lock irq_context: 0 &dev->mutex &desc->request_mutex &irq_desc_lock_class mask_lock tmp_mask_lock irq_context: 0 &dev->mutex &desc->request_mutex &irq_desc_lock_class mask_lock tmp_mask_lock vector_lock irq_context: 0 &dev->mutex register_lock irq_context: 0 &dev->mutex register_lock proc_subdir_lock irq_context: 0 &dev->mutex register_lock fs_reclaim irq_context: 0 &dev->mutex register_lock fs_reclaim mmu_notifier_invalidate_range_start irq_context: 0 &dev->mutex register_lock pool_lock#2 irq_context: 0 &dev->mutex register_lock proc_inum_ida.xa_lock irq_context: 0 &dev->mutex register_lock proc_subdir_lock irq_context: 0 &dev->mutex register_lock &c->lock irq_context: 0 &dev->mutex register_lock &____s->seqcount irq_context: 0 &dev->mutex &irq_desc_lock_class irq_context: 0 &dev->mutex proc_subdir_lock irq_context: 0 &dev->mutex proc_inum_ida.xa_lock irq_context: 0 &dev->mutex proc_subdir_lock irq_context: 0 &dev->mutex &zone->lock &____s->seqcount irq_context: 0 &dev->mutex &dev->vqs_list_lock irq_context: 0 &dev->mutex &vp_dev->lock irq_context: 0 &dev->mutex register_lock &pcp->lock &zone->lock irq_context: 0 &dev->mutex register_lock &zone->lock irq_context: 0 &dev->mutex cpu_hotplug_lock irq_context: 0 &dev->mutex &s->s_inode_list_lock irq_context: 0 &dev->mutex (oom_notify_list).rwsem irq_context: 0 &dev->mutex &dev->config_lock irq_context: 0 vdpa_dev_lock irq_context: 0 vdpa_dev_lock &rq->__lock irq_context: 0 vdpa_dev_lock &cfs_rq->removed.lock irq_context: 0 vdpa_dev_lock &obj_hash[i].lock irq_context: 0 &sig->cred_guard_mutex rcu_read_lock rcu_node_0 irq_context: 0 &sig->cred_guard_mutex rcu_read_lock &rq->__lock irq_context: 0 subsys mutex#33 irq_context: 0 subsys mutex#33 &k->k_lock irq_context: 0 &x->wait#11 &p->pi_lock &cfs_rq->removed.lock irq_context: 0 &x->wait#11 &p->pi_lock &rq->__lock &cfs_rq->removed.lock irq_context: 0 (wq_completion)events_unbound (work_completion)(&sub_info->work) &rq->__lock irq_context: 0 (wq_completion)events_unbound (work_completion)(&sub_info->work) &cfs_rq->removed.lock irq_context: 0 sb_writers &type->i_mutex_dir_key/1 quarantine_lock irq_context: 0 (wq_completion)events_unbound (work_completion)(&sub_info->work) fill_pool_map-wait-type-override &c->lock irq_context: 0 (wq_completion)events_unbound (work_completion)(&sub_info->work) fill_pool_map-wait-type-override &pcp->lock &zone->lock irq_context: 0 (wq_completion)events_unbound (work_completion)(&sub_info->work) fill_pool_map-wait-type-override &zone->lock irq_context: 0 (wq_completion)events_unbound (work_completion)(&sub_info->work) fill_pool_map-wait-type-override &____s->seqcount irq_context: 0 sb_writers &type->i_mutex_dir_key/1 &meta->lock irq_context: 0 sb_writers &type->i_mutex_dir_key/1 kfence_freelist_lock irq_context: 0 rcu_read_lock &pool->lock/1 &p->pi_lock &cfs_rq->removed.lock irq_context: 0 pcpu_alloc_mutex &rq->__lock irq_context: 0 rcu_read_lock &pool->lock/1 fill_pool_map-wait-type-override pool_lock#2 irq_context: 0 rcu_read_lock &pool->lock/1 fill_pool_map-wait-type-override pool_lock irq_context: 0 (wq_completion)events (work_completion)(&p->wq) &rq->__lock irq_context: 0 &child->perf_event_mutex &rq->__lock irq_context: 0 &sig->cred_guard_mutex &pcp->lock &zone->lock &____s->seqcount irq_context: 0 &mm->mmap_lock rcu_read_lock &rcu_state.gp_wq irq_context: 0 &mm->mmap_lock rcu_read_lock &rcu_state.gp_wq &p->pi_lock irq_context: 0 &mm->mmap_lock rcu_read_lock &rcu_state.gp_wq &p->pi_lock &cfs_rq->removed.lock irq_context: 0 &mm->mmap_lock rcu_read_lock &rcu_state.gp_wq &p->pi_lock &rq->__lock irq_context: 0 &mm->mmap_lock rcu_read_lock &rcu_state.gp_wq &p->pi_lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 &sig->cred_guard_mutex &mm->mmap_lock rcu_node_0 irq_context: 0 &sig->cred_guard_mutex &mm->mmap_lock &rq->__lock irq_context: 0 &sig->cred_guard_mutex &mm->mmap_lock &cfs_rq->removed.lock irq_context: 0 &sig->cred_guard_mutex &mm->mmap_lock rcu_read_lock rcu_node_0 irq_context: 0 &sig->cred_guard_mutex &n->list_lock irq_context: 0 &sig->cred_guard_mutex &cfs_rq->removed.lock irq_context: 0 (wq_completion)events_unbound (work_completion)(&sub_info->work) lock pidmap_lock &c->lock irq_context: 0 (wq_completion)events_unbound (work_completion)(&sub_info->work) lock pidmap_lock &pcp->lock &zone->lock irq_context: 0 (wq_completion)events_unbound (work_completion)(&sub_info->work) lock pidmap_lock &zone->lock irq_context: 0 (wq_completion)events_unbound (work_completion)(&sub_info->work) lock pidmap_lock &____s->seqcount irq_context: 0 sb_writers &type->i_mutex_dir_key/1 &sb->s_type->i_mutex_key#4 &rq->__lock irq_context: 0 &sig->cred_guard_mutex &mm->mmap_lock rcu_read_lock &rq->__lock irq_context: 0 sb_writers &type->i_mutex_dir_key/1 &sb->s_type->i_mutex_key#4 &cfs_rq->removed.lock irq_context: 0 sb_writers &type->i_mutex_dir_key/1 &sb->s_type->i_mutex_key#4 &obj_hash[i].lock irq_context: 0 &sig->cred_guard_mutex quarantine_lock irq_context: 0 sb_writers &type->i_mutex_dir_key/1 &rq->__lock irq_context: softirq rcu_callback put_task_map-wait-type-override quarantine_lock irq_context: 0 tasklist_lock &sighand->siglock &(&sig->stats_lock)->lock &____s->seqcount#5 fill_pool_map-wait-type-override pool_lock#2 irq_context: 0 tasklist_lock &sighand->siglock &(&sig->stats_lock)->lock &____s->seqcount#5 fill_pool_map-wait-type-override pool_lock irq_context: 0 &sig->cred_guard_mutex &mm->mmap_lock &meta->lock irq_context: 0 &sig->cred_guard_mutex &mm->mmap_lock kfence_freelist_lock irq_context: softirq rcu_callback rcu_read_lock &pool->lock &p->pi_lock &rq->__lock &cfs_rq->removed.lock irq_context: 0 &sig->cred_guard_mutex &mm->mmap_lock rcu_read_lock &rcu_state.gp_wq irq_context: 0 &sig->cred_guard_mutex &mm->mmap_lock rcu_read_lock &rcu_state.gp_wq &p->pi_lock irq_context: 0 &sig->cred_guard_mutex &mm->mmap_lock rcu_read_lock &rcu_state.gp_wq &p->pi_lock &rq->__lock irq_context: 0 &sig->cred_guard_mutex &mm->mmap_lock rcu_read_lock &rcu_state.gp_wq &p->pi_lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: hardirq &vb->stop_update_lock irq_context: hardirq &vb->stop_update_lock rcu_read_lock &pool->lock irq_context: hardirq &vb->stop_update_lock rcu_read_lock &pool->lock &obj_hash[i].lock irq_context: hardirq &vb->stop_update_lock rcu_read_lock &pool->lock pool_lock#2 irq_context: hardirq &vb->stop_update_lock rcu_read_lock &pool->lock &p->pi_lock irq_context: hardirq &vb->stop_update_lock rcu_read_lock &pool->lock &p->pi_lock &rq->__lock irq_context: hardirq &vb->stop_update_lock rcu_read_lock &pool->lock &p->pi_lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)events_freezable irq_context: 0 (wq_completion)events_freezable (work_completion)(&vb->update_balloon_stats_work) irq_context: 0 (wq_completion)events_freezable (work_completion)(&vb->update_balloon_stats_work) cpu_hotplug_lock irq_context: 0 (wq_completion)events_freezable (work_completion)(&vb->update_balloon_stats_work) &s->s_inode_list_lock irq_context: 0 (wq_completion)events_unbound (work_completion)(&sub_info->work) lock pidmap_lock &pcp->lock &zone->lock &____s->seqcount irq_context: 0 &mm->mmap_lock &pcp->lock &zone->lock &____s->seqcount irq_context: softirq &(&krcp->monitor_work)->timer irq_context: softirq &(&tbl->managed_work)->timer irq_context: softirq &(&krcp->monitor_work)->timer rcu_read_lock &pool->lock irq_context: softirq &(&tbl->managed_work)->timer rcu_read_lock &pool->lock irq_context: softirq &(&krcp->monitor_work)->timer rcu_read_lock &pool->lock &obj_hash[i].lock irq_context: softirq &(&tbl->managed_work)->timer rcu_read_lock &pool->lock &obj_hash[i].lock irq_context: softirq &(&krcp->monitor_work)->timer rcu_read_lock &pool->lock &p->pi_lock irq_context: softirq &(&tbl->managed_work)->timer rcu_read_lock &pool->lock &p->pi_lock irq_context: softirq &(&krcp->monitor_work)->timer rcu_read_lock &pool->lock &p->pi_lock &rq->__lock irq_context: softirq &(&tbl->managed_work)->timer rcu_read_lock &pool->lock &p->pi_lock &rq->__lock irq_context: softirq &(&krcp->monitor_work)->timer rcu_read_lock &pool->lock &p->pi_lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: softirq &(&tbl->managed_work)->timer rcu_read_lock &pool->lock &p->pi_lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 &sig->cred_guard_mutex &sb->s_type->i_mutex_key fs_reclaim &rq->__lock irq_context: 0 (wq_completion)events (work_completion)(&(&krcp->monitor_work)->work) irq_context: 0 (wq_completion)events (work_completion)(&(&krcp->monitor_work)->work) krc.lock irq_context: 0 (wq_completion)events (work_completion)(&(&krcp->monitor_work)->work) &obj_hash[i].lock irq_context: 0 (wq_completion)events (work_completion)(&(&krcp->monitor_work)->work) rcu_callback &obj_hash[i].lock irq_context: 0 (wq_completion)events (work_completion)(&(&krcp->monitor_work)->work) rcu_callback pool_lock#2 irq_context: 0 (wq_completion)events_unbound (work_completion)(&sub_info->work) batched_entropy_u8.lock irq_context: 0 (wq_completion)events_unbound (work_completion)(&sub_info->work) kfence_freelist_lock irq_context: 0 (wq_completion)events (work_completion)(&p->wq) &meta->lock irq_context: 0 (wq_completion)events (work_completion)(&p->wq) kfence_freelist_lock irq_context: 0 (wq_completion)events (work_completion)(&p->wq) rcu_read_lock &cfs_rq->removed.lock irq_context: 0 (wq_completion)events (work_completion)(&p->wq) rcu_read_lock &obj_hash[i].lock irq_context: 0 rcu_read_lock &pool->lock/1 fill_pool_map-wait-type-override &c->lock irq_context: 0 rcu_read_lock &pool->lock/1 fill_pool_map-wait-type-override &pcp->lock &zone->lock irq_context: 0 rcu_read_lock &pool->lock/1 fill_pool_map-wait-type-override &zone->lock irq_context: 0 rcu_read_lock &pool->lock/1 fill_pool_map-wait-type-override &____s->seqcount irq_context: 0 &dev->mutex pnp_lock irq_context: 0 &dev->mutex serial_mutex irq_context: 0 &dev->mutex serial_mutex gpio_lookup_lock irq_context: 0 &dev->mutex serial_mutex port_mutex irq_context: 0 &dev->mutex serial_mutex port_mutex fs_reclaim irq_context: 0 &dev->mutex serial_mutex port_mutex fs_reclaim mmu_notifier_invalidate_range_start irq_context: 0 &dev->mutex serial_mutex port_mutex pool_lock#2 irq_context: 0 &dev->mutex serial_mutex port_mutex &x->wait#9 irq_context: 0 &dev->mutex serial_mutex port_mutex &obj_hash[i].lock irq_context: 0 &dev->mutex serial_mutex port_mutex &k->list_lock irq_context: 0 &dev->mutex serial_mutex port_mutex &____s->seqcount irq_context: 0 &dev->mutex serial_mutex port_mutex lock irq_context: 0 &dev->mutex serial_mutex port_mutex lock kernfs_idr_lock irq_context: 0 &dev->mutex serial_mutex port_mutex &root->kernfs_rwsem irq_context: 0 &dev->mutex serial_mutex port_mutex &root->kernfs_rwsem &root->kernfs_iattr_rwsem irq_context: 0 &dev->mutex serial_mutex port_mutex bus_type_sem irq_context: 0 &dev->mutex serial_mutex port_mutex sysfs_symlink_target_lock irq_context: 0 &dev->mutex serial_mutex port_mutex &k->k_lock irq_context: 0 &dev->mutex serial_mutex port_mutex &root->kernfs_rwsem irq_context: 0 &dev->mutex serial_mutex port_mutex &c->lock irq_context: 0 &dev->mutex serial_mutex port_mutex &dev->power.lock irq_context: 0 &dev->mutex serial_mutex port_mutex dpm_list_mtx irq_context: 0 &dev->mutex serial_mutex port_mutex uevent_sock_mutex irq_context: 0 &dev->mutex serial_mutex port_mutex rcu_read_lock &pool->lock/1 irq_context: 0 &dev->mutex serial_mutex port_mutex rcu_read_lock &pool->lock/1 &obj_hash[i].lock irq_context: 0 &dev->mutex serial_mutex port_mutex rcu_read_lock &pool->lock/1 pool_lock#2 irq_context: 0 &dev->mutex serial_mutex port_mutex rcu_read_lock &pool->lock/1 &p->pi_lock irq_context: 0 &dev->mutex serial_mutex port_mutex rcu_read_lock &pool->lock/1 &p->pi_lock &rq->__lock irq_context: 0 &dev->mutex serial_mutex port_mutex rcu_read_lock &pool->lock/1 &p->pi_lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 &dev->mutex serial_mutex port_mutex running_helpers_waitq.lock irq_context: 0 &dev->mutex serial_mutex port_mutex &dev->mutex &dev->power.lock irq_context: 0 &dev->mutex serial_mutex port_mutex &dev->mutex &k->list_lock irq_context: 0 &dev->mutex serial_mutex port_mutex &dev->mutex &k->k_lock irq_context: 0 &dev->mutex serial_mutex port_mutex &dev->mutex device_links_srcu irq_context: 0 &dev->mutex serial_mutex port_mutex &dev->mutex fwnode_link_lock irq_context: 0 &dev->mutex serial_mutex port_mutex &dev->mutex device_links_lock irq_context: 0 &dev->mutex serial_mutex port_mutex &dev->mutex sysfs_symlink_target_lock irq_context: 0 &dev->mutex serial_mutex port_mutex &dev->mutex fs_reclaim irq_context: 0 &dev->mutex serial_mutex port_mutex &dev->mutex fs_reclaim mmu_notifier_invalidate_range_start irq_context: 0 &dev->mutex serial_mutex port_mutex &dev->mutex pool_lock#2 irq_context: 0 &dev->mutex serial_mutex port_mutex &dev->mutex lock irq_context: 0 &dev->mutex serial_mutex port_mutex &dev->mutex lock kernfs_idr_lock irq_context: 0 &dev->mutex serial_mutex port_mutex &dev->mutex &root->kernfs_rwsem irq_context: 0 &dev->mutex serial_mutex port_mutex &dev->mutex &root->kernfs_rwsem &root->kernfs_iattr_rwsem irq_context: 0 &dev->mutex serial_mutex port_mutex &dev->mutex deferred_probe_mutex irq_context: 0 &dev->mutex serial_mutex port_mutex &dev->mutex uevent_sock_mutex irq_context: 0 &dev->mutex serial_mutex port_mutex &dev->mutex &obj_hash[i].lock irq_context: 0 &dev->mutex serial_mutex port_mutex &dev->mutex rcu_read_lock &pool->lock/1 irq_context: 0 &dev->mutex serial_mutex port_mutex &dev->mutex rcu_read_lock &pool->lock/1 &obj_hash[i].lock irq_context: 0 &dev->mutex serial_mutex port_mutex &dev->mutex rcu_read_lock &pool->lock/1 pool_lock#2 irq_context: 0 &dev->mutex serial_mutex port_mutex &dev->mutex rcu_read_lock &pool->lock/1 &p->pi_lock irq_context: 0 &dev->mutex serial_mutex port_mutex &dev->mutex running_helpers_waitq.lock irq_context: 0 &dev->mutex serial_mutex port_mutex &dev->mutex probe_waitqueue.lock irq_context: 0 &dev->mutex serial_mutex port_mutex subsys mutex#34 irq_context: 0 &dev->mutex serial_mutex port_mutex &xa->xa_lock#6 irq_context: 0 &dev->mutex serial_mutex port_mutex &dev->mutex &dev->power.lock &dev->power.wait_queue irq_context: 0 &dev->mutex serial_mutex port_mutex &dev->mutex &dev->power.lock rcu_read_lock &pool->lock irq_context: 0 &dev->mutex serial_mutex port_mutex &dev->mutex &dev->power.lock rcu_read_lock &pool->lock &obj_hash[i].lock irq_context: 0 &dev->mutex serial_mutex port_mutex &dev->mutex &dev->power.lock rcu_read_lock &pool->lock pool_lock#2 irq_context: 0 &dev->mutex serial_mutex port_mutex &dev->mutex &dev->power.lock rcu_read_lock &pool->lock &p->pi_lock irq_context: 0 &dev->mutex serial_mutex port_mutex &dev->mutex &dev->power.lock rcu_read_lock &pool->lock &p->pi_lock &rq->__lock irq_context: 0 &dev->mutex serial_mutex port_mutex &dev->mutex &dev->power.lock rcu_read_lock &pool->lock &p->pi_lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 &dev->mutex serial_mutex port_mutex &dev->mutex &rq->__lock irq_context: 0 (wq_completion)pm irq_context: 0 (wq_completion)pm (work_completion)(&dev->power.work) irq_context: 0 (wq_completion)pm (work_completion)(&dev->power.work) &dev->power.lock irq_context: 0 (wq_completion)pm (work_completion)(&dev->power.work) &dev->power.lock &dev->power.wait_queue irq_context: 0 &dev->mutex serial_mutex port_mutex &port->mutex irq_context: 0 &dev->mutex serial_mutex port_mutex &port->mutex fs_reclaim irq_context: 0 &dev->mutex serial_mutex port_mutex &port->mutex fs_reclaim mmu_notifier_invalidate_range_start irq_context: 0 &dev->mutex serial_mutex port_mutex &port->mutex pool_lock#2 irq_context: 0 &dev->mutex serial_mutex port_mutex &port->mutex console_mutex irq_context: 0 &dev->mutex serial_mutex port_mutex &port->mutex resource_lock irq_context: 0 &dev->mutex serial_mutex port_mutex &port->mutex &port_lock_key irq_context: 0 &dev->mutex serial_mutex port_mutex &port->mutex (console_sem).lock irq_context: 0 &dev->mutex serial_mutex port_mutex &port->mutex console_lock console_srcu console_owner_lock irq_context: 0 &dev->mutex serial_mutex port_mutex &port->mutex console_lock console_srcu console_owner irq_context: 0 &dev->mutex serial_mutex port_mutex &port->mutex console_lock console_srcu console_owner &port_lock_key irq_context: 0 &dev->mutex serial_mutex port_mutex &port->mutex console_lock console_srcu console_owner console_owner_lock irq_context: 0 &dev->mutex serial_mutex port_mutex &port->mutex &rq->__lock irq_context: 0 &dev->mutex serial_mutex port_mutex &port->mutex &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 &dev->mutex serial_mutex port_mutex &port->mutex &cfs_rq->removed.lock irq_context: 0 &dev->mutex serial_mutex port_mutex &port->mutex &obj_hash[i].lock irq_context: 0 &dev->mutex serial_mutex port_mutex &port->mutex ctrl_ida.xa_lock irq_context: 0 &dev->mutex serial_mutex port_mutex &port->mutex &x->wait#9 irq_context: 0 &dev->mutex serial_mutex port_mutex &port->mutex &dev->power.lock irq_context: 0 &dev->mutex serial_mutex port_mutex &port->mutex &k->list_lock irq_context: 0 &dev->mutex serial_mutex port_mutex &port->mutex &c->lock irq_context: 0 &dev->mutex serial_mutex port_mutex &port->mutex &____s->seqcount irq_context: 0 &dev->mutex serial_mutex port_mutex &port->mutex lock irq_context: 0 &dev->mutex serial_mutex port_mutex &port->mutex lock kernfs_idr_lock irq_context: 0 &dev->mutex serial_mutex port_mutex &port->mutex &root->kernfs_rwsem irq_context: 0 &dev->mutex serial_mutex port_mutex &port->mutex &root->kernfs_rwsem &root->kernfs_iattr_rwsem irq_context: 0 &dev->mutex serial_mutex port_mutex &port->mutex bus_type_sem irq_context: 0 &dev->mutex serial_mutex port_mutex &port->mutex sysfs_symlink_target_lock irq_context: 0 &dev->mutex serial_mutex port_mutex &port->mutex &k->k_lock irq_context: 0 &dev->mutex serial_mutex port_mutex &port->mutex dpm_list_mtx irq_context: 0 &dev->mutex serial_mutex port_mutex &port->mutex &pcp->lock &zone->lock irq_context: 0 &dev->mutex serial_mutex port_mutex &port->mutex &pcp->lock &zone->lock &____s->seqcount irq_context: 0 &dev->mutex serial_mutex port_mutex &port->mutex &zone->lock irq_context: 0 &dev->mutex serial_mutex port_mutex &port->mutex rcu_read_lock pool_lock#2 irq_context: 0 &dev->mutex serial_mutex port_mutex &port->mutex uevent_sock_mutex irq_context: 0 &dev->mutex serial_mutex port_mutex &port->mutex rcu_read_lock &pool->lock/1 irq_context: 0 &dev->mutex serial_mutex port_mutex &port->mutex rcu_read_lock &pool->lock/1 &obj_hash[i].lock irq_context: 0 &dev->mutex serial_mutex port_mutex &port->mutex rcu_read_lock &pool->lock/1 pool_lock#2 irq_context: 0 &dev->mutex serial_mutex port_mutex &port->mutex rcu_read_lock &pool->lock/1 &p->pi_lock irq_context: 0 &dev->mutex serial_mutex port_mutex &port->mutex running_helpers_waitq.lock irq_context: 0 &dev->mutex serial_mutex port_mutex &port->mutex &dev->mutex &dev->power.lock irq_context: 0 &dev->mutex serial_mutex port_mutex &port->mutex &dev->mutex &k->list_lock irq_context: 0 &dev->mutex serial_mutex port_mutex &port->mutex &dev->mutex &k->k_lock irq_context: 0 &dev->mutex serial_mutex port_mutex &port->mutex subsys mutex#35 irq_context: 0 &dev->mutex serial_mutex port_mutex &port->mutex semaphore->lock irq_context: 0 &dev->mutex serial_mutex port_mutex &port->mutex *(&acpi_gbl_reference_count_lock) irq_context: 0 &dev->mutex serial_mutex port_mutex &port->mutex &root->kernfs_rwsem irq_context: 0 &dev->mutex serial_mutex port_mutex &port->mutex dev_pm_qos_sysfs_mtx irq_context: 0 &dev->mutex serial_mutex port_mutex &port->mutex dev_pm_qos_sysfs_mtx &root->kernfs_rwsem irq_context: 0 &dev->mutex serial_mutex port_mutex &port->mutex dev_pm_qos_sysfs_mtx &root->kernfs_rwsem irq_context: 0 &dev->mutex serial_mutex port_mutex &port->mutex dev_pm_qos_sysfs_mtx dev_pm_qos_mtx irq_context: 0 &dev->mutex serial_mutex port_mutex &port->mutex kernfs_idr_lock irq_context: 0 &dev->mutex serial_mutex port_mutex &port->mutex &k->k_lock klist_remove_lock irq_context: 0 &dev->mutex serial_mutex port_mutex &port->mutex &root->kernfs_rwsem kernfs_idr_lock irq_context: 0 &dev->mutex serial_mutex port_mutex &port->mutex &root->kernfs_rwsem &obj_hash[i].lock irq_context: 0 &dev->mutex serial_mutex port_mutex &port->mutex &root->kernfs_rwsem pool_lock#2 irq_context: 0 &dev->mutex serial_mutex port_mutex &port->mutex deferred_probe_mutex irq_context: 0 &dev->mutex serial_mutex port_mutex &port->mutex device_links_lock irq_context: 0 &dev->mutex serial_mutex port_mutex &port->mutex mmu_notifier_invalidate_range_start irq_context: 0 &dev->mutex serial_mutex port_mutex &port->mutex gdp_mutex irq_context: 0 &dev->mutex serial_mutex port_mutex &port->mutex gdp_mutex &k->list_lock irq_context: 0 &dev->mutex serial_mutex port_mutex &port->mutex gdp_mutex fs_reclaim irq_context: 0 &dev->mutex serial_mutex port_mutex &port->mutex gdp_mutex fs_reclaim mmu_notifier_invalidate_range_start irq_context: 0 &dev->mutex serial_mutex port_mutex &port->mutex gdp_mutex pool_lock#2 irq_context: 0 &dev->mutex serial_mutex port_mutex &port->mutex gdp_mutex lock irq_context: 0 &dev->mutex serial_mutex port_mutex &port->mutex gdp_mutex lock kernfs_idr_lock irq_context: 0 &dev->mutex serial_mutex port_mutex &port->mutex gdp_mutex &root->kernfs_rwsem irq_context: 0 &dev->mutex serial_mutex port_mutex &port->mutex gdp_mutex &root->kernfs_rwsem &root->kernfs_iattr_rwsem irq_context: 0 &dev->mutex serial_mutex port_mutex &port->mutex lock kernfs_idr_lock pool_lock#2 irq_context: 0 &dev->mutex serial_mutex port_mutex &port->mutex req_lock irq_context: 0 &dev->mutex serial_mutex port_mutex &port->mutex &p->pi_lock irq_context: 0 &dev->mutex serial_mutex port_mutex &port->mutex &p->pi_lock &rq->__lock irq_context: 0 &dev->mutex serial_mutex port_mutex &port->mutex &p->pi_lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 &dev->mutex serial_mutex port_mutex &port->mutex &p->pi_lock &rq->__lock &cfs_rq->removed.lock irq_context: 0 &dev->mutex serial_mutex port_mutex &port->mutex &x->wait#11 irq_context: 0 &dev->mutex serial_mutex port_mutex &port->mutex subsys mutex#21 irq_context: 0 &dev->mutex serial_mutex port_mutex &port->mutex subsys mutex#21 &k->k_lock irq_context: 0 &dev->mutex serial_mutex port_mutex &port->mutex chrdevs_lock irq_context: 0 &dev->mutex serial_mutex port_mutex &port->mutex &obj_hash[i].lock pool_lock irq_context: 0 &dev->mutex serial_mutex port_mutex &port->mutex rcu_read_lock &pool->lock/1 &p->pi_lock &rq->__lock irq_context: 0 &dev->mutex serial_mutex port_mutex &port->mutex rcu_read_lock &pool->lock/1 &p->pi_lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 &dev->mutex serial_mutex port_mutex &dev->mutex &c->lock irq_context: 0 &dev->mutex serial_mutex port_mutex &dev->mutex &____s->seqcount irq_context: 0 (wq_completion)events_unbound (work_completion)(&(&kfence_timer)->work) &rq->__lock &cfs_rq->removed.lock irq_context: 0 &dev->mutex serial_mutex port_mutex &port->mutex batched_entropy_u8.lock irq_context: 0 &dev->mutex serial_mutex port_mutex &port->mutex kfence_freelist_lock irq_context: 0 &dev->mutex serial_mutex port_mutex &port->mutex &meta->lock irq_context: 0 &dev->mutex &cfs_rq->removed.lock irq_context: 0 &dev->mutex serial_mutex port_mutex rcu_read_lock &pool->lock/1 &p->pi_lock &rq->__lock &cfs_rq->removed.lock irq_context: 0 &dev->mutex serial_mutex port_mutex &rq->__lock irq_context: 0 &dev->mutex serial_mutex port_mutex &dev->mutex rcu_read_lock &pool->lock/1 &p->pi_lock &rq->__lock irq_context: 0 &dev->mutex serial_mutex port_mutex &dev->mutex rcu_read_lock &pool->lock/1 &p->pi_lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 &dev->mutex serial_mutex port_mutex &dev->mutex &dev->power.lock rcu_read_lock &pool->lock &p->pi_lock &rq->__lock &cfs_rq->removed.lock irq_context: 0 &dev->mutex serial_mutex port_mutex &pcp->lock &zone->lock irq_context: 0 &dev->mutex serial_mutex port_mutex &zone->lock irq_context: 0 &dev->mutex serial_mutex port_mutex &dev->mutex &pcp->lock &zone->lock irq_context: 0 &dev->mutex serial_mutex port_mutex &dev->mutex &zone->lock irq_context: 0 &dev->mutex serial_mutex port_mutex &obj_hash[i].lock pool_lock irq_context: 0 fs_reclaim &rq->__lock irq_context: 0 &dev->mutex serial_mutex port_mutex &dev->mutex &cfs_rq->removed.lock irq_context: 0 &dev->mutex serial_mutex port_mutex &port->mutex &root->kernfs_rwsem kernfs_idr_lock &obj_hash[i].lock irq_context: 0 &dev->mutex serial_mutex port_mutex &port->mutex &root->kernfs_rwsem kernfs_idr_lock pool_lock#2 irq_context: 0 &dev->mutex serial_mutex port_mutex &port->mutex gdp_mutex lock kernfs_idr_lock pool_lock#2 irq_context: 0 fs_reclaim mmu_notifier_invalidate_range_start &rq->__lock irq_context: 0 fs_reclaim mmu_notifier_invalidate_range_start &cfs_rq->removed.lock irq_context: 0 fs_reclaim mmu_notifier_invalidate_range_start &obj_hash[i].lock irq_context: 0 misc_mtx &p->pi_lock &rq->__lock &cfs_rq->removed.lock irq_context: 0 &dev->mutex rng_index_ida.xa_lock irq_context: 0 pool_lock irq_context: hardirq &x->wait#12 irq_context: 0 &dev->mutex rng_mutex irq_context: 0 &dev->mutex rng_mutex &x->wait#13 irq_context: 0 &dev->mutex rng_mutex fs_reclaim irq_context: 0 &dev->mutex rng_mutex fs_reclaim mmu_notifier_invalidate_range_start irq_context: 0 &dev->mutex rng_mutex pool_lock#2 irq_context: 0 &dev->mutex rng_mutex kthread_create_lock irq_context: 0 &dev->mutex rng_mutex &p->pi_lock irq_context: 0 &dev->mutex rng_mutex &p->pi_lock &cfs_rq->removed.lock irq_context: 0 &dev->mutex rng_mutex &p->pi_lock &rq->__lock irq_context: 0 &dev->mutex rng_mutex &p->pi_lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 &dev->mutex rng_mutex &rq->__lock irq_context: 0 &dev->mutex rng_mutex &x->wait irq_context: 0 &dev->mutex rng_mutex &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 &dev->mutex rng_mutex &obj_hash[i].lock irq_context: 0 rng_mutex irq_context: 0 reading_mutex irq_context: 0 &dev->mutex reading_mutex irq_context: 0 &dev->mutex reading_mutex reading_mutex.wait_lock irq_context: 0 &dev->mutex reading_mutex &rq->__lock irq_context: 0 reading_mutex.wait_lock irq_context: 0 &dev->mutex reading_mutex &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 &dev->mutex input_pool.lock irq_context: 0 &root->kernfs_rwsem kernfs_idr_lock irq_context: 0 &root->kernfs_rwsem &obj_hash[i].lock irq_context: 0 &root->kernfs_rwsem pool_lock#2 irq_context: 0 klist_remove_lock irq_context: 0 &k->k_lock klist_remove_lock irq_context: 0 kernfs_idr_lock irq_context: 0 &dev->devres_lock irq_context: 0 &dev->managed.lock irq_context: 0 &type->s_umount_key#24/1 irq_context: 0 &type->s_umount_key#24/1 fs_reclaim irq_context: 0 &type->s_umount_key#24/1 fs_reclaim mmu_notifier_invalidate_range_start irq_context: 0 &type->s_umount_key#24/1 pool_lock#2 irq_context: 0 &type->s_umount_key#24/1 pcpu_alloc_mutex irq_context: 0 &type->s_umount_key#24/1 pcpu_alloc_mutex pcpu_lock irq_context: 0 &type->s_umount_key#24/1 shrinker_rwsem irq_context: 0 &type->s_umount_key#24/1 &c->lock irq_context: 0 &type->s_umount_key#24/1 &pcp->lock &zone->lock irq_context: 0 &type->s_umount_key#24/1 &zone->lock irq_context: 0 &type->s_umount_key#24/1 &____s->seqcount irq_context: 0 &type->s_umount_key#24/1 list_lrus_mutex irq_context: 0 &type->s_umount_key#24/1 sb_lock irq_context: 0 &type->s_umount_key#24/1 sb_lock unnamed_dev_ida.xa_lock irq_context: 0 &type->s_umount_key#24/1 mmu_notifier_invalidate_range_start irq_context: 0 &type->s_umount_key#24/1 &sb->s_type->i_lock_key#21 irq_context: 0 &type->s_umount_key#24/1 &s->s_inode_list_lock irq_context: 0 &type->s_umount_key#24/1 tk_core.seq.seqcount irq_context: 0 &type->s_umount_key#24/1 &sb->s_type->i_lock_key#21 &dentry->d_lock irq_context: 0 &type->s_umount_key#24/1 &dentry->d_lock irq_context: 0 &sb->s_type->i_lock_key#21 irq_context: 0 lock drm_minor_lock irq_context: 0 lock drm_minor_lock &c->lock irq_context: 0 lock drm_minor_lock &pcp->lock &zone->lock irq_context: 0 lock drm_minor_lock &zone->lock irq_context: 0 lock drm_minor_lock &____s->seqcount irq_context: 0 lock drm_minor_lock pool_lock#2 irq_context: 0 stack_depot_init_mutex irq_context: 0 &dev->debugfs_mutex irq_context: 0 subsys mutex#36 irq_context: 0 subsys mutex#36 &k->k_lock irq_context: 0 drm_minor_lock irq_context: 0 (worker)->lock irq_context: 0 &dev->mode_config.idr_mutex irq_context: 0 &dev->mode_config.idr_mutex fs_reclaim irq_context: 0 &dev->mode_config.idr_mutex fs_reclaim mmu_notifier_invalidate_range_start irq_context: 0 &dev->mode_config.idr_mutex pool_lock#2 irq_context: 0 crtc_ww_class_acquire irq_context: 0 crtc_ww_class_acquire crtc_ww_class_mutex irq_context: 0 crtc_ww_class_acquire crtc_ww_class_mutex reservation_ww_class_acquire irq_context: 0 crtc_ww_class_acquire crtc_ww_class_mutex reservation_ww_class_acquire reservation_ww_class_mutex irq_context: 0 &dev->mode_config.blob_lock irq_context: 0 &xa->xa_lock#7 irq_context: 0 &xa->xa_lock#8 irq_context: 0 &dev->mode_config.connector_list_lock irq_context: 0 &dev->vbl_lock irq_context: 0 drm_connector_list_iter &dev->mode_config.connector_list_lock irq_context: 0 drm_connector_list_iter fs_reclaim irq_context: 0 drm_connector_list_iter fs_reclaim mmu_notifier_invalidate_range_start irq_context: 0 drm_connector_list_iter pool_lock#2 irq_context: 0 drm_connector_list_iter &connector->mutex irq_context: 0 drm_connector_list_iter &connector->mutex fs_reclaim irq_context: 0 drm_connector_list_iter &connector->mutex fs_reclaim mmu_notifier_invalidate_range_start irq_context: 0 drm_connector_list_iter &connector->mutex pool_lock#2 irq_context: 0 drm_connector_list_iter &connector->mutex &x->wait#9 irq_context: 0 drm_connector_list_iter &connector->mutex &obj_hash[i].lock irq_context: 0 drm_connector_list_iter &connector->mutex &obj_hash[i].lock pool_lock irq_context: 0 &sig->cred_guard_mutex &mm->mmap_lock quarantine_lock irq_context: 0 drm_connector_list_iter &connector->mutex &k->list_lock irq_context: 0 drm_connector_list_iter &connector->mutex lock irq_context: 0 drm_connector_list_iter &connector->mutex lock kernfs_idr_lock irq_context: 0 drm_connector_list_iter &connector->mutex &root->kernfs_rwsem irq_context: 0 drm_connector_list_iter &connector->mutex &root->kernfs_rwsem &root->kernfs_iattr_rwsem irq_context: 0 drm_connector_list_iter &connector->mutex bus_type_sem irq_context: 0 drm_connector_list_iter &connector->mutex sysfs_symlink_target_lock irq_context: 0 drm_connector_list_iter &connector->mutex &root->kernfs_rwsem irq_context: 0 drm_connector_list_iter &connector->mutex &c->lock irq_context: 0 drm_connector_list_iter &connector->mutex &____s->seqcount irq_context: 0 drm_connector_list_iter &connector->mutex lock kernfs_idr_lock pool_lock#2 irq_context: 0 drm_connector_list_iter &connector->mutex &dev->power.lock irq_context: 0 drm_connector_list_iter &connector->mutex dpm_list_mtx irq_context: 0 drm_connector_list_iter &connector->mutex &pcp->lock &zone->lock irq_context: 0 drm_connector_list_iter &connector->mutex &zone->lock irq_context: 0 drm_connector_list_iter &connector->mutex uevent_sock_mutex irq_context: 0 drm_connector_list_iter &connector->mutex rcu_read_lock &pool->lock/1 irq_context: 0 drm_connector_list_iter &connector->mutex rcu_read_lock &pool->lock/1 &obj_hash[i].lock irq_context: 0 drm_connector_list_iter &connector->mutex rcu_read_lock &pool->lock/1 pool_lock#2 irq_context: 0 drm_connector_list_iter &connector->mutex rcu_read_lock &pool->lock/1 &p->pi_lock irq_context: 0 drm_connector_list_iter &connector->mutex rcu_read_lock &pool->lock/1 &p->pi_lock &rq->__lock irq_context: 0 drm_connector_list_iter &connector->mutex rcu_read_lock &pool->lock/1 &p->pi_lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 drm_connector_list_iter &connector->mutex running_helpers_waitq.lock irq_context: 0 drm_connector_list_iter &connector->mutex &rq->__lock irq_context: 0 drm_connector_list_iter &connector->mutex &k->k_lock irq_context: 0 drm_connector_list_iter &connector->mutex subsys mutex#36 irq_context: 0 drm_connector_list_iter &connector->mutex subsys mutex#36 &k->k_lock irq_context: 0 drm_connector_list_iter &connector->mutex pin_fs_lock irq_context: 0 drm_connector_list_iter &connector->mutex &sb->s_type->i_mutex_key#3 irq_context: 0 drm_connector_list_iter &connector->mutex &sb->s_type->i_mutex_key#3 rename_lock.seqcount irq_context: 0 drm_connector_list_iter &connector->mutex &sb->s_type->i_mutex_key#3 fs_reclaim irq_context: 0 drm_connector_list_iter &connector->mutex &sb->s_type->i_mutex_key#3 fs_reclaim mmu_notifier_invalidate_range_start irq_context: 0 drm_connector_list_iter &connector->mutex &sb->s_type->i_mutex_key#3 pool_lock#2 irq_context: 0 drm_connector_list_iter &connector->mutex &sb->s_type->i_mutex_key#3 &dentry->d_lock irq_context: 0 drm_connector_list_iter &connector->mutex &sb->s_type->i_mutex_key#3 rcu_read_lock rename_lock.seqcount irq_context: 0 drm_connector_list_iter &connector->mutex &sb->s_type->i_mutex_key#3 &dentry->d_lock &wq irq_context: 0 drm_connector_list_iter &connector->mutex &sb->s_type->i_mutex_key#3 mmu_notifier_invalidate_range_start irq_context: 0 drm_connector_list_iter &connector->mutex &sb->s_type->i_mutex_key#3 &sb->s_type->i_lock_key#7 irq_context: 0 drm_connector_list_iter &connector->mutex &sb->s_type->i_mutex_key#3 &s->s_inode_list_lock irq_context: 0 drm_connector_list_iter &connector->mutex &sb->s_type->i_mutex_key#3 tk_core.seq.seqcount irq_context: 0 drm_connector_list_iter &connector->mutex &sb->s_type->i_mutex_key#3 &sb->s_type->i_lock_key#7 &dentry->d_lock irq_context: 0 drm_connector_list_iter &connector->mutex &dev->mode_config.idr_mutex irq_context: 0 drm_connector_list_iter &connector->mutex connector_list_lock irq_context: 0 drm_connector_list_iter &connector->mutex &sb->s_type->i_mutex_key#3 &c->lock irq_context: 0 drm_connector_list_iter &connector->mutex &sb->s_type->i_mutex_key#3 &pcp->lock &zone->lock irq_context: 0 drm_connector_list_iter &connector->mutex &sb->s_type->i_mutex_key#3 &zone->lock irq_context: 0 drm_connector_list_iter &connector->mutex &sb->s_type->i_mutex_key#3 &____s->seqcount irq_context: 0 &dev->filelist_mutex irq_context: 0 &dev->clientlist_mutex irq_context: 0 &dev->clientlist_mutex &helper->lock irq_context: 0 &dev->clientlist_mutex &helper->lock drm_connector_list_iter &dev->mode_config.connector_list_lock irq_context: 0 &dev->clientlist_mutex &helper->lock drm_connector_list_iter fs_reclaim irq_context: 0 &dev->clientlist_mutex &helper->lock drm_connector_list_iter fs_reclaim mmu_notifier_invalidate_range_start irq_context: 0 &dev->clientlist_mutex &helper->lock drm_connector_list_iter pool_lock#2 irq_context: 0 &dev->clientlist_mutex &helper->lock fs_reclaim irq_context: 0 &dev->clientlist_mutex &helper->lock fs_reclaim mmu_notifier_invalidate_range_start irq_context: 0 &dev->clientlist_mutex &helper->lock pool_lock#2 irq_context: 0 &dev->clientlist_mutex &helper->lock &client->modeset_mutex irq_context: 0 &dev->clientlist_mutex &helper->lock &client->modeset_mutex &dev->mode_config.mutex irq_context: 0 &dev->clientlist_mutex &helper->lock &client->modeset_mutex &dev->mode_config.mutex crtc_ww_class_acquire irq_context: 0 &dev->clientlist_mutex &helper->lock &client->modeset_mutex &dev->mode_config.mutex crtc_ww_class_acquire crtc_ww_class_mutex irq_context: 0 &dev->clientlist_mutex &helper->lock &client->modeset_mutex &dev->mode_config.mutex crtc_ww_class_acquire crtc_ww_class_mutex fs_reclaim irq_context: 0 &dev->clientlist_mutex &helper->lock &client->modeset_mutex &dev->mode_config.mutex crtc_ww_class_acquire crtc_ww_class_mutex fs_reclaim mmu_notifier_invalidate_range_start irq_context: 0 &dev->clientlist_mutex &helper->lock &client->modeset_mutex &dev->mode_config.mutex crtc_ww_class_acquire crtc_ww_class_mutex pool_lock#2 irq_context: 0 &dev->clientlist_mutex &helper->lock &client->modeset_mutex &dev->mode_config.mutex crtc_ww_class_acquire crtc_ww_class_mutex &c->lock irq_context: 0 &dev->clientlist_mutex &helper->lock &client->modeset_mutex &dev->mode_config.mutex crtc_ww_class_acquire crtc_ww_class_mutex &pcp->lock &zone->lock irq_context: 0 &dev->clientlist_mutex &helper->lock &client->modeset_mutex &dev->mode_config.mutex crtc_ww_class_acquire crtc_ww_class_mutex &zone->lock irq_context: 0 &dev->clientlist_mutex &helper->lock &client->modeset_mutex &dev->mode_config.mutex crtc_ww_class_acquire crtc_ww_class_mutex &____s->seqcount irq_context: 0 &dev->clientlist_mutex &helper->lock &client->modeset_mutex &dev->mode_config.mutex crtc_ww_class_acquire crtc_ww_class_mutex &rq->__lock irq_context: 0 &dev->clientlist_mutex &helper->lock &client->modeset_mutex &dev->mode_config.mutex fs_reclaim irq_context: 0 &dev->clientlist_mutex &helper->lock &client->modeset_mutex &dev->mode_config.mutex fs_reclaim mmu_notifier_invalidate_range_start irq_context: 0 &dev->clientlist_mutex &helper->lock &client->modeset_mutex &dev->mode_config.mutex pool_lock#2 irq_context: 0 &dev->clientlist_mutex &helper->lock &client->modeset_mutex &dev->mode_config.mutex &obj_hash[i].lock irq_context: 0 &dev->clientlist_mutex &helper->lock &client->modeset_mutex fs_reclaim irq_context: 0 &dev->clientlist_mutex &helper->lock &client->modeset_mutex fs_reclaim mmu_notifier_invalidate_range_start irq_context: 0 &dev->clientlist_mutex &helper->lock &client->modeset_mutex pool_lock#2 irq_context: 0 &dev->clientlist_mutex &helper->lock &obj_hash[i].lock irq_context: 0 &dev->clientlist_mutex &helper->lock &client->modeset_mutex drm_connector_list_iter &dev->mode_config.connector_list_lock irq_context: 0 &dev->clientlist_mutex &helper->lock mmu_notifier_invalidate_range_start irq_context: 0 &dev->clientlist_mutex &helper->lock &sb->s_type->i_lock_key irq_context: 0 &dev->clientlist_mutex &helper->lock &s->s_inode_list_lock irq_context: 0 &dev->clientlist_mutex &helper->lock tk_core.seq.seqcount irq_context: 0 &dev->clientlist_mutex &helper->lock batched_entropy_u32.lock irq_context: 0 &dev->clientlist_mutex &helper->lock &sb->s_type->i_lock_key &dentry->d_lock irq_context: 0 &dev->clientlist_mutex &helper->lock &mgr->vm_lock irq_context: 0 &dev->clientlist_mutex &helper->lock &mgr->vm_lock &____s->seqcount irq_context: 0 &dev->clientlist_mutex &helper->lock &mgr->vm_lock &pcp->lock &zone->lock irq_context: 0 &dev->clientlist_mutex &helper->lock &mgr->vm_lock &zone->lock irq_context: 0 &dev->clientlist_mutex &helper->lock &mgr->vm_lock pool_lock#2 irq_context: 0 &dev->clientlist_mutex &helper->lock &mgr->vm_lock rcu_read_lock pool_lock#2 irq_context: 0 &dev->clientlist_mutex &helper->lock &mgr->vm_lock &obj_hash[i].lock irq_context: 0 &dev->clientlist_mutex &helper->lock &dev->object_name_lock irq_context: 0 &dev->clientlist_mutex &helper->lock &dev->object_name_lock lock irq_context: 0 &dev->clientlist_mutex &helper->lock &dev->object_name_lock lock &file_private->table_lock irq_context: 0 &dev->clientlist_mutex &helper->lock &dev->object_name_lock lock &file_private->table_lock pool_lock#2 irq_context: 0 &dev->clientlist_mutex &helper->lock &node->vm_lock irq_context: 0 &dev->clientlist_mutex &helper->lock &file_private->table_lock irq_context: 0 &dev->clientlist_mutex &helper->lock &dev->mode_config.idr_mutex irq_context: 0 &dev->clientlist_mutex &helper->lock &dev->mode_config.fb_lock irq_context: 0 &dev->clientlist_mutex &helper->lock &file->fbs_lock irq_context: 0 &dev->clientlist_mutex &helper->lock &prime_fpriv->lock irq_context: 0 &dev->clientlist_mutex &helper->lock &node->vm_lock &obj_hash[i].lock irq_context: 0 &dev->clientlist_mutex &helper->lock &node->vm_lock pool_lock#2 irq_context: 0 &dev->clientlist_mutex &helper->lock &file_private->table_lock &obj_hash[i].lock irq_context: 0 &dev->clientlist_mutex &helper->lock &file_private->table_lock pool_lock#2 irq_context: 0 &dev->clientlist_mutex &helper->lock free_vmap_area_lock irq_context: 0 &dev->clientlist_mutex &helper->lock vmap_area_lock irq_context: 0 &dev->clientlist_mutex &helper->lock &____s->seqcount irq_context: 0 &dev->clientlist_mutex &helper->lock init_mm.page_table_lock irq_context: 0 &dev->clientlist_mutex &helper->lock &pcp->lock &zone->lock irq_context: 0 &dev->clientlist_mutex &helper->lock &zone->lock irq_context: 0 &dev->clientlist_mutex &helper->lock &pcp->lock &zone->lock &____s->seqcount irq_context: 0 &dev->clientlist_mutex &helper->lock &rq->__lock irq_context: 0 &dev->clientlist_mutex registration_lock irq_context: 0 &dev->clientlist_mutex registration_lock fs_reclaim irq_context: 0 &dev->clientlist_mutex registration_lock fs_reclaim mmu_notifier_invalidate_range_start irq_context: 0 &dev->clientlist_mutex registration_lock pool_lock#2 irq_context: 0 &dev->clientlist_mutex registration_lock &x->wait#9 irq_context: 0 &dev->clientlist_mutex registration_lock &obj_hash[i].lock irq_context: 0 &dev->clientlist_mutex registration_lock &k->list_lock irq_context: 0 &dev->clientlist_mutex registration_lock gdp_mutex irq_context: 0 &dev->clientlist_mutex registration_lock gdp_mutex &k->list_lock irq_context: 0 &dev->clientlist_mutex registration_lock gdp_mutex fs_reclaim irq_context: 0 &dev->clientlist_mutex registration_lock gdp_mutex fs_reclaim mmu_notifier_invalidate_range_start irq_context: 0 &dev->clientlist_mutex registration_lock gdp_mutex pool_lock#2 irq_context: 0 &dev->clientlist_mutex registration_lock gdp_mutex lock irq_context: 0 &dev->clientlist_mutex registration_lock gdp_mutex lock kernfs_idr_lock irq_context: 0 &dev->clientlist_mutex registration_lock gdp_mutex &root->kernfs_rwsem irq_context: 0 &dev->clientlist_mutex registration_lock gdp_mutex &root->kernfs_rwsem &root->kernfs_iattr_rwsem irq_context: 0 &dev->clientlist_mutex registration_lock lock irq_context: 0 &dev->clientlist_mutex registration_lock lock kernfs_idr_lock irq_context: 0 &dev->clientlist_mutex registration_lock &root->kernfs_rwsem irq_context: 0 &dev->clientlist_mutex registration_lock &root->kernfs_rwsem &root->kernfs_iattr_rwsem irq_context: 0 &dev->clientlist_mutex registration_lock bus_type_sem irq_context: 0 &dev->clientlist_mutex registration_lock sysfs_symlink_target_lock irq_context: 0 &dev->clientlist_mutex registration_lock &root->kernfs_rwsem irq_context: 0 &dev->clientlist_mutex registration_lock &dev->power.lock irq_context: 0 &dev->clientlist_mutex registration_lock dpm_list_mtx irq_context: 0 &dev->clientlist_mutex registration_lock req_lock irq_context: 0 &dev->clientlist_mutex registration_lock &p->pi_lock irq_context: 0 &dev->clientlist_mutex registration_lock &p->pi_lock &rq->__lock irq_context: 0 &dev->clientlist_mutex registration_lock &p->pi_lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 &dev->clientlist_mutex registration_lock &p->pi_lock &rq->__lock &cfs_rq->removed.lock irq_context: 0 &dev->clientlist_mutex registration_lock &rq->__lock irq_context: 0 &dev->clientlist_mutex registration_lock &x->wait#11 irq_context: 0 &dev->clientlist_mutex registration_lock &c->lock irq_context: 0 &dev->clientlist_mutex registration_lock &pcp->lock &zone->lock irq_context: 0 &dev->clientlist_mutex registration_lock &zone->lock irq_context: 0 &dev->clientlist_mutex registration_lock &____s->seqcount irq_context: 0 &dev->clientlist_mutex registration_lock uevent_sock_mutex irq_context: 0 &dev->clientlist_mutex registration_lock rcu_read_lock &pool->lock/1 irq_context: 0 &dev->clientlist_mutex registration_lock rcu_read_lock &pool->lock/1 &obj_hash[i].lock irq_context: 0 &dev->clientlist_mutex registration_lock rcu_read_lock &pool->lock/1 pool_lock#2 irq_context: 0 &dev->clientlist_mutex registration_lock rcu_read_lock &pool->lock/1 &p->pi_lock irq_context: 0 &dev->clientlist_mutex registration_lock running_helpers_waitq.lock irq_context: 0 &dev->clientlist_mutex registration_lock &k->k_lock irq_context: 0 &dev->clientlist_mutex registration_lock subsys mutex#11 irq_context: 0 &dev->clientlist_mutex registration_lock subsys mutex#11 &k->k_lock irq_context: 0 &dev->clientlist_mutex registration_lock vt_switch_mutex irq_context: 0 &dev->clientlist_mutex registration_lock vt_switch_mutex fs_reclaim irq_context: 0 &dev->clientlist_mutex registration_lock vt_switch_mutex fs_reclaim mmu_notifier_invalidate_range_start irq_context: 0 &dev->clientlist_mutex registration_lock vt_switch_mutex pool_lock#2 irq_context: 0 &dev->clientlist_mutex registration_lock (console_sem).lock irq_context: 0 &dev->clientlist_mutex registration_lock console_lock irq_context: 0 &dev->clientlist_mutex registration_lock console_lock &fb_info->lock irq_context: 0 &dev->clientlist_mutex registration_lock console_lock fs_reclaim irq_context: 0 &dev->clientlist_mutex registration_lock console_lock fs_reclaim mmu_notifier_invalidate_range_start irq_context: 0 &dev->clientlist_mutex registration_lock console_lock pool_lock#2 irq_context: 0 &dev->clientlist_mutex registration_lock console_lock &obj_hash[i].lock irq_context: 0 &dev->clientlist_mutex registration_lock console_lock &pcp->lock &zone->lock irq_context: 0 &dev->clientlist_mutex registration_lock console_lock &zone->lock irq_context: 0 &dev->clientlist_mutex registration_lock console_lock &____s->seqcount irq_context: 0 &dev->clientlist_mutex registration_lock console_lock rcu_read_lock pool_lock#2 irq_context: 0 &dev->clientlist_mutex registration_lock console_lock &base->lock irq_context: 0 &dev->clientlist_mutex registration_lock console_lock &base->lock &obj_hash[i].lock irq_context: 0 &dev->clientlist_mutex registration_lock console_lock &x->wait#9 irq_context: 0 &dev->clientlist_mutex registration_lock console_lock &k->list_lock irq_context: 0 &dev->clientlist_mutex registration_lock console_lock gdp_mutex irq_context: 0 &dev->clientlist_mutex registration_lock console_lock gdp_mutex &k->list_lock irq_context: 0 &dev->clientlist_mutex registration_lock console_lock lock irq_context: 0 &dev->clientlist_mutex registration_lock console_lock lock kernfs_idr_lock irq_context: 0 &dev->clientlist_mutex registration_lock console_lock &root->kernfs_rwsem irq_context: 0 &dev->clientlist_mutex registration_lock console_lock &root->kernfs_rwsem &root->kernfs_iattr_rwsem irq_context: 0 &dev->clientlist_mutex registration_lock console_lock bus_type_sem irq_context: 0 &dev->clientlist_mutex registration_lock console_lock sysfs_symlink_target_lock irq_context: 0 &dev->clientlist_mutex registration_lock console_lock &c->lock irq_context: 0 &dev->clientlist_mutex registration_lock console_lock &root->kernfs_rwsem irq_context: 0 &dev->clientlist_mutex registration_lock console_lock &dev->power.lock irq_context: 0 &dev->clientlist_mutex registration_lock console_lock dpm_list_mtx irq_context: 0 &dev->clientlist_mutex registration_lock console_lock uevent_sock_mutex irq_context: 0 &dev->clientlist_mutex registration_lock console_lock rcu_read_lock &pool->lock/1 irq_context: 0 &dev->clientlist_mutex registration_lock console_lock rcu_read_lock &pool->lock/1 &obj_hash[i].lock irq_context: 0 &dev->clientlist_mutex registration_lock console_lock rcu_read_lock &pool->lock/1 pool_lock#2 irq_context: 0 &dev->clientlist_mutex registration_lock console_lock rcu_read_lock &pool->lock/1 &p->pi_lock irq_context: 0 &dev->clientlist_mutex registration_lock console_lock rcu_read_lock &pool->lock/1 &p->pi_lock &rq->__lock irq_context: 0 &dev->clientlist_mutex registration_lock console_lock rcu_read_lock &pool->lock/1 &p->pi_lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 &dev->clientlist_mutex registration_lock console_lock rcu_read_lock &pool->lock/1 &p->pi_lock &rq->__lock &cfs_rq->removed.lock irq_context: 0 &dev->clientlist_mutex registration_lock console_lock running_helpers_waitq.lock irq_context: 0 &dev->clientlist_mutex registration_lock console_lock subsys mutex#5 irq_context: 0 &dev->clientlist_mutex registration_lock console_lock subsys mutex#5 &k->k_lock irq_context: 0 &dev->clientlist_mutex registration_lock console_lock vga_lock irq_context: 0 &dev->clientlist_mutex registration_lock console_lock &helper->lock irq_context: 0 &dev->clientlist_mutex registration_lock console_lock &helper->lock &dev->master_mutex irq_context: 0 &dev->clientlist_mutex registration_lock console_lock &helper->lock &dev->master_mutex &client->modeset_mutex irq_context: 0 &dev->clientlist_mutex registration_lock console_lock &helper->lock &dev->master_mutex &client->modeset_mutex crtc_ww_class_acquire irq_context: 0 &dev->clientlist_mutex registration_lock console_lock &helper->lock &dev->master_mutex &client->modeset_mutex crtc_ww_class_acquire fs_reclaim irq_context: 0 &dev->clientlist_mutex registration_lock console_lock &helper->lock &dev->master_mutex &client->modeset_mutex crtc_ww_class_acquire fs_reclaim mmu_notifier_invalidate_range_start irq_context: 0 &dev->clientlist_mutex registration_lock console_lock &helper->lock &dev->master_mutex &client->modeset_mutex crtc_ww_class_acquire pool_lock#2 irq_context: 0 &dev->clientlist_mutex registration_lock console_lock &helper->lock &dev->master_mutex &client->modeset_mutex crtc_ww_class_acquire crtc_ww_class_mutex irq_context: 0 &dev->clientlist_mutex registration_lock console_lock &helper->lock &dev->master_mutex &client->modeset_mutex crtc_ww_class_acquire crtc_ww_class_mutex fs_reclaim irq_context: 0 &dev->clientlist_mutex registration_lock console_lock &helper->lock &dev->master_mutex &client->modeset_mutex crtc_ww_class_acquire crtc_ww_class_mutex fs_reclaim mmu_notifier_invalidate_range_start irq_context: 0 &dev->clientlist_mutex registration_lock console_lock &helper->lock &dev->master_mutex &client->modeset_mutex crtc_ww_class_acquire crtc_ww_class_mutex pool_lock#2 irq_context: 0 &dev->clientlist_mutex registration_lock console_lock &helper->lock &dev->master_mutex &client->modeset_mutex crtc_ww_class_acquire crtc_ww_class_mutex &obj_hash[i].lock irq_context: 0 &dev->clientlist_mutex registration_lock console_lock &helper->lock &dev->master_mutex &client->modeset_mutex crtc_ww_class_acquire crtc_ww_class_mutex &dev->mode_config.idr_mutex irq_context: 0 &dev->clientlist_mutex registration_lock console_lock &helper->lock &dev->master_mutex &client->modeset_mutex crtc_ww_class_acquire crtc_ww_class_mutex &dev->mode_config.blob_lock irq_context: 0 &dev->clientlist_mutex registration_lock console_lock &helper->lock &dev->master_mutex &client->modeset_mutex crtc_ww_class_acquire crtc_ww_class_mutex drm_connector_list_iter &dev->mode_config.connector_list_lock irq_context: 0 &dev->clientlist_mutex registration_lock console_lock &helper->lock &dev->master_mutex &client->modeset_mutex crtc_ww_class_acquire crtc_ww_class_mutex &crtc->commit_lock irq_context: 0 &dev->clientlist_mutex registration_lock console_lock &helper->lock &dev->master_mutex &client->modeset_mutex crtc_ww_class_acquire crtc_ww_class_mutex reservation_ww_class_mutex irq_context: 0 &dev->clientlist_mutex registration_lock console_lock &helper->lock &dev->master_mutex &client->modeset_mutex crtc_ww_class_acquire crtc_ww_class_mutex reservation_ww_class_mutex &shmem->vmap_lock irq_context: 0 &dev->clientlist_mutex registration_lock console_lock &helper->lock &dev->master_mutex &client->modeset_mutex crtc_ww_class_acquire crtc_ww_class_mutex reservation_ww_class_mutex &shmem->vmap_lock &shmem->pages_lock irq_context: 0 &dev->clientlist_mutex registration_lock console_lock &helper->lock &dev->master_mutex &client->modeset_mutex crtc_ww_class_acquire crtc_ww_class_mutex reservation_ww_class_mutex &shmem->vmap_lock &shmem->pages_lock fs_reclaim irq_context: 0 &dev->clientlist_mutex registration_lock console_lock &helper->lock &dev->master_mutex &client->modeset_mutex crtc_ww_class_acquire crtc_ww_class_mutex reservation_ww_class_mutex &shmem->vmap_lock &shmem->pages_lock fs_reclaim mmu_notifier_invalidate_range_start irq_context: 0 &dev->clientlist_mutex registration_lock console_lock &helper->lock &dev->master_mutex &client->modeset_mutex crtc_ww_class_acquire crtc_ww_class_mutex reservation_ww_class_mutex &shmem->vmap_lock &shmem->pages_lock pool_lock#2 irq_context: 0 &dev->clientlist_mutex registration_lock console_lock &helper->lock &dev->master_mutex &client->modeset_mutex crtc_ww_class_acquire crtc_ww_class_mutex reservation_ww_class_mutex &shmem->vmap_lock &shmem->pages_lock &sb->s_type->i_lock_key irq_context: 0 &dev->clientlist_mutex registration_lock console_lock &helper->lock &dev->master_mutex &client->modeset_mutex crtc_ww_class_acquire crtc_ww_class_mutex reservation_ww_class_mutex &shmem->vmap_lock &shmem->pages_lock &pcp->lock &zone->lock irq_context: 0 &dev->clientlist_mutex registration_lock console_lock &helper->lock &dev->master_mutex &client->modeset_mutex crtc_ww_class_acquire crtc_ww_class_mutex reservation_ww_class_mutex &shmem->vmap_lock &shmem->pages_lock &zone->lock irq_context: 0 &dev->clientlist_mutex registration_lock console_lock &helper->lock &dev->master_mutex &client->modeset_mutex crtc_ww_class_acquire crtc_ww_class_mutex reservation_ww_class_mutex &shmem->vmap_lock &shmem->pages_lock &____s->seqcount irq_context: 0 &dev->clientlist_mutex registration_lock console_lock &helper->lock &dev->master_mutex &client->modeset_mutex crtc_ww_class_acquire crtc_ww_class_mutex reservation_ww_class_mutex &shmem->vmap_lock &shmem->pages_lock &xa->xa_lock#9 irq_context: 0 &dev->clientlist_mutex registration_lock console_lock &helper->lock &dev->master_mutex &client->modeset_mutex crtc_ww_class_acquire crtc_ww_class_mutex reservation_ww_class_mutex &shmem->vmap_lock &shmem->pages_lock lock#4 irq_context: 0 &dev->clientlist_mutex registration_lock console_lock &helper->lock &dev->master_mutex &client->modeset_mutex crtc_ww_class_acquire crtc_ww_class_mutex reservation_ww_class_mutex &shmem->vmap_lock &shmem->pages_lock &info->lock irq_context: 0 &dev->clientlist_mutex registration_lock console_lock &helper->lock &dev->master_mutex &client->modeset_mutex crtc_ww_class_acquire crtc_ww_class_mutex reservation_ww_class_mutex &shmem->vmap_lock &shmem->pages_lock &xa->xa_lock#9 pool_lock#2 irq_context: 0 &dev->clientlist_mutex registration_lock console_lock &helper->lock &dev->master_mutex &client->modeset_mutex crtc_ww_class_acquire crtc_ww_class_mutex reservation_ww_class_mutex &shmem->vmap_lock &shmem->pages_lock lock#4 &lruvec->lru_lock irq_context: 0 &dev->clientlist_mutex registration_lock console_lock &helper->lock &dev->master_mutex &client->modeset_mutex crtc_ww_class_acquire crtc_ww_class_mutex reservation_ww_class_mutex &shmem->vmap_lock &shmem->pages_lock &pcp->lock &zone->lock &____s->seqcount irq_context: 0 &dev->clientlist_mutex registration_lock console_lock &helper->lock &dev->master_mutex &client->modeset_mutex crtc_ww_class_acquire crtc_ww_class_mutex reservation_ww_class_mutex &shmem->vmap_lock &shmem->pages_lock &rq->__lock irq_context: 0 &dev->clientlist_mutex registration_lock console_lock &helper->lock &dev->master_mutex &client->modeset_mutex crtc_ww_class_acquire crtc_ww_class_mutex reservation_ww_class_mutex &shmem->vmap_lock &shmem->pages_lock &xa->xa_lock#9 batched_entropy_u8.lock irq_context: 0 &dev->clientlist_mutex registration_lock console_lock &helper->lock &dev->master_mutex &client->modeset_mutex crtc_ww_class_acquire crtc_ww_class_mutex reservation_ww_class_mutex &shmem->vmap_lock &shmem->pages_lock &xa->xa_lock#9 kfence_freelist_lock irq_context: 0 &dev->clientlist_mutex registration_lock console_lock &helper->lock &dev->master_mutex &client->modeset_mutex crtc_ww_class_acquire crtc_ww_class_mutex reservation_ww_class_mutex &shmem->vmap_lock fs_reclaim irq_context: 0 &dev->clientlist_mutex registration_lock console_lock &helper->lock &dev->master_mutex &client->modeset_mutex crtc_ww_class_acquire crtc_ww_class_mutex reservation_ww_class_mutex &shmem->vmap_lock fs_reclaim mmu_notifier_invalidate_range_start irq_context: 0 &dev->clientlist_mutex registration_lock console_lock &helper->lock &dev->master_mutex &client->modeset_mutex crtc_ww_class_acquire crtc_ww_class_mutex reservation_ww_class_mutex &shmem->vmap_lock pool_lock#2 irq_context: 0 &dev->clientlist_mutex registration_lock console_lock &helper->lock &dev->master_mutex &client->modeset_mutex crtc_ww_class_acquire crtc_ww_class_mutex reservation_ww_class_mutex &shmem->vmap_lock free_vmap_area_lock irq_context: 0 &dev->clientlist_mutex registration_lock console_lock &helper->lock &dev->master_mutex &client->modeset_mutex crtc_ww_class_acquire crtc_ww_class_mutex reservation_ww_class_mutex &shmem->vmap_lock vmap_area_lock irq_context: 0 &dev->clientlist_mutex registration_lock console_lock &helper->lock &dev->master_mutex &client->modeset_mutex crtc_ww_class_acquire crtc_ww_class_mutex reservation_ww_class_mutex &shmem->vmap_lock &pcp->lock &zone->lock irq_context: 0 &dev->clientlist_mutex registration_lock console_lock &helper->lock &dev->master_mutex &client->modeset_mutex crtc_ww_class_acquire crtc_ww_class_mutex reservation_ww_class_mutex &shmem->vmap_lock &zone->lock irq_context: 0 &dev->clientlist_mutex registration_lock console_lock &helper->lock &dev->master_mutex &client->modeset_mutex crtc_ww_class_acquire crtc_ww_class_mutex reservation_ww_class_mutex &shmem->vmap_lock &____s->seqcount irq_context: 0 &dev->clientlist_mutex registration_lock console_lock &helper->lock &dev->master_mutex &client->modeset_mutex crtc_ww_class_acquire crtc_ww_class_mutex reservation_ww_class_mutex &shmem->vmap_lock init_mm.page_table_lock irq_context: 0 &dev->clientlist_mutex registration_lock console_lock &helper->lock &dev->master_mutex &client->modeset_mutex crtc_ww_class_acquire crtc_ww_class_mutex reservation_ww_class_mutex &shmem->vmap_lock &pcp->lock &zone->lock &____s->seqcount irq_context: 0 &dev->clientlist_mutex registration_lock console_lock &helper->lock &dev->master_mutex &client->modeset_mutex crtc_ww_class_acquire crtc_ww_class_mutex tk_core.seq.seqcount irq_context: 0 &dev->clientlist_mutex registration_lock console_lock &helper->lock &dev->master_mutex &client->modeset_mutex crtc_ww_class_acquire crtc_ww_class_mutex &vkms_out->lock irq_context: 0 &dev->clientlist_mutex registration_lock console_lock &helper->lock &dev->master_mutex &client->modeset_mutex crtc_ww_class_acquire crtc_ww_class_mutex &vkms_out->lock &dev->event_lock irq_context: 0 &dev->clientlist_mutex registration_lock console_lock &helper->lock &dev->master_mutex &client->modeset_mutex crtc_ww_class_acquire crtc_ww_class_mutex &vkms_out->lock &dev->event_lock &dev->vbl_lock irq_context: 0 &dev->clientlist_mutex registration_lock console_lock &helper->lock &dev->master_mutex &client->modeset_mutex crtc_ww_class_acquire crtc_ww_class_mutex &vkms_out->lock &dev->event_lock &____s->seqcount#6 irq_context: 0 &dev->clientlist_mutex registration_lock console_lock &helper->lock &dev->master_mutex &client->modeset_mutex crtc_ww_class_acquire crtc_ww_class_mutex &vkms_out->lock &dev->event_lock &x->wait#14 irq_context: 0 &dev->clientlist_mutex registration_lock console_lock &helper->lock &dev->master_mutex &client->modeset_mutex crtc_ww_class_acquire crtc_ww_class_mutex &vkms_out->lock &dev->event_lock &obj_hash[i].lock irq_context: 0 &dev->clientlist_mutex registration_lock console_lock &helper->lock &dev->master_mutex &client->modeset_mutex crtc_ww_class_acquire crtc_ww_class_mutex &vkms_out->lock &dev->event_lock pool_lock#2 irq_context: 0 &dev->clientlist_mutex registration_lock console_lock &helper->lock &dev->master_mutex &client->modeset_mutex crtc_ww_class_acquire crtc_ww_class_mutex &dev->vbl_lock irq_context: 0 &dev->clientlist_mutex registration_lock console_lock &helper->lock &dev->master_mutex &client->modeset_mutex crtc_ww_class_acquire crtc_ww_class_mutex &dev->vbl_lock &dev->vblank_time_lock irq_context: 0 &dev->clientlist_mutex registration_lock console_lock &helper->lock &dev->master_mutex &client->modeset_mutex crtc_ww_class_acquire crtc_ww_class_mutex &dev->vbl_lock &dev->vblank_time_lock tk_core.seq.seqcount irq_context: 0 &dev->clientlist_mutex registration_lock console_lock &helper->lock &dev->master_mutex &client->modeset_mutex crtc_ww_class_acquire crtc_ww_class_mutex &dev->vbl_lock &dev->vblank_time_lock &(&vblank->seqlock)->lock irq_context: 0 &dev->clientlist_mutex registration_lock console_lock &helper->lock &dev->master_mutex &client->modeset_mutex crtc_ww_class_acquire crtc_ww_class_mutex &dev->vbl_lock &dev->vblank_time_lock &(&vblank->seqlock)->lock &____s->seqcount#6 irq_context: 0 &dev->clientlist_mutex registration_lock console_lock &helper->lock &dev->master_mutex &client->modeset_mutex crtc_ww_class_acquire crtc_ww_class_mutex &x->wait#14 irq_context: 0 &dev->clientlist_mutex registration_lock console_lock &helper->lock &dev->master_mutex &client->modeset_mutex crtc_ww_class_acquire crtc_ww_class_mutex (work_completion)(&vkms_state->composer_work) irq_context: 0 &dev->clientlist_mutex registration_lock console_lock &helper->lock &dev->master_mutex &client->modeset_mutex crtc_ww_class_acquire crtc_ww_class_mutex &obj_hash[i].lock pool_lock irq_context: 0 &dev->clientlist_mutex registration_lock console_lock &helper->damage_lock irq_context: 0 &dev->clientlist_mutex registration_lock console_lock rcu_read_lock &pool->lock irq_context: 0 &dev->clientlist_mutex registration_lock console_lock rcu_read_lock &pool->lock &obj_hash[i].lock irq_context: 0 &dev->clientlist_mutex registration_lock console_lock rcu_read_lock &pool->lock pool_lock#2 irq_context: 0 &dev->clientlist_mutex registration_lock console_lock rcu_read_lock &pool->lock &p->pi_lock irq_context: 0 &dev->clientlist_mutex registration_lock console_lock rcu_read_lock &pool->lock &p->pi_lock &rq->__lock irq_context: 0 &dev->clientlist_mutex registration_lock console_lock rcu_read_lock &pool->lock &p->pi_lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 &dev->clientlist_mutex registration_lock console_lock &helper->lock &rq->__lock irq_context: 0 (wq_completion)events (work_completion)(&helper->damage_work) irq_context: 0 (wq_completion)events (work_completion)(&helper->damage_work) &helper->damage_lock irq_context: 0 (wq_completion)events (work_completion)(&helper->damage_work) &helper->lock irq_context: 0 (wq_completion)events (work_completion)(&helper->damage_work) &helper->lock &lock->wait_lock irq_context: 0 (wq_completion)events (work_completion)(&helper->damage_work) &helper->lock &pool->lock irq_context: 0 (wq_completion)events (work_completion)(&helper->damage_work) &helper->lock &rq->__lock irq_context: 0 (wq_completion)events (work_completion)(&helper->damage_work) &helper->lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 &dev->clientlist_mutex registration_lock console_lock &helper->lock &dev->master_mutex &client->modeset_mutex crtc_ww_class_acquire crtc_ww_class_mutex drm_connector_list_iter fs_reclaim irq_context: 0 &dev->clientlist_mutex registration_lock console_lock &helper->lock &dev->master_mutex &client->modeset_mutex crtc_ww_class_acquire crtc_ww_class_mutex drm_connector_list_iter fs_reclaim mmu_notifier_invalidate_range_start irq_context: 0 &dev->clientlist_mutex registration_lock console_lock &helper->lock &dev->master_mutex &client->modeset_mutex crtc_ww_class_acquire crtc_ww_class_mutex drm_connector_list_iter pool_lock#2 irq_context: 0 &dev->clientlist_mutex registration_lock console_lock &helper->lock &dev->master_mutex &client->modeset_mutex crtc_ww_class_acquire crtc_ww_class_mutex &vkms_out->lock &dev->event_lock &dev->vbl_lock &dev->vblank_time_lock irq_context: 0 &dev->clientlist_mutex registration_lock console_lock &helper->lock &dev->master_mutex &client->modeset_mutex crtc_ww_class_acquire crtc_ww_class_mutex &vkms_out->lock &dev->event_lock &dev->vbl_lock &dev->vblank_time_lock &obj_hash[i].lock irq_context: 0 &dev->clientlist_mutex registration_lock console_lock &helper->lock &dev->master_mutex &client->modeset_mutex crtc_ww_class_acquire crtc_ww_class_mutex &vkms_out->lock &dev->event_lock &dev->vbl_lock &dev->vblank_time_lock hrtimer_bases.lock irq_context: 0 &dev->clientlist_mutex registration_lock console_lock &helper->lock &dev->master_mutex &client->modeset_mutex crtc_ww_class_acquire crtc_ww_class_mutex &vkms_out->lock &dev->event_lock &dev->vbl_lock &dev->vblank_time_lock hrtimer_bases.lock tk_core.seq.seqcount irq_context: 0 &dev->clientlist_mutex registration_lock console_lock &helper->lock &dev->master_mutex &client->modeset_mutex crtc_ww_class_acquire crtc_ww_class_mutex &vkms_out->lock &dev->event_lock &dev->vbl_lock &dev->vblank_time_lock hrtimer_bases.lock &obj_hash[i].lock irq_context: 0 &dev->clientlist_mutex registration_lock console_lock &helper->lock &dev->master_mutex &client->modeset_mutex crtc_ww_class_acquire crtc_ww_class_mutex &vkms_out->lock &dev->event_lock &dev->vbl_lock &dev->vblank_time_lock tk_core.seq.seqcount irq_context: 0 &dev->clientlist_mutex registration_lock console_lock &helper->lock &dev->master_mutex &client->modeset_mutex crtc_ww_class_acquire crtc_ww_class_mutex &vkms_out->lock &dev->event_lock &dev->vblank_time_lock irq_context: 0 &dev->clientlist_mutex registration_lock console_lock &helper->lock &dev->master_mutex &client->modeset_mutex crtc_ww_class_acquire crtc_ww_class_mutex &base->lock irq_context: 0 &dev->clientlist_mutex registration_lock console_lock &helper->lock &dev->master_mutex &client->modeset_mutex crtc_ww_class_acquire crtc_ww_class_mutex &base->lock &obj_hash[i].lock irq_context: 0 &dev->clientlist_mutex registration_lock console_lock &helper->lock &dev->master_mutex &client->modeset_mutex crtc_ww_class_acquire crtc_ww_class_mutex &rq->__lock irq_context: 0 &dev->clientlist_mutex registration_lock console_lock &helper->lock &dev->master_mutex &client->modeset_mutex crtc_ww_class_acquire crtc_ww_class_mutex &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: hardirq &vkms_out->lock irq_context: hardirq &vkms_out->lock &dev->event_lock irq_context: hardirq &vkms_out->lock &dev->event_lock &dev->vblank_time_lock irq_context: hardirq &vkms_out->lock &dev->event_lock &dev->vblank_time_lock &(&vblank->seqlock)->lock irq_context: hardirq &vkms_out->lock &dev->event_lock &dev->vblank_time_lock &(&vblank->seqlock)->lock &____s->seqcount#6 irq_context: hardirq &vkms_out->lock &dev->event_lock &vblank->queue irq_context: hardirq &vkms_out->lock &dev->event_lock &____s->seqcount#6 irq_context: hardirq &vkms_out->lock &dev->event_lock &obj_hash[i].lock irq_context: hardirq &vkms_out->lock &dev->event_lock &base->lock irq_context: hardirq &vkms_out->lock &dev->event_lock &base->lock &obj_hash[i].lock irq_context: hardirq &vkms_out->lock &dev->event_lock &x->wait#14 irq_context: hardirq &vkms_out->lock &dev->event_lock &x->wait#14 &p->pi_lock irq_context: hardirq &vkms_out->lock &dev->event_lock &x->wait#14 &p->pi_lock &rq->__lock irq_context: hardirq &vkms_out->lock &dev->event_lock &x->wait#14 &p->pi_lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: hardirq &vkms_out->lock &dev->event_lock pool_lock#2 irq_context: 0 &dev->clientlist_mutex registration_lock console_lock &helper->lock &dev->master_mutex &client->modeset_mutex crtc_ww_class_acquire crtc_ww_class_mutex (&timer.timer) irq_context: 0 &dev->clientlist_mutex registration_lock console_lock &helper->lock &dev->master_mutex &client->modeset_mutex crtc_ww_class_acquire crtc_ww_class_mutex (work_completion)(&vkms_state->composer_work)#2 irq_context: 0 &dev->clientlist_mutex registration_lock console_lock &helper->lock &dev->master_mutex &client->modeset_mutex crtc_ww_class_acquire crtc_ww_class_mutex &pcp->lock &zone->lock irq_context: 0 &dev->clientlist_mutex registration_lock console_lock &helper->lock &dev->master_mutex &client->modeset_mutex crtc_ww_class_acquire crtc_ww_class_mutex &zone->lock irq_context: 0 &dev->clientlist_mutex registration_lock console_lock &helper->lock &dev->master_mutex &client->modeset_mutex crtc_ww_class_acquire crtc_ww_class_mutex &____s->seqcount irq_context: 0 &dev->clientlist_mutex registration_lock console_lock &helper->lock &dev->master_mutex &client->modeset_mutex crtc_ww_class_acquire crtc_ww_class_mutex rcu_read_lock pool_lock#2 irq_context: 0 &dev->clientlist_mutex registration_lock console_lock &lock->wait_lock irq_context: 0 &dev->clientlist_mutex registration_lock console_lock &p->pi_lock irq_context: 0 &dev->clientlist_mutex registration_lock console_lock &p->pi_lock &rq->__lock irq_context: 0 &dev->clientlist_mutex registration_lock console_lock &p->pi_lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 &dev->clientlist_mutex registration_lock console_lock &rq->__lock irq_context: 0 (wq_completion)events (work_completion)(&helper->damage_work) &helper->lock reservation_ww_class_mutex irq_context: 0 (wq_completion)events (work_completion)(&helper->damage_work) &helper->lock reservation_ww_class_mutex &shmem->vmap_lock irq_context: 0 &dev->clientlist_mutex registration_lock console_lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 &dev->clientlist_mutex registration_lock console_lock vt_event_lock irq_context: hardirq &vkms_out->lock &dev->event_lock &x->wait#14 &p->pi_lock &cfs_rq->removed.lock irq_context: 0 &dev->clientlist_mutex registration_lock console_lock (console_sem).lock irq_context: 0 &dev->clientlist_mutex registration_lock console_lock console_owner_lock irq_context: 0 &dev->clientlist_mutex registration_lock console_lock &helper->lock &dev->master_mutex &client->modeset_mutex crtc_ww_class_acquire crtc_ww_class_mutex &c->lock irq_context: 0 &pool->lock &base->lock irq_context: 0 &pool->lock &base->lock &obj_hash[i].lock irq_context: 0 &dev->clientlist_mutex registration_lock console_lock console_srcu console_owner_lock irq_context: 0 &dev->clientlist_mutex registration_lock console_lock console_srcu console_owner irq_context: 0 &dev->clientlist_mutex registration_lock console_lock console_srcu console_owner &port_lock_key irq_context: 0 &dev->clientlist_mutex registration_lock console_lock console_srcu console_owner console_owner_lock irq_context: 0 &dev->clientlist_mutex (console_sem).lock irq_context: 0 &dev->clientlist_mutex console_lock console_srcu console_owner_lock irq_context: 0 &dev->clientlist_mutex console_lock console_srcu console_owner irq_context: 0 &dev->clientlist_mutex console_lock console_srcu console_owner &port_lock_key irq_context: 0 &dev->clientlist_mutex console_lock console_srcu console_owner console_owner_lock irq_context: 0 &dev->clientlist_mutex kernel_fb_helper_lock irq_context: 0 &dev->queue_lock irq_context: 0 cpu_hotplug_lock cpuhp_state_mutex lock irq_context: 0 cpu_hotplug_lock cpuhp_state_mutex lock kernfs_idr_lock irq_context: 0 cpu_hotplug_lock cpuhp_state_mutex &root->kernfs_rwsem irq_context: 0 cpu_hotplug_lock cpuhp_state_mutex &root->kernfs_rwsem &root->kernfs_iattr_rwsem irq_context: 0 cpu_hotplug_lock cpuhp_state_mutex &c->lock irq_context: 0 cpu_hotplug_lock cpuhp_state_mutex &____s->seqcount irq_context: 0 &sig->cred_guard_mutex &mm->mmap_lock &cfs_rq->removed.lock irq_context: 0 &sig->cred_guard_mutex &mm->mmap_lock &obj_hash[i].lock irq_context: 0 &sig->cred_guard_mutex &mm->mmap_lock pool_lock#2 irq_context: 0 cpu_hotplug_lock cpuhp_state-up &k->k_lock irq_context: 0 cpu_hotplug_lock cpuhp_state_mutex &rq->__lock &cfs_rq->removed.lock irq_context: 0 cpu_hotplug_lock cpuhp_state-up lock kernfs_idr_lock pool_lock#2 irq_context: 0 cpu_hotplug_lock cpuhp_state-up &pcp->lock &zone->lock &____s->seqcount irq_context: 0 (wq_completion)events_unbound (work_completion)(&entry->work) cpu_hotplug_lock irq_context: 0 (wq_completion)events_unbound (work_completion)(&entry->work) cpu_hotplug_lock wq_pool_mutex irq_context: 0 (wq_completion)events_unbound (work_completion)(&entry->work) cpu_hotplug_lock wq_pool_mutex fs_reclaim irq_context: 0 (wq_completion)events_unbound (work_completion)(&entry->work) cpu_hotplug_lock wq_pool_mutex fs_reclaim mmu_notifier_invalidate_range_start irq_context: 0 (wq_completion)events_unbound (work_completion)(&entry->work) cpu_hotplug_lock wq_pool_mutex &pcp->lock &zone->lock irq_context: 0 (wq_completion)events_unbound (work_completion)(&entry->work) cpu_hotplug_lock wq_pool_mutex &zone->lock irq_context: 0 (wq_completion)events_unbound (work_completion)(&entry->work) cpu_hotplug_lock wq_pool_mutex &pcp->lock &zone->lock &____s->seqcount irq_context: 0 (wq_completion)events_unbound (work_completion)(&entry->work) cpu_hotplug_lock wq_pool_mutex &____s->seqcount irq_context: 0 (wq_completion)events_unbound (work_completion)(&entry->work) cpu_hotplug_lock wq_pool_mutex pool_lock#2 irq_context: 0 (wq_completion)events_unbound (work_completion)(&entry->work) cpu_hotplug_lock wq_pool_mutex rcu_read_lock pool_lock#2 irq_context: 0 (wq_completion)events_unbound (work_completion)(&entry->work) cpu_hotplug_lock wq_pool_mutex &obj_hash[i].lock irq_context: 0 (wq_completion)events_unbound (work_completion)(&entry->work) cpu_hotplug_lock wq_pool_mutex &wq->mutex irq_context: 0 (wq_completion)events_unbound (work_completion)(&entry->work) cpu_hotplug_lock wq_pool_mutex &wq->mutex &pool->lock/1 irq_context: 0 (wq_completion)events_unbound (work_completion)(&entry->work) wq_pool_mutex irq_context: 0 (wq_completion)events_unbound (work_completion)(&entry->work) wq_pool_mutex &wq->mutex irq_context: 0 (wq_completion)events_unbound (work_completion)(&entry->work) pcpu_alloc_mutex irq_context: 0 (wq_completion)events_unbound (work_completion)(&entry->work) pcpu_alloc_mutex pcpu_lock irq_context: 0 (wq_completion)events_unbound (work_completion)(&entry->work) batched_entropy_u32.lock irq_context: 0 (wq_completion)events_unbound (work_completion)(&entry->work) &n->list_lock irq_context: 0 (wq_completion)events_unbound (work_completion)(&entry->work) mmu_notifier_invalidate_range_start irq_context: 0 (wq_completion)events_unbound (work_completion)(&entry->work) blk_queue_ida.xa_lock irq_context: 0 (wq_completion)events_unbound (work_completion)(&entry->work) &q->sysfs_lock irq_context: 0 (wq_completion)events_unbound (work_completion)(&entry->work) &q->sysfs_lock &q->unused_hctx_lock irq_context: 0 (wq_completion)events_unbound (work_completion)(&entry->work) &q->sysfs_lock mmu_notifier_invalidate_range_start irq_context: 0 (wq_completion)events_unbound (work_completion)(&entry->work) &q->sysfs_lock pool_lock#2 irq_context: 0 (wq_completion)events_unbound (work_completion)(&entry->work) &q->sysfs_lock &obj_hash[i].lock irq_context: 0 (wq_completion)events_unbound (work_completion)(&entry->work) &q->sysfs_lock &obj_hash[i].lock pool_lock irq_context: 0 &sb->s_type->i_mutex_key#3 &cfs_rq->removed.lock irq_context: 0 (wq_completion)events_unbound (work_completion)(&entry->work) &q->sysfs_lock &c->lock irq_context: 0 (wq_completion)events_unbound (work_completion)(&entry->work) &q->sysfs_lock &n->list_lock irq_context: 0 (wq_completion)events_unbound (work_completion)(&entry->work) &q->sysfs_lock &pcp->lock &zone->lock irq_context: 0 (wq_completion)events_unbound (work_completion)(&entry->work) &q->sysfs_lock &zone->lock irq_context: 0 (wq_completion)events_unbound (work_completion)(&entry->work) &q->sysfs_lock &____s->seqcount irq_context: 0 (wq_completion)events_unbound (work_completion)(&entry->work) &q->sysfs_lock cpu_hotplug_lock irq_context: 0 (wq_completion)events_unbound (work_completion)(&entry->work) &q->sysfs_lock cpu_hotplug_lock cpuhp_state_mutex irq_context: 0 (wq_completion)events_unbound (work_completion)(&entry->work) &q->sysfs_lock fs_reclaim irq_context: 0 (wq_completion)events_unbound (work_completion)(&entry->work) &q->sysfs_lock fs_reclaim mmu_notifier_invalidate_range_start irq_context: 0 blk_queue_ida.xa_lock irq_context: 0 (wq_completion)events_unbound (work_completion)(&entry->work) &q->sysfs_lock &xa->xa_lock#10 irq_context: 0 (wq_completion)events_unbound (work_completion)(&entry->work) &set->tag_list_lock irq_context: 0 (wq_completion)events_unbound (work_completion)(&entry->work) bio_slab_lock irq_context: 0 (wq_completion)events_unbound (work_completion)(&entry->work) percpu_counters_lock irq_context: 0 &sb->s_type->i_lock_key#3 irq_context: 0 &xa->xa_lock#11 irq_context: 0 (wq_completion)events_unbound (work_completion)(&entry->work) &sb->s_type->i_lock_key#3 irq_context: 0 (wq_completion)events_unbound (work_completion)(&entry->work) &s->s_inode_list_lock irq_context: 0 (wq_completion)events_unbound (work_completion)(&entry->work) &xa->xa_lock#11 irq_context: 0 lock &q->queue_lock irq_context: 0 lock &q->queue_lock &blkcg->lock irq_context: 0 lock &q->queue_lock &blkcg->lock pool_lock#2 irq_context: 0 &q->queue_lock irq_context: 0 &q->queue_lock pool_lock#2 irq_context: 0 (wq_completion)events_unbound (work_completion)(&entry->work) lock &q->queue_lock irq_context: 0 (wq_completion)events_unbound (work_completion)(&entry->work) lock &q->queue_lock &blkcg->lock irq_context: 0 (wq_completion)events_unbound (work_completion)(&entry->work) &q->mq_freeze_lock irq_context: 0 &q->queue_lock pcpu_lock irq_context: 0 &q->queue_lock &obj_hash[i].lock irq_context: 0 (wq_completion)events_unbound (work_completion)(&entry->work) &q->mq_freeze_lock percpu_ref_switch_lock irq_context: 0 &q->queue_lock percpu_counters_lock irq_context: 0 (wq_completion)events_unbound (work_completion)(&entry->work) &q->mq_freeze_lock percpu_ref_switch_lock rcu_read_lock &q->mq_freeze_wq irq_context: 0 (wq_completion)events_unbound (work_completion)(&entry->work) percpu_ref_switch_lock irq_context: 0 (wq_completion)events_unbound (work_completion)(&entry->work) &q->queue_lock irq_context: 0 (wq_completion)events_unbound (work_completion)(&entry->work) &q->queue_lock pool_lock#2 irq_context: 0 (wq_completion)events_unbound (work_completion)(&entry->work) &q->mq_freeze_lock &q->mq_freeze_wq irq_context: 0 &q->queue_lock &obj_hash[i].lock pool_lock irq_context: 0 (wq_completion)events_unbound (work_completion)(&entry->work) &q->queue_lock pcpu_lock irq_context: 0 (wq_completion)events_unbound (work_completion)(&entry->work) &q->queue_lock &obj_hash[i].lock irq_context: 0 (wq_completion)events_unbound (work_completion)(&entry->work) &q->queue_lock &obj_hash[i].lock pool_lock irq_context: 0 (wq_completion)events_unbound (work_completion)(&entry->work) &q->queue_lock percpu_counters_lock irq_context: 0 &bdev->bd_size_lock irq_context: 0 (wq_completion)events_unbound (work_completion)(&entry->work) &x->wait#9 irq_context: 0 subsys mutex#37 irq_context: 0 subsys mutex#37 &k->k_lock irq_context: 0 dev_hotplug_mutex irq_context: 0 dev_hotplug_mutex &dev->power.lock irq_context: 0 &q->sysfs_dir_lock irq_context: 0 &q->sysfs_dir_lock fs_reclaim irq_context: 0 &q->sysfs_dir_lock fs_reclaim mmu_notifier_invalidate_range_start irq_context: 0 &q->sysfs_dir_lock pool_lock#2 irq_context: 0 &q->sysfs_dir_lock lock irq_context: 0 &q->sysfs_dir_lock lock kernfs_idr_lock irq_context: 0 &q->sysfs_dir_lock &root->kernfs_rwsem irq_context: 0 &q->sysfs_dir_lock &root->kernfs_rwsem &root->kernfs_iattr_rwsem irq_context: 0 &q->sysfs_dir_lock &c->lock irq_context: 0 &q->sysfs_dir_lock &____s->seqcount irq_context: 0 &q->sysfs_dir_lock &q->sysfs_lock irq_context: 0 &q->sysfs_dir_lock &q->sysfs_lock &q->debugfs_mutex irq_context: 0 &q->sysfs_dir_lock &q->sysfs_lock &q->debugfs_mutex pin_fs_lock irq_context: 0 &q->sysfs_dir_lock &q->sysfs_lock &q->debugfs_mutex &sb->s_type->i_mutex_key#3 irq_context: 0 &q->sysfs_dir_lock &q->sysfs_lock &q->debugfs_mutex &sb->s_type->i_mutex_key#3 &sb->s_type->i_lock_key#7 irq_context: 0 &q->sysfs_dir_lock &q->sysfs_lock &q->debugfs_mutex &sb->s_type->i_mutex_key#3 rename_lock.seqcount irq_context: 0 &q->sysfs_dir_lock &q->sysfs_lock &q->debugfs_mutex &sb->s_type->i_mutex_key#3 fs_reclaim irq_context: 0 &q->sysfs_dir_lock &q->sysfs_lock &q->debugfs_mutex &sb->s_type->i_mutex_key#3 fs_reclaim mmu_notifier_invalidate_range_start irq_context: 0 &q->sysfs_dir_lock &q->sysfs_lock &q->debugfs_mutex &sb->s_type->i_mutex_key#3 pool_lock#2 irq_context: 0 &q->sysfs_dir_lock &q->sysfs_lock &q->debugfs_mutex &sb->s_type->i_mutex_key#3 &dentry->d_lock irq_context: 0 &q->sysfs_dir_lock &q->sysfs_lock &q->debugfs_mutex &sb->s_type->i_mutex_key#3 rcu_read_lock rename_lock.seqcount irq_context: 0 &q->sysfs_dir_lock &q->sysfs_lock &q->debugfs_mutex &sb->s_type->i_mutex_key#3 &dentry->d_lock &wq irq_context: 0 &q->sysfs_dir_lock &q->sysfs_lock &q->debugfs_mutex &sb->s_type->i_mutex_key#3 mmu_notifier_invalidate_range_start irq_context: 0 &q->sysfs_dir_lock &q->sysfs_lock &q->debugfs_mutex &sb->s_type->i_mutex_key#3 &s->s_inode_list_lock irq_context: 0 &q->sysfs_dir_lock &q->sysfs_lock &q->debugfs_mutex &sb->s_type->i_mutex_key#3 tk_core.seq.seqcount irq_context: 0 &q->sysfs_dir_lock &q->sysfs_lock &q->debugfs_mutex &sb->s_type->i_mutex_key#3 &sb->s_type->i_lock_key#7 &dentry->d_lock irq_context: 0 percpu_ref_switch_lock irq_context: 0 subsys mutex#38 irq_context: 0 subsys mutex#38 &k->k_lock irq_context: 0 cgwb_lock irq_context: 0 bdi_lock irq_context: 0 inode_hash_lock irq_context: 0 inode_hash_lock &sb->s_type->i_lock_key#3 irq_context: 0 bdev_lock irq_context: 0 &disk->open_mutex irq_context: 0 &disk->open_mutex fs_reclaim irq_context: 0 &disk->open_mutex fs_reclaim mmu_notifier_invalidate_range_start irq_context: 0 &disk->open_mutex pool_lock#2 irq_context: 0 &disk->open_mutex &c->lock irq_context: 0 &disk->open_mutex &____s->seqcount irq_context: 0 &disk->open_mutex free_vmap_area_lock irq_context: 0 &disk->open_mutex vmap_area_lock irq_context: 0 &disk->open_mutex &pcp->lock &zone->lock irq_context: 0 &disk->open_mutex &zone->lock irq_context: 0 &disk->open_mutex init_mm.page_table_lock irq_context: 0 &disk->open_mutex &xa->xa_lock#9 irq_context: 0 &disk->open_mutex lock#4 irq_context: 0 &disk->open_mutex mmu_notifier_invalidate_range_start irq_context: 0 &disk->open_mutex &mapping->private_lock irq_context: 0 &disk->open_mutex tk_core.seq.seqcount irq_context: 0 &disk->open_mutex &ret->b_uptodate_lock irq_context: 0 &disk->open_mutex &obj_hash[i].lock irq_context: 0 &disk->open_mutex &xa->xa_lock#9 pool_lock#2 irq_context: 0 &disk->open_mutex rcu_read_lock pool_lock#2 irq_context: 0 &disk->open_mutex purge_vmap_area_lock irq_context: 0 &disk->open_mutex &sb->s_type->i_lock_key#3 irq_context: 0 &disk->open_mutex &sb->s_type->i_lock_key#3 &xa->xa_lock#9 irq_context: 0 &disk->open_mutex &sb->s_type->i_lock_key#3 &xa->xa_lock#9 &obj_hash[i].lock irq_context: 0 &disk->open_mutex &sb->s_type->i_lock_key#3 &xa->xa_lock#9 pool_lock#2 irq_context: 0 &disk->open_mutex lock#4 &lruvec->lru_lock irq_context: 0 &disk->open_mutex lock#5 irq_context: 0 &disk->open_mutex &lruvec->lru_lock irq_context: 0 (wq_completion)events_unbound (work_completion)(&entry->work) major_names_lock irq_context: 0 (wq_completion)events_unbound (work_completion)(&entry->work) major_names_lock fs_reclaim irq_context: 0 (wq_completion)events_unbound (work_completion)(&entry->work) major_names_lock fs_reclaim mmu_notifier_invalidate_range_start irq_context: 0 (wq_completion)events_unbound (work_completion)(&entry->work) major_names_lock pool_lock#2 irq_context: 0 (wq_completion)events_unbound (work_completion)(&entry->work) major_names_lock major_names_spinlock irq_context: 0 (wq_completion)events_unbound (work_completion)(&entry->work) floppy_lock irq_context: 0 (wq_completion)events_unbound (work_completion)(&entry->work) floppy_lock &obj_hash[i].lock irq_context: 0 (wq_completion)events_unbound (work_completion)(&entry->work) floppy_lock &base->lock irq_context: 0 (wq_completion)events_unbound (work_completion)(&entry->work) floppy_lock &base->lock &obj_hash[i].lock irq_context: 0 (wq_completion)events_unbound (work_completion)(&entry->work) rtc_lock irq_context: softirq rcu_callback rcu_read_lock &pool->lock pool_lock#2 irq_context: 0 (wq_completion)events_unbound (work_completion)(&entry->work) &wq->mutex irq_context: 0 (wq_completion)events_unbound (work_completion)(&entry->work) &wq->mutex &pool->lock/1 irq_context: 0 (wq_completion)events_unbound (work_completion)(&entry->work) &wq->mutex &x->wait#10 irq_context: 0 (wq_completion)events_unbound (work_completion)(&entry->work) &desc->request_mutex irq_context: 0 (wq_completion)events_unbound (work_completion)(&entry->work) &desc->request_mutex &irq_desc_lock_class irq_context: 0 (wq_completion)events_unbound (work_completion)(&entry->work) &desc->request_mutex &irq_desc_lock_class vector_lock irq_context: 0 (wq_completion)events_unbound (work_completion)(&entry->work) &desc->request_mutex &irq_desc_lock_class ioapic_lock irq_context: 0 (wq_completion)events_unbound (work_completion)(&entry->work) &desc->request_mutex &irq_desc_lock_class mask_lock irq_context: 0 (wq_completion)events_unbound (work_completion)(&entry->work) &desc->request_mutex &irq_desc_lock_class mask_lock tmp_mask_lock irq_context: 0 (wq_completion)events_unbound (work_completion)(&entry->work) &desc->request_mutex &irq_desc_lock_class mask_lock tmp_mask_lock vector_lock irq_context: 0 (wq_completion)events_unbound (work_completion)(&entry->work) &desc->request_mutex &irq_desc_lock_class mask_lock tmp_mask_lock ioapic_lock irq_context: 0 &q->sysfs_dir_lock lock kernfs_idr_lock pool_lock#2 irq_context: 0 (wq_completion)events_unbound (work_completion)(&entry->work) &desc->request_mutex &irq_desc_lock_class ioapic_lock i8259A_lock irq_context: 0 (wq_completion)events_unbound (work_completion)(&entry->work) register_lock irq_context: 0 &q->sysfs_dir_lock &pcp->lock &zone->lock irq_context: 0 (wq_completion)events_unbound (work_completion)(&entry->work) &irq_desc_lock_class irq_context: 0 &q->sysfs_dir_lock &zone->lock irq_context: 0 (wq_completion)events_unbound (work_completion)(&entry->work) proc_subdir_lock irq_context: 0 (wq_completion)events_unbound (work_completion)(&entry->work) proc_inum_ida.xa_lock irq_context: 0 (wq_completion)events_unbound (work_completion)(&entry->work) proc_subdir_lock irq_context: 0 (wq_completion)events_unbound (work_completion)(&entry->work) resource_lock irq_context: 0 (wq_completion)events_unbound (work_completion)(&entry->work) &base->lock irq_context: 0 (wq_completion)events_unbound (work_completion)(&entry->work) &base->lock &obj_hash[i].lock irq_context: 0 (wq_completion)events_unbound (work_completion)(&entry->work) &rq->__lock irq_context: 0 (wq_completion)events_unbound (work_completion)(&entry->work) &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 &q->sysfs_dir_lock &q->sysfs_lock &q->debugfs_mutex &sb->s_type->i_mutex_key#3 &c->lock irq_context: 0 &q->sysfs_dir_lock &q->sysfs_lock &q->debugfs_mutex &sb->s_type->i_mutex_key#3 &pcp->lock &zone->lock irq_context: 0 &q->sysfs_dir_lock &q->sysfs_lock &q->debugfs_mutex &sb->s_type->i_mutex_key#3 &zone->lock irq_context: 0 &q->sysfs_dir_lock &q->sysfs_lock &q->debugfs_mutex &sb->s_type->i_mutex_key#3 &____s->seqcount irq_context: 0 &disk->open_mutex purge_vmap_area_lock &obj_hash[i].lock irq_context: 0 &disk->open_mutex purge_vmap_area_lock pool_lock#2 irq_context: 0 &disk->open_mutex &xa->xa_lock#9 &c->lock irq_context: 0 &disk->open_mutex &xa->xa_lock#9 &pcp->lock &zone->lock irq_context: 0 &disk->open_mutex &xa->xa_lock#9 &zone->lock irq_context: 0 &disk->open_mutex &xa->xa_lock#9 &____s->seqcount irq_context: 0 &disk->open_mutex &pcp->lock &zone->lock &____s->seqcount irq_context: 0 &q->queue_lock fill_pool_map-wait-type-override pool_lock#2 irq_context: 0 &q->queue_lock fill_pool_map-wait-type-override pool_lock irq_context: 0 &q->queue_lock &c->lock irq_context: 0 &q->queue_lock &____s->seqcount irq_context: 0 (wq_completion)events_unbound (work_completion)(&entry->work) (&timer.timer) irq_context: 0 (wq_completion)events_unbound (work_completion)(&entry->work) rcu_read_lock &pool->lock/1 &p->pi_lock &rq->__lock irq_context: 0 (wq_completion)events_unbound (work_completion)(&entry->work) rcu_read_lock &pool->lock/1 &p->pi_lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)events (work_completion)(&blkg->free_work) irq_context: 0 (wq_completion)events_unbound (work_completion)(&entry->work) command_done.lock irq_context: 0 (wq_completion)events (work_completion)(&blkg->free_work) &q->blkcg_mutex irq_context: 0 (wq_completion)events (work_completion)(&blkg->free_work) &q->blkcg_mutex &obj_hash[i].lock irq_context: 0 (wq_completion)events_unbound (work_completion)(&entry->work) &cfs_rq->removed.lock irq_context: 0 (wq_completion)events_unbound (work_completion)(&entry->work) &rq->__lock &cfs_rq->removed.lock irq_context: 0 &disk->open_mutex &sb->s_type->i_lock_key#3 &xa->xa_lock#9 &obj_hash[i].lock pool_lock irq_context: 0 &disk->open_mutex &rq->__lock irq_context: 0 (wq_completion)events (work_completion)(&p->wq) purge_vmap_area_lock &obj_hash[i].lock irq_context: 0 (wq_completion)events (work_completion)(&p->wq) purge_vmap_area_lock pool_lock#2 irq_context: 0 &q->queue_lock &pcp->lock &zone->lock irq_context: 0 &q->queue_lock &zone->lock irq_context: 0 loop_ctl_mutex irq_context: 0 loop_ctl_mutex fs_reclaim irq_context: 0 loop_ctl_mutex fs_reclaim mmu_notifier_invalidate_range_start irq_context: 0 loop_ctl_mutex pool_lock#2 irq_context: 0 &q->sysfs_lock irq_context: 0 &q->sysfs_lock &q->unused_hctx_lock irq_context: 0 &q->sysfs_lock mmu_notifier_invalidate_range_start irq_context: 0 &q->sysfs_lock pool_lock#2 irq_context: 0 &q->sysfs_lock &obj_hash[i].lock irq_context: 0 &q->sysfs_lock &c->lock irq_context: 0 &q->sysfs_lock &n->list_lock irq_context: 0 &q->sysfs_lock &pcp->lock &zone->lock irq_context: 0 &q->sysfs_lock &zone->lock irq_context: 0 &q->sysfs_lock &____s->seqcount irq_context: 0 &q->sysfs_lock cpu_hotplug_lock irq_context: 0 &q->sysfs_lock cpu_hotplug_lock cpuhp_state_mutex irq_context: 0 &q->sysfs_lock fs_reclaim irq_context: 0 &q->sysfs_lock fs_reclaim mmu_notifier_invalidate_range_start irq_context: 0 &q->sysfs_lock &xa->xa_lock#10 irq_context: 0 &set->tag_list_lock irq_context: 0 &q->mq_freeze_lock irq_context: 0 &q->mq_freeze_lock percpu_ref_switch_lock irq_context: 0 &q->mq_freeze_lock percpu_ref_switch_lock rcu_read_lock &q->mq_freeze_wq irq_context: 0 &q->mq_freeze_lock &q->mq_freeze_wq irq_context: 0 &q->sysfs_dir_lock &q->sysfs_lock fs_reclaim irq_context: 0 &q->sysfs_dir_lock &q->sysfs_lock fs_reclaim mmu_notifier_invalidate_range_start irq_context: 0 &q->sysfs_dir_lock &q->sysfs_lock pool_lock#2 irq_context: 0 &q->sysfs_dir_lock &q->sysfs_lock pcpu_alloc_mutex irq_context: 0 &q->sysfs_dir_lock &q->sysfs_lock pcpu_alloc_mutex pcpu_lock irq_context: 0 &q->sysfs_dir_lock &q->sysfs_lock &obj_hash[i].lock irq_context: 0 &q->sysfs_dir_lock &q->sysfs_lock &q->rq_qos_mutex irq_context: 0 &q->sysfs_dir_lock &q->sysfs_lock &q->rq_qos_mutex &q->mq_freeze_lock irq_context: 0 &q->sysfs_dir_lock &q->sysfs_lock &q->rq_qos_mutex &q->mq_freeze_lock percpu_ref_switch_lock irq_context: 0 &q->sysfs_dir_lock &q->sysfs_lock &q->rq_qos_mutex &q->mq_freeze_lock percpu_ref_switch_lock rcu_read_lock &q->mq_freeze_wq irq_context: 0 &q->sysfs_dir_lock &q->sysfs_lock &q->rq_qos_mutex percpu_ref_switch_lock irq_context: 0 &q->sysfs_dir_lock &q->sysfs_lock &q->rq_qos_mutex &q->mq_freeze_lock &q->mq_freeze_wq irq_context: 0 &q->sysfs_dir_lock &q->sysfs_lock &q->rq_qos_mutex &q->debugfs_mutex irq_context: 0 &q->sysfs_dir_lock &q->sysfs_lock &q->rq_qos_mutex &q->debugfs_mutex pin_fs_lock irq_context: 0 &q->sysfs_dir_lock &q->sysfs_lock &q->rq_qos_mutex &q->debugfs_mutex &sb->s_type->i_mutex_key#3 irq_context: 0 &q->sysfs_dir_lock &q->sysfs_lock &q->rq_qos_mutex &q->debugfs_mutex &sb->s_type->i_mutex_key#3 rename_lock.seqcount irq_context: 0 &q->sysfs_dir_lock &q->sysfs_lock &q->rq_qos_mutex &q->debugfs_mutex &sb->s_type->i_mutex_key#3 fs_reclaim irq_context: 0 &q->sysfs_dir_lock &q->sysfs_lock &q->rq_qos_mutex &q->debugfs_mutex &sb->s_type->i_mutex_key#3 fs_reclaim mmu_notifier_invalidate_range_start irq_context: 0 &q->sysfs_dir_lock &q->sysfs_lock &q->rq_qos_mutex &q->debugfs_mutex &sb->s_type->i_mutex_key#3 pool_lock#2 irq_context: 0 &q->sysfs_dir_lock &q->sysfs_lock &q->rq_qos_mutex &q->debugfs_mutex &sb->s_type->i_mutex_key#3 &dentry->d_lock irq_context: 0 &q->sysfs_dir_lock &q->sysfs_lock &q->rq_qos_mutex &q->debugfs_mutex &sb->s_type->i_mutex_key#3 rcu_read_lock rename_lock.seqcount irq_context: 0 &q->sysfs_dir_lock &q->sysfs_lock &q->rq_qos_mutex &q->debugfs_mutex &sb->s_type->i_mutex_key#3 &dentry->d_lock &wq irq_context: 0 &q->sysfs_dir_lock &q->sysfs_lock &q->rq_qos_mutex &q->debugfs_mutex &sb->s_type->i_mutex_key#3 mmu_notifier_invalidate_range_start irq_context: 0 &q->sysfs_dir_lock &q->sysfs_lock &q->rq_qos_mutex &q->debugfs_mutex &sb->s_type->i_mutex_key#3 &sb->s_type->i_lock_key#7 irq_context: 0 &q->sysfs_dir_lock &q->sysfs_lock &q->rq_qos_mutex &q->debugfs_mutex &sb->s_type->i_mutex_key#3 &s->s_inode_list_lock irq_context: 0 &q->sysfs_dir_lock &q->sysfs_lock &q->rq_qos_mutex &q->debugfs_mutex &sb->s_type->i_mutex_key#3 tk_core.seq.seqcount irq_context: 0 &q->sysfs_dir_lock &q->sysfs_lock &q->rq_qos_mutex &q->debugfs_mutex &sb->s_type->i_mutex_key#3 &sb->s_type->i_lock_key#7 &dentry->d_lock irq_context: 0 &q->sysfs_dir_lock &q->sysfs_lock &q->rq_qos_mutex &q->debugfs_mutex &sb->s_type->i_mutex_key#3 &c->lock irq_context: 0 &q->sysfs_dir_lock &q->sysfs_lock &q->rq_qos_mutex &q->debugfs_mutex &sb->s_type->i_mutex_key#3 &pcp->lock &zone->lock irq_context: 0 &q->sysfs_dir_lock &q->sysfs_lock &q->rq_qos_mutex &q->debugfs_mutex &sb->s_type->i_mutex_key#3 &zone->lock irq_context: 0 &q->sysfs_dir_lock &q->sysfs_lock &q->rq_qos_mutex &q->debugfs_mutex &sb->s_type->i_mutex_key#3 &____s->seqcount irq_context: 0 &q->sysfs_dir_lock &q->sysfs_lock &stats->lock irq_context: 0 &q->sysfs_dir_lock &q->sysfs_lock &q->debugfs_mutex &sb->s_type->i_mutex_key#3 &pcp->lock &zone->lock &____s->seqcount irq_context: 0 &q->sysfs_dir_lock &q->sysfs_lock &obj_hash[i].lock pool_lock irq_context: 0 &q->sysfs_dir_lock &q->sysfs_lock &c->lock irq_context: 0 &q->sysfs_dir_lock &q->sysfs_lock &pcp->lock &zone->lock irq_context: 0 &q->sysfs_dir_lock &q->sysfs_lock &zone->lock irq_context: 0 &q->sysfs_dir_lock &q->sysfs_lock &____s->seqcount irq_context: 0 &mm->mmap_lock &rq->__lock irq_context: 0 &q->sysfs_dir_lock &pcp->lock &zone->lock &____s->seqcount irq_context: 0 &q->sysfs_lock &obj_hash[i].lock pool_lock irq_context: softirq &(&ops->cursor_work)->timer irq_context: softirq &(&ops->cursor_work)->timer rcu_read_lock &pool->lock irq_context: softirq &(&ops->cursor_work)->timer rcu_read_lock &pool->lock &obj_hash[i].lock irq_context: softirq &(&ops->cursor_work)->timer rcu_read_lock &pool->lock pool_lock#2 irq_context: softirq &(&ops->cursor_work)->timer rcu_read_lock &pool->lock &p->pi_lock irq_context: softirq &(&ops->cursor_work)->timer rcu_read_lock &pool->lock &p->pi_lock &rq->__lock irq_context: softirq &(&ops->cursor_work)->timer rcu_read_lock &pool->lock &p->pi_lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 &q->sysfs_lock &rq->__lock irq_context: 0 (wq_completion)events_power_efficient (work_completion)(&(&ops->cursor_work)->work) irq_context: 0 (wq_completion)events_power_efficient (work_completion)(&(&ops->cursor_work)->work) (console_sem).lock irq_context: 0 (wq_completion)events_power_efficient (work_completion)(&(&ops->cursor_work)->work) console_lock &helper->damage_lock irq_context: 0 (wq_completion)events_power_efficient (work_completion)(&(&ops->cursor_work)->work) console_lock rcu_read_lock &pool->lock irq_context: 0 (wq_completion)events_power_efficient (work_completion)(&(&ops->cursor_work)->work) console_lock rcu_read_lock &pool->lock &obj_hash[i].lock irq_context: 0 (wq_completion)events_power_efficient (work_completion)(&(&ops->cursor_work)->work) console_lock rcu_read_lock &pool->lock pool_lock#2 irq_context: 0 (wq_completion)events_power_efficient (work_completion)(&(&ops->cursor_work)->work) &obj_hash[i].lock irq_context: 0 (wq_completion)events_power_efficient (work_completion)(&(&ops->cursor_work)->work) &base->lock irq_context: 0 (wq_completion)events_power_efficient (work_completion)(&(&ops->cursor_work)->work) &base->lock &obj_hash[i].lock irq_context: 0 &q->sysfs_dir_lock &q->sysfs_lock &q->rq_qos_mutex &q->debugfs_mutex &sb->s_type->i_mutex_key#3 &pcp->lock &zone->lock &____s->seqcount irq_context: 0 &q->queue_lock fill_pool_map-wait-type-override &c->lock irq_context: 0 &q->queue_lock fill_pool_map-wait-type-override &____s->seqcount irq_context: softirq rcu_callback quarantine_lock irq_context: 0 rcu_read_lock &pcp->lock &zone->lock irq_context: 0 rcu_read_lock &zone->lock irq_context: 0 rcu_read_lock &____s->seqcount irq_context: 0 rcu_read_lock rcu_read_lock pool_lock#2 irq_context: 0 nbd_index_mutex irq_context: 0 nbd_index_mutex fs_reclaim irq_context: 0 nbd_index_mutex fs_reclaim mmu_notifier_invalidate_range_start irq_context: 0 nbd_index_mutex pool_lock#2 irq_context: 0 set->srcu irq_context: 0 (work_completion)(&(&q->requeue_work)->work) irq_context: 0 (work_completion)(&(&hctx->run_work)->work) irq_context: 0 &q->debugfs_mutex irq_context: 0 &q->sysfs_dir_lock rcu_read_lock pool_lock#2 irq_context: 0 &q->sysfs_dir_lock &obj_hash[i].lock irq_context: 0 &q->sysfs_dir_lock &q->sysfs_lock lock irq_context: 0 &q->sysfs_dir_lock &q->sysfs_lock lock kernfs_idr_lock irq_context: 0 &q->sysfs_dir_lock &q->sysfs_lock &root->kernfs_rwsem irq_context: 0 &q->sysfs_dir_lock &q->sysfs_lock &root->kernfs_rwsem &root->kernfs_iattr_rwsem irq_context: 0 &q->sysfs_dir_lock &q->sysfs_lock &q->rq_qos_mutex set->srcu irq_context: 0 &sig->cred_guard_mutex pool_lock irq_context: 0 lock kernfs_idr_lock rcu_read_lock pool_lock#2 irq_context: 0 lock kernfs_idr_lock &obj_hash[i].lock irq_context: 0 &q->sysfs_lock &pcp->lock &zone->lock &____s->seqcount irq_context: 0 &q->sysfs_dir_lock &q->sysfs_lock lock kernfs_idr_lock pool_lock#2 irq_context: 0 &x->wait &p->pi_lock &rq->__lock &cfs_rq->removed.lock irq_context: 0 cpu_hotplug_lock wq_pool_mutex fill_pool_map-wait-type-override pool_lock#2 irq_context: 0 cpu_hotplug_lock wq_pool_mutex fill_pool_map-wait-type-override pool_lock irq_context: 0 &q->sysfs_dir_lock lock kernfs_idr_lock &c->lock irq_context: 0 &q->sysfs_dir_lock lock kernfs_idr_lock &pcp->lock &zone->lock irq_context: 0 &q->sysfs_dir_lock lock kernfs_idr_lock &zone->lock irq_context: 0 &q->sysfs_dir_lock lock kernfs_idr_lock &____s->seqcount irq_context: 0 zram_index_mutex irq_context: 0 zram_index_mutex fs_reclaim irq_context: 0 zram_index_mutex fs_reclaim mmu_notifier_invalidate_range_start irq_context: 0 zram_index_mutex pool_lock#2 irq_context: 0 zram_index_mutex blk_queue_ida.xa_lock irq_context: 0 zram_index_mutex &obj_hash[i].lock irq_context: 0 zram_index_mutex pcpu_alloc_mutex irq_context: 0 zram_index_mutex pcpu_alloc_mutex pcpu_lock irq_context: 0 zram_index_mutex bio_slab_lock irq_context: 0 zram_index_mutex &c->lock irq_context: 0 zram_index_mutex &____s->seqcount irq_context: 0 zram_index_mutex percpu_counters_lock irq_context: 0 zram_index_mutex &obj_hash[i].lock pool_lock irq_context: 0 zram_index_mutex fill_pool_map-wait-type-override pool_lock#2 irq_context: 0 zram_index_mutex fill_pool_map-wait-type-override &c->lock irq_context: 0 zram_index_mutex fill_pool_map-wait-type-override &____s->seqcount irq_context: 0 zram_index_mutex fill_pool_map-wait-type-override pool_lock irq_context: 0 zram_index_mutex mmu_notifier_invalidate_range_start irq_context: 0 zram_index_mutex &sb->s_type->i_lock_key#3 irq_context: 0 zram_index_mutex &s->s_inode_list_lock irq_context: 0 zram_index_mutex &xa->xa_lock#11 irq_context: 0 zram_index_mutex lock irq_context: 0 zram_index_mutex lock &q->queue_lock irq_context: 0 zram_index_mutex lock &q->queue_lock &blkcg->lock irq_context: 0 zram_index_mutex &q->queue_lock irq_context: 0 zram_index_mutex &q->queue_lock pool_lock#2 irq_context: 0 zram_index_mutex &q->queue_lock &c->lock irq_context: 0 zram_index_mutex &q->queue_lock &____s->seqcount irq_context: 0 zram_index_mutex &q->queue_lock pcpu_lock irq_context: 0 zram_index_mutex &q->queue_lock &obj_hash[i].lock irq_context: 0 zram_index_mutex &q->queue_lock percpu_counters_lock irq_context: 0 zram_index_mutex &q->queue_lock &obj_hash[i].lock pool_lock irq_context: 0 zram_index_mutex &q->queue_lock fill_pool_map-wait-type-override pool_lock#2 irq_context: 0 zram_index_mutex &q->queue_lock fill_pool_map-wait-type-override pool_lock irq_context: 0 zram_index_mutex &x->wait#9 irq_context: 0 zram_index_mutex &bdev->bd_size_lock irq_context: 0 zram_index_mutex &n->list_lock irq_context: 0 zram_index_mutex &n->list_lock &c->lock irq_context: 0 zram_index_mutex &k->list_lock irq_context: 0 zram_index_mutex gdp_mutex irq_context: 0 zram_index_mutex gdp_mutex &k->list_lock irq_context: 0 zram_index_mutex &pcp->lock &zone->lock irq_context: 0 zram_index_mutex &zone->lock irq_context: 0 zram_index_mutex rcu_read_lock pool_lock#2 irq_context: 0 zram_index_mutex lock kernfs_idr_lock irq_context: 0 zram_index_mutex &root->kernfs_rwsem irq_context: 0 zram_index_mutex &root->kernfs_rwsem &root->kernfs_iattr_rwsem irq_context: 0 zram_index_mutex bus_type_sem irq_context: 0 zram_index_mutex sysfs_symlink_target_lock irq_context: 0 zram_index_mutex lock kernfs_idr_lock pool_lock#2 irq_context: 0 zram_index_mutex &root->kernfs_rwsem irq_context: 0 zram_index_mutex &dev->power.lock irq_context: 0 zram_index_mutex dpm_list_mtx irq_context: 0 zram_index_mutex req_lock irq_context: 0 zram_index_mutex &p->pi_lock irq_context: 0 zram_index_mutex &x->wait#11 irq_context: 0 zram_index_mutex &rq->__lock irq_context: 0 zram_index_mutex &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 zram_index_mutex subsys mutex#37 irq_context: 0 zram_index_mutex subsys mutex#37 &k->k_lock irq_context: 0 zram_index_mutex dev_hotplug_mutex irq_context: 0 zram_index_mutex dev_hotplug_mutex &dev->power.lock irq_context: 0 zram_index_mutex &q->sysfs_dir_lock irq_context: 0 zram_index_mutex &q->sysfs_dir_lock fs_reclaim irq_context: 0 zram_index_mutex &q->sysfs_dir_lock fs_reclaim mmu_notifier_invalidate_range_start irq_context: 0 zram_index_mutex &q->sysfs_dir_lock &c->lock irq_context: 0 zram_index_mutex &q->sysfs_dir_lock &pcp->lock &zone->lock irq_context: 0 zram_index_mutex &q->sysfs_dir_lock &zone->lock irq_context: 0 zram_index_mutex &q->sysfs_dir_lock &____s->seqcount irq_context: 0 zram_index_mutex &q->sysfs_dir_lock pool_lock#2 irq_context: 0 zram_index_mutex &q->sysfs_dir_lock lock irq_context: 0 zram_index_mutex &q->sysfs_dir_lock lock kernfs_idr_lock irq_context: 0 zram_index_mutex &q->sysfs_dir_lock &root->kernfs_rwsem irq_context: 0 zram_index_mutex &q->sysfs_dir_lock &root->kernfs_rwsem &root->kernfs_iattr_rwsem irq_context: 0 zram_index_mutex &q->sysfs_dir_lock &q->sysfs_lock irq_context: 0 zram_index_mutex &q->sysfs_dir_lock &q->sysfs_lock &q->debugfs_mutex irq_context: 0 zram_index_mutex &q->sysfs_dir_lock &q->sysfs_lock &q->debugfs_mutex pin_fs_lock irq_context: 0 zram_index_mutex &q->sysfs_dir_lock &q->sysfs_lock &q->debugfs_mutex &sb->s_type->i_mutex_key#3 irq_context: 0 zram_index_mutex &q->sysfs_dir_lock &q->sysfs_lock &q->debugfs_mutex &sb->s_type->i_mutex_key#3 rename_lock.seqcount irq_context: 0 zram_index_mutex &q->sysfs_dir_lock &q->sysfs_lock &q->debugfs_mutex &sb->s_type->i_mutex_key#3 fs_reclaim irq_context: 0 zram_index_mutex &q->sysfs_dir_lock &q->sysfs_lock &q->debugfs_mutex &sb->s_type->i_mutex_key#3 fs_reclaim mmu_notifier_invalidate_range_start irq_context: 0 zram_index_mutex &q->sysfs_dir_lock &q->sysfs_lock &q->debugfs_mutex &sb->s_type->i_mutex_key#3 pool_lock#2 irq_context: 0 zram_index_mutex &q->sysfs_dir_lock &q->sysfs_lock &q->debugfs_mutex &sb->s_type->i_mutex_key#3 &dentry->d_lock irq_context: 0 zram_index_mutex &q->sysfs_dir_lock &q->sysfs_lock &q->debugfs_mutex &sb->s_type->i_mutex_key#3 rcu_read_lock rename_lock.seqcount irq_context: 0 zram_index_mutex &q->sysfs_dir_lock &q->sysfs_lock &q->debugfs_mutex &sb->s_type->i_mutex_key#3 &dentry->d_lock &wq irq_context: 0 zram_index_mutex &q->sysfs_dir_lock &q->sysfs_lock &q->debugfs_mutex &sb->s_type->i_mutex_key#3 mmu_notifier_invalidate_range_start irq_context: 0 zram_index_mutex &q->sysfs_dir_lock &q->sysfs_lock &q->debugfs_mutex &sb->s_type->i_mutex_key#3 &sb->s_type->i_lock_key#7 irq_context: 0 zram_index_mutex &q->sysfs_dir_lock &q->sysfs_lock &q->debugfs_mutex &sb->s_type->i_mutex_key#3 &s->s_inode_list_lock irq_context: 0 zram_index_mutex &q->sysfs_dir_lock &q->sysfs_lock &q->debugfs_mutex &sb->s_type->i_mutex_key#3 tk_core.seq.seqcount irq_context: 0 zram_index_mutex &q->sysfs_dir_lock &q->sysfs_lock &q->debugfs_mutex &sb->s_type->i_mutex_key#3 &sb->s_type->i_lock_key#7 &dentry->d_lock irq_context: 0 zram_index_mutex percpu_ref_switch_lock irq_context: 0 zram_index_mutex lock kernfs_idr_lock &c->lock irq_context: 0 zram_index_mutex lock kernfs_idr_lock &pcp->lock &zone->lock irq_context: 0 zram_index_mutex lock kernfs_idr_lock &zone->lock irq_context: 0 zram_index_mutex lock kernfs_idr_lock &____s->seqcount irq_context: 0 zram_index_mutex uevent_sock_mutex irq_context: 0 zram_index_mutex rcu_read_lock &pool->lock/1 irq_context: 0 zram_index_mutex rcu_read_lock &pool->lock/1 &obj_hash[i].lock irq_context: 0 zram_index_mutex rcu_read_lock &pool->lock/1 pool_lock#2 irq_context: 0 zram_index_mutex rcu_read_lock &pool->lock/1 &p->pi_lock irq_context: 0 zram_index_mutex rcu_read_lock &pool->lock/1 &p->pi_lock &rq->__lock irq_context: 0 zram_index_mutex rcu_read_lock &pool->lock/1 &p->pi_lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 zram_index_mutex running_helpers_waitq.lock irq_context: 0 zram_index_mutex subsys mutex#38 irq_context: 0 zram_index_mutex subsys mutex#38 &k->k_lock irq_context: 0 zram_index_mutex cgwb_lock irq_context: 0 zram_index_mutex pin_fs_lock irq_context: 0 zram_index_mutex &sb->s_type->i_mutex_key#3 irq_context: 0 zram_index_mutex &sb->s_type->i_mutex_key#3 rename_lock.seqcount irq_context: 0 zram_index_mutex &sb->s_type->i_mutex_key#3 fs_reclaim irq_context: 0 zram_index_mutex &sb->s_type->i_mutex_key#3 fs_reclaim mmu_notifier_invalidate_range_start irq_context: 0 zram_index_mutex &sb->s_type->i_mutex_key#3 pool_lock#2 irq_context: 0 zram_index_mutex &sb->s_type->i_mutex_key#3 &dentry->d_lock irq_context: 0 zram_index_mutex &sb->s_type->i_mutex_key#3 rcu_read_lock rename_lock.seqcount irq_context: 0 zram_index_mutex &sb->s_type->i_mutex_key#3 &dentry->d_lock &wq irq_context: 0 zram_index_mutex &sb->s_type->i_mutex_key#3 mmu_notifier_invalidate_range_start irq_context: 0 zram_index_mutex &sb->s_type->i_mutex_key#3 &sb->s_type->i_lock_key#7 irq_context: 0 zram_index_mutex &sb->s_type->i_mutex_key#3 &s->s_inode_list_lock irq_context: 0 zram_index_mutex &sb->s_type->i_mutex_key#3 tk_core.seq.seqcount irq_context: 0 zram_index_mutex &sb->s_type->i_mutex_key#3 &sb->s_type->i_lock_key#7 &dentry->d_lock irq_context: 0 zram_index_mutex bdi_lock irq_context: 0 zram_index_mutex inode_hash_lock irq_context: 0 zram_index_mutex inode_hash_lock &sb->s_type->i_lock_key#3 irq_context: 0 zram_index_mutex (console_sem).lock irq_context: 0 zram_index_mutex console_lock console_srcu console_owner_lock irq_context: 0 zram_index_mutex console_lock console_srcu console_owner irq_context: 0 zram_index_mutex console_lock console_srcu console_owner &port_lock_key irq_context: 0 zram_index_mutex console_lock console_srcu console_owner console_owner_lock irq_context: 0 subsys mutex#39 irq_context: 0 subsys mutex#39 &k->k_lock irq_context: 0 &sb->s_type->i_mutex_key#7/1 &default_group_class[depth - 1]#2 irq_context: 0 &sb->s_type->i_mutex_key#7/1 &default_group_class[depth - 1]#2 fs_reclaim irq_context: 0 &sb->s_type->i_mutex_key#7/1 &default_group_class[depth - 1]#2 fs_reclaim mmu_notifier_invalidate_range_start irq_context: 0 &sb->s_type->i_mutex_key#7/1 &default_group_class[depth - 1]#2 pool_lock#2 irq_context: 0 &sb->s_type->i_mutex_key#7/1 &default_group_class[depth - 1]#2 configfs_dirent_lock irq_context: 0 &q->sysfs_lock &xa->xa_lock#10 pool_lock#2 irq_context: 0 &lock irq_context: 0 &lock nullb_indexes.xa_lock irq_context: 0 &q->sysfs_dir_lock &q->sysfs_lock &q->rq_qos_mutex &q->debugfs_mutex &sb->s_type->i_mutex_key#3 rcu_read_lock pool_lock#2 irq_context: 0 &q->sysfs_dir_lock &q->sysfs_lock &q->rq_qos_mutex &q->debugfs_mutex &sb->s_type->i_mutex_key#3 &obj_hash[i].lock irq_context: 0 &disk->open_mutex rcu_read_lock tk_core.seq.seqcount irq_context: 0 &disk->open_mutex rcu_read_lock &obj_hash[i].lock irq_context: 0 &disk->open_mutex rcu_read_lock &base->lock irq_context: 0 &disk->open_mutex rcu_read_lock &base->lock &obj_hash[i].lock irq_context: 0 &disk->open_mutex rcu_read_lock &ret->b_uptodate_lock irq_context: 0 ctx_list.lock irq_context: 0 nfc_index_ida.xa_lock irq_context: 0 nfc_devlist_mutex irq_context: 0 nfc_devlist_mutex fs_reclaim irq_context: 0 nfc_devlist_mutex fs_reclaim mmu_notifier_invalidate_range_start irq_context: 0 nfc_devlist_mutex &c->lock irq_context: 0 nfc_devlist_mutex &____s->seqcount irq_context: 0 nfc_devlist_mutex pool_lock#2 irq_context: 0 nfc_devlist_mutex &k->list_lock irq_context: 0 nfc_devlist_mutex gdp_mutex irq_context: 0 nfc_devlist_mutex gdp_mutex &k->list_lock irq_context: 0 nfc_devlist_mutex gdp_mutex fs_reclaim irq_context: 0 nfc_devlist_mutex gdp_mutex fs_reclaim mmu_notifier_invalidate_range_start irq_context: 0 nfc_devlist_mutex gdp_mutex pool_lock#2 irq_context: 0 nfc_devlist_mutex gdp_mutex lock irq_context: 0 nfc_devlist_mutex gdp_mutex lock kernfs_idr_lock irq_context: 0 nfc_devlist_mutex gdp_mutex &root->kernfs_rwsem irq_context: 0 nfc_devlist_mutex gdp_mutex &root->kernfs_rwsem &root->kernfs_iattr_rwsem irq_context: 0 nfc_devlist_mutex lock irq_context: 0 nfc_devlist_mutex lock kernfs_idr_lock irq_context: 0 nfc_devlist_mutex &root->kernfs_rwsem irq_context: 0 nfc_devlist_mutex &root->kernfs_rwsem &root->kernfs_iattr_rwsem irq_context: 0 nfc_devlist_mutex bus_type_sem irq_context: 0 nfc_devlist_mutex sysfs_symlink_target_lock irq_context: 0 nfc_devlist_mutex &root->kernfs_rwsem irq_context: 0 nfc_devlist_mutex &dev->power.lock irq_context: 0 nfc_devlist_mutex dpm_list_mtx irq_context: 0 nfc_devlist_mutex uevent_sock_mutex irq_context: 0 nfc_devlist_mutex &obj_hash[i].lock irq_context: 0 nfc_devlist_mutex rcu_read_lock &pool->lock/1 irq_context: 0 nfc_devlist_mutex rcu_read_lock &pool->lock/1 &obj_hash[i].lock irq_context: 0 nfc_devlist_mutex rcu_read_lock &pool->lock/1 pool_lock#2 irq_context: 0 nfc_devlist_mutex rcu_read_lock &pool->lock/1 &p->pi_lock irq_context: 0 nfc_devlist_mutex rcu_read_lock &pool->lock/1 &p->pi_lock &rq->__lock irq_context: 0 nfc_devlist_mutex rcu_read_lock &pool->lock/1 &p->pi_lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 nfc_devlist_mutex running_helpers_waitq.lock irq_context: 0 nfc_devlist_mutex subsys mutex#40 irq_context: 0 nfc_devlist_mutex subsys mutex#40 &k->k_lock irq_context: 0 &dev->mutex rfkill_global_mutex irq_context: 0 &dev->mutex rfkill_global_mutex fs_reclaim irq_context: 0 &dev->mutex rfkill_global_mutex fs_reclaim mmu_notifier_invalidate_range_start irq_context: 0 &dev->mutex rfkill_global_mutex pool_lock#2 irq_context: 0 &dev->mutex rfkill_global_mutex &k->list_lock irq_context: 0 &dev->mutex rfkill_global_mutex lock irq_context: 0 &dev->mutex rfkill_global_mutex lock kernfs_idr_lock irq_context: 0 &dev->mutex rfkill_global_mutex &root->kernfs_rwsem irq_context: 0 &dev->mutex rfkill_global_mutex &root->kernfs_rwsem &root->kernfs_iattr_rwsem irq_context: 0 &dev->mutex rfkill_global_mutex bus_type_sem irq_context: 0 &dev->mutex rfkill_global_mutex sysfs_symlink_target_lock irq_context: 0 &dev->mutex rfkill_global_mutex &c->lock irq_context: 0 &dev->mutex rfkill_global_mutex &____s->seqcount irq_context: 0 &dev->mutex rfkill_global_mutex &root->kernfs_rwsem irq_context: 0 &dev->mutex rfkill_global_mutex &dev->power.lock irq_context: 0 &dev->mutex rfkill_global_mutex dpm_list_mtx irq_context: 0 &dev->mutex rfkill_global_mutex &rfkill->lock irq_context: 0 &dev->mutex rfkill_global_mutex uevent_sock_mutex irq_context: 0 &dev->mutex rfkill_global_mutex &n->list_lock irq_context: 0 &dev->mutex rfkill_global_mutex &n->list_lock &c->lock irq_context: 0 &dev->mutex rfkill_global_mutex &obj_hash[i].lock irq_context: 0 &dev->mutex rfkill_global_mutex rcu_read_lock &pool->lock/1 irq_context: 0 &dev->mutex rfkill_global_mutex rcu_read_lock &pool->lock/1 &obj_hash[i].lock irq_context: 0 &dev->mutex rfkill_global_mutex rcu_read_lock &pool->lock/1 pool_lock#2 irq_context: 0 &dev->mutex rfkill_global_mutex rcu_read_lock &pool->lock/1 &p->pi_lock irq_context: 0 &dev->mutex rfkill_global_mutex running_helpers_waitq.lock irq_context: 0 &dev->mutex rfkill_global_mutex &k->k_lock irq_context: 0 &dev->mutex rfkill_global_mutex subsys mutex#41 irq_context: 0 &dev->mutex rfkill_global_mutex subsys mutex#41 &k->k_lock irq_context: 0 &dev->mutex rfkill_global_mutex triggers_list_lock irq_context: 0 &dev->mutex rfkill_global_mutex leds_list_lock irq_context: 0 &dev->mutex rfkill_global_mutex rcu_read_lock &pool->lock irq_context: 0 &dev->mutex rfkill_global_mutex rcu_read_lock &pool->lock &obj_hash[i].lock irq_context: 0 &dev->mutex rfkill_global_mutex rcu_read_lock &pool->lock pool_lock#2 irq_context: 0 &dev->mutex rfkill_global_mutex rcu_read_lock &pool->lock &p->pi_lock irq_context: 0 &dev->mutex rfkill_global_mutex rcu_read_lock &pool->lock &p->pi_lock &rq->__lock irq_context: 0 &dev->mutex rfkill_global_mutex rcu_read_lock &pool->lock &p->pi_lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 &dev->mutex rfkill_global_mutex &rq->__lock irq_context: 0 (wq_completion)events (work_completion)(&rfkill->sync_work) irq_context: 0 (wq_completion)events (work_completion)(&rfkill->sync_work) rfkill_global_mutex irq_context: 0 (wq_completion)events (work_completion)(&rfkill->sync_work) rfkill_global_mutex rfkill_global_mutex.wait_lock irq_context: 0 (wq_completion)events (work_completion)(&rfkill->sync_work) rfkill_global_mutex &pool->lock irq_context: 0 (wq_completion)events (work_completion)(&rfkill->sync_work) rfkill_global_mutex &pool->lock &p->pi_lock irq_context: 0 (wq_completion)events (work_completion)(&rfkill->sync_work) rfkill_global_mutex &pool->lock &p->pi_lock &rq->__lock irq_context: 0 (wq_completion)events (work_completion)(&rfkill->sync_work) rfkill_global_mutex &pool->lock &p->pi_lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)events (work_completion)(&rfkill->sync_work) rfkill_global_mutex &rq->__lock irq_context: 0 (wq_completion)events (work_completion)(&rfkill->sync_work) rfkill_global_mutex &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)events (work_completion)(&rfkill_global_led_trigger_work) rfkill_global_mutex rfkill_global_mutex.wait_lock irq_context: 0 (wq_completion)events (work_completion)(&rfkill_global_led_trigger_work) rfkill_global_mutex &pool->lock irq_context: 0 (wq_completion)events (work_completion)(&rfkill_global_led_trigger_work) rfkill_global_mutex &rq->__lock irq_context: 0 (wq_completion)events (work_completion)(&rfkill_global_led_trigger_work) rfkill_global_mutex &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 &dev->mutex rfkill_global_mutex.wait_lock irq_context: 0 &dev->mutex &p->pi_lock irq_context: 0 &dev->mutex &p->pi_lock &rq->__lock irq_context: 0 &dev->mutex &p->pi_lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)events (work_completion)(&rfkill->sync_work) rfkill_global_mutex &rfkill->lock irq_context: 0 (wq_completion)events (work_completion)(&rfkill->sync_work) rfkill_global_mutex rcu_read_lock &pool->lock irq_context: 0 (wq_completion)events (work_completion)(&rfkill->sync_work) rfkill_global_mutex rcu_read_lock &pool->lock &obj_hash[i].lock irq_context: 0 (wq_completion)events (work_completion)(&rfkill->sync_work) rfkill_global_mutex.wait_lock irq_context: 0 (wq_completion)events (work_completion)(&rfkill->sync_work) &p->pi_lock irq_context: 0 (wq_completion)events (work_completion)(&rfkill->sync_work) &p->pi_lock &rq->__lock irq_context: 0 (wq_completion)events (work_completion)(&rfkill->sync_work) &p->pi_lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)events (work_completion)(&rfkill->sync_work) &rq->__lock irq_context: 0 nfc_devlist_mutex &pcp->lock &zone->lock irq_context: 0 nfc_devlist_mutex &zone->lock irq_context: 0 nfc_devlist_mutex rcu_read_lock pool_lock#2 irq_context: 0 &dev->mutex rfkill_global_mutex &pcp->lock &zone->lock irq_context: 0 &dev->mutex rfkill_global_mutex &zone->lock irq_context: 0 &dev->mutex rfkill_global_mutex lock kernfs_idr_lock pool_lock#2 irq_context: 0 &dev->mutex rfkill_global_mutex &obj_hash[i].lock pool_lock irq_context: 0 &dev->mutex rfkill_global_mutex fill_pool_map-wait-type-override pool_lock#2 irq_context: 0 &dev->mutex rfkill_global_mutex fill_pool_map-wait-type-override pool_lock irq_context: 0 misc_mtx batched_entropy_u8.lock irq_context: 0 misc_mtx kfence_freelist_lock irq_context: 0 misc_mtx rcu_read_lock &rq->__lock irq_context: 0 misc_mtx rcu_read_lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 &mm->mmap_lock rcu_read_lock &cfs_rq->removed.lock irq_context: 0 &mm->mmap_lock rcu_read_lock &obj_hash[i].lock irq_context: 0 dma_heap_minors.xa_lock irq_context: 0 (wq_completion)events_unbound (work_completion)(&sub_info->work) rcu_node_0 irq_context: 0 (wq_completion)events_unbound (work_completion)(&sub_info->work) &n->list_lock irq_context: 0 subsys mutex#42 irq_context: 0 subsys mutex#42 &k->k_lock irq_context: 0 heap_list_lock irq_context: 0 &dev->mutex host_index_ida.xa_lock irq_context: 0 &dev->mutex kthread_create_lock irq_context: 0 &dev->mutex &p->pi_lock &rq->__lock &cfs_rq->removed.lock irq_context: 0 &sig->cred_guard_mutex &mm->mmap_lock &vma->vm_lock->lock &rq->__lock irq_context: 0 &dev->mutex &x->wait irq_context: softirq rcu_callback put_task_map-wait-type-override fill_pool_map-wait-type-override pool_lock#2 irq_context: softirq rcu_callback put_task_map-wait-type-override fill_pool_map-wait-type-override &c->lock irq_context: softirq rcu_callback put_task_map-wait-type-override fill_pool_map-wait-type-override &pcp->lock &zone->lock irq_context: softirq rcu_callback put_task_map-wait-type-override fill_pool_map-wait-type-override &zone->lock irq_context: softirq rcu_callback put_task_map-wait-type-override fill_pool_map-wait-type-override &____s->seqcount irq_context: softirq rcu_callback put_task_map-wait-type-override fill_pool_map-wait-type-override pool_lock irq_context: 0 &sig->cred_guard_mutex &mm->mmap_lock &vma->vm_lock->lock &cfs_rq->removed.lock irq_context: 0 &sig->cred_guard_mutex &mm->mmap_lock &vma->vm_lock->lock &obj_hash[i].lock irq_context: 0 &sig->cred_guard_mutex &mm->mmap_lock &pcp->lock &zone->lock irq_context: 0 &sig->cred_guard_mutex &mm->mmap_lock &pcp->lock &zone->lock &____s->seqcount irq_context: 0 &dev->mutex cpu_hotplug_lock wq_pool_mutex irq_context: 0 &dev->mutex cpu_hotplug_lock wq_pool_mutex fs_reclaim irq_context: 0 &dev->mutex cpu_hotplug_lock wq_pool_mutex fs_reclaim mmu_notifier_invalidate_range_start irq_context: 0 &dev->mutex cpu_hotplug_lock wq_pool_mutex pool_lock#2 irq_context: 0 &dev->mutex cpu_hotplug_lock wq_pool_mutex &c->lock irq_context: 0 &dev->mutex cpu_hotplug_lock wq_pool_mutex &pcp->lock &zone->lock irq_context: 0 &dev->mutex cpu_hotplug_lock wq_pool_mutex &zone->lock irq_context: 0 &dev->mutex cpu_hotplug_lock wq_pool_mutex &____s->seqcount irq_context: 0 &dev->mutex cpu_hotplug_lock wq_pool_mutex &obj_hash[i].lock irq_context: 0 &dev->mutex cpu_hotplug_lock wq_pool_mutex &wq->mutex irq_context: 0 &dev->mutex cpu_hotplug_lock wq_pool_mutex &wq->mutex &pool->lock/1 irq_context: 0 &dev->mutex rcu_read_lock &rq->__lock irq_context: 0 &dev->mutex rcu_read_lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 &dev->mutex subsys mutex#9 irq_context: 0 &dev->mutex wq_pool_mutex irq_context: 0 &dev->mutex wq_pool_mutex &wq->mutex irq_context: 0 &dev->mutex &md->mutex &pcp->lock &zone->lock irq_context: 0 &dev->mutex &md->mutex &zone->lock irq_context: 0 &dev->mutex &md->mutex &domain->mutex sparse_irq_lock &n->list_lock irq_context: 0 &dev->mutex &md->mutex &domain->mutex sparse_irq_lock &pcp->lock &zone->lock irq_context: 0 &dev->mutex &md->mutex &domain->mutex sparse_irq_lock &zone->lock irq_context: 0 &dev->mutex &md->mutex &domain->mutex sparse_irq_lock &pcp->lock &zone->lock &____s->seqcount irq_context: 0 &dev->mutex &md->mutex &domain->mutex sparse_irq_lock &obj_hash[i].lock pool_lock irq_context: 0 &dev->mutex &md->mutex &rq->__lock irq_context: 0 &dev->mutex &md->mutex &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 &dev->mutex register_lock rcu_read_lock pool_lock#2 irq_context: 0 &dev->mutex register_lock &obj_hash[i].lock irq_context: 0 &dev->mutex &desc->request_mutex &irq_desc_lock_class tmp_mask_lock irq_context: 0 &dev->mutex &desc->request_mutex &irq_desc_lock_class tmp_mask_lock vector_lock irq_context: 0 &dev->mutex scsi_sense_cache_mutex irq_context: 0 &dev->mutex scsi_sense_cache_mutex slab_mutex irq_context: 0 &dev->mutex scsi_sense_cache_mutex slab_mutex fs_reclaim irq_context: 0 &dev->mutex scsi_sense_cache_mutex slab_mutex fs_reclaim mmu_notifier_invalidate_range_start irq_context: 0 &dev->mutex scsi_sense_cache_mutex slab_mutex pool_lock#2 irq_context: 0 &dev->mutex scsi_sense_cache_mutex slab_mutex &c->lock irq_context: 0 &dev->mutex scsi_sense_cache_mutex slab_mutex &n->list_lock irq_context: 0 &dev->mutex scsi_sense_cache_mutex slab_mutex pcpu_alloc_mutex irq_context: 0 &dev->mutex scsi_sense_cache_mutex slab_mutex pcpu_alloc_mutex pcpu_lock irq_context: 0 &dev->mutex &n->list_lock irq_context: 0 &dev->mutex pcpu_alloc_mutex irq_context: 0 &dev->mutex pcpu_alloc_mutex pcpu_lock irq_context: 0 &dev->mutex batched_entropy_u32.lock irq_context: 0 &dev->mutex &dev->power.lock &dev->power.lock/1 irq_context: 0 &dev->mutex subsys mutex#43 irq_context: 0 &dev->mutex subsys mutex#44 irq_context: 0 &dev->mutex subsys mutex#44 &k->k_lock irq_context: 0 &dev->mutex attribute_container_mutex irq_context: 0 &dev->mutex &dev->power.lock &dev->power.wait_queue irq_context: 0 &dev->mutex &virtscsi_vq->vq_lock irq_context: 0 &dev->mutex &shost->scan_mutex irq_context: 0 &dev->mutex &shost->scan_mutex fs_reclaim irq_context: 0 &dev->mutex &shost->scan_mutex fs_reclaim mmu_notifier_invalidate_range_start irq_context: 0 &dev->mutex &shost->scan_mutex pool_lock#2 irq_context: 0 &dev->mutex &shost->scan_mutex shost->host_lock irq_context: 0 &dev->mutex async_scan_lock irq_context: 0 &dev->mutex async_scan_lock &x->wait#15 irq_context: 0 &dev->mutex async_lock irq_context: 0 major_names_lock &c->lock irq_context: 0 major_names_lock &____s->seqcount irq_context: 0 subsys mutex#45 irq_context: 0 subsys mutex#45 &k->list_lock irq_context: 0 subsys mutex#45 &k->k_lock irq_context: 0 (wq_completion)events_unbound (work_completion)(&entry->work) &shost->scan_mutex irq_context: 0 (wq_completion)events_unbound (work_completion)(&entry->work) &shost->scan_mutex &dev->power.lock irq_context: 0 (wq_completion)events_unbound (work_completion)(&entry->work) &shost->scan_mutex fs_reclaim irq_context: 0 (wq_completion)events_unbound (work_completion)(&entry->work) &shost->scan_mutex fs_reclaim mmu_notifier_invalidate_range_start irq_context: 0 (wq_completion)events_unbound (work_completion)(&entry->work) &shost->scan_mutex pool_lock#2 irq_context: 0 (wq_completion)events_unbound (work_completion)(&entry->work) &shost->scan_mutex &x->wait#9 irq_context: 0 (wq_completion)events_unbound (work_completion)(&entry->work) &shost->scan_mutex &obj_hash[i].lock irq_context: 0 (wq_completion)events_unbound (work_completion)(&entry->work) &shost->scan_mutex shost->host_lock irq_context: 0 (wq_completion)events_unbound (work_completion)(&entry->work) &shost->scan_mutex attribute_container_mutex irq_context: 0 (wq_completion)events_unbound (work_completion)(&entry->work) &shost->scan_mutex blk_queue_ida.xa_lock irq_context: 0 (wq_completion)events_unbound (work_completion)(&entry->work) &shost->scan_mutex pcpu_alloc_mutex irq_context: 0 (wq_completion)events_unbound (work_completion)(&entry->work) &shost->scan_mutex pcpu_alloc_mutex pcpu_lock irq_context: 0 (wq_completion)events_unbound (work_completion)(&entry->work) &shost->scan_mutex &q->sysfs_lock irq_context: 0 (wq_completion)events_unbound (work_completion)(&entry->work) &shost->scan_mutex &q->sysfs_lock &q->unused_hctx_lock irq_context: 0 (wq_completion)events_unbound (work_completion)(&entry->work) &shost->scan_mutex &q->sysfs_lock mmu_notifier_invalidate_range_start irq_context: 0 (wq_completion)events_unbound (work_completion)(&entry->work) &shost->scan_mutex &q->sysfs_lock &c->lock irq_context: 0 (wq_completion)events_unbound (work_completion)(&entry->work) &shost->scan_mutex &q->sysfs_lock &n->list_lock irq_context: 0 (wq_completion)events_unbound (work_completion)(&entry->work) &shost->scan_mutex &q->sysfs_lock &pcp->lock &zone->lock irq_context: 0 (wq_completion)events_unbound (work_completion)(&entry->work) &shost->scan_mutex &q->sysfs_lock &zone->lock irq_context: 0 (wq_completion)events_unbound (work_completion)(&entry->work) &shost->scan_mutex &q->sysfs_lock &____s->seqcount irq_context: 0 (wq_completion)events_unbound (work_completion)(&entry->work) &shost->scan_mutex &q->sysfs_lock pool_lock#2 irq_context: 0 (wq_completion)events_unbound (work_completion)(&entry->work) &shost->scan_mutex &q->sysfs_lock &obj_hash[i].lock irq_context: 0 (wq_completion)events_unbound (work_completion)(&entry->work) &shost->scan_mutex &q->sysfs_lock cpu_hotplug_lock irq_context: 0 (wq_completion)events_unbound (work_completion)(&entry->work) &shost->scan_mutex &q->sysfs_lock cpu_hotplug_lock cpuhp_state_mutex irq_context: 0 (wq_completion)events_unbound (work_completion)(&entry->work) &shost->scan_mutex &q->sysfs_lock fs_reclaim irq_context: 0 (wq_completion)events_unbound (work_completion)(&entry->work) &shost->scan_mutex &q->sysfs_lock fs_reclaim mmu_notifier_invalidate_range_start irq_context: 0 (wq_completion)events_unbound (work_completion)(&entry->work) &shost->scan_mutex &q->sysfs_lock &xa->xa_lock#10 irq_context: 0 (wq_completion)events_unbound (work_completion)(&entry->work) &shost->scan_mutex &set->tag_list_lock irq_context: 0 (wq_completion)events_unbound (work_completion)(&entry->work) &shost->scan_mutex batched_entropy_u32.lock irq_context: 0 (wq_completion)events_unbound (work_completion)(&entry->work) &shost->scan_mutex &c->lock irq_context: 0 (wq_completion)events_unbound (work_completion)(&entry->work) &shost->scan_mutex &pcp->lock &zone->lock irq_context: 0 (wq_completion)events_unbound (work_completion)(&entry->work) &shost->scan_mutex &zone->lock irq_context: 0 (wq_completion)events_unbound (work_completion)(&entry->work) &shost->scan_mutex &____s->seqcount irq_context: 0 (wq_completion)events_unbound (work_completion)(&entry->work) &shost->scan_mutex rcu_read_lock pool_lock#2 irq_context: 0 (wq_completion)events_unbound (work_completion)(&entry->work) &shost->scan_mutex tk_core.seq.seqcount irq_context: 0 (wq_completion)events_unbound (work_completion)(&entry->work) &shost->scan_mutex mmu_notifier_invalidate_range_start irq_context: 0 (wq_completion)events_unbound (work_completion)(&entry->work) &shost->scan_mutex &hctx->lock irq_context: 0 (wq_completion)events_unbound (work_completion)(&entry->work) &shost->scan_mutex rcu_read_lock &hctx->lock irq_context: 0 (wq_completion)events_unbound (work_completion)(&entry->work) &shost->scan_mutex rcu_read_lock batched_entropy_u32.lock irq_context: 0 (wq_completion)events_unbound (work_completion)(&entry->work) &shost->scan_mutex rcu_read_lock &obj_hash[i].lock irq_context: 0 (wq_completion)events_unbound (work_completion)(&entry->work) &shost->scan_mutex rcu_read_lock &base->lock irq_context: 0 (wq_completion)events_unbound (work_completion)(&entry->work) &shost->scan_mutex rcu_read_lock &base->lock &obj_hash[i].lock irq_context: 0 (wq_completion)events_unbound (work_completion)(&entry->work) &shost->scan_mutex rcu_read_lock &virtscsi_vq->vq_lock irq_context: 0 (wq_completion)events_unbound (work_completion)(&entry->work) &shost->scan_mutex &x->wait#16 irq_context: hardirq &virtscsi_vq->vq_lock irq_context: softirq &x->wait#16 irq_context: 0 (wq_completion)events_unbound (work_completion)(&entry->work) &shost->scan_mutex &base->lock irq_context: 0 (wq_completion)events_unbound (work_completion)(&entry->work) &shost->scan_mutex &base->lock &obj_hash[i].lock irq_context: 0 (wq_completion)events_unbound (work_completion)(&entry->work) &shost->scan_mutex &rq->__lock irq_context: 0 (wq_completion)events_unbound (work_completion)(&entry->work) &shost->scan_mutex &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)events_unbound (work_completion)(&entry->work) &shost->scan_mutex &cfs_rq->removed.lock irq_context: 0 (wq_completion)events_unbound (work_completion)(&entry->work) &shost->scan_mutex (&timer.timer) irq_context: 0 (wq_completion)events_unbound (work_completion)(&entry->work) &shost->scan_mutex &sdev->state_mutex irq_context: 0 (wq_completion)events_unbound (work_completion)(&entry->work) &shost->scan_mutex &q->mq_freeze_lock irq_context: 0 (wq_completion)events_unbound (work_completion)(&entry->work) &shost->scan_mutex &q->mq_freeze_lock percpu_ref_switch_lock irq_context: 0 (wq_completion)events_unbound (work_completion)(&entry->work) &shost->scan_mutex &q->mq_freeze_lock percpu_ref_switch_lock rcu_read_lock &q->mq_freeze_wq irq_context: 0 (wq_completion)events_unbound (work_completion)(&entry->work) &shost->scan_mutex &q->mq_freeze_wq irq_context: 0 (wq_completion)events_unbound (work_completion)(&entry->work) &shost->scan_mutex percpu_ref_switch_lock irq_context: 0 (wq_completion)events_unbound (work_completion)(&entry->work) &shost->scan_mutex (&q->timeout) irq_context: 0 (wq_completion)events_unbound (work_completion)(&entry->work) &shost->scan_mutex (work_completion)(&q->timeout_work) irq_context: 0 (wq_completion)events_unbound (work_completion)(&entry->work) &shost->scan_mutex (work_completion)(&(&q->requeue_work)->work) irq_context: 0 (wq_completion)events_unbound (work_completion)(&entry->work) &shost->scan_mutex (work_completion)(&(&hctx->run_work)->work) irq_context: 0 (wq_completion)events_unbound (work_completion)(&entry->work) &shost->scan_mutex cpu_hotplug_lock irq_context: 0 (wq_completion)events_unbound (work_completion)(&entry->work) &shost->scan_mutex cpu_hotplug_lock cpuhp_state_mutex irq_context: 0 (wq_completion)events_unbound (work_completion)(&entry->work) &shost->scan_mutex &xa->xa_lock#10 irq_context: 0 (wq_completion)events_unbound (work_completion)(&entry->work) &shost->scan_mutex &q->unused_hctx_lock irq_context: 0 (wq_completion)events_unbound (work_completion)(&entry->work) &shost->scan_mutex (work_completion)(&sdev->requeue_work) irq_context: 0 (wq_completion)events_unbound (work_completion)(&entry->work) &shost->scan_mutex (work_completion)(&sdev->event_work) irq_context: 0 (wq_completion)events_unbound (work_completion)(&entry->work) &shost->scan_mutex pcpu_lock irq_context: 0 (wq_completion)events_unbound (work_completion)(&entry->work) &shost->scan_mutex &sdev->inquiry_mutex irq_context: 0 (wq_completion)events_unbound (work_completion)(&entry->work) &shost->scan_mutex &rq->__lock &cfs_rq->removed.lock irq_context: softirq &x->wait#16 &p->pi_lock irq_context: 0 (wq_completion)events_unbound (work_completion)(&entry->work) &shost->scan_mutex rcu_read_lock &rq->__lock irq_context: 0 (wq_completion)events_unbound (work_completion)(&entry->work) &shost->scan_mutex rcu_read_lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)events_unbound (work_completion)(&entry->work) &shost->scan_mutex rcu_read_lock &rq->__lock &cfs_rq->removed.lock irq_context: softirq &x->wait#16 &p->pi_lock &rq->__lock irq_context: softirq &x->wait#16 &p->pi_lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: softirq &x->wait#16 &p->pi_lock &cfs_rq->removed.lock irq_context: 0 (wq_completion)events_unbound (work_completion)(&entry->work) &shost->scan_mutex (console_sem).lock irq_context: 0 (wq_completion)events_unbound (work_completion)(&entry->work) &shost->scan_mutex console_lock console_srcu console_owner_lock irq_context: 0 (wq_completion)events_unbound (work_completion)(&entry->work) &shost->scan_mutex console_lock console_srcu console_owner irq_context: 0 (wq_completion)events_unbound (work_completion)(&entry->work) &shost->scan_mutex console_lock console_srcu console_owner &port_lock_key irq_context: 0 (wq_completion)events_unbound (work_completion)(&entry->work) &shost->scan_mutex console_lock console_srcu console_owner console_owner_lock irq_context: 0 nvmf_hosts_mutex irq_context: 0 subsys mutex#46 irq_context: 0 subsys mutex#46 &k->k_lock irq_context: 0 nvmf_transports_rwsem irq_context: 0 subsys mutex#47 irq_context: 0 subsys mutex#47 &k->k_lock irq_context: 0 &sb->s_type->i_mutex_key#7/1 &default_group_class[depth - 1]/2 fs_reclaim irq_context: 0 &sb->s_type->i_mutex_key#7/1 &default_group_class[depth - 1]/2 fs_reclaim mmu_notifier_invalidate_range_start irq_context: 0 &sb->s_type->i_mutex_key#7/1 &default_group_class[depth - 1]/2 pool_lock#2 irq_context: 0 &sb->s_type->i_mutex_key#7/1 &default_group_class[depth - 1]/2 &dentry->d_lock irq_context: 0 &sb->s_type->i_mutex_key#7/1 &default_group_class[depth - 1]/2 configfs_dirent_lock irq_context: 0 &sb->s_type->i_mutex_key#7/1 &default_group_class[depth - 1]/2 mmu_notifier_invalidate_range_start irq_context: 0 &sb->s_type->i_mutex_key#7/1 &default_group_class[depth - 1]/2 &sb->s_type->i_lock_key#19 irq_context: 0 &sb->s_type->i_mutex_key#7/1 &default_group_class[depth - 1]/2 &s->s_inode_list_lock irq_context: 0 &sb->s_type->i_mutex_key#7/1 &default_group_class[depth - 1]/2 tk_core.seq.seqcount irq_context: 0 &sb->s_type->i_mutex_key#7/1 &default_group_class[depth - 1]/2 &sb->s_type->i_lock_key#19 &dentry->d_lock irq_context: 0 &sb->s_type->i_mutex_key#7/1 &default_group_class[depth - 1]/2 &default_group_class[depth - 1]#3/2 irq_context: 0 nvmet_config_sem irq_context: 0 gdp_mutex &c->lock irq_context: 0 gdp_mutex &____s->seqcount irq_context: 0 subsys mutex#48 irq_context: 0 subsys mutex#48 &k->k_lock irq_context: 0 (wq_completion)events_unbound (work_completion)(&entry->work) &shost->scan_mutex batched_entropy_u8.lock irq_context: 0 (wq_completion)events_unbound (work_completion)(&entry->work) &shost->scan_mutex kfence_freelist_lock irq_context: 0 (wq_completion)events_unbound (work_completion)(&entry->work) &shost->scan_mutex &obj_hash[i].lock pool_lock irq_context: 0 (wq_completion)events_unbound (work_completion)(&entry->work) &shost->scan_mutex &set->tag_list_lock &q->mq_freeze_lock irq_context: 0 (wq_completion)events_unbound (work_completion)(&entry->work) &shost->scan_mutex &set->tag_list_lock &q->mq_freeze_lock percpu_ref_switch_lock irq_context: 0 (wq_completion)events_unbound (work_completion)(&entry->work) &shost->scan_mutex &set->tag_list_lock &q->mq_freeze_lock percpu_ref_switch_lock rcu_read_lock &q->mq_freeze_wq irq_context: 0 (wq_completion)events_unbound (work_completion)(&entry->work) &shost->scan_mutex &set->tag_list_lock percpu_ref_switch_lock irq_context: 0 (wq_completion)events_unbound (work_completion)(&entry->work) &shost->scan_mutex &set->tag_list_lock &q->mq_freeze_lock &q->mq_freeze_wq irq_context: 0 (wq_completion)events_unbound (work_completion)(&entry->work) &shost->scan_mutex &tags->lock irq_context: 0 &sb->s_type->i_mutex_key#7/1 &default_group_class[depth - 1]/2 &default_group_class[depth - 1]#3/2 fs_reclaim irq_context: 0 &sb->s_type->i_mutex_key#7/1 &default_group_class[depth - 1]/2 &default_group_class[depth - 1]#3/2 fs_reclaim mmu_notifier_invalidate_range_start irq_context: 0 &sb->s_type->i_mutex_key#7/1 &default_group_class[depth - 1]/2 &default_group_class[depth - 1]#3/2 pool_lock#2 irq_context: 0 &sb->s_type->i_mutex_key#7/1 &default_group_class[depth - 1]/2 &default_group_class[depth - 1]#3/2 &dentry->d_lock irq_context: 0 &sb->s_type->i_mutex_key#7/1 &default_group_class[depth - 1]/2 &default_group_class[depth - 1]#3/2 configfs_dirent_lock irq_context: 0 &sb->s_type->i_mutex_key#7/1 &default_group_class[depth - 1]/2 &default_group_class[depth - 1]#3/2 mmu_notifier_invalidate_range_start irq_context: 0 &sb->s_type->i_mutex_key#7/1 &default_group_class[depth - 1]/2 &default_group_class[depth - 1]#3/2 &sb->s_type->i_lock_key#19 irq_context: 0 &sb->s_type->i_mutex_key#7/1 &default_group_class[depth - 1]/2 &default_group_class[depth - 1]#3/2 &s->s_inode_list_lock irq_context: 0 &sb->s_type->i_mutex_key#7/1 &default_group_class[depth - 1]/2 &default_group_class[depth - 1]#3/2 tk_core.seq.seqcount irq_context: 0 &sb->s_type->i_mutex_key#7/1 &default_group_class[depth - 1]/2 &default_group_class[depth - 1]#3/2 &sb->s_type->i_lock_key#19 &dentry->d_lock irq_context: 0 &sb->s_type->i_mutex_key#7/1 &default_group_class[depth - 1]/2 &default_group_class[depth - 1]#3/2 &default_group_class[depth - 1]#4/2 irq_context: 0 &sb->s_type->i_mutex_key#7/1 &default_group_class[depth - 1]/2 &default_group_class[depth - 1]#3/2 &default_group_class[depth - 1]#4/2 fs_reclaim irq_context: 0 &sb->s_type->i_mutex_key#7/1 &default_group_class[depth - 1]/2 &default_group_class[depth - 1]#3/2 &default_group_class[depth - 1]#4/2 fs_reclaim mmu_notifier_invalidate_range_start irq_context: 0 &sb->s_type->i_mutex_key#7/1 &default_group_class[depth - 1]/2 &default_group_class[depth - 1]#3/2 &default_group_class[depth - 1]#4/2 pool_lock#2 irq_context: 0 &sb->s_type->i_mutex_key#7/1 &default_group_class[depth - 1]/2 &default_group_class[depth - 1]#3/2 &default_group_class[depth - 1]#4/2 &dentry->d_lock irq_context: 0 &sb->s_type->i_mutex_key#7/1 &default_group_class[depth - 1]/2 &default_group_class[depth - 1]#3/2 &default_group_class[depth - 1]#4/2 configfs_dirent_lock irq_context: 0 &sb->s_type->i_mutex_key#7/1 &default_group_class[depth - 1]/2 &default_group_class[depth - 1]#3/2 &default_group_class[depth - 1]#4/2 mmu_notifier_invalidate_range_start irq_context: 0 &sb->s_type->i_mutex_key#7/1 &default_group_class[depth - 1]/2 &default_group_class[depth - 1]#3/2 &default_group_class[depth - 1]#4/2 &sb->s_type->i_lock_key#19 irq_context: 0 &sb->s_type->i_mutex_key#7/1 &default_group_class[depth - 1]/2 &default_group_class[depth - 1]#3/2 &default_group_class[depth - 1]#4/2 &s->s_inode_list_lock irq_context: 0 &sb->s_type->i_mutex_key#7/1 &default_group_class[depth - 1]/2 &default_group_class[depth - 1]#3/2 &default_group_class[depth - 1]#4/2 tk_core.seq.seqcount irq_context: 0 &sb->s_type->i_mutex_key#7/1 &default_group_class[depth - 1]/2 &default_group_class[depth - 1]#3/2 &default_group_class[depth - 1]#4/2 &sb->s_type->i_lock_key#19 &dentry->d_lock irq_context: 0 &sb->s_type->i_mutex_key#7/1 &default_group_class[depth - 1]/2 &default_group_class[depth - 1]#3/2 &default_group_class[depth - 1]#4/2 &default_group_class[depth - 1]#5/2 irq_context: 0 &sb->s_type->i_mutex_key#7/1 &default_group_class[depth - 1]/2 &default_group_class[depth - 1]#3/2 &default_group_class[depth - 1]#4/2 &default_group_class[depth - 1]#5/2 fs_reclaim irq_context: 0 &sb->s_type->i_mutex_key#7/1 &default_group_class[depth - 1]/2 &default_group_class[depth - 1]#3/2 &default_group_class[depth - 1]#4/2 &default_group_class[depth - 1]#5/2 fs_reclaim mmu_notifier_invalidate_range_start irq_context: 0 &sb->s_type->i_mutex_key#7/1 &default_group_class[depth - 1]/2 &default_group_class[depth - 1]#3/2 &default_group_class[depth - 1]#4/2 &default_group_class[depth - 1]#5/2 pool_lock#2 irq_context: 0 &sb->s_type->i_mutex_key#7/1 &default_group_class[depth - 1]/2 &default_group_class[depth - 1]#3/2 &default_group_class[depth - 1]#4/2 &default_group_class[depth - 1]#5/2 &dentry->d_lock irq_context: 0 &sb->s_type->i_mutex_key#7/1 &default_group_class[depth - 1]/2 &default_group_class[depth - 1]#3/2 &default_group_class[depth - 1]#4/2 &default_group_class[depth - 1]#5/2 configfs_dirent_lock irq_context: 0 &sb->s_type->i_mutex_key#7/1 &default_group_class[depth - 1]/2 &default_group_class[depth - 1]#3/2 &default_group_class[depth - 1]#4/2 &default_group_class[depth - 1]#5/2 mmu_notifier_invalidate_range_start irq_context: 0 &sb->s_type->i_mutex_key#7/1 &default_group_class[depth - 1]/2 &default_group_class[depth - 1]#3/2 &default_group_class[depth - 1]#4/2 &default_group_class[depth - 1]#5/2 &sb->s_type->i_lock_key#19 irq_context: 0 &sb->s_type->i_mutex_key#7/1 &default_group_class[depth - 1]/2 &default_group_class[depth - 1]#3/2 &default_group_class[depth - 1]#4/2 &default_group_class[depth - 1]#5/2 &s->s_inode_list_lock irq_context: 0 &sb->s_type->i_mutex_key#7/1 &default_group_class[depth - 1]/2 &default_group_class[depth - 1]#3/2 &default_group_class[depth - 1]#4/2 &default_group_class[depth - 1]#5/2 tk_core.seq.seqcount irq_context: 0 &sb->s_type->i_mutex_key#7/1 &default_group_class[depth - 1]/2 &default_group_class[depth - 1]#3/2 &default_group_class[depth - 1]#4/2 &default_group_class[depth - 1]#5/2 &sb->s_type->i_lock_key#19 &dentry->d_lock irq_context: 0 &sb->s_type->i_mutex_key#7/1 &default_group_class[depth - 1]/2 &default_group_class[depth - 1]#3/2 &default_group_class[depth - 1]#4/2 &default_group_class[depth - 1]#5/2 &default_group_class[depth - 1]#6 irq_context: 0 &sb->s_type->i_mutex_key#7/1 &default_group_class[depth - 1]/2 &default_group_class[depth - 1]#3/2 &default_group_class[depth - 1]#4/2 &default_group_class[depth - 1]#5/2 &default_group_class[depth - 1]#6 fs_reclaim irq_context: 0 &sb->s_type->i_mutex_key#7/1 &default_group_class[depth - 1]/2 &default_group_class[depth - 1]#3/2 &default_group_class[depth - 1]#4/2 &default_group_class[depth - 1]#5/2 &default_group_class[depth - 1]#6 fs_reclaim mmu_notifier_invalidate_range_start irq_context: 0 &sb->s_type->i_mutex_key#7/1 &default_group_class[depth - 1]/2 &default_group_class[depth - 1]#3/2 &default_group_class[depth - 1]#4/2 &default_group_class[depth - 1]#5/2 &default_group_class[depth - 1]#6 pool_lock#2 irq_context: 0 &sb->s_type->i_mutex_key#7/1 &default_group_class[depth - 1]/2 &default_group_class[depth - 1]#3/2 &default_group_class[depth - 1]#4/2 &default_group_class[depth - 1]#5/2 &default_group_class[depth - 1]#6 configfs_dirent_lock irq_context: 0 &sb->s_type->i_mutex_key#7/1 &default_group_class[depth - 1]/2 &default_group_class[depth - 1]#3/2 &default_group_class[depth - 1]#4/2 &default_group_class[depth - 1]#5/2 &default_group_class[depth - 1]#6/2 irq_context: 0 backend_mutex irq_context: 0 scsi_mib_index_lock irq_context: 0 hba_lock irq_context: 0 device_mutex irq_context: 0 device_mutex fs_reclaim irq_context: 0 device_mutex fs_reclaim mmu_notifier_invalidate_range_start irq_context: 0 device_mutex pool_lock#2 irq_context: 0 (wq_completion)events_unbound (work_completion)(&entry->work) &shost->scan_mutex rcu_read_lock &obj_hash[i].lock pool_lock irq_context: 0 (wq_completion)events_unbound (work_completion)(&entry->work) &shost->scan_mutex rcu_read_lock fill_pool_map-wait-type-override pool_lock#2 irq_context: 0 (wq_completion)events_unbound (work_completion)(&entry->work) &shost->scan_mutex rcu_read_lock fill_pool_map-wait-type-override pool_lock irq_context: 0 (wq_completion)events_unbound (work_completion)(&entry->work) &shost->scan_mutex batched_entropy_u32.lock crngs.lock irq_context: 0 &hba->device_lock irq_context: 0 rcu_read_lock init_fs.seq.seqcount irq_context: 0 &sb->s_type->i_mutex_key irq_context: 0 &sb->s_type->i_mutex_key fs_reclaim irq_context: 0 &sb->s_type->i_mutex_key fs_reclaim mmu_notifier_invalidate_range_start irq_context: 0 &sb->s_type->i_mutex_key pool_lock#2 irq_context: 0 &sb->s_type->i_mutex_key &dentry->d_lock irq_context: 0 &sb->s_type->i_mutex_key rcu_read_lock rename_lock.seqcount irq_context: 0 &sb->s_type->i_mutex_key &dentry->d_lock &wq irq_context: 0 (wq_completion)events_unbound (work_completion)(&entry->work) &shost->scan_mutex &pcp->lock &zone->lock &____s->seqcount irq_context: 0 (wq_completion)events_unbound (work_completion)(&entry->work) &shost->scan_mutex fill_pool_map-wait-type-override pool_lock#2 irq_context: 0 (wq_completion)events_unbound (work_completion)(&entry->work) &shost->scan_mutex fill_pool_map-wait-type-override &c->lock irq_context: 0 (wq_completion)events_unbound (work_completion)(&entry->work) &shost->scan_mutex fill_pool_map-wait-type-override &pcp->lock &zone->lock irq_context: 0 (wq_completion)events_unbound (work_completion)(&entry->work) &shost->scan_mutex fill_pool_map-wait-type-override &zone->lock irq_context: 0 (wq_completion)events_unbound (work_completion)(&entry->work) &shost->scan_mutex fill_pool_map-wait-type-override &____s->seqcount irq_context: 0 (wq_completion)events_unbound (work_completion)(&entry->work) &shost->scan_mutex fill_pool_map-wait-type-override pool_lock irq_context: softirq rcu_callback percpu_ref_switch_lock irq_context: 0 mtd_table_mutex irq_context: 0 part_parser_lock irq_context: 0 (kmod_concurrent_max).lock irq_context: 0 &x->wait#17 irq_context: 0 (wq_completion)events_unbound (work_completion)(&sub_info->work) &sig->wait_chldexit irq_context: 0 (wq_completion)events_unbound (work_completion)(&sub_info->work) tasklist_lock irq_context: 0 (wq_completion)events_unbound (work_completion)(&sub_info->work) &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 tasklist_lock &sighand->siglock &sig->wait_chldexit &p->pi_lock irq_context: 0 tasklist_lock &sighand->siglock &sig->wait_chldexit &p->pi_lock &cfs_rq->removed.lock irq_context: 0 tasklist_lock &sighand->siglock &sig->wait_chldexit &p->pi_lock &rq->__lock irq_context: 0 tasklist_lock &sighand->siglock &sig->wait_chldexit &p->pi_lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)events_unbound (work_completion)(&sub_info->work) rcu_read_lock &____s->seqcount#5 irq_context: 0 (wq_completion)events_unbound (work_completion)(&sub_info->work) &prev->lock irq_context: 0 (wq_completion)events_unbound (work_completion)(&sub_info->work) &sighand->siglock &(&sig->stats_lock)->lock irq_context: 0 (wq_completion)events_unbound (work_completion)(&sub_info->work) &sighand->siglock &(&sig->stats_lock)->lock &____s->seqcount#5 irq_context: 0 (wq_completion)events_unbound (work_completion)(&sub_info->work) css_set_lock irq_context: 0 (wq_completion)events_unbound (work_completion)(&sub_info->work) tasklist_lock irq_context: 0 (wq_completion)events_unbound (work_completion)(&sub_info->work) tasklist_lock &sighand->siglock irq_context: 0 (wq_completion)events_unbound (work_completion)(&sub_info->work) tasklist_lock &sighand->siglock input_pool.lock irq_context: 0 (wq_completion)events_unbound (work_completion)(&sub_info->work) tasklist_lock &sighand->siglock &(&sig->stats_lock)->lock irq_context: 0 (wq_completion)events_unbound (work_completion)(&sub_info->work) tasklist_lock &sighand->siglock &(&sig->stats_lock)->lock &____s->seqcount#5 irq_context: 0 (wq_completion)events_unbound (work_completion)(&sub_info->work) tasklist_lock &sighand->siglock &(&sig->stats_lock)->lock &____s->seqcount#5 pidmap_lock irq_context: 0 (wq_completion)events_unbound (work_completion)(&sub_info->work) tasklist_lock &sighand->siglock &(&sig->stats_lock)->lock &____s->seqcount#5 &obj_hash[i].lock irq_context: 0 (wq_completion)events_unbound (work_completion)(&sub_info->work) tasklist_lock &sighand->siglock &(&sig->stats_lock)->lock &____s->seqcount#5 pool_lock#2 irq_context: 0 (wq_completion)events_unbound (work_completion)(&sub_info->work) tasklist_lock &obj_hash[i].lock irq_context: 0 (wq_completion)events_unbound (work_completion)(&sub_info->work) &x->wait#17 irq_context: 0 (wq_completion)events_unbound (work_completion)(&sub_info->work) &x->wait#17 &p->pi_lock irq_context: 0 (wq_completion)events_unbound (work_completion)(&sub_info->work) &x->wait#17 &p->pi_lock &rq->__lock irq_context: 0 (wq_completion)events_unbound (work_completion)(&sub_info->work) &x->wait#17 &p->pi_lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)events_unbound (work_completion)(&sub_info->work) &x->wait#17 &p->pi_lock &rq->__lock &cfs_rq->removed.lock irq_context: 0 (wq_completion)events_unbound (work_completion)(&sub_info->work) rcu_read_lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 mtd_table_mutex fs_reclaim irq_context: 0 mtd_table_mutex fs_reclaim mmu_notifier_invalidate_range_start irq_context: 0 mtd_table_mutex pool_lock#2 irq_context: 0 mtd_table_mutex &x->wait#9 irq_context: 0 mtd_table_mutex &obj_hash[i].lock irq_context: 0 mtd_table_mutex &k->list_lock irq_context: 0 mtd_table_mutex gdp_mutex irq_context: 0 mtd_table_mutex gdp_mutex &k->list_lock irq_context: 0 mtd_table_mutex gdp_mutex fs_reclaim irq_context: 0 mtd_table_mutex gdp_mutex fs_reclaim mmu_notifier_invalidate_range_start irq_context: 0 mtd_table_mutex gdp_mutex pool_lock#2 irq_context: 0 mtd_table_mutex gdp_mutex lock irq_context: 0 mtd_table_mutex gdp_mutex lock kernfs_idr_lock irq_context: 0 mtd_table_mutex gdp_mutex &root->kernfs_rwsem irq_context: 0 mtd_table_mutex gdp_mutex &root->kernfs_rwsem &root->kernfs_iattr_rwsem irq_context: 0 mtd_table_mutex lock irq_context: 0 mtd_table_mutex lock kernfs_idr_lock irq_context: 0 mtd_table_mutex &root->kernfs_rwsem irq_context: 0 mtd_table_mutex &root->kernfs_rwsem &root->kernfs_iattr_rwsem irq_context: 0 mtd_table_mutex bus_type_sem irq_context: 0 mtd_table_mutex sysfs_symlink_target_lock irq_context: 0 mtd_table_mutex &pcp->lock &zone->lock irq_context: 0 mtd_table_mutex &zone->lock irq_context: 0 mtd_table_mutex &____s->seqcount irq_context: 0 mtd_table_mutex &c->lock irq_context: 0 mtd_table_mutex &root->kernfs_rwsem irq_context: 0 mtd_table_mutex &dev->power.lock irq_context: 0 mtd_table_mutex dpm_list_mtx irq_context: 0 mtd_table_mutex req_lock irq_context: 0 mtd_table_mutex &p->pi_lock irq_context: 0 mtd_table_mutex &p->pi_lock &rq->__lock irq_context: 0 mtd_table_mutex &p->pi_lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 mtd_table_mutex &x->wait#11 irq_context: 0 (wq_completion)events_unbound (work_completion)(&entry->work) &shost->scan_mutex rcu_read_lock rcu_node_0 irq_context: 0 mtd_table_mutex &rq->__lock irq_context: 0 mtd_table_mutex &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 mtd_table_mutex rcu_read_lock &rq->__lock irq_context: 0 mtd_table_mutex rcu_read_lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 mtd_table_mutex uevent_sock_mutex irq_context: 0 mtd_table_mutex rcu_read_lock &pool->lock/1 irq_context: 0 mtd_table_mutex rcu_read_lock &pool->lock/1 &obj_hash[i].lock irq_context: 0 mtd_table_mutex rcu_read_lock &pool->lock/1 pool_lock#2 irq_context: 0 mtd_table_mutex rcu_read_lock &pool->lock/1 &p->pi_lock irq_context: 0 mtd_table_mutex rcu_read_lock &pool->lock/1 &p->pi_lock &rq->__lock irq_context: 0 mtd_table_mutex rcu_read_lock &pool->lock/1 &p->pi_lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 mtd_table_mutex running_helpers_waitq.lock irq_context: 0 mtd_table_mutex subsys mutex#49 irq_context: 0 mtd_table_mutex subsys mutex#49 &k->k_lock irq_context: 0 mtd_table_mutex devtree_lock irq_context: 0 mtd_table_mutex nvmem_ida.xa_lock irq_context: 0 mtd_table_mutex nvmem_cell_mutex irq_context: 0 mtd_table_mutex lock kernfs_idr_lock pool_lock#2 irq_context: 0 mtd_table_mutex &k->k_lock irq_context: 0 mtd_table_mutex fill_pool_map-wait-type-override pool_lock#2 irq_context: 0 mtd_table_mutex fill_pool_map-wait-type-override pool_lock irq_context: 0 mtd_table_mutex &dev->mutex &dev->power.lock irq_context: 0 mtd_table_mutex &dev->mutex &k->list_lock irq_context: 0 mtd_table_mutex &dev->mutex &k->k_lock irq_context: 0 mtd_table_mutex subsys mutex#50 irq_context: 0 mtd_table_mutex pin_fs_lock irq_context: 0 mtd_table_mutex &sb->s_type->i_mutex_key#3 irq_context: 0 mtd_table_mutex &sb->s_type->i_mutex_key#3 rename_lock.seqcount irq_context: 0 mtd_table_mutex &sb->s_type->i_mutex_key#3 fs_reclaim irq_context: 0 mtd_table_mutex &sb->s_type->i_mutex_key#3 fs_reclaim mmu_notifier_invalidate_range_start irq_context: 0 mtd_table_mutex &sb->s_type->i_mutex_key#3 pool_lock#2 irq_context: 0 mtd_table_mutex &sb->s_type->i_mutex_key#3 &dentry->d_lock irq_context: 0 mtd_table_mutex &sb->s_type->i_mutex_key#3 rcu_read_lock rename_lock.seqcount irq_context: 0 mtd_table_mutex &sb->s_type->i_mutex_key#3 &dentry->d_lock &wq irq_context: 0 mtd_table_mutex &sb->s_type->i_mutex_key#3 mmu_notifier_invalidate_range_start irq_context: 0 mtd_table_mutex &sb->s_type->i_mutex_key#3 &sb->s_type->i_lock_key#7 irq_context: 0 mtd_table_mutex &sb->s_type->i_mutex_key#3 &s->s_inode_list_lock irq_context: 0 mtd_table_mutex &sb->s_type->i_mutex_key#3 tk_core.seq.seqcount irq_context: 0 mtd_table_mutex &sb->s_type->i_mutex_key#3 &sb->s_type->i_lock_key#7 &dentry->d_lock irq_context: 0 mtd_table_mutex &cfs_rq->removed.lock irq_context: 0 (wq_completion)events_unbound (work_completion)(&entry->work) &shost->scan_mutex &set->tag_list_lock &rq->__lock irq_context: 0 mtd_table_mutex (console_sem).lock irq_context: 0 mtd_table_mutex console_lock console_srcu console_owner_lock irq_context: 0 mtd_table_mutex console_lock console_srcu console_owner irq_context: 0 mtd_table_mutex console_lock console_srcu console_owner &port_lock_key irq_context: 0 (wq_completion)events_unbound (work_completion)(&entry->work) &shost->scan_mutex rcu_read_lock fill_pool_map-wait-type-override &c->lock irq_context: 0 (wq_completion)events_unbound (work_completion)(&entry->work) &shost->scan_mutex rcu_read_lock fill_pool_map-wait-type-override &pcp->lock &zone->lock irq_context: 0 (wq_completion)events_unbound (work_completion)(&entry->work) &shost->scan_mutex rcu_read_lock fill_pool_map-wait-type-override &zone->lock irq_context: 0 (wq_completion)events_unbound (work_completion)(&entry->work) &shost->scan_mutex rcu_read_lock fill_pool_map-wait-type-override &____s->seqcount irq_context: 0 &sig->cred_guard_mutex &sb->s_type->i_mutex_key &pcp->lock &zone->lock &____s->seqcount irq_context: 0 mtd_table_mutex console_lock console_srcu console_owner console_owner_lock irq_context: 0 mtd_table_mutex pcpu_alloc_mutex irq_context: 0 mtd_table_mutex pcpu_alloc_mutex pcpu_lock irq_context: 0 mtd_table_mutex &obj_hash[i].lock pool_lock irq_context: 0 mtd_table_mutex cpu_hotplug_lock irq_context: 0 mtd_table_mutex batched_entropy_u32.lock irq_context: 0 mtd_table_mutex batched_entropy_u32.lock crngs.lock irq_context: 0 mtd_table_mutex mmu_notifier_invalidate_range_start irq_context: 0 mtd_table_mutex blk_queue_ida.xa_lock irq_context: 0 mtd_table_mutex &q->sysfs_lock irq_context: 0 mtd_table_mutex &q->sysfs_lock &q->unused_hctx_lock irq_context: 0 mtd_table_mutex &q->sysfs_lock mmu_notifier_invalidate_range_start irq_context: 0 mtd_table_mutex &q->sysfs_lock pool_lock#2 irq_context: 0 mtd_table_mutex &q->sysfs_lock &obj_hash[i].lock irq_context: 0 mtd_table_mutex &q->sysfs_lock cpu_hotplug_lock irq_context: 0 mtd_table_mutex &q->sysfs_lock cpu_hotplug_lock cpuhp_state_mutex irq_context: 0 mtd_table_mutex &q->sysfs_lock fs_reclaim irq_context: 0 mtd_table_mutex &q->sysfs_lock fs_reclaim mmu_notifier_invalidate_range_start irq_context: 0 mtd_table_mutex &q->sysfs_lock &xa->xa_lock#10 irq_context: 0 mtd_table_mutex &set->tag_list_lock irq_context: 0 mtd_table_mutex bio_slab_lock irq_context: 0 mtd_table_mutex rcu_read_lock pool_lock#2 irq_context: 0 mtd_table_mutex percpu_counters_lock irq_context: 0 (wq_completion)events_unbound (work_completion)(&entry->work) &shost->scan_mutex rcu_read_lock batched_entropy_u32.lock crngs.lock irq_context: 0 mtd_table_mutex &sb->s_type->i_lock_key#3 irq_context: 0 mtd_table_mutex &s->s_inode_list_lock irq_context: 0 mtd_table_mutex &xa->xa_lock#11 irq_context: 0 mtd_table_mutex lock &q->queue_lock irq_context: 0 mtd_table_mutex lock &q->queue_lock &blkcg->lock irq_context: 0 mtd_table_mutex &q->mq_freeze_lock irq_context: 0 mtd_table_mutex &q->mq_freeze_lock percpu_ref_switch_lock irq_context: 0 mtd_table_mutex &q->mq_freeze_lock percpu_ref_switch_lock rcu_read_lock &q->mq_freeze_wq irq_context: 0 mtd_table_mutex set->srcu irq_context: 0 mtd_table_mutex percpu_ref_switch_lock irq_context: 0 mtd_table_mutex &q->queue_lock irq_context: 0 mtd_table_mutex &q->queue_lock pool_lock#2 irq_context: 0 mtd_table_mutex &q->mq_freeze_lock &q->mq_freeze_wq irq_context: 0 mtd_table_mutex &q->queue_lock &c->lock irq_context: 0 mtd_table_mutex &q->queue_lock &pcp->lock &zone->lock irq_context: 0 mtd_table_mutex &q->queue_lock &zone->lock irq_context: 0 mtd_table_mutex &q->queue_lock &____s->seqcount irq_context: 0 mtd_table_mutex &q->queue_lock pcpu_lock irq_context: 0 mtd_table_mutex &q->queue_lock &obj_hash[i].lock irq_context: 0 mtd_table_mutex &q->queue_lock percpu_counters_lock irq_context: 0 (wq_completion)events_unbound (work_completion)(&entry->work) &shost->scan_mutex &q->sysfs_lock fill_pool_map-wait-type-override pool_lock#2 irq_context: 0 mtd_table_mutex &bdev->bd_size_lock irq_context: 0 mtd_table_mutex elv_list_lock irq_context: 0 (wq_completion)events_unbound (work_completion)(&entry->work) &shost->scan_mutex &q->sysfs_lock fill_pool_map-wait-type-override pool_lock irq_context: 0 mtd_table_mutex (work_completion)(&(&q->requeue_work)->work) irq_context: 0 mtd_table_mutex (work_completion)(&(&hctx->run_work)->work) irq_context: 0 mtd_table_mutex &q->debugfs_mutex irq_context: 0 mtd_table_mutex subsys mutex#37 irq_context: 0 mtd_table_mutex subsys mutex#37 &k->k_lock irq_context: 0 mtd_table_mutex dev_hotplug_mutex irq_context: 0 mtd_table_mutex dev_hotplug_mutex &dev->power.lock irq_context: 0 mtd_table_mutex &q->sysfs_dir_lock irq_context: 0 mtd_table_mutex &q->sysfs_dir_lock fs_reclaim irq_context: 0 mtd_table_mutex &q->sysfs_dir_lock fs_reclaim mmu_notifier_invalidate_range_start irq_context: 0 mtd_table_mutex &q->sysfs_dir_lock pool_lock#2 irq_context: 0 mtd_table_mutex &q->sysfs_dir_lock lock irq_context: 0 mtd_table_mutex &q->sysfs_dir_lock lock kernfs_idr_lock irq_context: 0 mtd_table_mutex &q->sysfs_dir_lock &root->kernfs_rwsem irq_context: 0 mtd_table_mutex &q->sysfs_dir_lock &root->kernfs_rwsem &root->kernfs_iattr_rwsem irq_context: 0 mtd_table_mutex &q->sysfs_dir_lock &c->lock irq_context: 0 mtd_table_mutex &q->sysfs_dir_lock &pcp->lock &zone->lock irq_context: 0 mtd_table_mutex &q->sysfs_dir_lock &zone->lock irq_context: 0 mtd_table_mutex &q->sysfs_dir_lock &____s->seqcount irq_context: 0 mtd_table_mutex &q->sysfs_dir_lock &q->sysfs_lock irq_context: 0 mtd_table_mutex &q->sysfs_dir_lock &q->sysfs_lock &q->debugfs_mutex irq_context: 0 mtd_table_mutex &q->sysfs_dir_lock &q->sysfs_lock &q->debugfs_mutex pin_fs_lock irq_context: 0 mtd_table_mutex &q->sysfs_dir_lock &q->sysfs_lock &q->debugfs_mutex &sb->s_type->i_mutex_key#3 irq_context: 0 mtd_table_mutex &q->sysfs_dir_lock &q->sysfs_lock &q->debugfs_mutex &sb->s_type->i_mutex_key#3 rename_lock.seqcount irq_context: 0 mtd_table_mutex &q->sysfs_dir_lock &q->sysfs_lock &q->debugfs_mutex &sb->s_type->i_mutex_key#3 fs_reclaim irq_context: 0 mtd_table_mutex &q->sysfs_dir_lock &q->sysfs_lock &q->debugfs_mutex &sb->s_type->i_mutex_key#3 fs_reclaim mmu_notifier_invalidate_range_start irq_context: 0 mtd_table_mutex &q->sysfs_dir_lock &q->sysfs_lock &q->debugfs_mutex &sb->s_type->i_mutex_key#3 pool_lock#2 irq_context: 0 mtd_table_mutex &q->sysfs_dir_lock &q->sysfs_lock &q->debugfs_mutex &sb->s_type->i_mutex_key#3 &dentry->d_lock irq_context: 0 mtd_table_mutex &q->sysfs_dir_lock &q->sysfs_lock &q->debugfs_mutex &sb->s_type->i_mutex_key#3 rcu_read_lock rename_lock.seqcount irq_context: 0 mtd_table_mutex &q->sysfs_dir_lock &q->sysfs_lock &q->debugfs_mutex &sb->s_type->i_mutex_key#3 &dentry->d_lock &wq irq_context: 0 mtd_table_mutex &q->sysfs_dir_lock &q->sysfs_lock &q->debugfs_mutex &sb->s_type->i_mutex_key#3 mmu_notifier_invalidate_range_start irq_context: 0 mtd_table_mutex &q->sysfs_dir_lock &q->sysfs_lock &q->debugfs_mutex &sb->s_type->i_mutex_key#3 &sb->s_type->i_lock_key#7 irq_context: 0 mtd_table_mutex &q->sysfs_dir_lock &q->sysfs_lock &q->debugfs_mutex &sb->s_type->i_mutex_key#3 &s->s_inode_list_lock irq_context: 0 mtd_table_mutex &q->sysfs_dir_lock &q->sysfs_lock &q->debugfs_mutex &sb->s_type->i_mutex_key#3 tk_core.seq.seqcount irq_context: 0 mtd_table_mutex &q->sysfs_dir_lock &q->sysfs_lock &q->debugfs_mutex &sb->s_type->i_mutex_key#3 &sb->s_type->i_lock_key#7 &dentry->d_lock irq_context: 0 mtd_table_mutex &q->sysfs_dir_lock &q->sysfs_lock &q->debugfs_mutex &sb->s_type->i_mutex_key#3 &c->lock irq_context: 0 mtd_table_mutex &q->sysfs_dir_lock &q->sysfs_lock &q->debugfs_mutex &sb->s_type->i_mutex_key#3 &pcp->lock &zone->lock irq_context: 0 mtd_table_mutex &q->sysfs_dir_lock &q->sysfs_lock &q->debugfs_mutex &sb->s_type->i_mutex_key#3 &zone->lock irq_context: 0 mtd_table_mutex &q->sysfs_dir_lock &q->sysfs_lock &q->debugfs_mutex &sb->s_type->i_mutex_key#3 &____s->seqcount irq_context: 0 mtd_table_mutex &q->sysfs_dir_lock &q->sysfs_lock &q->debugfs_mutex &sb->s_type->i_mutex_key#3 rcu_read_lock pool_lock#2 irq_context: 0 mtd_table_mutex &q->sysfs_dir_lock &q->sysfs_lock &q->debugfs_mutex &sb->s_type->i_mutex_key#3 &obj_hash[i].lock irq_context: 0 mtd_table_mutex &q->sysfs_dir_lock &q->sysfs_lock fs_reclaim irq_context: 0 mtd_table_mutex &q->sysfs_dir_lock &q->sysfs_lock fs_reclaim mmu_notifier_invalidate_range_start irq_context: 0 mtd_table_mutex &q->sysfs_dir_lock &q->sysfs_lock pool_lock#2 irq_context: 0 mtd_table_mutex &q->sysfs_dir_lock &q->sysfs_lock lock irq_context: 0 mtd_table_mutex &q->sysfs_dir_lock &q->sysfs_lock lock kernfs_idr_lock irq_context: 0 mtd_table_mutex &q->sysfs_dir_lock &q->sysfs_lock &root->kernfs_rwsem irq_context: 0 mtd_table_mutex &q->sysfs_dir_lock &q->sysfs_lock &root->kernfs_rwsem &root->kernfs_iattr_rwsem irq_context: 0 mtd_table_mutex &q->sysfs_dir_lock &q->sysfs_lock &c->lock irq_context: 0 mtd_table_mutex &q->sysfs_dir_lock &q->sysfs_lock &pcp->lock &zone->lock irq_context: 0 mtd_table_mutex &q->sysfs_dir_lock &q->sysfs_lock &zone->lock irq_context: 0 mtd_table_mutex &q->sysfs_dir_lock &q->sysfs_lock &____s->seqcount irq_context: 0 mtd_table_mutex &q->sysfs_dir_lock &q->sysfs_lock pcpu_alloc_mutex irq_context: 0 mtd_table_mutex &q->sysfs_dir_lock &q->sysfs_lock pcpu_alloc_mutex pcpu_lock irq_context: 0 mtd_table_mutex &q->sysfs_dir_lock &q->sysfs_lock &obj_hash[i].lock irq_context: 0 mtd_table_mutex &q->sysfs_dir_lock &q->sysfs_lock &q->rq_qos_mutex irq_context: 0 mtd_table_mutex &q->sysfs_dir_lock &q->sysfs_lock &q->rq_qos_mutex &q->mq_freeze_lock irq_context: 0 mtd_table_mutex &q->sysfs_dir_lock &q->sysfs_lock &q->rq_qos_mutex &q->mq_freeze_lock percpu_ref_switch_lock irq_context: 0 mtd_table_mutex &q->sysfs_dir_lock &q->sysfs_lock &q->rq_qos_mutex &q->mq_freeze_lock percpu_ref_switch_lock rcu_read_lock &q->mq_freeze_wq irq_context: 0 mtd_table_mutex &q->sysfs_dir_lock &q->sysfs_lock &q->rq_qos_mutex set->srcu irq_context: 0 mtd_table_mutex &q->sysfs_dir_lock &q->sysfs_lock &q->rq_qos_mutex percpu_ref_switch_lock irq_context: 0 mtd_table_mutex &q->sysfs_dir_lock &q->sysfs_lock &q->rq_qos_mutex &q->mq_freeze_lock &q->mq_freeze_wq irq_context: 0 mtd_table_mutex &q->sysfs_dir_lock &q->sysfs_lock &q->rq_qos_mutex &q->debugfs_mutex irq_context: 0 mtd_table_mutex &q->sysfs_dir_lock &q->sysfs_lock &q->rq_qos_mutex &q->debugfs_mutex pin_fs_lock irq_context: 0 mtd_table_mutex &q->sysfs_dir_lock &q->sysfs_lock &q->rq_qos_mutex &q->debugfs_mutex &sb->s_type->i_mutex_key#3 irq_context: 0 mtd_table_mutex &q->sysfs_dir_lock &q->sysfs_lock &q->rq_qos_mutex &q->debugfs_mutex &sb->s_type->i_mutex_key#3 rename_lock.seqcount irq_context: 0 mtd_table_mutex &q->sysfs_dir_lock &q->sysfs_lock &q->rq_qos_mutex &q->debugfs_mutex &sb->s_type->i_mutex_key#3 fs_reclaim irq_context: 0 mtd_table_mutex &q->sysfs_dir_lock &q->sysfs_lock &q->rq_qos_mutex &q->debugfs_mutex &sb->s_type->i_mutex_key#3 fs_reclaim mmu_notifier_invalidate_range_start irq_context: 0 mtd_table_mutex &q->sysfs_dir_lock &q->sysfs_lock &q->rq_qos_mutex &q->debugfs_mutex &sb->s_type->i_mutex_key#3 pool_lock#2 irq_context: 0 mtd_table_mutex &q->sysfs_dir_lock &q->sysfs_lock &q->rq_qos_mutex &q->debugfs_mutex &sb->s_type->i_mutex_key#3 &dentry->d_lock irq_context: 0 mtd_table_mutex &q->sysfs_dir_lock &q->sysfs_lock &q->rq_qos_mutex &q->debugfs_mutex &sb->s_type->i_mutex_key#3 rcu_read_lock rename_lock.seqcount irq_context: 0 mtd_table_mutex &q->sysfs_dir_lock &q->sysfs_lock &q->rq_qos_mutex &q->debugfs_mutex &sb->s_type->i_mutex_key#3 &dentry->d_lock &wq irq_context: 0 mtd_table_mutex &q->sysfs_dir_lock &q->sysfs_lock &q->rq_qos_mutex &q->debugfs_mutex &sb->s_type->i_mutex_key#3 mmu_notifier_invalidate_range_start irq_context: 0 mtd_table_mutex &q->sysfs_dir_lock &q->sysfs_lock &q->rq_qos_mutex &q->debugfs_mutex &sb->s_type->i_mutex_key#3 &sb->s_type->i_lock_key#7 irq_context: 0 mtd_table_mutex &q->sysfs_dir_lock &q->sysfs_lock &q->rq_qos_mutex &q->debugfs_mutex &sb->s_type->i_mutex_key#3 &s->s_inode_list_lock irq_context: 0 mtd_table_mutex &q->sysfs_dir_lock &q->sysfs_lock &q->rq_qos_mutex &q->debugfs_mutex &sb->s_type->i_mutex_key#3 tk_core.seq.seqcount irq_context: 0 mtd_table_mutex &q->sysfs_dir_lock &q->sysfs_lock &q->rq_qos_mutex &q->debugfs_mutex &sb->s_type->i_mutex_key#3 &sb->s_type->i_lock_key#7 &dentry->d_lock irq_context: 0 mtd_table_mutex &q->sysfs_dir_lock &q->sysfs_lock &q->rq_qos_mutex &q->debugfs_mutex &sb->s_type->i_mutex_key#3 &c->lock irq_context: 0 mtd_table_mutex &q->sysfs_dir_lock &q->sysfs_lock &q->rq_qos_mutex &q->debugfs_mutex &sb->s_type->i_mutex_key#3 &pcp->lock &zone->lock irq_context: 0 mtd_table_mutex &q->sysfs_dir_lock &q->sysfs_lock &q->rq_qos_mutex &q->debugfs_mutex &sb->s_type->i_mutex_key#3 &zone->lock irq_context: 0 mtd_table_mutex &q->sysfs_dir_lock &q->sysfs_lock &q->rq_qos_mutex &q->debugfs_mutex &sb->s_type->i_mutex_key#3 &____s->seqcount irq_context: 0 mtd_table_mutex &q->sysfs_dir_lock &q->sysfs_lock &stats->lock irq_context: 0 (wq_completion)events_unbound (work_completion)(&entry->work) &shost->scan_mutex quarantine_lock irq_context: 0 mtd_table_mutex subsys mutex#38 irq_context: 0 mtd_table_mutex subsys mutex#38 &k->k_lock irq_context: 0 mtd_table_mutex cgwb_lock irq_context: 0 mtd_table_mutex bdi_lock irq_context: 0 mtd_table_mutex inode_hash_lock irq_context: 0 mtd_table_mutex inode_hash_lock &sb->s_type->i_lock_key#3 irq_context: 0 rtnl_mutex &pcp->lock &zone->lock &____s->seqcount irq_context: 0 rtnl_mutex stack_depot_init_mutex irq_context: 0 rtnl_mutex rcu_read_lock pool_lock#2 irq_context: 0 rtnl_mutex cpu_hotplug_lock irq_context: 0 rtnl_mutex cpu_hotplug_lock wq_pool_mutex irq_context: 0 rtnl_mutex cpu_hotplug_lock wq_pool_mutex fs_reclaim irq_context: 0 rtnl_mutex cpu_hotplug_lock wq_pool_mutex fs_reclaim mmu_notifier_invalidate_range_start irq_context: 0 rtnl_mutex cpu_hotplug_lock wq_pool_mutex pool_lock#2 irq_context: 0 rtnl_mutex cpu_hotplug_lock wq_pool_mutex &obj_hash[i].lock irq_context: 0 rtnl_mutex cpu_hotplug_lock wq_pool_mutex &wq->mutex irq_context: 0 rtnl_mutex cpu_hotplug_lock wq_pool_mutex &wq->mutex &pool->lock/1 irq_context: 0 rtnl_mutex kthread_create_lock irq_context: 0 rtnl_mutex &p->pi_lock irq_context: 0 rtnl_mutex &p->pi_lock &cfs_rq->removed.lock irq_context: 0 rtnl_mutex &p->pi_lock &rq->__lock irq_context: 0 rtnl_mutex &p->pi_lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 rtnl_mutex &x->wait irq_context: 0 rtnl_mutex &p->pi_lock &rq->__lock &cfs_rq->removed.lock irq_context: 0 rtnl_mutex wq_pool_mutex irq_context: 0 rtnl_mutex wq_pool_mutex &wq->mutex irq_context: 0 rtnl_mutex crngs.lock irq_context: 0 rtnl_mutex &xa->xa_lock#3 irq_context: 0 rtnl_mutex net_rwsem irq_context: 0 rtnl_mutex net_rwsem &list->lock#2 irq_context: 0 rtnl_mutex &x->wait#9 irq_context: 0 rtnl_mutex &k->list_lock irq_context: 0 rtnl_mutex gdp_mutex irq_context: 0 rtnl_mutex gdp_mutex &k->list_lock irq_context: 0 rtnl_mutex lock irq_context: 0 rtnl_mutex lock kernfs_idr_lock irq_context: 0 rtnl_mutex &root->kernfs_rwsem irq_context: 0 rtnl_mutex &root->kernfs_rwsem &root->kernfs_iattr_rwsem irq_context: 0 rtnl_mutex bus_type_sem irq_context: 0 rtnl_mutex sysfs_symlink_target_lock irq_context: 0 rtnl_mutex lock kernfs_idr_lock pool_lock#2 irq_context: 0 rtnl_mutex &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 rtnl_mutex &root->kernfs_rwsem irq_context: 0 rtnl_mutex &dev->power.lock irq_context: 0 rtnl_mutex dpm_list_mtx irq_context: 0 rtnl_mutex uevent_sock_mutex irq_context: 0 rtnl_mutex uevent_sock_mutex fs_reclaim irq_context: 0 rtnl_mutex uevent_sock_mutex fs_reclaim mmu_notifier_invalidate_range_start irq_context: 0 rtnl_mutex uevent_sock_mutex pool_lock#2 irq_context: 0 rtnl_mutex uevent_sock_mutex nl_table_lock irq_context: 0 rtnl_mutex uevent_sock_mutex &obj_hash[i].lock irq_context: 0 rtnl_mutex uevent_sock_mutex nl_table_wait.lock irq_context: 0 rtnl_mutex rcu_read_lock &pool->lock/1 irq_context: 0 rtnl_mutex rcu_read_lock &pool->lock/1 &obj_hash[i].lock irq_context: 0 rtnl_mutex rcu_read_lock &pool->lock/1 pool_lock#2 irq_context: 0 rtnl_mutex rcu_read_lock &pool->lock/1 &p->pi_lock irq_context: 0 rtnl_mutex rcu_read_lock &pool->lock/1 &p->pi_lock &rq->__lock irq_context: 0 rtnl_mutex rcu_read_lock &pool->lock/1 &p->pi_lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 rtnl_mutex running_helpers_waitq.lock irq_context: 0 rtnl_mutex subsys mutex#17 irq_context: 0 rtnl_mutex subsys mutex#17 &k->k_lock irq_context: 0 rtnl_mutex &dir->lock#2 irq_context: 0 rtnl_mutex uevent_sock_mutex &c->lock irq_context: 0 rtnl_mutex uevent_sock_mutex &pcp->lock &zone->lock irq_context: 0 rtnl_mutex uevent_sock_mutex &rq->__lock irq_context: 0 rtnl_mutex uevent_sock_mutex &zone->lock irq_context: 0 rtnl_mutex uevent_sock_mutex &____s->seqcount irq_context: 0 rtnl_mutex uevent_sock_mutex &n->list_lock irq_context: 0 rtnl_mutex rcu_read_lock &pool->lock/1 &p->pi_lock &rq->__lock &cfs_rq->removed.lock irq_context: 0 rtnl_mutex &obj_hash[i].lock pool_lock irq_context: 0 rtnl_mutex dev_hotplug_mutex irq_context: 0 rtnl_mutex dev_hotplug_mutex &dev->power.lock irq_context: 0 rtnl_mutex dev_base_lock irq_context: 0 rtnl_mutex input_pool.lock irq_context: 0 rtnl_mutex pcpu_alloc_mutex irq_context: 0 rtnl_mutex pcpu_alloc_mutex pcpu_lock irq_context: 0 rtnl_mutex batched_entropy_u32.lock irq_context: 0 rtnl_mutex &tbl->lock irq_context: 0 rtnl_mutex sysctl_lock irq_context: 0 rtnl_mutex nl_table_lock irq_context: 0 rtnl_mutex nl_table_wait.lock irq_context: 0 rtnl_mutex rcu_read_lock &bond->stats_lock irq_context: 0 rtnl_mutex lweventlist_lock irq_context: 0 rtnl_mutex lweventlist_lock pool_lock#2 irq_context: 0 rtnl_mutex lweventlist_lock &dir->lock#2 irq_context: 0 rtnl_mutex rcu_read_lock &pool->lock irq_context: 0 rtnl_mutex rcu_read_lock &pool->lock &obj_hash[i].lock irq_context: 0 rtnl_mutex rcu_read_lock &pool->lock &p->pi_lock irq_context: 0 rtnl_mutex rcu_read_lock &pool->lock &p->pi_lock &rq->__lock irq_context: 0 rtnl_mutex rcu_read_lock &pool->lock &p->pi_lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)events (linkwatch_work).work irq_context: 0 (wq_completion)events (linkwatch_work).work rtnl_mutex irq_context: 0 (wq_completion)events (linkwatch_work).work rtnl_mutex rtnl_mutex.wait_lock irq_context: 0 (wq_completion)events (linkwatch_work).work rtnl_mutex &pool->lock irq_context: 0 (wq_completion)events (linkwatch_work).work rtnl_mutex &rq->__lock irq_context: 0 (wq_completion)events (linkwatch_work).work rtnl_mutex &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)events_unbound (work_completion)(&sub_info->work) cgroup_threadgroup_rwsem &rq->__lock &cfs_rq->removed.lock irq_context: 0 (wq_completion)gid-cache-wq irq_context: 0 (wq_completion)gid-cache-wq (work_completion)(&ndev_work->work) irq_context: 0 (wq_completion)gid-cache-wq (work_completion)(&ndev_work->work) devices_rwsem irq_context: 0 (wq_completion)gid-cache-wq (work_completion)(&ndev_work->work) &obj_hash[i].lock irq_context: 0 (wq_completion)gid-cache-wq (work_completion)(&ndev_work->work) pool_lock#2 irq_context: 0 rtnl_mutex.wait_lock irq_context: 0 (wq_completion)events (linkwatch_work).work rtnl_mutex lweventlist_lock irq_context: 0 (wq_completion)events (linkwatch_work).work rtnl_mutex lweventlist_lock pool_lock#2 irq_context: 0 (wq_completion)events (linkwatch_work).work rtnl_mutex lweventlist_lock &dir->lock#2 irq_context: 0 (wq_completion)events (linkwatch_work).work rtnl_mutex dev_base_lock irq_context: 0 pernet_ops_rwsem rtnl_mutex proc_subdir_lock irq_context: 0 pernet_ops_rwsem rtnl_mutex proc_inum_ida.xa_lock irq_context: 0 pernet_ops_rwsem rtnl_mutex proc_subdir_lock irq_context: 0 once_lock irq_context: 0 once_lock crngs.lock irq_context: 0 rcu_read_lock &pool->lock pool_lock#2 irq_context: 0 (wq_completion)events (work_completion)(&w->work) irq_context: 0 (wq_completion)events (work_completion)(&w->work) cpu_hotplug_lock irq_context: 0 (wq_completion)events (work_completion)(&w->work) cpu_hotplug_lock jump_label_mutex irq_context: 0 (wq_completion)events (work_completion)(&w->work) cpu_hotplug_lock jump_label_mutex text_mutex irq_context: 0 (wq_completion)events (work_completion)(&w->work) cpu_hotplug_lock jump_label_mutex text_mutex ptlock_ptr(page)#2 irq_context: 0 (wq_completion)events (work_completion)(&w->work) &obj_hash[i].lock irq_context: 0 (inet6addr_validator_chain).rwsem irq_context: 0 (inetaddr_validator_chain).rwsem irq_context: 0 pernet_ops_rwsem rtnl_mutex stack_depot_init_mutex irq_context: 0 pernet_ops_rwsem rtnl_mutex crngs.lock irq_context: 0 pernet_ops_rwsem rtnl_mutex rcu_read_lock &pool->lock/1 irq_context: 0 pernet_ops_rwsem rtnl_mutex rcu_read_lock &pool->lock/1 &obj_hash[i].lock irq_context: 0 pernet_ops_rwsem rtnl_mutex rcu_read_lock &pool->lock/1 pool_lock#2 irq_context: 0 pernet_ops_rwsem rtnl_mutex rcu_read_lock &pool->lock/1 &p->pi_lock irq_context: 0 pernet_ops_rwsem rtnl_mutex &rq->__lock irq_context: 0 pernet_ops_rwsem rtnl_mutex uevent_sock_mutex &c->lock irq_context: 0 pernet_ops_rwsem rtnl_mutex uevent_sock_mutex &pcp->lock &zone->lock irq_context: 0 pernet_ops_rwsem rtnl_mutex uevent_sock_mutex &zone->lock irq_context: 0 pernet_ops_rwsem rtnl_mutex uevent_sock_mutex &____s->seqcount irq_context: 0 pernet_ops_rwsem rtnl_mutex rcu_read_lock &pool->lock/1 &p->pi_lock &rq->__lock irq_context: 0 pernet_ops_rwsem rtnl_mutex rcu_read_lock &pool->lock/1 &p->pi_lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)events_unbound (work_completion)(&entry->work) &shost->scan_mutex &q->sysfs_lock &pcp->lock &zone->lock &____s->seqcount irq_context: 0 pernet_ops_rwsem rtnl_mutex &cfs_rq->removed.lock irq_context: 0 pernet_ops_rwsem rtnl_mutex rcu_read_lock &pool->lock/1 &p->pi_lock &rq->__lock &cfs_rq->removed.lock irq_context: 0 subsys mutex#51 irq_context: 0 subsys mutex#51 &k->k_lock irq_context: 0 gpio_lookup_lock irq_context: 0 mdio_board_lock irq_context: 0 mode_list_lock irq_context: 0 &dev->mutex stack_depot_init_mutex irq_context: 0 &dev->mutex fill_pool_map-wait-type-override pool_lock#2 irq_context: 0 &dev->mutex fill_pool_map-wait-type-override &c->lock irq_context: 0 &dev->mutex fill_pool_map-wait-type-override &____s->seqcount irq_context: 0 &dev->mutex fill_pool_map-wait-type-override pool_lock irq_context: 0 &dev->mutex napi_hash_lock irq_context: 0 &sig->cred_guard_mutex rcu_read_lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 &dev->mutex &md->mutex &domain->mutex sparse_irq_lock rcu_read_lock pool_lock#2 irq_context: 0 &dev->mutex &md->mutex &domain->mutex &c->lock irq_context: 0 &dev->mutex &md->mutex &domain->mutex sparse_irq_lock fill_pool_map-wait-type-override pool_lock#2 irq_context: 0 &dev->mutex &md->mutex &domain->mutex sparse_irq_lock fill_pool_map-wait-type-override pool_lock irq_context: 0 &dev->mutex &md->mutex &domain->mutex sparse_irq_lock lock kernfs_idr_lock pool_lock#2 irq_context: 0 &dev->mutex &md->mutex &domain->mutex sparse_irq_lock &n->list_lock &c->lock irq_context: softirq &x->wait#16 &p->pi_lock &rq->__lock &cfs_rq->removed.lock irq_context: 0 (wq_completion)events_unbound (work_completion)(&entry->work) &shost->scan_mutex pool_lock irq_context: 0 &dev->mutex fs_reclaim &rq->__lock irq_context: 0 &dev->mutex cpu_hotplug_lock &md->mutex irq_context: 0 &dev->mutex cpu_hotplug_lock &irq_desc_lock_class irq_context: 0 &dev->mutex cpu_hotplug_lock xps_map_mutex irq_context: 0 &dev->mutex cpu_hotplug_lock xps_map_mutex fs_reclaim irq_context: 0 &dev->mutex cpu_hotplug_lock xps_map_mutex fs_reclaim mmu_notifier_invalidate_range_start irq_context: 0 &dev->mutex cpu_hotplug_lock xps_map_mutex pool_lock#2 irq_context: 0 &dev->mutex cpu_hotplug_lock xps_map_mutex &c->lock irq_context: 0 &dev->mutex cpu_hotplug_lock xps_map_mutex &n->list_lock irq_context: 0 &dev->mutex cpu_hotplug_lock xps_map_mutex &pcp->lock &zone->lock irq_context: 0 &dev->mutex cpu_hotplug_lock xps_map_mutex &zone->lock irq_context: 0 &dev->mutex cpu_hotplug_lock xps_map_mutex &____s->seqcount irq_context: 0 &dev->mutex cpu_hotplug_lock xps_map_mutex jump_label_mutex irq_context: 0 &dev->mutex cpu_hotplug_lock xps_map_mutex jump_label_mutex text_mutex irq_context: 0 &dev->mutex cpu_hotplug_lock xps_map_mutex jump_label_mutex text_mutex ptlock_ptr(page)#2 irq_context: 0 &dev->mutex cpu_hotplug_lock xps_map_mutex &obj_hash[i].lock irq_context: 0 &dev->mutex cpu_hotplug_lock xps_map_mutex krc.lock irq_context: 0 &dev->mutex rtnl_mutex irq_context: 0 &dev->mutex rtnl_mutex &pcp->lock &zone->lock irq_context: 0 &dev->mutex rtnl_mutex &zone->lock irq_context: 0 &dev->mutex rtnl_mutex &____s->seqcount irq_context: 0 &dev->mutex rtnl_mutex pool_lock#2 irq_context: 0 &dev->mutex rtnl_mutex rcu_read_lock pool_lock#2 irq_context: 0 &dev->mutex rtnl_mutex &obj_hash[i].lock irq_context: 0 &dev->mutex rtnl_mutex fs_reclaim irq_context: 0 &dev->mutex rtnl_mutex fs_reclaim mmu_notifier_invalidate_range_start irq_context: 0 &dev->mutex rtnl_mutex &xa->xa_lock#3 irq_context: 0 &dev->mutex rtnl_mutex net_rwsem irq_context: 0 &dev->mutex rtnl_mutex net_rwsem &list->lock#2 irq_context: 0 &dev->mutex rtnl_mutex &x->wait#9 irq_context: 0 &dev->mutex rtnl_mutex &k->list_lock irq_context: 0 &dev->mutex rtnl_mutex gdp_mutex irq_context: 0 &dev->mutex rtnl_mutex gdp_mutex &k->list_lock irq_context: 0 &dev->mutex rtnl_mutex gdp_mutex fs_reclaim irq_context: 0 &dev->mutex rtnl_mutex gdp_mutex fs_reclaim mmu_notifier_invalidate_range_start irq_context: 0 &dev->mutex rtnl_mutex gdp_mutex pool_lock#2 irq_context: 0 &dev->mutex rtnl_mutex gdp_mutex lock irq_context: 0 &dev->mutex rtnl_mutex gdp_mutex lock kernfs_idr_lock irq_context: 0 &dev->mutex rtnl_mutex gdp_mutex &root->kernfs_rwsem irq_context: 0 &dev->mutex rtnl_mutex gdp_mutex &root->kernfs_rwsem &root->kernfs_iattr_rwsem irq_context: 0 &dev->mutex rtnl_mutex gdp_mutex kobj_ns_type_lock irq_context: 0 &dev->mutex rtnl_mutex lock irq_context: 0 &dev->mutex rtnl_mutex lock kernfs_idr_lock irq_context: 0 &dev->mutex rtnl_mutex &root->kernfs_rwsem irq_context: 0 &dev->mutex rtnl_mutex &root->kernfs_rwsem &root->kernfs_iattr_rwsem irq_context: 0 &dev->mutex rtnl_mutex bus_type_sem irq_context: 0 &dev->mutex rtnl_mutex sysfs_symlink_target_lock irq_context: 0 &dev->mutex rtnl_mutex &c->lock irq_context: 0 &dev->mutex rtnl_mutex lock kernfs_idr_lock pool_lock#2 irq_context: 0 &dev->mutex rtnl_mutex &root->kernfs_rwsem irq_context: 0 &dev->mutex rtnl_mutex &dev->power.lock irq_context: 0 &dev->mutex rtnl_mutex dpm_list_mtx irq_context: 0 &dev->mutex rtnl_mutex &pcp->lock &zone->lock &____s->seqcount irq_context: 0 &dev->mutex rtnl_mutex uevent_sock_mutex irq_context: 0 &dev->mutex rtnl_mutex uevent_sock_mutex fs_reclaim irq_context: 0 &dev->mutex rtnl_mutex uevent_sock_mutex fs_reclaim mmu_notifier_invalidate_range_start irq_context: 0 &dev->mutex rtnl_mutex uevent_sock_mutex pool_lock#2 irq_context: 0 &dev->mutex rtnl_mutex uevent_sock_mutex nl_table_lock irq_context: 0 &dev->mutex rtnl_mutex uevent_sock_mutex &obj_hash[i].lock irq_context: 0 &dev->mutex rtnl_mutex uevent_sock_mutex nl_table_wait.lock irq_context: 0 &dev->mutex rtnl_mutex rcu_read_lock &pool->lock/1 irq_context: 0 &dev->mutex rtnl_mutex rcu_read_lock &pool->lock/1 &obj_hash[i].lock irq_context: 0 &dev->mutex rtnl_mutex rcu_read_lock &pool->lock/1 pool_lock#2 irq_context: 0 &dev->mutex rtnl_mutex rcu_read_lock &pool->lock/1 &p->pi_lock irq_context: 0 &dev->mutex rtnl_mutex rcu_read_lock &pool->lock/1 &p->pi_lock &rq->__lock irq_context: 0 &dev->mutex rtnl_mutex rcu_read_lock &pool->lock/1 &p->pi_lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)events_unbound (work_completion)(&entry->work) &shost->scan_mutex &q->sysfs_lock &rq->__lock irq_context: 0 &dev->mutex rtnl_mutex running_helpers_waitq.lock irq_context: 0 &dev->mutex rtnl_mutex &k->k_lock irq_context: 0 &dev->mutex rtnl_mutex subsys mutex#17 irq_context: 0 &dev->mutex rtnl_mutex subsys mutex#17 &k->k_lock irq_context: 0 &dev->mutex rtnl_mutex &dir->lock#2 irq_context: 0 &dev->mutex rtnl_mutex &rq->__lock irq_context: 0 &dev->mutex rtnl_mutex &cfs_rq->removed.lock irq_context: 0 &dev->mutex rtnl_mutex dev_hotplug_mutex irq_context: 0 &dev->mutex rtnl_mutex dev_hotplug_mutex &dev->power.lock irq_context: 0 &dev->mutex rtnl_mutex dev_base_lock irq_context: 0 &dev->mutex rtnl_mutex input_pool.lock irq_context: 0 &dev->mutex rtnl_mutex pcpu_alloc_mutex irq_context: 0 &dev->mutex rtnl_mutex pcpu_alloc_mutex pcpu_lock irq_context: 0 &dev->mutex rtnl_mutex batched_entropy_u32.lock irq_context: 0 &dev->mutex rtnl_mutex &tbl->lock irq_context: 0 &dev->mutex rtnl_mutex sysctl_lock irq_context: 0 &dev->mutex rtnl_mutex nl_table_lock irq_context: 0 &dev->mutex rtnl_mutex nl_table_wait.lock irq_context: hardirq &irq_desc_lock_class tmp_mask_lock irq_context: hardirq &irq_desc_lock_class tmp_mask_lock vector_lock irq_context: softirq rcu_callback &meta->lock irq_context: softirq rcu_callback kfence_freelist_lock irq_context: 0 &dev->mutex cpu_hotplug_lock cpuhp_state_mutex irq_context: 0 &dev->mutex lweventlist_lock irq_context: 0 &dev->mutex lweventlist_lock pool_lock#2 irq_context: 0 &dev->mutex lweventlist_lock &dir->lock#2 irq_context: 0 &dev->mutex &base->lock &obj_hash[i].lock irq_context: 0 &dev->mutex rcu_read_lock &pool->lock irq_context: 0 &dev->mutex rcu_read_lock &pool->lock &obj_hash[i].lock irq_context: 0 &dev->mutex rcu_read_lock &pool->lock pool_lock#2 irq_context: 0 &dev->mutex rcu_read_lock &pool->lock &p->pi_lock irq_context: 0 &dev->mutex rcu_read_lock &pool->lock &p->pi_lock &rq->__lock irq_context: 0 &dev->mutex rcu_read_lock &pool->lock &p->pi_lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 &dev->mutex rcu_read_lock &pool->lock &p->pi_lock &rq->__lock &cfs_rq->removed.lock irq_context: 0 l3mdev_lock irq_context: 0 (wq_completion)events (work_completion)(&vi->config_work) irq_context: 0 (wq_completion)events_unbound (work_completion)(&entry->work) &shost->scan_mutex rcu_read_lock &cfs_rq->removed.lock irq_context: 0 (wq_completion)events_unbound (work_completion)(&entry->work) &shost->scan_mutex rcu_read_lock fill_pool_map-wait-type-override rcu_node_0 irq_context: 0 (wq_completion)events_unbound (work_completion)(&entry->work) &shost->scan_mutex rcu_read_lock fill_pool_map-wait-type-override &rq->__lock irq_context: 0 subsys mutex#52 irq_context: 0 subsys mutex#52 &k->k_lock irq_context: 0 compressor_list_lock irq_context: 0 compressor_list_lock pool_lock#2 irq_context: 0 (wq_completion)events_unbound (work_completion)(&entry->work) &shost->scan_mutex &meta->lock irq_context: 0 (wq_completion)events_unbound (work_completion)(&entry->work) &shost->scan_mutex &q->mq_freeze_lock &rq->__lock irq_context: 0 pernet_ops_rwsem hwsim_netgroup_ida.xa_lock irq_context: 0 hwsim_radio_lock irq_context: 0 subsys mutex#53 irq_context: 0 subsys mutex#53 &k->k_lock irq_context: 0 deferred_probe_mutex irq_context: 0 rtnl_mutex param_lock irq_context: 0 rtnl_mutex param_lock rate_ctrl_mutex irq_context: 0 rtnl_mutex (console_sem).lock irq_context: 0 rtnl_mutex &rdev->wiphy.mtx irq_context: 0 rtnl_mutex &rdev->wiphy.mtx fs_reclaim irq_context: 0 rtnl_mutex &rdev->wiphy.mtx fs_reclaim mmu_notifier_invalidate_range_start irq_context: 0 rtnl_mutex &rdev->wiphy.mtx pool_lock#2 irq_context: 0 rtnl_mutex &rdev->wiphy.mtx &k->list_lock irq_context: 0 rtnl_mutex &rdev->wiphy.mtx gdp_mutex irq_context: 0 rtnl_mutex &rdev->wiphy.mtx gdp_mutex &k->list_lock irq_context: 0 rtnl_mutex &rdev->wiphy.mtx gdp_mutex fs_reclaim irq_context: 0 rtnl_mutex &rdev->wiphy.mtx gdp_mutex fs_reclaim mmu_notifier_invalidate_range_start irq_context: 0 rtnl_mutex &rdev->wiphy.mtx gdp_mutex pool_lock#2 irq_context: 0 rtnl_mutex &rdev->wiphy.mtx gdp_mutex lock irq_context: 0 rtnl_mutex &rdev->wiphy.mtx gdp_mutex lock kernfs_idr_lock irq_context: 0 rtnl_mutex &rdev->wiphy.mtx gdp_mutex &root->kernfs_rwsem irq_context: 0 rtnl_mutex &rdev->wiphy.mtx gdp_mutex &root->kernfs_rwsem &root->kernfs_iattr_rwsem irq_context: 0 rtnl_mutex &rdev->wiphy.mtx gdp_mutex kobj_ns_type_lock irq_context: 0 rtnl_mutex &rdev->wiphy.mtx lock irq_context: 0 rtnl_mutex &rdev->wiphy.mtx lock kernfs_idr_lock irq_context: 0 rtnl_mutex &rdev->wiphy.mtx &root->kernfs_rwsem irq_context: 0 rtnl_mutex &rdev->wiphy.mtx &root->kernfs_rwsem &root->kernfs_iattr_rwsem irq_context: 0 rtnl_mutex &rdev->wiphy.mtx bus_type_sem irq_context: 0 rtnl_mutex &rdev->wiphy.mtx sysfs_symlink_target_lock irq_context: 0 rtnl_mutex &rdev->wiphy.mtx &____s->seqcount irq_context: 0 rtnl_mutex &rdev->wiphy.mtx &c->lock irq_context: 0 rtnl_mutex &rdev->wiphy.mtx &pcp->lock &zone->lock irq_context: 0 rtnl_mutex &rdev->wiphy.mtx &zone->lock irq_context: 0 rtnl_mutex &rdev->wiphy.mtx &root->kernfs_rwsem irq_context: 0 rtnl_mutex &rdev->wiphy.mtx &dev->power.lock irq_context: 0 rtnl_mutex &rdev->wiphy.mtx dpm_list_mtx irq_context: 0 rtnl_mutex &rdev->wiphy.mtx uevent_sock_mutex irq_context: 0 rtnl_mutex &rdev->wiphy.mtx uevent_sock_mutex fs_reclaim irq_context: 0 rtnl_mutex &rdev->wiphy.mtx uevent_sock_mutex fs_reclaim mmu_notifier_invalidate_range_start irq_context: 0 rtnl_mutex &rdev->wiphy.mtx uevent_sock_mutex &c->lock irq_context: 0 rtnl_mutex &rdev->wiphy.mtx uevent_sock_mutex &pcp->lock &zone->lock irq_context: 0 rtnl_mutex &rdev->wiphy.mtx uevent_sock_mutex &zone->lock irq_context: 0 rtnl_mutex &rdev->wiphy.mtx uevent_sock_mutex &____s->seqcount irq_context: 0 rtnl_mutex &rdev->wiphy.mtx uevent_sock_mutex pool_lock#2 irq_context: 0 rtnl_mutex &rdev->wiphy.mtx uevent_sock_mutex nl_table_lock irq_context: 0 rtnl_mutex &rdev->wiphy.mtx uevent_sock_mutex &obj_hash[i].lock irq_context: 0 rtnl_mutex &rdev->wiphy.mtx uevent_sock_mutex nl_table_wait.lock irq_context: 0 rtnl_mutex &rdev->wiphy.mtx &obj_hash[i].lock irq_context: 0 rtnl_mutex &rdev->wiphy.mtx rcu_read_lock &pool->lock/1 irq_context: 0 rtnl_mutex &rdev->wiphy.mtx rcu_read_lock &pool->lock/1 &obj_hash[i].lock irq_context: 0 rtnl_mutex &rdev->wiphy.mtx rcu_read_lock &pool->lock/1 pool_lock#2 irq_context: 0 rtnl_mutex &rdev->wiphy.mtx rcu_read_lock &pool->lock/1 &p->pi_lock irq_context: 0 rtnl_mutex &rdev->wiphy.mtx rcu_read_lock &pool->lock/1 &p->pi_lock &rq->__lock irq_context: 0 rtnl_mutex &rdev->wiphy.mtx rcu_read_lock &pool->lock/1 &p->pi_lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 rtnl_mutex &rdev->wiphy.mtx running_helpers_waitq.lock irq_context: 0 rtnl_mutex &rdev->wiphy.mtx &k->k_lock irq_context: 0 rtnl_mutex &rdev->wiphy.mtx subsys mutex#54 irq_context: 0 rtnl_mutex &rdev->wiphy.mtx subsys mutex#54 &k->k_lock irq_context: 0 rtnl_mutex &rdev->wiphy.mtx pin_fs_lock irq_context: 0 rtnl_mutex &rdev->wiphy.mtx &sb->s_type->i_mutex_key#3 irq_context: 0 rtnl_mutex &rdev->wiphy.mtx &sb->s_type->i_mutex_key#3 &sb->s_type->i_lock_key#7 irq_context: 0 rtnl_mutex &rdev->wiphy.mtx &sb->s_type->i_mutex_key#3 rename_lock.seqcount irq_context: 0 rtnl_mutex &rdev->wiphy.mtx &sb->s_type->i_mutex_key#3 fs_reclaim irq_context: 0 rtnl_mutex &rdev->wiphy.mtx &sb->s_type->i_mutex_key#3 fs_reclaim mmu_notifier_invalidate_range_start irq_context: 0 rtnl_mutex &rdev->wiphy.mtx &sb->s_type->i_mutex_key#3 pool_lock#2 irq_context: 0 rtnl_mutex &rdev->wiphy.mtx &sb->s_type->i_mutex_key#3 &dentry->d_lock irq_context: 0 rtnl_mutex &rdev->wiphy.mtx &sb->s_type->i_mutex_key#3 rcu_read_lock rename_lock.seqcount irq_context: 0 rtnl_mutex &rdev->wiphy.mtx &sb->s_type->i_mutex_key#3 &dentry->d_lock &wq irq_context: 0 rtnl_mutex &rdev->wiphy.mtx &sb->s_type->i_mutex_key#3 mmu_notifier_invalidate_range_start irq_context: 0 rtnl_mutex &rdev->wiphy.mtx &sb->s_type->i_mutex_key#3 &s->s_inode_list_lock irq_context: 0 rtnl_mutex &rdev->wiphy.mtx &sb->s_type->i_mutex_key#3 tk_core.seq.seqcount irq_context: 0 rtnl_mutex &rdev->wiphy.mtx &sb->s_type->i_mutex_key#3 &sb->s_type->i_lock_key#7 &dentry->d_lock irq_context: 0 rtnl_mutex &rdev->wiphy.mtx nl_table_lock irq_context: 0 rtnl_mutex &rdev->wiphy.mtx nl_table_wait.lock irq_context: 0 rtnl_mutex &cfs_rq->removed.lock irq_context: 0 rtnl_mutex &rdev->wiphy.mtx reg_requests_lock irq_context: 0 rtnl_mutex &base->lock irq_context: 0 rtnl_mutex &base->lock &obj_hash[i].lock irq_context: 0 rfkill_global_mutex irq_context: 0 rfkill_global_mutex fs_reclaim irq_context: 0 rfkill_global_mutex fs_reclaim mmu_notifier_invalidate_range_start irq_context: 0 rfkill_global_mutex pool_lock#2 irq_context: 0 rfkill_global_mutex &k->list_lock irq_context: 0 rfkill_global_mutex lock irq_context: 0 rfkill_global_mutex lock kernfs_idr_lock irq_context: 0 rfkill_global_mutex &root->kernfs_rwsem irq_context: 0 rfkill_global_mutex &root->kernfs_rwsem &root->kernfs_iattr_rwsem irq_context: 0 rfkill_global_mutex bus_type_sem irq_context: 0 rfkill_global_mutex sysfs_symlink_target_lock irq_context: 0 rfkill_global_mutex &c->lock irq_context: 0 rfkill_global_mutex &pcp->lock &zone->lock irq_context: 0 rfkill_global_mutex &zone->lock irq_context: 0 rfkill_global_mutex &____s->seqcount irq_context: 0 rfkill_global_mutex &root->kernfs_rwsem irq_context: 0 rfkill_global_mutex lock kernfs_idr_lock pool_lock#2 irq_context: 0 rfkill_global_mutex &dev->power.lock irq_context: 0 rfkill_global_mutex dpm_list_mtx irq_context: 0 rfkill_global_mutex &rfkill->lock irq_context: 0 rfkill_global_mutex uevent_sock_mutex irq_context: 0 rfkill_global_mutex &obj_hash[i].lock irq_context: 0 rfkill_global_mutex rcu_read_lock &pool->lock/1 irq_context: 0 rfkill_global_mutex rcu_read_lock &pool->lock/1 &obj_hash[i].lock irq_context: 0 rfkill_global_mutex rcu_read_lock &pool->lock/1 pool_lock#2 irq_context: 0 rfkill_global_mutex rcu_read_lock &pool->lock/1 &p->pi_lock irq_context: 0 rfkill_global_mutex rcu_read_lock &pool->lock/1 &p->pi_lock &rq->__lock irq_context: 0 rfkill_global_mutex rcu_read_lock &pool->lock/1 &p->pi_lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 rfkill_global_mutex rcu_read_lock &pool->lock/1 &p->pi_lock &rq->__lock &cfs_rq->removed.lock irq_context: 0 rfkill_global_mutex running_helpers_waitq.lock irq_context: 0 rfkill_global_mutex &k->k_lock irq_context: 0 rfkill_global_mutex subsys mutex#41 irq_context: 0 rfkill_global_mutex subsys mutex#41 &k->k_lock irq_context: 0 rfkill_global_mutex triggers_list_lock irq_context: 0 rfkill_global_mutex leds_list_lock irq_context: 0 rfkill_global_mutex rcu_read_lock &pool->lock irq_context: 0 rfkill_global_mutex rcu_read_lock &pool->lock &obj_hash[i].lock irq_context: 0 rfkill_global_mutex rcu_read_lock &pool->lock pool_lock#2 irq_context: 0 rfkill_global_mutex rcu_read_lock &pool->lock &p->pi_lock irq_context: 0 rfkill_global_mutex rcu_read_lock &pool->lock &p->pi_lock &rq->__lock irq_context: 0 rfkill_global_mutex rcu_read_lock &pool->lock &p->pi_lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 rfkill_global_mutex rcu_read_lock &pool->lock &p->pi_lock &rq->__lock &cfs_rq->removed.lock irq_context: 0 rfkill_global_mutex &rq->__lock irq_context: 0 rfkill_global_mutex &cfs_rq->removed.lock irq_context: 0 rfkill_global_mutex.wait_lock irq_context: 0 rtnl_mutex &rdev->wiphy.mtx stack_depot_init_mutex irq_context: 0 rtnl_mutex &rdev->wiphy.mtx pcpu_alloc_mutex irq_context: 0 rtnl_mutex &rdev->wiphy.mtx pcpu_alloc_mutex pcpu_lock irq_context: 0 rtnl_mutex &rdev->wiphy.mtx rcu_read_lock pool_lock#2 irq_context: 0 rtnl_mutex &rdev->wiphy.mtx &local->iflist_mtx irq_context: 0 rtnl_mutex &rdev->wiphy.mtx &obj_hash[i].lock pool_lock irq_context: 0 rtnl_mutex &rdev->wiphy.mtx &sb->s_type->i_mutex_key#3 &c->lock irq_context: 0 rtnl_mutex &rdev->wiphy.mtx &sb->s_type->i_mutex_key#3 &pcp->lock &zone->lock irq_context: 0 rtnl_mutex &rdev->wiphy.mtx &sb->s_type->i_mutex_key#3 &zone->lock irq_context: 0 rtnl_mutex &rdev->wiphy.mtx &sb->s_type->i_mutex_key#3 &____s->seqcount irq_context: 0 rtnl_mutex &rdev->wiphy.mtx &xa->xa_lock#3 irq_context: 0 rtnl_mutex &rdev->wiphy.mtx net_rwsem irq_context: 0 rtnl_mutex &rdev->wiphy.mtx net_rwsem &list->lock#2 irq_context: 0 rtnl_mutex &rdev->wiphy.mtx &x->wait#9 irq_context: 0 rtnl_mutex &rdev->wiphy.mtx lock kernfs_idr_lock pool_lock#2 irq_context: 0 rtnl_mutex &rdev->wiphy.mtx subsys mutex#17 irq_context: 0 rtnl_mutex &rdev->wiphy.mtx subsys mutex#17 &k->k_lock irq_context: 0 rtnl_mutex &rdev->wiphy.mtx &dir->lock#2 irq_context: 0 rtnl_mutex &rdev->wiphy.mtx uevent_sock_mutex &rq->__lock irq_context: 0 rtnl_mutex &rdev->wiphy.mtx uevent_sock_mutex &cfs_rq->removed.lock irq_context: 0 rtnl_mutex &rdev->wiphy.mtx &root->kernfs_rwsem &root->kernfs_iattr_rwsem &rq->__lock irq_context: 0 rtnl_mutex &rdev->wiphy.mtx &pcp->lock &zone->lock &____s->seqcount irq_context: 0 rtnl_mutex &rdev->wiphy.mtx dev_hotplug_mutex irq_context: 0 rtnl_mutex &rdev->wiphy.mtx dev_hotplug_mutex &dev->power.lock irq_context: 0 rtnl_mutex &rdev->wiphy.mtx dev_base_lock irq_context: 0 rtnl_mutex &rdev->wiphy.mtx input_pool.lock irq_context: 0 rtnl_mutex &rdev->wiphy.mtx batched_entropy_u32.lock irq_context: 0 rtnl_mutex &rdev->wiphy.mtx &tbl->lock irq_context: 0 rtnl_mutex &rdev->wiphy.mtx sysctl_lock irq_context: 0 rtnl_mutex &rdev->wiphy.mtx &wdev->mtx irq_context: 0 rtnl_mutex &rdev->wiphy.mtx &fq->lock irq_context: 0 rtnl_mutex &rdev->wiphy.mtx quarantine_lock irq_context: 0 hwsim_radio_lock rcu_read_lock rhashtable_bucket irq_context: 0 rtnl_mutex &rdev->wiphy.mtx rcu_read_lock &pool->lock/1 fill_pool_map-wait-type-override pool_lock#2 irq_context: 0 (wq_completion)events_unbound (work_completion)(&sub_info->work) cgroup_threadgroup_rwsem fill_pool_map-wait-type-override pool_lock#2 irq_context: 0 (wq_completion)events_unbound (work_completion)(&sub_info->work) cgroup_threadgroup_rwsem fill_pool_map-wait-type-override &c->lock irq_context: 0 (wq_completion)events_unbound (work_completion)(&sub_info->work) cgroup_threadgroup_rwsem fill_pool_map-wait-type-override &pcp->lock &zone->lock irq_context: 0 rtnl_mutex &rdev->wiphy.mtx rcu_read_lock &pool->lock/1 fill_pool_map-wait-type-override pool_lock irq_context: 0 (wq_completion)events_unbound (work_completion)(&sub_info->work) cgroup_threadgroup_rwsem fill_pool_map-wait-type-override &zone->lock irq_context: 0 (wq_completion)events_unbound (work_completion)(&sub_info->work) cgroup_threadgroup_rwsem fill_pool_map-wait-type-override &____s->seqcount irq_context: 0 rtnl_mutex &rdev->wiphy.mtx rcu_read_lock &pool->lock/1 &p->pi_lock &rq->__lock &cfs_rq->removed.lock irq_context: 0 (wq_completion)events_unbound (work_completion)(&sub_info->work) cgroup_threadgroup_rwsem fill_pool_map-wait-type-override pool_lock irq_context: 0 rtnl_mutex &rdev->wiphy.mtx fill_pool_map-wait-type-override pool_lock#2 irq_context: 0 rtnl_mutex &rdev->wiphy.mtx fill_pool_map-wait-type-override &c->lock irq_context: 0 rtnl_mutex &rdev->wiphy.mtx fill_pool_map-wait-type-override &pcp->lock &zone->lock irq_context: 0 rtnl_mutex &rdev->wiphy.mtx fill_pool_map-wait-type-override &zone->lock irq_context: 0 rtnl_mutex &rdev->wiphy.mtx fill_pool_map-wait-type-override &____s->seqcount irq_context: 0 rtnl_mutex &rdev->wiphy.mtx fill_pool_map-wait-type-override pool_lock irq_context: 0 rtnl_mutex &rdev->wiphy.mtx &sb->s_type->i_mutex_key#3 &rq->__lock irq_context: hardirq &vb->stop_update_lock rcu_read_lock &pool->lock &p->pi_lock &rq->__lock &cfs_rq->removed.lock irq_context: 0 &dev->mutex crngs.lock irq_context: 0 &dev->mutex cpu_hotplug_lock wq_pool_mutex rcu_read_lock pool_lock#2 irq_context: 0 &dev->mutex rtnl_mutex subsys mutex#55 irq_context: 0 &dev->mutex rtnl_mutex subsys mutex#55 &k->k_lock irq_context: 0 &dev->mutex rtnl_mutex stack_depot_init_mutex irq_context: 0 &dev->mutex rtnl_mutex crngs.lock irq_context: 0 &dev->mutex rtnl_mutex &sdata->sec_mtx irq_context: 0 &dev->mutex rtnl_mutex &sdata->sec_mtx &sec->lock irq_context: 0 &dev->mutex rtnl_mutex uevent_sock_mutex &c->lock irq_context: 0 &dev->mutex rtnl_mutex uevent_sock_mutex &pcp->lock &zone->lock irq_context: 0 &dev->mutex rtnl_mutex uevent_sock_mutex &zone->lock irq_context: 0 &dev->mutex rtnl_mutex uevent_sock_mutex &____s->seqcount irq_context: 0 &dev->mutex rtnl_mutex &local->iflist_mtx#2 irq_context: 0 &dev->mutex hwsim_phys_lock irq_context: 0 &dev->mutex nl_table_lock irq_context: 0 &dev->mutex nl_table_wait.lock irq_context: 0 &dev->mutex &n->list_lock &c->lock irq_context: 0 &dev->mutex rtnl_mutex lock kernfs_idr_lock &c->lock irq_context: 0 &dev->mutex rtnl_mutex lock kernfs_idr_lock &____s->seqcount irq_context: 0 &dev->mutex rtnl_mutex batched_entropy_u32.lock crngs.lock irq_context: 0 &dev->mutex hwsim_phys_lock fs_reclaim irq_context: 0 &dev->mutex hwsim_phys_lock fs_reclaim mmu_notifier_invalidate_range_start irq_context: 0 &dev->mutex hwsim_phys_lock pool_lock#2 irq_context: 0 (wq_completion)events_unbound (work_completion)(&entry->work) &shost->scan_mutex &q->sysfs_lock mmu_notifier_invalidate_range_start &rq->__lock irq_context: 0 (wq_completion)events_unbound (work_completion)(&entry->work) &shost->scan_mutex &x->wait#15 irq_context: 0 (wq_completion)events_unbound (work_completion)(&entry->work) &shost->scan_mutex &k->list_lock irq_context: 0 (wq_completion)events_unbound (work_completion)(&entry->work) &shost->scan_mutex lock irq_context: 0 (wq_completion)events_unbound (work_completion)(&entry->work) &shost->scan_mutex lock kernfs_idr_lock irq_context: 0 (wq_completion)events_unbound (work_completion)(&entry->work) &shost->scan_mutex &root->kernfs_rwsem irq_context: 0 (wq_completion)events_unbound (work_completion)(&entry->work) &shost->scan_mutex &root->kernfs_rwsem &root->kernfs_iattr_rwsem irq_context: 0 (wq_completion)events_unbound (work_completion)(&entry->work) &shost->scan_mutex bus_type_sem irq_context: 0 (wq_completion)events_unbound (work_completion)(&entry->work) &shost->scan_mutex sysfs_symlink_target_lock irq_context: 0 (wq_completion)events_unbound (work_completion)(&entry->work) &shost->scan_mutex &k->k_lock irq_context: 0 (wq_completion)events_unbound (work_completion)(&entry->work) &shost->scan_mutex &root->kernfs_rwsem irq_context: 0 (wq_completion)events_unbound (work_completion)(&entry->work) &shost->scan_mutex dpm_list_mtx irq_context: 0 (wq_completion)events_unbound (work_completion)(&entry->work) &shost->scan_mutex uevent_sock_mutex irq_context: 0 (wq_completion)events_unbound (work_completion)(&entry->work) &shost->scan_mutex rcu_read_lock &pool->lock/1 irq_context: 0 (wq_completion)events_unbound (work_completion)(&entry->work) &shost->scan_mutex rcu_read_lock &pool->lock/1 &obj_hash[i].lock irq_context: 0 (wq_completion)events_unbound (work_completion)(&entry->work) &shost->scan_mutex rcu_read_lock &pool->lock/1 pool_lock#2 irq_context: 0 (wq_completion)events_unbound (work_completion)(&entry->work) &shost->scan_mutex rcu_read_lock &pool->lock/1 &p->pi_lock irq_context: 0 (wq_completion)events_unbound (work_completion)(&entry->work) &shost->scan_mutex rcu_read_lock &pool->lock/1 &p->pi_lock &rq->__lock irq_context: 0 (wq_completion)events_unbound (work_completion)(&entry->work) &shost->scan_mutex rcu_read_lock &pool->lock/1 &p->pi_lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)events_unbound (work_completion)(&entry->work) &shost->scan_mutex &dev->mutex &dev->power.lock irq_context: 0 (wq_completion)events_unbound (work_completion)(&entry->work) &shost->scan_mutex &dev->mutex &k->list_lock irq_context: 0 (wq_completion)events_unbound (work_completion)(&entry->work) &shost->scan_mutex &dev->mutex &k->k_lock irq_context: 0 (wq_completion)events_unbound (work_completion)(&entry->work) &shost->scan_mutex subsys mutex#43 irq_context: 0 (wq_completion)events_unbound (work_completion)(&entry->work) &shost->scan_mutex device_links_srcu irq_context: 0 (wq_completion)events_unbound (work_completion)(&entry->work) &shost->scan_mutex &dev->power.lock &dev->power.lock/1 irq_context: 0 (wq_completion)events_unbound (work_completion)(&entry->work) &shost->scan_mutex lock kernfs_idr_lock pool_lock#2 irq_context: 0 (wq_completion)events_unbound (work_completion)(&entry->work) &shost->scan_mutex running_helpers_waitq.lock irq_context: 0 (wq_completion)events_unbound (work_completion)(&entry->work) &shost->scan_mutex &dev->mutex device_links_srcu irq_context: 0 (wq_completion)events_unbound (work_completion)(&entry->work) &shost->scan_mutex &dev->mutex fwnode_link_lock irq_context: 0 (wq_completion)events_unbound (work_completion)(&entry->work) &shost->scan_mutex &dev->mutex device_links_lock irq_context: 0 (wq_completion)events_unbound (work_completion)(&entry->work) &shost->scan_mutex &dev->mutex sysfs_symlink_target_lock irq_context: 0 (wq_completion)events_unbound (work_completion)(&entry->work) &shost->scan_mutex &dev->mutex fs_reclaim irq_context: 0 (wq_completion)events_unbound (work_completion)(&entry->work) &shost->scan_mutex &dev->mutex fs_reclaim mmu_notifier_invalidate_range_start irq_context: 0 (wq_completion)events_unbound (work_completion)(&entry->work) &shost->scan_mutex &dev->mutex pool_lock#2 irq_context: 0 (wq_completion)events_unbound (work_completion)(&entry->work) &shost->scan_mutex &dev->mutex &c->lock irq_context: 0 (wq_completion)events_unbound (work_completion)(&entry->work) &shost->scan_mutex &dev->mutex &pcp->lock &zone->lock irq_context: 0 (wq_completion)events_unbound (work_completion)(&entry->work) &shost->scan_mutex &dev->mutex &zone->lock irq_context: 0 (wq_completion)events_unbound (work_completion)(&entry->work) &shost->scan_mutex &dev->mutex &____s->seqcount irq_context: 0 (wq_completion)events_unbound (work_completion)(&entry->work) &shost->scan_mutex &dev->mutex lock irq_context: 0 (wq_completion)events_unbound (work_completion)(&entry->work) &shost->scan_mutex &dev->mutex lock kernfs_idr_lock irq_context: 0 (wq_completion)events_unbound (work_completion)(&entry->work) &shost->scan_mutex &dev->mutex &root->kernfs_rwsem irq_context: 0 (wq_completion)events_unbound (work_completion)(&entry->work) &shost->scan_mutex &dev->mutex &root->kernfs_rwsem &root->kernfs_iattr_rwsem irq_context: 0 (wq_completion)events_unbound (work_completion)(&entry->work) &shost->scan_mutex &dev->mutex &root->kernfs_rwsem &obj_hash[i].lock irq_context: 0 (wq_completion)events_unbound (work_completion)(&entry->work) &shost->scan_mutex &dev->mutex &root->kernfs_rwsem pool_lock#2 irq_context: 0 (wq_completion)events_unbound (work_completion)(&entry->work) &shost->scan_mutex &dev->mutex &root->kernfs_rwsem kernfs_idr_lock irq_context: 0 (wq_completion)events_unbound (work_completion)(&entry->work) &shost->scan_mutex &dev->mutex probe_waitqueue.lock irq_context: 0 (wq_completion)events_unbound (work_completion)(&entry->work) &shost->scan_mutex async_lock irq_context: 0 (wq_completion)events_unbound (work_completion)(&entry->work) &shost->scan_mutex gdp_mutex irq_context: 0 (wq_completion)events_unbound (work_completion)(&entry->work) &shost->scan_mutex gdp_mutex &k->list_lock irq_context: 0 (wq_completion)events_unbound (work_completion)(&entry->work) &shost->scan_mutex gdp_mutex fs_reclaim irq_context: 0 (wq_completion)events_unbound (work_completion)(&entry->work) &shost->scan_mutex gdp_mutex fs_reclaim mmu_notifier_invalidate_range_start irq_context: 0 (wq_completion)events_unbound (work_completion)(&entry->work) &shost->scan_mutex gdp_mutex pool_lock#2 irq_context: 0 (wq_completion)events_unbound (work_completion)(&entry->work) &shost->scan_mutex gdp_mutex lock irq_context: 0 (wq_completion)events_unbound (work_completion)(&entry->work) &shost->scan_mutex gdp_mutex lock kernfs_idr_lock irq_context: 0 (wq_completion)events_unbound (work_completion)(&entry->work) &shost->scan_mutex gdp_mutex &root->kernfs_rwsem irq_context: 0 (wq_completion)events_unbound (work_completion)(&entry->work) &shost->scan_mutex gdp_mutex &root->kernfs_rwsem &root->kernfs_iattr_rwsem irq_context: 0 (wq_completion)events_unbound (work_completion)(&entry->work) &shost->scan_mutex subsys mutex#45 irq_context: 0 (wq_completion)events_unbound (work_completion)(&entry->work) &shost->scan_mutex subsys mutex#45 &k->k_lock irq_context: 0 (wq_completion)events_unbound (work_completion)(&entry->work) &shost->scan_mutex subsys mutex#45 fs_reclaim irq_context: 0 (wq_completion)events_unbound (work_completion)(&entry->work) &shost->scan_mutex subsys mutex#45 fs_reclaim mmu_notifier_invalidate_range_start irq_context: 0 (wq_completion)events_unbound (work_completion)(&entry->work) &shost->scan_mutex subsys mutex#45 pool_lock#2 irq_context: 0 (wq_completion)events_unbound (work_completion)(&entry->work) &shost->scan_mutex subsys mutex#45 lock irq_context: 0 (wq_completion)events_unbound (work_completion)(&entry->work) &shost->scan_mutex subsys mutex#45 lock sg_index_lock irq_context: 0 (wq_completion)events_unbound (work_completion)(&entry->work) &shost->scan_mutex subsys mutex#45 lock sg_index_lock pool_lock#2 irq_context: 0 (wq_completion)events_unbound (work_completion)(&entry->work) &shost->scan_mutex subsys mutex#45 chrdevs_lock irq_context: 0 (wq_completion)events_unbound (work_completion)(&entry->work) &shost->scan_mutex subsys mutex#45 &x->wait#9 irq_context: 0 (wq_completion)events_unbound (work_completion)(&entry->work) &shost->scan_mutex subsys mutex#45 &obj_hash[i].lock irq_context: 0 (wq_completion)events_unbound (work_completion)(&entry->work) &shost->scan_mutex subsys mutex#45 &k->list_lock irq_context: 0 (wq_completion)events_unbound (work_completion)(&entry->work) &shost->scan_mutex subsys mutex#45 gdp_mutex irq_context: 0 (wq_completion)events_unbound (work_completion)(&entry->work) &shost->scan_mutex subsys mutex#45 gdp_mutex &k->list_lock irq_context: 0 (wq_completion)events_unbound (work_completion)(&entry->work) &shost->scan_mutex subsys mutex#45 gdp_mutex fs_reclaim irq_context: 0 (wq_completion)events_unbound (work_completion)(&entry->work) &shost->scan_mutex subsys mutex#45 gdp_mutex fs_reclaim mmu_notifier_invalidate_range_start irq_context: 0 (wq_completion)events_unbound (work_completion)(&entry->work) &shost->scan_mutex subsys mutex#45 gdp_mutex &pcp->lock &zone->lock irq_context: 0 (wq_completion)events_unbound (work_completion)(&entry->work) &shost->scan_mutex subsys mutex#45 gdp_mutex &zone->lock irq_context: 0 (wq_completion)events_unbound (work_completion)(&entry->work) &shost->scan_mutex subsys mutex#45 gdp_mutex &____s->seqcount irq_context: 0 (wq_completion)events_unbound (work_completion)(&entry->work) &shost->scan_mutex subsys mutex#45 gdp_mutex pool_lock#2 irq_context: 0 (wq_completion)events_unbound (work_completion)(&entry->work) &shost->scan_mutex subsys mutex#45 gdp_mutex rcu_read_lock pool_lock#2 irq_context: 0 (wq_completion)events_unbound (work_completion)(&entry->work) &shost->scan_mutex subsys mutex#45 gdp_mutex &obj_hash[i].lock irq_context: 0 (wq_completion)events_unbound (work_completion)(&entry->work) &shost->scan_mutex subsys mutex#45 gdp_mutex lock irq_context: 0 (wq_completion)events_unbound (work_completion)(&entry->work) &shost->scan_mutex subsys mutex#45 gdp_mutex lock kernfs_idr_lock irq_context: 0 (wq_completion)events_unbound (work_completion)(&entry->work) &shost->scan_mutex subsys mutex#45 gdp_mutex &root->kernfs_rwsem irq_context: 0 (wq_completion)events_unbound (work_completion)(&entry->work) &shost->scan_mutex subsys mutex#45 gdp_mutex &root->kernfs_rwsem &root->kernfs_iattr_rwsem irq_context: 0 (wq_completion)events_unbound (work_completion)(&entry->work) &shost->scan_mutex subsys mutex#45 &c->lock irq_context: 0 (wq_completion)events_unbound (work_completion)(&entry->work) &shost->scan_mutex subsys mutex#45 &pcp->lock &zone->lock irq_context: 0 (wq_completion)events_unbound (work_completion)(&entry->work) &shost->scan_mutex subsys mutex#45 &zone->lock irq_context: 0 (wq_completion)events_unbound (work_completion)(&entry->work) &shost->scan_mutex subsys mutex#45 &____s->seqcount irq_context: 0 (wq_completion)events_unbound (work_completion)(&entry->work) &shost->scan_mutex subsys mutex#45 lock kernfs_idr_lock irq_context: 0 (wq_completion)events_unbound (work_completion)(&entry->work) &shost->scan_mutex subsys mutex#45 &root->kernfs_rwsem irq_context: 0 (wq_completion)events_unbound (work_completion)(&entry->work) &shost->scan_mutex subsys mutex#45 &root->kernfs_rwsem &root->kernfs_iattr_rwsem irq_context: 0 (wq_completion)events_unbound (work_completion)(&entry->work) &shost->scan_mutex subsys mutex#45 bus_type_sem irq_context: 0 (wq_completion)events_unbound (work_completion)(&entry->work) &shost->scan_mutex subsys mutex#45 sysfs_symlink_target_lock irq_context: 0 (wq_completion)events_unbound (work_completion)(&entry->work) &shost->scan_mutex subsys mutex#45 &root->kernfs_rwsem irq_context: 0 (wq_completion)events_unbound (work_completion)(&entry->work) &shost->scan_mutex subsys mutex#45 lock kernfs_idr_lock pool_lock#2 irq_context: 0 (wq_completion)events_unbound (work_completion)(&entry->work) &shost->scan_mutex subsys mutex#45 &dev->power.lock irq_context: 0 (wq_completion)events_unbound (work_completion)(&entry->work) &shost->scan_mutex subsys mutex#45 dpm_list_mtx irq_context: 0 (wq_completion)events_unbound (work_completion)(&entry->work) &shost->scan_mutex subsys mutex#45 req_lock irq_context: 0 (wq_completion)events_unbound (work_completion)(&entry->work) &shost->scan_mutex subsys mutex#45 &p->pi_lock irq_context: 0 (wq_completion)events_unbound (work_completion)(&entry->work) &shost->scan_mutex subsys mutex#45 &p->pi_lock &rq->__lock irq_context: 0 (wq_completion)events_unbound (work_completion)(&entry->work) &shost->scan_mutex subsys mutex#45 &p->pi_lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)events_unbound (work_completion)(&entry->work) &shost->scan_mutex subsys mutex#45 &x->wait#11 irq_context: 0 (wq_completion)events_unbound (work_completion)(&entry->work) &shost->scan_mutex subsys mutex#45 &rq->__lock irq_context: 0 (wq_completion)events_unbound (work_completion)(&entry->work) &shost->scan_mutex subsys mutex#45 &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)events_unbound (work_completion)(&entry->work) &dev->mutex bio_slab_lock irq_context: 0 (wq_completion)events_unbound (work_completion)(&entry->work) &dev->mutex &pcp->lock &zone->lock &____s->seqcount irq_context: 0 (wq_completion)events_unbound (work_completion)(&entry->work) &dev->mutex pcpu_alloc_mutex irq_context: 0 (wq_completion)events_unbound (work_completion)(&entry->work) &dev->mutex pcpu_alloc_mutex pcpu_lock irq_context: 0 (wq_completion)events_unbound (work_completion)(&entry->work) &dev->mutex percpu_counters_lock irq_context: 0 (wq_completion)events_unbound (work_completion)(&entry->work) &dev->mutex &obj_hash[i].lock pool_lock irq_context: 0 (wq_completion)events_unbound (work_completion)(&entry->work) &dev->mutex mmu_notifier_invalidate_range_start irq_context: 0 (wq_completion)events_unbound (work_completion)(&entry->work) &dev->mutex &sb->s_type->i_lock_key#3 irq_context: 0 (wq_completion)events_unbound (work_completion)(&entry->work) &dev->mutex &s->s_inode_list_lock irq_context: 0 (wq_completion)events_unbound (work_completion)(&entry->work) &dev->mutex &xa->xa_lock#11 irq_context: 0 (wq_completion)events_unbound (work_completion)(&entry->work) &dev->mutex lock &q->queue_lock irq_context: 0 (wq_completion)events_unbound (work_completion)(&entry->work) &dev->mutex lock &q->queue_lock &blkcg->lock irq_context: 0 (wq_completion)events_unbound (work_completion)(&entry->work) &dev->mutex &q->mq_freeze_lock irq_context: 0 (wq_completion)events_unbound (work_completion)(&entry->work) &dev->mutex &q->mq_freeze_lock percpu_ref_switch_lock irq_context: 0 (wq_completion)events_unbound (work_completion)(&entry->work) &dev->mutex &q->mq_freeze_lock percpu_ref_switch_lock rcu_read_lock &q->mq_freeze_wq irq_context: 0 (wq_completion)events_unbound (work_completion)(&entry->work) &dev->mutex percpu_ref_switch_lock irq_context: 0 (wq_completion)events_unbound (work_completion)(&entry->work) &dev->mutex &q->queue_lock irq_context: 0 (wq_completion)events_unbound (work_completion)(&entry->work) &dev->mutex &q->queue_lock pool_lock#2 irq_context: 0 (wq_completion)events_unbound (work_completion)(&entry->work) &dev->mutex &q->mq_freeze_lock &q->mq_freeze_wq irq_context: 0 (wq_completion)events_unbound (work_completion)(&entry->work) &dev->mutex &q->queue_lock &c->lock irq_context: 0 (wq_completion)events_unbound (work_completion)(&entry->work) &dev->mutex &q->queue_lock &pcp->lock &zone->lock irq_context: 0 (wq_completion)events_unbound (work_completion)(&entry->work) &dev->mutex &q->queue_lock &zone->lock irq_context: 0 (wq_completion)events_unbound (work_completion)(&entry->work) &dev->mutex &q->queue_lock &____s->seqcount irq_context: 0 (wq_completion)events_unbound (work_completion)(&entry->work) &dev->mutex &q->queue_lock pcpu_lock irq_context: 0 (wq_completion)events_unbound (work_completion)(&entry->work) &dev->mutex &q->queue_lock &obj_hash[i].lock irq_context: 0 (wq_completion)events_unbound (work_completion)(&entry->work) &dev->mutex &q->queue_lock percpu_counters_lock irq_context: 0 (wq_completion)events_unbound (work_completion)(&entry->work) &dev->mutex sd_index_ida.xa_lock irq_context: 0 (wq_completion)events_unbound (work_completion)(&entry->work) &dev->mutex rcu_read_lock &pool->lock/1 irq_context: 0 (wq_completion)events_unbound (work_completion)(&entry->work) &dev->mutex rcu_read_lock &pool->lock/1 &obj_hash[i].lock irq_context: 0 (wq_completion)events_unbound (work_completion)(&entry->work) &dev->mutex rcu_read_lock &pool->lock/1 pool_lock#2 irq_context: 0 (wq_completion)events_unbound (work_completion)(&entry->work) &dev->mutex rcu_read_lock &pool->lock/1 &p->pi_lock irq_context: 0 (wq_completion)events_unbound (work_completion)(&entry->work) &dev->mutex subsys mutex#56 irq_context: 0 (wq_completion)events_unbound (work_completion)(&entry->work) &dev->mutex subsys mutex#56 &k->k_lock irq_context: 0 (wq_completion)events_unbound (work_completion)(&entry->work) &dev->mutex tk_core.seq.seqcount irq_context: 0 (wq_completion)events_unbound (work_completion)(&entry->work) &dev->mutex &hctx->lock irq_context: 0 (wq_completion)events_unbound (work_completion)(&entry->work) &dev->mutex rcu_read_lock &hctx->lock irq_context: 0 (wq_completion)events_unbound (work_completion)(&entry->work) &dev->mutex rcu_read_lock &obj_hash[i].lock irq_context: 0 (wq_completion)events_unbound (work_completion)(&entry->work) &dev->mutex rcu_read_lock &virtscsi_vq->vq_lock irq_context: 0 (wq_completion)events_unbound (work_completion)(&entry->work) &dev->mutex &x->wait#16 irq_context: 0 (wq_completion)events_unbound (work_completion)(&entry->work) &dev->mutex &base->lock irq_context: 0 (wq_completion)events_unbound (work_completion)(&entry->work) &dev->mutex &base->lock &obj_hash[i].lock irq_context: 0 (wq_completion)events_unbound (work_completion)(&entry->work) &dev->mutex &rq->__lock irq_context: 0 (wq_completion)events_unbound (work_completion)(&entry->work) &dev->mutex &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)events_unbound (work_completion)(&entry->work) &dev->mutex (&timer.timer) irq_context: 0 (wq_completion)events_unbound (work_completion)(&entry->work) &dev->mutex (console_sem).lock irq_context: 0 (wq_completion)events_unbound (work_completion)(&entry->work) &dev->mutex console_lock console_srcu console_owner_lock irq_context: 0 (wq_completion)events_unbound (work_completion)(&entry->work) &dev->mutex console_lock console_srcu console_owner irq_context: 0 (wq_completion)events_unbound (work_completion)(&entry->work) &dev->mutex console_lock console_srcu console_owner &port_lock_key irq_context: 0 (wq_completion)events_unbound (work_completion)(&entry->work) &dev->mutex console_lock console_srcu console_owner console_owner_lock irq_context: 0 console_owner_lock irq_context: 0 console_owner irq_context: 0 (wq_completion)events_unbound (work_completion)(&entry->work) &shost->scan_mutex subsys mutex#45 uevent_sock_mutex irq_context: 0 (wq_completion)events_unbound (work_completion)(&entry->work) &shost->scan_mutex subsys mutex#45 rcu_read_lock &pool->lock/1 irq_context: 0 (wq_completion)events_unbound (work_completion)(&entry->work) &shost->scan_mutex subsys mutex#45 rcu_read_lock &pool->lock/1 &obj_hash[i].lock irq_context: 0 (wq_completion)events_unbound (work_completion)(&entry->work) &shost->scan_mutex subsys mutex#45 rcu_read_lock &pool->lock/1 pool_lock#2 irq_context: 0 (wq_completion)events_unbound (work_completion)(&entry->work) &shost->scan_mutex subsys mutex#45 rcu_read_lock &pool->lock/1 &p->pi_lock irq_context: 0 (wq_completion)events_unbound (work_completion)(&entry->work) &shost->scan_mutex subsys mutex#45 running_helpers_waitq.lock irq_context: 0 (wq_completion)events_unbound (work_completion)(&entry->work) &shost->scan_mutex subsys mutex#45 subsys mutex#57 irq_context: 0 (wq_completion)events_unbound (work_completion)(&entry->work) &shost->scan_mutex subsys mutex#45 subsys mutex#57 &k->k_lock irq_context: 0 (wq_completion)events_unbound (work_completion)(&entry->work) &shost->scan_mutex subsys mutex#45 (console_sem).lock irq_context: 0 (wq_completion)events_unbound (work_completion)(&entry->work) &shost->scan_mutex subsys mutex#45 console_owner_lock irq_context: 0 (wq_completion)events_unbound (work_completion)(&entry->work) &shost->scan_mutex subsys mutex#45 console_owner irq_context: 0 (wq_completion)events_unbound (work_completion)(&entry->work) &shost->scan_mutex subsys mutex#45 console_lock console_srcu console_owner_lock irq_context: 0 (wq_completion)events_unbound (work_completion)(&entry->work) &shost->scan_mutex subsys mutex#45 console_lock console_srcu console_owner irq_context: 0 (wq_completion)events_unbound (work_completion)(&entry->work) &shost->scan_mutex subsys mutex#45 console_lock console_srcu console_owner &port_lock_key irq_context: 0 (wq_completion)events_unbound (work_completion)(&entry->work) &shost->scan_mutex subsys mutex#45 console_lock console_srcu console_owner console_owner_lock irq_context: 0 (wq_completion)events_unbound (work_completion)(&entry->work) &shost->scan_mutex bsg_minor_ida.xa_lock irq_context: 0 (wq_completion)events_unbound (work_completion)(&entry->work) &shost->scan_mutex chrdevs_lock irq_context: 0 (wq_completion)events_unbound (work_completion)(&entry->work) &shost->scan_mutex req_lock irq_context: 0 (wq_completion)events_unbound (work_completion)(&entry->work) &shost->scan_mutex &p->pi_lock irq_context: 0 (wq_completion)events_unbound (work_completion)(&entry->work) &shost->scan_mutex &p->pi_lock &rq->__lock irq_context: 0 (wq_completion)events_unbound (work_completion)(&entry->work) &shost->scan_mutex &p->pi_lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)events_unbound (work_completion)(&entry->work) &shost->scan_mutex &p->pi_lock &rq->__lock &cfs_rq->removed.lock irq_context: 0 (wq_completion)events_unbound (work_completion)(&entry->work) &shost->scan_mutex &x->wait#11 irq_context: 0 (wq_completion)events_unbound (work_completion)(&entry->work) &shost->scan_mutex subsys mutex#58 irq_context: 0 (wq_completion)events_unbound (work_completion)(&entry->work) &shost->scan_mutex subsys mutex#58 &k->k_lock irq_context: 0 (wq_completion)events_unbound (work_completion)(&entry->work) async_scan_lock irq_context: 0 (wq_completion)events_unbound (work_completion)(&entry->work) &dev->power.lock irq_context: 0 (wq_completion)events_unbound (work_completion)(&entry->work) &dev->mutex &bdev->bd_size_lock irq_context: 0 (wq_completion)events_unbound (work_completion)(&entry->work) &dev->mutex elv_list_lock irq_context: 0 (wq_completion)events_unbound (work_completion)(&entry->work) &dev->mutex (work_completion)(&(&q->requeue_work)->work) irq_context: 0 (wq_completion)events_unbound (work_completion)(&entry->work) &dev->mutex (work_completion)(&(&hctx->run_work)->work) irq_context: 0 (wq_completion)events_unbound (work_completion)(&entry->work) &dev->mutex batched_entropy_u32.lock irq_context: 0 (wq_completion)events_unbound (work_completion)(&entry->work) &dev->mutex &zone->lock &____s->seqcount irq_context: 0 (wq_completion)events_unbound (work_completion)(&entry->work) &dev->mutex &q->debugfs_mutex irq_context: 0 (wq_completion)events_unbound (work_completion)(&entry->work) &dev->mutex req_lock irq_context: 0 (wq_completion)events_unbound (work_completion)(&entry->work) &dev->mutex &p->pi_lock irq_context: 0 (wq_completion)events_unbound (work_completion)(&entry->work) &dev->mutex &p->pi_lock &rq->__lock irq_context: 0 (wq_completion)events_unbound (work_completion)(&entry->work) &dev->mutex &p->pi_lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)events_unbound (work_completion)(&entry->work) &dev->mutex &p->pi_lock &rq->__lock &cfs_rq->removed.lock irq_context: 0 (wq_completion)events_unbound (work_completion)(&entry->work) &dev->mutex &x->wait#11 irq_context: 0 (wq_completion)events_unbound (work_completion)(&entry->work) &dev->mutex subsys mutex#37 irq_context: 0 (wq_completion)events_unbound (work_completion)(&entry->work) &dev->mutex subsys mutex#37 &k->k_lock irq_context: 0 (wq_completion)events_unbound (work_completion)(&entry->work) &dev->mutex dev_hotplug_mutex irq_context: 0 (wq_completion)events_unbound (work_completion)(&entry->work) &dev->mutex dev_hotplug_mutex &dev->power.lock irq_context: 0 (wq_completion)events_unbound (work_completion)(&entry->work) &dev->mutex &q->sysfs_dir_lock irq_context: 0 (wq_completion)events_unbound (work_completion)(&entry->work) &dev->mutex &q->sysfs_dir_lock fs_reclaim irq_context: 0 (wq_completion)events_unbound (work_completion)(&entry->work) &dev->mutex &q->sysfs_dir_lock fs_reclaim mmu_notifier_invalidate_range_start irq_context: 0 (wq_completion)events_unbound (work_completion)(&entry->work) &dev->mutex &q->sysfs_dir_lock pool_lock#2 irq_context: 0 (wq_completion)events_unbound (work_completion)(&entry->work) &dev->mutex &q->sysfs_dir_lock lock irq_context: 0 (wq_completion)events_unbound (work_completion)(&entry->work) &dev->mutex &q->sysfs_dir_lock lock kernfs_idr_lock irq_context: 0 (wq_completion)events_unbound (work_completion)(&entry->work) &dev->mutex &q->sysfs_dir_lock &root->kernfs_rwsem irq_context: 0 (wq_completion)events_unbound (work_completion)(&entry->work) &dev->mutex &q->sysfs_dir_lock &root->kernfs_rwsem &root->kernfs_iattr_rwsem irq_context: 0 (wq_completion)events_unbound (work_completion)(&entry->work) &dev->mutex &q->sysfs_dir_lock lock kernfs_idr_lock pool_lock#2 irq_context: 0 (wq_completion)events_unbound (work_completion)(&entry->work) &dev->mutex &q->sysfs_dir_lock &c->lock irq_context: 0 (wq_completion)events_unbound (work_completion)(&entry->work) &dev->mutex &q->sysfs_dir_lock &pcp->lock &zone->lock irq_context: 0 (wq_completion)events_unbound (work_completion)(&entry->work) &dev->mutex &q->sysfs_dir_lock &zone->lock irq_context: 0 (wq_completion)events_unbound (work_completion)(&entry->work) &dev->mutex &q->sysfs_dir_lock &____s->seqcount irq_context: 0 (wq_completion)events_unbound (work_completion)(&entry->work) &dev->mutex &q->sysfs_dir_lock rcu_read_lock pool_lock#2 irq_context: 0 (wq_completion)events_unbound (work_completion)(&entry->work) &dev->mutex &q->sysfs_dir_lock &obj_hash[i].lock irq_context: 0 (wq_completion)events_unbound (work_completion)(&entry->work) &dev->mutex &q->sysfs_dir_lock &q->sysfs_lock irq_context: 0 (wq_completion)events_unbound (work_completion)(&entry->work) &dev->mutex &q->sysfs_dir_lock &q->sysfs_lock &q->debugfs_mutex irq_context: 0 (wq_completion)events_unbound (work_completion)(&entry->work) &dev->mutex &q->sysfs_dir_lock &q->sysfs_lock &q->debugfs_mutex pin_fs_lock irq_context: 0 (wq_completion)events_unbound (work_completion)(&entry->work) &dev->mutex &q->sysfs_dir_lock &q->sysfs_lock &q->debugfs_mutex &sb->s_type->i_mutex_key#3 irq_context: 0 (wq_completion)events_unbound (work_completion)(&entry->work) &dev->mutex &q->sysfs_dir_lock &q->sysfs_lock &q->debugfs_mutex &sb->s_type->i_mutex_key#3 rename_lock.seqcount irq_context: 0 (wq_completion)events_unbound (work_completion)(&entry->work) &dev->mutex &q->sysfs_dir_lock &q->sysfs_lock &q->debugfs_mutex &sb->s_type->i_mutex_key#3 fs_reclaim irq_context: 0 (wq_completion)events_unbound (work_completion)(&entry->work) &dev->mutex &q->sysfs_dir_lock &q->sysfs_lock &q->debugfs_mutex &sb->s_type->i_mutex_key#3 fs_reclaim mmu_notifier_invalidate_range_start irq_context: 0 (wq_completion)events_unbound (work_completion)(&entry->work) &dev->mutex &q->sysfs_dir_lock &q->sysfs_lock &q->debugfs_mutex &sb->s_type->i_mutex_key#3 pool_lock#2 irq_context: 0 (wq_completion)events_unbound (work_completion)(&entry->work) &dev->mutex &q->sysfs_dir_lock &q->sysfs_lock &q->debugfs_mutex &sb->s_type->i_mutex_key#3 &dentry->d_lock irq_context: 0 (wq_completion)events_unbound (work_completion)(&entry->work) &dev->mutex &q->sysfs_dir_lock &q->sysfs_lock &q->debugfs_mutex &sb->s_type->i_mutex_key#3 rcu_read_lock rename_lock.seqcount irq_context: 0 (wq_completion)events_unbound (work_completion)(&entry->work) &dev->mutex &q->sysfs_dir_lock &q->sysfs_lock &q->debugfs_mutex &sb->s_type->i_mutex_key#3 &dentry->d_lock &wq irq_context: 0 (wq_completion)events_unbound (work_completion)(&entry->work) &dev->mutex &q->sysfs_dir_lock &q->sysfs_lock &q->debugfs_mutex &sb->s_type->i_mutex_key#3 mmu_notifier_invalidate_range_start irq_context: 0 (wq_completion)events_unbound (work_completion)(&entry->work) &dev->mutex &q->sysfs_dir_lock &q->sysfs_lock &q->debugfs_mutex &sb->s_type->i_mutex_key#3 &sb->s_type->i_lock_key#7 irq_context: 0 (wq_completion)events_unbound (work_completion)(&entry->work) &dev->mutex &q->sysfs_dir_lock &q->sysfs_lock &q->debugfs_mutex &sb->s_type->i_mutex_key#3 &s->s_inode_list_lock irq_context: 0 (wq_completion)events_unbound (work_completion)(&entry->work) &dev->mutex &q->sysfs_dir_lock &q->sysfs_lock &q->debugfs_mutex &sb->s_type->i_mutex_key#3 tk_core.seq.seqcount irq_context: 0 (wq_completion)events_unbound (work_completion)(&entry->work) &dev->mutex &q->sysfs_dir_lock &q->sysfs_lock &q->debugfs_mutex &sb->s_type->i_mutex_key#3 &sb->s_type->i_lock_key#7 &dentry->d_lock irq_context: 0 (wq_completion)events_unbound (work_completion)(&entry->work) &dev->mutex &q->sysfs_dir_lock &q->sysfs_lock &q->debugfs_mutex &sb->s_type->i_mutex_key#3 &c->lock irq_context: 0 (wq_completion)events_unbound (work_completion)(&entry->work) &dev->mutex &q->sysfs_dir_lock &q->sysfs_lock &q->debugfs_mutex &sb->s_type->i_mutex_key#3 &pcp->lock &zone->lock irq_context: 0 (wq_completion)events_unbound (work_completion)(&entry->work) &dev->mutex &q->sysfs_dir_lock &q->sysfs_lock &q->debugfs_mutex &sb->s_type->i_mutex_key#3 &zone->lock irq_context: 0 (wq_completion)events_unbound (work_completion)(&entry->work) &dev->mutex &q->sysfs_dir_lock &q->sysfs_lock &q->debugfs_mutex &sb->s_type->i_mutex_key#3 &____s->seqcount irq_context: 0 (wq_completion)events_unbound (work_completion)(&entry->work) &dev->mutex &q->sysfs_dir_lock &q->sysfs_lock &q->debugfs_mutex &sb->s_type->i_mutex_key#3 &pcp->lock &zone->lock &____s->seqcount irq_context: 0 (wq_completion)events_unbound (work_completion)(&entry->work) &dev->mutex &q->sysfs_dir_lock &q->sysfs_lock fs_reclaim irq_context: 0 (wq_completion)events_unbound (work_completion)(&entry->work) &dev->mutex &q->sysfs_dir_lock &q->sysfs_lock fs_reclaim mmu_notifier_invalidate_range_start irq_context: 0 (wq_completion)events_unbound (work_completion)(&entry->work) &dev->mutex &q->sysfs_dir_lock &q->sysfs_lock pool_lock#2 irq_context: 0 (wq_completion)events_unbound (work_completion)(&entry->work) &dev->mutex &q->sysfs_dir_lock &q->sysfs_lock lock irq_context: 0 (wq_completion)events_unbound (work_completion)(&entry->work) &dev->mutex &q->sysfs_dir_lock &q->sysfs_lock lock kernfs_idr_lock irq_context: 0 (wq_completion)events_unbound (work_completion)(&entry->work) &dev->mutex &q->sysfs_dir_lock &q->sysfs_lock &root->kernfs_rwsem irq_context: 0 (wq_completion)events_unbound (work_completion)(&entry->work) &dev->mutex &q->sysfs_dir_lock &q->sysfs_lock &root->kernfs_rwsem &root->kernfs_iattr_rwsem irq_context: 0 (wq_completion)events_unbound (work_completion)(&entry->work) &dev->mutex &q->sysfs_dir_lock &q->sysfs_lock &c->lock irq_context: 0 (wq_completion)events_unbound (work_completion)(&entry->work) &dev->mutex &q->sysfs_dir_lock &q->sysfs_lock &pcp->lock &zone->lock irq_context: 0 (wq_completion)events_unbound (work_completion)(&entry->work) &dev->mutex &q->sysfs_dir_lock &q->sysfs_lock &zone->lock irq_context: 0 (wq_completion)events_unbound (work_completion)(&entry->work) &dev->mutex &q->sysfs_dir_lock &q->sysfs_lock &____s->seqcount irq_context: 0 (wq_completion)events_unbound (work_completion)(&entry->work) &dev->mutex &q->sysfs_dir_lock &q->sysfs_lock pcpu_alloc_mutex irq_context: 0 (wq_completion)events_unbound (work_completion)(&entry->work) &dev->mutex &q->sysfs_dir_lock &q->sysfs_lock pcpu_alloc_mutex pcpu_lock irq_context: 0 (wq_completion)events_unbound (work_completion)(&entry->work) &dev->mutex &q->sysfs_dir_lock &q->sysfs_lock &obj_hash[i].lock irq_context: 0 (wq_completion)events_unbound (work_completion)(&entry->work) &dev->mutex &q->sysfs_dir_lock &q->sysfs_lock &q->rq_qos_mutex irq_context: 0 (wq_completion)events_unbound (work_completion)(&entry->work) &dev->mutex &q->sysfs_dir_lock &q->sysfs_lock &q->rq_qos_mutex &q->mq_freeze_lock irq_context: 0 (wq_completion)events_unbound (work_completion)(&entry->work) &dev->mutex &q->sysfs_dir_lock &q->sysfs_lock &q->rq_qos_mutex &q->mq_freeze_lock percpu_ref_switch_lock irq_context: 0 (wq_completion)events_unbound (work_completion)(&entry->work) &dev->mutex &q->sysfs_dir_lock &q->sysfs_lock &q->rq_qos_mutex &q->mq_freeze_lock percpu_ref_switch_lock rcu_read_lock &q->mq_freeze_wq irq_context: 0 (wq_completion)events_unbound (work_completion)(&entry->work) &dev->mutex &q->sysfs_dir_lock &q->sysfs_lock &q->rq_qos_mutex percpu_ref_switch_lock irq_context: 0 (wq_completion)events_unbound (work_completion)(&entry->work) &dev->mutex &q->sysfs_dir_lock &q->sysfs_lock &q->rq_qos_mutex &q->mq_freeze_lock &q->mq_freeze_wq irq_context: 0 (wq_completion)events_unbound (work_completion)(&entry->work) &dev->mutex &q->sysfs_dir_lock &q->sysfs_lock &q->rq_qos_mutex &q->debugfs_mutex irq_context: 0 (wq_completion)events_unbound (work_completion)(&entry->work) &dev->mutex &q->sysfs_dir_lock &q->sysfs_lock &q->rq_qos_mutex &q->debugfs_mutex pin_fs_lock irq_context: 0 (wq_completion)events_unbound (work_completion)(&entry->work) &dev->mutex &q->sysfs_dir_lock &q->sysfs_lock &q->rq_qos_mutex &q->debugfs_mutex &sb->s_type->i_mutex_key#3 irq_context: 0 (wq_completion)events_unbound (work_completion)(&entry->work) &dev->mutex &q->sysfs_dir_lock &q->sysfs_lock &q->rq_qos_mutex &q->debugfs_mutex &sb->s_type->i_mutex_key#3 rename_lock.seqcount irq_context: 0 (wq_completion)events_unbound (work_completion)(&entry->work) &dev->mutex &q->sysfs_dir_lock &q->sysfs_lock &q->rq_qos_mutex &q->debugfs_mutex &sb->s_type->i_mutex_key#3 fs_reclaim irq_context: 0 (wq_completion)events_unbound (work_completion)(&entry->work) &dev->mutex &q->sysfs_dir_lock &q->sysfs_lock &q->rq_qos_mutex &q->debugfs_mutex &sb->s_type->i_mutex_key#3 fs_reclaim mmu_notifier_invalidate_range_start irq_context: 0 (wq_completion)events_unbound (work_completion)(&entry->work) &dev->mutex &q->sysfs_dir_lock &q->sysfs_lock &q->rq_qos_mutex &q->debugfs_mutex &sb->s_type->i_mutex_key#3 pool_lock#2 irq_context: 0 (wq_completion)events_unbound (work_completion)(&entry->work) &dev->mutex &q->sysfs_dir_lock &q->sysfs_lock &q->rq_qos_mutex &q->debugfs_mutex &sb->s_type->i_mutex_key#3 &dentry->d_lock irq_context: 0 (wq_completion)events_unbound (work_completion)(&entry->work) &dev->mutex &q->sysfs_dir_lock &q->sysfs_lock &q->rq_qos_mutex &q->debugfs_mutex &sb->s_type->i_mutex_key#3 rcu_read_lock rename_lock.seqcount irq_context: 0 (wq_completion)events_unbound (work_completion)(&entry->work) &dev->mutex &q->sysfs_dir_lock &q->sysfs_lock &q->rq_qos_mutex &q->debugfs_mutex &sb->s_type->i_mutex_key#3 &dentry->d_lock &wq irq_context: 0 (wq_completion)events_unbound (work_completion)(&entry->work) &dev->mutex &q->sysfs_dir_lock &q->sysfs_lock &q->rq_qos_mutex &q->debugfs_mutex &sb->s_type->i_mutex_key#3 mmu_notifier_invalidate_range_start irq_context: 0 (wq_completion)events_unbound (work_completion)(&entry->work) &dev->mutex &q->sysfs_dir_lock &q->sysfs_lock &q->rq_qos_mutex &q->debugfs_mutex &sb->s_type->i_mutex_key#3 &sb->s_type->i_lock_key#7 irq_context: 0 (wq_completion)events_unbound (work_completion)(&entry->work) &dev->mutex &q->sysfs_dir_lock &q->sysfs_lock &q->rq_qos_mutex &q->debugfs_mutex &sb->s_type->i_mutex_key#3 &s->s_inode_list_lock irq_context: 0 (wq_completion)events_unbound (work_completion)(&entry->work) &dev->mutex &q->sysfs_dir_lock &q->sysfs_lock &q->rq_qos_mutex &q->debugfs_mutex &sb->s_type->i_mutex_key#3 tk_core.seq.seqcount irq_context: 0 (wq_completion)events_unbound (work_completion)(&entry->work) &dev->mutex &q->sysfs_dir_lock &q->sysfs_lock &q->rq_qos_mutex &q->debugfs_mutex &sb->s_type->i_mutex_key#3 &sb->s_type->i_lock_key#7 &dentry->d_lock irq_context: 0 (wq_completion)events_unbound (work_completion)(&entry->work) &dev->mutex &q->sysfs_dir_lock &q->sysfs_lock &q->rq_qos_mutex &q->debugfs_mutex &sb->s_type->i_mutex_key#3 &c->lock irq_context: 0 (wq_completion)events_unbound (work_completion)(&entry->work) &dev->mutex &q->sysfs_dir_lock &q->sysfs_lock &q->rq_qos_mutex &q->debugfs_mutex &sb->s_type->i_mutex_key#3 &pcp->lock &zone->lock irq_context: 0 (wq_completion)events_unbound (work_completion)(&entry->work) &dev->mutex &q->sysfs_dir_lock &q->sysfs_lock &q->rq_qos_mutex &q->debugfs_mutex &sb->s_type->i_mutex_key#3 &zone->lock irq_context: 0 (wq_completion)events_unbound (work_completion)(&entry->work) &dev->mutex &q->sysfs_dir_lock &q->sysfs_lock &q->rq_qos_mutex &q->debugfs_mutex &sb->s_type->i_mutex_key#3 &____s->seqcount irq_context: 0 (wq_completion)events_unbound (work_completion)(&entry->work) &dev->mutex &q->sysfs_dir_lock &q->sysfs_lock &stats->lock irq_context: 0 (wq_completion)events_unbound (work_completion)(&entry->work) &dev->mutex subsys mutex#38 irq_context: 0 (wq_completion)events_unbound (work_completion)(&entry->work) &dev->mutex subsys mutex#38 &k->k_lock irq_context: 0 (wq_completion)events_unbound (work_completion)(&entry->work) &dev->mutex cgwb_lock irq_context: 0 (wq_completion)events_unbound (work_completion)(&entry->work) &dev->mutex &sb->s_type->i_mutex_key#3 &pcp->lock &zone->lock irq_context: 0 (wq_completion)events_unbound (work_completion)(&entry->work) &dev->mutex &sb->s_type->i_mutex_key#3 &zone->lock irq_context: 0 (wq_completion)events_unbound (work_completion)(&entry->work) &dev->mutex &sb->s_type->i_mutex_key#3 &____s->seqcount irq_context: 0 (wq_completion)events_unbound (work_completion)(&entry->work) &dev->mutex &sb->s_type->i_mutex_key#3 rcu_read_lock pool_lock#2 irq_context: 0 (wq_completion)events_unbound (work_completion)(&entry->work) &dev->mutex &sb->s_type->i_mutex_key#3 &obj_hash[i].lock irq_context: 0 (wq_completion)events_unbound (work_completion)(&entry->work) &dev->mutex bdi_lock irq_context: 0 (wq_completion)events_unbound (work_completion)(&entry->work) &dev->mutex inode_hash_lock irq_context: 0 (wq_completion)events_unbound (work_completion)(&entry->work) &dev->mutex inode_hash_lock &sb->s_type->i_lock_key#3 irq_context: 0 (wq_completion)events_unbound (work_completion)(&entry->work) &dev->mutex bdev_lock irq_context: 0 (wq_completion)events_unbound (work_completion)(&entry->work) &dev->mutex &disk->open_mutex irq_context: 0 (wq_completion)events_unbound (work_completion)(&entry->work) &dev->mutex &disk->open_mutex fs_reclaim irq_context: 0 (wq_completion)events_unbound (work_completion)(&entry->work) &dev->mutex &disk->open_mutex fs_reclaim mmu_notifier_invalidate_range_start irq_context: 0 (wq_completion)events_unbound (work_completion)(&entry->work) &dev->mutex &disk->open_mutex pool_lock#2 irq_context: 0 (wq_completion)events_unbound (work_completion)(&entry->work) &dev->mutex &disk->open_mutex tk_core.seq.seqcount irq_context: 0 (wq_completion)events_unbound (work_completion)(&entry->work) &dev->mutex &disk->open_mutex &obj_hash[i].lock irq_context: 0 (wq_completion)events_unbound (work_completion)(&entry->work) &dev->mutex &disk->open_mutex &hctx->lock irq_context: 0 (wq_completion)events_unbound (work_completion)(&entry->work) &dev->mutex &disk->open_mutex rcu_read_lock &hctx->lock irq_context: 0 (wq_completion)events_unbound (work_completion)(&entry->work) &dev->mutex &disk->open_mutex rcu_read_lock &obj_hash[i].lock irq_context: 0 (wq_completion)events_unbound (work_completion)(&entry->work) &dev->mutex &disk->open_mutex rcu_read_lock tk_core.seq.seqcount irq_context: 0 (wq_completion)events_unbound (work_completion)(&entry->work) &dev->mutex &disk->open_mutex rcu_read_lock &virtscsi_vq->vq_lock irq_context: 0 (wq_completion)events_unbound (work_completion)(&entry->work) &dev->mutex &disk->open_mutex &x->wait#16 irq_context: 0 (wq_completion)events_unbound (work_completion)(&entry->work) &dev->mutex &disk->open_mutex &base->lock irq_context: 0 (wq_completion)events_unbound (work_completion)(&entry->work) &dev->mutex &disk->open_mutex &base->lock &obj_hash[i].lock irq_context: 0 (wq_completion)events_unbound (work_completion)(&entry->work) &dev->mutex &disk->open_mutex &rq->__lock irq_context: 0 (wq_completion)events_unbound (work_completion)(&entry->work) &dev->mutex &disk->open_mutex &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)events_unbound (work_completion)(&entry->work) &dev->mutex &disk->open_mutex (&timer.timer) irq_context: 0 (wq_completion)events_unbound (work_completion)(&entry->work) &dev->mutex &disk->open_mutex mmu_notifier_invalidate_range_start irq_context: 0 (wq_completion)events_unbound (work_completion)(&entry->work) &dev->mutex &disk->open_mutex &pcp->lock &zone->lock irq_context: 0 (wq_completion)events_unbound (work_completion)(&entry->work) &dev->mutex &disk->open_mutex &zone->lock irq_context: 0 (wq_completion)events_unbound (work_completion)(&entry->work) &dev->mutex &disk->open_mutex &____s->seqcount irq_context: 0 (wq_completion)events_unbound (work_completion)(&entry->work) &dev->mutex &disk->open_mutex &q->sysfs_dir_lock irq_context: 0 (wq_completion)events_unbound (work_completion)(&entry->work) &dev->mutex &disk->open_mutex &q->sysfs_dir_lock &q->sysfs_lock irq_context: 0 (wq_completion)events_unbound (work_completion)(&entry->work) &dev->mutex &disk->open_mutex rcu_read_lock pool_lock#2 irq_context: 0 (wq_completion)events_unbound (work_completion)(&entry->work) &dev->mutex &disk->open_mutex &bdev->bd_size_lock irq_context: 0 (wq_completion)events_unbound (work_completion)(&entry->work) &dev->mutex &disk->open_mutex free_vmap_area_lock irq_context: 0 (wq_completion)events_unbound (work_completion)(&entry->work) &dev->mutex &disk->open_mutex vmap_area_lock irq_context: 0 (wq_completion)events_unbound (work_completion)(&entry->work) &dev->mutex &disk->open_mutex init_mm.page_table_lock irq_context: 0 (wq_completion)events_unbound (work_completion)(&entry->work) &dev->mutex &disk->open_mutex &xa->xa_lock#9 irq_context: 0 (wq_completion)events_unbound (work_completion)(&entry->work) &dev->mutex &disk->open_mutex lock#4 irq_context: 0 (wq_completion)events_unbound (work_completion)(&entry->work) &dev->mutex &disk->open_mutex &c->lock irq_context: 0 (wq_completion)events_unbound (work_completion)(&entry->work) &dev->mutex &disk->open_mutex &mapping->private_lock irq_context: 0 (wq_completion)events_unbound (work_completion)(&entry->work) &dev->mutex &disk->open_mutex &dd->lock irq_context: 0 (wq_completion)events_unbound (work_completion)(&entry->work) &dev->mutex &disk->open_mutex rcu_read_lock &pool->lock irq_context: 0 (wq_completion)events_unbound (work_completion)(&entry->work) &dev->mutex &disk->open_mutex rcu_read_lock &pool->lock &obj_hash[i].lock irq_context: 0 (wq_completion)events_unbound (work_completion)(&entry->work) &dev->mutex &disk->open_mutex rcu_read_lock &pool->lock pool_lock#2 irq_context: 0 (wq_completion)events_unbound (work_completion)(&entry->work) &dev->mutex &disk->open_mutex rcu_read_lock &pool->lock &p->pi_lock irq_context: 0 (wq_completion)events_unbound (work_completion)(&entry->work) &dev->mutex &disk->open_mutex rcu_read_lock &pool->lock &p->pi_lock &rq->__lock irq_context: 0 (wq_completion)events_unbound (work_completion)(&entry->work) &dev->mutex &disk->open_mutex rcu_read_lock &pool->lock &p->pi_lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)events_unbound (work_completion)(&entry->work) &dev->mutex &disk->open_mutex &folio_wait_table[i] irq_context: 0 (wq_completion)kblockd irq_context: 0 (wq_completion)kblockd (work_completion)(&(&hctx->run_work)->work) irq_context: 0 (wq_completion)kblockd (work_completion)(&(&hctx->run_work)->work) rcu_read_lock &dd->lock irq_context: 0 (wq_completion)kblockd (work_completion)(&(&hctx->run_work)->work) rcu_read_lock &obj_hash[i].lock irq_context: 0 (wq_completion)kblockd (work_completion)(&(&hctx->run_work)->work) rcu_read_lock tk_core.seq.seqcount irq_context: 0 (wq_completion)kblockd (work_completion)(&(&hctx->run_work)->work) rcu_read_lock &virtscsi_vq->vq_lock irq_context: softirq &ret->b_uptodate_lock irq_context: softirq &folio_wait_table[i] irq_context: softirq &folio_wait_table[i] &p->pi_lock irq_context: softirq &folio_wait_table[i] &p->pi_lock &rq->__lock irq_context: softirq &folio_wait_table[i] &p->pi_lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)events_unbound (work_completion)(&entry->work) &dev->mutex &disk->open_mutex (console_sem).lock irq_context: 0 (wq_completion)events_unbound (work_completion)(&entry->work) &dev->mutex &disk->open_mutex console_owner_lock irq_context: 0 (wq_completion)events_unbound (work_completion)(&entry->work) &dev->mutex &disk->open_mutex console_owner irq_context: 0 (wq_completion)events_unbound (work_completion)(&entry->work) &dev->mutex &disk->open_mutex console_lock console_srcu console_owner_lock irq_context: 0 (wq_completion)events_unbound (work_completion)(&entry->work) &dev->mutex &disk->open_mutex console_lock console_srcu console_owner irq_context: 0 (wq_completion)events_unbound (work_completion)(&entry->work) &dev->mutex &disk->open_mutex console_lock console_srcu console_owner &port_lock_key irq_context: 0 (wq_completion)events_unbound (work_completion)(&entry->work) &dev->mutex &disk->open_mutex console_lock console_srcu console_owner console_owner_lock irq_context: 0 (wq_completion)events_unbound (work_completion)(&entry->work) &dev->mutex &disk->open_mutex &sb->s_type->i_lock_key#3 irq_context: 0 (wq_completion)events_unbound (work_completion)(&entry->work) &dev->mutex &disk->open_mutex &s->s_inode_list_lock irq_context: 0 (wq_completion)events_unbound (work_completion)(&entry->work) &dev->mutex &disk->open_mutex pcpu_alloc_mutex irq_context: 0 (wq_completion)events_unbound (work_completion)(&entry->work) &dev->mutex &disk->open_mutex pcpu_alloc_mutex pcpu_lock irq_context: 0 (wq_completion)events_unbound (work_completion)(&entry->work) &dev->mutex &disk->open_mutex &x->wait#9 irq_context: 0 (wq_completion)events_unbound (work_completion)(&entry->work) &dev->mutex &disk->open_mutex &k->list_lock irq_context: 0 (wq_completion)events_unbound (work_completion)(&entry->work) &dev->mutex &disk->open_mutex lock irq_context: 0 (wq_completion)events_unbound (work_completion)(&entry->work) &dev->mutex &disk->open_mutex lock kernfs_idr_lock irq_context: 0 (wq_completion)events_unbound (work_completion)(&entry->work) &dev->mutex &disk->open_mutex &root->kernfs_rwsem irq_context: 0 (wq_completion)events_unbound (work_completion)(&entry->work) &dev->mutex &disk->open_mutex &root->kernfs_rwsem &root->kernfs_iattr_rwsem irq_context: 0 (wq_completion)events_unbound (work_completion)(&entry->work) &dev->mutex &disk->open_mutex bus_type_sem irq_context: 0 (wq_completion)events_unbound (work_completion)(&entry->work) &dev->mutex &disk->open_mutex sysfs_symlink_target_lock irq_context: 0 (wq_completion)events_unbound (work_completion)(&entry->work) &dev->mutex &disk->open_mutex &root->kernfs_rwsem irq_context: 0 (wq_completion)events_unbound (work_completion)(&entry->work) &dev->mutex &disk->open_mutex &dev->power.lock irq_context: 0 (wq_completion)events_unbound (work_completion)(&entry->work) &dev->mutex &disk->open_mutex dpm_list_mtx irq_context: 0 (wq_completion)events_unbound (work_completion)(&entry->work) &dev->mutex &disk->open_mutex lock kernfs_idr_lock pool_lock#2 irq_context: 0 (wq_completion)events_unbound (work_completion)(&entry->work) &dev->mutex &disk->open_mutex req_lock irq_context: 0 (wq_completion)events_unbound (work_completion)(&entry->work) &dev->mutex &disk->open_mutex &p->pi_lock irq_context: 0 (wq_completion)events_unbound (work_completion)(&entry->work) &dev->mutex &disk->open_mutex &p->pi_lock &rq->__lock irq_context: 0 (wq_completion)events_unbound (work_completion)(&entry->work) &dev->mutex &disk->open_mutex &p->pi_lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)events_unbound (work_completion)(&entry->work) &dev->mutex &disk->open_mutex &p->pi_lock &rq->__lock &cfs_rq->removed.lock irq_context: 0 (wq_completion)events_unbound (work_completion)(&entry->work) &dev->mutex &disk->open_mutex &x->wait#11 irq_context: 0 (wq_completion)events_unbound (work_completion)(&entry->work) &dev->mutex &disk->open_mutex &cfs_rq->removed.lock irq_context: 0 (wq_completion)events_unbound (work_completion)(&entry->work) &dev->mutex &disk->open_mutex &k->k_lock irq_context: 0 (wq_completion)events_unbound (work_completion)(&entry->work) &dev->mutex &disk->open_mutex subsys mutex#37 irq_context: 0 (wq_completion)events_unbound (work_completion)(&entry->work) &dev->mutex &disk->open_mutex subsys mutex#37 &k->k_lock irq_context: 0 (wq_completion)events_unbound (work_completion)(&entry->work) &dev->mutex &disk->open_mutex &xa->xa_lock#11 irq_context: 0 (wq_completion)events_unbound (work_completion)(&entry->work) &dev->mutex &disk->open_mutex &xa->xa_lock#11 pool_lock#2 irq_context: 0 (wq_completion)events_unbound (work_completion)(&entry->work) &dev->mutex &disk->open_mutex inode_hash_lock irq_context: 0 (wq_completion)events_unbound (work_completion)(&entry->work) &dev->mutex &disk->open_mutex inode_hash_lock &sb->s_type->i_lock_key#3 irq_context: 0 (wq_completion)events_unbound (work_completion)(&entry->work) &dev->mutex &disk->open_mutex purge_vmap_area_lock irq_context: 0 (wq_completion)events_unbound (work_completion)(&entry->work) &dev->mutex &disk->open_mutex &sb->s_type->i_lock_key#3 &xa->xa_lock#9 irq_context: 0 (wq_completion)events_unbound (work_completion)(&entry->work) &dev->mutex &disk->open_mutex lock#5 irq_context: 0 (wq_completion)events_unbound (work_completion)(&entry->work) &dev->mutex &disk->open_mutex &lruvec->lru_lock irq_context: 0 (wq_completion)events_unbound (work_completion)(&entry->work) &dev->mutex &disk->open_mutex rcu_read_lock &pcp->lock &zone->lock irq_context: 0 (wq_completion)events_unbound (work_completion)(&entry->work) &dev->mutex &disk->open_mutex rcu_read_lock &zone->lock irq_context: 0 (wq_completion)events_unbound (work_completion)(&entry->work) &dev->mutex &disk->open_mutex rcu_read_lock &____s->seqcount irq_context: 0 (wq_completion)events_unbound (work_completion)(&entry->work) &dev->mutex console_owner_lock irq_context: 0 (wq_completion)events_unbound (work_completion)(&entry->work) &dev->mutex console_owner irq_context: 0 xdomain_lock irq_context: 0 xdomain_lock fs_reclaim irq_context: 0 xdomain_lock fs_reclaim mmu_notifier_invalidate_range_start irq_context: 0 xdomain_lock pool_lock#2 irq_context: 0 ioctl_mutex irq_context: 0 ioctl_mutex &rq->__lock irq_context: 0 ioctl_mutex &cfs_rq->removed.lock irq_context: 0 ioctl_mutex &obj_hash[i].lock irq_context: 0 ioctl_mutex pool_lock#2 irq_context: 0 address_handler_list_lock irq_context: 0 card_mutex irq_context: 0 &type->i_mutex_dir_key#2 &c->lock irq_context: 0 &type->i_mutex_dir_key#2 &pcp->lock &zone->lock irq_context: 0 &type->i_mutex_dir_key#2 &zone->lock irq_context: 0 &type->i_mutex_dir_key#2 &____s->seqcount irq_context: 0 subsys mutex#59 irq_context: 0 subsys mutex#59 &k->k_lock irq_context: 0 &x->wait#18 irq_context: 0 &x->wait#18 &p->pi_lock irq_context: 0 &x->wait#18 &p->pi_lock &rq->__lock irq_context: 0 &x->wait#18 &p->pi_lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 &txlock irq_context: 0 &txlock &list->lock#3 irq_context: 0 &txlock &txwq irq_context: 0 &iocq[i].lock irq_context: 0 &iocq[i].lock &ktiowq[i] irq_context: 0 rcu_read_lock &c->lock irq_context: 0 &txwq irq_context: 0 &txwq &p->pi_lock irq_context: 0 &txwq &p->pi_lock &rq->__lock irq_context: 0 &txwq &p->pi_lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 rcu_read_lock_bh noop_qdisc.q.lock irq_context: 0 rcu_read_lock_bh &obj_hash[i].lock irq_context: 0 rcu_read_lock_bh pool_lock#2 irq_context: 0 &root->kernfs_rwsem &rq->__lock irq_context: 0 &root->kernfs_rwsem &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 &root->kernfs_rwsem &rq->__lock &cfs_rq->removed.lock irq_context: 0 subsys mutex#60 irq_context: 0 subsys mutex#60 &k->k_lock irq_context: 0 usb_bus_idr_lock irq_context: 0 usb_bus_idr_lock (usb_notifier_list).rwsem irq_context: 0 tasklist_lock &n->list_lock irq_context: 0 rcu_read_lock &cfs_rq->removed.lock irq_context: 0 rcu_read_lock &rcu_state.gp_wq irq_context: 0 rcu_read_lock &rcu_state.gp_wq &p->pi_lock irq_context: 0 rcu_read_lock &rcu_state.gp_wq &p->pi_lock &rq->__lock irq_context: 0 rcu_read_lock &rcu_state.gp_wq &p->pi_lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 rcu_read_lock &rcu_state.gp_wq &p->pi_lock &rq->__lock &cfs_rq->removed.lock irq_context: 0 table_lock irq_context: 0 table_lock &k->list_lock irq_context: 0 table_lock fs_reclaim irq_context: 0 table_lock fs_reclaim mmu_notifier_invalidate_range_start irq_context: 0 table_lock pool_lock#2 irq_context: 0 table_lock lock irq_context: 0 table_lock lock kernfs_idr_lock irq_context: 0 table_lock &root->kernfs_rwsem irq_context: 0 table_lock &root->kernfs_rwsem &root->kernfs_iattr_rwsem irq_context: 0 table_lock &k->k_lock irq_context: 0 table_lock &c->lock irq_context: 0 table_lock &____s->seqcount irq_context: 0 table_lock uevent_sock_mutex irq_context: 0 table_lock &obj_hash[i].lock irq_context: 0 table_lock rcu_read_lock &pool->lock/1 irq_context: 0 table_lock rcu_read_lock &pool->lock/1 &obj_hash[i].lock irq_context: 0 table_lock rcu_read_lock &pool->lock/1 pool_lock#2 irq_context: 0 table_lock rcu_read_lock &pool->lock/1 &p->pi_lock irq_context: 0 table_lock running_helpers_waitq.lock irq_context: 0 table_lock (console_sem).lock irq_context: 0 table_lock console_lock console_srcu console_owner_lock irq_context: 0 table_lock console_lock console_srcu console_owner irq_context: 0 table_lock console_lock console_srcu console_owner &port_lock_key irq_context: 0 table_lock console_lock console_srcu console_owner console_owner_lock irq_context: 0 table_lock rcu_read_lock &pool->lock/1 &p->pi_lock &rq->__lock irq_context: 0 table_lock rcu_read_lock &pool->lock/1 &p->pi_lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 lock kernfs_idr_lock batched_entropy_u8.lock irq_context: 0 lock kernfs_idr_lock kfence_freelist_lock irq_context: 0 table_lock &pcp->lock &zone->lock irq_context: 0 table_lock &zone->lock irq_context: 0 table_lock &pcp->lock &zone->lock &____s->seqcount irq_context: 0 table_lock &rq->__lock irq_context: 0 table_lock &obj_hash[i].lock pool_lock irq_context: 0 table_lock rcu_read_lock &pool->lock/1 fill_pool_map-wait-type-override pool_lock#2 irq_context: 0 table_lock rcu_read_lock &pool->lock/1 fill_pool_map-wait-type-override pool_lock irq_context: 0 table_lock lock kernfs_idr_lock pool_lock#2 irq_context: 0 table_lock &cfs_rq->removed.lock irq_context: 0 table_lock rcu_read_lock &pool->lock/1 &p->pi_lock &rq->__lock &cfs_rq->removed.lock irq_context: 0 &dev->mutex devtree_lock irq_context: 0 &dev->mutex usb_bus_idr_lock irq_context: 0 &dev->mutex usb_bus_idr_lock fs_reclaim irq_context: 0 &dev->mutex usb_bus_idr_lock fs_reclaim mmu_notifier_invalidate_range_start irq_context: 0 &dev->mutex usb_bus_idr_lock pool_lock#2 irq_context: 0 &dev->mutex (usb_notifier_list).rwsem irq_context: 0 &dev->mutex (usb_notifier_list).rwsem fs_reclaim irq_context: 0 &dev->mutex (usb_notifier_list).rwsem fs_reclaim mmu_notifier_invalidate_range_start irq_context: 0 &dev->mutex (usb_notifier_list).rwsem pool_lock#2 irq_context: 0 &dev->mutex (usb_notifier_list).rwsem pin_fs_lock irq_context: 0 &dev->mutex (usb_notifier_list).rwsem &sb->s_type->i_mutex_key#3 irq_context: 0 &dev->mutex (usb_notifier_list).rwsem &sb->s_type->i_mutex_key#3 rename_lock.seqcount irq_context: 0 &dev->mutex (usb_notifier_list).rwsem &sb->s_type->i_mutex_key#3 fs_reclaim irq_context: 0 &dev->mutex (usb_notifier_list).rwsem &sb->s_type->i_mutex_key#3 fs_reclaim mmu_notifier_invalidate_range_start irq_context: 0 &dev->mutex (usb_notifier_list).rwsem &sb->s_type->i_mutex_key#3 pool_lock#2 irq_context: 0 &dev->mutex (usb_notifier_list).rwsem &sb->s_type->i_mutex_key#3 &dentry->d_lock irq_context: 0 &dev->mutex (usb_notifier_list).rwsem &sb->s_type->i_mutex_key#3 rcu_read_lock rename_lock.seqcount irq_context: 0 &dev->mutex (usb_notifier_list).rwsem &sb->s_type->i_mutex_key#3 &dentry->d_lock &wq irq_context: 0 &dev->mutex (usb_notifier_list).rwsem &sb->s_type->i_mutex_key#3 mmu_notifier_invalidate_range_start irq_context: 0 &dev->mutex (usb_notifier_list).rwsem &sb->s_type->i_mutex_key#3 &sb->s_type->i_lock_key#7 irq_context: 0 &dev->mutex (usb_notifier_list).rwsem &sb->s_type->i_mutex_key#3 &s->s_inode_list_lock irq_context: 0 &dev->mutex (usb_notifier_list).rwsem &sb->s_type->i_mutex_key#3 tk_core.seq.seqcount irq_context: 0 &dev->mutex (usb_notifier_list).rwsem &sb->s_type->i_mutex_key#3 &sb->s_type->i_lock_key#7 &dentry->d_lock irq_context: 0 &dev->mutex (usb_notifier_list).rwsem &sb->s_type->i_mutex_key#3 &c->lock irq_context: 0 &dev->mutex (usb_notifier_list).rwsem &sb->s_type->i_mutex_key#3 &pcp->lock &zone->lock irq_context: 0 &dev->mutex (usb_notifier_list).rwsem &sb->s_type->i_mutex_key#3 &zone->lock irq_context: 0 &dev->mutex (usb_notifier_list).rwsem &sb->s_type->i_mutex_key#3 &____s->seqcount irq_context: 0 &dev->mutex (usb_notifier_list).rwsem &x->wait#9 irq_context: 0 &dev->mutex (usb_notifier_list).rwsem &obj_hash[i].lock irq_context: 0 &dev->mutex (usb_notifier_list).rwsem &k->list_lock irq_context: 0 &dev->mutex (usb_notifier_list).rwsem gdp_mutex irq_context: 0 &dev->mutex (usb_notifier_list).rwsem gdp_mutex &k->list_lock irq_context: 0 &dev->mutex (usb_notifier_list).rwsem gdp_mutex fs_reclaim irq_context: 0 &dev->mutex (usb_notifier_list).rwsem gdp_mutex fs_reclaim mmu_notifier_invalidate_range_start irq_context: 0 &dev->mutex (usb_notifier_list).rwsem gdp_mutex pool_lock#2 irq_context: 0 &dev->mutex (usb_notifier_list).rwsem gdp_mutex lock irq_context: 0 &dev->mutex (usb_notifier_list).rwsem gdp_mutex lock kernfs_idr_lock irq_context: 0 &dev->mutex (usb_notifier_list).rwsem gdp_mutex &root->kernfs_rwsem irq_context: 0 &dev->mutex (usb_notifier_list).rwsem gdp_mutex &root->kernfs_rwsem &root->kernfs_iattr_rwsem irq_context: 0 &dev->mutex (usb_notifier_list).rwsem lock irq_context: 0 &dev->mutex (usb_notifier_list).rwsem lock kernfs_idr_lock irq_context: 0 &dev->mutex (usb_notifier_list).rwsem &root->kernfs_rwsem irq_context: 0 &dev->mutex (usb_notifier_list).rwsem &root->kernfs_rwsem &root->kernfs_iattr_rwsem irq_context: 0 &dev->mutex (usb_notifier_list).rwsem bus_type_sem irq_context: 0 &dev->mutex (usb_notifier_list).rwsem sysfs_symlink_target_lock irq_context: 0 &dev->mutex (usb_notifier_list).rwsem &c->lock irq_context: 0 &dev->mutex (usb_notifier_list).rwsem &pcp->lock &zone->lock irq_context: 0 &dev->mutex (usb_notifier_list).rwsem &zone->lock irq_context: 0 &dev->mutex (usb_notifier_list).rwsem &____s->seqcount irq_context: 0 &dev->mutex (usb_notifier_list).rwsem &root->kernfs_rwsem irq_context: 0 &dev->mutex (usb_notifier_list).rwsem &dev->power.lock irq_context: 0 &dev->mutex (usb_notifier_list).rwsem dpm_list_mtx irq_context: 0 &dev->mutex (usb_notifier_list).rwsem req_lock irq_context: 0 &dev->mutex (usb_notifier_list).rwsem &p->pi_lock irq_context: 0 &dev->mutex (usb_notifier_list).rwsem &x->wait#11 irq_context: 0 &dev->mutex (usb_notifier_list).rwsem &rq->__lock irq_context: 0 &dev->mutex (usb_notifier_list).rwsem &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 &dev->mutex (usb_notifier_list).rwsem uevent_sock_mutex irq_context: 0 &dev->mutex (usb_notifier_list).rwsem rcu_read_lock &pool->lock/1 irq_context: 0 &dev->mutex (usb_notifier_list).rwsem rcu_read_lock &pool->lock/1 &obj_hash[i].lock irq_context: 0 &dev->mutex (usb_notifier_list).rwsem rcu_read_lock &pool->lock/1 pool_lock#2 irq_context: 0 &dev->mutex (usb_notifier_list).rwsem rcu_read_lock &pool->lock/1 &p->pi_lock irq_context: 0 &dev->mutex (usb_notifier_list).rwsem rcu_read_lock &pool->lock/1 &p->pi_lock &rq->__lock irq_context: 0 &dev->mutex (usb_notifier_list).rwsem rcu_read_lock &pool->lock/1 &p->pi_lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 &dev->mutex (usb_notifier_list).rwsem rcu_read_lock &pool->lock/1 &p->pi_lock &rq->__lock &cfs_rq->removed.lock irq_context: 0 &dev->mutex (usb_notifier_list).rwsem running_helpers_waitq.lock irq_context: 0 &dev->mutex (usb_notifier_list).rwsem &k->k_lock irq_context: 0 &dev->mutex (usb_notifier_list).rwsem subsys mutex#60 irq_context: 0 &dev->mutex (usb_notifier_list).rwsem subsys mutex#60 &k->k_lock irq_context: 0 &dev->mutex (usb_notifier_list).rwsem mon_lock irq_context: 0 &dev->mutex usb_port_peer_mutex irq_context: 0 &dev->mutex device_state_lock irq_context: 0 &dev->mutex usb_bus_idr_lock mmu_notifier_invalidate_range_start irq_context: 0 &dev->mutex usb_bus_idr_lock hcd_root_hub_lock irq_context: 0 &dev->mutex usb_bus_idr_lock hcd_root_hub_lock hcd_urb_list_lock irq_context: 0 &dev->mutex usb_bus_idr_lock &obj_hash[i].lock irq_context: 0 &dev->mutex usb_bus_idr_lock hcd_root_hub_lock &bh->lock irq_context: 0 &dev->mutex usb_bus_idr_lock hcd_root_hub_lock &p->pi_lock irq_context: 0 &dev->mutex usb_bus_idr_lock hcd_root_hub_lock &p->pi_lock &rq->__lock irq_context: 0 &dev->mutex usb_bus_idr_lock hcd_root_hub_lock &p->pi_lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 &dev->mutex usb_bus_idr_lock &rq->__lock irq_context: softirq &bh->lock irq_context: softirq lock#6 irq_context: softirq lock#6 kcov_remote_lock irq_context: softirq &x->wait#19 irq_context: 0 &dev->mutex usb_bus_idr_lock &x->wait#19 irq_context: 0 &dev->mutex usb_bus_idr_lock &____s->seqcount irq_context: 0 &dev->mutex usb_bus_idr_lock &pcp->lock &zone->lock irq_context: 0 &dev->mutex usb_bus_idr_lock &zone->lock irq_context: 0 &dev->mutex usb_bus_idr_lock &pcp->lock &zone->lock &____s->seqcount irq_context: 0 &dev->mutex usb_bus_idr_lock rcu_read_lock pool_lock#2 irq_context: 0 &dev->mutex usb_bus_idr_lock &dev->power.lock irq_context: 0 &dev->mutex usb_bus_idr_lock device_links_srcu irq_context: 0 &dev->mutex usb_bus_idr_lock &dev->power.lock &dev->power.lock/1 irq_context: 0 &dev->mutex usb_bus_idr_lock &base->lock irq_context: 0 &dev->mutex usb_bus_idr_lock &base->lock &obj_hash[i].lock irq_context: 0 &dev->mutex usb_bus_idr_lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: softirq &x->wait#19 &p->pi_lock irq_context: softirq &x->wait#19 &p->pi_lock &rq->__lock irq_context: softirq &x->wait#19 &p->pi_lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 &dev->mutex usb_bus_idr_lock (&timer.timer) irq_context: 0 &dev->mutex usb_bus_idr_lock (console_sem).lock irq_context: 0 &dev->mutex usb_bus_idr_lock console_lock console_srcu console_owner_lock irq_context: 0 &dev->mutex usb_bus_idr_lock console_lock console_srcu console_owner irq_context: 0 &dev->mutex usb_bus_idr_lock console_lock console_srcu console_owner &port_lock_key irq_context: 0 &dev->mutex usb_bus_idr_lock console_lock console_srcu console_owner console_owner_lock irq_context: 0 &dev->mutex usb_bus_idr_lock input_pool.lock irq_context: 0 &dev->mutex usb_bus_idr_lock &k->list_lock irq_context: 0 &dev->mutex usb_bus_idr_lock lock irq_context: 0 &dev->mutex usb_bus_idr_lock lock kernfs_idr_lock irq_context: 0 &dev->mutex usb_bus_idr_lock &root->kernfs_rwsem irq_context: 0 &dev->mutex usb_bus_idr_lock &root->kernfs_rwsem &root->kernfs_iattr_rwsem irq_context: 0 &dev->mutex usb_bus_idr_lock bus_type_sem irq_context: 0 &dev->mutex usb_bus_idr_lock &c->lock irq_context: 0 &dev->mutex usb_bus_idr_lock lock kernfs_idr_lock pool_lock#2 irq_context: 0 &dev->mutex usb_bus_idr_lock sysfs_symlink_target_lock irq_context: 0 &dev->mutex usb_bus_idr_lock &k->k_lock irq_context: 0 &dev->mutex usb_bus_idr_lock &root->kernfs_rwsem irq_context: 0 &dev->mutex usb_bus_idr_lock dpm_list_mtx irq_context: 0 &dev->mutex usb_bus_idr_lock req_lock irq_context: 0 &dev->mutex usb_bus_idr_lock &p->pi_lock irq_context: 0 &dev->mutex usb_bus_idr_lock &x->wait#11 irq_context: 0 &dev->mutex usb_bus_idr_lock &(&priv->bus_notifier)->rwsem irq_context: 0 &dev->mutex usb_bus_idr_lock &(&priv->bus_notifier)->rwsem fs_reclaim irq_context: 0 &dev->mutex usb_bus_idr_lock &(&priv->bus_notifier)->rwsem fs_reclaim mmu_notifier_invalidate_range_start irq_context: 0 &dev->mutex usb_bus_idr_lock &(&priv->bus_notifier)->rwsem &c->lock irq_context: 0 &dev->mutex usb_bus_idr_lock &(&priv->bus_notifier)->rwsem &____s->seqcount irq_context: 0 &dev->mutex usb_bus_idr_lock &(&priv->bus_notifier)->rwsem pool_lock#2 irq_context: 0 &dev->mutex usb_bus_idr_lock &(&priv->bus_notifier)->rwsem lock irq_context: 0 &dev->mutex usb_bus_idr_lock &(&priv->bus_notifier)->rwsem lock kernfs_idr_lock irq_context: 0 &dev->mutex usb_bus_idr_lock &(&priv->bus_notifier)->rwsem &root->kernfs_rwsem irq_context: 0 &dev->mutex usb_bus_idr_lock &(&priv->bus_notifier)->rwsem &root->kernfs_rwsem &root->kernfs_iattr_rwsem irq_context: 0 &dev->mutex usb_bus_idr_lock &(&priv->bus_notifier)->rwsem &root->kernfs_rwsem irq_context: 0 &dev->mutex usb_bus_idr_lock uevent_sock_mutex irq_context: 0 &dev->mutex usb_bus_idr_lock rcu_read_lock &pool->lock/1 irq_context: 0 &dev->mutex usb_bus_idr_lock rcu_read_lock &pool->lock/1 &obj_hash[i].lock irq_context: 0 &dev->mutex usb_bus_idr_lock rcu_read_lock &pool->lock/1 pool_lock#2 irq_context: 0 &dev->mutex usb_bus_idr_lock rcu_read_lock &pool->lock/1 &p->pi_lock irq_context: 0 &dev->mutex usb_bus_idr_lock rcu_read_lock &pool->lock/1 &p->pi_lock &rq->__lock irq_context: 0 &dev->mutex usb_bus_idr_lock rcu_read_lock &pool->lock/1 &p->pi_lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 &dev->mutex usb_bus_idr_lock running_helpers_waitq.lock irq_context: 0 &dev->mutex usb_bus_idr_lock &dev->mutex &dev->power.lock irq_context: 0 &dev->mutex usb_bus_idr_lock &dev->mutex &k->list_lock irq_context: 0 &dev->mutex usb_bus_idr_lock &dev->mutex &k->k_lock irq_context: 0 &dev->mutex usb_bus_idr_lock &dev->mutex device_links_srcu irq_context: 0 &dev->mutex usb_bus_idr_lock &dev->mutex fwnode_link_lock irq_context: 0 &dev->mutex usb_bus_idr_lock &dev->mutex device_links_lock irq_context: 0 &dev->mutex usb_bus_idr_lock &dev->mutex &(&priv->bus_notifier)->rwsem irq_context: 0 &dev->mutex usb_bus_idr_lock &dev->mutex sysfs_symlink_target_lock irq_context: 0 &dev->mutex usb_bus_idr_lock &dev->mutex fs_reclaim irq_context: 0 &dev->mutex usb_bus_idr_lock &dev->mutex fs_reclaim mmu_notifier_invalidate_range_start irq_context: 0 &dev->mutex usb_bus_idr_lock &dev->mutex pool_lock#2 irq_context: 0 &dev->mutex usb_bus_idr_lock &dev->mutex lock irq_context: 0 &dev->mutex usb_bus_idr_lock &dev->mutex lock kernfs_idr_lock irq_context: 0 &dev->mutex usb_bus_idr_lock &dev->mutex &root->kernfs_rwsem irq_context: 0 &dev->mutex usb_bus_idr_lock &dev->mutex &root->kernfs_rwsem &root->kernfs_iattr_rwsem irq_context: 0 &dev->mutex usb_bus_idr_lock &dev->mutex mmu_notifier_invalidate_range_start irq_context: 0 &dev->mutex usb_bus_idr_lock &dev->mutex set_config_lock irq_context: 0 &dev->mutex usb_bus_idr_lock &dev->mutex hcd->bandwidth_mutex irq_context: 0 &dev->mutex usb_bus_idr_lock &dev->mutex hcd->bandwidth_mutex devtree_lock irq_context: 0 &dev->mutex usb_bus_idr_lock &dev->mutex hcd->bandwidth_mutex &obj_hash[i].lock irq_context: 0 &dev->mutex usb_bus_idr_lock &dev->mutex hcd->bandwidth_mutex &x->wait#9 irq_context: 0 &dev->mutex usb_bus_idr_lock &dev->mutex hcd->bandwidth_mutex &dev->power.lock irq_context: 0 &dev->mutex usb_bus_idr_lock &dev->mutex hcd->bandwidth_mutex fs_reclaim irq_context: 0 &dev->mutex usb_bus_idr_lock &dev->mutex hcd->bandwidth_mutex fs_reclaim mmu_notifier_invalidate_range_start irq_context: 0 &dev->mutex usb_bus_idr_lock &dev->mutex hcd->bandwidth_mutex pool_lock#2 irq_context: 0 &dev->mutex usb_bus_idr_lock &dev->mutex hcd->bandwidth_mutex mmu_notifier_invalidate_range_start irq_context: 0 &dev->mutex usb_bus_idr_lock &dev->mutex hcd->bandwidth_mutex hcd_root_hub_lock irq_context: 0 &dev->mutex usb_bus_idr_lock &dev->mutex hcd->bandwidth_mutex hcd_root_hub_lock hcd_urb_list_lock irq_context: 0 &dev->mutex usb_bus_idr_lock &dev->mutex hcd->bandwidth_mutex hcd_root_hub_lock &bh->lock irq_context: 0 &dev->mutex usb_bus_idr_lock &dev->mutex hcd->bandwidth_mutex hcd_root_hub_lock &p->pi_lock irq_context: 0 &dev->mutex usb_bus_idr_lock &dev->mutex hcd->bandwidth_mutex hcd_root_hub_lock &p->pi_lock &rq->__lock irq_context: 0 &dev->mutex usb_bus_idr_lock &dev->mutex hcd->bandwidth_mutex hcd_root_hub_lock &p->pi_lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 &dev->mutex usb_bus_idr_lock &dev->mutex hcd->bandwidth_mutex &rq->__lock irq_context: 0 &dev->mutex usb_bus_idr_lock &dev->mutex hcd->bandwidth_mutex &x->wait#19 irq_context: 0 &dev->mutex usb_bus_idr_lock &dev->mutex device_state_lock irq_context: 0 &dev->mutex usb_bus_idr_lock &dev->mutex bus_type_sem irq_context: 0 &dev->mutex usb_bus_idr_lock &dev->mutex &c->lock irq_context: 0 &dev->mutex usb_bus_idr_lock &dev->mutex &____s->seqcount irq_context: 0 &dev->mutex usb_bus_idr_lock &dev->mutex dpm_list_mtx irq_context: 0 &dev->mutex usb_bus_idr_lock &dev->mutex uevent_sock_mutex irq_context: 0 &dev->mutex usb_bus_idr_lock &dev->mutex &obj_hash[i].lock irq_context: 0 &dev->mutex usb_bus_idr_lock &dev->mutex rcu_read_lock &pool->lock/1 irq_context: 0 &dev->mutex usb_bus_idr_lock &dev->mutex rcu_read_lock &pool->lock/1 &obj_hash[i].lock irq_context: 0 &dev->mutex usb_bus_idr_lock &dev->mutex rcu_read_lock &pool->lock/1 pool_lock#2 irq_context: 0 &dev->mutex usb_bus_idr_lock &dev->mutex rcu_read_lock &pool->lock/1 &p->pi_lock irq_context: 0 &dev->mutex usb_bus_idr_lock &dev->mutex rcu_read_lock &pool->lock/1 &p->pi_lock &rq->__lock irq_context: 0 &dev->mutex usb_bus_idr_lock &dev->mutex rcu_read_lock &pool->lock/1 &p->pi_lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 &dev->mutex usb_bus_idr_lock &dev->mutex running_helpers_waitq.lock irq_context: 0 &dev->mutex usb_bus_idr_lock &dev->mutex &rq->__lock irq_context: 0 &dev->mutex usb_bus_idr_lock &dev->mutex &pcp->lock &zone->lock irq_context: 0 &dev->mutex usb_bus_idr_lock &dev->mutex &zone->lock irq_context: 0 &dev->mutex usb_bus_idr_lock &dev->mutex rcu_read_lock pool_lock#2 irq_context: 0 &dev->mutex usb_bus_idr_lock &dev->mutex &dev->mutex &dev->power.lock irq_context: 0 &dev->mutex usb_bus_idr_lock &dev->mutex &dev->mutex &k->list_lock irq_context: 0 &dev->mutex usb_bus_idr_lock &dev->mutex &dev->mutex &k->k_lock irq_context: 0 &dev->mutex usb_bus_idr_lock &dev->mutex &dev->mutex &new_driver->dynids.lock irq_context: 0 &dev->mutex usb_bus_idr_lock &dev->mutex &dev->mutex device_links_srcu irq_context: 0 &dev->mutex usb_bus_idr_lock &dev->mutex &dev->mutex fwnode_link_lock irq_context: 0 &dev->mutex usb_bus_idr_lock &dev->mutex &dev->mutex device_links_lock irq_context: 0 &dev->mutex usb_bus_idr_lock &dev->mutex &dev->mutex &(&priv->bus_notifier)->rwsem irq_context: 0 &dev->mutex usb_bus_idr_lock &dev->mutex &dev->mutex sysfs_symlink_target_lock irq_context: 0 &dev->mutex usb_bus_idr_lock &dev->mutex &dev->mutex fs_reclaim irq_context: 0 &dev->mutex usb_bus_idr_lock &dev->mutex &dev->mutex fs_reclaim mmu_notifier_invalidate_range_start irq_context: 0 &dev->mutex usb_bus_idr_lock &dev->mutex &dev->mutex pool_lock#2 irq_context: 0 &dev->mutex usb_bus_idr_lock &dev->mutex &dev->mutex lock irq_context: 0 &dev->mutex usb_bus_idr_lock &dev->mutex &dev->mutex lock kernfs_idr_lock irq_context: 0 &dev->mutex usb_bus_idr_lock &dev->mutex &dev->mutex &root->kernfs_rwsem irq_context: 0 &dev->mutex usb_bus_idr_lock &dev->mutex &dev->mutex &root->kernfs_rwsem &root->kernfs_iattr_rwsem irq_context: 0 &dev->mutex usb_bus_idr_lock &dev->mutex &dev->mutex &dev->power.lock &dev->power.lock/1 irq_context: 0 &dev->mutex usb_bus_idr_lock &dev->mutex &dev->mutex (console_sem).lock irq_context: 0 &dev->mutex usb_bus_idr_lock &dev->mutex &dev->mutex console_lock console_srcu console_owner_lock irq_context: 0 &dev->mutex usb_bus_idr_lock &dev->mutex &dev->mutex console_lock console_srcu console_owner irq_context: 0 &dev->mutex usb_bus_idr_lock &dev->mutex &dev->mutex console_lock console_srcu console_owner &port_lock_key irq_context: 0 &dev->mutex usb_bus_idr_lock &dev->mutex &dev->mutex console_lock console_srcu console_owner console_owner_lock irq_context: 0 &dev->mutex usb_bus_idr_lock &dev->mutex &dev->mutex &c->lock irq_context: 0 &dev->mutex usb_bus_idr_lock &dev->mutex &dev->mutex &____s->seqcount irq_context: 0 &dev->mutex usb_bus_idr_lock &dev->mutex &dev->mutex &obj_hash[i].lock irq_context: 0 &dev->mutex usb_bus_idr_lock &dev->mutex &dev->mutex mmu_notifier_invalidate_range_start irq_context: 0 &dev->mutex usb_bus_idr_lock &dev->mutex &dev->mutex hcd_root_hub_lock irq_context: 0 &dev->mutex usb_bus_idr_lock &dev->mutex &dev->mutex hcd_root_hub_lock hcd_urb_list_lock irq_context: 0 &dev->mutex usb_bus_idr_lock &dev->mutex &dev->mutex &dum_hcd->dum->lock irq_context: 0 &dev->mutex usb_bus_idr_lock &dev->mutex &dev->mutex hcd_root_hub_lock &bh->lock irq_context: 0 &dev->mutex usb_bus_idr_lock &dev->mutex &dev->mutex hcd_root_hub_lock &p->pi_lock irq_context: 0 &dev->mutex usb_bus_idr_lock &dev->mutex &dev->mutex hcd_root_hub_lock &p->pi_lock &rq->__lock irq_context: 0 &dev->mutex usb_bus_idr_lock &dev->mutex &dev->mutex hcd_root_hub_lock &p->pi_lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 &dev->mutex usb_bus_idr_lock &dev->mutex &dev->mutex &rq->__lock irq_context: 0 &dev->mutex usb_bus_idr_lock &dev->mutex &dev->mutex &x->wait#19 irq_context: hardirq allocation_wait.lock &p->pi_lock &cfs_rq->removed.lock irq_context: 0 &dev->mutex usb_bus_idr_lock &dev->mutex &dev->mutex batched_entropy_u8.lock irq_context: 0 &dev->mutex usb_bus_idr_lock &dev->mutex &dev->mutex kfence_freelist_lock irq_context: 0 &dev->mutex usb_bus_idr_lock &dev->mutex &dev->mutex hcd_root_hub_lock &p->pi_lock &rq->__lock &cfs_rq->removed.lock irq_context: 0 &dev->mutex usb_bus_idr_lock &dev->mutex &dev->mutex &hub->status_mutex irq_context: 0 &dev->mutex usb_bus_idr_lock &dev->mutex &dev->mutex &hub->status_mutex mmu_notifier_invalidate_range_start irq_context: 0 &dev->mutex usb_bus_idr_lock &dev->mutex &dev->mutex &hub->status_mutex pool_lock#2 irq_context: 0 &dev->mutex usb_bus_idr_lock &dev->mutex &dev->mutex &hub->status_mutex hcd_root_hub_lock irq_context: 0 &dev->mutex usb_bus_idr_lock &dev->mutex &dev->mutex &hub->status_mutex hcd_root_hub_lock hcd_urb_list_lock irq_context: 0 &dev->mutex usb_bus_idr_lock &dev->mutex &dev->mutex &hub->status_mutex fs_reclaim irq_context: 0 &dev->mutex usb_bus_idr_lock &dev->mutex &dev->mutex &hub->status_mutex fs_reclaim mmu_notifier_invalidate_range_start irq_context: 0 &dev->mutex usb_bus_idr_lock &dev->mutex &dev->mutex &hub->status_mutex &dum_hcd->dum->lock irq_context: 0 &dev->mutex usb_bus_idr_lock &dev->mutex &dev->mutex &hub->status_mutex &obj_hash[i].lock irq_context: 0 &dev->mutex usb_bus_idr_lock &dev->mutex &dev->mutex &hub->status_mutex hcd_root_hub_lock &bh->lock irq_context: 0 &dev->mutex usb_bus_idr_lock &dev->mutex &dev->mutex &hub->status_mutex hcd_root_hub_lock &p->pi_lock irq_context: 0 &dev->mutex usb_bus_idr_lock &dev->mutex &dev->mutex &hub->status_mutex hcd_root_hub_lock &p->pi_lock &rq->__lock irq_context: 0 &dev->mutex usb_bus_idr_lock &dev->mutex &dev->mutex &hub->status_mutex hcd_root_hub_lock &p->pi_lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 &dev->mutex usb_bus_idr_lock &dev->mutex &dev->mutex &hub->status_mutex &rq->__lock irq_context: 0 &dev->mutex usb_bus_idr_lock &dev->mutex &dev->mutex &hub->status_mutex &x->wait#19 irq_context: 0 &dev->mutex usb_bus_idr_lock &dev->mutex &dev->mutex usb_port_peer_mutex irq_context: 0 &dev->mutex usb_bus_idr_lock &dev->mutex &dev->mutex usb_port_peer_mutex fs_reclaim irq_context: 0 &dev->mutex usb_bus_idr_lock &dev->mutex &dev->mutex usb_port_peer_mutex fs_reclaim mmu_notifier_invalidate_range_start irq_context: 0 &dev->mutex usb_bus_idr_lock &dev->mutex &dev->mutex usb_port_peer_mutex pool_lock#2 irq_context: 0 &dev->mutex usb_bus_idr_lock &dev->mutex &dev->mutex usb_port_peer_mutex &x->wait#9 irq_context: 0 &dev->mutex usb_bus_idr_lock &dev->mutex &dev->mutex usb_port_peer_mutex &obj_hash[i].lock irq_context: 0 &dev->mutex usb_bus_idr_lock &dev->mutex &dev->mutex usb_port_peer_mutex &k->list_lock irq_context: 0 &dev->mutex usb_bus_idr_lock &dev->mutex &dev->mutex usb_port_peer_mutex lock irq_context: 0 &dev->mutex usb_bus_idr_lock &dev->mutex &dev->mutex usb_port_peer_mutex lock kernfs_idr_lock irq_context: 0 &dev->mutex usb_bus_idr_lock &dev->mutex &dev->mutex usb_port_peer_mutex &root->kernfs_rwsem irq_context: 0 &dev->mutex usb_bus_idr_lock &dev->mutex &dev->mutex usb_port_peer_mutex &root->kernfs_rwsem &root->kernfs_iattr_rwsem irq_context: 0 &dev->mutex usb_bus_idr_lock &dev->mutex &dev->mutex usb_port_peer_mutex bus_type_sem irq_context: 0 &dev->mutex usb_bus_idr_lock &dev->mutex &dev->mutex usb_port_peer_mutex lock kernfs_idr_lock &pcp->lock &zone->lock irq_context: 0 &dev->mutex usb_bus_idr_lock &dev->mutex &dev->mutex usb_port_peer_mutex lock kernfs_idr_lock &zone->lock irq_context: 0 &dev->mutex usb_bus_idr_lock &dev->mutex &dev->mutex usb_port_peer_mutex lock kernfs_idr_lock &____s->seqcount irq_context: 0 &dev->mutex usb_bus_idr_lock &dev->mutex &dev->mutex usb_port_peer_mutex lock kernfs_idr_lock pool_lock#2 irq_context: 0 &dev->mutex usb_bus_idr_lock &dev->mutex &dev->mutex usb_port_peer_mutex &c->lock irq_context: 0 &dev->mutex usb_bus_idr_lock &dev->mutex &dev->mutex usb_port_peer_mutex &pcp->lock &zone->lock irq_context: 0 &dev->mutex usb_bus_idr_lock &dev->mutex &dev->mutex usb_port_peer_mutex &zone->lock irq_context: 0 &dev->mutex usb_bus_idr_lock &dev->mutex &dev->mutex usb_port_peer_mutex &____s->seqcount irq_context: 0 &dev->mutex usb_bus_idr_lock &dev->mutex &dev->mutex usb_port_peer_mutex &root->kernfs_rwsem irq_context: 0 &dev->mutex usb_bus_idr_lock &dev->mutex &dev->mutex usb_port_peer_mutex &dev->power.lock irq_context: 0 &dev->mutex usb_bus_idr_lock &dev->mutex &dev->mutex usb_port_peer_mutex dpm_list_mtx irq_context: 0 &dev->mutex usb_bus_idr_lock &dev->mutex &dev->mutex usb_port_peer_mutex &k->k_lock irq_context: 0 &dev->mutex usb_bus_idr_lock &dev->mutex &dev->mutex usb_port_peer_mutex dev_pm_qos_mtx irq_context: 0 &dev->mutex usb_bus_idr_lock &dev->mutex &dev->mutex usb_port_peer_mutex dev_pm_qos_mtx fs_reclaim irq_context: 0 &dev->mutex usb_bus_idr_lock &dev->mutex &dev->mutex usb_port_peer_mutex dev_pm_qos_mtx fs_reclaim mmu_notifier_invalidate_range_start irq_context: 0 &dev->mutex usb_bus_idr_lock &dev->mutex &dev->mutex usb_port_peer_mutex dev_pm_qos_mtx pool_lock#2 irq_context: 0 &dev->mutex usb_bus_idr_lock &dev->mutex &dev->mutex usb_port_peer_mutex dev_pm_qos_mtx &dev->power.lock irq_context: 0 &dev->mutex usb_bus_idr_lock &dev->mutex &dev->mutex usb_port_peer_mutex dev_pm_qos_mtx pm_qos_lock irq_context: 0 &dev->mutex usb_bus_idr_lock &dev->mutex &dev->mutex usb_port_peer_mutex component_mutex irq_context: 0 &dev->mutex usb_bus_idr_lock &dev->mutex &dev->mutex usb_port_peer_mutex device_links_srcu irq_context: 0 &dev->mutex usb_bus_idr_lock &dev->mutex &dev->mutex usb_port_peer_mutex &dev->power.lock &dev->power.lock/1 irq_context: 0 &dev->mutex usb_bus_idr_lock &dev->mutex &dev->mutex usb_port_peer_mutex dev_pm_qos_sysfs_mtx irq_context: 0 &dev->mutex usb_bus_idr_lock &dev->mutex &dev->mutex usb_port_peer_mutex dev_pm_qos_sysfs_mtx dev_pm_qos_mtx irq_context: 0 &dev->mutex usb_bus_idr_lock &dev->mutex &dev->mutex usb_port_peer_mutex dev_pm_qos_sysfs_mtx &root->kernfs_rwsem irq_context: 0 &dev->mutex usb_bus_idr_lock &dev->mutex &dev->mutex usb_port_peer_mutex dev_pm_qos_sysfs_mtx fs_reclaim irq_context: 0 &dev->mutex usb_bus_idr_lock &dev->mutex &dev->mutex usb_port_peer_mutex dev_pm_qos_sysfs_mtx fs_reclaim mmu_notifier_invalidate_range_start irq_context: 0 &dev->mutex usb_bus_idr_lock &dev->mutex &dev->mutex usb_port_peer_mutex dev_pm_qos_sysfs_mtx pool_lock#2 irq_context: 0 &dev->mutex usb_bus_idr_lock &dev->mutex &dev->mutex usb_port_peer_mutex dev_pm_qos_sysfs_mtx lock irq_context: 0 &dev->mutex usb_bus_idr_lock &dev->mutex &dev->mutex usb_port_peer_mutex dev_pm_qos_sysfs_mtx lock kernfs_idr_lock irq_context: 0 &dev->mutex usb_bus_idr_lock &dev->mutex &dev->mutex usb_port_peer_mutex dev_pm_qos_sysfs_mtx &root->kernfs_rwsem irq_context: 0 &dev->mutex usb_bus_idr_lock &dev->mutex &dev->mutex usb_port_peer_mutex dev_pm_qos_sysfs_mtx &root->kernfs_rwsem &root->kernfs_iattr_rwsem irq_context: 0 &dev->mutex usb_bus_idr_lock &dev->mutex &dev->mutex usb_port_peer_mutex &dev->power.lock rcu_read_lock &pool->lock irq_context: 0 &dev->mutex usb_bus_idr_lock &dev->mutex &dev->mutex usb_port_peer_mutex &dev->power.lock rcu_read_lock &pool->lock &obj_hash[i].lock irq_context: 0 &dev->mutex usb_bus_idr_lock &dev->mutex &dev->mutex usb_port_peer_mutex &dev->power.lock rcu_read_lock &pool->lock pool_lock#2 irq_context: 0 &dev->mutex usb_bus_idr_lock &dev->mutex &dev->mutex usb_port_peer_mutex &dev->power.lock rcu_read_lock &pool->lock &p->pi_lock irq_context: 0 &dev->mutex usb_bus_idr_lock &dev->mutex &dev->mutex usb_port_peer_mutex &dev->power.lock rcu_read_lock &pool->lock &p->pi_lock &rq->__lock irq_context: 0 &dev->mutex usb_bus_idr_lock &dev->mutex &dev->mutex usb_port_peer_mutex &dev->power.lock rcu_read_lock &pool->lock &p->pi_lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 &dev->mutex usb_bus_idr_lock &dev->mutex &dev->mutex usb_port_peer_mutex &rq->__lock irq_context: 0 &dev->mutex usb_bus_idr_lock &dev->mutex &dev->mutex rcu_read_lock &pool->lock irq_context: 0 &dev->mutex usb_bus_idr_lock &dev->mutex &dev->mutex rcu_read_lock &pool->lock &obj_hash[i].lock irq_context: 0 &dev->mutex usb_bus_idr_lock &dev->mutex &dev->mutex rcu_read_lock &pool->lock pool_lock#2 irq_context: 0 &dev->mutex usb_bus_idr_lock &dev->mutex &dev->mutex rcu_read_lock &pool->lock &p->pi_lock irq_context: 0 &dev->mutex usb_bus_idr_lock &dev->mutex &dev->mutex rcu_read_lock &pool->lock &p->pi_lock &rq->__lock irq_context: 0 &dev->mutex usb_bus_idr_lock &dev->mutex &dev->mutex rcu_read_lock &pool->lock &p->pi_lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 &dev->mutex usb_bus_idr_lock &dev->mutex &dev->mutex deferred_probe_mutex irq_context: 0 &dev->mutex usb_bus_idr_lock &dev->mutex &dev->mutex &pcp->lock &zone->lock irq_context: 0 &dev->mutex usb_bus_idr_lock &dev->mutex &dev->mutex &zone->lock irq_context: 0 &dev->mutex usb_bus_idr_lock &dev->mutex &dev->mutex uevent_sock_mutex irq_context: 0 &dev->mutex usb_bus_idr_lock &dev->mutex &dev->mutex rcu_read_lock &pool->lock/1 irq_context: 0 &dev->mutex usb_bus_idr_lock &dev->mutex &dev->mutex rcu_read_lock &pool->lock/1 &obj_hash[i].lock irq_context: 0 &dev->mutex usb_bus_idr_lock &dev->mutex &dev->mutex rcu_read_lock &pool->lock/1 pool_lock#2 irq_context: 0 &dev->mutex usb_bus_idr_lock &dev->mutex &dev->mutex rcu_read_lock &pool->lock/1 &p->pi_lock irq_context: 0 &dev->mutex usb_bus_idr_lock &dev->mutex &dev->mutex running_helpers_waitq.lock irq_context: 0 &dev->mutex usb_bus_idr_lock &dev->mutex &dev->mutex probe_waitqueue.lock irq_context: 0 &dev->mutex usb_bus_idr_lock &dev->mutex subsys mutex#61 irq_context: 0 &dev->mutex usb_bus_idr_lock &dev->mutex &x->wait#9 irq_context: 0 &dev->mutex usb_bus_idr_lock &dev->mutex &root->kernfs_rwsem irq_context: 0 &dev->mutex usb_bus_idr_lock &dev->mutex (usb_notifier_list).rwsem irq_context: 0 &dev->mutex usb_bus_idr_lock &dev->mutex deferred_probe_mutex irq_context: 0 &dev->mutex usb_bus_idr_lock &dev->mutex probe_waitqueue.lock irq_context: 0 &dev->mutex usb_bus_idr_lock subsys mutex#61 irq_context: 0 &dev->mutex usb_bus_idr_lock &x->wait#9 irq_context: 0 (wq_completion)events_power_efficient (work_completion)(&(&hub->init_work)->work) irq_context: 0 (wq_completion)events_power_efficient (work_completion)(&(&hub->init_work)->work) &dev->mutex &hub->status_mutex irq_context: 0 (wq_completion)events_power_efficient (work_completion)(&(&hub->init_work)->work) &dev->mutex &hub->status_mutex mmu_notifier_invalidate_range_start irq_context: 0 (wq_completion)events_power_efficient (work_completion)(&(&hub->init_work)->work) &dev->mutex &hub->status_mutex pool_lock#2 irq_context: 0 (wq_completion)events_power_efficient (work_completion)(&(&hub->init_work)->work) &dev->mutex &hub->status_mutex hcd_root_hub_lock irq_context: 0 (wq_completion)events_power_efficient (work_completion)(&(&hub->init_work)->work) &dev->mutex &hub->status_mutex hcd_root_hub_lock hcd_urb_list_lock irq_context: 0 (wq_completion)events_power_efficient (work_completion)(&(&hub->init_work)->work) &dev->mutex &hub->status_mutex fs_reclaim irq_context: 0 (wq_completion)events_power_efficient (work_completion)(&(&hub->init_work)->work) &dev->mutex &hub->status_mutex fs_reclaim mmu_notifier_invalidate_range_start irq_context: 0 (wq_completion)events_power_efficient (work_completion)(&(&hub->init_work)->work) &dev->mutex &hub->status_mutex &dum_hcd->dum->lock irq_context: 0 (wq_completion)events_power_efficient (work_completion)(&(&hub->init_work)->work) &dev->mutex &hub->status_mutex &obj_hash[i].lock irq_context: 0 (wq_completion)events_power_efficient (work_completion)(&(&hub->init_work)->work) &dev->mutex &hub->status_mutex hcd_root_hub_lock &bh->lock irq_context: 0 (wq_completion)events_power_efficient (work_completion)(&(&hub->init_work)->work) &dev->mutex &hub->status_mutex hcd_root_hub_lock &p->pi_lock irq_context: 0 (wq_completion)events_power_efficient (work_completion)(&(&hub->init_work)->work) &dev->mutex &hub->status_mutex hcd_root_hub_lock &p->pi_lock &rq->__lock irq_context: 0 (wq_completion)events_power_efficient (work_completion)(&(&hub->init_work)->work) &dev->mutex &hub->status_mutex hcd_root_hub_lock &p->pi_lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)events_power_efficient (work_completion)(&(&hub->init_work)->work) &dev->mutex &hub->status_mutex &rq->__lock irq_context: 0 (wq_completion)events_power_efficient (work_completion)(&(&hub->init_work)->work) &dev->mutex &hub->status_mutex &x->wait#19 irq_context: 0 (wq_completion)events_power_efficient (work_completion)(&(&hub->init_work)->work) &dev->mutex hcd_root_hub_lock irq_context: 0 (wq_completion)events_power_efficient (work_completion)(&(&hub->init_work)->work) &dev->mutex hcd_root_hub_lock hcd_urb_list_lock irq_context: 0 (wq_completion)events_power_efficient (work_completion)(&(&hub->init_work)->work) &dev->mutex rcu_read_lock &pool->lock irq_context: 0 (wq_completion)events_power_efficient (work_completion)(&(&hub->init_work)->work) &dev->mutex rcu_read_lock &pool->lock &obj_hash[i].lock irq_context: 0 (wq_completion)events_power_efficient (work_completion)(&(&hub->init_work)->work) &dev->mutex rcu_read_lock &pool->lock pool_lock#2 irq_context: 0 (wq_completion)usb_hub_wq irq_context: 0 (wq_completion)usb_hub_wq (work_completion)(&hub->events) irq_context: 0 (wq_completion)usb_hub_wq (work_completion)(&hub->events) lock#6 irq_context: 0 (wq_completion)usb_hub_wq (work_completion)(&hub->events) lock#6 kcov_remote_lock irq_context: 0 (wq_completion)usb_hub_wq (work_completion)(&hub->events) &dev->mutex &dev->power.lock irq_context: 0 (wq_completion)usb_hub_wq (work_completion)(&hub->events) &dev->power.lock irq_context: 0 (wq_completion)usb_hub_wq (work_completion)(&hub->events) &dev->power.lock &dev->power.wait_queue irq_context: 0 (wq_completion)usb_hub_wq (work_completion)(&hub->events) &dev->power.lock rcu_read_lock &pool->lock irq_context: 0 (wq_completion)usb_hub_wq (work_completion)(&hub->events) &dev->power.lock rcu_read_lock &pool->lock &obj_hash[i].lock irq_context: 0 (wq_completion)usb_hub_wq (work_completion)(&hub->events) &dev->power.lock rcu_read_lock &pool->lock pool_lock#2 irq_context: 0 (wq_completion)pm (work_completion)(&dev->power.work) &hub->irq_urb_lock irq_context: 0 (wq_completion)pm (work_completion)(&dev->power.work) (&hub->irq_urb_retry) irq_context: 0 (wq_completion)pm (work_completion)(&dev->power.work) &obj_hash[i].lock irq_context: 0 (wq_completion)pm (work_completion)(&dev->power.work) &base->lock irq_context: 0 (wq_completion)pm (work_completion)(&dev->power.work) hcd_urb_unlink_lock irq_context: 0 (wq_completion)pm (work_completion)(&dev->power.work) hcd_root_hub_lock irq_context: 0 (wq_completion)pm (work_completion)(&dev->power.work) hcd_root_hub_lock hcd_urb_list_lock irq_context: 0 (wq_completion)pm (work_completion)(&dev->power.work) hcd_root_hub_lock &bh->lock irq_context: 0 (wq_completion)pm (work_completion)(&dev->power.work) hcd_root_hub_lock &p->pi_lock irq_context: 0 (wq_completion)pm (work_completion)(&dev->power.work) hcd_root_hub_lock &p->pi_lock &rq->__lock irq_context: 0 (wq_completion)pm (work_completion)(&dev->power.work) hcd_root_hub_lock &p->pi_lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)pm (work_completion)(&dev->power.work) &rq->__lock irq_context: softirq usb_kill_urb_queue.lock irq_context: 0 (wq_completion)pm (work_completion)(&dev->power.work) (work_completion)(&hub->tt.clear_work) irq_context: 0 (wq_completion)pm (work_completion)(&dev->power.work) &dum_hcd->dum->lock irq_context: 0 (wq_completion)pm (work_completion)(&dev->power.work) device_state_lock irq_context: 0 (wq_completion)pm (work_completion)(&dev->power.work) hcd_urb_list_lock irq_context: 0 &dev->mutex (usb_notifier_list).rwsem lock kernfs_idr_lock pool_lock#2 irq_context: 0 &dev->mutex (usb_notifier_list).rwsem &p->pi_lock &rq->__lock irq_context: 0 &dev->mutex (usb_notifier_list).rwsem &p->pi_lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 &dev->mutex (usb_notifier_list).rwsem &pcp->lock &zone->lock &____s->seqcount irq_context: 0 &dev->mutex usb_bus_idr_lock &p->pi_lock &rq->__lock irq_context: 0 &dev->mutex usb_bus_idr_lock &p->pi_lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 &dev->mutex usb_bus_idr_lock &(&priv->bus_notifier)->rwsem lock kernfs_idr_lock pool_lock#2 irq_context: 0 (wq_completion)events_power_efficient (work_completion)(&(&hub->init_work)->work) &dev->mutex &lock->wait_lock irq_context: 0 (wq_completion)events_power_efficient (work_completion)(&(&hub->init_work)->work) &dev->mutex &pool->lock irq_context: 0 (wq_completion)events_power_efficient (work_completion)(&(&hub->init_work)->work) &dev->mutex &rq->__lock irq_context: 0 (wq_completion)events_power_efficient (work_completion)(&(&hub->init_work)->work) &dev->mutex &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 &dev->mutex usb_bus_idr_lock &dev->mutex &dev->mutex rcu_read_lock &pool->lock/1 &p->pi_lock &rq->__lock irq_context: 0 &dev->mutex usb_bus_idr_lock &dev->mutex &dev->mutex rcu_read_lock &pool->lock/1 &p->pi_lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 &dev->mutex usb_bus_idr_lock &lock->wait_lock irq_context: 0 (wq_completion)events_power_efficient (work_completion)(&(&hub->init_work)->work) &dev->mutex &hub->status_mutex &c->lock irq_context: 0 (wq_completion)events_power_efficient (work_completion)(&(&hub->init_work)->work) &dev->mutex &hub->status_mutex &pcp->lock &zone->lock irq_context: 0 (wq_completion)events_power_efficient (work_completion)(&(&hub->init_work)->work) &dev->mutex &hub->status_mutex &zone->lock irq_context: 0 (wq_completion)events_power_efficient (work_completion)(&(&hub->init_work)->work) &dev->mutex &hub->status_mutex &____s->seqcount irq_context: 0 &dev->mutex usb_bus_idr_lock &hub->irq_urb_lock irq_context: 0 &dev->mutex usb_bus_idr_lock (&hub->irq_urb_retry) irq_context: 0 &dev->mutex usb_bus_idr_lock hcd_urb_unlink_lock irq_context: 0 &dev->mutex usb_bus_idr_lock (work_completion)(&hub->tt.clear_work) irq_context: 0 &dev->mutex usb_bus_idr_lock &dum_hcd->dum->lock irq_context: 0 &dev->mutex usb_bus_idr_lock device_state_lock irq_context: 0 &dev->mutex usb_bus_idr_lock hcd_urb_list_lock irq_context: 0 &dev->mutex usb_bus_idr_lock &dev->power.lock &dev->power.wait_queue irq_context: 0 &dev->mutex usb_bus_idr_lock &dev->mutex &dev->mutex &obj_hash[i].lock pool_lock irq_context: softirq rcu_callback put_task_map-wait-type-override &base->lock irq_context: softirq rcu_callback put_task_map-wait-type-override &base->lock &obj_hash[i].lock irq_context: 0 &dev->mutex usb_bus_idr_lock lock kernfs_idr_lock &c->lock irq_context: 0 &dev->mutex usb_bus_idr_lock lock kernfs_idr_lock &pcp->lock &zone->lock irq_context: 0 &dev->mutex usb_bus_idr_lock lock kernfs_idr_lock &zone->lock irq_context: 0 &dev->mutex usb_bus_idr_lock lock kernfs_idr_lock &____s->seqcount irq_context: 0 &dev->mutex usb_bus_idr_lock &dev->mutex lock kernfs_idr_lock pool_lock#2 irq_context: 0 &dev->mutex usb_bus_idr_lock &dev->mutex &dev->mutex usb_port_peer_mutex dev_pm_qos_mtx &c->lock irq_context: 0 &dev->mutex usb_bus_idr_lock &dev->mutex &dev->mutex usb_port_peer_mutex dev_pm_qos_mtx &pcp->lock &zone->lock irq_context: 0 &dev->mutex usb_bus_idr_lock &dev->mutex &dev->mutex usb_port_peer_mutex dev_pm_qos_mtx &zone->lock irq_context: 0 &dev->mutex usb_bus_idr_lock &dev->mutex &dev->mutex usb_port_peer_mutex dev_pm_qos_mtx &____s->seqcount irq_context: 0 &dev->mutex (usb_notifier_list).rwsem &obj_hash[i].lock pool_lock irq_context: 0 &dev->mutex usb_bus_idr_lock &(&priv->bus_notifier)->rwsem &pcp->lock &zone->lock irq_context: 0 &dev->mutex usb_bus_idr_lock &(&priv->bus_notifier)->rwsem &zone->lock irq_context: 0 &dev->mutex usb_bus_idr_lock &dev->mutex hcd->bandwidth_mutex &base->lock irq_context: 0 &dev->mutex usb_bus_idr_lock &dev->mutex hcd->bandwidth_mutex &base->lock &obj_hash[i].lock irq_context: 0 &dev->mutex usb_bus_idr_lock &dev->mutex hcd->bandwidth_mutex &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 &dev->mutex usb_bus_idr_lock &dev->mutex hcd->bandwidth_mutex (&timer.timer) irq_context: 0 &dev->mutex usb_bus_idr_lock &dev->mutex &dev->mutex &base->lock irq_context: 0 &dev->mutex usb_bus_idr_lock &dev->mutex &dev->mutex &base->lock &obj_hash[i].lock irq_context: 0 &dev->mutex usb_bus_idr_lock &dev->mutex &dev->mutex &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 &dev->mutex usb_bus_idr_lock &dev->mutex &dev->mutex (&timer.timer) irq_context: 0 &dev->mutex usb_bus_idr_lock &dev->mutex &dev->mutex &hub->status_mutex &base->lock irq_context: 0 &dev->mutex usb_bus_idr_lock &dev->mutex &dev->mutex &hub->status_mutex &base->lock &obj_hash[i].lock irq_context: 0 &dev->mutex usb_bus_idr_lock &dev->mutex &dev->mutex &hub->status_mutex &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 &dev->mutex usb_bus_idr_lock &dev->mutex &dev->mutex &hub->status_mutex (&timer.timer) irq_context: 0 (wq_completion)events_power_efficient (work_completion)(&(&hub->init_work)->work) &dev->mutex &hub->status_mutex &base->lock irq_context: 0 (wq_completion)events_power_efficient (work_completion)(&(&hub->init_work)->work) &dev->mutex &hub->status_mutex &base->lock &obj_hash[i].lock irq_context: 0 (wq_completion)events_power_efficient (work_completion)(&(&hub->init_work)->work) &dev->mutex &hub->status_mutex &pool->lock irq_context: 0 (wq_completion)events_power_efficient (work_completion)(&(&hub->init_work)->work) &dev->mutex &hub->status_mutex &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)events_power_efficient (work_completion)(&(&hub->init_work)->work) &dev->mutex &hub->status_mutex (&timer.timer) irq_context: 0 &dev->mutex usb_bus_idr_lock &obj_hash[i].lock pool_lock irq_context: 0 &dev->mutex usb_bus_idr_lock batched_entropy_u8.lock irq_context: 0 &dev->mutex usb_bus_idr_lock kfence_freelist_lock irq_context: 0 &dev->mutex usb_bus_idr_lock &dev->mutex &dev->mutex &pcp->lock &zone->lock &____s->seqcount irq_context: 0 &dev->mutex usb_bus_idr_lock &dev->mutex &obj_hash[i].lock pool_lock irq_context: 0 (wq_completion)pm (work_completion)(&dev->power.work) usb_kill_urb_queue.lock irq_context: 0 (wq_completion)pm (work_completion)(&dev->power.work) &pool->lock irq_context: 0 (wq_completion)pm (work_completion)(&dev->power.work) &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: softirq usb_kill_urb_queue.lock &p->pi_lock irq_context: softirq usb_kill_urb_queue.lock &p->pi_lock &rq->__lock irq_context: softirq usb_kill_urb_queue.lock &p->pi_lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 &dev->mutex usb_bus_idr_lock &dev->mutex rcu_read_lock &pool->lock/1 &p->pi_lock &cfs_rq->removed.lock irq_context: softirq lib/debugobjects.c:101 irq_context: softirq lib/debugobjects.c:101 rcu_read_lock &pool->lock irq_context: softirq lib/debugobjects.c:101 rcu_read_lock &pool->lock &obj_hash[i].lock irq_context: softirq lib/debugobjects.c:101 rcu_read_lock &pool->lock &p->pi_lock irq_context: softirq lib/debugobjects.c:101 rcu_read_lock &pool->lock &p->pi_lock &rq->__lock irq_context: softirq lib/debugobjects.c:101 rcu_read_lock &pool->lock &p->pi_lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)events (debug_obj_work).work irq_context: 0 &dev->mutex udc_lock irq_context: 0 &dev->mutex subsys mutex#62 irq_context: 0 &dev->mutex subsys mutex#62 &k->k_lock irq_context: 0 &dev->mutex gadget_id_numbers.xa_lock irq_context: 0 (wq_completion)events (work_completion)(&gadget->work) irq_context: 0 (wq_completion)events (work_completion)(&gadget->work) &root->kernfs_rwsem irq_context: 0 (wq_completion)events (work_completion)(&gadget->work) kernfs_notify_lock irq_context: 0 (wq_completion)events (work_completion)(&gadget->work) kernfs_notify_lock rcu_read_lock &pool->lock irq_context: 0 (wq_completion)events (work_completion)(&gadget->work) kernfs_notify_lock rcu_read_lock &pool->lock &obj_hash[i].lock irq_context: 0 (wq_completion)events kernfs_notify_work irq_context: 0 (wq_completion)events kernfs_notify_work kernfs_notify_lock irq_context: 0 (wq_completion)events kernfs_notify_work &root->kernfs_supers_rwsem irq_context: 0 &sig->cred_guard_mutex &mm->mmap_lock rcu_read_lock &rcu_state.gp_wq &p->pi_lock &cfs_rq->removed.lock irq_context: 0 &dev->mutex subsys mutex#63 irq_context: 0 func_lock irq_context: 0 g_tf_lock irq_context: 0 &sig->cred_guard_mutex &dentry->d_lock &dentry->d_lock/1 irq_context: 0 &sig->cred_guard_mutex &dentry->d_lock/1 irq_context: 0 &dev->mutex (usb_notifier_list).rwsem rcu_read_lock pool_lock#2 irq_context: 0 &dev->mutex usb_bus_idr_lock &dev->mutex &dev->mutex &vhci_hcd->vhci->lock irq_context: 0 &dev->mutex usb_bus_idr_lock &dev->mutex &dev->mutex rcu_read_lock pool_lock#2 irq_context: 0 &dev->mutex usb_bus_idr_lock &dev->mutex &dev->mutex &hub->status_mutex &vhci_hcd->vhci->lock irq_context: 0 (wq_completion)events_unbound (work_completion)(&(&kfence_timer)->work) &cfs_rq->removed.lock irq_context: 0 (wq_completion)events_unbound (work_completion)(&(&kfence_timer)->work) pool_lock#2 irq_context: 0 (wq_completion)events_power_efficient (work_completion)(&(&hub->init_work)->work) &dev->mutex &hub->status_mutex &vhci_hcd->vhci->lock irq_context: 0 (wq_completion)pm (work_completion)(&dev->power.work) &vhci_hcd->vhci->lock irq_context: 0 &dev->mutex usb_bus_idr_lock &vhci_hcd->vhci->lock irq_context: 0 &dev->mutex usb_bus_idr_lock quarantine_lock irq_context: 0 &dev->mutex usb_bus_idr_lock &dev->mutex &dev->mutex usb_port_peer_mutex sysfs_symlink_target_lock irq_context: 0 &dev->mutex usb_bus_idr_lock &dev->mutex &dev->mutex usb_port_peer_mutex rcu_read_lock pool_lock#2 irq_context: 0 &dev->mutex usb_bus_idr_lock &dev->mutex &dev->mutex usb_port_peer_mutex &pcp->lock &zone->lock &____s->seqcount irq_context: 0 &dev->mutex usb_bus_idr_lock &cfs_rq->removed.lock irq_context: 0 &dev->mutex usb_bus_idr_lock hcd_root_hub_lock &p->pi_lock &rq->__lock &cfs_rq->removed.lock irq_context: 0 &dev->mutex usb_bus_idr_lock usb_kill_urb_queue.lock irq_context: 0 &dev->mutex usb_bus_idr_lock &dev->mutex &dev->mutex usb_port_peer_mutex lock kernfs_idr_lock &c->lock irq_context: 0 &dev->mutex usb_bus_idr_lock &dev->mutex &cfs_rq->removed.lock irq_context: 0 &dev->mutex (usb_notifier_list).rwsem &rq->__lock &cfs_rq->removed.lock irq_context: 0 &dev->mutex usb_bus_idr_lock &dev->mutex hcd->bandwidth_mutex &c->lock irq_context: 0 &dev->mutex usb_bus_idr_lock &dev->mutex hcd->bandwidth_mutex &pcp->lock &zone->lock irq_context: 0 &dev->mutex usb_bus_idr_lock &dev->mutex hcd->bandwidth_mutex &zone->lock irq_context: 0 &dev->mutex usb_bus_idr_lock &dev->mutex hcd->bandwidth_mutex &____s->seqcount irq_context: 0 &dev->mutex usb_bus_idr_lock &dev->mutex &dev->mutex usb_port_peer_mutex dev_pm_qos_sysfs_mtx lock kernfs_idr_lock pool_lock#2 irq_context: 0 &dev->mutex usb_bus_idr_lock &dev->mutex &dev->mutex usb_port_peer_mutex dev_pm_qos_sysfs_mtx &c->lock irq_context: 0 &dev->mutex usb_bus_idr_lock &dev->mutex &dev->mutex usb_port_peer_mutex dev_pm_qos_sysfs_mtx &____s->seqcount irq_context: 0 &dev->mutex usb_bus_idr_lock &dev->mutex &dev->mutex usb_port_peer_mutex &obj_hash[i].lock pool_lock irq_context: 0 &dev->mutex (usb_notifier_list).rwsem &cfs_rq->removed.lock irq_context: softirq net/core/link_watch.c:31 irq_context: softirq net/core/link_watch.c:31 rcu_read_lock &pool->lock irq_context: softirq net/core/link_watch.c:31 rcu_read_lock &pool->lock &obj_hash[i].lock irq_context: softirq net/core/link_watch.c:31 rcu_read_lock &pool->lock &p->pi_lock irq_context: softirq net/core/link_watch.c:31 rcu_read_lock &pool->lock &p->pi_lock &rq->__lock irq_context: softirq net/core/link_watch.c:31 rcu_read_lock &pool->lock &p->pi_lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: softirq net/core/link_watch.c:31 rcu_read_lock &pool->lock &p->pi_lock &rq->__lock &cfs_rq->removed.lock irq_context: softirq &x->wait#19 &p->pi_lock &cfs_rq->removed.lock irq_context: 0 &dev->mutex rcu_read_lock &pool->lock/1 &p->pi_lock &cfs_rq->removed.lock irq_context: 0 &dev->mutex usb_bus_idr_lock &dev->mutex &dev->mutex usb_port_peer_mutex dev_pm_qos_sysfs_mtx &pcp->lock &zone->lock irq_context: 0 &dev->mutex usb_bus_idr_lock &dev->mutex &dev->mutex usb_port_peer_mutex dev_pm_qos_sysfs_mtx &zone->lock irq_context: 0 (wq_completion)pm (work_completion)(&dev->power.work) mmu_notifier_invalidate_range_start irq_context: 0 (wq_completion)pm (work_completion)(&dev->power.work) pool_lock#2 irq_context: 0 (wq_completion)pm (work_completion)(&dev->power.work) fs_reclaim irq_context: 0 (wq_completion)pm (work_completion)(&dev->power.work) fs_reclaim mmu_notifier_invalidate_range_start irq_context: 0 (wq_completion)pm (work_completion)(&dev->power.work) &x->wait#19 irq_context: 0 (wq_completion)pm (work_completion)(&dev->power.work) &base->lock &obj_hash[i].lock irq_context: 0 (wq_completion)pm (work_completion)(&dev->power.work) (&timer.timer) irq_context: 0 (wq_completion)pm (work_completion)(&dev->power.work) &c->lock irq_context: 0 (wq_completion)pm (work_completion)(&dev->power.work) &pcp->lock &zone->lock irq_context: 0 (wq_completion)pm (work_completion)(&dev->power.work) &zone->lock irq_context: 0 (wq_completion)pm (work_completion)(&dev->power.work) &____s->seqcount irq_context: 0 (wq_completion)pm (work_completion)(&dev->power.work) batched_entropy_u8.lock irq_context: 0 (wq_completion)pm (work_completion)(&dev->power.work) kfence_freelist_lock irq_context: 0 (wq_completion)pm (work_completion)(&dev->power.work) &meta->lock irq_context: 0 &dev->mutex (usb_notifier_list).rwsem &sb->s_type->i_mutex_key#3 mmu_notifier_invalidate_range_start &rq->__lock irq_context: 0 &dev->mutex (usb_notifier_list).rwsem rcu_read_lock &rq->__lock irq_context: 0 &dev->mutex (usb_notifier_list).rwsem rcu_read_lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 &dev->mutex usb_bus_idr_lock &p->pi_lock &rq->__lock &cfs_rq->removed.lock irq_context: 0 &dev->mutex usb_bus_idr_lock &dev->mutex &pcp->lock &zone->lock &____s->seqcount irq_context: softirq drivers/char/random.c:251 rcu_read_lock &pool->lock/1 &p->pi_lock &rq->__lock &cfs_rq->removed.lock irq_context: 0 sb_writers &type->i_mutex_dir_key/1 batched_entropy_u32.lock crngs.lock base_crng.lock irq_context: 0 &dev->mutex usb_bus_idr_lock &dev->mutex &dev->mutex usb_port_peer_mutex dev_pm_qos_mtx &pcp->lock &zone->lock &____s->seqcount irq_context: 0 &dev->mutex usb_bus_idr_lock &dev->mutex &dev->mutex lock kernfs_idr_lock pool_lock#2 irq_context: 0 &dev->mutex usb_bus_idr_lock &dev->mutex &dev->mutex usb_port_peer_mutex fill_pool_map-wait-type-override pool_lock#2 irq_context: 0 &dev->mutex usb_bus_idr_lock &dev->mutex &dev->mutex usb_port_peer_mutex fill_pool_map-wait-type-override pool_lock irq_context: 0 &dev->mutex usb_bus_idr_lock fill_pool_map-wait-type-override pool_lock#2 irq_context: 0 &dev->mutex usb_bus_idr_lock fill_pool_map-wait-type-override pool_lock irq_context: 0 &dev->mutex (usb_notifier_list).rwsem &sb->s_type->i_mutex_key#3 &pcp->lock &zone->lock &____s->seqcount irq_context: 0 batched_entropy_u8.lock crngs.lock base_crng.lock irq_context: 0 (wq_completion)pm (work_completion)(&dev->power.work) &cfs_rq->removed.lock irq_context: 0 &dev->mutex (usb_notifier_list).rwsem &p->pi_lock &rq->__lock &cfs_rq->removed.lock irq_context: 0 &dev->mutex usb_bus_idr_lock &dev->mutex fill_pool_map-wait-type-override pool_lock#2 irq_context: 0 &dev->mutex usb_bus_idr_lock &dev->mutex fill_pool_map-wait-type-override pool_lock irq_context: 0 &dev->mutex usb_bus_idr_lock &dev->mutex &dev->mutex fill_pool_map-wait-type-override pool_lock#2 irq_context: 0 &dev->mutex usb_bus_idr_lock &dev->mutex &dev->mutex fill_pool_map-wait-type-override pool_lock irq_context: 0 &dev->mutex usb_bus_idr_lock &dev->mutex &dev->mutex &cfs_rq->removed.lock irq_context: 0 &dev->mutex usb_bus_idr_lock &dev->mutex &dev->mutex usb_port_peer_mutex &dev->power.lock rcu_read_lock &pool->lock fill_pool_map-wait-type-override pool_lock#2 irq_context: 0 &dev->mutex usb_bus_idr_lock &dev->mutex &dev->mutex usb_port_peer_mutex &dev->power.lock rcu_read_lock &pool->lock fill_pool_map-wait-type-override pool_lock irq_context: 0 (wq_completion)events_power_efficient (work_completion)(&(&hub->init_work)->work) &dev->mutex &hub->status_mutex fill_pool_map-wait-type-override pool_lock#2 irq_context: 0 (wq_completion)events_power_efficient (work_completion)(&(&hub->init_work)->work) &dev->mutex &hub->status_mutex fill_pool_map-wait-type-override &c->lock irq_context: 0 (wq_completion)events_power_efficient (work_completion)(&(&hub->init_work)->work) &dev->mutex &hub->status_mutex fill_pool_map-wait-type-override &pcp->lock &zone->lock irq_context: 0 (wq_completion)events_power_efficient (work_completion)(&(&hub->init_work)->work) &dev->mutex &hub->status_mutex fill_pool_map-wait-type-override &zone->lock irq_context: 0 (wq_completion)events_power_efficient (work_completion)(&(&hub->init_work)->work) &dev->mutex &hub->status_mutex fill_pool_map-wait-type-override &____s->seqcount irq_context: 0 (wq_completion)events_power_efficient (work_completion)(&(&hub->init_work)->work) &dev->mutex &hub->status_mutex fill_pool_map-wait-type-override pool_lock irq_context: 0 &dev->mutex usb_bus_idr_lock &dev->mutex &dev->mutex usb_port_peer_mutex &dev->power.lock rcu_read_lock &pool->lock &p->pi_lock &rq->__lock &cfs_rq->removed.lock irq_context: 0 &dev->mutex usb_bus_idr_lock batched_entropy_u8.lock crngs.lock irq_context: 0 &dev->mutex (usb_notifier_list).rwsem &p->pi_lock &cfs_rq->removed.lock irq_context: 0 &dev->mutex usb_bus_idr_lock &(&priv->bus_notifier)->rwsem &pcp->lock &zone->lock &____s->seqcount irq_context: 0 &dev->mutex (usb_notifier_list).rwsem gdp_mutex &c->lock irq_context: 0 &dev->mutex (usb_notifier_list).rwsem gdp_mutex &____s->seqcount irq_context: 0 &dev->mutex usb_bus_idr_lock &dev->mutex &dev->mutex usb_port_peer_mutex fill_pool_map-wait-type-override &c->lock irq_context: 0 &dev->mutex usb_bus_idr_lock &dev->mutex &dev->mutex usb_port_peer_mutex fill_pool_map-wait-type-override &____s->seqcount irq_context: 0 &dev->mutex usb_bus_idr_lock &dev->mutex &dev->mutex &hub->status_mutex &c->lock irq_context: 0 &dev->mutex usb_bus_idr_lock &dev->mutex &dev->mutex &hub->status_mutex &____s->seqcount irq_context: 0 &dev->mutex usb_bus_idr_lock &dev->mutex &dev->mutex rcu_read_lock &pool->lock/1 fill_pool_map-wait-type-override pool_lock#2 irq_context: 0 &dev->mutex usb_bus_idr_lock &dev->mutex &dev->mutex rcu_read_lock &pool->lock/1 fill_pool_map-wait-type-override pool_lock irq_context: 0 &dev->mutex &dev->mutex probe_waitqueue.lock irq_context: 0 i8042_lock irq_context: 0 &dev->mutex i8042_lock irq_context: 0 &dev->mutex i8042_lock (console_sem).lock irq_context: 0 &dev->mutex i8042_lock console_lock console_srcu console_owner_lock irq_context: 0 &dev->mutex i8042_lock console_lock console_srcu console_owner irq_context: 0 &dev->mutex i8042_lock console_lock console_srcu console_owner &port_lock_key irq_context: 0 &dev->mutex i8042_lock console_lock console_srcu console_owner console_owner_lock irq_context: 0 &dev->mutex &desc->request_mutex &irq_desc_lock_class ioapic_lock irq_context: 0 &dev->mutex &desc->request_mutex &irq_desc_lock_class mask_lock tmp_mask_lock ioapic_lock irq_context: 0 &dev->mutex &desc->request_mutex &irq_desc_lock_class ioapic_lock i8259A_lock irq_context: 0 &dev->mutex &x->wait#20 irq_context: hardirq i8042_lock &x->wait#20 irq_context: 0 &dev->mutex &rq->__lock &cfs_rq->removed.lock irq_context: hardirq i8042_lock &x->wait#20 &p->pi_lock irq_context: hardirq i8042_lock &x->wait#20 &p->pi_lock &rq->__lock irq_context: hardirq i8042_lock &x->wait#20 &p->pi_lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 &dev->mutex (&timer.timer) irq_context: 0 &dev->mutex &desc->request_mutex &irq_desc_lock_class irq_resend_lock irq_context: 0 &dev->mutex &desc->request_mutex proc_subdir_lock irq_context: 0 &dev->mutex &desc->request_mutex &ent->pde_unload_lock irq_context: 0 &dev->mutex &desc->request_mutex proc_inum_ida.xa_lock irq_context: 0 &dev->mutex &desc->request_mutex &obj_hash[i].lock irq_context: 0 &dev->mutex &desc->request_mutex pool_lock#2 irq_context: 0 &dev->mutex serio_event_lock irq_context: 0 &dev->mutex serio_event_lock pool_lock#2 irq_context: 0 &dev->mutex serio_event_lock rcu_read_lock &pool->lock irq_context: 0 &dev->mutex serio_event_lock rcu_read_lock &pool->lock &obj_hash[i].lock irq_context: 0 &dev->mutex serio_event_lock rcu_read_lock &pool->lock &p->pi_lock irq_context: 0 &dev->mutex serio_event_lock rcu_read_lock &pool->lock &p->pi_lock &rq->__lock irq_context: 0 &dev->mutex serio_event_lock rcu_read_lock &pool->lock &p->pi_lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 &dev->mutex serio_event_lock rcu_read_lock &pool->lock &p->pi_lock &rq->__lock &cfs_rq->removed.lock irq_context: 0 (wq_completion)events_long irq_context: 0 (wq_completion)events_long serio_event_work irq_context: 0 (wq_completion)events_long serio_event_work serio_mutex irq_context: 0 (wq_completion)events_long serio_event_work serio_mutex serio_event_lock irq_context: 0 (wq_completion)events_long serio_event_work serio_mutex i8042_lock irq_context: 0 (wq_completion)events_long serio_event_work serio_mutex fs_reclaim irq_context: 0 (wq_completion)events_long serio_event_work serio_mutex fs_reclaim mmu_notifier_invalidate_range_start irq_context: 0 (wq_completion)events_long serio_event_work serio_mutex pool_lock#2 irq_context: 0 (wq_completion)events_long serio_event_work serio_mutex &k->list_lock irq_context: 0 (wq_completion)events_long serio_event_work serio_mutex lock irq_context: 0 (wq_completion)events_long serio_event_work serio_mutex lock kernfs_idr_lock irq_context: 0 (wq_completion)events_long serio_event_work serio_mutex &root->kernfs_rwsem irq_context: 0 (wq_completion)events_long serio_event_work serio_mutex &root->kernfs_rwsem &root->kernfs_iattr_rwsem irq_context: 0 (wq_completion)events_long serio_event_work serio_mutex &device->physical_node_lock irq_context: 0 (wq_completion)events_long serio_event_work serio_mutex &device->physical_node_lock sysfs_symlink_target_lock irq_context: 0 (wq_completion)events_long serio_event_work serio_mutex &device->physical_node_lock fs_reclaim irq_context: 0 (wq_completion)events_long serio_event_work serio_mutex &device->physical_node_lock fs_reclaim mmu_notifier_invalidate_range_start irq_context: 0 (wq_completion)events_long serio_event_work serio_mutex &device->physical_node_lock pool_lock#2 irq_context: 0 (wq_completion)events_long serio_event_work serio_mutex &device->physical_node_lock lock irq_context: 0 (wq_completion)events_long serio_event_work serio_mutex &device->physical_node_lock lock kernfs_idr_lock irq_context: 0 (wq_completion)events_long serio_event_work serio_mutex &device->physical_node_lock &root->kernfs_rwsem irq_context: 0 (wq_completion)events_long serio_event_work serio_mutex &device->physical_node_lock &root->kernfs_rwsem &root->kernfs_iattr_rwsem irq_context: 0 (wq_completion)events_long serio_event_work serio_mutex &c->lock irq_context: 0 (wq_completion)events_long serio_event_work serio_mutex &pcp->lock &zone->lock irq_context: 0 (wq_completion)events_long serio_event_work serio_mutex &zone->lock irq_context: 0 (wq_completion)events_long serio_event_work serio_mutex &____s->seqcount irq_context: 0 (wq_completion)events_long serio_event_work serio_mutex semaphore->lock irq_context: 0 (wq_completion)events_long serio_event_work serio_mutex &obj_hash[i].lock irq_context: 0 (wq_completion)events_long serio_event_work serio_mutex sysfs_symlink_target_lock irq_context: 0 (wq_completion)events_long serio_event_work serio_mutex &k->k_lock irq_context: 0 (wq_completion)events_long serio_event_work serio_mutex &root->kernfs_rwsem irq_context: 0 (wq_completion)events_long serio_event_work serio_mutex &dev->power.lock irq_context: 0 (wq_completion)events_long serio_event_work serio_mutex dpm_list_mtx irq_context: 0 (wq_completion)events_long serio_event_work serio_mutex &(&priv->bus_notifier)->rwsem irq_context: 0 (wq_completion)events_long serio_event_work serio_mutex uevent_sock_mutex irq_context: 0 (wq_completion)events_long serio_event_work serio_mutex rcu_read_lock &pool->lock/1 irq_context: 0 (wq_completion)events_long serio_event_work serio_mutex rcu_read_lock &pool->lock/1 &obj_hash[i].lock irq_context: 0 (wq_completion)events_long serio_event_work serio_mutex rcu_read_lock &pool->lock/1 pool_lock#2 irq_context: 0 (wq_completion)events_long serio_event_work serio_mutex rcu_read_lock &pool->lock/1 &p->pi_lock irq_context: 0 (wq_completion)events_long serio_event_work serio_mutex rcu_read_lock &pool->lock/1 &p->pi_lock &rq->__lock irq_context: 0 (wq_completion)events_long serio_event_work serio_mutex rcu_read_lock &pool->lock/1 &p->pi_lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)events_long serio_event_work serio_mutex rcu_read_lock &pool->lock/1 &p->pi_lock &rq->__lock &cfs_rq->removed.lock irq_context: 0 (wq_completion)events_long serio_event_work serio_mutex &dev->mutex &dev->power.lock irq_context: 0 (wq_completion)events_long serio_event_work serio_mutex &dev->mutex &k->list_lock irq_context: 0 (wq_completion)events_long serio_event_work serio_mutex &dev->mutex &k->k_lock irq_context: 0 input_ida.xa_lock irq_context: 0 input_ida.xa_lock pool_lock#2 irq_context: 0 (wq_completion)events_long serio_event_work serio_mutex subsys mutex#64 irq_context: 0 (wq_completion)events_long serio_event_work serio_mutex bus_type_sem irq_context: 0 &root->kernfs_rwsem &sem->wait_lock irq_context: 0 (wq_completion)events_long serio_event_work serio_mutex running_helpers_waitq.lock irq_context: 0 subsys mutex#30 irq_context: 0 subsys mutex#30 &k->k_lock irq_context: 0 input_mutex input_ida.xa_lock irq_context: 0 input_mutex fs_reclaim irq_context: 0 input_mutex fs_reclaim mmu_notifier_invalidate_range_start irq_context: 0 input_mutex &c->lock irq_context: 0 input_mutex &pcp->lock &zone->lock irq_context: 0 input_mutex &zone->lock irq_context: 0 input_mutex &____s->seqcount irq_context: 0 input_mutex pool_lock#2 irq_context: 0 input_mutex &x->wait#9 irq_context: 0 input_mutex &obj_hash[i].lock irq_context: 0 input_mutex &dev->mutex#2 irq_context: 0 input_mutex chrdevs_lock irq_context: 0 input_mutex &k->list_lock irq_context: 0 input_mutex lock irq_context: 0 input_mutex lock kernfs_idr_lock irq_context: 0 input_mutex &root->kernfs_rwsem irq_context: 0 input_mutex &root->kernfs_rwsem &root->kernfs_iattr_rwsem irq_context: 0 input_mutex bus_type_sem irq_context: 0 input_mutex sysfs_symlink_target_lock irq_context: 0 input_mutex &root->kernfs_rwsem irq_context: 0 input_mutex &dev->power.lock irq_context: 0 input_mutex dpm_list_mtx irq_context: 0 input_mutex req_lock irq_context: 0 input_mutex &p->pi_lock irq_context: 0 input_mutex &p->pi_lock &rq->__lock irq_context: 0 input_mutex &p->pi_lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 input_mutex &rq->__lock irq_context: 0 input_mutex &x->wait#11 irq_context: 0 input_mutex uevent_sock_mutex irq_context: 0 input_mutex rcu_read_lock &pool->lock/1 irq_context: 0 input_mutex rcu_read_lock &pool->lock/1 &obj_hash[i].lock irq_context: 0 input_mutex rcu_read_lock &pool->lock/1 pool_lock#2 irq_context: 0 input_mutex rcu_read_lock &pool->lock/1 &p->pi_lock irq_context: 0 input_mutex rcu_read_lock &pool->lock/1 &p->pi_lock &rq->__lock irq_context: 0 input_mutex rcu_read_lock &pool->lock/1 &p->pi_lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 input_mutex running_helpers_waitq.lock irq_context: 0 input_mutex &k->k_lock irq_context: 0 input_mutex subsys mutex#30 irq_context: 0 input_mutex subsys mutex#30 &k->k_lock irq_context: 0 serio_event_lock irq_context: 0 serio_event_lock pool_lock#2 irq_context: 0 serio_event_lock rcu_read_lock &pool->lock irq_context: 0 serio_event_lock rcu_read_lock &pool->lock &obj_hash[i].lock irq_context: 0 serio_event_lock rcu_read_lock &pool->lock &p->pi_lock irq_context: 0 serio_event_lock rcu_read_lock &pool->lock &p->pi_lock &rq->__lock irq_context: 0 serio_event_lock rcu_read_lock &pool->lock &p->pi_lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)events_long serio_event_work serio_mutex &dev->mutex device_links_srcu irq_context: 0 (wq_completion)events_long serio_event_work serio_mutex &dev->mutex fwnode_link_lock irq_context: 0 (wq_completion)events_long serio_event_work serio_mutex &dev->mutex device_links_lock irq_context: 0 (wq_completion)events_long serio_event_work serio_mutex &dev->mutex &(&priv->bus_notifier)->rwsem irq_context: 0 (wq_completion)events_long serio_event_work serio_mutex &dev->mutex sysfs_symlink_target_lock irq_context: 0 (wq_completion)events_long serio_event_work serio_mutex &dev->mutex fs_reclaim irq_context: 0 (wq_completion)events_long serio_event_work serio_mutex &dev->mutex fs_reclaim mmu_notifier_invalidate_range_start irq_context: 0 (wq_completion)events_long serio_event_work serio_mutex &dev->mutex pool_lock#2 irq_context: 0 (wq_completion)events_long serio_event_work serio_mutex &dev->mutex lock irq_context: 0 (wq_completion)events_long serio_event_work serio_mutex &dev->mutex lock kernfs_idr_lock irq_context: 0 (wq_completion)events_long serio_event_work serio_mutex &dev->mutex &root->kernfs_rwsem irq_context: 0 (wq_completion)events_long serio_event_work serio_mutex &dev->mutex &root->kernfs_rwsem &root->kernfs_iattr_rwsem irq_context: 0 (wq_completion)events_long serio_event_work serio_mutex &dev->mutex &serio->drv_mutex irq_context: 0 (wq_completion)events_long serio_event_work serio_mutex &dev->mutex &serio->drv_mutex fs_reclaim irq_context: 0 &new_driver->dynids.lock irq_context: 0 (wq_completion)events_long serio_event_work serio_mutex &dev->mutex &serio->drv_mutex fs_reclaim mmu_notifier_invalidate_range_start irq_context: 0 (wq_completion)events_long serio_event_work serio_mutex &dev->mutex &serio->drv_mutex pool_lock#2 irq_context: 0 (wq_completion)events_long serio_event_work serio_mutex &dev->mutex &serio->drv_mutex &x->wait#9 irq_context: 0 (wq_completion)events_long serio_event_work serio_mutex &dev->mutex &serio->drv_mutex &obj_hash[i].lock irq_context: 0 (wq_completion)events_long serio_event_work serio_mutex &dev->mutex &serio->drv_mutex &serio->lock irq_context: 0 (wq_completion)events_long serio_event_work serio_mutex &dev->mutex &serio->drv_mutex i8042_mutex irq_context: 0 (wq_completion)events_long serio_event_work serio_mutex &dev->mutex &serio->drv_mutex i8042_mutex &serio->lock irq_context: 0 (wq_completion)events_long serio_event_work serio_mutex &dev->mutex &serio->drv_mutex i8042_mutex i8042_lock irq_context: 0 (wq_completion)events_long serio_event_work serio_mutex &dev->mutex &serio->drv_mutex i8042_mutex &ps2dev->wait irq_context: 0 (wq_completion)events_long serio_event_work serio_mutex &dev->mutex &serio->drv_mutex i8042_mutex &obj_hash[i].lock irq_context: hardirq &serio->lock irq_context: 0 (wq_completion)events_long serio_event_work serio_mutex &dev->mutex &serio->drv_mutex i8042_mutex &obj_hash[i].lock pool_lock irq_context: 0 (wq_completion)events_long serio_event_work serio_mutex &dev->mutex &serio->drv_mutex i8042_mutex fill_pool_map-wait-type-override pool_lock#2 irq_context: 0 (wq_completion)events_long serio_event_work serio_mutex &dev->mutex &serio->drv_mutex i8042_mutex fill_pool_map-wait-type-override &c->lock irq_context: 0 (wq_completion)events_long serio_event_work serio_mutex &dev->mutex &serio->drv_mutex i8042_mutex fill_pool_map-wait-type-override &pcp->lock &zone->lock irq_context: 0 (wq_completion)events_long serio_event_work serio_mutex &dev->mutex &serio->drv_mutex i8042_mutex fill_pool_map-wait-type-override &zone->lock irq_context: 0 (wq_completion)events_long serio_event_work serio_mutex &dev->mutex &serio->drv_mutex i8042_mutex fill_pool_map-wait-type-override &____s->seqcount irq_context: 0 (wq_completion)events_long serio_event_work serio_mutex &dev->mutex &serio->drv_mutex i8042_mutex fill_pool_map-wait-type-override pool_lock irq_context: 0 (wq_completion)events_long serio_event_work serio_mutex &dev->mutex &serio->drv_mutex i8042_mutex &base->lock irq_context: 0 (wq_completion)events_long serio_event_work serio_mutex &dev->mutex &serio->drv_mutex i8042_mutex &base->lock &obj_hash[i].lock irq_context: 0 (wq_completion)events_long serio_event_work serio_mutex &dev->mutex &serio->drv_mutex i8042_mutex &pool->lock irq_context: 0 (wq_completion)events_long serio_event_work serio_mutex &dev->mutex &serio->drv_mutex i8042_mutex &rq->__lock irq_context: 0 (wq_completion)events_long serio_event_work serio_mutex &dev->mutex &serio->drv_mutex i8042_mutex &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 &dev->mutex rtc_ida.xa_lock irq_context: 0 &dev->mutex rtc_lock irq_context: 0 &dev->mutex &rtc->ops_lock irq_context: 0 &dev->mutex &rtc->ops_lock rtc_lock irq_context: 0 &dev->mutex chrdevs_lock irq_context: 0 &dev->mutex req_lock irq_context: 0 &dev->mutex &x->wait#11 irq_context: 0 &dev->mutex subsys mutex#27 irq_context: 0 &dev->mutex subsys mutex#27 &k->k_lock irq_context: 0 &dev->mutex subsys mutex#27 fs_reclaim irq_context: 0 &dev->mutex subsys mutex#27 fs_reclaim mmu_notifier_invalidate_range_start irq_context: 0 &dev->mutex subsys mutex#27 pool_lock#2 irq_context: 0 &dev->mutex subsys mutex#27 &x->wait#9 irq_context: 0 &dev->mutex subsys mutex#27 &obj_hash[i].lock irq_context: 0 &dev->mutex subsys mutex#27 platform_devid_ida.xa_lock irq_context: 0 &dev->mutex subsys mutex#27 &k->list_lock irq_context: 0 &dev->mutex subsys mutex#27 lock irq_context: 0 &dev->mutex subsys mutex#27 lock kernfs_idr_lock irq_context: 0 &dev->mutex subsys mutex#27 &root->kernfs_rwsem irq_context: 0 &dev->mutex subsys mutex#27 &root->kernfs_rwsem &root->kernfs_iattr_rwsem irq_context: 0 &dev->mutex subsys mutex#27 bus_type_sem irq_context: 0 &dev->mutex subsys mutex#27 sysfs_symlink_target_lock irq_context: 0 &dev->mutex subsys mutex#27 &root->kernfs_rwsem irq_context: 0 &dev->mutex subsys mutex#27 &dev->power.lock irq_context: 0 &dev->mutex subsys mutex#27 dpm_list_mtx irq_context: 0 &dev->mutex subsys mutex#27 &(&priv->bus_notifier)->rwsem irq_context: 0 &dev->mutex subsys mutex#27 uevent_sock_mutex irq_context: 0 &dev->mutex subsys mutex#27 rcu_read_lock &pool->lock/1 irq_context: 0 &dev->mutex subsys mutex#27 rcu_read_lock &pool->lock/1 &obj_hash[i].lock irq_context: 0 &dev->mutex subsys mutex#27 rcu_read_lock &pool->lock/1 pool_lock#2 irq_context: 0 &dev->mutex subsys mutex#27 rcu_read_lock &pool->lock/1 &p->pi_lock irq_context: 0 &dev->mutex subsys mutex#27 rcu_read_lock &pool->lock/1 &p->pi_lock &rq->__lock irq_context: 0 &dev->mutex subsys mutex#27 rcu_read_lock &pool->lock/1 &p->pi_lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 &dev->mutex subsys mutex#27 running_helpers_waitq.lock irq_context: 0 &dev->mutex subsys mutex#27 &dev->mutex &dev->power.lock irq_context: 0 &dev->mutex subsys mutex#27 &dev->mutex &k->list_lock irq_context: 0 &dev->mutex subsys mutex#27 &dev->mutex &k->k_lock irq_context: 0 &dev->mutex subsys mutex#27 &dev->mutex device_links_srcu irq_context: 0 &dev->mutex subsys mutex#27 &dev->mutex fwnode_link_lock irq_context: 0 &dev->mutex subsys mutex#27 &dev->mutex device_links_lock irq_context: 0 &dev->mutex subsys mutex#27 &dev->mutex &(&priv->bus_notifier)->rwsem irq_context: 0 &dev->mutex subsys mutex#27 &dev->mutex sysfs_symlink_target_lock irq_context: 0 &dev->mutex subsys mutex#27 &dev->mutex fs_reclaim irq_context: 0 &dev->mutex subsys mutex#27 &dev->mutex fs_reclaim mmu_notifier_invalidate_range_start irq_context: 0 &dev->mutex subsys mutex#27 &dev->mutex pool_lock#2 irq_context: 0 &dev->mutex subsys mutex#27 &dev->mutex lock irq_context: 0 &dev->mutex subsys mutex#27 &dev->mutex lock kernfs_idr_lock irq_context: 0 &dev->mutex subsys mutex#27 &dev->mutex &root->kernfs_rwsem irq_context: 0 &dev->mutex subsys mutex#27 &dev->mutex &root->kernfs_rwsem &root->kernfs_iattr_rwsem irq_context: 0 &dev->mutex subsys mutex#27 &dev->mutex &c->lock irq_context: 0 &dev->mutex subsys mutex#27 &dev->mutex &____s->seqcount irq_context: 0 &dev->mutex subsys mutex#27 &dev->mutex deferred_probe_mutex irq_context: 0 &dev->mutex subsys mutex#27 &dev->mutex uevent_sock_mutex irq_context: 0 &dev->mutex subsys mutex#27 &dev->mutex &obj_hash[i].lock irq_context: 0 &dev->mutex subsys mutex#27 &dev->mutex &obj_hash[i].lock pool_lock irq_context: 0 &dev->mutex subsys mutex#27 &dev->mutex rcu_read_lock &pool->lock/1 irq_context: 0 &dev->mutex subsys mutex#27 &dev->mutex rcu_read_lock &pool->lock/1 &obj_hash[i].lock irq_context: 0 &dev->mutex subsys mutex#27 &dev->mutex rcu_read_lock &pool->lock/1 pool_lock#2 irq_context: 0 &dev->mutex subsys mutex#27 &dev->mutex rcu_read_lock &pool->lock/1 &p->pi_lock irq_context: 0 &dev->mutex subsys mutex#27 &dev->mutex rcu_read_lock &pool->lock/1 &p->pi_lock &rq->__lock irq_context: 0 &dev->mutex subsys mutex#27 &dev->mutex rcu_read_lock &pool->lock/1 &p->pi_lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 &dev->mutex subsys mutex#27 &dev->mutex running_helpers_waitq.lock irq_context: 0 &dev->mutex subsys mutex#27 &dev->mutex probe_waitqueue.lock irq_context: 0 &dev->mutex subsys mutex#27 subsys mutex#3 irq_context: 0 &dev->mutex subsys mutex#27 &c->lock irq_context: 0 &dev->mutex subsys mutex#27 &____s->seqcount irq_context: 0 &dev->mutex subsys mutex#27 rcu_read_lock &pool->lock/1 &p->pi_lock &rq->__lock &cfs_rq->removed.lock irq_context: 0 &dev->mutex subsys mutex#27 &rq->__lock irq_context: 0 &dev->mutex subsys mutex#27 wakeup_ida.xa_lock irq_context: 0 &dev->mutex subsys mutex#27 gdp_mutex irq_context: 0 &dev->mutex subsys mutex#27 gdp_mutex &k->list_lock irq_context: 0 &dev->mutex subsys mutex#27 gdp_mutex fs_reclaim irq_context: 0 &dev->mutex subsys mutex#27 gdp_mutex fs_reclaim mmu_notifier_invalidate_range_start irq_context: 0 &dev->mutex subsys mutex#27 gdp_mutex pool_lock#2 irq_context: 0 &dev->mutex subsys mutex#27 gdp_mutex lock irq_context: 0 &dev->mutex subsys mutex#27 gdp_mutex lock kernfs_idr_lock irq_context: 0 &dev->mutex subsys mutex#27 gdp_mutex &root->kernfs_rwsem irq_context: 0 &dev->mutex subsys mutex#27 gdp_mutex &root->kernfs_rwsem &root->kernfs_iattr_rwsem irq_context: 0 &dev->mutex subsys mutex#27 &pcp->lock &zone->lock irq_context: 0 &dev->mutex subsys mutex#27 &zone->lock irq_context: 0 &dev->mutex subsys mutex#27 rcu_read_lock pool_lock#2 irq_context: 0 &dev->mutex subsys mutex#27 subsys mutex#15 irq_context: 0 &dev->mutex subsys mutex#27 subsys mutex#15 &k->k_lock irq_context: 0 &dev->mutex subsys mutex#27 events_lock irq_context: 0 &dev->mutex subsys mutex#27 rtcdev_lock irq_context: 0 &root->kernfs_rwsem &cfs_rq->removed.lock irq_context: 0 g_smscore_deviceslock irq_context: 0 g_smscore_deviceslock fs_reclaim irq_context: 0 g_smscore_deviceslock fs_reclaim mmu_notifier_invalidate_range_start irq_context: 0 g_smscore_deviceslock pool_lock#2 irq_context: 0 &sig->cred_guard_mutex &sb->s_type->i_mutex_key &rq->__lock irq_context: 0 &sig->cred_guard_mutex &sb->s_type->i_mutex_key &cfs_rq->removed.lock irq_context: 0 (wq_completion)events_unbound (work_completion)(&(&kfence_timer)->work) cpu_hotplug_lock jump_label_mutex text_mutex &cfs_rq->removed.lock irq_context: 0 (wq_completion)events_unbound (work_completion)(&(&kfence_timer)->work) cpu_hotplug_lock jump_label_mutex text_mutex &obj_hash[i].lock irq_context: 0 (wq_completion)events_long serio_event_work serio_mutex &dev->mutex &serio->drv_mutex i8042_mutex &cfs_rq->removed.lock irq_context: 0 (wq_completion)events_long serio_event_work serio_mutex &dev->mutex &serio->drv_mutex i8042_mutex pool_lock#2 irq_context: 0 (wq_completion)events_long serio_event_work serio_mutex &dev->mutex &serio->drv_mutex i8042_mutex (&timer.timer) irq_context: hardirq &serio->lock &ps2dev->wait irq_context: hardirq &serio->lock &ps2dev->wait &p->pi_lock irq_context: hardirq &serio->lock &ps2dev->wait &p->pi_lock &rq->__lock irq_context: hardirq &serio->lock &ps2dev->wait &p->pi_lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)events_long serio_event_work serio_mutex &dev->mutex &serio->drv_mutex i8042_mutex &rq->__lock &cfs_rq->removed.lock irq_context: 0 (wq_completion)events_long serio_event_work serio_mutex &dev->mutex &serio->drv_mutex &k->list_lock irq_context: 0 (wq_completion)events_long serio_event_work serio_mutex &dev->mutex &serio->drv_mutex gdp_mutex irq_context: 0 (wq_completion)events_long serio_event_work serio_mutex &dev->mutex &serio->drv_mutex gdp_mutex &k->list_lock irq_context: 0 (wq_completion)events_long serio_event_work serio_mutex &dev->mutex &serio->drv_mutex gdp_mutex fs_reclaim irq_context: 0 (wq_completion)events_long serio_event_work serio_mutex &dev->mutex &serio->drv_mutex gdp_mutex fs_reclaim mmu_notifier_invalidate_range_start irq_context: 0 (wq_completion)events_long serio_event_work serio_mutex &dev->mutex &serio->drv_mutex gdp_mutex pool_lock#2 irq_context: 0 (wq_completion)events_long serio_event_work serio_mutex &dev->mutex &serio->drv_mutex gdp_mutex lock irq_context: 0 (wq_completion)events_long serio_event_work serio_mutex &dev->mutex &serio->drv_mutex gdp_mutex lock kernfs_idr_lock irq_context: 0 (wq_completion)events_long serio_event_work serio_mutex &dev->mutex &serio->drv_mutex gdp_mutex &root->kernfs_rwsem irq_context: 0 (wq_completion)events_long serio_event_work serio_mutex &dev->mutex &serio->drv_mutex gdp_mutex &root->kernfs_rwsem &root->kernfs_iattr_rwsem irq_context: 0 (wq_completion)events_long serio_event_work serio_mutex &dev->mutex &serio->drv_mutex lock irq_context: 0 (wq_completion)events_long serio_event_work serio_mutex &dev->mutex &serio->drv_mutex lock kernfs_idr_lock irq_context: 0 (wq_completion)events_long serio_event_work serio_mutex &dev->mutex &serio->drv_mutex &root->kernfs_rwsem irq_context: 0 (wq_completion)events_long serio_event_work serio_mutex &dev->mutex &serio->drv_mutex &root->kernfs_rwsem &root->kernfs_iattr_rwsem irq_context: 0 (wq_completion)events_long serio_event_work serio_mutex &dev->mutex &serio->drv_mutex bus_type_sem irq_context: 0 (wq_completion)events_long serio_event_work serio_mutex &dev->mutex &serio->drv_mutex sysfs_symlink_target_lock irq_context: 0 (wq_completion)events_long serio_event_work serio_mutex &dev->mutex &serio->drv_mutex &c->lock irq_context: 0 (wq_completion)events_long serio_event_work serio_mutex &dev->mutex &serio->drv_mutex &pcp->lock &zone->lock irq_context: 0 (wq_completion)events_long serio_event_work serio_mutex &dev->mutex &serio->drv_mutex &zone->lock irq_context: 0 (wq_completion)events_long serio_event_work serio_mutex &dev->mutex &serio->drv_mutex &____s->seqcount irq_context: 0 (wq_completion)events_long serio_event_work serio_mutex &dev->mutex &serio->drv_mutex &root->kernfs_rwsem irq_context: 0 (wq_completion)events_long serio_event_work serio_mutex &dev->mutex &serio->drv_mutex rcu_read_lock pool_lock#2 irq_context: 0 (wq_completion)events_long serio_event_work serio_mutex &dev->mutex &serio->drv_mutex &dev->power.lock irq_context: 0 (wq_completion)events_long serio_event_work serio_mutex &dev->mutex &serio->drv_mutex dpm_list_mtx irq_context: 0 (wq_completion)events_long serio_event_work serio_mutex &dev->mutex &serio->drv_mutex uevent_sock_mutex irq_context: 0 (wq_completion)events_long serio_event_work serio_mutex &dev->mutex &serio->drv_mutex rcu_read_lock &pool->lock/1 irq_context: 0 (wq_completion)events_long serio_event_work serio_mutex &dev->mutex &serio->drv_mutex rcu_read_lock &pool->lock/1 &obj_hash[i].lock irq_context: 0 (wq_completion)events_long serio_event_work serio_mutex &dev->mutex &serio->drv_mutex rcu_read_lock &pool->lock/1 pool_lock#2 irq_context: 0 (wq_completion)events_long serio_event_work serio_mutex &dev->mutex &serio->drv_mutex rcu_read_lock &pool->lock/1 &p->pi_lock irq_context: 0 (wq_completion)events_long serio_event_work serio_mutex &dev->mutex &serio->drv_mutex running_helpers_waitq.lock irq_context: 0 (wq_completion)events_long serio_event_work serio_mutex &dev->mutex &serio->drv_mutex &k->k_lock irq_context: 0 (wq_completion)events_long serio_event_work serio_mutex &dev->mutex &serio->drv_mutex subsys mutex#30 irq_context: 0 (wq_completion)events_long serio_event_work serio_mutex &dev->mutex &serio->drv_mutex subsys mutex#30 &k->k_lock irq_context: 0 (wq_completion)events_long serio_event_work serio_mutex &dev->mutex &serio->drv_mutex (console_sem).lock irq_context: 0 (wq_completion)events_long serio_event_work serio_mutex &dev->mutex &serio->drv_mutex console_lock console_srcu console_owner_lock irq_context: 0 (wq_completion)events_long serio_event_work serio_mutex &dev->mutex &serio->drv_mutex console_lock console_srcu console_owner irq_context: 0 (wq_completion)events_long serio_event_work serio_mutex &dev->mutex &serio->drv_mutex console_lock console_srcu console_owner &port_lock_key irq_context: 0 (wq_completion)events_long serio_event_work serio_mutex &dev->mutex &serio->drv_mutex console_lock console_srcu console_owner console_owner_lock irq_context: 0 (wq_completion)events_long serio_event_work serio_mutex &dev->mutex &serio->drv_mutex input_mutex irq_context: 0 (wq_completion)events_long serio_event_work serio_mutex &dev->mutex &serio->drv_mutex input_mutex fs_reclaim irq_context: 0 (wq_completion)events_long serio_event_work serio_mutex &dev->mutex &serio->drv_mutex input_mutex fs_reclaim mmu_notifier_invalidate_range_start irq_context: 0 (wq_completion)events_long serio_event_work serio_mutex &dev->mutex &serio->drv_mutex input_mutex pool_lock#2 irq_context: 0 (wq_completion)events_long serio_event_work serio_mutex &dev->mutex &serio->drv_mutex input_mutex &dev->mutex#2 irq_context: 0 (wq_completion)events_long serio_event_work serio_mutex &dev->mutex &serio->drv_mutex input_mutex &k->list_lock irq_context: 0 (wq_completion)events_long serio_event_work serio_mutex &dev->mutex &serio->drv_mutex input_mutex &k->k_lock irq_context: 0 (wq_completion)events_long serio_event_work serio_mutex &dev->mutex &serio->drv_mutex input_mutex &led_cdev->led_access irq_context: 0 (wq_completion)events_long serio_event_work serio_mutex &dev->mutex &serio->drv_mutex input_mutex &led_cdev->led_access fs_reclaim irq_context: 0 (wq_completion)events_long serio_event_work serio_mutex &dev->mutex &serio->drv_mutex input_mutex &led_cdev->led_access fs_reclaim mmu_notifier_invalidate_range_start irq_context: 0 (wq_completion)events_long serio_event_work serio_mutex &dev->mutex &serio->drv_mutex input_mutex &led_cdev->led_access pool_lock#2 irq_context: 0 (wq_completion)events_long serio_event_work serio_mutex &dev->mutex &serio->drv_mutex input_mutex &led_cdev->led_access &x->wait#9 irq_context: 0 (wq_completion)events_long serio_event_work serio_mutex &dev->mutex &serio->drv_mutex input_mutex &led_cdev->led_access &obj_hash[i].lock irq_context: 0 (wq_completion)events_long serio_event_work serio_mutex &dev->mutex &serio->drv_mutex input_mutex &led_cdev->led_access &k->list_lock irq_context: 0 (wq_completion)events_long serio_event_work serio_mutex &dev->mutex &serio->drv_mutex input_mutex &led_cdev->led_access lock irq_context: 0 (wq_completion)events_long serio_event_work serio_mutex &dev->mutex &serio->drv_mutex input_mutex &led_cdev->led_access lock kernfs_idr_lock irq_context: 0 (wq_completion)events_long serio_event_work serio_mutex &dev->mutex &serio->drv_mutex input_mutex &led_cdev->led_access &root->kernfs_rwsem irq_context: 0 (wq_completion)events_long serio_event_work serio_mutex &dev->mutex &serio->drv_mutex input_mutex &led_cdev->led_access &root->kernfs_rwsem &root->kernfs_iattr_rwsem irq_context: 0 (wq_completion)events_long serio_event_work serio_mutex &dev->mutex &serio->drv_mutex input_mutex &led_cdev->led_access bus_type_sem irq_context: 0 (wq_completion)events_long serio_event_work serio_mutex &dev->mutex &serio->drv_mutex input_mutex &led_cdev->led_access sysfs_symlink_target_lock irq_context: 0 (wq_completion)events_long serio_event_work serio_mutex &dev->mutex &serio->drv_mutex input_mutex &led_cdev->led_access &c->lock irq_context: 0 (wq_completion)events_long serio_event_work serio_mutex &dev->mutex &serio->drv_mutex input_mutex &led_cdev->led_access &pcp->lock &zone->lock irq_context: 0 (wq_completion)events_long serio_event_work serio_mutex &dev->mutex &serio->drv_mutex input_mutex &led_cdev->led_access &zone->lock irq_context: 0 (wq_completion)events_long serio_event_work serio_mutex &dev->mutex &serio->drv_mutex input_mutex &led_cdev->led_access &____s->seqcount irq_context: 0 (wq_completion)events_long serio_event_work serio_mutex &dev->mutex &serio->drv_mutex input_mutex &led_cdev->led_access &root->kernfs_rwsem irq_context: 0 (wq_completion)events_long serio_event_work serio_mutex &dev->mutex &serio->drv_mutex input_mutex &led_cdev->led_access &dev->power.lock irq_context: 0 (wq_completion)events_long serio_event_work serio_mutex &dev->mutex &serio->drv_mutex input_mutex &led_cdev->led_access dpm_list_mtx irq_context: 0 (wq_completion)events_long serio_event_work serio_mutex &dev->mutex &serio->drv_mutex input_mutex &led_cdev->led_access uevent_sock_mutex irq_context: 0 (wq_completion)events_long serio_event_work serio_mutex &dev->mutex &serio->drv_mutex input_mutex &led_cdev->led_access rcu_read_lock &pool->lock/1 irq_context: 0 (wq_completion)events_long serio_event_work serio_mutex &dev->mutex &serio->drv_mutex input_mutex &led_cdev->led_access rcu_read_lock &pool->lock/1 &obj_hash[i].lock irq_context: 0 (wq_completion)events_long serio_event_work serio_mutex &dev->mutex &serio->drv_mutex input_mutex &led_cdev->led_access rcu_read_lock &pool->lock/1 pool_lock#2 irq_context: 0 (wq_completion)events_long serio_event_work serio_mutex &dev->mutex &serio->drv_mutex input_mutex &led_cdev->led_access rcu_read_lock &pool->lock/1 &p->pi_lock irq_context: 0 (wq_completion)events_long serio_event_work serio_mutex &dev->mutex &serio->drv_mutex input_mutex &led_cdev->led_access running_helpers_waitq.lock irq_context: 0 (wq_completion)events_long serio_event_work serio_mutex &dev->mutex &serio->drv_mutex input_mutex &led_cdev->led_access &k->k_lock irq_context: 0 (wq_completion)events_long serio_event_work serio_mutex &dev->mutex &serio->drv_mutex input_mutex &led_cdev->led_access subsys mutex#65 irq_context: 0 (wq_completion)events_long serio_event_work serio_mutex &dev->mutex &serio->drv_mutex input_mutex &led_cdev->led_access subsys mutex#65 &k->k_lock irq_context: 0 (wq_completion)events_long serio_event_work serio_mutex &dev->mutex &serio->drv_mutex input_mutex &led_cdev->led_access leds_list_lock irq_context: 0 (wq_completion)events_long serio_event_work serio_mutex &dev->mutex &serio->drv_mutex input_mutex &led_cdev->led_access triggers_list_lock irq_context: 0 (wq_completion)events_long serio_event_work serio_mutex &dev->mutex &serio->drv_mutex input_mutex &led_cdev->led_access triggers_list_lock &led_cdev->trigger_lock irq_context: 0 (wq_completion)events_long serio_event_work serio_mutex &dev->mutex &serio->drv_mutex input_mutex &led_cdev->led_access triggers_list_lock &led_cdev->trigger_lock fs_reclaim irq_context: 0 (wq_completion)events_long serio_event_work serio_mutex &dev->mutex &serio->drv_mutex input_mutex &led_cdev->led_access triggers_list_lock &led_cdev->trigger_lock fs_reclaim mmu_notifier_invalidate_range_start irq_context: 0 (wq_completion)events_long serio_event_work serio_mutex &dev->mutex &serio->drv_mutex input_mutex &led_cdev->led_access triggers_list_lock &led_cdev->trigger_lock pool_lock#2 irq_context: 0 (wq_completion)events_long serio_event_work serio_mutex &dev->mutex &serio->drv_mutex input_mutex &led_cdev->led_access triggers_list_lock &led_cdev->trigger_lock &trig->leddev_list_lock irq_context: 0 (wq_completion)events_long serio_event_work serio_mutex &dev->mutex &serio->drv_mutex input_mutex &led_cdev->led_access triggers_list_lock &led_cdev->trigger_lock rcu_read_lock &dev->event_lock#2 irq_context: 0 (wq_completion)events_long serio_event_work serio_mutex &dev->mutex &serio->drv_mutex input_mutex &led_cdev->led_access triggers_list_lock &led_cdev->trigger_lock &c->lock irq_context: 0 (wq_completion)events_long serio_event_work serio_mutex &dev->mutex &serio->drv_mutex input_mutex &led_cdev->led_access triggers_list_lock &led_cdev->trigger_lock &pcp->lock &zone->lock irq_context: 0 (wq_completion)events_long serio_event_work serio_mutex &dev->mutex &serio->drv_mutex input_mutex &led_cdev->led_access triggers_list_lock &led_cdev->trigger_lock &zone->lock irq_context: 0 (wq_completion)events_long serio_event_work serio_mutex &dev->mutex &serio->drv_mutex input_mutex &led_cdev->led_access triggers_list_lock &led_cdev->trigger_lock &____s->seqcount irq_context: 0 (wq_completion)events_long serio_event_work serio_mutex &dev->mutex &serio->drv_mutex input_mutex &led_cdev->led_access triggers_list_lock &led_cdev->trigger_lock uevent_sock_mutex irq_context: 0 (wq_completion)events_long serio_event_work serio_mutex &dev->mutex &serio->drv_mutex input_mutex &led_cdev->led_access triggers_list_lock &led_cdev->trigger_lock &obj_hash[i].lock irq_context: 0 (wq_completion)events_long serio_event_work serio_mutex &dev->mutex &serio->drv_mutex input_mutex &led_cdev->led_access triggers_list_lock &led_cdev->trigger_lock rcu_read_lock &pool->lock/1 irq_context: 0 (wq_completion)events_long serio_event_work serio_mutex &dev->mutex &serio->drv_mutex input_mutex &led_cdev->led_access triggers_list_lock &led_cdev->trigger_lock rcu_read_lock &pool->lock/1 &obj_hash[i].lock irq_context: 0 (wq_completion)events_long serio_event_work serio_mutex &dev->mutex &serio->drv_mutex input_mutex &led_cdev->led_access triggers_list_lock &led_cdev->trigger_lock rcu_read_lock &pool->lock/1 pool_lock#2 irq_context: 0 (wq_completion)events_long serio_event_work serio_mutex &dev->mutex &serio->drv_mutex input_mutex &led_cdev->led_access triggers_list_lock &led_cdev->trigger_lock rcu_read_lock &pool->lock/1 &p->pi_lock irq_context: 0 (wq_completion)events_long serio_event_work serio_mutex &dev->mutex &serio->drv_mutex input_mutex &led_cdev->led_access triggers_list_lock &led_cdev->trigger_lock running_helpers_waitq.lock irq_context: 0 (wq_completion)events_long serio_event_work serio_mutex &dev->mutex &serio->drv_mutex input_mutex &led_cdev->led_access &obj_hash[i].lock pool_lock irq_context: 0 (wq_completion)events_long serio_event_work serio_mutex &dev->mutex &serio->drv_mutex input_mutex &led_cdev->led_access fill_pool_map-wait-type-override pool_lock#2 irq_context: 0 (wq_completion)events_long serio_event_work serio_mutex &dev->mutex &serio->drv_mutex input_mutex &led_cdev->led_access fill_pool_map-wait-type-override pool_lock irq_context: 0 (wq_completion)events_long serio_event_work serio_mutex &dev->mutex &serio->drv_mutex input_mutex input_ida.xa_lock irq_context: 0 (wq_completion)events_long serio_event_work serio_mutex &dev->mutex &serio->drv_mutex input_mutex &x->wait#9 irq_context: 0 (wq_completion)events_long serio_event_work serio_mutex &dev->mutex &serio->drv_mutex input_mutex &obj_hash[i].lock irq_context: 0 (wq_completion)events_long serio_event_work serio_mutex &dev->mutex &serio->drv_mutex input_mutex chrdevs_lock irq_context: 0 (wq_completion)events_long serio_event_work serio_mutex &dev->mutex &serio->drv_mutex input_mutex &c->lock irq_context: 0 (wq_completion)events_long serio_event_work serio_mutex &dev->mutex &serio->drv_mutex input_mutex &pcp->lock &zone->lock irq_context: 0 (wq_completion)events_long serio_event_work serio_mutex &dev->mutex &serio->drv_mutex input_mutex &zone->lock irq_context: 0 (wq_completion)events_long serio_event_work serio_mutex &dev->mutex &serio->drv_mutex input_mutex &____s->seqcount irq_context: 0 (wq_completion)events_long serio_event_work serio_mutex &dev->mutex &serio->drv_mutex input_mutex lock irq_context: 0 (wq_completion)events_long serio_event_work serio_mutex &dev->mutex &serio->drv_mutex input_mutex lock kernfs_idr_lock irq_context: 0 (wq_completion)events_long serio_event_work serio_mutex &dev->mutex &serio->drv_mutex input_mutex &root->kernfs_rwsem irq_context: 0 (wq_completion)events_long serio_event_work serio_mutex &dev->mutex &serio->drv_mutex input_mutex &root->kernfs_rwsem &root->kernfs_iattr_rwsem irq_context: 0 (wq_completion)events_long serio_event_work serio_mutex &dev->mutex &serio->drv_mutex input_mutex bus_type_sem irq_context: 0 (wq_completion)events_long serio_event_work serio_mutex &dev->mutex &serio->drv_mutex input_mutex sysfs_symlink_target_lock irq_context: 0 (wq_completion)events_long serio_event_work serio_mutex &dev->mutex &serio->drv_mutex input_mutex &root->kernfs_rwsem irq_context: 0 (wq_completion)events_long serio_event_work serio_mutex &dev->mutex &serio->drv_mutex input_mutex lock kernfs_idr_lock pool_lock#2 irq_context: 0 (wq_completion)events_long serio_event_work serio_mutex &dev->mutex &serio->drv_mutex input_mutex &dev->power.lock irq_context: 0 (wq_completion)events_long serio_event_work serio_mutex &dev->mutex &serio->drv_mutex input_mutex dpm_list_mtx irq_context: 0 (wq_completion)events_long serio_event_work serio_mutex &dev->mutex &serio->drv_mutex input_mutex rcu_read_lock pool_lock#2 irq_context: 0 (wq_completion)events_long serio_event_work serio_mutex &dev->mutex &serio->drv_mutex input_mutex req_lock irq_context: 0 (wq_completion)events_long serio_event_work serio_mutex &dev->mutex &serio->drv_mutex input_mutex &p->pi_lock irq_context: 0 (wq_completion)events_long serio_event_work serio_mutex &dev->mutex &serio->drv_mutex input_mutex &p->pi_lock &rq->__lock irq_context: 0 (wq_completion)events_long serio_event_work serio_mutex &dev->mutex &serio->drv_mutex input_mutex &p->pi_lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)events_long serio_event_work serio_mutex &dev->mutex &serio->drv_mutex input_mutex &rq->__lock irq_context: 0 (wq_completion)events_long serio_event_work serio_mutex &dev->mutex &serio->drv_mutex input_mutex &cfs_rq->removed.lock irq_context: 0 (wq_completion)events_long serio_event_work serio_mutex &dev->mutex &serio->drv_mutex input_mutex &x->wait#11 irq_context: 0 (wq_completion)events_long serio_event_work serio_mutex &dev->mutex &serio->drv_mutex input_mutex uevent_sock_mutex irq_context: 0 (wq_completion)events_long serio_event_work serio_mutex &dev->mutex &serio->drv_mutex input_mutex fill_pool_map-wait-type-override pool_lock#2 irq_context: 0 (wq_completion)events_long serio_event_work serio_mutex &dev->mutex &serio->drv_mutex input_mutex fill_pool_map-wait-type-override pool_lock irq_context: 0 (wq_completion)events_long serio_event_work serio_mutex &dev->mutex &serio->drv_mutex input_mutex rcu_read_lock &pool->lock/1 irq_context: 0 (wq_completion)events_long serio_event_work serio_mutex &dev->mutex &serio->drv_mutex input_mutex rcu_read_lock &pool->lock/1 &obj_hash[i].lock irq_context: 0 (wq_completion)events_long serio_event_work serio_mutex &dev->mutex &serio->drv_mutex input_mutex rcu_read_lock &pool->lock/1 pool_lock#2 irq_context: 0 (wq_completion)events_long serio_event_work serio_mutex &dev->mutex &serio->drv_mutex input_mutex rcu_read_lock &pool->lock/1 &p->pi_lock irq_context: 0 (wq_completion)events_long serio_event_work serio_mutex &dev->mutex &serio->drv_mutex input_mutex rcu_read_lock &pool->lock/1 &p->pi_lock &rq->__lock irq_context: 0 (wq_completion)events_long serio_event_work serio_mutex &dev->mutex &serio->drv_mutex input_mutex rcu_read_lock &pool->lock/1 &p->pi_lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)events_long serio_event_work serio_mutex &dev->mutex &serio->drv_mutex input_mutex rcu_read_lock &pool->lock/1 &p->pi_lock &rq->__lock &cfs_rq->removed.lock irq_context: 0 (wq_completion)events_long serio_event_work serio_mutex &dev->mutex &serio->drv_mutex input_mutex subsys mutex#30 irq_context: 0 (wq_completion)events_long serio_event_work serio_mutex &dev->mutex &serio->drv_mutex input_mutex subsys mutex#30 &k->k_lock irq_context: 0 (wq_completion)events_long serio_event_work serio_mutex &dev->mutex &serio->drv_mutex input_mutex input_devices_poll_wait.lock irq_context: 0 (wq_completion)events_long serio_event_work serio_mutex &dev->mutex &c->lock irq_context: 0 (wq_completion)events_long serio_event_work serio_mutex &dev->mutex &pcp->lock &zone->lock irq_context: 0 (wq_completion)events_long serio_event_work serio_mutex &dev->mutex &zone->lock irq_context: 0 (wq_completion)events_long serio_event_work serio_mutex &dev->mutex &____s->seqcount irq_context: 0 (wq_completion)events_long serio_event_work serio_mutex &dev->mutex &root->kernfs_rwsem kernfs_idr_lock irq_context: 0 (wq_completion)events_long serio_event_work serio_mutex &dev->mutex &root->kernfs_rwsem &obj_hash[i].lock irq_context: 0 (wq_completion)events_long serio_event_work serio_mutex &dev->mutex &root->kernfs_rwsem pool_lock#2 irq_context: 0 (wq_completion)events_long serio_event_work serio_mutex &dev->mutex deferred_probe_mutex irq_context: 0 (wq_completion)events_long serio_event_work serio_mutex &dev->mutex deferred_probe_mutex &rq->__lock irq_context: 0 (wq_completion)events_long serio_event_work serio_mutex &dev->mutex uevent_sock_mutex irq_context: 0 (wq_completion)events_long serio_event_work serio_mutex &dev->mutex &obj_hash[i].lock irq_context: 0 (wq_completion)events_long serio_event_work serio_mutex &dev->mutex rcu_read_lock &pool->lock/1 irq_context: 0 (wq_completion)events_long serio_event_work serio_mutex &dev->mutex rcu_read_lock &pool->lock/1 &obj_hash[i].lock irq_context: 0 (wq_completion)events_long serio_event_work serio_mutex &dev->mutex rcu_read_lock &pool->lock/1 pool_lock#2 irq_context: 0 (wq_completion)events_long serio_event_work serio_mutex &dev->mutex rcu_read_lock &pool->lock/1 &p->pi_lock irq_context: 0 (wq_completion)events_long serio_event_work serio_mutex &dev->mutex rcu_read_lock &pool->lock/1 &p->pi_lock &rq->__lock irq_context: 0 (wq_completion)events_long serio_event_work serio_mutex &dev->mutex rcu_read_lock &pool->lock/1 &p->pi_lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)events_long serio_event_work serio_mutex &dev->mutex running_helpers_waitq.lock irq_context: 0 (wq_completion)events_long serio_event_work serio_mutex &dev->mutex &rq->__lock irq_context: 0 cx231xx_devlist_mutex irq_context: 0 (wq_completion)events_long serio_event_work serio_mutex &dev->mutex &cfs_rq->removed.lock irq_context: 0 (wq_completion)events_long serio_event_work serio_mutex &dev->mutex probe_waitqueue.lock irq_context: 0 (wq_completion)events_long serio_event_work serio_mutex &dev->mutex &serio->drv_mutex i8042_lock irq_context: 0 em28xx_devlist_mutex irq_context: 0 (wq_completion)events_long serio_event_work serio_mutex &dev->mutex &serio->drv_mutex psmouse_mutex irq_context: 0 (wq_completion)events_long serio_event_work serio_mutex &dev->mutex &serio->drv_mutex psmouse_mutex fs_reclaim irq_context: 0 (wq_completion)events_long serio_event_work serio_mutex &dev->mutex &serio->drv_mutex psmouse_mutex fs_reclaim mmu_notifier_invalidate_range_start irq_context: 0 (wq_completion)events_long serio_event_work serio_mutex &dev->mutex &serio->drv_mutex psmouse_mutex pool_lock#2 irq_context: 0 (wq_completion)events_long serio_event_work serio_mutex &dev->mutex &serio->drv_mutex psmouse_mutex &x->wait#9 irq_context: 0 (wq_completion)events_long serio_event_work serio_mutex &dev->mutex &serio->drv_mutex psmouse_mutex &obj_hash[i].lock irq_context: 0 (wq_completion)events_long serio_event_work serio_mutex &dev->mutex &serio->drv_mutex psmouse_mutex &serio->lock irq_context: 0 (wq_completion)events_long serio_event_work serio_mutex &dev->mutex &serio->drv_mutex psmouse_mutex i8042_mutex irq_context: 0 (wq_completion)events_long serio_event_work serio_mutex &dev->mutex &serio->drv_mutex psmouse_mutex i8042_mutex &serio->lock irq_context: 0 (wq_completion)events_long serio_event_work serio_mutex &dev->mutex &serio->drv_mutex psmouse_mutex i8042_mutex i8042_lock irq_context: 0 (wq_completion)events_long serio_event_work serio_mutex &dev->mutex &serio->drv_mutex psmouse_mutex i8042_mutex &ps2dev->wait irq_context: 0 (wq_completion)events_long serio_event_work serio_mutex &dev->mutex &serio->drv_mutex psmouse_mutex i8042_mutex &obj_hash[i].lock irq_context: 0 (wq_completion)events_long serio_event_work serio_mutex &dev->mutex &serio->drv_mutex psmouse_mutex i8042_mutex &base->lock irq_context: 0 (wq_completion)events_long serio_event_work serio_mutex &dev->mutex &serio->drv_mutex psmouse_mutex i8042_mutex &base->lock &obj_hash[i].lock irq_context: 0 (wq_completion)events_long serio_event_work serio_mutex &dev->mutex &serio->drv_mutex psmouse_mutex i8042_mutex &pool->lock irq_context: 0 (wq_completion)events_long serio_event_work serio_mutex &dev->mutex &serio->drv_mutex psmouse_mutex i8042_mutex &rq->__lock irq_context: 0 (wq_completion)events_long serio_event_work serio_mutex &dev->mutex &serio->drv_mutex psmouse_mutex i8042_mutex &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)events_long serio_event_work serio_mutex &dev->mutex &serio->drv_mutex psmouse_mutex i8042_mutex (&timer.timer) irq_context: hardirq &serio->lock &ps2dev->wait &p->pi_lock &rq->__lock &cfs_rq->removed.lock irq_context: 0 (wq_completion)events_long serio_event_work serio_mutex &dev->mutex &serio->drv_mutex psmouse_mutex i8042_mutex &cfs_rq->removed.lock irq_context: 0 (wq_completion)events_long serio_event_work serio_mutex &dev->mutex &serio->drv_mutex psmouse_mutex i8042_mutex pool_lock#2 irq_context: 0 (wq_completion)events_long serio_event_work serio_mutex &dev->mutex &serio->drv_mutex psmouse_mutex i8042_mutex &rq->__lock &cfs_rq->removed.lock irq_context: hardirq &serio->lock &dev->power.lock irq_context: hardirq &serio->lock &dev->event_lock#2 irq_context: 0 pvr2_context_sync_data.lock irq_context: 0 &dev->mutex core_lock irq_context: 0 &dev->mutex core_lock &rq->__lock irq_context: 0 &dev->mutex core_lock &cfs_rq->removed.lock irq_context: 0 &dev->mutex core_lock &obj_hash[i].lock irq_context: 0 &dev->mutex core_lock pool_lock#2 irq_context: 0 &dev->mutex core_lock fs_reclaim irq_context: 0 &dev->mutex core_lock fs_reclaim mmu_notifier_invalidate_range_start irq_context: 0 &dev->mutex &(&priv->bus_notifier)->rwsem fs_reclaim irq_context: 0 &dev->mutex &(&priv->bus_notifier)->rwsem fs_reclaim mmu_notifier_invalidate_range_start irq_context: 0 &dev->mutex &(&priv->bus_notifier)->rwsem pool_lock#2 irq_context: 0 &dev->mutex &(&priv->bus_notifier)->rwsem i2c_dev_list_lock irq_context: 0 &dev->mutex &(&priv->bus_notifier)->rwsem &x->wait#9 irq_context: 0 &dev->mutex &(&priv->bus_notifier)->rwsem &obj_hash[i].lock irq_context: 0 &dev->mutex &(&priv->bus_notifier)->rwsem chrdevs_lock irq_context: 0 &dev->mutex &(&priv->bus_notifier)->rwsem &k->list_lock irq_context: 0 &dev->mutex &(&priv->bus_notifier)->rwsem gdp_mutex irq_context: 0 &dev->mutex &(&priv->bus_notifier)->rwsem gdp_mutex &k->list_lock irq_context: 0 &dev->mutex &(&priv->bus_notifier)->rwsem gdp_mutex fs_reclaim irq_context: 0 &dev->mutex &(&priv->bus_notifier)->rwsem gdp_mutex fs_reclaim mmu_notifier_invalidate_range_start irq_context: 0 &dev->mutex &(&priv->bus_notifier)->rwsem gdp_mutex pool_lock#2 irq_context: 0 &dev->mutex &(&priv->bus_notifier)->rwsem gdp_mutex lock irq_context: 0 &dev->mutex &(&priv->bus_notifier)->rwsem gdp_mutex lock kernfs_idr_lock irq_context: 0 &dev->mutex &(&priv->bus_notifier)->rwsem gdp_mutex &root->kernfs_rwsem irq_context: 0 &dev->mutex &(&priv->bus_notifier)->rwsem gdp_mutex &root->kernfs_rwsem &root->kernfs_iattr_rwsem irq_context: 0 &dev->mutex &(&priv->bus_notifier)->rwsem lock irq_context: 0 &dev->mutex &(&priv->bus_notifier)->rwsem lock kernfs_idr_lock irq_context: 0 &dev->mutex &(&priv->bus_notifier)->rwsem &root->kernfs_rwsem irq_context: 0 &dev->mutex &(&priv->bus_notifier)->rwsem &root->kernfs_rwsem &root->kernfs_iattr_rwsem irq_context: 0 &dev->mutex &(&priv->bus_notifier)->rwsem bus_type_sem irq_context: 0 &dev->mutex &(&priv->bus_notifier)->rwsem sysfs_symlink_target_lock irq_context: 0 &dev->mutex &(&priv->bus_notifier)->rwsem &root->kernfs_rwsem irq_context: 0 &dev->mutex &(&priv->bus_notifier)->rwsem &dev->power.lock irq_context: 0 &dev->mutex &(&priv->bus_notifier)->rwsem dpm_list_mtx irq_context: 0 &dev->mutex &(&priv->bus_notifier)->rwsem &c->lock irq_context: 0 &dev->mutex &(&priv->bus_notifier)->rwsem &pcp->lock &zone->lock irq_context: 0 &dev->mutex &(&priv->bus_notifier)->rwsem &zone->lock irq_context: 0 &dev->mutex &(&priv->bus_notifier)->rwsem &____s->seqcount irq_context: 0 &dev->mutex &(&priv->bus_notifier)->rwsem req_lock irq_context: 0 &dev->mutex &(&priv->bus_notifier)->rwsem &p->pi_lock irq_context: 0 &dev->mutex &(&priv->bus_notifier)->rwsem &p->pi_lock &rq->__lock irq_context: 0 &dev->mutex &(&priv->bus_notifier)->rwsem &p->pi_lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 &dev->mutex &(&priv->bus_notifier)->rwsem &p->pi_lock &rq->__lock &cfs_rq->removed.lock irq_context: 0 &dev->mutex &(&priv->bus_notifier)->rwsem &rq->__lock irq_context: 0 &dev->mutex &(&priv->bus_notifier)->rwsem &x->wait#11 irq_context: 0 &dev->mutex &(&priv->bus_notifier)->rwsem uevent_sock_mutex irq_context: 0 &dev->mutex &(&priv->bus_notifier)->rwsem rcu_read_lock &pool->lock/1 irq_context: 0 &dev->mutex &(&priv->bus_notifier)->rwsem rcu_read_lock &pool->lock/1 &obj_hash[i].lock irq_context: 0 &dev->mutex &(&priv->bus_notifier)->rwsem rcu_read_lock &pool->lock/1 pool_lock#2 irq_context: 0 &dev->mutex &(&priv->bus_notifier)->rwsem rcu_read_lock &pool->lock/1 &p->pi_lock irq_context: 0 &dev->mutex &(&priv->bus_notifier)->rwsem rcu_read_lock &pool->lock/1 &p->pi_lock &rq->__lock irq_context: 0 &dev->mutex &(&priv->bus_notifier)->rwsem rcu_read_lock &pool->lock/1 &p->pi_lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 &dev->mutex &(&priv->bus_notifier)->rwsem running_helpers_waitq.lock irq_context: 0 &dev->mutex &(&priv->bus_notifier)->rwsem &k->k_lock irq_context: 0 &dev->mutex &(&priv->bus_notifier)->rwsem subsys mutex#66 irq_context: 0 &dev->mutex &(&priv->bus_notifier)->rwsem subsys mutex#66 &k->k_lock irq_context: 0 &dev->mutex subsys mutex#67 irq_context: 0 &dev->mutex core_lock &k->list_lock irq_context: 0 &dev->mutex core_lock &k->k_lock irq_context: 0 &dev->mutex dvbdev_register_lock irq_context: 0 &dev->mutex dvbdev_register_lock (console_sem).lock irq_context: 0 &dev->mutex dvbdev_register_lock console_lock console_srcu console_owner_lock irq_context: 0 &dev->mutex dvbdev_register_lock console_lock console_srcu console_owner irq_context: 0 &dev->mutex dvbdev_register_lock console_lock console_srcu console_owner &port_lock_key irq_context: 0 &dev->mutex dvbdev_register_lock console_lock console_srcu console_owner console_owner_lock irq_context: 0 &dev->mutex (kmod_concurrent_max).lock irq_context: 0 &dev->mutex &x->wait#17 irq_context: hardirq allocation_wait.lock &p->pi_lock &rq->__lock &cfs_rq->removed.lock irq_context: 0 &dev->mutex lock kernfs_idr_lock &c->lock irq_context: 0 &dev->mutex lock kernfs_idr_lock &pcp->lock &zone->lock irq_context: 0 &dev->mutex lock kernfs_idr_lock &zone->lock irq_context: 0 &dev->mutex lock kernfs_idr_lock &____s->seqcount irq_context: 0 &dev->mutex &dev->mutex &dev->power.lock &dev->power.lock/1 irq_context: 0 &dev->mutex &dev->mutex &dev->power.lock &dev->power.wait_queue irq_context: 0 &dev->mutex &dev->mutex device_links_srcu irq_context: 0 &dev->mutex &dev->mutex fwnode_link_lock irq_context: 0 &dev->mutex &dev->mutex device_links_lock irq_context: 0 &dev->mutex &dev->mutex &(&priv->bus_notifier)->rwsem irq_context: 0 &dev->mutex &dev->mutex sysfs_symlink_target_lock irq_context: 0 &dev->mutex &dev->mutex fs_reclaim irq_context: 0 &dev->mutex &dev->mutex fs_reclaim mmu_notifier_invalidate_range_start irq_context: 0 &dev->mutex &dev->mutex pool_lock#2 irq_context: 0 &dev->mutex &dev->mutex lock irq_context: 0 &dev->mutex &dev->mutex lock kernfs_idr_lock irq_context: 0 &dev->mutex &dev->mutex &root->kernfs_rwsem irq_context: 0 &dev->mutex &dev->mutex &root->kernfs_rwsem &root->kernfs_iattr_rwsem irq_context: 0 &dev->mutex &dev->mutex &dev->devres_lock irq_context: 0 &dev->mutex &dev->mutex deferred_probe_mutex irq_context: 0 &dev->mutex &dev->mutex uevent_sock_mutex irq_context: 0 &dev->mutex &dev->mutex &obj_hash[i].lock irq_context: 0 &dev->mutex &dev->mutex rcu_read_lock &pool->lock/1 irq_context: 0 &dev->mutex &dev->mutex rcu_read_lock &pool->lock/1 &obj_hash[i].lock irq_context: 0 &dev->mutex &dev->mutex rcu_read_lock &pool->lock/1 pool_lock#2 irq_context: 0 &dev->mutex &dev->mutex rcu_read_lock &pool->lock/1 &p->pi_lock irq_context: 0 &dev->mutex &dev->mutex running_helpers_waitq.lock irq_context: 0 &dev->mutex &dev->mutex rcu_read_lock &pool->lock/1 &p->pi_lock &rq->__lock irq_context: 0 &dev->mutex &dev->mutex rcu_read_lock &pool->lock/1 &p->pi_lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 &dev->mutex &dev->mutex &rq->__lock irq_context: 0 &dev->mutex &dev->mutex &____s->seqcount irq_context: 0 &dev->mutex &dev->mutex &pcp->lock &zone->lock irq_context: 0 &dev->mutex &dev->mutex &zone->lock irq_context: 0 &dev->mutex &dev->mutex rcu_read_lock pool_lock#2 irq_context: 0 &dev->mutex frontend_mutex irq_context: 0 &dev->mutex frontend_mutex fs_reclaim irq_context: 0 &dev->mutex frontend_mutex fs_reclaim mmu_notifier_invalidate_range_start irq_context: 0 &dev->mutex frontend_mutex pool_lock#2 irq_context: 0 &dev->mutex frontend_mutex (console_sem).lock irq_context: 0 &dev->mutex frontend_mutex console_lock console_srcu console_owner_lock irq_context: 0 &dev->mutex frontend_mutex console_lock console_srcu console_owner irq_context: 0 &dev->mutex frontend_mutex console_lock console_srcu console_owner &port_lock_key irq_context: 0 &dev->mutex frontend_mutex console_lock console_srcu console_owner console_owner_lock irq_context: 0 &dev->mutex frontend_mutex dvbdev_register_lock irq_context: 0 &dev->mutex frontend_mutex dvbdev_register_lock fs_reclaim irq_context: 0 &dev->mutex frontend_mutex dvbdev_register_lock fs_reclaim mmu_notifier_invalidate_range_start irq_context: 0 &dev->mutex frontend_mutex dvbdev_register_lock pool_lock#2 irq_context: 0 &dev->mutex frontend_mutex dvbdev_register_lock minor_rwsem irq_context: 0 &dev->mutex frontend_mutex dvbdev_register_lock &xa->xa_lock#12 irq_context: 0 &dev->mutex frontend_mutex dvbdev_register_lock &mdev->graph_mutex irq_context: 0 &dev->mutex frontend_mutex dvbdev_register_lock &mdev->graph_mutex fs_reclaim irq_context: 0 &dev->mutex frontend_mutex dvbdev_register_lock &mdev->graph_mutex fs_reclaim mmu_notifier_invalidate_range_start irq_context: 0 &dev->mutex frontend_mutex dvbdev_register_lock &mdev->graph_mutex pool_lock#2 irq_context: 0 &dev->mutex frontend_mutex dvbdev_register_lock (console_sem).lock irq_context: 0 &dev->mutex frontend_mutex dvbdev_register_lock console_lock console_srcu console_owner_lock irq_context: 0 &dev->mutex frontend_mutex dvbdev_register_lock console_lock console_srcu console_owner irq_context: 0 &dev->mutex frontend_mutex dvbdev_register_lock console_lock console_srcu console_owner &port_lock_key irq_context: 0 &dev->mutex frontend_mutex dvbdev_register_lock console_lock console_srcu console_owner console_owner_lock irq_context: 0 &dev->mutex frontend_mutex dvbdev_register_lock &x->wait#9 irq_context: 0 &dev->mutex frontend_mutex dvbdev_register_lock &obj_hash[i].lock irq_context: 0 &dev->mutex frontend_mutex dvbdev_register_lock &k->list_lock irq_context: 0 &dev->mutex frontend_mutex dvbdev_register_lock gdp_mutex irq_context: 0 &dev->mutex frontend_mutex dvbdev_register_lock gdp_mutex &k->list_lock irq_context: 0 &dev->mutex frontend_mutex dvbdev_register_lock gdp_mutex fs_reclaim irq_context: 0 &dev->mutex frontend_mutex dvbdev_register_lock gdp_mutex fs_reclaim mmu_notifier_invalidate_range_start irq_context: 0 &dev->mutex frontend_mutex dvbdev_register_lock gdp_mutex pool_lock#2 irq_context: 0 &dev->mutex frontend_mutex dvbdev_register_lock gdp_mutex lock irq_context: 0 &dev->mutex frontend_mutex dvbdev_register_lock gdp_mutex lock kernfs_idr_lock irq_context: 0 &dev->mutex frontend_mutex dvbdev_register_lock gdp_mutex &root->kernfs_rwsem irq_context: 0 &dev->mutex frontend_mutex dvbdev_register_lock gdp_mutex &root->kernfs_rwsem &root->kernfs_iattr_rwsem irq_context: 0 &dev->mutex frontend_mutex dvbdev_register_lock lock irq_context: 0 &dev->mutex frontend_mutex dvbdev_register_lock lock kernfs_idr_lock irq_context: 0 &dev->mutex frontend_mutex dvbdev_register_lock &root->kernfs_rwsem irq_context: 0 &dev->mutex frontend_mutex dvbdev_register_lock &root->kernfs_rwsem &root->kernfs_iattr_rwsem irq_context: 0 &dev->mutex frontend_mutex dvbdev_register_lock bus_type_sem irq_context: 0 &dev->mutex frontend_mutex dvbdev_register_lock sysfs_symlink_target_lock irq_context: 0 &dev->mutex frontend_mutex dvbdev_register_lock &c->lock irq_context: 0 &dev->mutex frontend_mutex dvbdev_register_lock &pcp->lock &zone->lock irq_context: 0 &dev->mutex frontend_mutex dvbdev_register_lock &zone->lock irq_context: 0 &dev->mutex frontend_mutex dvbdev_register_lock &____s->seqcount irq_context: 0 &dev->mutex frontend_mutex dvbdev_register_lock &root->kernfs_rwsem irq_context: 0 &dev->mutex frontend_mutex dvbdev_register_lock &dev->power.lock irq_context: 0 &dev->mutex frontend_mutex dvbdev_register_lock dpm_list_mtx irq_context: 0 &dev->mutex frontend_mutex dvbdev_register_lock req_lock irq_context: 0 &dev->mutex frontend_mutex dvbdev_register_lock &p->pi_lock irq_context: 0 &dev->mutex frontend_mutex dvbdev_register_lock &x->wait#11 irq_context: 0 &dev->mutex frontend_mutex dvbdev_register_lock &rq->__lock irq_context: 0 &dev->mutex frontend_mutex dvbdev_register_lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 &dev->mutex frontend_mutex dvbdev_register_lock uevent_sock_mutex irq_context: 0 &dev->mutex frontend_mutex dvbdev_register_lock rcu_read_lock &pool->lock/1 irq_context: 0 &dev->mutex frontend_mutex dvbdev_register_lock rcu_read_lock &pool->lock/1 &obj_hash[i].lock irq_context: 0 &dev->mutex frontend_mutex dvbdev_register_lock rcu_read_lock &pool->lock/1 pool_lock#2 irq_context: 0 &dev->mutex frontend_mutex dvbdev_register_lock rcu_read_lock &pool->lock/1 &p->pi_lock irq_context: 0 &dev->mutex frontend_mutex dvbdev_register_lock running_helpers_waitq.lock irq_context: 0 &dev->mutex frontend_mutex dvbdev_register_lock &k->k_lock irq_context: 0 &dev->mutex frontend_mutex dvbdev_register_lock subsys mutex#68 irq_context: 0 &dev->mutex frontend_mutex dvbdev_register_lock subsys mutex#68 &k->k_lock irq_context: 0 &dev->mutex init_mm.page_table_lock irq_context: 0 &dev->mutex &dmxdev->lock irq_context: 0 &dev->mutex dvbdev_register_lock fs_reclaim irq_context: 0 &dev->mutex dvbdev_register_lock fs_reclaim mmu_notifier_invalidate_range_start irq_context: 0 &dev->mutex dvbdev_register_lock &c->lock irq_context: 0 &dev->mutex dvbdev_register_lock &____s->seqcount irq_context: 0 &dev->mutex dvbdev_register_lock pool_lock#2 irq_context: 0 &dev->mutex dvbdev_register_lock minor_rwsem irq_context: 0 &dev->mutex dvbdev_register_lock &pcp->lock &zone->lock irq_context: 0 &dev->mutex dvbdev_register_lock &zone->lock irq_context: 0 &dev->mutex dvbdev_register_lock &xa->xa_lock#12 irq_context: 0 &dev->mutex dvbdev_register_lock &mdev->graph_mutex irq_context: 0 &dev->mutex dvbdev_register_lock &xa->xa_lock#12 &c->lock irq_context: 0 &dev->mutex dvbdev_register_lock &xa->xa_lock#12 &____s->seqcount irq_context: 0 &dev->mutex dvbdev_register_lock &xa->xa_lock#12 pool_lock#2 irq_context: 0 &dev->mutex dvbdev_register_lock &mdev->graph_mutex fs_reclaim irq_context: 0 &dev->mutex dvbdev_register_lock &mdev->graph_mutex fs_reclaim mmu_notifier_invalidate_range_start irq_context: 0 &dev->mutex dvbdev_register_lock &mdev->graph_mutex pool_lock#2 irq_context: 0 &dev->mutex dvbdev_register_lock &mdev->graph_mutex &obj_hash[i].lock irq_context: 0 &dev->mutex dvbdev_register_lock &rq->__lock irq_context: 0 &dev->mutex dvbdev_register_lock rcu_read_lock pool_lock#2 irq_context: 0 &dev->mutex dvbdev_register_lock &obj_hash[i].lock irq_context: 0 &dev->mutex dvbdev_register_lock &pcp->lock &zone->lock &____s->seqcount irq_context: 0 &dev->mutex dvbdev_register_lock &x->wait#9 irq_context: 0 &dev->mutex dvbdev_register_lock &k->list_lock irq_context: 0 &dev->mutex dvbdev_register_lock gdp_mutex irq_context: 0 &dev->mutex dvbdev_register_lock gdp_mutex &k->list_lock irq_context: 0 &dev->mutex dvbdev_register_lock lock irq_context: 0 &dev->mutex dvbdev_register_lock lock kernfs_idr_lock irq_context: 0 &dev->mutex dvbdev_register_lock &root->kernfs_rwsem irq_context: 0 &dev->mutex dvbdev_register_lock &root->kernfs_rwsem &root->kernfs_iattr_rwsem irq_context: 0 &dev->mutex dvbdev_register_lock bus_type_sem irq_context: 0 &dev->mutex dvbdev_register_lock sysfs_symlink_target_lock irq_context: 0 &dev->mutex dvbdev_register_lock &root->kernfs_rwsem irq_context: 0 &dev->mutex dvbdev_register_lock &dev->power.lock irq_context: 0 &dev->mutex dvbdev_register_lock dpm_list_mtx irq_context: 0 &dev->mutex dvbdev_register_lock req_lock irq_context: 0 &dev->mutex dvbdev_register_lock &p->pi_lock irq_context: 0 &dev->mutex dvbdev_register_lock &p->pi_lock &rq->__lock irq_context: 0 &dev->mutex dvbdev_register_lock &p->pi_lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 &dev->mutex dvbdev_register_lock &x->wait#11 irq_context: 0 &dev->mutex dvbdev_register_lock uevent_sock_mutex irq_context: 0 &dev->mutex dvbdev_register_lock rcu_read_lock &pool->lock/1 irq_context: 0 &dev->mutex dvbdev_register_lock rcu_read_lock &pool->lock/1 &obj_hash[i].lock irq_context: 0 &dev->mutex dvbdev_register_lock rcu_read_lock &pool->lock/1 pool_lock#2 irq_context: 0 &dev->mutex dvbdev_register_lock rcu_read_lock &pool->lock/1 &p->pi_lock irq_context: 0 &dev->mutex dvbdev_register_lock running_helpers_waitq.lock irq_context: 0 &dev->mutex dvbdev_register_lock &k->k_lock irq_context: 0 &dev->mutex dvbdev_register_lock subsys mutex#68 irq_context: 0 &dev->mutex dvbdev_register_lock subsys mutex#68 &k->k_lock irq_context: 0 &dev->mutex dvbdev_register_lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 &dev->mutex &dvbdemux->mutex irq_context: 0 &dev->mutex media_devnode_lock irq_context: 0 &dev->mutex subsys mutex#69 irq_context: 0 &dev->mutex videodev_lock irq_context: 0 &dev->mutex subsys mutex#70 irq_context: 0 &dev->mutex subsys mutex#70 &k->k_lock irq_context: 0 &dev->mutex &xa->xa_lock#12 irq_context: 0 &dev->mutex &mdev->graph_mutex irq_context: 0 &dev->mutex &mdev->graph_mutex fs_reclaim irq_context: 0 &dev->mutex &mdev->graph_mutex fs_reclaim mmu_notifier_invalidate_range_start irq_context: 0 &dev->mutex &mdev->graph_mutex pool_lock#2 irq_context: 0 &dev->mutex vimc_sensor:393:(&vsensor->hdl)->_lock irq_context: 0 &dev->mutex &v4l2_dev->lock irq_context: 0 &dev->mutex vimc_debayer:578:(&vdebayer->hdl)->_lock irq_context: 0 &dev->mutex gdp_mutex rcu_read_lock pool_lock#2 irq_context: 0 &dev->mutex gdp_mutex &obj_hash[i].lock irq_context: 0 &dev->mutex vimc_lens:61:(&vlens->hdl)->_lock irq_context: 0 &dev->mutex tk_core.seq.seqcount irq_context: 0 &dev->mutex vivid_ctrls:1606:(hdl_user_gen)->_lock irq_context: 0 &dev->mutex vivid_ctrls:1608:(hdl_user_vid)->_lock irq_context: 0 &dev->mutex vivid_ctrls:1610:(hdl_user_aud)->_lock irq_context: 0 &dev->mutex vivid_ctrls:1612:(hdl_streaming)->_lock irq_context: 0 &dev->mutex vivid_ctrls:1614:(hdl_sdtv_cap)->_lock irq_context: 0 &dev->mutex vivid_ctrls:1616:(hdl_loop_cap)->_lock irq_context: 0 &dev->mutex vivid_ctrls:1616:(hdl_loop_cap)->_lock &rq->__lock irq_context: 0 &dev->mutex vivid_ctrls:1618:(hdl_fb)->_lock irq_context: 0 &dev->mutex vivid_ctrls:1620:(hdl_vid_cap)->_lock irq_context: 0 &dev->mutex vivid_ctrls:1622:(hdl_vid_out)->_lock irq_context: 0 &dev->mutex vivid_ctrls:1625:(hdl_vbi_cap)->_lock irq_context: 0 &dev->mutex vivid_ctrls:1627:(hdl_vbi_out)->_lock irq_context: 0 &dev->mutex vivid_ctrls:1630:(hdl_radio_rx)->_lock irq_context: 0 &dev->mutex vivid_ctrls:1632:(hdl_radio_tx)->_lock irq_context: 0 &dev->mutex vivid_ctrls:1634:(hdl_sdr_cap)->_lock irq_context: 0 &dev->mutex vivid_ctrls:1636:(hdl_meta_cap)->_lock irq_context: 0 &dev->mutex vivid_ctrls:1638:(hdl_meta_out)->_lock irq_context: 0 &dev->mutex vivid_ctrls:1640:(hdl_tch_cap)->_lock irq_context: 0 &dev->mutex vivid_ctrls:1606:(hdl_user_gen)->_lock vivid_ctrls:1620:(hdl_vid_cap)->_lock irq_context: 0 &dev->mutex vivid_ctrls:1606:(hdl_user_gen)->_lock fs_reclaim irq_context: 0 &dev->mutex vivid_ctrls:1606:(hdl_user_gen)->_lock fs_reclaim mmu_notifier_invalidate_range_start irq_context: 0 &dev->mutex vivid_ctrls:1606:(hdl_user_gen)->_lock pool_lock#2 irq_context: 0 &dev->mutex vivid_ctrls:1606:(hdl_user_gen)->_lock &c->lock irq_context: 0 &dev->mutex vivid_ctrls:1606:(hdl_user_gen)->_lock &____s->seqcount irq_context: 0 &dev->mutex vivid_ctrls:1608:(hdl_user_vid)->_lock vivid_ctrls:1620:(hdl_vid_cap)->_lock irq_context: 0 &dev->mutex vivid_ctrls:1608:(hdl_user_vid)->_lock fs_reclaim irq_context: 0 &dev->mutex vivid_ctrls:1608:(hdl_user_vid)->_lock fs_reclaim mmu_notifier_invalidate_range_start irq_context: 0 &dev->mutex vivid_ctrls:1608:(hdl_user_vid)->_lock pool_lock#2 irq_context: 0 &dev->mutex vivid_ctrls:1610:(hdl_user_aud)->_lock vivid_ctrls:1620:(hdl_vid_cap)->_lock irq_context: 0 &dev->mutex vivid_ctrls:1610:(hdl_user_aud)->_lock fs_reclaim irq_context: 0 &dev->mutex vivid_ctrls:1610:(hdl_user_aud)->_lock fs_reclaim mmu_notifier_invalidate_range_start irq_context: 0 &dev->mutex vivid_ctrls:1610:(hdl_user_aud)->_lock pool_lock#2 irq_context: 0 &dev->mutex vivid_ctrls:1612:(hdl_streaming)->_lock vivid_ctrls:1620:(hdl_vid_cap)->_lock irq_context: 0 &dev->mutex vivid_ctrls:1612:(hdl_streaming)->_lock fs_reclaim irq_context: 0 &dev->mutex vivid_ctrls:1612:(hdl_streaming)->_lock fs_reclaim mmu_notifier_invalidate_range_start irq_context: 0 &dev->mutex vivid_ctrls:1612:(hdl_streaming)->_lock pool_lock#2 irq_context: 0 &dev->mutex vivid_ctrls:1612:(hdl_streaming)->_lock &c->lock irq_context: 0 &dev->mutex vivid_ctrls:1612:(hdl_streaming)->_lock &____s->seqcount irq_context: 0 &dev->mutex vivid_ctrls:1614:(hdl_sdtv_cap)->_lock vivid_ctrls:1620:(hdl_vid_cap)->_lock irq_context: 0 &dev->mutex vivid_ctrls:1614:(hdl_sdtv_cap)->_lock fs_reclaim irq_context: 0 &dev->mutex vivid_ctrls:1614:(hdl_sdtv_cap)->_lock fs_reclaim mmu_notifier_invalidate_range_start irq_context: 0 &dev->mutex vivid_ctrls:1614:(hdl_sdtv_cap)->_lock pool_lock#2 irq_context: 0 &dev->mutex vivid_ctrls:1616:(hdl_loop_cap)->_lock vivid_ctrls:1620:(hdl_vid_cap)->_lock irq_context: 0 &dev->mutex vivid_ctrls:1616:(hdl_loop_cap)->_lock fs_reclaim irq_context: 0 &dev->mutex vivid_ctrls:1616:(hdl_loop_cap)->_lock fs_reclaim mmu_notifier_invalidate_range_start irq_context: 0 &dev->mutex vivid_ctrls:1616:(hdl_loop_cap)->_lock pool_lock#2 irq_context: 0 &dev->mutex vivid_ctrls:1606:(hdl_user_gen)->_lock vivid_ctrls:1622:(hdl_vid_out)->_lock irq_context: 0 &dev->mutex vivid_ctrls:1610:(hdl_user_aud)->_lock vivid_ctrls:1622:(hdl_vid_out)->_lock irq_context: 0 &dev->mutex vivid_ctrls:1612:(hdl_streaming)->_lock vivid_ctrls:1622:(hdl_vid_out)->_lock irq_context: 0 &dev->mutex vivid_ctrls:1606:(hdl_user_gen)->_lock vivid_ctrls:1625:(hdl_vbi_cap)->_lock irq_context: 0 &dev->mutex vivid_ctrls:1612:(hdl_streaming)->_lock vivid_ctrls:1625:(hdl_vbi_cap)->_lock irq_context: 0 &dev->mutex vivid_ctrls:1612:(hdl_streaming)->_lock &pcp->lock &zone->lock irq_context: 0 &dev->mutex vivid_ctrls:1612:(hdl_streaming)->_lock &zone->lock irq_context: 0 &dev->mutex vivid_ctrls:1614:(hdl_sdtv_cap)->_lock vivid_ctrls:1625:(hdl_vbi_cap)->_lock irq_context: 0 &dev->mutex vivid_ctrls:1616:(hdl_loop_cap)->_lock vivid_ctrls:1625:(hdl_vbi_cap)->_lock irq_context: 0 &dev->mutex vivid_ctrls:1606:(hdl_user_gen)->_lock vivid_ctrls:1627:(hdl_vbi_out)->_lock irq_context: 0 &dev->mutex vivid_ctrls:1612:(hdl_streaming)->_lock vivid_ctrls:1627:(hdl_vbi_out)->_lock irq_context: 0 &dev->mutex vivid_ctrls:1606:(hdl_user_gen)->_lock vivid_ctrls:1630:(hdl_radio_rx)->_lock irq_context: 0 &dev->mutex vivid_ctrls:1610:(hdl_user_aud)->_lock vivid_ctrls:1630:(hdl_radio_rx)->_lock irq_context: 0 &dev->mutex vivid_ctrls:1606:(hdl_user_gen)->_lock vivid_ctrls:1632:(hdl_radio_tx)->_lock irq_context: 0 &dev->mutex vivid_ctrls:1606:(hdl_user_gen)->_lock &pcp->lock &zone->lock irq_context: 0 &dev->mutex vivid_ctrls:1606:(hdl_user_gen)->_lock &zone->lock irq_context: 0 &dev->mutex vivid_ctrls:1610:(hdl_user_aud)->_lock vivid_ctrls:1632:(hdl_radio_tx)->_lock irq_context: 0 &dev->mutex vivid_ctrls:1606:(hdl_user_gen)->_lock vivid_ctrls:1634:(hdl_sdr_cap)->_lock irq_context: 0 &dev->mutex vivid_ctrls:1612:(hdl_streaming)->_lock vivid_ctrls:1634:(hdl_sdr_cap)->_lock irq_context: 0 &dev->mutex vivid_ctrls:1606:(hdl_user_gen)->_lock vivid_ctrls:1636:(hdl_meta_cap)->_lock irq_context: 0 &dev->mutex vivid_ctrls:1612:(hdl_streaming)->_lock vivid_ctrls:1636:(hdl_meta_cap)->_lock irq_context: 0 &dev->mutex vivid_ctrls:1606:(hdl_user_gen)->_lock vivid_ctrls:1638:(hdl_meta_out)->_lock irq_context: 0 &dev->mutex vivid_ctrls:1612:(hdl_streaming)->_lock vivid_ctrls:1638:(hdl_meta_out)->_lock irq_context: 0 &dev->mutex vivid_ctrls:1606:(hdl_user_gen)->_lock vivid_ctrls:1640:(hdl_tch_cap)->_lock irq_context: 0 &dev->mutex vivid_ctrls:1612:(hdl_streaming)->_lock vivid_ctrls:1640:(hdl_tch_cap)->_lock irq_context: 0 &dev->mutex vivid_ctrls:1612:(hdl_streaming)->_lock rcu_read_lock pool_lock#2 irq_context: 0 &dev->mutex vivid_ctrls:1612:(hdl_streaming)->_lock &obj_hash[i].lock irq_context: 0 &dev->mutex vivid_ctrls:1606:(hdl_user_gen)->_lock &obj_hash[i].lock irq_context: 0 &adap->kthread_waitq irq_context: 0 &dev->cec_xfers_slock irq_context: 0 &dev->mutex cec_devnode_lock irq_context: 0 &dev->kthread_waitq_cec irq_context: 0 &dev->mutex subsys mutex#71 irq_context: 0 &dev->mutex pin_fs_lock irq_context: 0 &dev->mutex &sb->s_type->i_mutex_key#3 irq_context: 0 &dev->mutex &sb->s_type->i_mutex_key#3 &sb->s_type->i_lock_key#7 irq_context: 0 &dev->mutex &sb->s_type->i_mutex_key#3 rename_lock.seqcount irq_context: 0 &dev->mutex &sb->s_type->i_mutex_key#3 fs_reclaim irq_context: 0 &dev->mutex &sb->s_type->i_mutex_key#3 fs_reclaim mmu_notifier_invalidate_range_start irq_context: 0 &dev->mutex &sb->s_type->i_mutex_key#3 pool_lock#2 irq_context: 0 &dev->mutex &sb->s_type->i_mutex_key#3 &dentry->d_lock irq_context: 0 &dev->mutex &sb->s_type->i_mutex_key#3 rcu_read_lock rename_lock.seqcount irq_context: 0 &dev->mutex &sb->s_type->i_mutex_key#3 &dentry->d_lock &wq irq_context: 0 &dev->mutex &sb->s_type->i_mutex_key#3 mmu_notifier_invalidate_range_start irq_context: 0 &dev->mutex &sb->s_type->i_mutex_key#3 &s->s_inode_list_lock irq_context: 0 &dev->mutex &sb->s_type->i_mutex_key#3 tk_core.seq.seqcount irq_context: 0 &dev->mutex &sb->s_type->i_mutex_key#3 &sb->s_type->i_lock_key#7 &dentry->d_lock irq_context: 0 &dev->mutex &adap->lock irq_context: 0 &dev->mutex &adap->lock tk_core.seq.seqcount irq_context: 0 &dev->mutex &adap->lock &adap->devnode.lock_fhs irq_context: 0 (wq_completion)events_unbound (work_completion)(&(&kfence_timer)->work) rcu_read_lock &rq->__lock irq_context: 0 (wq_completion)events_unbound (work_completion)(&(&kfence_timer)->work) rcu_read_lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 &dev->mutex vivid_ctrls:1608:(hdl_user_vid)->_lock &c->lock irq_context: 0 &dev->mutex vivid_ctrls:1608:(hdl_user_vid)->_lock &pcp->lock &zone->lock irq_context: 0 &dev->mutex vivid_ctrls:1608:(hdl_user_vid)->_lock &zone->lock irq_context: 0 &dev->mutex vivid_ctrls:1608:(hdl_user_vid)->_lock &____s->seqcount irq_context: 0 &dev->mutex vivid_ctrls:1606:(hdl_user_gen)->_lock &pcp->lock &zone->lock &____s->seqcount irq_context: 0 &dev->mutex vivid_ctrls:1610:(hdl_user_aud)->_lock &c->lock irq_context: 0 &dev->mutex vivid_ctrls:1610:(hdl_user_aud)->_lock &____s->seqcount irq_context: 0 &dev->mutex &sb->s_type->i_mutex_key#3 &c->lock irq_context: 0 &dev->mutex &sb->s_type->i_mutex_key#3 &____s->seqcount irq_context: 0 (wq_completion)mm_percpu_wq (work_completion)(&(({ do { const void *__vpp_verify = (typeof((&vmstat_work) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __asm__ ("" : "=r"(__ptr) : "0"((typeof(*((&vmstat_work))) *)((&vmstat_work)))); (typeof((typeof(*((&vmstat_work))) *)((&vmstat_work)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); }))->work) &pcp->lock irq_context: 0 (wq_completion)mm_percpu_wq (work_completion)(&(({ do { const void *__vpp_verify = (typeof((&vmstat_work) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __asm__ ("" : "=r"(__ptr) : "0"((typeof(*((&vmstat_work))) *)((&vmstat_work)))); (typeof((typeof(*((&vmstat_work))) *)((&vmstat_work)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); }))->work) &pcp->lock &zone->lock irq_context: 0 (wq_completion)mm_percpu_wq (work_completion)(&(({ do { const void *__vpp_verify = (typeof((&vmstat_work) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __asm__ ("" : "=r"(__ptr) : "0"((typeof(*((&vmstat_work))) *)((&vmstat_work)))); (typeof((typeof(*((&vmstat_work))) *)((&vmstat_work)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); }))->work) &pcp->lock &zone->lock &____s->seqcount irq_context: 0 &dev->mutex vivid_ctrls:1614:(hdl_sdtv_cap)->_lock &c->lock irq_context: 0 &dev->mutex vivid_ctrls:1614:(hdl_sdtv_cap)->_lock &pcp->lock &zone->lock irq_context: 0 &dev->mutex vivid_ctrls:1614:(hdl_sdtv_cap)->_lock &zone->lock irq_context: 0 &dev->mutex vivid_ctrls:1614:(hdl_sdtv_cap)->_lock &____s->seqcount irq_context: 0 &dev->mutex vivid_ctrls:1606:(hdl_user_gen)->_lock &zone->lock &____s->seqcount irq_context: 0 (wq_completion)events_unbound (work_completion)(&sub_info->work) &n->list_lock &c->lock irq_context: 0 &dev->mutex fs_reclaim mmu_notifier_invalidate_range_start &rq->__lock irq_context: 0 &dev->mutex fs_reclaim mmu_notifier_invalidate_range_start &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 &dev->mutex rcu_read_lock rcu_node_0 irq_context: 0 &dev->mutex &p->pi_lock &cfs_rq->removed.lock irq_context: 0 &dev->mutex &sb->s_type->i_mutex_key#3 &pcp->lock &zone->lock irq_context: 0 &dev->mutex &sb->s_type->i_mutex_key#3 &zone->lock irq_context: 0 &dev->mutex vivid_ctrls:1616:(hdl_loop_cap)->_lock &c->lock irq_context: 0 &dev->mutex vivid_ctrls:1616:(hdl_loop_cap)->_lock &pcp->lock &zone->lock irq_context: 0 &dev->mutex vivid_ctrls:1616:(hdl_loop_cap)->_lock &zone->lock irq_context: 0 &dev->mutex vivid_ctrls:1616:(hdl_loop_cap)->_lock &____s->seqcount irq_context: 0 rcu_read_lock rcu_read_lock rcu_node_0 irq_context: 0 rcu_read_lock rcu_read_lock &rq->__lock irq_context: 0 &dev->mutex vivid_ctrls:1610:(hdl_user_aud)->_lock &pcp->lock &zone->lock irq_context: 0 &dev->mutex vivid_ctrls:1610:(hdl_user_aud)->_lock &zone->lock irq_context: 0 (wq_completion)events_long serio_event_work serio_mutex &dev->mutex &serio->drv_mutex psmouse_mutex &rq->__lock irq_context: 0 (wq_completion)events_long serio_event_work serio_mutex &dev->mutex &serio->drv_mutex psmouse_mutex &k->list_lock irq_context: 0 (wq_completion)events_long serio_event_work serio_mutex &dev->mutex &serio->drv_mutex psmouse_mutex gdp_mutex irq_context: 0 (wq_completion)events_long serio_event_work serio_mutex &dev->mutex &serio->drv_mutex psmouse_mutex gdp_mutex &k->list_lock irq_context: 0 (wq_completion)events_long serio_event_work serio_mutex &dev->mutex &serio->drv_mutex psmouse_mutex gdp_mutex fs_reclaim irq_context: 0 (wq_completion)events_long serio_event_work serio_mutex &dev->mutex &serio->drv_mutex psmouse_mutex gdp_mutex fs_reclaim mmu_notifier_invalidate_range_start irq_context: 0 (wq_completion)events_long serio_event_work serio_mutex &dev->mutex &serio->drv_mutex psmouse_mutex gdp_mutex pool_lock#2 irq_context: 0 (wq_completion)events_long serio_event_work serio_mutex &dev->mutex &serio->drv_mutex psmouse_mutex gdp_mutex lock irq_context: 0 (wq_completion)events_long serio_event_work serio_mutex &dev->mutex &serio->drv_mutex psmouse_mutex gdp_mutex lock kernfs_idr_lock irq_context: 0 (wq_completion)events_long serio_event_work serio_mutex &dev->mutex &serio->drv_mutex psmouse_mutex gdp_mutex &root->kernfs_rwsem irq_context: 0 (wq_completion)events_long serio_event_work serio_mutex &dev->mutex &serio->drv_mutex psmouse_mutex gdp_mutex &root->kernfs_rwsem &root->kernfs_iattr_rwsem irq_context: 0 (wq_completion)events_long serio_event_work serio_mutex &dev->mutex &serio->drv_mutex psmouse_mutex lock irq_context: 0 (wq_completion)events_long serio_event_work serio_mutex &dev->mutex &serio->drv_mutex psmouse_mutex lock kernfs_idr_lock irq_context: 0 (wq_completion)events_long serio_event_work serio_mutex &dev->mutex &serio->drv_mutex psmouse_mutex &root->kernfs_rwsem irq_context: 0 (wq_completion)events_long serio_event_work serio_mutex &dev->mutex &serio->drv_mutex psmouse_mutex &root->kernfs_rwsem &root->kernfs_iattr_rwsem irq_context: 0 (wq_completion)events_long serio_event_work serio_mutex &dev->mutex &serio->drv_mutex psmouse_mutex bus_type_sem irq_context: 0 (wq_completion)events_long serio_event_work serio_mutex &dev->mutex &serio->drv_mutex psmouse_mutex sysfs_symlink_target_lock irq_context: 0 (wq_completion)events_long serio_event_work serio_mutex &dev->mutex &serio->drv_mutex psmouse_mutex &c->lock irq_context: 0 (wq_completion)events_long serio_event_work serio_mutex &dev->mutex &serio->drv_mutex psmouse_mutex &pcp->lock &zone->lock irq_context: 0 (wq_completion)events_long serio_event_work serio_mutex &dev->mutex &serio->drv_mutex psmouse_mutex &zone->lock irq_context: 0 (wq_completion)events_long serio_event_work serio_mutex &dev->mutex &serio->drv_mutex psmouse_mutex &____s->seqcount irq_context: 0 (wq_completion)events_long serio_event_work serio_mutex &dev->mutex &serio->drv_mutex psmouse_mutex &root->kernfs_rwsem irq_context: 0 (wq_completion)events_long serio_event_work serio_mutex &dev->mutex &serio->drv_mutex psmouse_mutex lock kernfs_idr_lock pool_lock#2 irq_context: 0 (wq_completion)events_long serio_event_work serio_mutex &dev->mutex &serio->drv_mutex psmouse_mutex &dev->power.lock irq_context: 0 (wq_completion)events_long serio_event_work serio_mutex &dev->mutex &serio->drv_mutex psmouse_mutex dpm_list_mtx irq_context: 0 (wq_completion)events_long serio_event_work serio_mutex &dev->mutex &serio->drv_mutex psmouse_mutex uevent_sock_mutex irq_context: 0 (wq_completion)events_long serio_event_work serio_mutex &dev->mutex &serio->drv_mutex psmouse_mutex rcu_read_lock &pool->lock/1 irq_context: 0 (wq_completion)events_long serio_event_work serio_mutex &dev->mutex &serio->drv_mutex psmouse_mutex rcu_read_lock &pool->lock/1 &obj_hash[i].lock irq_context: 0 (wq_completion)events_long serio_event_work serio_mutex &dev->mutex &serio->drv_mutex psmouse_mutex rcu_read_lock &pool->lock/1 pool_lock#2 irq_context: 0 (wq_completion)events_long serio_event_work serio_mutex &dev->mutex &serio->drv_mutex psmouse_mutex rcu_read_lock &pool->lock/1 &p->pi_lock irq_context: 0 (wq_completion)events_long serio_event_work serio_mutex &dev->mutex &serio->drv_mutex psmouse_mutex running_helpers_waitq.lock irq_context: 0 (wq_completion)events_long serio_event_work serio_mutex &dev->mutex &serio->drv_mutex psmouse_mutex rcu_read_lock pool_lock#2 irq_context: 0 (wq_completion)events_long serio_event_work serio_mutex &dev->mutex &serio->drv_mutex psmouse_mutex &k->k_lock irq_context: 0 (wq_completion)events_long serio_event_work serio_mutex &dev->mutex &serio->drv_mutex psmouse_mutex subsys mutex#30 irq_context: 0 (wq_completion)events_long serio_event_work serio_mutex &dev->mutex &serio->drv_mutex psmouse_mutex subsys mutex#30 &k->k_lock irq_context: 0 (wq_completion)events_long serio_event_work serio_mutex &dev->mutex &serio->drv_mutex psmouse_mutex (console_sem).lock irq_context: 0 (wq_completion)events_long serio_event_work serio_mutex &dev->mutex &serio->drv_mutex psmouse_mutex console_owner_lock irq_context: 0 (wq_completion)events_long serio_event_work serio_mutex &dev->mutex &serio->drv_mutex psmouse_mutex console_owner irq_context: 0 (wq_completion)events_long serio_event_work serio_mutex &dev->mutex &serio->drv_mutex psmouse_mutex console_lock console_srcu console_owner_lock irq_context: 0 (wq_completion)events_long serio_event_work serio_mutex &dev->mutex &serio->drv_mutex psmouse_mutex console_lock console_srcu console_owner irq_context: 0 (wq_completion)events_long serio_event_work serio_mutex &dev->mutex &serio->drv_mutex psmouse_mutex console_lock console_srcu console_owner &port_lock_key irq_context: 0 (wq_completion)events_long serio_event_work serio_mutex &dev->mutex &serio->drv_mutex psmouse_mutex console_lock console_srcu console_owner console_owner_lock irq_context: 0 (wq_completion)events_long serio_event_work serio_mutex &dev->mutex &serio->drv_mutex psmouse_mutex &cfs_rq->removed.lock irq_context: 0 (wq_completion)events_long serio_event_work serio_mutex &dev->mutex &serio->drv_mutex psmouse_mutex input_mutex irq_context: 0 (wq_completion)events_long serio_event_work serio_mutex &dev->mutex &serio->drv_mutex psmouse_mutex input_mutex input_ida.xa_lock irq_context: 0 (wq_completion)events_long serio_event_work serio_mutex &dev->mutex &serio->drv_mutex psmouse_mutex input_mutex fs_reclaim irq_context: 0 (wq_completion)events_long serio_event_work serio_mutex &dev->mutex &serio->drv_mutex psmouse_mutex input_mutex fs_reclaim mmu_notifier_invalidate_range_start irq_context: 0 (wq_completion)events_long serio_event_work serio_mutex &dev->mutex &serio->drv_mutex psmouse_mutex input_mutex &c->lock irq_context: 0 (wq_completion)events_long serio_event_work serio_mutex &dev->mutex &serio->drv_mutex psmouse_mutex input_mutex &pcp->lock &zone->lock irq_context: 0 (wq_completion)events_long serio_event_work serio_mutex &dev->mutex &serio->drv_mutex psmouse_mutex input_mutex &zone->lock irq_context: 0 (wq_completion)events_long serio_event_work serio_mutex &dev->mutex &serio->drv_mutex psmouse_mutex input_mutex &____s->seqcount irq_context: 0 (wq_completion)events_long serio_event_work serio_mutex &dev->mutex &serio->drv_mutex psmouse_mutex input_mutex pool_lock#2 irq_context: 0 (wq_completion)events_long serio_event_work serio_mutex &dev->mutex &serio->drv_mutex psmouse_mutex input_mutex &x->wait#9 irq_context: 0 (wq_completion)events_long serio_event_work serio_mutex &dev->mutex &serio->drv_mutex psmouse_mutex input_mutex &obj_hash[i].lock irq_context: 0 (wq_completion)events_long serio_event_work serio_mutex &dev->mutex &serio->drv_mutex psmouse_mutex input_mutex &dev->mutex#2 irq_context: 0 (wq_completion)events_long serio_event_work serio_mutex &dev->mutex &serio->drv_mutex psmouse_mutex input_mutex chrdevs_lock irq_context: 0 (wq_completion)events_long serio_event_work serio_mutex &dev->mutex &serio->drv_mutex psmouse_mutex input_mutex &k->list_lock irq_context: 0 (wq_completion)events_long serio_event_work serio_mutex &dev->mutex &serio->drv_mutex psmouse_mutex input_mutex lock irq_context: 0 (wq_completion)events_long serio_event_work serio_mutex &dev->mutex &serio->drv_mutex psmouse_mutex input_mutex lock kernfs_idr_lock irq_context: 0 (wq_completion)events_long serio_event_work serio_mutex &dev->mutex &serio->drv_mutex psmouse_mutex input_mutex &root->kernfs_rwsem irq_context: 0 (wq_completion)events_long serio_event_work serio_mutex &dev->mutex &serio->drv_mutex psmouse_mutex input_mutex &root->kernfs_rwsem &root->kernfs_iattr_rwsem irq_context: 0 (wq_completion)events_long serio_event_work serio_mutex &dev->mutex &serio->drv_mutex psmouse_mutex input_mutex bus_type_sem irq_context: 0 (wq_completion)events_long serio_event_work serio_mutex &dev->mutex &serio->drv_mutex psmouse_mutex input_mutex sysfs_symlink_target_lock irq_context: 0 (wq_completion)events_long serio_event_work serio_mutex &dev->mutex &serio->drv_mutex psmouse_mutex input_mutex &root->kernfs_rwsem irq_context: 0 (wq_completion)events_long serio_event_work serio_mutex &dev->mutex &serio->drv_mutex psmouse_mutex input_mutex &dev->power.lock irq_context: 0 (wq_completion)events_long serio_event_work serio_mutex &dev->mutex &serio->drv_mutex psmouse_mutex input_mutex dpm_list_mtx irq_context: 0 (wq_completion)events_long serio_event_work serio_mutex &dev->mutex &serio->drv_mutex psmouse_mutex input_mutex req_lock irq_context: 0 (wq_completion)events_long serio_event_work serio_mutex &dev->mutex &serio->drv_mutex psmouse_mutex input_mutex &p->pi_lock irq_context: 0 (wq_completion)events_long serio_event_work serio_mutex &dev->mutex &serio->drv_mutex psmouse_mutex input_mutex &p->pi_lock &rq->__lock irq_context: 0 (wq_completion)events_long serio_event_work serio_mutex &dev->mutex &serio->drv_mutex psmouse_mutex input_mutex &p->pi_lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)events_long serio_event_work serio_mutex &dev->mutex &serio->drv_mutex psmouse_mutex input_mutex &p->pi_lock &rq->__lock &cfs_rq->removed.lock irq_context: 0 (wq_completion)events_long serio_event_work serio_mutex &dev->mutex &serio->drv_mutex psmouse_mutex input_mutex &x->wait#11 irq_context: 0 (wq_completion)events_long serio_event_work serio_mutex &dev->mutex &serio->drv_mutex psmouse_mutex input_mutex &pool->lock irq_context: 0 (wq_completion)events_long serio_event_work serio_mutex &dev->mutex &serio->drv_mutex psmouse_mutex input_mutex &rq->__lock irq_context: 0 (wq_completion)events_long serio_event_work serio_mutex &dev->mutex &serio->drv_mutex psmouse_mutex input_mutex &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)events_long serio_event_work serio_mutex &dev->mutex &serio->drv_mutex psmouse_mutex input_mutex uevent_sock_mutex irq_context: 0 (wq_completion)events_long serio_event_work serio_mutex &dev->mutex &serio->drv_mutex psmouse_mutex input_mutex rcu_read_lock &pool->lock/1 irq_context: 0 (wq_completion)events_long serio_event_work serio_mutex &dev->mutex &serio->drv_mutex psmouse_mutex input_mutex rcu_read_lock &pool->lock/1 &obj_hash[i].lock irq_context: 0 (wq_completion)events_long serio_event_work serio_mutex &dev->mutex &serio->drv_mutex psmouse_mutex input_mutex rcu_read_lock &pool->lock/1 pool_lock#2 irq_context: 0 (wq_completion)events_long serio_event_work serio_mutex &dev->mutex &serio->drv_mutex psmouse_mutex input_mutex rcu_read_lock &pool->lock/1 &p->pi_lock irq_context: 0 (wq_completion)events_long serio_event_work serio_mutex &dev->mutex &serio->drv_mutex psmouse_mutex input_mutex rcu_read_lock &pool->lock/1 &p->pi_lock &rq->__lock irq_context: 0 (wq_completion)events_long serio_event_work serio_mutex &dev->mutex &serio->drv_mutex psmouse_mutex input_mutex rcu_read_lock &pool->lock/1 &p->pi_lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)events_long serio_event_work serio_mutex &dev->mutex &serio->drv_mutex psmouse_mutex input_mutex running_helpers_waitq.lock irq_context: 0 (wq_completion)events_long serio_event_work serio_mutex &dev->mutex &serio->drv_mutex psmouse_mutex input_mutex &k->k_lock irq_context: 0 (wq_completion)events_long serio_event_work serio_mutex &dev->mutex &serio->drv_mutex psmouse_mutex input_mutex subsys mutex#30 irq_context: 0 (wq_completion)events_long serio_event_work serio_mutex &dev->mutex &serio->drv_mutex psmouse_mutex input_mutex subsys mutex#30 &k->k_lock irq_context: 0 (wq_completion)events_long serio_event_work serio_mutex &dev->mutex &serio->drv_mutex psmouse_mutex input_mutex &mousedev->mutex/1 irq_context: 0 (wq_completion)events_long serio_event_work serio_mutex &dev->mutex &serio->drv_mutex psmouse_mutex input_mutex &cfs_rq->removed.lock irq_context: 0 (wq_completion)events_long serio_event_work serio_mutex &dev->mutex &serio->drv_mutex psmouse_mutex input_mutex input_devices_poll_wait.lock irq_context: 0 &dev->mutex vivid_ctrls:1612:(hdl_streaming)->_lock &pcp->lock &zone->lock &____s->seqcount irq_context: 0 &dev->mutex vivid_ctrls:1606:(hdl_user_gen)->_lock batched_entropy_u8.lock irq_context: 0 &dev->mutex vivid_ctrls:1606:(hdl_user_gen)->_lock kfence_freelist_lock irq_context: softirq drivers/block/floppy.c:640 irq_context: softirq drivers/block/floppy.c:640 rcu_read_lock &pool->lock/1 irq_context: softirq drivers/block/floppy.c:640 rcu_read_lock &pool->lock/1 &obj_hash[i].lock irq_context: softirq drivers/block/floppy.c:640 rcu_read_lock &pool->lock/1 &p->pi_lock irq_context: softirq drivers/block/floppy.c:640 rcu_read_lock &pool->lock/1 &p->pi_lock &rq->__lock irq_context: softirq drivers/block/floppy.c:640 rcu_read_lock &pool->lock/1 &p->pi_lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: softirq drivers/block/floppy.c:640 rcu_read_lock &pool->lock/1 &p->pi_lock &rq->__lock &cfs_rq->removed.lock irq_context: 0 (wq_completion)events (work_completion)(&blkg->free_work) &q->blkcg_mutex pool_lock#2 irq_context: 0 (wq_completion)events (work_completion)(&blkg->free_work) &q->blkcg_mutex &q->queue_lock irq_context: 0 (wq_completion)events (work_completion)(&blkg->free_work) &obj_hash[i].lock irq_context: 0 (wq_completion)events (work_completion)(&blkg->free_work) pool_lock#2 irq_context: 0 (wq_completion)events (work_completion)(&blkg->free_work) &xa->xa_lock#10 irq_context: 0 (wq_completion)events (work_completion)(&blkg->free_work) pcpu_lock irq_context: 0 (wq_completion)events (work_completion)(&blkg->free_work) blk_queue_ida.xa_lock irq_context: 0 (wq_completion)events (work_completion)(&blkg->free_work) percpu_ref_switch_lock irq_context: 0 (wq_completion)events (work_completion)(&blkg->free_work) &obj_hash[i].lock pool_lock irq_context: 0 &dev->mutex lock kernfs_idr_lock &pcp->lock &zone->lock &____s->seqcount irq_context: softirq &(({ do { const void *__vpp_verify = (typeof((&vmstat_work) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __asm__ ("" : "=r"(__ptr) : "0"((typeof(*((&vmstat_work))) *)((&vmstat_work)))); (typeof((typeof(*((&vmstat_work))) *)((&vmstat_work)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); }))->timer rcu_read_lock &pool->lock &p->pi_lock &rq->__lock &cfs_rq->removed.lock irq_context: 0 ptp_clocks_map.xa_lock irq_context: 0 subsys mutex#72 irq_context: 0 (wq_completion)events_unbound (work_completion)(&entry->work) (console_sem).lock irq_context: 0 (wq_completion)events_unbound (work_completion)(&entry->work) console_owner_lock irq_context: 0 (wq_completion)events_unbound (work_completion)(&entry->work) console_owner irq_context: 0 (wq_completion)events_unbound (work_completion)(&entry->work) console_lock console_srcu console_owner_lock irq_context: 0 (wq_completion)events_unbound (work_completion)(&entry->work) console_lock console_srcu console_owner irq_context: 0 (wq_completion)events_unbound (work_completion)(&entry->work) console_lock console_srcu console_owner &port_lock_key irq_context: 0 &dev->mutex console_owner_lock irq_context: 0 &dev->mutex console_owner irq_context: 0 (wq_completion)events_unbound (work_completion)(&entry->work) console_lock console_srcu console_owner console_owner_lock irq_context: 0 (wq_completion)events_unbound (work_completion)(&entry->work) &desc->request_mutex &irq_desc_lock_class irq_resend_lock irq_context: 0 (wq_completion)events_unbound (work_completion)(&entry->work) &desc->request_mutex proc_subdir_lock irq_context: 0 (wq_completion)events_unbound (work_completion)(&entry->work) &desc->request_mutex &ent->pde_unload_lock irq_context: 0 (wq_completion)events_unbound (work_completion)(&entry->work) &desc->request_mutex proc_inum_ida.xa_lock irq_context: 0 (wq_completion)events_unbound (work_completion)(&entry->work) &desc->request_mutex &obj_hash[i].lock irq_context: 0 (wq_completion)events_unbound (work_completion)(&entry->work) &desc->request_mutex pool_lock#2 irq_context: 0 subsys mutex#72 &k->k_lock irq_context: 0 rcu_read_lock &stopper->lock irq_context: 0 rcu_read_lock &stop_pi_lock irq_context: 0 (wq_completion)events_unbound (work_completion)(&entry->work) &root->kernfs_rwsem kernfs_idr_lock irq_context: 0 (wq_completion)events_unbound (work_completion)(&entry->work) &root->kernfs_rwsem &obj_hash[i].lock irq_context: 0 (wq_completion)events_unbound (work_completion)(&entry->work) &root->kernfs_rwsem pool_lock#2 irq_context: 0 (wq_completion)events_unbound (work_completion)(&entry->work) klist_remove_lock irq_context: 0 (wq_completion)events_unbound (work_completion)(&entry->work) &k->k_lock klist_remove_lock irq_context: 0 (wq_completion)events_unbound (work_completion)(&entry->work) rcu_read_lock &pool->lock/1 &p->pi_lock &rq->__lock &cfs_rq->removed.lock irq_context: 0 (wq_completion)events_unbound (work_completion)(&entry->work) sysfs_symlink_target_lock irq_context: 0 (wq_completion)events_unbound (work_completion)(&entry->work) kernfs_idr_lock irq_context: 0 (wq_completion)events_unbound (work_completion)(&entry->work) wq_pool_mutex &wq->mutex &pool->lock/1 irq_context: 0 (wq_completion)events_unbound (work_completion)(&entry->work) &pool->lock/1 irq_context: 0 (wq_completion)events_unbound (work_completion)(&entry->work) &pool->lock/1 rcu_read_lock &pool->lock irq_context: 0 (wq_completion)events_unbound (work_completion)(&entry->work) &pool->lock/1 rcu_read_lock &pool->lock &obj_hash[i].lock irq_context: 0 (wq_completion)events_unbound (work_completion)(&entry->work) &pool->lock/1 rcu_read_lock &pool->lock pool_lock#2 irq_context: 0 (wq_completion)events_unbound (work_completion)(&entry->work) &pool->lock/1 rcu_read_lock &pool->lock &p->pi_lock irq_context: 0 (wq_completion)events_unbound (work_completion)(&entry->work) &pool->lock/1 rcu_read_lock &pool->lock &p->pi_lock &rq->__lock irq_context: 0 (wq_completion)events_unbound (work_completion)(&entry->work) &pool->lock/1 rcu_read_lock &pool->lock &p->pi_lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)events_unbound (work_completion)(&entry->work) &pool->lock/1 rcu_read_lock &pool->lock &p->pi_lock &rq->__lock &cfs_rq->removed.lock irq_context: 0 (wq_completion)events_unbound (work_completion)(&entry->work) (&motor_off_timer[drive]) irq_context: 0 (wq_completion)events_unbound (work_completion)(&entry->work) &q->debugfs_mutex irq_context: 0 (wq_completion)events_unbound (work_completion)(&entry->work) cpu_hotplug_lock cpuhp_state_mutex irq_context: 0 (wq_completion)events_unbound (work_completion)(&entry->work) &xa->xa_lock#10 irq_context: 0 (wq_completion)events_unbound (work_completion)(&entry->work) &q->unused_hctx_lock irq_context: 0 (wq_completion)events_unbound (work_completion)(&entry->work) &q->queue_lock &blkcg->lock irq_context: 0 (wq_completion)events_unbound (work_completion)(&entry->work) &q->queue_lock &blkcg->lock percpu_ref_switch_lock irq_context: 0 (wq_completion)events_unbound (work_completion)(&entry->work) &q->queue_lock &blkcg->lock percpu_ref_switch_lock &obj_hash[i].lock irq_context: 0 (wq_completion)events_unbound (work_completion)(&entry->work) &q->queue_lock &blkcg->lock percpu_ref_switch_lock pool_lock#2 irq_context: 0 (wq_completion)events_unbound (work_completion)(&entry->work) (&sq->pending_timer) irq_context: 0 (wq_completion)events_unbound (work_completion)(&entry->work) (work_completion)(&td->dispatch_work) irq_context: 0 (wq_completion)events_unbound (work_completion)(&entry->work) &q->blkcg_mutex irq_context: 0 (wq_completion)events_unbound (work_completion)(&entry->work) &q->blkcg_mutex &q->queue_lock irq_context: 0 (wq_completion)events_unbound (work_completion)(&entry->work) &q->blkcg_mutex &q->queue_lock &blkcg->lock irq_context: 0 (wq_completion)events_unbound (work_completion)(&entry->work) &q->blkcg_mutex &q->queue_lock &blkcg->lock (&sq->pending_timer) irq_context: 0 (wq_completion)events_unbound (work_completion)(&entry->work) &q->blkcg_mutex &q->queue_lock &blkcg->lock &obj_hash[i].lock irq_context: 0 (wq_completion)events_unbound (work_completion)(&entry->work) &q->blkcg_mutex &q->queue_lock &blkcg->lock &base->lock irq_context: 0 (wq_completion)events_unbound (work_completion)(&entry->work) &q->blkcg_mutex &q->queue_lock &blkcg->lock percpu_counters_lock irq_context: 0 (wq_completion)events_unbound (work_completion)(&entry->work) &q->blkcg_mutex &q->queue_lock &blkcg->lock pcpu_lock irq_context: 0 (wq_completion)events_unbound (work_completion)(&entry->work) &q->blkcg_mutex &q->queue_lock &blkcg->lock pool_lock#2 irq_context: 0 (wq_completion)events_unbound (work_completion)(&entry->work) pcpu_lock irq_context: 0 (wq_completion)events_unbound (work_completion)(&entry->work) &xa->xa_lock#9 irq_context: 0 (wq_completion)events_unbound (work_completion)(&entry->work) &fsnotify_mark_srcu irq_context: 0 (wq_completion)events_unbound (work_completion)(&entry->work) &q->blkcg_mutex &q->queue_lock &blkcg->lock pool_lock irq_context: 0 (wq_completion)events_unbound (work_completion)(&entry->work) &obj_hash[i].lock pool_lock irq_context: 0 (wq_completion)events (work_completion)(&pwq->unbound_release_work) rcu_state.exp_mutex rcu_node_0 irq_context: 0 (wq_completion)events (work_completion)(&pwq->unbound_release_work) rcu_state.exp_mutex &obj_hash[i].lock irq_context: 0 (wq_completion)events (work_completion)(&pwq->unbound_release_work) rcu_state.exp_mutex rcu_read_lock &pool->lock irq_context: 0 (wq_completion)events (work_completion)(&pwq->unbound_release_work) rcu_state.exp_mutex rcu_read_lock &pool->lock &obj_hash[i].lock irq_context: 0 (wq_completion)events (work_completion)(&pwq->unbound_release_work) rcu_state.exp_mutex &rnp->exp_wq[0] irq_context: 0 (wq_completion)events (work_completion)(&pwq->unbound_release_work) rcu_state.exp_mutex &pool->lock irq_context: 0 (wq_completion)events (work_completion)(&pwq->unbound_release_work) rcu_state.exp_mutex &pool->lock &p->pi_lock irq_context: 0 (wq_completion)events (work_completion)(&pwq->unbound_release_work) rcu_state.exp_mutex &pool->lock &p->pi_lock &rq->__lock irq_context: 0 (wq_completion)events (work_completion)(&pwq->unbound_release_work) rcu_state.exp_mutex &pool->lock &p->pi_lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)events (work_completion)(&pwq->unbound_release_work) rcu_state.exp_mutex &rq->__lock irq_context: 0 (wq_completion)events (work_completion)(&pwq->unbound_release_work) rcu_state.exp_mutex &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 &dev->mutex rcu_node_0 irq_context: 0 &dev->mutex &rcu_state.expedited_wq irq_context: 0 &dev->mutex &rcu_state.expedited_wq &p->pi_lock irq_context: softirq rcu_callback percpu_ref_switch_waitq.lock irq_context: softirq rcu_callback rcu_read_lock &obj_hash[i].lock irq_context: softirq rcu_callback rcu_read_lock pool_lock#2 irq_context: softirq rcu_callback percpu_counters_lock irq_context: softirq rcu_callback &pcp->lock &zone->lock irq_context: softirq rcu_callback &zone->lock irq_context: softirq rcu_callback &pcp->lock &zone->lock &____s->seqcount irq_context: softirq rcu_callback &____s->seqcount irq_context: softirq rcu_callback pool_lock irq_context: 0 rcu_read_lock &stop_pi_lock &rq->__lock irq_context: 0 rcu_read_lock &stop_pi_lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 pers_lock irq_context: 0 &rq->__lock rcu_read_lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 _lock irq_context: 0 dm_bufio_clients_lock irq_context: 0 _ps_lock irq_context: 0 _lock#2 irq_context: 0 _lock#3 irq_context: 0 register_lock#2 irq_context: 0 subsys mutex#73 irq_context: 0 subsys mutex#73 &k->k_lock irq_context: 0 bp_lock irq_context: 0 bp_lock irq_context: 0 subsys mutex#74 irq_context: 0 subsys mutex#74 &k->k_lock irq_context: softirq (&dsp_spl_tl) irq_context: softirq (&dsp_spl_tl) dsp_lock irq_context: softirq (&dsp_spl_tl) dsp_lock iclock_lock irq_context: softirq (&dsp_spl_tl) dsp_lock iclock_lock tk_core.seq.seqcount irq_context: softirq (&dsp_spl_tl) dsp_lock &obj_hash[i].lock irq_context: softirq (&dsp_spl_tl) dsp_lock &base->lock irq_context: softirq (&dsp_spl_tl) dsp_lock &base->lock &obj_hash[i].lock irq_context: 0 cpu_hotplug_lock wq_pool_mutex rcu_read_lock pool_lock#2 irq_context: 0 leds_list_lock &led_cdev->trigger_lock irq_context: 0 rtnl_mutex lock#7 irq_context: 0 iscsi_transport_lock irq_context: 0 subsys mutex#75 irq_context: 0 subsys mutex#75 &k->k_lock irq_context: 0 link_ops_rwsem irq_context: 0 &tx_task->waiting irq_context: 0 subsys mutex#76 irq_context: 0 subsys mutex#76 &k->k_lock irq_context: 0 service_lock irq_context: 0 tasklist_lock &sighand->siglock &sig->wait_chldexit &p->pi_lock &rq->__lock &cfs_rq->removed.lock irq_context: 0 tasklist_lock &sighand->siglock &(&sig->stats_lock)->lock &____s->seqcount#5 fill_pool_map-wait-type-override &c->lock irq_context: 0 tasklist_lock &sighand->siglock &(&sig->stats_lock)->lock &____s->seqcount#5 fill_pool_map-wait-type-override &pcp->lock &zone->lock irq_context: 0 tasklist_lock &sighand->siglock &(&sig->stats_lock)->lock &____s->seqcount#5 fill_pool_map-wait-type-override &zone->lock irq_context: 0 tasklist_lock &sighand->siglock &(&sig->stats_lock)->lock &____s->seqcount#5 fill_pool_map-wait-type-override &____s->seqcount irq_context: 0 &sig->cred_guard_mutex &mm->mmap_lock rcu_read_lock &cfs_rq->removed.lock irq_context: 0 &sig->cred_guard_mutex &mm->mmap_lock rcu_read_lock &obj_hash[i].lock irq_context: 0 &sig->cred_guard_mutex &mm->mmap_lock rcu_read_lock &rcu_state.gp_wq &p->pi_lock &rq->__lock &cfs_rq->removed.lock irq_context: 0 vsock_register_mutex irq_context: 0 comedi_drivers_list_lock irq_context: 0 reading_mutex &x->wait#12 irq_context: 0 subsys mutex#77 irq_context: 0 subsys mutex#77 &k->k_lock irq_context: 0 snd_ctl_layer_rwsem irq_context: 0 snd_card_mutex irq_context: 0 snd_ioctl_rwsem irq_context: 0 strings irq_context: 0 strings fs_reclaim irq_context: 0 strings fs_reclaim mmu_notifier_invalidate_range_start irq_context: 0 strings pool_lock#2 irq_context: 0 register_mutex irq_context: 0 sound_mutex irq_context: 0 sound_mutex fs_reclaim irq_context: 0 sound_mutex fs_reclaim mmu_notifier_invalidate_range_start irq_context: 0 sound_mutex pool_lock#2 irq_context: 0 sound_mutex &k->list_lock irq_context: 0 sound_mutex gdp_mutex irq_context: 0 sound_mutex gdp_mutex &k->list_lock irq_context: 0 sound_mutex lock irq_context: 0 sound_mutex lock kernfs_idr_lock irq_context: 0 sound_mutex &root->kernfs_rwsem irq_context: 0 sound_mutex &root->kernfs_rwsem &root->kernfs_iattr_rwsem irq_context: 0 sound_mutex bus_type_sem irq_context: 0 sound_mutex sysfs_symlink_target_lock irq_context: 0 sound_mutex lock kernfs_idr_lock pool_lock#2 irq_context: 0 sound_mutex &root->kernfs_rwsem irq_context: 0 sound_mutex fs_reclaim &rq->__lock irq_context: 0 sound_mutex fs_reclaim &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 sound_mutex &c->lock irq_context: 0 sound_mutex &____s->seqcount irq_context: 0 sound_mutex &dev->power.lock irq_context: 0 sound_mutex dpm_list_mtx irq_context: 0 sound_mutex req_lock irq_context: 0 sound_mutex &p->pi_lock irq_context: 0 sound_mutex &p->pi_lock &rq->__lock irq_context: 0 sound_mutex &p->pi_lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 sound_mutex &p->pi_lock &rq->__lock &cfs_rq->removed.lock irq_context: 0 sound_mutex &x->wait#11 irq_context: 0 sound_mutex &rq->__lock irq_context: 0 sound_mutex &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)events_unbound (work_completion)(&sub_info->work) fs_reclaim &rq->__lock irq_context: 0 sound_mutex &obj_hash[i].lock irq_context: 0 sound_mutex uevent_sock_mutex irq_context: 0 sound_mutex rcu_read_lock &pool->lock/1 irq_context: 0 sound_mutex rcu_read_lock &pool->lock/1 &obj_hash[i].lock irq_context: 0 sound_mutex rcu_read_lock &pool->lock/1 pool_lock#2 irq_context: 0 sound_mutex rcu_read_lock &pool->lock/1 &p->pi_lock irq_context: 0 sound_mutex running_helpers_waitq.lock irq_context: 0 sound_mutex subsys mutex#77 irq_context: 0 sound_mutex subsys mutex#77 &k->k_lock irq_context: 0 info_mutex &c->lock irq_context: 0 info_mutex &____s->seqcount irq_context: 0 register_mutex#2 irq_context: 0 register_mutex#3 irq_context: 0 register_mutex#3 fs_reclaim irq_context: 0 register_mutex#3 fs_reclaim mmu_notifier_invalidate_range_start irq_context: 0 register_mutex#3 pool_lock#2 irq_context: 0 register_mutex#3 sound_mutex irq_context: 0 register_mutex#3 sound_mutex fs_reclaim irq_context: 0 register_mutex#3 sound_mutex fs_reclaim mmu_notifier_invalidate_range_start irq_context: 0 register_mutex#3 sound_mutex pool_lock#2 irq_context: 0 register_mutex#3 sound_mutex &k->list_lock irq_context: 0 register_mutex#3 sound_mutex gdp_mutex irq_context: 0 register_mutex#3 sound_mutex gdp_mutex &k->list_lock irq_context: 0 register_mutex#3 sound_mutex lock irq_context: 0 register_mutex#3 sound_mutex lock kernfs_idr_lock irq_context: 0 register_mutex#3 sound_mutex &root->kernfs_rwsem irq_context: 0 register_mutex#3 sound_mutex &root->kernfs_rwsem &root->kernfs_iattr_rwsem irq_context: 0 register_mutex#3 sound_mutex bus_type_sem irq_context: 0 register_mutex#3 sound_mutex sysfs_symlink_target_lock irq_context: 0 register_mutex#3 sound_mutex &root->kernfs_rwsem irq_context: 0 register_mutex#3 sound_mutex &dev->power.lock irq_context: 0 register_mutex#3 sound_mutex dpm_list_mtx irq_context: 0 register_mutex#3 sound_mutex req_lock irq_context: 0 register_mutex#3 sound_mutex &p->pi_lock irq_context: 0 register_mutex#3 sound_mutex &p->pi_lock &rq->__lock irq_context: 0 register_mutex#3 sound_mutex &p->pi_lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 register_mutex#3 sound_mutex &x->wait#11 irq_context: 0 register_mutex#3 sound_mutex &rq->__lock irq_context: 0 register_mutex#3 sound_mutex &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)events_unbound (work_completion)(&sub_info->work) fs_reclaim &cfs_rq->removed.lock irq_context: 0 (wq_completion)events_unbound (work_completion)(&sub_info->work) fs_reclaim &obj_hash[i].lock irq_context: 0 &sig->cred_guard_mutex &sb->s_type->i_mutex_key &obj_hash[i].lock pool_lock irq_context: 0 register_mutex#3 sound_mutex &cfs_rq->removed.lock irq_context: 0 register_mutex#3 sound_mutex &obj_hash[i].lock irq_context: 0 register_mutex#3 sound_mutex uevent_sock_mutex irq_context: 0 register_mutex#3 sound_mutex rcu_read_lock &pool->lock/1 irq_context: 0 register_mutex#3 sound_mutex rcu_read_lock &pool->lock/1 &obj_hash[i].lock irq_context: 0 register_mutex#3 sound_mutex rcu_read_lock &pool->lock/1 pool_lock#2 irq_context: 0 register_mutex#3 sound_mutex rcu_read_lock &pool->lock/1 &p->pi_lock irq_context: 0 register_mutex#3 sound_mutex rcu_read_lock &pool->lock/1 &p->pi_lock &rq->__lock irq_context: 0 register_mutex#3 sound_mutex rcu_read_lock &pool->lock/1 &p->pi_lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 register_mutex#3 sound_mutex rcu_read_lock &pool->lock/1 &p->pi_lock &rq->__lock &cfs_rq->removed.lock irq_context: 0 register_mutex#3 sound_mutex running_helpers_waitq.lock irq_context: 0 register_mutex#3 sound_mutex subsys mutex#77 irq_context: 0 register_mutex#3 sound_mutex subsys mutex#77 &k->k_lock irq_context: 0 register_mutex#3 clients_lock irq_context: 0 &client->ports_mutex irq_context: 0 &client->ports_mutex &client->ports_lock irq_context: 0 register_mutex#4 irq_context: 0 register_mutex#4 fs_reclaim irq_context: 0 register_mutex#4 fs_reclaim mmu_notifier_invalidate_range_start irq_context: 0 register_mutex#4 pool_lock#2 irq_context: 0 register_mutex#4 sound_oss_mutex irq_context: 0 register_mutex#4 sound_oss_mutex fs_reclaim irq_context: 0 register_mutex#4 sound_oss_mutex fs_reclaim mmu_notifier_invalidate_range_start irq_context: 0 register_mutex#4 sound_oss_mutex pool_lock#2 irq_context: 0 register_mutex#4 sound_oss_mutex sound_loader_lock irq_context: 0 register_mutex#4 sound_oss_mutex &x->wait#9 irq_context: 0 register_mutex#4 sound_oss_mutex &obj_hash[i].lock irq_context: 0 register_mutex#4 sound_oss_mutex &pcp->lock &zone->lock irq_context: 0 register_mutex#4 sound_oss_mutex &zone->lock irq_context: 0 register_mutex#4 sound_oss_mutex &____s->seqcount irq_context: 0 register_mutex#4 sound_oss_mutex rcu_read_lock pool_lock#2 irq_context: 0 register_mutex#4 sound_oss_mutex &k->list_lock irq_context: 0 register_mutex#4 sound_oss_mutex gdp_mutex irq_context: 0 register_mutex#4 sound_oss_mutex gdp_mutex &k->list_lock irq_context: 0 register_mutex#4 sound_oss_mutex lock irq_context: 0 register_mutex#4 sound_oss_mutex lock kernfs_idr_lock irq_context: 0 register_mutex#4 sound_oss_mutex &root->kernfs_rwsem irq_context: 0 register_mutex#4 sound_oss_mutex &root->kernfs_rwsem &root->kernfs_iattr_rwsem irq_context: 0 register_mutex#4 sound_oss_mutex bus_type_sem irq_context: 0 register_mutex#4 sound_oss_mutex &c->lock irq_context: 0 register_mutex#4 sound_oss_mutex sysfs_symlink_target_lock irq_context: 0 register_mutex#4 sound_oss_mutex &root->kernfs_rwsem irq_context: 0 register_mutex#4 sound_oss_mutex &dev->power.lock irq_context: 0 register_mutex#4 sound_oss_mutex dpm_list_mtx irq_context: 0 register_mutex#4 sound_oss_mutex req_lock irq_context: 0 register_mutex#4 sound_oss_mutex &p->pi_lock irq_context: 0 register_mutex#4 sound_oss_mutex &p->pi_lock &rq->__lock irq_context: 0 register_mutex#4 sound_oss_mutex &p->pi_lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 register_mutex#4 sound_oss_mutex &p->pi_lock &rq->__lock &cfs_rq->removed.lock irq_context: 0 register_mutex#4 sound_oss_mutex &x->wait#11 irq_context: 0 register_mutex#4 sound_oss_mutex &rq->__lock irq_context: 0 register_mutex#4 sound_oss_mutex &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 register_mutex#4 sound_oss_mutex rcu_read_lock &rq->__lock irq_context: 0 register_mutex#4 sound_oss_mutex rcu_read_lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 register_mutex#4 sound_oss_mutex uevent_sock_mutex irq_context: 0 register_mutex#4 sound_oss_mutex rcu_read_lock &pool->lock/1 irq_context: 0 register_mutex#4 sound_oss_mutex rcu_read_lock &pool->lock/1 &obj_hash[i].lock irq_context: 0 register_mutex#4 sound_oss_mutex rcu_read_lock &pool->lock/1 pool_lock#2 irq_context: 0 register_mutex#4 sound_oss_mutex rcu_read_lock &pool->lock/1 &p->pi_lock irq_context: 0 register_mutex#4 sound_oss_mutex rcu_read_lock &pool->lock/1 &p->pi_lock &rq->__lock irq_context: 0 register_mutex#4 sound_oss_mutex rcu_read_lock &pool->lock/1 &p->pi_lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 register_mutex#4 sound_oss_mutex running_helpers_waitq.lock irq_context: 0 register_mutex#4 sound_oss_mutex subsys mutex#77 irq_context: 0 register_mutex#4 sound_oss_mutex subsys mutex#77 &k->k_lock irq_context: 0 register_mutex#4 sound_oss_mutex &obj_hash[i].lock pool_lock irq_context: 0 clients_lock irq_context: 0 &client->ports_lock irq_context: 0 &grp->list_mutex/1 irq_context: 0 &grp->list_mutex#2 irq_context: 0 &grp->list_mutex#2 &grp->list_lock irq_context: 0 &grp->list_mutex/1 clients_lock irq_context: 0 &grp->list_mutex/1 &client->ports_lock irq_context: 0 (wq_completion)events async_lookup_work irq_context: 0 (wq_completion)events async_lookup_work fs_reclaim irq_context: 0 (wq_completion)events async_lookup_work fs_reclaim mmu_notifier_invalidate_range_start irq_context: 0 (wq_completion)events async_lookup_work pool_lock#2 irq_context: 0 (wq_completion)events async_lookup_work clients_lock irq_context: 0 (wq_completion)events async_lookup_work &client->ports_lock irq_context: 0 (wq_completion)events async_lookup_work snd_card_mutex irq_context: 0 (wq_completion)events async_lookup_work (kmod_concurrent_max).lock irq_context: 0 (wq_completion)events async_lookup_work &obj_hash[i].lock irq_context: 0 (wq_completion)events async_lookup_work rcu_read_lock &pool->lock/1 irq_context: 0 (wq_completion)events async_lookup_work rcu_read_lock &pool->lock/1 &obj_hash[i].lock irq_context: 0 (wq_completion)events async_lookup_work rcu_read_lock &pool->lock/1 pool_lock#2 irq_context: 0 (wq_completion)events async_lookup_work rcu_read_lock &pool->lock/1 &p->pi_lock irq_context: 0 (wq_completion)events async_lookup_work rcu_read_lock &pool->lock/1 &p->pi_lock &rq->__lock irq_context: 0 (wq_completion)events async_lookup_work rcu_read_lock &pool->lock/1 &p->pi_lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)events async_lookup_work &x->wait#17 irq_context: 0 (wq_completion)events async_lookup_work &pool->lock irq_context: 0 (wq_completion)events async_lookup_work &rq->__lock irq_context: 0 (wq_completion)events async_lookup_work &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 &grp->list_mutex/1 register_lock#3 irq_context: 0 &grp->list_mutex/1 fs_reclaim irq_context: 0 &grp->list_mutex/1 fs_reclaim mmu_notifier_invalidate_range_start irq_context: 0 &grp->list_mutex/1 pool_lock#2 irq_context: 0 &dev->mutex snd_card_mutex irq_context: 0 &dev->mutex &entry->access irq_context: 0 &dev->mutex info_mutex irq_context: 0 &dev->mutex info_mutex proc_subdir_lock irq_context: 0 &dev->mutex info_mutex fs_reclaim irq_context: 0 &dev->mutex info_mutex fs_reclaim mmu_notifier_invalidate_range_start irq_context: 0 &dev->mutex info_mutex pool_lock#2 irq_context: 0 &dev->mutex info_mutex proc_inum_ida.xa_lock irq_context: 0 &dev->mutex info_mutex proc_subdir_lock irq_context: 0 (wq_completion)events async_lookup_work running_helpers_waitq.lock irq_context: 0 (wq_completion)events async_lookup_work rcu_read_lock &pool->lock irq_context: 0 (wq_completion)events async_lookup_work rcu_read_lock &pool->lock &obj_hash[i].lock irq_context: 0 (wq_completion)events async_lookup_work autoload_work irq_context: 0 (wq_completion)events async_lookup_work &x->wait#10 irq_context: 0 (wq_completion)events async_lookup_work &pool->lock &p->pi_lock irq_context: 0 (wq_completion)events async_lookup_work &pool->lock &p->pi_lock &rq->__lock irq_context: 0 (wq_completion)events async_lookup_work &pool->lock &p->pi_lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)events autoload_work irq_context: 0 (wq_completion)events autoload_work &k->list_lock irq_context: 0 (wq_completion)events autoload_work &k->k_lock irq_context: 0 (wq_completion)events (work_completion)(&barr->work) irq_context: 0 (wq_completion)events (work_completion)(&barr->work) &x->wait#10 irq_context: 0 (wq_completion)events (work_completion)(&barr->work) &x->wait#10 &p->pi_lock irq_context: 0 (wq_completion)events (work_completion)(&barr->work) &x->wait#10 &p->pi_lock &rq->__lock irq_context: 0 (wq_completion)events (work_completion)(&barr->work) &x->wait#10 &p->pi_lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 &dev->mutex &card->controls_rwsem irq_context: 0 &dev->mutex &card->controls_rwsem &xa->xa_lock#13 irq_context: 0 &dev->mutex &card->controls_rwsem &xa->xa_lock#13 pool_lock#2 irq_context: 0 &dev->mutex &card->controls_rwsem fs_reclaim irq_context: 0 &dev->mutex &card->controls_rwsem fs_reclaim mmu_notifier_invalidate_range_start irq_context: 0 &dev->mutex &card->controls_rwsem &card->ctl_files_rwlock irq_context: 0 &dev->mutex &card->controls_rwsem snd_ctl_layer_rwsem irq_context: 0 &dev->mutex &card->controls_rwsem &xa->xa_lock#13 &c->lock irq_context: 0 &dev->mutex &card->controls_rwsem &xa->xa_lock#13 &pcp->lock &zone->lock irq_context: 0 &dev->mutex &card->controls_rwsem &xa->xa_lock#13 &zone->lock irq_context: 0 &dev->mutex &card->controls_rwsem &xa->xa_lock#13 &____s->seqcount irq_context: 0 &dev->mutex &card->controls_rwsem &rq->__lock irq_context: 0 &dev->mutex subsys mutex#77 irq_context: 0 &dev->mutex subsys mutex#77 &k->k_lock irq_context: 0 &dev->mutex register_mutex#2 irq_context: 0 &dev->mutex register_mutex#2 fs_reclaim irq_context: 0 &dev->mutex register_mutex#2 fs_reclaim mmu_notifier_invalidate_range_start irq_context: 0 &dev->mutex register_mutex#2 pool_lock#2 irq_context: 0 &dev->mutex register_mutex#2 sound_mutex irq_context: 0 &dev->mutex register_mutex#2 sound_mutex fs_reclaim irq_context: 0 &dev->mutex register_mutex#2 sound_mutex fs_reclaim mmu_notifier_invalidate_range_start irq_context: 0 &dev->mutex register_mutex#2 sound_mutex &c->lock irq_context: 0 &dev->mutex register_mutex#2 sound_mutex &pcp->lock &zone->lock irq_context: 0 &dev->mutex register_mutex#2 sound_mutex &zone->lock irq_context: 0 &dev->mutex register_mutex#2 sound_mutex &____s->seqcount irq_context: 0 &dev->mutex register_mutex#2 sound_mutex pool_lock#2 irq_context: 0 &dev->mutex register_mutex#2 sound_mutex &k->list_lock irq_context: 0 &dev->mutex register_mutex#2 sound_mutex lock irq_context: 0 &dev->mutex register_mutex#2 sound_mutex lock kernfs_idr_lock irq_context: 0 &dev->mutex register_mutex#2 sound_mutex &root->kernfs_rwsem irq_context: 0 &dev->mutex register_mutex#2 sound_mutex &root->kernfs_rwsem &root->kernfs_iattr_rwsem irq_context: 0 &dev->mutex register_mutex#2 sound_mutex bus_type_sem irq_context: 0 &dev->mutex register_mutex#2 sound_mutex sysfs_symlink_target_lock irq_context: 0 &dev->mutex register_mutex#2 sound_mutex &root->kernfs_rwsem irq_context: 0 &dev->mutex register_mutex#2 sound_mutex &dev->power.lock irq_context: 0 &dev->mutex register_mutex#2 sound_mutex dpm_list_mtx irq_context: 0 &dev->mutex register_mutex#2 sound_mutex req_lock irq_context: 0 &dev->mutex register_mutex#2 sound_mutex &p->pi_lock irq_context: 0 &dev->mutex register_mutex#2 sound_mutex &p->pi_lock &rq->__lock irq_context: 0 &dev->mutex register_mutex#2 sound_mutex &p->pi_lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 &dev->mutex register_mutex#2 sound_mutex &rq->__lock irq_context: 0 &dev->mutex register_mutex#2 sound_mutex &x->wait#11 irq_context: 0 &dev->mutex register_mutex#2 sound_mutex &obj_hash[i].lock irq_context: 0 &dev->mutex register_mutex#2 sound_mutex uevent_sock_mutex irq_context: 0 &dev->mutex register_mutex#2 sound_mutex rcu_read_lock &pool->lock/1 irq_context: 0 &dev->mutex register_mutex#2 sound_mutex rcu_read_lock &pool->lock/1 &obj_hash[i].lock irq_context: 0 &dev->mutex register_mutex#2 sound_mutex rcu_read_lock &pool->lock/1 pool_lock#2 irq_context: 0 &dev->mutex register_mutex#2 sound_mutex rcu_read_lock &pool->lock/1 &p->pi_lock irq_context: 0 &dev->mutex register_mutex#2 sound_mutex rcu_read_lock &pool->lock/1 &p->pi_lock &rq->__lock irq_context: 0 &dev->mutex register_mutex#2 sound_mutex rcu_read_lock &pool->lock/1 &p->pi_lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 &dev->mutex register_mutex#2 sound_mutex running_helpers_waitq.lock irq_context: 0 &dev->mutex register_mutex#2 sound_mutex &k->k_lock irq_context: 0 &dev->mutex register_mutex#2 sound_mutex subsys mutex#77 irq_context: 0 &dev->mutex register_mutex#2 sound_mutex subsys mutex#77 &k->k_lock irq_context: 0 &dev->mutex register_mutex#2 &obj_hash[i].lock irq_context: 0 &dev->mutex register_mutex#2 register_mutex irq_context: 0 &dev->mutex register_mutex#2 &c->lock irq_context: 0 &dev->mutex register_mutex#2 &pcp->lock &zone->lock irq_context: 0 &dev->mutex register_mutex#2 &zone->lock irq_context: 0 &dev->mutex register_mutex#2 &____s->seqcount irq_context: 0 &dev->mutex register_mutex#2 sound_mutex &cfs_rq->removed.lock irq_context: 0 &dev->mutex register_mutex#2 sound_mutex &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 &dev->mutex register_mutex#2 sound_mutex &rq->__lock &cfs_rq->removed.lock irq_context: 0 &dev->mutex register_mutex#2 &obj_hash[i].lock pool_lock irq_context: 0 &dev->mutex register_mutex#2 rcu_read_lock pool_lock#2 irq_context: 0 &dev->mutex register_mutex#2 sound_oss_mutex irq_context: 0 &dev->mutex register_mutex#2 sound_oss_mutex fs_reclaim irq_context: 0 &dev->mutex register_mutex#2 sound_oss_mutex fs_reclaim mmu_notifier_invalidate_range_start irq_context: 0 &dev->mutex register_mutex#2 sound_oss_mutex pool_lock#2 irq_context: 0 &dev->mutex register_mutex#2 sound_oss_mutex sound_loader_lock irq_context: 0 &dev->mutex register_mutex#2 sound_oss_mutex &x->wait#9 irq_context: 0 &dev->mutex register_mutex#2 sound_oss_mutex &obj_hash[i].lock irq_context: 0 &dev->mutex register_mutex#2 sound_oss_mutex &k->list_lock irq_context: 0 &dev->mutex register_mutex#2 sound_oss_mutex lock irq_context: 0 &dev->mutex register_mutex#2 sound_oss_mutex lock kernfs_idr_lock irq_context: 0 &dev->mutex register_mutex#2 sound_oss_mutex &root->kernfs_rwsem irq_context: 0 &dev->mutex register_mutex#2 sound_oss_mutex &root->kernfs_rwsem &root->kernfs_iattr_rwsem irq_context: 0 &dev->mutex register_mutex#2 sound_oss_mutex bus_type_sem irq_context: 0 &dev->mutex register_mutex#2 sound_oss_mutex sysfs_symlink_target_lock irq_context: 0 &dev->mutex register_mutex#2 sound_oss_mutex &c->lock irq_context: 0 &dev->mutex register_mutex#2 sound_oss_mutex &pcp->lock &zone->lock irq_context: 0 &dev->mutex register_mutex#2 sound_oss_mutex &zone->lock irq_context: 0 &dev->mutex register_mutex#2 sound_oss_mutex &____s->seqcount irq_context: 0 &dev->mutex register_mutex#2 sound_oss_mutex &root->kernfs_rwsem irq_context: 0 &dev->mutex register_mutex#2 sound_oss_mutex &dev->power.lock irq_context: 0 &dev->mutex register_mutex#2 sound_oss_mutex dpm_list_mtx irq_context: 0 &dev->mutex register_mutex#2 sound_oss_mutex req_lock irq_context: 0 &dev->mutex register_mutex#2 sound_oss_mutex &p->pi_lock irq_context: 0 &dev->mutex register_mutex#2 sound_oss_mutex &p->pi_lock &rq->__lock irq_context: 0 &dev->mutex register_mutex#2 sound_oss_mutex &p->pi_lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 &dev->mutex register_mutex#2 sound_oss_mutex &x->wait#11 irq_context: 0 &dev->mutex register_mutex#2 sound_oss_mutex &rq->__lock irq_context: 0 &dev->mutex register_mutex#2 sound_oss_mutex &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 &dev->mutex register_mutex#2 sound_oss_mutex uevent_sock_mutex irq_context: 0 &dev->mutex register_mutex#2 sound_oss_mutex rcu_read_lock &pool->lock/1 irq_context: 0 &dev->mutex register_mutex#2 sound_oss_mutex rcu_read_lock &pool->lock/1 &obj_hash[i].lock irq_context: 0 &dev->mutex register_mutex#2 sound_oss_mutex rcu_read_lock &pool->lock/1 pool_lock#2 irq_context: 0 &dev->mutex register_mutex#2 sound_oss_mutex rcu_read_lock &pool->lock/1 &p->pi_lock irq_context: 0 &dev->mutex register_mutex#2 sound_oss_mutex running_helpers_waitq.lock irq_context: 0 &dev->mutex register_mutex#2 sound_oss_mutex &k->k_lock irq_context: 0 &dev->mutex register_mutex#2 sound_oss_mutex subsys mutex#77 irq_context: 0 &dev->mutex register_mutex#2 sound_oss_mutex subsys mutex#77 &k->k_lock irq_context: 0 &dev->mutex register_mutex#2 sound_oss_mutex lock kernfs_idr_lock pool_lock#2 irq_context: 0 &dev->mutex register_mutex#2 sound_oss_mutex rcu_read_lock &rq->__lock irq_context: 0 &dev->mutex register_mutex#2 sound_oss_mutex rcu_read_lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 &dev->mutex register_mutex#2 sound_oss_mutex rcu_read_lock &pool->lock/1 &p->pi_lock &rq->__lock irq_context: 0 &dev->mutex register_mutex#2 sound_oss_mutex rcu_read_lock &pool->lock/1 &p->pi_lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 &dev->mutex register_mutex#2 strings irq_context: 0 &dev->mutex register_mutex#2 strings fs_reclaim irq_context: 0 &dev->mutex register_mutex#2 strings fs_reclaim mmu_notifier_invalidate_range_start irq_context: 0 &dev->mutex register_mutex#2 strings pool_lock#2 irq_context: 0 &dev->mutex register_mutex#2 &entry->access irq_context: 0 &dev->mutex register_mutex#2 info_mutex irq_context: 0 &dev->mutex sound_mutex irq_context: 0 &dev->mutex sound_mutex fs_reclaim irq_context: 0 &dev->mutex sound_mutex fs_reclaim mmu_notifier_invalidate_range_start irq_context: 0 &dev->mutex sound_mutex pool_lock#2 irq_context: 0 &dev->mutex sound_mutex &k->list_lock irq_context: 0 &dev->mutex sound_mutex lock irq_context: 0 &dev->mutex sound_mutex lock kernfs_idr_lock irq_context: 0 &dev->mutex sound_mutex &root->kernfs_rwsem irq_context: 0 &dev->mutex sound_mutex &root->kernfs_rwsem &root->kernfs_iattr_rwsem irq_context: 0 &dev->mutex sound_mutex bus_type_sem irq_context: 0 &dev->mutex sound_mutex &____s->seqcount irq_context: 0 &dev->mutex sound_mutex sysfs_symlink_target_lock irq_context: 0 &dev->mutex sound_mutex &root->kernfs_rwsem irq_context: 0 &dev->mutex sound_mutex &c->lock irq_context: 0 &dev->mutex sound_mutex &dev->power.lock irq_context: 0 &dev->mutex sound_mutex dpm_list_mtx irq_context: 0 &dev->mutex sound_mutex req_lock irq_context: 0 &dev->mutex sound_mutex &p->pi_lock irq_context: 0 &dev->mutex sound_mutex &p->pi_lock &rq->__lock irq_context: 0 &dev->mutex sound_mutex &p->pi_lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 &dev->mutex sound_mutex &x->wait#11 irq_context: 0 &dev->mutex sound_mutex &rq->__lock irq_context: 0 &dev->mutex sound_mutex &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 &dev->mutex sound_mutex &obj_hash[i].lock irq_context: 0 &dev->mutex sound_mutex uevent_sock_mutex irq_context: 0 &dev->mutex sound_mutex rcu_read_lock &pool->lock/1 irq_context: 0 &dev->mutex sound_mutex rcu_read_lock &pool->lock/1 &obj_hash[i].lock irq_context: 0 &dev->mutex sound_mutex rcu_read_lock &pool->lock/1 pool_lock#2 irq_context: 0 &dev->mutex sound_mutex rcu_read_lock &pool->lock/1 &p->pi_lock irq_context: 0 &dev->mutex sound_mutex running_helpers_waitq.lock irq_context: 0 &dev->mutex sound_mutex &k->k_lock irq_context: 0 &dev->mutex sound_mutex subsys mutex#77 irq_context: 0 &dev->mutex sound_mutex subsys mutex#77 &k->k_lock irq_context: 0 &dev->mutex &card->controls_rwsem irq_context: 0 &dev->mutex &card->controls_rwsem snd_ctl_layer_rwsem irq_context: 0 &dev->mutex &card->controls_rwsem snd_ctl_layer_rwsem snd_ctl_led_mutex irq_context: 0 &dev->mutex &card->controls_rwsem snd_ctl_layer_rwsem fs_reclaim irq_context: 0 &dev->mutex &card->controls_rwsem snd_ctl_layer_rwsem fs_reclaim mmu_notifier_invalidate_range_start irq_context: 0 &dev->mutex &card->controls_rwsem snd_ctl_layer_rwsem &c->lock irq_context: 0 &dev->mutex &card->controls_rwsem snd_ctl_layer_rwsem &____s->seqcount irq_context: 0 &dev->mutex &card->controls_rwsem snd_ctl_layer_rwsem pool_lock#2 irq_context: 0 &dev->mutex &card->controls_rwsem snd_ctl_layer_rwsem &x->wait#9 irq_context: 0 &dev->mutex &card->controls_rwsem snd_ctl_layer_rwsem &obj_hash[i].lock irq_context: 0 &dev->mutex &card->controls_rwsem snd_ctl_layer_rwsem &k->list_lock irq_context: 0 &dev->mutex &card->controls_rwsem snd_ctl_layer_rwsem lock irq_context: 0 &dev->mutex &card->controls_rwsem snd_ctl_layer_rwsem lock kernfs_idr_lock irq_context: 0 &dev->mutex &card->controls_rwsem snd_ctl_layer_rwsem &root->kernfs_rwsem irq_context: 0 &dev->mutex &card->controls_rwsem snd_ctl_layer_rwsem &root->kernfs_rwsem &root->kernfs_iattr_rwsem irq_context: 0 &dev->mutex &card->controls_rwsem snd_ctl_layer_rwsem bus_type_sem irq_context: 0 &dev->mutex &card->controls_rwsem snd_ctl_layer_rwsem &root->kernfs_rwsem irq_context: 0 &dev->mutex &card->controls_rwsem snd_ctl_layer_rwsem &dev->power.lock irq_context: 0 &dev->mutex &card->controls_rwsem snd_ctl_layer_rwsem dpm_list_mtx irq_context: 0 &dev->mutex &card->controls_rwsem snd_ctl_layer_rwsem &k->k_lock irq_context: 0 &dev->mutex &card->controls_rwsem snd_ctl_layer_rwsem sysfs_symlink_target_lock irq_context: 0 &dev->mutex info_mutex &c->lock irq_context: 0 &dev->mutex info_mutex &____s->seqcount irq_context: 0 &dev->mutex sound_oss_mutex irq_context: 0 &dev->mutex sound_oss_mutex fs_reclaim irq_context: 0 &dev->mutex sound_oss_mutex fs_reclaim mmu_notifier_invalidate_range_start irq_context: 0 &dev->mutex sound_oss_mutex pool_lock#2 irq_context: 0 &dev->mutex sound_oss_mutex sound_loader_lock irq_context: 0 &dev->mutex sound_oss_mutex &x->wait#9 irq_context: 0 &dev->mutex sound_oss_mutex &obj_hash[i].lock irq_context: 0 &dev->mutex sound_oss_mutex &c->lock irq_context: 0 &dev->mutex sound_oss_mutex &____s->seqcount irq_context: 0 &dev->mutex sound_oss_mutex &k->list_lock irq_context: 0 &dev->mutex sound_oss_mutex lock irq_context: 0 &dev->mutex sound_oss_mutex lock kernfs_idr_lock irq_context: 0 &dev->mutex sound_oss_mutex &root->kernfs_rwsem irq_context: 0 &dev->mutex sound_oss_mutex &root->kernfs_rwsem &root->kernfs_iattr_rwsem irq_context: 0 &dev->mutex sound_oss_mutex bus_type_sem irq_context: 0 &dev->mutex sound_oss_mutex sysfs_symlink_target_lock irq_context: 0 &dev->mutex sound_oss_mutex &root->kernfs_rwsem irq_context: 0 &dev->mutex sound_oss_mutex &dev->power.lock irq_context: 0 &dev->mutex sound_oss_mutex dpm_list_mtx irq_context: 0 &dev->mutex sound_oss_mutex lock kernfs_idr_lock pool_lock#2 irq_context: 0 &dev->mutex sound_oss_mutex &pcp->lock &zone->lock irq_context: 0 &dev->mutex sound_oss_mutex &zone->lock irq_context: 0 &dev->mutex sound_oss_mutex req_lock irq_context: 0 &dev->mutex sound_oss_mutex &p->pi_lock irq_context: 0 &dev->mutex sound_oss_mutex &p->pi_lock &rq->__lock irq_context: 0 &dev->mutex sound_oss_mutex &p->pi_lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 &dev->mutex sound_oss_mutex &rq->__lock irq_context: 0 &dev->mutex sound_oss_mutex &x->wait#11 irq_context: 0 &dev->mutex sound_oss_mutex uevent_sock_mutex irq_context: 0 &dev->mutex sound_oss_mutex rcu_read_lock &pool->lock/1 irq_context: 0 &dev->mutex sound_oss_mutex rcu_read_lock &pool->lock/1 &obj_hash[i].lock irq_context: 0 &dev->mutex sound_oss_mutex rcu_read_lock &pool->lock/1 pool_lock#2 irq_context: 0 &dev->mutex sound_oss_mutex rcu_read_lock &pool->lock/1 &p->pi_lock irq_context: 0 &dev->mutex sound_oss_mutex running_helpers_waitq.lock irq_context: 0 &dev->mutex sound_oss_mutex &k->k_lock irq_context: 0 &dev->mutex sound_oss_mutex subsys mutex#77 irq_context: 0 &dev->mutex sound_oss_mutex subsys mutex#77 &k->k_lock irq_context: 0 &dev->mutex strings irq_context: 0 &dev->mutex strings fs_reclaim irq_context: 0 &dev->mutex strings fs_reclaim mmu_notifier_invalidate_range_start irq_context: 0 &dev->mutex strings pool_lock#2 irq_context: 0 &dev->mutex &card->controls_rwsem fs_reclaim irq_context: 0 &dev->mutex &card->controls_rwsem fs_reclaim mmu_notifier_invalidate_range_start irq_context: 0 &dev->mutex &card->controls_rwsem pool_lock#2 irq_context: 0 &dev->mutex &card->controls_rwsem &c->lock irq_context: 0 &dev->mutex &card->controls_rwsem &pcp->lock &zone->lock irq_context: 0 &dev->mutex &card->controls_rwsem &zone->lock irq_context: 0 &dev->mutex &card->controls_rwsem &____s->seqcount irq_context: 0 &dev->mutex &card->controls_rwsem &xa->xa_lock#13 &pcp->lock &zone->lock &____s->seqcount irq_context: 0 &dev->mutex register_mutex#2 sound_mutex &pcp->lock &zone->lock &____s->seqcount irq_context: 0 &dev->mutex register_mutex#2 sound_mutex rcu_read_lock pool_lock#2 irq_context: 0 &dev->mutex register_mutex#2 sound_oss_mutex rcu_read_lock pool_lock#2 irq_context: 0 &dev->mutex register_mutex#2 sound_oss_mutex &pcp->lock &zone->lock &____s->seqcount irq_context: 0 &dev->mutex register_mutex#2 sound_oss_mutex &obj_hash[i].lock pool_lock irq_context: 0 &dev->mutex sound_mutex &pcp->lock &zone->lock irq_context: 0 &dev->mutex sound_mutex &zone->lock irq_context: 0 &dev->mutex sound_mutex rcu_read_lock pool_lock#2 irq_context: 0 &dev->mutex info_mutex &pcp->lock &zone->lock irq_context: 0 &dev->mutex info_mutex &zone->lock irq_context: 0 &dev->mutex &root->kernfs_rwsem &rq->__lock irq_context: 0 &dev->mutex register_mutex#5 irq_context: 0 &dev->mutex sound_mutex rcu_read_lock &pool->lock/1 &p->pi_lock &rq->__lock irq_context: 0 &dev->mutex sound_mutex rcu_read_lock &pool->lock/1 &p->pi_lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 &dev->mutex register_mutex#3 irq_context: 0 &dev->mutex register_mutex#3 fs_reclaim irq_context: 0 &dev->mutex register_mutex#3 fs_reclaim mmu_notifier_invalidate_range_start irq_context: 0 &dev->mutex register_mutex#3 pool_lock#2 irq_context: 0 &dev->mutex register_mutex#3 clients_lock irq_context: 0 &dev->mutex clients_lock irq_context: 0 &dev->mutex &client->ports_lock irq_context: 0 &dev->mutex &grp->list_mutex/1 irq_context: 0 &dev->mutex &grp->list_mutex/1 clients_lock irq_context: 0 &dev->mutex &grp->list_mutex/1 &client->ports_lock irq_context: 0 &dev->mutex &client->ports_mutex irq_context: 0 &dev->mutex &client->ports_mutex &client->ports_lock irq_context: 0 &dev->mutex &grp->list_mutex/1 register_lock#3 irq_context: 0 &dev->mutex &grp->list_mutex/1 fs_reclaim irq_context: 0 &dev->mutex &grp->list_mutex/1 fs_reclaim mmu_notifier_invalidate_range_start irq_context: 0 &dev->mutex &grp->list_mutex/1 pool_lock#2 irq_context: 0 &dev->mutex sound_oss_mutex &root->kernfs_rwsem &rq->__lock irq_context: 0 &dev->mutex sound_oss_mutex &root->kernfs_rwsem &cfs_rq->removed.lock irq_context: 0 &dev->mutex sound_oss_mutex &root->kernfs_rwsem &obj_hash[i].lock irq_context: 0 &dev->mutex sound_oss_mutex &p->pi_lock &rq->__lock &cfs_rq->removed.lock irq_context: 0 &dev->mutex sound_oss_mutex rcu_read_lock &pool->lock/1 &p->pi_lock &rq->__lock irq_context: 0 &dev->mutex sound_oss_mutex rcu_read_lock &pool->lock/1 &p->pi_lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 &dev->mutex sound_mutex lock kernfs_idr_lock pool_lock#2 irq_context: 0 &dev->mutex sound_oss_mutex rcu_read_lock &pool->lock/1 &p->pi_lock &rq->__lock &cfs_rq->removed.lock irq_context: 0 &dev->mutex sound_mutex rcu_read_lock &pool->lock/1 &p->pi_lock &rq->__lock &cfs_rq->removed.lock irq_context: 0 &mm->mmap_lock rcu_read_lock rcu_node_0 irq_context: 0 &mm->mmap_lock rcu_read_lock &rq->__lock irq_context: 0 &dev->mutex sound_mutex &cfs_rq->removed.lock irq_context: 0 &dev->mutex &card->controls_rwsem snd_ctl_layer_rwsem &obj_hash[i].lock pool_lock irq_context: softirq &(&krcp->monitor_work)->timer rcu_read_lock &pool->lock &p->pi_lock &rq->__lock &cfs_rq->removed.lock irq_context: 0 pernet_ops_rwsem rtnl_mutex failover_lock irq_context: 0 llc_sap_list_lock irq_context: 0 llc_sap_list_lock pool_lock#2 irq_context: 0 act_id_mutex irq_context: 0 act_id_mutex fs_reclaim irq_context: 0 act_id_mutex fs_reclaim mmu_notifier_invalidate_range_start irq_context: 0 act_id_mutex pool_lock#2 irq_context: 0 act_mod_lock irq_context: 0 ife_mod_lock irq_context: 0 pernet_ops_rwsem nf_connlabels_lock irq_context: 0 cls_mod_lock irq_context: 0 ematch_mod_lock irq_context: 0 sock_diag_table_mutex irq_context: 0 nfnl_subsys_acct irq_context: 0 nfnl_subsys_queue irq_context: 0 nfnl_subsys_ulog irq_context: 0 nf_log_mutex irq_context: 0 nfnl_subsys_osf irq_context: 0 nf_sockopt_mutex irq_context: 0 nfnl_subsys_ctnetlink irq_context: 0 nfnl_subsys_ctnetlink_exp irq_context: 0 pernet_ops_rwsem nf_ct_ecache_mutex irq_context: 0 nfnl_subsys_cttimeout irq_context: 0 nfnl_subsys_cthelper irq_context: 0 nf_ct_helper_mutex irq_context: 0 rcu_state.exp_mutex &rnp->exp_wq[1] irq_context: 0 nf_conntrack_expect_lock irq_context: 0 net_rwsem irq_context: 0 nf_conntrack_mutex irq_context: 0 pernet_ops_rwsem nf_log_mutex irq_context: 0 pernet_ops_rwsem batched_entropy_u8.lock irq_context: 0 pernet_ops_rwsem kfence_freelist_lock irq_context: 0 nf_ct_nat_helpers_mutex irq_context: 0 nfnl_subsys_nftables irq_context: 0 nfnl_subsys_nftcompat irq_context: 0 masq_mutex irq_context: 0 masq_mutex pernet_ops_rwsem irq_context: 0 masq_mutex pernet_ops_rwsem rtnl_mutex irq_context: 0 masq_mutex (inetaddr_chain).rwsem irq_context: 0 masq_mutex inet6addr_chain.lock irq_context: 0 &xt[i].mutex irq_context: 0 pernet_ops_rwsem rtnl_mutex &tn->lock irq_context: 0 subsys mutex#78 irq_context: 0 subsys mutex#78 &k->k_lock irq_context: 0 nfnl_subsys_ipset irq_context: 0 ip_set_type_mutex irq_context: 0 pernet_ops_rwsem ipvs->est_mutex irq_context: 0 pernet_ops_rwsem ipvs->est_mutex fs_reclaim irq_context: 0 pernet_ops_rwsem ipvs->est_mutex fs_reclaim mmu_notifier_invalidate_range_start irq_context: 0 pernet_ops_rwsem ipvs->est_mutex pool_lock#2 irq_context: 0 pernet_ops_rwsem ipvs->est_mutex pcpu_alloc_mutex irq_context: 0 pernet_ops_rwsem ipvs->est_mutex pcpu_alloc_mutex pcpu_lock irq_context: 0 pernet_ops_rwsem &base->lock irq_context: 0 pernet_ops_rwsem &base->lock &obj_hash[i].lock irq_context: 0 ip_vs_sched_mutex irq_context: 0 pernet_ops_rwsem __ip_vs_app_mutex irq_context: 0 pernet_ops_rwsem __ip_vs_app_mutex fs_reclaim irq_context: 0 pernet_ops_rwsem __ip_vs_app_mutex fs_reclaim mmu_notifier_invalidate_range_start irq_context: 0 pernet_ops_rwsem __ip_vs_app_mutex pool_lock#2 irq_context: 0 ip_vs_pe_mutex irq_context: 0 tunnel4_mutex irq_context: 0 pernet_ops_rwsem rtnl_mutex lock kernfs_idr_lock &c->lock irq_context: 0 pernet_ops_rwsem rtnl_mutex lock kernfs_idr_lock &pcp->lock &zone->lock irq_context: 0 pernet_ops_rwsem rtnl_mutex lock kernfs_idr_lock &zone->lock irq_context: 0 pernet_ops_rwsem rtnl_mutex lock kernfs_idr_lock &____s->seqcount irq_context: 0 pernet_ops_rwsem rtnl_mutex &obj_hash[i].lock pool_lock irq_context: 0 pernet_ops_rwsem rtnl_mutex batched_entropy_u32.lock crngs.lock irq_context: 0 pernet_ops_rwsem net_generic_ids.xa_lock pool_lock#2 irq_context: 0 xfrm4_protocol_mutex irq_context: 0 pernet_ops_rwsem &cfs_rq->removed.lock irq_context: 0 &xt[i].mutex fs_reclaim irq_context: 0 &xt[i].mutex fs_reclaim mmu_notifier_invalidate_range_start irq_context: 0 &xt[i].mutex pool_lock#2 irq_context: 0 inet_diag_table_mutex irq_context: 0 xfrm_km_lock irq_context: 0 xfrm_translator_lock irq_context: 0 xfrm6_protocol_mutex irq_context: 0 tunnel6_mutex irq_context: 0 xfrm_if_cb_lock irq_context: 0 inetsw6_lock irq_context: 0 &hashinfo->lock#2 irq_context: 0 pernet_ops_rwsem &hashinfo->lock#2 irq_context: 0 pernet_ops_rwsem &net->ipv6.ip6addrlbl_table.lock irq_context: 0 pernet_ops_rwsem rtnl_mutex &idev->mc_lock irq_context: 0 pernet_ops_rwsem rtnl_mutex &idev->mc_lock fs_reclaim irq_context: 0 pernet_ops_rwsem rtnl_mutex &idev->mc_lock fs_reclaim mmu_notifier_invalidate_range_start irq_context: 0 pernet_ops_rwsem rtnl_mutex &idev->mc_lock pool_lock#2 irq_context: 0 pernet_ops_rwsem rtnl_mutex &idev->mc_lock &obj_hash[i].lock irq_context: 0 pernet_ops_rwsem rtnl_mutex &idev->mc_lock &c->lock irq_context: 0 pernet_ops_rwsem rtnl_mutex &idev->mc_lock &____s->seqcount irq_context: 0 pernet_ops_rwsem rtnl_mutex &idev->mc_lock &dev_addr_list_lock_key irq_context: 0 pernet_ops_rwsem rtnl_mutex &idev->mc_lock &dev_addr_list_lock_key pool_lock#2 irq_context: 0 pernet_ops_rwsem rtnl_mutex &idev->mc_lock &obj_hash[i].lock pool_lock irq_context: 0 pernet_ops_rwsem rtnl_mutex &idev->mc_lock _xmit_ETHER irq_context: 0 pernet_ops_rwsem rtnl_mutex rcu_read_lock &pool->lock irq_context: 0 pernet_ops_rwsem rtnl_mutex rcu_read_lock &pool->lock &obj_hash[i].lock irq_context: 0 pernet_ops_rwsem rtnl_mutex rcu_read_lock &pool->lock &p->pi_lock irq_context: 0 pernet_ops_rwsem rtnl_mutex rcu_read_lock &pool->lock &p->pi_lock &rq->__lock irq_context: 0 pernet_ops_rwsem rtnl_mutex rcu_read_lock &pool->lock &p->pi_lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 pernet_ops_rwsem rtnl_mutex &idev->mc_lock _xmit_ETHER &c->lock irq_context: 0 pernet_ops_rwsem rtnl_mutex &idev->mc_lock _xmit_ETHER &pcp->lock &zone->lock irq_context: 0 pernet_ops_rwsem rtnl_mutex &idev->mc_lock _xmit_ETHER &zone->lock irq_context: 0 pernet_ops_rwsem rtnl_mutex &idev->mc_lock _xmit_ETHER &____s->seqcount irq_context: 0 pernet_ops_rwsem rtnl_mutex &idev->mc_lock _xmit_ETHER pool_lock#2 irq_context: 0 pernet_ops_rwsem rtnl_mutex &idev->mc_lock &pcp->lock &zone->lock irq_context: 0 pernet_ops_rwsem rtnl_mutex &idev->mc_lock &zone->lock irq_context: 0 (wq_completion)ipv6_addrconf irq_context: 0 (wq_completion)ipv6_addrconf (work_completion)(&(&net->ipv6.addr_chk_work)->work) irq_context: 0 (wq_completion)ipv6_addrconf (work_completion)(&(&net->ipv6.addr_chk_work)->work) rtnl_mutex irq_context: 0 (wq_completion)ipv6_addrconf (work_completion)(&(&net->ipv6.addr_chk_work)->work) rtnl_mutex rcu_read_lock_bh &obj_hash[i].lock irq_context: 0 (wq_completion)ipv6_addrconf (work_completion)(&(&net->ipv6.addr_chk_work)->work) rtnl_mutex rcu_read_lock_bh &base->lock irq_context: 0 (wq_completion)ipv6_addrconf (work_completion)(&(&net->ipv6.addr_chk_work)->work) rtnl_mutex rcu_read_lock_bh &base->lock &obj_hash[i].lock irq_context: 0 (crypto_chain).rwsem irq_context: 0 (crypto_chain).rwsem fs_reclaim irq_context: 0 (crypto_chain).rwsem fs_reclaim mmu_notifier_invalidate_range_start irq_context: 0 (crypto_chain).rwsem &c->lock irq_context: 0 (crypto_chain).rwsem &____s->seqcount irq_context: 0 (crypto_chain).rwsem pool_lock#2 irq_context: 0 (crypto_chain).rwsem kthread_create_lock irq_context: 0 (crypto_chain).rwsem &p->pi_lock irq_context: 0 (crypto_chain).rwsem &p->pi_lock &rq->__lock irq_context: 0 (crypto_chain).rwsem &p->pi_lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (crypto_chain).rwsem &x->wait irq_context: 0 (crypto_chain).rwsem &rq->__lock irq_context: 0 (crypto_chain).rwsem &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (crypto_chain).rwsem &obj_hash[i].lock irq_context: 0 (crypto_chain).rwsem &pcp->lock &zone->lock irq_context: 0 (crypto_chain).rwsem &zone->lock irq_context: 0 (crypto_chain).rwsem rcu_read_lock pool_lock#2 irq_context: 0 &x->wait#21 irq_context: 0 &x->wait#21 &p->pi_lock irq_context: 0 &p->alloc_lock &x->wait irq_context: 0 pernet_ops_rwsem rtnl_mutex uevent_sock_mutex rcu_read_lock pool_lock#2 irq_context: 0 pernet_ops_rwsem rtnl_mutex &idev->mc_lock rcu_read_lock pool_lock#2 irq_context: 0 pernet_ops_rwsem rtnl_mutex rcu_read_lock &ndev->lock irq_context: 0 pernet_ops_rwsem rtnl_mutex batched_entropy_u8.lock irq_context: 0 pernet_ops_rwsem rtnl_mutex kfence_freelist_lock irq_context: 0 stp_proto_mutex irq_context: 0 stp_proto_mutex llc_sap_list_lock irq_context: 0 stp_proto_mutex llc_sap_list_lock pool_lock#2 irq_context: 0 switchdev_notif_chain.lock irq_context: 0 (switchdev_blocking_notif_chain).rwsem irq_context: 0 br_ioctl_mutex irq_context: 0 nf_ct_proto_mutex irq_context: 0 ebt_mutex irq_context: 0 ebt_mutex fs_reclaim irq_context: 0 ebt_mutex fs_reclaim mmu_notifier_invalidate_range_start irq_context: 0 ebt_mutex pool_lock#2 irq_context: 0 dsa_tag_drivers_lock irq_context: 0 rtnl_mutex &tn->lock irq_context: 0 rtnl_mutex lock kernfs_idr_lock &c->lock irq_context: 0 rtnl_mutex lock kernfs_idr_lock &____s->seqcount irq_context: 0 rtnl_mutex lock kernfs_idr_lock &pcp->lock &zone->lock irq_context: 0 rtnl_mutex lock kernfs_idr_lock &zone->lock irq_context: 0 rtnl_mutex fill_pool_map-wait-type-override pool_lock#2 irq_context: 0 rtnl_mutex fill_pool_map-wait-type-override pool_lock irq_context: 0 rtnl_mutex batched_entropy_u32.lock crngs.lock irq_context: 0 protocol_list_lock irq_context: 0 linkfail_lock irq_context: 0 rtnl_mutex &root->kernfs_rwsem &rq->__lock irq_context: 0 rtnl_mutex &root->kernfs_rwsem &cfs_rq->removed.lock irq_context: 0 rtnl_mutex &root->kernfs_rwsem &obj_hash[i].lock irq_context: 0 rtnl_mutex uevent_sock_mutex &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: softirq &(&tbl->managed_work)->timer rcu_read_lock &pool->lock &p->pi_lock &rq->__lock &cfs_rq->removed.lock irq_context: 0 rose_neigh_list_lock irq_context: 0 proto_tab_lock#2 irq_context: 0 bt_proto_lock irq_context: 0 bt_proto_lock pool_lock#2 irq_context: 0 bt_proto_lock &dir->lock irq_context: 0 bt_proto_lock &obj_hash[i].lock irq_context: 0 bt_proto_lock &c->lock irq_context: 0 bt_proto_lock &pcp->lock &zone->lock irq_context: 0 bt_proto_lock &zone->lock irq_context: 0 bt_proto_lock &____s->seqcount irq_context: 0 bt_proto_lock chan_list_lock irq_context: 0 bt_proto_lock l2cap_sk_list.lock irq_context: 0 sk_lock-AF_BLUETOOTH-BTPROTO_L2CAP irq_context: 0 sk_lock-AF_BLUETOOTH-BTPROTO_L2CAP slock-AF_BLUETOOTH-BTPROTO_L2CAP irq_context: 0 sk_lock-AF_BLUETOOTH-BTPROTO_L2CAP chan_list_lock irq_context: 0 slock-AF_BLUETOOTH-BTPROTO_L2CAP irq_context: 0 rfcomm_wq.lock irq_context: 0 rfcomm_mutex irq_context: 0 auth_domain_lock irq_context: 0 registered_mechs_lock irq_context: 0 &x->wait#21 &p->pi_lock &rq->__lock irq_context: 0 &x->wait#21 &p->pi_lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 key_types_sem &rq->__lock irq_context: 0 atm_dev_notify_chain.lock irq_context: 0 genl_mutex irq_context: 0 proto_tab_lock#3 irq_context: 0 vlan_ioctl_mutex irq_context: 0 pernet_ops_rwsem (console_sem).lock irq_context: 0 pernet_ops_rwsem console_lock console_srcu console_owner_lock irq_context: 0 pernet_ops_rwsem console_lock console_srcu console_owner irq_context: 0 pernet_ops_rwsem console_lock console_srcu console_owner &port_lock_key irq_context: 0 pernet_ops_rwsem console_lock console_srcu console_owner console_owner_lock irq_context: 0 pernet_ops_rwsem rcu_read_lock rcu_read_lock &ndev->lock irq_context: 0 rds_info_lock irq_context: 0 rds_trans_sem irq_context: 0 rds_trans_sem (console_sem).lock irq_context: 0 rds_trans_sem console_lock console_srcu console_owner_lock irq_context: 0 rds_trans_sem console_lock console_srcu console_owner irq_context: 0 rds_trans_sem console_lock console_srcu console_owner &port_lock_key irq_context: 0 rds_trans_sem console_lock console_srcu console_owner console_owner_lock irq_context: 0 &id_priv->lock irq_context: 0 lock#7 irq_context: 0 lock#7 fs_reclaim irq_context: 0 lock#7 fs_reclaim mmu_notifier_invalidate_range_start irq_context: 0 lock#7 pool_lock#2 irq_context: 0 lock#7 &xa->xa_lock#14 irq_context: 0 lock#7 &xa->xa_lock#14 pool_lock#2 irq_context: 0 pernet_ops_rwsem k-sk_lock-AF_INET6 irq_context: 0 pernet_ops_rwsem k-sk_lock-AF_INET6 k-slock-AF_INET6 irq_context: 0 pernet_ops_rwsem k-slock-AF_INET6 irq_context: 0 pernet_ops_rwsem k-clock-AF_INET6 irq_context: 0 pernet_ops_rwsem k-sk_lock-AF_INET6 &tcp_hashinfo.bhash[i].lock irq_context: 0 pernet_ops_rwsem k-sk_lock-AF_INET6 &tcp_hashinfo.bhash[i].lock &pcp->lock &zone->lock irq_context: 0 pernet_ops_rwsem k-sk_lock-AF_INET6 &tcp_hashinfo.bhash[i].lock &zone->lock irq_context: 0 pernet_ops_rwsem k-sk_lock-AF_INET6 &tcp_hashinfo.bhash[i].lock &____s->seqcount irq_context: 0 pernet_ops_rwsem k-sk_lock-AF_INET6 &tcp_hashinfo.bhash[i].lock pool_lock#2 irq_context: 0 pernet_ops_rwsem k-sk_lock-AF_INET6 &tcp_hashinfo.bhash[i].lock &c->lock irq_context: 0 pernet_ops_rwsem k-sk_lock-AF_INET6 &tcp_hashinfo.bhash[i].lock &tcp_hashinfo.bhash2[i].lock irq_context: 0 pernet_ops_rwsem k-sk_lock-AF_INET6 &tcp_hashinfo.bhash[i].lock &tcp_hashinfo.bhash2[i].lock &pcp->lock &zone->lock irq_context: 0 pernet_ops_rwsem k-sk_lock-AF_INET6 &tcp_hashinfo.bhash[i].lock &tcp_hashinfo.bhash2[i].lock &zone->lock irq_context: 0 pernet_ops_rwsem k-sk_lock-AF_INET6 &tcp_hashinfo.bhash[i].lock &tcp_hashinfo.bhash2[i].lock &____s->seqcount irq_context: 0 pernet_ops_rwsem k-sk_lock-AF_INET6 &tcp_hashinfo.bhash[i].lock &tcp_hashinfo.bhash2[i].lock pool_lock#2 irq_context: 0 pernet_ops_rwsem k-sk_lock-AF_INET6 &tcp_hashinfo.bhash[i].lock &tcp_hashinfo.bhash2[i].lock &c->lock irq_context: 0 pernet_ops_rwsem k-sk_lock-AF_INET6 &tcp_hashinfo.bhash[i].lock &tcp_hashinfo.bhash2[i].lock k-clock-AF_INET6 irq_context: 0 pernet_ops_rwsem k-sk_lock-AF_INET6 &tcp_hashinfo.bhash[i].lock k-clock-AF_INET6 irq_context: 0 pernet_ops_rwsem k-sk_lock-AF_INET6 &h->lhash2[i].lock irq_context: 0 rds_trans_sem &rq->__lock irq_context: 0 rds_trans_sem &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 pernet_ops_rwsem cpu_hotplug_lock wq_pool_mutex irq_context: 0 pernet_ops_rwsem cpu_hotplug_lock wq_pool_mutex fs_reclaim irq_context: 0 pernet_ops_rwsem cpu_hotplug_lock wq_pool_mutex fs_reclaim mmu_notifier_invalidate_range_start irq_context: 0 pernet_ops_rwsem cpu_hotplug_lock wq_pool_mutex pool_lock#2 irq_context: 0 pernet_ops_rwsem cpu_hotplug_lock wq_pool_mutex &obj_hash[i].lock irq_context: 0 pernet_ops_rwsem cpu_hotplug_lock wq_pool_mutex &wq->mutex irq_context: 0 pernet_ops_rwsem cpu_hotplug_lock wq_pool_mutex &wq->mutex &pool->lock/1 irq_context: 0 pernet_ops_rwsem wq_pool_mutex irq_context: 0 pernet_ops_rwsem wq_pool_mutex &wq->mutex irq_context: 0 pernet_ops_rwsem pcpu_lock irq_context: 0 pernet_ops_rwsem &list->lock#4 irq_context: 0 pernet_ops_rwsem &dir->lock#2 irq_context: 0 pernet_ops_rwsem ptype_lock irq_context: 0 pernet_ops_rwsem rcu_read_lock rhashtable_bucket irq_context: 0 pernet_ops_rwsem k-clock-AF_TIPC irq_context: 0 pernet_ops_rwsem k-sk_lock-AF_TIPC irq_context: 0 pernet_ops_rwsem k-sk_lock-AF_TIPC k-slock-AF_TIPC irq_context: 0 pernet_ops_rwsem k-slock-AF_TIPC irq_context: 0 pernet_ops_rwsem k-sk_lock-AF_TIPC &tn->nametbl_lock irq_context: 0 pernet_ops_rwsem k-sk_lock-AF_TIPC &tn->nametbl_lock pool_lock#2 irq_context: 0 pernet_ops_rwsem k-sk_lock-AF_TIPC &tn->nametbl_lock &service->lock irq_context: 0 pernet_ops_rwsem k-sk_lock-AF_TIPC &tn->nametbl_lock &service->lock pool_lock#2 irq_context: 0 rcu_read_lock &pcp->lock &zone->lock &____s->seqcount irq_context: 0 pernet_ops_rwsem rtnl_mutex &pnettable->lock irq_context: 0 pernet_ops_rwsem rtnl_mutex smc_ib_devices.mutex irq_context: 0 smc_wr_rx_hash_lock irq_context: 0 v9fs_trans_lock irq_context: 0 pernet_ops_rwsem &this->receive_lock irq_context: 0 &x->wait#17 &p->pi_lock irq_context: 0 &x->wait#17 &p->pi_lock &rq->__lock irq_context: 0 &x->wait#17 &p->pi_lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 lowpan_nhc_lock irq_context: softirq (&vblank->disable_timer) irq_context: softirq (&vblank->disable_timer) &dev->vbl_lock irq_context: softirq (&vblank->disable_timer) &dev->vbl_lock &dev->vblank_time_lock irq_context: softirq (&vblank->disable_timer) &dev->vbl_lock &dev->vblank_time_lock hrtimer_bases.lock irq_context: softirq (&vblank->disable_timer) &dev->vbl_lock &dev->vblank_time_lock hrtimer_bases.lock &obj_hash[i].lock irq_context: 0 ovs_mutex irq_context: 0 pernet_ops_rwsem once_lock irq_context: 0 pernet_ops_rwsem once_lock crngs.lock irq_context: 0 pernet_ops_rwsem rcu_read_lock &pool->lock pool_lock#2 irq_context: 0 pernet_ops_rwsem nf_ct_proto_mutex irq_context: 0 pernet_ops_rwsem nf_ct_proto_mutex defrag4_mutex irq_context: 0 pernet_ops_rwsem nf_ct_proto_mutex defrag4_mutex nf_hook_mutex irq_context: 0 pernet_ops_rwsem nf_ct_proto_mutex defrag4_mutex nf_hook_mutex fs_reclaim irq_context: 0 pernet_ops_rwsem nf_ct_proto_mutex defrag4_mutex nf_hook_mutex fs_reclaim mmu_notifier_invalidate_range_start irq_context: 0 pernet_ops_rwsem nf_ct_proto_mutex defrag4_mutex nf_hook_mutex &pcp->lock &zone->lock irq_context: 0 pernet_ops_rwsem nf_ct_proto_mutex defrag4_mutex nf_hook_mutex &zone->lock irq_context: 0 pernet_ops_rwsem nf_ct_proto_mutex defrag4_mutex nf_hook_mutex &____s->seqcount irq_context: 0 pernet_ops_rwsem nf_ct_proto_mutex defrag4_mutex nf_hook_mutex pool_lock#2 irq_context: 0 pernet_ops_rwsem nf_ct_proto_mutex defrag4_mutex nf_hook_mutex &c->lock irq_context: 0 pernet_ops_rwsem nf_ct_proto_mutex defrag4_mutex cpu_hotplug_lock irq_context: 0 pernet_ops_rwsem nf_ct_proto_mutex defrag4_mutex cpu_hotplug_lock jump_label_mutex irq_context: 0 pernet_ops_rwsem nf_ct_proto_mutex defrag4_mutex cpu_hotplug_lock jump_label_mutex text_mutex irq_context: 0 pernet_ops_rwsem nf_ct_proto_mutex defrag4_mutex cpu_hotplug_lock jump_label_mutex text_mutex ptlock_ptr(page)#2 irq_context: 0 pernet_ops_rwsem nf_ct_proto_mutex nf_hook_mutex irq_context: 0 pernet_ops_rwsem nf_ct_proto_mutex nf_hook_mutex fs_reclaim irq_context: 0 pernet_ops_rwsem nf_ct_proto_mutex nf_hook_mutex fs_reclaim mmu_notifier_invalidate_range_start irq_context: 0 pernet_ops_rwsem nf_ct_proto_mutex nf_hook_mutex pool_lock#2 irq_context: 0 pernet_ops_rwsem nf_ct_proto_mutex cpu_hotplug_lock irq_context: 0 pernet_ops_rwsem nf_ct_proto_mutex &obj_hash[i].lock irq_context: 0 pernet_ops_rwsem nf_ct_proto_mutex pool_lock#2 irq_context: 0 pernet_ops_rwsem nf_ct_proto_mutex &obj_hash[i].lock pool_lock irq_context: 0 pernet_ops_rwsem nf_ct_proto_mutex cpu_hotplug_lock jump_label_mutex irq_context: 0 pernet_ops_rwsem nf_ct_proto_mutex cpu_hotplug_lock jump_label_mutex text_mutex irq_context: 0 pernet_ops_rwsem nf_ct_proto_mutex cpu_hotplug_lock jump_label_mutex text_mutex ptlock_ptr(page)#2 irq_context: 0 pernet_ops_rwsem nf_ct_proto_mutex defrag6_mutex irq_context: 0 pernet_ops_rwsem nf_ct_proto_mutex defrag6_mutex nf_hook_mutex irq_context: 0 pernet_ops_rwsem nf_ct_proto_mutex defrag6_mutex nf_hook_mutex fs_reclaim irq_context: 0 pernet_ops_rwsem nf_ct_proto_mutex defrag6_mutex nf_hook_mutex fs_reclaim mmu_notifier_invalidate_range_start irq_context: 0 pernet_ops_rwsem nf_ct_proto_mutex defrag6_mutex nf_hook_mutex pool_lock#2 irq_context: 0 pernet_ops_rwsem nf_ct_proto_mutex defrag6_mutex cpu_hotplug_lock irq_context: 0 pernet_ops_rwsem nf_ct_proto_mutex defrag6_mutex cpu_hotplug_lock jump_label_mutex irq_context: 0 pernet_ops_rwsem nf_ct_proto_mutex defrag6_mutex cpu_hotplug_lock jump_label_mutex text_mutex irq_context: 0 pernet_ops_rwsem nf_ct_proto_mutex defrag6_mutex cpu_hotplug_lock jump_label_mutex text_mutex ptlock_ptr(page)#2 irq_context: 0 cpu_hotplug_lock cpuhp_state-up &dev->mutex &k->list_lock irq_context: 0 cpu_hotplug_lock cpuhp_state-up &dev->mutex &k->k_lock irq_context: 0 cpu_hotplug_lock cpuhp_state-up &dev->mutex &dev->power.lock irq_context: 0 cpu_hotplug_lock cpuhp_state-up subsys mutex#79 irq_context: 0 cpu_hotplug_lock cpuhp_state-up &base->lock irq_context: 0 cpu_hotplug_lock cpuhp_state-up &base->lock &obj_hash[i].lock irq_context: 0 cpu_hotplug_lock static_call_mutex text_mutex ptlock_ptr(page)#2 irq_context: 0 (wq_completion)events_unbound (work_completion)(&sub_info->work) rcu_read_lock pool_lock#2 irq_context: 0 fill_pool_map-wait-type-override &cfs_rq->removed.lock irq_context: 0 fill_pool_map-wait-type-override &obj_hash[i].lock irq_context: 0 &sig->cred_guard_mutex rcu_read_lock &cfs_rq->removed.lock irq_context: 0 &sig->cred_guard_mutex rcu_read_lock &obj_hash[i].lock irq_context: 0 &root->kernfs_rwsem kernfs_idr_lock &obj_hash[i].lock irq_context: 0 &root->kernfs_rwsem kernfs_idr_lock pool_lock#2 irq_context: 0 kernfs_idr_lock &obj_hash[i].lock irq_context: 0 kernfs_idr_lock pool_lock#2 irq_context: 0 &tsk->futex_exit_mutex &rq->__lock irq_context: 0 &root->kernfs_rwsem &rq->__lock irq_context: 0 &root->kernfs_rwsem &cfs_rq->removed.lock irq_context: 0 &root->kernfs_rwsem &obj_hash[i].lock irq_context: softirq (&q->timeout) irq_context: softirq &(&gc_work->dwork)->timer irq_context: softirq (&q->timeout) rcu_read_lock &pool->lock irq_context: softirq &(&gc_work->dwork)->timer rcu_read_lock &pool->lock irq_context: softirq (&q->timeout) rcu_read_lock &pool->lock &obj_hash[i].lock irq_context: softirq &(&gc_work->dwork)->timer rcu_read_lock &pool->lock &obj_hash[i].lock irq_context: softirq (&q->timeout) rcu_read_lock &pool->lock pool_lock#2 irq_context: softirq &(&gc_work->dwork)->timer rcu_read_lock &pool->lock &p->pi_lock irq_context: softirq (&q->timeout) rcu_read_lock &pool->lock &p->pi_lock irq_context: softirq &(&gc_work->dwork)->timer rcu_read_lock &pool->lock &p->pi_lock &rq->__lock irq_context: softirq (&q->timeout) rcu_read_lock &pool->lock &p->pi_lock &rq->__lock irq_context: softirq &(&gc_work->dwork)->timer rcu_read_lock &pool->lock &p->pi_lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: softirq (&q->timeout) rcu_read_lock &pool->lock &p->pi_lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)kblockd (work_completion)(&q->timeout_work) irq_context: 0 (wq_completion)events_power_efficient (work_completion)(&(&gc_work->dwork)->work) irq_context: 0 (wq_completion)events_power_efficient (work_completion)(&(&gc_work->dwork)->work) rcu_read_lock &____s->seqcount#7 irq_context: 0 kernfs_idr_lock &obj_hash[i].lock pool_lock irq_context: 0 (wq_completion)events_power_efficient (work_completion)(&(&gc_work->dwork)->work) rcu_read_lock rcu_node_0 irq_context: 0 (wq_completion)events_power_efficient (work_completion)(&(&gc_work->dwork)->work) rcu_read_lock &rq->__lock irq_context: 0 &root->kernfs_rwsem &root->kernfs_iattr_rwsem &rq->__lock irq_context: 0 (wq_completion)events_power_efficient (work_completion)(&(&gc_work->dwork)->work) rcu_read_lock &cfs_rq->removed.lock irq_context: 0 (wq_completion)events_power_efficient (work_completion)(&(&gc_work->dwork)->work) rcu_read_lock &obj_hash[i].lock irq_context: 0 (wq_completion)events_power_efficient (work_completion)(&(&gc_work->dwork)->work) rcu_read_lock &rcu_state.gp_wq irq_context: 0 (wq_completion)events_power_efficient (work_completion)(&(&gc_work->dwork)->work) rcu_read_lock &rcu_state.gp_wq &p->pi_lock irq_context: 0 (wq_completion)events_power_efficient (work_completion)(&(&gc_work->dwork)->work) rcu_read_lock &rcu_state.gp_wq &p->pi_lock &rq->__lock irq_context: 0 (wq_completion)events_power_efficient (work_completion)(&(&gc_work->dwork)->work) rcu_read_lock &rcu_state.gp_wq &p->pi_lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)events_power_efficient (work_completion)(&(&gc_work->dwork)->work) rcu_read_lock &rcu_state.gp_wq &p->pi_lock &rq->__lock &cfs_rq->removed.lock irq_context: 0 (wq_completion)events_power_efficient (work_completion)(&(&gc_work->dwork)->work) rcu_read_lock &pool->lock irq_context: 0 (wq_completion)events_power_efficient (work_completion)(&(&gc_work->dwork)->work) rcu_read_lock &pool->lock &obj_hash[i].lock irq_context: 0 (wq_completion)events_power_efficient (work_completion)(&(&gc_work->dwork)->work) &rq->__lock irq_context: 0 (wq_completion)events_power_efficient (work_completion)(&(&gc_work->dwork)->work) &cfs_rq->removed.lock irq_context: 0 (wq_completion)events_power_efficient (work_completion)(&(&gc_work->dwork)->work) &obj_hash[i].lock irq_context: 0 (wq_completion)events_power_efficient (work_completion)(&(&gc_work->dwork)->work) rcu_read_lock &pool->lock &p->pi_lock irq_context: 0 (wq_completion)events_power_efficient (work_completion)(&(&gc_work->dwork)->work) rcu_read_lock &pool->lock &p->pi_lock &rq->__lock irq_context: 0 (wq_completion)events_power_efficient (work_completion)(&(&gc_work->dwork)->work) rcu_read_lock &pool->lock &p->pi_lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)events_power_efficient (work_completion)(&(&gc_work->dwork)->work) rcu_read_lock &pool->lock &p->pi_lock &rq->__lock &cfs_rq->removed.lock irq_context: 0 (wq_completion)events (work_completion)(&p->wq) rcu_read_lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: softirq &(&ipvs->defense_work)->timer irq_context: softirq &(&ipvs->defense_work)->timer rcu_read_lock &pool->lock irq_context: softirq &(&ipvs->defense_work)->timer rcu_read_lock &pool->lock &obj_hash[i].lock irq_context: softirq &(&ipvs->defense_work)->timer rcu_read_lock &pool->lock &p->pi_lock irq_context: softirq &(&ipvs->defense_work)->timer rcu_read_lock &pool->lock &p->pi_lock &rq->__lock irq_context: softirq &(&ipvs->defense_work)->timer rcu_read_lock &pool->lock &p->pi_lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)events_long (work_completion)(&(&ipvs->defense_work)->work) irq_context: 0 (wq_completion)events_long (work_completion)(&(&ipvs->defense_work)->work) &s->s_inode_list_lock irq_context: 0 (wq_completion)events_long (work_completion)(&(&ipvs->defense_work)->work) &ipvs->dropentry_lock irq_context: 0 (wq_completion)events_long (work_completion)(&(&ipvs->defense_work)->work) &ipvs->droppacket_lock irq_context: 0 (wq_completion)events_long (work_completion)(&(&ipvs->defense_work)->work) &ipvs->securetcp_lock irq_context: 0 (wq_completion)events_long (work_completion)(&(&ipvs->defense_work)->work) &obj_hash[i].lock irq_context: 0 (wq_completion)events_long (work_completion)(&(&ipvs->defense_work)->work) &base->lock irq_context: 0 (wq_completion)events_long (work_completion)(&(&ipvs->defense_work)->work) &base->lock &obj_hash[i].lock irq_context: 0 &sig->cred_guard_mutex &sb->s_type->i_mutex_key &dentry->d_lock &dentry->d_lock/1 irq_context: 0 &sig->cred_guard_mutex &sb->s_type->i_mutex_key &dentry->d_lock/1 irq_context: 0 (wq_completion)events_power_efficient (work_completion)(&(&gc_work->dwork)->work) rcu_read_lock &rcu_state.gp_wq &p->pi_lock &cfs_rq->removed.lock irq_context: 0 fs_reclaim &cfs_rq->removed.lock irq_context: 0 fs_reclaim &obj_hash[i].lock irq_context: 0 uevent_sock_mutex &rq->__lock irq_context: 0 uevent_sock_mutex &cfs_rq->removed.lock irq_context: 0 uevent_sock_mutex &obj_hash[i].lock irq_context: 0 uevent_sock_mutex &obj_hash[i].lock pool_lock irq_context: 0 (wq_completion)events_power_efficient (work_completion)(&(&gc_work->dwork)->work) rcu_read_lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 &sig->cred_guard_mutex fs_reclaim mmu_notifier_invalidate_range_start &rq->__lock irq_context: 0 (wq_completion)events_power_efficient (work_completion)(&(&gc_work->dwork)->work) rcu_read_lock pool_lock#2 irq_context: softirq rcu_callback &base->lock irq_context: softirq rcu_callback &base->lock &obj_hash[i].lock irq_context: 0 &root->kernfs_rwsem &obj_hash[i].lock pool_lock irq_context: 0 (wq_completion)events_power_efficient (work_completion)(&(&gc_work->dwork)->work) rcu_node_0 irq_context: 0 (wq_completion)events_power_efficient (work_completion)(&(&gc_work->dwork)->work) pool_lock#2 irq_context: 0 lock kernfs_idr_lock &pcp->lock &zone->lock &____s->seqcount irq_context: 0 (wq_completion)events (work_completion)(&p->wq) rcu_read_lock &rcu_state.gp_wq irq_context: 0 (wq_completion)events (work_completion)(&p->wq) rcu_read_lock &rcu_state.gp_wq &p->pi_lock irq_context: 0 (wq_completion)events (work_completion)(&p->wq) rcu_read_lock &rcu_state.gp_wq &p->pi_lock &rq->__lock irq_context: 0 (wq_completion)events (work_completion)(&p->wq) rcu_read_lock &rcu_state.gp_wq &p->pi_lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 pcpu_alloc_mutex pcpu_alloc_mutex.wait_lock irq_context: 0 pcpu_alloc_mutex &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 pcpu_alloc_mutex.wait_lock irq_context: 0 pcpu_alloc_mutex &cfs_rq->removed.lock irq_context: 0 pcpu_alloc_mutex &obj_hash[i].lock irq_context: 0 pcpu_alloc_mutex pool_lock#2 irq_context: 0 (wq_completion)events (work_completion)(&p->wq) rcu_read_lock &pool->lock irq_context: 0 (wq_completion)events (work_completion)(&p->wq) rcu_read_lock &pool->lock &obj_hash[i].lock irq_context: 0 (wq_completion)events drain_vmap_work irq_context: 0 (wq_completion)events drain_vmap_work vmap_purge_lock irq_context: 0 (wq_completion)events drain_vmap_work vmap_purge_lock purge_vmap_area_lock irq_context: 0 (wq_completion)events drain_vmap_work vmap_purge_lock free_vmap_area_lock irq_context: 0 (wq_completion)events drain_vmap_work vmap_purge_lock free_vmap_area_lock &obj_hash[i].lock irq_context: 0 (wq_completion)events drain_vmap_work vmap_purge_lock free_vmap_area_lock pool_lock#2 irq_context: 0 (wq_completion)events drain_vmap_work vmap_purge_lock free_vmap_area_lock init_mm.page_table_lock irq_context: 0 (wq_completion)events drain_vmap_work vmap_purge_lock free_vmap_area_lock init_mm.page_table_lock rcu_read_lock pool_lock#2 irq_context: 0 (wq_completion)events drain_vmap_work vmap_purge_lock free_vmap_area_lock init_mm.page_table_lock &obj_hash[i].lock irq_context: 0 (wq_completion)events drain_vmap_work vmap_purge_lock free_vmap_area_lock quarantine_lock irq_context: 0 (wq_completion)events drain_vmap_work vmap_purge_lock &rq->__lock irq_context: 0 (wq_completion)events drain_vmap_work vmap_purge_lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)events drain_vmap_work vmap_purge_lock &cfs_rq->removed.lock irq_context: 0 (wq_completion)events drain_vmap_work vmap_purge_lock &obj_hash[i].lock irq_context: 0 (wq_completion)events drain_vmap_work vmap_purge_lock pool_lock#2 irq_context: 0 &sig->cred_guard_mutex &fs->lock &dentry->d_lock irq_context: 0 &sig->cred_guard_mutex delayed_uprobe_lock &rq->__lock irq_context: 0 &sig->cred_guard_mutex rcu_read_lock &rcu_state.gp_wq irq_context: 0 &sig->cred_guard_mutex rcu_read_lock &rcu_state.gp_wq &p->pi_lock irq_context: 0 &sig->cred_guard_mutex rcu_read_lock &rcu_state.gp_wq &p->pi_lock &rq->__lock irq_context: 0 &sig->cred_guard_mutex rcu_read_lock &rcu_state.gp_wq &p->pi_lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 tasklist_lock &sighand->siglock &(&sig->stats_lock)->lock &____s->seqcount#5 &obj_hash[i].lock pool_lock irq_context: softirq (&net->can.stattimer) irq_context: softirq (&net->can.stattimer) &obj_hash[i].lock irq_context: softirq (&net->can.stattimer) &base->lock irq_context: softirq (&net->can.stattimer) &base->lock &obj_hash[i].lock irq_context: 0 &root->kernfs_rwsem &root->kernfs_iattr_rwsem &cfs_rq->removed.lock irq_context: 0 &root->kernfs_rwsem &root->kernfs_iattr_rwsem &obj_hash[i].lock irq_context: 0 &root->kernfs_rwsem &root->kernfs_iattr_rwsem pool_lock#2 irq_context: softirq &(&ipvs->defense_work)->timer rcu_read_lock &pool->lock &p->pi_lock &rq->__lock &cfs_rq->removed.lock irq_context: 0 &mm->mmap_lock batched_entropy_u8.lock irq_context: 0 &mm->mmap_lock kfence_freelist_lock irq_context: 0 &root->kernfs_rwsem rcu_read_lock rcu_node_0 irq_context: 0 &root->kernfs_rwsem rcu_read_lock &rq->__lock irq_context: 0 &root->kernfs_rwsem rcu_read_lock &cfs_rq->removed.lock irq_context: 0 &root->kernfs_rwsem rcu_read_lock &obj_hash[i].lock irq_context: 0 &root->kernfs_rwsem rcu_read_lock &rcu_state.gp_wq irq_context: 0 &root->kernfs_rwsem rcu_read_lock &rcu_state.gp_wq &p->pi_lock irq_context: 0 &root->kernfs_rwsem rcu_read_lock &rcu_state.gp_wq &p->pi_lock &cfs_rq->removed.lock irq_context: 0 &root->kernfs_rwsem rcu_read_lock &rcu_state.gp_wq &p->pi_lock &rq->__lock irq_context: 0 &root->kernfs_rwsem rcu_read_lock &rcu_state.gp_wq &p->pi_lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 &root->kernfs_rwsem rcu_read_lock &rcu_state.gp_wq &p->pi_lock &rq->__lock &cfs_rq->removed.lock irq_context: 0 &root->kernfs_rwsem quarantine_lock irq_context: 0 (wq_completion)events_unbound (work_completion)(&sub_info->work) batched_entropy_u64.lock crngs.lock base_crng.lock irq_context: 0 lock map_idr_lock irq_context: 0 lock map_idr_lock pool_lock#2 irq_context: 0 &sb->s_type->i_lock_key#15 &dentry->d_lock irq_context: 0 purge_vmap_area_lock irq_context: 0 (wq_completion)events_power_efficient (work_completion)(&(&gc_work->dwork)->work) &base->lock irq_context: 0 (wq_completion)events_power_efficient (work_completion)(&(&gc_work->dwork)->work) &base->lock &obj_hash[i].lock irq_context: 0 lock prog_idr_lock irq_context: 0 lock prog_idr_lock pool_lock#2 irq_context: 0 bpf_lock irq_context: 0 rcu_read_lock_trace fs_reclaim irq_context: 0 rcu_read_lock_trace fs_reclaim mmu_notifier_invalidate_range_start irq_context: 0 rcu_read_lock_trace pool_lock#2 irq_context: 0 rcu_read_lock_trace &obj_hash[i].lock irq_context: 0 rcu_read_lock_trace lock irq_context: 0 rcu_read_lock_trace lock btf_idr_lock irq_context: 0 rcu_read_lock_trace lock btf_idr_lock pool_lock#2 irq_context: 0 rcu_read_lock_trace &newf->file_lock irq_context: 0 rcu_read_lock_trace &sb->s_type->i_lock_key#15 irq_context: 0 rcu_read_lock_trace &sb->s_type->i_lock_key#15 &dentry->d_lock irq_context: 0 rcu_read_lock_trace lock map_idr_lock irq_context: 0 rcu_read_lock_trace &map->freeze_mutex irq_context: 0 key_types_sem irq_context: 0 key_types_sem asymmetric_key_parsers_sem irq_context: 0 key_types_sem asymmetric_key_parsers_sem fs_reclaim irq_context: 0 key_types_sem asymmetric_key_parsers_sem fs_reclaim mmu_notifier_invalidate_range_start irq_context: 0 key_types_sem asymmetric_key_parsers_sem pool_lock#2 irq_context: 0 key_types_sem asymmetric_key_parsers_sem &c->lock irq_context: 0 key_types_sem asymmetric_key_parsers_sem &____s->seqcount irq_context: 0 key_types_sem asymmetric_key_parsers_sem &pcp->lock &zone->lock irq_context: 0 key_types_sem asymmetric_key_parsers_sem &zone->lock irq_context: 0 key_types_sem asymmetric_key_parsers_sem &pcp->lock &zone->lock &____s->seqcount irq_context: 0 key_types_sem asymmetric_key_parsers_sem rcu_read_lock pool_lock#2 irq_context: 0 key_types_sem asymmetric_key_parsers_sem &obj_hash[i].lock irq_context: 0 key_types_sem asymmetric_key_parsers_sem crypto_alg_sem irq_context: 0 key_types_sem asymmetric_key_parsers_sem crypto_alg_sem irq_context: 0 key_types_sem asymmetric_key_parsers_sem (crypto_chain).rwsem irq_context: 0 key_types_sem asymmetric_key_parsers_sem (crypto_chain).rwsem fs_reclaim irq_context: 0 key_types_sem asymmetric_key_parsers_sem (crypto_chain).rwsem fs_reclaim mmu_notifier_invalidate_range_start irq_context: 0 key_types_sem asymmetric_key_parsers_sem (crypto_chain).rwsem &c->lock irq_context: 0 key_types_sem asymmetric_key_parsers_sem (crypto_chain).rwsem &pcp->lock &zone->lock irq_context: 0 key_types_sem asymmetric_key_parsers_sem (crypto_chain).rwsem &pcp->lock &zone->lock &____s->seqcount irq_context: 0 key_types_sem asymmetric_key_parsers_sem (crypto_chain).rwsem &____s->seqcount irq_context: 0 key_types_sem asymmetric_key_parsers_sem (crypto_chain).rwsem pool_lock#2 irq_context: 0 key_types_sem asymmetric_key_parsers_sem (crypto_chain).rwsem kthread_create_lock irq_context: 0 key_types_sem asymmetric_key_parsers_sem (crypto_chain).rwsem &p->pi_lock irq_context: 0 key_types_sem asymmetric_key_parsers_sem (crypto_chain).rwsem &x->wait irq_context: 0 key_types_sem asymmetric_key_parsers_sem (crypto_chain).rwsem &rq->__lock irq_context: 0 key_types_sem asymmetric_key_parsers_sem (crypto_chain).rwsem &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 lock pidmap_lock &c->lock irq_context: 0 lock pidmap_lock &pcp->lock &zone->lock irq_context: 0 lock pidmap_lock &zone->lock irq_context: 0 lock pidmap_lock &____s->seqcount irq_context: 0 key_types_sem asymmetric_key_parsers_sem (crypto_chain).rwsem &obj_hash[i].lock irq_context: 0 key_types_sem asymmetric_key_parsers_sem (crypto_chain).rwsem &p->pi_lock &rq->__lock irq_context: 0 key_types_sem asymmetric_key_parsers_sem (crypto_chain).rwsem &p->pi_lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 key_types_sem asymmetric_key_parsers_sem &x->wait#21 irq_context: 0 key_types_sem asymmetric_key_parsers_sem &base->lock irq_context: 0 key_types_sem asymmetric_key_parsers_sem &base->lock &obj_hash[i].lock irq_context: 0 key_types_sem asymmetric_key_parsers_sem &rq->__lock irq_context: 0 key_types_sem asymmetric_key_parsers_sem &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 key_types_sem asymmetric_key_parsers_sem &cfs_rq->removed.lock irq_context: 0 key_types_sem asymmetric_key_parsers_sem (&timer.timer) irq_context: 0 key_types_sem &type->lock_class irq_context: 0 key_types_sem &type->lock_class fs_reclaim irq_context: 0 key_types_sem &type->lock_class fs_reclaim mmu_notifier_invalidate_range_start irq_context: 0 key_types_sem &type->lock_class pool_lock#2 irq_context: 0 key_types_sem &type->lock_class key_user_lock irq_context: 0 key_types_sem &type->lock_class crngs.lock irq_context: 0 key_types_sem &type->lock_class key_serial_lock irq_context: 0 key_types_sem &type->lock_class key_construction_mutex irq_context: 0 key_types_sem &type->lock_class key_construction_mutex &obj_hash[i].lock irq_context: 0 key_types_sem &type->lock_class key_construction_mutex pool_lock#2 irq_context: 0 key_types_sem &type->lock_class ima_keys_lock irq_context: 0 key_types_sem &obj_hash[i].lock irq_context: 0 key_types_sem pool_lock#2 irq_context: 0 cpu_hotplug_lock cpuhp_state_mutex &pcp->lock &zone->lock irq_context: 0 cpu_hotplug_lock cpuhp_state_mutex &zone->lock irq_context: 0 cpu_hotplug_lock cpuhp_state_mutex rcu_read_lock pool_lock#2 irq_context: 0 cpu_hotplug_lock cpuhp_state_mutex &n->list_lock irq_context: 0 cpu_hotplug_lock cpuhp_state_mutex crypto_alg_sem irq_context: 0 cpu_hotplug_lock cpuhp_state_mutex scomp_lock irq_context: 0 cpu_hotplug_lock cpuhp_state_mutex scomp_lock fs_reclaim irq_context: 0 cpu_hotplug_lock cpuhp_state_mutex scomp_lock fs_reclaim mmu_notifier_invalidate_range_start irq_context: 0 cpu_hotplug_lock cpuhp_state_mutex scomp_lock pool_lock#2 irq_context: 0 cpu_hotplug_lock cpuhp_state_mutex scomp_lock free_vmap_area_lock irq_context: 0 cpu_hotplug_lock cpuhp_state_mutex scomp_lock vmap_area_lock irq_context: 0 cpu_hotplug_lock cpuhp_state_mutex scomp_lock &pcp->lock &zone->lock irq_context: 0 cpu_hotplug_lock cpuhp_state_mutex scomp_lock &zone->lock irq_context: 0 cpu_hotplug_lock cpuhp_state_mutex scomp_lock &____s->seqcount irq_context: 0 cpu_hotplug_lock cpuhp_state_mutex scomp_lock init_mm.page_table_lock irq_context: 0 cpu_hotplug_lock cpuhp_state_mutex scomp_lock &c->lock irq_context: 0 slab_mutex lock irq_context: 0 slab_mutex lock kernfs_idr_lock irq_context: 0 slab_mutex &root->kernfs_rwsem irq_context: 0 slab_mutex &root->kernfs_rwsem &root->kernfs_iattr_rwsem irq_context: 0 slab_mutex &k->list_lock irq_context: 0 slab_mutex lock kernfs_idr_lock pool_lock#2 irq_context: 0 slab_mutex lock kernfs_idr_lock &c->lock irq_context: 0 slab_mutex lock kernfs_idr_lock &____s->seqcount irq_context: 0 slab_mutex lock kernfs_idr_lock &pcp->lock &zone->lock irq_context: 0 slab_mutex lock kernfs_idr_lock &zone->lock irq_context: 0 slab_mutex &root->kernfs_rwsem kernfs_idr_lock irq_context: 0 slab_mutex &root->kernfs_rwsem &obj_hash[i].lock irq_context: 0 slab_mutex &root->kernfs_rwsem pool_lock#2 irq_context: 0 &sb->s_type->i_mutex_key#3 &base->lock irq_context: 0 &sb->s_type->i_mutex_key#3 &base->lock &obj_hash[i].lock irq_context: 0 pcpu_drain_mutex &pcp->lock irq_context: 0 pcpu_drain_mutex &pcp->lock &zone->lock irq_context: 0 pcpu_drain_mutex &pcp->lock &zone->lock &____s->seqcount irq_context: 0 rcu_state.exp_mutex &rnp->exp_wq[0] irq_context: 0 (wq_completion)events (debug_obj_work).work pool_lock#2 irq_context: 0 pernet_ops_rwsem k-sk_lock-AF_RXRPC irq_context: 0 pernet_ops_rwsem k-sk_lock-AF_RXRPC k-slock-AF_RXRPC irq_context: 0 pernet_ops_rwsem k-slock-AF_RXRPC irq_context: 0 pernet_ops_rwsem k-sk_lock-AF_RXRPC &rxnet->local_mutex irq_context: 0 pernet_ops_rwsem k-sk_lock-AF_RXRPC &rxnet->local_mutex fs_reclaim irq_context: 0 pernet_ops_rwsem k-sk_lock-AF_RXRPC &rxnet->local_mutex fs_reclaim mmu_notifier_invalidate_range_start irq_context: 0 pernet_ops_rwsem k-sk_lock-AF_RXRPC &rxnet->local_mutex pool_lock#2 irq_context: 0 pernet_ops_rwsem k-sk_lock-AF_RXRPC &rxnet->local_mutex &obj_hash[i].lock irq_context: 0 pernet_ops_rwsem k-sk_lock-AF_RXRPC &rxnet->local_mutex crngs.lock irq_context: 0 pernet_ops_rwsem k-sk_lock-AF_RXRPC &rxnet->local_mutex mmu_notifier_invalidate_range_start irq_context: 0 pernet_ops_rwsem k-sk_lock-AF_RXRPC &rxnet->local_mutex &sb->s_type->i_lock_key#8 irq_context: 0 pernet_ops_rwsem k-sk_lock-AF_RXRPC &rxnet->local_mutex &pcp->lock &zone->lock irq_context: 0 pernet_ops_rwsem k-sk_lock-AF_RXRPC &rxnet->local_mutex &zone->lock irq_context: 0 pernet_ops_rwsem k-sk_lock-AF_RXRPC &rxnet->local_mutex &____s->seqcount irq_context: 0 pernet_ops_rwsem k-sk_lock-AF_RXRPC &rxnet->local_mutex &c->lock irq_context: 0 pernet_ops_rwsem k-sk_lock-AF_RXRPC &rxnet->local_mutex rcu_read_lock pool_lock#2 irq_context: 0 pernet_ops_rwsem k-sk_lock-AF_RXRPC &rxnet->local_mutex &dir->lock irq_context: 0 pernet_ops_rwsem k-sk_lock-AF_RXRPC &rxnet->local_mutex k-sk_lock-AF_INET6 irq_context: 0 pernet_ops_rwsem k-sk_lock-AF_RXRPC &rxnet->local_mutex k-sk_lock-AF_INET6 k-slock-AF_INET6 irq_context: 0 pernet_ops_rwsem k-sk_lock-AF_RXRPC &rxnet->local_mutex k-sk_lock-AF_INET6 &table->hash[i].lock irq_context: 0 pernet_ops_rwsem k-sk_lock-AF_RXRPC &rxnet->local_mutex k-sk_lock-AF_INET6 &table->hash[i].lock k-clock-AF_INET6 irq_context: 0 pernet_ops_rwsem k-sk_lock-AF_RXRPC &rxnet->local_mutex k-sk_lock-AF_INET6 &table->hash[i].lock &table->hash2[i].lock irq_context: 0 pernet_ops_rwsem k-sk_lock-AF_RXRPC &rxnet->local_mutex k-slock-AF_INET6 irq_context: 0 pernet_ops_rwsem k-sk_lock-AF_RXRPC &rxnet->local_mutex cpu_hotplug_lock irq_context: 0 pernet_ops_rwsem k-sk_lock-AF_RXRPC &rxnet->local_mutex cpu_hotplug_lock jump_label_mutex irq_context: 0 pernet_ops_rwsem k-sk_lock-AF_RXRPC &rxnet->local_mutex cpu_hotplug_lock jump_label_mutex text_mutex irq_context: 0 pernet_ops_rwsem k-sk_lock-AF_RXRPC &rxnet->local_mutex cpu_hotplug_lock jump_label_mutex text_mutex ptlock_ptr(page)#2 irq_context: 0 pernet_ops_rwsem k-sk_lock-AF_RXRPC &rxnet->local_mutex k-sk_lock-AF_INET6 rcu_read_lock &pool->lock irq_context: 0 pernet_ops_rwsem k-sk_lock-AF_RXRPC &rxnet->local_mutex k-sk_lock-AF_INET6 rcu_read_lock &pool->lock &obj_hash[i].lock irq_context: 0 pernet_ops_rwsem k-sk_lock-AF_RXRPC &rxnet->local_mutex k-sk_lock-AF_INET6 rcu_read_lock &pool->lock &p->pi_lock irq_context: 0 pernet_ops_rwsem k-sk_lock-AF_RXRPC &rxnet->local_mutex k-sk_lock-AF_INET6 rcu_read_lock &pool->lock &p->pi_lock &rq->__lock irq_context: 0 pernet_ops_rwsem k-sk_lock-AF_RXRPC &rxnet->local_mutex k-sk_lock-AF_INET6 rcu_read_lock &pool->lock &p->pi_lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 pernet_ops_rwsem k-sk_lock-AF_RXRPC &rxnet->local_mutex &rq->__lock irq_context: 0 (wq_completion)events netstamp_work irq_context: 0 (wq_completion)events netstamp_work cpu_hotplug_lock irq_context: 0 (wq_completion)events netstamp_work cpu_hotplug_lock jump_label_mutex irq_context: 0 (wq_completion)events netstamp_work cpu_hotplug_lock jump_label_mutex text_mutex irq_context: 0 (wq_completion)events netstamp_work cpu_hotplug_lock jump_label_mutex text_mutex ptlock_ptr(page)#2 irq_context: 0 pernet_ops_rwsem k-sk_lock-AF_RXRPC &rxnet->local_mutex kthread_create_lock irq_context: 0 pernet_ops_rwsem k-sk_lock-AF_RXRPC &rxnet->local_mutex &p->pi_lock irq_context: 0 pernet_ops_rwsem k-sk_lock-AF_RXRPC &rxnet->local_mutex &p->pi_lock &rq->__lock irq_context: 0 pernet_ops_rwsem k-sk_lock-AF_RXRPC &rxnet->local_mutex &p->pi_lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 pernet_ops_rwsem k-sk_lock-AF_RXRPC &rxnet->local_mutex &x->wait irq_context: 0 pernet_ops_rwsem k-sk_lock-AF_RXRPC &rxnet->local_mutex &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 &x->wait#22 irq_context: 0 pernet_ops_rwsem k-sk_lock-AF_RXRPC &rxnet->local_mutex &x->wait#22 irq_context: 0 pernet_ops_rwsem k-sk_lock-AF_RXRPC &local->services_lock irq_context: 0 pernet_ops_rwsem k-sk_lock-AF_RXRPC fs_reclaim irq_context: 0 pernet_ops_rwsem k-sk_lock-AF_RXRPC fs_reclaim mmu_notifier_invalidate_range_start irq_context: 0 pernet_ops_rwsem k-sk_lock-AF_RXRPC &c->lock irq_context: 0 pernet_ops_rwsem k-sk_lock-AF_RXRPC &pcp->lock &zone->lock irq_context: 0 pernet_ops_rwsem k-sk_lock-AF_RXRPC &zone->lock irq_context: 0 pernet_ops_rwsem k-sk_lock-AF_RXRPC &____s->seqcount irq_context: 0 pernet_ops_rwsem k-sk_lock-AF_RXRPC pool_lock#2 irq_context: 0 pernet_ops_rwsem &rxnet->conn_lock irq_context: 0 pernet_ops_rwsem &call->waitq irq_context: 0 pernet_ops_rwsem &rx->call_lock irq_context: 0 pernet_ops_rwsem &rxnet->call_lock irq_context: 0 bio_slab_lock slab_mutex &root->kernfs_rwsem irq_context: 0 bio_slab_lock slab_mutex &k->list_lock irq_context: 0 bio_slab_lock slab_mutex lock irq_context: 0 bio_slab_lock slab_mutex lock kernfs_idr_lock irq_context: 0 bio_slab_lock slab_mutex &root->kernfs_rwsem &root->kernfs_iattr_rwsem irq_context: 0 bio_slab_lock slab_mutex &____s->seqcount irq_context: 0 bio_slab_lock slab_mutex lock kernfs_idr_lock &c->lock irq_context: 0 bio_slab_lock slab_mutex lock kernfs_idr_lock &____s->seqcount irq_context: 0 bio_slab_lock slab_mutex lock kernfs_idr_lock pool_lock#2 irq_context: 0 bio_slab_lock slab_mutex &pcp->lock &zone->lock irq_context: 0 bio_slab_lock slab_mutex &zone->lock irq_context: softirq (&rxnet->peer_keepalive_timer) irq_context: softirq (&rxnet->peer_keepalive_timer) rcu_read_lock &pool->lock/1 irq_context: softirq (&rxnet->peer_keepalive_timer) rcu_read_lock &pool->lock/1 &obj_hash[i].lock irq_context: softirq (&rxnet->peer_keepalive_timer) rcu_read_lock &pool->lock/1 &p->pi_lock irq_context: 0 init_user_ns.keyring_sem irq_context: 0 init_user_ns.keyring_sem key_user_lock irq_context: 0 init_user_ns.keyring_sem root_key_user.lock irq_context: 0 init_user_ns.keyring_sem fs_reclaim irq_context: 0 init_user_ns.keyring_sem fs_reclaim mmu_notifier_invalidate_range_start irq_context: 0 init_user_ns.keyring_sem pool_lock#2 irq_context: 0 (wq_completion)krxrpcd irq_context: 0 init_user_ns.keyring_sem crngs.lock irq_context: 0 init_user_ns.keyring_sem key_serial_lock irq_context: 0 init_user_ns.keyring_sem key_construction_mutex irq_context: 0 (wq_completion)krxrpcd (work_completion)(&rxnet->peer_keepalive_work) irq_context: 0 init_user_ns.keyring_sem &type->lock_class irq_context: 0 init_user_ns.keyring_sem &type->lock_class keyring_serialise_link_lock irq_context: 0 init_user_ns.keyring_sem &type->lock_class keyring_serialise_link_lock fs_reclaim irq_context: 0 init_user_ns.keyring_sem &type->lock_class keyring_serialise_link_lock fs_reclaim mmu_notifier_invalidate_range_start irq_context: 0 (wq_completion)krxrpcd (work_completion)(&rxnet->peer_keepalive_work) &rxnet->peer_hash_lock irq_context: 0 init_user_ns.keyring_sem &type->lock_class keyring_serialise_link_lock pool_lock#2 irq_context: 0 (wq_completion)krxrpcd (work_completion)(&rxnet->peer_keepalive_work) &obj_hash[i].lock irq_context: 0 init_user_ns.keyring_sem &type->lock_class keyring_serialise_link_lock root_key_user.lock irq_context: 0 (wq_completion)krxrpcd (work_completion)(&rxnet->peer_keepalive_work) &base->lock irq_context: 0 init_user_ns.keyring_sem &type->lock_class keyring_serialise_link_lock key_construction_mutex irq_context: 0 (wq_completion)krxrpcd (work_completion)(&rxnet->peer_keepalive_work) &base->lock &obj_hash[i].lock irq_context: 0 init_user_ns.keyring_sem &type->lock_class keyring_serialise_link_lock key_construction_mutex keyring_name_lock irq_context: 0 init_user_ns.keyring_sem &type->lock_class keyring_serialise_link_lock key_construction_mutex &obj_hash[i].lock irq_context: 0 init_user_ns.keyring_sem &type->lock_class keyring_serialise_link_lock key_construction_mutex pool_lock#2 irq_context: 0 init_user_ns.keyring_sem keyring_serialise_link_lock irq_context: 0 init_user_ns.keyring_sem key_construction_mutex keyring_name_lock irq_context: 0 init_user_ns.keyring_sem &type->lock_class keyring_serialise_link_lock &obj_hash[i].lock irq_context: 0 (crypto_chain).rwsem &cfs_rq->removed.lock irq_context: 0 template_list irq_context: 0 idr_lock irq_context: 0 key_types_sem &type->lock_class &c->lock irq_context: 0 key_types_sem &type->lock_class &pcp->lock &zone->lock irq_context: 0 key_types_sem &type->lock_class &zone->lock irq_context: 0 key_types_sem &type->lock_class &____s->seqcount irq_context: 0 ima_extend_list_mutex irq_context: 0 ima_extend_list_mutex fs_reclaim irq_context: 0 ima_extend_list_mutex fs_reclaim mmu_notifier_invalidate_range_start irq_context: 0 ima_extend_list_mutex &c->lock irq_context: 0 ima_extend_list_mutex &pcp->lock &zone->lock irq_context: 0 ima_extend_list_mutex &zone->lock irq_context: 0 ima_extend_list_mutex &____s->seqcount irq_context: 0 ima_extend_list_mutex pool_lock#2 irq_context: 0 pci_bus_sem irq_context: 0 clk_debug_lock irq_context: 0 deferred_probe_work irq_context: 0 (wq_completion)events_unbound deferred_probe_work irq_context: 0 (wq_completion)events_unbound deferred_probe_work deferred_probe_mutex irq_context: 0 (wq_completion)events_unbound (work_completion)(&barr->work) irq_context: 0 &x->wait#10 irq_context: 0 (wq_completion)events_unbound (work_completion)(&barr->work) &x->wait#10 irq_context: 0 (wq_completion)events_unbound (work_completion)(&barr->work) &x->wait#10 &p->pi_lock irq_context: 0 console_mutex &root->kernfs_rwsem irq_context: 0 console_mutex kernfs_notify_lock irq_context: 0 console_mutex kernfs_notify_lock rcu_read_lock &pool->lock irq_context: 0 console_mutex kernfs_notify_lock rcu_read_lock &pool->lock &obj_hash[i].lock irq_context: 0 console_mutex kernfs_notify_lock rcu_read_lock &pool->lock &p->pi_lock irq_context: 0 console_mutex kernfs_notify_lock rcu_read_lock &pool->lock &p->pi_lock &rq->__lock irq_context: 0 console_mutex kernfs_notify_lock rcu_read_lock &pool->lock &p->pi_lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 console_mutex &rq->__lock irq_context: 0 pernet_ops_rwsem rtnl_mutex &(&net->nexthop.notifier_chain)->rwsem irq_context: 0 k-sk_lock-AF_INET irq_context: 0 k-sk_lock-AF_INET k-slock-AF_INET irq_context: 0 k-sk_lock-AF_INET &table->hash[i].lock irq_context: 0 k-sk_lock-AF_INET &table->hash[i].lock k-clock-AF_INET irq_context: 0 k-sk_lock-AF_INET &table->hash[i].lock &table->hash2[i].lock irq_context: 0 k-slock-AF_INET irq_context: 0 k-sk_lock-AF_INET6 irq_context: 0 k-sk_lock-AF_INET6 k-slock-AF_INET6 irq_context: 0 k-slock-AF_INET6 irq_context: 0 k-sk_lock-AF_INET6 &table->hash[i].lock irq_context: 0 k-sk_lock-AF_INET6 &table->hash[i].lock k-clock-AF_INET6 irq_context: 0 k-sk_lock-AF_INET6 &table->hash[i].lock &table->hash2[i].lock irq_context: 0 reg_requests_lock irq_context: 0 (wq_completion)events reg_work irq_context: 0 detector_work irq_context: 0 (wq_completion)events reg_work rtnl_mutex irq_context: 0 &wq->mutex &pool->lock/1 irq_context: 0 (wq_completion)events reg_work rtnl_mutex reg_requests_lock irq_context: 0 &wq->mutex &x->wait#10 irq_context: 0 (wq_completion)events reg_work rtnl_mutex fs_reclaim irq_context: 0 (wq_completion)events reg_work rtnl_mutex fs_reclaim mmu_notifier_invalidate_range_start irq_context: 0 &pool->lock/1 rcu_read_lock &pool->lock irq_context: 0 (wq_completion)events reg_work rtnl_mutex pool_lock#2 irq_context: 0 &pool->lock/1 rcu_read_lock &pool->lock &obj_hash[i].lock irq_context: 0 (wq_completion)events reg_work rtnl_mutex &obj_hash[i].lock irq_context: 0 &pool->lock/1 rcu_read_lock &pool->lock pool_lock#2 irq_context: 0 (wq_completion)events reg_work rtnl_mutex rcu_read_lock &pool->lock irq_context: 0 &pool->lock/1 rcu_read_lock &pool->lock &p->pi_lock irq_context: 0 (wq_completion)events reg_work rtnl_mutex rcu_read_lock &pool->lock &obj_hash[i].lock irq_context: 0 &pool->lock/1 rcu_read_lock &pool->lock &p->pi_lock &rq->__lock irq_context: 0 (wq_completion)events reg_work rtnl_mutex rcu_read_lock &pool->lock pool_lock#2 irq_context: 0 &pool->lock/1 rcu_read_lock &pool->lock &p->pi_lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)events reg_work rtnl_mutex reg_pending_beacons_lock irq_context: 0 (wq_completion)events reg_work rtnl_mutex &rdev->wiphy.mtx irq_context: 0 acpi_gpio_deferred_req_irqs_lock irq_context: 0 (wq_completion)events reg_work rtnl_mutex &rdev->wiphy.mtx reg_requests_lock irq_context: 0 (wq_completion)events reg_work rtnl_mutex &base->lock irq_context: 0 (wq_completion)events reg_work rtnl_mutex &base->lock &obj_hash[i].lock irq_context: 0 (wq_completion)events (work_completion)(&fw_work->work) irq_context: 0 (wq_completion)events (work_completion)(&fw_work->work) fs_reclaim irq_context: 0 (wq_completion)events (work_completion)(&fw_work->work) fs_reclaim mmu_notifier_invalidate_range_start irq_context: 0 (wq_completion)events (work_completion)(&fw_work->work) pool_lock#2 irq_context: 0 (wq_completion)events (work_completion)(&fw_work->work) &fw_cache.lock irq_context: 0 (wq_completion)events (work_completion)(&fw_work->work) &fw_cache.lock &c->lock irq_context: 0 (wq_completion)events (work_completion)(&fw_work->work) &fw_cache.lock &pcp->lock &zone->lock irq_context: 0 (wq_completion)events (work_completion)(&fw_work->work) &fw_cache.lock &zone->lock irq_context: 0 (wq_completion)events (work_completion)(&fw_work->work) &fw_cache.lock &____s->seqcount irq_context: 0 (wq_completion)events (work_completion)(&fw_work->work) &fw_cache.lock pool_lock#2 irq_context: 0 (wq_completion)events (work_completion)(&fw_work->work) &c->lock irq_context: 0 (wq_completion)events (work_completion)(&fw_work->work) &pcp->lock &zone->lock irq_context: 0 (wq_completion)events (work_completion)(&fw_work->work) &zone->lock irq_context: 0 (wq_completion)events (work_completion)(&fw_work->work) &____s->seqcount irq_context: 0 (wq_completion)events (work_completion)(&fw_work->work) tk_core.seq.seqcount irq_context: 0 (wq_completion)events (work_completion)(&fw_work->work) async_lock irq_context: 0 (wq_completion)events (work_completion)(&fw_work->work) init_task.alloc_lock irq_context: 0 (wq_completion)events (work_completion)(&fw_work->work) init_task.alloc_lock init_fs.lock irq_context: 0 (wq_completion)events (work_completion)(&fw_work->work) init_task.alloc_lock init_fs.lock &dentry->d_lock irq_context: 0 (wq_completion)events (work_completion)(&fw_work->work) rcu_read_lock pool_lock#2 irq_context: 0 (wq_completion)events (work_completion)(&fw_work->work) &obj_hash[i].lock irq_context: 0 (wq_completion)events (work_completion)(&fw_work->work) rcu_read_lock &____s->seqcount#4 irq_context: 0 (wq_completion)events (work_completion)(&fw_work->work) rcu_read_lock &dentry->d_lock irq_context: 0 (wq_completion)events (work_completion)(&fw_work->work) &sb->s_type->i_mutex_key irq_context: 0 (wq_completion)events (work_completion)(&fw_work->work) &sb->s_type->i_mutex_key fs_reclaim irq_context: 0 (wq_completion)events (work_completion)(&fw_work->work) &sb->s_type->i_mutex_key fs_reclaim mmu_notifier_invalidate_range_start irq_context: 0 (wq_completion)events (work_completion)(&fw_work->work) &sb->s_type->i_mutex_key pool_lock#2 irq_context: 0 (wq_completion)events (work_completion)(&fw_work->work) &sb->s_type->i_mutex_key &dentry->d_lock irq_context: 0 (wq_completion)events (work_completion)(&fw_work->work) &sb->s_type->i_mutex_key rcu_read_lock rename_lock.seqcount irq_context: 0 (wq_completion)events (work_completion)(&fw_work->work) &sb->s_type->i_mutex_key &dentry->d_lock &wq irq_context: 0 (wq_completion)events (work_completion)(&fw_work->work) &dentry->d_lock irq_context: 0 (wq_completion)events (work_completion)(&fw_work->work) &base->lock irq_context: 0 (wq_completion)events (work_completion)(&fw_work->work) &base->lock &obj_hash[i].lock irq_context: 0 (wq_completion)events (work_completion)(&fw_work->work) (console_sem).lock irq_context: 0 (wq_completion)events (work_completion)(&fw_work->work) console_owner_lock irq_context: 0 (wq_completion)events (work_completion)(&fw_work->work) console_owner irq_context: 0 (wq_completion)events (work_completion)(&fw_work->work) console_lock console_srcu console_owner_lock irq_context: 0 (wq_completion)events (work_completion)(&fw_work->work) console_lock console_srcu console_owner irq_context: 0 (wq_completion)events (work_completion)(&fw_work->work) console_lock console_srcu console_owner &port_lock_key irq_context: 0 (wq_completion)events (work_completion)(&fw_work->work) console_lock console_srcu console_owner console_owner_lock irq_context: 0 (wq_completion)events (work_completion)(&fw_work->work) umhelper_sem irq_context: 0 (wq_completion)events (work_completion)(&fw_work->work) umhelper_sem usermodehelper_disabled_waitq.lock irq_context: 0 (wq_completion)events (work_completion)(&fw_work->work) umhelper_sem fs_reclaim irq_context: 0 (wq_completion)events (work_completion)(&fw_work->work) umhelper_sem fs_reclaim mmu_notifier_invalidate_range_start irq_context: 0 (wq_completion)events (work_completion)(&fw_work->work) umhelper_sem &c->lock irq_context: 0 (wq_completion)events (work_completion)(&fw_work->work) umhelper_sem &pcp->lock &zone->lock irq_context: 0 (wq_completion)events (work_completion)(&fw_work->work) umhelper_sem &zone->lock irq_context: 0 (wq_completion)events (work_completion)(&fw_work->work) umhelper_sem &____s->seqcount irq_context: 0 (wq_completion)events (work_completion)(&fw_work->work) umhelper_sem pool_lock#2 irq_context: 0 (wq_completion)events (work_completion)(&fw_work->work) umhelper_sem &x->wait#9 irq_context: 0 (wq_completion)events (work_completion)(&fw_work->work) umhelper_sem &obj_hash[i].lock irq_context: 0 (wq_completion)events (work_completion)(&fw_work->work) umhelper_sem &k->list_lock irq_context: 0 (wq_completion)events (work_completion)(&fw_work->work) umhelper_sem gdp_mutex irq_context: 0 (wq_completion)events (work_completion)(&fw_work->work) umhelper_sem gdp_mutex &k->list_lock irq_context: 0 (wq_completion)events (work_completion)(&fw_work->work) umhelper_sem gdp_mutex fs_reclaim irq_context: 0 (wq_completion)events (work_completion)(&fw_work->work) umhelper_sem gdp_mutex fs_reclaim mmu_notifier_invalidate_range_start irq_context: 0 (wq_completion)events (work_completion)(&fw_work->work) umhelper_sem gdp_mutex pool_lock#2 irq_context: 0 (wq_completion)events (work_completion)(&fw_work->work) umhelper_sem gdp_mutex lock irq_context: 0 (wq_completion)events (work_completion)(&fw_work->work) umhelper_sem gdp_mutex lock kernfs_idr_lock irq_context: 0 (wq_completion)events (work_completion)(&fw_work->work) umhelper_sem gdp_mutex &root->kernfs_rwsem irq_context: 0 (wq_completion)events (work_completion)(&fw_work->work) umhelper_sem gdp_mutex &root->kernfs_rwsem &root->kernfs_iattr_rwsem irq_context: 0 (wq_completion)events (work_completion)(&fw_work->work) umhelper_sem lock irq_context: 0 (wq_completion)events (work_completion)(&fw_work->work) umhelper_sem lock kernfs_idr_lock irq_context: 0 (wq_completion)events (work_completion)(&fw_work->work) umhelper_sem &root->kernfs_rwsem irq_context: 0 (wq_completion)events (work_completion)(&fw_work->work) umhelper_sem &root->kernfs_rwsem &root->kernfs_iattr_rwsem irq_context: 0 (wq_completion)events (work_completion)(&fw_work->work) umhelper_sem bus_type_sem irq_context: 0 (wq_completion)events (work_completion)(&fw_work->work) umhelper_sem sysfs_symlink_target_lock irq_context: 0 (wq_completion)events (work_completion)(&fw_work->work) umhelper_sem &root->kernfs_rwsem irq_context: 0 (wq_completion)events (work_completion)(&fw_work->work) umhelper_sem &dev->power.lock irq_context: 0 (wq_completion)events (work_completion)(&fw_work->work) umhelper_sem dpm_list_mtx irq_context: 0 (wq_completion)events (work_completion)(&fw_work->work) umhelper_sem &k->k_lock irq_context: 0 (wq_completion)events (work_completion)(&fw_work->work) umhelper_sem subsys mutex#80 irq_context: 0 (wq_completion)events (work_completion)(&fw_work->work) umhelper_sem subsys mutex#80 &k->k_lock irq_context: 0 (wq_completion)events (work_completion)(&fw_work->work) umhelper_sem fw_lock irq_context: 0 (wq_completion)events (work_completion)(&fw_work->work) umhelper_sem uevent_sock_mutex irq_context: 0 (wq_completion)events (work_completion)(&fw_work->work) umhelper_sem rcu_read_lock &pool->lock/1 irq_context: 0 (wq_completion)events (work_completion)(&fw_work->work) umhelper_sem rcu_read_lock &pool->lock/1 &obj_hash[i].lock irq_context: 0 (wq_completion)events (work_completion)(&fw_work->work) umhelper_sem rcu_read_lock &pool->lock/1 pool_lock#2 irq_context: 0 (wq_completion)events (work_completion)(&fw_work->work) umhelper_sem rcu_read_lock &pool->lock/1 &p->pi_lock irq_context: 0 (wq_completion)events (work_completion)(&fw_work->work) umhelper_sem rcu_read_lock &pool->lock/1 &p->pi_lock &rq->__lock irq_context: 0 (wq_completion)events (work_completion)(&fw_work->work) umhelper_sem rcu_read_lock &pool->lock/1 &p->pi_lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)events (work_completion)(&fw_work->work) umhelper_sem running_helpers_waitq.lock irq_context: 0 (wq_completion)events (work_completion)(&fw_work->work) umhelper_sem &rq->__lock irq_context: 0 (wq_completion)events (work_completion)(&fw_work->work) umhelper_sem &cfs_rq->removed.lock irq_context: 0 (wq_completion)events (work_completion)(&fw_work->work) umhelper_sem &x->wait#23 irq_context: 0 (wq_completion)events (work_completion)(&fw_work->work) umhelper_sem &base->lock irq_context: 0 (wq_completion)events (work_completion)(&fw_work->work) umhelper_sem &base->lock &obj_hash[i].lock irq_context: 0 (wq_completion)events (work_completion)(&fw_work->work) umhelper_sem &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)events (work_completion)(&fw_work->work) umhelper_sem &rq->__lock &cfs_rq->removed.lock irq_context: 0 (wq_completion)events (work_completion)(&fw_work->work) umhelper_sem rcu_read_lock &rq->__lock irq_context: 0 (wq_completion)events (work_completion)(&fw_work->work) umhelper_sem rcu_read_lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: softirq fs/file_table.c:431 irq_context: softirq fs/file_table.c:431 rcu_read_lock &pool->lock irq_context: softirq fs/file_table.c:431 rcu_read_lock &pool->lock &obj_hash[i].lock irq_context: softirq fs/file_table.c:431 rcu_read_lock &pool->lock &p->pi_lock irq_context: softirq fs/file_table.c:431 rcu_read_lock &pool->lock &p->pi_lock &rq->__lock irq_context: softirq fs/file_table.c:431 rcu_read_lock &pool->lock &p->pi_lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)events (delayed_fput_work).work irq_context: 0 (wq_completion)events (delayed_fput_work).work &obj_hash[i].lock irq_context: 0 (wq_completion)events (delayed_fput_work).work pool_lock#2 irq_context: 0 (wq_completion)events (work_completion)(&pwq->unbound_release_work) rcu_state.exp_mutex &rnp->exp_wq[1] irq_context: 0 &sb->s_type->i_lock_key#2 irq_context: 0 tomoyo_ss &c->lock irq_context: 0 tomoyo_ss &pcp->lock &zone->lock irq_context: 0 tomoyo_ss &zone->lock irq_context: 0 tomoyo_ss &____s->seqcount irq_context: 0 tomoyo_ss rcu_read_lock mount_lock.seqcount irq_context: 0 tomoyo_ss rcu_read_lock rcu_read_lock rename_lock.seqcount irq_context: 0 tomoyo_ss &obj_hash[i].lock irq_context: 0 tomoyo_ss rcu_read_lock &dentry->d_lock irq_context: 0 tomoyo_ss tomoyo_log_lock irq_context: 0 tomoyo_ss tomoyo_log_wait.lock irq_context: 0 cdev_lock irq_context: 0 tty_mutex (console_sem).lock irq_context: 0 tty_mutex console_lock irq_context: 0 tty_mutex fs_reclaim irq_context: 0 tty_mutex fs_reclaim mmu_notifier_invalidate_range_start irq_context: 0 tty_mutex pool_lock#2 irq_context: 0 tty_mutex tty_ldiscs_lock irq_context: 0 tty_mutex &obj_hash[i].lock irq_context: 0 tty_mutex &k->list_lock irq_context: 0 tty_mutex &k->k_lock irq_context: 0 tty_mutex &tty->legacy_mutex irq_context: 0 tty_mutex &tty->legacy_mutex &tty->read_wait irq_context: 0 tty_mutex &tty->legacy_mutex &tty->write_wait irq_context: 0 tty_mutex &tty->legacy_mutex &tty->ldisc_sem irq_context: 0 tty_mutex &tty->legacy_mutex &tty->ldisc_sem fs_reclaim irq_context: 0 tty_mutex &tty->legacy_mutex &tty->ldisc_sem fs_reclaim mmu_notifier_invalidate_range_start irq_context: 0 tty_mutex &tty->legacy_mutex &tty->ldisc_sem pool_lock#2 irq_context: 0 tty_mutex &tty->legacy_mutex &tty->ldisc_sem free_vmap_area_lock irq_context: 0 tty_mutex &tty->legacy_mutex &tty->ldisc_sem free_vmap_area_lock &obj_hash[i].lock irq_context: 0 tty_mutex &tty->legacy_mutex &tty->ldisc_sem free_vmap_area_lock pool_lock#2 irq_context: 0 tty_mutex &tty->legacy_mutex &tty->ldisc_sem vmap_area_lock irq_context: 0 tty_mutex &tty->legacy_mutex &tty->ldisc_sem &____s->seqcount irq_context: 0 tty_mutex &tty->legacy_mutex &tty->ldisc_sem &pcp->lock &zone->lock irq_context: 0 tty_mutex &tty->legacy_mutex &tty->ldisc_sem &zone->lock irq_context: 0 tty_mutex &tty->legacy_mutex &tty->ldisc_sem &tty->write_wait irq_context: 0 tty_mutex &tty->legacy_mutex &tty->ldisc_sem &tty->read_wait irq_context: 0 tty_mutex &tty->legacy_mutex &tty->ldisc_sem &tty->termios_rwsem irq_context: 0 &tty->legacy_mutex irq_context: 0 &tty->legacy_mutex &tty->files_lock irq_context: 0 &tty->legacy_mutex &port->lock irq_context: 0 &tty->legacy_mutex &port->mutex irq_context: 0 &tty->legacy_mutex &port->mutex fs_reclaim irq_context: 0 &tty->legacy_mutex &port->mutex fs_reclaim mmu_notifier_invalidate_range_start irq_context: 0 &tty->legacy_mutex &port->mutex &pcp->lock &zone->lock irq_context: 0 &tty->legacy_mutex &port->mutex &zone->lock irq_context: 0 &tty->legacy_mutex &port->mutex &____s->seqcount irq_context: 0 &tty->legacy_mutex &port->mutex pool_lock#2 irq_context: 0 &tty->legacy_mutex &port->mutex &port_lock_key irq_context: 0 &tty->legacy_mutex &port->mutex hash_mutex irq_context: 0 &tty->legacy_mutex &port->mutex hash_mutex fs_reclaim irq_context: 0 &tty->legacy_mutex &port->mutex hash_mutex fs_reclaim mmu_notifier_invalidate_range_start irq_context: 0 &tty->legacy_mutex &port->mutex hash_mutex pool_lock#2 irq_context: 0 &tty->legacy_mutex &port->mutex &i->lock irq_context: 0 &tty->legacy_mutex &port->mutex &desc->request_mutex irq_context: 0 &tty->legacy_mutex &port->mutex &desc->request_mutex &irq_desc_lock_class irq_context: 0 &tty->legacy_mutex &port->mutex &desc->request_mutex &irq_desc_lock_class vector_lock irq_context: 0 &tty->legacy_mutex &port->mutex &desc->request_mutex &irq_desc_lock_class ioapic_lock irq_context: 0 &tty->legacy_mutex &port->mutex &desc->request_mutex &irq_desc_lock_class mask_lock irq_context: 0 &tty->legacy_mutex &port->mutex &desc->request_mutex &irq_desc_lock_class mask_lock tmp_mask_lock irq_context: 0 &tty->legacy_mutex &port->mutex &desc->request_mutex &irq_desc_lock_class mask_lock tmp_mask_lock vector_lock irq_context: 0 &tty->legacy_mutex &port->mutex &desc->request_mutex &irq_desc_lock_class mask_lock tmp_mask_lock ioapic_lock irq_context: 0 &tty->legacy_mutex &port->mutex &desc->request_mutex &irq_desc_lock_class ioapic_lock i8259A_lock irq_context: 0 &tty->legacy_mutex &port->mutex register_lock irq_context: 0 &tty->legacy_mutex &port->mutex &irq_desc_lock_class irq_context: 0 &tty->legacy_mutex &port->mutex proc_subdir_lock irq_context: 0 &tty->legacy_mutex &port->mutex proc_inum_ida.xa_lock irq_context: 0 &tty->legacy_mutex &port->mutex proc_subdir_lock irq_context: hardirq &i->lock irq_context: 0 &tty->legacy_mutex &port_lock_key irq_context: 0 detected_devices_mutex irq_context: 0 sb_writers#2 irq_context: 0 sb_writers#2 mount_lock irq_context: 0 sb_writers#2 &sb->s_type->i_mutex_key/1 irq_context: 0 sb_writers#2 &sb->s_type->i_mutex_key/1 rename_lock.seqcount irq_context: 0 sb_writers#2 &sb->s_type->i_mutex_key/1 fs_reclaim irq_context: 0 sb_writers#2 &sb->s_type->i_mutex_key/1 fs_reclaim mmu_notifier_invalidate_range_start irq_context: 0 sb_writers#2 &sb->s_type->i_mutex_key/1 pool_lock#2 irq_context: 0 sb_writers#2 &sb->s_type->i_mutex_key/1 &dentry->d_lock irq_context: 0 sb_writers#2 &sb->s_type->i_mutex_key/1 rcu_read_lock &dentry->d_lock irq_context: 0 sb_writers#2 &sb->s_type->i_mutex_key/1 &obj_hash[i].lock irq_context: 0 sb_writers#2 &sb->s_type->i_mutex_key/1 tomoyo_ss irq_context: 0 sb_writers#2 &sb->s_type->i_mutex_key/1 tomoyo_ss mmu_notifier_invalidate_range_start irq_context: 0 sb_writers#2 &sb->s_type->i_mutex_key/1 tomoyo_ss pool_lock#2 irq_context: 0 sb_writers#2 &sb->s_type->i_mutex_key/1 tomoyo_ss rcu_read_lock mount_lock.seqcount irq_context: 0 sb_writers#2 &sb->s_type->i_mutex_key/1 tomoyo_ss rcu_read_lock rcu_read_lock rename_lock.seqcount irq_context: 0 sb_writers#2 &sb->s_type->i_mutex_key/1 tomoyo_ss &obj_hash[i].lock irq_context: 0 sb_writers#2 &sb->s_type->i_mutex_key/1 tomoyo_ss rcu_read_lock &dentry->d_lock irq_context: 0 sb_writers#2 &sb->s_type->i_mutex_key/1 tomoyo_ss tomoyo_log_lock irq_context: 0 sb_writers#2 &sb->s_type->i_mutex_key/1 tomoyo_ss tomoyo_log_wait.lock irq_context: 0 sb_writers#2 &sb->s_type->i_mutex_key/1 tomoyo_ss &____s->seqcount irq_context: 0 sb_writers#2 &sb->s_type->i_mutex_key/1 tomoyo_ss &pcp->lock &zone->lock irq_context: 0 sb_writers#2 &sb->s_type->i_mutex_key/1 tomoyo_ss &zone->lock irq_context: 0 sb_writers#2 &sb->s_type->i_mutex_key/1 tomoyo_ss rcu_read_lock pool_lock#2 irq_context: 0 sb_writers#2 &sb->s_type->i_mutex_key/1 tomoyo_ss tomoyo_policy_lock irq_context: 0 sb_writers#2 &sb->s_type->i_mutex_key/1 tomoyo_ss tomoyo_policy_lock mmu_notifier_invalidate_range_start irq_context: 0 sb_writers#2 &sb->s_type->i_mutex_key/1 tomoyo_ss tomoyo_policy_lock pool_lock#2 irq_context: 0 sb_writers#2 &sb->s_type->i_mutex_key/1 &c->lock irq_context: 0 sb_writers#2 &sb->s_type->i_mutex_key/1 &pcp->lock &zone->lock irq_context: 0 sb_writers#2 &sb->s_type->i_mutex_key/1 &zone->lock irq_context: 0 sb_writers#2 &sb->s_type->i_mutex_key/1 &pcp->lock &zone->lock &____s->seqcount irq_context: 0 sb_writers#2 &sb->s_type->i_mutex_key/1 &____s->seqcount irq_context: 0 sb_writers#2 &sb->s_type->i_mutex_key/1 mmu_notifier_invalidate_range_start irq_context: 0 sb_writers#2 &sb->s_type->i_mutex_key/1 &sb->s_type->i_lock_key#2 irq_context: 0 sb_writers#2 &sb->s_type->i_mutex_key/1 &s->s_inode_list_lock irq_context: 0 sb_writers#2 &sb->s_type->i_mutex_key/1 tk_core.seq.seqcount irq_context: 0 sb_writers#2 &sb->s_type->i_mutex_key/1 &sb->s_type->i_lock_key#2 &dentry->d_lock irq_context: 0 &sb->s_type->i_mutex_key rename_lock.seqcount irq_context: 0 &sb->s_type->i_mutex_key &dentry->d_lock &wq#2 irq_context: 0 sb_writers#2 &sb->s_type->i_mutex_key/1 &sb->s_type->i_mutex_key irq_context: 0 sb_writers#2 &sb->s_type->i_mutex_key/1 &sb->s_type->i_mutex_key tk_core.seq.seqcount irq_context: 0 sb_writers#2 &sb->s_type->i_mutex_key/1 &sb->s_type->i_mutex_key rcu_read_lock &dentry->d_lock irq_context: 0 sb_writers#2 &sb->s_type->i_mutex_key/1 &sb->s_type->i_mutex_key &dentry->d_lock irq_context: 0 sb_writers#2 &sb->s_type->i_mutex_key/1 tomoyo_ss &c->lock irq_context: 0 tomoyo_ss file_systems_lock irq_context: 0 tomoyo_ss fs_reclaim irq_context: 0 tomoyo_ss fs_reclaim mmu_notifier_invalidate_range_start irq_context: 0 tomoyo_ss rcu_read_lock init_fs.seq.seqcount irq_context: 0 &type->s_umount_key#25/1 irq_context: 0 &type->s_umount_key#25/1 fs_reclaim irq_context: 0 &type->s_umount_key#25/1 fs_reclaim mmu_notifier_invalidate_range_start irq_context: 0 &type->s_umount_key#25/1 pool_lock#2 irq_context: 0 &type->s_umount_key#25/1 pcpu_alloc_mutex irq_context: 0 &type->s_umount_key#25/1 pcpu_alloc_mutex pcpu_lock irq_context: 0 &type->s_umount_key#25/1 shrinker_rwsem irq_context: 0 &type->s_umount_key#25/1 list_lrus_mutex irq_context: 0 &type->s_umount_key#25/1 sb_lock irq_context: 0 &disk->open_mutex bdev_lock irq_context: 0 &disk->open_mutex bdev_lock &bdev->bd_holder_lock irq_context: 0 &bdev->bd_fsfreeze_mutex irq_context: 0 &bdev->bd_fsfreeze_mutex sb_lock irq_context: 0 &type->s_umount_key#26 irq_context: 0 &type->s_umount_key#26 fs_reclaim irq_context: 0 &type->s_umount_key#26 fs_reclaim mmu_notifier_invalidate_range_start irq_context: 0 &type->s_umount_key#26 &c->lock irq_context: 0 &type->s_umount_key#26 &pcp->lock &zone->lock irq_context: 0 &type->s_umount_key#26 &zone->lock irq_context: 0 &type->s_umount_key#26 &____s->seqcount irq_context: 0 &type->s_umount_key#26 pool_lock#2 irq_context: 0 &type->s_umount_key#26 &obj_hash[i].lock irq_context: 0 &type->s_umount_key#26 pcpu_alloc_mutex irq_context: 0 &type->s_umount_key#26 pcpu_alloc_mutex pcpu_lock irq_context: 0 &type->s_umount_key#26 &wq->mutex irq_context: 0 &type->s_umount_key#26 &wq->mutex &pool->lock irq_context: 0 &type->s_umount_key#26 kthread_create_lock irq_context: 0 &type->s_umount_key#26 &p->pi_lock irq_context: 0 &type->s_umount_key#26 &p->pi_lock &rq->__lock irq_context: 0 &type->s_umount_key#26 &p->pi_lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 &type->s_umount_key#26 &rq->__lock irq_context: 0 &type->s_umount_key#26 &x->wait irq_context: 0 &type->s_umount_key#26 &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 &type->s_umount_key#26 wq_pool_mutex irq_context: 0 &type->s_umount_key#26 wq_pool_mutex &wq->mutex irq_context: 0 &type->s_umount_key#26 mmu_notifier_invalidate_range_start irq_context: 0 &type->s_umount_key#26 &pcp->lock &zone->lock &____s->seqcount irq_context: 0 &type->s_umount_key#26 &xa->xa_lock#9 irq_context: 0 &type->s_umount_key#26 &xa->xa_lock#9 pool_lock#2 irq_context: 0 &type->s_umount_key#26 lock#4 irq_context: 0 &type->s_umount_key#26 &mapping->private_lock irq_context: 0 &type->s_umount_key#26 tk_core.seq.seqcount irq_context: 0 &type->s_umount_key#26 &dd->lock irq_context: 0 &type->s_umount_key#26 rcu_read_lock &pool->lock irq_context: 0 &type->s_umount_key#26 rcu_read_lock &pool->lock &obj_hash[i].lock irq_context: 0 &type->s_umount_key#26 rcu_read_lock &pool->lock pool_lock#2 irq_context: 0 &type->s_umount_key#26 rcu_read_lock &pool->lock &p->pi_lock irq_context: 0 &type->s_umount_key#26 rcu_read_lock &pool->lock &p->pi_lock &rq->__lock irq_context: 0 &type->s_umount_key#26 rcu_read_lock &pool->lock &p->pi_lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 &type->s_umount_key#26 bit_wait_table + i irq_context: 0 (wq_completion)kblockd (work_completion)(&(&hctx->run_work)->work) rcu_read_lock &base->lock irq_context: 0 (wq_completion)kblockd (work_completion)(&(&hctx->run_work)->work) rcu_read_lock &base->lock &obj_hash[i].lock irq_context: softirq bit_wait_table + i irq_context: softirq bit_wait_table + i &p->pi_lock irq_context: softirq bit_wait_table + i &p->pi_lock &rq->__lock irq_context: softirq bit_wait_table + i &p->pi_lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 &type->s_umount_key#32 cpu_hotplug_lock wq_pool_mutex &zone->lock irq_context: 0 &type->s_umount_key#26 &wq->mutex &x->wait#10 irq_context: 0 &type->s_umount_key#26 wq_mayday_lock irq_context: 0 &p->alloc_lock &x->wait &p->pi_lock irq_context: 0 &type->s_umount_key#26 wq_pool_mutex &wq->mutex &pool->lock irq_context: 0 &type->s_umount_key#26 rcu_state.exp_mutex rcu_node_0 irq_context: 0 &type->s_umount_key#26 rcu_state.exp_mutex &obj_hash[i].lock irq_context: 0 &type->s_umount_key#26 rcu_state.exp_mutex rcu_read_lock &pool->lock irq_context: 0 &type->s_umount_key#26 rcu_state.exp_mutex rcu_read_lock &pool->lock &obj_hash[i].lock irq_context: 0 &type->s_umount_key#26 rcu_state.exp_mutex rcu_read_lock &pool->lock &p->pi_lock irq_context: 0 &type->s_umount_key#26 rcu_state.exp_mutex rcu_read_lock &pool->lock &p->pi_lock &rq->__lock irq_context: 0 &type->s_umount_key#26 rcu_state.exp_mutex rcu_read_lock &pool->lock &p->pi_lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 &type->s_umount_key#26 rcu_state.exp_mutex &rnp->exp_wq[2] irq_context: 0 &type->s_umount_key#26 rcu_state.exp_mutex &rq->__lock irq_context: 0 &type->s_umount_key#26 &sbi->old_work_lock irq_context: 0 &type->s_umount_key#26 (work_completion)(&(&sbi->old_work)->work) irq_context: 0 &type->s_umount_key#26 shrinker_rwsem irq_context: 0 &xa->xa_lock#4 irq_context: 0 sb_lock &obj_hash[i].lock irq_context: 0 sb_lock pool_lock#2 irq_context: 0 &type->s_umount_key#27/1 irq_context: 0 &type->s_umount_key#27/1 fs_reclaim irq_context: 0 &type->s_umount_key#27/1 fs_reclaim mmu_notifier_invalidate_range_start irq_context: 0 &type->s_umount_key#27/1 pool_lock#2 irq_context: 0 &type->s_umount_key#27/1 pcpu_alloc_mutex irq_context: 0 &type->s_umount_key#27/1 pcpu_alloc_mutex pcpu_lock irq_context: 0 &type->s_umount_key#27/1 shrinker_rwsem irq_context: 0 &type->s_umount_key#27/1 list_lrus_mutex irq_context: 0 &type->s_umount_key#27/1 sb_lock irq_context: 0 &type->s_umount_key#28 irq_context: 0 &type->s_umount_key#28 fs_reclaim irq_context: 0 &type->s_umount_key#28 fs_reclaim mmu_notifier_invalidate_range_start irq_context: 0 &type->s_umount_key#28 pool_lock#2 irq_context: 0 &type->s_umount_key#28 mmu_notifier_invalidate_range_start irq_context: 0 &type->s_umount_key#28 &____s->seqcount irq_context: 0 &type->s_umount_key#28 &xa->xa_lock#9 irq_context: 0 &type->s_umount_key#28 lock#4 irq_context: 0 &type->s_umount_key#28 &mapping->private_lock irq_context: 0 &type->s_umount_key#28 tk_core.seq.seqcount irq_context: 0 &type->s_umount_key#28 &dd->lock irq_context: 0 &type->s_umount_key#28 &obj_hash[i].lock irq_context: 0 &type->s_umount_key#28 rcu_read_lock &pool->lock irq_context: 0 &type->s_umount_key#28 rcu_read_lock &pool->lock &obj_hash[i].lock irq_context: 0 &type->s_umount_key#28 rcu_read_lock &pool->lock pool_lock#2 irq_context: 0 &type->s_umount_key#28 rcu_read_lock &pool->lock &p->pi_lock irq_context: 0 &type->s_umount_key#28 rcu_read_lock &pool->lock &p->pi_lock &rq->__lock irq_context: 0 &type->s_umount_key#28 rcu_read_lock &pool->lock &p->pi_lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 &type->s_umount_key#28 bit_wait_table + i irq_context: 0 &type->s_umount_key#28 &rq->__lock irq_context: 0 &type->s_umount_key#28 &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 &type->s_umount_key#32 cpu_hotplug_lock wq_pool_mutex rcu_read_lock pool_lock#2 irq_context: 0 &type->s_umount_key#32 cpu_hotplug_lock wq_pool_mutex &wq->mutex irq_context: 0 &type->s_umount_key#32 cpu_hotplug_lock wq_pool_mutex &wq->mutex &pool->lock/1 irq_context: 0 &type->s_umount_key#32 wq_pool_mutex irq_context: 0 &type->s_umount_key#32 wq_pool_mutex &wq->mutex irq_context: 0 &type->s_umount_key#32 &sb->s_type->i_lock_key#22 &dentry->d_lock irq_context: 0 &type->s_umount_key#32 &ei->i_es_lock irq_context: 0 &type->s_umount_key#32 ext4_grpinfo_slab_create_mutex irq_context: 0 &type->s_umount_key#32 ext4_grpinfo_slab_create_mutex slab_mutex irq_context: 0 &type->s_umount_key#32 ext4_grpinfo_slab_create_mutex slab_mutex fs_reclaim irq_context: 0 &type->s_umount_key#32 ext4_grpinfo_slab_create_mutex slab_mutex fs_reclaim mmu_notifier_invalidate_range_start irq_context: 0 &type->s_umount_key#32 ext4_grpinfo_slab_create_mutex slab_mutex pool_lock#2 irq_context: 0 &type->s_umount_key#32 ext4_grpinfo_slab_create_mutex slab_mutex &c->lock irq_context: 0 &type->s_umount_key#32 ext4_grpinfo_slab_create_mutex slab_mutex &n->list_lock irq_context: 0 &type->s_umount_key#32 ext4_grpinfo_slab_create_mutex slab_mutex &____s->seqcount irq_context: 0 &type->s_umount_key#32 ext4_grpinfo_slab_create_mutex slab_mutex pcpu_alloc_mutex irq_context: 0 &type->s_umount_key#32 ext4_grpinfo_slab_create_mutex slab_mutex pcpu_alloc_mutex pcpu_lock irq_context: 0 &type->s_umount_key#32 ext4_grpinfo_slab_create_mutex slab_mutex &root->kernfs_rwsem irq_context: 0 &type->s_umount_key#32 ext4_grpinfo_slab_create_mutex slab_mutex &k->list_lock irq_context: 0 &type->s_umount_key#32 ext4_grpinfo_slab_create_mutex slab_mutex lock irq_context: 0 &type->s_umount_key#32 ext4_grpinfo_slab_create_mutex slab_mutex lock kernfs_idr_lock irq_context: 0 &type->s_umount_key#32 ext4_grpinfo_slab_create_mutex slab_mutex &root->kernfs_rwsem &root->kernfs_iattr_rwsem irq_context: 0 &type->s_umount_key#32 ext4_grpinfo_slab_create_mutex slab_mutex &pcp->lock &zone->lock irq_context: 0 &type->s_umount_key#32 ext4_grpinfo_slab_create_mutex slab_mutex &zone->lock irq_context: 0 &type->s_umount_key#32 ext4_grpinfo_slab_create_mutex slab_mutex lock kernfs_idr_lock pool_lock#2 irq_context: 0 &type->s_umount_key#32 &s->s_inode_list_lock irq_context: 0 &type->s_umount_key#32 ext4_li_mtx irq_context: 0 &type->s_umount_key#32 lock irq_context: 0 &type->s_umount_key#32 lock kernfs_idr_lock irq_context: 0 &type->s_umount_key#32 &root->kernfs_rwsem irq_context: 0 &type->s_umount_key#32 &root->kernfs_rwsem &root->kernfs_iattr_rwsem irq_context: 0 &type->s_umount_key#32 (console_sem).lock irq_context: 0 &type->s_umount_key#32 console_lock console_srcu console_owner_lock irq_context: 0 &type->s_umount_key#32 console_lock console_srcu console_owner irq_context: 0 &type->s_umount_key#32 console_lock console_srcu console_owner &port_lock_key irq_context: 0 &type->s_umount_key#32 console_lock console_srcu console_owner console_owner_lock irq_context: 0 &type->s_umount_key#32 &dentry->d_lock irq_context: 0 &sb->s_type->i_lock_key#22 irq_context: 0 &type->i_mutex_dir_key#3 irq_context: 0 &type->i_mutex_dir_key#3 fs_reclaim irq_context: 0 &type->i_mutex_dir_key#3 fs_reclaim mmu_notifier_invalidate_range_start irq_context: 0 &type->i_mutex_dir_key#3 pool_lock#2 irq_context: 0 &type->i_mutex_dir_key#3 &dentry->d_lock irq_context: 0 &type->i_mutex_dir_key#3 rcu_read_lock rename_lock.seqcount irq_context: 0 &type->i_mutex_dir_key#3 &ei->i_es_lock irq_context: 0 &type->i_mutex_dir_key#3 &ei->i_data_sem irq_context: 0 &type->i_mutex_dir_key#3 &ei->i_data_sem mmu_notifier_invalidate_range_start irq_context: 0 &type->i_mutex_dir_key#3 &ei->i_data_sem pool_lock#2 irq_context: 0 &type->i_mutex_dir_key#3 &ei->i_data_sem &ei->i_es_lock irq_context: 0 &type->i_mutex_dir_key#3 &ei->i_data_sem &ei->i_es_lock pool_lock#2 irq_context: 0 &type->i_mutex_dir_key#3 &ei->i_data_sem &ei->i_es_lock &sbi->s_es_lock irq_context: 0 &type->i_mutex_dir_key#3 &ei->i_data_sem &obj_hash[i].lock irq_context: 0 &type->i_mutex_dir_key#3 &ei->i_data_sem &ei->i_es_lock &obj_hash[i].lock irq_context: 0 &type->i_mutex_dir_key#3 mmu_notifier_invalidate_range_start irq_context: 0 &type->i_mutex_dir_key#3 &____s->seqcount irq_context: 0 &type->i_mutex_dir_key#3 &xa->xa_lock#9 irq_context: 0 &type->i_mutex_dir_key#3 &xa->xa_lock#9 pool_lock#2 irq_context: 0 &type->i_mutex_dir_key#3 lock#4 irq_context: 0 &type->i_mutex_dir_key#3 &mapping->private_lock irq_context: 0 &type->i_mutex_dir_key#3 tk_core.seq.seqcount irq_context: 0 &type->i_mutex_dir_key#3 &dd->lock irq_context: 0 &type->i_mutex_dir_key#3 &obj_hash[i].lock irq_context: 0 &type->i_mutex_dir_key#3 rcu_read_lock &pool->lock irq_context: 0 &type->i_mutex_dir_key#3 rcu_read_lock &pool->lock &obj_hash[i].lock irq_context: 0 &type->i_mutex_dir_key#3 rcu_read_lock &pool->lock pool_lock#2 irq_context: 0 &type->i_mutex_dir_key#3 rcu_read_lock &pool->lock &p->pi_lock irq_context: 0 &type->i_mutex_dir_key#3 rcu_read_lock &pool->lock &p->pi_lock &rq->__lock irq_context: 0 &type->i_mutex_dir_key#3 rcu_read_lock &pool->lock &p->pi_lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 &type->i_mutex_dir_key#3 bit_wait_table + i irq_context: 0 &type->i_mutex_dir_key#3 &rq->__lock irq_context: 0 &type->i_mutex_dir_key#3 &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 &type->i_mutex_dir_key#3 inode_hash_lock irq_context: 0 &type->i_mutex_dir_key#3 inode_hash_lock &sb->s_type->i_lock_key#22 irq_context: 0 &type->i_mutex_dir_key#3 inode_hash_lock &s->s_inode_list_lock irq_context: 0 &type->i_mutex_dir_key#3 &journal->j_state_lock irq_context: 0 &type->i_mutex_dir_key#3 &sb->s_type->i_lock_key#22 irq_context: 0 &type->i_mutex_dir_key#3 &sb->s_type->i_lock_key#22 &dentry->d_lock irq_context: 0 &type->i_mutex_dir_key#3 &sb->s_type->i_lock_key#22 &dentry->d_lock &wq irq_context: 0 tomoyo_ss rcu_read_lock pool_lock#2 irq_context: 0 &type->i_mutex_dir_key#3 irq_context: 0 &type->i_mutex_dir_key#3 namespace_sem irq_context: 0 &type->i_mutex_dir_key#3 namespace_sem rcu_read_lock mount_lock.seqcount irq_context: 0 &type->i_mutex_dir_key#3 namespace_sem fs_reclaim irq_context: 0 &type->i_mutex_dir_key#3 namespace_sem fs_reclaim mmu_notifier_invalidate_range_start irq_context: 0 &type->i_mutex_dir_key#3 namespace_sem pool_lock#2 irq_context: 0 &type->i_mutex_dir_key#3 namespace_sem rename_lock irq_context: 0 &type->i_mutex_dir_key#3 namespace_sem rename_lock rename_lock.seqcount irq_context: 0 &type->i_mutex_dir_key#3 namespace_sem rename_lock rename_lock.seqcount &dentry->d_lock irq_context: 0 &type->i_mutex_dir_key#3 namespace_sem mount_lock irq_context: 0 &type->i_mutex_dir_key#3 namespace_sem mount_lock &dentry->d_lock irq_context: 0 &type->i_mutex_dir_key#3 namespace_sem mount_lock mount_lock.seqcount irq_context: 0 &type->i_mutex_dir_key#3 namespace_sem mount_lock mount_lock.seqcount &new_ns->poll irq_context: 0 &type->i_mutex_dir_key#3 namespace_sem mount_lock mount_lock.seqcount &dentry->d_lock irq_context: 0 &type->i_mutex_dir_key#3 namespace_sem mount_lock mount_lock.seqcount rcu_read_lock &dentry->d_lock irq_context: 0 &type->i_mutex_dir_key#3 namespace_sem mount_lock mount_lock.seqcount &obj_hash[i].lock irq_context: 0 &type->i_mutex_dir_key#3 namespace_sem mount_lock mount_lock.seqcount pool_lock#2 irq_context: 0 &sb->s_type->i_mutex_key namespace_sem &dentry->d_lock irq_context: 0 &sb->s_type->i_mutex_key namespace_sem rcu_read_lock &dentry->d_lock irq_context: 0 &sb->s_type->i_mutex_key namespace_sem &obj_hash[i].lock irq_context: 0 vmap_purge_lock free_vmap_area_lock irq_context: 0 vmap_purge_lock free_vmap_area_lock &obj_hash[i].lock irq_context: 0 vmap_purge_lock free_vmap_area_lock pool_lock#2 irq_context: 0 vmap_purge_lock free_vmap_area_lock init_mm.page_table_lock irq_context: 0 vmap_purge_lock free_vmap_area_lock init_mm.page_table_lock rcu_read_lock pool_lock#2 irq_context: 0 vmap_purge_lock free_vmap_area_lock init_mm.page_table_lock &obj_hash[i].lock irq_context: 0 vmap_purge_lock free_vmap_area_lock quarantine_lock irq_context: 0 rcu_state.barrier_mutex irq_context: 0 rcu_state.barrier_mutex rcu_state.barrier_lock irq_context: hardirq rcu_state.barrier_lock &obj_hash[i].lock irq_context: 0 rcu_state.barrier_mutex &x->wait#24 irq_context: 0 rcu_state.barrier_mutex &rq->__lock irq_context: 0 rcu_state.barrier_mutex &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: softirq rcu_callback &x->wait#24 irq_context: softirq rcu_callback &x->wait#24 &p->pi_lock irq_context: softirq rcu_callback &x->wait#24 &p->pi_lock &cfs_rq->removed.lock irq_context: softirq rcu_callback &x->wait#24 &p->pi_lock &rq->__lock irq_context: softirq rcu_callback &x->wait#24 &p->pi_lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (init_mm).mmap_lock irq_context: 0 (init_mm).mmap_lock &rq->__lock irq_context: 0 (init_mm).mmap_lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 &type->s_umount_key#33/1 irq_context: 0 &type->s_umount_key#33/1 fs_reclaim irq_context: 0 &type->s_umount_key#33/1 fs_reclaim mmu_notifier_invalidate_range_start irq_context: 0 &type->s_umount_key#33/1 pool_lock#2 irq_context: 0 &type->s_umount_key#33/1 pcpu_alloc_mutex irq_context: 0 &type->s_umount_key#33/1 pcpu_alloc_mutex pcpu_lock irq_context: 0 &type->s_umount_key#33/1 shrinker_rwsem irq_context: 0 &type->s_umount_key#33/1 list_lrus_mutex irq_context: 0 &type->s_umount_key#33/1 sb_lock irq_context: 0 &type->s_umount_key#33/1 sb_lock unnamed_dev_ida.xa_lock irq_context: 0 &type->s_umount_key#33/1 &pcp->lock &zone->lock irq_context: 0 &type->s_umount_key#33/1 &zone->lock irq_context: 0 &type->s_umount_key#33/1 &pcp->lock &zone->lock &____s->seqcount irq_context: 0 &type->s_umount_key#33/1 &____s->seqcount irq_context: 0 &type->s_umount_key#33/1 &c->lock irq_context: 0 &type->s_umount_key#33/1 mmu_notifier_invalidate_range_start irq_context: 0 &type->s_umount_key#33/1 &sb->s_type->i_lock_key#23 irq_context: 0 &type->s_umount_key#33/1 &s->s_inode_list_lock irq_context: 0 &type->s_umount_key#33/1 tk_core.seq.seqcount irq_context: 0 &type->s_umount_key#33/1 &sb->s_type->i_lock_key#23 &dentry->d_lock irq_context: 0 &type->s_umount_key#33/1 &sb->s_type->i_mutex_key#9 irq_context: 0 &type->s_umount_key#33/1 &sb->s_type->i_mutex_key#9 fs_reclaim irq_context: 0 &type->s_umount_key#33/1 &sb->s_type->i_mutex_key#9 fs_reclaim mmu_notifier_invalidate_range_start irq_context: 0 &type->s_umount_key#33/1 &sb->s_type->i_mutex_key#9 pool_lock#2 irq_context: 0 &type->s_umount_key#33/1 &sb->s_type->i_mutex_key#9 &dentry->d_lock irq_context: 0 &type->s_umount_key#33/1 &sb->s_type->i_mutex_key#9 mmu_notifier_invalidate_range_start irq_context: 0 &type->s_umount_key#33/1 &sb->s_type->i_mutex_key#9 &pcp->lock &zone->lock irq_context: 0 &type->s_umount_key#33/1 &sb->s_type->i_mutex_key#9 &zone->lock irq_context: 0 &type->s_umount_key#33/1 &sb->s_type->i_mutex_key#9 &____s->seqcount irq_context: 0 &type->s_umount_key#33/1 &sb->s_type->i_mutex_key#9 rcu_read_lock pool_lock#2 irq_context: 0 &type->s_umount_key#33/1 &sb->s_type->i_mutex_key#9 &obj_hash[i].lock irq_context: 0 &type->s_umount_key#33/1 &sb->s_type->i_mutex_key#9 &sb->s_type->i_lock_key#23 irq_context: 0 &type->s_umount_key#33/1 &sb->s_type->i_mutex_key#9 &s->s_inode_list_lock irq_context: 0 &type->s_umount_key#33/1 &sb->s_type->i_mutex_key#9 tk_core.seq.seqcount irq_context: 0 &type->s_umount_key#33/1 &sb->s_type->i_mutex_key#9 &sb->s_type->i_lock_key#23 &dentry->d_lock irq_context: 0 &type->s_umount_key#33/1 &dentry->d_lock irq_context: 0 rcu_read_lock &sb->s_type->i_lock_key#23 irq_context: 0 &sb->s_type->i_mutex_key#9 irq_context: 0 &sb->s_type->i_mutex_key#9 fs_reclaim irq_context: 0 &sb->s_type->i_mutex_key#9 fs_reclaim mmu_notifier_invalidate_range_start irq_context: 0 &sb->s_type->i_mutex_key#9 pool_lock#2 irq_context: 0 &sb->s_type->i_mutex_key#9 &dentry->d_lock irq_context: 0 &sb->s_type->i_mutex_key#9 rcu_read_lock rename_lock.seqcount irq_context: 0 &sb->s_type->i_mutex_key#9 proc_subdir_lock irq_context: 0 &sb->s_type->i_mutex_key#9 mmu_notifier_invalidate_range_start irq_context: 0 &sb->s_type->i_mutex_key#9 &sb->s_type->i_lock_key#23 irq_context: 0 &sb->s_type->i_mutex_key#9 &s->s_inode_list_lock irq_context: 0 &sb->s_type->i_mutex_key#9 tk_core.seq.seqcount irq_context: 0 &sb->s_type->i_mutex_key#9 &sb->s_type->i_lock_key#23 &dentry->d_lock irq_context: 0 &sb->s_type->i_mutex_key#9 &sb->s_type->i_lock_key#23 &dentry->d_lock &wq irq_context: 0 &sb->s_type->i_mutex_key#9 sysctl_lock irq_context: 0 sb_writers#3 irq_context: 0 sb_writers#3 mount_lock irq_context: 0 sb_writers#3 &sb->s_type->i_mutex_key#9 irq_context: 0 sb_writers#3 &sb->s_type->i_mutex_key#9 rename_lock.seqcount irq_context: 0 sb_writers#3 &sb->s_type->i_mutex_key#9 fs_reclaim irq_context: 0 sb_writers#3 &sb->s_type->i_mutex_key#9 fs_reclaim mmu_notifier_invalidate_range_start irq_context: 0 sb_writers#3 &sb->s_type->i_mutex_key#9 &c->lock irq_context: 0 sb_writers#3 &sb->s_type->i_mutex_key#9 &pcp->lock &zone->lock irq_context: 0 sb_writers#3 &sb->s_type->i_mutex_key#9 &zone->lock irq_context: 0 sb_writers#3 &sb->s_type->i_mutex_key#9 &____s->seqcount irq_context: 0 sb_writers#3 &sb->s_type->i_mutex_key#9 pool_lock#2 irq_context: 0 sb_writers#3 &sb->s_type->i_mutex_key#9 &dentry->d_lock irq_context: 0 sb_writers#3 &sb->s_type->i_mutex_key#9 rcu_read_lock rename_lock.seqcount irq_context: 0 sb_writers#3 &sb->s_type->i_mutex_key#9 sysctl_lock irq_context: 0 sb_writers#3 &sb->s_type->i_mutex_key#9 mmu_notifier_invalidate_range_start irq_context: 0 sb_writers#3 &sb->s_type->i_mutex_key#9 &sb->s_type->i_lock_key#23 irq_context: 0 sb_writers#3 &sb->s_type->i_mutex_key#9 &s->s_inode_list_lock irq_context: 0 sb_writers#3 &sb->s_type->i_mutex_key#9 tk_core.seq.seqcount irq_context: 0 sb_writers#3 &sb->s_type->i_mutex_key#9 &sb->s_type->i_lock_key#23 &dentry->d_lock irq_context: 0 sb_writers#3 &sb->s_type->i_mutex_key#9 &sb->s_type->i_lock_key#23 &dentry->d_lock &wq#2 irq_context: 0 tomoyo_ss rcu_read_lock rename_lock.seqcount irq_context: 0 &sb->s_type->i_mutex_key#9 irq_context: 0 sb_writers#3 sysctl_lock irq_context: 0 sb_writers#3 fs_reclaim irq_context: 0 sb_writers#3 fs_reclaim mmu_notifier_invalidate_range_start irq_context: 0 sb_writers#3 pool_lock#2 irq_context: 0 sb_writers#3 &obj_hash[i].lock irq_context: 0 sb_writers#3 &h->resize_lock irq_context: 0 sb_writers#3 &h->resize_lock free_hpage_work irq_context: 0 sb_writers#3 &h->resize_lock hugetlb_lock irq_context: 0 sb_writers#3 &h->resize_lock fs_reclaim irq_context: 0 sb_writers#3 &h->resize_lock fs_reclaim mmu_notifier_invalidate_range_start irq_context: 0 sb_writers#3 &h->resize_lock &pcp->lock &zone->lock irq_context: 0 sb_writers#3 &h->resize_lock &pcp->lock &zone->lock &____s->seqcount irq_context: 0 sb_writers#3 &h->resize_lock &____s->seqcount irq_context: 0 sb_writers#3 &h->resize_lock pool_lock#2 irq_context: 0 sb_writers#3 hugetlb_lock irq_context: 0 sb_writers#3 &sb->s_type->i_mutex_key#9 &dentry->d_lock &wq#2 irq_context: 0 sb_writers#3 &sb->s_type->i_mutex_key#9 rcu_read_lock &dentry->d_lock irq_context: 0 sb_writers#3 &sb->s_type->i_mutex_key#9 &obj_hash[i].lock irq_context: 0 tomoyo_ss tomoyo_policy_lock &c->lock irq_context: 0 tomoyo_ss tomoyo_policy_lock &pcp->lock &zone->lock irq_context: 0 tomoyo_ss tomoyo_policy_lock &zone->lock irq_context: 0 tomoyo_ss tomoyo_policy_lock &____s->seqcount irq_context: 0 &sig->cred_guard_mutex init_fs.lock irq_context: 0 &sig->cred_guard_mutex rcu_read_lock init_fs.seq.seqcount irq_context: 0 &sig->cred_guard_mutex &type->i_mutex_dir_key#3 irq_context: 0 &sig->cred_guard_mutex &type->i_mutex_dir_key#3 fs_reclaim irq_context: 0 &sig->cred_guard_mutex &type->i_mutex_dir_key#3 fs_reclaim mmu_notifier_invalidate_range_start irq_context: 0 &sig->cred_guard_mutex &type->i_mutex_dir_key#3 pool_lock#2 irq_context: 0 &sig->cred_guard_mutex &type->i_mutex_dir_key#3 &dentry->d_lock irq_context: 0 &sig->cred_guard_mutex &type->i_mutex_dir_key#3 rcu_read_lock rename_lock.seqcount irq_context: 0 &sig->cred_guard_mutex &type->i_mutex_dir_key#3 &ei->i_es_lock irq_context: 0 &sig->cred_guard_mutex &type->i_mutex_dir_key#3 &mapping->private_lock irq_context: 0 &sig->cred_guard_mutex &type->i_mutex_dir_key#3 inode_hash_lock irq_context: 0 &sig->cred_guard_mutex &type->i_mutex_dir_key#3 mmu_notifier_invalidate_range_start irq_context: 0 &sig->cred_guard_mutex &type->i_mutex_dir_key#3 &pcp->lock &zone->lock irq_context: 0 &sig->cred_guard_mutex &type->i_mutex_dir_key#3 &zone->lock irq_context: 0 &sig->cred_guard_mutex &type->i_mutex_dir_key#3 &____s->seqcount irq_context: 0 &sig->cred_guard_mutex &type->i_mutex_dir_key#3 &c->lock irq_context: 0 &sig->cred_guard_mutex &type->i_mutex_dir_key#3 &obj_hash[i].lock irq_context: 0 &sig->cred_guard_mutex &type->i_mutex_dir_key#3 inode_hash_lock &sb->s_type->i_lock_key#22 irq_context: 0 &sig->cred_guard_mutex &type->i_mutex_dir_key#3 inode_hash_lock &s->s_inode_list_lock irq_context: 0 &sig->cred_guard_mutex &type->i_mutex_dir_key#3 &xa->xa_lock#9 irq_context: 0 &sig->cred_guard_mutex &type->i_mutex_dir_key#3 &xa->xa_lock#9 pool_lock#2 irq_context: 0 &sig->cred_guard_mutex &type->i_mutex_dir_key#3 lock#4 irq_context: 0 &sig->cred_guard_mutex &type->i_mutex_dir_key#3 &pcp->lock &zone->lock &____s->seqcount irq_context: 0 &sig->cred_guard_mutex &type->i_mutex_dir_key#3 tk_core.seq.seqcount irq_context: 0 &sig->cred_guard_mutex &type->i_mutex_dir_key#3 &dd->lock irq_context: softirq &(&ovs_net->masks_rebalance)->timer irq_context: softirq &(&ovs_net->masks_rebalance)->timer rcu_read_lock &pool->lock irq_context: softirq &(&ovs_net->masks_rebalance)->timer rcu_read_lock &pool->lock &obj_hash[i].lock irq_context: softirq &(&ovs_net->masks_rebalance)->timer rcu_read_lock &pool->lock &p->pi_lock irq_context: softirq &(&ovs_net->masks_rebalance)->timer rcu_read_lock &pool->lock &p->pi_lock &rq->__lock irq_context: softirq &(&ovs_net->masks_rebalance)->timer rcu_read_lock &pool->lock &p->pi_lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)events (work_completion)(&(&ovs_net->masks_rebalance)->work) irq_context: 0 (wq_completion)events (work_completion)(&(&ovs_net->masks_rebalance)->work) ovs_mutex irq_context: 0 (wq_completion)events (work_completion)(&(&ovs_net->masks_rebalance)->work) &obj_hash[i].lock irq_context: 0 (wq_completion)events (work_completion)(&(&ovs_net->masks_rebalance)->work) &base->lock irq_context: 0 (wq_completion)events (work_completion)(&(&ovs_net->masks_rebalance)->work) &base->lock &obj_hash[i].lock irq_context: 0 &sig->cred_guard_mutex &type->i_mutex_dir_key#3 lock#4 &lruvec->lru_lock irq_context: 0 &sig->cred_guard_mutex &type->i_mutex_dir_key#3 rcu_read_lock &dd->lock irq_context: 0 &sig->cred_guard_mutex &type->i_mutex_dir_key#3 rcu_read_lock &obj_hash[i].lock irq_context: 0 &sig->cred_guard_mutex &type->i_mutex_dir_key#3 rcu_read_lock &pcp->lock &zone->lock irq_context: 0 &sig->cred_guard_mutex &type->i_mutex_dir_key#3 rcu_read_lock &zone->lock irq_context: 0 &sig->cred_guard_mutex &type->i_mutex_dir_key#3 rcu_read_lock &____s->seqcount irq_context: 0 &sig->cred_guard_mutex &type->i_mutex_dir_key#3 rcu_read_lock pool_lock#2 irq_context: 0 &sig->cred_guard_mutex &type->i_mutex_dir_key#3 rcu_read_lock &c->lock irq_context: 0 &sig->cred_guard_mutex &type->i_mutex_dir_key#3 rcu_read_lock tk_core.seq.seqcount irq_context: 0 &sig->cred_guard_mutex &type->i_mutex_dir_key#3 rcu_read_lock &virtscsi_vq->vq_lock irq_context: 0 &sig->cred_guard_mutex &type->i_mutex_dir_key#3 bit_wait_table + i irq_context: 0 &sig->cred_guard_mutex &type->i_mutex_dir_key#3 &rq->__lock irq_context: 0 &sig->cred_guard_mutex &type->i_mutex_dir_key#3 &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: softirq bit_wait_table + i &p->pi_lock &cfs_rq->removed.lock irq_context: 0 &sig->cred_guard_mutex &type->i_mutex_dir_key#3 &journal->j_state_lock irq_context: 0 &sig->cred_guard_mutex &type->i_mutex_dir_key#3 &sb->s_type->i_lock_key#22 irq_context: 0 &sig->cred_guard_mutex &type->i_mutex_dir_key#3 &sb->s_type->i_lock_key#22 &dentry->d_lock irq_context: 0 &sig->cred_guard_mutex &type->i_mutex_dir_key#3 &sb->s_type->i_lock_key#22 &dentry->d_lock &wq irq_context: 0 &sig->cred_guard_mutex &sb->s_type->i_lock_key#22 irq_context: 0 &sig->cred_guard_mutex &type->i_mutex_dir_key#3 rename_lock.seqcount irq_context: 0 &sig->cred_guard_mutex &type->i_mutex_dir_key#3 &ei->i_data_sem irq_context: 0 &sig->cred_guard_mutex &type->i_mutex_dir_key#3 &ei->i_data_sem mmu_notifier_invalidate_range_start irq_context: 0 &sig->cred_guard_mutex &type->i_mutex_dir_key#3 &ei->i_data_sem pool_lock#2 irq_context: 0 &sig->cred_guard_mutex &type->i_mutex_dir_key#3 &ei->i_data_sem &ei->i_es_lock irq_context: 0 &sig->cred_guard_mutex &type->i_mutex_dir_key#3 &ei->i_data_sem &ei->i_es_lock pool_lock#2 irq_context: 0 &sig->cred_guard_mutex &type->i_mutex_dir_key#3 &ei->i_data_sem &ei->i_es_lock &sbi->s_es_lock irq_context: 0 &sig->cred_guard_mutex &type->i_mutex_dir_key#3 &ei->i_data_sem &obj_hash[i].lock irq_context: 0 &sig->cred_guard_mutex &type->i_mutex_dir_key#3 &ei->i_data_sem &ei->i_es_lock &obj_hash[i].lock irq_context: 0 &sig->cred_guard_mutex &type->i_mutex_dir_key#3 rcu_read_lock &pool->lock irq_context: 0 &sig->cred_guard_mutex &type->i_mutex_dir_key#3 rcu_read_lock &pool->lock &obj_hash[i].lock irq_context: 0 &sig->cred_guard_mutex &type->i_mutex_dir_key#3 rcu_read_lock &pool->lock pool_lock#2 irq_context: 0 &sig->cred_guard_mutex &type->i_mutex_dir_key#3 rcu_read_lock &pool->lock &p->pi_lock irq_context: 0 &sig->cred_guard_mutex &type->i_mutex_dir_key#3 rcu_read_lock &pool->lock &p->pi_lock &rq->__lock irq_context: 0 &sig->cred_guard_mutex &type->i_mutex_dir_key#3 rcu_read_lock &pool->lock &p->pi_lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 &sig->cred_guard_mutex &type->i_mutex_dir_key#3 &sb->s_type->i_lock_key#22 &dentry->d_lock &wq#2 irq_context: 0 &sig->cred_guard_mutex &sb->s_type->i_mutex_key#8 irq_context: 0 &sig->cred_guard_mutex &p->pi_lock irq_context: 0 &sig->cred_guard_mutex aa_buffers_lock irq_context: 0 &sig->cred_guard_mutex rcu_read_lock mount_lock.seqcount irq_context: 0 &sig->cred_guard_mutex rcu_read_lock rcu_read_lock rename_lock.seqcount irq_context: 0 &sig->cred_guard_mutex mapping.invalidate_lock irq_context: 0 &sig->cred_guard_mutex mapping.invalidate_lock mmu_notifier_invalidate_range_start irq_context: 0 &sig->cred_guard_mutex mapping.invalidate_lock &pcp->lock &zone->lock irq_context: 0 &sig->cred_guard_mutex mapping.invalidate_lock &pcp->lock &zone->lock &____s->seqcount irq_context: 0 &sig->cred_guard_mutex mapping.invalidate_lock &____s->seqcount irq_context: 0 &sig->cred_guard_mutex mapping.invalidate_lock pool_lock#2 irq_context: 0 &sig->cred_guard_mutex mapping.invalidate_lock &xa->xa_lock#9 irq_context: 0 &sig->cred_guard_mutex mapping.invalidate_lock lock#4 irq_context: 0 &sig->cred_guard_mutex mapping.invalidate_lock &xa->xa_lock#9 pool_lock#2 irq_context: 0 &sig->cred_guard_mutex mapping.invalidate_lock &ei->i_es_lock irq_context: 0 &sig->cred_guard_mutex mapping.invalidate_lock &ei->i_data_sem irq_context: 0 &sig->cred_guard_mutex mapping.invalidate_lock &ei->i_data_sem mmu_notifier_invalidate_range_start irq_context: 0 &sig->cred_guard_mutex mapping.invalidate_lock &ei->i_data_sem pool_lock#2 irq_context: 0 &sig->cred_guard_mutex mapping.invalidate_lock &ei->i_data_sem &ei->i_es_lock irq_context: 0 &sig->cred_guard_mutex mapping.invalidate_lock &ei->i_data_sem &ei->i_es_lock pool_lock#2 irq_context: 0 &sig->cred_guard_mutex mapping.invalidate_lock &ei->i_data_sem &ei->i_es_lock &sbi->s_es_lock irq_context: 0 &sig->cred_guard_mutex mapping.invalidate_lock &ei->i_data_sem &obj_hash[i].lock irq_context: 0 &sig->cred_guard_mutex mapping.invalidate_lock &ei->i_data_sem &ei->i_es_lock &obj_hash[i].lock irq_context: 0 &sig->cred_guard_mutex mapping.invalidate_lock tk_core.seq.seqcount irq_context: 0 &sig->cred_guard_mutex mapping.invalidate_lock &dd->lock irq_context: 0 &sig->cred_guard_mutex mapping.invalidate_lock rcu_read_lock &dd->lock irq_context: 0 &sig->cred_guard_mutex mapping.invalidate_lock rcu_read_lock &obj_hash[i].lock irq_context: 0 &sig->cred_guard_mutex mapping.invalidate_lock rcu_read_lock tk_core.seq.seqcount irq_context: 0 &sig->cred_guard_mutex mapping.invalidate_lock rcu_read_lock &virtscsi_vq->vq_lock irq_context: 0 &sig->cred_guard_mutex &folio_wait_table[i] irq_context: 0 &sig->cred_guard_mutex &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 &sig->cred_guard_mutex tomoyo_ss irq_context: 0 &sig->cred_guard_mutex tomoyo_ss mmu_notifier_invalidate_range_start irq_context: 0 &sig->cred_guard_mutex tomoyo_ss pool_lock#2 irq_context: 0 &sig->cred_guard_mutex tomoyo_ss &c->lock irq_context: 0 &sig->cred_guard_mutex tomoyo_ss &pcp->lock &zone->lock irq_context: 0 &sig->cred_guard_mutex tomoyo_ss &zone->lock irq_context: 0 &sig->cred_guard_mutex tomoyo_ss &____s->seqcount irq_context: 0 &sig->cred_guard_mutex tomoyo_ss fs_reclaim irq_context: 0 &sig->cred_guard_mutex tomoyo_ss fs_reclaim mmu_notifier_invalidate_range_start irq_context: 0 &sig->cred_guard_mutex tomoyo_ss rcu_read_lock init_fs.seq.seqcount irq_context: 0 &sig->cred_guard_mutex tomoyo_ss rcu_read_lock &dentry->d_lock irq_context: 0 &sig->cred_guard_mutex tomoyo_ss &obj_hash[i].lock irq_context: 0 &sig->cred_guard_mutex tomoyo_ss rcu_read_lock mount_lock.seqcount irq_context: 0 &sig->cred_guard_mutex tomoyo_ss rcu_read_lock rcu_read_lock rename_lock.seqcount irq_context: 0 &sig->cred_guard_mutex tomoyo_ss &mm->mmap_lock irq_context: 0 &sig->cred_guard_mutex tomoyo_ss &mm->mmap_lock ptlock_ptr(page)#2 irq_context: 0 &sig->cred_guard_mutex tomoyo_ss tomoyo_log_lock irq_context: 0 &sig->cred_guard_mutex tomoyo_ss tomoyo_log_wait.lock irq_context: 0 &sig->cred_guard_mutex tomoyo_ss tomoyo_policy_lock irq_context: 0 &sig->cred_guard_mutex tomoyo_ss tomoyo_policy_lock mmu_notifier_invalidate_range_start irq_context: 0 &sig->cred_guard_mutex tomoyo_ss tomoyo_policy_lock pool_lock#2 irq_context: 0 &sig->cred_guard_mutex tomoyo_ss &pcp->lock &zone->lock &____s->seqcount irq_context: 0 &sig->cred_guard_mutex &sb->s_type->i_mutex_key#8 mmu_notifier_invalidate_range_start irq_context: 0 &sig->cred_guard_mutex &sb->s_type->i_mutex_key#8 &pcp->lock &zone->lock irq_context: 0 &sig->cred_guard_mutex &sb->s_type->i_mutex_key#8 &zone->lock irq_context: 0 &sig->cred_guard_mutex &sb->s_type->i_mutex_key#8 &____s->seqcount irq_context: 0 &sig->cred_guard_mutex &sb->s_type->i_mutex_key#8 pool_lock#2 irq_context: 0 &sig->cred_guard_mutex &sb->s_type->i_mutex_key#8 &c->lock irq_context: 0 &sig->cred_guard_mutex &sb->s_type->i_mutex_key#8 integrity_iint_lock irq_context: 0 &sig->cred_guard_mutex &iint->mutex irq_context: 0 &sig->cred_guard_mutex &iint->mutex &ei->xattr_sem irq_context: 0 &sig->cred_guard_mutex &iint->mutex fs_reclaim irq_context: 0 &sig->cred_guard_mutex &iint->mutex fs_reclaim mmu_notifier_invalidate_range_start irq_context: 0 &sig->cred_guard_mutex &iint->mutex pool_lock#2 irq_context: 0 &sig->cred_guard_mutex &iint->mutex &obj_hash[i].lock irq_context: 0 &sig->cred_guard_mutex &iint->mutex mmu_notifier_invalidate_range_start irq_context: 0 &sig->cred_guard_mutex &iint->mutex rcu_read_lock mount_lock.seqcount irq_context: 0 &sig->cred_guard_mutex &iint->mutex rcu_read_lock rcu_read_lock rename_lock.seqcount irq_context: 0 &sig->cred_guard_mutex &iint->mutex ima_extend_list_mutex irq_context: 0 &sig->cred_guard_mutex &iint->mutex ima_extend_list_mutex fs_reclaim irq_context: 0 &sig->cred_guard_mutex &iint->mutex ima_extend_list_mutex fs_reclaim mmu_notifier_invalidate_range_start irq_context: 0 &sig->cred_guard_mutex &iint->mutex ima_extend_list_mutex pool_lock#2 irq_context: 0 &sig->cred_guard_mutex binfmt_lock irq_context: 0 &sig->cred_guard_mutex entries_lock irq_context: 0 &sig->cred_guard_mutex &type->i_mutex_dir_key#3 &obj_hash[i].lock pool_lock irq_context: 0 &sig->cred_guard_mutex &dentry->d_lock &lru->node[i].lock irq_context: 0 &sig->cred_guard_mutex mapping.invalidate_lock &xa->xa_lock#9 &c->lock irq_context: 0 &sig->cred_guard_mutex mapping.invalidate_lock &xa->xa_lock#9 &pcp->lock &zone->lock irq_context: 0 &sig->cred_guard_mutex mapping.invalidate_lock &xa->xa_lock#9 &zone->lock irq_context: 0 &sig->cred_guard_mutex mapping.invalidate_lock &xa->xa_lock#9 &____s->seqcount irq_context: 0 &sig->cred_guard_mutex mapping.invalidate_lock lock#4 &lruvec->lru_lock irq_context: 0 &sig->cred_guard_mutex &ei->xattr_sem irq_context: 0 &sig->cred_guard_mutex &tsk->futex_exit_mutex irq_context: 0 &sig->cred_guard_mutex &tsk->futex_exit_mutex &p->pi_lock irq_context: 0 &sig->cred_guard_mutex &sig->exec_update_lock irq_context: 0 &sig->cred_guard_mutex &sig->exec_update_lock &p->alloc_lock irq_context: 0 &sig->cred_guard_mutex &sig->exec_update_lock &p->alloc_lock &memcg->mm_list.lock irq_context: 0 &sig->cred_guard_mutex &sig->exec_update_lock &sighand->siglock irq_context: 0 &sig->cred_guard_mutex &sig->exec_update_lock &newf->file_lock irq_context: 0 &sig->cred_guard_mutex &sig->exec_update_lock batched_entropy_u64.lock irq_context: 0 &sig->cred_guard_mutex &sig->exec_update_lock batched_entropy_u64.lock crngs.lock irq_context: 0 batched_entropy_u16.lock irq_context: 0 batched_entropy_u16.lock crngs.lock irq_context: 0 &mm->mmap_lock &anon_vma->rwsem irq_context: 0 &mm->mmap_lock &vma->vm_lock->lock irq_context: 0 &mm->mmap_lock mmu_notifier_invalidate_range_start irq_context: 0 &mm->mmap_lock ptlock_ptr(page)#2 irq_context: 0 &mm->mmap_lock &anon_vma->rwsem irq_context: 0 &mm->mmap_lock &anon_vma->rwsem &obj_hash[i].lock irq_context: 0 &mm->mmap_lock &anon_vma->rwsem pool_lock#2 irq_context: 0 &mm->mmap_lock &mm->page_table_lock irq_context: 0 &mm->mmap_lock ptlock_ptr(page) irq_context: 0 &mm->mmap_lock ptlock_ptr(page)#2 ptlock_ptr(page)#2/1 irq_context: 0 &mm->mmap_lock lock#4 irq_context: 0 &mm->mmap_lock lock#4 &lruvec->lru_lock irq_context: 0 &mm->mmap_lock lock#5 irq_context: 0 &mm->mmap_lock &obj_hash[i].lock irq_context: 0 &mm->mmap_lock rcu_read_lock pool_lock#2 irq_context: 0 &mm->mmap_lock &anon_vma->rwsem &mm->page_table_lock irq_context: 0 &mm->mmap_lock &anon_vma->rwsem &mm->page_table_lock &obj_hash[i].lock irq_context: 0 &mm->mmap_lock &anon_vma->rwsem &mm->page_table_lock pool_lock#2 irq_context: 0 &mm->mmap_lock &mapping->i_mmap_rwsem irq_context: 0 &mm->mmap_lock &mapping->i_mmap_rwsem &obj_hash[i].lock irq_context: 0 &mm->mmap_lock &mapping->i_mmap_rwsem pool_lock#2 irq_context: 0 &sb->s_type->i_mutex_key#8 irq_context: 0 &sb->s_type->i_mutex_key#8 integrity_iint_lock irq_context: 0 &iint->mutex irq_context: 0 &sb->s_type->i_mutex_key#8 mmu_notifier_invalidate_range_start irq_context: 0 &sb->s_type->i_mutex_key#8 pool_lock#2 irq_context: 0 &sb->s_type->i_mutex_key#8 integrity_iint_lock irq_context: 0 &iint->mutex &ei->xattr_sem irq_context: 0 &iint->mutex fs_reclaim irq_context: 0 &iint->mutex fs_reclaim mmu_notifier_invalidate_range_start irq_context: 0 &iint->mutex pool_lock#2 irq_context: 0 &iint->mutex mapping.invalidate_lock irq_context: 0 &iint->mutex mapping.invalidate_lock mmu_notifier_invalidate_range_start irq_context: 0 &iint->mutex mapping.invalidate_lock &pcp->lock &zone->lock irq_context: 0 &iint->mutex mapping.invalidate_lock &pcp->lock &zone->lock &____s->seqcount irq_context: 0 &iint->mutex mapping.invalidate_lock &____s->seqcount irq_context: 0 &iint->mutex mapping.invalidate_lock pool_lock#2 irq_context: 0 &iint->mutex mapping.invalidate_lock &xa->xa_lock#9 irq_context: 0 &iint->mutex mapping.invalidate_lock lock#4 irq_context: 0 &iint->mutex mapping.invalidate_lock &ei->i_es_lock irq_context: 0 &iint->mutex mapping.invalidate_lock &zone->lock irq_context: 0 &iint->mutex mapping.invalidate_lock &c->lock irq_context: 0 &iint->mutex mapping.invalidate_lock tk_core.seq.seqcount irq_context: 0 &iint->mutex mapping.invalidate_lock &dd->lock irq_context: 0 &iint->mutex mapping.invalidate_lock rcu_read_lock &dd->lock irq_context: 0 &iint->mutex mapping.invalidate_lock rcu_read_lock &obj_hash[i].lock irq_context: 0 &iint->mutex mapping.invalidate_lock rcu_read_lock &obj_hash[i].lock pool_lock irq_context: 0 &iint->mutex mapping.invalidate_lock rcu_read_lock tk_core.seq.seqcount irq_context: 0 &iint->mutex mapping.invalidate_lock rcu_read_lock &virtscsi_vq->vq_lock irq_context: 0 &iint->mutex mapping.invalidate_lock lock#4 &lruvec->lru_lock irq_context: 0 &iint->mutex &folio_wait_table[i] irq_context: 0 &iint->mutex &rq->__lock irq_context: 0 &iint->mutex &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: hardirq &rq->__lock pool_lock#2 irq_context: 0 &iint->mutex &obj_hash[i].lock irq_context: 0 &iint->mutex mmu_notifier_invalidate_range_start irq_context: 0 &iint->mutex rcu_read_lock mount_lock.seqcount irq_context: 0 &iint->mutex rcu_read_lock rcu_read_lock rename_lock.seqcount irq_context: 0 &iint->mutex ima_extend_list_mutex irq_context: 0 &iint->mutex ima_extend_list_mutex fs_reclaim irq_context: 0 &iint->mutex ima_extend_list_mutex fs_reclaim mmu_notifier_invalidate_range_start irq_context: 0 &iint->mutex ima_extend_list_mutex pool_lock#2 irq_context: 0 binfmt_lock irq_context: 0 &dentry->d_lock &lru->node[i].lock irq_context: 0 &fsnotify_mark_srcu irq_context: 0 &type->s_umount_key#34 irq_context: 0 &type->s_umount_key#34 shrinker_rwsem irq_context: 0 &type->s_umount_key#34 rcu_read_lock &dentry->d_lock irq_context: 0 &type->s_umount_key#34 &dentry->d_lock irq_context: 0 &type->s_umount_key#34 &dentry->d_lock &lru->node[i].lock irq_context: 0 &type->s_umount_key#34 rename_lock.seqcount irq_context: 0 &type->s_umount_key#34 &dentry->d_lock &dentry->d_lock/1 irq_context: 0 &type->s_umount_key#34 &dentry->d_lock/1 irq_context: 0 &type->s_umount_key#34 &dentry->d_lock &dentry->d_lock/1 &lru->node[i].lock irq_context: 0 &type->s_umount_key#34 &sb->s_type->i_lock_key#23 irq_context: 0 &type->s_umount_key#34 &s->s_inode_list_lock irq_context: 0 &type->s_umount_key#34 &xa->xa_lock#9 irq_context: 0 &type->s_umount_key#34 sysctl_lock irq_context: 0 &type->s_umount_key#34 &obj_hash[i].lock irq_context: 0 &type->s_umount_key#34 pool_lock#2 irq_context: 0 &type->s_umount_key#34 &fsnotify_mark_srcu irq_context: 0 &type->s_umount_key#34 &obj_hash[i].lock pool_lock irq_context: 0 unnamed_dev_ida.xa_lock irq_context: 0 sb_lock &obj_hash[i].lock pool_lock irq_context: 0 &xa->xa_lock#9 irq_context: 0 prog_idr_lock irq_context: 0 prog_idr_lock &obj_hash[i].lock irq_context: 0 prog_idr_lock pool_lock#2 irq_context: 0 map_idr_lock irq_context: 0 (wq_completion)events_unbound (work_completion)(&map->work) irq_context: 0 (wq_completion)events_unbound (work_completion)(&map->work) &obj_hash[i].lock irq_context: 0 (wq_completion)events_unbound (work_completion)(&map->work) pool_lock#2 irq_context: 0 btf_idr_lock irq_context: 0 btf_idr_lock &obj_hash[i].lock irq_context: 0 btf_idr_lock pool_lock#2 irq_context: 0 &mm->mmap_lock rcu_read_lock rcu_read_lock ptlock_ptr(page)#2 irq_context: 0 &mm->mmap_lock rcu_read_lock rcu_read_lock ptlock_ptr(page)#2 key irq_context: 0 &vma->vm_lock->lock fs_reclaim irq_context: 0 &vma->vm_lock->lock fs_reclaim mmu_notifier_invalidate_range_start irq_context: 0 &vma->vm_lock->lock &pcp->lock &zone->lock irq_context: 0 &vma->vm_lock->lock &____s->seqcount irq_context: 0 &vma->vm_lock->lock pool_lock#2 irq_context: 0 &vma->vm_lock->lock ptlock_ptr(page)#2 irq_context: 0 &vma->vm_lock->lock ptlock_ptr(page)#2 lock#4 irq_context: 0 &mm->mmap_lock mmu_notifier_invalidate_range_start irq_context: 0 &type->i_mutex_dir_key#3 &obj_hash[i].lock pool_lock irq_context: 0 &type->i_mutex_dir_key#3 &pcp->lock &zone->lock irq_context: 0 &type->i_mutex_dir_key#3 &pcp->lock &zone->lock &____s->seqcount irq_context: 0 &type->i_mutex_dir_key#3 &zone->lock irq_context: 0 &type->i_mutex_dir_key#3 &dentry->d_lock &wq irq_context: 0 &type->i_mutex_dir_key#3 rename_lock.seqcount irq_context: 0 &type->i_mutex_dir_key#3 &dentry->d_lock &wq#2 irq_context: 0 &type->i_mutex_dir_key#3 &c->lock irq_context: 0 &type->i_mutex_dir_key#3 &sb->s_type->i_lock_key#22 &dentry->d_lock &wq#2 irq_context: 0 &iint->mutex mapping.invalidate_lock &xa->xa_lock#9 pool_lock#2 irq_context: 0 &iint->mutex mapping.invalidate_lock &ei->i_data_sem irq_context: 0 &iint->mutex mapping.invalidate_lock &ei->i_data_sem mmu_notifier_invalidate_range_start irq_context: 0 &iint->mutex mapping.invalidate_lock &ei->i_data_sem pool_lock#2 irq_context: 0 &iint->mutex mapping.invalidate_lock &ei->i_data_sem &ei->i_es_lock irq_context: 0 &iint->mutex mapping.invalidate_lock &ei->i_data_sem &ei->i_es_lock pool_lock#2 irq_context: 0 &iint->mutex mapping.invalidate_lock &ei->i_data_sem &ei->i_es_lock &sbi->s_es_lock irq_context: 0 &iint->mutex mapping.invalidate_lock &ei->i_data_sem &obj_hash[i].lock irq_context: 0 &mm->mmap_lock &mapping->i_mmap_rwsem &pcp->lock &zone->lock irq_context: 0 &mm->mmap_lock &mapping->i_mmap_rwsem &zone->lock irq_context: 0 &mm->mmap_lock &mapping->i_mmap_rwsem &____s->seqcount irq_context: 0 &mm->mmap_lock &mapping->i_mmap_rwsem rcu_read_lock pool_lock#2 irq_context: 0 &iint->mutex mapping.invalidate_lock rcu_read_lock pool_lock#2 irq_context: 0 &iint->mutex mapping.invalidate_lock &ei->i_es_lock key#2 irq_context: 0 &iint->mutex mapping.invalidate_lock rcu_read_lock &c->lock irq_context: 0 &iint->mutex mapping.invalidate_lock rcu_read_lock &____s->seqcount irq_context: 0 &iint->mutex batched_entropy_u8.lock irq_context: 0 &iint->mutex kfence_freelist_lock irq_context: 0 &mm->mmap_lock rcu_read_lock rcu_read_lock ptlock_ptr(page) irq_context: 0 &iint->mutex &c->lock irq_context: 0 &iint->mutex &____s->seqcount irq_context: 0 &type->i_mutex_dir_key#3 lock#4 &lruvec->lru_lock irq_context: 0 &type->i_mutex_dir_key#3 rcu_read_lock &dd->lock irq_context: 0 &type->i_mutex_dir_key#3 rcu_read_lock &obj_hash[i].lock irq_context: 0 &type->i_mutex_dir_key#3 rcu_read_lock &obj_hash[i].lock pool_lock irq_context: 0 &type->i_mutex_dir_key#3 rcu_read_lock pool_lock#2 irq_context: 0 &type->i_mutex_dir_key#3 rcu_read_lock tk_core.seq.seqcount irq_context: 0 &type->i_mutex_dir_key#3 rcu_read_lock &virtscsi_vq->vq_lock irq_context: 0 &iint->mutex mapping.invalidate_lock &ei->i_data_sem &c->lock irq_context: 0 &iint->mutex mapping.invalidate_lock &ei->i_data_sem &____s->seqcount irq_context: softirq (&rtpcp->lazy_timer) rcu_tasks_trace__percpu.cbs_pcpu_lock &obj_hash[i].lock irq_context: softirq (&rtpcp->lazy_timer) rcu_tasks_trace__percpu.cbs_pcpu_lock &base->lock irq_context: softirq (&rtpcp->lazy_timer) rcu_tasks_trace__percpu.cbs_pcpu_lock &base->lock &obj_hash[i].lock irq_context: softirq (&rtpcp->lazy_timer) rcu_read_lock &p->pi_lock irq_context: softirq (&rtpcp->lazy_timer) rcu_read_lock &p->pi_lock &cfs_rq->removed.lock irq_context: softirq (&rtpcp->lazy_timer) rcu_read_lock &p->pi_lock &rq->__lock irq_context: softirq (&rtpcp->lazy_timer) rcu_read_lock &p->pi_lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 rcu_tasks_trace.tasks_gp_mutex rcu_state.exp_mutex &obj_hash[i].lock irq_context: 0 rcu_tasks_trace.tasks_gp_mutex rcu_state.exp_mutex rcu_read_lock &pool->lock irq_context: 0 rcu_tasks_trace.tasks_gp_mutex rcu_state.exp_mutex rcu_read_lock &pool->lock &obj_hash[i].lock irq_context: 0 rcu_tasks_trace.tasks_gp_mutex rcu_state.exp_mutex rcu_read_lock &pool->lock &p->pi_lock irq_context: 0 rcu_tasks_trace.tasks_gp_mutex rcu_state.exp_mutex rcu_read_lock &pool->lock &p->pi_lock &rq->__lock irq_context: 0 rcu_tasks_trace.tasks_gp_mutex rcu_state.exp_mutex rcu_read_lock &pool->lock &p->pi_lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 rcu_tasks_trace.tasks_gp_mutex rcu_state.exp_mutex &rnp->exp_wq[0] irq_context: 0 rcu_tasks_trace.tasks_gp_mutex rcu_state.exp_mutex &rq->__lock irq_context: 0 rcu_tasks_trace.tasks_gp_mutex rcu_state.exp_mutex &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 rcu_tasks_trace.tasks_gp_mutex rcu_read_lock &pool->lock irq_context: 0 rcu_tasks_trace.tasks_gp_mutex rcu_read_lock &pool->lock &obj_hash[i].lock irq_context: 0 rcu_tasks_trace.tasks_gp_mutex rcu_read_lock &pool->lock pool_lock#2 irq_context: 0 rcu_tasks_trace.tasks_gp_mutex rcu_read_lock &pool->lock &p->pi_lock irq_context: 0 rcu_tasks_trace.tasks_gp_mutex rcu_read_lock &pool->lock &p->pi_lock &rq->__lock irq_context: 0 rcu_tasks_trace.tasks_gp_mutex rcu_read_lock &pool->lock &p->pi_lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)events (work_completion)(&aux->work) irq_context: 0 (wq_completion)events (work_completion)(&aux->work) map_idr_lock irq_context: 0 (wq_completion)events (work_completion)(&aux->work) map_idr_lock &obj_hash[i].lock irq_context: 0 (wq_completion)events (work_completion)(&aux->work) map_idr_lock pool_lock#2 irq_context: 0 (wq_completion)events (work_completion)(&aux->work) &obj_hash[i].lock irq_context: 0 (wq_completion)events (work_completion)(&aux->work) rcu_read_lock &pool->lock/1 irq_context: 0 (wq_completion)events (work_completion)(&aux->work) rcu_read_lock &pool->lock/1 &obj_hash[i].lock irq_context: 0 (wq_completion)events (work_completion)(&aux->work) rcu_read_lock &pool->lock/1 pool_lock#2 irq_context: 0 (wq_completion)events (work_completion)(&aux->work) rcu_read_lock &pool->lock/1 &p->pi_lock irq_context: 0 (wq_completion)events (work_completion)(&aux->work) rcu_read_lock &pool->lock/1 &p->pi_lock &rq->__lock irq_context: 0 (wq_completion)events (work_completion)(&aux->work) rcu_read_lock &pool->lock/1 &p->pi_lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)events (work_completion)(&aux->work) pool_lock#2 irq_context: 0 (wq_completion)events (work_completion)(&aux->work) pack_mutex irq_context: 0 (wq_completion)events (work_completion)(&aux->work) pack_mutex text_mutex irq_context: 0 (wq_completion)events (work_completion)(&aux->work) pack_mutex text_mutex ptlock_ptr(page)#2 irq_context: 0 (wq_completion)events (work_completion)(&aux->work) pcpu_lock irq_context: 0 (wq_completion)events (work_completion)(&aux->work) vmap_area_lock irq_context: 0 (wq_completion)events (work_completion)(&aux->work) purge_vmap_area_lock irq_context: 0 (wq_completion)events (work_completion)(&aux->work) rcu_read_lock pool_lock#2 irq_context: 0 &iint->mutex mapping.invalidate_lock &xa->xa_lock#9 &c->lock irq_context: 0 &iint->mutex mapping.invalidate_lock &xa->xa_lock#9 &____s->seqcount irq_context: 0 &mm->mmap_lock &mapping->i_mmap_rwsem &anon_vma->rwsem irq_context: 0 &mm->mmap_lock &mapping->i_mmap_rwsem &anon_vma->rwsem &obj_hash[i].lock irq_context: 0 &mm->mmap_lock &mapping->i_mmap_rwsem &anon_vma->rwsem pool_lock#2 irq_context: 0 &mm->mmap_lock &mapping->i_mmap_rwsem &anon_vma->rwsem &obj_hash[i].lock pool_lock irq_context: 0 &vma->vm_lock->lock ptlock_ptr(page)#2 key irq_context: 0 &mm->mmap_lock ptlock_ptr(page)#2 lock#4 &lruvec->lru_lock irq_context: 0 &vma->vm_lock->lock mmu_notifier_invalidate_range_start irq_context: 0 tomoyo_ss &pcp->lock &zone->lock &____s->seqcount irq_context: 0 &port->mutex irq_context: 0 &tty->ldisc_sem irq_context: 0 &tty->ldisc_sem &tty->termios_rwsem irq_context: 0 &tty->ldisc_sem &mm->mmap_lock irq_context: 0 &tty->ldisc_sem &tty->termios_rwsem irq_context: 0 &tty->ldisc_sem &tty->termios_rwsem &port->mutex irq_context: 0 &tty->ldisc_sem &tty->termios_rwsem &tty->ldisc_sem &tty->write_wait irq_context: 0 &tty->ldisc_sem &tty->termios_rwsem &tty->ldisc_sem &tty->read_wait irq_context: 0 &iint->mutex mapping.invalidate_lock &ei->i_data_sem &ei->i_es_lock &obj_hash[i].lock irq_context: 0 integrity_iint_lock irq_context: 0 &sb->s_type->i_mutex_key#9 rename_lock.seqcount irq_context: 0 &sb->s_type->i_mutex_key#9 &sb->s_type->i_lock_key#23 &dentry->d_lock &wq#2 irq_context: 0 &sb->s_type->i_lock_key#23 irq_context: 0 &p->lock irq_context: 0 &p->lock fs_reclaim irq_context: 0 &p->lock fs_reclaim mmu_notifier_invalidate_range_start irq_context: 0 &p->lock pool_lock#2 irq_context: 0 &p->lock &mm->mmap_lock irq_context: 0 &type->s_umount_key#35/1 irq_context: 0 &type->s_umount_key#35/1 fs_reclaim irq_context: 0 &type->s_umount_key#35/1 fs_reclaim mmu_notifier_invalidate_range_start irq_context: 0 &type->s_umount_key#35/1 pool_lock#2 irq_context: 0 &type->s_umount_key#35/1 pcpu_alloc_mutex irq_context: 0 &type->s_umount_key#35/1 pcpu_alloc_mutex pcpu_lock irq_context: 0 &type->s_umount_key#35/1 shrinker_rwsem irq_context: 0 &type->s_umount_key#35/1 list_lrus_mutex irq_context: 0 &type->s_umount_key#35/1 &c->lock irq_context: 0 &type->s_umount_key#35/1 &____s->seqcount irq_context: 0 &type->s_umount_key#35/1 sb_lock irq_context: 0 &type->s_umount_key#35/1 sb_lock unnamed_dev_ida.xa_lock irq_context: 0 &type->s_umount_key#35/1 &root->kernfs_rwsem irq_context: 0 &type->s_umount_key#35/1 &root->kernfs_rwsem inode_hash_lock irq_context: 0 &type->s_umount_key#35/1 &root->kernfs_rwsem fs_reclaim irq_context: 0 &type->s_umount_key#35/1 &root->kernfs_rwsem fs_reclaim mmu_notifier_invalidate_range_start irq_context: 0 &type->s_umount_key#35/1 &root->kernfs_rwsem pool_lock#2 irq_context: 0 &type->s_umount_key#35/1 &root->kernfs_rwsem mmu_notifier_invalidate_range_start irq_context: 0 &type->s_umount_key#35/1 &root->kernfs_rwsem inode_hash_lock &sb->s_type->i_lock_key#24 irq_context: 0 &type->s_umount_key#35/1 &root->kernfs_rwsem inode_hash_lock &s->s_inode_list_lock irq_context: 0 &type->s_umount_key#35/1 &root->kernfs_rwsem tk_core.seq.seqcount irq_context: 0 &type->s_umount_key#35/1 &root->kernfs_rwsem &sb->s_type->i_lock_key#24 irq_context: 0 &type->s_umount_key#35/1 &sb->s_type->i_lock_key#24 irq_context: 0 &type->s_umount_key#35/1 &sb->s_type->i_lock_key#24 &dentry->d_lock irq_context: 0 &type->s_umount_key#35/1 &root->kernfs_supers_rwsem irq_context: 0 &type->s_umount_key#35/1 &dentry->d_lock irq_context: 0 &type->i_mutex_dir_key#3 namespace_sem &c->lock irq_context: 0 &type->i_mutex_dir_key#3 namespace_sem &____s->seqcount irq_context: 0 &root->kernfs_iattr_rwsem irq_context: 0 &type->i_mutex_dir_key#4 irq_context: 0 &type->i_mutex_dir_key#4 fs_reclaim irq_context: 0 &type->i_mutex_dir_key#4 fs_reclaim mmu_notifier_invalidate_range_start irq_context: 0 &type->i_mutex_dir_key#4 pool_lock#2 irq_context: 0 &type->i_mutex_dir_key#4 &dentry->d_lock irq_context: 0 &type->i_mutex_dir_key#4 rcu_read_lock rename_lock.seqcount irq_context: 0 &type->i_mutex_dir_key#4 &root->kernfs_rwsem irq_context: 0 &type->i_mutex_dir_key#4 &root->kernfs_rwsem inode_hash_lock irq_context: 0 &type->i_mutex_dir_key#4 &root->kernfs_rwsem fs_reclaim irq_context: 0 &type->i_mutex_dir_key#4 &root->kernfs_rwsem fs_reclaim mmu_notifier_invalidate_range_start irq_context: 0 &type->i_mutex_dir_key#4 &root->kernfs_rwsem pool_lock#2 irq_context: 0 &type->i_mutex_dir_key#4 &root->kernfs_rwsem mmu_notifier_invalidate_range_start irq_context: 0 &type->i_mutex_dir_key#4 &root->kernfs_rwsem inode_hash_lock &sb->s_type->i_lock_key#24 irq_context: 0 &type->i_mutex_dir_key#4 &root->kernfs_rwsem inode_hash_lock &s->s_inode_list_lock irq_context: 0 &type->i_mutex_dir_key#4 &root->kernfs_rwsem tk_core.seq.seqcount irq_context: 0 &type->i_mutex_dir_key#4 &root->kernfs_rwsem &sb->s_type->i_lock_key#24 irq_context: 0 &type->i_mutex_dir_key#4 &sb->s_type->i_lock_key#24 irq_context: 0 &type->i_mutex_dir_key#4 &sb->s_type->i_lock_key#24 &dentry->d_lock irq_context: 0 &type->i_mutex_dir_key#4 &sb->s_type->i_lock_key#24 &dentry->d_lock &wq irq_context: 0 &type->i_mutex_dir_key#4 &____s->seqcount irq_context: 0 &type->i_mutex_dir_key#4 rcu_read_lock pool_lock#2 irq_context: 0 &type->i_mutex_dir_key#4 &obj_hash[i].lock irq_context: 0 &type->i_mutex_dir_key#4 &dentry->d_lock &wq irq_context: 0 &ent->pde_unload_lock irq_context: 0 &p->lock file_systems_lock irq_context: 0 namespace_sem mount_lock mount_lock.seqcount &new_ns->poll irq_context: 0 namespace_sem mount_lock mount_lock.seqcount &dentry->d_lock irq_context: 0 namespace_sem mount_lock mount_lock.seqcount rcu_read_lock &dentry->d_lock irq_context: 0 namespace_sem mount_lock mount_lock.seqcount &obj_hash[i].lock irq_context: 0 namespace_sem mount_lock mount_lock.seqcount pool_lock#2 irq_context: 0 &x->wait#25 irq_context: 0 &mm->mmap_lock resource_lock irq_context: 0 &net->unx.table.locks[i] irq_context: 0 &sb->s_type->i_lock_key#8 &dentry->d_lock irq_context: 0 &sb->s_type->i_mutex_key#10 irq_context: 0 &sb->s_type->i_mutex_key#10 &net->unx.table.locks[i] irq_context: 0 &sb->s_type->i_mutex_key#10 &u->lock irq_context: 0 &sb->s_type->i_mutex_key#10 &u->lock clock-AF_UNIX irq_context: 0 &sb->s_type->i_mutex_key#10 &u->peer_wait irq_context: 0 &sb->s_type->i_mutex_key#10 rlock-AF_UNIX irq_context: 0 &sb->s_type->i_mutex_key#10 pool_lock#2 irq_context: 0 &sb->s_type->i_mutex_key#10 &dir->lock irq_context: 0 &sb->s_type->i_mutex_key#10 &obj_hash[i].lock irq_context: 0 &sig->cred_guard_mutex &type->i_mutex_dir_key#3 &ei->i_data_sem &ei->i_es_lock &____s->seqcount irq_context: 0 &sig->cred_guard_mutex &type->i_mutex_dir_key#3 &ei->i_data_sem &ei->i_es_lock &c->lock irq_context: 0 &sig->cred_guard_mutex &type->i_mutex_dir_key#3 &ei->i_data_sem &ei->i_es_lock rcu_read_lock pool_lock#2 irq_context: 0 &sig->cred_guard_mutex mapping.invalidate_lock rcu_read_lock &c->lock irq_context: 0 &sig->cred_guard_mutex mapping.invalidate_lock rcu_read_lock &____s->seqcount irq_context: 0 &sig->cred_guard_mutex mapping.invalidate_lock rcu_read_lock pool_lock#2 irq_context: 0 &sig->cred_guard_mutex tomoyo_ss rcu_read_lock &____s->seqcount#3 irq_context: 0 &sig->cred_guard_mutex &p->alloc_lock irq_context: 0 &sig->cred_guard_mutex &p->alloc_lock &x->wait#25 irq_context: 0 &sig->cred_guard_mutex &p->alloc_lock &x->wait#25 &p->pi_lock irq_context: 0 &sig->cred_guard_mutex &p->alloc_lock &x->wait#25 &p->pi_lock &rq->__lock irq_context: 0 &sig->cred_guard_mutex &p->alloc_lock &x->wait#25 &p->pi_lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 &sig->cred_guard_mutex &sig->exec_update_lock &mm->mmap_lock irq_context: 0 &sig->cred_guard_mutex &sig->exec_update_lock &mm->mmap_lock &p->alloc_lock irq_context: 0 &sig->cred_guard_mutex &sig->exec_update_lock &mm->mmap_lock &p->alloc_lock &memcg->mm_list.lock irq_context: 0 &mm->mmap_lock &mapping->i_mmap_rwsem &obj_hash[i].lock pool_lock irq_context: 0 &mm->mmap_lock fs_reclaim mmu_notifier_invalidate_range_start &rq->__lock irq_context: 0 &mm->mmap_lock fs_reclaim mmu_notifier_invalidate_range_start &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 &sig->wait_chldexit irq_context: 0 tasklist_lock irq_context: 0 &mm->mmap_lock ptlock_ptr(page)#2 key irq_context: 0 cgroup_threadgroup_rwsem &rq->__lock irq_context: 0 &tsk->futex_exit_mutex &mm->mmap_lock irq_context: 0 &mm->mmap_lock &p->alloc_lock irq_context: 0 &mm->mmap_lock lock#4 irq_context: 0 &mm->mmap_lock lock#4 &lruvec->lru_lock irq_context: 0 &mm->mmap_lock lock#5 irq_context: 0 &mm->mmap_lock ptlock_ptr(page)#2 &____s->seqcount irq_context: 0 &mm->mmap_lock ptlock_ptr(page)#2 pool_lock#2 irq_context: 0 &mm->mmap_lock &lruvec->lru_lock irq_context: 0 &memcg->mm_list.lock irq_context: 0 tasklist_lock &sighand->siglock &____s->seqcount irq_context: 0 tasklist_lock &sighand->siglock pool_lock#2 irq_context: 0 tasklist_lock &sighand->siglock &c->lock irq_context: 0 rcu_read_lock &____s->seqcount#5 irq_context: 0 &prev->lock irq_context: 0 &sighand->siglock &(&sig->stats_lock)->lock irq_context: 0 &sighand->siglock &(&sig->stats_lock)->lock &____s->seqcount#5 irq_context: 0 &type->i_mutex_dir_key#3 lock#4 rcu_read_lock pool_lock#2 irq_context: 0 &type->i_mutex_dir_key#3 lock#4 &obj_hash[i].lock irq_context: 0 &type->i_mutex_dir_key#3 rcu_read_lock &rq->__lock irq_context: 0 &type->i_mutex_dir_key#3 rcu_read_lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: softirq bit_wait_table + i &p->pi_lock &rq->__lock &cfs_rq->removed.lock irq_context: 0 &type->i_mutex_dir_key#3 pgd_lock irq_context: 0 &type->i_mutex_dir_key#3 key irq_context: 0 &type->i_mutex_dir_key#3 pcpu_lock irq_context: 0 &type->i_mutex_dir_key#3 percpu_counters_lock irq_context: 0 &type->i_mutex_dir_key#3 &ei->i_es_lock key#2 irq_context: 0 sb_writers#4 irq_context: 0 sb_writers#4 mount_lock irq_context: 0 &type->i_mutex_dir_key#3 rename_lock.seqcount irq_context: 0 &type->i_mutex_dir_key#3 rcu_read_lock &dentry->d_lock irq_context: 0 &type->i_mutex_dir_key#3 &dentry->d_lock irq_context: 0 &sig->cred_guard_mutex &sb->s_type->i_mutex_key#8 integrity_iint_lock irq_context: 0 &sig->cred_guard_mutex &ei->xattr_sem &mapping->private_lock irq_context: 0 &mm->mmap_lock &mapping->i_mmap_rwsem quarantine_lock irq_context: 0 &mm->mmap_lock &obj_hash[i].lock pool_lock irq_context: 0 sb_writers#3 tk_core.seq.seqcount irq_context: 0 sb_writers#3 &sb->s_type->i_lock_key#23 irq_context: 0 sb_writers#3 &wb->list_lock irq_context: 0 sb_writers#3 &wb->list_lock &sb->s_type->i_lock_key#23 irq_context: 0 &sb->s_type->i_mutex_key#9 &p->alloc_lock irq_context: 0 &sb->s_type->i_mutex_key#9 &pid->lock irq_context: 0 &sb->s_type->i_mutex_key#9 &c->lock irq_context: 0 &sb->s_type->i_mutex_key#9 &____s->seqcount irq_context: 0 &p->alloc_lock &fs->lock &dentry->d_lock irq_context: 0 &p->lock namespace_sem irq_context: 0 &p->lock namespace_sem &new_ns->ns_lock irq_context: 0 &p->lock namespace_sem rcu_read_lock mount_lock.seqcount irq_context: 0 &p->lock namespace_sem rcu_read_lock rcu_read_lock rename_lock.seqcount irq_context: 0 &type->s_umount_key#32 &lru->node[i].lock irq_context: 0 &type->s_umount_key#32 &sb->s_type->i_lock_key#22 &lru->node[i].lock irq_context: 0 &type->s_umount_key#32 &obj_hash[i].lock pool_lock irq_context: 0 &type->s_umount_key#32 (work_completion)(&sbi->s_error_work) irq_context: 0 &type->s_umount_key#32 key#3 irq_context: 0 &type->s_umount_key#32 key#4 irq_context: 0 &type->s_umount_key#32 &sbi->s_error_lock irq_context: 0 &type->s_umount_key#32 &base->lock irq_context: 0 &type->s_umount_key#32 &base->lock &obj_hash[i].lock irq_context: 0 &type->s_umount_key#32 &fq->mq_flush_lock irq_context: softirq &fq->mq_flush_lock irq_context: softirq &fq->mq_flush_lock tk_core.seq.seqcount irq_context: softirq &fq->mq_flush_lock &q->requeue_lock irq_context: softirq &fq->mq_flush_lock &obj_hash[i].lock irq_context: softirq &fq->mq_flush_lock rcu_read_lock &pool->lock irq_context: softirq &fq->mq_flush_lock rcu_read_lock &pool->lock &obj_hash[i].lock irq_context: softirq &fq->mq_flush_lock rcu_read_lock &pool->lock pool_lock#2 irq_context: softirq &fq->mq_flush_lock rcu_read_lock &pool->lock &p->pi_lock irq_context: softirq &fq->mq_flush_lock rcu_read_lock &pool->lock &p->pi_lock &rq->__lock irq_context: softirq &fq->mq_flush_lock rcu_read_lock &pool->lock &p->pi_lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)kblockd (work_completion)(&(&q->requeue_work)->work) irq_context: 0 (wq_completion)kblockd (work_completion)(&(&q->requeue_work)->work) &q->requeue_lock irq_context: 0 (wq_completion)kblockd (work_completion)(&(&q->requeue_work)->work) &hctx->lock irq_context: 0 (wq_completion)kblockd (work_completion)(&(&q->requeue_work)->work) rcu_read_lock &hctx->lock irq_context: 0 (wq_completion)kblockd (work_completion)(&(&q->requeue_work)->work) rcu_read_lock &obj_hash[i].lock irq_context: 0 (wq_completion)kblockd (work_completion)(&(&q->requeue_work)->work) rcu_read_lock tk_core.seq.seqcount irq_context: 0 (wq_completion)kblockd (work_completion)(&(&q->requeue_work)->work) rcu_read_lock &virtscsi_vq->vq_lock irq_context: softirq &fq->mq_flush_lock bit_wait_table + i irq_context: softirq &fq->mq_flush_lock bit_wait_table + i &p->pi_lock irq_context: softirq &fq->mq_flush_lock bit_wait_table + i &p->pi_lock &rq->__lock irq_context: softirq &fq->mq_flush_lock bit_wait_table + i &p->pi_lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: softirq &fq->mq_flush_lock bit_wait_table + i &p->pi_lock &cfs_rq->removed.lock irq_context: 0 &type->s_umount_key#32 ext4_li_mtx fs_reclaim irq_context: 0 &type->s_umount_key#32 ext4_li_mtx fs_reclaim mmu_notifier_invalidate_range_start irq_context: 0 &type->s_umount_key#32 ext4_li_mtx pool_lock#2 irq_context: 0 &type->s_umount_key#32 ext4_li_mtx batched_entropy_u16.lock irq_context: 0 &type->s_umount_key#32 ext4_li_mtx &eli->li_list_mtx irq_context: 0 &type->s_umount_key#32 ext4_li_mtx kthread_create_lock irq_context: 0 &type->s_umount_key#32 ext4_li_mtx &p->pi_lock irq_context: 0 &type->s_umount_key#32 ext4_li_mtx &p->pi_lock &rq->__lock irq_context: 0 &type->s_umount_key#32 ext4_li_mtx &p->pi_lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 &type->s_umount_key#32 ext4_li_mtx &x->wait irq_context: 0 &type->s_umount_key#32 ext4_li_mtx &rq->__lock irq_context: 0 &type->s_umount_key#32 ext4_li_mtx &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 &type->s_umount_key#32 ext4_li_mtx &obj_hash[i].lock irq_context: 0 &eli->li_list_mtx irq_context: 0 &type->s_umount_key#32 mount_lock irq_context: 0 &type->s_umount_key#32 mount_lock mount_lock.seqcount irq_context: 0 &type->s_umount_key#32 mount_lock mount_lock.seqcount &new_ns->poll irq_context: 0 namespace_sem irq_context: 0 namespace_sem &new_ns->ns_lock irq_context: 0 rcu_read_lock &pid->lock irq_context: 0 &sb->s_type->i_lock_key#23 &dentry->d_lock irq_context: 0 rename_lock.seqcount irq_context: 0 &dentry->d_lock &dentry->d_lock/1 irq_context: 0 &dentry->d_lock &dentry->d_lock/1 &lru->node[i].lock irq_context: 0 &pid->lock irq_context: 0 sb_writers#4 tk_core.seq.seqcount irq_context: 0 sb_writers#4 mmu_notifier_invalidate_range_start irq_context: 0 sb_writers#4 &____s->seqcount irq_context: 0 sb_writers#4 pool_lock#2 irq_context: 0 sb_writers#4 &c->lock irq_context: 0 sb_writers#4 &journal->j_state_lock irq_context: 0 sb_writers#4 &journal->j_state_lock irq_context: 0 sb_writers#4 &journal->j_state_lock tk_core.seq.seqcount irq_context: 0 sb_writers#4 &journal->j_state_lock &obj_hash[i].lock irq_context: 0 sb_writers#4 &journal->j_state_lock &base->lock irq_context: 0 sb_writers#4 &journal->j_state_lock &base->lock &obj_hash[i].lock irq_context: 0 sb_writers#4 jbd2_handle irq_context: 0 sb_writers#4 jbd2_handle mmu_notifier_invalidate_range_start irq_context: 0 sb_writers#4 jbd2_handle &____s->seqcount irq_context: 0 sb_writers#4 jbd2_handle pool_lock#2 irq_context: 0 sb_writers#4 jbd2_handle &c->lock irq_context: 0 sb_writers#4 jbd2_handle &ret->b_state_lock irq_context: 0 sb_writers#4 jbd2_handle &ret->b_state_lock &journal->j_list_lock irq_context: 0 sb_writers#4 jbd2_handle &journal->j_revoke_lock irq_context: 0 sb_writers#4 jbd2_handle &ei->i_raw_lock irq_context: 0 sb_writers#4 jbd2_handle &journal->j_wait_updates irq_context: 0 sb_writers#4 &obj_hash[i].lock irq_context: 0 sb_writers#4 &sb->s_type->i_lock_key#22 irq_context: 0 sb_writers#4 &wb->list_lock irq_context: 0 sb_writers#4 &wb->list_lock &sb->s_type->i_lock_key#22 irq_context: 0 sb_writers#4 &wb->work_lock irq_context: 0 sb_writers#4 &wb->work_lock &obj_hash[i].lock irq_context: 0 sb_writers#4 &wb->work_lock &base->lock irq_context: 0 sb_writers#4 &wb->work_lock &base->lock &obj_hash[i].lock irq_context: 0 rcu_read_lock tk_core.seq.seqcount irq_context: 0 sb_writers#4 &type->i_mutex_dir_key#3 irq_context: 0 sb_writers#4 &type->i_mutex_dir_key#3 rename_lock.seqcount irq_context: 0 sb_writers#4 &type->i_mutex_dir_key#3 rcu_read_lock &dentry->d_lock irq_context: 0 sb_writers#4 &type->i_mutex_dir_key#3 tomoyo_ss irq_context: 0 sb_writers#4 &type->i_mutex_dir_key#3 tomoyo_ss mmu_notifier_invalidate_range_start irq_context: 0 sb_writers#4 &type->i_mutex_dir_key#3 tomoyo_ss pool_lock#2 irq_context: 0 sb_writers#4 &type->i_mutex_dir_key#3 tomoyo_ss rcu_read_lock mount_lock.seqcount irq_context: 0 sb_writers#4 &type->i_mutex_dir_key#3 tomoyo_ss rcu_read_lock rcu_read_lock rename_lock.seqcount irq_context: 0 sb_writers#4 &type->i_mutex_dir_key#3 tomoyo_ss &obj_hash[i].lock irq_context: 0 sb_writers#4 &type->i_mutex_dir_key#3 tomoyo_ss &c->lock irq_context: 0 sb_writers#4 &type->i_mutex_dir_key#3 tomoyo_ss &____s->seqcount irq_context: 0 sb_writers#4 &type->i_mutex_dir_key#3 tomoyo_ss rcu_read_lock &dentry->d_lock irq_context: 0 sb_writers#4 &type->i_mutex_dir_key#3 tomoyo_ss rcu_read_lock pool_lock#2 irq_context: 0 sb_writers#4 &type->i_mutex_dir_key#3 tomoyo_ss tomoyo_policy_lock irq_context: 0 sb_writers#4 &type->i_mutex_dir_key#3 tomoyo_ss tomoyo_policy_lock mmu_notifier_invalidate_range_start irq_context: 0 sb_writers#4 &type->i_mutex_dir_key#3 tomoyo_ss tomoyo_policy_lock pool_lock#2 irq_context: 0 sb_writers#4 &type->i_mutex_dir_key#3 mmu_notifier_invalidate_range_start irq_context: 0 sb_writers#4 &type->i_mutex_dir_key#3 pool_lock#2 irq_context: 0 sb_writers#4 &type->i_mutex_dir_key#3 &obj_hash[i].lock irq_context: 0 sb_writers#4 &type->i_mutex_dir_key#3 &sb->s_type->i_lock_key#22 irq_context: 0 sb_writers#4 &type->i_mutex_dir_key#3 &s->s_inode_list_lock irq_context: 0 sb_writers#4 &type->i_mutex_dir_key#3 &ei->xattr_sem irq_context: 0 sb_writers#4 &type->i_mutex_dir_key#3 &____s->seqcount irq_context: 0 sb_writers#4 &type->i_mutex_dir_key#3 &xa->xa_lock#9 irq_context: 0 sb_writers#4 &type->i_mutex_dir_key#3 lock#4 irq_context: 0 sb_writers#4 &type->i_mutex_dir_key#3 &mapping->private_lock irq_context: 0 sb_writers#4 &type->i_mutex_dir_key#3 tk_core.seq.seqcount irq_context: 0 sb_writers#4 &type->i_mutex_dir_key#3 &dd->lock irq_context: 0 sb_writers#4 &type->i_mutex_dir_key#3 rcu_read_lock &pool->lock irq_context: 0 sb_writers#4 &type->i_mutex_dir_key#3 rcu_read_lock &pool->lock &obj_hash[i].lock irq_context: 0 sb_writers#4 &type->i_mutex_dir_key#3 rcu_read_lock &pool->lock pool_lock#2 irq_context: 0 sb_writers#4 &type->i_mutex_dir_key#3 rcu_read_lock &pool->lock &p->pi_lock irq_context: 0 sb_writers#4 &type->i_mutex_dir_key#3 rcu_read_lock &pool->lock &p->pi_lock &rq->__lock irq_context: 0 sb_writers#4 &type->i_mutex_dir_key#3 rcu_read_lock &pool->lock &p->pi_lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 sb_writers#4 &type->i_mutex_dir_key#3 bit_wait_table + i irq_context: 0 sb_writers#4 &type->i_mutex_dir_key#3 &rq->__lock irq_context: 0 sb_writers#4 &type->i_mutex_dir_key#3 &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 sb_writers#4 &type->i_mutex_dir_key#3 &journal->j_state_lock irq_context: 0 sb_writers#4 &type->i_mutex_dir_key#3 jbd2_handle irq_context: 0 sb_writers#4 &type->i_mutex_dir_key#3 jbd2_handle mmu_notifier_invalidate_range_start irq_context: 0 sb_writers#4 &type->i_mutex_dir_key#3 jbd2_handle pool_lock#2 irq_context: 0 sb_writers#4 &type->i_mutex_dir_key#3 jbd2_handle &ret->b_state_lock irq_context: 0 sb_writers#4 &type->i_mutex_dir_key#3 jbd2_handle &ret->b_state_lock &journal->j_list_lock irq_context: 0 sb_writers#4 &type->i_mutex_dir_key#3 jbd2_handle &journal->j_revoke_lock irq_context: 0 sb_writers#4 &type->i_mutex_dir_key#3 jbd2_handle &mapping->private_lock irq_context: 0 sb_writers#4 &type->i_mutex_dir_key#3 jbd2_handle &meta_group_info[i]->alloc_sem irq_context: 0 sb_writers#4 &type->i_mutex_dir_key#3 jbd2_handle tk_core.seq.seqcount irq_context: 0 sb_writers#4 &type->i_mutex_dir_key#3 jbd2_handle inode_hash_lock irq_context: 0 sb_writers#4 &type->i_mutex_dir_key#3 jbd2_handle inode_hash_lock &sb->s_type->i_lock_key#22 irq_context: 0 sb_writers#4 &type->i_mutex_dir_key#3 jbd2_handle batched_entropy_u32.lock irq_context: 0 sb_writers#4 &type->i_mutex_dir_key#3 jbd2_handle &ei->i_raw_lock irq_context: 0 sb_writers#4 &type->i_mutex_dir_key#3 jbd2_handle &ei->i_es_lock irq_context: 0 sb_writers#4 &type->i_mutex_dir_key#3 jbd2_handle &sb->s_type->i_lock_key#22 irq_context: 0 sb_writers#4 &type->i_mutex_dir_key#3 jbd2_handle &sb->s_type->i_lock_key#22 &dentry->d_lock irq_context: 0 sb_writers#4 &type->i_mutex_dir_key#3 jbd2_handle &journal->j_wait_updates irq_context: 0 sb_internal rcu_read_lock init_fs.seq.seqcount irq_context: 0 sb_internal rcu_read_lock rcu_read_lock mount_lock.seqcount irq_context: 0 sb_internal rcu_read_lock rcu_read_lock rcu_read_lock rename_lock.seqcount irq_context: 0 sb_internal mmu_notifier_invalidate_range_start irq_context: 0 sb_internal pool_lock#2 irq_context: 0 sb_internal &journal->j_state_lock irq_context: 0 sb_internal jbd2_handle irq_context: 0 sb_internal jbd2_handle mmu_notifier_invalidate_range_start irq_context: 0 sb_internal jbd2_handle pool_lock#2 irq_context: 0 sb_internal jbd2_handle &ret->b_state_lock irq_context: 0 sb_internal jbd2_handle &ret->b_state_lock &journal->j_list_lock irq_context: 0 sb_internal jbd2_handle &journal->j_revoke_lock irq_context: 0 sb_internal jbd2_handle &mapping->private_lock irq_context: 0 sb_internal jbd2_handle &journal->j_wait_updates irq_context: 0 sb_internal &obj_hash[i].lock irq_context: 0 &ei->i_data_sem irq_context: 0 &ei->i_data_sem &ei->i_prealloc_lock irq_context: 0 &sighand->siglock hrtimer_bases.lock irq_context: 0 &sighand->siglock hrtimer_bases.lock tk_core.seq.seqcount irq_context: 0 &sighand->siglock hrtimer_bases.lock &obj_hash[i].lock irq_context: 0 file_rwsem irq_context: 0 file_rwsem &ctx->flc_lock irq_context: 0 file_rwsem &ctx->flc_lock &fll->lock irq_context: 0 &ctx->flc_lock irq_context: softirq &folio_wait_table[i] &p->pi_lock &cfs_rq->removed.lock irq_context: 0 &sig->cred_guard_mutex tk_core.seq.seqcount irq_context: 0 &sig->cred_guard_mutex sb_writers#4 mount_lock irq_context: 0 &sig->cred_guard_mutex sb_writers#4 tk_core.seq.seqcount irq_context: 0 &sig->cred_guard_mutex sb_writers#4 mmu_notifier_invalidate_range_start irq_context: 0 &sig->cred_guard_mutex sb_writers#4 pool_lock#2 irq_context: 0 &sig->cred_guard_mutex sb_writers#4 &journal->j_state_lock irq_context: 0 &sig->cred_guard_mutex sb_writers#4 jbd2_handle irq_context: 0 &sig->cred_guard_mutex sb_writers#4 jbd2_handle &mapping->private_lock irq_context: 0 &sig->cred_guard_mutex sb_writers#4 jbd2_handle mmu_notifier_invalidate_range_start irq_context: 0 &sig->cred_guard_mutex sb_writers#4 jbd2_handle pool_lock#2 irq_context: 0 &sig->cred_guard_mutex sb_writers#4 jbd2_handle &ret->b_state_lock irq_context: 0 &sig->cred_guard_mutex sb_writers#4 jbd2_handle &ret->b_state_lock &journal->j_list_lock irq_context: 0 &sig->cred_guard_mutex sb_writers#4 jbd2_handle &journal->j_revoke_lock irq_context: 0 &sig->cred_guard_mutex sb_writers#4 jbd2_handle &ei->i_raw_lock irq_context: 0 &sig->cred_guard_mutex sb_writers#4 jbd2_handle &journal->j_wait_updates irq_context: 0 &sig->cred_guard_mutex sb_writers#4 &obj_hash[i].lock irq_context: 0 &sig->cred_guard_mutex sb_writers#4 &sb->s_type->i_lock_key#22 irq_context: 0 &sig->cred_guard_mutex sb_writers#4 &wb->list_lock irq_context: 0 &sig->cred_guard_mutex sb_writers#4 &wb->list_lock &sb->s_type->i_lock_key#22 irq_context: 0 &sig->cred_guard_mutex tomoyo_ss quarantine_lock irq_context: 0 &sig->cred_guard_mutex &iint->mutex &c->lock irq_context: 0 &sig->cred_guard_mutex &iint->mutex &____s->seqcount irq_context: 0 &sig->cred_guard_mutex &iint->mutex tk_core.seq.seqcount irq_context: 0 &mm->mmap_lock tk_core.seq.seqcount irq_context: 0 sb_writers#4 jbd2_handle &mapping->private_lock irq_context: 0 &p->lock &c->lock irq_context: 0 &p->lock &____s->seqcount irq_context: 0 sb_writers#4 &type->i_mutex_dir_key#3/1 irq_context: 0 sb_writers#4 &type->i_mutex_dir_key#3/1 rename_lock.seqcount irq_context: 0 sb_writers#4 &type->i_mutex_dir_key#3/1 rcu_read_lock &dentry->d_lock irq_context: 0 sb_writers &type->i_mutex_dir_key/1 tomoyo_ss irq_context: 0 sb_writers &type->i_mutex_dir_key/1 tomoyo_ss mmu_notifier_invalidate_range_start irq_context: 0 sb_writers &type->i_mutex_dir_key/1 tomoyo_ss pool_lock#2 irq_context: 0 sb_writers &type->i_mutex_dir_key/1 tomoyo_ss rcu_read_lock mount_lock.seqcount irq_context: 0 sb_writers &type->i_mutex_dir_key/1 tomoyo_ss rcu_read_lock rcu_read_lock rename_lock.seqcount irq_context: 0 sb_writers &type->i_mutex_dir_key/1 tomoyo_ss &obj_hash[i].lock irq_context: 0 sb_writers &type->i_mutex_dir_key/1 tomoyo_ss rcu_read_lock &dentry->d_lock irq_context: 0 sb_writers &type->i_mutex_dir_key/1 tomoyo_ss tomoyo_policy_lock irq_context: 0 sb_writers &type->i_mutex_dir_key/1 tomoyo_ss tomoyo_policy_lock mmu_notifier_invalidate_range_start irq_context: 0 sb_writers &type->i_mutex_dir_key/1 tomoyo_ss tomoyo_policy_lock pool_lock#2 irq_context: 0 &sig->cred_guard_mutex sb_writers#4 &____s->seqcount irq_context: 0 &sig->cred_guard_mutex sb_writers#4 &c->lock irq_context: 0 &sig->cred_guard_mutex sb_writers#4 jbd2_handle &____s->seqcount irq_context: 0 &sig->cred_guard_mutex sb_writers#4 jbd2_handle &c->lock irq_context: 0 &sig->cred_guard_mutex rcu_read_lock tk_core.seq.seqcount irq_context: 0 &mm->mmap_lock lock#4 rcu_read_lock pool_lock#2 irq_context: 0 &mm->mmap_lock lock#4 &obj_hash[i].lock irq_context: 0 tomoyo_ss quarantine_lock irq_context: 0 &sb->s_type->i_mutex_key#8 &____s->seqcount irq_context: 0 &sb->s_type->i_mutex_key#8 &c->lock irq_context: 0 &iint->mutex tk_core.seq.seqcount irq_context: 0 &iint->mutex sb_writers#4 mount_lock irq_context: 0 &iint->mutex sb_writers#4 tk_core.seq.seqcount irq_context: 0 &iint->mutex sb_writers#4 mmu_notifier_invalidate_range_start irq_context: 0 &iint->mutex sb_writers#4 pool_lock#2 irq_context: 0 &iint->mutex sb_writers#4 &journal->j_state_lock irq_context: 0 &iint->mutex sb_writers#4 jbd2_handle irq_context: 0 &iint->mutex sb_writers#4 jbd2_handle &mapping->private_lock irq_context: 0 &iint->mutex sb_writers#4 jbd2_handle mmu_notifier_invalidate_range_start irq_context: 0 &iint->mutex sb_writers#4 jbd2_handle &____s->seqcount irq_context: 0 &iint->mutex sb_writers#4 jbd2_handle pool_lock#2 irq_context: 0 &iint->mutex sb_writers#4 jbd2_handle rcu_read_lock pool_lock#2 irq_context: 0 &iint->mutex sb_writers#4 jbd2_handle &obj_hash[i].lock irq_context: 0 &iint->mutex sb_writers#4 jbd2_handle &ret->b_state_lock irq_context: 0 &iint->mutex sb_writers#4 jbd2_handle &ret->b_state_lock &journal->j_list_lock irq_context: 0 &iint->mutex sb_writers#4 jbd2_handle &journal->j_revoke_lock irq_context: 0 &iint->mutex sb_writers#4 jbd2_handle &ei->i_raw_lock irq_context: 0 &iint->mutex sb_writers#4 jbd2_handle &journal->j_wait_updates irq_context: 0 &iint->mutex sb_writers#4 &obj_hash[i].lock irq_context: 0 &iint->mutex sb_writers#4 &sb->s_type->i_lock_key#22 irq_context: 0 &iint->mutex sb_writers#4 &wb->list_lock irq_context: 0 &iint->mutex sb_writers#4 &wb->list_lock &sb->s_type->i_lock_key#22 irq_context: 0 &iint->mutex quarantine_lock irq_context: 0 rcu_read_lock &p->alloc_lock irq_context: 0 &type->s_umount_key#36/1 irq_context: 0 &type->s_umount_key#36/1 fs_reclaim irq_context: 0 &type->s_umount_key#36/1 fs_reclaim mmu_notifier_invalidate_range_start irq_context: 0 &type->s_umount_key#36/1 pool_lock#2 irq_context: 0 &type->s_umount_key#36/1 pcpu_alloc_mutex irq_context: 0 &type->s_umount_key#36/1 pcpu_alloc_mutex pcpu_lock irq_context: 0 &type->s_umount_key#36/1 shrinker_rwsem irq_context: 0 &type->s_umount_key#36/1 list_lrus_mutex irq_context: 0 &type->s_umount_key#36/1 sb_lock irq_context: 0 &type->s_umount_key#36/1 sb_lock unnamed_dev_ida.xa_lock irq_context: 0 &type->s_umount_key#36/1 mmu_notifier_invalidate_range_start irq_context: 0 &type->s_umount_key#36/1 &sb->s_type->i_lock_key#25 irq_context: 0 &type->s_umount_key#36/1 &s->s_inode_list_lock irq_context: 0 &type->s_umount_key#36/1 tk_core.seq.seqcount irq_context: 0 &type->s_umount_key#36/1 &sb->s_type->i_lock_key#25 &dentry->d_lock irq_context: 0 &type->s_umount_key#36/1 &sb->s_type->i_mutex_key#11 irq_context: 0 &type->s_umount_key#36/1 &sb->s_type->i_mutex_key#11 fs_reclaim irq_context: 0 &type->s_umount_key#36/1 &sb->s_type->i_mutex_key#11 fs_reclaim mmu_notifier_invalidate_range_start irq_context: 0 &type->s_umount_key#36/1 &sb->s_type->i_mutex_key#11 pool_lock#2 irq_context: 0 &type->s_umount_key#36/1 &sb->s_type->i_mutex_key#11 &dentry->d_lock irq_context: 0 &type->s_umount_key#36/1 &sb->s_type->i_mutex_key#11 mmu_notifier_invalidate_range_start irq_context: 0 &type->s_umount_key#36/1 &sb->s_type->i_mutex_key#11 &sb->s_type->i_lock_key#25 irq_context: 0 &type->s_umount_key#36/1 &sb->s_type->i_mutex_key#11 &s->s_inode_list_lock irq_context: 0 &type->s_umount_key#36/1 &sb->s_type->i_mutex_key#11 tk_core.seq.seqcount irq_context: 0 &type->s_umount_key#36/1 &sb->s_type->i_mutex_key#11 &sb->s_type->i_lock_key#25 &dentry->d_lock irq_context: 0 &type->s_umount_key#36/1 &dentry->d_lock irq_context: 0 &type->i_mutex_dir_key#2 irq_context: 0 &type->i_mutex_dir_key#2 namespace_sem irq_context: 0 &type->i_mutex_dir_key#2 namespace_sem rcu_read_lock mount_lock.seqcount irq_context: 0 &type->i_mutex_dir_key#2 namespace_sem fs_reclaim irq_context: 0 &type->i_mutex_dir_key#2 namespace_sem fs_reclaim mmu_notifier_invalidate_range_start irq_context: 0 &type->i_mutex_dir_key#2 namespace_sem rename_lock irq_context: 0 &type->i_mutex_dir_key#2 namespace_sem rename_lock rename_lock.seqcount irq_context: 0 &type->i_mutex_dir_key#2 namespace_sem rename_lock rename_lock.seqcount &dentry->d_lock irq_context: 0 &type->i_mutex_dir_key#2 namespace_sem mount_lock irq_context: 0 &type->i_mutex_dir_key#2 namespace_sem mount_lock &dentry->d_lock irq_context: 0 &type->i_mutex_dir_key#2 namespace_sem mount_lock mount_lock.seqcount irq_context: 0 &type->i_mutex_dir_key#2 namespace_sem mount_lock mount_lock.seqcount &new_ns->poll irq_context: 0 &type->i_mutex_dir_key#2 namespace_sem mount_lock mount_lock.seqcount &dentry->d_lock irq_context: 0 &type->i_mutex_dir_key#2 namespace_sem mount_lock mount_lock.seqcount rcu_read_lock &dentry->d_lock irq_context: 0 &type->i_mutex_dir_key#2 namespace_sem mount_lock mount_lock.seqcount &obj_hash[i].lock irq_context: 0 &type->s_umount_key/1 fs_reclaim irq_context: 0 &type->s_umount_key/1 fs_reclaim mmu_notifier_invalidate_range_start irq_context: 0 &type->s_umount_key/1 mmu_notifier_invalidate_range_start irq_context: 0 &type->s_umount_key#37 irq_context: 0 &type->s_umount_key#37 sb_lock irq_context: 0 &type->s_umount_key#37 fs_reclaim irq_context: 0 &type->s_umount_key#37 fs_reclaim mmu_notifier_invalidate_range_start irq_context: 0 &type->s_umount_key#37 pool_lock#2 irq_context: 0 &type->s_umount_key#37 &dentry->d_lock irq_context: 0 &type->s_umount_key#37 &lru->node[i].lock irq_context: 0 &type->s_umount_key#37 rcu_read_lock &dentry->d_lock irq_context: 0 &type->s_umount_key#37 &obj_hash[i].lock irq_context: 0 &type->i_mutex_dir_key#4 irq_context: 0 &type->i_mutex_dir_key#4 namespace_sem irq_context: 0 &type->i_mutex_dir_key#4 namespace_sem rcu_read_lock mount_lock.seqcount irq_context: 0 &type->i_mutex_dir_key#4 namespace_sem fs_reclaim irq_context: 0 &type->i_mutex_dir_key#4 namespace_sem fs_reclaim mmu_notifier_invalidate_range_start irq_context: 0 &type->i_mutex_dir_key#4 namespace_sem rename_lock irq_context: 0 &type->i_mutex_dir_key#4 namespace_sem rename_lock rename_lock.seqcount irq_context: 0 &type->i_mutex_dir_key#4 namespace_sem rename_lock rename_lock.seqcount &dentry->d_lock irq_context: 0 &type->i_mutex_dir_key#4 namespace_sem mount_lock irq_context: 0 &type->i_mutex_dir_key#4 namespace_sem mount_lock &dentry->d_lock irq_context: 0 &type->i_mutex_dir_key#4 namespace_sem mount_lock mount_lock.seqcount irq_context: 0 &type->i_mutex_dir_key#4 namespace_sem mount_lock mount_lock.seqcount &new_ns->poll irq_context: 0 &type->i_mutex_dir_key#4 namespace_sem mount_lock mount_lock.seqcount &dentry->d_lock irq_context: 0 &type->i_mutex_dir_key#4 namespace_sem mount_lock mount_lock.seqcount rcu_read_lock &dentry->d_lock irq_context: 0 &type->i_mutex_dir_key#4 namespace_sem mount_lock mount_lock.seqcount &obj_hash[i].lock irq_context: 0 &p->lock &pcp->lock &zone->lock irq_context: 0 &p->lock &pcp->lock &zone->lock &____s->seqcount irq_context: 0 &type->s_umount_key#38 irq_context: 0 &type->s_umount_key#38 sb_lock irq_context: 0 &type->s_umount_key#38 &dentry->d_lock irq_context: 0 rcu_read_lock &dentry->d_lock sysctl_lock irq_context: 0 &type->s_umount_key#39/1 irq_context: 0 &type->s_umount_key#39/1 fs_reclaim irq_context: 0 &type->s_umount_key#39/1 fs_reclaim mmu_notifier_invalidate_range_start irq_context: 0 &type->s_umount_key#39/1 pool_lock#2 irq_context: 0 &type->s_umount_key#39/1 pcpu_alloc_mutex irq_context: 0 &type->s_umount_key#39/1 pcpu_alloc_mutex pcpu_lock irq_context: 0 &type->s_umount_key#39/1 shrinker_rwsem irq_context: 0 &type->s_umount_key#39/1 list_lrus_mutex irq_context: 0 &type->s_umount_key#39/1 sb_lock irq_context: 0 &type->s_umount_key#39/1 sb_lock unnamed_dev_ida.xa_lock irq_context: 0 &type->s_umount_key#39/1 mmu_notifier_invalidate_range_start irq_context: 0 &type->s_umount_key#39/1 &c->lock irq_context: 0 &type->s_umount_key#39/1 &____s->seqcount irq_context: 0 &type->s_umount_key#39/1 &sb->s_type->i_lock_key#26 irq_context: 0 &type->s_umount_key#39/1 &s->s_inode_list_lock irq_context: 0 &type->s_umount_key#39/1 tk_core.seq.seqcount irq_context: 0 &type->s_umount_key#39/1 &sb->s_type->i_lock_key#26 &dentry->d_lock irq_context: 0 &type->s_umount_key#39/1 &dentry->d_lock irq_context: 0 &sb->s_type->i_mutex_key#9 namespace_sem irq_context: 0 &sb->s_type->i_mutex_key#9 namespace_sem rcu_read_lock mount_lock.seqcount irq_context: 0 &sb->s_type->i_mutex_key#9 namespace_sem fs_reclaim irq_context: 0 &sb->s_type->i_mutex_key#9 namespace_sem fs_reclaim mmu_notifier_invalidate_range_start irq_context: 0 &sb->s_type->i_mutex_key#9 namespace_sem rename_lock irq_context: 0 &sb->s_type->i_mutex_key#9 namespace_sem rename_lock rename_lock.seqcount irq_context: 0 &sb->s_type->i_mutex_key#9 namespace_sem rename_lock rename_lock.seqcount &dentry->d_lock irq_context: 0 &sb->s_type->i_mutex_key#9 namespace_sem mount_lock irq_context: 0 &sb->s_type->i_mutex_key#9 namespace_sem mount_lock &dentry->d_lock irq_context: 0 &sb->s_type->i_mutex_key#9 namespace_sem mount_lock mount_lock.seqcount irq_context: 0 &sb->s_type->i_mutex_key#9 namespace_sem mount_lock mount_lock.seqcount &new_ns->poll irq_context: 0 &sb->s_type->i_mutex_key#9 namespace_sem mount_lock mount_lock.seqcount &dentry->d_lock irq_context: 0 &sb->s_type->i_mutex_key#9 namespace_sem mount_lock mount_lock.seqcount rcu_read_lock &dentry->d_lock irq_context: 0 &sb->s_type->i_mutex_key#9 namespace_sem mount_lock mount_lock.seqcount &obj_hash[i].lock irq_context: 0 redirect_lock irq_context: 0 &tty->ldisc_sem &tty->atomic_write_lock fs_reclaim irq_context: 0 &tty->ldisc_sem &tty->atomic_write_lock fs_reclaim mmu_notifier_invalidate_range_start irq_context: 0 &tty->ldisc_sem &tty->atomic_write_lock pool_lock#2 irq_context: 0 &tty->ldisc_sem &tty->atomic_write_lock &mm->mmap_lock irq_context: 0 &tty->ldisc_sem &tty->atomic_write_lock &tty->termios_rwsem irq_context: 0 &tty->ldisc_sem &tty->atomic_write_lock &tty->termios_rwsem &tty->write_wait irq_context: 0 &tty->ldisc_sem &tty->atomic_write_lock &tty->termios_rwsem &ldata->output_lock irq_context: 0 &tty->ldisc_sem &tty->atomic_write_lock &tty->termios_rwsem &ldata->output_lock &port_lock_key irq_context: 0 &tty->ldisc_sem &tty->atomic_write_lock &tty->termios_rwsem &ldata->output_lock &port_lock_key &dev->power.lock irq_context: 0 &tty->ldisc_sem &tty->atomic_write_lock &tty->termios_rwsem &ldata->output_lock &port_lock_key &dev->power.lock rcu_read_lock &pool->lock irq_context: 0 &tty->ldisc_sem &tty->atomic_write_lock &tty->termios_rwsem &ldata->output_lock &port_lock_key &dev->power.lock rcu_read_lock &pool->lock &obj_hash[i].lock irq_context: 0 &tty->ldisc_sem &tty->atomic_write_lock &tty->termios_rwsem &ldata->output_lock &port_lock_key &dev->power.lock rcu_read_lock &pool->lock pool_lock#2 irq_context: 0 &tty->ldisc_sem &tty->atomic_write_lock &tty->termios_rwsem &ldata->output_lock &port_lock_key &dev->power.lock rcu_read_lock &pool->lock &p->pi_lock irq_context: 0 &tty->ldisc_sem &tty->atomic_write_lock &tty->termios_rwsem &ldata->output_lock &port_lock_key &dev->power.lock rcu_read_lock &pool->lock &p->pi_lock &rq->__lock irq_context: 0 &tty->ldisc_sem &tty->atomic_write_lock &tty->termios_rwsem &ldata->output_lock &port_lock_key &dev->power.lock rcu_read_lock &pool->lock &p->pi_lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 &tty->ldisc_sem &tty->atomic_write_lock &tty->termios_rwsem &ldata->output_lock &rq->__lock irq_context: 0 (wq_completion)pm (work_completion)(&dev->power.work) &port_lock_key irq_context: 0 (wq_completion)pm (work_completion)(&dev->power.work) &dev->power.lock hrtimer_bases.lock irq_context: 0 (wq_completion)pm (work_completion)(&dev->power.work) &dev->power.lock hrtimer_bases.lock &obj_hash[i].lock irq_context: hardirq &i->lock &port_lock_key irq_context: 0 &tty->ldisc_sem &tty->atomic_write_lock &tty->termios_rwsem &port_lock_key irq_context: hardirq &i->lock &port_lock_key &port->lock irq_context: hardirq &i->lock &port_lock_key &tty->write_wait irq_context: hardirq &i->lock &port_lock_key &tty->write_wait &p->pi_lock irq_context: 0 &tty->ldisc_sem &tty->atomic_write_lock &tty->termios_rwsem &port_lock_key &dev->power.lock irq_context: 0 &tty->ldisc_sem &tty->atomic_write_lock &tty->files_lock irq_context: 0 &tty->ldisc_sem &tty->write_wait irq_context: 0 &type->s_umount_key#40/1 irq_context: 0 &type->s_umount_key#40/1 fs_reclaim irq_context: 0 &type->s_umount_key#40/1 fs_reclaim mmu_notifier_invalidate_range_start irq_context: 0 &type->s_umount_key#40/1 pcpu_alloc_mutex irq_context: 0 &type->s_umount_key#40/1 pcpu_alloc_mutex pcpu_lock irq_context: 0 &type->s_umount_key#40/1 shrinker_rwsem irq_context: 0 &type->s_umount_key#40/1 list_lrus_mutex irq_context: 0 &type->s_umount_key#40/1 &c->lock irq_context: 0 &type->s_umount_key#40/1 &____s->seqcount irq_context: 0 &type->s_umount_key#40/1 pool_lock#2 irq_context: 0 &type->s_umount_key#40/1 sb_lock irq_context: 0 &type->s_umount_key#40/1 sb_lock unnamed_dev_ida.xa_lock irq_context: 0 &type->s_umount_key#40/1 mmu_notifier_invalidate_range_start irq_context: 0 &type->s_umount_key#40/1 &sb->s_type->i_lock_key#27 irq_context: 0 &type->s_umount_key#40/1 &s->s_inode_list_lock irq_context: 0 &type->s_umount_key#40/1 tk_core.seq.seqcount irq_context: 0 &type->s_umount_key#40/1 &sb->s_type->i_lock_key#27 &dentry->d_lock irq_context: 0 &type->s_umount_key#40/1 fuse_mutex irq_context: 0 &type->s_umount_key#40/1 &dentry->d_lock irq_context: 0 &type->s_umount_key#41/1 irq_context: 0 &type->s_umount_key#41/1 fs_reclaim irq_context: 0 &type->s_umount_key#41/1 fs_reclaim mmu_notifier_invalidate_range_start irq_context: 0 &type->s_umount_key#41/1 pool_lock#2 irq_context: 0 &type->s_umount_key#41/1 pcpu_alloc_mutex irq_context: 0 &type->s_umount_key#41/1 pcpu_alloc_mutex pcpu_lock irq_context: 0 &type->s_umount_key#41/1 shrinker_rwsem irq_context: 0 &type->s_umount_key#41/1 list_lrus_mutex irq_context: 0 &type->s_umount_key#41/1 sb_lock irq_context: 0 &type->s_umount_key#41/1 sb_lock unnamed_dev_ida.xa_lock irq_context: 0 &type->s_umount_key#41/1 mmu_notifier_invalidate_range_start irq_context: 0 &type->s_umount_key#41/1 &sb->s_type->i_lock_key#28 irq_context: 0 &type->s_umount_key#41/1 &s->s_inode_list_lock irq_context: 0 &type->s_umount_key#41/1 tk_core.seq.seqcount irq_context: 0 &type->s_umount_key#41/1 &sb->s_type->i_lock_key#28 &dentry->d_lock irq_context: 0 &type->s_umount_key#41/1 pstore_sb_lock irq_context: 0 &type->s_umount_key#41/1 &dentry->d_lock irq_context: 0 &type->i_mutex_dir_key#4 namespace_sem mount_lock mount_lock.seqcount quarantine_lock irq_context: 0 tomoyo_ss &rq->__lock irq_context: 0 &type->s_umount_key#42/1 irq_context: 0 &type->s_umount_key#42/1 fs_reclaim irq_context: 0 &type->s_umount_key#42/1 fs_reclaim mmu_notifier_invalidate_range_start irq_context: 0 &type->s_umount_key#42/1 pcpu_alloc_mutex irq_context: 0 &type->s_umount_key#42/1 pcpu_alloc_mutex pcpu_lock irq_context: 0 &type->s_umount_key#42/1 shrinker_rwsem irq_context: 0 &type->s_umount_key#42/1 list_lrus_mutex irq_context: 0 &type->s_umount_key#42/1 sb_lock irq_context: 0 &type->s_umount_key#42/1 sb_lock unnamed_dev_ida.xa_lock irq_context: 0 &type->s_umount_key#42/1 pool_lock#2 irq_context: 0 &type->s_umount_key#42/1 mmu_notifier_invalidate_range_start irq_context: 0 &type->s_umount_key#42/1 &sb->s_type->i_lock_key#29 irq_context: 0 &type->s_umount_key#42/1 &s->s_inode_list_lock irq_context: 0 &type->s_umount_key#42/1 tk_core.seq.seqcount irq_context: 0 &type->s_umount_key#42/1 &sb->s_type->i_lock_key#29 &dentry->d_lock irq_context: 0 &type->s_umount_key#42/1 bpf_preload_lock irq_context: 0 &type->s_umount_key#42/1 bpf_preload_lock (kmod_concurrent_max).lock irq_context: 0 &type->s_umount_key#42/1 bpf_preload_lock fs_reclaim irq_context: 0 &type->s_umount_key#42/1 bpf_preload_lock fs_reclaim mmu_notifier_invalidate_range_start irq_context: 0 &type->s_umount_key#42/1 bpf_preload_lock pool_lock#2 irq_context: 0 &type->s_umount_key#42/1 bpf_preload_lock &obj_hash[i].lock irq_context: 0 &type->s_umount_key#42/1 bpf_preload_lock rcu_read_lock &pool->lock/1 irq_context: 0 &type->s_umount_key#42/1 bpf_preload_lock rcu_read_lock &pool->lock/1 &obj_hash[i].lock irq_context: 0 &type->s_umount_key#42/1 bpf_preload_lock rcu_read_lock &pool->lock/1 pool_lock#2 irq_context: 0 &type->s_umount_key#42/1 bpf_preload_lock rcu_read_lock &pool->lock/1 &p->pi_lock irq_context: 0 &type->s_umount_key#42/1 bpf_preload_lock &x->wait#17 irq_context: 0 &type->s_umount_key#42/1 bpf_preload_lock &rq->__lock irq_context: 0 &type->s_umount_key#42/1 bpf_preload_lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 &sig->cred_guard_mutex tomoyo_ss rcu_read_lock pool_lock#2 irq_context: softirq (&cb->timer) irq_context: softirq (&cb->timer) &obj_hash[i].lock irq_context: softirq (&cb->timer) &base->lock irq_context: softirq (&cb->timer) &base->lock &obj_hash[i].lock irq_context: 0 &sig->cred_guard_mutex tomoyo_ss tomoyo_policy_lock &c->lock irq_context: 0 &sig->cred_guard_mutex tomoyo_ss tomoyo_policy_lock &____s->seqcount irq_context: 0 &sig->cred_guard_mutex tomoyo_ss batched_entropy_u8.lock irq_context: 0 &sig->cred_guard_mutex tomoyo_ss kfence_freelist_lock irq_context: 0 &sig->cred_guard_mutex tomoyo_ss &meta->lock irq_context: 0 uts_sem irq_context: 0 &mm->mmap_lock &anon_vma->rwsem quarantine_lock irq_context: 0 &type->s_umount_key#42/1 bpf_preload_lock running_helpers_waitq.lock irq_context: 0 &type->s_umount_key#42/1 &dentry->d_lock irq_context: 0 &type->s_umount_key#14 irq_context: 0 &type->s_umount_key#14 sb_lock irq_context: 0 &type->s_umount_key#14 fs_reclaim irq_context: 0 &type->s_umount_key#14 fs_reclaim mmu_notifier_invalidate_range_start irq_context: 0 &type->s_umount_key#14 pool_lock#2 irq_context: 0 &type->s_umount_key#14 &dentry->d_lock irq_context: 0 &type->s_umount_key#14 &lru->node[i].lock irq_context: 0 &type->s_umount_key#14 rcu_read_lock &dentry->d_lock irq_context: 0 &type->s_umount_key#14 &obj_hash[i].lock irq_context: 0 &mm->mmap_lock lock#4 rcu_read_lock pool_lock#2 irq_context: 0 &mm->mmap_lock lock#4 &obj_hash[i].lock irq_context: 0 rcu_read_lock &sb->s_type->i_lock_key irq_context: 0 &type->i_mutex_dir_key#5 irq_context: 0 &type->i_mutex_dir_key#5 fs_reclaim irq_context: 0 &type->i_mutex_dir_key#5 fs_reclaim mmu_notifier_invalidate_range_start irq_context: 0 &type->i_mutex_dir_key#5 &dentry->d_lock irq_context: 0 &type->i_mutex_dir_key#5 rcu_read_lock rename_lock.seqcount irq_context: 0 &type->i_mutex_dir_key#5 &dentry->d_lock &wq irq_context: 0 sb_writers#5 irq_context: 0 sb_writers#5 mount_lock irq_context: 0 sb_writers#5 &type->i_mutex_dir_key#5 irq_context: 0 sb_writers#5 &type->i_mutex_dir_key#5 rename_lock.seqcount irq_context: 0 sb_writers#5 &type->i_mutex_dir_key#5 fs_reclaim irq_context: 0 sb_writers#5 &type->i_mutex_dir_key#5 fs_reclaim mmu_notifier_invalidate_range_start irq_context: 0 sb_writers#5 &type->i_mutex_dir_key#5 &dentry->d_lock irq_context: 0 sb_writers#5 &type->i_mutex_dir_key#5 rcu_read_lock rename_lock.seqcount irq_context: 0 sb_writers#5 &type->i_mutex_dir_key#5 tomoyo_ss irq_context: 0 sb_writers#5 &type->i_mutex_dir_key#5 tomoyo_ss mmu_notifier_invalidate_range_start irq_context: 0 sb_writers#5 &type->i_mutex_dir_key#5 tomoyo_ss rcu_read_lock mount_lock.seqcount irq_context: 0 sb_writers#5 &type->i_mutex_dir_key#5 tomoyo_ss rcu_read_lock rcu_read_lock rename_lock.seqcount irq_context: 0 sb_writers#5 &type->i_mutex_dir_key#5 tomoyo_ss &obj_hash[i].lock irq_context: 0 sb_writers#5 &type->i_mutex_dir_key#5 &dentry->d_lock &wq#2 irq_context: 0 sb_writers#5 &type->i_mutex_dir_key#5 &sbinfo->stat_lock irq_context: 0 sb_writers#5 &type->i_mutex_dir_key#5 pool_lock#2 irq_context: 0 sb_writers#5 &type->i_mutex_dir_key#5 mmu_notifier_invalidate_range_start irq_context: 0 sb_writers#5 &type->i_mutex_dir_key#5 &sb->s_type->i_lock_key irq_context: 0 sb_writers#5 &type->i_mutex_dir_key#5 &s->s_inode_list_lock irq_context: 0 sb_writers#5 &type->i_mutex_dir_key#5 tk_core.seq.seqcount irq_context: 0 sb_writers#5 &type->i_mutex_dir_key#5 batched_entropy_u32.lock irq_context: 0 sb_writers#5 &type->i_mutex_dir_key#5 &simple_offset_xa_lock irq_context: 0 sb_writers#5 &type->i_mutex_dir_key#5 &simple_offset_xa_lock pool_lock#2 irq_context: 0 sb_writers#5 &type->i_mutex_dir_key#5 &sb->s_type->i_lock_key &dentry->d_lock irq_context: 0 &sb->s_type->i_lock_key irq_context: 0 &sb->s_type->i_mutex_key#12 irq_context: 0 sb_writers#5 &type->i_mutex_dir_key#5/1 irq_context: 0 sb_writers#5 &type->i_mutex_dir_key#5/1 rename_lock.seqcount irq_context: 0 sb_writers#5 &type->i_mutex_dir_key#5/1 fs_reclaim irq_context: 0 sb_writers#5 &type->i_mutex_dir_key#5/1 fs_reclaim mmu_notifier_invalidate_range_start irq_context: 0 sb_writers#5 &type->i_mutex_dir_key#5/1 &c->lock irq_context: 0 sb_writers#5 &type->i_mutex_dir_key#5/1 &____s->seqcount irq_context: 0 sb_writers#5 &type->i_mutex_dir_key#5/1 &dentry->d_lock irq_context: 0 sb_writers#5 &type->i_mutex_dir_key#5/1 tomoyo_ss irq_context: 0 sb_writers#5 &type->i_mutex_dir_key#5/1 tomoyo_ss mmu_notifier_invalidate_range_start irq_context: 0 sb_writers#5 &type->i_mutex_dir_key#5/1 tomoyo_ss rcu_read_lock mount_lock.seqcount irq_context: 0 sb_writers#5 &type->i_mutex_dir_key#5/1 tomoyo_ss rcu_read_lock rcu_read_lock rename_lock.seqcount irq_context: 0 sb_writers#5 &type->i_mutex_dir_key#5/1 tomoyo_ss &obj_hash[i].lock irq_context: 0 sb_writers#5 &type->i_mutex_dir_key#5/1 tomoyo_ss &c->lock irq_context: 0 sb_writers#5 &type->i_mutex_dir_key#5/1 tomoyo_ss &____s->seqcount irq_context: 0 sb_writers#5 &type->i_mutex_dir_key#5/1 tomoyo_ss pool_lock#2 irq_context: 0 sb_writers#5 &type->i_mutex_dir_key#5/1 tomoyo_ss rcu_read_lock &dentry->d_lock irq_context: 0 sb_writers#5 &type->i_mutex_dir_key#5/1 tomoyo_ss tomoyo_policy_lock irq_context: 0 sb_writers#5 &type->i_mutex_dir_key#5/1 tomoyo_ss tomoyo_policy_lock mmu_notifier_invalidate_range_start irq_context: 0 sb_writers#5 &type->i_mutex_dir_key#5/1 &sbinfo->stat_lock irq_context: 0 sb_writers#5 &type->i_mutex_dir_key#5/1 mmu_notifier_invalidate_range_start irq_context: 0 sb_writers#5 &type->i_mutex_dir_key#5/1 &sb->s_type->i_lock_key irq_context: 0 sb_writers#5 &type->i_mutex_dir_key#5/1 &s->s_inode_list_lock irq_context: 0 sb_writers#5 &type->i_mutex_dir_key#5/1 tk_core.seq.seqcount irq_context: 0 sb_writers#5 &type->i_mutex_dir_key#5/1 batched_entropy_u32.lock irq_context: 0 sb_writers#5 &type->i_mutex_dir_key#5/1 &simple_offset_xa_lock irq_context: 0 sb_writers#5 &type->i_mutex_dir_key#5/1 &sb->s_type->i_lock_key &dentry->d_lock irq_context: 0 sb_writers#5 &type->i_mutex_dir_key#5/1 rcu_read_lock &dentry->d_lock irq_context: 0 sb_writers#5 &type->i_mutex_dir_key#5/1 &simple_offset_xa_lock pool_lock#2 irq_context: 0 sb_writers#5 tk_core.seq.seqcount irq_context: 0 sb_writers#5 &sb->s_type->i_lock_key irq_context: 0 sb_writers#5 &wb->list_lock irq_context: 0 sb_writers#5 &wb->list_lock &sb->s_type->i_lock_key irq_context: 0 &f->f_lock irq_context: 0 &sig->cred_guard_mutex &stopper->lock irq_context: 0 &sig->cred_guard_mutex &stop_pi_lock irq_context: 0 &sig->cred_guard_mutex &stop_pi_lock &rq->__lock irq_context: 0 &sig->cred_guard_mutex &stop_pi_lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 &sig->cred_guard_mutex &x->wait#8 irq_context: 0 &sig->cred_guard_mutex &p->alloc_lock &x->wait#25 &p->pi_lock &cfs_rq->removed.lock irq_context: 0 &mm->mmap_lock rcu_read_lock rcu_read_lock rcu_node_0 irq_context: 0 &mm->mmap_lock rcu_read_lock rcu_read_lock &rq->__lock irq_context: 0 &mm->mmap_lock &mapping->i_mmap_rwsem &anon_vma->rwsem quarantine_lock irq_context: 0 sb_writers &type->i_mutex_dir_key/1 tomoyo_ss &c->lock irq_context: 0 sb_writers &type->i_mutex_dir_key/1 tomoyo_ss &____s->seqcount irq_context: 0 &mm->mmap_lock batched_entropy_u8.lock irq_context: 0 &mm->mmap_lock kfence_freelist_lock irq_context: 0 &mm->mmap_lock &mapping->i_mmap_rwsem &meta->lock irq_context: 0 &mm->mmap_lock &mapping->i_mmap_rwsem kfence_freelist_lock irq_context: 0 &mm->mmap_lock &anon_vma->rwsem &rq->__lock irq_context: 0 &mm->mmap_lock &anon_vma->rwsem &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 &sb->s_type->i_mutex_key#10 quarantine_lock irq_context: 0 &sig->cred_guard_mutex &iint->mutex &pcp->lock &zone->lock irq_context: 0 &sig->cred_guard_mutex &iint->mutex &pcp->lock &zone->lock &____s->seqcount irq_context: 0 &iint->mutex ima_extend_list_mutex &c->lock irq_context: 0 &iint->mutex ima_extend_list_mutex &____s->seqcount irq_context: 0 uts_sem irq_context: 0 uts_sem hostname_poll.wait.lock irq_context: 0 rcu_read_lock rcu_read_lock mount_lock.seqcount irq_context: 0 rcu_read_lock rcu_read_lock rcu_read_lock rename_lock.seqcount irq_context: 0 &f->f_pos_lock irq_context: 0 &f->f_pos_lock &type->i_mutex_dir_key#3 irq_context: 0 &f->f_pos_lock &type->i_mutex_dir_key#3 fs_reclaim irq_context: 0 &f->f_pos_lock &type->i_mutex_dir_key#3 fs_reclaim mmu_notifier_invalidate_range_start irq_context: 0 &f->f_pos_lock &type->i_mutex_dir_key#3 pool_lock#2 irq_context: 0 &f->f_pos_lock &type->i_mutex_dir_key#3 &ei->i_es_lock irq_context: 0 &f->f_pos_lock &type->i_mutex_dir_key#3 tk_core.seq.seqcount irq_context: 0 &f->f_pos_lock &type->i_mutex_dir_key#3 sb_writers#4 mount_lock irq_context: 0 &f->f_pos_lock &type->i_mutex_dir_key#3 sb_writers#4 tk_core.seq.seqcount irq_context: 0 &f->f_pos_lock &type->i_mutex_dir_key#3 sb_writers#4 mmu_notifier_invalidate_range_start irq_context: 0 &f->f_pos_lock &type->i_mutex_dir_key#3 sb_writers#4 pool_lock#2 irq_context: 0 &f->f_pos_lock &type->i_mutex_dir_key#3 sb_writers#4 &journal->j_state_lock irq_context: 0 &f->f_pos_lock &type->i_mutex_dir_key#3 sb_writers#4 jbd2_handle irq_context: 0 &f->f_pos_lock &type->i_mutex_dir_key#3 sb_writers#4 jbd2_handle &ei->i_raw_lock irq_context: 0 &f->f_pos_lock &type->i_mutex_dir_key#3 sb_writers#4 jbd2_handle &journal->j_wait_updates irq_context: 0 &f->f_pos_lock &type->i_mutex_dir_key#3 sb_writers#4 &obj_hash[i].lock irq_context: 0 &f->f_pos_lock &type->i_mutex_dir_key#3 sb_writers#4 &sb->s_type->i_lock_key#22 irq_context: 0 &f->f_pos_lock &type->i_mutex_dir_key#3 sb_writers#4 &wb->list_lock irq_context: 0 &f->f_pos_lock &type->i_mutex_dir_key#3 sb_writers#4 &wb->list_lock &sb->s_type->i_lock_key#22 irq_context: 0 &f->f_pos_lock &mm->mmap_lock irq_context: 0 &fs->lock &dentry->d_lock irq_context: 0 dup_mmap_sem irq_context: 0 dup_mmap_sem &mm->mmap_lock irq_context: 0 dup_mmap_sem &mm->mmap_lock &mm->mmap_lock/1 irq_context: 0 dup_mmap_sem &mm->mmap_lock &mm->mmap_lock/1 pool_lock#2 irq_context: 0 dup_mmap_sem &mm->mmap_lock &mm->mmap_lock/1 &c->lock irq_context: 0 dup_mmap_sem &mm->mmap_lock &mm->mmap_lock/1 &____s->seqcount irq_context: 0 dup_mmap_sem &mm->mmap_lock &mm->mmap_lock/1 &vma->vm_lock->lock irq_context: 0 dup_mmap_sem &mm->mmap_lock &mm->mmap_lock/1 fs_reclaim irq_context: 0 dup_mmap_sem &mm->mmap_lock &mm->mmap_lock/1 fs_reclaim mmu_notifier_invalidate_range_start irq_context: 0 dup_mmap_sem &mm->mmap_lock &mm->mmap_lock/1 &mapping->i_mmap_rwsem irq_context: 0 dup_mmap_sem &mm->mmap_lock &mm->mmap_lock/1 &anon_vma->rwsem irq_context: 0 dup_mmap_sem &mm->mmap_lock &mm->mmap_lock/1 mmu_notifier_invalidate_range_start irq_context: 0 dup_mmap_sem &mm->mmap_lock &mm->mmap_lock/1 &mm->page_table_lock irq_context: 0 dup_mmap_sem &mm->mmap_lock &mm->mmap_lock/1 ptlock_ptr(page) irq_context: 0 dup_mmap_sem &mm->mmap_lock &mm->mmap_lock/1 ptlock_ptr(page)#2 irq_context: 0 dup_mmap_sem &mm->mmap_lock &mm->mmap_lock/1 ptlock_ptr(page)#2 ptlock_ptr(page)#2/1 irq_context: 0 dup_mmap_sem &mm->mmap_lock &mm->mmap_lock/1 ptlock_ptr(page)#2 key irq_context: 0 dup_mmap_sem &mm->mmap_lock &mm->mmap_lock/1 &mm->context.lock irq_context: 0 dup_mmap_sem &mm->mmap_lock &mm->mmap_lock/1 &obj_hash[i].lock irq_context: 0 cgroup_threadgroup_rwsem &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 &p->alloc_lock &memcg->mm_list.lock irq_context: 0 &vma->vm_lock->lock ptlock_ptr(page)#2 lock#4 &lruvec->lru_lock irq_context: 0 &vma->vm_lock->lock ptlock_ptr(page)#2 lock#5 irq_context: 0 &sig->cred_guard_mutex &tsk->futex_exit_mutex &mm->mmap_lock irq_context: 0 &sig->cred_guard_mutex &sig->exec_update_lock delayed_uprobe_lock irq_context: 0 &sig->cred_guard_mutex &sig->exec_update_lock &mm->mmap_lock lock#4 irq_context: 0 &sig->cred_guard_mutex &sig->exec_update_lock &mm->mmap_lock lock#4 &lruvec->lru_lock irq_context: 0 &sig->cred_guard_mutex &sig->exec_update_lock &mm->mmap_lock lock#5 irq_context: 0 &sig->cred_guard_mutex &sig->exec_update_lock &mm->mmap_lock mmu_notifier_invalidate_range_start irq_context: 0 &sig->cred_guard_mutex &sig->exec_update_lock &mm->mmap_lock ptlock_ptr(page)#2 irq_context: 0 &sig->cred_guard_mutex &sig->exec_update_lock &mm->mmap_lock ptlock_ptr(page)#2 &____s->seqcount irq_context: 0 &sig->cred_guard_mutex &sig->exec_update_lock &mm->mmap_lock ptlock_ptr(page)#2 pool_lock#2 irq_context: 0 &sig->cred_guard_mutex &sig->exec_update_lock &mm->mmap_lock ptlock_ptr(page)#2 key irq_context: 0 &sig->cred_guard_mutex &sig->exec_update_lock &mm->mmap_lock irq_context: 0 &sig->cred_guard_mutex &sig->exec_update_lock &mm->mmap_lock &vma->vm_lock->lock irq_context: 0 &sig->cred_guard_mutex &sig->exec_update_lock &mm->mmap_lock &mapping->i_mmap_rwsem irq_context: 0 &sig->cred_guard_mutex &sig->exec_update_lock &mm->mmap_lock &anon_vma->rwsem irq_context: 0 &sig->cred_guard_mutex &sig->exec_update_lock &mm->mmap_lock &anon_vma->rwsem &obj_hash[i].lock irq_context: 0 &sig->cred_guard_mutex &sig->exec_update_lock &mm->mmap_lock &anon_vma->rwsem pool_lock#2 irq_context: 0 &sig->cred_guard_mutex &sig->exec_update_lock &mm->mmap_lock &obj_hash[i].lock irq_context: 0 &sig->cred_guard_mutex &sig->exec_update_lock &mm->mmap_lock &c->lock irq_context: 0 &sig->cred_guard_mutex &sig->exec_update_lock &mm->mmap_lock pool_lock#2 irq_context: 0 &sig->cred_guard_mutex &sig->exec_update_lock &mm->mmap_lock lock#4 irq_context: 0 &sig->cred_guard_mutex &sig->exec_update_lock &mm->mmap_lock lock#5 irq_context: 0 &sig->cred_guard_mutex &sig->exec_update_lock &mm->mmap_lock &lruvec->lru_lock irq_context: 0 &sig->cred_guard_mutex &sig->exec_update_lock &mm->mmap_lock rcu_read_lock pool_lock#2 irq_context: 0 &sig->cred_guard_mutex &sig->exec_update_lock &memcg->mm_list.lock irq_context: 0 &sig->cred_guard_mutex &sig->exec_update_lock pgd_lock irq_context: 0 &sig->cred_guard_mutex &sig->exec_update_lock rcu_read_lock pool_lock#2 irq_context: 0 &sig->cred_guard_mutex &sig->exec_update_lock &obj_hash[i].lock irq_context: 0 &sig->cred_guard_mutex &sig->exec_update_lock key irq_context: 0 &sig->cred_guard_mutex &sig->exec_update_lock pcpu_lock irq_context: 0 &sig->cred_guard_mutex &sig->exec_update_lock percpu_counters_lock irq_context: 0 &sig->cred_guard_mutex &sig->exec_update_lock pool_lock#2 irq_context: 0 &mm->mmap_lock quarantine_lock irq_context: 0 &type->i_mutex_dir_key#5 rename_lock.seqcount irq_context: 0 &type->i_mutex_dir_key#5 &dentry->d_lock &wq#2 irq_context: 0 sb_writers#5 &type->i_mutex_dir_key#5 tomoyo_ss &c->lock irq_context: 0 sb_writers#5 &type->i_mutex_dir_key#5 tomoyo_ss &____s->seqcount irq_context: 0 sb_writers#5 &type->i_mutex_dir_key#5 tomoyo_ss rcu_read_lock &dentry->d_lock irq_context: 0 sb_writers#5 &type->i_mutex_dir_key#5 tomoyo_ss tomoyo_policy_lock irq_context: 0 sb_writers#5 &type->i_mutex_dir_key#5 tomoyo_ss tomoyo_policy_lock mmu_notifier_invalidate_range_start irq_context: 0 sb_writers#5 &type->i_mutex_dir_key#5 tomoyo_ss tomoyo_policy_lock &c->lock irq_context: 0 sb_writers#5 &type->i_mutex_dir_key#5 tomoyo_ss tomoyo_policy_lock &____s->seqcount irq_context: 0 sb_writers#5 &type->i_mutex_dir_key#5 tomoyo_ss tomoyo_policy_lock pool_lock#2 irq_context: 0 sb_writers#5 &sb->s_type->i_mutex_key#12 irq_context: 0 sb_writers#5 &sb->s_type->i_mutex_key#12 &xattrs->lock irq_context: 0 sb_writers#5 &sb->s_type->i_mutex_key#12 tk_core.seq.seqcount irq_context: 0 sb_writers#5 &sb->s_type->i_mutex_key#12 &sb->s_type->i_lock_key irq_context: 0 sb_writers#5 &sb->s_type->i_mutex_key#12 &wb->list_lock irq_context: 0 sb_writers#5 &sb->s_type->i_mutex_key#12 &wb->list_lock &sb->s_type->i_lock_key irq_context: 0 sb_writers#5 &sb->s_type->i_mutex_key#12 fs_reclaim irq_context: 0 sb_writers#5 &sb->s_type->i_mutex_key#12 fs_reclaim mmu_notifier_invalidate_range_start irq_context: 0 sb_writers#5 &sb->s_type->i_mutex_key#12 &____s->seqcount irq_context: 0 sb_writers#5 &sb->s_type->i_mutex_key#12 pool_lock#2 irq_context: 0 sb_writers#5 &sb->s_type->i_mutex_key#12 &xa->xa_lock#9 irq_context: 0 sb_writers#5 &sb->s_type->i_mutex_key#12 lock#4 irq_context: 0 sb_writers#5 &sb->s_type->i_mutex_key#12 &info->lock irq_context: 0 &p->alloc_lock &x->wait#25 irq_context: 0 &p->alloc_lock &x->wait#25 &p->pi_lock irq_context: 0 &p->alloc_lock &x->wait#25 &p->pi_lock &rq->__lock irq_context: 0 &p->alloc_lock &x->wait#25 &p->pi_lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 &mm->mmap_lock &cfs_rq->removed.lock irq_context: 0 &mm->mmap_lock &n->list_lock irq_context: 0 tasklist_lock &sighand->siglock &p->pi_lock irq_context: 0 tasklist_lock &sighand->siglock &p->pi_lock &rq->__lock irq_context: 0 tasklist_lock &sighand->siglock &p->pi_lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 tasklist_lock &sighand->siglock &p->pi_lock &rq->__lock &cfs_rq->removed.lock irq_context: 0 &sighand->siglock &obj_hash[i].lock irq_context: 0 &sighand->siglock pool_lock#2 irq_context: 0 &mm->mmap_lock ptlock_ptr(page)#2 &pcp->lock &zone->lock irq_context: 0 tomoyo_ss &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 &mm->mmap_lock lock#4 &lruvec->lru_lock per_cpu_ptr(&cgroup_rstat_cpu_lock, cpu) irq_context: 0 &vma->vm_lock->lock &rq->__lock irq_context: 0 sb_writers &type->i_mutex_dir_key/1 &u->bindlock irq_context: 0 sb_writers &type->i_mutex_dir_key/1 &u->bindlock &net->unx.table.locks[i] irq_context: 0 sb_writers &type->i_mutex_dir_key/1 &u->bindlock &net->unx.table.locks[i] &net->unx.table.locks[i]/1 irq_context: 0 sb_writers &type->i_mutex_dir_key/1 &u->bindlock &net->unx.table.locks[i] &net->unx.table.locks[i]/1 &dentry->d_lock irq_context: 0 sb_writers &type->i_mutex_dir_key/1 &u->bindlock &net->unx.table.locks[i]/1 irq_context: 0 sb_writers &type->i_mutex_dir_key/1 &u->bindlock &bsd_socket_locks[i] irq_context: 0 sb_writers &sb->s_type->i_mutex_key#4 irq_context: 0 sb_writers &sb->s_type->i_mutex_key#4 tomoyo_ss irq_context: 0 sb_writers &sb->s_type->i_mutex_key#4 tomoyo_ss mmu_notifier_invalidate_range_start irq_context: 0 sb_writers &sb->s_type->i_mutex_key#4 tomoyo_ss pool_lock#2 irq_context: 0 sb_writers &sb->s_type->i_mutex_key#4 tomoyo_ss rcu_read_lock mount_lock.seqcount irq_context: 0 sb_writers &sb->s_type->i_mutex_key#4 tomoyo_ss rcu_read_lock rcu_read_lock rename_lock.seqcount irq_context: 0 sb_writers &sb->s_type->i_mutex_key#4 tomoyo_ss &obj_hash[i].lock irq_context: 0 sb_writers &sb->s_type->i_mutex_key#4 tomoyo_ss &c->lock irq_context: 0 sb_writers &sb->s_type->i_mutex_key#4 tomoyo_ss &pcp->lock &zone->lock irq_context: 0 sb_writers &sb->s_type->i_mutex_key#4 tomoyo_ss &pcp->lock &zone->lock &____s->seqcount irq_context: 0 sb_writers &sb->s_type->i_mutex_key#4 tomoyo_ss &____s->seqcount irq_context: 0 sb_writers &sb->s_type->i_mutex_key#4 tomoyo_ss rcu_read_lock &dentry->d_lock irq_context: 0 sb_writers &sb->s_type->i_mutex_key#4 tomoyo_ss tomoyo_policy_lock irq_context: 0 sb_writers &sb->s_type->i_mutex_key#4 tomoyo_ss tomoyo_policy_lock mmu_notifier_invalidate_range_start irq_context: 0 sb_writers &sb->s_type->i_mutex_key#4 tomoyo_ss tomoyo_policy_lock pool_lock#2 irq_context: 0 sb_writers &sb->s_type->i_mutex_key#4 tk_core.seq.seqcount irq_context: 0 rcu_read_lock mount_lock.seqcount irq_context: 0 &u->iolock irq_context: 0 &u->iolock rlock-AF_UNIX irq_context: 0 &ei->socket.wq.wait irq_context: 0 &mm->mmap_lock ptlock_ptr(page)#2 lock#5 irq_context: 0 &mm->mmap_lock &anon_vma->rwsem &meta->lock irq_context: 0 &mm->mmap_lock &anon_vma->rwsem kfence_freelist_lock irq_context: 0 &sig->cred_guard_mutex &type->i_mutex_dir_key#3 &ei->i_es_lock key#5 irq_context: 0 &sig->cred_guard_mutex &type->i_mutex_dir_key#3 &ei->i_data_sem &ei->i_es_lock key#6 irq_context: 0 &sig->cred_guard_mutex &type->i_mutex_dir_key#3 &ei->i_data_sem &ei->i_es_lock key#7 irq_context: 0 &sig->cred_guard_mutex &iint->mutex mapping.invalidate_lock irq_context: 0 &sig->cred_guard_mutex &iint->mutex mapping.invalidate_lock mmu_notifier_invalidate_range_start irq_context: 0 &sig->cred_guard_mutex &iint->mutex mapping.invalidate_lock &____s->seqcount irq_context: 0 &sig->cred_guard_mutex &iint->mutex mapping.invalidate_lock &xa->xa_lock#9 irq_context: 0 &sig->cred_guard_mutex &iint->mutex mapping.invalidate_lock lock#4 irq_context: 0 &sig->cred_guard_mutex &iint->mutex mapping.invalidate_lock lock#4 &lruvec->lru_lock irq_context: 0 &sig->cred_guard_mutex &iint->mutex mapping.invalidate_lock &ei->i_es_lock irq_context: 0 &sig->cred_guard_mutex &iint->mutex mapping.invalidate_lock pool_lock#2 irq_context: 0 &sig->cred_guard_mutex &iint->mutex mapping.invalidate_lock rcu_read_lock pool_lock#2 irq_context: 0 &sig->cred_guard_mutex &iint->mutex mapping.invalidate_lock &obj_hash[i].lock irq_context: 0 &sig->cred_guard_mutex &iint->mutex mapping.invalidate_lock tk_core.seq.seqcount irq_context: 0 &sig->cred_guard_mutex &iint->mutex mapping.invalidate_lock &dd->lock irq_context: 0 &sig->cred_guard_mutex &iint->mutex mapping.invalidate_lock rcu_read_lock &dd->lock irq_context: 0 &sig->cred_guard_mutex &iint->mutex mapping.invalidate_lock rcu_read_lock &obj_hash[i].lock irq_context: 0 &sig->cred_guard_mutex &iint->mutex mapping.invalidate_lock rcu_read_lock tk_core.seq.seqcount irq_context: 0 &sig->cred_guard_mutex &iint->mutex mapping.invalidate_lock rcu_read_lock &virtscsi_vq->vq_lock irq_context: 0 &sig->cred_guard_mutex &iint->mutex &folio_wait_table[i] irq_context: 0 &sig->cred_guard_mutex &iint->mutex &rq->__lock irq_context: 0 &sig->cred_guard_mutex &iint->mutex &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 &sig->cred_guard_mutex &sig->exec_update_lock &mm->mmap_lock &____s->seqcount irq_context: 0 &f->f_pos_lock &sb->s_type->i_mutex_key#9 irq_context: 0 &f->f_pos_lock &sb->s_type->i_mutex_key#9 &dentry->d_lock irq_context: 0 &f->f_pos_lock &sb->s_type->i_mutex_key#9 rcu_read_lock &p->alloc_lock irq_context: 0 &f->f_pos_lock &sb->s_type->i_mutex_key#9 rename_lock.seqcount irq_context: 0 &f->f_pos_lock &sb->s_type->i_mutex_key#9 fs_reclaim irq_context: 0 &f->f_pos_lock &sb->s_type->i_mutex_key#9 fs_reclaim mmu_notifier_invalidate_range_start irq_context: 0 &f->f_pos_lock &sb->s_type->i_mutex_key#9 pool_lock#2 irq_context: 0 &f->f_pos_lock &sb->s_type->i_mutex_key#9 rcu_read_lock rename_lock.seqcount irq_context: 0 &f->f_pos_lock &sb->s_type->i_mutex_key#9 mmu_notifier_invalidate_range_start irq_context: 0 &f->f_pos_lock &sb->s_type->i_mutex_key#9 &sb->s_type->i_lock_key#23 irq_context: 0 &f->f_pos_lock &sb->s_type->i_mutex_key#9 &s->s_inode_list_lock irq_context: 0 &f->f_pos_lock &sb->s_type->i_mutex_key#9 tk_core.seq.seqcount irq_context: 0 &f->f_pos_lock &sb->s_type->i_mutex_key#9 &p->alloc_lock irq_context: 0 &f->f_pos_lock &sb->s_type->i_mutex_key#9 &sb->s_type->i_lock_key#23 &dentry->d_lock irq_context: 0 &f->f_pos_lock &sb->s_type->i_mutex_key#9 &sb->s_type->i_lock_key#23 &dentry->d_lock &wq#3 irq_context: 0 &f->f_pos_lock &sb->s_type->i_mutex_key#9 rcu_read_lock &dentry->d_lock irq_context: 0 &f->f_pos_lock &sb->s_type->i_mutex_key#9 &dentry->d_lock &lru->node[i].lock irq_context: 0 &f->f_pos_lock &sb->s_type->i_mutex_key#9 &c->lock irq_context: 0 &f->f_pos_lock &sb->s_type->i_mutex_key#9 &____s->seqcount irq_context: 0 &bsd_socket_locks[i] irq_context: 0 sb_writers tk_core.seq.seqcount irq_context: 0 sb_writers &sb->s_type->i_lock_key#5 irq_context: 0 sb_writers &wb->list_lock irq_context: 0 sb_writers &wb->list_lock &sb->s_type->i_lock_key#5 irq_context: 0 &u->lock irq_context: 0 &u->lock &u->lock/1 irq_context: 0 &mm->mmap_lock &lruvec->lru_lock irq_context: 0 &mm->mmap_lock rcu_read_lock pool_lock#2 irq_context: 0 &group->mark_mutex irq_context: 0 &group->mark_mutex &fsnotify_mark_srcu irq_context: 0 &group->mark_mutex fs_reclaim irq_context: 0 &group->mark_mutex fs_reclaim mmu_notifier_invalidate_range_start irq_context: 0 &group->mark_mutex &____s->seqcount irq_context: 0 &group->mark_mutex pool_lock#2 irq_context: 0 &group->mark_mutex &c->lock irq_context: 0 &group->mark_mutex lock irq_context: 0 &group->mark_mutex lock &group->inotify_data.idr_lock irq_context: 0 &group->mark_mutex lock &group->inotify_data.idr_lock &c->lock irq_context: 0 &group->mark_mutex lock &group->inotify_data.idr_lock &____s->seqcount irq_context: 0 &group->mark_mutex lock &group->inotify_data.idr_lock pool_lock#2 irq_context: 0 &group->mark_mutex ucounts_lock irq_context: 0 &group->mark_mutex &mark->lock irq_context: 0 &group->mark_mutex &mark->lock &fsnotify_mark_srcu irq_context: 0 &group->mark_mutex &mark->lock &fsnotify_mark_srcu &conn->lock irq_context: 0 &group->mark_mutex &mark->lock &conn->lock irq_context: 0 &group->mark_mutex &conn->lock irq_context: 0 &group->mark_mutex &sb->s_type->i_lock_key#5 irq_context: 0 &group->mark_mutex &sb->s_type->i_lock_key#5 &dentry->d_lock irq_context: 0 &group->mark_mutex &sb->s_type->i_lock_key#5 &dentry->d_lock &dentry->d_lock/1 irq_context: 0 &vma->vm_lock->lock &lruvec->lru_lock irq_context: 0 &vma->vm_lock->lock rcu_read_lock pool_lock#2 irq_context: 0 &vma->vm_lock->lock &obj_hash[i].lock irq_context: 0 tomoyo_ss tomoyo_policy_lock mmu_notifier_invalidate_range_start &rq->__lock irq_context: 0 &dentry->d_lock/1 irq_context: 0 tomoyo_ss tomoyo_policy_lock mmu_notifier_invalidate_range_start pgd_lock irq_context: 0 tomoyo_ss tomoyo_policy_lock mmu_notifier_invalidate_range_start rcu_read_lock pool_lock#2 irq_context: 0 tomoyo_ss tomoyo_policy_lock mmu_notifier_invalidate_range_start &obj_hash[i].lock irq_context: 0 tomoyo_ss tomoyo_policy_lock mmu_notifier_invalidate_range_start key irq_context: 0 tomoyo_ss tomoyo_policy_lock mmu_notifier_invalidate_range_start pcpu_lock irq_context: 0 tomoyo_ss tomoyo_policy_lock mmu_notifier_invalidate_range_start percpu_counters_lock irq_context: 0 tomoyo_ss tomoyo_policy_lock mmu_notifier_invalidate_range_start pool_lock#2 irq_context: 0 &f->f_pos_lock &type->i_mutex_dir_key#2 irq_context: 0 &f->f_pos_lock &type->i_mutex_dir_key#2 &dentry->d_lock irq_context: 0 &f->f_pos_lock &type->i_mutex_dir_key#2 rcu_read_lock &dentry->d_lock irq_context: 0 &f->f_pos_lock &type->i_mutex_dir_key#2 tk_core.seq.seqcount irq_context: 0 &f->f_pos_lock &type->i_mutex_dir_key#2 sb_writers mount_lock irq_context: 0 &f->f_pos_lock &type->i_mutex_dir_key#2 sb_writers tk_core.seq.seqcount irq_context: 0 &f->f_pos_lock &type->i_mutex_dir_key#2 sb_writers &sb->s_type->i_lock_key#5 irq_context: 0 &f->f_pos_lock &type->i_mutex_dir_key#2 sb_writers &wb->list_lock irq_context: 0 &f->f_pos_lock &type->i_mutex_dir_key#2 sb_writers &wb->list_lock &sb->s_type->i_lock_key#5 irq_context: 0 &fsnotify_mark_srcu &conn->lock irq_context: 0 &conn->lock irq_context: 0 &evdev->client_lock irq_context: 0 &evdev->mutex irq_context: 0 &evdev->mutex &dev->mutex#2 irq_context: 0 &evdev->mutex &mm->mmap_lock irq_context: 0 tasklist_lock &sighand->siglock &p->pi_lock &cfs_rq->removed.lock irq_context: 0 rcu_state.exp_mutex rcu_read_lock &pool->lock &p->pi_lock &rq->__lock &cfs_rq->removed.lock irq_context: 0 dup_mmap_sem &mm->mmap_lock &mm->mmap_lock/1 rcu_node_0 irq_context: 0 dup_mmap_sem &mm->mmap_lock &mm->mmap_lock/1 &rcu_state.expedited_wq irq_context: 0 dup_mmap_sem &mm->mmap_lock &mm->mmap_lock/1 &rcu_state.expedited_wq &p->pi_lock irq_context: 0 dup_mmap_sem &mm->mmap_lock &mm->mmap_lock/1 &rq->__lock irq_context: 0 dup_mmap_sem &mm->mmap_lock &mm->mmap_lock/1 &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 &evdev->mutex &dev->mutex#2 rcu_state.exp_mutex rcu_node_0 irq_context: 0 &evdev->mutex &dev->mutex#2 rcu_state.exp_mutex &obj_hash[i].lock irq_context: 0 &evdev->mutex &dev->mutex#2 rcu_state.exp_mutex rcu_read_lock &pool->lock irq_context: 0 &evdev->mutex &dev->mutex#2 rcu_state.exp_mutex rcu_read_lock &pool->lock &obj_hash[i].lock irq_context: 0 &evdev->mutex &dev->mutex#2 rcu_state.exp_mutex rcu_read_lock &pool->lock &p->pi_lock irq_context: 0 &evdev->mutex &dev->mutex#2 rcu_state.exp_mutex rcu_read_lock &pool->lock &p->pi_lock &rq->__lock irq_context: 0 &evdev->mutex &dev->mutex#2 rcu_state.exp_mutex rcu_read_lock &pool->lock &p->pi_lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 &evdev->mutex &dev->mutex#2 rcu_state.exp_mutex &rnp->exp_wq[3] irq_context: 0 &evdev->mutex &dev->mutex#2 rcu_state.exp_mutex &rq->__lock irq_context: 0 &evdev->mutex &dev->mutex#2 rcu_state.exp_mutex &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 &evdev->mutex &dev->mutex#2 &obj_hash[i].lock irq_context: 0 sk_lock-AF_NETLINK irq_context: 0 sk_lock-AF_NETLINK slock-AF_NETLINK irq_context: 0 slock-AF_NETLINK irq_context: 0 sk_lock-AF_NETLINK rcu_read_lock rhashtable_bucket irq_context: 0 cb_lock irq_context: 0 cb_lock genl_mutex irq_context: 0 cb_lock genl_mutex fs_reclaim irq_context: 0 cb_lock genl_mutex fs_reclaim mmu_notifier_invalidate_range_start irq_context: 0 cb_lock genl_mutex pool_lock#2 irq_context: 0 cb_lock genl_mutex &c->lock irq_context: 0 cb_lock genl_mutex &____s->seqcount irq_context: 0 cb_lock genl_mutex &obj_hash[i].lock irq_context: 0 cb_lock genl_mutex rlock-AF_NETLINK irq_context: 0 cb_lock fs_reclaim irq_context: 0 cb_lock fs_reclaim mmu_notifier_invalidate_range_start irq_context: 0 cb_lock pool_lock#2 irq_context: 0 cb_lock rlock-AF_NETLINK irq_context: 0 rlock-AF_NETLINK irq_context: 0 &nlk->wait irq_context: 0 &sb->s_type->i_mutex_key#10 rcu_read_lock rhashtable_bucket irq_context: 0 &sb->s_type->i_mutex_key#10 rcu_read_lock rcu_read_lock &pool->lock irq_context: 0 &sb->s_type->i_mutex_key#10 rcu_read_lock rcu_read_lock &pool->lock &obj_hash[i].lock irq_context: 0 &sb->s_type->i_mutex_key#10 rcu_read_lock rcu_read_lock &pool->lock &p->pi_lock irq_context: 0 &sb->s_type->i_mutex_key#10 rcu_read_lock rcu_read_lock &pool->lock &p->pi_lock &rq->__lock irq_context: 0 &sb->s_type->i_mutex_key#10 rcu_read_lock rcu_read_lock &pool->lock &p->pi_lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 &sb->s_type->i_mutex_key#10 nl_table_lock irq_context: 0 &sb->s_type->i_mutex_key#10 nl_table_wait.lock irq_context: 0 &sb->s_type->i_mutex_key#10 clock-AF_NETLINK irq_context: 0 &sb->s_type->i_mutex_key#10 genl_sk_destructing_waitq.lock irq_context: 0 &sb->s_type->i_mutex_key#10 &nlk->wait irq_context: 0 &sb->s_type->i_mutex_key#10 wlock-AF_NETLINK irq_context: 0 &sb->s_type->i_mutex_key#10 (netlink_chain).rwsem irq_context: 0 &sb->s_type->i_mutex_key#10 (netlink_chain).rwsem pool_lock#2 irq_context: 0 &sb->s_type->i_mutex_key#10 (netlink_chain).rwsem &obj_hash[i].lock irq_context: 0 &sb->s_type->i_mutex_key#10 (netlink_chain).rwsem rcu_read_lock &pool->lock irq_context: 0 &sb->s_type->i_mutex_key#10 (netlink_chain).rwsem rcu_read_lock &pool->lock &obj_hash[i].lock irq_context: 0 &sb->s_type->i_mutex_key#10 (netlink_chain).rwsem rcu_read_lock &pool->lock pool_lock#2 irq_context: 0 &sb->s_type->i_mutex_key#10 (netlink_chain).rwsem rcu_read_lock &pool->lock &p->pi_lock irq_context: 0 &sb->s_type->i_mutex_key#10 (netlink_chain).rwsem rcu_read_lock &rdev->beacon_registrations_lock irq_context: 0 &sb->s_type->i_mutex_key#10 (netlink_chain).rwsem rcu_read_lock &rdev->mgmt_registrations_lock irq_context: 0 &sb->s_type->i_mutex_key#10 (netlink_chain).rwsem rcu_read_lock &wdev->pmsr_lock irq_context: 0 &sb->s_type->i_mutex_key#10 (netlink_chain).rwsem reg_indoor_lock irq_context: 0 &sb->s_type->i_mutex_key#10 (netlink_chain).rwsem hwsim_radio_lock irq_context: 0 nl_table_lock pool_lock#2 irq_context: 0 sb_writers#5 &type->i_mutex_dir_key#5/1 &obj_hash[i].lock irq_context: 0 sb_writers#6 irq_context: 0 sb_writers#6 mount_lock irq_context: 0 sb_writers#6 &sb->s_type->i_mutex_key#10 irq_context: 0 sb_writers#6 &sb->s_type->i_mutex_key#10 tomoyo_ss irq_context: 0 sb_writers#6 &sb->s_type->i_mutex_key#10 tomoyo_ss mmu_notifier_invalidate_range_start irq_context: 0 sb_writers#6 &sb->s_type->i_mutex_key#10 tomoyo_ss pool_lock#2 irq_context: 0 sb_writers#6 &sb->s_type->i_mutex_key#10 tomoyo_ss &obj_hash[i].lock irq_context: 0 sb_writers#6 &sb->s_type->i_mutex_key#10 tomoyo_ss rcu_read_lock &dentry->d_lock irq_context: 0 sb_writers#6 &sb->s_type->i_mutex_key#10 tomoyo_ss tomoyo_policy_lock irq_context: 0 sb_writers#6 &sb->s_type->i_mutex_key#10 tomoyo_ss tomoyo_policy_lock mmu_notifier_invalidate_range_start irq_context: 0 sb_writers#6 &sb->s_type->i_mutex_key#10 tomoyo_ss tomoyo_policy_lock pool_lock#2 irq_context: 0 sb_writers#6 &sb->s_type->i_mutex_key#10 tomoyo_ss &rq->__lock irq_context: 0 sb_writers#6 &sb->s_type->i_mutex_key#10 tomoyo_ss &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 sb_writers#6 &sb->s_type->i_mutex_key#10 tk_core.seq.seqcount irq_context: 0 sb_writers#6 &sb->s_type->i_mutex_key#10 &sb->s_type->i_lock_key#8 irq_context: 0 sb_writers#6 &sb->s_type->i_mutex_key#10 &wb->list_lock irq_context: 0 (wq_completion)events (work_completion)(&ht->run_work) irq_context: 0 sb_writers#6 &sb->s_type->i_mutex_key#10 &wb->list_lock &sb->s_type->i_lock_key#8 irq_context: 0 (wq_completion)events (work_completion)(&ht->run_work) &ht->mutex irq_context: 0 (wq_completion)events (work_completion)(&ht->run_work) &ht->mutex fs_reclaim irq_context: 0 (wq_completion)events (work_completion)(&ht->run_work) &ht->mutex fs_reclaim mmu_notifier_invalidate_range_start irq_context: 0 (wq_completion)events (work_completion)(&ht->run_work) &ht->mutex pool_lock#2 irq_context: 0 (wq_completion)events (work_completion)(&ht->run_work) &ht->mutex batched_entropy_u32.lock irq_context: 0 (wq_completion)events (work_completion)(&ht->run_work) &ht->mutex rhashtable_bucket irq_context: 0 (wq_completion)events (work_completion)(&ht->run_work) &ht->mutex rhashtable_bucket rhashtable_bucket/1 irq_context: 0 (wq_completion)events (work_completion)(&ht->run_work) &ht->mutex &ht->lock irq_context: 0 sb_writers#5 &type->i_mutex_dir_key#5/1 tomoyo_ss tomoyo_policy_lock pool_lock#2 irq_context: 0 (wq_completion)events (work_completion)(&ht->run_work) &ht->mutex &ht->lock &obj_hash[i].lock irq_context: 0 sb_writers#5 &type->i_mutex_dir_key#5/1 tomoyo_ss tomoyo_policy_lock &c->lock irq_context: 0 (wq_completion)events (work_completion)(&ht->run_work) &ht->mutex &ht->lock pool_lock#2 irq_context: 0 sb_writers#5 &type->i_mutex_dir_key#5/1 tomoyo_ss tomoyo_policy_lock &____s->seqcount irq_context: 0 (wq_completion)events (work_completion)(&w->w) irq_context: 0 (wq_completion)events (work_completion)(&w->w) nfc_devlist_mutex irq_context: 0 sb_writers#5 &type->i_mutex_dir_key#5/1 pool_lock#2 irq_context: 0 (wq_completion)events (work_completion)(&w->w) nfc_devlist_mutex &k->list_lock irq_context: 0 (wq_completion)events (work_completion)(&w->w) nfc_devlist_mutex &k->k_lock irq_context: 0 sb_writers#5 &type->i_mutex_dir_key#5/1 &u->bindlock irq_context: 0 sb_writers#5 &type->i_mutex_dir_key#5/1 &u->bindlock &net->unx.table.locks[i] irq_context: 0 sb_writers#5 &type->i_mutex_dir_key#5/1 &u->bindlock &net->unx.table.locks[i] &net->unx.table.locks[i]/1 irq_context: 0 sb_writers#5 &type->i_mutex_dir_key#5/1 &u->bindlock &net->unx.table.locks[i] &net->unx.table.locks[i]/1 &dentry->d_lock irq_context: 0 (wq_completion)events (work_completion)(&w->w) nfc_devlist_mutex &genl_data->genl_data_mutex irq_context: 0 sb_writers#5 &type->i_mutex_dir_key#5/1 &u->bindlock &bsd_socket_locks[i] irq_context: 0 (wq_completion)events (work_completion)(&w->w) &obj_hash[i].lock irq_context: 0 (wq_completion)events (work_completion)(&w->w) pool_lock#2 irq_context: 0 &u->lock &sk->sk_peer_lock irq_context: 0 &u->lock rlock-AF_UNIX irq_context: 0 rcu_read_lock &ei->socket.wq.wait irq_context: 0 rcu_read_lock &ei->socket.wq.wait &p->pi_lock irq_context: 0 rcu_read_lock &ei->socket.wq.wait &p->pi_lock &rq->__lock irq_context: 0 rcu_read_lock &ei->socket.wq.wait &p->pi_lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 &u->iolock &mm->mmap_lock irq_context: 0 &u->iolock &obj_hash[i].lock irq_context: 0 &u->iolock pool_lock#2 irq_context: 0 sb_writers#5 &type->i_mutex_dir_key#5 rcu_read_lock &dentry->d_lock irq_context: 0 &f->f_pos_lock &type->i_mutex_dir_key#3 &ei->i_data_sem irq_context: 0 &f->f_pos_lock &type->i_mutex_dir_key#3 &ei->i_data_sem mmu_notifier_invalidate_range_start irq_context: 0 &f->f_pos_lock &type->i_mutex_dir_key#3 &ei->i_data_sem pool_lock#2 irq_context: 0 &f->f_pos_lock &type->i_mutex_dir_key#3 &ei->i_data_sem &ei->i_es_lock irq_context: 0 &f->f_pos_lock &type->i_mutex_dir_key#3 &ei->i_data_sem &ei->i_es_lock &____s->seqcount irq_context: 0 &f->f_pos_lock &type->i_mutex_dir_key#3 &ei->i_data_sem &ei->i_es_lock pool_lock#2 irq_context: 0 &f->f_pos_lock &type->i_mutex_dir_key#3 &ei->i_data_sem &ei->i_es_lock rcu_read_lock pool_lock#2 irq_context: 0 &f->f_pos_lock &type->i_mutex_dir_key#3 &ei->i_data_sem &ei->i_es_lock &obj_hash[i].lock irq_context: 0 &f->f_pos_lock &type->i_mutex_dir_key#3 &ei->i_data_sem &ei->i_es_lock &sbi->s_es_lock irq_context: 0 &f->f_pos_lock &type->i_mutex_dir_key#3 &ei->i_data_sem &obj_hash[i].lock irq_context: 0 &f->f_pos_lock &type->i_mutex_dir_key#3 mmu_notifier_invalidate_range_start irq_context: 0 &f->f_pos_lock &type->i_mutex_dir_key#3 &____s->seqcount irq_context: 0 &f->f_pos_lock &type->i_mutex_dir_key#3 &xa->xa_lock#9 irq_context: 0 &f->f_pos_lock &type->i_mutex_dir_key#3 lock#4 irq_context: 0 &f->f_pos_lock &type->i_mutex_dir_key#3 &mapping->private_lock irq_context: 0 &f->f_pos_lock &type->i_mutex_dir_key#3 &dd->lock irq_context: 0 &f->f_pos_lock &type->i_mutex_dir_key#3 &obj_hash[i].lock irq_context: 0 &f->f_pos_lock &type->i_mutex_dir_key#3 rcu_read_lock &pool->lock irq_context: 0 &f->f_pos_lock &type->i_mutex_dir_key#3 rcu_read_lock &pool->lock &obj_hash[i].lock irq_context: 0 &f->f_pos_lock &type->i_mutex_dir_key#3 rcu_read_lock &pool->lock pool_lock#2 irq_context: 0 &f->f_pos_lock &type->i_mutex_dir_key#3 rcu_read_lock &pool->lock &p->pi_lock irq_context: 0 &f->f_pos_lock &type->i_mutex_dir_key#3 rcu_read_lock &pool->lock &p->pi_lock &rq->__lock irq_context: 0 &f->f_pos_lock &type->i_mutex_dir_key#3 rcu_read_lock &pool->lock &p->pi_lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 &f->f_pos_lock &type->i_mutex_dir_key#3 bit_wait_table + i irq_context: 0 &f->f_pos_lock &type->i_mutex_dir_key#3 &rq->__lock irq_context: 0 &f->f_pos_lock &type->i_mutex_dir_key#3 &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 &mm->mmap_lock &mapping->i_mmap_rwsem &rq->__lock irq_context: 0 sb_writers#5 &type->i_mutex_dir_key#5 tomoyo_ss pool_lock#2 irq_context: 0 &group->notification_waitq irq_context: 0 &group->notification_lock irq_context: 0 &client->wait irq_context: 0 key#8 irq_context: softirq rcu_callback rlock-AF_NETLINK irq_context: softirq rcu_callback &dir->lock irq_context: 0 syslog_lock irq_context: 0 &u->lock &u->peer_wait irq_context: 0 &u->iolock &u->peer_wait irq_context: 0 &u->iolock &u->peer_wait &p->pi_lock irq_context: 0 &u->iolock &u->peer_wait &p->pi_lock &rq->__lock irq_context: 0 &u->iolock &u->peer_wait &p->pi_lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 sb_writers#5 &sb->s_type->i_mutex_key#12 &xa->xa_lock#9 pool_lock#2 irq_context: 0 &u->iolock quarantine_lock irq_context: 0 sb_writers#5 &sb->s_type->i_mutex_key#12 lock#4 &lruvec->lru_lock irq_context: 0 &type->i_mutex_dir_key#3 &ei->i_data_sem &rq->__lock irq_context: 0 &u->iolock &u->peer_wait &p->pi_lock &cfs_rq->removed.lock irq_context: 0 rcu_read_lock &ei->socket.wq.wait &p->pi_lock &cfs_rq->removed.lock irq_context: 0 &sb->s_type->i_lock_key#14 irq_context: 0 &sb->s_type->i_lock_key#14 &dentry->d_lock irq_context: 0 &pipe->mutex/1 irq_context: 0 &pipe->rd_wait irq_context: 0 dup_mmap_sem &mm->mmap_lock &mm->mmap_lock/1 &pcp->lock &zone->lock irq_context: 0 sb_writers &type->i_mutex_dir_key#2 irq_context: 0 sb_writers &type->i_mutex_dir_key#2 rename_lock.seqcount irq_context: 0 sb_writers &type->i_mutex_dir_key#2 rcu_read_lock &dentry->d_lock irq_context: 0 dup_mmap_sem &mm->mmap_lock &mm->mmap_lock/1 &anon_vma->rwsem &c->lock irq_context: 0 sb_writers#5 &sb->s_type->i_mutex_key#12 lock#4 rcu_read_lock pool_lock#2 irq_context: 0 dup_mmap_sem &mm->mmap_lock &mm->mmap_lock/1 &anon_vma->rwsem &____s->seqcount irq_context: 0 sb_writers#5 &sb->s_type->i_mutex_key#12 lock#4 &obj_hash[i].lock irq_context: softirq (&cb->timer) tk_core.seq.seqcount irq_context: 0 rcu_read_lock &ei->socket.wq.wait &p->pi_lock &rq->__lock &cfs_rq->removed.lock irq_context: 0 &u->iolock &rq->__lock irq_context: hardirq &dev->power.lock hrtimer_bases.lock irq_context: hardirq &dev->power.lock hrtimer_bases.lock &obj_hash[i].lock irq_context: 0 &sig->cred_guard_mutex &sig->exec_update_lock &mm->mmap_lock &rq->__lock irq_context: 0 &sig->cred_guard_mutex &sig->exec_update_lock &mm->mmap_lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 &sig->cred_guard_mutex &sig->exec_update_lock &mm->mmap_lock &mapping->i_mmap_rwsem &sem->wait_lock irq_context: 0 &sig->cred_guard_mutex &sig->exec_update_lock &mm->mmap_lock &mapping->i_mmap_rwsem &rq->__lock irq_context: 0 &sig->cred_guard_mutex &sig->exec_update_lock &mm->mmap_lock &mapping->i_mmap_rwsem &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 &mm->mmap_lock &sem->wait_lock irq_context: 0 &mm->mmap_lock &p->pi_lock irq_context: 0 &mm->mmap_lock &p->pi_lock &rq->__lock irq_context: 0 &mm->mmap_lock &p->pi_lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 &mm->mmap_lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 &type->s_umount_key#32 sb_writers#4 mmu_notifier_invalidate_range_start irq_context: 0 &type->s_umount_key#32 sb_writers#4 &____s->seqcount irq_context: 0 &type->s_umount_key#32 sb_writers#4 pool_lock#2 irq_context: 0 &type->s_umount_key#32 sb_writers#4 &xa->xa_lock#9 irq_context: 0 &type->s_umount_key#32 sb_writers#4 lock#4 irq_context: 0 &type->s_umount_key#32 sb_writers#4 &mapping->private_lock irq_context: 0 &type->s_umount_key#32 sb_writers#4 tk_core.seq.seqcount irq_context: 0 &type->s_umount_key#32 sb_writers#4 &dd->lock irq_context: 0 &type->s_umount_key#32 sb_writers#4 &c->lock irq_context: 0 &type->s_umount_key#32 sb_writers#4 rcu_read_lock &dd->lock irq_context: 0 &type->s_umount_key#32 sb_writers#4 rcu_read_lock &obj_hash[i].lock irq_context: 0 &type->s_umount_key#32 sb_writers#4 rcu_read_lock pool_lock#2 irq_context: 0 &type->s_umount_key#32 sb_writers#4 rcu_read_lock tk_core.seq.seqcount irq_context: 0 &type->s_umount_key#32 sb_writers#4 rcu_read_lock &virtscsi_vq->vq_lock irq_context: 0 &type->s_umount_key#32 sb_writers#4 &xa->xa_lock#9 pool_lock#2 irq_context: 0 &type->s_umount_key#32 sb_writers#4 &pcp->lock &zone->lock irq_context: 0 &type->s_umount_key#32 sb_writers#4 bit_wait_table + i irq_context: 0 &type->s_umount_key#32 sb_writers#4 &rq->__lock irq_context: 0 &type->s_umount_key#32 sb_writers#4 &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 &type->s_umount_key#32 &eli->li_list_mtx irq_context: 0 &type->s_umount_key#32 sb_writers#4 &journal->j_state_lock irq_context: 0 &type->s_umount_key#32 sb_writers#4 jbd2_handle irq_context: 0 &type->s_umount_key#32 sb_writers#4 jbd2_handle &meta_group_info[i]->alloc_sem irq_context: 0 &type->s_umount_key#32 sb_writers#4 jbd2_handle &meta_group_info[i]->alloc_sem mmu_notifier_invalidate_range_start irq_context: 0 &type->s_umount_key#32 sb_writers#4 jbd2_handle &meta_group_info[i]->alloc_sem pool_lock#2 irq_context: 0 &type->s_umount_key#32 sb_writers#4 jbd2_handle &meta_group_info[i]->alloc_sem tk_core.seq.seqcount irq_context: 0 &type->s_umount_key#32 sb_writers#4 jbd2_handle &meta_group_info[i]->alloc_sem &dd->lock irq_context: 0 &type->s_umount_key#32 sb_writers#4 jbd2_handle &meta_group_info[i]->alloc_sem &x->wait#26 irq_context: 0 &type->s_umount_key#32 sb_writers#4 jbd2_handle &meta_group_info[i]->alloc_sem &obj_hash[i].lock irq_context: 0 &type->s_umount_key#32 sb_writers#4 jbd2_handle &meta_group_info[i]->alloc_sem rcu_read_lock &pool->lock irq_context: 0 &type->s_umount_key#32 sb_writers#4 jbd2_handle &meta_group_info[i]->alloc_sem rcu_read_lock &pool->lock &obj_hash[i].lock irq_context: 0 &type->s_umount_key#32 sb_writers#4 jbd2_handle &meta_group_info[i]->alloc_sem rcu_read_lock &pool->lock pool_lock#2 irq_context: 0 &type->s_umount_key#32 sb_writers#4 jbd2_handle &meta_group_info[i]->alloc_sem rcu_read_lock &pool->lock &p->pi_lock irq_context: 0 &type->s_umount_key#32 sb_writers#4 jbd2_handle &meta_group_info[i]->alloc_sem rcu_read_lock &pool->lock &p->pi_lock &rq->__lock irq_context: 0 &type->s_umount_key#32 sb_writers#4 jbd2_handle &meta_group_info[i]->alloc_sem rcu_read_lock &pool->lock &p->pi_lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 &type->s_umount_key#32 sb_writers#4 jbd2_handle &meta_group_info[i]->alloc_sem rcu_node_0 irq_context: 0 &type->s_umount_key#32 sb_writers#4 jbd2_handle &meta_group_info[i]->alloc_sem &rq->__lock irq_context: 0 (wq_completion)kblockd (work_completion)(&(&hctx->run_work)->work) rcu_read_lock &obj_hash[i].lock pool_lock irq_context: 0 (wq_completion)kblockd (work_completion)(&(&hctx->run_work)->work) rcu_read_lock &____s->seqcount irq_context: 0 (wq_completion)kblockd (work_completion)(&(&hctx->run_work)->work) rcu_read_lock pool_lock#2 irq_context: 0 &vma->vm_lock->lock &pcp->lock &zone->lock &____s->seqcount irq_context: 0 &sig->cred_guard_mutex &sig->exec_update_lock &mm->mmap_lock &sem->wait_lock irq_context: 0 &sig->cred_guard_mutex &sig->exec_update_lock &mm->mmap_lock &p->pi_lock irq_context: 0 &sig->cred_guard_mutex &sig->exec_update_lock &mm->mmap_lock &p->pi_lock &rq->__lock irq_context: 0 &sig->cred_guard_mutex &sig->exec_update_lock &mm->mmap_lock &p->pi_lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: softirq &x->wait#26 irq_context: softirq &x->wait#26 &p->pi_lock irq_context: softirq &x->wait#26 &p->pi_lock &rq->__lock irq_context: softirq &x->wait#26 &p->pi_lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 &type->s_umount_key#32 sb_writers#4 jbd2_handle &meta_group_info[i]->alloc_sem rcu_read_lock rcu_node_0 irq_context: 0 &type->s_umount_key#32 sb_writers#4 jbd2_handle &meta_group_info[i]->alloc_sem rcu_read_lock &rcu_state.gp_wq irq_context: 0 &type->s_umount_key#32 sb_writers#4 jbd2_handle &meta_group_info[i]->alloc_sem rcu_read_lock &rcu_state.gp_wq &p->pi_lock irq_context: 0 &type->s_umount_key#32 sb_writers#4 jbd2_handle &meta_group_info[i]->alloc_sem &base->lock irq_context: 0 &type->s_umount_key#32 sb_writers#4 jbd2_handle &meta_group_info[i]->alloc_sem &base->lock &obj_hash[i].lock irq_context: 0 &type->s_umount_key#32 sb_writers#4 jbd2_handle &meta_group_info[i]->alloc_sem &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 &type->s_umount_key#32 sb_writers#4 jbd2_handle &meta_group_info[i]->alloc_sem (&timer.timer) irq_context: 0 &type->s_umount_key#32 sb_writers#4 jbd2_handle &meta_group_info[i]->alloc_sem &fq->mq_flush_lock irq_context: 0 &type->s_umount_key#32 sb_writers#4 jbd2_handle &meta_group_info[i]->alloc_sem &fq->mq_flush_lock tk_core.seq.seqcount irq_context: 0 &type->s_umount_key#32 sb_writers#4 jbd2_handle &meta_group_info[i]->alloc_sem &fq->mq_flush_lock &q->requeue_lock irq_context: 0 &type->s_umount_key#32 sb_writers#4 jbd2_handle &meta_group_info[i]->alloc_sem &fq->mq_flush_lock &obj_hash[i].lock irq_context: 0 &type->s_umount_key#32 sb_writers#4 jbd2_handle &meta_group_info[i]->alloc_sem &fq->mq_flush_lock rcu_read_lock &pool->lock irq_context: 0 &type->s_umount_key#32 sb_writers#4 jbd2_handle &meta_group_info[i]->alloc_sem &fq->mq_flush_lock rcu_read_lock &pool->lock &obj_hash[i].lock irq_context: 0 &type->s_umount_key#32 sb_writers#4 jbd2_handle &meta_group_info[i]->alloc_sem &fq->mq_flush_lock rcu_read_lock &pool->lock pool_lock#2 irq_context: 0 &type->s_umount_key#32 sb_writers#4 jbd2_handle &meta_group_info[i]->alloc_sem &fq->mq_flush_lock rcu_read_lock &pool->lock &p->pi_lock irq_context: 0 &type->s_umount_key#32 sb_writers#4 jbd2_handle &meta_group_info[i]->alloc_sem &fq->mq_flush_lock rcu_read_lock &pool->lock &p->pi_lock &rq->__lock irq_context: 0 &type->s_umount_key#32 sb_writers#4 jbd2_handle &meta_group_info[i]->alloc_sem &fq->mq_flush_lock rcu_read_lock &pool->lock &p->pi_lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: softirq &fq->mq_flush_lock &x->wait#26 irq_context: softirq &fq->mq_flush_lock &x->wait#26 &p->pi_lock irq_context: softirq &fq->mq_flush_lock &x->wait#26 &p->pi_lock &rq->__lock irq_context: softirq &fq->mq_flush_lock &x->wait#26 &p->pi_lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 &type->s_umount_key#32 sb_writers#4 jbd2_handle &journal->j_wait_updates irq_context: 0 &type->s_umount_key#32 sb_writers#4 &obj_hash[i].lock irq_context: 0 &mm->mmap_lock &mapping->i_mmap_rwsem &sem->wait_lock irq_context: 0 &mm->mmap_lock &mapping->i_mmap_rwsem &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 &sig->cred_guard_mutex &sig->exec_update_lock &mm->mmap_lock lock#4 &lruvec->lru_lock irq_context: 0 tomoyo_ss &n->list_lock irq_context: 0 tomoyo_ss batched_entropy_u8.lock irq_context: 0 tomoyo_ss kfence_freelist_lock irq_context: 0 tomoyo_ss &meta->lock irq_context: 0 &type->i_mutex_dir_key#5 pool_lock#2 irq_context: 0 &u->lock clock-AF_UNIX irq_context: 0 &u->peer_wait irq_context: 0 rlock-AF_UNIX irq_context: 0 &mm->mmap_lock rcu_read_lock &rcu_state.gp_wq irq_context: 0 &mm->mmap_lock rcu_read_lock &rcu_state.gp_wq &p->pi_lock irq_context: 0 &mm->mmap_lock rcu_read_lock &rcu_state.gp_wq &p->pi_lock &rq->__lock irq_context: 0 &mm->mmap_lock rcu_read_lock &rcu_state.gp_wq &p->pi_lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 sb_writers#5 &sb->s_type->i_mutex_key#12 key#9 irq_context: 0 sb_writers#5 &type->i_mutex_dir_key#5/1 tomoyo_ss rcu_read_lock pool_lock#2 irq_context: 0 sb_writers#5 &type->i_mutex_dir_key#5/1 &sb->s_type->i_mutex_key#12 irq_context: 0 sb_writers#5 &type->i_mutex_dir_key#5/1 &sb->s_type->i_mutex_key#12 &dentry->d_lock irq_context: 0 sb_writers#5 &type->i_mutex_dir_key#5/1 &sb->s_type->i_mutex_key#12 &simple_offset_xa_lock irq_context: 0 sb_writers#5 &type->i_mutex_dir_key#5/1 &sb->s_type->i_mutex_key#12 &simple_offset_xa_lock &obj_hash[i].lock irq_context: 0 sb_writers#5 &type->i_mutex_dir_key#5/1 &sb->s_type->i_mutex_key#12 &simple_offset_xa_lock pool_lock#2 irq_context: 0 sb_writers#5 &type->i_mutex_dir_key#5/1 &sb->s_type->i_mutex_key#12 fs_reclaim irq_context: 0 sb_writers#5 &type->i_mutex_dir_key#5/1 &sb->s_type->i_mutex_key#12 fs_reclaim mmu_notifier_invalidate_range_start irq_context: 0 sb_writers#5 &type->i_mutex_dir_key#5/1 &sb->s_type->i_mutex_key#12 tk_core.seq.seqcount irq_context: 0 sb_writers#5 &type->i_mutex_dir_key#5/1 &sb->s_type->i_mutex_key#12 rename_lock irq_context: 0 sb_writers#5 &type->i_mutex_dir_key#5/1 &sb->s_type->i_mutex_key#12 rename_lock rename_lock.seqcount irq_context: 0 sb_writers#5 &type->i_mutex_dir_key#5/1 &sb->s_type->i_mutex_key#12 rename_lock rename_lock.seqcount &dentry->d_lock irq_context: 0 sb_writers#5 &type->i_mutex_dir_key#5/1 &sb->s_type->i_mutex_key#12 rename_lock rename_lock.seqcount &dentry->d_lock &dentry->d_lock/2 irq_context: 0 sb_writers#5 &type->i_mutex_dir_key#5/1 &sb->s_type->i_mutex_key#12 rename_lock rename_lock.seqcount &dentry->d_lock &dentry->d_lock/2 &dentry->d_lock/3 irq_context: 0 sb_writers#5 &type->i_mutex_dir_key#5/1 &sb->s_type->i_mutex_key#12 rename_lock rename_lock.seqcount &dentry->d_lock &dentry->d_lock/2 &dentry->d_lock/3 &____s->seqcount#4 irq_context: 0 sb_writers#5 &type->i_mutex_dir_key#5/1 &sb->s_type->i_mutex_key#12 rename_lock rename_lock.seqcount &dentry->d_lock &dentry->d_lock/2 &dentry->d_lock/3 &____s->seqcount#4 &____s->seqcount#4/1 irq_context: 0 sb_writers#5 &type->i_mutex_dir_key#5/1 &sb->s_type->i_mutex_key#12 rename_lock rename_lock.seqcount &dentry->d_lock/2 irq_context: 0 sb_writers#5 &type->i_mutex_dir_key#5/1 &sb->s_type->i_mutex_key#12 rename_lock rename_lock.seqcount &dentry->d_lock/2 &dentry->d_lock/3 irq_context: 0 sb_writers#5 &sb->s_type->i_mutex_key#12 &rq->__lock irq_context: 0 &type->i_mutex_dir_key#5 &c->lock irq_context: 0 &type->i_mutex_dir_key#5 &____s->seqcount irq_context: softirq &x->wait#26 &p->pi_lock &cfs_rq->removed.lock irq_context: 0 &pipe->mutex/1 &pipe->rd_wait irq_context: 0 &pipe->mutex/1 &pipe->rd_wait &p->pi_lock irq_context: 0 &pipe->mutex/1 &pipe->rd_wait &p->pi_lock &rq->__lock irq_context: 0 &pipe->mutex/1 &pipe->rd_wait &p->pi_lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 &pipe->mutex/1 &pipe->rd_wait &p->pi_lock &rq->__lock &cfs_rq->removed.lock irq_context: 0 &pipe->mutex/1 &pipe->wr_wait irq_context: 0 &pipe->mutex/1 &rq->__lock irq_context: 0 &pipe->mutex/1 &lock->wait_lock irq_context: 0 &pipe->mutex/1 &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 &lock->wait_lock irq_context: 0 &pipe->mutex/1 fs_reclaim irq_context: 0 &pipe->mutex/1 fs_reclaim mmu_notifier_invalidate_range_start irq_context: 0 &pipe->mutex/1 &____s->seqcount irq_context: 0 &pipe->mutex/1 pool_lock#2 irq_context: 0 &pipe->mutex/1 &mm->mmap_lock irq_context: 0 &pipe->rd_wait &p->pi_lock irq_context: 0 &pipe->rd_wait &p->pi_lock &rq->__lock irq_context: 0 &pipe->rd_wait &p->pi_lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 sb_writers#7 tk_core.seq.seqcount irq_context: 0 sb_writers#7 mount_lock irq_context: 0 rcu_read_lock &vma->vm_lock->lock rcu_node_0 irq_context: 0 &pipe->mutex/1 &cfs_rq->removed.lock irq_context: 0 &pipe->mutex/1 &obj_hash[i].lock irq_context: 0 &pipe->mutex/1 &pipe->rd_wait &p->pi_lock &cfs_rq->removed.lock irq_context: 0 &type->i_mutex_dir_key#3 rcu_read_lock &pool->lock &p->pi_lock &rq->__lock &cfs_rq->removed.lock irq_context: 0 sb_writers#3 &sb->s_type->i_mutex_key#9 irq_context: 0 sb_writers#3 &sb->s_type->i_mutex_key#9 rename_lock.seqcount irq_context: 0 sb_writers#3 &sb->s_type->i_mutex_key#9 rcu_read_lock &dentry->d_lock irq_context: 0 sb_writers#3 &sb->s_type->i_mutex_key#9 rcu_read_lock &dentry->d_lock sysctl_lock irq_context: 0 sb_writers#3 &dentry->d_lock irq_context: 0 sb_writers#3 tomoyo_ss irq_context: 0 sb_writers#3 tomoyo_ss mmu_notifier_invalidate_range_start irq_context: 0 sb_writers#3 tomoyo_ss rcu_read_lock rename_lock.seqcount irq_context: 0 sb_writers#3 tomoyo_ss &obj_hash[i].lock irq_context: 0 sb_writers#3 tomoyo_ss rcu_read_lock &dentry->d_lock irq_context: 0 sb_writers#3 tomoyo_ss tomoyo_policy_lock irq_context: 0 sb_writers#3 tomoyo_ss pool_lock#2 irq_context: 0 sb_writers#3 tomoyo_ss &c->lock irq_context: 0 sb_writers#3 tomoyo_ss &____s->seqcount irq_context: 0 sb_writers#3 &sb->s_type->i_mutex_key#9 tk_core.seq.seqcount irq_context: 0 sb_writers#3 &mm->mmap_lock irq_context: 0 &mm->mmap_lock ptlock_ptr(page)#2 lock#4 rcu_read_lock pool_lock#2 irq_context: 0 &mm->mmap_lock ptlock_ptr(page)#2 lock#4 &obj_hash[i].lock irq_context: 0 &sig->cred_guard_mutex &iint->mutex mapping.invalidate_lock &ei->i_es_lock key#2 irq_context: 0 &sig->cred_guard_mutex &iint->mutex mapping.invalidate_lock batched_entropy_u8.lock irq_context: 0 &sig->cred_guard_mutex &iint->mutex mapping.invalidate_lock kfence_freelist_lock irq_context: 0 &sig->cred_guard_mutex &iint->mutex mapping.invalidate_lock &c->lock irq_context: 0 &sig->cred_guard_mutex &iint->mutex mapping.invalidate_lock &xa->xa_lock#9 pool_lock#2 irq_context: 0 &iint->mutex mapping.invalidate_lock &ei->i_data_sem &ei->i_es_lock &c->lock irq_context: 0 &iint->mutex mapping.invalidate_lock &ei->i_data_sem &ei->i_es_lock &____s->seqcount irq_context: 0 &type->s_umount_key#32 sb_writers#4 batched_entropy_u8.lock irq_context: 0 &type->s_umount_key#32 sb_writers#4 kfence_freelist_lock irq_context: 0 &type->s_umount_key#32 sb_writers#4 &meta->lock irq_context: 0 &mm->mmap_lock &mapping->i_mmap_rwsem mmu_notifier_invalidate_range_start irq_context: 0 &mm->mmap_lock &mapping->i_mmap_rwsem ptlock_ptr(page) irq_context: 0 &iint->mutex sb_writers#4 &c->lock irq_context: 0 &iint->mutex sb_writers#4 &____s->seqcount irq_context: 0 &iint->mutex &pcp->lock &zone->lock irq_context: 0 &iint->mutex &pcp->lock &zone->lock &____s->seqcount irq_context: 0 &sb->s_type->i_mutex_key#9 &dentry->d_lock &wq irq_context: 0 &sb->s_type->i_mutex_key#9 rcu_read_lock &dentry->d_lock irq_context: 0 &sb->s_type->i_mutex_key#9 &obj_hash[i].lock irq_context: 0 sb_writers#5 &type->i_mutex_dir_key#5/1 tomoyo_ss &pcp->lock &zone->lock irq_context: 0 sb_writers#5 &type->i_mutex_dir_key#5/1 tomoyo_ss &pcp->lock &zone->lock &____s->seqcount irq_context: 0 sk_lock-AF_UNIX irq_context: 0 sk_lock-AF_UNIX slock-AF_UNIX irq_context: 0 slock-AF_UNIX irq_context: 0 sb_writers#5 &type->i_mutex_dir_key#5/1 &u->bindlock &net->unx.table.locks[i]/1 irq_context: hardirq log_wait.lock &p->pi_lock irq_context: hardirq log_wait.lock &p->pi_lock &cfs_rq->removed.lock irq_context: hardirq log_wait.lock &p->pi_lock &rq->__lock irq_context: hardirq log_wait.lock &p->pi_lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 &iint->mutex mapping.invalidate_lock batched_entropy_u8.lock irq_context: 0 &iint->mutex mapping.invalidate_lock kfence_freelist_lock irq_context: 0 &f->f_pos_lock &type->i_mutex_dir_key#3 batched_entropy_u8.lock irq_context: 0 &f->f_pos_lock &type->i_mutex_dir_key#3 kfence_freelist_lock irq_context: 0 &f->f_pos_lock &type->i_mutex_dir_key#3 &c->lock irq_context: 0 &f->f_pos_lock &type->i_mutex_dir_key#3 &xa->xa_lock#9 pool_lock#2 irq_context: 0 &f->f_pos_lock &type->i_mutex_dir_key#3 sb_writers#4 jbd2_handle mmu_notifier_invalidate_range_start irq_context: 0 &f->f_pos_lock &type->i_mutex_dir_key#3 sb_writers#4 jbd2_handle pool_lock#2 irq_context: 0 &f->f_pos_lock &type->i_mutex_dir_key#3 sb_writers#4 jbd2_handle &ret->b_state_lock irq_context: 0 &f->f_pos_lock &type->i_mutex_dir_key#3 sb_writers#4 jbd2_handle &ret->b_state_lock &journal->j_list_lock irq_context: 0 &f->f_pos_lock &type->i_mutex_dir_key#3 sb_writers#4 jbd2_handle &journal->j_revoke_lock irq_context: 0 &ei->xattr_sem irq_context: 0 &iint->mutex mapping.invalidate_lock &ei->i_es_lock key#5 irq_context: 0 &iint->mutex mapping.invalidate_lock &ei->i_data_sem &ei->i_es_lock key#6 irq_context: 0 &iint->mutex mapping.invalidate_lock &ei->i_data_sem &ei->i_es_lock key#7 irq_context: 0 sb_writers &sb->s_type->i_mutex_key#4 tomoyo_ss rcu_read_lock pool_lock#2 irq_context: 0 sb_writers &sb->s_type->i_mutex_key#4 &xattrs->lock irq_context: 0 sb_writers#3 &sb->s_type->i_mutex_key#9 fs_reclaim irq_context: 0 sb_writers#3 &sb->s_type->i_mutex_key#9 fs_reclaim mmu_notifier_invalidate_range_start irq_context: 0 sb_writers#3 &sb->s_type->i_mutex_key#9 &dentry->d_lock irq_context: 0 sb_writers#3 &sb->s_type->i_mutex_key#9 rcu_read_lock rename_lock.seqcount irq_context: 0 sb_writers#3 &sb->s_type->i_mutex_key#9 tomoyo_ss irq_context: 0 sb_writers#3 &sb->s_type->i_mutex_key#9 tomoyo_ss mmu_notifier_invalidate_range_start irq_context: 0 sb_writers#3 &sb->s_type->i_mutex_key#9 tomoyo_ss rcu_read_lock rename_lock.seqcount irq_context: 0 sb_writers#3 &sb->s_type->i_mutex_key#9 tomoyo_ss &obj_hash[i].lock irq_context: 0 sb_writers#3 &sb->s_type->i_mutex_key#9 mmu_notifier_invalidate_range_start irq_context: 0 sb_writers#3 &sb->s_type->i_mutex_key#9 &sb->s_type->i_lock_key#23 irq_context: 0 sb_writers#3 &sb->s_type->i_mutex_key#9 &s->s_inode_list_lock irq_context: 0 sb_writers#3 &sb->s_type->i_mutex_key#9 &p->alloc_lock irq_context: 0 sb_writers#3 &sb->s_type->i_mutex_key#9 &sb->s_type->i_lock_key#23 &dentry->d_lock irq_context: 0 sb_writers#3 &sb->s_type->i_mutex_key#9 &sb->s_type->i_lock_key#23 &dentry->d_lock &wq#2 irq_context: 0 sb_writers#3 oom_adj_mutex irq_context: 0 sb_writers#3 oom_adj_mutex rcu_read_lock &p->alloc_lock irq_context: 0 sb_writers#3 oom_adj_mutex &p->alloc_lock irq_context: 0 low_water_lock console_owner_lock irq_context: 0 low_water_lock console_owner irq_context: 0 &group->mark_mutex &sb->s_type->i_lock_key#22 irq_context: 0 &group->mark_mutex &sb->s_type->i_lock_key#22 &dentry->d_lock irq_context: 0 &group->mark_mutex &sb->s_type->i_lock_key#22 &dentry->d_lock &dentry->d_lock/1 irq_context: 0 &group->mark_mutex &sb->s_type->i_lock_key irq_context: 0 &group->mark_mutex &sb->s_type->i_lock_key &dentry->d_lock irq_context: 0 &sk->sk_peer_lock irq_context: 0 &ep->mtx irq_context: 0 epnested_mutex irq_context: 0 epnested_mutex &ep->mtx irq_context: 0 epnested_mutex &ep->mtx fs_reclaim irq_context: 0 epnested_mutex &ep->mtx fs_reclaim mmu_notifier_invalidate_range_start irq_context: 0 epnested_mutex &ep->mtx &____s->seqcount irq_context: 0 epnested_mutex &ep->mtx pool_lock#2 irq_context: 0 epnested_mutex &ep->mtx &c->lock irq_context: 0 epnested_mutex &ep->mtx &f->f_lock irq_context: 0 epnested_mutex &ep->mtx &ei->socket.wq.wait irq_context: 0 epnested_mutex &ep->mtx &ep->lock irq_context: 0 epnested_mutex rcu_read_lock &f->f_lock irq_context: 0 &ep->mtx fs_reclaim irq_context: 0 &ep->mtx fs_reclaim mmu_notifier_invalidate_range_start irq_context: 0 &ep->mtx &f->f_lock irq_context: 0 &ep->mtx pool_lock#2 irq_context: 0 &ep->mtx &group->notification_waitq irq_context: 0 &ep->mtx &group->notification_lock irq_context: 0 &ep->mtx &ep->lock irq_context: 0 &ep->mtx &sighand->signalfd_wqh irq_context: 0 &ep->mtx &sighand->siglock irq_context: 0 &ep->mtx &ei->socket.wq.wait irq_context: 0 &ep->lock irq_context: 0 &sig->cred_guard_mutex tomoyo_ss &dentry->d_lock irq_context: 0 &sig->cred_guard_mutex &iint->mutex mapping.invalidate_lock rcu_read_lock &c->lock irq_context: 0 &sig->cred_guard_mutex &iint->mutex mapping.invalidate_lock rcu_read_lock &____s->seqcount irq_context: 0 &eli->li_list_mtx &obj_hash[i].lock irq_context: 0 &eli->li_list_mtx pool_lock#2 irq_context: 0 ext4_li_mtx irq_context: 0 ext4_li_mtx &eli->li_list_mtx irq_context: 0 ext4_li_mtx &obj_hash[i].lock irq_context: 0 ext4_li_mtx pool_lock#2 irq_context: 0 &type->i_mutex_dir_key#4 rename_lock.seqcount irq_context: 0 &type->i_mutex_dir_key#4 &sb->s_type->i_lock_key#24 &dentry->d_lock &wq#2 irq_context: 0 &f->f_pos_lock &type->i_mutex_dir_key#4 irq_context: 0 &f->f_pos_lock &type->i_mutex_dir_key#4 &dentry->d_lock irq_context: 0 &f->f_pos_lock &type->i_mutex_dir_key#4 &root->kernfs_rwsem irq_context: 0 &f->f_pos_lock &type->i_mutex_dir_key#4 &mm->mmap_lock fs_reclaim irq_context: 0 &f->f_pos_lock &type->i_mutex_dir_key#4 &mm->mmap_lock fs_reclaim mmu_notifier_invalidate_range_start irq_context: 0 &f->f_pos_lock &type->i_mutex_dir_key#4 &mm->mmap_lock &____s->seqcount irq_context: 0 &f->f_pos_lock &type->i_mutex_dir_key#4 &mm->mmap_lock ptlock_ptr(page)#2 irq_context: 0 &f->f_pos_lock &type->i_mutex_dir_key#4 &mm->mmap_lock ptlock_ptr(page)#2 lock#4 irq_context: 0 &f->f_pos_lock &type->i_mutex_dir_key#4 tk_core.seq.seqcount irq_context: 0 &type->i_mutex_dir_key#4 &root->kernfs_rwsem &c->lock irq_context: 0 &type->i_mutex_dir_key#4 &root->kernfs_rwsem &____s->seqcount irq_context: 0 &type->i_mutex_dir_key#4 &c->lock irq_context: 0 &type->i_mutex_dir_key#4 &root->kernfs_rwsem &pcp->lock &zone->lock irq_context: 0 &type->i_mutex_dir_key#4 &root->kernfs_rwsem &pcp->lock &zone->lock &____s->seqcount irq_context: 0 &vma->vm_lock->lock ptlock_ptr(page)#2 lock#4 rcu_read_lock pool_lock#2 irq_context: 0 &vma->vm_lock->lock ptlock_ptr(page)#2 lock#4 &obj_hash[i].lock irq_context: 0 &type->i_mutex_dir_key#4 &pcp->lock &zone->lock irq_context: 0 &type->i_mutex_dir_key#4 &pcp->lock &zone->lock &____s->seqcount irq_context: 0 &f->f_pos_lock &type->i_mutex_dir_key#4 sb_writers#8 mount_lock irq_context: 0 &f->f_pos_lock &type->i_mutex_dir_key#4 sb_writers#8 tk_core.seq.seqcount irq_context: 0 &f->f_pos_lock &type->i_mutex_dir_key#4 sb_writers#8 &sb->s_type->i_lock_key#24 irq_context: 0 &f->f_pos_lock &type->i_mutex_dir_key#4 sb_writers#8 &wb->list_lock irq_context: 0 &f->f_pos_lock &type->i_mutex_dir_key#4 sb_writers#8 &wb->list_lock &sb->s_type->i_lock_key#24 irq_context: 0 remove_cache_srcu &c->lock irq_context: 0 remove_cache_srcu &n->list_lock irq_context: 0 remove_cache_srcu rcu_read_lock pool_lock#2 irq_context: 0 remove_cache_srcu &obj_hash[i].lock irq_context: 0 tomoyo_ss &n->list_lock &c->lock irq_context: 0 &type->i_mutex_dir_key#4 remove_cache_srcu irq_context: 0 &type->i_mutex_dir_key#4 remove_cache_srcu quarantine_lock irq_context: 0 &type->i_mutex_dir_key#4 remove_cache_srcu &c->lock irq_context: 0 &type->i_mutex_dir_key#4 remove_cache_srcu &n->list_lock irq_context: 0 &type->i_mutex_dir_key#4 remove_cache_srcu rcu_read_lock pool_lock#2 irq_context: 0 &type->i_mutex_dir_key#4 remove_cache_srcu &obj_hash[i].lock irq_context: 0 tomoyo_ss remove_cache_srcu irq_context: 0 tomoyo_ss remove_cache_srcu quarantine_lock irq_context: 0 tomoyo_ss remove_cache_srcu &c->lock irq_context: 0 tomoyo_ss remove_cache_srcu &n->list_lock irq_context: 0 tomoyo_ss remove_cache_srcu rcu_read_lock pool_lock#2 irq_context: 0 tomoyo_ss remove_cache_srcu &obj_hash[i].lock irq_context: 0 swap_lock irq_context: 0 sb_writers#8 irq_context: 0 sb_writers#8 mount_lock irq_context: 0 sb_writers#8 &type->i_mutex_dir_key#4 irq_context: 0 sb_writers#8 &type->i_mutex_dir_key#4 rename_lock.seqcount irq_context: 0 sb_writers#8 &type->i_mutex_dir_key#4 fs_reclaim irq_context: 0 sb_writers#8 &type->i_mutex_dir_key#4 fs_reclaim mmu_notifier_invalidate_range_start irq_context: 0 sb_writers#8 &type->i_mutex_dir_key#4 &dentry->d_lock irq_context: 0 sb_writers#8 &type->i_mutex_dir_key#4 rcu_read_lock rename_lock.seqcount irq_context: 0 sb_writers#8 &type->i_mutex_dir_key#4 &root->kernfs_rwsem irq_context: 0 sb_writers#8 &type->i_mutex_dir_key#4 &root->kernfs_rwsem inode_hash_lock irq_context: 0 sb_writers#8 &type->i_mutex_dir_key#4 &root->kernfs_rwsem fs_reclaim irq_context: 0 sb_writers#8 &type->i_mutex_dir_key#4 &root->kernfs_rwsem fs_reclaim mmu_notifier_invalidate_range_start irq_context: 0 sb_writers#8 &type->i_mutex_dir_key#4 &root->kernfs_rwsem mmu_notifier_invalidate_range_start irq_context: 0 sb_writers#8 &type->i_mutex_dir_key#4 &root->kernfs_rwsem inode_hash_lock &sb->s_type->i_lock_key#24 irq_context: 0 sb_writers#8 &type->i_mutex_dir_key#4 &root->kernfs_rwsem inode_hash_lock &s->s_inode_list_lock irq_context: 0 sb_writers#8 &type->i_mutex_dir_key#4 &root->kernfs_rwsem tk_core.seq.seqcount irq_context: 0 sb_writers#8 &type->i_mutex_dir_key#4 &root->kernfs_rwsem &sb->s_type->i_lock_key#24 irq_context: 0 sb_writers#8 &type->i_mutex_dir_key#4 &sb->s_type->i_lock_key#24 irq_context: 0 sb_writers#8 &type->i_mutex_dir_key#4 &sb->s_type->i_lock_key#24 &dentry->d_lock irq_context: 0 sb_writers#8 &type->i_mutex_dir_key#4 &sb->s_type->i_lock_key#24 &dentry->d_lock &wq#2 irq_context: 0 kn->active fs_reclaim irq_context: 0 kn->active fs_reclaim mmu_notifier_invalidate_range_start irq_context: 0 kn->active pool_lock#2 irq_context: 0 kn->active &kernfs_locks->open_file_mutex[count] irq_context: 0 kn->active &kernfs_locks->open_file_mutex[count] fs_reclaim irq_context: 0 kn->active &kernfs_locks->open_file_mutex[count] fs_reclaim mmu_notifier_invalidate_range_start irq_context: 0 kn->active &kernfs_locks->open_file_mutex[count] pool_lock#2 irq_context: 0 &sb->s_type->i_mutex_key#13 irq_context: 0 sb_writers#8 fs_reclaim irq_context: 0 sb_writers#8 fs_reclaim mmu_notifier_invalidate_range_start irq_context: 0 sb_writers#8 pool_lock#2 irq_context: 0 sb_writers#8 &mm->mmap_lock irq_context: 0 sb_writers#8 &of->mutex irq_context: 0 sb_writers#8 &of->mutex kn->active &k->list_lock irq_context: 0 sb_writers#8 &of->mutex kn->active fs_reclaim irq_context: 0 sb_writers#8 &of->mutex kn->active fs_reclaim mmu_notifier_invalidate_range_start irq_context: 0 sb_writers#8 &of->mutex kn->active pool_lock#2 irq_context: 0 sb_writers#8 &of->mutex kn->active uevent_sock_mutex irq_context: 0 sb_writers#8 &of->mutex kn->active uevent_sock_mutex fs_reclaim irq_context: 0 sb_writers#8 &of->mutex kn->active uevent_sock_mutex fs_reclaim mmu_notifier_invalidate_range_start irq_context: 0 sb_writers#8 &of->mutex kn->active uevent_sock_mutex pool_lock#2 irq_context: 0 sb_writers#8 &of->mutex kn->active uevent_sock_mutex nl_table_lock irq_context: 0 sb_writers#8 &of->mutex kn->active uevent_sock_mutex rlock-AF_NETLINK irq_context: 0 sb_writers#8 &of->mutex kn->active uevent_sock_mutex rcu_read_lock &ei->socket.wq.wait irq_context: 0 sb_writers#8 &of->mutex kn->active uevent_sock_mutex rcu_read_lock &ei->socket.wq.wait &ep->lock irq_context: 0 sb_writers#8 &of->mutex kn->active uevent_sock_mutex rcu_read_lock &ei->socket.wq.wait &ep->lock &ep->wq irq_context: 0 sb_writers#8 &of->mutex kn->active uevent_sock_mutex rcu_read_lock &ei->socket.wq.wait &ep->lock &ep->wq &p->pi_lock irq_context: 0 sb_writers#8 &of->mutex kn->active uevent_sock_mutex nl_table_wait.lock irq_context: 0 sb_writers#8 &of->mutex kn->active uevent_sock_mutex &obj_hash[i].lock irq_context: 0 sb_writers#8 &of->mutex kn->active &obj_hash[i].lock irq_context: 0 sb_writers#8 &obj_hash[i].lock irq_context: 0 &kernfs_locks->open_file_mutex[count] irq_context: 0 &kernfs_locks->open_file_mutex[count] &obj_hash[i].lock irq_context: 0 &kernfs_locks->open_file_mutex[count] pool_lock#2 irq_context: 0 &kernfs_locks->open_file_mutex[count] krc.lock irq_context: 0 &kernfs_locks->open_file_mutex[count] krc.lock &obj_hash[i].lock irq_context: 0 &kernfs_locks->open_file_mutex[count] krc.lock &base->lock irq_context: 0 &kernfs_locks->open_file_mutex[count] krc.lock &base->lock &obj_hash[i].lock irq_context: 0 kn->active#2 fs_reclaim irq_context: 0 kn->active#2 fs_reclaim mmu_notifier_invalidate_range_start irq_context: 0 kn->active#2 &kernfs_locks->open_file_mutex[count] irq_context: 0 kn->active#2 &kernfs_locks->open_file_mutex[count] fs_reclaim irq_context: 0 kn->active#2 &kernfs_locks->open_file_mutex[count] fs_reclaim mmu_notifier_invalidate_range_start irq_context: 0 sb_writers#8 &c->lock irq_context: 0 sb_writers#8 &of->mutex kn->active#2 fs_reclaim irq_context: 0 sb_writers#8 &of->mutex kn->active#2 fs_reclaim mmu_notifier_invalidate_range_start irq_context: 0 sb_writers#8 &of->mutex kn->active#2 &c->lock irq_context: 0 sb_writers#8 &of->mutex kn->active#2 &n->list_lock irq_context: 0 sb_writers#8 &of->mutex kn->active#2 &n->list_lock &c->lock irq_context: 0 sb_writers#8 &of->mutex kn->active#2 pool_lock#2 irq_context: 0 sb_writers#8 &of->mutex kn->active#2 uevent_sock_mutex irq_context: 0 sb_writers#8 &of->mutex kn->active#2 uevent_sock_mutex fs_reclaim irq_context: 0 sb_writers#8 &of->mutex kn->active#2 uevent_sock_mutex fs_reclaim mmu_notifier_invalidate_range_start irq_context: 0 sb_writers#8 &of->mutex kn->active#2 uevent_sock_mutex pool_lock#2 irq_context: 0 sb_writers#8 &of->mutex kn->active#2 uevent_sock_mutex nl_table_lock irq_context: 0 sb_writers#8 &of->mutex kn->active#2 uevent_sock_mutex &c->lock irq_context: 0 sb_writers#8 &of->mutex kn->active#2 uevent_sock_mutex rlock-AF_NETLINK irq_context: 0 sb_writers#8 &of->mutex kn->active#2 uevent_sock_mutex rcu_read_lock &ei->socket.wq.wait irq_context: 0 sb_writers#8 &of->mutex kn->active#2 uevent_sock_mutex rcu_read_lock &ei->socket.wq.wait &ep->lock irq_context: 0 sb_writers#8 &of->mutex kn->active#2 uevent_sock_mutex nl_table_wait.lock irq_context: 0 sb_writers#8 &of->mutex kn->active#2 uevent_sock_mutex &obj_hash[i].lock irq_context: 0 sb_writers#8 &of->mutex kn->active#2 &obj_hash[i].lock irq_context: 0 &mm->mmap_lock &mapping->i_mmap_rwsem irq_context: 0 &mm->mmap_lock &anon_vma->rwsem irq_context: 0 kn->active#2 &c->lock irq_context: 0 kn->active#2 &____s->seqcount irq_context: 0 kn->active#2 pool_lock#2 irq_context: 0 sb_writers#8 &type->i_mutex_dir_key#4 &c->lock irq_context: 0 kn->active#2 &kernfs_locks->open_file_mutex[count] &c->lock irq_context: 0 sb_writers#8 &type->i_mutex_dir_key#4 &root->kernfs_rwsem &c->lock irq_context: 0 sb_writers#8 &type->i_mutex_dir_key#4 &root->kernfs_rwsem &____s->seqcount irq_context: 0 kn->active &kernfs_locks->open_file_mutex[count] &c->lock irq_context: 0 sb_writers#8 &of->mutex kn->active &c->lock irq_context: 0 sb_writers#8 &of->mutex kn->active &____s->seqcount irq_context: 0 sb_writers#8 &of->mutex kn->active uevent_sock_mutex &c->lock irq_context: 0 &mm->mmap_lock &anon_vma->rwsem &obj_hash[i].lock irq_context: 0 &mm->mmap_lock &anon_vma->rwsem pool_lock#2 irq_context: 0 kn->active &c->lock irq_context: 0 &kernfs_locks->open_file_mutex[count] &obj_hash[i].lock pool_lock irq_context: 0 &f->f_pos_lock &type->i_mutex_dir_key#3 &n->list_lock irq_context: 0 &f->f_pos_lock &type->i_mutex_dir_key#3 &n->list_lock &c->lock irq_context: 0 sb_writers#8 &type->i_mutex_dir_key#4 &____s->seqcount irq_context: 0 &f->f_pos_lock &type->i_mutex_dir_key#5 irq_context: 0 &f->f_pos_lock &type->i_mutex_dir_key#5 &dentry->d_lock irq_context: 0 &f->f_pos_lock &type->i_mutex_dir_key#5 tk_core.seq.seqcount irq_context: 0 &f->f_pos_lock &type->i_mutex_dir_key#5 sb_writers#5 mount_lock irq_context: 0 &f->f_pos_lock &type->i_mutex_dir_key#5 sb_writers#5 tk_core.seq.seqcount irq_context: 0 &f->f_pos_lock &type->i_mutex_dir_key#5 sb_writers#5 &sb->s_type->i_lock_key irq_context: 0 &f->f_pos_lock &type->i_mutex_dir_key#5 sb_writers#5 &wb->list_lock irq_context: 0 &f->f_pos_lock &type->i_mutex_dir_key#5 sb_writers#5 &wb->list_lock &sb->s_type->i_lock_key irq_context: 0 kn->active &____s->seqcount irq_context: 0 kn->active#2 &n->list_lock irq_context: 0 kn->active#2 &n->list_lock &c->lock irq_context: 0 sb_writers#8 &of->mutex kn->active#2 &____s->seqcount irq_context: 0 sb_writers#8 &of->mutex kn->active quarantine_lock irq_context: 0 sb_writers#8 &of->mutex kn->active &n->list_lock irq_context: 0 sb_writers#8 &of->mutex kn->active &n->list_lock &c->lock irq_context: 0 sb_writers#8 &type->i_mutex_dir_key#4 &root->kernfs_rwsem pool_lock#2 irq_context: 0 sb_writers#8 &of->mutex kn->active#2 uevent_sock_mutex &____s->seqcount irq_context: 0 kn->active &kernfs_locks->open_file_mutex[count] &____s->seqcount irq_context: 0 sb_writers#8 &of->mutex kn->active uevent_sock_mutex &____s->seqcount irq_context: 0 sb_writers#8 &type->i_mutex_dir_key#4 &root->kernfs_rwsem remove_cache_srcu irq_context: 0 sb_writers#8 &type->i_mutex_dir_key#4 &root->kernfs_rwsem remove_cache_srcu quarantine_lock irq_context: 0 sb_writers#8 &type->i_mutex_dir_key#4 &root->kernfs_rwsem remove_cache_srcu &c->lock irq_context: 0 sb_writers#8 &type->i_mutex_dir_key#4 &root->kernfs_rwsem remove_cache_srcu &n->list_lock irq_context: 0 sb_writers#8 &type->i_mutex_dir_key#4 &root->kernfs_rwsem remove_cache_srcu rcu_read_lock pool_lock#2 irq_context: 0 sb_writers#8 &type->i_mutex_dir_key#4 &root->kernfs_rwsem remove_cache_srcu &obj_hash[i].lock irq_context: 0 sb_writers#8 &type->i_mutex_dir_key#4 &n->list_lock irq_context: 0 sb_writers#8 &type->i_mutex_dir_key#4 &n->list_lock &c->lock irq_context: 0 sb_writers#8 &type->i_mutex_dir_key#4 &root->kernfs_rwsem &pcp->lock &zone->lock irq_context: 0 sb_writers#8 &type->i_mutex_dir_key#4 &root->kernfs_rwsem &pcp->lock &zone->lock &____s->seqcount irq_context: 0 kn->active#2 &kernfs_locks->open_file_mutex[count] &____s->seqcount irq_context: 0 dup_mmap_sem &mm->mmap_lock &mm->mmap_lock/1 &n->list_lock irq_context: 0 dup_mmap_sem &mm->mmap_lock &mm->mmap_lock/1 &n->list_lock &c->lock irq_context: 0 sb_writers#8 &n->list_lock irq_context: 0 sb_writers#8 &n->list_lock &c->lock irq_context: 0 kn->active#2 remove_cache_srcu irq_context: 0 kn->active#2 remove_cache_srcu quarantine_lock irq_context: 0 rcu_read_lock &ei->socket.wq.wait &ep->lock irq_context: 0 rcu_read_lock &ei->socket.wq.wait &ep->lock &ep->wq irq_context: 0 rcu_read_lock &ei->socket.wq.wait &ep->lock &ep->wq &p->pi_lock irq_context: 0 rcu_read_lock &ei->socket.wq.wait &ep->lock &ep->wq &p->pi_lock &rq->__lock irq_context: 0 rcu_read_lock &ei->socket.wq.wait &ep->lock &ep->wq &p->pi_lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 dup_mmap_sem &mm->mmap_lock &mm->mmap_lock/1 remove_cache_srcu irq_context: 0 kn->active#2 remove_cache_srcu &c->lock irq_context: 0 dup_mmap_sem &mm->mmap_lock &mm->mmap_lock/1 remove_cache_srcu quarantine_lock irq_context: 0 kn->active#2 remove_cache_srcu &n->list_lock irq_context: 0 kn->active#2 remove_cache_srcu rcu_read_lock pool_lock#2 irq_context: 0 kn->active#2 remove_cache_srcu &obj_hash[i].lock irq_context: 0 &ep->mtx &____s->seqcount irq_context: 0 &ep->mtx &c->lock irq_context: hardirq &dev->power.lock rcu_read_lock &pool->lock irq_context: hardirq &dev->power.lock rcu_read_lock &pool->lock &obj_hash[i].lock irq_context: hardirq &dev->power.lock rcu_read_lock &pool->lock pool_lock#2 irq_context: hardirq &dev->power.lock rcu_read_lock &pool->lock &p->pi_lock irq_context: hardirq &dev->power.lock rcu_read_lock &pool->lock &p->pi_lock &rq->__lock irq_context: hardirq &dev->power.lock rcu_read_lock &pool->lock &p->pi_lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 &ep->mtx &rq->__lock irq_context: 0 dup_mmap_sem &mm->mmap_lock &mm->mmap_lock/1 remove_cache_srcu &n->list_lock irq_context: 0 dup_mmap_sem &mm->mmap_lock &mm->mmap_lock/1 remove_cache_srcu rcu_read_lock pool_lock#2 irq_context: 0 dup_mmap_sem &mm->mmap_lock &mm->mmap_lock/1 remove_cache_srcu &obj_hash[i].lock irq_context: 0 dup_mmap_sem &mm->mmap_lock &mm->mmap_lock/1 remove_cache_srcu &c->lock irq_context: 0 (wq_completion)pm (work_completion)(&dev->power.work) &dev->power.lock rcu_read_lock &pool->lock irq_context: 0 (wq_completion)pm (work_completion)(&dev->power.work) &dev->power.lock rcu_read_lock &pool->lock &obj_hash[i].lock irq_context: 0 (wq_completion)pm (work_completion)(&dev->power.work) &dev->power.lock rcu_read_lock &pool->lock pool_lock#2 irq_context: 0 sb_writers#3 &sb->s_type->i_mutex_key#9 &c->lock irq_context: 0 sb_writers#3 &sb->s_type->i_mutex_key#9 &____s->seqcount irq_context: 0 sb_writers#3 &sb->s_type->i_mutex_key#9 pool_lock#2 irq_context: 0 sb_writers#8 &rq->__lock irq_context: 0 sb_writers#3 &sb->s_type->i_mutex_key#9 tomoyo_ss &c->lock irq_context: 0 sb_writers#3 &sb->s_type->i_mutex_key#9 tomoyo_ss &pcp->lock &zone->lock irq_context: 0 sb_writers#3 &sb->s_type->i_mutex_key#9 tomoyo_ss &pcp->lock &zone->lock &____s->seqcount irq_context: 0 sb_writers#3 &sb->s_type->i_mutex_key#9 tomoyo_ss &____s->seqcount irq_context: 0 sb_writers#8 &of->mutex kn->active#2 &pcp->lock &zone->lock irq_context: 0 sb_writers#8 &of->mutex kn->active#2 &pcp->lock &zone->lock &____s->seqcount irq_context: 0 sb_writers#8 &of->mutex kn->active#2 uevent_sock_mutex batched_entropy_u8.lock irq_context: 0 sb_writers#8 &of->mutex kn->active#2 uevent_sock_mutex kfence_freelist_lock irq_context: 0 sb_writers#8 &of->mutex kn->active#2 quarantine_lock irq_context: 0 &root->kernfs_iattr_rwsem &rq->__lock irq_context: 0 sb_writers#8 &of->mutex kn->active &pcp->lock &zone->lock irq_context: 0 sb_writers#8 &of->mutex kn->active &pcp->lock &zone->lock &____s->seqcount irq_context: 0 &kernfs_locks->open_file_mutex[count] fill_pool_map-wait-type-override pool_lock#2 irq_context: 0 &kernfs_locks->open_file_mutex[count] fill_pool_map-wait-type-override pool_lock irq_context: 0 &type->i_mutex_dir_key#5 &n->list_lock irq_context: 0 sb_writers#8 &of->mutex kn->active#2 uevent_sock_mutex rcu_read_lock rcu_node_0 irq_context: 0 sb_writers#8 &of->mutex kn->active#2 uevent_sock_mutex rcu_read_lock &rq->__lock irq_context: 0 &kernfs_locks->open_file_mutex[count] fill_pool_map-wait-type-override &c->lock irq_context: 0 &kernfs_locks->open_file_mutex[count] fill_pool_map-wait-type-override &____s->seqcount irq_context: 0 &type->i_mutex_dir_key#5 &n->list_lock &c->lock irq_context: 0 sb_writers#8 &type->i_mutex_dir_key#4 fs_reclaim mmu_notifier_invalidate_range_start &rq->__lock irq_context: 0 sb_writers#8 &of->mutex kn->active uevent_sock_mutex rcu_node_0 irq_context: 0 sb_writers#8 &of->mutex kn->active uevent_sock_mutex &rq->__lock irq_context: 0 sb_writers#8 &of->mutex kn->active uevent_sock_mutex rcu_read_lock rcu_node_0 irq_context: 0 dup_mmap_sem &mm->mmap_lock &mm->mmap_lock/1 remove_cache_srcu rcu_read_lock rcu_read_lock rcu_node_0 irq_context: 0 dup_mmap_sem &mm->mmap_lock &mm->mmap_lock/1 remove_cache_srcu rcu_read_lock rcu_read_lock &rq->__lock irq_context: 0 dup_mmap_sem &mm->mmap_lock &mm->mmap_lock/1 remove_cache_srcu rcu_read_lock rcu_node_0 irq_context: 0 &mm->mmap_lock &anon_vma->rwsem &sem->wait_lock irq_context: 0 sb_writers#8 &of->mutex kn->active#2 uevent_sock_mutex remove_cache_srcu irq_context: 0 sb_writers#8 &of->mutex kn->active#2 uevent_sock_mutex remove_cache_srcu quarantine_lock irq_context: 0 sb_writers#8 &of->mutex kn->active#2 uevent_sock_mutex remove_cache_srcu &c->lock irq_context: 0 sb_writers#8 &of->mutex kn->active#2 uevent_sock_mutex remove_cache_srcu &n->list_lock irq_context: 0 sb_writers#8 &of->mutex kn->active#2 uevent_sock_mutex remove_cache_srcu rcu_read_lock pool_lock#2 irq_context: 0 sb_writers#8 &of->mutex kn->active#2 uevent_sock_mutex remove_cache_srcu &obj_hash[i].lock irq_context: 0 sb_writers#8 &of->mutex kn->active#2 uevent_sock_mutex &rq->__lock irq_context: 0 sb_writers#8 &of->mutex kn->active#2 uevent_sock_mutex &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 sb_writers#8 &of->mutex kn->active#2 &rq->__lock irq_context: 0 sb_writers#8 &of->mutex kn->active#2 uevent_sock_mutex &meta->lock irq_context: 0 sb_writers#8 &type->i_mutex_dir_key#4 &rq->__lock irq_context: 0 rcu_read_lock &ei->socket.wq.wait &ep->lock &ep->wq &p->pi_lock &cfs_rq->removed.lock irq_context: 0 tomoyo_ss rcu_read_lock rcu_read_lock rcu_node_0 irq_context: 0 tomoyo_ss rcu_read_lock rcu_read_lock &rq->__lock irq_context: 0 tomoyo_ss rcu_read_lock rcu_node_0 irq_context: 0 sb_writers#8 &of->mutex kn->active &rq->__lock irq_context: 0 sb_writers#5 &type->i_mutex_dir_key#5 &sem->wait_lock irq_context: 0 sb_writers#5 &rq->__lock irq_context: 0 kn->active#2 &rq->__lock irq_context: 0 &type->i_mutex_dir_key#5 rcu_read_lock &dentry->d_lock irq_context: 0 kn->active#2 &kernfs_locks->open_file_mutex[count] &n->list_lock irq_context: 0 kn->active#2 &kernfs_locks->open_file_mutex[count] &n->list_lock &c->lock irq_context: 0 &type->i_mutex_dir_key#5 &sem->wait_lock irq_context: 0 &type->i_mutex_dir_key#5 &obj_hash[i].lock irq_context: 0 sb_writers#8 &type->i_mutex_dir_key#4 &root->kernfs_rwsem &rq->__lock irq_context: 0 sb_writers#8 &type->i_mutex_dir_key#4 remove_cache_srcu irq_context: 0 sb_writers#8 &type->i_mutex_dir_key#4 remove_cache_srcu quarantine_lock irq_context: 0 sb_writers#8 &type->i_mutex_dir_key#4 remove_cache_srcu &c->lock irq_context: 0 sb_writers#8 &type->i_mutex_dir_key#4 remove_cache_srcu &n->list_lock irq_context: 0 sb_writers#8 &type->i_mutex_dir_key#4 remove_cache_srcu rcu_read_lock pool_lock#2 irq_context: 0 sb_writers#8 &type->i_mutex_dir_key#4 remove_cache_srcu &obj_hash[i].lock irq_context: 0 sb_writers#8 &type->i_mutex_dir_key#4 remove_cache_srcu &rq->__lock irq_context: 0 sb_writers#8 &type->i_mutex_dir_key#4 remove_cache_srcu rcu_read_lock rcu_node_0 irq_context: 0 sb_writers#8 &type->i_mutex_dir_key#4 remove_cache_srcu rcu_read_lock &rq->__lock irq_context: 0 &kernfs_locks->open_file_mutex[count] &rq->__lock irq_context: 0 (wq_completion)events (work_completion)(&(&krcp->monitor_work)->work) krc.lock &obj_hash[i].lock irq_context: 0 kn->active#2 &kernfs_locks->open_file_mutex[count] &rq->__lock irq_context: 0 sb_writers#8 &type->i_mutex_dir_key#4 &root->kernfs_rwsem rcu_read_lock rcu_node_0 irq_context: 0 sb_writers#8 &type->i_mutex_dir_key#4 &root->kernfs_rwsem rcu_read_lock &rq->__lock irq_context: 0 rcu_read_lock &rcu_state.gp_wq &p->pi_lock &cfs_rq->removed.lock irq_context: 0 rcu_read_lock &ei->socket.wq.wait &ep->lock &ep->wq &p->pi_lock &rq->__lock &cfs_rq->removed.lock irq_context: 0 (wq_completion)events (work_completion)(&(&krcp->krw_arr[i].rcu_work)->work) irq_context: 0 (wq_completion)events (work_completion)(&(&krcp->krw_arr[i].rcu_work)->work) krc.lock irq_context: 0 (wq_completion)events (work_completion)(&(&krcp->krw_arr[i].rcu_work)->work) &obj_hash[i].lock irq_context: 0 (wq_completion)events (work_completion)(&(&krcp->krw_arr[i].rcu_work)->work) rcu_callback &obj_hash[i].lock irq_context: 0 (wq_completion)events (work_completion)(&(&krcp->krw_arr[i].rcu_work)->work) rcu_callback pool_lock#2 irq_context: 0 (wq_completion)events (work_completion)(&(&krcp->krw_arr[i].rcu_work)->work) rcu_callback &obj_hash[i].lock pool_lock irq_context: 0 &sem->wait_lock irq_context: 0 sb_writers#5 &type->i_mutex_dir_key#5 &rq->__lock irq_context: 0 sb_writers#5 &type->i_mutex_dir_key#5 &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 remove_cache_srcu rcu_read_lock rcu_node_0 irq_context: 0 remove_cache_srcu rcu_read_lock &rq->__lock irq_context: 0 remove_cache_srcu &rq->__lock irq_context: 0 sb_writers#8 &of->mutex kn->active#2 fs_reclaim &rq->__lock irq_context: 0 sb_writers#8 &of->mutex kn->active#2 uevent_sock_mutex rcu_read_lock &rcu_state.gp_wq irq_context: 0 sb_writers#8 &of->mutex kn->active#2 uevent_sock_mutex rcu_read_lock &rcu_state.gp_wq &p->pi_lock irq_context: 0 sb_writers#8 &of->mutex kn->active#2 uevent_sock_mutex rcu_read_lock &rcu_state.gp_wq &p->pi_lock &cfs_rq->removed.lock irq_context: 0 sb_writers#8 &of->mutex kn->active#2 uevent_sock_mutex rcu_read_lock &rcu_state.gp_wq &p->pi_lock &rq->__lock irq_context: 0 sb_writers#8 &of->mutex kn->active#2 uevent_sock_mutex rcu_read_lock &rcu_state.gp_wq &p->pi_lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 sb_writers#5 &sb->s_type->i_mutex_key#12 &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 &type->i_mutex_dir_key#5 rcu_read_lock rcu_node_0 irq_context: 0 &type->i_mutex_dir_key#5 rcu_read_lock &rq->__lock irq_context: 0 sb_writers#5 &type->i_mutex_dir_key#5 rcu_read_lock &rq->__lock irq_context: 0 sb_writers#5 &type->i_mutex_dir_key#5 rcu_read_lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 &type->i_mutex_dir_key#5 &rq->__lock irq_context: 0 &type->i_mutex_dir_key#5 &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 &type->i_mutex_dir_key#5 rcu_node_0 irq_context: 0 sb_writers#5 &sem->wait_lock irq_context: 0 sb_writers#5 &p->pi_lock irq_context: 0 sb_writers#5 &p->pi_lock &rq->__lock irq_context: 0 sb_writers#5 &p->pi_lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 kn->active#2 &kernfs_locks->open_file_mutex[count] remove_cache_srcu irq_context: 0 kn->active#2 &kernfs_locks->open_file_mutex[count] remove_cache_srcu quarantine_lock irq_context: 0 kn->active#2 &kernfs_locks->open_file_mutex[count] remove_cache_srcu &c->lock irq_context: 0 kn->active#2 &kernfs_locks->open_file_mutex[count] remove_cache_srcu &n->list_lock irq_context: 0 kn->active#2 &kernfs_locks->open_file_mutex[count] remove_cache_srcu rcu_read_lock pool_lock#2 irq_context: 0 kn->active#2 &kernfs_locks->open_file_mutex[count] remove_cache_srcu &obj_hash[i].lock irq_context: 0 kn->active#2 &kernfs_locks->open_file_mutex[count] remove_cache_srcu &rq->__lock irq_context: 0 sb_writers#8 remove_cache_srcu irq_context: 0 sb_writers#8 remove_cache_srcu quarantine_lock irq_context: 0 sb_writers#8 remove_cache_srcu &c->lock irq_context: 0 sb_writers#8 remove_cache_srcu &n->list_lock irq_context: 0 sb_writers#8 remove_cache_srcu rcu_read_lock pool_lock#2 irq_context: 0 sb_writers#8 remove_cache_srcu &obj_hash[i].lock irq_context: 0 sb_writers#8 remove_cache_srcu rcu_read_lock rcu_node_0 irq_context: 0 sb_writers#8 remove_cache_srcu rcu_read_lock &rq->__lock irq_context: 0 &root->kernfs_iattr_rwsem &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 tomoyo_ss remove_cache_srcu &rq->__lock irq_context: 0 sb_writers#8 &of->mutex kn->active#2 uevent_sock_mutex fs_reclaim &rq->__lock irq_context: 0 &type->i_mutex_dir_key#5 &dentry->d_lock &dentry->d_lock/1 irq_context: 0 &type->i_mutex_dir_key#5 &dentry->d_lock/1 irq_context: 0 sb_writers#8 &of->mutex kn->active#2 uevent_sock_mutex &n->list_lock irq_context: 0 tomoyo_ss rcu_read_lock &rq->__lock irq_context: 0 kn->active#3 fs_reclaim irq_context: 0 kn->active#3 fs_reclaim mmu_notifier_invalidate_range_start irq_context: 0 kn->active#3 &kernfs_locks->open_file_mutex[count] irq_context: 0 kn->active#3 &kernfs_locks->open_file_mutex[count] fs_reclaim irq_context: 0 kn->active#3 &kernfs_locks->open_file_mutex[count] fs_reclaim mmu_notifier_invalidate_range_start irq_context: 0 kn->active#3 &kernfs_locks->open_file_mutex[count] &c->lock irq_context: 0 kn->active#3 &kernfs_locks->open_file_mutex[count] &____s->seqcount irq_context: 0 sb_writers#8 &of->mutex kn->active#3 fs_reclaim irq_context: 0 sb_writers#8 &of->mutex kn->active#3 fs_reclaim mmu_notifier_invalidate_range_start irq_context: 0 sb_writers#8 &of->mutex kn->active#3 pool_lock#2 irq_context: 0 sb_writers#8 &of->mutex kn->active#3 uevent_sock_mutex irq_context: 0 sb_writers#8 &of->mutex kn->active#3 uevent_sock_mutex fs_reclaim irq_context: 0 sb_writers#8 &of->mutex kn->active#3 uevent_sock_mutex fs_reclaim mmu_notifier_invalidate_range_start irq_context: 0 sb_writers#8 &of->mutex kn->active#3 uevent_sock_mutex pool_lock#2 irq_context: 0 sb_writers#8 &of->mutex kn->active#3 uevent_sock_mutex nl_table_lock irq_context: 0 sb_writers#8 &of->mutex kn->active#3 uevent_sock_mutex rlock-AF_NETLINK irq_context: 0 sb_writers#8 &of->mutex kn->active#3 uevent_sock_mutex rcu_read_lock &ei->socket.wq.wait irq_context: 0 sb_writers#8 &of->mutex kn->active#3 uevent_sock_mutex rcu_read_lock &ei->socket.wq.wait &ep->lock irq_context: 0 sb_writers#8 &of->mutex kn->active#3 uevent_sock_mutex nl_table_wait.lock irq_context: 0 sb_writers#8 &of->mutex kn->active#3 uevent_sock_mutex &obj_hash[i].lock irq_context: 0 sb_writers#8 &of->mutex kn->active#3 &obj_hash[i].lock irq_context: 0 kn->active#3 &c->lock irq_context: 0 kn->active#3 &____s->seqcount irq_context: 0 sb_writers#8 &of->mutex kn->active#3 uevent_sock_mutex &c->lock irq_context: 0 sb_writers#8 &of->mutex kn->active#3 uevent_sock_mutex &____s->seqcount irq_context: 0 sb_writers#8 &of->mutex kn->active#3 &rq->__lock irq_context: 0 sb_writers#8 &of->mutex kn->active#3 &c->lock irq_context: 0 &type->i_mutex_dir_key#5 remove_cache_srcu irq_context: 0 kn->active#3 &n->list_lock irq_context: 0 &type->i_mutex_dir_key#5 remove_cache_srcu quarantine_lock irq_context: 0 kn->active#3 &n->list_lock &c->lock irq_context: 0 &type->i_mutex_dir_key#5 remove_cache_srcu &c->lock irq_context: 0 &type->i_mutex_dir_key#5 remove_cache_srcu &n->list_lock irq_context: 0 &type->i_mutex_dir_key#5 remove_cache_srcu rcu_read_lock pool_lock#2 irq_context: 0 &type->i_mutex_dir_key#5 remove_cache_srcu &obj_hash[i].lock irq_context: 0 sb_writers#8 &of->mutex kn->active#3 &n->list_lock irq_context: 0 sb_writers#8 &of->mutex kn->active#3 &n->list_lock &c->lock irq_context: 0 sb_writers#8 &of->mutex kn->active#3 &____s->seqcount irq_context: 0 kn->active#3 remove_cache_srcu irq_context: 0 kn->active#3 remove_cache_srcu quarantine_lock irq_context: 0 sb_writers#8 &of->mutex kn->active#3 rcu_read_lock rcu_node_0 irq_context: 0 sb_writers#8 &of->mutex kn->active#3 rcu_read_lock &rq->__lock irq_context: 0 sb_writers#8 rcu_node_0 irq_context: 0 sb_writers#8 rcu_read_lock rcu_node_0 irq_context: 0 kn->active#3 &rq->__lock irq_context: 0 sb_writers#8 &of->mutex kn->active#3 uevent_sock_mutex &rq->__lock irq_context: 0 sb_writers#8 &of->mutex kn->active#3 quarantine_lock irq_context: 0 kn->active#3 &kernfs_locks->open_file_mutex[count] fs_reclaim mmu_notifier_invalidate_range_start &rq->__lock irq_context: 0 kn->active#3 remove_cache_srcu &c->lock irq_context: 0 kn->active#3 remove_cache_srcu &n->list_lock irq_context: 0 kn->active#3 remove_cache_srcu &obj_hash[i].lock irq_context: 0 sb_writers#8 &of->mutex kn->active#3 uevent_sock_mutex &n->list_lock irq_context: 0 sb_writers#8 &of->mutex kn->active#3 uevent_sock_mutex &n->list_lock &c->lock irq_context: 0 (wq_completion)events (work_completion)(&(&krcp->krw_arr[i].rcu_work)->work) rcu_callback &rq->__lock irq_context: 0 sb_writers#8 &of->mutex kn->active#3 uevent_sock_mutex rcu_read_lock rcu_node_0 irq_context: 0 sb_writers#8 &of->mutex kn->active#3 uevent_sock_mutex rcu_read_lock &rq->__lock irq_context: 0 sb_writers#8 &of->mutex kn->active#3 uevent_sock_mutex rcu_read_lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 kn->active#3 remove_cache_srcu &rq->__lock irq_context: 0 tomoyo_ss remove_cache_srcu &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 kn->active#3 &kernfs_locks->open_file_mutex[count] &rq->__lock irq_context: 0 &vma->vm_lock->lock rcu_read_lock per_cpu_ptr(&cgroup_rstat_cpu_lock, cpu) irq_context: 0 &mm->mmap_lock rcu_read_lock &rcu_state.gp_wq &p->pi_lock &cfs_rq->removed.lock irq_context: 0 &sig->cred_guard_mutex tomoyo_ss &n->list_lock irq_context: 0 &sig->cred_guard_mutex tomoyo_ss &n->list_lock &c->lock irq_context: 0 &mm->mmap_lock &n->list_lock &c->lock irq_context: 0 &mm->mmap_lock fs_reclaim &rq->__lock irq_context: 0 &mm->mmap_lock &mapping->i_mmap_rwsem rcu_read_lock rcu_node_0 irq_context: 0 &mm->mmap_lock &mapping->i_mmap_rwsem rcu_read_lock &rq->__lock irq_context: 0 &mm->mmap_lock remove_cache_srcu irq_context: 0 &mm->mmap_lock remove_cache_srcu quarantine_lock irq_context: 0 &mm->mmap_lock rcu_node_0 irq_context: 0 &mm->mmap_lock &anon_vma->rwsem &rq->__lock irq_context: 0 &sb->s_type->i_lock_key#24 irq_context: 0 &type->i_mutex_dir_key#4 &rq->__lock irq_context: 0 &type->i_mutex_dir_key#4 &n->list_lock irq_context: 0 &type->i_mutex_dir_key#4 &n->list_lock &c->lock irq_context: 0 &f->f_pos_lock &type->i_mutex_dir_key#4 &mm->mmap_lock rcu_read_lock rcu_node_0 irq_context: 0 &f->f_pos_lock &type->i_mutex_dir_key#4 &mm->mmap_lock rcu_read_lock &rq->__lock irq_context: 0 kn->active#4 fs_reclaim irq_context: 0 kn->active#4 fs_reclaim mmu_notifier_invalidate_range_start irq_context: 0 kn->active#4 &kernfs_locks->open_file_mutex[count] irq_context: 0 kn->active#4 &kernfs_locks->open_file_mutex[count] fs_reclaim irq_context: 0 kn->active#4 &kernfs_locks->open_file_mutex[count] fs_reclaim mmu_notifier_invalidate_range_start irq_context: 0 &p->lock &of->mutex irq_context: 0 &p->lock &of->mutex kn->active#4 param_lock irq_context: 0 &mm->mmap_lock rcu_read_lock per_cpu_ptr(&cgroup_rstat_cpu_lock, cpu) irq_context: 0 sb_writers#8 &type->i_mutex_dir_key#4 irq_context: 0 sb_writers#8 &type->i_mutex_dir_key#4 rename_lock.seqcount irq_context: 0 sb_writers#8 &type->i_mutex_dir_key#4 rcu_read_lock &dentry->d_lock irq_context: 0 sb_writers#8 &type->i_mutex_dir_key#4 &root->kernfs_rwsem irq_context: 0 sb_writers#8 &root->kernfs_iattr_rwsem irq_context: 0 sb_writers#8 &dentry->d_lock irq_context: 0 sb_writers#8 tomoyo_ss irq_context: 0 sb_writers#8 tomoyo_ss mmu_notifier_invalidate_range_start irq_context: 0 sb_writers#8 tomoyo_ss &c->lock irq_context: 0 sb_writers#8 tomoyo_ss &____s->seqcount irq_context: 0 sb_writers#8 tomoyo_ss rcu_read_lock mount_lock.seqcount irq_context: 0 sb_writers#8 tomoyo_ss rcu_read_lock rcu_read_lock rename_lock.seqcount irq_context: 0 sb_writers#8 tomoyo_ss &obj_hash[i].lock irq_context: 0 sb_writers#8 kn->active#4 fs_reclaim irq_context: 0 sb_writers#8 kn->active#4 fs_reclaim mmu_notifier_invalidate_range_start irq_context: 0 sb_writers#8 kn->active#4 &kernfs_locks->open_file_mutex[count] irq_context: 0 sb_writers#8 kn->active#4 &kernfs_locks->open_file_mutex[count] fs_reclaim irq_context: 0 sb_writers#8 kn->active#4 &kernfs_locks->open_file_mutex[count] fs_reclaim mmu_notifier_invalidate_range_start irq_context: 0 sb_writers#8 &sb->s_type->i_mutex_key#13 irq_context: 0 sb_writers#8 iattr_mutex irq_context: 0 sb_writers#8 &sb->s_type->i_mutex_key#13 tk_core.seq.seqcount irq_context: 0 sb_writers#8 &sb->s_type->i_mutex_key#13 &root->kernfs_iattr_rwsem irq_context: 0 sb_writers#8 &sb->s_type->i_mutex_key#13 &root->kernfs_iattr_rwsem iattr_mutex irq_context: 0 sb_writers#8 &sb->s_type->i_mutex_key#13 &root->kernfs_iattr_rwsem iattr_mutex fs_reclaim irq_context: 0 &root->kernfs_iattr_rwsem &sem->wait_lock irq_context: 0 sb_writers#8 &sb->s_type->i_mutex_key#13 &root->kernfs_iattr_rwsem iattr_mutex fs_reclaim mmu_notifier_invalidate_range_start irq_context: 0 &root->kernfs_iattr_rwsem rcu_read_lock &rq->__lock irq_context: 0 sb_writers#8 &sb->s_type->i_mutex_key#13 &root->kernfs_iattr_rwsem iattr_mutex &____s->seqcount irq_context: 0 &root->kernfs_iattr_rwsem rcu_read_lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 sb_writers#8 &sb->s_type->i_mutex_key#13 &root->kernfs_iattr_rwsem iattr_mutex pool_lock#2 irq_context: 0 sb_writers#8 &sb->s_type->i_mutex_key#13 &root->kernfs_iattr_rwsem iattr_mutex &rq->__lock irq_context: 0 sb_writers#8 &sb->s_type->i_mutex_key#13 &root->kernfs_iattr_rwsem iattr_mutex &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 sb_writers#8 &sb->s_type->i_mutex_key#13 &root->kernfs_iattr_rwsem iattr_mutex &c->lock irq_context: 0 sb_writers#8 &sb->s_type->i_mutex_key#13 &root->kernfs_iattr_rwsem iattr_mutex tk_core.seq.seqcount irq_context: 0 sb_writers#8 &sb->s_type->i_mutex_key#13 &sem->wait_lock irq_context: 0 sb_writers#8 &sb->s_type->i_mutex_key#13 &p->pi_lock irq_context: 0 sb_writers#8 &sb->s_type->i_mutex_key#13 &p->pi_lock &rq->__lock irq_context: 0 sb_writers#8 &sb->s_type->i_mutex_key#13 &p->pi_lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 sb_writers#8 &of->mutex kn->active#4 param_lock irq_context: 0 sb_writers#8 &of->mutex kn->active#4 param_lock disk_events_mutex irq_context: 0 &type->i_mutex_dir_key#4 &root->kernfs_rwsem remove_cache_srcu irq_context: 0 &type->i_mutex_dir_key#4 &root->kernfs_rwsem remove_cache_srcu quarantine_lock irq_context: 0 &type->i_mutex_dir_key#4 rcu_read_lock rcu_node_0 irq_context: 0 &type->i_mutex_dir_key#4 rcu_read_lock &rq->__lock irq_context: 0 &type->i_mutex_dir_key#4 rcu_node_0 irq_context: 0 &type->i_mutex_dir_key#4 &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 &type->i_mutex_dir_key#4 &root->kernfs_rwsem remove_cache_srcu &c->lock irq_context: 0 &type->i_mutex_dir_key#4 &root->kernfs_rwsem remove_cache_srcu &n->list_lock irq_context: 0 &type->i_mutex_dir_key#4 &root->kernfs_rwsem remove_cache_srcu rcu_read_lock pool_lock#2 irq_context: 0 &type->i_mutex_dir_key#4 &root->kernfs_rwsem remove_cache_srcu &obj_hash[i].lock irq_context: 0 &type->i_mutex_dir_key#4 &root->kernfs_rwsem mmu_notifier_invalidate_range_start &rq->__lock irq_context: 0 &type->i_mutex_dir_key#5 &pcp->lock &zone->lock irq_context: 0 &type->i_mutex_dir_key#5 &pcp->lock &zone->lock &____s->seqcount irq_context: 0 &type->i_mutex_dir_key#4 fs_reclaim &rq->__lock irq_context: 0 &type->i_mutex_dir_key#4 &root->kernfs_rwsem &rq->__lock irq_context: 0 sb_writers#8 tk_core.seq.seqcount irq_context: 0 sb_writers#8 &sb->s_type->i_lock_key#24 irq_context: 0 sb_writers#8 &wb->list_lock irq_context: 0 sb_writers#8 &wb->list_lock &sb->s_type->i_lock_key#24 irq_context: 0 remove_cache_srcu rcu_read_lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 &type->i_mutex_dir_key#4 fs_reclaim mmu_notifier_invalidate_range_start &rq->__lock irq_context: 0 &f->f_pos_lock &type->i_mutex_dir_key#4 &root->kernfs_rwsem &rq->__lock irq_context: 0 &type->i_mutex_dir_key#4 &root->kernfs_rwsem rcu_read_lock rcu_node_0 irq_context: 0 &type->i_mutex_dir_key#4 &root->kernfs_rwsem rcu_read_lock &rq->__lock irq_context: 0 &type->i_mutex_dir_key#4 &root->kernfs_rwsem &n->list_lock irq_context: 0 &type->i_mutex_dir_key#4 &root->kernfs_rwsem &n->list_lock &c->lock irq_context: 0 &type->i_mutex_dir_key#5 rcu_read_lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 &f->f_pos_lock &mm->mmap_lock &rq->__lock irq_context: 0 &f->f_pos_lock &type->i_mutex_dir_key#4 &rq->__lock irq_context: 0 tomoyo_ss remove_cache_srcu rcu_read_lock rcu_node_0 irq_context: 0 tomoyo_ss remove_cache_srcu rcu_read_lock &rq->__lock irq_context: 0 &type->i_mutex_dir_key#4 &root->kernfs_rwsem remove_cache_srcu rcu_read_lock &____s->seqcount irq_context: 0 &type->i_mutex_dir_key#4 remove_cache_srcu &rq->__lock irq_context: 0 &u->iolock &meta->lock irq_context: 0 &u->iolock kfence_freelist_lock irq_context: 0 tomoyo_ss rcu_node_0 irq_context: 0 sb_writers#5 &type->i_mutex_dir_key#5/1 &sb->s_type->i_mutex_key#12 rcu_read_lock &dentry->d_lock irq_context: 0 sb_writers#5 &type->i_mutex_dir_key#5/1 &fsnotify_mark_srcu irq_context: 0 sb_writers#5 &s->s_inode_list_lock irq_context: 0 sb_writers#5 &info->lock irq_context: 0 sb_writers#5 &sbinfo->stat_lock irq_context: 0 sb_writers#5 &xa->xa_lock#9 irq_context: 0 sb_writers#5 &obj_hash[i].lock irq_context: 0 sb_writers#5 pool_lock#2 irq_context: 0 sb_writers#5 &fsnotify_mark_srcu irq_context: 0 kn->active#5 fs_reclaim irq_context: 0 kn->active#5 fs_reclaim mmu_notifier_invalidate_range_start irq_context: 0 kn->active#5 &kernfs_locks->open_file_mutex[count] irq_context: 0 kn->active#5 &kernfs_locks->open_file_mutex[count] fs_reclaim irq_context: 0 kn->active#5 &kernfs_locks->open_file_mutex[count] fs_reclaim mmu_notifier_invalidate_range_start irq_context: 0 sb_writers#8 &of->mutex kn->active#5 fs_reclaim irq_context: 0 sb_writers#8 &of->mutex kn->active#5 fs_reclaim mmu_notifier_invalidate_range_start irq_context: 0 sb_writers#8 &of->mutex kn->active#5 pool_lock#2 irq_context: 0 sb_writers#8 &of->mutex kn->active#5 uevent_sock_mutex irq_context: 0 sb_writers#8 &of->mutex kn->active#5 uevent_sock_mutex fs_reclaim irq_context: 0 sb_writers#8 &of->mutex kn->active#5 uevent_sock_mutex fs_reclaim mmu_notifier_invalidate_range_start irq_context: 0 sb_writers#8 &of->mutex kn->active#5 uevent_sock_mutex pool_lock#2 irq_context: 0 sb_writers#8 &of->mutex kn->active#5 uevent_sock_mutex nl_table_lock irq_context: 0 sb_writers#8 &of->mutex kn->active#5 uevent_sock_mutex rlock-AF_NETLINK irq_context: 0 sb_writers#8 &of->mutex kn->active#5 uevent_sock_mutex rcu_read_lock &ei->socket.wq.wait irq_context: 0 sb_writers#8 &of->mutex kn->active#5 uevent_sock_mutex rcu_read_lock &ei->socket.wq.wait &ep->lock irq_context: 0 sb_writers#8 &of->mutex kn->active#5 uevent_sock_mutex rcu_read_lock &ei->socket.wq.wait &ep->lock &ep->wq irq_context: 0 sb_writers#8 &of->mutex kn->active#5 uevent_sock_mutex rcu_read_lock &ei->socket.wq.wait &ep->lock &ep->wq &p->pi_lock irq_context: 0 sb_writers#8 &of->mutex kn->active#5 uevent_sock_mutex rcu_read_lock &ei->socket.wq.wait &ep->lock &ep->wq &p->pi_lock &rq->__lock irq_context: 0 sb_writers#8 &of->mutex kn->active#5 uevent_sock_mutex rcu_read_lock &ei->socket.wq.wait &ep->lock &ep->wq &p->pi_lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 sb_writers#8 &of->mutex kn->active#5 uevent_sock_mutex rcu_read_lock rcu_node_0 irq_context: 0 sb_writers#8 &of->mutex kn->active#5 uevent_sock_mutex rcu_read_lock &rq->__lock irq_context: 0 sb_writers#8 &of->mutex kn->active#5 uevent_sock_mutex nl_table_wait.lock irq_context: 0 sb_writers#8 &of->mutex kn->active#5 uevent_sock_mutex &obj_hash[i].lock irq_context: 0 sb_writers#8 &of->mutex kn->active#5 &obj_hash[i].lock irq_context: 0 &p->lock &of->mutex kn->active#5 fs_reclaim irq_context: 0 &p->lock &of->mutex kn->active#5 fs_reclaim mmu_notifier_invalidate_range_start irq_context: 0 &p->lock &of->mutex kn->active#5 pool_lock#2 irq_context: 0 &p->lock &of->mutex kn->active#5 &obj_hash[i].lock irq_context: 0 sb_writers#8 &of->mutex kn->active#5 uevent_sock_mutex &c->lock irq_context: 0 kn->active#5 &c->lock irq_context: 0 sb_writers#8 &of->mutex kn->active#5 remove_cache_srcu irq_context: 0 kn->active#5 remove_cache_srcu irq_context: 0 sb_writers#8 &of->mutex kn->active#5 remove_cache_srcu quarantine_lock irq_context: 0 kn->active#5 remove_cache_srcu quarantine_lock irq_context: 0 sb_writers#8 &of->mutex kn->active#5 remove_cache_srcu &c->lock irq_context: 0 sb_writers#8 &of->mutex kn->active#5 remove_cache_srcu &n->list_lock irq_context: 0 sb_writers#8 &of->mutex kn->active#5 remove_cache_srcu rcu_read_lock pool_lock#2 irq_context: 0 sb_writers#8 &of->mutex kn->active#5 remove_cache_srcu &obj_hash[i].lock irq_context: 0 &p->lock &of->mutex kn->active#5 &c->lock irq_context: 0 &p->lock &of->mutex kn->active#5 &n->list_lock irq_context: 0 &p->lock &of->mutex kn->active#5 &n->list_lock &c->lock irq_context: 0 kn->active#5 &kernfs_locks->open_file_mutex[count] &c->lock irq_context: 0 sb_writers#8 &of->mutex kn->active#5 uevent_sock_mutex &rq->__lock irq_context: 0 kn->active#5 &kernfs_locks->open_file_mutex[count] &rq->__lock irq_context: 0 &p->lock &of->mutex kn->active#5 &____s->seqcount irq_context: 0 kn->active#5 &kernfs_locks->open_file_mutex[count] &____s->seqcount irq_context: 0 sb_writers#8 &of->mutex kn->active#5 uevent_sock_mutex &____s->seqcount irq_context: 0 &p->lock &n->list_lock irq_context: 0 &p->lock &n->list_lock &c->lock irq_context: 0 sb_writers#8 &of->mutex kn->active#5 uevent_sock_mutex &n->list_lock irq_context: 0 sb_writers#8 &of->mutex kn->active#5 uevent_sock_mutex &n->list_lock &c->lock irq_context: 0 kn->active#6 fs_reclaim irq_context: 0 kn->active#6 fs_reclaim mmu_notifier_invalidate_range_start irq_context: 0 kn->active#6 &kernfs_locks->open_file_mutex[count] irq_context: 0 kn->active#6 &kernfs_locks->open_file_mutex[count] fs_reclaim irq_context: 0 kn->active#6 &kernfs_locks->open_file_mutex[count] fs_reclaim mmu_notifier_invalidate_range_start irq_context: 0 kn->active#6 &kernfs_locks->open_file_mutex[count] &c->lock irq_context: 0 kn->active#5 &____s->seqcount irq_context: 0 kn->active#7 fs_reclaim irq_context: 0 kn->active#7 fs_reclaim mmu_notifier_invalidate_range_start irq_context: 0 kn->active#7 &c->lock irq_context: 0 kn->active#7 &n->list_lock irq_context: 0 kn->active#7 &n->list_lock &c->lock irq_context: 0 kn->active#7 &kernfs_locks->open_file_mutex[count] irq_context: 0 kn->active#7 &kernfs_locks->open_file_mutex[count] fs_reclaim irq_context: 0 kn->active#7 &kernfs_locks->open_file_mutex[count] fs_reclaim mmu_notifier_invalidate_range_start irq_context: 0 kn->active#8 fs_reclaim irq_context: 0 kn->active#8 fs_reclaim mmu_notifier_invalidate_range_start irq_context: 0 kn->active#8 &kernfs_locks->open_file_mutex[count] irq_context: 0 kn->active#8 &kernfs_locks->open_file_mutex[count] fs_reclaim irq_context: 0 kn->active#8 &kernfs_locks->open_file_mutex[count] fs_reclaim mmu_notifier_invalidate_range_start irq_context: 0 kn->active#9 fs_reclaim irq_context: 0 kn->active#9 fs_reclaim &rq->__lock irq_context: 0 kn->active#9 fs_reclaim &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 &type->i_mutex_dir_key#5 remove_cache_srcu &rq->__lock irq_context: 0 kn->active#5 &kernfs_locks->open_file_mutex[count] remove_cache_srcu irq_context: 0 kn->active#5 &kernfs_locks->open_file_mutex[count] remove_cache_srcu quarantine_lock irq_context: 0 kn->active#7 &kernfs_locks->open_file_mutex[count] &c->lock irq_context: 0 kn->active#8 &c->lock irq_context: 0 kn->active#9 fs_reclaim mmu_notifier_invalidate_range_start irq_context: 0 kn->active#9 &kernfs_locks->open_file_mutex[count] irq_context: 0 kn->active#9 &kernfs_locks->open_file_mutex[count] fs_reclaim irq_context: 0 kn->active#9 &kernfs_locks->open_file_mutex[count] fs_reclaim mmu_notifier_invalidate_range_start irq_context: 0 kn->active#10 fs_reclaim irq_context: 0 kn->active#10 fs_reclaim mmu_notifier_invalidate_range_start irq_context: 0 kn->active#10 &kernfs_locks->open_file_mutex[count] irq_context: 0 kn->active#10 &kernfs_locks->open_file_mutex[count] fs_reclaim irq_context: 0 kn->active#10 &kernfs_locks->open_file_mutex[count] fs_reclaim mmu_notifier_invalidate_range_start irq_context: 0 kn->active#11 fs_reclaim irq_context: 0 kn->active#11 fs_reclaim mmu_notifier_invalidate_range_start irq_context: 0 kn->active#11 &kernfs_locks->open_file_mutex[count] irq_context: 0 sb_writers#8 &of->mutex kn->active#5 &c->lock irq_context: 0 kn->active#11 &kernfs_locks->open_file_mutex[count] fs_reclaim irq_context: 0 kn->active#11 &kernfs_locks->open_file_mutex[count] fs_reclaim mmu_notifier_invalidate_range_start irq_context: 0 kn->active#5 &n->list_lock irq_context: 0 kn->active#5 &n->list_lock &c->lock irq_context: 0 kn->active#12 fs_reclaim irq_context: 0 kn->active#12 fs_reclaim mmu_notifier_invalidate_range_start irq_context: 0 kn->active#12 &kernfs_locks->open_file_mutex[count] irq_context: 0 kn->active#12 &kernfs_locks->open_file_mutex[count] fs_reclaim irq_context: 0 kn->active#12 &kernfs_locks->open_file_mutex[count] fs_reclaim mmu_notifier_invalidate_range_start irq_context: 0 sb_writers#8 &of->mutex kn->active#5 &____s->seqcount irq_context: 0 sb_writers#8 &of->mutex kn->active#5 &device->physical_node_lock irq_context: 0 sb_writers#8 remove_cache_srcu &rq->__lock irq_context: 0 sb_writers#5 &type->i_mutex_dir_key#5 tomoyo_ss &n->list_lock irq_context: 0 sb_writers#5 &type->i_mutex_dir_key#5 tomoyo_ss &n->list_lock &c->lock irq_context: 0 &p->lock &of->mutex kn->active#5 &pcp->lock &zone->lock irq_context: 0 &p->lock &of->mutex kn->active#5 &pcp->lock &zone->lock &____s->seqcount irq_context: 0 sb_writers#8 &of->mutex kn->active#5 &rq->__lock irq_context: 0 kn->active#5 remove_cache_srcu &c->lock irq_context: 0 kn->active#5 remove_cache_srcu &n->list_lock irq_context: 0 kn->active#5 remove_cache_srcu &obj_hash[i].lock irq_context: 0 sb_writers#8 &of->mutex kn->active#5 &n->list_lock irq_context: 0 sb_writers#8 &of->mutex kn->active#5 &n->list_lock &c->lock irq_context: 0 kn->active#13 fs_reclaim irq_context: 0 kn->active#13 fs_reclaim mmu_notifier_invalidate_range_start irq_context: 0 kn->active#13 &c->lock irq_context: 0 kn->active#13 &____s->seqcount irq_context: 0 kn->active#13 &kernfs_locks->open_file_mutex[count] irq_context: 0 kn->active#13 &kernfs_locks->open_file_mutex[count] fs_reclaim irq_context: 0 kn->active#13 &kernfs_locks->open_file_mutex[count] fs_reclaim mmu_notifier_invalidate_range_start irq_context: 0 sb_writers#8 &of->mutex kn->active#5 udc_lock irq_context: 0 kn->active#6 &c->lock irq_context: 0 kn->active#6 &____s->seqcount irq_context: 0 kn->active#14 fs_reclaim irq_context: 0 kn->active#14 fs_reclaim mmu_notifier_invalidate_range_start irq_context: 0 kn->active#14 &kernfs_locks->open_file_mutex[count] irq_context: 0 kn->active#14 &kernfs_locks->open_file_mutex[count] fs_reclaim irq_context: 0 kn->active#14 &kernfs_locks->open_file_mutex[count] fs_reclaim mmu_notifier_invalidate_range_start irq_context: 0 &p->lock &of->mutex kn->active#14 fs_reclaim irq_context: 0 &p->lock &of->mutex kn->active#14 fs_reclaim mmu_notifier_invalidate_range_start irq_context: 0 sb_writers#8 &of->mutex kn->active#5 quarantine_lock irq_context: 0 &p->lock &of->mutex kn->active#14 pool_lock#2 irq_context: 0 &p->lock &of->mutex kn->active#14 &obj_hash[i].lock irq_context: 0 kn->active#10 &c->lock irq_context: 0 kn->active#10 &____s->seqcount irq_context: 0 kn->active#10 &kernfs_locks->open_file_mutex[count] &c->lock irq_context: 0 kn->active#10 &kernfs_locks->open_file_mutex[count] &____s->seqcount irq_context: 0 sb_writers#8 &of->mutex kn->active#5 fw_lock irq_context: 0 sb_writers#5 &type->i_mutex_dir_key#5 batched_entropy_u32.lock crngs.lock irq_context: 0 sb_writers#5 &type->i_mutex_dir_key#5/1 tomoyo_ss rcu_read_lock rcu_node_0 irq_context: 0 sb_writers#5 &type->i_mutex_dir_key#5/1 tomoyo_ss rcu_read_lock &rq->__lock irq_context: 0 sb_writers#5 &type->i_mutex_dir_key#5/1 tomoyo_ss rcu_read_lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 kn->active#9 &c->lock irq_context: 0 kn->active#9 &____s->seqcount irq_context: 0 &rnp->exp_lock irq_context: 0 rcu_state.exp_mutex irq_context: 0 rcu_state.exp_mutex rcu_state.exp_mutex.wait_lock irq_context: 0 sb_writers#8 &of->mutex kn->active#5 uevent_sock_mutex &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 sb_writers#5 &type->i_mutex_dir_key#5/1 tomoyo_ss rcu_read_lock &rcu_state.gp_wq irq_context: 0 sb_writers#5 &type->i_mutex_dir_key#5/1 tomoyo_ss rcu_read_lock &rcu_state.gp_wq &p->pi_lock irq_context: 0 sb_writers#5 &type->i_mutex_dir_key#5/1 tomoyo_ss rcu_read_lock &rcu_state.gp_wq &p->pi_lock &rq->__lock irq_context: 0 sb_writers#5 &type->i_mutex_dir_key#5/1 tomoyo_ss rcu_read_lock &rcu_state.gp_wq &p->pi_lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 sb_writers#5 &type->i_mutex_dir_key#5/1 tomoyo_ss rcu_read_lock &rcu_state.expedited_wq irq_context: 0 sb_writers#5 &type->i_mutex_dir_key#5/1 tomoyo_ss rcu_read_lock &rcu_state.expedited_wq &p->pi_lock irq_context: 0 sb_writers#5 &type->i_mutex_dir_key#5/1 tomoyo_ss rcu_read_lock &rcu_state.expedited_wq &p->pi_lock &rq->__lock irq_context: 0 sb_writers#5 &type->i_mutex_dir_key#5/1 tomoyo_ss rcu_read_lock &rcu_state.expedited_wq &p->pi_lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 rcu_state.exp_mutex.wait_lock irq_context: 0 kn->active#11 &c->lock irq_context: 0 kn->active#11 &____s->seqcount irq_context: 0 sb_writers#5 &type->i_mutex_dir_key#5/1 &sem->wait_lock irq_context: 0 sb_writers#5 &type->i_mutex_dir_key#5/1 &rq->__lock irq_context: 0 sb_writers#5 &type->i_mutex_dir_key#5/1 &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 sb_writers &sb->s_type->i_mutex_key#4 rcu_read_lock &dentry->d_lock irq_context: 0 sb_writers &sb->s_type->i_mutex_key#4 &dentry->d_lock irq_context: 0 sb_writers &sb->s_type->i_mutex_key#4 &fsnotify_mark_srcu irq_context: 0 sb_writers &sb->s_type->i_mutex_key#4 &fsnotify_mark_srcu fs_reclaim irq_context: 0 sb_writers &sb->s_type->i_mutex_key#4 &fsnotify_mark_srcu fs_reclaim mmu_notifier_invalidate_range_start irq_context: 0 sb_writers &sb->s_type->i_mutex_key#4 &fsnotify_mark_srcu pool_lock#2 irq_context: 0 sb_writers &sb->s_type->i_mutex_key#4 &fsnotify_mark_srcu &group->notification_lock irq_context: 0 sb_writers &sb->s_type->i_mutex_key#4 &fsnotify_mark_srcu &group->notification_waitq irq_context: 0 sb_writers &sb->s_type->i_mutex_key#4 &fsnotify_mark_srcu &group->notification_waitq &p->pi_lock irq_context: 0 sb_writers &sb->s_type->i_mutex_key#4 &fsnotify_mark_srcu &group->notification_waitq &p->pi_lock &rq->__lock irq_context: 0 sb_writers &sb->s_type->i_mutex_key#4 &fsnotify_mark_srcu &group->notification_waitq &p->pi_lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 sb_writers &sb->s_type->i_mutex_key#4 &fsnotify_mark_srcu &obj_hash[i].lock irq_context: 0 &type->i_mutex_dir_key#5 remove_cache_srcu rcu_node_0 irq_context: 0 (wq_completion)rcu_gp (work_completion)(&rew->rew_work) &obj_hash[i].lock pool_lock irq_context: 0 &type->i_mutex_dir_key#5 remove_cache_srcu &rcu_state.expedited_wq irq_context: 0 &type->i_mutex_dir_key#5 remove_cache_srcu &rcu_state.expedited_wq &p->pi_lock irq_context: 0 &type->i_mutex_dir_key#5 remove_cache_srcu &rcu_state.expedited_wq &p->pi_lock &rq->__lock irq_context: 0 &type->i_mutex_dir_key#5 remove_cache_srcu &rcu_state.expedited_wq &p->pi_lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 &type->i_mutex_dir_key#5 remove_cache_srcu &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 sb_writers &type->i_mutex_dir_key/1 tomoyo_ss &n->list_lock irq_context: 0 sb_writers &type->i_mutex_dir_key/1 tomoyo_ss &n->list_lock &c->lock irq_context: 0 sb_writers#5 &p->pi_lock &cfs_rq->removed.lock irq_context: 0 sb_writers#5 &p->pi_lock &rq->__lock &cfs_rq->removed.lock irq_context: 0 kn->active#5 &kernfs_locks->open_file_mutex[count] rcu_read_lock rcu_node_0 irq_context: 0 kn->active#5 &rq->__lock irq_context: 0 kn->active#5 &kernfs_locks->open_file_mutex[count] rcu_read_lock &rq->__lock irq_context: 0 kn->active#5 &kernfs_locks->open_file_mutex[count] rcu_read_lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 sb_writers#5 &type->i_mutex_dir_key#5/1 tomoyo_ss quarantine_lock irq_context: 0 sb_writers#8 &of->mutex kn->active#5 uevent_sock_mutex remove_cache_srcu irq_context: 0 sb_writers#8 &of->mutex kn->active#5 uevent_sock_mutex remove_cache_srcu quarantine_lock irq_context: 0 sb_writers#8 &of->mutex kn->active#5 uevent_sock_mutex remove_cache_srcu &c->lock irq_context: 0 sb_writers#8 &of->mutex kn->active#5 uevent_sock_mutex remove_cache_srcu &n->list_lock irq_context: 0 sb_writers#8 &of->mutex kn->active#5 uevent_sock_mutex remove_cache_srcu rcu_read_lock pool_lock#2 irq_context: 0 sb_writers#8 &of->mutex kn->active#5 uevent_sock_mutex remove_cache_srcu &obj_hash[i].lock irq_context: 0 kn->active#5 fs_reclaim mmu_notifier_invalidate_range_start &rq->__lock irq_context: 0 (wq_completion)events (work_completion)(&(&krcp->monitor_work)->work) rcu_callback &obj_hash[i].lock pool_lock irq_context: 0 (wq_completion)events (work_completion)(&(&krcp->monitor_work)->work) rcu_callback &rq->__lock irq_context: 0 sb_writers#5 &type->i_mutex_dir_key#5/1 tomoyo_ss &n->list_lock irq_context: 0 sb_writers#5 &type->i_mutex_dir_key#5/1 tomoyo_ss &n->list_lock &c->lock irq_context: 0 &p->lock &of->mutex kn->active#5 quarantine_lock irq_context: 0 remove_cache_srcu &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 kn->active#5 fs_reclaim &rq->__lock irq_context: 0 kn->active#5 remove_cache_srcu &rq->__lock irq_context: 0 sb_writers#5 &type->i_mutex_dir_key#5/1 &obj_hash[i].lock pool_lock irq_context: 0 &p->lock remove_cache_srcu irq_context: 0 &p->lock remove_cache_srcu quarantine_lock irq_context: 0 &p->lock remove_cache_srcu &c->lock irq_context: 0 &p->lock remove_cache_srcu &n->list_lock irq_context: 0 &p->lock remove_cache_srcu rcu_read_lock pool_lock#2 irq_context: 0 &p->lock remove_cache_srcu &obj_hash[i].lock irq_context: 0 &p->lock &of->mutex kn->active#5 &rq->__lock irq_context: 0 &p->lock &of->mutex kn->active#5 batched_entropy_u8.lock irq_context: 0 &p->lock &of->mutex kn->active#5 kfence_freelist_lock irq_context: 0 &p->lock &of->mutex kn->active#5 &meta->lock irq_context: 0 sb_writers#5 &type->i_mutex_dir_key#5 &c->lock irq_context: 0 sb_writers#5 &type->i_mutex_dir_key#5 &____s->seqcount irq_context: 0 &type->i_mutex_dir_key#5 &p->pi_lock irq_context: 0 sb_writers#5 &type->i_mutex_dir_key#5/1 &p->pi_lock irq_context: 0 kn->active#5 rcu_read_lock rcu_node_0 irq_context: 0 kn->active#5 rcu_read_lock &rq->__lock irq_context: 0 kn->active#5 &kernfs_locks->open_file_mutex[count] &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 kn->active#15 fs_reclaim irq_context: 0 kn->active#15 fs_reclaim mmu_notifier_invalidate_range_start irq_context: 0 kn->active#15 &c->lock irq_context: 0 kn->active#15 &n->list_lock irq_context: 0 kn->active#15 &n->list_lock &c->lock irq_context: 0 kn->active#15 &kernfs_locks->open_file_mutex[count] irq_context: 0 sb_writers#5 &type->i_mutex_dir_key#5/1 &n->list_lock irq_context: 0 kn->active#15 &kernfs_locks->open_file_mutex[count] fs_reclaim irq_context: 0 sb_writers#5 &type->i_mutex_dir_key#5/1 &n->list_lock &c->lock irq_context: 0 kn->active#15 &kernfs_locks->open_file_mutex[count] fs_reclaim mmu_notifier_invalidate_range_start irq_context: 0 &p->lock &of->mutex kn->active#15 dev_base_lock irq_context: 0 kn->active#16 fs_reclaim irq_context: 0 kn->active#16 fs_reclaim mmu_notifier_invalidate_range_start irq_context: 0 kn->active#16 &kernfs_locks->open_file_mutex[count] irq_context: 0 kn->active#16 &kernfs_locks->open_file_mutex[count] fs_reclaim irq_context: 0 kn->active#16 &kernfs_locks->open_file_mutex[count] fs_reclaim mmu_notifier_invalidate_range_start irq_context: 0 &p->lock &of->mutex kn->active#16 dev_base_lock irq_context: 0 kn->active#17 fs_reclaim irq_context: 0 kn->active#17 fs_reclaim mmu_notifier_invalidate_range_start irq_context: 0 kn->active#17 &kernfs_locks->open_file_mutex[count] irq_context: 0 kn->active#17 &kernfs_locks->open_file_mutex[count] fs_reclaim irq_context: 0 kn->active#17 &kernfs_locks->open_file_mutex[count] fs_reclaim mmu_notifier_invalidate_range_start irq_context: 0 kn->active#18 fs_reclaim irq_context: 0 kn->active#18 fs_reclaim mmu_notifier_invalidate_range_start irq_context: 0 kn->active#18 &c->lock irq_context: 0 kn->active#18 &kernfs_locks->open_file_mutex[count] irq_context: 0 kn->active#18 &kernfs_locks->open_file_mutex[count] fs_reclaim irq_context: 0 kn->active#18 &kernfs_locks->open_file_mutex[count] fs_reclaim mmu_notifier_invalidate_range_start irq_context: 0 &p->lock &of->mutex kn->active#18 dev_base_lock irq_context: 0 kn->active#19 fs_reclaim irq_context: 0 kn->active#19 fs_reclaim mmu_notifier_invalidate_range_start irq_context: 0 kn->active#19 pool_lock#2 irq_context: 0 kn->active#19 &kernfs_locks->open_file_mutex[count] irq_context: 0 kn->active#19 &kernfs_locks->open_file_mutex[count] fs_reclaim irq_context: 0 kn->active#19 &kernfs_locks->open_file_mutex[count] fs_reclaim mmu_notifier_invalidate_range_start irq_context: 0 &of->mutex irq_context: 0 kn->active#20 fs_reclaim irq_context: 0 kn->active#20 fs_reclaim mmu_notifier_invalidate_range_start irq_context: 0 kn->active#20 &kernfs_locks->open_file_mutex[count] irq_context: 0 kn->active#20 &kernfs_locks->open_file_mutex[count] fs_reclaim irq_context: 0 kn->active#20 &kernfs_locks->open_file_mutex[count] fs_reclaim mmu_notifier_invalidate_range_start irq_context: 0 &p->lock &of->mutex kn->active#20 dev_base_lock irq_context: 0 kn->active#21 fs_reclaim irq_context: 0 kn->active#21 fs_reclaim mmu_notifier_invalidate_range_start irq_context: 0 kn->active#21 &kernfs_locks->open_file_mutex[count] irq_context: 0 kn->active#21 &kernfs_locks->open_file_mutex[count] fs_reclaim irq_context: 0 kn->active#21 &kernfs_locks->open_file_mutex[count] fs_reclaim mmu_notifier_invalidate_range_start irq_context: 0 kn->active#22 fs_reclaim irq_context: 0 kn->active#22 fs_reclaim mmu_notifier_invalidate_range_start irq_context: 0 kn->active#23 fs_reclaim irq_context: 0 kn->active#22 &kernfs_locks->open_file_mutex[count] irq_context: 0 kn->active#23 fs_reclaim mmu_notifier_invalidate_range_start irq_context: 0 kn->active#22 &kernfs_locks->open_file_mutex[count] fs_reclaim irq_context: 0 kn->active#23 &kernfs_locks->open_file_mutex[count] irq_context: 0 kn->active#22 &kernfs_locks->open_file_mutex[count] fs_reclaim mmu_notifier_invalidate_range_start irq_context: 0 kn->active#23 &kernfs_locks->open_file_mutex[count] fs_reclaim irq_context: 0 kn->active#23 &kernfs_locks->open_file_mutex[count] fs_reclaim mmu_notifier_invalidate_range_start irq_context: 0 &p->lock &of->mutex kn->active#23 dev_base_lock irq_context: 0 kn->active#24 fs_reclaim irq_context: 0 kn->active#24 fs_reclaim mmu_notifier_invalidate_range_start irq_context: 0 kn->active#24 &kernfs_locks->open_file_mutex[count] irq_context: 0 kn->active#24 &kernfs_locks->open_file_mutex[count] fs_reclaim irq_context: 0 kn->active#24 &kernfs_locks->open_file_mutex[count] fs_reclaim mmu_notifier_invalidate_range_start irq_context: 0 kn->active#25 fs_reclaim irq_context: 0 kn->active#25 fs_reclaim mmu_notifier_invalidate_range_start irq_context: 0 kn->active#25 &kernfs_locks->open_file_mutex[count] irq_context: 0 kn->active#25 &kernfs_locks->open_file_mutex[count] fs_reclaim irq_context: 0 kn->active#25 &kernfs_locks->open_file_mutex[count] fs_reclaim mmu_notifier_invalidate_range_start irq_context: 0 &of->mutex kn->active#25 &dev->power.lock irq_context: 0 &of->mutex kn->active#25 pci_lock irq_context: 0 &of->mutex kn->active#25 pci_lock pci_config_lock irq_context: 0 kn->active#26 fs_reclaim irq_context: 0 kn->active#26 fs_reclaim mmu_notifier_invalidate_range_start irq_context: 0 kn->active#26 &kernfs_locks->open_file_mutex[count] irq_context: 0 kn->active#26 &kernfs_locks->open_file_mutex[count] fs_reclaim irq_context: 0 kn->active#26 &kernfs_locks->open_file_mutex[count] fs_reclaim mmu_notifier_invalidate_range_start irq_context: 0 kn->active#27 fs_reclaim irq_context: 0 kn->active#27 fs_reclaim mmu_notifier_invalidate_range_start irq_context: 0 kn->active#27 &kernfs_locks->open_file_mutex[count] irq_context: 0 kn->active#27 &kernfs_locks->open_file_mutex[count] fs_reclaim irq_context: 0 kn->active#27 &kernfs_locks->open_file_mutex[count] fs_reclaim mmu_notifier_invalidate_range_start irq_context: 0 kn->active#28 fs_reclaim irq_context: 0 kn->active#28 fs_reclaim &rq->__lock irq_context: 0 kn->active#29 &rq->__lock irq_context: 0 kn->active#29 fs_reclaim irq_context: 0 kn->active#29 fs_reclaim mmu_notifier_invalidate_range_start irq_context: 0 kn->active#29 &kernfs_locks->open_file_mutex[count] irq_context: 0 kn->active#29 &kernfs_locks->open_file_mutex[count] fs_reclaim irq_context: 0 kn->active#29 &kernfs_locks->open_file_mutex[count] fs_reclaim mmu_notifier_invalidate_range_start irq_context: 0 &p->lock &of->mutex kn->active#5 &device->physical_node_lock irq_context: 0 sb_writers#8 &of->mutex kn->active#5 fs_reclaim &rq->__lock irq_context: 0 kn->active#19 &c->lock irq_context: 0 kn->active#24 &c->lock irq_context: 0 kn->active#29 &c->lock irq_context: 0 sb_writers#5 &type->i_mutex_dir_key#5/1 &sb->s_type->i_mutex_key#12 &rq->__lock irq_context: 0 kn->active#19 &kernfs_locks->open_file_mutex[count] &c->lock irq_context: 0 kn->active#19 &kernfs_locks->open_file_mutex[count] &____s->seqcount irq_context: 0 kn->active#28 fs_reclaim mmu_notifier_invalidate_range_start irq_context: 0 kn->active#28 &kernfs_locks->open_file_mutex[count] irq_context: 0 kn->active#28 &kernfs_locks->open_file_mutex[count] fs_reclaim irq_context: 0 kn->active#28 &kernfs_locks->open_file_mutex[count] fs_reclaim mmu_notifier_invalidate_range_start irq_context: 0 kn->active#30 fs_reclaim irq_context: 0 kn->active#30 fs_reclaim mmu_notifier_invalidate_range_start irq_context: 0 kn->active#30 &kernfs_locks->open_file_mutex[count] irq_context: 0 kn->active#30 &kernfs_locks->open_file_mutex[count] fs_reclaim irq_context: 0 kn->active#30 &kernfs_locks->open_file_mutex[count] fs_reclaim mmu_notifier_invalidate_range_start irq_context: 0 kn->active#27 &c->lock irq_context: 0 &type->i_mutex_dir_key#2 &n->list_lock irq_context: 0 &type->i_mutex_dir_key#2 &n->list_lock &c->lock irq_context: 0 sb_writers#5 &type->i_mutex_dir_key#5 batched_entropy_u8.lock irq_context: 0 sb_writers#5 &type->i_mutex_dir_key#5 kfence_freelist_lock irq_context: 0 sb_writers#8 &of->mutex kn->active#5 uevent_sock_mutex rcu_read_lock &rcu_state.gp_wq irq_context: 0 sb_writers#8 &of->mutex kn->active#5 uevent_sock_mutex rcu_read_lock &rcu_state.gp_wq &p->pi_lock irq_context: 0 sb_writers#8 &of->mutex kn->active#5 uevent_sock_mutex rcu_read_lock &rcu_state.gp_wq &p->pi_lock &cfs_rq->removed.lock irq_context: 0 sb_writers#8 &of->mutex kn->active#5 uevent_sock_mutex rcu_read_lock &rcu_state.gp_wq &p->pi_lock &rq->__lock irq_context: 0 sb_writers#8 &of->mutex kn->active#5 uevent_sock_mutex rcu_read_lock &rcu_state.gp_wq &p->pi_lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 tomoyo_ss rcu_read_lock rename_lock irq_context: 0 kn->active#21 &c->lock irq_context: 0 kn->active#26 &kernfs_locks->open_file_mutex[count] &c->lock irq_context: 0 kn->active#26 &kernfs_locks->open_file_mutex[count] &____s->seqcount irq_context: 0 sb_writers#5 &type->i_mutex_dir_key#5/1 tomoyo_ss &rq->__lock irq_context: 0 sb_writers#5 &type->i_mutex_dir_key#5/1 tomoyo_ss &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 &kernfs_locks->open_file_mutex[count] rcu_read_lock rcu_node_0 irq_context: 0 &kernfs_locks->open_file_mutex[count] rcu_read_lock &rq->__lock irq_context: 0 &p->lock &of->mutex kn->active#5 udc_lock irq_context: 0 kn->active#21 &____s->seqcount irq_context: 0 sb_writers &sb->s_type->i_mutex_key#4 tomoyo_ss &n->list_lock irq_context: 0 sb_writers &sb->s_type->i_mutex_key#4 tomoyo_ss &n->list_lock &c->lock irq_context: 0 &type->i_mutex_dir_key#2 &sem->wait_lock irq_context: 0 &type->i_mutex_dir_key#2 &rq->__lock irq_context: 0 &type->i_mutex_dir_key#2 &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 sb_writers &sem->wait_lock irq_context: 0 sb_writers &p->pi_lock irq_context: 0 sb_writers &p->pi_lock &rq->__lock irq_context: 0 sb_writers &p->pi_lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 kn->active#5 remove_cache_srcu rcu_read_lock pool_lock#2 irq_context: 0 kn->active#31 fs_reclaim irq_context: 0 kn->active#31 fs_reclaim mmu_notifier_invalidate_range_start irq_context: 0 kn->active#31 &kernfs_locks->open_file_mutex[count] irq_context: 0 kn->active#31 &kernfs_locks->open_file_mutex[count] fs_reclaim irq_context: 0 kn->active#31 &kernfs_locks->open_file_mutex[count] fs_reclaim mmu_notifier_invalidate_range_start irq_context: 0 tomoyo_ss mmu_notifier_invalidate_range_start &rq->__lock irq_context: 0 tomoyo_ss mmu_notifier_invalidate_range_start &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 kn->active#32 fs_reclaim irq_context: 0 kn->active#32 fs_reclaim mmu_notifier_invalidate_range_start irq_context: 0 kn->active#32 &kernfs_locks->open_file_mutex[count] irq_context: 0 kn->active#32 &kernfs_locks->open_file_mutex[count] fs_reclaim irq_context: 0 kn->active#32 &kernfs_locks->open_file_mutex[count] fs_reclaim mmu_notifier_invalidate_range_start irq_context: 0 sb_writers#5 &type->i_mutex_dir_key#5/1 &dentry->d_lock &dentry->d_lock/1 irq_context: 0 sb_writers#5 &type->i_mutex_dir_key#5/1 &dentry->d_lock/1 irq_context: 0 kn->active#32 &c->lock irq_context: 0 kn->active#33 fs_reclaim irq_context: 0 kn->active#33 fs_reclaim mmu_notifier_invalidate_range_start irq_context: 0 kn->active#33 &kernfs_locks->open_file_mutex[count] irq_context: 0 kn->active#33 &kernfs_locks->open_file_mutex[count] fs_reclaim irq_context: 0 kn->active#33 &kernfs_locks->open_file_mutex[count] fs_reclaim mmu_notifier_invalidate_range_start irq_context: 0 kn->active#5 &kernfs_locks->open_file_mutex[count] &n->list_lock irq_context: 0 kn->active#5 &kernfs_locks->open_file_mutex[count] &n->list_lock &c->lock irq_context: 0 dup_mmap_sem &mm->mmap_lock &mm->mmap_lock/1 remove_cache_srcu &rq->__lock irq_context: 0 dup_mmap_sem &mm->mmap_lock &mm->mmap_lock/1 remove_cache_srcu &rq->__lock &cfs_rq->removed.lock irq_context: 0 fill_pool_map-wait-type-override &n->list_lock irq_context: 0 fill_pool_map-wait-type-override &n->list_lock &c->lock irq_context: 0 &kernfs_locks->open_file_mutex[count] fill_pool_map-wait-type-override &n->list_lock irq_context: 0 &kernfs_locks->open_file_mutex[count] fill_pool_map-wait-type-override &n->list_lock &c->lock irq_context: 0 kn->active#5 batched_entropy_u8.lock irq_context: 0 kn->active#5 kfence_freelist_lock irq_context: 0 kn->active#9 &kernfs_locks->open_file_mutex[count] &c->lock irq_context: 0 kn->active#11 &kernfs_locks->open_file_mutex[count] &c->lock irq_context: 0 kn->active#12 &c->lock irq_context: 0 kn->active#12 &n->list_lock irq_context: 0 kn->active#12 &n->list_lock &c->lock irq_context: 0 &ep->mtx &pipe->rd_wait irq_context: 0 kn->active#8 &kernfs_locks->open_file_mutex[count] &c->lock irq_context: 0 kn->active#8 &kernfs_locks->open_file_mutex[count] &____s->seqcount irq_context: 0 &p->lock &of->mutex kn->active#5 fw_lock irq_context: 0 kn->active#14 &kernfs_locks->open_file_mutex[count] &c->lock irq_context: 0 kn->active#14 &kernfs_locks->open_file_mutex[count] &____s->seqcount irq_context: 0 &p->lock rcu_read_lock rcu_node_0 irq_context: 0 &p->lock rcu_read_lock &rq->__lock irq_context: 0 &p->lock rcu_read_lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 kn->active#6 &n->list_lock irq_context: 0 kn->active#6 &n->list_lock &c->lock irq_context: 0 sb_writers &type->i_mutex_dir_key/1 &fsnotify_mark_srcu irq_context: 0 sb_writers &type->i_mutex_dir_key/1 &fsnotify_mark_srcu fs_reclaim irq_context: 0 sb_writers &type->i_mutex_dir_key/1 &fsnotify_mark_srcu fs_reclaim mmu_notifier_invalidate_range_start irq_context: 0 sb_writers &type->i_mutex_dir_key/1 &fsnotify_mark_srcu pool_lock#2 irq_context: 0 sb_writers &type->i_mutex_dir_key/1 &fsnotify_mark_srcu &group->notification_lock irq_context: 0 sb_writers &type->i_mutex_dir_key/1 &fsnotify_mark_srcu &group->notification_waitq irq_context: 0 sb_writers &type->i_mutex_dir_key/1 &fsnotify_mark_srcu &group->notification_waitq &p->pi_lock irq_context: 0 sb_writers &type->i_mutex_dir_key/1 tomoyo_ss &pcp->lock &zone->lock irq_context: 0 sb_writers &type->i_mutex_dir_key/1 tomoyo_ss &pcp->lock &zone->lock &____s->seqcount irq_context: 0 &mousedev->client_lock irq_context: 0 &mousedev->mutex#2 irq_context: 0 &mousedev->mutex#2 &dev->mutex#2 irq_context: 0 &sig->cred_guard_mutex &sig->exec_update_lock &mm->mmap_lock rcu_read_lock rcu_node_0 irq_context: 0 &sig->cred_guard_mutex &sig->exec_update_lock &mm->mmap_lock rcu_read_lock &rq->__lock irq_context: 0 sb_writers &sb->s_type->i_mutex_key#4 tomoyo_ss quarantine_lock irq_context: 0 sb_writers &type->i_mutex_dir_key/1 &n->list_lock irq_context: 0 sb_writers &type->i_mutex_dir_key/1 &n->list_lock &c->lock irq_context: 0 &mm->mmap_lock remove_cache_srcu irq_context: 0 kn->active#10 remove_cache_srcu irq_context: 0 &mm->mmap_lock remove_cache_srcu quarantine_lock irq_context: 0 kn->active#10 remove_cache_srcu quarantine_lock irq_context: 0 &mm->mmap_lock remove_cache_srcu &c->lock irq_context: 0 &mm->mmap_lock remove_cache_srcu &n->list_lock irq_context: 0 &mm->mmap_lock remove_cache_srcu rcu_read_lock pool_lock#2 irq_context: 0 &mm->mmap_lock remove_cache_srcu &obj_hash[i].lock irq_context: 0 &mm->mmap_lock remove_cache_srcu &rq->__lock irq_context: 0 &mm->mmap_lock remove_cache_srcu &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 tomoyo_ss rcu_read_lock &rcu_state.gp_wq irq_context: 0 tomoyo_ss rcu_read_lock &rcu_state.gp_wq &p->pi_lock irq_context: 0 tomoyo_ss rcu_read_lock &rcu_state.gp_wq &p->pi_lock &rq->__lock irq_context: 0 tomoyo_ss rcu_read_lock &rcu_state.gp_wq &p->pi_lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 tomoyo_ss rcu_read_lock &rcu_state.expedited_wq irq_context: 0 tomoyo_ss rcu_read_lock &rcu_state.expedited_wq &p->pi_lock irq_context: 0 tomoyo_ss rcu_read_lock &rcu_state.expedited_wq &p->pi_lock &rq->__lock irq_context: 0 tomoyo_ss rcu_read_lock &rcu_state.expedited_wq &p->pi_lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 kn->active#7 &kernfs_locks->open_file_mutex[count] &____s->seqcount irq_context: 0 kn->active#9 &n->list_lock irq_context: 0 kn->active#9 &n->list_lock &c->lock irq_context: 0 &rnp->exp_wq[3] irq_context: 0 &mousedev->mutex#2 &dev->mutex#2 &rnp->exp_lock irq_context: 0 &mousedev->mutex#2 &dev->mutex#2 &rnp->exp_wq[3] irq_context: 0 &mousedev->mutex#2 &dev->mutex#2 &rq->__lock irq_context: 0 &mousedev->mutex#2 &dev->mutex#2 &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 sb_writers#8 kn->active#5 fs_reclaim irq_context: 0 sb_writers#8 kn->active#5 fs_reclaim mmu_notifier_invalidate_range_start irq_context: 0 sb_writers#8 kn->active#5 &kernfs_locks->open_file_mutex[count] irq_context: 0 sb_writers#8 kn->active#5 &kernfs_locks->open_file_mutex[count] fs_reclaim irq_context: 0 sb_writers#8 kn->active#5 &kernfs_locks->open_file_mutex[count] fs_reclaim mmu_notifier_invalidate_range_start irq_context: 0 sb_writers &type->i_mutex_dir_key/1 tomoyo_ss quarantine_lock irq_context: 0 sb_writers &type->i_mutex_dir_key/1 tomoyo_ss remove_cache_srcu irq_context: 0 sb_writers &type->i_mutex_dir_key/1 tomoyo_ss remove_cache_srcu quarantine_lock irq_context: 0 sb_writers &type->i_mutex_dir_key/1 tomoyo_ss remove_cache_srcu &c->lock irq_context: 0 sb_writers &type->i_mutex_dir_key/1 tomoyo_ss remove_cache_srcu &n->list_lock irq_context: 0 sb_writers &type->i_mutex_dir_key/1 tomoyo_ss remove_cache_srcu rcu_read_lock pool_lock#2 irq_context: 0 sb_writers &type->i_mutex_dir_key/1 tomoyo_ss remove_cache_srcu &obj_hash[i].lock irq_context: 0 sb_writers &type->i_mutex_dir_key/1 &sem->wait_lock irq_context: 0 sb_writers &type->i_mutex_dir_key/1 &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 rcu_read_lock &rcu_state.expedited_wq irq_context: 0 rcu_read_lock &rcu_state.expedited_wq &p->pi_lock irq_context: 0 rcu_read_lock &rcu_state.expedited_wq &p->pi_lock &rq->__lock irq_context: 0 rcu_read_lock &rcu_state.expedited_wq &p->pi_lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 &evdev->mutex &dev->mutex#2 &lock->wait_lock irq_context: 0 &evdev->mutex &dev->mutex#2 &rq->__lock irq_context: 0 &evdev->mutex &dev->mutex#2 &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 &mousedev->mutex#2 &lock->wait_lock irq_context: 0 &mousedev->mutex#2 &p->pi_lock irq_context: 0 &mousedev->mutex#2 &p->pi_lock &rq->__lock irq_context: 0 &mousedev->mutex#2 &p->pi_lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 &evdev->mutex &dev->mutex#2 rcu_state.exp_mutex &rnp->exp_wq[0] irq_context: 0 sb_writers#8 &sb->s_type->i_mutex_key#13 &p->pi_lock &cfs_rq->removed.lock irq_context: 0 sb_writers#8 &sb->s_type->i_mutex_key#13 &rq->__lock irq_context: 0 remove_cache_srcu rcu_read_lock &____s->seqcount irq_context: 0 &type->i_mutex_dir_key#2 &rq->__lock &cfs_rq->removed.lock irq_context: 0 (wq_completion)events (work_completion)(&(&krcp->krw_arr[i].rcu_work)->work) rcu_callback &base->lock irq_context: 0 (wq_completion)events (work_completion)(&(&krcp->krw_arr[i].rcu_work)->work) rcu_callback &base->lock &obj_hash[i].lock irq_context: 0 sb_writers &p->pi_lock &cfs_rq->removed.lock irq_context: 0 sb_writers &p->pi_lock &rq->__lock &cfs_rq->removed.lock irq_context: 0 sb_writers &sb->s_type->i_mutex_key#4 &fsnotify_mark_srcu &rq->__lock irq_context: 0 sb_writers &sb->s_type->i_mutex_key#4 &fsnotify_mark_srcu remove_cache_srcu irq_context: 0 sb_writers &sb->s_type->i_mutex_key#4 &fsnotify_mark_srcu remove_cache_srcu quarantine_lock irq_context: 0 rcu_state.exp_mutex &obj_hash[i].lock pool_lock irq_context: 0 sb_writers#5 &type->i_mutex_dir_key#5 &n->list_lock irq_context: 0 sb_writers#5 &type->i_mutex_dir_key#5 &n->list_lock &c->lock irq_context: 0 &evdev->mutex &dev->mutex#2 rcu_state.exp_mutex &rnp->exp_wq[2] irq_context: 0 sb_writers#5 &type->i_mutex_dir_key#5 remove_cache_srcu irq_context: 0 sb_writers#5 &type->i_mutex_dir_key#5 remove_cache_srcu quarantine_lock irq_context: 0 sb_writers#5 &type->i_mutex_dir_key#5 remove_cache_srcu &c->lock irq_context: 0 sb_writers#5 &type->i_mutex_dir_key#5 remove_cache_srcu &n->list_lock irq_context: 0 sb_writers#5 &type->i_mutex_dir_key#5 remove_cache_srcu rcu_read_lock pool_lock#2 irq_context: 0 sb_writers#5 &type->i_mutex_dir_key#5 remove_cache_srcu &obj_hash[i].lock irq_context: 0 sb_writers &rq->__lock irq_context: 0 sb_writers#5 &type->i_mutex_dir_key#5 &simple_offset_xa_lock &c->lock irq_context: 0 kn->active#29 &n->list_lock irq_context: 0 kn->active#29 &n->list_lock &c->lock irq_context: 0 kn->active#21 &n->list_lock irq_context: 0 kn->active#21 &n->list_lock &c->lock irq_context: 0 kn->active#26 &c->lock irq_context: 0 sb_writers &sb->s_type->i_mutex_key#4 tomoyo_ss &rq->__lock irq_context: 0 sb_writers &sb->s_type->i_mutex_key#4 tomoyo_ss &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 kn->active#21 &kernfs_locks->open_file_mutex[count] &c->lock irq_context: 0 kn->active#21 &kernfs_locks->open_file_mutex[count] &____s->seqcount irq_context: 0 tomoyo_ss &base->lock irq_context: 0 tomoyo_ss &base->lock &obj_hash[i].lock irq_context: 0 kn->active#27 &n->list_lock irq_context: 0 kn->active#27 &n->list_lock &c->lock irq_context: 0 kn->active#27 &____s->seqcount irq_context: 0 &mm->mmap_lock rcu_read_lock rcu_read_lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 sb_writers#8 &of->mutex kn->active#5 &rfkill->lock irq_context: 0 sb_writers#5 &type->i_mutex_dir_key#5 tomoyo_ss &pcp->lock &zone->lock irq_context: 0 sb_writers#5 &type->i_mutex_dir_key#5 tomoyo_ss &pcp->lock &zone->lock &____s->seqcount irq_context: 0 kn->active#22 &c->lock irq_context: 0 kn->active#26 remove_cache_srcu irq_context: 0 kn->active#26 remove_cache_srcu quarantine_lock irq_context: 0 &hctx->lock irq_context: 0 rcu_read_lock &hctx->lock irq_context: 0 rcu_read_lock &virtscsi_vq->vq_lock irq_context: 0 &pipe->rd_wait &ep->lock irq_context: 0 &pipe->rd_wait &ep->lock &ep->wq irq_context: 0 &pipe->rd_wait &ep->lock &ep->wq &p->pi_lock irq_context: 0 &pipe->rd_wait &ep->lock &ep->wq &p->pi_lock &rq->__lock irq_context: 0 &pipe->rd_wait &ep->lock &ep->wq &p->pi_lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 &pipe->mutex/1 &mm->mmap_lock ptlock_ptr(page)#2 irq_context: 0 &pipe->mutex/1 &pipe->rd_wait &ep->lock irq_context: 0 &pipe->mutex/1 &pipe->rd_wait &ep->lock &ep->wq irq_context: 0 &pipe->mutex/1 &pipe->rd_wait &ep->lock &ep->wq &p->pi_lock irq_context: 0 &pipe->mutex/1 &pipe->rd_wait &ep->lock &ep->wq &p->pi_lock &cfs_rq->removed.lock irq_context: 0 &pipe->mutex/1 &pipe->rd_wait &ep->lock &ep->wq &p->pi_lock &rq->__lock irq_context: 0 &pipe->mutex/1 &pipe->rd_wait &ep->lock &ep->wq &p->pi_lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 &type->i_mutex_dir_key#4 rcu_read_lock &dentry->d_lock irq_context: 0 &type->i_mutex_dir_key#4 &sb->s_type->i_lock_key#24 &dentry->d_lock &wq &p->pi_lock irq_context: 0 &type->i_mutex_dir_key#4 &sb->s_type->i_lock_key#24 &dentry->d_lock &wq &p->pi_lock &rq->__lock irq_context: 0 &type->i_mutex_dir_key#4 &sb->s_type->i_lock_key#24 &dentry->d_lock &wq &p->pi_lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 &ep->mtx rcu_read_lock &pipe->rd_wait irq_context: 0 &ep->mtx &obj_hash[i].lock irq_context: 0 tasklist_lock &sighand->siglock &sighand->signalfd_wqh irq_context: 0 tasklist_lock &sighand->siglock &sighand->signalfd_wqh &ep->lock irq_context: 0 &sighand->signalfd_wqh irq_context: 0 &sb->s_type->i_mutex_key#14 irq_context: 0 mapping.invalidate_lock#2 irq_context: 0 mapping.invalidate_lock#2 mmu_notifier_invalidate_range_start irq_context: 0 mapping.invalidate_lock#2 &____s->seqcount irq_context: 0 mapping.invalidate_lock#2 pool_lock#2 irq_context: 0 mapping.invalidate_lock#2 &xa->xa_lock#9 irq_context: 0 mapping.invalidate_lock#2 &xa->xa_lock#9 pool_lock#2 irq_context: 0 kn->active#24 &kernfs_locks->open_file_mutex[count] &c->lock irq_context: 0 mapping.invalidate_lock#2 lock#4 irq_context: 0 kn->active#24 &kernfs_locks->open_file_mutex[count] &____s->seqcount irq_context: 0 mapping.invalidate_lock#2 tk_core.seq.seqcount irq_context: 0 mapping.invalidate_lock#2 &dd->lock irq_context: 0 mapping.invalidate_lock#2 rcu_read_lock &dd->lock irq_context: 0 mapping.invalidate_lock#2 rcu_read_lock &obj_hash[i].lock irq_context: 0 mapping.invalidate_lock#2 rcu_read_lock tk_core.seq.seqcount irq_context: 0 mapping.invalidate_lock#2 rcu_read_lock &virtscsi_vq->vq_lock irq_context: 0 sb_writers#5 &type->i_mutex_dir_key#5 tomoyo_ss &rq->__lock irq_context: 0 mapping.invalidate_lock#2 &xa->xa_lock#9 &c->lock irq_context: 0 mapping.invalidate_lock#2 rcu_read_lock &obj_hash[i].lock pool_lock irq_context: 0 kn->active#24 &n->list_lock irq_context: 0 kn->active#24 &n->list_lock &c->lock irq_context: 0 sb_writers#5 &sb->s_type->i_mutex_key#12 &pcp->lock &zone->lock irq_context: 0 sb_writers#5 &sb->s_type->i_mutex_key#12 &pcp->lock &zone->lock &____s->seqcount irq_context: 0 sb_writers#5 &type->i_mutex_dir_key#5/1 tomoyo_ss remove_cache_srcu irq_context: 0 sb_writers#5 &type->i_mutex_dir_key#5/1 tomoyo_ss remove_cache_srcu quarantine_lock irq_context: 0 sb_writers#5 &type->i_mutex_dir_key#5/1 tomoyo_ss remove_cache_srcu &n->list_lock irq_context: 0 sb_writers#5 &type->i_mutex_dir_key#5/1 tomoyo_ss remove_cache_srcu rcu_read_lock pool_lock#2 irq_context: 0 sb_writers#5 &type->i_mutex_dir_key#5/1 tomoyo_ss remove_cache_srcu &obj_hash[i].lock irq_context: 0 sb_writers#5 &type->i_mutex_dir_key#5/1 tomoyo_ss remove_cache_srcu &c->lock irq_context: 0 mapping.invalidate_lock#2 &xa->xa_lock#9 &n->list_lock irq_context: 0 mapping.invalidate_lock#2 &xa->xa_lock#9 &n->list_lock &c->lock irq_context: 0 mapping.invalidate_lock#2 lock#4 &lruvec->lru_lock irq_context: 0 mapping.invalidate_lock#2 &c->lock irq_context: 0 sb_writers#8 &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 kn->active#29 &kernfs_locks->open_file_mutex[count] &c->lock irq_context: 0 kn->active#29 &kernfs_locks->open_file_mutex[count] &____s->seqcount irq_context: 0 kn->active#29 &kernfs_locks->open_file_mutex[count] remove_cache_srcu irq_context: 0 kn->active#29 &kernfs_locks->open_file_mutex[count] remove_cache_srcu quarantine_lock irq_context: 0 kn->active#29 &kernfs_locks->open_file_mutex[count] remove_cache_srcu &c->lock irq_context: 0 kn->active#29 &kernfs_locks->open_file_mutex[count] remove_cache_srcu &n->list_lock irq_context: 0 kn->active#29 &kernfs_locks->open_file_mutex[count] remove_cache_srcu rcu_read_lock pool_lock#2 irq_context: 0 kn->active#29 &kernfs_locks->open_file_mutex[count] remove_cache_srcu &obj_hash[i].lock irq_context: 0 &p->lock &rq->__lock irq_context: 0 mapping.invalidate_lock#2 rcu_read_lock pool_lock#2 irq_context: 0 mapping.invalidate_lock#2 rcu_read_lock &____s->seqcount irq_context: 0 mapping.invalidate_lock#2 rcu_read_lock &c->lock irq_context: 0 mapping.invalidate_lock#2 mmu_notifier_invalidate_range_start &rq->__lock irq_context: 0 mapping.invalidate_lock#2 &rq->__lock irq_context: 0 mapping.invalidate_lock#2 &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 &mm->mmap_lock remove_cache_srcu &c->lock irq_context: 0 &mm->mmap_lock remove_cache_srcu &n->list_lock irq_context: 0 &mm->mmap_lock remove_cache_srcu rcu_read_lock pool_lock#2 irq_context: 0 &mm->mmap_lock remove_cache_srcu &obj_hash[i].lock irq_context: 0 kn->active#27 &kernfs_locks->open_file_mutex[count] &c->lock irq_context: 0 sb_writers#8 rcu_read_lock &rq->__lock irq_context: 0 sb_writers#8 rcu_read_lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 kn->active#34 fs_reclaim irq_context: 0 kn->active#34 fs_reclaim mmu_notifier_invalidate_range_start irq_context: 0 kn->active#34 &kernfs_locks->open_file_mutex[count] irq_context: 0 kn->active#34 &kernfs_locks->open_file_mutex[count] fs_reclaim irq_context: 0 kn->active#34 &kernfs_locks->open_file_mutex[count] fs_reclaim mmu_notifier_invalidate_range_start irq_context: 0 kn->active#34 &c->lock irq_context: 0 sb_writers#5 &type->i_mutex_dir_key#5/1 &sb->s_type->i_mutex_key#12 &sb->s_type->i_mutex_key#12/4 irq_context: 0 sb_writers#5 &type->i_mutex_dir_key#5/1 &sb->s_type->i_mutex_key#12 &sb->s_type->i_mutex_key#12/4 &dentry->d_lock irq_context: 0 sb_writers#5 &type->i_mutex_dir_key#5/1 &sb->s_type->i_mutex_key#12 &sb->s_type->i_mutex_key#12/4 &simple_offset_xa_lock irq_context: 0 sb_writers#5 &type->i_mutex_dir_key#5/1 &sb->s_type->i_mutex_key#12 &sb->s_type->i_mutex_key#12/4 fs_reclaim irq_context: 0 sb_writers#5 &type->i_mutex_dir_key#5/1 &sb->s_type->i_mutex_key#12 &sb->s_type->i_mutex_key#12/4 fs_reclaim mmu_notifier_invalidate_range_start irq_context: 0 sb_writers#5 &type->i_mutex_dir_key#5/1 &sb->s_type->i_mutex_key#12 &sb->s_type->i_mutex_key#12/4 tk_core.seq.seqcount irq_context: 0 sb_writers#5 &type->i_mutex_dir_key#5/1 &sb->s_type->i_mutex_key#12 &sb->s_type->i_mutex_key#12/4 rcu_read_lock &dentry->d_lock irq_context: 0 sb_writers#5 &type->i_mutex_dir_key#5/1 &sb->s_type->i_mutex_key#12 &sb->s_type->i_mutex_key#12/4 rename_lock irq_context: 0 sb_writers#5 &type->i_mutex_dir_key#5/1 &sb->s_type->i_mutex_key#12 &sb->s_type->i_mutex_key#12/4 rename_lock rename_lock.seqcount irq_context: 0 sb_writers#5 &type->i_mutex_dir_key#5/1 &sb->s_type->i_mutex_key#12 &sb->s_type->i_mutex_key#12/4 rename_lock rename_lock.seqcount &dentry->d_lock irq_context: 0 sb_writers#5 &type->i_mutex_dir_key#5/1 &sb->s_type->i_mutex_key#12 &sb->s_type->i_mutex_key#12/4 rename_lock rename_lock.seqcount &dentry->d_lock &dentry->d_lock/2 irq_context: 0 sb_writers#5 &type->i_mutex_dir_key#5/1 &sb->s_type->i_mutex_key#12 &sb->s_type->i_mutex_key#12/4 rename_lock rename_lock.seqcount &dentry->d_lock &dentry->d_lock/2 &dentry->d_lock/3 irq_context: 0 sb_writers#5 &type->i_mutex_dir_key#5/1 &sb->s_type->i_mutex_key#12 &sb->s_type->i_mutex_key#12/4 rename_lock rename_lock.seqcount &dentry->d_lock &dentry->d_lock/2 &dentry->d_lock/3 &____s->seqcount#4 irq_context: 0 sb_writers#5 &type->i_mutex_dir_key#5/1 &sb->s_type->i_mutex_key#12 &sb->s_type->i_mutex_key#12/4 rename_lock rename_lock.seqcount &dentry->d_lock &dentry->d_lock/2 &dentry->d_lock/3 &____s->seqcount#4 &____s->seqcount#4/1 irq_context: 0 sb_writers#5 &type->i_mutex_dir_key#5/1 &sb->s_type->i_mutex_key#12 &sb->s_type->i_mutex_key#12/4 rename_lock rename_lock.seqcount &dentry->d_lock/2 irq_context: 0 sb_writers#5 &type->i_mutex_dir_key#5/1 &sb->s_type->i_mutex_key#12 &sb->s_type->i_mutex_key#12/4 rename_lock rename_lock.seqcount &dentry->d_lock/2 &dentry->d_lock/3 irq_context: 0 sb_writers#5 &type->i_mutex_dir_key#5/1 &sb->s_type->i_lock_key &xa->xa_lock#9 irq_context: 0 sb_writers#5 &type->i_mutex_dir_key#5/1 lock#4 irq_context: 0 sb_writers#5 &type->i_mutex_dir_key#5/1 lock#4 &lruvec->lru_lock irq_context: 0 sb_writers#5 &type->i_mutex_dir_key#5/1 lock#5 irq_context: 0 sb_writers#5 &type->i_mutex_dir_key#5/1 &lruvec->lru_lock irq_context: 0 sb_writers#5 &type->i_mutex_dir_key#5/1 rcu_read_lock pool_lock#2 irq_context: 0 sb_writers#5 &type->i_mutex_dir_key#5/1 &info->lock irq_context: 0 sb_writers#5 &type->i_mutex_dir_key#5/1 &xa->xa_lock#9 irq_context: 0 dup_mmap_sem &mm->mmap_lock &mm->mmap_lock/1 &anon_vma->rwsem &n->list_lock irq_context: 0 dup_mmap_sem &mm->mmap_lock &mm->mmap_lock/1 &anon_vma->rwsem &n->list_lock &c->lock irq_context: 0 fs_reclaim &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 kn->active#19 &kernfs_locks->open_file_mutex[count] &rq->__lock irq_context: 0 kn->active#26 &kernfs_locks->open_file_mutex[count] &n->list_lock irq_context: 0 kn->active#26 &kernfs_locks->open_file_mutex[count] &n->list_lock &c->lock irq_context: 0 kn->active#22 &n->list_lock irq_context: 0 kn->active#22 &n->list_lock &c->lock irq_context: 0 &sig->cred_guard_mutex sb_writers#4 jbd2_handle &rq->__lock irq_context: 0 &sig->cred_guard_mutex sb_writers#4 jbd2_handle &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 &sig->cred_guard_mutex &iint->mutex &lock->wait_lock irq_context: 0 &sig->cred_guard_mutex &lock->wait_lock irq_context: 0 &sig->cred_guard_mutex &p->pi_lock &rq->__lock irq_context: 0 &sig->cred_guard_mutex &p->pi_lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 kn->active#35 fs_reclaim irq_context: 0 kn->active#35 fs_reclaim mmu_notifier_invalidate_range_start irq_context: 0 kn->active#35 &kernfs_locks->open_file_mutex[count] irq_context: 0 kn->active#35 &kernfs_locks->open_file_mutex[count] fs_reclaim irq_context: 0 kn->active#35 &kernfs_locks->open_file_mutex[count] fs_reclaim mmu_notifier_invalidate_range_start irq_context: 0 &sig->cred_guard_mutex batched_entropy_u8.lock irq_context: 0 &sig->cred_guard_mutex kfence_freelist_lock irq_context: 0 &mm->mmap_lock &folio_wait_table[i] irq_context: 0 &mm->mmap_lock &folio_wait_table[i] &p->pi_lock irq_context: 0 &mm->mmap_lock &folio_wait_table[i] &p->pi_lock &rq->__lock irq_context: 0 &mm->mmap_lock &folio_wait_table[i] &p->pi_lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 mapping.invalidate_lock#2 per_cpu_ptr(&cgroup_rstat_cpu_lock, cpu) irq_context: 0 kn->active#26 &n->list_lock irq_context: 0 kn->active#26 &n->list_lock &c->lock irq_context: 0 &vma->vm_lock->lock per_cpu_ptr(&cgroup_rstat_cpu_lock, cpu) irq_context: 0 mapping.invalidate_lock#2 rcu_read_lock rcu_node_0 irq_context: 0 mapping.invalidate_lock#2 rcu_read_lock &rq->__lock irq_context: 0 mapping.invalidate_lock#2 rcu_read_lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 sb_writers &sb->s_type->i_mutex_key#4 tomoyo_ss remove_cache_srcu irq_context: 0 sb_writers &sb->s_type->i_mutex_key#4 tomoyo_ss remove_cache_srcu quarantine_lock irq_context: 0 sb_writers &sb->s_type->i_mutex_key#4 tomoyo_ss remove_cache_srcu &c->lock irq_context: 0 sb_writers &sb->s_type->i_mutex_key#4 tomoyo_ss remove_cache_srcu &n->list_lock irq_context: 0 sb_writers &sb->s_type->i_mutex_key#4 tomoyo_ss remove_cache_srcu rcu_read_lock pool_lock#2 irq_context: 0 sb_writers &sb->s_type->i_mutex_key#4 tomoyo_ss remove_cache_srcu &obj_hash[i].lock irq_context: 0 &p->lock &of->mutex kn->active#5 remove_cache_srcu irq_context: 0 &p->lock &of->mutex kn->active#5 remove_cache_srcu quarantine_lock irq_context: 0 &p->lock &of->mutex kn->active#5 remove_cache_srcu &c->lock irq_context: 0 &p->lock &of->mutex kn->active#5 remove_cache_srcu &n->list_lock irq_context: 0 &p->lock &of->mutex kn->active#5 remove_cache_srcu rcu_read_lock pool_lock#2 irq_context: 0 &p->lock &of->mutex kn->active#5 remove_cache_srcu &obj_hash[i].lock irq_context: 0 &p->lock &of->mutex kn->active#5 remove_cache_srcu &rq->__lock irq_context: 0 &p->lock &of->mutex kn->active#5 remove_cache_srcu &rq->__lock &cfs_rq->removed.lock irq_context: 0 kn->active#36 fs_reclaim irq_context: 0 kn->active#36 fs_reclaim mmu_notifier_invalidate_range_start irq_context: 0 kn->active#36 &kernfs_locks->open_file_mutex[count] irq_context: 0 kn->active#36 &kernfs_locks->open_file_mutex[count] fs_reclaim irq_context: 0 kn->active#36 &kernfs_locks->open_file_mutex[count] fs_reclaim mmu_notifier_invalidate_range_start irq_context: 0 &p->lock &of->mutex kn->active#36 i2c_dev_list_lock irq_context: 0 &type->i_mutex_dir_key#4 &root->kernfs_rwsem &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 sb_writers#8 &of->mutex kn->active#5 &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 dup_mmap_sem &mm->mmap_lock &mm->mmap_lock/1 &sem->wait_lock irq_context: 0 dup_mmap_sem &mm->mmap_lock &mm->mmap_lock/1 &p->pi_lock irq_context: 0 dup_mmap_sem &mm->mmap_lock &mm->mmap_lock/1 &p->pi_lock &rq->__lock irq_context: 0 dup_mmap_sem &mm->mmap_lock &mm->mmap_lock/1 &p->pi_lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 tomoyo_ss rcu_read_lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 &vma->vm_lock->lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 &mm->mmap_lock rcu_read_lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 kn->active#37 fs_reclaim irq_context: 0 kn->active#37 fs_reclaim mmu_notifier_invalidate_range_start irq_context: 0 kn->active#37 &kernfs_locks->open_file_mutex[count] irq_context: 0 kn->active#37 &kernfs_locks->open_file_mutex[count] fs_reclaim irq_context: 0 kn->active#37 &kernfs_locks->open_file_mutex[count] fs_reclaim mmu_notifier_invalidate_range_start irq_context: 0 kn->active#37 &kernfs_locks->open_file_mutex[count] &c->lock irq_context: 0 kn->active#37 &kernfs_locks->open_file_mutex[count] &____s->seqcount irq_context: 0 kn->active#38 fs_reclaim irq_context: 0 kn->active#38 fs_reclaim mmu_notifier_invalidate_range_start irq_context: 0 kn->active#38 &c->lock irq_context: 0 kn->active#38 &kernfs_locks->open_file_mutex[count] irq_context: 0 kn->active#38 &kernfs_locks->open_file_mutex[count] fs_reclaim irq_context: 0 kn->active#38 &kernfs_locks->open_file_mutex[count] fs_reclaim mmu_notifier_invalidate_range_start irq_context: 0 &mm->mmap_lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 kn->active#34 &n->list_lock irq_context: 0 kn->active#34 &n->list_lock &c->lock irq_context: 0 &vma->vm_lock->lock ptlock_ptr(page) irq_context: 0 &sig->cred_guard_mutex tomoyo_ss remove_cache_srcu irq_context: 0 &sig->cred_guard_mutex tomoyo_ss remove_cache_srcu quarantine_lock irq_context: 0 kn->active#34 &____s->seqcount irq_context: 0 &sig->cred_guard_mutex key#8 irq_context: 0 sb_writers#8 &of->mutex kn->active#5 uevent_sock_mutex quarantine_lock irq_context: 0 &ep->mtx batched_entropy_u8.lock irq_context: 0 &ep->mtx kfence_freelist_lock irq_context: 0 &sig->cred_guard_mutex &sig->exec_update_lock &mm->mmap_lock &n->list_lock irq_context: 0 &mm->mmap_lock remove_cache_srcu rcu_read_lock &____s->seqcount irq_context: 0 &sig->cred_guard_mutex &n->list_lock &c->lock irq_context: 0 &sig->cred_guard_mutex &sig->exec_update_lock &mm->mmap_lock quarantine_lock irq_context: 0 &mm->mmap_lock &mapping->i_mmap_rwsem fill_pool_map-wait-type-override pool_lock#2 irq_context: 0 &mm->mmap_lock &mapping->i_mmap_rwsem fill_pool_map-wait-type-override pool_lock irq_context: 0 videodev_lock irq_context: 0 &dev_instance->mutex irq_context: 0 &dev_instance->mutex fs_reclaim irq_context: 0 &dev_instance->mutex fs_reclaim mmu_notifier_invalidate_range_start irq_context: 0 &dev_instance->mutex &c->lock irq_context: 0 &dev_instance->mutex pool_lock#2 irq_context: 0 &dev_instance->mutex vicodec_core:1844:(hdl)->_lock irq_context: 0 &dev_instance->mutex &vdev->fh_lock irq_context: 0 &mdev->req_queue_mutex irq_context: 0 &mdev->req_queue_mutex &dev_instance->mutex irq_context: 0 &mdev->req_queue_mutex &dev_instance->mutex &m2m_dev->job_spinlock irq_context: 0 &mdev->req_queue_mutex &dev_instance->mutex &q->done_wq irq_context: 0 &mdev->req_queue_mutex &dev_instance->mutex &q->mmap_lock irq_context: 0 &mdev->req_queue_mutex &dev_instance->mutex &obj_hash[i].lock irq_context: 0 &mdev->req_queue_mutex &dev_instance->mutex pool_lock#2 irq_context: 0 &mdev->req_queue_mutex &dev_instance->mutex &rq->__lock irq_context: 0 &mdev->req_queue_mutex &lock->wait_lock irq_context: 0 &mdev->req_queue_mutex &rq->__lock irq_context: 0 &mdev->req_queue_mutex &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 &mdev->req_queue_mutex &vdev->fh_lock irq_context: 0 &mdev->req_queue_mutex &mdev->graph_mutex irq_context: 0 &mdev->req_queue_mutex vicodec_core:1844:(hdl)->_lock irq_context: 0 &mdev->req_queue_mutex vicodec_core:1844:(hdl)->_lock &obj_hash[i].lock irq_context: 0 &mdev->req_queue_mutex vicodec_core:1844:(hdl)->_lock pool_lock#2 irq_context: 0 &mdev->req_queue_mutex &obj_hash[i].lock irq_context: 0 &mdev->req_queue_mutex pool_lock#2 irq_context: 0 &mm->mmap_lock &n->list_lock irq_context: 0 &mm->mmap_lock &n->list_lock &c->lock irq_context: 0 dup_mmap_sem &mm->mmap_lock &mm->mmap_lock/1 &mapping->i_mmap_rwsem &sem->wait_lock irq_context: 0 dup_mmap_sem &mm->mmap_lock &mm->mmap_lock/1 &mapping->i_mmap_rwsem &rq->__lock irq_context: 0 dup_mmap_sem &mm->mmap_lock &mm->mmap_lock/1 &mapping->i_mmap_rwsem &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 &vma->vm_lock->lock pgd_lock irq_context: 0 &vma->vm_lock->lock key irq_context: 0 &vma->vm_lock->lock pcpu_lock irq_context: 0 &vma->vm_lock->lock percpu_counters_lock irq_context: 0 &vma->vm_lock->lock &cfs_rq->removed.lock irq_context: 0 &pipe->rd_wait &ep->lock &ep->wq &p->pi_lock &rq->__lock &cfs_rq->removed.lock irq_context: 0 &mm->mmap_lock &vma->vm_lock->lock &rq->__lock irq_context: 0 fs_reclaim mmu_notifier_invalidate_range_start &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 kn->active#39 fs_reclaim irq_context: 0 kn->active#39 fs_reclaim mmu_notifier_invalidate_range_start irq_context: 0 kn->active#39 &kernfs_locks->open_file_mutex[count] irq_context: 0 kn->active#39 &kernfs_locks->open_file_mutex[count] fs_reclaim irq_context: 0 kn->active#39 &kernfs_locks->open_file_mutex[count] fs_reclaim mmu_notifier_invalidate_range_start irq_context: 0 kn->active#39 &c->lock irq_context: 0 kn->active#39 &n->list_lock irq_context: 0 kn->active#39 &n->list_lock &c->lock irq_context: 0 &mm->mmap_lock &mapping->i_mmap_rwsem &p->pi_lock irq_context: 0 &p->lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 kn->active#5 pool_lock#2 irq_context: 0 &sig->cred_guard_mutex &sig->exec_update_lock &mm->mmap_lock &anon_vma->rwsem &sem->wait_lock irq_context: 0 &sig->cred_guard_mutex &sig->exec_update_lock &mm->mmap_lock &anon_vma->rwsem &rq->__lock irq_context: 0 &sig->cred_guard_mutex &sig->exec_update_lock &mm->mmap_lock &anon_vma->rwsem &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 fh->state->lock irq_context: 0 &vdev->fh_lock irq_context: softirq &(&wb->dwork)->timer irq_context: softirq &(&wb->dwork)->timer rcu_read_lock &pool->lock/1 irq_context: softirq &(&wb->dwork)->timer rcu_read_lock &pool->lock/1 &obj_hash[i].lock irq_context: softirq &(&wb->dwork)->timer rcu_read_lock &pool->lock/1 &p->pi_lock irq_context: softirq &(&wb->dwork)->timer rcu_read_lock &pool->lock/1 &p->pi_lock &rq->__lock irq_context: softirq &(&wb->dwork)->timer rcu_read_lock &pool->lock/1 &p->pi_lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)writeback irq_context: 0 (wq_completion)writeback (work_completion)(&(&wb->dwork)->work) irq_context: 0 (wq_completion)writeback (work_completion)(&(&wb->dwork)->work) &wb->work_lock irq_context: 0 (wq_completion)writeback (work_completion)(&(&wb->dwork)->work) &wb->list_lock irq_context: 0 (wq_completion)writeback (work_completion)(&(&wb->dwork)->work) &p->sequence irq_context: 0 (wq_completion)writeback (work_completion)(&(&wb->dwork)->work) key#10 irq_context: 0 (wq_completion)writeback (work_completion)(&(&wb->dwork)->work) &wb->work_lock &obj_hash[i].lock irq_context: 0 (wq_completion)writeback (work_completion)(&(&wb->dwork)->work) &wb->work_lock &base->lock irq_context: 0 (wq_completion)writeback (work_completion)(&(&wb->dwork)->work) &wb->work_lock &base->lock &obj_hash[i].lock irq_context: 0 &dev->dev_mutex irq_context: 0 &dev->dev_mutex fs_reclaim irq_context: 0 &dev->dev_mutex fs_reclaim mmu_notifier_invalidate_range_start irq_context: 0 &dev->dev_mutex pool_lock#2 irq_context: 0 &dev->dev_mutex vim2m:1183:(hdl)->_lock irq_context: 0 &dev->dev_mutex &c->lock irq_context: 0 &dev->dev_mutex &obj_hash[i].lock irq_context: 0 &dev->dev_mutex &vdev->fh_lock irq_context: 0 &mdev->req_queue_mutex vim2m:1183:(hdl)->_lock irq_context: 0 &mdev->req_queue_mutex vim2m:1183:(hdl)->_lock &obj_hash[i].lock irq_context: 0 &mdev->req_queue_mutex vim2m:1183:(hdl)->_lock pool_lock#2 irq_context: 0 &mdev->req_queue_mutex &dev->dev_mutex irq_context: 0 &mdev->req_queue_mutex &dev->dev_mutex &m2m_dev->job_spinlock irq_context: 0 &mdev->req_queue_mutex &dev->dev_mutex &q->done_wq irq_context: 0 &mdev->req_queue_mutex &dev->dev_mutex &q->mmap_lock irq_context: 0 &mdev->req_queue_mutex &dev->dev_mutex &obj_hash[i].lock irq_context: 0 &mdev->req_queue_mutex &dev->dev_mutex pool_lock#2 irq_context: 0 &mm->mmap_lock &mapping->i_mmap_rwsem &rq->__lock &cfs_rq->removed.lock irq_context: 0 &ep->mtx &meta->lock irq_context: 0 tasklist_lock &sighand->siglock &sighand->signalfd_wqh &p->pi_lock irq_context: 0 tasklist_lock &sighand->siglock &sighand->signalfd_wqh &p->pi_lock &rq->__lock irq_context: 0 tasklist_lock &sighand->siglock &sighand->signalfd_wqh &p->pi_lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 &vma->vm_lock->lock rcu_read_lock rcu_node_0 irq_context: 0 &vma->vm_lock->lock rcu_read_lock &rq->__lock irq_context: 0 &vma->vm_lock->lock rcu_read_lock &cfs_rq->removed.lock irq_context: 0 &vma->vm_lock->lock rcu_read_lock &obj_hash[i].lock irq_context: 0 (wq_completion)events (work_completion)(&(&krcp->monitor_work)->work) rcu_callback &base->lock irq_context: 0 (wq_completion)events (work_completion)(&(&krcp->monitor_work)->work) rcu_callback &base->lock &obj_hash[i].lock irq_context: 0 tasklist_lock &sighand->siglock &sighand->signalfd_wqh &p->pi_lock &cfs_rq->removed.lock irq_context: 0 &mm->mmap_lock rcu_read_lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 &vcapture->lock irq_context: 0 &mdev->graph_mutex irq_context: 0 sb_writers#8 &of->mutex kn->active#5 uevent_sock_mutex rcu_node_0 irq_context: 0 &kernfs_locks->open_file_mutex[count] &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 &pipe->mutex/1 &pipe->rd_wait &ep->lock &ep->wq &p->pi_lock &rq->__lock &cfs_rq->removed.lock irq_context: 0 &mm->mmap_lock rcu_read_lock &cfs_rq->removed.lock irq_context: 0 &mm->mmap_lock rcu_read_lock &obj_hash[i].lock irq_context: 0 &mm->mmap_lock &mapping->i_mmap_rwsem &anon_vma->rwsem &rq->__lock irq_context: 0 sb_writers#5 &type->i_mutex_dir_key#5/1 &sb->s_type->i_mutex_key#12 &simple_offset_xa_lock &c->lock irq_context: 0 sb_writers#5 &type->i_mutex_dir_key#5 tomoyo_ss quarantine_lock irq_context: 0 kn->active#5 &kernfs_locks->open_file_mutex[count] remove_cache_srcu &c->lock irq_context: 0 kn->active#5 &kernfs_locks->open_file_mutex[count] remove_cache_srcu &n->list_lock irq_context: 0 kn->active#5 &kernfs_locks->open_file_mutex[count] remove_cache_srcu &obj_hash[i].lock irq_context: 0 dup_mmap_sem &mm->mmap_lock &mm->mmap_lock/1 &anon_vma->rwsem &rq->__lock irq_context: 0 dup_mmap_sem &mm->mmap_lock &mm->mmap_lock/1 &anon_vma->rwsem &rq->__lock &cfs_rq->removed.lock irq_context: 0 dup_mmap_sem &mm->mmap_lock &mm->mmap_lock/1 &anon_vma->rwsem &sem->wait_lock irq_context: 0 dup_mmap_sem &mm->mmap_lock &mm->mmap_lock/1 &anon_vma->rwsem &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 &mm->mmap_lock &rq->__lock &cfs_rq->removed.lock irq_context: 0 &mm->mmap_lock &anon_vma->rwsem &rq->__lock irq_context: 0 dup_mmap_sem &mm->mmap_lock &mm->mmap_lock/1 fs_reclaim &rq->__lock irq_context: 0 dup_mmap_sem &mm->mmap_lock &mm->mmap_lock/1 fs_reclaim &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 &mm->mmap_lock &mapping->i_mmap_rwsem &base->lock irq_context: 0 &mm->mmap_lock &mapping->i_mmap_rwsem &base->lock &obj_hash[i].lock irq_context: 0 &mm->mmap_lock &p->pi_lock &rq->__lock &cfs_rq->removed.lock irq_context: 0 &mm->mmap_lock &mapping->i_mmap_rwsem &cfs_rq->removed.lock irq_context: 0 &dev->mutex#3 irq_context: 0 &mdev->req_queue_mutex &dev->mutex#3 irq_context: 0 &mdev->req_queue_mutex &dev->mutex#3 &vdev->fh_lock irq_context: 0 tomoyo_ss rcu_read_lock rcu_read_lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 &mm->mmap_lock pgd_lock irq_context: 0 &mm->mmap_lock key irq_context: 0 &mm->mmap_lock pcpu_lock irq_context: 0 &mm->mmap_lock percpu_counters_lock irq_context: 0 &ep->mtx &obj_hash[i].lock pool_lock irq_context: 0 kn->active#39 &____s->seqcount irq_context: 0 &mm->mmap_lock rcu_read_lock rcu_read_lock &cfs_rq->removed.lock irq_context: 0 dup_mmap_sem &mm->mmap_lock &mm->mmap_lock/1 &mapping->i_mmap_rwsem &cfs_rq->removed.lock irq_context: 0 dup_mmap_sem &mm->mmap_lock &mm->mmap_lock/1 &mapping->i_mmap_rwsem &obj_hash[i].lock irq_context: 0 dup_mmap_sem &mm->mmap_lock &mm->mmap_lock/1 &mapping->i_mmap_rwsem pool_lock#2 irq_context: 0 dup_mmap_sem &mm->mmap_lock &mm->mmap_lock/1 &p->pi_lock &rq->__lock &cfs_rq->removed.lock irq_context: 0 &mm->mmap_lock &meta->lock irq_context: 0 kn->active#39 &kernfs_locks->open_file_mutex[count] &c->lock irq_context: 0 kn->active#39 &kernfs_locks->open_file_mutex[count] &____s->seqcount irq_context: 0 kn->active#34 remove_cache_srcu irq_context: 0 kn->active#34 remove_cache_srcu quarantine_lock irq_context: 0 kn->active#34 remove_cache_srcu &c->lock irq_context: 0 kn->active#34 remove_cache_srcu &n->list_lock irq_context: 0 kn->active#34 remove_cache_srcu &obj_hash[i].lock irq_context: 0 rcu_read_lock &vma->vm_lock->lock &rq->__lock irq_context: 0 rcu_read_lock &vma->vm_lock->lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 &mm->mmap_lock rcu_read_lock pgd_lock irq_context: 0 &mm->mmap_lock rcu_read_lock key irq_context: 0 &mm->mmap_lock rcu_read_lock pcpu_lock irq_context: 0 &mm->mmap_lock rcu_read_lock percpu_counters_lock irq_context: 0 &mm->mmap_lock rcu_read_lock &rcu_state.gp_wq &p->pi_lock &rq->__lock &cfs_rq->removed.lock irq_context: 0 &sig->cred_guard_mutex &sig->exec_update_lock &mm->mmap_lock &rq->__lock irq_context: 0 &sig->cred_guard_mutex &sig->exec_update_lock &mm->mmap_lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 rcu_read_lock &vma->vm_lock->lock &rcu_state.gp_wq irq_context: 0 rcu_read_lock &vma->vm_lock->lock &rcu_state.gp_wq &p->pi_lock irq_context: 0 rcu_read_lock &vma->vm_lock->lock &rcu_state.gp_wq &p->pi_lock &cfs_rq->removed.lock irq_context: 0 rcu_read_lock &vma->vm_lock->lock &rcu_state.gp_wq &p->pi_lock &rq->__lock irq_context: 0 rcu_read_lock &vma->vm_lock->lock &rcu_state.gp_wq &p->pi_lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: softirq (&journal->j_commit_timer) irq_context: softirq (&journal->j_commit_timer) &p->pi_lock irq_context: softirq (&journal->j_commit_timer) &p->pi_lock &rq->__lock irq_context: softirq (&journal->j_commit_timer) &p->pi_lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 &journal->j_checkpoint_mutex irq_context: 0 &journal->j_checkpoint_mutex mmu_notifier_invalidate_range_start irq_context: 0 &journal->j_checkpoint_mutex pool_lock#2 irq_context: 0 &journal->j_checkpoint_mutex tk_core.seq.seqcount irq_context: 0 &journal->j_checkpoint_mutex &dd->lock irq_context: 0 &journal->j_checkpoint_mutex &obj_hash[i].lock irq_context: 0 &journal->j_checkpoint_mutex &base->lock irq_context: 0 &journal->j_checkpoint_mutex &base->lock &obj_hash[i].lock irq_context: 0 &journal->j_checkpoint_mutex rcu_read_lock &pool->lock irq_context: 0 &journal->j_checkpoint_mutex rcu_read_lock &pool->lock &obj_hash[i].lock irq_context: 0 &journal->j_checkpoint_mutex rcu_read_lock &pool->lock pool_lock#2 irq_context: 0 &journal->j_checkpoint_mutex rcu_read_lock &pool->lock &p->pi_lock irq_context: 0 &journal->j_checkpoint_mutex rcu_read_lock &pool->lock &p->pi_lock &rq->__lock irq_context: 0 &journal->j_checkpoint_mutex rcu_read_lock &pool->lock &p->pi_lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 &journal->j_checkpoint_mutex bit_wait_table + i irq_context: 0 &journal->j_checkpoint_mutex &rq->__lock irq_context: 0 &journal->j_checkpoint_mutex &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 &journal->j_checkpoint_mutex &journal->j_state_lock irq_context: 0 &journal->j_state_lock &journal->j_wait_updates irq_context: 0 &journal->j_list_lock irq_context: 0 &journal->j_state_lock tk_core.seq.seqcount irq_context: 0 &journal->j_state_lock &journal->j_wait_transaction_locked irq_context: 0 &ei->i_es_lock irq_context: 0 &ei->i_es_lock key#2 irq_context: 0 lock#4 irq_context: 0 &mapping->private_lock irq_context: 0 &ret->b_state_lock irq_context: 0 &ret->b_state_lock &journal->j_list_lock irq_context: 0 &dd->lock irq_context: 0 &journal->j_state_lock irq_context: 0 &journal->j_state_lock &journal->j_list_lock irq_context: 0 rcu_read_lock &dd->lock irq_context: 0 &sig->cred_guard_mutex &sig->exec_update_lock &mm->mmap_lock ptlock_ptr(page)#2 &pcp->lock &zone->lock irq_context: hardirq &x->wait#5 irq_context: 0 (wq_completion)kblockd (work_completion)(&(&q->requeue_work)->work) &dd->lock irq_context: 0 (wq_completion)kblockd (work_completion)(&(&q->requeue_work)->work) rcu_read_lock &dd->lock irq_context: 0 (wq_completion)kblockd (work_completion)(&(&q->requeue_work)->work) rcu_read_lock &obj_hash[i].lock pool_lock irq_context: 0 &mm->mmap_lock rcu_node_0 irq_context: 0 dup_mmap_sem &mm->mmap_lock &mm->mmap_lock/1 &p->pi_lock &cfs_rq->removed.lock irq_context: 0 &ret->b_state_lock &journal->j_list_lock rcu_read_lock &memcg->move_lock irq_context: 0 &ret->b_state_lock &journal->j_list_lock rcu_read_lock &xa->xa_lock#9 irq_context: 0 &ret->b_state_lock &journal->j_list_lock &sb->s_type->i_lock_key#3 irq_context: 0 &ret->b_state_lock &journal->j_list_lock &wb->list_lock irq_context: 0 &ret->b_state_lock &journal->j_list_lock &wb->list_lock &sb->s_type->i_lock_key#3 irq_context: 0 &ret->b_state_lock &journal->j_list_lock rcu_read_lock &xa->xa_lock#9 key#10 irq_context: 0 &journal->j_state_lock &journal->j_list_lock irq_context: 0 &sbi->s_md_lock irq_context: 0 &journal->j_fc_wait irq_context: 0 &journal->j_history_lock irq_context: 0 dup_mmap_sem &mm->mmap_lock &mm->mmap_lock/1 batched_entropy_u8.lock irq_context: 0 dup_mmap_sem &mm->mmap_lock &mm->mmap_lock/1 kfence_freelist_lock irq_context: 0 &u->bindlock irq_context: 0 &u->bindlock fs_reclaim irq_context: 0 &u->bindlock fs_reclaim mmu_notifier_invalidate_range_start irq_context: 0 &u->bindlock pool_lock#2 irq_context: 0 &u->bindlock batched_entropy_u32.lock irq_context: 0 &u->bindlock &net->unx.table.locks[i] irq_context: 0 &u->bindlock &net->unx.table.locks[i] &net->unx.table.locks[i]/1 irq_context: 0 &u->bindlock &net->unx.table.locks[i]/1 irq_context: 0 &u->lock &u->lock/1 &sk->sk_peer_lock irq_context: 0 &u->lock &u->lock/1 &dentry->d_lock irq_context: 0 &u->lock &u->lock/1 &sk->sk_peer_lock &sk->sk_peer_lock/1 irq_context: 0 &u->lock &u->lock/1 &sk->sk_peer_lock/1 irq_context: 0 &sig->cred_guard_mutex &sig->exec_update_lock &mm->mmap_lock &meta->lock irq_context: 0 &sig->cred_guard_mutex &sig->exec_update_lock &mm->mmap_lock kfence_freelist_lock irq_context: 0 &u->iolock rcu_read_lock &ei->socket.wq.wait irq_context: 0 &sb->s_type->i_mutex_key#10 rcu_read_lock &ei->socket.wq.wait irq_context: 0 &sb->s_type->i_mutex_key#10 rcu_read_lock &ei->socket.wq.wait &p->pi_lock irq_context: 0 &sb->s_type->i_mutex_key#10 rcu_read_lock &ei->socket.wq.wait &p->pi_lock &rq->__lock irq_context: 0 &sb->s_type->i_mutex_key#10 rcu_read_lock &ei->socket.wq.wait &p->pi_lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 &sb->s_type->i_mutex_key#10 rcu_read_lock &dentry->d_lock irq_context: 0 &group->mark_mutex &sb->s_type->i_lock_key &dentry->d_lock &dentry->d_lock/1 irq_context: 0 tomoyo_ss pgd_lock irq_context: 0 tomoyo_ss key irq_context: 0 tomoyo_ss pcpu_lock irq_context: 0 tomoyo_ss percpu_counters_lock irq_context: 0 tomoyo_ss &cfs_rq->removed.lock irq_context: 0 &mm->mmap_lock &mapping->i_mmap_rwsem rcu_read_lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 &mm->mmap_lock &anon_vma->rwsem &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 kn->active#34 &kernfs_locks->open_file_mutex[count] &c->lock irq_context: 0 dup_mmap_sem &mm->mmap_lock &mm->mmap_lock/1 &mapping->i_mmap_rwsem &rq->__lock &cfs_rq->removed.lock irq_context: 0 cgroup_threadgroup_rwsem &rq->__lock &cfs_rq->removed.lock irq_context: 0 dup_mmap_sem &mm->mmap_lock &mm->mmap_lock/1 &pcp->lock &zone->lock &____s->seqcount irq_context: 0 kn->active#34 batched_entropy_u8.lock irq_context: 0 kn->active#34 kfence_freelist_lock irq_context: 0 kn->active#5 &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 &sig->cred_guard_mutex &sig->exec_update_lock &mm->mmap_lock &vma->vm_lock->lock &rq->__lock irq_context: 0 &sig->cred_guard_mutex &sig->exec_update_lock &mm->mmap_lock &vma->vm_lock->lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 &sig->cred_guard_mutex &sig->exec_update_lock &mm->mmap_lock &anon_vma->rwsem quarantine_lock irq_context: 0 &sig->cred_guard_mutex &sig->exec_update_lock &mm->mmap_lock &mapping->i_mmap_rwsem &rq->__lock &cfs_rq->removed.lock irq_context: 0 &ep->mtx key#11 irq_context: 0 kn->active#39 remove_cache_srcu irq_context: 0 kn->active#39 remove_cache_srcu quarantine_lock irq_context: 0 kn->active#39 remove_cache_srcu &c->lock irq_context: 0 kn->active#39 remove_cache_srcu &n->list_lock irq_context: 0 kn->active#39 remove_cache_srcu &obj_hash[i].lock irq_context: 0 &mm->mmap_lock &mapping->i_mmap_rwsem &anon_vma->rwsem &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 remove_cache_srcu rcu_read_lock &rcu_state.gp_wq irq_context: 0 remove_cache_srcu rcu_read_lock &rcu_state.gp_wq &p->pi_lock irq_context: 0 remove_cache_srcu rcu_read_lock &rcu_state.gp_wq &p->pi_lock &cfs_rq->removed.lock irq_context: 0 remove_cache_srcu rcu_read_lock &rcu_state.gp_wq &p->pi_lock &rq->__lock irq_context: 0 remove_cache_srcu rcu_read_lock &rcu_state.gp_wq &p->pi_lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 kn->active#39 &kernfs_locks->open_file_mutex[count] &n->list_lock irq_context: 0 kn->active#39 &kernfs_locks->open_file_mutex[count] &n->list_lock &c->lock irq_context: 0 dup_mmap_sem &mm->mmap_lock &mm->mmap_lock/1 rcu_read_lock rcu_node_0 irq_context: 0 dup_mmap_sem &mm->mmap_lock &mm->mmap_lock/1 rcu_read_lock &rq->__lock irq_context: 0 dup_mmap_sem &mm->mmap_lock &mm->mmap_lock/1 rcu_read_lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 &sig->cred_guard_mutex tomoyo_ss remove_cache_srcu &c->lock irq_context: 0 &sig->cred_guard_mutex tomoyo_ss remove_cache_srcu &n->list_lock irq_context: 0 &sig->cred_guard_mutex tomoyo_ss remove_cache_srcu rcu_read_lock pool_lock#2 irq_context: 0 &sig->cred_guard_mutex tomoyo_ss remove_cache_srcu &obj_hash[i].lock irq_context: 0 &mm->mmap_lock fs_reclaim &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 &pipe->rd_wait &ep->lock &ep->wq &p->pi_lock &cfs_rq->removed.lock irq_context: 0 cgroup_threadgroup_rwsem &cfs_rq->removed.lock irq_context: 0 &mm->mmap_lock &mapping->i_mmap_rwsem rcu_node_0 irq_context: 0 &p->lock &of->mutex kn->active#39 &rq->__lock irq_context: 0 cgroup_threadgroup_rwsem &obj_hash[i].lock irq_context: 0 &sig->cred_guard_mutex &sig->exec_update_lock &mm->mmap_lock &mapping->i_mmap_rwsem &p->pi_lock irq_context: 0 kn->active#34 &kernfs_locks->open_file_mutex[count] &____s->seqcount irq_context: 0 remove_cache_srcu &pcp->lock &zone->lock irq_context: 0 remove_cache_srcu &pcp->lock &zone->lock &____s->seqcount irq_context: 0 &pipe->mutex/1 &pcp->lock &zone->lock irq_context: 0 &mm->mmap_lock fill_pool_map-wait-type-override &rq->__lock irq_context: 0 &mm->mmap_lock fill_pool_map-wait-type-override &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 &mm->mmap_lock &c->lock batched_entropy_u8.lock irq_context: 0 &mm->mmap_lock &c->lock kfence_freelist_lock irq_context: 0 &mm->mmap_lock &mapping->i_mmap_rwsem &anon_vma->rwsem &meta->lock irq_context: 0 &mm->mmap_lock &mapping->i_mmap_rwsem &anon_vma->rwsem kfence_freelist_lock irq_context: 0 dup_mmap_sem &mm->mmap_lock &mm->mmap_lock/1 rcu_read_lock &rcu_state.gp_wq irq_context: 0 dup_mmap_sem &mm->mmap_lock &mm->mmap_lock/1 rcu_read_lock &rcu_state.gp_wq &p->pi_lock irq_context: 0 dup_mmap_sem &mm->mmap_lock &mm->mmap_lock/1 rcu_read_lock &rcu_state.gp_wq &p->pi_lock &cfs_rq->removed.lock irq_context: 0 dup_mmap_sem &mm->mmap_lock &mm->mmap_lock/1 rcu_read_lock &rcu_state.gp_wq &p->pi_lock &rq->__lock irq_context: 0 dup_mmap_sem &mm->mmap_lock &mm->mmap_lock/1 rcu_read_lock &rcu_state.gp_wq &p->pi_lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 &mm->mmap_lock &rq->__lock &cfs_rq->removed.lock irq_context: 0 &mm->mmap_lock mmu_notifier_invalidate_range_start &rq->__lock irq_context: 0 rcu_read_lock rcu_read_lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 &sig->cred_guard_mutex &sig->exec_update_lock &mm->mmap_lock &rq->__lock &cfs_rq->removed.lock irq_context: 0 &sig->cred_guard_mutex &sig->exec_update_lock quarantine_lock irq_context: 0 &mm->mmap_lock &p->pi_lock &cfs_rq->removed.lock irq_context: 0 &mm->mmap_lock rcu_read_lock rcu_read_lock &obj_hash[i].lock irq_context: 0 &mm->mmap_lock rcu_read_lock rcu_read_lock &obj_hash[i].lock pool_lock irq_context: 0 &mm->mmap_lock rcu_read_lock rcu_read_lock pool_lock#2 irq_context: 0 &mm->mmap_lock rcu_read_lock &rcu_state.gp_wq &p->pi_lock &rq->__lock &cfs_rq->removed.lock irq_context: 0 &mm->mmap_lock fs_reclaim mmu_notifier_invalidate_range_start &rq->__lock irq_context: 0 cgroup_threadgroup_rwsem pool_lock#2 irq_context: 0 &sig->cred_guard_mutex &sig->exec_update_lock &mm->mmap_lock &mapping->i_mmap_rwsem rcu_read_lock &rq->__lock irq_context: 0 &sig->cred_guard_mutex &sig->exec_update_lock &mm->mmap_lock &mapping->i_mmap_rwsem rcu_read_lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 dup_mmap_sem &mm->mmap_lock &mm->mmap_lock/1 &anon_vma->rwsem rcu_read_lock rcu_node_0 irq_context: 0 dup_mmap_sem &mm->mmap_lock &mm->mmap_lock/1 &anon_vma->rwsem rcu_read_lock &rq->__lock irq_context: 0 &mm->mmap_lock &mapping->i_mmap_rwsem rcu_read_lock &rq->__lock &cfs_rq->removed.lock irq_context: 0 &f->f_pos_lock &type->i_mutex_dir_key#5 rcu_read_lock &dentry->d_lock irq_context: 0 sb_writers &type->i_mutex_dir_key/1 &sb->s_type->i_mutex_key#4 &sb->s_type->i_mutex_key#4/4 irq_context: 0 sb_writers &type->i_mutex_dir_key/1 &sb->s_type->i_mutex_key#4 &sb->s_type->i_mutex_key#4/4 &dentry->d_lock irq_context: 0 sb_writers &type->i_mutex_dir_key/1 &sb->s_type->i_mutex_key#4 &sb->s_type->i_mutex_key#4/4 &simple_offset_xa_lock irq_context: 0 sb_writers &type->i_mutex_dir_key/1 &sb->s_type->i_mutex_key#4 &sb->s_type->i_mutex_key#4/4 fs_reclaim irq_context: 0 sb_writers &type->i_mutex_dir_key/1 &sb->s_type->i_mutex_key#4 &sb->s_type->i_mutex_key#4/4 fs_reclaim mmu_notifier_invalidate_range_start irq_context: 0 sb_writers &type->i_mutex_dir_key/1 &sb->s_type->i_mutex_key#4 &sb->s_type->i_mutex_key#4/4 tk_core.seq.seqcount irq_context: 0 sb_writers &type->i_mutex_dir_key/1 &sb->s_type->i_mutex_key#4 &sb->s_type->i_mutex_key#4/4 rcu_read_lock &dentry->d_lock irq_context: 0 sb_writers &type->i_mutex_dir_key/1 &sb->s_type->i_mutex_key#4 &sb->s_type->i_mutex_key#4/4 rename_lock irq_context: 0 sb_writers &type->i_mutex_dir_key/1 &sb->s_type->i_mutex_key#4 &sb->s_type->i_mutex_key#4/4 rename_lock rename_lock.seqcount irq_context: 0 sb_writers &type->i_mutex_dir_key/1 &sb->s_type->i_mutex_key#4 &sb->s_type->i_mutex_key#4/4 rename_lock rename_lock.seqcount &dentry->d_lock irq_context: 0 sb_writers &type->i_mutex_dir_key/1 &sb->s_type->i_mutex_key#4 &sb->s_type->i_mutex_key#4/4 rename_lock rename_lock.seqcount &dentry->d_lock &dentry->d_lock/2 irq_context: 0 sb_writers &type->i_mutex_dir_key/1 &sb->s_type->i_mutex_key#4 &sb->s_type->i_mutex_key#4/4 rename_lock rename_lock.seqcount &dentry->d_lock &dentry->d_lock/2 &dentry->d_lock/3 irq_context: 0 sb_writers &type->i_mutex_dir_key/1 &sb->s_type->i_mutex_key#4 &sb->s_type->i_mutex_key#4/4 rename_lock rename_lock.seqcount &dentry->d_lock &dentry->d_lock/2 &dentry->d_lock/3 &____s->seqcount#4 irq_context: 0 sb_writers &type->i_mutex_dir_key/1 &sb->s_type->i_mutex_key#4 &sb->s_type->i_mutex_key#4/4 rename_lock rename_lock.seqcount &dentry->d_lock &dentry->d_lock/2 &dentry->d_lock/3 &____s->seqcount#4 &____s->seqcount#4/1 irq_context: 0 sb_writers &type->i_mutex_dir_key/1 &sb->s_type->i_mutex_key#4 &sb->s_type->i_mutex_key#4/4 rename_lock rename_lock.seqcount &dentry->d_lock/2 irq_context: 0 sb_writers &type->i_mutex_dir_key/1 &sb->s_type->i_mutex_key#4 &sb->s_type->i_mutex_key#4/4 rename_lock rename_lock.seqcount &dentry->d_lock/2 &dentry->d_lock/3 irq_context: 0 sb_writers &type->i_mutex_dir_key/1 krc.lock irq_context: 0 sb_writers &type->i_mutex_dir_key/1 &xa->xa_lock#9 irq_context: 0 &mm->mmap_lock &mapping->i_mmap_rwsem pgd_lock irq_context: 0 &mm->mmap_lock &mapping->i_mmap_rwsem key irq_context: 0 &mm->mmap_lock &mapping->i_mmap_rwsem pcpu_lock irq_context: 0 &mm->mmap_lock &mapping->i_mmap_rwsem percpu_counters_lock irq_context: 0 tasklist_lock &sighand->siglock batched_entropy_u8.lock irq_context: 0 tasklist_lock &sighand->siglock kfence_freelist_lock irq_context: 0 &sighand->siglock &meta->lock irq_context: 0 &sighand->siglock kfence_freelist_lock irq_context: 0 sb_writers &type->i_mutex_dir_key/1 &obj_hash[i].lock pool_lock irq_context: 0 tomoyo_ss remove_cache_srcu rcu_read_lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 &f->f_pos_lock &type->i_mutex_dir_key#5 &mm->mmap_lock fs_reclaim irq_context: 0 &f->f_pos_lock &type->i_mutex_dir_key#5 &mm->mmap_lock fs_reclaim mmu_notifier_invalidate_range_start irq_context: 0 &f->f_pos_lock &type->i_mutex_dir_key#5 &mm->mmap_lock &____s->seqcount irq_context: 0 &f->f_pos_lock &type->i_mutex_dir_key#5 &mm->mmap_lock ptlock_ptr(page)#2 irq_context: 0 &f->f_pos_lock &type->i_mutex_dir_key#5 &mm->mmap_lock ptlock_ptr(page)#2 lock#4 irq_context: 0 sb_writers &type->i_mutex_dir_key/1 &sb->s_type->i_mutex_key#4/4 irq_context: 0 &sig->cred_guard_mutex &sig->exec_update_lock &mm->mmap_lock lock#4 &lruvec->lru_lock per_cpu_ptr(&cgroup_rstat_cpu_lock, cpu) irq_context: 0 &mm->mmap_lock remove_cache_srcu rcu_read_lock rcu_node_0 irq_context: 0 &mm->mmap_lock remove_cache_srcu rcu_read_lock &rq->__lock irq_context: 0 &sig->cred_guard_mutex &sig->exec_update_lock &mm->mmap_lock &mapping->i_mmap_rwsem &cfs_rq->removed.lock irq_context: 0 &sig->cred_guard_mutex &sig->exec_update_lock &mm->mmap_lock &mapping->i_mmap_rwsem &obj_hash[i].lock irq_context: 0 &mm->mmap_lock remove_cache_srcu rcu_read_lock &rcu_state.gp_wq irq_context: 0 &mm->mmap_lock remove_cache_srcu rcu_read_lock &rcu_state.gp_wq &p->pi_lock irq_context: 0 &mm->mmap_lock remove_cache_srcu rcu_read_lock &rcu_state.gp_wq &p->pi_lock &rq->__lock irq_context: 0 &mm->mmap_lock remove_cache_srcu rcu_read_lock &rcu_state.gp_wq &p->pi_lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 &mm->mmap_lock fs_reclaim &rq->__lock irq_context: 0 &sig->cred_guard_mutex &sig->exec_update_lock &rq->__lock irq_context: 0 kn->active#5 remove_cache_srcu &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 sb_writers &type->i_mutex_dir_key/1 tomoyo_ss &rq->__lock irq_context: 0 sb_writers &type->i_mutex_dir_key/1 tomoyo_ss &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: softirq drivers/base/dd.c:321 irq_context: softirq drivers/base/dd.c:321 rcu_read_lock &pool->lock irq_context: softirq drivers/base/dd.c:321 rcu_read_lock &pool->lock &obj_hash[i].lock irq_context: softirq drivers/base/dd.c:321 rcu_read_lock &pool->lock &p->pi_lock irq_context: softirq drivers/base/dd.c:321 rcu_read_lock &pool->lock &p->pi_lock &rq->__lock irq_context: softirq drivers/base/dd.c:321 rcu_read_lock &pool->lock &p->pi_lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)events (deferred_probe_timeout_work).work irq_context: 0 (wq_completion)events (deferred_probe_timeout_work).work device_links_lock irq_context: 0 (wq_completion)events (deferred_probe_timeout_work).work device_links_lock &k->list_lock irq_context: 0 (wq_completion)events (deferred_probe_timeout_work).work device_links_lock &k->k_lock irq_context: 0 (wq_completion)events (deferred_probe_timeout_work).work deferred_probe_mutex irq_context: 0 (wq_completion)events (deferred_probe_timeout_work).work rcu_read_lock &pool->lock/1 irq_context: 0 (wq_completion)events (deferred_probe_timeout_work).work rcu_read_lock &pool->lock/1 &obj_hash[i].lock irq_context: 0 (wq_completion)events (deferred_probe_timeout_work).work rcu_read_lock &pool->lock/1 &p->pi_lock irq_context: 0 (wq_completion)events (deferred_probe_timeout_work).work rcu_read_lock &pool->lock/1 &p->pi_lock &rq->__lock irq_context: 0 (wq_completion)events (deferred_probe_timeout_work).work rcu_read_lock &pool->lock/1 &p->pi_lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)events (deferred_probe_timeout_work).work deferred_probe_work irq_context: 0 (wq_completion)events (deferred_probe_timeout_work).work &x->wait#10 irq_context: 0 (wq_completion)events (deferred_probe_timeout_work).work &pool->lock irq_context: 0 (wq_completion)events (deferred_probe_timeout_work).work &rq->__lock irq_context: 0 (wq_completion)events (deferred_probe_timeout_work).work &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)events_unbound (work_completion)(&barr->work) &x->wait#10 &p->pi_lock &rq->__lock irq_context: 0 (wq_completion)events_unbound (work_completion)(&barr->work) &x->wait#10 &p->pi_lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)events (deferred_probe_timeout_work).work &obj_hash[i].lock irq_context: 0 remove_cache_srcu &cfs_rq->removed.lock irq_context: 0 remove_cache_srcu pool_lock#2 irq_context: 0 &sig->cred_guard_mutex &sig->exec_update_lock &mm->mmap_lock ptlock_ptr(page)#2 &pcp->lock &zone->lock &____s->seqcount irq_context: 0 &p->lock &of->mutex kn->active#5 &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 &p->lock &mm->mmap_lock fs_reclaim irq_context: 0 &p->lock &mm->mmap_lock fs_reclaim mmu_notifier_invalidate_range_start irq_context: 0 &p->lock &mm->mmap_lock &____s->seqcount irq_context: 0 &p->lock &mm->mmap_lock ptlock_ptr(page)#2 irq_context: 0 &p->lock &mm->mmap_lock ptlock_ptr(page)#2 lock#4 irq_context: 0 dup_mmap_sem &mm->mmap_lock &mm->mmap_lock/1 &vma->vm_lock->lock &rq->__lock irq_context: 0 dup_mmap_sem &mm->mmap_lock &mm->mmap_lock/1 &vma->vm_lock->lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 cgroup_threadgroup_rwsem pgd_lock irq_context: 0 cgroup_threadgroup_rwsem rcu_read_lock pool_lock#2 irq_context: 0 cgroup_threadgroup_rwsem key irq_context: 0 cgroup_threadgroup_rwsem pcpu_lock irq_context: 0 cgroup_threadgroup_rwsem percpu_counters_lock irq_context: 0 &mm->mmap_lock &mapping->i_mmap_rwsem fill_pool_map-wait-type-override &c->lock irq_context: 0 &mm->mmap_lock &mapping->i_mmap_rwsem fill_pool_map-wait-type-override &____s->seqcount irq_context: 0 &mm->mmap_lock fill_pool_map-wait-type-override pool_lock#2 irq_context: 0 &mm->mmap_lock fill_pool_map-wait-type-override pool_lock irq_context: 0 &ep->mtx fill_pool_map-wait-type-override pool_lock#2 irq_context: 0 &ep->mtx fill_pool_map-wait-type-override pool_lock irq_context: 0 &mm->mmap_lock fs_reclaim mmu_notifier_invalidate_range_start &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 &mm->mmap_lock &mapping->i_mmap_rwsem &anon_vma->rwsem fill_pool_map-wait-type-override pool_lock#2 irq_context: 0 &mm->mmap_lock &mapping->i_mmap_rwsem &anon_vma->rwsem fill_pool_map-wait-type-override &c->lock irq_context: 0 &mm->mmap_lock &mapping->i_mmap_rwsem &anon_vma->rwsem fill_pool_map-wait-type-override &____s->seqcount irq_context: 0 &mm->mmap_lock &mapping->i_mmap_rwsem &anon_vma->rwsem fill_pool_map-wait-type-override pool_lock irq_context: 0 dup_mmap_sem &mm->mmap_lock &mm->mmap_lock/1 quarantine_lock irq_context: 0 sb_writers &type->i_mutex_dir_key/1 &sb->s_type->i_mutex_key#4 &sb->s_type->i_mutex_key#4/4 &simple_offset_xa_lock pool_lock#2 irq_context: 0 sb_writers &type->i_mutex_dir_key/1 &sb->s_type->i_mutex_key#4 &sb->s_type->i_mutex_key#4/4 &simple_offset_xa_lock &c->lock irq_context: 0 sb_writers &type->i_mutex_dir_key/1 krc.lock &obj_hash[i].lock irq_context: 0 sb_writers &type->i_mutex_dir_key/1 krc.lock &base->lock irq_context: 0 sb_writers &type->i_mutex_dir_key/1 krc.lock &base->lock &obj_hash[i].lock irq_context: 0 sb_writers &type->i_mutex_dir_key/1 &cfs_rq->removed.lock irq_context: 0 &f->f_pos_lock &type->i_mutex_dir_key#5 &sem->wait_lock irq_context: 0 &f->f_pos_lock &type->i_mutex_dir_key#5 &rq->__lock irq_context: 0 &f->f_pos_lock &type->i_mutex_dir_key#5 &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 &f->f_pos_lock &type->i_mutex_dir_key#5 &rq->__lock &cfs_rq->removed.lock irq_context: 0 (wq_completion)writeback (work_completion)(&(&wb->dwork)->work) &wb->list_lock &sb->s_type->i_lock_key#22 irq_context: 0 (wq_completion)writeback (work_completion)(&(&wb->dwork)->work) &wb->list_lock &sb->s_type->i_lock_key#3 irq_context: 0 (wq_completion)writeback (work_completion)(&(&wb->dwork)->work) &wb->list_lock &type->s_umount_key#32 &sb->s_type->i_lock_key#22 irq_context: 0 (wq_completion)writeback (work_completion)(&(&wb->dwork)->work) &type->s_umount_key#32 &sb->s_type->i_lock_key#22 irq_context: 0 (wq_completion)writeback (work_completion)(&(&wb->dwork)->work) &type->s_umount_key#32 &wb->list_lock irq_context: 0 (wq_completion)writeback (work_completion)(&(&wb->dwork)->work) &type->s_umount_key#32 &wb->list_lock &sb->s_type->i_lock_key#22 irq_context: 0 (wq_completion)writeback (work_completion)(&(&wb->dwork)->work) &type->s_umount_key#32 &sbi->s_writepages_rwsem irq_context: 0 (wq_completion)writeback (work_completion)(&(&wb->dwork)->work) &wb->list_lock &type->s_umount_key#43 &sb->s_type->i_lock_key#3 irq_context: 0 (wq_completion)writeback (work_completion)(&(&wb->dwork)->work) &type->s_umount_key#43 &sb->s_type->i_lock_key#3 irq_context: 0 (wq_completion)writeback (work_completion)(&(&wb->dwork)->work) &type->s_umount_key#43 rcu_read_lock &memcg->move_lock irq_context: 0 (wq_completion)writeback (work_completion)(&(&wb->dwork)->work) &type->s_umount_key#43 rcu_read_lock &xa->xa_lock#9 irq_context: 0 (wq_completion)writeback (work_completion)(&(&wb->dwork)->work) &type->s_umount_key#43 rcu_read_lock &xa->xa_lock#9 &s->s_inode_wblist_lock irq_context: 0 (wq_completion)writeback (work_completion)(&(&wb->dwork)->work) &type->s_umount_key#43 mmu_notifier_invalidate_range_start irq_context: 0 (wq_completion)writeback (work_completion)(&(&wb->dwork)->work) &type->s_umount_key#43 pool_lock#2 irq_context: 0 (wq_completion)writeback (work_completion)(&(&wb->dwork)->work) &type->s_umount_key#43 tk_core.seq.seqcount irq_context: 0 (wq_completion)writeback (work_completion)(&(&wb->dwork)->work) &type->s_umount_key#43 &dd->lock irq_context: 0 (wq_completion)writeback (work_completion)(&(&wb->dwork)->work) &type->s_umount_key#43 &obj_hash[i].lock irq_context: 0 (wq_completion)writeback (work_completion)(&(&wb->dwork)->work) &type->s_umount_key#43 &base->lock irq_context: 0 (wq_completion)writeback (work_completion)(&(&wb->dwork)->work) &type->s_umount_key#43 &base->lock &obj_hash[i].lock irq_context: 0 (wq_completion)writeback (work_completion)(&(&wb->dwork)->work) &type->s_umount_key#43 &c->lock irq_context: 0 (wq_completion)writeback (work_completion)(&(&wb->dwork)->work) &type->s_umount_key#43 lock#4 irq_context: 0 (wq_completion)writeback (work_completion)(&(&wb->dwork)->work) &type->s_umount_key#43 lock#5 irq_context: 0 (wq_completion)writeback (work_completion)(&(&wb->dwork)->work) &type->s_umount_key#43 rcu_read_lock key#10 irq_context: 0 (wq_completion)writeback (work_completion)(&(&wb->dwork)->work) &type->s_umount_key#43 rcu_read_lock &xa->xa_lock#9 key#10 irq_context: 0 (wq_completion)writeback (work_completion)(&(&wb->dwork)->work) &type->s_umount_key#43 &wb->list_lock irq_context: 0 (wq_completion)writeback (work_completion)(&(&wb->dwork)->work) &type->s_umount_key#43 &wb->list_lock &sb->s_type->i_lock_key#3 irq_context: 0 (wq_completion)writeback (work_completion)(&(&wb->dwork)->work) &dd->lock irq_context: 0 (wq_completion)writeback (work_completion)(&(&wb->dwork)->work) rcu_read_lock &dd->lock irq_context: 0 (wq_completion)writeback (work_completion)(&(&wb->dwork)->work) rcu_read_lock &obj_hash[i].lock irq_context: 0 (wq_completion)writeback (work_completion)(&(&wb->dwork)->work) rcu_read_lock tk_core.seq.seqcount irq_context: 0 (wq_completion)writeback (work_completion)(&(&wb->dwork)->work) rcu_read_lock &virtscsi_vq->vq_lock irq_context: 0 (wq_completion)writeback (work_completion)(&(&wb->dwork)->work) rcu_read_lock &____s->seqcount irq_context: 0 (wq_completion)writeback (work_completion)(&(&wb->dwork)->work) rcu_read_lock pool_lock#2 irq_context: 0 (wq_completion)writeback (work_completion)(&(&wb->dwork)->work) rcu_read_lock rcu_read_lock pool_lock#2 irq_context: softirq rcu_read_lock &memcg->move_lock irq_context: softirq rcu_read_lock &xa->xa_lock#9 irq_context: softirq rcu_read_lock &xa->xa_lock#9 &obj_hash[i].lock irq_context: softirq rcu_read_lock &xa->xa_lock#9 &base->lock irq_context: softirq rcu_read_lock &xa->xa_lock#9 &base->lock &obj_hash[i].lock irq_context: softirq rcu_read_lock &xa->xa_lock#9 key#10 irq_context: softirq rcu_read_lock &xa->xa_lock#9 key#12 irq_context: softirq rcu_read_lock &xa->xa_lock#9 &wb->work_lock irq_context: softirq rcu_read_lock &xa->xa_lock#9 &wb->work_lock &obj_hash[i].lock irq_context: softirq rcu_read_lock &xa->xa_lock#9 &wb->work_lock &base->lock irq_context: softirq rcu_read_lock &xa->xa_lock#9 &wb->work_lock &base->lock &obj_hash[i].lock irq_context: softirq rcu_read_lock &xa->xa_lock#9 &s->s_inode_wblist_lock irq_context: 0 sb_writers &type->i_mutex_dir_key/1 tomoyo_ss rcu_read_lock rcu_read_lock rcu_node_0 irq_context: 0 sb_writers &type->i_mutex_dir_key/1 tomoyo_ss rcu_read_lock rcu_read_lock &rq->__lock irq_context: 0 sb_writers &type->i_mutex_dir_key/1 tomoyo_ss rcu_read_lock rcu_node_0 irq_context: softirq &(&wb->bw_dwork)->timer irq_context: softirq &(&wb->bw_dwork)->timer rcu_read_lock &pool->lock/1 irq_context: softirq &(&wb->bw_dwork)->timer rcu_read_lock &pool->lock/1 &obj_hash[i].lock irq_context: softirq &(&wb->bw_dwork)->timer rcu_read_lock &pool->lock/1 &p->pi_lock irq_context: softirq &(&wb->bw_dwork)->timer rcu_read_lock &pool->lock/1 &p->pi_lock &cfs_rq->removed.lock irq_context: softirq &(&wb->bw_dwork)->timer rcu_read_lock &pool->lock/1 &p->pi_lock &rq->__lock irq_context: softirq &(&wb->bw_dwork)->timer rcu_read_lock &pool->lock/1 &p->pi_lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)writeback (work_completion)(&(&wb->bw_dwork)->work) irq_context: 0 (wq_completion)writeback (work_completion)(&(&wb->bw_dwork)->work) &wb->list_lock irq_context: 0 &mm->mmap_lock &anon_vma->rwsem pgd_lock irq_context: 0 &mm->mmap_lock &anon_vma->rwsem &obj_hash[i].lock irq_context: 0 &mm->mmap_lock &anon_vma->rwsem key irq_context: 0 &mm->mmap_lock &anon_vma->rwsem pcpu_lock irq_context: 0 &mm->mmap_lock &anon_vma->rwsem percpu_counters_lock irq_context: 0 sb_writers#7 &rq->__lock irq_context: 0 sb_writers#7 &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 tomoyo_ss rcu_read_lock &rcu_state.gp_wq &p->pi_lock &cfs_rq->removed.lock irq_context: 0 &sig->cred_guard_mutex &sig->exec_update_lock &mm->mmap_lock &p->pi_lock &rq->__lock &cfs_rq->removed.lock irq_context: 0 &sig->cred_guard_mutex &sig->exec_update_lock &mm->mmap_lock rcu_node_0 irq_context: 0 &sig->cred_guard_mutex &sig->exec_update_lock pool_lock irq_context: 0 &sig->cred_guard_mutex &sig->exec_update_lock &mm->mmap_lock rcu_read_lock &rcu_state.gp_wq irq_context: 0 &sig->cred_guard_mutex &sig->exec_update_lock &mm->mmap_lock rcu_read_lock &rcu_state.gp_wq &p->pi_lock irq_context: 0 &mm->mmap_lock &vma->vm_lock->lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 dup_mmap_sem &mm->mmap_lock &mm->mmap_lock/1 &mapping->i_mmap_rwsem &p->pi_lock irq_context: 0 tomoyo_ss rcu_read_lock rcu_read_lock &cfs_rq->removed.lock irq_context: 0 tomoyo_ss rcu_read_lock rcu_read_lock &obj_hash[i].lock irq_context: 0 tomoyo_ss rcu_read_lock &rcu_state.gp_wq &p->pi_lock &rq->__lock &cfs_rq->removed.lock irq_context: 0 &p->lock remove_cache_srcu &rq->__lock irq_context: 0 &ep->mtx remove_cache_srcu irq_context: 0 &ep->mtx remove_cache_srcu quarantine_lock irq_context: 0 &ep->mtx remove_cache_srcu &c->lock irq_context: 0 &ep->mtx remove_cache_srcu &n->list_lock irq_context: 0 &ep->mtx remove_cache_srcu rcu_read_lock pool_lock#2 irq_context: 0 &ep->mtx remove_cache_srcu &obj_hash[i].lock irq_context: 0 &sig->cred_guard_mutex &sig->exec_update_lock &mm->mmap_lock &cfs_rq->removed.lock irq_context: 0 &sig->cred_guard_mutex &sig->exec_update_lock &mm->mmap_lock &obj_hash[i].lock irq_context: 0 &p->lock &mm->mmap_lock &rq->__lock irq_context: 0 &sig->cred_guard_mutex tomoyo_ss &rq->__lock irq_context: 0 &sig->cred_guard_mutex tomoyo_ss &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 remove_cache_srcu rcu_read_lock rcu_read_lock rcu_node_0 irq_context: 0 remove_cache_srcu rcu_read_lock rcu_read_lock &rq->__lock irq_context: 0 remove_cache_srcu rcu_read_lock rcu_read_lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 &mm->mmap_lock fs_reclaim &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 &type->i_mutex_dir_key#2 pgd_lock irq_context: 0 &type->i_mutex_dir_key#2 rcu_read_lock pool_lock#2 irq_context: 0 &type->i_mutex_dir_key#2 &obj_hash[i].lock irq_context: 0 &type->i_mutex_dir_key#2 key irq_context: 0 &type->i_mutex_dir_key#2 pcpu_lock irq_context: 0 &type->i_mutex_dir_key#2 percpu_counters_lock irq_context: 0 &type->i_mutex_dir_key#2 &cfs_rq->removed.lock irq_context: 0 sb_writers &type->i_mutex_dir_key/1 tomoyo_ss batched_entropy_u8.lock irq_context: 0 sb_writers &type->i_mutex_dir_key/1 tomoyo_ss kfence_freelist_lock irq_context: 0 sb_writers &type->i_mutex_dir_key/1 tomoyo_ss &meta->lock irq_context: 0 kn->active#5 &kernfs_locks->open_file_mutex[count] &lock->wait_lock irq_context: 0 kn->active#5 remove_cache_srcu rcu_read_lock rcu_node_0 irq_context: 0 kn->active#5 remove_cache_srcu rcu_read_lock &rq->__lock irq_context: 0 &mm->mmap_lock &anon_vma->rwsem &rq->__lock &cfs_rq->removed.lock irq_context: 0 kn->active#40 fs_reclaim irq_context: 0 kn->active#40 fs_reclaim mmu_notifier_invalidate_range_start irq_context: 0 kn->active#40 &kernfs_locks->open_file_mutex[count] irq_context: 0 kn->active#40 &kernfs_locks->open_file_mutex[count] fs_reclaim irq_context: 0 kn->active#40 &kernfs_locks->open_file_mutex[count] fs_reclaim mmu_notifier_invalidate_range_start irq_context: 0 kn->active#41 fs_reclaim irq_context: 0 kn->active#41 fs_reclaim mmu_notifier_invalidate_range_start irq_context: 0 kn->active#41 &kernfs_locks->open_file_mutex[count] irq_context: 0 kn->active#41 &kernfs_locks->open_file_mutex[count] fs_reclaim irq_context: 0 kn->active#41 &kernfs_locks->open_file_mutex[count] fs_reclaim mmu_notifier_invalidate_range_start irq_context: 0 kn->active#42 fs_reclaim irq_context: 0 kn->active#42 fs_reclaim mmu_notifier_invalidate_range_start irq_context: 0 kn->active#42 &kernfs_locks->open_file_mutex[count] irq_context: 0 kn->active#42 &kernfs_locks->open_file_mutex[count] fs_reclaim irq_context: 0 kn->active#42 &kernfs_locks->open_file_mutex[count] fs_reclaim mmu_notifier_invalidate_range_start irq_context: 0 &vma->vm_lock->lock rcu_read_lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 &mm->mmap_lock &anon_vma->rwsem &cfs_rq->removed.lock irq_context: 0 &kernfs_locks->open_file_mutex[count] fill_pool_map-wait-type-override &rq->__lock irq_context: 0 kn->active#43 fs_reclaim irq_context: 0 kn->active#43 fs_reclaim mmu_notifier_invalidate_range_start irq_context: 0 kn->active#43 &kernfs_locks->open_file_mutex[count] irq_context: 0 kn->active#43 &kernfs_locks->open_file_mutex[count] fs_reclaim irq_context: 0 kn->active#43 &kernfs_locks->open_file_mutex[count] fs_reclaim mmu_notifier_invalidate_range_start irq_context: 0 sb_writers#5 &type->i_mutex_dir_key#5/1 &rq->__lock &cfs_rq->removed.lock irq_context: 0 sb_writers &sb->s_type->i_mutex_key#4 &rq->__lock irq_context: 0 sb_writers#5 &type->i_mutex_dir_key#5/1 fill_pool_map-wait-type-override &rq->__lock irq_context: 0 &lo->lo_mutex irq_context: 0 &disk->open_mutex &lo->lo_mutex irq_context: 0 &u->iolock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 &group->mark_mutex &rq->__lock irq_context: 0 &group->mark_mutex &lock->wait_lock irq_context: 0 &group->mark_mutex &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 sb_writers#5 &type->i_mutex_dir_key#5/1 &sb->s_type->i_mutex_key#12/4 irq_context: 0 &disk->open_mutex nbd_index_mutex irq_context: 0 &disk->open_mutex nbd_index_mutex &nbd->config_lock irq_context: 0 &disk->open_mutex nbd_index_mutex &nbd->config_lock mmu_notifier_invalidate_range_start irq_context: 0 &disk->open_mutex nbd_index_mutex &nbd->config_lock pool_lock#2 irq_context: 0 &disk->open_mutex &nbd->config_lock irq_context: 0 &disk->open_mutex &nbd->config_lock &bdev->bd_size_lock irq_context: 0 &disk->open_mutex &nbd->config_lock &q->queue_lock irq_context: 0 &disk->open_mutex &nbd->config_lock &ACCESS_PRIVATE(sdp, lock) irq_context: 0 &disk->open_mutex &nbd->config_lock set->srcu irq_context: 0 &disk->open_mutex &nbd->config_lock &obj_hash[i].lock irq_context: 0 &disk->open_mutex &nbd->config_lock &ACCESS_PRIVATE(ssp->srcu_sup, lock) &ACCESS_PRIVATE(sdp, lock) irq_context: 0 &disk->open_mutex &nbd->config_lock &ACCESS_PRIVATE(ssp->srcu_sup, lock) rcu_read_lock &pool->lock irq_context: 0 &disk->open_mutex &nbd->config_lock &ACCESS_PRIVATE(ssp->srcu_sup, lock) rcu_read_lock &pool->lock &obj_hash[i].lock irq_context: 0 &disk->open_mutex &nbd->config_lock &ACCESS_PRIVATE(ssp->srcu_sup, lock) rcu_read_lock &pool->lock pool_lock#2 irq_context: 0 &disk->open_mutex &nbd->config_lock &ACCESS_PRIVATE(ssp->srcu_sup, lock) rcu_read_lock &pool->lock &p->pi_lock irq_context: 0 &disk->open_mutex &nbd->config_lock &ACCESS_PRIVATE(ssp->srcu_sup, lock) rcu_read_lock &pool->lock &p->pi_lock &rq->__lock irq_context: 0 &disk->open_mutex &nbd->config_lock &ACCESS_PRIVATE(ssp->srcu_sup, lock) rcu_read_lock &pool->lock &p->pi_lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 &disk->open_mutex &nbd->config_lock &rq->__lock irq_context: 0 &group->mark_mutex batched_entropy_u8.lock irq_context: 0 (wq_completion)rcu_gp (work_completion)(&(&ssp->srcu_sup->work)->work) &ssp->srcu_sup->srcu_gp_mutex &ACCESS_PRIVATE(ssp->srcu_sup, lock) irq_context: 0 &group->mark_mutex kfence_freelist_lock irq_context: 0 (wq_completion)rcu_gp (work_completion)(&(&ssp->srcu_sup->work)->work) &ACCESS_PRIVATE(ssp->srcu_sup, lock) irq_context: 0 (wq_completion)rcu_gp (work_completion)(&(&ssp->srcu_sup->work)->work) rcu_read_lock &pool->lock pool_lock#2 irq_context: 0 sb_writers#5 &type->i_mutex_dir_key#5/1 remove_cache_srcu irq_context: 0 sb_writers#5 &type->i_mutex_dir_key#5/1 remove_cache_srcu quarantine_lock irq_context: 0 sb_writers#5 &type->i_mutex_dir_key#5/1 remove_cache_srcu &c->lock irq_context: 0 sb_writers#5 &type->i_mutex_dir_key#5/1 remove_cache_srcu &n->list_lock irq_context: 0 sb_writers#5 &type->i_mutex_dir_key#5/1 remove_cache_srcu rcu_read_lock pool_lock#2 irq_context: 0 sb_writers#5 &type->i_mutex_dir_key#5/1 remove_cache_srcu &obj_hash[i].lock irq_context: 0 &disk->open_mutex &nbd->config_lock &x->wait#2 irq_context: 0 &disk->open_mutex &nbd->config_lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)rcu_gp (work_completion)(&(&ssp->srcu_sup->work)->work) &ssp->srcu_sup->srcu_gp_mutex &ssp->srcu_sup->srcu_cb_mutex &ACCESS_PRIVATE(ssp->srcu_sup, lock) irq_context: 0 &disk->open_mutex &nbd->config_lock set->srcu irq_context: 0 &disk->open_mutex &nbd->config_lock pool_lock#2 irq_context: 0 &disk->open_mutex &nbd->config_lock rcu_read_lock &rq->__lock irq_context: 0 &disk->open_mutex &nbd->config_lock rcu_read_lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 sb_writers#5 &type->i_mutex_dir_key#5 &rq->__lock &cfs_rq->removed.lock irq_context: 0 &disk->open_mutex &nbd->config_lock set->srcu &rq->__lock irq_context: 0 &disk->open_mutex nbd_index_mutex &nbd->config_lock &c->lock irq_context: 0 sb_writers#5 &type->i_mutex_dir_key#5 tomoyo_ss remove_cache_srcu irq_context: 0 sb_writers#5 &type->i_mutex_dir_key#5 tomoyo_ss remove_cache_srcu quarantine_lock irq_context: 0 sb_writers#5 &type->i_mutex_dir_key#5 tomoyo_ss remove_cache_srcu &c->lock irq_context: 0 sb_writers#5 &type->i_mutex_dir_key#5 tomoyo_ss remove_cache_srcu &n->list_lock irq_context: 0 sb_writers#5 &type->i_mutex_dir_key#5 tomoyo_ss remove_cache_srcu rcu_read_lock pool_lock#2 irq_context: 0 sb_writers#5 &type->i_mutex_dir_key#5 tomoyo_ss remove_cache_srcu &obj_hash[i].lock irq_context: 0 &type->i_mutex_dir_key#5 &rq->__lock &cfs_rq->removed.lock irq_context: 0 sb_writers#5 &type->i_mutex_dir_key#5 tomoyo_ss rcu_read_lock rcu_read_lock rcu_node_0 irq_context: 0 sb_writers#5 &type->i_mutex_dir_key#5 tomoyo_ss rcu_read_lock rcu_read_lock &rq->__lock irq_context: 0 sb_writers#5 &type->i_mutex_dir_key#5 tomoyo_ss rcu_read_lock rcu_node_0 irq_context: 0 sb_writers#5 &type->i_mutex_dir_key#5/1 rcu_read_lock &rq->__lock irq_context: 0 sb_writers#5 &type->i_mutex_dir_key#5/1 rcu_read_lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 sb_writers#5 &type->i_mutex_dir_key#5 rcu_read_lock rcu_node_0 irq_context: 0 sb_writers#5 &type->i_mutex_dir_key#5 rcu_node_0 irq_context: 0 &mousedev->mutex/1 irq_context: 0 &mousedev->mutex/1 &mousedev->mutex#2 irq_context: 0 &mousedev->mutex/1 &mousedev->mutex#2 &dev->mutex#2 irq_context: 0 &p->lock rcu_node_0 irq_context: 0 &mousedev->mutex/1 &mousedev->mutex#2 &dev->mutex#2 rcu_state.exp_mutex rcu_node_0 irq_context: 0 &mousedev->mutex/1 &mousedev->mutex#2 &dev->mutex#2 rcu_state.exp_mutex &obj_hash[i].lock irq_context: 0 &mousedev->mutex/1 &mousedev->mutex#2 &dev->mutex#2 rcu_state.exp_mutex rcu_read_lock &pool->lock irq_context: 0 &mousedev->mutex/1 &mousedev->mutex#2 &dev->mutex#2 rcu_state.exp_mutex rcu_read_lock &pool->lock &obj_hash[i].lock irq_context: 0 &mousedev->mutex/1 &mousedev->mutex#2 &dev->mutex#2 rcu_state.exp_mutex &rnp->exp_wq[0] irq_context: 0 &mousedev->mutex/1 &mousedev->mutex#2 &dev->mutex#2 rcu_state.exp_mutex &rq->__lock irq_context: 0 &mousedev->mutex/1 &mousedev->mutex#2 &dev->mutex#2 rcu_state.exp_mutex &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 &mousedev->mutex/1 &mousedev->mutex#2 &dev->mutex#2 &obj_hash[i].lock irq_context: 0 &mousedev->mutex/1 &mousedev->mutex#2 &dev->mutex#2 rcu_state.exp_mutex &rnp->exp_wq[2] irq_context: 0 &p->lock &of->mutex kn->active#5 &rfkill->lock irq_context: 0 kn->active#18 &n->list_lock irq_context: 0 kn->active#18 &n->list_lock &c->lock irq_context: 0 kn->active#16 &c->lock irq_context: softirq (&dom->period_timer) irq_context: softirq (&dom->period_timer) key#13 irq_context: softirq (&dom->period_timer) &p->sequence irq_context: softirq (&dom->period_timer) &obj_hash[i].lock irq_context: softirq (&dom->period_timer) &base->lock irq_context: softirq (&dom->period_timer) &base->lock &obj_hash[i].lock irq_context: 0 kn->active#18 &kernfs_locks->open_file_mutex[count] &c->lock irq_context: 0 kn->active#18 &kernfs_locks->open_file_mutex[count] &____s->seqcount irq_context: 0 kn->active#20 &c->lock irq_context: 0 sb_writers#5 &type->i_mutex_dir_key#5/1 tomoyo_ss mmu_notifier_invalidate_range_start &rq->__lock irq_context: 0 &p->lock remove_cache_srcu &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 kn->active#44 fs_reclaim irq_context: 0 kn->active#44 fs_reclaim mmu_notifier_invalidate_range_start irq_context: 0 kn->active#44 &kernfs_locks->open_file_mutex[count] irq_context: 0 kn->active#44 &kernfs_locks->open_file_mutex[count] fs_reclaim irq_context: 0 kn->active#44 &kernfs_locks->open_file_mutex[count] fs_reclaim mmu_notifier_invalidate_range_start irq_context: 0 kn->active#15 &kernfs_locks->open_file_mutex[count] &c->lock irq_context: 0 kn->active#17 &c->lock irq_context: 0 &sig->cred_guard_mutex mapping.invalidate_lock rcu_read_lock &base->lock irq_context: 0 &sig->cred_guard_mutex mapping.invalidate_lock rcu_read_lock &base->lock &obj_hash[i].lock irq_context: 0 kn->active#16 &____s->seqcount irq_context: 0 kn->active#17 &kernfs_locks->open_file_mutex[count] &c->lock irq_context: 0 &disk->open_mutex &new->lock irq_context: 0 &disk->open_mutex &new->lock &mtdblk->cache_mutex irq_context: 0 &sig->cred_guard_mutex sb_writers#4 &journal->j_state_lock irq_context: 0 &sig->cred_guard_mutex sb_writers#4 &journal->j_state_lock tk_core.seq.seqcount irq_context: 0 &sig->cred_guard_mutex sb_writers#4 &journal->j_state_lock &obj_hash[i].lock irq_context: 0 &sig->cred_guard_mutex sb_writers#4 &journal->j_state_lock &base->lock irq_context: 0 &sig->cred_guard_mutex sb_writers#4 &journal->j_state_lock &base->lock &obj_hash[i].lock irq_context: 0 &sig->cred_guard_mutex sb_writers#4 &wb->work_lock irq_context: 0 &sig->cred_guard_mutex sb_writers#4 &wb->work_lock &obj_hash[i].lock irq_context: 0 &sig->cred_guard_mutex sb_writers#4 &wb->work_lock &base->lock irq_context: 0 &sig->cred_guard_mutex sb_writers#4 &wb->work_lock &base->lock &obj_hash[i].lock irq_context: 0 &sig->cred_guard_mutex tomoyo_ss rcu_read_lock rcu_node_0 irq_context: 0 &sig->cred_guard_mutex tomoyo_ss rcu_read_lock &rq->__lock irq_context: 0 kn->active#15 &____s->seqcount irq_context: 0 &sig->cred_guard_mutex &iint->mutex &n->list_lock irq_context: 0 &sig->cred_guard_mutex &iint->mutex &n->list_lock &c->lock irq_context: 0 &sig->cred_guard_mutex &iint->mutex mapping.invalidate_lock &rq->__lock irq_context: 0 kn->active#15 remove_cache_srcu irq_context: 0 kn->active#15 remove_cache_srcu quarantine_lock irq_context: 0 kn->active#15 remove_cache_srcu &c->lock irq_context: 0 kn->active#15 remove_cache_srcu &n->list_lock irq_context: 0 kn->active#15 remove_cache_srcu &obj_hash[i].lock irq_context: 0 kn->active#15 &kernfs_locks->open_file_mutex[count] &n->list_lock irq_context: 0 kn->active#15 &kernfs_locks->open_file_mutex[count] &n->list_lock &c->lock irq_context: 0 kn->active#45 fs_reclaim irq_context: 0 kn->active#45 fs_reclaim mmu_notifier_invalidate_range_start irq_context: 0 kn->active#45 &kernfs_locks->open_file_mutex[count] irq_context: 0 kn->active#45 &kernfs_locks->open_file_mutex[count] fs_reclaim irq_context: 0 kn->active#45 &kernfs_locks->open_file_mutex[count] fs_reclaim mmu_notifier_invalidate_range_start irq_context: 0 &mtd->master.chrdev_lock irq_context: 0 &mtd->master.chrdev_lock &mm->mmap_lock irq_context: 0 &type->i_mutex_dir_key#2 rcu_read_lock &rq->__lock irq_context: 0 &type->i_mutex_dir_key#2 rcu_read_lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)events (work_completion)(&(&krcp->krw_arr[i].rcu_work)->work) rcu_callback &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 sb_writers#5 &type->i_mutex_dir_key#5 tomoyo_ss rcu_node_0 irq_context: 0 sb_writers#5 &type->i_mutex_dir_key#5 &pcp->lock &zone->lock irq_context: 0 sb_writers#5 &type->i_mutex_dir_key#5 &pcp->lock &zone->lock &____s->seqcount irq_context: 0 sb_writers &type->i_mutex_dir_key/1 remove_cache_srcu irq_context: 0 sb_writers &type->i_mutex_dir_key/1 remove_cache_srcu quarantine_lock irq_context: 0 sb_writers#5 &type->i_mutex_dir_key#5/1 &pcp->lock &zone->lock irq_context: 0 sb_writers#5 &type->i_mutex_dir_key#5/1 &pcp->lock &zone->lock &____s->seqcount irq_context: 0 kn->active#5 &pcp->lock &zone->lock irq_context: 0 sb_writers#5 &type->i_mutex_dir_key#5/1 &sb->s_type->i_mutex_key#12 &simple_offset_xa_lock &n->list_lock irq_context: 0 sb_writers#5 &type->i_mutex_dir_key#5/1 &sb->s_type->i_mutex_key#12 &simple_offset_xa_lock &n->list_lock &c->lock irq_context: 0 sb_writers#5 &type->i_mutex_dir_key#5 remove_cache_srcu rcu_read_lock rcu_node_0 irq_context: 0 sb_writers#5 &type->i_mutex_dir_key#5 remove_cache_srcu rcu_read_lock &rq->__lock irq_context: 0 kn->active#5 &kernfs_locks->open_file_mutex[count] &pcp->lock &zone->lock irq_context: 0 sb_writers &type->i_mutex_dir_key/1 tomoyo_ss remove_cache_srcu &rq->__lock irq_context: 0 sb_writers#5 &type->i_mutex_dir_key#5 mmu_notifier_invalidate_range_start &rq->__lock irq_context: 0 sb_writers &type->i_mutex_dir_key/1 tomoyo_ss rcu_read_lock rename_lock irq_context: 0 sb_writers &type->i_mutex_dir_key/1 &rq->__lock &cfs_rq->removed.lock irq_context: 0 sb_writers &type->i_mutex_dir_key/1 remove_cache_srcu &c->lock irq_context: 0 sb_writers &type->i_mutex_dir_key/1 remove_cache_srcu &n->list_lock irq_context: 0 sb_writers &type->i_mutex_dir_key/1 remove_cache_srcu &obj_hash[i].lock irq_context: 0 &p->lock &of->mutex kn->active#5 rcu_read_lock &rq->__lock irq_context: 0 remove_cache_srcu rcu_read_lock rcu_read_lock &rq->__lock &cfs_rq->removed.lock irq_context: 0 sb_writers#5 &type->i_mutex_dir_key#5 &simple_offset_xa_lock &n->list_lock irq_context: 0 sb_writers#5 &type->i_mutex_dir_key#5 &simple_offset_xa_lock &n->list_lock &c->lock irq_context: 0 &type->i_mutex_dir_key#5 batched_entropy_u8.lock irq_context: 0 &type->i_mutex_dir_key#5 kfence_freelist_lock irq_context: 0 &p->lock fs_reclaim &rq->__lock irq_context: 0 &type->i_mutex_dir_key#2 &pcp->lock &zone->lock &____s->seqcount irq_context: 0 sb_writers &type->i_mutex_dir_key/1 remove_cache_srcu &rq->__lock irq_context: 0 sb_writers &type->i_mutex_dir_key/1 remove_cache_srcu rcu_read_lock pool_lock#2 irq_context: 0 sb_writers &type->i_mutex_dir_key/1 rcu_read_lock &rq->__lock irq_context: 0 sb_writers &type->i_mutex_dir_key/1 rcu_read_lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 sb_writers#5 &type->i_mutex_dir_key#5 tomoyo_ss &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 kn->active#5 remove_cache_srcu rcu_node_0 irq_context: 0 kn->active#5 remove_cache_srcu rcu_read_lock &rcu_state.gp_wq irq_context: 0 kn->active#5 remove_cache_srcu rcu_read_lock &rcu_state.gp_wq &p->pi_lock irq_context: 0 kn->active#5 remove_cache_srcu rcu_read_lock &rcu_state.gp_wq &p->pi_lock &cfs_rq->removed.lock irq_context: 0 kn->active#5 remove_cache_srcu rcu_read_lock &rcu_state.gp_wq &p->pi_lock &rq->__lock irq_context: 0 kn->active#5 remove_cache_srcu rcu_read_lock &rcu_state.gp_wq &p->pi_lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 sb_writers#5 &type->i_mutex_dir_key#5/1 &fsnotify_mark_srcu fs_reclaim irq_context: 0 sb_writers#5 &type->i_mutex_dir_key#5/1 &fsnotify_mark_srcu fs_reclaim mmu_notifier_invalidate_range_start irq_context: 0 sb_writers#5 &type->i_mutex_dir_key#5/1 &fsnotify_mark_srcu pool_lock#2 irq_context: 0 sb_writers#5 &type->i_mutex_dir_key#5/1 &fsnotify_mark_srcu &group->notification_lock irq_context: 0 sb_writers#5 &type->i_mutex_dir_key#5/1 &fsnotify_mark_srcu &group->notification_waitq irq_context: 0 sb_writers#5 &type->i_mutex_dir_key#5/1 &fsnotify_mark_srcu &group->notification_waitq &p->pi_lock irq_context: 0 sb_writers#5 &type->i_mutex_dir_key#5/1 &fsnotify_mark_srcu &group->notification_waitq &p->pi_lock &rq->__lock irq_context: 0 sb_writers#5 &type->i_mutex_dir_key#5/1 &fsnotify_mark_srcu &group->notification_waitq &p->pi_lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 sb_writers#5 &type->i_mutex_dir_key#5/1 &fsnotify_mark_srcu &rq->__lock irq_context: 0 &mark->lock irq_context: 0 &group->inotify_data.idr_lock irq_context: 0 &group->inotify_data.idr_lock &obj_hash[i].lock irq_context: 0 &group->inotify_data.idr_lock pool_lock#2 irq_context: 0 destroy_lock irq_context: 0 fs/notify/mark.c:89 irq_context: 0 (reaper_work).work irq_context: 0 (wq_completion)events_unbound connector_reaper_work irq_context: 0 (wq_completion)events_unbound connector_reaper_work destroy_lock irq_context: 0 (wq_completion)events_unbound connector_reaper_work &ACCESS_PRIVATE(sdp, lock) irq_context: 0 (wq_completion)events_unbound connector_reaper_work &fsnotify_mark_srcu irq_context: 0 (wq_completion)events_unbound connector_reaper_work &obj_hash[i].lock irq_context: 0 (wq_completion)events_unbound connector_reaper_work &ACCESS_PRIVATE(ssp->srcu_sup, lock) &ACCESS_PRIVATE(sdp, lock) irq_context: 0 (wq_completion)events_unbound connector_reaper_work &ACCESS_PRIVATE(ssp->srcu_sup, lock) rcu_read_lock &pool->lock irq_context: 0 (wq_completion)events_unbound connector_reaper_work &ACCESS_PRIVATE(ssp->srcu_sup, lock) rcu_read_lock &pool->lock &obj_hash[i].lock irq_context: 0 (wq_completion)events_unbound connector_reaper_work &ACCESS_PRIVATE(ssp->srcu_sup, lock) rcu_read_lock &pool->lock pool_lock#2 irq_context: 0 (wq_completion)events_unbound connector_reaper_work &ACCESS_PRIVATE(ssp->srcu_sup, lock) rcu_read_lock &pool->lock &p->pi_lock irq_context: 0 (wq_completion)events_unbound connector_reaper_work &ACCESS_PRIVATE(ssp->srcu_sup, lock) rcu_read_lock &pool->lock &p->pi_lock &rq->__lock irq_context: 0 (wq_completion)events_unbound connector_reaper_work &ACCESS_PRIVATE(ssp->srcu_sup, lock) rcu_read_lock &pool->lock &p->pi_lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)events_unbound connector_reaper_work &x->wait#2 irq_context: 0 (wq_completion)events_unbound connector_reaper_work &rq->__lock irq_context: 0 (wq_completion)events_unbound connector_reaper_work &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)events_unbound (reaper_work).work irq_context: 0 (wq_completion)events_unbound (reaper_work).work destroy_lock irq_context: 0 (wq_completion)events_unbound (reaper_work).work &ACCESS_PRIVATE(sdp, lock) irq_context: 0 (wq_completion)events_unbound (reaper_work).work &fsnotify_mark_srcu irq_context: 0 (wq_completion)events_unbound (reaper_work).work &obj_hash[i].lock irq_context: 0 (wq_completion)events_unbound (reaper_work).work &rq->__lock irq_context: 0 (wq_completion)events_unbound (reaper_work).work &ACCESS_PRIVATE(ssp->srcu_sup, lock) &ACCESS_PRIVATE(sdp, lock) irq_context: 0 (wq_completion)events_unbound (reaper_work).work &ACCESS_PRIVATE(ssp->srcu_sup, lock) rcu_read_lock &pool->lock irq_context: 0 (wq_completion)events_unbound connector_reaper_work pool_lock#2 irq_context: 0 (wq_completion)events_unbound (reaper_work).work &ACCESS_PRIVATE(ssp->srcu_sup, lock) rcu_read_lock &pool->lock &obj_hash[i].lock irq_context: 0 (wq_completion)events_unbound (reaper_work).work &ACCESS_PRIVATE(ssp->srcu_sup, lock) rcu_read_lock &pool->lock pool_lock#2 irq_context: 0 (wq_completion)events_unbound (reaper_work).work &ACCESS_PRIVATE(ssp->srcu_sup, lock) rcu_read_lock &pool->lock &p->pi_lock irq_context: 0 (wq_completion)events_unbound (reaper_work).work &ACCESS_PRIVATE(ssp->srcu_sup, lock) rcu_read_lock &pool->lock &p->pi_lock &rq->__lock irq_context: 0 (wq_completion)events_unbound (reaper_work).work &ACCESS_PRIVATE(ssp->srcu_sup, lock) rcu_read_lock &pool->lock &p->pi_lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)events_unbound (reaper_work).work &x->wait#2 irq_context: 0 (wq_completion)events_unbound (reaper_work).work &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)events_unbound (reaper_work).work pool_lock#2 irq_context: 0 (wq_completion)events_unbound (work_completion)(&barr->work) &rq->__lock irq_context: 0 (wq_completion)events_unbound (work_completion)(&barr->work) &cfs_rq->removed.lock irq_context: 0 (wq_completion)events_unbound (work_completion)(&barr->work) &obj_hash[i].lock irq_context: 0 (wq_completion)events_unbound (work_completion)(&barr->work) pool_lock#2 irq_context: 0 &type->i_mutex_dir_key#3 &n->list_lock irq_context: 0 &type->i_mutex_dir_key#3 &n->list_lock &c->lock irq_context: 0 &sig->cred_guard_mutex mapping.invalidate_lock &c->lock irq_context: 0 sb_writers#5 &type->i_mutex_dir_key#5/1 tomoyo_ss tomoyo_policy_lock &n->list_lock irq_context: 0 sb_writers#5 &type->i_mutex_dir_key#5/1 tomoyo_ss tomoyo_policy_lock &n->list_lock &c->lock irq_context: 0 &sig->cred_guard_mutex mapping.invalidate_lock &ei->i_data_sem &c->lock irq_context: 0 &type->i_mutex_dir_key#3 batched_entropy_u8.lock irq_context: 0 &type->i_mutex_dir_key#3 kfence_freelist_lock irq_context: 0 &iint->mutex &n->list_lock irq_context: 0 &iint->mutex &n->list_lock &c->lock irq_context: 0 sb_writers#5 &sb->s_type->i_mutex_key#12 tomoyo_ss irq_context: 0 sb_writers#5 &sb->s_type->i_mutex_key#12 tomoyo_ss mmu_notifier_invalidate_range_start irq_context: 0 sb_writers#5 &sb->s_type->i_mutex_key#12 tomoyo_ss &c->lock irq_context: 0 sb_writers#5 &sb->s_type->i_mutex_key#12 tomoyo_ss &n->list_lock irq_context: 0 sb_writers#5 &sb->s_type->i_mutex_key#12 tomoyo_ss &n->list_lock &c->lock irq_context: 0 sb_writers#5 &sb->s_type->i_mutex_key#12 tomoyo_ss rcu_read_lock mount_lock.seqcount irq_context: 0 sb_writers#5 &sb->s_type->i_mutex_key#12 tomoyo_ss rcu_read_lock rcu_read_lock rename_lock.seqcount irq_context: 0 sb_writers#5 &sb->s_type->i_mutex_key#12 tomoyo_ss &obj_hash[i].lock irq_context: 0 sb_writers#5 &sb->s_type->i_mutex_key#12 tomoyo_ss rcu_read_lock &dentry->d_lock irq_context: 0 sb_writers#5 &sb->s_type->i_mutex_key#12 tomoyo_ss tomoyo_policy_lock irq_context: 0 tomoyo_ss tomoyo_policy_lock &n->list_lock irq_context: 0 tomoyo_ss tomoyo_policy_lock &n->list_lock &c->lock irq_context: 0 sb_writers#5 &sb->s_type->i_mutex_key#12 tomoyo_ss tomoyo_policy_lock mmu_notifier_invalidate_range_start irq_context: 0 userns_state_mutex irq_context: 0 &ei->xattr_sem &mapping->private_lock irq_context: 0 &sig->cred_guard_mutex &iint->mutex ima_extend_list_mutex &c->lock irq_context: 0 &sig->cred_guard_mutex &rq->__lock &cfs_rq->removed.lock irq_context: 0 &iint->mutex mapping.invalidate_lock &xa->xa_lock#9 batched_entropy_u8.lock irq_context: 0 &iint->mutex mapping.invalidate_lock &xa->xa_lock#9 kfence_freelist_lock irq_context: 0 &sig->cred_guard_mutex sb_writers#4 jbd2_handle &pcp->lock &zone->lock irq_context: 0 &sig->cred_guard_mutex sb_writers#4 jbd2_handle &pcp->lock &zone->lock &____s->seqcount irq_context: 0 &sig->cred_guard_mutex mapping.invalidate_lock &ei->i_data_sem &ei->i_es_lock key#6 irq_context: 0 &sig->cred_guard_mutex mapping.invalidate_lock &ei->i_data_sem &ei->i_es_lock key#7 irq_context: 0 rtnl_mutex rcu_read_lock &ndev->lock irq_context: 0 rtnl_mutex rlock-AF_NETLINK irq_context: 0 rtnl_mutex (inetaddr_validator_chain).rwsem irq_context: 0 rtnl_mutex (inetaddr_chain).rwsem irq_context: 0 rtnl_mutex (inetaddr_chain).rwsem fs_reclaim irq_context: 0 rtnl_mutex (inetaddr_chain).rwsem fs_reclaim mmu_notifier_invalidate_range_start irq_context: 0 rtnl_mutex (inetaddr_chain).rwsem pool_lock#2 irq_context: 0 rtnl_mutex (inetaddr_chain).rwsem pcpu_alloc_mutex irq_context: 0 rtnl_mutex (inetaddr_chain).rwsem pcpu_alloc_mutex pcpu_lock irq_context: 0 rtnl_mutex (inetaddr_chain).rwsem fib_info_lock irq_context: 0 rtnl_mutex (inetaddr_chain).rwsem &c->lock irq_context: 0 rtnl_mutex (inetaddr_chain).rwsem &dir->lock#2 irq_context: 0 rtnl_mutex (inetaddr_chain).rwsem &____s->seqcount irq_context: 0 rtnl_mutex (inetaddr_chain).rwsem nl_table_lock irq_context: 0 rtnl_mutex (inetaddr_chain).rwsem &obj_hash[i].lock irq_context: 0 rtnl_mutex (inetaddr_chain).rwsem nl_table_wait.lock irq_context: 0 rtnl_mutex (inetaddr_chain).rwsem &net->sctp.local_addr_lock irq_context: 0 rtnl_mutex (inetaddr_chain).rwsem &net->sctp.local_addr_lock &net->sctp.addr_wq_lock irq_context: 0 rtnl_mutex (inetaddr_chain).rwsem &net->sctp.local_addr_lock &net->sctp.addr_wq_lock pool_lock#2 irq_context: 0 rtnl_mutex (inetaddr_chain).rwsem &net->sctp.local_addr_lock &net->sctp.addr_wq_lock &obj_hash[i].lock irq_context: 0 rtnl_mutex (inetaddr_chain).rwsem &net->sctp.local_addr_lock &net->sctp.addr_wq_lock &base->lock irq_context: 0 rtnl_mutex (inetaddr_chain).rwsem &net->sctp.local_addr_lock &net->sctp.addr_wq_lock &base->lock &obj_hash[i].lock irq_context: 0 &sb->s_type->i_mutex_key#10 rcu_read_lock rcu_node_0 irq_context: 0 &sb->s_type->i_mutex_key#10 rcu_read_lock &rq->__lock irq_context: 0 &sb->s_type->i_mutex_key#10 rcu_read_lock &rcu_state.gp_wq irq_context: 0 rtnl_mutex _xmit_LOOPBACK irq_context: 0 rtnl_mutex netpoll_srcu irq_context: 0 rtnl_mutex &in_dev->mc_tomb_lock irq_context: 0 rtnl_mutex &im->lock irq_context: 0 rtnl_mutex fib_info_lock irq_context: 0 rtnl_mutex rcu_read_lock &c->lock irq_context: 0 rtnl_mutex rcu_read_lock &dir->lock#2 irq_context: 0 rtnl_mutex cbs_list_lock irq_context: 0 rtnl_mutex &ndev->lock irq_context: 0 rtnl_mutex &idev->mc_lock irq_context: 0 rtnl_mutex (inet6addr_validator_chain).rwsem irq_context: 0 rtnl_mutex rcu_read_lock &net->ipv6.addrconf_hash_lock irq_context: 0 rtnl_mutex rcu_read_lock &ndev->lock irq_context: 0 rtnl_mutex rcu_read_lock &net->sctp.local_addr_lock irq_context: 0 rtnl_mutex rcu_read_lock &net->sctp.local_addr_lock &net->sctp.addr_wq_lock irq_context: 0 rtnl_mutex rcu_read_lock &net->sctp.local_addr_lock &net->sctp.addr_wq_lock pool_lock#2 irq_context: 0 rtnl_mutex &ifa->lock irq_context: 0 rtnl_mutex rcu_read_lock &tb->tb6_lock irq_context: 0 rtnl_mutex rcu_read_lock &tb->tb6_lock &net->ipv6.fib6_walker_lock irq_context: 0 rtnl_mutex &tb->tb6_lock irq_context: 0 rtnl_mutex &tb->tb6_lock &____s->seqcount irq_context: 0 rtnl_mutex &tb->tb6_lock pool_lock#2 irq_context: 0 rtnl_mutex &tb->tb6_lock &c->lock irq_context: 0 rtnl_mutex &tb->tb6_lock nl_table_lock irq_context: 0 rtnl_mutex &tb->tb6_lock &obj_hash[i].lock irq_context: 0 rtnl_mutex &tb->tb6_lock nl_table_wait.lock irq_context: 0 rtnl_mutex &ndev->lock irq_context: 0 rtnl_mutex &ndev->lock &ifa->lock irq_context: 0 &sb->s_type->i_mutex_key#10 sk_lock-AF_INET irq_context: 0 &sb->s_type->i_mutex_key#10 sk_lock-AF_INET slock-AF_INET irq_context: 0 &sb->s_type->i_mutex_key#10 slock-AF_INET irq_context: 0 &sb->s_type->i_mutex_key#10 clock-AF_INET irq_context: softirq rcu_callback &dir->lock#2 irq_context: 0 &sb->s_type->i_mutex_key#10 sk_lock-AF_INET6 irq_context: 0 &sb->s_type->i_mutex_key#10 sk_lock-AF_INET6 slock-AF_INET6 irq_context: 0 &sb->s_type->i_mutex_key#10 slock-AF_INET6 irq_context: 0 &sb->s_type->i_mutex_key#10 clock-AF_INET6 irq_context: 0 sb_writers#4 &type->i_mutex_dir_key#3/1 fs_reclaim irq_context: 0 sb_writers#4 &type->i_mutex_dir_key#3/1 fs_reclaim mmu_notifier_invalidate_range_start irq_context: 0 sb_writers#4 &type->i_mutex_dir_key#3/1 &dentry->d_lock irq_context: 0 sb_writers#4 &type->i_mutex_dir_key#3/1 &ei->i_es_lock irq_context: 0 sb_writers#4 &type->i_mutex_dir_key#3/1 &ei->i_data_sem irq_context: 0 sb_writers#4 &type->i_mutex_dir_key#3/1 &ei->i_data_sem mmu_notifier_invalidate_range_start irq_context: 0 sb_writers#4 &type->i_mutex_dir_key#3/1 &ei->i_data_sem pool_lock#2 irq_context: 0 sb_writers#4 &type->i_mutex_dir_key#3/1 &ei->i_data_sem &ei->i_es_lock irq_context: 0 sb_writers#4 &type->i_mutex_dir_key#3/1 &ei->i_data_sem &ei->i_es_lock pool_lock#2 irq_context: 0 sb_writers#4 &type->i_mutex_dir_key#3/1 &ei->i_data_sem &ei->i_es_lock &sbi->s_es_lock irq_context: 0 sb_writers#4 &type->i_mutex_dir_key#3/1 &ei->i_data_sem &obj_hash[i].lock irq_context: 0 sb_writers#4 &type->i_mutex_dir_key#3/1 &ei->i_data_sem &ei->i_es_lock &obj_hash[i].lock irq_context: 0 sb_writers#4 &type->i_mutex_dir_key#3/1 mmu_notifier_invalidate_range_start irq_context: 0 sb_writers#4 &type->i_mutex_dir_key#3/1 &____s->seqcount irq_context: 0 sb_writers#4 &type->i_mutex_dir_key#3/1 pool_lock#2 irq_context: 0 sb_writers#4 &type->i_mutex_dir_key#3/1 &xa->xa_lock#9 irq_context: 0 sb_writers#4 &type->i_mutex_dir_key#3/1 lock#4 irq_context: 0 sb_writers#4 &type->i_mutex_dir_key#3/1 &mapping->private_lock irq_context: 0 sb_writers#4 &type->i_mutex_dir_key#3/1 tk_core.seq.seqcount irq_context: 0 sb_writers#4 &type->i_mutex_dir_key#3/1 &dd->lock irq_context: 0 sb_writers#4 &type->i_mutex_dir_key#3/1 &obj_hash[i].lock irq_context: 0 sb_writers#4 &type->i_mutex_dir_key#3/1 rcu_read_lock &pool->lock irq_context: 0 sb_writers#4 &type->i_mutex_dir_key#3/1 rcu_read_lock &pool->lock &obj_hash[i].lock irq_context: 0 sb_writers#4 &type->i_mutex_dir_key#3/1 rcu_read_lock &pool->lock pool_lock#2 irq_context: 0 sb_writers#4 &type->i_mutex_dir_key#3/1 rcu_read_lock &pool->lock &p->pi_lock irq_context: 0 sb_writers#4 &type->i_mutex_dir_key#3/1 rcu_read_lock &pool->lock &p->pi_lock &rq->__lock irq_context: 0 sb_writers#4 &type->i_mutex_dir_key#3/1 rcu_read_lock &pool->lock &p->pi_lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 sb_writers#4 &type->i_mutex_dir_key#3/1 bit_wait_table + i irq_context: 0 sb_writers#4 &type->i_mutex_dir_key#3/1 &rq->__lock irq_context: 0 sb_writers#4 &type->i_mutex_dir_key#3/1 &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 sb_writers#4 &type->i_mutex_dir_key#3/1 inode_hash_lock irq_context: 0 sb_writers#4 &type->i_mutex_dir_key#3/1 &c->lock irq_context: 0 sb_writers#4 &type->i_mutex_dir_key#3/1 &n->list_lock irq_context: 0 sb_writers#4 &type->i_mutex_dir_key#3/1 &n->list_lock &c->lock irq_context: 0 sb_writers#4 &type->i_mutex_dir_key#3/1 inode_hash_lock &sb->s_type->i_lock_key#22 irq_context: 0 sb_writers#4 &type->i_mutex_dir_key#3/1 inode_hash_lock &s->s_inode_list_lock irq_context: 0 sb_writers#4 &type->i_mutex_dir_key#3/1 &journal->j_state_lock irq_context: 0 sb_writers#4 &type->i_mutex_dir_key#3/1 &sb->s_type->i_lock_key#22 irq_context: 0 sb_writers#4 &type->i_mutex_dir_key#3/1 &sb->s_type->i_lock_key#22 &dentry->d_lock irq_context: 0 sb_writers#4 &type->i_mutex_dir_key#3/1 &dentry->d_lock &lru->node[i].lock irq_context: 0 sb_writers#5 tomoyo_ss irq_context: 0 sb_writers#5 tomoyo_ss mmu_notifier_invalidate_range_start irq_context: 0 sb_writers#5 tomoyo_ss pool_lock#2 irq_context: 0 sb_writers#5 tomoyo_ss rcu_read_lock mount_lock.seqcount irq_context: 0 sb_writers#5 tomoyo_ss rcu_read_lock rcu_read_lock rename_lock.seqcount irq_context: 0 sb_writers#5 tomoyo_ss &obj_hash[i].lock irq_context: 0 sb_writers#5 tomoyo_ss rcu_read_lock &dentry->d_lock irq_context: 0 sb_writers#5 tomoyo_ss &c->lock irq_context: 0 sb_writers#5 tomoyo_ss tomoyo_policy_lock irq_context: 0 sb_writers#5 &xattrs->lock irq_context: 0 sb_writers#5 &sb->s_type->i_mutex_key#12 &mapping->i_mmap_rwsem irq_context: 0 sb_writers#5 &sb->s_type->i_mutex_key#12 &sb->s_type->i_lock_key &xa->xa_lock#9 irq_context: 0 sb_writers#5 &sb->s_type->i_mutex_key#12 lock#5 irq_context: 0 sb_writers#5 &sb->s_type->i_mutex_key#12 &lruvec->lru_lock irq_context: 0 sb_writers#5 &sb->s_type->i_mutex_key#12 rcu_read_lock pool_lock#2 irq_context: 0 sb_writers#5 &sb->s_type->i_mutex_key#12 &obj_hash[i].lock irq_context: 0 &f->f_pos_lock sb_writers#5 irq_context: 0 &f->f_pos_lock sb_writers#5 &sb->s_type->i_mutex_key#12 irq_context: 0 &f->f_pos_lock sb_writers#5 &sb->s_type->i_mutex_key#12 tk_core.seq.seqcount irq_context: 0 &f->f_pos_lock sb_writers#5 &sb->s_type->i_mutex_key#12 &sb->s_type->i_lock_key irq_context: 0 &f->f_pos_lock sb_writers#5 &sb->s_type->i_mutex_key#12 fs_reclaim irq_context: 0 &f->f_pos_lock sb_writers#5 &sb->s_type->i_mutex_key#12 fs_reclaim mmu_notifier_invalidate_range_start irq_context: 0 &f->f_pos_lock sb_writers#5 &sb->s_type->i_mutex_key#12 &____s->seqcount irq_context: 0 &f->f_pos_lock sb_writers#5 &sb->s_type->i_mutex_key#12 &xa->xa_lock#9 irq_context: 0 &f->f_pos_lock sb_writers#5 &sb->s_type->i_mutex_key#12 lock#4 irq_context: 0 &f->f_pos_lock sb_writers#5 &sb->s_type->i_mutex_key#12 &info->lock irq_context: 0 &sb->s_type->i_lock_key#4 irq_context: 0 &sb->s_type->i_lock_key#4 &dentry->d_lock irq_context: 0 sk_lock-AF_INET irq_context: 0 sk_lock-AF_INET slock-AF_INET irq_context: 0 slock-AF_INET irq_context: 0 sk_lock-AF_INET6 irq_context: 0 sk_lock-AF_INET6 slock-AF_INET6 irq_context: 0 slock-AF_INET6 irq_context: 0 &type->i_mutex_dir_key#3 &ei->i_es_lock key#5 irq_context: 0 &type->i_mutex_dir_key#3 &ei->i_data_sem &ei->i_es_lock key#6 irq_context: 0 &type->i_mutex_dir_key#3 &ei->i_data_sem &ei->i_es_lock key#7 irq_context: 0 sk_lock-AF_INET &table->hash[i].lock irq_context: 0 sk_lock-AF_INET &table->hash[i].lock clock-AF_INET irq_context: 0 sk_lock-AF_INET &table->hash[i].lock &table->hash2[i].lock irq_context: 0 sk_lock-AF_INET6 &table->hash[i].lock irq_context: 0 sk_lock-AF_INET6 &table->hash[i].lock clock-AF_INET6 irq_context: 0 sk_lock-AF_INET6 &table->hash[i].lock &table->hash2[i].lock irq_context: 0 &sb->s_type->i_mutex_key#9 &n->list_lock irq_context: 0 &sb->s_type->i_mutex_key#9 &n->list_lock &c->lock irq_context: 0 sk_lock-AF_NETLINK &mm->mmap_lock irq_context: 0 sk_lock-AF_NETLINK fs_reclaim irq_context: 0 sk_lock-AF_NETLINK fs_reclaim mmu_notifier_invalidate_range_start irq_context: 0 sk_lock-AF_NETLINK &c->lock irq_context: 0 sk_lock-AF_NETLINK pool_lock#2 irq_context: 0 sk_lock-AF_NETLINK free_vmap_area_lock irq_context: 0 sk_lock-AF_NETLINK free_vmap_area_lock &obj_hash[i].lock irq_context: 0 sk_lock-AF_NETLINK free_vmap_area_lock pool_lock#2 irq_context: 0 sk_lock-AF_NETLINK vmap_area_lock irq_context: 0 sk_lock-AF_NETLINK &____s->seqcount irq_context: 0 sk_lock-AF_NETLINK pcpu_alloc_mutex irq_context: 0 sk_lock-AF_NETLINK pcpu_alloc_mutex pcpu_lock irq_context: 0 sk_lock-AF_NETLINK &obj_hash[i].lock irq_context: 0 sk_lock-AF_NETLINK pack_mutex irq_context: 0 sk_lock-AF_NETLINK batched_entropy_u32.lock irq_context: 0 sk_lock-AF_NETLINK text_mutex irq_context: 0 sk_lock-AF_NETLINK text_mutex ptlock_ptr(page)#2 irq_context: 0 sk_lock-AF_NETLINK &fp->aux->used_maps_mutex irq_context: 0 nl_table_lock &c->lock irq_context: 0 rcu_read_lock &sb->s_type->i_lock_key#22 irq_context: 0 kn->active#46 fs_reclaim irq_context: 0 kn->active#46 fs_reclaim mmu_notifier_invalidate_range_start irq_context: 0 kn->active#46 &c->lock irq_context: 0 kn->active#46 &kernfs_locks->open_file_mutex[count] irq_context: 0 kn->active#46 &kernfs_locks->open_file_mutex[count] fs_reclaim irq_context: 0 kn->active#46 &kernfs_locks->open_file_mutex[count] fs_reclaim mmu_notifier_invalidate_range_start irq_context: 0 &p->lock cpufreq_driver_lock irq_context: 0 sk_lock-AF_NETLINK rcu_read_lock rcu_read_lock &pool->lock irq_context: 0 sk_lock-AF_NETLINK rcu_read_lock rcu_read_lock &pool->lock &obj_hash[i].lock irq_context: 0 sk_lock-AF_NETLINK rcu_read_lock rcu_read_lock &pool->lock &p->pi_lock irq_context: 0 sk_lock-AF_NETLINK rcu_read_lock rcu_read_lock &pool->lock &p->pi_lock &rq->__lock irq_context: 0 sk_lock-AF_NETLINK rcu_read_lock rcu_read_lock &pool->lock &p->pi_lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 rtnl_mutex &n->list_lock irq_context: 0 rtnl_mutex &n->list_lock &c->lock irq_context: 0 rtnl_mutex rcu_read_lock &ndev->lock &ifa->lock irq_context: 0 &type->i_mutex_dir_key#4 &dentry->d_lock &wq#2 irq_context: 0 vlan_ioctl_mutex &mm->mmap_lock irq_context: 0 vlan_ioctl_mutex rtnl_mutex irq_context: 0 cb_lock &c->lock irq_context: 0 cb_lock &n->list_lock irq_context: 0 cb_lock &n->list_lock &c->lock irq_context: 0 cb_lock rtnl_mutex irq_context: 0 cb_lock &obj_hash[i].lock irq_context: 0 cb_lock genl_mutex &n->list_lock irq_context: 0 cb_lock genl_mutex &n->list_lock &c->lock irq_context: 0 dev_addr_sem irq_context: 0 cb_lock rtnl_mutex &rdev->wiphy.mtx irq_context: 0 cb_lock &rdev->wiphy.mtx irq_context: 0 cb_lock &rdev->wiphy.mtx fs_reclaim irq_context: 0 cb_lock &rdev->wiphy.mtx fs_reclaim mmu_notifier_invalidate_range_start irq_context: 0 cb_lock &rdev->wiphy.mtx pool_lock#2 irq_context: 0 cb_lock &rdev->wiphy.mtx &c->lock irq_context: 0 cb_lock &rdev->wiphy.mtx &____s->seqcount irq_context: 0 cb_lock &rdev->wiphy.mtx &obj_hash[i].lock irq_context: 0 cb_lock &rdev->wiphy.mtx rlock-AF_NETLINK irq_context: 0 cb_lock nlk_cb_mutex-GENERIC irq_context: 0 cb_lock nlk_cb_mutex-GENERIC fs_reclaim irq_context: 0 cb_lock nlk_cb_mutex-GENERIC fs_reclaim mmu_notifier_invalidate_range_start irq_context: 0 cb_lock nlk_cb_mutex-GENERIC pool_lock#2 irq_context: 0 cb_lock nlk_cb_mutex-GENERIC &____s->seqcount irq_context: 0 cb_lock nlk_cb_mutex-GENERIC rtnl_mutex irq_context: 0 cb_lock nlk_cb_mutex-GENERIC rtnl_mutex &obj_hash[i].lock irq_context: 0 cb_lock nlk_cb_mutex-GENERIC rtnl_mutex pool_lock#2 irq_context: 0 cb_lock nlk_cb_mutex-GENERIC rtnl_mutex &rdev->wiphy.mtx irq_context: 0 cb_lock nlk_cb_mutex-GENERIC &rdev->wiphy.mtx irq_context: 0 cb_lock nlk_cb_mutex-GENERIC &rdev->wiphy.mtx &wdev->mtx irq_context: 0 cb_lock nlk_cb_mutex-GENERIC &rdev->wiphy.mtx &wdev->mtx &rdev->bss_lock irq_context: 0 cb_lock nlk_cb_mutex-GENERIC rlock-AF_NETLINK irq_context: 0 cb_lock nlk_cb_mutex-GENERIC &obj_hash[i].lock irq_context: 0 cb_lock &____s->seqcount irq_context: 0 sb_writers#3 tomoyo_ss &n->list_lock irq_context: 0 sb_writers#3 tomoyo_ss &n->list_lock &c->lock irq_context: 0 &u->iolock &mm->mmap_lock ptlock_ptr(page)#2 irq_context: 0 &u->iolock &mm->mmap_lock fs_reclaim irq_context: 0 &u->iolock &mm->mmap_lock fs_reclaim mmu_notifier_invalidate_range_start irq_context: 0 &u->iolock &mm->mmap_lock &____s->seqcount irq_context: 0 &u->iolock &mm->mmap_lock mmu_notifier_invalidate_range_start irq_context: 0 &u->iolock &mm->mmap_lock ptlock_ptr(page)#2 lock#4 irq_context: 0 sk_lock-AF_NETLINK rcu_read_lock rcu_node_0 irq_context: 0 sk_lock-AF_NETLINK rcu_read_lock &rq->__lock irq_context: 0 &sb->s_type->i_mutex_key#9 rcu_read_lock pool_lock#2 irq_context: 0 &sb->s_type->i_mutex_key#9 rcu_read_lock &p->alloc_lock irq_context: 0 &sig->cred_guard_mutex mapping.invalidate_lock &ei->i_es_lock key#5 irq_context: 0 &iint->mutex mapping.invalidate_lock &xa->xa_lock#9 &n->list_lock irq_context: 0 &iint->mutex mapping.invalidate_lock &xa->xa_lock#9 &n->list_lock &c->lock irq_context: 0 &pipe->rd_wait &p->pi_lock &cfs_rq->removed.lock irq_context: 0 sb_writers#5 fs_reclaim irq_context: 0 sb_writers#5 fs_reclaim mmu_notifier_invalidate_range_start irq_context: 0 sb_writers#5 sb_writers#5 mount_lock irq_context: 0 sb_writers#5 sb_writers#5 tk_core.seq.seqcount irq_context: 0 sb_writers#5 sb_writers#5 &sb->s_type->i_lock_key irq_context: 0 sb_writers#5 sb_writers#5 &wb->list_lock irq_context: 0 sb_writers#5 sb_writers#5 &wb->list_lock &sb->s_type->i_lock_key irq_context: 0 sb_writers#5 &c->lock irq_context: softirq (&net->sctp.addr_wq_timer) irq_context: softirq (&net->sctp.addr_wq_timer) &net->sctp.addr_wq_lock irq_context: softirq (&net->sctp.addr_wq_timer) &net->sctp.addr_wq_lock &obj_hash[i].lock irq_context: softirq (&net->sctp.addr_wq_timer) &net->sctp.addr_wq_lock pool_lock#2 irq_context: 0 &sig->cred_guard_mutex tomoyo_ss tomoyo_policy_lock &n->list_lock irq_context: 0 &sig->cred_guard_mutex tomoyo_ss tomoyo_policy_lock &n->list_lock &c->lock irq_context: softirq drivers/char/random.c:251 rcu_read_lock &pool->lock/1 &p->pi_lock &cfs_rq->removed.lock irq_context: 0 &mm->mmap_lock batched_entropy_u8.lock crngs.lock irq_context: 0 &mm->mmap_lock batched_entropy_u8.lock crngs.lock base_crng.lock irq_context: 0 sb_writers#5 &sb->s_type->i_lock_key &xa->xa_lock#9 irq_context: 0 sb_writers#5 lock#4 irq_context: 0 sb_writers#5 lock#4 &lruvec->lru_lock irq_context: 0 sb_writers#5 lock#5 irq_context: 0 sb_writers#5 &lruvec->lru_lock irq_context: 0 sb_writers#5 rcu_read_lock pool_lock#2 irq_context: 0 &sig->cred_guard_mutex &sig->exec_update_lock batched_entropy_u64.lock crngs.lock base_crng.lock irq_context: 0 &sig->cred_guard_mutex &sb->s_type->i_mutex_key#8 remove_cache_srcu irq_context: 0 &sig->cred_guard_mutex &sb->s_type->i_mutex_key#8 remove_cache_srcu quarantine_lock irq_context: 0 &sig->cred_guard_mutex &sb->s_type->i_mutex_key#8 remove_cache_srcu &c->lock irq_context: 0 &sig->cred_guard_mutex &sb->s_type->i_mutex_key#8 remove_cache_srcu &n->list_lock irq_context: 0 &sig->cred_guard_mutex &sb->s_type->i_mutex_key#8 remove_cache_srcu rcu_read_lock pool_lock#2 irq_context: 0 &sig->cred_guard_mutex &sb->s_type->i_mutex_key#8 remove_cache_srcu &obj_hash[i].lock irq_context: 0 cgroup_threadgroup_rwsem &obj_hash[i].lock pool_lock irq_context: 0 &sig->cred_guard_mutex remove_cache_srcu irq_context: 0 &sig->cred_guard_mutex remove_cache_srcu quarantine_lock irq_context: 0 &sig->cred_guard_mutex remove_cache_srcu &c->lock irq_context: 0 &sig->cred_guard_mutex remove_cache_srcu &n->list_lock irq_context: 0 &sig->cred_guard_mutex remove_cache_srcu rcu_read_lock pool_lock#2 irq_context: 0 &sig->cred_guard_mutex remove_cache_srcu &obj_hash[i].lock irq_context: 0 &mm->mmap_lock rcu_read_lock rcu_read_lock &rq->__lock &cfs_rq->removed.lock irq_context: 0 &pipe->rd_wait &p->pi_lock &rq->__lock &cfs_rq->removed.lock irq_context: 0 &f->f_pos_lock &type->i_mutex_dir_key#3 remove_cache_srcu irq_context: 0 &f->f_pos_lock &type->i_mutex_dir_key#3 remove_cache_srcu quarantine_lock irq_context: 0 &f->f_pos_lock &type->i_mutex_dir_key#3 remove_cache_srcu &c->lock irq_context: 0 &f->f_pos_lock &type->i_mutex_dir_key#3 remove_cache_srcu &n->list_lock irq_context: 0 &f->f_pos_lock &type->i_mutex_dir_key#3 remove_cache_srcu rcu_read_lock pool_lock#2 irq_context: 0 &f->f_pos_lock &type->i_mutex_dir_key#3 remove_cache_srcu &obj_hash[i].lock irq_context: 0 sb_writers#5 &obj_hash[i].lock pool_lock irq_context: softirq mm/vmstat.c:2018 rcu_read_lock &pool->lock &p->pi_lock &rq->__lock &cfs_rq->removed.lock irq_context: 0 &sig->cred_guard_mutex fs_reclaim &rq->__lock irq_context: 0 rtnl_mutex &dev_addr_list_lock_key irq_context: 0 rtnl_mutex &dev_addr_list_lock_key pool_lock#2 irq_context: 0 rtnl_mutex uevent_sock_mutex remove_cache_srcu irq_context: 0 rtnl_mutex uevent_sock_mutex remove_cache_srcu quarantine_lock irq_context: 0 rtnl_mutex uevent_sock_mutex remove_cache_srcu &c->lock irq_context: 0 rtnl_mutex uevent_sock_mutex remove_cache_srcu &n->list_lock irq_context: 0 rtnl_mutex uevent_sock_mutex remove_cache_srcu rcu_read_lock pool_lock#2 irq_context: 0 rtnl_mutex uevent_sock_mutex remove_cache_srcu &obj_hash[i].lock irq_context: 0 rtnl_mutex uevent_sock_mutex rlock-AF_NETLINK irq_context: 0 rtnl_mutex uevent_sock_mutex rcu_read_lock &ei->socket.wq.wait irq_context: 0 rtnl_mutex uevent_sock_mutex rcu_read_lock &ei->socket.wq.wait &ep->lock irq_context: 0 rtnl_mutex uevent_sock_mutex rcu_read_lock &ei->socket.wq.wait &ep->lock &ep->wq irq_context: 0 rtnl_mutex uevent_sock_mutex rcu_read_lock &ei->socket.wq.wait &ep->lock &ep->wq &p->pi_lock irq_context: 0 rtnl_mutex uevent_sock_mutex rcu_read_lock &ei->socket.wq.wait &ep->lock &ep->wq &p->pi_lock &rq->__lock irq_context: 0 rtnl_mutex uevent_sock_mutex rcu_read_lock &ei->socket.wq.wait &ep->lock &ep->wq &p->pi_lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 rtnl_mutex uevent_sock_mutex rcu_read_lock rcu_node_0 irq_context: 0 rtnl_mutex uevent_sock_mutex rcu_read_lock &rq->__lock irq_context: 0 &dentry->d_lock &dentry->d_lock &lru->node[i].lock irq_context: 0 rtnl_mutex &pnettable->lock irq_context: 0 rtnl_mutex smc_ib_devices.mutex irq_context: 0 rtnl_mutex napi_hash_lock irq_context: 0 rtnl_mutex lapb_list_lock irq_context: 0 rtnl_mutex &root->kernfs_rwsem &sem->wait_lock irq_context: 0 rtnl_mutex &root->kernfs_rwsem &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 &type->i_mutex_dir_key#4 &sem->wait_lock irq_context: 0 &type->i_mutex_dir_key#4 &p->pi_lock irq_context: 0 &type->i_mutex_dir_key#4 &root->kernfs_rwsem &sem->wait_lock irq_context: 0 rtnl_mutex x25_neigh_list_lock irq_context: 0 rtnl_mutex console_lock console_srcu console_owner_lock irq_context: 0 rtnl_mutex console_lock console_srcu console_owner irq_context: 0 rtnl_mutex console_lock console_srcu console_owner &port_lock_key irq_context: 0 rtnl_mutex console_lock console_srcu console_owner console_owner_lock irq_context: 0 &u->lock &ei->socket.wq.wait irq_context: 0 rtnl_mutex _xmit_ETHER irq_context: 0 rtnl_mutex &tb->tb6_lock rlock-AF_NETLINK irq_context: 0 rtnl_mutex _xmit_SLIP irq_context: softirq (&eql->timer) irq_context: softirq (&eql->timer) &eql->queue.lock irq_context: softirq (&eql->timer) &obj_hash[i].lock irq_context: softirq (&eql->timer) &base->lock irq_context: softirq (&eql->timer) &base->lock &obj_hash[i].lock irq_context: 0 rtnl_mutex &root->kernfs_rwsem rcu_read_lock &rq->__lock irq_context: 0 rtnl_mutex &root->kernfs_rwsem rcu_read_lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 &type->i_mutex_dir_key#4 &p->pi_lock &rq->__lock irq_context: 0 &type->i_mutex_dir_key#4 &p->pi_lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 rtnl_mutex &sem->wait_lock irq_context: 0 &root->kernfs_rwsem &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 &root->kernfs_rwsem &p->pi_lock irq_context: 0 &root->kernfs_rwsem rcu_read_lock &rq->__lock irq_context: 0 &root->kernfs_rwsem rcu_read_lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 rtnl_mutex uevent_sock_mutex &n->list_lock &c->lock irq_context: 0 rtnl_mutex &vi->refill_lock irq_context: softirq _xmit_ETHER#2 irq_context: 0 &sig->cred_guard_mutex &iint->mutex mapping.invalidate_lock &xa->xa_lock#9 &c->lock irq_context: 0 &sig->cred_guard_mutex &iint->mutex mapping.invalidate_lock &xa->xa_lock#9 &n->list_lock irq_context: 0 &sig->cred_guard_mutex &iint->mutex mapping.invalidate_lock &xa->xa_lock#9 &n->list_lock &c->lock irq_context: 0 rtnl_mutex noop_qdisc.q.lock irq_context: 0 &iint->mutex &ei->xattr_sem &mapping->private_lock irq_context: 0 &iint->mutex rcu_read_lock &rq->__lock irq_context: 0 &iint->mutex rcu_read_lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 rtnl_mutex &rfkill->lock irq_context: 0 rtnl_mutex &local->chanctx_mtx irq_context: 0 rtnl_mutex &rdev->wiphy.mtx lweventlist_lock irq_context: 0 rtnl_mutex &rdev->wiphy.mtx lweventlist_lock pool_lock#2 irq_context: 0 rtnl_mutex &rdev->wiphy.mtx lweventlist_lock &dir->lock#2 irq_context: 0 rtnl_mutex &rdev->wiphy.mtx rcu_read_lock &pool->lock irq_context: 0 rtnl_mutex &rdev->wiphy.mtx rcu_read_lock &pool->lock &obj_hash[i].lock irq_context: 0 rtnl_mutex &rdev->wiphy.mtx rcu_read_lock &pool->lock &p->pi_lock irq_context: 0 rtnl_mutex &rdev->wiphy.mtx rcu_read_lock &pool->lock &p->pi_lock &rq->__lock irq_context: 0 rtnl_mutex &rdev->wiphy.mtx rcu_read_lock &pool->lock &p->pi_lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 rtnl_mutex &rdev->wiphy.mtx &data->mutex irq_context: 0 rtnl_mutex _xmit_ETHER &local->filter_lock irq_context: 0 rtnl_mutex _xmit_ETHER &local->filter_lock pool_lock#2 irq_context: 0 rtnl_mutex _xmit_ETHER rcu_read_lock &pool->lock/1 irq_context: 0 rtnl_mutex _xmit_ETHER rcu_read_lock &pool->lock/1 &obj_hash[i].lock irq_context: 0 rtnl_mutex _xmit_ETHER rcu_read_lock &pool->lock/1 &p->pi_lock irq_context: 0 rtnl_mutex _xmit_ETHER rcu_read_lock &pool->lock/1 &p->pi_lock &rq->__lock irq_context: 0 rtnl_mutex _xmit_ETHER rcu_read_lock &pool->lock/1 &p->pi_lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)phy0 irq_context: 0 (wq_completion)phy0 (work_completion)(&local->reconfig_filter) irq_context: 0 (wq_completion)phy0 (work_completion)(&local->reconfig_filter) &local->filter_lock irq_context: 0 rtnl_mutex &xa->xa_lock#3 pool_lock#2 irq_context: 0 rtnl_mutex &xa->xa_lock#3 &c->lock irq_context: 0 (wq_completion)events (linkwatch_work).work rtnl_mutex &dev->tx_global_lock irq_context: 0 (wq_completion)events (linkwatch_work).work rtnl_mutex &dev->tx_global_lock _xmit_ETHER#2 irq_context: 0 (wq_completion)events (linkwatch_work).work rtnl_mutex &dev->tx_global_lock &obj_hash[i].lock irq_context: softirq rcu_read_lock noop_qdisc.q.lock irq_context: 0 (wq_completion)events (linkwatch_work).work rtnl_mutex rcu_state.exp_mutex rcu_node_0 irq_context: 0 (wq_completion)events (linkwatch_work).work rtnl_mutex rcu_state.exp_mutex &obj_hash[i].lock irq_context: 0 (wq_completion)events (linkwatch_work).work rtnl_mutex rcu_state.exp_mutex rcu_read_lock &pool->lock irq_context: 0 (wq_completion)events (linkwatch_work).work rtnl_mutex rcu_state.exp_mutex rcu_read_lock &pool->lock &obj_hash[i].lock irq_context: 0 (wq_completion)events (linkwatch_work).work rtnl_mutex rcu_state.exp_mutex &rnp->exp_wq[3] irq_context: 0 (wq_completion)events (linkwatch_work).work rtnl_mutex rcu_state.exp_mutex &pool->lock irq_context: 0 (wq_completion)events (linkwatch_work).work rtnl_mutex rcu_state.exp_mutex &pool->lock &p->pi_lock irq_context: 0 (wq_completion)events (linkwatch_work).work rtnl_mutex rcu_state.exp_mutex &pool->lock &p->pi_lock &rq->__lock irq_context: 0 (wq_completion)events (linkwatch_work).work rtnl_mutex rcu_state.exp_mutex &pool->lock &p->pi_lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)events (linkwatch_work).work rtnl_mutex rcu_state.exp_mutex &rq->__lock irq_context: 0 (wq_completion)events (linkwatch_work).work rtnl_mutex rcu_state.exp_mutex &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)events (linkwatch_work).work rtnl_mutex &obj_hash[i].lock irq_context: 0 (wq_completion)events (linkwatch_work).work rtnl_mutex &sch->q.lock irq_context: 0 (wq_completion)events (linkwatch_work).work rtnl_mutex net_rwsem irq_context: 0 (wq_completion)events (linkwatch_work).work rtnl_mutex net_rwsem &list->lock#2 irq_context: 0 (wq_completion)events (linkwatch_work).work rtnl_mutex &tbl->lock irq_context: 0 (wq_completion)events (linkwatch_work).work rtnl_mutex class irq_context: 0 (wq_completion)events (linkwatch_work).work rtnl_mutex (&tbl->proxy_timer) irq_context: 0 (wq_completion)events (linkwatch_work).work rtnl_mutex &base->lock irq_context: 0 (wq_completion)events (linkwatch_work).work rtnl_mutex cbs_list_lock irq_context: 0 (wq_completion)events (linkwatch_work).work rtnl_mutex &tn->lock irq_context: 0 (wq_completion)events (linkwatch_work).work rtnl_mutex rcu_read_lock &tb->tb6_lock irq_context: 0 (wq_completion)events (linkwatch_work).work rtnl_mutex rcu_read_lock &tb->tb6_lock &net->ipv6.fib6_walker_lock irq_context: 0 (wq_completion)events (linkwatch_work).work rtnl_mutex fs_reclaim irq_context: 0 (wq_completion)events (linkwatch_work).work rtnl_mutex fs_reclaim mmu_notifier_invalidate_range_start irq_context: 0 (wq_completion)events (linkwatch_work).work rtnl_mutex pool_lock#2 irq_context: 0 (wq_completion)events (linkwatch_work).work rtnl_mutex rcu_read_lock &ndev->lock irq_context: 0 (wq_completion)events (linkwatch_work).work rtnl_mutex nl_table_lock irq_context: 0 (wq_completion)events (linkwatch_work).work rtnl_mutex rlock-AF_NETLINK irq_context: 0 (wq_completion)events (linkwatch_work).work rtnl_mutex nl_table_wait.lock irq_context: 0 rtnl_mutex &rdev->wiphy.mtx &base->lock irq_context: 0 rtnl_mutex &rdev->wiphy.mtx &base->lock &obj_hash[i].lock irq_context: 0 (wq_completion)phy1 irq_context: 0 (wq_completion)phy1 (work_completion)(&local->reconfig_filter) irq_context: 0 (wq_completion)phy1 (work_completion)(&local->reconfig_filter) &local->filter_lock irq_context: 0 rtnl_mutex &rdev->wiphy.mtx &rq->__lock irq_context: 0 rtnl_mutex _xmit_VOID irq_context: 0 &u->iolock rcu_read_lock pool_lock#2 irq_context: 0 &u->iolock &u->lock irq_context: 0 &u->iolock &mm->mmap_lock ptlock_ptr(page)#2 lock#4 &lruvec->lru_lock irq_context: 0 &f->f_pos_lock &type->i_mutex_dir_key#3 &ei->i_es_lock key#2 irq_context: 0 &sig->cred_guard_mutex &iint->mutex mapping.invalidate_lock &pcp->lock &zone->lock irq_context: 0 &sig->cred_guard_mutex &iint->mutex mapping.invalidate_lock &pcp->lock &zone->lock &____s->seqcount irq_context: 0 &f->f_pos_lock &type->i_mutex_dir_key#5 rcu_read_lock rcu_node_0 irq_context: 0 &iint->mutex mapping.invalidate_lock rcu_read_lock rcu_node_0 irq_context: 0 &f->f_pos_lock &type->i_mutex_dir_key#5 rcu_read_lock &rq->__lock irq_context: 0 &iint->mutex mapping.invalidate_lock rcu_read_lock &rq->__lock irq_context: 0 &iint->mutex &rq->__lock &cfs_rq->removed.lock irq_context: 0 &iint->mutex remove_cache_srcu irq_context: 0 &iint->mutex remove_cache_srcu quarantine_lock irq_context: 0 &iint->mutex remove_cache_srcu &c->lock irq_context: 0 &iint->mutex remove_cache_srcu &n->list_lock irq_context: 0 &iint->mutex remove_cache_srcu rcu_read_lock pool_lock#2 irq_context: 0 &iint->mutex remove_cache_srcu &obj_hash[i].lock irq_context: 0 sk_lock-AF_INET6 &____s->seqcount#8 irq_context: 0 sk_lock-AF_INET6 batched_entropy_u32.lock irq_context: 0 sk_lock-AF_INET6 rcu_read_lock rcu_read_lock rcu_read_lock &____s->seqcount irq_context: 0 sk_lock-AF_INET6 rcu_read_lock rcu_read_lock rcu_read_lock pool_lock#2 irq_context: 0 sk_lock-AF_INET6 rcu_read_lock rcu_read_lock rcu_read_lock &c->lock irq_context: 0 sk_lock-AF_INET6 rcu_read_lock rcu_read_lock rcu_read_lock rcu_read_lock pool_lock#2 irq_context: 0 sk_lock-AF_INET6 rcu_read_lock rcu_read_lock rcu_read_lock &obj_hash[i].lock irq_context: 0 sk_lock-AF_INET6 rcu_read_lock rcu_read_lock rcu_read_lock &dir->lock#2 irq_context: 0 sk_lock-AF_INET6 rcu_read_lock &____s->seqcount irq_context: 0 sk_lock-AF_INET6 rcu_read_lock pool_lock#2 irq_context: 0 sk_lock-AF_INET6 rcu_read_lock &c->lock irq_context: 0 sk_lock-AF_INET6 rcu_read_lock &dir->lock#2 irq_context: 0 sk_lock-AF_INET6 rcu_read_lock &ul->lock irq_context: 0 sk_lock-AF_INET6 &obj_hash[i].lock irq_context: 0 sk_lock-AF_INET6 pool_lock#2 irq_context: 0 sk_lock-AF_INET6 batched_entropy_u16.lock irq_context: 0 &sb->s_type->i_mutex_key#10 &table->hash[i].lock irq_context: 0 &sb->s_type->i_mutex_key#10 &table->hash[i].lock &table->hash2[i].lock irq_context: 0 &iint->mutex ima_extend_list_mutex &n->list_lock irq_context: 0 &iint->mutex ima_extend_list_mutex &n->list_lock &c->lock irq_context: softirq rcu_callback &ul->lock irq_context: 0 dup_mmap_sem &mm->mmap_lock &mm->mmap_lock/1 pgd_lock irq_context: 0 dup_mmap_sem &mm->mmap_lock &mm->mmap_lock/1 rcu_read_lock pool_lock#2 irq_context: 0 dup_mmap_sem &mm->mmap_lock &mm->mmap_lock/1 key irq_context: 0 dup_mmap_sem &mm->mmap_lock &mm->mmap_lock/1 pcpu_lock irq_context: 0 dup_mmap_sem &mm->mmap_lock &mm->mmap_lock/1 percpu_counters_lock irq_context: 0 dup_mmap_sem &mm->mmap_lock &mm->mmap_lock/1 &cfs_rq->removed.lock irq_context: 0 sb_writers#3 &sb->s_type->i_mutex_key#9 &p->alloc_lock irq_context: 0 sk_lock-AF_INET &tcp_hashinfo.bhash[i].lock irq_context: 0 sk_lock-AF_INET &tcp_hashinfo.bhash[i].lock &____s->seqcount irq_context: 0 sk_lock-AF_INET &tcp_hashinfo.bhash[i].lock pool_lock#2 irq_context: 0 sk_lock-AF_INET &tcp_hashinfo.bhash[i].lock &c->lock irq_context: 0 sk_lock-AF_INET &tcp_hashinfo.bhash[i].lock &tcp_hashinfo.bhash2[i].lock irq_context: 0 sk_lock-AF_INET &tcp_hashinfo.bhash[i].lock &tcp_hashinfo.bhash2[i].lock &____s->seqcount irq_context: 0 sk_lock-AF_INET &tcp_hashinfo.bhash[i].lock &tcp_hashinfo.bhash2[i].lock pool_lock#2 irq_context: 0 sk_lock-AF_INET &tcp_hashinfo.bhash[i].lock &tcp_hashinfo.bhash2[i].lock &c->lock irq_context: 0 sk_lock-AF_INET &tcp_hashinfo.bhash[i].lock &tcp_hashinfo.bhash2[i].lock clock-AF_INET irq_context: 0 sk_lock-AF_INET &tcp_hashinfo.bhash[i].lock clock-AF_INET irq_context: 0 sk_lock-AF_INET &h->lhash2[i].lock irq_context: 0 sk_lock-AF_INET6 &tcp_hashinfo.bhash[i].lock irq_context: 0 sk_lock-AF_INET6 &tcp_hashinfo.bhash[i].lock clock-AF_INET6 irq_context: 0 sk_lock-AF_INET6 &tcp_hashinfo.bhash[i].lock &tcp_hashinfo.bhash2[i].lock irq_context: 0 sk_lock-AF_INET6 &tcp_hashinfo.bhash[i].lock &tcp_hashinfo.bhash2[i].lock pool_lock#2 irq_context: 0 sk_lock-AF_INET6 &tcp_hashinfo.bhash[i].lock &tcp_hashinfo.bhash2[i].lock clock-AF_INET6 irq_context: 0 sk_lock-AF_INET6 &h->lhash2[i].lock irq_context: 0 &u->iolock &dir->lock irq_context: 0 rtnl_mutex _xmit_X25 irq_context: 0 rtnl_mutex lapb_list_lock irq_context: 0 rtnl_mutex lapb_list_lock pool_lock#2 irq_context: 0 rtnl_mutex lapb_list_lock &obj_hash[i].lock irq_context: 0 rtnl_mutex lapb_list_lock &base->lock irq_context: 0 rtnl_mutex lapb_list_lock &base->lock &obj_hash[i].lock irq_context: 0 rtnl_mutex &lapbeth->up_lock irq_context: 0 rtnl_mutex &lapb->lock irq_context: 0 rtnl_mutex &lapb->lock &c->lock irq_context: 0 rtnl_mutex &lapb->lock pool_lock#2 irq_context: 0 rtnl_mutex &lapb->lock rcu_read_lock_bh noop_qdisc.q.lock irq_context: 0 rtnl_mutex &lapb->lock rcu_read_lock_bh &obj_hash[i].lock irq_context: 0 rtnl_mutex &lapb->lock rcu_read_lock_bh pool_lock#2 irq_context: 0 rtnl_mutex &lapb->lock &obj_hash[i].lock irq_context: 0 rtnl_mutex &lapb->lock &base->lock irq_context: 0 rtnl_mutex &lapb->lock &base->lock &obj_hash[i].lock irq_context: 0 &tty->legacy_mutex &tty->ldisc_sem irq_context: 0 &tty->legacy_mutex tasklist_lock irq_context: 0 &tty->legacy_mutex tasklist_lock &sighand->siglock irq_context: 0 &tty->legacy_mutex tasklist_lock &sighand->siglock &tty->ctrl.lock irq_context: 0 &tty->ldisc_sem rcu_read_lock &tty->ctrl.lock irq_context: 0 &sig->cred_guard_mutex mapping.invalidate_lock &ei->i_data_sem &n->list_lock irq_context: 0 &sig->cred_guard_mutex mapping.invalidate_lock &ei->i_data_sem &n->list_lock &c->lock irq_context: 0 &tty->ctrl.lock irq_context: 0 tasklist_lock rcu_read_lock &sighand->siglock irq_context: 0 tasklist_lock &sighand->siglock irq_context: 0 &tty->legacy_mutex &tty->ctrl.lock irq_context: 0 &tty->legacy_mutex &f->f_lock irq_context: 0 &tty->legacy_mutex &f->f_lock fasync_lock irq_context: 0 &tty->legacy_mutex &obj_hash[i].lock irq_context: 0 &tty->legacy_mutex pool_lock#2 irq_context: 0 rcu_read_lock &tty->ctrl.lock irq_context: 0 remove_cache_srcu pgd_lock irq_context: 0 remove_cache_srcu key irq_context: 0 remove_cache_srcu pcpu_lock irq_context: 0 remove_cache_srcu percpu_counters_lock irq_context: 0 sb_writers#5 &type->i_mutex_dir_key#5 irq_context: 0 sb_writers#5 &type->i_mutex_dir_key#5 rename_lock.seqcount irq_context: 0 sb_writers#5 &type->i_mutex_dir_key#5 fs_reclaim irq_context: 0 sb_writers#5 &type->i_mutex_dir_key#5 fs_reclaim mmu_notifier_invalidate_range_start irq_context: 0 sb_writers#5 &type->i_mutex_dir_key#5 &dentry->d_lock irq_context: 0 sb_writers#5 &type->i_mutex_dir_key#5 rcu_read_lock rename_lock.seqcount irq_context: 0 sb_writers#5 &type->i_mutex_dir_key#5 &dentry->d_lock &wq#2 irq_context: 0 &port_lock_key irq_context: 0 sb_writers#4 &type->i_mutex_dir_key#3 &ei->xattr_sem &mapping->private_lock irq_context: 0 &buf->lock irq_context: 0 sb_writers#4 &sb->s_type->i_mutex_key#8 irq_context: 0 &tty->ldisc_sem &port_lock_key irq_context: 0 sb_writers#4 &sb->s_type->i_mutex_key#8 &ei->xattr_sem irq_context: 0 &tty->ldisc_sem &port->lock irq_context: 0 sb_writers#4 &sb->s_type->i_mutex_key#8 tk_core.seq.seqcount irq_context: 0 sb_writers#4 &sb->s_type->i_mutex_key#8 fs_reclaim irq_context: 0 sb_writers#4 &sb->s_type->i_mutex_key#8 fs_reclaim mmu_notifier_invalidate_range_start irq_context: 0 sb_writers#4 &sb->s_type->i_mutex_key#8 &____s->seqcount irq_context: 0 sb_writers#4 &sb->s_type->i_mutex_key#8 pool_lock#2 irq_context: 0 sb_writers#4 &sb->s_type->i_mutex_key#8 &xa->xa_lock#9 irq_context: 0 &tty->ldisc_sem &tty->termios_rwsem &tty->ldisc_sem &tty->flow.lock irq_context: 0 sb_writers#4 &sb->s_type->i_mutex_key#8 lock#4 irq_context: 0 sb_writers#4 &sb->s_type->i_mutex_key#8 mmu_notifier_invalidate_range_start irq_context: 0 sb_writers#4 &sb->s_type->i_mutex_key#8 &mapping->private_lock irq_context: 0 sb_writers#4 &sb->s_type->i_mutex_key#8 &ei->i_es_lock irq_context: 0 sb_writers#4 &sb->s_type->i_mutex_key#8 &ei->i_data_sem irq_context: 0 sb_writers#4 &sb->s_type->i_mutex_key#8 &ei->i_data_sem mmu_notifier_invalidate_range_start irq_context: 0 sb_writers#4 &sb->s_type->i_mutex_key#8 &ei->i_data_sem pool_lock#2 irq_context: 0 sb_writers#4 &sb->s_type->i_mutex_key#8 &ei->i_data_sem &ei->i_es_lock irq_context: 0 sb_writers#4 &sb->s_type->i_mutex_key#8 &ei->i_data_sem &ei->i_es_lock irq_context: 0 sb_writers#4 &sb->s_type->i_mutex_key#8 &ei->i_data_sem &ei->i_es_lock pool_lock#2 irq_context: 0 sb_writers#4 &sb->s_type->i_mutex_key#8 &ei->i_data_sem &ei->i_es_lock &sbi->s_es_lock irq_context: 0 sb_writers#4 &sb->s_type->i_mutex_key#8 &ei->i_data_sem &obj_hash[i].lock irq_context: 0 sb_writers#4 &sb->s_type->i_mutex_key#8 &ei->i_data_sem &sb->s_type->i_lock_key#22 irq_context: 0 sb_writers#4 &sb->s_type->i_mutex_key#8 &ei->i_data_sem &(ei->i_block_reservation_lock) irq_context: 0 sb_writers#4 &sb->s_type->i_mutex_key#8 rcu_read_lock &memcg->move_lock irq_context: 0 sb_writers#4 &sb->s_type->i_mutex_key#8 rcu_read_lock &xa->xa_lock#9 irq_context: 0 sb_writers#4 &sb->s_type->i_mutex_key#8 &sb->s_type->i_lock_key#22 irq_context: 0 sb_writers#4 &sb->s_type->i_mutex_key#8 &wb->list_lock irq_context: 0 sb_writers#4 &sb->s_type->i_mutex_key#8 &wb->list_lock &sb->s_type->i_lock_key#22 irq_context: 0 sb_writers#4 &sb->s_type->i_mutex_key#8 &journal->j_state_lock irq_context: 0 sb_writers#4 &sb->s_type->i_mutex_key#8 jbd2_handle irq_context: 0 sb_writers#4 &sb->s_type->i_mutex_key#8 jbd2_handle &ei->i_raw_lock irq_context: 0 sb_writers#4 &sb->s_type->i_mutex_key#8 jbd2_handle &journal->j_wait_updates irq_context: 0 sb_writers#4 &sb->s_type->i_mutex_key#8 &obj_hash[i].lock irq_context: 0 rtnl_mutex rcu_read_lock &obj_hash[i].lock irq_context: 0 rtnl_mutex rcu_read_lock rcu_read_lock &pool->lock/1 irq_context: 0 rtnl_mutex rcu_read_lock rcu_read_lock &pool->lock/1 &obj_hash[i].lock irq_context: 0 rtnl_mutex rcu_read_lock rcu_read_lock &pool->lock/1 pool_lock#2 irq_context: 0 rtnl_mutex rcu_read_lock rcu_read_lock &pool->lock/1 &p->pi_lock irq_context: 0 rtnl_mutex rcu_read_lock &net->sctp.local_addr_lock &net->sctp.addr_wq_lock &obj_hash[i].lock irq_context: 0 rtnl_mutex rcu_read_lock &net->sctp.local_addr_lock &net->sctp.addr_wq_lock &base->lock irq_context: 0 rtnl_mutex rcu_read_lock &net->sctp.local_addr_lock &net->sctp.addr_wq_lock &base->lock &obj_hash[i].lock irq_context: 0 (wq_completion)gid-cache-wq (work_completion)(&work->work) irq_context: 0 (wq_completion)gid-cache-wq (work_completion)(&work->work) devices_rwsem irq_context: 0 (wq_completion)gid-cache-wq (work_completion)(&work->work) &obj_hash[i].lock irq_context: 0 rtnl_mutex rcu_read_lock &pool->lock pool_lock#2 irq_context: 0 rtnl_mutex rcu_read_lock_bh &obj_hash[i].lock irq_context: 0 rtnl_mutex rcu_read_lock_bh &base->lock irq_context: 0 rtnl_mutex rcu_read_lock_bh &base->lock &obj_hash[i].lock irq_context: 0 (wq_completion)ipv6_addrconf (work_completion)(&(&ifa->dad_work)->work) irq_context: 0 (wq_completion)ipv6_addrconf (work_completion)(&(&ifa->dad_work)->work) rtnl_mutex irq_context: 0 (wq_completion)ipv6_addrconf (work_completion)(&(&ifa->dad_work)->work) rtnl_mutex rtnl_mutex.wait_lock irq_context: 0 (wq_completion)ipv6_addrconf (work_completion)(&(&ifa->dad_work)->work) rtnl_mutex &pool->lock irq_context: 0 (wq_completion)ipv6_addrconf (work_completion)(&(&ifa->dad_work)->work) rtnl_mutex &rq->__lock irq_context: 0 (wq_completion)ipv6_addrconf (work_completion)(&(&ifa->dad_work)->work) rtnl_mutex &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)ipv6_addrconf (work_completion)(&(&ifa->dad_work)->work) rtnl_mutex &ifa->lock irq_context: 0 (wq_completion)ipv6_addrconf (work_completion)(&(&ifa->dad_work)->work) rtnl_mutex &ndev->lock irq_context: 0 (wq_completion)ipv6_addrconf (work_completion)(&(&ifa->dad_work)->work) rtnl_mutex &ndev->lock &ifa->lock irq_context: 0 (wq_completion)ipv6_addrconf (work_completion)(&(&ifa->dad_work)->work) rtnl_mutex &obj_hash[i].lock irq_context: 0 (wq_completion)ipv6_addrconf (work_completion)(&(&ifa->dad_work)->work) rtnl_mutex &____s->seqcount irq_context: 0 (wq_completion)ipv6_addrconf (work_completion)(&(&ifa->dad_work)->work) rtnl_mutex pool_lock#2 irq_context: 0 (wq_completion)ipv6_addrconf (work_completion)(&(&ifa->dad_work)->work) rtnl_mutex rcu_read_lock pool_lock#2 irq_context: 0 (wq_completion)ipv6_addrconf (work_completion)(&(&ifa->dad_work)->work) rtnl_mutex nl_table_lock irq_context: 0 (wq_completion)ipv6_addrconf (work_completion)(&(&ifa->dad_work)->work) rtnl_mutex rlock-AF_NETLINK irq_context: 0 (wq_completion)ipv6_addrconf (work_completion)(&(&ifa->dad_work)->work) rtnl_mutex nl_table_wait.lock irq_context: 0 (wq_completion)ipv6_addrconf (work_completion)(&(&ifa->dad_work)->work) rtnl_mutex &tb->tb6_lock irq_context: 0 (wq_completion)ipv6_addrconf (work_completion)(&(&ifa->dad_work)->work) rtnl_mutex &tb->tb6_lock pool_lock#2 irq_context: 0 (wq_completion)ipv6_addrconf (work_completion)(&(&ifa->dad_work)->work) rtnl_mutex &tb->tb6_lock nl_table_lock irq_context: 0 (wq_completion)ipv6_addrconf (work_completion)(&(&ifa->dad_work)->work) rtnl_mutex &tb->tb6_lock rlock-AF_NETLINK irq_context: 0 (wq_completion)ipv6_addrconf (work_completion)(&(&ifa->dad_work)->work) rtnl_mutex &tb->tb6_lock nl_table_wait.lock irq_context: 0 (wq_completion)ipv6_addrconf (work_completion)(&(&ifa->dad_work)->work) rtnl_mutex &idev->mc_lock irq_context: 0 (wq_completion)ipv6_addrconf (work_completion)(&(&ifa->dad_work)->work) rtnl_mutex &idev->mc_lock batched_entropy_u32.lock irq_context: 0 (wq_completion)ipv6_addrconf (work_completion)(&(&ifa->dad_work)->work) rtnl_mutex &idev->mc_lock batched_entropy_u32.lock crngs.lock irq_context: 0 (wq_completion)ipv6_addrconf (work_completion)(&(&ifa->dad_work)->work) rtnl_mutex &idev->mc_lock &obj_hash[i].lock irq_context: 0 (wq_completion)ipv6_addrconf (work_completion)(&(&ifa->dad_work)->work) rtnl_mutex &idev->mc_lock &base->lock irq_context: 0 (wq_completion)ipv6_addrconf (work_completion)(&(&ifa->dad_work)->work) rtnl_mutex &idev->mc_lock &base->lock &obj_hash[i].lock irq_context: 0 (wq_completion)ipv6_addrconf (work_completion)(&(&ifa->dad_work)->work) rtnl_mutex rcu_read_lock &tb->tb6_lock irq_context: 0 (wq_completion)ipv6_addrconf (work_completion)(&(&ifa->dad_work)->work) rtnl_mutex rcu_read_lock &tb->tb6_lock &net->ipv6.fib6_walker_lock irq_context: softirq rcu_callback rcu_read_lock rt6_exception_lock irq_context: 0 &tty->ldisc_sem &ldata->atomic_read_lock irq_context: 0 &tty->ldisc_sem &ldata->atomic_read_lock &tty->termios_rwsem irq_context: 0 &tty->ldisc_sem &ldata->atomic_read_lock &tty->termios_rwsem &tty->read_wait irq_context: 0 &tty->ldisc_sem &ldata->atomic_read_lock (work_completion)(&buf->work) irq_context: 0 &tty->ldisc_sem &ldata->atomic_read_lock &rq->__lock irq_context: 0 &tty->ldisc_sem &ldata->atomic_read_lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: softirq &(&idev->mc_dad_work)->timer irq_context: softirq &(&idev->mc_dad_work)->timer rcu_read_lock &pool->lock irq_context: softirq &(&idev->mc_dad_work)->timer rcu_read_lock &pool->lock &obj_hash[i].lock irq_context: softirq &(&idev->mc_dad_work)->timer rcu_read_lock &pool->lock &p->pi_lock irq_context: softirq &(&idev->mc_dad_work)->timer rcu_read_lock &pool->lock &p->pi_lock &rq->__lock irq_context: softirq &(&idev->mc_dad_work)->timer rcu_read_lock &pool->lock &p->pi_lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)mld irq_context: 0 (wq_completion)mld (work_completion)(&(&idev->mc_dad_work)->work) irq_context: 0 (wq_completion)mld (work_completion)(&(&idev->mc_dad_work)->work) &idev->mc_lock irq_context: 0 rtnl_mutex lapb_list_lock &c->lock irq_context: 0 rtnl_mutex &lapb->lock rcu_read_lock_bh dev->qdisc_tx_busylock ?: &qdisc_tx_busylock _xmit_ETHER#2 irq_context: 0 rtnl_mutex &lapb->lock rcu_read_lock_bh dev->qdisc_tx_busylock ?: &qdisc_tx_busylock _xmit_ETHER#2 &obj_hash[i].lock irq_context: 0 rtnl_mutex &lapb->lock rcu_read_lock_bh dev->qdisc_tx_busylock ?: &qdisc_tx_busylock _xmit_ETHER#2 pool_lock#2 irq_context: 0 &sb->s_type->i_mutex_key#9 &obj_hash[i].lock pool_lock irq_context: 0 rcu_read_lock pgd_lock irq_context: 0 rcu_read_lock key irq_context: 0 rcu_read_lock pcpu_lock irq_context: 0 rcu_read_lock pool_lock irq_context: 0 rcu_read_lock percpu_counters_lock irq_context: 0 &net->packet.sklist_lock irq_context: 0 &net->packet.sklist_lock &rq->__lock irq_context: 0 &net->packet.sklist_lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 sk_lock-AF_PACKET irq_context: 0 sk_lock-AF_PACKET slock-AF_PACKET irq_context: 0 sk_lock-AF_PACKET &po->bind_lock irq_context: 0 sk_lock-AF_PACKET &po->bind_lock ptype_lock irq_context: 0 sk_lock-AF_PACKET rcu_state.exp_mutex rcu_node_0 irq_context: 0 sk_lock-AF_PACKET rcu_state.exp_mutex &obj_hash[i].lock irq_context: 0 sk_lock-AF_PACKET rcu_state.exp_mutex rcu_read_lock &pool->lock irq_context: 0 sk_lock-AF_PACKET rcu_state.exp_mutex rcu_read_lock &pool->lock &obj_hash[i].lock irq_context: 0 sk_lock-AF_PACKET rcu_state.exp_mutex rcu_read_lock &pool->lock &p->pi_lock irq_context: 0 sk_lock-AF_PACKET rcu_state.exp_mutex rcu_read_lock &pool->lock &p->pi_lock &rq->__lock irq_context: 0 sk_lock-AF_PACKET rcu_state.exp_mutex rcu_read_lock &pool->lock &p->pi_lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 sk_lock-AF_PACKET rcu_state.exp_mutex &rnp->exp_wq[0] irq_context: 0 sk_lock-AF_PACKET rcu_state.exp_mutex &rq->__lock irq_context: 0 sk_lock-AF_PACKET &obj_hash[i].lock irq_context: 0 sk_lock-AF_PACKET &po->bind_lock rcu_read_lock pool_lock#2 irq_context: 0 sk_lock-AF_PACKET &po->bind_lock rcu_read_lock &dir->lock#2 irq_context: 0 sk_lock-AF_PACKET &po->bind_lock rcu_read_lock ptype_lock irq_context: 0 slock-AF_PACKET irq_context: 0 sk_lock-AF_PACKET &mm->mmap_lock irq_context: 0 sk_lock-AF_PACKET fs_reclaim irq_context: 0 sk_lock-AF_PACKET fs_reclaim mmu_notifier_invalidate_range_start irq_context: 0 sk_lock-AF_PACKET pool_lock#2 irq_context: 0 sk_lock-AF_PACKET free_vmap_area_lock irq_context: 0 sk_lock-AF_PACKET free_vmap_area_lock &obj_hash[i].lock irq_context: 0 sk_lock-AF_PACKET vmap_area_lock irq_context: 0 sk_lock-AF_PACKET &____s->seqcount irq_context: 0 sk_lock-AF_PACKET &c->lock irq_context: 0 sk_lock-AF_PACKET pcpu_alloc_mutex irq_context: 0 sk_lock-AF_PACKET pcpu_alloc_mutex pcpu_lock irq_context: 0 sk_lock-AF_PACKET pack_mutex irq_context: 0 sk_lock-AF_PACKET batched_entropy_u32.lock irq_context: 0 sk_lock-AF_PACKET text_mutex irq_context: 0 sk_lock-AF_PACKET text_mutex ptlock_ptr(page)#2 irq_context: 0 sk_lock-AF_PACKET &fp->aux->used_maps_mutex irq_context: 0 rlock-AF_PACKET irq_context: 0 wlock-AF_PACKET irq_context: 0 (wq_completion)ipv6_addrconf (work_completion)(&(&ifa->dad_work)->work) rtnl_mutex &c->lock irq_context: 0 (wq_completion)ipv6_addrconf (work_completion)(&(&ifa->dad_work)->work) rtnl_mutex &tb->tb6_lock &c->lock irq_context: softirq _xmit_ETHER#2 &obj_hash[i].lock irq_context: softirq _xmit_ETHER#2 pool_lock#2 irq_context: 0 rtnl_mutex pgd_lock irq_context: 0 rtnl_mutex key irq_context: 0 rtnl_mutex pcpu_lock irq_context: 0 rtnl_mutex percpu_counters_lock irq_context: softirq (&rxnet->peer_keepalive_timer) rcu_read_lock &pool->lock/1 &p->pi_lock &rq->__lock irq_context: softirq (&rxnet->peer_keepalive_timer) rcu_read_lock &pool->lock/1 &p->pi_lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)writeback (work_completion)(&(&wb->dwork)->work) &pl->lock irq_context: 0 (wq_completion)writeback (work_completion)(&(&wb->dwork)->work) &pl->lock key#12 irq_context: 0 (wq_completion)events (linkwatch_work).work rtnl_mutex rcu_state.exp_mutex &rnp->exp_wq[1] irq_context: 0 (wq_completion)events (linkwatch_work).work rtnl_mutex &c->lock irq_context: 0 &mm->mmap_lock &anon_vma->rwsem mmu_notifier_invalidate_range_start irq_context: 0 &mm->mmap_lock &anon_vma->rwsem ptlock_ptr(page) irq_context: 0 &mm->mmap_lock remove_cache_srcu &pcp->lock &zone->lock irq_context: 0 &mm->mmap_lock remove_cache_srcu &pcp->lock &zone->lock &____s->seqcount irq_context: 0 &mm->mmap_lock lock#4 &pcp->lock &zone->lock irq_context: 0 &mm->mmap_lock lock#4 &pcp->lock &zone->lock &____s->seqcount irq_context: 0 (wq_completion)ipv6_addrconf (work_completion)(&(&ifa->dad_work)->work) rtnl_mutex &idev->mc_lock fs_reclaim irq_context: 0 (wq_completion)ipv6_addrconf (work_completion)(&(&ifa->dad_work)->work) rtnl_mutex &idev->mc_lock fs_reclaim mmu_notifier_invalidate_range_start irq_context: 0 (wq_completion)ipv6_addrconf (work_completion)(&(&ifa->dad_work)->work) rtnl_mutex &idev->mc_lock pool_lock#2 irq_context: 0 (wq_completion)ipv6_addrconf (work_completion)(&(&ifa->dad_work)->work) rtnl_mutex &idev->mc_lock _xmit_ETHER irq_context: 0 (wq_completion)ipv6_addrconf (work_completion)(&(&ifa->dad_work)->work) rtnl_mutex &idev->mc_lock _xmit_ETHER pool_lock#2 irq_context: 0 (wq_completion)ipv6_addrconf (work_completion)(&(&ifa->dad_work)->work) rtnl_mutex &ndev->lock &ifa->lock batched_entropy_u32.lock irq_context: 0 (wq_completion)ipv6_addrconf (work_completion)(&(&ifa->dad_work)->work) rtnl_mutex &ndev->lock &ifa->lock crngs.lock irq_context: 0 (wq_completion)ipv6_addrconf (work_completion)(&(&ifa->dad_work)->work) rtnl_mutex &ndev->lock &ifa->lock &obj_hash[i].lock irq_context: 0 (wq_completion)ipv6_addrconf (work_completion)(&(&ifa->dad_work)->work) rtnl_mutex &ndev->lock &ifa->lock &base->lock irq_context: 0 (wq_completion)ipv6_addrconf (work_completion)(&(&ifa->dad_work)->work) rtnl_mutex &ndev->lock &ifa->lock &base->lock &obj_hash[i].lock irq_context: 0 cb_lock rtnl_mutex rtnl_mutex.wait_lock irq_context: 0 cb_lock rtnl_mutex &rq->__lock irq_context: 0 cb_lock rtnl_mutex &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)ipv6_addrconf (work_completion)(&(&ifa->dad_work)->work) rtnl_mutex.wait_lock irq_context: 0 (wq_completion)ipv6_addrconf (work_completion)(&(&ifa->dad_work)->work) &p->pi_lock irq_context: 0 rtnl_mutex class irq_context: 0 rtnl_mutex (&tbl->proxy_timer) irq_context: softirq &(&idev->mc_ifc_work)->timer irq_context: softirq &(&idev->mc_ifc_work)->timer rcu_read_lock &pool->lock irq_context: softirq &(&idev->mc_ifc_work)->timer rcu_read_lock &pool->lock &obj_hash[i].lock irq_context: softirq &(&idev->mc_ifc_work)->timer rcu_read_lock &pool->lock &p->pi_lock irq_context: softirq &(&idev->mc_ifc_work)->timer rcu_read_lock &pool->lock &p->pi_lock &rq->__lock irq_context: softirq &(&idev->mc_ifc_work)->timer rcu_read_lock &pool->lock &p->pi_lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)mld (work_completion)(&(&idev->mc_ifc_work)->work) irq_context: 0 (wq_completion)mld (work_completion)(&(&idev->mc_ifc_work)->work) &idev->mc_lock irq_context: 0 (wq_completion)mld (work_completion)(&(&idev->mc_ifc_work)->work) &idev->mc_lock fs_reclaim irq_context: 0 (wq_completion)mld (work_completion)(&(&idev->mc_ifc_work)->work) &idev->mc_lock fs_reclaim mmu_notifier_invalidate_range_start irq_context: 0 (wq_completion)mld (work_completion)(&(&idev->mc_ifc_work)->work) &idev->mc_lock pool_lock#2 irq_context: 0 (wq_completion)mld (work_completion)(&(&idev->mc_ifc_work)->work) &idev->mc_lock rcu_read_lock &ndev->lock irq_context: 0 (wq_completion)mld (work_completion)(&(&idev->mc_ifc_work)->work) &idev->mc_lock rcu_read_lock &____s->seqcount irq_context: 0 (wq_completion)mld (work_completion)(&(&idev->mc_ifc_work)->work) &idev->mc_lock rcu_read_lock pool_lock#2 irq_context: 0 (wq_completion)mld (work_completion)(&(&idev->mc_ifc_work)->work) &idev->mc_lock rcu_read_lock &c->lock irq_context: 0 (wq_completion)mld (work_completion)(&(&idev->mc_ifc_work)->work) &idev->mc_lock rcu_read_lock &dir->lock#2 irq_context: 0 (wq_completion)mld (work_completion)(&(&idev->mc_ifc_work)->work) &idev->mc_lock rcu_read_lock &ul->lock#2 irq_context: 0 (wq_completion)mld (work_completion)(&(&idev->mc_ifc_work)->work) &idev->mc_lock rcu_read_lock rcu_read_lock pool_lock#2 irq_context: 0 (wq_completion)mld (work_completion)(&(&idev->mc_ifc_work)->work) &idev->mc_lock rcu_read_lock rcu_read_lock &obj_hash[i].lock irq_context: 0 (wq_completion)mld (work_completion)(&(&idev->mc_ifc_work)->work) &idev->mc_lock rcu_read_lock rcu_read_lock &dir->lock#2 irq_context: 0 (wq_completion)mld (work_completion)(&(&idev->mc_ifc_work)->work) &idev->mc_lock rcu_read_lock rcu_read_lock &tbl->lock irq_context: 0 (wq_completion)mld (work_completion)(&(&idev->mc_ifc_work)->work) &idev->mc_lock rcu_read_lock rcu_read_lock &n->lock irq_context: 0 (wq_completion)mld (work_completion)(&(&idev->mc_ifc_work)->work) &idev->mc_lock rcu_read_lock rcu_read_lock &____s->seqcount#9 irq_context: 0 (wq_completion)mld (work_completion)(&(&idev->mc_ifc_work)->work) &idev->mc_lock rcu_read_lock rcu_read_lock rcu_read_lock_bh &obj_hash[i].lock irq_context: 0 (wq_completion)mld (work_completion)(&(&idev->mc_ifc_work)->work) &idev->mc_lock rcu_read_lock rcu_read_lock rcu_read_lock_bh pool_lock#2 irq_context: 0 (wq_completion)mld (work_completion)(&(&idev->mc_ifc_work)->work) &idev->mc_lock rcu_read_lock rcu_read_lock rcu_read_lock_bh dev->qdisc_tx_busylock ?: &qdisc_tx_busylock _xmit_ETHER#2 irq_context: 0 (wq_completion)mld (work_completion)(&(&idev->mc_ifc_work)->work) &idev->mc_lock batched_entropy_u32.lock irq_context: 0 (wq_completion)mld (work_completion)(&(&idev->mc_ifc_work)->work) &idev->mc_lock &obj_hash[i].lock irq_context: 0 (wq_completion)mld (work_completion)(&(&idev->mc_ifc_work)->work) &idev->mc_lock &base->lock irq_context: 0 (wq_completion)mld (work_completion)(&(&idev->mc_ifc_work)->work) &idev->mc_lock &base->lock &obj_hash[i].lock irq_context: 0 rtnl_mutex rcu_read_lock rcu_read_lock &pool->lock/1 &p->pi_lock &rq->__lock irq_context: 0 rtnl_mutex rcu_read_lock rcu_read_lock &pool->lock/1 &p->pi_lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)ipv6_addrconf (work_completion)(&(&ifa->dad_work)->work) rtnl_mutex &idev->mc_lock _xmit_ETHER &local->filter_lock irq_context: 0 (wq_completion)ipv6_addrconf (work_completion)(&(&ifa->dad_work)->work) rtnl_mutex &idev->mc_lock _xmit_ETHER &local->filter_lock pool_lock#2 irq_context: 0 (wq_completion)ipv6_addrconf (work_completion)(&(&ifa->dad_work)->work) rtnl_mutex &idev->mc_lock _xmit_ETHER rcu_read_lock &pool->lock/1 irq_context: 0 (wq_completion)ipv6_addrconf (work_completion)(&(&ifa->dad_work)->work) rtnl_mutex &idev->mc_lock _xmit_ETHER rcu_read_lock &pool->lock/1 &obj_hash[i].lock irq_context: 0 (wq_completion)ipv6_addrconf (work_completion)(&(&ifa->dad_work)->work) rtnl_mutex &idev->mc_lock _xmit_ETHER rcu_read_lock &pool->lock/1 &p->pi_lock irq_context: 0 (wq_completion)ipv6_addrconf (work_completion)(&(&ifa->dad_work)->work) rtnl_mutex &idev->mc_lock _xmit_ETHER rcu_read_lock &pool->lock/1 &p->pi_lock &rq->__lock irq_context: 0 (wq_completion)ipv6_addrconf (work_completion)(&(&ifa->dad_work)->work) rtnl_mutex &idev->mc_lock _xmit_ETHER rcu_read_lock &pool->lock/1 &p->pi_lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)ipv6_addrconf (work_completion)(&(&ifa->dad_work)->work) rtnl_mutex &ifa->lock &obj_hash[i].lock irq_context: softirq rcu_callback &ul->lock#2 irq_context: 0 (wq_completion)mld (work_completion)(&(&idev->mc_ifc_work)->work) &idev->mc_lock rcu_read_lock rcu_read_lock rcu_read_lock_bh noop_qdisc.q.lock irq_context: 0 rtnl_mutex &net->ipv6.addrconf_hash_lock irq_context: 0 rtnl_mutex &idev->mc_lock _xmit_ETHER irq_context: 0 rtnl_mutex &idev->mc_lock _xmit_ETHER &local->filter_lock irq_context: 0 rtnl_mutex &idev->mc_lock _xmit_ETHER &local->filter_lock &obj_hash[i].lock irq_context: 0 rtnl_mutex &idev->mc_lock _xmit_ETHER &local->filter_lock pool_lock#2 irq_context: 0 rtnl_mutex &idev->mc_lock _xmit_ETHER &local->filter_lock krc.lock irq_context: 0 rtnl_mutex &idev->mc_lock _xmit_ETHER rcu_read_lock &pool->lock/1 irq_context: 0 rtnl_mutex &idev->mc_lock _xmit_ETHER rcu_read_lock &pool->lock/1 &obj_hash[i].lock irq_context: 0 rtnl_mutex &idev->mc_lock _xmit_ETHER rcu_read_lock &pool->lock/1 &p->pi_lock irq_context: 0 rtnl_mutex &idev->mc_lock _xmit_ETHER rcu_read_lock &pool->lock/1 &p->pi_lock &rq->__lock irq_context: 0 rtnl_mutex &idev->mc_lock _xmit_ETHER rcu_read_lock &pool->lock/1 &p->pi_lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 rtnl_mutex &idev->mc_lock &rq->__lock irq_context: 0 rtnl_mutex &idev->mc_lock fs_reclaim irq_context: 0 rtnl_mutex &idev->mc_lock fs_reclaim mmu_notifier_invalidate_range_start irq_context: 0 rtnl_mutex &idev->mc_lock pool_lock#2 irq_context: 0 rtnl_mutex &idev->mc_lock &obj_hash[i].lock irq_context: 0 rtnl_mutex &idev->mc_lock &base->lock irq_context: 0 rtnl_mutex &idev->mc_lock &base->lock &obj_hash[i].lock irq_context: 0 rtnl_mutex rcu_read_lock &dir->lock irq_context: 0 rtnl_mutex rcu_read_lock rcu_read_lock &pool->lock irq_context: 0 rtnl_mutex rcu_read_lock rcu_read_lock &pool->lock &obj_hash[i].lock irq_context: 0 rtnl_mutex rcu_read_lock rcu_read_lock &pool->lock pool_lock#2 irq_context: 0 rtnl_mutex rcu_read_lock rcu_read_lock &pool->lock &p->pi_lock irq_context: 0 rtnl_mutex rcu_read_lock rcu_read_lock &pool->lock &p->pi_lock &rq->__lock irq_context: 0 rtnl_mutex rcu_read_lock rcu_read_lock &pool->lock &p->pi_lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 rtnl_mutex rcu_read_lock krc.lock irq_context: 0 rtnl_mutex &tb->tb6_lock &net->ipv6.fib6_walker_lock irq_context: 0 rtnl_mutex &tb->tb6_lock rt6_exception_lock irq_context: 0 (wq_completion)events (work_completion)(&w->work)#2 irq_context: 0 (wq_completion)events (work_completion)(&w->work)#2 pool_lock#2 irq_context: 0 (wq_completion)events (work_completion)(&w->work)#2 &dir->lock irq_context: 0 (wq_completion)events (work_completion)(&w->work)#2 &obj_hash[i].lock irq_context: 0 cb_lock nlk_cb_mutex-GENERIC &c->lock irq_context: 0 (wq_completion)mld (work_completion)(&(&idev->mc_ifc_work)->work) &idev->mc_lock rcu_read_lock rcu_read_lock &____s->seqcount#10 irq_context: 0 rtnl_mutex rcu_read_lock rcu_node_0 irq_context: 0 rtnl_mutex rcu_read_lock &rq->__lock irq_context: 0 rtnl_mutex rcu_read_lock &n->list_lock irq_context: 0 rtnl_mutex rcu_read_lock &n->list_lock &c->lock irq_context: 0 (wq_completion)ipv6_addrconf (work_completion)(&(&ifa->dad_work)->work) rtnl_mutex &idev->mc_lock &rq->__lock irq_context: 0 rcu_read_lock rcu_read_lock rcu_read_lock pool_lock#2 irq_context: 0 rcu_read_lock rcu_read_lock rcu_read_lock &dir->lock#2 irq_context: 0 sk_lock-AF_INET6 fs_reclaim irq_context: 0 sk_lock-AF_INET6 fs_reclaim mmu_notifier_invalidate_range_start irq_context: 0 sk_lock-AF_INET6 &c->lock irq_context: 0 sk_lock-AF_INET6 &mm->mmap_lock irq_context: 0 sk_lock-AF_INET6 once_lock irq_context: 0 sk_lock-AF_INET6 once_lock crngs.lock irq_context: 0 sk_lock-AF_INET6 rcu_read_lock &pool->lock irq_context: 0 sk_lock-AF_INET6 rcu_read_lock &pool->lock &obj_hash[i].lock irq_context: 0 sk_lock-AF_INET6 rcu_read_lock &pool->lock pool_lock#2 irq_context: 0 sk_lock-AF_INET6 rcu_read_lock &pool->lock &p->pi_lock irq_context: 0 sk_lock-AF_INET6 rcu_read_lock &pool->lock &p->pi_lock &rq->__lock irq_context: 0 sk_lock-AF_INET6 rcu_read_lock &pool->lock &p->pi_lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 sk_lock-AF_INET6 rcu_node_0 irq_context: 0 sk_lock-AF_INET6 &rq->__lock irq_context: 0 sk_lock-AF_INET6 pgd_lock irq_context: 0 sk_lock-AF_INET6 key irq_context: 0 sk_lock-AF_INET6 pcpu_lock irq_context: 0 sk_lock-AF_INET6 percpu_counters_lock irq_context: 0 sk_lock-AF_INET6 rcu_read_lock rcu_node_0 irq_context: 0 sk_lock-AF_INET6 rcu_read_lock &pcp->lock &zone->lock irq_context: 0 sk_lock-AF_INET6 rcu_read_lock &pcp->lock &zone->lock &____s->seqcount irq_context: 0 sk_lock-AF_INET6 rcu_read_lock rcu_read_lock pool_lock#2 irq_context: 0 sk_lock-AF_INET6 rcu_read_lock &obj_hash[i].lock irq_context: 0 sk_lock-AF_INET6 rcu_read_lock &tbl->lock irq_context: 0 sk_lock-AF_INET6 rcu_read_lock &n->lock irq_context: 0 sk_lock-AF_INET6 rcu_read_lock &____s->seqcount#9 irq_context: 0 sk_lock-AF_INET6 rcu_read_lock rcu_read_lock_bh rcu_read_lock &c->lock irq_context: 0 sk_lock-AF_INET6 rcu_read_lock rcu_read_lock_bh rcu_read_lock pool_lock#2 irq_context: 0 sk_lock-AF_INET6 rcu_read_lock rcu_read_lock_bh rcu_read_lock tk_core.seq.seqcount irq_context: 0 sk_lock-AF_INET6 rcu_read_lock rcu_read_lock_bh rcu_read_lock &obj_hash[i].lock irq_context: 0 sk_lock-AF_INET6 rcu_read_lock rcu_read_lock_bh &obj_hash[i].lock irq_context: 0 sk_lock-AF_INET6 rcu_read_lock rcu_read_lock_bh pool_lock#2 irq_context: 0 sk_lock-AF_INET6 rcu_read_lock rcu_read_lock_bh dev->qdisc_tx_busylock ?: &qdisc_tx_busylock _xmit_ETHER#2 irq_context: 0 sk_lock-AF_INET6 rcu_read_lock rcu_read_lock_bh dev->qdisc_tx_busylock ?: &qdisc_tx_busylock _xmit_ETHER#2 &obj_hash[i].lock irq_context: 0 sk_lock-AF_INET6 rcu_read_lock rcu_read_lock_bh dev->qdisc_tx_busylock ?: &qdisc_tx_busylock _xmit_ETHER#2 pool_lock#2 irq_context: hardirq &irq_desc_lock_class vector_lock irq_context: 0 (wq_completion)mld (work_completion)(&(&idev->mc_ifc_work)->work) &idev->mc_lock &c->lock irq_context: 0 (wq_completion)mld (work_completion)(&(&idev->mc_ifc_work)->work) &idev->mc_lock krc.lock irq_context: softirq &(&idev->mc_dad_work)->timer rcu_read_lock &pool->lock &p->pi_lock &rq->__lock &cfs_rq->removed.lock irq_context: softirq &(&ifa->dad_work)->timer irq_context: softirq &(&ifa->dad_work)->timer rcu_read_lock &pool->lock irq_context: softirq &(&ifa->dad_work)->timer rcu_read_lock &pool->lock &obj_hash[i].lock irq_context: softirq &(&ifa->dad_work)->timer rcu_read_lock &pool->lock &p->pi_lock irq_context: softirq &(&ifa->dad_work)->timer rcu_read_lock &pool->lock &p->pi_lock &rq->__lock irq_context: softirq &(&ifa->dad_work)->timer rcu_read_lock &pool->lock &p->pi_lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)ipv6_addrconf (work_completion)(&(&ifa->dad_work)->work) rtnl_mutex &ndev->lock irq_context: 0 (wq_completion)ipv6_addrconf (work_completion)(&(&ifa->dad_work)->work) rtnl_mutex &ndev->lock &ifa->lock irq_context: 0 (wq_completion)ipv6_addrconf (work_completion)(&(&ifa->dad_work)->work) rtnl_mutex &ndev->lock &ifa->lock &obj_hash[i].lock irq_context: 0 (wq_completion)ipv6_addrconf (work_completion)(&(&ifa->dad_work)->work) rtnl_mutex &ndev->lock &ifa->lock &base->lock irq_context: 0 (wq_completion)ipv6_addrconf (work_completion)(&(&ifa->dad_work)->work) rtnl_mutex &ndev->lock &ifa->lock &base->lock &obj_hash[i].lock irq_context: 0 (wq_completion)ipv6_addrconf (work_completion)(&(&ifa->dad_work)->work) rtnl_mutex &dir->lock#2 irq_context: 0 (wq_completion)ipv6_addrconf (work_completion)(&(&ifa->dad_work)->work) rtnl_mutex &ul->lock#2 irq_context: 0 (wq_completion)ipv6_addrconf (work_completion)(&(&ifa->dad_work)->work) rtnl_mutex rcu_read_lock rcu_read_lock pool_lock#2 irq_context: 0 (wq_completion)ipv6_addrconf (work_completion)(&(&ifa->dad_work)->work) rtnl_mutex rcu_read_lock rcu_read_lock &obj_hash[i].lock irq_context: 0 (wq_completion)ipv6_addrconf (work_completion)(&(&ifa->dad_work)->work) rtnl_mutex rcu_read_lock rcu_read_lock &dir->lock#2 irq_context: 0 (wq_completion)ipv6_addrconf (work_completion)(&(&ifa->dad_work)->work) rtnl_mutex rcu_read_lock rcu_read_lock &tbl->lock irq_context: 0 (wq_completion)ipv6_addrconf (work_completion)(&(&ifa->dad_work)->work) rtnl_mutex rcu_read_lock rcu_read_lock &n->lock irq_context: 0 (wq_completion)ipv6_addrconf (work_completion)(&(&ifa->dad_work)->work) rtnl_mutex rcu_read_lock rcu_read_lock &____s->seqcount#9 irq_context: 0 (wq_completion)ipv6_addrconf (work_completion)(&(&ifa->dad_work)->work) rtnl_mutex rcu_read_lock rcu_read_lock rcu_read_lock_bh &obj_hash[i].lock irq_context: 0 (wq_completion)ipv6_addrconf (work_completion)(&(&ifa->dad_work)->work) rtnl_mutex rcu_read_lock rcu_read_lock rcu_read_lock_bh pool_lock#2 irq_context: 0 (wq_completion)ipv6_addrconf (work_completion)(&(&ifa->dad_work)->work) rtnl_mutex rcu_read_lock rcu_read_lock rcu_read_lock_bh dev->qdisc_tx_busylock ?: &qdisc_tx_busylock _xmit_ETHER#2 irq_context: 0 sk_lock-AF_PACKET rcu_state.exp_mutex rcu_read_lock &pool->lock &p->pi_lock &rq->__lock &cfs_rq->removed.lock irq_context: 0 sk_lock-AF_PACKET rcu_state.exp_mutex &rnp->exp_wq[2] irq_context: 0 rcu_read_lock_bh dev->qdisc_tx_busylock ?: &qdisc_tx_busylock _xmit_ETHER#2 irq_context: 0 rcu_read_lock_bh dev->qdisc_tx_busylock ?: &qdisc_tx_busylock _xmit_ETHER#2 &obj_hash[i].lock irq_context: 0 rcu_read_lock_bh dev->qdisc_tx_busylock ?: &qdisc_tx_busylock _xmit_ETHER#2 pool_lock#2 irq_context: softirq mm/memcontrol.c:589 rcu_read_lock &pool->lock/1 &p->pi_lock &cfs_rq->removed.lock irq_context: 0 (wq_completion)events (work_completion)(&(&krcp->monitor_work)->work) rcu_callback quarantine_lock irq_context: 0 sk_lock-AF_PACKET rcu_state.exp_mutex &rnp->exp_wq[3] irq_context: 0 sk_lock-AF_PACKET rcu_state.exp_mutex &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 &sb->s_type->i_mutex_key#10 &net->packet.sklist_lock irq_context: 0 &sb->s_type->i_mutex_key#10 &po->bind_lock irq_context: 0 &sb->s_type->i_mutex_key#10 &po->bind_lock ptype_lock irq_context: 0 &sb->s_type->i_mutex_key#10 &po->bind_lock pool_lock#2 irq_context: 0 &sb->s_type->i_mutex_key#10 &po->bind_lock &dir->lock#2 irq_context: 0 &sb->s_type->i_mutex_key#10 sk_lock-AF_PACKET irq_context: 0 &sb->s_type->i_mutex_key#10 sk_lock-AF_PACKET slock-AF_PACKET irq_context: 0 &sb->s_type->i_mutex_key#10 slock-AF_PACKET irq_context: 0 &sb->s_type->i_mutex_key#10 fanout_mutex irq_context: 0 &sb->s_type->i_mutex_key#10 rcu_state.exp_mutex rcu_node_0 irq_context: 0 &sb->s_type->i_mutex_key#10 rcu_state.exp_mutex &obj_hash[i].lock irq_context: 0 &sb->s_type->i_mutex_key#10 rcu_state.exp_mutex rcu_read_lock &pool->lock irq_context: 0 &sb->s_type->i_mutex_key#10 rcu_state.exp_mutex rcu_read_lock &pool->lock &obj_hash[i].lock irq_context: 0 &sb->s_type->i_mutex_key#10 rcu_state.exp_mutex rcu_read_lock &pool->lock &p->pi_lock irq_context: 0 &sb->s_type->i_mutex_key#10 rcu_state.exp_mutex rcu_read_lock &pool->lock &p->pi_lock &rq->__lock irq_context: 0 &sb->s_type->i_mutex_key#10 rcu_state.exp_mutex rcu_read_lock &pool->lock &p->pi_lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 &sb->s_type->i_mutex_key#10 rcu_state.exp_mutex &rnp->exp_wq[0] irq_context: 0 &sb->s_type->i_mutex_key#10 rcu_state.exp_mutex &rq->__lock irq_context: 0 &sb->s_type->i_mutex_key#10 clock-AF_PACKET irq_context: 0 &sb->s_type->i_mutex_key#10 rlock-AF_PACKET irq_context: 0 &sb->s_type->i_mutex_key#10 pcpu_lock irq_context: 0 &sb->s_type->i_mutex_key#10 elock-AF_PACKET irq_context: 0 (wq_completion)rcu_gp (work_completion)(&rew->rew_work) rcu_state.exp_wake_mutex &cfs_rq->removed.lock irq_context: 0 sk_lock-AF_PACKET rcu_state.exp_mutex &rnp->exp_wq[1] irq_context: 0 &sb->s_type->i_mutex_key#10 rcu_state.exp_mutex &rnp->exp_wq[2] irq_context: 0 &sb->s_type->i_mutex_key#10 rcu_state.exp_mutex &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 &journal->j_list_lock &obj_hash[i].lock irq_context: 0 &journal->j_list_lock &c->lock irq_context: 0 &journal->j_list_lock pool_lock#2 irq_context: 0 &mm->mmap_lock &meta->lock irq_context: 0 &mm->mmap_lock &obj_hash[i].lock pool_lock irq_context: 0 (wq_completion)rcu_gp (work_completion)(&rew->rew_work) rcu_state.exp_wake_mutex &obj_hash[i].lock irq_context: 0 (wq_completion)events (work_completion)(&p->wq) &base->lock irq_context: 0 (wq_completion)events (work_completion)(&p->wq) &base->lock &obj_hash[i].lock irq_context: 0 (wq_completion)mld (work_completion)(&(&idev->mc_ifc_work)->work) &idev->mc_lock krc.lock &obj_hash[i].lock irq_context: 0 (wq_completion)mld (work_completion)(&(&idev->mc_ifc_work)->work) &idev->mc_lock krc.lock &base->lock irq_context: 0 (wq_completion)mld (work_completion)(&(&idev->mc_ifc_work)->work) &idev->mc_lock krc.lock &base->lock &obj_hash[i].lock irq_context: 0 sk_lock-AF_PACKET &n->list_lock irq_context: 0 sk_lock-AF_PACKET &n->list_lock &c->lock irq_context: softirq _xmit_ETHER#2 rcu_read_lock &ei->socket.wq.wait irq_context: softirq rcu_read_lock rlock-AF_PACKET irq_context: softirq rcu_read_lock rcu_read_lock &ei->socket.wq.wait irq_context: softirq rcu_read_lock rcu_read_lock &ei->socket.wq.wait &p->pi_lock irq_context: softirq rcu_read_lock rcu_read_lock &ei->socket.wq.wait &p->pi_lock &cfs_rq->removed.lock irq_context: softirq rcu_read_lock rcu_read_lock &ei->socket.wq.wait &p->pi_lock &rq->__lock irq_context: softirq rcu_read_lock rcu_read_lock &ei->socket.wq.wait &p->pi_lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: softirq rcu_read_lock rcu_read_lock once_lock irq_context: softirq rcu_read_lock rcu_read_lock once_lock crngs.lock irq_context: softirq rcu_read_lock rcu_read_lock &obj_hash[i].lock irq_context: softirq rcu_read_lock rcu_read_lock rcu_read_lock &pool->lock irq_context: softirq rcu_read_lock rcu_read_lock rcu_read_lock &pool->lock &obj_hash[i].lock irq_context: softirq rcu_read_lock rcu_read_lock rcu_read_lock &pool->lock pool_lock#2 irq_context: softirq rcu_read_lock rcu_read_lock rcu_read_lock &pool->lock &p->pi_lock irq_context: softirq rcu_read_lock rcu_read_lock rcu_read_lock &pool->lock &p->pi_lock &rq->__lock irq_context: softirq rcu_read_lock rcu_read_lock rcu_read_lock &pool->lock &p->pi_lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: softirq rcu_read_lock rcu_read_lock &____s->seqcount#7 irq_context: softirq rcu_read_lock rcu_read_lock &____s->seqcount irq_context: softirq rcu_read_lock rcu_read_lock &c->lock irq_context: softirq rcu_read_lock rcu_read_lock &dir->lock#2 irq_context: 0 rcu_read_lock_bh dev->qdisc_tx_busylock ?: &qdisc_tx_busylock _xmit_ETHER#2 rcu_read_lock pool_lock#2 irq_context: 0 rcu_read_lock_bh dev->qdisc_tx_busylock ?: &qdisc_tx_busylock _xmit_ETHER#2 rcu_read_lock tk_core.seq.seqcount irq_context: 0 rcu_read_lock_bh dev->qdisc_tx_busylock ?: &qdisc_tx_busylock _xmit_ETHER#2 rcu_read_lock &obj_hash[i].lock irq_context: 0 (wq_completion)ipv6_addrconf (work_completion)(&(&ifa->dad_work)->work) rtnl_mutex batched_entropy_u8.lock irq_context: 0 (wq_completion)ipv6_addrconf (work_completion)(&(&ifa->dad_work)->work) rtnl_mutex kfence_freelist_lock irq_context: 0 (wq_completion)ipv6_addrconf (work_completion)(&(&ifa->dad_work)->work) rtnl_mutex rcu_read_lock &ei->socket.wq.wait irq_context: 0 (wq_completion)ipv6_addrconf (work_completion)(&(&ifa->dad_work)->work) rtnl_mutex rcu_read_lock &ei->socket.wq.wait &p->pi_lock irq_context: 0 (wq_completion)ipv6_addrconf (work_completion)(&(&ifa->dad_work)->work) rtnl_mutex rcu_read_lock &ei->socket.wq.wait &p->pi_lock &rq->__lock irq_context: 0 (wq_completion)ipv6_addrconf (work_completion)(&(&ifa->dad_work)->work) rtnl_mutex rcu_read_lock &ei->socket.wq.wait &p->pi_lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)ipv6_addrconf (work_completion)(&(&ifa->dad_work)->work) rtnl_mutex &tb->tb6_lock rcu_read_lock &ei->socket.wq.wait irq_context: 0 (wq_completion)ipv6_addrconf (work_completion)(&(&ifa->dad_work)->work) rtnl_mutex &idev->mc_lock rcu_read_lock &ndev->lock irq_context: 0 (wq_completion)ipv6_addrconf (work_completion)(&(&ifa->dad_work)->work) rtnl_mutex &idev->mc_lock rcu_read_lock pool_lock#2 irq_context: 0 (wq_completion)ipv6_addrconf (work_completion)(&(&ifa->dad_work)->work) rtnl_mutex &idev->mc_lock rcu_read_lock &c->lock irq_context: 0 (wq_completion)ipv6_addrconf (work_completion)(&(&ifa->dad_work)->work) rtnl_mutex &idev->mc_lock rcu_read_lock &dir->lock#2 irq_context: 0 (wq_completion)ipv6_addrconf (work_completion)(&(&ifa->dad_work)->work) rtnl_mutex &idev->mc_lock rcu_read_lock &ul->lock#2 irq_context: 0 (wq_completion)ipv6_addrconf (work_completion)(&(&ifa->dad_work)->work) rtnl_mutex &idev->mc_lock rcu_read_lock rcu_read_lock &____s->seqcount#10 irq_context: 0 (wq_completion)ipv6_addrconf (work_completion)(&(&ifa->dad_work)->work) rtnl_mutex &idev->mc_lock rcu_read_lock rcu_read_lock rcu_read_lock_bh &obj_hash[i].lock irq_context: 0 (wq_completion)ipv6_addrconf (work_completion)(&(&ifa->dad_work)->work) rtnl_mutex &idev->mc_lock rcu_read_lock rcu_read_lock rcu_read_lock_bh pool_lock#2 irq_context: 0 (wq_completion)ipv6_addrconf (work_completion)(&(&ifa->dad_work)->work) rtnl_mutex &idev->mc_lock rcu_read_lock rcu_read_lock rcu_read_lock_bh dev->qdisc_tx_busylock ?: &qdisc_tx_busylock _xmit_ETHER#2 irq_context: 0 (wq_completion)ipv6_addrconf (work_completion)(&(&ifa->dad_work)->work) rtnl_mutex &idev->mc_lock rcu_read_lock rcu_read_lock rcu_read_lock_bh dev->qdisc_tx_busylock ?: &qdisc_tx_busylock _xmit_ETHER#2 rcu_read_lock pool_lock#2 irq_context: 0 (wq_completion)ipv6_addrconf (work_completion)(&(&ifa->dad_work)->work) rtnl_mutex &idev->mc_lock rcu_read_lock rcu_read_lock rcu_read_lock_bh dev->qdisc_tx_busylock ?: &qdisc_tx_busylock _xmit_ETHER#2 rcu_read_lock tk_core.seq.seqcount irq_context: 0 (wq_completion)ipv6_addrconf (work_completion)(&(&ifa->dad_work)->work) rtnl_mutex &idev->mc_lock rcu_read_lock rcu_read_lock rcu_read_lock_bh dev->qdisc_tx_busylock ?: &qdisc_tx_busylock _xmit_ETHER#2 rcu_read_lock &obj_hash[i].lock irq_context: 0 sk_lock-AF_INET6 rcu_read_lock rcu_read_lock_bh dev->qdisc_tx_busylock ?: &qdisc_tx_busylock _xmit_ETHER#2 rcu_read_lock pool_lock#2 irq_context: 0 sk_lock-AF_INET6 rcu_read_lock rcu_read_lock_bh dev->qdisc_tx_busylock ?: &qdisc_tx_busylock _xmit_ETHER#2 rcu_read_lock tk_core.seq.seqcount irq_context: 0 sk_lock-AF_INET6 rcu_read_lock rcu_read_lock_bh dev->qdisc_tx_busylock ?: &qdisc_tx_busylock _xmit_ETHER#2 rcu_read_lock &obj_hash[i].lock irq_context: 0 rcu_read_lock rcu_read_lock rcu_read_lock &sighand->siglock irq_context: 0 rcu_read_lock rcu_read_lock &sighand->siglock irq_context: 0 rcu_read_lock rcu_read_lock &sighand->siglock batched_entropy_u8.lock irq_context: 0 rcu_read_lock rcu_read_lock &sighand->siglock kfence_freelist_lock irq_context: 0 rcu_read_lock rcu_read_lock &sighand->siglock &sighand->signalfd_wqh irq_context: 0 rcu_read_lock rcu_read_lock &sighand->siglock &sighand->signalfd_wqh &ep->lock irq_context: 0 rcu_read_lock rcu_read_lock &sighand->siglock &sighand->signalfd_wqh &ep->lock &ep->wq irq_context: 0 rcu_read_lock rcu_read_lock &sighand->siglock &sighand->signalfd_wqh &ep->lock &ep->wq &p->pi_lock irq_context: 0 rcu_read_lock rcu_read_lock &sighand->siglock &sighand->signalfd_wqh &ep->lock &ep->wq &p->pi_lock &rq->__lock irq_context: 0 rcu_read_lock rcu_read_lock &sighand->siglock &sighand->signalfd_wqh &ep->lock &ep->wq &p->pi_lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 rcu_read_lock rcu_read_lock &sighand->siglock pool_lock#2 irq_context: 0 &ep->mtx rcu_read_lock &sighand->signalfd_wqh irq_context: 0 &ep->mtx rcu_read_lock &ei->socket.wq.wait irq_context: 0 rcu_read_lock rcu_read_lock pgd_lock irq_context: 0 rcu_read_lock rcu_read_lock key irq_context: 0 rcu_read_lock rcu_read_lock pcpu_lock irq_context: 0 rcu_read_lock rcu_read_lock percpu_counters_lock irq_context: 0 rcu_read_lock rcu_read_lock &sighand->siglock &c->lock irq_context: 0 rcu_read_lock rcu_read_lock &sighand->siglock &____s->seqcount irq_context: 0 tasklist_lock &sighand->siglock &sighand->signalfd_wqh &ep->lock &ep->wq irq_context: 0 tasklist_lock &sighand->siglock &sighand->signalfd_wqh &ep->lock &ep->wq &p->pi_lock irq_context: 0 tasklist_lock &sighand->siglock &sighand->signalfd_wqh &ep->lock &ep->wq &p->pi_lock &rq->__lock irq_context: 0 tasklist_lock &sighand->siglock &sighand->signalfd_wqh &ep->lock &ep->wq &p->pi_lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 &ep->mtx pgd_lock irq_context: 0 &ep->mtx rcu_read_lock pool_lock#2 irq_context: 0 &ep->mtx key irq_context: 0 &ep->mtx pcpu_lock irq_context: 0 &ep->mtx percpu_counters_lock irq_context: 0 &ep->mtx &cfs_rq->removed.lock irq_context: 0 &mm->mmap_lock &anon_vma->rwsem &sem->wait_lock irq_context: 0 &mm->mmap_lock &anon_vma->rwsem &rq->__lock irq_context: 0 &mm->mmap_lock &anon_vma->rwsem &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 &mm->mmap_lock &anon_vma->rwsem &sem->wait_lock irq_context: 0 &mm->mmap_lock &sem->wait_lock irq_context: 0 &mm->mmap_lock &p->pi_lock irq_context: 0 &mm->mmap_lock &p->pi_lock &rq->__lock irq_context: 0 &mm->mmap_lock &p->pi_lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 &mm->mmap_lock &anon_vma->rwsem pgd_lock irq_context: 0 &mm->mmap_lock &anon_vma->rwsem rcu_read_lock pool_lock#2 irq_context: 0 &mm->mmap_lock &anon_vma->rwsem &obj_hash[i].lock irq_context: 0 &mm->mmap_lock &anon_vma->rwsem key irq_context: 0 &mm->mmap_lock &anon_vma->rwsem pcpu_lock irq_context: 0 &mm->mmap_lock &anon_vma->rwsem percpu_counters_lock irq_context: 0 &mm->mmap_lock &anon_vma->rwsem pool_lock#2 irq_context: 0 tasklist_lock &sighand->siglock &sighand->signalfd_wqh &ep->lock &ep->wq &p->pi_lock &rq->__lock &cfs_rq->removed.lock irq_context: 0 (wq_completion)events (work_completion)(&ht->run_work) &ht->mutex remove_cache_srcu irq_context: 0 (wq_completion)events (work_completion)(&ht->run_work) &ht->mutex remove_cache_srcu quarantine_lock irq_context: 0 (wq_completion)events (work_completion)(&ht->run_work) &ht->mutex remove_cache_srcu &c->lock irq_context: 0 (wq_completion)events (work_completion)(&ht->run_work) &ht->mutex remove_cache_srcu &n->list_lock irq_context: 0 (wq_completion)events (work_completion)(&ht->run_work) &ht->mutex remove_cache_srcu rcu_read_lock pool_lock#2 irq_context: 0 (wq_completion)events (work_completion)(&ht->run_work) &ht->mutex remove_cache_srcu &obj_hash[i].lock irq_context: 0 sk_lock-AF_PACKET &po->bind_lock rcu_read_lock &c->lock irq_context: 0 &sb->s_type->i_mutex_key#10 rcu_state.exp_mutex &rnp->exp_wq[1] irq_context: 0 (wq_completion)mld (work_completion)(&(&idev->mc_dad_work)->work) &idev->mc_lock fs_reclaim irq_context: 0 (wq_completion)mld (work_completion)(&(&idev->mc_dad_work)->work) &idev->mc_lock fs_reclaim mmu_notifier_invalidate_range_start irq_context: 0 (wq_completion)mld (work_completion)(&(&idev->mc_dad_work)->work) &idev->mc_lock batched_entropy_u8.lock irq_context: 0 (wq_completion)mld (work_completion)(&(&idev->mc_dad_work)->work) &idev->mc_lock kfence_freelist_lock irq_context: 0 (wq_completion)mld (work_completion)(&(&idev->mc_dad_work)->work) &idev->mc_lock pool_lock#2 irq_context: 0 (wq_completion)mld (work_completion)(&(&idev->mc_dad_work)->work) &idev->mc_lock rcu_read_lock &ndev->lock irq_context: 0 (wq_completion)mld (work_completion)(&(&idev->mc_dad_work)->work) &idev->mc_lock rcu_read_lock pool_lock#2 irq_context: 0 (wq_completion)mld (work_completion)(&(&idev->mc_dad_work)->work) &idev->mc_lock rcu_read_lock &dir->lock#2 irq_context: 0 (wq_completion)mld (work_completion)(&(&idev->mc_dad_work)->work) &idev->mc_lock rcu_read_lock &ul->lock#2 irq_context: 0 (wq_completion)mld (work_completion)(&(&idev->mc_dad_work)->work) &idev->mc_lock rcu_read_lock rcu_read_lock &____s->seqcount#10 irq_context: 0 (wq_completion)mld (work_completion)(&(&idev->mc_dad_work)->work) &idev->mc_lock rcu_read_lock rcu_read_lock rcu_read_lock_bh &obj_hash[i].lock irq_context: 0 (wq_completion)mld (work_completion)(&(&idev->mc_dad_work)->work) &idev->mc_lock rcu_read_lock rcu_read_lock rcu_read_lock_bh pool_lock#2 irq_context: 0 (wq_completion)mld (work_completion)(&(&idev->mc_dad_work)->work) &idev->mc_lock rcu_read_lock rcu_read_lock rcu_read_lock_bh dev->qdisc_tx_busylock ?: &qdisc_tx_busylock _xmit_ETHER#2 irq_context: 0 (wq_completion)mld (work_completion)(&(&idev->mc_dad_work)->work) &idev->mc_lock rcu_read_lock rcu_read_lock rcu_read_lock_bh dev->qdisc_tx_busylock ?: &qdisc_tx_busylock _xmit_ETHER#2 rcu_read_lock pool_lock#2 irq_context: 0 (wq_completion)mld (work_completion)(&(&idev->mc_dad_work)->work) &idev->mc_lock rcu_read_lock rcu_read_lock rcu_read_lock_bh dev->qdisc_tx_busylock ?: &qdisc_tx_busylock _xmit_ETHER#2 rcu_read_lock tk_core.seq.seqcount irq_context: 0 (wq_completion)mld (work_completion)(&(&idev->mc_dad_work)->work) &idev->mc_lock rcu_read_lock rcu_read_lock rcu_read_lock_bh dev->qdisc_tx_busylock ?: &qdisc_tx_busylock _xmit_ETHER#2 rcu_read_lock &obj_hash[i].lock irq_context: softirq rcu_read_lock rcu_read_lock batched_entropy_u8.lock irq_context: softirq rcu_read_lock rcu_read_lock kfence_freelist_lock irq_context: softirq rcu_read_lock rcu_read_lock &ct->lock irq_context: softirq rcu_read_lock once_lock irq_context: softirq rcu_read_lock once_lock crngs.lock irq_context: softirq rcu_read_lock rcu_read_lock &pool->lock irq_context: softirq rcu_read_lock rcu_read_lock &pool->lock &obj_hash[i].lock irq_context: softirq rcu_read_lock rcu_read_lock &pool->lock pool_lock#2 irq_context: softirq rcu_read_lock rcu_read_lock &pool->lock &p->pi_lock irq_context: softirq rcu_read_lock rcu_read_lock &pool->lock &p->pi_lock &rq->__lock irq_context: softirq rcu_read_lock rcu_read_lock &pool->lock &p->pi_lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: softirq rcu_read_lock &meta->lock irq_context: softirq rcu_read_lock kfence_freelist_lock irq_context: softirq &(&tbl->gc_work)->timer irq_context: softirq &(&tbl->gc_work)->timer rcu_read_lock &pool->lock irq_context: softirq &(&tbl->gc_work)->timer rcu_read_lock &pool->lock &obj_hash[i].lock irq_context: softirq &(&tbl->gc_work)->timer rcu_read_lock &pool->lock &p->pi_lock irq_context: softirq &(&tbl->gc_work)->timer rcu_read_lock &pool->lock &p->pi_lock &rq->__lock irq_context: softirq &(&tbl->gc_work)->timer rcu_read_lock &pool->lock &p->pi_lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)events_power_efficient (work_completion)(&(&tbl->gc_work)->work) irq_context: 0 (wq_completion)events_power_efficient (work_completion)(&(&tbl->gc_work)->work) &tbl->lock irq_context: 0 (wq_completion)events_power_efficient (work_completion)(&(&tbl->gc_work)->work) &tbl->lock &obj_hash[i].lock irq_context: 0 (wq_completion)events_power_efficient (work_completion)(&(&tbl->gc_work)->work) &tbl->lock &base->lock irq_context: 0 (wq_completion)events_power_efficient (work_completion)(&(&tbl->gc_work)->work) &tbl->lock &base->lock &obj_hash[i].lock irq_context: softirq (&lapb->t1timer) irq_context: softirq (&lapb->t1timer) &lapb->lock irq_context: softirq (&lapb->t1timer) &lapb->lock batched_entropy_u8.lock irq_context: softirq (&lapb->t1timer) &lapb->lock kfence_freelist_lock irq_context: softirq (&lapb->t1timer) &lapb->lock pool_lock#2 irq_context: softirq (&lapb->t1timer) &lapb->lock rcu_read_lock_bh noop_qdisc.q.lock irq_context: softirq (&lapb->t1timer) &lapb->lock rcu_read_lock_bh &obj_hash[i].lock irq_context: softirq (&lapb->t1timer) &lapb->lock rcu_read_lock_bh pool_lock#2 irq_context: softirq (&lapb->t1timer) &lapb->lock rcu_read_lock_bh &meta->lock irq_context: softirq (&lapb->t1timer) &lapb->lock rcu_read_lock_bh kfence_freelist_lock irq_context: softirq (&lapb->t1timer) &lapb->lock &obj_hash[i].lock irq_context: softirq (&lapb->t1timer) &lapb->lock &base->lock irq_context: softirq (&lapb->t1timer) &lapb->lock &base->lock &obj_hash[i].lock irq_context: 0 sk_lock-AF_INET6 &____s->seqcount irq_context: 0 sk_lock-AF_INET6 rcu_read_lock &____s->seqcount#10 irq_context: softirq (&lapb->t1timer) &lapb->lock &c->lock irq_context: softirq (&lapb->t1timer) &lapb->lock &n->list_lock irq_context: softirq (&lapb->t1timer) &lapb->lock &n->list_lock &c->lock irq_context: softirq (&lapb->t1timer) &lapb->lock rcu_read_lock_bh dev->qdisc_tx_busylock ?: &qdisc_tx_busylock _xmit_ETHER#2 irq_context: softirq (&lapb->t1timer) &lapb->lock rcu_read_lock_bh dev->qdisc_tx_busylock ?: &qdisc_tx_busylock _xmit_ETHER#2 rcu_read_lock pool_lock#2 irq_context: softirq (&lapb->t1timer) &lapb->lock rcu_read_lock_bh dev->qdisc_tx_busylock ?: &qdisc_tx_busylock _xmit_ETHER#2 rcu_read_lock tk_core.seq.seqcount irq_context: softirq (&lapb->t1timer) &lapb->lock rcu_read_lock_bh dev->qdisc_tx_busylock ?: &qdisc_tx_busylock _xmit_ETHER#2 rcu_read_lock &obj_hash[i].lock irq_context: softirq (&lapb->t1timer) &lapb->lock rcu_read_lock_bh dev->qdisc_tx_busylock ?: &qdisc_tx_busylock _xmit_ETHER#2 &obj_hash[i].lock irq_context: softirq (&lapb->t1timer) &lapb->lock rcu_read_lock_bh dev->qdisc_tx_busylock ?: &qdisc_tx_busylock _xmit_ETHER#2 pool_lock#2 irq_context: 0 rcu_read_lock_bh rcu_read_lock pool_lock#2 irq_context: 0 rcu_read_lock_bh rcu_read_lock tk_core.seq.seqcount irq_context: 0 rcu_read_lock_bh rcu_read_lock &obj_hash[i].lock irq_context: softirq (&dev->watchdog_timer) irq_context: softirq (&dev->watchdog_timer) &dev->tx_global_lock irq_context: softirq (&dev->watchdog_timer) &dev->tx_global_lock &obj_hash[i].lock irq_context: softirq (&dev->watchdog_timer) &dev->tx_global_lock &base->lock irq_context: softirq (&dev->watchdog_timer) &dev->tx_global_lock &base->lock &obj_hash[i].lock irq_context: 0 sk_lock-AF_INET6 rcu_read_lock rcu_read_lock_bh dev->qdisc_tx_busylock ?: &qdisc_tx_busylock _xmit_ETHER#2 rcu_read_lock &c->lock irq_context: 0 sk_lock-AF_INET6 rcu_read_lock rcu_read_lock_bh dev->qdisc_tx_busylock ?: &qdisc_tx_busylock _xmit_ETHER#2 rcu_read_lock &____s->seqcount irq_context: 0 &sb->s_type->i_mutex_key#10 rcu_state.exp_mutex &rnp->exp_wq[3] irq_context: 0 reading_mutex &rq->__lock irq_context: 0 reading_mutex &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: softirq &(&dm_bufio_cleanup_old_work)->timer irq_context: softirq &(&dm_bufio_cleanup_old_work)->timer rcu_read_lock &pool->lock irq_context: softirq &(&dm_bufio_cleanup_old_work)->timer rcu_read_lock &pool->lock &obj_hash[i].lock irq_context: softirq &(&dm_bufio_cleanup_old_work)->timer rcu_read_lock &pool->lock &p->pi_lock irq_context: softirq &(&dm_bufio_cleanup_old_work)->timer rcu_read_lock &pool->lock &p->pi_lock &rq->__lock irq_context: softirq &(&dm_bufio_cleanup_old_work)->timer rcu_read_lock &pool->lock &p->pi_lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: hardirq|softirq &x->wait#12 &p->pi_lock irq_context: hardirq|softirq &x->wait#12 &p->pi_lock &cfs_rq->removed.lock irq_context: hardirq|softirq &x->wait#12 &p->pi_lock &rq->__lock irq_context: hardirq|softirq &x->wait#12 &p->pi_lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)dm_bufio_cache irq_context: 0 (wq_completion)dm_bufio_cache (work_completion)(&(&dm_bufio_cleanup_old_work)->work) irq_context: 0 (wq_completion)dm_bufio_cache (work_completion)(&(&dm_bufio_cleanup_old_work)->work) dm_bufio_clients_lock irq_context: 0 (wq_completion)dm_bufio_cache (work_completion)(&(&dm_bufio_cleanup_old_work)->work) &obj_hash[i].lock irq_context: 0 (wq_completion)dm_bufio_cache (work_completion)(&(&dm_bufio_cleanup_old_work)->work) &base->lock irq_context: 0 (wq_completion)dm_bufio_cache (work_completion)(&(&dm_bufio_cleanup_old_work)->work) &base->lock &obj_hash[i].lock irq_context: 0 sb_writers#4 &type->i_mutex_dir_key#3 &journal->j_state_lock irq_context: 0 sb_writers#4 &type->i_mutex_dir_key#3 &journal->j_state_lock tk_core.seq.seqcount irq_context: 0 sb_writers#4 &type->i_mutex_dir_key#3 &journal->j_state_lock &obj_hash[i].lock irq_context: 0 sb_writers#4 &type->i_mutex_dir_key#3 &journal->j_state_lock &base->lock irq_context: 0 sb_writers#4 &type->i_mutex_dir_key#3 &journal->j_state_lock &base->lock &obj_hash[i].lock irq_context: 0 sb_writers#4 &sb->s_type->i_mutex_key#8 per_cpu_ptr(&cgroup_rstat_cpu_lock, cpu) irq_context: 0 sb_writers#4 &sb->s_type->i_mutex_key#8 &c->lock irq_context: 0 sb_writers#5 &sb->s_type->i_mutex_key#12 per_cpu_ptr(&cgroup_rstat_cpu_lock, cpu) irq_context: 0 rtnl_mutex (inetaddr_chain).rwsem rcu_read_lock &pool->lock/1 irq_context: 0 rtnl_mutex (inetaddr_chain).rwsem rcu_read_lock &pool->lock/1 &obj_hash[i].lock irq_context: 0 rtnl_mutex (inetaddr_chain).rwsem rcu_read_lock &pool->lock/1 pool_lock#2 irq_context: 0 rtnl_mutex (inetaddr_chain).rwsem rcu_read_lock &pool->lock/1 &p->pi_lock irq_context: 0 rtnl_mutex (inetaddr_chain).rwsem rcu_read_lock &pool->lock/1 &p->pi_lock &rq->__lock irq_context: 0 rtnl_mutex (inetaddr_chain).rwsem rcu_read_lock &pool->lock/1 &p->pi_lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 rtnl_mutex (inetaddr_chain).rwsem rlock-AF_NETLINK irq_context: 0 rtnl_mutex (inetaddr_chain).rwsem rcu_read_lock pool_lock#2 irq_context: 0 rtnl_mutex (inetaddr_chain).rwsem rcu_read_lock &dir->lock#2 irq_context: 0 sb_writers#5 &dentry->d_lock irq_context: 0 hostname_poll.wait.lock irq_context: 0 rcu_read_lock &pool->lock &p->pi_lock &rq->__lock &cfs_rq->removed.lock irq_context: 0 (wq_completion)events (work_completion)(&p->wq) rcu_read_lock &rcu_state.expedited_wq irq_context: 0 rcu_read_lock_bh dev->qdisc_tx_busylock ?: &qdisc_tx_busylock _xmit_ETHER#2 rcu_read_lock &c->lock irq_context: softirq &(&group->avgs_work)->timer rcu_read_lock &pool->lock &p->pi_lock &rq->__lock &cfs_rq->removed.lock irq_context: softirq (&lapb->t1timer) &lapb->lock rcu_read_lock_bh dev->qdisc_tx_busylock ?: &qdisc_tx_busylock _xmit_ETHER#2 rcu_read_lock &c->lock irq_context: 0 &mm->mmap_lock ptlock_ptr(page)#2 &pcp->lock &zone->lock &____s->seqcount irq_context: 0 rcu_read_lock_bh rcu_read_lock &c->lock irq_context: softirq rcu_read_lock rcu_read_lock &nf_conntrack_locks[i] irq_context: softirq rcu_read_lock rcu_read_lock &nf_conntrack_locks[i] &nf_conntrack_locks[i]/1 irq_context: softirq rcu_read_lock rcu_read_lock &nf_conntrack_locks[i] &nf_conntrack_locks[i]/1 batched_entropy_u8.lock irq_context: softirq rcu_read_lock rcu_read_lock rcu_read_lock &____s->seqcount irq_context: softirq rcu_read_lock rcu_read_lock rcu_read_lock &c->lock irq_context: softirq rcu_read_lock rcu_read_lock rcu_read_lock rcu_read_lock rcu_read_lock pool_lock#2 irq_context: softirq rcu_read_lock rcu_read_lock rcu_read_lock rcu_read_lock rcu_read_lock &dir->lock#2 irq_context: softirq rcu_read_lock rcu_read_lock rcu_read_lock once_lock irq_context: softirq rcu_read_lock rcu_read_lock rcu_read_lock once_lock crngs.lock irq_context: softirq rcu_read_lock rcu_read_lock rcu_read_lock &obj_hash[i].lock irq_context: softirq rcu_read_lock rcu_read_lock rcu_read_lock rcu_read_lock &pool->lock irq_context: softirq rcu_read_lock rcu_read_lock rcu_read_lock rcu_read_lock &pool->lock &obj_hash[i].lock irq_context: softirq rcu_read_lock rcu_read_lock rcu_read_lock rcu_read_lock &pool->lock pool_lock#2 irq_context: softirq rcu_read_lock rcu_read_lock rcu_read_lock rcu_read_lock &pool->lock &p->pi_lock irq_context: softirq rcu_read_lock rcu_read_lock rcu_read_lock rcu_read_lock &pool->lock &p->pi_lock &rq->__lock irq_context: softirq rcu_read_lock rcu_read_lock rcu_read_lock rcu_read_lock &pool->lock &p->pi_lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: softirq rcu_read_lock rcu_read_lock rcu_read_lock tk_core.seq.seqcount irq_context: softirq rcu_read_lock rcu_read_lock rcu_read_lock batched_entropy_u32.lock irq_context: softirq rcu_read_lock rcu_read_lock rcu_read_lock &base->lock irq_context: softirq rcu_read_lock rcu_read_lock rcu_read_lock &base->lock &obj_hash[i].lock irq_context: softirq rcu_read_lock rcu_read_lock rcu_read_lock &hashinfo->ehash_locks[i] irq_context: softirq rcu_read_lock rcu_read_lock rcu_read_lock rcu_read_lock rcu_read_lock &____s->seqcount#7 irq_context: softirq rcu_read_lock rcu_read_lock rcu_read_lock rcu_read_lock rcu_read_lock &ct->lock irq_context: softirq rcu_read_lock rcu_read_lock rcu_read_lock rcu_read_lock rcu_read_lock &obj_hash[i].lock irq_context: softirq rcu_read_lock rcu_read_lock rcu_read_lock rcu_read_lock rcu_read_lock &tbl->lock irq_context: softirq rcu_read_lock rcu_read_lock rcu_read_lock rcu_read_lock rcu_read_lock &n->lock irq_context: softirq rcu_read_lock rcu_read_lock rcu_read_lock rcu_read_lock rcu_read_lock &n->lock &obj_hash[i].lock irq_context: softirq rcu_read_lock rcu_read_lock rcu_read_lock rcu_read_lock rcu_read_lock &n->lock &base->lock irq_context: softirq rcu_read_lock rcu_read_lock rcu_read_lock rcu_read_lock rcu_read_lock &n->lock &base->lock &obj_hash[i].lock irq_context: softirq rcu_read_lock rcu_read_lock rcu_read_lock rcu_read_lock rcu_read_lock &n->lock &c->lock irq_context: softirq rcu_read_lock rcu_read_lock rcu_read_lock rcu_read_lock rcu_read_lock &n->lock pool_lock#2 irq_context: softirq rcu_read_lock rcu_read_lock rcu_read_lock rcu_read_lock rcu_read_lock rcu_read_lock_bh dev->qdisc_tx_busylock ?: &qdisc_tx_busylock _xmit_ETHER#2 irq_context: softirq rcu_read_lock rcu_read_lock rcu_read_lock rcu_read_lock rcu_read_lock rcu_read_lock_bh dev->qdisc_tx_busylock ?: &qdisc_tx_busylock _xmit_ETHER#2 rcu_read_lock pool_lock#2 irq_context: softirq rcu_read_lock rcu_read_lock rcu_read_lock rcu_read_lock rcu_read_lock rcu_read_lock_bh dev->qdisc_tx_busylock ?: &qdisc_tx_busylock _xmit_ETHER#2 rcu_read_lock tk_core.seq.seqcount irq_context: softirq rcu_read_lock rcu_read_lock rcu_read_lock rcu_read_lock rcu_read_lock rcu_read_lock_bh dev->qdisc_tx_busylock ?: &qdisc_tx_busylock _xmit_ETHER#2 rcu_read_lock &obj_hash[i].lock irq_context: softirq rcu_read_lock &n->lock irq_context: softirq rcu_read_lock &n->lock &obj_hash[i].lock irq_context: softirq rcu_read_lock &n->lock &base->lock irq_context: softirq rcu_read_lock &n->lock &base->lock &obj_hash[i].lock irq_context: softirq rcu_read_lock &n->lock &(&n->ha_lock)->lock irq_context: softirq rcu_read_lock &n->lock &(&n->ha_lock)->lock &____s->seqcount#9 irq_context: softirq rcu_read_lock rcu_read_lock &n->lock irq_context: softirq rcu_read_lock rcu_read_lock &____s->seqcount#9 irq_context: softirq rcu_read_lock rcu_read_lock rcu_read_lock_bh dev->qdisc_tx_busylock ?: &qdisc_tx_busylock _xmit_ETHER#2 irq_context: softirq rcu_read_lock rcu_read_lock rcu_read_lock_bh dev->qdisc_tx_busylock ?: &qdisc_tx_busylock _xmit_ETHER#2 rcu_read_lock pool_lock#2 irq_context: softirq rcu_read_lock rcu_read_lock rcu_read_lock_bh dev->qdisc_tx_busylock ?: &qdisc_tx_busylock _xmit_ETHER#2 rcu_read_lock tk_core.seq.seqcount irq_context: softirq rcu_read_lock rcu_read_lock rcu_read_lock_bh dev->qdisc_tx_busylock ?: &qdisc_tx_busylock _xmit_ETHER#2 rcu_read_lock &obj_hash[i].lock irq_context: softirq rcu_read_lock rcu_read_lock lock#8 irq_context: softirq rcu_read_lock rcu_read_lock id_table_lock irq_context: softirq rcu_read_lock &n->lock irq_context: softirq rcu_read_lock &n->lock &____s->seqcount#9 irq_context: softirq rcu_read_lock nl_table_lock irq_context: softirq rcu_read_lock rlock-AF_NETLINK irq_context: softirq rcu_read_lock rcu_read_lock &dir->lock irq_context: softirq rcu_read_lock rcu_read_lock slock-AF_INET irq_context: softirq rcu_read_lock rcu_read_lock slock-AF_INET &obj_hash[i].lock irq_context: softirq rcu_read_lock rcu_read_lock slock-AF_INET batched_entropy_u16.lock irq_context: softirq rcu_read_lock rcu_read_lock slock-AF_INET &tcp_hashinfo.bhash[i].lock irq_context: softirq rcu_read_lock rcu_read_lock slock-AF_INET &tcp_hashinfo.bhash[i].lock &tcp_hashinfo.bhash2[i].lock irq_context: softirq rcu_read_lock rcu_read_lock slock-AF_INET &tcp_hashinfo.bhash[i].lock &tcp_hashinfo.bhash2[i].lock pool_lock#2 irq_context: softirq rcu_read_lock rcu_read_lock slock-AF_INET &hashinfo->ehash_locks[i] irq_context: softirq rcu_read_lock rcu_read_lock slock-AF_INET tk_core.seq.seqcount irq_context: softirq rcu_read_lock rcu_read_lock slock-AF_INET (&req->rsk_timer) irq_context: softirq rcu_read_lock rcu_read_lock slock-AF_INET &base->lock irq_context: softirq rcu_read_lock rcu_read_lock slock-AF_INET &base->lock &obj_hash[i].lock irq_context: softirq rcu_read_lock rcu_read_lock slock-AF_INET &queue->rskq_lock irq_context: softirq rcu_read_lock rcu_read_lock slock-AF_INET rcu_read_lock tcp_metrics_lock irq_context: softirq rcu_read_lock rcu_read_lock slock-AF_INET rcu_read_lock tcp_metrics_lock pool_lock#2 irq_context: softirq rcu_read_lock rcu_read_lock slock-AF_INET pool_lock#2 irq_context: softirq rcu_read_lock rcu_read_lock slock-AF_INET rcu_read_lock &ei->socket.wq.wait irq_context: softirq rcu_read_lock rcu_read_lock slock-AF_INET rcu_read_lock &ei->socket.wq.wait &p->pi_lock irq_context: softirq rcu_read_lock rcu_read_lock slock-AF_INET rcu_read_lock &ei->socket.wq.wait &p->pi_lock &cfs_rq->removed.lock irq_context: softirq rcu_read_lock rcu_read_lock slock-AF_INET rcu_read_lock &ei->socket.wq.wait &p->pi_lock &rq->__lock irq_context: softirq rcu_read_lock rcu_read_lock slock-AF_INET rcu_read_lock &ei->socket.wq.wait &p->pi_lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: softirq rcu_read_lock rcu_read_lock slock-AF_INET/1 irq_context: softirq rcu_read_lock rcu_read_lock slock-AF_INET/1 tk_core.seq.seqcount irq_context: softirq rcu_read_lock rcu_read_lock slock-AF_INET/1 pool_lock#2 irq_context: softirq rcu_read_lock rcu_read_lock slock-AF_INET/1 rcu_read_lock rcu_read_lock &____s->seqcount#7 irq_context: softirq rcu_read_lock rcu_read_lock slock-AF_INET/1 rcu_read_lock rcu_read_lock &ct->lock irq_context: softirq rcu_read_lock rcu_read_lock slock-AF_INET/1 rcu_read_lock rcu_read_lock &____s->seqcount#10 irq_context: softirq rcu_read_lock rcu_read_lock slock-AF_INET/1 rcu_read_lock rcu_read_lock rcu_read_lock_bh dev->qdisc_tx_busylock ?: &qdisc_tx_busylock _xmit_ETHER#2 irq_context: softirq rcu_read_lock rcu_read_lock slock-AF_INET/1 rcu_read_lock rcu_read_lock rcu_read_lock_bh dev->qdisc_tx_busylock ?: &qdisc_tx_busylock _xmit_ETHER#2 rcu_read_lock pool_lock#2 irq_context: softirq rcu_read_lock rcu_read_lock slock-AF_INET/1 rcu_read_lock rcu_read_lock rcu_read_lock_bh dev->qdisc_tx_busylock ?: &qdisc_tx_busylock _xmit_ETHER#2 rcu_read_lock tk_core.seq.seqcount irq_context: softirq rcu_read_lock rcu_read_lock slock-AF_INET/1 rcu_read_lock rcu_read_lock rcu_read_lock_bh dev->qdisc_tx_busylock ?: &qdisc_tx_busylock _xmit_ETHER#2 rcu_read_lock &obj_hash[i].lock irq_context: 0 sk_lock-AF_INET &queue->rskq_lock irq_context: 0 sk_lock-AF_INET clock-AF_INET irq_context: 0 sb_writers#3 &sb->s_type->i_mutex_key#9 tomoyo_ss &n->list_lock irq_context: 0 sb_writers#3 &sb->s_type->i_mutex_key#9 tomoyo_ss &n->list_lock &c->lock irq_context: 0 sk_lock-AF_INET &obj_hash[i].lock irq_context: 0 sk_lock-AF_INET &base->lock irq_context: 0 sk_lock-AF_INET &base->lock &obj_hash[i].lock irq_context: 0 sk_lock-AF_INET fs_reclaim irq_context: 0 sk_lock-AF_INET fs_reclaim mmu_notifier_invalidate_range_start irq_context: 0 sk_lock-AF_INET &____s->seqcount irq_context: 0 sk_lock-AF_INET pool_lock#2 irq_context: 0 sk_lock-AF_INET &c->lock irq_context: 0 sk_lock-AF_INET &mm->mmap_lock irq_context: 0 sk_lock-AF_INET tk_core.seq.seqcount irq_context: 0 sk_lock-AF_INET rcu_read_lock rcu_read_lock &____s->seqcount#7 irq_context: 0 sk_lock-AF_INET rcu_read_lock rcu_read_lock &ct->lock irq_context: 0 sk_lock-AF_INET rcu_read_lock rcu_read_lock &____s->seqcount#10 irq_context: 0 sk_lock-AF_INET rcu_read_lock rcu_read_lock rcu_read_lock_bh dev->qdisc_tx_busylock ?: &qdisc_tx_busylock _xmit_ETHER#2 irq_context: 0 sk_lock-AF_INET rcu_read_lock rcu_read_lock rcu_read_lock_bh dev->qdisc_tx_busylock ?: &qdisc_tx_busylock _xmit_ETHER#2 rcu_read_lock pool_lock#2 irq_context: 0 sk_lock-AF_INET rcu_read_lock rcu_read_lock rcu_read_lock_bh dev->qdisc_tx_busylock ?: &qdisc_tx_busylock _xmit_ETHER#2 rcu_read_lock tk_core.seq.seqcount irq_context: 0 sk_lock-AF_INET rcu_read_lock rcu_read_lock rcu_read_lock_bh dev->qdisc_tx_busylock ?: &qdisc_tx_busylock _xmit_ETHER#2 rcu_read_lock &obj_hash[i].lock irq_context: 0 sk_lock-AF_INET &sd->defer_lock irq_context: softirq &sd->defer_lock irq_context: 0 sk_lock-AF_INET rcu_read_lock pool_lock#2 irq_context: softirq rcu_read_lock rcu_read_lock slock-AF_INET/1 rcu_read_lock &ei->socket.wq.wait irq_context: softirq rcu_read_lock rcu_read_lock slock-AF_INET/1 rcu_read_lock &ei->socket.wq.wait &p->pi_lock irq_context: softirq rcu_read_lock rcu_read_lock slock-AF_INET/1 rcu_read_lock &ei->socket.wq.wait &p->pi_lock &rq->__lock irq_context: softirq rcu_read_lock rcu_read_lock slock-AF_INET/1 rcu_read_lock &ei->socket.wq.wait &p->pi_lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: softirq rcu_read_lock rcu_read_lock slock-AF_INET/1 &obj_hash[i].lock irq_context: softirq rcu_read_lock rcu_read_lock slock-AF_INET/1 &base->lock irq_context: softirq rcu_read_lock rcu_read_lock slock-AF_INET/1 &base->lock &obj_hash[i].lock irq_context: 0 sk_lock-AF_INET rcu_read_lock rcu_read_lock pool_lock#2 irq_context: 0 sk_lock-AF_INET rcu_read_lock rcu_read_lock &dir->lock#2 irq_context: softirq rcu_read_lock rcu_read_lock &ul->lock irq_context: softirq (&icsk->icsk_delack_timer) irq_context: softirq (&icsk->icsk_delack_timer) slock-AF_INET irq_context: softirq (&icsk->icsk_retransmit_timer) irq_context: softirq (&icsk->icsk_retransmit_timer) slock-AF_INET irq_context: softirq rcu_read_lock rcu_read_lock slock-AF_INET/1 rcu_read_lock &ei->socket.wq.wait &p->pi_lock &cfs_rq->removed.lock irq_context: 0 &pipe->mutex/1 &mm->mmap_lock fs_reclaim irq_context: 0 &pipe->mutex/1 &mm->mmap_lock fs_reclaim mmu_notifier_invalidate_range_start irq_context: 0 &pipe->mutex/1 &mm->mmap_lock &____s->seqcount irq_context: 0 &pipe->mutex/1 &mm->mmap_lock mmu_notifier_invalidate_range_start irq_context: 0 &pipe->mutex/1 &mm->mmap_lock ptlock_ptr(page)#2 lock#4 irq_context: softirq (&icsk->icsk_delack_timer) slock-AF_INET tk_core.seq.seqcount irq_context: softirq (&icsk->icsk_delack_timer) slock-AF_INET pool_lock#2 irq_context: softirq (&icsk->icsk_delack_timer) slock-AF_INET rcu_read_lock rcu_read_lock &____s->seqcount#7 irq_context: softirq (&icsk->icsk_delack_timer) slock-AF_INET rcu_read_lock rcu_read_lock &ct->lock irq_context: softirq (&icsk->icsk_delack_timer) slock-AF_INET rcu_read_lock rcu_read_lock &____s->seqcount#10 irq_context: softirq (&icsk->icsk_delack_timer) slock-AF_INET rcu_read_lock rcu_read_lock rcu_read_lock_bh dev->qdisc_tx_busylock ?: &qdisc_tx_busylock _xmit_ETHER#2 irq_context: softirq (&icsk->icsk_delack_timer) slock-AF_INET rcu_read_lock rcu_read_lock rcu_read_lock_bh dev->qdisc_tx_busylock ?: &qdisc_tx_busylock _xmit_ETHER#2 rcu_read_lock pool_lock#2 irq_context: softirq (&icsk->icsk_delack_timer) slock-AF_INET rcu_read_lock rcu_read_lock rcu_read_lock_bh dev->qdisc_tx_busylock ?: &qdisc_tx_busylock _xmit_ETHER#2 rcu_read_lock tk_core.seq.seqcount irq_context: softirq (&icsk->icsk_delack_timer) slock-AF_INET rcu_read_lock rcu_read_lock rcu_read_lock_bh dev->qdisc_tx_busylock ?: &qdisc_tx_busylock _xmit_ETHER#2 rcu_read_lock &obj_hash[i].lock irq_context: softirq rcu_read_lock rcu_read_lock slock-AF_INET/1 rcu_read_lock rcu_read_lock rcu_read_lock_bh dev->qdisc_tx_busylock ?: &qdisc_tx_busylock _xmit_ETHER#2 rcu_read_lock &c->lock irq_context: 0 sk_lock-AF_INET rcu_read_lock rcu_read_lock rcu_read_lock_bh dev->qdisc_tx_busylock ?: &qdisc_tx_busylock _xmit_ETHER#2 &obj_hash[i].lock irq_context: 0 sk_lock-AF_INET rcu_read_lock rcu_read_lock rcu_read_lock_bh dev->qdisc_tx_busylock ?: &qdisc_tx_busylock _xmit_ETHER#2 pool_lock#2 irq_context: softirq rcu_read_lock rcu_read_lock slock-AF_INET/1 rcu_read_lock pool_lock#2 irq_context: 0 &u->iolock rcu_read_lock &ei->socket.wq.wait &p->pi_lock irq_context: 0 &u->iolock rcu_read_lock &ei->socket.wq.wait &p->pi_lock &rq->__lock irq_context: 0 &u->iolock rcu_read_lock &ei->socket.wq.wait &p->pi_lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 &u->iolock rcu_read_lock rcu_node_0 irq_context: 0 &u->iolock rcu_read_lock &rq->__lock irq_context: 0 sk_lock-AF_INET rcu_read_lock rcu_node_0 irq_context: 0 sk_lock-AF_INET rcu_read_lock &rq->__lock irq_context: 0 sk_lock-AF_INET rcu_read_lock rcu_read_lock &ul->lock irq_context: 0 sk_lock-AF_INET rcu_read_lock rcu_read_lock &obj_hash[i].lock irq_context: 0 sk_lock-AF_INET &mm->mmap_lock ptlock_ptr(page)#2 irq_context: softirq (&icsk->icsk_delack_timer) slock-AF_INET &c->lock irq_context: softirq (&icsk->icsk_delack_timer) slock-AF_INET rcu_read_lock rcu_read_lock rcu_read_lock_bh dev->qdisc_tx_busylock ?: &qdisc_tx_busylock _xmit_ETHER#2 rcu_read_lock &c->lock irq_context: softirq rcu_callback uidhash_lock irq_context: softirq rcu_callback ucounts_lock irq_context: 0 sk_lock-AF_INET rcu_read_lock rcu_read_lock rcu_read_lock_bh dev->qdisc_tx_busylock ?: &qdisc_tx_busylock _xmit_ETHER#2 rcu_read_lock &c->lock irq_context: softirq rcu_read_lock rcu_read_lock slock-AF_INET/1 rcu_read_lock &ei->socket.wq.wait &p->pi_lock &rq->__lock &cfs_rq->removed.lock irq_context: 0 dup_mmap_sem &mm->mmap_lock &mm->mmap_lock/1 fs_reclaim mmu_notifier_invalidate_range_start &rq->__lock irq_context: softirq rcu_read_lock rcu_read_lock &nf_conntrack_locks[i]/1 irq_context: softirq rcu_read_lock rcu_read_lock rcu_read_lock rcu_read_lock rcu_read_lock &____s->seqcount#10 irq_context: softirq rcu_read_lock rcu_read_lock slock-AF_INET/1 &c->lock irq_context: 0 &sb->s_type->i_mutex_key#10 sk_lock-AF_INET rcu_read_lock pool_lock#2 irq_context: 0 &sb->s_type->i_mutex_key#10 sk_lock-AF_INET &obj_hash[i].lock irq_context: 0 &sb->s_type->i_mutex_key#10 sk_lock-AF_INET pool_lock#2 irq_context: 0 &sb->s_type->i_mutex_key#10 sk_lock-AF_INET fs_reclaim irq_context: 0 &sb->s_type->i_mutex_key#10 sk_lock-AF_INET fs_reclaim mmu_notifier_invalidate_range_start irq_context: 0 &sb->s_type->i_mutex_key#10 sk_lock-AF_INET tk_core.seq.seqcount irq_context: 0 &sb->s_type->i_mutex_key#10 sk_lock-AF_INET rcu_read_lock rcu_read_lock &____s->seqcount#7 irq_context: 0 &sb->s_type->i_mutex_key#10 sk_lock-AF_INET rcu_read_lock rcu_read_lock &ct->lock irq_context: 0 &sb->s_type->i_mutex_key#10 sk_lock-AF_INET rcu_read_lock rcu_read_lock &____s->seqcount#10 irq_context: 0 &sb->s_type->i_mutex_key#10 sk_lock-AF_INET rcu_read_lock rcu_read_lock rcu_read_lock_bh dev->qdisc_tx_busylock ?: &qdisc_tx_busylock _xmit_ETHER#2 irq_context: 0 &sb->s_type->i_mutex_key#10 sk_lock-AF_INET rcu_read_lock rcu_read_lock rcu_read_lock_bh dev->qdisc_tx_busylock ?: &qdisc_tx_busylock _xmit_ETHER#2 rcu_read_lock pool_lock#2 irq_context: 0 &sb->s_type->i_mutex_key#10 sk_lock-AF_INET rcu_read_lock rcu_read_lock rcu_read_lock_bh dev->qdisc_tx_busylock ?: &qdisc_tx_busylock _xmit_ETHER#2 rcu_read_lock tk_core.seq.seqcount irq_context: 0 &sb->s_type->i_mutex_key#10 sk_lock-AF_INET rcu_read_lock rcu_read_lock rcu_read_lock_bh dev->qdisc_tx_busylock ?: &qdisc_tx_busylock _xmit_ETHER#2 rcu_read_lock &obj_hash[i].lock irq_context: 0 &sb->s_type->i_mutex_key#10 sk_lock-AF_INET &base->lock irq_context: 0 &sb->s_type->i_mutex_key#10 sk_lock-AF_INET &base->lock &obj_hash[i].lock irq_context: 0 &sb->s_type->i_mutex_key#10 sk_lock-AF_INET clock-AF_INET irq_context: 0 &sb->s_type->i_mutex_key#10 sk_lock-AF_INET &hashinfo->ehash_locks[i] irq_context: 0 &sb->s_type->i_mutex_key#10 sk_lock-AF_INET &tcp_hashinfo.bhash[i].lock irq_context: 0 &sb->s_type->i_mutex_key#10 sk_lock-AF_INET &tcp_hashinfo.bhash[i].lock &tcp_hashinfo.bhash2[i].lock irq_context: 0 &sb->s_type->i_mutex_key#10 sk_lock-AF_INET elock-AF_INET irq_context: 0 &u->iolock rcu_read_lock &ei->socket.wq.wait &p->pi_lock &cfs_rq->removed.lock irq_context: softirq rcu_read_lock rcu_read_lock slock-AF_INET/1 rcu_read_lock rcu_read_lock rcu_read_lock_bh dev->qdisc_tx_busylock ?: &qdisc_tx_busylock _xmit_ETHER#2 rcu_read_lock &n->list_lock irq_context: softirq rcu_read_lock rcu_read_lock slock-AF_INET/1 rcu_read_lock rcu_read_lock rcu_read_lock_bh dev->qdisc_tx_busylock ?: &qdisc_tx_busylock _xmit_ETHER#2 rcu_read_lock &n->list_lock &c->lock irq_context: 0 sk_lock-AF_INET &mm->mmap_lock fs_reclaim irq_context: 0 sk_lock-AF_INET &mm->mmap_lock fs_reclaim mmu_notifier_invalidate_range_start irq_context: 0 sk_lock-AF_INET &mm->mmap_lock &____s->seqcount irq_context: 0 sk_lock-AF_INET &mm->mmap_lock mmu_notifier_invalidate_range_start irq_context: 0 sk_lock-AF_INET &mm->mmap_lock ptlock_ptr(page)#2 lock#4 irq_context: 0 &pipe->wr_wait irq_context: 0 &sig->cred_guard_mutex tomoyo_ss remove_cache_srcu rcu_read_lock &____s->seqcount irq_context: 0 sb_writers#4 &type->i_mutex_dir_key#3 tomoyo_ss &n->list_lock irq_context: 0 sb_writers#4 &type->i_mutex_dir_key#3 tomoyo_ss &n->list_lock &c->lock irq_context: softirq rcu_read_lock rcu_read_lock slock-AF_INET/1 rcu_read_lock rcu_read_lock rcu_read_lock_bh dev->qdisc_tx_busylock ?: &qdisc_tx_busylock _xmit_ETHER#2 &obj_hash[i].lock irq_context: softirq rcu_read_lock rcu_read_lock slock-AF_INET/1 rcu_read_lock rcu_read_lock rcu_read_lock_bh dev->qdisc_tx_busylock ?: &qdisc_tx_busylock _xmit_ETHER#2 pool_lock#2 irq_context: softirq rcu_read_lock rcu_read_lock slock-AF_INET/1 &n->list_lock irq_context: softirq rcu_read_lock rcu_read_lock slock-AF_INET/1 &n->list_lock &c->lock irq_context: 0 &pipe->mutex/1 rcu_read_lock pool_lock#2 irq_context: 0 &pipe->wr_wait &p->pi_lock irq_context: 0 &pipe->wr_wait &p->pi_lock &rq->__lock irq_context: 0 &pipe->wr_wait &p->pi_lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 &pipe->wr_wait &p->pi_lock &cfs_rq->removed.lock irq_context: softirq slock-AF_INET tk_core.seq.seqcount irq_context: softirq slock-AF_INET rcu_read_lock rcu_read_lock &____s->seqcount#7 irq_context: softirq slock-AF_INET rcu_read_lock rcu_read_lock &ct->lock irq_context: softirq slock-AF_INET rcu_read_lock rcu_read_lock &____s->seqcount#10 irq_context: softirq slock-AF_INET rcu_read_lock rcu_read_lock rcu_read_lock_bh dev->qdisc_tx_busylock ?: &qdisc_tx_busylock _xmit_ETHER#2 irq_context: softirq slock-AF_INET rcu_read_lock rcu_read_lock rcu_read_lock_bh dev->qdisc_tx_busylock ?: &qdisc_tx_busylock _xmit_ETHER#2 rcu_read_lock pool_lock#2 irq_context: softirq slock-AF_INET rcu_read_lock rcu_read_lock rcu_read_lock_bh dev->qdisc_tx_busylock ?: &qdisc_tx_busylock _xmit_ETHER#2 rcu_read_lock tk_core.seq.seqcount irq_context: softirq slock-AF_INET rcu_read_lock rcu_read_lock rcu_read_lock_bh dev->qdisc_tx_busylock ?: &qdisc_tx_busylock _xmit_ETHER#2 rcu_read_lock &obj_hash[i].lock irq_context: softirq slock-AF_INET &obj_hash[i].lock irq_context: softirq slock-AF_INET &base->lock irq_context: softirq slock-AF_INET &base->lock &obj_hash[i].lock irq_context: 0 sb_writers#4 &sb->s_type->i_mutex_key#8 rcu_read_lock &xa->xa_lock#9 key#10 irq_context: 0 sb_writers#4 &sb->s_type->i_mutex_key#8 &xa->xa_lock#9 pool_lock#2 irq_context: 0 sb_writers#4 &sb->s_type->i_mutex_key#8 lock#4 &lruvec->lru_lock irq_context: 0 sb_writers#4 &sb->s_type->i_mutex_key#8 &ei->i_es_lock key#2 irq_context: 0 sb_writers#4 &sb->s_type->i_mutex_key#8 &pcp->lock &zone->lock irq_context: 0 sb_writers#4 &sb->s_type->i_mutex_key#8 &pcp->lock &zone->lock &____s->seqcount irq_context: 0 sb_writers#4 &sb->s_type->i_mutex_key#8 &ei->i_data_sem &(ei->i_block_reservation_lock) key#14 irq_context: softirq rcu_read_lock rcu_read_lock slock-AF_INET/1 rcu_read_lock rcu_read_lock rcu_read_lock_bh dev->qdisc_tx_busylock ?: &qdisc_tx_busylock _xmit_ETHER#2 rcu_read_lock &____s->seqcount irq_context: 0 sk_lock-AF_INET rcu_read_lock rcu_read_lock rcu_read_lock_bh dev->qdisc_tx_busylock ?: &qdisc_tx_busylock _xmit_ETHER#2 rcu_read_lock &____s->seqcount irq_context: 0 &vma->vm_lock->lock fs_reclaim mmu_notifier_invalidate_range_start &rq->__lock irq_context: 0 sb_writers#4 &sb->s_type->i_mutex_key#8 jbd2_handle &rq->__lock irq_context: 0 sb_writers#4 &sb->s_type->i_mutex_key#8 jbd2_handle &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 &vma->vm_lock->lock fs_reclaim &rq->__lock irq_context: 0 rtnl_mutex (inetaddr_chain).rwsem &rq->__lock irq_context: softirq drivers/regulator/core.c:6262 irq_context: softirq drivers/regulator/core.c:6262 rcu_read_lock &pool->lock irq_context: softirq drivers/regulator/core.c:6262 rcu_read_lock &pool->lock &obj_hash[i].lock irq_context: softirq drivers/regulator/core.c:6262 rcu_read_lock &pool->lock &p->pi_lock irq_context: softirq drivers/regulator/core.c:6262 rcu_read_lock &pool->lock &p->pi_lock &rq->__lock irq_context: softirq drivers/regulator/core.c:6262 rcu_read_lock &pool->lock &p->pi_lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)events (regulator_init_complete_work).work irq_context: 0 (wq_completion)events (regulator_init_complete_work).work &k->list_lock irq_context: 0 (wq_completion)events (regulator_init_complete_work).work &k->k_lock irq_context: 0 &vma->vm_lock->lock rcu_node_0 irq_context: 0 &vma->vm_lock->lock rcu_read_lock pgd_lock irq_context: 0 &vma->vm_lock->lock rcu_read_lock key irq_context: 0 &vma->vm_lock->lock rcu_read_lock pcpu_lock irq_context: 0 &vma->vm_lock->lock rcu_read_lock percpu_counters_lock irq_context: softirq rcu_read_lock rcu_read_lock slock-AF_INET/1 rcu_read_lock rcu_read_lock rcu_read_lock_bh dev->qdisc_tx_busylock ?: &qdisc_tx_busylock _xmit_ETHER#2 &____s->seqcount irq_context: softirq rcu_read_lock rcu_read_lock slock-AF_INET/1 &____s->seqcount irq_context: 0 &vma->vm_lock->lock &c->lock irq_context: 0 &vma->vm_lock->lock rcu_read_lock &rcu_state.gp_wq irq_context: 0 &vma->vm_lock->lock rcu_read_lock &rcu_state.gp_wq &p->pi_lock irq_context: 0 &vma->vm_lock->lock rcu_read_lock &rcu_state.gp_wq &p->pi_lock &rq->__lock irq_context: 0 &vma->vm_lock->lock rcu_read_lock &rcu_state.gp_wq &p->pi_lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 rtnl_mutex (inetaddr_chain).rwsem &net->sctp.local_addr_lock &net->sctp.addr_wq_lock &c->lock irq_context: 0 sb_writers#4 &sb->s_type->i_mutex_key#8 jbd2_handle &journal->j_state_lock irq_context: 0 sb_writers#4 &sb->s_type->i_mutex_key#8 jbd2_handle &journal->j_state_lock &journal->j_wait_commit irq_context: 0 sb_writers#4 &sb->s_type->i_mutex_key#8 jbd2_handle &journal->j_state_lock &journal->j_wait_commit &p->pi_lock irq_context: 0 sb_writers#4 &sb->s_type->i_mutex_key#8 jbd2_handle &journal->j_state_lock &journal->j_wait_commit &p->pi_lock &rq->__lock irq_context: 0 sb_writers#4 &sb->s_type->i_mutex_key#8 jbd2_handle &journal->j_state_lock &journal->j_wait_commit &p->pi_lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 &xa->xa_lock#9 pool_lock#2 irq_context: 0 rtnl_mutex rcu_read_lock &pool->lock &p->pi_lock &rq->__lock &cfs_rq->removed.lock irq_context: 0 lock#4 &lruvec->lru_lock irq_context: 0 &mm->mmap_lock &anon_vma->rwsem &cfs_rq->removed.lock irq_context: 0 sb_writers#4 &sb->s_type->i_mutex_key#8 &journal->j_state_lock irq_context: 0 sb_writers#4 &sb->s_type->i_mutex_key#8 &journal->j_state_lock tk_core.seq.seqcount irq_context: 0 sb_writers#4 &sb->s_type->i_mutex_key#8 &journal->j_state_lock &obj_hash[i].lock irq_context: 0 sb_writers#4 &sb->s_type->i_mutex_key#8 &journal->j_state_lock &base->lock irq_context: 0 sb_writers#4 &sb->s_type->i_mutex_key#8 &journal->j_state_lock &base->lock &obj_hash[i].lock irq_context: 0 sb_writers#4 &sb->s_type->i_mutex_key#8 jbd2_handle &mapping->private_lock irq_context: 0 sb_writers#4 &sb->s_type->i_mutex_key#8 jbd2_handle &ret->b_state_lock irq_context: 0 sb_writers#4 &sb->s_type->i_mutex_key#8 jbd2_handle &ret->b_state_lock &journal->j_list_lock irq_context: 0 dup_mmap_sem &mm->mmap_lock &mm->mmap_lock/1 rcu_read_lock &rcu_state.gp_wq &p->pi_lock &rq->__lock &cfs_rq->removed.lock irq_context: 0 &pipe->mutex/1 &pcp->lock &zone->lock &____s->seqcount irq_context: 0 sb_writers#4 &sb->s_type->i_mutex_key#8 &xa->xa_lock#9 &c->lock irq_context: 0 sb_writers#4 &sb->s_type->i_mutex_key#8 &xa->xa_lock#9 &n->list_lock irq_context: 0 sb_writers#4 &sb->s_type->i_mutex_key#8 &xa->xa_lock#9 &n->list_lock &c->lock irq_context: 0 sb_writers#4 &sb->s_type->i_mutex_key#8 &rq->__lock irq_context: 0 sb_writers#4 &sb->s_type->i_mutex_key#8 &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 &sig->cred_guard_mutex &sig->exec_update_lock &mm->mmap_lock &anon_vma->rwsem &meta->lock irq_context: 0 &sig->cred_guard_mutex &sig->exec_update_lock &mm->mmap_lock &anon_vma->rwsem kfence_freelist_lock irq_context: 0 &mm->mmap_lock ptlock_ptr(page)#2 rcu_read_lock per_cpu_ptr(&cgroup_rstat_cpu_lock, cpu) irq_context: 0 sk_lock-AF_INET6 &n->list_lock irq_context: 0 sk_lock-AF_INET6 &n->list_lock &c->lock irq_context: 0 &mm->mmap_lock &rcu_state.expedited_wq irq_context: 0 &mm->mmap_lock &rcu_state.expedited_wq &p->pi_lock irq_context: 0 &mm->mmap_lock &rcu_state.expedited_wq &p->pi_lock &rq->__lock irq_context: 0 &mm->mmap_lock &rcu_state.expedited_wq &p->pi_lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 &sb->s_type->i_mutex_key#10 &obj_hash[i].lock pool_lock irq_context: 0 &sig->cred_guard_mutex &p->alloc_lock &x->wait#25 &p->pi_lock &rq->__lock &cfs_rq->removed.lock irq_context: 0 sb_writers#4 &sb->s_type->i_mutex_key#8 batched_entropy_u8.lock irq_context: 0 sb_writers#4 &sb->s_type->i_mutex_key#8 kfence_freelist_lock irq_context: 0 sb_writers#4 &sb->s_type->i_mutex_key#8 &meta->lock irq_context: 0 &vma->vm_lock->lock batched_entropy_u8.lock irq_context: 0 &vma->vm_lock->lock kfence_freelist_lock irq_context: 0 sk_lock-AF_INET remove_cache_srcu irq_context: 0 sk_lock-AF_INET remove_cache_srcu quarantine_lock irq_context: 0 sk_lock-AF_INET remove_cache_srcu &c->lock irq_context: 0 sk_lock-AF_INET remove_cache_srcu &n->list_lock irq_context: 0 sk_lock-AF_INET remove_cache_srcu rcu_read_lock pool_lock#2 irq_context: 0 sk_lock-AF_INET remove_cache_srcu &obj_hash[i].lock irq_context: 0 sb_writers#4 &sb->s_type->i_mutex_key#8 &rq->__lock &cfs_rq->removed.lock irq_context: 0 sk_lock-AF_INET &n->list_lock irq_context: 0 sk_lock-AF_INET &n->list_lock &c->lock irq_context: 0 sk_lock-AF_INET rcu_read_lock rcu_read_lock rcu_read_lock_bh dev->qdisc_tx_busylock ?: &qdisc_tx_busylock _xmit_ETHER#2 rcu_read_lock &n->list_lock irq_context: 0 sk_lock-AF_INET rcu_read_lock rcu_read_lock rcu_read_lock_bh dev->qdisc_tx_busylock ?: &qdisc_tx_busylock _xmit_ETHER#2 rcu_read_lock &n->list_lock &c->lock irq_context: 0 &pipe->mutex/1 rcu_read_lock rcu_node_0 irq_context: 0 &pipe->mutex/1 rcu_read_lock &rq->__lock irq_context: 0 &pipe->mutex/1 &rq->__lock &cfs_rq->removed.lock irq_context: 0 &pipe->mutex/1 rcu_read_lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 sk_lock-AF_INET slock-AF_INET tk_core.seq.seqcount irq_context: softirq rcu_read_lock rcu_read_lock slock-AF_INET/1 rcu_read_lock rcu_read_lock rcu_read_lock_bh dev->qdisc_tx_busylock ?: &qdisc_tx_busylock _xmit_ETHER#2 rcu_read_lock rlock-AF_PACKET irq_context: 0 sb_writers#4 &sb->s_type->i_mutex_key#8 rcu_read_lock rcu_node_0 irq_context: 0 sb_writers#4 &sb->s_type->i_mutex_key#8 rcu_read_lock &rcu_state.expedited_wq irq_context: 0 sb_writers#4 &sb->s_type->i_mutex_key#8 rcu_read_lock &rcu_state.expedited_wq &p->pi_lock irq_context: 0 sb_writers#4 &sb->s_type->i_mutex_key#8 rcu_read_lock &rcu_state.expedited_wq &p->pi_lock &rq->__lock irq_context: 0 sb_writers#4 &sb->s_type->i_mutex_key#8 rcu_read_lock &rcu_state.expedited_wq &p->pi_lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 &sb->s_type->i_mutex_key#10 rcu_read_lock pool_lock#2 irq_context: 0 sb_writers#4 &sb->s_type->i_mutex_key#8 &ei->i_data_sem &rq->__lock irq_context: 0 sb_writers#4 &sb->s_type->i_mutex_key#8 rcu_read_lock &rq->__lock irq_context: 0 sk_lock-AF_INET &rq->__lock irq_context: 0 sk_lock-AF_INET batched_entropy_u8.lock irq_context: 0 sk_lock-AF_INET kfence_freelist_lock irq_context: 0 &sb->s_type->i_mutex_key#10 rcu_state.exp_mutex rcu_read_lock &rq->__lock irq_context: 0 &sb->s_type->i_mutex_key#10 rcu_state.exp_mutex rcu_read_lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 &mm->mmap_lock &anon_vma->rwsem rcu_read_lock rcu_node_0 irq_context: 0 &mm->mmap_lock &anon_vma->rwsem rcu_read_lock &rq->__lock irq_context: 0 &mm->mmap_lock &anon_vma->rwsem rcu_read_lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 &mm->mmap_lock batched_entropy_u8.lock crngs.lock irq_context: 0 &mm->mmap_lock batched_entropy_u8.lock crngs.lock base_crng.lock irq_context: 0 &pipe->mutex/1 rcu_read_lock &rcu_state.expedited_wq irq_context: 0 &pipe->mutex/1 rcu_read_lock &rcu_state.expedited_wq &p->pi_lock irq_context: softirq rcu_read_lock rcu_read_lock slock-AF_INET/1 rcu_read_lock rcu_read_lock rcu_read_lock_bh dev->qdisc_tx_busylock ?: &qdisc_tx_busylock _xmit_ETHER#2 rcu_read_lock &zone->lock irq_context: softirq rcu_read_lock rcu_read_lock slock-AF_INET/1 rcu_read_lock rcu_read_lock rcu_read_lock_bh dev->qdisc_tx_busylock ?: &qdisc_tx_busylock _xmit_ETHER#2 rcu_read_lock &zone->lock &____s->seqcount irq_context: 0 sk_lock-AF_INET batched_entropy_u8.lock crngs.lock irq_context: softirq rcu_read_lock rcu_read_lock slock-AF_INET/1 &meta->lock irq_context: softirq rcu_read_lock rcu_read_lock slock-AF_INET/1 kfence_freelist_lock irq_context: 0 sb_writers#4 &sb->s_type->i_mutex_key#8 mmu_notifier_invalidate_range_start &rq->__lock irq_context: 0 sk_lock-AF_INET &pcp->lock &zone->lock irq_context: 0 sk_lock-AF_INET &pcp->lock &zone->lock &____s->seqcount irq_context: 0 &pipe->mutex/1 rcu_read_lock rcu_read_lock rcu_node_0 irq_context: 0 &pipe->mutex/1 rcu_read_lock rcu_read_lock &rq->__lock irq_context: 0 sk_lock-AF_INET rcu_read_lock rcu_read_lock rcu_read_lock_bh dev->qdisc_tx_busylock ?: &qdisc_tx_busylock _xmit_ETHER#2 quarantine_lock irq_context: 0 sb_writers#4 &sb->s_type->i_mutex_key#8 remove_cache_srcu irq_context: 0 sb_writers#4 &sb->s_type->i_mutex_key#8 remove_cache_srcu quarantine_lock irq_context: 0 sb_writers#4 &sb->s_type->i_mutex_key#8 remove_cache_srcu &c->lock irq_context: 0 sb_writers#4 &sb->s_type->i_mutex_key#8 remove_cache_srcu &n->list_lock irq_context: 0 sb_writers#4 &sb->s_type->i_mutex_key#8 remove_cache_srcu rcu_read_lock pool_lock#2 irq_context: 0 sb_writers#4 &sb->s_type->i_mutex_key#8 remove_cache_srcu &obj_hash[i].lock irq_context: 0 sb_writers#4 &sb->s_type->i_mutex_key#8 lock#4 rcu_read_lock pool_lock#2 irq_context: 0 sb_writers#4 &sb->s_type->i_mutex_key#8 lock#4 &obj_hash[i].lock irq_context: 0 sk_lock-AF_INET rcu_read_lock rcu_read_lock rcu_node_0 irq_context: 0 sk_lock-AF_INET rcu_read_lock rcu_read_lock &rq->__lock irq_context: 0 sk_lock-AF_INET rcu_read_lock rcu_read_lock rcu_read_lock_bh dev->qdisc_tx_busylock ?: &qdisc_tx_busylock _xmit_ETHER#2 rcu_read_lock quarantine_lock irq_context: 0 &pipe->wr_wait &p->pi_lock &rq->__lock &cfs_rq->removed.lock irq_context: 0 sb_writers#4 &sb->s_type->i_mutex_key#8 rcu_read_lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 sb_writers#4 &sb->s_type->i_mutex_key#8 jbd2_handle rcu_read_lock rcu_node_0 irq_context: 0 sb_writers#4 &sb->s_type->i_mutex_key#8 jbd2_handle rcu_read_lock &rq->__lock irq_context: 0 &pipe->mutex/1 rcu_read_lock &rcu_state.expedited_wq &p->pi_lock &rq->__lock irq_context: 0 &pipe->mutex/1 rcu_read_lock &rcu_state.expedited_wq &p->pi_lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 sb_writers#4 &sb->s_type->i_mutex_key#8 &ei->i_data_sem &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: softirq (&lapb->t1timer) &lapb->lock &____s->seqcount irq_context: 0 sk_lock-AF_INET rcu_read_lock rcu_read_lock rcu_read_lock_bh dev->qdisc_tx_busylock ?: &qdisc_tx_busylock _xmit_ETHER#2 &meta->lock irq_context: 0 sk_lock-AF_INET rcu_read_lock rcu_read_lock rcu_read_lock_bh dev->qdisc_tx_busylock ?: &qdisc_tx_busylock _xmit_ETHER#2 kfence_freelist_lock irq_context: 0 sb_writers#4 tomoyo_ss irq_context: 0 sb_writers#4 tomoyo_ss mmu_notifier_invalidate_range_start irq_context: 0 sb_writers#4 tomoyo_ss rcu_read_lock mount_lock.seqcount irq_context: 0 sb_writers#4 tomoyo_ss rcu_read_lock rcu_read_lock rename_lock.seqcount irq_context: 0 sb_writers#4 tomoyo_ss &obj_hash[i].lock irq_context: 0 sb_writers#4 tomoyo_ss &c->lock irq_context: 0 sb_writers#4 tomoyo_ss rcu_read_lock &dentry->d_lock irq_context: 0 sb_writers#4 tomoyo_ss tomoyo_policy_lock irq_context: 0 sb_writers#4 &sb->s_type->i_mutex_key#8 mapping.invalidate_lock irq_context: 0 sb_writers#4 &sb->s_type->i_mutex_key#8 mapping.invalidate_lock &mapping->i_mmap_rwsem irq_context: 0 sb_writers#4 &sb->s_type->i_mutex_key#8 mapping.invalidate_lock mmu_notifier_invalidate_range_start irq_context: 0 sb_writers#4 &sb->s_type->i_mutex_key#8 mapping.invalidate_lock pool_lock#2 irq_context: 0 sb_writers#4 &sb->s_type->i_mutex_key#8 mapping.invalidate_lock &journal->j_state_lock irq_context: 0 sb_writers#4 &sb->s_type->i_mutex_key#8 mapping.invalidate_lock jbd2_handle irq_context: 0 sb_writers#4 &sb->s_type->i_mutex_key#8 mapping.invalidate_lock jbd2_handle mmu_notifier_invalidate_range_start irq_context: 0 sb_writers#4 &sb->s_type->i_mutex_key#8 mapping.invalidate_lock jbd2_handle pool_lock#2 irq_context: 0 sb_writers#4 &sb->s_type->i_mutex_key#8 mapping.invalidate_lock jbd2_handle &ret->b_state_lock irq_context: 0 sb_writers#4 &sb->s_type->i_mutex_key#8 mapping.invalidate_lock jbd2_handle &ret->b_state_lock &journal->j_list_lock irq_context: 0 sb_writers#4 &sb->s_type->i_mutex_key#8 mapping.invalidate_lock jbd2_handle &sbi->s_orphan_lock irq_context: 0 sb_writers#4 &sb->s_type->i_mutex_key#8 mapping.invalidate_lock jbd2_handle &ei->i_raw_lock irq_context: 0 sb_writers#4 &sb->s_type->i_mutex_key#8 mapping.invalidate_lock jbd2_handle &ei->i_data_sem irq_context: 0 sb_writers#4 &sb->s_type->i_mutex_key#8 mapping.invalidate_lock jbd2_handle &ei->i_data_sem &ei->i_prealloc_lock irq_context: 0 sb_writers#4 &sb->s_type->i_mutex_key#8 mapping.invalidate_lock jbd2_handle &ei->i_data_sem &ei->i_raw_lock irq_context: 0 sb_writers#4 &sb->s_type->i_mutex_key#8 mapping.invalidate_lock jbd2_handle &ei->i_data_sem &ei->i_es_lock irq_context: 0 sb_writers#4 &sb->s_type->i_mutex_key#8 mapping.invalidate_lock jbd2_handle &ei->i_data_sem &ei->i_es_lock &sbi->s_es_lock irq_context: 0 sb_writers#4 &sb->s_type->i_mutex_key#8 mapping.invalidate_lock jbd2_handle &ei->i_data_sem &ei->i_es_lock &obj_hash[i].lock irq_context: 0 sb_writers#4 &sb->s_type->i_mutex_key#8 mapping.invalidate_lock jbd2_handle &ei->i_data_sem &ei->i_es_lock pool_lock#2 irq_context: 0 sb_writers#4 &sb->s_type->i_mutex_key#8 mapping.invalidate_lock jbd2_handle &ei->i_data_sem mmu_notifier_invalidate_range_start irq_context: 0 sb_writers#4 &sb->s_type->i_mutex_key#8 mapping.invalidate_lock jbd2_handle &ei->i_data_sem &c->lock irq_context: 0 sb_writers#4 &sb->s_type->i_mutex_key#8 mapping.invalidate_lock jbd2_handle &ei->i_data_sem pool_lock#2 irq_context: 0 sb_writers#4 &sb->s_type->i_mutex_key#8 mapping.invalidate_lock jbd2_handle &ei->i_data_sem &obj_hash[i].lock irq_context: 0 sb_writers#4 &sb->s_type->i_mutex_key#8 mapping.invalidate_lock jbd2_handle tk_core.seq.seqcount irq_context: 0 sb_writers#4 &sb->s_type->i_mutex_key#8 mapping.invalidate_lock jbd2_handle &journal->j_wait_updates irq_context: 0 sb_writers#4 &sb->s_type->i_mutex_key#8 mapping.invalidate_lock &obj_hash[i].lock irq_context: softirq rcu_read_lock rcu_read_lock slock-AF_INET/1 &hashinfo->ehash_locks[i] irq_context: softirq rcu_read_lock rcu_read_lock slock-AF_INET/1 &tcp_hashinfo.bhash[i].lock irq_context: softirq rcu_read_lock rcu_read_lock slock-AF_INET/1 &tcp_hashinfo.bhash[i].lock &tcp_hashinfo.bhash2[i].lock irq_context: softirq rcu_read_lock rcu_read_lock slock-AF_INET/1 &tcp_hashinfo.bhash[i].lock &tcp_hashinfo.bhash2[i].lock &obj_hash[i].lock irq_context: softirq rcu_read_lock rcu_read_lock slock-AF_INET/1 &tcp_hashinfo.bhash[i].lock &tcp_hashinfo.bhash2[i].lock pool_lock#2 irq_context: softirq rcu_read_lock rcu_read_lock slock-AF_INET/1 elock-AF_INET irq_context: softirq rcu_read_lock rcu_read_lock rcu_read_lock batched_entropy_u32.lock crngs.lock irq_context: softirq rcu_read_lock rcu_read_lock slock-AF_INET batched_entropy_u16.lock crngs.lock irq_context: softirq rcu_read_lock rcu_read_lock slock-AF_INET rcu_read_lock &ei->socket.wq.wait &p->pi_lock &rq->__lock &cfs_rq->removed.lock irq_context: 0 &sig->cred_guard_mutex &sig->exec_update_lock &mm->mmap_lock lock#4 rcu_read_lock pool_lock#2 irq_context: 0 &sig->cred_guard_mutex &sig->exec_update_lock &mm->mmap_lock lock#4 &obj_hash[i].lock irq_context: softirq (&icsk->icsk_delack_timer) slock-AF_INET batched_entropy_u8.lock irq_context: softirq (&icsk->icsk_delack_timer) slock-AF_INET kfence_freelist_lock irq_context: softirq (&icsk->icsk_delack_timer) slock-AF_INET &____s->seqcount irq_context: 0 sk_lock-AF_INET &meta->lock irq_context: 0 sb_writers#4 &sb->s_type->i_mutex_key#8 mapping.invalidate_lock jbd2_handle &journal->j_list_lock irq_context: softirq (&rxnet->peer_keepalive_timer) rcu_read_lock &pool->lock/1 &p->pi_lock &cfs_rq->removed.lock irq_context: 0 rcu_read_lock rcu_read_lock per_cpu_ptr(&cgroup_rstat_cpu_lock, cpu) irq_context: 0 lock#5 irq_context: 0 rcu_read_lock &base->lock irq_context: 0 rcu_read_lock &base->lock &obj_hash[i].lock irq_context: softirq (&lapb->t1timer) &lapb->lock rcu_read_lock_bh dev->qdisc_tx_busylock ?: &qdisc_tx_busylock _xmit_ETHER#2 &meta->lock irq_context: softirq (&lapb->t1timer) &lapb->lock rcu_read_lock_bh dev->qdisc_tx_busylock ?: &qdisc_tx_busylock _xmit_ETHER#2 kfence_freelist_lock irq_context: softirq (&icsk->icsk_delack_timer) slock-AF_INET rcu_read_lock rcu_read_lock rcu_read_lock_bh dev->qdisc_tx_busylock ?: &qdisc_tx_busylock _xmit_ETHER#2 rcu_read_lock &____s->seqcount irq_context: 0 &sig->cred_guard_mutex &iint->mutex pgd_lock irq_context: 0 &sig->cred_guard_mutex &iint->mutex key irq_context: 0 &sig->cred_guard_mutex &iint->mutex pcpu_lock irq_context: 0 &sig->cred_guard_mutex &iint->mutex percpu_counters_lock irq_context: 0 &sig->cred_guard_mutex &iint->mutex batched_entropy_u8.lock irq_context: 0 &sig->cred_guard_mutex &iint->mutex kfence_freelist_lock irq_context: 0 kn->active#47 fs_reclaim irq_context: 0 kn->active#47 fs_reclaim mmu_notifier_invalidate_range_start irq_context: 0 kn->active#47 &kernfs_locks->open_file_mutex[count] irq_context: 0 kn->active#47 &kernfs_locks->open_file_mutex[count] fs_reclaim irq_context: 0 kn->active#47 &kernfs_locks->open_file_mutex[count] fs_reclaim mmu_notifier_invalidate_range_start irq_context: 0 &mm->mmap_lock khugepaged_mm_lock irq_context: 0 &mm->mmap_lock khugepaged_wait.lock irq_context: 0 &mm->mmap_lock khugepaged_wait.lock &p->pi_lock irq_context: 0 lock#3 &obj_hash[i].lock irq_context: 0 lock#3 rcu_read_lock &pool->lock irq_context: 0 lock#3 rcu_read_lock &pool->lock &obj_hash[i].lock irq_context: 0 lock#3 rcu_read_lock &pool->lock &p->pi_lock irq_context: 0 lock#3 rcu_read_lock &pool->lock &p->pi_lock &rq->__lock irq_context: 0 lock#3 rcu_read_lock &pool->lock &p->pi_lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)mm_percpu_wq (work_completion)(work) lock#4 rcu_read_lock pool_lock#2 irq_context: 0 (wq_completion)mm_percpu_wq (work_completion)(work) lock#4 &obj_hash[i].lock irq_context: 0 lock#3 (work_completion)(work) irq_context: 0 rcu_read_lock &sighand->siglock irq_context: 0 rcu_read_lock &sighand->siglock pool_lock#2 irq_context: 0 rcu_read_lock &sighand->siglock &p->pi_lock irq_context: 0 &futex_queues[i].lock irq_context: 0 rcu_read_lock &sighand->siglock &p->pi_lock &rq->__lock irq_context: 0 rcu_read_lock &sighand->siglock batched_entropy_u8.lock irq_context: 0 rcu_read_lock &sighand->siglock kfence_freelist_lock irq_context: 0 (wq_completion)events (shepherd).work cpu_hotplug_lock rcu_read_lock &pool->lock &p->pi_lock &rq->__lock &cfs_rq->removed.lock irq_context: 0 rcu_read_lock &sighand->siglock &c->lock irq_context: 0 &ep->mtx &ep->lock &ep->wq irq_context: 0 &ep->mtx &ep->lock &ep->wq &p->pi_lock irq_context: 0 &ep->mtx &ep->lock &ep->wq &p->pi_lock &rq->__lock irq_context: 0 &ep->mtx &ep->lock &ep->wq &p->pi_lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 &f->f_pos_lock &p->lock irq_context: 0 &f->f_pos_lock &p->lock fs_reclaim irq_context: 0 &f->f_pos_lock &p->lock fs_reclaim mmu_notifier_invalidate_range_start irq_context: 0 &f->f_pos_lock &p->lock cpufreq_driver_lock irq_context: 0 &f->f_pos_lock &p->lock &mm->mmap_lock irq_context: 0 &ep->mtx kn->active#4 fs_reclaim irq_context: 0 &ep->mtx kn->active#4 fs_reclaim mmu_notifier_invalidate_range_start irq_context: 0 &ep->mtx kn->active#4 pool_lock#2 irq_context: 0 &ep->mtx kn->active#4 &on->poll irq_context: 0 &ep->mtx &ep->lock &ep->wq &p->pi_lock &cfs_rq->removed.lock irq_context: 0 &ep->mtx &lock->wait_lock irq_context: 0 &ep->mtx &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 &f->f_pos_lock &p->lock &c->lock irq_context: 0 &f->f_pos_lock &p->lock &of->mutex irq_context: 0 &f->f_pos_lock &p->lock &of->mutex kn->active#4 param_lock irq_context: 0 &ep->mtx rcu_read_lock &on->poll irq_context: 0 kn->active#4 &kernfs_locks->open_file_mutex[count] &c->lock irq_context: 0 &ep->mtx &ep->lock &ep->wq &p->pi_lock &rq->__lock &cfs_rq->removed.lock irq_context: 0 &f->f_pos_lock &p->lock &n->list_lock irq_context: 0 &f->f_pos_lock &p->lock &n->list_lock &c->lock irq_context: 0 kn->active#4 &c->lock irq_context: 0 kn->active#4 &kernfs_locks->open_file_mutex[count] &n->list_lock irq_context: 0 kn->active#4 &kernfs_locks->open_file_mutex[count] &n->list_lock &c->lock irq_context: 0 &ep->mtx kn->active#4 &c->lock irq_context: 0 kn->active#4 &rq->__lock irq_context: 0 &f->f_pos_lock &p->lock &of->mutex kn->active#4 &rq->__lock irq_context: 0 kn->active#4 &n->list_lock irq_context: 0 kn->active#4 &n->list_lock &c->lock irq_context: 0 &ep->mtx kn->active#4 &rq->__lock irq_context: 0 kn->active#4 remove_cache_srcu irq_context: 0 kn->active#4 remove_cache_srcu quarantine_lock irq_context: 0 kn->active#4 remove_cache_srcu &c->lock irq_context: 0 kn->active#4 remove_cache_srcu &n->list_lock irq_context: 0 kn->active#4 remove_cache_srcu &obj_hash[i].lock irq_context: 0 &f->f_pos_lock &p->lock pool_lock#2 irq_context: 0 &f->f_pos_lock &p->lock module_mutex irq_context: 0 sk_lock-AF_INET rcu_read_lock &dir->lock#2 irq_context: 0 sk_lock-AF_INET rcu_read_lock &ul->lock irq_context: 0 sk_lock-AF_INET rcu_read_lock &obj_hash[i].lock irq_context: 0 sk_lock-AF_INET &____s->seqcount#8 irq_context: 0 sk_lock-AF_INET once_mutex irq_context: 0 sk_lock-AF_INET once_mutex crngs.lock irq_context: 0 sk_lock-AF_INET rcu_read_lock &pool->lock irq_context: 0 sk_lock-AF_INET rcu_read_lock &pool->lock &obj_hash[i].lock irq_context: 0 sk_lock-AF_INET rcu_read_lock &pool->lock pool_lock#2 irq_context: 0 sk_lock-AF_INET rcu_read_lock &pool->lock &p->pi_lock irq_context: 0 sk_lock-AF_INET rcu_read_lock &pool->lock &p->pi_lock &rq->__lock irq_context: 0 sk_lock-AF_INET rcu_read_lock &pool->lock &p->pi_lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 sk_lock-AF_INET &tcp_hashinfo.bhash[i].lock &tcp_hashinfo.bhash2[i].lock batched_entropy_u8.lock irq_context: 0 sk_lock-AF_INET &tcp_hashinfo.bhash[i].lock &tcp_hashinfo.bhash2[i].lock &hashinfo->ehash_locks[i] irq_context: 0 sk_lock-AF_INET batched_entropy_u32.lock irq_context: 0 sk_lock-AF_INET batched_entropy_u16.lock irq_context: 0 sk_lock-AF_INET rcu_read_lock rcu_read_lock &nf_conntrack_locks[i] irq_context: 0 sk_lock-AF_INET rcu_read_lock rcu_read_lock &nf_conntrack_locks[i] &nf_conntrack_locks[i]/1 irq_context: 0 sk_lock-AF_INET rcu_read_lock rcu_read_lock &nf_conntrack_locks[i] &nf_conntrack_locks[i]/1 batched_entropy_u8.lock irq_context: softirq rcu_read_lock rcu_read_lock slock-AF_INET/1 rcu_read_lock &ei->socket.wq.wait &ep->lock irq_context: softirq rcu_read_lock rcu_read_lock slock-AF_INET/1 rcu_read_lock &ei->socket.wq.wait &ep->lock &ep->wq irq_context: softirq rcu_read_lock rcu_read_lock slock-AF_INET/1 rcu_read_lock &ei->socket.wq.wait &ep->lock &ep->wq &p->pi_lock irq_context: softirq rcu_read_lock rcu_read_lock slock-AF_INET/1 rcu_read_lock &ei->socket.wq.wait &ep->lock &ep->wq &p->pi_lock &rq->__lock irq_context: softirq rcu_read_lock rcu_read_lock slock-AF_INET/1 rcu_read_lock &ei->socket.wq.wait &ep->lock &ep->wq &p->pi_lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 rcu_read_lock &sighand->siglock &____s->seqcount irq_context: 0 &ep->mtx &pipe->wr_wait irq_context: 0 rcu_read_lock tasklist_lock irq_context: 0 &ep->mtx rcu_read_lock &pipe->wr_wait irq_context: 0 &sb->s_type->i_lock_key#7 irq_context: 0 sb_writers#9 irq_context: 0 sb_writers#9 &attr->mutex irq_context: 0 sb_writers#9 &attr->mutex &mm->mmap_lock irq_context: 0 sb_writers#3 &p->pi_lock irq_context: 0 sb_writers#3 &p->pi_lock &rq->__lock irq_context: 0 sb_writers#3 &p->pi_lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 sb_writers#3 &rq->__lock irq_context: 0 sb_writers#3 &sb->s_type->i_mutex_key#9 &n->list_lock irq_context: 0 sb_writers#3 &sb->s_type->i_mutex_key#9 &n->list_lock &c->lock irq_context: 0 sb_writers#4 &type->i_mutex_dir_key#3/1 tomoyo_ss irq_context: 0 sb_writers#4 &type->i_mutex_dir_key#3/1 tomoyo_ss mmu_notifier_invalidate_range_start irq_context: 0 sb_writers#4 &type->i_mutex_dir_key#3/1 tomoyo_ss &c->lock irq_context: 0 sb_writers#4 &type->i_mutex_dir_key#3/1 tomoyo_ss &____s->seqcount irq_context: 0 sb_writers#4 &type->i_mutex_dir_key#3/1 tomoyo_ss pool_lock#2 irq_context: 0 sb_writers#4 &type->i_mutex_dir_key#3/1 tomoyo_ss rcu_read_lock mount_lock.seqcount irq_context: 0 sb_writers#4 &type->i_mutex_dir_key#3/1 tomoyo_ss rcu_read_lock rcu_read_lock rename_lock.seqcount irq_context: 0 sb_writers#4 &type->i_mutex_dir_key#3/1 tomoyo_ss &obj_hash[i].lock irq_context: 0 sb_writers#4 &type->i_mutex_dir_key#3/1 tomoyo_ss rcu_read_lock &dentry->d_lock irq_context: 0 sb_writers#4 &type->i_mutex_dir_key#3/1 tomoyo_ss tomoyo_policy_lock irq_context: 0 sb_writers#4 &type->i_mutex_dir_key#3/1 tomoyo_ss tomoyo_policy_lock mmu_notifier_invalidate_range_start irq_context: 0 sb_writers#4 &type->i_mutex_dir_key#3/1 tomoyo_ss tomoyo_policy_lock pool_lock#2 irq_context: 0 sb_writers#4 &type->i_mutex_dir_key#3/1 &s->s_inode_list_lock irq_context: 0 sb_writers#4 &type->i_mutex_dir_key#3/1 &ei->xattr_sem irq_context: 0 sb_writers#4 &type->i_mutex_dir_key#3/1 &ei->xattr_sem &mapping->private_lock irq_context: 0 sb_writers#4 &type->i_mutex_dir_key#3/1 jbd2_handle irq_context: 0 sb_writers#4 &type->i_mutex_dir_key#3/1 jbd2_handle mmu_notifier_invalidate_range_start irq_context: 0 sb_writers#4 &type->i_mutex_dir_key#3/1 jbd2_handle pool_lock#2 irq_context: 0 sb_writers#4 &type->i_mutex_dir_key#3/1 jbd2_handle &ret->b_state_lock irq_context: 0 sb_writers#4 &type->i_mutex_dir_key#3/1 jbd2_handle &ret->b_state_lock &journal->j_list_lock irq_context: 0 sb_writers#4 &type->i_mutex_dir_key#3/1 jbd2_handle &journal->j_revoke_lock irq_context: 0 sb_writers#4 &type->i_mutex_dir_key#3/1 jbd2_handle &meta_group_info[i]->alloc_sem irq_context: 0 sb_writers#4 &type->i_mutex_dir_key#3/1 jbd2_handle tk_core.seq.seqcount irq_context: 0 sb_writers#4 &type->i_mutex_dir_key#3/1 jbd2_handle inode_hash_lock irq_context: 0 sb_writers#4 &type->i_mutex_dir_key#3/1 jbd2_handle inode_hash_lock &sb->s_type->i_lock_key#22 irq_context: 0 sb_writers#4 &type->i_mutex_dir_key#3/1 jbd2_handle batched_entropy_u32.lock irq_context: 0 sb_writers#4 &type->i_mutex_dir_key#3/1 jbd2_handle &____s->seqcount irq_context: 0 sb_writers#4 &type->i_mutex_dir_key#3/1 jbd2_handle &xa->xa_lock#9 irq_context: 0 sb_writers#4 &type->i_mutex_dir_key#3/1 jbd2_handle &xa->xa_lock#9 pool_lock#2 irq_context: 0 sb_writers#4 &type->i_mutex_dir_key#3/1 jbd2_handle lock#4 irq_context: 0 sb_writers#4 &type->i_mutex_dir_key#3/1 jbd2_handle &mapping->private_lock irq_context: 0 sb_writers#4 &type->i_mutex_dir_key#3/1 jbd2_handle &ei->i_raw_lock irq_context: 0 sb_writers#4 &type->i_mutex_dir_key#3/1 jbd2_handle &ei->i_es_lock irq_context: 0 sb_writers#4 &type->i_mutex_dir_key#3/1 jbd2_handle &ei->i_data_sem irq_context: 0 sb_writers#4 &type->i_mutex_dir_key#3/1 jbd2_handle &ei->i_data_sem mmu_notifier_invalidate_range_start irq_context: 0 sb_writers#4 &type->i_mutex_dir_key#3/1 jbd2_handle &ei->i_data_sem pool_lock#2 irq_context: 0 sb_writers#4 &type->i_mutex_dir_key#3/1 jbd2_handle &ei->i_data_sem &ei->i_es_lock irq_context: 0 sb_writers#4 &type->i_mutex_dir_key#3/1 jbd2_handle &ei->i_data_sem &ei->i_es_lock irq_context: 0 sb_writers#4 &type->i_mutex_dir_key#3/1 jbd2_handle &ei->i_data_sem &ei->i_es_lock pool_lock#2 irq_context: 0 sb_writers#4 &type->i_mutex_dir_key#3/1 jbd2_handle &ei->i_data_sem &ei->i_es_lock &sbi->s_es_lock irq_context: 0 sb_writers#4 &type->i_mutex_dir_key#3/1 jbd2_handle &ei->i_data_sem &obj_hash[i].lock irq_context: 0 sb_writers#4 &type->i_mutex_dir_key#3/1 jbd2_handle &ei->i_data_sem irq_context: 0 sb_writers#4 &type->i_mutex_dir_key#3/1 jbd2_handle &ei->i_data_sem mmu_notifier_invalidate_range_start irq_context: 0 sb_writers#4 &type->i_mutex_dir_key#3/1 jbd2_handle &ei->i_data_sem pool_lock#2 irq_context: 0 sb_writers#4 &type->i_mutex_dir_key#3/1 jbd2_handle &ei->i_data_sem &sb->s_type->i_lock_key#22 irq_context: 0 sb_writers#4 &type->i_mutex_dir_key#3/1 jbd2_handle &ei->i_data_sem &ei->i_raw_lock irq_context: 0 sb_writers#4 &type->i_mutex_dir_key#3/1 jbd2_handle &ei->i_data_sem &wb->list_lock irq_context: 0 sb_writers#4 &type->i_mutex_dir_key#3/1 jbd2_handle &ei->i_data_sem &wb->list_lock &sb->s_type->i_lock_key#22 irq_context: 0 sb_writers#4 &type->i_mutex_dir_key#3/1 jbd2_handle &ei->i_data_sem &____s->seqcount irq_context: 0 sb_writers#4 &type->i_mutex_dir_key#3/1 jbd2_handle &ei->i_data_sem &c->lock irq_context: 0 sb_writers#4 &type->i_mutex_dir_key#3/1 jbd2_handle &ei->i_data_sem &obj_hash[i].lock irq_context: 0 sb_writers#4 &type->i_mutex_dir_key#3/1 jbd2_handle &ei->i_data_sem &mapping->private_lock irq_context: 0 sb_writers#4 &type->i_mutex_dir_key#3/1 jbd2_handle &ei->i_data_sem &ret->b_state_lock irq_context: 0 sb_writers#4 &type->i_mutex_dir_key#3/1 jbd2_handle &ei->i_data_sem &ret->b_state_lock &journal->j_list_lock irq_context: 0 sb_writers#4 &type->i_mutex_dir_key#3/1 jbd2_handle &ei->i_data_sem &journal->j_revoke_lock irq_context: 0 sb_writers#4 &type->i_mutex_dir_key#3/1 jbd2_handle &ei->i_data_sem &ei->i_es_lock irq_context: 0 sb_writers#4 &type->i_mutex_dir_key#3/1 jbd2_handle &ei->i_data_sem &ei->i_es_lock irq_context: 0 sb_writers#4 &type->i_mutex_dir_key#3/1 jbd2_handle &ei->i_data_sem &ei->i_es_lock pool_lock#2 irq_context: 0 sb_writers#4 &type->i_mutex_dir_key#3/1 jbd2_handle &sb->s_type->i_lock_key#22 irq_context: 0 sb_writers#4 &type->i_mutex_dir_key#3/1 jbd2_handle &sb->s_type->i_lock_key#22 &dentry->d_lock irq_context: 0 sb_writers#4 &type->i_mutex_dir_key#3/1 jbd2_handle &journal->j_wait_updates irq_context: 0 &type->s_umount_key#44/1 irq_context: 0 &type->s_umount_key#44/1 fs_reclaim irq_context: 0 &type->s_umount_key#44/1 fs_reclaim mmu_notifier_invalidate_range_start irq_context: 0 &type->s_umount_key#44/1 pool_lock#2 irq_context: 0 &type->s_umount_key#44/1 pcpu_alloc_mutex irq_context: 0 &type->s_umount_key#44/1 pcpu_alloc_mutex pcpu_lock irq_context: 0 &type->s_umount_key#44/1 shrinker_rwsem irq_context: 0 &type->s_umount_key#44/1 list_lrus_mutex irq_context: 0 &type->s_umount_key#44/1 &c->lock irq_context: 0 &type->s_umount_key#44/1 sb_lock irq_context: 0 &type->s_umount_key#44/1 sb_lock unnamed_dev_ida.xa_lock irq_context: 0 &type->s_umount_key#44/1 &root->kernfs_rwsem irq_context: 0 &type->s_umount_key#44/1 &root->kernfs_rwsem inode_hash_lock irq_context: 0 &type->s_umount_key#44/1 &root->kernfs_rwsem fs_reclaim irq_context: 0 &type->s_umount_key#44/1 &root->kernfs_rwsem fs_reclaim mmu_notifier_invalidate_range_start irq_context: 0 &type->s_umount_key#44/1 &root->kernfs_rwsem &c->lock irq_context: 0 &type->s_umount_key#44/1 &root->kernfs_rwsem pool_lock#2 irq_context: 0 &type->s_umount_key#44/1 &root->kernfs_rwsem mmu_notifier_invalidate_range_start irq_context: 0 &type->s_umount_key#44/1 &root->kernfs_rwsem inode_hash_lock &sb->s_type->i_lock_key#30 irq_context: 0 &type->s_umount_key#44/1 &root->kernfs_rwsem inode_hash_lock &s->s_inode_list_lock irq_context: 0 &type->s_umount_key#44/1 &root->kernfs_rwsem tk_core.seq.seqcount irq_context: 0 &type->s_umount_key#44/1 &root->kernfs_rwsem &sb->s_type->i_lock_key#30 irq_context: 0 &type->s_umount_key#44/1 &sb->s_type->i_lock_key#30 irq_context: 0 &type->s_umount_key#44/1 &sb->s_type->i_lock_key#30 &dentry->d_lock irq_context: 0 &type->s_umount_key#44/1 &root->kernfs_supers_rwsem irq_context: 0 &type->s_umount_key#44/1 &dentry->d_lock irq_context: 0 sb_writers#10 irq_context: 0 sb_writers#10 mount_lock irq_context: 0 sb_writers#10 &type->i_mutex_dir_key#6 irq_context: 0 sb_writers#10 &type->i_mutex_dir_key#6 tomoyo_ss irq_context: 0 sb_writers#10 &type->i_mutex_dir_key#6 tomoyo_ss mmu_notifier_invalidate_range_start irq_context: 0 sb_writers#10 &type->i_mutex_dir_key#6 tomoyo_ss rcu_read_lock mount_lock.seqcount irq_context: 0 sb_writers#10 &type->i_mutex_dir_key#6 tomoyo_ss rcu_read_lock rcu_read_lock rename_lock.seqcount irq_context: 0 sb_writers#10 &type->i_mutex_dir_key#6 tomoyo_ss &obj_hash[i].lock irq_context: 0 sb_writers#10 &type->i_mutex_dir_key#6 tomoyo_ss &c->lock irq_context: 0 sb_writers#10 &type->i_mutex_dir_key#6 tomoyo_ss &____s->seqcount irq_context: 0 sb_writers#10 &type->i_mutex_dir_key#6 tomoyo_ss rcu_read_lock &dentry->d_lock irq_context: 0 sb_writers#10 &type->i_mutex_dir_key#6 tomoyo_ss tomoyo_policy_lock irq_context: 0 sb_writers#10 &type->i_mutex_dir_key#6 tk_core.seq.seqcount irq_context: 0 sb_writers#10 &type->i_mutex_dir_key#6 &root->kernfs_iattr_rwsem irq_context: 0 sb_writers#10 &type->i_mutex_dir_key#6 &root->kernfs_iattr_rwsem iattr_mutex irq_context: 0 sb_writers#10 &type->i_mutex_dir_key#6 &root->kernfs_iattr_rwsem iattr_mutex fs_reclaim irq_context: 0 sb_writers#10 &type->i_mutex_dir_key#6 &root->kernfs_iattr_rwsem iattr_mutex fs_reclaim mmu_notifier_invalidate_range_start irq_context: 0 sb_writers#10 &type->i_mutex_dir_key#6 &root->kernfs_iattr_rwsem iattr_mutex pool_lock#2 irq_context: 0 sb_writers#10 &type->i_mutex_dir_key#6 &root->kernfs_iattr_rwsem iattr_mutex tk_core.seq.seqcount irq_context: 0 sb_writers#10 &type->i_mutex_dir_key#6 irq_context: 0 sb_writers#10 &type->i_mutex_dir_key#6 rename_lock.seqcount irq_context: 0 sb_writers#10 &type->i_mutex_dir_key#6 fs_reclaim irq_context: 0 sb_writers#10 &type->i_mutex_dir_key#6 fs_reclaim mmu_notifier_invalidate_range_start irq_context: 0 sb_writers#10 &type->i_mutex_dir_key#6 &dentry->d_lock irq_context: 0 sb_writers#10 &type->i_mutex_dir_key#6 rcu_read_lock rename_lock.seqcount irq_context: 0 sb_writers#10 &type->i_mutex_dir_key#6 &root->kernfs_rwsem irq_context: 0 sb_writers#10 &type->i_mutex_dir_key#6 &root->kernfs_rwsem inode_hash_lock irq_context: 0 sb_writers#10 &type->i_mutex_dir_key#6 &root->kernfs_rwsem fs_reclaim irq_context: 0 sb_writers#10 &type->i_mutex_dir_key#6 &root->kernfs_rwsem fs_reclaim mmu_notifier_invalidate_range_start irq_context: 0 sb_writers#10 &type->i_mutex_dir_key#6 &root->kernfs_rwsem &c->lock irq_context: 0 sb_writers#10 &type->i_mutex_dir_key#6 &root->kernfs_rwsem &____s->seqcount irq_context: 0 sb_writers#10 &type->i_mutex_dir_key#6 &root->kernfs_rwsem mmu_notifier_invalidate_range_start irq_context: 0 sb_writers#10 &type->i_mutex_dir_key#6 &root->kernfs_rwsem inode_hash_lock &sb->s_type->i_lock_key#30 irq_context: 0 sb_writers#10 &type->i_mutex_dir_key#6 &root->kernfs_rwsem inode_hash_lock &s->s_inode_list_lock irq_context: 0 sb_writers#10 &type->i_mutex_dir_key#6 &root->kernfs_rwsem tk_core.seq.seqcount irq_context: 0 sb_writers#10 &type->i_mutex_dir_key#6 &root->kernfs_rwsem &sb->s_type->i_lock_key#30 irq_context: 0 sb_writers#10 &type->i_mutex_dir_key#6 &sb->s_type->i_lock_key#30 irq_context: 0 sb_writers#10 &type->i_mutex_dir_key#6 &sb->s_type->i_lock_key#30 &dentry->d_lock irq_context: 0 sb_writers#10 &type->i_mutex_dir_key#6 &sb->s_type->i_lock_key#30 &dentry->d_lock &wq#2 irq_context: 0 kn->active#48 fs_reclaim irq_context: 0 kn->active#48 fs_reclaim mmu_notifier_invalidate_range_start irq_context: 0 kn->active#48 &kernfs_locks->open_file_mutex[count] irq_context: 0 kn->active#48 &kernfs_locks->open_file_mutex[count] fs_reclaim irq_context: 0 kn->active#48 &kernfs_locks->open_file_mutex[count] fs_reclaim mmu_notifier_invalidate_range_start irq_context: 0 kn->active#48 pool_lock#2 irq_context: 0 &sb->s_type->i_mutex_key#15 irq_context: 0 sb_writers#10 fs_reclaim irq_context: 0 sb_writers#10 fs_reclaim mmu_notifier_invalidate_range_start irq_context: 0 sb_writers#10 &mm->mmap_lock irq_context: 0 sb_writers#10 &of->mutex irq_context: 0 sb_writers#10 &of->mutex cgroup_mutex irq_context: 0 sb_writers#10 &of->mutex cgroup_mutex css_set_lock irq_context: 0 sb_writers#10 &of->mutex cgroup_mutex cpu_hotplug_lock irq_context: 0 sb_writers#10 &of->mutex cgroup_mutex cpu_hotplug_lock css_set_lock irq_context: 0 sb_writers#10 &of->mutex cgroup_mutex &root->kernfs_rwsem irq_context: 0 sb_writers#10 &obj_hash[i].lock irq_context: 0 cgroup_mutex fs_reclaim irq_context: 0 cgroup_mutex fs_reclaim mmu_notifier_invalidate_range_start irq_context: 0 cgroup_mutex cpu_hotplug_lock css_set_lock irq_context: 0 cgroup_mutex cpu_hotplug_lock jump_label_mutex irq_context: 0 cgroup_mutex css_set_lock cgroup_file_kn_lock irq_context: 0 &type->s_umount_key#45/1 irq_context: 0 &type->s_umount_key#45/1 fs_reclaim irq_context: 0 &type->s_umount_key#45/1 fs_reclaim mmu_notifier_invalidate_range_start irq_context: 0 &type->s_umount_key#45/1 pool_lock#2 irq_context: 0 &type->s_umount_key#45/1 pcpu_alloc_mutex irq_context: 0 &type->s_umount_key#45/1 pcpu_alloc_mutex pcpu_lock irq_context: 0 &type->s_umount_key#45/1 shrinker_rwsem irq_context: 0 &type->s_umount_key#45/1 &____s->seqcount irq_context: 0 &type->s_umount_key#45/1 rcu_read_lock pool_lock#2 irq_context: 0 &type->s_umount_key#45/1 &obj_hash[i].lock irq_context: 0 &type->s_umount_key#45/1 list_lrus_mutex irq_context: 0 &type->s_umount_key#45/1 sb_lock irq_context: 0 &type->s_umount_key#45/1 sb_lock unnamed_dev_ida.xa_lock irq_context: 0 &type->s_umount_key#45/1 &root->kernfs_rwsem irq_context: 0 &type->s_umount_key#45/1 &root->kernfs_rwsem inode_hash_lock irq_context: 0 &type->s_umount_key#45/1 &root->kernfs_rwsem fs_reclaim irq_context: 0 &type->s_umount_key#45/1 &root->kernfs_rwsem fs_reclaim mmu_notifier_invalidate_range_start irq_context: 0 &type->s_umount_key#45/1 &root->kernfs_rwsem pool_lock#2 irq_context: 0 &type->s_umount_key#45/1 &root->kernfs_rwsem mmu_notifier_invalidate_range_start irq_context: 0 &type->s_umount_key#45/1 &root->kernfs_rwsem inode_hash_lock &sb->s_type->i_lock_key#31 irq_context: 0 &type->s_umount_key#45/1 &root->kernfs_rwsem inode_hash_lock &s->s_inode_list_lock irq_context: 0 &type->s_umount_key#45/1 &root->kernfs_rwsem tk_core.seq.seqcount irq_context: 0 &type->s_umount_key#45/1 &root->kernfs_rwsem &sb->s_type->i_lock_key#31 irq_context: 0 &type->s_umount_key#45/1 &sb->s_type->i_lock_key#31 irq_context: 0 &type->s_umount_key#45/1 &sb->s_type->i_lock_key#31 &dentry->d_lock irq_context: 0 &type->s_umount_key#45/1 &root->kernfs_supers_rwsem irq_context: 0 &type->s_umount_key#45/1 &dentry->d_lock irq_context: 0 &type->s_umount_key#46 irq_context: 0 &type->s_umount_key#46 shrinker_rwsem irq_context: 0 &type->s_umount_key#46 percpu_ref_switch_lock irq_context: 0 &type->s_umount_key#46 percpu_ref_switch_lock &obj_hash[i].lock irq_context: 0 &type->s_umount_key#46 percpu_ref_switch_lock pool_lock#2 irq_context: 0 &type->s_umount_key#46 &root->kernfs_supers_rwsem irq_context: 0 &type->s_umount_key#46 rename_lock.seqcount irq_context: 0 &type->s_umount_key#46 &dentry->d_lock irq_context: 0 &type->s_umount_key#46 rcu_read_lock &dentry->d_lock irq_context: 0 &type->s_umount_key#46 &sb->s_type->i_lock_key#31 irq_context: 0 &type->s_umount_key#46 &s->s_inode_list_lock irq_context: 0 &type->s_umount_key#46 &xa->xa_lock#9 irq_context: 0 &type->s_umount_key#46 inode_hash_lock irq_context: 0 &type->s_umount_key#46 inode_hash_lock &sb->s_type->i_lock_key#31 irq_context: 0 &type->s_umount_key#46 &obj_hash[i].lock irq_context: 0 &type->s_umount_key#46 pool_lock#2 irq_context: 0 &type->s_umount_key#46 &fsnotify_mark_srcu irq_context: 0 cgroup_mutex &n->list_lock &c->lock irq_context: 0 cgroup_mutex &root->kernfs_rwsem &root->kernfs_iattr_rwsem tk_core.seq.seqcount irq_context: 0 cgroup_mutex &root->kernfs_rwsem &obj_hash[i].lock irq_context: 0 cgroup_mutex &root->kernfs_rwsem pool_lock#2 irq_context: 0 cgroup_mutex &root->kernfs_rwsem kernfs_idr_lock irq_context: 0 cgroup_mutex rcu_state.exp_mutex rcu_node_0 irq_context: 0 cgroup_mutex rcu_state.exp_mutex &obj_hash[i].lock irq_context: 0 cgroup_mutex rcu_state.exp_mutex rcu_read_lock &pool->lock irq_context: 0 cgroup_mutex rcu_state.exp_mutex rcu_read_lock &pool->lock &obj_hash[i].lock irq_context: 0 cgroup_mutex rcu_state.exp_mutex rcu_read_lock &pool->lock &p->pi_lock irq_context: 0 cgroup_mutex rcu_state.exp_mutex rcu_read_lock &pool->lock &p->pi_lock &rq->__lock irq_context: 0 cgroup_mutex rcu_state.exp_mutex rcu_read_lock &pool->lock &p->pi_lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 cgroup_mutex rcu_state.exp_mutex &rq->__lock irq_context: 0 cgroup_mutex cpu_hotplug_lock jump_label_mutex text_mutex irq_context: 0 cgroup_mutex cpu_hotplug_lock jump_label_mutex text_mutex ptlock_ptr(page)#2 irq_context: 0 &type->s_umount_key#45/1 &c->lock irq_context: softirq rcu_callback rcu_read_lock rcu_read_lock &pool->lock irq_context: softirq rcu_callback rcu_read_lock rcu_read_lock &pool->lock &obj_hash[i].lock irq_context: softirq rcu_callback rcu_read_lock rcu_read_lock &pool->lock pool_lock#2 irq_context: softirq rcu_callback rcu_read_lock rcu_read_lock &pool->lock &p->pi_lock irq_context: softirq rcu_callback rcu_read_lock rcu_read_lock &pool->lock &p->pi_lock &rq->__lock irq_context: softirq rcu_callback rcu_read_lock rcu_read_lock &pool->lock &p->pi_lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)events (work_completion)(&cgrp->bpf.release_work) irq_context: 0 (wq_completion)events (work_completion)(&cgrp->bpf.release_work) cgroup_mutex irq_context: 0 (wq_completion)events (work_completion)(&cgrp->bpf.release_work) percpu_ref_switch_lock irq_context: 0 (wq_completion)events (work_completion)(&cgrp->bpf.release_work) &obj_hash[i].lock irq_context: 0 (wq_completion)events (work_completion)(&cgrp->bpf.release_work) pool_lock#2 irq_context: 0 (wq_completion)events (work_completion)(&cgrp->bpf.release_work) rcu_read_lock &obj_hash[i].lock irq_context: 0 (wq_completion)events (work_completion)(&cgrp->bpf.release_work) rcu_read_lock rcu_read_lock &pool->lock irq_context: 0 (wq_completion)events (work_completion)(&cgrp->bpf.release_work) rcu_read_lock rcu_read_lock &pool->lock &obj_hash[i].lock irq_context: 0 (wq_completion)events (work_completion)(&cgrp->bpf.release_work) rcu_read_lock rcu_read_lock &pool->lock pool_lock#2 irq_context: 0 (wq_completion)cgroup_destroy irq_context: 0 (wq_completion)cgroup_destroy (work_completion)(&css->destroy_work) irq_context: 0 (wq_completion)cgroup_destroy (work_completion)(&css->destroy_work) cgroup_mutex irq_context: 0 (wq_completion)cgroup_destroy (work_completion)(&css->destroy_work) cgroup_mutex cgroup_rstat_lock irq_context: 0 (wq_completion)cgroup_destroy (work_completion)(&css->destroy_work) cgroup_mutex cgroup_rstat_lock per_cpu_ptr(&cgroup_rstat_cpu_lock, cpu) irq_context: 0 (wq_completion)cgroup_destroy (work_completion)(&css->destroy_work) cgroup_mutex css_set_lock irq_context: 0 (wq_completion)cgroup_destroy (work_completion)(&css->destroy_work) &obj_hash[i].lock irq_context: 0 (wq_completion)cgroup_destroy (work_completion)(&css->destroy_work) pool_lock#2 irq_context: 0 (wq_completion)cgroup_destroy (work_completion)(&(&css->destroy_rwork)->work) irq_context: 0 (wq_completion)cgroup_destroy (work_completion)(&(&css->destroy_rwork)->work) percpu_ref_switch_lock irq_context: 0 (wq_completion)cgroup_destroy (work_completion)(&(&css->destroy_rwork)->work) &obj_hash[i].lock irq_context: 0 (wq_completion)cgroup_destroy (work_completion)(&(&css->destroy_rwork)->work) pool_lock#2 irq_context: 0 (wq_completion)cgroup_destroy (work_completion)(&(&css->destroy_rwork)->work) &cgrp->pidlist_mutex irq_context: 0 (wq_completion)cgroup_destroy (work_completion)(&(&css->destroy_rwork)->work) (wq_completion)cgroup_pidlist_destroy irq_context: 0 (wq_completion)cgroup_destroy (work_completion)(&(&css->destroy_rwork)->work) &wq->mutex irq_context: 0 (wq_completion)cgroup_destroy (work_completion)(&(&css->destroy_rwork)->work) &wq->mutex &pool->lock irq_context: 0 (wq_completion)cgroup_destroy (work_completion)(&(&css->destroy_rwork)->work) &wq->mutex &x->wait#10 irq_context: 0 (wq_completion)cgroup_destroy (work_completion)(&(&css->destroy_rwork)->work) (work_completion)(&cgrp->release_agent_work) irq_context: 0 (wq_completion)cgroup_destroy (work_completion)(&(&css->destroy_rwork)->work) cgroup_mutex irq_context: 0 (wq_completion)cgroup_destroy (work_completion)(&(&css->destroy_rwork)->work) cgroup_mutex css_set_lock irq_context: 0 (wq_completion)cgroup_destroy (work_completion)(&(&css->destroy_rwork)->work) cgroup_mutex cpu_hotplug_lock irq_context: 0 (wq_completion)cgroup_destroy (work_completion)(&(&css->destroy_rwork)->work) cgroup_mutex cpu_hotplug_lock css_set_lock irq_context: 0 (wq_completion)cgroup_destroy (work_completion)(&(&css->destroy_rwork)->work) cgroup_mutex &root->kernfs_rwsem irq_context: 0 (wq_completion)cgroup_destroy (work_completion)(&(&css->destroy_rwork)->work) cgroup_mutex &root->kernfs_rwsem &root->kernfs_iattr_rwsem irq_context: 0 (wq_completion)cgroup_destroy (work_completion)(&(&css->destroy_rwork)->work) cgroup_mutex &root->kernfs_rwsem &obj_hash[i].lock irq_context: 0 (wq_completion)cgroup_destroy (work_completion)(&(&css->destroy_rwork)->work) cgroup_mutex &root->kernfs_rwsem pool_lock#2 irq_context: 0 (wq_completion)cgroup_destroy (work_completion)(&(&css->destroy_rwork)->work) cgroup_mutex &root->kernfs_rwsem kernfs_idr_lock irq_context: 0 (wq_completion)cgroup_destroy (work_completion)(&(&css->destroy_rwork)->work) cgroup_mutex cpu_hotplug_lock jump_label_mutex irq_context: 0 (wq_completion)cgroup_destroy (work_completion)(&(&css->destroy_rwork)->work) cgroup_mutex css_set_lock &obj_hash[i].lock irq_context: 0 (wq_completion)cgroup_destroy (work_completion)(&(&css->destroy_rwork)->work) cgroup_mutex css_set_lock pool_lock#2 irq_context: 0 (wq_completion)cgroup_destroy (work_completion)(&(&css->destroy_rwork)->work) cgroup_rstat_lock irq_context: 0 (wq_completion)cgroup_destroy (work_completion)(&(&css->destroy_rwork)->work) cgroup_rstat_lock per_cpu_ptr(&cgroup_rstat_cpu_lock, cpu) irq_context: 0 (wq_completion)cgroup_destroy (work_completion)(&(&css->destroy_rwork)->work) pcpu_lock irq_context: 0 (wq_completion)cgroup_destroy (work_completion)(&(&css->destroy_rwork)->work) &root->kernfs_rwsem irq_context: 0 (wq_completion)cgroup_destroy (work_completion)(&(&css->destroy_rwork)->work) &root->kernfs_rwsem &root->kernfs_iattr_rwsem irq_context: 0 (wq_completion)cgroup_destroy (work_completion)(&(&css->destroy_rwork)->work) &root->kernfs_rwsem &obj_hash[i].lock irq_context: 0 (wq_completion)cgroup_destroy (work_completion)(&(&css->destroy_rwork)->work) &root->kernfs_rwsem pool_lock#2 irq_context: 0 (wq_completion)cgroup_destroy (work_completion)(&(&css->destroy_rwork)->work) &root->kernfs_rwsem kernfs_idr_lock irq_context: 0 (wq_completion)cgroup_destroy (work_completion)(&(&css->destroy_rwork)->work) kernfs_idr_lock irq_context: 0 (wq_completion)cgroup_destroy (work_completion)(&(&css->destroy_rwork)->work) kernfs_idr_lock &obj_hash[i].lock irq_context: 0 (wq_completion)cgroup_destroy (work_completion)(&(&css->destroy_rwork)->work) kernfs_idr_lock pool_lock#2 irq_context: 0 (wq_completion)cgroup_destroy (work_completion)(&(&css->destroy_rwork)->work) cgroup_mutex rcu_state.exp_mutex rcu_node_0 irq_context: 0 (wq_completion)cgroup_destroy (work_completion)(&(&css->destroy_rwork)->work) cgroup_mutex rcu_state.exp_mutex &obj_hash[i].lock irq_context: 0 (wq_completion)cgroup_destroy (work_completion)(&(&css->destroy_rwork)->work) cgroup_mutex rcu_state.exp_mutex rcu_read_lock &pool->lock irq_context: 0 (wq_completion)cgroup_destroy (work_completion)(&(&css->destroy_rwork)->work) cgroup_mutex rcu_state.exp_mutex rcu_read_lock &pool->lock &obj_hash[i].lock irq_context: 0 (wq_completion)cgroup_destroy (work_completion)(&(&css->destroy_rwork)->work) cgroup_mutex rcu_state.exp_mutex &rnp->exp_wq[3] irq_context: 0 (wq_completion)cgroup_destroy (work_completion)(&(&css->destroy_rwork)->work) cgroup_mutex rcu_state.exp_mutex &pool->lock irq_context: 0 (wq_completion)cgroup_destroy (work_completion)(&(&css->destroy_rwork)->work) cgroup_mutex rcu_state.exp_mutex &pool->lock &p->pi_lock irq_context: 0 (wq_completion)cgroup_destroy (work_completion)(&(&css->destroy_rwork)->work) cgroup_mutex rcu_state.exp_mutex &pool->lock &p->pi_lock &rq->__lock irq_context: 0 (wq_completion)cgroup_destroy (work_completion)(&(&css->destroy_rwork)->work) cgroup_mutex rcu_state.exp_mutex &pool->lock &p->pi_lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)cgroup_destroy (work_completion)(&(&css->destroy_rwork)->work) cgroup_mutex rcu_state.exp_mutex &rq->__lock irq_context: 0 (wq_completion)cgroup_destroy (work_completion)(&(&css->destroy_rwork)->work) cgroup_mutex rcu_state.exp_mutex &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)cgroup_destroy (work_completion)(&(&css->destroy_rwork)->work) cgroup_mutex cgroup_mutex.wait_lock irq_context: 0 (wq_completion)cgroup_destroy (work_completion)(&(&css->destroy_rwork)->work) cgroup_mutex &pool->lock irq_context: 0 (wq_completion)cgroup_destroy (work_completion)(&(&css->destroy_rwork)->work) cgroup_mutex rcu_node_0 irq_context: 0 (wq_completion)cgroup_destroy (work_completion)(&(&css->destroy_rwork)->work) cgroup_mutex &rcu_state.expedited_wq irq_context: 0 (wq_completion)cgroup_destroy (work_completion)(&(&css->destroy_rwork)->work) cgroup_mutex &rcu_state.expedited_wq &p->pi_lock irq_context: 0 (wq_completion)cgroup_destroy (work_completion)(&(&css->destroy_rwork)->work) cgroup_mutex &rq->__lock irq_context: 0 (wq_completion)cgroup_destroy (work_completion)(&(&css->destroy_rwork)->work) cgroup_mutex &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)cgroup_destroy (work_completion)(&(&css->destroy_rwork)->work) cgroup_mutex &obj_hash[i].lock irq_context: 0 (wq_completion)cgroup_destroy (work_completion)(&(&css->destroy_rwork)->work) cgroup_mutex cpu_hotplug_lock jump_label_mutex text_mutex irq_context: 0 (wq_completion)cgroup_destroy (work_completion)(&(&css->destroy_rwork)->work) cgroup_mutex cpu_hotplug_lock jump_label_mutex text_mutex ptlock_ptr(page)#2 irq_context: 0 (wq_completion)cgroup_destroy (work_completion)(&(&css->destroy_rwork)->work) cgroup_mutex fs_reclaim irq_context: 0 (wq_completion)cgroup_destroy (work_completion)(&(&css->destroy_rwork)->work) cgroup_mutex fs_reclaim mmu_notifier_invalidate_range_start irq_context: 0 (wq_completion)cgroup_destroy (work_completion)(&(&css->destroy_rwork)->work) cgroup_mutex pool_lock#2 irq_context: 0 (wq_completion)cgroup_destroy (work_completion)(&(&css->destroy_rwork)->work) cgroup_mutex lock irq_context: 0 (wq_completion)cgroup_destroy (work_completion)(&(&css->destroy_rwork)->work) cgroup_mutex lock kernfs_idr_lock irq_context: 0 (wq_completion)cgroup_destroy (work_completion)(&(&css->destroy_rwork)->work) cgroup_mutex &root->kernfs_rwsem &root->kernfs_iattr_rwsem tk_core.seq.seqcount irq_context: 0 (wq_completion)cgroup_destroy (work_completion)(&(&css->destroy_rwork)->work) cgroup_mutex &c->lock irq_context: 0 (wq_completion)cgroup_destroy (work_completion)(&(&css->destroy_rwork)->work) cgroup_mutex.wait_lock irq_context: 0 (wq_completion)cgroup_destroy (work_completion)(&(&css->destroy_rwork)->work) &p->pi_lock irq_context: 0 cgroup_mutex rcu_state.exp_mutex &rnp->exp_wq[0] irq_context: 0 cgroup_mutex rcu_state.exp_mutex &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 sb_writers#11 irq_context: 0 sb_writers#11 mount_lock irq_context: 0 sb_writers#11 &type->i_mutex_dir_key#7 irq_context: 0 sb_writers#11 &type->i_mutex_dir_key#7 tomoyo_ss irq_context: 0 sb_writers#11 &type->i_mutex_dir_key#7 tomoyo_ss mmu_notifier_invalidate_range_start irq_context: 0 sb_writers#11 &type->i_mutex_dir_key#7 tomoyo_ss &c->lock irq_context: 0 sb_writers#11 &type->i_mutex_dir_key#7 tomoyo_ss &____s->seqcount irq_context: 0 sb_writers#11 &type->i_mutex_dir_key#7 tomoyo_ss rcu_read_lock mount_lock.seqcount irq_context: 0 sb_writers#11 &type->i_mutex_dir_key#7 tomoyo_ss rcu_read_lock rcu_read_lock rename_lock.seqcount irq_context: 0 sb_writers#11 &type->i_mutex_dir_key#7 tomoyo_ss &obj_hash[i].lock irq_context: 0 sb_writers#11 &type->i_mutex_dir_key#7 tomoyo_ss rcu_read_lock &dentry->d_lock irq_context: 0 sb_writers#11 &type->i_mutex_dir_key#7 tomoyo_ss tomoyo_policy_lock irq_context: 0 sb_writers#11 &type->i_mutex_dir_key#7 tk_core.seq.seqcount irq_context: 0 sb_writers#11 &type->i_mutex_dir_key#7 &root->kernfs_iattr_rwsem irq_context: 0 sb_writers#11 &type->i_mutex_dir_key#7 &root->kernfs_iattr_rwsem iattr_mutex irq_context: 0 sb_writers#11 &type->i_mutex_dir_key#7 &root->kernfs_iattr_rwsem iattr_mutex fs_reclaim irq_context: 0 sb_writers#11 &type->i_mutex_dir_key#7 &root->kernfs_iattr_rwsem iattr_mutex fs_reclaim mmu_notifier_invalidate_range_start irq_context: 0 sb_writers#11 &type->i_mutex_dir_key#7 &root->kernfs_iattr_rwsem iattr_mutex tk_core.seq.seqcount irq_context: 0 cgroup_mutex cpuset_mutex irq_context: 0 cgroup_mutex cpuset_mutex callback_lock irq_context: 0 &type->s_umount_key#45/1 &n->list_lock irq_context: 0 &type->s_umount_key#45/1 &n->list_lock &c->lock irq_context: 0 cgroup_mutex lock kernfs_idr_lock &c->lock irq_context: 0 cgroup_mutex lock kernfs_idr_lock &n->list_lock irq_context: 0 cgroup_mutex lock kernfs_idr_lock &n->list_lock &c->lock irq_context: 0 cgroup_mutex rcu_read_lock &pool->lock irq_context: 0 cgroup_mutex rcu_read_lock &pool->lock &obj_hash[i].lock irq_context: 0 cgroup_mutex rcu_read_lock &pool->lock &p->pi_lock irq_context: 0 cgroup_mutex rcu_read_lock &pool->lock &p->pi_lock &rq->__lock irq_context: 0 cgroup_mutex rcu_read_lock &pool->lock &p->pi_lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 cgroup_mutex &rq->__lock irq_context: 0 cgroup_mutex &dom->lock irq_context: 0 (wq_completion)cgroup_destroy (work_completion)(&(&css->destroy_rwork)->work) cgroup_mutex cpuset_mutex irq_context: 0 (wq_completion)cgroup_destroy (work_completion)(&(&css->destroy_rwork)->work) cgroup_mutex cpuset_mutex callback_lock irq_context: 0 (wq_completion)cgroup_destroy (work_completion)(&(&css->destroy_rwork)->work) cgroup_mutex &dom->lock irq_context: 0 (wq_completion)cgroup_destroy (work_completion)(&(&css->destroy_rwork)->work) cgroup_mutex rcu_state.exp_mutex &rnp->exp_wq[2] irq_context: 0 cgroup_mutex rcu_state.exp_mutex &rnp->exp_wq[3] irq_context: 0 sb_writers#11 &type->i_mutex_dir_key#7 irq_context: 0 sb_writers#11 &type->i_mutex_dir_key#7 rename_lock.seqcount irq_context: 0 sb_writers#11 &type->i_mutex_dir_key#7 fs_reclaim irq_context: 0 sb_writers#11 &type->i_mutex_dir_key#7 fs_reclaim mmu_notifier_invalidate_range_start irq_context: 0 sb_writers#11 &type->i_mutex_dir_key#7 &dentry->d_lock irq_context: 0 sb_writers#11 &type->i_mutex_dir_key#7 rcu_read_lock rename_lock.seqcount irq_context: 0 sb_writers#11 &type->i_mutex_dir_key#7 &root->kernfs_rwsem irq_context: 0 sb_writers#11 &type->i_mutex_dir_key#7 &root->kernfs_rwsem inode_hash_lock irq_context: 0 sb_writers#11 &type->i_mutex_dir_key#7 &root->kernfs_rwsem fs_reclaim irq_context: 0 sb_writers#11 &type->i_mutex_dir_key#7 &root->kernfs_rwsem fs_reclaim mmu_notifier_invalidate_range_start irq_context: 0 sb_writers#11 &type->i_mutex_dir_key#7 &root->kernfs_rwsem mmu_notifier_invalidate_range_start irq_context: 0 sb_writers#11 &type->i_mutex_dir_key#7 &root->kernfs_rwsem &c->lock irq_context: 0 sb_writers#11 &type->i_mutex_dir_key#7 &root->kernfs_rwsem inode_hash_lock &sb->s_type->i_lock_key#31 irq_context: 0 sb_writers#11 &type->i_mutex_dir_key#7 &root->kernfs_rwsem inode_hash_lock &s->s_inode_list_lock irq_context: 0 sb_writers#11 &type->i_mutex_dir_key#7 &root->kernfs_rwsem tk_core.seq.seqcount irq_context: 0 sb_writers#11 &type->i_mutex_dir_key#7 &root->kernfs_rwsem &sb->s_type->i_lock_key#31 irq_context: 0 sb_writers#11 &type->i_mutex_dir_key#7 &sb->s_type->i_lock_key#31 irq_context: 0 sb_writers#11 &type->i_mutex_dir_key#7 &sb->s_type->i_lock_key#31 &dentry->d_lock irq_context: 0 sb_writers#11 &type->i_mutex_dir_key#7 &sb->s_type->i_lock_key#31 &dentry->d_lock &wq#2 irq_context: 0 kn->active#49 fs_reclaim irq_context: 0 kn->active#49 fs_reclaim mmu_notifier_invalidate_range_start irq_context: 0 kn->active#49 &kernfs_locks->open_file_mutex[count] irq_context: 0 kn->active#49 &kernfs_locks->open_file_mutex[count] fs_reclaim irq_context: 0 kn->active#49 &kernfs_locks->open_file_mutex[count] fs_reclaim mmu_notifier_invalidate_range_start irq_context: 0 &sb->s_type->i_mutex_key#16 irq_context: 0 sb_writers#11 fs_reclaim irq_context: 0 sb_writers#11 fs_reclaim mmu_notifier_invalidate_range_start irq_context: 0 sb_writers#11 &mm->mmap_lock irq_context: 0 sb_writers#11 &of->mutex irq_context: 0 sb_writers#11 &obj_hash[i].lock irq_context: 0 kn->active#50 fs_reclaim irq_context: 0 kn->active#50 fs_reclaim mmu_notifier_invalidate_range_start irq_context: 0 kn->active#50 &kernfs_locks->open_file_mutex[count] irq_context: 0 kn->active#50 &kernfs_locks->open_file_mutex[count] fs_reclaim irq_context: 0 kn->active#50 &kernfs_locks->open_file_mutex[count] fs_reclaim mmu_notifier_invalidate_range_start irq_context: 0 sb_writers#11 &of->mutex kn->active#50 cpu_hotplug_lock irq_context: 0 sb_writers#11 &of->mutex kn->active#50 cpu_hotplug_lock cpuset_mutex irq_context: 0 sb_writers#9 &mm->mmap_lock irq_context: 0 &type->s_umount_key#47 irq_context: 0 &type->s_umount_key#47 sb_lock irq_context: 0 &type->s_umount_key#47 &dentry->d_lock irq_context: 0 &sb->s_type->i_mutex_key#17 irq_context: 0 &sb->s_type->i_mutex_key#17 namespace_sem irq_context: 0 &sb->s_type->i_mutex_key#17 namespace_sem rcu_read_lock mount_lock.seqcount irq_context: 0 &sb->s_type->i_mutex_key#17 namespace_sem fs_reclaim irq_context: 0 &sb->s_type->i_mutex_key#17 namespace_sem fs_reclaim mmu_notifier_invalidate_range_start irq_context: 0 &sb->s_type->i_mutex_key#17 namespace_sem rename_lock irq_context: 0 &sb->s_type->i_mutex_key#17 namespace_sem rename_lock rename_lock.seqcount irq_context: 0 &sb->s_type->i_mutex_key#17 namespace_sem rename_lock rename_lock.seqcount &dentry->d_lock irq_context: 0 &sb->s_type->i_mutex_key#17 namespace_sem mount_lock irq_context: 0 &sb->s_type->i_mutex_key#17 namespace_sem mount_lock &dentry->d_lock irq_context: 0 &sb->s_type->i_mutex_key#17 namespace_sem mount_lock rcu_read_lock &dentry->d_lock irq_context: 0 &sb->s_type->i_mutex_key#17 namespace_sem mount_lock &obj_hash[i].lock irq_context: 0 &sb->s_type->i_mutex_key#17 namespace_sem mount_lock pool_lock#2 irq_context: 0 &sb->s_type->i_lock_key#26 irq_context: 0 sb_writers#12 irq_context: 0 sb_writers#12 fs_reclaim irq_context: 0 sb_writers#12 fs_reclaim mmu_notifier_invalidate_range_start irq_context: 0 sb_writers#12 &c->lock irq_context: 0 sb_writers#12 pool_lock#2 irq_context: 0 sb_writers#12 &mm->mmap_lock irq_context: 0 sb_writers#12 &sb->s_type->i_mutex_key#17 irq_context: 0 sb_writers#12 &sb->s_type->i_mutex_key#17 rename_lock.seqcount irq_context: 0 sb_writers#12 &sb->s_type->i_mutex_key#17 fs_reclaim irq_context: 0 sb_writers#12 &sb->s_type->i_mutex_key#17 fs_reclaim mmu_notifier_invalidate_range_start irq_context: 0 sb_writers#12 &sb->s_type->i_mutex_key#17 pool_lock#2 irq_context: 0 sb_writers#12 &sb->s_type->i_mutex_key#17 &dentry->d_lock irq_context: 0 sb_writers#12 &sb->s_type->i_mutex_key#17 rcu_read_lock rename_lock.seqcount irq_context: 0 sb_writers#12 &sb->s_type->i_mutex_key#17 &dentry->d_lock &wq irq_context: 0 sb_writers#12 &sb->s_type->i_mutex_key#17 mmu_notifier_invalidate_range_start irq_context: 0 sb_writers#12 &sb->s_type->i_mutex_key#17 &sb->s_type->i_lock_key#26 irq_context: 0 sb_writers#12 &sb->s_type->i_mutex_key#17 &s->s_inode_list_lock irq_context: 0 sb_writers#12 &sb->s_type->i_mutex_key#17 tk_core.seq.seqcount irq_context: 0 sb_writers#12 &sb->s_type->i_mutex_key#17 pin_fs_lock irq_context: 0 sb_writers#12 &sb->s_type->i_mutex_key#17 sb_lock irq_context: 0 sb_writers#12 &sb->s_type->i_mutex_key#17 &type->s_umount_key#47 irq_context: 0 sb_writers#12 &sb->s_type->i_mutex_key#17 &type->s_umount_key#47 sb_lock irq_context: 0 sb_writers#12 &sb->s_type->i_mutex_key#17 &type->s_umount_key#47 &dentry->d_lock irq_context: 0 sb_writers#12 &sb->s_type->i_mutex_key#17 mnt_id_ida.xa_lock irq_context: 0 sb_writers#12 &sb->s_type->i_mutex_key#17 pcpu_alloc_mutex irq_context: 0 sb_writers#12 &sb->s_type->i_mutex_key#17 pcpu_alloc_mutex pcpu_lock irq_context: 0 sb_writers#12 &sb->s_type->i_mutex_key#17 mount_lock irq_context: 0 sb_writers#12 &sb->s_type->i_mutex_key#17 mount_lock mount_lock.seqcount irq_context: 0 sb_writers#12 &sb->s_type->i_mutex_key#17 rcu_read_lock &dentry->d_lock irq_context: 0 sb_writers#12 &sb->s_type->i_mutex_key#17 &obj_hash[i].lock irq_context: 0 sb_writers#12 &sb->s_type->i_mutex_key#17 rcu_read_lock mount_lock irq_context: 0 sb_writers#12 &sb->s_type->i_mutex_key#17 rcu_read_lock mount_lock mount_lock.seqcount irq_context: 0 sb_writers#12 &sb->s_type->i_mutex_key#17 &sb->s_type->i_lock_key#26 &dentry->d_lock irq_context: 0 sb_writers#12 &sb->s_type->i_mutex_key#17 entries_lock irq_context: 0 sb_writers#12 &sb->s_type->i_mutex_key#17 &c->lock irq_context: 0 sb_writers#12 &sb->s_type->i_mutex_key#17 &____s->seqcount irq_context: 0 cb_lock genl_mutex rtnl_mutex irq_context: 0 rtnl_mutex dev_addr_sem irq_context: 0 rtnl_mutex dev_addr_sem net_rwsem irq_context: 0 rtnl_mutex dev_addr_sem net_rwsem &list->lock#2 irq_context: 0 rtnl_mutex dev_addr_sem &tn->lock irq_context: 0 rtnl_mutex dev_addr_sem &sdata->sec_mtx irq_context: 0 rtnl_mutex dev_addr_sem &sdata->sec_mtx &sec->lock irq_context: 0 rtnl_mutex dev_addr_sem fs_reclaim irq_context: 0 rtnl_mutex dev_addr_sem fs_reclaim mmu_notifier_invalidate_range_start irq_context: 0 rtnl_mutex dev_addr_sem pool_lock#2 irq_context: 0 rtnl_mutex dev_addr_sem nl_table_lock irq_context: 0 rtnl_mutex dev_addr_sem rlock-AF_NETLINK irq_context: 0 rtnl_mutex dev_addr_sem rcu_read_lock &ei->socket.wq.wait irq_context: 0 rtnl_mutex dev_addr_sem rcu_read_lock &ei->socket.wq.wait &p->pi_lock irq_context: 0 rtnl_mutex dev_addr_sem rcu_read_lock &ei->socket.wq.wait &p->pi_lock &cfs_rq->removed.lock irq_context: 0 rtnl_mutex dev_addr_sem rcu_read_lock &ei->socket.wq.wait &p->pi_lock &rq->__lock irq_context: 0 rtnl_mutex dev_addr_sem rcu_read_lock &ei->socket.wq.wait &p->pi_lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 rtnl_mutex dev_addr_sem nl_table_wait.lock irq_context: 0 rtnl_mutex dev_addr_sem &tbl->lock irq_context: 0 rtnl_mutex dev_addr_sem &pn->hash_lock irq_context: 0 rtnl_mutex dev_addr_sem &net->ipv6.fib6_gc_lock rcu_read_lock &tb->tb6_lock irq_context: 0 rtnl_mutex dev_addr_sem &net->ipv6.fib6_gc_lock &obj_hash[i].lock irq_context: 0 rtnl_mutex dev_addr_sem input_pool.lock irq_context: 0 rtnl_mutex _xmit_IEEE802154 irq_context: 0 rtnl_mutex rcu_read_lock &ei->socket.wq.wait irq_context: 0 rtnl_mutex rcu_read_lock &ei->socket.wq.wait &p->pi_lock irq_context: 0 rtnl_mutex lock kernfs_idr_lock &obj_hash[i].lock irq_context: 0 rtnl_mutex proc_subdir_lock irq_context: 0 rtnl_mutex proc_inum_ida.xa_lock irq_context: 0 rtnl_mutex proc_subdir_lock irq_context: 0 rtnl_mutex &idev->mc_lock &c->lock irq_context: 0 rtnl_mutex &idev->mc_lock &obj_hash[i].lock pool_lock irq_context: 0 &sb->s_type->i_mutex_key#10 (netlink_chain).rwsem rcu_read_lock &pool->lock &p->pi_lock &rq->__lock irq_context: 0 &sb->s_type->i_mutex_key#10 (netlink_chain).rwsem rcu_read_lock &pool->lock &p->pi_lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 &sb->s_type->i_mutex_key#10 (netlink_chain).rwsem &rq->__lock irq_context: 0 sb_writers#4 &type->i_mutex_dir_key#3 irq_context: 0 sb_writers#4 &type->i_mutex_dir_key#3 rename_lock.seqcount irq_context: 0 sb_writers#4 &type->i_mutex_dir_key#3 fs_reclaim irq_context: 0 sb_writers#4 &type->i_mutex_dir_key#3 fs_reclaim mmu_notifier_invalidate_range_start irq_context: 0 sb_writers#4 &type->i_mutex_dir_key#3 pool_lock#2 irq_context: 0 sb_writers#4 &type->i_mutex_dir_key#3 &dentry->d_lock irq_context: 0 sb_writers#4 &type->i_mutex_dir_key#3 rcu_read_lock rename_lock.seqcount irq_context: 0 sb_writers#4 &type->i_mutex_dir_key#3 &ei->i_es_lock irq_context: 0 sb_writers#4 &type->i_mutex_dir_key#3 &mapping->private_lock irq_context: 0 sb_writers#4 &type->i_mutex_dir_key#3 &dentry->d_lock &wq#2 irq_context: 0 sb_writers#4 &sb->s_type->i_mutex_key#8 mapping.invalidate_lock tk_core.seq.seqcount irq_context: 0 sb_writers#4 &sb->s_type->i_mutex_key#8 mapping.invalidate_lock jbd2_handle &ei->i_es_lock irq_context: 0 sb_writers#4 &sb->s_type->i_mutex_key#8 mapping.invalidate_lock jbd2_handle &ei->i_data_sem irq_context: 0 sb_writers#4 &sb->s_type->i_mutex_key#8 mapping.invalidate_lock jbd2_handle &ei->i_data_sem mmu_notifier_invalidate_range_start irq_context: 0 sb_writers#4 &sb->s_type->i_mutex_key#8 mapping.invalidate_lock jbd2_handle &ei->i_data_sem pool_lock#2 irq_context: 0 sb_writers#4 &sb->s_type->i_mutex_key#8 mapping.invalidate_lock jbd2_handle &ei->i_data_sem &ei->i_es_lock irq_context: 0 sb_writers#4 &sb->s_type->i_mutex_key#8 mapping.invalidate_lock jbd2_handle &ei->i_data_sem &ei->i_es_lock irq_context: 0 sb_writers#4 &sb->s_type->i_mutex_key#8 mapping.invalidate_lock jbd2_handle &ei->i_data_sem &ei->i_es_lock pool_lock#2 irq_context: 0 sb_writers#4 &sb->s_type->i_mutex_key#8 mapping.invalidate_lock jbd2_handle &ei->i_data_sem &ei->i_es_lock &sbi->s_es_lock irq_context: 0 sb_writers#4 &sb->s_type->i_mutex_key#8 mapping.invalidate_lock jbd2_handle &ei->i_data_sem &obj_hash[i].lock irq_context: 0 sb_writers#4 &sb->s_type->i_mutex_key#8 mapping.invalidate_lock jbd2_handle &ei->i_data_sem key#14 irq_context: 0 sb_writers#4 &sb->s_type->i_mutex_key#8 mapping.invalidate_lock jbd2_handle &ei->i_data_sem &sb->s_type->i_lock_key#22 irq_context: 0 sb_writers#4 &sb->s_type->i_mutex_key#8 mapping.invalidate_lock jbd2_handle &ei->i_data_sem &wb->list_lock irq_context: 0 sb_writers#4 &sb->s_type->i_mutex_key#8 mapping.invalidate_lock jbd2_handle &ei->i_data_sem &wb->list_lock &sb->s_type->i_lock_key#22 irq_context: 0 sb_writers#4 &sb->s_type->i_mutex_key#8 mapping.invalidate_lock jbd2_handle &ei->i_data_sem &ei->i_prealloc_lock irq_context: 0 sb_writers#4 &sb->s_type->i_mutex_key#8 mapping.invalidate_lock jbd2_handle &ei->i_data_sem &sbi->s_md_lock irq_context: 0 sb_writers#4 &sb->s_type->i_mutex_key#8 mapping.invalidate_lock jbd2_handle &ei->i_data_sem &bgl->locks[i].lock &sbi->s_md_lock irq_context: 0 sb_writers#4 &sb->s_type->i_mutex_key#8 mapping.invalidate_lock jbd2_handle &ei->i_data_sem &mapping->private_lock irq_context: 0 sb_writers#4 &sb->s_type->i_mutex_key#8 mapping.invalidate_lock jbd2_handle &ei->i_data_sem &____s->seqcount irq_context: 0 sb_writers#4 &sb->s_type->i_mutex_key#8 mapping.invalidate_lock jbd2_handle &ei->i_data_sem &ret->b_state_lock irq_context: 0 sb_writers#4 &sb->s_type->i_mutex_key#8 mapping.invalidate_lock jbd2_handle &ei->i_data_sem &ret->b_state_lock &journal->j_list_lock irq_context: 0 sb_writers#4 &sb->s_type->i_mutex_key#8 mapping.invalidate_lock jbd2_handle &ei->i_data_sem &journal->j_revoke_lock irq_context: 0 sb_writers#4 &sb->s_type->i_mutex_key#8 mapping.invalidate_lock jbd2_handle &ei->i_data_sem key#3 irq_context: 0 sb_writers#4 &sb->s_type->i_mutex_key#8 mapping.invalidate_lock jbd2_handle &ei->i_data_sem &ei->i_es_lock irq_context: 0 sb_writers#4 &sb->s_type->i_mutex_key#8 mapping.invalidate_lock jbd2_handle &ei->i_data_sem &ei->i_es_lock key#7 irq_context: 0 &sb->s_type->i_mutex_key#10 nl_table_lock nl_table_wait.lock irq_context: 0 &sb->s_type->i_mutex_key#10 &rq->__lock irq_context: 0 &sb->s_type->i_mutex_key#10 &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 nl_table_wait.lock &p->pi_lock irq_context: 0 nl_table_wait.lock &p->pi_lock &rq->__lock irq_context: 0 nl_table_wait.lock &p->pi_lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 &wb->list_lock irq_context: 0 &sbi->s_writepages_rwsem irq_context: 0 &sbi->s_writepages_rwsem &xa->xa_lock#9 irq_context: 0 &sbi->s_writepages_rwsem mmu_notifier_invalidate_range_start irq_context: 0 &sbi->s_writepages_rwsem &____s->seqcount irq_context: 0 &sbi->s_writepages_rwsem pool_lock#2 irq_context: 0 &sbi->s_writepages_rwsem &c->lock irq_context: 0 &sbi->s_writepages_rwsem lock#4 irq_context: 0 &sbi->s_writepages_rwsem lock#4 &lruvec->lru_lock irq_context: 0 &sbi->s_writepages_rwsem lock#5 irq_context: 0 &sbi->s_writepages_rwsem &obj_hash[i].lock irq_context: 0 &sbi->s_writepages_rwsem &journal->j_state_lock irq_context: 0 &sbi->s_writepages_rwsem jbd2_handle irq_context: 0 &sbi->s_writepages_rwsem jbd2_handle lock#4 irq_context: 0 &sbi->s_writepages_rwsem jbd2_handle lock#5 irq_context: 0 &sbi->s_writepages_rwsem jbd2_handle mmu_notifier_invalidate_range_start irq_context: 0 &sbi->s_writepages_rwsem jbd2_handle &____s->seqcount irq_context: 0 &sbi->s_writepages_rwsem jbd2_handle pool_lock#2 irq_context: 0 &sbi->s_writepages_rwsem jbd2_handle &c->lock irq_context: 0 &sbi->s_writepages_rwsem jbd2_handle &ei->i_es_lock irq_context: 0 &sbi->s_writepages_rwsem jbd2_handle &ei->i_data_sem irq_context: 0 &sbi->s_writepages_rwsem jbd2_handle &ei->i_data_sem mmu_notifier_invalidate_range_start irq_context: 0 &sbi->s_writepages_rwsem jbd2_handle &ei->i_data_sem pool_lock#2 irq_context: 0 &sbi->s_writepages_rwsem jbd2_handle &ei->i_data_sem &ei->i_es_lock irq_context: 0 &sbi->s_writepages_rwsem jbd2_handle &ei->i_data_sem &ei->i_raw_lock irq_context: 0 &sbi->s_writepages_rwsem jbd2_handle &ei->i_data_sem &obj_hash[i].lock irq_context: 0 &sbi->s_writepages_rwsem jbd2_handle &ei->i_data_sem &ei->i_es_lock irq_context: 0 &sbi->s_writepages_rwsem jbd2_handle rcu_read_lock &memcg->move_lock irq_context: 0 &sbi->s_writepages_rwsem jbd2_handle rcu_read_lock &xa->xa_lock#9 irq_context: 0 &sbi->s_writepages_rwsem jbd2_handle rcu_read_lock &xa->xa_lock#9 &s->s_inode_wblist_lock irq_context: 0 &sbi->s_writepages_rwsem jbd2_handle &journal->j_wait_updates irq_context: 0 &sbi->s_writepages_rwsem tk_core.seq.seqcount irq_context: 0 &sbi->s_writepages_rwsem &dd->lock irq_context: 0 &sbi->s_writepages_rwsem &base->lock irq_context: 0 &sbi->s_writepages_rwsem &base->lock &obj_hash[i].lock irq_context: 0 &sbi->s_writepages_rwsem rcu_read_lock &dd->lock irq_context: 0 &sbi->s_writepages_rwsem rcu_read_lock &obj_hash[i].lock irq_context: 0 &sbi->s_writepages_rwsem rcu_read_lock tk_core.seq.seqcount irq_context: 0 &sbi->s_writepages_rwsem rcu_read_lock &virtscsi_vq->vq_lock irq_context: softirq &ei->i_completed_io_lock irq_context: softirq &ei->i_completed_io_lock rcu_read_lock &pool->lock/1 irq_context: softirq &ei->i_completed_io_lock rcu_read_lock &pool->lock/1 &obj_hash[i].lock irq_context: softirq &ei->i_completed_io_lock rcu_read_lock &pool->lock/1 pool_lock#2 irq_context: softirq &ei->i_completed_io_lock rcu_read_lock &pool->lock/1 &p->pi_lock irq_context: softirq &ei->i_completed_io_lock rcu_read_lock &pool->lock/1 &p->pi_lock &cfs_rq->removed.lock irq_context: softirq &ei->i_completed_io_lock rcu_read_lock &pool->lock/1 &p->pi_lock &rq->__lock irq_context: softirq &ei->i_completed_io_lock rcu_read_lock &pool->lock/1 &p->pi_lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)ext4-rsv-conversion irq_context: 0 (wq_completion)ext4-rsv-conversion (work_completion)(&ei->i_rsv_conversion_work) irq_context: 0 (wq_completion)ext4-rsv-conversion (work_completion)(&ei->i_rsv_conversion_work) &ei->i_completed_io_lock irq_context: 0 (wq_completion)ext4-rsv-conversion (work_completion)(&ei->i_rsv_conversion_work) &journal->j_state_lock irq_context: 0 (wq_completion)ext4-rsv-conversion (work_completion)(&ei->i_rsv_conversion_work) &journal->j_state_lock &journal->j_wait_reserved irq_context: 0 (wq_completion)ext4-rsv-conversion (work_completion)(&ei->i_rsv_conversion_work) jbd2_handle irq_context: 0 (wq_completion)ext4-rsv-conversion (work_completion)(&ei->i_rsv_conversion_work) jbd2_handle &ei->i_es_lock irq_context: 0 (wq_completion)ext4-rsv-conversion (work_completion)(&ei->i_rsv_conversion_work) jbd2_handle &ei->i_data_sem irq_context: 0 (wq_completion)ext4-rsv-conversion (work_completion)(&ei->i_rsv_conversion_work) jbd2_handle &ei->i_data_sem mmu_notifier_invalidate_range_start irq_context: 0 (wq_completion)ext4-rsv-conversion (work_completion)(&ei->i_rsv_conversion_work) jbd2_handle &ei->i_data_sem pool_lock#2 irq_context: 0 (wq_completion)ext4-rsv-conversion (work_completion)(&ei->i_rsv_conversion_work) jbd2_handle &ei->i_data_sem &ei->i_es_lock irq_context: 0 (wq_completion)ext4-rsv-conversion (work_completion)(&ei->i_rsv_conversion_work) jbd2_handle &ei->i_data_sem &ei->i_raw_lock irq_context: 0 (wq_completion)ext4-rsv-conversion (work_completion)(&ei->i_rsv_conversion_work) jbd2_handle &ei->i_data_sem &obj_hash[i].lock irq_context: 0 (wq_completion)ext4-rsv-conversion (work_completion)(&ei->i_rsv_conversion_work) jbd2_handle &ei->i_data_sem &ei->i_es_lock pool_lock#2 irq_context: 0 (wq_completion)ext4-rsv-conversion (work_completion)(&ei->i_rsv_conversion_work) jbd2_handle &ei->i_raw_lock irq_context: 0 (wq_completion)ext4-rsv-conversion (work_completion)(&ei->i_rsv_conversion_work) jbd2_handle &journal->j_wait_updates irq_context: 0 (wq_completion)ext4-rsv-conversion (work_completion)(&ei->i_rsv_conversion_work) &obj_hash[i].lock irq_context: 0 (wq_completion)ext4-rsv-conversion (work_completion)(&ei->i_rsv_conversion_work) pool_lock#2 irq_context: 0 (wq_completion)ext4-rsv-conversion (work_completion)(&ei->i_rsv_conversion_work) &ext4__ioend_wq[i] irq_context: 0 (wq_completion)ext4-rsv-conversion (work_completion)(&ei->i_rsv_conversion_work) &ret->b_uptodate_lock irq_context: 0 (wq_completion)ext4-rsv-conversion (work_completion)(&ei->i_rsv_conversion_work) rcu_read_lock &memcg->move_lock irq_context: 0 (wq_completion)ext4-rsv-conversion (work_completion)(&ei->i_rsv_conversion_work) rcu_read_lock &xa->xa_lock#9 irq_context: 0 (wq_completion)ext4-rsv-conversion (work_completion)(&ei->i_rsv_conversion_work) rcu_read_lock &xa->xa_lock#9 key#10 irq_context: 0 (wq_completion)ext4-rsv-conversion (work_completion)(&ei->i_rsv_conversion_work) rcu_read_lock &xa->xa_lock#9 &obj_hash[i].lock irq_context: 0 (wq_completion)ext4-rsv-conversion (work_completion)(&ei->i_rsv_conversion_work) rcu_read_lock &xa->xa_lock#9 &base->lock irq_context: 0 (wq_completion)ext4-rsv-conversion (work_completion)(&ei->i_rsv_conversion_work) rcu_read_lock &xa->xa_lock#9 &base->lock &obj_hash[i].lock irq_context: 0 (wq_completion)ext4-rsv-conversion (work_completion)(&ei->i_rsv_conversion_work) rcu_read_lock &xa->xa_lock#9 &wb->work_lock irq_context: 0 (wq_completion)ext4-rsv-conversion (work_completion)(&ei->i_rsv_conversion_work) rcu_read_lock &xa->xa_lock#9 &wb->work_lock &obj_hash[i].lock irq_context: 0 (wq_completion)ext4-rsv-conversion (work_completion)(&ei->i_rsv_conversion_work) rcu_read_lock &xa->xa_lock#9 &wb->work_lock &base->lock irq_context: 0 (wq_completion)ext4-rsv-conversion (work_completion)(&ei->i_rsv_conversion_work) rcu_read_lock &xa->xa_lock#9 &wb->work_lock &base->lock &obj_hash[i].lock irq_context: 0 (wq_completion)ext4-rsv-conversion (work_completion)(&ei->i_rsv_conversion_work) rcu_read_lock &xa->xa_lock#9 &s->s_inode_wblist_lock irq_context: 0 (wq_completion)ext4-rsv-conversion (work_completion)(&ei->i_rsv_conversion_work) &folio_wait_table[i] irq_context: 0 (wq_completion)ext4-rsv-conversion (work_completion)(&ei->i_rsv_conversion_work) &folio_wait_table[i] &p->pi_lock irq_context: 0 (wq_completion)ext4-rsv-conversion (work_completion)(&ei->i_rsv_conversion_work) &folio_wait_table[i] &p->pi_lock &rq->__lock irq_context: 0 (wq_completion)ext4-rsv-conversion (work_completion)(&ei->i_rsv_conversion_work) &folio_wait_table[i] &p->pi_lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 &journal->j_state_lock &journal->j_wait_commit &p->pi_lock irq_context: 0 &journal->j_state_lock &journal->j_wait_commit &p->pi_lock &rq->__lock irq_context: 0 &journal->j_state_lock &journal->j_wait_commit &p->pi_lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 jbd2_handle irq_context: 0 &journal->j_wait_commit irq_context: 0 &ret->b_state_lock &journal->j_list_lock key#15 irq_context: 0 &journal->j_wait_done_commit &p->pi_lock irq_context: 0 &journal->j_wait_done_commit &p->pi_lock &rq->__lock irq_context: 0 &journal->j_wait_done_commit &p->pi_lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 &iint->mutex sb_writers#4 &journal->j_state_lock irq_context: 0 &iint->mutex sb_writers#4 &journal->j_state_lock tk_core.seq.seqcount irq_context: 0 &iint->mutex sb_writers#4 &journal->j_state_lock &obj_hash[i].lock irq_context: 0 &iint->mutex sb_writers#4 &journal->j_state_lock &base->lock irq_context: 0 &iint->mutex sb_writers#4 &journal->j_state_lock &base->lock &obj_hash[i].lock irq_context: 0 &iint->mutex mapping.invalidate_lock lock#4 rcu_read_lock pool_lock#2 irq_context: 0 &iint->mutex mapping.invalidate_lock lock#4 &obj_hash[i].lock irq_context: 0 &iint->mutex mapping.invalidate_lock &rq->__lock irq_context: 0 &iint->mutex mapping.invalidate_lock rcu_node_0 irq_context: 0 &sb->s_type->i_mutex_key#8 fs_reclaim irq_context: 0 &sb->s_type->i_mutex_key#8 fs_reclaim mmu_notifier_invalidate_range_start irq_context: 0 &sb->s_type->i_mutex_key#8 free_vmap_area_lock irq_context: 0 &sb->s_type->i_mutex_key#8 vmap_area_lock irq_context: 0 &sb->s_type->i_mutex_key#8 init_mm.page_table_lock irq_context: 0 &sb->s_type->i_mutex_key#8 free_vmap_area_lock &obj_hash[i].lock irq_context: 0 &sb->s_type->i_mutex_key#8 free_vmap_area_lock pool_lock#2 irq_context: 0 &sb->s_type->i_mutex_key#8 swap_cgroup_mutex irq_context: 0 &sb->s_type->i_mutex_key#8 swap_cgroup_mutex fs_reclaim irq_context: 0 &sb->s_type->i_mutex_key#8 swap_cgroup_mutex fs_reclaim mmu_notifier_invalidate_range_start irq_context: 0 &sb->s_type->i_mutex_key#8 swap_cgroup_mutex &____s->seqcount irq_context: 0 &sb->s_type->i_mutex_key#8 swap_cgroup_mutex pool_lock#2 irq_context: 0 &sb->s_type->i_mutex_key#8 &journal->j_state_lock irq_context: 0 &sb->s_type->i_mutex_key#8 tk_core.seq.seqcount irq_context: 0 &sb->s_type->i_mutex_key#8 &obj_hash[i].lock irq_context: 0 &sb->s_type->i_mutex_key#8 &base->lock irq_context: 0 &sb->s_type->i_mutex_key#8 &base->lock &obj_hash[i].lock irq_context: 0 &sb->s_type->i_mutex_key#8 &fq->mq_flush_lock irq_context: 0 &sb->s_type->i_mutex_key#8 &fq->mq_flush_lock tk_core.seq.seqcount irq_context: 0 &sb->s_type->i_mutex_key#8 &fq->mq_flush_lock &q->requeue_lock irq_context: 0 &sb->s_type->i_mutex_key#8 &fq->mq_flush_lock &obj_hash[i].lock irq_context: 0 &sb->s_type->i_mutex_key#8 &fq->mq_flush_lock rcu_read_lock &pool->lock irq_context: 0 &sb->s_type->i_mutex_key#8 &fq->mq_flush_lock rcu_read_lock &pool->lock &obj_hash[i].lock irq_context: 0 &sb->s_type->i_mutex_key#8 &fq->mq_flush_lock rcu_read_lock &pool->lock pool_lock#2 irq_context: 0 &sb->s_type->i_mutex_key#8 &fq->mq_flush_lock rcu_read_lock &pool->lock &p->pi_lock irq_context: 0 &sb->s_type->i_mutex_key#8 &fq->mq_flush_lock rcu_read_lock &pool->lock &p->pi_lock &rq->__lock irq_context: 0 &sb->s_type->i_mutex_key#8 &fq->mq_flush_lock rcu_read_lock &pool->lock &p->pi_lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 &sb->s_type->i_mutex_key#8 &rq->__lock irq_context: 0 &sb->s_type->i_mutex_key#8 &x->wait#26 irq_context: 0 &sb->s_type->i_mutex_key#8 &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 &sb->s_type->i_mutex_key#8 (&timer.timer) irq_context: 0 &sb->s_type->i_mutex_key#8 &ei->i_es_lock irq_context: 0 &sb->s_type->i_mutex_key#8 swapon_mutex irq_context: 0 &sb->s_type->i_mutex_key#8 swapon_mutex fs_reclaim irq_context: 0 &sb->s_type->i_mutex_key#8 swapon_mutex fs_reclaim mmu_notifier_invalidate_range_start irq_context: 0 &sb->s_type->i_mutex_key#8 swapon_mutex &c->lock irq_context: 0 &sb->s_type->i_mutex_key#8 swapon_mutex pool_lock#2 irq_context: 0 &sb->s_type->i_mutex_key#8 swapon_mutex swap_lock irq_context: 0 &sb->s_type->i_mutex_key#8 swapon_mutex swap_lock &p->lock#2 irq_context: 0 &sb->s_type->i_mutex_key#8 swapon_mutex percpu_ref_switch_lock irq_context: 0 &sb->s_type->i_mutex_key#8 swapon_mutex swap_lock &p->lock#2 swap_avail_lock irq_context: 0 &sb->s_type->i_mutex_key#8 swapon_mutex (console_sem).lock irq_context: 0 &sb->s_type->i_mutex_key#8 swapon_mutex console_lock console_srcu console_owner_lock irq_context: 0 &sb->s_type->i_mutex_key#8 swapon_mutex console_lock console_srcu console_owner irq_context: 0 &sb->s_type->i_mutex_key#8 swapon_mutex console_lock console_srcu console_owner &port_lock_key irq_context: 0 &sb->s_type->i_mutex_key#8 swapon_mutex console_lock console_srcu console_owner console_owner_lock irq_context: 0 &sb->s_type->i_mutex_key#8 proc_poll_wait.lock irq_context: 0 swap_slots_cache_enable_mutex irq_context: 0 swap_slots_cache_enable_mutex cpu_hotplug_lock irq_context: 0 swap_slots_cache_enable_mutex cpu_hotplug_lock cpuhp_state_mutex irq_context: 0 swap_slots_cache_enable_mutex cpu_hotplug_lock cpuhp_state_mutex cpuhp_state-down irq_context: 0 swap_slots_cache_enable_mutex cpu_hotplug_lock cpuhp_state_mutex cpuhp_state-up irq_context: 0 swap_slots_cache_enable_mutex cpu_hotplug_lock cpuhp_state_mutex &p->pi_lock irq_context: 0 swap_slots_cache_enable_mutex cpu_hotplug_lock cpuhp_state_mutex &p->pi_lock &rq->__lock irq_context: 0 swap_slots_cache_enable_mutex cpu_hotplug_lock cpuhp_state_mutex &p->pi_lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 swap_slots_cache_enable_mutex cpu_hotplug_lock cpuhp_state_mutex &x->wait#6 irq_context: 0 swap_slots_cache_enable_mutex cpu_hotplug_lock cpuhp_state_mutex &rq->__lock irq_context: 0 swap_slots_cache_enable_mutex cpu_hotplug_lock cpuhp_state_mutex &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 cpu_hotplug_lock cpuhp_state-up swap_slots_cache_mutex irq_context: 0 swap_slots_cache_enable_mutex swap_lock irq_context: 0 &sighand->siglock rcu_read_lock &____s->seqcount#5 irq_context: 0 &sighand->siglock &prev->lock irq_context: 0 sk_lock-AF_INET rcu_read_lock &ei->socket.wq.wait irq_context: 0 sk_lock-AF_INET rcu_read_lock &ei->socket.wq.wait &ep->lock irq_context: softirq rcu_read_lock rcu_read_lock slock-AF_INET/1 rcu_read_lock &ei->socket.wq.wait &ep->lock &ep->wq &p->pi_lock &cfs_rq->removed.lock irq_context: softirq rcu_read_lock rcu_read_lock slock-AF_INET/1 &pcp->lock &zone->lock irq_context: softirq rcu_read_lock rcu_read_lock slock-AF_INET/1 &pcp->lock &zone->lock &____s->seqcount irq_context: 0 sk_lock-AF_INET &mm->mmap_lock &rq->__lock irq_context: 0 sk_lock-AF_INET remove_cache_srcu &rq->__lock irq_context: softirq rcu_read_lock rcu_read_lock slock-AF_INET/1 &zone->lock irq_context: softirq rcu_read_lock rcu_read_lock slock-AF_INET/1 &zone->lock &____s->seqcount irq_context: 0 sk_lock-AF_INET rcu_read_lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: softirq _xmit_ETHER#2 quarantine_lock irq_context: 0 sk_lock-AF_INET &rq->__lock &cfs_rq->removed.lock irq_context: 0 (wq_completion)events_unbound (work_completion)(&(&kfence_timer)->work) rcu_read_lock &stopper->lock irq_context: 0 (wq_completion)events_unbound (work_completion)(&(&kfence_timer)->work) rcu_read_lock &stop_pi_lock irq_context: 0 (wq_completion)events_unbound (work_completion)(&(&kfence_timer)->work) rcu_read_lock &stop_pi_lock &rq->__lock irq_context: 0 (wq_completion)events_unbound (work_completion)(&(&kfence_timer)->work) rcu_read_lock &stop_pi_lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: softirq (&lapb->t1timer) &lapb->lock rcu_read_lock_bh dev->qdisc_tx_busylock ?: &qdisc_tx_busylock _xmit_ETHER#2 rcu_read_lock &____s->seqcount irq_context: softirq slock-AF_INET rcu_read_lock rcu_read_lock rcu_read_lock_bh dev->qdisc_tx_busylock ?: &qdisc_tx_busylock _xmit_ETHER#2 rcu_read_lock batched_entropy_u8.lock irq_context: softirq slock-AF_INET rcu_read_lock rcu_read_lock rcu_read_lock_bh dev->qdisc_tx_busylock ?: &qdisc_tx_busylock _xmit_ETHER#2 rcu_read_lock kfence_freelist_lock irq_context: softirq slock-AF_INET rcu_read_lock rcu_read_lock rcu_read_lock_bh dev->qdisc_tx_busylock ?: &qdisc_tx_busylock _xmit_ETHER#2 rcu_read_lock &meta->lock irq_context: softirq rcu_read_lock rcu_read_lock slock-AF_INET/1 rcu_read_lock rcu_read_lock rcu_read_lock_bh dev->qdisc_tx_busylock ?: &qdisc_tx_busylock _xmit_ETHER#2 quarantine_lock irq_context: softirq rcu_read_lock rcu_read_lock slock-AF_INET/1 rcu_read_lock rcu_read_lock rcu_read_lock_bh dev->qdisc_tx_busylock ?: &qdisc_tx_busylock _xmit_ETHER#2 rcu_read_lock batched_entropy_u8.lock irq_context: softirq rcu_read_lock rcu_read_lock slock-AF_INET/1 rcu_read_lock rcu_read_lock rcu_read_lock_bh dev->qdisc_tx_busylock ?: &qdisc_tx_busylock _xmit_ETHER#2 rcu_read_lock kfence_freelist_lock irq_context: softirq rcu_read_lock rcu_read_lock slock-AF_INET/1 rcu_read_lock rcu_read_lock rcu_read_lock_bh dev->qdisc_tx_busylock ?: &qdisc_tx_busylock _xmit_ETHER#2 rcu_read_lock &meta->lock irq_context: softirq _xmit_ETHER#2 &meta->lock irq_context: softirq _xmit_ETHER#2 kfence_freelist_lock irq_context: 0 sk_lock-AF_INET rcu_read_lock rcu_read_lock rcu_read_lock_bh dev->qdisc_tx_busylock ?: &qdisc_tx_busylock _xmit_ETHER#2 rcu_read_lock &pcp->lock &zone->lock irq_context: 0 sk_lock-AF_INET rcu_read_lock rcu_read_lock rcu_read_lock_bh dev->qdisc_tx_busylock ?: &qdisc_tx_busylock _xmit_ETHER#2 rcu_read_lock &pcp->lock &zone->lock &____s->seqcount irq_context: softirq &(&ovs_net->masks_rebalance)->timer rcu_read_lock &pool->lock &p->pi_lock &rq->__lock &cfs_rq->removed.lock irq_context: softirq slock-AF_INET rcu_read_lock rcu_read_lock rcu_read_lock_bh dev->qdisc_tx_busylock ?: &qdisc_tx_busylock _xmit_ETHER#2 rcu_read_lock &c->lock irq_context: 0 sk_lock-AF_INET rcu_read_lock rcu_read_lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 &vma->vm_lock->lock &p->pi_lock irq_context: 0 &vma->vm_lock->lock &p->pi_lock &rq->__lock irq_context: 0 &vma->vm_lock->lock &p->pi_lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 sk_lock-AF_INET rcu_read_lock &ei->socket.wq.wait &ep->lock &ep->wq irq_context: 0 sk_lock-AF_INET rcu_read_lock &ei->socket.wq.wait &ep->lock &ep->wq &p->pi_lock irq_context: 0 sk_lock-AF_INET rcu_read_lock &ei->socket.wq.wait &ep->lock &ep->wq &p->pi_lock &rq->__lock irq_context: 0 sk_lock-AF_INET rcu_read_lock &ei->socket.wq.wait &ep->lock &ep->wq &p->pi_lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 sk_lock-AF_INET &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: softirq (&journal->j_commit_timer) &p->pi_lock &rq->__lock &cfs_rq->removed.lock irq_context: 0 sk_lock-AF_INET rcu_read_lock rcu_read_lock rcu_read_lock_bh dev->qdisc_tx_busylock ?: &qdisc_tx_busylock _xmit_ETHER#2 rcu_read_lock batched_entropy_u8.lock irq_context: 0 sk_lock-AF_INET rcu_read_lock rcu_read_lock rcu_read_lock_bh dev->qdisc_tx_busylock ?: &qdisc_tx_busylock _xmit_ETHER#2 rcu_read_lock kfence_freelist_lock irq_context: 0 sk_lock-AF_INET rcu_read_lock rcu_read_lock rcu_read_lock_bh dev->qdisc_tx_busylock ?: &qdisc_tx_busylock _xmit_ETHER#2 rcu_read_lock &meta->lock irq_context: softirq rcu_read_lock rcu_read_lock slock-AF_INET/1 rcu_read_lock rcu_read_lock rcu_read_lock_bh dev->qdisc_tx_busylock ?: &qdisc_tx_busylock _xmit_ETHER#2 rcu_read_lock quarantine_lock irq_context: softirq rcu_read_lock rcu_read_lock slock-AF_INET/1 quarantine_lock irq_context: 0 &rq->__lock rcu_read_lock &cfs_rq->removed.lock irq_context: softirq rcu_read_lock rcu_read_lock slock-AF_INET/1 batched_entropy_u8.lock irq_context: softirq rcu_read_lock rcu_read_lock slock-AF_INET/1 rcu_read_lock rcu_read_lock rcu_read_lock_bh dev->qdisc_tx_busylock ?: &qdisc_tx_busylock _xmit_ETHER#2 &meta->lock irq_context: softirq rcu_read_lock rcu_read_lock slock-AF_INET/1 rcu_read_lock rcu_read_lock rcu_read_lock_bh dev->qdisc_tx_busylock ?: &qdisc_tx_busylock _xmit_ETHER#2 kfence_freelist_lock irq_context: 0 (wq_completion)writeback (work_completion)(&(&wb->dwork)->work) &type->s_umount_key#32 &sbi->s_writepages_rwsem mmu_notifier_invalidate_range_start irq_context: 0 (wq_completion)writeback (work_completion)(&(&wb->dwork)->work) &type->s_umount_key#32 &sbi->s_writepages_rwsem &____s->seqcount irq_context: 0 (wq_completion)writeback (work_completion)(&(&wb->dwork)->work) &type->s_umount_key#32 &sbi->s_writepages_rwsem pool_lock#2 irq_context: 0 (wq_completion)writeback (work_completion)(&(&wb->dwork)->work) &type->s_umount_key#32 &sbi->s_writepages_rwsem &c->lock irq_context: 0 (wq_completion)writeback (work_completion)(&(&wb->dwork)->work) &type->s_umount_key#32 &sbi->s_writepages_rwsem lock#4 irq_context: 0 (wq_completion)writeback (work_completion)(&(&wb->dwork)->work) &type->s_umount_key#32 &sbi->s_writepages_rwsem lock#4 &lruvec->lru_lock irq_context: 0 (wq_completion)writeback (work_completion)(&(&wb->dwork)->work) &type->s_umount_key#32 &sbi->s_writepages_rwsem lock#5 irq_context: 0 (wq_completion)writeback (work_completion)(&(&wb->dwork)->work) &type->s_umount_key#32 &sbi->s_writepages_rwsem &obj_hash[i].lock irq_context: 0 (wq_completion)writeback (work_completion)(&(&wb->dwork)->work) &type->s_umount_key#32 &sbi->s_writepages_rwsem &journal->j_state_lock irq_context: 0 (wq_completion)writeback (work_completion)(&(&wb->dwork)->work) &type->s_umount_key#32 &sbi->s_writepages_rwsem &journal->j_state_lock irq_context: 0 (wq_completion)writeback (work_completion)(&(&wb->dwork)->work) &type->s_umount_key#32 &sbi->s_writepages_rwsem &journal->j_state_lock tk_core.seq.seqcount irq_context: 0 (wq_completion)writeback (work_completion)(&(&wb->dwork)->work) &type->s_umount_key#32 &sbi->s_writepages_rwsem &journal->j_state_lock &obj_hash[i].lock irq_context: 0 (wq_completion)writeback (work_completion)(&(&wb->dwork)->work) &type->s_umount_key#32 &sbi->s_writepages_rwsem &journal->j_state_lock &base->lock irq_context: 0 (wq_completion)writeback (work_completion)(&(&wb->dwork)->work) &type->s_umount_key#32 &sbi->s_writepages_rwsem &journal->j_state_lock &base->lock &obj_hash[i].lock irq_context: 0 (wq_completion)writeback (work_completion)(&(&wb->dwork)->work) &type->s_umount_key#32 &sbi->s_writepages_rwsem &rq->__lock irq_context: 0 (wq_completion)writeback (work_completion)(&(&wb->dwork)->work) &type->s_umount_key#32 &sbi->s_writepages_rwsem jbd2_handle irq_context: 0 (wq_completion)writeback (work_completion)(&(&wb->dwork)->work) &type->s_umount_key#32 &sbi->s_writepages_rwsem jbd2_handle lock#4 irq_context: 0 (wq_completion)writeback (work_completion)(&(&wb->dwork)->work) &type->s_umount_key#32 &sbi->s_writepages_rwsem jbd2_handle lock#4 &lruvec->lru_lock irq_context: 0 (wq_completion)writeback (work_completion)(&(&wb->dwork)->work) &type->s_umount_key#32 &sbi->s_writepages_rwsem jbd2_handle lock#5 irq_context: 0 (wq_completion)writeback (work_completion)(&(&wb->dwork)->work) &type->s_umount_key#32 &sbi->s_writepages_rwsem jbd2_handle mmu_notifier_invalidate_range_start irq_context: 0 (wq_completion)writeback (work_completion)(&(&wb->dwork)->work) &type->s_umount_key#32 &sbi->s_writepages_rwsem jbd2_handle pool_lock#2 irq_context: 0 (wq_completion)writeback (work_completion)(&(&wb->dwork)->work) &type->s_umount_key#32 &sbi->s_writepages_rwsem jbd2_handle &ei->i_es_lock irq_context: 0 (wq_completion)writeback (work_completion)(&(&wb->dwork)->work) &type->s_umount_key#32 &sbi->s_writepages_rwsem jbd2_handle &ei->i_data_sem irq_context: 0 (wq_completion)writeback (work_completion)(&(&wb->dwork)->work) &type->s_umount_key#32 &sbi->s_writepages_rwsem jbd2_handle &ei->i_data_sem mmu_notifier_invalidate_range_start irq_context: 0 (wq_completion)writeback (work_completion)(&(&wb->dwork)->work) &type->s_umount_key#32 &sbi->s_writepages_rwsem jbd2_handle &ei->i_data_sem pool_lock#2 irq_context: 0 (wq_completion)writeback (work_completion)(&(&wb->dwork)->work) &type->s_umount_key#32 &sbi->s_writepages_rwsem jbd2_handle &ei->i_data_sem &lg->lg_mutex irq_context: 0 (wq_completion)writeback (work_completion)(&(&wb->dwork)->work) &type->s_umount_key#32 &sbi->s_writepages_rwsem jbd2_handle &ei->i_data_sem &lg->lg_mutex &ei->i_prealloc_lock irq_context: 0 (wq_completion)writeback (work_completion)(&(&wb->dwork)->work) &type->s_umount_key#32 &sbi->s_writepages_rwsem jbd2_handle &ei->i_data_sem &lg->lg_mutex mmu_notifier_invalidate_range_start irq_context: 0 (wq_completion)writeback (work_completion)(&(&wb->dwork)->work) &type->s_umount_key#32 &sbi->s_writepages_rwsem jbd2_handle &ei->i_data_sem &lg->lg_mutex pool_lock#2 irq_context: 0 (wq_completion)writeback (work_completion)(&(&wb->dwork)->work) &type->s_umount_key#32 &sbi->s_writepages_rwsem jbd2_handle &ei->i_data_sem &lg->lg_mutex &mapping->private_lock irq_context: 0 (wq_completion)writeback (work_completion)(&(&wb->dwork)->work) &type->s_umount_key#32 &sbi->s_writepages_rwsem jbd2_handle &ei->i_data_sem &lg->lg_mutex &ret->b_state_lock irq_context: 0 (wq_completion)writeback (work_completion)(&(&wb->dwork)->work) &type->s_umount_key#32 &sbi->s_writepages_rwsem jbd2_handle &ei->i_data_sem &lg->lg_mutex &ret->b_state_lock &journal->j_list_lock irq_context: 0 (wq_completion)writeback (work_completion)(&(&wb->dwork)->work) &type->s_umount_key#32 &sbi->s_writepages_rwsem jbd2_handle &ei->i_data_sem &lg->lg_mutex &journal->j_revoke_lock irq_context: 0 (wq_completion)writeback (work_completion)(&(&wb->dwork)->work) &type->s_umount_key#32 &sbi->s_writepages_rwsem jbd2_handle &ei->i_data_sem &lg->lg_mutex &pa->pa_lock irq_context: 0 (wq_completion)writeback (work_completion)(&(&wb->dwork)->work) &type->s_umount_key#32 &sbi->s_writepages_rwsem jbd2_handle &ei->i_data_sem &lg->lg_mutex &lg->lg_prealloc_lock irq_context: 0 (wq_completion)writeback (work_completion)(&(&wb->dwork)->work) &type->s_umount_key#32 &sbi->s_writepages_rwsem jbd2_handle &ei->i_data_sem &obj_hash[i].lock irq_context: 0 (wq_completion)writeback (work_completion)(&(&wb->dwork)->work) &type->s_umount_key#32 &sbi->s_writepages_rwsem jbd2_handle &ei->i_data_sem &ret->b_state_lock irq_context: 0 (wq_completion)writeback (work_completion)(&(&wb->dwork)->work) &type->s_umount_key#32 &sbi->s_writepages_rwsem jbd2_handle &ei->i_data_sem &ret->b_state_lock &journal->j_list_lock irq_context: 0 (wq_completion)writeback (work_completion)(&(&wb->dwork)->work) &type->s_umount_key#32 &sbi->s_writepages_rwsem jbd2_handle &ei->i_data_sem &ei->i_raw_lock irq_context: 0 (wq_completion)writeback (work_completion)(&(&wb->dwork)->work) &type->s_umount_key#32 &sbi->s_writepages_rwsem jbd2_handle &ei->i_data_sem &(ei->i_block_reservation_lock) irq_context: 0 (wq_completion)writeback (work_completion)(&(&wb->dwork)->work) &type->s_umount_key#32 &sbi->s_writepages_rwsem jbd2_handle &ei->i_data_sem &sb->s_type->i_lock_key#22 irq_context: 0 (wq_completion)writeback (work_completion)(&(&wb->dwork)->work) &type->s_umount_key#32 &sbi->s_writepages_rwsem jbd2_handle &ei->i_data_sem &ei->i_prealloc_lock irq_context: 0 (wq_completion)writeback (work_completion)(&(&wb->dwork)->work) &type->s_umount_key#32 &sbi->s_writepages_rwsem jbd2_handle &ei->i_data_sem &ei->i_es_lock irq_context: 0 (wq_completion)writeback (work_completion)(&(&wb->dwork)->work) &type->s_umount_key#32 &sbi->s_writepages_rwsem jbd2_handle &ei->i_data_sem &ei->i_es_lock irq_context: 0 (wq_completion)writeback (work_completion)(&(&wb->dwork)->work) &type->s_umount_key#32 &sbi->s_writepages_rwsem jbd2_handle &ei->i_data_sem &ei->i_es_lock &obj_hash[i].lock irq_context: 0 (wq_completion)writeback (work_completion)(&(&wb->dwork)->work) &type->s_umount_key#32 &sbi->s_writepages_rwsem jbd2_handle &ei->i_data_sem &ei->i_es_lock pool_lock#2 irq_context: 0 (wq_completion)writeback (work_completion)(&(&wb->dwork)->work) &type->s_umount_key#32 &sbi->s_writepages_rwsem jbd2_handle rcu_read_lock &memcg->move_lock irq_context: 0 (wq_completion)writeback (work_completion)(&(&wb->dwork)->work) &type->s_umount_key#32 &sbi->s_writepages_rwsem jbd2_handle rcu_read_lock &xa->xa_lock#9 irq_context: 0 (wq_completion)writeback (work_completion)(&(&wb->dwork)->work) &type->s_umount_key#32 &sbi->s_writepages_rwsem jbd2_handle rcu_read_lock &xa->xa_lock#9 &s->s_inode_wblist_lock irq_context: 0 (wq_completion)writeback (work_completion)(&(&wb->dwork)->work) &type->s_umount_key#32 &sbi->s_writepages_rwsem jbd2_handle &ei->i_raw_lock irq_context: 0 (wq_completion)writeback (work_completion)(&(&wb->dwork)->work) &type->s_umount_key#32 &sbi->s_writepages_rwsem jbd2_handle &journal->j_wait_updates irq_context: 0 (wq_completion)writeback (work_completion)(&(&wb->dwork)->work) &type->s_umount_key#32 &sbi->s_writepages_rwsem tk_core.seq.seqcount irq_context: 0 (wq_completion)writeback (work_completion)(&(&wb->dwork)->work) &type->s_umount_key#32 &sbi->s_writepages_rwsem &dd->lock irq_context: 0 (wq_completion)writeback (work_completion)(&(&wb->dwork)->work) &type->s_umount_key#32 &sbi->s_writepages_rwsem &base->lock irq_context: 0 (wq_completion)writeback (work_completion)(&(&wb->dwork)->work) &type->s_umount_key#32 &sbi->s_writepages_rwsem &base->lock &obj_hash[i].lock irq_context: 0 (wq_completion)ext4-rsv-conversion (work_completion)(&ei->i_rsv_conversion_work) jbd2_handle &ei->i_data_sem &c->lock irq_context: 0 (wq_completion)ext4-rsv-conversion (work_completion)(&ei->i_rsv_conversion_work) jbd2_handle &ei->i_data_sem &ei->i_es_lock &obj_hash[i].lock irq_context: 0 sk_lock-AF_INET quarantine_lock irq_context: softirq &(&wb->bw_dwork)->timer rcu_read_lock &pool->lock/1 &p->pi_lock &rq->__lock &cfs_rq->removed.lock irq_context: 0 (wq_completion)events_long (work_completion)(&(&ipvs->defense_work)->work) &rq->__lock irq_context: softirq rcu_read_lock rcu_read_lock slock-AF_INET/1 rcu_read_lock &ei->socket.wq.wait &ep->lock &ep->wq &p->pi_lock &rq->__lock &cfs_rq->removed.lock irq_context: 0 &vma->vm_lock->lock fs_reclaim &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 &mm->mmap_lock &anon_vma->rwsem ptlock_ptr(page)#2 irq_context: 0 &mm->mmap_lock &anon_vma->rwsem ptlock_ptr(page)#2 &lruvec->lru_lock irq_context: 0 &mm->mmap_lock ptlock_ptr(page) lock#4 irq_context: 0 &mm->mmap_lock ptlock_ptr(page) lock#4 &lruvec->lru_lock irq_context: 0 &mm->mmap_lock ptlock_ptr(page) lock#4 rcu_read_lock pool_lock#2 irq_context: 0 &mm->mmap_lock ptlock_ptr(page) lock#4 &obj_hash[i].lock irq_context: 0 &vma->vm_lock->lock mmu_notifier_invalidate_range_start &rq->__lock irq_context: 0 sk_lock-AF_INET rcu_node_0 irq_context: 0 (wq_completion)events_freezable (work_completion)(&vb->update_balloon_stats_work) &rq->__lock irq_context: 0 sk_lock-AF_INET remove_cache_srcu &pcp->lock &zone->lock irq_context: 0 sk_lock-AF_INET remove_cache_srcu &pcp->lock &zone->lock &____s->seqcount irq_context: softirq rcu_read_lock rcu_read_lock slock-AF_INET/1 rcu_read_lock rcu_read_lock rcu_read_lock_bh dev->qdisc_tx_busylock ?: &qdisc_tx_busylock _xmit_ETHER#2 rcu_read_lock &pcp->lock &zone->lock irq_context: softirq rcu_read_lock rcu_read_lock slock-AF_INET/1 rcu_read_lock rcu_read_lock rcu_read_lock_bh dev->qdisc_tx_busylock ?: &qdisc_tx_busylock _xmit_ETHER#2 rcu_read_lock &pcp->lock &zone->lock &____s->seqcount irq_context: 0 sk_lock-AF_INET &mm->mmap_lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: softirq (&q->timeout) rcu_read_lock &pool->lock &p->pi_lock &rq->__lock &cfs_rq->removed.lock irq_context: softirq (&timer) irq_context: softirq (&timer) &obj_hash[i].lock irq_context: softirq (&timer) &base->lock irq_context: softirq (&timer) &base->lock &obj_hash[i].lock irq_context: softirq (&timer) rcu_read_lock pool_lock#2 irq_context: softirq (&timer) rcu_read_lock &c->lock irq_context: softirq (&timer) rcu_read_lock &____s->seqcount irq_context: softirq (&timer) &txlock irq_context: softirq (&timer) &txlock &list->lock#3 irq_context: softirq (&timer) &txwq irq_context: softirq (&timer) &txwq &p->pi_lock irq_context: softirq (&timer) &txwq &p->pi_lock &rq->__lock irq_context: softirq (&timer) &txwq &p->pi_lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 rcu_read_lock_bh tk_core.seq.seqcount irq_context: 0 rcu_read_lock_bh &list->lock#5 irq_context: softirq &list->lock#5 irq_context: 0 rcu_read_lock_bh rcu_read_lock &____s->seqcount irq_context: 0 rcu_read_lock_bh dev->qdisc_tx_busylock ?: &qdisc_tx_busylock _xmit_SLIP#2 irq_context: 0 rcu_read_lock_bh dev->qdisc_tx_busylock ?: &qdisc_tx_busylock _xmit_SLIP#2 &eql->queue.lock irq_context: 0 rcu_read_lock_bh dev->qdisc_tx_busylock ?: &qdisc_tx_busylock _xmit_SLIP#2 &eql->queue.lock &obj_hash[i].lock irq_context: 0 rcu_read_lock_bh dev->qdisc_tx_busylock ?: &qdisc_tx_busylock _xmit_SLIP#2 &eql->queue.lock &____s->seqcount irq_context: 0 rcu_read_lock_bh dev->qdisc_tx_busylock ?: &qdisc_tx_busylock _xmit_SLIP#2 &eql->queue.lock pool_lock#2 irq_context: 0 rcu_read_lock_bh dev->qdisc_tx_busylock ?: &qdisc_tx_busylock _xmit_SLIP#2 &eql->queue.lock rcu_read_lock pool_lock#2 irq_context: 0 rcu_read_lock_bh dev->qdisc_tx_busylock ?: &qdisc_tx_busylock _xmit_NETROM irq_context: 0 rcu_read_lock_bh dev->qdisc_tx_busylock ?: &qdisc_tx_busylock _xmit_NETROM (console_sem).lock irq_context: 0 rcu_read_lock_bh dev->qdisc_tx_busylock ?: &qdisc_tx_busylock _xmit_NETROM console_lock console_srcu console_owner_lock irq_context: 0 rcu_read_lock_bh dev->qdisc_tx_busylock ?: &qdisc_tx_busylock _xmit_NETROM console_lock console_srcu console_owner irq_context: 0 rcu_read_lock_bh dev->qdisc_tx_busylock ?: &qdisc_tx_busylock _xmit_NETROM console_lock console_srcu console_owner &port_lock_key irq_context: 0 rcu_read_lock_bh dev->qdisc_tx_busylock ?: &qdisc_tx_busylock _xmit_NETROM console_lock console_srcu console_owner console_owner_lock irq_context: 0 rcu_read_lock_bh dev->qdisc_tx_busylock ?: &qdisc_tx_busylock _xmit_NETROM &obj_hash[i].lock irq_context: 0 rcu_read_lock_bh dev->qdisc_tx_busylock ?: &qdisc_tx_busylock _xmit_NETROM pool_lock#2 irq_context: 0 rcu_read_lock_bh _xmit_X25#2 irq_context: 0 rcu_read_lock_bh _xmit_X25#2 &lapbeth->up_lock irq_context: 0 rcu_read_lock_bh _xmit_X25#2 &lapbeth->up_lock &obj_hash[i].lock irq_context: 0 rcu_read_lock_bh _xmit_X25#2 &lapbeth->up_lock pool_lock#2 irq_context: softirq &c->lock batched_entropy_u8.lock irq_context: softirq &c->lock kfence_freelist_lock irq_context: 0 (wq_completion)writeback (work_completion)(&(&wb->dwork)->work) &type->s_umount_key#43 rcu_read_lock &xa->xa_lock#9 &pl->lock irq_context: 0 (wq_completion)writeback (work_completion)(&(&wb->dwork)->work) &type->s_umount_key#43 rcu_read_lock &xa->xa_lock#9 &pl->lock key#12 irq_context: 0 (wq_completion)writeback (work_completion)(&(&wb->dwork)->work) &type->s_umount_key#43 &____s->seqcount irq_context: 0 (wq_completion)writeback (work_completion)(&(&wb->dwork)->work) &type->s_umount_key#43 &rq_wait->wait irq_context: 0 (wq_completion)writeback (work_completion)(&(&wb->dwork)->work) &type->s_umount_key#43 rcu_read_lock &pool->lock irq_context: 0 (wq_completion)writeback (work_completion)(&(&wb->dwork)->work) &type->s_umount_key#43 rcu_read_lock &pool->lock &obj_hash[i].lock irq_context: 0 (wq_completion)writeback (work_completion)(&(&wb->dwork)->work) &type->s_umount_key#43 rcu_read_lock &pool->lock pool_lock#2 irq_context: 0 (wq_completion)writeback (work_completion)(&(&wb->dwork)->work) &type->s_umount_key#43 rcu_read_lock &pool->lock &p->pi_lock irq_context: 0 (wq_completion)writeback (work_completion)(&(&wb->dwork)->work) &type->s_umount_key#43 rcu_read_lock &pool->lock &p->pi_lock &rq->__lock irq_context: 0 (wq_completion)writeback (work_completion)(&(&wb->dwork)->work) &type->s_umount_key#43 rcu_read_lock &pool->lock &p->pi_lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)writeback (work_completion)(&(&wb->dwork)->work) &type->s_umount_key#43 rcu_node_0 irq_context: 0 (wq_completion)writeback (work_completion)(&(&wb->dwork)->work) &type->s_umount_key#43 &rq->__lock irq_context: 0 (wq_completion)writeback (work_completion)(&(&wb->dwork)->work) &type->s_umount_key#43 rcu_read_lock rcu_node_0 irq_context: 0 (wq_completion)writeback (work_completion)(&(&wb->dwork)->work) &type->s_umount_key#43 &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: softirq &rq_wait->wait irq_context: softirq &rq_wait->wait &p->pi_lock irq_context: softirq &rq_wait->wait &p->pi_lock &rq->__lock irq_context: softirq &rq_wait->wait &p->pi_lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: softirq &rq_wait->wait &p->pi_lock &cfs_rq->removed.lock irq_context: 0 (wq_completion)writeback (work_completion)(&(&wb->dwork)->work) &type->s_umount_key#43 &n->list_lock irq_context: 0 (wq_completion)writeback (work_completion)(&(&wb->dwork)->work) &type->s_umount_key#43 &n->list_lock &c->lock irq_context: softirq rcu_read_lock &xa->xa_lock#9 key#13 irq_context: 0 &journal->j_list_lock key#15 irq_context: softirq (&n->timer) irq_context: softirq (&n->timer) &n->lock irq_context: softirq (&n->timer) &n->lock &obj_hash[i].lock irq_context: softirq (&n->timer) &n->lock &base->lock irq_context: softirq (&n->timer) &n->lock &base->lock &obj_hash[i].lock irq_context: softirq mm/memcontrol.c:589 rcu_read_lock &pool->lock/1 &p->pi_lock &rq->__lock &cfs_rq->removed.lock irq_context: softirq rcu_read_lock rcu_read_lock slock-AF_INET/1 batched_entropy_u8.lock crngs.lock irq_context: softirq rcu_read_lock rcu_read_lock slock-AF_INET/1 batched_entropy_u8.lock crngs.lock base_crng.lock irq_context: 0 &ep->mtx &mm->mmap_lock ptlock_ptr(page)#2 irq_context: 0 &ep->mtx &mm->mmap_lock &p->pi_lock irq_context: 0 &ep->mtx &mm->mmap_lock &p->pi_lock &rq->__lock irq_context: 0 &ep->mtx &mm->mmap_lock &p->pi_lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: softirq &(&gc_work->dwork)->timer rcu_read_lock &pool->lock &p->pi_lock &rq->__lock &cfs_rq->removed.lock irq_context: 0 (wq_completion)events_power_efficient (work_completion)(&(&gc_work->dwork)->work) &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)writeback (work_completion)(&(&wb->dwork)->work) &type->s_umount_key#32 &sbi->s_writepages_rwsem jbd2_handle &____s->seqcount irq_context: 0 (wq_completion)writeback (work_completion)(&(&wb->dwork)->work) &type->s_umount_key#32 &sbi->s_writepages_rwsem jbd2_handle &c->lock irq_context: 0 (wq_completion)writeback (work_completion)(&(&wb->dwork)->work) &type->s_umount_key#32 &sbi->s_writepages_rwsem jbd2_handle &ei->i_data_sem &____s->seqcount irq_context: 0 (wq_completion)writeback (work_completion)(&(&wb->dwork)->work) &type->s_umount_key#32 &sbi->s_writepages_rwsem jbd2_handle &ei->i_data_sem &c->lock irq_context: 0 (wq_completion)writeback (work_completion)(&(&wb->dwork)->work) &type->s_umount_key#32 &sbi->s_writepages_rwsem jbd2_handle &ei->i_data_sem &lg->lg_mutex &____s->seqcount irq_context: 0 (wq_completion)writeback (work_completion)(&(&wb->dwork)->work) &type->s_umount_key#32 &sbi->s_writepages_rwsem jbd2_handle &ei->i_data_sem &lg->lg_mutex &c->lock irq_context: 0 (wq_completion)writeback (work_completion)(&(&wb->dwork)->work) &type->s_umount_key#32 &sbi->s_writepages_rwsem jbd2_handle &ei->i_data_sem &mapping->private_lock irq_context: 0 (wq_completion)ext4-rsv-conversion (work_completion)(&ei->i_rsv_conversion_work) rcu_read_lock rcu_read_lock per_cpu_ptr(&cgroup_rstat_cpu_lock, cpu) irq_context: 0 rcu_read_lock &sighand->siglock batched_entropy_u8.lock crngs.lock irq_context: 0 rcu_read_lock &sighand->siglock batched_entropy_u8.lock crngs.lock base_crng.lock irq_context: 0 sk_lock-AF_INET remove_cache_srcu rcu_read_lock rcu_node_0 irq_context: 0 sk_lock-AF_INET remove_cache_srcu rcu_read_lock &rq->__lock irq_context: 0 sk_lock-AF_INET remove_cache_srcu rcu_read_lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 sk_lock-AF_INET rcu_read_lock &ei->socket.wq.wait &ep->lock &ep->wq &p->pi_lock &cfs_rq->removed.lock irq_context: 0 sk_lock-AF_INET fill_pool_map-wait-type-override &rq->__lock irq_context: 0 lock#3 rcu_read_lock (wq_completion)mm_percpu_wq irq_context: 0 lock#3 &x->wait#10 irq_context: 0 lock#3 &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)mm_percpu_wq (work_completion)(&barr->work) irq_context: 0 (wq_completion)mm_percpu_wq (work_completion)(&barr->work) &x->wait#10 irq_context: 0 (wq_completion)mm_percpu_wq (work_completion)(&barr->work) &x->wait#10 &p->pi_lock irq_context: 0 (wq_completion)events_unbound (work_completion)(&(&kfence_timer)->work) cpu_hotplug_lock jump_label_mutex &rq->__lock irq_context: 0 &vma->vm_lock->lock remove_cache_srcu irq_context: 0 &vma->vm_lock->lock remove_cache_srcu quarantine_lock irq_context: 0 &vma->vm_lock->lock remove_cache_srcu &c->lock irq_context: 0 &vma->vm_lock->lock remove_cache_srcu &n->list_lock irq_context: 0 &vma->vm_lock->lock remove_cache_srcu rcu_read_lock pool_lock#2 irq_context: 0 &vma->vm_lock->lock remove_cache_srcu &obj_hash[i].lock irq_context: softirq (&dom->period_timer) &p->sequence key#13 irq_context: softirq net/wireless/reg.c:236 irq_context: softirq net/wireless/reg.c:236 rcu_read_lock &pool->lock irq_context: softirq net/wireless/reg.c:236 rcu_read_lock &pool->lock &obj_hash[i].lock irq_context: softirq net/wireless/reg.c:236 rcu_read_lock &pool->lock &p->pi_lock irq_context: softirq net/wireless/reg.c:236 rcu_read_lock &pool->lock &p->pi_lock &rq->__lock irq_context: softirq net/wireless/reg.c:236 rcu_read_lock &pool->lock &p->pi_lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)events_power_efficient (reg_check_chans).work irq_context: 0 (wq_completion)events_power_efficient (reg_check_chans).work rtnl_mutex irq_context: 0 (wq_completion)events_power_efficient (reg_check_chans).work rtnl_mutex &rdev->wiphy.mtx irq_context: 0 (wq_completion)events_power_efficient (reg_check_chans).work rtnl_mutex &rdev->wiphy.mtx &wdev->mtx irq_context: 0 (wq_completion)events (work_completion)(&fw_work->work) umhelper_sem (&timer.timer) irq_context: 0 (wq_completion)events (work_completion)(&fw_work->work) umhelper_sem fw_lock &x->wait#23 irq_context: 0 (wq_completion)events (work_completion)(&fw_work->work) umhelper_sem dev_pm_qos_sysfs_mtx irq_context: 0 (wq_completion)events (work_completion)(&fw_work->work) umhelper_sem dev_pm_qos_sysfs_mtx &root->kernfs_rwsem irq_context: 0 (wq_completion)events (work_completion)(&fw_work->work) umhelper_sem dev_pm_qos_sysfs_mtx &root->kernfs_rwsem irq_context: 0 (wq_completion)events (work_completion)(&fw_work->work) umhelper_sem dev_pm_qos_sysfs_mtx dev_pm_qos_mtx irq_context: 0 (wq_completion)events (work_completion)(&fw_work->work) umhelper_sem &root->kernfs_rwsem kernfs_idr_lock irq_context: 0 (wq_completion)events (work_completion)(&fw_work->work) umhelper_sem &root->kernfs_rwsem &obj_hash[i].lock irq_context: 0 (wq_completion)events (work_completion)(&fw_work->work) umhelper_sem &root->kernfs_rwsem pool_lock#2 irq_context: 0 (wq_completion)events (work_completion)(&fw_work->work) umhelper_sem kernfs_idr_lock irq_context: 0 (wq_completion)events (work_completion)(&fw_work->work) umhelper_sem &k->k_lock klist_remove_lock irq_context: 0 (wq_completion)events (work_completion)(&fw_work->work) umhelper_sem subsys mutex#80 &k->k_lock klist_remove_lock irq_context: 0 (wq_completion)events (work_completion)(&fw_work->work) umhelper_sem deferred_probe_mutex irq_context: 0 (wq_completion)events (work_completion)(&fw_work->work) umhelper_sem device_links_lock irq_context: 0 (wq_completion)events (work_completion)(&fw_work->work) umhelper_sem mmu_notifier_invalidate_range_start irq_context: 0 (wq_completion)events (work_completion)(&fw_work->work) umhelper_sem uevent_sock_mutex mmu_notifier_invalidate_range_start irq_context: 0 (wq_completion)events (work_completion)(&fw_work->work) umhelper_sem uevent_sock_mutex pool_lock#2 irq_context: 0 (wq_completion)events (work_completion)(&fw_work->work) umhelper_sem uevent_sock_mutex nl_table_lock irq_context: 0 (wq_completion)events (work_completion)(&fw_work->work) umhelper_sem uevent_sock_mutex rlock-AF_NETLINK irq_context: 0 (wq_completion)events (work_completion)(&fw_work->work) umhelper_sem uevent_sock_mutex rcu_read_lock &ei->socket.wq.wait irq_context: 0 (wq_completion)events (work_completion)(&fw_work->work) umhelper_sem uevent_sock_mutex rcu_read_lock &ei->socket.wq.wait &ep->lock irq_context: 0 (wq_completion)events (work_completion)(&fw_work->work) umhelper_sem uevent_sock_mutex rcu_read_lock &ei->socket.wq.wait &ep->lock &ep->wq irq_context: 0 (wq_completion)events (work_completion)(&fw_work->work) umhelper_sem uevent_sock_mutex rcu_read_lock &ei->socket.wq.wait &ep->lock &ep->wq &p->pi_lock irq_context: 0 (wq_completion)events (work_completion)(&fw_work->work) umhelper_sem uevent_sock_mutex rcu_read_lock &ei->socket.wq.wait &ep->lock &ep->wq &p->pi_lock &cfs_rq->removed.lock irq_context: 0 (wq_completion)events (work_completion)(&fw_work->work) umhelper_sem uevent_sock_mutex nl_table_wait.lock irq_context: 0 (wq_completion)events (work_completion)(&fw_work->work) umhelper_sem uevent_sock_mutex &obj_hash[i].lock irq_context: 0 (wq_completion)events (work_completion)(&fw_work->work) umhelper_sem gdp_mutex sysfs_symlink_target_lock irq_context: 0 (wq_completion)events (work_completion)(&fw_work->work) umhelper_sem gdp_mutex &obj_hash[i].lock irq_context: 0 (wq_completion)events (work_completion)(&fw_work->work) fw_lock irq_context: 0 (wq_completion)events (work_completion)(&fw_work->work) &rq->__lock irq_context: 0 &ep->mtx fs_reclaim &rq->__lock irq_context: 0 (wq_completion)events (work_completion)(&fw_work->work) rtnl_mutex irq_context: 0 (wq_completion)events (work_completion)(&fw_work->work) rtnl_mutex reg_indoor_lock irq_context: 0 (wq_completion)events (work_completion)(&fw_work->work) rtnl_mutex &obj_hash[i].lock irq_context: 0 (wq_completion)events (work_completion)(&fw_work->work) rtnl_mutex pool_lock#2 irq_context: 0 (wq_completion)events (work_completion)(&fw_work->work) rtnl_mutex krc.lock irq_context: 0 (wq_completion)events (work_completion)(&fw_work->work) rtnl_mutex krc.lock &obj_hash[i].lock irq_context: 0 (wq_completion)events (work_completion)(&fw_work->work) rtnl_mutex krc.lock &base->lock irq_context: 0 (wq_completion)events (work_completion)(&fw_work->work) rtnl_mutex krc.lock &base->lock &obj_hash[i].lock irq_context: 0 (wq_completion)events (work_completion)(&fw_work->work) rtnl_mutex reg_requests_lock irq_context: 0 (wq_completion)events (work_completion)(&fw_work->work) rtnl_mutex reg_pending_beacons_lock irq_context: 0 (wq_completion)events (work_completion)(&fw_work->work) rtnl_mutex fs_reclaim irq_context: 0 (wq_completion)events (work_completion)(&fw_work->work) rtnl_mutex fs_reclaim mmu_notifier_invalidate_range_start irq_context: 0 (wq_completion)events (work_completion)(&fw_work->work) rtnl_mutex rcu_read_lock &pool->lock irq_context: 0 (wq_completion)events (work_completion)(&fw_work->work) rtnl_mutex rcu_read_lock &pool->lock &obj_hash[i].lock irq_context: 0 (wq_completion)events (work_completion)(&fw_work->work) rtnl_mutex rcu_read_lock &pool->lock &p->pi_lock irq_context: 0 (wq_completion)events (work_completion)(&fw_work->work) rtnl_mutex rcu_read_lock &pool->lock &p->pi_lock &rq->__lock irq_context: 0 (wq_completion)events (work_completion)(&fw_work->work) rtnl_mutex rcu_read_lock &pool->lock &p->pi_lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)events (work_completion)(&fw_work->work) rtnl_mutex &rq->__lock irq_context: 0 (wq_completion)events reg_work rtnl_mutex rtnl_mutex.wait_lock irq_context: 0 (wq_completion)events reg_work rtnl_mutex &pool->lock irq_context: 0 (wq_completion)events reg_work rtnl_mutex &rq->__lock irq_context: 0 (wq_completion)events reg_work rtnl_mutex &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)events (work_completion)(&fw_work->work) rtnl_mutex.wait_lock irq_context: 0 (wq_completion)events (work_completion)(&fw_work->work) &p->pi_lock irq_context: 0 (wq_completion)events (work_completion)(&fw_work->work) &p->pi_lock &rq->__lock irq_context: 0 (wq_completion)events (work_completion)(&fw_work->work) &p->pi_lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)events reg_work rtnl_mutex uevent_sock_mutex irq_context: 0 (wq_completion)events reg_work rtnl_mutex uevent_sock_mutex fs_reclaim irq_context: 0 (wq_completion)events reg_work rtnl_mutex uevent_sock_mutex fs_reclaim mmu_notifier_invalidate_range_start irq_context: 0 (wq_completion)events reg_work rtnl_mutex uevent_sock_mutex pool_lock#2 irq_context: 0 (wq_completion)events reg_work rtnl_mutex uevent_sock_mutex &c->lock irq_context: 0 (wq_completion)events reg_work rtnl_mutex uevent_sock_mutex &____s->seqcount irq_context: 0 (wq_completion)events reg_work rtnl_mutex uevent_sock_mutex nl_table_lock irq_context: 0 (wq_completion)events reg_work rtnl_mutex uevent_sock_mutex rlock-AF_NETLINK irq_context: 0 (wq_completion)events reg_work rtnl_mutex uevent_sock_mutex rcu_read_lock &ei->socket.wq.wait irq_context: 0 (wq_completion)events reg_work rtnl_mutex uevent_sock_mutex rcu_read_lock &ei->socket.wq.wait &ep->lock irq_context: 0 (wq_completion)events reg_work rtnl_mutex uevent_sock_mutex rcu_read_lock &ei->socket.wq.wait &ep->lock &ep->wq irq_context: 0 (wq_completion)events reg_work rtnl_mutex uevent_sock_mutex rcu_read_lock &ei->socket.wq.wait &ep->lock &ep->wq &p->pi_lock irq_context: 0 (wq_completion)events reg_work rtnl_mutex uevent_sock_mutex rcu_read_lock &ei->socket.wq.wait &ep->lock &ep->wq &p->pi_lock &rq->__lock irq_context: 0 inode_hash_lock &sb->s_type->i_lock_key#24 irq_context: 0 (wq_completion)events reg_work rtnl_mutex uevent_sock_mutex rcu_read_lock &ei->socket.wq.wait &ep->lock &ep->wq &p->pi_lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)events reg_work rtnl_mutex uevent_sock_mutex nl_table_wait.lock irq_context: 0 (wq_completion)events reg_work rtnl_mutex uevent_sock_mutex &obj_hash[i].lock irq_context: softirq &(&wb->dwork)->timer rcu_read_lock &pool->lock/1 &p->pi_lock &cfs_rq->removed.lock irq_context: 0 (wq_completion)writeback (work_completion)(&(&wb->dwork)->work) &type->s_umount_key#32 &sbi->s_writepages_rwsem jbd2_handle &ei->i_data_sem &ei->i_prealloc_lock irq_context: 0 (wq_completion)writeback (work_completion)(&(&wb->dwork)->work) &type->s_umount_key#32 &sbi->s_writepages_rwsem jbd2_handle &ei->i_data_sem &sbi->s_md_lock irq_context: 0 (wq_completion)writeback (work_completion)(&(&wb->dwork)->work) &type->s_umount_key#32 &sbi->s_writepages_rwsem jbd2_handle &ei->i_data_sem &bgl->locks[i].lock &sbi->s_md_lock irq_context: 0 (wq_completion)writeback (work_completion)(&(&wb->dwork)->work) &type->s_umount_key#32 &sbi->s_writepages_rwsem jbd2_handle &ei->i_data_sem &journal->j_revoke_lock irq_context: 0 (wq_completion)writeback (work_completion)(&(&wb->dwork)->work) &type->s_umount_key#32 &sbi->s_writepages_rwsem jbd2_handle &ei->i_data_sem key#3 irq_context: 0 (wq_completion)writeback (work_completion)(&(&wb->dwork)->work) &type->s_umount_key#32 &sbi->s_writepages_rwsem jbd2_handle &ei->i_data_sem &(ei->i_block_reservation_lock) key#14 irq_context: 0 (wq_completion)writeback (work_completion)(&(&wb->dwork)->work) &type->s_umount_key#32 &sbi->s_writepages_rwsem jbd2_handle &ei->i_data_sem &ei->i_es_lock &sbi->s_es_lock irq_context: 0 (wq_completion)writeback (work_completion)(&(&wb->dwork)->work) &type->s_umount_key#32 &sbi->s_writepages_rwsem jbd2_handle rcu_read_lock key#10 irq_context: 0 (wq_completion)writeback (work_completion)(&(&wb->dwork)->work) &type->s_umount_key#32 &sbi->s_writepages_rwsem jbd2_handle rcu_read_lock &xa->xa_lock#9 key#10 irq_context: 0 (wq_completion)writeback (work_completion)(&(&wb->dwork)->work) &type->s_umount_key#32 &sbi->s_writepages_rwsem jbd2_handle rcu_read_lock rcu_node_0 irq_context: 0 (wq_completion)writeback (work_completion)(&(&wb->dwork)->work) &type->s_umount_key#32 &sbi->s_writepages_rwsem jbd2_handle rcu_read_lock &rq->__lock irq_context: 0 (wq_completion)writeback (work_completion)(&(&wb->dwork)->work) &type->s_umount_key#32 &sbi->s_writepages_rwsem jbd2_handle tk_core.seq.seqcount irq_context: 0 (wq_completion)writeback (work_completion)(&(&wb->dwork)->work) &type->s_umount_key#32 &sbi->s_writepages_rwsem jbd2_handle &obj_hash[i].lock irq_context: 0 (wq_completion)writeback (work_completion)(&(&wb->dwork)->work) &type->s_umount_key#32 &sbi->s_writepages_rwsem jbd2_handle &base->lock irq_context: 0 (wq_completion)writeback (work_completion)(&(&wb->dwork)->work) &type->s_umount_key#32 &sbi->s_writepages_rwsem jbd2_handle &base->lock &obj_hash[i].lock irq_context: 0 (wq_completion)writeback (work_completion)(&(&wb->dwork)->work) &type->s_umount_key#32 &sbi->s_writepages_rwsem jbd2_handle &dd->lock irq_context: 0 (wq_completion)writeback (work_completion)(&(&wb->dwork)->work) &type->s_umount_key#32 &sbi->s_writepages_rwsem jbd2_handle rcu_read_lock &dd->lock irq_context: 0 (wq_completion)writeback (work_completion)(&(&wb->dwork)->work) &type->s_umount_key#32 &sbi->s_writepages_rwsem jbd2_handle rcu_read_lock &obj_hash[i].lock irq_context: 0 (wq_completion)writeback (work_completion)(&(&wb->dwork)->work) &type->s_umount_key#32 &sbi->s_writepages_rwsem jbd2_handle rcu_read_lock pool_lock#2 irq_context: 0 (wq_completion)writeback (work_completion)(&(&wb->dwork)->work) &type->s_umount_key#32 &sbi->s_writepages_rwsem jbd2_handle rcu_read_lock tk_core.seq.seqcount irq_context: 0 (wq_completion)writeback (work_completion)(&(&wb->dwork)->work) &type->s_umount_key#32 &sbi->s_writepages_rwsem jbd2_handle rcu_read_lock &base->lock irq_context: 0 (wq_completion)writeback (work_completion)(&(&wb->dwork)->work) &type->s_umount_key#32 &sbi->s_writepages_rwsem jbd2_handle rcu_read_lock &base->lock &obj_hash[i].lock irq_context: 0 (wq_completion)writeback (work_completion)(&(&wb->dwork)->work) &type->s_umount_key#32 &sbi->s_writepages_rwsem jbd2_handle rcu_read_lock &virtscsi_vq->vq_lock irq_context: 0 (wq_completion)writeback (work_completion)(&(&wb->dwork)->work) &type->s_umount_key#32 &sbi->s_writepages_rwsem jbd2_handle &rq_wait->wait irq_context: 0 (wq_completion)writeback (work_completion)(&(&wb->dwork)->work) &type->s_umount_key#32 &sbi->s_writepages_rwsem jbd2_handle rcu_read_lock &pool->lock irq_context: 0 (wq_completion)writeback (work_completion)(&(&wb->dwork)->work) &type->s_umount_key#32 &sbi->s_writepages_rwsem jbd2_handle rcu_read_lock &pool->lock &obj_hash[i].lock irq_context: 0 (wq_completion)writeback (work_completion)(&(&wb->dwork)->work) &type->s_umount_key#32 &sbi->s_writepages_rwsem jbd2_handle rcu_read_lock &pool->lock pool_lock#2 irq_context: 0 (wq_completion)writeback (work_completion)(&(&wb->dwork)->work) &type->s_umount_key#32 &sbi->s_writepages_rwsem jbd2_handle rcu_read_lock &pool->lock &p->pi_lock irq_context: 0 (wq_completion)writeback (work_completion)(&(&wb->dwork)->work) &type->s_umount_key#32 &sbi->s_writepages_rwsem jbd2_handle rcu_read_lock &pool->lock &p->pi_lock &rq->__lock irq_context: 0 (wq_completion)writeback (work_completion)(&(&wb->dwork)->work) &type->s_umount_key#32 &sbi->s_writepages_rwsem jbd2_handle rcu_read_lock &pool->lock &p->pi_lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)writeback (work_completion)(&(&wb->dwork)->work) &type->s_umount_key#32 &sbi->s_writepages_rwsem jbd2_handle rcu_node_0 irq_context: 0 (wq_completion)writeback (work_completion)(&(&wb->dwork)->work) &type->s_umount_key#32 &sbi->s_writepages_rwsem jbd2_handle &rq->__lock irq_context: 0 (wq_completion)writeback (work_completion)(&(&wb->dwork)->work) &type->s_umount_key#32 &sbi->s_writepages_rwsem jbd2_handle &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)writeback (work_completion)(&(&wb->dwork)->work) &type->s_umount_key#32 &sbi->s_writepages_rwsem jbd2_handle rcu_read_lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)writeback (work_completion)(&(&wb->dwork)->work) &type->s_umount_key#32 &sbi->s_writepages_rwsem jbd2_handle rcu_read_lock &c->lock irq_context: 0 (wq_completion)writeback (work_completion)(&(&wb->dwork)->work) &type->s_umount_key#32 &sbi->s_writepages_rwsem jbd2_handle &journal->j_state_lock irq_context: 0 (wq_completion)writeback (work_completion)(&(&wb->dwork)->work) &type->s_umount_key#32 &sbi->s_writepages_rwsem jbd2_handle &journal->j_state_lock &journal->j_wait_commit irq_context: 0 (wq_completion)writeback (work_completion)(&(&wb->dwork)->work) &type->s_umount_key#32 &sbi->s_writepages_rwsem jbd2_handle &journal->j_state_lock &journal->j_wait_commit &p->pi_lock irq_context: 0 (wq_completion)writeback (work_completion)(&(&wb->dwork)->work) &type->s_umount_key#32 &sbi->s_writepages_rwsem jbd2_handle &journal->j_state_lock &journal->j_wait_commit &p->pi_lock &rq->__lock irq_context: 0 (wq_completion)writeback (work_completion)(&(&wb->dwork)->work) &type->s_umount_key#32 &sbi->s_writepages_rwsem jbd2_handle &journal->j_state_lock &journal->j_wait_commit &p->pi_lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)writeback (work_completion)(&(&wb->dwork)->work) &type->s_umount_key#32 &sbi->s_writepages_rwsem jbd2_handle &journal->j_wait_updates &p->pi_lock irq_context: 0 (wq_completion)writeback (work_completion)(&(&wb->dwork)->work) &type->s_umount_key#32 &sbi->s_writepages_rwsem jbd2_handle &journal->j_wait_updates &p->pi_lock &rq->__lock irq_context: 0 (wq_completion)writeback (work_completion)(&(&wb->dwork)->work) &type->s_umount_key#32 &sbi->s_writepages_rwsem jbd2_handle &journal->j_wait_updates &p->pi_lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)writeback (work_completion)(&(&wb->dwork)->work) &type->s_umount_key#32 &sbi->s_writepages_rwsem rcu_read_lock &dd->lock irq_context: 0 (wq_completion)writeback (work_completion)(&(&wb->dwork)->work) &type->s_umount_key#32 &sbi->s_writepages_rwsem rcu_read_lock &obj_hash[i].lock irq_context: 0 (wq_completion)writeback (work_completion)(&(&wb->dwork)->work) &type->s_umount_key#32 &sbi->s_writepages_rwsem rcu_read_lock tk_core.seq.seqcount irq_context: 0 (wq_completion)writeback (work_completion)(&(&wb->dwork)->work) &type->s_umount_key#32 &sbi->s_writepages_rwsem rcu_read_lock &virtscsi_vq->vq_lock irq_context: 0 (wq_completion)writeback (work_completion)(&(&wb->dwork)->work) &type->s_umount_key#32 &sbi->s_writepages_rwsem rcu_read_lock pool_lock#2 irq_context: 0 (wq_completion)writeback (work_completion)(&(&wb->dwork)->work) &type->s_umount_key#32 &sbi->s_writepages_rwsem rcu_read_lock &c->lock irq_context: 0 (wq_completion)writeback (work_completion)(&(&wb->dwork)->work) &type->s_umount_key#32 &sbi->s_writepages_rwsem rcu_read_lock &____s->seqcount irq_context: 0 (wq_completion)writeback (work_completion)(&(&wb->dwork)->work) &type->s_umount_key#32 &sbi->s_writepages_rwsem &rq_wait->wait irq_context: 0 (wq_completion)writeback (work_completion)(&(&wb->dwork)->work) &type->s_umount_key#32 &sbi->s_writepages_rwsem rcu_read_lock &pool->lock irq_context: 0 (wq_completion)writeback (work_completion)(&(&wb->dwork)->work) &type->s_umount_key#32 &sbi->s_writepages_rwsem rcu_read_lock &pool->lock &obj_hash[i].lock irq_context: 0 (wq_completion)writeback (work_completion)(&(&wb->dwork)->work) &type->s_umount_key#32 &sbi->s_writepages_rwsem rcu_read_lock &pool->lock pool_lock#2 irq_context: 0 (wq_completion)writeback (work_completion)(&(&wb->dwork)->work) &type->s_umount_key#32 &sbi->s_writepages_rwsem rcu_read_lock &pool->lock &p->pi_lock irq_context: 0 (wq_completion)writeback (work_completion)(&(&wb->dwork)->work) &type->s_umount_key#32 &sbi->s_writepages_rwsem rcu_read_lock &pool->lock &p->pi_lock &rq->__lock irq_context: 0 (wq_completion)writeback (work_completion)(&(&wb->dwork)->work) &type->s_umount_key#32 &sbi->s_writepages_rwsem rcu_read_lock &pool->lock &p->pi_lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)writeback (work_completion)(&(&wb->dwork)->work) &type->s_umount_key#32 &sbi->s_writepages_rwsem &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)writeback (work_completion)(&(&wb->dwork)->work) &type->s_umount_key#32 &sbi->s_writepages_rwsem rcu_read_lock &rq->__lock irq_context: 0 (wq_completion)writeback (work_completion)(&(&wb->dwork)->work) &type->s_umount_key#32 &sbi->s_writepages_rwsem rcu_read_lock &stopper->lock irq_context: 0 (wq_completion)writeback (work_completion)(&(&wb->dwork)->work) &type->s_umount_key#32 &sbi->s_writepages_rwsem rcu_read_lock &stop_pi_lock irq_context: 0 (wq_completion)writeback (work_completion)(&(&wb->dwork)->work) &type->s_umount_key#32 &sbi->s_writepages_rwsem rcu_read_lock &stop_pi_lock &rq->__lock irq_context: 0 (wq_completion)writeback (work_completion)(&(&wb->dwork)->work) &type->s_umount_key#32 &sbi->s_writepages_rwsem rcu_read_lock &stop_pi_lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)writeback (work_completion)(&(&wb->dwork)->work) &type->s_umount_key#32 &sbi->s_writepages_rwsem jbd2_handle &ei->i_data_sem bit_wait_table + i irq_context: 0 (wq_completion)writeback (work_completion)(&(&wb->dwork)->work) &type->s_umount_key#32 &sbi->s_writepages_rwsem jbd2_handle &ei->i_data_sem &dd->lock irq_context: 0 (wq_completion)writeback (work_completion)(&(&wb->dwork)->work) &type->s_umount_key#32 &sbi->s_writepages_rwsem jbd2_handle &ei->i_data_sem rcu_read_lock &pool->lock irq_context: 0 (wq_completion)writeback (work_completion)(&(&wb->dwork)->work) &type->s_umount_key#32 &sbi->s_writepages_rwsem jbd2_handle &ei->i_data_sem rcu_read_lock &pool->lock &obj_hash[i].lock irq_context: 0 (wq_completion)writeback (work_completion)(&(&wb->dwork)->work) &type->s_umount_key#32 &sbi->s_writepages_rwsem jbd2_handle &ei->i_data_sem rcu_read_lock &pool->lock pool_lock#2 irq_context: 0 (wq_completion)writeback (work_completion)(&(&wb->dwork)->work) &type->s_umount_key#32 &sbi->s_writepages_rwsem jbd2_handle &ei->i_data_sem rcu_read_lock &pool->lock &p->pi_lock irq_context: 0 (wq_completion)writeback (work_completion)(&(&wb->dwork)->work) &type->s_umount_key#32 &sbi->s_writepages_rwsem jbd2_handle &ei->i_data_sem rcu_read_lock &pool->lock &p->pi_lock &rq->__lock irq_context: 0 (wq_completion)writeback (work_completion)(&(&wb->dwork)->work) &type->s_umount_key#32 &sbi->s_writepages_rwsem jbd2_handle &ei->i_data_sem rcu_read_lock &pool->lock &p->pi_lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)writeback (work_completion)(&(&wb->dwork)->work) &type->s_umount_key#32 &sbi->s_writepages_rwsem jbd2_handle &ei->i_data_sem &rq->__lock irq_context: 0 (wq_completion)writeback (work_completion)(&(&wb->dwork)->work) &type->s_umount_key#32 &sbi->s_writepages_rwsem jbd2_handle &ei->i_data_sem &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)writeback (work_completion)(&(&wb->dwork)->work) &type->s_umount_key#32 &sbi->s_writepages_rwsem jbd2_handle &ei->i_data_sem rcu_read_lock &rq->__lock irq_context: 0 (wq_completion)writeback (work_completion)(&(&wb->dwork)->work) &type->s_umount_key#32 &sbi->s_writepages_rwsem jbd2_handle &ei->i_data_sem rcu_read_lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)writeback (work_completion)(&(&wb->dwork)->work) &type->s_umount_key#32 &sbi->s_writepages_rwsem jbd2_handle &ei->i_data_sem rcu_read_lock &stopper->lock irq_context: 0 (wq_completion)writeback (work_completion)(&(&wb->dwork)->work) &type->s_umount_key#32 &sbi->s_writepages_rwsem jbd2_handle &ei->i_data_sem rcu_read_lock &stop_pi_lock irq_context: 0 (wq_completion)writeback (work_completion)(&(&wb->dwork)->work) &type->s_umount_key#32 &sbi->s_writepages_rwsem jbd2_handle &ei->i_data_sem rcu_read_lock &stop_pi_lock &rq->__lock irq_context: 0 (wq_completion)writeback (work_completion)(&(&wb->dwork)->work) &type->s_umount_key#32 &sbi->s_writepages_rwsem jbd2_handle &ei->i_data_sem rcu_read_lock &stop_pi_lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: softirq &ei->i_completed_io_lock rcu_read_lock &pool->lock/1 &p->pi_lock &rq->__lock &cfs_rq->removed.lock irq_context: 0 (wq_completion)ext4-rsv-conversion (work_completion)(&ei->i_rsv_conversion_work) jbd2_handle &ei->i_data_sem &mapping->private_lock irq_context: 0 (wq_completion)ext4-rsv-conversion (work_completion)(&ei->i_rsv_conversion_work) jbd2_handle &ei->i_data_sem &ei->i_es_lock key#6 irq_context: 0 (wq_completion)ext4-rsv-conversion (work_completion)(&ei->i_rsv_conversion_work) rcu_read_lock &xa->xa_lock#9 &pl->lock irq_context: 0 (wq_completion)ext4-rsv-conversion (work_completion)(&ei->i_rsv_conversion_work) rcu_read_lock &xa->xa_lock#9 &pl->lock key#12 irq_context: 0 (wq_completion)ext4-rsv-conversion (work_completion)(&ei->i_rsv_conversion_work) rcu_read_lock &xa->xa_lock#9 key#12 irq_context: 0 (wq_completion)ext4-rsv-conversion (work_completion)(&ei->i_rsv_conversion_work) rcu_read_lock &xa->xa_lock#9 key#13 irq_context: softirq &(&hctx->run_work)->timer irq_context: softirq &(&hctx->run_work)->timer rcu_read_lock &pool->lock irq_context: softirq &(&hctx->run_work)->timer rcu_read_lock &pool->lock &obj_hash[i].lock irq_context: softirq &(&hctx->run_work)->timer rcu_read_lock &pool->lock &p->pi_lock irq_context: softirq &(&hctx->run_work)->timer rcu_read_lock &pool->lock &p->pi_lock &rq->__lock irq_context: softirq &(&hctx->run_work)->timer rcu_read_lock &pool->lock &p->pi_lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)ext4-rsv-conversion (work_completion)(&ei->i_rsv_conversion_work) rcu_read_lock rcu_node_0 irq_context: 0 (wq_completion)ext4-rsv-conversion (work_completion)(&ei->i_rsv_conversion_work) rcu_read_lock &rq->__lock irq_context: 0 (wq_completion)writeback (work_completion)(&(&wb->dwork)->work) &type->s_umount_key#32 &sbi->s_writepages_rwsem jbd2_handle &mapping->i_mmap_rwsem &rq->__lock irq_context: 0 (wq_completion)ext4-rsv-conversion (work_completion)(&ei->i_rsv_conversion_work) rcu_read_lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)writeback (work_completion)(&(&wb->dwork)->work) &type->s_umount_key#32 &sbi->s_writepages_rwsem rcu_node_0 irq_context: 0 (wq_completion)writeback (work_completion)(&(&wb->dwork)->work) &type->s_umount_key#32 &sbi->s_writepages_rwsem rcu_read_lock rcu_node_0 irq_context: softirq (&cb->timer) &rq_wait->wait irq_context: softirq (&cb->timer) &rq_wait->wait &p->pi_lock irq_context: softirq (&cb->timer) &rq_wait->wait &p->pi_lock &rq->__lock irq_context: softirq (&cb->timer) &rq_wait->wait &p->pi_lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: softirq &rq_wait->wait &p->pi_lock &rq->__lock &cfs_rq->removed.lock irq_context: 0 (wq_completion)writeback (work_completion)(&(&wb->dwork)->work) &type->s_umount_key#32 &sbi->s_writepages_rwsem rcu_read_lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)ext4-rsv-conversion (work_completion)(&ei->i_rsv_conversion_work) &rq->__lock irq_context: 0 (wq_completion)ext4-rsv-conversion (work_completion)(&ei->i_rsv_conversion_work) &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)ext4-rsv-conversion (work_completion)(&ei->i_rsv_conversion_work) jbd2_handle &ei->i_data_sem batched_entropy_u8.lock irq_context: 0 (wq_completion)ext4-rsv-conversion (work_completion)(&ei->i_rsv_conversion_work) jbd2_handle &ei->i_data_sem kfence_freelist_lock irq_context: 0 (wq_completion)ext4-rsv-conversion (work_completion)(&ei->i_rsv_conversion_work) jbd2_handle &ei->i_data_sem &meta->lock irq_context: 0 (wq_completion)ext4-rsv-conversion (work_completion)(&ei->i_rsv_conversion_work) jbd2_handle &ei->i_data_sem &ei->i_es_lock &sbi->s_es_lock irq_context: 0 sk_lock-AF_INET rcu_read_lock rcu_read_lock &rq->__lock &cfs_rq->removed.lock irq_context: 0 tasklist_lock &sighand->siglock &sighand->signalfd_wqh &ep->lock &ep->wq &p->pi_lock &cfs_rq->removed.lock irq_context: 0 sb_writers#5 &type->i_mutex_dir_key#5/1 &cfs_rq->removed.lock irq_context: softirq net/wireless/reg.c:533 irq_context: softirq net/wireless/reg.c:533 rcu_read_lock &pool->lock irq_context: softirq net/wireless/reg.c:533 rcu_read_lock &pool->lock &obj_hash[i].lock irq_context: softirq net/wireless/reg.c:533 rcu_read_lock &pool->lock &p->pi_lock irq_context: softirq net/wireless/reg.c:533 rcu_read_lock &pool->lock &p->pi_lock &rq->__lock irq_context: softirq net/wireless/reg.c:533 rcu_read_lock &pool->lock &p->pi_lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)events_power_efficient (crda_timeout).work irq_context: 0 (wq_completion)events_power_efficient (crda_timeout).work rtnl_mutex irq_context: 0 (wq_completion)events_power_efficient (crda_timeout).work rtnl_mutex reg_indoor_lock irq_context: 0 (wq_completion)events_power_efficient (crda_timeout).work rtnl_mutex &obj_hash[i].lock irq_context: 0 (wq_completion)events_power_efficient (crda_timeout).work rtnl_mutex pool_lock#2 irq_context: 0 (wq_completion)events_power_efficient (crda_timeout).work rtnl_mutex krc.lock irq_context: 0 (wq_completion)events_power_efficient (crda_timeout).work rtnl_mutex reg_requests_lock irq_context: 0 (wq_completion)events_power_efficient (crda_timeout).work rtnl_mutex reg_pending_beacons_lock irq_context: 0 (wq_completion)events_power_efficient (crda_timeout).work rtnl_mutex fs_reclaim irq_context: 0 (wq_completion)events_power_efficient (crda_timeout).work rtnl_mutex fs_reclaim mmu_notifier_invalidate_range_start irq_context: 0 (wq_completion)events_power_efficient (crda_timeout).work rtnl_mutex remove_cache_srcu irq_context: 0 (wq_completion)events_power_efficient (crda_timeout).work rtnl_mutex remove_cache_srcu quarantine_lock irq_context: 0 (wq_completion)events_power_efficient (crda_timeout).work rtnl_mutex remove_cache_srcu &c->lock irq_context: 0 (wq_completion)events_power_efficient (crda_timeout).work rtnl_mutex remove_cache_srcu &n->list_lock irq_context: 0 (wq_completion)events_power_efficient (crda_timeout).work rtnl_mutex remove_cache_srcu rcu_read_lock pool_lock#2 irq_context: 0 (wq_completion)events_power_efficient (crda_timeout).work rtnl_mutex remove_cache_srcu &obj_hash[i].lock irq_context: 0 (wq_completion)events_power_efficient (crda_timeout).work rtnl_mutex rcu_read_lock &pool->lock irq_context: 0 (wq_completion)events_power_efficient (crda_timeout).work rtnl_mutex rcu_read_lock &pool->lock &obj_hash[i].lock irq_context: 0 (wq_completion)events reg_work rtnl_mutex &c->lock irq_context: 0 &mm->mmap_lock rcu_read_lock rcu_read_lock rcu_node_0 irq_context: 0 &mm->mmap_lock rcu_read_lock rcu_read_lock &rq->__lock irq_context: 0 (wq_completion)events_power_efficient (crda_timeout).work rtnl_mutex krc.lock &obj_hash[i].lock irq_context: 0 (wq_completion)events_power_efficient (crda_timeout).work rtnl_mutex krc.lock &base->lock irq_context: 0 (wq_completion)events_power_efficient (crda_timeout).work rtnl_mutex krc.lock &base->lock &obj_hash[i].lock irq_context: softirq slock-AF_INET rcu_read_lock rcu_read_lock rcu_read_lock_bh dev->qdisc_tx_busylock ?: &qdisc_tx_busylock _xmit_ETHER#2 rcu_read_lock &____s->seqcount irq_context: 0 (wq_completion)mm_percpu_wq (work_completion)(&barr->work) &x->wait#10 &p->pi_lock &rq->__lock irq_context: 0 (wq_completion)mm_percpu_wq (work_completion)(&barr->work) &x->wait#10 &p->pi_lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)events reg_work rtnl_mutex uevent_sock_mutex rcu_read_lock &ei->socket.wq.wait &ep->lock &ep->wq &p->pi_lock &cfs_rq->removed.lock irq_context: 0 &mm->mmap_lock &anon_vma->rwsem &obj_hash[i].lock pool_lock irq_context: 0 sk_lock-AF_INET remove_cache_srcu &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)events reg_work rtnl_mutex uevent_sock_mutex rcu_read_lock rcu_node_0 irq_context: 0 (wq_completion)events reg_work rtnl_mutex uevent_sock_mutex rcu_read_lock &rq->__lock irq_context: 0 sb_writers#3 &sb->s_type->i_mutex_key#9 &n->list_lock irq_context: 0 sb_writers#3 &sb->s_type->i_mutex_key#9 &n->list_lock &c->lock irq_context: 0 sb_writers#3 &sb->s_type->i_mutex_key#9 tomoyo_ss batched_entropy_u8.lock irq_context: 0 sb_writers#3 &sb->s_type->i_mutex_key#9 tomoyo_ss kfence_freelist_lock irq_context: 0 sb_writers#3 &sb->s_type->i_mutex_key#9 tomoyo_ss &meta->lock irq_context: 0 &sb->s_type->i_mutex_key#10 rcu_state.exp_mutex rcu_read_lock &stopper->lock irq_context: 0 &sb->s_type->i_mutex_key#10 rcu_state.exp_mutex rcu_read_lock &stop_pi_lock irq_context: 0 sk_lock-AF_INET rcu_read_lock &rcu_state.expedited_wq irq_context: 0 &sb->s_type->i_mutex_key#10 rcu_state.exp_mutex rcu_read_lock &stop_pi_lock &rq->__lock irq_context: 0 sk_lock-AF_INET rcu_read_lock &rcu_state.expedited_wq &p->pi_lock irq_context: 0 &sb->s_type->i_mutex_key#10 rcu_state.exp_mutex rcu_read_lock &stop_pi_lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 &ep->mtx rcu_node_0 irq_context: 0 &ep->mtx &rcu_state.expedited_wq irq_context: 0 &ep->mtx &rcu_state.expedited_wq &p->pi_lock irq_context: 0 &rq->__lock rcu_read_lock &obj_hash[i].lock irq_context: 0 &rq->__lock rcu_read_lock &base->lock irq_context: 0 &rq->__lock rcu_read_lock &base->lock &obj_hash[i].lock irq_context: 0 sk_lock-AF_PACKET rcu_state.exp_mutex rcu_read_lock &rq->__lock irq_context: 0 sk_lock-AF_PACKET rcu_state.exp_mutex rcu_read_lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 sk_lock-AF_PACKET &rq->__lock irq_context: 0 &vma->vm_lock->lock rcu_read_lock &rcu_state.expedited_wq irq_context: 0 &mm->mmap_lock fill_pool_map-wait-type-override &rq->__lock irq_context: 0 sk_lock-AF_INET &rcu_state.expedited_wq irq_context: 0 rcu_read_lock rcu_read_lock &sighand->siglock &sighand->signalfd_wqh &ep->lock &ep->wq &p->pi_lock &rq->__lock &cfs_rq->removed.lock irq_context: softirq net/wireless/reg.c:533 rcu_read_lock &pool->lock &p->pi_lock &rq->__lock &cfs_rq->removed.lock irq_context: 0 (wq_completion)writeback (work_completion)(&(&wb->dwork)->work) &type->s_umount_key#43 rcu_read_lock rcu_read_lock per_cpu_ptr(&cgroup_rstat_cpu_lock, cpu) irq_context: 0 (wq_completion)writeback (work_completion)(&(&wb->dwork)->work) &type->s_umount_key#43 lock#4 &lruvec->lru_lock irq_context: 0 (wq_completion)writeback (work_completion)(&(&wb->dwork)->work) rcu_read_lock &base->lock irq_context: 0 (wq_completion)writeback (work_completion)(&(&wb->dwork)->work) rcu_read_lock &base->lock &obj_hash[i].lock irq_context: 0 (wq_completion)events_power_efficient (crda_timeout).work rtnl_mutex &c->lock irq_context: 0 sk_lock-AF_INET remove_cache_srcu &rq->__lock &cfs_rq->removed.lock irq_context: 0 lock#3 rcu_read_lock &rq->__lock irq_context: 0 lock#3 rcu_read_lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 &sb->s_type->i_mutex_key#9 &rq->__lock irq_context: 0 sb_writers#3 &sb->s_type->i_mutex_key#9 &rq->__lock irq_context: 0 sb_writers#3 &sb->s_type->i_mutex_key#9 tomoyo_ss &rq->__lock irq_context: 0 sb_writers#3 tomoyo_ss &rq->__lock irq_context: 0 &vma->vm_lock->lock batched_entropy_u8.lock crngs.lock irq_context: 0 &vma->vm_lock->lock batched_entropy_u8.lock crngs.lock base_crng.lock irq_context: softirq (&lapb->t1timer) &lapb->lock &pcp->lock &zone->lock irq_context: softirq (&lapb->t1timer) &lapb->lock &pcp->lock &zone->lock &____s->seqcount irq_context: softirq (&sk->sk_timer) irq_context: softirq (&sk->sk_timer) slock-AF_INET irq_context: softirq (&sk->sk_timer) slock-AF_INET tk_core.seq.seqcount irq_context: softirq (&sk->sk_timer) slock-AF_INET &obj_hash[i].lock irq_context: softirq (&sk->sk_timer) slock-AF_INET &base->lock irq_context: softirq (&sk->sk_timer) slock-AF_INET &base->lock &obj_hash[i].lock irq_context: softirq (&icsk->icsk_delack_timer) slock-AF_INET &pcp->lock &zone->lock irq_context: 0 &vma->vm_lock->lock &rq->__lock &cfs_rq->removed.lock irq_context: softirq rcu_read_lock rcu_read_lock slock-AF_INET/1 rcu_read_lock rcu_read_lock rcu_read_lock_bh dev->qdisc_tx_busylock ?: &qdisc_tx_busylock _xmit_ETHER#2 rcu_read_lock batched_entropy_u8.lock crngs.lock irq_context: 0 sk_lock-AF_INET fs_reclaim &rq->__lock irq_context: 0 sk_lock-AF_INET fs_reclaim &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)mm_percpu_wq (work_completion)(&barr->work) &x->wait#10 &p->pi_lock &cfs_rq->removed.lock irq_context: 0 sk_lock-AF_INET slock-AF_INET rcu_read_lock rcu_read_lock &____s->seqcount#7 irq_context: 0 sk_lock-AF_INET slock-AF_INET rcu_read_lock rcu_read_lock &ct->lock irq_context: 0 sk_lock-AF_INET slock-AF_INET rcu_read_lock rcu_read_lock &____s->seqcount#10 irq_context: 0 sk_lock-AF_INET slock-AF_INET rcu_read_lock rcu_read_lock rcu_read_lock_bh dev->qdisc_tx_busylock ?: &qdisc_tx_busylock _xmit_ETHER#2 irq_context: 0 sk_lock-AF_INET slock-AF_INET rcu_read_lock rcu_read_lock rcu_read_lock_bh dev->qdisc_tx_busylock ?: &qdisc_tx_busylock _xmit_ETHER#2 rcu_read_lock &____s->seqcount irq_context: 0 sk_lock-AF_INET slock-AF_INET rcu_read_lock rcu_read_lock rcu_read_lock_bh dev->qdisc_tx_busylock ?: &qdisc_tx_busylock _xmit_ETHER#2 rcu_read_lock pool_lock#2 irq_context: 0 sk_lock-AF_INET slock-AF_INET rcu_read_lock rcu_read_lock rcu_read_lock_bh dev->qdisc_tx_busylock ?: &qdisc_tx_busylock _xmit_ETHER#2 rcu_read_lock rcu_read_lock pool_lock#2 irq_context: 0 sk_lock-AF_INET slock-AF_INET rcu_read_lock rcu_read_lock rcu_read_lock_bh dev->qdisc_tx_busylock ?: &qdisc_tx_busylock _xmit_ETHER#2 rcu_read_lock &obj_hash[i].lock irq_context: 0 sk_lock-AF_INET slock-AF_INET rcu_read_lock rcu_read_lock rcu_read_lock_bh dev->qdisc_tx_busylock ?: &qdisc_tx_busylock _xmit_ETHER#2 rcu_read_lock tk_core.seq.seqcount irq_context: 0 sk_lock-AF_INET slock-AF_INET &obj_hash[i].lock irq_context: 0 sk_lock-AF_INET slock-AF_INET &base->lock irq_context: 0 sk_lock-AF_INET slock-AF_INET &base->lock &obj_hash[i].lock irq_context: softirq (&rxnet->peer_keepalive_timer) rcu_read_lock &pool->lock/1 &p->pi_lock &rq->__lock &cfs_rq->removed.lock irq_context: 0 &sbinfo->stat_lock irq_context: 0 &sb->s_type->i_lock_key &dentry->d_lock irq_context: 0 sb_writers#4 &type->i_mutex_dir_key#3/1 &sb->s_type->i_mutex_key#8 irq_context: 0 sb_writers#4 &type->i_mutex_dir_key#3/1 &sb->s_type->i_mutex_key#8 mmu_notifier_invalidate_range_start irq_context: 0 sb_writers#4 &type->i_mutex_dir_key#3/1 &sb->s_type->i_mutex_key#8 pool_lock#2 irq_context: 0 sb_writers#4 &type->i_mutex_dir_key#3/1 &sb->s_type->i_mutex_key#8 &journal->j_state_lock irq_context: 0 sb_writers#4 &type->i_mutex_dir_key#3/1 &sb->s_type->i_mutex_key#8 &journal->j_state_lock irq_context: 0 sb_writers#4 &type->i_mutex_dir_key#3/1 &sb->s_type->i_mutex_key#8 &journal->j_state_lock tk_core.seq.seqcount irq_context: 0 sb_writers#4 &type->i_mutex_dir_key#3/1 &sb->s_type->i_mutex_key#8 &journal->j_state_lock &obj_hash[i].lock irq_context: 0 sb_writers#4 &type->i_mutex_dir_key#3/1 &sb->s_type->i_mutex_key#8 &journal->j_state_lock &base->lock irq_context: 0 sb_writers#4 &type->i_mutex_dir_key#3/1 &sb->s_type->i_mutex_key#8 &journal->j_state_lock &base->lock &obj_hash[i].lock irq_context: 0 sb_writers#4 &type->i_mutex_dir_key#3/1 &sb->s_type->i_mutex_key#8 jbd2_handle irq_context: 0 sb_writers#4 &type->i_mutex_dir_key#3/1 &sb->s_type->i_mutex_key#8 jbd2_handle tk_core.seq.seqcount irq_context: 0 sb_writers#4 &type->i_mutex_dir_key#3/1 &sb->s_type->i_mutex_key#8 jbd2_handle &ei->i_es_lock irq_context: 0 sb_writers#4 &type->i_mutex_dir_key#3/1 &sb->s_type->i_mutex_key#8 jbd2_handle mmu_notifier_invalidate_range_start irq_context: 0 sb_writers#4 &type->i_mutex_dir_key#3/1 &sb->s_type->i_mutex_key#8 jbd2_handle pool_lock#2 irq_context: 0 sb_writers#4 &type->i_mutex_dir_key#3/1 &sb->s_type->i_mutex_key#8 jbd2_handle &ret->b_state_lock irq_context: 0 sb_writers#4 &type->i_mutex_dir_key#3/1 &sb->s_type->i_mutex_key#8 jbd2_handle &ret->b_state_lock &journal->j_list_lock irq_context: 0 sb_writers#4 &type->i_mutex_dir_key#3/1 &sb->s_type->i_mutex_key#8 jbd2_handle &mapping->private_lock irq_context: 0 sb_writers#4 &type->i_mutex_dir_key#3/1 &sb->s_type->i_mutex_key#8 jbd2_handle &ei->i_raw_lock irq_context: 0 sb_writers#4 &type->i_mutex_dir_key#3/1 &sb->s_type->i_mutex_key#8 jbd2_handle &sb->s_type->i_lock_key#22 irq_context: 0 sb_writers#4 &type->i_mutex_dir_key#3/1 &sb->s_type->i_mutex_key#8 jbd2_handle &sb->s_type->i_lock_key#22 &dentry->d_lock irq_context: 0 sb_writers#4 &type->i_mutex_dir_key#3/1 &sb->s_type->i_mutex_key#8 jbd2_handle &journal->j_wait_updates irq_context: 0 sb_writers#4 &type->i_mutex_dir_key#3/1 &sb->s_type->i_mutex_key#8 &obj_hash[i].lock irq_context: 0 sb_writers#5 tomoyo_ss &n->list_lock irq_context: 0 sb_writers#5 tomoyo_ss &n->list_lock &c->lock irq_context: 0 sb_writers#4 &type->i_mutex_dir_key#3/1 tomoyo_ss &n->list_lock irq_context: 0 sb_writers#4 &type->i_mutex_dir_key#3/1 tomoyo_ss &n->list_lock &c->lock irq_context: 0 &mm->mmap_lock &sb->s_type->i_lock_key irq_context: 0 &mm->mmap_lock &xa->xa_lock#9 irq_context: 0 &mm->mmap_lock &info->lock irq_context: 0 &mm->mmap_lock tk_core.seq.seqcount irq_context: 0 &mm->mmap_lock mount_lock irq_context: 0 &mm->mmap_lock &wb->list_lock irq_context: 0 &mm->mmap_lock &wb->list_lock &sb->s_type->i_lock_key irq_context: 0 &mm->mmap_lock &xa->xa_lock#9 &c->lock irq_context: 0 &mm->mmap_lock &xa->xa_lock#9 pool_lock#2 irq_context: 0 sb_writers#4 &type->i_mutex_dir_key#3/1 jbd2_handle &ei->i_data_sem &wb->work_lock irq_context: 0 sb_writers#4 &type->i_mutex_dir_key#3/1 jbd2_handle &ei->i_data_sem &wb->work_lock &obj_hash[i].lock irq_context: 0 sb_writers#4 &type->i_mutex_dir_key#3/1 jbd2_handle &ei->i_data_sem &wb->work_lock &base->lock irq_context: 0 sb_writers#4 &type->i_mutex_dir_key#3/1 jbd2_handle &ei->i_data_sem &wb->work_lock &base->lock &obj_hash[i].lock irq_context: 0 sb_writers#4 &type->i_mutex_dir_key#3/1 jbd2_handle &ei->i_es_lock key#2 irq_context: 0 sb_writers#4 &type->i_mutex_dir_key#3/1 &ei->i_es_lock key#2 irq_context: 0 sb_writers#4 &type->i_mutex_dir_key#3/1 rcu_read_lock &rq->__lock irq_context: 0 sb_writers#4 &type->i_mutex_dir_key#3/1 &sem->wait_lock irq_context: 0 &pipe->mutex/1 fs_reclaim &rq->__lock irq_context: 0 &pipe->mutex/1 &mm->mmap_lock &rq->__lock irq_context: 0 sb_writers#4 &type->i_mutex_dir_key#3/1 tomoyo_ss &rq->__lock irq_context: 0 sb_writers#4 &type->i_mutex_dir_key#3/1 &pcp->lock &zone->lock irq_context: 0 sb_writers#4 &type->i_mutex_dir_key#3/1 jbd2_handle &rq->__lock irq_context: 0 sb_writers#4 &type->i_mutex_dir_key#3/1 jbd2_handle &ei->i_data_sem &rq->__lock irq_context: 0 sb_writers#4 &type->i_mutex_dir_key#3/1 jbd2_handle &ei->i_data_sem &rq->__lock irq_context: 0 sb_writers#4 &type->i_mutex_dir_key#3/1 jbd2_handle &ei->i_data_sem rcu_read_lock rcu_node_0 irq_context: 0 sb_writers#4 &type->i_mutex_dir_key#3/1 jbd2_handle &ei->i_data_sem rcu_read_lock &rq->__lock irq_context: 0 sb_writers#4 &type->i_mutex_dir_key#3/1 jbd2_handle &ei->i_data_sem rcu_read_lock &rcu_state.gp_wq irq_context: 0 sb_writers#4 &type->i_mutex_dir_key#3/1 jbd2_handle &ei->i_data_sem rcu_read_lock &rcu_state.gp_wq &p->pi_lock irq_context: 0 sb_writers#4 &type->i_mutex_dir_key#3/1 jbd2_handle &ei->i_data_sem rcu_read_lock &rcu_state.gp_wq &p->pi_lock &cfs_rq->removed.lock irq_context: 0 sb_writers#4 &type->i_mutex_dir_key#3/1 jbd2_handle &ei->i_data_sem rcu_read_lock &rcu_state.gp_wq &p->pi_lock &rq->__lock irq_context: 0 sb_writers#4 &type->i_mutex_dir_key#3/1 jbd2_handle &ei->i_data_sem rcu_read_lock &rcu_state.gp_wq &p->pi_lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 sb_writers#4 &type->i_mutex_dir_key#3/1 jbd2_handle rcu_node_0 irq_context: 0 sb_writers#4 &type->i_mutex_dir_key#3/1 jbd2_handle rcu_read_lock rcu_node_0 irq_context: 0 sb_writers#4 &sem->wait_lock irq_context: 0 sb_writers#4 &p->pi_lock irq_context: 0 sb_writers#4 &p->pi_lock &rq->__lock irq_context: 0 sb_writers#4 &p->pi_lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 sb_writers#4 &rq->__lock irq_context: 0 sb_writers#4 &type->i_mutex_dir_key#3/1 &rq->__lock &cfs_rq->removed.lock irq_context: 0 sb_writers#4 &type->i_mutex_dir_key#3/1 tomoyo_ss quarantine_lock irq_context: 0 sb_writers#4 &type->i_mutex_dir_key#3/1 remove_cache_srcu irq_context: 0 sb_writers#4 &type->i_mutex_dir_key#3/1 remove_cache_srcu quarantine_lock irq_context: 0 sb_writers#4 &type->i_mutex_dir_key#3/1 remove_cache_srcu &rq->__lock irq_context: 0 sb_writers#4 &type->i_mutex_dir_key#3/1 remove_cache_srcu &c->lock irq_context: 0 sb_writers#4 &type->i_mutex_dir_key#3/1 remove_cache_srcu &n->list_lock irq_context: 0 sb_writers#4 &type->i_mutex_dir_key#3/1 remove_cache_srcu rcu_read_lock rcu_node_0 irq_context: 0 sb_writers#4 &type->i_mutex_dir_key#3/1 remove_cache_srcu rcu_read_lock &rq->__lock irq_context: 0 sb_writers#4 &type->i_mutex_dir_key#3/1 remove_cache_srcu rcu_read_lock pool_lock#2 irq_context: 0 sb_writers#4 &type->i_mutex_dir_key#3/1 remove_cache_srcu &obj_hash[i].lock irq_context: 0 sb_writers#4 &type->i_mutex_dir_key#3/1 remove_cache_srcu rcu_node_0 irq_context: 0 sb_writers#4 &type->i_mutex_dir_key#3/1 jbd2_handle rcu_read_lock &rq->__lock irq_context: 0 sb_writers#4 &type->i_mutex_dir_key#3 tomoyo_ss &rq->__lock irq_context: 0 sb_writers#4 &type->i_mutex_dir_key#3 jbd2_handle &rq->__lock irq_context: 0 sb_writers#4 &type->i_mutex_dir_key#3/1 jbd2_handle &c->lock irq_context: 0 rcu_read_lock &sighand->siglock &p->pi_lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 &mm->mmap_lock sb_writers#5 mount_lock irq_context: 0 &mm->mmap_lock sb_writers#5 tk_core.seq.seqcount irq_context: 0 &newf->file_lock &newf->resize_wait irq_context: 0 sb_writers#4 &type->i_mutex_dir_key#3/1 jbd2_handle &ei->i_data_sem &c->lock irq_context: 0 &kcov->lock irq_context: 0 &mm->mmap_lock &kcov->lock irq_context: 0 &mm->mmap_lock ptlock_ptr(page)#2 key irq_context: 0 sb_writers#4 &type->i_mutex_dir_key#3/1 jbd2_handle &ei->i_data_sem &ei->i_es_lock key#6 irq_context: 0 sb_writers#4 &type->i_mutex_dir_key#3/1 jbd2_handle &ei->i_data_sem &ei->i_es_lock key#7 irq_context: 0 sb_writers#4 &type->i_mutex_dir_key#3/1 jbd2_handle &ei->i_data_sem &n->list_lock irq_context: 0 sb_writers#4 &type->i_mutex_dir_key#3/1 jbd2_handle &ei->i_data_sem &n->list_lock &c->lock irq_context: 0 sb_writers#4 &type->i_mutex_dir_key#3/1 jbd2_handle &ei->i_es_lock key#5 irq_context: 0 &kcov->lock kcov_remote_lock irq_context: 0 &kcov->lock kcov_remote_lock pool_lock#2 irq_context: 0 pid_caches_mutex irq_context: 0 pid_caches_mutex slab_mutex irq_context: 0 pid_caches_mutex slab_mutex fs_reclaim irq_context: 0 pid_caches_mutex slab_mutex fs_reclaim mmu_notifier_invalidate_range_start irq_context: 0 pid_caches_mutex slab_mutex pool_lock#2 irq_context: 0 pid_caches_mutex slab_mutex &c->lock irq_context: 0 pid_caches_mutex slab_mutex &n->list_lock irq_context: 0 pid_caches_mutex slab_mutex &rq->__lock irq_context: 0 pid_caches_mutex slab_mutex pcpu_alloc_mutex irq_context: 0 pid_caches_mutex slab_mutex pcpu_alloc_mutex pcpu_lock irq_context: 0 pid_caches_mutex slab_mutex &root->kernfs_rwsem irq_context: 0 pid_caches_mutex slab_mutex &k->list_lock irq_context: 0 pid_caches_mutex slab_mutex &____s->seqcount irq_context: 0 pid_caches_mutex slab_mutex rcu_read_lock pool_lock#2 irq_context: 0 pid_caches_mutex slab_mutex &obj_hash[i].lock irq_context: 0 pid_caches_mutex slab_mutex lock irq_context: 0 pid_caches_mutex slab_mutex lock kernfs_idr_lock irq_context: 0 pid_caches_mutex slab_mutex &root->kernfs_rwsem &root->kernfs_iattr_rwsem irq_context: 0 pid_caches_mutex slab_mutex &n->list_lock &c->lock irq_context: 0 pid_caches_mutex slab_mutex lock kernfs_idr_lock pool_lock#2 irq_context: 0 &kcov->lock kcov_remote_lock &c->lock irq_context: 0 &type->s_umount_key#48 irq_context: 0 &type->s_umount_key#48 sb_lock irq_context: 0 &type->s_umount_key#48 &dentry->d_lock irq_context: 0 &sb->s_type->i_mutex_key#18 irq_context: 0 &sb->s_type->i_mutex_key#18 namespace_sem irq_context: 0 &sb->s_type->i_mutex_key#18 namespace_sem rcu_read_lock mount_lock.seqcount irq_context: 0 &sb->s_type->i_mutex_key#18 namespace_sem fs_reclaim irq_context: 0 &sb->s_type->i_mutex_key#18 namespace_sem fs_reclaim mmu_notifier_invalidate_range_start irq_context: 0 &sb->s_type->i_mutex_key#18 namespace_sem rename_lock irq_context: 0 &sb->s_type->i_mutex_key#18 namespace_sem rename_lock rename_lock.seqcount irq_context: 0 &sb->s_type->i_mutex_key#18 namespace_sem rename_lock rename_lock.seqcount &dentry->d_lock irq_context: 0 &sb->s_type->i_mutex_key#18 namespace_sem mount_lock irq_context: 0 &sb->s_type->i_mutex_key#18 namespace_sem mount_lock &dentry->d_lock irq_context: 0 &sb->s_type->i_mutex_key#18 namespace_sem mount_lock rcu_read_lock &dentry->d_lock irq_context: 0 &sb->s_type->i_mutex_key#18 namespace_sem mount_lock &obj_hash[i].lock irq_context: 0 bt_proto_lock &sk->sk_peer_lock irq_context: 0 bt_proto_lock hci_sk_list.lock irq_context: 0 misc_mtx &base->lock irq_context: 0 misc_mtx &base->lock &obj_hash[i].lock irq_context: 0 (work_completion)(&(&data->open_timeout)->work) irq_context: 0 &data->open_mutex irq_context: 0 &data->open_mutex fs_reclaim irq_context: 0 &data->open_mutex fs_reclaim mmu_notifier_invalidate_range_start irq_context: 0 &data->open_mutex pool_lock#2 irq_context: 0 &data->open_mutex &____s->seqcount irq_context: 0 &data->open_mutex &obj_hash[i].lock irq_context: 0 &data->open_mutex &obj_hash[i].lock pool_lock irq_context: 0 &data->open_mutex &x->wait#9 irq_context: 0 &data->open_mutex hci_index_ida.xa_lock irq_context: 0 &data->open_mutex cpu_hotplug_lock irq_context: 0 &data->open_mutex cpu_hotplug_lock wq_pool_mutex irq_context: 0 &data->open_mutex cpu_hotplug_lock wq_pool_mutex fs_reclaim irq_context: 0 &data->open_mutex cpu_hotplug_lock wq_pool_mutex fs_reclaim mmu_notifier_invalidate_range_start irq_context: 0 &data->open_mutex cpu_hotplug_lock wq_pool_mutex pool_lock#2 irq_context: 0 &data->open_mutex cpu_hotplug_lock wq_pool_mutex &obj_hash[i].lock irq_context: 0 &data->open_mutex cpu_hotplug_lock wq_pool_mutex &wq->mutex irq_context: 0 &data->open_mutex cpu_hotplug_lock wq_pool_mutex &wq->mutex &pool->lock/1 irq_context: 0 &data->open_mutex wq_pool_mutex irq_context: 0 &data->open_mutex wq_pool_mutex &wq->mutex irq_context: 0 &data->open_mutex pin_fs_lock irq_context: 0 &data->open_mutex &sb->s_type->i_mutex_key#3 irq_context: 0 &data->open_mutex &sb->s_type->i_mutex_key#3 rename_lock.seqcount irq_context: 0 &data->open_mutex &sb->s_type->i_mutex_key#3 fs_reclaim irq_context: 0 &data->open_mutex &sb->s_type->i_mutex_key#3 fs_reclaim mmu_notifier_invalidate_range_start irq_context: 0 &data->open_mutex &sb->s_type->i_mutex_key#3 pool_lock#2 irq_context: 0 &data->open_mutex &sb->s_type->i_mutex_key#3 &dentry->d_lock irq_context: 0 &data->open_mutex &sb->s_type->i_mutex_key#3 rcu_read_lock rename_lock.seqcount irq_context: 0 &data->open_mutex &sb->s_type->i_mutex_key#3 &dentry->d_lock &wq irq_context: 0 &data->open_mutex &sb->s_type->i_mutex_key#3 mmu_notifier_invalidate_range_start irq_context: 0 &data->open_mutex &sb->s_type->i_mutex_key#3 &sb->s_type->i_lock_key#7 irq_context: 0 &data->open_mutex &sb->s_type->i_mutex_key#3 &s->s_inode_list_lock irq_context: 0 &data->open_mutex &sb->s_type->i_mutex_key#3 tk_core.seq.seqcount irq_context: 0 &data->open_mutex &sb->s_type->i_mutex_key#3 &sb->s_type->i_lock_key#7 &dentry->d_lock irq_context: 0 &data->open_mutex &k->list_lock irq_context: 0 &data->open_mutex gdp_mutex irq_context: 0 &data->open_mutex gdp_mutex &k->list_lock irq_context: 0 &data->open_mutex gdp_mutex fs_reclaim irq_context: 0 &data->open_mutex gdp_mutex fs_reclaim mmu_notifier_invalidate_range_start irq_context: 0 &data->open_mutex gdp_mutex pool_lock#2 irq_context: 0 &data->open_mutex gdp_mutex lock irq_context: 0 &data->open_mutex gdp_mutex lock kernfs_idr_lock irq_context: 0 &data->open_mutex gdp_mutex &root->kernfs_rwsem irq_context: 0 &data->open_mutex gdp_mutex &root->kernfs_rwsem &root->kernfs_iattr_rwsem irq_context: 0 &data->open_mutex lock irq_context: 0 &data->open_mutex lock kernfs_idr_lock irq_context: 0 &data->open_mutex &root->kernfs_rwsem irq_context: 0 &data->open_mutex &root->kernfs_rwsem &root->kernfs_iattr_rwsem irq_context: 0 &data->open_mutex bus_type_sem irq_context: 0 &data->open_mutex sysfs_symlink_target_lock irq_context: 0 &data->open_mutex &c->lock irq_context: 0 &data->open_mutex &root->kernfs_rwsem irq_context: 0 &data->open_mutex &dev->power.lock irq_context: 0 &data->open_mutex dpm_list_mtx irq_context: 0 &data->open_mutex uevent_sock_mutex irq_context: 0 &data->open_mutex uevent_sock_mutex fs_reclaim irq_context: 0 &data->open_mutex uevent_sock_mutex fs_reclaim mmu_notifier_invalidate_range_start irq_context: 0 &data->open_mutex uevent_sock_mutex pool_lock#2 irq_context: 0 &data->open_mutex uevent_sock_mutex nl_table_lock irq_context: 0 &data->open_mutex uevent_sock_mutex rlock-AF_NETLINK irq_context: 0 &data->open_mutex uevent_sock_mutex rcu_read_lock &ei->socket.wq.wait irq_context: 0 &data->open_mutex uevent_sock_mutex rcu_read_lock &ei->socket.wq.wait &ep->lock irq_context: 0 &data->open_mutex uevent_sock_mutex rcu_read_lock &ei->socket.wq.wait &ep->lock &ep->wq irq_context: 0 &data->open_mutex uevent_sock_mutex rcu_read_lock &ei->socket.wq.wait &ep->lock &ep->wq &p->pi_lock irq_context: 0 &data->open_mutex uevent_sock_mutex rcu_read_lock &ei->socket.wq.wait &ep->lock &ep->wq &p->pi_lock &rq->__lock irq_context: 0 &data->open_mutex uevent_sock_mutex rcu_read_lock &ei->socket.wq.wait &ep->lock &ep->wq &p->pi_lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 &data->open_mutex uevent_sock_mutex rcu_read_lock rcu_node_0 irq_context: 0 &data->open_mutex uevent_sock_mutex rcu_read_lock &rq->__lock irq_context: 0 bt_proto_lock &n->list_lock irq_context: 0 bt_proto_lock &n->list_lock &c->lock irq_context: 0 &data->open_mutex &n->list_lock irq_context: 0 &data->open_mutex &n->list_lock &c->lock irq_context: 0 &data->open_mutex cpu_hotplug_lock wq_pool_mutex &obj_hash[i].lock pool_lock irq_context: 0 &data->open_mutex cpu_hotplug_lock wq_pool_mutex &c->lock irq_context: 0 &data->open_mutex cpu_hotplug_lock wq_pool_mutex &____s->seqcount irq_context: 0 &data->open_mutex uevent_sock_mutex uevent_sock_mutex.wait_lock irq_context: 0 &data->open_mutex uevent_sock_mutex &rq->__lock irq_context: 0 &data->open_mutex uevent_sock_mutex &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 &data->open_mutex uevent_sock_mutex nl_table_wait.lock irq_context: 0 &data->open_mutex uevent_sock_mutex &obj_hash[i].lock irq_context: 0 &data->open_mutex uevent_sock_mutex.wait_lock irq_context: 0 &data->open_mutex &p->pi_lock irq_context: 0 &data->open_mutex &p->pi_lock &rq->__lock irq_context: 0 &data->open_mutex &p->pi_lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 &data->open_mutex subsys mutex#81 irq_context: 0 &data->open_mutex subsys mutex#81 &k->k_lock irq_context: 0 &data->open_mutex &dev->devres_lock irq_context: 0 &data->open_mutex triggers_list_lock irq_context: 0 &data->open_mutex leds_list_lock irq_context: 0 &data->open_mutex leds_list_lock &led_cdev->trigger_lock irq_context: 0 &data->open_mutex rfkill_global_mutex irq_context: 0 &data->open_mutex rfkill_global_mutex fs_reclaim irq_context: 0 &data->open_mutex rfkill_global_mutex fs_reclaim mmu_notifier_invalidate_range_start irq_context: 0 &data->open_mutex rfkill_global_mutex pool_lock#2 irq_context: 0 &data->open_mutex rfkill_global_mutex &c->lock irq_context: 0 &data->open_mutex rfkill_global_mutex &k->list_lock irq_context: 0 &data->open_mutex rfkill_global_mutex lock irq_context: 0 &data->open_mutex rfkill_global_mutex lock kernfs_idr_lock irq_context: 0 &data->open_mutex rfkill_global_mutex &root->kernfs_rwsem irq_context: 0 &data->open_mutex rfkill_global_mutex &root->kernfs_rwsem &root->kernfs_iattr_rwsem irq_context: 0 &data->open_mutex rfkill_global_mutex bus_type_sem irq_context: 0 &data->open_mutex rfkill_global_mutex sysfs_symlink_target_lock irq_context: 0 &data->open_mutex rfkill_global_mutex &root->kernfs_rwsem irq_context: 0 &data->open_mutex rfkill_global_mutex &dev->power.lock irq_context: 0 &data->open_mutex rfkill_global_mutex dpm_list_mtx irq_context: 0 &data->open_mutex rfkill_global_mutex &rfkill->lock irq_context: 0 &data->open_mutex rfkill_global_mutex uevent_sock_mutex irq_context: 0 &data->open_mutex rfkill_global_mutex uevent_sock_mutex fs_reclaim irq_context: 0 &data->open_mutex rfkill_global_mutex uevent_sock_mutex fs_reclaim mmu_notifier_invalidate_range_start irq_context: 0 &data->open_mutex rfkill_global_mutex uevent_sock_mutex pool_lock#2 irq_context: 0 &data->open_mutex rfkill_global_mutex uevent_sock_mutex nl_table_lock irq_context: 0 &data->open_mutex rfkill_global_mutex uevent_sock_mutex &c->lock irq_context: 0 &data->open_mutex rfkill_global_mutex uevent_sock_mutex rlock-AF_NETLINK irq_context: 0 &data->open_mutex rfkill_global_mutex uevent_sock_mutex rcu_read_lock &ei->socket.wq.wait irq_context: 0 &data->open_mutex rfkill_global_mutex uevent_sock_mutex rcu_read_lock &ei->socket.wq.wait &ep->lock irq_context: 0 &data->open_mutex rfkill_global_mutex uevent_sock_mutex nl_table_wait.lock irq_context: 0 &data->open_mutex rfkill_global_mutex uevent_sock_mutex &obj_hash[i].lock irq_context: 0 &data->open_mutex rfkill_global_mutex uevent_sock_mutex.wait_lock irq_context: 0 &data->open_mutex rfkill_global_mutex &p->pi_lock irq_context: 0 &data->open_mutex rfkill_global_mutex &obj_hash[i].lock irq_context: 0 &data->open_mutex rfkill_global_mutex &k->k_lock irq_context: 0 &data->open_mutex rfkill_global_mutex subsys mutex#41 irq_context: 0 &data->open_mutex rfkill_global_mutex subsys mutex#41 &k->k_lock irq_context: 0 &data->open_mutex rfkill_global_mutex triggers_list_lock irq_context: 0 &data->open_mutex rfkill_global_mutex leds_list_lock irq_context: 0 &data->open_mutex rfkill_global_mutex leds_list_lock &led_cdev->trigger_lock irq_context: 0 &data->open_mutex rfkill_global_mutex rcu_read_lock &pool->lock irq_context: 0 &data->open_mutex rfkill_global_mutex rcu_read_lock &pool->lock &obj_hash[i].lock irq_context: 0 &data->open_mutex rfkill_global_mutex rcu_read_lock &pool->lock pool_lock#2 irq_context: 0 &data->open_mutex rfkill_global_mutex rcu_read_lock &pool->lock &p->pi_lock irq_context: 0 &data->open_mutex rfkill_global_mutex rcu_read_lock &pool->lock &p->pi_lock &rq->__lock irq_context: 0 &data->open_mutex rfkill_global_mutex rcu_read_lock &pool->lock &p->pi_lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 &data->open_mutex rfkill_global_mutex &rq->__lock irq_context: 0 &data->open_mutex &rfkill->lock irq_context: 0 &data->open_mutex hci_dev_list_lock irq_context: 0 &data->open_mutex tk_core.seq.seqcount irq_context: 0 &data->open_mutex hci_sk_list.lock irq_context: 0 &data->open_mutex (pm_chain_head).rwsem irq_context: 0 &data->open_mutex rcu_read_lock &pool->lock/1 irq_context: 0 &data->open_mutex rcu_read_lock &pool->lock/1 &obj_hash[i].lock irq_context: 0 &data->open_mutex rcu_read_lock &pool->lock/1 &p->pi_lock irq_context: 0 &data->open_mutex rcu_read_lock &pool->lock/1 &p->pi_lock &rq->__lock irq_context: 0 &data->open_mutex rcu_read_lock &pool->lock/1 &p->pi_lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 &data->open_mutex &sb->s_type->i_mutex_key#3 &c->lock irq_context: 0 &data->open_mutex &list->lock#6 irq_context: 0 &data->open_mutex &data->read_wait irq_context: 0 &list->lock#6 irq_context: 0 (wq_completion)hci0 irq_context: 0 (wq_completion)hci0 (work_completion)(&hdev->power_on) irq_context: 0 (wq_completion)hci0 (work_completion)(&hdev->power_on) &hdev->req_lock irq_context: 0 (wq_completion)hci0 (work_completion)(&hdev->power_on) &hdev->req_lock &obj_hash[i].lock irq_context: 0 (wq_completion)hci0 (work_completion)(&hdev->power_on) &hdev->req_lock &list->lock#7 irq_context: 0 (wq_completion)hci0 (work_completion)(&hdev->power_on) &hdev->req_lock &c->lock irq_context: 0 (wq_completion)hci0 (work_completion)(&hdev->power_on) &hdev->req_lock pool_lock#2 irq_context: 0 (wq_completion)hci0 (work_completion)(&hdev->power_on) &hdev->req_lock &list->lock#8 irq_context: 0 (wq_completion)hci0 (work_completion)(&hdev->power_on) &hdev->req_lock rcu_read_lock &pool->lock/1 irq_context: 0 (wq_completion)hci0 (work_completion)(&hdev->power_on) &hdev->req_lock rcu_read_lock &pool->lock/1 &obj_hash[i].lock irq_context: 0 (wq_completion)hci0 (work_completion)(&hdev->power_on) &hdev->req_lock rcu_read_lock &pool->lock/1 &p->pi_lock irq_context: 0 (wq_completion)hci0 (work_completion)(&hdev->power_on) &hdev->req_lock rcu_read_lock &pool->lock/1 &p->pi_lock &rq->__lock irq_context: 0 (wq_completion)hci0 (work_completion)(&hdev->power_on) &hdev->req_lock rcu_read_lock &pool->lock/1 &p->pi_lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)hci0 (work_completion)(&hdev->power_on) &hdev->req_lock &hdev->req_wait_q irq_context: 0 (wq_completion)hci0 (work_completion)(&hdev->power_on) &hdev->req_lock &base->lock irq_context: 0 (wq_completion)hci0 (work_completion)(&hdev->power_on) &hdev->req_lock &base->lock &obj_hash[i].lock irq_context: 0 (wq_completion)hci0 (work_completion)(&hdev->power_on) &hdev->req_lock &rq->__lock irq_context: 0 (wq_completion)hci0 (work_completion)(&hdev->power_on) &hdev->req_lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 &data->open_mutex uevent_sock_mutex batched_entropy_u8.lock irq_context: 0 &data->open_mutex uevent_sock_mutex kfence_freelist_lock irq_context: 0 &data->open_mutex uevent_sock_mutex &meta->lock irq_context: 0 &data->open_mutex rfkill_global_mutex &n->list_lock irq_context: 0 &data->open_mutex rfkill_global_mutex &n->list_lock &c->lock irq_context: 0 &data->open_mutex rfkill_global_mutex lock kernfs_idr_lock pool_lock#2 irq_context: 0 &data->open_mutex rfkill_global_mutex.wait_lock irq_context: 0 &data->open_mutex &rq->__lock irq_context: 0 sk_lock-AF_BLUETOOTH-BTPROTO_HCI irq_context: 0 sk_lock-AF_BLUETOOTH-BTPROTO_HCI slock-AF_BLUETOOTH-BTPROTO_HCI irq_context: 0 sk_lock-AF_BLUETOOTH-BTPROTO_HCI sock_cookie_ida.xa_lock irq_context: 0 sk_lock-AF_BLUETOOTH-BTPROTO_HCI &p->alloc_lock irq_context: 0 sk_lock-AF_BLUETOOTH-BTPROTO_HCI pool_lock#2 irq_context: 0 sk_lock-AF_BLUETOOTH-BTPROTO_HCI tk_core.seq.seqcount irq_context: 0 sk_lock-AF_BLUETOOTH-BTPROTO_HCI hci_sk_list.lock irq_context: 0 sk_lock-AF_BLUETOOTH-BTPROTO_HCI &obj_hash[i].lock irq_context: 0 slock-AF_BLUETOOTH-BTPROTO_HCI irq_context: 0 hci_dev_list_lock irq_context: 0 &data->open_mutex rfkill_global_mutex &____s->seqcount irq_context: 0 (wq_completion)hci0#2 irq_context: 0 (wq_completion)hci0#2 (work_completion)(&hdev->cmd_work) irq_context: 0 (wq_completion)hci0#2 (work_completion)(&hdev->cmd_work) &list->lock#7 irq_context: 0 (wq_completion)hci0#2 (work_completion)(&hdev->cmd_work) fs_reclaim irq_context: 0 (wq_completion)hci0#2 (work_completion)(&hdev->cmd_work) fs_reclaim mmu_notifier_invalidate_range_start irq_context: 0 (wq_completion)hci0#2 (work_completion)(&hdev->cmd_work) pool_lock#2 irq_context: 0 (wq_completion)hci0#2 (work_completion)(&hdev->cmd_work) tk_core.seq.seqcount irq_context: 0 (wq_completion)hci0#2 (work_completion)(&hdev->cmd_work) &list->lock#6 irq_context: 0 (wq_completion)hci0#2 (work_completion)(&hdev->cmd_work) &data->read_wait irq_context: 0 (wq_completion)hci0#2 (work_completion)(&hdev->cmd_work) rcu_read_lock &obj_hash[i].lock irq_context: 0 (wq_completion)hci1 irq_context: 0 (wq_completion)hci1 (work_completion)(&hdev->power_on) irq_context: 0 (wq_completion)hci1 (work_completion)(&hdev->power_on) &hdev->req_lock irq_context: 0 (wq_completion)hci1 (work_completion)(&hdev->power_on) &hdev->req_lock &obj_hash[i].lock irq_context: 0 (wq_completion)hci1 (work_completion)(&hdev->power_on) &hdev->req_lock &list->lock#7 irq_context: 0 (wq_completion)hci1 (work_completion)(&hdev->power_on) &hdev->req_lock &c->lock irq_context: 0 (wq_completion)hci1 (work_completion)(&hdev->power_on) &hdev->req_lock &list->lock#8 irq_context: 0 (wq_completion)hci1 (work_completion)(&hdev->power_on) &hdev->req_lock rcu_read_lock &pool->lock/1 irq_context: 0 (wq_completion)hci1 (work_completion)(&hdev->power_on) &hdev->req_lock rcu_read_lock &pool->lock/1 &obj_hash[i].lock irq_context: 0 (wq_completion)hci1 (work_completion)(&hdev->power_on) &hdev->req_lock rcu_read_lock &pool->lock/1 &p->pi_lock irq_context: 0 (wq_completion)hci1 (work_completion)(&hdev->power_on) &hdev->req_lock &hdev->req_wait_q irq_context: 0 (wq_completion)hci1 (work_completion)(&hdev->power_on) &hdev->req_lock &base->lock irq_context: 0 (wq_completion)hci1 (work_completion)(&hdev->power_on) &hdev->req_lock &base->lock &obj_hash[i].lock irq_context: 0 (wq_completion)hci1 (work_completion)(&hdev->power_on) &hdev->req_lock &rq->__lock irq_context: 0 (wq_completion)hci1 (work_completion)(&hdev->power_on) &hdev->req_lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 &list->lock#7 irq_context: 0 &data->read_wait irq_context: 0 &data->open_mutex &sb->s_type->i_mutex_key#3 &n->list_lock irq_context: 0 &data->open_mutex &sb->s_type->i_mutex_key#3 &n->list_lock &c->lock irq_context: 0 (wq_completion)hci1#2 irq_context: 0 (wq_completion)hci1#2 (work_completion)(&hdev->cmd_work) irq_context: 0 (wq_completion)hci1#2 (work_completion)(&hdev->cmd_work) &list->lock#7 irq_context: 0 (wq_completion)hci1#2 (work_completion)(&hdev->cmd_work) fs_reclaim irq_context: 0 (wq_completion)hci1#2 (work_completion)(&hdev->cmd_work) fs_reclaim mmu_notifier_invalidate_range_start irq_context: 0 (wq_completion)hci1#2 (work_completion)(&hdev->cmd_work) tk_core.seq.seqcount irq_context: 0 (wq_completion)hci1#2 (work_completion)(&hdev->cmd_work) &list->lock#6 irq_context: 0 (wq_completion)hci1#2 (work_completion)(&hdev->cmd_work) &data->read_wait irq_context: 0 (wq_completion)hci1#2 (work_completion)(&hdev->cmd_work) rcu_read_lock &obj_hash[i].lock irq_context: 0 (wq_completion)hci0#2 (work_completion)(&hdev->rx_work) irq_context: 0 (wq_completion)hci0#2 (work_completion)(&hdev->rx_work) &list->lock#7 irq_context: 0 (wq_completion)hci0#2 (work_completion)(&hdev->rx_work) lock#6 irq_context: 0 (wq_completion)hci0#2 (work_completion)(&hdev->rx_work) lock#6 kcov_remote_lock irq_context: 0 (wq_completion)hci0#2 (work_completion)(&hdev->rx_work) fs_reclaim irq_context: 0 (wq_completion)hci0#2 (work_completion)(&hdev->rx_work) fs_reclaim mmu_notifier_invalidate_range_start irq_context: 0 (wq_completion)hci0#2 (work_completion)(&hdev->rx_work) pool_lock#2 irq_context: 0 (wq_completion)hci0#2 (work_completion)(&hdev->rx_work) free_vmap_area_lock irq_context: 0 (wq_completion)hci0#2 (work_completion)(&hdev->rx_work) vmap_area_lock irq_context: 0 (wq_completion)hci0#2 (work_completion)(&hdev->rx_work) &____s->seqcount irq_context: 0 (wq_completion)hci0#2 (work_completion)(&hdev->rx_work) init_mm.page_table_lock irq_context: 0 (wq_completion)hci0#2 (work_completion)(&hdev->rx_work) &pcp->lock &zone->lock irq_context: 0 (wq_completion)hci0#2 (work_completion)(&hdev->rx_work) &pcp->lock &zone->lock &____s->seqcount irq_context: 0 sk_lock-AF_BLUETOOTH-BTPROTO_HCI &c->lock irq_context: 0 (wq_completion)hci2 irq_context: 0 (wq_completion)hci0#2 (work_completion)(&hdev->rx_work) &c->lock irq_context: 0 (wq_completion)hci0#2 (work_completion)(&hdev->rx_work) &hdev->lock irq_context: 0 (wq_completion)hci0#2 (work_completion)(&hdev->rx_work) (console_sem).lock irq_context: 0 (wq_completion)hci0#2 (work_completion)(&hdev->rx_work) console_lock console_srcu console_owner_lock irq_context: 0 (wq_completion)hci0#2 (work_completion)(&hdev->rx_work) console_lock console_srcu console_owner irq_context: 0 (wq_completion)hci0#2 (work_completion)(&hdev->rx_work) console_lock console_srcu console_owner &port_lock_key irq_context: 0 &data->open_mutex rfkill_global_mutex &root->kernfs_rwsem &rq->__lock irq_context: 0 &data->open_mutex rfkill_global_mutex &root->kernfs_rwsem &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)hci0#2 (work_completion)(&hdev->rx_work) console_lock console_srcu console_owner console_owner_lock irq_context: 0 (wq_completion)hci0#2 (work_completion)(&hdev->rx_work) &rq->__lock irq_context: 0 (wq_completion)hci0#2 (work_completion)(&hdev->rx_work) &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)hci0#2 (work_completion)(&hdev->rx_work) &obj_hash[i].lock irq_context: 0 (wq_completion)hci0#2 (work_completion)(&hdev->rx_work) rcu_read_lock &obj_hash[i].lock irq_context: 0 (wq_completion)hci0#2 (work_completion)(&hdev->rx_work) &hdev->req_wait_q irq_context: 0 (wq_completion)hci0#2 (work_completion)(&hdev->rx_work) &hdev->req_wait_q &p->pi_lock irq_context: 0 (wq_completion)hci0#2 (work_completion)(&hdev->rx_work) &hdev->req_wait_q &p->pi_lock &rq->__lock irq_context: 0 (wq_completion)hci0#2 (work_completion)(&hdev->rx_work) &hdev->req_wait_q &p->pi_lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)hci0 (work_completion)(&hdev->power_on) &hdev->req_lock (&timer.timer) irq_context: 0 (wq_completion)hci0#2 (work_completion)(&hdev->rx_work) lock#6 &kcov->lock irq_context: 0 (wq_completion)hci2 (work_completion)(&hdev->power_on) irq_context: 0 (wq_completion)hci2 (work_completion)(&hdev->power_on) &hdev->req_lock irq_context: 0 (wq_completion)hci2 (work_completion)(&hdev->power_on) &hdev->req_lock &obj_hash[i].lock irq_context: 0 (wq_completion)hci2 (work_completion)(&hdev->power_on) &hdev->req_lock &list->lock#7 irq_context: 0 (wq_completion)hci2 (work_completion)(&hdev->power_on) &hdev->req_lock &list->lock#8 irq_context: 0 (wq_completion)hci2 (work_completion)(&hdev->power_on) &hdev->req_lock rcu_read_lock &pool->lock/1 irq_context: 0 (wq_completion)hci2 (work_completion)(&hdev->power_on) &hdev->req_lock rcu_read_lock &pool->lock/1 &obj_hash[i].lock irq_context: 0 (wq_completion)hci2 (work_completion)(&hdev->power_on) &hdev->req_lock &hdev->req_wait_q irq_context: 0 (wq_completion)hci2 (work_completion)(&hdev->power_on) &hdev->req_lock &base->lock irq_context: 0 (wq_completion)hci2 (work_completion)(&hdev->power_on) &hdev->req_lock &base->lock &obj_hash[i].lock irq_context: 0 (wq_completion)hci2 (work_completion)(&hdev->power_on) &hdev->req_lock &rq->__lock irq_context: 0 (wq_completion)hci2 (work_completion)(&hdev->power_on) &hdev->req_lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 &data->open_mutex lock kernfs_idr_lock pool_lock#2 irq_context: 0 &data->open_mutex &root->kernfs_rwsem &sem->wait_lock irq_context: 0 &data->open_mutex &root->kernfs_rwsem &rq->__lock irq_context: 0 &data->open_mutex &root->kernfs_rwsem &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)hci1#2 (work_completion)(&hdev->rx_work) irq_context: 0 (wq_completion)hci1#2 (work_completion)(&hdev->rx_work) &list->lock#7 irq_context: 0 (wq_completion)hci1#2 (work_completion)(&hdev->rx_work) lock#6 irq_context: 0 (wq_completion)hci1#2 (work_completion)(&hdev->rx_work) lock#6 kcov_remote_lock irq_context: 0 (wq_completion)hci1#2 (work_completion)(&hdev->rx_work) fs_reclaim irq_context: 0 (wq_completion)hci1#2 (work_completion)(&hdev->rx_work) fs_reclaim mmu_notifier_invalidate_range_start irq_context: 0 (wq_completion)hci1#2 (work_completion)(&hdev->rx_work) &c->lock irq_context: 0 (wq_completion)hci1#2 (work_completion)(&hdev->rx_work) &hdev->lock irq_context: 0 (wq_completion)hci1#2 (work_completion)(&hdev->rx_work) (console_sem).lock irq_context: 0 (wq_completion)hci1#2 (work_completion)(&hdev->rx_work) console_lock console_srcu console_owner_lock irq_context: 0 (wq_completion)hci1#2 (work_completion)(&hdev->rx_work) console_lock console_srcu console_owner irq_context: 0 (wq_completion)hci1#2 (work_completion)(&hdev->rx_work) console_lock console_srcu console_owner &port_lock_key irq_context: 0 (wq_completion)hci1#2 (work_completion)(&hdev->rx_work) console_lock console_srcu console_owner console_owner_lock irq_context: 0 (wq_completion)hci1#2 (work_completion)(&hdev->rx_work) &rq->__lock irq_context: 0 (wq_completion)hci1#2 (work_completion)(&hdev->rx_work) &obj_hash[i].lock irq_context: 0 (wq_completion)hci1#2 (work_completion)(&hdev->rx_work) rcu_read_lock &obj_hash[i].lock irq_context: 0 (wq_completion)hci1#2 (work_completion)(&hdev->rx_work) &hdev->req_wait_q irq_context: 0 (wq_completion)hci1#2 (work_completion)(&hdev->rx_work) &hdev->req_wait_q &p->pi_lock irq_context: 0 (wq_completion)hci1#2 (work_completion)(&hdev->rx_work) &hdev->req_wait_q &p->pi_lock &cfs_rq->removed.lock irq_context: 0 (wq_completion)hci1#2 (work_completion)(&hdev->rx_work) &hdev->req_wait_q &p->pi_lock &rq->__lock irq_context: 0 (wq_completion)hci1#2 (work_completion)(&hdev->rx_work) &hdev->req_wait_q &p->pi_lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)hci1#2 (work_completion)(&hdev->rx_work) lock#6 &kcov->lock irq_context: 0 (wq_completion)hci3 irq_context: 0 (wq_completion)hci3 (work_completion)(&hdev->power_on) irq_context: 0 (wq_completion)hci3 (work_completion)(&hdev->power_on) &hdev->req_lock irq_context: 0 (wq_completion)hci3 (work_completion)(&hdev->power_on) &hdev->req_lock &obj_hash[i].lock irq_context: 0 (wq_completion)hci3 (work_completion)(&hdev->power_on) &hdev->req_lock &list->lock#7 irq_context: 0 (wq_completion)hci3 (work_completion)(&hdev->power_on) &hdev->req_lock &list->lock#8 irq_context: 0 (wq_completion)hci3 (work_completion)(&hdev->power_on) &hdev->req_lock rcu_read_lock &pool->lock/1 irq_context: 0 (wq_completion)hci3 (work_completion)(&hdev->power_on) &hdev->req_lock rcu_read_lock &pool->lock/1 &obj_hash[i].lock irq_context: 0 (wq_completion)hci3 (work_completion)(&hdev->power_on) &hdev->req_lock &hdev->req_wait_q irq_context: 0 (wq_completion)hci3 (work_completion)(&hdev->power_on) &hdev->req_lock &base->lock irq_context: 0 (wq_completion)hci3 (work_completion)(&hdev->power_on) &hdev->req_lock &base->lock &obj_hash[i].lock irq_context: 0 (wq_completion)hci3 (work_completion)(&hdev->power_on) &hdev->req_lock &rq->__lock irq_context: 0 (wq_completion)hci3 (work_completion)(&hdev->power_on) &hdev->req_lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)hci1 (work_completion)(&hdev->power_on) &hdev->req_lock (&timer.timer) irq_context: 0 (wq_completion)hci0#2 (work_completion)(&hdev->cmd_work) &obj_hash[i].lock irq_context: 0 (wq_completion)hci0#2 (work_completion)(&hdev->cmd_work) &data->read_wait &p->pi_lock irq_context: 0 (wq_completion)hci0#2 (work_completion)(&hdev->cmd_work) &data->read_wait &p->pi_lock &rq->__lock irq_context: 0 (wq_completion)hci0#2 (work_completion)(&hdev->cmd_work) &data->read_wait &p->pi_lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 sk_lock-AF_BLUETOOTH-BTPROTO_HCI &____s->seqcount irq_context: 0 (wq_completion)hci0#2 (work_completion)(&hdev->cmd_work) rcu_read_lock &base->lock irq_context: 0 (wq_completion)hci0#2 (work_completion)(&hdev->cmd_work) rcu_read_lock &base->lock &obj_hash[i].lock irq_context: 0 (wq_completion)hci2#2 irq_context: 0 (wq_completion)hci2#2 (work_completion)(&hdev->cmd_work) irq_context: 0 (wq_completion)hci2#2 (work_completion)(&hdev->cmd_work) &list->lock#7 irq_context: 0 (wq_completion)hci2#2 (work_completion)(&hdev->cmd_work) fs_reclaim irq_context: 0 (wq_completion)hci2#2 (work_completion)(&hdev->cmd_work) fs_reclaim mmu_notifier_invalidate_range_start irq_context: 0 (wq_completion)hci2#2 (work_completion)(&hdev->cmd_work) tk_core.seq.seqcount irq_context: 0 (wq_completion)hci2#2 (work_completion)(&hdev->cmd_work) &list->lock#6 irq_context: 0 (wq_completion)hci2#2 (work_completion)(&hdev->cmd_work) &data->read_wait irq_context: 0 (wq_completion)hci2#2 (work_completion)(&hdev->cmd_work) &data->read_wait &p->pi_lock irq_context: 0 (wq_completion)hci2#2 (work_completion)(&hdev->cmd_work) &data->read_wait &p->pi_lock &rq->__lock irq_context: 0 (wq_completion)hci2#2 (work_completion)(&hdev->cmd_work) &data->read_wait &p->pi_lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)hci2#2 (work_completion)(&hdev->cmd_work) rcu_read_lock &obj_hash[i].lock irq_context: 0 (wq_completion)hci4 irq_context: 0 (wq_completion)hci4 (work_completion)(&hdev->power_on) irq_context: 0 (wq_completion)hci4 (work_completion)(&hdev->power_on) &hdev->req_lock irq_context: 0 (wq_completion)hci4 (work_completion)(&hdev->power_on) &hdev->req_lock &obj_hash[i].lock irq_context: 0 (wq_completion)hci4 (work_completion)(&hdev->power_on) &hdev->req_lock &list->lock#7 irq_context: 0 (wq_completion)hci4 (work_completion)(&hdev->power_on) &hdev->req_lock &c->lock irq_context: 0 (wq_completion)hci4 (work_completion)(&hdev->power_on) &hdev->req_lock &list->lock#8 irq_context: 0 (wq_completion)hci4 (work_completion)(&hdev->power_on) &hdev->req_lock rcu_read_lock &pool->lock/1 irq_context: 0 (wq_completion)hci4 (work_completion)(&hdev->power_on) &hdev->req_lock rcu_read_lock &pool->lock/1 &obj_hash[i].lock irq_context: 0 (wq_completion)hci4 (work_completion)(&hdev->power_on) &hdev->req_lock &hdev->req_wait_q irq_context: 0 (wq_completion)hci4 (work_completion)(&hdev->power_on) &hdev->req_lock &base->lock irq_context: 0 (wq_completion)hci4 (work_completion)(&hdev->power_on) &hdev->req_lock &base->lock &obj_hash[i].lock irq_context: 0 (wq_completion)hci4 (work_completion)(&hdev->power_on) &hdev->req_lock &rq->__lock irq_context: 0 (wq_completion)hci4 (work_completion)(&hdev->power_on) &hdev->req_lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)events (work_completion)(&ht->run_work) &ht->mutex batched_entropy_u32.lock crngs.lock irq_context: 0 bt_proto_lock &obj_hash[i].lock pool_lock irq_context: 0 (wq_completion)hci3#2 irq_context: 0 (wq_completion)hci3#2 (work_completion)(&hdev->cmd_work) irq_context: 0 (wq_completion)hci3#2 (work_completion)(&hdev->cmd_work) &list->lock#7 irq_context: 0 (wq_completion)hci3#2 (work_completion)(&hdev->cmd_work) fs_reclaim irq_context: 0 (wq_completion)hci3#2 (work_completion)(&hdev->cmd_work) fs_reclaim mmu_notifier_invalidate_range_start irq_context: 0 (wq_completion)hci3#2 (work_completion)(&hdev->cmd_work) &c->lock irq_context: 0 (wq_completion)hci3#2 (work_completion)(&hdev->cmd_work) &____s->seqcount irq_context: 0 (wq_completion)hci3#2 (work_completion)(&hdev->cmd_work) pool_lock#2 irq_context: 0 (wq_completion)hci3#2 (work_completion)(&hdev->cmd_work) tk_core.seq.seqcount irq_context: 0 (wq_completion)hci3#2 (work_completion)(&hdev->cmd_work) &list->lock#6 irq_context: 0 (wq_completion)hci3#2 (work_completion)(&hdev->cmd_work) &data->read_wait irq_context: 0 (wq_completion)hci1#2 (work_completion)(&hdev->cmd_work) &obj_hash[i].lock irq_context: 0 (wq_completion)hci3#2 (work_completion)(&hdev->cmd_work) &data->read_wait &p->pi_lock irq_context: 0 (wq_completion)hci3#2 (work_completion)(&hdev->cmd_work) &data->read_wait &p->pi_lock &rq->__lock irq_context: 0 (wq_completion)hci1#2 (work_completion)(&hdev->cmd_work) &data->read_wait &p->pi_lock irq_context: 0 (wq_completion)hci3#2 (work_completion)(&hdev->cmd_work) &data->read_wait &p->pi_lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)hci1#2 (work_completion)(&hdev->cmd_work) &data->read_wait &p->pi_lock &rq->__lock irq_context: 0 (wq_completion)hci3#2 (work_completion)(&hdev->cmd_work) rcu_read_lock &obj_hash[i].lock irq_context: 0 (wq_completion)hci1#2 (work_completion)(&hdev->cmd_work) &data->read_wait &p->pi_lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)hci2#2 (work_completion)(&hdev->rx_work) irq_context: 0 (wq_completion)hci1#2 (work_completion)(&hdev->cmd_work) rcu_read_lock &base->lock irq_context: 0 (wq_completion)hci2#2 (work_completion)(&hdev->rx_work) &list->lock#7 irq_context: 0 (wq_completion)hci1#2 (work_completion)(&hdev->cmd_work) rcu_read_lock &base->lock &obj_hash[i].lock irq_context: 0 (wq_completion)hci2#2 (work_completion)(&hdev->rx_work) lock#6 irq_context: 0 (wq_completion)hci2#2 (work_completion)(&hdev->rx_work) lock#6 kcov_remote_lock irq_context: 0 (wq_completion)hci2#2 (work_completion)(&hdev->rx_work) fs_reclaim irq_context: 0 (wq_completion)hci2#2 (work_completion)(&hdev->rx_work) fs_reclaim mmu_notifier_invalidate_range_start irq_context: 0 (wq_completion)hci4#2 irq_context: 0 (wq_completion)hci4#2 (work_completion)(&hdev->cmd_work) irq_context: 0 (wq_completion)hci2#2 (work_completion)(&hdev->rx_work) &hdev->lock irq_context: 0 (wq_completion)hci4#2 (work_completion)(&hdev->cmd_work) &list->lock#7 irq_context: 0 (wq_completion)hci4#2 (work_completion)(&hdev->cmd_work) fs_reclaim irq_context: 0 (wq_completion)hci4#2 (work_completion)(&hdev->cmd_work) fs_reclaim mmu_notifier_invalidate_range_start irq_context: 0 (wq_completion)hci2#2 (work_completion)(&hdev->rx_work) (console_sem).lock irq_context: 0 (wq_completion)hci4#2 (work_completion)(&hdev->cmd_work) tk_core.seq.seqcount irq_context: 0 (wq_completion)hci4#2 (work_completion)(&hdev->cmd_work) &list->lock#6 irq_context: 0 (wq_completion)hci2#2 (work_completion)(&hdev->rx_work) console_lock console_srcu console_owner_lock irq_context: 0 (wq_completion)hci4#2 (work_completion)(&hdev->cmd_work) &data->read_wait irq_context: 0 (wq_completion)hci2#2 (work_completion)(&hdev->rx_work) console_lock console_srcu console_owner irq_context: 0 (wq_completion)hci4#2 (work_completion)(&hdev->cmd_work) &data->read_wait &p->pi_lock irq_context: 0 (wq_completion)hci2#2 (work_completion)(&hdev->rx_work) console_lock console_srcu console_owner &port_lock_key irq_context: 0 (wq_completion)hci4#2 (work_completion)(&hdev->cmd_work) &data->read_wait &p->pi_lock &rq->__lock irq_context: 0 (wq_completion)hci4#2 (work_completion)(&hdev->cmd_work) &data->read_wait &p->pi_lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)hci4#2 (work_completion)(&hdev->cmd_work) rcu_read_lock &obj_hash[i].lock irq_context: 0 (wq_completion)hci2#2 (work_completion)(&hdev->rx_work) console_lock console_srcu console_owner console_owner_lock irq_context: 0 (wq_completion)hci2#2 (work_completion)(&hdev->rx_work) &rq->__lock irq_context: 0 (wq_completion)hci2#2 (work_completion)(&hdev->rx_work) &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)hci2#2 (work_completion)(&hdev->rx_work) &obj_hash[i].lock irq_context: 0 (wq_completion)hci2#2 (work_completion)(&hdev->rx_work) rcu_read_lock &obj_hash[i].lock irq_context: 0 (wq_completion)hci2#2 (work_completion)(&hdev->rx_work) &hdev->req_wait_q irq_context: 0 (wq_completion)hci2#2 (work_completion)(&hdev->rx_work) &hdev->req_wait_q &p->pi_lock irq_context: 0 (wq_completion)hci2#2 (work_completion)(&hdev->rx_work) &hdev->req_wait_q &p->pi_lock &rq->__lock irq_context: 0 (wq_completion)hci2#2 (work_completion)(&hdev->rx_work) &hdev->req_wait_q &p->pi_lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)hci2#2 (work_completion)(&hdev->rx_work) lock#6 &kcov->lock irq_context: 0 (wq_completion)hci2 (work_completion)(&hdev->power_on) &hdev->req_lock (&timer.timer) irq_context: 0 (wq_completion)hci0#2 (work_completion)(&hdev->rx_work) console_owner_lock irq_context: 0 (wq_completion)hci0#2 (work_completion)(&hdev->rx_work) console_owner irq_context: 0 (wq_completion)hci1#2 (work_completion)(&hdev->rx_work) &base->lock irq_context: 0 (wq_completion)hci1#2 (work_completion)(&hdev->rx_work) &base->lock &obj_hash[i].lock irq_context: 0 (wq_completion)hci1 (work_completion)(&hdev->power_on) &hdev->req_lock pool_lock#2 irq_context: 0 (wq_completion)hci1#2 (work_completion)(&hdev->rx_work) pool_lock#2 irq_context: 0 (wq_completion)hci2#2 (work_completion)(&hdev->cmd_work) &obj_hash[i].lock irq_context: 0 (wq_completion)hci2#2 (work_completion)(&hdev->cmd_work) rcu_read_lock &base->lock irq_context: 0 (wq_completion)hci2#2 (work_completion)(&hdev->cmd_work) rcu_read_lock &base->lock &obj_hash[i].lock irq_context: 0 (wq_completion)hci1#2 (work_completion)(&hdev->rx_work) console_owner_lock irq_context: 0 (wq_completion)hci1#2 (work_completion)(&hdev->rx_work) console_owner irq_context: 0 (wq_completion)hci0#2 (work_completion)(&hdev->rx_work) &base->lock irq_context: 0 (wq_completion)hci0#2 (work_completion)(&hdev->rx_work) &base->lock &obj_hash[i].lock irq_context: 0 (wq_completion)hci0#2 (work_completion)(&hdev->rx_work) &hdev->req_wait_q &p->pi_lock &cfs_rq->removed.lock irq_context: 0 (wq_completion)hci0#2 (work_completion)(&hdev->cmd_work) &rq->__lock irq_context: 0 (wq_completion)hci1#2 (work_completion)(&hdev->cmd_work) rcu_read_lock rcu_node_0 irq_context: 0 (wq_completion)hci1#2 (work_completion)(&hdev->cmd_work) rcu_read_lock &rq->__lock irq_context: 0 (wq_completion)hci4#2 (work_completion)(&hdev->rx_work) irq_context: 0 (wq_completion)hci4#2 (work_completion)(&hdev->rx_work) &list->lock#7 irq_context: 0 (wq_completion)hci4#2 (work_completion)(&hdev->rx_work) lock#6 irq_context: 0 (wq_completion)hci4#2 (work_completion)(&hdev->rx_work) lock#6 kcov_remote_lock irq_context: 0 (wq_completion)hci4#2 (work_completion)(&hdev->rx_work) fs_reclaim irq_context: 0 (wq_completion)hci4#2 (work_completion)(&hdev->rx_work) fs_reclaim mmu_notifier_invalidate_range_start irq_context: 0 (wq_completion)hci4#2 (work_completion)(&hdev->rx_work) &hdev->lock irq_context: 0 (wq_completion)hci4#2 (work_completion)(&hdev->rx_work) (console_sem).lock irq_context: 0 (wq_completion)hci4#2 (work_completion)(&hdev->rx_work) console_owner_lock irq_context: 0 (wq_completion)hci4#2 (work_completion)(&hdev->rx_work) console_owner irq_context: softirq (&pool->mayday_timer) &pool->lock/1 irq_context: softirq (&pool->mayday_timer) &pool->lock/1 wq_mayday_lock irq_context: softirq (&pool->mayday_timer) &obj_hash[i].lock irq_context: softirq (&pool->mayday_timer) &base->lock irq_context: softirq (&pool->mayday_timer) &base->lock &obj_hash[i].lock irq_context: 0 (wq_completion)hci4#2 (work_completion)(&hdev->rx_work) console_lock console_srcu console_owner_lock irq_context: 0 (wq_completion)hci0#2 (work_completion)(&hdev->cmd_work) rcu_read_lock rcu_node_0 irq_context: 0 (wq_completion)hci4#2 (work_completion)(&hdev->rx_work) console_lock console_srcu console_owner irq_context: 0 (wq_completion)hci0#2 (work_completion)(&hdev->cmd_work) rcu_read_lock &rq->__lock irq_context: 0 (wq_completion)hci4#2 (work_completion)(&hdev->rx_work) console_lock console_srcu console_owner &port_lock_key irq_context: 0 (wq_completion)hci1#2 (work_completion)(&hdev->cmd_work) &c->lock irq_context: 0 (wq_completion)hci1#2 (work_completion)(&hdev->cmd_work) &rq->__lock irq_context: 0 (wq_completion)hci1#2 (work_completion)(&hdev->cmd_work) &____s->seqcount irq_context: 0 (wq_completion)hci2#2 (work_completion)(&hdev->rx_work) console_owner_lock irq_context: 0 (wq_completion)hci2#2 (work_completion)(&hdev->rx_work) console_owner irq_context: 0 (wq_completion)hci4#2 (work_completion)(&hdev->rx_work) console_lock console_srcu console_owner console_owner_lock irq_context: 0 (wq_completion)hci4#2 (work_completion)(&hdev->rx_work) &rq->__lock irq_context: 0 (wq_completion)hci4#2 (work_completion)(&hdev->rx_work) &obj_hash[i].lock irq_context: 0 (wq_completion)hci4#2 (work_completion)(&hdev->rx_work) rcu_read_lock &obj_hash[i].lock irq_context: 0 (wq_completion)hci4#2 (work_completion)(&hdev->rx_work) &hdev->req_wait_q irq_context: 0 (wq_completion)hci4#2 (work_completion)(&hdev->rx_work) &hdev->req_wait_q &p->pi_lock irq_context: 0 (wq_completion)hci4#2 (work_completion)(&hdev->rx_work) &hdev->req_wait_q &p->pi_lock &rq->__lock irq_context: 0 (wq_completion)hci4#2 (work_completion)(&hdev->rx_work) &hdev->req_wait_q &p->pi_lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)hci4 (work_completion)(&hdev->power_on) &hdev->req_lock (&timer.timer) irq_context: 0 (wq_completion)hci4#2 (work_completion)(&hdev->rx_work) lock#6 &kcov->lock irq_context: 0 (wq_completion)hci4#2 (work_completion)(&hdev->cmd_work) &obj_hash[i].lock irq_context: 0 (wq_completion)hci4#2 (work_completion)(&hdev->cmd_work) &rq->__lock irq_context: 0 (wq_completion)hci4#2 (work_completion)(&hdev->cmd_work) rcu_read_lock &base->lock irq_context: 0 (wq_completion)hci4#2 (work_completion)(&hdev->cmd_work) rcu_read_lock &base->lock &obj_hash[i].lock irq_context: 0 (wq_completion)hci2#2 (work_completion)(&hdev->rx_work) &base->lock irq_context: 0 (wq_completion)hci2#2 (work_completion)(&hdev->rx_work) &base->lock &obj_hash[i].lock irq_context: 0 (wq_completion)hci0 (work_completion)(&hdev->power_on) &hdev->req_lock batched_entropy_u8.lock irq_context: 0 (wq_completion)hci0 (work_completion)(&hdev->power_on) &hdev->req_lock kfence_freelist_lock irq_context: 0 (wq_completion)hci0 (work_completion)(&hdev->power_on) &hdev->req_lock &____s->seqcount irq_context: 0 (wq_completion)hci2 (work_completion)(&hdev->power_on) &hdev->req_lock pool_lock#2 irq_context: 0 (wq_completion)hci4#2 (work_completion)(&hdev->rx_work) &base->lock irq_context: 0 (wq_completion)hci4#2 (work_completion)(&hdev->rx_work) &base->lock &obj_hash[i].lock irq_context: 0 (wq_completion)hci0#2 (work_completion)(&hdev->cmd_work) &data->read_wait &p->pi_lock &cfs_rq->removed.lock irq_context: 0 (wq_completion)hci4 (work_completion)(&hdev->power_on) &hdev->req_lock rcu_read_lock &pool->lock/1 &p->pi_lock irq_context: 0 (wq_completion)hci4 (work_completion)(&hdev->power_on) &hdev->req_lock rcu_read_lock &pool->lock/1 &p->pi_lock &rq->__lock irq_context: 0 (wq_completion)hci4 (work_completion)(&hdev->power_on) &hdev->req_lock rcu_read_lock &pool->lock/1 &p->pi_lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)hci4#2 (work_completion)(&hdev->rx_work) &c->lock irq_context: 0 (wq_completion)hci4#2 (work_completion)(&hdev->rx_work) &____s->seqcount irq_context: 0 (wq_completion)hci4#2 (work_completion)(&hdev->rx_work) pool_lock#2 irq_context: 0 (wq_completion)hci2#2 (work_completion)(&hdev->rx_work) &c->lock irq_context: 0 (wq_completion)hci2#2 (work_completion)(&hdev->rx_work) &____s->seqcount irq_context: 0 (wq_completion)hci4 (work_completion)(&hdev->power_on) &hdev->req_lock pool_lock#2 irq_context: 0 (wq_completion)hci4#2 (work_completion)(&hdev->cmd_work) &c->lock irq_context: 0 (wq_completion)hci4#2 (work_completion)(&hdev->cmd_work) &____s->seqcount irq_context: 0 (wq_completion)hci5 irq_context: 0 (wq_completion)hci5 (work_completion)(&hdev->power_on) irq_context: 0 (wq_completion)hci5 (work_completion)(&hdev->power_on) &hdev->req_lock irq_context: 0 (wq_completion)hci5 (work_completion)(&hdev->power_on) &hdev->req_lock &obj_hash[i].lock irq_context: 0 (wq_completion)hci5 (work_completion)(&hdev->power_on) &hdev->req_lock &list->lock#7 irq_context: 0 (wq_completion)hci5 (work_completion)(&hdev->power_on) &hdev->req_lock &list->lock#8 irq_context: 0 (wq_completion)hci5 (work_completion)(&hdev->power_on) &hdev->req_lock rcu_read_lock &pool->lock/1 irq_context: 0 (wq_completion)hci5 (work_completion)(&hdev->power_on) &hdev->req_lock rcu_read_lock &pool->lock/1 &obj_hash[i].lock irq_context: 0 (wq_completion)hci5 (work_completion)(&hdev->power_on) &hdev->req_lock &hdev->req_wait_q irq_context: 0 (wq_completion)hci5 (work_completion)(&hdev->power_on) &hdev->req_lock &base->lock irq_context: 0 (wq_completion)hci5 (work_completion)(&hdev->power_on) &hdev->req_lock &base->lock &obj_hash[i].lock irq_context: 0 (wq_completion)hci5 (work_completion)(&hdev->power_on) &hdev->req_lock &rq->__lock irq_context: 0 (wq_completion)hci5 (work_completion)(&hdev->power_on) &hdev->req_lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)hci5#2 irq_context: 0 (wq_completion)hci5#2 (work_completion)(&hdev->cmd_work) irq_context: 0 (wq_completion)hci5#2 (work_completion)(&hdev->cmd_work) &list->lock#7 irq_context: 0 (wq_completion)hci5#2 (work_completion)(&hdev->cmd_work) fs_reclaim irq_context: 0 (wq_completion)hci5#2 (work_completion)(&hdev->cmd_work) fs_reclaim mmu_notifier_invalidate_range_start irq_context: 0 (wq_completion)hci5#2 (work_completion)(&hdev->cmd_work) tk_core.seq.seqcount irq_context: 0 (wq_completion)hci5#2 (work_completion)(&hdev->cmd_work) &list->lock#6 irq_context: 0 (wq_completion)hci5#2 (work_completion)(&hdev->cmd_work) &data->read_wait irq_context: 0 (wq_completion)hci5#2 (work_completion)(&hdev->cmd_work) &data->read_wait &p->pi_lock irq_context: 0 (wq_completion)hci5#2 (work_completion)(&hdev->cmd_work) &data->read_wait &p->pi_lock &rq->__lock irq_context: 0 (wq_completion)hci5#2 (work_completion)(&hdev->cmd_work) &data->read_wait &p->pi_lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)hci5#2 (work_completion)(&hdev->cmd_work) rcu_read_lock &obj_hash[i].lock irq_context: 0 (wq_completion)hci3#2 (work_completion)(&hdev->rx_work) irq_context: 0 (wq_completion)hci3#2 (work_completion)(&hdev->rx_work) &list->lock#7 irq_context: 0 (wq_completion)hci3#2 (work_completion)(&hdev->rx_work) lock#6 irq_context: 0 (wq_completion)hci3#2 (work_completion)(&hdev->rx_work) lock#6 kcov_remote_lock irq_context: 0 (wq_completion)hci3#2 (work_completion)(&hdev->rx_work) fs_reclaim irq_context: 0 (wq_completion)hci3#2 (work_completion)(&hdev->rx_work) fs_reclaim mmu_notifier_invalidate_range_start irq_context: 0 (wq_completion)hci3#2 (work_completion)(&hdev->rx_work) &c->lock irq_context: 0 (wq_completion)hci3#2 (work_completion)(&hdev->rx_work) &____s->seqcount irq_context: 0 (wq_completion)hci3#2 (work_completion)(&hdev->rx_work) &hdev->lock irq_context: 0 (wq_completion)hci3#2 (work_completion)(&hdev->rx_work) (console_sem).lock irq_context: 0 (wq_completion)hci3#2 (work_completion)(&hdev->rx_work) console_lock console_srcu console_owner_lock irq_context: 0 (wq_completion)hci3#2 (work_completion)(&hdev->rx_work) console_lock console_srcu console_owner irq_context: 0 (wq_completion)hci3#2 (work_completion)(&hdev->rx_work) console_lock console_srcu console_owner &port_lock_key irq_context: 0 (wq_completion)hci3#2 (work_completion)(&hdev->rx_work) console_lock console_srcu console_owner console_owner_lock irq_context: 0 (wq_completion)hci3#2 (work_completion)(&hdev->rx_work) &rq->__lock irq_context: 0 (wq_completion)hci5#2 (work_completion)(&hdev->rx_work) irq_context: 0 (wq_completion)hci5#2 (work_completion)(&hdev->rx_work) &list->lock#7 irq_context: 0 (wq_completion)hci5#2 (work_completion)(&hdev->rx_work) lock#6 irq_context: 0 (wq_completion)hci5#2 (work_completion)(&hdev->rx_work) lock#6 kcov_remote_lock irq_context: 0 (wq_completion)hci5#2 (work_completion)(&hdev->rx_work) fs_reclaim irq_context: 0 (wq_completion)hci5#2 (work_completion)(&hdev->rx_work) fs_reclaim mmu_notifier_invalidate_range_start irq_context: 0 (wq_completion)hci5#2 (work_completion)(&hdev->rx_work) &hdev->lock irq_context: 0 (wq_completion)hci5#2 (work_completion)(&hdev->rx_work) (console_sem).lock irq_context: 0 (wq_completion)hci5#2 (work_completion)(&hdev->rx_work) console_lock console_srcu console_owner_lock irq_context: 0 (wq_completion)hci5#2 (work_completion)(&hdev->rx_work) console_lock console_srcu console_owner irq_context: 0 (wq_completion)hci5#2 (work_completion)(&hdev->rx_work) console_lock console_srcu console_owner &port_lock_key irq_context: 0 (wq_completion)hci5#2 (work_completion)(&hdev->rx_work) console_lock console_srcu console_owner console_owner_lock irq_context: 0 (wq_completion)hci5#2 (work_completion)(&hdev->rx_work) &rq->__lock irq_context: 0 (wq_completion)hci3#2 (work_completion)(&hdev->rx_work) &obj_hash[i].lock irq_context: 0 (wq_completion)hci3#2 (work_completion)(&hdev->rx_work) rcu_read_lock &obj_hash[i].lock irq_context: 0 (wq_completion)hci3#2 (work_completion)(&hdev->rx_work) &hdev->req_wait_q irq_context: 0 (wq_completion)hci3#2 (work_completion)(&hdev->rx_work) &hdev->req_wait_q &p->pi_lock irq_context: 0 (wq_completion)hci3#2 (work_completion)(&hdev->rx_work) &hdev->req_wait_q &p->pi_lock &rq->__lock irq_context: 0 (wq_completion)hci3#2 (work_completion)(&hdev->rx_work) &hdev->req_wait_q &p->pi_lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)hci3#2 (work_completion)(&hdev->rx_work) lock#6 &kcov->lock irq_context: 0 (wq_completion)hci5#2 (work_completion)(&hdev->rx_work) &obj_hash[i].lock irq_context: 0 (wq_completion)hci5#2 (work_completion)(&hdev->rx_work) rcu_read_lock &obj_hash[i].lock irq_context: 0 (wq_completion)hci5#2 (work_completion)(&hdev->rx_work) &hdev->req_wait_q irq_context: 0 (wq_completion)hci5#2 (work_completion)(&hdev->rx_work) &hdev->req_wait_q &p->pi_lock irq_context: 0 (wq_completion)hci5#2 (work_completion)(&hdev->rx_work) &hdev->req_wait_q &p->pi_lock &rq->__lock irq_context: 0 (wq_completion)hci5#2 (work_completion)(&hdev->rx_work) &hdev->req_wait_q &p->pi_lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)hci5#2 (work_completion)(&hdev->rx_work) lock#6 &kcov->lock irq_context: 0 (wq_completion)hci3 (work_completion)(&hdev->power_on) &hdev->req_lock (&timer.timer) irq_context: 0 (wq_completion)hci3 (work_completion)(&hdev->power_on) &hdev->req_lock rcu_read_lock &pool->lock/1 &p->pi_lock irq_context: 0 (wq_completion)hci3 (work_completion)(&hdev->power_on) &hdev->req_lock rcu_read_lock &pool->lock/1 &p->pi_lock &rq->__lock irq_context: 0 (wq_completion)hci3 (work_completion)(&hdev->power_on) &hdev->req_lock rcu_read_lock &pool->lock/1 &p->pi_lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)hci5 (work_completion)(&hdev->power_on) &hdev->req_lock (&timer.timer) irq_context: 0 (wq_completion)hci5 (work_completion)(&hdev->power_on) &hdev->req_lock &c->lock irq_context: 0 (wq_completion)hci5 (work_completion)(&hdev->power_on) &hdev->req_lock &____s->seqcount irq_context: 0 (wq_completion)hci5 (work_completion)(&hdev->power_on) &hdev->req_lock pool_lock#2 irq_context: 0 (wq_completion)hci5 (work_completion)(&hdev->power_on) &hdev->req_lock rcu_read_lock &pool->lock/1 &p->pi_lock irq_context: 0 (wq_completion)hci3#2 (work_completion)(&hdev->cmd_work) &obj_hash[i].lock irq_context: 0 (wq_completion)hci3#2 (work_completion)(&hdev->cmd_work) rcu_read_lock &base->lock irq_context: 0 (wq_completion)hci3#2 (work_completion)(&hdev->cmd_work) rcu_read_lock &base->lock &obj_hash[i].lock irq_context: 0 (wq_completion)hci5#2 (work_completion)(&hdev->cmd_work) &obj_hash[i].lock irq_context: 0 (wq_completion)hci5#2 (work_completion)(&hdev->cmd_work) &rq->__lock irq_context: 0 (wq_completion)hci5#2 (work_completion)(&hdev->cmd_work) rcu_read_lock &base->lock irq_context: 0 (wq_completion)hci5#2 (work_completion)(&hdev->cmd_work) rcu_read_lock &base->lock &obj_hash[i].lock irq_context: 0 (wq_completion)hci5#2 (work_completion)(&hdev->rx_work) console_owner_lock irq_context: 0 (wq_completion)hci5#2 (work_completion)(&hdev->rx_work) console_owner irq_context: 0 (wq_completion)hci2 (work_completion)(&hdev->power_on) &hdev->req_lock rcu_read_lock &pool->lock/1 &p->pi_lock irq_context: 0 (wq_completion)hci0 (work_completion)(&hdev->power_on) &hdev->req_lock tk_core.seq.seqcount irq_context: 0 (wq_completion)hci0 (work_completion)(&hdev->power_on) &hdev->req_lock hci_sk_list.lock irq_context: 0 (wq_completion)hci0 (work_completion)(&hdev->power_on) fs_reclaim irq_context: 0 (wq_completion)hci0 (work_completion)(&hdev->power_on) fs_reclaim mmu_notifier_invalidate_range_start irq_context: 0 (wq_completion)hci0 (work_completion)(&hdev->power_on) pool_lock#2 irq_context: 0 (wq_completion)hci0 (work_completion)(&hdev->power_on) tk_core.seq.seqcount irq_context: 0 (wq_completion)hci0 (work_completion)(&hdev->power_on) hci_sk_list.lock irq_context: 0 (wq_completion)hci0 (work_completion)(&hdev->power_on) &obj_hash[i].lock irq_context: 0 &pool->lock/1 &x->wait#10 irq_context: 0 &pool->lock/1 &x->wait#10 &p->pi_lock irq_context: 0 &pool->lock/1 &x->wait#10 &p->pi_lock &rq->__lock irq_context: 0 &pool->lock/1 &x->wait#10 &p->pi_lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 &hdev->req_lock irq_context: 0 &hdev->req_lock pool_lock#2 irq_context: 0 &hdev->req_lock &list->lock#8 irq_context: 0 &hdev->req_lock &list->lock#7 irq_context: 0 &hdev->req_lock rcu_read_lock &pool->lock/1 irq_context: 0 &hdev->req_lock rcu_read_lock &pool->lock/1 &obj_hash[i].lock irq_context: 0 &hdev->req_lock rcu_read_lock &pool->lock/1 &p->pi_lock irq_context: 0 &hdev->req_lock rcu_read_lock &pool->lock/1 &p->pi_lock &rq->__lock irq_context: 0 &hdev->req_lock rcu_read_lock &pool->lock/1 &p->pi_lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 &hdev->req_lock &hdev->req_wait_q irq_context: 0 &hdev->req_lock &obj_hash[i].lock irq_context: 0 &hdev->req_lock &base->lock irq_context: 0 &hdev->req_lock &base->lock &obj_hash[i].lock irq_context: 0 &hdev->req_lock &rq->__lock irq_context: 0 &hdev->req_lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)hci5#2 (work_completion)(&hdev->rx_work) &base->lock irq_context: 0 (wq_completion)hci5#2 (work_completion)(&hdev->rx_work) &base->lock &obj_hash[i].lock irq_context: 0 cgroup_threadgroup_rwsem &sighand->siglock irq_context: 0 cgroup_threadgroup_rwsem &sighand->siglock &rq->__lock irq_context: 0 &hdev->req_lock (&timer.timer) irq_context: 0 &sb->s_type->i_mutex_key#10 sk_lock-AF_BLUETOOTH-BTPROTO_HCI irq_context: 0 &sb->s_type->i_mutex_key#10 sk_lock-AF_BLUETOOTH-BTPROTO_HCI slock-AF_BLUETOOTH-BTPROTO_HCI irq_context: 0 &sb->s_type->i_mutex_key#10 sk_lock-AF_BLUETOOTH-BTPROTO_HCI pool_lock#2 irq_context: 0 &sb->s_type->i_mutex_key#10 sk_lock-AF_BLUETOOTH-BTPROTO_HCI tk_core.seq.seqcount irq_context: 0 &sb->s_type->i_mutex_key#10 sk_lock-AF_BLUETOOTH-BTPROTO_HCI hci_sk_list.lock irq_context: 0 &sb->s_type->i_mutex_key#10 sk_lock-AF_BLUETOOTH-BTPROTO_HCI &obj_hash[i].lock irq_context: 0 &sb->s_type->i_mutex_key#10 sk_lock-AF_BLUETOOTH-BTPROTO_HCI sock_cookie_ida.xa_lock irq_context: 0 &sb->s_type->i_mutex_key#10 sk_lock-AF_BLUETOOTH-BTPROTO_HCI hci_sk_list.lock irq_context: 0 &sb->s_type->i_mutex_key#10 sk_lock-AF_BLUETOOTH-BTPROTO_HCI clock-AF_BLUETOOTH irq_context: 0 &sb->s_type->i_mutex_key#10 slock-AF_BLUETOOTH-BTPROTO_HCI irq_context: 0 &sb->s_type->i_mutex_key#10 hci_dev_list_lock irq_context: 0 &sb->s_type->i_mutex_key#10 rlock-AF_BLUETOOTH irq_context: 0 &sb->s_type->i_mutex_key#10 wlock-AF_BLUETOOTH irq_context: 0 (wq_completion)hci0#2 (work_completion)(&hdev->rx_work) chan_list_lock irq_context: 0 (wq_completion)hci0#2 (work_completion)(&hdev->rx_work) &hdev->lock fs_reclaim irq_context: 0 (wq_completion)hci0#2 (work_completion)(&hdev->rx_work) &hdev->lock fs_reclaim mmu_notifier_invalidate_range_start irq_context: 0 (wq_completion)hci0#2 (work_completion)(&hdev->rx_work) &hdev->lock pool_lock#2 irq_context: 0 (wq_completion)hci0#2 (work_completion)(&hdev->rx_work) &hdev->lock &obj_hash[i].lock irq_context: 0 (wq_completion)hci0#2 (work_completion)(&hdev->rx_work) &hdev->lock &x->wait#9 irq_context: 0 (wq_completion)hci0#2 (work_completion)(&hdev->rx_work) &hdev->lock &obj_hash[i].lock pool_lock irq_context: 0 (wq_completion)hci0#2 (work_completion)(&hdev->rx_work) rcu_read_lock &pool->lock/1 irq_context: 0 (wq_completion)hci0#2 (work_completion)(&hdev->rx_work) rcu_read_lock &pool->lock/1 &obj_hash[i].lock irq_context: 0 (wq_completion)hci0#2 (work_completion)(&hdev->rx_work) rcu_read_lock &pool->lock/1 &p->pi_lock irq_context: 0 (wq_completion)hci0#2 (work_completion)(&hdev->rx_work) &hdev->lock &k->list_lock irq_context: 0 (wq_completion)hci0#2 (work_completion)(&hdev->rx_work) &hdev->lock lock irq_context: 0 (wq_completion)hci0#2 (work_completion)(&hdev->rx_work) &hdev->lock lock kernfs_idr_lock irq_context: 0 (wq_completion)hci0#2 (work_completion)(&hdev->rx_work) &hdev->lock &root->kernfs_rwsem irq_context: 0 (wq_completion)hci0#2 (work_completion)(&hdev->rx_work) &hdev->lock &root->kernfs_rwsem &root->kernfs_iattr_rwsem irq_context: 0 (wq_completion)hci0#2 (work_completion)(&hdev->rx_work) &hdev->lock bus_type_sem irq_context: 0 (wq_completion)hci0#2 (work_completion)(&hdev->rx_work) &hdev->lock sysfs_symlink_target_lock irq_context: 0 (wq_completion)hci0#2 (work_completion)(&hdev->rx_work) &hdev->lock &____s->seqcount irq_context: 0 (wq_completion)hci0#2 (work_completion)(&hdev->rx_work) &hdev->lock rcu_read_lock pool_lock#2 irq_context: 0 (wq_completion)hci0#2 (work_completion)(&hdev->rx_work) &hdev->lock &root->kernfs_rwsem irq_context: 0 (wq_completion)hci0#2 (work_completion)(&hdev->rx_work) &hdev->lock &c->lock irq_context: 0 (wq_completion)hci0#2 (work_completion)(&hdev->rx_work) &hdev->lock lock kernfs_idr_lock pool_lock#2 irq_context: 0 (wq_completion)hci0#2 (work_completion)(&hdev->rx_work) &hdev->lock &dev->power.lock irq_context: 0 (wq_completion)hci0#2 (work_completion)(&hdev->rx_work) &hdev->lock dpm_list_mtx irq_context: 0 (wq_completion)hci0#2 (work_completion)(&hdev->rx_work) &hdev->lock uevent_sock_mutex irq_context: 0 (wq_completion)hci0#2 (work_completion)(&hdev->rx_work) &hdev->lock uevent_sock_mutex fs_reclaim irq_context: 0 (wq_completion)hci0#2 (work_completion)(&hdev->rx_work) &hdev->lock uevent_sock_mutex fs_reclaim mmu_notifier_invalidate_range_start irq_context: 0 (wq_completion)hci0#2 (work_completion)(&hdev->rx_work) &hdev->lock uevent_sock_mutex pool_lock#2 irq_context: 0 (wq_completion)hci0#2 (work_completion)(&hdev->rx_work) &hdev->lock uevent_sock_mutex nl_table_lock irq_context: 0 (wq_completion)hci0#2 (work_completion)(&hdev->rx_work) &hdev->lock uevent_sock_mutex rlock-AF_NETLINK irq_context: 0 (wq_completion)hci0#2 (work_completion)(&hdev->rx_work) &hdev->lock uevent_sock_mutex rcu_read_lock &ei->socket.wq.wait irq_context: 0 (wq_completion)hci0#2 (work_completion)(&hdev->rx_work) &hdev->lock uevent_sock_mutex rcu_read_lock &ei->socket.wq.wait &ep->lock irq_context: 0 (wq_completion)hci0#2 (work_completion)(&hdev->rx_work) &hdev->lock uevent_sock_mutex nl_table_wait.lock irq_context: 0 (wq_completion)hci0#2 (work_completion)(&hdev->rx_work) &hdev->lock uevent_sock_mutex &obj_hash[i].lock irq_context: 0 (wq_completion)hci0#2 (work_completion)(&hdev->rx_work) &hdev->lock &k->k_lock irq_context: 0 (wq_completion)hci0#2 (work_completion)(&hdev->rx_work) &hdev->lock subsys mutex#81 irq_context: 0 (wq_completion)hci0#2 (work_completion)(&hdev->rx_work) &hdev->lock subsys mutex#81 &k->k_lock irq_context: 0 (wq_completion)hci0#2 (work_completion)(&hdev->rx_work) &hdev->lock &list->lock#7 irq_context: 0 (wq_completion)hci0#2 (work_completion)(&hdev->rx_work) &hdev->lock &hdev->unregister_lock irq_context: 0 (wq_completion)hci0#2 (work_completion)(&hdev->rx_work) &hdev->lock &hdev->unregister_lock fs_reclaim irq_context: 0 (wq_completion)hci0#2 (work_completion)(&hdev->rx_work) &hdev->lock &hdev->unregister_lock fs_reclaim mmu_notifier_invalidate_range_start irq_context: 0 (wq_completion)hci0#2 (work_completion)(&hdev->rx_work) &hdev->lock &hdev->unregister_lock pool_lock#2 irq_context: 0 (wq_completion)hci0#2 (work_completion)(&hdev->rx_work) &hdev->lock &hdev->unregister_lock &hdev->cmd_sync_work_lock irq_context: 0 (wq_completion)hci0#2 (work_completion)(&hdev->rx_work) &hdev->lock &hdev->unregister_lock rcu_read_lock &pool->lock/1 irq_context: 0 (wq_completion)hci0#2 (work_completion)(&hdev->rx_work) &hdev->lock &hdev->unregister_lock rcu_read_lock &pool->lock/1 &obj_hash[i].lock irq_context: 0 (wq_completion)hci0#2 (work_completion)(&hdev->rx_work) &hdev->lock &hdev->unregister_lock rcu_read_lock &pool->lock/1 &p->pi_lock irq_context: 0 (wq_completion)hci0#2 (work_completion)(&hdev->rx_work) &hdev->lock hci_cb_list_lock irq_context: 0 (wq_completion)hci0#2 (work_completion)(&hdev->rx_work) &hdev->lock hci_cb_list_lock fs_reclaim irq_context: 0 (wq_completion)hci0#2 (work_completion)(&hdev->rx_work) &hdev->lock hci_cb_list_lock fs_reclaim mmu_notifier_invalidate_range_start irq_context: 0 (wq_completion)hci0#2 (work_completion)(&hdev->rx_work) &hdev->lock hci_cb_list_lock pool_lock#2 irq_context: 0 (wq_completion)hci0#2 (work_completion)(&hdev->rx_work) &hdev->lock hci_cb_list_lock &obj_hash[i].lock irq_context: 0 (wq_completion)hci0#2 (work_completion)(&hdev->rx_work) &hdev->lock hci_cb_list_lock chan_list_lock irq_context: 0 (wq_completion)hci0#2 (work_completion)(&hdev->rx_work) &hdev->lock hci_cb_list_lock &conn->ident_lock irq_context: 0 (wq_completion)hci0#2 (work_completion)(&hdev->rx_work) &hdev->lock hci_cb_list_lock &base->lock irq_context: 0 (wq_completion)hci0#2 (work_completion)(&hdev->rx_work) &hdev->lock hci_cb_list_lock &base->lock &obj_hash[i].lock irq_context: 0 (wq_completion)hci0#2 (work_completion)(&hdev->rx_work) &hdev->lock hci_cb_list_lock &list->lock#9 irq_context: 0 (wq_completion)hci0#2 (work_completion)(&hdev->rx_work) &hdev->lock hci_cb_list_lock rcu_read_lock &pool->lock/1 irq_context: 0 (wq_completion)hci0#2 (work_completion)(&hdev->rx_work) &hdev->lock hci_cb_list_lock rcu_read_lock &pool->lock/1 &obj_hash[i].lock irq_context: 0 (wq_completion)hci0#2 (work_completion)(&hdev->rx_work) &hdev->lock hci_cb_list_lock rcu_read_lock &pool->lock/1 &p->pi_lock irq_context: 0 (wq_completion)hci0#2 (work_completion)(&hdev->rx_work) &hdev->lock hci_cb_list_lock &conn->chan_lock irq_context: 0 (wq_completion)hci0#2 (work_completion)(&hdev->rx_work) &hdev->lock hci_cb_list_lock rcu_read_lock &pool->lock/1 pool_lock#2 irq_context: 0 (wq_completion)hci0#2 (work_completion)(&hdev->rx_work) &hdev->lock &base->lock irq_context: 0 (wq_completion)hci0#2 (work_completion)(&hdev->rx_work) &hdev->lock &base->lock &obj_hash[i].lock irq_context: 0 (wq_completion)hci0#2 (work_completion)(&hdev->rx_work) &hdev->lock tk_core.seq.seqcount irq_context: 0 (wq_completion)hci0#2 (work_completion)(&hdev->rx_work) &hdev->lock hci_sk_list.lock irq_context: 0 (wq_completion)hci0#2 (work_completion)(&hdev->rx_work) &hdev->lock &n->list_lock irq_context: 0 (wq_completion)hci0#2 (work_completion)(&hdev->rx_work) &hdev->lock &n->list_lock &c->lock irq_context: 0 (wq_completion)hci0#2 (work_completion)(&hdev->rx_work) &hdev->lock hci_cb_list_lock &obj_hash[i].lock pool_lock irq_context: 0 (wq_completion)hci0 (work_completion)(&hdev->cmd_sync_work) irq_context: 0 (wq_completion)hci0 (work_completion)(&hdev->cmd_sync_work) &hdev->cmd_sync_work_lock irq_context: 0 (wq_completion)hci0 (work_completion)(&hdev->cmd_sync_work) &hdev->req_lock irq_context: 0 (wq_completion)hci0 (work_completion)(&hdev->cmd_sync_work) &obj_hash[i].lock irq_context: 0 (wq_completion)hci0 (work_completion)(&hdev->cmd_sync_work) pool_lock#2 irq_context: 0 (wq_completion)hci0#2 (work_completion)(&hdev->tx_work) irq_context: 0 (wq_completion)hci0#2 (work_completion)(&hdev->tx_work) &list->lock#9 irq_context: 0 (wq_completion)hci0#2 (work_completion)(&hdev->tx_work) tk_core.seq.seqcount irq_context: 0 (wq_completion)hci0#2 (work_completion)(&hdev->tx_work) &list->lock#6 irq_context: 0 (wq_completion)hci0#2 (work_completion)(&hdev->tx_work) &data->read_wait irq_context: 0 (wq_completion)hci0#2 (work_completion)(&hdev->tx_work) &list->lock#7 irq_context: 0 (wq_completion)hci0#2 (work_completion)(&conn->pending_rx_work) irq_context: 0 (wq_completion)hci0#2 (work_completion)(&conn->pending_rx_work) &list->lock#10 irq_context: 0 (wq_completion)hci5#2 (work_completion)(&hdev->rx_work) &c->lock irq_context: 0 (wq_completion)hci5#2 (work_completion)(&hdev->rx_work) &____s->seqcount irq_context: 0 (wq_completion)hci2#2 (work_completion)(&hdev->cmd_work) &data->read_wait &p->pi_lock &cfs_rq->removed.lock irq_context: 0 (wq_completion)hci2 (work_completion)(&hdev->power_on) &hdev->req_lock rcu_read_lock &pool->lock/1 &p->pi_lock &rq->__lock irq_context: 0 (wq_completion)hci2 (work_completion)(&hdev->power_on) &hdev->req_lock rcu_read_lock &pool->lock/1 &p->pi_lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)hci5#2 (work_completion)(&hdev->rx_work) &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)hci5 (work_completion)(&hdev->power_on) &hdev->req_lock rcu_read_lock &pool->lock/1 &p->pi_lock &rq->__lock irq_context: 0 (wq_completion)hci5 (work_completion)(&hdev->power_on) &hdev->req_lock rcu_read_lock &pool->lock/1 &p->pi_lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)hci5#2 (work_completion)(&hdev->cmd_work) &c->lock irq_context: 0 (wq_completion)hci5#2 (work_completion)(&hdev->cmd_work) &____s->seqcount irq_context: 0 (wq_completion)hci4 (work_completion)(&hdev->power_on) &hdev->req_lock tk_core.seq.seqcount irq_context: 0 (wq_completion)hci4 (work_completion)(&hdev->power_on) &hdev->req_lock hci_sk_list.lock irq_context: 0 (wq_completion)hci4 (work_completion)(&hdev->power_on) fs_reclaim irq_context: 0 (wq_completion)hci4 (work_completion)(&hdev->power_on) fs_reclaim mmu_notifier_invalidate_range_start irq_context: 0 (wq_completion)hci4 (work_completion)(&hdev->power_on) tk_core.seq.seqcount irq_context: 0 (wq_completion)hci4 (work_completion)(&hdev->power_on) hci_sk_list.lock irq_context: 0 (wq_completion)hci4 (work_completion)(&hdev->power_on) &obj_hash[i].lock irq_context: 0 (wq_completion)hci4#2 (work_completion)(&hdev->rx_work) &hdev->req_wait_q &p->pi_lock &rq->__lock &cfs_rq->removed.lock irq_context: 0 &sb->s_type->i_mutex_key#19 irq_context: 0 namespace_sem mnt_id_ida.xa_lock pool_lock#2 irq_context: 0 rcu_read_lock &undo_list->lock irq_context: 0 (wq_completion)hci4#2 (work_completion)(&hdev->rx_work) chan_list_lock irq_context: 0 (wq_completion)hci4#2 (work_completion)(&hdev->rx_work) &hdev->lock fs_reclaim irq_context: 0 (wq_completion)hci4#2 (work_completion)(&hdev->rx_work) &hdev->lock fs_reclaim mmu_notifier_invalidate_range_start irq_context: 0 (wq_completion)hci4#2 (work_completion)(&hdev->rx_work) &hdev->lock &c->lock irq_context: 0 (wq_completion)hci4#2 (work_completion)(&hdev->rx_work) &hdev->lock &____s->seqcount irq_context: 0 (wq_completion)hci4#2 (work_completion)(&hdev->rx_work) &hdev->lock pool_lock#2 irq_context: 0 (wq_completion)hci4#2 (work_completion)(&hdev->rx_work) &hdev->lock &obj_hash[i].lock irq_context: 0 (wq_completion)hci4#2 (work_completion)(&hdev->rx_work) &hdev->lock &obj_hash[i].lock pool_lock irq_context: 0 (wq_completion)hci4#2 (work_completion)(&hdev->rx_work) &hdev->lock &x->wait#9 irq_context: 0 (wq_completion)hci4#2 (work_completion)(&hdev->rx_work) rcu_read_lock &pool->lock/1 irq_context: 0 (wq_completion)hci4#2 (work_completion)(&hdev->rx_work) rcu_read_lock &pool->lock/1 &obj_hash[i].lock irq_context: 0 (wq_completion)hci4#2 (work_completion)(&hdev->rx_work) rcu_read_lock &pool->lock/1 &p->pi_lock irq_context: 0 (wq_completion)hci4#2 (work_completion)(&hdev->rx_work) rcu_read_lock &pool->lock/1 &p->pi_lock &rq->__lock irq_context: 0 (wq_completion)hci4#2 (work_completion)(&hdev->rx_work) rcu_read_lock &pool->lock/1 &p->pi_lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)hci4#2 (work_completion)(&hdev->rx_work) &hdev->lock &k->list_lock irq_context: 0 (wq_completion)hci4#2 (work_completion)(&hdev->rx_work) &hdev->lock &n->list_lock irq_context: 0 (wq_completion)hci4#2 (work_completion)(&hdev->rx_work) &hdev->lock &n->list_lock &c->lock irq_context: 0 (wq_completion)hci4#2 (work_completion)(&hdev->rx_work) &hdev->lock lock irq_context: 0 (wq_completion)hci4#2 (work_completion)(&hdev->rx_work) &hdev->lock lock kernfs_idr_lock irq_context: 0 (wq_completion)hci4#2 (work_completion)(&hdev->rx_work) &hdev->lock &root->kernfs_rwsem irq_context: 0 (wq_completion)hci4#2 (work_completion)(&hdev->rx_work) &hdev->lock &root->kernfs_rwsem &root->kernfs_iattr_rwsem irq_context: 0 (wq_completion)hci4#2 (work_completion)(&hdev->rx_work) &hdev->lock bus_type_sem irq_context: 0 (wq_completion)hci4#2 (work_completion)(&hdev->rx_work) &hdev->lock sysfs_symlink_target_lock irq_context: 0 (wq_completion)hci4#2 (work_completion)(&hdev->rx_work) &hdev->lock &root->kernfs_rwsem irq_context: 0 (wq_completion)hci4#2 (work_completion)(&hdev->rx_work) &hdev->lock &dev->power.lock irq_context: 0 (wq_completion)hci4#2 (work_completion)(&hdev->rx_work) &hdev->lock dpm_list_mtx irq_context: 0 (wq_completion)hci4#2 (work_completion)(&hdev->rx_work) &hdev->lock uevent_sock_mutex irq_context: 0 (wq_completion)hci4#2 (work_completion)(&hdev->rx_work) &hdev->lock uevent_sock_mutex fs_reclaim irq_context: 0 (wq_completion)hci4#2 (work_completion)(&hdev->rx_work) &hdev->lock uevent_sock_mutex fs_reclaim mmu_notifier_invalidate_range_start irq_context: 0 (wq_completion)hci4#2 (work_completion)(&hdev->rx_work) &hdev->lock uevent_sock_mutex nl_table_lock irq_context: 0 (wq_completion)hci4#2 (work_completion)(&hdev->rx_work) &hdev->lock uevent_sock_mutex rlock-AF_NETLINK irq_context: 0 (wq_completion)hci4#2 (work_completion)(&hdev->rx_work) &hdev->lock uevent_sock_mutex rcu_read_lock &ei->socket.wq.wait irq_context: 0 (wq_completion)hci4#2 (work_completion)(&hdev->rx_work) &hdev->lock uevent_sock_mutex rcu_read_lock &ei->socket.wq.wait &ep->lock irq_context: 0 (wq_completion)hci4#2 (work_completion)(&hdev->rx_work) &hdev->lock uevent_sock_mutex nl_table_wait.lock irq_context: 0 (wq_completion)hci4#2 (work_completion)(&hdev->rx_work) &hdev->lock uevent_sock_mutex &obj_hash[i].lock irq_context: 0 (wq_completion)hci4#2 (work_completion)(&hdev->rx_work) &hdev->lock &k->k_lock irq_context: 0 (wq_completion)hci4#2 (work_completion)(&hdev->rx_work) &hdev->lock subsys mutex#81 irq_context: 0 (wq_completion)hci4#2 (work_completion)(&hdev->rx_work) &hdev->lock subsys mutex#81 &k->k_lock irq_context: 0 (wq_completion)hci4#2 (work_completion)(&hdev->rx_work) &hdev->lock &list->lock#7 irq_context: 0 (wq_completion)hci4#2 (work_completion)(&hdev->rx_work) &hdev->lock &hdev->unregister_lock irq_context: 0 (wq_completion)hci4#2 (work_completion)(&hdev->rx_work) &hdev->lock &hdev->unregister_lock fs_reclaim irq_context: 0 (wq_completion)hci4#2 (work_completion)(&hdev->rx_work) &hdev->lock &hdev->unregister_lock fs_reclaim mmu_notifier_invalidate_range_start irq_context: 0 (wq_completion)hci4#2 (work_completion)(&hdev->rx_work) &hdev->lock &hdev->unregister_lock &hdev->cmd_sync_work_lock irq_context: 0 (wq_completion)hci4#2 (work_completion)(&hdev->rx_work) &hdev->lock &hdev->unregister_lock rcu_read_lock &pool->lock/1 irq_context: 0 (wq_completion)hci4#2 (work_completion)(&hdev->rx_work) &hdev->lock &hdev->unregister_lock rcu_read_lock &pool->lock/1 &obj_hash[i].lock irq_context: 0 (wq_completion)hci4#2 (work_completion)(&hdev->rx_work) &hdev->lock &hdev->unregister_lock rcu_read_lock &pool->lock/1 &p->pi_lock irq_context: 0 (wq_completion)hci4#2 (work_completion)(&hdev->rx_work) &hdev->lock &hdev->unregister_lock rcu_read_lock &pool->lock/1 &p->pi_lock &rq->__lock irq_context: 0 (wq_completion)hci4#2 (work_completion)(&hdev->rx_work) &hdev->lock &hdev->unregister_lock rcu_read_lock &pool->lock/1 &p->pi_lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)hci4#2 (work_completion)(&hdev->rx_work) &hdev->lock hci_cb_list_lock irq_context: 0 (wq_completion)hci4#2 (work_completion)(&hdev->rx_work) &hdev->lock hci_cb_list_lock fs_reclaim irq_context: 0 (wq_completion)hci4#2 (work_completion)(&hdev->rx_work) &hdev->lock hci_cb_list_lock fs_reclaim mmu_notifier_invalidate_range_start irq_context: 0 (wq_completion)hci4#2 (work_completion)(&hdev->rx_work) &hdev->lock hci_cb_list_lock &obj_hash[i].lock irq_context: 0 (wq_completion)hci4#2 (work_completion)(&hdev->rx_work) &hdev->lock hci_cb_list_lock chan_list_lock irq_context: 0 (wq_completion)hci4#2 (work_completion)(&hdev->rx_work) &hdev->lock hci_cb_list_lock &conn->ident_lock irq_context: 0 (wq_completion)hci4#2 (work_completion)(&hdev->rx_work) &hdev->lock hci_cb_list_lock &base->lock irq_context: 0 (wq_completion)hci4#2 (work_completion)(&hdev->rx_work) &hdev->lock hci_cb_list_lock &base->lock &obj_hash[i].lock irq_context: 0 (wq_completion)hci4#2 (work_completion)(&hdev->rx_work) &hdev->lock hci_cb_list_lock &list->lock#9 irq_context: 0 (wq_completion)hci4#2 (work_completion)(&hdev->rx_work) &hdev->lock hci_cb_list_lock rcu_read_lock &pool->lock/1 irq_context: 0 (wq_completion)hci4#2 (work_completion)(&hdev->rx_work) &hdev->lock hci_cb_list_lock rcu_read_lock &pool->lock/1 &obj_hash[i].lock irq_context: 0 (wq_completion)hci4#2 (work_completion)(&hdev->rx_work) &hdev->lock hci_cb_list_lock rcu_read_lock &pool->lock/1 &p->pi_lock irq_context: 0 (wq_completion)hci4#2 (work_completion)(&hdev->rx_work) &hdev->lock hci_cb_list_lock &conn->chan_lock irq_context: 0 (wq_completion)hci4#2 (work_completion)(&hdev->rx_work) &hdev->lock &base->lock irq_context: 0 (wq_completion)hci4#2 (work_completion)(&hdev->rx_work) &hdev->lock &base->lock &obj_hash[i].lock irq_context: 0 (wq_completion)hci4#2 (work_completion)(&hdev->rx_work) &hdev->lock tk_core.seq.seqcount irq_context: 0 (wq_completion)hci4#2 (work_completion)(&hdev->rx_work) &hdev->lock hci_sk_list.lock irq_context: 0 (wq_completion)hci4 (work_completion)(&hdev->cmd_sync_work) irq_context: 0 (wq_completion)hci4 (work_completion)(&hdev->cmd_sync_work) &hdev->cmd_sync_work_lock irq_context: 0 (wq_completion)hci4 (work_completion)(&hdev->cmd_sync_work) &hdev->req_lock irq_context: 0 (wq_completion)hci4 (work_completion)(&hdev->cmd_sync_work) &obj_hash[i].lock irq_context: 0 (wq_completion)hci4#2 (work_completion)(&hdev->tx_work) irq_context: 0 (wq_completion)hci4#2 (work_completion)(&hdev->tx_work) &list->lock#9 irq_context: 0 (wq_completion)hci4#2 (work_completion)(&hdev->tx_work) tk_core.seq.seqcount irq_context: 0 (wq_completion)hci4#2 (work_completion)(&hdev->tx_work) &list->lock#6 irq_context: 0 (wq_completion)hci4#2 (work_completion)(&hdev->tx_work) &data->read_wait irq_context: 0 (wq_completion)hci4#2 (work_completion)(&hdev->tx_work) &list->lock#7 irq_context: 0 (wq_completion)hci4#2 (work_completion)(&conn->pending_rx_work) irq_context: 0 (wq_completion)hci4#2 (work_completion)(&conn->pending_rx_work) &list->lock#10 irq_context: 0 (wq_completion)hci2 (work_completion)(&hdev->power_on) &hdev->req_lock tk_core.seq.seqcount irq_context: 0 (wq_completion)hci2 (work_completion)(&hdev->power_on) &hdev->req_lock hci_sk_list.lock irq_context: 0 (wq_completion)hci2 (work_completion)(&hdev->power_on) fs_reclaim irq_context: 0 (wq_completion)hci2 (work_completion)(&hdev->power_on) fs_reclaim mmu_notifier_invalidate_range_start irq_context: 0 (wq_completion)hci2 (work_completion)(&hdev->power_on) tk_core.seq.seqcount irq_context: 0 (wq_completion)hci2 (work_completion)(&hdev->power_on) hci_sk_list.lock irq_context: 0 (wq_completion)hci2 (work_completion)(&hdev->power_on) &obj_hash[i].lock irq_context: 0 &pool->lock/1 &x->wait#10 &p->pi_lock &cfs_rq->removed.lock irq_context: 0 (wq_completion)hci2#2 (work_completion)(&hdev->rx_work) &hdev->req_wait_q &p->pi_lock &rq->__lock &cfs_rq->removed.lock irq_context: 0 sb_writers#3 &sb->s_type->i_mutex_key#9 rcu_read_lock &dentry->d_lock sysctl_lock irq_context: 0 sb_writers#3 &sb->s_type->i_mutex_key#9 rcu_read_lock sysctl_lock irq_context: 0 (wq_completion)hci2#2 (work_completion)(&hdev->rx_work) chan_list_lock irq_context: 0 (wq_completion)hci2#2 (work_completion)(&hdev->rx_work) &hdev->lock fs_reclaim irq_context: 0 (wq_completion)hci2#2 (work_completion)(&hdev->rx_work) &hdev->lock fs_reclaim mmu_notifier_invalidate_range_start irq_context: 0 (wq_completion)hci2#2 (work_completion)(&hdev->rx_work) &hdev->lock &obj_hash[i].lock irq_context: 0 (wq_completion)hci2#2 (work_completion)(&hdev->rx_work) &hdev->lock &x->wait#9 irq_context: 0 (wq_completion)hci2#2 (work_completion)(&hdev->rx_work) rcu_read_lock &pool->lock/1 irq_context: 0 (wq_completion)hci2#2 (work_completion)(&hdev->rx_work) rcu_read_lock &pool->lock/1 &obj_hash[i].lock irq_context: 0 (wq_completion)hci2#2 (work_completion)(&hdev->rx_work) rcu_read_lock &pool->lock/1 &p->pi_lock irq_context: 0 (wq_completion)hci2#2 (work_completion)(&hdev->rx_work) rcu_read_lock &pool->lock/1 &p->pi_lock &rq->__lock irq_context: 0 (wq_completion)hci2#2 (work_completion)(&hdev->rx_work) rcu_read_lock &pool->lock/1 &p->pi_lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)hci2#2 (work_completion)(&hdev->rx_work) &hdev->lock &k->list_lock irq_context: 0 (wq_completion)hci2#2 (work_completion)(&hdev->rx_work) &hdev->lock lock irq_context: 0 (wq_completion)hci2#2 (work_completion)(&hdev->rx_work) &hdev->lock lock kernfs_idr_lock irq_context: 0 (wq_completion)hci2#2 (work_completion)(&hdev->rx_work) &hdev->lock &root->kernfs_rwsem irq_context: 0 (wq_completion)hci2#2 (work_completion)(&hdev->rx_work) &hdev->lock &root->kernfs_rwsem &root->kernfs_iattr_rwsem irq_context: 0 (wq_completion)hci2#2 (work_completion)(&hdev->rx_work) &hdev->lock bus_type_sem irq_context: 0 (wq_completion)hci2#2 (work_completion)(&hdev->rx_work) &hdev->lock &c->lock irq_context: 0 (wq_completion)hci2#2 (work_completion)(&hdev->rx_work) &hdev->lock sysfs_symlink_target_lock irq_context: 0 (wq_completion)hci2#2 (work_completion)(&hdev->rx_work) &hdev->lock &root->kernfs_rwsem irq_context: 0 (wq_completion)hci2#2 (work_completion)(&hdev->rx_work) &hdev->lock &dev->power.lock irq_context: 0 (wq_completion)hci2#2 (work_completion)(&hdev->rx_work) &hdev->lock dpm_list_mtx irq_context: 0 (wq_completion)hci2#2 (work_completion)(&hdev->rx_work) &hdev->lock uevent_sock_mutex irq_context: 0 (wq_completion)hci2#2 (work_completion)(&hdev->rx_work) &hdev->lock uevent_sock_mutex fs_reclaim irq_context: 0 (wq_completion)hci2#2 (work_completion)(&hdev->rx_work) &hdev->lock uevent_sock_mutex fs_reclaim mmu_notifier_invalidate_range_start irq_context: 0 (wq_completion)hci2#2 (work_completion)(&hdev->rx_work) &hdev->lock uevent_sock_mutex &c->lock irq_context: 0 (wq_completion)hci2#2 (work_completion)(&hdev->rx_work) &hdev->lock uevent_sock_mutex &pcp->lock &zone->lock irq_context: 0 (wq_completion)hci2#2 (work_completion)(&hdev->rx_work) &hdev->lock uevent_sock_mutex &pcp->lock &zone->lock &____s->seqcount irq_context: 0 (wq_completion)hci2#2 (work_completion)(&hdev->rx_work) &hdev->lock uevent_sock_mutex &____s->seqcount irq_context: 0 (wq_completion)hci2#2 (work_completion)(&hdev->rx_work) &hdev->lock uevent_sock_mutex pool_lock#2 irq_context: 0 (wq_completion)hci2#2 (work_completion)(&hdev->rx_work) &hdev->lock uevent_sock_mutex nl_table_lock irq_context: 0 (wq_completion)hci2#2 (work_completion)(&hdev->rx_work) &hdev->lock uevent_sock_mutex rlock-AF_NETLINK irq_context: 0 (wq_completion)hci2#2 (work_completion)(&hdev->rx_work) &hdev->lock uevent_sock_mutex rcu_read_lock &ei->socket.wq.wait irq_context: 0 (wq_completion)hci2#2 (work_completion)(&hdev->rx_work) &hdev->lock uevent_sock_mutex rcu_read_lock &ei->socket.wq.wait &ep->lock irq_context: 0 (wq_completion)hci2#2 (work_completion)(&hdev->rx_work) &hdev->lock uevent_sock_mutex nl_table_wait.lock irq_context: 0 (wq_completion)hci2#2 (work_completion)(&hdev->rx_work) &hdev->lock uevent_sock_mutex &obj_hash[i].lock irq_context: 0 (wq_completion)hci2#2 (work_completion)(&hdev->rx_work) &hdev->lock &k->k_lock irq_context: 0 (wq_completion)hci2#2 (work_completion)(&hdev->rx_work) &hdev->lock subsys mutex#81 irq_context: 0 (wq_completion)hci2#2 (work_completion)(&hdev->rx_work) &hdev->lock subsys mutex#81 &k->k_lock irq_context: 0 (wq_completion)hci2#2 (work_completion)(&hdev->rx_work) &hdev->lock &list->lock#7 irq_context: 0 (wq_completion)hci2#2 (work_completion)(&hdev->rx_work) &hdev->lock &hdev->unregister_lock irq_context: 0 (wq_completion)hci2#2 (work_completion)(&hdev->rx_work) &hdev->lock &hdev->unregister_lock fs_reclaim irq_context: 0 (wq_completion)hci2#2 (work_completion)(&hdev->rx_work) &hdev->lock &hdev->unregister_lock fs_reclaim mmu_notifier_invalidate_range_start irq_context: 0 (wq_completion)hci2#2 (work_completion)(&hdev->rx_work) &hdev->lock &hdev->unregister_lock &c->lock irq_context: 0 (wq_completion)hci2#2 (work_completion)(&hdev->rx_work) &hdev->lock &hdev->unregister_lock &hdev->cmd_sync_work_lock irq_context: 0 (wq_completion)hci2#2 (work_completion)(&hdev->rx_work) &hdev->lock &hdev->unregister_lock rcu_read_lock &pool->lock/1 irq_context: 0 (wq_completion)hci2#2 (work_completion)(&hdev->rx_work) &hdev->lock &hdev->unregister_lock rcu_read_lock &pool->lock/1 &obj_hash[i].lock irq_context: 0 (wq_completion)hci2#2 (work_completion)(&hdev->rx_work) &hdev->lock &hdev->unregister_lock rcu_read_lock &pool->lock/1 &p->pi_lock irq_context: 0 (wq_completion)hci2#2 (work_completion)(&hdev->rx_work) &hdev->lock &hdev->unregister_lock rcu_read_lock &pool->lock/1 &p->pi_lock &rq->__lock irq_context: 0 (wq_completion)hci2#2 (work_completion)(&hdev->rx_work) &hdev->lock &hdev->unregister_lock rcu_read_lock &pool->lock/1 &p->pi_lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 sb_writers#3 &sb->s_type->i_mutex_key#9 &rq->__lock irq_context: 0 (wq_completion)hci2 (work_completion)(&hdev->cmd_sync_work) irq_context: 0 (wq_completion)hci2 (work_completion)(&hdev->cmd_sync_work) &hdev->cmd_sync_work_lock irq_context: 0 (wq_completion)hci2#2 (work_completion)(&hdev->rx_work) &hdev->lock hci_cb_list_lock irq_context: 0 (wq_completion)hci2 (work_completion)(&hdev->cmd_sync_work) &hdev->req_lock irq_context: 0 (wq_completion)hci2#2 (work_completion)(&hdev->rx_work) &hdev->lock hci_cb_list_lock fs_reclaim irq_context: 0 (wq_completion)hci2 (work_completion)(&hdev->cmd_sync_work) &obj_hash[i].lock irq_context: 0 (wq_completion)hci2#2 (work_completion)(&hdev->rx_work) &hdev->lock hci_cb_list_lock fs_reclaim mmu_notifier_invalidate_range_start irq_context: 0 (wq_completion)hci2#2 (work_completion)(&hdev->rx_work) &hdev->lock hci_cb_list_lock &obj_hash[i].lock irq_context: 0 (wq_completion)hci2#2 (work_completion)(&hdev->rx_work) &hdev->lock hci_cb_list_lock &obj_hash[i].lock pool_lock irq_context: 0 (wq_completion)hci2#2 (work_completion)(&hdev->rx_work) &hdev->lock hci_cb_list_lock chan_list_lock irq_context: 0 (wq_completion)hci2#2 (work_completion)(&hdev->rx_work) &hdev->lock hci_cb_list_lock &conn->ident_lock irq_context: 0 (wq_completion)hci2#2 (work_completion)(&hdev->rx_work) &hdev->lock hci_cb_list_lock &base->lock irq_context: 0 (wq_completion)hci2#2 (work_completion)(&hdev->rx_work) &hdev->lock hci_cb_list_lock &base->lock &obj_hash[i].lock irq_context: 0 (wq_completion)hci2#2 (work_completion)(&hdev->rx_work) &hdev->lock hci_cb_list_lock &list->lock#9 irq_context: 0 (wq_completion)hci2#2 (work_completion)(&hdev->rx_work) &hdev->lock hci_cb_list_lock rcu_read_lock &pool->lock/1 irq_context: 0 (wq_completion)hci2#2 (work_completion)(&hdev->rx_work) &hdev->lock hci_cb_list_lock rcu_read_lock &pool->lock/1 &obj_hash[i].lock irq_context: 0 (wq_completion)hci2#2 (work_completion)(&hdev->rx_work) &hdev->lock hci_cb_list_lock rcu_read_lock &pool->lock/1 &p->pi_lock irq_context: 0 (wq_completion)hci2#2 (work_completion)(&hdev->rx_work) &hdev->lock hci_cb_list_lock rcu_read_lock &pool->lock/1 &p->pi_lock &rq->__lock irq_context: 0 (wq_completion)hci2#2 (work_completion)(&hdev->rx_work) &hdev->lock hci_cb_list_lock rcu_read_lock &pool->lock/1 &p->pi_lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)hci2#2 (work_completion)(&hdev->rx_work) &hdev->lock hci_cb_list_lock &conn->chan_lock irq_context: 0 (wq_completion)hci2#2 (work_completion)(&hdev->rx_work) &hdev->lock &base->lock irq_context: 0 (wq_completion)hci2#2 (work_completion)(&hdev->rx_work) &hdev->lock &base->lock &obj_hash[i].lock irq_context: 0 (wq_completion)hci2#2 (work_completion)(&hdev->rx_work) &hdev->lock tk_core.seq.seqcount irq_context: 0 (wq_completion)hci2#2 (work_completion)(&hdev->rx_work) &hdev->lock hci_sk_list.lock irq_context: 0 (wq_completion)hci2#2 (work_completion)(&hdev->rx_work) &hdev->lock hci_cb_list_lock &c->lock irq_context: 0 (wq_completion)hci2#2 (work_completion)(&hdev->cmd_work) &c->lock irq_context: 0 (wq_completion)hci2#2 (work_completion)(&hdev->cmd_work) &____s->seqcount irq_context: 0 (wq_completion)hci2#2 (work_completion)(&hdev->tx_work) irq_context: 0 (wq_completion)hci2#2 (work_completion)(&hdev->tx_work) &list->lock#9 irq_context: 0 (wq_completion)hci2#2 (work_completion)(&hdev->tx_work) tk_core.seq.seqcount irq_context: 0 (wq_completion)hci2#2 (work_completion)(&hdev->tx_work) &list->lock#6 irq_context: 0 (wq_completion)hci2#2 (work_completion)(&hdev->tx_work) &data->read_wait irq_context: 0 (wq_completion)hci2#2 (work_completion)(&hdev->tx_work) &list->lock#7 irq_context: 0 (wq_completion)hci2#2 (work_completion)(&conn->pending_rx_work) irq_context: 0 (wq_completion)hci2#2 (work_completion)(&conn->pending_rx_work) &list->lock#10 irq_context: 0 rtnl_mutex rtnl_mutex.wait_lock irq_context: 0 rtnl_mutex rcu_read_lock &ei->socket.wq.wait &p->pi_lock &cfs_rq->removed.lock irq_context: 0 rtnl_mutex rcu_read_lock &ei->socket.wq.wait &p->pi_lock &rq->__lock irq_context: 0 rtnl_mutex rcu_read_lock &ei->socket.wq.wait &p->pi_lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 rtnl_mutex dev_addr_sem &rq->__lock irq_context: 0 rtnl_mutex dev_addr_sem &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 rtnl_mutex &nr_netdev_addr_lock_key irq_context: 0 rtnl_mutex listen_lock irq_context: 0 (wq_completion)hci5 (work_completion)(&hdev->power_on) &hdev->req_lock tk_core.seq.seqcount irq_context: 0 (wq_completion)hci5 (work_completion)(&hdev->power_on) &hdev->req_lock hci_sk_list.lock irq_context: 0 (wq_completion)hci5 (work_completion)(&hdev->power_on) fs_reclaim irq_context: 0 (wq_completion)hci5 (work_completion)(&hdev->power_on) fs_reclaim mmu_notifier_invalidate_range_start irq_context: 0 (wq_completion)hci5 (work_completion)(&hdev->power_on) tk_core.seq.seqcount irq_context: 0 (wq_completion)hci5 (work_completion)(&hdev->power_on) hci_sk_list.lock irq_context: 0 (wq_completion)hci5 (work_completion)(&hdev->power_on) &obj_hash[i].lock irq_context: 0 rtnl_mutex dev_addr_sem &c->lock irq_context: 0 rtnl_mutex dev_addr_sem &n->list_lock irq_context: 0 rtnl_mutex dev_addr_sem &n->list_lock &c->lock irq_context: 0 rtnl_mutex dev_addr_sem rcu_read_lock rcu_node_0 irq_context: 0 rtnl_mutex dev_addr_sem rcu_read_lock &rq->__lock irq_context: 0 rtnl_mutex dev_addr_sem &obj_hash[i].lock irq_context: 0 pernet_ops_rwsem irq_context: 0 pernet_ops_rwsem stack_depot_init_mutex irq_context: 0 pernet_ops_rwsem crngs.lock irq_context: 0 pernet_ops_rwsem proc_inum_ida.xa_lock irq_context: 0 pernet_ops_rwsem fs_reclaim irq_context: 0 pernet_ops_rwsem fs_reclaim mmu_notifier_invalidate_range_start irq_context: 0 pernet_ops_rwsem pool_lock#2 irq_context: 0 pernet_ops_rwsem proc_subdir_lock irq_context: 0 pernet_ops_rwsem proc_subdir_lock irq_context: 0 pernet_ops_rwsem &c->lock irq_context: 0 pernet_ops_rwsem sysctl_lock irq_context: 0 pernet_ops_rwsem pcpu_alloc_mutex irq_context: 0 pernet_ops_rwsem pcpu_alloc_mutex pcpu_lock irq_context: 0 pernet_ops_rwsem mmu_notifier_invalidate_range_start irq_context: 0 pernet_ops_rwsem &sb->s_type->i_lock_key#8 irq_context: 0 pernet_ops_rwsem &dir->lock irq_context: 0 pernet_ops_rwsem &obj_hash[i].lock irq_context: 0 pernet_ops_rwsem &obj_hash[i].lock pool_lock irq_context: 0 pernet_ops_rwsem k-sk_lock-AF_NETLINK irq_context: 0 pernet_ops_rwsem k-sk_lock-AF_NETLINK k-slock-AF_NETLINK irq_context: 0 pernet_ops_rwsem k-sk_lock-AF_NETLINK rcu_read_lock rhashtable_bucket irq_context: 0 pernet_ops_rwsem k-slock-AF_NETLINK irq_context: 0 pernet_ops_rwsem nl_table_lock irq_context: 0 pernet_ops_rwsem nl_table_lock &obj_hash[i].lock irq_context: 0 pernet_ops_rwsem nl_table_lock pool_lock#2 irq_context: 0 pernet_ops_rwsem nl_table_wait.lock irq_context: 0 pernet_ops_rwsem nl_table_lock &____s->seqcount irq_context: 0 pernet_ops_rwsem nl_table_lock rcu_read_lock pool_lock#2 irq_context: 0 pernet_ops_rwsem uevent_sock_mutex irq_context: 0 pernet_ops_rwsem &____s->seqcount irq_context: 0 pernet_ops_rwsem &n->list_lock irq_context: 0 pernet_ops_rwsem &n->list_lock &c->lock irq_context: 0 pernet_ops_rwsem nl_table_lock irq_context: 0 pernet_ops_rwsem &rq->__lock irq_context: 0 pernet_ops_rwsem &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 pernet_ops_rwsem &net->rules_mod_lock irq_context: 0 pernet_ops_rwsem proc_inum_ida.xa_lock pool_lock#2 irq_context: 0 pernet_ops_rwsem percpu_counters_lock irq_context: 0 pernet_ops_rwsem batched_entropy_u32.lock irq_context: 0 (wq_completion)hci3#2 (work_completion)(&hdev->rx_work) &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)hci3#2 (work_completion)(&hdev->rx_work) &base->lock irq_context: 0 (wq_completion)hci3#2 (work_completion)(&hdev->rx_work) &base->lock &obj_hash[i].lock irq_context: 0 pernet_ops_rwsem cache_list_lock irq_context: 0 pernet_ops_rwsem tk_core.seq.seqcount irq_context: 0 pernet_ops_rwsem &k->list_lock irq_context: 0 pernet_ops_rwsem lock irq_context: 0 pernet_ops_rwsem lock kernfs_idr_lock irq_context: 0 pernet_ops_rwsem &root->kernfs_rwsem irq_context: 0 pernet_ops_rwsem &root->kernfs_rwsem &root->kernfs_iattr_rwsem irq_context: 0 pernet_ops_rwsem uevent_sock_mutex fs_reclaim irq_context: 0 pernet_ops_rwsem uevent_sock_mutex fs_reclaim mmu_notifier_invalidate_range_start irq_context: 0 pernet_ops_rwsem uevent_sock_mutex pool_lock#2 irq_context: 0 pernet_ops_rwsem uevent_sock_mutex nl_table_lock irq_context: 0 pernet_ops_rwsem uevent_sock_mutex rlock-AF_NETLINK irq_context: 0 pernet_ops_rwsem uevent_sock_mutex rcu_read_lock &ei->socket.wq.wait irq_context: 0 pernet_ops_rwsem uevent_sock_mutex rcu_read_lock &ei->socket.wq.wait &ep->lock irq_context: 0 pernet_ops_rwsem uevent_sock_mutex nl_table_wait.lock irq_context: 0 pernet_ops_rwsem uevent_sock_mutex &obj_hash[i].lock irq_context: 0 pernet_ops_rwsem &sn->pipefs_sb_lock irq_context: 0 pernet_ops_rwsem &s->s_inode_list_lock irq_context: 0 pernet_ops_rwsem nf_hook_mutex irq_context: 0 pernet_ops_rwsem nf_hook_mutex fs_reclaim irq_context: 0 pernet_ops_rwsem nf_hook_mutex fs_reclaim mmu_notifier_invalidate_range_start irq_context: 0 pernet_ops_rwsem nf_hook_mutex pool_lock#2 irq_context: 0 pernet_ops_rwsem cpu_hotplug_lock irq_context: 0 pernet_ops_rwsem rcu_read_lock pool_lock#2 irq_context: 0 pernet_ops_rwsem nf_connlabels_lock irq_context: 0 pernet_ops_rwsem nf_ct_ecache_mutex irq_context: 0 pernet_ops_rwsem nf_log_mutex irq_context: 0 pernet_ops_rwsem ipvs->est_mutex irq_context: 0 pernet_ops_rwsem ipvs->est_mutex fs_reclaim irq_context: 0 pernet_ops_rwsem ipvs->est_mutex fs_reclaim mmu_notifier_invalidate_range_start irq_context: 0 pernet_ops_rwsem ipvs->est_mutex pool_lock#2 irq_context: 0 pernet_ops_rwsem ipvs->est_mutex pcpu_alloc_mutex irq_context: 0 pernet_ops_rwsem ipvs->est_mutex pcpu_alloc_mutex pcpu_lock irq_context: 0 pernet_ops_rwsem &base->lock irq_context: 0 pernet_ops_rwsem &base->lock &obj_hash[i].lock irq_context: 0 pernet_ops_rwsem __ip_vs_app_mutex irq_context: 0 pernet_ops_rwsem __ip_vs_app_mutex fs_reclaim irq_context: 0 pernet_ops_rwsem __ip_vs_app_mutex fs_reclaim mmu_notifier_invalidate_range_start irq_context: 0 pernet_ops_rwsem __ip_vs_app_mutex &c->lock irq_context: 0 pernet_ops_rwsem __ip_vs_app_mutex &n->list_lock irq_context: 0 pernet_ops_rwsem __ip_vs_app_mutex &n->list_lock &c->lock irq_context: 0 pernet_ops_rwsem __ip_vs_app_mutex pool_lock#2 irq_context: 0 pernet_ops_rwsem &hashinfo->lock#2 irq_context: 0 pernet_ops_rwsem &net->ipv6.ip6addrlbl_table.lock irq_context: 0 pernet_ops_rwsem batched_entropy_u32.lock crngs.lock irq_context: 0 pernet_ops_rwsem rtnl_mutex irq_context: 0 pernet_ops_rwsem &this->receive_lock irq_context: 0 pernet_ops_rwsem rtnl_mutex &(&net->nexthop.notifier_chain)->rwsem irq_context: 0 pernet_ops_rwsem rtnl_mutex fs_reclaim irq_context: 0 pernet_ops_rwsem rtnl_mutex fs_reclaim mmu_notifier_invalidate_range_start irq_context: 0 pernet_ops_rwsem rtnl_mutex pool_lock#2 irq_context: 0 pernet_ops_rwsem rtnl_mutex pcpu_alloc_mutex irq_context: 0 pernet_ops_rwsem rtnl_mutex pcpu_alloc_mutex pcpu_lock irq_context: 0 pernet_ops_rwsem rtnl_mutex &xa->xa_lock#3 irq_context: 0 pernet_ops_rwsem rtnl_mutex &xa->xa_lock#3 pool_lock#2 irq_context: 0 pernet_ops_rwsem rtnl_mutex net_rwsem irq_context: 0 pernet_ops_rwsem rtnl_mutex net_rwsem &list->lock#2 irq_context: 0 pernet_ops_rwsem rtnl_mutex &tn->lock irq_context: 0 pernet_ops_rwsem rtnl_mutex &x->wait#9 irq_context: 0 pernet_ops_rwsem rtnl_mutex &obj_hash[i].lock irq_context: 0 pernet_ops_rwsem rtnl_mutex &c->lock irq_context: 0 pernet_ops_rwsem rtnl_mutex &k->list_lock irq_context: 0 pernet_ops_rwsem rtnl_mutex gdp_mutex irq_context: 0 pernet_ops_rwsem rtnl_mutex gdp_mutex &k->list_lock irq_context: 0 pernet_ops_rwsem rtnl_mutex lock irq_context: 0 pernet_ops_rwsem rtnl_mutex lock kernfs_idr_lock irq_context: 0 pernet_ops_rwsem rtnl_mutex &root->kernfs_rwsem irq_context: 0 pernet_ops_rwsem rtnl_mutex &root->kernfs_rwsem &root->kernfs_iattr_rwsem irq_context: 0 pernet_ops_rwsem rtnl_mutex bus_type_sem irq_context: 0 pernet_ops_rwsem rtnl_mutex sysfs_symlink_target_lock irq_context: 0 pernet_ops_rwsem rtnl_mutex &root->kernfs_rwsem &sem->wait_lock irq_context: 0 pernet_ops_rwsem rtnl_mutex lock kernfs_idr_lock pool_lock#2 irq_context: 0 pernet_ops_rwsem rtnl_mutex &root->kernfs_rwsem &rq->__lock irq_context: 0 pernet_ops_rwsem rtnl_mutex &n->list_lock irq_context: 0 pernet_ops_rwsem rtnl_mutex &n->list_lock &c->lock irq_context: 0 pernet_ops_rwsem rtnl_mutex &root->kernfs_rwsem irq_context: 0 pernet_ops_rwsem rtnl_mutex &dev->power.lock irq_context: 0 pernet_ops_rwsem rtnl_mutex dpm_list_mtx irq_context: 0 pernet_ops_rwsem rtnl_mutex uevent_sock_mutex irq_context: 0 pernet_ops_rwsem rtnl_mutex uevent_sock_mutex fs_reclaim irq_context: 0 pernet_ops_rwsem rtnl_mutex uevent_sock_mutex fs_reclaim mmu_notifier_invalidate_range_start irq_context: 0 pernet_ops_rwsem rtnl_mutex uevent_sock_mutex pool_lock#2 irq_context: 0 pernet_ops_rwsem rtnl_mutex uevent_sock_mutex nl_table_lock irq_context: 0 pernet_ops_rwsem rtnl_mutex uevent_sock_mutex &obj_hash[i].lock irq_context: 0 pernet_ops_rwsem rtnl_mutex uevent_sock_mutex nl_table_wait.lock irq_context: 0 pernet_ops_rwsem rtnl_mutex subsys mutex#17 irq_context: 0 pernet_ops_rwsem rtnl_mutex subsys mutex#17 &k->k_lock irq_context: 0 pernet_ops_rwsem rtnl_mutex &dir->lock#2 irq_context: 0 pernet_ops_rwsem rtnl_mutex dev_hotplug_mutex irq_context: 0 pernet_ops_rwsem rtnl_mutex dev_hotplug_mutex &dev->power.lock irq_context: 0 pernet_ops_rwsem rtnl_mutex dev_base_lock irq_context: 0 pernet_ops_rwsem rtnl_mutex input_pool.lock irq_context: 0 pernet_ops_rwsem rtnl_mutex batched_entropy_u32.lock irq_context: 0 pernet_ops_rwsem rtnl_mutex &tbl->lock irq_context: 0 pernet_ops_rwsem rtnl_mutex sysctl_lock irq_context: 0 pernet_ops_rwsem rtnl_mutex &____s->seqcount irq_context: 0 pernet_ops_rwsem rtnl_mutex nl_table_lock irq_context: 0 pernet_ops_rwsem rtnl_mutex nl_table_wait.lock irq_context: 0 pernet_ops_rwsem rtnl_mutex &obj_hash[i].lock pool_lock irq_context: 0 pernet_ops_rwsem rtnl_mutex proc_subdir_lock irq_context: 0 pernet_ops_rwsem rtnl_mutex proc_inum_ida.xa_lock irq_context: 0 pernet_ops_rwsem rtnl_mutex proc_subdir_lock irq_context: 0 pernet_ops_rwsem rtnl_mutex &idev->mc_lock irq_context: 0 pernet_ops_rwsem rtnl_mutex &idev->mc_lock fs_reclaim irq_context: 0 pernet_ops_rwsem rtnl_mutex &idev->mc_lock fs_reclaim mmu_notifier_invalidate_range_start irq_context: 0 pernet_ops_rwsem rtnl_mutex &idev->mc_lock &c->lock irq_context: 0 pernet_ops_rwsem rtnl_mutex &idev->mc_lock pool_lock#2 irq_context: 0 pernet_ops_rwsem rtnl_mutex &idev->mc_lock &obj_hash[i].lock irq_context: 0 pernet_ops_rwsem rtnl_mutex &pnettable->lock irq_context: 0 pernet_ops_rwsem rtnl_mutex smc_ib_devices.mutex irq_context: 0 pernet_ops_rwsem rtnl_mutex rcu_read_lock &ndev->lock irq_context: 0 pernet_ops_rwsem rdma_nets.xa_lock irq_context: 0 pernet_ops_rwsem devices_rwsem irq_context: 0 pernet_ops_rwsem hwsim_netgroup_ida.xa_lock irq_context: 0 pernet_ops_rwsem rtnl_mutex stack_depot_init_mutex irq_context: 0 pernet_ops_rwsem rtnl_mutex fill_pool_map-wait-type-override pool_lock#2 irq_context: 0 pernet_ops_rwsem rtnl_mutex fill_pool_map-wait-type-override pool_lock irq_context: 0 pernet_ops_rwsem rtnl_mutex rcu_read_lock pool_lock#2 irq_context: 0 pernet_ops_rwsem rtnl_mutex lock kernfs_idr_lock &c->lock irq_context: 0 pernet_ops_rwsem rtnl_mutex fill_pool_map-wait-type-override &c->lock irq_context: 0 pernet_ops_rwsem rtnl_mutex fill_pool_map-wait-type-override &n->list_lock irq_context: 0 pernet_ops_rwsem rtnl_mutex fill_pool_map-wait-type-override &n->list_lock &c->lock irq_context: 0 pernet_ops_rwsem rtnl_mutex rcu_read_lock &pool->lock/1 irq_context: 0 pernet_ops_rwsem rtnl_mutex rcu_read_lock &pool->lock/1 &obj_hash[i].lock irq_context: 0 pernet_ops_rwsem rtnl_mutex rcu_read_lock &pool->lock/1 pool_lock#2 irq_context: 0 pernet_ops_rwsem rtnl_mutex rcu_read_lock &pool->lock/1 &p->pi_lock irq_context: 0 pernet_ops_rwsem rtnl_mutex failover_lock irq_context: 0 pernet_ops_rwsem rtnl_mutex &idev->mc_lock &obj_hash[i].lock pool_lock irq_context: 0 pernet_ops_rwsem rtnl_mutex &idev->mc_lock fill_pool_map-wait-type-override pool_lock#2 irq_context: 0 pernet_ops_rwsem rtnl_mutex &idev->mc_lock fill_pool_map-wait-type-override pool_lock irq_context: 0 pernet_ops_rwsem rtnl_mutex &idev->mc_lock _xmit_ETHER irq_context: 0 pernet_ops_rwsem rtnl_mutex &idev->mc_lock _xmit_ETHER &c->lock irq_context: 0 pernet_ops_rwsem rtnl_mutex &idev->mc_lock _xmit_ETHER &n->list_lock irq_context: 0 pernet_ops_rwsem rtnl_mutex &idev->mc_lock _xmit_ETHER &n->list_lock &c->lock irq_context: 0 pernet_ops_rwsem rtnl_mutex &idev->mc_lock _xmit_ETHER pool_lock#2 irq_context: 0 (wq_completion)hci1 (work_completion)(&hdev->power_on) &hdev->req_lock &pcp->lock &zone->lock irq_context: 0 (wq_completion)hci1 (work_completion)(&hdev->power_on) &hdev->req_lock &pcp->lock &zone->lock &____s->seqcount irq_context: 0 (wq_completion)hci1 (work_completion)(&hdev->power_on) &hdev->req_lock &____s->seqcount irq_context: 0 (wq_completion)hci1 (work_completion)(&hdev->power_on) &hdev->req_lock rcu_read_lock &pool->lock/1 &p->pi_lock &rq->__lock irq_context: 0 (wq_completion)hci1 (work_completion)(&hdev->power_on) &hdev->req_lock rcu_read_lock &pool->lock/1 &p->pi_lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)hci1 (work_completion)(&hdev->power_on) &hdev->req_lock tk_core.seq.seqcount irq_context: 0 (wq_completion)hci1 (work_completion)(&hdev->power_on) &hdev->req_lock hci_sk_list.lock irq_context: 0 (wq_completion)hci1 (work_completion)(&hdev->power_on) fs_reclaim irq_context: 0 (wq_completion)hci1 (work_completion)(&hdev->power_on) fs_reclaim mmu_notifier_invalidate_range_start irq_context: 0 (wq_completion)hci1 (work_completion)(&hdev->power_on) tk_core.seq.seqcount irq_context: 0 (wq_completion)hci1 (work_completion)(&hdev->power_on) hci_sk_list.lock irq_context: 0 (wq_completion)hci1 (work_completion)(&hdev->power_on) &obj_hash[i].lock irq_context: 0 pernet_ops_rwsem rtnl_mutex &rq->__lock irq_context: 0 pernet_ops_rwsem rtnl_mutex uevent_sock_mutex &c->lock irq_context: 0 pernet_ops_rwsem rtnl_mutex uevent_sock_mutex &____s->seqcount irq_context: 0 (wq_completion)hci1#2 (work_completion)(&hdev->rx_work) &____s->seqcount irq_context: 0 (wq_completion)hci1#2 (work_completion)(&hdev->rx_work) chan_list_lock irq_context: 0 (wq_completion)hci1#2 (work_completion)(&hdev->rx_work) &hdev->lock fs_reclaim irq_context: 0 (wq_completion)hci1#2 (work_completion)(&hdev->rx_work) &hdev->lock fs_reclaim mmu_notifier_invalidate_range_start irq_context: 0 (wq_completion)hci1#2 (work_completion)(&hdev->rx_work) &hdev->lock &obj_hash[i].lock irq_context: 0 (wq_completion)hci1#2 (work_completion)(&hdev->rx_work) &hdev->lock &x->wait#9 irq_context: 0 (wq_completion)hci1#2 (work_completion)(&hdev->rx_work) rcu_read_lock &pool->lock/1 irq_context: 0 (wq_completion)hci1#2 (work_completion)(&hdev->rx_work) rcu_read_lock &pool->lock/1 &obj_hash[i].lock irq_context: 0 (wq_completion)hci1#2 (work_completion)(&hdev->rx_work) rcu_read_lock &pool->lock/1 &p->pi_lock irq_context: 0 (wq_completion)hci1#2 (work_completion)(&hdev->rx_work) &hdev->lock &c->lock irq_context: 0 (wq_completion)hci1#2 (work_completion)(&hdev->rx_work) &hdev->lock &n->list_lock irq_context: 0 (wq_completion)hci1#2 (work_completion)(&hdev->rx_work) &hdev->lock &n->list_lock &c->lock irq_context: 0 (wq_completion)hci1#2 (work_completion)(&hdev->rx_work) &hdev->lock &k->list_lock irq_context: 0 (wq_completion)hci1#2 (work_completion)(&hdev->rx_work) &hdev->lock lock irq_context: 0 (wq_completion)hci1#2 (work_completion)(&hdev->rx_work) &hdev->lock lock kernfs_idr_lock irq_context: 0 (wq_completion)hci1#2 (work_completion)(&hdev->rx_work) &hdev->lock &root->kernfs_rwsem irq_context: 0 (wq_completion)hci1#2 (work_completion)(&hdev->rx_work) &hdev->lock &root->kernfs_rwsem &root->kernfs_iattr_rwsem irq_context: 0 (wq_completion)hci1#2 (work_completion)(&hdev->rx_work) &hdev->lock bus_type_sem irq_context: 0 (wq_completion)hci1#2 (work_completion)(&hdev->rx_work) &hdev->lock sysfs_symlink_target_lock irq_context: 0 namespace_sem pcpu_alloc_mutex pcpu_alloc_mutex.wait_lock irq_context: 0 namespace_sem pcpu_alloc_mutex &rq->__lock irq_context: 0 namespace_sem pcpu_alloc_mutex &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)hci1#2 (work_completion)(&hdev->rx_work) &hdev->lock &root->kernfs_rwsem irq_context: 0 pernet_ops_rwsem rtnl_mutex pcpu_alloc_mutex.wait_lock irq_context: 0 pernet_ops_rwsem rtnl_mutex &p->pi_lock irq_context: 0 pernet_ops_rwsem rtnl_mutex &p->pi_lock &rq->__lock irq_context: 0 (wq_completion)hci1#2 (work_completion)(&hdev->rx_work) &hdev->lock &dev->power.lock irq_context: 0 pernet_ops_rwsem rtnl_mutex &p->pi_lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)hci1#2 (work_completion)(&hdev->rx_work) &hdev->lock dpm_list_mtx irq_context: 0 (wq_completion)hci1#2 (work_completion)(&hdev->rx_work) &hdev->lock &____s->seqcount irq_context: 0 (wq_completion)hci1#2 (work_completion)(&hdev->rx_work) &hdev->lock pool_lock#2 irq_context: 0 (wq_completion)hci1#2 (work_completion)(&hdev->rx_work) &hdev->lock uevent_sock_mutex irq_context: 0 (wq_completion)hci1#2 (work_completion)(&hdev->rx_work) &hdev->lock uevent_sock_mutex fs_reclaim irq_context: 0 (wq_completion)hci1#2 (work_completion)(&hdev->rx_work) &hdev->lock uevent_sock_mutex fs_reclaim mmu_notifier_invalidate_range_start irq_context: 0 (wq_completion)hci1#2 (work_completion)(&hdev->rx_work) &hdev->lock uevent_sock_mutex nl_table_lock irq_context: 0 (wq_completion)hci1#2 (work_completion)(&hdev->rx_work) &hdev->lock uevent_sock_mutex rlock-AF_NETLINK irq_context: 0 (wq_completion)hci1#2 (work_completion)(&hdev->rx_work) &hdev->lock uevent_sock_mutex rcu_read_lock &ei->socket.wq.wait irq_context: 0 (wq_completion)hci1#2 (work_completion)(&hdev->rx_work) &hdev->lock uevent_sock_mutex rcu_read_lock &ei->socket.wq.wait &ep->lock irq_context: 0 (wq_completion)hci1#2 (work_completion)(&hdev->rx_work) &hdev->lock uevent_sock_mutex nl_table_wait.lock irq_context: 0 (wq_completion)hci1#2 (work_completion)(&hdev->rx_work) &hdev->lock uevent_sock_mutex &obj_hash[i].lock irq_context: 0 (wq_completion)hci1#2 (work_completion)(&hdev->rx_work) &hdev->lock &k->k_lock irq_context: 0 (wq_completion)hci1#2 (work_completion)(&hdev->rx_work) &hdev->lock subsys mutex#81 irq_context: 0 (wq_completion)hci1#2 (work_completion)(&hdev->rx_work) &hdev->lock subsys mutex#81 &k->k_lock irq_context: 0 (wq_completion)hci1#2 (work_completion)(&hdev->rx_work) &hdev->lock &list->lock#7 irq_context: 0 (wq_completion)hci1#2 (work_completion)(&hdev->rx_work) &hdev->lock &hdev->unregister_lock irq_context: 0 (wq_completion)hci1#2 (work_completion)(&hdev->rx_work) &hdev->lock &hdev->unregister_lock fs_reclaim irq_context: 0 (wq_completion)hci1#2 (work_completion)(&hdev->rx_work) &hdev->lock &hdev->unregister_lock fs_reclaim mmu_notifier_invalidate_range_start irq_context: 0 (wq_completion)hci1#2 (work_completion)(&hdev->rx_work) &hdev->lock &hdev->unregister_lock &hdev->cmd_sync_work_lock irq_context: 0 (wq_completion)hci1#2 (work_completion)(&hdev->rx_work) &hdev->lock &hdev->unregister_lock rcu_read_lock &pool->lock/1 irq_context: 0 (wq_completion)hci1#2 (work_completion)(&hdev->rx_work) &hdev->lock &hdev->unregister_lock rcu_read_lock &pool->lock/1 &obj_hash[i].lock irq_context: 0 (wq_completion)hci1#2 (work_completion)(&hdev->rx_work) &hdev->lock &hdev->unregister_lock rcu_read_lock &pool->lock/1 &p->pi_lock irq_context: 0 (wq_completion)hci1#2 (work_completion)(&hdev->rx_work) &hdev->lock hci_cb_list_lock irq_context: 0 (wq_completion)hci1#2 (work_completion)(&hdev->rx_work) &hdev->lock hci_cb_list_lock fs_reclaim irq_context: 0 (wq_completion)hci1#2 (work_completion)(&hdev->rx_work) &hdev->lock hci_cb_list_lock fs_reclaim mmu_notifier_invalidate_range_start irq_context: 0 (wq_completion)hci1#2 (work_completion)(&hdev->rx_work) &hdev->lock hci_cb_list_lock &obj_hash[i].lock irq_context: 0 (wq_completion)hci1#2 (work_completion)(&hdev->rx_work) &hdev->lock hci_cb_list_lock &obj_hash[i].lock pool_lock irq_context: 0 (wq_completion)hci1#2 (work_completion)(&hdev->rx_work) &hdev->lock hci_cb_list_lock &rq->__lock irq_context: 0 (wq_completion)hci1#2 (work_completion)(&hdev->rx_work) &hdev->lock hci_cb_list_lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 pernet_ops_rwsem rtnl_mutex.wait_lock irq_context: 0 pernet_ops_rwsem &p->pi_lock irq_context: 0 pernet_ops_rwsem &p->pi_lock &rq->__lock irq_context: 0 pernet_ops_rwsem &p->pi_lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)hci3#2 (work_completion)(&hdev->rx_work) fill_pool_map-wait-type-override pool_lock#2 irq_context: 0 (wq_completion)hci3#2 (work_completion)(&hdev->rx_work) fill_pool_map-wait-type-override &c->lock irq_context: 0 (wq_completion)hci3#2 (work_completion)(&hdev->rx_work) fill_pool_map-wait-type-override &n->list_lock irq_context: 0 (wq_completion)hci3#2 (work_completion)(&hdev->rx_work) fill_pool_map-wait-type-override &n->list_lock &c->lock irq_context: 0 (wq_completion)hci3#2 (work_completion)(&hdev->rx_work) fill_pool_map-wait-type-override pool_lock irq_context: 0 (wq_completion)hci1 (work_completion)(&hdev->cmd_sync_work) irq_context: 0 (wq_completion)hci1 (work_completion)(&hdev->cmd_sync_work) &hdev->cmd_sync_work_lock irq_context: 0 (wq_completion)hci1 (work_completion)(&hdev->cmd_sync_work) &hdev->req_lock irq_context: 0 (wq_completion)hci1 (work_completion)(&hdev->cmd_sync_work) &obj_hash[i].lock irq_context: 0 (wq_completion)hci1#2 (work_completion)(&hdev->rx_work) &hdev->lock hci_cb_list_lock chan_list_lock irq_context: 0 (wq_completion)hci1#2 (work_completion)(&hdev->rx_work) &hdev->lock hci_cb_list_lock &conn->ident_lock irq_context: 0 (wq_completion)hci1#2 (work_completion)(&hdev->rx_work) &hdev->lock hci_cb_list_lock &base->lock irq_context: 0 (wq_completion)hci1#2 (work_completion)(&hdev->rx_work) &hdev->lock hci_cb_list_lock &base->lock &obj_hash[i].lock irq_context: 0 (wq_completion)hci1#2 (work_completion)(&hdev->rx_work) &hdev->lock hci_cb_list_lock &list->lock#9 irq_context: 0 (wq_completion)hci1#2 (work_completion)(&hdev->rx_work) &hdev->lock hci_cb_list_lock rcu_read_lock &pool->lock/1 irq_context: 0 (wq_completion)hci1#2 (work_completion)(&hdev->rx_work) &hdev->lock hci_cb_list_lock rcu_read_lock &pool->lock/1 &obj_hash[i].lock irq_context: 0 (wq_completion)hci1#2 (work_completion)(&hdev->rx_work) &hdev->lock hci_cb_list_lock rcu_read_lock &pool->lock/1 &p->pi_lock irq_context: 0 (wq_completion)hci1#2 (work_completion)(&hdev->rx_work) &hdev->lock hci_cb_list_lock rcu_read_lock &pool->lock/1 &p->pi_lock &rq->__lock irq_context: 0 (wq_completion)hci1#2 (work_completion)(&hdev->rx_work) &hdev->lock hci_cb_list_lock rcu_read_lock &pool->lock/1 &p->pi_lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)hci1#2 (work_completion)(&hdev->rx_work) &hdev->lock hci_cb_list_lock &conn->chan_lock irq_context: 0 (wq_completion)hci1#2 (work_completion)(&hdev->rx_work) &hdev->lock &base->lock irq_context: 0 (wq_completion)hci1#2 (work_completion)(&hdev->rx_work) &hdev->lock &base->lock &obj_hash[i].lock irq_context: 0 (wq_completion)hci1#2 (work_completion)(&hdev->rx_work) &hdev->lock tk_core.seq.seqcount irq_context: 0 (wq_completion)hci1#2 (work_completion)(&hdev->rx_work) &hdev->lock hci_sk_list.lock irq_context: 0 (wq_completion)hci1#2 (work_completion)(&hdev->tx_work) irq_context: 0 (wq_completion)hci1#2 (work_completion)(&hdev->tx_work) &list->lock#9 irq_context: 0 (wq_completion)hci1#2 (work_completion)(&hdev->tx_work) tk_core.seq.seqcount irq_context: 0 (wq_completion)hci1#2 (work_completion)(&hdev->tx_work) &list->lock#6 irq_context: 0 (wq_completion)hci1#2 (work_completion)(&hdev->tx_work) &data->read_wait irq_context: 0 (wq_completion)hci1#2 (work_completion)(&hdev->tx_work) &list->lock#7 irq_context: 0 (wq_completion)hci1#2 (work_completion)(&conn->pending_rx_work) irq_context: 0 pernet_ops_rwsem rtnl_mutex uevent_sock_mutex &pcp->lock &zone->lock irq_context: 0 (wq_completion)hci1#2 (work_completion)(&conn->pending_rx_work) &list->lock#10 irq_context: 0 pernet_ops_rwsem rtnl_mutex uevent_sock_mutex &pcp->lock &zone->lock &____s->seqcount irq_context: 0 pernet_ops_rwsem rtnl_mutex uevent_sock_mutex rcu_read_lock pool_lock#2 irq_context: 0 pernet_ops_rwsem rtnl_mutex batched_entropy_u8.lock irq_context: 0 pernet_ops_rwsem rtnl_mutex kfence_freelist_lock irq_context: 0 &hdev->req_lock rcu_read_lock &pool->lock/1 &p->pi_lock &cfs_rq->removed.lock irq_context: 0 (wq_completion)hci5#2 (work_completion)(&hdev->cmd_work) &data->read_wait &p->pi_lock &cfs_rq->removed.lock irq_context: 0 (wq_completion)hci5#2 (work_completion)(&hdev->rx_work) &hdev->req_wait_q &p->pi_lock &rq->__lock &cfs_rq->removed.lock irq_context: 0 &sb->s_type->i_mutex_key#9 &pcp->lock &zone->lock irq_context: 0 &sb->s_type->i_mutex_key#9 &pcp->lock &zone->lock &____s->seqcount irq_context: 0 &type->s_umount_key#23/1 &n->list_lock irq_context: 0 &type->s_umount_key#23/1 &n->list_lock &c->lock irq_context: 0 (wq_completion)hci3#2 (work_completion)(&hdev->rx_work) &hdev->req_wait_q &p->pi_lock &rq->__lock &cfs_rq->removed.lock irq_context: 0 (wq_completion)hci3 (work_completion)(&hdev->power_on) &hdev->req_lock &c->lock irq_context: 0 (wq_completion)hci3 (work_completion)(&hdev->power_on) &hdev->req_lock &____s->seqcount irq_context: 0 (wq_completion)hci5#2 (work_completion)(&hdev->rx_work) chan_list_lock irq_context: 0 (wq_completion)hci5#2 (work_completion)(&hdev->rx_work) &hdev->lock fs_reclaim irq_context: 0 (wq_completion)hci5#2 (work_completion)(&hdev->rx_work) &hdev->lock fs_reclaim mmu_notifier_invalidate_range_start irq_context: 0 (wq_completion)hci5#2 (work_completion)(&hdev->rx_work) &hdev->lock &obj_hash[i].lock irq_context: 0 (wq_completion)hci5#2 (work_completion)(&hdev->rx_work) &hdev->lock &x->wait#9 irq_context: 0 (wq_completion)hci5#2 (work_completion)(&hdev->rx_work) pool_lock#2 irq_context: 0 (wq_completion)hci5#2 (work_completion)(&hdev->rx_work) rcu_read_lock &pool->lock/1 irq_context: 0 (wq_completion)hci5#2 (work_completion)(&hdev->rx_work) rcu_read_lock &pool->lock/1 &obj_hash[i].lock irq_context: 0 (wq_completion)hci5#2 (work_completion)(&hdev->rx_work) rcu_read_lock &pool->lock/1 &p->pi_lock irq_context: 0 (wq_completion)hci5#2 (work_completion)(&hdev->rx_work) &hdev->lock &k->list_lock irq_context: 0 (wq_completion)hci5#2 (work_completion)(&hdev->rx_work) &hdev->lock &c->lock irq_context: 0 (wq_completion)hci5#2 (work_completion)(&hdev->rx_work) &hdev->lock lock irq_context: 0 (wq_completion)hci5#2 (work_completion)(&hdev->rx_work) &hdev->lock lock kernfs_idr_lock irq_context: 0 (wq_completion)hci5#2 (work_completion)(&hdev->rx_work) &hdev->lock &root->kernfs_rwsem irq_context: 0 (wq_completion)hci5#2 (work_completion)(&hdev->rx_work) &hdev->lock &root->kernfs_rwsem &root->kernfs_iattr_rwsem irq_context: 0 (wq_completion)hci5#2 (work_completion)(&hdev->rx_work) &hdev->lock bus_type_sem irq_context: 0 (wq_completion)hci5#2 (work_completion)(&hdev->rx_work) &hdev->lock sysfs_symlink_target_lock irq_context: 0 (wq_completion)hci5#2 (work_completion)(&hdev->rx_work) &hdev->lock &root->kernfs_rwsem irq_context: 0 (wq_completion)hci5#2 (work_completion)(&hdev->rx_work) &hdev->lock &dev->power.lock irq_context: 0 (wq_completion)hci5#2 (work_completion)(&hdev->rx_work) &hdev->lock dpm_list_mtx irq_context: 0 (wq_completion)hci5#2 (work_completion)(&hdev->rx_work) &hdev->lock &____s->seqcount irq_context: 0 (wq_completion)hci5#2 (work_completion)(&hdev->rx_work) &hdev->lock uevent_sock_mutex irq_context: 0 (wq_completion)hci5#2 (work_completion)(&hdev->rx_work) &hdev->lock uevent_sock_mutex fs_reclaim irq_context: 0 (wq_completion)hci5#2 (work_completion)(&hdev->rx_work) &hdev->lock uevent_sock_mutex fs_reclaim mmu_notifier_invalidate_range_start irq_context: 0 (wq_completion)hci5#2 (work_completion)(&hdev->rx_work) &hdev->lock uevent_sock_mutex nl_table_lock irq_context: 0 (wq_completion)hci5#2 (work_completion)(&hdev->rx_work) &hdev->lock uevent_sock_mutex rlock-AF_NETLINK irq_context: 0 (wq_completion)hci5#2 (work_completion)(&hdev->rx_work) &hdev->lock uevent_sock_mutex rcu_read_lock &ei->socket.wq.wait irq_context: 0 (wq_completion)hci5#2 (work_completion)(&hdev->rx_work) &hdev->lock uevent_sock_mutex rcu_read_lock &ei->socket.wq.wait &ep->lock irq_context: 0 (wq_completion)hci5#2 (work_completion)(&hdev->rx_work) &hdev->lock uevent_sock_mutex nl_table_wait.lock irq_context: 0 (wq_completion)hci5#2 (work_completion)(&hdev->rx_work) &hdev->lock uevent_sock_mutex &obj_hash[i].lock irq_context: 0 (wq_completion)hci5#2 (work_completion)(&hdev->rx_work) &hdev->lock &k->k_lock irq_context: 0 (wq_completion)hci5#2 (work_completion)(&hdev->rx_work) &hdev->lock subsys mutex#81 irq_context: 0 (wq_completion)hci5#2 (work_completion)(&hdev->rx_work) &hdev->lock subsys mutex#81 &k->k_lock irq_context: 0 (wq_completion)hci5#2 (work_completion)(&hdev->rx_work) &hdev->lock &list->lock#7 irq_context: 0 (wq_completion)hci5#2 (work_completion)(&hdev->rx_work) &hdev->lock &hdev->unregister_lock irq_context: 0 (wq_completion)hci5#2 (work_completion)(&hdev->rx_work) &hdev->lock &hdev->unregister_lock fs_reclaim irq_context: 0 (wq_completion)hci5#2 (work_completion)(&hdev->rx_work) &hdev->lock &hdev->unregister_lock fs_reclaim mmu_notifier_invalidate_range_start irq_context: 0 (wq_completion)hci5#2 (work_completion)(&hdev->rx_work) &hdev->lock &hdev->unregister_lock &hdev->cmd_sync_work_lock irq_context: 0 (wq_completion)hci5#2 (work_completion)(&hdev->rx_work) &hdev->lock &hdev->unregister_lock rcu_read_lock &pool->lock/1 irq_context: 0 (wq_completion)hci5#2 (work_completion)(&hdev->rx_work) &hdev->lock &hdev->unregister_lock rcu_read_lock &pool->lock/1 &obj_hash[i].lock irq_context: 0 (wq_completion)hci5#2 (work_completion)(&hdev->rx_work) &hdev->lock &hdev->unregister_lock rcu_read_lock &pool->lock/1 &p->pi_lock irq_context: 0 (wq_completion)hci5#2 (work_completion)(&hdev->rx_work) &hdev->lock hci_cb_list_lock irq_context: 0 (wq_completion)hci5#2 (work_completion)(&hdev->rx_work) &hdev->lock hci_cb_list_lock fs_reclaim irq_context: 0 (wq_completion)hci5#2 (work_completion)(&hdev->rx_work) &hdev->lock hci_cb_list_lock fs_reclaim mmu_notifier_invalidate_range_start irq_context: 0 (wq_completion)hci5#2 (work_completion)(&hdev->rx_work) &hdev->lock hci_cb_list_lock &obj_hash[i].lock irq_context: 0 (wq_completion)hci5#2 (work_completion)(&hdev->rx_work) &hdev->lock hci_cb_list_lock chan_list_lock irq_context: 0 (wq_completion)hci5#2 (work_completion)(&hdev->rx_work) &hdev->lock hci_cb_list_lock &conn->ident_lock irq_context: 0 (wq_completion)hci5#2 (work_completion)(&hdev->rx_work) &hdev->lock hci_cb_list_lock &base->lock irq_context: 0 (wq_completion)hci5#2 (work_completion)(&hdev->rx_work) &hdev->lock hci_cb_list_lock &base->lock &obj_hash[i].lock irq_context: 0 (wq_completion)hci5#2 (work_completion)(&hdev->rx_work) &hdev->lock hci_cb_list_lock &list->lock#9 irq_context: 0 (wq_completion)hci5#2 (work_completion)(&hdev->rx_work) &hdev->lock hci_cb_list_lock rcu_read_lock &pool->lock/1 irq_context: 0 (wq_completion)hci5#2 (work_completion)(&hdev->rx_work) &hdev->lock hci_cb_list_lock rcu_read_lock &pool->lock/1 &obj_hash[i].lock irq_context: 0 (wq_completion)hci5#2 (work_completion)(&hdev->rx_work) &hdev->lock hci_cb_list_lock rcu_read_lock &pool->lock/1 &p->pi_lock irq_context: 0 (wq_completion)hci5#2 (work_completion)(&hdev->rx_work) &hdev->lock hci_cb_list_lock &conn->chan_lock irq_context: 0 (wq_completion)hci5#2 (work_completion)(&hdev->rx_work) &hdev->lock &base->lock irq_context: 0 (wq_completion)hci5#2 (work_completion)(&hdev->rx_work) &hdev->lock &base->lock &obj_hash[i].lock irq_context: 0 (wq_completion)hci5 (work_completion)(&hdev->cmd_sync_work) irq_context: 0 (wq_completion)hci5 (work_completion)(&hdev->cmd_sync_work) &hdev->cmd_sync_work_lock irq_context: 0 (wq_completion)hci5 (work_completion)(&hdev->cmd_sync_work) &hdev->req_lock irq_context: 0 (wq_completion)hci5 (work_completion)(&hdev->cmd_sync_work) &obj_hash[i].lock irq_context: 0 (wq_completion)hci5#2 (work_completion)(&hdev->rx_work) &hdev->lock &rq->__lock irq_context: 0 (wq_completion)hci5#2 (work_completion)(&hdev->rx_work) &hdev->lock &obj_hash[i].lock pool_lock irq_context: 0 (wq_completion)hci3#2 (work_completion)(&hdev->rx_work) &hdev->req_wait_q &p->pi_lock &cfs_rq->removed.lock irq_context: 0 (wq_completion)hci5#2 (work_completion)(&hdev->rx_work) &hdev->lock fill_pool_map-wait-type-override pool_lock#2 irq_context: 0 (wq_completion)hci3 (work_completion)(&hdev->power_on) &hdev->req_lock fill_pool_map-wait-type-override pool_lock#2 irq_context: 0 (wq_completion)hci5#2 (work_completion)(&hdev->rx_work) &hdev->lock fill_pool_map-wait-type-override pool_lock irq_context: 0 (wq_completion)hci3 (work_completion)(&hdev->power_on) &hdev->req_lock fill_pool_map-wait-type-override &c->lock irq_context: 0 (wq_completion)hci5#2 (work_completion)(&hdev->rx_work) &hdev->lock tk_core.seq.seqcount irq_context: 0 (wq_completion)hci3 (work_completion)(&hdev->power_on) &hdev->req_lock fill_pool_map-wait-type-override pool_lock irq_context: 0 (wq_completion)hci5#2 (work_completion)(&hdev->rx_work) &hdev->lock hci_sk_list.lock irq_context: 0 (wq_completion)hci3 (work_completion)(&hdev->power_on) &hdev->req_lock tk_core.seq.seqcount irq_context: 0 (wq_completion)hci5#2 (work_completion)(&hdev->rx_work) &hdev->lock uevent_sock_mutex &c->lock irq_context: 0 (wq_completion)hci3 (work_completion)(&hdev->power_on) &hdev->req_lock hci_sk_list.lock irq_context: 0 (wq_completion)hci5#2 (work_completion)(&hdev->rx_work) &hdev->lock uevent_sock_mutex &____s->seqcount irq_context: 0 (wq_completion)hci3 (work_completion)(&hdev->power_on) fs_reclaim irq_context: 0 (wq_completion)hci5#2 (work_completion)(&hdev->rx_work) &hdev->lock uevent_sock_mutex pool_lock#2 irq_context: 0 (wq_completion)hci3 (work_completion)(&hdev->power_on) fs_reclaim mmu_notifier_invalidate_range_start irq_context: 0 (wq_completion)hci3 (work_completion)(&hdev->power_on) tk_core.seq.seqcount irq_context: 0 (wq_completion)hci5#2 (work_completion)(&hdev->rx_work) &hdev->lock hci_cb_list_lock &c->lock irq_context: 0 (wq_completion)hci3 (work_completion)(&hdev->power_on) hci_sk_list.lock irq_context: 0 (wq_completion)hci3 (work_completion)(&hdev->power_on) &obj_hash[i].lock irq_context: 0 (wq_completion)hci5#2 (work_completion)(&hdev->rx_work) &hdev->lock hci_cb_list_lock rcu_read_lock &pool->lock/1 &p->pi_lock &rq->__lock irq_context: 0 (wq_completion)hci5#2 (work_completion)(&hdev->rx_work) &hdev->lock hci_cb_list_lock rcu_read_lock &pool->lock/1 &p->pi_lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)hci5#2 (work_completion)(&hdev->rx_work) &hdev->lock &hdev->unregister_lock rcu_read_lock &pool->lock/1 &p->pi_lock &rq->__lock irq_context: 0 (wq_completion)hci5#2 (work_completion)(&hdev->rx_work) &hdev->lock &hdev->unregister_lock rcu_read_lock &pool->lock/1 &p->pi_lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)hci5#2 (work_completion)(&hdev->tx_work) irq_context: 0 (wq_completion)hci5#2 (work_completion)(&hdev->tx_work) &list->lock#9 irq_context: 0 (wq_completion)hci5#2 (work_completion)(&hdev->tx_work) tk_core.seq.seqcount irq_context: 0 (wq_completion)hci5#2 (work_completion)(&hdev->tx_work) &list->lock#6 irq_context: 0 (wq_completion)hci5#2 (work_completion)(&hdev->tx_work) &data->read_wait irq_context: 0 (wq_completion)hci5#2 (work_completion)(&hdev->tx_work) &list->lock#7 irq_context: 0 (wq_completion)hci5#2 (work_completion)(&conn->pending_rx_work) irq_context: 0 (wq_completion)hci5#2 (work_completion)(&conn->pending_rx_work) &list->lock#10 irq_context: 0 &hdev->req_lock &obj_hash[i].lock pool_lock irq_context: 0 (wq_completion)hci3#2 (work_completion)(&hdev->rx_work) chan_list_lock irq_context: 0 (wq_completion)hci3#2 (work_completion)(&hdev->rx_work) &hdev->lock fs_reclaim irq_context: 0 (wq_completion)hci3#2 (work_completion)(&hdev->rx_work) &hdev->lock fs_reclaim mmu_notifier_invalidate_range_start irq_context: 0 (wq_completion)hci3#2 (work_completion)(&hdev->rx_work) &hdev->lock &obj_hash[i].lock irq_context: 0 (wq_completion)hci3#2 (work_completion)(&hdev->rx_work) &hdev->lock &obj_hash[i].lock pool_lock irq_context: 0 (wq_completion)hci3#2 (work_completion)(&hdev->rx_work) &hdev->lock fill_pool_map-wait-type-override pool_lock#2 irq_context: 0 (wq_completion)hci3#2 (work_completion)(&hdev->rx_work) &hdev->lock fill_pool_map-wait-type-override &c->lock irq_context: 0 (wq_completion)hci3#2 (work_completion)(&hdev->rx_work) &hdev->lock fill_pool_map-wait-type-override pool_lock irq_context: 0 (wq_completion)hci3#2 (work_completion)(&hdev->rx_work) &hdev->lock &x->wait#9 irq_context: 0 (wq_completion)hci3#2 (work_completion)(&hdev->rx_work) rcu_read_lock &pool->lock/1 irq_context: 0 (wq_completion)hci3#2 (work_completion)(&hdev->rx_work) rcu_read_lock &pool->lock/1 &obj_hash[i].lock irq_context: 0 (wq_completion)hci3#2 (work_completion)(&hdev->rx_work) rcu_read_lock &pool->lock/1 &p->pi_lock irq_context: 0 (wq_completion)hci3#2 (work_completion)(&hdev->rx_work) &hdev->lock &c->lock irq_context: 0 (wq_completion)hci3#2 (work_completion)(&hdev->rx_work) &hdev->lock &k->list_lock irq_context: 0 (wq_completion)hci3#2 (work_completion)(&hdev->rx_work) &hdev->lock lock irq_context: 0 (wq_completion)hci3#2 (work_completion)(&hdev->rx_work) &hdev->lock lock kernfs_idr_lock irq_context: 0 (wq_completion)hci3#2 (work_completion)(&hdev->rx_work) &hdev->lock &root->kernfs_rwsem irq_context: 0 (wq_completion)hci3#2 (work_completion)(&hdev->rx_work) &hdev->lock &root->kernfs_rwsem &root->kernfs_iattr_rwsem irq_context: 0 (wq_completion)hci3#2 (work_completion)(&hdev->rx_work) &hdev->lock bus_type_sem irq_context: 0 (wq_completion)hci3#2 (work_completion)(&hdev->rx_work) &hdev->lock sysfs_symlink_target_lock irq_context: 0 (wq_completion)hci3#2 (work_completion)(&hdev->rx_work) &hdev->lock &root->kernfs_rwsem irq_context: 0 (wq_completion)hci3#2 (work_completion)(&hdev->rx_work) &hdev->lock &dev->power.lock irq_context: 0 (wq_completion)hci3#2 (work_completion)(&hdev->rx_work) &hdev->lock dpm_list_mtx irq_context: 0 (wq_completion)hci3#2 (work_completion)(&hdev->rx_work) &hdev->lock &____s->seqcount irq_context: 0 (wq_completion)hci3#2 (work_completion)(&hdev->rx_work) &hdev->lock uevent_sock_mutex irq_context: 0 (wq_completion)hci3#2 (work_completion)(&hdev->rx_work) &hdev->lock uevent_sock_mutex fs_reclaim irq_context: 0 (wq_completion)hci3#2 (work_completion)(&hdev->rx_work) &hdev->lock uevent_sock_mutex fs_reclaim mmu_notifier_invalidate_range_start irq_context: 0 (wq_completion)hci3#2 (work_completion)(&hdev->rx_work) &hdev->lock uevent_sock_mutex nl_table_lock irq_context: 0 (wq_completion)hci3#2 (work_completion)(&hdev->rx_work) &hdev->lock uevent_sock_mutex rlock-AF_NETLINK irq_context: 0 (wq_completion)hci3#2 (work_completion)(&hdev->rx_work) &hdev->lock uevent_sock_mutex rcu_read_lock &ei->socket.wq.wait irq_context: 0 (wq_completion)hci3#2 (work_completion)(&hdev->rx_work) &hdev->lock uevent_sock_mutex rcu_read_lock &ei->socket.wq.wait &ep->lock irq_context: 0 (wq_completion)hci3#2 (work_completion)(&hdev->rx_work) &hdev->lock uevent_sock_mutex nl_table_wait.lock irq_context: 0 (wq_completion)hci3#2 (work_completion)(&hdev->rx_work) &hdev->lock uevent_sock_mutex &obj_hash[i].lock irq_context: 0 (wq_completion)hci3#2 (work_completion)(&hdev->rx_work) &hdev->lock &k->k_lock irq_context: 0 (wq_completion)hci3#2 (work_completion)(&hdev->rx_work) &hdev->lock subsys mutex#81 irq_context: 0 (wq_completion)hci3#2 (work_completion)(&hdev->rx_work) &hdev->lock subsys mutex#81 &k->k_lock irq_context: 0 (wq_completion)hci3#2 (work_completion)(&hdev->rx_work) &hdev->lock &list->lock#7 irq_context: 0 (wq_completion)hci3#2 (work_completion)(&hdev->rx_work) &hdev->lock &hdev->unregister_lock irq_context: 0 (wq_completion)hci3#2 (work_completion)(&hdev->rx_work) &hdev->lock &hdev->unregister_lock fs_reclaim irq_context: 0 (wq_completion)hci3#2 (work_completion)(&hdev->rx_work) &hdev->lock &hdev->unregister_lock fs_reclaim mmu_notifier_invalidate_range_start irq_context: 0 (wq_completion)hci3#2 (work_completion)(&hdev->rx_work) &hdev->lock &hdev->unregister_lock &hdev->cmd_sync_work_lock irq_context: 0 (wq_completion)hci3#2 (work_completion)(&hdev->rx_work) &hdev->lock &hdev->unregister_lock rcu_read_lock &pool->lock/1 irq_context: 0 (wq_completion)hci3#2 (work_completion)(&hdev->rx_work) &hdev->lock &hdev->unregister_lock rcu_read_lock &pool->lock/1 &obj_hash[i].lock irq_context: 0 (wq_completion)hci3#2 (work_completion)(&hdev->rx_work) &hdev->lock &hdev->unregister_lock rcu_read_lock &pool->lock/1 &p->pi_lock irq_context: 0 (wq_completion)hci3#2 (work_completion)(&hdev->rx_work) &hdev->lock &n->list_lock irq_context: 0 (wq_completion)hci3#2 (work_completion)(&hdev->rx_work) &hdev->lock &n->list_lock &c->lock irq_context: 0 (wq_completion)hci3#2 (work_completion)(&hdev->rx_work) &hdev->lock hci_cb_list_lock irq_context: 0 (wq_completion)hci3#2 (work_completion)(&hdev->rx_work) &hdev->lock hci_cb_list_lock fs_reclaim irq_context: 0 (wq_completion)hci3#2 (work_completion)(&hdev->rx_work) &hdev->lock hci_cb_list_lock fs_reclaim &rq->__lock irq_context: 0 (wq_completion)hci3 (work_completion)(&hdev->cmd_sync_work) irq_context: 0 (wq_completion)hci3 (work_completion)(&hdev->cmd_sync_work) &hdev->cmd_sync_work_lock irq_context: 0 (wq_completion)hci3 (work_completion)(&hdev->cmd_sync_work) &hdev->req_lock irq_context: 0 (wq_completion)hci3 (work_completion)(&hdev->cmd_sync_work) &obj_hash[i].lock irq_context: 0 (wq_completion)hci3#2 (work_completion)(&hdev->rx_work) &hdev->lock hci_cb_list_lock fs_reclaim mmu_notifier_invalidate_range_start irq_context: 0 (wq_completion)hci3#2 (work_completion)(&hdev->rx_work) &hdev->lock hci_cb_list_lock &obj_hash[i].lock irq_context: 0 (wq_completion)hci3#2 (work_completion)(&hdev->rx_work) &hdev->lock hci_cb_list_lock chan_list_lock irq_context: 0 (wq_completion)hci3#2 (work_completion)(&hdev->rx_work) &hdev->lock hci_cb_list_lock &conn->ident_lock irq_context: 0 (wq_completion)hci3#2 (work_completion)(&hdev->rx_work) &hdev->lock hci_cb_list_lock &base->lock irq_context: 0 (wq_completion)hci3#2 (work_completion)(&hdev->rx_work) &hdev->lock hci_cb_list_lock &base->lock &obj_hash[i].lock irq_context: 0 (wq_completion)hci3#2 (work_completion)(&hdev->rx_work) &hdev->lock hci_cb_list_lock &list->lock#9 irq_context: 0 (wq_completion)hci3#2 (work_completion)(&hdev->rx_work) &hdev->lock hci_cb_list_lock rcu_read_lock &pool->lock/1 irq_context: 0 (wq_completion)hci3#2 (work_completion)(&hdev->rx_work) &hdev->lock hci_cb_list_lock rcu_read_lock &pool->lock/1 &obj_hash[i].lock irq_context: 0 (wq_completion)hci3#2 (work_completion)(&hdev->rx_work) &hdev->lock hci_cb_list_lock rcu_read_lock &pool->lock/1 &p->pi_lock irq_context: 0 (wq_completion)hci3#2 (work_completion)(&hdev->rx_work) &hdev->lock hci_cb_list_lock rcu_read_lock &pool->lock/1 &p->pi_lock &rq->__lock irq_context: 0 (wq_completion)hci3#2 (work_completion)(&hdev->rx_work) &hdev->lock hci_cb_list_lock rcu_read_lock &pool->lock/1 &p->pi_lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)hci3#2 (work_completion)(&hdev->rx_work) &hdev->lock hci_cb_list_lock &conn->chan_lock irq_context: 0 (wq_completion)hci3#2 (work_completion)(&hdev->rx_work) &hdev->lock &base->lock irq_context: 0 (wq_completion)hci3#2 (work_completion)(&hdev->rx_work) &hdev->lock &base->lock &obj_hash[i].lock irq_context: 0 (wq_completion)hci3#2 (work_completion)(&hdev->rx_work) &hdev->lock tk_core.seq.seqcount irq_context: 0 (wq_completion)hci3#2 (work_completion)(&hdev->rx_work) &hdev->lock hci_sk_list.lock irq_context: 0 (wq_completion)hci3#2 (work_completion)(&hdev->rx_work) &hdev->lock lock kernfs_idr_lock pool_lock#2 irq_context: 0 (wq_completion)hci3#2 (work_completion)(&hdev->rx_work) &hdev->lock uevent_sock_mutex &c->lock irq_context: 0 pernet_ops_rwsem rtnl_mutex batched_entropy_u32.lock crngs.lock irq_context: 0 (wq_completion)hci3#2 (work_completion)(&hdev->tx_work) irq_context: 0 (wq_completion)hci3#2 (work_completion)(&hdev->tx_work) &list->lock#9 irq_context: 0 (wq_completion)hci3#2 (work_completion)(&hdev->tx_work) tk_core.seq.seqcount irq_context: 0 (wq_completion)hci3#2 (work_completion)(&hdev->tx_work) &list->lock#6 irq_context: 0 (wq_completion)hci3#2 (work_completion)(&hdev->tx_work) &data->read_wait irq_context: 0 (wq_completion)hci3#2 (work_completion)(&hdev->tx_work) &list->lock#7 irq_context: 0 (wq_completion)hci3#2 (work_completion)(&conn->pending_rx_work) irq_context: 0 (wq_completion)hci3#2 (work_completion)(&conn->pending_rx_work) &list->lock#10 irq_context: 0 pernet_ops_rwsem rtnl_mutex &idev->mc_lock &____s->seqcount irq_context: 0 pernet_ops_rwsem rtnl_mutex &idev->mc_lock rcu_read_lock rcu_node_0 irq_context: 0 pernet_ops_rwsem rtnl_mutex &idev->mc_lock rcu_read_lock &rq->__lock irq_context: 0 sb_writers#3 &c->lock irq_context: 0 pernet_ops_rwsem rtnl_mutex &idev->mc_lock rcu_read_lock pool_lock#2 irq_context: 0 pernet_ops_rwsem rtnl_mutex rtnl_mutex.wait_lock irq_context: 0 pernet_ops_rwsem rtnl_mutex &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 rtnl_mutex (inetaddr_chain).rwsem rcu_read_lock &ei->socket.wq.wait irq_context: 0 rtnl_mutex (inetaddr_chain).rwsem rcu_read_lock &ei->socket.wq.wait &p->pi_lock irq_context: 0 pernet_ops_rwsem rcu_read_lock &pool->lock irq_context: 0 pernet_ops_rwsem rcu_read_lock &pool->lock &obj_hash[i].lock irq_context: 0 pernet_ops_rwsem rcu_read_lock &pool->lock &p->pi_lock irq_context: 0 pernet_ops_rwsem fill_pool_map-wait-type-override pool_lock#2 irq_context: 0 pernet_ops_rwsem fill_pool_map-wait-type-override pool_lock irq_context: 0 pernet_ops_rwsem fill_pool_map-wait-type-override &c->lock irq_context: 0 pernet_ops_rwsem fill_pool_map-wait-type-override &n->list_lock irq_context: 0 pernet_ops_rwsem fill_pool_map-wait-type-override &n->list_lock &c->lock irq_context: 0 rtnl_mutex fill_pool_map-wait-type-override &c->lock irq_context: 0 pernet_ops_rwsem k-sk_lock-AF_NETLINK rcu_read_lock rcu_read_lock &pool->lock irq_context: 0 pernet_ops_rwsem k-sk_lock-AF_NETLINK rcu_read_lock rcu_read_lock &pool->lock &obj_hash[i].lock irq_context: 0 pernet_ops_rwsem k-sk_lock-AF_NETLINK rcu_read_lock rcu_read_lock &pool->lock &p->pi_lock irq_context: 0 pernet_ops_rwsem k-sk_lock-AF_NETLINK rcu_read_lock rcu_read_lock &pool->lock &p->pi_lock &rq->__lock irq_context: 0 pernet_ops_rwsem k-sk_lock-AF_NETLINK rcu_read_lock rcu_read_lock &pool->lock &p->pi_lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 pernet_ops_rwsem uevent_sock_mutex &c->lock irq_context: 0 pernet_ops_rwsem ipvs->est_mutex &c->lock irq_context: 0 pernet_ops_rwsem rcu_read_lock &pool->lock &p->pi_lock &rq->__lock irq_context: 0 pernet_ops_rwsem rcu_read_lock &pool->lock &p->pi_lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 sk_lock-AF_NETLINK rcu_read_lock &rcu_state.gp_wq irq_context: 0 sk_lock-AF_NETLINK rcu_read_lock &rcu_state.gp_wq &p->pi_lock irq_context: 0 sk_lock-AF_NETLINK rcu_read_lock &rcu_state.gp_wq &p->pi_lock &rq->__lock irq_context: 0 sk_lock-AF_NETLINK rcu_read_lock &rcu_state.gp_wq &p->pi_lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 rtnl_mutex dev_addr_sem &____s->seqcount irq_context: 0 pernet_ops_rwsem batched_entropy_u8.lock irq_context: 0 pernet_ops_rwsem kfence_freelist_lock irq_context: 0 pernet_ops_rwsem rtnl_mutex &xa->xa_lock#3 &c->lock irq_context: 0 pernet_ops_rwsem rtnl_mutex &xa->xa_lock#3 &n->list_lock irq_context: 0 pernet_ops_rwsem rtnl_mutex &xa->xa_lock#3 &n->list_lock &c->lock irq_context: 0 pernet_ops_rwsem rtnl_mutex uevent_sock_mutex &n->list_lock irq_context: 0 pernet_ops_rwsem rtnl_mutex uevent_sock_mutex &n->list_lock &c->lock irq_context: 0 pernet_ops_rwsem rtnl_mutex &idev->mc_lock &n->list_lock irq_context: 0 pernet_ops_rwsem rtnl_mutex &idev->mc_lock &n->list_lock &c->lock irq_context: 0 pernet_ops_rwsem nl_table_lock nl_table_wait.lock irq_context: 0 pernet_ops_rwsem rdma_nets.xa_lock pool_lock#2 irq_context: 0 pernet_ops_rwsem nf_hook_mutex &c->lock irq_context: 0 pernet_ops_rwsem &p->pi_lock &cfs_rq->removed.lock irq_context: 0 pernet_ops_rwsem rtnl_mutex rcu_read_lock &pool->lock/1 &p->pi_lock &rq->__lock irq_context: 0 pernet_ops_rwsem rtnl_mutex rcu_read_lock &pool->lock/1 &p->pi_lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 pernet_ops_rwsem rtnl_mutex remove_cache_srcu irq_context: 0 pernet_ops_rwsem rtnl_mutex remove_cache_srcu quarantine_lock irq_context: 0 pernet_ops_rwsem rtnl_mutex net_rwsem &rq->__lock irq_context: 0 pernet_ops_rwsem rtnl_mutex net_rwsem &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 pernet_ops_rwsem rtnl_mutex quarantine_lock irq_context: 0 pernet_ops_rwsem remove_cache_srcu irq_context: 0 pernet_ops_rwsem remove_cache_srcu quarantine_lock irq_context: 0 pernet_ops_rwsem remove_cache_srcu &n->list_lock irq_context: 0 pernet_ops_rwsem remove_cache_srcu rcu_read_lock pool_lock#2 irq_context: 0 pernet_ops_rwsem remove_cache_srcu &obj_hash[i].lock irq_context: 0 pernet_ops_rwsem remove_cache_srcu &c->lock irq_context: 0 pernet_ops_rwsem rtnl_mutex &meta->lock irq_context: 0 pernet_ops_rwsem k-sk_lock-AF_INET6 irq_context: 0 pernet_ops_rwsem k-sk_lock-AF_INET6 k-slock-AF_INET6 irq_context: 0 pernet_ops_rwsem k-slock-AF_INET6 irq_context: 0 pernet_ops_rwsem k-clock-AF_INET6 irq_context: 0 pernet_ops_rwsem k-sk_lock-AF_INET6 &tcp_hashinfo.bhash[i].lock irq_context: 0 pernet_ops_rwsem k-sk_lock-AF_INET6 &tcp_hashinfo.bhash[i].lock pool_lock#2 irq_context: 0 pernet_ops_rwsem k-sk_lock-AF_INET6 &tcp_hashinfo.bhash[i].lock &tcp_hashinfo.bhash2[i].lock irq_context: 0 pernet_ops_rwsem k-sk_lock-AF_INET6 &tcp_hashinfo.bhash[i].lock &tcp_hashinfo.bhash2[i].lock pool_lock#2 irq_context: 0 pernet_ops_rwsem k-sk_lock-AF_INET6 &tcp_hashinfo.bhash[i].lock &tcp_hashinfo.bhash2[i].lock k-clock-AF_INET6 irq_context: 0 pernet_ops_rwsem k-sk_lock-AF_INET6 &tcp_hashinfo.bhash[i].lock k-clock-AF_INET6 irq_context: 0 pernet_ops_rwsem k-sk_lock-AF_INET6 &h->lhash2[i].lock irq_context: 0 pernet_ops_rwsem cpu_hotplug_lock wq_pool_mutex irq_context: 0 pernet_ops_rwsem cpu_hotplug_lock wq_pool_mutex fs_reclaim irq_context: 0 pernet_ops_rwsem cpu_hotplug_lock wq_pool_mutex fs_reclaim mmu_notifier_invalidate_range_start irq_context: 0 pernet_ops_rwsem cpu_hotplug_lock wq_pool_mutex pool_lock#2 irq_context: 0 pernet_ops_rwsem cpu_hotplug_lock wq_pool_mutex &c->lock irq_context: 0 pernet_ops_rwsem cpu_hotplug_lock wq_pool_mutex &____s->seqcount irq_context: 0 pernet_ops_rwsem cpu_hotplug_lock wq_pool_mutex &obj_hash[i].lock irq_context: 0 pernet_ops_rwsem cpu_hotplug_lock wq_pool_mutex rcu_read_lock pool_lock#2 irq_context: 0 pernet_ops_rwsem cpu_hotplug_lock wq_pool_mutex &wq->mutex irq_context: 0 pernet_ops_rwsem cpu_hotplug_lock wq_pool_mutex &wq->mutex &pool->lock/1 irq_context: 0 pernet_ops_rwsem wq_pool_mutex irq_context: 0 pernet_ops_rwsem wq_pool_mutex &wq->mutex irq_context: 0 pernet_ops_rwsem pcpu_lock irq_context: 0 pernet_ops_rwsem &list->lock#4 irq_context: 0 pernet_ops_rwsem &dir->lock#2 irq_context: 0 pernet_ops_rwsem ptype_lock irq_context: 0 pernet_ops_rwsem rcu_read_lock rhashtable_bucket irq_context: 0 pernet_ops_rwsem k-clock-AF_TIPC irq_context: 0 pernet_ops_rwsem k-sk_lock-AF_TIPC irq_context: 0 pernet_ops_rwsem k-sk_lock-AF_TIPC k-slock-AF_TIPC irq_context: 0 pernet_ops_rwsem k-slock-AF_TIPC irq_context: 0 pernet_ops_rwsem k-sk_lock-AF_TIPC &tn->nametbl_lock irq_context: 0 pernet_ops_rwsem k-sk_lock-AF_TIPC &tn->nametbl_lock pool_lock#2 irq_context: 0 pernet_ops_rwsem k-sk_lock-AF_TIPC &tn->nametbl_lock &service->lock irq_context: 0 pernet_ops_rwsem k-sk_lock-AF_TIPC &tn->nametbl_lock &service->lock pool_lock#2 irq_context: 0 pernet_ops_rwsem nf_ct_proto_mutex irq_context: 0 pernet_ops_rwsem nf_ct_proto_mutex defrag4_mutex irq_context: 0 pernet_ops_rwsem nf_ct_proto_mutex defrag4_mutex nf_hook_mutex irq_context: 0 pernet_ops_rwsem nf_ct_proto_mutex defrag4_mutex nf_hook_mutex fs_reclaim irq_context: 0 pernet_ops_rwsem nf_ct_proto_mutex defrag4_mutex nf_hook_mutex fs_reclaim mmu_notifier_invalidate_range_start irq_context: 0 pernet_ops_rwsem nf_ct_proto_mutex defrag4_mutex nf_hook_mutex pool_lock#2 irq_context: 0 pernet_ops_rwsem nf_ct_proto_mutex defrag4_mutex cpu_hotplug_lock irq_context: 0 pernet_ops_rwsem nf_ct_proto_mutex nf_hook_mutex irq_context: 0 pernet_ops_rwsem nf_ct_proto_mutex nf_hook_mutex fs_reclaim irq_context: 0 pernet_ops_rwsem nf_ct_proto_mutex nf_hook_mutex fs_reclaim mmu_notifier_invalidate_range_start irq_context: 0 pernet_ops_rwsem nf_ct_proto_mutex nf_hook_mutex pool_lock#2 irq_context: 0 pernet_ops_rwsem nf_ct_proto_mutex cpu_hotplug_lock irq_context: 0 pernet_ops_rwsem nf_ct_proto_mutex &obj_hash[i].lock irq_context: 0 pernet_ops_rwsem nf_ct_proto_mutex pool_lock#2 irq_context: 0 pernet_ops_rwsem nf_ct_proto_mutex defrag6_mutex irq_context: 0 pernet_ops_rwsem nf_ct_proto_mutex defrag6_mutex nf_hook_mutex irq_context: 0 pernet_ops_rwsem nf_ct_proto_mutex defrag6_mutex nf_hook_mutex fs_reclaim irq_context: 0 pernet_ops_rwsem nf_ct_proto_mutex defrag6_mutex nf_hook_mutex fs_reclaim mmu_notifier_invalidate_range_start irq_context: 0 pernet_ops_rwsem nf_ct_proto_mutex defrag6_mutex nf_hook_mutex pool_lock#2 irq_context: 0 pernet_ops_rwsem nf_ct_proto_mutex defrag6_mutex cpu_hotplug_lock irq_context: 0 pernet_ops_rwsem k-sk_lock-AF_RXRPC irq_context: 0 pernet_ops_rwsem k-sk_lock-AF_RXRPC k-slock-AF_RXRPC irq_context: 0 pernet_ops_rwsem k-slock-AF_RXRPC irq_context: 0 pernet_ops_rwsem k-sk_lock-AF_RXRPC &rxnet->local_mutex irq_context: 0 pernet_ops_rwsem k-sk_lock-AF_RXRPC &rxnet->local_mutex fs_reclaim irq_context: 0 pernet_ops_rwsem k-sk_lock-AF_RXRPC &rxnet->local_mutex fs_reclaim mmu_notifier_invalidate_range_start irq_context: 0 pernet_ops_rwsem k-sk_lock-AF_RXRPC &rxnet->local_mutex pool_lock#2 irq_context: 0 pernet_ops_rwsem k-sk_lock-AF_RXRPC &rxnet->local_mutex &obj_hash[i].lock irq_context: 0 pernet_ops_rwsem k-sk_lock-AF_RXRPC &rxnet->local_mutex crngs.lock irq_context: 0 pernet_ops_rwsem k-sk_lock-AF_RXRPC &rxnet->local_mutex mmu_notifier_invalidate_range_start irq_context: 0 pernet_ops_rwsem k-sk_lock-AF_RXRPC &rxnet->local_mutex &sb->s_type->i_lock_key#8 irq_context: 0 pernet_ops_rwsem k-sk_lock-AF_RXRPC &rxnet->local_mutex &dir->lock irq_context: 0 pernet_ops_rwsem k-sk_lock-AF_RXRPC &rxnet->local_mutex k-sk_lock-AF_INET6 irq_context: 0 pernet_ops_rwsem k-sk_lock-AF_RXRPC &rxnet->local_mutex k-sk_lock-AF_INET6 k-slock-AF_INET6 irq_context: 0 pernet_ops_rwsem k-sk_lock-AF_RXRPC &rxnet->local_mutex k-sk_lock-AF_INET6 &table->hash[i].lock irq_context: 0 pernet_ops_rwsem k-sk_lock-AF_RXRPC &rxnet->local_mutex k-sk_lock-AF_INET6 &table->hash[i].lock k-clock-AF_INET6 irq_context: 0 pernet_ops_rwsem k-sk_lock-AF_RXRPC &rxnet->local_mutex k-sk_lock-AF_INET6 &table->hash[i].lock &table->hash2[i].lock irq_context: 0 pernet_ops_rwsem k-sk_lock-AF_RXRPC &rxnet->local_mutex k-slock-AF_INET6 irq_context: 0 pernet_ops_rwsem k-sk_lock-AF_RXRPC &rxnet->local_mutex cpu_hotplug_lock irq_context: 0 pernet_ops_rwsem k-sk_lock-AF_RXRPC &rxnet->local_mutex kthread_create_lock irq_context: 0 pernet_ops_rwsem k-sk_lock-AF_RXRPC &rxnet->local_mutex &p->pi_lock irq_context: 0 pernet_ops_rwsem k-sk_lock-AF_RXRPC &rxnet->local_mutex &p->pi_lock &rq->__lock irq_context: 0 pernet_ops_rwsem k-sk_lock-AF_RXRPC &rxnet->local_mutex &p->pi_lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 pernet_ops_rwsem k-sk_lock-AF_RXRPC &rxnet->local_mutex &x->wait irq_context: 0 pernet_ops_rwsem k-sk_lock-AF_RXRPC &rxnet->local_mutex &rq->__lock irq_context: 0 pernet_ops_rwsem k-sk_lock-AF_RXRPC &rxnet->local_mutex &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 pernet_ops_rwsem rtnl_mutex &pcp->lock &zone->lock irq_context: 0 pernet_ops_rwsem rtnl_mutex &pcp->lock &zone->lock &____s->seqcount irq_context: 0 pernet_ops_rwsem k-sk_lock-AF_RXRPC &rxnet->local_mutex &x->wait#22 irq_context: 0 &x->wait#22 &p->pi_lock irq_context: 0 &x->wait#22 &p->pi_lock &rq->__lock irq_context: 0 &x->wait#22 &p->pi_lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 pernet_ops_rwsem k-sk_lock-AF_RXRPC &local->services_lock irq_context: 0 pernet_ops_rwsem k-sk_lock-AF_RXRPC fs_reclaim irq_context: 0 pernet_ops_rwsem k-sk_lock-AF_RXRPC fs_reclaim mmu_notifier_invalidate_range_start irq_context: 0 pernet_ops_rwsem k-sk_lock-AF_RXRPC &c->lock irq_context: 0 pernet_ops_rwsem k-sk_lock-AF_RXRPC pool_lock#2 irq_context: 0 pernet_ops_rwsem &rxnet->conn_lock irq_context: 0 pernet_ops_rwsem &call->waitq irq_context: 0 pernet_ops_rwsem &rx->call_lock irq_context: 0 pernet_ops_rwsem &rxnet->call_lock irq_context: 0 pernet_ops_rwsem fill_pool_map-wait-type-override &____s->seqcount irq_context: 0 pernet_ops_rwsem net_rwsem irq_context: 0 &sb->s_type->i_mutex_key#9 rcu_read_lock sysctl_lock irq_context: 0 sb_writers#3 &____s->seqcount#11 irq_context: 0 sb_writers#3 &(&net->ipv4.ping_group_range.lock)->lock irq_context: 0 sb_writers#3 &(&net->ipv4.ping_group_range.lock)->lock &____s->seqcount#11 irq_context: 0 misc_mtx &dir->lock irq_context: 0 pernet_ops_rwsem rtnl_mutex fill_pool_map-wait-type-override &____s->seqcount irq_context: 0 rtnl_mutex &r->consumer_lock irq_context: 0 rtnl_mutex &r->consumer_lock &r->producer_lock irq_context: 0 pernet_ops_rwsem rtnl_mutex rcu_read_lock &rq->__lock irq_context: 0 pernet_ops_rwsem rtnl_mutex rcu_read_lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 rtnl_mutex failover_lock irq_context: 0 rtnl_mutex &idev->mc_lock _xmit_ETHER pool_lock#2 irq_context: 0 rtnl_mutex &mm->mmap_lock irq_context: 0 pernet_ops_rwsem rtnl_mutex uevent_sock_mutex &rq->__lock irq_context: 0 pernet_ops_rwsem &pcp->lock &zone->lock irq_context: 0 pernet_ops_rwsem &pcp->lock &zone->lock &____s->seqcount irq_context: 0 pernet_ops_rwsem rtnl_mutex lock kernfs_idr_lock &n->list_lock irq_context: 0 pernet_ops_rwsem rtnl_mutex lock kernfs_idr_lock &n->list_lock &c->lock irq_context: 0 pernet_ops_rwsem rtnl_mutex rcu_read_lock &pool->lock/1 &p->pi_lock &rq->__lock &cfs_rq->removed.lock irq_context: 0 pernet_ops_rwsem rtnl_mutex rcu_read_lock &pool->lock irq_context: 0 pernet_ops_rwsem rtnl_mutex rcu_read_lock &pool->lock &obj_hash[i].lock irq_context: 0 pernet_ops_rwsem rtnl_mutex rcu_read_lock &pool->lock &p->pi_lock irq_context: 0 pernet_ops_rwsem rtnl_mutex rcu_read_lock &pool->lock &p->pi_lock &rq->__lock irq_context: 0 pernet_ops_rwsem rtnl_mutex rcu_read_lock &pool->lock &p->pi_lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 pernet_ops_rwsem rtnl_mutex uevent_sock_mutex &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)ipv6_addrconf (work_completion)(&(&ifa->dad_work)->work) rtnl_mutex &idev->mc_lock batched_entropy_u8.lock irq_context: 0 (wq_completion)ipv6_addrconf (work_completion)(&(&ifa->dad_work)->work) rtnl_mutex &idev->mc_lock kfence_freelist_lock irq_context: 0 (wq_completion)ipv6_addrconf (work_completion)(&(&ifa->dad_work)->work) rtnl_mutex &tb->tb6_lock &obj_hash[i].lock irq_context: 0 (wq_completion)ipv6_addrconf (work_completion)(&(&ifa->dad_work)->work) rtnl_mutex &idev->mc_lock rcu_read_lock rcu_read_lock pool_lock#2 irq_context: 0 (wq_completion)ipv6_addrconf (work_completion)(&(&ifa->dad_work)->work) rtnl_mutex &idev->mc_lock rcu_read_lock rcu_read_lock &obj_hash[i].lock irq_context: 0 (wq_completion)ipv6_addrconf (work_completion)(&(&ifa->dad_work)->work) rtnl_mutex &idev->mc_lock rcu_read_lock rcu_read_lock &c->lock irq_context: 0 (wq_completion)ipv6_addrconf (work_completion)(&(&ifa->dad_work)->work) rtnl_mutex &idev->mc_lock rcu_read_lock rcu_read_lock &dir->lock#2 irq_context: 0 (wq_completion)ipv6_addrconf (work_completion)(&(&ifa->dad_work)->work) rtnl_mutex &idev->mc_lock rcu_read_lock rcu_read_lock &tbl->lock irq_context: 0 (wq_completion)ipv6_addrconf (work_completion)(&(&ifa->dad_work)->work) rtnl_mutex &idev->mc_lock rcu_read_lock rcu_read_lock &tbl->lock pool_lock#2 irq_context: 0 (wq_completion)ipv6_addrconf (work_completion)(&(&ifa->dad_work)->work) rtnl_mutex &idev->mc_lock rcu_read_lock rcu_read_lock &tbl->lock batched_entropy_u32.lock irq_context: 0 (wq_completion)ipv6_addrconf (work_completion)(&(&ifa->dad_work)->work) rtnl_mutex &idev->mc_lock rcu_read_lock rcu_read_lock &tbl->lock &obj_hash[i].lock irq_context: 0 (wq_completion)ipv6_addrconf (work_completion)(&(&ifa->dad_work)->work) rtnl_mutex &idev->mc_lock rcu_read_lock rcu_read_lock &n->lock irq_context: 0 (wq_completion)ipv6_addrconf (work_completion)(&(&ifa->dad_work)->work) rtnl_mutex &idev->mc_lock rcu_read_lock rcu_read_lock &____s->seqcount#9 irq_context: 0 (wq_completion)ipv6_addrconf (work_completion)(&(&ifa->dad_work)->work) rtnl_mutex &idev->mc_lock rcu_read_lock rcu_read_lock rcu_read_lock_bh noop_qdisc.q.lock irq_context: 0 pernet_ops_rwsem k-sk_lock-AF_TIPC &tn->nametbl_lock &c->lock irq_context: 0 pernet_ops_rwsem nf_ct_proto_mutex defrag4_mutex nf_hook_mutex &c->lock irq_context: 0 pernet_ops_rwsem nf_ct_proto_mutex defrag4_mutex nf_hook_mutex &____s->seqcount irq_context: 0 fill_pool_map-wait-type-override rcu_read_lock rcu_node_0 irq_context: 0 fill_pool_map-wait-type-override rcu_read_lock &rq->__lock irq_context: 0 rtnl_mutex &n->lock irq_context: 0 rtnl_mutex &n->lock &(&n->ha_lock)->lock irq_context: 0 rtnl_mutex &n->lock &(&n->ha_lock)->lock &____s->seqcount#9 irq_context: 0 rtnl_mutex &tbl->lock &n->lock irq_context: 0 rtnl_mutex rcu_read_lock lock#8 irq_context: 0 rtnl_mutex rcu_read_lock id_table_lock irq_context: 0 rtnl_mutex &n->lock irq_context: 0 rtnl_mutex &n->lock &____s->seqcount#9 irq_context: 0 rtnl_mutex dev_addr_sem rcu_read_lock &ndev->lock irq_context: 0 rtnl_mutex dev_addr_sem rcu_read_lock &pool->lock/1 irq_context: 0 rtnl_mutex dev_addr_sem rcu_read_lock &pool->lock/1 &obj_hash[i].lock irq_context: 0 rtnl_mutex dev_addr_sem rcu_read_lock &pool->lock/1 pool_lock#2 irq_context: 0 rtnl_mutex dev_addr_sem rcu_read_lock &pool->lock/1 &p->pi_lock irq_context: 0 rtnl_mutex dev_addr_sem rcu_read_lock &pool->lock/1 &p->pi_lock &rq->__lock irq_context: 0 rtnl_mutex dev_addr_sem rcu_read_lock &pool->lock/1 &p->pi_lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 rtnl_mutex dev_addr_sem &tbl->lock &n->lock irq_context: 0 rtnl_mutex dev_addr_sem &tbl->lock pool_lock#2 irq_context: 0 rtnl_mutex dev_addr_sem &tbl->lock &c->lock irq_context: 0 rtnl_mutex dev_addr_sem &tbl->lock &n->list_lock irq_context: 0 rtnl_mutex dev_addr_sem &tbl->lock &n->lock irq_context: 0 rtnl_mutex dev_addr_sem &tbl->lock &n->lock &____s->seqcount#9 irq_context: 0 rtnl_mutex dev_addr_sem &tbl->lock nl_table_lock irq_context: 0 rtnl_mutex dev_addr_sem &tbl->lock &obj_hash[i].lock irq_context: 0 rtnl_mutex dev_addr_sem &tbl->lock nl_table_wait.lock irq_context: 0 rtnl_mutex dev_addr_sem &tbl->lock rcu_read_lock lock#8 irq_context: 0 rtnl_mutex dev_addr_sem &tbl->lock rcu_read_lock id_table_lock irq_context: 0 rtnl_mutex dev_addr_sem &tbl->lock &dir->lock#2 irq_context: 0 rtnl_mutex dev_addr_sem &tbl->lock krc.lock irq_context: 0 rtnl_mutex _xmit_ETHER pool_lock#2 irq_context: 0 rtnl_mutex &ndev->lock &____s->seqcount irq_context: 0 rtnl_mutex &ndev->lock pool_lock#2 irq_context: 0 rtnl_mutex &ndev->lock rcu_read_lock pool_lock#2 irq_context: 0 rtnl_mutex &ndev->lock &obj_hash[i].lock irq_context: 0 rtnl_mutex &ndev->lock &dir->lock#2 irq_context: 0 rtnl_mutex &ndev->lock pcpu_lock irq_context: 0 rtnl_mutex &ndev->lock &tb->tb6_lock irq_context: 0 rtnl_mutex &ndev->lock &tb->tb6_lock pool_lock#2 irq_context: 0 rtnl_mutex &ndev->lock &tb->tb6_lock &c->lock irq_context: 0 rtnl_mutex &ndev->lock &tb->tb6_lock nl_table_lock irq_context: 0 rtnl_mutex &ndev->lock &tb->tb6_lock &obj_hash[i].lock irq_context: 0 rtnl_mutex &ndev->lock &tb->tb6_lock nl_table_wait.lock irq_context: 0 rtnl_mutex &idev->mc_lock _xmit_ETHER &____s->seqcount irq_context: 0 rtnl_mutex &idev->mc_lock _xmit_ETHER rcu_read_lock pool_lock#2 irq_context: 0 rtnl_mutex &idev->mc_lock _xmit_ETHER &obj_hash[i].lock irq_context: 0 rtnl_mutex &idev->mc_lock _xmit_ETHER &c->lock irq_context: 0 (wq_completion)mld (work_completion)(&(&idev->mc_ifc_work)->work) &idev->mc_lock &pcp->lock &zone->lock irq_context: 0 (wq_completion)mld (work_completion)(&(&idev->mc_ifc_work)->work) &idev->mc_lock &pcp->lock &zone->lock &____s->seqcount irq_context: 0 (wq_completion)mld (work_completion)(&(&idev->mc_ifc_work)->work) &idev->mc_lock &____s->seqcount irq_context: 0 (wq_completion)mld (work_completion)(&(&idev->mc_ifc_work)->work) &idev->mc_lock rcu_read_lock rcu_read_lock &c->lock irq_context: 0 (wq_completion)mld (work_completion)(&(&idev->mc_ifc_work)->work) &idev->mc_lock rcu_read_lock rcu_read_lock rcu_read_lock_bh dev->qdisc_tx_busylock ?: &qdisc_tx_busylock rcu_read_lock &r->producer_lock irq_context: 0 (wq_completion)ipv6_addrconf (work_completion)(&(&ifa->dad_work)->work) rtnl_mutex &idev->mc_lock &obj_hash[i].lock pool_lock irq_context: 0 (wq_completion)ipv6_addrconf (work_completion)(&(&ifa->dad_work)->work) rtnl_mutex &idev->mc_lock fill_pool_map-wait-type-override pool_lock#2 irq_context: 0 (wq_completion)ipv6_addrconf (work_completion)(&(&ifa->dad_work)->work) rtnl_mutex &idev->mc_lock fill_pool_map-wait-type-override pool_lock irq_context: 0 (wq_completion)ipv6_addrconf (work_completion)(&(&ifa->dad_work)->work) &p->pi_lock &rq->__lock irq_context: 0 (wq_completion)ipv6_addrconf (work_completion)(&(&ifa->dad_work)->work) &p->pi_lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 rtnl_mutex dev_addr_sem &obj_hash[i].lock pool_lock irq_context: 0 rtnl_mutex dev_addr_sem rcu_read_lock &pool->lock/1 fill_pool_map-wait-type-override pool_lock#2 irq_context: 0 rtnl_mutex dev_addr_sem rcu_read_lock &pool->lock/1 fill_pool_map-wait-type-override &c->lock irq_context: 0 rtnl_mutex dev_addr_sem rcu_read_lock &pool->lock/1 fill_pool_map-wait-type-override &____s->seqcount irq_context: 0 rtnl_mutex dev_addr_sem rcu_read_lock &pool->lock/1 fill_pool_map-wait-type-override pool_lock irq_context: 0 rtnl_mutex &ndev->lock &c->lock irq_context: 0 pernet_ops_rwsem rtnl_mutex remove_cache_srcu &c->lock irq_context: 0 pernet_ops_rwsem rtnl_mutex remove_cache_srcu &n->list_lock irq_context: 0 pernet_ops_rwsem rtnl_mutex remove_cache_srcu rcu_read_lock pool_lock#2 irq_context: 0 pernet_ops_rwsem rtnl_mutex remove_cache_srcu &obj_hash[i].lock irq_context: 0 rtnl_mutex uevent_sock_mutex rcu_read_lock pool_lock#2 irq_context: 0 rtnl_mutex rcu_read_lock &pool->lock/1 fill_pool_map-wait-type-override pool_lock#2 irq_context: 0 rtnl_mutex rcu_read_lock &pool->lock/1 fill_pool_map-wait-type-override pool_lock irq_context: 0 rtnl_mutex &idev->mc_lock &bridge_netdev_addr_lock_key irq_context: 0 rtnl_mutex &idev->mc_lock &bridge_netdev_addr_lock_key pool_lock#2 irq_context: 0 rtnl_mutex (switchdev_blocking_notif_chain).rwsem irq_context: 0 rtnl_mutex &br->hash_lock irq_context: 0 rtnl_mutex &br->hash_lock &____s->seqcount irq_context: 0 rtnl_mutex &br->hash_lock pool_lock#2 irq_context: 0 rtnl_mutex &br->hash_lock &c->lock irq_context: 0 rtnl_mutex &br->hash_lock rcu_read_lock rhashtable_bucket irq_context: 0 rtnl_mutex &br->hash_lock nl_table_lock irq_context: 0 rtnl_mutex &br->hash_lock &obj_hash[i].lock irq_context: 0 rtnl_mutex &br->hash_lock nl_table_wait.lock irq_context: 0 rtnl_mutex rcu_read_lock rhashtable_bucket irq_context: 0 rtnl_mutex nf_hook_mutex irq_context: 0 rtnl_mutex nf_hook_mutex fs_reclaim irq_context: 0 rtnl_mutex nf_hook_mutex fs_reclaim mmu_notifier_invalidate_range_start irq_context: 0 rtnl_mutex nf_hook_mutex pool_lock#2 irq_context: 0 rtnl_mutex cpu_hotplug_lock jump_label_mutex irq_context: 0 rtnl_mutex cpu_hotplug_lock jump_label_mutex text_mutex irq_context: 0 rtnl_mutex cpu_hotplug_lock jump_label_mutex text_mutex ptlock_ptr(page)#2 irq_context: 0 rtnl_mutex fill_pool_map-wait-type-override &____s->seqcount irq_context: 0 (wq_completion)mld (work_completion)(&(&idev->mc_ifc_work)->work) &idev->mc_lock rcu_read_lock rcu_read_lock rcu_read_lock_bh &obj_hash[i].lock pool_lock irq_context: 0 (wq_completion)mld (work_completion)(&(&idev->mc_ifc_work)->work) &idev->mc_lock fill_pool_map-wait-type-override pool_lock#2 irq_context: 0 (wq_completion)mld (work_completion)(&(&idev->mc_ifc_work)->work) &idev->mc_lock fill_pool_map-wait-type-override pool_lock irq_context: 0 rtnl_mutex j1939_netdev_lock irq_context: 0 pernet_ops_rwsem k-sk_lock-AF_RXRPC &rxnet->local_mutex &c->lock irq_context: 0 pernet_ops_rwsem rtnl_mutex &idev->mc_lock &rq->__lock irq_context: 0 pernet_ops_rwsem nf_ct_proto_mutex nf_hook_mutex &c->lock irq_context: 0 pernet_ops_rwsem nf_ct_proto_mutex nf_hook_mutex &____s->seqcount irq_context: 0 rtnl_mutex &idev->mc_lock &bridge_netdev_addr_lock_key &c->lock irq_context: 0 rtnl_mutex &root->kernfs_rwsem &root->kernfs_iattr_rwsem &rq->__lock irq_context: 0 rtnl_mutex cpu_hotplug_lock wq_pool_mutex &c->lock irq_context: 0 rtnl_mutex cpu_hotplug_lock wq_pool_mutex &____s->seqcount irq_context: 0 rtnl_mutex uevent_sock_mutex quarantine_lock irq_context: 0 rtnl_mutex remove_cache_srcu irq_context: 0 rtnl_mutex remove_cache_srcu quarantine_lock irq_context: 0 rtnl_mutex remove_cache_srcu &n->list_lock irq_context: 0 rtnl_mutex remove_cache_srcu rcu_read_lock pool_lock#2 irq_context: 0 rtnl_mutex remove_cache_srcu &obj_hash[i].lock irq_context: 0 rtnl_mutex remove_cache_srcu &c->lock irq_context: 0 rtnl_mutex &idev->mc_lock &dev_addr_list_lock_key irq_context: 0 rtnl_mutex &idev->mc_lock &dev_addr_list_lock_key pool_lock#2 irq_context: 0 rtnl_mutex uevent_sock_mutex &pcp->lock &zone->lock &____s->seqcount irq_context: 0 rtnl_mutex &idev->mc_lock &dev_addr_list_lock_key#2 irq_context: 0 rtnl_mutex &idev->mc_lock &dev_addr_list_lock_key#2 pool_lock#2 irq_context: 0 pernet_ops_rwsem rtnl_mutex fs_reclaim &rq->__lock irq_context: 0 (wq_completion)events (linkwatch_work).work rtnl_mutex.wait_lock irq_context: 0 (wq_completion)events (linkwatch_work).work &p->pi_lock irq_context: 0 (wq_completion)events (linkwatch_work).work &p->pi_lock &rq->__lock irq_context: 0 (wq_completion)events (linkwatch_work).work &p->pi_lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 pernet_ops_rwsem nf_ct_proto_mutex &obj_hash[i].lock pool_lock irq_context: 0 pernet_ops_rwsem nf_ct_proto_mutex fill_pool_map-wait-type-override pool_lock#2 irq_context: 0 pernet_ops_rwsem nf_ct_proto_mutex fill_pool_map-wait-type-override pool_lock irq_context: 0 (wq_completion)mld (work_completion)(&(&idev->mc_dad_work)->work) &idev->mc_lock rcu_read_lock &c->lock irq_context: 0 (wq_completion)mld (work_completion)(&(&idev->mc_dad_work)->work) &idev->mc_lock rcu_read_lock &____s->seqcount irq_context: 0 (wq_completion)mld (work_completion)(&(&idev->mc_dad_work)->work) &idev->mc_lock rcu_read_lock rcu_read_lock rcu_read_lock_bh dev->qdisc_tx_busylock ?: &qdisc_tx_busylock rcu_read_lock &r->producer_lock irq_context: 0 &mm->mmap_lock fill_pool_map-wait-type-override &c->lock irq_context: 0 &mm->mmap_lock fill_pool_map-wait-type-override &____s->seqcount irq_context: 0 pernet_ops_rwsem nf_ct_proto_mutex &rq->__lock irq_context: 0 rtnl_mutex dev_addr_sem &tbl->lock &obj_hash[i].lock pool_lock irq_context: 0 rtnl_mutex dev_addr_sem &tbl->lock fill_pool_map-wait-type-override pool_lock#2 irq_context: 0 rtnl_mutex dev_addr_sem &tbl->lock fill_pool_map-wait-type-override &c->lock irq_context: 0 rtnl_mutex dev_addr_sem &tbl->lock fill_pool_map-wait-type-override &____s->seqcount irq_context: 0 rtnl_mutex dev_addr_sem &tbl->lock fill_pool_map-wait-type-override pool_lock irq_context: 0 rtnl_mutex &ndev->lock &ifa->lock irq_context: 0 rtnl_mutex &ndev->lock rcu_read_lock &pool->lock irq_context: 0 rtnl_mutex &ndev->lock rcu_read_lock &pool->lock &obj_hash[i].lock irq_context: 0 rtnl_mutex &ndev->lock rcu_read_lock &pool->lock pool_lock#2 irq_context: 0 rtnl_mutex &ndev->lock rcu_read_lock &pool->lock &p->pi_lock irq_context: 0 pernet_ops_rwsem k-sk_lock-AF_RXRPC &pcp->lock &zone->lock irq_context: 0 pernet_ops_rwsem k-sk_lock-AF_RXRPC &pcp->lock &zone->lock &____s->seqcount irq_context: 0 pernet_ops_rwsem k-sk_lock-AF_RXRPC &____s->seqcount irq_context: 0 rtnl_mutex quarantine_lock irq_context: 0 rtnl_mutex &idev->mc_lock &dev_addr_list_lock_key &c->lock irq_context: 0 rtnl_mutex &idev->mc_lock &dev_addr_list_lock_key &n->list_lock irq_context: 0 rtnl_mutex &idev->mc_lock &dev_addr_list_lock_key &n->list_lock &c->lock irq_context: 0 rtnl_mutex lock kernfs_idr_lock &n->list_lock irq_context: 0 rtnl_mutex lock kernfs_idr_lock &n->list_lock &c->lock irq_context: 0 rtnl_mutex rcu_read_lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)ipv6_addrconf (work_completion)(&(&ifa->dad_work)->work) rtnl_mutex &idev->mc_lock rcu_read_lock rcu_read_lock rcu_read_lock_bh dev->qdisc_tx_busylock ?: &qdisc_tx_busylock rcu_read_lock &r->producer_lock irq_context: 0 (wq_completion)ipv6_addrconf (work_completion)(&(&ifa->dad_work)->work) rtnl_mutex &idev->mc_lock &c->lock irq_context: 0 (wq_completion)ipv6_addrconf (work_completion)(&(&ifa->dad_work)->work) rtnl_mutex &idev->mc_lock &n->list_lock irq_context: 0 (wq_completion)ipv6_addrconf (work_completion)(&(&ifa->dad_work)->work) rtnl_mutex &idev->mc_lock &n->list_lock &c->lock irq_context: 0 (wq_completion)ipv6_addrconf (work_completion)(&(&ifa->dad_work)->work) rtnl_mutex &n->list_lock irq_context: 0 (wq_completion)ipv6_addrconf (work_completion)(&(&ifa->dad_work)->work) rtnl_mutex &n->list_lock &c->lock irq_context: 0 (wq_completion)events_unbound (work_completion)(&sub_info->work) tasklist_lock &c->lock irq_context: 0 rtnl_mutex &rq->__lock &cfs_rq->removed.lock irq_context: 0 rtnl_mutex nf_hook_mutex &c->lock irq_context: 0 rtnl_mutex nf_hook_mutex &____s->seqcount irq_context: 0 rtnl_mutex rcu_read_lock &pool->lock/1 &p->pi_lock &cfs_rq->removed.lock irq_context: 0 rtnl_mutex &idev->mc_lock fill_pool_map-wait-type-override pool_lock#2 irq_context: 0 rtnl_mutex &idev->mc_lock fill_pool_map-wait-type-override pool_lock irq_context: 0 rtnl_mutex dev_addr_sem rcu_read_lock &pool->lock/1 &p->pi_lock &rq->__lock &cfs_rq->removed.lock irq_context: 0 rtnl_mutex &ndev->lock &n->list_lock irq_context: 0 rtnl_mutex &ndev->lock &n->list_lock &c->lock irq_context: 0 rtnl_mutex &ndev->lock &tb->tb6_lock &____s->seqcount irq_context: 0 rtnl_mutex rcu_read_lock &obj_hash[i].lock pool_lock irq_context: 0 rtnl_mutex rcu_read_lock rcu_read_lock &pool->lock/1 fill_pool_map-wait-type-override pool_lock#2 irq_context: 0 rtnl_mutex rcu_read_lock rcu_read_lock &pool->lock/1 fill_pool_map-wait-type-override pool_lock irq_context: 0 (wq_completion)ipv6_addrconf (work_completion)(&(&ifa->dad_work)->work) rtnl_mutex &idev->mc_lock _xmit_ETHER &c->lock irq_context: 0 (wq_completion)ipv6_addrconf (work_completion)(&(&ifa->dad_work)->work) rtnl_mutex &idev->mc_lock _xmit_ETHER &____s->seqcount irq_context: 0 (wq_completion)ipv6_addrconf (work_completion)(&(&ifa->dad_work)->work) rtnl_mutex fill_pool_map-wait-type-override pool_lock#2 irq_context: 0 (wq_completion)ipv6_addrconf (work_completion)(&(&ifa->dad_work)->work) rtnl_mutex fill_pool_map-wait-type-override &c->lock irq_context: 0 (wq_completion)ipv6_addrconf (work_completion)(&(&ifa->dad_work)->work) rtnl_mutex fill_pool_map-wait-type-override &____s->seqcount irq_context: 0 (wq_completion)ipv6_addrconf (work_completion)(&(&ifa->dad_work)->work) rtnl_mutex fill_pool_map-wait-type-override pool_lock irq_context: 0 rtnl_mutex &zone->lock &____s->seqcount irq_context: 0 rtnl_mutex &bat_priv->tvlv.handler_list_lock irq_context: 0 rtnl_mutex &bat_priv->tvlv.handler_list_lock pool_lock#2 irq_context: 0 rtnl_mutex &bat_priv->tvlv.container_list_lock irq_context: 0 rtnl_mutex &idev->mc_lock &batadv_netdev_addr_lock_key irq_context: 0 rtnl_mutex &idev->mc_lock &batadv_netdev_addr_lock_key pool_lock#2 irq_context: 0 rtnl_mutex &bat_priv->softif_vlan_list_lock irq_context: 0 rtnl_mutex &bat_priv->softif_vlan_list_lock pool_lock#2 irq_context: 0 rtnl_mutex key#16 irq_context: 0 rtnl_mutex &bat_priv->tt.changes_list_lock irq_context: softirq &(&bat_priv->nc.work)->timer irq_context: softirq &(&bat_priv->nc.work)->timer rcu_read_lock &pool->lock/1 irq_context: softirq &(&bat_priv->nc.work)->timer rcu_read_lock &pool->lock/1 &obj_hash[i].lock irq_context: softirq &(&bat_priv->nc.work)->timer rcu_read_lock &pool->lock/1 &p->pi_lock irq_context: softirq &(&bat_priv->nc.work)->timer rcu_read_lock &pool->lock/1 &p->pi_lock &cfs_rq->removed.lock irq_context: softirq &(&bat_priv->nc.work)->timer rcu_read_lock &pool->lock/1 &p->pi_lock &rq->__lock irq_context: softirq &(&bat_priv->nc.work)->timer rcu_read_lock &pool->lock/1 &p->pi_lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: softirq &(&bat_priv->nc.work)->timer rcu_read_lock &pool->lock/1 &p->pi_lock &rq->__lock &cfs_rq->removed.lock irq_context: 0 (wq_completion)bat_events irq_context: 0 (wq_completion)bat_events (work_completion)(&(&bat_priv->nc.work)->work) irq_context: 0 (wq_completion)bat_events (work_completion)(&(&bat_priv->nc.work)->work) rcu_read_lock rcu_node_0 irq_context: 0 (wq_completion)bat_events (work_completion)(&(&bat_priv->nc.work)->work) rcu_read_lock &rq->__lock irq_context: 0 (wq_completion)bat_events (work_completion)(&(&bat_priv->nc.work)->work) key#17 irq_context: 0 (wq_completion)bat_events (work_completion)(&(&bat_priv->nc.work)->work) key#18 irq_context: 0 (wq_completion)bat_events (work_completion)(&(&bat_priv->nc.work)->work) &obj_hash[i].lock irq_context: 0 (wq_completion)bat_events (work_completion)(&(&bat_priv->nc.work)->work) &base->lock irq_context: 0 (wq_completion)bat_events (work_completion)(&(&bat_priv->nc.work)->work) &base->lock &obj_hash[i].lock irq_context: 0 rtnl_mutex &idev->mc_lock fill_pool_map-wait-type-override &c->lock irq_context: 0 rtnl_mutex &idev->mc_lock fill_pool_map-wait-type-override &____s->seqcount irq_context: 0 rtnl_mutex remove_cache_srcu &rq->__lock irq_context: 0 rtnl_mutex remove_cache_srcu &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 rtnl_mutex remove_cache_srcu rcu_read_lock &pcp->lock &zone->lock irq_context: 0 rtnl_mutex remove_cache_srcu rcu_read_lock &pcp->lock &zone->lock &____s->seqcount irq_context: 0 rtnl_mutex remove_cache_srcu rcu_read_lock &____s->seqcount irq_context: 0 rtnl_mutex &root->kernfs_rwsem kernfs_idr_lock irq_context: 0 rtnl_mutex &root->kernfs_rwsem pool_lock#2 irq_context: 0 rtnl_mutex kernfs_idr_lock irq_context: 0 rtnl_mutex rcu_state.exp_mutex rcu_node_0 irq_context: 0 rtnl_mutex rcu_state.exp_mutex &obj_hash[i].lock irq_context: 0 rtnl_mutex rcu_state.exp_mutex rcu_read_lock &pool->lock irq_context: 0 rtnl_mutex rcu_state.exp_mutex rcu_read_lock &pool->lock &obj_hash[i].lock irq_context: 0 rtnl_mutex rcu_state.exp_mutex rcu_read_lock &pool->lock &p->pi_lock irq_context: 0 rtnl_mutex rcu_state.exp_mutex rcu_read_lock &pool->lock &p->pi_lock &rq->__lock irq_context: 0 rtnl_mutex rcu_state.exp_mutex rcu_read_lock &pool->lock &p->pi_lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 rtnl_mutex rcu_state.exp_mutex &rnp->exp_wq[0] irq_context: 0 rtnl_mutex rcu_state.exp_mutex &rq->__lock irq_context: 0 rtnl_mutex rcu_state.exp_mutex &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 rtnl_mutex cpu_hotplug_lock xps_map_mutex irq_context: 0 rtnl_mutex rcu_state.exp_mutex &rnp->exp_wq[1] irq_context: 0 rtnl_mutex &bat_priv->softif_vlan_list_lock &c->lock irq_context: 0 rtnl_mutex &idev->mc_lock &n->list_lock irq_context: 0 rtnl_mutex &idev->mc_lock &n->list_lock &c->lock irq_context: 0 rtnl_mutex rcu_state.exp_mutex &rnp->exp_wq[2] irq_context: 0 rtnl_mutex rcu_state.exp_mutex &rnp->exp_wq[3] irq_context: 0 rtnl_mutex tk_core.seq.seqcount irq_context: 0 rtnl_mutex &wq->mutex irq_context: 0 rtnl_mutex &wq->mutex &pool->lock irq_context: 0 rtnl_mutex wq_pool_mutex &wq->mutex &pool->lock irq_context: 0 rtnl_mutex wq_pool_mutex &wq->mutex &pool->lock/1 irq_context: 0 rtnl_mutex init_lock irq_context: 0 rtnl_mutex init_lock slab_mutex irq_context: 0 rtnl_mutex init_lock slab_mutex fs_reclaim irq_context: 0 rtnl_mutex init_lock slab_mutex fs_reclaim mmu_notifier_invalidate_range_start irq_context: 0 rtnl_mutex init_lock slab_mutex &c->lock irq_context: 0 rtnl_mutex init_lock slab_mutex &____s->seqcount irq_context: 0 rtnl_mutex init_lock slab_mutex pool_lock#2 irq_context: 0 rtnl_mutex init_lock slab_mutex &n->list_lock irq_context: 0 rtnl_mutex init_lock slab_mutex pcpu_alloc_mutex irq_context: 0 rtnl_mutex init_lock slab_mutex pcpu_alloc_mutex pcpu_lock irq_context: 0 rtnl_mutex init_lock slab_mutex &root->kernfs_rwsem irq_context: 0 rtnl_mutex init_lock slab_mutex &k->list_lock irq_context: 0 rtnl_mutex init_lock slab_mutex lock irq_context: 0 rtnl_mutex init_lock slab_mutex lock kernfs_idr_lock irq_context: 0 rtnl_mutex init_lock slab_mutex &root->kernfs_rwsem &root->kernfs_iattr_rwsem irq_context: 0 rtnl_mutex init_lock slab_mutex lock kernfs_idr_lock pool_lock#2 irq_context: 0 rtnl_mutex init_lock slab_mutex &rq->__lock irq_context: 0 rtnl_mutex init_lock fs_reclaim irq_context: 0 rtnl_mutex init_lock fs_reclaim mmu_notifier_invalidate_range_start irq_context: 0 rtnl_mutex init_lock &zone->lock irq_context: 0 rtnl_mutex init_lock &____s->seqcount irq_context: 0 rtnl_mutex init_lock pool_lock#2 irq_context: 0 rtnl_mutex init_lock &zone->lock &____s->seqcount irq_context: 0 rtnl_mutex init_lock &obj_hash[i].lock irq_context: 0 rtnl_mutex init_lock &base->lock irq_context: 0 rtnl_mutex init_lock &base->lock &obj_hash[i].lock irq_context: 0 rtnl_mutex init_lock crngs.lock irq_context: 0 rtnl_mutex &idev->mc_lock &dev_addr_list_lock_key#2 &c->lock irq_context: 0 rtnl_mutex &idev->mc_lock &dev_addr_list_lock_key#2 &____s->seqcount irq_context: 0 rtnl_mutex fs_reclaim mmu_notifier_invalidate_range_start &rq->__lock irq_context: 0 rtnl_mutex fs_reclaim mmu_notifier_invalidate_range_start &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 rtnl_mutex &bat_priv->tvlv.handler_list_lock &c->lock irq_context: 0 rtnl_mutex &meta->lock irq_context: 0 rtnl_mutex &root->kernfs_rwsem kernfs_idr_lock &obj_hash[i].lock irq_context: 0 rtnl_mutex &root->kernfs_rwsem kernfs_idr_lock pool_lock#2 irq_context: 0 &mm->mmap_lock rcu_read_lock &rcu_state.expedited_wq irq_context: 0 &mm->mmap_lock rcu_read_lock &rcu_state.expedited_wq &p->pi_lock irq_context: 0 rtnl_mutex rcu_state.exp_mutex rcu_read_lock &rq->__lock irq_context: 0 rtnl_mutex rcu_state.exp_mutex rcu_read_lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 &mm->mmap_lock rcu_read_lock &rcu_state.expedited_wq irq_context: 0 rtnl_mutex fs_reclaim &rq->__lock irq_context: 0 rtnl_mutex fs_reclaim &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 rtnl_mutex fill_pool_map-wait-type-override &rq->__lock irq_context: 0 &mm->mmap_lock &mapping->i_mmap_rwsem &anon_vma->rwsem fill_pool_map-wait-type-override &rq->__lock irq_context: 0 &mm->mmap_lock rcu_read_lock &rcu_state.expedited_wq &p->pi_lock irq_context: 0 &mm->mmap_lock ptlock_ptr(page)#2 rcu_read_lock &p->pi_lock irq_context: 0 &mm->mmap_lock ptlock_ptr(page)#2 rcu_read_lock &p->pi_lock &rq->__lock irq_context: 0 &mm->mmap_lock ptlock_ptr(page)#2 rcu_read_lock &p->pi_lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 rtnl_mutex cpu_hotplug_lock wq_pool_mutex &n->list_lock irq_context: 0 rtnl_mutex cpu_hotplug_lock wq_pool_mutex &n->list_lock &c->lock irq_context: 0 (wq_completion)bat_events (work_completion)(&(&bat_priv->nc.work)->work) &rq->__lock irq_context: 0 rtnl_mutex cpu_hotplug_lock wq_pool_mutex &pcp->lock &zone->lock irq_context: 0 rtnl_mutex cpu_hotplug_lock wq_pool_mutex &pcp->lock &zone->lock &____s->seqcount irq_context: 0 (wq_completion)bat_events (work_completion)(&(&bat_priv->nc.work)->work) rcu_read_lock &rcu_state.expedited_wq irq_context: 0 (wq_completion)bat_events (work_completion)(&(&bat_priv->nc.work)->work) rcu_read_lock &rcu_state.expedited_wq &p->pi_lock irq_context: 0 rtnl_mutex &idev->mc_lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 rtnl_mutex &idev->mc_lock batched_entropy_u8.lock irq_context: 0 rtnl_mutex &idev->mc_lock kfence_freelist_lock irq_context: 0 rtnl_mutex deferred_lock irq_context: 0 (wq_completion)events deferred_process_work irq_context: 0 (wq_completion)events deferred_process_work rtnl_mutex irq_context: 0 (wq_completion)events deferred_process_work rtnl_mutex rtnl_mutex.wait_lock irq_context: 0 (wq_completion)events deferred_process_work rtnl_mutex &pool->lock irq_context: 0 (wq_completion)events deferred_process_work rtnl_mutex &rq->__lock irq_context: 0 (wq_completion)events deferred_process_work rtnl_mutex &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 rtnl_mutex target_list_lock irq_context: 0 (wq_completion)bat_events (work_completion)(&(&bat_priv->nc.work)->work) rcu_read_lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 rtnl_mutex rcu_read_lock _xmit_ETHER irq_context: 0 rtnl_mutex &br->lock irq_context: 0 rtnl_mutex &br->lock &br->hash_lock irq_context: 0 rtnl_mutex &br->lock &br->hash_lock rcu_read_lock rhashtable_bucket irq_context: 0 rtnl_mutex &br->lock &br->hash_lock rcu_read_lock rcu_read_lock &pool->lock irq_context: 0 rtnl_mutex &br->lock &br->hash_lock rcu_read_lock rcu_read_lock &pool->lock &obj_hash[i].lock irq_context: 0 rtnl_mutex &br->lock &br->hash_lock rcu_read_lock rcu_read_lock &pool->lock pool_lock#2 irq_context: 0 rtnl_mutex &br->lock &br->hash_lock rcu_read_lock rcu_read_lock &pool->lock &p->pi_lock irq_context: 0 rtnl_mutex &br->lock &br->hash_lock rcu_read_lock rcu_read_lock &pool->lock &p->pi_lock &rq->__lock irq_context: 0 rtnl_mutex &br->lock &br->hash_lock rcu_read_lock rcu_read_lock &pool->lock &p->pi_lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 rtnl_mutex &br->lock &br->hash_lock &c->lock irq_context: 0 rtnl_mutex &br->lock &br->hash_lock &____s->seqcount irq_context: 0 rtnl_mutex &br->lock &br->hash_lock pool_lock#2 irq_context: 0 rtnl_mutex &br->lock &br->hash_lock nl_table_lock irq_context: 0 rtnl_mutex &br->lock &br->hash_lock &obj_hash[i].lock irq_context: 0 rtnl_mutex &br->lock &br->hash_lock nl_table_wait.lock irq_context: 0 rtnl_mutex &br->lock lweventlist_lock irq_context: 0 rtnl_mutex &br->lock lweventlist_lock pool_lock#2 irq_context: 0 rtnl_mutex &br->lock lweventlist_lock &dir->lock#2 irq_context: 0 rtnl_mutex &pn->hash_lock irq_context: 0 rtnl_mutex &net->ipv6.fib6_gc_lock rcu_read_lock &tb->tb6_lock irq_context: 0 rtnl_mutex &net->ipv6.fib6_gc_lock &obj_hash[i].lock irq_context: 0 (wq_completion)bat_events (work_completion)(&(&bat_priv->nc.work)->work) fill_pool_map-wait-type-override pool_lock#2 irq_context: 0 (wq_completion)bat_events (work_completion)(&(&bat_priv->nc.work)->work) fill_pool_map-wait-type-override pool_lock irq_context: 0 rtnl_mutex &br->hash_lock rcu_read_lock rcu_read_lock &pool->lock irq_context: 0 rtnl_mutex &br->hash_lock rcu_read_lock rcu_read_lock &pool->lock &obj_hash[i].lock irq_context: 0 rtnl_mutex &br->hash_lock rcu_read_lock rcu_read_lock &pool->lock pool_lock#2 irq_context: 0 rtnl_mutex &br->hash_lock rcu_read_lock rcu_read_lock &pool->lock &p->pi_lock irq_context: 0 rtnl_mutex &br->hash_lock rcu_read_lock rcu_read_lock &pool->lock &p->pi_lock &rq->__lock irq_context: 0 rtnl_mutex &br->hash_lock rcu_read_lock rcu_read_lock &pool->lock &p->pi_lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)events_unbound (work_completion)(&(&kfence_timer)->work) fill_pool_map-wait-type-override pool_lock#2 irq_context: 0 (wq_completion)events_unbound (work_completion)(&(&kfence_timer)->work) fill_pool_map-wait-type-override pool_lock irq_context: 0 rtnl_mutex net_rwsem &rq->__lock irq_context: 0 rtnl_mutex net_rwsem &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 rtnl_mutex &br->hash_lock &n->list_lock irq_context: 0 rtnl_mutex &br->hash_lock &n->list_lock &c->lock irq_context: 0 (wq_completion)events deferred_process_work rtnl_mutex deferred_lock irq_context: 0 (wq_completion)events deferred_process_work rtnl_mutex (switchdev_blocking_notif_chain).rwsem irq_context: 0 (wq_completion)events deferred_process_work rtnl_mutex pool_lock#2 irq_context: 0 (wq_completion)events deferred_process_work rtnl_mutex &dir->lock#2 irq_context: 0 (wq_completion)events deferred_process_work rtnl_mutex &obj_hash[i].lock irq_context: 0 (wq_completion)events deferred_process_work rtnl_mutex &meta->lock irq_context: 0 (wq_completion)events deferred_process_work rtnl_mutex kfence_freelist_lock irq_context: 0 (wq_completion)events deferred_process_work rtnl_mutex.wait_lock irq_context: 0 (wq_completion)events deferred_process_work &p->pi_lock irq_context: 0 (wq_completion)events deferred_process_work &p->pi_lock &rq->__lock irq_context: 0 (wq_completion)events deferred_process_work &p->pi_lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)events deferred_process_work rtnl_mutex &pool->lock &p->pi_lock irq_context: 0 (wq_completion)events deferred_process_work rtnl_mutex &pool->lock &p->pi_lock &rq->__lock irq_context: 0 (wq_completion)events deferred_process_work rtnl_mutex &pool->lock &p->pi_lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: softirq &(&bat_priv->mcast.work)->timer irq_context: softirq &(&bat_priv->mcast.work)->timer rcu_read_lock &pool->lock/1 irq_context: softirq &(&bat_priv->mcast.work)->timer rcu_read_lock &pool->lock/1 &obj_hash[i].lock irq_context: softirq &(&bat_priv->mcast.work)->timer rcu_read_lock &pool->lock/1 &p->pi_lock irq_context: softirq &(&bat_priv->mcast.work)->timer rcu_read_lock &pool->lock/1 &p->pi_lock &rq->__lock irq_context: softirq &(&bat_priv->mcast.work)->timer rcu_read_lock &pool->lock/1 &p->pi_lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)bat_events (work_completion)(&(&bat_priv->mcast.work)->work) irq_context: 0 (wq_completion)bat_events (work_completion)(&(&bat_priv->mcast.work)->work) rcu_read_lock pool_lock#2 irq_context: 0 (wq_completion)bat_events (work_completion)(&(&bat_priv->mcast.work)->work) &bat_priv->mcast.mla_lock irq_context: 0 (wq_completion)bat_events (work_completion)(&(&bat_priv->mcast.work)->work) &bat_priv->mcast.mla_lock pool_lock#2 irq_context: 0 (wq_completion)bat_events (work_completion)(&(&bat_priv->mcast.work)->work) &bat_priv->mcast.mla_lock key#16 irq_context: 0 (wq_completion)bat_events (work_completion)(&(&bat_priv->mcast.work)->work) &bat_priv->mcast.mla_lock &bat_priv->tt.changes_list_lock irq_context: 0 (wq_completion)bat_events (work_completion)(&(&bat_priv->mcast.work)->work) &bat_priv->mcast.mla_lock &bat_priv->tvlv.container_list_lock irq_context: 0 (wq_completion)bat_events (work_completion)(&(&bat_priv->mcast.work)->work) &obj_hash[i].lock irq_context: 0 (wq_completion)bat_events (work_completion)(&(&bat_priv->mcast.work)->work) &base->lock irq_context: 0 (wq_completion)bat_events (work_completion)(&(&bat_priv->mcast.work)->work) &base->lock &obj_hash[i].lock irq_context: 0 (wq_completion)events (work_completion)(&ht->run_work) &ht->mutex &c->lock irq_context: 0 rtnl_mutex rcu_state.exp_mutex rcu_read_lock &pool->lock &p->pi_lock &rq->__lock &cfs_rq->removed.lock irq_context: 0 (wq_completion)bat_events (work_completion)(&(&bat_priv->nc.work)->work) rcu_read_lock &rcu_state.expedited_wq &p->pi_lock &rq->__lock irq_context: 0 (wq_completion)rcu_gp (work_completion)(&rew->rew_work) fill_pool_map-wait-type-override pool_lock#2 irq_context: 0 (wq_completion)rcu_gp (work_completion)(&rew->rew_work) fill_pool_map-wait-type-override pool_lock irq_context: 0 (wq_completion)bat_events (work_completion)(&(&bat_priv->nc.work)->work) rcu_node_0 irq_context: 0 (wq_completion)bat_events (work_completion)(&(&bat_priv->nc.work)->work) &rcu_state.expedited_wq irq_context: 0 (wq_completion)rcu_gp (work_completion)(&rew->rew_work) rcu_read_lock &rq->__lock irq_context: 0 (wq_completion)bat_events (work_completion)(&(&bat_priv->nc.work)->work) &rcu_state.expedited_wq &p->pi_lock irq_context: 0 (wq_completion)rcu_gp (work_completion)(&rew->rew_work) rcu_read_lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)bat_events (work_completion)(&(&bat_priv->nc.work)->work) &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 rtnl_mutex lweventlist_lock &c->lock irq_context: 0 rtnl_mutex rcu_read_lock &bond->stats_lock/1 irq_context: 0 (wq_completion)bond0 irq_context: 0 (wq_completion)bond0 (work_completion)(&(&slave->notify_work)->work) irq_context: 0 (wq_completion)bond0 (work_completion)(&(&slave->notify_work)->work) &obj_hash[i].lock irq_context: 0 (wq_completion)bond0 (work_completion)(&(&slave->notify_work)->work) &base->lock irq_context: 0 (wq_completion)bond0 (work_completion)(&(&slave->notify_work)->work) &base->lock &obj_hash[i].lock irq_context: 0 rtnl_mutex lweventlist_lock &n->list_lock irq_context: 0 rtnl_mutex lweventlist_lock &n->list_lock &c->lock irq_context: softirq &(&slave->notify_work)->timer irq_context: softirq &(&slave->notify_work)->timer rcu_read_lock &pool->lock/1 irq_context: softirq &(&slave->notify_work)->timer rcu_read_lock &pool->lock/1 &obj_hash[i].lock irq_context: softirq &(&slave->notify_work)->timer rcu_read_lock &pool->lock/1 &p->pi_lock irq_context: softirq &(&slave->notify_work)->timer rcu_read_lock &pool->lock/1 &p->pi_lock &rq->__lock irq_context: softirq &(&slave->notify_work)->timer rcu_read_lock &pool->lock/1 &p->pi_lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)mld (work_completion)(&(&idev->mc_dad_work)->work) &idev->mc_lock &c->lock irq_context: 0 (wq_completion)mld (work_completion)(&(&idev->mc_dad_work)->work) &idev->mc_lock &____s->seqcount irq_context: 0 rtnl_mutex rcu_state.exp_mutex fill_pool_map-wait-type-override pool_lock#2 irq_context: 0 rtnl_mutex rcu_state.exp_mutex fill_pool_map-wait-type-override &c->lock irq_context: 0 rtnl_mutex rcu_state.exp_mutex fill_pool_map-wait-type-override &____s->seqcount irq_context: 0 rtnl_mutex rcu_state.exp_mutex fill_pool_map-wait-type-override pool_lock irq_context: hardirq log_wait.lock &p->pi_lock &rq->__lock &cfs_rq->removed.lock irq_context: 0 rtnl_mutex fill_pool_map-wait-type-override rcu_read_lock pool_lock#2 irq_context: 0 rtnl_mutex fill_pool_map-wait-type-override &obj_hash[i].lock irq_context: 0 rtnl_mutex fill_pool_map-wait-type-override &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 rtnl_mutex fill_pool_map-wait-type-override batched_entropy_u8.lock irq_context: 0 rtnl_mutex fill_pool_map-wait-type-override kfence_freelist_lock irq_context: 0 (wq_completion)bat_events (work_completion)(&(&bat_priv->nc.work)->work) fill_pool_map-wait-type-override &c->lock irq_context: 0 (wq_completion)bat_events (work_completion)(&(&bat_priv->nc.work)->work) fill_pool_map-wait-type-override &____s->seqcount irq_context: 0 (wq_completion)bond0#2 irq_context: 0 (wq_completion)bond0#2 (work_completion)(&(&slave->notify_work)->work) irq_context: 0 (wq_completion)bond0#2 (work_completion)(&(&slave->notify_work)->work) &obj_hash[i].lock irq_context: 0 (wq_completion)bond0#2 (work_completion)(&(&slave->notify_work)->work) &base->lock irq_context: 0 (wq_completion)bond0#2 (work_completion)(&(&slave->notify_work)->work) &base->lock &obj_hash[i].lock irq_context: 0 rtnl_mutex cpu_hotplug_lock wq_pool_mutex fill_pool_map-wait-type-override pool_lock#2 irq_context: 0 rtnl_mutex cpu_hotplug_lock wq_pool_mutex fill_pool_map-wait-type-override pool_lock irq_context: 0 (wq_completion)bond0#3 irq_context: 0 (wq_completion)bond0#3 (work_completion)(&(&slave->notify_work)->work) irq_context: 0 (wq_completion)bond0#3 (work_completion)(&(&slave->notify_work)->work) &obj_hash[i].lock irq_context: 0 (wq_completion)bond0#3 (work_completion)(&(&slave->notify_work)->work) &base->lock irq_context: 0 (wq_completion)bond0#3 (work_completion)(&(&slave->notify_work)->work) &base->lock &obj_hash[i].lock irq_context: 0 (wq_completion)mld (work_completion)(&(&idev->mc_ifc_work)->work) &idev->mc_lock &n->list_lock irq_context: 0 (wq_completion)mld (work_completion)(&(&idev->mc_ifc_work)->work) &idev->mc_lock &n->list_lock &c->lock irq_context: softirq &(&slave->notify_work)->timer rcu_read_lock &pool->lock/1 &p->pi_lock &rq->__lock &cfs_rq->removed.lock irq_context: 0 rtnl_mutex team->team_lock_key irq_context: 0 rtnl_mutex team->team_lock_key fs_reclaim irq_context: 0 rtnl_mutex team->team_lock_key fs_reclaim mmu_notifier_invalidate_range_start irq_context: 0 rtnl_mutex team->team_lock_key pool_lock#2 irq_context: 0 rtnl_mutex team->team_lock_key netpoll_srcu irq_context: 0 rtnl_mutex team->team_lock_key net_rwsem irq_context: 0 rtnl_mutex team->team_lock_key net_rwsem &list->lock#2 irq_context: 0 rtnl_mutex team->team_lock_key &tn->lock irq_context: 0 rtnl_mutex team->team_lock_key _xmit_ETHER irq_context: 0 rtnl_mutex team->team_lock_key &dir->lock#2 irq_context: 0 rtnl_mutex team->team_lock_key input_pool.lock irq_context: 0 rtnl_mutex team->team_lock_key rcu_read_lock &ndev->lock irq_context: 0 rtnl_mutex team->team_lock_key &obj_hash[i].lock irq_context: 0 rtnl_mutex team->team_lock_key nl_table_lock irq_context: 0 rtnl_mutex team->team_lock_key nl_table_wait.lock irq_context: 0 rtnl_mutex team->team_lock_key rcu_read_lock &pool->lock/1 irq_context: 0 rtnl_mutex team->team_lock_key rcu_read_lock &pool->lock/1 &obj_hash[i].lock irq_context: 0 rtnl_mutex team->team_lock_key rcu_read_lock &pool->lock/1 pool_lock#2 irq_context: 0 rtnl_mutex team->team_lock_key rcu_read_lock &pool->lock/1 &p->pi_lock irq_context: 0 rtnl_mutex team->team_lock_key &in_dev->mc_tomb_lock irq_context: 0 rtnl_mutex team->team_lock_key &im->lock irq_context: 0 rtnl_mutex team->team_lock_key _xmit_ETHER pool_lock#2 irq_context: 0 rtnl_mutex team->team_lock_key cbs_list_lock irq_context: 0 rtnl_mutex team->team_lock_key &ndev->lock irq_context: 0 rtnl_mutex team->team_lock_key sysfs_symlink_target_lock irq_context: 0 rtnl_mutex team->team_lock_key lock irq_context: 0 rtnl_mutex team->team_lock_key lock kernfs_idr_lock irq_context: 0 rtnl_mutex team->team_lock_key &root->kernfs_rwsem irq_context: 0 rtnl_mutex team->team_lock_key &root->kernfs_rwsem &root->kernfs_iattr_rwsem irq_context: 0 rtnl_mutex team->team_lock_key &c->lock irq_context: 0 rtnl_mutex team->team_lock_key &n->list_lock irq_context: 0 rtnl_mutex team->team_lock_key &n->list_lock &c->lock irq_context: 0 rtnl_mutex team->team_lock_key fill_pool_map-wait-type-override pool_lock#2 irq_context: 0 rtnl_mutex team->team_lock_key fill_pool_map-wait-type-override pool_lock irq_context: 0 rtnl_mutex team->team_lock_key &____s->seqcount irq_context: 0 rtnl_mutex team->team_lock_key rcu_read_lock pool_lock#2 irq_context: 0 rtnl_mutex team->team_lock_key lweventlist_lock irq_context: 0 rtnl_mutex team->team_lock_key lweventlist_lock pool_lock#2 irq_context: 0 rtnl_mutex team->team_lock_key lweventlist_lock &dir->lock#2 irq_context: 0 rtnl_mutex team->team_lock_key (console_sem).lock irq_context: 0 rtnl_mutex team->team_lock_key console_lock console_srcu console_owner_lock irq_context: 0 rtnl_mutex team->team_lock_key console_lock console_srcu console_owner irq_context: 0 rtnl_mutex team->team_lock_key console_lock console_srcu console_owner &port_lock_key irq_context: 0 rtnl_mutex team->team_lock_key console_lock console_srcu console_owner console_owner_lock irq_context: 0 rtnl_mutex team->team_lock_key rcu_read_lock &pool->lock/1 &p->pi_lock &rq->__lock irq_context: 0 rtnl_mutex team->team_lock_key rcu_read_lock &pool->lock/1 &p->pi_lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 rtnl_mutex cpu_hotplug_lock wq_pool_mutex &obj_hash[i].lock pool_lock irq_context: 0 (wq_completion)bond0#4 irq_context: 0 (wq_completion)bond0#4 (work_completion)(&(&slave->notify_work)->work) irq_context: 0 (wq_completion)bond0#4 (work_completion)(&(&slave->notify_work)->work) &obj_hash[i].lock irq_context: 0 (wq_completion)bond0#4 (work_completion)(&(&slave->notify_work)->work) &base->lock irq_context: 0 (wq_completion)bond0#4 (work_completion)(&(&slave->notify_work)->work) &base->lock &obj_hash[i].lock irq_context: 0 rtnl_mutex rcu_node_0 irq_context: 0 rtnl_mutex team->team_lock_key#2 irq_context: 0 rtnl_mutex team->team_lock_key#2 fs_reclaim irq_context: 0 rtnl_mutex team->team_lock_key#2 fs_reclaim mmu_notifier_invalidate_range_start irq_context: 0 rtnl_mutex team->team_lock_key#2 netpoll_srcu irq_context: 0 rtnl_mutex team->team_lock_key#2 net_rwsem irq_context: 0 rtnl_mutex team->team_lock_key#2 net_rwsem &list->lock#2 irq_context: 0 rtnl_mutex team->team_lock_key#2 &tn->lock irq_context: 0 rtnl_mutex team->team_lock_key#2 _xmit_ETHER irq_context: 0 rtnl_mutex team->team_lock_key#2 &dir->lock#2 irq_context: 0 rtnl_mutex team->team_lock_key#2 input_pool.lock irq_context: 0 rtnl_mutex team->team_lock_key#2 rcu_read_lock &ndev->lock irq_context: 0 rtnl_mutex team->team_lock_key#2 &obj_hash[i].lock irq_context: 0 rtnl_mutex team->team_lock_key#2 nl_table_lock irq_context: 0 rtnl_mutex team->team_lock_key#2 nl_table_wait.lock irq_context: 0 rtnl_mutex team->team_lock_key#2 &c->lock irq_context: 0 rtnl_mutex team->team_lock_key#2 rcu_read_lock &pool->lock/1 irq_context: 0 rtnl_mutex team->team_lock_key#2 rcu_read_lock &pool->lock/1 &obj_hash[i].lock irq_context: 0 rtnl_mutex team->team_lock_key#2 rcu_read_lock &pool->lock/1 &p->pi_lock irq_context: 0 rtnl_mutex team->team_lock_key#2 &in_dev->mc_tomb_lock irq_context: 0 rtnl_mutex team->team_lock_key#2 &im->lock irq_context: 0 rtnl_mutex team->team_lock_key#2 cbs_list_lock irq_context: 0 rtnl_mutex team->team_lock_key#2 &ndev->lock irq_context: 0 rtnl_mutex team->team_lock_key#2 &rq->__lock irq_context: 0 rtnl_mutex team->team_lock_key#2 sysfs_symlink_target_lock irq_context: 0 rtnl_mutex team->team_lock_key#2 lock irq_context: 0 rtnl_mutex team->team_lock_key#2 lock kernfs_idr_lock irq_context: 0 rtnl_mutex team->team_lock_key#2 &root->kernfs_rwsem irq_context: 0 rtnl_mutex team->team_lock_key#2 &root->kernfs_rwsem &root->kernfs_iattr_rwsem irq_context: 0 rtnl_mutex team->team_lock_key#2 rcu_read_lock &pool->lock/1 &p->pi_lock &rq->__lock irq_context: 0 rtnl_mutex team->team_lock_key#2 rcu_read_lock &pool->lock/1 &p->pi_lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 rtnl_mutex team->team_lock_key#2 &pcp->lock &zone->lock irq_context: 0 rtnl_mutex team->team_lock_key#2 &pcp->lock &zone->lock &____s->seqcount irq_context: 0 rtnl_mutex team->team_lock_key#2 &____s->seqcount irq_context: 0 rtnl_mutex team->team_lock_key#2 quarantine_lock irq_context: 0 rtnl_mutex team->team_lock_key#2 remove_cache_srcu irq_context: 0 rtnl_mutex team->team_lock_key#2 remove_cache_srcu quarantine_lock irq_context: 0 rtnl_mutex team->team_lock_key#2 remove_cache_srcu &c->lock irq_context: 0 rtnl_mutex team->team_lock_key#2 remove_cache_srcu &n->list_lock irq_context: 0 rtnl_mutex team->team_lock_key#2 remove_cache_srcu rcu_read_lock pool_lock#2 irq_context: 0 rtnl_mutex team->team_lock_key#2 remove_cache_srcu &obj_hash[i].lock irq_context: 0 rtnl_mutex team->team_lock_key#2 lweventlist_lock irq_context: 0 rtnl_mutex team->team_lock_key#2 lweventlist_lock &dir->lock#2 irq_context: 0 rtnl_mutex team->team_lock_key#2 (console_sem).lock irq_context: 0 rtnl_mutex team->team_lock_key#2 console_lock console_srcu console_owner_lock irq_context: 0 rtnl_mutex team->team_lock_key#2 console_lock console_srcu console_owner irq_context: 0 rtnl_mutex team->team_lock_key#2 console_lock console_srcu console_owner &port_lock_key irq_context: 0 rtnl_mutex team->team_lock_key#2 console_lock console_srcu console_owner console_owner_lock irq_context: 0 syslog_lock &rq->__lock irq_context: 0 (wq_completion)bond0#2 (work_completion)(&(&slave->notify_work)->work) rtnl_mutex net_rwsem irq_context: 0 (wq_completion)bond0#2 (work_completion)(&(&slave->notify_work)->work) rtnl_mutex net_rwsem &list->lock#2 irq_context: 0 (wq_completion)bond0#2 (work_completion)(&(&slave->notify_work)->work) rtnl_mutex &tn->lock irq_context: 0 (wq_completion)bond0#2 (work_completion)(&(&slave->notify_work)->work) rtnl_mutex.wait_lock irq_context: 0 (wq_completion)bond0#2 (work_completion)(&(&slave->notify_work)->work) &p->pi_lock irq_context: 0 (wq_completion)bond0 (work_completion)(&(&slave->notify_work)->work) rtnl_mutex net_rwsem irq_context: 0 (wq_completion)bond0 (work_completion)(&(&slave->notify_work)->work) rtnl_mutex net_rwsem &list->lock#2 irq_context: 0 (wq_completion)bond0 (work_completion)(&(&slave->notify_work)->work) rtnl_mutex &tn->lock irq_context: 0 (wq_completion)bond0 (work_completion)(&(&slave->notify_work)->work) rtnl_mutex.wait_lock irq_context: 0 (wq_completion)bond0 (work_completion)(&(&slave->notify_work)->work) &p->pi_lock irq_context: 0 rtnl_mutex &idev->mc_lock &____s->seqcount irq_context: 0 rtnl_mutex rcu_state.exp_mutex &rcu_state.expedited_wq irq_context: 0 rtnl_mutex rcu_state.exp_mutex &rcu_state.expedited_wq &p->pi_lock irq_context: 0 (wq_completion)events (linkwatch_work).work rtnl_mutex rcu_state.exp_mutex &rnp->exp_wq[0] irq_context: 0 (wq_completion)events (linkwatch_work).work rtnl_mutex rcu_state.exp_mutex &rnp->exp_wq[2] irq_context: 0 (wq_completion)events (linkwatch_work).work rtnl_mutex &____s->seqcount irq_context: 0 (wq_completion)events (linkwatch_work).work rtnl_mutex team->team_lock_key irq_context: 0 (wq_completion)events (linkwatch_work).work rtnl_mutex team->team_lock_key fs_reclaim irq_context: 0 (wq_completion)events (linkwatch_work).work rtnl_mutex team->team_lock_key fs_reclaim mmu_notifier_invalidate_range_start irq_context: 0 (wq_completion)events (linkwatch_work).work rtnl_mutex team->team_lock_key pool_lock#2 irq_context: 0 (wq_completion)events (linkwatch_work).work rtnl_mutex team->team_lock_key rcu_read_lock &ndev->lock irq_context: 0 (wq_completion)events (linkwatch_work).work rtnl_mutex team->team_lock_key &obj_hash[i].lock irq_context: 0 (wq_completion)events (linkwatch_work).work rtnl_mutex team->team_lock_key nl_table_lock irq_context: 0 (wq_completion)events (linkwatch_work).work rtnl_mutex team->team_lock_key nl_table_wait.lock irq_context: 0 (wq_completion)events (linkwatch_work).work rtnl_mutex team->team_lock_key net_rwsem irq_context: 0 (wq_completion)events (linkwatch_work).work rtnl_mutex team->team_lock_key net_rwsem &list->lock#2 irq_context: 0 (wq_completion)events (linkwatch_work).work rtnl_mutex team->team_lock_key &tn->lock irq_context: 0 (wq_completion)events (linkwatch_work).work rtnl_mutex team->team_lock_key &c->lock irq_context: 0 (wq_completion)events (linkwatch_work).work rtnl_mutex team->team_lock_key &____s->seqcount irq_context: 0 (wq_completion)events (linkwatch_work).work rtnl_mutex team->team_lock_key#2 irq_context: 0 (wq_completion)events (linkwatch_work).work rtnl_mutex team->team_lock_key#2 fs_reclaim irq_context: 0 (wq_completion)events (linkwatch_work).work rtnl_mutex team->team_lock_key#2 fs_reclaim mmu_notifier_invalidate_range_start irq_context: 0 (wq_completion)events (linkwatch_work).work rtnl_mutex team->team_lock_key#2 &c->lock irq_context: 0 (wq_completion)events (linkwatch_work).work rtnl_mutex team->team_lock_key#2 rcu_read_lock &ndev->lock irq_context: 0 (wq_completion)events (linkwatch_work).work rtnl_mutex team->team_lock_key#2 &obj_hash[i].lock irq_context: 0 (wq_completion)events (linkwatch_work).work rtnl_mutex team->team_lock_key#2 nl_table_lock irq_context: 0 (wq_completion)events (linkwatch_work).work rtnl_mutex team->team_lock_key#2 nl_table_wait.lock irq_context: 0 (wq_completion)events (linkwatch_work).work rtnl_mutex team->team_lock_key#2 net_rwsem irq_context: 0 (wq_completion)events (linkwatch_work).work rtnl_mutex team->team_lock_key#2 net_rwsem &list->lock#2 irq_context: 0 (wq_completion)events (linkwatch_work).work rtnl_mutex team->team_lock_key#2 &tn->lock irq_context: 0 (wq_completion)events (linkwatch_work).work &rq->__lock irq_context: softirq &(&slave->notify_work)->timer rcu_read_lock &pool->lock/1 fill_pool_map-wait-type-override pool_lock#2 irq_context: softirq &(&slave->notify_work)->timer rcu_read_lock &pool->lock/1 fill_pool_map-wait-type-override &c->lock irq_context: softirq &(&slave->notify_work)->timer rcu_read_lock &pool->lock/1 fill_pool_map-wait-type-override &____s->seqcount irq_context: softirq &(&slave->notify_work)->timer rcu_read_lock &pool->lock/1 fill_pool_map-wait-type-override pool_lock irq_context: 0 rtnl_mutex &hard_iface->bat_iv.ogm_buff_mutex irq_context: 0 rtnl_mutex &hard_iface->bat_iv.ogm_buff_mutex crngs.lock irq_context: 0 rtnl_mutex &hard_iface->bat_iv.ogm_buff_mutex pool_lock#2 irq_context: 0 rtnl_mutex ptype_lock irq_context: softirq &(&bat_priv->orig_work)->timer irq_context: softirq &(&bat_priv->orig_work)->timer rcu_read_lock &pool->lock/1 irq_context: softirq &(&bat_priv->orig_work)->timer rcu_read_lock &pool->lock/1 &obj_hash[i].lock irq_context: softirq &(&bat_priv->orig_work)->timer rcu_read_lock &pool->lock/1 &p->pi_lock irq_context: softirq &(&bat_priv->orig_work)->timer rcu_read_lock &pool->lock/1 &p->pi_lock &rq->__lock irq_context: softirq &(&bat_priv->orig_work)->timer rcu_read_lock &pool->lock/1 &p->pi_lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)bat_events (work_completion)(&(&bat_priv->orig_work)->work) irq_context: 0 (wq_completion)bat_events (work_completion)(&(&bat_priv->orig_work)->work) key#19 irq_context: 0 (wq_completion)bat_events (work_completion)(&(&bat_priv->orig_work)->work) &obj_hash[i].lock irq_context: 0 (wq_completion)bat_events (work_completion)(&(&bat_priv->orig_work)->work) &base->lock irq_context: 0 (wq_completion)bat_events (work_completion)(&(&bat_priv->orig_work)->work) &base->lock &obj_hash[i].lock irq_context: 0 rtnl_mutex &hard_iface->bat_iv.ogm_buff_mutex batched_entropy_u8.lock irq_context: 0 rtnl_mutex &hard_iface->bat_iv.ogm_buff_mutex &bat_priv->forw_bat_list_lock irq_context: 0 rtnl_mutex &hard_iface->bat_iv.ogm_buff_mutex &c->lock irq_context: 0 rtnl_mutex &hard_iface->bat_iv.ogm_buff_mutex &obj_hash[i].lock irq_context: 0 rtnl_mutex &hard_iface->bat_iv.ogm_buff_mutex &bat_priv->forw_bat_list_lock &obj_hash[i].lock irq_context: 0 rtnl_mutex &hard_iface->bat_iv.ogm_buff_mutex &bat_priv->forw_bat_list_lock &base->lock irq_context: 0 rtnl_mutex &hard_iface->bat_iv.ogm_buff_mutex &bat_priv->forw_bat_list_lock &base->lock &obj_hash[i].lock irq_context: 0 (wq_completion)bat_events (work_completion)(&(&bat_priv->mcast.work)->work) pool_lock#2 irq_context: 0 rtnl_mutex &br->lock &br->hash_lock &n->list_lock irq_context: 0 rtnl_mutex &br->lock &br->hash_lock &n->list_lock &c->lock irq_context: 0 rtnl_mutex &br->lock &obj_hash[i].lock irq_context: 0 rtnl_mutex &br->lock &base->lock irq_context: 0 rtnl_mutex &br->lock &base->lock &obj_hash[i].lock irq_context: 0 rtnl_mutex team->team_lock_key#3 irq_context: 0 rtnl_mutex team->team_lock_key#3 fs_reclaim irq_context: 0 rtnl_mutex team->team_lock_key#3 fs_reclaim mmu_notifier_invalidate_range_start irq_context: 0 rtnl_mutex team->team_lock_key#3 netpoll_srcu irq_context: 0 rtnl_mutex team->team_lock_key#3 net_rwsem irq_context: 0 rtnl_mutex team->team_lock_key#3 net_rwsem &list->lock#2 irq_context: 0 rtnl_mutex team->team_lock_key#3 &tn->lock irq_context: 0 rtnl_mutex team->team_lock_key#3 _xmit_ETHER irq_context: 0 rtnl_mutex team->team_lock_key#3 &dir->lock#2 irq_context: 0 rtnl_mutex team->team_lock_key#3 input_pool.lock irq_context: 0 rtnl_mutex team->team_lock_key#3 rcu_read_lock &ndev->lock irq_context: 0 rtnl_mutex team->team_lock_key#3 &obj_hash[i].lock irq_context: 0 rtnl_mutex team->team_lock_key#3 nl_table_lock irq_context: 0 rtnl_mutex team->team_lock_key#3 nl_table_wait.lock irq_context: 0 rtnl_mutex team->team_lock_key#3 &c->lock irq_context: 0 rtnl_mutex team->team_lock_key#3 rcu_read_lock &pool->lock/1 irq_context: 0 rtnl_mutex team->team_lock_key#3 rcu_read_lock &pool->lock/1 &obj_hash[i].lock irq_context: 0 rtnl_mutex team->team_lock_key#3 rcu_read_lock &pool->lock/1 &p->pi_lock irq_context: 0 rtnl_mutex team->team_lock_key#3 rcu_read_lock &pool->lock/1 &p->pi_lock &rq->__lock irq_context: 0 rtnl_mutex team->team_lock_key#3 rcu_read_lock &pool->lock/1 &p->pi_lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 rtnl_mutex team->team_lock_key#3 &in_dev->mc_tomb_lock irq_context: 0 rtnl_mutex team->team_lock_key#3 &im->lock irq_context: 0 rtnl_mutex team->team_lock_key#3 cbs_list_lock irq_context: 0 rtnl_mutex team->team_lock_key#3 &ndev->lock irq_context: 0 rtnl_mutex team->team_lock_key#3 sysfs_symlink_target_lock irq_context: 0 rtnl_mutex team->team_lock_key#3 lock irq_context: 0 rtnl_mutex team->team_lock_key#3 lock kernfs_idr_lock irq_context: 0 rtnl_mutex team->team_lock_key#3 &root->kernfs_rwsem irq_context: 0 rtnl_mutex team->team_lock_key#3 &root->kernfs_rwsem &root->kernfs_iattr_rwsem irq_context: 0 rtnl_mutex team->team_lock_key#3 &rq->__lock irq_context: 0 rtnl_mutex team->team_lock_key#3 quarantine_lock irq_context: 0 rtnl_mutex team->team_lock_key#3 remove_cache_srcu irq_context: 0 rtnl_mutex team->team_lock_key#3 remove_cache_srcu quarantine_lock irq_context: 0 rtnl_mutex team->team_lock_key#3 remove_cache_srcu &c->lock irq_context: 0 rtnl_mutex team->team_lock_key#3 remove_cache_srcu &n->list_lock irq_context: 0 rtnl_mutex team->team_lock_key#3 remove_cache_srcu rcu_read_lock pool_lock#2 irq_context: 0 rtnl_mutex team->team_lock_key#3 remove_cache_srcu &obj_hash[i].lock irq_context: 0 rtnl_mutex team->team_lock_key#3 &____s->seqcount irq_context: 0 rtnl_mutex team->team_lock_key#3 (console_sem).lock irq_context: 0 rtnl_mutex team->team_lock_key#3 console_lock console_srcu console_owner_lock irq_context: 0 rtnl_mutex team->team_lock_key#3 console_lock console_srcu console_owner irq_context: 0 rtnl_mutex team->team_lock_key#3 console_lock console_srcu console_owner &port_lock_key irq_context: 0 rtnl_mutex team->team_lock_key#3 console_lock console_srcu console_owner console_owner_lock irq_context: 0 rtnl_mutex team->team_lock_key#3 &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 rtnl_mutex &idev->mc_lock _xmit_ETHER &n->list_lock irq_context: 0 rtnl_mutex &idev->mc_lock _xmit_ETHER &n->list_lock &c->lock irq_context: 0 rtnl_mutex rcu_state.exp_mutex &obj_hash[i].lock pool_lock irq_context: 0 rtnl_mutex rcu_state.exp_mutex rcu_read_lock &pool->lock fill_pool_map-wait-type-override pool_lock#2 irq_context: 0 rtnl_mutex rcu_state.exp_mutex rcu_read_lock &pool->lock fill_pool_map-wait-type-override pool_lock irq_context: 0 rtnl_mutex _xmit_NONE irq_context: 0 rtnl_mutex &tb->tb6_lock rcu_read_lock pool_lock#2 irq_context: 0 rtnl_mutex lock#9 irq_context: softirq &(&bat_priv->orig_work)->timer rcu_read_lock &pool->lock/1 &p->pi_lock &cfs_rq->removed.lock irq_context: softirq drivers/net/wireguard/ratelimiter.c:20 irq_context: softirq drivers/net/wireguard/ratelimiter.c:20 rcu_read_lock &pool->lock irq_context: softirq drivers/net/wireguard/ratelimiter.c:20 rcu_read_lock &pool->lock &obj_hash[i].lock irq_context: softirq drivers/net/wireguard/ratelimiter.c:20 rcu_read_lock &pool->lock &p->pi_lock irq_context: 0 (wq_completion)events_power_efficient (gc_work).work irq_context: 0 (wq_completion)events_power_efficient (gc_work).work tk_core.seq.seqcount irq_context: 0 (wq_completion)events_power_efficient (gc_work).work "ratelimiter_table_lock" irq_context: 0 (wq_completion)events_power_efficient (gc_work).work &rq->__lock irq_context: 0 (wq_completion)events_power_efficient (gc_work).work &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)events_power_efficient (gc_work).work &obj_hash[i].lock irq_context: 0 (wq_completion)events_power_efficient (gc_work).work &base->lock irq_context: 0 (wq_completion)events_power_efficient (gc_work).work &base->lock &obj_hash[i].lock irq_context: 0 rtnl_mutex team->team_lock_key#3 &pcp->lock &zone->lock irq_context: 0 rtnl_mutex team->team_lock_key#3 &pcp->lock &zone->lock &____s->seqcount irq_context: 0 rtnl_mutex team->team_lock_key#3 pool_lock#2 irq_context: 0 rtnl_mutex team->team_lock_key#4 irq_context: 0 rtnl_mutex team->team_lock_key#4 fs_reclaim irq_context: 0 rtnl_mutex team->team_lock_key#4 fs_reclaim mmu_notifier_invalidate_range_start irq_context: 0 rtnl_mutex team->team_lock_key#4 netpoll_srcu irq_context: 0 rtnl_mutex team->team_lock_key#4 net_rwsem irq_context: 0 rtnl_mutex team->team_lock_key#4 net_rwsem &list->lock#2 irq_context: 0 rtnl_mutex team->team_lock_key#4 &tn->lock irq_context: 0 rtnl_mutex team->team_lock_key#4 _xmit_ETHER irq_context: 0 rtnl_mutex team->team_lock_key#4 &dir->lock#2 irq_context: 0 rtnl_mutex team->team_lock_key#4 input_pool.lock irq_context: 0 rtnl_mutex team->team_lock_key#4 rcu_read_lock &ndev->lock irq_context: 0 rtnl_mutex team->team_lock_key#4 &obj_hash[i].lock irq_context: 0 rtnl_mutex team->team_lock_key#4 nl_table_lock irq_context: 0 rtnl_mutex team->team_lock_key#4 nl_table_wait.lock irq_context: 0 rtnl_mutex team->team_lock_key#4 rcu_read_lock &pool->lock/1 irq_context: 0 rtnl_mutex team->team_lock_key#4 rcu_read_lock &pool->lock/1 &obj_hash[i].lock irq_context: 0 rtnl_mutex team->team_lock_key#4 rcu_read_lock &pool->lock/1 &p->pi_lock irq_context: 0 rtnl_mutex team->team_lock_key#4 &in_dev->mc_tomb_lock irq_context: 0 rtnl_mutex team->team_lock_key#4 &im->lock irq_context: 0 rtnl_mutex team->team_lock_key#4 cbs_list_lock irq_context: 0 rtnl_mutex team->team_lock_key#4 &rq->__lock irq_context: 0 rtnl_mutex team->team_lock_key#4 &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 rtnl_mutex team->team_lock_key#4 &ndev->lock irq_context: 0 rtnl_mutex team->team_lock_key#4 &c->lock irq_context: 0 rtnl_mutex team->team_lock_key#4 sysfs_symlink_target_lock irq_context: 0 rtnl_mutex team->team_lock_key#4 lock irq_context: 0 rtnl_mutex team->team_lock_key#4 lock kernfs_idr_lock irq_context: 0 rtnl_mutex team->team_lock_key#4 &root->kernfs_rwsem irq_context: 0 rtnl_mutex team->team_lock_key#4 &root->kernfs_rwsem &root->kernfs_iattr_rwsem irq_context: 0 rtnl_mutex team->team_lock_key#4 rcu_read_lock &pool->lock/1 &p->pi_lock &rq->__lock irq_context: 0 rtnl_mutex team->team_lock_key#4 rcu_read_lock &pool->lock/1 &p->pi_lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 rtnl_mutex team->team_lock_key#4 &____s->seqcount irq_context: 0 rtnl_mutex team->team_lock_key#4 lweventlist_lock irq_context: 0 rtnl_mutex team->team_lock_key#4 lweventlist_lock &dir->lock#2 irq_context: 0 rtnl_mutex team->team_lock_key#4 (console_sem).lock irq_context: 0 rtnl_mutex team->team_lock_key#4 console_lock console_srcu console_owner_lock irq_context: 0 rtnl_mutex team->team_lock_key#4 console_lock console_srcu console_owner irq_context: 0 rtnl_mutex team->team_lock_key#4 console_lock console_srcu console_owner &port_lock_key irq_context: 0 rtnl_mutex team->team_lock_key#4 console_lock console_srcu console_owner console_owner_lock irq_context: 0 rtnl_mutex team->team_lock_key#4 &pcp->lock &zone->lock irq_context: 0 rtnl_mutex team->team_lock_key#4 &pcp->lock &zone->lock &____s->seqcount irq_context: 0 rtnl_mutex team->team_lock_key#4 &n->list_lock irq_context: 0 rtnl_mutex team->team_lock_key#4 &n->list_lock &c->lock irq_context: 0 (wq_completion)bond0#5 irq_context: 0 (wq_completion)bond0#5 (work_completion)(&(&slave->notify_work)->work) irq_context: 0 (wq_completion)bond0#5 (work_completion)(&(&slave->notify_work)->work) &obj_hash[i].lock irq_context: 0 (wq_completion)bond0#5 (work_completion)(&(&slave->notify_work)->work) &base->lock irq_context: 0 (wq_completion)bond0#5 (work_completion)(&(&slave->notify_work)->work) &base->lock &obj_hash[i].lock irq_context: 0 (wq_completion)bond0#4 (work_completion)(&(&slave->notify_work)->work) rtnl_mutex net_rwsem irq_context: 0 (wq_completion)bond0#4 (work_completion)(&(&slave->notify_work)->work) rtnl_mutex net_rwsem &list->lock#2 irq_context: 0 (wq_completion)bond0#4 (work_completion)(&(&slave->notify_work)->work) rtnl_mutex &tn->lock irq_context: 0 (wq_completion)bond0#4 (work_completion)(&(&slave->notify_work)->work) rtnl_mutex.wait_lock irq_context: 0 (wq_completion)bond0#4 (work_completion)(&(&slave->notify_work)->work) &p->pi_lock irq_context: 0 (wq_completion)bond0#3 (work_completion)(&(&slave->notify_work)->work) rtnl_mutex net_rwsem irq_context: 0 (wq_completion)bond0#3 (work_completion)(&(&slave->notify_work)->work) rtnl_mutex net_rwsem &list->lock#2 irq_context: 0 (wq_completion)bond0#3 (work_completion)(&(&slave->notify_work)->work) rtnl_mutex &tn->lock irq_context: 0 (wq_completion)bond0#3 (work_completion)(&(&slave->notify_work)->work) rtnl_mutex.wait_lock irq_context: 0 (wq_completion)bond0#3 (work_completion)(&(&slave->notify_work)->work) &p->pi_lock irq_context: 0 (wq_completion)ipv6_addrconf (work_completion)(&(&ifa->dad_work)->work) rtnl_mutex rcu_read_lock &ndev->lock irq_context: 0 (wq_completion)ipv6_addrconf (work_completion)(&(&ifa->dad_work)->work) rtnl_mutex rcu_read_lock rcu_read_lock rcu_read_lock_bh rcu_read_lock &____s->seqcount#12 irq_context: 0 (wq_completion)ipv6_addrconf (work_completion)(&(&ifa->dad_work)->work) rtnl_mutex rcu_read_lock rcu_read_lock rcu_read_lock_bh rcu_read_lock &ndev->lock irq_context: 0 (wq_completion)ipv6_addrconf (work_completion)(&(&ifa->dad_work)->work) rtnl_mutex &ndev->lock &ifa->lock batched_entropy_u32.lock irq_context: 0 (wq_completion)bat_events (work_completion)(&(&bat_priv->orig_work)->work) &rq->__lock irq_context: 0 (wq_completion)bat_events (work_completion)(&(&bat_priv->orig_work)->work) &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)mm_percpu_wq (work_completion)(work) lock#4 &lruvec->lru_lock per_cpu_ptr(&cgroup_rstat_cpu_lock, cpu) irq_context: softirq &(&hdev->cmd_timer)->timer irq_context: softirq &(&hdev->cmd_timer)->timer rcu_read_lock &pool->lock/1 irq_context: softirq &(&hdev->cmd_timer)->timer rcu_read_lock &pool->lock/1 &obj_hash[i].lock irq_context: softirq &(&hdev->cmd_timer)->timer rcu_read_lock &pool->lock/1 &p->pi_lock irq_context: softirq &(&hdev->cmd_timer)->timer rcu_read_lock &pool->lock/1 &p->pi_lock &rq->__lock irq_context: softirq &(&hdev->cmd_timer)->timer rcu_read_lock &pool->lock/1 &p->pi_lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)hci0#2 (work_completion)(&(&hdev->cmd_timer)->work) irq_context: 0 (wq_completion)hci0#2 (work_completion)(&(&hdev->cmd_timer)->work) (console_sem).lock irq_context: 0 (wq_completion)hci0#2 (work_completion)(&(&hdev->cmd_timer)->work) console_lock console_srcu console_owner_lock irq_context: 0 (wq_completion)hci0#2 (work_completion)(&(&hdev->cmd_timer)->work) console_lock console_srcu console_owner irq_context: 0 (wq_completion)hci0#2 (work_completion)(&(&hdev->cmd_timer)->work) console_lock console_srcu console_owner &port_lock_key irq_context: 0 (wq_completion)hci0#2 (work_completion)(&(&hdev->cmd_timer)->work) console_lock console_srcu console_owner console_owner_lock irq_context: 0 (wq_completion)hci0#2 (work_completion)(&(&hdev->cmd_timer)->work) rcu_read_lock &pool->lock/1 irq_context: 0 (wq_completion)hci0#2 (work_completion)(&(&hdev->cmd_timer)->work) rcu_read_lock &pool->lock/1 &obj_hash[i].lock irq_context: 0 (wq_completion)hci0#2 (work_completion)(&(&hdev->cmd_timer)->work) rcu_read_lock &pool->lock/1 &p->pi_lock irq_context: 0 (wq_completion)hci0#2 (work_completion)(&(&hdev->cmd_timer)->work) rcu_read_lock &pool->lock/1 &p->pi_lock &rq->__lock irq_context: 0 (wq_completion)hci0#2 (work_completion)(&(&hdev->cmd_timer)->work) rcu_read_lock &pool->lock/1 &p->pi_lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)hci0#2 (work_completion)(&hdev->cmd_work) batched_entropy_u8.lock irq_context: 0 (wq_completion)hci0#2 (work_completion)(&hdev->cmd_work) kfence_freelist_lock irq_context: 0 rtnl_mutex &hsr->list_lock irq_context: 0 rtnl_mutex rcu_read_lock &pool->lock/1 fill_pool_map-wait-type-override &c->lock irq_context: 0 rtnl_mutex rcu_read_lock &pool->lock/1 fill_pool_map-wait-type-override &____s->seqcount irq_context: 0 rtnl_mutex pin_fs_lock irq_context: 0 rtnl_mutex &sb->s_type->i_mutex_key#3 irq_context: 0 rtnl_mutex &sb->s_type->i_mutex_key#3 &sb->s_type->i_lock_key#7 irq_context: 0 rtnl_mutex &sb->s_type->i_mutex_key#3 rename_lock.seqcount irq_context: 0 rtnl_mutex &sb->s_type->i_mutex_key#3 fs_reclaim irq_context: 0 rtnl_mutex &sb->s_type->i_mutex_key#3 fs_reclaim mmu_notifier_invalidate_range_start irq_context: 0 rtnl_mutex &sb->s_type->i_mutex_key#3 &rq->__lock irq_context: 0 rtnl_mutex &sb->s_type->i_mutex_key#3 pool_lock#2 irq_context: 0 rtnl_mutex &sb->s_type->i_mutex_key#3 &dentry->d_lock irq_context: 0 rtnl_mutex &sb->s_type->i_mutex_key#3 rcu_read_lock rename_lock.seqcount irq_context: 0 rtnl_mutex &sb->s_type->i_mutex_key#3 &dentry->d_lock &wq irq_context: 0 rtnl_mutex &sb->s_type->i_mutex_key#3 mmu_notifier_invalidate_range_start irq_context: 0 rtnl_mutex &sb->s_type->i_mutex_key#3 &s->s_inode_list_lock irq_context: 0 rtnl_mutex &sb->s_type->i_mutex_key#3 tk_core.seq.seqcount irq_context: 0 rtnl_mutex &sb->s_type->i_mutex_key#3 &sb->s_type->i_lock_key#7 &dentry->d_lock irq_context: 0 (wq_completion)bond0#6 irq_context: 0 (wq_completion)bond0#6 (work_completion)(&(&slave->notify_work)->work) irq_context: 0 (wq_completion)bond0#6 (work_completion)(&(&slave->notify_work)->work) &obj_hash[i].lock irq_context: 0 (wq_completion)bond0#6 (work_completion)(&(&slave->notify_work)->work) &base->lock irq_context: 0 (wq_completion)bond0#6 (work_completion)(&(&slave->notify_work)->work) &base->lock &obj_hash[i].lock irq_context: 0 (wq_completion)hci4#2 (work_completion)(&(&hdev->cmd_timer)->work) irq_context: 0 (wq_completion)hci4#2 (work_completion)(&(&hdev->cmd_timer)->work) (console_sem).lock irq_context: 0 (wq_completion)hci4#2 (work_completion)(&(&hdev->cmd_timer)->work) console_owner_lock irq_context: 0 (wq_completion)hci4#2 (work_completion)(&(&hdev->cmd_timer)->work) console_owner irq_context: 0 (wq_completion)hci4#2 (work_completion)(&(&hdev->cmd_timer)->work) console_lock console_srcu console_owner_lock irq_context: 0 (wq_completion)hci4#2 (work_completion)(&(&hdev->cmd_timer)->work) console_lock console_srcu console_owner irq_context: 0 (wq_completion)hci4#2 (work_completion)(&(&hdev->cmd_timer)->work) console_lock console_srcu console_owner &port_lock_key irq_context: 0 rtnl_mutex console_owner_lock irq_context: 0 rtnl_mutex console_owner irq_context: 0 (wq_completion)hci4#2 (work_completion)(&(&hdev->cmd_timer)->work) console_lock console_srcu console_owner console_owner_lock irq_context: 0 (wq_completion)hci4#2 (work_completion)(&(&hdev->cmd_timer)->work) &rq->__lock irq_context: 0 (wq_completion)hci2#2 (work_completion)(&(&hdev->cmd_timer)->work) irq_context: 0 (wq_completion)hci2#2 (work_completion)(&(&hdev->cmd_timer)->work) (console_sem).lock irq_context: 0 (wq_completion)hci2#2 (work_completion)(&(&hdev->cmd_timer)->work) console_owner_lock irq_context: 0 (wq_completion)hci2#2 (work_completion)(&(&hdev->cmd_timer)->work) console_owner irq_context: 0 (wq_completion)hci2#2 (work_completion)(&(&hdev->cmd_timer)->work) console_lock console_srcu console_owner_lock irq_context: 0 (wq_completion)hci2#2 (work_completion)(&(&hdev->cmd_timer)->work) console_lock console_srcu console_owner irq_context: 0 (wq_completion)hci2#2 (work_completion)(&(&hdev->cmd_timer)->work) console_lock console_srcu console_owner &port_lock_key irq_context: 0 (wq_completion)hci2#2 (work_completion)(&(&hdev->cmd_timer)->work) console_lock console_srcu console_owner console_owner_lock irq_context: 0 (wq_completion)hci2#2 (work_completion)(&(&hdev->cmd_timer)->work) &rq->__lock irq_context: 0 (wq_completion)hci2#2 (work_completion)(&(&hdev->cmd_timer)->work) &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)hci1#2 (work_completion)(&(&hdev->cmd_timer)->work) irq_context: 0 (wq_completion)hci1#2 (work_completion)(&(&hdev->cmd_timer)->work) (console_sem).lock irq_context: 0 (wq_completion)hci1#2 (work_completion)(&(&hdev->cmd_timer)->work) console_owner_lock irq_context: 0 (wq_completion)hci1#2 (work_completion)(&(&hdev->cmd_timer)->work) console_owner irq_context: 0 (wq_completion)hci1#2 (work_completion)(&(&hdev->cmd_timer)->work) console_lock console_srcu console_owner_lock irq_context: 0 (wq_completion)hci1#2 (work_completion)(&(&hdev->cmd_timer)->work) console_lock console_srcu console_owner irq_context: 0 (wq_completion)hci1#2 (work_completion)(&(&hdev->cmd_timer)->work) console_lock console_srcu console_owner &port_lock_key irq_context: 0 (wq_completion)hci1#2 (work_completion)(&(&hdev->cmd_timer)->work) console_lock console_srcu console_owner console_owner_lock irq_context: 0 (wq_completion)hci1#2 (work_completion)(&(&hdev->cmd_timer)->work) &rq->__lock irq_context: 0 (wq_completion)hci1#2 (work_completion)(&(&hdev->cmd_timer)->work) &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)hci4#2 (work_completion)(&(&hdev->cmd_timer)->work) rcu_read_lock &pool->lock/1 irq_context: 0 (wq_completion)hci4#2 (work_completion)(&(&hdev->cmd_timer)->work) rcu_read_lock &pool->lock/1 &obj_hash[i].lock irq_context: 0 (wq_completion)hci4#2 (work_completion)(&(&hdev->cmd_timer)->work) rcu_read_lock &pool->lock/1 &p->pi_lock irq_context: 0 (wq_completion)hci4#2 (work_completion)(&(&hdev->cmd_timer)->work) rcu_read_lock &pool->lock/1 &p->pi_lock &rq->__lock irq_context: 0 (wq_completion)hci4#2 (work_completion)(&(&hdev->cmd_timer)->work) rcu_read_lock &pool->lock/1 &p->pi_lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)hci1#2 (work_completion)(&(&hdev->cmd_timer)->work) rcu_read_lock &pool->lock/1 irq_context: 0 (wq_completion)hci1#2 (work_completion)(&(&hdev->cmd_timer)->work) rcu_read_lock &pool->lock/1 &obj_hash[i].lock irq_context: 0 (wq_completion)hci1#2 (work_completion)(&(&hdev->cmd_timer)->work) rcu_read_lock &pool->lock/1 &p->pi_lock irq_context: 0 (wq_completion)hci1#2 (work_completion)(&(&hdev->cmd_timer)->work) rcu_read_lock &pool->lock/1 &p->pi_lock &rq->__lock irq_context: 0 (wq_completion)hci1#2 (work_completion)(&(&hdev->cmd_timer)->work) rcu_read_lock &pool->lock/1 &p->pi_lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)hci2#2 (work_completion)(&(&hdev->cmd_timer)->work) rcu_read_lock &pool->lock/1 irq_context: 0 (wq_completion)hci2#2 (work_completion)(&(&hdev->cmd_timer)->work) rcu_read_lock &pool->lock/1 &obj_hash[i].lock irq_context: 0 (wq_completion)hci2#2 (work_completion)(&(&hdev->cmd_timer)->work) rcu_read_lock &pool->lock/1 &p->pi_lock irq_context: 0 (wq_completion)hci2#2 (work_completion)(&(&hdev->cmd_timer)->work) rcu_read_lock &pool->lock/1 &p->pi_lock &rq->__lock irq_context: 0 (wq_completion)hci2#2 (work_completion)(&(&hdev->cmd_timer)->work) rcu_read_lock &pool->lock/1 &p->pi_lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)hci2#2 (work_completion)(&hdev->cmd_work) batched_entropy_u8.lock irq_context: 0 (wq_completion)hci2#2 (work_completion)(&hdev->cmd_work) kfence_freelist_lock irq_context: softirq &(&hdev->cmd_timer)->timer rcu_read_lock &pool->lock/1 &p->pi_lock &cfs_rq->removed.lock irq_context: 0 (wq_completion)hci3#2 (work_completion)(&(&hdev->cmd_timer)->work) irq_context: 0 (wq_completion)hci3#2 (work_completion)(&(&hdev->cmd_timer)->work) (console_sem).lock irq_context: 0 (wq_completion)hci3#2 (work_completion)(&(&hdev->cmd_timer)->work) console_owner_lock irq_context: 0 (wq_completion)hci3#2 (work_completion)(&(&hdev->cmd_timer)->work) console_owner irq_context: 0 (wq_completion)hci3#2 (work_completion)(&(&hdev->cmd_timer)->work) console_lock console_srcu console_owner_lock irq_context: 0 (wq_completion)hci3#2 (work_completion)(&(&hdev->cmd_timer)->work) console_lock console_srcu console_owner irq_context: 0 (wq_completion)hci3#2 (work_completion)(&(&hdev->cmd_timer)->work) console_lock console_srcu console_owner &port_lock_key irq_context: 0 (wq_completion)hci5#2 (work_completion)(&(&hdev->cmd_timer)->work) irq_context: 0 (wq_completion)hci5#2 (work_completion)(&(&hdev->cmd_timer)->work) (console_sem).lock irq_context: 0 (wq_completion)hci5#2 (work_completion)(&(&hdev->cmd_timer)->work) console_owner_lock irq_context: 0 (wq_completion)hci5#2 (work_completion)(&(&hdev->cmd_timer)->work) console_owner irq_context: 0 (wq_completion)hci3#2 (work_completion)(&(&hdev->cmd_timer)->work) console_lock console_srcu console_owner console_owner_lock irq_context: 0 (wq_completion)hci3#2 (work_completion)(&(&hdev->cmd_timer)->work) &rq->__lock irq_context: 0 (wq_completion)hci5#2 (work_completion)(&(&hdev->cmd_timer)->work) console_lock console_srcu console_owner_lock irq_context: 0 (wq_completion)hci5#2 (work_completion)(&(&hdev->cmd_timer)->work) console_lock console_srcu console_owner irq_context: 0 (wq_completion)hci5#2 (work_completion)(&(&hdev->cmd_timer)->work) console_lock console_srcu console_owner &port_lock_key irq_context: 0 (wq_completion)hci3#2 (work_completion)(&(&hdev->cmd_timer)->work) rcu_read_lock &pool->lock/1 irq_context: 0 (wq_completion)hci3#2 (work_completion)(&(&hdev->cmd_timer)->work) rcu_read_lock &pool->lock/1 &obj_hash[i].lock irq_context: 0 (wq_completion)hci3#2 (work_completion)(&(&hdev->cmd_timer)->work) rcu_read_lock &pool->lock/1 &p->pi_lock irq_context: 0 (wq_completion)hci3#2 (work_completion)(&(&hdev->cmd_timer)->work) rcu_read_lock &pool->lock/1 &p->pi_lock &rq->__lock irq_context: 0 (wq_completion)hci3#2 (work_completion)(&(&hdev->cmd_timer)->work) rcu_read_lock &pool->lock/1 &p->pi_lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)hci5#2 (work_completion)(&(&hdev->cmd_timer)->work) console_lock console_srcu console_owner console_owner_lock irq_context: 0 (wq_completion)hci5#2 (work_completion)(&(&hdev->cmd_timer)->work) &rq->__lock irq_context: 0 sk_lock-AF_INET rcu_read_lock rcu_read_lock &c->lock irq_context: 0 sk_lock-AF_INET rcu_read_lock rcu_read_lock &____s->seqcount irq_context: 0 (wq_completion)hci5#2 (work_completion)(&(&hdev->cmd_timer)->work) rcu_read_lock &pool->lock/1 irq_context: 0 (wq_completion)hci5#2 (work_completion)(&(&hdev->cmd_timer)->work) rcu_read_lock &pool->lock/1 &obj_hash[i].lock irq_context: 0 (wq_completion)hci5#2 (work_completion)(&(&hdev->cmd_timer)->work) rcu_read_lock &pool->lock/1 &p->pi_lock irq_context: 0 (wq_completion)hci5#2 (work_completion)(&(&hdev->cmd_timer)->work) rcu_read_lock &pool->lock/1 &p->pi_lock &rq->__lock irq_context: 0 (wq_completion)hci5#2 (work_completion)(&(&hdev->cmd_timer)->work) rcu_read_lock &pool->lock/1 &p->pi_lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 rtnl_mutex &hard_iface->bat_iv.ogm_buff_mutex rcu_read_lock rcu_node_0 irq_context: 0 rtnl_mutex &hard_iface->bat_iv.ogm_buff_mutex rcu_read_lock &rq->__lock irq_context: 0 rtnl_mutex proc_inum_ida.xa_lock pool_lock#2 irq_context: 0 (wq_completion)bat_events (work_completion)(&(&bat_priv->nc.work)->work) &base->lock fill_pool_map-wait-type-override pool_lock#2 irq_context: 0 (wq_completion)bat_events (work_completion)(&(&bat_priv->nc.work)->work) &base->lock fill_pool_map-wait-type-override pool_lock irq_context: 0 (wq_completion)rcu_gp (work_completion)(&rew->rew_work) rcu_state.exp_wake_mutex rcu_state.exp_wake_mutex.wait_lock irq_context: 0 (wq_completion)rcu_gp (work_completion)(&rew->rew_work) rcu_state.exp_wake_mutex &pool->lock irq_context: 0 (wq_completion)rcu_gp (work_completion)(&rew->rew_work) rcu_state.exp_wake_mutex &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)rcu_gp (work_completion)(&rew->rew_work) rcu_state.exp_wake_mutex.wait_lock irq_context: 0 (wq_completion)rcu_gp (work_completion)(&rew->rew_work) &p->pi_lock irq_context: 0 (wq_completion)bond0#6 (work_completion)(&(&slave->notify_work)->work) rtnl_mutex net_rwsem irq_context: 0 (wq_completion)bond0#6 (work_completion)(&(&slave->notify_work)->work) rtnl_mutex net_rwsem &list->lock#2 irq_context: 0 (wq_completion)bond0#6 (work_completion)(&(&slave->notify_work)->work) rtnl_mutex &tn->lock irq_context: 0 (wq_completion)bond0#6 (work_completion)(&(&slave->notify_work)->work) rtnl_mutex.wait_lock irq_context: 0 (wq_completion)bond0#6 (work_completion)(&(&slave->notify_work)->work) &p->pi_lock irq_context: 0 rtnl_mutex rcu_state.exp_mutex &rcu_state.expedited_wq &p->pi_lock &rq->__lock irq_context: softirq &(&bat_priv->mcast.work)->timer rcu_read_lock &pool->lock/1 &p->pi_lock &cfs_rq->removed.lock irq_context: 0 (wq_completion)bat_events (work_completion)(&(&bat_priv->mcast.work)->work) rcu_read_lock batched_entropy_u8.lock irq_context: 0 (wq_completion)bat_events (work_completion)(&(&bat_priv->mcast.work)->work) rcu_read_lock kfence_freelist_lock irq_context: 0 (wq_completion)bat_events (work_completion)(&(&bat_priv->mcast.work)->work) &meta->lock irq_context: 0 (wq_completion)bat_events (work_completion)(&(&bat_priv->mcast.work)->work) kfence_freelist_lock irq_context: 0 rtnl_mutex team->team_lock_key#5 irq_context: 0 rtnl_mutex team->team_lock_key#5 fs_reclaim irq_context: 0 rtnl_mutex team->team_lock_key#5 fs_reclaim mmu_notifier_invalidate_range_start irq_context: 0 rtnl_mutex team->team_lock_key#5 netpoll_srcu irq_context: 0 rtnl_mutex team->team_lock_key#5 net_rwsem irq_context: 0 rtnl_mutex team->team_lock_key#5 net_rwsem &list->lock#2 irq_context: 0 rtnl_mutex team->team_lock_key#5 &tn->lock irq_context: 0 rtnl_mutex team->team_lock_key#5 _xmit_ETHER irq_context: 0 rtnl_mutex team->team_lock_key#5 &dir->lock#2 irq_context: 0 rtnl_mutex team->team_lock_key#5 input_pool.lock irq_context: 0 rtnl_mutex team->team_lock_key#5 &c->lock irq_context: 0 rtnl_mutex team->team_lock_key#5 &____s->seqcount irq_context: 0 rtnl_mutex team->team_lock_key#5 rcu_read_lock &ndev->lock irq_context: 0 rtnl_mutex team->team_lock_key#5 &obj_hash[i].lock irq_context: 0 rtnl_mutex team->team_lock_key#5 nl_table_lock irq_context: 0 rtnl_mutex team->team_lock_key#5 nl_table_wait.lock irq_context: 0 rtnl_mutex team->team_lock_key#5 rcu_read_lock &pool->lock/1 irq_context: 0 rtnl_mutex team->team_lock_key#5 rcu_read_lock &pool->lock/1 &obj_hash[i].lock irq_context: 0 rtnl_mutex team->team_lock_key#5 rcu_read_lock &pool->lock/1 &p->pi_lock irq_context: 0 rtnl_mutex team->team_lock_key#5 rcu_read_lock &pool->lock/1 &p->pi_lock &rq->__lock irq_context: 0 rtnl_mutex team->team_lock_key#5 rcu_read_lock &pool->lock/1 &p->pi_lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 rtnl_mutex team->team_lock_key#5 &rq->__lock irq_context: 0 rtnl_mutex team->team_lock_key#5 &in_dev->mc_tomb_lock irq_context: 0 rtnl_mutex team->team_lock_key#5 &im->lock irq_context: 0 rtnl_mutex team->team_lock_key#5 cbs_list_lock irq_context: 0 rtnl_mutex team->team_lock_key#5 &ndev->lock irq_context: 0 rtnl_mutex team->team_lock_key#5 sysfs_symlink_target_lock irq_context: 0 rtnl_mutex team->team_lock_key#5 lock irq_context: 0 rtnl_mutex team->team_lock_key#5 lock kernfs_idr_lock irq_context: 0 rtnl_mutex team->team_lock_key#5 &root->kernfs_rwsem irq_context: 0 rtnl_mutex team->team_lock_key#5 &root->kernfs_rwsem &root->kernfs_iattr_rwsem irq_context: 0 rtnl_mutex team->team_lock_key#5 lweventlist_lock irq_context: 0 rtnl_mutex team->team_lock_key#5 lweventlist_lock &dir->lock#2 irq_context: 0 rtnl_mutex team->team_lock_key#5 (console_sem).lock irq_context: 0 rtnl_mutex team->team_lock_key#5 console_lock console_srcu console_owner_lock irq_context: 0 rtnl_mutex team->team_lock_key#5 console_lock console_srcu console_owner irq_context: 0 rtnl_mutex team->team_lock_key#5 console_lock console_srcu console_owner &port_lock_key irq_context: 0 rtnl_mutex team->team_lock_key#5 console_lock console_srcu console_owner console_owner_lock irq_context: 0 rtnl_mutex _xmit_ETHER &c->lock irq_context: 0 rtnl_mutex _xmit_ETHER &____s->seqcount irq_context: 0 rtnl_mutex team->team_lock_key#6 irq_context: 0 rtnl_mutex team->team_lock_key#6 fs_reclaim irq_context: 0 rtnl_mutex team->team_lock_key#6 fs_reclaim mmu_notifier_invalidate_range_start irq_context: 0 rtnl_mutex team->team_lock_key#6 netpoll_srcu irq_context: 0 rtnl_mutex team->team_lock_key#6 net_rwsem irq_context: 0 rtnl_mutex team->team_lock_key#6 net_rwsem &list->lock#2 irq_context: 0 rtnl_mutex team->team_lock_key#6 &tn->lock irq_context: 0 rtnl_mutex team->team_lock_key#6 _xmit_ETHER irq_context: 0 rtnl_mutex team->team_lock_key#6 &dir->lock#2 irq_context: 0 rtnl_mutex team->team_lock_key#6 input_pool.lock irq_context: 0 rtnl_mutex team->team_lock_key#6 rcu_read_lock &ndev->lock irq_context: 0 rtnl_mutex team->team_lock_key#6 &obj_hash[i].lock irq_context: 0 rtnl_mutex team->team_lock_key#6 nl_table_lock irq_context: 0 rtnl_mutex team->team_lock_key#6 nl_table_wait.lock irq_context: 0 rtnl_mutex team->team_lock_key#6 rcu_read_lock &pool->lock/1 irq_context: 0 rtnl_mutex team->team_lock_key#6 rcu_read_lock &pool->lock/1 &obj_hash[i].lock irq_context: 0 rtnl_mutex team->team_lock_key#6 rcu_read_lock &pool->lock/1 &p->pi_lock irq_context: 0 rtnl_mutex team->team_lock_key#6 &in_dev->mc_tomb_lock irq_context: 0 rtnl_mutex team->team_lock_key#6 &im->lock irq_context: 0 rtnl_mutex team->team_lock_key#6 cbs_list_lock irq_context: 0 rtnl_mutex team->team_lock_key#6 &ndev->lock irq_context: 0 rtnl_mutex team->team_lock_key#6 &c->lock irq_context: 0 rtnl_mutex team->team_lock_key#6 sysfs_symlink_target_lock irq_context: 0 rtnl_mutex team->team_lock_key#6 lock irq_context: 0 rtnl_mutex team->team_lock_key#6 lock kernfs_idr_lock irq_context: 0 rtnl_mutex team->team_lock_key#6 &root->kernfs_rwsem irq_context: 0 rtnl_mutex team->team_lock_key#6 &root->kernfs_rwsem &root->kernfs_iattr_rwsem irq_context: 0 rtnl_mutex team->team_lock_key#6 &n->list_lock irq_context: 0 rtnl_mutex team->team_lock_key#6 &n->list_lock &c->lock irq_context: 0 rtnl_mutex team->team_lock_key#6 rcu_read_lock &pool->lock/1 &p->pi_lock &cfs_rq->removed.lock irq_context: 0 rtnl_mutex team->team_lock_key#6 rcu_read_lock &pool->lock/1 &p->pi_lock &rq->__lock irq_context: 0 rtnl_mutex team->team_lock_key#6 rcu_read_lock &pool->lock/1 &p->pi_lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 rtnl_mutex team->team_lock_key#6 &rq->__lock irq_context: 0 rtnl_mutex team->team_lock_key#6 &____s->seqcount irq_context: 0 rtnl_mutex team->team_lock_key#6 lweventlist_lock irq_context: 0 rtnl_mutex team->team_lock_key#6 lweventlist_lock &dir->lock#2 irq_context: 0 rtnl_mutex team->team_lock_key#6 (console_sem).lock irq_context: 0 rtnl_mutex team->team_lock_key#6 console_lock console_srcu console_owner_lock irq_context: 0 rtnl_mutex team->team_lock_key#6 console_lock console_srcu console_owner irq_context: 0 rtnl_mutex team->team_lock_key#6 console_lock console_srcu console_owner &port_lock_key irq_context: 0 rtnl_mutex team->team_lock_key#6 console_lock console_srcu console_owner console_owner_lock irq_context: 0 (wq_completion)ipv6_addrconf (work_completion)(&(&ifa->dad_work)->work) rtnl_mutex rcu_read_lock rcu_read_lock &tbl->lock pool_lock#2 irq_context: 0 (wq_completion)ipv6_addrconf (work_completion)(&(&ifa->dad_work)->work) rtnl_mutex rcu_read_lock rcu_read_lock &tbl->lock batched_entropy_u32.lock irq_context: 0 (wq_completion)ipv6_addrconf (work_completion)(&(&ifa->dad_work)->work) rtnl_mutex rcu_read_lock rcu_read_lock &tbl->lock &obj_hash[i].lock irq_context: 0 (wq_completion)bat_events (work_completion)(&(&bat_priv->nc.work)->work) &rcu_state.expedited_wq &p->pi_lock &rq->__lock irq_context: 0 rtnl_mutex rcu_read_lock &net->sctp.local_addr_lock &net->sctp.addr_wq_lock &c->lock irq_context: 0 rtnl_mutex rcu_read_lock &net->sctp.local_addr_lock &net->sctp.addr_wq_lock &n->list_lock irq_context: 0 rtnl_mutex rcu_read_lock &net->sctp.local_addr_lock &net->sctp.addr_wq_lock &n->list_lock &c->lock irq_context: 0 rtnl_mutex &tb->tb6_lock &n->list_lock irq_context: 0 rtnl_mutex &tb->tb6_lock &n->list_lock &c->lock irq_context: 0 rtnl_mutex team->team_lock_key#5 _xmit_ETHER &c->lock irq_context: 0 rtnl_mutex &root->kernfs_rwsem kernfs_idr_lock &obj_hash[i].lock pool_lock irq_context: 0 (wq_completion)rcu_gp (work_completion)(&rew->rew_work) &pool->lock &p->pi_lock irq_context: 0 (wq_completion)rcu_gp (work_completion)(&rew->rew_work) &pool->lock &p->pi_lock &rq->__lock irq_context: 0 (wq_completion)rcu_gp (work_completion)(&rew->rew_work) &pool->lock &p->pi_lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 rtnl_mutex gdp_mutex fs_reclaim irq_context: 0 rtnl_mutex gdp_mutex fs_reclaim mmu_notifier_invalidate_range_start irq_context: 0 rtnl_mutex gdp_mutex pool_lock#2 irq_context: 0 rtnl_mutex gdp_mutex lock irq_context: 0 rtnl_mutex gdp_mutex lock kernfs_idr_lock irq_context: 0 rtnl_mutex gdp_mutex lock kernfs_idr_lock pool_lock#2 irq_context: 0 rtnl_mutex gdp_mutex &root->kernfs_rwsem irq_context: 0 rtnl_mutex gdp_mutex &root->kernfs_rwsem &root->kernfs_iattr_rwsem irq_context: 0 rtnl_mutex gdp_mutex kobj_ns_type_lock irq_context: 0 rtnl_mutex &k->k_lock irq_context: softirq &(&slave->notify_work)->timer rcu_read_lock &pool->lock/1 &p->pi_lock &cfs_rq->removed.lock irq_context: 0 rtnl_mutex &rcu_state.expedited_wq irq_context: 0 rtnl_mutex &rcu_state.expedited_wq &p->pi_lock irq_context: 0 rtnl_mutex &rcu_state.expedited_wq &p->pi_lock &rq->__lock irq_context: 0 rtnl_mutex &rcu_state.expedited_wq &p->pi_lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 rtnl_mutex &hard_iface->bat_iv.ogm_buff_mutex &obj_hash[i].lock pool_lock irq_context: 0 rtnl_mutex &hard_iface->bat_iv.ogm_buff_mutex fill_pool_map-wait-type-override pool_lock#2 irq_context: 0 rtnl_mutex &hard_iface->bat_iv.ogm_buff_mutex fill_pool_map-wait-type-override pool_lock irq_context: 0 rtnl_mutex &idev->mc_lock &vlan_netdev_addr_lock_key irq_context: 0 rtnl_mutex &idev->mc_lock &vlan_netdev_addr_lock_key pool_lock#2 irq_context: 0 rtnl_mutex &sb->s_type->i_mutex_key#3 rcu_read_lock &dentry->d_lock irq_context: 0 rtnl_mutex &sb->s_type->i_mutex_key#3 (console_sem).lock irq_context: 0 rtnl_mutex &sb->s_type->i_mutex_key#3 console_lock console_srcu console_owner_lock irq_context: 0 rtnl_mutex &sb->s_type->i_mutex_key#3 console_lock console_srcu console_owner irq_context: 0 rtnl_mutex &sb->s_type->i_mutex_key#3 console_lock console_srcu console_owner &port_lock_key irq_context: 0 rtnl_mutex &sb->s_type->i_mutex_key#3 console_lock console_srcu console_owner console_owner_lock irq_context: 0 rtnl_mutex rcu_read_lock mount_lock irq_context: 0 rtnl_mutex rcu_read_lock mount_lock mount_lock.seqcount irq_context: 0 rtnl_mutex mount_lock irq_context: 0 rtnl_mutex mount_lock mount_lock.seqcount irq_context: 0 rtnl_mutex &hard_iface->bat_iv.ogm_buff_mutex &n->list_lock irq_context: 0 rtnl_mutex &hard_iface->bat_iv.ogm_buff_mutex &n->list_lock &c->lock irq_context: softirq (&app->join_timer) irq_context: softirq (&app->join_timer) &app->lock irq_context: softirq (&app->join_timer) &list->lock#11 irq_context: softirq (&app->join_timer) &app->lock batched_entropy_u32.lock irq_context: softirq (&app->join_timer) &app->lock &obj_hash[i].lock irq_context: softirq (&app->join_timer) &app->lock &base->lock irq_context: softirq (&app->join_timer) &app->lock &base->lock &obj_hash[i].lock irq_context: softirq (&app->join_timer)#2 irq_context: softirq (&app->join_timer)#2 &app->lock#2 irq_context: softirq (&app->join_timer)#2 &list->lock#12 irq_context: softirq (&app->join_timer)#2 batched_entropy_u32.lock irq_context: softirq (&app->join_timer)#2 &obj_hash[i].lock irq_context: softirq (&app->join_timer)#2 &base->lock irq_context: softirq (&app->join_timer)#2 &base->lock &obj_hash[i].lock irq_context: 0 rtnl_mutex rcu_read_lock &net->sctp.local_addr_lock &net->sctp.addr_wq_lock fill_pool_map-wait-type-override pool_lock#2 irq_context: 0 rtnl_mutex rcu_read_lock &net->sctp.local_addr_lock &net->sctp.addr_wq_lock fill_pool_map-wait-type-override pool_lock irq_context: 0 &sb->s_type->i_mutex_key#10 rcu_node_0 irq_context: 0 &sb->s_type->i_mutex_key#10 &rcu_state.expedited_wq irq_context: 0 &sb->s_type->i_mutex_key#10 &rcu_state.expedited_wq &p->pi_lock irq_context: 0 &sb->s_type->i_mutex_key#10 &rcu_state.expedited_wq &p->pi_lock &rq->__lock irq_context: 0 &sb->s_type->i_mutex_key#10 &rcu_state.expedited_wq &p->pi_lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)events (linkwatch_work).work rtnl_mutex team->team_lock_key#4 irq_context: 0 (wq_completion)events (linkwatch_work).work rtnl_mutex team->team_lock_key#4 fs_reclaim irq_context: 0 (wq_completion)events (linkwatch_work).work rtnl_mutex team->team_lock_key#4 fs_reclaim mmu_notifier_invalidate_range_start irq_context: 0 (wq_completion)events (linkwatch_work).work rtnl_mutex team->team_lock_key#4 &c->lock irq_context: 0 (wq_completion)events (linkwatch_work).work rtnl_mutex team->team_lock_key#4 &____s->seqcount irq_context: 0 (wq_completion)events (linkwatch_work).work rtnl_mutex team->team_lock_key#4 rcu_read_lock &ndev->lock irq_context: 0 (wq_completion)events (linkwatch_work).work rtnl_mutex team->team_lock_key#4 &obj_hash[i].lock irq_context: 0 (wq_completion)events (linkwatch_work).work rtnl_mutex team->team_lock_key#4 nl_table_lock irq_context: softirq &(&forw_packet_aggr->delayed_work)->timer irq_context: softirq &(&forw_packet_aggr->delayed_work)->timer rcu_read_lock &pool->lock/1 irq_context: softirq &(&forw_packet_aggr->delayed_work)->timer rcu_read_lock &pool->lock/1 &obj_hash[i].lock irq_context: softirq &(&forw_packet_aggr->delayed_work)->timer rcu_read_lock &pool->lock/1 &p->pi_lock irq_context: softirq &(&forw_packet_aggr->delayed_work)->timer rcu_read_lock &pool->lock/1 &p->pi_lock &rq->__lock irq_context: softirq &(&forw_packet_aggr->delayed_work)->timer rcu_read_lock &pool->lock/1 &p->pi_lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)events (linkwatch_work).work rtnl_mutex team->team_lock_key#4 &rq->__lock irq_context: 0 (wq_completion)bat_events (work_completion)(&(&forw_packet_aggr->delayed_work)->work) irq_context: 0 (wq_completion)bat_events (work_completion)(&(&forw_packet_aggr->delayed_work)->work) &hard_iface->bat_iv.ogm_buff_mutex irq_context: 0 (wq_completion)events (linkwatch_work).work rtnl_mutex team->team_lock_key#4 nl_table_wait.lock irq_context: 0 (wq_completion)events (linkwatch_work).work rtnl_mutex team->team_lock_key#4 net_rwsem irq_context: 0 (wq_completion)events (linkwatch_work).work rtnl_mutex team->team_lock_key#4 net_rwsem &list->lock#2 irq_context: 0 (wq_completion)events (linkwatch_work).work rtnl_mutex team->team_lock_key#4 &tn->lock irq_context: 0 (wq_completion)events (linkwatch_work).work rtnl_mutex &pcp->lock &zone->lock irq_context: 0 (wq_completion)events (linkwatch_work).work rtnl_mutex &pcp->lock &zone->lock &____s->seqcount irq_context: 0 (wq_completion)bat_events (work_completion)(&(&forw_packet_aggr->delayed_work)->work) &hard_iface->bat_iv.ogm_buff_mutex batched_entropy_u8.lock irq_context: 0 (wq_completion)bat_events (work_completion)(&(&forw_packet_aggr->delayed_work)->work) &hard_iface->bat_iv.ogm_buff_mutex &bat_priv->forw_bat_list_lock irq_context: 0 (wq_completion)bat_events (work_completion)(&(&forw_packet_aggr->delayed_work)->work) &hard_iface->bat_iv.ogm_buff_mutex pool_lock#2 irq_context: 0 (wq_completion)bat_events (work_completion)(&(&forw_packet_aggr->delayed_work)->work) &hard_iface->bat_iv.ogm_buff_mutex &obj_hash[i].lock irq_context: 0 (wq_completion)bat_events (work_completion)(&(&forw_packet_aggr->delayed_work)->work) &hard_iface->bat_iv.ogm_buff_mutex &bat_priv->forw_bat_list_lock &obj_hash[i].lock irq_context: 0 (wq_completion)bat_events (work_completion)(&(&forw_packet_aggr->delayed_work)->work) &hard_iface->bat_iv.ogm_buff_mutex &bat_priv->forw_bat_list_lock &base->lock irq_context: 0 (wq_completion)bat_events (work_completion)(&(&forw_packet_aggr->delayed_work)->work) &hard_iface->bat_iv.ogm_buff_mutex &bat_priv->forw_bat_list_lock &base->lock &obj_hash[i].lock irq_context: 0 (wq_completion)bat_events (work_completion)(&(&forw_packet_aggr->delayed_work)->work) &bat_priv->forw_bat_list_lock irq_context: 0 (wq_completion)bat_events (work_completion)(&(&forw_packet_aggr->delayed_work)->work) &obj_hash[i].lock irq_context: 0 (wq_completion)bat_events (work_completion)(&(&forw_packet_aggr->delayed_work)->work) rcu_node_0 irq_context: 0 (wq_completion)bat_events (work_completion)(&(&forw_packet_aggr->delayed_work)->work) &rcu_state.expedited_wq irq_context: 0 (wq_completion)bat_events (work_completion)(&(&forw_packet_aggr->delayed_work)->work) &rcu_state.expedited_wq &p->pi_lock irq_context: 0 (wq_completion)bat_events (work_completion)(&(&forw_packet_aggr->delayed_work)->work) &rq->__lock irq_context: 0 (wq_completion)bat_events (work_completion)(&(&forw_packet_aggr->delayed_work)->work) &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)bat_events (work_completion)(&(&forw_packet_aggr->delayed_work)->work) pool_lock#2 irq_context: 0 (wq_completion)bat_events (work_completion)(&(&forw_packet_aggr->delayed_work)->work) &hard_iface->bat_iv.ogm_buff_mutex rcu_read_lock rcu_node_0 irq_context: 0 (wq_completion)bat_events (work_completion)(&(&forw_packet_aggr->delayed_work)->work) &hard_iface->bat_iv.ogm_buff_mutex rcu_read_lock &rcu_state.expedited_wq irq_context: 0 (wq_completion)bat_events (work_completion)(&(&forw_packet_aggr->delayed_work)->work) &hard_iface->bat_iv.ogm_buff_mutex rcu_read_lock &rcu_state.expedited_wq &p->pi_lock irq_context: 0 (wq_completion)bat_events (work_completion)(&(&forw_packet_aggr->delayed_work)->work) &hard_iface->bat_iv.ogm_buff_mutex rcu_read_lock &rcu_state.expedited_wq &p->pi_lock &rq->__lock irq_context: 0 (wq_completion)bat_events (work_completion)(&(&forw_packet_aggr->delayed_work)->work) &hard_iface->bat_iv.ogm_buff_mutex &c->lock irq_context: 0 (wq_completion)bat_events (work_completion)(&(&forw_packet_aggr->delayed_work)->work) &hard_iface->bat_iv.ogm_buff_mutex rcu_node_0 irq_context: 0 (wq_completion)bat_events (work_completion)(&(&forw_packet_aggr->delayed_work)->work) &hard_iface->bat_iv.ogm_buff_mutex &rcu_state.expedited_wq irq_context: 0 (wq_completion)bat_events (work_completion)(&(&forw_packet_aggr->delayed_work)->work) &hard_iface->bat_iv.ogm_buff_mutex &rcu_state.expedited_wq &p->pi_lock irq_context: 0 (wq_completion)bat_events (work_completion)(&(&forw_packet_aggr->delayed_work)->work) &hard_iface->bat_iv.ogm_buff_mutex &rq->__lock irq_context: 0 (wq_completion)bat_events (work_completion)(&(&forw_packet_aggr->delayed_work)->work) &hard_iface->bat_iv.ogm_buff_mutex &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)bat_events (work_completion)(&(&forw_packet_aggr->delayed_work)->work) &hard_iface->bat_iv.ogm_buff_mutex &____s->seqcount irq_context: 0 (wq_completion)events (linkwatch_work).work rtnl_mutex team->team_lock_key#5 irq_context: 0 (wq_completion)events (linkwatch_work).work rtnl_mutex team->team_lock_key#5 fs_reclaim irq_context: 0 (wq_completion)events (linkwatch_work).work rtnl_mutex team->team_lock_key#5 fs_reclaim mmu_notifier_invalidate_range_start irq_context: 0 (wq_completion)events (linkwatch_work).work rtnl_mutex team->team_lock_key#5 &c->lock irq_context: 0 (wq_completion)events (linkwatch_work).work rtnl_mutex team->team_lock_key#5 &____s->seqcount irq_context: 0 (wq_completion)events (linkwatch_work).work rtnl_mutex team->team_lock_key#5 rcu_read_lock &ndev->lock irq_context: 0 (wq_completion)events (linkwatch_work).work rtnl_mutex team->team_lock_key#5 &obj_hash[i].lock irq_context: 0 (wq_completion)events (linkwatch_work).work rtnl_mutex team->team_lock_key#5 nl_table_lock irq_context: 0 (wq_completion)events (linkwatch_work).work rtnl_mutex team->team_lock_key#5 nl_table_wait.lock irq_context: 0 (wq_completion)events (linkwatch_work).work rtnl_mutex team->team_lock_key#5 net_rwsem irq_context: 0 (wq_completion)events (linkwatch_work).work rtnl_mutex team->team_lock_key#5 net_rwsem &list->lock#2 irq_context: 0 (wq_completion)events (linkwatch_work).work rtnl_mutex team->team_lock_key#5 &tn->lock irq_context: 0 (wq_completion)events (linkwatch_work).work rtnl_mutex team->team_lock_key#6 irq_context: 0 (wq_completion)events (linkwatch_work).work rtnl_mutex team->team_lock_key#6 fs_reclaim irq_context: 0 (wq_completion)events (linkwatch_work).work rtnl_mutex team->team_lock_key#6 fs_reclaim mmu_notifier_invalidate_range_start irq_context: 0 (wq_completion)events (linkwatch_work).work rtnl_mutex team->team_lock_key#6 rcu_read_lock &ndev->lock irq_context: 0 (wq_completion)events (linkwatch_work).work rtnl_mutex team->team_lock_key#6 &obj_hash[i].lock irq_context: 0 (wq_completion)events (linkwatch_work).work rtnl_mutex team->team_lock_key#6 nl_table_lock irq_context: 0 (wq_completion)events (linkwatch_work).work rtnl_mutex team->team_lock_key#6 nl_table_wait.lock irq_context: 0 (wq_completion)events (linkwatch_work).work rtnl_mutex team->team_lock_key#6 net_rwsem irq_context: 0 (wq_completion)events (linkwatch_work).work rtnl_mutex team->team_lock_key#6 net_rwsem &list->lock#2 irq_context: 0 (wq_completion)events (linkwatch_work).work rtnl_mutex team->team_lock_key#6 &tn->lock irq_context: 0 (wq_completion)events (linkwatch_work).work rtnl_mutex quarantine_lock irq_context: 0 (wq_completion)events (linkwatch_work).work rtnl_mutex team->team_lock_key#6 remove_cache_srcu irq_context: 0 (wq_completion)events (linkwatch_work).work rtnl_mutex team->team_lock_key#6 remove_cache_srcu quarantine_lock irq_context: 0 (wq_completion)events (linkwatch_work).work rtnl_mutex team->team_lock_key#6 remove_cache_srcu &c->lock irq_context: 0 (wq_completion)events (linkwatch_work).work rtnl_mutex team->team_lock_key#6 remove_cache_srcu &n->list_lock irq_context: 0 (wq_completion)events (linkwatch_work).work rtnl_mutex team->team_lock_key#6 remove_cache_srcu rcu_read_lock pool_lock#2 irq_context: 0 (wq_completion)events (linkwatch_work).work rtnl_mutex team->team_lock_key#6 remove_cache_srcu &obj_hash[i].lock irq_context: 0 rtnl_mutex kernfs_idr_lock &obj_hash[i].lock irq_context: 0 rtnl_mutex kernfs_idr_lock pool_lock#2 irq_context: 0 tomoyo_ss tomoyo_policy_lock &rq->__lock irq_context: 0 fs_reclaim mmu_notifier_invalidate_range_start rcu_node_0 irq_context: 0 fs_reclaim mmu_notifier_invalidate_range_start &rcu_state.expedited_wq irq_context: 0 fs_reclaim mmu_notifier_invalidate_range_start &rcu_state.expedited_wq &p->pi_lock irq_context: 0 rtnl_mutex &idev->mc_lock &macvlan_netdev_addr_lock_key irq_context: 0 rtnl_mutex &idev->mc_lock &macvlan_netdev_addr_lock_key pool_lock#2 irq_context: 0 (wq_completion)bond0#5 (work_completion)(&(&slave->notify_work)->work) rtnl_mutex net_rwsem irq_context: 0 (wq_completion)bond0#5 (work_completion)(&(&slave->notify_work)->work) rtnl_mutex net_rwsem &list->lock#2 irq_context: 0 (wq_completion)bond0#5 (work_completion)(&(&slave->notify_work)->work) rtnl_mutex &tn->lock irq_context: 0 (wq_completion)bond0#5 (work_completion)(&(&slave->notify_work)->work) rtnl_mutex.wait_lock irq_context: 0 (wq_completion)bond0#5 (work_completion)(&(&slave->notify_work)->work) &p->pi_lock irq_context: 0 (wq_completion)bat_events (work_completion)(&(&bat_priv->orig_work)->work) fill_pool_map-wait-type-override pool_lock#2 irq_context: 0 (wq_completion)bat_events (work_completion)(&(&bat_priv->orig_work)->work) fill_pool_map-wait-type-override pool_lock irq_context: 0 rtnl_mutex &idev->mc_lock &dev_addr_list_lock_key#3 irq_context: 0 rtnl_mutex &idev->mc_lock &dev_addr_list_lock_key#3 pool_lock#2 irq_context: 0 rtnl_mutex &xa->xa_lock#15 irq_context: 0 rtnl_mutex &dev_addr_list_lock_key#3/1 irq_context: 0 (wq_completion)rcu_gp (work_completion)(&rew->rew_work) fill_pool_map-wait-type-override &c->lock irq_context: 0 (wq_completion)rcu_gp (work_completion)(&rew->rew_work) fill_pool_map-wait-type-override &____s->seqcount irq_context: 0 rtnl_mutex &root->kernfs_rwsem quarantine_lock irq_context: 0 rtnl_mutex &idev->mc_lock &vlan_netdev_addr_lock_key &c->lock irq_context: 0 rtnl_mutex &idev->mc_lock &vlan_netdev_addr_lock_key &____s->seqcount irq_context: softirq &(&bat_priv->orig_work)->timer rcu_read_lock &pool->lock/1 &p->pi_lock &rq->__lock &cfs_rq->removed.lock irq_context: 0 rtnl_mutex lweventlist_lock &____s->seqcount irq_context: 0 rtnl_mutex rcu_read_lock &tap_major->minor_lock irq_context: 0 rtnl_mutex rcu_read_lock &tap_major->minor_lock pool_lock#2 irq_context: 0 rtnl_mutex req_lock irq_context: 0 rtnl_mutex &x->wait#11 irq_context: 0 rtnl_mutex subsys mutex#82 irq_context: 0 rtnl_mutex subsys mutex#82 &k->k_lock irq_context: 0 (wq_completion)gid-cache-wq (work_completion)(&ndev_work->work) &rq->__lock irq_context: 0 rtnl_mutex rcu_read_lock &rcu_state.gp_wq irq_context: 0 rtnl_mutex rcu_read_lock &rcu_state.gp_wq &p->pi_lock irq_context: 0 rtnl_mutex rcu_read_lock &rcu_state.gp_wq &p->pi_lock &cfs_rq->removed.lock irq_context: 0 rtnl_mutex rcu_read_lock &rcu_state.gp_wq &p->pi_lock &rq->__lock irq_context: 0 rtnl_mutex rcu_read_lock &rcu_state.gp_wq &p->pi_lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 rtnl_mutex &idev->mc_lock &dev_addr_list_lock_key#3 &c->lock irq_context: 0 rtnl_mutex &idev->mc_lock &dev_addr_list_lock_key#3 &____s->seqcount irq_context: 0 kn->active#51 fs_reclaim irq_context: 0 kn->active#51 fs_reclaim mmu_notifier_invalidate_range_start irq_context: 0 kn->active#51 &kernfs_locks->open_file_mutex[count] irq_context: 0 kn->active#51 &kernfs_locks->open_file_mutex[count] fs_reclaim irq_context: 0 kn->active#51 &kernfs_locks->open_file_mutex[count] fs_reclaim mmu_notifier_invalidate_range_start irq_context: 0 sb_writers#8 &of->mutex kn->active#51 nsim_bus_dev_list_lock irq_context: 0 sb_writers#8 &type->i_mutex_dir_key#4 &sem->wait_lock irq_context: 0 sb_writers#8 &type->i_mutex_dir_key#4 &p->pi_lock irq_context: 0 sb_writers#8 &type->i_mutex_dir_key#4 &p->pi_lock &rq->__lock irq_context: 0 sb_writers#8 &type->i_mutex_dir_key#4 &p->pi_lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 sb_writers#8 &type->i_mutex_dir_key#4 &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 kn->active#52 fs_reclaim irq_context: 0 kn->active#52 fs_reclaim mmu_notifier_invalidate_range_start irq_context: 0 kn->active#52 &kernfs_locks->open_file_mutex[count] irq_context: 0 kn->active#52 &kernfs_locks->open_file_mutex[count] fs_reclaim irq_context: 0 kn->active#52 &kernfs_locks->open_file_mutex[count] fs_reclaim mmu_notifier_invalidate_range_start irq_context: 0 sb_writers#8 &of->mutex kn->active#52 nsim_bus_dev_list_lock irq_context: 0 sb_writers#8 &of->mutex kn->active#52 nsim_bus_dev_list_lock fs_reclaim irq_context: 0 sb_writers#8 &of->mutex kn->active#52 nsim_bus_dev_list_lock fs_reclaim mmu_notifier_invalidate_range_start irq_context: 0 sb_writers#8 &of->mutex kn->active#52 nsim_bus_dev_list_lock pool_lock#2 irq_context: 0 sb_writers#8 &of->mutex kn->active#52 nsim_bus_dev_list_lock nsim_bus_dev_ids.xa_lock irq_context: 0 sb_writers#8 &of->mutex kn->active#52 nsim_bus_dev_list_lock &x->wait#9 irq_context: 0 sb_writers#8 &of->mutex kn->active#52 nsim_bus_dev_list_lock &obj_hash[i].lock irq_context: 0 sb_writers#8 &of->mutex kn->active#52 nsim_bus_dev_list_lock &k->list_lock irq_context: 0 sb_writers#8 &of->mutex kn->active#52 nsim_bus_dev_list_lock lock irq_context: 0 sb_writers#8 &of->mutex kn->active#52 nsim_bus_dev_list_lock lock kernfs_idr_lock irq_context: 0 sb_writers#8 &of->mutex kn->active#52 nsim_bus_dev_list_lock &root->kernfs_rwsem irq_context: 0 sb_writers#8 &of->mutex kn->active#52 nsim_bus_dev_list_lock &root->kernfs_rwsem &root->kernfs_iattr_rwsem irq_context: 0 sb_writers#8 &of->mutex kn->active#52 nsim_bus_dev_list_lock bus_type_sem irq_context: 0 sb_writers#8 &of->mutex kn->active#52 nsim_bus_dev_list_lock sysfs_symlink_target_lock irq_context: 0 sb_writers#8 &of->mutex kn->active#52 nsim_bus_dev_list_lock &k->k_lock irq_context: 0 sb_writers#8 &of->mutex kn->active#52 nsim_bus_dev_list_lock &root->kernfs_rwsem irq_context: 0 sb_writers#8 &of->mutex kn->active#52 nsim_bus_dev_list_lock &root->kernfs_rwsem &rq->__lock irq_context: 0 sb_writers#8 &of->mutex kn->active#52 nsim_bus_dev_list_lock &c->lock irq_context: 0 sb_writers#8 &of->mutex kn->active#52 nsim_bus_dev_list_lock &____s->seqcount irq_context: 0 sb_writers#8 &of->mutex kn->active#52 nsim_bus_dev_list_lock &dev->power.lock irq_context: 0 sb_writers#8 &of->mutex kn->active#52 nsim_bus_dev_list_lock dpm_list_mtx irq_context: 0 sb_writers#8 &of->mutex kn->active#52 nsim_bus_dev_list_lock uevent_sock_mutex irq_context: 0 sb_writers#8 &of->mutex kn->active#52 nsim_bus_dev_list_lock uevent_sock_mutex fs_reclaim irq_context: 0 sb_writers#8 &of->mutex kn->active#52 nsim_bus_dev_list_lock uevent_sock_mutex fs_reclaim mmu_notifier_invalidate_range_start irq_context: 0 sb_writers#8 &of->mutex kn->active#52 nsim_bus_dev_list_lock uevent_sock_mutex &c->lock irq_context: 0 sb_writers#8 &of->mutex kn->active#52 nsim_bus_dev_list_lock uevent_sock_mutex pool_lock#2 irq_context: 0 sb_writers#8 &of->mutex kn->active#52 nsim_bus_dev_list_lock uevent_sock_mutex nl_table_lock irq_context: 0 sb_writers#8 &of->mutex kn->active#52 nsim_bus_dev_list_lock uevent_sock_mutex rlock-AF_NETLINK irq_context: 0 sb_writers#8 &of->mutex kn->active#52 nsim_bus_dev_list_lock uevent_sock_mutex rcu_read_lock &ei->socket.wq.wait irq_context: 0 sb_writers#8 &of->mutex kn->active#52 nsim_bus_dev_list_lock uevent_sock_mutex rcu_read_lock &ei->socket.wq.wait &ep->lock irq_context: 0 sb_writers#8 &of->mutex kn->active#52 nsim_bus_dev_list_lock uevent_sock_mutex rcu_read_lock &ei->socket.wq.wait &ep->lock &ep->wq irq_context: 0 sb_writers#8 &of->mutex kn->active#52 nsim_bus_dev_list_lock uevent_sock_mutex rcu_read_lock &ei->socket.wq.wait &ep->lock &ep->wq &p->pi_lock irq_context: 0 sb_writers#8 &of->mutex kn->active#52 nsim_bus_dev_list_lock uevent_sock_mutex rcu_read_lock &ei->socket.wq.wait &ep->lock &ep->wq &p->pi_lock &rq->__lock irq_context: 0 sb_writers#8 &of->mutex kn->active#52 nsim_bus_dev_list_lock uevent_sock_mutex rcu_read_lock &ei->socket.wq.wait &ep->lock &ep->wq &p->pi_lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 sb_writers#8 &of->mutex kn->active#52 nsim_bus_dev_list_lock uevent_sock_mutex rcu_read_lock rcu_node_0 irq_context: 0 sb_writers#8 &of->mutex kn->active#52 nsim_bus_dev_list_lock uevent_sock_mutex rcu_read_lock &rq->__lock irq_context: 0 sb_writers#5 fill_pool_map-wait-type-override pool_lock#2 irq_context: 0 sb_writers#5 fill_pool_map-wait-type-override pool_lock irq_context: 0 sb_writers#8 &of->mutex kn->active#52 nsim_bus_dev_list_lock uevent_sock_mutex nl_table_wait.lock irq_context: 0 sb_writers#8 &of->mutex kn->active#52 nsim_bus_dev_list_lock uevent_sock_mutex &obj_hash[i].lock irq_context: 0 sb_writers#8 &of->mutex kn->active#52 nsim_bus_dev_list_lock &dev->mutex &k->list_lock irq_context: 0 sb_writers#8 &of->mutex kn->active#52 nsim_bus_dev_list_lock &dev->mutex &k->k_lock irq_context: 0 sb_writers#8 &of->mutex kn->active#52 nsim_bus_dev_list_lock &dev->mutex device_links_srcu irq_context: 0 sb_writers#8 &of->mutex kn->active#52 nsim_bus_dev_list_lock &dev->mutex &dev->power.lock irq_context: 0 sb_writers#8 &of->mutex kn->active#52 nsim_bus_dev_list_lock &dev->mutex fwnode_link_lock irq_context: 0 sb_writers#8 &of->mutex kn->active#52 nsim_bus_dev_list_lock &dev->mutex device_links_lock irq_context: 0 sb_writers#8 &of->mutex kn->active#52 nsim_bus_dev_list_lock &dev->mutex sysfs_symlink_target_lock irq_context: 0 sb_writers#8 &of->mutex kn->active#52 nsim_bus_dev_list_lock &dev->mutex fs_reclaim irq_context: 0 sb_writers#8 &of->mutex kn->active#52 nsim_bus_dev_list_lock &dev->mutex fs_reclaim mmu_notifier_invalidate_range_start irq_context: 0 sb_writers#8 &of->mutex kn->active#52 nsim_bus_dev_list_lock &dev->mutex pool_lock#2 irq_context: 0 sb_writers#8 &of->mutex kn->active#52 nsim_bus_dev_list_lock &dev->mutex lock irq_context: 0 sb_writers#8 &of->mutex kn->active#52 nsim_bus_dev_list_lock &dev->mutex lock kernfs_idr_lock irq_context: 0 sb_writers#8 &of->mutex kn->active#52 nsim_bus_dev_list_lock &dev->mutex &root->kernfs_rwsem irq_context: 0 sb_writers#8 &of->mutex kn->active#52 nsim_bus_dev_list_lock &dev->mutex &root->kernfs_rwsem &root->kernfs_iattr_rwsem irq_context: 0 sb_writers#8 &of->mutex kn->active#52 nsim_bus_dev_list_lock &dev->mutex devlinks.xa_lock irq_context: 0 sb_writers#8 &of->mutex kn->active#52 nsim_bus_dev_list_lock &dev->mutex &obj_hash[i].lock irq_context: 0 sb_writers#8 &of->mutex kn->active#52 nsim_bus_dev_list_lock &dev->mutex &devlink->lock_key irq_context: 0 sb_writers#8 &of->mutex kn->active#52 nsim_bus_dev_list_lock &dev->mutex &devlink->lock_key crngs.lock irq_context: 0 sb_writers#8 &of->mutex kn->active#52 nsim_bus_dev_list_lock &dev->mutex &devlink->lock_key fs_reclaim irq_context: 0 sb_writers#8 &of->mutex kn->active#52 nsim_bus_dev_list_lock &dev->mutex &devlink->lock_key fs_reclaim mmu_notifier_invalidate_range_start irq_context: 0 sb_writers#8 &of->mutex kn->active#52 nsim_bus_dev_list_lock &dev->mutex &devlink->lock_key pool_lock#2 irq_context: 0 sb_writers#8 &of->mutex kn->active#52 nsim_bus_dev_list_lock &dev->mutex &devlink->lock_key devlinks.xa_lock irq_context: 0 sb_writers#8 &of->mutex kn->active#52 nsim_bus_dev_list_lock &dev->mutex &devlink->lock_key &obj_hash[i].lock irq_context: 0 sb_writers#8 &of->mutex kn->active#52 nsim_bus_dev_list_lock &dev->mutex &devlink->lock_key nl_table_lock irq_context: 0 sb_writers#8 &of->mutex kn->active#52 nsim_bus_dev_list_lock &dev->mutex &devlink->lock_key nl_table_wait.lock irq_context: 0 sb_writers#8 &of->mutex kn->active#52 nsim_bus_dev_list_lock &dev->mutex &devlink->lock_key &c->lock irq_context: 0 sb_writers#8 &of->mutex kn->active#52 nsim_bus_dev_list_lock &dev->mutex &devlink->lock_key &____s->seqcount irq_context: 0 sb_writers#8 &of->mutex kn->active#52 nsim_bus_dev_list_lock &dev->mutex &devlink->lock_key &xa->xa_lock#16 irq_context: 0 sb_writers#8 &of->mutex kn->active#52 nsim_bus_dev_list_lock &dev->mutex &devlink->lock_key &xa->xa_lock#16 pool_lock#2 irq_context: 0 sb_writers#8 &of->mutex kn->active#52 nsim_bus_dev_list_lock &dev->mutex &devlink->lock_key rcu_read_lock pool_lock#2 irq_context: 0 sb_writers#8 &of->mutex kn->active#52 nsim_bus_dev_list_lock &dev->mutex &devlink->lock_key pcpu_alloc_mutex irq_context: 0 sb_writers#8 &of->mutex kn->active#52 nsim_bus_dev_list_lock &dev->mutex &devlink->lock_key pcpu_alloc_mutex pcpu_lock irq_context: 0 sb_writers#8 &of->mutex kn->active#52 nsim_bus_dev_list_lock &dev->mutex &devlink->lock_key quarantine_lock irq_context: 0 sb_writers#8 &of->mutex kn->active#52 nsim_bus_dev_list_lock &dev->mutex &devlink->lock_key &n->list_lock irq_context: 0 sb_writers#8 &of->mutex kn->active#52 nsim_bus_dev_list_lock &dev->mutex &devlink->lock_key &n->list_lock &c->lock irq_context: 0 sb_writers#8 &of->mutex kn->active#52 nsim_bus_dev_list_lock &dev->mutex &devlink->lock_key &pcp->lock &zone->lock irq_context: 0 sb_writers#8 &of->mutex kn->active#52 nsim_bus_dev_list_lock &dev->mutex &devlink->lock_key &pcp->lock &zone->lock &____s->seqcount irq_context: 0 sb_writers#8 &of->mutex kn->active#52 nsim_bus_dev_list_lock &dev->mutex &devlink->lock_key &base->lock irq_context: 0 sb_writers#8 &of->mutex kn->active#52 nsim_bus_dev_list_lock &dev->mutex &devlink->lock_key &base->lock &obj_hash[i].lock irq_context: 0 sb_writers#8 &of->mutex kn->active#52 nsim_bus_dev_list_lock &dev->mutex &devlink->lock_key pin_fs_lock irq_context: 0 sb_writers#8 &of->mutex kn->active#52 nsim_bus_dev_list_lock &dev->mutex &devlink->lock_key &sb->s_type->i_mutex_key#3 irq_context: 0 sb_writers#8 &of->mutex kn->active#52 nsim_bus_dev_list_lock &dev->mutex &devlink->lock_key &sb->s_type->i_mutex_key#3 &sb->s_type->i_lock_key#7 irq_context: 0 sb_writers#8 &of->mutex kn->active#52 nsim_bus_dev_list_lock &dev->mutex &devlink->lock_key &sb->s_type->i_mutex_key#3 rename_lock.seqcount irq_context: 0 rtnl_mutex remove_cache_srcu rcu_read_lock rcu_node_0 irq_context: 0 sb_writers#8 &of->mutex kn->active#52 nsim_bus_dev_list_lock &dev->mutex &devlink->lock_key &sb->s_type->i_mutex_key#3 fs_reclaim irq_context: 0 rtnl_mutex remove_cache_srcu rcu_read_lock &rq->__lock irq_context: 0 sb_writers#8 &of->mutex kn->active#52 nsim_bus_dev_list_lock &dev->mutex &devlink->lock_key &sb->s_type->i_mutex_key#3 fs_reclaim mmu_notifier_invalidate_range_start irq_context: 0 rtnl_mutex remove_cache_srcu rcu_read_lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 sb_writers#8 &of->mutex kn->active#52 nsim_bus_dev_list_lock &dev->mutex &devlink->lock_key &sb->s_type->i_mutex_key#3 pool_lock#2 irq_context: 0 sb_writers#8 &of->mutex kn->active#52 nsim_bus_dev_list_lock &dev->mutex &devlink->lock_key &sb->s_type->i_mutex_key#3 &dentry->d_lock irq_context: 0 sb_writers#8 &of->mutex kn->active#52 nsim_bus_dev_list_lock &dev->mutex &devlink->lock_key &sb->s_type->i_mutex_key#3 rcu_read_lock rename_lock.seqcount irq_context: 0 sb_writers#8 &of->mutex kn->active#52 nsim_bus_dev_list_lock &dev->mutex &devlink->lock_key &sb->s_type->i_mutex_key#3 rcu_read_lock rcu_node_0 irq_context: 0 sb_writers#8 &of->mutex kn->active#52 nsim_bus_dev_list_lock &dev->mutex &devlink->lock_key &sb->s_type->i_mutex_key#3 rcu_read_lock &rq->__lock irq_context: 0 sb_writers#8 &of->mutex kn->active#52 nsim_bus_dev_list_lock &dev->mutex &devlink->lock_key &sb->s_type->i_mutex_key#3 rcu_node_0 irq_context: 0 sb_writers#8 &of->mutex kn->active#52 nsim_bus_dev_list_lock &dev->mutex &devlink->lock_key &sb->s_type->i_mutex_key#3 &rq->__lock irq_context: 0 sb_writers#8 &of->mutex kn->active#52 nsim_bus_dev_list_lock &dev->mutex &devlink->lock_key &sb->s_type->i_mutex_key#3 &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 sb_writers#8 &of->mutex kn->active#52 nsim_bus_dev_list_lock &dev->mutex &devlink->lock_key &sb->s_type->i_mutex_key#3 &dentry->d_lock &wq irq_context: 0 sb_writers#8 &of->mutex kn->active#52 nsim_bus_dev_list_lock &dev->mutex &devlink->lock_key &sb->s_type->i_mutex_key#3 mmu_notifier_invalidate_range_start irq_context: 0 sb_writers#8 &of->mutex kn->active#52 nsim_bus_dev_list_lock &dev->mutex &devlink->lock_key &sb->s_type->i_mutex_key#3 &s->s_inode_list_lock irq_context: 0 sb_writers#8 &of->mutex kn->active#52 nsim_bus_dev_list_lock &dev->mutex &devlink->lock_key &sb->s_type->i_mutex_key#3 tk_core.seq.seqcount irq_context: 0 sb_writers#8 &of->mutex kn->active#52 nsim_bus_dev_list_lock &dev->mutex &devlink->lock_key &sb->s_type->i_mutex_key#3 &sb->s_type->i_lock_key#7 &dentry->d_lock irq_context: 0 sb_writers#8 &of->mutex kn->active#52 nsim_bus_dev_list_lock &dev->mutex &devlink->lock_key &sb->s_type->i_mutex_key#3 &c->lock irq_context: 0 sb_writers#8 &of->mutex kn->active#52 nsim_bus_dev_list_lock &dev->mutex &devlink->lock_key &sb->s_type->i_mutex_key#3 &n->list_lock irq_context: 0 sb_writers#8 &of->mutex kn->active#52 nsim_bus_dev_list_lock &dev->mutex &devlink->lock_key &sb->s_type->i_mutex_key#3 &n->list_lock &c->lock irq_context: 0 sb_writers#8 &of->mutex kn->active#52 nsim_bus_dev_list_lock &dev->mutex &devlink->lock_key &sb->s_type->i_mutex_key#3 &____s->seqcount irq_context: 0 sb_writers#8 &of->mutex kn->active#52 nsim_bus_dev_list_lock &dev->mutex &devlink->lock_key &sb->s_type->i_mutex_key#3 rcu_read_lock pool_lock#2 irq_context: 0 sb_writers#8 &of->mutex kn->active#52 nsim_bus_dev_list_lock &dev->mutex &devlink->lock_key &sb->s_type->i_mutex_key#3 &obj_hash[i].lock irq_context: 0 sb_writers#8 &of->mutex kn->active#52 nsim_bus_dev_list_lock &dev->mutex &devlink->lock_key batched_entropy_u32.lock irq_context: 0 sb_writers#8 &of->mutex kn->active#52 nsim_bus_dev_list_lock &dev->mutex &devlink->lock_key rtnl_mutex irq_context: 0 sb_writers#8 &of->mutex kn->active#52 nsim_bus_dev_list_lock &dev->mutex &devlink->lock_key rtnl_mutex rtnl_mutex.wait_lock irq_context: 0 sb_writers#8 &of->mutex kn->active#52 nsim_bus_dev_list_lock &dev->mutex &devlink->lock_key rtnl_mutex &rq->__lock irq_context: 0 sb_writers#8 &of->mutex kn->active#52 nsim_bus_dev_list_lock &dev->mutex &devlink->lock_key rtnl_mutex &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 sb_writers#8 &of->mutex kn->active#52 nsim_bus_dev_list_lock &dev->mutex &devlink->lock_key rtnl_mutex rcu_read_lock &rq->__lock irq_context: 0 sb_writers#8 &of->mutex kn->active#52 nsim_bus_dev_list_lock &dev->mutex &devlink->lock_key rtnl_mutex rcu_read_lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 rtnl_mutex &root->kernfs_rwsem &meta->lock irq_context: 0 rtnl_mutex &root->kernfs_rwsem kfence_freelist_lock irq_context: 0 rtnl_mutex &idev->mc_lock &macvlan_netdev_addr_lock_key &c->lock irq_context: 0 rtnl_mutex &idev->mc_lock &macvlan_netdev_addr_lock_key &____s->seqcount irq_context: 0 sb_writers#8 &of->mutex kn->active#52 nsim_bus_dev_list_lock &dev->mutex &devlink->lock_key rtnl_mutex &(&net->nexthop.notifier_chain)->rwsem irq_context: 0 sb_writers#8 &of->mutex kn->active#52 nsim_bus_dev_list_lock &dev->mutex &devlink->lock_key rtnl_mutex.wait_lock irq_context: 0 sb_writers#8 &of->mutex kn->active#52 nsim_bus_dev_list_lock &dev->mutex &devlink->lock_key &p->pi_lock irq_context: 0 sb_writers#8 &of->mutex kn->active#52 nsim_bus_dev_list_lock &dev->mutex &devlink->lock_key &p->pi_lock &rq->__lock irq_context: 0 sb_writers#8 &of->mutex kn->active#52 nsim_bus_dev_list_lock &dev->mutex &devlink->lock_key &p->pi_lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 sb_writers#8 &of->mutex kn->active#52 nsim_bus_dev_list_lock &dev->mutex &devlink->lock_key rcu_read_lock &data->fib_event_queue_lock irq_context: 0 sb_writers#8 &of->mutex kn->active#52 nsim_bus_dev_list_lock &dev->mutex &devlink->lock_key rcu_read_lock rcu_read_lock &pool->lock irq_context: 0 sb_writers#8 &of->mutex kn->active#52 nsim_bus_dev_list_lock &dev->mutex &devlink->lock_key rcu_read_lock rcu_read_lock &pool->lock &obj_hash[i].lock irq_context: 0 sb_writers#8 &of->mutex kn->active#52 nsim_bus_dev_list_lock &dev->mutex &devlink->lock_key rcu_read_lock rcu_read_lock &pool->lock pool_lock#2 irq_context: 0 sb_writers#8 &of->mutex kn->active#52 nsim_bus_dev_list_lock &dev->mutex &devlink->lock_key rcu_read_lock rcu_read_lock &pool->lock &p->pi_lock irq_context: 0 sb_writers#8 &of->mutex kn->active#52 nsim_bus_dev_list_lock &dev->mutex &devlink->lock_key rcu_read_lock rcu_read_lock &pool->lock &p->pi_lock &rq->__lock irq_context: 0 sb_writers#8 &of->mutex kn->active#52 nsim_bus_dev_list_lock &dev->mutex &devlink->lock_key rcu_read_lock rcu_read_lock &pool->lock &p->pi_lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 sb_writers#8 &of->mutex kn->active#52 nsim_bus_dev_list_lock &dev->mutex &devlink->lock_key rcu_read_lock rcu_node_0 irq_context: 0 sb_writers#8 &of->mutex kn->active#52 nsim_bus_dev_list_lock &dev->mutex &devlink->lock_key rcu_read_lock &rq->__lock irq_context: 0 (wq_completion)events (work_completion)(&data->fib_event_work) irq_context: 0 (wq_completion)events (work_completion)(&data->fib_event_work) &data->fib_event_queue_lock irq_context: 0 (wq_completion)events (work_completion)(&data->fib_event_work) &data->fib_lock irq_context: 0 (wq_completion)events (work_completion)(&data->fib_event_work) &data->fib_lock fs_reclaim irq_context: 0 (wq_completion)events (work_completion)(&data->fib_event_work) &data->fib_lock fs_reclaim mmu_notifier_invalidate_range_start irq_context: 0 (wq_completion)events (work_completion)(&data->fib_event_work) &data->fib_lock pool_lock#2 irq_context: 0 (wq_completion)events (work_completion)(&data->fib_event_work) &data->fib_lock rcu_read_lock rhashtable_bucket irq_context: 0 (wq_completion)events (work_completion)(&data->fib_event_work) &data->fib_lock &obj_hash[i].lock irq_context: 0 (wq_completion)events (work_completion)(&data->fib_event_work) &data->fib_lock &base->lock irq_context: 0 (wq_completion)events (work_completion)(&data->fib_event_work) &data->fib_lock &base->lock &obj_hash[i].lock irq_context: 0 (wq_completion)events (work_completion)(&data->fib_event_work) &data->fib_lock &pool->lock irq_context: 0 (wq_completion)events (work_completion)(&data->fib_event_work) &data->fib_lock &rq->__lock irq_context: 0 (wq_completion)events (work_completion)(&data->fib_event_work) &data->fib_lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 sb_writers#8 &of->mutex kn->active#52 nsim_bus_dev_list_lock &dev->mutex &devlink->lock_key rcu_read_lock &tb->tb6_lock irq_context: 0 sb_writers#8 &of->mutex kn->active#52 nsim_bus_dev_list_lock &dev->mutex &devlink->lock_key rcu_read_lock &tb->tb6_lock &net->ipv6.fib6_walker_lock irq_context: 0 sb_writers#8 &of->mutex kn->active#52 nsim_bus_dev_list_lock &dev->mutex &devlink->lock_key rcu_read_lock &tb->tb6_lock pool_lock#2 irq_context: 0 sb_writers#8 &of->mutex kn->active#52 nsim_bus_dev_list_lock &dev->mutex &devlink->lock_key rcu_read_lock &tb->tb6_lock &data->fib_event_queue_lock irq_context: 0 sb_writers#8 &of->mutex kn->active#52 nsim_bus_dev_list_lock &dev->mutex &devlink->lock_key rcu_read_lock &obj_hash[i].lock irq_context: 0 sb_writers#8 &of->mutex kn->active#52 nsim_bus_dev_list_lock &dev->mutex &devlink->lock_key &(&fn_net->fib_chain)->lock irq_context: 0 sb_writers#8 &of->mutex kn->active#52 nsim_bus_dev_list_lock &dev->mutex &devlink->lock_key &rq->__lock irq_context: 0 sb_writers#8 &of->mutex kn->active#51 nsim_bus_dev_list_lock nsim_bus_dev_list_lock.wait_lock irq_context: 0 sb_writers#8 &of->mutex kn->active#51 nsim_bus_dev_list_lock &rq->__lock irq_context: 0 sb_writers#8 &of->mutex kn->active#51 nsim_bus_dev_list_lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 sb_writers#8 &of->mutex kn->active#51 nsim_bus_dev_list_lock rcu_read_lock &rq->__lock irq_context: 0 sb_writers#8 &of->mutex kn->active#51 nsim_bus_dev_list_lock rcu_read_lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: softirq &(&nsim_dev->trap_data->trap_report_dw)->timer irq_context: softirq &(&nsim_dev->trap_data->trap_report_dw)->timer rcu_read_lock &pool->lock irq_context: softirq &(&nsim_dev->trap_data->trap_report_dw)->timer rcu_read_lock &pool->lock &obj_hash[i].lock irq_context: softirq &(&nsim_dev->trap_data->trap_report_dw)->timer rcu_read_lock &pool->lock &p->pi_lock irq_context: 0 (wq_completion)events (work_completion)(&data->fib_event_work) &data->fib_lock (&timer.timer) irq_context: 0 (wq_completion)events (work_completion)(&data->fib_event_work) &data->fib_lock &pool->lock &p->pi_lock irq_context: 0 (wq_completion)events (work_completion)(&(&nsim_dev->trap_data->trap_report_dw)->work) irq_context: 0 (wq_completion)events (work_completion)(&(&nsim_dev->trap_data->trap_report_dw)->work) rcu_read_lock &pool->lock irq_context: 0 (wq_completion)events (work_completion)(&(&nsim_dev->trap_data->trap_report_dw)->work) rcu_read_lock &pool->lock &obj_hash[i].lock irq_context: 0 (wq_completion)events (work_completion)(&(&nsim_dev->trap_data->trap_report_dw)->work) rcu_read_lock &pool->lock pool_lock#2 irq_context: 0 (wq_completion)events (work_completion)(&(&nsim_dev->trap_data->trap_report_dw)->work) &rq->__lock irq_context: 0 (wq_completion)events (work_completion)(&(&nsim_dev->trap_data->trap_report_dw)->work) &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)events (work_completion)(&(&nsim_dev->trap_data->trap_report_dw)->work) rcu_read_lock &pool->lock fill_pool_map-wait-type-override pool_lock#2 irq_context: 0 (wq_completion)events (work_completion)(&(&nsim_dev->trap_data->trap_report_dw)->work) rcu_read_lock &pool->lock fill_pool_map-wait-type-override pool_lock irq_context: 0 sb_writers#8 &of->mutex kn->active#52 nsim_bus_dev_list_lock &dev->mutex &devlink->lock_key &devlink_port->type_lock irq_context: 0 sb_writers#8 &of->mutex kn->active#52 nsim_bus_dev_list_lock &dev->mutex &devlink->lock_key stack_depot_init_mutex irq_context: 0 sb_writers#8 &of->mutex kn->active#52 nsim_bus_dev_list_lock &dev->mutex &devlink->lock_key rtnl_mutex bpf_devs_lock irq_context: 0 sb_writers#8 &of->mutex kn->active#52 nsim_bus_dev_list_lock &dev->mutex &devlink->lock_key rtnl_mutex bpf_devs_lock fs_reclaim irq_context: 0 sb_writers#8 &of->mutex kn->active#52 nsim_bus_dev_list_lock &dev->mutex &devlink->lock_key rtnl_mutex bpf_devs_lock fs_reclaim mmu_notifier_invalidate_range_start irq_context: 0 sb_writers#8 &of->mutex kn->active#52 nsim_bus_dev_list_lock &dev->mutex &devlink->lock_key rtnl_mutex bpf_devs_lock pool_lock#2 irq_context: 0 sb_writers#8 &of->mutex kn->active#52 nsim_bus_dev_list_lock &dev->mutex &devlink->lock_key rtnl_mutex bpf_devs_lock rcu_read_lock rhashtable_bucket irq_context: 0 sb_writers#8 &of->mutex kn->active#52 nsim_bus_dev_list_lock &dev->mutex &devlink->lock_key rtnl_mutex pin_fs_lock irq_context: 0 sb_writers#8 &of->mutex kn->active#52 nsim_bus_dev_list_lock &dev->mutex &devlink->lock_key rtnl_mutex &sb->s_type->i_mutex_key#3 irq_context: 0 sb_writers#8 &of->mutex kn->active#52 nsim_bus_dev_list_lock &dev->mutex &devlink->lock_key rtnl_mutex &sb->s_type->i_mutex_key#3 rename_lock.seqcount irq_context: 0 sb_writers#8 &of->mutex kn->active#52 nsim_bus_dev_list_lock &dev->mutex &devlink->lock_key rtnl_mutex &sb->s_type->i_mutex_key#3 fs_reclaim irq_context: 0 sb_writers#8 &of->mutex kn->active#52 nsim_bus_dev_list_lock &dev->mutex &devlink->lock_key rtnl_mutex &sb->s_type->i_mutex_key#3 fs_reclaim mmu_notifier_invalidate_range_start irq_context: 0 sb_writers#8 &of->mutex kn->active#52 nsim_bus_dev_list_lock &dev->mutex &devlink->lock_key rtnl_mutex &sb->s_type->i_mutex_key#3 pool_lock#2 irq_context: 0 sb_writers#8 &of->mutex kn->active#52 nsim_bus_dev_list_lock &dev->mutex &devlink->lock_key rtnl_mutex &sb->s_type->i_mutex_key#3 &dentry->d_lock irq_context: 0 sb_writers#8 &of->mutex kn->active#52 nsim_bus_dev_list_lock &dev->mutex &devlink->lock_key rtnl_mutex &sb->s_type->i_mutex_key#3 rcu_read_lock rename_lock.seqcount irq_context: 0 sb_writers#8 &of->mutex kn->active#52 nsim_bus_dev_list_lock &dev->mutex &devlink->lock_key rtnl_mutex &sb->s_type->i_mutex_key#3 &dentry->d_lock &wq irq_context: 0 sb_writers#8 &of->mutex kn->active#52 nsim_bus_dev_list_lock &dev->mutex &devlink->lock_key rtnl_mutex &sb->s_type->i_mutex_key#3 mmu_notifier_invalidate_range_start irq_context: 0 sb_writers#8 &of->mutex kn->active#52 nsim_bus_dev_list_lock &dev->mutex &devlink->lock_key rtnl_mutex &sb->s_type->i_mutex_key#3 &sb->s_type->i_lock_key#7 irq_context: 0 sb_writers#8 &of->mutex kn->active#52 nsim_bus_dev_list_lock &dev->mutex &devlink->lock_key rtnl_mutex &sb->s_type->i_mutex_key#3 &s->s_inode_list_lock irq_context: 0 sb_writers#8 &of->mutex kn->active#52 nsim_bus_dev_list_lock &dev->mutex &devlink->lock_key rtnl_mutex &sb->s_type->i_mutex_key#3 tk_core.seq.seqcount irq_context: 0 sb_writers#8 &of->mutex kn->active#52 nsim_bus_dev_list_lock &dev->mutex &devlink->lock_key rtnl_mutex &sb->s_type->i_mutex_key#3 &sb->s_type->i_lock_key#7 &dentry->d_lock irq_context: 0 sb_writers#8 &of->mutex kn->active#52 nsim_bus_dev_list_lock &dev->mutex &devlink->lock_key rtnl_mutex &sb->s_type->i_mutex_key#3 &c->lock irq_context: 0 sb_writers#8 &of->mutex kn->active#52 nsim_bus_dev_list_lock &dev->mutex &devlink->lock_key rtnl_mutex &sb->s_type->i_mutex_key#3 &n->list_lock irq_context: 0 sb_writers#8 &of->mutex kn->active#52 nsim_bus_dev_list_lock &dev->mutex &devlink->lock_key rtnl_mutex &sb->s_type->i_mutex_key#3 &n->list_lock &c->lock irq_context: 0 sb_writers#8 &of->mutex kn->active#52 nsim_bus_dev_list_lock &dev->mutex &devlink->lock_key rtnl_mutex &____s->seqcount irq_context: 0 sb_writers#8 &of->mutex kn->active#52 nsim_bus_dev_list_lock &dev->mutex &devlink->lock_key rtnl_mutex pool_lock#2 irq_context: 0 sb_writers#8 &of->mutex kn->active#52 nsim_bus_dev_list_lock &dev->mutex &devlink->lock_key rtnl_mutex rcu_read_lock pool_lock#2 irq_context: 0 sb_writers#8 &of->mutex kn->active#52 nsim_bus_dev_list_lock &dev->mutex &devlink->lock_key rtnl_mutex &obj_hash[i].lock irq_context: 0 sb_writers#8 &of->mutex kn->active#52 nsim_bus_dev_list_lock &dev->mutex &devlink->lock_key rtnl_mutex fs_reclaim irq_context: 0 sb_writers#8 &of->mutex kn->active#52 nsim_bus_dev_list_lock &dev->mutex &devlink->lock_key rtnl_mutex fs_reclaim mmu_notifier_invalidate_range_start irq_context: 0 sb_writers#8 &of->mutex kn->active#52 nsim_bus_dev_list_lock &dev->mutex &devlink->lock_key rtnl_mutex &xa->xa_lock#3 irq_context: 0 sb_writers#8 &of->mutex kn->active#52 nsim_bus_dev_list_lock &dev->mutex &devlink->lock_key rtnl_mutex &base->lock irq_context: 0 sb_writers#8 &of->mutex kn->active#52 nsim_bus_dev_list_lock &dev->mutex &devlink->lock_key rtnl_mutex &base->lock &obj_hash[i].lock irq_context: 0 sb_writers#8 &of->mutex kn->active#52 nsim_bus_dev_list_lock &dev->mutex &devlink->lock_key rtnl_mutex (work_completion)(&(&devlink_port->type_warn_dw)->work) irq_context: 0 sb_writers#8 &of->mutex kn->active#52 nsim_bus_dev_list_lock &dev->mutex &devlink->lock_key rtnl_mutex &devlink_port->type_lock irq_context: 0 sb_writers#8 &of->mutex kn->active#52 nsim_bus_dev_list_lock &dev->mutex &devlink->lock_key rtnl_mutex nl_table_lock irq_context: 0 sb_writers#8 &of->mutex kn->active#52 nsim_bus_dev_list_lock &dev->mutex &devlink->lock_key rtnl_mutex nl_table_wait.lock irq_context: 0 sb_writers#8 &of->mutex kn->active#52 nsim_bus_dev_list_lock &dev->mutex &devlink->lock_key rtnl_mutex net_rwsem irq_context: 0 sb_writers#8 &of->mutex kn->active#52 nsim_bus_dev_list_lock &dev->mutex &devlink->lock_key rtnl_mutex net_rwsem &list->lock#2 irq_context: 0 sb_writers#8 &of->mutex kn->active#52 nsim_bus_dev_list_lock &dev->mutex &devlink->lock_key rtnl_mutex &tn->lock irq_context: 0 sb_writers#8 &of->mutex kn->active#52 nsim_bus_dev_list_lock &dev->mutex &devlink->lock_key rtnl_mutex &x->wait#9 irq_context: 0 sb_writers#8 &of->mutex kn->active#52 nsim_bus_dev_list_lock &dev->mutex &devlink->lock_key rtnl_mutex &k->list_lock irq_context: 0 sb_writers#8 &of->mutex kn->active#52 nsim_bus_dev_list_lock &dev->mutex &devlink->lock_key rtnl_mutex gdp_mutex irq_context: 0 sb_writers#8 &of->mutex kn->active#52 nsim_bus_dev_list_lock &dev->mutex &devlink->lock_key rtnl_mutex gdp_mutex &k->list_lock irq_context: 0 sb_writers#8 &of->mutex kn->active#52 nsim_bus_dev_list_lock &dev->mutex &devlink->lock_key rtnl_mutex gdp_mutex fs_reclaim irq_context: 0 sb_writers#8 &of->mutex kn->active#52 nsim_bus_dev_list_lock &dev->mutex &devlink->lock_key rtnl_mutex gdp_mutex fs_reclaim mmu_notifier_invalidate_range_start irq_context: 0 sb_writers#8 &of->mutex kn->active#52 nsim_bus_dev_list_lock &dev->mutex &devlink->lock_key rtnl_mutex gdp_mutex pool_lock#2 irq_context: 0 sb_writers#8 &of->mutex kn->active#52 nsim_bus_dev_list_lock &dev->mutex &devlink->lock_key rtnl_mutex gdp_mutex lock irq_context: 0 sb_writers#8 &of->mutex kn->active#52 nsim_bus_dev_list_lock &dev->mutex &devlink->lock_key rtnl_mutex gdp_mutex lock kernfs_idr_lock irq_context: 0 sb_writers#8 &of->mutex kn->active#52 nsim_bus_dev_list_lock &dev->mutex &devlink->lock_key rtnl_mutex gdp_mutex &root->kernfs_rwsem irq_context: 0 sb_writers#8 &of->mutex kn->active#52 nsim_bus_dev_list_lock &dev->mutex &devlink->lock_key rtnl_mutex gdp_mutex &root->kernfs_rwsem &root->kernfs_iattr_rwsem irq_context: 0 sb_writers#8 &of->mutex kn->active#52 nsim_bus_dev_list_lock &dev->mutex &devlink->lock_key rtnl_mutex gdp_mutex kobj_ns_type_lock irq_context: 0 sb_writers#8 &of->mutex kn->active#52 nsim_bus_dev_list_lock &dev->mutex &devlink->lock_key rtnl_mutex lock irq_context: 0 sb_writers#8 &of->mutex kn->active#52 nsim_bus_dev_list_lock &dev->mutex &devlink->lock_key rtnl_mutex lock kernfs_idr_lock irq_context: 0 sb_writers#8 &of->mutex kn->active#52 nsim_bus_dev_list_lock &dev->mutex &devlink->lock_key rtnl_mutex &root->kernfs_rwsem irq_context: 0 sb_writers#8 &of->mutex kn->active#52 nsim_bus_dev_list_lock &dev->mutex &devlink->lock_key rtnl_mutex &root->kernfs_rwsem &root->kernfs_iattr_rwsem irq_context: 0 sb_writers#8 &of->mutex kn->active#52 nsim_bus_dev_list_lock &dev->mutex &devlink->lock_key rtnl_mutex bus_type_sem irq_context: 0 sb_writers#8 &of->mutex kn->active#52 nsim_bus_dev_list_lock &dev->mutex &devlink->lock_key rtnl_mutex sysfs_symlink_target_lock irq_context: 0 sb_writers#8 &of->mutex kn->active#52 nsim_bus_dev_list_lock &dev->mutex &devlink->lock_key rtnl_mutex lock kernfs_idr_lock pool_lock#2 irq_context: 0 sb_writers#8 &of->mutex kn->active#52 nsim_bus_dev_list_lock &dev->mutex &devlink->lock_key rtnl_mutex &c->lock irq_context: 0 sb_writers#8 &of->mutex kn->active#52 nsim_bus_dev_list_lock &dev->mutex &devlink->lock_key rtnl_mutex &root->kernfs_rwsem irq_context: 0 sb_writers#8 &of->mutex kn->active#52 nsim_bus_dev_list_lock &dev->mutex &devlink->lock_key rtnl_mutex &dev->power.lock irq_context: 0 sb_writers#8 &of->mutex kn->active#52 nsim_bus_dev_list_lock &dev->mutex &devlink->lock_key rtnl_mutex dpm_list_mtx irq_context: 0 sb_writers#8 &of->mutex kn->active#52 nsim_bus_dev_list_lock &dev->mutex &devlink->lock_key rtnl_mutex uevent_sock_mutex irq_context: 0 sb_writers#8 &of->mutex kn->active#52 nsim_bus_dev_list_lock &dev->mutex &devlink->lock_key rtnl_mutex uevent_sock_mutex fs_reclaim irq_context: 0 sb_writers#8 &of->mutex kn->active#52 nsim_bus_dev_list_lock &dev->mutex &devlink->lock_key rtnl_mutex uevent_sock_mutex fs_reclaim mmu_notifier_invalidate_range_start irq_context: 0 sb_writers#8 &of->mutex kn->active#52 nsim_bus_dev_list_lock &dev->mutex &devlink->lock_key rtnl_mutex uevent_sock_mutex pool_lock#2 irq_context: 0 sb_writers#8 &of->mutex kn->active#52 nsim_bus_dev_list_lock &dev->mutex &devlink->lock_key rtnl_mutex uevent_sock_mutex nl_table_lock irq_context: 0 sb_writers#8 &of->mutex kn->active#52 nsim_bus_dev_list_lock &dev->mutex &devlink->lock_key rtnl_mutex uevent_sock_mutex &obj_hash[i].lock irq_context: 0 sb_writers#8 &of->mutex kn->active#52 nsim_bus_dev_list_lock &dev->mutex &devlink->lock_key rtnl_mutex uevent_sock_mutex nl_table_wait.lock irq_context: 0 sb_writers#8 &of->mutex kn->active#52 nsim_bus_dev_list_lock &dev->mutex &devlink->lock_key rtnl_mutex &k->k_lock irq_context: 0 sb_writers#8 &of->mutex kn->active#52 nsim_bus_dev_list_lock &dev->mutex &devlink->lock_key rtnl_mutex subsys mutex#17 irq_context: 0 sb_writers#8 &of->mutex kn->active#52 nsim_bus_dev_list_lock &dev->mutex &devlink->lock_key rtnl_mutex subsys mutex#17 &k->k_lock irq_context: 0 sb_writers#8 &of->mutex kn->active#52 nsim_bus_dev_list_lock &dev->mutex &devlink->lock_key rtnl_mutex &dir->lock#2 irq_context: 0 sb_writers#8 &of->mutex kn->active#52 nsim_bus_dev_list_lock &dev->mutex &devlink->lock_key rtnl_mutex dev_hotplug_mutex irq_context: 0 sb_writers#8 &of->mutex kn->active#52 nsim_bus_dev_list_lock &dev->mutex &devlink->lock_key rtnl_mutex dev_hotplug_mutex &dev->power.lock irq_context: 0 sb_writers#8 &of->mutex kn->active#52 nsim_bus_dev_list_lock &dev->mutex &devlink->lock_key rtnl_mutex dev_base_lock irq_context: 0 sb_writers#8 &of->mutex kn->active#52 nsim_bus_dev_list_lock &dev->mutex &devlink->lock_key rtnl_mutex input_pool.lock irq_context: 0 sb_writers#8 &of->mutex kn->active#52 nsim_bus_dev_list_lock &dev->mutex &devlink->lock_key rtnl_mutex rcu_read_lock &pool->lock/1 irq_context: 0 sb_writers#8 &of->mutex kn->active#52 nsim_bus_dev_list_lock &dev->mutex &devlink->lock_key rtnl_mutex rcu_read_lock &pool->lock/1 &obj_hash[i].lock irq_context: 0 sb_writers#8 &of->mutex kn->active#52 nsim_bus_dev_list_lock &dev->mutex &devlink->lock_key rtnl_mutex rcu_read_lock &pool->lock/1 pool_lock#2 irq_context: 0 sb_writers#8 &of->mutex kn->active#52 nsim_bus_dev_list_lock &dev->mutex &devlink->lock_key rtnl_mutex rcu_read_lock &pool->lock/1 &p->pi_lock irq_context: 0 sb_writers#8 &of->mutex kn->active#52 nsim_bus_dev_list_lock &dev->mutex &devlink->lock_key rtnl_mutex batched_entropy_u32.lock irq_context: 0 sb_writers#8 &of->mutex kn->active#52 nsim_bus_dev_list_lock &dev->mutex &devlink->lock_key rtnl_mutex &tbl->lock irq_context: 0 sb_writers#8 &of->mutex kn->active#52 nsim_bus_dev_list_lock &dev->mutex &devlink->lock_key rtnl_mutex sysctl_lock irq_context: 0 sb_writers#8 &of->mutex kn->active#52 nsim_bus_dev_list_lock &dev->mutex &devlink->lock_key rtnl_mutex &n->list_lock irq_context: 0 sb_writers#8 &of->mutex kn->active#52 nsim_bus_dev_list_lock &dev->mutex &devlink->lock_key rtnl_mutex &n->list_lock &c->lock irq_context: 0 sb_writers#8 &of->mutex kn->active#52 nsim_bus_dev_list_lock &dev->mutex &devlink->lock_key rtnl_mutex failover_lock irq_context: 0 sb_writers#8 &of->mutex kn->active#52 nsim_bus_dev_list_lock &dev->mutex &devlink->lock_key rtnl_mutex pcpu_alloc_mutex irq_context: 0 sb_writers#8 &of->mutex kn->active#52 nsim_bus_dev_list_lock &dev->mutex &devlink->lock_key rtnl_mutex pcpu_alloc_mutex pcpu_lock irq_context: 0 sb_writers#8 &of->mutex kn->active#52 nsim_bus_dev_list_lock &dev->mutex &devlink->lock_key rtnl_mutex proc_subdir_lock irq_context: 0 sb_writers#8 &of->mutex kn->active#52 nsim_bus_dev_list_lock &dev->mutex &devlink->lock_key rtnl_mutex proc_inum_ida.xa_lock irq_context: 0 sb_writers#8 &of->mutex kn->active#52 nsim_bus_dev_list_lock &dev->mutex &devlink->lock_key rtnl_mutex proc_subdir_lock irq_context: 0 sb_writers#8 &of->mutex kn->active#52 nsim_bus_dev_list_lock &dev->mutex &devlink->lock_key rtnl_mutex &obj_hash[i].lock pool_lock irq_context: 0 sb_writers#8 &of->mutex kn->active#52 nsim_bus_dev_list_lock &dev->mutex &devlink->lock_key rtnl_mutex fill_pool_map-wait-type-override pool_lock#2 irq_context: 0 sb_writers#8 &of->mutex kn->active#52 nsim_bus_dev_list_lock &dev->mutex &devlink->lock_key rtnl_mutex fill_pool_map-wait-type-override pool_lock irq_context: 0 sb_writers#8 &of->mutex kn->active#52 nsim_bus_dev_list_lock &dev->mutex &devlink->lock_key rtnl_mutex &idev->mc_lock irq_context: 0 sb_writers#8 &of->mutex kn->active#52 nsim_bus_dev_list_lock &dev->mutex &devlink->lock_key rtnl_mutex &idev->mc_lock fs_reclaim irq_context: 0 sb_writers#8 &of->mutex kn->active#52 nsim_bus_dev_list_lock &dev->mutex &devlink->lock_key rtnl_mutex &idev->mc_lock fs_reclaim mmu_notifier_invalidate_range_start irq_context: 0 sb_writers#8 &of->mutex kn->active#52 nsim_bus_dev_list_lock &dev->mutex &devlink->lock_key rtnl_mutex &idev->mc_lock pool_lock#2 irq_context: 0 sb_writers#8 &of->mutex kn->active#52 nsim_bus_dev_list_lock &dev->mutex &devlink->lock_key rtnl_mutex &idev->mc_lock &obj_hash[i].lock irq_context: 0 sb_writers#8 &of->mutex kn->active#52 nsim_bus_dev_list_lock &dev->mutex &devlink->lock_key rtnl_mutex &idev->mc_lock _xmit_ETHER irq_context: 0 sb_writers#8 &of->mutex kn->active#52 nsim_bus_dev_list_lock &dev->mutex &devlink->lock_key rtnl_mutex &idev->mc_lock _xmit_ETHER pool_lock#2 irq_context: 0 sb_writers#8 &of->mutex kn->active#52 nsim_bus_dev_list_lock &dev->mutex &devlink->lock_key rtnl_mutex &pnettable->lock irq_context: 0 sb_writers#8 &of->mutex kn->active#52 nsim_bus_dev_list_lock &dev->mutex &devlink->lock_key rtnl_mutex smc_ib_devices.mutex irq_context: 0 sb_writers#8 &of->mutex kn->active#52 nsim_bus_dev_list_lock &dev->mutex &devlink->lock_key rtnl_mutex &vn->sock_lock irq_context: 0 sb_writers#8 &of->mutex kn->active#52 nsim_bus_dev_list_lock &dev->mutex &devlink->lock_key rtnl_mutex rcu_read_lock &ndev->lock irq_context: 0 (wq_completion)events (work_completion)(&(&nsim_dev->trap_data->trap_report_dw)->work) rcu_read_lock &pool->lock fill_pool_map-wait-type-override &c->lock irq_context: 0 (wq_completion)events (work_completion)(&(&nsim_dev->trap_data->trap_report_dw)->work) rcu_read_lock &pool->lock fill_pool_map-wait-type-override &____s->seqcount irq_context: 0 kn->active#51 &c->lock irq_context: 0 sb_writers#8 &of->mutex kn->active#52 nsim_bus_dev_list_lock &dev->mutex &devlink->lock_key rtnl_mutex uevent_sock_mutex &c->lock irq_context: 0 sb_writers#8 &of->mutex kn->active#52 nsim_bus_dev_list_lock &dev->mutex &devlink->lock_key rtnl_mutex rcu_read_lock &pool->lock/1 &p->pi_lock &rq->__lock irq_context: 0 sb_writers#8 &of->mutex kn->active#52 nsim_bus_dev_list_lock &dev->mutex &devlink->lock_key rtnl_mutex rcu_read_lock &pool->lock/1 &p->pi_lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 sb_writers#8 &of->mutex kn->active#52 nsim_bus_dev_list_lock &dev->mutex &devlink->lock_key rtnl_mutex fill_pool_map-wait-type-override &c->lock irq_context: 0 sb_writers#8 &of->mutex kn->active#52 nsim_bus_dev_list_lock &dev->mutex &devlink->lock_key rtnl_mutex fill_pool_map-wait-type-override &____s->seqcount irq_context: 0 sb_writers#8 &of->mutex kn->active#52 nsim_bus_dev_list_lock &dev->mutex &devlink->lock_key &sb->s_type->i_mutex_key#3 &pcp->lock &zone->lock irq_context: 0 sb_writers#8 &of->mutex kn->active#52 nsim_bus_dev_list_lock &dev->mutex &devlink->lock_key &sb->s_type->i_mutex_key#3 &pcp->lock &zone->lock &____s->seqcount irq_context: softirq (&app->join_timer)#2 batched_entropy_u32.lock crngs.lock irq_context: 0 sb_writers#8 &of->mutex kn->active#52 nsim_bus_dev_list_lock &dev->mutex &devlink->lock_key rtnl_mutex &sb->s_type->i_mutex_key#3 &____s->seqcount irq_context: 0 sb_writers#8 &of->mutex kn->active#52 nsim_bus_dev_list_lock &dev->mutex &devlink->lock_key rtnl_mutex uevent_sock_mutex quarantine_lock irq_context: 0 sb_writers#8 &of->mutex kn->active#52 nsim_bus_dev_list_lock &dev->mutex &devlink->lock_key rtnl_mutex remove_cache_srcu irq_context: 0 sb_writers#8 &of->mutex kn->active#52 nsim_bus_dev_list_lock &dev->mutex &devlink->lock_key rtnl_mutex remove_cache_srcu quarantine_lock irq_context: 0 sb_writers#8 &of->mutex kn->active#52 nsim_bus_dev_list_lock &dev->mutex &devlink->lock_key rtnl_mutex remove_cache_srcu &c->lock irq_context: 0 sb_writers#8 &of->mutex kn->active#51 nsim_bus_dev_list_lock &rq->__lock &cfs_rq->removed.lock irq_context: 0 sb_writers#8 &of->mutex kn->active#52 nsim_bus_dev_list_lock &dev->mutex &devlink->lock_key rtnl_mutex remove_cache_srcu &rq->__lock irq_context: 0 sb_writers#8 &of->mutex kn->active#52 nsim_bus_dev_list_lock &dev->mutex &devlink->lock_key rtnl_mutex remove_cache_srcu &n->list_lock irq_context: 0 sb_writers#8 &of->mutex kn->active#52 nsim_bus_dev_list_lock &dev->mutex &devlink->lock_key rtnl_mutex remove_cache_srcu rcu_read_lock pool_lock#2 irq_context: 0 sb_writers#8 &of->mutex kn->active#52 nsim_bus_dev_list_lock &dev->mutex &devlink->lock_key rtnl_mutex remove_cache_srcu &obj_hash[i].lock irq_context: 0 sb_writers#8 &of->mutex kn->active#52 nsim_bus_dev_list_lock &dev->mutex &devlink->lock_key rtnl_mutex batched_entropy_u32.lock crngs.lock irq_context: 0 sb_writers#8 &of->mutex kn->active#52 nsim_bus_dev_list_lock &dev->mutex &devlink->lock_key &p->pi_lock &cfs_rq->removed.lock irq_context: 0 sb_writers#8 &of->mutex kn->active#52 nsim_bus_dev_list_lock &dev->mutex deferred_probe_mutex irq_context: 0 sb_writers#8 &of->mutex kn->active#52 nsim_bus_dev_list_lock &dev->mutex rcu_read_lock &pool->lock/1 irq_context: 0 sb_writers#8 &of->mutex kn->active#52 nsim_bus_dev_list_lock &dev->mutex rcu_read_lock &pool->lock/1 &obj_hash[i].lock irq_context: 0 sb_writers#8 &of->mutex kn->active#52 nsim_bus_dev_list_lock &dev->mutex rcu_read_lock &pool->lock/1 &p->pi_lock irq_context: 0 sb_writers#8 &of->mutex kn->active#52 nsim_bus_dev_list_lock &dev->mutex rcu_read_lock &pool->lock/1 &p->pi_lock &rq->__lock irq_context: 0 sb_writers#8 &of->mutex kn->active#52 nsim_bus_dev_list_lock &dev->mutex rcu_read_lock &pool->lock/1 &p->pi_lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 sb_writers#8 &of->mutex kn->active#52 nsim_bus_dev_list_lock &dev->mutex uevent_sock_mutex irq_context: 0 sb_writers#8 &of->mutex kn->active#52 nsim_bus_dev_list_lock &dev->mutex uevent_sock_mutex fs_reclaim irq_context: 0 sb_writers#8 &of->mutex kn->active#52 nsim_bus_dev_list_lock &dev->mutex uevent_sock_mutex fs_reclaim mmu_notifier_invalidate_range_start irq_context: 0 sb_writers#8 &of->mutex kn->active#52 nsim_bus_dev_list_lock &dev->mutex uevent_sock_mutex pool_lock#2 irq_context: 0 sb_writers#8 &of->mutex kn->active#52 nsim_bus_dev_list_lock &dev->mutex uevent_sock_mutex &c->lock irq_context: 0 sb_writers#8 &of->mutex kn->active#52 nsim_bus_dev_list_lock &dev->mutex uevent_sock_mutex nl_table_lock irq_context: 0 sb_writers#8 &of->mutex kn->active#52 nsim_bus_dev_list_lock &dev->mutex uevent_sock_mutex rlock-AF_NETLINK irq_context: 0 sb_writers#8 &of->mutex kn->active#52 nsim_bus_dev_list_lock &dev->mutex uevent_sock_mutex rcu_read_lock &ei->socket.wq.wait irq_context: 0 sb_writers#8 &of->mutex kn->active#52 nsim_bus_dev_list_lock &dev->mutex uevent_sock_mutex rcu_read_lock &ei->socket.wq.wait &ep->lock irq_context: 0 sb_writers#8 &of->mutex kn->active#52 nsim_bus_dev_list_lock &dev->mutex uevent_sock_mutex rcu_read_lock &ei->socket.wq.wait &ep->lock &ep->wq irq_context: 0 sb_writers#8 &of->mutex kn->active#52 nsim_bus_dev_list_lock &dev->mutex uevent_sock_mutex rcu_read_lock &ei->socket.wq.wait &ep->lock &ep->wq &p->pi_lock irq_context: 0 sb_writers#8 &of->mutex kn->active#52 nsim_bus_dev_list_lock &dev->mutex uevent_sock_mutex rcu_read_lock &ei->socket.wq.wait &ep->lock &ep->wq &p->pi_lock &rq->__lock irq_context: 0 sb_writers#8 &of->mutex kn->active#52 nsim_bus_dev_list_lock &dev->mutex uevent_sock_mutex rcu_read_lock &ei->socket.wq.wait &ep->lock &ep->wq &p->pi_lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 sb_writers#8 &of->mutex kn->active#52 nsim_bus_dev_list_lock &dev->mutex uevent_sock_mutex rcu_read_lock rcu_node_0 irq_context: 0 sb_writers#8 &of->mutex kn->active#52 nsim_bus_dev_list_lock &dev->mutex uevent_sock_mutex rcu_read_lock &rq->__lock irq_context: 0 rtnl_mutex uevent_sock_mutex uevent_sock_mutex.wait_lock irq_context: 0 (wq_completion)events (work_completion)(&(&nsim_dev->trap_data->trap_report_dw)->work) &obj_hash[i].lock irq_context: 0 rtnl_mutex uevent_sock_mutex rcu_read_lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)events (work_completion)(&(&nsim_dev->trap_data->trap_report_dw)->work) &base->lock irq_context: 0 (wq_completion)events (work_completion)(&(&nsim_dev->trap_data->trap_report_dw)->work) &base->lock &obj_hash[i].lock irq_context: 0 sb_writers#8 &of->mutex kn->active#52 nsim_bus_dev_list_lock &dev->mutex uevent_sock_mutex nl_table_wait.lock irq_context: softirq (&lapb->t1timer) &lapb->lock &list->lock#13 irq_context: softirq (&lapb->t1timer) &lapb->lock &list->lock#14 irq_context: softirq &list->lock#14 irq_context: softirq rcu_read_lock x25_neigh_list_lock irq_context: softirq rcu_read_lock &list->lock#15 irq_context: softirq rcu_read_lock x25_list_lock irq_context: softirq rcu_read_lock x25_forward_list_lock irq_context: 0 sb_writers#8 &of->mutex kn->active#52 nsim_bus_dev_list_lock &dev->mutex uevent_sock_mutex &obj_hash[i].lock irq_context: 0 sb_writers#8 &of->mutex kn->active#52 nsim_bus_dev_list_lock &dev->mutex uevent_sock_mutex.wait_lock irq_context: 0 sb_writers#8 &of->mutex kn->active#52 nsim_bus_dev_list_lock &dev->mutex &p->pi_lock irq_context: 0 sb_writers#8 &of->mutex kn->active#52 nsim_bus_dev_list_lock &dev->mutex &p->pi_lock &rq->__lock irq_context: 0 sb_writers#8 &of->mutex kn->active#52 nsim_bus_dev_list_lock &dev->mutex &p->pi_lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 sb_writers#8 &of->mutex kn->active#52 nsim_bus_dev_list_lock &dev->mutex probe_waitqueue.lock irq_context: 0 sb_writers#8 &of->mutex kn->active#52 nsim_bus_dev_list_lock subsys mutex#83 irq_context: 0 sb_writers#8 &of->mutex kn->active#52 nsim_bus_dev_list_lock.wait_lock irq_context: 0 sb_writers#8 &of->mutex kn->active#52 &p->pi_lock irq_context: 0 sb_writers#8 &of->mutex kn->active#52 &p->pi_lock &rq->__lock irq_context: 0 sb_writers#8 &of->mutex kn->active#52 &p->pi_lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 cb_lock &rq->__lock irq_context: 0 cb_lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 sb_writers#8 &of->mutex kn->active#51 nsim_bus_dev_list_lock.wait_lock irq_context: 0 sb_writers#8 &of->mutex kn->active#51 &p->pi_lock irq_context: 0 sb_writers#8 &of->mutex kn->active#51 &p->pi_lock &rq->__lock irq_context: 0 sb_writers#8 &of->mutex kn->active#51 &p->pi_lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 sb_writers#8 &of->mutex kn->active#51 &p->pi_lock &rq->__lock &cfs_rq->removed.lock irq_context: 0 sb_writers#8 &of->mutex kn->active#52 nsim_bus_dev_list_lock &dev->mutex &c->lock irq_context: 0 sb_writers#8 &of->mutex kn->active#52 nsim_bus_dev_list_lock &dev->mutex &____s->seqcount irq_context: 0 sb_writers#8 &of->mutex kn->active#52 nsim_bus_dev_list_lock &dev->mutex devlinks.xa_lock &c->lock irq_context: 0 sb_writers#8 &of->mutex kn->active#52 nsim_bus_dev_list_lock &dev->mutex devlinks.xa_lock &____s->seqcount irq_context: 0 sb_writers#8 &of->mutex kn->active#52 nsim_bus_dev_list_lock &dev->mutex devlinks.xa_lock pool_lock#2 irq_context: 0 sb_writers#8 &of->mutex kn->active#52 nsim_bus_dev_list_lock &dev->mutex &devlink->lock_key#2 irq_context: 0 sb_writers#8 &of->mutex kn->active#52 nsim_bus_dev_list_lock &dev->mutex &devlink->lock_key#2 crngs.lock irq_context: 0 sb_writers#8 &of->mutex kn->active#52 nsim_bus_dev_list_lock &dev->mutex &devlink->lock_key#2 fs_reclaim irq_context: 0 sb_writers#8 &of->mutex kn->active#52 nsim_bus_dev_list_lock &dev->mutex &devlink->lock_key#2 fs_reclaim mmu_notifier_invalidate_range_start irq_context: 0 sb_writers#8 &of->mutex kn->active#52 nsim_bus_dev_list_lock &dev->mutex &devlink->lock_key#2 devlinks.xa_lock irq_context: 0 sb_writers#8 &of->mutex kn->active#52 nsim_bus_dev_list_lock &dev->mutex &devlink->lock_key#2 &c->lock irq_context: 0 sb_writers#8 &of->mutex kn->active#52 nsim_bus_dev_list_lock &dev->mutex &devlink->lock_key#2 &obj_hash[i].lock irq_context: 0 sb_writers#8 &of->mutex kn->active#52 nsim_bus_dev_list_lock &dev->mutex &devlink->lock_key#2 nl_table_lock irq_context: 0 sb_writers#8 &of->mutex kn->active#52 nsim_bus_dev_list_lock &dev->mutex &devlink->lock_key#2 nl_table_wait.lock irq_context: 0 sb_writers#8 &of->mutex kn->active#52 nsim_bus_dev_list_lock &dev->mutex &devlink->lock_key#2 &____s->seqcount irq_context: 0 sb_writers#8 &of->mutex kn->active#52 nsim_bus_dev_list_lock &dev->mutex &devlink->lock_key#2 pool_lock#2 irq_context: 0 sb_writers#8 &of->mutex kn->active#52 nsim_bus_dev_list_lock &dev->mutex &devlink->lock_key#2 &xa->xa_lock#16 irq_context: 0 sb_writers#8 &of->mutex kn->active#52 nsim_bus_dev_list_lock &dev->mutex &devlink->lock_key#2 &pcp->lock &zone->lock irq_context: 0 sb_writers#8 &of->mutex kn->active#52 nsim_bus_dev_list_lock &dev->mutex &devlink->lock_key#2 &pcp->lock &zone->lock &____s->seqcount irq_context: 0 sb_writers#8 &of->mutex kn->active#52 nsim_bus_dev_list_lock &dev->mutex &devlink->lock_key#2 pcpu_alloc_mutex irq_context: 0 sb_writers#8 &of->mutex kn->active#52 nsim_bus_dev_list_lock &dev->mutex &devlink->lock_key#2 pcpu_alloc_mutex pcpu_lock irq_context: 0 sb_writers#8 &of->mutex kn->active#52 nsim_bus_dev_list_lock &dev->mutex &devlink->lock_key#2 &n->list_lock irq_context: 0 sb_writers#8 &of->mutex kn->active#52 nsim_bus_dev_list_lock &dev->mutex &devlink->lock_key#2 &n->list_lock &c->lock irq_context: 0 sb_writers#5 &type->i_mutex_dir_key#5/1 fill_pool_map-wait-type-override pool_lock#2 irq_context: 0 sb_writers#5 &type->i_mutex_dir_key#5/1 fill_pool_map-wait-type-override pool_lock irq_context: 0 sb_writers#8 &of->mutex kn->active#52 nsim_bus_dev_list_lock &dev->mutex &devlink->lock_key#2 &rq->__lock irq_context: 0 sb_writers#8 &of->mutex kn->active#52 nsim_bus_dev_list_lock &dev->mutex &devlink->lock_key#2 &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 cb_lock nlk_cb_mutex-GENERIC &devlink->lock_key irq_context: 0 cb_lock nlk_cb_mutex-GENERIC &devlink->lock_key rcu_node_0 irq_context: 0 cb_lock nlk_cb_mutex-GENERIC &devlink->lock_key &rcu_state.expedited_wq irq_context: 0 cb_lock nlk_cb_mutex-GENERIC &devlink->lock_key &rcu_state.expedited_wq &p->pi_lock irq_context: 0 cb_lock nlk_cb_mutex-GENERIC &devlink->lock_key &rcu_state.expedited_wq &p->pi_lock &rq->__lock irq_context: 0 cb_lock nlk_cb_mutex-GENERIC &devlink->lock_key &rcu_state.expedited_wq &p->pi_lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 cb_lock nlk_cb_mutex-GENERIC &devlink->lock_key &rq->__lock irq_context: 0 cb_lock nlk_cb_mutex-GENERIC &devlink->lock_key &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 cb_lock nlk_cb_mutex-GENERIC &devlink->lock_key &devlink_port->type_lock irq_context: 0 sb_writers#8 &of->mutex kn->active#52 nsim_bus_dev_list_lock &dev->mutex &devlink->lock_key#2 &base->lock irq_context: 0 sb_writers#8 &of->mutex kn->active#52 nsim_bus_dev_list_lock &dev->mutex &devlink->lock_key#2 &base->lock &obj_hash[i].lock irq_context: 0 sb_writers#8 &of->mutex kn->active#52 nsim_bus_dev_list_lock &dev->mutex &devlink->lock_key#2 pin_fs_lock irq_context: 0 sb_writers#8 &of->mutex kn->active#52 nsim_bus_dev_list_lock &dev->mutex &devlink->lock_key#2 &sb->s_type->i_mutex_key#3 irq_context: 0 sb_writers#8 &of->mutex kn->active#52 nsim_bus_dev_list_lock &dev->mutex &devlink->lock_key#2 &sb->s_type->i_mutex_key#3 rename_lock.seqcount irq_context: 0 sb_writers#8 &of->mutex kn->active#52 nsim_bus_dev_list_lock &dev->mutex &devlink->lock_key#2 &sb->s_type->i_mutex_key#3 fs_reclaim irq_context: 0 sb_writers#8 &of->mutex kn->active#52 nsim_bus_dev_list_lock &dev->mutex &devlink->lock_key#2 &sb->s_type->i_mutex_key#3 fs_reclaim mmu_notifier_invalidate_range_start irq_context: 0 sb_writers#8 &of->mutex kn->active#52 nsim_bus_dev_list_lock &dev->mutex &devlink->lock_key#2 &sb->s_type->i_mutex_key#3 &c->lock irq_context: 0 sb_writers#8 &of->mutex kn->active#52 nsim_bus_dev_list_lock &dev->mutex &devlink->lock_key#2 &sb->s_type->i_mutex_key#3 &n->list_lock irq_context: 0 sb_writers#8 &of->mutex kn->active#52 nsim_bus_dev_list_lock &dev->mutex &devlink->lock_key#2 &sb->s_type->i_mutex_key#3 &n->list_lock &c->lock irq_context: 0 sb_writers#8 &of->mutex kn->active#52 nsim_bus_dev_list_lock &dev->mutex &devlink->lock_key#2 &sb->s_type->i_mutex_key#3 &dentry->d_lock irq_context: 0 sb_writers#8 &of->mutex kn->active#52 nsim_bus_dev_list_lock &dev->mutex &devlink->lock_key#2 &sb->s_type->i_mutex_key#3 rcu_read_lock rename_lock.seqcount irq_context: 0 sb_writers#8 &of->mutex kn->active#52 nsim_bus_dev_list_lock &dev->mutex &devlink->lock_key#2 &sb->s_type->i_mutex_key#3 &dentry->d_lock &wq irq_context: 0 sb_writers#8 &of->mutex kn->active#52 nsim_bus_dev_list_lock &dev->mutex &devlink->lock_key#2 &sb->s_type->i_mutex_key#3 mmu_notifier_invalidate_range_start irq_context: 0 sb_writers#8 &of->mutex kn->active#52 nsim_bus_dev_list_lock &dev->mutex &devlink->lock_key#2 &sb->s_type->i_mutex_key#3 &sb->s_type->i_lock_key#7 irq_context: 0 sb_writers#8 &of->mutex kn->active#52 nsim_bus_dev_list_lock &dev->mutex &devlink->lock_key#2 &sb->s_type->i_mutex_key#3 &s->s_inode_list_lock irq_context: 0 sb_writers#8 &of->mutex kn->active#52 nsim_bus_dev_list_lock &dev->mutex &devlink->lock_key#2 &sb->s_type->i_mutex_key#3 tk_core.seq.seqcount irq_context: 0 sb_writers#8 &of->mutex kn->active#52 nsim_bus_dev_list_lock &dev->mutex &devlink->lock_key#2 &sb->s_type->i_mutex_key#3 &sb->s_type->i_lock_key#7 &dentry->d_lock irq_context: 0 sb_writers#8 &of->mutex kn->active#52 nsim_bus_dev_list_lock &dev->mutex &devlink->lock_key#2 &sb->s_type->i_mutex_key#3 &____s->seqcount irq_context: 0 sb_writers#8 &of->mutex kn->active#52 nsim_bus_dev_list_lock &dev->mutex &devlink->lock_key#2 &sb->s_type->i_mutex_key#3 pool_lock#2 irq_context: 0 sb_writers#8 &of->mutex kn->active#52 nsim_bus_dev_list_lock &dev->mutex &devlink->lock_key#2 batched_entropy_u32.lock irq_context: 0 sb_writers#8 &of->mutex kn->active#52 nsim_bus_dev_list_lock &dev->mutex &devlink->lock_key#2 rtnl_mutex irq_context: 0 sb_writers#8 &of->mutex kn->active#52 nsim_bus_dev_list_lock &dev->mutex &devlink->lock_key#2 rtnl_mutex &rq->__lock irq_context: 0 sb_writers#8 &of->mutex kn->active#52 nsim_bus_dev_list_lock &dev->mutex &devlink->lock_key#2 rtnl_mutex rtnl_mutex.wait_lock irq_context: 0 sb_writers#8 &of->mutex kn->active#52 nsim_bus_dev_list_lock &dev->mutex &devlink->lock_key#2 rtnl_mutex &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 rtnl_mutex devnet_rename_sem irq_context: 0 rtnl_mutex devnet_rename_sem (console_sem).lock irq_context: 0 rtnl_mutex devnet_rename_sem console_lock console_srcu console_owner_lock irq_context: 0 rtnl_mutex devnet_rename_sem console_lock console_srcu console_owner irq_context: 0 rtnl_mutex devnet_rename_sem console_lock console_srcu console_owner &port_lock_key irq_context: 0 rtnl_mutex devnet_rename_sem console_lock console_srcu console_owner console_owner_lock irq_context: 0 rtnl_mutex devnet_rename_sem fs_reclaim irq_context: 0 rtnl_mutex devnet_rename_sem fs_reclaim mmu_notifier_invalidate_range_start irq_context: 0 rtnl_mutex devnet_rename_sem pool_lock#2 irq_context: 0 rtnl_mutex devnet_rename_sem &k->list_lock irq_context: 0 rtnl_mutex devnet_rename_sem &root->kernfs_rwsem irq_context: 0 rtnl_mutex devnet_rename_sem &root->kernfs_rwsem irq_context: 0 rtnl_mutex devnet_rename_sem &root->kernfs_rwsem fs_reclaim irq_context: 0 rtnl_mutex devnet_rename_sem &root->kernfs_rwsem fs_reclaim mmu_notifier_invalidate_range_start irq_context: 0 rtnl_mutex devnet_rename_sem &root->kernfs_rwsem pool_lock#2 irq_context: 0 rtnl_mutex devnet_rename_sem &root->kernfs_rwsem kernfs_rename_lock irq_context: 0 rtnl_mutex devnet_rename_sem &root->kernfs_rwsem &obj_hash[i].lock irq_context: 0 rtnl_mutex devnet_rename_sem &c->lock irq_context: 0 rtnl_mutex devnet_rename_sem &____s->seqcount irq_context: 0 rtnl_mutex devnet_rename_sem kernfs_rename_lock irq_context: 0 rtnl_mutex devnet_rename_sem uevent_sock_mutex irq_context: 0 rtnl_mutex devnet_rename_sem uevent_sock_mutex fs_reclaim irq_context: 0 rtnl_mutex devnet_rename_sem uevent_sock_mutex fs_reclaim mmu_notifier_invalidate_range_start irq_context: 0 rtnl_mutex devnet_rename_sem uevent_sock_mutex pool_lock#2 irq_context: 0 rtnl_mutex devnet_rename_sem uevent_sock_mutex nl_table_lock irq_context: 0 rtnl_mutex devnet_rename_sem uevent_sock_mutex &obj_hash[i].lock irq_context: 0 rtnl_mutex devnet_rename_sem uevent_sock_mutex nl_table_wait.lock irq_context: 0 rtnl_mutex devnet_rename_sem &obj_hash[i].lock irq_context: 0 rtnl_mutex (work_completion)(&(&devlink_port->type_warn_dw)->work) irq_context: 0 rtnl_mutex &devlink_port->type_lock irq_context: 0 rtnl_mutex sysctl_lock &obj_hash[i].lock irq_context: 0 rtnl_mutex sysctl_lock pool_lock#2 irq_context: 0 rtnl_mutex sysctl_lock krc.lock irq_context: 0 rtnl_mutex &nft_net->commit_mutex irq_context: 0 rtnl_mutex &ent->pde_unload_lock irq_context: 0 sb_writers#8 &of->mutex kn->active#52 nsim_bus_dev_list_lock &dev->mutex &devlink->lock_key#2 rtnl_mutex &(&net->nexthop.notifier_chain)->rwsem irq_context: 0 sb_writers#8 &of->mutex kn->active#52 nsim_bus_dev_list_lock &dev->mutex &devlink->lock_key#2 rtnl_mutex.wait_lock irq_context: 0 sb_writers#8 &of->mutex kn->active#52 nsim_bus_dev_list_lock &dev->mutex &devlink->lock_key#2 &p->pi_lock irq_context: 0 sb_writers#8 &of->mutex kn->active#52 nsim_bus_dev_list_lock &dev->mutex &devlink->lock_key#2 &p->pi_lock &rq->__lock irq_context: 0 sb_writers#8 &of->mutex kn->active#52 nsim_bus_dev_list_lock &dev->mutex &devlink->lock_key#2 &p->pi_lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: softirq &(&hwstats->traffic_dw)->timer irq_context: softirq &(&hwstats->traffic_dw)->timer rcu_read_lock &pool->lock irq_context: softirq &(&hwstats->traffic_dw)->timer rcu_read_lock &pool->lock &obj_hash[i].lock irq_context: 0 rtnl_mutex devnet_rename_sem &rq->__lock irq_context: softirq &(&hwstats->traffic_dw)->timer rcu_read_lock &pool->lock &p->pi_lock irq_context: softirq &(&hwstats->traffic_dw)->timer rcu_read_lock &pool->lock &p->pi_lock &rq->__lock irq_context: softirq &(&hwstats->traffic_dw)->timer rcu_read_lock &pool->lock &p->pi_lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)events (work_completion)(&(&hwstats->traffic_dw)->work) irq_context: 0 (wq_completion)events (work_completion)(&(&hwstats->traffic_dw)->work) &hwstats->hwsdev_list_lock irq_context: 0 (wq_completion)events (work_completion)(&(&hwstats->traffic_dw)->work) &obj_hash[i].lock irq_context: 0 (wq_completion)events (work_completion)(&(&hwstats->traffic_dw)->work) &base->lock irq_context: 0 (wq_completion)events (work_completion)(&(&hwstats->traffic_dw)->work) &base->lock &obj_hash[i].lock irq_context: 0 rtnl_mutex sysctl_lock &obj_hash[i].lock pool_lock irq_context: 0 rtnl_mutex sysctl_lock fill_pool_map-wait-type-override pool_lock#2 irq_context: 0 rtnl_mutex sysctl_lock fill_pool_map-wait-type-override &c->lock irq_context: 0 rtnl_mutex sysctl_lock fill_pool_map-wait-type-override &____s->seqcount irq_context: 0 rtnl_mutex sysctl_lock fill_pool_map-wait-type-override pool_lock irq_context: 0 (wq_completion)ipv6_addrconf (work_completion)(&(&ifa->dad_work)->work) rtnl_mutex &tb->tb6_lock rcu_read_lock pool_lock#2 irq_context: 0 (wq_completion)ipv6_addrconf (work_completion)(&(&ifa->dad_work)->work) rtnl_mutex &tb->tb6_lock rcu_read_lock &data->fib_event_queue_lock irq_context: 0 (wq_completion)ipv6_addrconf (work_completion)(&(&ifa->dad_work)->work) rtnl_mutex &tb->tb6_lock rcu_read_lock rcu_read_lock &pool->lock irq_context: 0 (wq_completion)ipv6_addrconf (work_completion)(&(&ifa->dad_work)->work) rtnl_mutex &tb->tb6_lock rcu_read_lock rcu_read_lock &pool->lock &obj_hash[i].lock irq_context: 0 (wq_completion)ipv6_addrconf (work_completion)(&(&ifa->dad_work)->work) rtnl_mutex &tb->tb6_lock rcu_read_lock rcu_read_lock &pool->lock pool_lock#2 irq_context: 0 (wq_completion)ipv6_addrconf (work_completion)(&(&ifa->dad_work)->work) rtnl_mutex rcu_read_lock rcu_read_lock &obj_hash[i].lock pool_lock irq_context: 0 (wq_completion)ipv6_addrconf (work_completion)(&(&ifa->dad_work)->work) rtnl_mutex rcu_read_lock rcu_read_lock rcu_read_lock_bh fill_pool_map-wait-type-override &____s->seqcount irq_context: 0 (wq_completion)ipv6_addrconf (work_completion)(&(&ifa->dad_work)->work) rtnl_mutex rcu_read_lock rcu_read_lock rcu_read_lock_bh fill_pool_map-wait-type-override pool_lock#2 irq_context: 0 (wq_completion)ipv6_addrconf (work_completion)(&(&ifa->dad_work)->work) rtnl_mutex rcu_read_lock rcu_read_lock rcu_read_lock_bh fill_pool_map-wait-type-override rcu_read_lock pool_lock#2 irq_context: 0 (wq_completion)ipv6_addrconf (work_completion)(&(&ifa->dad_work)->work) rtnl_mutex rcu_read_lock rcu_read_lock rcu_read_lock_bh fill_pool_map-wait-type-override &obj_hash[i].lock irq_context: 0 (wq_completion)ipv6_addrconf (work_completion)(&(&ifa->dad_work)->work) rtnl_mutex rcu_read_lock rcu_read_lock rcu_read_lock_bh fill_pool_map-wait-type-override pool_lock irq_context: 0 (wq_completion)ipv6_addrconf (work_completion)(&(&ifa->dad_work)->work) rtnl_mutex rcu_read_lock rcu_read_lock rcu_read_lock_bh _xmit_ETHER#2 irq_context: 0 (wq_completion)ipv6_addrconf (work_completion)(&(&ifa->dad_work)->work) rtnl_mutex rcu_read_lock rcu_read_lock rcu_read_lock_bh _xmit_ETHER#2 &obj_hash[i].lock irq_context: 0 (wq_completion)ipv6_addrconf (work_completion)(&(&ifa->dad_work)->work) rtnl_mutex rcu_read_lock rcu_read_lock rcu_read_lock_bh _xmit_ETHER#2 pool_lock#2 irq_context: 0 sb_writers#8 &of->mutex kn->active#52 nsim_bus_dev_list_lock &dev->mutex &devlink->lock_key#2 rcu_read_lock &c->lock irq_context: 0 sb_writers#8 &of->mutex kn->active#52 nsim_bus_dev_list_lock &dev->mutex &devlink->lock_key#2 rcu_read_lock &data->fib_event_queue_lock irq_context: 0 sb_writers#8 &of->mutex kn->active#52 nsim_bus_dev_list_lock &dev->mutex &devlink->lock_key#2 rcu_read_lock rcu_read_lock &pool->lock irq_context: 0 sb_writers#8 &of->mutex kn->active#52 nsim_bus_dev_list_lock &dev->mutex &devlink->lock_key#2 rcu_read_lock rcu_read_lock &pool->lock &obj_hash[i].lock irq_context: 0 sb_writers#8 &of->mutex kn->active#52 nsim_bus_dev_list_lock &dev->mutex &devlink->lock_key#2 rcu_read_lock rcu_read_lock &pool->lock &p->pi_lock irq_context: 0 sb_writers#8 &of->mutex kn->active#52 nsim_bus_dev_list_lock &dev->mutex &devlink->lock_key#2 rcu_read_lock rcu_read_lock &pool->lock &p->pi_lock &rq->__lock irq_context: 0 sb_writers#8 &of->mutex kn->active#52 nsim_bus_dev_list_lock &dev->mutex &devlink->lock_key#2 rcu_read_lock rcu_read_lock &pool->lock &p->pi_lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 sb_writers#8 &of->mutex kn->active#52 nsim_bus_dev_list_lock &dev->mutex &devlink->lock_key#2 rcu_read_lock rcu_node_0 irq_context: 0 sb_writers#8 &of->mutex kn->active#52 nsim_bus_dev_list_lock &dev->mutex &devlink->lock_key#2 rcu_read_lock &rq->__lock irq_context: 0 (wq_completion)events (work_completion)(&data->fib_event_work) &data->fib_lock &c->lock irq_context: 0 sb_writers#8 &of->mutex kn->active#52 nsim_bus_dev_list_lock &dev->mutex &devlink->lock_key#2 rcu_read_lock &tb->tb6_lock irq_context: 0 sb_writers#8 &of->mutex kn->active#52 nsim_bus_dev_list_lock &dev->mutex &devlink->lock_key#2 rcu_read_lock &tb->tb6_lock &net->ipv6.fib6_walker_lock irq_context: 0 sb_writers#8 &of->mutex kn->active#52 nsim_bus_dev_list_lock &dev->mutex &devlink->lock_key#2 rcu_read_lock &tb->tb6_lock &data->fib_event_queue_lock irq_context: 0 sb_writers#8 &of->mutex kn->active#52 nsim_bus_dev_list_lock &dev->mutex &devlink->lock_key#2 rcu_read_lock &tb->tb6_lock &c->lock irq_context: 0 sb_writers#8 &of->mutex kn->active#52 nsim_bus_dev_list_lock &dev->mutex &devlink->lock_key#2 rcu_read_lock &obj_hash[i].lock irq_context: 0 sb_writers#8 &of->mutex kn->active#52 nsim_bus_dev_list_lock &dev->mutex &devlink->lock_key#2 &(&fn_net->fib_chain)->lock irq_context: 0 sb_writers#8 &of->mutex kn->active#52 nsim_bus_dev_list_lock &dev->mutex &devlink->lock_key#2 rtnl_mutex rcu_read_lock &rq->__lock irq_context: 0 sb_writers#8 &of->mutex kn->active#52 nsim_bus_dev_list_lock &dev->mutex &devlink->lock_key#2 rtnl_mutex rcu_read_lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 rtnl_mutex devnet_rename_sem batched_entropy_u8.lock irq_context: 0 rtnl_mutex devnet_rename_sem kfence_freelist_lock irq_context: 0 rtnl_mutex devnet_rename_sem remove_cache_srcu irq_context: 0 rtnl_mutex devnet_rename_sem remove_cache_srcu quarantine_lock irq_context: 0 rtnl_mutex devnet_rename_sem &meta->lock irq_context: 0 cb_lock genl_mutex rtnl_mutex &wg->device_update_lock irq_context: 0 cb_lock genl_mutex rtnl_mutex &wg->device_update_lock &wg->static_identity.lock irq_context: 0 cb_lock genl_mutex rtnl_mutex &wg->device_update_lock &wg->static_identity.lock irq_context: 0 cb_lock genl_mutex rtnl_mutex &wg->device_update_lock fs_reclaim irq_context: 0 cb_lock genl_mutex rtnl_mutex &wg->device_update_lock fs_reclaim mmu_notifier_invalidate_range_start irq_context: 0 cb_lock genl_mutex rtnl_mutex &wg->device_update_lock &____s->seqcount irq_context: 0 cb_lock genl_mutex rtnl_mutex &wg->device_update_lock pool_lock#2 irq_context: 0 cb_lock genl_mutex rtnl_mutex &wg->device_update_lock &c->lock irq_context: 0 cb_lock genl_mutex rtnl_mutex &wg->device_update_lock pcpu_alloc_mutex irq_context: 0 cb_lock genl_mutex rtnl_mutex &wg->device_update_lock pcpu_alloc_mutex pcpu_lock irq_context: 0 cb_lock genl_mutex rtnl_mutex &wg->device_update_lock &handshake->lock irq_context: 0 cb_lock genl_mutex rtnl_mutex &wg->device_update_lock &obj_hash[i].lock irq_context: 0 cb_lock genl_mutex rtnl_mutex &wg->device_update_lock &obj_hash[i].lock pool_lock irq_context: 0 cb_lock genl_mutex rtnl_mutex &wg->device_update_lock fill_pool_map-wait-type-override pool_lock#2 irq_context: 0 cb_lock genl_mutex rtnl_mutex &wg->device_update_lock fill_pool_map-wait-type-override pool_lock irq_context: 0 cb_lock genl_mutex rtnl_mutex &wg->device_update_lock tk_core.seq.seqcount irq_context: 0 cb_lock genl_mutex rtnl_mutex &wg->device_update_lock &table->lock irq_context: 0 cb_lock genl_mutex rtnl_mutex &wg->device_update_lock &peer->endpoint_lock irq_context: 0 cb_lock genl_mutex rtnl_mutex &wg->device_update_lock rcu_read_lock pool_lock#2 irq_context: 0 cb_lock genl_mutex rtnl_mutex.wait_lock irq_context: 0 cb_lock genl_mutex &p->pi_lock irq_context: 0 sb_writers#8 &of->mutex kn->active#52 nsim_bus_dev_list_lock &dev->mutex &devlink->lock_key#2 fill_pool_map-wait-type-override pool_lock#2 irq_context: 0 sb_writers#8 &of->mutex kn->active#52 nsim_bus_dev_list_lock &dev->mutex &devlink->lock_key#2 fill_pool_map-wait-type-override pool_lock irq_context: 0 sb_writers#8 &of->mutex kn->active#52 nsim_bus_dev_list_lock &dev->mutex &devlink->lock_key#2 &devlink_port->type_lock irq_context: 0 sb_writers#8 &of->mutex kn->active#52 nsim_bus_dev_list_lock &dev->mutex &devlink->lock_key#2 stack_depot_init_mutex irq_context: 0 sb_writers#8 &of->mutex kn->active#52 nsim_bus_dev_list_lock &dev->mutex &devlink->lock_key#2 &sb->s_type->i_mutex_key#3 &rq->__lock irq_context: 0 sb_writers#8 &of->mutex kn->active#52 nsim_bus_dev_list_lock &dev->mutex &devlink->lock_key#2 rtnl_mutex bpf_devs_lock irq_context: 0 sb_writers#8 &of->mutex kn->active#52 nsim_bus_dev_list_lock &dev->mutex &devlink->lock_key#2 rtnl_mutex bpf_devs_lock fs_reclaim irq_context: 0 sb_writers#8 &of->mutex kn->active#52 nsim_bus_dev_list_lock &dev->mutex &devlink->lock_key#2 rtnl_mutex bpf_devs_lock fs_reclaim mmu_notifier_invalidate_range_start irq_context: 0 sb_writers#8 &of->mutex kn->active#52 nsim_bus_dev_list_lock &dev->mutex &devlink->lock_key#2 rtnl_mutex bpf_devs_lock rcu_read_lock rhashtable_bucket irq_context: 0 sb_writers#8 &of->mutex kn->active#52 nsim_bus_dev_list_lock &dev->mutex &devlink->lock_key#2 rtnl_mutex pin_fs_lock irq_context: 0 sb_writers#8 &of->mutex kn->active#52 nsim_bus_dev_list_lock &dev->mutex &devlink->lock_key#2 rtnl_mutex &sb->s_type->i_mutex_key#3 irq_context: 0 sb_writers#8 &of->mutex kn->active#52 nsim_bus_dev_list_lock &dev->mutex &devlink->lock_key#2 rtnl_mutex &sb->s_type->i_mutex_key#3 rename_lock.seqcount irq_context: 0 sb_writers#8 &of->mutex kn->active#52 nsim_bus_dev_list_lock &dev->mutex &devlink->lock_key#2 rtnl_mutex &sb->s_type->i_mutex_key#3 fs_reclaim irq_context: 0 sb_writers#8 &of->mutex kn->active#52 nsim_bus_dev_list_lock &dev->mutex &devlink->lock_key#2 rtnl_mutex &sb->s_type->i_mutex_key#3 fs_reclaim mmu_notifier_invalidate_range_start irq_context: 0 sb_writers#8 &of->mutex kn->active#52 nsim_bus_dev_list_lock &dev->mutex &devlink->lock_key#2 rtnl_mutex &sb->s_type->i_mutex_key#3 &dentry->d_lock irq_context: 0 sb_writers#8 &of->mutex kn->active#52 nsim_bus_dev_list_lock &dev->mutex &devlink->lock_key#2 rtnl_mutex &sb->s_type->i_mutex_key#3 rcu_read_lock rename_lock.seqcount irq_context: 0 sb_writers#8 &of->mutex kn->active#52 nsim_bus_dev_list_lock &dev->mutex &devlink->lock_key#2 rtnl_mutex &sb->s_type->i_mutex_key#3 &dentry->d_lock &wq irq_context: 0 sb_writers#8 &of->mutex kn->active#52 nsim_bus_dev_list_lock &dev->mutex &devlink->lock_key#2 rtnl_mutex &sb->s_type->i_mutex_key#3 mmu_notifier_invalidate_range_start irq_context: 0 sb_writers#8 &of->mutex kn->active#52 nsim_bus_dev_list_lock &dev->mutex &devlink->lock_key#2 rtnl_mutex &sb->s_type->i_mutex_key#3 &sb->s_type->i_lock_key#7 irq_context: 0 sb_writers#8 &of->mutex kn->active#52 nsim_bus_dev_list_lock &dev->mutex &devlink->lock_key#2 rtnl_mutex &sb->s_type->i_mutex_key#3 &s->s_inode_list_lock irq_context: 0 sb_writers#8 &of->mutex kn->active#52 nsim_bus_dev_list_lock &dev->mutex &devlink->lock_key#2 rtnl_mutex &sb->s_type->i_mutex_key#3 tk_core.seq.seqcount irq_context: 0 sb_writers#8 &of->mutex kn->active#52 nsim_bus_dev_list_lock &dev->mutex &devlink->lock_key#2 rtnl_mutex &sb->s_type->i_mutex_key#3 &sb->s_type->i_lock_key#7 &dentry->d_lock irq_context: 0 sb_writers#8 &of->mutex kn->active#52 nsim_bus_dev_list_lock &dev->mutex &devlink->lock_key#2 rtnl_mutex &sb->s_type->i_mutex_key#3 &c->lock irq_context: 0 sb_writers#8 &of->mutex kn->active#52 nsim_bus_dev_list_lock &dev->mutex &devlink->lock_key#2 rtnl_mutex &sb->s_type->i_mutex_key#3 &____s->seqcount irq_context: 0 sb_writers#8 &of->mutex kn->active#52 nsim_bus_dev_list_lock &dev->mutex &devlink->lock_key#2 rtnl_mutex &sb->s_type->i_mutex_key#3 pool_lock#2 irq_context: 0 sb_writers#8 &of->mutex kn->active#52 nsim_bus_dev_list_lock &dev->mutex &devlink->lock_key#2 rtnl_mutex &____s->seqcount irq_context: 0 sb_writers#8 &of->mutex kn->active#52 nsim_bus_dev_list_lock &dev->mutex &devlink->lock_key#2 rtnl_mutex &obj_hash[i].lock irq_context: 0 sb_writers#8 &of->mutex kn->active#52 nsim_bus_dev_list_lock &dev->mutex &devlink->lock_key#2 rtnl_mutex fs_reclaim irq_context: 0 sb_writers#8 &of->mutex kn->active#52 nsim_bus_dev_list_lock &dev->mutex &devlink->lock_key#2 rtnl_mutex fs_reclaim mmu_notifier_invalidate_range_start irq_context: 0 sb_writers#8 &of->mutex kn->active#52 nsim_bus_dev_list_lock &dev->mutex &devlink->lock_key#2 rtnl_mutex &xa->xa_lock#3 irq_context: 0 sb_writers#8 &of->mutex kn->active#52 nsim_bus_dev_list_lock &dev->mutex &devlink->lock_key#2 rtnl_mutex &base->lock irq_context: 0 sb_writers#8 &of->mutex kn->active#52 nsim_bus_dev_list_lock &dev->mutex &devlink->lock_key#2 rtnl_mutex &base->lock &obj_hash[i].lock irq_context: 0 sb_writers#8 &of->mutex kn->active#52 nsim_bus_dev_list_lock &dev->mutex &devlink->lock_key#2 rtnl_mutex (work_completion)(&(&devlink_port->type_warn_dw)->work) irq_context: 0 sb_writers#8 &of->mutex kn->active#52 nsim_bus_dev_list_lock &dev->mutex &devlink->lock_key#2 rtnl_mutex &devlink_port->type_lock irq_context: 0 sb_writers#8 &of->mutex kn->active#52 nsim_bus_dev_list_lock &dev->mutex &devlink->lock_key#2 rtnl_mutex &c->lock irq_context: 0 sb_writers#8 &of->mutex kn->active#52 nsim_bus_dev_list_lock &dev->mutex &devlink->lock_key#2 rtnl_mutex nl_table_lock irq_context: 0 sb_writers#8 &of->mutex kn->active#52 nsim_bus_dev_list_lock &dev->mutex &devlink->lock_key#2 rtnl_mutex nl_table_wait.lock irq_context: 0 sb_writers#8 &of->mutex kn->active#52 nsim_bus_dev_list_lock &dev->mutex &devlink->lock_key#2 rtnl_mutex net_rwsem irq_context: 0 sb_writers#8 &of->mutex kn->active#52 nsim_bus_dev_list_lock &dev->mutex &devlink->lock_key#2 rtnl_mutex net_rwsem &list->lock#2 irq_context: 0 sb_writers#8 &of->mutex kn->active#52 nsim_bus_dev_list_lock &dev->mutex &devlink->lock_key#2 rtnl_mutex &tn->lock irq_context: 0 sb_writers#8 &of->mutex kn->active#52 nsim_bus_dev_list_lock &dev->mutex &devlink->lock_key#2 rtnl_mutex &x->wait#9 irq_context: 0 sb_writers#8 &of->mutex kn->active#52 nsim_bus_dev_list_lock &dev->mutex &devlink->lock_key#2 rtnl_mutex &k->list_lock irq_context: 0 sb_writers#8 &of->mutex kn->active#52 nsim_bus_dev_list_lock &dev->mutex &devlink->lock_key#2 rtnl_mutex gdp_mutex irq_context: 0 sb_writers#8 &of->mutex kn->active#52 nsim_bus_dev_list_lock &dev->mutex &devlink->lock_key#2 rtnl_mutex gdp_mutex &k->list_lock irq_context: 0 sb_writers#8 &of->mutex kn->active#52 nsim_bus_dev_list_lock &dev->mutex &devlink->lock_key#2 rtnl_mutex gdp_mutex fs_reclaim irq_context: 0 sb_writers#8 &of->mutex kn->active#52 nsim_bus_dev_list_lock &dev->mutex &devlink->lock_key#2 rtnl_mutex gdp_mutex fs_reclaim mmu_notifier_invalidate_range_start irq_context: 0 sb_writers#8 &of->mutex kn->active#52 nsim_bus_dev_list_lock &dev->mutex &devlink->lock_key#2 rtnl_mutex gdp_mutex lock irq_context: 0 sb_writers#8 &of->mutex kn->active#52 nsim_bus_dev_list_lock &dev->mutex &devlink->lock_key#2 rtnl_mutex gdp_mutex lock kernfs_idr_lock irq_context: 0 sb_writers#8 &of->mutex kn->active#52 nsim_bus_dev_list_lock &dev->mutex &devlink->lock_key#2 rtnl_mutex gdp_mutex &root->kernfs_rwsem irq_context: 0 sb_writers#8 &of->mutex kn->active#52 nsim_bus_dev_list_lock &dev->mutex &devlink->lock_key#2 rtnl_mutex gdp_mutex &root->kernfs_rwsem &root->kernfs_iattr_rwsem irq_context: 0 sb_writers#8 &of->mutex kn->active#52 nsim_bus_dev_list_lock &dev->mutex &devlink->lock_key#2 rtnl_mutex gdp_mutex kobj_ns_type_lock irq_context: 0 sb_writers#8 &of->mutex kn->active#52 nsim_bus_dev_list_lock &dev->mutex &devlink->lock_key#2 rtnl_mutex lock irq_context: 0 sb_writers#8 &of->mutex kn->active#52 nsim_bus_dev_list_lock &dev->mutex &devlink->lock_key#2 rtnl_mutex lock kernfs_idr_lock irq_context: 0 sb_writers#8 &of->mutex kn->active#52 nsim_bus_dev_list_lock &dev->mutex &devlink->lock_key#2 rtnl_mutex &root->kernfs_rwsem irq_context: 0 sb_writers#8 &of->mutex kn->active#52 nsim_bus_dev_list_lock &dev->mutex &devlink->lock_key#2 rtnl_mutex &root->kernfs_rwsem &root->kernfs_iattr_rwsem irq_context: 0 sb_writers#8 &of->mutex kn->active#52 nsim_bus_dev_list_lock &dev->mutex &devlink->lock_key#2 rtnl_mutex bus_type_sem irq_context: 0 sb_writers#8 &of->mutex kn->active#52 nsim_bus_dev_list_lock &dev->mutex &devlink->lock_key#2 rtnl_mutex sysfs_symlink_target_lock irq_context: 0 sb_writers#8 &of->mutex kn->active#52 nsim_bus_dev_list_lock &dev->mutex &devlink->lock_key#2 rtnl_mutex &root->kernfs_rwsem irq_context: 0 sb_writers#8 &of->mutex kn->active#52 nsim_bus_dev_list_lock &dev->mutex &devlink->lock_key#2 rtnl_mutex &dev->power.lock irq_context: 0 sb_writers#8 &of->mutex kn->active#52 nsim_bus_dev_list_lock &dev->mutex &devlink->lock_key#2 rtnl_mutex dpm_list_mtx irq_context: 0 sb_writers#8 &of->mutex kn->active#52 nsim_bus_dev_list_lock &dev->mutex &devlink->lock_key#2 rtnl_mutex uevent_sock_mutex irq_context: 0 sb_writers#8 &of->mutex kn->active#52 nsim_bus_dev_list_lock &dev->mutex &devlink->lock_key#2 rtnl_mutex uevent_sock_mutex fs_reclaim irq_context: 0 sb_writers#8 &of->mutex kn->active#52 nsim_bus_dev_list_lock &dev->mutex &devlink->lock_key#2 rtnl_mutex uevent_sock_mutex fs_reclaim mmu_notifier_invalidate_range_start irq_context: 0 sb_writers#8 &of->mutex kn->active#52 nsim_bus_dev_list_lock &dev->mutex &devlink->lock_key#2 rtnl_mutex uevent_sock_mutex nl_table_lock irq_context: 0 sb_writers#8 &of->mutex kn->active#52 nsim_bus_dev_list_lock &dev->mutex &devlink->lock_key#2 rtnl_mutex uevent_sock_mutex &obj_hash[i].lock irq_context: 0 sb_writers#8 &of->mutex kn->active#52 nsim_bus_dev_list_lock &dev->mutex &devlink->lock_key#2 rtnl_mutex uevent_sock_mutex nl_table_wait.lock irq_context: 0 sb_writers#8 &of->mutex kn->active#52 nsim_bus_dev_list_lock &dev->mutex &devlink->lock_key#2 rtnl_mutex &k->k_lock irq_context: 0 sb_writers#8 &of->mutex kn->active#52 nsim_bus_dev_list_lock &dev->mutex &devlink->lock_key#2 rtnl_mutex subsys mutex#17 irq_context: 0 sb_writers#8 &of->mutex kn->active#52 nsim_bus_dev_list_lock &dev->mutex &devlink->lock_key#2 rtnl_mutex subsys mutex#17 &k->k_lock irq_context: 0 sb_writers#8 &of->mutex kn->active#52 nsim_bus_dev_list_lock &dev->mutex &devlink->lock_key#2 rtnl_mutex &dir->lock#2 irq_context: 0 sb_writers#8 &of->mutex kn->active#52 nsim_bus_dev_list_lock &dev->mutex &devlink->lock_key#2 rtnl_mutex uevent_sock_mutex &c->lock irq_context: 0 sb_writers#8 &of->mutex kn->active#52 nsim_bus_dev_list_lock &dev->mutex &devlink->lock_key#2 rtnl_mutex uevent_sock_mutex &n->list_lock irq_context: 0 sb_writers#8 &of->mutex kn->active#52 nsim_bus_dev_list_lock &dev->mutex &devlink->lock_key#2 rtnl_mutex uevent_sock_mutex &n->list_lock &c->lock irq_context: 0 sb_writers#8 &of->mutex kn->active#52 nsim_bus_dev_list_lock &dev->mutex &devlink->lock_key#2 rtnl_mutex dev_hotplug_mutex irq_context: 0 sb_writers#8 &of->mutex kn->active#52 nsim_bus_dev_list_lock &dev->mutex &devlink->lock_key#2 rtnl_mutex dev_hotplug_mutex &dev->power.lock irq_context: 0 sb_writers#8 &of->mutex kn->active#52 nsim_bus_dev_list_lock &dev->mutex &devlink->lock_key#2 rtnl_mutex dev_base_lock irq_context: 0 sb_writers#8 &of->mutex kn->active#52 nsim_bus_dev_list_lock &dev->mutex &devlink->lock_key#2 rtnl_mutex &obj_hash[i].lock pool_lock irq_context: 0 sb_writers#8 &of->mutex kn->active#52 nsim_bus_dev_list_lock &dev->mutex &devlink->lock_key#2 rtnl_mutex input_pool.lock irq_context: 0 sb_writers#8 &of->mutex kn->active#52 nsim_bus_dev_list_lock &dev->mutex &devlink->lock_key#2 rtnl_mutex fill_pool_map-wait-type-override pool_lock#2 irq_context: 0 sb_writers#8 &of->mutex kn->active#52 nsim_bus_dev_list_lock &dev->mutex &devlink->lock_key#2 rtnl_mutex fill_pool_map-wait-type-override pool_lock irq_context: 0 sb_writers#8 &of->mutex kn->active#52 nsim_bus_dev_list_lock &dev->mutex &devlink->lock_key#2 rtnl_mutex quarantine_lock irq_context: 0 sb_writers#8 &of->mutex kn->active#52 nsim_bus_dev_list_lock &dev->mutex &devlink->lock_key#2 rtnl_mutex remove_cache_srcu irq_context: 0 sb_writers#8 &of->mutex kn->active#52 nsim_bus_dev_list_lock &dev->mutex &devlink->lock_key#2 rtnl_mutex remove_cache_srcu quarantine_lock irq_context: 0 sb_writers#8 &of->mutex kn->active#52 nsim_bus_dev_list_lock &dev->mutex &devlink->lock_key#2 rtnl_mutex remove_cache_srcu &c->lock irq_context: 0 sb_writers#8 &of->mutex kn->active#52 nsim_bus_dev_list_lock &dev->mutex &devlink->lock_key#2 rtnl_mutex remove_cache_srcu &n->list_lock irq_context: 0 sb_writers#8 &of->mutex kn->active#52 nsim_bus_dev_list_lock &dev->mutex &devlink->lock_key#2 rtnl_mutex remove_cache_srcu rcu_read_lock pool_lock#2 irq_context: 0 sb_writers#8 &of->mutex kn->active#52 nsim_bus_dev_list_lock &dev->mutex &devlink->lock_key#2 rtnl_mutex remove_cache_srcu &obj_hash[i].lock irq_context: 0 sb_writers#8 &of->mutex kn->active#52 nsim_bus_dev_list_lock &dev->mutex &devlink->lock_key#2 rtnl_mutex rcu_read_lock &pool->lock/1 irq_context: 0 sb_writers#8 &of->mutex kn->active#52 nsim_bus_dev_list_lock &dev->mutex &devlink->lock_key#2 rtnl_mutex rcu_read_lock &pool->lock/1 &obj_hash[i].lock irq_context: 0 sb_writers#8 &of->mutex kn->active#52 nsim_bus_dev_list_lock &dev->mutex &devlink->lock_key#2 rtnl_mutex rcu_read_lock &pool->lock/1 &p->pi_lock irq_context: 0 sb_writers#8 &of->mutex kn->active#52 nsim_bus_dev_list_lock &dev->mutex &devlink->lock_key#2 rtnl_mutex pool_lock#2 irq_context: 0 sb_writers#8 &of->mutex kn->active#52 nsim_bus_dev_list_lock &dev->mutex &devlink->lock_key#2 rtnl_mutex batched_entropy_u32.lock irq_context: 0 sb_writers#8 &of->mutex kn->active#52 nsim_bus_dev_list_lock &dev->mutex &devlink->lock_key#2 rtnl_mutex &tbl->lock irq_context: 0 sb_writers#8 &of->mutex kn->active#52 nsim_bus_dev_list_lock &dev->mutex &devlink->lock_key#2 rtnl_mutex sysctl_lock irq_context: 0 sb_writers#8 &of->mutex kn->active#52 nsim_bus_dev_list_lock &dev->mutex &devlink->lock_key#2 rtnl_mutex failover_lock irq_context: 0 sb_writers#8 &of->mutex kn->active#52 nsim_bus_dev_list_lock &dev->mutex &devlink->lock_key#2 rtnl_mutex pcpu_alloc_mutex irq_context: 0 sb_writers#8 &of->mutex kn->active#52 nsim_bus_dev_list_lock &dev->mutex &devlink->lock_key#2 rtnl_mutex pcpu_alloc_mutex pcpu_lock irq_context: 0 sb_writers#8 &of->mutex kn->active#52 nsim_bus_dev_list_lock &dev->mutex &devlink->lock_key#2 rtnl_mutex proc_subdir_lock irq_context: 0 sb_writers#8 &of->mutex kn->active#52 nsim_bus_dev_list_lock &dev->mutex &devlink->lock_key#2 rtnl_mutex proc_inum_ida.xa_lock irq_context: 0 sb_writers#8 &of->mutex kn->active#52 nsim_bus_dev_list_lock &dev->mutex &devlink->lock_key#2 rtnl_mutex proc_subdir_lock irq_context: 0 sb_writers#8 &of->mutex kn->active#52 nsim_bus_dev_list_lock &dev->mutex &devlink->lock_key#2 rtnl_mutex fill_pool_map-wait-type-override &c->lock irq_context: 0 sb_writers#8 &of->mutex kn->active#52 nsim_bus_dev_list_lock &dev->mutex &devlink->lock_key#2 rtnl_mutex fill_pool_map-wait-type-override &____s->seqcount irq_context: 0 sb_writers#8 &of->mutex kn->active#52 nsim_bus_dev_list_lock &dev->mutex &devlink->lock_key#2 rtnl_mutex &idev->mc_lock irq_context: 0 sb_writers#8 &of->mutex kn->active#52 nsim_bus_dev_list_lock &dev->mutex &devlink->lock_key#2 rtnl_mutex &idev->mc_lock fs_reclaim irq_context: 0 sb_writers#8 &of->mutex kn->active#52 nsim_bus_dev_list_lock &dev->mutex &devlink->lock_key#2 rtnl_mutex &idev->mc_lock fs_reclaim mmu_notifier_invalidate_range_start irq_context: 0 sb_writers#8 &of->mutex kn->active#52 nsim_bus_dev_list_lock &dev->mutex &devlink->lock_key#2 rtnl_mutex &idev->mc_lock &obj_hash[i].lock irq_context: 0 sb_writers#8 &of->mutex kn->active#52 nsim_bus_dev_list_lock &dev->mutex &devlink->lock_key#2 rtnl_mutex &idev->mc_lock _xmit_ETHER irq_context: 0 sb_writers#8 &of->mutex kn->active#52 nsim_bus_dev_list_lock &dev->mutex &devlink->lock_key#2 rtnl_mutex &pnettable->lock irq_context: 0 sb_writers#8 &of->mutex kn->active#52 nsim_bus_dev_list_lock &dev->mutex &devlink->lock_key#2 rtnl_mutex smc_ib_devices.mutex irq_context: 0 sb_writers#8 &of->mutex kn->active#52 nsim_bus_dev_list_lock &dev->mutex &devlink->lock_key#2 rtnl_mutex &vn->sock_lock irq_context: 0 sb_writers#8 &of->mutex kn->active#52 nsim_bus_dev_list_lock &dev->mutex &devlink->lock_key#2 rtnl_mutex rcu_read_lock &ndev->lock irq_context: 0 sb_writers#8 &of->mutex kn->active#52 nsim_bus_dev_list_lock &dev->mutex &devlink->lock_key#2 rtnl_mutex &pcp->lock &zone->lock irq_context: 0 sb_writers#8 &of->mutex kn->active#52 nsim_bus_dev_list_lock &dev->mutex &devlink->lock_key#2 rtnl_mutex &pcp->lock &zone->lock &____s->seqcount irq_context: 0 sb_writers#8 &of->mutex kn->active#52 nsim_bus_dev_list_lock &dev->mutex &devlink->lock_key#2 &sb->s_type->i_mutex_key#3 &pcp->lock &zone->lock irq_context: 0 sb_writers#8 &of->mutex kn->active#52 nsim_bus_dev_list_lock &dev->mutex &devlink->lock_key#2 &sb->s_type->i_mutex_key#3 &pcp->lock &zone->lock &____s->seqcount irq_context: 0 rtnl_mutex (inetaddr_chain).rwsem rcu_read_lock &data->fib_event_queue_lock irq_context: 0 rtnl_mutex rcu_read_lock &data->fib_event_queue_lock irq_context: 0 rtnl_mutex &tb->tb6_lock rcu_read_lock &data->fib_event_queue_lock irq_context: 0 (wq_completion)ipv6_addrconf (work_completion)(&(&ifa->dad_work)->work) rtnl_mutex rcu_read_lock &rq->__lock irq_context: 0 (wq_completion)ipv6_addrconf (work_completion)(&(&ifa->dad_work)->work) rtnl_mutex rcu_read_lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 sb_writers#8 &of->mutex kn->active#52 nsim_bus_dev_list_lock &dev->mutex &devlink->lock_key#2 rtnl_mutex &n->list_lock irq_context: 0 sb_writers#8 &of->mutex kn->active#52 nsim_bus_dev_list_lock &dev->mutex &devlink->lock_key#2 rtnl_mutex &n->list_lock &c->lock irq_context: 0 sb_writers#8 &of->mutex kn->active#52 nsim_bus_dev_list_lock &dev->mutex &devlink->lock_key#2 rtnl_mutex bpf_devs_lock &c->lock irq_context: 0 sb_writers#8 &of->mutex kn->active#52 nsim_bus_dev_list_lock &dev->mutex &devlink->lock_key#2 rtnl_mutex bpf_devs_lock &n->list_lock irq_context: 0 sb_writers#8 &of->mutex kn->active#52 nsim_bus_dev_list_lock &dev->mutex &devlink->lock_key#2 rtnl_mutex bpf_devs_lock &n->list_lock &c->lock irq_context: 0 sb_writers#8 &of->mutex kn->active#52 nsim_bus_dev_list_lock &dev->mutex &devlink->lock_key#2 rtnl_mutex bpf_devs_lock rcu_read_lock rcu_read_lock &pool->lock irq_context: 0 sb_writers#8 &of->mutex kn->active#52 nsim_bus_dev_list_lock &dev->mutex &devlink->lock_key#2 rtnl_mutex bpf_devs_lock rcu_read_lock rcu_read_lock &pool->lock &obj_hash[i].lock irq_context: 0 sb_writers#8 &of->mutex kn->active#52 nsim_bus_dev_list_lock &dev->mutex &devlink->lock_key#2 rtnl_mutex bpf_devs_lock rcu_read_lock rcu_read_lock &pool->lock &p->pi_lock irq_context: 0 sb_writers#8 &of->mutex kn->active#52 nsim_bus_dev_list_lock &dev->mutex &devlink->lock_key#2 rtnl_mutex bpf_devs_lock rcu_read_lock rcu_read_lock &pool->lock &p->pi_lock &rq->__lock irq_context: 0 sb_writers#8 &of->mutex kn->active#52 nsim_bus_dev_list_lock &dev->mutex &devlink->lock_key#2 rtnl_mutex bpf_devs_lock rcu_read_lock rcu_read_lock &pool->lock &p->pi_lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 sb_writers#8 &of->mutex kn->active#52 nsim_bus_dev_list_lock &dev->mutex &devlink->lock_key#2 rtnl_mutex bpf_devs_lock rcu_read_lock rcu_node_0 irq_context: 0 sb_writers#8 &of->mutex kn->active#52 nsim_bus_dev_list_lock &dev->mutex &devlink->lock_key#2 rtnl_mutex bpf_devs_lock rcu_read_lock &rq->__lock irq_context: 0 (wq_completion)ipv6_addrconf (work_completion)(&(&ifa->dad_work)->work) rtnl_mutex &pool->lock &p->pi_lock irq_context: 0 sb_writers#8 &of->mutex kn->active#52 nsim_bus_dev_list_lock &dev->mutex &devlink->lock_key#2 rtnl_mutex lock kernfs_idr_lock pool_lock#2 irq_context: 0 sb_writers#8 &of->mutex kn->active#52 nsim_bus_dev_list_lock &dev->mutex &devlink->lock_key#2 rtnl_mutex &sb->s_type->i_mutex_key#3 &n->list_lock irq_context: 0 sb_writers#8 &of->mutex kn->active#52 nsim_bus_dev_list_lock &dev->mutex &devlink->lock_key#2 rtnl_mutex &sb->s_type->i_mutex_key#3 &n->list_lock &c->lock irq_context: softirq &(&nsim_dev->trap_data->trap_report_dw)->timer rcu_read_lock &pool->lock &p->pi_lock &rq->__lock irq_context: softirq &(&nsim_dev->trap_data->trap_report_dw)->timer rcu_read_lock &pool->lock &p->pi_lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)events (work_completion)(&(&nsim_dev->trap_data->trap_report_dw)->work) &devlink->lock_key &nsim_trap_data->trap_lock irq_context: 0 (wq_completion)events (work_completion)(&(&nsim_dev->trap_data->trap_report_dw)->work) &devlink->lock_key &nsim_trap_data->trap_lock pool_lock#2 irq_context: 0 (wq_completion)events (work_completion)(&(&nsim_dev->trap_data->trap_report_dw)->work) &devlink->lock_key &nsim_trap_data->trap_lock crngs.lock irq_context: 0 (wq_completion)events (work_completion)(&(&nsim_dev->trap_data->trap_report_dw)->work) &devlink->lock_key &nsim_trap_data->trap_lock &nsim_dev->fa_cookie_lock irq_context: 0 (wq_completion)events (work_completion)(&(&nsim_dev->trap_data->trap_report_dw)->work) &devlink->lock_key &nsim_trap_data->trap_lock &obj_hash[i].lock irq_context: 0 (wq_completion)events (work_completion)(&(&nsim_dev->trap_data->trap_report_dw)->work) &devlink->lock_key &nsim_trap_data->trap_lock &c->lock irq_context: 0 (wq_completion)events (work_completion)(&(&nsim_dev->trap_data->trap_report_dw)->work) &devlink->lock_key &nsim_trap_data->trap_lock &pcp->lock &zone->lock irq_context: 0 (wq_completion)events (work_completion)(&(&nsim_dev->trap_data->trap_report_dw)->work) &devlink->lock_key &nsim_trap_data->trap_lock &pcp->lock &zone->lock &____s->seqcount irq_context: 0 (wq_completion)events (work_completion)(&(&nsim_dev->trap_data->trap_report_dw)->work) &devlink->lock_key &nsim_trap_data->trap_lock &____s->seqcount irq_context: 0 sb_writers#8 &of->mutex kn->active#52 nsim_bus_dev_list_lock &dev->mutex &devlink->lock_key#2 rtnl_mutex rcu_read_lock &pool->lock/1 &p->pi_lock &rq->__lock irq_context: 0 sb_writers#8 &of->mutex kn->active#52 nsim_bus_dev_list_lock &dev->mutex &devlink->lock_key#2 rtnl_mutex rcu_read_lock &pool->lock/1 &p->pi_lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 sb_writers#8 &of->mutex kn->active#52 nsim_bus_dev_list_lock &dev->mutex &devlink->lock_key#2 rtnl_mutex &idev->mc_lock &obj_hash[i].lock pool_lock irq_context: 0 sb_writers#8 &of->mutex kn->active#52 nsim_bus_dev_list_lock &dev->mutex &devlink->lock_key#2 &p->pi_lock &rq->__lock &cfs_rq->removed.lock irq_context: 0 rtnl_mutex _xmit_SIT irq_context: 0 rtnl_mutex &ndev->lock &obj_hash[i].lock pool_lock irq_context: 0 rtnl_mutex &ndev->lock fill_pool_map-wait-type-override pool_lock#2 irq_context: 0 rtnl_mutex &ndev->lock fill_pool_map-wait-type-override pool_lock irq_context: 0 rtnl_mutex &ndev->lock rcu_read_lock &pool->lock &p->pi_lock &rq->__lock irq_context: 0 rtnl_mutex &ndev->lock rcu_read_lock &pool->lock &p->pi_lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 cb_lock nlk_cb_mutex-GENERIC &devlink->lock_key#2 irq_context: 0 cb_lock nlk_cb_mutex-GENERIC &devlink->lock_key#2 &devlink_port->type_lock irq_context: 0 kn->active#52 &c->lock irq_context: 0 sb_writers#8 &of->mutex kn->active#52 nsim_bus_dev_list_lock &dev->mutex &devlink->lock_key#3 irq_context: 0 sb_writers#8 &of->mutex kn->active#52 nsim_bus_dev_list_lock &dev->mutex &devlink->lock_key#3 crngs.lock irq_context: 0 sb_writers#8 &of->mutex kn->active#52 nsim_bus_dev_list_lock &dev->mutex &devlink->lock_key#3 fs_reclaim irq_context: 0 sb_writers#8 &of->mutex kn->active#52 nsim_bus_dev_list_lock &dev->mutex &devlink->lock_key#3 fs_reclaim mmu_notifier_invalidate_range_start irq_context: 0 sb_writers#8 &of->mutex kn->active#52 nsim_bus_dev_list_lock &dev->mutex &devlink->lock_key#3 devlinks.xa_lock irq_context: 0 sb_writers#8 &of->mutex kn->active#52 nsim_bus_dev_list_lock &dev->mutex &devlink->lock_key#3 &obj_hash[i].lock irq_context: 0 sb_writers#8 &of->mutex kn->active#52 nsim_bus_dev_list_lock &dev->mutex &devlink->lock_key#3 nl_table_lock irq_context: 0 sb_writers#8 &of->mutex kn->active#52 nsim_bus_dev_list_lock &dev->mutex &devlink->lock_key#3 nl_table_wait.lock irq_context: 0 sb_writers#8 &of->mutex kn->active#52 nsim_bus_dev_list_lock &dev->mutex &devlink->lock_key#3 &c->lock irq_context: 0 sb_writers#8 &of->mutex kn->active#52 nsim_bus_dev_list_lock &dev->mutex &devlink->lock_key#3 &____s->seqcount irq_context: 0 sb_writers#8 &of->mutex kn->active#52 nsim_bus_dev_list_lock &dev->mutex &devlink->lock_key#3 pool_lock#2 irq_context: 0 sb_writers#8 &of->mutex kn->active#52 nsim_bus_dev_list_lock &dev->mutex &devlink->lock_key#3 &xa->xa_lock#16 irq_context: 0 sb_writers#8 &of->mutex kn->active#52 nsim_bus_dev_list_lock &dev->mutex &devlink->lock_key#3 &pcp->lock &zone->lock irq_context: 0 sb_writers#8 &of->mutex kn->active#52 nsim_bus_dev_list_lock &dev->mutex &devlink->lock_key#3 &pcp->lock &zone->lock &____s->seqcount irq_context: 0 sb_writers#8 &of->mutex kn->active#52 nsim_bus_dev_list_lock &dev->mutex &devlink->lock_key#3 pcpu_alloc_mutex irq_context: 0 sb_writers#8 &of->mutex kn->active#52 nsim_bus_dev_list_lock &dev->mutex &devlink->lock_key#3 pcpu_alloc_mutex pcpu_lock irq_context: 0 sb_writers#8 &of->mutex kn->active#52 nsim_bus_dev_list_lock &dev->mutex &devlink->lock_key#3 rcu_read_lock &rq->__lock irq_context: 0 sb_writers#8 &of->mutex kn->active#52 nsim_bus_dev_list_lock &dev->mutex &devlink->lock_key#3 rcu_read_lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 rtnl_mutex &tb->tb6_lock rcu_read_lock &____s->seqcount irq_context: 0 rtnl_mutex &tb->tb6_lock rcu_read_lock rcu_read_lock pool_lock#2 irq_context: 0 rtnl_mutex &tb->tb6_lock rcu_read_lock &obj_hash[i].lock irq_context: 0 rtnl_mutex &tb->tb6_lock rcu_read_lock &c->lock irq_context: 0 (wq_completion)ipv6_addrconf (work_completion)(&(&ifa->dad_work)->work) rtnl_mutex rcu_read_lock &pool->lock irq_context: 0 (wq_completion)ipv6_addrconf (work_completion)(&(&ifa->dad_work)->work) rtnl_mutex rcu_read_lock &pool->lock &obj_hash[i].lock irq_context: 0 (wq_completion)ipv6_addrconf (work_completion)(&(&ifa->dad_work)->work) rtnl_mutex rcu_read_lock rcu_read_lock rcu_read_lock_bh &obj_hash[i].lock pool_lock irq_context: 0 (wq_completion)ipv6_addrconf (work_completion)(&(&ifa->dad_work)->work) rtnl_mutex &ndev->lock &ifa->lock fill_pool_map-wait-type-override pool_lock#2 irq_context: 0 (wq_completion)ipv6_addrconf (work_completion)(&(&ifa->dad_work)->work) rtnl_mutex &ndev->lock &ifa->lock fill_pool_map-wait-type-override pool_lock irq_context: 0 sb_writers#8 &of->mutex kn->active#52 nsim_bus_dev_list_lock &dev->mutex &devlink->lock_key#3 &base->lock irq_context: 0 sb_writers#8 &of->mutex kn->active#52 nsim_bus_dev_list_lock &dev->mutex &devlink->lock_key#3 &base->lock &obj_hash[i].lock irq_context: 0 sb_writers#8 &of->mutex kn->active#52 nsim_bus_dev_list_lock &dev->mutex &devlink->lock_key#3 pin_fs_lock irq_context: 0 sb_writers#8 &of->mutex kn->active#52 nsim_bus_dev_list_lock &dev->mutex &devlink->lock_key#3 &sb->s_type->i_mutex_key#3 irq_context: 0 sb_writers#8 &of->mutex kn->active#52 nsim_bus_dev_list_lock &dev->mutex &devlink->lock_key#3 &sb->s_type->i_mutex_key#3 rename_lock.seqcount irq_context: 0 sb_writers#8 &of->mutex kn->active#52 nsim_bus_dev_list_lock &dev->mutex &devlink->lock_key#3 &sb->s_type->i_mutex_key#3 fs_reclaim irq_context: 0 sb_writers#8 &of->mutex kn->active#52 nsim_bus_dev_list_lock &dev->mutex &devlink->lock_key#3 &sb->s_type->i_mutex_key#3 fs_reclaim mmu_notifier_invalidate_range_start irq_context: 0 sb_writers#8 &of->mutex kn->active#52 nsim_bus_dev_list_lock &dev->mutex &devlink->lock_key#3 &sb->s_type->i_mutex_key#3 &c->lock irq_context: 0 sb_writers#8 &of->mutex kn->active#52 nsim_bus_dev_list_lock &dev->mutex &devlink->lock_key#3 &sb->s_type->i_mutex_key#3 &n->list_lock irq_context: 0 sb_writers#8 &of->mutex kn->active#52 nsim_bus_dev_list_lock &dev->mutex &devlink->lock_key#3 &sb->s_type->i_mutex_key#3 &n->list_lock &c->lock irq_context: 0 sb_writers#8 &of->mutex kn->active#52 nsim_bus_dev_list_lock &dev->mutex &devlink->lock_key#3 &sb->s_type->i_mutex_key#3 &dentry->d_lock irq_context: 0 sb_writers#8 &of->mutex kn->active#52 nsim_bus_dev_list_lock &dev->mutex &devlink->lock_key#3 &sb->s_type->i_mutex_key#3 rcu_read_lock rename_lock.seqcount irq_context: 0 sb_writers#8 &of->mutex kn->active#52 nsim_bus_dev_list_lock &dev->mutex &devlink->lock_key#3 &sb->s_type->i_mutex_key#3 &dentry->d_lock &wq irq_context: 0 sb_writers#8 &of->mutex kn->active#52 nsim_bus_dev_list_lock &dev->mutex &devlink->lock_key#3 &sb->s_type->i_mutex_key#3 mmu_notifier_invalidate_range_start irq_context: 0 sb_writers#8 &of->mutex kn->active#52 nsim_bus_dev_list_lock &dev->mutex &devlink->lock_key#3 &sb->s_type->i_mutex_key#3 &sb->s_type->i_lock_key#7 irq_context: 0 sb_writers#8 &of->mutex kn->active#52 nsim_bus_dev_list_lock &dev->mutex &devlink->lock_key#3 &sb->s_type->i_mutex_key#3 &s->s_inode_list_lock irq_context: 0 sb_writers#8 &of->mutex kn->active#52 nsim_bus_dev_list_lock &dev->mutex &devlink->lock_key#3 &sb->s_type->i_mutex_key#3 tk_core.seq.seqcount irq_context: 0 sb_writers#8 &of->mutex kn->active#52 nsim_bus_dev_list_lock &dev->mutex &devlink->lock_key#3 &sb->s_type->i_mutex_key#3 &sb->s_type->i_lock_key#7 &dentry->d_lock irq_context: 0 sb_writers#8 &of->mutex kn->active#52 nsim_bus_dev_list_lock &dev->mutex &devlink->lock_key#3 &sb->s_type->i_mutex_key#3 &____s->seqcount irq_context: 0 sb_writers#8 &of->mutex kn->active#52 nsim_bus_dev_list_lock &dev->mutex &devlink->lock_key#3 &sb->s_type->i_mutex_key#3 pool_lock#2 irq_context: softirq (&app->join_timer) &app->lock batched_entropy_u32.lock crngs.lock irq_context: 0 sb_writers#8 &of->mutex kn->active#52 nsim_bus_dev_list_lock &dev->mutex &devlink->lock_key#3 batched_entropy_u32.lock irq_context: 0 sb_writers#8 &of->mutex kn->active#52 nsim_bus_dev_list_lock &dev->mutex &devlink->lock_key#3 rtnl_mutex irq_context: 0 sb_writers#8 &of->mutex kn->active#52 nsim_bus_dev_list_lock &dev->mutex &devlink->lock_key#3 rtnl_mutex rtnl_mutex.wait_lock irq_context: 0 sb_writers#8 &of->mutex kn->active#52 nsim_bus_dev_list_lock &dev->mutex &devlink->lock_key#3 rtnl_mutex &rq->__lock irq_context: 0 sb_writers#8 &of->mutex kn->active#52 nsim_bus_dev_list_lock &dev->mutex &devlink->lock_key#3 rtnl_mutex &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 sb_writers#8 &of->mutex kn->active#52 nsim_bus_dev_list_lock &dev->mutex &devlink->lock_key#3 rtnl_mutex &(&net->nexthop.notifier_chain)->rwsem irq_context: 0 sb_writers#8 &of->mutex kn->active#52 nsim_bus_dev_list_lock &dev->mutex &devlink->lock_key#3 rtnl_mutex.wait_lock irq_context: 0 sb_writers#8 &of->mutex kn->active#52 nsim_bus_dev_list_lock &dev->mutex &devlink->lock_key#3 &p->pi_lock irq_context: 0 sb_writers#8 &of->mutex kn->active#52 nsim_bus_dev_list_lock &dev->mutex &devlink->lock_key#3 &p->pi_lock &rq->__lock irq_context: 0 sb_writers#8 &of->mutex kn->active#52 nsim_bus_dev_list_lock &dev->mutex &devlink->lock_key#3 &p->pi_lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 sb_writers#8 &of->mutex kn->active#52 nsim_bus_dev_list_lock &dev->mutex &devlink->lock_key#3 rcu_read_lock &data->fib_event_queue_lock irq_context: 0 sb_writers#8 &of->mutex kn->active#52 nsim_bus_dev_list_lock &dev->mutex &devlink->lock_key#3 rcu_read_lock rcu_read_lock &pool->lock irq_context: 0 sb_writers#8 &of->mutex kn->active#52 nsim_bus_dev_list_lock &dev->mutex &devlink->lock_key#3 rcu_read_lock rcu_read_lock &pool->lock &obj_hash[i].lock irq_context: 0 sb_writers#8 &of->mutex kn->active#52 nsim_bus_dev_list_lock &dev->mutex &devlink->lock_key#3 rcu_read_lock rcu_read_lock &pool->lock &p->pi_lock irq_context: 0 sb_writers#8 &of->mutex kn->active#52 nsim_bus_dev_list_lock &dev->mutex &devlink->lock_key#3 rcu_read_lock rcu_read_lock &pool->lock &p->pi_lock &rq->__lock irq_context: 0 sb_writers#8 &of->mutex kn->active#52 nsim_bus_dev_list_lock &dev->mutex &devlink->lock_key#3 rcu_read_lock rcu_read_lock &pool->lock &p->pi_lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 sb_writers#8 &of->mutex kn->active#52 nsim_bus_dev_list_lock &dev->mutex &devlink->lock_key#3 rcu_read_lock &c->lock irq_context: 0 sb_writers#8 &of->mutex kn->active#52 nsim_bus_dev_list_lock &dev->mutex &devlink->lock_key#3 rcu_read_lock &tb->tb6_lock irq_context: 0 sb_writers#8 &of->mutex kn->active#52 nsim_bus_dev_list_lock &dev->mutex &devlink->lock_key#3 rcu_read_lock &tb->tb6_lock &net->ipv6.fib6_walker_lock irq_context: 0 sb_writers#8 &of->mutex kn->active#52 nsim_bus_dev_list_lock &dev->mutex &devlink->lock_key#3 rcu_read_lock &tb->tb6_lock &data->fib_event_queue_lock irq_context: 0 sb_writers#8 &of->mutex kn->active#52 nsim_bus_dev_list_lock &dev->mutex &devlink->lock_key#3 rcu_read_lock &obj_hash[i].lock irq_context: 0 sb_writers#8 &of->mutex kn->active#52 nsim_bus_dev_list_lock &dev->mutex &devlink->lock_key#3 &(&fn_net->fib_chain)->lock irq_context: 0 sb_writers#8 &of->mutex kn->active#52 nsim_bus_dev_list_lock &dev->mutex &devlink->lock_key#3 &obj_hash[i].lock pool_lock irq_context: 0 sb_writers#8 &of->mutex kn->active#52 nsim_bus_dev_list_lock &dev->mutex &devlink->lock_key#3 fill_pool_map-wait-type-override pool_lock#2 irq_context: 0 sb_writers#8 &of->mutex kn->active#52 nsim_bus_dev_list_lock &dev->mutex &devlink->lock_key#3 fill_pool_map-wait-type-override pool_lock irq_context: 0 sb_writers#8 &of->mutex kn->active#52 nsim_bus_dev_list_lock &dev->mutex &devlink->lock_key#3 &devlink_port->type_lock irq_context: 0 sb_writers#8 &of->mutex kn->active#52 nsim_bus_dev_list_lock &dev->mutex &devlink->lock_key#3 stack_depot_init_mutex irq_context: 0 sb_writers#8 &of->mutex kn->active#52 nsim_bus_dev_list_lock &dev->mutex &devlink->lock_key#3 rtnl_mutex bpf_devs_lock irq_context: 0 sb_writers#8 &of->mutex kn->active#52 nsim_bus_dev_list_lock &dev->mutex &devlink->lock_key#3 rtnl_mutex bpf_devs_lock fs_reclaim irq_context: 0 sb_writers#8 &of->mutex kn->active#52 nsim_bus_dev_list_lock &dev->mutex &devlink->lock_key#3 rtnl_mutex bpf_devs_lock fs_reclaim mmu_notifier_invalidate_range_start irq_context: 0 sb_writers#8 &of->mutex kn->active#52 nsim_bus_dev_list_lock &dev->mutex &devlink->lock_key#3 rtnl_mutex bpf_devs_lock rcu_read_lock rhashtable_bucket irq_context: 0 sb_writers#8 &of->mutex kn->active#52 nsim_bus_dev_list_lock &dev->mutex &devlink->lock_key#3 rtnl_mutex pin_fs_lock irq_context: 0 sb_writers#8 &of->mutex kn->active#52 nsim_bus_dev_list_lock &dev->mutex &devlink->lock_key#3 rtnl_mutex &sb->s_type->i_mutex_key#3 irq_context: 0 sb_writers#8 &of->mutex kn->active#52 nsim_bus_dev_list_lock &dev->mutex &devlink->lock_key#3 rtnl_mutex &sb->s_type->i_mutex_key#3 rename_lock.seqcount irq_context: 0 sb_writers#8 &of->mutex kn->active#52 nsim_bus_dev_list_lock &dev->mutex &devlink->lock_key#3 rtnl_mutex &sb->s_type->i_mutex_key#3 fs_reclaim irq_context: 0 sb_writers#8 &of->mutex kn->active#52 nsim_bus_dev_list_lock &dev->mutex &devlink->lock_key#3 rtnl_mutex &sb->s_type->i_mutex_key#3 fs_reclaim mmu_notifier_invalidate_range_start irq_context: 0 sb_writers#8 &of->mutex kn->active#52 nsim_bus_dev_list_lock &dev->mutex &devlink->lock_key#3 rtnl_mutex &sb->s_type->i_mutex_key#3 &dentry->d_lock irq_context: 0 sb_writers#8 &of->mutex kn->active#52 nsim_bus_dev_list_lock &dev->mutex &devlink->lock_key#3 rtnl_mutex &sb->s_type->i_mutex_key#3 rcu_read_lock rename_lock.seqcount irq_context: 0 sb_writers#8 &of->mutex kn->active#52 nsim_bus_dev_list_lock &dev->mutex &devlink->lock_key#3 rtnl_mutex &sb->s_type->i_mutex_key#3 &dentry->d_lock &wq irq_context: 0 sb_writers#8 &of->mutex kn->active#52 nsim_bus_dev_list_lock &dev->mutex &devlink->lock_key#3 rtnl_mutex &sb->s_type->i_mutex_key#3 mmu_notifier_invalidate_range_start irq_context: 0 sb_writers#8 &of->mutex kn->active#52 nsim_bus_dev_list_lock &dev->mutex &devlink->lock_key#3 rtnl_mutex &sb->s_type->i_mutex_key#3 &sb->s_type->i_lock_key#7 irq_context: 0 sb_writers#8 &of->mutex kn->active#52 nsim_bus_dev_list_lock &dev->mutex &devlink->lock_key#3 rtnl_mutex &sb->s_type->i_mutex_key#3 &s->s_inode_list_lock irq_context: 0 sb_writers#8 &of->mutex kn->active#52 nsim_bus_dev_list_lock &dev->mutex &devlink->lock_key#3 rtnl_mutex &sb->s_type->i_mutex_key#3 tk_core.seq.seqcount irq_context: 0 sb_writers#8 &of->mutex kn->active#52 nsim_bus_dev_list_lock &dev->mutex &devlink->lock_key#3 rtnl_mutex &sb->s_type->i_mutex_key#3 &sb->s_type->i_lock_key#7 &dentry->d_lock irq_context: 0 sb_writers#8 &of->mutex kn->active#52 nsim_bus_dev_list_lock &dev->mutex &devlink->lock_key#3 rtnl_mutex &____s->seqcount irq_context: 0 sb_writers#8 &of->mutex kn->active#52 nsim_bus_dev_list_lock &dev->mutex &devlink->lock_key#3 rtnl_mutex &obj_hash[i].lock irq_context: 0 sb_writers#8 &of->mutex kn->active#52 nsim_bus_dev_list_lock &dev->mutex &devlink->lock_key#3 rtnl_mutex fs_reclaim irq_context: 0 sb_writers#8 &of->mutex kn->active#52 nsim_bus_dev_list_lock &dev->mutex &devlink->lock_key#3 rtnl_mutex fs_reclaim mmu_notifier_invalidate_range_start irq_context: 0 sb_writers#8 &of->mutex kn->active#52 nsim_bus_dev_list_lock &dev->mutex &devlink->lock_key#3 rtnl_mutex &xa->xa_lock#3 irq_context: 0 sb_writers#8 &of->mutex kn->active#52 nsim_bus_dev_list_lock &dev->mutex &devlink->lock_key#3 rtnl_mutex &base->lock irq_context: 0 sb_writers#8 &of->mutex kn->active#52 nsim_bus_dev_list_lock &dev->mutex &devlink->lock_key#3 rtnl_mutex &base->lock &obj_hash[i].lock irq_context: 0 sb_writers#8 &of->mutex kn->active#52 nsim_bus_dev_list_lock &dev->mutex &devlink->lock_key#3 rtnl_mutex (work_completion)(&(&devlink_port->type_warn_dw)->work) irq_context: 0 sb_writers#8 &of->mutex kn->active#52 nsim_bus_dev_list_lock &dev->mutex &devlink->lock_key#3 rtnl_mutex &devlink_port->type_lock irq_context: 0 sb_writers#8 &of->mutex kn->active#52 nsim_bus_dev_list_lock &dev->mutex &devlink->lock_key#3 rtnl_mutex nl_table_lock irq_context: 0 sb_writers#8 &of->mutex kn->active#52 nsim_bus_dev_list_lock &dev->mutex &devlink->lock_key#3 rtnl_mutex nl_table_wait.lock irq_context: 0 sb_writers#8 &of->mutex kn->active#52 nsim_bus_dev_list_lock &dev->mutex &devlink->lock_key#3 rtnl_mutex net_rwsem irq_context: 0 sb_writers#8 &of->mutex kn->active#52 nsim_bus_dev_list_lock &dev->mutex &devlink->lock_key#3 rtnl_mutex net_rwsem &list->lock#2 irq_context: 0 sb_writers#8 &of->mutex kn->active#52 nsim_bus_dev_list_lock &dev->mutex &devlink->lock_key#3 rtnl_mutex &tn->lock irq_context: 0 sb_writers#8 &of->mutex kn->active#52 nsim_bus_dev_list_lock &dev->mutex &devlink->lock_key#3 rtnl_mutex &x->wait#9 irq_context: 0 sb_writers#8 &of->mutex kn->active#52 nsim_bus_dev_list_lock &dev->mutex &devlink->lock_key#3 rtnl_mutex &k->list_lock irq_context: 0 sb_writers#8 &of->mutex kn->active#52 nsim_bus_dev_list_lock &dev->mutex &devlink->lock_key#3 rtnl_mutex gdp_mutex irq_context: 0 sb_writers#8 &of->mutex kn->active#52 nsim_bus_dev_list_lock &dev->mutex &devlink->lock_key#3 rtnl_mutex gdp_mutex &k->list_lock irq_context: 0 sb_writers#8 &of->mutex kn->active#52 nsim_bus_dev_list_lock &dev->mutex &devlink->lock_key#3 rtnl_mutex gdp_mutex fs_reclaim irq_context: 0 sb_writers#8 &of->mutex kn->active#52 nsim_bus_dev_list_lock &dev->mutex &devlink->lock_key#3 rtnl_mutex gdp_mutex fs_reclaim mmu_notifier_invalidate_range_start irq_context: 0 sb_writers#8 &of->mutex kn->active#52 nsim_bus_dev_list_lock &dev->mutex &devlink->lock_key#3 rtnl_mutex gdp_mutex lock irq_context: 0 sb_writers#8 &of->mutex kn->active#52 nsim_bus_dev_list_lock &dev->mutex &devlink->lock_key#3 rtnl_mutex gdp_mutex lock kernfs_idr_lock irq_context: 0 sb_writers#8 &of->mutex kn->active#52 nsim_bus_dev_list_lock &dev->mutex &devlink->lock_key#3 rtnl_mutex gdp_mutex &root->kernfs_rwsem irq_context: 0 sb_writers#8 &of->mutex kn->active#52 nsim_bus_dev_list_lock &dev->mutex &devlink->lock_key#3 rtnl_mutex gdp_mutex &root->kernfs_rwsem &root->kernfs_iattr_rwsem irq_context: 0 sb_writers#8 &of->mutex kn->active#52 nsim_bus_dev_list_lock &dev->mutex &devlink->lock_key#3 rtnl_mutex gdp_mutex kobj_ns_type_lock irq_context: 0 sb_writers#8 &of->mutex kn->active#52 nsim_bus_dev_list_lock &dev->mutex &devlink->lock_key#3 rtnl_mutex lock irq_context: 0 sb_writers#8 &of->mutex kn->active#52 nsim_bus_dev_list_lock &dev->mutex &devlink->lock_key#3 rtnl_mutex lock kernfs_idr_lock irq_context: 0 sb_writers#8 &of->mutex kn->active#52 nsim_bus_dev_list_lock &dev->mutex &devlink->lock_key#3 rtnl_mutex &root->kernfs_rwsem irq_context: 0 sb_writers#8 &of->mutex kn->active#52 nsim_bus_dev_list_lock &dev->mutex &devlink->lock_key#3 rtnl_mutex &root->kernfs_rwsem &root->kernfs_iattr_rwsem irq_context: 0 sb_writers#8 &of->mutex kn->active#52 nsim_bus_dev_list_lock &dev->mutex &devlink->lock_key#3 rtnl_mutex bus_type_sem irq_context: 0 sb_writers#8 &of->mutex kn->active#52 nsim_bus_dev_list_lock &dev->mutex &devlink->lock_key#3 rtnl_mutex sysfs_symlink_target_lock irq_context: 0 sb_writers#8 &of->mutex kn->active#52 nsim_bus_dev_list_lock &dev->mutex &devlink->lock_key#3 rtnl_mutex &c->lock irq_context: 0 sb_writers#8 &of->mutex kn->active#52 nsim_bus_dev_list_lock &dev->mutex &devlink->lock_key#3 rtnl_mutex &root->kernfs_rwsem irq_context: 0 sb_writers#8 &of->mutex kn->active#52 nsim_bus_dev_list_lock &dev->mutex &devlink->lock_key#3 rtnl_mutex &dev->power.lock irq_context: 0 sb_writers#8 &of->mutex kn->active#52 nsim_bus_dev_list_lock &dev->mutex &devlink->lock_key#3 rtnl_mutex dpm_list_mtx irq_context: 0 sb_writers#8 &of->mutex kn->active#52 nsim_bus_dev_list_lock &dev->mutex &devlink->lock_key#3 rtnl_mutex uevent_sock_mutex irq_context: 0 sb_writers#8 &of->mutex kn->active#52 nsim_bus_dev_list_lock &dev->mutex &devlink->lock_key#3 rtnl_mutex uevent_sock_mutex fs_reclaim irq_context: 0 sb_writers#8 &of->mutex kn->active#52 nsim_bus_dev_list_lock &dev->mutex &devlink->lock_key#3 rtnl_mutex uevent_sock_mutex fs_reclaim mmu_notifier_invalidate_range_start irq_context: 0 sb_writers#8 &of->mutex kn->active#52 nsim_bus_dev_list_lock &dev->mutex &devlink->lock_key#3 rtnl_mutex uevent_sock_mutex nl_table_lock irq_context: 0 sb_writers#8 &of->mutex kn->active#52 nsim_bus_dev_list_lock &dev->mutex &devlink->lock_key#3 rtnl_mutex uevent_sock_mutex &obj_hash[i].lock irq_context: 0 sb_writers#8 &of->mutex kn->active#52 nsim_bus_dev_list_lock &dev->mutex &devlink->lock_key#3 rtnl_mutex uevent_sock_mutex nl_table_wait.lock irq_context: 0 sb_writers#8 &of->mutex kn->active#52 nsim_bus_dev_list_lock &dev->mutex &devlink->lock_key#3 rtnl_mutex &k->k_lock irq_context: 0 sb_writers#8 &of->mutex kn->active#52 nsim_bus_dev_list_lock &dev->mutex &devlink->lock_key#3 rtnl_mutex subsys mutex#17 irq_context: 0 sb_writers#8 &of->mutex kn->active#52 nsim_bus_dev_list_lock &dev->mutex &devlink->lock_key#3 rtnl_mutex subsys mutex#17 &k->k_lock irq_context: 0 sb_writers#8 &of->mutex kn->active#52 nsim_bus_dev_list_lock &dev->mutex &devlink->lock_key#3 rtnl_mutex &dir->lock#2 irq_context: 0 sb_writers#8 &of->mutex kn->active#52 nsim_bus_dev_list_lock &dev->mutex &devlink->lock_key#3 rtnl_mutex lock kernfs_idr_lock pool_lock#2 irq_context: 0 sb_writers#8 &of->mutex kn->active#52 nsim_bus_dev_list_lock &dev->mutex &devlink->lock_key#3 rtnl_mutex dev_hotplug_mutex irq_context: 0 sb_writers#8 &of->mutex kn->active#52 nsim_bus_dev_list_lock &dev->mutex &devlink->lock_key#3 rtnl_mutex dev_hotplug_mutex &dev->power.lock irq_context: 0 sb_writers#8 &of->mutex kn->active#52 nsim_bus_dev_list_lock &dev->mutex &devlink->lock_key#3 rtnl_mutex dev_base_lock irq_context: 0 sb_writers#8 &of->mutex kn->active#52 nsim_bus_dev_list_lock &dev->mutex &devlink->lock_key#3 rtnl_mutex input_pool.lock irq_context: 0 sb_writers#8 &of->mutex kn->active#52 nsim_bus_dev_list_lock &dev->mutex &devlink->lock_key#3 rtnl_mutex rcu_read_lock &pool->lock/1 irq_context: 0 sb_writers#8 &of->mutex kn->active#52 nsim_bus_dev_list_lock &dev->mutex &devlink->lock_key#3 rtnl_mutex rcu_read_lock &pool->lock/1 &obj_hash[i].lock irq_context: 0 sb_writers#8 &of->mutex kn->active#52 nsim_bus_dev_list_lock &dev->mutex &devlink->lock_key#3 rtnl_mutex rcu_read_lock &pool->lock/1 &p->pi_lock irq_context: 0 sb_writers#8 &of->mutex kn->active#52 nsim_bus_dev_list_lock &dev->mutex &devlink->lock_key#3 rtnl_mutex &n->list_lock irq_context: 0 sb_writers#8 &of->mutex kn->active#52 nsim_bus_dev_list_lock &dev->mutex &devlink->lock_key#3 rtnl_mutex &n->list_lock &c->lock irq_context: 0 sb_writers#8 &of->mutex kn->active#52 nsim_bus_dev_list_lock &dev->mutex &devlink->lock_key#3 rtnl_mutex batched_entropy_u32.lock irq_context: 0 sb_writers#8 &of->mutex kn->active#52 nsim_bus_dev_list_lock &dev->mutex &devlink->lock_key#3 rtnl_mutex &tbl->lock irq_context: 0 sb_writers#8 &of->mutex kn->active#52 nsim_bus_dev_list_lock &dev->mutex &devlink->lock_key#3 rtnl_mutex &pcp->lock &zone->lock irq_context: 0 sb_writers#8 &of->mutex kn->active#52 nsim_bus_dev_list_lock &dev->mutex &devlink->lock_key#3 rtnl_mutex &pcp->lock &zone->lock &____s->seqcount irq_context: 0 sb_writers#8 &of->mutex kn->active#52 nsim_bus_dev_list_lock &dev->mutex &devlink->lock_key#3 rtnl_mutex sysctl_lock irq_context: 0 sb_writers#8 &of->mutex kn->active#52 nsim_bus_dev_list_lock &dev->mutex &devlink->lock_key#3 rtnl_mutex failover_lock irq_context: 0 sb_writers#8 &of->mutex kn->active#52 nsim_bus_dev_list_lock &dev->mutex &devlink->lock_key#3 rtnl_mutex pcpu_alloc_mutex irq_context: 0 sb_writers#8 &of->mutex kn->active#52 nsim_bus_dev_list_lock &dev->mutex &devlink->lock_key#3 rtnl_mutex pcpu_alloc_mutex pcpu_lock irq_context: 0 sb_writers#8 &of->mutex kn->active#52 nsim_bus_dev_list_lock &dev->mutex &devlink->lock_key#3 rtnl_mutex proc_subdir_lock irq_context: 0 sb_writers#8 &of->mutex kn->active#52 nsim_bus_dev_list_lock &dev->mutex &devlink->lock_key#3 rtnl_mutex proc_inum_ida.xa_lock irq_context: 0 sb_writers#8 &of->mutex kn->active#52 nsim_bus_dev_list_lock &dev->mutex &devlink->lock_key#3 rtnl_mutex proc_subdir_lock irq_context: 0 sb_writers#8 &of->mutex kn->active#52 nsim_bus_dev_list_lock &dev->mutex &devlink->lock_key#3 rtnl_mutex &obj_hash[i].lock pool_lock irq_context: 0 sb_writers#8 &of->mutex kn->active#52 nsim_bus_dev_list_lock &dev->mutex &devlink->lock_key#3 rtnl_mutex fill_pool_map-wait-type-override pool_lock#2 irq_context: 0 sb_writers#8 &of->mutex kn->active#52 nsim_bus_dev_list_lock &dev->mutex &devlink->lock_key#3 rtnl_mutex fill_pool_map-wait-type-override pool_lock irq_context: 0 sb_writers#8 &of->mutex kn->active#52 nsim_bus_dev_list_lock &dev->mutex &devlink->lock_key#3 rtnl_mutex &idev->mc_lock irq_context: 0 sb_writers#8 &of->mutex kn->active#52 nsim_bus_dev_list_lock &dev->mutex &devlink->lock_key#3 rtnl_mutex &idev->mc_lock fs_reclaim irq_context: 0 sb_writers#8 &of->mutex kn->active#52 nsim_bus_dev_list_lock &dev->mutex &devlink->lock_key#3 rtnl_mutex &idev->mc_lock fs_reclaim mmu_notifier_invalidate_range_start irq_context: 0 sb_writers#8 &of->mutex kn->active#52 nsim_bus_dev_list_lock &dev->mutex &devlink->lock_key#3 rtnl_mutex &idev->mc_lock &obj_hash[i].lock irq_context: 0 sb_writers#8 &of->mutex kn->active#52 nsim_bus_dev_list_lock &dev->mutex &devlink->lock_key#3 rtnl_mutex &idev->mc_lock _xmit_ETHER irq_context: 0 sb_writers#8 &of->mutex kn->active#52 nsim_bus_dev_list_lock &dev->mutex &devlink->lock_key#3 rtnl_mutex &idev->mc_lock _xmit_ETHER &c->lock irq_context: 0 sb_writers#8 &of->mutex kn->active#52 nsim_bus_dev_list_lock &dev->mutex &devlink->lock_key#3 rtnl_mutex &idev->mc_lock _xmit_ETHER &____s->seqcount irq_context: 0 sb_writers#8 &of->mutex kn->active#52 nsim_bus_dev_list_lock &dev->mutex &devlink->lock_key#3 rtnl_mutex &pnettable->lock irq_context: 0 sb_writers#8 &of->mutex kn->active#52 nsim_bus_dev_list_lock &dev->mutex &devlink->lock_key#3 rtnl_mutex smc_ib_devices.mutex irq_context: 0 sb_writers#8 &of->mutex kn->active#52 nsim_bus_dev_list_lock &dev->mutex &devlink->lock_key#3 rtnl_mutex &vn->sock_lock irq_context: 0 sb_writers#8 &of->mutex kn->active#52 nsim_bus_dev_list_lock &dev->mutex &devlink->lock_key#3 rtnl_mutex rcu_read_lock &ndev->lock irq_context: 0 sb_writers#8 &of->mutex kn->active#52 nsim_bus_dev_list_lock &dev->mutex &devlink->lock_key#3 &rq->__lock irq_context: 0 (wq_completion)ipv6_addrconf (work_completion)(&(&ifa->dad_work)->work) rtnl_mutex &tb->tb6_lock rcu_read_lock rcu_read_lock &pool->lock &p->pi_lock irq_context: 0 (wq_completion)ipv6_addrconf (work_completion)(&(&ifa->dad_work)->work) rtnl_mutex &tb->tb6_lock rcu_read_lock rcu_read_lock &pool->lock &p->pi_lock &rq->__lock irq_context: 0 (wq_completion)ipv6_addrconf (work_completion)(&(&ifa->dad_work)->work) rtnl_mutex &tb->tb6_lock rcu_read_lock rcu_read_lock &pool->lock &p->pi_lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)ipv6_addrconf (work_completion)(&(&ifa->dad_work)->work) rtnl_mutex &ndev->lock &ifa->lock batched_entropy_u32.lock crngs.lock irq_context: 0 sb_writers#8 &of->mutex kn->active#52 nsim_bus_dev_list_lock &dev->mutex &devlink->lock_key#3 &xa->xa_lock#16 &c->lock irq_context: 0 sb_writers#8 &of->mutex kn->active#52 nsim_bus_dev_list_lock &dev->mutex &devlink->lock_key#3 &xa->xa_lock#16 &____s->seqcount irq_context: 0 sb_writers#8 &of->mutex kn->active#52 nsim_bus_dev_list_lock &dev->mutex &devlink->lock_key#3 &xa->xa_lock#16 pool_lock#2 irq_context: 0 (wq_completion)bat_events (work_completion)(&(&bat_priv->mcast.work)->work) rcu_read_lock &c->lock irq_context: 0 (wq_completion)bat_events (work_completion)(&(&bat_priv->mcast.work)->work) rcu_read_lock &____s->seqcount irq_context: 0 sb_writers#8 &of->mutex kn->active#52 nsim_bus_dev_list_lock &dev->mutex &devlink->lock_key#3 &sb->s_type->i_mutex_key#3 rcu_read_lock rcu_node_0 irq_context: 0 sb_writers#8 &of->mutex kn->active#52 nsim_bus_dev_list_lock &dev->mutex &devlink->lock_key#3 &sb->s_type->i_mutex_key#3 rcu_read_lock &rq->__lock irq_context: 0 rtnl_mutex dev_addr_sem &br->lock irq_context: 0 rtnl_mutex dev_addr_sem &br->lock &br->hash_lock irq_context: 0 rtnl_mutex dev_addr_sem &br->lock &br->hash_lock pool_lock#2 irq_context: 0 rtnl_mutex dev_addr_sem &br->lock &br->hash_lock rcu_read_lock rhashtable_bucket irq_context: 0 rtnl_mutex dev_addr_sem &br->lock &br->hash_lock &c->lock irq_context: 0 rtnl_mutex dev_addr_sem &br->lock &br->hash_lock nl_table_lock irq_context: 0 rtnl_mutex dev_addr_sem &br->lock &br->hash_lock &obj_hash[i].lock irq_context: 0 rtnl_mutex dev_addr_sem &br->lock &br->hash_lock nl_table_wait.lock irq_context: 0 rtnl_mutex &bridge_netdev_addr_lock_key/1 irq_context: 0 rtnl_mutex &bridge_netdev_addr_lock_key/1 pool_lock#2 irq_context: 0 rtnl_mutex &idev->mc_lock &bridge_netdev_addr_lock_key/1 irq_context: 0 rtnl_mutex &idev->mc_lock &bridge_netdev_addr_lock_key/1 pool_lock#2 irq_context: 0 (wq_completion)ipv6_addrconf (work_completion)(&(&ifa->dad_work)->work) rtnl_mutex &idev->mc_lock &bridge_netdev_addr_lock_key/1 irq_context: 0 (wq_completion)ipv6_addrconf (work_completion)(&(&ifa->dad_work)->work) rtnl_mutex &ifa->lock rcu_read_lock &pool->lock irq_context: 0 (wq_completion)ipv6_addrconf (work_completion)(&(&ifa->dad_work)->work) rtnl_mutex &ifa->lock rcu_read_lock &pool->lock &obj_hash[i].lock irq_context: 0 sb_writers#8 &of->mutex kn->active#52 nsim_bus_dev_list_lock &dev->mutex &devlink->lock_key#3 rtnl_mutex &sb->s_type->i_mutex_key#3 &c->lock irq_context: 0 sb_writers#8 &of->mutex kn->active#52 nsim_bus_dev_list_lock &dev->mutex &devlink->lock_key#3 rtnl_mutex &sb->s_type->i_mutex_key#3 &____s->seqcount irq_context: 0 sb_writers#8 &of->mutex kn->active#52 nsim_bus_dev_list_lock &dev->mutex &devlink->lock_key#3 rtnl_mutex &sb->s_type->i_mutex_key#3 pool_lock#2 irq_context: 0 sb_writers#8 &of->mutex kn->active#52 nsim_bus_dev_list_lock &dev->mutex &devlink->lock_key#3 rtnl_mutex uevent_sock_mutex &c->lock irq_context: 0 sb_writers#8 &of->mutex kn->active#52 nsim_bus_dev_list_lock &dev->mutex &devlink->lock_key#3 rtnl_mutex pool_lock#2 irq_context: 0 sb_writers#8 &of->mutex kn->active#52 nsim_bus_dev_list_lock &dev->mutex &devlink->lock_key#3 rtnl_mutex rcu_read_lock &pool->lock/1 &p->pi_lock &rq->__lock irq_context: 0 sb_writers#8 &of->mutex kn->active#52 nsim_bus_dev_list_lock &dev->mutex &devlink->lock_key#3 rtnl_mutex rcu_read_lock &pool->lock/1 &p->pi_lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: softirq (&brmctx->ip6_own_query.timer) irq_context: softirq (&brmctx->ip6_own_query.timer) &br->multicast_lock irq_context: softirq (&brmctx->ip4_own_query.timer) irq_context: softirq (&brmctx->ip4_own_query.timer) &br->multicast_lock irq_context: softirq (&app->join_timer)#2 fill_pool_map-wait-type-override pool_lock#2 irq_context: 0 sb_writers#8 &of->mutex kn->active#52 nsim_bus_dev_list_lock &dev->mutex &devlink->lock_key#3 rtnl_mutex fill_pool_map-wait-type-override &c->lock irq_context: 0 sb_writers#8 &of->mutex kn->active#52 nsim_bus_dev_list_lock &dev->mutex &devlink->lock_key#3 rtnl_mutex fill_pool_map-wait-type-override &____s->seqcount irq_context: softirq (&app->join_timer)#2 fill_pool_map-wait-type-override pool_lock irq_context: 0 sb_writers#8 &of->mutex kn->active#52 nsim_bus_dev_list_lock &dev->mutex &devlink->lock_key#3 rtnl_mutex &idev->mc_lock _xmit_ETHER batched_entropy_u8.lock irq_context: 0 sb_writers#8 &of->mutex kn->active#52 nsim_bus_dev_list_lock &dev->mutex &devlink->lock_key#3 rtnl_mutex &idev->mc_lock _xmit_ETHER kfence_freelist_lock irq_context: 0 sb_writers#8 &of->mutex kn->active#52 nsim_bus_dev_list_lock &dev->mutex &devlink->lock_key#3 &p->pi_lock &rq->__lock &cfs_rq->removed.lock irq_context: 0 (wq_completion)ipv6_addrconf (work_completion)(&(&ifa->dad_work)->work) rtnl_mutex rcu_read_lock rcu_read_lock &c->lock irq_context: 0 (wq_completion)ipv6_addrconf (work_completion)(&(&ifa->dad_work)->work) rtnl_mutex rcu_read_lock rcu_read_lock &____s->seqcount irq_context: 0 rtnl_mutex devnet_rename_sem &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)events (work_completion)(&data->fib_event_work) &data->fib_lock &rq->__lock &cfs_rq->removed.lock irq_context: 0 rtnl_mutex (inetaddr_chain).rwsem rcu_read_lock rcu_read_lock &pool->lock irq_context: 0 rtnl_mutex (inetaddr_chain).rwsem rcu_read_lock rcu_read_lock &pool->lock &obj_hash[i].lock irq_context: 0 rtnl_mutex (inetaddr_chain).rwsem rcu_read_lock rcu_read_lock &pool->lock pool_lock#2 irq_context: 0 rtnl_mutex (inetaddr_chain).rwsem rcu_read_lock rcu_read_lock &pool->lock &p->pi_lock irq_context: 0 cb_lock genl_mutex rtnl_mutex rtnl_mutex.wait_lock irq_context: 0 cb_lock genl_mutex rtnl_mutex &rq->__lock irq_context: 0 cb_lock genl_mutex rtnl_mutex &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 rtnl_mutex (inetaddr_chain).rwsem &net->sctp.local_addr_lock &net->sctp.addr_wq_lock &n->list_lock irq_context: 0 rtnl_mutex (inetaddr_chain).rwsem &net->sctp.local_addr_lock &net->sctp.addr_wq_lock &n->list_lock &c->lock irq_context: 0 sb_writers#8 &of->mutex kn->active#52 nsim_bus_dev_list_lock &dev->mutex &devlink->lock_key#3 rtnl_mutex &sb->s_type->i_mutex_key#3 &n->list_lock irq_context: 0 sb_writers#8 &of->mutex kn->active#52 nsim_bus_dev_list_lock &dev->mutex &devlink->lock_key#3 rtnl_mutex &sb->s_type->i_mutex_key#3 &n->list_lock &c->lock irq_context: 0 sb_writers#8 &of->mutex kn->active#52 nsim_bus_dev_list_lock &dev->mutex &devlink->lock_key#3 rtnl_mutex uevent_sock_mutex &n->list_lock irq_context: 0 sb_writers#8 &of->mutex kn->active#52 nsim_bus_dev_list_lock &dev->mutex &devlink->lock_key#3 rtnl_mutex uevent_sock_mutex &n->list_lock &c->lock irq_context: softirq (&in_dev->mr_ifc_timer) irq_context: softirq (&in_dev->mr_ifc_timer) rcu_read_lock &in_dev->mc_tomb_lock irq_context: softirq (&in_dev->mr_ifc_timer) rcu_read_lock &im->lock irq_context: softirq (&in_dev->mr_ifc_timer) rcu_read_lock &im->lock pool_lock#2 irq_context: softirq (&in_dev->mr_ifc_timer) rcu_read_lock &im->lock rcu_read_lock pool_lock#2 irq_context: softirq (&in_dev->mr_ifc_timer) rcu_read_lock &im->lock rcu_read_lock &dir->lock#2 irq_context: softirq (&in_dev->mr_ifc_timer) rcu_read_lock &im->lock rcu_read_lock &ul->lock irq_context: softirq (&in_dev->mr_ifc_timer) rcu_read_lock &____s->seqcount#7 irq_context: softirq (&in_dev->mr_ifc_timer) rcu_read_lock pool_lock#2 irq_context: 0 (wq_completion)ipv6_addrconf (work_completion)(&(&ifa->dad_work)->work) rtnl_mutex &tb->tb6_lock &____s->seqcount irq_context: softirq (&in_dev->mr_ifc_timer) rcu_read_lock &nf_conntrack_locks[i] irq_context: softirq (&in_dev->mr_ifc_timer) rcu_read_lock &nf_conntrack_locks[i] &nf_conntrack_locks[i]/1 irq_context: softirq (&in_dev->mr_ifc_timer) rcu_read_lock &nf_conntrack_locks[i] &nf_conntrack_locks[i]/1 batched_entropy_u8.lock irq_context: 0 (wq_completion)ipv6_addrconf (work_completion)(&(&ifa->dad_work)->work) rtnl_mutex &tb->tb6_lock &n->list_lock irq_context: 0 (wq_completion)ipv6_addrconf (work_completion)(&(&ifa->dad_work)->work) rtnl_mutex &tb->tb6_lock &n->list_lock &c->lock irq_context: softirq (&in_dev->mr_ifc_timer) rcu_read_lock &obj_hash[i].lock irq_context: softirq (&in_dev->mr_ifc_timer) rcu_read_lock &dir->lock#2 irq_context: softirq (&in_dev->mr_ifc_timer) rcu_read_lock &tbl->lock irq_context: softirq (&in_dev->mr_ifc_timer) rcu_read_lock &n->lock irq_context: softirq (&in_dev->mr_ifc_timer) rcu_read_lock &____s->seqcount#9 irq_context: softirq (&in_dev->mr_ifc_timer) rcu_read_lock rcu_read_lock_bh noop_qdisc.q.lock irq_context: softirq (&in_dev->mr_ifc_timer) rcu_read_lock rcu_read_lock_bh &obj_hash[i].lock irq_context: softirq (&in_dev->mr_ifc_timer) rcu_read_lock rcu_read_lock_bh pool_lock#2 irq_context: softirq (&in_dev->mr_ifc_timer) rcu_read_lock rcu_read_lock_bh &____s->seqcount irq_context: softirq (&in_dev->mr_ifc_timer) rcu_read_lock rcu_read_lock_bh rcu_read_lock pool_lock#2 irq_context: softirq (&in_dev->mr_ifc_timer) batched_entropy_u32.lock irq_context: softirq (&in_dev->mr_ifc_timer) &obj_hash[i].lock irq_context: softirq (&in_dev->mr_ifc_timer) &base->lock irq_context: softirq (&in_dev->mr_ifc_timer) &base->lock &obj_hash[i].lock irq_context: 0 cb_lock genl_mutex rtnl_mutex &wg->device_update_lock &pcp->lock &zone->lock irq_context: 0 cb_lock genl_mutex rtnl_mutex &wg->device_update_lock &pcp->lock &zone->lock &____s->seqcount irq_context: 0 cb_lock genl_mutex &p->pi_lock &rq->__lock irq_context: 0 cb_lock genl_mutex &p->pi_lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 sb_writers#8 &of->mutex kn->active#52 nsim_bus_dev_list_lock &dev->mutex uevent_sock_mutex nl_table_wait.lock &p->pi_lock irq_context: 0 sb_writers#8 &of->mutex kn->active#52 nsim_bus_dev_list_lock &dev->mutex uevent_sock_mutex nl_table_wait.lock &p->pi_lock &rq->__lock irq_context: 0 sb_writers#8 &of->mutex kn->active#52 nsim_bus_dev_list_lock &dev->mutex uevent_sock_mutex nl_table_wait.lock &p->pi_lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 sb_writers#8 &of->mutex kn->active#52 &rq->__lock irq_context: 0 sb_writers#8 &of->mutex kn->active#52 nsim_bus_dev_list_lock lock kernfs_idr_lock pool_lock#2 irq_context: 0 sb_writers#8 &of->mutex kn->active#52 nsim_bus_dev_list_lock uevent_sock_mutex nl_table_wait.lock &p->pi_lock irq_context: 0 sb_writers#8 &of->mutex kn->active#52 nsim_bus_dev_list_lock &dev->mutex &devlink->lock_key#4 irq_context: 0 sb_writers#8 &of->mutex kn->active#52 nsim_bus_dev_list_lock &dev->mutex &devlink->lock_key#4 crngs.lock irq_context: 0 sb_writers#8 &of->mutex kn->active#52 nsim_bus_dev_list_lock &dev->mutex &devlink->lock_key#4 fs_reclaim irq_context: 0 sb_writers#8 &of->mutex kn->active#52 nsim_bus_dev_list_lock &dev->mutex &devlink->lock_key#4 fs_reclaim mmu_notifier_invalidate_range_start irq_context: 0 sb_writers#8 &of->mutex kn->active#52 nsim_bus_dev_list_lock &dev->mutex &devlink->lock_key#4 devlinks.xa_lock irq_context: 0 sb_writers#8 &of->mutex kn->active#52 nsim_bus_dev_list_lock &dev->mutex &devlink->lock_key#4 &obj_hash[i].lock irq_context: 0 sb_writers#8 &of->mutex kn->active#52 nsim_bus_dev_list_lock &dev->mutex &devlink->lock_key#4 nl_table_lock irq_context: 0 sb_writers#8 &of->mutex kn->active#52 nsim_bus_dev_list_lock &dev->mutex &devlink->lock_key#4 nl_table_wait.lock irq_context: 0 sb_writers#8 &of->mutex kn->active#52 nsim_bus_dev_list_lock &dev->mutex &devlink->lock_key#4 nl_table_wait.lock &p->pi_lock irq_context: 0 sb_writers#8 &of->mutex kn->active#52 nsim_bus_dev_list_lock &dev->mutex &devlink->lock_key#4 &xa->xa_lock#16 irq_context: 0 sb_writers#8 &of->mutex kn->active#52 nsim_bus_dev_list_lock &dev->mutex &devlink->lock_key#4 &c->lock irq_context: 0 sb_writers#8 &of->mutex kn->active#52 nsim_bus_dev_list_lock &dev->mutex &devlink->lock_key#4 &____s->seqcount irq_context: 0 rtnl_mutex nl_table_wait.lock &p->pi_lock irq_context: 0 sb_writers#8 &of->mutex kn->active#52 nsim_bus_dev_list_lock &dev->mutex &devlink->lock_key#4 pcpu_alloc_mutex irq_context: 0 sb_writers#8 &of->mutex kn->active#52 nsim_bus_dev_list_lock &dev->mutex &devlink->lock_key#4 pcpu_alloc_mutex pcpu_lock irq_context: 0 sb_writers#8 &of->mutex kn->active#52 nsim_bus_dev_list_lock &dev->mutex &devlink->lock_key#4 remove_cache_srcu irq_context: 0 sb_writers#8 &of->mutex kn->active#52 nsim_bus_dev_list_lock &dev->mutex &devlink->lock_key#4 remove_cache_srcu quarantine_lock irq_context: 0 sb_writers#8 &of->mutex kn->active#52 nsim_bus_dev_list_lock &dev->mutex &devlink->lock_key#4 &pcp->lock &zone->lock irq_context: 0 sb_writers#8 &of->mutex kn->active#52 nsim_bus_dev_list_lock &dev->mutex &devlink->lock_key#4 &pcp->lock &zone->lock &____s->seqcount irq_context: 0 sb_writers#8 &of->mutex kn->active#52 nsim_bus_dev_list_lock &dev->mutex &devlink->lock_key#4 &base->lock irq_context: 0 sb_writers#8 &of->mutex kn->active#52 nsim_bus_dev_list_lock &dev->mutex &devlink->lock_key#4 &base->lock &obj_hash[i].lock irq_context: 0 sb_writers#8 &of->mutex kn->active#52 nsim_bus_dev_list_lock &dev->mutex &devlink->lock_key#4 pin_fs_lock irq_context: 0 sb_writers#8 &of->mutex kn->active#52 nsim_bus_dev_list_lock &dev->mutex &devlink->lock_key#4 &sb->s_type->i_mutex_key#3 irq_context: 0 sb_writers#8 &of->mutex kn->active#52 nsim_bus_dev_list_lock &dev->mutex &devlink->lock_key#4 &sb->s_type->i_mutex_key#3 rename_lock.seqcount irq_context: 0 sb_writers#8 &of->mutex kn->active#52 nsim_bus_dev_list_lock &dev->mutex &devlink->lock_key#4 &sb->s_type->i_mutex_key#3 fs_reclaim irq_context: 0 sb_writers#8 &of->mutex kn->active#52 nsim_bus_dev_list_lock &dev->mutex &devlink->lock_key#4 &sb->s_type->i_mutex_key#3 fs_reclaim mmu_notifier_invalidate_range_start irq_context: 0 sb_writers#8 &of->mutex kn->active#52 nsim_bus_dev_list_lock &dev->mutex &devlink->lock_key#4 &sb->s_type->i_mutex_key#3 &dentry->d_lock irq_context: 0 sb_writers#8 &of->mutex kn->active#52 nsim_bus_dev_list_lock &dev->mutex &devlink->lock_key#4 &sb->s_type->i_mutex_key#3 rcu_read_lock rename_lock.seqcount irq_context: 0 sb_writers#8 &of->mutex kn->active#52 nsim_bus_dev_list_lock &dev->mutex &devlink->lock_key#4 &sb->s_type->i_mutex_key#3 &dentry->d_lock &wq irq_context: 0 sb_writers#8 &of->mutex kn->active#52 nsim_bus_dev_list_lock &dev->mutex &devlink->lock_key#4 &sb->s_type->i_mutex_key#3 mmu_notifier_invalidate_range_start irq_context: 0 sb_writers#8 &of->mutex kn->active#52 nsim_bus_dev_list_lock &dev->mutex &devlink->lock_key#4 &sb->s_type->i_mutex_key#3 &sb->s_type->i_lock_key#7 irq_context: 0 sb_writers#8 &of->mutex kn->active#52 nsim_bus_dev_list_lock &dev->mutex &devlink->lock_key#4 &sb->s_type->i_mutex_key#3 &s->s_inode_list_lock irq_context: 0 sb_writers#8 &of->mutex kn->active#52 nsim_bus_dev_list_lock &dev->mutex &devlink->lock_key#4 &sb->s_type->i_mutex_key#3 tk_core.seq.seqcount irq_context: 0 sb_writers#8 &of->mutex kn->active#52 nsim_bus_dev_list_lock &dev->mutex &devlink->lock_key#4 &sb->s_type->i_mutex_key#3 &sb->s_type->i_lock_key#7 &dentry->d_lock irq_context: 0 sb_writers#8 &of->mutex kn->active#52 nsim_bus_dev_list_lock &dev->mutex &devlink->lock_key#4 &sb->s_type->i_mutex_key#3 &c->lock irq_context: 0 sb_writers#8 &of->mutex kn->active#52 nsim_bus_dev_list_lock &dev->mutex &devlink->lock_key#4 &sb->s_type->i_mutex_key#3 &____s->seqcount irq_context: 0 sb_writers#8 &of->mutex kn->active#52 nsim_bus_dev_list_lock &dev->mutex &devlink->lock_key#4 &sb->s_type->i_mutex_key#3 pool_lock#2 irq_context: 0 sb_writers#8 &of->mutex kn->active#52 nsim_bus_dev_list_lock &dev->mutex &devlink->lock_key#4 pool_lock#2 irq_context: 0 sb_writers#8 &of->mutex kn->active#52 nsim_bus_dev_list_lock &dev->mutex &devlink->lock_key#4 &sb->s_type->i_mutex_key#3 &n->list_lock irq_context: 0 sb_writers#8 &of->mutex kn->active#52 nsim_bus_dev_list_lock &dev->mutex &devlink->lock_key#4 &sb->s_type->i_mutex_key#3 &n->list_lock &c->lock irq_context: 0 sb_writers#8 &of->mutex kn->active#52 nsim_bus_dev_list_lock &dev->mutex &devlink->lock_key#4 batched_entropy_u32.lock irq_context: 0 sb_writers#8 &of->mutex kn->active#52 nsim_bus_dev_list_lock &dev->mutex &devlink->lock_key#4 rtnl_mutex irq_context: 0 sb_writers#8 &of->mutex kn->active#52 nsim_bus_dev_list_lock &dev->mutex &devlink->lock_key#4 rtnl_mutex rtnl_mutex.wait_lock irq_context: 0 sb_writers#8 &of->mutex kn->active#52 nsim_bus_dev_list_lock &dev->mutex &devlink->lock_key#4 rtnl_mutex &rq->__lock irq_context: 0 sb_writers#8 &of->mutex kn->active#52 nsim_bus_dev_list_lock &dev->mutex &devlink->lock_key#4 rtnl_mutex &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 cb_lock nlk_cb_mutex-GENERIC &devlink->lock_key#3 irq_context: 0 cb_lock nlk_cb_mutex-GENERIC &devlink->lock_key#3 &devlink_port->type_lock irq_context: 0 rtnl_mutex _xmit_TUNNEL irq_context: 0 rtnl_mutex &ndev->lock &tb->tb6_lock nl_table_wait.lock &p->pi_lock irq_context: 0 rtnl_mutex &tb->tb6_lock &pcp->lock &zone->lock irq_context: 0 rtnl_mutex &tb->tb6_lock &pcp->lock &zone->lock &____s->seqcount irq_context: 0 rtnl_mutex &tb->tb6_lock nl_table_wait.lock &p->pi_lock irq_context: 0 (wq_completion)ipv6_addrconf (work_completion)(&(&ifa->dad_work)->work) rtnl_mutex nl_table_wait.lock &p->pi_lock irq_context: 0 (wq_completion)ipv6_addrconf (work_completion)(&(&ifa->dad_work)->work) rtnl_mutex &tb->tb6_lock nl_table_wait.lock &p->pi_lock irq_context: 0 rtnl_mutex (inetaddr_chain).rwsem rcu_read_lock &c->lock irq_context: 0 rtnl_mutex (inetaddr_chain).rwsem &n->list_lock irq_context: 0 rtnl_mutex (inetaddr_chain).rwsem &n->list_lock &c->lock irq_context: 0 sb_writers#8 &of->mutex kn->active#52 nsim_bus_dev_list_lock &dev->mutex &devlink->lock_key#4 rtnl_mutex &(&net->nexthop.notifier_chain)->rwsem irq_context: 0 sb_writers#8 &of->mutex kn->active#52 nsim_bus_dev_list_lock &dev->mutex &devlink->lock_key#4 rtnl_mutex.wait_lock irq_context: 0 sb_writers#8 &of->mutex kn->active#52 nsim_bus_dev_list_lock &dev->mutex &devlink->lock_key#4 &p->pi_lock irq_context: 0 sb_writers#8 &of->mutex kn->active#52 nsim_bus_dev_list_lock &dev->mutex &devlink->lock_key#4 &p->pi_lock &rq->__lock irq_context: 0 sb_writers#8 &of->mutex kn->active#52 nsim_bus_dev_list_lock &dev->mutex &devlink->lock_key#4 &p->pi_lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 rtnl_mutex devnet_rename_sem uevent_sock_mutex &c->lock irq_context: 0 (wq_completion)events (work_completion)(&(&nsim_dev->trap_data->trap_report_dw)->work) &devlink->lock_key#3 &nsim_trap_data->trap_lock irq_context: 0 (wq_completion)events (work_completion)(&(&nsim_dev->trap_data->trap_report_dw)->work) &devlink->lock_key#3 &nsim_trap_data->trap_lock crngs.lock irq_context: 0 (wq_completion)events (work_completion)(&(&nsim_dev->trap_data->trap_report_dw)->work) &devlink->lock_key#3 &nsim_trap_data->trap_lock &nsim_dev->fa_cookie_lock irq_context: 0 (wq_completion)events (work_completion)(&(&nsim_dev->trap_data->trap_report_dw)->work) &devlink->lock_key#3 &nsim_trap_data->trap_lock &obj_hash[i].lock irq_context: 0 (wq_completion)events (work_completion)(&(&nsim_dev->trap_data->trap_report_dw)->work) &devlink->lock_key#3 &nsim_trap_data->trap_lock &c->lock irq_context: 0 (wq_completion)events (work_completion)(&(&nsim_dev->trap_data->trap_report_dw)->work) &devlink->lock_key#3 &nsim_trap_data->trap_lock &____s->seqcount irq_context: 0 (wq_completion)events (work_completion)(&(&nsim_dev->trap_data->trap_report_dw)->work) &devlink->lock_key#2 &nsim_trap_data->trap_lock irq_context: 0 (wq_completion)events (work_completion)(&(&nsim_dev->trap_data->trap_report_dw)->work) &devlink->lock_key#2 &nsim_trap_data->trap_lock &c->lock irq_context: 0 (wq_completion)events (work_completion)(&(&nsim_dev->trap_data->trap_report_dw)->work) &devlink->lock_key#2 &nsim_trap_data->trap_lock crngs.lock irq_context: 0 (wq_completion)events (work_completion)(&(&nsim_dev->trap_data->trap_report_dw)->work) &devlink->lock_key#2 &nsim_trap_data->trap_lock &nsim_dev->fa_cookie_lock irq_context: 0 (wq_completion)events (work_completion)(&(&nsim_dev->trap_data->trap_report_dw)->work) &devlink->lock_key#2 &nsim_trap_data->trap_lock &obj_hash[i].lock irq_context: 0 (wq_completion)events (work_completion)(&(&nsim_dev->trap_data->trap_report_dw)->work) &devlink->lock_key#2 &nsim_trap_data->trap_lock &____s->seqcount irq_context: 0 (wq_completion)events (work_completion)(&(&nsim_dev->trap_data->trap_report_dw)->work) &devlink->lock_key#2 &nsim_trap_data->trap_lock &n->list_lock irq_context: 0 (wq_completion)events (work_completion)(&(&nsim_dev->trap_data->trap_report_dw)->work) &devlink->lock_key#2 &nsim_trap_data->trap_lock &n->list_lock &c->lock irq_context: 0 (wq_completion)events (work_completion)(&(&nsim_dev->trap_data->trap_report_dw)->work) &devlink->lock_key#2 &nsim_trap_data->trap_lock &pcp->lock &zone->lock irq_context: 0 (wq_completion)events (work_completion)(&(&nsim_dev->trap_data->trap_report_dw)->work) &devlink->lock_key#2 &nsim_trap_data->trap_lock &pcp->lock &zone->lock &____s->seqcount irq_context: 0 rtnl_mutex _xmit_IPGRE irq_context: 0 (wq_completion)ipv6_addrconf (work_completion)(&(&ifa->dad_work)->work) rtnl_mutex &tb->tb6_lock rcu_read_lock &c->lock irq_context: 0 sb_writers#8 &of->mutex kn->active#52 nsim_bus_dev_list_lock &dev->mutex &devlink->lock_key#4 rcu_read_lock &data->fib_event_queue_lock irq_context: 0 sb_writers#8 &of->mutex kn->active#52 nsim_bus_dev_list_lock &dev->mutex &devlink->lock_key#4 rcu_read_lock rcu_read_lock &pool->lock irq_context: 0 sb_writers#8 &of->mutex kn->active#52 nsim_bus_dev_list_lock &dev->mutex &devlink->lock_key#4 rcu_read_lock rcu_read_lock &pool->lock &obj_hash[i].lock irq_context: 0 sb_writers#8 &of->mutex kn->active#52 nsim_bus_dev_list_lock &dev->mutex &devlink->lock_key#4 rcu_read_lock rcu_read_lock &pool->lock &p->pi_lock irq_context: 0 sb_writers#8 &of->mutex kn->active#52 nsim_bus_dev_list_lock &dev->mutex &devlink->lock_key#4 rcu_read_lock rcu_read_lock &pool->lock &p->pi_lock &rq->__lock irq_context: 0 sb_writers#8 &of->mutex kn->active#52 nsim_bus_dev_list_lock &dev->mutex &devlink->lock_key#4 rcu_read_lock rcu_read_lock &pool->lock &p->pi_lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 sb_writers#8 &of->mutex kn->active#52 nsim_bus_dev_list_lock &dev->mutex &devlink->lock_key#4 rcu_read_lock &c->lock irq_context: 0 sb_writers#8 &of->mutex kn->active#52 nsim_bus_dev_list_lock &dev->mutex &devlink->lock_key#4 rcu_read_lock &n->list_lock irq_context: 0 sb_writers#8 &of->mutex kn->active#52 nsim_bus_dev_list_lock &dev->mutex &devlink->lock_key#4 rcu_read_lock &n->list_lock &c->lock irq_context: 0 sb_writers#8 &of->mutex kn->active#52 nsim_bus_dev_list_lock &dev->mutex &devlink->lock_key#4 rcu_read_lock &tb->tb6_lock irq_context: 0 sb_writers#8 &of->mutex kn->active#52 nsim_bus_dev_list_lock &dev->mutex &devlink->lock_key#4 rcu_read_lock &tb->tb6_lock &net->ipv6.fib6_walker_lock irq_context: 0 sb_writers#8 &of->mutex kn->active#52 nsim_bus_dev_list_lock &dev->mutex &devlink->lock_key#4 rcu_read_lock &tb->tb6_lock &data->fib_event_queue_lock irq_context: 0 sb_writers#8 &of->mutex kn->active#52 nsim_bus_dev_list_lock &dev->mutex &devlink->lock_key#4 rcu_read_lock &obj_hash[i].lock irq_context: 0 sb_writers#8 &of->mutex kn->active#52 nsim_bus_dev_list_lock &dev->mutex &devlink->lock_key#4 &(&fn_net->fib_chain)->lock irq_context: 0 sb_writers#8 &of->mutex kn->active#52 nsim_bus_dev_list_lock &dev->mutex &devlink->lock_key#4 &sb->s_type->i_mutex_key#3 &pcp->lock &zone->lock irq_context: 0 sb_writers#8 &of->mutex kn->active#52 nsim_bus_dev_list_lock &dev->mutex &devlink->lock_key#4 &sb->s_type->i_mutex_key#3 &pcp->lock &zone->lock &____s->seqcount irq_context: 0 rtnl_mutex rcu_read_lock &tb->tb6_lock rt6_exception_lock irq_context: 0 sb_writers#8 &of->mutex kn->active#52 nsim_bus_dev_list_lock &dev->mutex &devlink->lock_key#4 &n->list_lock irq_context: 0 sb_writers#8 &of->mutex kn->active#52 nsim_bus_dev_list_lock &dev->mutex &devlink->lock_key#4 &n->list_lock &c->lock irq_context: 0 sb_writers#8 &of->mutex kn->active#52 nsim_bus_dev_list_lock &dev->mutex &devlink->lock_key#4 &devlink_port->type_lock irq_context: 0 sb_writers#8 &of->mutex kn->active#52 nsim_bus_dev_list_lock &dev->mutex &devlink->lock_key#4 stack_depot_init_mutex irq_context: 0 sb_writers#8 &of->mutex kn->active#52 nsim_bus_dev_list_lock &dev->mutex &devlink->lock_key#4 &sb->s_type->i_mutex_key#3 &rq->__lock irq_context: 0 (wq_completion)ipv6_addrconf (work_completion)(&(&ifa->dad_work)->work) rtnl_mutex &pool->lock &p->pi_lock &rq->__lock irq_context: 0 (wq_completion)ipv6_addrconf (work_completion)(&(&ifa->dad_work)->work) rtnl_mutex &pool->lock &p->pi_lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 rtnl_mutex &ndev->lock &tb->tb6_lock &n->list_lock irq_context: 0 rtnl_mutex &ndev->lock &tb->tb6_lock &n->list_lock &c->lock irq_context: 0 rtnl_mutex _xmit_TUNNEL6 irq_context: 0 rtnl_mutex rcu_read_lock &____s->seqcount irq_context: 0 sb_writers#8 &of->mutex kn->active#52 nsim_bus_dev_list_lock &dev->mutex &devlink->lock_key#4 rtnl_mutex rcu_read_lock &rq->__lock irq_context: 0 sb_writers#8 &of->mutex kn->active#52 nsim_bus_dev_list_lock &dev->mutex &devlink->lock_key#4 rtnl_mutex rcu_read_lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 rtnl_mutex devnet_rename_sem &pcp->lock &zone->lock irq_context: 0 rtnl_mutex devnet_rename_sem &pcp->lock &zone->lock &____s->seqcount irq_context: softirq &(&br->gc_work)->timer irq_context: softirq &(&br->gc_work)->timer rcu_read_lock &pool->lock irq_context: softirq &(&br->gc_work)->timer rcu_read_lock &pool->lock &obj_hash[i].lock irq_context: softirq &(&br->gc_work)->timer rcu_read_lock &pool->lock &p->pi_lock irq_context: 0 (wq_completion)events_long (work_completion)(&(&br->gc_work)->work) irq_context: 0 (wq_completion)events_long (work_completion)(&(&br->gc_work)->work) &obj_hash[i].lock irq_context: 0 (wq_completion)events_long (work_completion)(&(&br->gc_work)->work) &base->lock irq_context: 0 (wq_completion)events_long (work_completion)(&(&br->gc_work)->work) &base->lock &obj_hash[i].lock irq_context: 0 (wq_completion)ipv6_addrconf (work_completion)(&(&ifa->dad_work)->work) rtnl_mutex rcu_read_lock rcu_read_lock rcu_read_lock_bh _xmit_TUNNEL6#2 irq_context: 0 (wq_completion)ipv6_addrconf (work_completion)(&(&ifa->dad_work)->work) rtnl_mutex rcu_read_lock rcu_read_lock rcu_read_lock_bh _xmit_TUNNEL6#2 rcu_read_lock &ndev->lock irq_context: 0 (wq_completion)ipv6_addrconf (work_completion)(&(&ifa->dad_work)->work) rtnl_mutex rcu_read_lock rcu_read_lock rcu_read_lock_bh _xmit_TUNNEL6#2 &obj_hash[i].lock irq_context: 0 (wq_completion)ipv6_addrconf (work_completion)(&(&ifa->dad_work)->work) rtnl_mutex rcu_read_lock rcu_read_lock rcu_read_lock_bh _xmit_TUNNEL6#2 pool_lock#2 irq_context: 0 (wq_completion)events (work_completion)(&data->fib_event_work) &data->fib_lock batched_entropy_u8.lock irq_context: 0 (wq_completion)events (work_completion)(&data->fib_event_work) &data->fib_lock kfence_freelist_lock irq_context: 0 sb_writers#8 &of->mutex kn->active#52 nsim_bus_dev_list_lock &dev->mutex &devlink->lock_key#4 rtnl_mutex bpf_devs_lock irq_context: 0 sb_writers#8 &of->mutex kn->active#52 nsim_bus_dev_list_lock &dev->mutex &devlink->lock_key#4 rtnl_mutex bpf_devs_lock fs_reclaim irq_context: 0 sb_writers#8 &of->mutex kn->active#52 nsim_bus_dev_list_lock &dev->mutex &devlink->lock_key#4 rtnl_mutex bpf_devs_lock fs_reclaim mmu_notifier_invalidate_range_start irq_context: 0 sb_writers#8 &of->mutex kn->active#52 nsim_bus_dev_list_lock &dev->mutex &devlink->lock_key#4 rtnl_mutex bpf_devs_lock rcu_read_lock rhashtable_bucket irq_context: 0 sb_writers#8 &of->mutex kn->active#52 nsim_bus_dev_list_lock &dev->mutex &devlink->lock_key#4 rtnl_mutex bpf_devs_lock rcu_read_lock rcu_read_lock &pool->lock irq_context: 0 sb_writers#8 &of->mutex kn->active#52 nsim_bus_dev_list_lock &dev->mutex &devlink->lock_key#4 rtnl_mutex bpf_devs_lock rcu_read_lock rcu_read_lock &pool->lock &obj_hash[i].lock irq_context: 0 sb_writers#8 &of->mutex kn->active#52 nsim_bus_dev_list_lock &dev->mutex &devlink->lock_key#4 rtnl_mutex bpf_devs_lock rcu_read_lock rcu_read_lock &pool->lock &p->pi_lock irq_context: 0 sb_writers#8 &of->mutex kn->active#52 nsim_bus_dev_list_lock &dev->mutex &devlink->lock_key#4 rtnl_mutex bpf_devs_lock rcu_read_lock rcu_read_lock &pool->lock &p->pi_lock &rq->__lock irq_context: 0 sb_writers#8 &of->mutex kn->active#52 nsim_bus_dev_list_lock &dev->mutex &devlink->lock_key#4 rtnl_mutex bpf_devs_lock rcu_read_lock rcu_read_lock &pool->lock &p->pi_lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 sb_writers#8 &of->mutex kn->active#52 nsim_bus_dev_list_lock &dev->mutex &devlink->lock_key#4 rtnl_mutex bpf_devs_lock rcu_read_lock rcu_node_0 irq_context: 0 sb_writers#8 &of->mutex kn->active#52 nsim_bus_dev_list_lock &dev->mutex &devlink->lock_key#4 rtnl_mutex bpf_devs_lock rcu_read_lock &rq->__lock irq_context: 0 sb_writers#8 &of->mutex kn->active#52 nsim_bus_dev_list_lock &dev->mutex &devlink->lock_key#4 rtnl_mutex pin_fs_lock irq_context: 0 sb_writers#8 &of->mutex kn->active#52 nsim_bus_dev_list_lock &dev->mutex &devlink->lock_key#4 rtnl_mutex &sb->s_type->i_mutex_key#3 irq_context: 0 sb_writers#8 &of->mutex kn->active#52 nsim_bus_dev_list_lock &dev->mutex &devlink->lock_key#4 rtnl_mutex &sb->s_type->i_mutex_key#3 rename_lock.seqcount irq_context: 0 sb_writers#8 &of->mutex kn->active#52 nsim_bus_dev_list_lock &dev->mutex &devlink->lock_key#4 rtnl_mutex &sb->s_type->i_mutex_key#3 fs_reclaim irq_context: 0 sb_writers#8 &of->mutex kn->active#52 nsim_bus_dev_list_lock &dev->mutex &devlink->lock_key#4 rtnl_mutex &sb->s_type->i_mutex_key#3 fs_reclaim mmu_notifier_invalidate_range_start irq_context: 0 sb_writers#8 &of->mutex kn->active#52 nsim_bus_dev_list_lock &dev->mutex &devlink->lock_key#4 rtnl_mutex &sb->s_type->i_mutex_key#3 &dentry->d_lock irq_context: 0 sb_writers#8 &of->mutex kn->active#52 nsim_bus_dev_list_lock &dev->mutex &devlink->lock_key#4 rtnl_mutex &sb->s_type->i_mutex_key#3 rcu_read_lock rename_lock.seqcount irq_context: 0 sb_writers#8 &of->mutex kn->active#52 nsim_bus_dev_list_lock &dev->mutex &devlink->lock_key#4 rtnl_mutex &sb->s_type->i_mutex_key#3 &dentry->d_lock &wq irq_context: 0 sb_writers#8 &of->mutex kn->active#52 nsim_bus_dev_list_lock &dev->mutex &devlink->lock_key#4 rtnl_mutex &sb->s_type->i_mutex_key#3 mmu_notifier_invalidate_range_start irq_context: 0 sb_writers#8 &of->mutex kn->active#52 nsim_bus_dev_list_lock &dev->mutex &devlink->lock_key#4 rtnl_mutex &sb->s_type->i_mutex_key#3 &sb->s_type->i_lock_key#7 irq_context: 0 sb_writers#8 &of->mutex kn->active#52 nsim_bus_dev_list_lock &dev->mutex &devlink->lock_key#4 rtnl_mutex &sb->s_type->i_mutex_key#3 &s->s_inode_list_lock irq_context: 0 sb_writers#8 &of->mutex kn->active#52 nsim_bus_dev_list_lock &dev->mutex &devlink->lock_key#4 rtnl_mutex &sb->s_type->i_mutex_key#3 tk_core.seq.seqcount irq_context: 0 sb_writers#8 &of->mutex kn->active#52 nsim_bus_dev_list_lock &dev->mutex &devlink->lock_key#4 rtnl_mutex &sb->s_type->i_mutex_key#3 &sb->s_type->i_lock_key#7 &dentry->d_lock irq_context: 0 sb_writers#8 &of->mutex kn->active#52 nsim_bus_dev_list_lock &dev->mutex &devlink->lock_key#4 rtnl_mutex &sb->s_type->i_mutex_key#3 &c->lock irq_context: 0 sb_writers#8 &of->mutex kn->active#52 nsim_bus_dev_list_lock &dev->mutex &devlink->lock_key#4 rtnl_mutex &____s->seqcount irq_context: 0 sb_writers#8 &of->mutex kn->active#52 nsim_bus_dev_list_lock &dev->mutex &devlink->lock_key#4 rtnl_mutex &obj_hash[i].lock irq_context: 0 sb_writers#8 &of->mutex kn->active#52 nsim_bus_dev_list_lock &dev->mutex &devlink->lock_key#4 rtnl_mutex fs_reclaim irq_context: 0 sb_writers#8 &of->mutex kn->active#52 nsim_bus_dev_list_lock &dev->mutex &devlink->lock_key#4 rtnl_mutex fs_reclaim mmu_notifier_invalidate_range_start irq_context: 0 sb_writers#8 &of->mutex kn->active#52 nsim_bus_dev_list_lock &dev->mutex &devlink->lock_key#4 rtnl_mutex &xa->xa_lock#3 irq_context: 0 sb_writers#8 &of->mutex kn->active#52 nsim_bus_dev_list_lock &dev->mutex &devlink->lock_key#4 rtnl_mutex &base->lock irq_context: 0 sb_writers#8 &of->mutex kn->active#52 nsim_bus_dev_list_lock &dev->mutex &devlink->lock_key#4 rtnl_mutex &base->lock &obj_hash[i].lock irq_context: 0 sb_writers#8 &of->mutex kn->active#52 nsim_bus_dev_list_lock &dev->mutex &devlink->lock_key#4 rtnl_mutex (work_completion)(&(&devlink_port->type_warn_dw)->work) irq_context: 0 sb_writers#8 &of->mutex kn->active#52 nsim_bus_dev_list_lock &dev->mutex &devlink->lock_key#4 rtnl_mutex &devlink_port->type_lock irq_context: 0 sb_writers#8 &of->mutex kn->active#52 nsim_bus_dev_list_lock &dev->mutex &devlink->lock_key#4 rtnl_mutex &c->lock irq_context: 0 sb_writers#8 &of->mutex kn->active#52 nsim_bus_dev_list_lock &dev->mutex &devlink->lock_key#4 rtnl_mutex nl_table_lock irq_context: 0 sb_writers#8 &of->mutex kn->active#52 nsim_bus_dev_list_lock &dev->mutex &devlink->lock_key#4 rtnl_mutex nl_table_wait.lock irq_context: 0 sb_writers#8 &of->mutex kn->active#52 nsim_bus_dev_list_lock &dev->mutex &devlink->lock_key#4 rtnl_mutex net_rwsem irq_context: 0 sb_writers#8 &of->mutex kn->active#52 nsim_bus_dev_list_lock &dev->mutex &devlink->lock_key#4 rtnl_mutex net_rwsem &list->lock#2 irq_context: 0 sb_writers#8 &of->mutex kn->active#52 nsim_bus_dev_list_lock &dev->mutex &devlink->lock_key#4 rtnl_mutex &tn->lock irq_context: 0 sb_writers#8 &of->mutex kn->active#52 nsim_bus_dev_list_lock &dev->mutex &devlink->lock_key#4 rtnl_mutex &x->wait#9 irq_context: 0 sb_writers#8 &of->mutex kn->active#52 nsim_bus_dev_list_lock &dev->mutex &devlink->lock_key#4 rtnl_mutex &k->list_lock irq_context: 0 sb_writers#8 &of->mutex kn->active#52 nsim_bus_dev_list_lock &dev->mutex &devlink->lock_key#4 rtnl_mutex gdp_mutex irq_context: 0 sb_writers#8 &of->mutex kn->active#52 nsim_bus_dev_list_lock &dev->mutex &devlink->lock_key#4 rtnl_mutex gdp_mutex &k->list_lock irq_context: 0 sb_writers#8 &of->mutex kn->active#52 nsim_bus_dev_list_lock &dev->mutex &devlink->lock_key#4 rtnl_mutex gdp_mutex fs_reclaim irq_context: 0 sb_writers#8 &of->mutex kn->active#52 nsim_bus_dev_list_lock &dev->mutex &devlink->lock_key#4 rtnl_mutex gdp_mutex fs_reclaim mmu_notifier_invalidate_range_start irq_context: 0 sb_writers#8 &of->mutex kn->active#52 nsim_bus_dev_list_lock &dev->mutex &devlink->lock_key#4 rtnl_mutex gdp_mutex lock irq_context: 0 sb_writers#8 &of->mutex kn->active#52 nsim_bus_dev_list_lock &dev->mutex &devlink->lock_key#4 rtnl_mutex gdp_mutex lock kernfs_idr_lock irq_context: 0 sb_writers#8 &of->mutex kn->active#52 nsim_bus_dev_list_lock &dev->mutex &devlink->lock_key#4 rtnl_mutex gdp_mutex &root->kernfs_rwsem irq_context: 0 sb_writers#8 &of->mutex kn->active#52 nsim_bus_dev_list_lock &dev->mutex &devlink->lock_key#4 rtnl_mutex gdp_mutex &root->kernfs_rwsem &root->kernfs_iattr_rwsem irq_context: 0 sb_writers#8 &of->mutex kn->active#52 nsim_bus_dev_list_lock &dev->mutex &devlink->lock_key#4 rtnl_mutex gdp_mutex kobj_ns_type_lock irq_context: 0 sb_writers#8 &of->mutex kn->active#52 nsim_bus_dev_list_lock &dev->mutex &devlink->lock_key#4 rtnl_mutex lock irq_context: 0 sb_writers#8 &of->mutex kn->active#52 nsim_bus_dev_list_lock &dev->mutex &devlink->lock_key#4 rtnl_mutex lock kernfs_idr_lock irq_context: 0 sb_writers#8 &of->mutex kn->active#52 nsim_bus_dev_list_lock &dev->mutex &devlink->lock_key#4 rtnl_mutex &root->kernfs_rwsem irq_context: 0 sb_writers#8 &of->mutex kn->active#52 nsim_bus_dev_list_lock &dev->mutex &devlink->lock_key#4 rtnl_mutex &root->kernfs_rwsem &root->kernfs_iattr_rwsem irq_context: 0 sb_writers#8 &of->mutex kn->active#52 nsim_bus_dev_list_lock &dev->mutex &devlink->lock_key#4 rtnl_mutex &root->kernfs_rwsem &root->kernfs_iattr_rwsem &rq->__lock irq_context: 0 sb_writers#8 &of->mutex kn->active#52 nsim_bus_dev_list_lock &dev->mutex &devlink->lock_key#4 rtnl_mutex bus_type_sem irq_context: 0 sb_writers#8 &of->mutex kn->active#52 nsim_bus_dev_list_lock &dev->mutex &devlink->lock_key#4 rtnl_mutex sysfs_symlink_target_lock irq_context: 0 sb_writers#8 &of->mutex kn->active#52 nsim_bus_dev_list_lock &dev->mutex &devlink->lock_key#4 rtnl_mutex &root->kernfs_rwsem irq_context: 0 sb_writers#8 &of->mutex kn->active#52 nsim_bus_dev_list_lock &dev->mutex &devlink->lock_key#4 rtnl_mutex &dev->power.lock irq_context: 0 sb_writers#8 &of->mutex kn->active#52 nsim_bus_dev_list_lock &dev->mutex &devlink->lock_key#4 rtnl_mutex dpm_list_mtx irq_context: 0 sb_writers#8 &of->mutex kn->active#52 nsim_bus_dev_list_lock &dev->mutex &devlink->lock_key#4 rtnl_mutex uevent_sock_mutex irq_context: 0 sb_writers#8 &of->mutex kn->active#52 nsim_bus_dev_list_lock &dev->mutex &devlink->lock_key#4 rtnl_mutex uevent_sock_mutex fs_reclaim irq_context: 0 sb_writers#8 &of->mutex kn->active#52 nsim_bus_dev_list_lock &dev->mutex &devlink->lock_key#4 rtnl_mutex uevent_sock_mutex fs_reclaim mmu_notifier_invalidate_range_start irq_context: 0 sb_writers#8 &of->mutex kn->active#52 nsim_bus_dev_list_lock &dev->mutex &devlink->lock_key#4 rtnl_mutex uevent_sock_mutex &c->lock irq_context: 0 sb_writers#8 &of->mutex kn->active#52 nsim_bus_dev_list_lock &dev->mutex &devlink->lock_key#4 rtnl_mutex uevent_sock_mutex nl_table_lock irq_context: 0 sb_writers#8 &of->mutex kn->active#52 nsim_bus_dev_list_lock &dev->mutex &devlink->lock_key#4 rtnl_mutex uevent_sock_mutex &obj_hash[i].lock irq_context: 0 sb_writers#8 &of->mutex kn->active#52 nsim_bus_dev_list_lock &dev->mutex &devlink->lock_key#4 rtnl_mutex uevent_sock_mutex nl_table_wait.lock irq_context: 0 sb_writers#8 &of->mutex kn->active#52 nsim_bus_dev_list_lock &dev->mutex &devlink->lock_key#4 rtnl_mutex &k->k_lock irq_context: 0 sb_writers#8 &of->mutex kn->active#52 nsim_bus_dev_list_lock &dev->mutex &devlink->lock_key#4 rtnl_mutex subsys mutex#17 irq_context: 0 sb_writers#8 &of->mutex kn->active#52 nsim_bus_dev_list_lock &dev->mutex &devlink->lock_key#4 rtnl_mutex subsys mutex#17 &k->k_lock irq_context: 0 sb_writers#8 &of->mutex kn->active#52 nsim_bus_dev_list_lock &dev->mutex &devlink->lock_key#4 rtnl_mutex &dir->lock#2 irq_context: 0 sb_writers#8 &of->mutex kn->active#52 nsim_bus_dev_list_lock &dev->mutex &devlink->lock_key#4 rtnl_mutex dev_hotplug_mutex irq_context: 0 sb_writers#8 &of->mutex kn->active#52 nsim_bus_dev_list_lock &dev->mutex &devlink->lock_key#4 rtnl_mutex dev_hotplug_mutex &dev->power.lock irq_context: 0 sb_writers#8 &of->mutex kn->active#52 nsim_bus_dev_list_lock &dev->mutex &devlink->lock_key#4 rtnl_mutex dev_base_lock irq_context: 0 sb_writers#8 &of->mutex kn->active#52 nsim_bus_dev_list_lock &dev->mutex &devlink->lock_key#4 rtnl_mutex input_pool.lock irq_context: 0 sb_writers#8 &of->mutex kn->active#52 nsim_bus_dev_list_lock &dev->mutex &devlink->lock_key#4 rtnl_mutex &pcp->lock &zone->lock irq_context: 0 sb_writers#8 &of->mutex kn->active#52 nsim_bus_dev_list_lock &dev->mutex &devlink->lock_key#4 rtnl_mutex &pcp->lock &zone->lock &____s->seqcount irq_context: 0 sb_writers#8 &of->mutex kn->active#52 nsim_bus_dev_list_lock &dev->mutex &devlink->lock_key#4 rtnl_mutex rcu_read_lock &pool->lock/1 irq_context: 0 sb_writers#8 &of->mutex kn->active#52 nsim_bus_dev_list_lock &dev->mutex &devlink->lock_key#4 rtnl_mutex rcu_read_lock &pool->lock/1 &obj_hash[i].lock irq_context: 0 sb_writers#8 &of->mutex kn->active#52 nsim_bus_dev_list_lock &dev->mutex &devlink->lock_key#4 rtnl_mutex rcu_read_lock &pool->lock/1 &p->pi_lock irq_context: 0 sb_writers#8 &of->mutex kn->active#52 nsim_bus_dev_list_lock &dev->mutex &devlink->lock_key#4 rtnl_mutex rcu_read_lock &pool->lock/1 &p->pi_lock &rq->__lock irq_context: 0 sb_writers#8 &of->mutex kn->active#52 nsim_bus_dev_list_lock &dev->mutex &devlink->lock_key#4 rtnl_mutex rcu_read_lock &pool->lock/1 &p->pi_lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 sb_writers#8 &of->mutex kn->active#52 nsim_bus_dev_list_lock &dev->mutex &devlink->lock_key#4 rtnl_mutex batched_entropy_u32.lock irq_context: 0 sb_writers#8 &of->mutex kn->active#52 nsim_bus_dev_list_lock &dev->mutex &devlink->lock_key#4 rtnl_mutex &tbl->lock irq_context: 0 sb_writers#8 &of->mutex kn->active#52 nsim_bus_dev_list_lock &dev->mutex &devlink->lock_key#4 rtnl_mutex sysctl_lock irq_context: 0 sb_writers#8 &of->mutex kn->active#52 nsim_bus_dev_list_lock &dev->mutex &devlink->lock_key#4 rtnl_mutex failover_lock irq_context: 0 sb_writers#8 &of->mutex kn->active#52 nsim_bus_dev_list_lock &dev->mutex &devlink->lock_key#4 rtnl_mutex pcpu_alloc_mutex irq_context: 0 sb_writers#8 &of->mutex kn->active#52 nsim_bus_dev_list_lock &dev->mutex &devlink->lock_key#4 rtnl_mutex pcpu_alloc_mutex pcpu_lock irq_context: 0 sb_writers#8 &of->mutex kn->active#52 nsim_bus_dev_list_lock &dev->mutex &devlink->lock_key#4 rtnl_mutex proc_subdir_lock irq_context: 0 sb_writers#8 &of->mutex kn->active#52 nsim_bus_dev_list_lock &dev->mutex &devlink->lock_key#4 rtnl_mutex proc_inum_ida.xa_lock irq_context: 0 sb_writers#8 &of->mutex kn->active#52 nsim_bus_dev_list_lock &dev->mutex &devlink->lock_key#4 rtnl_mutex proc_subdir_lock irq_context: 0 sb_writers#8 &of->mutex kn->active#52 nsim_bus_dev_list_lock &dev->mutex &devlink->lock_key#4 rtnl_mutex &obj_hash[i].lock pool_lock irq_context: 0 sb_writers#8 &of->mutex kn->active#52 nsim_bus_dev_list_lock &dev->mutex &devlink->lock_key#4 rtnl_mutex fill_pool_map-wait-type-override pool_lock#2 irq_context: 0 sb_writers#8 &of->mutex kn->active#52 nsim_bus_dev_list_lock &dev->mutex &devlink->lock_key#4 rtnl_mutex fill_pool_map-wait-type-override pool_lock irq_context: 0 sb_writers#8 &of->mutex kn->active#52 nsim_bus_dev_list_lock &dev->mutex &devlink->lock_key#4 rtnl_mutex &n->list_lock irq_context: 0 sb_writers#8 &of->mutex kn->active#52 nsim_bus_dev_list_lock &dev->mutex &devlink->lock_key#4 rtnl_mutex &n->list_lock &c->lock irq_context: 0 sb_writers#8 &of->mutex kn->active#52 nsim_bus_dev_list_lock &dev->mutex &devlink->lock_key#4 rtnl_mutex &idev->mc_lock irq_context: 0 sb_writers#8 &of->mutex kn->active#52 nsim_bus_dev_list_lock &dev->mutex &devlink->lock_key#4 rtnl_mutex &idev->mc_lock fs_reclaim irq_context: 0 sb_writers#8 &of->mutex kn->active#52 nsim_bus_dev_list_lock &dev->mutex &devlink->lock_key#4 rtnl_mutex &idev->mc_lock fs_reclaim mmu_notifier_invalidate_range_start irq_context: 0 sb_writers#8 &of->mutex kn->active#52 nsim_bus_dev_list_lock &dev->mutex &devlink->lock_key#4 rtnl_mutex &idev->mc_lock &obj_hash[i].lock irq_context: 0 sb_writers#8 &of->mutex kn->active#52 nsim_bus_dev_list_lock &dev->mutex &devlink->lock_key#4 rtnl_mutex &idev->mc_lock _xmit_ETHER irq_context: 0 sb_writers#8 &of->mutex kn->active#52 nsim_bus_dev_list_lock &dev->mutex &devlink->lock_key#4 rtnl_mutex &pnettable->lock irq_context: 0 sb_writers#8 &of->mutex kn->active#52 nsim_bus_dev_list_lock &dev->mutex &devlink->lock_key#4 rtnl_mutex smc_ib_devices.mutex irq_context: 0 sb_writers#8 &of->mutex kn->active#52 nsim_bus_dev_list_lock &dev->mutex &devlink->lock_key#4 rtnl_mutex &vn->sock_lock irq_context: 0 sb_writers#8 &of->mutex kn->active#52 nsim_bus_dev_list_lock &dev->mutex &devlink->lock_key#4 rtnl_mutex rcu_read_lock &ndev->lock irq_context: 0 sb_writers#8 &of->mutex kn->active#52 nsim_bus_dev_list_lock &dev->mutex &devlink->lock_key#4 &rq->__lock irq_context: 0 sb_writers#8 &of->mutex kn->active#52 nsim_bus_dev_list_lock &dev->mutex &devlink->lock_key#4 &obj_hash[i].lock pool_lock irq_context: 0 sb_writers#8 &of->mutex kn->active#52 nsim_bus_dev_list_lock &dev->mutex &devlink->lock_key#4 fill_pool_map-wait-type-override pool_lock#2 irq_context: 0 sb_writers#8 &of->mutex kn->active#52 nsim_bus_dev_list_lock &dev->mutex &devlink->lock_key#4 fill_pool_map-wait-type-override &c->lock irq_context: 0 sb_writers#8 &of->mutex kn->active#52 nsim_bus_dev_list_lock &dev->mutex &devlink->lock_key#4 fill_pool_map-wait-type-override &____s->seqcount irq_context: 0 sb_writers#8 &of->mutex kn->active#52 nsim_bus_dev_list_lock &dev->mutex &devlink->lock_key#4 fill_pool_map-wait-type-override pool_lock irq_context: 0 (wq_completion)mld (work_completion)(&(&idev->mc_ifc_work)->work) &idev->mc_lock rcu_read_lock rcu_read_lock rcu_read_lock_bh dev->qdisc_tx_busylock ?: &qdisc_tx_busylock rcu_read_lock &ndev->lock irq_context: 0 (wq_completion)mld (work_completion)(&(&idev->mc_ifc_work)->work) &idev->mc_lock rcu_read_lock rcu_read_lock rcu_read_lock_bh dev->qdisc_tx_busylock ?: &qdisc_tx_busylock &obj_hash[i].lock irq_context: 0 (wq_completion)mld (work_completion)(&(&idev->mc_ifc_work)->work) &idev->mc_lock rcu_read_lock rcu_read_lock rcu_read_lock_bh dev->qdisc_tx_busylock ?: &qdisc_tx_busylock pool_lock#2 irq_context: 0 (wq_completion)mld (work_completion)(&(&idev->mc_ifc_work)->work) &idev->mc_lock batched_entropy_u32.lock crngs.lock irq_context: softirq (&app->periodic_timer) irq_context: softirq (&app->periodic_timer) &app->lock irq_context: softirq (&app->periodic_timer) &app->lock &obj_hash[i].lock irq_context: softirq (&app->periodic_timer) &app->lock &base->lock irq_context: softirq (&app->periodic_timer) &app->lock &base->lock &obj_hash[i].lock irq_context: 0 sb_writers#8 &of->mutex kn->active#52 nsim_bus_dev_list_lock &dev->mutex &devlink->lock_key#4 rtnl_mutex pool_lock#2 irq_context: 0 rtnl_mutex dev_addr_sem &rq->__lock &cfs_rq->removed.lock irq_context: 0 (wq_completion)events (work_completion)(&(&nsim_dev->trap_data->trap_report_dw)->work) &devlink->lock_key#3 &nsim_trap_data->trap_lock &n->list_lock irq_context: 0 (wq_completion)events (work_completion)(&(&nsim_dev->trap_data->trap_report_dw)->work) &devlink->lock_key#3 &nsim_trap_data->trap_lock &n->list_lock &c->lock irq_context: 0 rtnl_mutex (inetaddr_chain).rwsem &obj_hash[i].lock pool_lock irq_context: 0 rtnl_mutex (inetaddr_chain).rwsem fill_pool_map-wait-type-override pool_lock#2 irq_context: 0 rtnl_mutex (inetaddr_chain).rwsem fill_pool_map-wait-type-override pool_lock irq_context: softirq (&in_dev->mr_ifc_timer) rcu_read_lock &____s->seqcount irq_context: softirq (&in_dev->mr_ifc_timer) rcu_read_lock &c->lock irq_context: softirq (&in_dev->mr_ifc_timer) rcu_read_lock &nf_conntrack_locks[i]/1 irq_context: 0 sb_writers#8 &of->mutex kn->active#52 nsim_bus_dev_list_lock &dev->mutex &devlink->lock_key#4 rtnl_mutex &sb->s_type->i_mutex_key#3 &____s->seqcount irq_context: 0 sb_writers#8 &of->mutex kn->active#52 nsim_bus_dev_list_lock &dev->mutex &devlink->lock_key#4 rtnl_mutex &sb->s_type->i_mutex_key#3 pool_lock#2 irq_context: 0 (wq_completion)ipv6_addrconf (work_completion)(&(&ifa->dad_work)->work) rtnl_mutex rcu_read_lock rcu_read_lock rcu_read_lock_bh dev->qdisc_tx_busylock ?: &qdisc_tx_busylock rcu_read_lock &ndev->lock irq_context: 0 (wq_completion)ipv6_addrconf (work_completion)(&(&ifa->dad_work)->work) rtnl_mutex rcu_read_lock rcu_read_lock rcu_read_lock_bh dev->qdisc_tx_busylock ?: &qdisc_tx_busylock &obj_hash[i].lock irq_context: 0 (wq_completion)ipv6_addrconf (work_completion)(&(&ifa->dad_work)->work) rtnl_mutex rcu_read_lock rcu_read_lock rcu_read_lock_bh dev->qdisc_tx_busylock ?: &qdisc_tx_busylock pool_lock#2 irq_context: 0 rtnl_mutex dev_addr_sem &pcp->lock &zone->lock irq_context: 0 rtnl_mutex dev_addr_sem &pcp->lock &zone->lock &____s->seqcount irq_context: 0 rtnl_mutex dev_addr_sem rcu_read_lock &bond->stats_lock/1 irq_context: 0 rtnl_mutex &dev_addr_list_lock_key/1 irq_context: 0 rtnl_mutex &dev_addr_list_lock_key/1 rcu_read_lock _xmit_ETHER irq_context: 0 (wq_completion)ipv6_addrconf (work_completion)(&(&ifa->dad_work)->work) rtnl_mutex &idev->mc_lock &dev_addr_list_lock_key/1 irq_context: 0 (wq_completion)ipv6_addrconf (work_completion)(&(&ifa->dad_work)->work) rtnl_mutex &idev->mc_lock &dev_addr_list_lock_key/1 rcu_read_lock _xmit_ETHER irq_context: 0 (wq_completion)ipv6_addrconf (work_completion)(&(&ifa->dad_work)->work) rtnl_mutex &idev->mc_lock &dev_addr_list_lock_key/1 rcu_read_lock _xmit_ETHER pool_lock#2 irq_context: 0 (wq_completion)ipv6_addrconf (work_completion)(&(&ifa->dad_work)->work) rtnl_mutex &idev->mc_lock &dev_addr_list_lock_key/1 rcu_read_lock _xmit_ETHER &c->lock irq_context: 0 (wq_completion)ipv6_addrconf (work_completion)(&(&ifa->dad_work)->work) rtnl_mutex &idev->mc_lock &dev_addr_list_lock_key/1 rcu_read_lock _xmit_ETHER &____s->seqcount irq_context: 0 sb_writers#8 &of->mutex kn->active#52 nsim_bus_dev_list_lock &dev->mutex &devlink->lock_key#4 rtnl_mutex batched_entropy_u8.lock irq_context: 0 sb_writers#8 &of->mutex kn->active#52 nsim_bus_dev_list_lock &dev->mutex &devlink->lock_key#4 rtnl_mutex kfence_freelist_lock irq_context: 0 sb_writers#8 &of->mutex kn->active#52 nsim_bus_dev_list_lock &dev->mutex &devlink->lock_key#4 rtnl_mutex &idev->mc_lock &obj_hash[i].lock pool_lock irq_context: 0 (wq_completion)ipv6_addrconf (work_completion)(&(&ifa->dad_work)->work) rtnl_mutex &tb->tb6_lock rcu_read_lock &n->list_lock irq_context: 0 (wq_completion)ipv6_addrconf (work_completion)(&(&ifa->dad_work)->work) rtnl_mutex &tb->tb6_lock rcu_read_lock &n->list_lock &c->lock irq_context: 0 (wq_completion)ipv6_addrconf (work_completion)(&(&ifa->dad_work)->work) rtnl_mutex &idev->mc_lock &bridge_netdev_addr_lock_key/1 &c->lock irq_context: 0 (wq_completion)ipv6_addrconf (work_completion)(&(&ifa->dad_work)->work) rtnl_mutex &idev->mc_lock &bridge_netdev_addr_lock_key/1 &____s->seqcount irq_context: 0 cb_lock nlk_cb_mutex-GENERIC &devlink->lock_key#4 irq_context: 0 cb_lock nlk_cb_mutex-GENERIC &devlink->lock_key#4 &devlink_port->type_lock irq_context: 0 sb_writers#8 &of->mutex kn->active#52 nsim_bus_dev_list_lock &dev->mutex &devlink->lock_key#5 irq_context: 0 sb_writers#8 &of->mutex kn->active#52 nsim_bus_dev_list_lock &dev->mutex &devlink->lock_key#5 crngs.lock irq_context: 0 sb_writers#8 &of->mutex kn->active#52 nsim_bus_dev_list_lock &dev->mutex &devlink->lock_key#5 fs_reclaim irq_context: 0 sb_writers#8 &of->mutex kn->active#52 nsim_bus_dev_list_lock &dev->mutex &devlink->lock_key#5 fs_reclaim mmu_notifier_invalidate_range_start irq_context: 0 sb_writers#8 &of->mutex kn->active#52 nsim_bus_dev_list_lock &dev->mutex &devlink->lock_key#5 devlinks.xa_lock irq_context: 0 sb_writers#8 &of->mutex kn->active#52 nsim_bus_dev_list_lock &dev->mutex &devlink->lock_key#5 &c->lock irq_context: 0 sb_writers#8 &of->mutex kn->active#52 nsim_bus_dev_list_lock &dev->mutex &devlink->lock_key#5 &pcp->lock &zone->lock irq_context: 0 sb_writers#8 &of->mutex kn->active#52 nsim_bus_dev_list_lock &dev->mutex &devlink->lock_key#5 &pcp->lock &zone->lock &____s->seqcount irq_context: 0 sb_writers#8 &of->mutex kn->active#52 nsim_bus_dev_list_lock &dev->mutex &devlink->lock_key#5 &____s->seqcount irq_context: 0 sb_writers#8 &of->mutex kn->active#52 nsim_bus_dev_list_lock &dev->mutex &devlink->lock_key#5 pool_lock#2 irq_context: 0 sb_writers#8 &of->mutex kn->active#52 nsim_bus_dev_list_lock &dev->mutex &devlink->lock_key#5 &obj_hash[i].lock irq_context: 0 sb_writers#8 &of->mutex kn->active#52 nsim_bus_dev_list_lock &dev->mutex &devlink->lock_key#5 nl_table_lock irq_context: 0 sb_writers#8 &of->mutex kn->active#52 nsim_bus_dev_list_lock &dev->mutex &devlink->lock_key#5 nl_table_wait.lock irq_context: 0 sb_writers#8 &of->mutex kn->active#52 nsim_bus_dev_list_lock &dev->mutex &devlink->lock_key#5 &xa->xa_lock#16 irq_context: 0 sb_writers#8 &of->mutex kn->active#52 nsim_bus_dev_list_lock &dev->mutex &devlink->lock_key#5 pcpu_alloc_mutex irq_context: 0 sb_writers#8 &of->mutex kn->active#52 nsim_bus_dev_list_lock &dev->mutex &devlink->lock_key#5 pcpu_alloc_mutex pcpu_lock irq_context: 0 sb_writers#8 &of->mutex kn->active#52 nsim_bus_dev_list_lock &dev->mutex &devlink->lock_key#5 &base->lock irq_context: 0 sb_writers#8 &of->mutex kn->active#52 nsim_bus_dev_list_lock &dev->mutex &devlink->lock_key#5 &base->lock &obj_hash[i].lock irq_context: 0 sb_writers#8 &of->mutex kn->active#52 nsim_bus_dev_list_lock &dev->mutex &devlink->lock_key#5 pin_fs_lock irq_context: 0 sb_writers#8 &of->mutex kn->active#52 nsim_bus_dev_list_lock &dev->mutex &devlink->lock_key#5 &sb->s_type->i_mutex_key#3 irq_context: 0 sb_writers#8 &of->mutex kn->active#52 nsim_bus_dev_list_lock &dev->mutex &devlink->lock_key#5 &sb->s_type->i_mutex_key#3 rename_lock.seqcount irq_context: 0 sb_writers#8 &of->mutex kn->active#52 nsim_bus_dev_list_lock &dev->mutex &devlink->lock_key#5 &sb->s_type->i_mutex_key#3 fs_reclaim irq_context: 0 sb_writers#8 &of->mutex kn->active#52 nsim_bus_dev_list_lock &dev->mutex &devlink->lock_key#5 &sb->s_type->i_mutex_key#3 fs_reclaim mmu_notifier_invalidate_range_start irq_context: 0 sb_writers#8 &of->mutex kn->active#52 nsim_bus_dev_list_lock &dev->mutex &devlink->lock_key#5 &sb->s_type->i_mutex_key#3 &dentry->d_lock irq_context: 0 sb_writers#8 &of->mutex kn->active#52 nsim_bus_dev_list_lock &dev->mutex &devlink->lock_key#5 &sb->s_type->i_mutex_key#3 rcu_read_lock rename_lock.seqcount irq_context: 0 sb_writers#8 &of->mutex kn->active#52 nsim_bus_dev_list_lock &dev->mutex &devlink->lock_key#5 &sb->s_type->i_mutex_key#3 &dentry->d_lock &wq irq_context: 0 sb_writers#8 &of->mutex kn->active#52 nsim_bus_dev_list_lock &dev->mutex &devlink->lock_key#5 &sb->s_type->i_mutex_key#3 mmu_notifier_invalidate_range_start irq_context: 0 sb_writers#8 &of->mutex kn->active#52 nsim_bus_dev_list_lock &dev->mutex &devlink->lock_key#5 &sb->s_type->i_mutex_key#3 &sb->s_type->i_lock_key#7 irq_context: 0 sb_writers#8 &of->mutex kn->active#52 nsim_bus_dev_list_lock &dev->mutex &devlink->lock_key#5 &sb->s_type->i_mutex_key#3 &s->s_inode_list_lock irq_context: 0 sb_writers#8 &of->mutex kn->active#52 nsim_bus_dev_list_lock &dev->mutex &devlink->lock_key#5 &sb->s_type->i_mutex_key#3 tk_core.seq.seqcount irq_context: 0 sb_writers#8 &of->mutex kn->active#52 nsim_bus_dev_list_lock &dev->mutex &devlink->lock_key#5 &sb->s_type->i_mutex_key#3 &sb->s_type->i_lock_key#7 &dentry->d_lock irq_context: 0 sb_writers#8 &of->mutex kn->active#52 nsim_bus_dev_list_lock &dev->mutex &devlink->lock_key#5 &sb->s_type->i_mutex_key#3 &c->lock irq_context: 0 sb_writers#8 &of->mutex kn->active#52 nsim_bus_dev_list_lock &dev->mutex &devlink->lock_key#5 &sb->s_type->i_mutex_key#3 &____s->seqcount irq_context: 0 sb_writers#8 &of->mutex kn->active#52 nsim_bus_dev_list_lock &dev->mutex &devlink->lock_key#5 &sb->s_type->i_mutex_key#3 pool_lock#2 irq_context: 0 sb_writers#8 &of->mutex kn->active#52 nsim_bus_dev_list_lock &dev->mutex &devlink->lock_key#5 &sb->s_type->i_mutex_key#3 &n->list_lock irq_context: 0 sb_writers#8 &of->mutex kn->active#52 nsim_bus_dev_list_lock &dev->mutex &devlink->lock_key#5 &sb->s_type->i_mutex_key#3 &n->list_lock &c->lock irq_context: 0 sb_writers#8 &of->mutex kn->active#52 nsim_bus_dev_list_lock &dev->mutex &devlink->lock_key#5 batched_entropy_u32.lock irq_context: 0 sb_writers#8 &of->mutex kn->active#52 nsim_bus_dev_list_lock &dev->mutex &devlink->lock_key#5 rtnl_mutex irq_context: 0 sb_writers#8 &of->mutex kn->active#52 nsim_bus_dev_list_lock &dev->mutex &devlink->lock_key#5 rtnl_mutex &rq->__lock irq_context: 0 sb_writers#8 &of->mutex kn->active#52 nsim_bus_dev_list_lock &dev->mutex &devlink->lock_key#5 rtnl_mutex &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)mld (work_completion)(&(&idev->mc_ifc_work)->work) &idev->mc_lock rcu_read_lock rcu_read_lock rcu_read_lock_bh dev->qdisc_tx_busylock ?: &qdisc_tx_busylock _xmit_ETHER#2 rcu_read_lock &ndev->lock irq_context: 0 (wq_completion)mld (work_completion)(&(&idev->mc_ifc_work)->work) &idev->mc_lock rcu_read_lock rcu_read_lock rcu_read_lock_bh dev->qdisc_tx_busylock ?: &qdisc_tx_busylock _xmit_ETHER#2 &obj_hash[i].lock irq_context: 0 (wq_completion)mld (work_completion)(&(&idev->mc_ifc_work)->work) &idev->mc_lock rcu_read_lock rcu_read_lock rcu_read_lock_bh dev->qdisc_tx_busylock ?: &qdisc_tx_busylock _xmit_ETHER#2 pool_lock#2 irq_context: 0 sb_writers#8 &of->mutex kn->active#52 nsim_bus_dev_list_lock &dev->mutex &devlink->lock_key#5 rtnl_mutex rtnl_mutex.wait_lock irq_context: 0 sb_writers#8 &of->mutex kn->active#52 nsim_bus_dev_list_lock &dev->mutex &devlink->lock_key#5 rtnl_mutex &(&net->nexthop.notifier_chain)->rwsem irq_context: 0 sb_writers#8 &of->mutex kn->active#52 nsim_bus_dev_list_lock &dev->mutex &devlink->lock_key#5 rtnl_mutex.wait_lock irq_context: 0 sb_writers#8 &of->mutex kn->active#52 nsim_bus_dev_list_lock &dev->mutex &devlink->lock_key#5 &p->pi_lock irq_context: 0 sb_writers#8 &of->mutex kn->active#52 nsim_bus_dev_list_lock &dev->mutex &devlink->lock_key#5 rcu_read_lock &data->fib_event_queue_lock irq_context: 0 sb_writers#8 &of->mutex kn->active#52 nsim_bus_dev_list_lock &dev->mutex &devlink->lock_key#5 rcu_read_lock rcu_read_lock &pool->lock irq_context: 0 sb_writers#8 &of->mutex kn->active#52 nsim_bus_dev_list_lock &dev->mutex &devlink->lock_key#5 rcu_read_lock rcu_read_lock &pool->lock &obj_hash[i].lock irq_context: 0 sb_writers#8 &of->mutex kn->active#52 nsim_bus_dev_list_lock &dev->mutex &devlink->lock_key#5 rcu_read_lock rcu_read_lock &pool->lock &p->pi_lock irq_context: 0 sb_writers#8 &of->mutex kn->active#52 nsim_bus_dev_list_lock &dev->mutex &devlink->lock_key#5 rcu_read_lock rcu_read_lock &pool->lock &p->pi_lock &rq->__lock irq_context: 0 sb_writers#8 &of->mutex kn->active#52 nsim_bus_dev_list_lock &dev->mutex &devlink->lock_key#5 rcu_read_lock rcu_read_lock &pool->lock &p->pi_lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 sb_writers#8 &of->mutex kn->active#52 nsim_bus_dev_list_lock &dev->mutex &devlink->lock_key#5 rcu_read_lock &tb->tb6_lock irq_context: 0 sb_writers#8 &of->mutex kn->active#52 nsim_bus_dev_list_lock &dev->mutex &devlink->lock_key#5 rcu_read_lock &tb->tb6_lock &net->ipv6.fib6_walker_lock irq_context: 0 sb_writers#8 &of->mutex kn->active#52 nsim_bus_dev_list_lock &dev->mutex &devlink->lock_key#5 rcu_read_lock &tb->tb6_lock &c->lock irq_context: 0 sb_writers#8 &of->mutex kn->active#52 nsim_bus_dev_list_lock &dev->mutex &devlink->lock_key#5 rcu_read_lock &tb->tb6_lock &data->fib_event_queue_lock irq_context: 0 sb_writers#8 &of->mutex kn->active#52 nsim_bus_dev_list_lock &dev->mutex &devlink->lock_key#5 rcu_read_lock &obj_hash[i].lock irq_context: 0 sb_writers#8 &of->mutex kn->active#52 nsim_bus_dev_list_lock &dev->mutex &devlink->lock_key#5 &(&fn_net->fib_chain)->lock irq_context: 0 (wq_completion)mld (work_completion)(&(&idev->mc_ifc_work)->work) &idev->mc_lock rcu_read_lock rcu_read_lock &____s->seqcount irq_context: 0 (wq_completion)mld (work_completion)(&(&idev->mc_ifc_work)->work) &idev->mc_lock rcu_read_lock rcu_read_lock rcu_read_lock_bh rcu_read_lock pcpu_lock irq_context: 0 (wq_completion)mld (work_completion)(&(&idev->mc_ifc_work)->work) &idev->mc_lock rcu_read_lock rcu_read_lock rcu_read_lock_bh rcu_read_lock &obj_hash[i].lock irq_context: 0 (wq_completion)mld (work_completion)(&(&idev->mc_ifc_work)->work) &idev->mc_lock rcu_read_lock rcu_read_lock rcu_read_lock_bh rcu_read_lock pool_lock#2 irq_context: 0 rtnl_mutex devnet_rename_sem uevent_sock_mutex &n->list_lock irq_context: 0 rtnl_mutex devnet_rename_sem uevent_sock_mutex &n->list_lock &c->lock irq_context: 0 (wq_completion)events (linkwatch_work).work rtnl_mutex &idev->mc_lock irq_context: 0 (wq_completion)events (linkwatch_work).work rtnl_mutex &dir->lock#2 irq_context: 0 (wq_completion)events (linkwatch_work).work rtnl_mutex pcpu_alloc_mutex irq_context: 0 (wq_completion)events (linkwatch_work).work rtnl_mutex pcpu_alloc_mutex pcpu_lock irq_context: 0 (wq_completion)events (linkwatch_work).work rtnl_mutex &tb->tb6_lock irq_context: 0 (wq_completion)events (linkwatch_work).work rtnl_mutex &tb->tb6_lock pool_lock#2 irq_context: 0 (wq_completion)events (linkwatch_work).work rtnl_mutex &tb->tb6_lock &c->lock irq_context: 0 (wq_completion)events (linkwatch_work).work rtnl_mutex &tb->tb6_lock &____s->seqcount irq_context: 0 (wq_completion)events (linkwatch_work).work rtnl_mutex &tb->tb6_lock nl_table_lock irq_context: 0 (wq_completion)events (linkwatch_work).work rtnl_mutex &tb->tb6_lock &obj_hash[i].lock irq_context: 0 (wq_completion)events (linkwatch_work).work rtnl_mutex &tb->tb6_lock nl_table_wait.lock irq_context: 0 (wq_completion)events (linkwatch_work).work rtnl_mutex (inet6addr_validator_chain).rwsem irq_context: 0 (wq_completion)events (linkwatch_work).work rtnl_mutex rcu_read_lock &net->ipv6.addrconf_hash_lock irq_context: 0 (wq_completion)events (linkwatch_work).work rtnl_mutex rcu_read_lock &ndev->lock irq_context: 0 (wq_completion)events (linkwatch_work).work rtnl_mutex rcu_read_lock pool_lock#2 irq_context: 0 (wq_completion)events (linkwatch_work).work rtnl_mutex rcu_read_lock &obj_hash[i].lock irq_context: 0 (wq_completion)events (linkwatch_work).work rtnl_mutex rcu_read_lock rcu_read_lock &pool->lock/1 irq_context: 0 (wq_completion)events (linkwatch_work).work rtnl_mutex rcu_read_lock rcu_read_lock &pool->lock/1 &obj_hash[i].lock irq_context: 0 (wq_completion)events (linkwatch_work).work rtnl_mutex rcu_read_lock rcu_read_lock &pool->lock/1 pool_lock#2 irq_context: 0 (wq_completion)events (linkwatch_work).work rtnl_mutex rcu_read_lock rcu_read_lock &pool->lock/1 &p->pi_lock irq_context: 0 (wq_completion)events (linkwatch_work).work rtnl_mutex rcu_read_lock rcu_read_lock &pool->lock/1 &p->pi_lock &rq->__lock irq_context: 0 (wq_completion)events (linkwatch_work).work rtnl_mutex rcu_read_lock rcu_read_lock &pool->lock/1 &p->pi_lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)events (linkwatch_work).work rtnl_mutex rcu_read_lock rcu_node_0 irq_context: 0 (wq_completion)events (linkwatch_work).work rtnl_mutex rcu_read_lock &rq->__lock irq_context: 0 (wq_completion)events (linkwatch_work).work rtnl_mutex rcu_read_lock &net->sctp.local_addr_lock irq_context: 0 (wq_completion)events (linkwatch_work).work rtnl_mutex rcu_read_lock &net->sctp.local_addr_lock &net->sctp.addr_wq_lock irq_context: 0 (wq_completion)events (linkwatch_work).work rtnl_mutex rcu_read_lock &net->sctp.local_addr_lock &net->sctp.addr_wq_lock pool_lock#2 irq_context: 0 (wq_completion)events (linkwatch_work).work rtnl_mutex pcpu_lock irq_context: 0 (wq_completion)events (linkwatch_work).work rtnl_mutex &ifa->lock irq_context: 0 (wq_completion)events (linkwatch_work).work rtnl_mutex rcu_read_lock &pool->lock irq_context: 0 (wq_completion)events (linkwatch_work).work rtnl_mutex rcu_read_lock &pool->lock &obj_hash[i].lock irq_context: 0 (wq_completion)events (linkwatch_work).work rtnl_mutex rcu_read_lock &pool->lock pool_lock#2 irq_context: 0 (wq_completion)events (linkwatch_work).work rtnl_mutex &ndev->lock irq_context: 0 (wq_completion)events (linkwatch_work).work rtnl_mutex &ndev->lock &ifa->lock irq_context: 0 (wq_completion)events (linkwatch_work).work rtnl_mutex &idev->mc_lock &obj_hash[i].lock irq_context: 0 (wq_completion)events (linkwatch_work).work rtnl_mutex &idev->mc_lock &base->lock irq_context: 0 (wq_completion)events (linkwatch_work).work rtnl_mutex &idev->mc_lock &base->lock &obj_hash[i].lock irq_context: 0 (wq_completion)events (linkwatch_work).work rtnl_mutex &ndev->lock &ifa->lock batched_entropy_u32.lock irq_context: 0 (wq_completion)events (linkwatch_work).work rtnl_mutex &ndev->lock &ifa->lock crngs.lock irq_context: 0 (wq_completion)events (linkwatch_work).work rtnl_mutex &ndev->lock &ifa->lock &obj_hash[i].lock irq_context: 0 (wq_completion)events (linkwatch_work).work rtnl_mutex &ndev->lock &ifa->lock &base->lock irq_context: 0 (wq_completion)events (linkwatch_work).work rtnl_mutex &ndev->lock &ifa->lock &base->lock &obj_hash[i].lock irq_context: 0 sb_writers#8 &of->mutex kn->active#52 nsim_bus_dev_list_lock &dev->mutex &devlink->lock_key#5 &p->pi_lock &rq->__lock irq_context: 0 sb_writers#8 &of->mutex kn->active#52 nsim_bus_dev_list_lock &dev->mutex &devlink->lock_key#5 &p->pi_lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 sb_writers#8 &of->mutex kn->active#52 nsim_bus_dev_list_lock &dev->mutex &devlink->lock_key#5 &sb->s_type->i_mutex_key#3 &rq->__lock irq_context: 0 sb_writers#8 &of->mutex kn->active#52 nsim_bus_dev_list_lock &dev->mutex &devlink->lock_key#5 &sb->s_type->i_mutex_key#3 &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 sb_writers#8 &of->mutex kn->active#52 nsim_bus_dev_list_lock &dev->mutex &devlink->lock_key#5 &devlink_port->type_lock irq_context: 0 sb_writers#8 &of->mutex kn->active#52 nsim_bus_dev_list_lock &dev->mutex &devlink->lock_key#5 stack_depot_init_mutex irq_context: 0 sb_writers#8 &of->mutex kn->active#52 nsim_bus_dev_list_lock &dev->mutex &devlink->lock_key#5 rtnl_mutex bpf_devs_lock irq_context: 0 sb_writers#8 &of->mutex kn->active#52 nsim_bus_dev_list_lock &dev->mutex &devlink->lock_key#5 rtnl_mutex bpf_devs_lock fs_reclaim irq_context: 0 sb_writers#8 &of->mutex kn->active#52 nsim_bus_dev_list_lock &dev->mutex &devlink->lock_key#5 rtnl_mutex bpf_devs_lock fs_reclaim mmu_notifier_invalidate_range_start irq_context: 0 sb_writers#8 &of->mutex kn->active#52 nsim_bus_dev_list_lock &dev->mutex &devlink->lock_key#5 rtnl_mutex bpf_devs_lock rcu_read_lock rhashtable_bucket irq_context: 0 sb_writers#8 &of->mutex kn->active#52 nsim_bus_dev_list_lock &dev->mutex &devlink->lock_key#5 rtnl_mutex pin_fs_lock irq_context: 0 sb_writers#8 &of->mutex kn->active#52 nsim_bus_dev_list_lock &dev->mutex &devlink->lock_key#5 rtnl_mutex &sb->s_type->i_mutex_key#3 irq_context: 0 sb_writers#8 &of->mutex kn->active#52 nsim_bus_dev_list_lock &dev->mutex &devlink->lock_key#5 rtnl_mutex &sb->s_type->i_mutex_key#3 rename_lock.seqcount irq_context: 0 sb_writers#8 &of->mutex kn->active#52 nsim_bus_dev_list_lock &dev->mutex &devlink->lock_key#5 rtnl_mutex &sb->s_type->i_mutex_key#3 fs_reclaim irq_context: 0 sb_writers#8 &of->mutex kn->active#52 nsim_bus_dev_list_lock &dev->mutex &devlink->lock_key#5 rtnl_mutex &sb->s_type->i_mutex_key#3 fs_reclaim mmu_notifier_invalidate_range_start irq_context: 0 sb_writers#8 &of->mutex kn->active#52 nsim_bus_dev_list_lock &dev->mutex &devlink->lock_key#5 rtnl_mutex &sb->s_type->i_mutex_key#3 &dentry->d_lock irq_context: 0 sb_writers#8 &of->mutex kn->active#52 nsim_bus_dev_list_lock &dev->mutex &devlink->lock_key#5 rtnl_mutex &sb->s_type->i_mutex_key#3 &rq->__lock irq_context: 0 (wq_completion)events (work_completion)(&data->fib_event_work) &data->fib_lock &n->list_lock irq_context: 0 (wq_completion)events (work_completion)(&data->fib_event_work) &data->fib_lock &n->list_lock &c->lock irq_context: 0 (wq_completion)events (work_completion)(&(&nsim_dev->trap_data->trap_report_dw)->work) &devlink->lock_key &nsim_trap_data->trap_lock quarantine_lock irq_context: 0 (wq_completion)events (work_completion)(&(&nsim_dev->trap_data->trap_report_dw)->work) &devlink->lock_key#2 &nsim_trap_data->trap_lock quarantine_lock irq_context: 0 sb_writers#8 &of->mutex kn->active#52 nsim_bus_dev_list_lock &dev->mutex &devlink->lock_key#5 rtnl_mutex &sb->s_type->i_mutex_key#3 rcu_read_lock rename_lock.seqcount irq_context: 0 sb_writers#8 &of->mutex kn->active#52 nsim_bus_dev_list_lock &dev->mutex &devlink->lock_key#5 rtnl_mutex &sb->s_type->i_mutex_key#3 &dentry->d_lock &wq irq_context: 0 sb_writers#8 &of->mutex kn->active#52 nsim_bus_dev_list_lock &dev->mutex &devlink->lock_key#5 rtnl_mutex &sb->s_type->i_mutex_key#3 remove_cache_srcu irq_context: 0 sb_writers#8 &of->mutex kn->active#52 nsim_bus_dev_list_lock &dev->mutex &devlink->lock_key#5 rtnl_mutex &sb->s_type->i_mutex_key#3 remove_cache_srcu quarantine_lock irq_context: 0 sb_writers#8 &of->mutex kn->active#52 nsim_bus_dev_list_lock &dev->mutex &devlink->lock_key#5 rtnl_mutex &sb->s_type->i_mutex_key#3 remove_cache_srcu &c->lock irq_context: 0 sb_writers#8 &of->mutex kn->active#52 nsim_bus_dev_list_lock &dev->mutex &devlink->lock_key#5 rtnl_mutex &sb->s_type->i_mutex_key#3 remove_cache_srcu &n->list_lock irq_context: 0 sb_writers#8 &of->mutex kn->active#52 nsim_bus_dev_list_lock &dev->mutex &devlink->lock_key#5 rtnl_mutex &sb->s_type->i_mutex_key#3 remove_cache_srcu rcu_read_lock pool_lock#2 irq_context: 0 sb_writers#8 &of->mutex kn->active#52 nsim_bus_dev_list_lock &dev->mutex &devlink->lock_key#5 rtnl_mutex &sb->s_type->i_mutex_key#3 remove_cache_srcu &obj_hash[i].lock irq_context: 0 sb_writers#8 &of->mutex kn->active#52 nsim_bus_dev_list_lock &dev->mutex &devlink->lock_key#5 rtnl_mutex &sb->s_type->i_mutex_key#3 mmu_notifier_invalidate_range_start irq_context: 0 sb_writers#8 &of->mutex kn->active#52 nsim_bus_dev_list_lock &dev->mutex &devlink->lock_key#5 rtnl_mutex &sb->s_type->i_mutex_key#3 &sb->s_type->i_lock_key#7 irq_context: 0 sb_writers#8 &of->mutex kn->active#52 nsim_bus_dev_list_lock &dev->mutex &devlink->lock_key#5 rtnl_mutex &sb->s_type->i_mutex_key#3 &s->s_inode_list_lock irq_context: 0 sb_writers#8 &of->mutex kn->active#52 nsim_bus_dev_list_lock &dev->mutex &devlink->lock_key#5 rtnl_mutex &sb->s_type->i_mutex_key#3 tk_core.seq.seqcount irq_context: 0 sb_writers#8 &of->mutex kn->active#52 nsim_bus_dev_list_lock &dev->mutex &devlink->lock_key#5 rtnl_mutex &sb->s_type->i_mutex_key#3 &sb->s_type->i_lock_key#7 &dentry->d_lock irq_context: 0 sb_writers#8 &of->mutex kn->active#52 nsim_bus_dev_list_lock &dev->mutex &devlink->lock_key#5 rtnl_mutex &sb->s_type->i_mutex_key#3 &c->lock irq_context: 0 sb_writers#8 &of->mutex kn->active#52 nsim_bus_dev_list_lock &dev->mutex &devlink->lock_key#5 rtnl_mutex &____s->seqcount irq_context: 0 sb_writers#8 &of->mutex kn->active#52 nsim_bus_dev_list_lock &dev->mutex &devlink->lock_key#5 rtnl_mutex &obj_hash[i].lock irq_context: 0 sb_writers#8 &of->mutex kn->active#52 nsim_bus_dev_list_lock &dev->mutex &devlink->lock_key#5 rtnl_mutex fs_reclaim irq_context: 0 sb_writers#8 &of->mutex kn->active#52 nsim_bus_dev_list_lock &dev->mutex &devlink->lock_key#5 rtnl_mutex fs_reclaim mmu_notifier_invalidate_range_start irq_context: 0 sb_writers#8 &of->mutex kn->active#52 nsim_bus_dev_list_lock &dev->mutex &devlink->lock_key#5 rtnl_mutex &xa->xa_lock#3 irq_context: 0 sb_writers#8 &of->mutex kn->active#52 nsim_bus_dev_list_lock &dev->mutex &devlink->lock_key#5 rtnl_mutex &base->lock irq_context: 0 sb_writers#8 &of->mutex kn->active#52 nsim_bus_dev_list_lock &dev->mutex &devlink->lock_key#5 rtnl_mutex &base->lock &obj_hash[i].lock irq_context: 0 sb_writers#8 &of->mutex kn->active#52 nsim_bus_dev_list_lock &dev->mutex &devlink->lock_key#5 rtnl_mutex (work_completion)(&(&devlink_port->type_warn_dw)->work) irq_context: 0 sb_writers#8 &of->mutex kn->active#52 nsim_bus_dev_list_lock &dev->mutex &devlink->lock_key#5 rtnl_mutex &devlink_port->type_lock irq_context: 0 sb_writers#8 &of->mutex kn->active#52 nsim_bus_dev_list_lock &dev->mutex &devlink->lock_key#5 rtnl_mutex nl_table_lock irq_context: 0 sb_writers#8 &of->mutex kn->active#52 nsim_bus_dev_list_lock &dev->mutex &devlink->lock_key#5 rtnl_mutex nl_table_wait.lock irq_context: 0 sb_writers#8 &of->mutex kn->active#52 nsim_bus_dev_list_lock &dev->mutex &devlink->lock_key#5 rtnl_mutex net_rwsem irq_context: 0 sb_writers#8 &of->mutex kn->active#52 nsim_bus_dev_list_lock &dev->mutex &devlink->lock_key#5 rtnl_mutex net_rwsem &list->lock#2 irq_context: 0 sb_writers#8 &of->mutex kn->active#52 nsim_bus_dev_list_lock &dev->mutex &devlink->lock_key#5 rtnl_mutex &tn->lock irq_context: 0 sb_writers#8 &of->mutex kn->active#52 nsim_bus_dev_list_lock &dev->mutex &devlink->lock_key#5 rtnl_mutex &x->wait#9 irq_context: 0 sb_writers#8 &of->mutex kn->active#52 nsim_bus_dev_list_lock &dev->mutex &devlink->lock_key#5 rtnl_mutex &k->list_lock irq_context: 0 sb_writers#8 &of->mutex kn->active#52 nsim_bus_dev_list_lock &dev->mutex &devlink->lock_key#5 rtnl_mutex gdp_mutex irq_context: 0 sb_writers#8 &of->mutex kn->active#52 nsim_bus_dev_list_lock &dev->mutex &devlink->lock_key#5 rtnl_mutex gdp_mutex &k->list_lock irq_context: 0 sb_writers#8 &of->mutex kn->active#52 nsim_bus_dev_list_lock &dev->mutex &devlink->lock_key#5 rtnl_mutex gdp_mutex fs_reclaim irq_context: 0 sb_writers#8 &of->mutex kn->active#52 nsim_bus_dev_list_lock &dev->mutex &devlink->lock_key#5 rtnl_mutex gdp_mutex fs_reclaim mmu_notifier_invalidate_range_start irq_context: 0 sb_writers#8 &of->mutex kn->active#52 nsim_bus_dev_list_lock &dev->mutex &devlink->lock_key#5 rtnl_mutex gdp_mutex lock irq_context: 0 sb_writers#8 &of->mutex kn->active#52 nsim_bus_dev_list_lock &dev->mutex &devlink->lock_key#5 rtnl_mutex gdp_mutex lock kernfs_idr_lock irq_context: 0 sb_writers#8 &of->mutex kn->active#52 nsim_bus_dev_list_lock &dev->mutex &devlink->lock_key#5 rtnl_mutex gdp_mutex &root->kernfs_rwsem irq_context: 0 sb_writers#8 &of->mutex kn->active#52 nsim_bus_dev_list_lock &dev->mutex &devlink->lock_key#5 rtnl_mutex gdp_mutex &root->kernfs_rwsem &root->kernfs_iattr_rwsem irq_context: 0 sb_writers#8 &of->mutex kn->active#52 nsim_bus_dev_list_lock &dev->mutex &devlink->lock_key#5 rtnl_mutex gdp_mutex kobj_ns_type_lock irq_context: 0 sb_writers#8 &of->mutex kn->active#52 nsim_bus_dev_list_lock &dev->mutex &devlink->lock_key#5 rtnl_mutex lock irq_context: 0 sb_writers#8 &of->mutex kn->active#52 nsim_bus_dev_list_lock &dev->mutex &devlink->lock_key#5 rtnl_mutex lock kernfs_idr_lock irq_context: 0 sb_writers#8 &of->mutex kn->active#52 nsim_bus_dev_list_lock &dev->mutex &devlink->lock_key#5 rtnl_mutex &root->kernfs_rwsem irq_context: 0 sb_writers#8 &of->mutex kn->active#52 nsim_bus_dev_list_lock &dev->mutex &devlink->lock_key#5 rtnl_mutex &root->kernfs_rwsem &root->kernfs_iattr_rwsem irq_context: 0 sb_writers#8 &of->mutex kn->active#52 nsim_bus_dev_list_lock &dev->mutex &devlink->lock_key#5 rtnl_mutex bus_type_sem irq_context: 0 sb_writers#8 &of->mutex kn->active#52 nsim_bus_dev_list_lock &dev->mutex &devlink->lock_key#5 rtnl_mutex sysfs_symlink_target_lock irq_context: 0 sb_writers#8 &of->mutex kn->active#52 nsim_bus_dev_list_lock &dev->mutex &devlink->lock_key#5 rtnl_mutex &c->lock irq_context: 0 sb_writers#8 &of->mutex kn->active#52 nsim_bus_dev_list_lock &dev->mutex &devlink->lock_key#5 rtnl_mutex &root->kernfs_rwsem irq_context: 0 sb_writers#8 &of->mutex kn->active#52 nsim_bus_dev_list_lock &dev->mutex &devlink->lock_key#5 rtnl_mutex &dev->power.lock irq_context: 0 sb_writers#8 &of->mutex kn->active#52 nsim_bus_dev_list_lock &dev->mutex &devlink->lock_key#5 rtnl_mutex dpm_list_mtx irq_context: 0 sb_writers#8 &of->mutex kn->active#52 nsim_bus_dev_list_lock &dev->mutex &devlink->lock_key#5 rtnl_mutex uevent_sock_mutex irq_context: 0 sb_writers#8 &of->mutex kn->active#52 nsim_bus_dev_list_lock &dev->mutex &devlink->lock_key#5 rtnl_mutex uevent_sock_mutex fs_reclaim irq_context: 0 sb_writers#8 &of->mutex kn->active#52 nsim_bus_dev_list_lock &dev->mutex &devlink->lock_key#5 rtnl_mutex uevent_sock_mutex fs_reclaim mmu_notifier_invalidate_range_start irq_context: 0 sb_writers#8 &of->mutex kn->active#52 nsim_bus_dev_list_lock &dev->mutex &devlink->lock_key#5 rtnl_mutex uevent_sock_mutex nl_table_lock irq_context: 0 sb_writers#8 &of->mutex kn->active#52 nsim_bus_dev_list_lock &dev->mutex &devlink->lock_key#5 rtnl_mutex uevent_sock_mutex &obj_hash[i].lock irq_context: 0 sb_writers#8 &of->mutex kn->active#52 nsim_bus_dev_list_lock &dev->mutex &devlink->lock_key#5 rtnl_mutex uevent_sock_mutex nl_table_wait.lock irq_context: 0 sb_writers#8 &of->mutex kn->active#52 nsim_bus_dev_list_lock &dev->mutex &devlink->lock_key#5 rtnl_mutex &k->k_lock irq_context: 0 sb_writers#8 &of->mutex kn->active#52 nsim_bus_dev_list_lock &dev->mutex &devlink->lock_key#5 rtnl_mutex subsys mutex#17 irq_context: 0 sb_writers#8 &of->mutex kn->active#52 nsim_bus_dev_list_lock &dev->mutex &devlink->lock_key#5 rtnl_mutex subsys mutex#17 &k->k_lock irq_context: 0 sb_writers#8 &of->mutex kn->active#52 nsim_bus_dev_list_lock &dev->mutex &devlink->lock_key#5 rtnl_mutex &dir->lock#2 irq_context: 0 sb_writers#8 &of->mutex kn->active#52 nsim_bus_dev_list_lock &dev->mutex &devlink->lock_key#5 rtnl_mutex dev_hotplug_mutex irq_context: 0 sb_writers#8 &of->mutex kn->active#52 nsim_bus_dev_list_lock &dev->mutex &devlink->lock_key#5 rtnl_mutex dev_hotplug_mutex &dev->power.lock irq_context: 0 sb_writers#8 &of->mutex kn->active#52 nsim_bus_dev_list_lock &dev->mutex &devlink->lock_key#5 rtnl_mutex dev_base_lock irq_context: 0 sb_writers#8 &of->mutex kn->active#52 nsim_bus_dev_list_lock &dev->mutex &devlink->lock_key#5 rtnl_mutex pool_lock#2 irq_context: 0 sb_writers#8 &of->mutex kn->active#52 nsim_bus_dev_list_lock &dev->mutex &devlink->lock_key#5 rtnl_mutex input_pool.lock irq_context: 0 sb_writers#8 &of->mutex kn->active#52 nsim_bus_dev_list_lock &dev->mutex &devlink->lock_key#5 rtnl_mutex rcu_read_lock &pool->lock/1 irq_context: 0 sb_writers#8 &of->mutex kn->active#52 nsim_bus_dev_list_lock &dev->mutex &devlink->lock_key#5 rtnl_mutex rcu_read_lock &pool->lock/1 &obj_hash[i].lock irq_context: 0 sb_writers#8 &of->mutex kn->active#52 nsim_bus_dev_list_lock &dev->mutex &devlink->lock_key#5 rtnl_mutex rcu_read_lock &pool->lock/1 &p->pi_lock irq_context: 0 sb_writers#8 &of->mutex kn->active#52 nsim_bus_dev_list_lock &dev->mutex &devlink->lock_key#5 rtnl_mutex rcu_read_lock &pool->lock/1 &p->pi_lock &rq->__lock irq_context: 0 sb_writers#8 &of->mutex kn->active#52 nsim_bus_dev_list_lock &dev->mutex &devlink->lock_key#5 rtnl_mutex rcu_read_lock &pool->lock/1 &p->pi_lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 sb_writers#8 &of->mutex kn->active#52 nsim_bus_dev_list_lock &dev->mutex &devlink->lock_key#5 rtnl_mutex batched_entropy_u32.lock irq_context: 0 sb_writers#8 &of->mutex kn->active#52 nsim_bus_dev_list_lock &dev->mutex &devlink->lock_key#5 rtnl_mutex &tbl->lock irq_context: 0 sb_writers#8 &of->mutex kn->active#52 nsim_bus_dev_list_lock &dev->mutex &devlink->lock_key#5 rtnl_mutex sysctl_lock irq_context: 0 sb_writers#8 &of->mutex kn->active#52 nsim_bus_dev_list_lock &dev->mutex &devlink->lock_key#5 rtnl_mutex failover_lock irq_context: 0 sb_writers#8 &of->mutex kn->active#52 nsim_bus_dev_list_lock &dev->mutex &devlink->lock_key#5 rtnl_mutex pcpu_alloc_mutex irq_context: 0 sb_writers#8 &of->mutex kn->active#52 nsim_bus_dev_list_lock &dev->mutex &devlink->lock_key#5 rtnl_mutex pcpu_alloc_mutex pcpu_lock irq_context: 0 sb_writers#8 &of->mutex kn->active#52 nsim_bus_dev_list_lock &dev->mutex &devlink->lock_key#5 rtnl_mutex pcpu_alloc_mutex &rq->__lock irq_context: 0 sb_writers#8 &of->mutex kn->active#52 nsim_bus_dev_list_lock &dev->mutex &devlink->lock_key#5 rtnl_mutex proc_subdir_lock irq_context: 0 sb_writers#8 &of->mutex kn->active#52 nsim_bus_dev_list_lock &dev->mutex &devlink->lock_key#5 rtnl_mutex proc_inum_ida.xa_lock irq_context: 0 sb_writers#8 &of->mutex kn->active#52 nsim_bus_dev_list_lock &dev->mutex &devlink->lock_key#5 rtnl_mutex proc_subdir_lock irq_context: 0 sb_writers#8 &of->mutex kn->active#52 nsim_bus_dev_list_lock &dev->mutex &devlink->lock_key#5 rtnl_mutex &idev->mc_lock irq_context: 0 sb_writers#8 &of->mutex kn->active#52 nsim_bus_dev_list_lock &dev->mutex &devlink->lock_key#5 rtnl_mutex &idev->mc_lock fs_reclaim irq_context: 0 sb_writers#8 &of->mutex kn->active#52 nsim_bus_dev_list_lock &dev->mutex &devlink->lock_key#5 rtnl_mutex &idev->mc_lock fs_reclaim mmu_notifier_invalidate_range_start irq_context: 0 sb_writers#8 &of->mutex kn->active#52 nsim_bus_dev_list_lock &dev->mutex &devlink->lock_key#5 rtnl_mutex &idev->mc_lock &obj_hash[i].lock irq_context: 0 sb_writers#8 &of->mutex kn->active#52 nsim_bus_dev_list_lock &dev->mutex &devlink->lock_key#5 rtnl_mutex &idev->mc_lock _xmit_ETHER irq_context: 0 sb_writers#8 &of->mutex kn->active#52 nsim_bus_dev_list_lock &dev->mutex &devlink->lock_key#5 rtnl_mutex &pnettable->lock irq_context: 0 sb_writers#8 &of->mutex kn->active#52 nsim_bus_dev_list_lock &dev->mutex &devlink->lock_key#5 rtnl_mutex smc_ib_devices.mutex irq_context: 0 sb_writers#8 &of->mutex kn->active#52 nsim_bus_dev_list_lock &dev->mutex &devlink->lock_key#5 rtnl_mutex &vn->sock_lock irq_context: 0 sb_writers#8 &of->mutex kn->active#52 nsim_bus_dev_list_lock &dev->mutex &devlink->lock_key#5 rtnl_mutex rcu_read_lock &ndev->lock irq_context: 0 sb_writers#8 &of->mutex kn->active#52 nsim_bus_dev_list_lock &dev->mutex &devlink->lock_key#5 &rq->__lock irq_context: 0 rtnl_mutex devnet_rename_sem &root->kernfs_rwsem &rq->__lock irq_context: 0 rtnl_mutex devnet_rename_sem &root->kernfs_rwsem &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 sb_writers#8 &of->mutex kn->active#52 nsim_bus_dev_list_lock &dev->mutex &devlink->lock_key#5 &sb->s_type->i_mutex_key#3 rcu_node_0 irq_context: 0 sb_writers#8 &of->mutex kn->active#52 nsim_bus_dev_list_lock &dev->mutex &devlink->lock_key#5 &sb->s_type->i_mutex_key#3 &rcu_state.expedited_wq irq_context: 0 sb_writers#8 &of->mutex kn->active#52 nsim_bus_dev_list_lock &dev->mutex &devlink->lock_key#5 &sb->s_type->i_mutex_key#3 &rcu_state.expedited_wq &p->pi_lock irq_context: 0 sb_writers#8 &of->mutex kn->active#52 nsim_bus_dev_list_lock &dev->mutex &devlink->lock_key#5 &sb->s_type->i_mutex_key#3 &pcp->lock &zone->lock irq_context: 0 sb_writers#8 &of->mutex kn->active#52 nsim_bus_dev_list_lock &dev->mutex &devlink->lock_key#5 &sb->s_type->i_mutex_key#3 &pcp->lock &zone->lock &____s->seqcount irq_context: 0 (wq_completion)mld (work_completion)(&(&idev->mc_ifc_work)->work) &idev->mc_lock rcu_read_lock rcu_read_lock rcu_read_lock_bh rcu_read_lock tk_core.seq.seqcount irq_context: 0 (wq_completion)mld (work_completion)(&(&idev->mc_ifc_work)->work) &idev->mc_lock rcu_read_lock rcu_read_lock rcu_read_lock_bh rcu_read_lock &list->lock#5 irq_context: 0 sb_writers#8 &of->mutex kn->active#52 nsim_bus_dev_list_lock &dev->mutex &devlink->lock_key#5 rtnl_mutex rcu_read_lock &rq->__lock irq_context: 0 sb_writers#8 &of->mutex kn->active#52 nsim_bus_dev_list_lock &dev->mutex &devlink->lock_key#5 rtnl_mutex rcu_read_lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)ipv6_addrconf (work_completion)(&(&ifa->dad_work)->work) rtnl_mutex rcu_read_lock rcu_read_lock rcu_read_lock_bh dev->qdisc_tx_busylock ?: &qdisc_tx_busylock _xmit_ETHER#2 rcu_read_lock &ndev->lock irq_context: 0 (wq_completion)ipv6_addrconf (work_completion)(&(&ifa->dad_work)->work) rtnl_mutex rcu_read_lock rcu_read_lock rcu_read_lock_bh dev->qdisc_tx_busylock ?: &qdisc_tx_busylock _xmit_ETHER#2 &obj_hash[i].lock irq_context: 0 (wq_completion)ipv6_addrconf (work_completion)(&(&ifa->dad_work)->work) rtnl_mutex rcu_read_lock rcu_read_lock rcu_read_lock_bh dev->qdisc_tx_busylock ?: &qdisc_tx_busylock _xmit_ETHER#2 pool_lock#2 irq_context: 0 (wq_completion)bat_events (work_completion)(&(&forw_packet_aggr->delayed_work)->work) &hard_iface->bat_iv.ogm_buff_mutex batched_entropy_u8.lock crngs.lock irq_context: 0 sb_writers#8 &of->mutex kn->active#52 nsim_bus_dev_list_lock &dev->mutex &devlink->lock_key#5 rtnl_mutex &pcp->lock &zone->lock irq_context: 0 sb_writers#8 &of->mutex kn->active#52 nsim_bus_dev_list_lock &dev->mutex &devlink->lock_key#5 rtnl_mutex &pcp->lock &zone->lock &____s->seqcount irq_context: 0 sb_writers#8 &of->mutex kn->active#52 nsim_bus_dev_list_lock &dev->mutex &devlink->lock_key#5 rtnl_mutex uevent_sock_mutex &c->lock irq_context: 0 cb_lock genl_mutex rtnl_mutex rcu_read_lock &rq->__lock irq_context: 0 cb_lock genl_mutex rtnl_mutex rcu_read_lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 rtnl_mutex rcu_read_lock fill_pool_map-wait-type-override pool_lock#2 irq_context: 0 rtnl_mutex rcu_read_lock fill_pool_map-wait-type-override pool_lock irq_context: 0 rtnl_mutex rcu_read_lock &net->sctp.local_addr_lock &net->sctp.addr_wq_lock &____s->seqcount irq_context: 0 (wq_completion)ipv6_addrconf (work_completion)(&(&ifa->dad_work)->work) &rq->__lock irq_context: 0 sb_writers#8 &of->mutex kn->active#52 nsim_bus_dev_list_lock &dev->mutex &devlink->lock_key#5 rtnl_mutex &sb->s_type->i_mutex_key#3 &____s->seqcount irq_context: 0 sb_writers#8 &of->mutex kn->active#52 nsim_bus_dev_list_lock &dev->mutex &devlink->lock_key#5 rtnl_mutex &sb->s_type->i_mutex_key#3 pool_lock#2 irq_context: 0 sb_writers#8 &of->mutex kn->active#52 nsim_bus_dev_list_lock &dev->mutex &devlink->lock_key#5 rtnl_mutex &n->list_lock irq_context: 0 sb_writers#8 &of->mutex kn->active#52 nsim_bus_dev_list_lock &dev->mutex &devlink->lock_key#5 rtnl_mutex &n->list_lock &c->lock irq_context: 0 sb_writers#8 &of->mutex kn->active#52 nsim_bus_dev_list_lock &dev->mutex &devlink->lock_key#5 rtnl_mutex &obj_hash[i].lock pool_lock irq_context: 0 sb_writers#8 &of->mutex kn->active#52 nsim_bus_dev_list_lock &dev->mutex &devlink->lock_key#5 rtnl_mutex fill_pool_map-wait-type-override pool_lock#2 irq_context: 0 sb_writers#8 &of->mutex kn->active#52 nsim_bus_dev_list_lock &dev->mutex &devlink->lock_key#5 rtnl_mutex fill_pool_map-wait-type-override pool_lock irq_context: 0 sb_writers#8 &of->mutex kn->active#52 nsim_bus_dev_list_lock &dev->mutex &devlink->lock_key#5 &n->list_lock irq_context: 0 sb_writers#8 &of->mutex kn->active#52 nsim_bus_dev_list_lock &dev->mutex &devlink->lock_key#5 &n->list_lock &c->lock irq_context: 0 sb_writers#8 &of->mutex kn->active#52 nsim_bus_dev_list_lock &dev->mutex &devlink->lock_key#5 rtnl_mutex &idev->mc_lock &c->lock irq_context: 0 sb_writers#8 &of->mutex kn->active#52 nsim_bus_dev_list_lock &dev->mutex &devlink->lock_key#5 rtnl_mutex &idev->mc_lock &____s->seqcount irq_context: 0 sb_writers#8 &of->mutex kn->active#52 nsim_bus_dev_list_lock &dev->mutex &devlink->lock_key#5 rtnl_mutex &idev->mc_lock pool_lock#2 irq_context: 0 sb_writers#8 &of->mutex kn->active#52 nsim_bus_dev_list_lock &dev->mutex &devlink->lock_key#5 rtnl_mutex &idev->mc_lock &obj_hash[i].lock pool_lock irq_context: 0 sb_writers#8 &of->mutex kn->active#52 nsim_bus_dev_list_lock &dev->mutex &devlink->lock_key#5 rtnl_mutex &idev->mc_lock fill_pool_map-wait-type-override &c->lock irq_context: 0 sb_writers#8 &of->mutex kn->active#52 nsim_bus_dev_list_lock &dev->mutex &devlink->lock_key#5 rtnl_mutex &idev->mc_lock fill_pool_map-wait-type-override &____s->seqcount irq_context: 0 sb_writers#8 &of->mutex kn->active#52 nsim_bus_dev_list_lock &dev->mutex &devlink->lock_key#5 rtnl_mutex &idev->mc_lock fill_pool_map-wait-type-override pool_lock#2 irq_context: 0 sb_writers#8 &of->mutex kn->active#52 nsim_bus_dev_list_lock &dev->mutex &devlink->lock_key#5 rtnl_mutex &idev->mc_lock fill_pool_map-wait-type-override pool_lock irq_context: 0 sb_writers#8 &of->mutex kn->active#52 nsim_bus_dev_list_lock &dev->mutex &rq->__lock irq_context: 0 sb_writers#8 &of->mutex kn->active#52 nsim_bus_dev_list_lock &dev->mutex uevent_sock_mutex &n->list_lock irq_context: 0 sb_writers#8 &of->mutex kn->active#52 nsim_bus_dev_list_lock &dev->mutex uevent_sock_mutex &n->list_lock &c->lock irq_context: 0 cb_lock genl_mutex genl_mutex.wait_lock irq_context: 0 cb_lock genl_mutex &rq->__lock irq_context: 0 cb_lock genl_mutex &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)ipv6_addrconf (work_completion)(&(&ifa->dad_work)->work) rtnl_mutex &idev->mc_lock &dev_addr_list_lock_key#2/1 irq_context: 0 cb_lock genl_mutex.wait_lock irq_context: 0 cb_lock &p->pi_lock irq_context: 0 cb_lock &p->pi_lock &rq->__lock irq_context: 0 cb_lock &p->pi_lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 sb_writers#8 &of->mutex kn->active#52 nsim_bus_dev_list_lock &dev->mutex &devlink->lock_key#6 irq_context: 0 sb_writers#8 &of->mutex kn->active#52 nsim_bus_dev_list_lock &dev->mutex &devlink->lock_key#6 crngs.lock irq_context: 0 sb_writers#8 &of->mutex kn->active#52 nsim_bus_dev_list_lock &dev->mutex &devlink->lock_key#6 fs_reclaim irq_context: 0 sb_writers#8 &of->mutex kn->active#52 nsim_bus_dev_list_lock &dev->mutex &devlink->lock_key#6 fs_reclaim mmu_notifier_invalidate_range_start irq_context: 0 sb_writers#8 &of->mutex kn->active#52 nsim_bus_dev_list_lock &dev->mutex &devlink->lock_key#6 devlinks.xa_lock irq_context: 0 sb_writers#8 &of->mutex kn->active#52 nsim_bus_dev_list_lock &dev->mutex &devlink->lock_key#6 &c->lock irq_context: 0 sb_writers#8 &of->mutex kn->active#52 nsim_bus_dev_list_lock &dev->mutex &devlink->lock_key#6 &n->list_lock irq_context: 0 sb_writers#8 &of->mutex kn->active#52 nsim_bus_dev_list_lock &dev->mutex &devlink->lock_key#6 &n->list_lock &c->lock irq_context: 0 sb_writers#8 &of->mutex kn->active#52 nsim_bus_dev_list_lock &dev->mutex &devlink->lock_key#6 &obj_hash[i].lock irq_context: 0 sb_writers#8 &of->mutex kn->active#52 nsim_bus_dev_list_lock &dev->mutex &devlink->lock_key#6 nl_table_lock irq_context: 0 sb_writers#8 &of->mutex kn->active#52 nsim_bus_dev_list_lock &dev->mutex &devlink->lock_key#6 nl_table_wait.lock irq_context: 0 sb_writers#8 &of->mutex kn->active#52 nsim_bus_dev_list_lock &dev->mutex &devlink->lock_key#6 &xa->xa_lock#16 irq_context: 0 sb_writers#8 &of->mutex kn->active#52 nsim_bus_dev_list_lock &dev->mutex &devlink->lock_key#6 &____s->seqcount irq_context: 0 sb_writers#8 &of->mutex kn->active#52 nsim_bus_dev_list_lock &dev->mutex &devlink->lock_key#6 pcpu_alloc_mutex irq_context: 0 sb_writers#8 &of->mutex kn->active#52 nsim_bus_dev_list_lock &dev->mutex &devlink->lock_key#6 pcpu_alloc_mutex pcpu_lock irq_context: 0 sb_writers#8 &of->mutex kn->active#52 nsim_bus_dev_list_lock &dev->mutex &devlink->lock_key#6 &pcp->lock &zone->lock irq_context: 0 sb_writers#8 &of->mutex kn->active#52 nsim_bus_dev_list_lock &dev->mutex &devlink->lock_key#6 &pcp->lock &zone->lock &____s->seqcount irq_context: 0 sb_writers#8 &of->mutex kn->active#52 nsim_bus_dev_list_lock &dev->mutex &devlink->lock_key#6 &base->lock irq_context: 0 sb_writers#8 &of->mutex kn->active#52 nsim_bus_dev_list_lock &dev->mutex &devlink->lock_key#6 &base->lock &obj_hash[i].lock irq_context: 0 sb_writers#8 &of->mutex kn->active#52 nsim_bus_dev_list_lock &dev->mutex &devlink->lock_key#6 &rq->__lock irq_context: 0 sb_writers#8 &of->mutex kn->active#52 nsim_bus_dev_list_lock &dev->mutex &devlink->lock_key#6 &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 rtnl_mutex dev_addr_sem team->team_lock_key irq_context: 0 rtnl_mutex &dev_addr_list_lock_key#2/1 irq_context: 0 rtnl_mutex &dev_addr_list_lock_key#2/1 rcu_read_lock _xmit_ETHER irq_context: 0 rtnl_mutex &dev_addr_list_lock_key#2/1 rcu_read_lock _xmit_ETHER pool_lock#2 irq_context: 0 sb_writers#8 &of->mutex kn->active#52 nsim_bus_dev_list_lock &dev->mutex &devlink->lock_key#6 pin_fs_lock irq_context: 0 sb_writers#8 &of->mutex kn->active#52 nsim_bus_dev_list_lock &dev->mutex &devlink->lock_key#6 &sb->s_type->i_mutex_key#3 irq_context: 0 sb_writers#8 &of->mutex kn->active#52 nsim_bus_dev_list_lock &dev->mutex &devlink->lock_key#6 &sb->s_type->i_mutex_key#3 rename_lock.seqcount irq_context: 0 sb_writers#8 &of->mutex kn->active#52 nsim_bus_dev_list_lock &dev->mutex &devlink->lock_key#6 &sb->s_type->i_mutex_key#3 fs_reclaim irq_context: 0 sb_writers#8 &of->mutex kn->active#52 nsim_bus_dev_list_lock &dev->mutex &devlink->lock_key#6 &sb->s_type->i_mutex_key#3 fs_reclaim mmu_notifier_invalidate_range_start irq_context: 0 sb_writers#8 &of->mutex kn->active#52 nsim_bus_dev_list_lock &dev->mutex &devlink->lock_key#6 &sb->s_type->i_mutex_key#3 &dentry->d_lock irq_context: 0 sb_writers#8 &of->mutex kn->active#52 nsim_bus_dev_list_lock &dev->mutex &devlink->lock_key#6 &sb->s_type->i_mutex_key#3 rcu_read_lock rename_lock.seqcount irq_context: 0 sb_writers#8 &of->mutex kn->active#52 nsim_bus_dev_list_lock &dev->mutex &devlink->lock_key#6 &sb->s_type->i_mutex_key#3 &dentry->d_lock &wq irq_context: 0 sb_writers#8 &of->mutex kn->active#52 nsim_bus_dev_list_lock &dev->mutex &devlink->lock_key#6 &sb->s_type->i_mutex_key#3 mmu_notifier_invalidate_range_start irq_context: 0 sb_writers#8 &of->mutex kn->active#52 nsim_bus_dev_list_lock &dev->mutex &devlink->lock_key#6 &sb->s_type->i_mutex_key#3 &sb->s_type->i_lock_key#7 irq_context: 0 sb_writers#8 &of->mutex kn->active#52 nsim_bus_dev_list_lock &dev->mutex &devlink->lock_key#6 &sb->s_type->i_mutex_key#3 &s->s_inode_list_lock irq_context: 0 sb_writers#8 &of->mutex kn->active#52 nsim_bus_dev_list_lock &dev->mutex &devlink->lock_key#6 &sb->s_type->i_mutex_key#3 tk_core.seq.seqcount irq_context: 0 sb_writers#8 &of->mutex kn->active#52 nsim_bus_dev_list_lock &dev->mutex &devlink->lock_key#6 &sb->s_type->i_mutex_key#3 &sb->s_type->i_lock_key#7 &dentry->d_lock irq_context: 0 sb_writers#8 &of->mutex kn->active#52 nsim_bus_dev_list_lock &dev->mutex &devlink->lock_key#6 &sb->s_type->i_mutex_key#3 &c->lock irq_context: 0 sb_writers#8 &of->mutex kn->active#52 nsim_bus_dev_list_lock &dev->mutex &devlink->lock_key#6 &sb->s_type->i_mutex_key#3 &____s->seqcount irq_context: 0 sb_writers#8 &of->mutex kn->active#52 nsim_bus_dev_list_lock &dev->mutex &devlink->lock_key#6 &sb->s_type->i_mutex_key#3 pool_lock#2 irq_context: 0 sb_writers#8 &of->mutex kn->active#52 nsim_bus_dev_list_lock &dev->mutex &devlink->lock_key#6 batched_entropy_u32.lock irq_context: 0 sb_writers#8 &of->mutex kn->active#52 nsim_bus_dev_list_lock &dev->mutex &devlink->lock_key#6 rtnl_mutex irq_context: 0 sb_writers#8 &of->mutex kn->active#52 nsim_bus_dev_list_lock &dev->mutex &devlink->lock_key#6 rtnl_mutex rtnl_mutex.wait_lock irq_context: 0 sb_writers#8 &of->mutex kn->active#52 nsim_bus_dev_list_lock &dev->mutex &devlink->lock_key#6 rtnl_mutex &rq->__lock irq_context: 0 sb_writers#8 &of->mutex kn->active#52 nsim_bus_dev_list_lock &dev->mutex &devlink->lock_key#6 rtnl_mutex &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 cb_lock genl_mutex rtnl_mutex &wg->device_update_lock fill_pool_map-wait-type-override &c->lock irq_context: 0 cb_lock genl_mutex rtnl_mutex &wg->device_update_lock fill_pool_map-wait-type-override &____s->seqcount irq_context: 0 cb_lock genl_mutex rcu_read_lock rcu_node_0 irq_context: 0 cb_lock genl_mutex rcu_read_lock &rq->__lock irq_context: 0 cb_lock nlk_cb_mutex-GENERIC &devlink->lock_key#5 irq_context: 0 cb_lock nlk_cb_mutex-GENERIC &devlink->lock_key#5 &devlink_port->type_lock irq_context: 0 &sb->s_type->i_mutex_key#10 (netlink_chain).rwsem &obj_hash[i].lock pool_lock irq_context: 0 (wq_completion)events (work_completion)(&(&nsim_dev->trap_data->trap_report_dw)->work) &devlink->lock_key#4 &nsim_trap_data->trap_lock irq_context: 0 (wq_completion)events (work_completion)(&(&nsim_dev->trap_data->trap_report_dw)->work) &devlink->lock_key#4 &nsim_trap_data->trap_lock crngs.lock irq_context: 0 (wq_completion)events (work_completion)(&(&nsim_dev->trap_data->trap_report_dw)->work) &devlink->lock_key#4 &nsim_trap_data->trap_lock &nsim_dev->fa_cookie_lock irq_context: 0 (wq_completion)events (work_completion)(&(&nsim_dev->trap_data->trap_report_dw)->work) &devlink->lock_key#4 &nsim_trap_data->trap_lock &obj_hash[i].lock irq_context: 0 (wq_completion)events (work_completion)(&(&nsim_dev->trap_data->trap_report_dw)->work) &devlink->lock_key#4 &nsim_trap_data->trap_lock &c->lock irq_context: 0 (wq_completion)events (work_completion)(&(&nsim_dev->trap_data->trap_report_dw)->work) &devlink->lock_key#4 &nsim_trap_data->trap_lock &____s->seqcount irq_context: 0 (wq_completion)events (work_completion)(&(&nsim_dev->trap_data->trap_report_dw)->work) &devlink->lock_key#4 &nsim_trap_data->trap_lock &n->list_lock irq_context: 0 (wq_completion)events (work_completion)(&(&nsim_dev->trap_data->trap_report_dw)->work) &devlink->lock_key#4 &nsim_trap_data->trap_lock &n->list_lock &c->lock irq_context: 0 sb_writers#8 &of->mutex kn->active#52 nsim_bus_dev_list_lock &dev->mutex &devlink->lock_key#6 rtnl_mutex rcu_read_lock &rq->__lock irq_context: 0 sb_writers#8 &of->mutex kn->active#52 nsim_bus_dev_list_lock &dev->mutex &devlink->lock_key#6 rtnl_mutex rcu_read_lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 sb_writers#8 &of->mutex kn->active#52 nsim_bus_dev_list_lock &dev->mutex &devlink->lock_key#6 rtnl_mutex &(&net->nexthop.notifier_chain)->rwsem irq_context: 0 sb_writers#8 &of->mutex kn->active#52 nsim_bus_dev_list_lock &dev->mutex &devlink->lock_key#6 rtnl_mutex.wait_lock irq_context: 0 sb_writers#8 &of->mutex kn->active#52 nsim_bus_dev_list_lock &dev->mutex &devlink->lock_key#6 &p->pi_lock irq_context: 0 sb_writers#8 &of->mutex kn->active#52 nsim_bus_dev_list_lock &dev->mutex &devlink->lock_key#6 rcu_read_lock &data->fib_event_queue_lock irq_context: 0 sb_writers#8 &of->mutex kn->active#52 nsim_bus_dev_list_lock &dev->mutex &devlink->lock_key#6 rcu_read_lock rcu_read_lock &pool->lock irq_context: 0 sb_writers#8 &of->mutex kn->active#52 nsim_bus_dev_list_lock &dev->mutex &devlink->lock_key#6 rcu_read_lock rcu_read_lock &pool->lock &obj_hash[i].lock irq_context: 0 sb_writers#8 &of->mutex kn->active#52 nsim_bus_dev_list_lock &dev->mutex &devlink->lock_key#6 rcu_read_lock rcu_read_lock &pool->lock &p->pi_lock irq_context: 0 sb_writers#8 &of->mutex kn->active#52 nsim_bus_dev_list_lock &dev->mutex &devlink->lock_key#6 rcu_read_lock rcu_read_lock &pool->lock &p->pi_lock &rq->__lock irq_context: 0 sb_writers#8 &of->mutex kn->active#52 nsim_bus_dev_list_lock &dev->mutex &devlink->lock_key#6 rcu_read_lock rcu_read_lock &pool->lock &p->pi_lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 sb_writers#8 &of->mutex kn->active#52 nsim_bus_dev_list_lock &dev->mutex &devlink->lock_key#6 rcu_read_lock rcu_node_0 irq_context: 0 sb_writers#8 &of->mutex kn->active#52 nsim_bus_dev_list_lock &dev->mutex &devlink->lock_key#6 rcu_read_lock &rq->__lock irq_context: 0 sb_writers#8 &of->mutex kn->active#52 nsim_bus_dev_list_lock &dev->mutex &devlink->lock_key#6 rcu_read_lock &tb->tb6_lock irq_context: 0 sb_writers#8 &of->mutex kn->active#52 nsim_bus_dev_list_lock &dev->mutex &devlink->lock_key#6 rcu_read_lock &tb->tb6_lock &net->ipv6.fib6_walker_lock irq_context: 0 sb_writers#8 &of->mutex kn->active#52 nsim_bus_dev_list_lock &dev->mutex &devlink->lock_key#6 rcu_read_lock &tb->tb6_lock &data->fib_event_queue_lock irq_context: 0 sb_writers#8 &of->mutex kn->active#52 nsim_bus_dev_list_lock &dev->mutex &devlink->lock_key#6 rcu_read_lock &obj_hash[i].lock irq_context: 0 sb_writers#8 &of->mutex kn->active#52 nsim_bus_dev_list_lock &dev->mutex &devlink->lock_key#6 &(&fn_net->fib_chain)->lock irq_context: 0 (wq_completion)ipv6_addrconf (work_completion)(&(&ifa->dad_work)->work) rtnl_mutex rcu_read_lock rcu_read_lock &pcp->lock &zone->lock irq_context: 0 sb_writers#8 &of->mutex kn->active#52 nsim_bus_dev_list_lock &dev->mutex &devlink->lock_key#6 &devlink_port->type_lock irq_context: 0 (wq_completion)ipv6_addrconf (work_completion)(&(&ifa->dad_work)->work) rtnl_mutex rcu_read_lock rcu_read_lock &pcp->lock &zone->lock &____s->seqcount irq_context: 0 sb_writers#8 &of->mutex kn->active#52 nsim_bus_dev_list_lock &dev->mutex &devlink->lock_key#6 stack_depot_init_mutex irq_context: 0 sb_writers#8 &of->mutex kn->active#52 nsim_bus_dev_list_lock &dev->mutex &devlink->lock_key#6 &sb->s_type->i_mutex_key#3 &rq->__lock irq_context: 0 sb_writers#8 &of->mutex kn->active#52 nsim_bus_dev_list_lock &dev->mutex &devlink->lock_key#6 rtnl_mutex bpf_devs_lock irq_context: 0 sb_writers#8 &of->mutex kn->active#52 nsim_bus_dev_list_lock &dev->mutex &devlink->lock_key#6 rtnl_mutex bpf_devs_lock fs_reclaim irq_context: 0 sb_writers#8 &of->mutex kn->active#52 nsim_bus_dev_list_lock &dev->mutex &devlink->lock_key#6 rtnl_mutex bpf_devs_lock fs_reclaim mmu_notifier_invalidate_range_start irq_context: 0 sb_writers#8 &of->mutex kn->active#52 nsim_bus_dev_list_lock &dev->mutex &devlink->lock_key#6 rtnl_mutex bpf_devs_lock rcu_read_lock rhashtable_bucket irq_context: 0 sb_writers#8 &of->mutex kn->active#52 nsim_bus_dev_list_lock &dev->mutex &devlink->lock_key#6 rtnl_mutex pin_fs_lock irq_context: 0 sb_writers#8 &of->mutex kn->active#52 nsim_bus_dev_list_lock &dev->mutex &devlink->lock_key#6 rtnl_mutex &sb->s_type->i_mutex_key#3 irq_context: 0 sb_writers#8 &of->mutex kn->active#52 nsim_bus_dev_list_lock &dev->mutex &devlink->lock_key#6 rtnl_mutex &sb->s_type->i_mutex_key#3 rename_lock.seqcount irq_context: 0 sb_writers#8 &of->mutex kn->active#52 nsim_bus_dev_list_lock &dev->mutex &devlink->lock_key#6 rtnl_mutex &sb->s_type->i_mutex_key#3 fs_reclaim irq_context: 0 sb_writers#8 &of->mutex kn->active#52 nsim_bus_dev_list_lock &dev->mutex &devlink->lock_key#6 rtnl_mutex &sb->s_type->i_mutex_key#3 fs_reclaim mmu_notifier_invalidate_range_start irq_context: 0 sb_writers#8 &of->mutex kn->active#52 nsim_bus_dev_list_lock &dev->mutex &devlink->lock_key#6 rtnl_mutex &sb->s_type->i_mutex_key#3 &dentry->d_lock irq_context: 0 sb_writers#8 &of->mutex kn->active#52 nsim_bus_dev_list_lock &dev->mutex &devlink->lock_key#6 rtnl_mutex &sb->s_type->i_mutex_key#3 rcu_read_lock rename_lock.seqcount irq_context: 0 sb_writers#8 &of->mutex kn->active#52 nsim_bus_dev_list_lock &dev->mutex &devlink->lock_key#6 rtnl_mutex &sb->s_type->i_mutex_key#3 &dentry->d_lock &wq irq_context: 0 sb_writers#8 &of->mutex kn->active#52 nsim_bus_dev_list_lock &dev->mutex &devlink->lock_key#6 rtnl_mutex &sb->s_type->i_mutex_key#3 &c->lock irq_context: 0 sb_writers#8 &of->mutex kn->active#52 nsim_bus_dev_list_lock &dev->mutex &devlink->lock_key#6 rtnl_mutex &sb->s_type->i_mutex_key#3 &____s->seqcount irq_context: 0 sb_writers#8 &of->mutex kn->active#52 nsim_bus_dev_list_lock &dev->mutex &devlink->lock_key#6 rtnl_mutex &sb->s_type->i_mutex_key#3 pool_lock#2 irq_context: 0 sb_writers#8 &of->mutex kn->active#52 nsim_bus_dev_list_lock &dev->mutex &devlink->lock_key#6 rtnl_mutex &sb->s_type->i_mutex_key#3 mmu_notifier_invalidate_range_start irq_context: 0 sb_writers#8 &of->mutex kn->active#52 nsim_bus_dev_list_lock &dev->mutex &devlink->lock_key#6 rtnl_mutex &sb->s_type->i_mutex_key#3 &sb->s_type->i_lock_key#7 irq_context: 0 sb_writers#8 &of->mutex kn->active#52 nsim_bus_dev_list_lock &dev->mutex &devlink->lock_key#6 rtnl_mutex &sb->s_type->i_mutex_key#3 &s->s_inode_list_lock irq_context: 0 sb_writers#8 &of->mutex kn->active#52 nsim_bus_dev_list_lock &dev->mutex &devlink->lock_key#6 rtnl_mutex &sb->s_type->i_mutex_key#3 tk_core.seq.seqcount irq_context: 0 sb_writers#8 &of->mutex kn->active#52 nsim_bus_dev_list_lock &dev->mutex &devlink->lock_key#6 rtnl_mutex &sb->s_type->i_mutex_key#3 &sb->s_type->i_lock_key#7 &dentry->d_lock irq_context: 0 sb_writers#8 &of->mutex kn->active#52 nsim_bus_dev_list_lock &dev->mutex &devlink->lock_key#6 rtnl_mutex &____s->seqcount irq_context: 0 sb_writers#8 &of->mutex kn->active#52 nsim_bus_dev_list_lock &dev->mutex &devlink->lock_key#6 rtnl_mutex &obj_hash[i].lock irq_context: 0 sb_writers#8 &of->mutex kn->active#52 nsim_bus_dev_list_lock &dev->mutex &devlink->lock_key#6 rtnl_mutex fs_reclaim irq_context: 0 sb_writers#8 &of->mutex kn->active#52 nsim_bus_dev_list_lock &dev->mutex &devlink->lock_key#6 rtnl_mutex fs_reclaim mmu_notifier_invalidate_range_start irq_context: 0 sb_writers#8 &of->mutex kn->active#52 nsim_bus_dev_list_lock &dev->mutex &devlink->lock_key#6 rtnl_mutex &xa->xa_lock#3 irq_context: 0 sb_writers#8 &of->mutex kn->active#52 nsim_bus_dev_list_lock &dev->mutex &devlink->lock_key#6 rtnl_mutex &base->lock irq_context: 0 sb_writers#8 &of->mutex kn->active#52 nsim_bus_dev_list_lock &dev->mutex &devlink->lock_key#6 rtnl_mutex &base->lock &obj_hash[i].lock irq_context: 0 sb_writers#8 &of->mutex kn->active#52 nsim_bus_dev_list_lock &dev->mutex &devlink->lock_key#6 rtnl_mutex (work_completion)(&(&devlink_port->type_warn_dw)->work) irq_context: 0 sb_writers#8 &of->mutex kn->active#52 nsim_bus_dev_list_lock &dev->mutex &devlink->lock_key#6 rtnl_mutex &devlink_port->type_lock irq_context: 0 sb_writers#8 &of->mutex kn->active#52 nsim_bus_dev_list_lock &dev->mutex &devlink->lock_key#6 rtnl_mutex nl_table_lock irq_context: 0 sb_writers#8 &of->mutex kn->active#52 nsim_bus_dev_list_lock &dev->mutex &devlink->lock_key#6 rtnl_mutex nl_table_wait.lock irq_context: 0 sb_writers#8 &of->mutex kn->active#52 nsim_bus_dev_list_lock &dev->mutex &devlink->lock_key#6 rtnl_mutex net_rwsem irq_context: 0 sb_writers#8 &of->mutex kn->active#52 nsim_bus_dev_list_lock &dev->mutex &devlink->lock_key#6 rtnl_mutex net_rwsem &list->lock#2 irq_context: 0 sb_writers#8 &of->mutex kn->active#52 nsim_bus_dev_list_lock &dev->mutex &devlink->lock_key#6 rtnl_mutex &tn->lock irq_context: 0 sb_writers#8 &of->mutex kn->active#52 nsim_bus_dev_list_lock &dev->mutex &devlink->lock_key#6 rtnl_mutex &x->wait#9 irq_context: 0 sb_writers#8 &of->mutex kn->active#52 nsim_bus_dev_list_lock &dev->mutex &devlink->lock_key#6 rtnl_mutex &k->list_lock irq_context: 0 sb_writers#8 &of->mutex kn->active#52 nsim_bus_dev_list_lock &dev->mutex &devlink->lock_key#6 rtnl_mutex gdp_mutex irq_context: 0 sb_writers#8 &of->mutex kn->active#52 nsim_bus_dev_list_lock &dev->mutex &devlink->lock_key#6 rtnl_mutex gdp_mutex &k->list_lock irq_context: 0 sb_writers#8 &of->mutex kn->active#52 nsim_bus_dev_list_lock &dev->mutex &devlink->lock_key#6 rtnl_mutex gdp_mutex fs_reclaim irq_context: 0 sb_writers#8 &of->mutex kn->active#52 nsim_bus_dev_list_lock &dev->mutex &devlink->lock_key#6 rtnl_mutex gdp_mutex fs_reclaim mmu_notifier_invalidate_range_start irq_context: 0 sb_writers#8 &of->mutex kn->active#52 nsim_bus_dev_list_lock &dev->mutex &devlink->lock_key#6 rtnl_mutex gdp_mutex lock irq_context: 0 sb_writers#8 &of->mutex kn->active#52 nsim_bus_dev_list_lock &dev->mutex &devlink->lock_key#6 rtnl_mutex gdp_mutex lock kernfs_idr_lock irq_context: 0 sb_writers#8 &of->mutex kn->active#52 nsim_bus_dev_list_lock &dev->mutex &devlink->lock_key#6 rtnl_mutex gdp_mutex &root->kernfs_rwsem irq_context: 0 sb_writers#8 &of->mutex kn->active#52 nsim_bus_dev_list_lock &dev->mutex &devlink->lock_key#6 rtnl_mutex gdp_mutex &root->kernfs_rwsem &root->kernfs_iattr_rwsem irq_context: 0 sb_writers#8 &of->mutex kn->active#52 nsim_bus_dev_list_lock &dev->mutex &devlink->lock_key#6 rtnl_mutex gdp_mutex kobj_ns_type_lock irq_context: 0 sb_writers#8 &of->mutex kn->active#52 nsim_bus_dev_list_lock &dev->mutex &devlink->lock_key#6 rtnl_mutex &c->lock irq_context: 0 sb_writers#8 &of->mutex kn->active#52 nsim_bus_dev_list_lock &dev->mutex &devlink->lock_key#6 rtnl_mutex lock irq_context: 0 sb_writers#8 &of->mutex kn->active#52 nsim_bus_dev_list_lock &dev->mutex &devlink->lock_key#6 rtnl_mutex lock kernfs_idr_lock irq_context: 0 sb_writers#8 &of->mutex kn->active#52 nsim_bus_dev_list_lock &dev->mutex &devlink->lock_key#6 rtnl_mutex &root->kernfs_rwsem irq_context: 0 sb_writers#8 &of->mutex kn->active#52 nsim_bus_dev_list_lock &dev->mutex &devlink->lock_key#6 rtnl_mutex &root->kernfs_rwsem &root->kernfs_iattr_rwsem irq_context: 0 sb_writers#8 &of->mutex kn->active#52 nsim_bus_dev_list_lock &dev->mutex &devlink->lock_key#6 rtnl_mutex bus_type_sem irq_context: 0 sb_writers#8 &of->mutex kn->active#52 nsim_bus_dev_list_lock &dev->mutex &devlink->lock_key#6 rtnl_mutex sysfs_symlink_target_lock irq_context: 0 sb_writers#8 &of->mutex kn->active#52 nsim_bus_dev_list_lock &dev->mutex &devlink->lock_key#6 rtnl_mutex &root->kernfs_rwsem irq_context: 0 sb_writers#8 &of->mutex kn->active#52 nsim_bus_dev_list_lock &dev->mutex &devlink->lock_key#6 rtnl_mutex &dev->power.lock irq_context: 0 sb_writers#8 &of->mutex kn->active#52 nsim_bus_dev_list_lock &dev->mutex &devlink->lock_key#6 rtnl_mutex dpm_list_mtx irq_context: 0 sb_writers#8 &of->mutex kn->active#52 nsim_bus_dev_list_lock &dev->mutex &devlink->lock_key#6 rtnl_mutex uevent_sock_mutex irq_context: 0 sb_writers#8 &of->mutex kn->active#52 nsim_bus_dev_list_lock &dev->mutex &devlink->lock_key#6 rtnl_mutex uevent_sock_mutex fs_reclaim irq_context: 0 sb_writers#8 &of->mutex kn->active#52 nsim_bus_dev_list_lock &dev->mutex &devlink->lock_key#6 rtnl_mutex uevent_sock_mutex fs_reclaim mmu_notifier_invalidate_range_start irq_context: 0 sb_writers#8 &of->mutex kn->active#52 nsim_bus_dev_list_lock &dev->mutex &devlink->lock_key#6 rtnl_mutex uevent_sock_mutex &c->lock irq_context: 0 sb_writers#8 &of->mutex kn->active#52 nsim_bus_dev_list_lock &dev->mutex &devlink->lock_key#6 rtnl_mutex uevent_sock_mutex nl_table_lock irq_context: 0 sb_writers#8 &of->mutex kn->active#52 nsim_bus_dev_list_lock &dev->mutex &devlink->lock_key#6 rtnl_mutex uevent_sock_mutex &obj_hash[i].lock irq_context: 0 sb_writers#8 &of->mutex kn->active#52 nsim_bus_dev_list_lock &dev->mutex &devlink->lock_key#6 rtnl_mutex uevent_sock_mutex nl_table_wait.lock irq_context: 0 sb_writers#8 &of->mutex kn->active#52 nsim_bus_dev_list_lock &dev->mutex &devlink->lock_key#6 rtnl_mutex &k->k_lock irq_context: 0 sb_writers#8 &of->mutex kn->active#52 nsim_bus_dev_list_lock &dev->mutex &devlink->lock_key#6 rtnl_mutex subsys mutex#17 irq_context: 0 sb_writers#8 &of->mutex kn->active#52 nsim_bus_dev_list_lock &dev->mutex &devlink->lock_key#6 rtnl_mutex subsys mutex#17 &k->k_lock irq_context: 0 sb_writers#8 &of->mutex kn->active#52 nsim_bus_dev_list_lock &dev->mutex &devlink->lock_key#6 rtnl_mutex &dir->lock#2 irq_context: 0 sb_writers#8 &of->mutex kn->active#52 nsim_bus_dev_list_lock &dev->mutex &devlink->lock_key#6 rtnl_mutex dev_hotplug_mutex irq_context: 0 sb_writers#8 &of->mutex kn->active#52 nsim_bus_dev_list_lock &dev->mutex &devlink->lock_key#6 rtnl_mutex dev_hotplug_mutex &dev->power.lock irq_context: 0 sb_writers#8 &of->mutex kn->active#52 nsim_bus_dev_list_lock &dev->mutex &devlink->lock_key#6 rtnl_mutex dev_base_lock irq_context: 0 sb_writers#8 &of->mutex kn->active#52 nsim_bus_dev_list_lock &dev->mutex &devlink->lock_key#6 rtnl_mutex input_pool.lock irq_context: 0 sb_writers#8 &of->mutex kn->active#52 nsim_bus_dev_list_lock &dev->mutex &devlink->lock_key#6 rtnl_mutex rcu_read_lock &pool->lock/1 irq_context: 0 sb_writers#8 &of->mutex kn->active#52 nsim_bus_dev_list_lock &dev->mutex &devlink->lock_key#6 rtnl_mutex rcu_read_lock &pool->lock/1 &obj_hash[i].lock irq_context: 0 sb_writers#8 &of->mutex kn->active#52 nsim_bus_dev_list_lock &dev->mutex &devlink->lock_key#6 rtnl_mutex rcu_read_lock &pool->lock/1 &p->pi_lock irq_context: 0 sb_writers#8 &of->mutex kn->active#52 nsim_bus_dev_list_lock &dev->mutex &devlink->lock_key#6 rtnl_mutex batched_entropy_u32.lock irq_context: 0 sb_writers#8 &of->mutex kn->active#52 nsim_bus_dev_list_lock &dev->mutex &devlink->lock_key#6 rtnl_mutex &tbl->lock irq_context: 0 sb_writers#8 &of->mutex kn->active#52 nsim_bus_dev_list_lock &dev->mutex &devlink->lock_key#6 rtnl_mutex sysctl_lock irq_context: 0 sb_writers#8 &of->mutex kn->active#52 nsim_bus_dev_list_lock &dev->mutex &devlink->lock_key#6 rtnl_mutex failover_lock irq_context: 0 sb_writers#8 &of->mutex kn->active#52 nsim_bus_dev_list_lock &dev->mutex &devlink->lock_key#6 rtnl_mutex batched_entropy_u32.lock crngs.lock irq_context: 0 sb_writers#8 &of->mutex kn->active#52 nsim_bus_dev_list_lock &dev->mutex &devlink->lock_key#6 rtnl_mutex pcpu_alloc_mutex irq_context: 0 sb_writers#8 &of->mutex kn->active#52 nsim_bus_dev_list_lock &dev->mutex &devlink->lock_key#6 rtnl_mutex pcpu_alloc_mutex pcpu_lock irq_context: 0 sb_writers#8 &of->mutex kn->active#52 nsim_bus_dev_list_lock &dev->mutex &devlink->lock_key#6 rtnl_mutex pool_lock#2 irq_context: 0 sb_writers#8 &of->mutex kn->active#52 nsim_bus_dev_list_lock &dev->mutex &devlink->lock_key#6 rtnl_mutex proc_subdir_lock irq_context: 0 sb_writers#8 &of->mutex kn->active#52 nsim_bus_dev_list_lock &dev->mutex &devlink->lock_key#6 rtnl_mutex proc_inum_ida.xa_lock irq_context: 0 sb_writers#8 &of->mutex kn->active#52 nsim_bus_dev_list_lock &dev->mutex &devlink->lock_key#6 rtnl_mutex proc_subdir_lock irq_context: 0 sb_writers#8 &of->mutex kn->active#52 nsim_bus_dev_list_lock &dev->mutex &devlink->lock_key#6 rtnl_mutex &obj_hash[i].lock pool_lock irq_context: 0 sb_writers#8 &of->mutex kn->active#52 nsim_bus_dev_list_lock &dev->mutex &devlink->lock_key#6 rtnl_mutex fill_pool_map-wait-type-override pool_lock irq_context: 0 sb_writers#8 &of->mutex kn->active#52 nsim_bus_dev_list_lock &dev->mutex &devlink->lock_key#6 rtnl_mutex &idev->mc_lock irq_context: 0 sb_writers#8 &of->mutex kn->active#52 nsim_bus_dev_list_lock &dev->mutex &devlink->lock_key#6 rtnl_mutex &idev->mc_lock fs_reclaim irq_context: 0 sb_writers#8 &of->mutex kn->active#52 nsim_bus_dev_list_lock &dev->mutex &devlink->lock_key#6 rtnl_mutex &idev->mc_lock fs_reclaim mmu_notifier_invalidate_range_start irq_context: 0 sb_writers#8 &of->mutex kn->active#52 nsim_bus_dev_list_lock &dev->mutex &devlink->lock_key#6 rtnl_mutex &idev->mc_lock &obj_hash[i].lock irq_context: 0 sb_writers#8 &of->mutex kn->active#52 nsim_bus_dev_list_lock &dev->mutex &devlink->lock_key#6 rtnl_mutex &idev->mc_lock _xmit_ETHER irq_context: 0 sb_writers#8 &of->mutex kn->active#52 nsim_bus_dev_list_lock &dev->mutex &devlink->lock_key#6 rtnl_mutex &pnettable->lock irq_context: 0 sb_writers#8 &of->mutex kn->active#52 nsim_bus_dev_list_lock &dev->mutex &devlink->lock_key#6 rtnl_mutex smc_ib_devices.mutex irq_context: 0 sb_writers#8 &of->mutex kn->active#52 nsim_bus_dev_list_lock &dev->mutex &devlink->lock_key#6 rtnl_mutex &vn->sock_lock irq_context: 0 sb_writers#8 &of->mutex kn->active#52 nsim_bus_dev_list_lock &dev->mutex &devlink->lock_key#6 rtnl_mutex &n->list_lock irq_context: 0 sb_writers#8 &of->mutex kn->active#52 nsim_bus_dev_list_lock &dev->mutex &devlink->lock_key#6 rtnl_mutex &n->list_lock &c->lock irq_context: 0 sb_writers#8 &of->mutex kn->active#52 nsim_bus_dev_list_lock &dev->mutex &devlink->lock_key#6 rtnl_mutex rcu_read_lock &ndev->lock irq_context: softirq (&tun->flow_gc_timer) irq_context: softirq (&tun->flow_gc_timer) &tun->lock irq_context: 0 (wq_completion)events (linkwatch_work).work rtnl_mutex rcu_read_lock &net->sctp.local_addr_lock &net->sctp.addr_wq_lock &obj_hash[i].lock irq_context: 0 (wq_completion)events (linkwatch_work).work rtnl_mutex rcu_read_lock &net->sctp.local_addr_lock &net->sctp.addr_wq_lock &base->lock irq_context: 0 (wq_completion)events (linkwatch_work).work rtnl_mutex rcu_read_lock &net->sctp.local_addr_lock &net->sctp.addr_wq_lock &base->lock &obj_hash[i].lock irq_context: 0 (wq_completion)events (linkwatch_work).work rtnl_mutex &n->list_lock irq_context: 0 (wq_completion)events (linkwatch_work).work rtnl_mutex &n->list_lock &c->lock irq_context: 0 (wq_completion)events (linkwatch_work).work rtnl_mutex &br->lock irq_context: 0 (wq_completion)events (linkwatch_work).work rtnl_mutex &br->lock pool_lock#2 irq_context: 0 (wq_completion)events (linkwatch_work).work rtnl_mutex &br->lock &dir->lock#2 irq_context: 0 (wq_completion)events (linkwatch_work).work rtnl_mutex &br->lock deferred_lock irq_context: 0 (wq_completion)events (linkwatch_work).work rtnl_mutex &br->lock rcu_read_lock &pool->lock irq_context: 0 (wq_completion)events (linkwatch_work).work rtnl_mutex &br->lock rcu_read_lock &pool->lock &obj_hash[i].lock irq_context: 0 (wq_completion)events (linkwatch_work).work rtnl_mutex &br->lock (console_sem).lock irq_context: 0 (wq_completion)events (linkwatch_work).work rtnl_mutex &br->lock console_lock console_srcu console_owner_lock irq_context: 0 (wq_completion)events (linkwatch_work).work rtnl_mutex &br->lock console_lock console_srcu console_owner irq_context: 0 (wq_completion)events (linkwatch_work).work rtnl_mutex &br->lock console_lock console_srcu console_owner &port_lock_key irq_context: softirq (&in_dev->mr_ifc_timer) rcu_read_lock &____s->seqcount#10 irq_context: 0 (wq_completion)events (linkwatch_work).work rtnl_mutex &br->lock console_lock console_srcu console_owner console_owner_lock irq_context: 0 (wq_completion)events (linkwatch_work).work rtnl_mutex &br->lock &c->lock irq_context: 0 (wq_completion)events (linkwatch_work).work rtnl_mutex &br->lock &____s->seqcount irq_context: 0 (wq_completion)events (linkwatch_work).work rtnl_mutex &br->lock &obj_hash[i].lock irq_context: 0 (wq_completion)events (linkwatch_work).work rtnl_mutex &br->lock nl_table_lock irq_context: 0 (wq_completion)events (linkwatch_work).work rtnl_mutex &br->lock nl_table_wait.lock irq_context: 0 (wq_completion)events (linkwatch_work).work rtnl_mutex &br->lock &base->lock irq_context: 0 (wq_completion)events (linkwatch_work).work rtnl_mutex &br->lock &base->lock &obj_hash[i].lock irq_context: 0 (wq_completion)events (linkwatch_work).work rtnl_mutex &br->lock &br->multicast_lock irq_context: 0 (wq_completion)events (linkwatch_work).work rtnl_mutex &br->lock &br->multicast_lock &obj_hash[i].lock irq_context: 0 (wq_completion)events (linkwatch_work).work rtnl_mutex &br->lock &br->multicast_lock &base->lock irq_context: 0 (wq_completion)events (linkwatch_work).work rtnl_mutex &br->lock &br->multicast_lock &base->lock &obj_hash[i].lock irq_context: 0 (wq_completion)events (linkwatch_work).work rtnl_mutex &br->lock lweventlist_lock irq_context: 0 (wq_completion)events (linkwatch_work).work rtnl_mutex &br->lock lweventlist_lock pool_lock#2 irq_context: 0 (wq_completion)events (linkwatch_work).work rtnl_mutex &br->lock lweventlist_lock &dir->lock#2 irq_context: 0 (wq_completion)events (linkwatch_work).work rtnl_mutex &br->lock rcu_read_lock &pool->lock &p->pi_lock irq_context: 0 (wq_completion)events (linkwatch_work).work rtnl_mutex &br->lock &pcp->lock &zone->lock irq_context: 0 (wq_completion)events (linkwatch_work).work rtnl_mutex &br->lock &pcp->lock &zone->lock &____s->seqcount irq_context: 0 rtnl_mutex (inetaddr_chain).rwsem rcu_read_lock rcu_read_lock &pool->lock &p->pi_lock &rq->__lock irq_context: 0 rtnl_mutex (inetaddr_chain).rwsem rcu_read_lock rcu_read_lock &pool->lock &p->pi_lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 rtnl_mutex (inetaddr_chain).rwsem rcu_read_lock rcu_node_0 irq_context: 0 rtnl_mutex (inetaddr_chain).rwsem rcu_read_lock &rq->__lock irq_context: 0 (wq_completion)ipv6_addrconf (work_completion)(&(&ifa->dad_work)->work) rtnl_mutex rcu_read_lock rcu_read_lock rcu_read_lock_bh rcu_read_lock tk_core.seq.seqcount irq_context: 0 (wq_completion)ipv6_addrconf (work_completion)(&(&ifa->dad_work)->work) rtnl_mutex rcu_read_lock rcu_read_lock rcu_read_lock_bh rcu_read_lock &list->lock#5 irq_context: 0 (wq_completion)ipv6_addrconf (work_completion)(&(&ifa->dad_work)->work) rtnl_mutex &idev->mc_lock fill_pool_map-wait-type-override &c->lock irq_context: 0 (wq_completion)ipv6_addrconf (work_completion)(&(&ifa->dad_work)->work) rtnl_mutex &idev->mc_lock fill_pool_map-wait-type-override &____s->seqcount irq_context: 0 (wq_completion)ipv6_addrconf (work_completion)(&(&ifa->dad_work)->work) rtnl_mutex &idev->mc_lock &____s->seqcount irq_context: 0 (wq_completion)ipv6_addrconf (work_completion)(&(&ifa->dad_work)->work) rtnl_mutex rcu_read_lock rcu_read_lock rcu_read_lock_bh rcu_read_lock &obj_hash[i].lock irq_context: 0 (wq_completion)ipv6_addrconf (work_completion)(&(&ifa->dad_work)->work) rtnl_mutex rcu_read_lock rcu_read_lock rcu_read_lock_bh rcu_read_lock &____s->seqcount irq_context: 0 (wq_completion)ipv6_addrconf (work_completion)(&(&ifa->dad_work)->work) rtnl_mutex rcu_read_lock rcu_read_lock rcu_read_lock_bh rcu_read_lock pool_lock#2 irq_context: 0 (wq_completion)ipv6_addrconf (work_completion)(&(&ifa->dad_work)->work) rtnl_mutex rcu_read_lock rcu_read_lock rcu_read_lock_bh rcu_read_lock rcu_read_lock pool_lock#2 irq_context: 0 (wq_completion)ipv6_addrconf (work_completion)(&(&ifa->dad_work)->work) rtnl_mutex rcu_read_lock rcu_read_lock rcu_read_lock_bh noop_qdisc.q.lock irq_context: 0 (wq_completion)ipv6_addrconf (work_completion)(&(&ifa->dad_work)->work) rtnl_mutex &ndev->lock &ifa->lock batched_entropy_u32.lock crngs.lock irq_context: softirq (&pmctx->ip6_own_query.timer) irq_context: softirq (&pmctx->ip6_own_query.timer) &br->multicast_lock irq_context: softirq (&pmctx->ip4_own_query.timer) irq_context: softirq (&pmctx->ip4_own_query.timer) &br->multicast_lock irq_context: 0 sb_writers#8 &of->mutex kn->active#52 nsim_bus_dev_list_lock &dev->mutex &devlink->lock_key#6 rtnl_mutex rcu_read_lock &pool->lock/1 &p->pi_lock &rq->__lock irq_context: 0 sb_writers#8 &of->mutex kn->active#52 nsim_bus_dev_list_lock &dev->mutex &devlink->lock_key#6 rtnl_mutex rcu_read_lock &pool->lock/1 &p->pi_lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 sb_writers#8 &of->mutex kn->active#52 nsim_bus_dev_list_lock &dev->mutex &devlink->lock_key#6 rtnl_mutex &idev->mc_lock _xmit_ETHER &c->lock irq_context: 0 sb_writers#8 &of->mutex kn->active#52 nsim_bus_dev_list_lock &dev->mutex &devlink->lock_key#6 rtnl_mutex &idev->mc_lock _xmit_ETHER &____s->seqcount irq_context: 0 sb_writers#8 &of->mutex kn->active#52 nsim_bus_dev_list_lock &dev->mutex &devlink->lock_key#6 &p->pi_lock &rq->__lock irq_context: 0 sb_writers#8 &of->mutex kn->active#52 nsim_bus_dev_list_lock &dev->mutex &devlink->lock_key#6 &p->pi_lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 rtnl_mutex &ndev->lock &ifa->lock batched_entropy_u32.lock irq_context: 0 rtnl_mutex &ndev->lock &ifa->lock crngs.lock irq_context: 0 rtnl_mutex &ndev->lock &ifa->lock &obj_hash[i].lock irq_context: 0 rtnl_mutex &ndev->lock &ifa->lock &base->lock irq_context: 0 rtnl_mutex &ndev->lock &ifa->lock &base->lock &obj_hash[i].lock irq_context: softirq rcu_read_lock &br->hash_lock irq_context: softirq rcu_read_lock &br->hash_lock pool_lock#2 irq_context: softirq rcu_read_lock &br->hash_lock rcu_read_lock rhashtable_bucket irq_context: softirq rcu_read_lock &br->hash_lock rcu_read_lock rcu_read_lock &pool->lock irq_context: softirq rcu_read_lock &br->hash_lock rcu_read_lock rcu_read_lock &pool->lock &obj_hash[i].lock irq_context: softirq rcu_read_lock &br->hash_lock rcu_read_lock rcu_read_lock &pool->lock pool_lock#2 irq_context: softirq rcu_read_lock &br->hash_lock batched_entropy_u8.lock irq_context: softirq rcu_read_lock &br->hash_lock kfence_freelist_lock irq_context: softirq rcu_read_lock &br->hash_lock nl_table_lock irq_context: softirq rcu_read_lock &br->hash_lock &obj_hash[i].lock irq_context: softirq rcu_read_lock &br->hash_lock &meta->lock irq_context: softirq rcu_read_lock &br->hash_lock nl_table_wait.lock irq_context: softirq rcu_read_lock &br->multicast_lock irq_context: softirq rcu_read_lock &br->multicast_lock pool_lock#2 irq_context: softirq rcu_read_lock &br->multicast_lock &obj_hash[i].lock irq_context: softirq rcu_read_lock &br->multicast_lock rcu_read_lock rhashtable_bucket irq_context: softirq rcu_read_lock &br->multicast_lock &dir->lock#2 irq_context: softirq rcu_read_lock &br->multicast_lock deferred_lock irq_context: softirq rcu_read_lock &br->multicast_lock rcu_read_lock &pool->lock irq_context: softirq rcu_read_lock &br->multicast_lock rcu_read_lock &pool->lock &obj_hash[i].lock irq_context: softirq rcu_read_lock &br->multicast_lock nl_table_lock irq_context: softirq rcu_read_lock &br->multicast_lock nl_table_wait.lock irq_context: softirq rcu_read_lock &br->multicast_lock &base->lock irq_context: softirq rcu_read_lock &br->multicast_lock &base->lock &obj_hash[i].lock irq_context: softirq rcu_read_lock &br->multicast_lock &c->lock irq_context: softirq rcu_read_lock &br->multicast_lock &____s->seqcount irq_context: 0 (wq_completion)mld (work_completion)(&(&idev->mc_ifc_work)->work) &idev->mc_lock rcu_read_lock rcu_read_lock rcu_read_lock_bh rcu_read_lock &br->multicast_lock irq_context: 0 (wq_completion)mld (work_completion)(&(&idev->mc_ifc_work)->work) &idev->mc_lock rcu_read_lock rcu_read_lock rcu_read_lock_bh rcu_read_lock &br->multicast_lock pool_lock#2 irq_context: 0 (wq_completion)mld (work_completion)(&(&idev->mc_ifc_work)->work) &idev->mc_lock rcu_read_lock rcu_read_lock rcu_read_lock_bh rcu_read_lock &br->multicast_lock &obj_hash[i].lock irq_context: 0 (wq_completion)mld (work_completion)(&(&idev->mc_ifc_work)->work) &idev->mc_lock rcu_read_lock rcu_read_lock rcu_read_lock_bh rcu_read_lock &br->multicast_lock rcu_read_lock rhashtable_bucket irq_context: 0 (wq_completion)mld (work_completion)(&(&idev->mc_ifc_work)->work) &idev->mc_lock rcu_read_lock rcu_read_lock rcu_read_lock_bh rcu_read_lock &br->multicast_lock &c->lock irq_context: 0 (wq_completion)mld (work_completion)(&(&idev->mc_ifc_work)->work) &idev->mc_lock rcu_read_lock rcu_read_lock rcu_read_lock_bh rcu_read_lock &br->multicast_lock &____s->seqcount irq_context: 0 (wq_completion)mld (work_completion)(&(&idev->mc_ifc_work)->work) &idev->mc_lock rcu_read_lock rcu_read_lock rcu_read_lock_bh rcu_read_lock &br->multicast_lock &dir->lock#2 irq_context: 0 (wq_completion)mld (work_completion)(&(&idev->mc_ifc_work)->work) &idev->mc_lock rcu_read_lock rcu_read_lock rcu_read_lock_bh rcu_read_lock &br->multicast_lock deferred_lock irq_context: 0 (wq_completion)mld (work_completion)(&(&idev->mc_ifc_work)->work) &idev->mc_lock rcu_read_lock rcu_read_lock rcu_read_lock_bh rcu_read_lock &br->multicast_lock nl_table_lock irq_context: 0 (wq_completion)mld (work_completion)(&(&idev->mc_ifc_work)->work) &idev->mc_lock rcu_read_lock rcu_read_lock rcu_read_lock_bh rcu_read_lock &br->multicast_lock nl_table_wait.lock irq_context: 0 (wq_completion)mld (work_completion)(&(&idev->mc_ifc_work)->work) &idev->mc_lock rcu_read_lock rcu_read_lock rcu_read_lock_bh rcu_read_lock &br->multicast_lock &base->lock irq_context: 0 (wq_completion)mld (work_completion)(&(&idev->mc_ifc_work)->work) &idev->mc_lock rcu_read_lock rcu_read_lock rcu_read_lock_bh rcu_read_lock &br->multicast_lock &base->lock &obj_hash[i].lock irq_context: 0 (wq_completion)mld (work_completion)(&(&idev->mc_ifc_work)->work) &idev->mc_lock rcu_read_lock rcu_read_lock rcu_read_lock_bh rcu_read_lock rcu_read_lock_bh &obj_hash[i].lock irq_context: 0 (wq_completion)mld (work_completion)(&(&idev->mc_ifc_work)->work) &idev->mc_lock rcu_read_lock rcu_read_lock rcu_read_lock_bh rcu_read_lock rcu_read_lock_bh pool_lock#2 irq_context: 0 (wq_completion)mld (work_completion)(&(&idev->mc_ifc_work)->work) &idev->mc_lock rcu_read_lock rcu_read_lock rcu_read_lock_bh rcu_read_lock rcu_read_lock_bh rcu_read_lock tk_core.seq.seqcount irq_context: 0 (wq_completion)mld (work_completion)(&(&idev->mc_ifc_work)->work) &idev->mc_lock rcu_read_lock rcu_read_lock rcu_read_lock_bh rcu_read_lock rcu_read_lock_bh rcu_read_lock &list->lock#5 irq_context: 0 (wq_completion)events (work_completion)(&(&nsim_dev->trap_data->trap_report_dw)->work) &devlink->lock_key#5 &nsim_trap_data->trap_lock irq_context: 0 (wq_completion)events (work_completion)(&(&nsim_dev->trap_data->trap_report_dw)->work) &devlink->lock_key#5 &nsim_trap_data->trap_lock &c->lock irq_context: 0 (wq_completion)events (work_completion)(&(&nsim_dev->trap_data->trap_report_dw)->work) &devlink->lock_key#5 &nsim_trap_data->trap_lock crngs.lock irq_context: 0 (wq_completion)events (work_completion)(&(&nsim_dev->trap_data->trap_report_dw)->work) &devlink->lock_key#5 &nsim_trap_data->trap_lock &nsim_dev->fa_cookie_lock irq_context: 0 (wq_completion)events (work_completion)(&(&nsim_dev->trap_data->trap_report_dw)->work) &devlink->lock_key#5 &nsim_trap_data->trap_lock &obj_hash[i].lock irq_context: 0 (wq_completion)events (work_completion)(&(&nsim_dev->trap_data->trap_report_dw)->work) &devlink->lock_key#5 &nsim_trap_data->trap_lock &____s->seqcount irq_context: softirq &(&conn->info_timer)->timer irq_context: softirq &(&conn->info_timer)->timer rcu_read_lock &pool->lock irq_context: softirq &(&conn->info_timer)->timer rcu_read_lock &pool->lock &obj_hash[i].lock irq_context: softirq &(&conn->info_timer)->timer rcu_read_lock &pool->lock &p->pi_lock irq_context: softirq &(&conn->info_timer)->timer rcu_read_lock &pool->lock &p->pi_lock &rq->__lock irq_context: softirq &(&conn->info_timer)->timer rcu_read_lock &pool->lock &p->pi_lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)events (work_completion)(&(&conn->info_timer)->work) irq_context: 0 (wq_completion)events (work_completion)(&(&conn->info_timer)->work) &conn->chan_lock irq_context: 0 (wq_completion)events (work_completion)(&(&nsim_dev->trap_data->trap_report_dw)->work) &devlink->lock_key#4 &nsim_trap_data->trap_lock &pcp->lock &zone->lock irq_context: 0 (wq_completion)events (work_completion)(&(&nsim_dev->trap_data->trap_report_dw)->work) &devlink->lock_key#4 &nsim_trap_data->trap_lock &pcp->lock &zone->lock &____s->seqcount irq_context: 0 (wq_completion)events (linkwatch_work).work rtnl_mutex &pool->lock &p->pi_lock irq_context: 0 (wq_completion)events (linkwatch_work).work rtnl_mutex &pool->lock &p->pi_lock &rq->__lock irq_context: 0 (wq_completion)events (linkwatch_work).work rtnl_mutex &pool->lock &p->pi_lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 sb_writers#8 &of->mutex kn->active#52 nsim_bus_dev_list_lock &dev->mutex &devlink->lock_key#6 rtnl_mutex &pcp->lock &zone->lock irq_context: 0 sb_writers#8 &of->mutex kn->active#52 nsim_bus_dev_list_lock &dev->mutex &devlink->lock_key#6 rtnl_mutex &pcp->lock &zone->lock &____s->seqcount irq_context: 0 cb_lock nlk_cb_mutex-GENERIC &devlink->lock_key#6 irq_context: 0 cb_lock nlk_cb_mutex-GENERIC &devlink->lock_key#6 &devlink_port->type_lock irq_context: 0 (wq_completion)events (work_completion)(&(&nsim_dev->trap_data->trap_report_dw)->work) &devlink->lock_key#3 &nsim_trap_data->trap_lock quarantine_lock irq_context: 0 (wq_completion)events (work_completion)(&data->fib_event_work) &data->fib_lock remove_cache_srcu irq_context: 0 (wq_completion)events (work_completion)(&data->fib_event_work) &data->fib_lock remove_cache_srcu quarantine_lock irq_context: 0 (wq_completion)events (work_completion)(&data->fib_event_work) &data->fib_lock remove_cache_srcu &c->lock irq_context: 0 (wq_completion)events (work_completion)(&data->fib_event_work) &data->fib_lock remove_cache_srcu &n->list_lock irq_context: 0 (wq_completion)events (work_completion)(&data->fib_event_work) &data->fib_lock remove_cache_srcu rcu_read_lock pool_lock#2 irq_context: 0 (wq_completion)events (work_completion)(&data->fib_event_work) &data->fib_lock remove_cache_srcu &obj_hash[i].lock irq_context: 0 (wq_completion)mld (work_completion)(&(&idev->mc_ifc_work)->work) &idev->mc_lock rcu_read_lock rcu_read_lock rcu_read_lock_bh rcu_read_lock &br->multicast_lock rcu_read_lock &pool->lock irq_context: 0 (wq_completion)mld (work_completion)(&(&idev->mc_ifc_work)->work) &idev->mc_lock rcu_read_lock rcu_read_lock rcu_read_lock_bh rcu_read_lock &br->multicast_lock rcu_read_lock &pool->lock &obj_hash[i].lock irq_context: 0 (wq_completion)events (work_completion)(&data->fib_event_work) &data->fib_lock rcu_read_lock &rq->__lock irq_context: 0 (wq_completion)events (work_completion)(&data->fib_event_work) &data->fib_lock rcu_read_lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)events (linkwatch_work).work rtnl_mutex rcu_read_lock &pool->lock &p->pi_lock irq_context: 0 (wq_completion)events (linkwatch_work).work rtnl_mutex rcu_read_lock &pool->lock &p->pi_lock &rq->__lock irq_context: 0 (wq_completion)events (linkwatch_work).work rtnl_mutex rcu_read_lock &pool->lock &p->pi_lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)events (work_completion)(&data->fib_event_work) &data->fib_lock &____s->seqcount irq_context: 0 (wq_completion)hci0#2 (work_completion)(&(&hdev->cmd_timer)->work) &rq->__lock irq_context: 0 (wq_completion)hci0#2 (work_completion)(&hdev->cmd_work) &meta->lock irq_context: 0 rtnl_mutex &dev_addr_list_lock_key/1 rcu_read_lock _xmit_ETHER pool_lock#2 irq_context: 0 (wq_completion)events (work_completion)(&(&nsim_dev->trap_data->trap_report_dw)->work) &devlink->lock_key#5 &nsim_trap_data->trap_lock &pcp->lock &zone->lock irq_context: 0 (wq_completion)events (work_completion)(&(&nsim_dev->trap_data->trap_report_dw)->work) &devlink->lock_key#5 &nsim_trap_data->trap_lock &pcp->lock &zone->lock &____s->seqcount irq_context: 0 (wq_completion)mld (work_completion)(&(&idev->mc_ifc_work)->work) &idev->mc_lock rcu_read_lock &n->list_lock irq_context: softirq rcu_read_lock rcu_read_lock rcu_read_lock rcu_read_lock_bh rcu_read_lock tk_core.seq.seqcount irq_context: softirq rcu_read_lock rcu_read_lock rcu_read_lock rcu_read_lock_bh rcu_read_lock &list->lock#5 irq_context: 0 rtnl_mutex (inetaddr_chain).rwsem &net->sctp.local_addr_lock &net->sctp.addr_wq_lock &____s->seqcount irq_context: 0 rtnl_mutex &ndev->lock rcu_read_lock &pool->lock &p->pi_lock &rq->__lock &cfs_rq->removed.lock irq_context: 0 cb_lock genl_mutex rtnl_mutex &rq->__lock &cfs_rq->removed.lock irq_context: 0 (wq_completion)events (work_completion)(&(&nsim_dev->trap_data->trap_report_dw)->work) fill_pool_map-wait-type-override pool_lock#2 irq_context: 0 &sb->s_type->i_mutex_key#10 (netlink_chain).rwsem fill_pool_map-wait-type-override pool_lock#2 irq_context: 0 (wq_completion)events (work_completion)(&(&nsim_dev->trap_data->trap_report_dw)->work) fill_pool_map-wait-type-override pool_lock irq_context: 0 &sb->s_type->i_mutex_key#10 (netlink_chain).rwsem fill_pool_map-wait-type-override &c->lock irq_context: 0 &sb->s_type->i_mutex_key#10 (netlink_chain).rwsem fill_pool_map-wait-type-override &____s->seqcount irq_context: 0 &sb->s_type->i_mutex_key#10 (netlink_chain).rwsem fill_pool_map-wait-type-override pool_lock irq_context: 0 (wq_completion)events (work_completion)(&(&nsim_dev->trap_data->trap_report_dw)->work) &devlink->lock_key#6 &nsim_trap_data->trap_lock irq_context: 0 (wq_completion)events (work_completion)(&(&nsim_dev->trap_data->trap_report_dw)->work) &devlink->lock_key#6 &nsim_trap_data->trap_lock crngs.lock irq_context: 0 (wq_completion)events (work_completion)(&(&nsim_dev->trap_data->trap_report_dw)->work) &devlink->lock_key#6 &nsim_trap_data->trap_lock &nsim_dev->fa_cookie_lock irq_context: 0 (wq_completion)events (work_completion)(&(&nsim_dev->trap_data->trap_report_dw)->work) &devlink->lock_key#6 &nsim_trap_data->trap_lock &obj_hash[i].lock irq_context: 0 (wq_completion)events (work_completion)(&(&nsim_dev->trap_data->trap_report_dw)->work) &devlink->lock_key#6 &nsim_trap_data->trap_lock &c->lock irq_context: 0 (wq_completion)events (work_completion)(&(&nsim_dev->trap_data->trap_report_dw)->work) &devlink->lock_key#6 &nsim_trap_data->trap_lock &____s->seqcount irq_context: 0 rtnl_mutex dev_addr_sem fill_pool_map-wait-type-override pool_lock#2 irq_context: 0 rtnl_mutex dev_addr_sem fill_pool_map-wait-type-override pool_lock irq_context: 0 (wq_completion)events (work_completion)(&(&nsim_dev->trap_data->trap_report_dw)->work) &devlink->lock_key &nsim_trap_data->trap_lock &n->list_lock irq_context: 0 (wq_completion)events (work_completion)(&(&nsim_dev->trap_data->trap_report_dw)->work) &devlink->lock_key &nsim_trap_data->trap_lock &n->list_lock &c->lock irq_context: 0 rtnl_mutex dev_addr_sem &br->lock &br->hash_lock &n->list_lock irq_context: 0 rtnl_mutex dev_addr_sem &br->lock &br->hash_lock &n->list_lock &c->lock irq_context: 0 rtnl_mutex (inetaddr_chain).rwsem rcu_read_lock &pool->lock/1 fill_pool_map-wait-type-override pool_lock#2 irq_context: 0 rtnl_mutex (inetaddr_chain).rwsem rcu_read_lock &pool->lock/1 fill_pool_map-wait-type-override &c->lock irq_context: 0 rtnl_mutex (inetaddr_chain).rwsem rcu_read_lock &pool->lock/1 fill_pool_map-wait-type-override &____s->seqcount irq_context: 0 rtnl_mutex (inetaddr_chain).rwsem rcu_read_lock &pool->lock/1 fill_pool_map-wait-type-override pool_lock irq_context: 0 (wq_completion)hci2#2 (work_completion)(&hdev->cmd_work) &meta->lock irq_context: 0 (wq_completion)ipv6_addrconf (work_completion)(&(&ifa->dad_work)->work) rtnl_mutex &idev->mc_lock _xmit_ETHER/1 irq_context: 0 cb_lock genl_mutex rtnl_mutex &wg->device_update_lock &n->list_lock irq_context: 0 cb_lock genl_mutex rtnl_mutex &wg->device_update_lock &n->list_lock &c->lock irq_context: 0 (wq_completion)events (linkwatch_work).work rtnl_mutex team->team_lock_key lweventlist_lock irq_context: 0 (wq_completion)events (linkwatch_work).work rtnl_mutex team->team_lock_key lweventlist_lock &c->lock irq_context: 0 (wq_completion)events (linkwatch_work).work rtnl_mutex team->team_lock_key lweventlist_lock &____s->seqcount irq_context: 0 (wq_completion)events (linkwatch_work).work rtnl_mutex team->team_lock_key lweventlist_lock pool_lock#2 irq_context: 0 (wq_completion)events (linkwatch_work).work rtnl_mutex team->team_lock_key lweventlist_lock &dir->lock#2 irq_context: 0 (wq_completion)events (linkwatch_work).work rtnl_mutex team->team_lock_key rcu_read_lock &pool->lock irq_context: 0 (wq_completion)events (linkwatch_work).work rtnl_mutex team->team_lock_key rcu_read_lock &pool->lock &obj_hash[i].lock irq_context: 0 (wq_completion)events (linkwatch_work).work rtnl_mutex &idev->mc_lock rcu_read_lock &pool->lock irq_context: 0 (wq_completion)events (linkwatch_work).work rtnl_mutex &idev->mc_lock rcu_read_lock &pool->lock &obj_hash[i].lock irq_context: 0 (wq_completion)events (linkwatch_work).work rtnl_mutex rcu_read_lock &net->sctp.local_addr_lock &net->sctp.addr_wq_lock &c->lock irq_context: 0 (wq_completion)events (linkwatch_work).work rtnl_mutex rcu_read_lock &net->sctp.local_addr_lock &net->sctp.addr_wq_lock &____s->seqcount irq_context: 0 (wq_completion)events (linkwatch_work).work rtnl_mutex remove_cache_srcu irq_context: 0 (wq_completion)events (linkwatch_work).work rtnl_mutex remove_cache_srcu quarantine_lock irq_context: 0 (wq_completion)events (linkwatch_work).work rtnl_mutex remove_cache_srcu &c->lock irq_context: 0 (wq_completion)events (linkwatch_work).work rtnl_mutex remove_cache_srcu &n->list_lock irq_context: 0 (wq_completion)events (linkwatch_work).work rtnl_mutex remove_cache_srcu rcu_read_lock pool_lock#2 irq_context: 0 (wq_completion)events (linkwatch_work).work rtnl_mutex remove_cache_srcu &obj_hash[i].lock irq_context: 0 (wq_completion)events (linkwatch_work).work rtnl_mutex remove_cache_srcu &pcp->lock &zone->lock irq_context: 0 (wq_completion)events (linkwatch_work).work rtnl_mutex remove_cache_srcu &pcp->lock &zone->lock &____s->seqcount irq_context: 0 (wq_completion)events (linkwatch_work).work rtnl_mutex remove_cache_srcu &rq->__lock irq_context: 0 (wq_completion)ipv6_addrconf (work_completion)(&(&ifa->dad_work)->work) rtnl_mutex &ndev->lock &ifa->lock rcu_read_lock &pool->lock irq_context: 0 (wq_completion)ipv6_addrconf (work_completion)(&(&ifa->dad_work)->work) rtnl_mutex &ndev->lock &ifa->lock rcu_read_lock &pool->lock &obj_hash[i].lock irq_context: 0 (wq_completion)ipv6_addrconf (work_completion)(&(&ifa->dad_work)->work) rtnl_mutex &ndev->lock &ifa->lock rcu_read_lock &pool->lock pool_lock#2 irq_context: 0 (wq_completion)events (linkwatch_work).work rtnl_mutex rcu_read_lock &obj_hash[i].lock pool_lock irq_context: 0 (wq_completion)events (linkwatch_work).work rtnl_mutex rcu_read_lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)events (linkwatch_work).work rtnl_mutex lweventlist_lock &obj_hash[i].lock irq_context: 0 (wq_completion)events (linkwatch_work).work rtnl_mutex lweventlist_lock &base->lock irq_context: 0 (wq_completion)events (linkwatch_work).work rtnl_mutex lweventlist_lock &base->lock &obj_hash[i].lock irq_context: 0 (wq_completion)ipv6_addrconf (work_completion)(&(&ifa->dad_work)->work) rtnl_mutex &idev->mc_lock &pcp->lock &zone->lock irq_context: 0 (wq_completion)ipv6_addrconf (work_completion)(&(&ifa->dad_work)->work) rtnl_mutex &idev->mc_lock &dev_addr_list_lock_key#2/1 rcu_read_lock _xmit_ETHER irq_context: 0 (wq_completion)ipv6_addrconf (work_completion)(&(&ifa->dad_work)->work) rtnl_mutex &idev->mc_lock &dev_addr_list_lock_key#2/1 rcu_read_lock _xmit_ETHER pool_lock#2 irq_context: 0 (wq_completion)events (work_completion)(&(&nsim_dev->trap_data->trap_report_dw)->work) &devlink->lock_key#5 &nsim_trap_data->trap_lock &n->list_lock irq_context: 0 (wq_completion)events (work_completion)(&(&nsim_dev->trap_data->trap_report_dw)->work) &devlink->lock_key#5 &nsim_trap_data->trap_lock &n->list_lock &c->lock irq_context: 0 rtnl_mutex _xmit_ETHER/1 irq_context: 0 rtnl_mutex remove_cache_srcu rcu_node_0 irq_context: 0 (wq_completion)mld (work_completion)(&(&idev->mc_ifc_work)->work) &idev->mc_lock rcu_read_lock rcu_read_lock &obj_hash[i].lock pool_lock irq_context: 0 (wq_completion)mld (work_completion)(&(&idev->mc_ifc_work)->work) &idev->mc_lock rcu_read_lock rcu_read_lock rcu_read_lock_bh fill_pool_map-wait-type-override pool_lock#2 irq_context: 0 (wq_completion)mld (work_completion)(&(&idev->mc_ifc_work)->work) &idev->mc_lock rcu_read_lock rcu_read_lock rcu_read_lock_bh fill_pool_map-wait-type-override &c->lock irq_context: 0 (wq_completion)mld (work_completion)(&(&idev->mc_ifc_work)->work) &idev->mc_lock rcu_read_lock rcu_read_lock rcu_read_lock_bh fill_pool_map-wait-type-override &____s->seqcount irq_context: 0 (wq_completion)mld (work_completion)(&(&idev->mc_ifc_work)->work) &idev->mc_lock rcu_read_lock rcu_read_lock rcu_read_lock_bh fill_pool_map-wait-type-override pool_lock irq_context: 0 (wq_completion)mld (work_completion)(&(&idev->mc_ifc_work)->work) &idev->mc_lock rcu_read_lock rcu_read_lock &pcp->lock &zone->lock irq_context: 0 (wq_completion)events (linkwatch_work).work rtnl_mutex rcu_read_lock fill_pool_map-wait-type-override pool_lock#2 irq_context: 0 (wq_completion)events (linkwatch_work).work rtnl_mutex rcu_read_lock fill_pool_map-wait-type-override pool_lock irq_context: 0 rtnl_mutex dev_addr_sem team->team_lock_key#3 irq_context: 0 (wq_completion)mld (work_completion)(&(&idev->mc_ifc_work)->work) &idev->mc_lock &rq->__lock irq_context: 0 (wq_completion)mld (work_completion)(&(&idev->mc_ifc_work)->work) &idev->mc_lock rcu_read_lock rcu_read_lock rcu_read_lock_bh &hsr->seqnr_lock irq_context: 0 (wq_completion)mld (work_completion)(&(&idev->mc_ifc_work)->work) &idev->mc_lock rcu_read_lock rcu_read_lock rcu_read_lock_bh &hsr->seqnr_lock rcu_read_lock pool_lock#2 irq_context: 0 (wq_completion)mld (work_completion)(&(&idev->mc_ifc_work)->work) &idev->mc_lock rcu_read_lock rcu_read_lock rcu_read_lock_bh &hsr->seqnr_lock rcu_read_lock &hsr->list_lock irq_context: 0 (wq_completion)mld (work_completion)(&(&idev->mc_ifc_work)->work) &idev->mc_lock rcu_read_lock rcu_read_lock rcu_read_lock_bh &hsr->seqnr_lock rcu_read_lock &new_node->seq_out_lock irq_context: 0 (wq_completion)mld (work_completion)(&(&idev->mc_ifc_work)->work) &idev->mc_lock rcu_read_lock rcu_read_lock rcu_read_lock_bh &hsr->seqnr_lock &obj_hash[i].lock irq_context: 0 (wq_completion)mld (work_completion)(&(&idev->mc_ifc_work)->work) &idev->mc_lock rcu_read_lock rcu_read_lock rcu_read_lock_bh &hsr->seqnr_lock pool_lock#2 irq_context: 0 (wq_completion)events (work_completion)(&(&nsim_dev->trap_data->trap_report_dw)->work) &devlink->lock_key#3 &nsim_trap_data->trap_lock &pcp->lock &zone->lock irq_context: 0 (wq_completion)events (work_completion)(&(&nsim_dev->trap_data->trap_report_dw)->work) &devlink->lock_key#6 &nsim_trap_data->trap_lock &pcp->lock &zone->lock irq_context: 0 (wq_completion)mld (work_completion)(&(&idev->mc_ifc_work)->work) &idev->mc_lock rcu_read_lock rcu_read_lock rcu_read_lock_bh dev->qdisc_tx_busylock ?: &qdisc_tx_busylock &____s->seqcount irq_context: 0 (wq_completion)mld (work_completion)(&(&idev->mc_ifc_work)->work) &idev->mc_lock rcu_read_lock rcu_read_lock rcu_read_lock_bh dev->qdisc_tx_busylock ?: &qdisc_tx_busylock rcu_read_lock pool_lock#2 irq_context: 0 (wq_completion)ipv6_addrconf (work_completion)(&(&ifa->dad_work)->work) rtnl_mutex rcu_read_lock rcu_read_lock rcu_read_lock_bh rcu_read_lock rcu_read_lock_bh rcu_read_lock tk_core.seq.seqcount irq_context: 0 (wq_completion)ipv6_addrconf (work_completion)(&(&ifa->dad_work)->work) rtnl_mutex rcu_read_lock rcu_read_lock rcu_read_lock_bh rcu_read_lock rcu_read_lock_bh rcu_read_lock &list->lock#5 irq_context: 0 (wq_completion)ipv6_addrconf (work_completion)(&(&ifa->dad_work)->work) rtnl_mutex rcu_read_lock rcu_read_lock rcu_read_lock_bh rcu_read_lock rcu_read_lock_bh &obj_hash[i].lock irq_context: 0 (wq_completion)ipv6_addrconf (work_completion)(&(&ifa->dad_work)->work) rtnl_mutex rcu_read_lock rcu_read_lock rcu_read_lock_bh rcu_read_lock rcu_read_lock_bh pool_lock#2 irq_context: 0 (wq_completion)ipv6_addrconf (work_completion)(&(&ifa->dad_work)->work) rtnl_mutex &idev->mc_lock _xmit_ETHER/1 &c->lock irq_context: 0 (wq_completion)ipv6_addrconf (work_completion)(&(&ifa->dad_work)->work) rtnl_mutex &idev->mc_lock _xmit_ETHER/1 &____s->seqcount irq_context: softirq &(&idev->mc_ifc_work)->timer rcu_read_lock &pool->lock &p->pi_lock &rq->__lock &cfs_rq->removed.lock irq_context: 0 (wq_completion)gid-cache-wq (work_completion)(&ndev_work->work) devices_rwsem rcu_read_lock rcu_node_0 irq_context: 0 (wq_completion)gid-cache-wq (work_completion)(&ndev_work->work) devices_rwsem rcu_read_lock &rq->__lock irq_context: 0 (wq_completion)mld (work_completion)(&(&idev->mc_ifc_work)->work) &idev->mc_lock rcu_read_lock rcu_read_lock rcu_read_lock_bh &hsr->seqnr_lock rcu_read_lock &c->lock irq_context: 0 (wq_completion)mld (work_completion)(&(&idev->mc_ifc_work)->work) &idev->mc_lock rcu_read_lock rcu_read_lock rcu_read_lock_bh &hsr->seqnr_lock rcu_read_lock rcu_read_lock_bh rcu_read_lock tk_core.seq.seqcount irq_context: 0 (wq_completion)mld (work_completion)(&(&idev->mc_ifc_work)->work) &idev->mc_lock rcu_read_lock rcu_read_lock rcu_read_lock_bh &hsr->seqnr_lock rcu_read_lock rcu_read_lock_bh rcu_read_lock &list->lock#5 irq_context: 0 rtnl_mutex (inetaddr_chain).rwsem rcu_read_lock &____s->seqcount irq_context: 0 (wq_completion)events (linkwatch_work).work rtnl_mutex fill_pool_map-wait-type-override pool_lock#2 irq_context: 0 (wq_completion)events (linkwatch_work).work rtnl_mutex fill_pool_map-wait-type-override pool_lock irq_context: softirq (&hsr->announce_timer) irq_context: softirq (&hsr->announce_timer) rcu_read_lock pool_lock#2 irq_context: softirq (&hsr->announce_timer) rcu_read_lock &hsr->seqnr_lock irq_context: softirq (&hsr->announce_timer) rcu_read_lock &hsr->seqnr_lock rcu_read_lock &new_node->seq_out_lock irq_context: softirq (&hsr->announce_timer) rcu_read_lock &hsr->seqnr_lock rcu_read_lock pool_lock#2 irq_context: softirq (&hsr->announce_timer) rcu_read_lock &hsr->seqnr_lock rcu_read_lock rcu_read_lock_bh rcu_read_lock tk_core.seq.seqcount irq_context: softirq (&hsr->announce_timer) rcu_read_lock &hsr->seqnr_lock rcu_read_lock rcu_read_lock_bh rcu_read_lock &list->lock#5 irq_context: softirq (&hsr->announce_timer) rcu_read_lock &hsr->seqnr_lock &obj_hash[i].lock irq_context: softirq (&hsr->announce_timer) rcu_read_lock &hsr->seqnr_lock pool_lock#2 irq_context: softirq (&hsr->announce_timer) rcu_read_lock &obj_hash[i].lock irq_context: softirq (&hsr->announce_timer) rcu_read_lock &base->lock irq_context: softirq (&hsr->announce_timer) rcu_read_lock &base->lock &obj_hash[i].lock irq_context: 0 (wq_completion)events (work_completion)(&(&nsim_dev->trap_data->trap_report_dw)->work) &devlink->lock_key#5 &nsim_trap_data->trap_lock quarantine_lock irq_context: softirq rcu_read_lock &br->hash_lock &c->lock irq_context: 0 (wq_completion)events (work_completion)(&data->fib_event_work) &data->fib_lock remove_cache_srcu &rq->__lock irq_context: 0 (wq_completion)events (work_completion)(&data->fib_event_work) &data->fib_lock &pool->lock &p->pi_lock &rq->__lock irq_context: 0 (wq_completion)events (work_completion)(&data->fib_event_work) &data->fib_lock &pool->lock &p->pi_lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)ipv6_addrconf (work_completion)(&(&ifa->dad_work)->work) rtnl_mutex rcu_read_lock rcu_read_lock &tbl->lock &c->lock irq_context: 0 (wq_completion)ipv6_addrconf (work_completion)(&(&ifa->dad_work)->work) rtnl_mutex rcu_read_lock rcu_read_lock &tbl->lock &____s->seqcount irq_context: 0 (wq_completion)ipv6_addrconf (work_completion)(&(&ifa->dad_work)->work) rtnl_mutex rcu_read_lock rcu_read_lock &tbl->lock batched_entropy_u32.lock crngs.lock irq_context: 0 rtnl_mutex dev_addr_sem team->team_lock_key#4 irq_context: softirq rcu_read_lock quarantine_lock irq_context: 0 (wq_completion)mld (work_completion)(&(&idev->mc_ifc_work)->work) &idev->mc_lock remove_cache_srcu irq_context: 0 (wq_completion)mld (work_completion)(&(&idev->mc_ifc_work)->work) &idev->mc_lock remove_cache_srcu quarantine_lock irq_context: softirq rcu_read_lock &n->list_lock irq_context: softirq rcu_read_lock &n->list_lock &c->lock irq_context: 0 rtnl_mutex &tb->tb6_lock quarantine_lock irq_context: 0 rtnl_mutex remove_cache_srcu &pcp->lock &zone->lock irq_context: 0 rtnl_mutex remove_cache_srcu &pcp->lock &zone->lock &____s->seqcount irq_context: softirq &(&br->gc_work)->timer rcu_read_lock &pool->lock &p->pi_lock &rq->__lock irq_context: softirq &(&br->gc_work)->timer rcu_read_lock &pool->lock &p->pi_lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)bat_events (work_completion)(&(&bat_priv->nc.work)->work) rcu_read_lock &rcu_state.gp_wq irq_context: 0 (wq_completion)bat_events (work_completion)(&(&bat_priv->nc.work)->work) rcu_read_lock &rcu_state.gp_wq &p->pi_lock irq_context: 0 (wq_completion)events (linkwatch_work).work rtnl_mutex &ndev->lock &ifa->lock batched_entropy_u32.lock crngs.lock irq_context: 0 rtnl_mutex rcu_read_lock rcu_read_lock rcu_node_0 irq_context: 0 rtnl_mutex rcu_read_lock rcu_read_lock &rq->__lock irq_context: 0 (wq_completion)gid-cache-wq (work_completion)(&work->work) &obj_hash[i].lock pool_lock irq_context: 0 (wq_completion)ipv6_addrconf (work_completion)(&(&ifa->dad_work)->work) rtnl_mutex rcu_read_lock rcu_read_lock rcu_read_lock_bh dev->qdisc_tx_busylock ?: &qdisc_tx_busylock _xmit_ETHER#2 &obj_hash[i].lock pool_lock irq_context: 0 rtnl_mutex pcpu_alloc_mutex &rq->__lock irq_context: 0 (wq_completion)events (linkwatch_work).work rtnl_mutex team->team_lock_key#3 irq_context: 0 (wq_completion)events (linkwatch_work).work rtnl_mutex team->team_lock_key#3 fs_reclaim irq_context: 0 (wq_completion)events (linkwatch_work).work rtnl_mutex team->team_lock_key#3 fs_reclaim mmu_notifier_invalidate_range_start irq_context: 0 (wq_completion)events (linkwatch_work).work rtnl_mutex team->team_lock_key#3 &c->lock irq_context: 0 (wq_completion)events (linkwatch_work).work rtnl_mutex team->team_lock_key#3 rcu_read_lock &ndev->lock irq_context: 0 (wq_completion)events (linkwatch_work).work rtnl_mutex team->team_lock_key#3 &obj_hash[i].lock irq_context: 0 (wq_completion)events (linkwatch_work).work rtnl_mutex team->team_lock_key#3 nl_table_lock irq_context: 0 (wq_completion)events (linkwatch_work).work rtnl_mutex team->team_lock_key#3 nl_table_wait.lock irq_context: 0 (wq_completion)events (linkwatch_work).work rtnl_mutex team->team_lock_key#3 net_rwsem irq_context: 0 (wq_completion)events (linkwatch_work).work rtnl_mutex team->team_lock_key#3 net_rwsem &list->lock#2 irq_context: 0 (wq_completion)events (linkwatch_work).work rtnl_mutex team->team_lock_key#3 &tn->lock irq_context: 0 (wq_completion)events (linkwatch_work).work rtnl_mutex team->team_lock_key#3 &____s->seqcount irq_context: 0 (wq_completion)events (linkwatch_work).work rtnl_mutex team->team_lock_key#3 lweventlist_lock irq_context: 0 (wq_completion)events (linkwatch_work).work rtnl_mutex team->team_lock_key#3 lweventlist_lock &dir->lock#2 irq_context: 0 (wq_completion)events (linkwatch_work).work rtnl_mutex team->team_lock_key#3 rcu_read_lock &pool->lock irq_context: 0 (wq_completion)events (linkwatch_work).work rtnl_mutex team->team_lock_key#3 rcu_read_lock &pool->lock &obj_hash[i].lock irq_context: softirq (&hsr->announce_timer) rcu_read_lock &c->lock irq_context: softirq (&in_dev->mr_ifc_timer) rcu_read_lock &im->lock rcu_read_lock &c->lock irq_context: 0 (wq_completion)ipv6_addrconf (work_completion)(&(&ifa->dad_work)->work) rtnl_mutex &idev->mc_lock &dev_addr_list_lock_key#2/1 rcu_read_lock _xmit_ETHER &c->lock irq_context: 0 (wq_completion)ipv6_addrconf (work_completion)(&(&ifa->dad_work)->work) rtnl_mutex &idev->mc_lock &dev_addr_list_lock_key#2/1 rcu_read_lock _xmit_ETHER &____s->seqcount irq_context: 0 rtnl_mutex (inetaddr_chain).rwsem remove_cache_srcu irq_context: 0 rtnl_mutex (inetaddr_chain).rwsem remove_cache_srcu quarantine_lock irq_context: 0 rtnl_mutex (inetaddr_chain).rwsem remove_cache_srcu &c->lock irq_context: 0 rtnl_mutex (inetaddr_chain).rwsem remove_cache_srcu &n->list_lock irq_context: 0 rtnl_mutex (inetaddr_chain).rwsem remove_cache_srcu rcu_read_lock pool_lock#2 irq_context: 0 rtnl_mutex (inetaddr_chain).rwsem remove_cache_srcu &obj_hash[i].lock irq_context: 0 rtnl_mutex (inetaddr_chain).rwsem remove_cache_srcu &pcp->lock &zone->lock irq_context: 0 rtnl_mutex (inetaddr_chain).rwsem remove_cache_srcu &pcp->lock &zone->lock &____s->seqcount irq_context: 0 rtnl_mutex (inetaddr_chain).rwsem remove_cache_srcu &rq->__lock irq_context: 0 rtnl_mutex (inetaddr_chain).rwsem remove_cache_srcu &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: softirq rcu_read_lock rcu_read_lock &pcp->lock &zone->lock irq_context: 0 rtnl_mutex (inetaddr_chain).rwsem remove_cache_srcu rcu_read_lock rcu_node_0 irq_context: 0 rtnl_mutex (inetaddr_chain).rwsem remove_cache_srcu rcu_read_lock &rq->__lock irq_context: 0 (wq_completion)mld (work_completion)(&(&idev->mc_ifc_work)->work) &idev->mc_lock rcu_read_lock rcu_read_lock rcu_read_lock_bh rcu_read_lock &c->lock irq_context: softirq rcu_read_lock &br->multicast_lock &obj_hash[i].lock pool_lock irq_context: softirq rcu_read_lock &br->multicast_lock fill_pool_map-wait-type-override pool_lock#2 irq_context: softirq rcu_read_lock &br->multicast_lock fill_pool_map-wait-type-override pool_lock irq_context: 0 (wq_completion)events (linkwatch_work).work rtnl_mutex rcu_read_lock &c->lock irq_context: 0 (wq_completion)events (linkwatch_work).work rtnl_mutex rcu_read_lock &____s->seqcount irq_context: 0 (wq_completion)events (linkwatch_work).work rtnl_mutex rcu_read_lock rcu_read_lock &pool->lock/1 &p->pi_lock &rq->__lock &cfs_rq->removed.lock irq_context: 0 (wq_completion)events (linkwatch_work).work rtnl_mutex &base->lock &obj_hash[i].lock irq_context: 0 remove_cache_srcu &rq->__lock &cfs_rq->removed.lock irq_context: 0 rtnl_mutex dev_addr_sem batched_entropy_u8.lock irq_context: 0 rtnl_mutex dev_addr_sem kfence_freelist_lock irq_context: 0 (wq_completion)gid-cache-wq (work_completion)(&ndev_work->work) &meta->lock irq_context: 0 (wq_completion)gid-cache-wq (work_completion)(&ndev_work->work) kfence_freelist_lock irq_context: 0 rtnl_mutex &nn->netlink_tap_lock irq_context: 0 rcu_read_lock rcu_read_lock_bh &obj_hash[i].lock irq_context: 0 rcu_read_lock rcu_read_lock_bh pool_lock#2 irq_context: 0 (wq_completion)events (linkwatch_work).work rtnl_mutex team->team_lock_key#4 lweventlist_lock irq_context: 0 (wq_completion)events (linkwatch_work).work rtnl_mutex team->team_lock_key#4 lweventlist_lock &dir->lock#2 irq_context: 0 (wq_completion)events (linkwatch_work).work rtnl_mutex team->team_lock_key#4 rcu_read_lock &pool->lock irq_context: 0 (wq_completion)events (linkwatch_work).work rtnl_mutex team->team_lock_key#4 rcu_read_lock &pool->lock &obj_hash[i].lock irq_context: 0 (wq_completion)events (linkwatch_work).work rtnl_mutex &tb->tb6_lock &n->list_lock irq_context: 0 (wq_completion)events (linkwatch_work).work rtnl_mutex &tb->tb6_lock &n->list_lock &c->lock irq_context: 0 (wq_completion)events (linkwatch_work).work rtnl_mutex rcu_read_lock &n->list_lock irq_context: 0 (wq_completion)events (linkwatch_work).work rtnl_mutex rcu_read_lock &n->list_lock &c->lock irq_context: 0 (wq_completion)events (work_completion)(&(&nsim_dev->trap_data->trap_report_dw)->work) &devlink->lock_key#4 &nsim_trap_data->trap_lock quarantine_lock irq_context: 0 rtnl_mutex rcu_read_lock rcu_read_lock &pool->lock/1 &p->pi_lock &rq->__lock &cfs_rq->removed.lock irq_context: 0 &mm->mmap_lock fill_pool_map-wait-type-override rcu_read_lock rcu_node_0 irq_context: 0 &mm->mmap_lock fill_pool_map-wait-type-override rcu_read_lock &rq->__lock irq_context: 0 (wq_completion)events (work_completion)(&(&nsim_dev->trap_data->trap_report_dw)->work) &devlink->lock_key#6 &nsim_trap_data->trap_lock quarantine_lock irq_context: 0 &mm->mmap_lock fill_pool_map-wait-type-override rcu_read_lock &rcu_state.gp_wq irq_context: 0 &mm->mmap_lock fill_pool_map-wait-type-override rcu_read_lock &rcu_state.gp_wq &p->pi_lock irq_context: 0 &mm->mmap_lock fill_pool_map-wait-type-override rcu_read_lock &rcu_state.gp_wq &p->pi_lock &rq->__lock irq_context: 0 &mm->mmap_lock fill_pool_map-wait-type-override rcu_read_lock &rcu_state.gp_wq &p->pi_lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)ipv6_addrconf (work_completion)(&(&ifa->dad_work)->work) rtnl_mutex &tb->tb6_lock rcu_read_lock &____s->seqcount irq_context: 0 rtnl_mutex (inetaddr_chain).rwsem rcu_read_lock &pool->lock/1 &p->pi_lock &rq->__lock &cfs_rq->removed.lock irq_context: 0 rtnl_mutex dev_addr_sem rcu_read_lock pool_lock#2 irq_context: 0 rtnl_mutex dev_addr_sem rcu_read_lock &bat_priv->tt.changes_list_lock irq_context: 0 rtnl_mutex dev_addr_sem rcu_read_lock &bat_priv->tt.changes_list_lock &obj_hash[i].lock irq_context: 0 rtnl_mutex dev_addr_sem rcu_read_lock &bat_priv->tt.changes_list_lock pool_lock#2 irq_context: 0 rtnl_mutex dev_addr_sem rcu_read_lock key#16 irq_context: 0 rtnl_mutex dev_addr_sem rcu_read_lock &obj_hash[i].lock irq_context: softirq &(&forw_packet_aggr->delayed_work)->timer rcu_read_lock &pool->lock/1 &p->pi_lock &rq->__lock &cfs_rq->removed.lock irq_context: 0 (wq_completion)mld (work_completion)(&(&idev->mc_ifc_work)->work) &idev->mc_lock rcu_read_lock rcu_read_lock rcu_read_lock_bh dev->qdisc_tx_busylock ?: &qdisc_tx_busylock _xmit_ETHER#2 fill_pool_map-wait-type-override pool_lock#2 irq_context: 0 (wq_completion)mld (work_completion)(&(&idev->mc_ifc_work)->work) &idev->mc_lock rcu_read_lock rcu_read_lock rcu_read_lock_bh dev->qdisc_tx_busylock ?: &qdisc_tx_busylock _xmit_ETHER#2 fill_pool_map-wait-type-override pool_lock irq_context: 0 rtnl_mutex &batadv_netdev_addr_lock_key/1 irq_context: softirq (&hsr->announce_timer) rcu_read_lock &hsr->seqnr_lock rcu_read_lock &c->lock irq_context: 0 (wq_completion)ipv6_addrconf (work_completion)(&(&ifa->dad_work)->work) rtnl_mutex &idev->mc_lock &dev_addr_list_lock_key/1 &c->lock irq_context: 0 (wq_completion)ipv6_addrconf (work_completion)(&(&ifa->dad_work)->work) rtnl_mutex &idev->mc_lock &dev_addr_list_lock_key/1 &____s->seqcount irq_context: 0 (wq_completion)ipv6_addrconf (work_completion)(&(&ifa->dad_work)->work) rtnl_mutex &idev->mc_lock &batadv_netdev_addr_lock_key/1 irq_context: 0 (wq_completion)events (linkwatch_work).work rtnl_mutex j1939_netdev_lock irq_context: 0 (wq_completion)ipv6_addrconf (work_completion)(&(&ifa->dad_work)->work) rtnl_mutex &pool->lock &p->pi_lock &rq->__lock &cfs_rq->removed.lock irq_context: 0 (wq_completion)events (linkwatch_work).work rtnl_mutex &ndev->lock &ifa->lock rcu_read_lock &pool->lock irq_context: 0 (wq_completion)events (linkwatch_work).work rtnl_mutex &ndev->lock &ifa->lock rcu_read_lock &pool->lock &obj_hash[i].lock irq_context: 0 (wq_completion)events (linkwatch_work).work rtnl_mutex &ndev->lock &ifa->lock rcu_read_lock &pool->lock pool_lock#2 irq_context: 0 (wq_completion)events (linkwatch_work).work rtnl_mutex &obj_hash[i].lock pool_lock irq_context: 0 rtnl_mutex dev_addr_sem team->team_lock_key#6 irq_context: 0 (wq_completion)events (work_completion)(&(&nsim_dev->trap_data->trap_report_dw)->work) &devlink->lock_key#5 &nsim_trap_data->trap_lock batched_entropy_u8.lock irq_context: 0 (wq_completion)events (work_completion)(&(&nsim_dev->trap_data->trap_report_dw)->work) &devlink->lock_key#5 &nsim_trap_data->trap_lock batched_entropy_u8.lock crngs.lock irq_context: 0 (wq_completion)events (work_completion)(&(&nsim_dev->trap_data->trap_report_dw)->work) &devlink->lock_key#5 &nsim_trap_data->trap_lock kfence_freelist_lock irq_context: 0 (wq_completion)events (work_completion)(&(&nsim_dev->trap_data->trap_report_dw)->work) &devlink->lock_key#5 &nsim_trap_data->trap_lock &meta->lock irq_context: 0 (wq_completion)events (linkwatch_work).work rtnl_mutex rcu_read_lock rcu_read_lock &pool->lock/1 fill_pool_map-wait-type-override pool_lock#2 irq_context: 0 (wq_completion)events (linkwatch_work).work rtnl_mutex rcu_read_lock rcu_read_lock &pool->lock/1 fill_pool_map-wait-type-override pool_lock irq_context: 0 (wq_completion)mld (work_completion)(&(&idev->mc_ifc_work)->work) &idev->mc_lock rcu_read_lock rcu_read_lock rcu_read_lock_bh &hsr->seqnr_lock rcu_read_lock &____s->seqcount irq_context: 0 (wq_completion)events (work_completion)(&data->fib_event_work) &data->fib_lock rcu_read_lock rcu_read_lock &pool->lock irq_context: 0 (wq_completion)events (work_completion)(&data->fib_event_work) &data->fib_lock rcu_read_lock rcu_read_lock &pool->lock &obj_hash[i].lock irq_context: 0 (wq_completion)events (work_completion)(&data->fib_event_work) &data->fib_lock rcu_read_lock rcu_read_lock &pool->lock pool_lock#2 irq_context: 0 (wq_completion)mld (work_completion)(&(&idev->mc_ifc_work)->work) &idev->mc_lock rcu_read_lock rcu_read_lock rcu_read_lock_bh dev->qdisc_tx_busylock ?: &qdisc_tx_busylock _xmit_ETHER#2 quarantine_lock irq_context: 0 (wq_completion)events (linkwatch_work).work rtnl_mutex rcu_read_lock &rcu_state.gp_wq irq_context: 0 (wq_completion)events (linkwatch_work).work rtnl_mutex rcu_read_lock &rcu_state.gp_wq &p->pi_lock irq_context: 0 (wq_completion)events (linkwatch_work).work rtnl_mutex rcu_read_lock &rcu_state.gp_wq &p->pi_lock &rq->__lock irq_context: 0 (wq_completion)events (linkwatch_work).work rtnl_mutex rcu_read_lock &rcu_state.gp_wq &p->pi_lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)events (work_completion)(&(&nsim_dev->trap_data->trap_report_dw)->work) &devlink->lock_key#6 &nsim_trap_data->trap_lock &n->list_lock irq_context: 0 (wq_completion)events (work_completion)(&(&nsim_dev->trap_data->trap_report_dw)->work) &devlink->lock_key#6 &nsim_trap_data->trap_lock &n->list_lock &c->lock irq_context: 0 (wq_completion)mld (work_completion)(&(&idev->mc_ifc_work)->work) &idev->mc_lock rcu_read_lock rcu_read_lock rcu_node_0 irq_context: 0 (wq_completion)mld (work_completion)(&(&idev->mc_ifc_work)->work) &idev->mc_lock rcu_read_lock rcu_read_lock &rq->__lock irq_context: 0 (wq_completion)mld (work_completion)(&(&idev->mc_ifc_work)->work) &idev->mc_lock rcu_read_lock rcu_node_0 irq_context: softirq rcu_read_lock rcu_read_lock rcu_read_lock rcu_read_lock &ndev->lock irq_context: 0 &mm->mmap_lock remove_cache_srcu rcu_read_lock &rcu_state.gp_wq &p->pi_lock &cfs_rq->removed.lock irq_context: 0 (wq_completion)mld (work_completion)(&(&idev->mc_ifc_work)->work) &idev->mc_lock rcu_read_lock rcu_read_lock rcu_read_lock_bh rcu_read_lock rcu_read_lock_bh noop_qdisc.q.lock irq_context: softirq &(&nsim_dev->trap_data->trap_report_dw)->timer rcu_read_lock &pool->lock &p->pi_lock &rq->__lock &cfs_rq->removed.lock irq_context: 0 (wq_completion)ipv6_addrconf (work_completion)(&(&ifa->dad_work)->work) rtnl_mutex rcu_read_lock rcu_read_lock rcu_read_lock_bh &hsr->seqnr_lock irq_context: 0 (wq_completion)ipv6_addrconf (work_completion)(&(&ifa->dad_work)->work) rtnl_mutex rcu_read_lock rcu_read_lock rcu_read_lock_bh &hsr->seqnr_lock rcu_read_lock &new_node->seq_out_lock irq_context: 0 (wq_completion)ipv6_addrconf (work_completion)(&(&ifa->dad_work)->work) rtnl_mutex rcu_read_lock rcu_read_lock rcu_read_lock_bh &hsr->seqnr_lock rcu_read_lock pool_lock#2 irq_context: 0 (wq_completion)ipv6_addrconf (work_completion)(&(&ifa->dad_work)->work) rtnl_mutex rcu_read_lock rcu_read_lock rcu_read_lock_bh &hsr->seqnr_lock rcu_read_lock rcu_read_lock_bh rcu_read_lock tk_core.seq.seqcount irq_context: 0 (wq_completion)ipv6_addrconf (work_completion)(&(&ifa->dad_work)->work) rtnl_mutex rcu_read_lock rcu_read_lock rcu_read_lock_bh &hsr->seqnr_lock rcu_read_lock rcu_read_lock_bh rcu_read_lock &list->lock#5 irq_context: 0 (wq_completion)ipv6_addrconf (work_completion)(&(&ifa->dad_work)->work) rtnl_mutex rcu_read_lock rcu_read_lock rcu_read_lock_bh &hsr->seqnr_lock &obj_hash[i].lock irq_context: 0 (wq_completion)ipv6_addrconf (work_completion)(&(&ifa->dad_work)->work) rtnl_mutex rcu_read_lock rcu_read_lock rcu_read_lock_bh &hsr->seqnr_lock pool_lock#2 irq_context: 0 rtnl_mutex dev_addr_sem team->team_lock_key#2 irq_context: 0 (wq_completion)mld (work_completion)(&(&idev->mc_ifc_work)->work) &idev->mc_lock rcu_read_lock rcu_read_lock rcu_read_lock_bh &hsr->seqnr_lock rcu_read_lock &n->list_lock irq_context: 0 (wq_completion)mld (work_completion)(&(&idev->mc_ifc_work)->work) &idev->mc_lock rcu_read_lock rcu_read_lock rcu_read_lock_bh &hsr->seqnr_lock rcu_read_lock &n->list_lock &c->lock irq_context: 0 (wq_completion)mld (work_completion)(&(&idev->mc_ifc_work)->work) &idev->mc_lock rcu_read_lock rcu_read_lock rcu_read_lock_bh &hsr->seqnr_lock quarantine_lock irq_context: 0 (wq_completion)mld (work_completion)(&(&idev->mc_ifc_work)->work) &idev->mc_lock remove_cache_srcu &c->lock irq_context: 0 (wq_completion)mld (work_completion)(&(&idev->mc_ifc_work)->work) &idev->mc_lock remove_cache_srcu &n->list_lock irq_context: 0 (wq_completion)mld (work_completion)(&(&idev->mc_ifc_work)->work) &idev->mc_lock remove_cache_srcu rcu_read_lock pool_lock#2 irq_context: 0 (wq_completion)mld (work_completion)(&(&idev->mc_ifc_work)->work) &idev->mc_lock remove_cache_srcu &obj_hash[i].lock irq_context: 0 (wq_completion)ipv6_addrconf (work_completion)(&(&ifa->dad_work)->work) &p->pi_lock &rq->__lock &cfs_rq->removed.lock irq_context: 0 (wq_completion)events (work_completion)(&(&nsim_dev->trap_data->trap_report_dw)->work) &devlink->lock_key#3 &nsim_trap_data->trap_lock &pcp->lock &zone->lock &____s->seqcount irq_context: 0 (wq_completion)ipv6_addrconf (work_completion)(&(&ifa->dad_work)->work) rtnl_mutex &idev->mc_lock rcu_read_lock rcu_read_lock rcu_read_lock_bh dev->qdisc_tx_busylock ?: &qdisc_tx_busylock rcu_read_lock &ndev->lock irq_context: 0 (wq_completion)ipv6_addrconf (work_completion)(&(&ifa->dad_work)->work) rtnl_mutex &idev->mc_lock rcu_read_lock rcu_read_lock rcu_read_lock_bh dev->qdisc_tx_busylock ?: &qdisc_tx_busylock &obj_hash[i].lock irq_context: 0 (wq_completion)ipv6_addrconf (work_completion)(&(&ifa->dad_work)->work) rtnl_mutex &idev->mc_lock rcu_read_lock rcu_read_lock rcu_read_lock_bh dev->qdisc_tx_busylock ?: &qdisc_tx_busylock pool_lock#2 irq_context: 0 rtnl_mutex (inetaddr_chain).rwsem &pcp->lock &zone->lock irq_context: 0 rtnl_mutex (inetaddr_chain).rwsem &pcp->lock &zone->lock &____s->seqcount irq_context: 0 rtnl_mutex (inetaddr_chain).rwsem quarantine_lock irq_context: 0 (wq_completion)ipv6_addrconf (work_completion)(&(&ifa->dad_work)->work) rtnl_mutex rcu_read_lock rcu_read_lock rcu_node_0 irq_context: 0 (wq_completion)ipv6_addrconf (work_completion)(&(&ifa->dad_work)->work) rtnl_mutex rcu_read_lock rcu_read_lock &rq->__lock irq_context: 0 (wq_completion)ipv6_addrconf (work_completion)(&(&ifa->dad_work)->work) rtnl_mutex rcu_read_lock rcu_node_0 irq_context: 0 (wq_completion)ipv6_addrconf (work_completion)(&(&ifa->dad_work)->work) rtnl_mutex rcu_read_lock &rcu_state.gp_wq irq_context: 0 (wq_completion)ipv6_addrconf (work_completion)(&(&ifa->dad_work)->work) rtnl_mutex rcu_read_lock &rcu_state.gp_wq &p->pi_lock irq_context: 0 (wq_completion)ipv6_addrconf (work_completion)(&(&ifa->dad_work)->work) rtnl_mutex rcu_read_lock &rcu_state.gp_wq &p->pi_lock &rq->__lock irq_context: 0 (wq_completion)ipv6_addrconf (work_completion)(&(&ifa->dad_work)->work) rtnl_mutex rcu_read_lock &rcu_state.gp_wq &p->pi_lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)ipv6_addrconf (work_completion)(&(&ifa->dad_work)->work) rtnl_mutex &idev->mc_lock rcu_read_lock rcu_read_lock rcu_read_lock_bh dev->qdisc_tx_busylock ?: &qdisc_tx_busylock _xmit_ETHER#2 rcu_read_lock &ndev->lock irq_context: 0 (wq_completion)ipv6_addrconf (work_completion)(&(&ifa->dad_work)->work) rtnl_mutex &idev->mc_lock rcu_read_lock rcu_read_lock rcu_read_lock_bh dev->qdisc_tx_busylock ?: &qdisc_tx_busylock _xmit_ETHER#2 &obj_hash[i].lock irq_context: 0 (wq_completion)ipv6_addrconf (work_completion)(&(&ifa->dad_work)->work) rtnl_mutex &idev->mc_lock rcu_read_lock rcu_read_lock rcu_read_lock_bh dev->qdisc_tx_busylock ?: &qdisc_tx_busylock _xmit_ETHER#2 pool_lock#2 irq_context: 0 (wq_completion)events (work_completion)(&(&nsim_dev->trap_data->trap_report_dw)->work) &devlink->lock_key#6 &nsim_trap_data->trap_lock &pcp->lock &zone->lock &____s->seqcount irq_context: 0 rtnl_mutex dev_addr_sem team->team_lock_key#5 irq_context: 0 (wq_completion)events (linkwatch_work).work rtnl_mutex &br->lock rcu_read_lock &pool->lock &p->pi_lock &rq->__lock irq_context: 0 (wq_completion)events (linkwatch_work).work rtnl_mutex &br->lock rcu_read_lock &pool->lock &p->pi_lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)events (linkwatch_work).work rtnl_mutex rcu_read_lock rcu_read_lock &pool->lock/1 &p->pi_lock &cfs_rq->removed.lock irq_context: 0 (wq_completion)mld (work_completion)(&(&idev->mc_ifc_work)->work) &idev->mc_lock batched_entropy_u8.lock irq_context: 0 (wq_completion)mld (work_completion)(&(&idev->mc_ifc_work)->work) &idev->mc_lock kfence_freelist_lock irq_context: softirq rcu_read_lock rcu_read_lock rcu_read_lock &meta->lock irq_context: softirq rcu_read_lock rcu_read_lock rcu_read_lock kfence_freelist_lock irq_context: 0 (wq_completion)events (linkwatch_work).work &p->pi_lock &rq->__lock &cfs_rq->removed.lock irq_context: 0 rtnl_mutex &vlan_netdev_addr_lock_key/1 irq_context: 0 rtnl_mutex &vlan_netdev_addr_lock_key/1 _xmit_ETHER irq_context: 0 rtnl_mutex &vlan_netdev_addr_lock_key/1 &c->lock irq_context: 0 rtnl_mutex &vlan_netdev_addr_lock_key/1 &____s->seqcount irq_context: 0 rtnl_mutex &vlan_netdev_addr_lock_key/1 pool_lock#2 irq_context: 0 (wq_completion)ipv6_addrconf (work_completion)(&(&ifa->dad_work)->work) rtnl_mutex &idev->mc_lock rcu_read_lock rcu_read_lock rcu_read_lock_bh rcu_read_lock rcu_read_lock_bh rcu_read_lock tk_core.seq.seqcount irq_context: 0 (wq_completion)ipv6_addrconf (work_completion)(&(&ifa->dad_work)->work) rtnl_mutex &idev->mc_lock rcu_read_lock rcu_read_lock rcu_read_lock_bh rcu_read_lock rcu_read_lock_bh rcu_read_lock &list->lock#5 irq_context: 0 (wq_completion)ipv6_addrconf (work_completion)(&(&ifa->dad_work)->work) rtnl_mutex &idev->mc_lock rcu_read_lock rcu_read_lock rcu_read_lock_bh &obj_hash[i].lock pool_lock irq_context: 0 (wq_completion)ipv6_addrconf (work_completion)(&(&ifa->dad_work)->work) rtnl_mutex &idev->mc_lock rcu_read_lock rcu_read_lock rcu_read_lock_bh rcu_read_lock tk_core.seq.seqcount irq_context: 0 (wq_completion)ipv6_addrconf (work_completion)(&(&ifa->dad_work)->work) rtnl_mutex &idev->mc_lock rcu_read_lock rcu_read_lock rcu_read_lock_bh rcu_read_lock &list->lock#5 irq_context: softirq rcu_read_lock rcu_read_lock &ndev->lock irq_context: softirq rcu_read_lock &br->multicast_lock &n->list_lock irq_context: softirq rcu_read_lock &br->multicast_lock &n->list_lock &c->lock irq_context: 0 (wq_completion)events (linkwatch_work).work rtnl_mutex team->team_lock_key#6 &c->lock irq_context: 0 (wq_completion)events (linkwatch_work).work rtnl_mutex team->team_lock_key#6 &____s->seqcount irq_context: 0 (wq_completion)events (linkwatch_work).work rtnl_mutex team->team_lock_key#6 lweventlist_lock irq_context: 0 (wq_completion)events (linkwatch_work).work rtnl_mutex team->team_lock_key#6 lweventlist_lock &dir->lock#2 irq_context: 0 (wq_completion)events (linkwatch_work).work rtnl_mutex team->team_lock_key#6 rcu_read_lock &pool->lock irq_context: 0 (wq_completion)events (linkwatch_work).work rtnl_mutex team->team_lock_key#6 rcu_read_lock &pool->lock &obj_hash[i].lock irq_context: 0 (wq_completion)events (linkwatch_work).work &p->pi_lock &cfs_rq->removed.lock irq_context: 0 (wq_completion)mld (work_completion)(&(&idev->mc_ifc_work)->work) &idev->mc_lock rcu_read_lock batched_entropy_u8.lock irq_context: 0 (wq_completion)mld (work_completion)(&(&idev->mc_ifc_work)->work) &idev->mc_lock rcu_read_lock kfence_freelist_lock irq_context: softirq (&hsr->prune_timer) irq_context: softirq (&hsr->prune_timer) &hsr->list_lock irq_context: softirq (&hsr->prune_timer) &obj_hash[i].lock irq_context: softirq (&hsr->prune_timer) &base->lock irq_context: softirq (&hsr->prune_timer) &base->lock &obj_hash[i].lock irq_context: softirq (&wq_watchdog_timer) &obj_hash[i].lock irq_context: softirq (&wq_watchdog_timer) &base->lock irq_context: softirq (&wq_watchdog_timer) &base->lock &obj_hash[i].lock irq_context: softirq rcu_read_lock &br->hash_lock &n->list_lock irq_context: softirq rcu_read_lock &br->hash_lock &n->list_lock &c->lock irq_context: 0 (wq_completion)ipv6_addrconf (work_completion)(&(&ifa->dad_work)->work) rtnl_mutex rcu_read_lock rcu_read_lock rcu_read_lock_bh rcu_read_lock &c->lock irq_context: 0 (wq_completion)ipv6_addrconf (work_completion)(&(&ifa->dad_work)->work) rtnl_mutex &idev->mc_lock &vlan_netdev_addr_lock_key/1 irq_context: 0 (wq_completion)ipv6_addrconf (work_completion)(&(&ifa->dad_work)->work) rtnl_mutex &idev->mc_lock &vlan_netdev_addr_lock_key/1 _xmit_ETHER irq_context: 0 (wq_completion)ipv6_addrconf (work_completion)(&(&ifa->dad_work)->work) rtnl_mutex &idev->mc_lock &vlan_netdev_addr_lock_key/1 _xmit_ETHER pool_lock#2 irq_context: 0 (wq_completion)events (linkwatch_work).work rtnl_mutex team->team_lock_key#5 quarantine_lock irq_context: 0 (wq_completion)events (linkwatch_work).work rtnl_mutex team->team_lock_key#5 remove_cache_srcu irq_context: 0 (wq_completion)events (linkwatch_work).work rtnl_mutex team->team_lock_key#5 remove_cache_srcu quarantine_lock irq_context: 0 (wq_completion)events (linkwatch_work).work rtnl_mutex team->team_lock_key#5 remove_cache_srcu &c->lock irq_context: 0 (wq_completion)events (linkwatch_work).work rtnl_mutex team->team_lock_key#5 remove_cache_srcu &n->list_lock irq_context: 0 (wq_completion)events (linkwatch_work).work rtnl_mutex team->team_lock_key#5 remove_cache_srcu rcu_read_lock pool_lock#2 irq_context: 0 (wq_completion)events (linkwatch_work).work rtnl_mutex team->team_lock_key#5 remove_cache_srcu &obj_hash[i].lock irq_context: 0 (wq_completion)events (linkwatch_work).work rtnl_mutex team->team_lock_key#5 lweventlist_lock irq_context: 0 (wq_completion)events (linkwatch_work).work rtnl_mutex team->team_lock_key#5 lweventlist_lock &dir->lock#2 irq_context: 0 (wq_completion)events (linkwatch_work).work rtnl_mutex team->team_lock_key#5 rcu_read_lock &pool->lock irq_context: 0 (wq_completion)events (linkwatch_work).work rtnl_mutex team->team_lock_key#5 rcu_read_lock &pool->lock &obj_hash[i].lock irq_context: 0 (wq_completion)events (linkwatch_work).work rtnl_mutex fs_reclaim &rq->__lock irq_context: 0 &xa->xa_lock#9 &c->lock irq_context: 0 &xa->xa_lock#9 &pcp->lock &zone->lock irq_context: 0 &xa->xa_lock#9 &pcp->lock &zone->lock &____s->seqcount irq_context: 0 &xa->xa_lock#9 &____s->seqcount irq_context: 0 &mm->mmap_lock &mapping->i_mmap_rwsem fill_pool_map-wait-type-override &rq->__lock irq_context: 0 (wq_completion)mld (work_completion)(&(&idev->mc_ifc_work)->work) &idev->mc_lock rcu_read_lock rcu_read_lock &pcp->lock &zone->lock &____s->seqcount irq_context: 0 (wq_completion)mld (work_completion)(&(&idev->mc_ifc_work)->work) &idev->mc_lock rcu_read_lock rcu_read_lock rcu_read_lock_bh rcu_read_lock_bh rcu_read_lock tk_core.seq.seqcount irq_context: 0 (wq_completion)mld (work_completion)(&(&idev->mc_ifc_work)->work) &idev->mc_lock rcu_read_lock rcu_read_lock rcu_read_lock_bh rcu_read_lock_bh rcu_read_lock &list->lock#5 irq_context: 0 rtnl_mutex dev_addr_sem _xmit_ETHER irq_context: 0 rtnl_mutex dev_addr_sem _xmit_ETHER pool_lock#2 irq_context: 0 rtnl_mutex dev_addr_sem _xmit_ETHER (console_sem).lock irq_context: 0 rtnl_mutex dev_addr_sem _xmit_ETHER console_lock console_srcu console_owner_lock irq_context: 0 rtnl_mutex dev_addr_sem _xmit_ETHER console_lock console_srcu console_owner irq_context: 0 rtnl_mutex dev_addr_sem _xmit_ETHER console_lock console_srcu console_owner &port_lock_key irq_context: softirq (&in_dev->mr_ifc_timer) rcu_read_lock rcu_read_lock_bh rcu_read_lock rcu_read_lock_bh rcu_read_lock tk_core.seq.seqcount irq_context: softirq (&in_dev->mr_ifc_timer) rcu_read_lock rcu_read_lock_bh rcu_read_lock rcu_read_lock_bh rcu_read_lock &list->lock#5 irq_context: softirq (&in_dev->mr_ifc_timer) rcu_read_lock rcu_read_lock_bh rcu_read_lock rcu_read_lock_bh &obj_hash[i].lock irq_context: softirq (&in_dev->mr_ifc_timer) rcu_read_lock rcu_read_lock_bh rcu_read_lock rcu_read_lock_bh pool_lock#2 irq_context: 0 rtnl_mutex dev_addr_sem _xmit_ETHER console_lock console_srcu console_owner console_owner_lock irq_context: 0 rtnl_mutex dev_addr_sem &tbl->lock krc.lock &obj_hash[i].lock irq_context: 0 rtnl_mutex dev_addr_sem &tbl->lock krc.lock &base->lock irq_context: 0 rtnl_mutex dev_addr_sem &tbl->lock krc.lock &base->lock &obj_hash[i].lock irq_context: 0 (wq_completion)ipv6_addrconf (work_completion)(&(&ifa->dad_work)->work) rtnl_mutex &idev->mc_lock &vlan_netdev_addr_lock_key/1 &c->lock irq_context: 0 (wq_completion)ipv6_addrconf (work_completion)(&(&ifa->dad_work)->work) rtnl_mutex &idev->mc_lock &vlan_netdev_addr_lock_key/1 &____s->seqcount irq_context: 0 (wq_completion)ipv6_addrconf (work_completion)(&(&ifa->dad_work)->work) rtnl_mutex rcu_read_lock rcu_read_lock rcu_read_lock_bh &hsr->seqnr_lock rcu_read_lock &c->lock irq_context: 0 (wq_completion)ipv6_addrconf (work_completion)(&(&ifa->dad_work)->work) rtnl_mutex rcu_read_lock rcu_read_lock rcu_read_lock_bh fill_pool_map-wait-type-override &c->lock irq_context: 0 (wq_completion)events (linkwatch_work).work rtnl_mutex team->team_lock_key#5 &pcp->lock &zone->lock irq_context: 0 (wq_completion)events (linkwatch_work).work rtnl_mutex team->team_lock_key#5 &pcp->lock &zone->lock &____s->seqcount irq_context: 0 (wq_completion)events (linkwatch_work).work rtnl_mutex batched_entropy_u8.lock irq_context: 0 (wq_completion)events (linkwatch_work).work rtnl_mutex kfence_freelist_lock irq_context: 0 (wq_completion)events (linkwatch_work).work rtnl_mutex &meta->lock irq_context: 0 (wq_completion)events (linkwatch_work).work rtnl_mutex team->team_lock_key#2 &____s->seqcount irq_context: 0 (wq_completion)events (linkwatch_work).work rtnl_mutex team->team_lock_key#2 lweventlist_lock irq_context: 0 (wq_completion)events (linkwatch_work).work rtnl_mutex team->team_lock_key#2 lweventlist_lock &dir->lock#2 irq_context: 0 (wq_completion)events (linkwatch_work).work rtnl_mutex team->team_lock_key#2 &base->lock irq_context: 0 (wq_completion)events (linkwatch_work).work rtnl_mutex team->team_lock_key#2 &base->lock &obj_hash[i].lock irq_context: 0 (wq_completion)events (linkwatch_work).work rtnl_mutex team->team_lock_key#2 rcu_read_lock &pool->lock irq_context: 0 (wq_completion)events (linkwatch_work).work rtnl_mutex team->team_lock_key#2 rcu_read_lock &pool->lock &obj_hash[i].lock irq_context: 0 rtnl_mutex &batadv_netdev_addr_lock_key/1 &c->lock irq_context: 0 rtnl_mutex &batadv_netdev_addr_lock_key/1 &____s->seqcount irq_context: 0 (wq_completion)mld (work_completion)(&(&idev->mc_ifc_work)->work) &idev->mc_lock rcu_read_lock rcu_read_lock fill_pool_map-wait-type-override pool_lock#2 irq_context: 0 (wq_completion)mld (work_completion)(&(&idev->mc_ifc_work)->work) &idev->mc_lock rcu_read_lock rcu_read_lock fill_pool_map-wait-type-override pool_lock irq_context: 0 (wq_completion)mld (work_completion)(&(&idev->mc_dad_work)->work) &idev->mc_lock rcu_read_lock rcu_read_lock rcu_read_lock_bh dev->qdisc_tx_busylock ?: &qdisc_tx_busylock rcu_read_lock &ndev->lock irq_context: 0 (wq_completion)mld (work_completion)(&(&idev->mc_dad_work)->work) &idev->mc_lock rcu_read_lock rcu_read_lock rcu_read_lock_bh dev->qdisc_tx_busylock ?: &qdisc_tx_busylock &obj_hash[i].lock irq_context: 0 (wq_completion)mld (work_completion)(&(&idev->mc_dad_work)->work) &idev->mc_lock rcu_read_lock rcu_read_lock rcu_read_lock_bh dev->qdisc_tx_busylock ?: &qdisc_tx_busylock pool_lock#2 irq_context: 0 rtnl_mutex dev_addr_sem &net->ipv6.fib6_gc_lock fill_pool_map-wait-type-override pool_lock#2 irq_context: 0 rtnl_mutex dev_addr_sem &net->ipv6.fib6_gc_lock fill_pool_map-wait-type-override pool_lock irq_context: 0 &mm->mmap_lock remove_cache_srcu rcu_read_lock rcu_node_0 irq_context: 0 &mm->mmap_lock remove_cache_srcu rcu_read_lock &rq->__lock irq_context: 0 (wq_completion)mld (work_completion)(&(&idev->mc_dad_work)->work) &idev->mc_lock rcu_read_lock rcu_read_lock rcu_read_lock_bh rcu_read_lock tk_core.seq.seqcount irq_context: 0 (wq_completion)mld (work_completion)(&(&idev->mc_dad_work)->work) &idev->mc_lock rcu_read_lock rcu_read_lock rcu_read_lock_bh rcu_read_lock &list->lock#5 irq_context: 0 (wq_completion)ipv6_addrconf (work_completion)(&(&ifa->dad_work)->work) rtnl_mutex rcu_read_lock rcu_read_lock rcu_read_lock_bh rcu_read_lock &n->list_lock irq_context: 0 (wq_completion)ipv6_addrconf (work_completion)(&(&ifa->dad_work)->work) rtnl_mutex rcu_read_lock rcu_read_lock rcu_read_lock_bh rcu_read_lock &n->list_lock &c->lock irq_context: 0 &mm->mmap_lock remove_cache_srcu rcu_read_lock &rcu_state.gp_wq irq_context: 0 &mm->mmap_lock remove_cache_srcu rcu_read_lock &rcu_state.gp_wq &p->pi_lock irq_context: 0 &mm->mmap_lock remove_cache_srcu rcu_read_lock &rcu_state.gp_wq &p->pi_lock &rq->__lock irq_context: 0 &mm->mmap_lock remove_cache_srcu rcu_read_lock &rcu_state.gp_wq &p->pi_lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)ipv6_addrconf (work_completion)(&(&ifa->dad_work)->work) rtnl_mutex &ndev->lock &ifa->lock fill_pool_map-wait-type-override &c->lock irq_context: 0 (wq_completion)ipv6_addrconf (work_completion)(&(&ifa->dad_work)->work) rtnl_mutex &ndev->lock &ifa->lock fill_pool_map-wait-type-override &____s->seqcount irq_context: 0 rtnl_mutex &macvlan_netdev_addr_lock_key/1 irq_context: 0 rtnl_mutex _xmit_ETHER (console_sem).lock irq_context: 0 rtnl_mutex _xmit_ETHER console_lock console_srcu console_owner_lock irq_context: 0 rtnl_mutex _xmit_ETHER console_lock console_srcu console_owner irq_context: 0 rtnl_mutex _xmit_ETHER console_lock console_srcu console_owner &port_lock_key irq_context: 0 rtnl_mutex _xmit_ETHER console_lock console_srcu console_owner console_owner_lock irq_context: 0 rtnl_mutex &macvlan_netdev_addr_lock_key/1 _xmit_ETHER irq_context: softirq &(&idev->mc_ifc_work)->timer rcu_read_lock &pool->lock fill_pool_map-wait-type-override pool_lock irq_context: softirq &(&bat_priv->mcast.work)->timer rcu_read_lock &pool->lock/1 &p->pi_lock &rq->__lock &cfs_rq->removed.lock irq_context: softirq rcu_read_lock &list->lock#16 irq_context: softirq rcu_read_lock rcu_read_lock &pool->lock/1 irq_context: softirq rcu_read_lock rcu_read_lock &pool->lock/1 &obj_hash[i].lock irq_context: softirq rcu_read_lock rcu_read_lock &pool->lock/1 pool_lock#2 irq_context: softirq rcu_read_lock rcu_read_lock &pool->lock/1 &p->pi_lock irq_context: 0 (wq_completion)events_unbound (work_completion)(&port->bc_work) irq_context: 0 (wq_completion)events_unbound (work_completion)(&port->bc_work) &list->lock#16 irq_context: 0 (wq_completion)events_unbound (work_completion)(&port->bc_work) &obj_hash[i].lock irq_context: 0 (wq_completion)events_unbound (work_completion)(&port->bc_work) pool_lock#2 irq_context: 0 (wq_completion)ipv6_addrconf (work_completion)(&(&ifa->dad_work)->work) rtnl_mutex &idev->mc_lock rcu_read_lock rcu_read_lock rcu_read_lock_bh rcu_read_lock &br->multicast_lock irq_context: 0 (wq_completion)ipv6_addrconf (work_completion)(&(&ifa->dad_work)->work) rtnl_mutex &idev->mc_lock rcu_read_lock rcu_read_lock rcu_read_lock_bh rcu_read_lock &br->multicast_lock &obj_hash[i].lock irq_context: 0 (wq_completion)ipv6_addrconf (work_completion)(&(&ifa->dad_work)->work) rtnl_mutex &idev->mc_lock rcu_read_lock rcu_read_lock rcu_read_lock_bh rcu_read_lock &br->multicast_lock &base->lock irq_context: 0 (wq_completion)ipv6_addrconf (work_completion)(&(&ifa->dad_work)->work) rtnl_mutex &idev->mc_lock rcu_read_lock rcu_read_lock rcu_read_lock_bh rcu_read_lock pool_lock#2 irq_context: 0 (wq_completion)ipv6_addrconf (work_completion)(&(&ifa->dad_work)->work) rtnl_mutex &idev->mc_lock rcu_read_lock rcu_read_lock rcu_read_lock_bh rcu_read_lock rcu_read_lock_bh &obj_hash[i].lock irq_context: 0 (wq_completion)ipv6_addrconf (work_completion)(&(&ifa->dad_work)->work) rtnl_mutex &idev->mc_lock rcu_read_lock rcu_read_lock rcu_read_lock_bh rcu_read_lock rcu_read_lock_bh pool_lock#2 irq_context: 0 (wq_completion)ipv6_addrconf (work_completion)(&(&ifa->dad_work)->work) rtnl_mutex &idev->mc_lock &macvlan_netdev_addr_lock_key/1 irq_context: 0 (wq_completion)ipv6_addrconf (work_completion)(&(&ifa->dad_work)->work) rtnl_mutex &idev->mc_lock &macvlan_netdev_addr_lock_key/1 _xmit_ETHER irq_context: 0 (wq_completion)ipv6_addrconf (work_completion)(&(&ifa->dad_work)->work) rtnl_mutex &idev->mc_lock &macvlan_netdev_addr_lock_key/1 _xmit_ETHER pool_lock#2 irq_context: 0 rtnl_mutex (inetaddr_chain).rwsem &ipvlan->addrs_lock irq_context: 0 rtnl_mutex (inetaddr_chain).rwsem &ipvlan->addrs_lock pool_lock#2 irq_context: 0 rtnl_mutex rcu_read_lock &ipvlan->addrs_lock irq_context: 0 rtnl_mutex rcu_read_lock &ipvlan->addrs_lock pool_lock#2 irq_context: 0 (wq_completion)mld (work_completion)(&(&idev->mc_ifc_work)->work) &idev->mc_lock rcu_read_lock rcu_read_lock rcu_read_lock_bh rcu_read_lock_bh &obj_hash[i].lock irq_context: 0 (wq_completion)mld (work_completion)(&(&idev->mc_ifc_work)->work) &idev->mc_lock rcu_read_lock rcu_read_lock rcu_read_lock_bh rcu_read_lock_bh &obj_hash[i].lock pool_lock irq_context: 0 (wq_completion)mld (work_completion)(&(&idev->mc_ifc_work)->work) &idev->mc_lock rcu_read_lock rcu_read_lock rcu_read_lock_bh rcu_read_lock_bh pool_lock#2 irq_context: 0 (wq_completion)mld (work_completion)(&(&idev->mc_dad_work)->work) &idev->mc_lock rcu_read_lock rcu_read_lock rcu_read_lock_bh rcu_read_lock rcu_read_lock_bh rcu_read_lock tk_core.seq.seqcount irq_context: 0 (wq_completion)mld (work_completion)(&(&idev->mc_dad_work)->work) &idev->mc_lock rcu_read_lock rcu_read_lock rcu_read_lock_bh rcu_read_lock rcu_read_lock_bh rcu_read_lock &list->lock#5 irq_context: 0 (wq_completion)ipv6_addrconf (work_completion)(&(&ifa->dad_work)->work) rtnl_mutex &idev->mc_lock &dev_addr_list_lock_key#3/1 irq_context: 0 &mm->mmap_lock fill_pool_map-wait-type-override rcu_read_lock pool_lock#2 irq_context: 0 &mm->mmap_lock fill_pool_map-wait-type-override &obj_hash[i].lock irq_context: 0 rtnl_mutex &macsec_netdev_addr_lock_key/1 irq_context: 0 rtnl_mutex &macsec_netdev_addr_lock_key/1 _xmit_ETHER irq_context: 0 (wq_completion)ipv6_addrconf (work_completion)(&(&ifa->dad_work)->work) rtnl_mutex &idev->mc_lock &macsec_netdev_addr_lock_key/1 irq_context: 0 (wq_completion)ipv6_addrconf (work_completion)(&(&ifa->dad_work)->work) rtnl_mutex &idev->mc_lock &macsec_netdev_addr_lock_key/1 _xmit_ETHER irq_context: 0 (wq_completion)ipv6_addrconf (work_completion)(&(&ifa->dad_work)->work) rtnl_mutex &idev->mc_lock &macsec_netdev_addr_lock_key/1 _xmit_ETHER &c->lock irq_context: 0 (wq_completion)ipv6_addrconf (work_completion)(&(&ifa->dad_work)->work) rtnl_mutex &idev->mc_lock &macsec_netdev_addr_lock_key/1 _xmit_ETHER &____s->seqcount irq_context: 0 (wq_completion)ipv6_addrconf (work_completion)(&(&ifa->dad_work)->work) rtnl_mutex &idev->mc_lock &macsec_netdev_addr_lock_key/1 _xmit_ETHER pool_lock#2 irq_context: 0 rtnl_mutex dev_addr_sem quarantine_lock irq_context: 0 rtnl_mutex dev_addr_sem remove_cache_srcu irq_context: 0 rtnl_mutex dev_addr_sem remove_cache_srcu quarantine_lock irq_context: 0 rtnl_mutex dev_addr_sem remove_cache_srcu &c->lock irq_context: 0 rtnl_mutex dev_addr_sem remove_cache_srcu &n->list_lock irq_context: 0 rtnl_mutex dev_addr_sem remove_cache_srcu rcu_read_lock pool_lock#2 irq_context: 0 rtnl_mutex dev_addr_sem remove_cache_srcu &obj_hash[i].lock irq_context: 0 rtnl_mutex dev_addr_sem remove_cache_srcu rcu_read_lock rcu_node_0 irq_context: 0 rtnl_mutex dev_addr_sem remove_cache_srcu rcu_read_lock &rq->__lock irq_context: 0 rtnl_mutex dev_addr_sem remove_cache_srcu rcu_read_lock pgd_lock irq_context: 0 rtnl_mutex dev_addr_sem remove_cache_srcu rcu_read_lock &obj_hash[i].lock irq_context: 0 rtnl_mutex dev_addr_sem remove_cache_srcu rcu_read_lock key irq_context: 0 rtnl_mutex dev_addr_sem remove_cache_srcu rcu_read_lock pcpu_lock irq_context: 0 rtnl_mutex dev_addr_sem remove_cache_srcu rcu_read_lock percpu_counters_lock irq_context: 0 (wq_completion)events (work_completion)(&(&nsim_dev->trap_data->trap_report_dw)->work) &devlink->lock_key#2 &nsim_trap_data->trap_lock batched_entropy_u8.lock irq_context: 0 (wq_completion)events (work_completion)(&(&nsim_dev->trap_data->trap_report_dw)->work) &devlink->lock_key#2 &nsim_trap_data->trap_lock kfence_freelist_lock irq_context: 0 (wq_completion)events (work_completion)(&(&nsim_dev->trap_data->trap_report_dw)->work) &devlink->lock_key#2 &nsim_trap_data->trap_lock &meta->lock irq_context: 0 (wq_completion)mld (work_completion)(&(&idev->mc_ifc_work)->work) &idev->mc_lock rcu_read_lock rcu_read_lock rcu_read_lock_bh dev->qdisc_tx_busylock ?: &qdisc_tx_busylock#2 rcu_read_lock_bh &obj_hash[i].lock irq_context: 0 (wq_completion)mld (work_completion)(&(&idev->mc_ifc_work)->work) &idev->mc_lock rcu_read_lock rcu_read_lock rcu_read_lock_bh dev->qdisc_tx_busylock ?: &qdisc_tx_busylock#2 rcu_read_lock_bh pool_lock#2 irq_context: 0 (wq_completion)mld (work_completion)(&(&idev->mc_ifc_work)->work) &idev->mc_lock rcu_read_lock rcu_read_lock rcu_read_lock_bh dev->qdisc_tx_busylock ?: &qdisc_tx_busylock#2 rcu_read_lock_bh rcu_read_lock tk_core.seq.seqcount irq_context: 0 (wq_completion)mld (work_completion)(&(&idev->mc_ifc_work)->work) &idev->mc_lock rcu_read_lock rcu_read_lock rcu_read_lock_bh dev->qdisc_tx_busylock ?: &qdisc_tx_busylock#2 rcu_read_lock_bh rcu_read_lock &list->lock#5 irq_context: 0 (wq_completion)ipv6_addrconf (work_completion)(&(&ifa->dad_work)->work) rtnl_mutex &cfs_rq->removed.lock irq_context: 0 (wq_completion)ipv6_addrconf (work_completion)(&(&ifa->dad_work)->work) rtnl_mutex &idev->mc_lock rcu_read_lock &____s->seqcount irq_context: 0 (wq_completion)ipv6_addrconf (work_completion)(&(&ifa->dad_work)->work) rtnl_mutex &idev->mc_lock &pcp->lock &zone->lock &____s->seqcount irq_context: 0 rtnl_mutex dev_addr_sem &hard_iface->bat_iv.ogm_buff_mutex irq_context: 0 rtnl_mutex key#20 irq_context: 0 (wq_completion)mld (work_completion)(&(&idev->mc_dad_work)->work) &idev->mc_lock rcu_read_lock rcu_read_lock rcu_read_lock_bh dev->qdisc_tx_busylock ?: &qdisc_tx_busylock _xmit_ETHER#2 rcu_read_lock &ndev->lock irq_context: 0 (wq_completion)mld (work_completion)(&(&idev->mc_dad_work)->work) &idev->mc_lock rcu_read_lock rcu_read_lock rcu_read_lock_bh dev->qdisc_tx_busylock ?: &qdisc_tx_busylock _xmit_ETHER#2 &obj_hash[i].lock irq_context: 0 (wq_completion)mld (work_completion)(&(&idev->mc_dad_work)->work) &idev->mc_lock rcu_read_lock rcu_read_lock rcu_read_lock_bh dev->qdisc_tx_busylock ?: &qdisc_tx_busylock _xmit_ETHER#2 pool_lock#2 irq_context: 0 rtnl_mutex &bat_priv->tt.commit_lock irq_context: 0 (wq_completion)bat_events (work_completion)(&(&bat_priv->nc.work)->work) rcu_read_lock &rcu_state.gp_wq &p->pi_lock &rq->__lock irq_context: 0 (wq_completion)bat_events (work_completion)(&(&bat_priv->nc.work)->work) rcu_read_lock &rcu_state.gp_wq &p->pi_lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: softirq (&hsr->announce_timer) rcu_read_lock &hsr->seqnr_lock rcu_read_lock &n->list_lock irq_context: softirq (&hsr->announce_timer) rcu_read_lock &hsr->seqnr_lock rcu_read_lock &n->list_lock &c->lock irq_context: 0 (wq_completion)events_unbound (work_completion)(&sub_info->work) &x->wait#17 &p->pi_lock &cfs_rq->removed.lock irq_context: 0 rtnl_mutex mmu_notifier_invalidate_range_start irq_context: 0 rtnl_mutex &sb->s_type->i_lock_key#8 irq_context: 0 rtnl_mutex &dir->lock irq_context: 0 rtnl_mutex k-sk_lock-AF_INET irq_context: 0 rtnl_mutex k-sk_lock-AF_INET k-slock-AF_INET irq_context: 0 rtnl_mutex k-sk_lock-AF_INET &table->hash[i].lock irq_context: 0 rtnl_mutex k-sk_lock-AF_INET &table->hash[i].lock k-clock-AF_INET irq_context: 0 rtnl_mutex k-sk_lock-AF_INET &table->hash[i].lock &table->hash2[i].lock irq_context: 0 rtnl_mutex k-slock-AF_INET irq_context: 0 rtnl_mutex rcu_read_lock (console_sem).lock irq_context: 0 rtnl_mutex rcu_read_lock console_lock console_srcu console_owner_lock irq_context: 0 rtnl_mutex rcu_read_lock console_lock console_srcu console_owner irq_context: 0 rtnl_mutex rcu_read_lock console_lock console_srcu console_owner &port_lock_key irq_context: 0 rtnl_mutex rcu_read_lock console_lock console_srcu console_owner console_owner_lock irq_context: 0 (wq_completion)events (work_completion)(&data->fib_event_work) &data->fib_lock &cfs_rq->removed.lock irq_context: 0 (wq_completion)events (work_completion)(&data->fib_event_work) &data->fib_lock fill_pool_map-wait-type-override pool_lock#2 irq_context: 0 (wq_completion)events (work_completion)(&data->fib_event_work) &data->fib_lock fill_pool_map-wait-type-override pool_lock irq_context: softirq &(&bat_priv->tt.work)->timer irq_context: softirq &(&bat_priv->tt.work)->timer rcu_read_lock &pool->lock/1 irq_context: softirq &(&bat_priv->tt.work)->timer rcu_read_lock &pool->lock/1 &obj_hash[i].lock irq_context: softirq &(&bat_priv->tt.work)->timer rcu_read_lock &pool->lock/1 &p->pi_lock irq_context: 0 rtnl_mutex k-sk_lock-AF_INET6 irq_context: 0 rtnl_mutex k-sk_lock-AF_INET6 k-slock-AF_INET6 irq_context: 0 rtnl_mutex k-slock-AF_INET6 irq_context: 0 rtnl_mutex k-sk_lock-AF_INET6 &table->hash[i].lock irq_context: 0 rtnl_mutex k-sk_lock-AF_INET6 &table->hash[i].lock k-clock-AF_INET6 irq_context: 0 rtnl_mutex k-sk_lock-AF_INET6 &table->hash[i].lock &table->hash2[i].lock irq_context: 0 (wq_completion)bat_events (work_completion)(&(&bat_priv->tt.work)->work) irq_context: 0 (wq_completion)bat_events (work_completion)(&(&bat_priv->tt.work)->work) key#16 irq_context: 0 (wq_completion)bat_events (work_completion)(&(&bat_priv->tt.work)->work) key#21 irq_context: 0 (wq_completion)bat_events (work_completion)(&(&bat_priv->tt.work)->work) &rq->__lock irq_context: 0 (wq_completion)events (work_completion)(&data->fib_event_work) &data->fib_lock pool_lock irq_context: 0 (wq_completion)bat_events (work_completion)(&(&bat_priv->tt.work)->work) &bat_priv->tt.req_list_lock irq_context: 0 (wq_completion)bat_events (work_completion)(&(&bat_priv->tt.work)->work) &bat_priv->tt.roam_list_lock irq_context: 0 (wq_completion)bat_events (work_completion)(&(&bat_priv->tt.work)->work) &obj_hash[i].lock irq_context: 0 (wq_completion)bat_events (work_completion)(&(&bat_priv->tt.work)->work) &base->lock irq_context: 0 (wq_completion)bat_events (work_completion)(&(&bat_priv->tt.work)->work) &base->lock &obj_hash[i].lock irq_context: 0 (wq_completion)mld (work_completion)(&(&idev->mc_ifc_work)->work) &idev->mc_lock rcu_read_lock rcu_read_lock rcu_read_lock_bh rcu_read_lock rcu_read_lock pool_lock#2 irq_context: 0 (wq_completion)mld (work_completion)(&(&idev->mc_ifc_work)->work) &idev->mc_lock rcu_read_lock rcu_read_lock rcu_read_lock_bh rcu_read_lock rcu_read_lock &dir->lock#2 irq_context: 0 (wq_completion)mld (work_completion)(&(&idev->mc_ifc_work)->work) &idev->mc_lock rcu_read_lock rcu_read_lock rcu_read_lock_bh rcu_read_lock rcu_read_lock &ul->lock irq_context: 0 (wq_completion)mld (work_completion)(&(&idev->mc_ifc_work)->work) &idev->mc_lock rcu_read_lock rcu_read_lock rcu_read_lock_bh rcu_read_lock crngs.lock irq_context: 0 (wq_completion)mld (work_completion)(&(&idev->mc_ifc_work)->work) &idev->mc_lock rcu_read_lock rcu_read_lock rcu_read_lock_bh rcu_read_lock batched_entropy_u32.lock irq_context: 0 (wq_completion)mld (work_completion)(&(&idev->mc_ifc_work)->work) &idev->mc_lock rcu_read_lock rcu_read_lock rcu_read_lock_bh rcu_read_lock rcu_read_lock &____s->seqcount#7 irq_context: 0 (wq_completion)mld (work_completion)(&(&idev->mc_ifc_work)->work) &idev->mc_lock rcu_read_lock rcu_read_lock rcu_read_lock_bh rcu_read_lock rcu_read_lock &nf_conntrack_locks[i] irq_context: 0 (wq_completion)mld (work_completion)(&(&idev->mc_ifc_work)->work) &idev->mc_lock rcu_read_lock rcu_read_lock rcu_read_lock_bh rcu_read_lock rcu_read_lock &nf_conntrack_locks[i] &nf_conntrack_locks[i]/1 irq_context: 0 (wq_completion)mld (work_completion)(&(&idev->mc_ifc_work)->work) &idev->mc_lock rcu_read_lock rcu_read_lock rcu_read_lock_bh rcu_read_lock rcu_read_lock &nf_conntrack_locks[i] &nf_conntrack_locks[i]/1 batched_entropy_u8.lock irq_context: 0 (wq_completion)mld (work_completion)(&(&idev->mc_ifc_work)->work) &idev->mc_lock rcu_read_lock rcu_read_lock rcu_read_lock_bh rcu_read_lock rcu_read_lock &obj_hash[i].lock irq_context: 0 (wq_completion)mld (work_completion)(&(&idev->mc_ifc_work)->work) &idev->mc_lock rcu_read_lock rcu_read_lock rcu_read_lock_bh rcu_read_lock rcu_read_lock &tbl->lock irq_context: 0 (wq_completion)mld (work_completion)(&(&idev->mc_ifc_work)->work) &idev->mc_lock rcu_read_lock rcu_read_lock rcu_read_lock_bh rcu_read_lock rcu_read_lock &n->lock irq_context: 0 (wq_completion)mld (work_completion)(&(&idev->mc_ifc_work)->work) &idev->mc_lock rcu_read_lock rcu_read_lock rcu_read_lock_bh rcu_read_lock rcu_read_lock &____s->seqcount#9 irq_context: 0 (wq_completion)mld (work_completion)(&(&idev->mc_ifc_work)->work) &idev->mc_lock rcu_read_lock rcu_read_lock rcu_read_lock_bh rcu_read_lock rcu_read_lock rcu_read_lock_bh tk_core.seq.seqcount irq_context: 0 (wq_completion)mld (work_completion)(&(&idev->mc_ifc_work)->work) &idev->mc_lock rcu_read_lock rcu_read_lock rcu_read_lock_bh rcu_read_lock rcu_read_lock rcu_read_lock_bh &list->lock#5 irq_context: softirq rcu_read_lock rcu_read_lock &pool->lock/1 &p->pi_lock &rq->__lock irq_context: softirq rcu_read_lock rcu_read_lock &pool->lock/1 &p->pi_lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)ipv6_addrconf (work_completion)(&(&ifa->dad_work)->work) rtnl_mutex rcu_read_lock rcu_read_lock rcu_read_lock_bh rcu_read_lock_bh &obj_hash[i].lock irq_context: 0 (wq_completion)ipv6_addrconf (work_completion)(&(&ifa->dad_work)->work) rtnl_mutex rcu_read_lock rcu_read_lock rcu_read_lock_bh rcu_read_lock_bh &obj_hash[i].lock pool_lock irq_context: 0 (wq_completion)ipv6_addrconf (work_completion)(&(&ifa->dad_work)->work) rtnl_mutex rcu_read_lock rcu_read_lock rcu_read_lock_bh rcu_read_lock_bh pool_lock#2 irq_context: 0 (wq_completion)ipv6_addrconf (work_completion)(&(&ifa->dad_work)->work) rtnl_mutex rcu_read_lock rcu_read_lock rcu_read_lock_bh rcu_read_lock_bh rcu_read_lock tk_core.seq.seqcount irq_context: 0 (wq_completion)ipv6_addrconf (work_completion)(&(&ifa->dad_work)->work) rtnl_mutex rcu_read_lock rcu_read_lock rcu_read_lock_bh rcu_read_lock_bh rcu_read_lock &list->lock#5 irq_context: 0 rtnl_mutex &wg->device_update_lock irq_context: 0 rtnl_mutex &wg->device_update_lock fs_reclaim irq_context: 0 rtnl_mutex &wg->device_update_lock fs_reclaim mmu_notifier_invalidate_range_start irq_context: 0 rtnl_mutex &wg->device_update_lock pool_lock#2 irq_context: 0 rtnl_mutex &wg->device_update_lock mmu_notifier_invalidate_range_start irq_context: 0 rtnl_mutex &wg->device_update_lock &sb->s_type->i_lock_key#8 irq_context: 0 rtnl_mutex &wg->device_update_lock &dir->lock irq_context: 0 rtnl_mutex &wg->device_update_lock &obj_hash[i].lock irq_context: 0 rtnl_mutex &wg->device_update_lock k-sk_lock-AF_INET irq_context: 0 rtnl_mutex &wg->device_update_lock k-sk_lock-AF_INET k-slock-AF_INET irq_context: 0 rtnl_mutex &wg->device_update_lock k-sk_lock-AF_INET &table->hash[i].lock irq_context: 0 rtnl_mutex &wg->device_update_lock k-sk_lock-AF_INET &table->hash[i].lock k-clock-AF_INET irq_context: 0 rtnl_mutex &wg->device_update_lock k-sk_lock-AF_INET &table->hash[i].lock &table->hash2[i].lock irq_context: 0 rtnl_mutex &wg->device_update_lock k-slock-AF_INET irq_context: 0 rtnl_mutex &wg->device_update_lock cpu_hotplug_lock irq_context: 0 rtnl_mutex &wg->device_update_lock cpu_hotplug_lock jump_label_mutex irq_context: 0 rtnl_mutex &wg->device_update_lock cpu_hotplug_lock jump_label_mutex text_mutex irq_context: 0 rtnl_mutex &wg->device_update_lock cpu_hotplug_lock jump_label_mutex text_mutex ptlock_ptr(page)#2 irq_context: 0 rtnl_mutex &wg->device_update_lock k-sk_lock-AF_INET6 irq_context: 0 rtnl_mutex &wg->device_update_lock k-sk_lock-AF_INET6 k-slock-AF_INET6 irq_context: 0 rtnl_mutex &wg->device_update_lock k-slock-AF_INET6 irq_context: 0 rtnl_mutex &wg->device_update_lock k-sk_lock-AF_INET6 &table->hash[i].lock irq_context: 0 rtnl_mutex &wg->device_update_lock k-sk_lock-AF_INET6 &table->hash[i].lock k-clock-AF_INET6 irq_context: 0 rtnl_mutex &wg->device_update_lock k-sk_lock-AF_INET6 &table->hash[i].lock &table->hash2[i].lock irq_context: 0 rtnl_mutex &wg->device_update_lock &wg->socket_update_lock irq_context: 0 rtnl_mutex &wg->device_update_lock rcu_state.exp_mutex rcu_node_0 irq_context: 0 rtnl_mutex &wg->device_update_lock rcu_state.exp_mutex &obj_hash[i].lock irq_context: 0 rtnl_mutex &wg->device_update_lock rcu_state.exp_mutex rcu_read_lock &pool->lock irq_context: 0 rtnl_mutex &wg->device_update_lock rcu_state.exp_mutex rcu_read_lock &pool->lock &obj_hash[i].lock irq_context: 0 rtnl_mutex &wg->device_update_lock rcu_state.exp_mutex &rnp->exp_wq[1] irq_context: 0 rtnl_mutex &wg->device_update_lock rcu_state.exp_mutex &rq->__lock irq_context: 0 rtnl_mutex &wg->device_update_lock rcu_state.exp_mutex &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 &mm->mmap_lock rcu_read_lock &rcu_state.expedited_wq &p->pi_lock &rq->__lock irq_context: 0 &mm->mmap_lock rcu_read_lock &rcu_state.expedited_wq &p->pi_lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 rtnl_mutex &wg->device_update_lock &list->lock#17 irq_context: 0 rtnl_mutex &wg->device_update_lock rcu_read_lock_bh tk_core.seq.seqcount irq_context: 0 rtnl_mutex &wg->device_update_lock rcu_read_lock_bh rcu_read_lock &pool->lock/1 irq_context: 0 rtnl_mutex &wg->device_update_lock rcu_read_lock_bh rcu_read_lock &pool->lock/1 &obj_hash[i].lock irq_context: 0 rtnl_mutex &wg->device_update_lock rcu_read_lock_bh rcu_read_lock &pool->lock/1 pool_lock#2 irq_context: 0 rtnl_mutex &wg->device_update_lock rcu_read_lock_bh rcu_read_lock &pool->lock/1 &p->pi_lock irq_context: 0 rtnl_mutex &wg->device_update_lock rcu_read_lock_bh rcu_read_lock &pool->lock/1 &p->pi_lock &rq->__lock irq_context: 0 rtnl_mutex &wg->device_update_lock rcu_read_lock_bh rcu_read_lock &pool->lock/1 &p->pi_lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)wg-kex-wg0 irq_context: 0 (wq_completion)wg-kex-wg0 (work_completion)(&peer->transmit_handshake_work) irq_context: 0 (wq_completion)wg-kex-wg0 (work_completion)(&peer->transmit_handshake_work) tk_core.seq.seqcount irq_context: 0 (wq_completion)wg-kex-wg0 (work_completion)(&peer->transmit_handshake_work) &wg->static_identity.lock irq_context: 0 (wq_completion)wg-kex-wg0 (work_completion)(&peer->transmit_handshake_work) &wg->static_identity.lock &handshake->lock irq_context: 0 (wq_completion)wg-kex-wg0 (work_completion)(&peer->transmit_handshake_work) &wg->static_identity.lock &handshake->lock crngs.lock irq_context: 0 (wq_completion)wg-kex-wg0 (work_completion)(&peer->transmit_handshake_work) &wg->static_identity.lock &handshake->lock tk_core.seq.seqcount irq_context: 0 (wq_completion)wg-kex-wg0 (work_completion)(&peer->transmit_handshake_work) &wg->static_identity.lock &handshake->lock &table->lock#2 irq_context: 0 (wq_completion)wg-kex-wg0 (work_completion)(&peer->transmit_handshake_work) &wg->static_identity.lock &handshake->lock rcu_read_lock_bh batched_entropy_u32.lock irq_context: 0 (wq_completion)wg-kex-wg0 (work_completion)(&peer->transmit_handshake_work) &wg->static_identity.lock &handshake->lock rcu_read_lock_bh &table->lock#2 irq_context: 0 (wq_completion)wg-kex-wg0 (work_completion)(&peer->transmit_handshake_work) &cookie->lock irq_context: 0 (wq_completion)wg-kex-wg0 (work_completion)(&peer->transmit_handshake_work) &cookie->lock irq_context: 0 (wq_completion)wg-kex-wg0 (work_completion)(&peer->transmit_handshake_work) rcu_read_lock_bh &obj_hash[i].lock irq_context: 0 (wq_completion)wg-kex-wg0 (work_completion)(&peer->transmit_handshake_work) rcu_read_lock_bh &base->lock irq_context: 0 (wq_completion)wg-kex-wg0 (work_completion)(&peer->transmit_handshake_work) rcu_read_lock_bh &base->lock &obj_hash[i].lock irq_context: 0 (wq_completion)wg-kex-wg0 (work_completion)(&peer->transmit_handshake_work) &obj_hash[i].lock irq_context: 0 (wq_completion)wg-kex-wg0 (work_completion)(&peer->transmit_handshake_work) pool_lock#2 irq_context: 0 (wq_completion)wg-kex-wg0 (work_completion)(&peer->transmit_handshake_work) &peer->endpoint_lock irq_context: 0 (wq_completion)wg-kex-wg0 (work_completion)(&peer->transmit_handshake_work) &peer->endpoint_lock rcu_read_lock_bh rcu_read_lock rcu_read_lock rcu_read_lock pool_lock#2 irq_context: 0 (wq_completion)wg-kex-wg0 (work_completion)(&peer->transmit_handshake_work) &peer->endpoint_lock rcu_read_lock_bh rcu_read_lock rcu_read_lock rcu_read_lock &dir->lock#2 irq_context: 0 (wq_completion)wg-kex-wg0 (work_completion)(&peer->transmit_handshake_work) &peer->endpoint_lock rcu_read_lock_bh rcu_read_lock &____s->seqcount#7 irq_context: 0 (wq_completion)wg-kex-wg0 (work_completion)(&peer->transmit_handshake_work) &peer->endpoint_lock rcu_read_lock_bh rcu_read_lock pool_lock#2 irq_context: 0 (wq_completion)wg-kex-wg0 (work_completion)(&peer->transmit_handshake_work) &peer->endpoint_lock rcu_read_lock_bh rcu_read_lock &nf_conntrack_locks[i] irq_context: 0 (wq_completion)wg-kex-wg0 (work_completion)(&peer->transmit_handshake_work) &peer->endpoint_lock rcu_read_lock_bh rcu_read_lock &nf_conntrack_locks[i] &nf_conntrack_locks[i]/1 irq_context: 0 (wq_completion)wg-kex-wg0 (work_completion)(&peer->transmit_handshake_work) &peer->endpoint_lock rcu_read_lock_bh rcu_read_lock &nf_conntrack_locks[i] &nf_conntrack_locks[i]/1 batched_entropy_u8.lock irq_context: 0 (wq_completion)wg-kex-wg0 (work_completion)(&peer->transmit_handshake_work) &peer->endpoint_lock rcu_read_lock_bh rcu_read_lock &nf_conntrack_locks[i]/1 irq_context: 0 (wq_completion)wg-kex-wg0 (work_completion)(&peer->transmit_handshake_work) &peer->endpoint_lock rcu_read_lock_bh rcu_read_lock &obj_hash[i].lock irq_context: 0 (wq_completion)wg-kex-wg0 (work_completion)(&peer->transmit_handshake_work) &peer->endpoint_lock rcu_read_lock_bh rcu_read_lock &dir->lock#2 irq_context: 0 (wq_completion)wg-kex-wg0 (work_completion)(&peer->transmit_handshake_work) &peer->endpoint_lock rcu_read_lock_bh rcu_read_lock &tbl->lock irq_context: 0 (wq_completion)wg-kex-wg0 (work_completion)(&peer->transmit_handshake_work) &peer->endpoint_lock rcu_read_lock_bh rcu_read_lock &n->lock irq_context: 0 (wq_completion)wg-kex-wg0 (work_completion)(&peer->transmit_handshake_work) &peer->endpoint_lock rcu_read_lock_bh rcu_read_lock &____s->seqcount#9 irq_context: 0 (wq_completion)wg-kex-wg0 (work_completion)(&peer->transmit_handshake_work) &peer->endpoint_lock rcu_read_lock_bh rcu_read_lock rcu_read_lock_bh tk_core.seq.seqcount irq_context: 0 (wq_completion)wg-kex-wg0 (work_completion)(&peer->transmit_handshake_work) &peer->endpoint_lock rcu_read_lock_bh rcu_read_lock rcu_read_lock_bh &list->lock#5 irq_context: softirq rcu_read_lock rcu_read_lock k-slock-AF_INET6 rcu_read_lock pool_lock#2 irq_context: softirq rcu_read_lock rcu_read_lock k-slock-AF_INET6 rcu_read_lock rcu_read_lock rcu_read_lock &____s->seqcount#7 irq_context: softirq rcu_read_lock rcu_read_lock k-slock-AF_INET6 rcu_read_lock rcu_read_lock &____s->seqcount#10 irq_context: softirq rcu_read_lock rcu_read_lock k-slock-AF_INET6 rcu_read_lock rcu_read_lock rcu_read_lock_bh tk_core.seq.seqcount irq_context: softirq rcu_read_lock rcu_read_lock k-slock-AF_INET6 rcu_read_lock rcu_read_lock rcu_read_lock_bh &list->lock#5 irq_context: 0 rtnl_mutex &wg->device_update_lock rcu_state.exp_mutex rcu_read_lock &pool->lock &p->pi_lock irq_context: 0 (wq_completion)wg-kex-wg0 (work_completion)(&peer->transmit_handshake_work) batched_entropy_u8.lock irq_context: 0 rtnl_mutex &wg->device_update_lock rcu_state.exp_mutex &rnp->exp_wq[2] irq_context: 0 (wq_completion)wg-kex-wg0 (work_completion)(&peer->transmit_handshake_work) &c->lock irq_context: 0 (wq_completion)wg-kex-wg0 (work_completion)(&peer->transmit_handshake_work) &peer->endpoint_lock rcu_read_lock_bh batched_entropy_u32.lock irq_context: 0 (wq_completion)wg-kex-wg0 (work_completion)(&peer->transmit_handshake_work) &peer->endpoint_lock rcu_read_lock_bh rcu_read_lock &____s->seqcount#10 irq_context: softirq rcu_read_lock rcu_read_lock &r->producer_lock#2 irq_context: 0 (wq_completion)wg-kex-wg1 irq_context: 0 (wq_completion)wg-kex-wg1 (work_completion)(&({ do { const void *__vpp_verify = (typeof((worker) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __asm__ ("" : "=r"(__ptr) : "0"((typeof(*((worker))) *)((worker)))); (typeof((typeof(*((worker))) *)((worker)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); })->work) irq_context: 0 (wq_completion)wg-kex-wg1 (work_completion)(&({ do { const void *__vpp_verify = (typeof((worker) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __asm__ ("" : "=r"(__ptr) : "0"((typeof(*((worker))) *)((worker)))); (typeof((typeof(*((worker))) *)((worker)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); })->work) &r->consumer_lock#2 irq_context: 0 (wq_completion)wg-kex-wg1 (work_completion)(&({ do { const void *__vpp_verify = (typeof((worker) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __asm__ ("" : "=r"(__ptr) : "0"((typeof(*((worker))) *)((worker)))); (typeof((typeof(*((worker))) *)((worker)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); })->work) &wg->static_identity.lock irq_context: 0 (wq_completion)wg-kex-wg1 (work_completion)(&({ do { const void *__vpp_verify = (typeof((worker) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __asm__ ("" : "=r"(__ptr) : "0"((typeof(*((worker))) *)((worker)))); (typeof((typeof(*((worker))) *)((worker)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); })->work) &wg->static_identity.lock &handshake->lock irq_context: 0 (wq_completion)wg-kex-wg1 (work_completion)(&({ do { const void *__vpp_verify = (typeof((worker) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __asm__ ("" : "=r"(__ptr) : "0"((typeof(*((worker))) *)((worker)))); (typeof((typeof(*((worker))) *)((worker)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); })->work) &wg->static_identity.lock &handshake->lock tk_core.seq.seqcount irq_context: 0 (wq_completion)wg-kex-wg1 (work_completion)(&({ do { const void *__vpp_verify = (typeof((worker) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __asm__ ("" : "=r"(__ptr) : "0"((typeof(*((worker))) *)((worker)))); (typeof((typeof(*((worker))) *)((worker)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); })->work) &wg->static_identity.lock &handshake->lock irq_context: 0 (wq_completion)wg-kex-wg1 (work_completion)(&({ do { const void *__vpp_verify = (typeof((worker) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __asm__ ("" : "=r"(__ptr) : "0"((typeof(*((worker))) *)((worker)))); (typeof((typeof(*((worker))) *)((worker)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); })->work) &wg->static_identity.lock &handshake->lock tk_core.seq.seqcount irq_context: 0 (wq_completion)wg-kex-wg1 (work_completion)(&({ do { const void *__vpp_verify = (typeof((worker) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __asm__ ("" : "=r"(__ptr) : "0"((typeof(*((worker))) *)((worker)))); (typeof((typeof(*((worker))) *)((worker)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); })->work) &peer->endpoint_lock irq_context: 0 (wq_completion)wg-kex-wg1 (work_completion)(&({ do { const void *__vpp_verify = (typeof((worker) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __asm__ ("" : "=r"(__ptr) : "0"((typeof(*((worker))) *)((worker)))); (typeof((typeof(*((worker))) *)((worker)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); })->work) tk_core.seq.seqcount irq_context: 0 (wq_completion)wg-kex-wg1 (work_completion)(&({ do { const void *__vpp_verify = (typeof((worker) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __asm__ ("" : "=r"(__ptr) : "0"((typeof(*((worker))) *)((worker)))); (typeof((typeof(*((worker))) *)((worker)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); })->work) &wg->static_identity.lock &handshake->lock crngs.lock irq_context: 0 (wq_completion)wg-kex-wg1 (work_completion)(&({ do { const void *__vpp_verify = (typeof((worker) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __asm__ ("" : "=r"(__ptr) : "0"((typeof(*((worker))) *)((worker)))); (typeof((typeof(*((worker))) *)((worker)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); })->work) &wg->static_identity.lock &handshake->lock &table->lock#2 irq_context: 0 (wq_completion)wg-kex-wg1 (work_completion)(&({ do { const void *__vpp_verify = (typeof((worker) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __asm__ ("" : "=r"(__ptr) : "0"((typeof(*((worker))) *)((worker)))); (typeof((typeof(*((worker))) *)((worker)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); })->work) &wg->static_identity.lock &handshake->lock rcu_read_lock_bh batched_entropy_u32.lock irq_context: 0 (wq_completion)wg-kex-wg1 (work_completion)(&({ do { const void *__vpp_verify = (typeof((worker) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __asm__ ("" : "=r"(__ptr) : "0"((typeof(*((worker))) *)((worker)))); (typeof((typeof(*((worker))) *)((worker)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); })->work) &wg->static_identity.lock &handshake->lock rcu_read_lock_bh &table->lock#2 irq_context: 0 (wq_completion)wg-kex-wg1 (work_completion)(&({ do { const void *__vpp_verify = (typeof((worker) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __asm__ ("" : "=r"(__ptr) : "0"((typeof(*((worker))) *)((worker)))); (typeof((typeof(*((worker))) *)((worker)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); })->work) &cookie->lock irq_context: 0 (wq_completion)wg-kex-wg1 (work_completion)(&({ do { const void *__vpp_verify = (typeof((worker) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __asm__ ("" : "=r"(__ptr) : "0"((typeof(*((worker))) *)((worker)))); (typeof((typeof(*((worker))) *)((worker)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); })->work) &cookie->lock irq_context: 0 (wq_completion)wg-kex-wg1 (work_completion)(&({ do { const void *__vpp_verify = (typeof((worker) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __asm__ ("" : "=r"(__ptr) : "0"((typeof(*((worker))) *)((worker)))); (typeof((typeof(*((worker))) *)((worker)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); })->work) &handshake->lock irq_context: 0 (wq_completion)wg-kex-wg1 (work_completion)(&({ do { const void *__vpp_verify = (typeof((worker) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __asm__ ("" : "=r"(__ptr) : "0"((typeof(*((worker))) *)((worker)))); (typeof((typeof(*((worker))) *)((worker)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); })->work) &handshake->lock fs_reclaim irq_context: 0 (wq_completion)wg-kex-wg1 (work_completion)(&({ do { const void *__vpp_verify = (typeof((worker) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __asm__ ("" : "=r"(__ptr) : "0"((typeof(*((worker))) *)((worker)))); (typeof((typeof(*((worker))) *)((worker)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); })->work) &handshake->lock fs_reclaim mmu_notifier_invalidate_range_start irq_context: 0 (wq_completion)wg-kex-wg1 (work_completion)(&({ do { const void *__vpp_verify = (typeof((worker) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __asm__ ("" : "=r"(__ptr) : "0"((typeof(*((worker))) *)((worker)))); (typeof((typeof(*((worker))) *)((worker)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); })->work) &handshake->lock pool_lock#2 irq_context: 0 (wq_completion)wg-kex-wg1 (work_completion)(&({ do { const void *__vpp_verify = (typeof((worker) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __asm__ ("" : "=r"(__ptr) : "0"((typeof(*((worker))) *)((worker)))); (typeof((typeof(*((worker))) *)((worker)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); })->work) &handshake->lock tk_core.seq.seqcount irq_context: 0 (wq_completion)wg-kex-wg1 (work_completion)(&({ do { const void *__vpp_verify = (typeof((worker) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __asm__ ("" : "=r"(__ptr) : "0"((typeof(*((worker))) *)((worker)))); (typeof((typeof(*((worker))) *)((worker)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); })->work) &handshake->lock rcu_read_lock_bh &peer->keypairs.keypair_update_lock irq_context: 0 (wq_completion)wg-kex-wg1 (work_completion)(&({ do { const void *__vpp_verify = (typeof((worker) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __asm__ ("" : "=r"(__ptr) : "0"((typeof(*((worker))) *)((worker)))); (typeof((typeof(*((worker))) *)((worker)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); })->work) &handshake->lock rcu_read_lock_bh &table->lock#2 irq_context: 0 (wq_completion)wg-kex-wg1 (work_completion)(&({ do { const void *__vpp_verify = (typeof((worker) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __asm__ ("" : "=r"(__ptr) : "0"((typeof(*((worker))) *)((worker)))); (typeof((typeof(*((worker))) *)((worker)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); })->work) rcu_read_lock_bh &obj_hash[i].lock irq_context: 0 (wq_completion)wg-kex-wg1 (work_completion)(&({ do { const void *__vpp_verify = (typeof((worker) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __asm__ ("" : "=r"(__ptr) : "0"((typeof(*((worker))) *)((worker)))); (typeof((typeof(*((worker))) *)((worker)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); })->work) rcu_read_lock_bh &base->lock irq_context: 0 (wq_completion)wg-kex-wg1 (work_completion)(&({ do { const void *__vpp_verify = (typeof((worker) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __asm__ ("" : "=r"(__ptr) : "0"((typeof(*((worker))) *)((worker)))); (typeof((typeof(*((worker))) *)((worker)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); })->work) rcu_read_lock_bh &base->lock &obj_hash[i].lock irq_context: 0 (wq_completion)wg-kex-wg1 (work_completion)(&({ do { const void *__vpp_verify = (typeof((worker) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __asm__ ("" : "=r"(__ptr) : "0"((typeof(*((worker))) *)((worker)))); (typeof((typeof(*((worker))) *)((worker)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); })->work) &obj_hash[i].lock irq_context: 0 (wq_completion)wg-kex-wg1 (work_completion)(&({ do { const void *__vpp_verify = (typeof((worker) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __asm__ ("" : "=r"(__ptr) : "0"((typeof(*((worker))) *)((worker)))); (typeof((typeof(*((worker))) *)((worker)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); })->work) pool_lock#2 irq_context: 0 (wq_completion)wg-kex-wg1 (work_completion)(&({ do { const void *__vpp_verify = (typeof((worker) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __asm__ ("" : "=r"(__ptr) : "0"((typeof(*((worker))) *)((worker)))); (typeof((typeof(*((worker))) *)((worker)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); })->work) &peer->endpoint_lock irq_context: 0 (wq_completion)wg-kex-wg1 (work_completion)(&({ do { const void *__vpp_verify = (typeof((worker) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __asm__ ("" : "=r"(__ptr) : "0"((typeof(*((worker))) *)((worker)))); (typeof((typeof(*((worker))) *)((worker)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); })->work) &peer->endpoint_lock rcu_read_lock_bh rcu_read_lock &c->lock irq_context: 0 (wq_completion)wg-kex-wg1 (work_completion)(&({ do { const void *__vpp_verify = (typeof((worker) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __asm__ ("" : "=r"(__ptr) : "0"((typeof(*((worker))) *)((worker)))); (typeof((typeof(*((worker))) *)((worker)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); })->work) &peer->endpoint_lock rcu_read_lock_bh rcu_read_lock pool_lock#2 irq_context: 0 (wq_completion)wg-kex-wg1 (work_completion)(&({ do { const void *__vpp_verify = (typeof((worker) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __asm__ ("" : "=r"(__ptr) : "0"((typeof(*((worker))) *)((worker)))); (typeof((typeof(*((worker))) *)((worker)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); })->work) &peer->endpoint_lock rcu_read_lock_bh rcu_read_lock &dir->lock#2 irq_context: 0 (wq_completion)wg-kex-wg1 (work_completion)(&({ do { const void *__vpp_verify = (typeof((worker) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __asm__ ("" : "=r"(__ptr) : "0"((typeof(*((worker))) *)((worker)))); (typeof((typeof(*((worker))) *)((worker)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); })->work) &peer->endpoint_lock rcu_read_lock_bh rcu_read_lock &____s->seqcount#7 irq_context: 0 (wq_completion)wg-kex-wg1 (work_completion)(&({ do { const void *__vpp_verify = (typeof((worker) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __asm__ ("" : "=r"(__ptr) : "0"((typeof(*((worker))) *)((worker)))); (typeof((typeof(*((worker))) *)((worker)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); })->work) &peer->endpoint_lock rcu_read_lock_bh rcu_read_lock &____s->seqcount#10 irq_context: 0 (wq_completion)wg-kex-wg1 (work_completion)(&({ do { const void *__vpp_verify = (typeof((worker) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __asm__ ("" : "=r"(__ptr) : "0"((typeof(*((worker))) *)((worker)))); (typeof((typeof(*((worker))) *)((worker)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); })->work) &peer->endpoint_lock rcu_read_lock_bh rcu_read_lock rcu_read_lock_bh tk_core.seq.seqcount irq_context: 0 (wq_completion)wg-kex-wg1 (work_completion)(&({ do { const void *__vpp_verify = (typeof((worker) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __asm__ ("" : "=r"(__ptr) : "0"((typeof(*((worker))) *)((worker)))); (typeof((typeof(*((worker))) *)((worker)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); })->work) &peer->endpoint_lock rcu_read_lock_bh rcu_read_lock rcu_read_lock_bh &list->lock#5 irq_context: 0 (wq_completion)wg-kex-wg1 (work_completion)(&({ do { const void *__vpp_verify = (typeof((worker) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __asm__ ("" : "=r"(__ptr) : "0"((typeof(*((worker))) *)((worker)))); (typeof((typeof(*((worker))) *)((worker)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); })->work) &rq->__lock irq_context: 0 (wq_completion)wg-kex-wg0#2 irq_context: 0 (wq_completion)wg-kex-wg0#2 (work_completion)(&({ do { const void *__vpp_verify = (typeof((worker) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __asm__ ("" : "=r"(__ptr) : "0"((typeof(*((worker))) *)((worker)))); (typeof((typeof(*((worker))) *)((worker)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); })->work) irq_context: 0 (wq_completion)wg-kex-wg0#2 (work_completion)(&({ do { const void *__vpp_verify = (typeof((worker) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __asm__ ("" : "=r"(__ptr) : "0"((typeof(*((worker))) *)((worker)))); (typeof((typeof(*((worker))) *)((worker)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); })->work) &r->consumer_lock#2 irq_context: 0 (wq_completion)wg-kex-wg0#2 (work_completion)(&({ do { const void *__vpp_verify = (typeof((worker) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __asm__ ("" : "=r"(__ptr) : "0"((typeof(*((worker))) *)((worker)))); (typeof((typeof(*((worker))) *)((worker)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); })->work) &wg->static_identity.lock irq_context: 0 (wq_completion)wg-kex-wg0#2 (work_completion)(&({ do { const void *__vpp_verify = (typeof((worker) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __asm__ ("" : "=r"(__ptr) : "0"((typeof(*((worker))) *)((worker)))); (typeof((typeof(*((worker))) *)((worker)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); })->work) &wg->static_identity.lock &handshake->lock irq_context: 0 (wq_completion)wg-kex-wg0#2 (work_completion)(&({ do { const void *__vpp_verify = (typeof((worker) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __asm__ ("" : "=r"(__ptr) : "0"((typeof(*((worker))) *)((worker)))); (typeof((typeof(*((worker))) *)((worker)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); })->work) &wg->static_identity.lock &handshake->lock irq_context: 0 (wq_completion)wg-kex-wg0#2 (work_completion)(&({ do { const void *__vpp_verify = (typeof((worker) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __asm__ ("" : "=r"(__ptr) : "0"((typeof(*((worker))) *)((worker)))); (typeof((typeof(*((worker))) *)((worker)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); })->work) &peer->endpoint_lock irq_context: 0 (wq_completion)wg-kex-wg0#2 (work_completion)(&({ do { const void *__vpp_verify = (typeof((worker) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __asm__ ("" : "=r"(__ptr) : "0"((typeof(*((worker))) *)((worker)))); (typeof((typeof(*((worker))) *)((worker)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); })->work) &handshake->lock irq_context: 0 (wq_completion)wg-kex-wg0#2 (work_completion)(&({ do { const void *__vpp_verify = (typeof((worker) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __asm__ ("" : "=r"(__ptr) : "0"((typeof(*((worker))) *)((worker)))); (typeof((typeof(*((worker))) *)((worker)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); })->work) &handshake->lock fs_reclaim irq_context: 0 (wq_completion)wg-kex-wg0#2 (work_completion)(&({ do { const void *__vpp_verify = (typeof((worker) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __asm__ ("" : "=r"(__ptr) : "0"((typeof(*((worker))) *)((worker)))); (typeof((typeof(*((worker))) *)((worker)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); })->work) &handshake->lock fs_reclaim mmu_notifier_invalidate_range_start irq_context: 0 (wq_completion)wg-kex-wg0#2 (work_completion)(&({ do { const void *__vpp_verify = (typeof((worker) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __asm__ ("" : "=r"(__ptr) : "0"((typeof(*((worker))) *)((worker)))); (typeof((typeof(*((worker))) *)((worker)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); })->work) &handshake->lock &c->lock irq_context: 0 (wq_completion)wg-kex-wg0#2 (work_completion)(&({ do { const void *__vpp_verify = (typeof((worker) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __asm__ ("" : "=r"(__ptr) : "0"((typeof(*((worker))) *)((worker)))); (typeof((typeof(*((worker))) *)((worker)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); })->work) &handshake->lock &pcp->lock &zone->lock irq_context: 0 (wq_completion)wg-kex-wg0#2 (work_completion)(&({ do { const void *__vpp_verify = (typeof((worker) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __asm__ ("" : "=r"(__ptr) : "0"((typeof(*((worker))) *)((worker)))); (typeof((typeof(*((worker))) *)((worker)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); })->work) &handshake->lock &____s->seqcount irq_context: 0 (wq_completion)wg-kex-wg0#2 (work_completion)(&({ do { const void *__vpp_verify = (typeof((worker) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __asm__ ("" : "=r"(__ptr) : "0"((typeof(*((worker))) *)((worker)))); (typeof((typeof(*((worker))) *)((worker)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); })->work) &handshake->lock pool_lock#2 irq_context: 0 (wq_completion)wg-kex-wg0#2 (work_completion)(&({ do { const void *__vpp_verify = (typeof((worker) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __asm__ ("" : "=r"(__ptr) : "0"((typeof(*((worker))) *)((worker)))); (typeof((typeof(*((worker))) *)((worker)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); })->work) &handshake->lock tk_core.seq.seqcount irq_context: 0 (wq_completion)wg-kex-wg0#2 (work_completion)(&({ do { const void *__vpp_verify = (typeof((worker) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __asm__ ("" : "=r"(__ptr) : "0"((typeof(*((worker))) *)((worker)))); (typeof((typeof(*((worker))) *)((worker)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); })->work) &handshake->lock rcu_read_lock_bh &peer->keypairs.keypair_update_lock irq_context: 0 (wq_completion)wg-kex-wg0#2 (work_completion)(&({ do { const void *__vpp_verify = (typeof((worker) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __asm__ ("" : "=r"(__ptr) : "0"((typeof(*((worker))) *)((worker)))); (typeof((typeof(*((worker))) *)((worker)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); })->work) &handshake->lock rcu_read_lock_bh &table->lock#2 irq_context: 0 (wq_completion)wg-kex-wg0#2 (work_completion)(&({ do { const void *__vpp_verify = (typeof((worker) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __asm__ ("" : "=r"(__ptr) : "0"((typeof(*((worker))) *)((worker)))); (typeof((typeof(*((worker))) *)((worker)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); })->work) rcu_read_lock_bh &obj_hash[i].lock irq_context: 0 (wq_completion)wg-kex-wg0#2 (work_completion)(&({ do { const void *__vpp_verify = (typeof((worker) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __asm__ ("" : "=r"(__ptr) : "0"((typeof(*((worker))) *)((worker)))); (typeof((typeof(*((worker))) *)((worker)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); })->work) rcu_read_lock_bh &base->lock irq_context: 0 (wq_completion)wg-kex-wg0#2 (work_completion)(&({ do { const void *__vpp_verify = (typeof((worker) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __asm__ ("" : "=r"(__ptr) : "0"((typeof(*((worker))) *)((worker)))); (typeof((typeof(*((worker))) *)((worker)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); })->work) rcu_read_lock_bh &base->lock &obj_hash[i].lock irq_context: 0 (wq_completion)wg-kex-wg0#2 (work_completion)(&({ do { const void *__vpp_verify = (typeof((worker) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __asm__ ("" : "=r"(__ptr) : "0"((typeof(*((worker))) *)((worker)))); (typeof((typeof(*((worker))) *)((worker)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); })->work) &obj_hash[i].lock irq_context: 0 (wq_completion)wg-kex-wg0#2 (work_completion)(&({ do { const void *__vpp_verify = (typeof((worker) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __asm__ ("" : "=r"(__ptr) : "0"((typeof(*((worker))) *)((worker)))); (typeof((typeof(*((worker))) *)((worker)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); })->work) &base->lock irq_context: 0 (wq_completion)wg-kex-wg0#2 (work_completion)(&({ do { const void *__vpp_verify = (typeof((worker) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __asm__ ("" : "=r"(__ptr) : "0"((typeof(*((worker))) *)((worker)))); (typeof((typeof(*((worker))) *)((worker)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); })->work) &base->lock &obj_hash[i].lock irq_context: 0 (wq_completion)wg-kex-wg0#2 (work_completion)(&({ do { const void *__vpp_verify = (typeof((worker) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __asm__ ("" : "=r"(__ptr) : "0"((typeof(*((worker))) *)((worker)))); (typeof((typeof(*((worker))) *)((worker)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); })->work) tk_core.seq.seqcount irq_context: 0 (wq_completion)wg-kex-wg0#2 (work_completion)(&({ do { const void *__vpp_verify = (typeof((worker) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __asm__ ("" : "=r"(__ptr) : "0"((typeof(*((worker))) *)((worker)))); (typeof((typeof(*((worker))) *)((worker)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); })->work) &list->lock#17 irq_context: 0 (wq_completion)wg-kex-wg0#2 (work_completion)(&({ do { const void *__vpp_verify = (typeof((worker) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __asm__ ("" : "=r"(__ptr) : "0"((typeof(*((worker))) *)((worker)))); (typeof((typeof(*((worker))) *)((worker)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); })->work) rcu_read_lock_bh &r->producer_lock#2 irq_context: 0 (wq_completion)wg-kex-wg0#2 (work_completion)(&({ do { const void *__vpp_verify = (typeof((worker) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __asm__ ("" : "=r"(__ptr) : "0"((typeof(*((worker))) *)((worker)))); (typeof((typeof(*((worker))) *)((worker)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); })->work) rcu_read_lock_bh rcu_read_lock &pool->lock irq_context: 0 (wq_completion)wg-kex-wg0#2 (work_completion)(&({ do { const void *__vpp_verify = (typeof((worker) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __asm__ ("" : "=r"(__ptr) : "0"((typeof(*((worker))) *)((worker)))); (typeof((typeof(*((worker))) *)((worker)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); })->work) rcu_read_lock_bh rcu_read_lock &pool->lock &obj_hash[i].lock irq_context: 0 (wq_completion)wg-kex-wg0#2 (work_completion)(&({ do { const void *__vpp_verify = (typeof((worker) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __asm__ ("" : "=r"(__ptr) : "0"((typeof(*((worker))) *)((worker)))); (typeof((typeof(*((worker))) *)((worker)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); })->work) rcu_read_lock_bh rcu_read_lock &pool->lock &p->pi_lock irq_context: 0 (wq_completion)wg-kex-wg0#2 (work_completion)(&({ do { const void *__vpp_verify = (typeof((worker) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __asm__ ("" : "=r"(__ptr) : "0"((typeof(*((worker))) *)((worker)))); (typeof((typeof(*((worker))) *)((worker)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); })->work) rcu_read_lock_bh rcu_read_lock &pool->lock &p->pi_lock &rq->__lock irq_context: 0 (wq_completion)wg-kex-wg0#2 (work_completion)(&({ do { const void *__vpp_verify = (typeof((worker) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __asm__ ("" : "=r"(__ptr) : "0"((typeof(*((worker))) *)((worker)))); (typeof((typeof(*((worker))) *)((worker)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); })->work) rcu_read_lock_bh rcu_read_lock &pool->lock &p->pi_lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)wg-kex-wg0#2 (work_completion)(&({ do { const void *__vpp_verify = (typeof((worker) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __asm__ ("" : "=r"(__ptr) : "0"((typeof(*((worker))) *)((worker)))); (typeof((typeof(*((worker))) *)((worker)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); })->work) &rq->__lock irq_context: 0 (wq_completion)wg-crypt-wg0 irq_context: 0 (wq_completion)wg-crypt-wg0 (work_completion)(&({ do { const void *__vpp_verify = (typeof((worker) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __asm__ ("" : "=r"(__ptr) : "0"((typeof(*((worker))) *)((worker)))); (typeof((typeof(*((worker))) *)((worker)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); })->work) irq_context: 0 (wq_completion)wg-crypt-wg0 (work_completion)(&({ do { const void *__vpp_verify = (typeof((worker) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __asm__ ("" : "=r"(__ptr) : "0"((typeof(*((worker))) *)((worker)))); (typeof((typeof(*((worker))) *)((worker)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); })->work) &r->consumer_lock#2 irq_context: 0 (wq_completion)wg-crypt-wg0 (work_completion)(&({ do { const void *__vpp_verify = (typeof((worker) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __asm__ ("" : "=r"(__ptr) : "0"((typeof(*((worker))) *)((worker)))); (typeof((typeof(*((worker))) *)((worker)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); })->work) rcu_read_lock &pool->lock irq_context: 0 (wq_completion)wg-crypt-wg0 (work_completion)(&({ do { const void *__vpp_verify = (typeof((worker) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __asm__ ("" : "=r"(__ptr) : "0"((typeof(*((worker))) *)((worker)))); (typeof((typeof(*((worker))) *)((worker)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); })->work) rcu_read_lock &pool->lock &obj_hash[i].lock irq_context: 0 (wq_completion)wg-crypt-wg0 (work_completion)(&({ do { const void *__vpp_verify = (typeof((worker) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __asm__ ("" : "=r"(__ptr) : "0"((typeof(*((worker))) *)((worker)))); (typeof((typeof(*((worker))) *)((worker)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); })->work) rcu_read_lock &pool->lock pool_lock#2 irq_context: 0 (wq_completion)wg-crypt-wg0 (work_completion)(&({ do { const void *__vpp_verify = (typeof((worker) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __asm__ ("" : "=r"(__ptr) : "0"((typeof(*((worker))) *)((worker)))); (typeof((typeof(*((worker))) *)((worker)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); })->work) rcu_read_lock &pool->lock &p->pi_lock irq_context: 0 (wq_completion)wg-crypt-wg0 (work_completion)(&({ do { const void *__vpp_verify = (typeof((worker) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __asm__ ("" : "=r"(__ptr) : "0"((typeof(*((worker))) *)((worker)))); (typeof((typeof(*((worker))) *)((worker)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); })->work) rcu_read_lock &pool->lock &p->pi_lock &rq->__lock irq_context: 0 (wq_completion)wg-crypt-wg0 (work_completion)(&({ do { const void *__vpp_verify = (typeof((worker) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __asm__ ("" : "=r"(__ptr) : "0"((typeof(*((worker))) *)((worker)))); (typeof((typeof(*((worker))) *)((worker)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); })->work) rcu_read_lock &pool->lock &p->pi_lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)wg-crypt-wg0 (work_completion)(&peer->transmit_packet_work) irq_context: 0 (wq_completion)wg-crypt-wg0 (work_completion)(&peer->transmit_packet_work) rcu_read_lock_bh &obj_hash[i].lock irq_context: 0 (wq_completion)wg-crypt-wg0 (work_completion)(&peer->transmit_packet_work) &obj_hash[i].lock irq_context: 0 (wq_completion)wg-crypt-wg0 (work_completion)(&peer->transmit_packet_work) &peer->endpoint_lock irq_context: 0 (wq_completion)wg-crypt-wg0 (work_completion)(&peer->transmit_packet_work) &peer->endpoint_lock rcu_read_lock_bh rcu_read_lock &____s->seqcount#7 irq_context: 0 (wq_completion)wg-crypt-wg0 (work_completion)(&peer->transmit_packet_work) &peer->endpoint_lock rcu_read_lock_bh rcu_read_lock &____s->seqcount#10 irq_context: 0 (wq_completion)wg-crypt-wg0 (work_completion)(&peer->transmit_packet_work) &peer->endpoint_lock rcu_read_lock_bh rcu_read_lock rcu_read_lock_bh tk_core.seq.seqcount irq_context: 0 (wq_completion)wg-crypt-wg0 (work_completion)(&peer->transmit_packet_work) &peer->endpoint_lock rcu_read_lock_bh rcu_read_lock rcu_read_lock_bh &list->lock#5 irq_context: softirq rcu_read_lock rcu_read_lock rcu_read_lock_bh &r->producer_lock#2 irq_context: softirq rcu_read_lock rcu_read_lock rcu_read_lock_bh rcu_read_lock &pool->lock irq_context: softirq rcu_read_lock rcu_read_lock rcu_read_lock_bh rcu_read_lock &pool->lock &obj_hash[i].lock irq_context: softirq rcu_read_lock rcu_read_lock rcu_read_lock_bh rcu_read_lock &pool->lock &p->pi_lock irq_context: softirq rcu_read_lock rcu_read_lock rcu_read_lock_bh rcu_read_lock &pool->lock &p->pi_lock &rq->__lock irq_context: softirq rcu_read_lock rcu_read_lock rcu_read_lock_bh rcu_read_lock &pool->lock &p->pi_lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)wg-crypt-wg0 (work_completion)(&peer->transmit_packet_work) rcu_read_lock_bh tk_core.seq.seqcount irq_context: 0 (wq_completion)wg-crypt-wg1 irq_context: 0 (wq_completion)wg-crypt-wg1 (work_completion)(&({ do { const void *__vpp_verify = (typeof((worker) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __asm__ ("" : "=r"(__ptr) : "0"((typeof(*((worker))) *)((worker)))); (typeof((typeof(*((worker))) *)((worker)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); })->work) irq_context: 0 (wq_completion)bat_events (work_completion)(&(&bat_priv->nc.work)->work) rcu_read_lock &cfs_rq->removed.lock irq_context: 0 (wq_completion)wg-crypt-wg1 (work_completion)(&({ do { const void *__vpp_verify = (typeof((worker) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __asm__ ("" : "=r"(__ptr) : "0"((typeof(*((worker))) *)((worker)))); (typeof((typeof(*((worker))) *)((worker)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); })->work) &r->consumer_lock#2 irq_context: 0 (wq_completion)bat_events (work_completion)(&(&bat_priv->nc.work)->work) rcu_read_lock &obj_hash[i].lock irq_context: 0 (wq_completion)wg-crypt-wg1 (work_completion)(&({ do { const void *__vpp_verify = (typeof((worker) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __asm__ ("" : "=r"(__ptr) : "0"((typeof(*((worker))) *)((worker)))); (typeof((typeof(*((worker))) *)((worker)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); })->work) tk_core.seq.seqcount irq_context: softirq &keypair->receiving_counter.lock irq_context: softirq &peer->keypairs.keypair_update_lock irq_context: softirq &list->lock#17 irq_context: 0 rtnl_mutex &wg->device_update_lock tk_core.seq.seqcount irq_context: 0 rtnl_mutex &wg->device_update_lock rcu_read_lock_bh &r->producer_lock#2 irq_context: 0 rtnl_mutex &wg->device_update_lock rcu_read_lock_bh rcu_read_lock &pool->lock irq_context: 0 rtnl_mutex &wg->device_update_lock rcu_read_lock_bh rcu_read_lock &pool->lock &obj_hash[i].lock irq_context: 0 rtnl_mutex &wg->device_update_lock rcu_read_lock_bh rcu_read_lock &pool->lock &p->pi_lock irq_context: 0 rtnl_mutex &wg->device_update_lock rcu_read_lock_bh rcu_read_lock &pool->lock &p->pi_lock &rq->__lock irq_context: 0 rtnl_mutex &wg->device_update_lock rcu_read_lock_bh rcu_read_lock &pool->lock &p->pi_lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)wg-crypt-wg1 (work_completion)(&({ do { const void *__vpp_verify = (typeof((worker) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __asm__ ("" : "=r"(__ptr) : "0"((typeof(*((worker))) *)((worker)))); (typeof((typeof(*((worker))) *)((worker)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); })->work) rcu_read_lock &pool->lock irq_context: 0 (wq_completion)wg-crypt-wg1 (work_completion)(&({ do { const void *__vpp_verify = (typeof((worker) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __asm__ ("" : "=r"(__ptr) : "0"((typeof(*((worker))) *)((worker)))); (typeof((typeof(*((worker))) *)((worker)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); })->work) rcu_read_lock &pool->lock &obj_hash[i].lock irq_context: 0 (wq_completion)wg-crypt-wg1 (work_completion)(&({ do { const void *__vpp_verify = (typeof((worker) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __asm__ ("" : "=r"(__ptr) : "0"((typeof(*((worker))) *)((worker)))); (typeof((typeof(*((worker))) *)((worker)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); })->work) rcu_read_lock &pool->lock &p->pi_lock irq_context: 0 (wq_completion)wg-crypt-wg1 (work_completion)(&({ do { const void *__vpp_verify = (typeof((worker) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __asm__ ("" : "=r"(__ptr) : "0"((typeof(*((worker))) *)((worker)))); (typeof((typeof(*((worker))) *)((worker)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); })->work) rcu_read_lock &pool->lock &p->pi_lock &rq->__lock irq_context: 0 (wq_completion)wg-crypt-wg1 (work_completion)(&({ do { const void *__vpp_verify = (typeof((worker) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __asm__ ("" : "=r"(__ptr) : "0"((typeof(*((worker))) *)((worker)))); (typeof((typeof(*((worker))) *)((worker)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); })->work) rcu_read_lock &pool->lock &p->pi_lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)wg-crypt-wg1 (work_completion)(&peer->transmit_packet_work) irq_context: 0 (wq_completion)wg-crypt-wg1 (work_completion)(&peer->transmit_packet_work) rcu_read_lock_bh &obj_hash[i].lock irq_context: 0 (wq_completion)wg-crypt-wg1 (work_completion)(&peer->transmit_packet_work) rcu_read_lock_bh &base->lock irq_context: 0 (wq_completion)wg-crypt-wg1 (work_completion)(&peer->transmit_packet_work) &obj_hash[i].lock irq_context: 0 (wq_completion)wg-crypt-wg1 (work_completion)(&peer->transmit_packet_work) &peer->endpoint_lock irq_context: 0 (wq_completion)wg-crypt-wg1 (work_completion)(&peer->transmit_packet_work) &peer->endpoint_lock rcu_read_lock_bh rcu_read_lock pool_lock#2 irq_context: 0 (wq_completion)wg-crypt-wg1 (work_completion)(&peer->transmit_packet_work) &peer->endpoint_lock rcu_read_lock_bh rcu_read_lock &dir->lock#2 irq_context: 0 (wq_completion)wg-crypt-wg1 (work_completion)(&peer->transmit_packet_work) &peer->endpoint_lock rcu_read_lock_bh rcu_read_lock &ul->lock irq_context: 0 (wq_completion)wg-crypt-wg1 (work_completion)(&peer->transmit_packet_work) &peer->endpoint_lock rcu_read_lock_bh batched_entropy_u32.lock irq_context: 0 (wq_completion)wg-crypt-wg1 (work_completion)(&peer->transmit_packet_work) &peer->endpoint_lock rcu_read_lock_bh rcu_read_lock &____s->seqcount#7 irq_context: 0 (wq_completion)wg-crypt-wg1 (work_completion)(&peer->transmit_packet_work) &peer->endpoint_lock rcu_read_lock_bh rcu_read_lock &____s->seqcount#10 irq_context: 0 (wq_completion)wg-crypt-wg1 (work_completion)(&peer->transmit_packet_work) &peer->endpoint_lock rcu_read_lock_bh rcu_read_lock rcu_read_lock_bh tk_core.seq.seqcount irq_context: 0 (wq_completion)wg-crypt-wg1 (work_completion)(&peer->transmit_packet_work) &peer->endpoint_lock rcu_read_lock_bh rcu_read_lock rcu_read_lock_bh &list->lock#5 irq_context: 0 (wq_completion)wg-crypt-wg0 (work_completion)(&({ do { const void *__vpp_verify = (typeof((worker) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __asm__ ("" : "=r"(__ptr) : "0"((typeof(*((worker))) *)((worker)))); (typeof((typeof(*((worker))) *)((worker)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); })->work) tk_core.seq.seqcount irq_context: softirq rcu_read_lock_bh &base->lock irq_context: 0 (wq_completion)ipv6_addrconf (work_completion)(&(&ifa->dad_work)->work) rtnl_mutex &meta->lock irq_context: 0 (wq_completion)ipv6_addrconf (work_completion)(&(&ifa->dad_work)->work) rtnl_mutex rcu_read_lock rcu_read_lock &tbl->lock &n->lock irq_context: 0 (wq_completion)ipv6_addrconf (work_completion)(&(&ifa->dad_work)->work) rtnl_mutex rcu_read_lock rcu_read_lock &tbl->lock &n->lock irq_context: 0 (wq_completion)ipv6_addrconf (work_completion)(&(&ifa->dad_work)->work) rtnl_mutex rcu_read_lock rcu_read_lock &tbl->lock &n->lock &____s->seqcount#9 irq_context: 0 (wq_completion)ipv6_addrconf (work_completion)(&(&ifa->dad_work)->work) rtnl_mutex rcu_read_lock rcu_read_lock &tbl->lock nl_table_lock irq_context: 0 (wq_completion)ipv6_addrconf (work_completion)(&(&ifa->dad_work)->work) rtnl_mutex rcu_read_lock rcu_read_lock &tbl->lock rlock-AF_NETLINK irq_context: 0 (wq_completion)ipv6_addrconf (work_completion)(&(&ifa->dad_work)->work) rtnl_mutex rcu_read_lock rcu_read_lock &tbl->lock rcu_read_lock &ei->socket.wq.wait irq_context: 0 (wq_completion)ipv6_addrconf (work_completion)(&(&ifa->dad_work)->work) rtnl_mutex rcu_read_lock rcu_read_lock &tbl->lock rcu_read_lock &ei->socket.wq.wait &p->pi_lock irq_context: 0 (wq_completion)ipv6_addrconf (work_completion)(&(&ifa->dad_work)->work) rtnl_mutex rcu_read_lock rcu_read_lock &tbl->lock rcu_read_lock &ei->socket.wq.wait &p->pi_lock &rq->__lock irq_context: 0 (wq_completion)ipv6_addrconf (work_completion)(&(&ifa->dad_work)->work) rtnl_mutex rcu_read_lock rcu_read_lock &tbl->lock rcu_read_lock &ei->socket.wq.wait &p->pi_lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)ipv6_addrconf (work_completion)(&(&ifa->dad_work)->work) rtnl_mutex rcu_read_lock rcu_read_lock &tbl->lock nl_table_wait.lock irq_context: 0 (wq_completion)ipv6_addrconf (work_completion)(&(&ifa->dad_work)->work) rtnl_mutex rcu_read_lock rcu_read_lock &tbl->lock rcu_read_lock lock#8 irq_context: 0 (wq_completion)ipv6_addrconf (work_completion)(&(&ifa->dad_work)->work) rtnl_mutex rcu_read_lock rcu_read_lock &tbl->lock rcu_read_lock id_table_lock irq_context: 0 (wq_completion)ipv6_addrconf (work_completion)(&(&ifa->dad_work)->work) rtnl_mutex rcu_read_lock rcu_read_lock &tbl->lock &dir->lock#2 irq_context: 0 (wq_completion)ipv6_addrconf (work_completion)(&(&ifa->dad_work)->work) rtnl_mutex rcu_read_lock rcu_read_lock &tbl->lock krc.lock irq_context: 0 (wq_completion)wg-kex-wg1#2 irq_context: 0 (wq_completion)wg-kex-wg1#2 (work_completion)(&peer->transmit_handshake_work) irq_context: 0 (wq_completion)wg-kex-wg1#2 (work_completion)(&peer->transmit_handshake_work) tk_core.seq.seqcount irq_context: 0 (wq_completion)wg-kex-wg1#2 (work_completion)(&peer->transmit_handshake_work) &wg->static_identity.lock irq_context: 0 (wq_completion)wg-kex-wg1#2 (work_completion)(&peer->transmit_handshake_work) &wg->static_identity.lock &handshake->lock irq_context: 0 (wq_completion)wg-kex-wg1#2 (work_completion)(&peer->transmit_handshake_work) &wg->static_identity.lock &handshake->lock crngs.lock irq_context: 0 (wq_completion)wg-kex-wg1#2 (work_completion)(&peer->transmit_handshake_work) &wg->static_identity.lock &handshake->lock tk_core.seq.seqcount irq_context: 0 (wq_completion)wg-kex-wg1#2 (work_completion)(&peer->transmit_handshake_work) &wg->static_identity.lock &handshake->lock &table->lock#2 irq_context: 0 (wq_completion)wg-kex-wg1#2 (work_completion)(&peer->transmit_handshake_work) &wg->static_identity.lock &handshake->lock rcu_read_lock_bh batched_entropy_u32.lock irq_context: 0 (wq_completion)wg-kex-wg1#2 (work_completion)(&peer->transmit_handshake_work) &wg->static_identity.lock &handshake->lock rcu_read_lock_bh &table->lock#2 irq_context: 0 (wq_completion)wg-kex-wg1#2 (work_completion)(&peer->transmit_handshake_work) &cookie->lock irq_context: 0 (wq_completion)wg-kex-wg1#2 (work_completion)(&peer->transmit_handshake_work) &cookie->lock irq_context: 0 (wq_completion)wg-kex-wg1#2 (work_completion)(&peer->transmit_handshake_work) rcu_read_lock_bh &obj_hash[i].lock irq_context: 0 (wq_completion)wg-kex-wg1#2 (work_completion)(&peer->transmit_handshake_work) rcu_read_lock_bh &base->lock irq_context: 0 (wq_completion)wg-kex-wg1#2 (work_completion)(&peer->transmit_handshake_work) rcu_read_lock_bh &base->lock &obj_hash[i].lock irq_context: 0 (wq_completion)wg-kex-wg1#2 (work_completion)(&peer->transmit_handshake_work) &obj_hash[i].lock irq_context: 0 (wq_completion)wg-kex-wg1#2 (work_completion)(&peer->transmit_handshake_work) &c->lock irq_context: 0 (wq_completion)wg-kex-wg1#2 (work_completion)(&peer->transmit_handshake_work) &peer->endpoint_lock irq_context: 0 (wq_completion)wg-kex-wg1#2 (work_completion)(&peer->transmit_handshake_work) &peer->endpoint_lock rcu_read_lock_bh rcu_read_lock &dir->lock#2 irq_context: 0 (wq_completion)wg-kex-wg1#2 (work_completion)(&peer->transmit_handshake_work) &peer->endpoint_lock rcu_read_lock_bh rcu_read_lock &ul->lock irq_context: 0 (wq_completion)wg-kex-wg1#2 (work_completion)(&peer->transmit_handshake_work) &peer->endpoint_lock rcu_read_lock_bh batched_entropy_u32.lock irq_context: 0 (wq_completion)wg-kex-wg1#2 (work_completion)(&peer->transmit_handshake_work) &peer->endpoint_lock rcu_read_lock_bh rcu_read_lock &____s->seqcount#7 irq_context: 0 (wq_completion)wg-kex-wg1#2 (work_completion)(&peer->transmit_handshake_work) &peer->endpoint_lock rcu_read_lock_bh rcu_read_lock &nf_conntrack_locks[i] irq_context: 0 (wq_completion)wg-kex-wg1#2 (work_completion)(&peer->transmit_handshake_work) &peer->endpoint_lock rcu_read_lock_bh rcu_read_lock &nf_conntrack_locks[i] &nf_conntrack_locks[i]/1 irq_context: 0 (wq_completion)wg-kex-wg1#2 (work_completion)(&peer->transmit_handshake_work) &peer->endpoint_lock rcu_read_lock_bh rcu_read_lock &nf_conntrack_locks[i] &nf_conntrack_locks[i]/1 batched_entropy_u8.lock irq_context: 0 (wq_completion)wg-kex-wg1#2 (work_completion)(&peer->transmit_handshake_work) &peer->endpoint_lock rcu_read_lock_bh rcu_read_lock &____s->seqcount#10 irq_context: 0 (wq_completion)wg-kex-wg1#2 (work_completion)(&peer->transmit_handshake_work) &peer->endpoint_lock rcu_read_lock_bh rcu_read_lock rcu_read_lock_bh tk_core.seq.seqcount irq_context: 0 (wq_completion)wg-kex-wg1#2 (work_completion)(&peer->transmit_handshake_work) &peer->endpoint_lock rcu_read_lock_bh rcu_read_lock rcu_read_lock_bh &list->lock#5 irq_context: softirq rcu_read_lock rcu_read_lock k-slock-AF_INET pool_lock#2 irq_context: softirq rcu_read_lock rcu_read_lock k-slock-AF_INET &c->lock irq_context: softirq rcu_read_lock rcu_read_lock k-slock-AF_INET batched_entropy_u32.lock irq_context: softirq rcu_read_lock rcu_read_lock k-slock-AF_INET &obj_hash[i].lock irq_context: softirq rcu_read_lock rcu_read_lock k-slock-AF_INET rcu_read_lock rcu_read_lock &____s->seqcount#7 irq_context: softirq rcu_read_lock rcu_read_lock k-slock-AF_INET rcu_read_lock &____s->seqcount#10 irq_context: softirq rcu_read_lock rcu_read_lock k-slock-AF_INET rcu_read_lock rcu_read_lock_bh tk_core.seq.seqcount irq_context: softirq rcu_read_lock rcu_read_lock k-slock-AF_INET rcu_read_lock rcu_read_lock_bh &list->lock#5 irq_context: 0 (wq_completion)wg-kex-wg1#2 (work_completion)(&peer->transmit_handshake_work) batched_entropy_u8.lock irq_context: 0 (wq_completion)bat_events (work_completion)(&(&bat_priv->nc.work)->work) rcu_read_lock &rcu_state.gp_wq &p->pi_lock &cfs_rq->removed.lock irq_context: softirq (&ndev->rs_timer) irq_context: softirq (&ndev->rs_timer) &ndev->lock irq_context: softirq (&ndev->rs_timer) rcu_read_lock &ndev->lock irq_context: softirq (&ndev->rs_timer) pool_lock#2 irq_context: softirq (&ndev->rs_timer) &c->lock irq_context: softirq (&ndev->rs_timer) &____s->seqcount irq_context: softirq (&ndev->rs_timer) &dir->lock#2 irq_context: softirq (&ndev->rs_timer) &ul->lock#2 irq_context: softirq (&ndev->rs_timer) rcu_read_lock rcu_read_lock &____s->seqcount#9 irq_context: softirq (&ndev->rs_timer) rcu_read_lock rcu_read_lock rcu_read_lock_bh rcu_read_lock &____s->seqcount#12 irq_context: softirq (&ndev->rs_timer) rcu_read_lock rcu_read_lock rcu_read_lock_bh rcu_read_lock &ndev->lock irq_context: softirq (&ndev->rs_timer) rcu_read_lock rcu_read_lock rcu_read_lock_bh &obj_hash[i].lock irq_context: softirq (&ndev->rs_timer) rcu_read_lock rcu_read_lock rcu_read_lock_bh pool_lock#2 irq_context: softirq (&ndev->rs_timer) &ndev->lock batched_entropy_u32.lock irq_context: softirq (&ndev->rs_timer) &ndev->lock batched_entropy_u32.lock crngs.lock irq_context: softirq (&ndev->rs_timer) &ndev->lock &obj_hash[i].lock irq_context: softirq (&ndev->rs_timer) &ndev->lock &base->lock irq_context: softirq (&ndev->rs_timer) &ndev->lock &base->lock &obj_hash[i].lock irq_context: 0 (wq_completion)ipv6_addrconf (work_completion)(&(&ifa->dad_work)->work) rtnl_mutex rcu_read_lock rcu_read_lock rcu_read_lock_bh &list->lock#17 irq_context: 0 (wq_completion)ipv6_addrconf (work_completion)(&(&ifa->dad_work)->work) rtnl_mutex rcu_read_lock rcu_read_lock rcu_read_lock_bh rcu_read_lock_bh tk_core.seq.seqcount irq_context: 0 (wq_completion)mld (work_completion)(&(&idev->mc_ifc_work)->work) &idev->mc_lock rcu_read_lock rcu_read_lock rcu_read_lock_bh rcu_read_lock rcu_read_lock &____s->seqcount irq_context: 0 (wq_completion)mld (work_completion)(&(&idev->mc_ifc_work)->work) &idev->mc_lock rcu_read_lock rcu_read_lock rcu_read_lock_bh rcu_read_lock rcu_read_lock rcu_read_lock pool_lock#2 irq_context: 0 (wq_completion)mld (work_completion)(&(&idev->mc_ifc_work)->work) &idev->mc_lock rcu_read_lock rcu_read_lock rcu_read_lock_bh rcu_read_lock rcu_read_lock &____s->seqcount#10 irq_context: softirq rcu_read_lock rcu_read_lock &obj_hash[i].lock pool_lock irq_context: 0 rtnl_mutex &wg->device_update_lock &c->lock irq_context: 0 rtnl_mutex &wg->device_update_lock &obj_hash[i].lock pool_lock irq_context: 0 rtnl_mutex &wg->device_update_lock rcu_state.exp_mutex &rnp->exp_wq[3] irq_context: 0 (wq_completion)mld (work_completion)(&(&idev->mc_dad_work)->work) &idev->mc_lock remove_cache_srcu irq_context: 0 (wq_completion)mld (work_completion)(&(&idev->mc_dad_work)->work) &idev->mc_lock remove_cache_srcu quarantine_lock irq_context: 0 (wq_completion)mld (work_completion)(&(&idev->mc_dad_work)->work) &idev->mc_lock remove_cache_srcu &c->lock irq_context: 0 (wq_completion)mld (work_completion)(&(&idev->mc_dad_work)->work) &idev->mc_lock remove_cache_srcu &n->list_lock irq_context: 0 (wq_completion)mld (work_completion)(&(&idev->mc_dad_work)->work) &idev->mc_lock remove_cache_srcu rcu_read_lock pool_lock#2 irq_context: 0 (wq_completion)mld (work_completion)(&(&idev->mc_dad_work)->work) &idev->mc_lock remove_cache_srcu &obj_hash[i].lock irq_context: 0 (wq_completion)bat_events (work_completion)(&(&bat_priv->nc.work)->work) rcu_read_lock &rcu_state.expedited_wq &p->pi_lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 misc_mtx rfkill_global_mutex irq_context: 0 misc_mtx rfkill_global_mutex &data->mtx irq_context: 0 misc_mtx rfkill_global_mutex &data->mtx fs_reclaim irq_context: 0 misc_mtx rfkill_global_mutex &data->mtx fs_reclaim mmu_notifier_invalidate_range_start irq_context: 0 misc_mtx rfkill_global_mutex &data->mtx pool_lock#2 irq_context: 0 misc_mtx rfkill_global_mutex &data->mtx &rfkill->lock irq_context: 0 cb_lock genl_mutex rcu_read_lock pool_lock#2 irq_context: 0 cb_lock genl_mutex rcu_read_lock rcu_read_lock_bh &obj_hash[i].lock irq_context: 0 cb_lock genl_mutex rcu_read_lock rcu_read_lock_bh pool_lock#2 irq_context: 0 cb_lock rcu_read_lock &c->lock irq_context: 0 cb_lock rcu_read_lock &n->list_lock irq_context: 0 cb_lock rcu_read_lock &n->list_lock &c->lock irq_context: 0 cb_lock rcu_read_lock pool_lock#2 irq_context: 0 cb_lock rcu_read_lock rcu_read_lock_bh &obj_hash[i].lock irq_context: 0 cb_lock rcu_read_lock rcu_read_lock_bh pool_lock#2 irq_context: 0 cb_lock genl_mutex hwsim_radio_lock irq_context: 0 cb_lock genl_mutex &x->wait#9 irq_context: 0 cb_lock genl_mutex &pcp->lock &zone->lock irq_context: 0 cb_lock genl_mutex batched_entropy_u32.lock irq_context: 0 cb_lock genl_mutex &obj_hash[i].lock pool_lock irq_context: 0 cb_lock genl_mutex &k->list_lock irq_context: 0 cb_lock genl_mutex gdp_mutex irq_context: 0 cb_lock genl_mutex gdp_mutex &k->list_lock irq_context: 0 cb_lock genl_mutex lock irq_context: 0 cb_lock genl_mutex lock kernfs_idr_lock irq_context: 0 cb_lock genl_mutex &root->kernfs_rwsem irq_context: 0 cb_lock genl_mutex &root->kernfs_rwsem &root->kernfs_iattr_rwsem irq_context: 0 cb_lock genl_mutex bus_type_sem irq_context: 0 cb_lock genl_mutex sysfs_symlink_target_lock irq_context: 0 cb_lock genl_mutex &root->kernfs_rwsem irq_context: 0 cb_lock genl_mutex &dev->power.lock irq_context: 0 cb_lock genl_mutex dpm_list_mtx irq_context: 0 cb_lock genl_mutex uevent_sock_mutex irq_context: 0 cb_lock genl_mutex uevent_sock_mutex fs_reclaim irq_context: 0 cb_lock genl_mutex uevent_sock_mutex fs_reclaim mmu_notifier_invalidate_range_start irq_context: 0 cb_lock genl_mutex uevent_sock_mutex pool_lock#2 irq_context: 0 cb_lock genl_mutex uevent_sock_mutex nl_table_lock irq_context: 0 cb_lock genl_mutex uevent_sock_mutex rlock-AF_NETLINK irq_context: 0 cb_lock genl_mutex uevent_sock_mutex rcu_read_lock &ei->socket.wq.wait irq_context: 0 cb_lock genl_mutex uevent_sock_mutex rcu_read_lock &ei->socket.wq.wait &ep->lock irq_context: 0 cb_lock genl_mutex uevent_sock_mutex rcu_read_lock &ei->socket.wq.wait &ep->lock &ep->wq irq_context: 0 cb_lock genl_mutex uevent_sock_mutex rcu_read_lock &ei->socket.wq.wait &ep->lock &ep->wq &p->pi_lock irq_context: 0 cb_lock genl_mutex uevent_sock_mutex rcu_read_lock &ei->socket.wq.wait &ep->lock &ep->wq &p->pi_lock &rq->__lock irq_context: 0 cb_lock genl_mutex uevent_sock_mutex rcu_read_lock &ei->socket.wq.wait &ep->lock &ep->wq &p->pi_lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 cb_lock genl_mutex uevent_sock_mutex rcu_read_lock rcu_node_0 irq_context: 0 cb_lock genl_mutex uevent_sock_mutex rcu_read_lock &rq->__lock irq_context: 0 (wq_completion)wg-kex-wg2 irq_context: 0 (wq_completion)wg-kex-wg2 (work_completion)(&peer->transmit_handshake_work) irq_context: 0 (wq_completion)wg-kex-wg2 (work_completion)(&peer->transmit_handshake_work) tk_core.seq.seqcount irq_context: 0 (wq_completion)wg-kex-wg2 (work_completion)(&peer->transmit_handshake_work) &wg->static_identity.lock irq_context: 0 (wq_completion)wg-kex-wg2 (work_completion)(&peer->transmit_handshake_work) &wg->static_identity.lock &handshake->lock irq_context: 0 (wq_completion)wg-kex-wg2 (work_completion)(&peer->transmit_handshake_work) &wg->static_identity.lock &handshake->lock crngs.lock irq_context: 0 (wq_completion)wg-kex-wg2 (work_completion)(&peer->transmit_handshake_work) &wg->static_identity.lock &handshake->lock tk_core.seq.seqcount irq_context: 0 (wq_completion)wg-kex-wg2 (work_completion)(&peer->transmit_handshake_work) &wg->static_identity.lock &handshake->lock &table->lock#2 irq_context: 0 (wq_completion)wg-kex-wg2 (work_completion)(&peer->transmit_handshake_work) &wg->static_identity.lock &handshake->lock rcu_read_lock_bh batched_entropy_u32.lock irq_context: 0 (wq_completion)wg-kex-wg2 (work_completion)(&peer->transmit_handshake_work) &wg->static_identity.lock &handshake->lock rcu_read_lock_bh &table->lock#2 irq_context: 0 (wq_completion)wg-kex-wg2 (work_completion)(&peer->transmit_handshake_work) &cookie->lock irq_context: 0 (wq_completion)wg-kex-wg2 (work_completion)(&peer->transmit_handshake_work) &cookie->lock irq_context: 0 (wq_completion)wg-kex-wg2 (work_completion)(&peer->transmit_handshake_work) rcu_read_lock_bh &obj_hash[i].lock irq_context: 0 (wq_completion)wg-kex-wg2 (work_completion)(&peer->transmit_handshake_work) rcu_read_lock_bh &base->lock irq_context: 0 (wq_completion)wg-kex-wg2 (work_completion)(&peer->transmit_handshake_work) rcu_read_lock_bh &base->lock &obj_hash[i].lock irq_context: 0 (wq_completion)wg-kex-wg2 (work_completion)(&peer->transmit_handshake_work) &obj_hash[i].lock irq_context: 0 (wq_completion)wg-kex-wg2 (work_completion)(&peer->transmit_handshake_work) &peer->endpoint_lock irq_context: 0 (wq_completion)wg-kex-wg2 (work_completion)(&peer->transmit_handshake_work) &peer->endpoint_lock rcu_read_lock_bh rcu_read_lock &____s->seqcount#7 irq_context: 0 (wq_completion)wg-kex-wg2 (work_completion)(&peer->transmit_handshake_work) &peer->endpoint_lock rcu_read_lock_bh rcu_read_lock &____s->seqcount#10 irq_context: 0 (wq_completion)wg-kex-wg2 (work_completion)(&peer->transmit_handshake_work) &peer->endpoint_lock rcu_read_lock_bh rcu_read_lock rcu_read_lock_bh tk_core.seq.seqcount irq_context: 0 (wq_completion)wg-kex-wg2 (work_completion)(&peer->transmit_handshake_work) &peer->endpoint_lock rcu_read_lock_bh rcu_read_lock rcu_read_lock_bh &list->lock#5 irq_context: 0 (wq_completion)wg-kex-wg2 (work_completion)(&peer->transmit_handshake_work) batched_entropy_u8.lock irq_context: 0 (wq_completion)wg-kex-wg2 (work_completion)(&peer->transmit_handshake_work) &peer->endpoint_lock rcu_read_lock_bh rcu_read_lock &dir->lock#2 irq_context: 0 (wq_completion)wg-kex-wg2 (work_completion)(&peer->transmit_handshake_work) &peer->endpoint_lock rcu_read_lock_bh rcu_read_lock &ul->lock irq_context: 0 (wq_completion)wg-kex-wg2 (work_completion)(&peer->transmit_handshake_work) &peer->endpoint_lock rcu_read_lock_bh batched_entropy_u32.lock irq_context: 0 (wq_completion)wg-kex-wg0#2 (work_completion)(&({ do { const void *__vpp_verify = (typeof((worker) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __asm__ ("" : "=r"(__ptr) : "0"((typeof(*((worker))) *)((worker)))); (typeof((typeof(*((worker))) *)((worker)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); })->work) &wg->static_identity.lock &handshake->lock tk_core.seq.seqcount irq_context: 0 (wq_completion)wg-kex-wg0#2 (work_completion)(&({ do { const void *__vpp_verify = (typeof((worker) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __asm__ ("" : "=r"(__ptr) : "0"((typeof(*((worker))) *)((worker)))); (typeof((typeof(*((worker))) *)((worker)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); })->work) &wg->static_identity.lock &handshake->lock tk_core.seq.seqcount irq_context: 0 (wq_completion)wg-kex-wg0#2 (work_completion)(&({ do { const void *__vpp_verify = (typeof((worker) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __asm__ ("" : "=r"(__ptr) : "0"((typeof(*((worker))) *)((worker)))); (typeof((typeof(*((worker))) *)((worker)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); })->work) &wg->static_identity.lock &handshake->lock crngs.lock irq_context: 0 (wq_completion)wg-kex-wg0#2 (work_completion)(&({ do { const void *__vpp_verify = (typeof((worker) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __asm__ ("" : "=r"(__ptr) : "0"((typeof(*((worker))) *)((worker)))); (typeof((typeof(*((worker))) *)((worker)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); })->work) &wg->static_identity.lock &handshake->lock &table->lock#2 irq_context: 0 (wq_completion)wg-kex-wg0#2 (work_completion)(&({ do { const void *__vpp_verify = (typeof((worker) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __asm__ ("" : "=r"(__ptr) : "0"((typeof(*((worker))) *)((worker)))); (typeof((typeof(*((worker))) *)((worker)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); })->work) &wg->static_identity.lock &handshake->lock rcu_read_lock_bh batched_entropy_u32.lock irq_context: 0 (wq_completion)wg-kex-wg0#2 (work_completion)(&({ do { const void *__vpp_verify = (typeof((worker) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __asm__ ("" : "=r"(__ptr) : "0"((typeof(*((worker))) *)((worker)))); (typeof((typeof(*((worker))) *)((worker)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); })->work) &wg->static_identity.lock &handshake->lock rcu_read_lock_bh &table->lock#2 irq_context: 0 (wq_completion)wg-kex-wg0#2 (work_completion)(&({ do { const void *__vpp_verify = (typeof((worker) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __asm__ ("" : "=r"(__ptr) : "0"((typeof(*((worker))) *)((worker)))); (typeof((typeof(*((worker))) *)((worker)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); })->work) &cookie->lock irq_context: 0 (wq_completion)wg-kex-wg0#2 (work_completion)(&({ do { const void *__vpp_verify = (typeof((worker) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __asm__ ("" : "=r"(__ptr) : "0"((typeof(*((worker))) *)((worker)))); (typeof((typeof(*((worker))) *)((worker)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); })->work) &cookie->lock irq_context: 0 (wq_completion)wg-kex-wg0#2 (work_completion)(&({ do { const void *__vpp_verify = (typeof((worker) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __asm__ ("" : "=r"(__ptr) : "0"((typeof(*((worker))) *)((worker)))); (typeof((typeof(*((worker))) *)((worker)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); })->work) &peer->endpoint_lock irq_context: 0 (wq_completion)wg-kex-wg0#2 (work_completion)(&({ do { const void *__vpp_verify = (typeof((worker) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __asm__ ("" : "=r"(__ptr) : "0"((typeof(*((worker))) *)((worker)))); (typeof((typeof(*((worker))) *)((worker)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); })->work) &peer->endpoint_lock rcu_read_lock_bh rcu_read_lock &____s->seqcount#7 irq_context: 0 (wq_completion)wg-kex-wg0#2 (work_completion)(&({ do { const void *__vpp_verify = (typeof((worker) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __asm__ ("" : "=r"(__ptr) : "0"((typeof(*((worker))) *)((worker)))); (typeof((typeof(*((worker))) *)((worker)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); })->work) &peer->endpoint_lock rcu_read_lock_bh rcu_read_lock &____s->seqcount#10 irq_context: 0 (wq_completion)wg-kex-wg0#2 (work_completion)(&({ do { const void *__vpp_verify = (typeof((worker) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __asm__ ("" : "=r"(__ptr) : "0"((typeof(*((worker))) *)((worker)))); (typeof((typeof(*((worker))) *)((worker)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); })->work) &peer->endpoint_lock rcu_read_lock_bh rcu_read_lock rcu_read_lock_bh tk_core.seq.seqcount irq_context: 0 (wq_completion)wg-kex-wg0#2 (work_completion)(&({ do { const void *__vpp_verify = (typeof((worker) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __asm__ ("" : "=r"(__ptr) : "0"((typeof(*((worker))) *)((worker)))); (typeof((typeof(*((worker))) *)((worker)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); })->work) &peer->endpoint_lock rcu_read_lock_bh rcu_read_lock rcu_read_lock_bh &list->lock#5 irq_context: 0 (wq_completion)wg-kex-wg1 (work_completion)(&({ do { const void *__vpp_verify = (typeof((worker) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __asm__ ("" : "=r"(__ptr) : "0"((typeof(*((worker))) *)((worker)))); (typeof((typeof(*((worker))) *)((worker)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); })->work) &peer->endpoint_lock rcu_read_lock_bh &obj_hash[i].lock irq_context: 0 (wq_completion)wg-kex-wg1 (work_completion)(&({ do { const void *__vpp_verify = (typeof((worker) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __asm__ ("" : "=r"(__ptr) : "0"((typeof(*((worker))) *)((worker)))); (typeof((typeof(*((worker))) *)((worker)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); })->work) &peer->endpoint_lock rcu_read_lock_bh pool_lock#2 irq_context: 0 (wq_completion)wg-kex-wg2#2 irq_context: 0 (wq_completion)wg-kex-wg2#2 (work_completion)(&({ do { const void *__vpp_verify = (typeof((worker) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __asm__ ("" : "=r"(__ptr) : "0"((typeof(*((worker))) *)((worker)))); (typeof((typeof(*((worker))) *)((worker)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); })->work) irq_context: 0 (wq_completion)wg-kex-wg2#2 (work_completion)(&({ do { const void *__vpp_verify = (typeof((worker) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __asm__ ("" : "=r"(__ptr) : "0"((typeof(*((worker))) *)((worker)))); (typeof((typeof(*((worker))) *)((worker)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); })->work) &r->consumer_lock#2 irq_context: 0 (wq_completion)wg-kex-wg2#2 (work_completion)(&({ do { const void *__vpp_verify = (typeof((worker) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __asm__ ("" : "=r"(__ptr) : "0"((typeof(*((worker))) *)((worker)))); (typeof((typeof(*((worker))) *)((worker)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); })->work) &wg->static_identity.lock irq_context: 0 (wq_completion)wg-kex-wg2#2 (work_completion)(&({ do { const void *__vpp_verify = (typeof((worker) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __asm__ ("" : "=r"(__ptr) : "0"((typeof(*((worker))) *)((worker)))); (typeof((typeof(*((worker))) *)((worker)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); })->work) &wg->static_identity.lock &handshake->lock irq_context: 0 (wq_completion)wg-kex-wg2#2 (work_completion)(&({ do { const void *__vpp_verify = (typeof((worker) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __asm__ ("" : "=r"(__ptr) : "0"((typeof(*((worker))) *)((worker)))); (typeof((typeof(*((worker))) *)((worker)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); })->work) &wg->static_identity.lock &handshake->lock irq_context: 0 (wq_completion)wg-kex-wg2#2 (work_completion)(&({ do { const void *__vpp_verify = (typeof((worker) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __asm__ ("" : "=r"(__ptr) : "0"((typeof(*((worker))) *)((worker)))); (typeof((typeof(*((worker))) *)((worker)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); })->work) &peer->endpoint_lock irq_context: 0 (wq_completion)wg-kex-wg2#2 (work_completion)(&({ do { const void *__vpp_verify = (typeof((worker) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __asm__ ("" : "=r"(__ptr) : "0"((typeof(*((worker))) *)((worker)))); (typeof((typeof(*((worker))) *)((worker)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); })->work) &handshake->lock irq_context: 0 (wq_completion)wg-kex-wg2#2 (work_completion)(&({ do { const void *__vpp_verify = (typeof((worker) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __asm__ ("" : "=r"(__ptr) : "0"((typeof(*((worker))) *)((worker)))); (typeof((typeof(*((worker))) *)((worker)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); })->work) &handshake->lock fs_reclaim irq_context: 0 (wq_completion)wg-kex-wg2#2 (work_completion)(&({ do { const void *__vpp_verify = (typeof((worker) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __asm__ ("" : "=r"(__ptr) : "0"((typeof(*((worker))) *)((worker)))); (typeof((typeof(*((worker))) *)((worker)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); })->work) &handshake->lock fs_reclaim mmu_notifier_invalidate_range_start irq_context: 0 (wq_completion)wg-kex-wg2#2 (work_completion)(&({ do { const void *__vpp_verify = (typeof((worker) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __asm__ ("" : "=r"(__ptr) : "0"((typeof(*((worker))) *)((worker)))); (typeof((typeof(*((worker))) *)((worker)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); })->work) &handshake->lock tk_core.seq.seqcount irq_context: 0 (wq_completion)wg-kex-wg2#2 (work_completion)(&({ do { const void *__vpp_verify = (typeof((worker) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __asm__ ("" : "=r"(__ptr) : "0"((typeof(*((worker))) *)((worker)))); (typeof((typeof(*((worker))) *)((worker)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); })->work) &handshake->lock rcu_read_lock_bh &peer->keypairs.keypair_update_lock irq_context: 0 (wq_completion)wg-kex-wg2#2 (work_completion)(&({ do { const void *__vpp_verify = (typeof((worker) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __asm__ ("" : "=r"(__ptr) : "0"((typeof(*((worker))) *)((worker)))); (typeof((typeof(*((worker))) *)((worker)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); })->work) &handshake->lock rcu_read_lock_bh &table->lock#2 irq_context: 0 (wq_completion)wg-kex-wg2#2 (work_completion)(&({ do { const void *__vpp_verify = (typeof((worker) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __asm__ ("" : "=r"(__ptr) : "0"((typeof(*((worker))) *)((worker)))); (typeof((typeof(*((worker))) *)((worker)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); })->work) rcu_read_lock_bh &obj_hash[i].lock irq_context: 0 (wq_completion)wg-kex-wg2#2 (work_completion)(&({ do { const void *__vpp_verify = (typeof((worker) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __asm__ ("" : "=r"(__ptr) : "0"((typeof(*((worker))) *)((worker)))); (typeof((typeof(*((worker))) *)((worker)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); })->work) rcu_read_lock_bh &base->lock irq_context: 0 (wq_completion)wg-kex-wg2#2 (work_completion)(&({ do { const void *__vpp_verify = (typeof((worker) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __asm__ ("" : "=r"(__ptr) : "0"((typeof(*((worker))) *)((worker)))); (typeof((typeof(*((worker))) *)((worker)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); })->work) rcu_read_lock_bh &base->lock &obj_hash[i].lock irq_context: 0 (wq_completion)wg-kex-wg2#2 (work_completion)(&({ do { const void *__vpp_verify = (typeof((worker) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __asm__ ("" : "=r"(__ptr) : "0"((typeof(*((worker))) *)((worker)))); (typeof((typeof(*((worker))) *)((worker)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); })->work) &obj_hash[i].lock irq_context: 0 (wq_completion)wg-kex-wg2#2 (work_completion)(&({ do { const void *__vpp_verify = (typeof((worker) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __asm__ ("" : "=r"(__ptr) : "0"((typeof(*((worker))) *)((worker)))); (typeof((typeof(*((worker))) *)((worker)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); })->work) &base->lock irq_context: 0 (wq_completion)wg-kex-wg2#2 (work_completion)(&({ do { const void *__vpp_verify = (typeof((worker) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __asm__ ("" : "=r"(__ptr) : "0"((typeof(*((worker))) *)((worker)))); (typeof((typeof(*((worker))) *)((worker)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); })->work) &base->lock &obj_hash[i].lock irq_context: 0 (wq_completion)wg-kex-wg2#2 (work_completion)(&({ do { const void *__vpp_verify = (typeof((worker) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __asm__ ("" : "=r"(__ptr) : "0"((typeof(*((worker))) *)((worker)))); (typeof((typeof(*((worker))) *)((worker)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); })->work) tk_core.seq.seqcount irq_context: 0 (wq_completion)wg-kex-wg2#2 (work_completion)(&({ do { const void *__vpp_verify = (typeof((worker) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __asm__ ("" : "=r"(__ptr) : "0"((typeof(*((worker))) *)((worker)))); (typeof((typeof(*((worker))) *)((worker)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); })->work) &list->lock#17 irq_context: 0 (wq_completion)wg-kex-wg2#2 (work_completion)(&({ do { const void *__vpp_verify = (typeof((worker) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __asm__ ("" : "=r"(__ptr) : "0"((typeof(*((worker))) *)((worker)))); (typeof((typeof(*((worker))) *)((worker)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); })->work) rcu_read_lock_bh &r->producer_lock#2 irq_context: 0 (wq_completion)wg-kex-wg2#2 (work_completion)(&({ do { const void *__vpp_verify = (typeof((worker) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __asm__ ("" : "=r"(__ptr) : "0"((typeof(*((worker))) *)((worker)))); (typeof((typeof(*((worker))) *)((worker)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); })->work) rcu_read_lock_bh rcu_read_lock &pool->lock irq_context: 0 (wq_completion)wg-kex-wg2#2 (work_completion)(&({ do { const void *__vpp_verify = (typeof((worker) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __asm__ ("" : "=r"(__ptr) : "0"((typeof(*((worker))) *)((worker)))); (typeof((typeof(*((worker))) *)((worker)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); })->work) rcu_read_lock_bh rcu_read_lock &pool->lock &obj_hash[i].lock irq_context: 0 (wq_completion)wg-kex-wg2#2 (work_completion)(&({ do { const void *__vpp_verify = (typeof((worker) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __asm__ ("" : "=r"(__ptr) : "0"((typeof(*((worker))) *)((worker)))); (typeof((typeof(*((worker))) *)((worker)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); })->work) rcu_read_lock_bh rcu_read_lock &pool->lock &p->pi_lock irq_context: 0 (wq_completion)wg-kex-wg2#2 (work_completion)(&({ do { const void *__vpp_verify = (typeof((worker) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __asm__ ("" : "=r"(__ptr) : "0"((typeof(*((worker))) *)((worker)))); (typeof((typeof(*((worker))) *)((worker)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); })->work) &handshake->lock &c->lock irq_context: 0 (wq_completion)wg-kex-wg2#2 (work_completion)(&({ do { const void *__vpp_verify = (typeof((worker) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __asm__ ("" : "=r"(__ptr) : "0"((typeof(*((worker))) *)((worker)))); (typeof((typeof(*((worker))) *)((worker)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); })->work) &handshake->lock &____s->seqcount irq_context: 0 (wq_completion)wg-crypt-wg2 irq_context: 0 (wq_completion)wg-crypt-wg2 (work_completion)(&({ do { const void *__vpp_verify = (typeof((worker) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __asm__ ("" : "=r"(__ptr) : "0"((typeof(*((worker))) *)((worker)))); (typeof((typeof(*((worker))) *)((worker)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); })->work) irq_context: 0 (wq_completion)wg-crypt-wg2 (work_completion)(&({ do { const void *__vpp_verify = (typeof((worker) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __asm__ ("" : "=r"(__ptr) : "0"((typeof(*((worker))) *)((worker)))); (typeof((typeof(*((worker))) *)((worker)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); })->work) &r->consumer_lock#2 irq_context: 0 (wq_completion)wg-crypt-wg2 (work_completion)(&({ do { const void *__vpp_verify = (typeof((worker) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __asm__ ("" : "=r"(__ptr) : "0"((typeof(*((worker))) *)((worker)))); (typeof((typeof(*((worker))) *)((worker)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); })->work) rcu_read_lock &pool->lock irq_context: 0 (wq_completion)wg-crypt-wg2 (work_completion)(&({ do { const void *__vpp_verify = (typeof((worker) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __asm__ ("" : "=r"(__ptr) : "0"((typeof(*((worker))) *)((worker)))); (typeof((typeof(*((worker))) *)((worker)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); })->work) rcu_read_lock &pool->lock &obj_hash[i].lock irq_context: 0 (wq_completion)wg-crypt-wg2 (work_completion)(&({ do { const void *__vpp_verify = (typeof((worker) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __asm__ ("" : "=r"(__ptr) : "0"((typeof(*((worker))) *)((worker)))); (typeof((typeof(*((worker))) *)((worker)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); })->work) rcu_read_lock &pool->lock &p->pi_lock irq_context: 0 (wq_completion)wg-crypt-wg2 (work_completion)(&peer->transmit_packet_work) irq_context: 0 (wq_completion)wg-crypt-wg2 (work_completion)(&peer->transmit_packet_work) rcu_read_lock_bh &obj_hash[i].lock irq_context: 0 (wq_completion)wg-crypt-wg2 (work_completion)(&peer->transmit_packet_work) &obj_hash[i].lock irq_context: 0 (wq_completion)wg-crypt-wg2 (work_completion)(&peer->transmit_packet_work) &peer->endpoint_lock irq_context: 0 (wq_completion)wg-crypt-wg2 (work_completion)(&peer->transmit_packet_work) &peer->endpoint_lock rcu_read_lock_bh rcu_read_lock &____s->seqcount#7 irq_context: 0 (wq_completion)wg-crypt-wg2 (work_completion)(&peer->transmit_packet_work) &peer->endpoint_lock rcu_read_lock_bh rcu_read_lock &____s->seqcount#10 irq_context: 0 (wq_completion)wg-crypt-wg2 (work_completion)(&peer->transmit_packet_work) &peer->endpoint_lock rcu_read_lock_bh rcu_read_lock rcu_read_lock_bh tk_core.seq.seqcount irq_context: 0 (wq_completion)wg-crypt-wg2 (work_completion)(&peer->transmit_packet_work) &peer->endpoint_lock rcu_read_lock_bh rcu_read_lock rcu_read_lock_bh &list->lock#5 irq_context: 0 (wq_completion)wg-crypt-wg2 (work_completion)(&peer->transmit_packet_work) rcu_read_lock_bh tk_core.seq.seqcount irq_context: softirq rcu_read_lock_bh &r->producer_lock#2 irq_context: softirq rcu_read_lock_bh rcu_read_lock &pool->lock irq_context: softirq rcu_read_lock_bh rcu_read_lock &pool->lock &obj_hash[i].lock irq_context: softirq rcu_read_lock_bh rcu_read_lock &pool->lock &p->pi_lock irq_context: 0 (wq_completion)wg-crypt-wg2 (work_completion)(&peer->transmit_packet_work) &peer->endpoint_lock rcu_read_lock_bh rcu_read_lock &dir->lock#2 irq_context: 0 (wq_completion)wg-crypt-wg2 (work_completion)(&peer->transmit_packet_work) &peer->endpoint_lock rcu_read_lock_bh rcu_read_lock &ul->lock irq_context: 0 (wq_completion)wg-crypt-wg0 (work_completion)(&peer->transmit_packet_work) &peer->endpoint_lock rcu_read_lock_bh rcu_read_lock rcu_read_lock rcu_read_lock pool_lock#2 irq_context: softirq rcu_read_lock_bh rcu_read_lock &pool->lock &p->pi_lock &rq->__lock irq_context: 0 (wq_completion)wg-crypt-wg0 (work_completion)(&peer->transmit_packet_work) &peer->endpoint_lock rcu_read_lock_bh rcu_read_lock rcu_read_lock rcu_read_lock &dir->lock#2 irq_context: softirq rcu_read_lock_bh rcu_read_lock &pool->lock &p->pi_lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)wg-crypt-wg0 (work_completion)(&peer->transmit_packet_work) batched_entropy_u8.lock irq_context: 0 (wq_completion)wg-crypt-wg0 (work_completion)(&peer->transmit_packet_work) rcu_read_lock_bh &base->lock irq_context: 0 (wq_completion)wg-crypt-wg0 (work_completion)(&peer->transmit_packet_work) rcu_read_lock_bh &base->lock &obj_hash[i].lock irq_context: 0 (wq_completion)wg-crypt-wg2 (work_completion)(&({ do { const void *__vpp_verify = (typeof((worker) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __asm__ ("" : "=r"(__ptr) : "0"((typeof(*((worker))) *)((worker)))); (typeof((typeof(*((worker))) *)((worker)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); })->work) tk_core.seq.seqcount irq_context: softirq rcu_read_lock_bh &base->lock &obj_hash[i].lock irq_context: 0 (wq_completion)wg-crypt-wg1 (work_completion)(&peer->transmit_packet_work) batched_entropy_u8.lock irq_context: 0 (wq_completion)wg-crypt-wg1 (work_completion)(&peer->transmit_packet_work) rcu_read_lock_bh &base->lock &obj_hash[i].lock irq_context: 0 (wq_completion)ipv6_addrconf (work_completion)(&(&ifa->dad_work)->work) rtnl_mutex rcu_read_lock rcu_read_lock rcu_read_lock_bh tk_core.seq.seqcount irq_context: 0 (wq_completion)ipv6_addrconf (work_completion)(&(&ifa->dad_work)->work) rtnl_mutex rcu_read_lock rcu_read_lock rcu_read_lock_bh rcu_read_lock_bh &r->producer_lock#2 irq_context: 0 (wq_completion)ipv6_addrconf (work_completion)(&(&ifa->dad_work)->work) rtnl_mutex rcu_read_lock rcu_read_lock rcu_read_lock_bh rcu_read_lock_bh rcu_read_lock &pool->lock irq_context: 0 (wq_completion)ipv6_addrconf (work_completion)(&(&ifa->dad_work)->work) rtnl_mutex rcu_read_lock rcu_read_lock rcu_read_lock_bh rcu_read_lock_bh rcu_read_lock &pool->lock &obj_hash[i].lock irq_context: 0 (wq_completion)wg-crypt-wg2 (work_completion)(&peer->transmit_packet_work) &base->lock irq_context: 0 (wq_completion)wg-crypt-wg2 (work_completion)(&peer->transmit_packet_work) &base->lock &obj_hash[i].lock irq_context: 0 (wq_completion)wg-crypt-wg2 (work_completion)(&peer->transmit_packet_work) batched_entropy_u8.lock irq_context: 0 (wq_completion)wg-crypt-wg2 (work_completion)(&peer->transmit_packet_work) rcu_read_lock_bh &base->lock irq_context: 0 (wq_completion)wg-crypt-wg2 (work_completion)(&peer->transmit_packet_work) rcu_read_lock_bh &base->lock &obj_hash[i].lock irq_context: 0 (wq_completion)ipv6_addrconf (work_completion)(&(&ifa->dad_work)->work) rtnl_mutex &idev->mc_lock rcu_read_lock rcu_read_lock rcu_read_lock_bh key#20 irq_context: 0 (wq_completion)ipv6_addrconf (work_completion)(&(&ifa->dad_work)->work) rtnl_mutex &idev->mc_lock rcu_read_lock rcu_read_lock rcu_read_lock_bh &entry->crc_lock irq_context: 0 (wq_completion)ipv6_addrconf (work_completion)(&(&ifa->dad_work)->work) rtnl_mutex &idev->mc_lock rcu_read_lock rcu_read_lock rcu_read_lock_bh tk_core.seq.seqcount irq_context: 0 (wq_completion)ipv6_addrconf (work_completion)(&(&ifa->dad_work)->work) rtnl_mutex &idev->mc_lock rcu_read_lock rcu_read_lock rcu_read_lock_bh &list->lock#5 irq_context: 0 (wq_completion)ipv6_addrconf (work_completion)(&(&ifa->dad_work)->work) rtnl_mutex &idev->mc_lock rcu_read_lock rcu_read_lock rcu_read_lock_bh &hsr->seqnr_lock irq_context: 0 (wq_completion)ipv6_addrconf (work_completion)(&(&ifa->dad_work)->work) rtnl_mutex &idev->mc_lock rcu_read_lock rcu_read_lock rcu_read_lock_bh &hsr->seqnr_lock rcu_read_lock &new_node->seq_out_lock irq_context: 0 (wq_completion)ipv6_addrconf (work_completion)(&(&ifa->dad_work)->work) rtnl_mutex &idev->mc_lock rcu_read_lock rcu_read_lock rcu_read_lock_bh &hsr->seqnr_lock rcu_read_lock pool_lock#2 irq_context: 0 (wq_completion)ipv6_addrconf (work_completion)(&(&ifa->dad_work)->work) rtnl_mutex &idev->mc_lock rcu_read_lock rcu_read_lock rcu_read_lock_bh &hsr->seqnr_lock rcu_read_lock rcu_read_lock_bh rcu_read_lock tk_core.seq.seqcount irq_context: 0 (wq_completion)ipv6_addrconf (work_completion)(&(&ifa->dad_work)->work) rtnl_mutex &idev->mc_lock rcu_read_lock rcu_read_lock rcu_read_lock_bh &hsr->seqnr_lock rcu_read_lock rcu_read_lock_bh rcu_read_lock &list->lock#5 irq_context: 0 (wq_completion)ipv6_addrconf (work_completion)(&(&ifa->dad_work)->work) rtnl_mutex &idev->mc_lock rcu_read_lock rcu_read_lock rcu_read_lock_bh &hsr->seqnr_lock rcu_read_lock &c->lock irq_context: 0 (wq_completion)ipv6_addrconf (work_completion)(&(&ifa->dad_work)->work) rtnl_mutex &idev->mc_lock rcu_read_lock rcu_read_lock rcu_read_lock_bh &hsr->seqnr_lock rcu_read_lock &n->list_lock irq_context: 0 (wq_completion)ipv6_addrconf (work_completion)(&(&ifa->dad_work)->work) rtnl_mutex &idev->mc_lock rcu_read_lock rcu_read_lock rcu_read_lock_bh &hsr->seqnr_lock rcu_read_lock &n->list_lock &c->lock irq_context: 0 (wq_completion)ipv6_addrconf (work_completion)(&(&ifa->dad_work)->work) rtnl_mutex &idev->mc_lock rcu_read_lock rcu_read_lock rcu_read_lock_bh &hsr->seqnr_lock &obj_hash[i].lock irq_context: 0 (wq_completion)ipv6_addrconf (work_completion)(&(&ifa->dad_work)->work) rtnl_mutex &idev->mc_lock rcu_read_lock rcu_read_lock rcu_read_lock_bh &hsr->seqnr_lock pool_lock#2 irq_context: 0 (wq_completion)ipv6_addrconf (work_completion)(&(&ifa->dad_work)->work) rtnl_mutex &tb->tb6_lock &pcp->lock &zone->lock irq_context: 0 (wq_completion)ipv6_addrconf (work_completion)(&(&ifa->dad_work)->work) rtnl_mutex &idev->mc_lock rcu_read_lock rcu_read_lock rcu_read_lock_bh dev->qdisc_tx_busylock ?: &qdisc_tx_busylock &obj_hash[i].lock pool_lock irq_context: 0 rtnl_mutex dev_addr_sem rcu_read_lock &obj_hash[i].lock pool_lock irq_context: 0 cb_lock genl_mutex uevent_sock_mutex nl_table_wait.lock irq_context: 0 cb_lock genl_mutex uevent_sock_mutex &obj_hash[i].lock irq_context: 0 cb_lock genl_mutex subsys mutex#53 irq_context: 0 cb_lock genl_mutex subsys mutex#53 &k->k_lock irq_context: 0 cb_lock genl_mutex device_links_lock irq_context: 0 cb_lock genl_mutex &k->k_lock irq_context: 0 cb_lock genl_mutex deferred_probe_mutex irq_context: 0 cb_lock genl_mutex rcu_read_lock &pool->lock/1 irq_context: 0 cb_lock genl_mutex rcu_read_lock &pool->lock/1 &obj_hash[i].lock irq_context: 0 cb_lock genl_mutex rcu_read_lock &pool->lock/1 &p->pi_lock irq_context: 0 cb_lock genl_mutex uevent_sock_mutex &____s->seqcount irq_context: 0 cb_lock genl_mutex uevent_sock_mutex rcu_read_lock pool_lock#2 irq_context: 0 cb_lock genl_mutex cpu_hotplug_lock irq_context: 0 cb_lock genl_mutex cpu_hotplug_lock wq_pool_mutex irq_context: 0 cb_lock genl_mutex cpu_hotplug_lock wq_pool_mutex fs_reclaim irq_context: 0 cb_lock genl_mutex cpu_hotplug_lock wq_pool_mutex fs_reclaim mmu_notifier_invalidate_range_start irq_context: 0 cb_lock genl_mutex cpu_hotplug_lock wq_pool_mutex pool_lock#2 irq_context: 0 cb_lock genl_mutex cpu_hotplug_lock wq_pool_mutex &obj_hash[i].lock irq_context: 0 cb_lock genl_mutex cpu_hotplug_lock wq_pool_mutex &wq->mutex irq_context: 0 cb_lock genl_mutex cpu_hotplug_lock wq_pool_mutex &wq->mutex &pool->lock/1 irq_context: 0 cb_lock genl_mutex wq_pool_mutex irq_context: 0 cb_lock genl_mutex wq_pool_mutex &wq->mutex irq_context: 0 cb_lock genl_mutex crngs.lock irq_context: 0 cb_lock genl_mutex triggers_list_lock irq_context: 0 cb_lock genl_mutex leds_list_lock irq_context: 0 cb_lock genl_mutex leds_list_lock &led_cdev->trigger_lock irq_context: 0 cb_lock genl_mutex &zone->lock irq_context: 0 (wq_completion)bat_events (work_completion)(&(&forw_packet_aggr->delayed_work)->work) &hard_iface->bat_iv.ogm_buff_mutex &bat_priv->tt.commit_lock irq_context: 0 (wq_completion)bat_events (work_completion)(&(&forw_packet_aggr->delayed_work)->work) &hard_iface->bat_iv.ogm_buff_mutex &bat_priv->tt.commit_lock key#16 irq_context: 0 (wq_completion)bat_events (work_completion)(&(&forw_packet_aggr->delayed_work)->work) &hard_iface->bat_iv.ogm_buff_mutex &bat_priv->tt.commit_lock &bat_priv->softif_vlan_list_lock irq_context: 0 (wq_completion)bat_events (work_completion)(&(&forw_packet_aggr->delayed_work)->work) &hard_iface->bat_iv.ogm_buff_mutex &bat_priv->tt.commit_lock &bat_priv->softif_vlan_list_lock pool_lock#2 irq_context: 0 (wq_completion)bat_events (work_completion)(&(&forw_packet_aggr->delayed_work)->work) &hard_iface->bat_iv.ogm_buff_mutex &bat_priv->tt.commit_lock &bat_priv->tt.changes_list_lock irq_context: 0 (wq_completion)bat_events (work_completion)(&(&forw_packet_aggr->delayed_work)->work) &hard_iface->bat_iv.ogm_buff_mutex &bat_priv->tt.commit_lock &bat_priv->tt.changes_list_lock &obj_hash[i].lock irq_context: 0 (wq_completion)bat_events (work_completion)(&(&forw_packet_aggr->delayed_work)->work) &hard_iface->bat_iv.ogm_buff_mutex &bat_priv->tt.commit_lock &bat_priv->tt.changes_list_lock pool_lock#2 irq_context: 0 (wq_completion)bat_events (work_completion)(&(&forw_packet_aggr->delayed_work)->work) &hard_iface->bat_iv.ogm_buff_mutex &bat_priv->tt.commit_lock &bat_priv->tt.last_changeset_lock irq_context: 0 (wq_completion)bat_events (work_completion)(&(&forw_packet_aggr->delayed_work)->work) &hard_iface->bat_iv.ogm_buff_mutex &bat_priv->tt.commit_lock &bat_priv->tt.last_changeset_lock pool_lock#2 irq_context: 0 (wq_completion)bat_events (work_completion)(&(&forw_packet_aggr->delayed_work)->work) &hard_iface->bat_iv.ogm_buff_mutex &bat_priv->tt.commit_lock pool_lock#2 irq_context: 0 (wq_completion)bat_events (work_completion)(&(&forw_packet_aggr->delayed_work)->work) &hard_iface->bat_iv.ogm_buff_mutex &bat_priv->tt.commit_lock &bat_priv->tvlv.container_list_lock irq_context: 0 (wq_completion)bat_events (work_completion)(&(&forw_packet_aggr->delayed_work)->work) &hard_iface->bat_iv.ogm_buff_mutex &bat_priv->tt.commit_lock &obj_hash[i].lock irq_context: 0 (wq_completion)bat_events (work_completion)(&(&forw_packet_aggr->delayed_work)->work) &hard_iface->bat_iv.ogm_buff_mutex &bat_priv->tvlv.container_list_lock irq_context: 0 (wq_completion)bat_events (work_completion)(&(&forw_packet_aggr->delayed_work)->work) &hard_iface->bat_iv.ogm_buff_mutex &bat_priv->tvlv.container_list_lock pool_lock#2 irq_context: 0 (wq_completion)bat_events (work_completion)(&(&forw_packet_aggr->delayed_work)->work) &hard_iface->bat_iv.ogm_buff_mutex &bat_priv->tvlv.container_list_lock &obj_hash[i].lock irq_context: 0 (wq_completion)bat_events (work_completion)(&(&forw_packet_aggr->delayed_work)->work) &hard_iface->bat_iv.ogm_buff_mutex rcu_read_lock &rq->__lock irq_context: 0 (wq_completion)bat_events (work_completion)(&(&forw_packet_aggr->delayed_work)->work) &hard_iface->bat_iv.ogm_buff_mutex rcu_read_lock &bat_priv->forw_bat_list_lock irq_context: 0 (wq_completion)bat_events (work_completion)(&(&forw_packet_aggr->delayed_work)->work) &hard_iface->bat_iv.ogm_buff_mutex rcu_read_lock batched_entropy_u8.lock irq_context: 0 (wq_completion)bat_events (work_completion)(&(&forw_packet_aggr->delayed_work)->work) &hard_iface->bat_iv.ogm_buff_mutex rcu_read_lock kfence_freelist_lock irq_context: 0 (wq_completion)bat_events (work_completion)(&(&forw_packet_aggr->delayed_work)->work) &hard_iface->bat_iv.ogm_buff_mutex rcu_read_lock pool_lock#2 irq_context: 0 (wq_completion)bat_events (work_completion)(&(&forw_packet_aggr->delayed_work)->work) &hard_iface->bat_iv.ogm_buff_mutex rcu_read_lock &obj_hash[i].lock irq_context: 0 (wq_completion)bat_events (work_completion)(&(&forw_packet_aggr->delayed_work)->work) &hard_iface->bat_iv.ogm_buff_mutex rcu_read_lock &bat_priv->forw_bat_list_lock &obj_hash[i].lock irq_context: 0 (wq_completion)bat_events (work_completion)(&(&forw_packet_aggr->delayed_work)->work) &hard_iface->bat_iv.ogm_buff_mutex rcu_read_lock &bat_priv->forw_bat_list_lock &base->lock irq_context: 0 (wq_completion)bat_events (work_completion)(&(&forw_packet_aggr->delayed_work)->work) &hard_iface->bat_iv.ogm_buff_mutex rcu_read_lock &bat_priv->forw_bat_list_lock &base->lock &obj_hash[i].lock irq_context: 0 cb_lock genl_mutex rtnl_mutex fs_reclaim irq_context: 0 cb_lock genl_mutex rtnl_mutex fs_reclaim mmu_notifier_invalidate_range_start irq_context: 0 cb_lock genl_mutex rtnl_mutex pool_lock#2 irq_context: 0 cb_lock genl_mutex rtnl_mutex param_lock irq_context: 0 cb_lock genl_mutex rtnl_mutex param_lock rate_ctrl_mutex irq_context: 0 cb_lock genl_mutex rtnl_mutex (console_sem).lock irq_context: 0 cb_lock genl_mutex rtnl_mutex &rdev->wiphy.mtx irq_context: 0 cb_lock genl_mutex rtnl_mutex &rdev->wiphy.mtx fs_reclaim irq_context: 0 cb_lock genl_mutex rtnl_mutex &rdev->wiphy.mtx fs_reclaim mmu_notifier_invalidate_range_start irq_context: 0 cb_lock genl_mutex rtnl_mutex &rdev->wiphy.mtx pool_lock#2 irq_context: 0 cb_lock genl_mutex rtnl_mutex &rdev->wiphy.mtx &k->list_lock irq_context: 0 cb_lock genl_mutex rtnl_mutex &rdev->wiphy.mtx gdp_mutex irq_context: 0 cb_lock genl_mutex rtnl_mutex &rdev->wiphy.mtx gdp_mutex &k->list_lock irq_context: 0 cb_lock genl_mutex rtnl_mutex &rdev->wiphy.mtx gdp_mutex fs_reclaim irq_context: 0 cb_lock genl_mutex rtnl_mutex &rdev->wiphy.mtx gdp_mutex fs_reclaim mmu_notifier_invalidate_range_start irq_context: 0 cb_lock genl_mutex rtnl_mutex &rdev->wiphy.mtx gdp_mutex pool_lock#2 irq_context: 0 cb_lock genl_mutex rtnl_mutex &rdev->wiphy.mtx gdp_mutex lock irq_context: 0 cb_lock genl_mutex rtnl_mutex &rdev->wiphy.mtx gdp_mutex lock kernfs_idr_lock irq_context: 0 cb_lock genl_mutex rtnl_mutex &rdev->wiphy.mtx gdp_mutex &root->kernfs_rwsem irq_context: 0 cb_lock genl_mutex rtnl_mutex &rdev->wiphy.mtx gdp_mutex &root->kernfs_rwsem &root->kernfs_iattr_rwsem irq_context: 0 cb_lock genl_mutex rtnl_mutex &rdev->wiphy.mtx gdp_mutex kobj_ns_type_lock irq_context: 0 cb_lock genl_mutex rtnl_mutex &rdev->wiphy.mtx lock irq_context: 0 cb_lock genl_mutex rtnl_mutex &rdev->wiphy.mtx lock kernfs_idr_lock irq_context: 0 cb_lock genl_mutex rtnl_mutex &rdev->wiphy.mtx &root->kernfs_rwsem irq_context: 0 cb_lock genl_mutex rtnl_mutex &rdev->wiphy.mtx &root->kernfs_rwsem &root->kernfs_iattr_rwsem irq_context: 0 cb_lock genl_mutex rtnl_mutex &rdev->wiphy.mtx bus_type_sem irq_context: 0 cb_lock genl_mutex rtnl_mutex &rdev->wiphy.mtx sysfs_symlink_target_lock irq_context: 0 cb_lock genl_mutex rtnl_mutex &rdev->wiphy.mtx &c->lock irq_context: 0 cb_lock genl_mutex rtnl_mutex &rdev->wiphy.mtx &____s->seqcount irq_context: 0 cb_lock genl_mutex rtnl_mutex &rdev->wiphy.mtx &root->kernfs_rwsem irq_context: 0 cb_lock genl_mutex rtnl_mutex &rdev->wiphy.mtx &dev->power.lock irq_context: 0 cb_lock genl_mutex rtnl_mutex &rdev->wiphy.mtx dpm_list_mtx irq_context: 0 cb_lock genl_mutex rtnl_mutex &rdev->wiphy.mtx uevent_sock_mutex irq_context: 0 cb_lock genl_mutex rtnl_mutex &rdev->wiphy.mtx uevent_sock_mutex fs_reclaim irq_context: 0 cb_lock genl_mutex rtnl_mutex &rdev->wiphy.mtx uevent_sock_mutex fs_reclaim mmu_notifier_invalidate_range_start irq_context: 0 cb_lock genl_mutex rtnl_mutex &rdev->wiphy.mtx uevent_sock_mutex pool_lock#2 irq_context: 0 cb_lock genl_mutex rtnl_mutex &rdev->wiphy.mtx uevent_sock_mutex &c->lock irq_context: 0 cb_lock genl_mutex rtnl_mutex &rdev->wiphy.mtx uevent_sock_mutex nl_table_lock irq_context: 0 cb_lock genl_mutex rtnl_mutex &rdev->wiphy.mtx uevent_sock_mutex &obj_hash[i].lock irq_context: 0 cb_lock genl_mutex rtnl_mutex &rdev->wiphy.mtx uevent_sock_mutex nl_table_wait.lock irq_context: 0 cb_lock genl_mutex rtnl_mutex &rdev->wiphy.mtx &obj_hash[i].lock irq_context: 0 cb_lock genl_mutex rtnl_mutex &rdev->wiphy.mtx &k->k_lock irq_context: 0 cb_lock genl_mutex rtnl_mutex &rdev->wiphy.mtx subsys mutex#54 irq_context: 0 cb_lock genl_mutex rtnl_mutex &rdev->wiphy.mtx subsys mutex#54 &k->k_lock irq_context: 0 cb_lock genl_mutex rtnl_mutex &rdev->wiphy.mtx pin_fs_lock irq_context: 0 cb_lock genl_mutex rtnl_mutex &rdev->wiphy.mtx &sb->s_type->i_mutex_key#3 irq_context: 0 cb_lock genl_mutex rtnl_mutex &rdev->wiphy.mtx &sb->s_type->i_mutex_key#3 rename_lock.seqcount irq_context: 0 cb_lock genl_mutex rtnl_mutex &rdev->wiphy.mtx &sb->s_type->i_mutex_key#3 fs_reclaim irq_context: 0 cb_lock genl_mutex rtnl_mutex &rdev->wiphy.mtx &sb->s_type->i_mutex_key#3 fs_reclaim mmu_notifier_invalidate_range_start irq_context: 0 cb_lock genl_mutex rtnl_mutex &rdev->wiphy.mtx &sb->s_type->i_mutex_key#3 pool_lock#2 irq_context: 0 cb_lock genl_mutex rtnl_mutex &rdev->wiphy.mtx &sb->s_type->i_mutex_key#3 &dentry->d_lock irq_context: 0 cb_lock genl_mutex rtnl_mutex &rdev->wiphy.mtx &sb->s_type->i_mutex_key#3 rcu_read_lock rename_lock.seqcount irq_context: 0 cb_lock genl_mutex rtnl_mutex &rdev->wiphy.mtx &sb->s_type->i_mutex_key#3 &dentry->d_lock &wq irq_context: 0 cb_lock genl_mutex rtnl_mutex &rdev->wiphy.mtx &sb->s_type->i_mutex_key#3 mmu_notifier_invalidate_range_start irq_context: 0 cb_lock genl_mutex rtnl_mutex &rdev->wiphy.mtx &sb->s_type->i_mutex_key#3 &sb->s_type->i_lock_key#7 irq_context: 0 cb_lock genl_mutex rtnl_mutex &rdev->wiphy.mtx &sb->s_type->i_mutex_key#3 &s->s_inode_list_lock irq_context: 0 cb_lock genl_mutex rtnl_mutex &rdev->wiphy.mtx &sb->s_type->i_mutex_key#3 tk_core.seq.seqcount irq_context: 0 cb_lock genl_mutex rtnl_mutex &rdev->wiphy.mtx &sb->s_type->i_mutex_key#3 &sb->s_type->i_lock_key#7 &dentry->d_lock irq_context: 0 cb_lock genl_mutex rtnl_mutex &rdev->wiphy.mtx nl_table_lock irq_context: 0 cb_lock genl_mutex rtnl_mutex &rdev->wiphy.mtx rcu_read_lock pool_lock#2 irq_context: 0 cb_lock genl_mutex rtnl_mutex &rdev->wiphy.mtx nl_table_wait.lock irq_context: 0 cb_lock genl_mutex rtnl_mutex &rdev->wiphy.mtx reg_requests_lock irq_context: 0 cb_lock genl_mutex rtnl_mutex &obj_hash[i].lock irq_context: 0 cb_lock genl_mutex rtnl_mutex &base->lock irq_context: 0 cb_lock genl_mutex rtnl_mutex &base->lock &obj_hash[i].lock irq_context: 0 cb_lock genl_mutex rfkill_global_mutex irq_context: 0 cb_lock genl_mutex rfkill_global_mutex fs_reclaim irq_context: 0 cb_lock genl_mutex rfkill_global_mutex fs_reclaim mmu_notifier_invalidate_range_start irq_context: 0 cb_lock genl_mutex rfkill_global_mutex pool_lock#2 irq_context: 0 cb_lock genl_mutex rfkill_global_mutex &k->list_lock irq_context: 0 cb_lock genl_mutex rfkill_global_mutex lock irq_context: 0 cb_lock genl_mutex rfkill_global_mutex lock kernfs_idr_lock irq_context: 0 cb_lock genl_mutex rfkill_global_mutex &root->kernfs_rwsem irq_context: 0 cb_lock genl_mutex rfkill_global_mutex &root->kernfs_rwsem &root->kernfs_iattr_rwsem irq_context: 0 cb_lock genl_mutex rfkill_global_mutex bus_type_sem irq_context: 0 cb_lock genl_mutex rfkill_global_mutex sysfs_symlink_target_lock irq_context: 0 cb_lock genl_mutex rfkill_global_mutex lock kernfs_idr_lock pool_lock#2 irq_context: 0 cb_lock genl_mutex rfkill_global_mutex &c->lock irq_context: 0 cb_lock genl_mutex rfkill_global_mutex &____s->seqcount irq_context: 0 cb_lock genl_mutex rfkill_global_mutex &root->kernfs_rwsem irq_context: 0 cb_lock genl_mutex rfkill_global_mutex &dev->power.lock irq_context: 0 cb_lock genl_mutex rfkill_global_mutex dpm_list_mtx irq_context: 0 cb_lock genl_mutex rfkill_global_mutex &rfkill->lock irq_context: 0 cb_lock genl_mutex rfkill_global_mutex uevent_sock_mutex irq_context: 0 cb_lock genl_mutex rfkill_global_mutex uevent_sock_mutex fs_reclaim irq_context: 0 cb_lock genl_mutex rfkill_global_mutex uevent_sock_mutex fs_reclaim mmu_notifier_invalidate_range_start irq_context: 0 cb_lock genl_mutex rfkill_global_mutex uevent_sock_mutex &c->lock irq_context: 0 cb_lock genl_mutex rfkill_global_mutex uevent_sock_mutex pool_lock#2 irq_context: 0 cb_lock genl_mutex rfkill_global_mutex uevent_sock_mutex nl_table_lock irq_context: 0 cb_lock genl_mutex rfkill_global_mutex uevent_sock_mutex rlock-AF_NETLINK irq_context: 0 cb_lock genl_mutex rfkill_global_mutex uevent_sock_mutex rcu_read_lock &ei->socket.wq.wait irq_context: 0 cb_lock genl_mutex rfkill_global_mutex uevent_sock_mutex rcu_read_lock &ei->socket.wq.wait &ep->lock irq_context: 0 cb_lock genl_mutex rfkill_global_mutex uevent_sock_mutex rcu_read_lock &ei->socket.wq.wait &ep->lock &ep->wq irq_context: 0 cb_lock genl_mutex rfkill_global_mutex uevent_sock_mutex rcu_read_lock &ei->socket.wq.wait &ep->lock &ep->wq &p->pi_lock irq_context: 0 cb_lock genl_mutex rfkill_global_mutex uevent_sock_mutex rcu_read_lock &ei->socket.wq.wait &ep->lock &ep->wq &p->pi_lock &rq->__lock irq_context: 0 cb_lock genl_mutex rfkill_global_mutex uevent_sock_mutex rcu_read_lock &ei->socket.wq.wait &ep->lock &ep->wq &p->pi_lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 cb_lock genl_mutex rfkill_global_mutex uevent_sock_mutex rcu_read_lock rcu_node_0 irq_context: 0 cb_lock genl_mutex rfkill_global_mutex uevent_sock_mutex rcu_read_lock &rq->__lock irq_context: 0 cb_lock genl_mutex rfkill_global_mutex uevent_sock_mutex nl_table_wait.lock irq_context: 0 cb_lock genl_mutex rfkill_global_mutex uevent_sock_mutex &obj_hash[i].lock irq_context: 0 cb_lock genl_mutex rfkill_global_mutex &obj_hash[i].lock irq_context: 0 cb_lock genl_mutex rfkill_global_mutex &k->k_lock irq_context: 0 cb_lock genl_mutex rfkill_global_mutex subsys mutex#41 irq_context: 0 cb_lock genl_mutex rfkill_global_mutex subsys mutex#41 &k->k_lock irq_context: 0 cb_lock genl_mutex rfkill_global_mutex triggers_list_lock irq_context: 0 cb_lock genl_mutex rfkill_global_mutex leds_list_lock irq_context: 0 cb_lock genl_mutex rfkill_global_mutex leds_list_lock &led_cdev->trigger_lock irq_context: 0 cb_lock genl_mutex rfkill_global_mutex rcu_read_lock &pool->lock irq_context: 0 cb_lock genl_mutex rfkill_global_mutex rcu_read_lock &pool->lock &obj_hash[i].lock irq_context: 0 cb_lock genl_mutex rfkill_global_mutex rcu_read_lock &pool->lock pool_lock#2 irq_context: 0 cb_lock genl_mutex rfkill_global_mutex rcu_read_lock &pool->lock &p->pi_lock irq_context: 0 cb_lock genl_mutex rfkill_global_mutex rcu_read_lock &pool->lock &p->pi_lock &rq->__lock irq_context: 0 cb_lock genl_mutex rfkill_global_mutex rcu_read_lock &pool->lock &p->pi_lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 cb_lock genl_mutex rfkill_global_mutex &rq->__lock irq_context: 0 cb_lock genl_mutex rfkill_global_mutex.wait_lock irq_context: 0 cb_lock genl_mutex pin_fs_lock irq_context: 0 cb_lock genl_mutex &sb->s_type->i_mutex_key#3 irq_context: 0 cb_lock genl_mutex &sb->s_type->i_mutex_key#3 rename_lock.seqcount irq_context: 0 cb_lock genl_mutex &sb->s_type->i_mutex_key#3 fs_reclaim irq_context: 0 cb_lock genl_mutex &sb->s_type->i_mutex_key#3 fs_reclaim mmu_notifier_invalidate_range_start irq_context: 0 cb_lock genl_mutex &sb->s_type->i_mutex_key#3 &c->lock irq_context: 0 cb_lock genl_mutex &sb->s_type->i_mutex_key#3 &____s->seqcount irq_context: 0 cb_lock genl_mutex &sb->s_type->i_mutex_key#3 pool_lock#2 irq_context: 0 cb_lock genl_mutex &sb->s_type->i_mutex_key#3 &dentry->d_lock irq_context: 0 cb_lock genl_mutex &sb->s_type->i_mutex_key#3 rcu_read_lock rename_lock.seqcount irq_context: 0 cb_lock genl_mutex &sb->s_type->i_mutex_key#3 &dentry->d_lock &wq irq_context: 0 cb_lock genl_mutex &sb->s_type->i_mutex_key#3 mmu_notifier_invalidate_range_start irq_context: 0 cb_lock genl_mutex &sb->s_type->i_mutex_key#3 &sb->s_type->i_lock_key#7 irq_context: 0 cb_lock genl_mutex &sb->s_type->i_mutex_key#3 &s->s_inode_list_lock irq_context: 0 cb_lock genl_mutex &sb->s_type->i_mutex_key#3 tk_core.seq.seqcount irq_context: 0 cb_lock genl_mutex &sb->s_type->i_mutex_key#3 &sb->s_type->i_lock_key#7 &dentry->d_lock irq_context: 0 cb_lock genl_mutex &sb->s_type->i_mutex_key#3 &pcp->lock &zone->lock irq_context: 0 cb_lock genl_mutex &sb->s_type->i_mutex_key#3 &pcp->lock &zone->lock &____s->seqcount irq_context: 0 cb_lock genl_mutex &sb->s_type->i_mutex_key#3 rcu_read_lock pool_lock#2 irq_context: 0 cb_lock genl_mutex &sb->s_type->i_mutex_key#3 &obj_hash[i].lock irq_context: 0 cb_lock genl_mutex &sb->s_type->i_mutex_key#3 &rq->__lock irq_context: 0 (wq_completion)mld (work_completion)(&(&idev->mc_dad_work)->work) &idev->mc_lock rcu_read_lock rcu_read_lock rcu_read_lock_bh &obj_hash[i].lock pool_lock irq_context: 0 cb_lock genl_mutex rtnl_mutex &rdev->wiphy.mtx stack_depot_init_mutex irq_context: 0 cb_lock genl_mutex rtnl_mutex &rdev->wiphy.mtx pcpu_alloc_mutex irq_context: 0 cb_lock genl_mutex rtnl_mutex &rdev->wiphy.mtx pcpu_alloc_mutex pcpu_lock irq_context: 0 cb_lock genl_mutex rtnl_mutex &rdev->wiphy.mtx &local->iflist_mtx irq_context: 0 cb_lock genl_mutex rtnl_mutex &rdev->wiphy.mtx &obj_hash[i].lock pool_lock irq_context: 0 cb_lock genl_mutex rtnl_mutex &rdev->wiphy.mtx &sb->s_type->i_mutex_key#3 &c->lock irq_context: 0 cb_lock genl_mutex rtnl_mutex &rdev->wiphy.mtx &sb->s_type->i_mutex_key#3 &pcp->lock &zone->lock irq_context: 0 cb_lock genl_mutex rtnl_mutex &rdev->wiphy.mtx &sb->s_type->i_mutex_key#3 &pcp->lock &zone->lock &____s->seqcount irq_context: 0 cb_lock genl_mutex rtnl_mutex &rdev->wiphy.mtx &sb->s_type->i_mutex_key#3 &____s->seqcount irq_context: 0 cb_lock genl_mutex rtnl_mutex &rdev->wiphy.mtx &sb->s_type->i_mutex_key#3 rcu_read_lock pool_lock#2 irq_context: 0 cb_lock genl_mutex rtnl_mutex &rdev->wiphy.mtx &sb->s_type->i_mutex_key#3 &obj_hash[i].lock irq_context: 0 cb_lock genl_mutex rtnl_mutex &rdev->wiphy.mtx &xa->xa_lock#3 irq_context: 0 cb_lock genl_mutex rtnl_mutex &rdev->wiphy.mtx net_rwsem irq_context: 0 cb_lock genl_mutex rtnl_mutex &rdev->wiphy.mtx net_rwsem &list->lock#2 irq_context: 0 cb_lock genl_mutex rtnl_mutex &rdev->wiphy.mtx &tn->lock irq_context: 0 cb_lock genl_mutex rtnl_mutex &rdev->wiphy.mtx &x->wait#9 irq_context: 0 cb_lock genl_mutex rtnl_mutex &rdev->wiphy.mtx lock kernfs_idr_lock pool_lock#2 irq_context: 0 cb_lock genl_mutex rtnl_mutex &rdev->wiphy.mtx subsys mutex#17 irq_context: 0 cb_lock genl_mutex rtnl_mutex &rdev->wiphy.mtx subsys mutex#17 &k->k_lock irq_context: 0 cb_lock genl_mutex rtnl_mutex &rdev->wiphy.mtx &dir->lock#2 irq_context: 0 cb_lock genl_mutex rtnl_mutex &rdev->wiphy.mtx quarantine_lock irq_context: 0 cb_lock genl_mutex rtnl_mutex &rdev->wiphy.mtx remove_cache_srcu irq_context: 0 cb_lock genl_mutex rtnl_mutex &rdev->wiphy.mtx remove_cache_srcu quarantine_lock irq_context: 0 cb_lock genl_mutex rtnl_mutex &rdev->wiphy.mtx remove_cache_srcu &c->lock irq_context: 0 cb_lock genl_mutex rtnl_mutex &rdev->wiphy.mtx remove_cache_srcu &n->list_lock irq_context: 0 cb_lock genl_mutex rtnl_mutex &rdev->wiphy.mtx remove_cache_srcu rcu_read_lock pool_lock#2 irq_context: 0 cb_lock genl_mutex rtnl_mutex &rdev->wiphy.mtx remove_cache_srcu &obj_hash[i].lock irq_context: 0 cb_lock genl_mutex rtnl_mutex &rdev->wiphy.mtx dev_hotplug_mutex irq_context: 0 cb_lock genl_mutex rtnl_mutex &rdev->wiphy.mtx dev_hotplug_mutex &dev->power.lock irq_context: 0 cb_lock genl_mutex rtnl_mutex &rdev->wiphy.mtx dev_base_lock irq_context: 0 cb_lock genl_mutex rtnl_mutex &rdev->wiphy.mtx input_pool.lock irq_context: 0 cb_lock genl_mutex rtnl_mutex &rdev->wiphy.mtx rcu_read_lock &pool->lock/1 irq_context: 0 cb_lock genl_mutex rtnl_mutex &rdev->wiphy.mtx rcu_read_lock &pool->lock/1 &obj_hash[i].lock irq_context: 0 cb_lock genl_mutex rtnl_mutex &rdev->wiphy.mtx rcu_read_lock &pool->lock/1 pool_lock#2 irq_context: 0 cb_lock genl_mutex rtnl_mutex &rdev->wiphy.mtx rcu_read_lock &pool->lock/1 &p->pi_lock irq_context: 0 cb_lock genl_mutex rtnl_mutex &rdev->wiphy.mtx rcu_read_lock &pool->lock/1 &p->pi_lock &rq->__lock irq_context: 0 cb_lock genl_mutex rtnl_mutex &rdev->wiphy.mtx rcu_read_lock &pool->lock/1 &p->pi_lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 cb_lock genl_mutex rtnl_mutex &rdev->wiphy.mtx &rq->__lock irq_context: 0 (wq_completion)mld (work_completion)(&(&idev->mc_dad_work)->work) &idev->mc_lock &n->list_lock irq_context: 0 (wq_completion)mld (work_completion)(&(&idev->mc_dad_work)->work) &idev->mc_lock &n->list_lock &c->lock irq_context: 0 cb_lock genl_mutex rtnl_mutex &rdev->wiphy.mtx batched_entropy_u32.lock irq_context: 0 cb_lock genl_mutex rtnl_mutex &rdev->wiphy.mtx &tbl->lock irq_context: 0 cb_lock genl_mutex rtnl_mutex &rdev->wiphy.mtx sysctl_lock irq_context: 0 cb_lock genl_mutex rtnl_mutex &rdev->wiphy.mtx failover_lock irq_context: 0 cb_lock genl_mutex rtnl_mutex &rdev->wiphy.mtx proc_subdir_lock irq_context: 0 cb_lock genl_mutex rtnl_mutex &rdev->wiphy.mtx proc_inum_ida.xa_lock irq_context: 0 cb_lock genl_mutex rtnl_mutex &rdev->wiphy.mtx proc_subdir_lock irq_context: 0 cb_lock genl_mutex rtnl_mutex &rdev->wiphy.mtx &idev->mc_lock irq_context: 0 cb_lock genl_mutex rtnl_mutex &rdev->wiphy.mtx &idev->mc_lock fs_reclaim irq_context: 0 cb_lock genl_mutex rtnl_mutex &rdev->wiphy.mtx &idev->mc_lock fs_reclaim mmu_notifier_invalidate_range_start irq_context: 0 cb_lock genl_mutex rtnl_mutex &rdev->wiphy.mtx &idev->mc_lock pool_lock#2 irq_context: 0 cb_lock genl_mutex rtnl_mutex &rdev->wiphy.mtx &idev->mc_lock &obj_hash[i].lock irq_context: 0 cb_lock genl_mutex rtnl_mutex &rdev->wiphy.mtx &idev->mc_lock _xmit_ETHER irq_context: 0 cb_lock genl_mutex rtnl_mutex &rdev->wiphy.mtx &idev->mc_lock _xmit_ETHER pool_lock#2 irq_context: 0 cb_lock genl_mutex rtnl_mutex &rdev->wiphy.mtx &pnettable->lock irq_context: 0 cb_lock genl_mutex rtnl_mutex &rdev->wiphy.mtx smc_ib_devices.mutex irq_context: 0 cb_lock genl_mutex rtnl_mutex &rdev->wiphy.mtx rcu_read_lock &ndev->lock irq_context: 0 cb_lock genl_mutex rtnl_mutex &rdev->wiphy.mtx &wdev->mtx irq_context: 0 cb_lock genl_mutex rtnl_mutex &rdev->wiphy.mtx &fq->lock irq_context: 0 cb_lock genl_mutex (inetaddr_chain).rwsem irq_context: 0 cb_lock genl_mutex inet6addr_chain.lock irq_context: 0 cb_lock genl_mutex hwsim_radio_lock rcu_read_lock rhashtable_bucket irq_context: 0 cb_lock genl_mutex nl_table_lock irq_context: 0 cb_lock genl_mutex nl_table_wait.lock irq_context: 0 cb_lock rtnl_mutex &rdev->wiphy.mtx &wdev->mtx irq_context: 0 cb_lock rtnl_mutex &rdev->wiphy.mtx &wdev->mtx &rq->__lock irq_context: 0 cb_lock rtnl_mutex &rdev->wiphy.mtx &wdev->event_lock irq_context: 0 cb_lock rtnl_mutex &rdev->wiphy.mtx &rdev->mgmt_registrations_lock irq_context: 0 cb_lock rtnl_mutex &rdev->wiphy.mtx &obj_hash[i].lock irq_context: 0 cb_lock rtnl_mutex &rdev->wiphy.mtx (work_completion)(&(&sdata->dec_tailroom_needed_wk)->work) irq_context: 0 cb_lock rtnl_mutex &rdev->wiphy.mtx &local->key_mtx irq_context: 0 cb_lock rtnl_mutex &rdev->wiphy.mtx pin_fs_lock irq_context: 0 cb_lock rtnl_mutex &rdev->wiphy.mtx &dentry->d_lock irq_context: 0 cb_lock rtnl_mutex &rdev->wiphy.mtx &sb->s_type->i_mutex_key#3 irq_context: 0 cb_lock rtnl_mutex &rdev->wiphy.mtx &sb->s_type->i_mutex_key#3 &dentry->d_lock irq_context: 0 cb_lock rtnl_mutex &rdev->wiphy.mtx &sb->s_type->i_mutex_key#3 &dentry->d_lock &dentry->d_lock/1 irq_context: 0 cb_lock rtnl_mutex &rdev->wiphy.mtx &sb->s_type->i_mutex_key#3 tk_core.seq.seqcount irq_context: 0 cb_lock rtnl_mutex &rdev->wiphy.mtx &sb->s_type->i_mutex_key#3 rename_lock.seqcount irq_context: 0 cb_lock rtnl_mutex &rdev->wiphy.mtx &sb->s_type->i_mutex_key#3 pin_fs_lock irq_context: 0 cb_lock rtnl_mutex &rdev->wiphy.mtx &sb->s_type->i_mutex_key#3 rcu_read_lock mount_lock irq_context: 0 cb_lock rtnl_mutex &rdev->wiphy.mtx &sb->s_type->i_mutex_key#3 rcu_read_lock mount_lock mount_lock.seqcount irq_context: 0 cb_lock rtnl_mutex &rdev->wiphy.mtx &sb->s_type->i_mutex_key#3 mount_lock irq_context: 0 cb_lock rtnl_mutex &rdev->wiphy.mtx &sb->s_type->i_mutex_key#3 mount_lock mount_lock.seqcount irq_context: 0 cb_lock rtnl_mutex &rdev->wiphy.mtx &sb->s_type->i_mutex_key#3 rcu_read_lock &dentry->d_lock irq_context: 0 cb_lock rtnl_mutex &rdev->wiphy.mtx &sb->s_type->i_mutex_key#3 &fsnotify_mark_srcu irq_context: 0 cb_lock rtnl_mutex &rdev->wiphy.mtx &sb->s_type->i_mutex_key#3 &sb->s_type->i_lock_key#7 irq_context: 0 cb_lock rtnl_mutex &rdev->wiphy.mtx &sb->s_type->i_mutex_key#3 &s->s_inode_list_lock irq_context: 0 cb_lock rtnl_mutex &rdev->wiphy.mtx &sb->s_type->i_mutex_key#3 &xa->xa_lock#9 irq_context: 0 cb_lock rtnl_mutex &rdev->wiphy.mtx &sb->s_type->i_mutex_key#3 &obj_hash[i].lock irq_context: 0 cb_lock rtnl_mutex &rdev->wiphy.mtx &sb->s_type->i_mutex_key#3 pool_lock#2 irq_context: 0 cb_lock rtnl_mutex &rdev->wiphy.mtx &sb->s_type->i_mutex_key#3 &obj_hash[i].lock pool_lock irq_context: 0 cb_lock rtnl_mutex &rdev->wiphy.mtx rcu_read_lock &dentry->d_lock irq_context: 0 cb_lock rtnl_mutex &rdev->wiphy.mtx &fsnotify_mark_srcu irq_context: 0 cb_lock rtnl_mutex &rdev->wiphy.mtx &sb->s_type->i_lock_key#7 irq_context: 0 cb_lock rtnl_mutex &rdev->wiphy.mtx &s->s_inode_list_lock irq_context: 0 cb_lock rtnl_mutex &rdev->wiphy.mtx &xa->xa_lock#9 irq_context: 0 cb_lock rtnl_mutex &rdev->wiphy.mtx pool_lock#2 irq_context: 0 cb_lock rtnl_mutex &rdev->wiphy.mtx rcu_read_lock mount_lock irq_context: 0 cb_lock rtnl_mutex &rdev->wiphy.mtx rcu_read_lock mount_lock mount_lock.seqcount irq_context: 0 cb_lock rtnl_mutex &rdev->wiphy.mtx mount_lock irq_context: 0 cb_lock rtnl_mutex &rdev->wiphy.mtx mount_lock mount_lock.seqcount irq_context: 0 cb_lock rtnl_mutex &rdev->wiphy.mtx &rdev->wiphy_work_lock irq_context: 0 cb_lock rtnl_mutex &rdev->wiphy.mtx (&dwork->timer) irq_context: 0 cb_lock rtnl_mutex &rdev->wiphy.mtx &base->lock irq_context: 0 cb_lock rtnl_mutex &rdev->wiphy.mtx (work_completion)(&(&link->color_collision_detect_work)->work) irq_context: 0 cb_lock rtnl_mutex &rdev->wiphy.mtx &local->chanctx_mtx irq_context: 0 cb_lock rtnl_mutex &rdev->wiphy.mtx &sb->s_type->i_mutex_key#3 fs_reclaim irq_context: 0 cb_lock rtnl_mutex &rdev->wiphy.mtx &sb->s_type->i_mutex_key#3 fs_reclaim mmu_notifier_invalidate_range_start irq_context: 0 cb_lock rtnl_mutex &rdev->wiphy.mtx &sb->s_type->i_mutex_key#3 &____s->seqcount irq_context: 0 cb_lock rtnl_mutex &rdev->wiphy.mtx &sb->s_type->i_mutex_key#3 rcu_read_lock pool_lock#2 irq_context: 0 cb_lock rtnl_mutex &rdev->wiphy.mtx &sb->s_type->i_mutex_key#3 rcu_read_lock rename_lock.seqcount irq_context: 0 cb_lock rtnl_mutex &rdev->wiphy.mtx &sb->s_type->i_mutex_key#3 &dentry->d_lock &wq irq_context: 0 cb_lock rtnl_mutex &rdev->wiphy.mtx &sb->s_type->i_mutex_key#3 mmu_notifier_invalidate_range_start irq_context: 0 cb_lock rtnl_mutex &rdev->wiphy.mtx &sb->s_type->i_mutex_key#3 &sb->s_type->i_lock_key#7 &dentry->d_lock irq_context: 0 cb_lock rtnl_mutex &rdev->wiphy.mtx &sb->s_type->i_mutex_key#3 &c->lock irq_context: 0 cb_lock rtnl_mutex &rdev->wiphy.mtx fs_reclaim irq_context: 0 cb_lock rtnl_mutex &rdev->wiphy.mtx fs_reclaim mmu_notifier_invalidate_range_start irq_context: 0 cb_lock rtnl_mutex &rdev->wiphy.mtx &c->lock irq_context: 0 cb_lock rtnl_mutex &rdev->wiphy.mtx &fq->lock irq_context: 0 cb_lock rtnl_mutex &rdev->wiphy.mtx nl_table_lock irq_context: 0 cb_lock rtnl_mutex &rdev->wiphy.mtx nl_table_wait.lock irq_context: 0 cb_lock rtnl_mutex.wait_lock irq_context: 0 (wq_completion)ipv6_addrconf (work_completion)(&(&ifa->dad_work)->work) rtnl_mutex &idev->mc_lock &batadv_netdev_addr_lock_key/1 &c->lock irq_context: 0 (wq_completion)phy3 irq_context: 0 (wq_completion)phy3 (work_completion)(&local->reconfig_filter) irq_context: 0 (wq_completion)phy3 (work_completion)(&local->reconfig_filter) &local->filter_lock irq_context: 0 cb_lock &rdev->wiphy.mtx rtnl_mutex.wait_lock irq_context: 0 cb_lock &rdev->wiphy.mtx &p->pi_lock irq_context: 0 cb_lock &rdev->wiphy.mtx &p->pi_lock &rq->__lock irq_context: 0 cb_lock &rdev->wiphy.mtx &p->pi_lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 cb_lock &rdev->wiphy.mtx &wdev->mtx irq_context: 0 cb_lock &rdev->wiphy.mtx &wdev->mtx &local->chanctx_mtx irq_context: 0 cb_lock &rdev->wiphy.mtx &wdev->mtx &rdev->wiphy_work_lock irq_context: 0 cb_lock &rdev->wiphy.mtx &wdev->mtx rcu_read_lock &pool->lock irq_context: 0 cb_lock &rdev->wiphy.mtx &wdev->mtx rcu_read_lock &pool->lock &obj_hash[i].lock irq_context: 0 cb_lock &rdev->wiphy.mtx &wdev->mtx rcu_read_lock &pool->lock &p->pi_lock irq_context: 0 cb_lock &rdev->wiphy.mtx &wdev->mtx rcu_read_lock &pool->lock &p->pi_lock &rq->__lock irq_context: 0 cb_lock &rdev->wiphy.mtx &wdev->mtx rcu_read_lock &pool->lock &p->pi_lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 cb_lock &rdev->wiphy.mtx &wdev->mtx &rq->__lock irq_context: 0 (wq_completion)events (work_completion)(&rdev->wiphy_work) irq_context: 0 (wq_completion)events (work_completion)(&rdev->wiphy_work) &rq->__lock irq_context: 0 (wq_completion)events (work_completion)(&rdev->wiphy_work) &rdev->wiphy.mtx irq_context: 0 (wq_completion)events (work_completion)(&rdev->wiphy_work) &rdev->wiphy.mtx &rdev->wiphy_work_lock irq_context: 0 (wq_completion)events (work_completion)(&rdev->wiphy_work) &rdev->wiphy.mtx &list->lock#18 irq_context: 0 (wq_completion)events (work_completion)(&rdev->wiphy_work) &rdev->wiphy.mtx &wdev->mtx irq_context: 0 (wq_completion)events (work_completion)(&rdev->wiphy_work) &rdev->wiphy.mtx &wdev->mtx &ifibss->incomplete_lock irq_context: 0 (wq_completion)events (work_completion)(&rdev->wiphy_work) &rdev->wiphy.mtx &wdev->mtx &rdev->bss_lock irq_context: 0 (wq_completion)events (work_completion)(&rdev->wiphy_work) &rdev->wiphy.mtx &wdev->mtx (console_sem).lock irq_context: 0 (wq_completion)events (work_completion)(&rdev->wiphy_work) &rdev->wiphy.mtx &wdev->mtx console_lock console_srcu console_owner_lock irq_context: 0 (wq_completion)events (work_completion)(&rdev->wiphy_work) &rdev->wiphy.mtx &wdev->mtx console_lock console_srcu console_owner irq_context: 0 (wq_completion)events (work_completion)(&rdev->wiphy_work) &rdev->wiphy.mtx &wdev->mtx console_lock console_srcu console_owner &port_lock_key irq_context: 0 (wq_completion)events (work_completion)(&rdev->wiphy_work) &rdev->wiphy.mtx &wdev->mtx console_lock console_srcu console_owner console_owner_lock irq_context: 0 (wq_completion)events (work_completion)(&rdev->wiphy_work) &rdev->wiphy.mtx &wdev->mtx &rq->__lock irq_context: 0 cb_lock genl_mutex &zone->lock &____s->seqcount irq_context: 0 (wq_completion)ipv6_addrconf (work_completion)(&(&ifa->dad_work)->work) &p->pi_lock &cfs_rq->removed.lock irq_context: 0 (wq_completion)mld (work_completion)(&(&idev->mc_ifc_work)->work) &idev->mc_lock rcu_read_lock rcu_read_lock rcu_read_lock_bh quarantine_lock irq_context: 0 (wq_completion)ipv6_addrconf (work_completion)(&(&ifa->dad_work)->work) rtnl_mutex &idev->mc_lock remove_cache_srcu irq_context: 0 (wq_completion)ipv6_addrconf (work_completion)(&(&ifa->dad_work)->work) rtnl_mutex &idev->mc_lock remove_cache_srcu quarantine_lock irq_context: 0 (wq_completion)ipv6_addrconf (work_completion)(&(&ifa->dad_work)->work) rtnl_mutex &idev->mc_lock remove_cache_srcu &c->lock irq_context: 0 (wq_completion)ipv6_addrconf (work_completion)(&(&ifa->dad_work)->work) rtnl_mutex &idev->mc_lock remove_cache_srcu &n->list_lock irq_context: 0 (wq_completion)ipv6_addrconf (work_completion)(&(&ifa->dad_work)->work) rtnl_mutex &idev->mc_lock remove_cache_srcu rcu_read_lock pool_lock#2 irq_context: 0 (wq_completion)ipv6_addrconf (work_completion)(&(&ifa->dad_work)->work) rtnl_mutex &idev->mc_lock remove_cache_srcu &obj_hash[i].lock irq_context: 0 (wq_completion)ipv6_addrconf (work_completion)(&(&ifa->dad_work)->work) rtnl_mutex &idev->mc_lock rcu_read_lock rcu_read_lock rcu_node_0 irq_context: 0 (wq_completion)ipv6_addrconf (work_completion)(&(&ifa->dad_work)->work) rtnl_mutex &idev->mc_lock rcu_read_lock rcu_read_lock &rq->__lock irq_context: 0 (wq_completion)ipv6_addrconf (work_completion)(&(&ifa->dad_work)->work) rtnl_mutex &idev->mc_lock rcu_read_lock rcu_read_lock &rq->__lock &cfs_rq->removed.lock irq_context: 0 (wq_completion)ipv6_addrconf (work_completion)(&(&ifa->dad_work)->work) rtnl_mutex &idev->mc_lock rcu_read_lock rcu_node_0 irq_context: 0 (wq_completion)ipv6_addrconf (work_completion)(&(&ifa->dad_work)->work) rtnl_mutex &idev->mc_lock rcu_read_lock rcu_read_lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)events (work_completion)(&rdev->wiphy_work) &rdev->wiphy.mtx &wdev->mtx &local->mtx irq_context: 0 (wq_completion)events (work_completion)(&rdev->wiphy_work) &rdev->wiphy.mtx &wdev->mtx &local->mtx &local->chanctx_mtx irq_context: 0 (wq_completion)events (work_completion)(&rdev->wiphy_work) &rdev->wiphy.mtx &wdev->mtx &local->mtx &local->chanctx_mtx fs_reclaim irq_context: 0 (wq_completion)events (work_completion)(&rdev->wiphy_work) &rdev->wiphy.mtx &wdev->mtx &local->mtx &local->chanctx_mtx fs_reclaim mmu_notifier_invalidate_range_start irq_context: 0 (wq_completion)events (work_completion)(&rdev->wiphy_work) &rdev->wiphy.mtx &wdev->mtx &local->mtx &local->chanctx_mtx pool_lock#2 irq_context: 0 (wq_completion)events (work_completion)(&rdev->wiphy_work) &rdev->wiphy.mtx &wdev->mtx &local->mtx &local->chanctx_mtx &data->mutex irq_context: 0 (wq_completion)events (work_completion)(&rdev->wiphy_work) &rdev->wiphy.mtx &wdev->mtx fs_reclaim irq_context: 0 (wq_completion)events (work_completion)(&rdev->wiphy_work) &rdev->wiphy.mtx &wdev->mtx fs_reclaim mmu_notifier_invalidate_range_start irq_context: 0 (wq_completion)events (work_completion)(&rdev->wiphy_work) &rdev->wiphy.mtx &wdev->mtx pool_lock#2 irq_context: 0 (wq_completion)events (work_completion)(&rdev->wiphy_work) &rdev->wiphy.mtx &wdev->mtx tk_core.seq.seqcount irq_context: 0 (wq_completion)events (work_completion)(&rdev->wiphy_work) &rdev->wiphy.mtx &wdev->mtx hrtimer_bases.lock irq_context: 0 (wq_completion)events (work_completion)(&rdev->wiphy_work) &rdev->wiphy.mtx &wdev->mtx hrtimer_bases.lock tk_core.seq.seqcount irq_context: 0 (wq_completion)events (work_completion)(&rdev->wiphy_work) &rdev->wiphy.mtx &wdev->mtx hrtimer_bases.lock &obj_hash[i].lock irq_context: 0 (wq_completion)events (work_completion)(&rdev->wiphy_work) &rdev->wiphy.mtx &wdev->mtx &obj_hash[i].lock irq_context: 0 (wq_completion)events (work_completion)(&rdev->wiphy_work) &rdev->wiphy.mtx &wdev->mtx &base->lock irq_context: 0 (wq_completion)events (work_completion)(&rdev->wiphy_work) &rdev->wiphy.mtx &wdev->mtx &base->lock &obj_hash[i].lock irq_context: 0 (wq_completion)events (work_completion)(&rdev->wiphy_work) &rdev->wiphy.mtx &wdev->mtx &rdev->bss_lock pool_lock#2 irq_context: 0 (wq_completion)events (work_completion)(&rdev->wiphy_work) &rdev->wiphy.mtx &wdev->mtx rcu_read_lock &pool->lock irq_context: 0 (wq_completion)events (work_completion)(&rdev->wiphy_work) &rdev->wiphy.mtx &wdev->mtx rcu_read_lock &pool->lock &obj_hash[i].lock irq_context: 0 (wq_completion)events (work_completion)(&rdev->wiphy_work) &rdev->wiphy.mtx &wdev->mtx rcu_read_lock &pool->lock &p->pi_lock irq_context: 0 (wq_completion)events (work_completion)(&rdev->wiphy_work) &rdev->wiphy.mtx &wdev->mtx rcu_read_lock &pool->lock &p->pi_lock &rq->__lock irq_context: 0 (wq_completion)events (work_completion)(&rdev->wiphy_work) &rdev->wiphy.mtx &wdev->mtx rcu_read_lock &pool->lock &p->pi_lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)events (work_completion)(&rdev->wiphy_work) &rdev->wiphy.mtx &wdev->mtx &____s->seqcount irq_context: 0 (wq_completion)events (work_completion)(&rdev->wiphy_work) &rdev->wiphy.mtx &wdev->mtx rcu_read_lock pool_lock#2 irq_context: 0 (wq_completion)events (work_completion)(&rdev->wiphy_work) &rdev->wiphy.mtx &wdev->mtx &wdev->event_lock irq_context: 0 (wq_completion)events (work_completion)(&rdev->wiphy_work) &rdev->wiphy.mtx &wdev->mtx rcu_read_lock &pool->lock/1 irq_context: 0 (wq_completion)events (work_completion)(&rdev->wiphy_work) &rdev->wiphy.mtx &wdev->mtx rcu_read_lock &pool->lock/1 &obj_hash[i].lock irq_context: 0 (wq_completion)events (work_completion)(&rdev->wiphy_work) &rdev->wiphy.mtx &wdev->mtx rcu_read_lock &pool->lock/1 &p->pi_lock irq_context: 0 (wq_completion)cfg80211 irq_context: 0 (wq_completion)cfg80211 (work_completion)(&rdev->event_work) irq_context: 0 (wq_completion)cfg80211 (work_completion)(&rdev->event_work) &rdev->wiphy.mtx irq_context: 0 (wq_completion)cfg80211 (work_completion)(&rdev->event_work) &rdev->wiphy.mtx &wdev->event_lock irq_context: 0 (wq_completion)cfg80211 (work_completion)(&rdev->event_work) &rdev->wiphy.mtx &wdev->mtx irq_context: 0 (wq_completion)cfg80211 (work_completion)(&rdev->event_work) &rdev->wiphy.mtx &wdev->mtx &rdev->bss_lock irq_context: 0 (wq_completion)cfg80211 (work_completion)(&rdev->event_work) &rdev->wiphy.mtx &wdev->mtx fs_reclaim irq_context: 0 (wq_completion)cfg80211 (work_completion)(&rdev->event_work) &rdev->wiphy.mtx &wdev->mtx fs_reclaim mmu_notifier_invalidate_range_start irq_context: 0 (wq_completion)cfg80211 (work_completion)(&rdev->event_work) &rdev->wiphy.mtx &wdev->mtx &c->lock irq_context: 0 (wq_completion)cfg80211 (work_completion)(&rdev->event_work) &rdev->wiphy.mtx &wdev->mtx pool_lock#2 irq_context: 0 (wq_completion)cfg80211 (work_completion)(&rdev->event_work) &rdev->wiphy.mtx &wdev->mtx &obj_hash[i].lock irq_context: 0 (wq_completion)cfg80211 (work_completion)(&rdev->event_work) &rdev->wiphy.mtx &wdev->mtx nl_table_lock irq_context: 0 (wq_completion)cfg80211 (work_completion)(&rdev->event_work) &rdev->wiphy.mtx &wdev->mtx nl_table_wait.lock irq_context: 0 (wq_completion)cfg80211 (work_completion)(&rdev->event_work) &rdev->wiphy.mtx &wdev->mtx &____s->seqcount irq_context: 0 (wq_completion)cfg80211 (work_completion)(&rdev->event_work) &rdev->wiphy.mtx &wdev->mtx &list->lock#2 irq_context: 0 (wq_completion)cfg80211 (work_completion)(&rdev->event_work) &rdev->wiphy.mtx &wdev->mtx rcu_read_lock &pool->lock irq_context: 0 (wq_completion)cfg80211 (work_completion)(&rdev->event_work) &rdev->wiphy.mtx &wdev->mtx rcu_read_lock &pool->lock &obj_hash[i].lock irq_context: 0 cb_lock genl_mutex rtnl_mutex &rdev->wiphy.mtx &sb->s_type->i_mutex_key#3 remove_cache_srcu irq_context: 0 (wq_completion)cfg80211 (work_completion)(&rdev->event_work) &rdev->wiphy.mtx &wdev->mtx rcu_read_lock &pool->lock &p->pi_lock irq_context: 0 cb_lock genl_mutex rtnl_mutex &rdev->wiphy.mtx &sb->s_type->i_mutex_key#3 remove_cache_srcu quarantine_lock irq_context: 0 (wq_completion)cfg80211 (work_completion)(&rdev->event_work) &rdev->wiphy.mtx &wdev->mtx rcu_read_lock &pool->lock &p->pi_lock &rq->__lock irq_context: 0 cb_lock genl_mutex rtnl_mutex &rdev->wiphy.mtx &sb->s_type->i_mutex_key#3 remove_cache_srcu &c->lock irq_context: 0 (wq_completion)cfg80211 (work_completion)(&rdev->event_work) &rdev->wiphy.mtx &wdev->mtx rcu_read_lock &pool->lock &p->pi_lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 cb_lock genl_mutex rtnl_mutex &rdev->wiphy.mtx &sb->s_type->i_mutex_key#3 remove_cache_srcu &n->list_lock irq_context: 0 (wq_completion)cfg80211 (work_completion)(&rdev->event_work) &rdev->wiphy.mtx &obj_hash[i].lock irq_context: 0 cb_lock genl_mutex rtnl_mutex &rdev->wiphy.mtx &sb->s_type->i_mutex_key#3 remove_cache_srcu rcu_read_lock pool_lock#2 irq_context: 0 (wq_completion)cfg80211 (work_completion)(&rdev->event_work) &rdev->wiphy.mtx &rq->__lock irq_context: 0 (wq_completion)events wireless_nlevent_work irq_context: 0 (wq_completion)events wireless_nlevent_work net_rwsem irq_context: 0 (wq_completion)events wireless_nlevent_work net_rwsem &list->lock#2 irq_context: 0 (wq_completion)events wireless_nlevent_work net_rwsem &c->lock irq_context: 0 (wq_completion)events wireless_nlevent_work net_rwsem pool_lock#2 irq_context: 0 (wq_completion)events wireless_nlevent_work net_rwsem &obj_hash[i].lock irq_context: 0 (wq_completion)events wireless_nlevent_work net_rwsem nl_table_lock irq_context: 0 (wq_completion)events wireless_nlevent_work net_rwsem nl_table_wait.lock irq_context: 0 cb_lock genl_mutex rtnl_mutex &rdev->wiphy.mtx &sb->s_type->i_mutex_key#3 remove_cache_srcu &obj_hash[i].lock irq_context: softirq rcu_read_lock hwsim_radio_lock irq_context: 0 cb_lock genl_mutex rtnl_mutex &rdev->wiphy.mtx &lock->wait_lock irq_context: 0 cb_lock genl_mutex rtnl_mutex &rdev->wiphy.mtx &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)cfg80211 (work_completion)(&rdev->event_work) &rdev->wiphy.mtx pool_lock#2 irq_context: 0 (wq_completion)cfg80211 (work_completion)(&rdev->event_work) &lock->wait_lock irq_context: 0 (wq_completion)cfg80211 (work_completion)(&rdev->event_work) &p->pi_lock irq_context: 0 (wq_completion)cfg80211 (work_completion)(&rdev->event_work) &p->pi_lock &rq->__lock irq_context: 0 (wq_completion)cfg80211 (work_completion)(&rdev->event_work) &p->pi_lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 cb_lock genl_mutex rfkill_global_mutex &obj_hash[i].lock pool_lock irq_context: 0 cb_lock genl_mutex rtnl_mutex &rdev->wiphy.mtx &pcp->lock &zone->lock irq_context: 0 cb_lock genl_mutex rtnl_mutex &rdev->wiphy.mtx &idev->mc_lock _xmit_ETHER &c->lock irq_context: 0 cb_lock genl_mutex rtnl_mutex &rdev->wiphy.mtx &idev->mc_lock _xmit_ETHER &____s->seqcount irq_context: 0 cb_lock genl_mutex rtnl_mutex &rdev->wiphy.mtx pcpu_alloc_mutex &rq->__lock irq_context: 0 cb_lock genl_mutex hwsim_radio_lock rcu_read_lock rcu_read_lock &pool->lock irq_context: 0 cb_lock genl_mutex hwsim_radio_lock rcu_read_lock rcu_read_lock &pool->lock &obj_hash[i].lock irq_context: 0 cb_lock genl_mutex hwsim_radio_lock rcu_read_lock rcu_read_lock &pool->lock &p->pi_lock irq_context: 0 cb_lock genl_mutex hwsim_radio_lock rcu_read_lock rcu_read_lock &pool->lock &p->pi_lock &rq->__lock irq_context: 0 cb_lock genl_mutex hwsim_radio_lock rcu_read_lock rcu_read_lock &pool->lock &p->pi_lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 cb_lock genl_mutex hwsim_radio_lock rcu_read_lock rcu_read_lock &pool->lock &p->pi_lock &rq->__lock &cfs_rq->removed.lock irq_context: 0 cb_lock rtnl_mutex &rdev->wiphy.mtx &____s->seqcount irq_context: 0 (wq_completion)phy4 irq_context: 0 (wq_completion)phy4 (work_completion)(&local->reconfig_filter) irq_context: 0 (wq_completion)phy4 (work_completion)(&local->reconfig_filter) &local->filter_lock irq_context: 0 (wq_completion)events (work_completion)(&rdev->wiphy_work) &rdev->wiphy.mtx &lock->wait_lock irq_context: 0 (wq_completion)events (work_completion)(&rdev->wiphy_work) &rdev->wiphy.mtx &pool->lock irq_context: 0 (wq_completion)events (work_completion)(&rdev->wiphy_work) &rdev->wiphy.mtx &rq->__lock irq_context: 0 (wq_completion)events (work_completion)(&rdev->wiphy_work) &rdev->wiphy.mtx &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 cb_lock &lock->wait_lock irq_context: 0 (wq_completion)events_power_efficient (check_lifetime_work).work &rq->__lock irq_context: 0 (wq_completion)events (work_completion)(&rdev->wiphy_work) &rdev->wiphy.mtx &wdev->mtx &c->lock irq_context: 0 (wq_completion)cfg80211 (work_completion)(&rdev->event_work) &rdev->wiphy.mtx &wdev->mtx &n->list_lock irq_context: 0 (wq_completion)cfg80211 (work_completion)(&rdev->event_work) &rdev->wiphy.mtx &wdev->mtx &n->list_lock &c->lock irq_context: 0 (wq_completion)ipv6_addrconf (work_completion)(&(&ifa->dad_work)->work) rtnl_mutex rcu_read_lock rcu_read_lock rcu_read_lock_bh &hsr->seqnr_lock rcu_read_lock &n->list_lock irq_context: 0 (wq_completion)ipv6_addrconf (work_completion)(&(&ifa->dad_work)->work) rtnl_mutex rcu_read_lock rcu_read_lock rcu_read_lock_bh &hsr->seqnr_lock rcu_read_lock &n->list_lock &c->lock irq_context: 0 rtnl_mutex rcu_read_lock rcu_read_lock_bh &obj_hash[i].lock irq_context: 0 rtnl_mutex rcu_read_lock rcu_read_lock_bh pool_lock#2 irq_context: 0 &type->s_umount_key#49/1 irq_context: 0 &type->s_umount_key#49/1 fs_reclaim irq_context: 0 &type->s_umount_key#49/1 fs_reclaim mmu_notifier_invalidate_range_start irq_context: 0 &type->s_umount_key#49/1 pcpu_alloc_mutex irq_context: 0 &type->s_umount_key#49/1 pcpu_alloc_mutex pcpu_lock irq_context: 0 &type->s_umount_key#49/1 shrinker_rwsem irq_context: 0 &type->s_umount_key#49/1 list_lrus_mutex irq_context: 0 &type->s_umount_key#49/1 sb_lock irq_context: 0 &type->s_umount_key#49/1 sb_lock unnamed_dev_ida.xa_lock irq_context: 0 &type->s_umount_key#49/1 pool_lock#2 irq_context: 0 &type->s_umount_key#49/1 mmu_notifier_invalidate_range_start irq_context: 0 &type->s_umount_key#49/1 &sb->s_type->i_lock_key#32 irq_context: 0 &type->s_umount_key#49/1 &rq->__lock irq_context: 0 (wq_completion)mld (work_completion)(&(&idev->mc_dad_work)->work) &idev->mc_lock rcu_read_lock rcu_read_lock rcu_read_lock_bh rcu_read_lock &br->multicast_lock irq_context: 0 (wq_completion)mld (work_completion)(&(&idev->mc_dad_work)->work) &idev->mc_lock rcu_read_lock rcu_read_lock rcu_read_lock_bh rcu_read_lock &br->multicast_lock &obj_hash[i].lock irq_context: 0 (wq_completion)mld (work_completion)(&(&idev->mc_dad_work)->work) &idev->mc_lock rcu_read_lock rcu_read_lock rcu_read_lock_bh rcu_read_lock &br->multicast_lock &base->lock irq_context: 0 (wq_completion)mld (work_completion)(&(&idev->mc_dad_work)->work) &idev->mc_lock rcu_read_lock rcu_read_lock rcu_read_lock_bh rcu_read_lock &br->multicast_lock &base->lock &obj_hash[i].lock irq_context: 0 (wq_completion)mld (work_completion)(&(&idev->mc_dad_work)->work) &idev->mc_lock rcu_read_lock rcu_read_lock rcu_read_lock_bh rcu_read_lock pool_lock#2 irq_context: 0 (wq_completion)mld (work_completion)(&(&idev->mc_dad_work)->work) &idev->mc_lock rcu_read_lock rcu_read_lock rcu_read_lock_bh rcu_read_lock rcu_read_lock_bh &obj_hash[i].lock irq_context: 0 (wq_completion)mld (work_completion)(&(&idev->mc_dad_work)->work) &idev->mc_lock rcu_read_lock rcu_read_lock rcu_read_lock_bh rcu_read_lock rcu_read_lock_bh pool_lock#2 irq_context: 0 &type->s_umount_key#49/1 &s->s_inode_list_lock irq_context: 0 &type->s_umount_key#49/1 tk_core.seq.seqcount irq_context: 0 &type->s_umount_key#49/1 &sb->s_type->i_lock_key#32 &dentry->d_lock irq_context: 0 &type->s_umount_key#49/1 binderfs_minors_mutex irq_context: 0 &type->s_umount_key#49/1 binderfs_minors_mutex binderfs_minors.xa_lock irq_context: 0 &type->s_umount_key#49/1 &dentry->d_lock irq_context: 0 &type->s_umount_key#49/1 &sb->s_type->i_mutex_key#20 irq_context: 0 &type->s_umount_key#49/1 &sb->s_type->i_mutex_key#20 &sb->s_type->i_lock_key#32 irq_context: 0 &type->s_umount_key#49/1 &sb->s_type->i_mutex_key#20 rename_lock.seqcount irq_context: 0 &type->s_umount_key#49/1 &sb->s_type->i_mutex_key#20 fs_reclaim irq_context: 0 &type->s_umount_key#49/1 &sb->s_type->i_mutex_key#20 fs_reclaim mmu_notifier_invalidate_range_start irq_context: 0 &type->s_umount_key#49/1 &sb->s_type->i_mutex_key#20 &c->lock irq_context: 0 &type->s_umount_key#49/1 &sb->s_type->i_mutex_key#20 pool_lock#2 irq_context: 0 &type->s_umount_key#49/1 &sb->s_type->i_mutex_key#20 &dentry->d_lock irq_context: 0 &type->s_umount_key#49/1 &sb->s_type->i_mutex_key#20 rcu_read_lock rename_lock.seqcount irq_context: 0 &type->s_umount_key#49/1 &sb->s_type->i_mutex_key#20 &dentry->d_lock &wq irq_context: 0 &type->s_umount_key#49/1 &sb->s_type->i_mutex_key#20 &sb->s_type->i_lock_key#32 &dentry->d_lock irq_context: 0 &type->s_umount_key#49/1 &sb->s_type->i_mutex_key#20 mmu_notifier_invalidate_range_start irq_context: 0 &type->s_umount_key#49/1 &sb->s_type->i_mutex_key#20 &s->s_inode_list_lock irq_context: 0 &type->s_umount_key#49/1 &sb->s_type->i_mutex_key#20 rcu_read_lock iunique_lock irq_context: 0 &type->s_umount_key#49/1 &sb->s_type->i_mutex_key#20 tk_core.seq.seqcount irq_context: 0 &type->s_umount_key#49/1 &sb->s_type->i_mutex_key#20 &____s->seqcount irq_context: 0 &type->i_mutex_dir_key#2 namespace_sem &c->lock irq_context: 0 sb_writers#10 &type->i_mutex_dir_key#6/1 irq_context: 0 sb_writers#10 &type->i_mutex_dir_key#6/1 rename_lock.seqcount irq_context: 0 sb_writers#10 &type->i_mutex_dir_key#6/1 fs_reclaim irq_context: 0 sb_writers#10 &type->i_mutex_dir_key#6/1 fs_reclaim mmu_notifier_invalidate_range_start irq_context: 0 sb_writers#10 &type->i_mutex_dir_key#6/1 &dentry->d_lock irq_context: 0 sb_writers#10 &type->i_mutex_dir_key#6/1 &root->kernfs_rwsem irq_context: 0 sb_writers#10 &type->i_mutex_dir_key#6/1 tomoyo_ss irq_context: 0 sb_writers#10 &type->i_mutex_dir_key#6/1 tomoyo_ss mmu_notifier_invalidate_range_start irq_context: 0 sb_writers#10 &type->i_mutex_dir_key#6/1 tomoyo_ss &c->lock irq_context: 0 sb_writers#10 &type->i_mutex_dir_key#6/1 tomoyo_ss &____s->seqcount irq_context: 0 sb_writers#10 &type->i_mutex_dir_key#6/1 tomoyo_ss rcu_read_lock mount_lock.seqcount irq_context: 0 sb_writers#10 &type->i_mutex_dir_key#6/1 tomoyo_ss rcu_read_lock rcu_read_lock rename_lock.seqcount irq_context: 0 sb_writers#10 &type->i_mutex_dir_key#6/1 tomoyo_ss &obj_hash[i].lock irq_context: 0 sb_writers#10 &type->i_mutex_dir_key#6/1 &root->kernfs_iattr_rwsem irq_context: 0 sb_writers#10 &type->i_mutex_dir_key#6/1 cgroup_mutex irq_context: 0 sb_writers#10 &type->i_mutex_dir_key#6/1 cgroup_mutex fs_reclaim irq_context: 0 sb_writers#10 &type->i_mutex_dir_key#6/1 cgroup_mutex fs_reclaim mmu_notifier_invalidate_range_start irq_context: 0 sb_writers#10 &type->i_mutex_dir_key#6/1 cgroup_mutex pool_lock#2 irq_context: 0 sb_writers#10 &type->i_mutex_dir_key#6/1 cgroup_mutex pcpu_alloc_mutex irq_context: 0 sb_writers#10 &type->i_mutex_dir_key#6/1 cgroup_mutex pcpu_alloc_mutex pcpu_lock irq_context: 0 sb_writers#10 &type->i_mutex_dir_key#6/1 cgroup_mutex &c->lock irq_context: 0 sb_writers#10 &type->i_mutex_dir_key#6/1 cgroup_mutex &____s->seqcount irq_context: 0 sb_writers#10 &type->i_mutex_dir_key#6/1 cgroup_mutex lock irq_context: 0 sb_writers#10 &type->i_mutex_dir_key#6/1 cgroup_mutex lock kernfs_idr_lock irq_context: 0 sb_writers#10 &type->i_mutex_dir_key#6/1 cgroup_mutex &root->kernfs_rwsem irq_context: 0 sb_writers#10 &type->i_mutex_dir_key#6/1 cgroup_mutex &root->kernfs_rwsem &root->kernfs_iattr_rwsem irq_context: 0 sb_writers#10 &type->i_mutex_dir_key#6/1 cgroup_mutex &root->kernfs_rwsem &root->kernfs_iattr_rwsem tk_core.seq.seqcount irq_context: 0 sb_writers#10 &type->i_mutex_dir_key#6/1 cgroup_mutex &obj_hash[i].lock irq_context: 0 sb_writers#10 &type->i_mutex_dir_key#6/1 cgroup_mutex css_set_lock irq_context: 0 sb_writers#10 &type->i_mutex_dir_key#6/1 cgroup_mutex cgroup_file_kn_lock irq_context: 0 sb_writers#10 &type->i_mutex_dir_key#6/1 cgroup_mutex &obj_hash[i].lock pool_lock irq_context: 0 sb_writers#10 &type->i_mutex_dir_key#6/1 cgroup_mutex rcu_read_lock pool_lock#2 irq_context: 0 sb_writers#10 &type->i_mutex_dir_key#6/1 cgroup_mutex batched_entropy_u32.lock irq_context: 0 sb_writers#10 &type->i_mutex_dir_key#6/1 cgroup_mutex lock cgroup_idr_lock irq_context: 0 sb_writers#10 &type->i_mutex_dir_key#6/1 cgroup_mutex cgroup_idr_lock irq_context: 0 sb_writers#10 &type->i_mutex_dir_key#6/1 cgroup_mutex task_group_lock irq_context: 0 sb_writers#10 &type->i_mutex_dir_key#6/1 cgroup_mutex &rq->__lock irq_context: 0 sb_writers#10 &type->i_mutex_dir_key#6/1 cgroup_mutex &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 sb_writers#10 &type->i_mutex_dir_key#6/1 rcu_read_lock &dentry->d_lock irq_context: 0 sb_writers#10 &type->i_mutex_dir_key#6/1 &dentry->d_lock &lru->node[i].lock irq_context: 0 &type->i_mutex_dir_key#6 irq_context: 0 &type->i_mutex_dir_key#6 fs_reclaim irq_context: 0 &type->i_mutex_dir_key#6 fs_reclaim mmu_notifier_invalidate_range_start irq_context: 0 &type->i_mutex_dir_key#6 &dentry->d_lock irq_context: 0 &type->i_mutex_dir_key#6 rcu_read_lock rename_lock.seqcount irq_context: 0 &type->i_mutex_dir_key#6 &root->kernfs_rwsem irq_context: 0 &type->i_mutex_dir_key#6 &root->kernfs_rwsem inode_hash_lock irq_context: 0 &type->i_mutex_dir_key#6 &root->kernfs_rwsem fs_reclaim irq_context: 0 &type->i_mutex_dir_key#6 &root->kernfs_rwsem fs_reclaim mmu_notifier_invalidate_range_start irq_context: 0 &type->i_mutex_dir_key#6 &root->kernfs_rwsem &c->lock irq_context: 0 &type->i_mutex_dir_key#6 &root->kernfs_rwsem &____s->seqcount irq_context: 0 &type->i_mutex_dir_key#6 &root->kernfs_rwsem mmu_notifier_invalidate_range_start irq_context: 0 &type->i_mutex_dir_key#6 &root->kernfs_rwsem inode_hash_lock &sb->s_type->i_lock_key#30 irq_context: 0 &type->i_mutex_dir_key#6 &root->kernfs_rwsem inode_hash_lock &s->s_inode_list_lock irq_context: 0 &type->i_mutex_dir_key#6 &root->kernfs_rwsem tk_core.seq.seqcount irq_context: 0 &type->i_mutex_dir_key#6 &root->kernfs_rwsem &sb->s_type->i_lock_key#30 irq_context: 0 &type->i_mutex_dir_key#6 &sb->s_type->i_lock_key#30 irq_context: 0 &type->i_mutex_dir_key#6 &sb->s_type->i_lock_key#30 &dentry->d_lock irq_context: 0 &type->i_mutex_dir_key#6 &sb->s_type->i_lock_key#30 &dentry->d_lock &wq irq_context: 0 kn->active#53 fs_reclaim irq_context: 0 kn->active#53 fs_reclaim mmu_notifier_invalidate_range_start irq_context: 0 kn->active#53 &kernfs_locks->open_file_mutex[count] irq_context: 0 kn->active#53 &kernfs_locks->open_file_mutex[count] fs_reclaim irq_context: 0 kn->active#53 &kernfs_locks->open_file_mutex[count] fs_reclaim mmu_notifier_invalidate_range_start irq_context: 0 kn->active#54 fs_reclaim irq_context: 0 kn->active#54 fs_reclaim mmu_notifier_invalidate_range_start irq_context: 0 kn->active#54 &kernfs_locks->open_file_mutex[count] irq_context: 0 kn->active#54 &kernfs_locks->open_file_mutex[count] fs_reclaim irq_context: 0 kn->active#54 &kernfs_locks->open_file_mutex[count] fs_reclaim mmu_notifier_invalidate_range_start irq_context: 0 kn->active#54 &kernfs_locks->open_file_mutex[count] &c->lock irq_context: 0 kn->active#54 &kernfs_locks->open_file_mutex[count] &____s->seqcount irq_context: 0 sb_writers#10 &of->mutex cgroup_mutex cpu_hotplug_lock cgroup_threadgroup_rwsem irq_context: 0 sb_writers#10 &of->mutex cgroup_mutex cpu_hotplug_lock cgroup_threadgroup_rwsem cgroup_threadgroup_rwsem.rss.gp_wait.lock irq_context: 0 sb_writers#10 &of->mutex cgroup_mutex cpu_hotplug_lock cgroup_threadgroup_rwsem rcu_state.exp_mutex rcu_node_0 irq_context: 0 sb_writers#10 &of->mutex cgroup_mutex cpu_hotplug_lock cgroup_threadgroup_rwsem rcu_state.exp_mutex &obj_hash[i].lock irq_context: 0 sb_writers#10 &of->mutex cgroup_mutex cpu_hotplug_lock cgroup_threadgroup_rwsem rcu_state.exp_mutex rcu_read_lock &pool->lock irq_context: 0 sb_writers#10 &of->mutex cgroup_mutex cpu_hotplug_lock cgroup_threadgroup_rwsem rcu_state.exp_mutex rcu_read_lock &pool->lock &obj_hash[i].lock irq_context: 0 sb_writers#10 &of->mutex cgroup_mutex cpu_hotplug_lock cgroup_threadgroup_rwsem rcu_state.exp_mutex rcu_read_lock &pool->lock &p->pi_lock irq_context: 0 sb_writers#10 &of->mutex cgroup_mutex cpu_hotplug_lock cgroup_threadgroup_rwsem rcu_state.exp_mutex rcu_read_lock &pool->lock &p->pi_lock &rq->__lock irq_context: 0 sb_writers#10 &of->mutex cgroup_mutex cpu_hotplug_lock cgroup_threadgroup_rwsem rcu_state.exp_mutex rcu_read_lock &pool->lock &p->pi_lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 sb_writers#10 &of->mutex cgroup_mutex cpu_hotplug_lock cgroup_threadgroup_rwsem rcu_state.exp_mutex &rnp->exp_wq[0] irq_context: 0 sb_writers#10 &of->mutex cgroup_mutex cpu_hotplug_lock cgroup_threadgroup_rwsem rcu_state.exp_mutex &rq->__lock irq_context: 0 rtnl_mutex _xmit_ETHER console_owner_lock irq_context: 0 rtnl_mutex _xmit_ETHER console_owner irq_context: 0 (wq_completion)hci2#2 (work_completion)(&(&hdev->cmd_timer)->work) rcu_node_0 irq_context: 0 (wq_completion)hci2#2 (work_completion)(&(&hdev->cmd_timer)->work) &rcu_state.expedited_wq irq_context: 0 (wq_completion)hci2#2 (work_completion)(&(&hdev->cmd_timer)->work) &rcu_state.expedited_wq &p->pi_lock irq_context: 0 (wq_completion)hci2#2 (work_completion)(&(&hdev->cmd_timer)->work) &rcu_state.expedited_wq &p->pi_lock &rq->__lock irq_context: 0 (wq_completion)hci2#2 (work_completion)(&(&hdev->cmd_timer)->work) &rcu_state.expedited_wq &p->pi_lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 sb_writers#10 &of->mutex cgroup_mutex cpu_hotplug_lock cgroup_threadgroup_rwsem rcu_state.exp_mutex &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 sb_writers#10 &of->mutex cgroup_mutex cpu_hotplug_lock cgroup_threadgroup_rwsem rcu_state.exp_mutex rcu_read_lock &rq->__lock irq_context: 0 sb_writers#10 &of->mutex cgroup_mutex cpu_hotplug_lock cgroup_threadgroup_rwsem rcu_state.exp_mutex rcu_read_lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 sb_writers#10 &of->mutex cgroup_mutex cpu_hotplug_lock cgroup_threadgroup_rwsem &obj_hash[i].lock irq_context: 0 sb_writers#10 &of->mutex cgroup_mutex cpu_hotplug_lock cgroup_threadgroup_rwsem css_set_lock irq_context: 0 sb_writers#10 &of->mutex cgroup_mutex cpu_hotplug_lock cgroup_threadgroup_rwsem inode_hash_lock irq_context: 0 sb_writers#10 &of->mutex cgroup_mutex cpu_hotplug_lock cgroup_threadgroup_rwsem fs_reclaim irq_context: 0 sb_writers#10 &of->mutex cgroup_mutex cpu_hotplug_lock cgroup_threadgroup_rwsem fs_reclaim mmu_notifier_invalidate_range_start irq_context: softirq rcu_read_lock hwsim_radio_lock pool_lock#2 irq_context: softirq rcu_read_lock hwsim_radio_lock &list->lock#19 irq_context: softirq rcu_read_lock hwsim_radio_lock &c->lock irq_context: softirq &list->lock#19 irq_context: softirq rcu_read_lock lock#6 irq_context: softirq rcu_read_lock lock#6 kcov_remote_lock irq_context: softirq rcu_read_lock &local->rx_path_lock irq_context: softirq rcu_read_lock &local->rx_path_lock &list->lock#18 irq_context: softirq rcu_read_lock &local->rx_path_lock &rdev->wiphy_work_lock irq_context: softirq rcu_read_lock &local->rx_path_lock rcu_read_lock &pool->lock irq_context: softirq rcu_read_lock &local->rx_path_lock rcu_read_lock &pool->lock &obj_hash[i].lock irq_context: softirq rcu_read_lock &local->rx_path_lock rcu_read_lock &pool->lock &p->pi_lock irq_context: softirq rcu_read_lock &local->rx_path_lock rcu_read_lock &pool->lock &p->pi_lock &rq->__lock irq_context: softirq rcu_read_lock &local->rx_path_lock rcu_read_lock &pool->lock &p->pi_lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: softirq rcu_read_lock &local->rx_path_lock rcu_read_lock &pool->lock &p->pi_lock &rq->__lock &cfs_rq->removed.lock irq_context: 0 sb_writers#10 &of->mutex cgroup_mutex cpu_hotplug_lock cgroup_threadgroup_rwsem fs_reclaim mmu_notifier_invalidate_range_start &rq->__lock irq_context: 0 (wq_completion)events (work_completion)(&rdev->wiphy_work) &rdev->wiphy.mtx lock#6 irq_context: 0 (wq_completion)events (work_completion)(&rdev->wiphy_work) &rdev->wiphy.mtx lock#6 kcov_remote_lock irq_context: 0 (wq_completion)events (work_completion)(&rdev->wiphy_work) &rdev->wiphy.mtx &wdev->mtx &local->chanctx_mtx irq_context: 0 (wq_completion)events (work_completion)(&rdev->wiphy_work) &rdev->wiphy.mtx &wdev->mtx &sta->lock irq_context: 0 (wq_completion)events (work_completion)(&rdev->wiphy_work) &rdev->wiphy.mtx &wdev->mtx &sta->lock pool_lock#2 irq_context: 0 (wq_completion)events (work_completion)(&rdev->wiphy_work) &rdev->wiphy.mtx &wdev->mtx rcu_read_lock &sta->rate_ctrl_lock irq_context: 0 (wq_completion)events (work_completion)(&rdev->wiphy_work) &rdev->wiphy.mtx &wdev->mtx rcu_read_lock &sta->rate_ctrl_lock pool_lock#2 irq_context: 0 (wq_completion)events (work_completion)(&rdev->wiphy_work) &rdev->wiphy.mtx &wdev->mtx &local->sta_mtx irq_context: 0 (wq_completion)events (work_completion)(&rdev->wiphy_work) &rdev->wiphy.mtx &wdev->mtx &local->sta_mtx fs_reclaim irq_context: 0 (wq_completion)events (work_completion)(&rdev->wiphy_work) &rdev->wiphy.mtx &wdev->mtx &local->sta_mtx fs_reclaim mmu_notifier_invalidate_range_start irq_context: 0 (wq_completion)events (work_completion)(&rdev->wiphy_work) &rdev->wiphy.mtx &wdev->mtx &local->sta_mtx pool_lock#2 irq_context: 0 (wq_completion)events (work_completion)(&rdev->wiphy_work) &rdev->wiphy.mtx &wdev->mtx &local->sta_mtx rcu_read_lock rhashtable_bucket irq_context: 0 (wq_completion)events (work_completion)(&rdev->wiphy_work) &rdev->wiphy.mtx &wdev->mtx &local->sta_mtx &local->chanctx_mtx irq_context: 0 (wq_completion)events (work_completion)(&rdev->wiphy_work) &rdev->wiphy.mtx &wdev->mtx &local->sta_mtx pin_fs_lock irq_context: 0 (wq_completion)events (work_completion)(&rdev->wiphy_work) &rdev->wiphy.mtx &wdev->mtx &local->sta_mtx &sb->s_type->i_mutex_key#3 irq_context: 0 (wq_completion)events (work_completion)(&rdev->wiphy_work) &rdev->wiphy.mtx &wdev->mtx &local->sta_mtx &sb->s_type->i_mutex_key#3 &sb->s_type->i_lock_key#7 irq_context: 0 (wq_completion)events (work_completion)(&rdev->wiphy_work) &rdev->wiphy.mtx &wdev->mtx &local->sta_mtx &sb->s_type->i_mutex_key#3 rename_lock.seqcount irq_context: 0 (wq_completion)events (work_completion)(&rdev->wiphy_work) &rdev->wiphy.mtx &wdev->mtx &local->sta_mtx &sb->s_type->i_mutex_key#3 fs_reclaim irq_context: 0 (wq_completion)events (work_completion)(&rdev->wiphy_work) &rdev->wiphy.mtx &wdev->mtx &local->sta_mtx &sb->s_type->i_mutex_key#3 fs_reclaim mmu_notifier_invalidate_range_start irq_context: 0 (wq_completion)events (work_completion)(&rdev->wiphy_work) &rdev->wiphy.mtx &wdev->mtx &local->sta_mtx &sb->s_type->i_mutex_key#3 pool_lock#2 irq_context: 0 (wq_completion)events (work_completion)(&rdev->wiphy_work) &rdev->wiphy.mtx &wdev->mtx &local->sta_mtx &sb->s_type->i_mutex_key#3 &dentry->d_lock irq_context: 0 (wq_completion)events (work_completion)(&rdev->wiphy_work) &rdev->wiphy.mtx &wdev->mtx &local->sta_mtx &sb->s_type->i_mutex_key#3 rcu_read_lock rename_lock.seqcount irq_context: 0 (wq_completion)events (work_completion)(&rdev->wiphy_work) &rdev->wiphy.mtx &wdev->mtx &local->sta_mtx &sb->s_type->i_mutex_key#3 &dentry->d_lock &wq irq_context: 0 (wq_completion)events (work_completion)(&rdev->wiphy_work) &rdev->wiphy.mtx &wdev->mtx &local->sta_mtx &sb->s_type->i_mutex_key#3 mmu_notifier_invalidate_range_start irq_context: 0 (wq_completion)events (work_completion)(&rdev->wiphy_work) &rdev->wiphy.mtx &wdev->mtx &local->sta_mtx &sb->s_type->i_mutex_key#3 &s->s_inode_list_lock irq_context: 0 (wq_completion)events (work_completion)(&rdev->wiphy_work) &rdev->wiphy.mtx &wdev->mtx &local->sta_mtx &sb->s_type->i_mutex_key#3 tk_core.seq.seqcount irq_context: 0 (wq_completion)events (work_completion)(&rdev->wiphy_work) &rdev->wiphy.mtx &wdev->mtx &local->sta_mtx &sb->s_type->i_mutex_key#3 &sb->s_type->i_lock_key#7 &dentry->d_lock irq_context: 0 (wq_completion)events (work_completion)(&rdev->wiphy_work) &rdev->wiphy.mtx &wdev->mtx &local->sta_mtx &sb->s_type->i_mutex_key#3 &c->lock irq_context: 0 (wq_completion)events (work_completion)(&rdev->wiphy_work) &rdev->wiphy.mtx &wdev->mtx &local->sta_mtx &sb->s_type->i_mutex_key#3 &____s->seqcount irq_context: 0 (wq_completion)events (work_completion)(&rdev->wiphy_work) &rdev->wiphy.mtx &wdev->mtx &local->sta_mtx &obj_hash[i].lock irq_context: 0 (wq_completion)events (work_completion)(&rdev->wiphy_work) &rdev->wiphy.mtx &wdev->mtx &local->sta_mtx nl_table_lock irq_context: 0 (wq_completion)events (work_completion)(&rdev->wiphy_work) &rdev->wiphy.mtx &wdev->mtx &local->sta_mtx nl_table_wait.lock irq_context: 0 (wq_completion)events (work_completion)(&rdev->wiphy_work) &rdev->wiphy.mtx &wdev->mtx rcu_read_lock &sta->lock irq_context: 0 (wq_completion)events (work_completion)(&rdev->wiphy_work) &rdev->wiphy.mtx &wdev->mtx rcu_read_lock &sta->lock pool_lock#2 irq_context: 0 (wq_completion)events (work_completion)(&rdev->wiphy_work) &rdev->wiphy.mtx &wdev->mtx rcu_read_lock &sta->lock &obj_hash[i].lock irq_context: 0 (wq_completion)events (work_completion)(&rdev->wiphy_work) &rdev->wiphy.mtx &wdev->mtx rcu_read_lock &sta->lock krc.lock irq_context: 0 (wq_completion)events (work_completion)(&rdev->wiphy_work) &rdev->wiphy.mtx &wdev->mtx &rdev->bss_lock &obj_hash[i].lock irq_context: 0 (wq_completion)events (work_completion)(&rdev->wiphy_work) &rdev->wiphy.mtx &obj_hash[i].lock irq_context: 0 (wq_completion)events (work_completion)(&rdev->wiphy_work) &rdev->wiphy.mtx pool_lock#2 irq_context: 0 sb_writers#10 &of->mutex cgroup_mutex cpu_hotplug_lock cgroup_threadgroup_rwsem pool_lock#2 irq_context: 0 sb_writers#10 &of->mutex cgroup_mutex cpu_hotplug_lock cgroup_threadgroup_rwsem mmu_notifier_invalidate_range_start irq_context: 0 sb_writers#10 &of->mutex cgroup_mutex cpu_hotplug_lock cgroup_threadgroup_rwsem inode_hash_lock &sb->s_type->i_lock_key#30 irq_context: 0 sb_writers#10 &of->mutex cgroup_mutex cpu_hotplug_lock cgroup_threadgroup_rwsem inode_hash_lock &s->s_inode_list_lock irq_context: 0 sb_writers#10 &of->mutex cgroup_mutex cpu_hotplug_lock cgroup_threadgroup_rwsem tk_core.seq.seqcount irq_context: 0 sb_writers#10 &of->mutex cgroup_mutex cpu_hotplug_lock cgroup_threadgroup_rwsem &sb->s_type->i_lock_key#30 irq_context: 0 sb_writers#10 &of->mutex cgroup_mutex cpu_hotplug_lock cgroup_threadgroup_rwsem &root->kernfs_iattr_rwsem irq_context: 0 sb_writers#10 &of->mutex cgroup_mutex cpu_hotplug_lock cgroup_threadgroup_rwsem &s->s_inode_list_lock irq_context: 0 sb_writers#10 &of->mutex cgroup_mutex cpu_hotplug_lock cgroup_threadgroup_rwsem &xa->xa_lock#9 irq_context: 0 sb_writers#10 &of->mutex cgroup_mutex cpu_hotplug_lock cgroup_threadgroup_rwsem &fsnotify_mark_srcu irq_context: 0 sb_writers#10 &of->mutex cgroup_mutex cpu_hotplug_lock cgroup_threadgroup_rwsem css_set_lock cgroup_file_kn_lock irq_context: 0 sb_writers#10 &of->mutex cgroup_mutex cpu_hotplug_lock cgroup_threadgroup_rwsem css_set_lock cgroup_file_kn_lock kernfs_notify_lock irq_context: 0 sb_writers#10 &of->mutex cgroup_mutex cpu_hotplug_lock cgroup_threadgroup_rwsem css_set_lock cgroup_file_kn_lock kernfs_notify_lock rcu_read_lock &pool->lock irq_context: 0 sb_writers#10 &of->mutex cgroup_mutex cpu_hotplug_lock cgroup_threadgroup_rwsem css_set_lock cgroup_file_kn_lock kernfs_notify_lock rcu_read_lock &pool->lock &obj_hash[i].lock irq_context: 0 sb_writers#10 &of->mutex cgroup_mutex cpu_hotplug_lock cgroup_threadgroup_rwsem css_set_lock cgroup_file_kn_lock kernfs_notify_lock rcu_read_lock &pool->lock &p->pi_lock irq_context: 0 sb_writers#10 &of->mutex cgroup_mutex cpu_hotplug_lock cgroup_threadgroup_rwsem css_set_lock cgroup_file_kn_lock kernfs_notify_lock rcu_read_lock &pool->lock &p->pi_lock &rq->__lock irq_context: 0 sb_writers#10 &of->mutex cgroup_mutex cpu_hotplug_lock cgroup_threadgroup_rwsem css_set_lock cgroup_file_kn_lock kernfs_notify_lock rcu_read_lock &pool->lock &p->pi_lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 sb_writers#10 &of->mutex cgroup_mutex cpu_hotplug_lock cgroup_threadgroup_rwsem css_set_lock &p->pi_lock irq_context: 0 sb_writers#10 &of->mutex cgroup_mutex cpu_hotplug_lock cgroup_threadgroup_rwsem css_set_lock &p->pi_lock &rq->__lock irq_context: 0 sb_writers#10 &of->mutex cgroup_mutex cpu_hotplug_lock cgroup_threadgroup_rwsem css_set_lock &p->pi_lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 sb_writers#10 &of->mutex cgroup_mutex cpu_hotplug_lock cgroup_threadgroup_rwsem css_set_lock &p->pi_lock &rq->__lock &obj_hash[i].lock irq_context: 0 sb_writers#10 &of->mutex cgroup_mutex cpu_hotplug_lock cgroup_threadgroup_rwsem css_set_lock &p->pi_lock &rq->__lock &base->lock irq_context: 0 sb_writers#10 &of->mutex cgroup_mutex cpu_hotplug_lock cgroup_threadgroup_rwsem css_set_lock &p->pi_lock &rq->__lock &base->lock &obj_hash[i].lock irq_context: 0 sb_writers#10 &of->mutex cgroup_mutex cpu_hotplug_lock cgroup_threadgroup_rwsem &p->pi_lock irq_context: 0 sb_writers#10 &of->mutex cgroup_mutex cpu_hotplug_lock cgroup_threadgroup_rwsem &p->pi_lock &rq->__lock irq_context: 0 sb_writers#10 &of->mutex cgroup_mutex cpu_hotplug_lock cgroup_threadgroup_rwsem &p->pi_lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 sb_writers#10 &of->mutex cgroup_mutex cpu_hotplug_lock cgroup_threadgroup_rwsem &p->pi_lock &rq->__lock per_cpu_ptr(&cgroup_rstat_cpu_lock, cpu) irq_context: 0 sb_writers#10 &of->mutex cgroup_mutex cpu_hotplug_lock cgroup_threadgroup_rwsem &rq->__lock irq_context: 0 sb_writers#10 &of->mutex cgroup_mutex cpu_hotplug_lock cgroup_threadgroup_rwsem &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)events kernfs_notify_work &root->kernfs_supers_rwsem inode_hash_lock irq_context: 0 sb_writers#10 &of->mutex cgroup_mutex cpu_hotplug_lock cgroup_threadgroup_rwsem.waiters.lock irq_context: 0 sb_writers#10 &of->mutex cgroup_mutex cpu_hotplug_lock cgroup_threadgroup_rwsem.rss.gp_wait.lock irq_context: 0 sb_writers#10 &of->mutex cgroup_mutex cpu_hotplug_lock cgroup_threadgroup_rwsem.rss.gp_wait.lock &obj_hash[i].lock irq_context: 0 sb_writers#10 &of->mutex cgroup_mutex (wq_completion)cpuset_migrate_mm irq_context: 0 sb_writers#10 &of->mutex cgroup_mutex &wq->mutex irq_context: 0 sb_writers#10 &of->mutex cgroup_mutex &wq->mutex &pool->lock/1 irq_context: 0 sb_writers#10 &of->mutex cgroup_mutex &wq->mutex &x->wait#10 irq_context: 0 sb_writers#11 &type->i_mutex_dir_key#7/1 irq_context: 0 sb_writers#11 &type->i_mutex_dir_key#7/1 &rq->__lock irq_context: 0 sb_writers#11 &type->i_mutex_dir_key#7/1 &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 sb_writers#11 &type->i_mutex_dir_key#7/1 rename_lock.seqcount irq_context: 0 sb_writers#11 &type->i_mutex_dir_key#7/1 fs_reclaim irq_context: 0 sb_writers#11 &type->i_mutex_dir_key#7/1 fs_reclaim mmu_notifier_invalidate_range_start irq_context: 0 sb_writers#11 &type->i_mutex_dir_key#7/1 &dentry->d_lock irq_context: 0 sb_writers#11 &type->i_mutex_dir_key#7/1 &root->kernfs_rwsem irq_context: 0 sb_writers#11 &type->i_mutex_dir_key#7/1 tomoyo_ss irq_context: 0 sb_writers#11 &type->i_mutex_dir_key#7/1 tomoyo_ss mmu_notifier_invalidate_range_start irq_context: 0 sb_writers#11 &type->i_mutex_dir_key#7/1 tomoyo_ss rcu_read_lock mount_lock.seqcount irq_context: 0 sb_writers#11 &type->i_mutex_dir_key#7/1 tomoyo_ss rcu_read_lock rcu_read_lock rename_lock.seqcount irq_context: 0 sb_writers#11 &type->i_mutex_dir_key#7/1 tomoyo_ss &obj_hash[i].lock irq_context: 0 sb_writers#11 &type->i_mutex_dir_key#7/1 &root->kernfs_iattr_rwsem irq_context: 0 sb_writers#11 &type->i_mutex_dir_key#7/1 cgroup_mutex irq_context: 0 sb_writers#11 &type->i_mutex_dir_key#7/1 cgroup_mutex fs_reclaim irq_context: 0 sb_writers#11 &type->i_mutex_dir_key#7/1 cgroup_mutex fs_reclaim mmu_notifier_invalidate_range_start irq_context: 0 sb_writers#11 &type->i_mutex_dir_key#7/1 cgroup_mutex &c->lock irq_context: 0 sb_writers#11 &type->i_mutex_dir_key#7/1 cgroup_mutex &____s->seqcount irq_context: 0 sb_writers#11 &type->i_mutex_dir_key#7/1 cgroup_mutex pool_lock#2 irq_context: 0 sb_writers#11 &type->i_mutex_dir_key#7/1 cgroup_mutex pcpu_alloc_mutex irq_context: 0 sb_writers#11 &type->i_mutex_dir_key#7/1 cgroup_mutex pcpu_alloc_mutex pcpu_lock irq_context: 0 sb_writers#11 &type->i_mutex_dir_key#7/1 cgroup_mutex lock irq_context: 0 sb_writers#11 &type->i_mutex_dir_key#7/1 cgroup_mutex lock kernfs_idr_lock irq_context: 0 sb_writers#11 &type->i_mutex_dir_key#7/1 cgroup_mutex &root->kernfs_rwsem irq_context: 0 sb_writers#11 &type->i_mutex_dir_key#7/1 cgroup_mutex &root->kernfs_rwsem &root->kernfs_iattr_rwsem irq_context: 0 sb_writers#11 &type->i_mutex_dir_key#7/1 cgroup_mutex &root->kernfs_rwsem &root->kernfs_iattr_rwsem tk_core.seq.seqcount irq_context: 0 sb_writers#11 &type->i_mutex_dir_key#7/1 cgroup_mutex &obj_hash[i].lock irq_context: 0 sb_writers#11 &type->i_mutex_dir_key#7/1 cgroup_mutex css_set_lock irq_context: 0 sb_writers#11 &type->i_mutex_dir_key#7/1 cgroup_mutex lock cgroup_idr_lock irq_context: 0 sb_writers#11 &type->i_mutex_dir_key#7/1 cgroup_mutex cgroup_idr_lock irq_context: 0 sb_writers#11 &type->i_mutex_dir_key#7/1 cgroup_mutex cpu_hotplug_lock irq_context: 0 sb_writers#11 &type->i_mutex_dir_key#7/1 cgroup_mutex cpu_hotplug_lock cpuset_mutex irq_context: 0 sb_writers#11 &type->i_mutex_dir_key#7/1 cgroup_mutex cpu_hotplug_lock cpuset_mutex jump_label_mutex irq_context: 0 sb_writers#11 &type->i_mutex_dir_key#7/1 cgroup_mutex cpu_hotplug_lock cpuset_mutex jump_label_mutex text_mutex irq_context: 0 sb_writers#11 &type->i_mutex_dir_key#7/1 cgroup_mutex cpu_hotplug_lock cpuset_mutex jump_label_mutex text_mutex ptlock_ptr(page)#2 irq_context: 0 sb_writers#11 &type->i_mutex_dir_key#7/1 cgroup_mutex cpu_hotplug_lock cpuset_mutex callback_lock irq_context: 0 sb_writers#11 &type->i_mutex_dir_key#7/1 cgroup_mutex &____s->seqcount#2 irq_context: 0 sb_writers#11 &type->i_mutex_dir_key#7/1 cgroup_mutex rcu_read_lock &pool->lock irq_context: 0 sb_writers#11 &type->i_mutex_dir_key#7/1 cgroup_mutex rcu_read_lock &pool->lock &obj_hash[i].lock irq_context: 0 sb_writers#11 &type->i_mutex_dir_key#7/1 cgroup_mutex rcu_read_lock &pool->lock &p->pi_lock irq_context: 0 sb_writers#11 &type->i_mutex_dir_key#7/1 cgroup_mutex rcu_read_lock &pool->lock &p->pi_lock &rq->__lock irq_context: 0 rtnl_mutex rcu_read_lock &____s->seqcount#2 irq_context: 0 sb_writers#11 &type->i_mutex_dir_key#7/1 cgroup_mutex rcu_read_lock &pool->lock &p->pi_lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 sb_writers#11 &type->i_mutex_dir_key#7/1 cgroup_mutex &rq->__lock irq_context: 0 sb_writers#11 &type->i_mutex_dir_key#7/1 cgroup_mutex &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 tomoyo_ss &____s->seqcount#2 irq_context: 0 rtnl_mutex &____s->seqcount#2 irq_context: 0 sb_writers#11 &type->i_mutex_dir_key#7/1 cgroup_mutex &n->list_lock irq_context: 0 sb_writers#11 &type->i_mutex_dir_key#7/1 cgroup_mutex &n->list_lock &c->lock irq_context: hardirq per_cpu_ptr(&cgroup_rstat_cpu_lock, cpu) irq_context: 0 (wq_completion)mld (work_completion)(&(&idev->mc_dad_work)->work) &idev->mc_lock &____s->seqcount#2 irq_context: 0 rtnl_mutex &ndev->lock &tb->tb6_lock &____s->seqcount#2 irq_context: 0 sb_writers#11 &type->i_mutex_dir_key#7/1 cgroup_mutex &obj_hash[i].lock pool_lock irq_context: 0 sb_writers#11 &type->i_mutex_dir_key#7/1 cgroup_mutex percpu_counters_lock irq_context: 0 sb_writers#11 &type->i_mutex_dir_key#7/1 cgroup_mutex cpu_hotplug_lock jump_label_mutex irq_context: 0 sb_writers#11 &type->i_mutex_dir_key#7/1 cgroup_mutex cpu_hotplug_lock jump_label_mutex text_mutex irq_context: 0 sb_writers#11 &type->i_mutex_dir_key#7/1 cgroup_mutex cpu_hotplug_lock jump_label_mutex text_mutex ptlock_ptr(page)#2 irq_context: 0 sb_writers#11 &type->i_mutex_dir_key#7/1 cgroup_mutex shrinker_rwsem irq_context: 0 sb_writers#11 &type->i_mutex_dir_key#7/1 cgroup_mutex shrinker_rwsem fs_reclaim irq_context: 0 sb_writers#11 &type->i_mutex_dir_key#7/1 cgroup_mutex shrinker_rwsem fs_reclaim mmu_notifier_invalidate_range_start irq_context: 0 sb_writers#11 &type->i_mutex_dir_key#7/1 cgroup_mutex shrinker_rwsem pool_lock#2 irq_context: 0 sb_writers#11 &type->i_mutex_dir_key#7/1 cgroup_mutex shrinker_rwsem &c->lock irq_context: 0 sb_writers#11 &type->i_mutex_dir_key#7/1 cgroup_mutex shrinker_rwsem &n->list_lock irq_context: 0 sb_writers#11 &type->i_mutex_dir_key#7/1 cgroup_mutex shrinker_rwsem &n->list_lock &c->lock irq_context: 0 sb_writers#11 &type->i_mutex_dir_key#7/1 cgroup_mutex batched_entropy_u8.lock irq_context: 0 sb_writers#11 &type->i_mutex_dir_key#7/1 cgroup_mutex &pgdat->memcg_lru.lock irq_context: 0 sb_writers#11 &type->i_mutex_dir_key#7/1 cgroup_mutex lock kernfs_idr_lock pool_lock#2 irq_context: 0 (wq_completion)ipv6_addrconf (work_completion)(&(&ifa->dad_work)->work) rtnl_mutex &____s->seqcount#2 irq_context: 0 sb_writers#11 &type->i_mutex_dir_key#7/1 rcu_read_lock &dentry->d_lock irq_context: 0 sb_writers#11 &type->i_mutex_dir_key#7/1 &dentry->d_lock &lru->node[i].lock irq_context: 0 &____s->seqcount#2 irq_context: 0 &type->i_mutex_dir_key#7 irq_context: 0 &type->i_mutex_dir_key#7 fs_reclaim irq_context: 0 &type->i_mutex_dir_key#7 fs_reclaim mmu_notifier_invalidate_range_start irq_context: 0 &type->i_mutex_dir_key#7 &dentry->d_lock irq_context: 0 &type->i_mutex_dir_key#7 rcu_read_lock rename_lock.seqcount irq_context: 0 &type->i_mutex_dir_key#7 &root->kernfs_rwsem irq_context: 0 &type->i_mutex_dir_key#7 &root->kernfs_rwsem inode_hash_lock irq_context: 0 &type->i_mutex_dir_key#7 &root->kernfs_rwsem fs_reclaim irq_context: 0 &type->i_mutex_dir_key#7 &root->kernfs_rwsem fs_reclaim mmu_notifier_invalidate_range_start irq_context: 0 &type->i_mutex_dir_key#7 &root->kernfs_rwsem mmu_notifier_invalidate_range_start irq_context: 0 &type->i_mutex_dir_key#7 &root->kernfs_rwsem inode_hash_lock &sb->s_type->i_lock_key#31 irq_context: 0 &type->i_mutex_dir_key#7 &root->kernfs_rwsem inode_hash_lock &s->s_inode_list_lock irq_context: 0 &type->i_mutex_dir_key#7 &root->kernfs_rwsem tk_core.seq.seqcount irq_context: 0 &type->i_mutex_dir_key#7 &root->kernfs_rwsem &sb->s_type->i_lock_key#31 irq_context: 0 &type->i_mutex_dir_key#7 &sb->s_type->i_lock_key#31 irq_context: 0 &type->i_mutex_dir_key#7 &sb->s_type->i_lock_key#31 &dentry->d_lock irq_context: 0 &type->i_mutex_dir_key#7 &sb->s_type->i_lock_key#31 &dentry->d_lock &wq irq_context: 0 (wq_completion)ipv6_addrconf (work_completion)(&(&ifa->dad_work)->work) rtnl_mutex &idev->mc_lock &____s->seqcount#2 irq_context: 0 kn->active#55 fs_reclaim irq_context: 0 kn->active#55 fs_reclaim mmu_notifier_invalidate_range_start irq_context: 0 kn->active#55 &kernfs_locks->open_file_mutex[count] irq_context: 0 kn->active#55 &kernfs_locks->open_file_mutex[count] fs_reclaim irq_context: 0 kn->active#55 &kernfs_locks->open_file_mutex[count] fs_reclaim mmu_notifier_invalidate_range_start irq_context: 0 kn->active#55 &kernfs_locks->open_file_mutex[count] &c->lock irq_context: 0 kn->active#55 &kernfs_locks->open_file_mutex[count] &____s->seqcount#2 irq_context: 0 kn->active#55 &kernfs_locks->open_file_mutex[count] &____s->seqcount irq_context: 0 sb_writers#11 &of->mutex cgroup_mutex irq_context: 0 sb_writers#11 &of->mutex cgroup_mutex cpu_hotplug_lock irq_context: 0 sb_writers#11 &of->mutex cgroup_mutex cpu_hotplug_lock cgroup_threadgroup_rwsem irq_context: 0 sb_writers#11 &of->mutex cgroup_mutex cpu_hotplug_lock cgroup_threadgroup_rwsem cgroup_threadgroup_rwsem.rss.gp_wait.lock irq_context: 0 sb_writers#11 &of->mutex cgroup_mutex cpu_hotplug_lock cgroup_threadgroup_rwsem css_set_lock irq_context: 0 sb_writers#11 &of->mutex cgroup_mutex cpu_hotplug_lock cgroup_threadgroup_rwsem fs_reclaim irq_context: 0 sb_writers#11 &of->mutex cgroup_mutex cpu_hotplug_lock cgroup_threadgroup_rwsem fs_reclaim mmu_notifier_invalidate_range_start irq_context: 0 (wq_completion)ipv6_addrconf (work_completion)(&(&ifa->dad_work)->work) rtnl_mutex &idev->mc_lock rcu_read_lock rcu_read_lock rcu_read_lock_bh rcu_read_lock &br->multicast_lock &base->lock &obj_hash[i].lock irq_context: 0 sb_writers#11 &of->mutex cgroup_mutex cpu_hotplug_lock cgroup_threadgroup_rwsem pool_lock#2 irq_context: 0 sb_writers#11 &of->mutex cgroup_mutex cpu_hotplug_lock cgroup_threadgroup_rwsem cpuset_mutex irq_context: 0 sb_writers#11 &of->mutex cgroup_mutex cpu_hotplug_lock cgroup_threadgroup_rwsem css_set_lock cgroup_file_kn_lock irq_context: 0 sb_writers#11 &of->mutex cgroup_mutex cpu_hotplug_lock cgroup_threadgroup_rwsem css_set_lock &p->pi_lock irq_context: 0 sb_writers#11 &of->mutex cgroup_mutex cpu_hotplug_lock cgroup_threadgroup_rwsem css_set_lock &p->pi_lock &rq->__lock irq_context: 0 sb_writers#11 &of->mutex cgroup_mutex cpu_hotplug_lock cgroup_threadgroup_rwsem css_set_lock &p->pi_lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 sb_writers#11 &of->mutex cgroup_mutex cpu_hotplug_lock cgroup_threadgroup_rwsem cpuset_mutex &p->pi_lock irq_context: 0 sb_writers#11 &of->mutex cgroup_mutex cpu_hotplug_lock cgroup_threadgroup_rwsem cpuset_mutex &p->pi_lock &rq->__lock irq_context: 0 sb_writers#11 &of->mutex cgroup_mutex cpu_hotplug_lock cgroup_threadgroup_rwsem cpuset_mutex &p->pi_lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 sb_writers#11 &of->mutex cgroup_mutex cpu_hotplug_lock cgroup_threadgroup_rwsem cpuset_mutex &p->alloc_lock irq_context: 0 sb_writers#11 &of->mutex cgroup_mutex cpu_hotplug_lock cgroup_threadgroup_rwsem cpuset_mutex &p->alloc_lock &____s->seqcount#2 irq_context: 0 sb_writers#11 &of->mutex cgroup_mutex cpu_hotplug_lock cgroup_threadgroup_rwsem cpuset_mutex cpuset_attach_wq.lock irq_context: 0 sb_writers#11 &of->mutex cgroup_mutex cpu_hotplug_lock cgroup_threadgroup_rwsem &p->alloc_lock irq_context: 0 sb_writers#11 &of->mutex cgroup_mutex cpu_hotplug_lock cgroup_threadgroup_rwsem &p->alloc_lock &memcg->mm_list.lock irq_context: 0 sb_writers#11 &of->mutex cgroup_mutex cpu_hotplug_lock cgroup_threadgroup_rwsem css_set_lock &obj_hash[i].lock irq_context: 0 sb_writers#11 &of->mutex cgroup_mutex cpu_hotplug_lock cgroup_threadgroup_rwsem css_set_lock pool_lock#2 irq_context: softirq rcu_read_lock rcu_read_lock &____s->seqcount#2 irq_context: 0 sb_writers#11 &of->mutex cgroup_mutex cpu_hotplug_lock cgroup_threadgroup_rwsem css_set_lock krc.lock irq_context: 0 sb_writers#11 &of->mutex cgroup_mutex cpu_hotplug_lock cgroup_threadgroup_rwsem.waiters.lock irq_context: 0 sb_writers#11 &of->mutex cgroup_mutex cpu_hotplug_lock cgroup_threadgroup_rwsem.rss.gp_wait.lock irq_context: 0 sb_writers#11 &of->mutex cgroup_mutex (wq_completion)cpuset_migrate_mm irq_context: 0 sb_writers#11 &of->mutex cgroup_mutex &wq->mutex irq_context: 0 sb_writers#11 &of->mutex cgroup_mutex &wq->mutex &pool->lock/1 irq_context: 0 sb_writers#11 &of->mutex cgroup_mutex &wq->mutex &x->wait#10 irq_context: 0 (wq_completion)ipv6_addrconf (work_completion)(&(&ifa->dad_work)->work) rtnl_mutex &tb->tb6_lock &____s->seqcount#2 irq_context: 0 stock_lock irq_context: 0 sb_writers#11 &type->i_mutex_dir_key#7 pool_lock#2 irq_context: 0 sb_writers#11 &type->i_mutex_dir_key#7 &xa->xa_lock#4 irq_context: 0 sb_writers#11 &type->i_mutex_dir_key#7 &xa->xa_lock#4 pool_lock#2 irq_context: 0 sb_writers#11 &type->i_mutex_dir_key#7 &obj_hash[i].lock irq_context: 0 sb_writers#11 &type->i_mutex_dir_key#7 stock_lock irq_context: 0 sb_writers#11 &type->i_mutex_dir_key#7 &c->lock irq_context: 0 sb_writers#11 &type->i_mutex_dir_key#7 &n->list_lock irq_context: 0 sb_writers#11 &type->i_mutex_dir_key#7 &____s->seqcount irq_context: 0 sb_writers#11 &type->i_mutex_dir_key#7 &root->kernfs_rwsem pool_lock#2 irq_context: 0 sb_writers#11 &type->i_mutex_dir_key#7 &root->kernfs_rwsem &xa->xa_lock#4 irq_context: 0 sb_writers#11 &type->i_mutex_dir_key#7 &root->kernfs_rwsem &xa->xa_lock#4 pool_lock#2 irq_context: 0 sb_writers#11 &type->i_mutex_dir_key#7 &root->kernfs_rwsem &obj_hash[i].lock irq_context: 0 sb_writers#11 &type->i_mutex_dir_key#7 &root->kernfs_rwsem stock_lock irq_context: 0 kn->active#56 fs_reclaim irq_context: 0 kn->active#56 fs_reclaim mmu_notifier_invalidate_range_start irq_context: 0 kn->active#56 stock_lock irq_context: 0 kn->active#56 pool_lock#2 irq_context: 0 kn->active#56 &kernfs_locks->open_file_mutex[count] irq_context: 0 kn->active#56 &kernfs_locks->open_file_mutex[count] fs_reclaim irq_context: 0 kn->active#56 &kernfs_locks->open_file_mutex[count] fs_reclaim mmu_notifier_invalidate_range_start irq_context: 0 kn->active#57 fs_reclaim irq_context: 0 kn->active#57 fs_reclaim mmu_notifier_invalidate_range_start irq_context: 0 (wq_completion)ipv6_addrconf (work_completion)(&(&ifa->dad_work)->work) rtnl_mutex &idev->mc_lock fill_pool_map-wait-type-override &____s->seqcount#2 irq_context: 0 kn->active#57 stock_lock irq_context: 0 kn->active#57 &kernfs_locks->open_file_mutex[count] irq_context: 0 kn->active#57 &kernfs_locks->open_file_mutex[count] fs_reclaim irq_context: 0 kn->active#57 &kernfs_locks->open_file_mutex[count] fs_reclaim mmu_notifier_invalidate_range_start irq_context: 0 sb_writers#11 &of->mutex kn->active#57 memcg_max_mutex irq_context: 0 (wq_completion)ipv6_addrconf (work_completion)(&(&ifa->dad_work)->work) rtnl_mutex rcu_read_lock rcu_read_lock rcu_read_lock_bh dev->qdisc_tx_busylock ?: &qdisc_tx_busylock#2 rcu_read_lock_bh &obj_hash[i].lock irq_context: 0 sb_writers#11 &type->i_mutex_dir_key#7/1 pool_lock#2 irq_context: 0 (wq_completion)ipv6_addrconf (work_completion)(&(&ifa->dad_work)->work) rtnl_mutex rcu_read_lock rcu_read_lock rcu_read_lock_bh dev->qdisc_tx_busylock ?: &qdisc_tx_busylock#2 rcu_read_lock_bh pool_lock#2 irq_context: 0 sb_writers#11 &type->i_mutex_dir_key#7/1 &xa->xa_lock#4 irq_context: 0 (wq_completion)ipv6_addrconf (work_completion)(&(&ifa->dad_work)->work) rtnl_mutex rcu_read_lock rcu_read_lock rcu_read_lock_bh dev->qdisc_tx_busylock ?: &qdisc_tx_busylock#2 rcu_read_lock_bh rcu_read_lock tk_core.seq.seqcount irq_context: 0 sb_writers#11 &type->i_mutex_dir_key#7/1 &xa->xa_lock#4 pool_lock#2 irq_context: 0 (wq_completion)ipv6_addrconf (work_completion)(&(&ifa->dad_work)->work) rtnl_mutex rcu_read_lock rcu_read_lock rcu_read_lock_bh dev->qdisc_tx_busylock ?: &qdisc_tx_busylock#2 rcu_read_lock_bh rcu_read_lock &list->lock#5 irq_context: 0 sb_writers#11 &type->i_mutex_dir_key#7/1 &obj_hash[i].lock irq_context: 0 sb_writers#11 &type->i_mutex_dir_key#7/1 stock_lock irq_context: softirq rcu_callback cgroup_threadgroup_rwsem.rss.gp_wait.lock irq_context: softirq rcu_callback cgroup_threadgroup_rwsem.rss.gp_wait.lock &obj_hash[i].lock irq_context: 0 sb_writers#11 &type->i_mutex_dir_key#7/1 cgroup_mutex blkcg_pol_mutex irq_context: 0 sb_writers#11 &type->i_mutex_dir_key#7/1 cgroup_mutex blkcg_pol_mutex fs_reclaim irq_context: 0 sb_writers#11 &type->i_mutex_dir_key#7/1 cgroup_mutex blkcg_pol_mutex fs_reclaim mmu_notifier_invalidate_range_start irq_context: 0 sb_writers#11 &type->i_mutex_dir_key#7/1 cgroup_mutex blkcg_pol_mutex pool_lock#2 irq_context: 0 sb_writers#11 &type->i_mutex_dir_key#7/1 cgroup_mutex blkcg_pol_mutex pcpu_alloc_mutex irq_context: 0 sb_writers#11 &type->i_mutex_dir_key#7/1 cgroup_mutex blkcg_pol_mutex pcpu_alloc_mutex pcpu_lock irq_context: 0 sb_writers#11 &type->i_mutex_dir_key#7/1 cgroup_mutex kfence_freelist_lock irq_context: 0 sb_writers#11 &type->i_mutex_dir_key#7/1 cgroup_mutex devcgroup_mutex irq_context: 0 sb_writers#11 &type->i_mutex_dir_key#7/1 cgroup_mutex cpu_hotplug_lock freezer_mutex irq_context: 0 sb_writers#11 &type->i_mutex_dir_key#7/1 cgroup_mutex rtnl_mutex irq_context: 0 sb_writers#11 &type->i_mutex_dir_key#7/1 cgroup_mutex rtnl_mutex rtnl_mutex.wait_lock irq_context: 0 sb_writers#11 &type->i_mutex_dir_key#7/1 cgroup_mutex rtnl_mutex &rq->__lock irq_context: 0 sb_writers#11 &type->i_mutex_dir_key#7/1 cgroup_mutex rtnl_mutex &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)events (linkwatch_work).work rtnl_mutex &____s->seqcount#2 irq_context: 0 (wq_completion)events (work_completion)(&(&nsim_dev->trap_data->trap_report_dw)->work) &devlink->lock_key &nsim_trap_data->trap_lock &____s->seqcount#2 irq_context: 0 (wq_completion)ipv6_addrconf (work_completion)(&(&ifa->dad_work)->work) rtnl_mutex &idev->mc_lock rcu_read_lock &____s->seqcount#2 irq_context: 0 (wq_completion)events (work_completion)(&(&nsim_dev->trap_data->trap_report_dw)->work) &devlink->lock_key#6 &nsim_trap_data->trap_lock &____s->seqcount#2 irq_context: 0 (wq_completion)ipv6_addrconf (work_completion)(&(&ifa->dad_work)->work) rtnl_mutex &idev->mc_lock remove_cache_srcu &rq->__lock irq_context: softirq rcu_callback stock_lock irq_context: 0 (wq_completion)events (work_completion)(&(&nsim_dev->trap_data->trap_report_dw)->work) &devlink->lock_key#5 &nsim_trap_data->trap_lock &____s->seqcount#2 irq_context: 0 rtnl_mutex (inetaddr_chain).rwsem &____s->seqcount#2 irq_context: 0 (wq_completion)ipv6_addrconf (work_completion)(&(&ifa->dad_work)->work) rtnl_mutex &rq->__lock &cfs_rq->removed.lock irq_context: 0 rtnl_mutex (inetaddr_chain).rwsem rcu_read_lock &____s->seqcount#2 irq_context: 0 rtnl_mutex dev_addr_sem &____s->seqcount#2 irq_context: 0 rtnl_mutex &ndev->lock &____s->seqcount#2 irq_context: 0 (wq_completion)mld (work_completion)(&(&idev->mc_ifc_work)->work) &idev->mc_lock rcu_read_lock rcu_read_lock rcu_read_lock_bh _xmit_ETHER#2 irq_context: 0 (wq_completion)mld (work_completion)(&(&idev->mc_ifc_work)->work) &idev->mc_lock rcu_read_lock rcu_read_lock rcu_read_lock_bh _xmit_ETHER#2 rcu_read_lock tk_core.seq.seqcount irq_context: 0 (wq_completion)mld (work_completion)(&(&idev->mc_ifc_work)->work) &idev->mc_lock rcu_read_lock rcu_read_lock rcu_read_lock_bh _xmit_ETHER#2 rcu_read_lock &fq->lock irq_context: 0 (wq_completion)mld (work_completion)(&(&idev->mc_ifc_work)->work) &idev->mc_lock rcu_read_lock rcu_read_lock rcu_read_lock_bh _xmit_ETHER#2 rcu_read_lock &local->active_txq_lock[i] irq_context: 0 (wq_completion)mld (work_completion)(&(&idev->mc_ifc_work)->work) &idev->mc_lock rcu_read_lock rcu_read_lock rcu_read_lock_bh _xmit_ETHER#2 rcu_read_lock &local->handle_wake_tx_queue_lock irq_context: 0 (wq_completion)mld (work_completion)(&(&idev->mc_ifc_work)->work) &idev->mc_lock rcu_read_lock rcu_read_lock rcu_read_lock_bh _xmit_ETHER#2 rcu_read_lock &local->handle_wake_tx_queue_lock &local->active_txq_lock[i] irq_context: 0 (wq_completion)mld (work_completion)(&(&idev->mc_ifc_work)->work) &idev->mc_lock rcu_read_lock rcu_read_lock rcu_read_lock_bh _xmit_ETHER#2 rcu_read_lock &local->handle_wake_tx_queue_lock &local->queue_stop_reason_lock irq_context: 0 (wq_completion)mld (work_completion)(&(&idev->mc_ifc_work)->work) &idev->mc_lock rcu_read_lock rcu_read_lock rcu_read_lock_bh _xmit_ETHER#2 rcu_read_lock &local->handle_wake_tx_queue_lock &fq->lock irq_context: 0 (wq_completion)mld (work_completion)(&(&idev->mc_ifc_work)->work) &idev->mc_lock rcu_read_lock rcu_read_lock rcu_read_lock_bh _xmit_ETHER#2 rcu_read_lock &local->handle_wake_tx_queue_lock &fq->lock tk_core.seq.seqcount irq_context: 0 (wq_completion)mld (work_completion)(&(&idev->mc_ifc_work)->work) &idev->mc_lock rcu_read_lock rcu_read_lock rcu_read_lock_bh _xmit_ETHER#2 rcu_read_lock &local->handle_wake_tx_queue_lock tk_core.seq.seqcount irq_context: 0 (wq_completion)mld (work_completion)(&(&idev->mc_ifc_work)->work) &idev->mc_lock rcu_read_lock rcu_read_lock rcu_read_lock_bh _xmit_ETHER#2 rcu_read_lock &local->handle_wake_tx_queue_lock hwsim_radio_lock irq_context: 0 (wq_completion)mld (work_completion)(&(&idev->mc_ifc_work)->work) &idev->mc_lock rcu_read_lock rcu_read_lock rcu_read_lock_bh _xmit_ETHER#2 rcu_read_lock &local->handle_wake_tx_queue_lock hwsim_radio_lock pool_lock#2 irq_context: 0 (wq_completion)mld (work_completion)(&(&idev->mc_ifc_work)->work) &idev->mc_lock rcu_read_lock rcu_read_lock rcu_read_lock_bh _xmit_ETHER#2 rcu_read_lock &local->handle_wake_tx_queue_lock hwsim_radio_lock &c->lock irq_context: 0 (wq_completion)mld (work_completion)(&(&idev->mc_ifc_work)->work) &idev->mc_lock rcu_read_lock rcu_read_lock rcu_read_lock_bh _xmit_ETHER#2 rcu_read_lock &local->handle_wake_tx_queue_lock hwsim_radio_lock &____s->seqcount#2 irq_context: 0 (wq_completion)mld (work_completion)(&(&idev->mc_ifc_work)->work) &idev->mc_lock rcu_read_lock rcu_read_lock rcu_read_lock_bh _xmit_ETHER#2 rcu_read_lock &local->handle_wake_tx_queue_lock hwsim_radio_lock &pcp->lock &zone->lock irq_context: 0 (wq_completion)mld (work_completion)(&(&idev->mc_ifc_work)->work) &idev->mc_lock rcu_read_lock rcu_read_lock rcu_read_lock_bh _xmit_ETHER#2 rcu_read_lock &local->handle_wake_tx_queue_lock hwsim_radio_lock &____s->seqcount irq_context: 0 (wq_completion)mld (work_completion)(&(&idev->mc_ifc_work)->work) &idev->mc_lock rcu_read_lock rcu_read_lock rcu_read_lock_bh _xmit_ETHER#2 rcu_read_lock &local->handle_wake_tx_queue_lock hwsim_radio_lock &list->lock#19 irq_context: 0 (wq_completion)mld (work_completion)(&(&idev->mc_ifc_work)->work) &idev->mc_lock rcu_read_lock rcu_read_lock rcu_read_lock_bh _xmit_ETHER#2 rcu_read_lock &local->handle_wake_tx_queue_lock &list->lock#19 irq_context: 0 (wq_completion)mld (work_completion)(&(&idev->mc_ifc_work)->work) &idev->mc_lock rcu_read_lock rcu_read_lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)mld (work_completion)(&(&idev->mc_ifc_work)->work) &idev->mc_lock rcu_read_lock &____s->seqcount#2 irq_context: 0 (wq_completion)events (work_completion)(&(&nsim_dev->trap_data->trap_report_dw)->work) &devlink->lock_key#4 &nsim_trap_data->trap_lock &____s->seqcount#2 irq_context: 0 sb_writers#11 &type->i_mutex_dir_key#7/1 cgroup_mutex rtnl_mutex.wait_lock irq_context: 0 sb_writers#11 &type->i_mutex_dir_key#7/1 cgroup_mutex &p->pi_lock irq_context: 0 sb_writers#11 &type->i_mutex_dir_key#7/1 cgroup_mutex &p->pi_lock &rq->__lock irq_context: 0 sb_writers#11 &type->i_mutex_dir_key#7/1 cgroup_mutex &p->pi_lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 &type->i_mutex_dir_key#7 stock_lock irq_context: 0 &type->i_mutex_dir_key#7 &c->lock irq_context: 0 &type->i_mutex_dir_key#7 &n->list_lock irq_context: 0 &type->i_mutex_dir_key#7 &____s->seqcount irq_context: 0 &type->i_mutex_dir_key#7 pool_lock#2 irq_context: 0 &type->i_mutex_dir_key#7 &root->kernfs_rwsem pool_lock#2 irq_context: 0 &type->i_mutex_dir_key#7 &root->kernfs_rwsem &xa->xa_lock#4 irq_context: 0 &type->i_mutex_dir_key#7 &root->kernfs_rwsem &xa->xa_lock#4 pool_lock#2 irq_context: 0 &type->i_mutex_dir_key#7 &root->kernfs_rwsem &obj_hash[i].lock irq_context: 0 &type->i_mutex_dir_key#7 &root->kernfs_rwsem stock_lock irq_context: 0 kn->active#55 stock_lock irq_context: 0 kn->active#55 &c->lock irq_context: 0 sb_writers#11 &of->mutex cgroup_mutex cpu_hotplug_lock cgroup_threadgroup_rwsem rcu_state.exp_mutex rcu_node_0 irq_context: 0 sb_writers#11 &of->mutex cgroup_mutex cpu_hotplug_lock cgroup_threadgroup_rwsem rcu_state.exp_mutex &obj_hash[i].lock irq_context: 0 sb_writers#11 &of->mutex cgroup_mutex cpu_hotplug_lock cgroup_threadgroup_rwsem rcu_state.exp_mutex rcu_read_lock &pool->lock irq_context: 0 sb_writers#11 &of->mutex cgroup_mutex cpu_hotplug_lock cgroup_threadgroup_rwsem rcu_state.exp_mutex rcu_read_lock &pool->lock &obj_hash[i].lock irq_context: 0 sb_writers#11 &of->mutex cgroup_mutex cpu_hotplug_lock cgroup_threadgroup_rwsem rcu_state.exp_mutex rcu_read_lock &pool->lock &p->pi_lock irq_context: 0 sb_writers#11 &of->mutex cgroup_mutex cpu_hotplug_lock cgroup_threadgroup_rwsem rcu_state.exp_mutex rcu_read_lock &pool->lock &p->pi_lock &rq->__lock irq_context: 0 sb_writers#11 &of->mutex cgroup_mutex cpu_hotplug_lock cgroup_threadgroup_rwsem rcu_state.exp_mutex rcu_read_lock &pool->lock &p->pi_lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 sb_writers#11 &of->mutex cgroup_mutex cpu_hotplug_lock cgroup_threadgroup_rwsem rcu_state.exp_mutex &rnp->exp_wq[1] irq_context: 0 sb_writers#11 &of->mutex cgroup_mutex cpu_hotplug_lock cgroup_threadgroup_rwsem rcu_state.exp_mutex &rq->__lock irq_context: 0 sb_writers#11 &of->mutex cgroup_mutex cpu_hotplug_lock cgroup_threadgroup_rwsem rcu_state.exp_mutex &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 sb_writers#11 &of->mutex cgroup_mutex cpu_hotplug_lock cgroup_threadgroup_rwsem &obj_hash[i].lock irq_context: 0 sb_writers#11 &of->mutex cgroup_mutex cpu_hotplug_lock cgroup_threadgroup_rwsem freezer_mutex irq_context: 0 sb_writers#11 &of->mutex cgroup_mutex cpu_hotplug_lock cgroup_threadgroup_rwsem freezer_mutex freezer_lock irq_context: 0 sb_writers#11 &of->mutex cgroup_mutex cpu_hotplug_lock cgroup_threadgroup_rwsem freezer_mutex freezer_lock rcu_read_lock &sighand->siglock irq_context: 0 sb_writers#11 &of->mutex cgroup_mutex cpu_hotplug_lock cgroup_threadgroup_rwsem freezer_mutex freezer_lock &sighand->siglock irq_context: 0 sb_writers#11 &of->mutex cgroup_mutex cpu_hotplug_lock cgroup_threadgroup_rwsem freezer_mutex freezer_lock &sighand->siglock &p->pi_lock irq_context: 0 sb_writers#11 &of->mutex cgroup_mutex cpu_hotplug_lock cgroup_threadgroup_rwsem freezer_mutex freezer_lock &sighand->siglock &p->pi_lock &rq->__lock irq_context: 0 sb_writers#11 &of->mutex cgroup_mutex cpu_hotplug_lock cgroup_threadgroup_rwsem &p->alloc_lock &newf->file_lock irq_context: 0 sb_writers#11 &of->mutex cgroup_mutex cpu_hotplug_lock cgroup_threadgroup_rwsem.rss.gp_wait.lock &obj_hash[i].lock irq_context: 0 &xa->xa_lock#4 pool_lock#2 irq_context: 0 (wq_completion)events (work_completion)(&ht->run_work) &ht->mutex &____s->seqcount#2 irq_context: 0 (wq_completion)events (work_completion)(&ht->run_work) &ht->mutex &____s->seqcount irq_context: 0 nf_hook_mutex irq_context: 0 nf_hook_mutex fs_reclaim irq_context: 0 nf_hook_mutex fs_reclaim mmu_notifier_invalidate_range_start irq_context: 0 nf_hook_mutex stock_lock irq_context: 0 nf_hook_mutex pool_lock#2 irq_context: 0 nf_hook_mutex &c->lock irq_context: 0 nf_hook_mutex &____s->seqcount#2 irq_context: 0 nf_hook_mutex &____s->seqcount irq_context: 0 ebt_mutex &mm->mmap_lock irq_context: 0 &sb->s_type->i_mutex_key#10 sk_lock-AF_INET slock-AF_INET &obj_hash[i].lock irq_context: 0 &sb->s_type->i_mutex_key#10 sk_lock-AF_INET slock-AF_INET elock-AF_INET irq_context: 0 &sb->s_type->i_mutex_key#10 stock_lock irq_context: 0 &mm->mmap_lock stock_lock irq_context: 0 &xt[i].mutex &c->lock irq_context: 0 &xt[i].mutex &n->list_lock irq_context: 0 &xt[i].mutex &n->list_lock &c->lock irq_context: 0 &xt[i].mutex &mm->mmap_lock irq_context: 0 &xt[i].mutex free_vmap_area_lock irq_context: 0 &xt[i].mutex free_vmap_area_lock &obj_hash[i].lock irq_context: 0 &xt[i].mutex free_vmap_area_lock pool_lock#2 irq_context: 0 &xt[i].mutex vmap_area_lock irq_context: 0 &xt[i].mutex &____s->seqcount irq_context: 0 &xt[i].mutex &per_cpu(xt_recseq, i) irq_context: 0 &xt[i].mutex &obj_hash[i].lock irq_context: 0 &xt[i].mutex purge_vmap_area_lock irq_context: 0 &xt[i].mutex rcu_read_lock pool_lock#2 irq_context: 0 nf_nat_proto_mutex irq_context: 0 nf_nat_proto_mutex fs_reclaim irq_context: 0 nf_nat_proto_mutex fs_reclaim mmu_notifier_invalidate_range_start irq_context: 0 nf_nat_proto_mutex pool_lock#2 irq_context: 0 nf_nat_proto_mutex nf_hook_mutex irq_context: 0 nf_nat_proto_mutex nf_hook_mutex fs_reclaim irq_context: 0 nf_nat_proto_mutex nf_hook_mutex fs_reclaim mmu_notifier_invalidate_range_start irq_context: 0 nf_nat_proto_mutex nf_hook_mutex stock_lock irq_context: 0 nf_nat_proto_mutex nf_hook_mutex pool_lock#2 irq_context: 0 nf_nat_proto_mutex cpu_hotplug_lock irq_context: 0 nf_nat_proto_mutex &obj_hash[i].lock irq_context: 0 nf_nat_proto_mutex stock_lock irq_context: 0 &sb->s_type->i_mutex_key#10 rcu_read_lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)mld (work_completion)(&(&idev->mc_ifc_work)->work) &idev->mc_lock &____s->seqcount#2 irq_context: 0 (wq_completion)mld (work_completion)(&(&idev->mc_ifc_work)->work) &idev->mc_lock rcu_read_lock rcu_read_lock fill_pool_map-wait-type-override &c->lock irq_context: 0 (wq_completion)mld (work_completion)(&(&idev->mc_ifc_work)->work) &idev->mc_lock rcu_read_lock rcu_read_lock fill_pool_map-wait-type-override &____s->seqcount#2 irq_context: 0 (wq_completion)mld (work_completion)(&(&idev->mc_ifc_work)->work) &idev->mc_lock rcu_read_lock rcu_read_lock fill_pool_map-wait-type-override &____s->seqcount irq_context: 0 &xt[i].mutex &rq->__lock irq_context: 0 &xt[i].mutex &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 nf_nat_proto_mutex nf_hook_mutex &c->lock irq_context: 0 nf_nat_proto_mutex nf_hook_mutex &____s->seqcount#2 irq_context: 0 nf_nat_proto_mutex nf_hook_mutex &____s->seqcount irq_context: 0 nf_hook_mutex &pcp->lock &zone->lock irq_context: 0 &sb->s_type->i_mutex_key#10 sk_lock-AF_INET6 clock-AF_INET6 irq_context: 0 &sb->s_type->i_mutex_key#10 sk_lock-AF_INET6 slock-AF_INET6 &obj_hash[i].lock irq_context: 0 &sb->s_type->i_mutex_key#10 sk_lock-AF_INET6 slock-AF_INET6 elock-AF_INET6 irq_context: 0 &pipe->mutex/1 stock_lock irq_context: 0 sb_writers#4 &type->i_mutex_dir_key#3/1 &xa->xa_lock#4 irq_context: 0 sb_writers#4 &type->i_mutex_dir_key#3/1 stock_lock irq_context: 0 sb_writers#4 &type->i_mutex_dir_key#3/1 tomoyo_ss &____s->seqcount#2 irq_context: 0 sb_writers#4 &type->i_mutex_dir_key#3/1 &xa->xa_lock#4 pool_lock#2 irq_context: 0 sb_writers#4 &type->i_mutex_dir_key#3/1 &journal->j_state_lock irq_context: 0 sb_writers#4 &type->i_mutex_dir_key#3/1 &journal->j_state_lock tk_core.seq.seqcount irq_context: 0 sb_writers#4 &type->i_mutex_dir_key#3/1 &journal->j_state_lock &obj_hash[i].lock irq_context: 0 sb_writers#4 &type->i_mutex_dir_key#3/1 &journal->j_state_lock &base->lock irq_context: 0 sb_writers#4 &type->i_mutex_dir_key#3/1 &journal->j_state_lock &base->lock &obj_hash[i].lock irq_context: 0 sb_writers#4 &type->i_mutex_dir_key#3/1 jbd2_handle stock_lock irq_context: 0 loop_validate_mutex irq_context: 0 loop_validate_mutex &lo->lo_mutex irq_context: 0 &fsnotify_mark_srcu fs_reclaim irq_context: 0 &fsnotify_mark_srcu fs_reclaim mmu_notifier_invalidate_range_start irq_context: 0 &fsnotify_mark_srcu &rq->__lock irq_context: 0 &fsnotify_mark_srcu &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 &fsnotify_mark_srcu pool_lock#2 irq_context: 0 &fsnotify_mark_srcu &group->notification_lock irq_context: 0 &fsnotify_mark_srcu &group->notification_waitq irq_context: 0 &fsnotify_mark_srcu &group->notification_waitq &ep->lock irq_context: 0 &fsnotify_mark_srcu &group->notification_waitq &ep->lock &ep->wq irq_context: 0 &fsnotify_mark_srcu &group->notification_waitq &ep->lock &ep->wq &p->pi_lock irq_context: 0 &fsnotify_mark_srcu &group->notification_waitq &ep->lock &ep->wq &p->pi_lock &rq->__lock irq_context: 0 &fsnotify_mark_srcu &group->notification_waitq &ep->lock &ep->wq &p->pi_lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)events (work_completion)(&(&nsim_dev->trap_data->trap_report_dw)->work) &devlink->lock_key#2 &nsim_trap_data->trap_lock &____s->seqcount#2 irq_context: 0 (wq_completion)events (work_completion)(&(&nsim_dev->trap_data->trap_report_dw)->work) &devlink->lock_key#3 &nsim_trap_data->trap_lock &____s->seqcount#2 irq_context: 0 &type->i_mutex_dir_key#4 &root->kernfs_rwsem &____s->seqcount#2 irq_context: 0 &f->f_pos_lock &type->i_mutex_dir_key#4 &mm->mmap_lock mmu_notifier_invalidate_range_start irq_context: 0 &type->i_mutex_dir_key#4 &____s->seqcount#2 irq_context: 0 (wq_completion)events (work_completion)(&rdev->wiphy_work) &rdev->wiphy.mtx &wdev->mtx &rdev->bss_lock krc.lock irq_context: 0 (wq_completion)events (work_completion)(&rdev->wiphy_work) &rdev->wiphy.mtx &wdev->mtx &rdev->bss_lock &c->lock irq_context: 0 (wq_completion)events (work_completion)(&rdev->wiphy_work) &rdev->wiphy.mtx &wdev->mtx &rdev->bss_lock &____s->seqcount#2 irq_context: 0 (wq_completion)events (work_completion)(&rdev->wiphy_work) &rdev->wiphy.mtx &wdev->mtx &rdev->bss_lock &____s->seqcount irq_context: 0 &xt[i].mutex &____s->seqcount#2 irq_context: 0 &xt[i].mutex purge_vmap_area_lock &obj_hash[i].lock irq_context: 0 &xt[i].mutex purge_vmap_area_lock pool_lock#2 irq_context: 0 purge_vmap_area_lock &obj_hash[i].lock irq_context: 0 purge_vmap_area_lock pool_lock#2 irq_context: 0 &pipe->mutex/1 &mm->mmap_lock stock_lock irq_context: 0 (wq_completion)events (linkwatch_work).work rtnl_mutex rcu_read_lock &____s->seqcount#2 irq_context: 0 dup_mmap_sem &mm->mmap_lock &mm->mmap_lock/1 stock_lock irq_context: 0 dup_mmap_sem &mm->mmap_lock &mm->mmap_lock/1 &anon_vma->rwsem stock_lock irq_context: 0 (wq_completion)events (linkwatch_work).work rtnl_mutex &tb->tb6_lock &____s->seqcount#2 irq_context: 0 (wq_completion)events (linkwatch_work).work rtnl_mutex &tb->tb6_lock &pcp->lock &zone->lock irq_context: 0 sb_writers#8 tomoyo_ss &____s->seqcount#2 irq_context: 0 &group->notification_waitq &ep->lock irq_context: 0 dup_mmap_sem &mm->mmap_lock &mm->mmap_lock/1 &____s->seqcount#2 irq_context: 0 cgroup_threadgroup_rwsem freezer_mutex irq_context: 0 &vma->vm_lock->lock stock_lock irq_context: softirq fs/notify/mark.c:89 rcu_read_lock &pool->lock/1 irq_context: softirq fs/notify/mark.c:89 rcu_read_lock &pool->lock/1 &obj_hash[i].lock irq_context: softirq fs/notify/mark.c:89 rcu_read_lock &pool->lock/1 &p->pi_lock irq_context: 0 (wq_completion)mld (work_completion)(&(&idev->mc_dad_work)->work) &idev->mc_lock rcu_read_lock rcu_read_lock &____s->seqcount#9 irq_context: 0 (wq_completion)mld (work_completion)(&(&idev->mc_dad_work)->work) &idev->mc_lock rcu_read_lock rcu_read_lock rcu_read_lock_bh &hsr->seqnr_lock irq_context: 0 (wq_completion)mld (work_completion)(&(&idev->mc_dad_work)->work) &idev->mc_lock rcu_read_lock rcu_read_lock rcu_read_lock_bh &hsr->seqnr_lock rcu_read_lock &new_node->seq_out_lock irq_context: 0 (wq_completion)mld (work_completion)(&(&idev->mc_dad_work)->work) &idev->mc_lock rcu_read_lock rcu_read_lock rcu_read_lock_bh &hsr->seqnr_lock rcu_read_lock pool_lock#2 irq_context: 0 (wq_completion)mld (work_completion)(&(&idev->mc_dad_work)->work) &idev->mc_lock rcu_read_lock rcu_read_lock rcu_read_lock_bh &hsr->seqnr_lock rcu_read_lock rcu_read_lock_bh rcu_read_lock tk_core.seq.seqcount irq_context: 0 (wq_completion)mld (work_completion)(&(&idev->mc_dad_work)->work) &idev->mc_lock rcu_read_lock rcu_read_lock rcu_read_lock_bh &hsr->seqnr_lock rcu_read_lock rcu_read_lock_bh rcu_read_lock &list->lock#5 irq_context: 0 (wq_completion)mld (work_completion)(&(&idev->mc_dad_work)->work) &idev->mc_lock rcu_read_lock rcu_read_lock rcu_read_lock_bh &hsr->seqnr_lock &obj_hash[i].lock irq_context: 0 (wq_completion)mld (work_completion)(&(&idev->mc_dad_work)->work) &idev->mc_lock rcu_read_lock rcu_read_lock rcu_read_lock_bh &hsr->seqnr_lock pool_lock#2 irq_context: 0 (wq_completion)mld (work_completion)(&(&idev->mc_dad_work)->work) &idev->mc_lock rcu_read_lock rcu_read_lock rcu_read_lock_bh &hsr->seqnr_lock &____s->seqcount irq_context: 0 rtnl_mutex (inetaddr_chain).rwsem &net->sctp.local_addr_lock &net->sctp.addr_wq_lock &____s->seqcount#2 irq_context: 0 (wq_completion)events (linkwatch_work).work rtnl_mutex &tb->tb6_lock quarantine_lock irq_context: 0 sb_writers#4 &type->i_mutex_dir_key#3/1 jbd2_handle batched_entropy_u32.lock crngs.lock irq_context: 0 sb_writers#4 &type->i_mutex_dir_key#3/1 &____s->seqcount#2 irq_context: 0 sb_writers#4 &type->i_mutex_dir_key#3/1 jbd2_handle rcu_read_lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 sb_writers#5 &type->i_mutex_dir_key#5/1 tomoyo_ss &____s->seqcount#2 irq_context: 0 sb_writers#5 &type->i_mutex_dir_key#5 &____s->seqcount#2 irq_context: 0 &sb->s_type->i_mutex_key#9 &xa->xa_lock#4 irq_context: 0 &sb->s_type->i_mutex_key#9 &xa->xa_lock#4 pool_lock#2 irq_context: 0 &sb->s_type->i_mutex_key#9 stock_lock irq_context: 0 sb_writers#3 &sb->s_type->i_mutex_key#9 stock_lock irq_context: 0 sb_writers#3 &sb->s_type->i_mutex_key#9 &____s->seqcount#2 irq_context: 0 &r->consumer_lock irq_context: 0 &mm->mmap_lock stock_lock irq_context: hardirq hrtimer_bases.lock fill_pool_map-wait-type-override &____s->seqcount#2 irq_context: 0 (wq_completion)events (work_completion)(&(&nsim_dev->trap_data->trap_report_dw)->work) &devlink->lock_key#6 &nsim_trap_data->trap_lock batched_entropy_u8.lock irq_context: 0 (wq_completion)events (work_completion)(&(&nsim_dev->trap_data->trap_report_dw)->work) &devlink->lock_key#6 &nsim_trap_data->trap_lock kfence_freelist_lock irq_context: 0 (wq_completion)events (work_completion)(&(&nsim_dev->trap_data->trap_report_dw)->work) &devlink->lock_key#6 &nsim_trap_data->trap_lock &meta->lock irq_context: 0 (wq_completion)events (work_completion)(&(&nsim_dev->trap_data->trap_report_dw)->work) &devlink->lock_key#6 &rq->__lock irq_context: 0 rtnl_mutex (inetaddr_chain).rwsem rcu_read_lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 &type->i_mutex_dir_key#6 pool_lock#2 irq_context: 0 &type->i_mutex_dir_key#6 &xa->xa_lock#4 irq_context: 0 &type->i_mutex_dir_key#6 &xa->xa_lock#4 pool_lock#2 irq_context: 0 &type->i_mutex_dir_key#6 &obj_hash[i].lock irq_context: 0 &type->i_mutex_dir_key#6 stock_lock irq_context: 0 &type->i_mutex_dir_key#6 &dentry->d_lock &wq irq_context: 0 &mm->mmap_lock &xa->xa_lock#9 stock_lock irq_context: 0 &mm->mmap_lock &xa->xa_lock#9 &____s->seqcount#2 irq_context: 0 &mm->mmap_lock &xa->xa_lock#9 &____s->seqcount irq_context: 0 fill_pool_map-wait-type-override &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 rtnl_mutex _xmit_ETHER &____s->seqcount#2 irq_context: 0 cb_lock genl_mutex &____s->seqcount#2 irq_context: 0 &sb->s_type->i_mutex_key#10 (netlink_chain).rwsem &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 &sb->s_type->i_mutex_key#10 &pnsocks.lock irq_context: 0 &sb->s_type->i_mutex_key#10 resource_mutex irq_context: 0 &sb->s_type->i_mutex_key#10 clock-AF_PHONET irq_context: 0 &sb->s_type->i_mutex_key#10 rlock-AF_PHONET irq_context: 0 &sighand->siglock &p->pi_lock irq_context: 0 &sighand->siglock &p->pi_lock &rq->__lock irq_context: 0 &sighand->siglock &p->pi_lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 rtnl_mutex rcu_read_lock &cfs_rq->removed.lock irq_context: softirq &(&bat_priv->nc.work)->timer rcu_read_lock &pool->lock/1 fill_pool_map-wait-type-override &c->lock irq_context: softirq &(&bat_priv->nc.work)->timer rcu_read_lock &pool->lock/1 fill_pool_map-wait-type-override &____s->seqcount#2 irq_context: softirq &(&bat_priv->nc.work)->timer rcu_read_lock &pool->lock/1 fill_pool_map-wait-type-override &____s->seqcount irq_context: softirq &(&bat_priv->nc.work)->timer rcu_read_lock &pool->lock/1 fill_pool_map-wait-type-override pool_lock irq_context: 0 (wq_completion)mld (work_completion)(&(&idev->mc_ifc_work)->work) &idev->mc_lock rcu_read_lock rcu_read_lock &____s->seqcount#2 irq_context: 0 &mm->mmap_lock &anon_vma->rwsem stock_lock irq_context: 0 (wq_completion)events (linkwatch_work).work rtnl_mutex lweventlist_lock &c->lock irq_context: 0 (wq_completion)events (linkwatch_work).work rtnl_mutex lweventlist_lock &____s->seqcount#2 irq_context: 0 (wq_completion)events (linkwatch_work).work rtnl_mutex lweventlist_lock &____s->seqcount irq_context: 0 (wq_completion)ipv6_addrconf (work_completion)(&(&ifa->dad_work)->work) rtnl_mutex rcu_read_lock rcu_read_lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 pcpu_lock stock_lock irq_context: 0 (wq_completion)ipv6_addrconf (work_completion)(&(&ifa->dad_work)->work) rtnl_mutex rcu_read_lock rcu_read_lock &cfs_rq->removed.lock irq_context: 0 tasklist_lock stock_lock irq_context: 0 (wq_completion)ipv6_addrconf (work_completion)(&(&ifa->dad_work)->work) rtnl_mutex &tb->tb6_lock rcu_read_lock &____s->seqcount#2 irq_context: 0 (wq_completion)ipv6_addrconf (work_completion)(&(&ifa->dad_work)->work) rtnl_mutex rcu_read_lock rcu_read_lock rcu_read_lock_bh rcu_read_lock rcu_read_lock &dir->lock#2 irq_context: 0 (wq_completion)ipv6_addrconf (work_completion)(&(&ifa->dad_work)->work) rtnl_mutex rcu_read_lock rcu_read_lock rcu_read_lock_bh rcu_read_lock rcu_read_lock &ul->lock irq_context: 0 (wq_completion)ipv6_addrconf (work_completion)(&(&ifa->dad_work)->work) rtnl_mutex rcu_read_lock rcu_read_lock rcu_read_lock_bh rcu_read_lock batched_entropy_u32.lock irq_context: 0 (wq_completion)ipv6_addrconf (work_completion)(&(&ifa->dad_work)->work) rtnl_mutex rcu_read_lock rcu_read_lock rcu_read_lock_bh rcu_read_lock rcu_read_lock &____s->seqcount#7 irq_context: 0 (wq_completion)ipv6_addrconf (work_completion)(&(&ifa->dad_work)->work) rtnl_mutex rcu_read_lock rcu_read_lock rcu_read_lock_bh rcu_read_lock rcu_read_lock rcu_read_lock &____s->seqcount#7 irq_context: 0 (wq_completion)ipv6_addrconf (work_completion)(&(&ifa->dad_work)->work) rtnl_mutex rcu_read_lock rcu_read_lock rcu_read_lock_bh rcu_read_lock rcu_read_lock once_lock irq_context: 0 (wq_completion)ipv6_addrconf (work_completion)(&(&ifa->dad_work)->work) rtnl_mutex rcu_read_lock rcu_read_lock rcu_read_lock_bh rcu_read_lock rcu_read_lock once_lock crngs.lock irq_context: 0 (wq_completion)ipv6_addrconf (work_completion)(&(&ifa->dad_work)->work) rtnl_mutex rcu_read_lock rcu_read_lock rcu_read_lock_bh rcu_read_lock rcu_read_lock &obj_hash[i].lock irq_context: 0 (wq_completion)ipv6_addrconf (work_completion)(&(&ifa->dad_work)->work) rtnl_mutex rcu_read_lock rcu_read_lock rcu_read_lock_bh rcu_read_lock rcu_read_lock rcu_read_lock &pool->lock irq_context: 0 (wq_completion)ipv6_addrconf (work_completion)(&(&ifa->dad_work)->work) rtnl_mutex rcu_read_lock rcu_read_lock rcu_read_lock_bh rcu_read_lock rcu_read_lock rcu_read_lock &pool->lock &obj_hash[i].lock irq_context: 0 (wq_completion)ipv6_addrconf (work_completion)(&(&ifa->dad_work)->work) rtnl_mutex rcu_read_lock rcu_read_lock rcu_read_lock_bh rcu_read_lock rcu_read_lock rcu_read_lock &pool->lock pool_lock#2 irq_context: 0 (wq_completion)ipv6_addrconf (work_completion)(&(&ifa->dad_work)->work) rtnl_mutex rcu_read_lock rcu_read_lock rcu_read_lock_bh rcu_read_lock rcu_read_lock &nf_nat_locks[i] irq_context: 0 (wq_completion)ipv6_addrconf (work_completion)(&(&ifa->dad_work)->work) rtnl_mutex rcu_read_lock rcu_read_lock rcu_read_lock_bh rcu_read_lock rcu_read_lock &nf_conntrack_locks[i] irq_context: 0 (wq_completion)ipv6_addrconf (work_completion)(&(&ifa->dad_work)->work) rtnl_mutex rcu_read_lock rcu_read_lock rcu_read_lock_bh rcu_read_lock rcu_read_lock &nf_conntrack_locks[i] &nf_conntrack_locks[i]/1 irq_context: 0 (wq_completion)ipv6_addrconf (work_completion)(&(&ifa->dad_work)->work) rtnl_mutex rcu_read_lock rcu_read_lock rcu_read_lock_bh rcu_read_lock rcu_read_lock &nf_conntrack_locks[i] &nf_conntrack_locks[i]/1 batched_entropy_u8.lock irq_context: 0 (wq_completion)ipv6_addrconf (work_completion)(&(&ifa->dad_work)->work) rtnl_mutex rcu_read_lock rcu_read_lock rcu_read_lock_bh rcu_read_lock rcu_read_lock &nf_conntrack_locks[i]/1 irq_context: 0 (wq_completion)ipv6_addrconf (work_completion)(&(&ifa->dad_work)->work) rtnl_mutex rcu_read_lock rcu_read_lock rcu_read_lock_bh rcu_read_lock rcu_read_lock &____s->seqcount#10 irq_context: 0 (wq_completion)ipv6_addrconf (work_completion)(&(&ifa->dad_work)->work) rtnl_mutex rcu_read_lock rcu_read_lock rcu_read_lock_bh rcu_read_lock rcu_read_lock rcu_read_lock_bh tk_core.seq.seqcount irq_context: 0 (wq_completion)ipv6_addrconf (work_completion)(&(&ifa->dad_work)->work) rtnl_mutex rcu_read_lock rcu_read_lock rcu_read_lock_bh rcu_read_lock rcu_read_lock rcu_read_lock_bh &list->lock#5 irq_context: 0 (wq_completion)ipv6_addrconf (work_completion)(&(&ifa->dad_work)->work) rtnl_mutex &idev->mc_lock &vlan_netdev_addr_lock_key/1 _xmit_ETHER &c->lock irq_context: 0 (wq_completion)ipv6_addrconf (work_completion)(&(&ifa->dad_work)->work) rtnl_mutex &idev->mc_lock &vlan_netdev_addr_lock_key/1 _xmit_ETHER &____s->seqcount#2 irq_context: 0 (wq_completion)ipv6_addrconf (work_completion)(&(&ifa->dad_work)->work) rtnl_mutex &idev->mc_lock &vlan_netdev_addr_lock_key/1 _xmit_ETHER &____s->seqcount irq_context: 0 (wq_completion)ipv6_addrconf (work_completion)(&(&ifa->dad_work)->work) rtnl_mutex rcu_read_lock rcu_read_lock &____s->seqcount#2 irq_context: 0 (wq_completion)ipv6_addrconf (work_completion)(&(&ifa->dad_work)->work) rtnl_mutex &idev->mc_lock _xmit_ETHER &____s->seqcount#2 irq_context: 0 (wq_completion)ipv6_addrconf (work_completion)(&(&ifa->dad_work)->work) rtnl_mutex &idev->mc_lock rcu_read_lock &rq->__lock irq_context: softirq rcu_callback put_task_map-wait-type-override stock_lock irq_context: 0 remove_cache_srcu rcu_node_0 irq_context: 0 &f->f_pos_lock &type->i_mutex_dir_key#3 &____s->seqcount#2 irq_context: 0 sb_writers#4 &type->i_mutex_dir_key#3/1 &sb->s_type->i_mutex_key#8 &ei->i_es_lock irq_context: 0 sb_writers#4 &type->i_mutex_dir_key#3/1 &sb->s_type->i_mutex_key#8 jbd2_handle &sbi->s_orphan_lock irq_context: 0 sb_writers#4 &type->i_mutex_dir_key#3/1 &sb->s_type->i_mutex_key#8 &dentry->d_lock irq_context: 0 sb_writers#4 &type->i_mutex_dir_key#3/1 &fsnotify_mark_srcu irq_context: 0 sb_writers#4 &s->s_inode_list_lock irq_context: 0 sb_writers#4 sb_internal irq_context: 0 sb_writers#4 sb_internal mmu_notifier_invalidate_range_start irq_context: 0 sb_writers#4 sb_internal pool_lock#2 irq_context: 0 sb_writers#4 sb_internal &journal->j_state_lock irq_context: 0 sb_writers#4 sb_internal jbd2_handle irq_context: 0 sb_writers#4 sb_internal jbd2_handle &ei->i_raw_lock irq_context: 0 sb_writers#4 sb_internal jbd2_handle &sbi->s_orphan_lock irq_context: 0 sb_writers#4 sb_internal jbd2_handle &xa->xa_lock#9 irq_context: 0 sb_writers#4 sb_internal jbd2_handle &ei->i_es_lock irq_context: 0 sb_writers#4 sb_internal jbd2_handle &journal->j_wait_updates irq_context: 0 sb_writers#4 sb_internal &obj_hash[i].lock irq_context: 0 sb_writers#4 inode_hash_lock irq_context: 0 sb_writers#4 inode_hash_lock &sb->s_type->i_lock_key#22 irq_context: 0 sb_writers#4 &fsnotify_mark_srcu irq_context: 0 sb_writers#4 &type->i_mutex_dir_key#3/1 tomoyo_ss rcu_read_lock pool_lock#2 irq_context: 0 sb_writers#4 &type->i_mutex_dir_key#3/1 &type->i_mutex_dir_key#3 irq_context: 0 sb_writers#4 &type->i_mutex_dir_key#3/1 &type->i_mutex_dir_key#3 &ei->i_es_lock irq_context: 0 sb_writers#4 &type->i_mutex_dir_key#3/1 &type->i_mutex_dir_key#3 mmu_notifier_invalidate_range_start irq_context: 0 sb_writers#4 &type->i_mutex_dir_key#3/1 &type->i_mutex_dir_key#3 pool_lock#2 irq_context: 0 sb_writers#4 &type->i_mutex_dir_key#3/1 &type->i_mutex_dir_key#3 &journal->j_state_lock irq_context: 0 sb_writers#4 &type->i_mutex_dir_key#3/1 &type->i_mutex_dir_key#3 jbd2_handle irq_context: 0 sb_writers#4 &type->i_mutex_dir_key#3/1 &type->i_mutex_dir_key#3 jbd2_handle &sbi->s_orphan_lock irq_context: 0 sb_writers#4 &type->i_mutex_dir_key#3/1 &type->i_mutex_dir_key#3 jbd2_handle &ei->i_raw_lock irq_context: 0 sb_writers#4 &type->i_mutex_dir_key#3/1 &type->i_mutex_dir_key#3 jbd2_handle tk_core.seq.seqcount irq_context: 0 sb_writers#4 &type->i_mutex_dir_key#3/1 &type->i_mutex_dir_key#3 jbd2_handle &journal->j_wait_updates irq_context: 0 sb_writers#4 &type->i_mutex_dir_key#3/1 &type->i_mutex_dir_key#3 &obj_hash[i].lock irq_context: 0 sb_writers#4 &type->i_mutex_dir_key#3/1 &type->i_mutex_dir_key#3 rename_lock.seqcount irq_context: 0 sb_writers#4 &type->i_mutex_dir_key#3/1 &type->i_mutex_dir_key#3 &dentry->d_lock irq_context: 0 sb_writers#4 &type->i_mutex_dir_key#3/1 &type->i_mutex_dir_key#3 &dentry->d_lock &dentry->d_lock/1 irq_context: 0 sb_writers#4 &type->i_mutex_dir_key#3/1 &type->i_mutex_dir_key#3 &dentry->d_lock &dentry->d_lock/1 &lru->node[i].lock irq_context: 0 sb_writers#4 &type->i_mutex_dir_key#3/1 &type->i_mutex_dir_key#3 &dentry->d_lock &dentry->d_lock &lru->node[i].lock irq_context: 0 sb_writers#4 &type->i_mutex_dir_key#3/1 &wb->list_lock irq_context: 0 sb_writers#4 &type->i_mutex_dir_key#3/1 &wb->list_lock &sb->s_type->i_lock_key#22 irq_context: 0 sb_writers#4 &type->i_mutex_dir_key#3/1 sb_internal irq_context: 0 sb_writers#4 &type->i_mutex_dir_key#3/1 sb_internal mmu_notifier_invalidate_range_start irq_context: 0 sb_writers#4 &type->i_mutex_dir_key#3/1 sb_internal pool_lock#2 irq_context: 0 sb_writers#4 &type->i_mutex_dir_key#3/1 sb_internal &journal->j_state_lock irq_context: 0 sb_writers#4 &type->i_mutex_dir_key#3/1 sb_internal jbd2_handle irq_context: 0 sb_writers#4 &type->i_mutex_dir_key#3/1 sb_internal jbd2_handle &ei->i_raw_lock irq_context: 0 sb_writers#4 &type->i_mutex_dir_key#3/1 sb_internal jbd2_handle &ei->i_data_sem irq_context: 0 sb_writers#4 &type->i_mutex_dir_key#3/1 sb_internal jbd2_handle &ei->i_data_sem &ei->i_raw_lock irq_context: 0 sb_writers#4 &type->i_mutex_dir_key#3/1 sb_internal jbd2_handle &ei->i_data_sem &ei->i_es_lock irq_context: 0 sb_writers#4 &type->i_mutex_dir_key#3/1 sb_internal jbd2_handle &ei->i_data_sem &ei->i_es_lock &obj_hash[i].lock irq_context: 0 sb_writers#4 &type->i_mutex_dir_key#3/1 sb_internal jbd2_handle &ei->i_data_sem &ei->i_es_lock pool_lock#2 irq_context: 0 sb_writers#4 &type->i_mutex_dir_key#3/1 sb_internal jbd2_handle &ei->i_data_sem &ei->i_es_lock &sbi->s_es_lock irq_context: 0 sb_writers#4 &type->i_mutex_dir_key#3/1 sb_internal jbd2_handle &ei->i_data_sem mmu_notifier_invalidate_range_start irq_context: 0 sb_writers#4 &type->i_mutex_dir_key#3/1 sb_internal jbd2_handle &ei->i_data_sem pool_lock#2 irq_context: 0 sb_writers#4 &type->i_mutex_dir_key#3/1 sb_internal jbd2_handle &ei->i_data_sem &ret->b_state_lock irq_context: 0 sb_writers#4 &type->i_mutex_dir_key#3/1 sb_internal jbd2_handle &ei->i_data_sem &ret->b_state_lock &journal->j_list_lock irq_context: 0 sb_writers#4 &type->i_mutex_dir_key#3/1 sb_internal jbd2_handle &ei->i_data_sem &obj_hash[i].lock irq_context: 0 sb_writers#4 &type->i_mutex_dir_key#3/1 sb_internal jbd2_handle &ei->i_data_sem &____s->seqcount#2 irq_context: 0 sb_writers#4 &type->i_mutex_dir_key#3/1 sb_internal jbd2_handle &ei->i_data_sem &pcp->lock &zone->lock irq_context: 0 sb_writers#4 &type->i_mutex_dir_key#3/1 sb_internal jbd2_handle &ei->i_data_sem &____s->seqcount irq_context: 0 sb_writers#4 &type->i_mutex_dir_key#3/1 sb_internal jbd2_handle &ei->i_data_sem &c->lock irq_context: 0 sb_writers#4 &type->i_mutex_dir_key#3/1 sb_internal jbd2_handle &ei->i_data_sem &journal->j_revoke_lock irq_context: 0 sb_writers#4 &type->i_mutex_dir_key#3/1 sb_internal jbd2_handle &ei->i_data_sem &bgl->locks[i].lock &sbi->s_md_lock irq_context: 0 sb_writers#4 &type->i_mutex_dir_key#3/1 sb_internal jbd2_handle &ei->i_data_sem &sb->s_type->i_lock_key#22 irq_context: 0 sb_writers#4 &type->i_mutex_dir_key#3/1 sb_internal jbd2_handle tk_core.seq.seqcount irq_context: 0 sb_writers#4 &type->i_mutex_dir_key#3/1 sb_internal jbd2_handle &sbi->s_orphan_lock irq_context: 0 sb_writers#4 &type->i_mutex_dir_key#3/1 sb_internal jbd2_handle &xa->xa_lock#9 irq_context: 0 sb_writers#4 &type->i_mutex_dir_key#3/1 sb_internal jbd2_handle &ei->i_es_lock irq_context: 0 sb_writers#4 &type->i_mutex_dir_key#3/1 sb_internal jbd2_handle &journal->j_wait_updates irq_context: 0 sb_writers#4 &type->i_mutex_dir_key#3/1 sb_internal &obj_hash[i].lock irq_context: 0 &xt[i].mutex init_mm.page_table_lock irq_context: softirq rcu_read_lock rcu_read_lock rcu_read_lock &____s->seqcount#2 irq_context: softirq rcu_read_lock rcu_read_lock &pool->lock/1 &p->pi_lock &cfs_rq->removed.lock irq_context: 0 &p->lock &of->mutex kn->active#5 &____s->seqcount#2 irq_context: 0 sb_writers#8 &xattrs->lock irq_context: 0 sb_writers#8 &of->mutex kn->active#5 &____s->seqcount#2 irq_context: 0 sb_writers#8 &of->mutex kn->active#5 &pcp->lock &zone->lock irq_context: 0 sb_writers#8 &of->mutex kn->active#5 &pcp->lock &zone->lock &____s->seqcount irq_context: 0 (wq_completion)ipv6_addrconf (work_completion)(&(&ifa->dad_work)->work) rtnl_mutex &idev->mc_lock rcu_read_lock rcu_read_lock rcu_read_lock_bh rcu_read_lock &c->lock irq_context: 0 &group->notification_waitq &ep->lock &ep->wq irq_context: 0 &group->notification_waitq &ep->lock &ep->wq &p->pi_lock irq_context: 0 &group->notification_waitq &ep->lock &ep->wq &p->pi_lock &rq->__lock irq_context: 0 &group->notification_waitq &ep->lock &ep->wq &p->pi_lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 sb_writers#5 &type->i_mutex_dir_key#5 tomoyo_ss &____s->seqcount#2 irq_context: 0 sb_writers#5 &type->i_mutex_dir_key#5/1 &sb->s_type->i_mutex_key#12 &sb->s_type->i_mutex_key#12/4 &rq->__lock irq_context: 0 sb_writers#5 &type->i_mutex_dir_key#5/1 &sb->s_type->i_mutex_key#12 &sb->s_type->i_mutex_key#12/4 &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 &sb->s_type->i_mutex_key#9 &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 &sb->s_type->i_mutex_key#10 (netlink_chain).rwsem &c->lock irq_context: 0 &sb->s_type->i_mutex_key#10 (netlink_chain).rwsem &____s->seqcount#2 irq_context: 0 &sb->s_type->i_mutex_key#10 (netlink_chain).rwsem &____s->seqcount irq_context: 0 &sb->s_type->i_mutex_key#10 rcu_state.exp_mutex &obj_hash[i].lock pool_lock irq_context: 0 &sb->s_type->i_mutex_key#10 rcu_state.exp_mutex rcu_read_lock &pool->lock fill_pool_map-wait-type-override pool_lock#2 irq_context: softirq (&dsp_spl_tl) dsp_lock fill_pool_map-wait-type-override pool_lock#2 irq_context: 0 &sb->s_type->i_mutex_key#10 rcu_state.exp_mutex rcu_read_lock &pool->lock fill_pool_map-wait-type-override pool_lock irq_context: softirq (&dsp_spl_tl) dsp_lock fill_pool_map-wait-type-override pool_lock irq_context: 0 (wq_completion)rcu_gp (work_completion)(&rew->rew_work) &rq->__lock &cfs_rq->removed.lock irq_context: 0 rtnl_mutex dev_addr_sem rcu_node_0 irq_context: 0 rtnl_mutex dev_addr_sem &rcu_state.expedited_wq irq_context: 0 rtnl_mutex dev_addr_sem &rcu_state.expedited_wq &p->pi_lock irq_context: 0 rtnl_mutex dev_addr_sem &rcu_state.expedited_wq &p->pi_lock &rq->__lock irq_context: 0 (wq_completion)events_unbound (reaper_work).work rcu_read_lock &rq->__lock irq_context: 0 rtnl_mutex dev_addr_sem &rcu_state.expedited_wq &p->pi_lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)events_unbound (reaper_work).work rcu_read_lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 &sighand->siglock &p->pi_lock &rq->__lock &cfs_rq->removed.lock irq_context: 0 (wq_completion)rcu_gp (work_completion)(&(&ssp->srcu_sup->work)->work) &ACCESS_PRIVATE(ssp->srcu_sup, lock) &ACCESS_PRIVATE(sdp, lock) irq_context: 0 rtnl_mutex dev_addr_sem &cfs_rq->removed.lock irq_context: 0 fill_pool_map-wait-type-override &____s->seqcount#2 irq_context: 0 sb_writers#4 &type->i_mutex_dir_key#3/1 &sb->s_type->i_mutex_key#8 &c->lock irq_context: 0 sb_writers#4 &type->i_mutex_dir_key#3/1 tomoyo_ss &pcp->lock &zone->lock irq_context: 0 sb_writers#4 &type->i_mutex_dir_key#3/1 tomoyo_ss &pcp->lock &zone->lock &____s->seqcount irq_context: 0 rtnl_mutex dev_addr_sem rcu_read_lock &pool->lock/1 fill_pool_map-wait-type-override &____s->seqcount#2 irq_context: 0 sb_writers#4 sb_internal &c->lock irq_context: 0 sb_writers#4 &type->i_mutex_dir_key#3/1 &type->i_mutex_dir_key#3 &ei->i_es_lock key#2 irq_context: 0 &p->lock &____s->seqcount#2 irq_context: 0 kn->active#5 &kernfs_locks->open_file_mutex[count] &____s->seqcount#2 irq_context: 0 &mm->mmap_lock &____s->seqcount#2 irq_context: softirq rcu_callback put_task_map-wait-type-override &meta->lock irq_context: softirq rcu_callback put_task_map-wait-type-override kfence_freelist_lock irq_context: 0 &sighand->siglock stock_lock irq_context: 0 tasklist_lock &p->alloc_lock irq_context: 0 (wq_completion)events (linkwatch_work).work rtnl_mutex &tb->tb6_lock &pcp->lock &zone->lock &____s->seqcount irq_context: 0 kn->active#5 &____s->seqcount#2 irq_context: 0 rtnl_mutex &macvlan_netdev_addr_lock_key/1 &c->lock irq_context: 0 rtnl_mutex &macvlan_netdev_addr_lock_key/1 &____s->seqcount#2 irq_context: 0 rtnl_mutex &macvlan_netdev_addr_lock_key/1 &____s->seqcount irq_context: 0 sb_writers#4 &type->i_mutex_dir_key#3/1 sb_internal jbd2_handle &ei->i_data_sem &bgl->locks[i].lock &obj_hash[i].lock irq_context: 0 sb_writers#4 &type->i_mutex_dir_key#3/1 sb_internal jbd2_handle &ei->i_data_sem &bgl->locks[i].lock pool_lock#2 irq_context: 0 sb_writers#4 &type->i_mutex_dir_key#3/1 rcu_read_lock rcu_node_0 irq_context: 0 sb_writers#4 &type->i_mutex_dir_key#3/1 rcu_read_lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 rtnl_mutex dev_addr_sem rcu_read_lock (console_sem).lock irq_context: 0 rtnl_mutex dev_addr_sem rcu_read_lock console_lock console_srcu console_owner_lock irq_context: 0 rtnl_mutex dev_addr_sem rcu_read_lock console_lock console_srcu console_owner irq_context: 0 rtnl_mutex dev_addr_sem rcu_read_lock console_lock console_srcu console_owner &port_lock_key irq_context: 0 rtnl_mutex dev_addr_sem rcu_read_lock console_lock console_srcu console_owner console_owner_lock irq_context: 0 &mm->mmap_lock remove_cache_srcu &rq->__lock irq_context: 0 &mm->mmap_lock remove_cache_srcu &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)mld (work_completion)(&(&idev->mc_ifc_work)->work) &idev->mc_lock rcu_read_lock rcu_read_lock rcu_read_lock_bh &____s->seqcount irq_context: softirq fs/notify/mark.c:89 rcu_read_lock &pool->lock/1 &p->pi_lock &rq->__lock irq_context: softirq fs/notify/mark.c:89 rcu_read_lock &pool->lock/1 &p->pi_lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)events (work_completion)(&ht->run_work) &ht->mutex &rq->__lock irq_context: 0 (wq_completion)events (work_completion)(&rdev->wiphy_work) &rdev->wiphy.mtx &wdev->mtx &rdev->bss_lock &pcp->lock &zone->lock irq_context: 0 (wq_completion)events (work_completion)(&rdev->wiphy_work) &rdev->wiphy.mtx &wdev->mtx &rdev->bss_lock &pcp->lock &zone->lock &____s->seqcount irq_context: 0 rtnl_mutex dev_addr_sem rcu_read_lock &rcu_state.gp_wq irq_context: 0 rtnl_mutex dev_addr_sem rcu_read_lock &rcu_state.gp_wq &p->pi_lock irq_context: 0 rtnl_mutex dev_addr_sem rcu_read_lock &rcu_state.gp_wq &p->pi_lock &rq->__lock irq_context: 0 rtnl_mutex dev_addr_sem rcu_read_lock &rcu_state.gp_wq &p->pi_lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 sb_writers#4 &type->i_mutex_dir_key#3/1 tomoyo_ss &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 loop_validate_mutex &lo->lo_mutex &rq->__lock irq_context: 0 loop_validate_mutex &lo->lo_mutex &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)mld (work_completion)(&(&idev->mc_dad_work)->work) &idev->mc_lock rcu_read_lock &____s->seqcount#2 irq_context: 0 (wq_completion)ipv6_addrconf (work_completion)(&(&ifa->dad_work)->work) rtnl_mutex &idev->mc_lock &vlan_netdev_addr_lock_key/1 &____s->seqcount#2 irq_context: softirq (&ndev->rs_timer) rcu_read_lock rcu_read_lock rcu_read_lock_bh &obj_hash[i].lock pool_lock irq_context: 0 (wq_completion)bat_events (work_completion)(&(&forw_packet_aggr->delayed_work)->work) &hard_iface->bat_iv.ogm_buff_mutex &____s->seqcount#2 irq_context: 0 rtnl_mutex fill_pool_map-wait-type-override &____s->seqcount#2 irq_context: 0 (wq_completion)events_power_efficient (check_lifetime_work).work fill_pool_map-wait-type-override pool_lock irq_context: 0 (wq_completion)events (work_completion)(&(&nsim_dev->trap_data->trap_report_dw)->work) &devlink->lock_key#5 &rq->__lock irq_context: 0 sb_writers#5 &type->i_mutex_dir_key#5/1 lock#4 rcu_read_lock pool_lock#2 irq_context: 0 sb_writers#5 &type->i_mutex_dir_key#5/1 lock#4 &obj_hash[i].lock irq_context: 0 (wq_completion)bat_events (work_completion)(&(&bat_priv->nc.work)->work) &cfs_rq->removed.lock irq_context: 0 (wq_completion)events (work_completion)(&(&nsim_dev->trap_data->trap_report_dw)->work) &devlink->lock_key &nsim_trap_data->trap_lock batched_entropy_u8.lock irq_context: 0 (wq_completion)events (work_completion)(&(&nsim_dev->trap_data->trap_report_dw)->work) &devlink->lock_key &nsim_trap_data->trap_lock kfence_freelist_lock irq_context: 0 (wq_completion)events (work_completion)(&(&nsim_dev->trap_data->trap_report_dw)->work) &devlink->lock_key &nsim_trap_data->trap_lock &meta->lock irq_context: 0 (wq_completion)events (work_completion)(&data->fib_event_work) &data->fib_lock &____s->seqcount#2 irq_context: 0 &vma->vm_lock->lock &____s->seqcount#2 irq_context: 0 sb_writers#4 &obj_hash[i].lock pool_lock irq_context: 0 sb_writers#4 fill_pool_map-wait-type-override pool_lock irq_context: 0 (wq_completion)events (work_completion)(&rdev->wiphy_work) &rdev->wiphy.mtx &wdev->mtx &____s->seqcount#2 irq_context: 0 sb_writers#4 &type->i_mutex_dir_key#3/1 &type->i_mutex_dir_key#3 &mapping->private_lock irq_context: 0 sb_writers#4 &type->i_mutex_dir_key#3/1 sb_internal jbd2_handle &ei->i_data_sem &mapping->private_lock irq_context: softirq (&ndev->rs_timer) &ndev->lock fill_pool_map-wait-type-override pool_lock#2 irq_context: softirq (&ndev->rs_timer) &ndev->lock fill_pool_map-wait-type-override pool_lock irq_context: 0 &mm->mmap_lock remove_cache_srcu rcu_read_lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 sb_writers#8 kn->active#5 &c->lock irq_context: 0 &group->notification_waitq &ep->lock &ep->wq &p->pi_lock &cfs_rq->removed.lock irq_context: 0 (wq_completion)events (linkwatch_work).work rtnl_mutex rcu_read_lock &net->sctp.local_addr_lock &net->sctp.addr_wq_lock &____s->seqcount#2 irq_context: 0 tomoyo_ss remove_cache_srcu rcu_read_lock &rcu_state.gp_wq irq_context: 0 tomoyo_ss remove_cache_srcu rcu_read_lock &rcu_state.gp_wq &p->pi_lock irq_context: 0 tomoyo_ss remove_cache_srcu rcu_read_lock &rcu_state.gp_wq &p->pi_lock &rq->__lock irq_context: 0 tomoyo_ss remove_cache_srcu rcu_read_lock &rcu_state.gp_wq &p->pi_lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: softirq (&peer->timer_persistent_keepalive) irq_context: softirq (&peer->timer_persistent_keepalive) pool_lock#2 irq_context: softirq (&peer->timer_persistent_keepalive) &list->lock#17 irq_context: softirq (&peer->timer_persistent_keepalive) tk_core.seq.seqcount irq_context: softirq (&peer->timer_persistent_keepalive) rcu_read_lock_bh &r->producer_lock#2 irq_context: softirq (&peer->timer_persistent_keepalive) rcu_read_lock_bh rcu_read_lock &pool->lock irq_context: softirq (&peer->timer_persistent_keepalive) rcu_read_lock_bh rcu_read_lock &pool->lock &obj_hash[i].lock irq_context: softirq (&peer->timer_persistent_keepalive) rcu_read_lock_bh rcu_read_lock &pool->lock &p->pi_lock irq_context: softirq (&peer->timer_persistent_keepalive) rcu_read_lock_bh rcu_read_lock &pool->lock &p->pi_lock &rq->__lock irq_context: softirq (&peer->timer_persistent_keepalive) rcu_read_lock_bh rcu_read_lock &pool->lock &p->pi_lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)wg-crypt-wg0 (work_completion)(&peer->transmit_packet_work) &peer->endpoint_lock rcu_read_lock_bh &obj_hash[i].lock irq_context: 0 (wq_completion)wg-crypt-wg0 (work_completion)(&peer->transmit_packet_work) &peer->endpoint_lock rcu_read_lock_bh pool_lock#2 irq_context: 0 (wq_completion)wg-crypt-wg0 (work_completion)(&peer->transmit_packet_work) &peer->endpoint_lock rcu_read_lock_bh rcu_read_lock &c->lock irq_context: 0 (wq_completion)wg-crypt-wg0 (work_completion)(&peer->transmit_packet_work) &peer->endpoint_lock rcu_read_lock_bh rcu_read_lock &dir->lock#2 irq_context: 0 (wq_completion)wg-crypt-wg0 (work_completion)(&peer->transmit_packet_work) &peer->endpoint_lock rcu_read_lock_bh rcu_read_lock &ul->lock irq_context: 0 (wq_completion)wg-crypt-wg0 (work_completion)(&peer->transmit_packet_work) &peer->endpoint_lock rcu_read_lock_bh batched_entropy_u32.lock irq_context: 0 (wq_completion)wg-crypt-wg0 (work_completion)(&peer->transmit_packet_work) &rq->__lock irq_context: 0 (wq_completion)mld (work_completion)(&(&idev->mc_ifc_work)->work) &idev->mc_lock remove_cache_srcu rcu_read_lock rcu_node_0 irq_context: 0 (wq_completion)mld (work_completion)(&(&idev->mc_ifc_work)->work) &idev->mc_lock remove_cache_srcu rcu_read_lock &rq->__lock irq_context: 0 (wq_completion)mld (work_completion)(&(&idev->mc_ifc_work)->work) &idev->mc_lock remove_cache_srcu rcu_read_lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)mld (work_completion)(&(&idev->mc_ifc_work)->work) &idev->mc_lock rcu_read_lock rcu_read_lock rcu_read_lock_bh key#20 irq_context: 0 (wq_completion)mld (work_completion)(&(&idev->mc_ifc_work)->work) &idev->mc_lock rcu_read_lock rcu_read_lock rcu_read_lock_bh &entry->crc_lock irq_context: 0 (wq_completion)mld (work_completion)(&(&idev->mc_ifc_work)->work) &idev->mc_lock rcu_read_lock rcu_read_lock rcu_read_lock_bh &c->lock irq_context: 0 (wq_completion)mld (work_completion)(&(&idev->mc_ifc_work)->work) &idev->mc_lock rcu_read_lock rcu_read_lock rcu_read_lock_bh tk_core.seq.seqcount irq_context: 0 (wq_completion)mld (work_completion)(&(&idev->mc_ifc_work)->work) &idev->mc_lock rcu_read_lock rcu_read_lock rcu_read_lock_bh &list->lock#5 irq_context: 0 sb_writers#4 &type->i_mutex_dir_key#3/1 jbd2_handle &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 sb_writers#4 &type->i_mutex_dir_key#3/1 mmu_notifier_invalidate_range_start &rq->__lock irq_context: 0 rcu_read_lock rcu_read_lock rcu_read_lock &____s->seqcount#7 irq_context: 0 rcu_read_lock rcu_read_lock rcu_read_lock rcu_read_lock &____s->seqcount#7 irq_context: 0 rcu_read_lock rcu_read_lock rcu_read_lock &nf_nat_locks[i] irq_context: 0 rcu_read_lock rcu_read_lock rcu_read_lock &nf_conntrack_locks[i] irq_context: 0 rcu_read_lock rcu_read_lock rcu_read_lock &nf_conntrack_locks[i] &nf_conntrack_locks[i]/1 irq_context: 0 rcu_read_lock rcu_read_lock rcu_read_lock &nf_conntrack_locks[i] &nf_conntrack_locks[i]/1 batched_entropy_u8.lock irq_context: 0 rcu_read_lock rcu_read_lock rcu_read_lock &nf_conntrack_locks[i]/1 irq_context: 0 sb_writers#4 &type->i_mutex_dir_key#3/1 &type->i_mutex_dir_key#3 &c->lock irq_context: 0 sb_writers#8 tomoyo_ss &pcp->lock &zone->lock irq_context: 0 sb_writers#8 tomoyo_ss &pcp->lock &zone->lock &____s->seqcount irq_context: 0 (wq_completion)mld (work_completion)(&(&idev->mc_dad_work)->work) &idev->mc_lock &pcp->lock &zone->lock irq_context: 0 (wq_completion)mld (work_completion)(&(&idev->mc_dad_work)->work) &idev->mc_lock &pcp->lock &zone->lock &____s->seqcount irq_context: 0 (wq_completion)bat_events (work_completion)(&(&forw_packet_aggr->delayed_work)->work) rcu_read_lock_bh rcu_read_lock tk_core.seq.seqcount irq_context: 0 (wq_completion)bat_events (work_completion)(&(&forw_packet_aggr->delayed_work)->work) rcu_read_lock_bh rcu_read_lock &list->lock#5 irq_context: 0 (wq_completion)ipv6_addrconf (work_completion)(&(&ifa->dad_work)->work) rtnl_mutex &idev->mc_lock &macvlan_netdev_addr_lock_key/1 _xmit_ETHER &c->lock irq_context: 0 (wq_completion)ipv6_addrconf (work_completion)(&(&ifa->dad_work)->work) rtnl_mutex &idev->mc_lock &macvlan_netdev_addr_lock_key/1 _xmit_ETHER &____s->seqcount#2 irq_context: 0 (wq_completion)ipv6_addrconf (work_completion)(&(&ifa->dad_work)->work) rtnl_mutex &idev->mc_lock &macvlan_netdev_addr_lock_key/1 _xmit_ETHER &____s->seqcount irq_context: 0 (wq_completion)ipv6_addrconf (work_completion)(&(&ifa->dad_work)->work) rtnl_mutex rcu_read_lock rcu_read_lock rcu_read_lock_bh _xmit_ETHER#2 rcu_read_lock tk_core.seq.seqcount irq_context: 0 (wq_completion)ipv6_addrconf (work_completion)(&(&ifa->dad_work)->work) rtnl_mutex rcu_read_lock rcu_read_lock rcu_read_lock_bh _xmit_ETHER#2 rcu_read_lock &fq->lock irq_context: 0 (wq_completion)ipv6_addrconf (work_completion)(&(&ifa->dad_work)->work) rtnl_mutex rcu_read_lock rcu_read_lock rcu_read_lock_bh _xmit_ETHER#2 rcu_read_lock &local->active_txq_lock[i] irq_context: 0 (wq_completion)ipv6_addrconf (work_completion)(&(&ifa->dad_work)->work) rtnl_mutex rcu_read_lock rcu_read_lock rcu_read_lock_bh _xmit_ETHER#2 rcu_read_lock &local->handle_wake_tx_queue_lock irq_context: 0 (wq_completion)ipv6_addrconf (work_completion)(&(&ifa->dad_work)->work) rtnl_mutex rcu_read_lock rcu_read_lock rcu_read_lock_bh _xmit_ETHER#2 rcu_read_lock &local->handle_wake_tx_queue_lock &local->active_txq_lock[i] irq_context: 0 (wq_completion)ipv6_addrconf (work_completion)(&(&ifa->dad_work)->work) rtnl_mutex rcu_read_lock rcu_read_lock rcu_read_lock_bh _xmit_ETHER#2 rcu_read_lock &local->handle_wake_tx_queue_lock &local->queue_stop_reason_lock irq_context: 0 (wq_completion)ipv6_addrconf (work_completion)(&(&ifa->dad_work)->work) rtnl_mutex rcu_read_lock rcu_read_lock rcu_read_lock_bh _xmit_ETHER#2 rcu_read_lock &local->handle_wake_tx_queue_lock &fq->lock irq_context: 0 (wq_completion)ipv6_addrconf (work_completion)(&(&ifa->dad_work)->work) rtnl_mutex rcu_read_lock rcu_read_lock rcu_read_lock_bh _xmit_ETHER#2 rcu_read_lock &local->handle_wake_tx_queue_lock &fq->lock tk_core.seq.seqcount irq_context: 0 (wq_completion)ipv6_addrconf (work_completion)(&(&ifa->dad_work)->work) rtnl_mutex rcu_read_lock rcu_read_lock rcu_read_lock_bh _xmit_ETHER#2 rcu_read_lock &local->handle_wake_tx_queue_lock tk_core.seq.seqcount irq_context: 0 (wq_completion)ipv6_addrconf (work_completion)(&(&ifa->dad_work)->work) rtnl_mutex rcu_read_lock rcu_read_lock rcu_read_lock_bh _xmit_ETHER#2 rcu_read_lock &local->handle_wake_tx_queue_lock hwsim_radio_lock irq_context: 0 (wq_completion)ipv6_addrconf (work_completion)(&(&ifa->dad_work)->work) rtnl_mutex rcu_read_lock rcu_read_lock rcu_read_lock_bh _xmit_ETHER#2 rcu_read_lock &local->handle_wake_tx_queue_lock hwsim_radio_lock &c->lock irq_context: 0 (wq_completion)ipv6_addrconf (work_completion)(&(&ifa->dad_work)->work) rtnl_mutex rcu_read_lock rcu_read_lock rcu_read_lock_bh _xmit_ETHER#2 rcu_read_lock &local->handle_wake_tx_queue_lock hwsim_radio_lock pool_lock#2 irq_context: 0 (wq_completion)ipv6_addrconf (work_completion)(&(&ifa->dad_work)->work) rtnl_mutex rcu_read_lock rcu_read_lock rcu_read_lock_bh _xmit_ETHER#2 rcu_read_lock &local->handle_wake_tx_queue_lock hwsim_radio_lock &list->lock#19 irq_context: 0 (wq_completion)ipv6_addrconf (work_completion)(&(&ifa->dad_work)->work) rtnl_mutex rcu_read_lock rcu_read_lock rcu_read_lock_bh _xmit_ETHER#2 rcu_read_lock &local->handle_wake_tx_queue_lock &list->lock#19 irq_context: 0 sb_writers#5 &type->i_mutex_dir_key#5/1 &____s->seqcount#2 irq_context: 0 rtnl_mutex (inetaddr_validator_chain).rwsem &rq->__lock irq_context: 0 rtnl_mutex &wg->device_update_lock rcu_state.exp_mutex rcu_read_lock &pool->lock &p->pi_lock &rq->__lock irq_context: 0 rtnl_mutex &wg->device_update_lock rcu_state.exp_mutex rcu_read_lock &pool->lock &p->pi_lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 rtnl_mutex &wg->device_update_lock rcu_state.exp_mutex &rnp->exp_wq[0] irq_context: 0 (wq_completion)ipv6_addrconf (work_completion)(&(&ifa->dad_work)->work) rcu_node_0 irq_context: 0 (wq_completion)ipv6_addrconf (work_completion)(&(&ifa->dad_work)->work) &rcu_state.expedited_wq irq_context: 0 (wq_completion)ipv6_addrconf (work_completion)(&(&ifa->dad_work)->work) &rcu_state.expedited_wq &p->pi_lock irq_context: 0 rtnl_mutex &wg->device_update_lock rcu_state.exp_mutex rcu_read_lock &rq->__lock irq_context: 0 (wq_completion)ipv6_addrconf (work_completion)(&(&ifa->dad_work)->work) &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 rtnl_mutex &wg->device_update_lock rcu_state.exp_mutex rcu_read_lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)wg-kex-wg0#3 irq_context: 0 (wq_completion)wg-kex-wg0#3 (work_completion)(&peer->transmit_handshake_work) irq_context: 0 (wq_completion)wg-kex-wg0#3 (work_completion)(&peer->transmit_handshake_work) &rq->__lock irq_context: 0 (wq_completion)wg-kex-wg0#3 (work_completion)(&peer->transmit_handshake_work) tk_core.seq.seqcount irq_context: 0 (wq_completion)wg-kex-wg0#3 (work_completion)(&peer->transmit_handshake_work) &wg->static_identity.lock irq_context: 0 (wq_completion)wg-kex-wg0#3 (work_completion)(&peer->transmit_handshake_work) &wg->static_identity.lock &handshake->lock irq_context: 0 (wq_completion)wg-kex-wg0#3 (work_completion)(&peer->transmit_handshake_work) &wg->static_identity.lock &handshake->lock crngs.lock irq_context: 0 (wq_completion)wg-kex-wg0#3 (work_completion)(&peer->transmit_handshake_work) &wg->static_identity.lock &handshake->lock tk_core.seq.seqcount irq_context: 0 (wq_completion)wg-kex-wg0#3 (work_completion)(&peer->transmit_handshake_work) &wg->static_identity.lock &handshake->lock &table->lock#2 irq_context: 0 (wq_completion)wg-kex-wg0#3 (work_completion)(&peer->transmit_handshake_work) &wg->static_identity.lock &handshake->lock rcu_read_lock_bh batched_entropy_u32.lock irq_context: 0 (wq_completion)wg-kex-wg0#3 (work_completion)(&peer->transmit_handshake_work) &wg->static_identity.lock &handshake->lock rcu_read_lock_bh &table->lock#2 irq_context: 0 (wq_completion)wg-kex-wg0#3 (work_completion)(&peer->transmit_handshake_work) &cookie->lock irq_context: 0 (wq_completion)wg-kex-wg0#3 (work_completion)(&peer->transmit_handshake_work) &cookie->lock irq_context: 0 (wq_completion)wg-kex-wg0#3 (work_completion)(&peer->transmit_handshake_work) rcu_read_lock_bh &obj_hash[i].lock irq_context: 0 (wq_completion)wg-kex-wg0#3 (work_completion)(&peer->transmit_handshake_work) rcu_read_lock_bh &base->lock irq_context: 0 (wq_completion)wg-kex-wg0#3 (work_completion)(&peer->transmit_handshake_work) rcu_read_lock_bh &base->lock &obj_hash[i].lock irq_context: 0 (wq_completion)wg-kex-wg0#3 (work_completion)(&peer->transmit_handshake_work) &obj_hash[i].lock irq_context: 0 (wq_completion)wg-kex-wg0#3 (work_completion)(&peer->transmit_handshake_work) &c->lock irq_context: 0 (wq_completion)wg-kex-wg0#3 (work_completion)(&peer->transmit_handshake_work) &peer->endpoint_lock irq_context: 0 (wq_completion)wg-kex-wg0#3 (work_completion)(&peer->transmit_handshake_work) &peer->endpoint_lock rcu_read_lock_bh rcu_read_lock &c->lock irq_context: 0 (wq_completion)wg-kex-wg0#3 (work_completion)(&peer->transmit_handshake_work) &peer->endpoint_lock rcu_read_lock_bh rcu_read_lock &____s->seqcount#2 irq_context: 0 (wq_completion)wg-kex-wg0#3 (work_completion)(&peer->transmit_handshake_work) &peer->endpoint_lock rcu_read_lock_bh rcu_read_lock &____s->seqcount irq_context: 0 (wq_completion)wg-kex-wg0#3 (work_completion)(&peer->transmit_handshake_work) &peer->endpoint_lock rcu_read_lock_bh rcu_read_lock pool_lock#2 irq_context: 0 (wq_completion)wg-kex-wg0#3 (work_completion)(&peer->transmit_handshake_work) &peer->endpoint_lock rcu_read_lock_bh rcu_read_lock &dir->lock#2 irq_context: 0 (wq_completion)wg-kex-wg0#3 (work_completion)(&peer->transmit_handshake_work) &peer->endpoint_lock rcu_read_lock_bh crngs.lock irq_context: 0 (wq_completion)wg-kex-wg0#3 (work_completion)(&peer->transmit_handshake_work) &peer->endpoint_lock rcu_read_lock_bh batched_entropy_u32.lock irq_context: 0 (wq_completion)wg-kex-wg0#3 (work_completion)(&peer->transmit_handshake_work) &peer->endpoint_lock rcu_read_lock_bh rcu_read_lock &____s->seqcount#7 irq_context: 0 (wq_completion)wg-kex-wg0#3 (work_completion)(&peer->transmit_handshake_work) &peer->endpoint_lock rcu_read_lock_bh rcu_read_lock &nf_conntrack_locks[i] irq_context: 0 (wq_completion)wg-kex-wg0#3 (work_completion)(&peer->transmit_handshake_work) &peer->endpoint_lock rcu_read_lock_bh rcu_read_lock &nf_conntrack_locks[i] &nf_conntrack_locks[i]/1 irq_context: 0 (wq_completion)wg-kex-wg0#3 (work_completion)(&peer->transmit_handshake_work) &peer->endpoint_lock rcu_read_lock_bh rcu_read_lock &nf_conntrack_locks[i] &nf_conntrack_locks[i]/1 batched_entropy_u8.lock irq_context: 0 (wq_completion)wg-kex-wg0#3 (work_completion)(&peer->transmit_handshake_work) &peer->endpoint_lock rcu_read_lock_bh rcu_read_lock &nf_conntrack_locks[i]/1 irq_context: 0 (wq_completion)wg-kex-wg0#3 (work_completion)(&peer->transmit_handshake_work) &peer->endpoint_lock rcu_read_lock_bh rcu_read_lock &obj_hash[i].lock irq_context: 0 (wq_completion)wg-kex-wg0#3 (work_completion)(&peer->transmit_handshake_work) &peer->endpoint_lock rcu_read_lock_bh rcu_read_lock &tbl->lock irq_context: 0 (wq_completion)wg-kex-wg0#3 (work_completion)(&peer->transmit_handshake_work) &peer->endpoint_lock rcu_read_lock_bh rcu_read_lock &tbl->lock pool_lock#2 irq_context: 0 (wq_completion)wg-kex-wg0#3 (work_completion)(&peer->transmit_handshake_work) &peer->endpoint_lock rcu_read_lock_bh rcu_read_lock &tbl->lock batched_entropy_u32.lock irq_context: 0 (wq_completion)wg-kex-wg0#3 (work_completion)(&peer->transmit_handshake_work) &peer->endpoint_lock rcu_read_lock_bh rcu_read_lock &tbl->lock &obj_hash[i].lock irq_context: 0 (wq_completion)wg-kex-wg0#3 (work_completion)(&peer->transmit_handshake_work) &peer->endpoint_lock rcu_read_lock_bh rcu_read_lock &n->lock irq_context: 0 (wq_completion)wg-kex-wg0#3 (work_completion)(&peer->transmit_handshake_work) &peer->endpoint_lock rcu_read_lock_bh rcu_read_lock &____s->seqcount#9 irq_context: 0 (wq_completion)wg-kex-wg0#3 (work_completion)(&peer->transmit_handshake_work) &peer->endpoint_lock rcu_read_lock_bh rcu_read_lock rcu_read_lock_bh tk_core.seq.seqcount irq_context: 0 (wq_completion)wg-kex-wg0#3 (work_completion)(&peer->transmit_handshake_work) &peer->endpoint_lock rcu_read_lock_bh rcu_read_lock rcu_read_lock_bh &list->lock#5 irq_context: 0 (wq_completion)wg-kex-wg0#3 (work_completion)(&peer->transmit_handshake_work) batched_entropy_u8.lock irq_context: 0 (wq_completion)wg-kex-wg0#3 (work_completion)(&peer->transmit_handshake_work) &peer->endpoint_lock rcu_read_lock_bh rcu_read_lock rcu_read_lock rcu_read_lock &dir->lock#2 irq_context: 0 (wq_completion)ipv6_addrconf (work_completion)(&(&ifa->dad_work)->work) rtnl_mutex &idev->mc_lock &macvlan_netdev_addr_lock_key/1 &c->lock irq_context: 0 (wq_completion)ipv6_addrconf (work_completion)(&(&ifa->dad_work)->work) rtnl_mutex &idev->mc_lock &macvlan_netdev_addr_lock_key/1 &____s->seqcount#2 irq_context: 0 (wq_completion)ipv6_addrconf (work_completion)(&(&ifa->dad_work)->work) rtnl_mutex &idev->mc_lock &macvlan_netdev_addr_lock_key/1 &____s->seqcount irq_context: 0 (wq_completion)bat_events (work_completion)(&(&forw_packet_aggr->delayed_work)->work) &meta->lock irq_context: 0 (wq_completion)bat_events (work_completion)(&(&forw_packet_aggr->delayed_work)->work) kfence_freelist_lock irq_context: 0 (wq_completion)wg-kex-wg1#3 irq_context: 0 (wq_completion)wg-kex-wg1#3 (work_completion)(&peer->transmit_handshake_work) irq_context: 0 (wq_completion)wg-kex-wg1#3 (work_completion)(&peer->transmit_handshake_work) tk_core.seq.seqcount irq_context: 0 (wq_completion)wg-kex-wg1#3 (work_completion)(&peer->transmit_handshake_work) &wg->static_identity.lock irq_context: 0 (wq_completion)wg-kex-wg1#3 (work_completion)(&peer->transmit_handshake_work) &wg->static_identity.lock &handshake->lock irq_context: 0 (wq_completion)wg-kex-wg1#3 (work_completion)(&peer->transmit_handshake_work) &wg->static_identity.lock &handshake->lock crngs.lock irq_context: 0 (wq_completion)wg-kex-wg1#3 (work_completion)(&peer->transmit_handshake_work) &wg->static_identity.lock &handshake->lock tk_core.seq.seqcount irq_context: 0 (wq_completion)wg-kex-wg1#3 (work_completion)(&peer->transmit_handshake_work) &wg->static_identity.lock &handshake->lock &table->lock#2 irq_context: 0 (wq_completion)wg-kex-wg1#3 (work_completion)(&peer->transmit_handshake_work) &wg->static_identity.lock &handshake->lock rcu_read_lock_bh batched_entropy_u32.lock irq_context: 0 (wq_completion)wg-kex-wg1#3 (work_completion)(&peer->transmit_handshake_work) &wg->static_identity.lock &handshake->lock rcu_read_lock_bh &table->lock#2 irq_context: 0 (wq_completion)wg-kex-wg1#3 (work_completion)(&peer->transmit_handshake_work) &cookie->lock irq_context: 0 (wq_completion)wg-kex-wg1#3 (work_completion)(&peer->transmit_handshake_work) &cookie->lock irq_context: 0 (wq_completion)wg-kex-wg1#3 (work_completion)(&peer->transmit_handshake_work) rcu_read_lock_bh &obj_hash[i].lock irq_context: 0 (wq_completion)wg-kex-wg1#3 (work_completion)(&peer->transmit_handshake_work) rcu_read_lock_bh &base->lock irq_context: 0 (wq_completion)wg-kex-wg1#3 (work_completion)(&peer->transmit_handshake_work) rcu_read_lock_bh &base->lock &obj_hash[i].lock irq_context: 0 (wq_completion)wg-kex-wg1#3 (work_completion)(&peer->transmit_handshake_work) &obj_hash[i].lock irq_context: 0 (wq_completion)wg-kex-wg1#3 (work_completion)(&peer->transmit_handshake_work) &peer->endpoint_lock irq_context: 0 (wq_completion)wg-kex-wg1#3 (work_completion)(&peer->transmit_handshake_work) &peer->endpoint_lock rcu_read_lock_bh rcu_read_lock &dir->lock#2 irq_context: 0 (wq_completion)wg-kex-wg1#3 (work_completion)(&peer->transmit_handshake_work) &peer->endpoint_lock rcu_read_lock_bh rcu_read_lock &ul->lock irq_context: 0 (wq_completion)wg-kex-wg1#3 (work_completion)(&peer->transmit_handshake_work) &peer->endpoint_lock rcu_read_lock_bh batched_entropy_u32.lock irq_context: 0 (wq_completion)wg-kex-wg1#3 (work_completion)(&peer->transmit_handshake_work) &peer->endpoint_lock rcu_read_lock_bh rcu_read_lock &____s->seqcount#7 irq_context: 0 (wq_completion)wg-kex-wg1#3 (work_completion)(&peer->transmit_handshake_work) &peer->endpoint_lock rcu_read_lock_bh rcu_read_lock &nf_conntrack_locks[i] irq_context: 0 (wq_completion)wg-kex-wg1#3 (work_completion)(&peer->transmit_handshake_work) &peer->endpoint_lock rcu_read_lock_bh rcu_read_lock &nf_conntrack_locks[i] &nf_conntrack_locks[i]/1 irq_context: 0 (wq_completion)wg-kex-wg1#3 (work_completion)(&peer->transmit_handshake_work) &peer->endpoint_lock rcu_read_lock_bh rcu_read_lock &nf_conntrack_locks[i] &nf_conntrack_locks[i]/1 batched_entropy_u8.lock irq_context: 0 (wq_completion)wg-kex-wg1#3 (work_completion)(&peer->transmit_handshake_work) &peer->endpoint_lock rcu_read_lock_bh rcu_read_lock &nf_conntrack_locks[i]/1 irq_context: 0 (wq_completion)wg-kex-wg1#3 (work_completion)(&peer->transmit_handshake_work) &peer->endpoint_lock rcu_read_lock_bh rcu_read_lock &____s->seqcount#10 irq_context: 0 (wq_completion)wg-kex-wg1#3 (work_completion)(&peer->transmit_handshake_work) &peer->endpoint_lock rcu_read_lock_bh rcu_read_lock rcu_read_lock_bh tk_core.seq.seqcount irq_context: 0 (wq_completion)wg-kex-wg1#3 (work_completion)(&peer->transmit_handshake_work) &peer->endpoint_lock rcu_read_lock_bh rcu_read_lock rcu_read_lock_bh &list->lock#5 irq_context: 0 (wq_completion)wg-kex-wg1#3 (work_completion)(&peer->transmit_handshake_work) batched_entropy_u8.lock irq_context: 0 (wq_completion)wg-kex-wg1#3 (work_completion)(&peer->transmit_handshake_work) &peer->endpoint_lock rcu_read_lock_bh rcu_read_lock rcu_read_lock rcu_read_lock &dir->lock#2 irq_context: 0 (wq_completion)wg-kex-wg1#3 (work_completion)(&peer->transmit_handshake_work) &rq->__lock irq_context: 0 (wq_completion)wg-kex-wg0#4 irq_context: 0 (wq_completion)wg-kex-wg0#4 (work_completion)(&({ do { const void *__vpp_verify = (typeof((worker) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __asm__ ("" : "=r"(__ptr) : "0"((typeof(*((worker))) *)((worker)))); (typeof((typeof(*((worker))) *)((worker)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); })->work) irq_context: 0 (wq_completion)wg-kex-wg0#4 (work_completion)(&({ do { const void *__vpp_verify = (typeof((worker) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __asm__ ("" : "=r"(__ptr) : "0"((typeof(*((worker))) *)((worker)))); (typeof((typeof(*((worker))) *)((worker)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); })->work) &rq->__lock irq_context: 0 (wq_completion)events_power_efficient (check_lifetime_work).work rcu_node_0 irq_context: 0 (wq_completion)events_power_efficient (check_lifetime_work).work rcu_read_lock rcu_node_0 irq_context: 0 rtnl_mutex rcu_read_lock &net->sctp.local_addr_lock &net->sctp.addr_wq_lock &____s->seqcount#2 irq_context: 0 (wq_completion)wg-kex-wg0#4 (work_completion)(&({ do { const void *__vpp_verify = (typeof((worker) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __asm__ ("" : "=r"(__ptr) : "0"((typeof(*((worker))) *)((worker)))); (typeof((typeof(*((worker))) *)((worker)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); })->work) &r->consumer_lock#2 irq_context: 0 (wq_completion)wg-kex-wg0#4 (work_completion)(&({ do { const void *__vpp_verify = (typeof((worker) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __asm__ ("" : "=r"(__ptr) : "0"((typeof(*((worker))) *)((worker)))); (typeof((typeof(*((worker))) *)((worker)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); })->work) &wg->static_identity.lock irq_context: 0 (wq_completion)wg-kex-wg0#4 (work_completion)(&({ do { const void *__vpp_verify = (typeof((worker) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __asm__ ("" : "=r"(__ptr) : "0"((typeof(*((worker))) *)((worker)))); (typeof((typeof(*((worker))) *)((worker)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); })->work) &wg->static_identity.lock &handshake->lock irq_context: 0 (wq_completion)wg-kex-wg0#4 (work_completion)(&({ do { const void *__vpp_verify = (typeof((worker) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __asm__ ("" : "=r"(__ptr) : "0"((typeof(*((worker))) *)((worker)))); (typeof((typeof(*((worker))) *)((worker)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); })->work) &wg->static_identity.lock &handshake->lock tk_core.seq.seqcount irq_context: 0 (wq_completion)wg-kex-wg0#4 (work_completion)(&({ do { const void *__vpp_verify = (typeof((worker) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __asm__ ("" : "=r"(__ptr) : "0"((typeof(*((worker))) *)((worker)))); (typeof((typeof(*((worker))) *)((worker)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); })->work) &wg->static_identity.lock &handshake->lock irq_context: 0 (wq_completion)wg-kex-wg0#4 (work_completion)(&({ do { const void *__vpp_verify = (typeof((worker) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __asm__ ("" : "=r"(__ptr) : "0"((typeof(*((worker))) *)((worker)))); (typeof((typeof(*((worker))) *)((worker)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); })->work) &wg->static_identity.lock &handshake->lock tk_core.seq.seqcount irq_context: 0 (wq_completion)wg-kex-wg0#4 (work_completion)(&({ do { const void *__vpp_verify = (typeof((worker) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __asm__ ("" : "=r"(__ptr) : "0"((typeof(*((worker))) *)((worker)))); (typeof((typeof(*((worker))) *)((worker)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); })->work) &peer->endpoint_lock irq_context: 0 (wq_completion)wg-kex-wg0#4 (work_completion)(&({ do { const void *__vpp_verify = (typeof((worker) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __asm__ ("" : "=r"(__ptr) : "0"((typeof(*((worker))) *)((worker)))); (typeof((typeof(*((worker))) *)((worker)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); })->work) tk_core.seq.seqcount irq_context: 0 (wq_completion)wg-kex-wg0#4 (work_completion)(&({ do { const void *__vpp_verify = (typeof((worker) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __asm__ ("" : "=r"(__ptr) : "0"((typeof(*((worker))) *)((worker)))); (typeof((typeof(*((worker))) *)((worker)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); })->work) &wg->static_identity.lock &handshake->lock crngs.lock irq_context: 0 (wq_completion)wg-kex-wg0#4 (work_completion)(&({ do { const void *__vpp_verify = (typeof((worker) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __asm__ ("" : "=r"(__ptr) : "0"((typeof(*((worker))) *)((worker)))); (typeof((typeof(*((worker))) *)((worker)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); })->work) &wg->static_identity.lock &handshake->lock &table->lock#2 irq_context: 0 (wq_completion)wg-kex-wg0#4 (work_completion)(&({ do { const void *__vpp_verify = (typeof((worker) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __asm__ ("" : "=r"(__ptr) : "0"((typeof(*((worker))) *)((worker)))); (typeof((typeof(*((worker))) *)((worker)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); })->work) &wg->static_identity.lock &handshake->lock rcu_read_lock_bh batched_entropy_u32.lock irq_context: 0 (wq_completion)wg-kex-wg0#4 (work_completion)(&({ do { const void *__vpp_verify = (typeof((worker) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __asm__ ("" : "=r"(__ptr) : "0"((typeof(*((worker))) *)((worker)))); (typeof((typeof(*((worker))) *)((worker)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); })->work) &wg->static_identity.lock &handshake->lock rcu_read_lock_bh &table->lock#2 irq_context: 0 (wq_completion)wg-kex-wg0#4 (work_completion)(&({ do { const void *__vpp_verify = (typeof((worker) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __asm__ ("" : "=r"(__ptr) : "0"((typeof(*((worker))) *)((worker)))); (typeof((typeof(*((worker))) *)((worker)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); })->work) &cookie->lock irq_context: 0 (wq_completion)wg-kex-wg0#4 (work_completion)(&({ do { const void *__vpp_verify = (typeof((worker) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __asm__ ("" : "=r"(__ptr) : "0"((typeof(*((worker))) *)((worker)))); (typeof((typeof(*((worker))) *)((worker)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); })->work) &cookie->lock irq_context: 0 (wq_completion)wg-kex-wg0#4 (work_completion)(&({ do { const void *__vpp_verify = (typeof((worker) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __asm__ ("" : "=r"(__ptr) : "0"((typeof(*((worker))) *)((worker)))); (typeof((typeof(*((worker))) *)((worker)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); })->work) &handshake->lock irq_context: 0 (wq_completion)wg-kex-wg0#4 (work_completion)(&({ do { const void *__vpp_verify = (typeof((worker) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __asm__ ("" : "=r"(__ptr) : "0"((typeof(*((worker))) *)((worker)))); (typeof((typeof(*((worker))) *)((worker)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); })->work) &handshake->lock fs_reclaim irq_context: 0 (wq_completion)wg-kex-wg0#4 (work_completion)(&({ do { const void *__vpp_verify = (typeof((worker) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __asm__ ("" : "=r"(__ptr) : "0"((typeof(*((worker))) *)((worker)))); (typeof((typeof(*((worker))) *)((worker)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); })->work) &handshake->lock fs_reclaim mmu_notifier_invalidate_range_start irq_context: 0 (wq_completion)wg-kex-wg0#4 (work_completion)(&({ do { const void *__vpp_verify = (typeof((worker) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __asm__ ("" : "=r"(__ptr) : "0"((typeof(*((worker))) *)((worker)))); (typeof((typeof(*((worker))) *)((worker)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); })->work) &handshake->lock tk_core.seq.seqcount irq_context: 0 (wq_completion)wg-kex-wg0#4 (work_completion)(&({ do { const void *__vpp_verify = (typeof((worker) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __asm__ ("" : "=r"(__ptr) : "0"((typeof(*((worker))) *)((worker)))); (typeof((typeof(*((worker))) *)((worker)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); })->work) &handshake->lock rcu_read_lock_bh &peer->keypairs.keypair_update_lock irq_context: 0 (wq_completion)wg-kex-wg0#4 (work_completion)(&({ do { const void *__vpp_verify = (typeof((worker) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __asm__ ("" : "=r"(__ptr) : "0"((typeof(*((worker))) *)((worker)))); (typeof((typeof(*((worker))) *)((worker)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); })->work) &handshake->lock rcu_read_lock_bh &table->lock#2 irq_context: 0 (wq_completion)wg-kex-wg0#4 (work_completion)(&({ do { const void *__vpp_verify = (typeof((worker) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __asm__ ("" : "=r"(__ptr) : "0"((typeof(*((worker))) *)((worker)))); (typeof((typeof(*((worker))) *)((worker)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); })->work) rcu_read_lock_bh &obj_hash[i].lock irq_context: 0 (wq_completion)wg-kex-wg0#4 (work_completion)(&({ do { const void *__vpp_verify = (typeof((worker) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __asm__ ("" : "=r"(__ptr) : "0"((typeof(*((worker))) *)((worker)))); (typeof((typeof(*((worker))) *)((worker)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); })->work) rcu_read_lock_bh &base->lock irq_context: 0 (wq_completion)wg-kex-wg0#4 (work_completion)(&({ do { const void *__vpp_verify = (typeof((worker) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __asm__ ("" : "=r"(__ptr) : "0"((typeof(*((worker))) *)((worker)))); (typeof((typeof(*((worker))) *)((worker)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); })->work) rcu_read_lock_bh &base->lock &obj_hash[i].lock irq_context: 0 (wq_completion)wg-kex-wg0#4 (work_completion)(&({ do { const void *__vpp_verify = (typeof((worker) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __asm__ ("" : "=r"(__ptr) : "0"((typeof(*((worker))) *)((worker)))); (typeof((typeof(*((worker))) *)((worker)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); })->work) &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)wg-kex-wg0#4 (work_completion)(&({ do { const void *__vpp_verify = (typeof((worker) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __asm__ ("" : "=r"(__ptr) : "0"((typeof(*((worker))) *)((worker)))); (typeof((typeof(*((worker))) *)((worker)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); })->work) &obj_hash[i].lock irq_context: 0 (wq_completion)wg-kex-wg0#4 (work_completion)(&({ do { const void *__vpp_verify = (typeof((worker) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __asm__ ("" : "=r"(__ptr) : "0"((typeof(*((worker))) *)((worker)))); (typeof((typeof(*((worker))) *)((worker)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); })->work) &peer->endpoint_lock irq_context: 0 (wq_completion)wg-kex-wg0#4 (work_completion)(&({ do { const void *__vpp_verify = (typeof((worker) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __asm__ ("" : "=r"(__ptr) : "0"((typeof(*((worker))) *)((worker)))); (typeof((typeof(*((worker))) *)((worker)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); })->work) &peer->endpoint_lock rcu_read_lock_bh rcu_read_lock &____s->seqcount#7 irq_context: 0 (wq_completion)wg-kex-wg0#4 (work_completion)(&({ do { const void *__vpp_verify = (typeof((worker) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __asm__ ("" : "=r"(__ptr) : "0"((typeof(*((worker))) *)((worker)))); (typeof((typeof(*((worker))) *)((worker)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); })->work) &peer->endpoint_lock rcu_read_lock_bh rcu_read_lock &____s->seqcount#10 irq_context: 0 (wq_completion)wg-kex-wg0#4 (work_completion)(&({ do { const void *__vpp_verify = (typeof((worker) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __asm__ ("" : "=r"(__ptr) : "0"((typeof(*((worker))) *)((worker)))); (typeof((typeof(*((worker))) *)((worker)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); })->work) &peer->endpoint_lock rcu_read_lock_bh rcu_read_lock rcu_read_lock_bh tk_core.seq.seqcount irq_context: 0 (wq_completion)wg-kex-wg0#4 (work_completion)(&({ do { const void *__vpp_verify = (typeof((worker) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __asm__ ("" : "=r"(__ptr) : "0"((typeof(*((worker))) *)((worker)))); (typeof((typeof(*((worker))) *)((worker)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); })->work) &peer->endpoint_lock rcu_read_lock_bh rcu_read_lock rcu_read_lock_bh &list->lock#5 irq_context: 0 (wq_completion)wg-kex-wg1#4 irq_context: 0 (wq_completion)wg-kex-wg1#4 (work_completion)(&({ do { const void *__vpp_verify = (typeof((worker) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __asm__ ("" : "=r"(__ptr) : "0"((typeof(*((worker))) *)((worker)))); (typeof((typeof(*((worker))) *)((worker)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); })->work) irq_context: 0 (wq_completion)wg-kex-wg1#4 (work_completion)(&({ do { const void *__vpp_verify = (typeof((worker) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __asm__ ("" : "=r"(__ptr) : "0"((typeof(*((worker))) *)((worker)))); (typeof((typeof(*((worker))) *)((worker)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); })->work) &r->consumer_lock#2 irq_context: 0 (wq_completion)wg-kex-wg1#4 (work_completion)(&({ do { const void *__vpp_verify = (typeof((worker) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __asm__ ("" : "=r"(__ptr) : "0"((typeof(*((worker))) *)((worker)))); (typeof((typeof(*((worker))) *)((worker)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); })->work) &wg->static_identity.lock irq_context: 0 (wq_completion)wg-kex-wg1#4 (work_completion)(&({ do { const void *__vpp_verify = (typeof((worker) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __asm__ ("" : "=r"(__ptr) : "0"((typeof(*((worker))) *)((worker)))); (typeof((typeof(*((worker))) *)((worker)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); })->work) &wg->static_identity.lock &handshake->lock irq_context: 0 (wq_completion)wg-kex-wg1#4 (work_completion)(&({ do { const void *__vpp_verify = (typeof((worker) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __asm__ ("" : "=r"(__ptr) : "0"((typeof(*((worker))) *)((worker)))); (typeof((typeof(*((worker))) *)((worker)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); })->work) &wg->static_identity.lock &handshake->lock irq_context: 0 (wq_completion)wg-kex-wg1#4 (work_completion)(&({ do { const void *__vpp_verify = (typeof((worker) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __asm__ ("" : "=r"(__ptr) : "0"((typeof(*((worker))) *)((worker)))); (typeof((typeof(*((worker))) *)((worker)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); })->work) &peer->endpoint_lock irq_context: 0 (wq_completion)wg-kex-wg1#4 (work_completion)(&({ do { const void *__vpp_verify = (typeof((worker) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __asm__ ("" : "=r"(__ptr) : "0"((typeof(*((worker))) *)((worker)))); (typeof((typeof(*((worker))) *)((worker)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); })->work) &handshake->lock irq_context: 0 (wq_completion)wg-kex-wg1#4 (work_completion)(&({ do { const void *__vpp_verify = (typeof((worker) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __asm__ ("" : "=r"(__ptr) : "0"((typeof(*((worker))) *)((worker)))); (typeof((typeof(*((worker))) *)((worker)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); })->work) &handshake->lock fs_reclaim irq_context: 0 (wq_completion)wg-kex-wg1#4 (work_completion)(&({ do { const void *__vpp_verify = (typeof((worker) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __asm__ ("" : "=r"(__ptr) : "0"((typeof(*((worker))) *)((worker)))); (typeof((typeof(*((worker))) *)((worker)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); })->work) &handshake->lock fs_reclaim mmu_notifier_invalidate_range_start irq_context: 0 (wq_completion)wg-kex-wg1#4 (work_completion)(&({ do { const void *__vpp_verify = (typeof((worker) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __asm__ ("" : "=r"(__ptr) : "0"((typeof(*((worker))) *)((worker)))); (typeof((typeof(*((worker))) *)((worker)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); })->work) &handshake->lock tk_core.seq.seqcount irq_context: 0 (wq_completion)wg-kex-wg1#4 (work_completion)(&({ do { const void *__vpp_verify = (typeof((worker) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __asm__ ("" : "=r"(__ptr) : "0"((typeof(*((worker))) *)((worker)))); (typeof((typeof(*((worker))) *)((worker)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); })->work) &handshake->lock rcu_read_lock_bh &peer->keypairs.keypair_update_lock irq_context: 0 (wq_completion)wg-kex-wg1#4 (work_completion)(&({ do { const void *__vpp_verify = (typeof((worker) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __asm__ ("" : "=r"(__ptr) : "0"((typeof(*((worker))) *)((worker)))); (typeof((typeof(*((worker))) *)((worker)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); })->work) &handshake->lock rcu_read_lock_bh &table->lock#2 irq_context: 0 (wq_completion)wg-kex-wg1#4 (work_completion)(&({ do { const void *__vpp_verify = (typeof((worker) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __asm__ ("" : "=r"(__ptr) : "0"((typeof(*((worker))) *)((worker)))); (typeof((typeof(*((worker))) *)((worker)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); })->work) rcu_read_lock_bh &obj_hash[i].lock irq_context: 0 (wq_completion)wg-kex-wg1#4 (work_completion)(&({ do { const void *__vpp_verify = (typeof((worker) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __asm__ ("" : "=r"(__ptr) : "0"((typeof(*((worker))) *)((worker)))); (typeof((typeof(*((worker))) *)((worker)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); })->work) rcu_read_lock_bh &base->lock irq_context: 0 (wq_completion)wg-kex-wg1#4 (work_completion)(&({ do { const void *__vpp_verify = (typeof((worker) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __asm__ ("" : "=r"(__ptr) : "0"((typeof(*((worker))) *)((worker)))); (typeof((typeof(*((worker))) *)((worker)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); })->work) rcu_read_lock_bh &base->lock &obj_hash[i].lock irq_context: 0 (wq_completion)wg-kex-wg1#4 (work_completion)(&({ do { const void *__vpp_verify = (typeof((worker) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __asm__ ("" : "=r"(__ptr) : "0"((typeof(*((worker))) *)((worker)))); (typeof((typeof(*((worker))) *)((worker)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); })->work) &obj_hash[i].lock irq_context: 0 (wq_completion)wg-kex-wg1#4 (work_completion)(&({ do { const void *__vpp_verify = (typeof((worker) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __asm__ ("" : "=r"(__ptr) : "0"((typeof(*((worker))) *)((worker)))); (typeof((typeof(*((worker))) *)((worker)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); })->work) &base->lock irq_context: 0 (wq_completion)wg-kex-wg1#4 (work_completion)(&({ do { const void *__vpp_verify = (typeof((worker) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __asm__ ("" : "=r"(__ptr) : "0"((typeof(*((worker))) *)((worker)))); (typeof((typeof(*((worker))) *)((worker)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); })->work) &base->lock &obj_hash[i].lock irq_context: 0 (wq_completion)wg-kex-wg1#4 (work_completion)(&({ do { const void *__vpp_verify = (typeof((worker) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __asm__ ("" : "=r"(__ptr) : "0"((typeof(*((worker))) *)((worker)))); (typeof((typeof(*((worker))) *)((worker)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); })->work) tk_core.seq.seqcount irq_context: 0 (wq_completion)wg-kex-wg1#4 (work_completion)(&({ do { const void *__vpp_verify = (typeof((worker) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __asm__ ("" : "=r"(__ptr) : "0"((typeof(*((worker))) *)((worker)))); (typeof((typeof(*((worker))) *)((worker)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); })->work) &list->lock#17 irq_context: 0 (wq_completion)wg-kex-wg1#4 (work_completion)(&({ do { const void *__vpp_verify = (typeof((worker) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __asm__ ("" : "=r"(__ptr) : "0"((typeof(*((worker))) *)((worker)))); (typeof((typeof(*((worker))) *)((worker)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); })->work) rcu_read_lock_bh &r->producer_lock#2 irq_context: 0 (wq_completion)wg-kex-wg1#4 (work_completion)(&({ do { const void *__vpp_verify = (typeof((worker) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __asm__ ("" : "=r"(__ptr) : "0"((typeof(*((worker))) *)((worker)))); (typeof((typeof(*((worker))) *)((worker)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); })->work) rcu_read_lock_bh rcu_read_lock &pool->lock irq_context: 0 (wq_completion)wg-kex-wg1#4 (work_completion)(&({ do { const void *__vpp_verify = (typeof((worker) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __asm__ ("" : "=r"(__ptr) : "0"((typeof(*((worker))) *)((worker)))); (typeof((typeof(*((worker))) *)((worker)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); })->work) rcu_read_lock_bh rcu_read_lock &pool->lock &obj_hash[i].lock irq_context: 0 (wq_completion)wg-kex-wg1#4 (work_completion)(&({ do { const void *__vpp_verify = (typeof((worker) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __asm__ ("" : "=r"(__ptr) : "0"((typeof(*((worker))) *)((worker)))); (typeof((typeof(*((worker))) *)((worker)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); })->work) rcu_read_lock_bh rcu_read_lock &pool->lock &p->pi_lock irq_context: 0 (wq_completion)wg-crypt-wg1#2 irq_context: 0 (wq_completion)wg-crypt-wg1#2 (work_completion)(&({ do { const void *__vpp_verify = (typeof((worker) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __asm__ ("" : "=r"(__ptr) : "0"((typeof(*((worker))) *)((worker)))); (typeof((typeof(*((worker))) *)((worker)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); })->work) irq_context: 0 (wq_completion)wg-crypt-wg1#2 (work_completion)(&({ do { const void *__vpp_verify = (typeof((worker) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __asm__ ("" : "=r"(__ptr) : "0"((typeof(*((worker))) *)((worker)))); (typeof((typeof(*((worker))) *)((worker)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); })->work) &r->consumer_lock#2 irq_context: 0 (wq_completion)wg-crypt-wg1#2 (work_completion)(&({ do { const void *__vpp_verify = (typeof((worker) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __asm__ ("" : "=r"(__ptr) : "0"((typeof(*((worker))) *)((worker)))); (typeof((typeof(*((worker))) *)((worker)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); })->work) rcu_read_lock &pool->lock irq_context: 0 (wq_completion)wg-crypt-wg1#2 (work_completion)(&({ do { const void *__vpp_verify = (typeof((worker) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __asm__ ("" : "=r"(__ptr) : "0"((typeof(*((worker))) *)((worker)))); (typeof((typeof(*((worker))) *)((worker)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); })->work) rcu_read_lock &pool->lock &obj_hash[i].lock irq_context: 0 (wq_completion)wg-crypt-wg1#2 (work_completion)(&({ do { const void *__vpp_verify = (typeof((worker) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __asm__ ("" : "=r"(__ptr) : "0"((typeof(*((worker))) *)((worker)))); (typeof((typeof(*((worker))) *)((worker)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); })->work) rcu_read_lock &pool->lock &p->pi_lock irq_context: 0 (wq_completion)wg-crypt-wg1#2 (work_completion)(&peer->transmit_packet_work) irq_context: 0 (wq_completion)wg-crypt-wg1#2 (work_completion)(&peer->transmit_packet_work) rcu_read_lock_bh &obj_hash[i].lock irq_context: 0 (wq_completion)wg-crypt-wg1#2 (work_completion)(&peer->transmit_packet_work) &obj_hash[i].lock irq_context: 0 (wq_completion)wg-crypt-wg1#2 (work_completion)(&peer->transmit_packet_work) &peer->endpoint_lock irq_context: 0 (wq_completion)wg-crypt-wg1#2 (work_completion)(&peer->transmit_packet_work) &peer->endpoint_lock rcu_read_lock_bh rcu_read_lock &____s->seqcount#7 irq_context: 0 (wq_completion)wg-crypt-wg1#2 (work_completion)(&peer->transmit_packet_work) &peer->endpoint_lock rcu_read_lock_bh rcu_read_lock &____s->seqcount#10 irq_context: 0 (wq_completion)wg-crypt-wg1#2 (work_completion)(&peer->transmit_packet_work) &peer->endpoint_lock rcu_read_lock_bh rcu_read_lock rcu_read_lock_bh tk_core.seq.seqcount irq_context: 0 (wq_completion)wg-crypt-wg1#2 (work_completion)(&peer->transmit_packet_work) &peer->endpoint_lock rcu_read_lock_bh rcu_read_lock rcu_read_lock_bh &list->lock#5 irq_context: 0 (wq_completion)wg-crypt-wg1#2 (work_completion)(&peer->transmit_packet_work) rcu_read_lock_bh tk_core.seq.seqcount irq_context: 0 (wq_completion)wg-crypt-wg0#2 irq_context: 0 (wq_completion)wg-crypt-wg0#2 (work_completion)(&({ do { const void *__vpp_verify = (typeof((worker) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __asm__ ("" : "=r"(__ptr) : "0"((typeof(*((worker))) *)((worker)))); (typeof((typeof(*((worker))) *)((worker)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); })->work) irq_context: 0 (wq_completion)wg-crypt-wg0#2 (work_completion)(&({ do { const void *__vpp_verify = (typeof((worker) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __asm__ ("" : "=r"(__ptr) : "0"((typeof(*((worker))) *)((worker)))); (typeof((typeof(*((worker))) *)((worker)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); })->work) &r->consumer_lock#2 irq_context: 0 (wq_completion)wg-crypt-wg0#2 (work_completion)(&({ do { const void *__vpp_verify = (typeof((worker) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __asm__ ("" : "=r"(__ptr) : "0"((typeof(*((worker))) *)((worker)))); (typeof((typeof(*((worker))) *)((worker)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); })->work) tk_core.seq.seqcount irq_context: 0 (wq_completion)wg-crypt-wg0#2 (work_completion)(&({ do { const void *__vpp_verify = (typeof((worker) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __asm__ ("" : "=r"(__ptr) : "0"((typeof(*((worker))) *)((worker)))); (typeof((typeof(*((worker))) *)((worker)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); })->work) rcu_read_lock &pool->lock irq_context: 0 (wq_completion)wg-crypt-wg0#2 (work_completion)(&({ do { const void *__vpp_verify = (typeof((worker) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __asm__ ("" : "=r"(__ptr) : "0"((typeof(*((worker))) *)((worker)))); (typeof((typeof(*((worker))) *)((worker)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); })->work) rcu_read_lock &pool->lock &obj_hash[i].lock irq_context: 0 (wq_completion)wg-crypt-wg0#2 (work_completion)(&({ do { const void *__vpp_verify = (typeof((worker) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __asm__ ("" : "=r"(__ptr) : "0"((typeof(*((worker))) *)((worker)))); (typeof((typeof(*((worker))) *)((worker)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); })->work) rcu_read_lock &pool->lock &p->pi_lock irq_context: 0 (wq_completion)wg-crypt-wg0#2 (work_completion)(&({ do { const void *__vpp_verify = (typeof((worker) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __asm__ ("" : "=r"(__ptr) : "0"((typeof(*((worker))) *)((worker)))); (typeof((typeof(*((worker))) *)((worker)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); })->work) rcu_read_lock &pool->lock &p->pi_lock &rq->__lock irq_context: 0 (wq_completion)wg-crypt-wg0#2 (work_completion)(&({ do { const void *__vpp_verify = (typeof((worker) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __asm__ ("" : "=r"(__ptr) : "0"((typeof(*((worker))) *)((worker)))); (typeof((typeof(*((worker))) *)((worker)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); })->work) rcu_read_lock &pool->lock &p->pi_lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)wg-crypt-wg0#2 (work_completion)(&peer->transmit_packet_work) irq_context: 0 (wq_completion)wg-crypt-wg0#2 (work_completion)(&peer->transmit_packet_work) rcu_read_lock_bh &obj_hash[i].lock irq_context: 0 (wq_completion)wg-crypt-wg0#2 (work_completion)(&peer->transmit_packet_work) &obj_hash[i].lock irq_context: 0 (wq_completion)wg-crypt-wg0#2 (work_completion)(&peer->transmit_packet_work) &peer->endpoint_lock irq_context: 0 (wq_completion)wg-crypt-wg0#2 (work_completion)(&peer->transmit_packet_work) &peer->endpoint_lock rcu_read_lock_bh &obj_hash[i].lock irq_context: 0 (wq_completion)wg-crypt-wg0#2 (work_completion)(&peer->transmit_packet_work) &peer->endpoint_lock rcu_read_lock_bh pool_lock#2 irq_context: 0 rtnl_mutex &wg->device_update_lock &____s->seqcount#2 irq_context: 0 (wq_completion)wg-crypt-wg0#2 (work_completion)(&peer->transmit_packet_work) &peer->endpoint_lock rcu_read_lock_bh rcu_read_lock &____s->seqcount#7 irq_context: 0 rtnl_mutex &wg->device_update_lock &____s->seqcount irq_context: 0 (wq_completion)wg-crypt-wg0#2 (work_completion)(&peer->transmit_packet_work) &peer->endpoint_lock rcu_read_lock_bh rcu_read_lock &____s->seqcount#10 irq_context: 0 (wq_completion)wg-crypt-wg0#2 (work_completion)(&peer->transmit_packet_work) &peer->endpoint_lock rcu_read_lock_bh rcu_read_lock rcu_read_lock_bh tk_core.seq.seqcount irq_context: 0 (wq_completion)wg-crypt-wg0#2 (work_completion)(&peer->transmit_packet_work) &peer->endpoint_lock rcu_read_lock_bh rcu_read_lock rcu_read_lock_bh &list->lock#5 irq_context: 0 (wq_completion)wg-crypt-wg1#2 (work_completion)(&({ do { const void *__vpp_verify = (typeof((worker) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __asm__ ("" : "=r"(__ptr) : "0"((typeof(*((worker))) *)((worker)))); (typeof((typeof(*((worker))) *)((worker)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); })->work) tk_core.seq.seqcount irq_context: 0 (wq_completion)wg-kex-wg2#3 irq_context: 0 (wq_completion)wg-kex-wg2#3 (work_completion)(&peer->transmit_handshake_work) irq_context: 0 (wq_completion)wg-kex-wg2#3 (work_completion)(&peer->transmit_handshake_work) &rq->__lock irq_context: 0 (wq_completion)wg-kex-wg2#3 (work_completion)(&peer->transmit_handshake_work) tk_core.seq.seqcount irq_context: 0 (wq_completion)wg-kex-wg2#3 (work_completion)(&peer->transmit_handshake_work) &wg->static_identity.lock irq_context: 0 (wq_completion)wg-kex-wg2#3 (work_completion)(&peer->transmit_handshake_work) &wg->static_identity.lock &handshake->lock irq_context: 0 (wq_completion)wg-kex-wg2#3 (work_completion)(&peer->transmit_handshake_work) &wg->static_identity.lock &handshake->lock crngs.lock irq_context: 0 (wq_completion)wg-kex-wg2#3 (work_completion)(&peer->transmit_handshake_work) &wg->static_identity.lock &handshake->lock tk_core.seq.seqcount irq_context: 0 (wq_completion)wg-kex-wg2#3 (work_completion)(&peer->transmit_handshake_work) &wg->static_identity.lock &handshake->lock &table->lock#2 irq_context: 0 (wq_completion)wg-kex-wg2#3 (work_completion)(&peer->transmit_handshake_work) &wg->static_identity.lock &handshake->lock rcu_read_lock_bh batched_entropy_u32.lock irq_context: 0 (wq_completion)wg-kex-wg2#3 (work_completion)(&peer->transmit_handshake_work) &wg->static_identity.lock &handshake->lock rcu_read_lock_bh batched_entropy_u32.lock crngs.lock irq_context: 0 (wq_completion)wg-kex-wg2#3 (work_completion)(&peer->transmit_handshake_work) &wg->static_identity.lock &handshake->lock rcu_read_lock_bh &table->lock#2 irq_context: 0 (wq_completion)wg-kex-wg2#3 (work_completion)(&peer->transmit_handshake_work) &cookie->lock irq_context: 0 (wq_completion)wg-kex-wg2#3 (work_completion)(&peer->transmit_handshake_work) &cookie->lock irq_context: 0 (wq_completion)wg-kex-wg2#3 (work_completion)(&peer->transmit_handshake_work) rcu_read_lock_bh &obj_hash[i].lock irq_context: 0 (wq_completion)wg-kex-wg2#3 (work_completion)(&peer->transmit_handshake_work) rcu_read_lock_bh &base->lock irq_context: 0 (wq_completion)wg-kex-wg2#3 (work_completion)(&peer->transmit_handshake_work) rcu_read_lock_bh &base->lock &obj_hash[i].lock irq_context: 0 (wq_completion)wg-kex-wg2#3 (work_completion)(&peer->transmit_handshake_work) &obj_hash[i].lock irq_context: 0 (wq_completion)wg-kex-wg2#3 (work_completion)(&peer->transmit_handshake_work) &peer->endpoint_lock irq_context: 0 (wq_completion)wg-kex-wg2#3 (work_completion)(&peer->transmit_handshake_work) &peer->endpoint_lock rcu_read_lock_bh rcu_read_lock &____s->seqcount#7 irq_context: 0 (wq_completion)wg-kex-wg2#3 (work_completion)(&peer->transmit_handshake_work) &peer->endpoint_lock rcu_read_lock_bh rcu_read_lock &____s->seqcount#10 irq_context: 0 (wq_completion)wg-kex-wg2#3 (work_completion)(&peer->transmit_handshake_work) &peer->endpoint_lock rcu_read_lock_bh rcu_read_lock rcu_read_lock_bh tk_core.seq.seqcount irq_context: 0 (wq_completion)wg-kex-wg2#3 (work_completion)(&peer->transmit_handshake_work) &peer->endpoint_lock rcu_read_lock_bh rcu_read_lock rcu_read_lock_bh &list->lock#5 irq_context: 0 (wq_completion)wg-kex-wg2#3 (work_completion)(&peer->transmit_handshake_work) batched_entropy_u8.lock irq_context: 0 (wq_completion)wg-kex-wg2#3 (work_completion)(&peer->transmit_handshake_work) &peer->endpoint_lock rcu_read_lock_bh rcu_read_lock &dir->lock#2 irq_context: 0 (wq_completion)wg-kex-wg2#3 (work_completion)(&peer->transmit_handshake_work) &peer->endpoint_lock rcu_read_lock_bh rcu_read_lock &ul->lock irq_context: 0 (wq_completion)wg-kex-wg2#3 (work_completion)(&peer->transmit_handshake_work) &peer->endpoint_lock rcu_read_lock_bh batched_entropy_u32.lock irq_context: 0 rtnl_mutex &tb->tb6_lock &____s->seqcount#2 irq_context: softirq rcu_read_lock rcu_read_lock &pool->lock/1 &p->pi_lock &rq->__lock &cfs_rq->removed.lock irq_context: 0 cb_lock genl_mutex rcu_read_lock &c->lock irq_context: 0 cb_lock genl_mutex uevent_sock_mutex &c->lock irq_context: 0 cb_lock genl_mutex uevent_sock_mutex &rq->__lock irq_context: 0 (wq_completion)wg-kex-wg1#4 (work_completion)(&({ do { const void *__vpp_verify = (typeof((worker) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __asm__ ("" : "=r"(__ptr) : "0"((typeof(*((worker))) *)((worker)))); (typeof((typeof(*((worker))) *)((worker)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); })->work) &wg->static_identity.lock &handshake->lock tk_core.seq.seqcount irq_context: 0 (wq_completion)wg-kex-wg1#4 (work_completion)(&({ do { const void *__vpp_verify = (typeof((worker) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __asm__ ("" : "=r"(__ptr) : "0"((typeof(*((worker))) *)((worker)))); (typeof((typeof(*((worker))) *)((worker)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); })->work) &wg->static_identity.lock &handshake->lock tk_core.seq.seqcount irq_context: 0 (wq_completion)wg-kex-wg2#4 irq_context: 0 (wq_completion)wg-kex-wg1#4 (work_completion)(&({ do { const void *__vpp_verify = (typeof((worker) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __asm__ ("" : "=r"(__ptr) : "0"((typeof(*((worker))) *)((worker)))); (typeof((typeof(*((worker))) *)((worker)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); })->work) &wg->static_identity.lock &handshake->lock crngs.lock irq_context: 0 (wq_completion)wg-kex-wg2#4 (work_completion)(&({ do { const void *__vpp_verify = (typeof((worker) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __asm__ ("" : "=r"(__ptr) : "0"((typeof(*((worker))) *)((worker)))); (typeof((typeof(*((worker))) *)((worker)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); })->work) irq_context: 0 (wq_completion)wg-kex-wg2#4 (work_completion)(&({ do { const void *__vpp_verify = (typeof((worker) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __asm__ ("" : "=r"(__ptr) : "0"((typeof(*((worker))) *)((worker)))); (typeof((typeof(*((worker))) *)((worker)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); })->work) &r->consumer_lock#2 irq_context: 0 (wq_completion)wg-kex-wg2#4 (work_completion)(&({ do { const void *__vpp_verify = (typeof((worker) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __asm__ ("" : "=r"(__ptr) : "0"((typeof(*((worker))) *)((worker)))); (typeof((typeof(*((worker))) *)((worker)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); })->work) &wg->static_identity.lock irq_context: 0 (wq_completion)wg-kex-wg2#4 (work_completion)(&({ do { const void *__vpp_verify = (typeof((worker) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __asm__ ("" : "=r"(__ptr) : "0"((typeof(*((worker))) *)((worker)))); (typeof((typeof(*((worker))) *)((worker)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); })->work) &wg->static_identity.lock &handshake->lock irq_context: 0 (wq_completion)wg-kex-wg1#4 (work_completion)(&({ do { const void *__vpp_verify = (typeof((worker) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __asm__ ("" : "=r"(__ptr) : "0"((typeof(*((worker))) *)((worker)))); (typeof((typeof(*((worker))) *)((worker)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); })->work) &wg->static_identity.lock &handshake->lock &table->lock#2 irq_context: 0 (wq_completion)wg-kex-wg1#4 (work_completion)(&({ do { const void *__vpp_verify = (typeof((worker) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __asm__ ("" : "=r"(__ptr) : "0"((typeof(*((worker))) *)((worker)))); (typeof((typeof(*((worker))) *)((worker)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); })->work) &wg->static_identity.lock &handshake->lock rcu_read_lock_bh batched_entropy_u32.lock irq_context: 0 (wq_completion)wg-kex-wg1#4 (work_completion)(&({ do { const void *__vpp_verify = (typeof((worker) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __asm__ ("" : "=r"(__ptr) : "0"((typeof(*((worker))) *)((worker)))); (typeof((typeof(*((worker))) *)((worker)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); })->work) &wg->static_identity.lock &handshake->lock rcu_read_lock_bh &table->lock#2 irq_context: 0 (wq_completion)wg-kex-wg2#4 (work_completion)(&({ do { const void *__vpp_verify = (typeof((worker) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __asm__ ("" : "=r"(__ptr) : "0"((typeof(*((worker))) *)((worker)))); (typeof((typeof(*((worker))) *)((worker)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); })->work) &wg->static_identity.lock &handshake->lock irq_context: 0 (wq_completion)wg-kex-wg1#4 (work_completion)(&({ do { const void *__vpp_verify = (typeof((worker) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __asm__ ("" : "=r"(__ptr) : "0"((typeof(*((worker))) *)((worker)))); (typeof((typeof(*((worker))) *)((worker)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); })->work) &cookie->lock irq_context: 0 (wq_completion)wg-kex-wg2#4 (work_completion)(&({ do { const void *__vpp_verify = (typeof((worker) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __asm__ ("" : "=r"(__ptr) : "0"((typeof(*((worker))) *)((worker)))); (typeof((typeof(*((worker))) *)((worker)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); })->work) &peer->endpoint_lock irq_context: 0 (wq_completion)wg-kex-wg1#4 (work_completion)(&({ do { const void *__vpp_verify = (typeof((worker) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __asm__ ("" : "=r"(__ptr) : "0"((typeof(*((worker))) *)((worker)))); (typeof((typeof(*((worker))) *)((worker)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); })->work) &cookie->lock irq_context: 0 (wq_completion)wg-kex-wg2#4 (work_completion)(&({ do { const void *__vpp_verify = (typeof((worker) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __asm__ ("" : "=r"(__ptr) : "0"((typeof(*((worker))) *)((worker)))); (typeof((typeof(*((worker))) *)((worker)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); })->work) &handshake->lock irq_context: 0 (wq_completion)wg-kex-wg2#4 (work_completion)(&({ do { const void *__vpp_verify = (typeof((worker) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __asm__ ("" : "=r"(__ptr) : "0"((typeof(*((worker))) *)((worker)))); (typeof((typeof(*((worker))) *)((worker)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); })->work) &handshake->lock fs_reclaim irq_context: 0 (wq_completion)wg-kex-wg1#4 (work_completion)(&({ do { const void *__vpp_verify = (typeof((worker) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __asm__ ("" : "=r"(__ptr) : "0"((typeof(*((worker))) *)((worker)))); (typeof((typeof(*((worker))) *)((worker)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); })->work) &peer->endpoint_lock irq_context: 0 (wq_completion)wg-kex-wg2#4 (work_completion)(&({ do { const void *__vpp_verify = (typeof((worker) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __asm__ ("" : "=r"(__ptr) : "0"((typeof(*((worker))) *)((worker)))); (typeof((typeof(*((worker))) *)((worker)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); })->work) &handshake->lock fs_reclaim mmu_notifier_invalidate_range_start irq_context: 0 (wq_completion)wg-kex-wg1#4 (work_completion)(&({ do { const void *__vpp_verify = (typeof((worker) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __asm__ ("" : "=r"(__ptr) : "0"((typeof(*((worker))) *)((worker)))); (typeof((typeof(*((worker))) *)((worker)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); })->work) &peer->endpoint_lock rcu_read_lock_bh &obj_hash[i].lock irq_context: 0 (wq_completion)wg-kex-wg2#4 (work_completion)(&({ do { const void *__vpp_verify = (typeof((worker) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __asm__ ("" : "=r"(__ptr) : "0"((typeof(*((worker))) *)((worker)))); (typeof((typeof(*((worker))) *)((worker)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); })->work) &handshake->lock &c->lock irq_context: 0 (wq_completion)wg-kex-wg1#4 (work_completion)(&({ do { const void *__vpp_verify = (typeof((worker) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __asm__ ("" : "=r"(__ptr) : "0"((typeof(*((worker))) *)((worker)))); (typeof((typeof(*((worker))) *)((worker)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); })->work) &peer->endpoint_lock rcu_read_lock_bh batched_entropy_u32.lock irq_context: 0 (wq_completion)wg-kex-wg2#4 (work_completion)(&({ do { const void *__vpp_verify = (typeof((worker) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __asm__ ("" : "=r"(__ptr) : "0"((typeof(*((worker))) *)((worker)))); (typeof((typeof(*((worker))) *)((worker)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); })->work) &handshake->lock &____s->seqcount#2 irq_context: 0 (wq_completion)wg-kex-wg1#4 (work_completion)(&({ do { const void *__vpp_verify = (typeof((worker) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __asm__ ("" : "=r"(__ptr) : "0"((typeof(*((worker))) *)((worker)))); (typeof((typeof(*((worker))) *)((worker)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); })->work) &peer->endpoint_lock rcu_read_lock_bh rcu_read_lock &____s->seqcount#7 irq_context: 0 (wq_completion)wg-kex-wg2#4 (work_completion)(&({ do { const void *__vpp_verify = (typeof((worker) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __asm__ ("" : "=r"(__ptr) : "0"((typeof(*((worker))) *)((worker)))); (typeof((typeof(*((worker))) *)((worker)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); })->work) &handshake->lock &____s->seqcount irq_context: 0 (wq_completion)wg-kex-wg1#4 (work_completion)(&({ do { const void *__vpp_verify = (typeof((worker) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __asm__ ("" : "=r"(__ptr) : "0"((typeof(*((worker))) *)((worker)))); (typeof((typeof(*((worker))) *)((worker)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); })->work) &peer->endpoint_lock rcu_read_lock_bh rcu_read_lock &____s->seqcount#10 irq_context: 0 (wq_completion)wg-kex-wg1#4 (work_completion)(&({ do { const void *__vpp_verify = (typeof((worker) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __asm__ ("" : "=r"(__ptr) : "0"((typeof(*((worker))) *)((worker)))); (typeof((typeof(*((worker))) *)((worker)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); })->work) &peer->endpoint_lock rcu_read_lock_bh rcu_read_lock rcu_read_lock_bh tk_core.seq.seqcount irq_context: 0 (wq_completion)wg-kex-wg2#4 (work_completion)(&({ do { const void *__vpp_verify = (typeof((worker) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __asm__ ("" : "=r"(__ptr) : "0"((typeof(*((worker))) *)((worker)))); (typeof((typeof(*((worker))) *)((worker)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); })->work) &handshake->lock tk_core.seq.seqcount irq_context: 0 (wq_completion)wg-kex-wg1#4 (work_completion)(&({ do { const void *__vpp_verify = (typeof((worker) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __asm__ ("" : "=r"(__ptr) : "0"((typeof(*((worker))) *)((worker)))); (typeof((typeof(*((worker))) *)((worker)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); })->work) &peer->endpoint_lock rcu_read_lock_bh rcu_read_lock rcu_read_lock_bh &list->lock#5 irq_context: 0 (wq_completion)wg-kex-wg2#4 (work_completion)(&({ do { const void *__vpp_verify = (typeof((worker) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __asm__ ("" : "=r"(__ptr) : "0"((typeof(*((worker))) *)((worker)))); (typeof((typeof(*((worker))) *)((worker)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); })->work) &handshake->lock rcu_read_lock_bh &peer->keypairs.keypair_update_lock irq_context: 0 (wq_completion)wg-kex-wg2#4 (work_completion)(&({ do { const void *__vpp_verify = (typeof((worker) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __asm__ ("" : "=r"(__ptr) : "0"((typeof(*((worker))) *)((worker)))); (typeof((typeof(*((worker))) *)((worker)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); })->work) &handshake->lock rcu_read_lock_bh &table->lock#2 irq_context: 0 (wq_completion)wg-kex-wg2#4 (work_completion)(&({ do { const void *__vpp_verify = (typeof((worker) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __asm__ ("" : "=r"(__ptr) : "0"((typeof(*((worker))) *)((worker)))); (typeof((typeof(*((worker))) *)((worker)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); })->work) rcu_read_lock_bh &obj_hash[i].lock irq_context: 0 (wq_completion)wg-kex-wg2#4 (work_completion)(&({ do { const void *__vpp_verify = (typeof((worker) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __asm__ ("" : "=r"(__ptr) : "0"((typeof(*((worker))) *)((worker)))); (typeof((typeof(*((worker))) *)((worker)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); })->work) rcu_read_lock_bh &base->lock irq_context: 0 (wq_completion)wg-kex-wg2#4 (work_completion)(&({ do { const void *__vpp_verify = (typeof((worker) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __asm__ ("" : "=r"(__ptr) : "0"((typeof(*((worker))) *)((worker)))); (typeof((typeof(*((worker))) *)((worker)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); })->work) rcu_read_lock_bh &base->lock &obj_hash[i].lock irq_context: 0 (wq_completion)wg-kex-wg2#4 (work_completion)(&({ do { const void *__vpp_verify = (typeof((worker) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __asm__ ("" : "=r"(__ptr) : "0"((typeof(*((worker))) *)((worker)))); (typeof((typeof(*((worker))) *)((worker)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); })->work) &obj_hash[i].lock irq_context: 0 (wq_completion)wg-kex-wg2#4 (work_completion)(&({ do { const void *__vpp_verify = (typeof((worker) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __asm__ ("" : "=r"(__ptr) : "0"((typeof(*((worker))) *)((worker)))); (typeof((typeof(*((worker))) *)((worker)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); })->work) &base->lock irq_context: 0 (wq_completion)wg-kex-wg2#4 (work_completion)(&({ do { const void *__vpp_verify = (typeof((worker) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __asm__ ("" : "=r"(__ptr) : "0"((typeof(*((worker))) *)((worker)))); (typeof((typeof(*((worker))) *)((worker)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); })->work) &base->lock &obj_hash[i].lock irq_context: 0 (wq_completion)wg-kex-wg2#4 (work_completion)(&({ do { const void *__vpp_verify = (typeof((worker) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __asm__ ("" : "=r"(__ptr) : "0"((typeof(*((worker))) *)((worker)))); (typeof((typeof(*((worker))) *)((worker)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); })->work) tk_core.seq.seqcount irq_context: 0 (wq_completion)wg-kex-wg2#4 (work_completion)(&({ do { const void *__vpp_verify = (typeof((worker) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __asm__ ("" : "=r"(__ptr) : "0"((typeof(*((worker))) *)((worker)))); (typeof((typeof(*((worker))) *)((worker)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); })->work) &list->lock#17 irq_context: 0 (wq_completion)wg-kex-wg2#4 (work_completion)(&({ do { const void *__vpp_verify = (typeof((worker) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __asm__ ("" : "=r"(__ptr) : "0"((typeof(*((worker))) *)((worker)))); (typeof((typeof(*((worker))) *)((worker)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); })->work) rcu_read_lock_bh &r->producer_lock#2 irq_context: 0 (wq_completion)wg-kex-wg2#4 (work_completion)(&({ do { const void *__vpp_verify = (typeof((worker) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __asm__ ("" : "=r"(__ptr) : "0"((typeof(*((worker))) *)((worker)))); (typeof((typeof(*((worker))) *)((worker)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); })->work) rcu_read_lock_bh rcu_read_lock &pool->lock irq_context: 0 (wq_completion)wg-kex-wg2#4 (work_completion)(&({ do { const void *__vpp_verify = (typeof((worker) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __asm__ ("" : "=r"(__ptr) : "0"((typeof(*((worker))) *)((worker)))); (typeof((typeof(*((worker))) *)((worker)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); })->work) rcu_read_lock_bh rcu_read_lock &pool->lock &obj_hash[i].lock irq_context: 0 (wq_completion)wg-kex-wg2#4 (work_completion)(&({ do { const void *__vpp_verify = (typeof((worker) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __asm__ ("" : "=r"(__ptr) : "0"((typeof(*((worker))) *)((worker)))); (typeof((typeof(*((worker))) *)((worker)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); })->work) rcu_read_lock_bh rcu_read_lock &pool->lock &p->pi_lock irq_context: 0 (wq_completion)wg-crypt-wg2#2 irq_context: 0 (wq_completion)wg-crypt-wg2#2 &rq->__lock irq_context: 0 rcu_read_lock rcu_read_lock_bh quarantine_lock irq_context: 0 (wq_completion)wg-crypt-wg2#2 (work_completion)(&({ do { const void *__vpp_verify = (typeof((worker) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __asm__ ("" : "=r"(__ptr) : "0"((typeof(*((worker))) *)((worker)))); (typeof((typeof(*((worker))) *)((worker)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); })->work) irq_context: 0 (wq_completion)wg-crypt-wg2#2 (work_completion)(&({ do { const void *__vpp_verify = (typeof((worker) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __asm__ ("" : "=r"(__ptr) : "0"((typeof(*((worker))) *)((worker)))); (typeof((typeof(*((worker))) *)((worker)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); })->work) &r->consumer_lock#2 irq_context: 0 (wq_completion)wg-crypt-wg2#2 (work_completion)(&({ do { const void *__vpp_verify = (typeof((worker) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __asm__ ("" : "=r"(__ptr) : "0"((typeof(*((worker))) *)((worker)))); (typeof((typeof(*((worker))) *)((worker)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); })->work) rcu_read_lock &pool->lock irq_context: 0 (wq_completion)wg-crypt-wg2#2 (work_completion)(&({ do { const void *__vpp_verify = (typeof((worker) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __asm__ ("" : "=r"(__ptr) : "0"((typeof(*((worker))) *)((worker)))); (typeof((typeof(*((worker))) *)((worker)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); })->work) rcu_read_lock &pool->lock &obj_hash[i].lock irq_context: 0 (wq_completion)wg-crypt-wg2#2 (work_completion)(&({ do { const void *__vpp_verify = (typeof((worker) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __asm__ ("" : "=r"(__ptr) : "0"((typeof(*((worker))) *)((worker)))); (typeof((typeof(*((worker))) *)((worker)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); })->work) rcu_read_lock &pool->lock &p->pi_lock irq_context: 0 (wq_completion)wg-crypt-wg2#2 (work_completion)(&peer->transmit_packet_work) irq_context: 0 (wq_completion)wg-crypt-wg2#2 (work_completion)(&peer->transmit_packet_work) rcu_read_lock_bh &obj_hash[i].lock irq_context: 0 (wq_completion)wg-crypt-wg2#2 (work_completion)(&peer->transmit_packet_work) &obj_hash[i].lock irq_context: 0 (wq_completion)wg-crypt-wg2#2 (work_completion)(&peer->transmit_packet_work) &peer->endpoint_lock irq_context: 0 (wq_completion)wg-crypt-wg2#2 (work_completion)(&peer->transmit_packet_work) &peer->endpoint_lock rcu_read_lock_bh rcu_read_lock &____s->seqcount#7 irq_context: 0 (wq_completion)wg-crypt-wg2#2 (work_completion)(&peer->transmit_packet_work) &peer->endpoint_lock rcu_read_lock_bh rcu_read_lock &____s->seqcount#10 irq_context: 0 (wq_completion)wg-crypt-wg2#2 (work_completion)(&peer->transmit_packet_work) &peer->endpoint_lock rcu_read_lock_bh rcu_read_lock rcu_read_lock_bh tk_core.seq.seqcount irq_context: 0 (wq_completion)wg-crypt-wg2#2 (work_completion)(&peer->transmit_packet_work) &peer->endpoint_lock rcu_read_lock_bh rcu_read_lock rcu_read_lock_bh &list->lock#5 irq_context: 0 (wq_completion)wg-crypt-wg2#2 (work_completion)(&peer->transmit_packet_work) rcu_read_lock_bh tk_core.seq.seqcount irq_context: 0 (wq_completion)wg-crypt-wg2#2 (work_completion)(&peer->transmit_packet_work) rcu_read_lock_bh &base->lock irq_context: 0 (wq_completion)wg-crypt-wg2#2 (work_completion)(&peer->transmit_packet_work) &peer->endpoint_lock rcu_read_lock_bh rcu_read_lock &dir->lock#2 irq_context: 0 (wq_completion)wg-crypt-wg2#2 (work_completion)(&peer->transmit_packet_work) &peer->endpoint_lock rcu_read_lock_bh batched_entropy_u32.lock irq_context: 0 (wq_completion)wg-crypt-wg0#2 (work_completion)(&peer->transmit_packet_work) rcu_read_lock_bh &base->lock irq_context: 0 (wq_completion)wg-crypt-wg0#2 (work_completion)(&peer->transmit_packet_work) batched_entropy_u8.lock irq_context: 0 (wq_completion)wg-crypt-wg0#2 (work_completion)(&peer->transmit_packet_work) rcu_read_lock_bh &base->lock &obj_hash[i].lock irq_context: 0 (wq_completion)wg-crypt-wg2#2 (work_completion)(&({ do { const void *__vpp_verify = (typeof((worker) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __asm__ ("" : "=r"(__ptr) : "0"((typeof(*((worker))) *)((worker)))); (typeof((typeof(*((worker))) *)((worker)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); })->work) tk_core.seq.seqcount irq_context: 0 (wq_completion)ipv6_addrconf (work_completion)(&(&ifa->dad_work)->work) rtnl_mutex &idev->mc_lock rcu_read_lock rcu_read_lock rcu_read_lock_bh rcu_read_lock_bh rcu_read_lock tk_core.seq.seqcount irq_context: 0 (wq_completion)ipv6_addrconf (work_completion)(&(&ifa->dad_work)->work) rtnl_mutex &idev->mc_lock rcu_read_lock rcu_read_lock rcu_read_lock_bh rcu_read_lock_bh rcu_read_lock &list->lock#5 irq_context: 0 cb_lock genl_mutex rcu_read_lock &pool->lock/1 &p->pi_lock &rq->__lock irq_context: 0 cb_lock genl_mutex rcu_read_lock &pool->lock/1 &p->pi_lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 cb_lock genl_mutex fill_pool_map-wait-type-override pool_lock#2 irq_context: 0 cb_lock genl_mutex fill_pool_map-wait-type-override pool_lock irq_context: 0 cb_lock genl_mutex fill_pool_map-wait-type-override &c->lock irq_context: 0 cb_lock genl_mutex fill_pool_map-wait-type-override &____s->seqcount#2 irq_context: 0 cb_lock genl_mutex fill_pool_map-wait-type-override &____s->seqcount irq_context: 0 cb_lock genl_mutex &pcp->lock &zone->lock &____s->seqcount irq_context: 0 (wq_completion)ipv6_addrconf (work_completion)(&(&ifa->dad_work)->work) rtnl_mutex &idev->mc_lock rcu_read_lock rcu_read_lock rcu_read_lock_bh rcu_read_lock_bh &obj_hash[i].lock irq_context: 0 (wq_completion)ipv6_addrconf (work_completion)(&(&ifa->dad_work)->work) rtnl_mutex &idev->mc_lock rcu_read_lock rcu_read_lock rcu_read_lock_bh rcu_read_lock_bh pool_lock#2 irq_context: 0 cb_lock genl_mutex rtnl_mutex &rdev->wiphy.mtx &____s->seqcount#2 irq_context: 0 cb_lock genl_mutex rtnl_mutex &rdev->wiphy.mtx &sb->s_type->i_mutex_key#3 &____s->seqcount#2 irq_context: 0 cb_lock genl_mutex rfkill_global_mutex &____s->seqcount#2 irq_context: 0 cb_lock genl_mutex &sb->s_type->i_mutex_key#3 &____s->seqcount#2 irq_context: softirq rcu_read_lock rcu_read_lock &n->list_lock irq_context: softirq rcu_read_lock rcu_read_lock &n->list_lock &c->lock irq_context: 0 rtnl_mutex &ndev->lock &tb->tb6_lock quarantine_lock irq_context: softirq (&ndev->rs_timer) rcu_read_lock rcu_read_lock pool_lock#2 irq_context: softirq (&ndev->rs_timer) rcu_read_lock rcu_read_lock &obj_hash[i].lock irq_context: softirq (&ndev->rs_timer) rcu_read_lock rcu_read_lock &dir->lock#2 irq_context: softirq (&ndev->rs_timer) rcu_read_lock rcu_read_lock &tbl->lock irq_context: softirq (&ndev->rs_timer) rcu_read_lock rcu_read_lock &n->lock irq_context: softirq (&ndev->rs_timer) rcu_read_lock rcu_read_lock rcu_read_lock_bh _xmit_ETHER#2 irq_context: softirq (&ndev->rs_timer) rcu_read_lock rcu_read_lock rcu_read_lock_bh _xmit_ETHER#2 &obj_hash[i].lock irq_context: softirq (&ndev->rs_timer) rcu_read_lock rcu_read_lock rcu_read_lock_bh _xmit_ETHER#2 pool_lock#2 irq_context: 0 (wq_completion)ipv6_addrconf (work_completion)(&(&ifa->dad_work)->work) rtnl_mutex &tb->tb6_lock &pcp->lock &zone->lock &____s->seqcount irq_context: 0 (wq_completion)ipv6_addrconf (work_completion)(&(&ifa->dad_work)->work) rtnl_mutex &idev->mc_lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)mld (work_completion)(&(&idev->mc_dad_work)->work) &idev->mc_lock rcu_read_lock rcu_read_lock rcu_read_lock_bh &hsr->seqnr_lock rcu_read_lock &c->lock irq_context: softirq (&ndev->rs_timer) rcu_read_lock rcu_read_lock &____s->seqcount#10 irq_context: 0 rtnl_mutex lweventlist_lock &____s->seqcount#2 irq_context: 0 (wq_completion)bat_events (work_completion)(&(&forw_packet_aggr->delayed_work)->work) &hard_iface->bat_iv.ogm_buff_mutex rcu_read_lock &c->lock irq_context: 0 (wq_completion)bat_events (work_completion)(&(&forw_packet_aggr->delayed_work)->work) &hard_iface->bat_iv.ogm_buff_mutex rcu_read_lock &obj_hash[i].lock pool_lock irq_context: 0 (wq_completion)bat_events (work_completion)(&(&forw_packet_aggr->delayed_work)->work) &hard_iface->bat_iv.ogm_buff_mutex rcu_read_lock &bat_priv->forw_bat_list_lock fill_pool_map-wait-type-override pool_lock#2 irq_context: 0 (wq_completion)bat_events (work_completion)(&(&forw_packet_aggr->delayed_work)->work) &hard_iface->bat_iv.ogm_buff_mutex rcu_read_lock &bat_priv->forw_bat_list_lock fill_pool_map-wait-type-override pool_lock irq_context: 0 cb_lock genl_mutex rtnl_mutex &rdev->wiphy.mtx &sb->s_type->i_mutex_key#3 &n->list_lock irq_context: 0 cb_lock genl_mutex rtnl_mutex &rdev->wiphy.mtx fill_pool_map-wait-type-override pool_lock#2 irq_context: 0 cb_lock genl_mutex rtnl_mutex &rdev->wiphy.mtx fill_pool_map-wait-type-override pool_lock irq_context: 0 (wq_completion)mld (work_completion)(&(&idev->mc_dad_work)->work) &idev->mc_lock rcu_read_lock rcu_read_lock rcu_read_lock_bh fill_pool_map-wait-type-override pool_lock#2 irq_context: 0 (wq_completion)mld (work_completion)(&(&idev->mc_dad_work)->work) &idev->mc_lock rcu_read_lock rcu_read_lock rcu_read_lock_bh fill_pool_map-wait-type-override &c->lock irq_context: 0 (wq_completion)mld (work_completion)(&(&idev->mc_dad_work)->work) &idev->mc_lock rcu_read_lock rcu_read_lock rcu_read_lock_bh fill_pool_map-wait-type-override &____s->seqcount#2 irq_context: 0 (wq_completion)mld (work_completion)(&(&idev->mc_dad_work)->work) &idev->mc_lock rcu_read_lock rcu_read_lock rcu_read_lock_bh fill_pool_map-wait-type-override &____s->seqcount irq_context: 0 (wq_completion)mld (work_completion)(&(&idev->mc_dad_work)->work) &idev->mc_lock rcu_read_lock rcu_read_lock rcu_read_lock_bh fill_pool_map-wait-type-override pool_lock irq_context: 0 (wq_completion)ipv6_addrconf (work_completion)(&(&ifa->dad_work)->work) rtnl_mutex &ndev->lock &ifa->lock fill_pool_map-wait-type-override &____s->seqcount#2 irq_context: 0 rtnl_mutex dev_addr_sem rcu_read_lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)bat_events (work_completion)(&(&bat_priv->mcast.work)->work) rcu_read_lock &____s->seqcount#2 irq_context: 0 (wq_completion)mld (work_completion)(&(&idev->mc_dad_work)->work) &idev->mc_lock remove_cache_srcu rcu_read_lock rcu_node_0 irq_context: 0 (wq_completion)mld (work_completion)(&(&idev->mc_dad_work)->work) &idev->mc_lock remove_cache_srcu rcu_read_lock &rq->__lock irq_context: 0 (wq_completion)mld (work_completion)(&(&idev->mc_dad_work)->work) &idev->mc_lock remove_cache_srcu rcu_read_lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)events (work_completion)(&(&nsim_dev->trap_data->trap_report_dw)->work) &devlink->lock_key &rq->__lock irq_context: 0 (wq_completion)mld (work_completion)(&(&idev->mc_dad_work)->work) &idev->mc_lock rcu_read_lock rcu_read_lock rcu_read_lock_bh rcu_read_lock_bh &obj_hash[i].lock irq_context: 0 (wq_completion)mld (work_completion)(&(&idev->mc_dad_work)->work) &idev->mc_lock rcu_read_lock rcu_read_lock rcu_read_lock_bh rcu_read_lock_bh pool_lock#2 irq_context: 0 (wq_completion)mld (work_completion)(&(&idev->mc_dad_work)->work) &idev->mc_lock rcu_read_lock rcu_read_lock rcu_read_lock_bh rcu_read_lock_bh rcu_read_lock tk_core.seq.seqcount irq_context: 0 (wq_completion)mld (work_completion)(&(&idev->mc_dad_work)->work) &idev->mc_lock rcu_read_lock rcu_read_lock rcu_read_lock_bh rcu_read_lock_bh rcu_read_lock &list->lock#5 irq_context: 0 cb_lock rtnl_mutex &rdev->wiphy.mtx &sb->s_type->i_mutex_key#3 fill_pool_map-wait-type-override pool_lock#2 irq_context: 0 cb_lock rtnl_mutex &rdev->wiphy.mtx &sb->s_type->i_mutex_key#3 fill_pool_map-wait-type-override pool_lock irq_context: 0 cb_lock rtnl_mutex &rdev->wiphy.mtx &sb->s_type->i_mutex_key#3 fill_pool_map-wait-type-override &____s->seqcount irq_context: 0 cb_lock rtnl_mutex &rdev->wiphy.mtx &sb->s_type->i_mutex_key#3 fill_pool_map-wait-type-override rcu_read_lock pool_lock#2 irq_context: 0 cb_lock rtnl_mutex &rdev->wiphy.mtx &sb->s_type->i_mutex_key#3 fill_pool_map-wait-type-override &obj_hash[i].lock irq_context: 0 cb_lock rtnl_mutex &rdev->wiphy.mtx &obj_hash[i].lock pool_lock irq_context: 0 cb_lock rtnl_mutex &rdev->wiphy.mtx &sb->s_type->i_mutex_key#3 &____s->seqcount#2 irq_context: 0 cb_lock rtnl_mutex &rdev->wiphy.mtx &sb->s_type->i_mutex_key#3 &rq->__lock irq_context: 0 cb_lock rtnl_mutex &rdev->wiphy.mtx &sb->s_type->i_mutex_key#3 &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 cb_lock rtnl_mutex &rdev->wiphy.mtx &sb->s_type->i_mutex_key#3 &n->list_lock irq_context: 0 (wq_completion)mld (work_completion)(&(&idev->mc_dad_work)->work) &idev->mc_lock rcu_read_lock rcu_read_lock rcu_read_lock_bh rcu_read_lock &c->lock irq_context: 0 (wq_completion)events (work_completion)(&data->fib_event_work) &data->fib_lock &pool->lock &p->pi_lock &rq->__lock &cfs_rq->removed.lock irq_context: softirq rcu_read_lock hwsim_radio_lock &n->list_lock irq_context: softirq rcu_read_lock hwsim_radio_lock &n->list_lock &c->lock irq_context: softirq (&ndev->rs_timer) &n->list_lock irq_context: softirq (&ndev->rs_timer) &n->list_lock &c->lock irq_context: softirq (&ndev->rs_timer) rcu_read_lock rcu_read_lock rcu_read_lock_bh _xmit_TUNNEL6#2 irq_context: softirq (&ndev->rs_timer) rcu_read_lock rcu_read_lock rcu_read_lock_bh _xmit_TUNNEL6#2 rcu_read_lock &ndev->lock irq_context: softirq (&ndev->rs_timer) rcu_read_lock rcu_read_lock rcu_read_lock_bh _xmit_TUNNEL6#2 &obj_hash[i].lock irq_context: softirq (&ndev->rs_timer) rcu_read_lock rcu_read_lock rcu_read_lock_bh _xmit_TUNNEL6#2 pool_lock#2 irq_context: softirq (&ndev->rs_timer) &____s->seqcount#2 irq_context: 0 (wq_completion)events (linkwatch_work).work rtnl_mutex remove_cache_srcu rcu_read_lock rcu_node_0 irq_context: 0 (wq_completion)events (linkwatch_work).work rtnl_mutex remove_cache_srcu rcu_read_lock &rq->__lock irq_context: 0 (wq_completion)events (linkwatch_work).work rtnl_mutex remove_cache_srcu rcu_read_lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: softirq rcu_read_lock rcu_read_lock rcu_read_lock &pcp->lock &zone->lock irq_context: 0 (wq_completion)phy5 irq_context: 0 (wq_completion)phy5 (work_completion)(&local->reconfig_filter) irq_context: 0 (wq_completion)phy5 (work_completion)(&local->reconfig_filter) &local->filter_lock irq_context: 0 cb_lock &____s->seqcount#2 irq_context: 0 cb_lock genl_mutex lock kernfs_idr_lock pool_lock#2 irq_context: 0 sb_writers#5 &type->i_mutex_dir_key#5/1 fill_pool_map-wait-type-override &c->lock irq_context: 0 sb_writers#5 &type->i_mutex_dir_key#5/1 fill_pool_map-wait-type-override &____s->seqcount#2 irq_context: 0 sb_writers#5 &type->i_mutex_dir_key#5/1 fill_pool_map-wait-type-override &____s->seqcount irq_context: 0 (wq_completion)ipv6_addrconf (work_completion)(&(&ifa->dad_work)->work) rtnl_mutex &idev->mc_lock rcu_read_lock &rcu_state.gp_wq irq_context: 0 (wq_completion)ipv6_addrconf (work_completion)(&(&ifa->dad_work)->work) rtnl_mutex &idev->mc_lock rcu_read_lock &rcu_state.gp_wq &p->pi_lock irq_context: 0 (wq_completion)ipv6_addrconf (work_completion)(&(&ifa->dad_work)->work) rtnl_mutex &idev->mc_lock rcu_read_lock &rcu_state.gp_wq &p->pi_lock &rq->__lock irq_context: 0 (wq_completion)ipv6_addrconf (work_completion)(&(&ifa->dad_work)->work) rtnl_mutex &idev->mc_lock rcu_read_lock &rcu_state.gp_wq &p->pi_lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 rtnl_mutex rcu_read_lock console_owner_lock irq_context: 0 rtnl_mutex rcu_read_lock console_owner irq_context: 0 (wq_completion)wg-crypt-wg0 (work_completion)(&({ do { const void *__vpp_verify = (typeof((worker) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __asm__ ("" : "=r"(__ptr) : "0"((typeof(*((worker))) *)((worker)))); (typeof((typeof(*((worker))) *)((worker)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); })->work) &rq->__lock irq_context: softirq &(&forw_packet_aggr->delayed_work)->timer rcu_read_lock &pool->lock/1 &p->pi_lock &cfs_rq->removed.lock irq_context: 0 (wq_completion)wg-kex-wg0#5 irq_context: 0 (wq_completion)wg-kex-wg0#5 (work_completion)(&peer->transmit_handshake_work) irq_context: 0 (wq_completion)wg-kex-wg0#5 (work_completion)(&peer->transmit_handshake_work) tk_core.seq.seqcount irq_context: 0 (wq_completion)wg-kex-wg0#5 (work_completion)(&peer->transmit_handshake_work) &wg->static_identity.lock irq_context: 0 (wq_completion)wg-kex-wg0#5 (work_completion)(&peer->transmit_handshake_work) &wg->static_identity.lock &handshake->lock irq_context: 0 (wq_completion)wg-kex-wg0#5 (work_completion)(&peer->transmit_handshake_work) &wg->static_identity.lock &handshake->lock crngs.lock irq_context: 0 (wq_completion)wg-kex-wg0#5 (work_completion)(&peer->transmit_handshake_work) &wg->static_identity.lock &handshake->lock tk_core.seq.seqcount irq_context: 0 (wq_completion)wg-kex-wg0#5 (work_completion)(&peer->transmit_handshake_work) &wg->static_identity.lock &handshake->lock &table->lock#2 irq_context: 0 (wq_completion)wg-kex-wg0#5 (work_completion)(&peer->transmit_handshake_work) &wg->static_identity.lock &handshake->lock rcu_read_lock_bh batched_entropy_u32.lock irq_context: 0 (wq_completion)wg-kex-wg0#5 (work_completion)(&peer->transmit_handshake_work) &wg->static_identity.lock &handshake->lock rcu_read_lock_bh &table->lock#2 irq_context: 0 (wq_completion)wg-kex-wg0#5 (work_completion)(&peer->transmit_handshake_work) &cookie->lock irq_context: 0 (wq_completion)wg-kex-wg0#5 (work_completion)(&peer->transmit_handshake_work) &cookie->lock irq_context: 0 (wq_completion)wg-kex-wg0#5 (work_completion)(&peer->transmit_handshake_work) rcu_read_lock_bh &obj_hash[i].lock irq_context: 0 (wq_completion)wg-kex-wg0#5 (work_completion)(&peer->transmit_handshake_work) rcu_read_lock_bh &base->lock irq_context: 0 (wq_completion)wg-kex-wg0#5 (work_completion)(&peer->transmit_handshake_work) rcu_read_lock_bh &base->lock &obj_hash[i].lock irq_context: 0 (wq_completion)wg-kex-wg0#5 (work_completion)(&peer->transmit_handshake_work) &obj_hash[i].lock irq_context: 0 (wq_completion)wg-kex-wg0#5 (work_completion)(&peer->transmit_handshake_work) &c->lock irq_context: 0 (wq_completion)wg-kex-wg0#5 (work_completion)(&peer->transmit_handshake_work) &peer->endpoint_lock irq_context: 0 (wq_completion)wg-kex-wg0#5 (work_completion)(&peer->transmit_handshake_work) &peer->endpoint_lock rcu_read_lock_bh rcu_read_lock rcu_read_lock rcu_read_lock &dir->lock#2 irq_context: 0 (wq_completion)wg-kex-wg0#5 (work_completion)(&peer->transmit_handshake_work) &peer->endpoint_lock rcu_read_lock_bh rcu_read_lock &____s->seqcount#7 irq_context: 0 (wq_completion)wg-kex-wg0#5 (work_completion)(&peer->transmit_handshake_work) &peer->endpoint_lock rcu_read_lock_bh rcu_read_lock &nf_conntrack_locks[i] irq_context: 0 (wq_completion)wg-kex-wg0#5 (work_completion)(&peer->transmit_handshake_work) &peer->endpoint_lock rcu_read_lock_bh rcu_read_lock &nf_conntrack_locks[i] &nf_conntrack_locks[i]/1 irq_context: 0 (wq_completion)wg-kex-wg0#5 (work_completion)(&peer->transmit_handshake_work) &peer->endpoint_lock rcu_read_lock_bh rcu_read_lock &nf_conntrack_locks[i] &nf_conntrack_locks[i]/1 batched_entropy_u8.lock irq_context: 0 (wq_completion)wg-kex-wg0#5 (work_completion)(&peer->transmit_handshake_work) &peer->endpoint_lock rcu_read_lock_bh rcu_read_lock &nf_conntrack_locks[i]/1 irq_context: 0 (wq_completion)wg-kex-wg0#5 (work_completion)(&peer->transmit_handshake_work) &peer->endpoint_lock rcu_read_lock_bh rcu_read_lock &obj_hash[i].lock irq_context: 0 (wq_completion)wg-kex-wg0#5 (work_completion)(&peer->transmit_handshake_work) &peer->endpoint_lock rcu_read_lock_bh rcu_read_lock &dir->lock#2 irq_context: 0 (wq_completion)wg-kex-wg0#5 (work_completion)(&peer->transmit_handshake_work) &peer->endpoint_lock rcu_read_lock_bh rcu_read_lock &tbl->lock irq_context: 0 (wq_completion)wg-kex-wg0#5 (work_completion)(&peer->transmit_handshake_work) &peer->endpoint_lock rcu_read_lock_bh rcu_read_lock &n->lock irq_context: 0 (wq_completion)wg-kex-wg0#5 (work_completion)(&peer->transmit_handshake_work) &peer->endpoint_lock rcu_read_lock_bh rcu_read_lock &____s->seqcount#9 irq_context: 0 (wq_completion)wg-kex-wg0#5 (work_completion)(&peer->transmit_handshake_work) &peer->endpoint_lock rcu_read_lock_bh crngs.lock irq_context: 0 (wq_completion)wg-kex-wg0#5 (work_completion)(&peer->transmit_handshake_work) &peer->endpoint_lock rcu_read_lock_bh rcu_read_lock rcu_read_lock_bh tk_core.seq.seqcount irq_context: 0 (wq_completion)wg-kex-wg0#5 (work_completion)(&peer->transmit_handshake_work) &peer->endpoint_lock rcu_read_lock_bh batched_entropy_u32.lock irq_context: 0 (wq_completion)wg-kex-wg0#5 (work_completion)(&peer->transmit_handshake_work) &peer->endpoint_lock rcu_read_lock_bh rcu_read_lock rcu_read_lock_bh &list->lock#5 irq_context: softirq rcu_read_lock rcu_read_lock k-slock-AF_INET &____s->seqcount#2 irq_context: softirq rcu_read_lock rcu_read_lock k-slock-AF_INET &____s->seqcount irq_context: 0 (wq_completion)wg-kex-wg0#5 (work_completion)(&peer->transmit_handshake_work) batched_entropy_u8.lock irq_context: 0 (wq_completion)bat_events (work_completion)(&(&forw_packet_aggr->delayed_work)->work) &c->lock irq_context: 0 rtnl_mutex &wg->device_update_lock &n->list_lock irq_context: 0 rtnl_mutex &wg->device_update_lock &n->list_lock &c->lock irq_context: 0 rtnl_mutex &wg->device_update_lock &rq->__lock irq_context: 0 (wq_completion)wg-kex-wg1#5 irq_context: 0 (wq_completion)wg-kex-wg1#5 (work_completion)(&peer->transmit_handshake_work) irq_context: 0 (wq_completion)wg-kex-wg1#5 (work_completion)(&peer->transmit_handshake_work) tk_core.seq.seqcount irq_context: 0 (wq_completion)wg-kex-wg1#5 (work_completion)(&peer->transmit_handshake_work) &wg->static_identity.lock irq_context: 0 (wq_completion)wg-kex-wg1#5 (work_completion)(&peer->transmit_handshake_work) &wg->static_identity.lock &handshake->lock irq_context: 0 (wq_completion)wg-kex-wg1#5 (work_completion)(&peer->transmit_handshake_work) &wg->static_identity.lock &handshake->lock crngs.lock irq_context: 0 (wq_completion)wg-kex-wg1#5 (work_completion)(&peer->transmit_handshake_work) &wg->static_identity.lock &handshake->lock tk_core.seq.seqcount irq_context: 0 (wq_completion)wg-kex-wg1#5 (work_completion)(&peer->transmit_handshake_work) &wg->static_identity.lock &handshake->lock &table->lock#2 irq_context: 0 (wq_completion)wg-kex-wg1#5 (work_completion)(&peer->transmit_handshake_work) &wg->static_identity.lock &handshake->lock rcu_read_lock_bh batched_entropy_u32.lock irq_context: 0 (wq_completion)wg-kex-wg1#5 (work_completion)(&peer->transmit_handshake_work) &wg->static_identity.lock &handshake->lock rcu_read_lock_bh &table->lock#2 irq_context: 0 (wq_completion)wg-kex-wg1#5 (work_completion)(&peer->transmit_handshake_work) &cookie->lock irq_context: 0 (wq_completion)wg-kex-wg1#5 (work_completion)(&peer->transmit_handshake_work) &cookie->lock irq_context: 0 (wq_completion)wg-kex-wg1#5 (work_completion)(&peer->transmit_handshake_work) rcu_read_lock_bh &obj_hash[i].lock irq_context: 0 (wq_completion)wg-kex-wg1#5 (work_completion)(&peer->transmit_handshake_work) rcu_read_lock_bh &base->lock irq_context: 0 (wq_completion)wg-kex-wg1#5 (work_completion)(&peer->transmit_handshake_work) rcu_read_lock_bh &base->lock &obj_hash[i].lock irq_context: 0 (wq_completion)wg-kex-wg1#5 (work_completion)(&peer->transmit_handshake_work) &obj_hash[i].lock irq_context: 0 (wq_completion)wg-kex-wg1#5 (work_completion)(&peer->transmit_handshake_work) &c->lock irq_context: 0 (wq_completion)wg-kex-wg1#5 (work_completion)(&peer->transmit_handshake_work) &peer->endpoint_lock irq_context: 0 (wq_completion)wg-kex-wg1#5 (work_completion)(&peer->transmit_handshake_work) &peer->endpoint_lock rcu_read_lock_bh rcu_read_lock &dir->lock#2 irq_context: 0 (wq_completion)wg-kex-wg1#5 (work_completion)(&peer->transmit_handshake_work) &peer->endpoint_lock rcu_read_lock_bh rcu_read_lock rcu_read_lock rcu_read_lock &dir->lock#2 irq_context: 0 (wq_completion)wg-kex-wg1#5 (work_completion)(&peer->transmit_handshake_work) &peer->endpoint_lock rcu_read_lock_bh batched_entropy_u32.lock irq_context: 0 (wq_completion)wg-kex-wg1#5 (work_completion)(&peer->transmit_handshake_work) &peer->endpoint_lock rcu_read_lock_bh rcu_read_lock &____s->seqcount#7 irq_context: 0 (wq_completion)wg-kex-wg1#5 (work_completion)(&peer->transmit_handshake_work) &peer->endpoint_lock rcu_read_lock_bh rcu_read_lock &nf_conntrack_locks[i] irq_context: 0 (wq_completion)wg-kex-wg1#5 (work_completion)(&peer->transmit_handshake_work) &peer->endpoint_lock rcu_read_lock_bh rcu_read_lock &nf_conntrack_locks[i] &nf_conntrack_locks[i]/1 irq_context: 0 (wq_completion)wg-kex-wg1#5 (work_completion)(&peer->transmit_handshake_work) &peer->endpoint_lock rcu_read_lock_bh rcu_read_lock &nf_conntrack_locks[i] &nf_conntrack_locks[i]/1 batched_entropy_u8.lock irq_context: 0 (wq_completion)wg-kex-wg1#5 (work_completion)(&peer->transmit_handshake_work) &peer->endpoint_lock rcu_read_lock_bh rcu_read_lock &____s->seqcount#10 irq_context: 0 (wq_completion)wg-kex-wg1#5 (work_completion)(&peer->transmit_handshake_work) &peer->endpoint_lock rcu_read_lock_bh rcu_read_lock rcu_read_lock_bh tk_core.seq.seqcount irq_context: 0 (wq_completion)wg-kex-wg1#5 (work_completion)(&peer->transmit_handshake_work) &peer->endpoint_lock rcu_read_lock_bh rcu_read_lock rcu_read_lock_bh &list->lock#5 irq_context: 0 (wq_completion)wg-kex-wg1#5 (work_completion)(&peer->transmit_handshake_work) batched_entropy_u8.lock irq_context: 0 (wq_completion)wg-kex-wg0#6 irq_context: 0 (wq_completion)wg-kex-wg0#6 (work_completion)(&({ do { const void *__vpp_verify = (typeof((worker) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __asm__ ("" : "=r"(__ptr) : "0"((typeof(*((worker))) *)((worker)))); (typeof((typeof(*((worker))) *)((worker)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); })->work) irq_context: 0 (wq_completion)wg-kex-wg0#6 (work_completion)(&({ do { const void *__vpp_verify = (typeof((worker) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __asm__ ("" : "=r"(__ptr) : "0"((typeof(*((worker))) *)((worker)))); (typeof((typeof(*((worker))) *)((worker)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); })->work) &r->consumer_lock#2 irq_context: 0 (wq_completion)wg-kex-wg0#6 (work_completion)(&({ do { const void *__vpp_verify = (typeof((worker) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __asm__ ("" : "=r"(__ptr) : "0"((typeof(*((worker))) *)((worker)))); (typeof((typeof(*((worker))) *)((worker)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); })->work) &wg->static_identity.lock irq_context: 0 (wq_completion)wg-kex-wg0#6 (work_completion)(&({ do { const void *__vpp_verify = (typeof((worker) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __asm__ ("" : "=r"(__ptr) : "0"((typeof(*((worker))) *)((worker)))); (typeof((typeof(*((worker))) *)((worker)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); })->work) &wg->static_identity.lock &handshake->lock irq_context: 0 (wq_completion)wg-kex-wg0#6 (work_completion)(&({ do { const void *__vpp_verify = (typeof((worker) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __asm__ ("" : "=r"(__ptr) : "0"((typeof(*((worker))) *)((worker)))); (typeof((typeof(*((worker))) *)((worker)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); })->work) &wg->static_identity.lock &handshake->lock tk_core.seq.seqcount irq_context: 0 (wq_completion)wg-kex-wg0#6 (work_completion)(&({ do { const void *__vpp_verify = (typeof((worker) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __asm__ ("" : "=r"(__ptr) : "0"((typeof(*((worker))) *)((worker)))); (typeof((typeof(*((worker))) *)((worker)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); })->work) &wg->static_identity.lock &handshake->lock irq_context: 0 (wq_completion)wg-kex-wg0#6 (work_completion)(&({ do { const void *__vpp_verify = (typeof((worker) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __asm__ ("" : "=r"(__ptr) : "0"((typeof(*((worker))) *)((worker)))); (typeof((typeof(*((worker))) *)((worker)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); })->work) &wg->static_identity.lock &handshake->lock tk_core.seq.seqcount irq_context: 0 (wq_completion)wg-kex-wg0#6 (work_completion)(&({ do { const void *__vpp_verify = (typeof((worker) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __asm__ ("" : "=r"(__ptr) : "0"((typeof(*((worker))) *)((worker)))); (typeof((typeof(*((worker))) *)((worker)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); })->work) &peer->endpoint_lock irq_context: 0 (wq_completion)wg-kex-wg0#6 (work_completion)(&({ do { const void *__vpp_verify = (typeof((worker) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __asm__ ("" : "=r"(__ptr) : "0"((typeof(*((worker))) *)((worker)))); (typeof((typeof(*((worker))) *)((worker)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); })->work) tk_core.seq.seqcount irq_context: 0 (wq_completion)wg-kex-wg0#6 (work_completion)(&({ do { const void *__vpp_verify = (typeof((worker) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __asm__ ("" : "=r"(__ptr) : "0"((typeof(*((worker))) *)((worker)))); (typeof((typeof(*((worker))) *)((worker)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); })->work) &wg->static_identity.lock &handshake->lock crngs.lock irq_context: 0 (wq_completion)wg-kex-wg0#6 (work_completion)(&({ do { const void *__vpp_verify = (typeof((worker) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __asm__ ("" : "=r"(__ptr) : "0"((typeof(*((worker))) *)((worker)))); (typeof((typeof(*((worker))) *)((worker)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); })->work) &wg->static_identity.lock &handshake->lock &table->lock#2 irq_context: 0 (wq_completion)wg-kex-wg0#6 (work_completion)(&({ do { const void *__vpp_verify = (typeof((worker) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __asm__ ("" : "=r"(__ptr) : "0"((typeof(*((worker))) *)((worker)))); (typeof((typeof(*((worker))) *)((worker)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); })->work) &wg->static_identity.lock &handshake->lock rcu_read_lock_bh batched_entropy_u32.lock irq_context: 0 (wq_completion)wg-kex-wg0#6 (work_completion)(&({ do { const void *__vpp_verify = (typeof((worker) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __asm__ ("" : "=r"(__ptr) : "0"((typeof(*((worker))) *)((worker)))); (typeof((typeof(*((worker))) *)((worker)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); })->work) &wg->static_identity.lock &handshake->lock rcu_read_lock_bh &table->lock#2 irq_context: 0 (wq_completion)wg-kex-wg0#6 (work_completion)(&({ do { const void *__vpp_verify = (typeof((worker) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __asm__ ("" : "=r"(__ptr) : "0"((typeof(*((worker))) *)((worker)))); (typeof((typeof(*((worker))) *)((worker)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); })->work) &cookie->lock irq_context: 0 (wq_completion)wg-kex-wg0#6 (work_completion)(&({ do { const void *__vpp_verify = (typeof((worker) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __asm__ ("" : "=r"(__ptr) : "0"((typeof(*((worker))) *)((worker)))); (typeof((typeof(*((worker))) *)((worker)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); })->work) &cookie->lock irq_context: 0 (wq_completion)wg-kex-wg0#6 (work_completion)(&({ do { const void *__vpp_verify = (typeof((worker) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __asm__ ("" : "=r"(__ptr) : "0"((typeof(*((worker))) *)((worker)))); (typeof((typeof(*((worker))) *)((worker)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); })->work) &handshake->lock irq_context: 0 (wq_completion)wg-kex-wg0#6 (work_completion)(&({ do { const void *__vpp_verify = (typeof((worker) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __asm__ ("" : "=r"(__ptr) : "0"((typeof(*((worker))) *)((worker)))); (typeof((typeof(*((worker))) *)((worker)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); })->work) &handshake->lock fs_reclaim irq_context: 0 (wq_completion)wg-kex-wg0#6 (work_completion)(&({ do { const void *__vpp_verify = (typeof((worker) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __asm__ ("" : "=r"(__ptr) : "0"((typeof(*((worker))) *)((worker)))); (typeof((typeof(*((worker))) *)((worker)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); })->work) &handshake->lock fs_reclaim &rq->__lock irq_context: 0 cb_lock genl_mutex rtnl_mutex &rdev->wiphy.mtx fill_pool_map-wait-type-override &c->lock irq_context: 0 cb_lock genl_mutex rtnl_mutex &rdev->wiphy.mtx fill_pool_map-wait-type-override &____s->seqcount#2 irq_context: 0 cb_lock genl_mutex rtnl_mutex &rdev->wiphy.mtx fill_pool_map-wait-type-override &____s->seqcount irq_context: 0 cb_lock genl_mutex rtnl_mutex &rdev->wiphy.mtx &sb->s_type->i_mutex_key#3 rcu_read_lock rcu_node_0 irq_context: 0 cb_lock genl_mutex rtnl_mutex &rdev->wiphy.mtx &sb->s_type->i_mutex_key#3 rcu_read_lock &rq->__lock irq_context: 0 cb_lock genl_mutex rtnl_mutex &rdev->wiphy.mtx &idev->mc_lock fill_pool_map-wait-type-override pool_lock#2 irq_context: 0 cb_lock genl_mutex rtnl_mutex &rdev->wiphy.mtx &idev->mc_lock fill_pool_map-wait-type-override pool_lock irq_context: 0 (wq_completion)wg-kex-wg0#6 (work_completion)(&({ do { const void *__vpp_verify = (typeof((worker) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __asm__ ("" : "=r"(__ptr) : "0"((typeof(*((worker))) *)((worker)))); (typeof((typeof(*((worker))) *)((worker)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); })->work) &handshake->lock fs_reclaim mmu_notifier_invalidate_range_start irq_context: 0 (wq_completion)wg-kex-wg0#6 (work_completion)(&({ do { const void *__vpp_verify = (typeof((worker) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __asm__ ("" : "=r"(__ptr) : "0"((typeof(*((worker))) *)((worker)))); (typeof((typeof(*((worker))) *)((worker)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); })->work) &handshake->lock tk_core.seq.seqcount irq_context: 0 (wq_completion)wg-kex-wg0#6 (work_completion)(&({ do { const void *__vpp_verify = (typeof((worker) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __asm__ ("" : "=r"(__ptr) : "0"((typeof(*((worker))) *)((worker)))); (typeof((typeof(*((worker))) *)((worker)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); })->work) &handshake->lock rcu_read_lock_bh &peer->keypairs.keypair_update_lock irq_context: 0 (wq_completion)wg-kex-wg0#6 (work_completion)(&({ do { const void *__vpp_verify = (typeof((worker) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __asm__ ("" : "=r"(__ptr) : "0"((typeof(*((worker))) *)((worker)))); (typeof((typeof(*((worker))) *)((worker)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); })->work) &handshake->lock rcu_read_lock_bh &table->lock#2 irq_context: 0 (wq_completion)wg-kex-wg0#6 (work_completion)(&({ do { const void *__vpp_verify = (typeof((worker) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __asm__ ("" : "=r"(__ptr) : "0"((typeof(*((worker))) *)((worker)))); (typeof((typeof(*((worker))) *)((worker)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); })->work) rcu_read_lock_bh &obj_hash[i].lock irq_context: 0 (wq_completion)wg-kex-wg0#6 (work_completion)(&({ do { const void *__vpp_verify = (typeof((worker) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __asm__ ("" : "=r"(__ptr) : "0"((typeof(*((worker))) *)((worker)))); (typeof((typeof(*((worker))) *)((worker)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); })->work) rcu_read_lock_bh &base->lock irq_context: 0 (wq_completion)wg-kex-wg0#6 (work_completion)(&({ do { const void *__vpp_verify = (typeof((worker) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __asm__ ("" : "=r"(__ptr) : "0"((typeof(*((worker))) *)((worker)))); (typeof((typeof(*((worker))) *)((worker)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); })->work) rcu_read_lock_bh &base->lock &obj_hash[i].lock irq_context: 0 (wq_completion)wg-kex-wg0#6 (work_completion)(&({ do { const void *__vpp_verify = (typeof((worker) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __asm__ ("" : "=r"(__ptr) : "0"((typeof(*((worker))) *)((worker)))); (typeof((typeof(*((worker))) *)((worker)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); })->work) &obj_hash[i].lock irq_context: 0 (wq_completion)wg-kex-wg0#6 (work_completion)(&({ do { const void *__vpp_verify = (typeof((worker) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __asm__ ("" : "=r"(__ptr) : "0"((typeof(*((worker))) *)((worker)))); (typeof((typeof(*((worker))) *)((worker)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); })->work) &peer->endpoint_lock irq_context: 0 (wq_completion)wg-kex-wg0#6 (work_completion)(&({ do { const void *__vpp_verify = (typeof((worker) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __asm__ ("" : "=r"(__ptr) : "0"((typeof(*((worker))) *)((worker)))); (typeof((typeof(*((worker))) *)((worker)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); })->work) &peer->endpoint_lock rcu_read_lock_bh rcu_read_lock &____s->seqcount#7 irq_context: 0 (wq_completion)wg-kex-wg0#6 (work_completion)(&({ do { const void *__vpp_verify = (typeof((worker) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __asm__ ("" : "=r"(__ptr) : "0"((typeof(*((worker))) *)((worker)))); (typeof((typeof(*((worker))) *)((worker)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); })->work) &peer->endpoint_lock rcu_read_lock_bh rcu_read_lock &____s->seqcount#10 irq_context: 0 (wq_completion)wg-kex-wg0#6 (work_completion)(&({ do { const void *__vpp_verify = (typeof((worker) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __asm__ ("" : "=r"(__ptr) : "0"((typeof(*((worker))) *)((worker)))); (typeof((typeof(*((worker))) *)((worker)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); })->work) &peer->endpoint_lock rcu_read_lock_bh rcu_read_lock rcu_read_lock_bh tk_core.seq.seqcount irq_context: 0 (wq_completion)wg-kex-wg0#6 (work_completion)(&({ do { const void *__vpp_verify = (typeof((worker) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __asm__ ("" : "=r"(__ptr) : "0"((typeof(*((worker))) *)((worker)))); (typeof((typeof(*((worker))) *)((worker)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); })->work) &peer->endpoint_lock rcu_read_lock_bh rcu_read_lock rcu_read_lock_bh &list->lock#5 irq_context: 0 (wq_completion)wg-kex-wg0#6 (work_completion)(&({ do { const void *__vpp_verify = (typeof((worker) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __asm__ ("" : "=r"(__ptr) : "0"((typeof(*((worker))) *)((worker)))); (typeof((typeof(*((worker))) *)((worker)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); })->work) &rq->__lock irq_context: 0 (wq_completion)wg-kex-wg1#6 irq_context: 0 (wq_completion)wg-kex-wg1#6 (work_completion)(&({ do { const void *__vpp_verify = (typeof((worker) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __asm__ ("" : "=r"(__ptr) : "0"((typeof(*((worker))) *)((worker)))); (typeof((typeof(*((worker))) *)((worker)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); })->work) irq_context: 0 (wq_completion)wg-kex-wg1#6 (work_completion)(&({ do { const void *__vpp_verify = (typeof((worker) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __asm__ ("" : "=r"(__ptr) : "0"((typeof(*((worker))) *)((worker)))); (typeof((typeof(*((worker))) *)((worker)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); })->work) &r->consumer_lock#2 irq_context: 0 (wq_completion)wg-kex-wg1#6 (work_completion)(&({ do { const void *__vpp_verify = (typeof((worker) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __asm__ ("" : "=r"(__ptr) : "0"((typeof(*((worker))) *)((worker)))); (typeof((typeof(*((worker))) *)((worker)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); })->work) &wg->static_identity.lock irq_context: 0 (wq_completion)wg-kex-wg1#6 (work_completion)(&({ do { const void *__vpp_verify = (typeof((worker) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __asm__ ("" : "=r"(__ptr) : "0"((typeof(*((worker))) *)((worker)))); (typeof((typeof(*((worker))) *)((worker)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); })->work) &wg->static_identity.lock &handshake->lock irq_context: 0 cb_lock rtnl_mutex &rdev->wiphy.mtx &sb->s_type->i_mutex_key#3 fill_pool_map-wait-type-override &c->lock irq_context: 0 cb_lock rtnl_mutex &rdev->wiphy.mtx &sb->s_type->i_mutex_key#3 fill_pool_map-wait-type-override &____s->seqcount#2 irq_context: 0 (wq_completion)wg-kex-wg1#6 (work_completion)(&({ do { const void *__vpp_verify = (typeof((worker) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __asm__ ("" : "=r"(__ptr) : "0"((typeof(*((worker))) *)((worker)))); (typeof((typeof(*((worker))) *)((worker)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); })->work) &wg->static_identity.lock &handshake->lock irq_context: 0 (wq_completion)wg-kex-wg1#6 (work_completion)(&({ do { const void *__vpp_verify = (typeof((worker) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __asm__ ("" : "=r"(__ptr) : "0"((typeof(*((worker))) *)((worker)))); (typeof((typeof(*((worker))) *)((worker)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); })->work) &peer->endpoint_lock irq_context: 0 (wq_completion)wg-kex-wg1#6 (work_completion)(&({ do { const void *__vpp_verify = (typeof((worker) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __asm__ ("" : "=r"(__ptr) : "0"((typeof(*((worker))) *)((worker)))); (typeof((typeof(*((worker))) *)((worker)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); })->work) &handshake->lock irq_context: 0 (wq_completion)wg-kex-wg1#6 (work_completion)(&({ do { const void *__vpp_verify = (typeof((worker) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __asm__ ("" : "=r"(__ptr) : "0"((typeof(*((worker))) *)((worker)))); (typeof((typeof(*((worker))) *)((worker)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); })->work) &handshake->lock fs_reclaim irq_context: 0 (wq_completion)wg-kex-wg1#6 (work_completion)(&({ do { const void *__vpp_verify = (typeof((worker) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __asm__ ("" : "=r"(__ptr) : "0"((typeof(*((worker))) *)((worker)))); (typeof((typeof(*((worker))) *)((worker)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); })->work) &handshake->lock fs_reclaim mmu_notifier_invalidate_range_start irq_context: 0 (wq_completion)wg-kex-wg1#6 (work_completion)(&({ do { const void *__vpp_verify = (typeof((worker) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __asm__ ("" : "=r"(__ptr) : "0"((typeof(*((worker))) *)((worker)))); (typeof((typeof(*((worker))) *)((worker)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); })->work) &handshake->lock tk_core.seq.seqcount irq_context: 0 (wq_completion)wg-kex-wg1#6 (work_completion)(&({ do { const void *__vpp_verify = (typeof((worker) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __asm__ ("" : "=r"(__ptr) : "0"((typeof(*((worker))) *)((worker)))); (typeof((typeof(*((worker))) *)((worker)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); })->work) &handshake->lock rcu_read_lock_bh &peer->keypairs.keypair_update_lock irq_context: 0 (wq_completion)wg-kex-wg1#6 (work_completion)(&({ do { const void *__vpp_verify = (typeof((worker) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __asm__ ("" : "=r"(__ptr) : "0"((typeof(*((worker))) *)((worker)))); (typeof((typeof(*((worker))) *)((worker)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); })->work) &handshake->lock rcu_read_lock_bh &table->lock#2 irq_context: 0 (wq_completion)wg-kex-wg1#6 (work_completion)(&({ do { const void *__vpp_verify = (typeof((worker) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __asm__ ("" : "=r"(__ptr) : "0"((typeof(*((worker))) *)((worker)))); (typeof((typeof(*((worker))) *)((worker)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); })->work) rcu_read_lock_bh &obj_hash[i].lock irq_context: 0 (wq_completion)wg-kex-wg1#6 (work_completion)(&({ do { const void *__vpp_verify = (typeof((worker) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __asm__ ("" : "=r"(__ptr) : "0"((typeof(*((worker))) *)((worker)))); (typeof((typeof(*((worker))) *)((worker)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); })->work) rcu_read_lock_bh &base->lock irq_context: 0 (wq_completion)wg-kex-wg1#6 (work_completion)(&({ do { const void *__vpp_verify = (typeof((worker) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __asm__ ("" : "=r"(__ptr) : "0"((typeof(*((worker))) *)((worker)))); (typeof((typeof(*((worker))) *)((worker)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); })->work) rcu_read_lock_bh &base->lock &obj_hash[i].lock irq_context: 0 (wq_completion)wg-kex-wg1#6 (work_completion)(&({ do { const void *__vpp_verify = (typeof((worker) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __asm__ ("" : "=r"(__ptr) : "0"((typeof(*((worker))) *)((worker)))); (typeof((typeof(*((worker))) *)((worker)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); })->work) fill_pool_map-wait-type-override pool_lock#2 irq_context: 0 (wq_completion)wg-kex-wg1#6 (work_completion)(&({ do { const void *__vpp_verify = (typeof((worker) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __asm__ ("" : "=r"(__ptr) : "0"((typeof(*((worker))) *)((worker)))); (typeof((typeof(*((worker))) *)((worker)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); })->work) fill_pool_map-wait-type-override pool_lock irq_context: 0 (wq_completion)wg-kex-wg1#6 (work_completion)(&({ do { const void *__vpp_verify = (typeof((worker) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __asm__ ("" : "=r"(__ptr) : "0"((typeof(*((worker))) *)((worker)))); (typeof((typeof(*((worker))) *)((worker)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); })->work) &obj_hash[i].lock irq_context: 0 (wq_completion)wg-kex-wg1#6 (work_completion)(&({ do { const void *__vpp_verify = (typeof((worker) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __asm__ ("" : "=r"(__ptr) : "0"((typeof(*((worker))) *)((worker)))); (typeof((typeof(*((worker))) *)((worker)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); })->work) &base->lock irq_context: 0 (wq_completion)wg-kex-wg1#6 (work_completion)(&({ do { const void *__vpp_verify = (typeof((worker) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __asm__ ("" : "=r"(__ptr) : "0"((typeof(*((worker))) *)((worker)))); (typeof((typeof(*((worker))) *)((worker)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); })->work) &base->lock &obj_hash[i].lock irq_context: 0 (wq_completion)wg-kex-wg1#6 (work_completion)(&({ do { const void *__vpp_verify = (typeof((worker) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __asm__ ("" : "=r"(__ptr) : "0"((typeof(*((worker))) *)((worker)))); (typeof((typeof(*((worker))) *)((worker)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); })->work) tk_core.seq.seqcount irq_context: 0 (wq_completion)wg-kex-wg1#6 (work_completion)(&({ do { const void *__vpp_verify = (typeof((worker) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __asm__ ("" : "=r"(__ptr) : "0"((typeof(*((worker))) *)((worker)))); (typeof((typeof(*((worker))) *)((worker)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); })->work) &list->lock#17 irq_context: 0 (wq_completion)wg-kex-wg1#6 (work_completion)(&({ do { const void *__vpp_verify = (typeof((worker) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __asm__ ("" : "=r"(__ptr) : "0"((typeof(*((worker))) *)((worker)))); (typeof((typeof(*((worker))) *)((worker)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); })->work) rcu_read_lock_bh &r->producer_lock#2 irq_context: 0 (wq_completion)wg-kex-wg1#6 (work_completion)(&({ do { const void *__vpp_verify = (typeof((worker) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __asm__ ("" : "=r"(__ptr) : "0"((typeof(*((worker))) *)((worker)))); (typeof((typeof(*((worker))) *)((worker)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); })->work) rcu_read_lock_bh rcu_read_lock &pool->lock irq_context: 0 (wq_completion)wg-kex-wg1#6 (work_completion)(&({ do { const void *__vpp_verify = (typeof((worker) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __asm__ ("" : "=r"(__ptr) : "0"((typeof(*((worker))) *)((worker)))); (typeof((typeof(*((worker))) *)((worker)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); })->work) rcu_read_lock_bh rcu_read_lock &pool->lock &obj_hash[i].lock irq_context: 0 (wq_completion)wg-kex-wg1#6 (work_completion)(&({ do { const void *__vpp_verify = (typeof((worker) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __asm__ ("" : "=r"(__ptr) : "0"((typeof(*((worker))) *)((worker)))); (typeof((typeof(*((worker))) *)((worker)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); })->work) rcu_read_lock_bh rcu_read_lock &pool->lock &p->pi_lock irq_context: 0 (wq_completion)wg-kex-wg1#6 (work_completion)(&({ do { const void *__vpp_verify = (typeof((worker) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __asm__ ("" : "=r"(__ptr) : "0"((typeof(*((worker))) *)((worker)))); (typeof((typeof(*((worker))) *)((worker)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); })->work) rcu_read_lock_bh rcu_read_lock &pool->lock &p->pi_lock &rq->__lock irq_context: 0 (wq_completion)wg-kex-wg1#6 (work_completion)(&({ do { const void *__vpp_verify = (typeof((worker) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __asm__ ("" : "=r"(__ptr) : "0"((typeof(*((worker))) *)((worker)))); (typeof((typeof(*((worker))) *)((worker)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); })->work) rcu_read_lock_bh rcu_read_lock &pool->lock &p->pi_lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)wg-kex-wg1#6 (work_completion)(&({ do { const void *__vpp_verify = (typeof((worker) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __asm__ ("" : "=r"(__ptr) : "0"((typeof(*((worker))) *)((worker)))); (typeof((typeof(*((worker))) *)((worker)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); })->work) &rq->__lock irq_context: 0 (wq_completion)wg-crypt-wg1#3 irq_context: 0 (wq_completion)wg-crypt-wg1#3 (work_completion)(&({ do { const void *__vpp_verify = (typeof((worker) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __asm__ ("" : "=r"(__ptr) : "0"((typeof(*((worker))) *)((worker)))); (typeof((typeof(*((worker))) *)((worker)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); })->work) irq_context: 0 (wq_completion)wg-crypt-wg1#3 (work_completion)(&({ do { const void *__vpp_verify = (typeof((worker) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __asm__ ("" : "=r"(__ptr) : "0"((typeof(*((worker))) *)((worker)))); (typeof((typeof(*((worker))) *)((worker)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); })->work) &r->consumer_lock#2 irq_context: 0 (wq_completion)wg-crypt-wg1#3 (work_completion)(&({ do { const void *__vpp_verify = (typeof((worker) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __asm__ ("" : "=r"(__ptr) : "0"((typeof(*((worker))) *)((worker)))); (typeof((typeof(*((worker))) *)((worker)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); })->work) rcu_read_lock &pool->lock irq_context: 0 (wq_completion)wg-crypt-wg1#3 (work_completion)(&({ do { const void *__vpp_verify = (typeof((worker) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __asm__ ("" : "=r"(__ptr) : "0"((typeof(*((worker))) *)((worker)))); (typeof((typeof(*((worker))) *)((worker)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); })->work) rcu_read_lock &pool->lock &obj_hash[i].lock irq_context: 0 (wq_completion)wg-crypt-wg1#3 (work_completion)(&({ do { const void *__vpp_verify = (typeof((worker) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __asm__ ("" : "=r"(__ptr) : "0"((typeof(*((worker))) *)((worker)))); (typeof((typeof(*((worker))) *)((worker)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); })->work) rcu_read_lock &pool->lock &p->pi_lock irq_context: 0 (wq_completion)wg-crypt-wg1#3 (work_completion)(&({ do { const void *__vpp_verify = (typeof((worker) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __asm__ ("" : "=r"(__ptr) : "0"((typeof(*((worker))) *)((worker)))); (typeof((typeof(*((worker))) *)((worker)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); })->work) rcu_read_lock &pool->lock &p->pi_lock &rq->__lock irq_context: 0 (wq_completion)wg-crypt-wg1#3 (work_completion)(&({ do { const void *__vpp_verify = (typeof((worker) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __asm__ ("" : "=r"(__ptr) : "0"((typeof(*((worker))) *)((worker)))); (typeof((typeof(*((worker))) *)((worker)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); })->work) rcu_read_lock &pool->lock &p->pi_lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)wg-crypt-wg1#3 (work_completion)(&peer->transmit_packet_work) irq_context: 0 (wq_completion)wg-crypt-wg1#3 (work_completion)(&peer->transmit_packet_work) rcu_read_lock_bh &obj_hash[i].lock irq_context: 0 (wq_completion)wg-crypt-wg1#3 (work_completion)(&peer->transmit_packet_work) &obj_hash[i].lock irq_context: 0 (wq_completion)wg-crypt-wg1#3 (work_completion)(&peer->transmit_packet_work) &peer->endpoint_lock irq_context: 0 (wq_completion)wg-crypt-wg1#3 (work_completion)(&peer->transmit_packet_work) &peer->endpoint_lock rcu_read_lock_bh rcu_read_lock &____s->seqcount#7 irq_context: 0 (wq_completion)wg-crypt-wg1#3 (work_completion)(&peer->transmit_packet_work) &peer->endpoint_lock rcu_read_lock_bh rcu_read_lock &____s->seqcount#10 irq_context: 0 (wq_completion)wg-crypt-wg1#3 (work_completion)(&peer->transmit_packet_work) &peer->endpoint_lock rcu_read_lock_bh rcu_read_lock rcu_read_lock_bh tk_core.seq.seqcount irq_context: 0 (wq_completion)wg-crypt-wg1#3 (work_completion)(&peer->transmit_packet_work) &peer->endpoint_lock rcu_read_lock_bh rcu_read_lock rcu_read_lock_bh &list->lock#5 irq_context: 0 (wq_completion)wg-crypt-wg1#3 (work_completion)(&peer->transmit_packet_work) rcu_read_lock_bh tk_core.seq.seqcount irq_context: 0 cb_lock &pcp->lock &zone->lock irq_context: 0 (wq_completion)wg-crypt-wg0#3 irq_context: 0 (wq_completion)wg-crypt-wg0#3 (work_completion)(&({ do { const void *__vpp_verify = (typeof((worker) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __asm__ ("" : "=r"(__ptr) : "0"((typeof(*((worker))) *)((worker)))); (typeof((typeof(*((worker))) *)((worker)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); })->work) irq_context: 0 (wq_completion)wg-crypt-wg0#3 (work_completion)(&({ do { const void *__vpp_verify = (typeof((worker) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __asm__ ("" : "=r"(__ptr) : "0"((typeof(*((worker))) *)((worker)))); (typeof((typeof(*((worker))) *)((worker)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); })->work) &r->consumer_lock#2 irq_context: 0 (wq_completion)wg-crypt-wg0#3 (work_completion)(&({ do { const void *__vpp_verify = (typeof((worker) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __asm__ ("" : "=r"(__ptr) : "0"((typeof(*((worker))) *)((worker)))); (typeof((typeof(*((worker))) *)((worker)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); })->work) tk_core.seq.seqcount irq_context: 0 (wq_completion)ipv6_addrconf (work_completion)(&(&ifa->dad_work)->work) rtnl_mutex &idev->mc_lock remove_cache_srcu &pcp->lock &zone->lock irq_context: 0 (wq_completion)ipv6_addrconf (work_completion)(&(&ifa->dad_work)->work) rtnl_mutex &idev->mc_lock remove_cache_srcu &pcp->lock &zone->lock &____s->seqcount irq_context: softirq rcu_read_lock rcu_read_lock &pool->lock/1 fill_pool_map-wait-type-override pool_lock#2 irq_context: softirq rcu_read_lock rcu_read_lock &pool->lock/1 fill_pool_map-wait-type-override &c->lock irq_context: softirq rcu_read_lock rcu_read_lock &pool->lock/1 fill_pool_map-wait-type-override &____s->seqcount#2 irq_context: softirq rcu_read_lock rcu_read_lock &pool->lock/1 fill_pool_map-wait-type-override &____s->seqcount irq_context: softirq rcu_read_lock rcu_read_lock &pool->lock/1 fill_pool_map-wait-type-override pool_lock irq_context: 0 (wq_completion)phy6 irq_context: 0 (wq_completion)phy6 (work_completion)(&local->reconfig_filter) irq_context: 0 (wq_completion)phy6 (work_completion)(&local->reconfig_filter) &local->filter_lock irq_context: 0 rtnl_mutex &wg->device_update_lock rcu_state.exp_mutex &obj_hash[i].lock pool_lock irq_context: 0 rtnl_mutex &wg->device_update_lock rcu_state.exp_mutex rcu_read_lock &pool->lock fill_pool_map-wait-type-override pool_lock#2 irq_context: 0 rtnl_mutex &wg->device_update_lock rcu_state.exp_mutex rcu_read_lock &pool->lock fill_pool_map-wait-type-override pool_lock irq_context: 0 (wq_completion)wg-kex-wg0#7 irq_context: 0 (wq_completion)wg-kex-wg0#7 (work_completion)(&peer->transmit_handshake_work) irq_context: 0 (wq_completion)wg-kex-wg0#7 (work_completion)(&peer->transmit_handshake_work) tk_core.seq.seqcount irq_context: 0 (wq_completion)wg-kex-wg0#7 (work_completion)(&peer->transmit_handshake_work) &wg->static_identity.lock irq_context: 0 (wq_completion)wg-kex-wg0#7 (work_completion)(&peer->transmit_handshake_work) &wg->static_identity.lock &handshake->lock irq_context: 0 (wq_completion)wg-kex-wg0#7 (work_completion)(&peer->transmit_handshake_work) &wg->static_identity.lock &handshake->lock crngs.lock irq_context: 0 (wq_completion)wg-kex-wg0#7 (work_completion)(&peer->transmit_handshake_work) &wg->static_identity.lock &handshake->lock tk_core.seq.seqcount irq_context: 0 (wq_completion)wg-kex-wg0#7 (work_completion)(&peer->transmit_handshake_work) &wg->static_identity.lock &handshake->lock &table->lock#2 irq_context: 0 (wq_completion)wg-kex-wg0#7 (work_completion)(&peer->transmit_handshake_work) &wg->static_identity.lock &handshake->lock rcu_read_lock_bh batched_entropy_u32.lock irq_context: 0 (wq_completion)wg-kex-wg0#7 (work_completion)(&peer->transmit_handshake_work) &wg->static_identity.lock &handshake->lock rcu_read_lock_bh &table->lock#2 irq_context: 0 (wq_completion)wg-kex-wg0#7 (work_completion)(&peer->transmit_handshake_work) &cookie->lock irq_context: 0 (wq_completion)wg-kex-wg0#7 (work_completion)(&peer->transmit_handshake_work) &cookie->lock irq_context: 0 (wq_completion)wg-kex-wg0#7 (work_completion)(&peer->transmit_handshake_work) rcu_read_lock_bh &obj_hash[i].lock irq_context: 0 (wq_completion)wg-kex-wg0#7 (work_completion)(&peer->transmit_handshake_work) rcu_read_lock_bh &base->lock irq_context: 0 (wq_completion)wg-kex-wg0#7 (work_completion)(&peer->transmit_handshake_work) rcu_read_lock_bh &base->lock &obj_hash[i].lock irq_context: 0 (wq_completion)wg-kex-wg0#7 (work_completion)(&peer->transmit_handshake_work) &obj_hash[i].lock irq_context: 0 (wq_completion)wg-kex-wg0#7 (work_completion)(&peer->transmit_handshake_work) &peer->endpoint_lock irq_context: 0 (wq_completion)wg-kex-wg0#7 (work_completion)(&peer->transmit_handshake_work) &peer->endpoint_lock rcu_read_lock_bh rcu_read_lock &dir->lock#2 irq_context: 0 (wq_completion)wg-kex-wg0#7 (work_completion)(&peer->transmit_handshake_work) &peer->endpoint_lock rcu_read_lock_bh crngs.lock irq_context: 0 (wq_completion)wg-kex-wg0#7 (work_completion)(&peer->transmit_handshake_work) &peer->endpoint_lock rcu_read_lock_bh batched_entropy_u32.lock irq_context: 0 (wq_completion)wg-kex-wg0#7 (work_completion)(&peer->transmit_handshake_work) &peer->endpoint_lock rcu_read_lock_bh rcu_read_lock &____s->seqcount#7 irq_context: 0 (wq_completion)wg-kex-wg0#7 (work_completion)(&peer->transmit_handshake_work) &peer->endpoint_lock rcu_read_lock_bh rcu_read_lock &nf_conntrack_locks[i] irq_context: 0 (wq_completion)wg-kex-wg0#7 (work_completion)(&peer->transmit_handshake_work) &peer->endpoint_lock rcu_read_lock_bh rcu_read_lock &nf_conntrack_locks[i] &nf_conntrack_locks[i]/1 irq_context: 0 (wq_completion)wg-kex-wg0#7 (work_completion)(&peer->transmit_handshake_work) &peer->endpoint_lock rcu_read_lock_bh rcu_read_lock &nf_conntrack_locks[i] &nf_conntrack_locks[i]/1 batched_entropy_u8.lock irq_context: 0 (wq_completion)wg-kex-wg0#7 (work_completion)(&peer->transmit_handshake_work) &peer->endpoint_lock rcu_read_lock_bh rcu_read_lock &nf_conntrack_locks[i]/1 irq_context: 0 (wq_completion)wg-kex-wg0#7 (work_completion)(&peer->transmit_handshake_work) &peer->endpoint_lock rcu_read_lock_bh rcu_read_lock &obj_hash[i].lock irq_context: 0 (wq_completion)wg-kex-wg0#7 (work_completion)(&peer->transmit_handshake_work) &peer->endpoint_lock rcu_read_lock_bh rcu_read_lock &tbl->lock irq_context: 0 (wq_completion)wg-kex-wg0#7 (work_completion)(&peer->transmit_handshake_work) &peer->endpoint_lock rcu_read_lock_bh rcu_read_lock &n->lock irq_context: 0 (wq_completion)wg-kex-wg0#7 (work_completion)(&peer->transmit_handshake_work) &peer->endpoint_lock rcu_read_lock_bh rcu_read_lock &____s->seqcount#9 irq_context: 0 (wq_completion)wg-kex-wg0#7 (work_completion)(&peer->transmit_handshake_work) &peer->endpoint_lock rcu_read_lock_bh rcu_read_lock rcu_read_lock_bh tk_core.seq.seqcount irq_context: 0 (wq_completion)wg-kex-wg0#7 (work_completion)(&peer->transmit_handshake_work) &peer->endpoint_lock rcu_read_lock_bh rcu_read_lock rcu_read_lock_bh &list->lock#5 irq_context: 0 (wq_completion)wg-kex-wg0#7 (work_completion)(&peer->transmit_handshake_work) batched_entropy_u8.lock irq_context: 0 (wq_completion)wg-kex-wg0#7 (work_completion)(&peer->transmit_handshake_work) &peer->endpoint_lock rcu_read_lock_bh rcu_read_lock rcu_read_lock rcu_read_lock &dir->lock#2 irq_context: softirq (&peer->timer_persistent_keepalive) batched_entropy_u8.lock irq_context: softirq (&peer->timer_persistent_keepalive) kfence_freelist_lock irq_context: 0 (wq_completion)wg-crypt-wg0#3 (work_completion)(&({ do { const void *__vpp_verify = (typeof((worker) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __asm__ ("" : "=r"(__ptr) : "0"((typeof(*((worker))) *)((worker)))); (typeof((typeof(*((worker))) *)((worker)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); })->work) rcu_read_lock &pool->lock irq_context: 0 (wq_completion)wg-crypt-wg0#3 (work_completion)(&({ do { const void *__vpp_verify = (typeof((worker) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __asm__ ("" : "=r"(__ptr) : "0"((typeof(*((worker))) *)((worker)))); (typeof((typeof(*((worker))) *)((worker)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); })->work) rcu_read_lock &pool->lock &obj_hash[i].lock irq_context: 0 (wq_completion)wg-crypt-wg0#3 (work_completion)(&({ do { const void *__vpp_verify = (typeof((worker) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __asm__ ("" : "=r"(__ptr) : "0"((typeof(*((worker))) *)((worker)))); (typeof((typeof(*((worker))) *)((worker)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); })->work) rcu_read_lock &pool->lock &p->pi_lock irq_context: 0 (wq_completion)wg-crypt-wg0#3 (work_completion)(&({ do { const void *__vpp_verify = (typeof((worker) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __asm__ ("" : "=r"(__ptr) : "0"((typeof(*((worker))) *)((worker)))); (typeof((typeof(*((worker))) *)((worker)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); })->work) rcu_read_lock &pool->lock &p->pi_lock &rq->__lock irq_context: 0 (wq_completion)wg-crypt-wg0#3 (work_completion)(&({ do { const void *__vpp_verify = (typeof((worker) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __asm__ ("" : "=r"(__ptr) : "0"((typeof(*((worker))) *)((worker)))); (typeof((typeof(*((worker))) *)((worker)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); })->work) rcu_read_lock &pool->lock &p->pi_lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)wg-crypt-wg0#3 (work_completion)(&peer->transmit_packet_work) irq_context: 0 (wq_completion)wg-crypt-wg0#3 (work_completion)(&peer->transmit_packet_work) rcu_read_lock_bh &obj_hash[i].lock irq_context: 0 (wq_completion)wg-crypt-wg0#3 (work_completion)(&peer->transmit_packet_work) rcu_read_lock_bh &base->lock irq_context: 0 (wq_completion)wg-crypt-wg0#3 (work_completion)(&peer->transmit_packet_work) &obj_hash[i].lock irq_context: 0 (wq_completion)wg-crypt-wg0#3 (work_completion)(&peer->transmit_packet_work) &peer->endpoint_lock irq_context: 0 (wq_completion)wg-crypt-wg0#3 (work_completion)(&peer->transmit_packet_work) &peer->endpoint_lock rcu_read_lock_bh rcu_read_lock &____s->seqcount#7 irq_context: 0 (wq_completion)wg-crypt-wg0#3 (work_completion)(&peer->transmit_packet_work) &peer->endpoint_lock rcu_read_lock_bh rcu_read_lock &____s->seqcount#10 irq_context: 0 (wq_completion)wg-crypt-wg0#3 (work_completion)(&peer->transmit_packet_work) &peer->endpoint_lock rcu_read_lock_bh rcu_read_lock rcu_read_lock_bh tk_core.seq.seqcount irq_context: 0 (wq_completion)wg-crypt-wg0#3 (work_completion)(&peer->transmit_packet_work) &peer->endpoint_lock rcu_read_lock_bh rcu_read_lock rcu_read_lock_bh &list->lock#5 irq_context: 0 (wq_completion)wg-crypt-wg1#3 (work_completion)(&({ do { const void *__vpp_verify = (typeof((worker) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __asm__ ("" : "=r"(__ptr) : "0"((typeof(*((worker))) *)((worker)))); (typeof((typeof(*((worker))) *)((worker)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); })->work) tk_core.seq.seqcount irq_context: 0 (wq_completion)wg-crypt-wg0#2 (work_completion)(&peer->transmit_packet_work) &rq->__lock irq_context: 0 (wq_completion)wg-kex-wg0#8 irq_context: 0 (wq_completion)wg-kex-wg0#8 (work_completion)(&peer->transmit_handshake_work) irq_context: 0 (wq_completion)wg-kex-wg0#8 (work_completion)(&peer->transmit_handshake_work) tk_core.seq.seqcount irq_context: 0 (wq_completion)wg-kex-wg0#8 (work_completion)(&peer->transmit_handshake_work) &wg->static_identity.lock irq_context: 0 (wq_completion)wg-kex-wg0#8 (work_completion)(&peer->transmit_handshake_work) &wg->static_identity.lock &handshake->lock irq_context: 0 (wq_completion)wg-kex-wg0#8 (work_completion)(&peer->transmit_handshake_work) &wg->static_identity.lock &handshake->lock crngs.lock irq_context: 0 (wq_completion)wg-kex-wg0#8 (work_completion)(&peer->transmit_handshake_work) &wg->static_identity.lock &handshake->lock tk_core.seq.seqcount irq_context: 0 (wq_completion)wg-kex-wg0#8 (work_completion)(&peer->transmit_handshake_work) &wg->static_identity.lock &handshake->lock &table->lock#2 irq_context: 0 (wq_completion)wg-kex-wg0#8 (work_completion)(&peer->transmit_handshake_work) &wg->static_identity.lock &handshake->lock rcu_read_lock_bh batched_entropy_u32.lock irq_context: 0 (wq_completion)wg-kex-wg0#8 (work_completion)(&peer->transmit_handshake_work) &wg->static_identity.lock &handshake->lock rcu_read_lock_bh &table->lock#2 irq_context: 0 (wq_completion)wg-kex-wg0#8 (work_completion)(&peer->transmit_handshake_work) &cookie->lock irq_context: 0 (wq_completion)wg-kex-wg0#8 (work_completion)(&peer->transmit_handshake_work) &cookie->lock irq_context: 0 (wq_completion)wg-kex-wg0#8 (work_completion)(&peer->transmit_handshake_work) rcu_read_lock_bh &obj_hash[i].lock irq_context: 0 (wq_completion)wg-kex-wg0#8 (work_completion)(&peer->transmit_handshake_work) rcu_read_lock_bh &base->lock irq_context: 0 (wq_completion)wg-kex-wg0#8 (work_completion)(&peer->transmit_handshake_work) rcu_read_lock_bh &base->lock &obj_hash[i].lock irq_context: 0 (wq_completion)mld (work_completion)(&(&idev->mc_ifc_work)->work) &idev->mc_lock rcu_read_lock rcu_read_lock rcu_read_lock_bh rcu_read_lock rcu_read_lock &nf_conntrack_locks[i]/1 irq_context: 0 (wq_completion)wg-kex-wg0#8 (work_completion)(&peer->transmit_handshake_work) &obj_hash[i].lock irq_context: 0 (wq_completion)wg-kex-wg0#8 (work_completion)(&peer->transmit_handshake_work) &peer->endpoint_lock irq_context: 0 (wq_completion)wg-kex-wg0#8 (work_completion)(&peer->transmit_handshake_work) &peer->endpoint_lock rcu_read_lock_bh rcu_read_lock &dir->lock#2 irq_context: 0 (wq_completion)wg-kex-wg0#8 (work_completion)(&peer->transmit_handshake_work) &peer->endpoint_lock rcu_read_lock_bh crngs.lock irq_context: 0 (wq_completion)wg-kex-wg0#8 (work_completion)(&peer->transmit_handshake_work) &peer->endpoint_lock rcu_read_lock_bh batched_entropy_u32.lock irq_context: 0 (wq_completion)wg-kex-wg0#8 (work_completion)(&peer->transmit_handshake_work) &peer->endpoint_lock rcu_read_lock_bh rcu_read_lock &____s->seqcount#7 irq_context: 0 (wq_completion)wg-kex-wg0#8 (work_completion)(&peer->transmit_handshake_work) &peer->endpoint_lock rcu_read_lock_bh rcu_read_lock &nf_conntrack_locks[i] irq_context: 0 (wq_completion)wg-kex-wg0#8 (work_completion)(&peer->transmit_handshake_work) &peer->endpoint_lock rcu_read_lock_bh rcu_read_lock &nf_conntrack_locks[i] &nf_conntrack_locks[i]/1 irq_context: 0 (wq_completion)wg-kex-wg0#8 (work_completion)(&peer->transmit_handshake_work) &peer->endpoint_lock rcu_read_lock_bh rcu_read_lock &nf_conntrack_locks[i] &nf_conntrack_locks[i]/1 batched_entropy_u8.lock irq_context: 0 (wq_completion)wg-kex-wg0#8 (work_completion)(&peer->transmit_handshake_work) &peer->endpoint_lock rcu_read_lock_bh rcu_read_lock &obj_hash[i].lock irq_context: 0 (wq_completion)wg-kex-wg0#8 (work_completion)(&peer->transmit_handshake_work) &peer->endpoint_lock rcu_read_lock_bh rcu_read_lock &tbl->lock irq_context: 0 (wq_completion)wg-kex-wg0#8 (work_completion)(&peer->transmit_handshake_work) &peer->endpoint_lock rcu_read_lock_bh rcu_read_lock &n->lock irq_context: 0 (wq_completion)wg-kex-wg0#8 (work_completion)(&peer->transmit_handshake_work) &peer->endpoint_lock rcu_read_lock_bh rcu_read_lock &____s->seqcount#9 irq_context: 0 (wq_completion)wg-kex-wg0#8 (work_completion)(&peer->transmit_handshake_work) &peer->endpoint_lock rcu_read_lock_bh rcu_read_lock rcu_read_lock_bh tk_core.seq.seqcount irq_context: 0 (wq_completion)wg-kex-wg0#8 (work_completion)(&peer->transmit_handshake_work) &peer->endpoint_lock rcu_read_lock_bh rcu_read_lock rcu_read_lock_bh &list->lock#5 irq_context: 0 (wq_completion)wg-kex-wg0#8 (work_completion)(&peer->transmit_handshake_work) &rq->__lock irq_context: 0 (wq_completion)events_power_efficient (check_lifetime_work).work rcu_read_lock &rq->__lock irq_context: 0 rtnl_mutex netpoll_srcu &rq->__lock irq_context: 0 (wq_completion)wg-kex-wg0#8 (work_completion)(&peer->transmit_handshake_work) &peer->endpoint_lock rcu_read_lock_bh rcu_read_lock rcu_read_lock rcu_read_lock &dir->lock#2 irq_context: 0 (wq_completion)wg-kex-wg0#8 (work_completion)(&peer->transmit_handshake_work) batched_entropy_u8.lock irq_context: 0 (wq_completion)cfg80211 (work_completion)(&rdev->event_work) &rdev->wiphy.mtx &wdev->mtx &____s->seqcount#2 irq_context: 0 (wq_completion)wg-kex-wg0#9 irq_context: 0 (wq_completion)wg-kex-wg0#9 (work_completion)(&peer->transmit_handshake_work) irq_context: 0 (wq_completion)wg-kex-wg0#9 (work_completion)(&peer->transmit_handshake_work) tk_core.seq.seqcount irq_context: 0 (wq_completion)wg-kex-wg0#9 (work_completion)(&peer->transmit_handshake_work) &wg->static_identity.lock irq_context: 0 (wq_completion)wg-kex-wg0#9 (work_completion)(&peer->transmit_handshake_work) &wg->static_identity.lock &handshake->lock irq_context: 0 (wq_completion)wg-kex-wg0#9 (work_completion)(&peer->transmit_handshake_work) &wg->static_identity.lock &handshake->lock crngs.lock irq_context: 0 (wq_completion)wg-kex-wg0#9 (work_completion)(&peer->transmit_handshake_work) &wg->static_identity.lock &handshake->lock tk_core.seq.seqcount irq_context: 0 (wq_completion)wg-kex-wg0#9 (work_completion)(&peer->transmit_handshake_work) &wg->static_identity.lock &handshake->lock &table->lock#2 irq_context: 0 (wq_completion)wg-kex-wg0#9 (work_completion)(&peer->transmit_handshake_work) &wg->static_identity.lock &handshake->lock rcu_read_lock_bh batched_entropy_u32.lock irq_context: 0 (wq_completion)wg-kex-wg0#9 (work_completion)(&peer->transmit_handshake_work) &wg->static_identity.lock &handshake->lock rcu_read_lock_bh &table->lock#2 irq_context: 0 (wq_completion)wg-kex-wg0#9 (work_completion)(&peer->transmit_handshake_work) &cookie->lock irq_context: 0 (wq_completion)wg-kex-wg0#9 (work_completion)(&peer->transmit_handshake_work) &cookie->lock irq_context: 0 (wq_completion)wg-kex-wg0#9 (work_completion)(&peer->transmit_handshake_work) rcu_read_lock_bh &obj_hash[i].lock irq_context: 0 (wq_completion)wg-kex-wg0#9 (work_completion)(&peer->transmit_handshake_work) rcu_read_lock_bh &base->lock irq_context: 0 (wq_completion)wg-kex-wg0#9 (work_completion)(&peer->transmit_handshake_work) rcu_read_lock_bh &base->lock &obj_hash[i].lock irq_context: 0 (wq_completion)wg-kex-wg0#9 (work_completion)(&peer->transmit_handshake_work) &obj_hash[i].lock irq_context: 0 (wq_completion)wg-kex-wg0#9 (work_completion)(&peer->transmit_handshake_work) &peer->endpoint_lock irq_context: 0 (wq_completion)wg-kex-wg0#9 (work_completion)(&peer->transmit_handshake_work) &peer->endpoint_lock rcu_read_lock_bh rcu_read_lock &dir->lock#2 irq_context: 0 (wq_completion)wg-kex-wg0#9 (work_completion)(&peer->transmit_handshake_work) &peer->endpoint_lock rcu_read_lock_bh rcu_read_lock rcu_read_lock rcu_read_lock &dir->lock#2 irq_context: 0 (wq_completion)wg-kex-wg0#9 (work_completion)(&peer->transmit_handshake_work) &peer->endpoint_lock rcu_read_lock_bh crngs.lock irq_context: 0 (wq_completion)wg-kex-wg0#9 (work_completion)(&peer->transmit_handshake_work) &peer->endpoint_lock rcu_read_lock_bh batched_entropy_u32.lock irq_context: 0 (wq_completion)wg-kex-wg0#9 (work_completion)(&peer->transmit_handshake_work) &peer->endpoint_lock rcu_read_lock_bh rcu_read_lock &____s->seqcount#7 irq_context: 0 (wq_completion)wg-kex-wg0#9 (work_completion)(&peer->transmit_handshake_work) &peer->endpoint_lock rcu_read_lock_bh rcu_read_lock &nf_conntrack_locks[i] irq_context: 0 (wq_completion)wg-kex-wg0#9 (work_completion)(&peer->transmit_handshake_work) &peer->endpoint_lock rcu_read_lock_bh rcu_read_lock &nf_conntrack_locks[i] &nf_conntrack_locks[i]/1 irq_context: 0 (wq_completion)wg-kex-wg0#9 (work_completion)(&peer->transmit_handshake_work) &peer->endpoint_lock rcu_read_lock_bh rcu_read_lock &nf_conntrack_locks[i] &nf_conntrack_locks[i]/1 batched_entropy_u8.lock irq_context: 0 (wq_completion)wg-kex-wg0#9 (work_completion)(&peer->transmit_handshake_work) &peer->endpoint_lock rcu_read_lock_bh rcu_read_lock &nf_conntrack_locks[i]/1 irq_context: 0 (wq_completion)wg-kex-wg0#9 (work_completion)(&peer->transmit_handshake_work) &peer->endpoint_lock rcu_read_lock_bh rcu_read_lock &obj_hash[i].lock irq_context: 0 (wq_completion)wg-kex-wg0#9 (work_completion)(&peer->transmit_handshake_work) &peer->endpoint_lock rcu_read_lock_bh rcu_read_lock &tbl->lock irq_context: 0 (wq_completion)wg-kex-wg0#9 (work_completion)(&peer->transmit_handshake_work) &peer->endpoint_lock rcu_read_lock_bh rcu_read_lock &n->lock irq_context: 0 (wq_completion)wg-kex-wg0#9 (work_completion)(&peer->transmit_handshake_work) &peer->endpoint_lock rcu_read_lock_bh rcu_read_lock &____s->seqcount#9 irq_context: 0 (wq_completion)wg-kex-wg0#9 (work_completion)(&peer->transmit_handshake_work) &peer->endpoint_lock rcu_read_lock_bh rcu_read_lock rcu_read_lock_bh tk_core.seq.seqcount irq_context: 0 (wq_completion)wg-kex-wg0#9 (work_completion)(&peer->transmit_handshake_work) &peer->endpoint_lock rcu_read_lock_bh rcu_read_lock rcu_read_lock_bh &list->lock#5 irq_context: softirq rcu_read_lock rcu_read_lock k-slock-AF_INET6 rcu_read_lock &c->lock irq_context: softirq rcu_read_lock rcu_read_lock k-slock-AF_INET batched_entropy_u32.lock crngs.lock irq_context: 0 (wq_completion)wg-kex-wg0#9 (work_completion)(&peer->transmit_handshake_work) batched_entropy_u8.lock irq_context: 0 &wg->device_update_lock rcu_state.barrier_mutex.wait_lock irq_context: 0 &wg->device_update_lock init_lock irq_context: 0 &wg->device_update_lock &zone->lock irq_context: 0 &wg->device_update_lock &zone->lock &____s->seqcount irq_context: 0 sb_writers#4 sb_internal jbd2_handle &ei->i_data_sem &ei->i_es_lock key#7 irq_context: 0 sb_writers#4 sb_internal jbd2_handle &ei->i_data_sem &c->lock irq_context: 0 sb_writers#4 &type->i_mutex_dir_key#3/1 sb_internal jbd2_handle &ei->i_data_sem &ei->i_es_lock key#6 irq_context: 0 &xt[i].mutex &pcp->lock &zone->lock irq_context: 0 (wq_completion)events_freezable &rq->__lock irq_context: 0 sb_writers#5 &type->i_mutex_dir_key#5/1 tomoyo_ss rcu_node_0 irq_context: 0 sb_writers#5 &type->i_mutex_dir_key#5/1 tomoyo_ss &rcu_state.expedited_wq irq_context: 0 sb_writers#5 &type->i_mutex_dir_key#5/1 tomoyo_ss &rcu_state.expedited_wq &p->pi_lock irq_context: 0 sb_writers#5 &type->i_mutex_dir_key#5/1 tomoyo_ss &rcu_state.expedited_wq &p->pi_lock &rq->__lock irq_context: 0 sb_writers#5 &type->i_mutex_dir_key#5/1 tomoyo_ss &rcu_state.expedited_wq &p->pi_lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)events_unbound connector_reaper_work rcu_read_lock &rq->__lock irq_context: 0 (wq_completion)events_unbound connector_reaper_work rcu_read_lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)events (work_completion)(&(&nsim_dev->trap_data->trap_report_dw)->work) &devlink->lock_key#3 rcu_node_0 irq_context: 0 (wq_completion)events (work_completion)(&(&nsim_dev->trap_data->trap_report_dw)->work) &devlink->lock_key#3 &rcu_state.expedited_wq irq_context: 0 (wq_completion)events (work_completion)(&(&nsim_dev->trap_data->trap_report_dw)->work) &devlink->lock_key#3 &rcu_state.expedited_wq &p->pi_lock irq_context: 0 pernet_ops_rwsem rtnl_mutex &ul->lock#2 irq_context: 0 pernet_ops_rwsem rtnl_mutex _xmit_IPGRE irq_context: 0 pernet_ops_rwsem &fn->fou_lock irq_context: 0 pernet_ops_rwsem ipvs->sync_mutex irq_context: 0 pernet_ops_rwsem hwsim_radio_lock irq_context: 0 pernet_ops_rwsem &ent->pde_unload_lock irq_context: 0 pernet_ops_rwsem rdma_nets_rwsem irq_context: 0 pernet_ops_rwsem rdma_nets_rwsem rdma_nets.xa_lock irq_context: 0 pernet_ops_rwsem k-clock-AF_NETLINK irq_context: 0 pernet_ops_rwsem &nlk->wait irq_context: 0 pernet_ops_rwsem wlock-AF_NETLINK irq_context: 0 pernet_ops_rwsem &xa->xa_lock#9 irq_context: 0 pernet_ops_rwsem &fsnotify_mark_srcu irq_context: 0 pernet_ops_rwsem rtnl_mutex dev_base_lock &xa->xa_lock#3 &obj_hash[i].lock irq_context: 0 (wq_completion)wg-kex-wg0#10 irq_context: 0 (wq_completion)wg-kex-wg0#10 (work_completion)(&({ do { const void *__vpp_verify = (typeof((worker) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __asm__ ("" : "=r"(__ptr) : "0"((typeof(*((worker))) *)((worker)))); (typeof((typeof(*((worker))) *)((worker)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); })->work) irq_context: 0 (wq_completion)wg-kex-wg0#10 (work_completion)(&({ do { const void *__vpp_verify = (typeof((worker) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __asm__ ("" : "=r"(__ptr) : "0"((typeof(*((worker))) *)((worker)))); (typeof((typeof(*((worker))) *)((worker)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); })->work) &r->consumer_lock#2 irq_context: 0 (wq_completion)wg-kex-wg0#10 (work_completion)(&({ do { const void *__vpp_verify = (typeof((worker) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __asm__ ("" : "=r"(__ptr) : "0"((typeof(*((worker))) *)((worker)))); (typeof((typeof(*((worker))) *)((worker)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); })->work) &wg->static_identity.lock irq_context: 0 (wq_completion)wg-kex-wg0#10 (work_completion)(&({ do { const void *__vpp_verify = (typeof((worker) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __asm__ ("" : "=r"(__ptr) : "0"((typeof(*((worker))) *)((worker)))); (typeof((typeof(*((worker))) *)((worker)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); })->work) &wg->static_identity.lock &handshake->lock irq_context: 0 (wq_completion)wg-kex-wg0#10 (work_completion)(&({ do { const void *__vpp_verify = (typeof((worker) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __asm__ ("" : "=r"(__ptr) : "0"((typeof(*((worker))) *)((worker)))); (typeof((typeof(*((worker))) *)((worker)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); })->work) &wg->static_identity.lock &handshake->lock tk_core.seq.seqcount irq_context: 0 (wq_completion)wg-kex-wg0#10 (work_completion)(&({ do { const void *__vpp_verify = (typeof((worker) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __asm__ ("" : "=r"(__ptr) : "0"((typeof(*((worker))) *)((worker)))); (typeof((typeof(*((worker))) *)((worker)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); })->work) &wg->static_identity.lock &handshake->lock irq_context: 0 (wq_completion)wg-kex-wg0#10 (work_completion)(&({ do { const void *__vpp_verify = (typeof((worker) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __asm__ ("" : "=r"(__ptr) : "0"((typeof(*((worker))) *)((worker)))); (typeof((typeof(*((worker))) *)((worker)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); })->work) &wg->static_identity.lock &handshake->lock tk_core.seq.seqcount irq_context: 0 (wq_completion)wg-kex-wg0#10 (work_completion)(&({ do { const void *__vpp_verify = (typeof((worker) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __asm__ ("" : "=r"(__ptr) : "0"((typeof(*((worker))) *)((worker)))); (typeof((typeof(*((worker))) *)((worker)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); })->work) &peer->endpoint_lock irq_context: 0 (wq_completion)wg-kex-wg0#10 (work_completion)(&({ do { const void *__vpp_verify = (typeof((worker) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __asm__ ("" : "=r"(__ptr) : "0"((typeof(*((worker))) *)((worker)))); (typeof((typeof(*((worker))) *)((worker)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); })->work) tk_core.seq.seqcount irq_context: 0 (wq_completion)wg-kex-wg0#10 (work_completion)(&({ do { const void *__vpp_verify = (typeof((worker) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __asm__ ("" : "=r"(__ptr) : "0"((typeof(*((worker))) *)((worker)))); (typeof((typeof(*((worker))) *)((worker)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); })->work) &wg->static_identity.lock &handshake->lock crngs.lock irq_context: 0 (wq_completion)wg-kex-wg0#10 (work_completion)(&({ do { const void *__vpp_verify = (typeof((worker) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __asm__ ("" : "=r"(__ptr) : "0"((typeof(*((worker))) *)((worker)))); (typeof((typeof(*((worker))) *)((worker)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); })->work) &wg->static_identity.lock &handshake->lock &table->lock#2 irq_context: 0 (wq_completion)wg-kex-wg0#10 (work_completion)(&({ do { const void *__vpp_verify = (typeof((worker) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __asm__ ("" : "=r"(__ptr) : "0"((typeof(*((worker))) *)((worker)))); (typeof((typeof(*((worker))) *)((worker)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); })->work) &wg->static_identity.lock &handshake->lock rcu_read_lock_bh batched_entropy_u32.lock irq_context: 0 (wq_completion)wg-kex-wg0#10 (work_completion)(&({ do { const void *__vpp_verify = (typeof((worker) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __asm__ ("" : "=r"(__ptr) : "0"((typeof(*((worker))) *)((worker)))); (typeof((typeof(*((worker))) *)((worker)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); })->work) &wg->static_identity.lock &handshake->lock rcu_read_lock_bh &table->lock#2 irq_context: 0 (wq_completion)wg-kex-wg0#10 (work_completion)(&({ do { const void *__vpp_verify = (typeof((worker) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __asm__ ("" : "=r"(__ptr) : "0"((typeof(*((worker))) *)((worker)))); (typeof((typeof(*((worker))) *)((worker)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); })->work) &cookie->lock irq_context: 0 (wq_completion)wg-kex-wg0#10 (work_completion)(&({ do { const void *__vpp_verify = (typeof((worker) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __asm__ ("" : "=r"(__ptr) : "0"((typeof(*((worker))) *)((worker)))); (typeof((typeof(*((worker))) *)((worker)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); })->work) &cookie->lock irq_context: 0 (wq_completion)wg-kex-wg0#10 (work_completion)(&({ do { const void *__vpp_verify = (typeof((worker) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __asm__ ("" : "=r"(__ptr) : "0"((typeof(*((worker))) *)((worker)))); (typeof((typeof(*((worker))) *)((worker)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); })->work) &handshake->lock irq_context: 0 (wq_completion)wg-kex-wg0#10 (work_completion)(&({ do { const void *__vpp_verify = (typeof((worker) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __asm__ ("" : "=r"(__ptr) : "0"((typeof(*((worker))) *)((worker)))); (typeof((typeof(*((worker))) *)((worker)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); })->work) &handshake->lock fs_reclaim irq_context: 0 (wq_completion)wg-kex-wg0#10 (work_completion)(&({ do { const void *__vpp_verify = (typeof((worker) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __asm__ ("" : "=r"(__ptr) : "0"((typeof(*((worker))) *)((worker)))); (typeof((typeof(*((worker))) *)((worker)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); })->work) &handshake->lock fs_reclaim mmu_notifier_invalidate_range_start irq_context: 0 (wq_completion)wg-kex-wg0#10 (work_completion)(&({ do { const void *__vpp_verify = (typeof((worker) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __asm__ ("" : "=r"(__ptr) : "0"((typeof(*((worker))) *)((worker)))); (typeof((typeof(*((worker))) *)((worker)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); })->work) &handshake->lock tk_core.seq.seqcount irq_context: 0 (wq_completion)wg-kex-wg0#10 (work_completion)(&({ do { const void *__vpp_verify = (typeof((worker) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __asm__ ("" : "=r"(__ptr) : "0"((typeof(*((worker))) *)((worker)))); (typeof((typeof(*((worker))) *)((worker)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); })->work) &handshake->lock rcu_read_lock_bh &peer->keypairs.keypair_update_lock irq_context: 0 (wq_completion)wg-kex-wg0#10 (work_completion)(&({ do { const void *__vpp_verify = (typeof((worker) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __asm__ ("" : "=r"(__ptr) : "0"((typeof(*((worker))) *)((worker)))); (typeof((typeof(*((worker))) *)((worker)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); })->work) &handshake->lock rcu_read_lock_bh &table->lock#2 irq_context: 0 (wq_completion)wg-kex-wg0#10 (work_completion)(&({ do { const void *__vpp_verify = (typeof((worker) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __asm__ ("" : "=r"(__ptr) : "0"((typeof(*((worker))) *)((worker)))); (typeof((typeof(*((worker))) *)((worker)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); })->work) rcu_read_lock_bh &obj_hash[i].lock irq_context: 0 (wq_completion)wg-kex-wg0#10 (work_completion)(&({ do { const void *__vpp_verify = (typeof((worker) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __asm__ ("" : "=r"(__ptr) : "0"((typeof(*((worker))) *)((worker)))); (typeof((typeof(*((worker))) *)((worker)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); })->work) rcu_read_lock_bh &base->lock irq_context: 0 (wq_completion)wg-kex-wg0#10 (work_completion)(&({ do { const void *__vpp_verify = (typeof((worker) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __asm__ ("" : "=r"(__ptr) : "0"((typeof(*((worker))) *)((worker)))); (typeof((typeof(*((worker))) *)((worker)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); })->work) rcu_read_lock_bh &base->lock &obj_hash[i].lock irq_context: 0 (wq_completion)wg-kex-wg0#10 (work_completion)(&({ do { const void *__vpp_verify = (typeof((worker) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __asm__ ("" : "=r"(__ptr) : "0"((typeof(*((worker))) *)((worker)))); (typeof((typeof(*((worker))) *)((worker)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); })->work) &obj_hash[i].lock irq_context: 0 (wq_completion)wg-kex-wg0#10 (work_completion)(&({ do { const void *__vpp_verify = (typeof((worker) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __asm__ ("" : "=r"(__ptr) : "0"((typeof(*((worker))) *)((worker)))); (typeof((typeof(*((worker))) *)((worker)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); })->work) &peer->endpoint_lock irq_context: 0 (wq_completion)wg-kex-wg0#10 (work_completion)(&({ do { const void *__vpp_verify = (typeof((worker) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __asm__ ("" : "=r"(__ptr) : "0"((typeof(*((worker))) *)((worker)))); (typeof((typeof(*((worker))) *)((worker)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); })->work) &peer->endpoint_lock rcu_read_lock_bh rcu_read_lock &____s->seqcount#7 irq_context: 0 (wq_completion)wg-kex-wg0#10 (work_completion)(&({ do { const void *__vpp_verify = (typeof((worker) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __asm__ ("" : "=r"(__ptr) : "0"((typeof(*((worker))) *)((worker)))); (typeof((typeof(*((worker))) *)((worker)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); })->work) &peer->endpoint_lock rcu_read_lock_bh rcu_read_lock &____s->seqcount#10 irq_context: 0 (wq_completion)wg-kex-wg0#10 (work_completion)(&({ do { const void *__vpp_verify = (typeof((worker) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __asm__ ("" : "=r"(__ptr) : "0"((typeof(*((worker))) *)((worker)))); (typeof((typeof(*((worker))) *)((worker)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); })->work) &peer->endpoint_lock rcu_read_lock_bh rcu_read_lock rcu_read_lock_bh tk_core.seq.seqcount irq_context: 0 (wq_completion)wg-kex-wg0#10 (work_completion)(&({ do { const void *__vpp_verify = (typeof((worker) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __asm__ ("" : "=r"(__ptr) : "0"((typeof(*((worker))) *)((worker)))); (typeof((typeof(*((worker))) *)((worker)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); })->work) &peer->endpoint_lock rcu_read_lock_bh rcu_read_lock rcu_read_lock_bh &list->lock#5 irq_context: 0 (wq_completion)wg-kex-wg0#10 (work_completion)(&({ do { const void *__vpp_verify = (typeof((worker) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __asm__ ("" : "=r"(__ptr) : "0"((typeof(*((worker))) *)((worker)))); (typeof((typeof(*((worker))) *)((worker)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); })->work) &rq->__lock irq_context: 0 pernet_ops_rwsem napi_hash_lock irq_context: 0 &wg->device_update_lock rcu_state.barrier_mutex rcu_state.barrier_lock irq_context: 0 &wg->device_update_lock rcu_state.barrier_mutex rcu_state.barrier_lock &obj_hash[i].lock irq_context: 0 &wg->device_update_lock rcu_state.barrier_mutex &x->wait#24 irq_context: 0 pernet_ops_rwsem rtnl_mutex &idev->mc_lock _xmit_ETHER &obj_hash[i].lock irq_context: 0 pernet_ops_rwsem rtnl_mutex &idev->mc_lock _xmit_ETHER krc.lock irq_context: 0 pernet_ops_rwsem rtnl_mutex _xmit_ETHER irq_context: 0 pernet_ops_rwsem rtnl_mutex remove_cache_srcu rcu_read_lock rcu_node_0 irq_context: 0 pernet_ops_rwsem rtnl_mutex remove_cache_srcu rcu_read_lock &rq->__lock irq_context: 0 pernet_ops_rwsem rtnl_mutex remove_cache_srcu rcu_read_lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 cb_lock fs_reclaim mmu_notifier_invalidate_range_start &rq->__lock irq_context: 0 cb_lock fs_reclaim mmu_notifier_invalidate_range_start &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 &sb->s_type->i_mutex_key#8 rcu_read_lock rcu_read_lock rcu_node_0 irq_context: 0 &sb->s_type->i_mutex_key#8 rcu_read_lock rcu_read_lock &rq->__lock irq_context: 0 &sb->s_type->i_mutex_key#8 rcu_read_lock rcu_read_lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 &sbi->s_writepages_rwsem jbd2_handle &ei->i_data_sem &c->lock irq_context: 0 &sbi->s_writepages_rwsem jbd2_handle &ei->i_data_sem &n->list_lock irq_context: 0 &sbi->s_writepages_rwsem jbd2_handle &ei->i_data_sem &n->list_lock &c->lock irq_context: 0 &sbi->s_writepages_rwsem jbd2_handle &ei->i_data_sem &ei->i_prealloc_lock irq_context: 0 &sbi->s_writepages_rwsem jbd2_handle &ei->i_data_sem &ei->i_prealloc_lock &pa->pa_lock#2 irq_context: 0 &sbi->s_writepages_rwsem jbd2_handle &ei->i_data_sem key#3 irq_context: 0 &sbi->s_writepages_rwsem jbd2_handle &ei->i_data_sem &pa->pa_lock#2 irq_context: 0 &sbi->s_writepages_rwsem jbd2_handle &ei->i_data_sem &(ei->i_block_reservation_lock) irq_context: 0 &sbi->s_writepages_rwsem jbd2_handle &ei->i_data_sem &(ei->i_block_reservation_lock) key#14 irq_context: 0 &sbi->s_writepages_rwsem jbd2_handle &ei->i_data_sem &sb->s_type->i_lock_key#22 irq_context: 0 &sbi->s_writepages_rwsem jbd2_handle &ei->i_data_sem &ei->i_es_lock &obj_hash[i].lock irq_context: 0 &sbi->s_writepages_rwsem jbd2_handle &ei->i_data_sem &ei->i_es_lock pool_lock#2 irq_context: 0 &sbi->s_writepages_rwsem jbd2_handle rcu_read_lock key#10 irq_context: 0 &sbi->s_writepages_rwsem jbd2_handle rcu_read_lock &xa->xa_lock#9 key#10 irq_context: 0 &sbi->s_writepages_rwsem jbd2_handle &ei->i_raw_lock irq_context: 0 &sbi->s_writepages_rwsem rcu_read_lock pool_lock#2 irq_context: 0 pernet_ops_rwsem rtnl_mutex dev_base_lock &xa->xa_lock#3 pool_lock#2 irq_context: 0 pernet_ops_rwsem rtnl_mutex _xmit_LOOPBACK irq_context: 0 pernet_ops_rwsem &hn->hn_lock irq_context: 0 pernet_ops_rwsem sysctl_lock &obj_hash[i].lock irq_context: 0 pernet_ops_rwsem sysctl_lock pool_lock#2 irq_context: 0 pernet_ops_rwsem sysctl_lock krc.lock irq_context: 0 pernet_ops_rwsem rcu_read_lock &pool->lock pool_lock#2 irq_context: 0 pernet_ops_rwsem rcu_node_0 irq_context: 0 (wq_completion)inet_frag_wq irq_context: 0 (wq_completion)inet_frag_wq (work_completion)(&fqdir->destroy_work) irq_context: 0 (wq_completion)inet_frag_wq (work_completion)(&fqdir->destroy_work) (work_completion)(&ht->run_work) irq_context: 0 (wq_completion)inet_frag_wq (work_completion)(&fqdir->destroy_work) &ht->mutex irq_context: 0 (wq_completion)inet_frag_wq (work_completion)(&fqdir->destroy_work) &ht->mutex &obj_hash[i].lock irq_context: 0 (wq_completion)inet_frag_wq (work_completion)(&fqdir->destroy_work) &ht->mutex pool_lock#2 irq_context: 0 (wq_completion)inet_frag_wq (work_completion)(&fqdir->destroy_work) rcu_read_lock &pool->lock irq_context: 0 (wq_completion)inet_frag_wq (work_completion)(&fqdir->destroy_work) rcu_read_lock &pool->lock &obj_hash[i].lock irq_context: 0 (wq_completion)events fqdir_free_work irq_context: 0 (wq_completion)events fqdir_free_work rcu_state.barrier_mutex irq_context: 0 (wq_completion)wg-kex-wg2#5 irq_context: 0 (wq_completion)wg-kex-wg2#5 (work_completion)(&peer->transmit_handshake_work) irq_context: 0 (wq_completion)wg-kex-wg2#5 (work_completion)(&peer->transmit_handshake_work) tk_core.seq.seqcount irq_context: 0 (wq_completion)wg-kex-wg2#5 (work_completion)(&peer->transmit_handshake_work) &wg->static_identity.lock irq_context: 0 (wq_completion)wg-kex-wg2#5 (work_completion)(&peer->transmit_handshake_work) &wg->static_identity.lock &handshake->lock irq_context: 0 (wq_completion)wg-kex-wg2#5 (work_completion)(&peer->transmit_handshake_work) &wg->static_identity.lock &handshake->lock crngs.lock irq_context: 0 (wq_completion)wg-kex-wg2#5 (work_completion)(&peer->transmit_handshake_work) &wg->static_identity.lock &handshake->lock &rq->__lock irq_context: 0 (wq_completion)wg-crypt-wg0#4 irq_context: 0 (wq_completion)wg-crypt-wg0#4 &rq->__lock irq_context: 0 (wq_completion)wg-crypt-wg0#4 (work_completion)(&({ do { const void *__vpp_verify = (typeof((worker) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __asm__ ("" : "=r"(__ptr) : "0"((typeof(*((worker))) *)((worker)))); (typeof((typeof(*((worker))) *)((worker)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); })->work) irq_context: 0 (wq_completion)wg-crypt-wg0#4 (work_completion)(&({ do { const void *__vpp_verify = (typeof((worker) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __asm__ ("" : "=r"(__ptr) : "0"((typeof(*((worker))) *)((worker)))); (typeof((typeof(*((worker))) *)((worker)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); })->work) &r->consumer_lock#2 irq_context: 0 (wq_completion)wg-crypt-wg0#4 (work_completion)(&({ do { const void *__vpp_verify = (typeof((worker) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __asm__ ("" : "=r"(__ptr) : "0"((typeof(*((worker))) *)((worker)))); (typeof((typeof(*((worker))) *)((worker)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); })->work) tk_core.seq.seqcount irq_context: 0 (wq_completion)wg-crypt-wg0#4 (work_completion)(&({ do { const void *__vpp_verify = (typeof((worker) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __asm__ ("" : "=r"(__ptr) : "0"((typeof(*((worker))) *)((worker)))); (typeof((typeof(*((worker))) *)((worker)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); })->work) &rq->__lock irq_context: 0 (wq_completion)wg-kex-wg2#5 (work_completion)(&peer->transmit_handshake_work) &wg->static_identity.lock &handshake->lock tk_core.seq.seqcount irq_context: 0 (wq_completion)wg-crypt-wg0#4 (work_completion)(&({ do { const void *__vpp_verify = (typeof((worker) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __asm__ ("" : "=r"(__ptr) : "0"((typeof(*((worker))) *)((worker)))); (typeof((typeof(*((worker))) *)((worker)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); })->work) rcu_read_lock &pool->lock irq_context: 0 (wq_completion)wg-crypt-wg0#4 (work_completion)(&({ do { const void *__vpp_verify = (typeof((worker) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __asm__ ("" : "=r"(__ptr) : "0"((typeof(*((worker))) *)((worker)))); (typeof((typeof(*((worker))) *)((worker)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); })->work) rcu_read_lock &pool->lock &obj_hash[i].lock irq_context: 0 (wq_completion)wg-kex-wg2#5 (work_completion)(&peer->transmit_handshake_work) &wg->static_identity.lock &handshake->lock &table->lock#2 irq_context: 0 (wq_completion)wg-crypt-wg0#4 (work_completion)(&({ do { const void *__vpp_verify = (typeof((worker) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __asm__ ("" : "=r"(__ptr) : "0"((typeof(*((worker))) *)((worker)))); (typeof((typeof(*((worker))) *)((worker)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); })->work) rcu_read_lock &pool->lock &p->pi_lock irq_context: 0 (wq_completion)wg-kex-wg2#5 (work_completion)(&peer->transmit_handshake_work) &wg->static_identity.lock &handshake->lock rcu_read_lock_bh batched_entropy_u32.lock irq_context: 0 (wq_completion)wg-kex-wg2#5 (work_completion)(&peer->transmit_handshake_work) &wg->static_identity.lock &handshake->lock rcu_read_lock_bh &table->lock#2 irq_context: 0 (wq_completion)wg-kex-wg2#5 (work_completion)(&peer->transmit_handshake_work) &cookie->lock irq_context: 0 (wq_completion)wg-kex-wg2#5 (work_completion)(&peer->transmit_handshake_work) &cookie->lock irq_context: 0 (wq_completion)wg-kex-wg2#5 (work_completion)(&peer->transmit_handshake_work) rcu_read_lock_bh &obj_hash[i].lock irq_context: 0 (wq_completion)wg-kex-wg2#5 (work_completion)(&peer->transmit_handshake_work) rcu_read_lock_bh &base->lock irq_context: 0 (wq_completion)wg-kex-wg2#5 (work_completion)(&peer->transmit_handshake_work) rcu_read_lock_bh &base->lock &obj_hash[i].lock irq_context: 0 (wq_completion)wg-kex-wg2#5 (work_completion)(&peer->transmit_handshake_work) &rq->__lock irq_context: 0 (wq_completion)wg-crypt-wg0#4 (work_completion)(&peer->transmit_packet_work) irq_context: 0 (wq_completion)wg-crypt-wg0#4 (work_completion)(&peer->transmit_packet_work) rcu_read_lock_bh &obj_hash[i].lock irq_context: 0 (wq_completion)wg-crypt-wg0#4 (work_completion)(&peer->transmit_packet_work) rcu_read_lock_bh &base->lock irq_context: 0 (wq_completion)wg-crypt-wg0#4 (work_completion)(&peer->transmit_packet_work) &obj_hash[i].lock irq_context: 0 (wq_completion)wg-crypt-wg0#4 (work_completion)(&peer->transmit_packet_work) &peer->endpoint_lock irq_context: 0 (wq_completion)wg-crypt-wg0#4 (work_completion)(&peer->transmit_packet_work) &peer->endpoint_lock rcu_read_lock_bh &obj_hash[i].lock irq_context: 0 (wq_completion)wg-crypt-wg0#4 (work_completion)(&peer->transmit_packet_work) &peer->endpoint_lock rcu_read_lock_bh rcu_read_lock &____s->seqcount#7 irq_context: 0 (wq_completion)wg-crypt-wg0#4 (work_completion)(&peer->transmit_packet_work) &peer->endpoint_lock rcu_read_lock_bh rcu_read_lock &____s->seqcount#10 irq_context: 0 (wq_completion)wg-crypt-wg0#4 (work_completion)(&peer->transmit_packet_work) &peer->endpoint_lock rcu_read_lock_bh rcu_read_lock rcu_read_lock_bh tk_core.seq.seqcount irq_context: 0 (wq_completion)wg-crypt-wg0#4 (work_completion)(&peer->transmit_packet_work) &peer->endpoint_lock rcu_read_lock_bh rcu_read_lock rcu_read_lock_bh &list->lock#5 irq_context: 0 (wq_completion)events fqdir_free_work rcu_state.barrier_mutex rcu_state.barrier_lock irq_context: 0 (wq_completion)events fqdir_free_work rcu_state.barrier_mutex &x->wait#24 irq_context: 0 (wq_completion)wg-kex-wg2#5 (work_completion)(&peer->transmit_handshake_work) &obj_hash[i].lock irq_context: 0 (wq_completion)wg-kex-wg2#5 (work_completion)(&peer->transmit_handshake_work) &peer->endpoint_lock irq_context: 0 (wq_completion)wg-kex-wg2#5 (work_completion)(&peer->transmit_handshake_work) &peer->endpoint_lock rcu_read_lock_bh rcu_read_lock &____s->seqcount#7 irq_context: 0 (wq_completion)wg-kex-wg2#5 (work_completion)(&peer->transmit_handshake_work) &peer->endpoint_lock rcu_read_lock_bh rcu_read_lock &____s->seqcount#10 irq_context: 0 (wq_completion)wg-kex-wg2#5 (work_completion)(&peer->transmit_handshake_work) &peer->endpoint_lock rcu_read_lock_bh rcu_read_lock rcu_read_lock_bh tk_core.seq.seqcount irq_context: 0 (wq_completion)wg-kex-wg2#5 (work_completion)(&peer->transmit_handshake_work) &peer->endpoint_lock rcu_read_lock_bh rcu_read_lock rcu_read_lock_bh &list->lock#5 irq_context: 0 (wq_completion)wg-kex-wg2#5 (work_completion)(&peer->transmit_handshake_work) batched_entropy_u8.lock irq_context: 0 (wq_completion)wg-kex-wg2#6 irq_context: 0 (wq_completion)wg-kex-wg2#6 (work_completion)(&({ do { const void *__vpp_verify = (typeof((worker) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __asm__ ("" : "=r"(__ptr) : "0"((typeof(*((worker))) *)((worker)))); (typeof((typeof(*((worker))) *)((worker)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); })->work) irq_context: 0 (wq_completion)wg-kex-wg2#6 (work_completion)(&({ do { const void *__vpp_verify = (typeof((worker) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __asm__ ("" : "=r"(__ptr) : "0"((typeof(*((worker))) *)((worker)))); (typeof((typeof(*((worker))) *)((worker)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); })->work) &r->consumer_lock#2 irq_context: 0 (wq_completion)wg-kex-wg2#6 (work_completion)(&({ do { const void *__vpp_verify = (typeof((worker) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __asm__ ("" : "=r"(__ptr) : "0"((typeof(*((worker))) *)((worker)))); (typeof((typeof(*((worker))) *)((worker)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); })->work) &wg->static_identity.lock irq_context: 0 (wq_completion)wg-kex-wg2#6 (work_completion)(&({ do { const void *__vpp_verify = (typeof((worker) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __asm__ ("" : "=r"(__ptr) : "0"((typeof(*((worker))) *)((worker)))); (typeof((typeof(*((worker))) *)((worker)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); })->work) &wg->static_identity.lock &handshake->lock irq_context: 0 (wq_completion)wg-kex-wg2#5 (work_completion)(&peer->transmit_handshake_work) &peer->endpoint_lock rcu_read_lock_bh rcu_read_lock &dir->lock#2 irq_context: 0 (wq_completion)wg-kex-wg2#5 (work_completion)(&peer->transmit_handshake_work) &peer->endpoint_lock rcu_read_lock_bh rcu_read_lock &ul->lock irq_context: 0 (wq_completion)wg-kex-wg2#5 (work_completion)(&peer->transmit_handshake_work) &peer->endpoint_lock rcu_read_lock_bh rcu_read_lock &obj_hash[i].lock irq_context: 0 (wq_completion)wg-kex-wg2#6 (work_completion)(&({ do { const void *__vpp_verify = (typeof((worker) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __asm__ ("" : "=r"(__ptr) : "0"((typeof(*((worker))) *)((worker)))); (typeof((typeof(*((worker))) *)((worker)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); })->work) &wg->static_identity.lock &handshake->lock irq_context: 0 (wq_completion)wg-kex-wg2#5 (work_completion)(&peer->transmit_handshake_work) &peer->endpoint_lock rcu_read_lock_bh rcu_read_lock pool_lock#2 irq_context: 0 (wq_completion)wg-kex-wg2#6 (work_completion)(&({ do { const void *__vpp_verify = (typeof((worker) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __asm__ ("" : "=r"(__ptr) : "0"((typeof(*((worker))) *)((worker)))); (typeof((typeof(*((worker))) *)((worker)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); })->work) &peer->endpoint_lock irq_context: 0 (wq_completion)wg-kex-wg2#5 (work_completion)(&peer->transmit_handshake_work) &peer->endpoint_lock rcu_read_lock_bh batched_entropy_u32.lock irq_context: 0 (wq_completion)wg-kex-wg2#6 (work_completion)(&({ do { const void *__vpp_verify = (typeof((worker) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __asm__ ("" : "=r"(__ptr) : "0"((typeof(*((worker))) *)((worker)))); (typeof((typeof(*((worker))) *)((worker)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); })->work) &handshake->lock irq_context: 0 (wq_completion)wg-kex-wg2#6 (work_completion)(&({ do { const void *__vpp_verify = (typeof((worker) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __asm__ ("" : "=r"(__ptr) : "0"((typeof(*((worker))) *)((worker)))); (typeof((typeof(*((worker))) *)((worker)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); })->work) &handshake->lock fs_reclaim irq_context: 0 (wq_completion)wg-kex-wg2#6 (work_completion)(&({ do { const void *__vpp_verify = (typeof((worker) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __asm__ ("" : "=r"(__ptr) : "0"((typeof(*((worker))) *)((worker)))); (typeof((typeof(*((worker))) *)((worker)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); })->work) &handshake->lock fs_reclaim mmu_notifier_invalidate_range_start irq_context: 0 (wq_completion)wg-kex-wg2#6 (work_completion)(&({ do { const void *__vpp_verify = (typeof((worker) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __asm__ ("" : "=r"(__ptr) : "0"((typeof(*((worker))) *)((worker)))); (typeof((typeof(*((worker))) *)((worker)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); })->work) &handshake->lock tk_core.seq.seqcount irq_context: 0 (wq_completion)wg-kex-wg2#6 (work_completion)(&({ do { const void *__vpp_verify = (typeof((worker) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __asm__ ("" : "=r"(__ptr) : "0"((typeof(*((worker))) *)((worker)))); (typeof((typeof(*((worker))) *)((worker)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); })->work) &handshake->lock rcu_read_lock_bh &peer->keypairs.keypair_update_lock irq_context: 0 (wq_completion)wg-kex-wg2#6 (work_completion)(&({ do { const void *__vpp_verify = (typeof((worker) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __asm__ ("" : "=r"(__ptr) : "0"((typeof(*((worker))) *)((worker)))); (typeof((typeof(*((worker))) *)((worker)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); })->work) &handshake->lock rcu_read_lock_bh &table->lock#2 irq_context: 0 (wq_completion)wg-kex-wg2#6 (work_completion)(&({ do { const void *__vpp_verify = (typeof((worker) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __asm__ ("" : "=r"(__ptr) : "0"((typeof(*((worker))) *)((worker)))); (typeof((typeof(*((worker))) *)((worker)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); })->work) rcu_read_lock_bh &obj_hash[i].lock irq_context: 0 (wq_completion)wg-kex-wg2#6 (work_completion)(&({ do { const void *__vpp_verify = (typeof((worker) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __asm__ ("" : "=r"(__ptr) : "0"((typeof(*((worker))) *)((worker)))); (typeof((typeof(*((worker))) *)((worker)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); })->work) rcu_read_lock_bh &base->lock irq_context: 0 (wq_completion)wg-kex-wg1#6 (work_completion)(&({ do { const void *__vpp_verify = (typeof((worker) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __asm__ ("" : "=r"(__ptr) : "0"((typeof(*((worker))) *)((worker)))); (typeof((typeof(*((worker))) *)((worker)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); })->work) &wg->static_identity.lock &handshake->lock tk_core.seq.seqcount irq_context: 0 (wq_completion)wg-kex-wg2#6 (work_completion)(&({ do { const void *__vpp_verify = (typeof((worker) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __asm__ ("" : "=r"(__ptr) : "0"((typeof(*((worker))) *)((worker)))); (typeof((typeof(*((worker))) *)((worker)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); })->work) rcu_read_lock_bh &base->lock &obj_hash[i].lock irq_context: 0 (wq_completion)wg-kex-wg1#6 (work_completion)(&({ do { const void *__vpp_verify = (typeof((worker) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __asm__ ("" : "=r"(__ptr) : "0"((typeof(*((worker))) *)((worker)))); (typeof((typeof(*((worker))) *)((worker)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); })->work) &wg->static_identity.lock &handshake->lock tk_core.seq.seqcount irq_context: 0 (wq_completion)wg-kex-wg2#6 (work_completion)(&({ do { const void *__vpp_verify = (typeof((worker) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __asm__ ("" : "=r"(__ptr) : "0"((typeof(*((worker))) *)((worker)))); (typeof((typeof(*((worker))) *)((worker)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); })->work) &obj_hash[i].lock irq_context: 0 (wq_completion)wg-kex-wg1#6 (work_completion)(&({ do { const void *__vpp_verify = (typeof((worker) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __asm__ ("" : "=r"(__ptr) : "0"((typeof(*((worker))) *)((worker)))); (typeof((typeof(*((worker))) *)((worker)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); })->work) &wg->static_identity.lock &handshake->lock crngs.lock irq_context: 0 (wq_completion)wg-kex-wg2#6 (work_completion)(&({ do { const void *__vpp_verify = (typeof((worker) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __asm__ ("" : "=r"(__ptr) : "0"((typeof(*((worker))) *)((worker)))); (typeof((typeof(*((worker))) *)((worker)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); })->work) &base->lock irq_context: 0 (wq_completion)wg-kex-wg2#6 (work_completion)(&({ do { const void *__vpp_verify = (typeof((worker) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __asm__ ("" : "=r"(__ptr) : "0"((typeof(*((worker))) *)((worker)))); (typeof((typeof(*((worker))) *)((worker)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); })->work) &base->lock &obj_hash[i].lock irq_context: 0 (wq_completion)wg-kex-wg2#6 (work_completion)(&({ do { const void *__vpp_verify = (typeof((worker) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __asm__ ("" : "=r"(__ptr) : "0"((typeof(*((worker))) *)((worker)))); (typeof((typeof(*((worker))) *)((worker)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); })->work) tk_core.seq.seqcount irq_context: 0 (wq_completion)wg-kex-wg2#6 (work_completion)(&({ do { const void *__vpp_verify = (typeof((worker) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __asm__ ("" : "=r"(__ptr) : "0"((typeof(*((worker))) *)((worker)))); (typeof((typeof(*((worker))) *)((worker)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); })->work) &list->lock#17 irq_context: 0 (wq_completion)wg-kex-wg2#6 (work_completion)(&({ do { const void *__vpp_verify = (typeof((worker) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __asm__ ("" : "=r"(__ptr) : "0"((typeof(*((worker))) *)((worker)))); (typeof((typeof(*((worker))) *)((worker)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); })->work) rcu_read_lock_bh &r->producer_lock#2 irq_context: 0 (wq_completion)wg-kex-wg2#6 (work_completion)(&({ do { const void *__vpp_verify = (typeof((worker) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __asm__ ("" : "=r"(__ptr) : "0"((typeof(*((worker))) *)((worker)))); (typeof((typeof(*((worker))) *)((worker)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); })->work) rcu_read_lock_bh rcu_read_lock &pool->lock irq_context: 0 (wq_completion)wg-kex-wg2#6 (work_completion)(&({ do { const void *__vpp_verify = (typeof((worker) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __asm__ ("" : "=r"(__ptr) : "0"((typeof(*((worker))) *)((worker)))); (typeof((typeof(*((worker))) *)((worker)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); })->work) rcu_read_lock_bh rcu_read_lock &pool->lock &obj_hash[i].lock irq_context: 0 (wq_completion)wg-kex-wg2#6 (work_completion)(&({ do { const void *__vpp_verify = (typeof((worker) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __asm__ ("" : "=r"(__ptr) : "0"((typeof(*((worker))) *)((worker)))); (typeof((typeof(*((worker))) *)((worker)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); })->work) rcu_read_lock_bh rcu_read_lock &pool->lock &p->pi_lock irq_context: 0 (wq_completion)wg-kex-wg1#6 (work_completion)(&({ do { const void *__vpp_verify = (typeof((worker) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __asm__ ("" : "=r"(__ptr) : "0"((typeof(*((worker))) *)((worker)))); (typeof((typeof(*((worker))) *)((worker)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); })->work) &wg->static_identity.lock &handshake->lock &table->lock#2 irq_context: 0 (wq_completion)wg-kex-wg2#6 (work_completion)(&({ do { const void *__vpp_verify = (typeof((worker) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __asm__ ("" : "=r"(__ptr) : "0"((typeof(*((worker))) *)((worker)))); (typeof((typeof(*((worker))) *)((worker)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); })->work) rcu_read_lock_bh rcu_read_lock &pool->lock &p->pi_lock &rq->__lock irq_context: 0 (wq_completion)wg-kex-wg1#6 (work_completion)(&({ do { const void *__vpp_verify = (typeof((worker) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __asm__ ("" : "=r"(__ptr) : "0"((typeof(*((worker))) *)((worker)))); (typeof((typeof(*((worker))) *)((worker)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); })->work) &wg->static_identity.lock &handshake->lock rcu_read_lock_bh batched_entropy_u32.lock irq_context: 0 (wq_completion)wg-kex-wg2#6 (work_completion)(&({ do { const void *__vpp_verify = (typeof((worker) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __asm__ ("" : "=r"(__ptr) : "0"((typeof(*((worker))) *)((worker)))); (typeof((typeof(*((worker))) *)((worker)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); })->work) rcu_read_lock_bh rcu_read_lock &pool->lock &p->pi_lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)wg-kex-wg1#6 (work_completion)(&({ do { const void *__vpp_verify = (typeof((worker) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __asm__ ("" : "=r"(__ptr) : "0"((typeof(*((worker))) *)((worker)))); (typeof((typeof(*((worker))) *)((worker)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); })->work) &wg->static_identity.lock &handshake->lock rcu_read_lock_bh &table->lock#2 irq_context: 0 (wq_completion)wg-kex-wg2#6 (work_completion)(&({ do { const void *__vpp_verify = (typeof((worker) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __asm__ ("" : "=r"(__ptr) : "0"((typeof(*((worker))) *)((worker)))); (typeof((typeof(*((worker))) *)((worker)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); })->work) &rq->__lock irq_context: 0 (wq_completion)wg-kex-wg1#6 (work_completion)(&({ do { const void *__vpp_verify = (typeof((worker) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __asm__ ("" : "=r"(__ptr) : "0"((typeof(*((worker))) *)((worker)))); (typeof((typeof(*((worker))) *)((worker)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); })->work) &cookie->lock irq_context: 0 (wq_completion)wg-kex-wg1#6 (work_completion)(&({ do { const void *__vpp_verify = (typeof((worker) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __asm__ ("" : "=r"(__ptr) : "0"((typeof(*((worker))) *)((worker)))); (typeof((typeof(*((worker))) *)((worker)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); })->work) &cookie->lock irq_context: 0 (wq_completion)wg-kex-wg1#6 (work_completion)(&({ do { const void *__vpp_verify = (typeof((worker) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __asm__ ("" : "=r"(__ptr) : "0"((typeof(*((worker))) *)((worker)))); (typeof((typeof(*((worker))) *)((worker)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); })->work) &peer->endpoint_lock irq_context: 0 (wq_completion)wg-kex-wg1#6 (work_completion)(&({ do { const void *__vpp_verify = (typeof((worker) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __asm__ ("" : "=r"(__ptr) : "0"((typeof(*((worker))) *)((worker)))); (typeof((typeof(*((worker))) *)((worker)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); })->work) &peer->endpoint_lock rcu_read_lock_bh rcu_read_lock &____s->seqcount#7 irq_context: 0 (wq_completion)wg-kex-wg1#6 (work_completion)(&({ do { const void *__vpp_verify = (typeof((worker) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __asm__ ("" : "=r"(__ptr) : "0"((typeof(*((worker))) *)((worker)))); (typeof((typeof(*((worker))) *)((worker)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); })->work) &peer->endpoint_lock rcu_read_lock_bh rcu_read_lock &____s->seqcount#10 irq_context: 0 (wq_completion)wg-kex-wg1#6 (work_completion)(&({ do { const void *__vpp_verify = (typeof((worker) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __asm__ ("" : "=r"(__ptr) : "0"((typeof(*((worker))) *)((worker)))); (typeof((typeof(*((worker))) *)((worker)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); })->work) &peer->endpoint_lock rcu_read_lock_bh rcu_read_lock rcu_read_lock_bh tk_core.seq.seqcount irq_context: 0 (wq_completion)wg-kex-wg1#6 (work_completion)(&({ do { const void *__vpp_verify = (typeof((worker) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __asm__ ("" : "=r"(__ptr) : "0"((typeof(*((worker))) *)((worker)))); (typeof((typeof(*((worker))) *)((worker)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); })->work) &peer->endpoint_lock rcu_read_lock_bh rcu_read_lock rcu_read_lock_bh &list->lock#5 irq_context: 0 (wq_completion)wg-crypt-wg2#3 irq_context: 0 (wq_completion)wg-crypt-wg2#3 (work_completion)(&({ do { const void *__vpp_verify = (typeof((worker) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __asm__ ("" : "=r"(__ptr) : "0"((typeof(*((worker))) *)((worker)))); (typeof((typeof(*((worker))) *)((worker)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); })->work) irq_context: 0 (wq_completion)wg-crypt-wg2#3 (work_completion)(&({ do { const void *__vpp_verify = (typeof((worker) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __asm__ ("" : "=r"(__ptr) : "0"((typeof(*((worker))) *)((worker)))); (typeof((typeof(*((worker))) *)((worker)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); })->work) &r->consumer_lock#2 irq_context: 0 (wq_completion)wg-crypt-wg2#3 (work_completion)(&({ do { const void *__vpp_verify = (typeof((worker) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __asm__ ("" : "=r"(__ptr) : "0"((typeof(*((worker))) *)((worker)))); (typeof((typeof(*((worker))) *)((worker)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); })->work) rcu_read_lock &pool->lock irq_context: 0 (wq_completion)wg-crypt-wg2#3 (work_completion)(&({ do { const void *__vpp_verify = (typeof((worker) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __asm__ ("" : "=r"(__ptr) : "0"((typeof(*((worker))) *)((worker)))); (typeof((typeof(*((worker))) *)((worker)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); })->work) rcu_read_lock &pool->lock &obj_hash[i].lock irq_context: 0 (wq_completion)wg-crypt-wg2#3 (work_completion)(&({ do { const void *__vpp_verify = (typeof((worker) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __asm__ ("" : "=r"(__ptr) : "0"((typeof(*((worker))) *)((worker)))); (typeof((typeof(*((worker))) *)((worker)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); })->work) rcu_read_lock &pool->lock &p->pi_lock irq_context: 0 (wq_completion)wg-kex-wg2#6 (work_completion)(&({ do { const void *__vpp_verify = (typeof((worker) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __asm__ ("" : "=r"(__ptr) : "0"((typeof(*((worker))) *)((worker)))); (typeof((typeof(*((worker))) *)((worker)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); })->work) &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)wg-crypt-wg2#3 (work_completion)(&({ do { const void *__vpp_verify = (typeof((worker) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __asm__ ("" : "=r"(__ptr) : "0"((typeof(*((worker))) *)((worker)))); (typeof((typeof(*((worker))) *)((worker)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); })->work) rcu_read_lock &pool->lock &p->pi_lock &rq->__lock irq_context: 0 (wq_completion)wg-crypt-wg2#3 (work_completion)(&({ do { const void *__vpp_verify = (typeof((worker) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __asm__ ("" : "=r"(__ptr) : "0"((typeof(*((worker))) *)((worker)))); (typeof((typeof(*((worker))) *)((worker)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); })->work) rcu_read_lock &pool->lock &p->pi_lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)wg-crypt-wg2#3 (work_completion)(&peer->transmit_packet_work) irq_context: 0 (wq_completion)wg-crypt-wg2#3 (work_completion)(&peer->transmit_packet_work) rcu_read_lock_bh &obj_hash[i].lock irq_context: 0 (wq_completion)wg-crypt-wg2#3 (work_completion)(&peer->transmit_packet_work) &obj_hash[i].lock irq_context: 0 (wq_completion)wg-crypt-wg2#3 (work_completion)(&peer->transmit_packet_work) &peer->endpoint_lock irq_context: 0 (wq_completion)wg-crypt-wg2#3 (work_completion)(&peer->transmit_packet_work) &peer->endpoint_lock rcu_read_lock_bh rcu_read_lock &____s->seqcount#7 irq_context: 0 (wq_completion)wg-crypt-wg2#3 (work_completion)(&peer->transmit_packet_work) &peer->endpoint_lock rcu_read_lock_bh rcu_read_lock &____s->seqcount#10 irq_context: 0 (wq_completion)wg-crypt-wg2#3 (work_completion)(&peer->transmit_packet_work) &peer->endpoint_lock rcu_read_lock_bh rcu_read_lock rcu_read_lock_bh tk_core.seq.seqcount irq_context: 0 (wq_completion)wg-crypt-wg2#3 (work_completion)(&peer->transmit_packet_work) &peer->endpoint_lock rcu_read_lock_bh rcu_read_lock rcu_read_lock_bh &list->lock#5 irq_context: 0 (wq_completion)wg-crypt-wg2#3 (work_completion)(&peer->transmit_packet_work) &rq->__lock irq_context: 0 (wq_completion)wg-crypt-wg2#3 (work_completion)(&peer->transmit_packet_work) rcu_read_lock_bh tk_core.seq.seqcount irq_context: 0 (wq_completion)ipv6_addrconf (work_completion)(&(&ifa->dad_work)->work) rtnl_mutex &idev->mc_lock rcu_read_lock rcu_read_lock rcu_read_lock_bh dev->qdisc_tx_busylock ?: &qdisc_tx_busylock#2 rcu_read_lock_bh &obj_hash[i].lock irq_context: 0 (wq_completion)ipv6_addrconf (work_completion)(&(&ifa->dad_work)->work) rtnl_mutex &idev->mc_lock rcu_read_lock rcu_read_lock rcu_read_lock_bh dev->qdisc_tx_busylock ?: &qdisc_tx_busylock#2 rcu_read_lock_bh pool_lock#2 irq_context: 0 (wq_completion)ipv6_addrconf (work_completion)(&(&ifa->dad_work)->work) rtnl_mutex &idev->mc_lock rcu_read_lock rcu_read_lock rcu_read_lock_bh dev->qdisc_tx_busylock ?: &qdisc_tx_busylock#2 rcu_read_lock_bh rcu_read_lock tk_core.seq.seqcount irq_context: 0 (wq_completion)ipv6_addrconf (work_completion)(&(&ifa->dad_work)->work) rtnl_mutex &idev->mc_lock rcu_read_lock rcu_read_lock rcu_read_lock_bh dev->qdisc_tx_busylock ?: &qdisc_tx_busylock#2 rcu_read_lock_bh rcu_read_lock &list->lock#5 irq_context: 0 (wq_completion)wg-crypt-wg2#3 (work_completion)(&peer->transmit_packet_work) &peer->endpoint_lock rcu_read_lock_bh rcu_read_lock &dir->lock#2 irq_context: 0 (wq_completion)wg-crypt-wg2#3 (work_completion)(&peer->transmit_packet_work) &peer->endpoint_lock rcu_read_lock_bh rcu_read_lock &ul->lock irq_context: 0 (wq_completion)wg-crypt-wg2#3 (work_completion)(&({ do { const void *__vpp_verify = (typeof((worker) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __asm__ ("" : "=r"(__ptr) : "0"((typeof(*((worker))) *)((worker)))); (typeof((typeof(*((worker))) *)((worker)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); })->work) tk_core.seq.seqcount irq_context: 0 (wq_completion)wg-crypt-wg1#3 (work_completion)(&peer->transmit_packet_work) rcu_read_lock_bh &base->lock irq_context: 0 (wq_completion)wg-crypt-wg1#3 (work_completion)(&peer->transmit_packet_work) &peer->endpoint_lock rcu_read_lock_bh &obj_hash[i].lock irq_context: 0 (wq_completion)wg-crypt-wg1#3 (work_completion)(&peer->transmit_packet_work) &peer->endpoint_lock rcu_read_lock_bh batched_entropy_u32.lock irq_context: 0 &type->i_mutex_dir_key#5 &____s->seqcount#2 irq_context: 0 sb_writers#10 &type->i_mutex_dir_key#6/1 cgroup_mutex lock kernfs_idr_lock &____s->seqcount irq_context: 0 sb_writers#10 &type->i_mutex_dir_key#6/1 cgroup_mutex lock kernfs_idr_lock pool_lock#2 irq_context: 0 sb_writers#10 &type->i_mutex_dir_key#6/1 cgroup_mutex lock kernfs_idr_lock rcu_read_lock pool_lock#2 irq_context: 0 sb_writers#10 &type->i_mutex_dir_key#6/1 cgroup_mutex lock kernfs_idr_lock &obj_hash[i].lock irq_context: 0 sb_writers#10 &type->i_mutex_dir_key#6/1 cgroup_mutex &____s->seqcount#2 irq_context: 0 sb_writers#10 &type->i_mutex_dir_key#6/1 cgroup_mutex batched_entropy_u32.lock crngs.lock irq_context: 0 sb_writers#10 &type->i_mutex_dir_key#6/1 cgroup_mutex &n->list_lock irq_context: 0 sb_writers#10 &type->i_mutex_dir_key#6/1 cgroup_mutex &n->list_lock &c->lock irq_context: 0 (wq_completion)events (work_completion)(&rdev->wiphy_work) &rdev->wiphy.mtx &wdev->mtx &local->sta_mtx &sb->s_type->i_mutex_key#3 &____s->seqcount#2 irq_context: 0 (wq_completion)events (work_completion)(&rdev->wiphy_work) &rdev->wiphy.mtx &wdev->mtx &local->sta_mtx &sb->s_type->i_mutex_key#3 &pcp->lock &zone->lock irq_context: 0 (wq_completion)events (work_completion)(&rdev->wiphy_work) &rdev->wiphy.mtx &wdev->mtx &local->sta_mtx &sb->s_type->i_mutex_key#3 &pcp->lock &zone->lock &____s->seqcount irq_context: 0 (wq_completion)events (work_completion)(&rdev->wiphy_work) &rdev->wiphy.mtx &wdev->mtx &local->sta_mtx &c->lock irq_context: 0 (wq_completion)events (work_completion)(&rdev->wiphy_work) &rdev->wiphy.mtx &wdev->mtx &local->sta_mtx &____s->seqcount#2 irq_context: 0 (wq_completion)events (work_completion)(&rdev->wiphy_work) &rdev->wiphy.mtx &wdev->mtx &local->sta_mtx &pcp->lock &zone->lock irq_context: 0 (wq_completion)events (work_completion)(&rdev->wiphy_work) &rdev->wiphy.mtx &wdev->mtx &local->sta_mtx &____s->seqcount irq_context: 0 sb_writers#10 &type->i_mutex_dir_key#6 &c->lock irq_context: 0 sb_writers#10 &type->i_mutex_dir_key#6 &____s->seqcount#2 irq_context: 0 sb_writers#10 &type->i_mutex_dir_key#6 &____s->seqcount irq_context: 0 sb_writers#10 &of->mutex cgroup_mutex cpu_hotplug_lock cgroup_threadgroup_rwsem rcu_state.exp_mutex &rnp->exp_wq[2] irq_context: 0 (wq_completion)events (work_completion)(&rdev->wiphy_work) &rdev->wiphy.mtx &wdev->mtx fill_pool_map-wait-type-override pool_lock#2 irq_context: 0 (wq_completion)events (work_completion)(&rdev->wiphy_work) &rdev->wiphy.mtx &wdev->mtx fill_pool_map-wait-type-override &c->lock irq_context: 0 (wq_completion)events (work_completion)(&rdev->wiphy_work) &rdev->wiphy.mtx &wdev->mtx fill_pool_map-wait-type-override &____s->seqcount#2 irq_context: 0 (wq_completion)events (work_completion)(&rdev->wiphy_work) &rdev->wiphy.mtx &wdev->mtx fill_pool_map-wait-type-override &____s->seqcount irq_context: 0 (wq_completion)events (work_completion)(&rdev->wiphy_work) &rdev->wiphy.mtx &wdev->mtx fill_pool_map-wait-type-override pool_lock irq_context: 0 rtnl_mutex rcu_read_lock_bh fill_pool_map-wait-type-override pool_lock#2 irq_context: 0 rtnl_mutex rcu_read_lock_bh fill_pool_map-wait-type-override pool_lock irq_context: 0 sb_writers#11 &type->i_mutex_dir_key#7/1 cgroup_mutex shrinker_rwsem &____s->seqcount irq_context: 0 (wq_completion)mld (work_completion)(&(&idev->mc_ifc_work)->work) &idev->mc_lock rcu_read_lock rcu_read_lock &n->list_lock irq_context: 0 (wq_completion)mld (work_completion)(&(&idev->mc_ifc_work)->work) &idev->mc_lock rcu_read_lock rcu_read_lock rcu_read_lock_bh rcu_read_lock rcu_read_lock &c->lock irq_context: 0 (wq_completion)mld (work_completion)(&(&idev->mc_ifc_work)->work) &idev->mc_lock rcu_read_lock rcu_read_lock rcu_read_lock_bh rcu_read_lock rcu_read_lock &____s->seqcount#2 irq_context: 0 (wq_completion)mld (work_completion)(&(&idev->mc_ifc_work)->work) &idev->mc_lock fill_pool_map-wait-type-override &c->lock irq_context: 0 (wq_completion)mld (work_completion)(&(&idev->mc_ifc_work)->work) &idev->mc_lock fill_pool_map-wait-type-override &____s->seqcount#2 irq_context: 0 (wq_completion)mld (work_completion)(&(&idev->mc_ifc_work)->work) &idev->mc_lock fill_pool_map-wait-type-override &____s->seqcount irq_context: 0 (wq_completion)wg-kex-wg1#9 irq_context: 0 (wq_completion)wg-kex-wg1#9 (work_completion)(&peer->transmit_handshake_work) irq_context: 0 (wq_completion)wg-kex-wg1#9 (work_completion)(&peer->transmit_handshake_work) tk_core.seq.seqcount irq_context: 0 (wq_completion)wg-kex-wg1#9 (work_completion)(&peer->transmit_handshake_work) &wg->static_identity.lock irq_context: 0 (wq_completion)wg-kex-wg1#9 (work_completion)(&peer->transmit_handshake_work) &wg->static_identity.lock &handshake->lock irq_context: 0 (wq_completion)wg-kex-wg1#9 (work_completion)(&peer->transmit_handshake_work) &wg->static_identity.lock &handshake->lock crngs.lock irq_context: 0 (wq_completion)wg-kex-wg1#9 (work_completion)(&peer->transmit_handshake_work) &wg->static_identity.lock &handshake->lock tk_core.seq.seqcount irq_context: 0 (wq_completion)wg-kex-wg1#9 (work_completion)(&peer->transmit_handshake_work) &wg->static_identity.lock &handshake->lock &table->lock#2 irq_context: 0 (wq_completion)wg-kex-wg1#9 (work_completion)(&peer->transmit_handshake_work) &wg->static_identity.lock &handshake->lock rcu_read_lock_bh batched_entropy_u32.lock irq_context: 0 (wq_completion)wg-kex-wg1#9 (work_completion)(&peer->transmit_handshake_work) &wg->static_identity.lock &handshake->lock rcu_read_lock_bh &table->lock#2 irq_context: 0 (wq_completion)wg-kex-wg1#9 (work_completion)(&peer->transmit_handshake_work) &cookie->lock irq_context: 0 (wq_completion)wg-kex-wg1#9 (work_completion)(&peer->transmit_handshake_work) &cookie->lock irq_context: 0 (wq_completion)wg-kex-wg1#9 (work_completion)(&peer->transmit_handshake_work) rcu_read_lock_bh &obj_hash[i].lock irq_context: 0 (wq_completion)wg-kex-wg1#9 (work_completion)(&peer->transmit_handshake_work) rcu_read_lock_bh &base->lock irq_context: 0 (wq_completion)wg-kex-wg1#9 (work_completion)(&peer->transmit_handshake_work) rcu_read_lock_bh &base->lock &obj_hash[i].lock irq_context: 0 (wq_completion)wg-kex-wg1#9 (work_completion)(&peer->transmit_handshake_work) &obj_hash[i].lock irq_context: 0 (wq_completion)wg-kex-wg1#9 (work_completion)(&peer->transmit_handshake_work) &c->lock irq_context: 0 (wq_completion)wg-kex-wg1#9 (work_completion)(&peer->transmit_handshake_work) &peer->endpoint_lock irq_context: 0 (wq_completion)wg-kex-wg1#9 (work_completion)(&peer->transmit_handshake_work) &peer->endpoint_lock rcu_read_lock_bh rcu_read_lock rcu_read_lock rcu_read_lock &dir->lock#2 irq_context: 0 (wq_completion)wg-kex-wg1#9 (work_completion)(&peer->transmit_handshake_work) &peer->endpoint_lock rcu_read_lock_bh rcu_read_lock &____s->seqcount#7 irq_context: 0 (wq_completion)wg-kex-wg1#9 (work_completion)(&peer->transmit_handshake_work) &peer->endpoint_lock rcu_read_lock_bh rcu_read_lock &nf_conntrack_locks[i] irq_context: 0 (wq_completion)wg-kex-wg1#9 (work_completion)(&peer->transmit_handshake_work) &peer->endpoint_lock rcu_read_lock_bh rcu_read_lock &nf_conntrack_locks[i] &nf_conntrack_locks[i]/1 irq_context: 0 (wq_completion)wg-kex-wg1#9 (work_completion)(&peer->transmit_handshake_work) &peer->endpoint_lock rcu_read_lock_bh rcu_read_lock &nf_conntrack_locks[i] &nf_conntrack_locks[i]/1 batched_entropy_u8.lock irq_context: 0 (wq_completion)wg-kex-wg1#9 (work_completion)(&peer->transmit_handshake_work) &peer->endpoint_lock rcu_read_lock_bh rcu_read_lock &nf_conntrack_locks[i]/1 irq_context: 0 (wq_completion)wg-kex-wg1#9 (work_completion)(&peer->transmit_handshake_work) &peer->endpoint_lock rcu_read_lock_bh rcu_read_lock &____s->seqcount#10 irq_context: 0 (wq_completion)wg-kex-wg1#9 (work_completion)(&peer->transmit_handshake_work) &peer->endpoint_lock rcu_read_lock_bh rcu_read_lock rcu_read_lock_bh tk_core.seq.seqcount irq_context: 0 (wq_completion)wg-kex-wg1#9 (work_completion)(&peer->transmit_handshake_work) &peer->endpoint_lock rcu_read_lock_bh rcu_read_lock rcu_read_lock_bh &list->lock#5 irq_context: 0 (wq_completion)wg-kex-wg1#9 (work_completion)(&peer->transmit_handshake_work) batched_entropy_u8.lock irq_context: 0 (wq_completion)wg-kex-wg0#11 irq_context: 0 (wq_completion)wg-kex-wg0#11 (work_completion)(&({ do { const void *__vpp_verify = (typeof((worker) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __asm__ ("" : "=r"(__ptr) : "0"((typeof(*((worker))) *)((worker)))); (typeof((typeof(*((worker))) *)((worker)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); })->work) irq_context: 0 (wq_completion)wg-kex-wg0#11 (work_completion)(&({ do { const void *__vpp_verify = (typeof((worker) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __asm__ ("" : "=r"(__ptr) : "0"((typeof(*((worker))) *)((worker)))); (typeof((typeof(*((worker))) *)((worker)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); })->work) &r->consumer_lock#2 irq_context: 0 (wq_completion)wg-kex-wg0#11 (work_completion)(&({ do { const void *__vpp_verify = (typeof((worker) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __asm__ ("" : "=r"(__ptr) : "0"((typeof(*((worker))) *)((worker)))); (typeof((typeof(*((worker))) *)((worker)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); })->work) &wg->static_identity.lock irq_context: 0 (wq_completion)wg-kex-wg1#9 (work_completion)(&peer->transmit_handshake_work) &peer->endpoint_lock rcu_read_lock_bh rcu_read_lock &dir->lock#2 irq_context: 0 (wq_completion)wg-kex-wg1#9 (work_completion)(&peer->transmit_handshake_work) &peer->endpoint_lock rcu_read_lock_bh rcu_read_lock &ul->lock irq_context: 0 (wq_completion)wg-kex-wg0#11 (work_completion)(&({ do { const void *__vpp_verify = (typeof((worker) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __asm__ ("" : "=r"(__ptr) : "0"((typeof(*((worker))) *)((worker)))); (typeof((typeof(*((worker))) *)((worker)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); })->work) &wg->static_identity.lock &handshake->lock irq_context: 0 (wq_completion)wg-kex-wg1#9 (work_completion)(&peer->transmit_handshake_work) &peer->endpoint_lock rcu_read_lock_bh batched_entropy_u32.lock irq_context: 0 (wq_completion)wg-kex-wg0#11 (work_completion)(&({ do { const void *__vpp_verify = (typeof((worker) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __asm__ ("" : "=r"(__ptr) : "0"((typeof(*((worker))) *)((worker)))); (typeof((typeof(*((worker))) *)((worker)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); })->work) &wg->static_identity.lock &handshake->lock tk_core.seq.seqcount irq_context: 0 (wq_completion)wg-kex-wg0#11 (work_completion)(&({ do { const void *__vpp_verify = (typeof((worker) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __asm__ ("" : "=r"(__ptr) : "0"((typeof(*((worker))) *)((worker)))); (typeof((typeof(*((worker))) *)((worker)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); })->work) &wg->static_identity.lock &handshake->lock irq_context: 0 (wq_completion)wg-kex-wg0#11 (work_completion)(&({ do { const void *__vpp_verify = (typeof((worker) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __asm__ ("" : "=r"(__ptr) : "0"((typeof(*((worker))) *)((worker)))); (typeof((typeof(*((worker))) *)((worker)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); })->work) &wg->static_identity.lock &handshake->lock tk_core.seq.seqcount irq_context: 0 (wq_completion)wg-kex-wg0#11 (work_completion)(&({ do { const void *__vpp_verify = (typeof((worker) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __asm__ ("" : "=r"(__ptr) : "0"((typeof(*((worker))) *)((worker)))); (typeof((typeof(*((worker))) *)((worker)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); })->work) &peer->endpoint_lock irq_context: 0 (wq_completion)wg-kex-wg0#11 (work_completion)(&({ do { const void *__vpp_verify = (typeof((worker) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __asm__ ("" : "=r"(__ptr) : "0"((typeof(*((worker))) *)((worker)))); (typeof((typeof(*((worker))) *)((worker)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); })->work) tk_core.seq.seqcount irq_context: 0 (wq_completion)wg-kex-wg0#11 (work_completion)(&({ do { const void *__vpp_verify = (typeof((worker) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __asm__ ("" : "=r"(__ptr) : "0"((typeof(*((worker))) *)((worker)))); (typeof((typeof(*((worker))) *)((worker)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); })->work) &wg->static_identity.lock &handshake->lock crngs.lock irq_context: 0 (wq_completion)wg-kex-wg0#11 (work_completion)(&({ do { const void *__vpp_verify = (typeof((worker) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __asm__ ("" : "=r"(__ptr) : "0"((typeof(*((worker))) *)((worker)))); (typeof((typeof(*((worker))) *)((worker)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); })->work) &wg->static_identity.lock &handshake->lock &table->lock#2 irq_context: 0 (wq_completion)wg-kex-wg0#11 (work_completion)(&({ do { const void *__vpp_verify = (typeof((worker) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __asm__ ("" : "=r"(__ptr) : "0"((typeof(*((worker))) *)((worker)))); (typeof((typeof(*((worker))) *)((worker)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); })->work) &wg->static_identity.lock &handshake->lock rcu_read_lock_bh batched_entropy_u32.lock irq_context: 0 (wq_completion)wg-kex-wg0#11 (work_completion)(&({ do { const void *__vpp_verify = (typeof((worker) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __asm__ ("" : "=r"(__ptr) : "0"((typeof(*((worker))) *)((worker)))); (typeof((typeof(*((worker))) *)((worker)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); })->work) &wg->static_identity.lock &handshake->lock rcu_read_lock_bh &table->lock#2 irq_context: 0 (wq_completion)wg-kex-wg0#11 (work_completion)(&({ do { const void *__vpp_verify = (typeof((worker) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __asm__ ("" : "=r"(__ptr) : "0"((typeof(*((worker))) *)((worker)))); (typeof((typeof(*((worker))) *)((worker)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); })->work) &cookie->lock irq_context: 0 (wq_completion)wg-kex-wg0#11 (work_completion)(&({ do { const void *__vpp_verify = (typeof((worker) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __asm__ ("" : "=r"(__ptr) : "0"((typeof(*((worker))) *)((worker)))); (typeof((typeof(*((worker))) *)((worker)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); })->work) &cookie->lock irq_context: 0 (wq_completion)wg-kex-wg0#11 (work_completion)(&({ do { const void *__vpp_verify = (typeof((worker) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __asm__ ("" : "=r"(__ptr) : "0"((typeof(*((worker))) *)((worker)))); (typeof((typeof(*((worker))) *)((worker)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); })->work) &handshake->lock irq_context: 0 (wq_completion)wg-kex-wg0#11 (work_completion)(&({ do { const void *__vpp_verify = (typeof((worker) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __asm__ ("" : "=r"(__ptr) : "0"((typeof(*((worker))) *)((worker)))); (typeof((typeof(*((worker))) *)((worker)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); })->work) &handshake->lock fs_reclaim irq_context: 0 (wq_completion)wg-kex-wg0#11 (work_completion)(&({ do { const void *__vpp_verify = (typeof((worker) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __asm__ ("" : "=r"(__ptr) : "0"((typeof(*((worker))) *)((worker)))); (typeof((typeof(*((worker))) *)((worker)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); })->work) &handshake->lock fs_reclaim mmu_notifier_invalidate_range_start irq_context: 0 (wq_completion)wg-kex-wg0#11 (work_completion)(&({ do { const void *__vpp_verify = (typeof((worker) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __asm__ ("" : "=r"(__ptr) : "0"((typeof(*((worker))) *)((worker)))); (typeof((typeof(*((worker))) *)((worker)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); })->work) &handshake->lock tk_core.seq.seqcount irq_context: 0 (wq_completion)wg-kex-wg0#11 (work_completion)(&({ do { const void *__vpp_verify = (typeof((worker) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __asm__ ("" : "=r"(__ptr) : "0"((typeof(*((worker))) *)((worker)))); (typeof((typeof(*((worker))) *)((worker)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); })->work) &handshake->lock rcu_read_lock_bh &peer->keypairs.keypair_update_lock irq_context: 0 (wq_completion)wg-kex-wg0#11 (work_completion)(&({ do { const void *__vpp_verify = (typeof((worker) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __asm__ ("" : "=r"(__ptr) : "0"((typeof(*((worker))) *)((worker)))); (typeof((typeof(*((worker))) *)((worker)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); })->work) &handshake->lock rcu_read_lock_bh &table->lock#2 irq_context: 0 (wq_completion)wg-kex-wg0#11 (work_completion)(&({ do { const void *__vpp_verify = (typeof((worker) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __asm__ ("" : "=r"(__ptr) : "0"((typeof(*((worker))) *)((worker)))); (typeof((typeof(*((worker))) *)((worker)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); })->work) rcu_read_lock_bh &obj_hash[i].lock irq_context: 0 (wq_completion)wg-kex-wg0#11 (work_completion)(&({ do { const void *__vpp_verify = (typeof((worker) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __asm__ ("" : "=r"(__ptr) : "0"((typeof(*((worker))) *)((worker)))); (typeof((typeof(*((worker))) *)((worker)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); })->work) rcu_read_lock_bh &base->lock irq_context: 0 (wq_completion)wg-kex-wg0#11 (work_completion)(&({ do { const void *__vpp_verify = (typeof((worker) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __asm__ ("" : "=r"(__ptr) : "0"((typeof(*((worker))) *)((worker)))); (typeof((typeof(*((worker))) *)((worker)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); })->work) rcu_read_lock_bh &base->lock &obj_hash[i].lock irq_context: 0 (wq_completion)wg-kex-wg0#11 (work_completion)(&({ do { const void *__vpp_verify = (typeof((worker) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __asm__ ("" : "=r"(__ptr) : "0"((typeof(*((worker))) *)((worker)))); (typeof((typeof(*((worker))) *)((worker)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); })->work) &obj_hash[i].lock irq_context: 0 (wq_completion)wg-kex-wg0#11 (work_completion)(&({ do { const void *__vpp_verify = (typeof((worker) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __asm__ ("" : "=r"(__ptr) : "0"((typeof(*((worker))) *)((worker)))); (typeof((typeof(*((worker))) *)((worker)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); })->work) &peer->endpoint_lock irq_context: 0 (wq_completion)wg-kex-wg0#11 (work_completion)(&({ do { const void *__vpp_verify = (typeof((worker) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __asm__ ("" : "=r"(__ptr) : "0"((typeof(*((worker))) *)((worker)))); (typeof((typeof(*((worker))) *)((worker)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); })->work) &peer->endpoint_lock rcu_read_lock_bh rcu_read_lock &____s->seqcount#7 irq_context: 0 (wq_completion)wg-kex-wg0#11 (work_completion)(&({ do { const void *__vpp_verify = (typeof((worker) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __asm__ ("" : "=r"(__ptr) : "0"((typeof(*((worker))) *)((worker)))); (typeof((typeof(*((worker))) *)((worker)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); })->work) &peer->endpoint_lock rcu_read_lock_bh rcu_read_lock &____s->seqcount#10 irq_context: 0 (wq_completion)wg-kex-wg0#11 (work_completion)(&({ do { const void *__vpp_verify = (typeof((worker) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __asm__ ("" : "=r"(__ptr) : "0"((typeof(*((worker))) *)((worker)))); (typeof((typeof(*((worker))) *)((worker)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); })->work) &peer->endpoint_lock rcu_read_lock_bh rcu_read_lock rcu_read_lock_bh tk_core.seq.seqcount irq_context: 0 (wq_completion)wg-kex-wg0#11 (work_completion)(&({ do { const void *__vpp_verify = (typeof((worker) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __asm__ ("" : "=r"(__ptr) : "0"((typeof(*((worker))) *)((worker)))); (typeof((typeof(*((worker))) *)((worker)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); })->work) &peer->endpoint_lock rcu_read_lock_bh rcu_read_lock rcu_read_lock_bh &list->lock#5 irq_context: 0 (wq_completion)wg-kex-wg0#11 (work_completion)(&({ do { const void *__vpp_verify = (typeof((worker) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __asm__ ("" : "=r"(__ptr) : "0"((typeof(*((worker))) *)((worker)))); (typeof((typeof(*((worker))) *)((worker)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); })->work) &rq->__lock irq_context: 0 (wq_completion)wg-kex-wg0#11 (work_completion)(&({ do { const void *__vpp_verify = (typeof((worker) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __asm__ ("" : "=r"(__ptr) : "0"((typeof(*((worker))) *)((worker)))); (typeof((typeof(*((worker))) *)((worker)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); })->work) &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 sb_writers#11 &type->i_mutex_dir_key#7 &____s->seqcount#2 irq_context: 0 stock_lock rcu_read_lock per_cpu_ptr(&cgroup_rstat_cpu_lock, cpu) irq_context: 0 sb_writers#11 &type->i_mutex_dir_key#7/1 tomoyo_ss &c->lock irq_context: 0 (wq_completion)wg-kex-wg1#10 irq_context: 0 (wq_completion)wg-kex-wg1#11 irq_context: 0 (wq_completion)wg-kex-wg1#10 (work_completion)(&peer->transmit_handshake_work) irq_context: 0 (wq_completion)wg-kex-wg1#11 (work_completion)(&({ do { const void *__vpp_verify = (typeof((worker) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __asm__ ("" : "=r"(__ptr) : "0"((typeof(*((worker))) *)((worker)))); (typeof((typeof(*((worker))) *)((worker)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); })->work) irq_context: 0 (wq_completion)wg-kex-wg1#10 (work_completion)(&peer->transmit_handshake_work) &rq->__lock irq_context: 0 (wq_completion)wg-kex-wg1#11 (work_completion)(&({ do { const void *__vpp_verify = (typeof((worker) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __asm__ ("" : "=r"(__ptr) : "0"((typeof(*((worker))) *)((worker)))); (typeof((typeof(*((worker))) *)((worker)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); })->work) &rq->__lock irq_context: 0 (wq_completion)wg-kex-wg1#10 (work_completion)(&peer->transmit_handshake_work) tk_core.seq.seqcount irq_context: 0 (wq_completion)wg-kex-wg1#10 (work_completion)(&peer->transmit_handshake_work) &wg->static_identity.lock irq_context: 0 (wq_completion)wg-kex-wg1#10 (work_completion)(&peer->transmit_handshake_work) &wg->static_identity.lock &handshake->lock irq_context: 0 (wq_completion)wg-kex-wg1#10 (work_completion)(&peer->transmit_handshake_work) &wg->static_identity.lock &handshake->lock crngs.lock irq_context: 0 (wq_completion)wg-kex-wg1#10 (work_completion)(&peer->transmit_handshake_work) &wg->static_identity.lock &handshake->lock tk_core.seq.seqcount irq_context: 0 (wq_completion)wg-kex-wg1#10 (work_completion)(&peer->transmit_handshake_work) &wg->static_identity.lock &handshake->lock &table->lock#2 irq_context: 0 (wq_completion)wg-kex-wg1#10 (work_completion)(&peer->transmit_handshake_work) &wg->static_identity.lock &handshake->lock rcu_read_lock_bh batched_entropy_u32.lock irq_context: 0 (wq_completion)wg-kex-wg1#10 (work_completion)(&peer->transmit_handshake_work) &wg->static_identity.lock &handshake->lock rcu_read_lock_bh &table->lock#2 irq_context: 0 (wq_completion)wg-kex-wg1#10 (work_completion)(&peer->transmit_handshake_work) &cookie->lock irq_context: 0 (wq_completion)wg-kex-wg1#10 (work_completion)(&peer->transmit_handshake_work) &cookie->lock irq_context: 0 (wq_completion)wg-kex-wg1#10 (work_completion)(&peer->transmit_handshake_work) rcu_read_lock_bh &obj_hash[i].lock irq_context: 0 (wq_completion)wg-kex-wg1#10 (work_completion)(&peer->transmit_handshake_work) rcu_read_lock_bh &base->lock irq_context: 0 (wq_completion)wg-kex-wg1#10 (work_completion)(&peer->transmit_handshake_work) rcu_read_lock_bh &base->lock &obj_hash[i].lock irq_context: 0 (wq_completion)wg-kex-wg1#10 (work_completion)(&peer->transmit_handshake_work) &obj_hash[i].lock irq_context: 0 (wq_completion)wg-kex-wg1#10 (work_completion)(&peer->transmit_handshake_work) &peer->endpoint_lock irq_context: 0 (wq_completion)wg-kex-wg1#10 (work_completion)(&peer->transmit_handshake_work) &peer->endpoint_lock rcu_read_lock_bh rcu_read_lock &dir->lock#2 irq_context: 0 (wq_completion)wg-kex-wg1#10 (work_completion)(&peer->transmit_handshake_work) &peer->endpoint_lock rcu_read_lock_bh rcu_read_lock &____s->seqcount#7 irq_context: 0 (wq_completion)wg-kex-wg1#10 (work_completion)(&peer->transmit_handshake_work) &peer->endpoint_lock rcu_read_lock_bh rcu_read_lock &ul->lock irq_context: 0 (wq_completion)wg-kex-wg1#10 (work_completion)(&peer->transmit_handshake_work) &peer->endpoint_lock rcu_read_lock_bh batched_entropy_u32.lock irq_context: 0 (wq_completion)wg-kex-wg1#10 (work_completion)(&peer->transmit_handshake_work) &peer->endpoint_lock rcu_read_lock_bh rcu_read_lock &nf_conntrack_locks[i] irq_context: 0 (wq_completion)wg-kex-wg1#10 (work_completion)(&peer->transmit_handshake_work) &peer->endpoint_lock rcu_read_lock_bh rcu_read_lock &nf_conntrack_locks[i] &nf_conntrack_locks[i]/1 irq_context: 0 (wq_completion)wg-kex-wg1#10 (work_completion)(&peer->transmit_handshake_work) &peer->endpoint_lock rcu_read_lock_bh rcu_read_lock &nf_conntrack_locks[i] &nf_conntrack_locks[i]/1 batched_entropy_u8.lock irq_context: 0 (wq_completion)wg-kex-wg1#10 (work_completion)(&peer->transmit_handshake_work) &peer->endpoint_lock rcu_read_lock_bh rcu_read_lock &nf_conntrack_locks[i]/1 irq_context: 0 (wq_completion)wg-kex-wg1#10 (work_completion)(&peer->transmit_handshake_work) &peer->endpoint_lock rcu_read_lock_bh rcu_read_lock &____s->seqcount#10 irq_context: 0 (wq_completion)wg-kex-wg1#10 (work_completion)(&peer->transmit_handshake_work) &peer->endpoint_lock rcu_read_lock_bh rcu_read_lock rcu_read_lock_bh tk_core.seq.seqcount irq_context: 0 (wq_completion)wg-kex-wg1#10 (work_completion)(&peer->transmit_handshake_work) &peer->endpoint_lock rcu_read_lock_bh rcu_read_lock rcu_read_lock_bh &list->lock#5 irq_context: 0 (wq_completion)wg-kex-wg1#10 (work_completion)(&peer->transmit_handshake_work) batched_entropy_u8.lock irq_context: 0 (wq_completion)mld (work_completion)(&(&idev->mc_ifc_work)->work) &idev->mc_lock rcu_read_lock &rcu_state.expedited_wq irq_context: 0 (wq_completion)mld (work_completion)(&(&idev->mc_ifc_work)->work) &idev->mc_lock rcu_read_lock &rcu_state.expedited_wq &p->pi_lock irq_context: 0 (wq_completion)mld (work_completion)(&(&idev->mc_ifc_work)->work) &idev->mc_lock rcu_read_lock &rcu_state.expedited_wq &p->pi_lock &rq->__lock irq_context: 0 (wq_completion)mld (work_completion)(&(&idev->mc_ifc_work)->work) &idev->mc_lock rcu_read_lock &rcu_state.expedited_wq &p->pi_lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)wg-kex-wg0#12 irq_context: 0 (wq_completion)wg-kex-wg0#12 (work_completion)(&({ do { const void *__vpp_verify = (typeof((worker) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __asm__ ("" : "=r"(__ptr) : "0"((typeof(*((worker))) *)((worker)))); (typeof((typeof(*((worker))) *)((worker)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); })->work) irq_context: 0 (wq_completion)wg-kex-wg0#12 (work_completion)(&({ do { const void *__vpp_verify = (typeof((worker) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __asm__ ("" : "=r"(__ptr) : "0"((typeof(*((worker))) *)((worker)))); (typeof((typeof(*((worker))) *)((worker)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); })->work) &r->consumer_lock#2 irq_context: 0 (wq_completion)wg-kex-wg0#12 (work_completion)(&({ do { const void *__vpp_verify = (typeof((worker) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __asm__ ("" : "=r"(__ptr) : "0"((typeof(*((worker))) *)((worker)))); (typeof((typeof(*((worker))) *)((worker)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); })->work) &wg->static_identity.lock irq_context: 0 (wq_completion)wg-kex-wg0#12 (work_completion)(&({ do { const void *__vpp_verify = (typeof((worker) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __asm__ ("" : "=r"(__ptr) : "0"((typeof(*((worker))) *)((worker)))); (typeof((typeof(*((worker))) *)((worker)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); })->work) &wg->static_identity.lock &handshake->lock irq_context: 0 (wq_completion)wg-kex-wg0#12 (work_completion)(&({ do { const void *__vpp_verify = (typeof((worker) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __asm__ ("" : "=r"(__ptr) : "0"((typeof(*((worker))) *)((worker)))); (typeof((typeof(*((worker))) *)((worker)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); })->work) &wg->static_identity.lock &handshake->lock tk_core.seq.seqcount irq_context: 0 (wq_completion)wg-kex-wg0#12 (work_completion)(&({ do { const void *__vpp_verify = (typeof((worker) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __asm__ ("" : "=r"(__ptr) : "0"((typeof(*((worker))) *)((worker)))); (typeof((typeof(*((worker))) *)((worker)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); })->work) &wg->static_identity.lock &handshake->lock irq_context: 0 (wq_completion)wg-kex-wg0#12 (work_completion)(&({ do { const void *__vpp_verify = (typeof((worker) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __asm__ ("" : "=r"(__ptr) : "0"((typeof(*((worker))) *)((worker)))); (typeof((typeof(*((worker))) *)((worker)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); })->work) &wg->static_identity.lock &handshake->lock tk_core.seq.seqcount irq_context: 0 (wq_completion)wg-kex-wg0#12 (work_completion)(&({ do { const void *__vpp_verify = (typeof((worker) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __asm__ ("" : "=r"(__ptr) : "0"((typeof(*((worker))) *)((worker)))); (typeof((typeof(*((worker))) *)((worker)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); })->work) &peer->endpoint_lock irq_context: 0 (wq_completion)wg-kex-wg0#12 (work_completion)(&({ do { const void *__vpp_verify = (typeof((worker) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __asm__ ("" : "=r"(__ptr) : "0"((typeof(*((worker))) *)((worker)))); (typeof((typeof(*((worker))) *)((worker)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); })->work) tk_core.seq.seqcount irq_context: 0 (wq_completion)wg-kex-wg0#12 (work_completion)(&({ do { const void *__vpp_verify = (typeof((worker) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __asm__ ("" : "=r"(__ptr) : "0"((typeof(*((worker))) *)((worker)))); (typeof((typeof(*((worker))) *)((worker)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); })->work) &wg->static_identity.lock &handshake->lock crngs.lock irq_context: 0 (wq_completion)wg-kex-wg0#12 (work_completion)(&({ do { const void *__vpp_verify = (typeof((worker) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __asm__ ("" : "=r"(__ptr) : "0"((typeof(*((worker))) *)((worker)))); (typeof((typeof(*((worker))) *)((worker)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); })->work) &wg->static_identity.lock &handshake->lock &table->lock#2 irq_context: 0 (wq_completion)wg-kex-wg0#12 (work_completion)(&({ do { const void *__vpp_verify = (typeof((worker) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __asm__ ("" : "=r"(__ptr) : "0"((typeof(*((worker))) *)((worker)))); (typeof((typeof(*((worker))) *)((worker)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); })->work) &wg->static_identity.lock &handshake->lock rcu_read_lock_bh batched_entropy_u32.lock irq_context: 0 (wq_completion)wg-kex-wg0#12 (work_completion)(&({ do { const void *__vpp_verify = (typeof((worker) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __asm__ ("" : "=r"(__ptr) : "0"((typeof(*((worker))) *)((worker)))); (typeof((typeof(*((worker))) *)((worker)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); })->work) &wg->static_identity.lock &handshake->lock rcu_read_lock_bh &table->lock#2 irq_context: 0 (wq_completion)wg-kex-wg0#12 (work_completion)(&({ do { const void *__vpp_verify = (typeof((worker) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __asm__ ("" : "=r"(__ptr) : "0"((typeof(*((worker))) *)((worker)))); (typeof((typeof(*((worker))) *)((worker)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); })->work) &cookie->lock irq_context: 0 (wq_completion)wg-kex-wg0#12 (work_completion)(&({ do { const void *__vpp_verify = (typeof((worker) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __asm__ ("" : "=r"(__ptr) : "0"((typeof(*((worker))) *)((worker)))); (typeof((typeof(*((worker))) *)((worker)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); })->work) &cookie->lock irq_context: 0 (wq_completion)wg-kex-wg0#12 (work_completion)(&({ do { const void *__vpp_verify = (typeof((worker) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __asm__ ("" : "=r"(__ptr) : "0"((typeof(*((worker))) *)((worker)))); (typeof((typeof(*((worker))) *)((worker)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); })->work) &handshake->lock irq_context: 0 (wq_completion)wg-kex-wg0#12 (work_completion)(&({ do { const void *__vpp_verify = (typeof((worker) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __asm__ ("" : "=r"(__ptr) : "0"((typeof(*((worker))) *)((worker)))); (typeof((typeof(*((worker))) *)((worker)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); })->work) &handshake->lock fs_reclaim irq_context: 0 (wq_completion)wg-kex-wg0#12 (work_completion)(&({ do { const void *__vpp_verify = (typeof((worker) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __asm__ ("" : "=r"(__ptr) : "0"((typeof(*((worker))) *)((worker)))); (typeof((typeof(*((worker))) *)((worker)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); })->work) &handshake->lock fs_reclaim mmu_notifier_invalidate_range_start irq_context: 0 (wq_completion)wg-kex-wg0#12 (work_completion)(&({ do { const void *__vpp_verify = (typeof((worker) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __asm__ ("" : "=r"(__ptr) : "0"((typeof(*((worker))) *)((worker)))); (typeof((typeof(*((worker))) *)((worker)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); })->work) &handshake->lock tk_core.seq.seqcount irq_context: 0 (wq_completion)wg-kex-wg0#12 (work_completion)(&({ do { const void *__vpp_verify = (typeof((worker) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __asm__ ("" : "=r"(__ptr) : "0"((typeof(*((worker))) *)((worker)))); (typeof((typeof(*((worker))) *)((worker)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); })->work) &handshake->lock rcu_read_lock_bh &peer->keypairs.keypair_update_lock irq_context: 0 (wq_completion)wg-kex-wg0#12 (work_completion)(&({ do { const void *__vpp_verify = (typeof((worker) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __asm__ ("" : "=r"(__ptr) : "0"((typeof(*((worker))) *)((worker)))); (typeof((typeof(*((worker))) *)((worker)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); })->work) &handshake->lock rcu_read_lock_bh &table->lock#2 irq_context: 0 (wq_completion)wg-kex-wg0#12 (work_completion)(&({ do { const void *__vpp_verify = (typeof((worker) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __asm__ ("" : "=r"(__ptr) : "0"((typeof(*((worker))) *)((worker)))); (typeof((typeof(*((worker))) *)((worker)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); })->work) rcu_read_lock_bh &obj_hash[i].lock irq_context: 0 (wq_completion)wg-kex-wg0#12 (work_completion)(&({ do { const void *__vpp_verify = (typeof((worker) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __asm__ ("" : "=r"(__ptr) : "0"((typeof(*((worker))) *)((worker)))); (typeof((typeof(*((worker))) *)((worker)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); })->work) rcu_read_lock_bh &base->lock irq_context: 0 (wq_completion)wg-kex-wg0#12 (work_completion)(&({ do { const void *__vpp_verify = (typeof((worker) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __asm__ ("" : "=r"(__ptr) : "0"((typeof(*((worker))) *)((worker)))); (typeof((typeof(*((worker))) *)((worker)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); })->work) rcu_read_lock_bh &base->lock &obj_hash[i].lock irq_context: 0 (wq_completion)wg-kex-wg0#12 (work_completion)(&({ do { const void *__vpp_verify = (typeof((worker) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __asm__ ("" : "=r"(__ptr) : "0"((typeof(*((worker))) *)((worker)))); (typeof((typeof(*((worker))) *)((worker)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); })->work) &obj_hash[i].lock irq_context: 0 (wq_completion)wg-kex-wg0#12 (work_completion)(&({ do { const void *__vpp_verify = (typeof((worker) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __asm__ ("" : "=r"(__ptr) : "0"((typeof(*((worker))) *)((worker)))); (typeof((typeof(*((worker))) *)((worker)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); })->work) &peer->endpoint_lock irq_context: 0 (wq_completion)wg-kex-wg0#12 (work_completion)(&({ do { const void *__vpp_verify = (typeof((worker) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __asm__ ("" : "=r"(__ptr) : "0"((typeof(*((worker))) *)((worker)))); (typeof((typeof(*((worker))) *)((worker)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); })->work) &peer->endpoint_lock rcu_read_lock_bh &obj_hash[i].lock irq_context: 0 (wq_completion)wg-kex-wg0#12 (work_completion)(&({ do { const void *__vpp_verify = (typeof((worker) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __asm__ ("" : "=r"(__ptr) : "0"((typeof(*((worker))) *)((worker)))); (typeof((typeof(*((worker))) *)((worker)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); })->work) &peer->endpoint_lock rcu_read_lock_bh &obj_hash[i].lock pool_lock irq_context: 0 (wq_completion)wg-kex-wg0#12 (work_completion)(&({ do { const void *__vpp_verify = (typeof((worker) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __asm__ ("" : "=r"(__ptr) : "0"((typeof(*((worker))) *)((worker)))); (typeof((typeof(*((worker))) *)((worker)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); })->work) &peer->endpoint_lock rcu_read_lock_bh pool_lock#2 irq_context: 0 (wq_completion)wg-kex-wg0#12 (work_completion)(&({ do { const void *__vpp_verify = (typeof((worker) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __asm__ ("" : "=r"(__ptr) : "0"((typeof(*((worker))) *)((worker)))); (typeof((typeof(*((worker))) *)((worker)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); })->work) &peer->endpoint_lock rcu_read_lock_bh rcu_read_lock rcu_read_lock rcu_read_lock pool_lock#2 irq_context: 0 (wq_completion)wg-kex-wg0#12 (work_completion)(&({ do { const void *__vpp_verify = (typeof((worker) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __asm__ ("" : "=r"(__ptr) : "0"((typeof(*((worker))) *)((worker)))); (typeof((typeof(*((worker))) *)((worker)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); })->work) &peer->endpoint_lock rcu_read_lock_bh rcu_read_lock rcu_read_lock rcu_read_lock &dir->lock#2 irq_context: 0 (wq_completion)wg-kex-wg0#12 (work_completion)(&({ do { const void *__vpp_verify = (typeof((worker) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __asm__ ("" : "=r"(__ptr) : "0"((typeof(*((worker))) *)((worker)))); (typeof((typeof(*((worker))) *)((worker)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); })->work) &peer->endpoint_lock rcu_read_lock_bh rcu_read_lock &____s->seqcount#7 irq_context: 0 (wq_completion)wg-kex-wg0#12 (work_completion)(&({ do { const void *__vpp_verify = (typeof((worker) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __asm__ ("" : "=r"(__ptr) : "0"((typeof(*((worker))) *)((worker)))); (typeof((typeof(*((worker))) *)((worker)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); })->work) &peer->endpoint_lock rcu_read_lock_bh rcu_read_lock &____s->seqcount#10 irq_context: 0 (wq_completion)wg-kex-wg0#12 (work_completion)(&({ do { const void *__vpp_verify = (typeof((worker) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __asm__ ("" : "=r"(__ptr) : "0"((typeof(*((worker))) *)((worker)))); (typeof((typeof(*((worker))) *)((worker)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); })->work) &peer->endpoint_lock rcu_read_lock_bh rcu_read_lock rcu_read_lock_bh tk_core.seq.seqcount irq_context: 0 (wq_completion)wg-kex-wg0#12 (work_completion)(&({ do { const void *__vpp_verify = (typeof((worker) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __asm__ ("" : "=r"(__ptr) : "0"((typeof(*((worker))) *)((worker)))); (typeof((typeof(*((worker))) *)((worker)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); })->work) &peer->endpoint_lock rcu_read_lock_bh rcu_read_lock rcu_read_lock_bh &list->lock#5 irq_context: softirq rcu_read_lock rcu_read_lock rcu_read_lock &pool->lock fill_pool_map-wait-type-override pool_lock#2 irq_context: softirq rcu_read_lock rcu_read_lock rcu_read_lock &pool->lock fill_pool_map-wait-type-override pool_lock irq_context: 0 (wq_completion)wg-kex-wg1#12 irq_context: 0 (wq_completion)wg-kex-wg1#12 (work_completion)(&({ do { const void *__vpp_verify = (typeof((worker) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __asm__ ("" : "=r"(__ptr) : "0"((typeof(*((worker))) *)((worker)))); (typeof((typeof(*((worker))) *)((worker)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); })->work) irq_context: 0 (wq_completion)wg-kex-wg1#12 (work_completion)(&({ do { const void *__vpp_verify = (typeof((worker) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __asm__ ("" : "=r"(__ptr) : "0"((typeof(*((worker))) *)((worker)))); (typeof((typeof(*((worker))) *)((worker)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); })->work) &r->consumer_lock#2 irq_context: 0 (wq_completion)wg-kex-wg1#12 (work_completion)(&({ do { const void *__vpp_verify = (typeof((worker) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __asm__ ("" : "=r"(__ptr) : "0"((typeof(*((worker))) *)((worker)))); (typeof((typeof(*((worker))) *)((worker)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); })->work) &wg->static_identity.lock irq_context: 0 (wq_completion)wg-kex-wg1#12 (work_completion)(&({ do { const void *__vpp_verify = (typeof((worker) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __asm__ ("" : "=r"(__ptr) : "0"((typeof(*((worker))) *)((worker)))); (typeof((typeof(*((worker))) *)((worker)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); })->work) &wg->static_identity.lock &handshake->lock irq_context: 0 (wq_completion)wg-kex-wg1#12 (work_completion)(&({ do { const void *__vpp_verify = (typeof((worker) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __asm__ ("" : "=r"(__ptr) : "0"((typeof(*((worker))) *)((worker)))); (typeof((typeof(*((worker))) *)((worker)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); })->work) &wg->static_identity.lock &handshake->lock irq_context: 0 (wq_completion)wg-kex-wg1#12 (work_completion)(&({ do { const void *__vpp_verify = (typeof((worker) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __asm__ ("" : "=r"(__ptr) : "0"((typeof(*((worker))) *)((worker)))); (typeof((typeof(*((worker))) *)((worker)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); })->work) &peer->endpoint_lock irq_context: 0 (wq_completion)wg-kex-wg1#12 (work_completion)(&({ do { const void *__vpp_verify = (typeof((worker) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __asm__ ("" : "=r"(__ptr) : "0"((typeof(*((worker))) *)((worker)))); (typeof((typeof(*((worker))) *)((worker)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); })->work) &handshake->lock irq_context: 0 (wq_completion)wg-kex-wg1#12 (work_completion)(&({ do { const void *__vpp_verify = (typeof((worker) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __asm__ ("" : "=r"(__ptr) : "0"((typeof(*((worker))) *)((worker)))); (typeof((typeof(*((worker))) *)((worker)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); })->work) &handshake->lock fs_reclaim irq_context: 0 (wq_completion)wg-kex-wg1#12 (work_completion)(&({ do { const void *__vpp_verify = (typeof((worker) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __asm__ ("" : "=r"(__ptr) : "0"((typeof(*((worker))) *)((worker)))); (typeof((typeof(*((worker))) *)((worker)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); })->work) &handshake->lock fs_reclaim mmu_notifier_invalidate_range_start irq_context: 0 (wq_completion)wg-kex-wg1#12 (work_completion)(&({ do { const void *__vpp_verify = (typeof((worker) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __asm__ ("" : "=r"(__ptr) : "0"((typeof(*((worker))) *)((worker)))); (typeof((typeof(*((worker))) *)((worker)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); })->work) &handshake->lock &c->lock irq_context: 0 (wq_completion)wg-kex-wg1#12 (work_completion)(&({ do { const void *__vpp_verify = (typeof((worker) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __asm__ ("" : "=r"(__ptr) : "0"((typeof(*((worker))) *)((worker)))); (typeof((typeof(*((worker))) *)((worker)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); })->work) &handshake->lock &____s->seqcount#2 irq_context: 0 (wq_completion)wg-kex-wg1#12 (work_completion)(&({ do { const void *__vpp_verify = (typeof((worker) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __asm__ ("" : "=r"(__ptr) : "0"((typeof(*((worker))) *)((worker)))); (typeof((typeof(*((worker))) *)((worker)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); })->work) &handshake->lock &____s->seqcount irq_context: 0 (wq_completion)wg-kex-wg1#12 (work_completion)(&({ do { const void *__vpp_verify = (typeof((worker) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __asm__ ("" : "=r"(__ptr) : "0"((typeof(*((worker))) *)((worker)))); (typeof((typeof(*((worker))) *)((worker)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); })->work) &handshake->lock tk_core.seq.seqcount irq_context: 0 (wq_completion)wg-kex-wg1#12 (work_completion)(&({ do { const void *__vpp_verify = (typeof((worker) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __asm__ ("" : "=r"(__ptr) : "0"((typeof(*((worker))) *)((worker)))); (typeof((typeof(*((worker))) *)((worker)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); })->work) &handshake->lock rcu_read_lock_bh &peer->keypairs.keypair_update_lock irq_context: 0 (wq_completion)wg-kex-wg1#12 (work_completion)(&({ do { const void *__vpp_verify = (typeof((worker) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __asm__ ("" : "=r"(__ptr) : "0"((typeof(*((worker))) *)((worker)))); (typeof((typeof(*((worker))) *)((worker)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); })->work) &handshake->lock rcu_read_lock_bh &table->lock#2 irq_context: 0 (wq_completion)wg-kex-wg1#12 (work_completion)(&({ do { const void *__vpp_verify = (typeof((worker) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __asm__ ("" : "=r"(__ptr) : "0"((typeof(*((worker))) *)((worker)))); (typeof((typeof(*((worker))) *)((worker)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); })->work) rcu_read_lock_bh &obj_hash[i].lock irq_context: 0 (wq_completion)wg-kex-wg1#12 (work_completion)(&({ do { const void *__vpp_verify = (typeof((worker) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __asm__ ("" : "=r"(__ptr) : "0"((typeof(*((worker))) *)((worker)))); (typeof((typeof(*((worker))) *)((worker)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); })->work) rcu_read_lock_bh &base->lock irq_context: 0 (wq_completion)wg-kex-wg1#12 (work_completion)(&({ do { const void *__vpp_verify = (typeof((worker) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __asm__ ("" : "=r"(__ptr) : "0"((typeof(*((worker))) *)((worker)))); (typeof((typeof(*((worker))) *)((worker)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); })->work) rcu_read_lock_bh &base->lock &obj_hash[i].lock irq_context: 0 (wq_completion)wg-kex-wg1#12 (work_completion)(&({ do { const void *__vpp_verify = (typeof((worker) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __asm__ ("" : "=r"(__ptr) : "0"((typeof(*((worker))) *)((worker)))); (typeof((typeof(*((worker))) *)((worker)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); })->work) &obj_hash[i].lock irq_context: 0 (wq_completion)wg-kex-wg1#12 (work_completion)(&({ do { const void *__vpp_verify = (typeof((worker) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __asm__ ("" : "=r"(__ptr) : "0"((typeof(*((worker))) *)((worker)))); (typeof((typeof(*((worker))) *)((worker)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); })->work) &base->lock irq_context: 0 (wq_completion)wg-kex-wg1#12 (work_completion)(&({ do { const void *__vpp_verify = (typeof((worker) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __asm__ ("" : "=r"(__ptr) : "0"((typeof(*((worker))) *)((worker)))); (typeof((typeof(*((worker))) *)((worker)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); })->work) &base->lock &obj_hash[i].lock irq_context: 0 (wq_completion)wg-kex-wg1#12 (work_completion)(&({ do { const void *__vpp_verify = (typeof((worker) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __asm__ ("" : "=r"(__ptr) : "0"((typeof(*((worker))) *)((worker)))); (typeof((typeof(*((worker))) *)((worker)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); })->work) tk_core.seq.seqcount irq_context: 0 (wq_completion)wg-kex-wg1#12 (work_completion)(&({ do { const void *__vpp_verify = (typeof((worker) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __asm__ ("" : "=r"(__ptr) : "0"((typeof(*((worker))) *)((worker)))); (typeof((typeof(*((worker))) *)((worker)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); })->work) &list->lock#17 irq_context: 0 (wq_completion)wg-kex-wg1#12 (work_completion)(&({ do { const void *__vpp_verify = (typeof((worker) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __asm__ ("" : "=r"(__ptr) : "0"((typeof(*((worker))) *)((worker)))); (typeof((typeof(*((worker))) *)((worker)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); })->work) rcu_read_lock_bh &r->producer_lock#2 irq_context: 0 (wq_completion)wg-kex-wg1#12 (work_completion)(&({ do { const void *__vpp_verify = (typeof((worker) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __asm__ ("" : "=r"(__ptr) : "0"((typeof(*((worker))) *)((worker)))); (typeof((typeof(*((worker))) *)((worker)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); })->work) rcu_read_lock_bh rcu_read_lock &pool->lock irq_context: 0 (wq_completion)wg-kex-wg1#12 (work_completion)(&({ do { const void *__vpp_verify = (typeof((worker) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __asm__ ("" : "=r"(__ptr) : "0"((typeof(*((worker))) *)((worker)))); (typeof((typeof(*((worker))) *)((worker)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); })->work) rcu_read_lock_bh rcu_read_lock &pool->lock &obj_hash[i].lock irq_context: 0 (wq_completion)wg-kex-wg1#12 (work_completion)(&({ do { const void *__vpp_verify = (typeof((worker) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __asm__ ("" : "=r"(__ptr) : "0"((typeof(*((worker))) *)((worker)))); (typeof((typeof(*((worker))) *)((worker)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); })->work) rcu_read_lock_bh rcu_read_lock &pool->lock &p->pi_lock irq_context: 0 (wq_completion)wg-kex-wg2#7 irq_context: 0 (wq_completion)wg-kex-wg2#7 (work_completion)(&peer->transmit_handshake_work) irq_context: 0 (wq_completion)wg-kex-wg2#7 (work_completion)(&peer->transmit_handshake_work) tk_core.seq.seqcount irq_context: 0 (wq_completion)wg-kex-wg2#7 (work_completion)(&peer->transmit_handshake_work) &wg->static_identity.lock irq_context: 0 (wq_completion)wg-kex-wg2#7 (work_completion)(&peer->transmit_handshake_work) &wg->static_identity.lock &handshake->lock irq_context: 0 (wq_completion)wg-kex-wg2#7 (work_completion)(&peer->transmit_handshake_work) &wg->static_identity.lock &handshake->lock crngs.lock irq_context: 0 (wq_completion)wg-kex-wg2#7 (work_completion)(&peer->transmit_handshake_work) &wg->static_identity.lock &handshake->lock tk_core.seq.seqcount irq_context: 0 (wq_completion)wg-kex-wg2#7 (work_completion)(&peer->transmit_handshake_work) &wg->static_identity.lock &handshake->lock &table->lock#2 irq_context: 0 (wq_completion)wg-crypt-wg1#5 irq_context: 0 (wq_completion)wg-kex-wg2#7 (work_completion)(&peer->transmit_handshake_work) &wg->static_identity.lock &handshake->lock rcu_read_lock_bh batched_entropy_u32.lock irq_context: 0 (wq_completion)wg-crypt-wg1#5 (work_completion)(&({ do { const void *__vpp_verify = (typeof((worker) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __asm__ ("" : "=r"(__ptr) : "0"((typeof(*((worker))) *)((worker)))); (typeof((typeof(*((worker))) *)((worker)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); })->work) irq_context: 0 (wq_completion)wg-kex-wg2#7 (work_completion)(&peer->transmit_handshake_work) &wg->static_identity.lock &handshake->lock rcu_read_lock_bh &table->lock#2 irq_context: 0 (wq_completion)wg-crypt-wg1#5 (work_completion)(&({ do { const void *__vpp_verify = (typeof((worker) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __asm__ ("" : "=r"(__ptr) : "0"((typeof(*((worker))) *)((worker)))); (typeof((typeof(*((worker))) *)((worker)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); })->work) &r->consumer_lock#2 irq_context: 0 (wq_completion)wg-kex-wg2#7 (work_completion)(&peer->transmit_handshake_work) &cookie->lock irq_context: 0 (wq_completion)wg-kex-wg2#7 (work_completion)(&peer->transmit_handshake_work) &cookie->lock irq_context: 0 (wq_completion)wg-crypt-wg1#5 (work_completion)(&({ do { const void *__vpp_verify = (typeof((worker) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __asm__ ("" : "=r"(__ptr) : "0"((typeof(*((worker))) *)((worker)))); (typeof((typeof(*((worker))) *)((worker)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); })->work) rcu_read_lock &pool->lock irq_context: 0 (wq_completion)wg-kex-wg2#7 (work_completion)(&peer->transmit_handshake_work) rcu_read_lock_bh &obj_hash[i].lock irq_context: 0 (wq_completion)wg-crypt-wg1#5 (work_completion)(&({ do { const void *__vpp_verify = (typeof((worker) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __asm__ ("" : "=r"(__ptr) : "0"((typeof(*((worker))) *)((worker)))); (typeof((typeof(*((worker))) *)((worker)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); })->work) rcu_read_lock &pool->lock &obj_hash[i].lock irq_context: 0 (wq_completion)wg-kex-wg2#7 (work_completion)(&peer->transmit_handshake_work) rcu_read_lock_bh &base->lock irq_context: 0 (wq_completion)wg-crypt-wg1#5 (work_completion)(&({ do { const void *__vpp_verify = (typeof((worker) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __asm__ ("" : "=r"(__ptr) : "0"((typeof(*((worker))) *)((worker)))); (typeof((typeof(*((worker))) *)((worker)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); })->work) rcu_read_lock &pool->lock &p->pi_lock irq_context: 0 (wq_completion)wg-kex-wg2#7 (work_completion)(&peer->transmit_handshake_work) rcu_read_lock_bh &base->lock &obj_hash[i].lock irq_context: 0 (wq_completion)wg-crypt-wg1#5 (work_completion)(&({ do { const void *__vpp_verify = (typeof((worker) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __asm__ ("" : "=r"(__ptr) : "0"((typeof(*((worker))) *)((worker)))); (typeof((typeof(*((worker))) *)((worker)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); })->work) rcu_read_lock &pool->lock &p->pi_lock &rq->__lock irq_context: 0 (wq_completion)wg-kex-wg2#7 (work_completion)(&peer->transmit_handshake_work) &obj_hash[i].lock irq_context: 0 (wq_completion)wg-crypt-wg1#5 (work_completion)(&({ do { const void *__vpp_verify = (typeof((worker) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __asm__ ("" : "=r"(__ptr) : "0"((typeof(*((worker))) *)((worker)))); (typeof((typeof(*((worker))) *)((worker)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); })->work) rcu_read_lock &pool->lock &p->pi_lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)wg-kex-wg2#7 (work_completion)(&peer->transmit_handshake_work) &peer->endpoint_lock irq_context: 0 (wq_completion)wg-kex-wg2#7 (work_completion)(&peer->transmit_handshake_work) &peer->endpoint_lock rcu_read_lock_bh rcu_read_lock &____s->seqcount#7 irq_context: 0 (wq_completion)wg-kex-wg2#7 (work_completion)(&peer->transmit_handshake_work) &peer->endpoint_lock rcu_read_lock_bh rcu_read_lock &____s->seqcount#10 irq_context: 0 (wq_completion)wg-kex-wg2#7 (work_completion)(&peer->transmit_handshake_work) &peer->endpoint_lock rcu_read_lock_bh rcu_read_lock rcu_read_lock_bh tk_core.seq.seqcount irq_context: 0 (wq_completion)wg-kex-wg2#7 (work_completion)(&peer->transmit_handshake_work) &peer->endpoint_lock rcu_read_lock_bh rcu_read_lock rcu_read_lock_bh &list->lock#5 irq_context: 0 (wq_completion)wg-kex-wg2#7 (work_completion)(&peer->transmit_handshake_work) &rq->__lock irq_context: 0 (wq_completion)wg-crypt-wg1#5 (work_completion)(&peer->transmit_packet_work) irq_context: 0 (wq_completion)wg-crypt-wg1#5 (work_completion)(&peer->transmit_packet_work) rcu_read_lock_bh &obj_hash[i].lock irq_context: 0 (wq_completion)wg-crypt-wg1#5 (work_completion)(&peer->transmit_packet_work) &obj_hash[i].lock irq_context: 0 (wq_completion)wg-crypt-wg1#5 (work_completion)(&peer->transmit_packet_work) &peer->endpoint_lock irq_context: 0 (wq_completion)wg-crypt-wg1#5 (work_completion)(&peer->transmit_packet_work) &peer->endpoint_lock rcu_read_lock_bh rcu_read_lock &____s->seqcount#7 irq_context: 0 (wq_completion)wg-kex-wg1#11 (work_completion)(&({ do { const void *__vpp_verify = (typeof((worker) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __asm__ ("" : "=r"(__ptr) : "0"((typeof(*((worker))) *)((worker)))); (typeof((typeof(*((worker))) *)((worker)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); })->work) &r->consumer_lock#2 irq_context: 0 (wq_completion)wg-crypt-wg1#5 (work_completion)(&peer->transmit_packet_work) &peer->endpoint_lock rcu_read_lock_bh rcu_read_lock &____s->seqcount#10 irq_context: 0 (wq_completion)wg-kex-wg1#11 (work_completion)(&({ do { const void *__vpp_verify = (typeof((worker) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __asm__ ("" : "=r"(__ptr) : "0"((typeof(*((worker))) *)((worker)))); (typeof((typeof(*((worker))) *)((worker)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); })->work) &wg->static_identity.lock irq_context: 0 (wq_completion)wg-crypt-wg1#5 (work_completion)(&peer->transmit_packet_work) &peer->endpoint_lock rcu_read_lock_bh rcu_read_lock rcu_read_lock_bh tk_core.seq.seqcount irq_context: 0 (wq_completion)wg-kex-wg1#11 (work_completion)(&({ do { const void *__vpp_verify = (typeof((worker) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __asm__ ("" : "=r"(__ptr) : "0"((typeof(*((worker))) *)((worker)))); (typeof((typeof(*((worker))) *)((worker)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); })->work) &wg->static_identity.lock &handshake->lock irq_context: 0 (wq_completion)wg-crypt-wg1#5 (work_completion)(&peer->transmit_packet_work) &peer->endpoint_lock rcu_read_lock_bh rcu_read_lock rcu_read_lock_bh &list->lock#5 irq_context: 0 (wq_completion)wg-kex-wg1#11 (work_completion)(&({ do { const void *__vpp_verify = (typeof((worker) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __asm__ ("" : "=r"(__ptr) : "0"((typeof(*((worker))) *)((worker)))); (typeof((typeof(*((worker))) *)((worker)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); })->work) &wg->static_identity.lock &rq->__lock irq_context: 0 (wq_completion)wg-crypt-wg1#5 (work_completion)(&peer->transmit_packet_work) rcu_read_lock_bh tk_core.seq.seqcount irq_context: 0 (wq_completion)wg-crypt-wg0#5 irq_context: 0 (wq_completion)wg-crypt-wg0#5 (work_completion)(&({ do { const void *__vpp_verify = (typeof((worker) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __asm__ ("" : "=r"(__ptr) : "0"((typeof(*((worker))) *)((worker)))); (typeof((typeof(*((worker))) *)((worker)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); })->work) irq_context: 0 (wq_completion)wg-crypt-wg0#5 (work_completion)(&({ do { const void *__vpp_verify = (typeof((worker) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __asm__ ("" : "=r"(__ptr) : "0"((typeof(*((worker))) *)((worker)))); (typeof((typeof(*((worker))) *)((worker)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); })->work) &r->consumer_lock#2 irq_context: 0 (wq_completion)wg-crypt-wg0#5 (work_completion)(&({ do { const void *__vpp_verify = (typeof((worker) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __asm__ ("" : "=r"(__ptr) : "0"((typeof(*((worker))) *)((worker)))); (typeof((typeof(*((worker))) *)((worker)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); })->work) tk_core.seq.seqcount irq_context: 0 (wq_completion)wg-crypt-wg0#5 (work_completion)(&({ do { const void *__vpp_verify = (typeof((worker) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __asm__ ("" : "=r"(__ptr) : "0"((typeof(*((worker))) *)((worker)))); (typeof((typeof(*((worker))) *)((worker)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); })->work) rcu_read_lock &pool->lock irq_context: 0 (wq_completion)wg-crypt-wg0#5 (work_completion)(&({ do { const void *__vpp_verify = (typeof((worker) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __asm__ ("" : "=r"(__ptr) : "0"((typeof(*((worker))) *)((worker)))); (typeof((typeof(*((worker))) *)((worker)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); })->work) rcu_read_lock &pool->lock &obj_hash[i].lock irq_context: 0 (wq_completion)wg-crypt-wg0#5 (work_completion)(&({ do { const void *__vpp_verify = (typeof((worker) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __asm__ ("" : "=r"(__ptr) : "0"((typeof(*((worker))) *)((worker)))); (typeof((typeof(*((worker))) *)((worker)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); })->work) rcu_read_lock &pool->lock &p->pi_lock irq_context: 0 (wq_completion)wg-kex-wg1#11 (work_completion)(&({ do { const void *__vpp_verify = (typeof((worker) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __asm__ ("" : "=r"(__ptr) : "0"((typeof(*((worker))) *)((worker)))); (typeof((typeof(*((worker))) *)((worker)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); })->work) &wg->static_identity.lock &handshake->lock irq_context: 0 (wq_completion)wg-kex-wg1#11 (work_completion)(&({ do { const void *__vpp_verify = (typeof((worker) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __asm__ ("" : "=r"(__ptr) : "0"((typeof(*((worker))) *)((worker)))); (typeof((typeof(*((worker))) *)((worker)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); })->work) &peer->endpoint_lock irq_context: 0 (wq_completion)wg-crypt-wg0#5 (work_completion)(&peer->transmit_packet_work) irq_context: 0 (wq_completion)wg-kex-wg1#11 (work_completion)(&({ do { const void *__vpp_verify = (typeof((worker) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __asm__ ("" : "=r"(__ptr) : "0"((typeof(*((worker))) *)((worker)))); (typeof((typeof(*((worker))) *)((worker)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); })->work) &handshake->lock irq_context: 0 (wq_completion)wg-crypt-wg0#5 (work_completion)(&peer->transmit_packet_work) rcu_read_lock_bh &obj_hash[i].lock irq_context: 0 (wq_completion)wg-kex-wg1#11 (work_completion)(&({ do { const void *__vpp_verify = (typeof((worker) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __asm__ ("" : "=r"(__ptr) : "0"((typeof(*((worker))) *)((worker)))); (typeof((typeof(*((worker))) *)((worker)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); })->work) &handshake->lock fs_reclaim irq_context: 0 (wq_completion)wg-crypt-wg0#5 (work_completion)(&peer->transmit_packet_work) &obj_hash[i].lock irq_context: 0 (wq_completion)wg-kex-wg1#11 (work_completion)(&({ do { const void *__vpp_verify = (typeof((worker) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __asm__ ("" : "=r"(__ptr) : "0"((typeof(*((worker))) *)((worker)))); (typeof((typeof(*((worker))) *)((worker)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); })->work) &handshake->lock fs_reclaim mmu_notifier_invalidate_range_start irq_context: 0 (wq_completion)wg-crypt-wg0#5 (work_completion)(&peer->transmit_packet_work) &peer->endpoint_lock irq_context: 0 (wq_completion)wg-kex-wg1#11 (work_completion)(&({ do { const void *__vpp_verify = (typeof((worker) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __asm__ ("" : "=r"(__ptr) : "0"((typeof(*((worker))) *)((worker)))); (typeof((typeof(*((worker))) *)((worker)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); })->work) &handshake->lock tk_core.seq.seqcount irq_context: 0 (wq_completion)wg-crypt-wg0#5 (work_completion)(&peer->transmit_packet_work) &peer->endpoint_lock rcu_read_lock_bh rcu_read_lock &____s->seqcount#7 irq_context: 0 (wq_completion)wg-kex-wg1#11 (work_completion)(&({ do { const void *__vpp_verify = (typeof((worker) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __asm__ ("" : "=r"(__ptr) : "0"((typeof(*((worker))) *)((worker)))); (typeof((typeof(*((worker))) *)((worker)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); })->work) &handshake->lock rcu_read_lock_bh &peer->keypairs.keypair_update_lock irq_context: 0 (wq_completion)wg-crypt-wg0#5 (work_completion)(&peer->transmit_packet_work) &peer->endpoint_lock rcu_read_lock_bh rcu_read_lock &____s->seqcount#10 irq_context: 0 (wq_completion)wg-kex-wg1#11 (work_completion)(&({ do { const void *__vpp_verify = (typeof((worker) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __asm__ ("" : "=r"(__ptr) : "0"((typeof(*((worker))) *)((worker)))); (typeof((typeof(*((worker))) *)((worker)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); })->work) &handshake->lock rcu_read_lock_bh &table->lock#2 irq_context: 0 (wq_completion)wg-crypt-wg0#5 (work_completion)(&peer->transmit_packet_work) &peer->endpoint_lock rcu_read_lock_bh rcu_read_lock rcu_read_lock_bh tk_core.seq.seqcount irq_context: 0 (wq_completion)wg-kex-wg1#11 (work_completion)(&({ do { const void *__vpp_verify = (typeof((worker) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __asm__ ("" : "=r"(__ptr) : "0"((typeof(*((worker))) *)((worker)))); (typeof((typeof(*((worker))) *)((worker)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); })->work) rcu_read_lock_bh &obj_hash[i].lock irq_context: 0 (wq_completion)wg-crypt-wg0#5 (work_completion)(&peer->transmit_packet_work) &peer->endpoint_lock rcu_read_lock_bh rcu_read_lock rcu_read_lock_bh &list->lock#5 irq_context: 0 (wq_completion)wg-kex-wg1#11 (work_completion)(&({ do { const void *__vpp_verify = (typeof((worker) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __asm__ ("" : "=r"(__ptr) : "0"((typeof(*((worker))) *)((worker)))); (typeof((typeof(*((worker))) *)((worker)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); })->work) rcu_read_lock_bh &base->lock irq_context: 0 (wq_completion)wg-kex-wg1#11 (work_completion)(&({ do { const void *__vpp_verify = (typeof((worker) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __asm__ ("" : "=r"(__ptr) : "0"((typeof(*((worker))) *)((worker)))); (typeof((typeof(*((worker))) *)((worker)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); })->work) rcu_read_lock_bh &base->lock &obj_hash[i].lock irq_context: 0 (wq_completion)wg-kex-wg1#11 (work_completion)(&({ do { const void *__vpp_verify = (typeof((worker) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __asm__ ("" : "=r"(__ptr) : "0"((typeof(*((worker))) *)((worker)))); (typeof((typeof(*((worker))) *)((worker)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); })->work) &obj_hash[i].lock irq_context: 0 (wq_completion)wg-kex-wg1#11 (work_completion)(&({ do { const void *__vpp_verify = (typeof((worker) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __asm__ ("" : "=r"(__ptr) : "0"((typeof(*((worker))) *)((worker)))); (typeof((typeof(*((worker))) *)((worker)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); })->work) &base->lock irq_context: 0 (wq_completion)wg-kex-wg1#11 (work_completion)(&({ do { const void *__vpp_verify = (typeof((worker) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __asm__ ("" : "=r"(__ptr) : "0"((typeof(*((worker))) *)((worker)))); (typeof((typeof(*((worker))) *)((worker)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); })->work) &base->lock &obj_hash[i].lock irq_context: 0 (wq_completion)wg-kex-wg1#11 (work_completion)(&({ do { const void *__vpp_verify = (typeof((worker) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __asm__ ("" : "=r"(__ptr) : "0"((typeof(*((worker))) *)((worker)))); (typeof((typeof(*((worker))) *)((worker)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); })->work) tk_core.seq.seqcount irq_context: 0 (wq_completion)wg-kex-wg2#8 irq_context: 0 (wq_completion)wg-kex-wg1#11 (work_completion)(&({ do { const void *__vpp_verify = (typeof((worker) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __asm__ ("" : "=r"(__ptr) : "0"((typeof(*((worker))) *)((worker)))); (typeof((typeof(*((worker))) *)((worker)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); })->work) &list->lock#17 irq_context: 0 (wq_completion)wg-kex-wg2#8 (work_completion)(&({ do { const void *__vpp_verify = (typeof((worker) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __asm__ ("" : "=r"(__ptr) : "0"((typeof(*((worker))) *)((worker)))); (typeof((typeof(*((worker))) *)((worker)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); })->work) irq_context: 0 (wq_completion)wg-kex-wg1#11 (work_completion)(&({ do { const void *__vpp_verify = (typeof((worker) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __asm__ ("" : "=r"(__ptr) : "0"((typeof(*((worker))) *)((worker)))); (typeof((typeof(*((worker))) *)((worker)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); })->work) rcu_read_lock_bh &r->producer_lock#2 irq_context: 0 (wq_completion)wg-kex-wg2#8 (work_completion)(&({ do { const void *__vpp_verify = (typeof((worker) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __asm__ ("" : "=r"(__ptr) : "0"((typeof(*((worker))) *)((worker)))); (typeof((typeof(*((worker))) *)((worker)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); })->work) &r->consumer_lock#2 irq_context: 0 (wq_completion)wg-kex-wg1#11 (work_completion)(&({ do { const void *__vpp_verify = (typeof((worker) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __asm__ ("" : "=r"(__ptr) : "0"((typeof(*((worker))) *)((worker)))); (typeof((typeof(*((worker))) *)((worker)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); })->work) rcu_read_lock_bh rcu_read_lock &pool->lock irq_context: 0 (wq_completion)wg-kex-wg2#8 (work_completion)(&({ do { const void *__vpp_verify = (typeof((worker) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __asm__ ("" : "=r"(__ptr) : "0"((typeof(*((worker))) *)((worker)))); (typeof((typeof(*((worker))) *)((worker)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); })->work) &wg->static_identity.lock irq_context: 0 (wq_completion)wg-kex-wg1#11 (work_completion)(&({ do { const void *__vpp_verify = (typeof((worker) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __asm__ ("" : "=r"(__ptr) : "0"((typeof(*((worker))) *)((worker)))); (typeof((typeof(*((worker))) *)((worker)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); })->work) rcu_read_lock_bh rcu_read_lock &pool->lock &obj_hash[i].lock irq_context: 0 (wq_completion)wg-kex-wg1#11 (work_completion)(&({ do { const void *__vpp_verify = (typeof((worker) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __asm__ ("" : "=r"(__ptr) : "0"((typeof(*((worker))) *)((worker)))); (typeof((typeof(*((worker))) *)((worker)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); })->work) rcu_read_lock_bh rcu_read_lock &pool->lock &p->pi_lock irq_context: 0 (wq_completion)wg-kex-wg2#8 (work_completion)(&({ do { const void *__vpp_verify = (typeof((worker) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __asm__ ("" : "=r"(__ptr) : "0"((typeof(*((worker))) *)((worker)))); (typeof((typeof(*((worker))) *)((worker)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); })->work) &wg->static_identity.lock &handshake->lock irq_context: 0 (wq_completion)wg-crypt-wg1#5 (work_completion)(&({ do { const void *__vpp_verify = (typeof((worker) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __asm__ ("" : "=r"(__ptr) : "0"((typeof(*((worker))) *)((worker)))); (typeof((typeof(*((worker))) *)((worker)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); })->work) tk_core.seq.seqcount irq_context: 0 (wq_completion)wg-kex-wg2#8 (work_completion)(&({ do { const void *__vpp_verify = (typeof((worker) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __asm__ ("" : "=r"(__ptr) : "0"((typeof(*((worker))) *)((worker)))); (typeof((typeof(*((worker))) *)((worker)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); })->work) &wg->static_identity.lock &handshake->lock tk_core.seq.seqcount irq_context: 0 (wq_completion)wg-kex-wg2#8 (work_completion)(&({ do { const void *__vpp_verify = (typeof((worker) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __asm__ ("" : "=r"(__ptr) : "0"((typeof(*((worker))) *)((worker)))); (typeof((typeof(*((worker))) *)((worker)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); })->work) &wg->static_identity.lock &handshake->lock irq_context: 0 (wq_completion)wg-kex-wg2#8 (work_completion)(&({ do { const void *__vpp_verify = (typeof((worker) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __asm__ ("" : "=r"(__ptr) : "0"((typeof(*((worker))) *)((worker)))); (typeof((typeof(*((worker))) *)((worker)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); })->work) &wg->static_identity.lock &handshake->lock tk_core.seq.seqcount irq_context: 0 (wq_completion)wg-crypt-wg1#6 irq_context: 0 (wq_completion)wg-kex-wg2#8 (work_completion)(&({ do { const void *__vpp_verify = (typeof((worker) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __asm__ ("" : "=r"(__ptr) : "0"((typeof(*((worker))) *)((worker)))); (typeof((typeof(*((worker))) *)((worker)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); })->work) &wg->static_identity.lock &handshake->lock &rq->__lock irq_context: 0 (wq_completion)wg-crypt-wg1#6 &rq->__lock irq_context: 0 (wq_completion)wg-crypt-wg1#6 &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)wg-kex-wg2#8 (work_completion)(&({ do { const void *__vpp_verify = (typeof((worker) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __asm__ ("" : "=r"(__ptr) : "0"((typeof(*((worker))) *)((worker)))); (typeof((typeof(*((worker))) *)((worker)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); })->work) &peer->endpoint_lock irq_context: 0 (wq_completion)wg-kex-wg2#8 (work_completion)(&({ do { const void *__vpp_verify = (typeof((worker) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __asm__ ("" : "=r"(__ptr) : "0"((typeof(*((worker))) *)((worker)))); (typeof((typeof(*((worker))) *)((worker)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); })->work) &handshake->lock irq_context: 0 (wq_completion)wg-kex-wg2#8 (work_completion)(&({ do { const void *__vpp_verify = (typeof((worker) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __asm__ ("" : "=r"(__ptr) : "0"((typeof(*((worker))) *)((worker)))); (typeof((typeof(*((worker))) *)((worker)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); })->work) &handshake->lock fs_reclaim irq_context: 0 (wq_completion)wg-kex-wg2#8 (work_completion)(&({ do { const void *__vpp_verify = (typeof((worker) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __asm__ ("" : "=r"(__ptr) : "0"((typeof(*((worker))) *)((worker)))); (typeof((typeof(*((worker))) *)((worker)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); })->work) &handshake->lock fs_reclaim mmu_notifier_invalidate_range_start irq_context: 0 (wq_completion)wg-kex-wg2#8 (work_completion)(&({ do { const void *__vpp_verify = (typeof((worker) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __asm__ ("" : "=r"(__ptr) : "0"((typeof(*((worker))) *)((worker)))); (typeof((typeof(*((worker))) *)((worker)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); })->work) &handshake->lock tk_core.seq.seqcount irq_context: 0 (wq_completion)wg-kex-wg2#8 (work_completion)(&({ do { const void *__vpp_verify = (typeof((worker) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __asm__ ("" : "=r"(__ptr) : "0"((typeof(*((worker))) *)((worker)))); (typeof((typeof(*((worker))) *)((worker)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); })->work) &handshake->lock rcu_read_lock_bh &peer->keypairs.keypair_update_lock irq_context: 0 (wq_completion)wg-kex-wg2#8 (work_completion)(&({ do { const void *__vpp_verify = (typeof((worker) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __asm__ ("" : "=r"(__ptr) : "0"((typeof(*((worker))) *)((worker)))); (typeof((typeof(*((worker))) *)((worker)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); })->work) &handshake->lock rcu_read_lock_bh &table->lock#2 irq_context: 0 (wq_completion)wg-kex-wg2#8 (work_completion)(&({ do { const void *__vpp_verify = (typeof((worker) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __asm__ ("" : "=r"(__ptr) : "0"((typeof(*((worker))) *)((worker)))); (typeof((typeof(*((worker))) *)((worker)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); })->work) rcu_read_lock_bh &obj_hash[i].lock irq_context: 0 (wq_completion)wg-kex-wg2#8 (work_completion)(&({ do { const void *__vpp_verify = (typeof((worker) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __asm__ ("" : "=r"(__ptr) : "0"((typeof(*((worker))) *)((worker)))); (typeof((typeof(*((worker))) *)((worker)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); })->work) rcu_read_lock_bh &base->lock irq_context: 0 (wq_completion)wg-kex-wg2#8 (work_completion)(&({ do { const void *__vpp_verify = (typeof((worker) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __asm__ ("" : "=r"(__ptr) : "0"((typeof(*((worker))) *)((worker)))); (typeof((typeof(*((worker))) *)((worker)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); })->work) rcu_read_lock_bh &base->lock &obj_hash[i].lock irq_context: 0 (wq_completion)wg-kex-wg2#8 (work_completion)(&({ do { const void *__vpp_verify = (typeof((worker) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __asm__ ("" : "=r"(__ptr) : "0"((typeof(*((worker))) *)((worker)))); (typeof((typeof(*((worker))) *)((worker)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); })->work) &obj_hash[i].lock irq_context: 0 (wq_completion)wg-kex-wg2#8 (work_completion)(&({ do { const void *__vpp_verify = (typeof((worker) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __asm__ ("" : "=r"(__ptr) : "0"((typeof(*((worker))) *)((worker)))); (typeof((typeof(*((worker))) *)((worker)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); })->work) tk_core.seq.seqcount irq_context: 0 (wq_completion)wg-kex-wg2#8 (work_completion)(&({ do { const void *__vpp_verify = (typeof((worker) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __asm__ ("" : "=r"(__ptr) : "0"((typeof(*((worker))) *)((worker)))); (typeof((typeof(*((worker))) *)((worker)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); })->work) &list->lock#17 irq_context: 0 (wq_completion)wg-kex-wg2#8 (work_completion)(&({ do { const void *__vpp_verify = (typeof((worker) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __asm__ ("" : "=r"(__ptr) : "0"((typeof(*((worker))) *)((worker)))); (typeof((typeof(*((worker))) *)((worker)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); })->work) rcu_read_lock_bh &r->producer_lock#2 irq_context: 0 (wq_completion)wg-kex-wg2#8 (work_completion)(&({ do { const void *__vpp_verify = (typeof((worker) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __asm__ ("" : "=r"(__ptr) : "0"((typeof(*((worker))) *)((worker)))); (typeof((typeof(*((worker))) *)((worker)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); })->work) rcu_read_lock_bh rcu_read_lock &pool->lock irq_context: 0 (wq_completion)wg-kex-wg2#8 (work_completion)(&({ do { const void *__vpp_verify = (typeof((worker) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __asm__ ("" : "=r"(__ptr) : "0"((typeof(*((worker))) *)((worker)))); (typeof((typeof(*((worker))) *)((worker)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); })->work) rcu_read_lock_bh rcu_read_lock &pool->lock &obj_hash[i].lock irq_context: 0 (wq_completion)wg-kex-wg2#8 (work_completion)(&({ do { const void *__vpp_verify = (typeof((worker) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __asm__ ("" : "=r"(__ptr) : "0"((typeof(*((worker))) *)((worker)))); (typeof((typeof(*((worker))) *)((worker)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); })->work) rcu_read_lock_bh rcu_read_lock &pool->lock &p->pi_lock irq_context: 0 (wq_completion)wg-kex-wg2#8 (work_completion)(&({ do { const void *__vpp_verify = (typeof((worker) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __asm__ ("" : "=r"(__ptr) : "0"((typeof(*((worker))) *)((worker)))); (typeof((typeof(*((worker))) *)((worker)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); })->work) rcu_read_lock_bh rcu_read_lock &pool->lock &p->pi_lock &rq->__lock irq_context: 0 (wq_completion)wg-kex-wg2#8 (work_completion)(&({ do { const void *__vpp_verify = (typeof((worker) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __asm__ ("" : "=r"(__ptr) : "0"((typeof(*((worker))) *)((worker)))); (typeof((typeof(*((worker))) *)((worker)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); })->work) rcu_read_lock_bh rcu_read_lock &pool->lock &p->pi_lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)wg-kex-wg2#8 (work_completion)(&({ do { const void *__vpp_verify = (typeof((worker) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __asm__ ("" : "=r"(__ptr) : "0"((typeof(*((worker))) *)((worker)))); (typeof((typeof(*((worker))) *)((worker)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); })->work) &rq->__lock irq_context: 0 (wq_completion)wg-crypt-wg2#4 irq_context: 0 (wq_completion)wg-crypt-wg2#4 (work_completion)(&({ do { const void *__vpp_verify = (typeof((worker) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __asm__ ("" : "=r"(__ptr) : "0"((typeof(*((worker))) *)((worker)))); (typeof((typeof(*((worker))) *)((worker)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); })->work) irq_context: 0 (wq_completion)wg-crypt-wg2#4 (work_completion)(&({ do { const void *__vpp_verify = (typeof((worker) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __asm__ ("" : "=r"(__ptr) : "0"((typeof(*((worker))) *)((worker)))); (typeof((typeof(*((worker))) *)((worker)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); })->work) &r->consumer_lock#2 irq_context: 0 (wq_completion)wg-crypt-wg2#4 (work_completion)(&({ do { const void *__vpp_verify = (typeof((worker) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __asm__ ("" : "=r"(__ptr) : "0"((typeof(*((worker))) *)((worker)))); (typeof((typeof(*((worker))) *)((worker)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); })->work) rcu_read_lock &pool->lock irq_context: 0 (wq_completion)wg-crypt-wg2#4 (work_completion)(&({ do { const void *__vpp_verify = (typeof((worker) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __asm__ ("" : "=r"(__ptr) : "0"((typeof(*((worker))) *)((worker)))); (typeof((typeof(*((worker))) *)((worker)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); })->work) rcu_read_lock &pool->lock &obj_hash[i].lock irq_context: 0 (wq_completion)wg-crypt-wg2#4 (work_completion)(&({ do { const void *__vpp_verify = (typeof((worker) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __asm__ ("" : "=r"(__ptr) : "0"((typeof(*((worker))) *)((worker)))); (typeof((typeof(*((worker))) *)((worker)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); })->work) rcu_read_lock &pool->lock &p->pi_lock irq_context: 0 (wq_completion)wg-crypt-wg1#6 (work_completion)(&({ do { const void *__vpp_verify = (typeof((worker) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __asm__ ("" : "=r"(__ptr) : "0"((typeof(*((worker))) *)((worker)))); (typeof((typeof(*((worker))) *)((worker)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); })->work) irq_context: 0 (wq_completion)wg-crypt-wg1#6 (work_completion)(&({ do { const void *__vpp_verify = (typeof((worker) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __asm__ ("" : "=r"(__ptr) : "0"((typeof(*((worker))) *)((worker)))); (typeof((typeof(*((worker))) *)((worker)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); })->work) &r->consumer_lock#2 irq_context: 0 (wq_completion)wg-crypt-wg1#6 (work_completion)(&({ do { const void *__vpp_verify = (typeof((worker) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __asm__ ("" : "=r"(__ptr) : "0"((typeof(*((worker))) *)((worker)))); (typeof((typeof(*((worker))) *)((worker)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); })->work) rcu_read_lock &pool->lock irq_context: 0 (wq_completion)wg-crypt-wg1#6 (work_completion)(&({ do { const void *__vpp_verify = (typeof((worker) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __asm__ ("" : "=r"(__ptr) : "0"((typeof(*((worker))) *)((worker)))); (typeof((typeof(*((worker))) *)((worker)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); })->work) rcu_read_lock &pool->lock &obj_hash[i].lock irq_context: 0 (wq_completion)wg-crypt-wg1#6 (work_completion)(&({ do { const void *__vpp_verify = (typeof((worker) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __asm__ ("" : "=r"(__ptr) : "0"((typeof(*((worker))) *)((worker)))); (typeof((typeof(*((worker))) *)((worker)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); })->work) rcu_read_lock &pool->lock &p->pi_lock irq_context: 0 misc_mtx rfkill_global_mutex &data->mtx &c->lock irq_context: 0 misc_mtx rfkill_global_mutex &data->mtx &____s->seqcount irq_context: 0 (wq_completion)wg-kex-wg2#7 (work_completion)(&peer->transmit_handshake_work) &wg->static_identity.lock &handshake->lock &sem->wait_lock irq_context: 0 (wq_completion)wg-kex-wg2#7 (work_completion)(&peer->transmit_handshake_work) &wg->static_identity.lock &handshake->lock &rq->__lock irq_context: 0 (wq_completion)wg-kex-wg2#7 (work_completion)(&peer->transmit_handshake_work) &wg->static_identity.lock &handshake->lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)wg-crypt-wg1#6 (work_completion)(&peer->transmit_packet_work) irq_context: 0 (wq_completion)wg-crypt-wg1#6 (work_completion)(&peer->transmit_packet_work) rcu_read_lock_bh &obj_hash[i].lock irq_context: 0 (wq_completion)wg-crypt-wg1#6 (work_completion)(&peer->transmit_packet_work) rcu_read_lock_bh &base->lock irq_context: 0 (wq_completion)wg-crypt-wg1#6 (work_completion)(&peer->transmit_packet_work) &obj_hash[i].lock irq_context: 0 (wq_completion)wg-crypt-wg1#6 (work_completion)(&peer->transmit_packet_work) &peer->endpoint_lock irq_context: 0 (wq_completion)wg-crypt-wg1#6 (work_completion)(&peer->transmit_packet_work) &peer->endpoint_lock rcu_read_lock_bh rcu_read_lock &____s->seqcount#7 irq_context: 0 (wq_completion)wg-crypt-wg1#6 (work_completion)(&peer->transmit_packet_work) &peer->endpoint_lock rcu_read_lock_bh rcu_read_lock &____s->seqcount#10 irq_context: 0 (wq_completion)wg-crypt-wg1#6 (work_completion)(&peer->transmit_packet_work) &peer->endpoint_lock rcu_read_lock_bh rcu_read_lock rcu_read_lock_bh tk_core.seq.seqcount irq_context: 0 (wq_completion)wg-crypt-wg1#6 (work_completion)(&peer->transmit_packet_work) &peer->endpoint_lock rcu_read_lock_bh rcu_read_lock rcu_read_lock_bh &list->lock#5 irq_context: 0 (wq_completion)bat_events (work_completion)(&(&forw_packet_aggr->delayed_work)->work) &hard_iface->bat_iv.ogm_buff_mutex &bat_priv->tvlv.container_list_lock &c->lock irq_context: 0 (wq_completion)bat_events (work_completion)(&(&forw_packet_aggr->delayed_work)->work) &hard_iface->bat_iv.ogm_buff_mutex &bat_priv->tvlv.container_list_lock &____s->seqcount#2 irq_context: 0 (wq_completion)bat_events (work_completion)(&(&forw_packet_aggr->delayed_work)->work) &hard_iface->bat_iv.ogm_buff_mutex &bat_priv->tvlv.container_list_lock &n->list_lock irq_context: 0 (wq_completion)wg-crypt-wg1#6 (work_completion)(&peer->transmit_packet_work) rcu_read_lock_bh tk_core.seq.seqcount irq_context: 0 (wq_completion)bat_events (work_completion)(&(&forw_packet_aggr->delayed_work)->work) &hard_iface->bat_iv.ogm_buff_mutex &bat_priv->tvlv.container_list_lock &n->list_lock &c->lock irq_context: 0 (wq_completion)wg-crypt-wg2#4 (work_completion)(&peer->transmit_packet_work) irq_context: 0 (wq_completion)wg-crypt-wg2#4 (work_completion)(&peer->transmit_packet_work) rcu_read_lock_bh &obj_hash[i].lock irq_context: 0 (wq_completion)wg-crypt-wg2#4 (work_completion)(&peer->transmit_packet_work) &obj_hash[i].lock irq_context: 0 (wq_completion)wg-crypt-wg2#4 (work_completion)(&peer->transmit_packet_work) &peer->endpoint_lock irq_context: 0 (wq_completion)wg-crypt-wg0#6 irq_context: 0 (wq_completion)wg-crypt-wg0#6 (work_completion)(&({ do { const void *__vpp_verify = (typeof((worker) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __asm__ ("" : "=r"(__ptr) : "0"((typeof(*((worker))) *)((worker)))); (typeof((typeof(*((worker))) *)((worker)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); })->work) irq_context: 0 (wq_completion)wg-crypt-wg2#4 (work_completion)(&peer->transmit_packet_work) &peer->endpoint_lock rcu_read_lock_bh rcu_read_lock &____s->seqcount#7 irq_context: 0 (wq_completion)wg-crypt-wg0#6 (work_completion)(&({ do { const void *__vpp_verify = (typeof((worker) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __asm__ ("" : "=r"(__ptr) : "0"((typeof(*((worker))) *)((worker)))); (typeof((typeof(*((worker))) *)((worker)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); })->work) &r->consumer_lock#2 irq_context: 0 (wq_completion)wg-crypt-wg2#4 (work_completion)(&peer->transmit_packet_work) &peer->endpoint_lock rcu_read_lock_bh rcu_read_lock &____s->seqcount#10 irq_context: 0 (wq_completion)wg-crypt-wg0#6 (work_completion)(&({ do { const void *__vpp_verify = (typeof((worker) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __asm__ ("" : "=r"(__ptr) : "0"((typeof(*((worker))) *)((worker)))); (typeof((typeof(*((worker))) *)((worker)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); })->work) tk_core.seq.seqcount irq_context: 0 (wq_completion)wg-crypt-wg2#4 (work_completion)(&peer->transmit_packet_work) &peer->endpoint_lock rcu_read_lock_bh rcu_read_lock rcu_read_lock_bh tk_core.seq.seqcount irq_context: 0 (wq_completion)wg-crypt-wg2#4 (work_completion)(&peer->transmit_packet_work) &peer->endpoint_lock rcu_read_lock_bh rcu_read_lock rcu_read_lock_bh &list->lock#5 irq_context: 0 (wq_completion)wg-crypt-wg2#4 (work_completion)(&peer->transmit_packet_work) &rq->__lock irq_context: 0 (wq_completion)wg-crypt-wg0#6 (work_completion)(&({ do { const void *__vpp_verify = (typeof((worker) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __asm__ ("" : "=r"(__ptr) : "0"((typeof(*((worker))) *)((worker)))); (typeof((typeof(*((worker))) *)((worker)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); })->work) &rq->__lock irq_context: 0 (wq_completion)wg-crypt-wg0#6 (work_completion)(&({ do { const void *__vpp_verify = (typeof((worker) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __asm__ ("" : "=r"(__ptr) : "0"((typeof(*((worker))) *)((worker)))); (typeof((typeof(*((worker))) *)((worker)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); })->work) rcu_read_lock &pool->lock irq_context: 0 (wq_completion)wg-crypt-wg0#6 (work_completion)(&({ do { const void *__vpp_verify = (typeof((worker) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __asm__ ("" : "=r"(__ptr) : "0"((typeof(*((worker))) *)((worker)))); (typeof((typeof(*((worker))) *)((worker)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); })->work) rcu_read_lock &pool->lock &obj_hash[i].lock irq_context: 0 (wq_completion)wg-crypt-wg0#6 (work_completion)(&({ do { const void *__vpp_verify = (typeof((worker) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __asm__ ("" : "=r"(__ptr) : "0"((typeof(*((worker))) *)((worker)))); (typeof((typeof(*((worker))) *)((worker)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); })->work) rcu_read_lock &pool->lock &p->pi_lock irq_context: 0 (wq_completion)wg-crypt-wg0#5 (work_completion)(&({ do { const void *__vpp_verify = (typeof((worker) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __asm__ ("" : "=r"(__ptr) : "0"((typeof(*((worker))) *)((worker)))); (typeof((typeof(*((worker))) *)((worker)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); })->work) rcu_read_lock &pool->lock &p->pi_lock &rq->__lock irq_context: 0 (wq_completion)wg-crypt-wg0#5 (work_completion)(&({ do { const void *__vpp_verify = (typeof((worker) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __asm__ ("" : "=r"(__ptr) : "0"((typeof(*((worker))) *)((worker)))); (typeof((typeof(*((worker))) *)((worker)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); })->work) rcu_read_lock &pool->lock &p->pi_lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)wg-crypt-wg0#6 (work_completion)(&peer->transmit_packet_work) irq_context: 0 (wq_completion)wg-crypt-wg0#6 (work_completion)(&peer->transmit_packet_work) rcu_read_lock_bh &obj_hash[i].lock irq_context: 0 (wq_completion)wg-crypt-wg0#6 (work_completion)(&peer->transmit_packet_work) &obj_hash[i].lock irq_context: 0 (wq_completion)wg-crypt-wg0#6 (work_completion)(&peer->transmit_packet_work) &peer->endpoint_lock irq_context: 0 (wq_completion)wg-crypt-wg0#6 (work_completion)(&peer->transmit_packet_work) &peer->endpoint_lock rcu_read_lock_bh &obj_hash[i].lock irq_context: 0 (wq_completion)wg-crypt-wg2#4 (work_completion)(&({ do { const void *__vpp_verify = (typeof((worker) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __asm__ ("" : "=r"(__ptr) : "0"((typeof(*((worker))) *)((worker)))); (typeof((typeof(*((worker))) *)((worker)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); })->work) tk_core.seq.seqcount irq_context: 0 (wq_completion)wg-crypt-wg0#6 (work_completion)(&peer->transmit_packet_work) &peer->endpoint_lock rcu_read_lock_bh rcu_read_lock &____s->seqcount#7 irq_context: 0 (wq_completion)wg-crypt-wg0#6 (work_completion)(&peer->transmit_packet_work) &peer->endpoint_lock rcu_read_lock_bh rcu_read_lock &____s->seqcount#10 irq_context: 0 (wq_completion)wg-crypt-wg0#6 (work_completion)(&peer->transmit_packet_work) &peer->endpoint_lock rcu_read_lock_bh rcu_read_lock rcu_read_lock_bh tk_core.seq.seqcount irq_context: 0 (wq_completion)wg-crypt-wg0#6 (work_completion)(&peer->transmit_packet_work) &peer->endpoint_lock rcu_read_lock_bh rcu_read_lock rcu_read_lock_bh &list->lock#5 irq_context: 0 (wq_completion)wg-crypt-wg1#6 (work_completion)(&({ do { const void *__vpp_verify = (typeof((worker) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __asm__ ("" : "=r"(__ptr) : "0"((typeof(*((worker))) *)((worker)))); (typeof((typeof(*((worker))) *)((worker)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); })->work) tk_core.seq.seqcount irq_context: 0 (wq_completion)wg-kex-wg2#7 (work_completion)(&peer->transmit_handshake_work) batched_entropy_u8.lock irq_context: 0 (wq_completion)ipv6_addrconf (work_completion)(&(&ifa->dad_work)->work) rtnl_mutex rcu_read_lock rcu_read_lock rcu_read_lock_bh rcu_read_lock_bh rcu_read_lock &pool->lock &p->pi_lock irq_context: 0 (wq_completion)bat_events (work_completion)(&(&forw_packet_aggr->delayed_work)->work) &hard_iface->bat_iv.ogm_buff_mutex rcu_read_lock &n->list_lock irq_context: 0 (wq_completion)ipv6_addrconf (work_completion)(&(&ifa->dad_work)->work) rtnl_mutex rcu_read_lock rcu_read_lock rcu_read_lock_bh rcu_read_lock_bh rcu_read_lock &pool->lock &p->pi_lock &rq->__lock irq_context: 0 (wq_completion)bat_events (work_completion)(&(&forw_packet_aggr->delayed_work)->work) &hard_iface->bat_iv.ogm_buff_mutex rcu_read_lock &n->list_lock &c->lock irq_context: 0 (wq_completion)ipv6_addrconf (work_completion)(&(&ifa->dad_work)->work) rtnl_mutex rcu_read_lock rcu_read_lock rcu_read_lock_bh rcu_read_lock_bh rcu_read_lock &pool->lock &p->pi_lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)wg-crypt-wg0#3 (work_completion)(&peer->transmit_packet_work) rcu_read_lock_bh &base->lock &obj_hash[i].lock irq_context: 0 (wq_completion)wg-crypt-wg0#3 (work_completion)(&peer->transmit_packet_work) batched_entropy_u8.lock irq_context: 0 (wq_completion)wg-crypt-wg1#3 (work_completion)(&({ do { const void *__vpp_verify = (typeof((worker) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __asm__ ("" : "=r"(__ptr) : "0"((typeof(*((worker))) *)((worker)))); (typeof((typeof(*((worker))) *)((worker)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); })->work) &rq->__lock irq_context: 0 (wq_completion)wg-crypt-wg1#3 (work_completion)(&peer->transmit_packet_work) batched_entropy_u8.lock irq_context: 0 (wq_completion)wg-crypt-wg1#3 (work_completion)(&peer->transmit_packet_work) rcu_read_lock_bh &base->lock &obj_hash[i].lock irq_context: 0 &pool->lock &p->pi_lock &rq->__lock &cfs_rq->removed.lock irq_context: 0 (wq_completion)ipv6_addrconf (work_completion)(&(&ifa->dad_work)->work) rtnl_mutex &tb->tb6_lock quarantine_lock irq_context: 0 (wq_completion)ipv6_addrconf (work_completion)(&(&ifa->dad_work)->work) rtnl_mutex &idev->mc_lock remove_cache_srcu rcu_read_lock rcu_node_0 irq_context: 0 (wq_completion)ipv6_addrconf (work_completion)(&(&ifa->dad_work)->work) rtnl_mutex &idev->mc_lock remove_cache_srcu rcu_read_lock &rq->__lock irq_context: 0 (wq_completion)wg-crypt-wg2#4 (work_completion)(&peer->transmit_packet_work) rcu_read_lock_bh tk_core.seq.seqcount irq_context: 0 (wq_completion)wg-kex-wg2#8 (work_completion)(&({ do { const void *__vpp_verify = (typeof((worker) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __asm__ ("" : "=r"(__ptr) : "0"((typeof(*((worker))) *)((worker)))); (typeof((typeof(*((worker))) *)((worker)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); })->work) &wg->static_identity.lock &sem->wait_lock irq_context: 0 (wq_completion)wg-kex-wg2#8 (work_completion)(&({ do { const void *__vpp_verify = (typeof((worker) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __asm__ ("" : "=r"(__ptr) : "0"((typeof(*((worker))) *)((worker)))); (typeof((typeof(*((worker))) *)((worker)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); })->work) &wg->static_identity.lock &p->pi_lock irq_context: 0 (wq_completion)wg-kex-wg2#8 (work_completion)(&({ do { const void *__vpp_verify = (typeof((worker) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __asm__ ("" : "=r"(__ptr) : "0"((typeof(*((worker))) *)((worker)))); (typeof((typeof(*((worker))) *)((worker)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); })->work) &wg->static_identity.lock &handshake->lock crngs.lock irq_context: 0 (wq_completion)wg-kex-wg2#8 (work_completion)(&({ do { const void *__vpp_verify = (typeof((worker) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __asm__ ("" : "=r"(__ptr) : "0"((typeof(*((worker))) *)((worker)))); (typeof((typeof(*((worker))) *)((worker)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); })->work) &wg->static_identity.lock &handshake->lock &table->lock#2 irq_context: 0 (wq_completion)wg-kex-wg2#8 (work_completion)(&({ do { const void *__vpp_verify = (typeof((worker) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __asm__ ("" : "=r"(__ptr) : "0"((typeof(*((worker))) *)((worker)))); (typeof((typeof(*((worker))) *)((worker)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); })->work) &wg->static_identity.lock &handshake->lock rcu_read_lock_bh batched_entropy_u32.lock irq_context: 0 (wq_completion)wg-kex-wg2#8 (work_completion)(&({ do { const void *__vpp_verify = (typeof((worker) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __asm__ ("" : "=r"(__ptr) : "0"((typeof(*((worker))) *)((worker)))); (typeof((typeof(*((worker))) *)((worker)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); })->work) &wg->static_identity.lock &handshake->lock rcu_read_lock_bh &table->lock#2 irq_context: 0 (wq_completion)wg-kex-wg2#8 (work_completion)(&({ do { const void *__vpp_verify = (typeof((worker) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __asm__ ("" : "=r"(__ptr) : "0"((typeof(*((worker))) *)((worker)))); (typeof((typeof(*((worker))) *)((worker)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); })->work) &cookie->lock irq_context: 0 (wq_completion)wg-kex-wg2#8 (work_completion)(&({ do { const void *__vpp_verify = (typeof((worker) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __asm__ ("" : "=r"(__ptr) : "0"((typeof(*((worker))) *)((worker)))); (typeof((typeof(*((worker))) *)((worker)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); })->work) &cookie->lock irq_context: 0 (wq_completion)wg-kex-wg2#7 (work_completion)(&peer->transmit_handshake_work) &peer->endpoint_lock rcu_read_lock_bh rcu_read_lock &dir->lock#2 irq_context: 0 (wq_completion)wg-kex-wg2#7 (work_completion)(&peer->transmit_handshake_work) &peer->endpoint_lock rcu_read_lock_bh rcu_read_lock &ul->lock irq_context: 0 (wq_completion)wg-kex-wg2#7 (work_completion)(&peer->transmit_handshake_work) &peer->endpoint_lock rcu_read_lock_bh batched_entropy_u32.lock irq_context: 0 (wq_completion)ipv6_addrconf (work_completion)(&(&ifa->dad_work)->work) rtnl_mutex &idev->mc_lock remove_cache_srcu rcu_read_lock &rcu_state.gp_wq irq_context: 0 (wq_completion)ipv6_addrconf (work_completion)(&(&ifa->dad_work)->work) rtnl_mutex &idev->mc_lock remove_cache_srcu rcu_read_lock &rcu_state.gp_wq &p->pi_lock irq_context: 0 cb_lock genl_mutex rfkill_global_mutex &root->kernfs_rwsem &root->kernfs_iattr_rwsem &rq->__lock irq_context: 0 (wq_completion)wg-kex-wg1#12 (work_completion)(&({ do { const void *__vpp_verify = (typeof((worker) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __asm__ ("" : "=r"(__ptr) : "0"((typeof(*((worker))) *)((worker)))); (typeof((typeof(*((worker))) *)((worker)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); })->work) &wg->static_identity.lock &handshake->lock tk_core.seq.seqcount irq_context: 0 (wq_completion)wg-kex-wg1#12 (work_completion)(&({ do { const void *__vpp_verify = (typeof((worker) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __asm__ ("" : "=r"(__ptr) : "0"((typeof(*((worker))) *)((worker)))); (typeof((typeof(*((worker))) *)((worker)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); })->work) &wg->static_identity.lock &handshake->lock tk_core.seq.seqcount irq_context: 0 (wq_completion)wg-kex-wg1#12 (work_completion)(&({ do { const void *__vpp_verify = (typeof((worker) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __asm__ ("" : "=r"(__ptr) : "0"((typeof(*((worker))) *)((worker)))); (typeof((typeof(*((worker))) *)((worker)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); })->work) &wg->static_identity.lock &handshake->lock crngs.lock irq_context: 0 (wq_completion)wg-kex-wg1#12 (work_completion)(&({ do { const void *__vpp_verify = (typeof((worker) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __asm__ ("" : "=r"(__ptr) : "0"((typeof(*((worker))) *)((worker)))); (typeof((typeof(*((worker))) *)((worker)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); })->work) &wg->static_identity.lock &handshake->lock &table->lock#2 irq_context: 0 (wq_completion)wg-kex-wg1#12 (work_completion)(&({ do { const void *__vpp_verify = (typeof((worker) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __asm__ ("" : "=r"(__ptr) : "0"((typeof(*((worker))) *)((worker)))); (typeof((typeof(*((worker))) *)((worker)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); })->work) &wg->static_identity.lock &handshake->lock rcu_read_lock_bh batched_entropy_u32.lock irq_context: 0 (wq_completion)wg-kex-wg1#12 (work_completion)(&({ do { const void *__vpp_verify = (typeof((worker) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __asm__ ("" : "=r"(__ptr) : "0"((typeof(*((worker))) *)((worker)))); (typeof((typeof(*((worker))) *)((worker)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); })->work) &wg->static_identity.lock &handshake->lock rcu_read_lock_bh &table->lock#2 irq_context: 0 (wq_completion)wg-kex-wg1#12 (work_completion)(&({ do { const void *__vpp_verify = (typeof((worker) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __asm__ ("" : "=r"(__ptr) : "0"((typeof(*((worker))) *)((worker)))); (typeof((typeof(*((worker))) *)((worker)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); })->work) &cookie->lock irq_context: 0 (wq_completion)wg-kex-wg1#12 (work_completion)(&({ do { const void *__vpp_verify = (typeof((worker) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __asm__ ("" : "=r"(__ptr) : "0"((typeof(*((worker))) *)((worker)))); (typeof((typeof(*((worker))) *)((worker)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); })->work) &cookie->lock irq_context: 0 (wq_completion)wg-kex-wg1#12 (work_completion)(&({ do { const void *__vpp_verify = (typeof((worker) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __asm__ ("" : "=r"(__ptr) : "0"((typeof(*((worker))) *)((worker)))); (typeof((typeof(*((worker))) *)((worker)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); })->work) &peer->endpoint_lock irq_context: 0 (wq_completion)wg-kex-wg1#12 (work_completion)(&({ do { const void *__vpp_verify = (typeof((worker) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __asm__ ("" : "=r"(__ptr) : "0"((typeof(*((worker))) *)((worker)))); (typeof((typeof(*((worker))) *)((worker)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); })->work) &peer->endpoint_lock rcu_read_lock_bh rcu_read_lock &dir->lock#2 irq_context: 0 (wq_completion)wg-kex-wg1#12 (work_completion)(&({ do { const void *__vpp_verify = (typeof((worker) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __asm__ ("" : "=r"(__ptr) : "0"((typeof(*((worker))) *)((worker)))); (typeof((typeof(*((worker))) *)((worker)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); })->work) &peer->endpoint_lock rcu_read_lock_bh rcu_read_lock &____s->seqcount#7 irq_context: 0 (wq_completion)wg-kex-wg1#12 (work_completion)(&({ do { const void *__vpp_verify = (typeof((worker) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __asm__ ("" : "=r"(__ptr) : "0"((typeof(*((worker))) *)((worker)))); (typeof((typeof(*((worker))) *)((worker)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); })->work) &peer->endpoint_lock rcu_read_lock_bh rcu_read_lock &____s->seqcount#10 irq_context: 0 (wq_completion)wg-kex-wg1#12 (work_completion)(&({ do { const void *__vpp_verify = (typeof((worker) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __asm__ ("" : "=r"(__ptr) : "0"((typeof(*((worker))) *)((worker)))); (typeof((typeof(*((worker))) *)((worker)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); })->work) &peer->endpoint_lock rcu_read_lock_bh rcu_read_lock rcu_read_lock_bh tk_core.seq.seqcount irq_context: 0 (wq_completion)wg-kex-wg1#12 (work_completion)(&({ do { const void *__vpp_verify = (typeof((worker) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __asm__ ("" : "=r"(__ptr) : "0"((typeof(*((worker))) *)((worker)))); (typeof((typeof(*((worker))) *)((worker)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); })->work) &peer->endpoint_lock rcu_read_lock_bh rcu_read_lock rcu_read_lock_bh &list->lock#5 irq_context: 0 (wq_completion)ipv6_addrconf (work_completion)(&(&ifa->dad_work)->work) rtnl_mutex &idev->mc_lock rcu_read_lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 cb_lock genl_mutex &sb->s_type->i_mutex_key#3 batched_entropy_u8.lock irq_context: 0 cb_lock genl_mutex &sb->s_type->i_mutex_key#3 kfence_freelist_lock irq_context: 0 (wq_completion)events_unbound (work_completion)(&port->bc_work) quarantine_lock irq_context: 0 (wq_completion)wg-kex-wg2#8 (work_completion)(&({ do { const void *__vpp_verify = (typeof((worker) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __asm__ ("" : "=r"(__ptr) : "0"((typeof(*((worker))) *)((worker)))); (typeof((typeof(*((worker))) *)((worker)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); })->work) &base->lock irq_context: 0 (wq_completion)wg-kex-wg2#8 (work_completion)(&({ do { const void *__vpp_verify = (typeof((worker) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __asm__ ("" : "=r"(__ptr) : "0"((typeof(*((worker))) *)((worker)))); (typeof((typeof(*((worker))) *)((worker)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); })->work) &base->lock &obj_hash[i].lock irq_context: 0 (wq_completion)rcu_gp (work_completion)(&rew->rew_work) rcu_state.exp_wake_mutex &rnp->exp_wq[2] &p->pi_lock &rq->__lock &cfs_rq->removed.lock irq_context: 0 (wq_completion)wg-crypt-wg2#4 (work_completion)(&peer->transmit_packet_work) &peer->endpoint_lock rcu_read_lock_bh batched_entropy_u32.lock irq_context: 0 (wq_completion)wg-kex-wg2#9 irq_context: 0 (wq_completion)wg-kex-wg2#9 (work_completion)(&peer->transmit_handshake_work) irq_context: 0 (wq_completion)wg-kex-wg2#9 (work_completion)(&peer->transmit_handshake_work) tk_core.seq.seqcount irq_context: 0 (wq_completion)wg-kex-wg2#9 (work_completion)(&peer->transmit_handshake_work) &wg->static_identity.lock irq_context: 0 (wq_completion)wg-kex-wg2#9 (work_completion)(&peer->transmit_handshake_work) &wg->static_identity.lock &handshake->lock irq_context: 0 (wq_completion)wg-kex-wg2#9 (work_completion)(&peer->transmit_handshake_work) &wg->static_identity.lock &handshake->lock crngs.lock irq_context: 0 (wq_completion)wg-kex-wg2#9 (work_completion)(&peer->transmit_handshake_work) &wg->static_identity.lock &handshake->lock tk_core.seq.seqcount irq_context: 0 (wq_completion)wg-kex-wg2#9 (work_completion)(&peer->transmit_handshake_work) &wg->static_identity.lock &handshake->lock &table->lock#2 irq_context: 0 (wq_completion)wg-kex-wg2#9 (work_completion)(&peer->transmit_handshake_work) &wg->static_identity.lock &handshake->lock rcu_read_lock_bh batched_entropy_u32.lock irq_context: 0 (wq_completion)wg-kex-wg2#9 (work_completion)(&peer->transmit_handshake_work) &wg->static_identity.lock &handshake->lock rcu_read_lock_bh &table->lock#2 irq_context: 0 (wq_completion)wg-kex-wg2#9 (work_completion)(&peer->transmit_handshake_work) &cookie->lock irq_context: 0 (wq_completion)wg-kex-wg2#9 (work_completion)(&peer->transmit_handshake_work) &cookie->lock irq_context: 0 (wq_completion)wg-kex-wg2#9 (work_completion)(&peer->transmit_handshake_work) rcu_read_lock_bh &obj_hash[i].lock irq_context: 0 (wq_completion)wg-kex-wg2#9 (work_completion)(&peer->transmit_handshake_work) rcu_read_lock_bh &base->lock irq_context: 0 (wq_completion)wg-kex-wg2#9 (work_completion)(&peer->transmit_handshake_work) rcu_read_lock_bh &base->lock &obj_hash[i].lock irq_context: 0 (wq_completion)wg-kex-wg2#9 (work_completion)(&peer->transmit_handshake_work) &obj_hash[i].lock irq_context: 0 (wq_completion)wg-kex-wg2#9 (work_completion)(&peer->transmit_handshake_work) &peer->endpoint_lock irq_context: 0 (wq_completion)wg-kex-wg2#9 (work_completion)(&peer->transmit_handshake_work) &peer->endpoint_lock rcu_read_lock_bh rcu_read_lock &____s->seqcount#7 irq_context: 0 (wq_completion)wg-kex-wg2#9 (work_completion)(&peer->transmit_handshake_work) &peer->endpoint_lock rcu_read_lock_bh rcu_read_lock &____s->seqcount#10 irq_context: 0 (wq_completion)wg-kex-wg2#9 (work_completion)(&peer->transmit_handshake_work) &peer->endpoint_lock rcu_read_lock_bh rcu_read_lock rcu_read_lock_bh tk_core.seq.seqcount irq_context: 0 (wq_completion)wg-kex-wg2#9 (work_completion)(&peer->transmit_handshake_work) &peer->endpoint_lock rcu_read_lock_bh rcu_read_lock rcu_read_lock_bh &list->lock#5 irq_context: 0 (wq_completion)wg-kex-wg2#9 (work_completion)(&peer->transmit_handshake_work) batched_entropy_u8.lock irq_context: 0 (wq_completion)wg-kex-wg2#9 (work_completion)(&peer->transmit_handshake_work) &peer->endpoint_lock rcu_read_lock_bh rcu_read_lock &dir->lock#2 irq_context: 0 (wq_completion)wg-kex-wg2#9 (work_completion)(&peer->transmit_handshake_work) &peer->endpoint_lock rcu_read_lock_bh batched_entropy_u32.lock irq_context: 0 (wq_completion)events fqdir_free_work &obj_hash[i].lock irq_context: 0 (wq_completion)events fqdir_free_work pool_lock#2 irq_context: 0 pernet_ops_rwsem rtnl_mutex &caifn->caifdevs.lock irq_context: 0 pernet_ops_rwsem rtnl_mutex &caifn->caifdevs.lock rcu_state.exp_mutex rcu_node_0 irq_context: 0 pernet_ops_rwsem rtnl_mutex &caifn->caifdevs.lock rcu_state.exp_mutex &obj_hash[i].lock irq_context: 0 pernet_ops_rwsem rtnl_mutex &caifn->caifdevs.lock rcu_state.exp_mutex rcu_read_lock &pool->lock irq_context: 0 pernet_ops_rwsem rtnl_mutex &caifn->caifdevs.lock rcu_state.exp_mutex rcu_read_lock &pool->lock &obj_hash[i].lock irq_context: 0 pernet_ops_rwsem rtnl_mutex &caifn->caifdevs.lock rcu_state.exp_mutex rcu_read_lock &pool->lock &p->pi_lock irq_context: 0 pernet_ops_rwsem rtnl_mutex &caifn->caifdevs.lock rcu_state.exp_mutex rcu_read_lock &pool->lock &p->pi_lock &rq->__lock irq_context: 0 pernet_ops_rwsem rtnl_mutex &caifn->caifdevs.lock rcu_state.exp_mutex rcu_read_lock &pool->lock &p->pi_lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 pernet_ops_rwsem rtnl_mutex &caifn->caifdevs.lock rcu_state.exp_mutex &rnp->exp_wq[3] irq_context: 0 pernet_ops_rwsem rtnl_mutex &caifn->caifdevs.lock rcu_state.exp_mutex &rq->__lock irq_context: 0 pernet_ops_rwsem rtnl_mutex &caifn->caifdevs.lock rcu_state.exp_mutex &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 pernet_ops_rwsem rtnl_mutex &caifn->caifdevs.lock &obj_hash[i].lock irq_context: 0 pernet_ops_rwsem rtnl_mutex &caifn->caifdevs.lock pool_lock#2 irq_context: 0 pernet_ops_rwsem rtnl_mutex &caifn->caifdevs.lock &this->info_list_lock irq_context: 0 pernet_ops_rwsem &pnettable->lock irq_context: 0 (wq_completion)wg-crypt-wg2#4 (work_completion)(&peer->transmit_packet_work) batched_entropy_u8.lock irq_context: 0 (wq_completion)wg-crypt-wg2#4 (work_completion)(&peer->transmit_packet_work) rcu_read_lock_bh &base->lock irq_context: 0 (wq_completion)wg-crypt-wg2#4 (work_completion)(&peer->transmit_packet_work) rcu_read_lock_bh &base->lock &obj_hash[i].lock irq_context: 0 (wq_completion)wg-crypt-wg1#5 (work_completion)(&peer->transmit_packet_work) rcu_read_lock_bh &base->lock irq_context: 0 (wq_completion)wg-crypt-wg1#5 (work_completion)(&peer->transmit_packet_work) rcu_read_lock_bh &base->lock &obj_hash[i].lock irq_context: 0 (wq_completion)wg-kex-wg0#10 (work_completion)(&({ do { const void *__vpp_verify = (typeof((worker) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __asm__ ("" : "=r"(__ptr) : "0"((typeof(*((worker))) *)((worker)))); (typeof((typeof(*((worker))) *)((worker)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); })->work) &handshake->lock &c->lock irq_context: 0 (wq_completion)wg-crypt-wg1#5 (work_completion)(&peer->transmit_packet_work) &peer->endpoint_lock rcu_read_lock_bh &obj_hash[i].lock irq_context: 0 (wq_completion)wg-kex-wg0#10 (work_completion)(&({ do { const void *__vpp_verify = (typeof((worker) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __asm__ ("" : "=r"(__ptr) : "0"((typeof(*((worker))) *)((worker)))); (typeof((typeof(*((worker))) *)((worker)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); })->work) &handshake->lock &____s->seqcount#2 irq_context: 0 (wq_completion)wg-kex-wg0#10 (work_completion)(&({ do { const void *__vpp_verify = (typeof((worker) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __asm__ ("" : "=r"(__ptr) : "0"((typeof(*((worker))) *)((worker)))); (typeof((typeof(*((worker))) *)((worker)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); })->work) &handshake->lock &____s->seqcount irq_context: 0 (wq_completion)wg-kex-wg0#10 (work_completion)(&({ do { const void *__vpp_verify = (typeof((worker) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __asm__ ("" : "=r"(__ptr) : "0"((typeof(*((worker))) *)((worker)))); (typeof((typeof(*((worker))) *)((worker)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); })->work) &handshake->lock &rq->__lock irq_context: 0 (wq_completion)wg-crypt-wg1#5 (work_completion)(&peer->transmit_packet_work) batched_entropy_u8.lock irq_context: 0 (wq_completion)wg-kex-wg2#10 irq_context: 0 (wq_completion)wg-kex-wg0#10 (work_completion)(&({ do { const void *__vpp_verify = (typeof((worker) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __asm__ ("" : "=r"(__ptr) : "0"((typeof(*((worker))) *)((worker)))); (typeof((typeof(*((worker))) *)((worker)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); })->work) &handshake->lock pool_lock#2 irq_context: 0 (wq_completion)wg-kex-wg2#10 (work_completion)(&({ do { const void *__vpp_verify = (typeof((worker) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __asm__ ("" : "=r"(__ptr) : "0"((typeof(*((worker))) *)((worker)))); (typeof((typeof(*((worker))) *)((worker)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); })->work) irq_context: 0 (wq_completion)wg-kex-wg2#10 (work_completion)(&({ do { const void *__vpp_verify = (typeof((worker) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __asm__ ("" : "=r"(__ptr) : "0"((typeof(*((worker))) *)((worker)))); (typeof((typeof(*((worker))) *)((worker)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); })->work) &r->consumer_lock#2 irq_context: 0 (wq_completion)wg-kex-wg0#10 (work_completion)(&({ do { const void *__vpp_verify = (typeof((worker) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __asm__ ("" : "=r"(__ptr) : "0"((typeof(*((worker))) *)((worker)))); (typeof((typeof(*((worker))) *)((worker)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); })->work) &c->lock irq_context: 0 (wq_completion)wg-kex-wg2#10 (work_completion)(&({ do { const void *__vpp_verify = (typeof((worker) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __asm__ ("" : "=r"(__ptr) : "0"((typeof(*((worker))) *)((worker)))); (typeof((typeof(*((worker))) *)((worker)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); })->work) &wg->static_identity.lock irq_context: 0 (wq_completion)wg-kex-wg2#10 (work_completion)(&({ do { const void *__vpp_verify = (typeof((worker) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __asm__ ("" : "=r"(__ptr) : "0"((typeof(*((worker))) *)((worker)))); (typeof((typeof(*((worker))) *)((worker)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); })->work) &wg->static_identity.lock &handshake->lock irq_context: 0 (wq_completion)wg-kex-wg2#10 (work_completion)(&({ do { const void *__vpp_verify = (typeof((worker) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __asm__ ("" : "=r"(__ptr) : "0"((typeof(*((worker))) *)((worker)))); (typeof((typeof(*((worker))) *)((worker)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); })->work) &wg->static_identity.lock &handshake->lock irq_context: 0 (wq_completion)wg-kex-wg2#10 (work_completion)(&({ do { const void *__vpp_verify = (typeof((worker) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __asm__ ("" : "=r"(__ptr) : "0"((typeof(*((worker))) *)((worker)))); (typeof((typeof(*((worker))) *)((worker)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); })->work) &peer->endpoint_lock irq_context: 0 (wq_completion)wg-kex-wg2#10 (work_completion)(&({ do { const void *__vpp_verify = (typeof((worker) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __asm__ ("" : "=r"(__ptr) : "0"((typeof(*((worker))) *)((worker)))); (typeof((typeof(*((worker))) *)((worker)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); })->work) &handshake->lock irq_context: 0 (wq_completion)wg-kex-wg2#10 (work_completion)(&({ do { const void *__vpp_verify = (typeof((worker) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __asm__ ("" : "=r"(__ptr) : "0"((typeof(*((worker))) *)((worker)))); (typeof((typeof(*((worker))) *)((worker)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); })->work) &handshake->lock fs_reclaim irq_context: 0 (wq_completion)wg-kex-wg2#10 (work_completion)(&({ do { const void *__vpp_verify = (typeof((worker) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __asm__ ("" : "=r"(__ptr) : "0"((typeof(*((worker))) *)((worker)))); (typeof((typeof(*((worker))) *)((worker)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); })->work) &handshake->lock fs_reclaim mmu_notifier_invalidate_range_start irq_context: 0 (wq_completion)wg-kex-wg2#10 (work_completion)(&({ do { const void *__vpp_verify = (typeof((worker) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __asm__ ("" : "=r"(__ptr) : "0"((typeof(*((worker))) *)((worker)))); (typeof((typeof(*((worker))) *)((worker)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); })->work) &handshake->lock tk_core.seq.seqcount irq_context: 0 (wq_completion)wg-kex-wg2#10 (work_completion)(&({ do { const void *__vpp_verify = (typeof((worker) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __asm__ ("" : "=r"(__ptr) : "0"((typeof(*((worker))) *)((worker)))); (typeof((typeof(*((worker))) *)((worker)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); })->work) &handshake->lock rcu_read_lock_bh &peer->keypairs.keypair_update_lock irq_context: 0 (wq_completion)wg-kex-wg2#10 (work_completion)(&({ do { const void *__vpp_verify = (typeof((worker) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __asm__ ("" : "=r"(__ptr) : "0"((typeof(*((worker))) *)((worker)))); (typeof((typeof(*((worker))) *)((worker)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); })->work) &handshake->lock rcu_read_lock_bh &table->lock#2 irq_context: 0 (wq_completion)wg-kex-wg2#10 (work_completion)(&({ do { const void *__vpp_verify = (typeof((worker) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __asm__ ("" : "=r"(__ptr) : "0"((typeof(*((worker))) *)((worker)))); (typeof((typeof(*((worker))) *)((worker)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); })->work) rcu_read_lock_bh &obj_hash[i].lock irq_context: 0 (wq_completion)wg-kex-wg2#10 (work_completion)(&({ do { const void *__vpp_verify = (typeof((worker) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __asm__ ("" : "=r"(__ptr) : "0"((typeof(*((worker))) *)((worker)))); (typeof((typeof(*((worker))) *)((worker)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); })->work) rcu_read_lock_bh &base->lock irq_context: 0 (wq_completion)wg-kex-wg2#10 (work_completion)(&({ do { const void *__vpp_verify = (typeof((worker) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __asm__ ("" : "=r"(__ptr) : "0"((typeof(*((worker))) *)((worker)))); (typeof((typeof(*((worker))) *)((worker)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); })->work) rcu_read_lock_bh &base->lock &obj_hash[i].lock irq_context: 0 (wq_completion)wg-kex-wg2#10 (work_completion)(&({ do { const void *__vpp_verify = (typeof((worker) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __asm__ ("" : "=r"(__ptr) : "0"((typeof(*((worker))) *)((worker)))); (typeof((typeof(*((worker))) *)((worker)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); })->work) &obj_hash[i].lock irq_context: 0 (wq_completion)wg-kex-wg2#10 (work_completion)(&({ do { const void *__vpp_verify = (typeof((worker) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __asm__ ("" : "=r"(__ptr) : "0"((typeof(*((worker))) *)((worker)))); (typeof((typeof(*((worker))) *)((worker)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); })->work) &base->lock irq_context: 0 (wq_completion)wg-kex-wg2#10 (work_completion)(&({ do { const void *__vpp_verify = (typeof((worker) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __asm__ ("" : "=r"(__ptr) : "0"((typeof(*((worker))) *)((worker)))); (typeof((typeof(*((worker))) *)((worker)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); })->work) &base->lock &obj_hash[i].lock irq_context: 0 (wq_completion)wg-kex-wg2#10 (work_completion)(&({ do { const void *__vpp_verify = (typeof((worker) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __asm__ ("" : "=r"(__ptr) : "0"((typeof(*((worker))) *)((worker)))); (typeof((typeof(*((worker))) *)((worker)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); })->work) tk_core.seq.seqcount irq_context: 0 (wq_completion)wg-kex-wg2#10 (work_completion)(&({ do { const void *__vpp_verify = (typeof((worker) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __asm__ ("" : "=r"(__ptr) : "0"((typeof(*((worker))) *)((worker)))); (typeof((typeof(*((worker))) *)((worker)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); })->work) &list->lock#17 irq_context: 0 (wq_completion)wg-kex-wg2#10 (work_completion)(&({ do { const void *__vpp_verify = (typeof((worker) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __asm__ ("" : "=r"(__ptr) : "0"((typeof(*((worker))) *)((worker)))); (typeof((typeof(*((worker))) *)((worker)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); })->work) rcu_read_lock_bh &r->producer_lock#2 irq_context: 0 (wq_completion)wg-kex-wg2#10 (work_completion)(&({ do { const void *__vpp_verify = (typeof((worker) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __asm__ ("" : "=r"(__ptr) : "0"((typeof(*((worker))) *)((worker)))); (typeof((typeof(*((worker))) *)((worker)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); })->work) rcu_read_lock_bh rcu_read_lock &pool->lock irq_context: 0 (wq_completion)wg-kex-wg2#10 (work_completion)(&({ do { const void *__vpp_verify = (typeof((worker) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __asm__ ("" : "=r"(__ptr) : "0"((typeof(*((worker))) *)((worker)))); (typeof((typeof(*((worker))) *)((worker)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); })->work) rcu_read_lock_bh rcu_read_lock &pool->lock &obj_hash[i].lock irq_context: 0 (wq_completion)wg-kex-wg2#10 (work_completion)(&({ do { const void *__vpp_verify = (typeof((worker) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __asm__ ("" : "=r"(__ptr) : "0"((typeof(*((worker))) *)((worker)))); (typeof((typeof(*((worker))) *)((worker)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); })->work) rcu_read_lock_bh rcu_read_lock &pool->lock &p->pi_lock irq_context: 0 (wq_completion)wg-crypt-wg1#5 (work_completion)(&({ do { const void *__vpp_verify = (typeof((worker) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __asm__ ("" : "=r"(__ptr) : "0"((typeof(*((worker))) *)((worker)))); (typeof((typeof(*((worker))) *)((worker)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); })->work) &rq->__lock irq_context: 0 (wq_completion)wg-crypt-wg1#5 (work_completion)(&({ do { const void *__vpp_verify = (typeof((worker) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __asm__ ("" : "=r"(__ptr) : "0"((typeof(*((worker))) *)((worker)))); (typeof((typeof(*((worker))) *)((worker)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); })->work) &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)wg-crypt-wg2#5 irq_context: 0 (wq_completion)wg-crypt-wg2#5 (work_completion)(&({ do { const void *__vpp_verify = (typeof((worker) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __asm__ ("" : "=r"(__ptr) : "0"((typeof(*((worker))) *)((worker)))); (typeof((typeof(*((worker))) *)((worker)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); })->work) irq_context: 0 (wq_completion)wg-crypt-wg2#5 (work_completion)(&({ do { const void *__vpp_verify = (typeof((worker) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __asm__ ("" : "=r"(__ptr) : "0"((typeof(*((worker))) *)((worker)))); (typeof((typeof(*((worker))) *)((worker)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); })->work) &r->consumer_lock#2 irq_context: 0 (wq_completion)wg-crypt-wg2#5 (work_completion)(&({ do { const void *__vpp_verify = (typeof((worker) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __asm__ ("" : "=r"(__ptr) : "0"((typeof(*((worker))) *)((worker)))); (typeof((typeof(*((worker))) *)((worker)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); })->work) rcu_read_lock &pool->lock irq_context: 0 (wq_completion)wg-crypt-wg2#5 (work_completion)(&({ do { const void *__vpp_verify = (typeof((worker) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __asm__ ("" : "=r"(__ptr) : "0"((typeof(*((worker))) *)((worker)))); (typeof((typeof(*((worker))) *)((worker)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); })->work) rcu_read_lock &pool->lock &obj_hash[i].lock irq_context: 0 (wq_completion)wg-crypt-wg2#5 (work_completion)(&({ do { const void *__vpp_verify = (typeof((worker) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __asm__ ("" : "=r"(__ptr) : "0"((typeof(*((worker))) *)((worker)))); (typeof((typeof(*((worker))) *)((worker)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); })->work) rcu_read_lock &pool->lock &p->pi_lock irq_context: 0 (wq_completion)wg-crypt-wg2#5 (work_completion)(&({ do { const void *__vpp_verify = (typeof((worker) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __asm__ ("" : "=r"(__ptr) : "0"((typeof(*((worker))) *)((worker)))); (typeof((typeof(*((worker))) *)((worker)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); })->work) rcu_read_lock &pool->lock &p->pi_lock &rq->__lock irq_context: 0 (wq_completion)wg-crypt-wg2#5 (work_completion)(&({ do { const void *__vpp_verify = (typeof((worker) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __asm__ ("" : "=r"(__ptr) : "0"((typeof(*((worker))) *)((worker)))); (typeof((typeof(*((worker))) *)((worker)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); })->work) rcu_read_lock &pool->lock &p->pi_lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)wg-crypt-wg2#5 (work_completion)(&peer->transmit_packet_work) irq_context: 0 (wq_completion)wg-crypt-wg2#5 (work_completion)(&peer->transmit_packet_work) rcu_read_lock_bh &obj_hash[i].lock irq_context: 0 (wq_completion)wg-crypt-wg2#5 (work_completion)(&peer->transmit_packet_work) &obj_hash[i].lock irq_context: 0 (wq_completion)wg-crypt-wg2#5 (work_completion)(&peer->transmit_packet_work) &peer->endpoint_lock irq_context: 0 (wq_completion)wg-crypt-wg2#5 (work_completion)(&peer->transmit_packet_work) &peer->endpoint_lock rcu_read_lock_bh rcu_read_lock &____s->seqcount#7 irq_context: 0 (wq_completion)wg-crypt-wg2#5 (work_completion)(&peer->transmit_packet_work) &peer->endpoint_lock rcu_read_lock_bh rcu_read_lock &____s->seqcount#10 irq_context: 0 (wq_completion)wg-crypt-wg2#5 (work_completion)(&peer->transmit_packet_work) &peer->endpoint_lock rcu_read_lock_bh rcu_read_lock rcu_read_lock_bh tk_core.seq.seqcount irq_context: 0 (wq_completion)wg-crypt-wg2#5 (work_completion)(&peer->transmit_packet_work) &peer->endpoint_lock rcu_read_lock_bh rcu_read_lock rcu_read_lock_bh &list->lock#5 irq_context: 0 (wq_completion)wg-crypt-wg2#5 (work_completion)(&peer->transmit_packet_work) rcu_read_lock_bh tk_core.seq.seqcount irq_context: 0 &type->i_mutex_dir_key#7 &root->kernfs_rwsem &c->lock irq_context: 0 &type->i_mutex_dir_key#7 &root->kernfs_rwsem &____s->seqcount#2 irq_context: 0 &type->i_mutex_dir_key#7 &root->kernfs_rwsem &____s->seqcount irq_context: 0 pernet_ops_rwsem &pnetids_ndev->lock irq_context: 0 sb_writers#11 &type->i_mutex_dir_key#7 &root->kernfs_rwsem &____s->seqcount#2 irq_context: 0 pernet_ops_rwsem k-sk_lock-AF_INET6/1 irq_context: 0 sb_writers#11 &type->i_mutex_dir_key#7 &root->kernfs_rwsem &____s->seqcount irq_context: 0 pernet_ops_rwsem k-sk_lock-AF_INET6/1 k-slock-AF_INET6 irq_context: 0 kn->active#55 fs_reclaim &rq->__lock irq_context: 0 kn->active#55 fs_reclaim &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 pernet_ops_rwsem k-sk_lock-AF_INET6/1 rlock-AF_INET6 irq_context: 0 (wq_completion)wg-crypt-wg2#5 (work_completion)(&({ do { const void *__vpp_verify = (typeof((worker) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __asm__ ("" : "=r"(__ptr) : "0"((typeof(*((worker))) *)((worker)))); (typeof((typeof(*((worker))) *)((worker)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); })->work) tk_core.seq.seqcount irq_context: 0 pernet_ops_rwsem k-sk_lock-AF_INET6/1 &list->lock#21 irq_context: 0 pernet_ops_rwsem &net->sctp.addr_wq_lock irq_context: 0 (wq_completion)wg-crypt-wg0#4 (work_completion)(&peer->transmit_packet_work) batched_entropy_u8.lock irq_context: 0 (wq_completion)wg-crypt-wg0#4 (work_completion)(&peer->transmit_packet_work) rcu_read_lock_bh &base->lock &obj_hash[i].lock irq_context: 0 sb_writers#11 &of->mutex cgroup_mutex cpu_hotplug_lock cgroup_threadgroup_rwsem rcu_state.exp_mutex &rnp->exp_wq[0] irq_context: 0 (wq_completion)wg-kex-wg2#11 irq_context: 0 (wq_completion)wg-kex-wg2#11 (work_completion)(&peer->transmit_handshake_work) irq_context: 0 (wq_completion)wg-kex-wg2#11 (work_completion)(&peer->transmit_handshake_work) tk_core.seq.seqcount irq_context: 0 (wq_completion)wg-kex-wg2#11 (work_completion)(&peer->transmit_handshake_work) &wg->static_identity.lock irq_context: 0 (wq_completion)wg-kex-wg2#11 (work_completion)(&peer->transmit_handshake_work) &wg->static_identity.lock &handshake->lock irq_context: 0 (wq_completion)wg-kex-wg2#11 (work_completion)(&peer->transmit_handshake_work) &wg->static_identity.lock &handshake->lock crngs.lock irq_context: 0 (wq_completion)wg-kex-wg2#11 (work_completion)(&peer->transmit_handshake_work) &wg->static_identity.lock &handshake->lock tk_core.seq.seqcount irq_context: 0 (wq_completion)wg-kex-wg2#11 (work_completion)(&peer->transmit_handshake_work) &wg->static_identity.lock &handshake->lock &table->lock#2 irq_context: 0 (wq_completion)wg-kex-wg2#11 (work_completion)(&peer->transmit_handshake_work) &wg->static_identity.lock &handshake->lock rcu_read_lock_bh batched_entropy_u32.lock irq_context: 0 (wq_completion)wg-kex-wg2#11 (work_completion)(&peer->transmit_handshake_work) &wg->static_identity.lock &handshake->lock rcu_read_lock_bh &table->lock#2 irq_context: 0 (wq_completion)wg-kex-wg2#11 (work_completion)(&peer->transmit_handshake_work) &cookie->lock irq_context: 0 (wq_completion)wg-kex-wg2#11 (work_completion)(&peer->transmit_handshake_work) &cookie->lock irq_context: 0 (wq_completion)wg-kex-wg2#11 (work_completion)(&peer->transmit_handshake_work) rcu_read_lock_bh &obj_hash[i].lock irq_context: 0 (wq_completion)wg-kex-wg2#11 (work_completion)(&peer->transmit_handshake_work) rcu_read_lock_bh &base->lock irq_context: 0 (wq_completion)wg-kex-wg2#11 (work_completion)(&peer->transmit_handshake_work) rcu_read_lock_bh &base->lock &obj_hash[i].lock irq_context: 0 (wq_completion)wg-kex-wg2#11 (work_completion)(&peer->transmit_handshake_work) &obj_hash[i].lock irq_context: 0 (wq_completion)wg-kex-wg2#11 (work_completion)(&peer->transmit_handshake_work) &peer->endpoint_lock irq_context: 0 (wq_completion)wg-kex-wg2#11 (work_completion)(&peer->transmit_handshake_work) &peer->endpoint_lock rcu_read_lock_bh rcu_read_lock &____s->seqcount#7 irq_context: 0 (wq_completion)wg-kex-wg2#11 (work_completion)(&peer->transmit_handshake_work) &peer->endpoint_lock rcu_read_lock_bh rcu_read_lock &____s->seqcount#10 irq_context: 0 (wq_completion)wg-kex-wg2#11 (work_completion)(&peer->transmit_handshake_work) &peer->endpoint_lock rcu_read_lock_bh rcu_read_lock rcu_read_lock_bh tk_core.seq.seqcount irq_context: 0 (wq_completion)wg-kex-wg2#11 (work_completion)(&peer->transmit_handshake_work) &peer->endpoint_lock rcu_read_lock_bh rcu_read_lock rcu_read_lock_bh &list->lock#5 irq_context: 0 (wq_completion)wg-kex-wg2#11 (work_completion)(&peer->transmit_handshake_work) &rq->__lock irq_context: 0 (wq_completion)wg-kex-wg0#11 (work_completion)(&({ do { const void *__vpp_verify = (typeof((worker) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __asm__ ("" : "=r"(__ptr) : "0"((typeof(*((worker))) *)((worker)))); (typeof((typeof(*((worker))) *)((worker)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); })->work) rcu_read_lock_bh fill_pool_map-wait-type-override pool_lock#2 irq_context: 0 (wq_completion)wg-kex-wg0#11 (work_completion)(&({ do { const void *__vpp_verify = (typeof((worker) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __asm__ ("" : "=r"(__ptr) : "0"((typeof(*((worker))) *)((worker)))); (typeof((typeof(*((worker))) *)((worker)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); })->work) rcu_read_lock_bh fill_pool_map-wait-type-override pool_lock irq_context: 0 (wq_completion)wg-kex-wg2#11 (work_completion)(&peer->transmit_handshake_work) batched_entropy_u8.lock irq_context: 0 (wq_completion)wg-kex-wg2#11 (work_completion)(&peer->transmit_handshake_work) &peer->endpoint_lock rcu_read_lock_bh rcu_read_lock &dir->lock#2 irq_context: 0 (wq_completion)wg-kex-wg2#11 (work_completion)(&peer->transmit_handshake_work) &peer->endpoint_lock rcu_read_lock_bh batched_entropy_u32.lock irq_context: softirq rcu_read_lock rcu_read_lock rcu_read_lock &pool->lock fill_pool_map-wait-type-override &c->lock irq_context: softirq rcu_read_lock rcu_read_lock rcu_read_lock &pool->lock fill_pool_map-wait-type-override &____s->seqcount#2 irq_context: softirq rcu_read_lock rcu_read_lock rcu_read_lock &pool->lock fill_pool_map-wait-type-override &____s->seqcount irq_context: 0 (wq_completion)wg-kex-wg1#11 (work_completion)(&({ do { const void *__vpp_verify = (typeof((worker) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __asm__ ("" : "=r"(__ptr) : "0"((typeof(*((worker))) *)((worker)))); (typeof((typeof(*((worker))) *)((worker)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); })->work) &wg->static_identity.lock &handshake->lock tk_core.seq.seqcount irq_context: 0 (wq_completion)wg-kex-wg1#11 (work_completion)(&({ do { const void *__vpp_verify = (typeof((worker) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __asm__ ("" : "=r"(__ptr) : "0"((typeof(*((worker))) *)((worker)))); (typeof((typeof(*((worker))) *)((worker)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); })->work) &wg->static_identity.lock &handshake->lock tk_core.seq.seqcount irq_context: 0 (wq_completion)wg-kex-wg1#11 (work_completion)(&({ do { const void *__vpp_verify = (typeof((worker) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __asm__ ("" : "=r"(__ptr) : "0"((typeof(*((worker))) *)((worker)))); (typeof((typeof(*((worker))) *)((worker)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); })->work) &wg->static_identity.lock &handshake->lock crngs.lock irq_context: 0 (wq_completion)wg-kex-wg1#11 (work_completion)(&({ do { const void *__vpp_verify = (typeof((worker) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __asm__ ("" : "=r"(__ptr) : "0"((typeof(*((worker))) *)((worker)))); (typeof((typeof(*((worker))) *)((worker)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); })->work) &wg->static_identity.lock &handshake->lock &table->lock#2 irq_context: 0 (wq_completion)wg-kex-wg1#11 (work_completion)(&({ do { const void *__vpp_verify = (typeof((worker) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __asm__ ("" : "=r"(__ptr) : "0"((typeof(*((worker))) *)((worker)))); (typeof((typeof(*((worker))) *)((worker)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); })->work) &wg->static_identity.lock &handshake->lock rcu_read_lock_bh batched_entropy_u32.lock irq_context: 0 (wq_completion)wg-kex-wg1#11 (work_completion)(&({ do { const void *__vpp_verify = (typeof((worker) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __asm__ ("" : "=r"(__ptr) : "0"((typeof(*((worker))) *)((worker)))); (typeof((typeof(*((worker))) *)((worker)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); })->work) &wg->static_identity.lock &handshake->lock rcu_read_lock_bh &table->lock#2 irq_context: 0 (wq_completion)wg-kex-wg1#11 (work_completion)(&({ do { const void *__vpp_verify = (typeof((worker) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __asm__ ("" : "=r"(__ptr) : "0"((typeof(*((worker))) *)((worker)))); (typeof((typeof(*((worker))) *)((worker)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); })->work) &cookie->lock irq_context: 0 (wq_completion)wg-kex-wg1#11 (work_completion)(&({ do { const void *__vpp_verify = (typeof((worker) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __asm__ ("" : "=r"(__ptr) : "0"((typeof(*((worker))) *)((worker)))); (typeof((typeof(*((worker))) *)((worker)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); })->work) &cookie->lock irq_context: 0 (wq_completion)wg-kex-wg1#11 (work_completion)(&({ do { const void *__vpp_verify = (typeof((worker) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __asm__ ("" : "=r"(__ptr) : "0"((typeof(*((worker))) *)((worker)))); (typeof((typeof(*((worker))) *)((worker)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); })->work) &handshake->lock &c->lock irq_context: 0 (wq_completion)wg-kex-wg1#11 (work_completion)(&({ do { const void *__vpp_verify = (typeof((worker) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __asm__ ("" : "=r"(__ptr) : "0"((typeof(*((worker))) *)((worker)))); (typeof((typeof(*((worker))) *)((worker)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); })->work) &handshake->lock &____s->seqcount#2 irq_context: 0 (wq_completion)wg-kex-wg1#11 (work_completion)(&({ do { const void *__vpp_verify = (typeof((worker) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __asm__ ("" : "=r"(__ptr) : "0"((typeof(*((worker))) *)((worker)))); (typeof((typeof(*((worker))) *)((worker)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); })->work) &handshake->lock &pcp->lock &zone->lock irq_context: 0 (wq_completion)wg-kex-wg1#11 (work_completion)(&({ do { const void *__vpp_verify = (typeof((worker) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __asm__ ("" : "=r"(__ptr) : "0"((typeof(*((worker))) *)((worker)))); (typeof((typeof(*((worker))) *)((worker)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); })->work) &handshake->lock &____s->seqcount irq_context: 0 (wq_completion)wg-kex-wg1#11 (work_completion)(&({ do { const void *__vpp_verify = (typeof((worker) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __asm__ ("" : "=r"(__ptr) : "0"((typeof(*((worker))) *)((worker)))); (typeof((typeof(*((worker))) *)((worker)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); })->work) &peer->endpoint_lock irq_context: 0 (wq_completion)wg-kex-wg1#11 (work_completion)(&({ do { const void *__vpp_verify = (typeof((worker) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __asm__ ("" : "=r"(__ptr) : "0"((typeof(*((worker))) *)((worker)))); (typeof((typeof(*((worker))) *)((worker)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); })->work) &peer->endpoint_lock rcu_read_lock_bh rcu_read_lock &dir->lock#2 irq_context: 0 (wq_completion)wg-kex-wg1#11 (work_completion)(&({ do { const void *__vpp_verify = (typeof((worker) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __asm__ ("" : "=r"(__ptr) : "0"((typeof(*((worker))) *)((worker)))); (typeof((typeof(*((worker))) *)((worker)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); })->work) &peer->endpoint_lock rcu_read_lock_bh rcu_read_lock &ul->lock irq_context: 0 (wq_completion)wg-kex-wg1#11 (work_completion)(&({ do { const void *__vpp_verify = (typeof((worker) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __asm__ ("" : "=r"(__ptr) : "0"((typeof(*((worker))) *)((worker)))); (typeof((typeof(*((worker))) *)((worker)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); })->work) &peer->endpoint_lock rcu_read_lock_bh rcu_read_lock &obj_hash[i].lock irq_context: 0 (wq_completion)wg-kex-wg1#11 (work_completion)(&({ do { const void *__vpp_verify = (typeof((worker) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __asm__ ("" : "=r"(__ptr) : "0"((typeof(*((worker))) *)((worker)))); (typeof((typeof(*((worker))) *)((worker)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); })->work) &peer->endpoint_lock rcu_read_lock_bh batched_entropy_u32.lock irq_context: 0 (wq_completion)wg-kex-wg1#11 (work_completion)(&({ do { const void *__vpp_verify = (typeof((worker) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __asm__ ("" : "=r"(__ptr) : "0"((typeof(*((worker))) *)((worker)))); (typeof((typeof(*((worker))) *)((worker)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); })->work) &peer->endpoint_lock rcu_read_lock_bh rcu_read_lock &____s->seqcount#7 irq_context: 0 (wq_completion)wg-kex-wg1#11 (work_completion)(&({ do { const void *__vpp_verify = (typeof((worker) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __asm__ ("" : "=r"(__ptr) : "0"((typeof(*((worker))) *)((worker)))); (typeof((typeof(*((worker))) *)((worker)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); })->work) &peer->endpoint_lock rcu_read_lock_bh rcu_read_lock &____s->seqcount#10 irq_context: 0 (wq_completion)wg-kex-wg1#11 (work_completion)(&({ do { const void *__vpp_verify = (typeof((worker) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __asm__ ("" : "=r"(__ptr) : "0"((typeof(*((worker))) *)((worker)))); (typeof((typeof(*((worker))) *)((worker)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); })->work) &peer->endpoint_lock rcu_read_lock_bh rcu_read_lock rcu_read_lock_bh tk_core.seq.seqcount irq_context: 0 (wq_completion)wg-kex-wg1#11 (work_completion)(&({ do { const void *__vpp_verify = (typeof((worker) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __asm__ ("" : "=r"(__ptr) : "0"((typeof(*((worker))) *)((worker)))); (typeof((typeof(*((worker))) *)((worker)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); })->work) &peer->endpoint_lock rcu_read_lock_bh rcu_read_lock rcu_read_lock_bh &list->lock#5 irq_context: 0 (wq_completion)mld (work_completion)(&(&idev->mc_ifc_work)->work) &idev->mc_lock rcu_read_lock &rcu_state.gp_wq irq_context: 0 (wq_completion)mld (work_completion)(&(&idev->mc_ifc_work)->work) &idev->mc_lock rcu_read_lock &rcu_state.gp_wq &p->pi_lock irq_context: 0 (wq_completion)wg-crypt-wg2#5 (work_completion)(&peer->transmit_packet_work) rcu_read_lock_bh &base->lock irq_context: 0 (wq_completion)wg-crypt-wg2#5 (work_completion)(&peer->transmit_packet_work) &base->lock irq_context: 0 (wq_completion)wg-crypt-wg2#5 (work_completion)(&peer->transmit_packet_work) &base->lock &obj_hash[i].lock irq_context: 0 (wq_completion)wg-crypt-wg2#5 (work_completion)(&peer->transmit_packet_work) &peer->endpoint_lock rcu_read_lock_bh batched_entropy_u32.lock irq_context: 0 (wq_completion)wg-crypt-wg2#5 (work_completion)(&peer->transmit_packet_work) batched_entropy_u8.lock irq_context: 0 (wq_completion)wg-crypt-wg2#5 (work_completion)(&peer->transmit_packet_work) rcu_read_lock_bh &base->lock &obj_hash[i].lock irq_context: 0 (wq_completion)wg-kex-wg2#12 irq_context: 0 (wq_completion)wg-kex-wg2#12 (work_completion)(&({ do { const void *__vpp_verify = (typeof((worker) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __asm__ ("" : "=r"(__ptr) : "0"((typeof(*((worker))) *)((worker)))); (typeof((typeof(*((worker))) *)((worker)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); })->work) irq_context: 0 (wq_completion)wg-kex-wg2#12 (work_completion)(&({ do { const void *__vpp_verify = (typeof((worker) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __asm__ ("" : "=r"(__ptr) : "0"((typeof(*((worker))) *)((worker)))); (typeof((typeof(*((worker))) *)((worker)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); })->work) &r->consumer_lock#2 irq_context: 0 (wq_completion)wg-kex-wg2#12 (work_completion)(&({ do { const void *__vpp_verify = (typeof((worker) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __asm__ ("" : "=r"(__ptr) : "0"((typeof(*((worker))) *)((worker)))); (typeof((typeof(*((worker))) *)((worker)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); })->work) &wg->static_identity.lock irq_context: 0 (wq_completion)wg-kex-wg2#12 (work_completion)(&({ do { const void *__vpp_verify = (typeof((worker) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __asm__ ("" : "=r"(__ptr) : "0"((typeof(*((worker))) *)((worker)))); (typeof((typeof(*((worker))) *)((worker)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); })->work) &wg->static_identity.lock &handshake->lock irq_context: 0 (wq_completion)wg-kex-wg2#12 (work_completion)(&({ do { const void *__vpp_verify = (typeof((worker) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __asm__ ("" : "=r"(__ptr) : "0"((typeof(*((worker))) *)((worker)))); (typeof((typeof(*((worker))) *)((worker)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); })->work) &wg->static_identity.lock &handshake->lock irq_context: 0 (wq_completion)wg-kex-wg2#12 (work_completion)(&({ do { const void *__vpp_verify = (typeof((worker) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __asm__ ("" : "=r"(__ptr) : "0"((typeof(*((worker))) *)((worker)))); (typeof((typeof(*((worker))) *)((worker)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); })->work) &peer->endpoint_lock irq_context: 0 (wq_completion)wg-kex-wg2#12 (work_completion)(&({ do { const void *__vpp_verify = (typeof((worker) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __asm__ ("" : "=r"(__ptr) : "0"((typeof(*((worker))) *)((worker)))); (typeof((typeof(*((worker))) *)((worker)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); })->work) &handshake->lock irq_context: 0 (wq_completion)wg-kex-wg2#12 (work_completion)(&({ do { const void *__vpp_verify = (typeof((worker) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __asm__ ("" : "=r"(__ptr) : "0"((typeof(*((worker))) *)((worker)))); (typeof((typeof(*((worker))) *)((worker)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); })->work) &handshake->lock fs_reclaim irq_context: 0 (wq_completion)wg-kex-wg2#12 (work_completion)(&({ do { const void *__vpp_verify = (typeof((worker) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __asm__ ("" : "=r"(__ptr) : "0"((typeof(*((worker))) *)((worker)))); (typeof((typeof(*((worker))) *)((worker)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); })->work) &handshake->lock fs_reclaim mmu_notifier_invalidate_range_start irq_context: 0 (wq_completion)wg-kex-wg2#12 (work_completion)(&({ do { const void *__vpp_verify = (typeof((worker) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __asm__ ("" : "=r"(__ptr) : "0"((typeof(*((worker))) *)((worker)))); (typeof((typeof(*((worker))) *)((worker)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); })->work) &handshake->lock tk_core.seq.seqcount irq_context: 0 (wq_completion)wg-kex-wg2#12 (work_completion)(&({ do { const void *__vpp_verify = (typeof((worker) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __asm__ ("" : "=r"(__ptr) : "0"((typeof(*((worker))) *)((worker)))); (typeof((typeof(*((worker))) *)((worker)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); })->work) &handshake->lock rcu_read_lock_bh &peer->keypairs.keypair_update_lock irq_context: 0 (wq_completion)wg-kex-wg2#12 (work_completion)(&({ do { const void *__vpp_verify = (typeof((worker) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __asm__ ("" : "=r"(__ptr) : "0"((typeof(*((worker))) *)((worker)))); (typeof((typeof(*((worker))) *)((worker)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); })->work) &handshake->lock rcu_read_lock_bh &table->lock#2 irq_context: 0 (wq_completion)wg-kex-wg2#12 (work_completion)(&({ do { const void *__vpp_verify = (typeof((worker) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __asm__ ("" : "=r"(__ptr) : "0"((typeof(*((worker))) *)((worker)))); (typeof((typeof(*((worker))) *)((worker)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); })->work) rcu_read_lock_bh &obj_hash[i].lock irq_context: 0 (wq_completion)wg-kex-wg2#12 (work_completion)(&({ do { const void *__vpp_verify = (typeof((worker) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __asm__ ("" : "=r"(__ptr) : "0"((typeof(*((worker))) *)((worker)))); (typeof((typeof(*((worker))) *)((worker)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); })->work) rcu_read_lock_bh &base->lock irq_context: 0 (wq_completion)wg-kex-wg2#12 (work_completion)(&({ do { const void *__vpp_verify = (typeof((worker) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __asm__ ("" : "=r"(__ptr) : "0"((typeof(*((worker))) *)((worker)))); (typeof((typeof(*((worker))) *)((worker)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); })->work) rcu_read_lock_bh &base->lock fill_pool_map-wait-type-override pool_lock#2 irq_context: 0 (wq_completion)wg-kex-wg2#12 (work_completion)(&({ do { const void *__vpp_verify = (typeof((worker) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __asm__ ("" : "=r"(__ptr) : "0"((typeof(*((worker))) *)((worker)))); (typeof((typeof(*((worker))) *)((worker)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); })->work) rcu_read_lock_bh &base->lock fill_pool_map-wait-type-override &c->lock irq_context: 0 (wq_completion)wg-kex-wg2#12 (work_completion)(&({ do { const void *__vpp_verify = (typeof((worker) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __asm__ ("" : "=r"(__ptr) : "0"((typeof(*((worker))) *)((worker)))); (typeof((typeof(*((worker))) *)((worker)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); })->work) rcu_read_lock_bh &base->lock fill_pool_map-wait-type-override &____s->seqcount#2 irq_context: 0 (wq_completion)wg-kex-wg2#12 (work_completion)(&({ do { const void *__vpp_verify = (typeof((worker) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __asm__ ("" : "=r"(__ptr) : "0"((typeof(*((worker))) *)((worker)))); (typeof((typeof(*((worker))) *)((worker)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); })->work) rcu_read_lock_bh &base->lock fill_pool_map-wait-type-override &____s->seqcount irq_context: 0 (wq_completion)wg-kex-wg2#12 (work_completion)(&({ do { const void *__vpp_verify = (typeof((worker) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __asm__ ("" : "=r"(__ptr) : "0"((typeof(*((worker))) *)((worker)))); (typeof((typeof(*((worker))) *)((worker)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); })->work) rcu_read_lock_bh &base->lock fill_pool_map-wait-type-override pool_lock irq_context: 0 (wq_completion)wg-kex-wg2#12 (work_completion)(&({ do { const void *__vpp_verify = (typeof((worker) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __asm__ ("" : "=r"(__ptr) : "0"((typeof(*((worker))) *)((worker)))); (typeof((typeof(*((worker))) *)((worker)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); })->work) rcu_read_lock_bh &base->lock &obj_hash[i].lock irq_context: 0 (wq_completion)wg-kex-wg2#12 (work_completion)(&({ do { const void *__vpp_verify = (typeof((worker) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __asm__ ("" : "=r"(__ptr) : "0"((typeof(*((worker))) *)((worker)))); (typeof((typeof(*((worker))) *)((worker)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); })->work) &obj_hash[i].lock irq_context: 0 (wq_completion)wg-kex-wg2#12 (work_completion)(&({ do { const void *__vpp_verify = (typeof((worker) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __asm__ ("" : "=r"(__ptr) : "0"((typeof(*((worker))) *)((worker)))); (typeof((typeof(*((worker))) *)((worker)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); })->work) &base->lock irq_context: 0 (wq_completion)wg-kex-wg2#12 (work_completion)(&({ do { const void *__vpp_verify = (typeof((worker) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __asm__ ("" : "=r"(__ptr) : "0"((typeof(*((worker))) *)((worker)))); (typeof((typeof(*((worker))) *)((worker)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); })->work) &base->lock &obj_hash[i].lock irq_context: 0 (wq_completion)wg-kex-wg2#12 (work_completion)(&({ do { const void *__vpp_verify = (typeof((worker) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __asm__ ("" : "=r"(__ptr) : "0"((typeof(*((worker))) *)((worker)))); (typeof((typeof(*((worker))) *)((worker)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); })->work) tk_core.seq.seqcount irq_context: 0 (wq_completion)wg-kex-wg2#12 (work_completion)(&({ do { const void *__vpp_verify = (typeof((worker) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __asm__ ("" : "=r"(__ptr) : "0"((typeof(*((worker))) *)((worker)))); (typeof((typeof(*((worker))) *)((worker)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); })->work) &list->lock#17 irq_context: 0 (wq_completion)wg-kex-wg2#12 (work_completion)(&({ do { const void *__vpp_verify = (typeof((worker) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __asm__ ("" : "=r"(__ptr) : "0"((typeof(*((worker))) *)((worker)))); (typeof((typeof(*((worker))) *)((worker)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); })->work) rcu_read_lock_bh &r->producer_lock#2 irq_context: 0 (wq_completion)wg-kex-wg2#12 (work_completion)(&({ do { const void *__vpp_verify = (typeof((worker) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __asm__ ("" : "=r"(__ptr) : "0"((typeof(*((worker))) *)((worker)))); (typeof((typeof(*((worker))) *)((worker)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); })->work) rcu_read_lock_bh rcu_read_lock &pool->lock irq_context: 0 (wq_completion)ipv6_addrconf (work_completion)(&(&ifa->dad_work)->work) rtnl_mutex &idev->mc_lock &macsec_netdev_addr_lock_key/1 _xmit_ETHER &____s->seqcount#2 irq_context: 0 (wq_completion)wg-kex-wg2#12 (work_completion)(&({ do { const void *__vpp_verify = (typeof((worker) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __asm__ ("" : "=r"(__ptr) : "0"((typeof(*((worker))) *)((worker)))); (typeof((typeof(*((worker))) *)((worker)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); })->work) rcu_read_lock_bh rcu_read_lock &pool->lock &obj_hash[i].lock irq_context: 0 (wq_completion)wg-crypt-wg2#6 irq_context: 0 (wq_completion)wg-crypt-wg2#6 (work_completion)(&({ do { const void *__vpp_verify = (typeof((worker) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __asm__ ("" : "=r"(__ptr) : "0"((typeof(*((worker))) *)((worker)))); (typeof((typeof(*((worker))) *)((worker)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); })->work) irq_context: 0 (wq_completion)wg-crypt-wg2#6 (work_completion)(&({ do { const void *__vpp_verify = (typeof((worker) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __asm__ ("" : "=r"(__ptr) : "0"((typeof(*((worker))) *)((worker)))); (typeof((typeof(*((worker))) *)((worker)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); })->work) &r->consumer_lock#2 irq_context: 0 (wq_completion)wg-crypt-wg2#6 (work_completion)(&({ do { const void *__vpp_verify = (typeof((worker) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __asm__ ("" : "=r"(__ptr) : "0"((typeof(*((worker))) *)((worker)))); (typeof((typeof(*((worker))) *)((worker)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); })->work) rcu_read_lock &pool->lock irq_context: 0 (wq_completion)wg-crypt-wg2#6 (work_completion)(&({ do { const void *__vpp_verify = (typeof((worker) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __asm__ ("" : "=r"(__ptr) : "0"((typeof(*((worker))) *)((worker)))); (typeof((typeof(*((worker))) *)((worker)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); })->work) rcu_read_lock &pool->lock &obj_hash[i].lock irq_context: 0 (wq_completion)wg-crypt-wg2#6 (work_completion)(&({ do { const void *__vpp_verify = (typeof((worker) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __asm__ ("" : "=r"(__ptr) : "0"((typeof(*((worker))) *)((worker)))); (typeof((typeof(*((worker))) *)((worker)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); })->work) rcu_read_lock &pool->lock &p->pi_lock irq_context: 0 (wq_completion)wg-kex-wg2#12 (work_completion)(&({ do { const void *__vpp_verify = (typeof((worker) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __asm__ ("" : "=r"(__ptr) : "0"((typeof(*((worker))) *)((worker)))); (typeof((typeof(*((worker))) *)((worker)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); })->work) rcu_read_lock_bh rcu_read_lock &pool->lock &p->pi_lock irq_context: 0 (wq_completion)wg-crypt-wg2#6 (work_completion)(&peer->transmit_packet_work) irq_context: 0 (wq_completion)wg-crypt-wg2#6 (work_completion)(&peer->transmit_packet_work) rcu_read_lock_bh &obj_hash[i].lock irq_context: 0 (wq_completion)wg-crypt-wg2#6 (work_completion)(&peer->transmit_packet_work) &obj_hash[i].lock irq_context: 0 (wq_completion)wg-crypt-wg2#6 (work_completion)(&peer->transmit_packet_work) &peer->endpoint_lock irq_context: 0 (wq_completion)wg-crypt-wg2#6 (work_completion)(&peer->transmit_packet_work) &peer->endpoint_lock rcu_read_lock_bh rcu_read_lock &____s->seqcount#7 irq_context: 0 (wq_completion)wg-crypt-wg2#6 (work_completion)(&peer->transmit_packet_work) &peer->endpoint_lock rcu_read_lock_bh rcu_read_lock &____s->seqcount#10 irq_context: 0 (wq_completion)wg-crypt-wg2#6 (work_completion)(&peer->transmit_packet_work) &peer->endpoint_lock rcu_read_lock_bh rcu_read_lock rcu_read_lock_bh tk_core.seq.seqcount irq_context: 0 (wq_completion)wg-crypt-wg2#6 (work_completion)(&peer->transmit_packet_work) &peer->endpoint_lock rcu_read_lock_bh rcu_read_lock rcu_read_lock_bh &list->lock#5 irq_context: 0 (wq_completion)wg-crypt-wg2#6 (work_completion)(&peer->transmit_packet_work) rcu_read_lock_bh tk_core.seq.seqcount irq_context: 0 (wq_completion)wg-crypt-wg1#2 (work_completion)(&peer->transmit_packet_work) rcu_read_lock_bh &base->lock irq_context: 0 (wq_completion)wg-crypt-wg1#2 (work_completion)(&peer->transmit_packet_work) rcu_read_lock_bh &base->lock &obj_hash[i].lock irq_context: 0 (wq_completion)wg-crypt-wg1#2 (work_completion)(&peer->transmit_packet_work) &peer->endpoint_lock rcu_read_lock_bh rcu_read_lock &dir->lock#2 irq_context: 0 (wq_completion)wg-crypt-wg1#2 (work_completion)(&peer->transmit_packet_work) &peer->endpoint_lock rcu_read_lock_bh rcu_read_lock &ul->lock irq_context: 0 (wq_completion)wg-crypt-wg1#2 (work_completion)(&peer->transmit_packet_work) &peer->endpoint_lock rcu_read_lock_bh batched_entropy_u32.lock irq_context: 0 (wq_completion)wg-crypt-wg1#2 (work_completion)(&peer->transmit_packet_work) batched_entropy_u8.lock irq_context: 0 (wq_completion)wg-crypt-wg2#6 (work_completion)(&peer->transmit_packet_work) batched_entropy_u8.lock irq_context: 0 (wq_completion)wg-crypt-wg2#6 (work_completion)(&peer->transmit_packet_work) rcu_read_lock_bh &base->lock irq_context: 0 (wq_completion)wg-crypt-wg2#6 (work_completion)(&peer->transmit_packet_work) rcu_read_lock_bh &base->lock &obj_hash[i].lock irq_context: 0 misc_mtx rfkill_global_mutex &data->mtx &____s->seqcount#2 irq_context: 0 (wq_completion)wg-crypt-wg1#6 (work_completion)(&peer->transmit_packet_work) batched_entropy_u8.lock irq_context: 0 (wq_completion)wg-crypt-wg1#6 (work_completion)(&peer->transmit_packet_work) rcu_read_lock_bh &base->lock &obj_hash[i].lock irq_context: 0 (wq_completion)wg-crypt-wg2#6 (work_completion)(&({ do { const void *__vpp_verify = (typeof((worker) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __asm__ ("" : "=r"(__ptr) : "0"((typeof(*((worker))) *)((worker)))); (typeof((typeof(*((worker))) *)((worker)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); })->work) tk_core.seq.seqcount irq_context: 0 (wq_completion)wg-crypt-wg2#2 (work_completion)(&peer->transmit_packet_work) &base->lock irq_context: 0 (wq_completion)wg-crypt-wg2#2 (work_completion)(&peer->transmit_packet_work) &base->lock &obj_hash[i].lock irq_context: 0 (wq_completion)wg-crypt-wg2#2 (work_completion)(&peer->transmit_packet_work) &peer->endpoint_lock rcu_read_lock_bh &obj_hash[i].lock irq_context: 0 (wq_completion)wg-crypt-wg2#2 (work_completion)(&peer->transmit_packet_work) &peer->endpoint_lock rcu_read_lock_bh &obj_hash[i].lock pool_lock irq_context: 0 (wq_completion)wg-crypt-wg2#2 (work_completion)(&peer->transmit_packet_work) &rq->__lock irq_context: 0 cb_lock rtnl_mutex rcu_read_lock &rq->__lock irq_context: 0 cb_lock rtnl_mutex rcu_read_lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 &rq->__lock per_cpu_ptr(&cgroup_rstat_cpu_lock, cpu) irq_context: 0 (wq_completion)phy7 irq_context: 0 (wq_completion)phy7 (work_completion)(&local->reconfig_filter) irq_context: 0 (wq_completion)phy7 (work_completion)(&local->reconfig_filter) &local->filter_lock irq_context: 0 cb_lock genl_mutex rcu_read_lock &pool->lock/1 fill_pool_map-wait-type-override pool_lock#2 irq_context: 0 cb_lock genl_mutex rcu_read_lock &pool->lock/1 fill_pool_map-wait-type-override pool_lock irq_context: 0 fill_pool_map-wait-type-override rcu_read_lock pool_lock#2 irq_context: 0 &xt[i].mutex rcu_read_lock rcu_node_0 irq_context: 0 &xt[i].mutex rcu_read_lock &rq->__lock irq_context: 0 &xt[i].mutex rcu_read_lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)wg-crypt-wg2#2 (work_completion)(&peer->transmit_packet_work) batched_entropy_u8.lock irq_context: 0 (wq_completion)wg-crypt-wg2#2 (work_completion)(&peer->transmit_packet_work) rcu_read_lock_bh &base->lock &obj_hash[i].lock irq_context: 0 &sb->s_type->i_mutex_key#10 sk_lock-AF_INET slock-AF_INET fill_pool_map-wait-type-override pool_lock#2 irq_context: 0 &sb->s_type->i_mutex_key#10 sk_lock-AF_INET slock-AF_INET fill_pool_map-wait-type-override pool_lock irq_context: 0 (wq_completion)events (work_completion)(&rdev->wiphy_work) &rdev->wiphy.mtx &wdev->mtx rcu_read_lock &pool->lock/1 &p->pi_lock &rq->__lock irq_context: 0 (wq_completion)events (work_completion)(&rdev->wiphy_work) &rdev->wiphy.mtx &wdev->mtx rcu_read_lock &pool->lock/1 &p->pi_lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 sb_writers#5 &type->i_mutex_dir_key#5 remove_cache_srcu &rq->__lock irq_context: 0 (wq_completion)mld (work_completion)(&(&idev->mc_ifc_work)->work) &idev->mc_lock rcu_read_lock rcu_read_lock rcu_read_lock_bh rcu_read_lock batched_entropy_u32.lock crngs.lock irq_context: 0 cb_lock genl_mutex rtnl_mutex &rdev->wiphy.mtx &idev->mc_lock _xmit_ETHER &____s->seqcount#2 irq_context: 0 (wq_completion)ipv6_addrconf (work_completion)(&(&ifa->dad_work)->work) rtnl_mutex rcu_read_lock rcu_read_lock rcu_read_lock_bh rcu_read_lock rcu_read_lock &c->lock irq_context: 0 (wq_completion)ipv6_addrconf (work_completion)(&(&ifa->dad_work)->work) rtnl_mutex rcu_read_lock rcu_read_lock rcu_read_lock_bh rcu_read_lock rcu_read_lock &____s->seqcount#2 irq_context: 0 (wq_completion)ipv6_addrconf (work_completion)(&(&ifa->dad_work)->work) rtnl_mutex rcu_read_lock rcu_read_lock rcu_read_lock_bh rcu_read_lock rcu_read_lock &____s->seqcount irq_context: 0 cb_lock genl_mutex rtnl_mutex &rdev->wiphy.mtx &sb->s_type->i_mutex_key#3 &rq->__lock irq_context: 0 sb_writers#5 &type->i_mutex_dir_key#5/1 &sb->s_type->i_lock_key &xa->xa_lock#9 rcu_read_lock per_cpu_ptr(&cgroup_rstat_cpu_lock, cpu) irq_context: 0 sb_writers#5 &type->i_mutex_dir_key#5/1 lock#4 &lruvec->lru_lock per_cpu_ptr(&cgroup_rstat_cpu_lock, cpu) irq_context: 0 dup_mmap_sem &mm->mmap_lock &mm->mmap_lock/1 per_cpu_ptr(&cgroup_rstat_cpu_lock, cpu) irq_context: 0 pfkey_mutex irq_context: 0 vlan_ioctl_mutex rtnl_mutex rtnl_mutex.wait_lock irq_context: 0 vlan_ioctl_mutex rtnl_mutex &rq->__lock irq_context: 0 vlan_ioctl_mutex rtnl_mutex &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 vlan_ioctl_mutex rtnl_mutex rcu_read_lock &rq->__lock irq_context: 0 vlan_ioctl_mutex rtnl_mutex rcu_read_lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 cb_lock genl_mutex rfkill_global_mutex uevent_sock_mutex rcu_read_lock &rcu_state.gp_wq irq_context: 0 cb_lock genl_mutex rfkill_global_mutex uevent_sock_mutex rcu_read_lock &rcu_state.gp_wq &p->pi_lock irq_context: 0 cb_lock genl_mutex rfkill_global_mutex uevent_sock_mutex rcu_read_lock &rcu_state.gp_wq &p->pi_lock &rq->__lock irq_context: 0 cb_lock genl_mutex rfkill_global_mutex uevent_sock_mutex rcu_read_lock &rcu_state.gp_wq &p->pi_lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 cb_lock genl_mutex rfkill_global_mutex uevent_sock_mutex rcu_read_lock &rcu_state.gp_wq &p->pi_lock &rq->__lock &cfs_rq->removed.lock irq_context: 0 (wq_completion)phy8 irq_context: 0 (wq_completion)phy8 (work_completion)(&local->reconfig_filter) irq_context: 0 (wq_completion)phy8 (work_completion)(&local->reconfig_filter) &local->filter_lock irq_context: 0 vlan_ioctl_mutex rtnl_mutex.wait_lock irq_context: 0 vlan_ioctl_mutex &p->pi_lock irq_context: 0 vlan_ioctl_mutex &p->pi_lock &rq->__lock irq_context: 0 vlan_ioctl_mutex &p->pi_lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)events (work_completion)(&rdev->wiphy_work) &rdev->wiphy.mtx &wdev->mtx &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 cb_lock genl_mutex rtnl_mutex &rdev->wiphy.mtx gdp_mutex &c->lock irq_context: 0 cb_lock genl_mutex rtnl_mutex &rdev->wiphy.mtx gdp_mutex &____s->seqcount#2 irq_context: 0 cb_lock genl_mutex rtnl_mutex &rdev->wiphy.mtx gdp_mutex &____s->seqcount irq_context: 0 cb_lock genl_mutex rtnl_mutex &rdev->wiphy.mtx gdp_mutex &rq->__lock irq_context: 0 cb_lock genl_mutex &sb->s_type->i_mutex_key#3 fs_reclaim &rq->__lock irq_context: 0 cb_lock genl_mutex &sb->s_type->i_mutex_key#3 fs_reclaim &rq->__lock &cfs_rq->removed.lock irq_context: 0 cb_lock &p->pi_lock &cfs_rq->removed.lock irq_context: 0 cb_lock rtnl_mutex &rq->__lock &cfs_rq->removed.lock irq_context: 0 cb_lock genl_mutex &root->kernfs_rwsem &sem->wait_lock irq_context: 0 cb_lock genl_mutex &root->kernfs_rwsem &rq->__lock irq_context: 0 cb_lock genl_mutex &root->kernfs_rwsem &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 &sb->s_type->i_mutex_key#10 (netlink_chain).rwsem rcu_read_lock &pool->lock &p->pi_lock &rq->__lock &cfs_rq->removed.lock irq_context: 0 &type->s_umount_key#49/1 &sb->s_type->i_mutex_key#20 &____s->seqcount#2 irq_context: 0 &type->s_umount_key#49/1 &sb->s_type->i_mutex_key#20 &pcp->lock &zone->lock irq_context: 0 &type->s_umount_key#49/1 &sb->s_type->i_mutex_key#20 &pcp->lock &zone->lock &____s->seqcount irq_context: 0 sb_writers#10 &type->i_mutex_dir_key#6/1 tomoyo_ss &____s->seqcount#2 irq_context: 0 sb_writers#10 &type->i_mutex_dir_key#6/1 cgroup_mutex fill_pool_map-wait-type-override pool_lock#2 irq_context: 0 sb_writers#10 &type->i_mutex_dir_key#6/1 cgroup_mutex fill_pool_map-wait-type-override &c->lock irq_context: 0 sb_writers#10 &type->i_mutex_dir_key#6/1 cgroup_mutex fill_pool_map-wait-type-override &____s->seqcount#2 irq_context: 0 sb_writers#10 &type->i_mutex_dir_key#6/1 cgroup_mutex fill_pool_map-wait-type-override &____s->seqcount irq_context: 0 sb_writers#10 &type->i_mutex_dir_key#6/1 cgroup_mutex fill_pool_map-wait-type-override pool_lock irq_context: 0 cb_lock rtnl_mutex &rdev->wiphy.mtx &sb->s_type->i_mutex_key#3 fill_pool_map-wait-type-override &rq->__lock irq_context: 0 sb_writers#10 &type->i_mutex_dir_key#6/1 cgroup_mutex fill_pool_map-wait-type-override &rq->__lock irq_context: 0 sb_writers#10 &type->i_mutex_dir_key#6 &n->list_lock irq_context: 0 kn->active#54 &c->lock irq_context: 0 sb_writers#10 &of->mutex cgroup_mutex cpu_hotplug_lock cgroup_threadgroup_rwsem rcu_state.exp_mutex &rnp->exp_wq[1] irq_context: 0 sb_writers#11 &type->i_mutex_dir_key#7/1 cgroup_mutex &pcp->lock &zone->lock irq_context: 0 kn->active#55 &____s->seqcount#2 irq_context: 0 kn->active#55 &____s->seqcount irq_context: 0 sb_writers#11 &c->lock irq_context: 0 kn->active#57 &kernfs_locks->open_file_mutex[count] &c->lock irq_context: 0 kn->active#57 &kernfs_locks->open_file_mutex[count] &____s->seqcount#2 irq_context: 0 kn->active#57 &kernfs_locks->open_file_mutex[count] &____s->seqcount irq_context: 0 sb_writers#11 &type->i_mutex_dir_key#7/1 cgroup_mutex rtnl_mutex rcu_read_lock &rq->__lock irq_context: 0 sb_writers#11 &type->i_mutex_dir_key#7/1 cgroup_mutex rtnl_mutex rcu_read_lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 cb_lock rtnl_mutex &rdev->wiphy.mtx &rq->__lock irq_context: 0 cb_lock rtnl_mutex &rdev->wiphy.mtx &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)phy9 irq_context: 0 (wq_completion)phy9 (work_completion)(&local->reconfig_filter) irq_context: 0 (wq_completion)phy9 (work_completion)(&local->reconfig_filter) &local->filter_lock irq_context: 0 cb_lock genl_mutex rtnl_mutex &rdev->wiphy.mtx uevent_sock_mutex &n->list_lock irq_context: 0 cb_lock genl_mutex rtnl_mutex &rdev->wiphy.mtx uevent_sock_mutex &n->list_lock &c->lock irq_context: 0 (wq_completion)ipv6_addrconf (work_completion)(&(&ifa->dad_work)->work) rtnl_mutex &idev->mc_lock rcu_read_lock rcu_read_lock rcu_read_lock_bh rcu_read_lock rcu_read_lock pool_lock#2 irq_context: 0 (wq_completion)ipv6_addrconf (work_completion)(&(&ifa->dad_work)->work) rtnl_mutex &idev->mc_lock rcu_read_lock rcu_read_lock rcu_read_lock_bh rcu_read_lock rcu_read_lock &____s->seqcount irq_context: 0 (wq_completion)ipv6_addrconf (work_completion)(&(&ifa->dad_work)->work) rtnl_mutex &idev->mc_lock rcu_read_lock rcu_read_lock rcu_read_lock_bh rcu_read_lock rcu_read_lock rcu_read_lock pool_lock#2 irq_context: 0 (wq_completion)ipv6_addrconf (work_completion)(&(&ifa->dad_work)->work) rtnl_mutex &idev->mc_lock rcu_read_lock rcu_read_lock rcu_read_lock_bh rcu_read_lock rcu_read_lock &obj_hash[i].lock irq_context: 0 (wq_completion)ipv6_addrconf (work_completion)(&(&ifa->dad_work)->work) rtnl_mutex &idev->mc_lock rcu_read_lock rcu_read_lock rcu_read_lock_bh rcu_read_lock rcu_read_lock &dir->lock#2 irq_context: 0 (wq_completion)ipv6_addrconf (work_completion)(&(&ifa->dad_work)->work) rtnl_mutex &idev->mc_lock rcu_read_lock rcu_read_lock rcu_read_lock_bh rcu_read_lock rcu_read_lock &ul->lock irq_context: 0 (wq_completion)ipv6_addrconf (work_completion)(&(&ifa->dad_work)->work) rtnl_mutex &idev->mc_lock rcu_read_lock rcu_read_lock rcu_read_lock_bh rcu_read_lock &obj_hash[i].lock irq_context: 0 (wq_completion)ipv6_addrconf (work_completion)(&(&ifa->dad_work)->work) rtnl_mutex &idev->mc_lock rcu_read_lock rcu_read_lock rcu_read_lock_bh rcu_read_lock batched_entropy_u32.lock irq_context: 0 (wq_completion)ipv6_addrconf (work_completion)(&(&ifa->dad_work)->work) rtnl_mutex &idev->mc_lock rcu_read_lock rcu_read_lock rcu_read_lock_bh rcu_read_lock rcu_read_lock &____s->seqcount#7 irq_context: 0 (wq_completion)ipv6_addrconf (work_completion)(&(&ifa->dad_work)->work) rtnl_mutex &idev->mc_lock rcu_read_lock rcu_read_lock rcu_read_lock_bh rcu_read_lock rcu_read_lock rcu_read_lock &____s->seqcount#7 irq_context: 0 (wq_completion)ipv6_addrconf (work_completion)(&(&ifa->dad_work)->work) rtnl_mutex &idev->mc_lock rcu_read_lock rcu_read_lock rcu_read_lock_bh rcu_read_lock rcu_read_lock &nf_nat_locks[i] irq_context: 0 (wq_completion)ipv6_addrconf (work_completion)(&(&ifa->dad_work)->work) rtnl_mutex &idev->mc_lock rcu_read_lock rcu_read_lock rcu_read_lock_bh rcu_read_lock rcu_read_lock &nf_conntrack_locks[i] irq_context: 0 (wq_completion)ipv6_addrconf (work_completion)(&(&ifa->dad_work)->work) rtnl_mutex &idev->mc_lock rcu_read_lock rcu_read_lock rcu_read_lock_bh rcu_read_lock rcu_read_lock &nf_conntrack_locks[i] &nf_conntrack_locks[i]/1 irq_context: 0 (wq_completion)ipv6_addrconf (work_completion)(&(&ifa->dad_work)->work) rtnl_mutex &idev->mc_lock rcu_read_lock rcu_read_lock rcu_read_lock_bh rcu_read_lock rcu_read_lock &nf_conntrack_locks[i] &nf_conntrack_locks[i]/1 batched_entropy_u8.lock irq_context: 0 (wq_completion)ipv6_addrconf (work_completion)(&(&ifa->dad_work)->work) rtnl_mutex &idev->mc_lock rcu_read_lock rcu_read_lock rcu_read_lock_bh rcu_read_lock rcu_read_lock &____s->seqcount#10 irq_context: 0 (wq_completion)ipv6_addrconf (work_completion)(&(&ifa->dad_work)->work) rtnl_mutex &idev->mc_lock rcu_read_lock rcu_read_lock rcu_read_lock_bh rcu_read_lock rcu_read_lock rcu_read_lock_bh tk_core.seq.seqcount irq_context: 0 (wq_completion)ipv6_addrconf (work_completion)(&(&ifa->dad_work)->work) rtnl_mutex &idev->mc_lock rcu_read_lock rcu_read_lock rcu_read_lock_bh rcu_read_lock rcu_read_lock rcu_read_lock_bh &list->lock#5 irq_context: 0 (wq_completion)ipv6_addrconf (work_completion)(&(&ifa->dad_work)->work) rtnl_mutex rcu_read_lock rcu_read_lock rcu_read_lock_bh rcu_read_lock rcu_read_lock &nf_conntrack_locks[i] &nf_conntrack_locks[i]/1 batched_entropy_u8.lock crngs.lock irq_context: 0 (wq_completion)ipv6_addrconf (work_completion)(&(&ifa->dad_work)->work) rtnl_mutex rcu_read_lock rcu_read_lock rcu_read_lock_bh rcu_read_lock_bh fill_pool_map-wait-type-override pool_lock#2 irq_context: 0 (wq_completion)ipv6_addrconf (work_completion)(&(&ifa->dad_work)->work) rtnl_mutex rcu_read_lock rcu_read_lock rcu_read_lock_bh rcu_read_lock_bh fill_pool_map-wait-type-override pool_lock irq_context: 0 nf_nat_proto_mutex &c->lock irq_context: 0 nf_nat_proto_mutex &____s->seqcount#2 irq_context: 0 nf_nat_proto_mutex &____s->seqcount irq_context: 0 nf_hook_mutex per_cpu_ptr(&cgroup_rstat_cpu_lock, cpu) irq_context: 0 pcpu_alloc_mutex fs_reclaim irq_context: 0 pcpu_alloc_mutex fs_reclaim mmu_notifier_invalidate_range_start irq_context: 0 pcpu_alloc_mutex &____s->seqcount irq_context: 0 sb_writers#4 &type->i_mutex_dir_key#3/1 jbd2_handle &ei->i_data_sem &____s->seqcount#2 irq_context: 0 (wq_completion)events (work_completion)(&rdev->wiphy_work) &rdev->wiphy.mtx &wdev->mtx &local->mtx &local->chanctx_mtx rcu_read_lock rcu_read_lock rcu_node_0 irq_context: 0 (wq_completion)events (work_completion)(&rdev->wiphy_work) &rdev->wiphy.mtx &wdev->mtx &local->mtx &local->chanctx_mtx rcu_read_lock rcu_read_lock &rq->__lock irq_context: 0 (wq_completion)events (work_completion)(&rdev->wiphy_work) &rdev->wiphy.mtx &wdev->mtx &local->mtx &local->chanctx_mtx rcu_read_lock rcu_node_0 irq_context: softirq rcu_read_lock &____s->seqcount#2 irq_context: softirq rcu_read_lock &ifibss->incomplete_lock irq_context: softirq rcu_read_lock &rdev->wiphy_work_lock irq_context: softirq rcu_read_lock &local->rx_path_lock &obj_hash[i].lock irq_context: softirq rcu_read_lock &local->rx_path_lock pool_lock#2 irq_context: 0 (wq_completion)events (work_completion)(&rdev->wiphy_work) &rdev->wiphy.mtx &wdev->mtx &local->sta_mtx &sb->s_type->i_mutex_key#3 rcu_read_lock pool_lock#2 irq_context: 0 (wq_completion)events (work_completion)(&rdev->wiphy_work) &rdev->wiphy.mtx &wdev->mtx &local->sta_mtx &sb->s_type->i_mutex_key#3 &obj_hash[i].lock irq_context: 0 &type->s_umount_key#50 irq_context: 0 &type->s_umount_key#50 &rq->__lock irq_context: 0 (wq_completion)events_unbound (work_completion)(&port->bc_work) rcu_read_lock pool_lock#2 irq_context: 0 (wq_completion)events_unbound (work_completion)(&port->bc_work) rcu_read_lock &list->lock#5 irq_context: 0 (wq_completion)bat_events (work_completion)(&(&forw_packet_aggr->delayed_work)->work) &hard_iface->bat_iv.ogm_buff_mutex fill_pool_map-wait-type-override pool_lock#2 irq_context: 0 (wq_completion)bat_events (work_completion)(&(&forw_packet_aggr->delayed_work)->work) &hard_iface->bat_iv.ogm_buff_mutex fill_pool_map-wait-type-override pool_lock irq_context: 0 &sig->cred_guard_mutex tomoyo_ss &____s->seqcount#2 irq_context: 0 sb_writers#8 tomoyo_ss &rq->__lock irq_context: 0 cb_lock genl_mutex uevent_sock_mutex &n->list_lock irq_context: 0 cb_lock genl_mutex uevent_sock_mutex &n->list_lock &c->lock irq_context: 0 (wq_completion)events (work_completion)(&(&nsim_dev->trap_data->trap_report_dw)->work) &devlink->lock_key &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)events (work_completion)(&rdev->wiphy_work) &rdev->wiphy.mtx &wdev->mtx rcu_read_lock rcu_read_lock &sta->rate_ctrl_lock irq_context: 0 (wq_completion)events (work_completion)(&rdev->wiphy_work) &rdev->wiphy.mtx &wdev->mtx rcu_read_lock rcu_read_lock &sta->rate_ctrl_lock pool_lock#2 irq_context: 0 (wq_completion)events (work_completion)(&rdev->wiphy_work) &rdev->wiphy.mtx &wdev->mtx rcu_read_lock rcu_read_lock &sta->rate_ctrl_lock &obj_hash[i].lock irq_context: 0 (wq_completion)events (work_completion)(&rdev->wiphy_work) &rdev->wiphy.mtx &wdev->mtx rcu_read_lock rcu_read_lock &sta->rate_ctrl_lock krc.lock irq_context: 0 &group->mark_mutex &____s->seqcount#2 irq_context: 0 cb_lock rtnl_mutex &rdev->wiphy.mtx &____s->seqcount#2 irq_context: 0 cb_lock rtnl_mutex &rdev->wiphy.mtx &pcp->lock &zone->lock irq_context: 0 rtnl_mutex _xmit_ETHER &local->filter_lock &c->lock irq_context: 0 rtnl_mutex _xmit_ETHER &local->filter_lock &____s->seqcount#2 irq_context: 0 rtnl_mutex _xmit_ETHER &local->filter_lock &____s->seqcount irq_context: 0 cb_lock genl_mutex rtnl_mutex &rdev->wiphy.mtx &sb->s_type->i_mutex_key#3 &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)phy10 irq_context: 0 (wq_completion)phy10 (work_completion)(&local->reconfig_filter) irq_context: 0 (wq_completion)phy10 (work_completion)(&local->reconfig_filter) &local->filter_lock irq_context: 0 (wq_completion)bat_events (work_completion)(&(&forw_packet_aggr->delayed_work)->work) &hard_iface->bat_iv.ogm_buff_mutex rcu_read_lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 cb_lock genl_mutex rtnl_mutex &rdev->wiphy.mtx &idev->mc_lock &c->lock irq_context: 0 cb_lock genl_mutex rtnl_mutex &rdev->wiphy.mtx &idev->mc_lock &____s->seqcount#2 irq_context: 0 cb_lock genl_mutex rtnl_mutex &rdev->wiphy.mtx &idev->mc_lock &____s->seqcount irq_context: 0 (wq_completion)ipv6_addrconf (work_completion)(&(&ifa->dad_work)->work) rtnl_mutex &idev->mc_lock rcu_read_lock rcu_read_lock rcu_read_lock_bh _xmit_ETHER#2 irq_context: 0 (wq_completion)ipv6_addrconf (work_completion)(&(&ifa->dad_work)->work) rtnl_mutex &idev->mc_lock rcu_read_lock rcu_read_lock rcu_read_lock_bh _xmit_ETHER#2 rcu_read_lock tk_core.seq.seqcount irq_context: 0 (wq_completion)ipv6_addrconf (work_completion)(&(&ifa->dad_work)->work) rtnl_mutex &idev->mc_lock rcu_read_lock rcu_read_lock rcu_read_lock_bh _xmit_ETHER#2 rcu_read_lock &fq->lock irq_context: 0 (wq_completion)ipv6_addrconf (work_completion)(&(&ifa->dad_work)->work) rtnl_mutex &idev->mc_lock rcu_read_lock rcu_read_lock rcu_read_lock_bh _xmit_ETHER#2 rcu_read_lock &local->active_txq_lock[i] irq_context: 0 (wq_completion)ipv6_addrconf (work_completion)(&(&ifa->dad_work)->work) rtnl_mutex &idev->mc_lock rcu_read_lock rcu_read_lock rcu_read_lock_bh _xmit_ETHER#2 rcu_read_lock &local->handle_wake_tx_queue_lock irq_context: 0 (wq_completion)ipv6_addrconf (work_completion)(&(&ifa->dad_work)->work) rtnl_mutex &idev->mc_lock rcu_read_lock rcu_read_lock rcu_read_lock_bh _xmit_ETHER#2 rcu_read_lock &local->handle_wake_tx_queue_lock &local->active_txq_lock[i] irq_context: 0 (wq_completion)ipv6_addrconf (work_completion)(&(&ifa->dad_work)->work) rtnl_mutex &idev->mc_lock rcu_read_lock rcu_read_lock rcu_read_lock_bh _xmit_ETHER#2 rcu_read_lock &local->handle_wake_tx_queue_lock &local->queue_stop_reason_lock irq_context: 0 (wq_completion)ipv6_addrconf (work_completion)(&(&ifa->dad_work)->work) rtnl_mutex &idev->mc_lock rcu_read_lock rcu_read_lock rcu_read_lock_bh _xmit_ETHER#2 rcu_read_lock &local->handle_wake_tx_queue_lock &fq->lock irq_context: 0 (wq_completion)ipv6_addrconf (work_completion)(&(&ifa->dad_work)->work) rtnl_mutex &idev->mc_lock rcu_read_lock rcu_read_lock rcu_read_lock_bh _xmit_ETHER#2 rcu_read_lock &local->handle_wake_tx_queue_lock &fq->lock tk_core.seq.seqcount irq_context: 0 (wq_completion)ipv6_addrconf (work_completion)(&(&ifa->dad_work)->work) rtnl_mutex &idev->mc_lock rcu_read_lock rcu_read_lock rcu_read_lock_bh _xmit_ETHER#2 rcu_read_lock &local->handle_wake_tx_queue_lock tk_core.seq.seqcount irq_context: 0 (wq_completion)ipv6_addrconf (work_completion)(&(&ifa->dad_work)->work) rtnl_mutex &idev->mc_lock rcu_read_lock rcu_read_lock rcu_read_lock_bh _xmit_ETHER#2 rcu_read_lock &local->handle_wake_tx_queue_lock hwsim_radio_lock irq_context: 0 (wq_completion)ipv6_addrconf (work_completion)(&(&ifa->dad_work)->work) rtnl_mutex &idev->mc_lock rcu_read_lock rcu_read_lock rcu_read_lock_bh _xmit_ETHER#2 rcu_read_lock &local->handle_wake_tx_queue_lock hwsim_radio_lock pool_lock#2 irq_context: 0 (wq_completion)ipv6_addrconf (work_completion)(&(&ifa->dad_work)->work) rtnl_mutex &idev->mc_lock rcu_read_lock rcu_read_lock rcu_read_lock_bh _xmit_ETHER#2 rcu_read_lock &local->handle_wake_tx_queue_lock hwsim_radio_lock &list->lock#19 irq_context: 0 (wq_completion)ipv6_addrconf (work_completion)(&(&ifa->dad_work)->work) rtnl_mutex &idev->mc_lock rcu_read_lock rcu_read_lock rcu_read_lock_bh _xmit_ETHER#2 rcu_read_lock &local->handle_wake_tx_queue_lock &list->lock#19 irq_context: 0 cb_lock genl_mutex uevent_sock_mutex rcu_read_lock &ei->socket.wq.wait &ep->lock &ep->wq &p->pi_lock &cfs_rq->removed.lock irq_context: 0 base_sockets.lock irq_context: 0 cb_lock genl_mutex cpu_hotplug_lock wq_pool_mutex &c->lock irq_context: 0 cb_lock genl_mutex cpu_hotplug_lock wq_pool_mutex &____s->seqcount#2 irq_context: 0 cb_lock genl_mutex cpu_hotplug_lock wq_pool_mutex &____s->seqcount irq_context: 0 cb_lock genl_mutex rtnl_mutex &rdev->wiphy.mtx rcu_read_lock &rq->__lock irq_context: 0 cb_lock genl_mutex rtnl_mutex &rdev->wiphy.mtx rcu_read_lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 cb_lock genl_mutex rtnl_mutex &rdev->wiphy.mtx rcu_read_lock &rq->__lock &cfs_rq->removed.lock irq_context: 0 (wq_completion)events (work_completion)(&rdev->wiphy_work) &rdev->wiphy.mtx &wdev->mtx &local->mtx &local->chanctx_mtx &c->lock irq_context: 0 (wq_completion)events (work_completion)(&rdev->wiphy_work) &rdev->wiphy.mtx &wdev->mtx &local->mtx &local->chanctx_mtx &____s->seqcount#2 irq_context: 0 (wq_completion)events (work_completion)(&rdev->wiphy_work) &rdev->wiphy.mtx &wdev->mtx &local->mtx &local->chanctx_mtx &n->list_lock irq_context: 0 (wq_completion)events (work_completion)(&rdev->wiphy_work) &rdev->wiphy.mtx &wdev->mtx &local->mtx &local->chanctx_mtx &n->list_lock &c->lock irq_context: 0 (wq_completion)events (work_completion)(&rdev->wiphy_work) &lock->wait_lock irq_context: 0 (wq_completion)events (work_completion)(&rdev->wiphy_work) &p->pi_lock irq_context: 0 (wq_completion)events (work_completion)(&rdev->wiphy_work) &p->pi_lock &rq->__lock irq_context: 0 (wq_completion)cfg80211 (work_completion)(&rdev->event_work) &rdev->wiphy.mtx &wdev->mtx &rq->__lock irq_context: 0 (wq_completion)events (work_completion)(&rdev->wiphy_work) &p->pi_lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)events (work_completion)(&rdev->wiphy_work) &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)events (work_completion)(&(&nsim_dev->trap_data->trap_report_dw)->work) &devlink->lock_key#3 &rq->__lock irq_context: 0 (wq_completion)events (work_completion)(&(&nsim_dev->trap_data->trap_report_dw)->work) &devlink->lock_key#3 &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)events (work_completion)(&(&nsim_dev->trap_data->trap_report_dw)->work) &devlink->lock_key#2 &rq->__lock irq_context: 0 (wq_completion)events (work_completion)(&(&nsim_dev->trap_data->trap_report_dw)->work) &devlink->lock_key#2 &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)events (work_completion)(&rdev->wiphy_work) &rdev->wiphy.mtx &wdev->mtx &pcp->lock &zone->lock irq_context: 0 cb_lock genl_mutex rfkill_global_mutex uevent_sock_mutex rcu_read_lock &ei->socket.wq.wait &ep->lock &ep->wq &p->pi_lock &cfs_rq->removed.lock irq_context: 0 (wq_completion)events_unbound (work_completion)(&port->bc_work) &rq->__lock irq_context: 0 (wq_completion)events_unbound (work_completion)(&port->bc_work) &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)phy11 irq_context: 0 (wq_completion)phy11 (work_completion)(&local->reconfig_filter) irq_context: 0 (wq_completion)phy11 (work_completion)(&local->reconfig_filter) &local->filter_lock irq_context: 0 (wq_completion)ipv6_addrconf (work_completion)(&(&ifa->dad_work)->work) rtnl_mutex rcu_read_lock rcu_read_lock rcu_read_lock_bh _xmit_ETHER#2 rcu_read_lock &local->handle_wake_tx_queue_lock hwsim_radio_lock &n->list_lock irq_context: 0 (wq_completion)ipv6_addrconf (work_completion)(&(&ifa->dad_work)->work) rtnl_mutex rcu_read_lock rcu_read_lock rcu_read_lock_bh _xmit_ETHER#2 rcu_read_lock &local->handle_wake_tx_queue_lock hwsim_radio_lock &n->list_lock &c->lock irq_context: softirq (&ndev->rs_timer) rcu_read_lock rcu_read_lock rcu_read_lock_bh rcu_read_lock rcu_read_lock_bh rcu_read_lock tk_core.seq.seqcount irq_context: softirq (&ndev->rs_timer) rcu_read_lock rcu_read_lock rcu_read_lock_bh rcu_read_lock rcu_read_lock_bh rcu_read_lock &list->lock#5 irq_context: 0 (wq_completion)events (work_completion)(&(&nsim_dev->trap_data->trap_report_dw)->work) &devlink->lock_key#5 &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 cb_lock genl_mutex rtnl_mutex &rdev->wiphy.mtx &pcp->lock &zone->lock &____s->seqcount irq_context: 0 &mm->mmap_lock rcu_read_lock rcu_read_lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 cb_lock genl_mutex &cfs_rq->removed.lock irq_context: 0 cb_lock genl_mutex &sb->s_type->i_mutex_key#3 fs_reclaim &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 cb_lock rtnl_mutex &rdev->wiphy.mtx &sb->s_type->i_mutex_key#3 fill_pool_map-wait-type-override &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)wg-crypt-wg0#6 (work_completion)(&peer->transmit_packet_work) rcu_read_lock_bh &base->lock irq_context: 0 (wq_completion)wg-crypt-wg0#6 (work_completion)(&peer->transmit_packet_work) rcu_read_lock_bh &base->lock &obj_hash[i].lock irq_context: 0 (wq_completion)wg-crypt-wg0#6 (work_completion)(&peer->transmit_packet_work) batched_entropy_u8.lock irq_context: 0 (wq_completion)mld (work_completion)(&(&idev->mc_dad_work)->work) &idev->mc_lock rcu_read_lock rcu_read_lock rcu_read_lock_bh rcu_read_lock rcu_read_lock pool_lock#2 irq_context: 0 (wq_completion)mld (work_completion)(&(&idev->mc_dad_work)->work) &idev->mc_lock rcu_read_lock rcu_read_lock rcu_read_lock_bh rcu_read_lock rcu_read_lock &dir->lock#2 irq_context: 0 (wq_completion)mld (work_completion)(&(&idev->mc_dad_work)->work) &idev->mc_lock rcu_read_lock rcu_read_lock rcu_read_lock_bh rcu_read_lock rcu_read_lock &ul->lock irq_context: 0 (wq_completion)mld (work_completion)(&(&idev->mc_dad_work)->work) &idev->mc_lock rcu_read_lock rcu_read_lock rcu_read_lock_bh rcu_read_lock fill_pool_map-wait-type-override pool_lock#2 irq_context: 0 cb_lock rtnl_mutex &rdev->wiphy.mtx fill_pool_map-wait-type-override pool_lock#2 irq_context: 0 (wq_completion)mld (work_completion)(&(&idev->mc_dad_work)->work) &idev->mc_lock rcu_read_lock rcu_read_lock rcu_read_lock_bh rcu_read_lock fill_pool_map-wait-type-override pool_lock irq_context: 0 (wq_completion)mld (work_completion)(&(&idev->mc_dad_work)->work) &idev->mc_lock rcu_read_lock rcu_read_lock rcu_read_lock_bh rcu_read_lock &obj_hash[i].lock irq_context: 0 cb_lock rtnl_mutex &rdev->wiphy.mtx fill_pool_map-wait-type-override pool_lock irq_context: 0 (wq_completion)mld (work_completion)(&(&idev->mc_dad_work)->work) &idev->mc_lock rcu_read_lock rcu_read_lock rcu_read_lock_bh rcu_read_lock batched_entropy_u32.lock irq_context: 0 (wq_completion)mld (work_completion)(&(&idev->mc_dad_work)->work) &idev->mc_lock rcu_read_lock rcu_read_lock rcu_read_lock_bh rcu_read_lock rcu_read_lock &____s->seqcount#7 irq_context: 0 (wq_completion)mld (work_completion)(&(&idev->mc_dad_work)->work) &idev->mc_lock rcu_read_lock rcu_read_lock rcu_read_lock_bh rcu_read_lock rcu_read_lock rcu_read_lock &____s->seqcount#7 irq_context: 0 (wq_completion)mld (work_completion)(&(&idev->mc_dad_work)->work) &idev->mc_lock rcu_read_lock rcu_read_lock rcu_read_lock_bh rcu_read_lock rcu_read_lock &nf_nat_locks[i] irq_context: 0 (wq_completion)mld (work_completion)(&(&idev->mc_dad_work)->work) &idev->mc_lock rcu_read_lock rcu_read_lock rcu_read_lock_bh rcu_read_lock rcu_read_lock &nf_conntrack_locks[i] irq_context: 0 (wq_completion)mld (work_completion)(&(&idev->mc_dad_work)->work) &idev->mc_lock rcu_read_lock rcu_read_lock rcu_read_lock_bh rcu_read_lock rcu_read_lock &nf_conntrack_locks[i] &nf_conntrack_locks[i]/1 irq_context: 0 (wq_completion)mld (work_completion)(&(&idev->mc_dad_work)->work) &idev->mc_lock rcu_read_lock rcu_read_lock rcu_read_lock_bh rcu_read_lock rcu_read_lock &nf_conntrack_locks[i] &nf_conntrack_locks[i]/1 batched_entropy_u8.lock irq_context: 0 (wq_completion)mld (work_completion)(&(&idev->mc_dad_work)->work) &idev->mc_lock rcu_read_lock rcu_read_lock rcu_read_lock_bh rcu_read_lock rcu_read_lock &nf_conntrack_locks[i]/1 irq_context: 0 (wq_completion)mld (work_completion)(&(&idev->mc_dad_work)->work) &idev->mc_lock rcu_read_lock rcu_read_lock rcu_read_lock_bh rcu_read_lock rcu_read_lock &____s->seqcount#10 irq_context: 0 (wq_completion)mld (work_completion)(&(&idev->mc_dad_work)->work) &idev->mc_lock rcu_read_lock rcu_read_lock rcu_read_lock_bh rcu_read_lock rcu_read_lock rcu_read_lock_bh tk_core.seq.seqcount irq_context: 0 (wq_completion)mld (work_completion)(&(&idev->mc_dad_work)->work) &idev->mc_lock rcu_read_lock rcu_read_lock rcu_read_lock_bh rcu_read_lock rcu_read_lock rcu_read_lock_bh &list->lock#5 irq_context: 0 (wq_completion)wg-crypt-wg0#5 (work_completion)(&peer->transmit_packet_work) rcu_read_lock_bh &base->lock irq_context: 0 (wq_completion)wg-crypt-wg0#5 (work_completion)(&peer->transmit_packet_work) rcu_read_lock_bh &base->lock &obj_hash[i].lock irq_context: 0 cb_lock rtnl_mutex &rdev->wiphy.mtx &sb->s_type->i_mutex_key#3 mmu_notifier_invalidate_range_start &rq->__lock irq_context: 0 cb_lock rtnl_mutex &rdev->wiphy.mtx &sb->s_type->i_mutex_key#3 mmu_notifier_invalidate_range_start &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)wg-crypt-wg0#5 (work_completion)(&peer->transmit_packet_work) &rq->__lock irq_context: softirq rcu_read_lock &table->lock#3 irq_context: softirq rcu_read_lock rcu_read_lock &table->lock#3 irq_context: softirq rcu_read_lock rcu_read_lock rcu_read_lock &table->lock#3 irq_context: 0 (wq_completion)wg-crypt-wg0#5 (work_completion)(&peer->transmit_packet_work) batched_entropy_u8.lock irq_context: 0 (wq_completion)cfg80211 (work_completion)(&rdev->event_work) &rdev->wiphy.mtx &lock->wait_lock irq_context: 0 (wq_completion)cfg80211 (work_completion)(&rdev->event_work) &rdev->wiphy.mtx &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 cb_lock rtnl_mutex &rdev->wiphy.mtx &pcp->lock &zone->lock &____s->seqcount irq_context: 0 cb_lock genl_mutex rtnl_mutex &rdev->wiphy.mtx gdp_mutex &root->kernfs_rwsem &root->kernfs_iattr_rwsem &sem->wait_lock irq_context: 0 cb_lock genl_mutex rtnl_mutex &rdev->wiphy.mtx gdp_mutex &root->kernfs_rwsem &root->kernfs_iattr_rwsem &rq->__lock irq_context: 0 cb_lock genl_mutex rtnl_mutex &rdev->wiphy.mtx gdp_mutex &root->kernfs_rwsem &root->kernfs_iattr_rwsem &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)wg-crypt-wg2#3 (work_completion)(&peer->transmit_packet_work) rcu_read_lock_bh &base->lock irq_context: 0 (wq_completion)wg-crypt-wg2#3 (work_completion)(&peer->transmit_packet_work) rcu_read_lock_bh &base->lock &obj_hash[i].lock irq_context: 0 &sb->s_type->i_mutex_key#10 pfkey_mutex irq_context: 0 (wq_completion)wg-crypt-wg2#3 (work_completion)(&peer->transmit_packet_work) &base->lock irq_context: 0 (wq_completion)wg-crypt-wg2#3 (work_completion)(&peer->transmit_packet_work) &base->lock &obj_hash[i].lock irq_context: 0 &sb->s_type->i_mutex_key#10 clock-AF_KEY irq_context: 0 (wq_completion)wg-crypt-wg2#3 (work_completion)(&peer->transmit_packet_work) &peer->endpoint_lock rcu_read_lock_bh batched_entropy_u32.lock irq_context: 0 &sb->s_type->i_mutex_key#10 wlock-AF_KEY irq_context: 0 (wq_completion)wg-crypt-wg1#3 (work_completion)(&({ do { const void *__vpp_verify = (typeof((worker) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __asm__ ("" : "=r"(__ptr) : "0"((typeof(*((worker))) *)((worker)))); (typeof((typeof(*((worker))) *)((worker)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); })->work) rcu_node_0 irq_context: 0 (wq_completion)wg-crypt-wg1#3 (work_completion)(&({ do { const void *__vpp_verify = (typeof((worker) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __asm__ ("" : "=r"(__ptr) : "0"((typeof(*((worker))) *)((worker)))); (typeof((typeof(*((worker))) *)((worker)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); })->work) &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 cb_lock genl_mutex rtnl_mutex &rdev->wiphy.mtx gdp_mutex &sem->wait_lock irq_context: 0 cb_lock genl_mutex rtnl_mutex &rdev->wiphy.mtx gdp_mutex &p->pi_lock irq_context: 0 cb_lock genl_mutex rtnl_mutex &rdev->wiphy.mtx gdp_mutex &p->pi_lock &rq->__lock irq_context: 0 cb_lock genl_mutex rtnl_mutex &rdev->wiphy.mtx gdp_mutex &p->pi_lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)wg-crypt-wg2#3 (work_completion)(&peer->transmit_packet_work) batched_entropy_u8.lock irq_context: 0 cb_lock genl_mutex rtnl_mutex &rdev->wiphy.mtx gdp_mutex &root->kernfs_rwsem &sem->wait_lock irq_context: 0 cb_lock genl_mutex rtnl_mutex &rdev->wiphy.mtx gdp_mutex &root->kernfs_rwsem &rq->__lock irq_context: 0 cb_lock genl_mutex rtnl_mutex &rdev->wiphy.mtx gdp_mutex &root->kernfs_rwsem &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 &sb->s_type->i_mutex_key#10 rlock-AF_KEY irq_context: 0 &sb->s_type->i_mutex_key#10 &net->ipv4.ra_mutex irq_context: 0 &sb->s_type->i_mutex_key#10 &hashinfo->lock irq_context: 0 &sighand->siglock &p->pi_lock &cfs_rq->removed.lock irq_context: softirq &(&ifa->dad_work)->timer rcu_read_lock &pool->lock &p->pi_lock &rq->__lock &cfs_rq->removed.lock irq_context: 0 cb_lock genl_mutex rtnl_mutex &rdev->wiphy.mtx gdp_mutex &root->kernfs_rwsem &cfs_rq->removed.lock irq_context: 0 cb_lock genl_mutex rfkill_global_mutex &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)phy12 irq_context: 0 (wq_completion)phy12 (work_completion)(&local->reconfig_filter) irq_context: 0 (wq_completion)phy12 (work_completion)(&local->reconfig_filter) &local->filter_lock irq_context: 0 cb_lock genl_mutex rfkill_global_mutex rcu_read_lock &rq->__lock irq_context: 0 cb_lock genl_mutex rfkill_global_mutex uevent_sock_mutex &rq->__lock irq_context: 0 cb_lock genl_mutex rfkill_global_mutex triggers_list_lock &rq->__lock irq_context: 0 &f->f_pos_lock &type->i_mutex_dir_key#3 fs_reclaim &rq->__lock irq_context: 0 &f->f_pos_lock &type->i_mutex_dir_key#3 fs_reclaim &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 &f->f_pos_lock &type->i_mutex_dir_key#3 sb_writers#4 jbd2_handle &rq->__lock irq_context: 0 &f->f_pos_lock &type->i_mutex_dir_key#3 sb_writers#4 jbd2_handle &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: softirq (&ndev->rs_timer) rcu_read_lock rcu_read_lock rcu_read_lock_bh rcu_read_lock tk_core.seq.seqcount irq_context: softirq (&ndev->rs_timer) rcu_read_lock rcu_read_lock rcu_read_lock_bh rcu_read_lock &list->lock#5 irq_context: softirq (&ndev->rs_timer) rcu_read_lock rcu_read_lock rcu_read_lock_bh dev->qdisc_tx_busylock ?: &qdisc_tx_busylock rcu_read_lock &ndev->lock irq_context: softirq (&ndev->rs_timer) rcu_read_lock rcu_read_lock rcu_read_lock_bh dev->qdisc_tx_busylock ?: &qdisc_tx_busylock &obj_hash[i].lock irq_context: softirq (&ndev->rs_timer) rcu_read_lock rcu_read_lock rcu_read_lock_bh dev->qdisc_tx_busylock ?: &qdisc_tx_busylock pool_lock#2 irq_context: 0 &sb->s_type->i_mutex_key#9 &____s->seqcount#2 irq_context: 0 (wq_completion)events (work_completion)(&aux->work) stock_lock irq_context: 0 (wq_completion)events (work_completion)(&aux->work) pcpu_lock stock_lock irq_context: 0 (wq_completion)events (work_completion)(&aux->work) purge_vmap_area_lock &obj_hash[i].lock irq_context: 0 (wq_completion)events (work_completion)(&aux->work) purge_vmap_area_lock pool_lock#2 irq_context: 0 sb_writers#4 &type->i_mutex_dir_key#3/1 jbd2_handle &ei->i_data_sem rcu_read_lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 &fsnotify_mark_srcu &group->notification_waitq &ep->lock &ep->wq &p->pi_lock &cfs_rq->removed.lock irq_context: 0 cb_lock genl_mutex rfkill_global_mutex fill_pool_map-wait-type-override pool_lock#2 irq_context: 0 cb_lock genl_mutex rfkill_global_mutex fill_pool_map-wait-type-override &c->lock irq_context: 0 cb_lock genl_mutex rfkill_global_mutex fill_pool_map-wait-type-override &____s->seqcount#2 irq_context: 0 cb_lock genl_mutex rfkill_global_mutex fill_pool_map-wait-type-override &____s->seqcount irq_context: 0 cb_lock genl_mutex rfkill_global_mutex fill_pool_map-wait-type-override pool_lock irq_context: 0 &type->s_umount_key#49/1 &c->lock irq_context: 0 &type->s_umount_key#49/1 &____s->seqcount#2 irq_context: 0 &type->s_umount_key#49/1 &____s->seqcount irq_context: 0 sb_writers#10 &type->i_mutex_dir_key#6/1 cgroup_mutex lock kernfs_idr_lock &c->lock irq_context: 0 sb_writers#10 &type->i_mutex_dir_key#6/1 cgroup_mutex lock kernfs_idr_lock &____s->seqcount#2 irq_context: 0 &xt[i].mutex rcu_read_lock &rcu_state.gp_wq irq_context: 0 &xt[i].mutex rcu_read_lock &rcu_state.gp_wq &p->pi_lock irq_context: 0 &xt[i].mutex rcu_read_lock &rcu_state.gp_wq &p->pi_lock &rq->__lock irq_context: 0 &xt[i].mutex rcu_read_lock &rcu_state.gp_wq &p->pi_lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)phy13 irq_context: 0 (wq_completion)phy13 (work_completion)(&local->reconfig_filter) irq_context: 0 (wq_completion)phy13 (work_completion)(&local->reconfig_filter) &local->filter_lock irq_context: 0 kn->active#54 &____s->seqcount#2 irq_context: 0 kn->active#54 &____s->seqcount irq_context: 0 (wq_completion)events (work_completion)(&rfkill_global_led_trigger_work) rfkill_global_mutex &pool->lock &p->pi_lock irq_context: 0 (wq_completion)events (work_completion)(&rfkill_global_led_trigger_work) rfkill_global_mutex &pool->lock &p->pi_lock &rq->__lock irq_context: 0 (wq_completion)events (work_completion)(&rfkill_global_led_trigger_work) rfkill_global_mutex &pool->lock &p->pi_lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 cb_lock rtnl_mutex &rdev->wiphy.mtx &sb->s_type->i_mutex_key#3 rcu_node_0 irq_context: 0 cb_lock rtnl_mutex &rdev->wiphy.mtx &sb->s_type->i_mutex_key#3 &rcu_state.expedited_wq irq_context: 0 cb_lock rtnl_mutex &rdev->wiphy.mtx &sb->s_type->i_mutex_key#3 &rcu_state.expedited_wq &p->pi_lock irq_context: 0 cb_lock rtnl_mutex &rdev->wiphy.mtx &sb->s_type->i_mutex_key#3 &rcu_state.expedited_wq &p->pi_lock &rq->__lock irq_context: 0 cb_lock rtnl_mutex &rdev->wiphy.mtx &sb->s_type->i_mutex_key#3 &rcu_state.expedited_wq &p->pi_lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)mld (work_completion)(&(&idev->mc_dad_work)->work) &idev->mc_lock rcu_read_lock rcu_read_lock rcu_read_lock_bh rcu_read_lock_bh &obj_hash[i].lock pool_lock irq_context: 0 sb_writers#10 &of->mutex cgroup_mutex cpu_hotplug_lock cgroup_threadgroup_rwsem fill_pool_map-wait-type-override pool_lock#2 irq_context: 0 sb_writers#10 &of->mutex cgroup_mutex cpu_hotplug_lock cgroup_threadgroup_rwsem fill_pool_map-wait-type-override &c->lock irq_context: 0 sb_writers#10 &of->mutex cgroup_mutex cpu_hotplug_lock cgroup_threadgroup_rwsem fill_pool_map-wait-type-override &____s->seqcount#2 irq_context: 0 sb_writers#10 &of->mutex cgroup_mutex cpu_hotplug_lock cgroup_threadgroup_rwsem fill_pool_map-wait-type-override &____s->seqcount irq_context: 0 sb_writers#10 &of->mutex cgroup_mutex cpu_hotplug_lock cgroup_threadgroup_rwsem fill_pool_map-wait-type-override pool_lock irq_context: 0 (wq_completion)phy14 irq_context: 0 (wq_completion)phy14 (work_completion)(&local->reconfig_filter) irq_context: 0 (wq_completion)phy14 (work_completion)(&local->reconfig_filter) &local->filter_lock irq_context: 0 &sb->s_type->i_mutex_key#10 (console_sem).lock irq_context: 0 &sb->s_type->i_mutex_key#10 base_sockets.lock irq_context: 0 &sb->s_type->i_mutex_key#10 clock-AF_ISDN irq_context: 0 &sb->s_type->i_mutex_key#10 clock-AF_ROSE irq_context: 0 &sb->s_type->i_mutex_key#10 sk_lock-AF_ROSE irq_context: 0 &sb->s_type->i_mutex_key#10 sk_lock-AF_ROSE slock-AF_ROSE irq_context: 0 &sb->s_type->i_mutex_key#10 slock-AF_ROSE irq_context: 0 &sb->s_type->i_mutex_key#10 wlock-AF_ROSE irq_context: 0 &sb->s_type->i_mutex_key#10 &list->lock#20 irq_context: 0 &sb->s_type->i_mutex_key#10 sk_lock-AF_ROSE rose_list_lock irq_context: 0 &sb->s_type->i_mutex_key#10 sk_lock-AF_ROSE &obj_hash[i].lock irq_context: 0 &sb->s_type->i_mutex_key#10 sk_lock-AF_ROSE wlock-AF_ROSE irq_context: 0 &sb->s_type->i_mutex_key#10 sk_lock-AF_ROSE &list->lock#20 irq_context: 0 &sb->s_type->i_mutex_key#10 sk_lock-AF_ROSE rlock-AF_ROSE irq_context: 0 sb_writers#11 &of->mutex cgroup_mutex cpu_hotplug_lock cgroup_threadgroup_rwsem &rq->__lock irq_context: 0 cgroup_threadgroup_rwsem cgroup_threadgroup_rwsem.waiters.lock irq_context: 0 sb_writers#11 &of->mutex cgroup_mutex cpu_hotplug_lock cgroup_threadgroup_rwsem &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 sb_writers#11 &of->mutex cgroup_mutex cpu_hotplug_lock cgroup_threadgroup_rwsem.waiters.lock &p->pi_lock irq_context: 0 sb_writers#11 &of->mutex cgroup_mutex cpu_hotplug_lock cgroup_threadgroup_rwsem.waiters.lock &p->pi_lock &cfs_rq->removed.lock irq_context: 0 sb_writers#11 &of->mutex cgroup_mutex cpu_hotplug_lock cgroup_threadgroup_rwsem.waiters.lock &p->pi_lock &rq->__lock irq_context: 0 sb_writers#11 &of->mutex cgroup_mutex cpu_hotplug_lock cgroup_threadgroup_rwsem.waiters.lock &p->pi_lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 sb_writers#11 &type->i_mutex_dir_key#7/1 cgroup_mutex pcpu_alloc_mutex &rq->__lock irq_context: 0 sb_writers#11 &type->i_mutex_dir_key#7/1 cgroup_mutex pcpu_alloc_mutex &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 ebt_mutex &c->lock irq_context: 0 ebt_mutex &____s->seqcount#2 irq_context: 0 ebt_mutex &____s->seqcount irq_context: 0 nf_nat_proto_mutex fill_pool_map-wait-type-override &rq->__lock irq_context: 0 nf_nat_proto_mutex fill_pool_map-wait-type-override &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 nf_hook_mutex &rq->__lock irq_context: 0 nf_hook_mutex &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)events (work_completion)(&rdev->wiphy_work) &rdev->wiphy.mtx &wdev->mtx &pcp->lock &zone->lock &____s->seqcount irq_context: 0 (wq_completion)events (work_completion)(&rdev->wiphy_work) &rdev->wiphy.mtx &wdev->mtx &sta->lock batched_entropy_u8.lock irq_context: 0 (wq_completion)events (work_completion)(&rdev->wiphy_work) &rdev->wiphy.mtx &wdev->mtx &sta->lock kfence_freelist_lock irq_context: 0 (wq_completion)events (work_completion)(&rdev->wiphy_work) &rdev->wiphy.mtx &wdev->mtx rcu_read_lock &sta->lock &c->lock irq_context: 0 (wq_completion)events (work_completion)(&rdev->wiphy_work) &rdev->wiphy.mtx &wdev->mtx rcu_read_lock &sta->lock &____s->seqcount#2 irq_context: 0 (wq_completion)events (work_completion)(&rdev->wiphy_work) &rdev->wiphy.mtx &wdev->mtx rcu_read_lock &sta->lock &____s->seqcount irq_context: 0 (wq_completion)events (work_completion)(&rdev->wiphy_work) &rdev->wiphy.mtx &wdev->mtx &rdev->bss_lock &n->list_lock irq_context: 0 (wq_completion)events (work_completion)(&rdev->wiphy_work) &rdev->wiphy.mtx &wdev->mtx rcu_read_lock &sta->lock &obj_hash[i].lock pool_lock irq_context: 0 (wq_completion)mld (work_completion)(&(&idev->mc_dad_work)->work) &idev->mc_lock rcu_read_lock rcu_read_lock rcu_read_lock_bh _xmit_ETHER#2 irq_context: 0 (wq_completion)mld (work_completion)(&(&idev->mc_dad_work)->work) &idev->mc_lock rcu_read_lock rcu_read_lock rcu_read_lock_bh _xmit_ETHER#2 rcu_read_lock tk_core.seq.seqcount irq_context: 0 (wq_completion)mld (work_completion)(&(&idev->mc_dad_work)->work) &idev->mc_lock rcu_read_lock rcu_read_lock rcu_read_lock_bh _xmit_ETHER#2 rcu_read_lock &fq->lock irq_context: 0 (wq_completion)mld (work_completion)(&(&idev->mc_dad_work)->work) &idev->mc_lock rcu_read_lock rcu_read_lock rcu_read_lock_bh _xmit_ETHER#2 rcu_read_lock &local->active_txq_lock[i] irq_context: 0 (wq_completion)mld (work_completion)(&(&idev->mc_dad_work)->work) &idev->mc_lock rcu_read_lock rcu_read_lock rcu_read_lock_bh _xmit_ETHER#2 rcu_read_lock &local->handle_wake_tx_queue_lock irq_context: 0 (wq_completion)mld (work_completion)(&(&idev->mc_dad_work)->work) &idev->mc_lock rcu_read_lock rcu_read_lock rcu_read_lock_bh _xmit_ETHER#2 rcu_read_lock &local->handle_wake_tx_queue_lock &local->active_txq_lock[i] irq_context: 0 (wq_completion)mld (work_completion)(&(&idev->mc_dad_work)->work) &idev->mc_lock rcu_read_lock rcu_read_lock rcu_read_lock_bh _xmit_ETHER#2 rcu_read_lock &local->handle_wake_tx_queue_lock &local->queue_stop_reason_lock irq_context: 0 (wq_completion)mld (work_completion)(&(&idev->mc_dad_work)->work) &idev->mc_lock rcu_read_lock rcu_read_lock rcu_read_lock_bh _xmit_ETHER#2 rcu_read_lock &local->handle_wake_tx_queue_lock &fq->lock irq_context: 0 (wq_completion)mld (work_completion)(&(&idev->mc_dad_work)->work) &idev->mc_lock rcu_read_lock rcu_read_lock rcu_read_lock_bh _xmit_ETHER#2 rcu_read_lock &local->handle_wake_tx_queue_lock &fq->lock tk_core.seq.seqcount irq_context: 0 (wq_completion)mld (work_completion)(&(&idev->mc_dad_work)->work) &idev->mc_lock rcu_read_lock rcu_read_lock rcu_read_lock_bh _xmit_ETHER#2 rcu_read_lock &local->handle_wake_tx_queue_lock tk_core.seq.seqcount irq_context: 0 (wq_completion)mld (work_completion)(&(&idev->mc_dad_work)->work) &idev->mc_lock rcu_read_lock rcu_read_lock rcu_read_lock_bh _xmit_ETHER#2 rcu_read_lock &local->handle_wake_tx_queue_lock hwsim_radio_lock irq_context: 0 (wq_completion)mld (work_completion)(&(&idev->mc_dad_work)->work) &idev->mc_lock rcu_read_lock rcu_read_lock rcu_read_lock_bh _xmit_ETHER#2 rcu_read_lock &local->handle_wake_tx_queue_lock hwsim_radio_lock &c->lock irq_context: 0 (wq_completion)mld (work_completion)(&(&idev->mc_dad_work)->work) &idev->mc_lock rcu_read_lock rcu_read_lock rcu_read_lock_bh _xmit_ETHER#2 rcu_read_lock &local->handle_wake_tx_queue_lock hwsim_radio_lock pool_lock#2 irq_context: 0 (wq_completion)mld (work_completion)(&(&idev->mc_dad_work)->work) &idev->mc_lock rcu_read_lock rcu_read_lock rcu_read_lock_bh _xmit_ETHER#2 rcu_read_lock &local->handle_wake_tx_queue_lock hwsim_radio_lock &list->lock#19 irq_context: 0 (wq_completion)mld (work_completion)(&(&idev->mc_dad_work)->work) &idev->mc_lock rcu_read_lock rcu_read_lock rcu_read_lock_bh _xmit_ETHER#2 rcu_read_lock &local->handle_wake_tx_queue_lock &list->lock#19 irq_context: 0 (wq_completion)mld (work_completion)(&(&idev->mc_dad_work)->work) &idev->mc_lock rcu_read_lock rcu_read_lock rcu_node_0 irq_context: 0 (wq_completion)mld (work_completion)(&(&idev->mc_dad_work)->work) &idev->mc_lock rcu_read_lock rcu_read_lock &rq->__lock irq_context: 0 kn->active#53 &c->lock irq_context: 0 kn->active#53 &____s->seqcount#2 irq_context: 0 kn->active#53 &____s->seqcount irq_context: 0 sb_writers#11 &type->i_mutex_dir_key#7/1 cgroup_mutex remove_cache_srcu irq_context: 0 sb_writers#11 &type->i_mutex_dir_key#7/1 cgroup_mutex remove_cache_srcu quarantine_lock irq_context: 0 sb_writers#11 &type->i_mutex_dir_key#7/1 cgroup_mutex remove_cache_srcu &c->lock irq_context: 0 sb_writers#11 &type->i_mutex_dir_key#7/1 cgroup_mutex remove_cache_srcu &n->list_lock irq_context: 0 sb_writers#11 &type->i_mutex_dir_key#7/1 cgroup_mutex remove_cache_srcu rcu_read_lock pool_lock#2 irq_context: 0 sb_writers#11 &type->i_mutex_dir_key#7/1 cgroup_mutex remove_cache_srcu &obj_hash[i].lock irq_context: 0 &f->f_pos_lock &type->i_mutex_dir_key#4 &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 sb_writers#11 &type->i_mutex_dir_key#7/1 cgroup_mutex remove_cache_srcu &rq->__lock irq_context: 0 sb_writers#11 &type->i_mutex_dir_key#7/1 cgroup_mutex remove_cache_srcu &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)mld (work_completion)(&(&idev->mc_dad_work)->work) &idev->mc_lock rcu_read_lock rcu_node_0 irq_context: 0 (wq_completion)mld (work_completion)(&(&idev->mc_dad_work)->work) &idev->mc_lock rcu_read_lock rcu_read_lock rcu_read_lock_bh dev->qdisc_tx_busylock ?: &qdisc_tx_busylock#2 rcu_read_lock_bh &obj_hash[i].lock irq_context: 0 (wq_completion)mld (work_completion)(&(&idev->mc_dad_work)->work) &idev->mc_lock rcu_read_lock rcu_read_lock rcu_read_lock_bh dev->qdisc_tx_busylock ?: &qdisc_tx_busylock#2 rcu_read_lock_bh pool_lock#2 irq_context: 0 (wq_completion)mld (work_completion)(&(&idev->mc_dad_work)->work) &idev->mc_lock rcu_read_lock rcu_read_lock rcu_read_lock_bh dev->qdisc_tx_busylock ?: &qdisc_tx_busylock#2 rcu_read_lock_bh rcu_read_lock tk_core.seq.seqcount irq_context: 0 (wq_completion)mld (work_completion)(&(&idev->mc_dad_work)->work) &idev->mc_lock rcu_read_lock rcu_read_lock rcu_read_lock_bh dev->qdisc_tx_busylock ?: &qdisc_tx_busylock#2 rcu_read_lock_bh rcu_read_lock &list->lock#5 irq_context: 0 kn->active#57 &c->lock irq_context: 0 kn->active#57 &____s->seqcount#2 irq_context: 0 kn->active#57 &n->list_lock irq_context: 0 kn->active#57 &n->list_lock &c->lock irq_context: 0 sb_writers#11 &type->i_mutex_dir_key#7/1 tomoyo_ss &____s->seqcount#2 irq_context: 0 sb_writers#11 &type->i_mutex_dir_key#7/1 tomoyo_ss &pcp->lock &zone->lock irq_context: 0 sb_writers#11 &type->i_mutex_dir_key#7/1 tomoyo_ss &pcp->lock &zone->lock &____s->seqcount irq_context: 0 sb_writers#10 &type->i_mutex_dir_key#6 &root->kernfs_rwsem &____s->seqcount#2 irq_context: 0 sb_writers#11 &type->i_mutex_dir_key#7/1 tomoyo_ss &____s->seqcount irq_context: 0 (wq_completion)events wireless_nlevent_work net_rwsem &n->list_lock irq_context: 0 (wq_completion)events wireless_nlevent_work net_rwsem &n->list_lock &c->lock irq_context: 0 sb_writers#11 &of->mutex cgroup_mutex cgroup_mutex.wait_lock irq_context: 0 sb_writers#11 &of->mutex cgroup_mutex &rq->__lock irq_context: 0 sb_writers#11 &of->mutex cgroup_mutex &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 sb_writers#10 &of->mutex cgroup_mutex.wait_lock irq_context: 0 sb_writers#10 &of->mutex &p->pi_lock irq_context: 0 sb_writers#10 &of->mutex &p->pi_lock &rq->__lock irq_context: 0 sb_writers#10 &of->mutex &p->pi_lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 sb_writers#11 &type->i_mutex_dir_key#7/1 cgroup_mutex.wait_lock irq_context: 0 sb_writers#11 &type->i_mutex_dir_key#7/1 &p->pi_lock irq_context: 0 sb_writers#11 &of->mutex cgroup_mutex.wait_lock irq_context: 0 sb_writers#11 &of->mutex &p->pi_lock irq_context: 0 kn->active#56 &c->lock irq_context: 0 kn->active#56 &____s->seqcount#2 irq_context: 0 kn->active#56 &____s->seqcount irq_context: 0 kn->active#56 &n->list_lock irq_context: 0 kn->active#56 &n->list_lock &c->lock irq_context: 0 &xt[i].mutex &lock->wait_lock irq_context: 0 nf_nat_proto_mutex nf_hook_mutex remove_cache_srcu irq_context: 0 nf_nat_proto_mutex nf_hook_mutex remove_cache_srcu quarantine_lock irq_context: 0 (wq_completion)events pcpu_balance_work pcpu_alloc_mutex pcpu_alloc_mutex.wait_lock irq_context: 0 (wq_completion)events pcpu_balance_work pcpu_alloc_mutex &pool->lock irq_context: 0 sb_writers#4 &type->i_mutex_dir_key#3/1 tomoyo_ss remove_cache_srcu irq_context: 0 sb_writers#4 &type->i_mutex_dir_key#3/1 tomoyo_ss remove_cache_srcu quarantine_lock irq_context: 0 sb_writers#4 &type->i_mutex_dir_key#3/1 tomoyo_ss remove_cache_srcu &c->lock irq_context: 0 sb_writers#4 &type->i_mutex_dir_key#3/1 tomoyo_ss remove_cache_srcu &n->list_lock irq_context: 0 sb_writers#4 &type->i_mutex_dir_key#3/1 tomoyo_ss remove_cache_srcu rcu_read_lock pool_lock#2 irq_context: 0 sb_writers#4 &type->i_mutex_dir_key#3/1 tomoyo_ss remove_cache_srcu &obj_hash[i].lock irq_context: 0 sb_writers#4 &type->i_mutex_dir_key#3/1 fs_reclaim mmu_notifier_invalidate_range_start &rq->__lock irq_context: 0 sb_writers#4 &type->i_mutex_dir_key#3/1 fs_reclaim mmu_notifier_invalidate_range_start &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 sb_writers#4 &type->i_mutex_dir_key#3/1 jbd2_handle &ei->i_data_sem &____s->seqcount#2 irq_context: 0 sb_writers#4 &type->i_mutex_dir_key#3/1 jbd2_handle &ei->i_data_sem &____s->seqcount irq_context: 0 (wq_completion)mld (work_completion)(&(&idev->mc_ifc_work)->work) &idev->mc_lock rcu_read_lock rcu_read_lock rcu_read_lock_bh _xmit_ETHER#2 rcu_read_lock &local->handle_wake_tx_queue_lock hwsim_radio_lock &pcp->lock &zone->lock &____s->seqcount irq_context: 0 sb_writers#5 &type->i_mutex_dir_key#5/1 &sb->s_type->i_mutex_key#12 &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 sb_writers#8 &____s->seqcount#2 irq_context: 0 sb_writers#8 &____s->seqcount irq_context: 0 sb_writers#5 &type->i_mutex_dir_key#5/1 &sb->s_type->i_mutex_key#12 &sb->s_type->i_mutex_key#12/4 &simple_offset_xa_lock &obj_hash[i].lock irq_context: 0 sb_writers#5 &type->i_mutex_dir_key#5/1 &sb->s_type->i_mutex_key#12 &sb->s_type->i_mutex_key#12/4 &simple_offset_xa_lock pool_lock#2 irq_context: 0 (wq_completion)events_unbound (reaper_work).work &meta->lock irq_context: 0 (wq_completion)events_unbound (reaper_work).work kfence_freelist_lock irq_context: 0 nfnl_subsys_nftables &nft_net->commit_mutex irq_context: 0 &nft_net->commit_mutex irq_context: 0 &nft_net->commit_mutex fs_reclaim irq_context: 0 &nft_net->commit_mutex fs_reclaim mmu_notifier_invalidate_range_start irq_context: 0 &nft_net->commit_mutex pool_lock#2 irq_context: 0 &nft_net->commit_mutex stock_lock irq_context: 0 &nft_net->commit_mutex batched_entropy_u32.lock irq_context: 0 &nft_net->commit_mutex &obj_hash[i].lock irq_context: 0 &nft_net->commit_mutex rcu_state.exp_mutex rcu_node_0 irq_context: 0 &nft_net->commit_mutex rcu_state.exp_mutex &obj_hash[i].lock irq_context: 0 &nft_net->commit_mutex rcu_state.exp_mutex rcu_read_lock &pool->lock irq_context: 0 &nft_net->commit_mutex rcu_state.exp_mutex rcu_read_lock &pool->lock &obj_hash[i].lock irq_context: 0 &nft_net->commit_mutex rcu_state.exp_mutex rcu_read_lock &pool->lock &p->pi_lock irq_context: 0 &nft_net->commit_mutex rcu_state.exp_mutex rcu_read_lock &pool->lock &p->pi_lock &rq->__lock irq_context: 0 &nft_net->commit_mutex rcu_state.exp_mutex rcu_read_lock &pool->lock &p->pi_lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 &nft_net->commit_mutex rcu_state.exp_mutex &rnp->exp_wq[1] irq_context: 0 &nft_net->commit_mutex rcu_state.exp_mutex &rq->__lock irq_context: 0 &nft_net->commit_mutex rcu_state.exp_mutex &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 &mm->mmap_lock rcu_read_lock &rcu_state.expedited_wq &p->pi_lock &rq->__lock irq_context: 0 &mm->mmap_lock rcu_read_lock &rcu_state.expedited_wq &p->pi_lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 &nft_net->commit_mutex (work_completion)(&ht->run_work) irq_context: 0 &nft_net->commit_mutex &ht->mutex irq_context: 0 &nft_net->commit_mutex &ht->mutex &obj_hash[i].lock irq_context: 0 &nft_net->commit_mutex &ht->mutex pool_lock#2 irq_context: 0 l2tp_ip_lock irq_context: 0 &sig->cred_guard_mutex &____s->seqcount#2 irq_context: 0 (wq_completion)bat_events (work_completion)(&(&bat_priv->mcast.work)->work) rcu_read_lock rcu_node_0 irq_context: 0 (wq_completion)bat_events (work_completion)(&(&bat_priv->mcast.work)->work) rcu_read_lock &rq->__lock irq_context: 0 (wq_completion)bat_events (work_completion)(&(&bat_priv->mcast.work)->work) rcu_read_lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)events_unbound (work_completion)(&sub_info->work) &____s->seqcount#2 irq_context: 0 (wq_completion)ipv6_addrconf (work_completion)(&(&ifa->dad_work)->work) rtnl_mutex rcu_read_lock rcu_read_lock &tbl->lock rcu_read_lock &ei->socket.wq.wait &p->pi_lock &cfs_rq->removed.lock irq_context: 0 (wq_completion)ipv6_addrconf (work_completion)(&(&ifa->dad_work)->work) rtnl_mutex rcu_read_lock rcu_read_lock &tbl->lock &n->list_lock irq_context: 0 (wq_completion)ipv6_addrconf (work_completion)(&(&ifa->dad_work)->work) rtnl_mutex rcu_read_lock rcu_read_lock &tbl->lock &n->list_lock &c->lock irq_context: 0 (wq_completion)ipv6_addrconf (work_completion)(&(&ifa->dad_work)->work) rtnl_mutex rcu_read_lock rcu_read_lock &tbl->lock &obj_hash[i].lock pool_lock irq_context: 0 (wq_completion)ipv6_addrconf (work_completion)(&(&ifa->dad_work)->work) rtnl_mutex rcu_read_lock rcu_read_lock &tbl->lock fill_pool_map-wait-type-override pool_lock#2 irq_context: 0 (wq_completion)ipv6_addrconf (work_completion)(&(&ifa->dad_work)->work) rtnl_mutex rcu_read_lock rcu_read_lock &tbl->lock fill_pool_map-wait-type-override &c->lock irq_context: 0 (wq_completion)ipv6_addrconf (work_completion)(&(&ifa->dad_work)->work) rtnl_mutex rcu_read_lock rcu_read_lock &tbl->lock fill_pool_map-wait-type-override &____s->seqcount#2 irq_context: 0 (wq_completion)ipv6_addrconf (work_completion)(&(&ifa->dad_work)->work) rtnl_mutex rcu_read_lock rcu_read_lock &tbl->lock fill_pool_map-wait-type-override &____s->seqcount irq_context: 0 (wq_completion)ipv6_addrconf (work_completion)(&(&ifa->dad_work)->work) rtnl_mutex rcu_read_lock rcu_read_lock &tbl->lock fill_pool_map-wait-type-override pool_lock irq_context: 0 (wq_completion)ipv6_addrconf (work_completion)(&(&ifa->dad_work)->work) rtnl_mutex rcu_read_lock rcu_read_lock &tbl->lock &____s->seqcount#2 irq_context: 0 (wq_completion)bat_events (work_completion)(&(&forw_packet_aggr->delayed_work)->work) &hard_iface->bat_iv.ogm_buff_mutex &pcp->lock &zone->lock irq_context: 0 (wq_completion)bat_events (work_completion)(&(&forw_packet_aggr->delayed_work)->work) &hard_iface->bat_iv.ogm_buff_mutex &pcp->lock &zone->lock &____s->seqcount irq_context: 0 &mm->mmap_lock &mapping->i_mmap_rwsem fill_pool_map-wait-type-override &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 &mm->mmap_lock &mapping->i_mmap_rwsem fill_pool_map-wait-type-override &____s->seqcount#2 irq_context: 0 (wq_completion)ipv6_addrconf (work_completion)(&(&ifa->dad_work)->work) rtnl_mutex rcu_read_lock rcu_read_lock &tbl->lock krc.lock &obj_hash[i].lock irq_context: 0 (wq_completion)ipv6_addrconf (work_completion)(&(&ifa->dad_work)->work) rtnl_mutex rcu_read_lock rcu_read_lock &tbl->lock krc.lock &base->lock irq_context: 0 (wq_completion)ipv6_addrconf (work_completion)(&(&ifa->dad_work)->work) rtnl_mutex rcu_read_lock rcu_read_lock &tbl->lock krc.lock &base->lock &obj_hash[i].lock irq_context: 0 &mm->mmap_lock &____s->seqcount#2 irq_context: 0 rcu_read_lock rcu_read_lock &p->alloc_lock irq_context: 0 &sb->s_type->i_mutex_key#10 (netlink_chain).rwsem &q->instances_lock irq_context: 0 &sb->s_type->i_mutex_key#10 (netlink_chain).rwsem &log->instances_lock irq_context: 0 &sb->s_type->i_mutex_key#10 (netlink_chain).rwsem &nft_net->commit_mutex irq_context: 0 cb_lock genl_mutex rcu_read_lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)events (work_completion)(&(&nsim_dev->trap_data->trap_report_dw)->work) &devlink->lock_key#6 &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)bat_events (work_completion)(&(&bat_priv->nc.work)->work) pool_lock#2 irq_context: softirq (&ndev->rs_timer) rcu_read_lock rcu_read_lock rcu_read_lock_bh dev->qdisc_tx_busylock ?: &qdisc_tx_busylock _xmit_ETHER#2 irq_context: softirq (&ndev->rs_timer) rcu_read_lock rcu_read_lock rcu_read_lock_bh dev->qdisc_tx_busylock ?: &qdisc_tx_busylock _xmit_ETHER#2 rcu_read_lock &ndev->lock irq_context: softirq (&ndev->rs_timer) rcu_read_lock rcu_read_lock rcu_read_lock_bh dev->qdisc_tx_busylock ?: &qdisc_tx_busylock _xmit_ETHER#2 &obj_hash[i].lock irq_context: softirq (&ndev->rs_timer) rcu_read_lock rcu_read_lock rcu_read_lock_bh dev->qdisc_tx_busylock ?: &qdisc_tx_busylock _xmit_ETHER#2 pool_lock#2 irq_context: 0 (wq_completion)bat_events (work_completion)(&(&forw_packet_aggr->delayed_work)->work) &hard_iface->bat_iv.ogm_buff_mutex rcu_read_lock &rcu_state.gp_wq irq_context: 0 (wq_completion)bat_events (work_completion)(&(&forw_packet_aggr->delayed_work)->work) &hard_iface->bat_iv.ogm_buff_mutex rcu_read_lock &rcu_state.gp_wq &p->pi_lock irq_context: 0 (wq_completion)events (work_completion)(&(&krcp->krw_arr[i].rcu_work)->work) rcu_callback quarantine_lock irq_context: 0 (wq_completion)bat_events (work_completion)(&(&forw_packet_aggr->delayed_work)->work) &hard_iface->bat_iv.ogm_buff_mutex rcu_read_lock &rcu_state.gp_wq &p->pi_lock &rq->__lock irq_context: 0 (wq_completion)bat_events (work_completion)(&(&forw_packet_aggr->delayed_work)->work) &hard_iface->bat_iv.ogm_buff_mutex rcu_read_lock &rcu_state.gp_wq &p->pi_lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 sb_writers#4 &type->i_mutex_dir_key#3/1 &sb->s_type->i_mutex_key#8 &ei->i_es_lock key#2 irq_context: 0 sb_writers#4 &type->i_mutex_dir_key#3/1 &type->i_mutex_dir_key#3 jbd2_handle &mapping->private_lock irq_context: 0 &f->f_pos_lock &rq->__lock irq_context: 0 &f->f_pos_lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 &f->f_pos_lock &type->i_mutex_dir_key#3 fs_reclaim mmu_notifier_invalidate_range_start &rq->__lock irq_context: 0 &f->f_pos_lock &type->i_mutex_dir_key#3 fs_reclaim mmu_notifier_invalidate_range_start &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 &f->f_pos_lock &type->i_mutex_dir_key#3 sb_writers#4 &rq->__lock irq_context: 0 &f->f_pos_lock &type->i_mutex_dir_key#3 sb_writers#4 &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)events (work_completion)(&(&krcp->krw_arr[i].rcu_work)->work) rcu_callback &meta->lock irq_context: 0 (wq_completion)events (work_completion)(&(&krcp->krw_arr[i].rcu_work)->work) rcu_callback kfence_freelist_lock irq_context: softirq _xmit_ETHER#2 rcu_read_lock pool_lock#2 irq_context: 0 (wq_completion)bat_events (work_completion)(&(&forw_packet_aggr->delayed_work)->work) &hard_iface->bat_iv.ogm_buff_mutex rcu_read_lock &rcu_state.gp_wq &p->pi_lock &rq->__lock &cfs_rq->removed.lock irq_context: 0 sb_writers#4 &type->i_mutex_dir_key#3/1 tomoyo_ss rcu_read_lock rcu_read_lock rcu_node_0 irq_context: 0 sb_writers#4 &type->i_mutex_dir_key#3/1 tomoyo_ss rcu_read_lock rcu_read_lock &rq->__lock irq_context: 0 sb_writers#4 &type->i_mutex_dir_key#3/1 tomoyo_ss rcu_read_lock rcu_read_lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 sb_writers#4 &type->i_mutex_dir_key#3/1 tomoyo_ss rcu_read_lock rcu_node_0 irq_context: 0 (wq_completion)events (work_completion)(&rdev->wiphy_work) &rdev->wiphy.mtx &wdev->mtx &rdev->bss_lock krc.lock &obj_hash[i].lock irq_context: 0 (wq_completion)events (work_completion)(&rdev->wiphy_work) &rdev->wiphy.mtx &wdev->mtx &rdev->bss_lock krc.lock &base->lock irq_context: 0 sb_writers#4 &type->i_mutex_dir_key#3/1 &sb->s_type->i_mutex_key#8 jbd2_handle &rq->__lock irq_context: 0 (wq_completion)events (work_completion)(&rdev->wiphy_work) &rdev->wiphy.mtx &wdev->mtx &rdev->bss_lock krc.lock &base->lock &obj_hash[i].lock irq_context: 0 sb_writers#4 &type->i_mutex_dir_key#3/1 &sb->s_type->i_mutex_key#8 jbd2_handle &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 sb_writers#4 sb_internal jbd2_handle rcu_read_lock rcu_node_0 irq_context: 0 sb_writers#4 sb_internal jbd2_handle rcu_read_lock &rq->__lock irq_context: 0 sb_writers#4 sb_internal jbd2_handle rcu_read_lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 sb_writers#4 sb_internal jbd2_handle rcu_read_lock &rcu_state.gp_wq irq_context: 0 sb_writers#4 sb_internal jbd2_handle rcu_read_lock &rcu_state.gp_wq &p->pi_lock irq_context: 0 sb_writers#4 sb_internal jbd2_handle rcu_read_lock &rcu_state.gp_wq &p->pi_lock &rq->__lock irq_context: 0 sb_writers#4 sb_internal jbd2_handle rcu_read_lock &rcu_state.gp_wq &p->pi_lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 &pipe->mutex/1 rcu_node_0 irq_context: 0 sb_writers#4 &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 sb_writers#4 &type->i_mutex_dir_key#3/1 jbd2_handle &ei->i_data_sem &ei->i_es_lock &c->lock irq_context: 0 sb_writers#4 &type->i_mutex_dir_key#3/1 jbd2_handle &ei->i_data_sem &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 sb_writers#4 &type->i_mutex_dir_key#3/1 jbd2_handle mmu_notifier_invalidate_range_start &rq->__lock irq_context: 0 sb_writers#4 &type->i_mutex_dir_key#3/1 jbd2_handle mmu_notifier_invalidate_range_start &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 sb_writers#8 tomoyo_ss pool_lock#2 irq_context: softirq (&peer->timer_persistent_keepalive) &c->lock irq_context: 0 (wq_completion)mld (work_completion)(&(&idev->mc_dad_work)->work) &idev->mc_lock rcu_read_lock rcu_read_lock pool_lock#2 irq_context: 0 (wq_completion)mld (work_completion)(&(&idev->mc_dad_work)->work) &idev->mc_lock rcu_read_lock rcu_read_lock &obj_hash[i].lock irq_context: 0 (wq_completion)mld (work_completion)(&(&idev->mc_dad_work)->work) &idev->mc_lock rcu_read_lock rcu_read_lock &dir->lock#2 irq_context: 0 (wq_completion)mld (work_completion)(&(&idev->mc_dad_work)->work) &idev->mc_lock rcu_read_lock rcu_read_lock &tbl->lock irq_context: 0 (wq_completion)mld (work_completion)(&(&idev->mc_dad_work)->work) &idev->mc_lock rcu_read_lock rcu_read_lock &n->lock irq_context: 0 sb_writers#5 rcu_read_lock rcu_node_0 irq_context: 0 sb_writers#5 rcu_read_lock &rq->__lock irq_context: 0 sb_writers#5 rcu_read_lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 &f->f_pos_lock &type->i_mutex_dir_key#4 &root->kernfs_rwsem &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 sb_writers#5 rcu_read_lock &rcu_state.gp_wq irq_context: 0 sb_writers#5 rcu_read_lock &rcu_state.gp_wq &p->pi_lock irq_context: 0 sb_writers#5 rcu_read_lock &rcu_state.gp_wq &p->pi_lock &rq->__lock irq_context: 0 sb_writers#5 rcu_read_lock &rcu_state.gp_wq &p->pi_lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)wg-crypt-wg0#3 (work_completion)(&peer->transmit_packet_work) &rq->__lock irq_context: 0 sb_writers#3 &sb->s_type->i_mutex_key#9 &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 sb_writers#4 &type->i_mutex_dir_key#3/1 remove_cache_srcu &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 &sb->s_type->i_mutex_key#9 rcu_read_lock rcu_node_0 irq_context: 0 &sb->s_type->i_mutex_key#9 rcu_read_lock &rq->__lock irq_context: 0 &sb->s_type->i_mutex_key#9 rcu_read_lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)events (work_completion)(&rdev->wiphy_work) &rdev->wiphy.mtx &wdev->mtx &rdev->bss_lock &obj_hash[i].lock pool_lock irq_context: 0 &sb->s_type->i_mutex_key#9 rcu_node_0 irq_context: 0 &nft_net->commit_mutex &c->lock irq_context: 0 &nft_net->commit_mutex &____s->seqcount#2 irq_context: 0 &nft_net->commit_mutex &n->list_lock irq_context: 0 &nft_net->commit_mutex &n->list_lock &c->lock irq_context: 0 &nft_net->commit_mutex rcu_state.exp_mutex &rnp->exp_wq[2] irq_context: 0 nfnl_subsys_nftables &nft_net->commit_mutex &lock->wait_lock irq_context: 0 nfnl_subsys_nftables &nft_net->commit_mutex &rq->__lock irq_context: 0 nfnl_subsys_nftables &nft_net->commit_mutex &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 &nft_net->commit_mutex &____s->seqcount irq_context: 0 &nft_net->commit_mutex rcu_state.exp_mutex &rnp->exp_wq[3] irq_context: 0 rcu_read_lock &vma->vm_lock->lock rcu_read_lock rcu_node_0 irq_context: 0 rcu_read_lock &vma->vm_lock->lock rcu_read_lock &rq->__lock irq_context: 0 rcu_read_lock &vma->vm_lock->lock rcu_read_lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 sb_writers#8 kn->active#5 &kernfs_locks->open_file_mutex[count] &c->lock irq_context: 0 sb_writers#8 kn->active#5 &kernfs_locks->open_file_mutex[count] &____s->seqcount#2 irq_context: 0 sb_writers#8 kn->active#5 &kernfs_locks->open_file_mutex[count] &rq->__lock irq_context: 0 sb_writers#8 kn->active#5 &kernfs_locks->open_file_mutex[count] &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 sb_writers#8 kn->active#5 &kernfs_locks->open_file_mutex[count] &____s->seqcount irq_context: 0 sb_writers#8 tomoyo_ss &n->list_lock irq_context: 0 sb_writers#8 tomoyo_ss &n->list_lock &c->lock irq_context: 0 sb_writers#4 sb_internal jbd2_handle &rq->__lock irq_context: 0 sb_writers#4 sb_internal jbd2_handle &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 sb_writers#4 &type->i_mutex_dir_key#3/1 sb_internal jbd2_handle &rq->__lock irq_context: 0 sb_writers#4 &type->i_mutex_dir_key#3/1 sb_internal jbd2_handle &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 sb_writers#4 &type->i_mutex_dir_key#3/1 &obj_hash[i].lock pool_lock irq_context: softirq (&ndev->rs_timer) rcu_read_lock rcu_read_lock rcu_read_lock_bh dev->qdisc_tx_busylock ?: &qdisc_tx_busylock &obj_hash[i].lock pool_lock irq_context: 0 (wq_completion)mld (work_completion)(&(&idev->mc_dad_work)->work) &idev->mc_lock rcu_read_lock rcu_read_lock &c->lock irq_context: 0 (wq_completion)mld (work_completion)(&(&idev->mc_dad_work)->work) &idev->mc_lock rcu_read_lock rcu_read_lock &____s->seqcount#2 irq_context: 0 (wq_completion)mld (work_completion)(&(&idev->mc_dad_work)->work) &idev->mc_lock rcu_read_lock rcu_read_lock &____s->seqcount irq_context: 0 (wq_completion)ipv6_addrconf (work_completion)(&(&ifa->dad_work)->work) rtnl_mutex &idev->mc_lock rcu_read_lock rcu_read_lock rcu_read_lock_bh _xmit_ETHER#2 rcu_read_lock &local->handle_wake_tx_queue_lock hwsim_radio_lock &c->lock irq_context: 0 (wq_completion)ipv6_addrconf (work_completion)(&(&ifa->dad_work)->work) rtnl_mutex &idev->mc_lock rcu_read_lock rcu_read_lock rcu_read_lock_bh _xmit_ETHER#2 rcu_read_lock &local->handle_wake_tx_queue_lock hwsim_radio_lock &____s->seqcount#2 irq_context: 0 (wq_completion)ipv6_addrconf (work_completion)(&(&ifa->dad_work)->work) rtnl_mutex &idev->mc_lock rcu_read_lock rcu_read_lock rcu_read_lock_bh _xmit_ETHER#2 rcu_read_lock &local->handle_wake_tx_queue_lock hwsim_radio_lock &____s->seqcount irq_context: 0 sb_writers &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)events (work_completion)(&(&nsim_dev->trap_data->trap_report_dw)->work) &devlink->lock_key#3 &nsim_trap_data->trap_lock batched_entropy_u8.lock irq_context: 0 (wq_completion)events (work_completion)(&(&nsim_dev->trap_data->trap_report_dw)->work) &devlink->lock_key#3 &nsim_trap_data->trap_lock kfence_freelist_lock irq_context: 0 (wq_completion)events (work_completion)(&(&nsim_dev->trap_data->trap_report_dw)->work) &devlink->lock_key#3 &nsim_trap_data->trap_lock &meta->lock irq_context: 0 &u->iolock rcu_read_lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 &u->iolock rcu_read_lock &cfs_rq->removed.lock irq_context: 0 &u->iolock rcu_read_lock &obj_hash[i].lock irq_context: 0 cb_lock genl_mutex rcu_read_lock &rcu_state.gp_wq irq_context: 0 cb_lock genl_mutex rcu_read_lock &rcu_state.gp_wq &p->pi_lock irq_context: 0 cb_lock genl_mutex rcu_read_lock &rcu_state.gp_wq &p->pi_lock &rq->__lock irq_context: 0 cb_lock genl_mutex rcu_read_lock &rcu_state.gp_wq &p->pi_lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 &mm->mmap_lock &mapping->i_mmap_rwsem &anon_vma->rwsem rcu_read_lock rcu_node_0 irq_context: 0 &mm->mmap_lock &mapping->i_mmap_rwsem &anon_vma->rwsem rcu_read_lock &rq->__lock irq_context: 0 &mm->mmap_lock &mapping->i_mmap_rwsem &anon_vma->rwsem rcu_read_lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 sb_writers#4 &type->i_mutex_dir_key#3/1 tomoyo_ss &rq->__lock &cfs_rq->removed.lock irq_context: 0 sb_writers#4 &type->i_mutex_dir_key#3/1 &sb->s_type->i_mutex_key#8 jbd2_handle rcu_read_lock rcu_node_0 irq_context: 0 sb_writers#4 &type->i_mutex_dir_key#3/1 &sb->s_type->i_mutex_key#8 jbd2_handle rcu_read_lock &rq->__lock irq_context: 0 sb_writers#4 &type->i_mutex_dir_key#3/1 &sb->s_type->i_mutex_key#8 jbd2_handle rcu_read_lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 sb_writers#4 &type->i_mutex_dir_key#3/1 &type->i_mutex_dir_key#3 &rq->__lock irq_context: 0 sb_writers#4 &type->i_mutex_dir_key#3/1 &type->i_mutex_dir_key#3 &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 &xt[i].mutex &mm->mmap_lock &rq->__lock irq_context: 0 &xt[i].mutex &mm->mmap_lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 &xt[i].mutex &cfs_rq->removed.lock irq_context: 0 sb_writers#4 sb_internal &rq->__lock irq_context: 0 sb_writers#4 sb_internal &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 &mm->mmap_lock rcu_read_lock stock_lock irq_context: 0 &mm->mmap_lock rcu_read_lock pcpu_lock stock_lock irq_context: softirq (&ndev->rs_timer) rcu_read_lock rcu_read_lock &c->lock irq_context: softirq (&ndev->rs_timer) rcu_read_lock rcu_read_lock rcu_read_lock_bh rcu_read_lock pool_lock#2 irq_context: softirq (&ndev->rs_timer) rcu_read_lock rcu_read_lock rcu_read_lock_bh rcu_read_lock rcu_read_lock &table->lock#3 irq_context: softirq (&ndev->rs_timer) rcu_read_lock rcu_read_lock rcu_read_lock_bh rcu_read_lock rcu_read_lock_bh &obj_hash[i].lock irq_context: softirq (&ndev->rs_timer) rcu_read_lock rcu_read_lock rcu_read_lock_bh rcu_read_lock rcu_read_lock_bh pool_lock#2 irq_context: 0 &f->f_pos_lock &type->i_mutex_dir_key#3 sb_writers#4 &c->lock irq_context: 0 sb_writers#4 &type->i_mutex_dir_key#3/1 tomoyo_ss mmu_notifier_invalidate_range_start &rq->__lock irq_context: 0 sb_writers#4 &type->i_mutex_dir_key#3/1 tomoyo_ss mmu_notifier_invalidate_range_start &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)ipv6_addrconf (work_completion)(&(&ifa->dad_work)->work) rtnl_mutex rcu_read_lock rcu_read_lock &n->list_lock irq_context: 0 (wq_completion)ipv6_addrconf (work_completion)(&(&ifa->dad_work)->work) rtnl_mutex rcu_read_lock rcu_read_lock &n->list_lock &c->lock irq_context: 0 (wq_completion)events_unbound (work_completion)(&sub_info->work) tasklist_lock &sighand->siglock &(&sig->stats_lock)->lock &____s->seqcount#5 &obj_hash[i].lock pool_lock irq_context: 0 (wq_completion)ipv6_addrconf (work_completion)(&(&ifa->dad_work)->work) rtnl_mutex &idev->mc_lock rcu_read_lock rcu_read_lock rcu_read_lock_bh rcu_read_lock rcu_read_lock &nf_conntrack_locks[i]/1 irq_context: 0 (wq_completion)ipv6_addrconf (work_completion)(&(&ifa->dad_work)->work) rtnl_mutex rcu_read_lock rcu_read_lock rcu_read_lock_bh rcu_read_lock &obj_hash[i].lock pool_lock irq_context: softirq rcu_callback per_cpu_ptr(&cgroup_rstat_cpu_lock, cpu) irq_context: 0 &sig->cred_guard_mutex tomoyo_ss remove_cache_srcu &rq->__lock irq_context: 0 &sig->cred_guard_mutex tomoyo_ss remove_cache_srcu &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 &sig->cred_guard_mutex fs_reclaim &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 sb_writers#4 &type->i_mutex_dir_key#3/1 per_cpu_ptr(&cgroup_rstat_cpu_lock, cpu) irq_context: 0 sb_writers#4 &type->i_mutex_dir_key#3/1 jbd2_handle &ei->i_data_sem mmu_notifier_invalidate_range_start &rq->__lock irq_context: 0 sb_writers#4 &type->i_mutex_dir_key#3/1 jbd2_handle &ei->i_data_sem mmu_notifier_invalidate_range_start &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 ebt_mutex &rq->__lock irq_context: 0 ebt_mutex &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 sb_writers#8 kn->active#5 &kernfs_locks->open_file_mutex[count] &n->list_lock irq_context: 0 (wq_completion)bat_events (work_completion)(&(&forw_packet_aggr->delayed_work)->work) &hard_iface->bat_iv.ogm_buff_mutex rcu_read_lock &____s->seqcount#2 irq_context: 0 (wq_completion)bat_events (work_completion)(&(&forw_packet_aggr->delayed_work)->work) &hard_iface->bat_iv.ogm_buff_mutex rcu_read_lock &____s->seqcount irq_context: 0 (wq_completion)mld (work_completion)(&(&idev->mc_dad_work)->work) &idev->mc_lock rcu_read_lock rcu_read_lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 sb_writers#5 &type->i_mutex_dir_key#5 &simple_offset_xa_lock &____s->seqcount#2 irq_context: 0 sb_writers#5 &type->i_mutex_dir_key#5 &simple_offset_xa_lock &____s->seqcount irq_context: 0 (wq_completion)ipv6_addrconf (work_completion)(&(&ifa->dad_work)->work) rtnl_mutex &idev->mc_lock rcu_read_lock rcu_read_lock rcu_read_lock_bh rcu_read_lock batched_entropy_u32.lock crngs.lock irq_context: 0 sb_writers#5 &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 &group->mark_mutex remove_cache_srcu irq_context: 0 &group->mark_mutex remove_cache_srcu quarantine_lock irq_context: 0 &group->mark_mutex remove_cache_srcu &c->lock irq_context: 0 &group->mark_mutex remove_cache_srcu &n->list_lock irq_context: 0 &group->mark_mutex remove_cache_srcu rcu_read_lock pool_lock#2 irq_context: 0 &group->mark_mutex remove_cache_srcu &obj_hash[i].lock irq_context: 0 &group->mark_mutex remove_cache_srcu &rq->__lock irq_context: 0 &group->mark_mutex remove_cache_srcu &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)events (work_completion)(&rdev->wiphy_work) &rdev->wiphy.mtx &wdev->mtx &rdev->bss_lock quarantine_lock irq_context: 0 (wq_completion)ipv6_addrconf (work_completion)(&(&ifa->dad_work)->work) rtnl_mutex &tb->tb6_lock batched_entropy_u8.lock irq_context: 0 (wq_completion)ipv6_addrconf (work_completion)(&(&ifa->dad_work)->work) rtnl_mutex &tb->tb6_lock kfence_freelist_lock irq_context: 0 (wq_completion)ipv6_addrconf (work_completion)(&(&ifa->dad_work)->work) rtnl_mutex &tb->tb6_lock &meta->lock irq_context: 0 &sb->s_type->i_mutex_key#10 per_cpu_ptr(&cgroup_rstat_cpu_lock, cpu) irq_context: 0 sb_writers#4 &type->i_mutex_dir_key#3/1 &type->i_mutex_dir_key#3 &obj_hash[i].lock pool_lock irq_context: 0 sb_writers#4 &type->i_mutex_dir_key#3/1 jbd2_handle &ei->i_data_sem &ei->i_es_lock &c->lock irq_context: 0 sb_writers#8 &of->mutex kn->active#5 uevent_sock_mutex remove_cache_srcu &rq->__lock irq_context: 0 sb_writers#8 &of->mutex kn->active#5 uevent_sock_mutex remove_cache_srcu &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)mld (work_completion)(&(&idev->mc_dad_work)->work) &idev->mc_lock rcu_read_lock &n->list_lock irq_context: 0 (wq_completion)mld (work_completion)(&(&idev->mc_dad_work)->work) &idev->mc_lock rcu_read_lock &n->list_lock &c->lock irq_context: softirq (&ndev->rs_timer) rcu_read_lock rcu_read_lock rcu_read_lock_bh &list->lock#17 irq_context: softirq (&ndev->rs_timer) rcu_read_lock rcu_read_lock rcu_read_lock_bh tk_core.seq.seqcount irq_context: softirq (&ndev->rs_timer) rcu_read_lock rcu_read_lock rcu_read_lock_bh rcu_read_lock_bh &r->producer_lock#2 irq_context: softirq (&ndev->rs_timer) rcu_read_lock rcu_read_lock rcu_read_lock_bh rcu_read_lock_bh rcu_read_lock &pool->lock irq_context: softirq (&ndev->rs_timer) rcu_read_lock rcu_read_lock rcu_read_lock_bh rcu_read_lock_bh rcu_read_lock &pool->lock &obj_hash[i].lock irq_context: softirq (&ndev->rs_timer) rcu_read_lock rcu_read_lock rcu_read_lock_bh rcu_read_lock_bh rcu_read_lock &pool->lock &p->pi_lock irq_context: 0 (wq_completion)wg-crypt-wg1 (work_completion)(&peer->transmit_packet_work) &base->lock irq_context: 0 (wq_completion)wg-crypt-wg1 (work_completion)(&peer->transmit_packet_work) &base->lock &obj_hash[i].lock irq_context: 0 (wq_completion)ipv6_addrconf (work_completion)(&(&ifa->dad_work)->work) rtnl_mutex &idev->mc_lock rcu_read_lock rcu_read_lock rcu_read_lock_bh rcu_read_lock rcu_read_lock &c->lock irq_context: 0 (wq_completion)ipv6_addrconf (work_completion)(&(&ifa->dad_work)->work) rtnl_mutex &idev->mc_lock rcu_read_lock rcu_read_lock rcu_read_lock_bh rcu_read_lock rcu_read_lock &____s->seqcount#2 irq_context: 0 (wq_completion)events (debug_obj_work).work &rq->__lock irq_context: 0 (wq_completion)events (debug_obj_work).work &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 &sb->s_type->i_mutex_key#10 l2tp_ip_lock irq_context: 0 rcu_read_lock per_cpu_ptr(&cgroup_rstat_cpu_lock, cpu) irq_context: 0 (wq_completion)mld (work_completion)(&(&idev->mc_dad_work)->work) &idev->mc_lock rcu_read_lock rcu_read_lock &cfs_rq->removed.lock irq_context: 0 (wq_completion)mld (work_completion)(&(&idev->mc_dad_work)->work) &idev->mc_lock remove_cache_srcu &rq->__lock irq_context: 0 (wq_completion)mld (work_completion)(&(&idev->mc_dad_work)->work) &idev->mc_lock remove_cache_srcu &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 &pipe->mutex/1 &mm->mmap_lock rcu_read_lock per_cpu_ptr(&cgroup_rstat_cpu_lock, cpu) irq_context: 0 (wq_completion)wg-crypt-wg2 (work_completion)(&peer->transmit_packet_work) &peer->endpoint_lock rcu_read_lock_bh &obj_hash[i].lock irq_context: 0 (wq_completion)wg-crypt-wg2 (work_completion)(&peer->transmit_packet_work) &peer->endpoint_lock rcu_read_lock_bh batched_entropy_u32.lock irq_context: 0 sb_writers#4 &type->i_mutex_dir_key#3/1 &sb->s_type->i_mutex_key#8 jbd2_handle rcu_node_0 irq_context: 0 sb_writers#4 &type->i_mutex_dir_key#3/1 &sb->s_type->i_mutex_key#8 &rq->__lock irq_context: 0 sb_writers#4 &type->i_mutex_dir_key#3/1 &sb->s_type->i_mutex_key#8 &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 sb_writers#4 &type->i_mutex_dir_key#3/1 sb_internal jbd2_handle &ei->i_data_sem &rq->__lock irq_context: 0 sb_writers#4 &type->i_mutex_dir_key#3/1 sb_internal jbd2_handle &ei->i_data_sem &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 sb_writers#4 &type->i_mutex_dir_key#3/1 &sb->s_type->i_mutex_key#8 remove_cache_srcu irq_context: 0 sb_writers#4 &type->i_mutex_dir_key#3/1 &sb->s_type->i_mutex_key#8 remove_cache_srcu quarantine_lock irq_context: 0 sb_writers#4 &type->i_mutex_dir_key#3/1 &sb->s_type->i_mutex_key#8 remove_cache_srcu &c->lock irq_context: 0 sb_writers#4 &type->i_mutex_dir_key#3/1 &sb->s_type->i_mutex_key#8 remove_cache_srcu &n->list_lock irq_context: 0 sb_writers#4 &type->i_mutex_dir_key#3/1 &sb->s_type->i_mutex_key#8 remove_cache_srcu rcu_read_lock pool_lock#2 irq_context: 0 sb_writers#4 &type->i_mutex_dir_key#3/1 &sb->s_type->i_mutex_key#8 remove_cache_srcu &obj_hash[i].lock irq_context: 0 sb_writers#4 &type->i_mutex_dir_key#3/1 jbd2_handle &____s->seqcount#2 irq_context: 0 sb_writers#4 &type->i_mutex_dir_key#3/1 jbd2_handle &pcp->lock &zone->lock irq_context: 0 sb_writers#4 &type->i_mutex_dir_key#3/1 jbd2_handle key#4 irq_context: 0 (wq_completion)ipv6_addrconf (work_completion)(&(&ifa->dad_work)->work) rtnl_mutex &idev->mc_lock rcu_read_lock &n->list_lock irq_context: 0 (wq_completion)ipv6_addrconf (work_completion)(&(&ifa->dad_work)->work) rtnl_mutex &idev->mc_lock rcu_read_lock &n->list_lock &c->lock irq_context: 0 sb_writers#8 kn->active#5 &rq->__lock irq_context: 0 &sb->s_type->i_mutex_key#10 (netlink_chain).rwsem rcu_read_lock rcu_node_0 irq_context: 0 &sb->s_type->i_mutex_key#10 (netlink_chain).rwsem rcu_read_lock &rq->__lock irq_context: 0 &sb->s_type->i_mutex_key#10 (netlink_chain).rwsem rcu_read_lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)bat_events (work_completion)(&(&bat_priv->mcast.work)->work) &rq->__lock irq_context: softirq (&ndev->rs_timer) rcu_read_lock rcu_read_lock rcu_read_lock_bh &hsr->seqnr_lock irq_context: softirq (&ndev->rs_timer) rcu_read_lock rcu_read_lock rcu_read_lock_bh &hsr->seqnr_lock rcu_read_lock &new_node->seq_out_lock irq_context: softirq (&ndev->rs_timer) rcu_read_lock rcu_read_lock rcu_read_lock_bh &hsr->seqnr_lock rcu_read_lock pool_lock#2 irq_context: softirq (&ndev->rs_timer) rcu_read_lock rcu_read_lock rcu_read_lock_bh &hsr->seqnr_lock rcu_read_lock rcu_read_lock_bh rcu_read_lock tk_core.seq.seqcount irq_context: softirq (&ndev->rs_timer) rcu_read_lock rcu_read_lock rcu_read_lock_bh &hsr->seqnr_lock rcu_read_lock rcu_read_lock_bh rcu_read_lock &list->lock#5 irq_context: softirq (&ndev->rs_timer) rcu_read_lock rcu_read_lock rcu_read_lock_bh &hsr->seqnr_lock &obj_hash[i].lock irq_context: softirq (&ndev->rs_timer) rcu_read_lock rcu_read_lock rcu_read_lock_bh &hsr->seqnr_lock pool_lock#2 irq_context: 0 (wq_completion)wg-crypt-wg0 (work_completion)(&peer->transmit_packet_work) &peer->endpoint_lock rcu_read_lock_bh rcu_read_lock pool_lock#2 irq_context: 0 (wq_completion)wg-crypt-wg0 (work_completion)(&peer->transmit_packet_work) &peer->endpoint_lock rcu_read_lock_bh rcu_read_lock &obj_hash[i].lock irq_context: 0 (wq_completion)wg-crypt-wg0 (work_completion)(&peer->transmit_packet_work) &peer->endpoint_lock rcu_read_lock_bh rcu_read_lock &tbl->lock irq_context: 0 (wq_completion)wg-crypt-wg0 (work_completion)(&peer->transmit_packet_work) &peer->endpoint_lock rcu_read_lock_bh rcu_read_lock &n->lock irq_context: 0 (wq_completion)wg-crypt-wg0 (work_completion)(&peer->transmit_packet_work) &peer->endpoint_lock rcu_read_lock_bh rcu_read_lock &____s->seqcount#9 irq_context: 0 (wq_completion)events (work_completion)(&rdev->wiphy_work) &rdev->wiphy.mtx &wdev->mtx &rdev->bss_lock &n->list_lock &c->lock irq_context: 0 &mm->mmap_lock &mapping->i_mmap_rwsem &anon_vma->rwsem &cfs_rq->removed.lock irq_context: 0 dup_mmap_sem &mm->mmap_lock &mm->mmap_lock/1 rcu_read_lock rcu_read_lock rcu_node_0 irq_context: 0 dup_mmap_sem &mm->mmap_lock &mm->mmap_lock/1 rcu_read_lock rcu_read_lock &rq->__lock irq_context: 0 dup_mmap_sem &mm->mmap_lock &mm->mmap_lock/1 rcu_read_lock rcu_read_lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)rcu_gp (work_completion)(&(&ssp->srcu_sup->work)->work) &ssp->srcu_sup->srcu_cb_mutex &rq->__lock irq_context: 0 (wq_completion)rcu_gp (work_completion)(&(&ssp->srcu_sup->work)->work) &ssp->srcu_sup->srcu_cb_mutex &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 sb_writers#4 &type->i_mutex_dir_key#3 fs_reclaim irq_context: 0 sb_writers#4 &type->i_mutex_dir_key#3 fs_reclaim mmu_notifier_invalidate_range_start irq_context: 0 sb_writers#4 &type->i_mutex_dir_key#3 stock_lock irq_context: 0 sb_writers#4 &type->i_mutex_dir_key#3 &c->lock irq_context: 0 sb_writers#4 &type->i_mutex_dir_key#3 &dentry->d_lock irq_context: 0 sb_writers#4 &type->i_mutex_dir_key#3 rcu_read_lock rename_lock.seqcount irq_context: 0 sb_writers#4 &type->i_mutex_dir_key#3 &ei->i_es_lock irq_context: 0 sb_writers#4 &type->i_mutex_dir_key#3 &dentry->d_lock &wq#2 irq_context: 0 &iint->mutex &____s->seqcount#2 irq_context: 0 sb_writers#4 &type->i_mutex_dir_key#3/1 jbd2_handle &ei->i_data_sem rcu_node_0 irq_context: 0 sb_writers#4 &sb->s_type->i_mutex_key#8 jbd2_handle &sb->s_type->i_lock_key#22 irq_context: 0 sb_writers#4 &sb->s_type->i_mutex_key#8 jbd2_handle &sb->s_type->i_lock_key#22 &dentry->d_lock irq_context: 0 sb_writers#4 &sb->s_type->i_mutex_key#8 jbd2_handle tk_core.seq.seqcount irq_context: 0 sb_writers#4 &sb->s_type->i_mutex_key#8 &sbi->s_writepages_rwsem irq_context: 0 sb_writers#4 &sb->s_type->i_mutex_key#8 &sbi->s_writepages_rwsem &rsp->gp_wait irq_context: 0 sb_writers#4 &sb->s_type->i_mutex_key#8 &sbi->s_writepages_rwsem rcu_state.exp_mutex rcu_node_0 irq_context: 0 sb_writers#4 &sb->s_type->i_mutex_key#8 &sbi->s_writepages_rwsem rcu_state.exp_mutex &obj_hash[i].lock irq_context: 0 sb_writers#4 &sb->s_type->i_mutex_key#8 &sbi->s_writepages_rwsem rcu_state.exp_mutex rcu_read_lock &pool->lock irq_context: 0 sb_writers#4 &sb->s_type->i_mutex_key#8 &sbi->s_writepages_rwsem rcu_state.exp_mutex rcu_read_lock &pool->lock &obj_hash[i].lock irq_context: 0 sb_writers#4 &sb->s_type->i_mutex_key#8 &sbi->s_writepages_rwsem rcu_state.exp_mutex rcu_read_lock &pool->lock &p->pi_lock irq_context: 0 sb_writers#4 &sb->s_type->i_mutex_key#8 &sbi->s_writepages_rwsem rcu_state.exp_mutex rcu_read_lock &pool->lock &p->pi_lock &rq->__lock irq_context: 0 sb_writers#4 &sb->s_type->i_mutex_key#8 &sbi->s_writepages_rwsem rcu_state.exp_mutex rcu_read_lock &pool->lock &p->pi_lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 sb_writers#4 &sb->s_type->i_mutex_key#8 &sbi->s_writepages_rwsem rcu_state.exp_mutex &rq->__lock irq_context: 0 sb_writers#4 &sb->s_type->i_mutex_key#8 &sbi->s_writepages_rwsem rcu_state.exp_mutex &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 sb_writers#4 &sb->s_type->i_mutex_key#8 &sbi->s_writepages_rwsem rcu_state.exp_mutex &rnp->exp_wq[2] irq_context: 0 dup_mmap_sem &mm->mmap_lock &mm->mmap_lock/1 rcu_read_lock &rcu_state.expedited_wq irq_context: 0 dup_mmap_sem &mm->mmap_lock &mm->mmap_lock/1 rcu_read_lock &rcu_state.expedited_wq &p->pi_lock irq_context: 0 dup_mmap_sem &mm->mmap_lock &mm->mmap_lock/1 rcu_read_lock &rcu_state.expedited_wq &p->pi_lock &rq->__lock irq_context: 0 dup_mmap_sem &mm->mmap_lock &mm->mmap_lock/1 rcu_read_lock &rcu_state.expedited_wq &p->pi_lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 sb_writers#4 &sb->s_type->i_mutex_key#8 &sbi->s_writepages_rwsem &obj_hash[i].lock irq_context: 0 sb_writers#4 &sb->s_type->i_mutex_key#8 &sbi->s_writepages_rwsem mmu_notifier_invalidate_range_start irq_context: 0 sb_writers#4 &sb->s_type->i_mutex_key#8 &sbi->s_writepages_rwsem pool_lock#2 irq_context: 0 sb_writers#4 &sb->s_type->i_mutex_key#8 &sbi->s_writepages_rwsem &journal->j_state_lock irq_context: 0 sb_writers#4 &sb->s_type->i_mutex_key#8 &sbi->s_writepages_rwsem jbd2_handle irq_context: 0 sb_writers#4 &sb->s_type->i_mutex_key#8 &sbi->s_writepages_rwsem jbd2_handle &ei->i_data_sem irq_context: 0 sb_writers#4 &sb->s_type->i_mutex_key#8 &sbi->s_writepages_rwsem jbd2_handle &ei->i_data_sem &ei->i_raw_lock irq_context: 0 sb_writers#4 &sb->s_type->i_mutex_key#8 &sbi->s_writepages_rwsem jbd2_handle &ei->i_data_sem &journal->j_wait_updates irq_context: 0 sb_writers#4 &sb->s_type->i_mutex_key#8 &sbi->s_writepages_rwsem &ei->i_data_sem irq_context: 0 sb_writers#4 &sb->s_type->i_mutex_key#8 &sbi->s_writepages_rwsem &ei->i_data_sem &obj_hash[i].lock irq_context: 0 sb_writers#4 &sb->s_type->i_mutex_key#8 &sbi->s_writepages_rwsem &ei->i_data_sem pool_lock#2 irq_context: 0 sb_writers#4 &sb->s_type->i_mutex_key#8 &sem->waiters irq_context: 0 sb_writers#4 &sb->s_type->i_mutex_key#8 &rsp->gp_wait irq_context: 0 sb_writers#4 &sb->s_type->i_mutex_key#8 &rsp->gp_wait &obj_hash[i].lock irq_context: 0 sb_writers#4 &sb->s_type->i_mutex_key#8 &rsp->gp_wait pool_lock#2 irq_context: 0 &ei->i_prealloc_lock irq_context: 0 &ei->i_es_lock irq_context: 0 inode_hash_lock &sb->s_type->i_lock_key#22 irq_context: 0 integrity_iint_lock irq_context: 0 rtnl_mutex &dev->tx_global_lock irq_context: 0 rtnl_mutex &dev->tx_global_lock _xmit_NONE#2 irq_context: 0 rtnl_mutex &dev->tx_global_lock &obj_hash[i].lock irq_context: 0 rtnl_mutex &sch->q.lock irq_context: 0 rtnl_mutex __ip_vs_mutex irq_context: 0 rtnl_mutex __ip_vs_mutex &ipvs->dest_trash_lock irq_context: softirq (&ndev->rs_timer) rcu_read_lock rcu_read_lock &____s->seqcount#2 irq_context: softirq (&ndev->rs_timer) rcu_read_lock rcu_read_lock &n->list_lock irq_context: softirq (&ndev->rs_timer) rcu_read_lock rcu_read_lock &n->list_lock &c->lock irq_context: softirq rcu_read_lock rcu_read_lock rcu_read_lock rcu_read_lock &ct->lock irq_context: softirq rcu_read_lock rcu_read_lock rcu_read_lock rcu_read_lock &nf_conntrack_locks[i] irq_context: softirq rcu_read_lock rcu_read_lock rcu_read_lock rcu_read_lock &nf_conntrack_locks[i] &nf_conntrack_locks[i]/1 irq_context: softirq rcu_read_lock rcu_read_lock rcu_read_lock rcu_read_lock &nf_conntrack_locks[i]/1 irq_context: softirq rcu_read_lock rcu_read_lock rcu_read_lock rcu_read_lock &____s->seqcount#10 irq_context: softirq rcu_read_lock rcu_read_lock rcu_read_lock rcu_read_lock rcu_read_lock_bh dev->qdisc_tx_busylock ?: &qdisc_tx_busylock _xmit_ETHER#2 irq_context: softirq rcu_read_lock rcu_read_lock rcu_read_lock rcu_read_lock rcu_read_lock_bh dev->qdisc_tx_busylock ?: &qdisc_tx_busylock _xmit_ETHER#2 rcu_read_lock pool_lock#2 irq_context: softirq rcu_read_lock rcu_read_lock rcu_read_lock rcu_read_lock rcu_read_lock_bh dev->qdisc_tx_busylock ?: &qdisc_tx_busylock _xmit_ETHER#2 rcu_read_lock tk_core.seq.seqcount irq_context: softirq rcu_read_lock rcu_read_lock rcu_read_lock rcu_read_lock rcu_read_lock_bh dev->qdisc_tx_busylock ?: &qdisc_tx_busylock _xmit_ETHER#2 rcu_read_lock &obj_hash[i].lock irq_context: softirq _xmit_ETHER#2 rcu_read_lock &obj_hash[i].lock irq_context: 0 rtnl_mutex fib_info_lock &obj_hash[i].lock irq_context: 0 rtnl_mutex fib_info_lock pool_lock#2 irq_context: 0 rtnl_mutex flowtable_lock irq_context: 0 rtnl_mutex &ul->lock#2 irq_context: 0 rtnl_mutex raw_notifier_lock irq_context: 0 rtnl_mutex bcm_notifier_lock irq_context: 0 rtnl_mutex isotp_notifier_lock irq_context: 0 (wq_completion)bat_events (work_completion)(&(&forw_packet_aggr->delayed_work)->work) &____s->seqcount#2 irq_context: 0 (wq_completion)bat_events (work_completion)(&(&forw_packet_aggr->delayed_work)->work) &____s->seqcount irq_context: softirq rcu_callback &rsp->gp_wait irq_context: 0 &f->f_pos_lock sb_writers#4 irq_context: 0 &f->f_pos_lock sb_writers#4 &sb->s_type->i_mutex_key#8 irq_context: 0 &f->f_pos_lock sb_writers#4 &sb->s_type->i_mutex_key#8 &ei->xattr_sem irq_context: 0 &f->f_pos_lock sb_writers#4 &sb->s_type->i_mutex_key#8 tk_core.seq.seqcount irq_context: 0 &f->f_pos_lock sb_writers#4 &sb->s_type->i_mutex_key#8 mmu_notifier_invalidate_range_start irq_context: 0 &f->f_pos_lock sb_writers#4 &sb->s_type->i_mutex_key#8 &journal->j_state_lock irq_context: 0 &f->f_pos_lock sb_writers#4 &sb->s_type->i_mutex_key#8 jbd2_handle irq_context: 0 &f->f_pos_lock sb_writers#4 &sb->s_type->i_mutex_key#8 jbd2_handle &ei->i_raw_lock irq_context: 0 &f->f_pos_lock sb_writers#4 &sb->s_type->i_mutex_key#8 jbd2_handle &journal->j_wait_updates irq_context: 0 &f->f_pos_lock sb_writers#4 &sb->s_type->i_mutex_key#8 &obj_hash[i].lock irq_context: 0 &f->f_pos_lock sb_writers#4 &sb->s_type->i_mutex_key#8 &sb->s_type->i_lock_key#22 irq_context: 0 &f->f_pos_lock sb_writers#4 &sb->s_type->i_mutex_key#8 &wb->list_lock irq_context: 0 &f->f_pos_lock sb_writers#4 &sb->s_type->i_mutex_key#8 &wb->list_lock &sb->s_type->i_lock_key#22 irq_context: 0 &f->f_pos_lock sb_writers#4 &sb->s_type->i_mutex_key#8 &mm->mmap_lock ptlock_ptr(page)#2 irq_context: 0 &f->f_pos_lock sb_writers#4 &sb->s_type->i_mutex_key#8 fs_reclaim irq_context: 0 &f->f_pos_lock sb_writers#4 &sb->s_type->i_mutex_key#8 fs_reclaim mmu_notifier_invalidate_range_start irq_context: 0 &f->f_pos_lock sb_writers#4 &sb->s_type->i_mutex_key#8 fs_reclaim mmu_notifier_invalidate_range_start &rq->__lock irq_context: 0 &f->f_pos_lock sb_writers#4 &sb->s_type->i_mutex_key#8 fs_reclaim mmu_notifier_invalidate_range_start &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 &f->f_pos_lock sb_writers#4 &sb->s_type->i_mutex_key#8 &____s->seqcount irq_context: 0 &f->f_pos_lock sb_writers#4 &sb->s_type->i_mutex_key#8 stock_lock irq_context: 0 &f->f_pos_lock sb_writers#4 &sb->s_type->i_mutex_key#8 &xa->xa_lock#9 irq_context: 0 &f->f_pos_lock sb_writers#4 &sb->s_type->i_mutex_key#8 lock#4 irq_context: 0 &f->f_pos_lock sb_writers#4 &sb->s_type->i_mutex_key#8 &mapping->private_lock irq_context: 0 &f->f_pos_lock sb_writers#4 &sb->s_type->i_mutex_key#8 &ei->i_es_lock irq_context: 0 &f->f_pos_lock sb_writers#4 &sb->s_type->i_mutex_key#8 &ei->i_data_sem irq_context: 0 &f->f_pos_lock sb_writers#4 &sb->s_type->i_mutex_key#8 &ei->i_data_sem mmu_notifier_invalidate_range_start irq_context: 0 &f->f_pos_lock sb_writers#4 &sb->s_type->i_mutex_key#8 &ei->i_data_sem &ei->i_es_lock irq_context: 0 &f->f_pos_lock sb_writers#4 &sb->s_type->i_mutex_key#8 &ei->i_data_sem &ei->i_es_lock irq_context: 0 &f->f_pos_lock sb_writers#4 &sb->s_type->i_mutex_key#8 &ei->i_data_sem &ei->i_es_lock &sbi->s_es_lock irq_context: 0 &f->f_pos_lock sb_writers#4 &sb->s_type->i_mutex_key#8 &ei->i_data_sem &obj_hash[i].lock irq_context: 0 &f->f_pos_lock sb_writers#4 &sb->s_type->i_mutex_key#8 &ei->i_data_sem &sb->s_type->i_lock_key#22 irq_context: 0 &f->f_pos_lock sb_writers#4 &sb->s_type->i_mutex_key#8 &ei->i_data_sem &(ei->i_block_reservation_lock) irq_context: 0 &f->f_pos_lock sb_writers#4 &sb->s_type->i_mutex_key#8 rcu_read_lock &memcg->move_lock irq_context: 0 &f->f_pos_lock sb_writers#4 &sb->s_type->i_mutex_key#8 rcu_read_lock &xa->xa_lock#9 irq_context: 0 &f->f_pos_lock sb_writers#4 &sb->s_type->i_mutex_key#8 &c->lock irq_context: 0 &f->f_pos_lock sb_writers#4 &sb->s_type->i_mutex_key#8 &____s->seqcount#2 irq_context: 0 &f->f_pos_lock sb_writers#4 &sb->s_type->i_mutex_key#8 &xa->xa_lock#9 pool_lock#2 irq_context: 0 &f->f_pos_lock sb_writers#4 &sb->s_type->i_mutex_key#8 &xa->xa_lock#9 &xa->xa_lock#4 irq_context: 0 &f->f_pos_lock sb_writers#4 &sb->s_type->i_mutex_key#8 &xa->xa_lock#9 &xa->xa_lock#4 pool_lock#2 irq_context: 0 &f->f_pos_lock sb_writers#4 &sb->s_type->i_mutex_key#8 &xa->xa_lock#9 &obj_hash[i].lock irq_context: 0 &f->f_pos_lock sb_writers#4 &sb->s_type->i_mutex_key#8 &xa->xa_lock#9 stock_lock irq_context: 0 &f->f_pos_lock sb_writers#4 &sb->s_type->i_mutex_key#8 &xa->xa_lock#9 &c->lock irq_context: 0 &f->f_pos_lock sb_writers#4 &sb->s_type->i_mutex_key#8 &rq->__lock irq_context: 0 &f->f_pos_lock sb_writers#4 &sb->s_type->i_mutex_key#8 &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)events (work_completion)(&w->work)#2 nf_conntrack_mutex irq_context: 0 (wq_completion)events (work_completion)(&w->work)#2 nf_conntrack_mutex &nf_conntrack_locks[i] irq_context: 0 (wq_completion)events (work_completion)(&w->work)#2 nf_conntrack_mutex &rq->__lock irq_context: 0 (wq_completion)events (work_completion)(&w->work)#2 nf_conntrack_mutex &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)events (linkwatch_work).work rtnl_mutex &dev->tx_global_lock _xmit_NONE#2 irq_context: 0 (wq_completion)events (work_completion)(&w->work)#2 nf_conntrack_mutex &cfs_rq->removed.lock irq_context: 0 &f->f_pos_lock sb_writers#4 &sb->s_type->i_mutex_key#8 rcu_read_lock &xa->xa_lock#9 key#10 irq_context: 0 &f->f_pos_lock sb_writers#4 &sb->s_type->i_mutex_key#8 rcu_read_lock rcu_node_0 irq_context: 0 &f->f_pos_lock sb_writers#4 &sb->s_type->i_mutex_key#8 rcu_read_lock &rq->__lock irq_context: 0 &f->f_pos_lock sb_writers#4 &sb->s_type->i_mutex_key#8 pool_lock#2 irq_context: 0 &f->f_pos_lock sb_writers#4 &sb->s_type->i_mutex_key#8 lock#4 &lruvec->lru_lock irq_context: 0 &f->f_pos_lock sb_writers#4 &sb->s_type->i_mutex_key#8 rcu_read_lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 &f->f_pos_lock sb_writers#4 &sb->s_type->i_mutex_key#8 &ei->i_es_lock key#2 irq_context: 0 &f->f_pos_lock sb_writers#4 &sb->s_type->i_mutex_key#8 &pcp->lock &zone->lock irq_context: 0 &f->f_pos_lock sb_writers#4 &sb->s_type->i_mutex_key#8 &pcp->lock &zone->lock &____s->seqcount irq_context: 0 &f->f_pos_lock sb_writers#4 &sb->s_type->i_mutex_key#8 &ei->i_data_sem &(ei->i_block_reservation_lock) key#14 irq_context: 0 &f->f_pos_lock sb_writers#4 &sb->s_type->i_mutex_key#8 &mm->mmap_lock rcu_read_lock rcu_node_0 irq_context: 0 &f->f_pos_lock sb_writers#4 &sb->s_type->i_mutex_key#8 &mm->mmap_lock rcu_read_lock &rq->__lock irq_context: 0 &f->f_pos_lock sb_writers#4 &sb->s_type->i_mutex_key#8 &mm->mmap_lock rcu_read_lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: softirq (&ndev->rs_timer) rcu_read_lock rcu_read_lock rcu_read_lock_bh &____s->seqcount irq_context: softirq (&ndev->rs_timer) rcu_read_lock rcu_read_lock rcu_read_lock_bh dev->qdisc_tx_busylock ?: &qdisc_tx_busylock _xmit_ETHER#2 &obj_hash[i].lock pool_lock irq_context: softirq (&ndev->rs_timer) rcu_read_lock rcu_read_lock rcu_read_lock_bh dev->qdisc_tx_busylock ?: &qdisc_tx_busylock _xmit_ETHER#2 quarantine_lock irq_context: 0 &f->f_pos_lock sb_writers#4 &sb->s_type->i_mutex_key#8 &sem->wait_lock irq_context: 0 &f->f_pos_lock sb_writers#4 &sb->s_type->i_mutex_key#8 &p->pi_lock irq_context: 0 &f->f_pos_lock sb_writers#4 &sb->s_type->i_mutex_key#8 &p->pi_lock &rq->__lock irq_context: 0 &f->f_pos_lock sb_writers#4 &sb->s_type->i_mutex_key#8 &p->pi_lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: softirq fs/notify/mark.c:89 rcu_read_lock &pool->lock/1 &p->pi_lock &cfs_rq->removed.lock irq_context: 0 &f->f_pos_lock sb_writers#4 &sb->s_type->i_mutex_key#8 jbd2_handle &rq->__lock irq_context: 0 &f->f_pos_lock sb_writers#4 &sb->s_type->i_mutex_key#8 jbd2_handle &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 &mm->mmap_lock mmu_notifier_invalidate_range_start &rq->__lock irq_context: 0 &mm->mmap_lock mmu_notifier_invalidate_range_start &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)events_power_efficient (work_completion)(&(&tbl->gc_work)->work) &tbl->lock &n->lock irq_context: 0 &mm->mmap_lock sb_writers#4 mount_lock irq_context: 0 &mm->mmap_lock sb_writers#4 tk_core.seq.seqcount irq_context: 0 &mm->mmap_lock sb_writers#4 mmu_notifier_invalidate_range_start irq_context: 0 &mm->mmap_lock sb_writers#4 pool_lock#2 irq_context: 0 &mm->mmap_lock sb_writers#4 &journal->j_state_lock irq_context: 0 &mm->mmap_lock sb_writers#4 jbd2_handle irq_context: 0 &mm->mmap_lock sb_writers#4 jbd2_handle &ei->i_raw_lock irq_context: 0 &mm->mmap_lock sb_writers#4 &obj_hash[i].lock irq_context: 0 &f->f_pos_lock sb_writers#4 &sb->s_type->i_mutex_key#8 &mm->mmap_lock irq_context: 0 &f->f_pos_lock sb_writers#4 &sb->s_type->i_mutex_key#8 &mm->mmap_lock &rq->__lock irq_context: 0 &f->f_pos_lock sb_writers#4 &sb->s_type->i_mutex_key#8 &mm->mmap_lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 &f->f_pos_lock sb_writers#4 &sb->s_type->i_mutex_key#8 &mm->mmap_lock &sem->wait_lock irq_context: 0 &f->f_pos_lock sb_writers#4 &sb->s_type->i_mutex_key#8 &mm->mmap_lock fs_reclaim irq_context: 0 &f->f_pos_lock sb_writers#4 &sb->s_type->i_mutex_key#8 &mm->mmap_lock fs_reclaim mmu_notifier_invalidate_range_start irq_context: 0 &f->f_pos_lock sb_writers#4 &sb->s_type->i_mutex_key#8 &mm->mmap_lock fs_reclaim mmu_notifier_invalidate_range_start &rq->__lock irq_context: 0 &f->f_pos_lock sb_writers#4 &sb->s_type->i_mutex_key#8 &mm->mmap_lock fs_reclaim mmu_notifier_invalidate_range_start &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 &f->f_pos_lock sb_writers#4 &sb->s_type->i_mutex_key#8 &mm->mmap_lock &____s->seqcount irq_context: 0 &f->f_pos_lock sb_writers#4 &sb->s_type->i_mutex_key#8 &mm->mmap_lock stock_lock irq_context: 0 &f->f_pos_lock sb_writers#4 &sb->s_type->i_mutex_key#8 &mm->mmap_lock rcu_read_lock rcu_read_lock ptlock_ptr(page) irq_context: 0 &f->f_pos_lock sb_writers#4 &sb->s_type->i_mutex_key#8 &mm->mmap_lock rcu_read_lock rcu_read_lock ptlock_ptr(page)#2 irq_context: 0 &f->f_pos_lock sb_writers#4 &sb->s_type->i_mutex_key#8 &mm->mmap_lock rcu_read_lock rcu_read_lock rcu_node_0 irq_context: 0 &f->f_pos_lock sb_writers#4 &sb->s_type->i_mutex_key#8 &mm->mmap_lock rcu_read_lock rcu_read_lock &rq->__lock irq_context: 0 &f->f_pos_lock sb_writers#4 &sb->s_type->i_mutex_key#8 &mm->mmap_lock rcu_read_lock rcu_read_lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)bat_events (work_completion)(&(&forw_packet_aggr->delayed_work)->work) &hard_iface->bat_iv.ogm_buff_mutex &bat_priv->tt.commit_lock &bat_priv->tvlv.container_list_lock &obj_hash[i].lock irq_context: 0 (wq_completion)bat_events (work_completion)(&(&forw_packet_aggr->delayed_work)->work) &hard_iface->bat_iv.ogm_buff_mutex &bat_priv->tt.commit_lock &bat_priv->tvlv.container_list_lock pool_lock#2 irq_context: 0 sb_writers#8 tomoyo_ss &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 &nft_net->commit_mutex rcu_read_lock rhashtable_bucket irq_context: 0 &mm->mmap_lock sb_pagefaults irq_context: 0 &mm->mmap_lock sb_pagefaults tk_core.seq.seqcount irq_context: 0 &mm->mmap_lock sb_pagefaults mmu_notifier_invalidate_range_start irq_context: 0 &mm->mmap_lock sb_pagefaults pool_lock#2 irq_context: 0 &mm->mmap_lock sb_pagefaults &journal->j_state_lock irq_context: 0 &mm->mmap_lock sb_pagefaults jbd2_handle irq_context: 0 &mm->mmap_lock sb_pagefaults jbd2_handle &rq->__lock irq_context: 0 &mm->mmap_lock sb_pagefaults jbd2_handle &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 &nft_net->commit_mutex nf_ct_proto_mutex irq_context: 0 &nft_net->commit_mutex rcu_read_lock pool_lock#2 irq_context: 0 &nft_net->commit_mutex &rq->__lock irq_context: 0 &nft_net->commit_mutex &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 &nft_net->commit_mutex nl_table_lock irq_context: 0 &nft_net->commit_mutex nl_table_wait.lock irq_context: 0 &nft_net->commit_mutex rcu_read_lock rcu_read_lock_bh &obj_hash[i].lock irq_context: 0 &nft_net->commit_mutex rcu_read_lock rcu_read_lock_bh pool_lock#2 irq_context: 0 &nft_net->commit_mutex rlock-AF_NETLINK irq_context: 0 &nft_net->commit_mutex &p->alloc_lock irq_context: 0 &mm->mmap_lock sb_pagefaults jbd2_handle &ei->i_raw_lock irq_context: 0 &mm->mmap_lock sb_pagefaults jbd2_handle &journal->j_wait_updates irq_context: 0 &mm->mmap_lock sb_pagefaults &obj_hash[i].lock irq_context: 0 &mm->mmap_lock sb_pagefaults mapping.invalidate_lock irq_context: 0 &mm->mmap_lock sb_pagefaults mapping.invalidate_lock &mapping->private_lock irq_context: 0 &mm->mmap_lock sb_pagefaults mapping.invalidate_lock &mapping->private_lock rcu_read_lock &memcg->move_lock irq_context: 0 &mm->mmap_lock &mapping->private_lock irq_context: 0 &mm->mmap_lock &mapping->private_lock rcu_read_lock &memcg->move_lock irq_context: 0 &sb->s_type->i_mutex_key#8 rcu_read_lock mount_lock.seqcount irq_context: 0 &sb->s_type->i_mutex_key#8 rcu_read_lock rcu_read_lock rename_lock.seqcount irq_context: 0 &sb->s_type->i_mutex_key#8 ima_extend_list_mutex irq_context: 0 &sb->s_type->i_mutex_key#8 ima_extend_list_mutex fs_reclaim irq_context: 0 &sb->s_type->i_mutex_key#8 ima_extend_list_mutex fs_reclaim mmu_notifier_invalidate_range_start irq_context: 0 &sb->s_type->i_mutex_key#8 ima_extend_list_mutex &c->lock irq_context: 0 &sb->s_type->i_mutex_key#8 ima_extend_list_mutex &rq->__lock irq_context: 0 &sb->s_type->i_mutex_key#8 ima_extend_list_mutex &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 &type->i_mutex_dir_key#3 stock_lock irq_context: 0 &type->i_mutex_dir_key#3 rcu_read_lock rcu_node_0 irq_context: 0 &type->i_mutex_dir_key#3 rcu_read_lock &rq->__lock &cfs_rq->removed.lock irq_context: 0 &sb->s_type->i_mutex_key#8 ima_extend_list_mutex pool_lock#2 irq_context: 0 sb_writers#4 sb_internal jbd2_handle &ei->i_prealloc_lock irq_context: 0 sb_writers#4 &type->i_mutex_dir_key#3/1 rcu_node_0 irq_context: 0 &sb->s_type->i_mutex_key#8 &p->alloc_lock irq_context: 0 &sb->s_type->i_mutex_key#8 &list->lock irq_context: 0 &sb->s_type->i_mutex_key#8 kauditd_wait.lock irq_context: 0 &sb->s_type->i_mutex_key#8 kauditd_wait.lock &p->pi_lock irq_context: 0 &sb->s_type->i_mutex_key#8 kauditd_wait.lock &p->pi_lock &rq->__lock irq_context: 0 &sb->s_type->i_mutex_key#8 kauditd_wait.lock &p->pi_lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 &f->f_pos_lock sb_writers#4 &sb->s_type->i_mutex_key#8 rcu_node_0 irq_context: 0 sb_writers#4 &type->i_mutex_dir_key#3/1 jbd2_handle &ei->i_data_sem key#14 irq_context: 0 &f->f_pos_lock sb_writers#4 &sb->s_type->i_mutex_key#8 mmu_notifier_invalidate_range_start &rq->__lock irq_context: 0 &f->f_pos_lock sb_writers#4 &sb->s_type->i_mutex_key#8 &ei->i_data_sem &rq->__lock irq_context: 0 &f->f_pos_lock sb_writers#4 &sb->s_type->i_mutex_key#8 &ei->i_data_sem &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 &f->f_pos_lock sb_writers#4 &sb->s_type->i_mutex_key#8 remove_cache_srcu irq_context: 0 &f->f_pos_lock sb_writers#4 &sb->s_type->i_mutex_key#8 remove_cache_srcu quarantine_lock irq_context: 0 &f->f_pos_lock sb_writers#4 &sb->s_type->i_mutex_key#8 remove_cache_srcu &c->lock irq_context: 0 &f->f_pos_lock sb_writers#4 &sb->s_type->i_mutex_key#8 remove_cache_srcu &n->list_lock irq_context: 0 &f->f_pos_lock sb_writers#4 &sb->s_type->i_mutex_key#8 remove_cache_srcu rcu_read_lock pool_lock#2 irq_context: 0 &f->f_pos_lock sb_writers#4 &sb->s_type->i_mutex_key#8 remove_cache_srcu &obj_hash[i].lock irq_context: 0 &f->f_pos_lock sb_writers#4 &sb->s_type->i_mutex_key#8 remove_cache_srcu &rq->__lock irq_context: 0 &f->f_pos_lock sb_writers#4 &sb->s_type->i_mutex_key#8 remove_cache_srcu &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 sb_writers#4 &dentry->d_lock irq_context: 0 sb_writers#4 rcu_read_lock &dentry->d_lock irq_context: 0 sb_writers#4 rcu_read_lock rcu_node_0 irq_context: 0 sb_writers#4 rcu_read_lock &rq->__lock irq_context: 0 sb_writers#4 rcu_read_lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: softirq (&ndev->rs_timer) rcu_read_lock rcu_read_lock rcu_read_lock_bh &hsr->seqnr_lock rcu_read_lock &c->lock irq_context: 0 &sb->s_type->i_mutex_key#10 (netlink_chain).rwsem &nft_net->commit_mutex rcu_state.exp_mutex rcu_node_0 irq_context: 0 &sb->s_type->i_mutex_key#10 (netlink_chain).rwsem &nft_net->commit_mutex rcu_state.exp_mutex &rq->__lock irq_context: 0 &sb->s_type->i_mutex_key#10 (netlink_chain).rwsem &nft_net->commit_mutex rcu_state.exp_mutex &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 &sb->s_type->i_mutex_key#10 (netlink_chain).rwsem &nft_net->commit_mutex rcu_state.exp_mutex &obj_hash[i].lock irq_context: 0 &sb->s_type->i_mutex_key#10 (netlink_chain).rwsem &nft_net->commit_mutex rcu_state.exp_mutex rcu_read_lock &pool->lock irq_context: 0 &sb->s_type->i_mutex_key#10 (netlink_chain).rwsem &nft_net->commit_mutex rcu_state.exp_mutex rcu_read_lock &pool->lock &obj_hash[i].lock irq_context: 0 &sb->s_type->i_mutex_key#10 (netlink_chain).rwsem &nft_net->commit_mutex rcu_state.exp_mutex rcu_read_lock &pool->lock &p->pi_lock irq_context: 0 &sb->s_type->i_mutex_key#10 (netlink_chain).rwsem &nft_net->commit_mutex rcu_state.exp_mutex &rnp->exp_wq[2] irq_context: 0 sb_writers#4 &sb->s_type->i_mutex_key#8 &sem->wait_lock irq_context: 0 &sb->s_type->i_mutex_key#8 &n->list_lock irq_context: 0 &sb->s_type->i_mutex_key#8 &n->list_lock &c->lock irq_context: 0 &sb->s_type->i_mutex_key#8 irq_context: 0 &sb->s_type->i_mutex_key#8 fs_reclaim irq_context: 0 &sb->s_type->i_mutex_key#8 fs_reclaim mmu_notifier_invalidate_range_start irq_context: 0 &sb->s_type->i_mutex_key#8 pool_lock#2 irq_context: 0 &sb->s_type->i_mutex_key#8 &sb->s_type->i_lock_key#22 irq_context: 0 &sb->s_type->i_mutex_key#8 &wb->list_lock irq_context: 0 &sb->s_type->i_mutex_key#8 &sbi->s_writepages_rwsem irq_context: 0 &sb->s_type->i_mutex_key#8 &sbi->s_writepages_rwsem &xa->xa_lock#9 irq_context: 0 &sb->s_type->i_mutex_key#8 &sbi->s_writepages_rwsem mmu_notifier_invalidate_range_start irq_context: 0 &sb->s_type->i_mutex_key#8 &sbi->s_writepages_rwsem mmu_notifier_invalidate_range_start &rq->__lock irq_context: 0 &sb->s_type->i_mutex_key#8 &sbi->s_writepages_rwsem mmu_notifier_invalidate_range_start &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 &sb->s_type->i_mutex_key#8 &sbi->s_writepages_rwsem pool_lock#2 irq_context: 0 &sb->s_type->i_mutex_key#8 &sbi->s_writepages_rwsem lock#4 irq_context: 0 &sb->s_type->i_mutex_key#8 &sbi->s_writepages_rwsem lock#4 &lruvec->lru_lock irq_context: 0 &sb->s_type->i_mutex_key#8 &sbi->s_writepages_rwsem lock#5 irq_context: 0 &sb->s_type->i_mutex_key#8 &sbi->s_writepages_rwsem &obj_hash[i].lock irq_context: 0 &sb->s_type->i_mutex_key#8 &sbi->s_writepages_rwsem &journal->j_state_lock irq_context: 0 &sb->s_type->i_mutex_key#8 &sbi->s_writepages_rwsem jbd2_handle irq_context: 0 &sb->s_type->i_mutex_key#8 &sbi->s_writepages_rwsem jbd2_handle lock#4 irq_context: 0 &sb->s_type->i_mutex_key#8 &sbi->s_writepages_rwsem jbd2_handle lock#5 irq_context: 0 &sb->s_type->i_mutex_key#8 &sbi->s_writepages_rwsem jbd2_handle mmu_notifier_invalidate_range_start irq_context: 0 &sb->s_type->i_mutex_key#8 &sbi->s_writepages_rwsem jbd2_handle pool_lock#2 irq_context: 0 &sb->s_type->i_mutex_key#8 &sbi->s_writepages_rwsem jbd2_handle &ei->i_es_lock irq_context: 0 &sb->s_type->i_mutex_key#8 &sbi->s_writepages_rwsem jbd2_handle &ei->i_data_sem irq_context: 0 &sb->s_type->i_mutex_key#8 &sbi->s_writepages_rwsem jbd2_handle &ei->i_data_sem mmu_notifier_invalidate_range_start irq_context: 0 &sb->s_type->i_mutex_key#8 &sbi->s_writepages_rwsem jbd2_handle &ei->i_data_sem &c->lock irq_context: 0 &sb->s_type->i_mutex_key#8 &sbi->s_writepages_rwsem jbd2_handle &ei->i_data_sem pool_lock#2 irq_context: 0 &sb->s_type->i_mutex_key#8 &sbi->s_writepages_rwsem jbd2_handle &ei->i_data_sem &____s->seqcount#2 irq_context: 0 &sb->s_type->i_mutex_key#8 &sbi->s_writepages_rwsem jbd2_handle &ei->i_data_sem &rq->__lock irq_context: 0 &sb->s_type->i_mutex_key#8 &sbi->s_writepages_rwsem jbd2_handle &ei->i_data_sem &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 &sb->s_type->i_mutex_key#8 &sbi->s_writepages_rwsem jbd2_handle &ei->i_data_sem &____s->seqcount irq_context: 0 &sb->s_type->i_mutex_key#8 &sbi->s_writepages_rwsem jbd2_handle &ei->i_data_sem &ei->i_prealloc_lock irq_context: 0 &sb->s_type->i_mutex_key#8 &sbi->s_writepages_rwsem jbd2_handle &ei->i_data_sem &sbi->s_md_lock irq_context: 0 &sb->s_type->i_mutex_key#8 &sbi->s_writepages_rwsem jbd2_handle &ei->i_data_sem &bgl->locks[i].lock &sbi->s_md_lock irq_context: 0 &sb->s_type->i_mutex_key#8 &sbi->s_writepages_rwsem jbd2_handle &ei->i_data_sem &bgl->locks[i].lock &ei->i_prealloc_lock irq_context: 0 &sb->s_type->i_mutex_key#8 &sbi->s_writepages_rwsem jbd2_handle &ei->i_data_sem &mapping->private_lock irq_context: 0 &sb->s_type->i_mutex_key#8 &sbi->s_writepages_rwsem jbd2_handle &ei->i_data_sem &ret->b_state_lock irq_context: 0 &sb->s_type->i_mutex_key#8 &sbi->s_writepages_rwsem jbd2_handle &ei->i_data_sem &ret->b_state_lock &journal->j_list_lock irq_context: 0 &sb->s_type->i_mutex_key#8 &sbi->s_writepages_rwsem jbd2_handle &ei->i_data_sem &pa->pa_lock#2 irq_context: 0 &sb->s_type->i_mutex_key#8 &sbi->s_writepages_rwsem jbd2_handle &ei->i_data_sem &obj_hash[i].lock irq_context: 0 &sb->s_type->i_mutex_key#8 &sbi->s_writepages_rwsem jbd2_handle &ei->i_data_sem &ei->i_raw_lock irq_context: 0 &sb->s_type->i_mutex_key#8 &sbi->s_writepages_rwsem jbd2_handle &ei->i_data_sem &(ei->i_block_reservation_lock) irq_context: 0 &sb->s_type->i_mutex_key#8 &sbi->s_writepages_rwsem jbd2_handle &ei->i_data_sem &sb->s_type->i_lock_key#22 irq_context: 0 &sb->s_type->i_mutex_key#8 &sbi->s_writepages_rwsem jbd2_handle &ei->i_data_sem &ei->i_es_lock irq_context: 0 &sb->s_type->i_mutex_key#8 &sbi->s_writepages_rwsem jbd2_handle &ei->i_data_sem &ei->i_es_lock irq_context: 0 &sb->s_type->i_mutex_key#8 &sbi->s_writepages_rwsem jbd2_handle &ei->i_data_sem &ei->i_es_lock &____s->seqcount irq_context: 0 &sb->s_type->i_mutex_key#8 &sbi->s_writepages_rwsem jbd2_handle &ei->i_data_sem &ei->i_es_lock pool_lock#2 irq_context: 0 &sb->s_type->i_mutex_key#8 &sbi->s_writepages_rwsem jbd2_handle &ei->i_data_sem &ei->i_es_lock rcu_read_lock pool_lock#2 irq_context: 0 &sb->s_type->i_mutex_key#8 &sbi->s_writepages_rwsem jbd2_handle &ei->i_data_sem &ei->i_es_lock &obj_hash[i].lock irq_context: 0 &sb->s_type->i_mutex_key#10 (netlink_chain).rwsem &nft_net->commit_mutex &obj_hash[i].lock irq_context: 0 &sb->s_type->i_mutex_key#10 (netlink_chain).rwsem &nft_net->commit_mutex nf_ct_proto_mutex irq_context: 0 &sb->s_type->i_mutex_key#10 (netlink_chain).rwsem &nft_net->commit_mutex stock_lock irq_context: 0 &sb->s_type->i_mutex_key#10 (netlink_chain).rwsem &nft_net->commit_mutex pool_lock#2 irq_context: 0 &sb->s_type->i_mutex_key#10 (netlink_chain).rwsem &nft_net->commit_mutex rcu_read_lock rhashtable_bucket irq_context: 0 &sb->s_type->i_mutex_key#10 (netlink_chain).rwsem &nft_net->commit_mutex rcu_read_lock rcu_read_lock &pool->lock irq_context: 0 &sb->s_type->i_mutex_key#10 (netlink_chain).rwsem &nft_net->commit_mutex rcu_read_lock rcu_read_lock &pool->lock &obj_hash[i].lock irq_context: 0 &sb->s_type->i_mutex_key#10 (netlink_chain).rwsem &nft_net->commit_mutex rcu_read_lock rcu_read_lock &pool->lock pool_lock#2 irq_context: 0 &sb->s_type->i_mutex_key#10 (netlink_chain).rwsem &nft_net->commit_mutex rcu_read_lock &pool->lock irq_context: 0 &sb->s_type->i_mutex_key#10 (netlink_chain).rwsem &nft_net->commit_mutex rcu_read_lock &pool->lock &obj_hash[i].lock irq_context: 0 &sb->s_type->i_mutex_key#10 (netlink_chain).rwsem &nft_net->commit_mutex (work_completion)(&ht->run_work) irq_context: 0 &sb->s_type->i_mutex_key#10 (netlink_chain).rwsem &nft_net->commit_mutex &ht->mutex irq_context: 0 &sb->s_type->i_mutex_key#10 (netlink_chain).rwsem &nft_net->commit_mutex &ht->mutex &obj_hash[i].lock irq_context: 0 &sb->s_type->i_mutex_key#10 (netlink_chain).rwsem &nft_net->commit_mutex &ht->mutex pool_lock#2 irq_context: 0 &sb->s_type->i_mutex_key#8 &sbi->s_writepages_rwsem jbd2_handle &mapping->i_mmap_rwsem mmu_notifier_invalidate_range_start irq_context: 0 &sb->s_type->i_mutex_key#8 &sbi->s_writepages_rwsem jbd2_handle &mapping->i_mmap_rwsem ptlock_ptr(page)#2 irq_context: 0 &sb->s_type->i_mutex_key#8 &sbi->s_writepages_rwsem jbd2_handle &mapping->private_lock irq_context: 0 &sb->s_type->i_mutex_key#8 &sbi->s_writepages_rwsem jbd2_handle &mapping->private_lock rcu_read_lock &memcg->move_lock irq_context: 0 &sb->s_type->i_mutex_key#8 &sbi->s_writepages_rwsem jbd2_handle rcu_read_lock &memcg->move_lock irq_context: 0 &sb->s_type->i_mutex_key#8 &sbi->s_writepages_rwsem jbd2_handle rcu_read_lock &xa->xa_lock#9 irq_context: 0 &sb->s_type->i_mutex_key#8 &sbi->s_writepages_rwsem jbd2_handle rcu_read_lock &xa->xa_lock#9 &s->s_inode_wblist_lock irq_context: 0 &sb->s_type->i_mutex_key#8 &sbi->s_writepages_rwsem jbd2_handle rcu_read_lock rcu_read_lock &q->queue_lock irq_context: 0 &sb->s_type->i_mutex_key#8 &sbi->s_writepages_rwsem jbd2_handle rcu_read_lock rcu_read_lock &q->queue_lock &c->lock irq_context: 0 &sb->s_type->i_mutex_key#8 &sbi->s_writepages_rwsem jbd2_handle rcu_read_lock rcu_read_lock &q->queue_lock pool_lock#2 irq_context: 0 &sb->s_type->i_mutex_key#8 &sbi->s_writepages_rwsem jbd2_handle rcu_read_lock rcu_read_lock &q->queue_lock pcpu_lock irq_context: 0 &sb->s_type->i_mutex_key#8 &sbi->s_writepages_rwsem jbd2_handle rcu_read_lock rcu_read_lock &q->queue_lock &obj_hash[i].lock irq_context: 0 &sb->s_type->i_mutex_key#8 &sbi->s_writepages_rwsem jbd2_handle rcu_read_lock rcu_read_lock &q->queue_lock percpu_counters_lock irq_context: 0 &sb->s_type->i_mutex_key#8 &sbi->s_writepages_rwsem jbd2_handle rcu_read_lock rcu_read_lock &q->queue_lock &blkcg->lock irq_context: 0 &sb->s_type->i_mutex_key#8 &sbi->s_writepages_rwsem jbd2_handle rcu_read_lock rcu_read_lock &q->queue_lock &blkcg->lock pool_lock#2 irq_context: 0 &sb->s_type->i_mutex_key#8 &sbi->s_writepages_rwsem jbd2_handle rcu_read_lock rcu_read_lock rcu_node_0 irq_context: 0 &sb->s_type->i_mutex_key#8 &sbi->s_writepages_rwsem jbd2_handle rcu_read_lock rcu_read_lock &rq->__lock irq_context: 0 &sb->s_type->i_mutex_key#8 &sbi->s_writepages_rwsem jbd2_handle rcu_read_lock rcu_read_lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 &sb->s_type->i_mutex_key#8 &sbi->s_writepages_rwsem jbd2_handle rcu_read_lock rcu_node_0 irq_context: 0 &sb->s_type->i_mutex_key#8 &sbi->s_writepages_rwsem jbd2_handle &ei->i_raw_lock irq_context: 0 &sb->s_type->i_mutex_key#8 &sbi->s_writepages_rwsem jbd2_handle &journal->j_wait_updates irq_context: 0 &sb->s_type->i_mutex_key#8 &sbi->s_writepages_rwsem tk_core.seq.seqcount irq_context: 0 &sb->s_type->i_mutex_key#8 &sbi->s_writepages_rwsem &dd->lock irq_context: 0 &sb->s_type->i_mutex_key#8 &sbi->s_writepages_rwsem &base->lock irq_context: 0 &sb->s_type->i_mutex_key#8 &sbi->s_writepages_rwsem &base->lock &obj_hash[i].lock irq_context: 0 &sb->s_type->i_mutex_key#8 &sbi->s_writepages_rwsem rcu_read_lock &dd->lock irq_context: 0 &sb->s_type->i_mutex_key#8 &sbi->s_writepages_rwsem rcu_read_lock &obj_hash[i].lock irq_context: 0 &sb->s_type->i_mutex_key#8 &sbi->s_writepages_rwsem rcu_read_lock tk_core.seq.seqcount irq_context: 0 &sb->s_type->i_mutex_key#8 &sbi->s_writepages_rwsem rcu_read_lock &virtscsi_vq->vq_lock irq_context: 0 &sb->s_type->i_mutex_key#8 &folio_wait_table[i] irq_context: 0 &sb->s_type->i_mutex_key#8 &rq->__lock irq_context: 0 &sb->s_type->i_mutex_key#8 &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 sb_writers#8 kn->active#5 &n->list_lock irq_context: 0 sb_writers#8 kn->active#5 &n->list_lock &c->lock irq_context: 0 &p->lock &of->mutex &rq->__lock irq_context: 0 &p->lock &of->mutex &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 &sb->s_type->i_mutex_key#8 lock#4 irq_context: 0 &sb->s_type->i_mutex_key#8 lock#4 &lruvec->lru_lock irq_context: 0 &sb->s_type->i_mutex_key#8 lock#5 irq_context: 0 &sb->s_type->i_mutex_key#8 &ei->i_es_lock irq_context: 0 &sb->s_type->i_mutex_key#8 &journal->j_state_lock irq_context: 0 &sb->s_type->i_mutex_key#8 tk_core.seq.seqcount irq_context: 0 &sb->s_type->i_mutex_key#8 &dd->lock irq_context: 0 &sb->s_type->i_mutex_key#8 rcu_read_lock &dd->lock irq_context: 0 &sb->s_type->i_mutex_key#8 rcu_read_lock &obj_hash[i].lock irq_context: 0 &sb->s_type->i_mutex_key#8 rcu_read_lock tk_core.seq.seqcount irq_context: 0 &sb->s_type->i_mutex_key#8 rcu_read_lock &virtscsi_vq->vq_lock irq_context: 0 &sb->s_type->i_mutex_key#8 &obj_hash[i].lock irq_context: 0 &sb->s_type->i_mutex_key#8 &base->lock irq_context: 0 &sb->s_type->i_mutex_key#8 &base->lock &obj_hash[i].lock irq_context: 0 &sb->s_type->i_mutex_key#8 (&timer.timer) irq_context: 0 cb_lock genl_mutex (console_sem).lock irq_context: 0 cb_lock genl_mutex console_lock console_srcu console_owner_lock irq_context: 0 cb_lock genl_mutex console_lock console_srcu console_owner irq_context: 0 cb_lock genl_mutex console_lock console_srcu console_owner &port_lock_key irq_context: 0 cb_lock genl_mutex console_lock console_srcu console_owner console_owner_lock irq_context: 0 sb_writers#5 &type->i_mutex_dir_key#5/1 fill_pool_map-wait-type-override &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 &f->f_pos_lock sb_writers#4 &sb->s_type->i_mutex_key#8 rcu_read_lock &rcu_state.gp_wq irq_context: 0 &f->f_pos_lock sb_writers#4 &sb->s_type->i_mutex_key#8 rcu_read_lock &rcu_state.gp_wq &p->pi_lock irq_context: 0 (wq_completion)events (work_completion)(&data->fib_event_work) &data->fib_lock &obj_hash[i].lock pool_lock irq_context: 0 &mm->mmap_lock sb_pagefaults mapping.invalidate_lock rcu_read_lock &memcg->move_lock irq_context: 0 &mm->mmap_lock sb_pagefaults mapping.invalidate_lock rcu_read_lock &xa->xa_lock#9 irq_context: 0 vlan_ioctl_mutex rtnl_mutex fs_reclaim irq_context: 0 vlan_ioctl_mutex rtnl_mutex fs_reclaim mmu_notifier_invalidate_range_start irq_context: 0 vlan_ioctl_mutex rtnl_mutex stock_lock irq_context: 0 vlan_ioctl_mutex rtnl_mutex pool_lock#2 irq_context: 0 vlan_ioctl_mutex rtnl_mutex stack_depot_init_mutex irq_context: 0 vlan_ioctl_mutex rtnl_mutex &c->lock irq_context: 0 vlan_ioctl_mutex rtnl_mutex &____s->seqcount#2 irq_context: 0 vlan_ioctl_mutex rtnl_mutex &____s->seqcount irq_context: 0 vlan_ioctl_mutex rtnl_mutex &macvlan_netdev_addr_lock_key/1 irq_context: 0 vlan_ioctl_mutex rtnl_mutex &macvlan_netdev_addr_lock_key/1 pool_lock#2 irq_context: 0 vlan_ioctl_mutex rtnl_mutex &macvlan_netdev_addr_lock_key/1 _xmit_ETHER irq_context: 0 vlan_ioctl_mutex rtnl_mutex &macvlan_netdev_addr_lock_key/1 _xmit_ETHER pool_lock#2 irq_context: 0 vlan_ioctl_mutex rtnl_mutex &obj_hash[i].lock irq_context: 0 vlan_ioctl_mutex rtnl_mutex batched_entropy_u32.lock irq_context: 0 vlan_ioctl_mutex rtnl_mutex &base->lock irq_context: 0 vlan_ioctl_mutex rtnl_mutex &base->lock &obj_hash[i].lock irq_context: 0 vlan_ioctl_mutex rtnl_mutex pcpu_alloc_mutex irq_context: 0 vlan_ioctl_mutex rtnl_mutex pcpu_alloc_mutex pcpu_lock irq_context: 0 vlan_ioctl_mutex rtnl_mutex &dir->lock#2 irq_context: 0 vlan_ioctl_mutex rtnl_mutex &xa->xa_lock#3 irq_context: 0 vlan_ioctl_mutex rtnl_mutex net_rwsem irq_context: 0 vlan_ioctl_mutex rtnl_mutex net_rwsem &list->lock#2 irq_context: 0 vlan_ioctl_mutex rtnl_mutex &tn->lock irq_context: 0 vlan_ioctl_mutex rtnl_mutex &x->wait#9 irq_context: 0 vlan_ioctl_mutex rtnl_mutex &k->list_lock irq_context: 0 vlan_ioctl_mutex rtnl_mutex gdp_mutex irq_context: 0 vlan_ioctl_mutex rtnl_mutex gdp_mutex &k->list_lock irq_context: 0 vlan_ioctl_mutex rtnl_mutex lock irq_context: 0 vlan_ioctl_mutex rtnl_mutex lock kernfs_idr_lock irq_context: 0 vlan_ioctl_mutex rtnl_mutex &root->kernfs_rwsem irq_context: 0 vlan_ioctl_mutex rtnl_mutex &root->kernfs_rwsem &root->kernfs_iattr_rwsem irq_context: 0 vlan_ioctl_mutex rtnl_mutex bus_type_sem irq_context: 0 vlan_ioctl_mutex rtnl_mutex sysfs_symlink_target_lock irq_context: 0 vlan_ioctl_mutex rtnl_mutex lock kernfs_idr_lock pool_lock#2 irq_context: 0 vlan_ioctl_mutex rtnl_mutex &root->kernfs_rwsem irq_context: 0 vlan_ioctl_mutex rtnl_mutex &dev->power.lock irq_context: 0 vlan_ioctl_mutex rtnl_mutex dpm_list_mtx irq_context: 0 vlan_ioctl_mutex rtnl_mutex uevent_sock_mutex irq_context: 0 vlan_ioctl_mutex rtnl_mutex uevent_sock_mutex fs_reclaim irq_context: 0 vlan_ioctl_mutex rtnl_mutex uevent_sock_mutex fs_reclaim mmu_notifier_invalidate_range_start irq_context: 0 vlan_ioctl_mutex rtnl_mutex uevent_sock_mutex pool_lock#2 irq_context: 0 vlan_ioctl_mutex rtnl_mutex uevent_sock_mutex nl_table_lock irq_context: 0 vlan_ioctl_mutex rtnl_mutex uevent_sock_mutex &obj_hash[i].lock irq_context: 0 vlan_ioctl_mutex rtnl_mutex uevent_sock_mutex nl_table_wait.lock irq_context: 0 vlan_ioctl_mutex rtnl_mutex uevent_sock_mutex &rq->__lock irq_context: 0 cb_lock genl_mutex tcp_metrics_lock irq_context: 0 sb_writers#4 &type->i_mutex_dir_key#3 tomoyo_ss &____s->seqcount#2 irq_context: 0 sb_writers#4 &type->i_mutex_dir_key#3 jbd2_handle &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 vlan_ioctl_mutex rtnl_mutex subsys mutex#17 irq_context: 0 vlan_ioctl_mutex rtnl_mutex subsys mutex#17 &k->k_lock irq_context: 0 vlan_ioctl_mutex rtnl_mutex rcu_read_lock pool_lock#2 irq_context: 0 vlan_ioctl_mutex rtnl_mutex &n->list_lock irq_context: 0 vlan_ioctl_mutex rtnl_mutex &n->list_lock &c->lock irq_context: 0 sb_writers#4 &sb->s_type->i_mutex_key#8 &sbi->s_writepages_rwsem rcu_state.exp_mutex &rnp->exp_wq[3] irq_context: 0 vlan_ioctl_mutex rtnl_mutex dev_hotplug_mutex irq_context: 0 vlan_ioctl_mutex rtnl_mutex dev_hotplug_mutex &dev->power.lock irq_context: 0 vlan_ioctl_mutex rtnl_mutex dev_base_lock irq_context: 0 vlan_ioctl_mutex rtnl_mutex input_pool.lock irq_context: 0 vlan_ioctl_mutex rtnl_mutex rcu_read_lock &pool->lock/1 irq_context: 0 vlan_ioctl_mutex rtnl_mutex rcu_read_lock &pool->lock/1 &obj_hash[i].lock irq_context: 0 vlan_ioctl_mutex rtnl_mutex rcu_read_lock &pool->lock/1 pool_lock#2 irq_context: 0 vlan_ioctl_mutex rtnl_mutex rcu_read_lock &pool->lock/1 &p->pi_lock irq_context: 0 vlan_ioctl_mutex rtnl_mutex rcu_read_lock &pool->lock/1 &p->pi_lock &rq->__lock irq_context: 0 vlan_ioctl_mutex rtnl_mutex rcu_read_lock &pool->lock/1 &p->pi_lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 vlan_ioctl_mutex rtnl_mutex &tbl->lock irq_context: 0 vlan_ioctl_mutex rtnl_mutex sysctl_lock irq_context: 0 vlan_ioctl_mutex rtnl_mutex nl_table_lock irq_context: 0 vlan_ioctl_mutex rtnl_mutex nl_table_wait.lock irq_context: 0 vlan_ioctl_mutex rtnl_mutex failover_lock irq_context: 0 vlan_ioctl_mutex rtnl_mutex proc_subdir_lock irq_context: 0 vlan_ioctl_mutex rtnl_mutex proc_inum_ida.xa_lock irq_context: 0 vlan_ioctl_mutex rtnl_mutex proc_subdir_lock irq_context: 0 vlan_ioctl_mutex rtnl_mutex &idev->mc_lock irq_context: 0 vlan_ioctl_mutex rtnl_mutex &idev->mc_lock fs_reclaim irq_context: 0 vlan_ioctl_mutex rtnl_mutex &idev->mc_lock fs_reclaim mmu_notifier_invalidate_range_start irq_context: 0 vlan_ioctl_mutex rtnl_mutex &idev->mc_lock pool_lock#2 irq_context: 0 vlan_ioctl_mutex rtnl_mutex &idev->mc_lock &obj_hash[i].lock irq_context: 0 vlan_ioctl_mutex rtnl_mutex &idev->mc_lock &c->lock irq_context: 0 vlan_ioctl_mutex rtnl_mutex &idev->mc_lock &vlan_netdev_addr_lock_key irq_context: 0 vlan_ioctl_mutex rtnl_mutex &idev->mc_lock &vlan_netdev_addr_lock_key pool_lock#2 irq_context: 0 vlan_ioctl_mutex rtnl_mutex &pnettable->lock irq_context: 0 vlan_ioctl_mutex rtnl_mutex smc_ib_devices.mutex irq_context: 0 &f->f_pos_lock sb_writers#4 &sb->s_type->i_mutex_key#8 rcu_read_lock rcu_read_lock rcu_node_0 irq_context: 0 &f->f_pos_lock sb_writers#4 &sb->s_type->i_mutex_key#8 rcu_read_lock rcu_read_lock &rq->__lock irq_context: 0 &f->f_pos_lock sb_writers#4 &sb->s_type->i_mutex_key#8 rcu_read_lock rcu_read_lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 vlan_ioctl_mutex rtnl_mutex rcu_read_lock &ndev->lock irq_context: 0 vlan_ioctl_mutex rtnl_mutex lweventlist_lock irq_context: 0 vlan_ioctl_mutex rtnl_mutex lweventlist_lock pool_lock#2 irq_context: 0 vlan_ioctl_mutex rtnl_mutex lweventlist_lock &____s->seqcount irq_context: 0 vlan_ioctl_mutex rtnl_mutex lweventlist_lock rcu_read_lock pool_lock#2 irq_context: 0 vlan_ioctl_mutex rtnl_mutex lweventlist_lock &obj_hash[i].lock irq_context: 0 vlan_ioctl_mutex rtnl_mutex lweventlist_lock &dir->lock#2 irq_context: 0 vlan_ioctl_mutex &rq->__lock irq_context: 0 vlan_ioctl_mutex &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)events (work_completion)(&rdev->wiphy_work) &rdev->wiphy.mtx &wdev->mtx &n->list_lock irq_context: 0 (wq_completion)events (work_completion)(&rdev->wiphy_work) &rdev->wiphy.mtx &wdev->mtx &n->list_lock &c->lock irq_context: 0 &mm->mmap_lock sb_writers#4 jbd2_handle &journal->j_wait_updates irq_context: 0 &sb->s_type->i_mutex_key#8 &sem->wait_lock irq_context: 0 &f->f_pos_lock sb_writers#4 &sem->wait_lock irq_context: 0 &f->f_pos_lock sb_writers#4 &p->pi_lock irq_context: 0 &f->f_pos_lock sb_writers#4 &p->pi_lock &rq->__lock irq_context: 0 &f->f_pos_lock sb_writers#4 &p->pi_lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 &f->f_pos_lock sb_writers#4 &rq->__lock irq_context: 0 sb_writers#4 &sb->s_type->i_mutex_key#8 integrity_iint_lock irq_context: 0 sb_writers#4 &iint->mutex irq_context: 0 sb_writers#4 &iint->mutex &rq->__lock irq_context: 0 sb_writers#4 &iint->mutex &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 sb_writers#4 &sb->s_type->i_mutex_key#8 mapping.invalidate_lock &mapping->i_mmap_rwsem lock#4 irq_context: 0 sb_writers#4 &sb->s_type->i_mutex_key#8 mapping.invalidate_lock &mapping->i_mmap_rwsem lock#4 &lruvec->lru_lock irq_context: 0 sb_writers#4 &sb->s_type->i_mutex_key#8 mapping.invalidate_lock &mapping->i_mmap_rwsem lock#5 irq_context: 0 sb_writers#4 &sb->s_type->i_mutex_key#8 mapping.invalidate_lock &mapping->i_mmap_rwsem mmu_notifier_invalidate_range_start irq_context: 0 sb_writers#4 &sb->s_type->i_mutex_key#8 mapping.invalidate_lock &mapping->i_mmap_rwsem ptlock_ptr(page)#2 irq_context: 0 sb_writers#4 &sb->s_type->i_mutex_key#8 mapping.invalidate_lock &mapping->i_mmap_rwsem ptlock_ptr(page)#2 &mapping->private_lock irq_context: 0 sb_writers#4 &sb->s_type->i_mutex_key#8 mapping.invalidate_lock &mapping->i_mmap_rwsem ptlock_ptr(page)#2 &mapping->private_lock rcu_read_lock &memcg->move_lock irq_context: 0 sb_writers#4 &sb->s_type->i_mutex_key#8 mapping.invalidate_lock &mapping->i_mmap_rwsem ptlock_ptr(page)#2 &____s->seqcount irq_context: 0 sb_writers#4 &sb->s_type->i_mutex_key#8 mapping.invalidate_lock &mapping->i_mmap_rwsem ptlock_ptr(page)#2 pool_lock#2 irq_context: 0 sb_writers#4 &sb->s_type->i_mutex_key#8 mapping.invalidate_lock &mapping->i_mmap_rwsem &rq->__lock irq_context: 0 sb_writers#4 &sb->s_type->i_mutex_key#8 mapping.invalidate_lock &mapping->i_mmap_rwsem &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 sb_writers#4 &sb->s_type->i_mutex_key#8 mapping.invalidate_lock &mapping->i_mmap_rwsem ptlock_ptr(page)#2 key irq_context: 0 sb_writers#4 &sb->s_type->i_mutex_key#8 mapping.invalidate_lock &mapping->i_mmap_rwsem rcu_read_lock pool_lock#2 irq_context: 0 sb_writers#4 &sb->s_type->i_mutex_key#8 mapping.invalidate_lock &mapping->i_mmap_rwsem &obj_hash[i].lock irq_context: 0 sb_writers#4 &sb->s_type->i_mutex_key#8 mapping.invalidate_lock &mapping->private_lock irq_context: 0 sb_writers#4 &sb->s_type->i_mutex_key#8 mapping.invalidate_lock &mapping->private_lock rcu_read_lock &memcg->move_lock irq_context: 0 sb_writers#4 &sb->s_type->i_mutex_key#8 mapping.invalidate_lock stock_lock irq_context: 0 sb_writers#4 &sb->s_type->i_mutex_key#8 mapping.invalidate_lock rcu_read_lock rcu_node_0 irq_context: 0 sb_writers#4 &sb->s_type->i_mutex_key#8 mapping.invalidate_lock rcu_read_lock &rq->__lock irq_context: 0 sb_writers#4 &sb->s_type->i_mutex_key#8 mapping.invalidate_lock rcu_read_lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 sb_writers#4 &type->i_mutex_dir_key#3/1 &sb->s_type->i_mutex_key#8 remove_cache_srcu &rq->__lock irq_context: 0 sb_writers#4 &type->i_mutex_dir_key#3/1 &sb->s_type->i_mutex_key#8 remove_cache_srcu &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 sb_writers#4 &sb->s_type->i_mutex_key#8 mapping.invalidate_lock &sb->s_type->i_lock_key#22 irq_context: 0 sb_writers#4 &sb->s_type->i_mutex_key#8 mapping.invalidate_lock &sb->s_type->i_lock_key#22 &xa->xa_lock#9 irq_context: 0 sb_writers#4 &sb->s_type->i_mutex_key#8 mapping.invalidate_lock lock#4 irq_context: 0 sb_writers#4 &sb->s_type->i_mutex_key#8 mapping.invalidate_lock lock#5 irq_context: 0 sb_writers#4 &sb->s_type->i_mutex_key#8 mapping.invalidate_lock &lruvec->lru_lock irq_context: 0 sb_writers#4 &sb->s_type->i_mutex_key#8 mapping.invalidate_lock rcu_read_lock pool_lock#2 irq_context: 0 sb_writers#4 &sb->s_type->i_mutex_key#8 mapping.invalidate_lock &mapping->private_lock rcu_read_lock rcu_read_lock key#10 irq_context: 0 sb_writers#4 &sb->s_type->i_mutex_key#8 mapping.invalidate_lock &rq->__lock irq_context: 0 sb_writers#4 &sb->s_type->i_mutex_key#8 mapping.invalidate_lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 sb_writers#4 &type->i_mutex_dir_key#3/1 &type->i_mutex_dir_key#3 jbd2_handle &rq->__lock irq_context: 0 sb_writers#4 &type->i_mutex_dir_key#3/1 &type->i_mutex_dir_key#3 jbd2_handle &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 &nft_net->commit_mutex rcu_state.exp_mutex &rnp->exp_wq[0] irq_context: 0 &mm->mmap_lock &anon_vma->rwsem pool_lock#2 irq_context: 0 sb_writers#4 sb_internal jbd2_handle rcu_node_0 irq_context: 0 sb_writers#4 &sb->s_type->i_mutex_key#8 mapping.invalidate_lock rcu_read_lock &rcu_state.gp_wq irq_context: 0 sb_writers#4 &sb->s_type->i_mutex_key#8 mapping.invalidate_lock rcu_read_lock &rcu_state.gp_wq &p->pi_lock irq_context: 0 sb_writers#4 &sb->s_type->i_mutex_key#8 mapping.invalidate_lock rcu_read_lock &rcu_state.gp_wq &p->pi_lock &rq->__lock irq_context: 0 sb_writers#4 &sb->s_type->i_mutex_key#8 mapping.invalidate_lock rcu_read_lock &rcu_state.gp_wq &p->pi_lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 sb_writers#4 &type->i_mutex_dir_key#3/1 tomoyo_ss rcu_read_lock &rq->__lock irq_context: 0 sb_writers#4 &type->i_mutex_dir_key#3/1 tomoyo_ss rcu_read_lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 sb_writers#4 &sb->s_type->i_mutex_key#8 mapping.invalidate_lock &sb->s_type->i_lock_key#22 &xa->xa_lock#9 &obj_hash[i].lock irq_context: 0 sb_writers#4 &sb->s_type->i_mutex_key#8 mapping.invalidate_lock &sb->s_type->i_lock_key#22 &xa->xa_lock#9 pool_lock#2 irq_context: 0 &sb->s_type->i_mutex_key#10 (netlink_chain).rwsem &nft_net->commit_mutex rcu_state.exp_mutex &rnp->exp_wq[1] irq_context: 0 &iint->mutex ima_extend_list_mutex fs_reclaim &rq->__lock irq_context: 0 &iint->mutex ima_extend_list_mutex fs_reclaim &rq->__lock &cfs_rq->removed.lock irq_context: 0 &iint->mutex ima_extend_list_mutex fs_reclaim &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 &sb->s_type->i_mutex_key#10 (netlink_chain).rwsem &nft_net->commit_mutex rcu_read_lock rcu_read_lock &pool->lock &p->pi_lock irq_context: 0 &sb->s_type->i_mutex_key#10 (netlink_chain).rwsem &nft_net->commit_mutex rcu_read_lock rcu_read_lock &pool->lock &p->pi_lock &rq->__lock irq_context: 0 &sb->s_type->i_mutex_key#10 (netlink_chain).rwsem &nft_net->commit_mutex rcu_read_lock rcu_read_lock &pool->lock &p->pi_lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 &sb->s_type->i_mutex_key#10 (netlink_chain).rwsem &nft_net->commit_mutex rcu_read_lock rcu_node_0 irq_context: 0 &sb->s_type->i_mutex_key#10 (netlink_chain).rwsem &nft_net->commit_mutex rcu_read_lock &rq->__lock irq_context: 0 &mm->mmap_lock ptlock_ptr(page)#2 &mapping->private_lock irq_context: 0 &sb->s_type->i_mutex_key#10 (netlink_chain).rwsem &nft_net->commit_mutex rcu_read_lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 &mm->mmap_lock ptlock_ptr(page)#2 &mapping->private_lock rcu_read_lock &memcg->move_lock irq_context: 0 &mm->mmap_lock ptlock_ptr(page)#2 &____s->seqcount irq_context: 0 &mm->mmap_lock ptlock_ptr(page)#2 pool_lock#2 irq_context: 0 &mm->mmap_lock sb_writers#4 &sb->s_type->i_lock_key#22 irq_context: 0 &mm->mmap_lock sb_writers#4 &wb->list_lock irq_context: 0 &mm->mmap_lock sb_writers#4 &wb->list_lock &sb->s_type->i_lock_key#22 irq_context: 0 &mm->mmap_lock sb_writers#4 &rq->__lock irq_context: 0 &mm->mmap_lock sb_writers#4 &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 &sighand->siglock &c->lock irq_context: 0 &xt[i].mutex &mm->mmap_lock &cfs_rq->removed.lock irq_context: 0 &xt[i].mutex &mm->mmap_lock &obj_hash[i].lock irq_context: softirq (&ndev->rs_timer) rcu_read_lock rcu_read_lock rcu_read_lock_bh key#20 irq_context: softirq (&ndev->rs_timer) rcu_read_lock rcu_read_lock rcu_read_lock_bh &entry->crc_lock irq_context: softirq (&ndev->rs_timer) rcu_read_lock rcu_read_lock rcu_read_lock_bh &c->lock irq_context: softirq (&ndev->rs_timer) rcu_read_lock rcu_read_lock rcu_read_lock_bh &list->lock#5 irq_context: 0 sk_lock-AF_INET6 &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 &sb->s_type->i_mutex_key#8 &sbi->s_writepages_rwsem jbd2_handle &ei->i_data_sem &ei->i_prealloc_lock &pa->pa_lock#2 irq_context: 0 sb_writers#4 &sb->s_type->i_mutex_key#8 mapping.invalidate_lock jbd2_handle &ei->i_data_sem &(ei->i_block_reservation_lock) irq_context: 0 sb_writers#4 &sb->s_type->i_mutex_key#8 mapping.invalidate_lock jbd2_handle &ei->i_data_sem &(ei->i_block_reservation_lock) key#14 irq_context: 0 &sb->s_type->i_mutex_key#8 &sbi->s_writepages_rwsem jbd2_handle &ei->i_data_sem &ei->i_es_lock &c->lock irq_context: 0 &sb->s_type->i_mutex_key#8 &sbi->s_writepages_rwsem jbd2_handle &ei->i_data_sem &ei->i_es_lock &____s->seqcount#2 irq_context: 0 &sb->s_type->i_mutex_key#8 &ei->i_es_lock key#2 irq_context: 0 sb_writers#8 &of->mutex kn->active#5 uevent_sock_mutex rcu_read_lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 &sb->s_type->i_mutex_key#8 &sbi->s_writepages_rwsem jbd2_handle &c->lock irq_context: 0 &sb->s_type->i_mutex_key#8 &c->lock irq_context: 0 &sb->s_type->i_mutex_key#8 &sbi->s_writepages_rwsem &c->lock irq_context: 0 &sb->s_type->i_mutex_key#8 &sbi->s_writepages_rwsem &____s->seqcount#2 irq_context: 0 &sb->s_type->i_mutex_key#8 &sbi->s_writepages_rwsem &____s->seqcount irq_context: 0 (wq_completion)bat_events (work_completion)(&(&bat_priv->mcast.work)->work) &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)bat_events (work_completion)(&(&bat_priv->mcast.work)->work) &cfs_rq->removed.lock irq_context: 0 &sb->s_type->i_mutex_key#8 &____s->seqcount#2 irq_context: 0 &tsk->futex_exit_mutex &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 sb_writers#4 &type->i_mutex_dir_key#3/1 rcu_read_lock &rcu_state.gp_wq irq_context: 0 sb_writers#4 &type->i_mutex_dir_key#3/1 rcu_read_lock &rcu_state.gp_wq &p->pi_lock irq_context: 0 sb_writers#4 &type->i_mutex_dir_key#3 &____s->seqcount#2 irq_context: 0 sb_writers#4 &type->i_mutex_dir_key#3/1 rcu_read_lock &rcu_state.gp_wq &p->pi_lock &rq->__lock irq_context: 0 sb_writers#4 &type->i_mutex_dir_key#3/1 rcu_read_lock &rcu_state.gp_wq &p->pi_lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 sb_writers#4 &type->i_mutex_dir_key#3 inode_hash_lock irq_context: 0 sb_writers#4 &type->i_mutex_dir_key#3 inode_hash_lock &sb->s_type->i_lock_key#22 irq_context: 0 sb_writers#4 &type->i_mutex_dir_key#3 inode_hash_lock &s->s_inode_list_lock irq_context: 0 sb_writers#4 &type->i_mutex_dir_key#3 &sb->s_type->i_lock_key#22 &dentry->d_lock irq_context: 0 sb_writers#4 &type->i_mutex_dir_key#3 &sb->s_type->i_lock_key#22 &dentry->d_lock &wq#2 irq_context: 0 sb_writers#4 fs_reclaim irq_context: 0 sb_writers#4 fs_reclaim mmu_notifier_invalidate_range_start irq_context: 0 sb_writers#4 &sb->s_type->i_mutex_key#8 integrity_iint_lock irq_context: 0 sb_writers#4 &iint->mutex &ei->xattr_sem irq_context: 0 sb_writers#4 &iint->mutex mmu_notifier_invalidate_range_start irq_context: 0 sb_writers#4 &iint->mutex fs_reclaim irq_context: 0 sb_writers#4 &iint->mutex fs_reclaim &rq->__lock irq_context: 0 sb_writers#4 &iint->mutex fs_reclaim &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 sb_writers#4 &iint->mutex fs_reclaim mmu_notifier_invalidate_range_start irq_context: 0 sb_writers#4 &iint->mutex &c->lock irq_context: 0 sb_writers#4 &iint->mutex &____s->seqcount#2 irq_context: 0 sb_writers#4 &iint->mutex &____s->seqcount irq_context: 0 sb_writers#4 &iint->mutex rcu_read_lock mount_lock.seqcount irq_context: 0 sb_writers#4 &iint->mutex rcu_read_lock rcu_read_lock rename_lock.seqcount irq_context: 0 sb_writers#4 &iint->mutex &n->list_lock irq_context: 0 sb_writers#4 &iint->mutex &n->list_lock &c->lock irq_context: 0 sb_writers#4 &iint->mutex ima_extend_list_mutex irq_context: 0 sb_writers#4 &iint->mutex &obj_hash[i].lock irq_context: 0 sb_writers#4 &iint->mutex pool_lock#2 irq_context: 0 sb_writers#4 &ei->xattr_sem irq_context: 0 (wq_completion)bat_events (work_completion)(&(&forw_packet_aggr->delayed_work)->work) &hard_iface->bat_iv.ogm_buff_mutex &n->list_lock irq_context: 0 (wq_completion)bat_events (work_completion)(&(&forw_packet_aggr->delayed_work)->work) &hard_iface->bat_iv.ogm_buff_mutex &n->list_lock &c->lock irq_context: 0 sb_writers#5 &sb->s_type->i_mutex_key#12 rcu_read_lock rcu_node_0 irq_context: 0 sb_writers#5 &sb->s_type->i_mutex_key#12 rcu_read_lock &rq->__lock irq_context: 0 sb_writers#5 &sb->s_type->i_mutex_key#12 rcu_read_lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 &nft_net->commit_mutex &pcp->lock &zone->lock irq_context: 0 &nft_net->commit_mutex &pcp->lock &zone->lock &____s->seqcount irq_context: 0 &sb->s_type->i_mutex_key#10 (netlink_chain).rwsem &nft_net->commit_mutex rcu_state.exp_mutex rcu_read_lock &pool->lock &p->pi_lock &rq->__lock irq_context: 0 &sb->s_type->i_mutex_key#10 (netlink_chain).rwsem &nft_net->commit_mutex rcu_state.exp_mutex rcu_read_lock &pool->lock &p->pi_lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 &sb->s_type->i_mutex_key#10 (netlink_chain).rwsem &nft_net->commit_mutex rcu_state.exp_mutex &rnp->exp_wq[0] irq_context: 0 &tsk->futex_exit_mutex &cfs_rq->removed.lock irq_context: 0 &tsk->futex_exit_mutex &obj_hash[i].lock irq_context: 0 sb_writers#4 sb_internal jbd2_handle &journal->j_list_lock irq_context: 0 sb_writers#4 sb_internal jbd2_handle &obj_hash[i].lock irq_context: 0 sb_writers#4 sb_internal jbd2_handle pool_lock#2 irq_context: 0 sb_writers#4 integrity_iint_lock irq_context: 0 &sb->s_type->i_mutex_key#8 &n->list_lock irq_context: 0 &sb->s_type->i_mutex_key#8 &n->list_lock &c->lock irq_context: 0 &sb->s_type->i_mutex_key#8 &sbi->s_writepages_rwsem jbd2_handle &rq->__lock irq_context: 0 &sb->s_type->i_mutex_key#8 &sbi->s_writepages_rwsem &rq->__lock irq_context: 0 &sb->s_type->i_mutex_key#8 &sbi->s_writepages_rwsem jbd2_handle &____s->seqcount#2 irq_context: 0 &sb->s_type->i_mutex_key#8 &sbi->s_writepages_rwsem jbd2_handle &____s->seqcount irq_context: 0 sb_writers#4 &type->i_mutex_dir_key#3/1 sb_internal &c->lock irq_context: 0 sb_writers#4 &type->i_mutex_dir_key#3/1 sb_internal &____s->seqcount#2 irq_context: 0 sb_writers#4 &type->i_mutex_dir_key#3/1 sb_internal &____s->seqcount irq_context: 0 sb_writers#4 &type->i_mutex_dir_key#3/1 sb_internal jbd2_handle &ei->i_data_sem &n->list_lock irq_context: 0 sb_writers#4 &type->i_mutex_dir_key#3/1 sb_internal jbd2_handle &ei->i_data_sem &n->list_lock &c->lock irq_context: 0 sb_writers#4 sb_internal &____s->seqcount#2 irq_context: 0 sb_writers#4 sb_internal &____s->seqcount irq_context: 0 &sb->s_type->i_mutex_key#8 &sbi->s_writepages_rwsem jbd2_handle &ei->i_data_sem &ei->i_es_lock key#7 irq_context: 0 sb_writers#4 &type->i_mutex_dir_key#3/1 sb_internal jbd2_handle &ei->i_data_sem rcu_read_lock rcu_node_0 irq_context: 0 sb_writers#4 &type->i_mutex_dir_key#3/1 sb_internal jbd2_handle &ei->i_data_sem rcu_read_lock &rq->__lock irq_context: 0 sb_writers#4 &type->i_mutex_dir_key#3/1 sb_internal jbd2_handle &ei->i_data_sem rcu_read_lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 &mm->mmap_lock &anon_vma->rwsem rcu_read_lock &rcu_state.gp_wq irq_context: 0 &mm->mmap_lock &anon_vma->rwsem rcu_read_lock &rcu_state.gp_wq &p->pi_lock irq_context: 0 &mm->mmap_lock &anon_vma->rwsem rcu_read_lock &rcu_state.gp_wq &p->pi_lock &rq->__lock irq_context: 0 &mm->mmap_lock &anon_vma->rwsem rcu_read_lock &rcu_state.gp_wq &p->pi_lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 &sb->s_type->i_mutex_key#8 &sbi->s_writepages_rwsem jbd2_handle &ei->i_data_sem &ei->i_es_lock key#6 irq_context: 0 &sb->s_type->i_mutex_key#8 &sbi->s_writepages_rwsem jbd2_handle rcu_read_lock key#10 irq_context: 0 &xt[i].mutex fs_reclaim &rq->__lock irq_context: 0 &xt[i].mutex fs_reclaim &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 &sb->s_type->i_mutex_key#8 &obj_hash[i].lock pool_lock irq_context: 0 &sb->s_type->i_mutex_key#8 &sbi->s_writepages_rwsem jbd2_handle &ei->i_data_sem key#3 irq_context: 0 &sb->s_type->i_mutex_key#8 &sbi->s_writepages_rwsem jbd2_handle rcu_read_lock &xa->xa_lock#9 key#10 irq_context: 0 (wq_completion)ext4-rsv-conversion (work_completion)(&ei->i_rsv_conversion_work) jbd2_handle &ei->i_es_lock key#2 irq_context: 0 &sb->s_type->i_mutex_key#8 rcu_read_lock rcu_node_0 irq_context: 0 &sb->s_type->i_mutex_key#8 rcu_read_lock &rq->__lock irq_context: 0 &sb->s_type->i_mutex_key#8 rcu_read_lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)bat_events (work_completion)(&(&bat_priv->tt.work)->work) &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 sb_writers#3 &sb->s_type->i_mutex_key#9 rcu_read_lock rcu_node_0 irq_context: 0 sb_writers#3 &sb->s_type->i_mutex_key#9 rcu_read_lock &rq->__lock irq_context: 0 sb_writers#3 &sb->s_type->i_mutex_key#9 rcu_read_lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 &sb->s_type->i_mutex_key#8 &sbi->s_writepages_rwsem &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)ext4-rsv-conversion (work_completion)(&ei->i_rsv_conversion_work) jbd2_handle &ei->i_data_sem remove_cache_srcu irq_context: 0 (wq_completion)ext4-rsv-conversion (work_completion)(&ei->i_rsv_conversion_work) jbd2_handle &ei->i_data_sem remove_cache_srcu quarantine_lock irq_context: 0 (wq_completion)ext4-rsv-conversion (work_completion)(&ei->i_rsv_conversion_work) jbd2_handle &ei->i_data_sem remove_cache_srcu &c->lock irq_context: 0 (wq_completion)ext4-rsv-conversion (work_completion)(&ei->i_rsv_conversion_work) jbd2_handle &ei->i_data_sem remove_cache_srcu &n->list_lock irq_context: 0 (wq_completion)ext4-rsv-conversion (work_completion)(&ei->i_rsv_conversion_work) jbd2_handle &ei->i_data_sem remove_cache_srcu rcu_read_lock pool_lock#2 irq_context: 0 (wq_completion)ext4-rsv-conversion (work_completion)(&ei->i_rsv_conversion_work) jbd2_handle &ei->i_data_sem remove_cache_srcu &obj_hash[i].lock irq_context: 0 &sb->s_type->i_mutex_key#8 &sbi->s_writepages_rwsem jbd2_handle &ei->i_data_sem &ei->i_es_lock key#2 irq_context: 0 sb_writers#4 &type->i_mutex_dir_key#3 &n->list_lock irq_context: 0 sb_writers#4 &type->i_mutex_dir_key#3 &n->list_lock &c->lock irq_context: 0 sb_writers#4 &type->i_mutex_dir_key#3 tomoyo_ss &pcp->lock &zone->lock irq_context: 0 sb_writers#4 &type->i_mutex_dir_key#3 tomoyo_ss &pcp->lock &zone->lock &____s->seqcount irq_context: 0 (wq_completion)mld (work_completion)(&(&idev->mc_dad_work)->work) &idev->mc_lock rcu_read_lock rcu_read_lock rcu_read_lock_bh rcu_read_lock batched_entropy_u32.lock crngs.lock irq_context: 0 sb_writers#5 &type->i_mutex_dir_key#5/1 rcu_read_lock rcu_node_0 irq_context: 0 &mm->mmap_lock remove_cache_srcu rcu_read_lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 &f->f_pos_lock sb_writers#4 &sb->s_type->i_mutex_key#8 remove_cache_srcu rcu_read_lock rcu_node_0 irq_context: 0 &f->f_pos_lock sb_writers#4 &sb->s_type->i_mutex_key#8 remove_cache_srcu rcu_read_lock &rq->__lock irq_context: 0 &f->f_pos_lock sb_writers#4 &sb->s_type->i_mutex_key#8 remove_cache_srcu rcu_read_lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 &f->f_pos_lock sb_writers#4 &sb->s_type->i_mutex_key#8 &n->list_lock irq_context: 0 sb_writers#4 &type->i_mutex_dir_key#3/1 &pcp->lock &zone->lock &____s->seqcount irq_context: 0 &sb->s_type->i_mutex_key#8 &sbi->s_writepages_rwsem jbd2_handle &ei->i_es_lock key#2 irq_context: 0 sb_writers#5 &type->i_mutex_dir_key#5/1 rcu_read_lock &rcu_state.gp_wq irq_context: 0 sb_writers#5 &type->i_mutex_dir_key#5/1 rcu_read_lock &rcu_state.gp_wq &p->pi_lock irq_context: 0 sb_writers#5 &type->i_mutex_dir_key#5/1 rcu_read_lock &rcu_state.gp_wq &p->pi_lock &rq->__lock irq_context: 0 (wq_completion)ext4-rsv-conversion (work_completion)(&ei->i_rsv_conversion_work) jbd2_handle &ei->i_data_sem &____s->seqcount#2 irq_context: 0 sb_writers#5 &type->i_mutex_dir_key#5/1 rcu_read_lock &rcu_state.gp_wq &p->pi_lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)ext4-rsv-conversion (work_completion)(&ei->i_rsv_conversion_work) jbd2_handle &ei->i_data_sem &____s->seqcount irq_context: 0 (wq_completion)ext4-rsv-conversion (work_completion)(&ei->i_rsv_conversion_work) jbd2_handle &ei->i_data_sem &rq->__lock irq_context: 0 (wq_completion)ext4-rsv-conversion (work_completion)(&ei->i_rsv_conversion_work) jbd2_handle &ei->i_data_sem &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 &f->f_pos_lock sb_writers#4 &sb->s_type->i_mutex_key#8 rcu_read_lock &rcu_state.gp_wq &p->pi_lock &rq->__lock irq_context: 0 &f->f_pos_lock sb_writers#4 &sb->s_type->i_mutex_key#8 rcu_read_lock &rcu_state.gp_wq &p->pi_lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 sb_writers#5 &type->i_mutex_dir_key#5/1 &sb->s_type->i_mutex_key#12 rcu_read_lock rcu_node_0 irq_context: 0 sb_writers#5 &type->i_mutex_dir_key#5/1 &sb->s_type->i_mutex_key#12 rcu_read_lock &rq->__lock irq_context: 0 &p->lock &of->mutex kn->active#5 rcu_read_lock rcu_node_0 irq_context: 0 &p->lock &of->mutex kn->active#5 rcu_read_lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 alg_types_sem irq_context: 0 (wq_completion)events (work_completion)(&rdev->wiphy_work) &rdev->wiphy.mtx &wdev->mtx rcu_read_lock rcu_node_0 irq_context: 0 (wq_completion)events (work_completion)(&rdev->wiphy_work) &rdev->wiphy.mtx &wdev->mtx rcu_read_lock &rq->__lock irq_context: 0 (wq_completion)events (work_completion)(&rdev->wiphy_work) &rdev->wiphy.mtx &wdev->mtx rcu_read_lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 &f->f_pos_lock sb_writers#4 &sb->s_type->i_mutex_key#8 jbd2_handle rcu_read_lock &rq->__lock irq_context: 0 &f->f_pos_lock sb_writers#4 &sb->s_type->i_mutex_key#8 jbd2_handle rcu_read_lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: softirq &(&bat_priv->dat.work)->timer irq_context: softirq &(&bat_priv->dat.work)->timer rcu_read_lock &pool->lock/1 irq_context: softirq &(&bat_priv->dat.work)->timer rcu_read_lock &pool->lock/1 &obj_hash[i].lock irq_context: softirq &(&bat_priv->dat.work)->timer rcu_read_lock &pool->lock/1 &p->pi_lock irq_context: softirq &(&bat_priv->bla.work)->timer irq_context: softirq &(&bat_priv->bla.work)->timer rcu_read_lock &pool->lock/1 irq_context: softirq &(&bat_priv->bla.work)->timer rcu_read_lock &pool->lock/1 &obj_hash[i].lock irq_context: softirq &(&bat_priv->bla.work)->timer rcu_read_lock &pool->lock/1 &p->pi_lock irq_context: 0 &f->f_pos_lock sb_writers#4 &sb->s_type->i_mutex_key#8 &n->list_lock &c->lock irq_context: 0 &f->f_pos_lock sb_writers#4 &sb->s_type->i_mutex_key#8 mmu_notifier_invalidate_range_start &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 sk_lock-AF_ALG irq_context: 0 sk_lock-AF_ALG slock-AF_ALG irq_context: 0 slock-AF_ALG irq_context: 0 (wq_completion)bat_events (work_completion)(&(&bat_priv->dat.work)->work) irq_context: 0 (wq_completion)bat_events (work_completion)(&(&bat_priv->dat.work)->work) &hash->list_locks[i] irq_context: 0 (wq_completion)bat_events (work_completion)(&(&bat_priv->dat.work)->work) &rq->__lock irq_context: 0 (wq_completion)bat_events (work_completion)(&(&bat_priv->dat.work)->work) &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)bat_events (work_completion)(&(&bat_priv->dat.work)->work) &obj_hash[i].lock irq_context: 0 (wq_completion)bat_events (work_completion)(&(&bat_priv->dat.work)->work) &base->lock irq_context: 0 (wq_completion)bat_events (work_completion)(&(&bat_priv->dat.work)->work) &base->lock &obj_hash[i].lock irq_context: 0 &f->f_pos_lock sb_writers#4 &sb->s_type->i_mutex_key#8 jbd2_handle rcu_read_lock rcu_node_0 irq_context: 0 (wq_completion)bat_events (work_completion)(&(&bat_priv->bla.work)->work) irq_context: 0 (wq_completion)bat_events (work_completion)(&(&bat_priv->bla.work)->work) key#20 irq_context: 0 (wq_completion)bat_events (work_completion)(&(&bat_priv->bla.work)->work) rcu_read_lock &entry->crc_lock irq_context: 0 (wq_completion)bat_events (work_completion)(&(&bat_priv->bla.work)->work) rcu_read_lock pool_lock#2 irq_context: 0 (wq_completion)bat_events (work_completion)(&(&bat_priv->bla.work)->work) rcu_read_lock tk_core.seq.seqcount irq_context: 0 (wq_completion)bat_events (work_completion)(&(&bat_priv->bla.work)->work) rcu_read_lock &list->lock#5 irq_context: 0 (wq_completion)bat_events (work_completion)(&(&bat_priv->bla.work)->work) &obj_hash[i].lock irq_context: 0 (wq_completion)bat_events (work_completion)(&(&bat_priv->bla.work)->work) &base->lock irq_context: 0 (wq_completion)bat_events (work_completion)(&(&bat_priv->bla.work)->work) &base->lock &obj_hash[i].lock irq_context: 0 (wq_completion)bat_events (work_completion)(&(&bat_priv->bla.work)->work) &rq->__lock irq_context: 0 (wq_completion)bat_events (work_completion)(&(&bat_priv->bla.work)->work) &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 sb_writers#4 tomoyo_ss &rq->__lock irq_context: 0 sb_writers#4 tomoyo_ss &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 sb_writers#4 &sb->s_type->i_mutex_key#8 mapping.invalidate_lock rcu_node_0 irq_context: softirq &(&bat_priv->tt.work)->timer rcu_read_lock &pool->lock/1 &p->pi_lock &rq->__lock irq_context: softirq &(&bat_priv->tt.work)->timer rcu_read_lock &pool->lock/1 &p->pi_lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 sb_writers#4 &sb->s_type->i_mutex_key#8 mapping.invalidate_lock jbd2_handle &ei->i_data_sem &rq->__lock irq_context: 0 sb_writers#4 &sb->s_type->i_mutex_key#8 mapping.invalidate_lock jbd2_handle &ei->i_data_sem &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 sb_writers#4 &sb->s_type->i_mutex_key#8 mapping.invalidate_lock jbd2_handle &ei->i_data_sem &n->list_lock irq_context: 0 sb_writers#4 &sb->s_type->i_mutex_key#8 mapping.invalidate_lock jbd2_handle &ei->i_data_sem &n->list_lock &c->lock irq_context: 0 &sig->cred_guard_mutex tomoyo_ss &mm->mmap_lock &rq->__lock irq_context: 0 &sig->cred_guard_mutex tomoyo_ss &mm->mmap_lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)events (work_completion)(&(&nsim_dev->trap_data->trap_report_dw)->work) &devlink->lock_key#4 &rq->__lock irq_context: 0 (wq_completion)events (work_completion)(&(&nsim_dev->trap_data->trap_report_dw)->work) &devlink->lock_key#4 &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 &ei->i_data_sem &ei->i_prealloc_lock &pa->pa_lock#2 irq_context: 0 &ei->i_data_sem &obj_hash[i].lock irq_context: 0 &ei->i_data_sem pool_lock#2 irq_context: 0 &sb->s_type->i_mutex_key#10 sk_lock-AF_INET6/1 irq_context: 0 &sb->s_type->i_mutex_key#10 sk_lock-AF_INET6/1 slock-AF_INET6 irq_context: 0 &sb->s_type->i_mutex_key#10 sk_lock-AF_INET6/1 rlock-AF_INET6 irq_context: 0 &sb->s_type->i_mutex_key#10 sk_lock-AF_INET6/1 &list->lock#21 irq_context: 0 &sb->s_type->i_mutex_key#10 &net->sctp.addr_wq_lock irq_context: 0 &sb->s_type->i_mutex_key#10 &net->sctp.addr_wq_lock slock-AF_INET6/1 irq_context: 0 &sb->s_type->i_mutex_key#10 &net->sctp.addr_wq_lock slock-AF_INET6/1 &sctp_ep_hashtable[i].lock irq_context: 0 &sb->s_type->i_mutex_key#10 &net->sctp.addr_wq_lock slock-AF_INET6/1 &obj_hash[i].lock irq_context: 0 &sb->s_type->i_mutex_key#10 &net->sctp.addr_wq_lock slock-AF_INET6/1 pool_lock#2 irq_context: 0 &sb->s_type->i_mutex_key#10 &net->sctp.addr_wq_lock slock-AF_INET6/1 clock-AF_INET6 irq_context: 0 &sb->s_type->i_mutex_key#10 sk_lock-AF_INET/1 irq_context: 0 &sb->s_type->i_mutex_key#10 sk_lock-AF_INET/1 slock-AF_INET irq_context: 0 &sb->s_type->i_mutex_key#10 sk_lock-AF_INET/1 rlock-AF_INET irq_context: 0 &sb->s_type->i_mutex_key#10 sk_lock-AF_INET/1 &list->lock#21 irq_context: 0 &sb->s_type->i_mutex_key#10 &net->sctp.addr_wq_lock slock-AF_INET/1 irq_context: 0 &sb->s_type->i_mutex_key#10 &net->sctp.addr_wq_lock slock-AF_INET/1 &sctp_ep_hashtable[i].lock irq_context: 0 &sb->s_type->i_mutex_key#10 &net->sctp.addr_wq_lock slock-AF_INET/1 &obj_hash[i].lock irq_context: 0 &sb->s_type->i_mutex_key#10 &net->sctp.addr_wq_lock slock-AF_INET/1 pool_lock#2 irq_context: 0 &sb->s_type->i_mutex_key#10 &net->sctp.addr_wq_lock slock-AF_INET/1 clock-AF_INET irq_context: 0 &sb->s_type->i_mutex_key#10 &rq->__lock &cfs_rq->removed.lock irq_context: 0 sb_writers#4 &type->i_mutex_dir_key#3/1 &sb->s_type->i_mutex_key#8 batched_entropy_u8.lock irq_context: 0 sb_writers#4 &type->i_mutex_dir_key#3/1 &sb->s_type->i_mutex_key#8 kfence_freelist_lock irq_context: 0 sb_writers#4 &type->i_mutex_dir_key#3/1 &sb->s_type->i_mutex_key#8 &meta->lock irq_context: 0 sb_writers#4 &type->i_mutex_dir_key#3/1 &sb->s_type->i_mutex_key#8 &____s->seqcount#2 irq_context: 0 sb_writers#4 &type->i_mutex_dir_key#3/1 &sb->s_type->i_mutex_key#8 &____s->seqcount irq_context: 0 sb_writers#4 &xa->xa_lock#9 irq_context: 0 sb_writers#4 &mapping->private_lock irq_context: 0 sb_writers#4 &mapping->private_lock rcu_read_lock &memcg->move_lock irq_context: 0 sb_writers#4 stock_lock irq_context: 0 sb_writers#4 &sb->s_type->i_lock_key#22 &xa->xa_lock#9 irq_context: 0 sb_writers#4 lock#4 irq_context: 0 sb_writers#4 lock#5 irq_context: 0 sb_writers#4 &lruvec->lru_lock irq_context: 0 sb_writers#4 rcu_read_lock pool_lock#2 irq_context: 0 sb_writers#4 &sb->s_type->i_lock_key#22 &xa->xa_lock#9 &obj_hash[i].lock irq_context: 0 sb_writers#4 &sb->s_type->i_lock_key#22 &xa->xa_lock#9 pool_lock#2 irq_context: 0 sb_writers#4 sb_internal jbd2_handle &ei->i_data_sem irq_context: 0 sb_writers#4 sb_internal jbd2_handle &ei->i_data_sem &ei->i_prealloc_lock irq_context: 0 sb_writers#4 sb_internal jbd2_handle &ei->i_data_sem &ei->i_raw_lock irq_context: 0 sb_writers#4 sb_internal jbd2_handle &ei->i_data_sem &ei->i_es_lock irq_context: 0 sb_writers#4 sb_internal jbd2_handle &ei->i_data_sem &ei->i_es_lock &obj_hash[i].lock irq_context: 0 sb_writers#4 sb_internal jbd2_handle &ei->i_data_sem &ei->i_es_lock pool_lock#2 irq_context: 0 sb_writers#4 sb_internal jbd2_handle &ei->i_data_sem &ei->i_es_lock &sbi->s_es_lock irq_context: 0 sb_writers#4 sb_internal jbd2_handle &ei->i_data_sem mmu_notifier_invalidate_range_start irq_context: 0 sb_writers#4 sb_internal jbd2_handle &ei->i_data_sem pool_lock#2 irq_context: 0 sb_writers#4 sb_internal jbd2_handle &ei->i_data_sem &mapping->private_lock irq_context: 0 sb_writers#4 sb_internal jbd2_handle &ei->i_data_sem &bgl->locks[i].lock &sbi->s_md_lock irq_context: 0 sb_writers#4 sb_internal jbd2_handle &ei->i_data_sem &sb->s_type->i_lock_key#22 irq_context: 0 sb_writers#4 sb_internal jbd2_handle &ei->i_data_sem key#3 irq_context: 0 sb_writers#4 sb_internal jbd2_handle &ei->i_data_sem &obj_hash[i].lock irq_context: 0 sb_writers#4 sb_internal jbd2_handle tk_core.seq.seqcount irq_context: 0 (wq_completion)events (work_completion)(&rdev->wiphy_work) &rdev->wiphy.mtx &wdev->mtx quarantine_lock irq_context: 0 &sb->s_type->i_mutex_key#9 rcu_read_lock rcu_read_lock rcu_node_0 irq_context: 0 &sb->s_type->i_mutex_key#9 rcu_read_lock rcu_read_lock &rq->__lock irq_context: 0 &sb->s_type->i_mutex_key#9 rcu_read_lock rcu_read_lock &rq->__lock &cfs_rq->removed.lock irq_context: 0 &sb->s_type->i_mutex_key#9 rcu_read_lock rcu_read_lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 &sb->s_type->i_mutex_key#10 sk_lock-AF_INET &rq->__lock irq_context: 0 &sb->s_type->i_mutex_key#10 sk_lock-AF_INET &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 &f->f_pos_lock &type->i_mutex_dir_key#3 sb_writers#4 &____s->seqcount#2 irq_context: 0 &f->f_pos_lock &type->i_mutex_dir_key#3 sb_writers#4 &____s->seqcount irq_context: 0 sb_writers#4 &type->i_mutex_dir_key#3/1 jbd2_handle &ei->i_data_sem &ei->i_es_lock &____s->seqcount#2 irq_context: 0 sb_writers#4 &type->i_mutex_dir_key#3/1 jbd2_handle &ei->i_data_sem &ei->i_es_lock &____s->seqcount irq_context: 0 sb_writers#4 tomoyo_ss &____s->seqcount#2 irq_context: 0 sb_writers#4 tomoyo_ss &____s->seqcount irq_context: 0 &xt[i].mutex rcu_read_lock rcu_read_lock rcu_node_0 irq_context: 0 &xt[i].mutex rcu_read_lock rcu_read_lock &rq->__lock irq_context: 0 &xt[i].mutex rcu_read_lock rcu_read_lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 crypto_cfg_mutex irq_context: 0 crypto_cfg_mutex (console_sem).lock irq_context: 0 crypto_cfg_mutex console_lock console_srcu console_owner_lock irq_context: 0 crypto_cfg_mutex console_lock console_srcu console_owner irq_context: 0 crypto_cfg_mutex console_lock console_srcu console_owner &port_lock_key irq_context: 0 crypto_cfg_mutex console_lock console_srcu console_owner console_owner_lock irq_context: 0 crypto_cfg_mutex &rq->__lock irq_context: 0 crypto_cfg_mutex &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 &sb->s_type->i_mutex_key#10 sk_lock-AF_INET6 &rq->__lock irq_context: 0 &sb->s_type->i_mutex_key#10 sk_lock-AF_INET6 &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 &f->f_pos_lock sb_writers#4 &sb->s_type->i_mutex_key#8 &xa->xa_lock#9 &n->list_lock irq_context: 0 &f->f_pos_lock sb_writers#4 &sb->s_type->i_mutex_key#8 &xa->xa_lock#9 &n->list_lock &c->lock irq_context: 0 &f->f_pos_lock sb_writers#4 &sb->s_type->i_mutex_key#8 &cfs_rq->removed.lock irq_context: 0 sb_writers#4 &sb->s_type->i_mutex_key#8 mapping.invalidate_lock lock#4 &lruvec->lru_lock irq_context: 0 crypto_cfg_mutex crypto_alg_sem irq_context: 0 crypto_cfg_mutex crypto_alg_sem irq_context: 0 &nft_net->commit_mutex rcu_read_lock rcu_read_lock rcu_node_0 irq_context: 0 &nft_net->commit_mutex rcu_read_lock rcu_read_lock &rq->__lock irq_context: 0 &nft_net->commit_mutex rcu_read_lock rcu_node_0 irq_context: 0 &f->f_pos_lock sb_writers#4 &sb->s_type->i_mutex_key#8 &p->pi_lock &cfs_rq->removed.lock irq_context: 0 &f->f_pos_lock sb_writers#4 &sb->s_type->i_mutex_key#8 &xa->xa_lock#9 &____s->seqcount#2 irq_context: 0 &f->f_pos_lock sb_writers#4 &sb->s_type->i_mutex_key#8 &xa->xa_lock#9 &____s->seqcount irq_context: 0 &mm->mmap_lock ptlock_ptr(page)#2 &mapping->private_lock irq_context: 0 &mm->mmap_lock ptlock_ptr(page)#2 &mapping->private_lock rcu_read_lock &memcg->move_lock irq_context: 0 cb_lock (console_sem).lock irq_context: 0 cb_lock console_lock console_srcu console_owner_lock irq_context: 0 cb_lock console_lock console_srcu console_owner irq_context: 0 cb_lock console_lock console_srcu console_owner &port_lock_key irq_context: 0 (wq_completion)bat_events (work_completion)(&(&bat_priv->mcast.work)->work) rcu_read_lock &n->list_lock irq_context: 0 (wq_completion)bat_events (work_completion)(&(&bat_priv->mcast.work)->work) rcu_read_lock &n->list_lock &c->lock irq_context: 0 cb_lock console_lock console_srcu console_owner console_owner_lock irq_context: 0 sb_writers#4 &cfs_rq->removed.lock irq_context: 0 cb_lock rcu_read_lock &____s->seqcount irq_context: 0 cb_lock rcu_read_lock rcu_read_lock pool_lock#2 irq_context: 0 cb_lock rcu_read_lock &obj_hash[i].lock irq_context: 0 sb_writers#4 &type->i_mutex_dir_key#3/1 jbd2_handle &journal->j_state_lock irq_context: 0 sb_writers#4 &type->i_mutex_dir_key#3/1 jbd2_handle &journal->j_state_lock &journal->j_wait_commit irq_context: 0 sb_writers#4 &type->i_mutex_dir_key#3/1 jbd2_handle &journal->j_state_lock &journal->j_wait_commit &p->pi_lock irq_context: 0 sb_writers#4 &type->i_mutex_dir_key#3/1 jbd2_handle &journal->j_state_lock &journal->j_wait_commit &p->pi_lock &rq->__lock irq_context: 0 sb_writers#4 &type->i_mutex_dir_key#3/1 jbd2_handle &journal->j_state_lock &journal->j_wait_commit &p->pi_lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: softirq (&ndev->rs_timer) rcu_read_lock rcu_read_lock rcu_read_lock_bh rcu_read_lock_bh &obj_hash[i].lock irq_context: softirq (&ndev->rs_timer) rcu_read_lock rcu_read_lock rcu_read_lock_bh rcu_read_lock_bh pool_lock#2 irq_context: softirq (&ndev->rs_timer) rcu_read_lock rcu_read_lock rcu_read_lock_bh rcu_read_lock_bh rcu_read_lock tk_core.seq.seqcount irq_context: softirq (&ndev->rs_timer) rcu_read_lock rcu_read_lock rcu_read_lock_bh rcu_read_lock_bh rcu_read_lock &list->lock#5 irq_context: 0 sb_writers#4 &type->i_mutex_dir_key#3/1 &sb->s_type->i_mutex_key#8 jbd2_handle &journal->j_state_lock irq_context: 0 sb_writers#4 sb_internal jbd2_handle &journal->j_state_lock irq_context: 0 sb_writers#4 &mapping->private_lock rcu_read_lock rcu_read_lock key#10 irq_context: 0 sb_writers#4 sb_internal jbd2_handle &ei->i_es_lock &obj_hash[i].lock irq_context: 0 sb_writers#4 sb_internal jbd2_handle &ei->i_es_lock pool_lock#2 irq_context: 0 sb_writers#4 sb_internal jbd2_handle &ei->i_es_lock &sbi->s_es_lock irq_context: 0 sb_writers#4 sb_internal jbd2_handle &(ei->i_block_reservation_lock) irq_context: 0 sb_writers#4 sb_internal jbd2_handle &(ei->i_block_reservation_lock) key#14 irq_context: 0 sb_writers#4 sb_internal jbd2_handle &sb->s_type->i_lock_key#22 irq_context: 0 &f->f_pos_lock &type->i_mutex_dir_key#3 sb_writers#4 jbd2_handle &journal->j_state_lock irq_context: 0 sb_writers#4 &type->i_mutex_dir_key#3/1 &type->i_mutex_dir_key#3 jbd2_handle &journal->j_state_lock irq_context: 0 sb_writers#4 &type->i_mutex_dir_key#3/1 sb_internal jbd2_handle &journal->j_state_lock irq_context: 0 sb_writers#4 &type->i_mutex_dir_key#3/1 &sb->s_type->i_mutex_key#8 &journal->j_state_lock &journal->j_wait_transaction_locked irq_context: 0 sb_writers#4 &type->i_mutex_dir_key#3/1 &sb->s_type->i_mutex_key#8 jbd2_handle irq_context: 0 &journal->j_state_lock &journal->j_wait_transaction_locked &p->pi_lock irq_context: 0 &journal->j_state_lock &journal->j_wait_transaction_locked &p->pi_lock &rq->__lock irq_context: 0 &journal->j_state_lock &journal->j_wait_transaction_locked &p->pi_lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 sb_writers#4 &type->i_mutex_dir_key#3/1 &sb->s_type->i_mutex_key#8 jbd2_handle &____s->seqcount irq_context: 0 sb_writers#4 &type->i_mutex_dir_key#3/1 &sb->s_type->i_mutex_key#8 jbd2_handle bit_wait_table + i irq_context: 0 &mm->mmap_lock sb_pagefaults jbd2_handle &mapping->private_lock irq_context: 0 &mm->mmap_lock sb_pagefaults jbd2_handle &ret->b_state_lock irq_context: 0 &mm->mmap_lock sb_pagefaults jbd2_handle bit_wait_table + i irq_context: 0 sb_writers#4 &type->i_mutex_dir_key#3/1 jbd2_handle bit_wait_table + i irq_context: 0 cgroup_threadgroup_rwsem rcu_read_lock rcu_node_0 irq_context: 0 cgroup_threadgroup_rwsem rcu_read_lock &rq->__lock irq_context: 0 cgroup_threadgroup_rwsem rcu_read_lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 sb_writers#4 &type->i_mutex_dir_key#3/1 &type->i_mutex_dir_key#3 jbd2_handle &ret->b_state_lock irq_context: 0 sb_writers#4 &type->i_mutex_dir_key#3/1 jbd2_handle &bgl->locks[i].lock irq_context: 0 &fsnotify_mark_srcu &c->lock irq_context: 0 &ret->b_state_lock rcu_read_lock pool_lock#2 irq_context: 0 &ret->b_state_lock &obj_hash[i].lock irq_context: 0 &sb->s_type->i_mutex_key#8 &sbi->s_writepages_rwsem jbd2_handle rcu_read_lock rcu_read_lock &q->queue_lock &obj_hash[i].lock pool_lock irq_context: 0 cb_lock rcu_read_lock rcu_node_0 irq_context: 0 cb_lock rcu_read_lock &rq->__lock irq_context: 0 cb_lock rcu_read_lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 &sb->s_type->i_mutex_key#10 (netlink_chain).rwsem &nft_net->commit_mutex rcu_state.exp_mutex &rnp->exp_wq[3] irq_context: 0 (wq_completion)ext4-rsv-conversion (work_completion)(&ei->i_rsv_conversion_work) &folio_wait_table[i] &p->pi_lock &cfs_rq->removed.lock irq_context: 0 &mm->mmap_lock sb_pagefaults mapping.invalidate_lock rcu_read_lock &xa->xa_lock#9 rcu_read_lock per_cpu_ptr(&cgroup_rstat_cpu_lock, cpu) irq_context: 0 vlan_ioctl_mutex rtnl_mutex stock_lock rcu_read_lock per_cpu_ptr(&cgroup_rstat_cpu_lock, cpu) irq_context: 0 vlan_ioctl_mutex rtnl_mutex uevent_sock_mutex &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 &sb->s_type->i_mutex_key#8 lock#4 &lruvec->lru_lock per_cpu_ptr(&cgroup_rstat_cpu_lock, cpu) irq_context: 0 &sb->s_type->i_mutex_key#8 rcu_read_lock &base->lock irq_context: 0 &sb->s_type->i_mutex_key#8 rcu_read_lock &base->lock &obj_hash[i].lock irq_context: 0 (wq_completion)bat_events (work_completion)(&(&bat_priv->bla.work)->work) rcu_read_lock &c->lock irq_context: 0 &f->f_pos_lock sb_writers#4 &sb->s_type->i_mutex_key#8 &ei->i_data_sem &c->lock irq_context: 0 vlan_ioctl_mutex rtnl_mutex quarantine_lock irq_context: 0 vlan_ioctl_mutex rtnl_mutex remove_cache_srcu irq_context: 0 vlan_ioctl_mutex rtnl_mutex remove_cache_srcu quarantine_lock irq_context: 0 vlan_ioctl_mutex rtnl_mutex remove_cache_srcu &c->lock irq_context: 0 vlan_ioctl_mutex rtnl_mutex remove_cache_srcu &n->list_lock irq_context: 0 vlan_ioctl_mutex rtnl_mutex remove_cache_srcu rcu_read_lock pool_lock#2 irq_context: 0 vlan_ioctl_mutex rtnl_mutex remove_cache_srcu &obj_hash[i].lock irq_context: 0 vlan_ioctl_mutex rtnl_mutex uevent_sock_mutex &c->lock irq_context: 0 vlan_ioctl_mutex rtnl_mutex &obj_hash[i].lock pool_lock irq_context: 0 vlan_ioctl_mutex rtnl_mutex &idev->mc_lock &n->list_lock irq_context: 0 vlan_ioctl_mutex rtnl_mutex &idev->mc_lock &n->list_lock &c->lock irq_context: 0 vlan_ioctl_mutex rtnl_mutex lweventlist_lock &c->lock irq_context: 0 &mm->mmap_lock sb_writers#4 &c->lock irq_context: 0 &mm->mmap_lock sb_writers#4 &____s->seqcount#2 irq_context: 0 &mm->mmap_lock sb_writers#4 &____s->seqcount irq_context: 0 &f->f_pos_lock sb_writers#4 &sb->s_type->i_mutex_key#8 &mm->mmap_lock &obj_hash[i].lock irq_context: 0 &sb->s_type->i_mutex_key#10 (netlink_chain).rwsem rcu_read_lock &rcu_state.gp_wq irq_context: 0 &sb->s_type->i_mutex_key#10 (netlink_chain).rwsem rcu_read_lock &rcu_state.gp_wq &p->pi_lock irq_context: 0 &sb->s_type->i_mutex_key#10 (netlink_chain).rwsem rcu_read_lock &rcu_state.gp_wq &p->pi_lock &rq->__lock irq_context: 0 &sb->s_type->i_mutex_key#10 (netlink_chain).rwsem rcu_read_lock &rcu_state.gp_wq &p->pi_lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 &sb->s_type->i_mutex_key#10 (netlink_chain).rwsem rcu_read_lock &rcu_state.expedited_wq irq_context: 0 &sb->s_type->i_mutex_key#10 (netlink_chain).rwsem rcu_read_lock &rcu_state.expedited_wq &p->pi_lock irq_context: 0 &sb->s_type->i_mutex_key#10 (netlink_chain).rwsem rcu_read_lock &rcu_state.expedited_wq &p->pi_lock &rq->__lock irq_context: 0 &sb->s_type->i_mutex_key#10 (netlink_chain).rwsem rcu_read_lock &rcu_state.expedited_wq &p->pi_lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: softirq rcu_callback stock_lock rcu_read_lock per_cpu_ptr(&cgroup_rstat_cpu_lock, cpu) irq_context: 0 &sb->s_type->i_mutex_key#8 &sbi->s_writepages_rwsem jbd2_handle &ei->i_data_sem &lg->lg_mutex irq_context: 0 &sb->s_type->i_mutex_key#8 &sbi->s_writepages_rwsem jbd2_handle &ei->i_data_sem &lg->lg_mutex &ei->i_prealloc_lock irq_context: 0 &sb->s_type->i_mutex_key#8 &sbi->s_writepages_rwsem jbd2_handle &ei->i_data_sem &lg->lg_mutex rcu_read_lock &pa->pa_lock irq_context: 0 &sb->s_type->i_mutex_key#8 &sbi->s_writepages_rwsem jbd2_handle &ei->i_data_sem &lg->lg_mutex &mapping->private_lock irq_context: 0 &sb->s_type->i_mutex_key#8 &sbi->s_writepages_rwsem jbd2_handle &ei->i_data_sem &lg->lg_mutex mmu_notifier_invalidate_range_start irq_context: 0 &sb->s_type->i_mutex_key#8 &sbi->s_writepages_rwsem jbd2_handle &ei->i_data_sem &lg->lg_mutex &ret->b_state_lock irq_context: 0 &sb->s_type->i_mutex_key#8 &sbi->s_writepages_rwsem jbd2_handle &ei->i_data_sem &lg->lg_mutex &ret->b_state_lock &journal->j_list_lock irq_context: 0 &sb->s_type->i_mutex_key#8 &sbi->s_writepages_rwsem jbd2_handle &ei->i_data_sem &lg->lg_mutex &pa->pa_lock irq_context: 0 &sb->s_type->i_mutex_key#8 &sbi->s_writepages_rwsem jbd2_handle &ei->i_data_sem &lg->lg_mutex &lg->lg_prealloc_lock irq_context: 0 (wq_completion)bat_events (work_completion)(&(&forw_packet_aggr->delayed_work)->work) &obj_hash[i].lock pool_lock irq_context: 0 (wq_completion)bat_events (work_completion)(&(&forw_packet_aggr->delayed_work)->work) &hard_iface->bat_iv.ogm_buff_mutex &bat_priv->tt.commit_lock &c->lock irq_context: 0 sb_writers#4 &type->i_mutex_dir_key#3/1 &type->i_mutex_dir_key#3 jbd2_handle &ret->b_state_lock &journal->j_list_lock irq_context: 0 sb_writers#4 &type->i_mutex_dir_key#3 tomoyo_ss &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 sb_writers#5 &type->i_mutex_dir_key#5 tomoyo_ss rcu_read_lock &rq->__lock irq_context: 0 sb_writers#5 &type->i_mutex_dir_key#5 tomoyo_ss rcu_read_lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 sb_writers#4 lock#4 &lruvec->lru_lock irq_context: 0 sb_writers#4 sb_internal jbd2_handle &ei->i_data_sem &(ei->i_block_reservation_lock) irq_context: 0 (wq_completion)rcu_gp (work_completion)(&sdp->work) &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)ext4-rsv-conversion (work_completion)(&ei->i_rsv_conversion_work) jbd2_handle &ei->i_data_sem &n->list_lock irq_context: 0 (wq_completion)ext4-rsv-conversion (work_completion)(&ei->i_rsv_conversion_work) jbd2_handle &ei->i_data_sem &n->list_lock &c->lock irq_context: 0 &fsnotify_mark_srcu &____s->seqcount#2 irq_context: 0 &fsnotify_mark_srcu &____s->seqcount irq_context: 0 sb_writers#8 &of->mutex kn->active#5 batched_entropy_u8.lock irq_context: 0 sb_writers#8 &of->mutex kn->active#5 kfence_freelist_lock irq_context: 0 sb_writers#8 &of->mutex kn->active#5 &meta->lock irq_context: 0 kn->active#5 &kernfs_locks->open_file_mutex[count] fs_reclaim mmu_notifier_invalidate_range_start &rq->__lock irq_context: 0 kn->active#5 &kernfs_locks->open_file_mutex[count] fs_reclaim mmu_notifier_invalidate_range_start &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 &sb->s_type->i_mutex_key#8 &sbi->s_writepages_rwsem remove_cache_srcu irq_context: 0 &sb->s_type->i_mutex_key#8 &sbi->s_writepages_rwsem remove_cache_srcu &rq->__lock irq_context: 0 &sb->s_type->i_mutex_key#8 &sbi->s_writepages_rwsem remove_cache_srcu &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 &sb->s_type->i_mutex_key#8 &sbi->s_writepages_rwsem remove_cache_srcu quarantine_lock irq_context: 0 &sb->s_type->i_mutex_key#8 &sbi->s_writepages_rwsem remove_cache_srcu &c->lock irq_context: 0 &sb->s_type->i_mutex_key#8 &sbi->s_writepages_rwsem remove_cache_srcu &n->list_lock irq_context: 0 &sb->s_type->i_mutex_key#8 &sbi->s_writepages_rwsem remove_cache_srcu rcu_read_lock pool_lock#2 irq_context: 0 &sb->s_type->i_mutex_key#8 &sbi->s_writepages_rwsem remove_cache_srcu &obj_hash[i].lock irq_context: 0 sb_writers#4 &type->i_mutex_dir_key#3/1 jbd2_handle &ei->i_data_sem &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 sb_writers#8 tomoyo_ss rcu_read_lock rcu_node_0 irq_context: 0 sb_writers#8 tomoyo_ss rcu_read_lock &rq->__lock irq_context: 0 sb_writers#8 tomoyo_ss rcu_read_lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 namespace_sem stock_lock irq_context: 0 namespace_sem &____s->seqcount#2 irq_context: 0 namespace_sem rcu_read_lock rcu_read_lock rcu_node_0 irq_context: 0 namespace_sem rcu_read_lock rcu_read_lock &rq->__lock irq_context: 0 namespace_sem rcu_read_lock rcu_read_lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 &sb->s_type->i_mutex_key#8 &sbi->s_writepages_rwsem jbd2_handle rcu_read_lock &rq->__lock irq_context: 0 &sb->s_type->i_mutex_key#8 &sbi->s_writepages_rwsem jbd2_handle rcu_read_lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 sb_writers#4 &sb->s_type->i_mutex_key#8 &sbi->s_writepages_rwsem rcu_state.exp_mutex &rnp->exp_wq[0] irq_context: 0 namespace_sem rcu_read_lock rcu_node_0 irq_context: 0 namespace_sem rcu_read_lock &rcu_state.gp_wq irq_context: 0 namespace_sem rcu_read_lock &rcu_state.gp_wq &p->pi_lock irq_context: 0 namespace_sem rcu_read_lock &rcu_state.gp_wq &p->pi_lock &rq->__lock irq_context: 0 namespace_sem rcu_read_lock &rcu_state.gp_wq &p->pi_lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 namespace_sem rcu_read_lock &rcu_state.expedited_wq irq_context: 0 namespace_sem rcu_read_lock &rcu_state.expedited_wq &p->pi_lock irq_context: 0 namespace_sem rcu_read_lock &rcu_state.expedited_wq &p->pi_lock &rq->__lock irq_context: 0 namespace_sem rcu_read_lock &rcu_state.expedited_wq &p->pi_lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 namespace_sem rcu_read_lock &rq->__lock irq_context: 0 namespace_sem rcu_read_lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 &type->s_umount_key#23/1 &xa->xa_lock#4 irq_context: 0 &type->s_umount_key#23/1 &xa->xa_lock#4 pool_lock#2 irq_context: 0 &type->s_umount_key#23/1 &obj_hash[i].lock irq_context: 0 &type->s_umount_key#23/1 stock_lock irq_context: 0 pernet_ops_rwsem stock_lock irq_context: 0 pernet_ops_rwsem k-sk_lock-AF_NETLINK rcu_read_lock rcu_node_0 irq_context: 0 pernet_ops_rwsem k-sk_lock-AF_NETLINK rcu_read_lock &rq->__lock irq_context: 0 pernet_ops_rwsem k-sk_lock-AF_NETLINK rcu_read_lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 pernet_ops_rwsem &____s->seqcount#2 irq_context: 0 pernet_ops_rwsem uevent_sock_mutex rcu_read_lock &ei->socket.wq.wait &ep->lock &ep->wq irq_context: 0 pernet_ops_rwsem uevent_sock_mutex rcu_read_lock &ei->socket.wq.wait &ep->lock &ep->wq &p->pi_lock irq_context: 0 pernet_ops_rwsem uevent_sock_mutex rcu_read_lock &ei->socket.wq.wait &ep->lock &ep->wq &p->pi_lock &rq->__lock irq_context: 0 pernet_ops_rwsem uevent_sock_mutex rcu_read_lock &ei->socket.wq.wait &ep->lock &ep->wq &p->pi_lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 pernet_ops_rwsem uevent_sock_mutex rcu_read_lock rcu_node_0 irq_context: 0 pernet_ops_rwsem uevent_sock_mutex rcu_read_lock &rq->__lock irq_context: 0 pernet_ops_rwsem uevent_sock_mutex rcu_read_lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 &sb->s_type->i_mutex_key#8 &cfs_rq->removed.lock irq_context: 0 pernet_ops_rwsem uevent_sock_mutex nl_table_wait.lock &p->pi_lock irq_context: 0 pernet_ops_rwsem uevent_sock_mutex nl_table_wait.lock &p->pi_lock &rq->__lock irq_context: 0 pernet_ops_rwsem uevent_sock_mutex nl_table_wait.lock &p->pi_lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 pernet_ops_rwsem uevent_sock_mutex &rq->__lock irq_context: 0 pernet_ops_rwsem uevent_sock_mutex &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 &sb->s_type->i_mutex_key#10 (netlink_chain).rwsem batched_entropy_u8.lock irq_context: 0 &sb->s_type->i_mutex_key#10 (netlink_chain).rwsem kfence_freelist_lock irq_context: 0 (wq_completion)events (work_completion)(&w->w) &meta->lock irq_context: 0 (wq_completion)events (work_completion)(&w->w) kfence_freelist_lock irq_context: softirq (&ndev->rs_timer) rcu_read_lock rcu_read_lock rcu_read_lock_bh rcu_read_lock_bh rcu_read_lock &pool->lock &p->pi_lock &rq->__lock irq_context: softirq (&ndev->rs_timer) rcu_read_lock rcu_read_lock rcu_read_lock_bh rcu_read_lock_bh rcu_read_lock &pool->lock &p->pi_lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 &sb->s_type->i_mutex_key#10 fill_pool_map-wait-type-override &rq->__lock irq_context: 0 &sb->s_type->i_mutex_key#10 fill_pool_map-wait-type-override &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 sb_writers#4 &type->i_mutex_dir_key#3/1 &sb->s_type->i_mutex_key#8 jbd2_handle &sbi->s_orphan_lock &rq->__lock irq_context: 0 sb_writers#4 &type->i_mutex_dir_key#3/1 &sb->s_type->i_mutex_key#8 jbd2_handle &sbi->s_orphan_lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 rtnl_mutex qdisc_mod_lock irq_context: 0 rtnl_mutex &block->lock irq_context: 0 rtnl_mutex &block->cb_lock irq_context: 0 rtnl_mutex &block->cb_lock flow_indr_block_lock irq_context: 0 rtnl_mutex &block->cb_lock flow_indr_block_lock fs_reclaim irq_context: 0 rtnl_mutex &block->cb_lock flow_indr_block_lock fs_reclaim mmu_notifier_invalidate_range_start irq_context: 0 rtnl_mutex &block->cb_lock flow_indr_block_lock pool_lock#2 irq_context: 0 rtnl_mutex &block->cb_lock flow_indr_block_lock &c->lock irq_context: 0 rtnl_mutex &dev->tx_global_lock _xmit_ETHER#2 irq_context: 0 pernet_ops_rwsem nf_hook_mutex stock_lock irq_context: 0 pernet_ops_rwsem rcu_read_lock rcu_read_lock rcu_node_0 irq_context: 0 pernet_ops_rwsem rcu_read_lock rcu_read_lock &rq->__lock irq_context: 0 pernet_ops_rwsem rcu_read_lock rcu_read_lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 &sb->s_type->i_mutex_key#8 &sbi->s_writepages_rwsem jbd2_handle &ei->i_data_sem &n->list_lock irq_context: 0 &sb->s_type->i_mutex_key#8 &sbi->s_writepages_rwsem jbd2_handle &ei->i_data_sem &n->list_lock &c->lock irq_context: 0 pernet_ops_rwsem rcu_read_lock rcu_node_0 irq_context: 0 lock link_idr_lock irq_context: 0 lock link_idr_lock pool_lock#2 irq_context: 0 tracepoints_mutex irq_context: 0 tracepoints_mutex fs_reclaim irq_context: 0 tracepoints_mutex fs_reclaim mmu_notifier_invalidate_range_start irq_context: 0 tracepoints_mutex pool_lock#2 irq_context: 0 tracepoints_mutex cpu_hotplug_lock irq_context: 0 tracepoints_mutex cpu_hotplug_lock &rq->__lock irq_context: 0 tracepoints_mutex cpu_hotplug_lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 &sb->s_type->i_mutex_key#8 &sbi->s_writepages_rwsem jbd2_handle &ei->i_data_sem &(ei->i_block_reservation_lock) key#14 irq_context: 0 tracepoints_mutex cpu_hotplug_lock static_call_mutex irq_context: 0 tracepoints_mutex cpu_hotplug_lock static_call_mutex text_mutex irq_context: 0 tracepoints_mutex cpu_hotplug_lock static_call_mutex text_mutex ptlock_ptr(page)#2 irq_context: 0 tracepoints_mutex cpu_hotplug_lock jump_label_mutex irq_context: 0 tracepoints_mutex cpu_hotplug_lock jump_label_mutex text_mutex irq_context: 0 tracepoints_mutex cpu_hotplug_lock jump_label_mutex text_mutex ptlock_ptr(page)#2 irq_context: 0 link_idr_lock irq_context: 0 pernet_ops_rwsem rcu_read_lock &rq->__lock irq_context: 0 pernet_ops_rwsem rcu_read_lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 sk_lock-AF_INET6 rcu_read_lock &____s->seqcount#7 irq_context: 0 sk_lock-AF_INET6 rcu_read_lock rcu_read_lock &____s->seqcount#7 irq_context: 0 sk_lock-AF_INET6 rcu_read_lock &nf_nat_locks[i] irq_context: 0 sk_lock-AF_INET6 rcu_read_lock &rq->__lock irq_context: 0 sk_lock-AF_INET6 rcu_read_lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 sk_lock-AF_INET6 rcu_read_lock &nf_conntrack_locks[i] irq_context: 0 sk_lock-AF_INET6 rcu_read_lock &nf_conntrack_locks[i] batched_entropy_u8.lock irq_context: 0 sk_lock-AF_INET6 rcu_read_lock rcu_read_lock_bh tk_core.seq.seqcount irq_context: 0 sk_lock-AF_INET6 rcu_read_lock rcu_read_lock_bh &list->lock#5 irq_context: softirq rcu_read_lock rcu_read_lock k-slock-AF_INET6 rcu_read_lock &n->list_lock irq_context: softirq rcu_read_lock rcu_read_lock k-slock-AF_INET6 rcu_read_lock &n->list_lock &c->lock irq_context: 0 rtnl_mutex _xmit_ETHER &obj_hash[i].lock irq_context: 0 rtnl_mutex _xmit_ETHER krc.lock irq_context: 0 &sb->s_type->i_mutex_key#8 &____s->seqcount#2 irq_context: 0 rtnl_mutex rcu_read_lock &tb->tb6_lock &net->ipv6.fib6_walker_lock irq_context: 0 rtnl_mutex rcu_read_lock &tb->tb6_lock pool_lock#2 irq_context: 0 rtnl_mutex rcu_read_lock &tb->tb6_lock &c->lock irq_context: 0 sk_lock-AF_INET6 rcu_read_lock &rcu_state.gp_wq irq_context: 0 rtnl_mutex rcu_read_lock &tb->tb6_lock nl_table_lock irq_context: 0 sk_lock-AF_INET6 rcu_read_lock &rcu_state.gp_wq &p->pi_lock irq_context: 0 rtnl_mutex rcu_read_lock &tb->tb6_lock &obj_hash[i].lock irq_context: 0 sk_lock-AF_INET6 rcu_read_lock &rcu_state.gp_wq &p->pi_lock &rq->__lock irq_context: 0 rtnl_mutex rcu_read_lock &tb->tb6_lock &____s->seqcount irq_context: 0 sk_lock-AF_INET6 rcu_read_lock &rcu_state.gp_wq &p->pi_lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 rtnl_mutex rcu_read_lock &tb->tb6_lock rcu_read_lock pool_lock#2 irq_context: 0 rtnl_mutex rcu_read_lock &tb->tb6_lock nl_table_wait.lock irq_context: 0 rtnl_mutex rcu_read_lock &tb->tb6_lock rcu_read_lock &data->fib_event_queue_lock irq_context: 0 rtnl_mutex rcu_read_lock &tb->tb6_lock rcu_read_lock rcu_read_lock &pool->lock irq_context: 0 rtnl_mutex rcu_read_lock &tb->tb6_lock rcu_read_lock rcu_read_lock &pool->lock &obj_hash[i].lock irq_context: 0 rtnl_mutex rcu_read_lock &tb->tb6_lock rcu_read_lock rcu_read_lock &pool->lock pool_lock#2 irq_context: 0 rtnl_mutex rcu_read_lock &tb->tb6_lock rcu_read_lock rcu_read_lock &pool->lock &p->pi_lock irq_context: 0 rtnl_mutex rcu_read_lock &tb->tb6_lock rcu_read_lock rcu_read_lock &pool->lock &p->pi_lock &rq->__lock irq_context: 0 rtnl_mutex rcu_read_lock &tb->tb6_lock rcu_read_lock rcu_read_lock &pool->lock &p->pi_lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 &type->i_mutex_dir_key#3 &____s->seqcount#2 irq_context: 0 &type->i_mutex_dir_key#3 rcu_read_lock &dentry->d_lock irq_context: 0 rtnl_mutex &tbl->lock &c->lock irq_context: 0 rtnl_mutex &tbl->lock pool_lock#2 irq_context: 0 rtnl_mutex &tbl->lock &n->lock irq_context: 0 rtnl_mutex &tbl->lock &n->lock &____s->seqcount#9 irq_context: 0 rtnl_mutex &tbl->lock nl_table_lock irq_context: 0 rtnl_mutex &tbl->lock &obj_hash[i].lock irq_context: 0 rtnl_mutex &tbl->lock nl_table_wait.lock irq_context: 0 rtnl_mutex &tbl->lock rcu_read_lock lock#8 irq_context: 0 rtnl_mutex &tbl->lock rcu_read_lock id_table_lock irq_context: 0 rtnl_mutex &tbl->lock &dir->lock#2 irq_context: 0 rtnl_mutex &tbl->lock krc.lock irq_context: 0 rtnl_mutex &net->ipv6.addrconf_hash_lock &obj_hash[i].lock irq_context: 0 (wq_completion)events (work_completion)(&aux->work) rtnl_mutex irq_context: 0 rtnl_mutex &ndev->lock &base->lock irq_context: 0 rtnl_mutex &ndev->lock &base->lock &obj_hash[i].lock irq_context: 0 (wq_completion)events (work_completion)(&aux->work) rtnl_mutex &rq->__lock irq_context: 0 (wq_completion)events (work_completion)(&aux->work) rtnl_mutex &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)events (work_completion)(&aux->work) rtnl_mutex rtnl_mutex.wait_lock irq_context: 0 (wq_completion)events (work_completion)(&aux->work) rtnl_mutex &pool->lock irq_context: 0 rtnl_mutex &idev->mc_lock _xmit_ETHER krc.lock irq_context: 0 sb_writers#4 &sb->s_type->i_mutex_key#8 &sbi->s_writepages_rwsem &rnp->exp_lock irq_context: 0 sb_writers#4 &sb->s_type->i_mutex_key#8 &sbi->s_writepages_rwsem rcu_state.exp_mutex irq_context: 0 sb_writers#4 &sb->s_type->i_mutex_key#8 &sbi->s_writepages_rwsem rcu_state.exp_mutex rcu_state.exp_mutex.wait_lock irq_context: 0 (wq_completion)wg-crypt-wg1#6 (work_completion)(&({ do { const void *__vpp_verify = (typeof((worker) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __asm__ ("" : "=r"(__ptr) : "0"((typeof(*((worker))) *)((worker)))); (typeof((typeof(*((worker))) *)((worker)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); })->work) rcu_read_lock &pool->lock &p->pi_lock &rq->__lock irq_context: 0 (wq_completion)wg-crypt-wg1#6 (work_completion)(&({ do { const void *__vpp_verify = (typeof((worker) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __asm__ ("" : "=r"(__ptr) : "0"((typeof(*((worker))) *)((worker)))); (typeof((typeof(*((worker))) *)((worker)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); })->work) rcu_read_lock &pool->lock &p->pi_lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)wg-crypt-wg1#6 (work_completion)(&peer->transmit_packet_work) &base->lock irq_context: 0 (wq_completion)wg-crypt-wg1#6 (work_completion)(&peer->transmit_packet_work) &base->lock &obj_hash[i].lock irq_context: 0 (wq_completion)wg-crypt-wg1#6 (work_completion)(&peer->transmit_packet_work) &peer->endpoint_lock rcu_read_lock_bh rcu_read_lock &dir->lock#2 irq_context: 0 (wq_completion)wg-crypt-wg1#6 (work_completion)(&peer->transmit_packet_work) &peer->endpoint_lock rcu_read_lock_bh rcu_read_lock &ul->lock irq_context: 0 (wq_completion)wg-crypt-wg1#6 (work_completion)(&peer->transmit_packet_work) &peer->endpoint_lock rcu_read_lock_bh batched_entropy_u32.lock irq_context: 0 sk_lock-AF_INET6 rcu_read_lock &rcu_state.expedited_wq irq_context: 0 sk_lock-AF_INET6 rcu_read_lock &rcu_state.expedited_wq &p->pi_lock irq_context: 0 sk_lock-AF_INET6 rcu_read_lock &rcu_state.expedited_wq &p->pi_lock &rq->__lock irq_context: 0 sk_lock-AF_INET6 rcu_read_lock &rcu_state.expedited_wq &p->pi_lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (crypto_chain).rwsem &____s->seqcount#2 irq_context: 0 sb_writers#5 &sb->s_type->i_mutex_key#12 &cfs_rq->removed.lock irq_context: 0 (wq_completion)ext4-rsv-conversion (work_completion)(&ei->i_rsv_conversion_work) &cfs_rq->removed.lock irq_context: 0 &sb->s_type->i_mutex_key#8 &rq->__lock &cfs_rq->removed.lock irq_context: 0 (wq_completion)rcu_gp (work_completion)(&rew->rew_work) &cfs_rq->removed.lock irq_context: 0 (wq_completion)rcu_gp (work_completion)(&rew->rew_work) pool_lock#2 irq_context: 0 rtnl_mutex rcu_state.exp_mutex.wait_lock irq_context: 0 sb_writers#4 &sb->s_type->i_mutex_key#8 &sbi->s_writepages_rwsem rcu_state.exp_mutex &rnp->exp_wq[1] irq_context: 0 rtnl_mutex &idev->mc_query_lock irq_context: 0 rtnl_mutex &idev->mc_query_lock &obj_hash[i].lock irq_context: 0 rtnl_mutex (work_completion)(&(&idev->mc_report_work)->work) irq_context: 0 rtnl_mutex &br->lock pool_lock#2 irq_context: 0 rtnl_mutex &br->lock &dir->lock#2 irq_context: 0 rtnl_mutex &br->lock deferred_lock irq_context: 0 rtnl_mutex &br->lock rcu_read_lock &pool->lock irq_context: 0 rtnl_mutex &br->lock rcu_read_lock &pool->lock &obj_hash[i].lock irq_context: 0 rtnl_mutex &br->lock rcu_read_lock &pool->lock &p->pi_lock irq_context: 0 rtnl_mutex &br->lock rcu_read_lock &pool->lock &p->pi_lock &rq->__lock irq_context: 0 rtnl_mutex &br->lock rcu_read_lock &pool->lock &p->pi_lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 rtnl_mutex &br->lock (console_sem).lock irq_context: 0 rtnl_mutex &br->lock console_lock console_srcu console_owner_lock irq_context: 0 rtnl_mutex &br->lock console_lock console_srcu console_owner irq_context: 0 rtnl_mutex &br->lock console_lock console_srcu console_owner &port_lock_key irq_context: 0 rtnl_mutex &br->lock console_lock console_srcu console_owner console_owner_lock irq_context: 0 rtnl_mutex &br->lock nl_table_lock irq_context: 0 rtnl_mutex &br->lock nl_table_wait.lock irq_context: 0 rtnl_mutex &br->lock &br->multicast_lock irq_context: 0 rtnl_mutex &br->lock &br->multicast_lock pool_lock#2 irq_context: 0 rtnl_mutex &br->lock &br->multicast_lock &dir->lock#2 irq_context: 0 rtnl_mutex &br->lock &br->multicast_lock deferred_lock irq_context: 0 rtnl_mutex &br->lock &br->multicast_lock nl_table_lock irq_context: 0 rtnl_mutex &br->lock &br->multicast_lock &obj_hash[i].lock irq_context: 0 rtnl_mutex &br->lock &br->multicast_lock nl_table_wait.lock irq_context: 0 rtnl_mutex &br->lock &br->multicast_lock rcu_read_lock &pool->lock irq_context: 0 rtnl_mutex &br->lock &br->multicast_lock rcu_read_lock &pool->lock &obj_hash[i].lock irq_context: 0 rtnl_mutex &br->lock &br->multicast_lock rcu_read_lock &pool->lock pool_lock#2 irq_context: 0 rtnl_mutex &br->lock &br->multicast_lock rcu_read_lock &pool->lock &p->pi_lock irq_context: 0 rtnl_mutex &br->lock &br->multicast_lock &base->lock irq_context: 0 rtnl_mutex &br->lock &br->multicast_lock &base->lock &obj_hash[i].lock irq_context: softirq (&mp->timer) irq_context: softirq (&mp->timer) &br->multicast_lock irq_context: softirq (&mp->timer) &br->multicast_lock rcu_read_lock rhashtable_bucket irq_context: softirq (&mp->timer) &br->multicast_lock rcu_read_lock rcu_read_lock &pool->lock irq_context: softirq (&mp->timer) &br->multicast_lock rcu_read_lock rcu_read_lock &pool->lock &obj_hash[i].lock irq_context: softirq (&mp->timer) &br->multicast_lock rcu_read_lock rcu_read_lock &pool->lock pool_lock#2 irq_context: softirq (&mp->timer) &br->multicast_lock rcu_read_lock rcu_read_lock &pool->lock &p->pi_lock irq_context: 0 (wq_completion)events_long (work_completion)(&br->mcast_gc_work) irq_context: 0 (wq_completion)events_long (work_completion)(&br->mcast_gc_work) &br->multicast_lock irq_context: 0 (wq_completion)events_long (work_completion)(&br->mcast_gc_work) (&mp->timer) irq_context: 0 (wq_completion)events_long (work_completion)(&br->mcast_gc_work) &obj_hash[i].lock irq_context: 0 (wq_completion)events_long (work_completion)(&br->mcast_gc_work) &base->lock irq_context: 0 (wq_completion)events_long (work_completion)(&br->mcast_gc_work) pool_lock#2 irq_context: 0 (wq_completion)events_long (work_completion)(&br->mcast_gc_work) krc.lock irq_context: 0 (wq_completion)events_long (work_completion)(&br->mcast_gc_work) (&p->rexmit_timer) irq_context: 0 (wq_completion)events_long (work_completion)(&br->mcast_gc_work) (&p->timer) irq_context: 0 (wq_completion)events_long (work_completion)(&br->mcast_gc_work) &base->lock &obj_hash[i].lock irq_context: 0 rtnl_mutex &rnp->exp_lock irq_context: 0 rtnl_mutex rcu_state.exp_mutex irq_context: 0 rtnl_mutex rcu_state.exp_mutex rcu_state.exp_mutex.wait_lock irq_context: 0 sb_writers#4 &sb->s_type->i_mutex_key#8 &sbi->s_writepages_rwsem rcu_state.exp_mutex.wait_lock irq_context: 0 sb_writers#4 &sb->s_type->i_mutex_key#8 &sbi->s_writepages_rwsem &p->pi_lock irq_context: 0 sb_writers#4 &sb->s_type->i_mutex_key#8 &sbi->s_writepages_rwsem &p->pi_lock &cfs_rq->removed.lock irq_context: 0 sb_writers#4 &sb->s_type->i_mutex_key#8 &sbi->s_writepages_rwsem &p->pi_lock &rq->__lock irq_context: 0 sb_writers#4 &sb->s_type->i_mutex_key#8 &sbi->s_writepages_rwsem &p->pi_lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 sb_writers#4 &sb->s_type->i_mutex_key#8 &sbi->s_writepages_rwsem &rq->__lock irq_context: 0 sb_writers#4 &sb->s_type->i_mutex_key#8 &sbi->s_writepages_rwsem &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 rtnl_mutex &tbl->lock &____s->seqcount irq_context: 0 rtnl_mutex rcu_read_lock batched_entropy_u8.lock irq_context: 0 rtnl_mutex rcu_read_lock kfence_freelist_lock irq_context: 0 sb_writers#4 &sb->s_type->i_mutex_key#8 &sbi->s_writepages_rwsem &cfs_rq->removed.lock irq_context: softirq rcu_read_lock rcu_read_lock k-slock-AF_INET6 rcu_read_lock &____s->seqcount#2 irq_context: softirq rcu_read_lock rcu_read_lock k-slock-AF_INET6 rcu_read_lock &____s->seqcount irq_context: 0 sb_writers#4 quarantine_lock irq_context: 0 sk_lock-AF_INET6 remove_cache_srcu irq_context: 0 sk_lock-AF_INET6 remove_cache_srcu quarantine_lock irq_context: 0 sk_lock-AF_INET6 remove_cache_srcu &c->lock irq_context: 0 sk_lock-AF_INET6 remove_cache_srcu &n->list_lock irq_context: 0 sk_lock-AF_INET6 remove_cache_srcu rcu_read_lock pool_lock#2 irq_context: 0 sk_lock-AF_INET6 remove_cache_srcu &obj_hash[i].lock irq_context: 0 sk_lock-AF_INET6 remove_cache_srcu &rq->__lock irq_context: 0 sk_lock-AF_INET6 remove_cache_srcu &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 sb_writers#4 sb_internal jbd2_handle &ei->i_data_sem &ei->i_prealloc_lock &pa->pa_lock#2 irq_context: 0 sb_writers#4 sb_internal jbd2_handle &ei->i_data_sem &(ei->i_block_reservation_lock) key#14 irq_context: 0 sb_writers#8 kn->active#5 &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 sb_writers#8 kn->active#5 &rq->__lock &cfs_rq->removed.lock irq_context: 0 (wq_completion)events (work_completion)(&w->work)#2 &meta->lock irq_context: 0 (wq_completion)events (work_completion)(&w->work)#2 kfence_freelist_lock irq_context: softirq (&ndev->rs_timer) rcu_read_lock rcu_read_lock rcu_read_lock_bh dev->qdisc_tx_busylock ?: &qdisc_tx_busylock#2 rcu_read_lock_bh &obj_hash[i].lock irq_context: softirq (&ndev->rs_timer) rcu_read_lock rcu_read_lock rcu_read_lock_bh dev->qdisc_tx_busylock ?: &qdisc_tx_busylock#2 rcu_read_lock_bh pool_lock#2 irq_context: softirq (&ndev->rs_timer) rcu_read_lock rcu_read_lock rcu_read_lock_bh dev->qdisc_tx_busylock ?: &qdisc_tx_busylock#2 rcu_read_lock_bh rcu_read_lock tk_core.seq.seqcount irq_context: softirq (&ndev->rs_timer) rcu_read_lock rcu_read_lock rcu_read_lock_bh dev->qdisc_tx_busylock ?: &qdisc_tx_busylock#2 rcu_read_lock_bh rcu_read_lock &list->lock#5 irq_context: softirq drivers/net/wireguard/ratelimiter.c:20 rcu_read_lock &pool->lock &p->pi_lock &rq->__lock irq_context: softirq drivers/net/wireguard/ratelimiter.c:20 rcu_read_lock &pool->lock &p->pi_lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 pernet_ops_rwsem &net->sctp.addr_wq_lock k-slock-AF_INET6/1 irq_context: 0 pernet_ops_rwsem &net->sctp.addr_wq_lock k-slock-AF_INET6/1 &sctp_ep_hashtable[i].lock irq_context: 0 pernet_ops_rwsem &net->sctp.addr_wq_lock k-slock-AF_INET6/1 &obj_hash[i].lock irq_context: 0 pernet_ops_rwsem &net->sctp.addr_wq_lock k-slock-AF_INET6/1 pool_lock#2 irq_context: 0 pernet_ops_rwsem &net->sctp.addr_wq_lock k-slock-AF_INET6/1 k-clock-AF_INET6 irq_context: 0 (wq_completion)events_unbound (work_completion)(&sub_info->work) remove_cache_srcu irq_context: 0 (wq_completion)events_unbound (work_completion)(&sub_info->work) remove_cache_srcu quarantine_lock irq_context: 0 (wq_completion)events_unbound (work_completion)(&sub_info->work) remove_cache_srcu &c->lock irq_context: 0 (wq_completion)events_unbound (work_completion)(&sub_info->work) remove_cache_srcu &n->list_lock irq_context: 0 (wq_completion)events_unbound (work_completion)(&sub_info->work) remove_cache_srcu rcu_read_lock pool_lock#2 irq_context: 0 (wq_completion)events_unbound (work_completion)(&sub_info->work) remove_cache_srcu &obj_hash[i].lock irq_context: 0 &f->f_pos_lock sb_writers#4 &sb->s_type->i_mutex_key#8 jbd2_handle &mapping->private_lock irq_context: 0 (wq_completion)bat_events (work_completion)(&(&forw_packet_aggr->delayed_work)->work) &n->list_lock irq_context: 0 (wq_completion)bat_events (work_completion)(&(&forw_packet_aggr->delayed_work)->work) &n->list_lock &c->lock irq_context: 0 rtnl_mutex dev_base_lock &xa->xa_lock#3 irq_context: 0 rtnl_mutex cpu_hotplug_lock &list->lock#5 irq_context: 0 &f->f_pos_lock sb_writers#4 &sb->s_type->i_mutex_key#8 rcu_read_lock &rcu_state.expedited_wq irq_context: 0 &f->f_pos_lock sb_writers#4 &sb->s_type->i_mutex_key#8 rcu_read_lock &rcu_state.expedited_wq &p->pi_lock irq_context: 0 &f->f_pos_lock sb_writers#4 &sb->s_type->i_mutex_key#8 rcu_read_lock &rcu_state.expedited_wq &p->pi_lock &rq->__lock irq_context: 0 &f->f_pos_lock sb_writers#4 &sb->s_type->i_mutex_key#8 rcu_read_lock &rcu_state.expedited_wq &p->pi_lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 rtnl_mutex &block->cb_lock flow_indr_block_lock &obj_hash[i].lock irq_context: 0 rtnl_mutex &block->cb_lock flow_indr_block_lock &rq->__lock irq_context: 0 rtnl_mutex &block->cb_lock flow_indr_block_lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 rtnl_mutex bpf_devs_lock irq_context: 0 rtnl_mutex &hwstats->hwsdev_list_lock irq_context: 0 rtnl_mutex &ul->lock irq_context: 0 rtnl_mutex &net->xdp.lock irq_context: 0 rtnl_mutex &net->xdp.lock &rq->__lock irq_context: 0 rtnl_mutex &net->xdp.lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 rtnl_mutex mirred_list_lock irq_context: 0 sb_writers#5 &type->i_mutex_dir_key#5 mmu_notifier_invalidate_range_start &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 rtnl_mutex &idev->mc_lock krc.lock irq_context: 0 rtnl_mutex &idev->mc_report_lock irq_context: 0 rtnl_mutex &root->kernfs_rwsem &____s->seqcount irq_context: 0 rtnl_mutex &root->kernfs_rwsem rcu_read_lock pool_lock#2 irq_context: 0 (wq_completion)wg-crypt-wg2#5 (work_completion)(&peer->transmit_packet_work) &peer->endpoint_lock rcu_read_lock_bh &obj_hash[i].lock irq_context: 0 sk_lock-AF_INET6 &____s->seqcount#2 irq_context: 0 sb_writers#4 &sb->s_type->i_mutex_key#8 mapping.invalidate_lock jbd2_handle &mapping->private_lock irq_context: 0 &tsk->futex_exit_mutex &mm->mmap_lock &rq->__lock irq_context: 0 &tsk->futex_exit_mutex &mm->mmap_lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 &tsk->futex_exit_mutex &mm->mmap_lock &cfs_rq->removed.lock irq_context: 0 &tsk->futex_exit_mutex &mm->mmap_lock &obj_hash[i].lock irq_context: 0 rtnl_mutex &br->lock &c->lock irq_context: 0 sb_writers#4 &sb->s_type->i_mutex_key#8 rcu_read_lock mount_lock.seqcount irq_context: 0 sb_writers#4 &sb->s_type->i_mutex_key#8 rcu_read_lock rcu_read_lock rename_lock.seqcount irq_context: 0 sb_writers#4 &sb->s_type->i_mutex_key#8 ima_extend_list_mutex irq_context: 0 sb_writers#4 &sb->s_type->i_mutex_key#8 ima_extend_list_mutex fs_reclaim irq_context: 0 sb_writers#4 &sb->s_type->i_mutex_key#8 ima_extend_list_mutex fs_reclaim mmu_notifier_invalidate_range_start irq_context: 0 sb_writers#4 &sb->s_type->i_mutex_key#8 ima_extend_list_mutex fs_reclaim mmu_notifier_invalidate_range_start &rq->__lock irq_context: 0 sb_writers#4 &sb->s_type->i_mutex_key#8 ima_extend_list_mutex fs_reclaim mmu_notifier_invalidate_range_start &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 sb_writers#4 &sb->s_type->i_mutex_key#8 ima_extend_list_mutex pool_lock#2 irq_context: 0 sb_writers#4 &sb->s_type->i_mutex_key#8 &p->alloc_lock irq_context: 0 sb_writers#4 &sb->s_type->i_mutex_key#8 &list->lock irq_context: 0 sb_writers#4 &sb->s_type->i_mutex_key#8 kauditd_wait.lock irq_context: 0 &f->f_pos_lock &type->i_mutex_dir_key#3 sb_writers#4 jbd2_handle &mapping->private_lock irq_context: 0 sb_writers#4 sb_internal jbd2_handle &mapping->private_lock irq_context: 0 (wq_completion)bat_events (work_completion)(&(&forw_packet_aggr->delayed_work)->work) &hard_iface->bat_iv.ogm_buff_mutex &bat_priv->tt.commit_lock &bat_priv->softif_vlan_list_lock &c->lock irq_context: 0 rtnl_mutex (&pmctx->ip6_mc_router_timer) irq_context: 0 rtnl_mutex (&pmctx->ip4_mc_router_timer) irq_context: 0 rtnl_mutex (work_completion)(&ht->run_work) irq_context: 0 rtnl_mutex &ht->mutex irq_context: 0 rtnl_mutex &ht->mutex &obj_hash[i].lock irq_context: 0 rtnl_mutex &ht->mutex pool_lock#2 irq_context: 0 rtnl_mutex &br->multicast_lock irq_context: 0 rtnl_mutex &pnn->pndevs.lock irq_context: 0 rtnl_mutex &pnn->routes.lock irq_context: 0 rtnl_mutex dev_pm_qos_sysfs_mtx irq_context: 0 rtnl_mutex dev_pm_qos_sysfs_mtx &root->kernfs_rwsem irq_context: 0 rtnl_mutex dev_pm_qos_sysfs_mtx &root->kernfs_rwsem irq_context: 0 rtnl_mutex dev_pm_qos_sysfs_mtx dev_pm_qos_mtx irq_context: 0 rtnl_mutex subsys mutex#17 &k->k_lock klist_remove_lock irq_context: 0 rtnl_mutex deferred_probe_mutex irq_context: 0 rtnl_mutex device_links_lock irq_context: 0 rtnl_mutex uevent_sock_mutex mmu_notifier_invalidate_range_start irq_context: 0 sb_writers#4 &type->i_mutex_dir_key#3/1 jbd2_handle &meta_group_info[i]->alloc_sem &bgl->locks[i].lock irq_context: 0 rtnl_mutex (inetaddr_chain).rwsem &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 rtnl_mutex (inetaddr_chain).rwsem fib_info_lock &obj_hash[i].lock irq_context: 0 rtnl_mutex (inetaddr_chain).rwsem fib_info_lock pool_lock#2 irq_context: 0 rtnl_mutex (inetaddr_chain).rwsem &tbl->lock irq_context: 0 rtnl_mutex (inetaddr_chain).rwsem class irq_context: 0 rtnl_mutex (inetaddr_chain).rwsem (&tbl->proxy_timer) irq_context: 0 rtnl_mutex (inetaddr_chain).rwsem &base->lock irq_context: 0 rtnl_mutex (inetaddr_chain).rwsem krc.lock irq_context: 0 sb_writers#8 kn->active#5 &____s->seqcount#2 irq_context: 0 sb_writers#8 kn->active#5 &____s->seqcount irq_context: 0 (wq_completion)wg-crypt-wg1#2 (work_completion)(&({ do { const void *__vpp_verify = (typeof((worker) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __asm__ ("" : "=r"(__ptr) : "0"((typeof(*((worker))) *)((worker)))); (typeof((typeof(*((worker))) *)((worker)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); })->work) rcu_read_lock &pool->lock &p->pi_lock &rq->__lock irq_context: 0 (wq_completion)wg-crypt-wg1#2 (work_completion)(&({ do { const void *__vpp_verify = (typeof((worker) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __asm__ ("" : "=r"(__ptr) : "0"((typeof(*((worker))) *)((worker)))); (typeof((typeof(*((worker))) *)((worker)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); })->work) rcu_read_lock &pool->lock &p->pi_lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)wg-crypt-wg1#2 (work_completion)(&peer->transmit_packet_work) &base->lock irq_context: 0 (wq_completion)wg-crypt-wg1#2 (work_completion)(&peer->transmit_packet_work) &base->lock &obj_hash[i].lock irq_context: 0 &sb->s_type->i_mutex_key#8 &sem->wait_lock irq_context: 0 rtnl_mutex remove_cache_srcu rcu_read_lock &rcu_state.gp_wq irq_context: 0 rtnl_mutex remove_cache_srcu rcu_read_lock &rcu_state.gp_wq &p->pi_lock irq_context: 0 rtnl_mutex remove_cache_srcu rcu_read_lock &rcu_state.gp_wq &p->pi_lock &rq->__lock irq_context: 0 rtnl_mutex remove_cache_srcu rcu_read_lock &rcu_state.gp_wq &p->pi_lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: softirq rcu_read_lock rcu_read_lock k-slock-AF_INET6 rcu_read_lock batched_entropy_u8.lock irq_context: softirq rcu_read_lock rcu_read_lock k-slock-AF_INET6 rcu_read_lock kfence_freelist_lock irq_context: softirq rcu_read_lock rcu_read_lock &meta->lock irq_context: 0 rtnl_mutex &root->kernfs_rwsem rcu_read_lock rcu_node_0 irq_context: 0 rtnl_mutex uevent_sock_mutex nl_table_wait.lock &p->pi_lock irq_context: 0 rtnl_mutex uevent_sock_mutex nl_table_wait.lock &p->pi_lock &rq->__lock irq_context: 0 rtnl_mutex uevent_sock_mutex nl_table_wait.lock &p->pi_lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 sk_lock-AF_INET6 rcu_read_lock &cfs_rq->removed.lock irq_context: 0 &f->f_pos_lock sb_writers#4 &p->pi_lock &cfs_rq->removed.lock irq_context: 0 &f->f_pos_lock sb_writers#4 &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 pernet_ops_rwsem rtnl_mutex &____s->seqcount#2 irq_context: 0 pernet_ops_rwsem rtnl_mutex stock_lock irq_context: 0 pernet_ops_rwsem rtnl_mutex lock kernfs_idr_lock &____s->seqcount#2 irq_context: 0 pernet_ops_rwsem rtnl_mutex lock kernfs_idr_lock &pcp->lock &zone->lock irq_context: 0 pernet_ops_rwsem rtnl_mutex lock kernfs_idr_lock &pcp->lock &zone->lock &____s->seqcount irq_context: 0 pernet_ops_rwsem rtnl_mutex lock kernfs_idr_lock &____s->seqcount irq_context: 0 rcu_state.barrier_mutex rcu_state.barrier_lock &obj_hash[i].lock irq_context: softirq (&ndev->rs_timer) rcu_read_lock rcu_read_lock rcu_read_lock_bh _xmit_ETHER#2 rcu_read_lock tk_core.seq.seqcount irq_context: softirq (&ndev->rs_timer) rcu_read_lock rcu_read_lock rcu_read_lock_bh _xmit_ETHER#2 rcu_read_lock &fq->lock irq_context: softirq (&ndev->rs_timer) rcu_read_lock rcu_read_lock rcu_read_lock_bh _xmit_ETHER#2 rcu_read_lock &local->active_txq_lock[i] irq_context: softirq (&ndev->rs_timer) rcu_read_lock rcu_read_lock rcu_read_lock_bh _xmit_ETHER#2 rcu_read_lock &local->handle_wake_tx_queue_lock irq_context: softirq (&ndev->rs_timer) rcu_read_lock rcu_read_lock rcu_read_lock_bh _xmit_ETHER#2 rcu_read_lock &local->handle_wake_tx_queue_lock &local->active_txq_lock[i] irq_context: softirq (&ndev->rs_timer) rcu_read_lock rcu_read_lock rcu_read_lock_bh _xmit_ETHER#2 rcu_read_lock &local->handle_wake_tx_queue_lock &local->queue_stop_reason_lock irq_context: softirq (&ndev->rs_timer) rcu_read_lock rcu_read_lock rcu_read_lock_bh _xmit_ETHER#2 rcu_read_lock &local->handle_wake_tx_queue_lock &fq->lock irq_context: softirq (&ndev->rs_timer) rcu_read_lock rcu_read_lock rcu_read_lock_bh _xmit_ETHER#2 rcu_read_lock &local->handle_wake_tx_queue_lock &fq->lock tk_core.seq.seqcount irq_context: softirq (&ndev->rs_timer) rcu_read_lock rcu_read_lock rcu_read_lock_bh _xmit_ETHER#2 rcu_read_lock &local->handle_wake_tx_queue_lock tk_core.seq.seqcount irq_context: softirq (&ndev->rs_timer) rcu_read_lock rcu_read_lock rcu_read_lock_bh _xmit_ETHER#2 rcu_read_lock &local->handle_wake_tx_queue_lock hwsim_radio_lock irq_context: softirq (&ndev->rs_timer) rcu_read_lock rcu_read_lock rcu_read_lock_bh _xmit_ETHER#2 rcu_read_lock &local->handle_wake_tx_queue_lock hwsim_radio_lock pool_lock#2 irq_context: softirq (&ndev->rs_timer) rcu_read_lock rcu_read_lock rcu_read_lock_bh _xmit_ETHER#2 rcu_read_lock &local->handle_wake_tx_queue_lock hwsim_radio_lock &list->lock#19 irq_context: softirq (&ndev->rs_timer) rcu_read_lock rcu_read_lock rcu_read_lock_bh _xmit_ETHER#2 rcu_read_lock &local->handle_wake_tx_queue_lock &list->lock#19 irq_context: 0 (wq_completion)events (work_completion)(&aux->work) rtnl_mutex &pool->lock &p->pi_lock irq_context: 0 &mm->mmap_lock &lruvec->lru_lock per_cpu_ptr(&cgroup_rstat_cpu_lock, cpu) irq_context: 0 (wq_completion)wg-crypt-wg2#6 (work_completion)(&peer->transmit_packet_work) &base->lock irq_context: 0 (wq_completion)wg-crypt-wg2#6 (work_completion)(&peer->transmit_packet_work) &base->lock &obj_hash[i].lock irq_context: 0 (wq_completion)wg-crypt-wg2#6 (work_completion)(&peer->transmit_packet_work) &peer->endpoint_lock rcu_read_lock_bh &obj_hash[i].lock irq_context: 0 (wq_completion)wg-crypt-wg2#6 (work_completion)(&peer->transmit_packet_work) &peer->endpoint_lock rcu_read_lock_bh batched_entropy_u32.lock irq_context: 0 (wq_completion)wg-crypt-wg1#3 (work_completion)(&peer->transmit_packet_work) &base->lock irq_context: 0 (wq_completion)wg-crypt-wg1#3 (work_completion)(&peer->transmit_packet_work) &base->lock &obj_hash[i].lock irq_context: 0 (wq_completion)wg-crypt-wg1#3 (work_completion)(&peer->transmit_packet_work) &peer->endpoint_lock rcu_read_lock_bh rcu_read_lock &dir->lock#2 irq_context: 0 (wq_completion)wg-crypt-wg1#3 (work_completion)(&peer->transmit_packet_work) &peer->endpoint_lock rcu_read_lock_bh rcu_read_lock &ul->lock irq_context: 0 (wq_completion)wg-crypt-wg2#4 (work_completion)(&peer->transmit_packet_work) &base->lock irq_context: 0 (wq_completion)wg-crypt-wg2#4 (work_completion)(&peer->transmit_packet_work) &base->lock &obj_hash[i].lock irq_context: 0 (wq_completion)wg-crypt-wg2#4 (work_completion)(&peer->transmit_packet_work) &peer->endpoint_lock rcu_read_lock_bh rcu_read_lock &c->lock irq_context: 0 (wq_completion)wg-crypt-wg2#4 (work_completion)(&peer->transmit_packet_work) &peer->endpoint_lock rcu_read_lock_bh rcu_read_lock &dir->lock#2 irq_context: 0 (wq_completion)wg-crypt-wg2#4 (work_completion)(&peer->transmit_packet_work) &peer->endpoint_lock rcu_read_lock_bh rcu_read_lock &ul->lock irq_context: 0 (wq_completion)wg-crypt-wg1#5 (work_completion)(&peer->transmit_packet_work) &base->lock irq_context: 0 (wq_completion)wg-crypt-wg1#5 (work_completion)(&peer->transmit_packet_work) &base->lock &obj_hash[i].lock irq_context: 0 dev_base_lock irq_context: 0 lweventlist_lock irq_context: 0 lweventlist_lock pool_lock#2 irq_context: 0 lweventlist_lock &dir->lock#2 irq_context: 0 krc.lock irq_context: 0 &dir->lock#2 irq_context: 0 &dir->lock#2 &obj_hash[i].lock irq_context: 0 &dir->lock#2 pool_lock#2 irq_context: 0 netdev_unregistering_wq.lock irq_context: 0 sb_writers#4 sb_internal jbd2_handle &sbi->s_orphan_lock &ei->i_raw_lock irq_context: 0 sb_writers#4 sb_internal jbd2_handle &bgl->locks[i].lock irq_context: 0 sb_writers#4 &type->i_mutex_dir_key#3/1 jbd2_handle &ei->i_data_sem &bgl->locks[i].lock irq_context: 0 map_idr_lock &obj_hash[i].lock irq_context: 0 map_idr_lock pool_lock#2 irq_context: 0 (wq_completion)events_unbound (work_completion)(&map->work) cgroup_mutex irq_context: 0 (wq_completion)events_unbound (work_completion)(&map->work) stock_lock irq_context: 0 pernet_ops_rwsem rtnl_mutex uevent_sock_mutex uevent_sock_mutex.wait_lock irq_context: 0 (wq_completion)events (work_completion)(&rdev->wiphy_work) &rdev->wiphy.mtx quarantine_lock irq_context: 0 &xt[i].mutex remove_cache_srcu irq_context: 0 &xt[i].mutex remove_cache_srcu quarantine_lock irq_context: 0 &xt[i].mutex remove_cache_srcu &c->lock irq_context: 0 &xt[i].mutex remove_cache_srcu &n->list_lock irq_context: 0 &xt[i].mutex remove_cache_srcu rcu_read_lock pool_lock#2 irq_context: 0 &xt[i].mutex remove_cache_srcu &obj_hash[i].lock irq_context: 0 &ret->b_state_lock &journal->j_list_lock &obj_hash[i].lock irq_context: 0 &ret->b_state_lock &journal->j_list_lock pool_lock#2 irq_context: 0 &wb->list_lock &sb->s_type->i_lock_key#22 irq_context: 0 sk_lock-AF_INET6 batched_entropy_u8.lock irq_context: 0 sk_lock-AF_INET6 kfence_freelist_lock irq_context: 0 (wq_completion)wg-crypt-wg0#3 (work_completion)(&({ do { const void *__vpp_verify = (typeof((worker) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __asm__ ("" : "=r"(__ptr) : "0"((typeof(*((worker))) *)((worker)))); (typeof((typeof(*((worker))) *)((worker)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); })->work) rcu_read_lock &pool->lock &p->pi_lock &rq->__lock &cfs_rq->removed.lock irq_context: 0 sb_writers#8 &of->mutex kn->active#5 uevent_sock_mutex.wait_lock irq_context: 0 sb_writers#8 &of->mutex kn->active#5 &p->pi_lock irq_context: 0 sb_writers#8 &of->mutex kn->active#5 &p->pi_lock &rq->__lock irq_context: 0 sb_writers#8 &of->mutex kn->active#5 &p->pi_lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)events (work_completion)(&aux->work) rtnl_mutex bpf_devs_lock irq_context: 0 (wq_completion)events (work_completion)(&aux->work) rtnl_mutex.wait_lock irq_context: 0 (wq_completion)events (work_completion)(&aux->work) &p->pi_lock irq_context: 0 (wq_completion)events (work_completion)(&aux->work) &p->pi_lock &rq->__lock irq_context: 0 (wq_completion)events (work_completion)(&aux->work) &p->pi_lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 &sb->s_type->i_mutex_key#8 &sbi->s_writepages_rwsem &journal->j_state_lock irq_context: 0 &sb->s_type->i_mutex_key#8 &sbi->s_writepages_rwsem &journal->j_state_lock tk_core.seq.seqcount irq_context: 0 &sb->s_type->i_mutex_key#8 &sbi->s_writepages_rwsem &journal->j_state_lock &obj_hash[i].lock irq_context: 0 &sb->s_type->i_mutex_key#8 &sbi->s_writepages_rwsem &journal->j_state_lock &base->lock irq_context: 0 &sb->s_type->i_mutex_key#8 &sbi->s_writepages_rwsem &journal->j_state_lock &base->lock &obj_hash[i].lock irq_context: 0 sb_writers#4 sb_internal jbd2_handle &ret->b_state_lock irq_context: 0 sb_writers#4 sb_internal jbd2_handle &ret->b_state_lock &journal->j_list_lock irq_context: 0 pernet_ops_rwsem rtnl_mutex uevent_sock_mutex &____s->seqcount#2 irq_context: 0 (wq_completion)bat_events (work_completion)(&(&forw_packet_aggr->delayed_work)->work) &hard_iface->bat_iv.ogm_buff_mutex rcu_read_lock fill_pool_map-wait-type-override pool_lock#2 irq_context: 0 (wq_completion)bat_events (work_completion)(&(&forw_packet_aggr->delayed_work)->work) &hard_iface->bat_iv.ogm_buff_mutex rcu_read_lock fill_pool_map-wait-type-override pool_lock irq_context: 0 cb_lock genl_mutex rcu_read_lock &n->list_lock irq_context: 0 cb_lock genl_mutex rcu_read_lock &n->list_lock &c->lock irq_context: softirq (&ndev->rs_timer) rcu_read_lock rcu_read_lock rcu_read_lock_bh dev->qdisc_tx_busylock ?: &qdisc_tx_busylock#2 rcu_read_lock_bh &obj_hash[i].lock pool_lock irq_context: 0 (wq_completion)wg-crypt-wg0#5 (work_completion)(&peer->transmit_packet_work) &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)rcu_gp (work_completion)(&sdp->work) rcu_node_0 irq_context: 0 (wq_completion)rcu_gp (work_completion)(&sdp->work) &rcu_state.expedited_wq irq_context: 0 link_idr_lock &obj_hash[i].lock irq_context: 0 link_idr_lock pool_lock#2 irq_context: 0 tracepoints_mutex tracepoint_srcu_srcu_usage.lock irq_context: 0 tracepoints_mutex tracepoint_srcu_srcu_usage.lock &obj_hash[i].lock irq_context: 0 tracepoints_mutex tracepoint_srcu_srcu_usage.lock &ACCESS_PRIVATE(sdp, lock) irq_context: 0 tracepoints_mutex tracepoint_srcu_srcu_usage.lock &base->lock irq_context: 0 tracepoints_mutex tracepoint_srcu_srcu_usage.lock &base->lock &obj_hash[i].lock irq_context: 0 tracepoints_mutex &rq->__lock irq_context: 0 tracepoints_mutex &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 tracepoints_mutex &obj_hash[i].lock irq_context: 0 (wq_completion)rcu_gp (work_completion)(&(&ssp->srcu_sup->work)->work) &ssp->srcu_sup->srcu_gp_mutex tracepoint_srcu_srcu_usage.lock irq_context: 0 (wq_completion)rcu_gp (work_completion)(&(&ssp->srcu_sup->work)->work) &ssp->srcu_sup->srcu_gp_mutex &ssp->srcu_sup->srcu_cb_mutex tracepoint_srcu_srcu_usage.lock irq_context: 0 (wq_completion)rcu_gp (work_completion)(&(&ssp->srcu_sup->work)->work) &ssp->srcu_sup->srcu_cb_mutex &obj_hash[i].lock irq_context: 0 (wq_completion)rcu_gp (work_completion)(&(&ssp->srcu_sup->work)->work) &ssp->srcu_sup->srcu_cb_mutex &base->lock irq_context: 0 (wq_completion)rcu_gp (work_completion)(&(&ssp->srcu_sup->work)->work) &ssp->srcu_sup->srcu_cb_mutex &base->lock &obj_hash[i].lock irq_context: 0 (wq_completion)rcu_gp (work_completion)(&(&ssp->srcu_sup->work)->work) tracepoint_srcu_srcu_usage.lock irq_context: 0 pernet_ops_rwsem rtnl_mutex &cfs_rq->removed.lock irq_context: 0 pernet_ops_rwsem krc.lock irq_context: 0 cb_lock &rdev->wiphy.mtx &n->list_lock irq_context: 0 cb_lock &rdev->wiphy.mtx &n->list_lock &c->lock irq_context: 0 (wq_completion)bat_events (work_completion)(&(&forw_packet_aggr->delayed_work)->work) &hard_iface->bat_iv.ogm_buff_mutex &bat_priv->tt.commit_lock &bat_priv->softif_vlan_list_lock &n->list_lock irq_context: 0 (wq_completion)bat_events (work_completion)(&(&forw_packet_aggr->delayed_work)->work) &hard_iface->bat_iv.ogm_buff_mutex &bat_priv->tt.commit_lock &bat_priv->softif_vlan_list_lock &n->list_lock &c->lock irq_context: softirq (&sdp->delay_work) irq_context: softirq (&sdp->delay_work) rcu_read_lock &pool->lock irq_context: softirq (&sdp->delay_work) rcu_read_lock &pool->lock &obj_hash[i].lock irq_context: softirq (&sdp->delay_work) rcu_read_lock &pool->lock &p->pi_lock irq_context: softirq (&sdp->delay_work) rcu_read_lock &pool->lock &p->pi_lock &rq->__lock irq_context: softirq (&sdp->delay_work) rcu_read_lock &pool->lock &p->pi_lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: softirq rcu_callback tracepoint_srcu_srcu_usage.lock &ACCESS_PRIVATE(sdp, lock) irq_context: softirq rcu_callback tracepoint_srcu_srcu_usage.lock &obj_hash[i].lock irq_context: softirq rcu_callback tracepoint_srcu_srcu_usage.lock &base->lock irq_context: softirq rcu_callback tracepoint_srcu_srcu_usage.lock &base->lock &obj_hash[i].lock irq_context: 0 pernet_ops_rwsem rcu_read_lock &tn->node_list_lock irq_context: 0 pernet_ops_rwsem ebt_mutex irq_context: 0 pernet_ops_rwsem &xt[i].mutex irq_context: 0 pernet_ops_rwsem &nft_net->commit_mutex irq_context: 0 pernet_ops_rwsem netns_bpf_mutex irq_context: 0 pernet_ops_rwsem rcu_state.exp_mutex rcu_node_0 irq_context: 0 pernet_ops_rwsem rcu_state.exp_mutex &obj_hash[i].lock irq_context: 0 pernet_ops_rwsem rcu_state.exp_mutex rcu_read_lock &pool->lock irq_context: 0 pernet_ops_rwsem rcu_state.exp_mutex rcu_read_lock &pool->lock &obj_hash[i].lock irq_context: 0 pernet_ops_rwsem rcu_state.exp_mutex rcu_read_lock &pool->lock &p->pi_lock irq_context: 0 pernet_ops_rwsem rcu_state.exp_mutex rcu_read_lock &pool->lock &p->pi_lock &rq->__lock irq_context: 0 pernet_ops_rwsem rcu_state.exp_mutex rcu_read_lock &pool->lock &p->pi_lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 pernet_ops_rwsem rcu_state.exp_mutex &rnp->exp_wq[1] irq_context: 0 pernet_ops_rwsem rcu_state.exp_mutex &rq->__lock irq_context: 0 pernet_ops_rwsem rcu_state.exp_mutex &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 sb_writers#4 rcu_node_0 irq_context: 0 sb_writers#4 rcu_read_lock rcu_read_lock rcu_node_0 irq_context: 0 sb_writers#4 rcu_read_lock rcu_read_lock &rq->__lock irq_context: 0 sb_writers#4 rcu_read_lock rcu_read_lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 pernet_ops_rwsem rtnl_mutex rcu_state.exp_mutex rcu_node_0 irq_context: 0 pernet_ops_rwsem rtnl_mutex rcu_state.exp_mutex &obj_hash[i].lock irq_context: 0 pernet_ops_rwsem rtnl_mutex rcu_state.exp_mutex rcu_read_lock &pool->lock irq_context: 0 pernet_ops_rwsem rtnl_mutex rcu_state.exp_mutex rcu_read_lock &pool->lock &obj_hash[i].lock irq_context: 0 pernet_ops_rwsem rtnl_mutex rcu_state.exp_mutex rcu_read_lock &pool->lock &p->pi_lock irq_context: 0 pernet_ops_rwsem rtnl_mutex rcu_state.exp_mutex rcu_read_lock &pool->lock &p->pi_lock &rq->__lock irq_context: 0 pernet_ops_rwsem rtnl_mutex rcu_state.exp_mutex rcu_read_lock &pool->lock &p->pi_lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 pernet_ops_rwsem rtnl_mutex rcu_state.exp_mutex &rnp->exp_wq[2] irq_context: 0 pernet_ops_rwsem rtnl_mutex rcu_state.exp_mutex &rq->__lock irq_context: 0 pernet_ops_rwsem rtnl_mutex rcu_state.exp_mutex &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 sb_writers#4 sb_internal jbd2_handle &cfs_rq->removed.lock irq_context: 0 &sb->s_type->i_mutex_key#10 &hashinfo->lock#2 irq_context: 0 pernet_ops_rwsem rtnl_mutex dev_base_lock &xa->xa_lock#3 irq_context: 0 pernet_ops_rwsem rtnl_mutex cpu_hotplug_lock irq_context: 0 pernet_ops_rwsem rtnl_mutex cpu_hotplug_lock &list->lock#5 irq_context: 0 pernet_ops_rwsem rtnl_mutex rcu_state.exp_mutex &rnp->exp_wq[3] irq_context: 0 (wq_completion)rcu_gp (work_completion)(&(&ssp->srcu_sup->work)->work) &ssp->srcu_sup->srcu_gp_mutex &ssp->srcu_sup->srcu_cb_mutex &rq->__lock irq_context: 0 (wq_completion)rcu_gp (work_completion)(&(&ssp->srcu_sup->work)->work) &ssp->srcu_sup->srcu_gp_mutex &ssp->srcu_sup->srcu_cb_mutex &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)events (work_completion)(&(&hwstats->traffic_dw)->work) &rq->__lock irq_context: 0 (wq_completion)rcu_gp (work_completion)(&sdp->work) pool_lock#2 irq_context: 0 pernet_ops_rwsem rtnl_mutex rcu_state.exp_mutex.wait_lock irq_context: 0 pernet_ops_rwsem rtnl_mutex bpf_devs_lock irq_context: 0 pernet_ops_rwsem rtnl_mutex &in_dev->mc_tomb_lock irq_context: 0 pernet_ops_rwsem rtnl_mutex sysctl_lock &obj_hash[i].lock irq_context: 0 pernet_ops_rwsem rtnl_mutex sysctl_lock pool_lock#2 irq_context: 0 sb_writers#8 tomoyo_ss rcu_read_lock rcu_read_lock rcu_node_0 irq_context: 0 pernet_ops_rwsem rtnl_mutex sysctl_lock krc.lock irq_context: 0 sb_writers#8 tomoyo_ss rcu_read_lock rcu_read_lock &rq->__lock irq_context: 0 sb_writers#8 tomoyo_ss rcu_read_lock rcu_read_lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 pernet_ops_rwsem rtnl_mutex class irq_context: 0 pernet_ops_rwsem rtnl_mutex (&tbl->proxy_timer) irq_context: 0 pernet_ops_rwsem rtnl_mutex &base->lock irq_context: 0 pernet_ops_rwsem rtnl_mutex &ul->lock irq_context: 0 pernet_ops_rwsem rtnl_mutex &net->xdp.lock irq_context: 0 pernet_ops_rwsem rtnl_mutex krc.lock irq_context: 0 pernet_ops_rwsem rtnl_mutex mirred_list_lock irq_context: 0 pernet_ops_rwsem rtnl_mutex &nft_net->commit_mutex irq_context: 0 pernet_ops_rwsem rtnl_mutex rcu_read_lock &tb->tb6_lock irq_context: 0 pernet_ops_rwsem rtnl_mutex rcu_read_lock &tb->tb6_lock &net->ipv6.fib6_walker_lock irq_context: 0 pernet_ops_rwsem rtnl_mutex &ent->pde_unload_lock irq_context: 0 pernet_ops_rwsem rtnl_mutex &net->ipv6.addrconf_hash_lock irq_context: 0 pernet_ops_rwsem rtnl_mutex &ndev->lock irq_context: 0 pernet_ops_rwsem rtnl_mutex &ndev->lock &obj_hash[i].lock irq_context: 0 pernet_ops_rwsem rtnl_mutex &rnp->exp_lock irq_context: 0 pernet_ops_rwsem rtnl_mutex &rnp->exp_wq[0] irq_context: 0 sb_writers#8 &of->mutex kn->active#5 rcu_node_0 irq_context: 0 sb_writers#8 &of->mutex kn->active#5 &rcu_state.expedited_wq irq_context: 0 sb_writers#8 &of->mutex kn->active#5 &rcu_state.expedited_wq &p->pi_lock irq_context: 0 sb_writers#8 &of->mutex kn->active#5 &rcu_state.expedited_wq &p->pi_lock &rq->__lock irq_context: 0 sb_writers#8 &of->mutex kn->active#5 &rcu_state.expedited_wq &p->pi_lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 pernet_ops_rwsem rtnl_mutex &idev->mc_query_lock irq_context: 0 pernet_ops_rwsem rtnl_mutex &idev->mc_query_lock &obj_hash[i].lock irq_context: 0 pernet_ops_rwsem rtnl_mutex (work_completion)(&(&idev->mc_report_work)->work) irq_context: 0 pernet_ops_rwsem rtnl_mutex &idev->mc_report_lock irq_context: 0 &p->lock &of->mutex kn->active#5 fs_reclaim &rq->__lock irq_context: 0 pernet_ops_rwsem rtnl_mutex &idev->mc_lock krc.lock irq_context: 0 &p->lock &of->mutex kn->active#5 fs_reclaim &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 pernet_ops_rwsem rtnl_mutex &pnn->pndevs.lock irq_context: 0 pernet_ops_rwsem rtnl_mutex &pnn->routes.lock irq_context: 0 pernet_ops_rwsem rtnl_mutex target_list_lock irq_context: 0 pernet_ops_rwsem rtnl_mutex _xmit_TUNNEL6 irq_context: 0 pernet_ops_rwsem rtnl_mutex &root->kernfs_rwsem kernfs_idr_lock irq_context: 0 pernet_ops_rwsem rtnl_mutex &root->kernfs_rwsem &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 &root->kernfs_rwsem &rq->__lock &cfs_rq->removed.lock irq_context: 0 pernet_ops_rwsem rtnl_mutex &root->kernfs_rwsem &obj_hash[i].lock irq_context: 0 pernet_ops_rwsem rtnl_mutex &root->kernfs_rwsem pool_lock#2 irq_context: 0 pernet_ops_rwsem rtnl_mutex &sem->wait_lock irq_context: 0 pernet_ops_rwsem rtnl_mutex kernfs_idr_lock irq_context: 0 pernet_ops_rwsem rtnl_mutex &root->kernfs_rwsem &rq->__lock irq_context: 0 sb_writers#5 &type->i_mutex_dir_key#5/1 fs_reclaim &rq->__lock irq_context: 0 pernet_ops_rwsem rtnl_mutex &root->kernfs_rwsem &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 sb_writers#5 &type->i_mutex_dir_key#5/1 fs_reclaim &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 pernet_ops_rwsem rtnl_mutex dev_pm_qos_sysfs_mtx irq_context: 0 pernet_ops_rwsem rtnl_mutex dev_pm_qos_sysfs_mtx &root->kernfs_rwsem irq_context: 0 pernet_ops_rwsem rtnl_mutex dev_pm_qos_sysfs_mtx &root->kernfs_rwsem irq_context: 0 pernet_ops_rwsem rtnl_mutex dev_pm_qos_sysfs_mtx dev_pm_qos_mtx irq_context: 0 pernet_ops_rwsem rtnl_mutex subsys mutex#17 &k->k_lock klist_remove_lock irq_context: 0 &f->f_pos_lock &type->i_mutex_dir_key#3 sb_writers#4 &journal->j_state_lock irq_context: 0 &f->f_pos_lock &type->i_mutex_dir_key#3 sb_writers#4 &journal->j_state_lock tk_core.seq.seqcount irq_context: 0 &f->f_pos_lock &type->i_mutex_dir_key#3 sb_writers#4 &journal->j_state_lock &obj_hash[i].lock irq_context: 0 &f->f_pos_lock &type->i_mutex_dir_key#3 sb_writers#4 &journal->j_state_lock &base->lock irq_context: 0 &f->f_pos_lock &type->i_mutex_dir_key#3 sb_writers#4 &journal->j_state_lock &base->lock &obj_hash[i].lock irq_context: 0 pernet_ops_rwsem rtnl_mutex &root->kernfs_rwsem kernfs_idr_lock &obj_hash[i].lock irq_context: 0 pernet_ops_rwsem rtnl_mutex &root->kernfs_rwsem kernfs_idr_lock pool_lock#2 irq_context: 0 pernet_ops_rwsem rtnl_mutex deferred_probe_mutex irq_context: 0 pernet_ops_rwsem rtnl_mutex device_links_lock irq_context: 0 pernet_ops_rwsem rtnl_mutex mmu_notifier_invalidate_range_start irq_context: 0 pernet_ops_rwsem rtnl_mutex uevent_sock_mutex mmu_notifier_invalidate_range_start irq_context: 0 pernet_ops_rwsem rtnl_mutex cpu_hotplug_lock xps_map_mutex irq_context: 0 pernet_ops_rwsem rtnl_mutex rcu_state.exp_mutex &rnp->exp_wq[1] irq_context: 0 pernet_ops_rwsem rcu_state.barrier_mutex irq_context: 0 pernet_ops_rwsem rcu_state.barrier_mutex rcu_state.barrier_lock irq_context: 0 pernet_ops_rwsem rcu_state.barrier_mutex rcu_state.barrier_lock &obj_hash[i].lock irq_context: 0 pernet_ops_rwsem rcu_state.barrier_mutex &x->wait#24 irq_context: 0 pernet_ops_rwsem rcu_state.barrier_mutex &rq->__lock irq_context: 0 pernet_ops_rwsem rcu_state.barrier_mutex &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 pernet_ops_rwsem rcu_state.barrier_mutex rcu_read_lock &rq->__lock irq_context: 0 pernet_ops_rwsem rcu_state.barrier_mutex rcu_read_lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)ext4-rsv-conversion (work_completion)(&ei->i_rsv_conversion_work) quarantine_lock irq_context: 0 rtnl_mutex &wg->device_update_lock (&peer->timer_retransmit_handshake) irq_context: 0 rtnl_mutex &wg->device_update_lock &base->lock irq_context: 0 rtnl_mutex &wg->device_update_lock (&peer->timer_send_keepalive) irq_context: 0 rtnl_mutex &wg->device_update_lock (&peer->timer_new_handshake) irq_context: 0 rtnl_mutex &wg->device_update_lock (&peer->timer_zero_key_material) irq_context: 0 rtnl_mutex &wg->device_update_lock &base->lock &obj_hash[i].lock irq_context: 0 rtnl_mutex &wg->device_update_lock (&peer->timer_persistent_keepalive) irq_context: 0 rtnl_mutex &wg->device_update_lock (work_completion)(&peer->clear_peer_work) irq_context: 0 rtnl_mutex &wg->device_update_lock &handshake->lock irq_context: 0 rtnl_mutex &wg->device_update_lock &handshake->lock &table->lock#2 irq_context: 0 rtnl_mutex &wg->device_update_lock &peer->keypairs.keypair_update_lock irq_context: 0 rtnl_mutex &wg->device_update_lock &peer->keypairs.keypair_update_lock &table->lock#2 irq_context: 0 rtnl_mutex &wg->device_update_lock &peer->keypairs.keypair_update_lock &obj_hash[i].lock irq_context: 0 rtnl_mutex &wg->device_update_lock &peer->keypairs.keypair_update_lock pool_lock#2 irq_context: 0 rtnl_mutex &r->consumer_lock#2 irq_context: 0 rtnl_mutex &wg->socket_update_lock irq_context: softirq (&ndev->rs_timer) rcu_read_lock rcu_read_lock rcu_read_lock_bh rcu_read_lock &obj_hash[i].lock irq_context: softirq (&ndev->rs_timer) rcu_read_lock rcu_read_lock rcu_read_lock_bh rcu_read_lock rcu_read_lock pool_lock#2 irq_context: softirq (&ndev->rs_timer) rcu_read_lock rcu_read_lock rcu_read_lock_bh rcu_read_lock rcu_read_lock &dir->lock#2 irq_context: softirq (&ndev->rs_timer) rcu_read_lock rcu_read_lock rcu_read_lock_bh rcu_read_lock rcu_read_lock &ul->lock irq_context: softirq (&ndev->rs_timer) rcu_read_lock rcu_read_lock rcu_read_lock_bh rcu_read_lock batched_entropy_u32.lock irq_context: softirq (&ndev->rs_timer) rcu_read_lock rcu_read_lock rcu_read_lock_bh rcu_read_lock rcu_read_lock &____s->seqcount#7 irq_context: softirq (&ndev->rs_timer) rcu_read_lock rcu_read_lock rcu_read_lock_bh rcu_read_lock rcu_read_lock rcu_read_lock &____s->seqcount#7 irq_context: softirq (&ndev->rs_timer) rcu_read_lock rcu_read_lock rcu_read_lock_bh rcu_read_lock rcu_read_lock &nf_nat_locks[i] irq_context: softirq (&ndev->rs_timer) rcu_read_lock rcu_read_lock rcu_read_lock_bh rcu_read_lock rcu_read_lock &nf_conntrack_locks[i] irq_context: softirq (&ndev->rs_timer) rcu_read_lock rcu_read_lock rcu_read_lock_bh rcu_read_lock rcu_read_lock &nf_conntrack_locks[i] &nf_conntrack_locks[i]/1 irq_context: softirq (&ndev->rs_timer) rcu_read_lock rcu_read_lock rcu_read_lock_bh rcu_read_lock rcu_read_lock &nf_conntrack_locks[i] &nf_conntrack_locks[i]/1 batched_entropy_u8.lock irq_context: softirq (&ndev->rs_timer) rcu_read_lock rcu_read_lock rcu_read_lock_bh rcu_read_lock rcu_read_lock &nf_conntrack_locks[i]/1 irq_context: softirq (&ndev->rs_timer) rcu_read_lock rcu_read_lock rcu_read_lock_bh rcu_read_lock rcu_read_lock &____s->seqcount#10 irq_context: softirq (&ndev->rs_timer) rcu_read_lock rcu_read_lock rcu_read_lock_bh rcu_read_lock rcu_read_lock rcu_read_lock_bh tk_core.seq.seqcount irq_context: softirq (&ndev->rs_timer) rcu_read_lock rcu_read_lock rcu_read_lock_bh rcu_read_lock rcu_read_lock rcu_read_lock_bh &list->lock#5 irq_context: 0 rtnl_mutex &table->hash[i].lock irq_context: 0 rtnl_mutex &table->hash[i].lock &table->hash2[i].lock irq_context: 0 rtnl_mutex k-clock-AF_INET irq_context: 0 rtnl_mutex &xa->xa_lock#9 irq_context: 0 rtnl_mutex &fsnotify_mark_srcu irq_context: 0 rtnl_mutex k-clock-AF_INET6 irq_context: 0 rtnl_mutex rcu_read_lock &tb->tb6_lock &n->list_lock irq_context: 0 rtnl_mutex rcu_read_lock &tb->tb6_lock &n->list_lock &c->lock irq_context: softirq rcu_callback pcpu_lock stock_lock irq_context: 0 rcu_state.barrier_mutex rcu_state.barrier_mutex.wait_lock irq_context: 0 pernet_ops_rwsem rcu_state.barrier_mutex.wait_lock irq_context: 0 pernet_ops_rwsem dev_base_lock irq_context: 0 pernet_ops_rwsem lweventlist_lock irq_context: 0 pernet_ops_rwsem &dir->lock#2 &obj_hash[i].lock irq_context: 0 pernet_ops_rwsem &dir->lock#2 pool_lock#2 irq_context: 0 pernet_ops_rwsem netdev_unregistering_wq.lock irq_context: 0 pernet_ops_rwsem (work_completion)(&ht->run_work) irq_context: 0 pernet_ops_rwsem &ht->mutex irq_context: 0 pernet_ops_rwsem &ht->mutex rcu_read_lock pool_lock#2 irq_context: 0 pernet_ops_rwsem &ht->mutex &obj_hash[i].lock irq_context: 0 pernet_ops_rwsem rtnl_mutex rcu_state.exp_mutex &rnp->exp_wq[0] irq_context: 0 pernet_ops_rwsem rtnl_mutex _xmit_TUNNEL irq_context: 0 pernet_ops_rwsem rcu_state.barrier_mutex rcu_state.barrier_mutex.wait_lock irq_context: 0 sb_writers#4 tomoyo_ss &n->list_lock irq_context: 0 sb_writers#4 tomoyo_ss &n->list_lock &c->lock irq_context: 0 (wq_completion)wg-crypt-wg2#3 (work_completion)(&peer->transmit_packet_work) &peer->endpoint_lock rcu_read_lock_bh &obj_hash[i].lock irq_context: 0 &sb->s_type->i_mutex_key#8 batched_entropy_u8.lock irq_context: 0 &sb->s_type->i_mutex_key#8 kfence_freelist_lock irq_context: 0 cb_lock &pcp->lock &zone->lock &____s->seqcount irq_context: 0 rcu_state.barrier_mutex.wait_lock irq_context: 0 &wg->device_update_lock irq_context: 0 &wg->device_update_lock &wg->socket_update_lock irq_context: 0 &wg->device_update_lock rcu_state.exp_mutex rcu_node_0 irq_context: 0 &wg->device_update_lock rcu_state.exp_mutex &obj_hash[i].lock irq_context: 0 &wg->device_update_lock rcu_state.exp_mutex rcu_read_lock &pool->lock irq_context: 0 &wg->device_update_lock rcu_state.exp_mutex rcu_read_lock &pool->lock &obj_hash[i].lock irq_context: 0 &wg->device_update_lock rcu_state.exp_mutex rcu_read_lock &pool->lock &p->pi_lock irq_context: 0 &wg->device_update_lock rcu_state.exp_mutex rcu_read_lock &pool->lock &p->pi_lock &rq->__lock irq_context: 0 &wg->device_update_lock rcu_state.exp_mutex rcu_read_lock &pool->lock &p->pi_lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 &wg->device_update_lock rcu_state.exp_mutex &rnp->exp_wq[2] irq_context: 0 &wg->device_update_lock rcu_state.exp_mutex &rq->__lock irq_context: 0 &wg->device_update_lock rcu_state.exp_mutex &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 &wg->device_update_lock &obj_hash[i].lock irq_context: 0 &wg->device_update_lock pool_lock#2 irq_context: 0 &wg->device_update_lock &table->lock irq_context: 0 &wg->device_update_lock rcu_state.exp_mutex &rnp->exp_wq[3] irq_context: 0 &wg->device_update_lock &peer->keypairs.keypair_update_lock irq_context: 0 &wg->device_update_lock (&peer->timer_retransmit_handshake) irq_context: 0 &wg->device_update_lock &base->lock irq_context: 0 &wg->device_update_lock (&peer->timer_send_keepalive) irq_context: 0 &wg->device_update_lock (&peer->timer_new_handshake) irq_context: 0 &wg->device_update_lock (&peer->timer_zero_key_material) irq_context: 0 &wg->device_update_lock (&peer->timer_persistent_keepalive) irq_context: 0 &wg->device_update_lock (work_completion)(&peer->clear_peer_work) irq_context: 0 pernet_ops_rwsem &net->sctp.addr_wq_lock &obj_hash[i].lock irq_context: 0 &wg->device_update_lock &wq->mutex irq_context: 0 &wg->device_update_lock &wq->mutex &pool->lock irq_context: 0 &wg->device_update_lock &wq->mutex &x->wait#10 irq_context: 0 &wg->device_update_lock napi_hash_lock irq_context: 0 &wg->device_update_lock rcu_state.exp_mutex &rnp->exp_wq[0] irq_context: 0 &sb->s_type->i_mutex_key#8 &sbi->s_writepages_rwsem jbd2_handle &ei->i_data_sem rcu_node_0 irq_context: 0 &sb->s_type->i_mutex_key#8 &sbi->s_writepages_rwsem jbd2_handle &ei->i_data_sem &rcu_state.expedited_wq irq_context: 0 &sb->s_type->i_mutex_key#8 &sbi->s_writepages_rwsem jbd2_handle &ei->i_data_sem &rcu_state.expedited_wq &p->pi_lock irq_context: 0 pernet_ops_rwsem k-sk_lock-AF_INET6 &obj_hash[i].lock irq_context: 0 &wg->device_update_lock &wq->mutex &pool->lock/1 irq_context: 0 &wg->device_update_lock &table->lock#2 irq_context: 0 &wg->device_update_lock &list->lock#17 irq_context: 0 &wg->device_update_lock rcu_state.exp_mutex &rnp->exp_wq[1] irq_context: 0 pernet_ops_rwsem k-sk_lock-AF_INET6 k-clock-AF_INET6 irq_context: 0 &wg->device_update_lock wq_pool_mutex irq_context: 0 &wg->device_update_lock wq_pool_mutex &wq->mutex irq_context: 0 &wg->device_update_lock wq_pool_mutex &wq->mutex &pool->lock irq_context: 0 &wg->device_update_lock &rq->__lock irq_context: 0 &wg->device_update_lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 &sb->s_type->i_mutex_key#8 &sbi->s_writepages_rwsem rcu_read_lock rcu_node_0 irq_context: 0 &sb->s_type->i_mutex_key#8 &sbi->s_writepages_rwsem rcu_read_lock &rcu_state.expedited_wq irq_context: 0 &sb->s_type->i_mutex_key#8 &sbi->s_writepages_rwsem rcu_read_lock &rcu_state.expedited_wq &p->pi_lock irq_context: 0 &wg->device_update_lock wq_pool_mutex &wq->mutex &pool->lock/1 irq_context: 0 &wg->device_update_lock &pool->lock/1 irq_context: 0 &wg->device_update_lock &pool->lock/1 rcu_read_lock &pool->lock irq_context: 0 &wg->device_update_lock &pool->lock/1 rcu_read_lock &pool->lock &obj_hash[i].lock irq_context: 0 &wg->device_update_lock &pool->lock/1 rcu_read_lock &pool->lock pool_lock#2 irq_context: 0 &wg->device_update_lock &pool->lock/1 rcu_read_lock &pool->lock &p->pi_lock irq_context: 0 &wg->device_update_lock &pool->lock/1 rcu_read_lock &pool->lock &p->pi_lock &rq->__lock irq_context: 0 &wg->device_update_lock &pool->lock/1 rcu_read_lock &pool->lock &p->pi_lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 &wg->device_update_lock wq_mayday_lock irq_context: 0 &wg->device_update_lock &p->pi_lock irq_context: 0 &wg->device_update_lock &p->pi_lock &rq->__lock irq_context: 0 &wg->device_update_lock &p->pi_lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 &wg->device_update_lock &cfs_rq->removed.lock irq_context: 0 &wg->device_update_lock &x->wait irq_context: 0 (wq_completion)events (work_completion)(&pwq->unbound_release_work) &rq->__lock irq_context: 0 (wq_completion)events (work_completion)(&pwq->unbound_release_work) &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 &wg->device_update_lock rcu_state.exp_mutex rcu_read_lock &pool->lock &p->pi_lock &rq->__lock &cfs_rq->removed.lock irq_context: 0 (wq_completion)events (work_completion)(&pwq->unbound_release_work) &rnp->exp_lock irq_context: 0 (wq_completion)events (work_completion)(&pwq->unbound_release_work) rcu_state.exp_mutex irq_context: 0 (wq_completion)events (work_completion)(&pwq->unbound_release_work) rcu_state.exp_mutex rcu_state.exp_mutex.wait_lock irq_context: 0 &sb->s_type->i_mutex_key#8 &sbi->s_writepages_rwsem jbd2_handle &pcp->lock &zone->lock irq_context: 0 &sb->s_type->i_mutex_key#8 &sbi->s_writepages_rwsem jbd2_handle &pcp->lock &zone->lock &____s->seqcount irq_context: 0 &wg->device_update_lock rcu_state.exp_mutex.wait_lock irq_context: 0 &wg->device_update_lock pcpu_lock irq_context: 0 &wg->device_update_lock &r->consumer_lock#2 irq_context: 0 &wg->device_update_lock rcu_read_lock pool_lock#2 irq_context: 0 &wg->device_update_lock rcu_read_lock rcu_node_0 irq_context: 0 &wg->device_update_lock rcu_read_lock &rcu_state.expedited_wq irq_context: 0 &wg->device_update_lock rcu_read_lock &rcu_state.expedited_wq &p->pi_lock irq_context: 0 &wg->device_update_lock rcu_state.barrier_mutex irq_context: 0 &wg->device_update_lock rcu_state.barrier_mutex rcu_state.barrier_mutex.wait_lock irq_context: 0 &wg->device_update_lock rcu_state.barrier_mutex &rq->__lock irq_context: 0 &wg->device_update_lock rcu_state.barrier_mutex &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 cb_lock &rdev->wiphy.mtx quarantine_lock irq_context: 0 cb_lock remove_cache_srcu irq_context: 0 cb_lock remove_cache_srcu quarantine_lock irq_context: 0 cb_lock remove_cache_srcu &n->list_lock irq_context: 0 cb_lock remove_cache_srcu rcu_read_lock pool_lock#2 irq_context: 0 cb_lock remove_cache_srcu &obj_hash[i].lock irq_context: 0 cb_lock remove_cache_srcu &c->lock irq_context: 0 pernet_ops_rwsem k-slock-AF_INET6 &obj_hash[i].lock irq_context: 0 pernet_ops_rwsem k-slock-AF_INET6 pool_lock#2 irq_context: 0 pernet_ops_rwsem k-slock-AF_INET6 elock-AF_INET6 irq_context: softirq (&peer->timer_persistent_keepalive) &n->list_lock irq_context: softirq (&peer->timer_persistent_keepalive) &n->list_lock &c->lock irq_context: 0 pernet_ops_rwsem rcu_read_lock &rq->__lock &cfs_rq->removed.lock irq_context: 0 (wq_completion)wg-crypt-wg0#3 (work_completion)(&({ do { const void *__vpp_verify = (typeof((worker) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __asm__ ("" : "=r"(__ptr) : "0"((typeof(*((worker))) *)((worker)))); (typeof((typeof(*((worker))) *)((worker)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); })->work) &rq->__lock irq_context: softirq (&peer->timer_retransmit_handshake) irq_context: softirq (&peer->timer_retransmit_handshake) &peer->endpoint_lock irq_context: softirq (&peer->timer_retransmit_handshake) rcu_read_lock_bh tk_core.seq.seqcount irq_context: softirq (&peer->timer_retransmit_handshake) rcu_read_lock_bh rcu_read_lock &pool->lock/1 irq_context: softirq (&peer->timer_retransmit_handshake) rcu_read_lock_bh rcu_read_lock &pool->lock/1 &obj_hash[i].lock irq_context: softirq (&peer->timer_retransmit_handshake) rcu_read_lock_bh rcu_read_lock &pool->lock/1 pool_lock#2 irq_context: softirq (&peer->timer_retransmit_handshake) rcu_read_lock_bh rcu_read_lock &pool->lock/1 &p->pi_lock irq_context: softirq (&peer->timer_retransmit_handshake) rcu_read_lock_bh rcu_read_lock &pool->lock/1 &p->pi_lock &rq->__lock irq_context: softirq (&peer->timer_retransmit_handshake) rcu_read_lock_bh rcu_read_lock &pool->lock/1 &p->pi_lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)wg-crypt-wg0#6 (work_completion)(&({ do { const void *__vpp_verify = (typeof((worker) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __asm__ ("" : "=r"(__ptr) : "0"((typeof(*((worker))) *)((worker)))); (typeof((typeof(*((worker))) *)((worker)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); })->work) rcu_read_lock &pool->lock &p->pi_lock &rq->__lock irq_context: 0 (wq_completion)wg-crypt-wg0#6 (work_completion)(&({ do { const void *__vpp_verify = (typeof((worker) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __asm__ ("" : "=r"(__ptr) : "0"((typeof(*((worker))) *)((worker)))); (typeof((typeof(*((worker))) *)((worker)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); })->work) rcu_read_lock &pool->lock &p->pi_lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)wg-crypt-wg0#5 (work_completion)(&({ do { const void *__vpp_verify = (typeof((worker) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __asm__ ("" : "=r"(__ptr) : "0"((typeof(*((worker))) *)((worker)))); (typeof((typeof(*((worker))) *)((worker)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); })->work) &rq->__lock irq_context: 0 (wq_completion)bat_events (work_completion)(&(&forw_packet_aggr->delayed_work)->work) batched_entropy_u8.lock irq_context: 0 (wq_completion)wg-kex-wg2#7 (work_completion)(&peer->transmit_handshake_work) &base->lock irq_context: 0 (wq_completion)wg-kex-wg2#7 (work_completion)(&peer->transmit_handshake_work) &base->lock &obj_hash[i].lock irq_context: 0 (wq_completion)wg-kex-wg0#12 (work_completion)(&({ do { const void *__vpp_verify = (typeof((worker) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __asm__ ("" : "=r"(__ptr) : "0"((typeof(*((worker))) *)((worker)))); (typeof((typeof(*((worker))) *)((worker)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); })->work) &handshake->lock remove_cache_srcu irq_context: 0 (wq_completion)wg-kex-wg0#12 (work_completion)(&({ do { const void *__vpp_verify = (typeof((worker) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __asm__ ("" : "=r"(__ptr) : "0"((typeof(*((worker))) *)((worker)))); (typeof((typeof(*((worker))) *)((worker)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); })->work) &handshake->lock remove_cache_srcu quarantine_lock irq_context: 0 (wq_completion)wg-kex-wg0#12 (work_completion)(&({ do { const void *__vpp_verify = (typeof((worker) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __asm__ ("" : "=r"(__ptr) : "0"((typeof(*((worker))) *)((worker)))); (typeof((typeof(*((worker))) *)((worker)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); })->work) &handshake->lock remove_cache_srcu &c->lock irq_context: 0 (wq_completion)wg-kex-wg0#12 (work_completion)(&({ do { const void *__vpp_verify = (typeof((worker) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __asm__ ("" : "=r"(__ptr) : "0"((typeof(*((worker))) *)((worker)))); (typeof((typeof(*((worker))) *)((worker)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); })->work) &handshake->lock remove_cache_srcu &n->list_lock irq_context: 0 (wq_completion)wg-kex-wg0#12 (work_completion)(&({ do { const void *__vpp_verify = (typeof((worker) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __asm__ ("" : "=r"(__ptr) : "0"((typeof(*((worker))) *)((worker)))); (typeof((typeof(*((worker))) *)((worker)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); })->work) &handshake->lock remove_cache_srcu rcu_read_lock pool_lock#2 irq_context: 0 (wq_completion)wg-kex-wg0#12 (work_completion)(&({ do { const void *__vpp_verify = (typeof((worker) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __asm__ ("" : "=r"(__ptr) : "0"((typeof(*((worker))) *)((worker)))); (typeof((typeof(*((worker))) *)((worker)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); })->work) &handshake->lock remove_cache_srcu &obj_hash[i].lock irq_context: 0 (wq_completion)wg-kex-wg0#12 (work_completion)(&({ do { const void *__vpp_verify = (typeof((worker) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __asm__ ("" : "=r"(__ptr) : "0"((typeof(*((worker))) *)((worker)))); (typeof((typeof(*((worker))) *)((worker)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); })->work) &handshake->lock remove_cache_srcu pool_lock#2 irq_context: 0 (wq_completion)wg-kex-wg0#12 (work_completion)(&({ do { const void *__vpp_verify = (typeof((worker) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __asm__ ("" : "=r"(__ptr) : "0"((typeof(*((worker))) *)((worker)))); (typeof((typeof(*((worker))) *)((worker)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); })->work) &base->lock irq_context: 0 (wq_completion)wg-kex-wg0#12 (work_completion)(&({ do { const void *__vpp_verify = (typeof((worker) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __asm__ ("" : "=r"(__ptr) : "0"((typeof(*((worker))) *)((worker)))); (typeof((typeof(*((worker))) *)((worker)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); })->work) &base->lock &obj_hash[i].lock irq_context: 0 (wq_completion)wg-kex-wg2#8 (work_completion)(&({ do { const void *__vpp_verify = (typeof((worker) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __asm__ ("" : "=r"(__ptr) : "0"((typeof(*((worker))) *)((worker)))); (typeof((typeof(*((worker))) *)((worker)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); })->work) pool_lock#2 irq_context: 0 (wq_completion)wg-kex-wg2#8 (work_completion)(&({ do { const void *__vpp_verify = (typeof((worker) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __asm__ ("" : "=r"(__ptr) : "0"((typeof(*((worker))) *)((worker)))); (typeof((typeof(*((worker))) *)((worker)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); })->work) &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)wg-crypt-wg2#4 (work_completion)(&({ do { const void *__vpp_verify = (typeof((worker) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __asm__ ("" : "=r"(__ptr) : "0"((typeof(*((worker))) *)((worker)))); (typeof((typeof(*((worker))) *)((worker)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); })->work) rcu_read_lock &pool->lock &p->pi_lock &rq->__lock irq_context: 0 (wq_completion)wg-crypt-wg2#4 (work_completion)(&({ do { const void *__vpp_verify = (typeof((worker) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __asm__ ("" : "=r"(__ptr) : "0"((typeof(*((worker))) *)((worker)))); (typeof((typeof(*((worker))) *)((worker)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); })->work) rcu_read_lock &pool->lock &p->pi_lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)wg-crypt-wg2#4 (work_completion)(&({ do { const void *__vpp_verify = (typeof((worker) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __asm__ ("" : "=r"(__ptr) : "0"((typeof(*((worker))) *)((worker)))); (typeof((typeof(*((worker))) *)((worker)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); })->work) &rq->__lock irq_context: 0 pernet_ops_rwsem k-sk_lock-AF_INET irq_context: 0 pernet_ops_rwsem k-sk_lock-AF_INET k-slock-AF_INET irq_context: 0 pernet_ops_rwsem k-sk_lock-AF_INET &obj_hash[i].lock irq_context: 0 pernet_ops_rwsem k-sk_lock-AF_INET k-clock-AF_INET irq_context: 0 pernet_ops_rwsem k-slock-AF_INET irq_context: 0 pernet_ops_rwsem k-slock-AF_INET &obj_hash[i].lock irq_context: 0 pernet_ops_rwsem k-slock-AF_INET pool_lock#2 irq_context: 0 pernet_ops_rwsem k-slock-AF_INET elock-AF_INET irq_context: 0 pernet_ops_rwsem per_cpu_ptr(&cgroup_rstat_cpu_lock, cpu) irq_context: 0 (wq_completion)events (work_completion)(&(&krcp->krw_arr[i].rcu_work)->work) &rq->__lock irq_context: 0 (wq_completion)events (work_completion)(&(&krcp->krw_arr[i].rcu_work)->work) &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)events (work_completion)(&(&krcp->krw_arr[i].rcu_work)->work) rcu_callback stock_lock irq_context: 0 (wq_completion)events (work_completion)(&(&krcp->krw_arr[i].rcu_work)->work) rcu_callback stock_lock rcu_read_lock per_cpu_ptr(&cgroup_rstat_cpu_lock, cpu) irq_context: 0 (wq_completion)events (work_completion)(&(&krcp->krw_arr[i].rcu_work)->work) rcu_callback per_cpu_ptr(&cgroup_rstat_cpu_lock, cpu) irq_context: softirq (&ndev->rs_timer) rcu_read_lock rcu_read_lock rcu_read_lock_bh rcu_read_lock rcu_read_lock &c->lock irq_context: softirq (&hsr->announce_timer) rcu_read_lock &n->list_lock irq_context: softirq (&hsr->announce_timer) rcu_read_lock &n->list_lock &c->lock irq_context: 0 pernet_ops_rwsem &sn->gssp_lock irq_context: 0 pernet_ops_rwsem &cd->hash_lock irq_context: 0 pernet_ops_rwsem cache_list_lock &cd->hash_lock irq_context: 0 pernet_ops_rwsem (&net->can.stattimer) irq_context: 0 (wq_completion)events fqdir_free_work rcu_state.barrier_mutex rcu_state.barrier_lock &obj_hash[i].lock irq_context: 0 (wq_completion)events fqdir_free_work rcu_state.barrier_mutex &pool->lock irq_context: 0 (wq_completion)events fqdir_free_work rcu_state.barrier_mutex &rq->__lock irq_context: 0 (wq_completion)events fqdir_free_work rcu_state.barrier_mutex &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 pernet_ops_rwsem xfrm_state_gc_work irq_context: 0 pernet_ops_rwsem &net->xfrm.xfrm_state_lock irq_context: 0 pernet_ops_rwsem &ht->mutex pool_lock#2 irq_context: 0 (wq_completion)inet_frag_wq (work_completion)(&fqdir->destroy_work) rcu_read_lock &pool->lock &p->pi_lock irq_context: 0 pernet_ops_rwsem remove_cache_srcu &pcp->lock &zone->lock irq_context: 0 pernet_ops_rwsem remove_cache_srcu &pcp->lock &zone->lock &____s->seqcount irq_context: 0 pernet_ops_rwsem remove_cache_srcu pool_lock#2 irq_context: 0 pernet_ops_rwsem (work_completion)(&(&net->ipv6.addr_chk_work)->work) irq_context: 0 pernet_ops_rwsem &net->ipv6.ip6addrlbl_table.lock &obj_hash[i].lock irq_context: 0 pernet_ops_rwsem &net->ipv6.ip6addrlbl_table.lock pool_lock#2 irq_context: 0 pernet_ops_rwsem &net->ipv6.ip6addrlbl_table.lock krc.lock irq_context: 0 pernet_ops_rwsem ip6_fl_lock irq_context: 0 pernet_ops_rwsem rtnl_mutex &net->rules_mod_lock irq_context: 0 pernet_ops_rwsem (&net->ipv6.ip6_fib_timer) irq_context: 0 pernet_ops_rwsem rtnl_mutex (&mrt->ipmr_expire_timer) irq_context: 0 pernet_ops_rwsem rtnl_mutex (work_completion)(&ht->run_work) irq_context: 0 pernet_ops_rwsem rtnl_mutex &ht->mutex irq_context: 0 pernet_ops_rwsem rtnl_mutex &ht->mutex &obj_hash[i].lock irq_context: 0 pernet_ops_rwsem rtnl_mutex &ht->mutex pool_lock#2 irq_context: 0 pernet_ops_rwsem rcu_read_lock rcu_read_lock &pool->lock irq_context: 0 pernet_ops_rwsem rcu_read_lock rcu_read_lock &pool->lock &obj_hash[i].lock irq_context: 0 pernet_ops_rwsem rcu_read_lock rcu_read_lock &pool->lock &p->pi_lock irq_context: 0 pernet_ops_rwsem rcu_read_lock rcu_read_lock &pool->lock &p->pi_lock &rq->__lock irq_context: 0 pernet_ops_rwsem rcu_read_lock rcu_read_lock &pool->lock &p->pi_lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 pernet_ops_rwsem __ip_vs_app_mutex &obj_hash[i].lock irq_context: 0 pernet_ops_rwsem __ip_vs_mutex irq_context: 0 pernet_ops_rwsem (&ipvs->dest_trash_timer) irq_context: 0 pernet_ops_rwsem (work_completion)(&(&ipvs->expire_nodest_conn_work)->work) irq_context: 0 pernet_ops_rwsem (work_completion)(&(&ipvs->defense_work)->work) irq_context: 0 pernet_ops_rwsem ipvs->est_mutex pcpu_lock irq_context: 0 pernet_ops_rwsem ipvs->est_mutex &obj_hash[i].lock irq_context: 0 pernet_ops_rwsem (work_completion)(&(&ipvs->est_reload_work)->work) irq_context: 0 pernet_ops_rwsem nfnl_subsys_ipset irq_context: 0 pernet_ops_rwsem recent_lock irq_context: 0 pernet_ops_rwsem hashlimit_mutex irq_context: 0 pernet_ops_rwsem trans_gc_work irq_context: 0 pernet_ops_rwsem rcu_state.exp_mutex &rnp->exp_wq[0] irq_context: 0 pernet_ops_rwsem (work_completion)(&(&cnet->ecache.dwork)->work) irq_context: 0 pernet_ops_rwsem rtnl_mutex nf_connlabels_lock irq_context: 0 pernet_ops_rwsem &root->kernfs_rwsem kernfs_idr_lock irq_context: 0 pernet_ops_rwsem &root->kernfs_rwsem &obj_hash[i].lock irq_context: 0 pernet_ops_rwsem &root->kernfs_rwsem pool_lock#2 irq_context: 0 pernet_ops_rwsem nf_hook_mutex cpu_hotplug_lock irq_context: 0 pernet_ops_rwsem uevent_sock_mutex rcu_read_lock &rq->__lock &cfs_rq->removed.lock irq_context: 0 pernet_ops_rwsem sysfs_symlink_target_lock irq_context: 0 pernet_ops_rwsem kernfs_idr_lock irq_context: 0 pernet_ops_rwsem tcp_metrics_lock irq_context: 0 (wq_completion)events (work_completion)(&ht->run_work) &ht->mutex &ht->lock &obj_hash[i].lock pool_lock irq_context: 0 pernet_ops_rwsem k-clock-AF_INET irq_context: 0 pernet_ops_rwsem sysctl_lock &obj_hash[i].lock pool_lock irq_context: 0 pernet_ops_rwsem (work_completion)(&net->xfrm.policy_hash_work) irq_context: 0 pernet_ops_rwsem &net->xfrm.xfrm_policy_lock irq_context: 0 pernet_ops_rwsem (work_completion)(&net->xfrm.state_hash_work) irq_context: 0 (wq_completion)events (work_completion)(&ht->run_work) &ht->mutex batched_entropy_u8.lock irq_context: 0 (wq_completion)events (work_completion)(&ht->run_work) &ht->mutex kfence_freelist_lock irq_context: 0 (wq_completion)events (work_completion)(&ht->run_work) &ht->mutex remove_cache_srcu pool_lock#2 irq_context: 0 pernet_ops_rwsem &list->lock#2 irq_context: 0 pernet_ops_rwsem &xa->xa_lock#3 irq_context: 0 pernet_ops_rwsem genl_sk_destructing_waitq.lock irq_context: 0 (wq_completion)events fqdir_free_work rcu_state.barrier_mutex.wait_lock irq_context: 0 (wq_completion)events fqdir_free_work &p->pi_lock irq_context: 0 (wq_completion)events fqdir_free_work &p->pi_lock &rq->__lock irq_context: 0 (wq_completion)events fqdir_free_work &p->pi_lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)events fqdir_free_work &obj_hash[i].lock pool_lock irq_context: 0 &dir->lock &obj_hash[i].lock irq_context: 0 &dir->lock pool_lock#2 irq_context: 0 &mq_lock irq_context: 0 (wq_completion)events free_ipc_work irq_context: 0 (wq_completion)events free_ipc_work &rnp->exp_lock irq_context: 0 (wq_completion)events free_ipc_work rcu_state.exp_mutex irq_context: 0 (wq_completion)events free_ipc_work rcu_state.exp_mutex rcu_state.exp_mutex.wait_lock irq_context: 0 (wq_completion)events free_ipc_work rcu_state.exp_mutex &pool->lock irq_context: 0 (wq_completion)events free_ipc_work rcu_state.exp_mutex &pool->lock &p->pi_lock irq_context: 0 (wq_completion)events free_ipc_work rcu_state.exp_mutex &pool->lock &p->pi_lock &rq->__lock irq_context: 0 (wq_completion)events free_ipc_work rcu_state.exp_mutex &pool->lock &p->pi_lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)events free_ipc_work rcu_state.exp_mutex &rq->__lock irq_context: 0 (wq_completion)events free_ipc_work rcu_state.exp_mutex &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)events free_ipc_work rcu_state.exp_mutex rcu_node_0 irq_context: 0 (wq_completion)events free_ipc_work rcu_state.exp_mutex &obj_hash[i].lock irq_context: 0 (wq_completion)events free_ipc_work rcu_state.exp_mutex rcu_read_lock &pool->lock irq_context: 0 (wq_completion)events free_ipc_work rcu_state.exp_mutex rcu_read_lock &pool->lock &obj_hash[i].lock irq_context: 0 (wq_completion)events free_ipc_work rcu_state.exp_mutex &rnp->exp_wq[2] irq_context: 0 (wq_completion)events free_ipc_work &obj_hash[i].lock irq_context: 0 (wq_completion)events free_ipc_work rcu_read_lock mount_lock irq_context: 0 (wq_completion)events free_ipc_work rcu_read_lock mount_lock mount_lock.seqcount irq_context: 0 (wq_completion)events free_ipc_work mount_lock irq_context: 0 (wq_completion)events free_ipc_work mount_lock mount_lock.seqcount irq_context: 0 (wq_completion)events free_ipc_work &fsnotify_mark_srcu irq_context: 0 (wq_completion)events free_ipc_work rcu_read_lock &dentry->d_lock irq_context: 0 (wq_completion)events free_ipc_work &type->s_umount_key#51 irq_context: 0 (wq_completion)events free_ipc_work &type->s_umount_key#51 shrinker_rwsem irq_context: 0 (wq_completion)events free_ipc_work &type->s_umount_key#51 rename_lock.seqcount irq_context: 0 (wq_completion)events free_ipc_work &type->s_umount_key#51 &dentry->d_lock irq_context: 0 (wq_completion)events free_ipc_work &type->s_umount_key#51 rcu_read_lock &dentry->d_lock irq_context: 0 (wq_completion)events free_ipc_work &type->s_umount_key#51 &sb->s_type->i_lock_key#20 irq_context: 0 (wq_completion)events free_ipc_work &type->s_umount_key#51 &s->s_inode_list_lock irq_context: 0 (wq_completion)events free_ipc_work &type->s_umount_key#51 &xa->xa_lock#9 irq_context: 0 (wq_completion)events free_ipc_work &type->s_umount_key#51 &obj_hash[i].lock irq_context: 0 (wq_completion)events free_ipc_work &type->s_umount_key#51 pool_lock#2 irq_context: 0 (wq_completion)events free_ipc_work &type->s_umount_key#51 &fsnotify_mark_srcu irq_context: 0 (wq_completion)events free_ipc_work unnamed_dev_ida.xa_lock irq_context: 0 (wq_completion)events free_ipc_work list_lrus_mutex irq_context: 0 (wq_completion)events free_ipc_work &xa->xa_lock#4 irq_context: 0 (wq_completion)events free_ipc_work &xa->xa_lock#4 &obj_hash[i].lock irq_context: 0 (wq_completion)events free_ipc_work &xa->xa_lock#4 &____s->seqcount irq_context: 0 (wq_completion)events free_ipc_work &xa->xa_lock#4 pool_lock#2 irq_context: 0 (wq_completion)events free_ipc_work &xa->xa_lock#4 rcu_read_lock pool_lock#2 irq_context: 0 (wq_completion)events free_ipc_work pool_lock#2 irq_context: 0 (wq_completion)events free_ipc_work sb_lock irq_context: 0 (wq_completion)events free_ipc_work sb_lock &obj_hash[i].lock irq_context: 0 (wq_completion)events free_ipc_work sb_lock pool_lock#2 irq_context: 0 (wq_completion)events free_ipc_work mnt_id_ida.xa_lock irq_context: 0 (wq_completion)events free_ipc_work &ids->rwsem irq_context: 0 (wq_completion)events free_ipc_work (work_completion)(&ht->run_work) irq_context: 0 (wq_completion)events free_ipc_work &ht->mutex irq_context: 0 (wq_completion)events free_ipc_work &ht->mutex &obj_hash[i].lock irq_context: 0 (wq_completion)events free_ipc_work &ht->mutex pool_lock#2 irq_context: 0 (wq_completion)events free_ipc_work percpu_counters_lock irq_context: 0 (wq_completion)events free_ipc_work pcpu_lock irq_context: 0 (wq_completion)events free_ipc_work sysctl_lock irq_context: 0 (wq_completion)events free_ipc_work sysctl_lock &obj_hash[i].lock irq_context: 0 (wq_completion)events free_ipc_work sysctl_lock pool_lock#2 irq_context: 0 (wq_completion)events free_ipc_work sysctl_lock krc.lock irq_context: 0 (wq_completion)events free_ipc_work proc_inum_ida.xa_lock irq_context: 0 (wq_completion)events free_ipc_work stock_lock irq_context: 0 &type->i_mutex_dir_key#3 rcu_node_0 irq_context: 0 sb_writers#4 sb_internal rcu_read_lock rcu_node_0 irq_context: 0 sb_writers#4 sb_internal rcu_read_lock &rq->__lock irq_context: 0 sb_writers#4 sb_internal rcu_read_lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 sb_writers#5 &type->i_mutex_dir_key#5/1 batched_entropy_u32.lock crngs.lock irq_context: 0 sb_writers#5 &type->i_mutex_dir_key#5 &simple_offset_xa_lock &pcp->lock &zone->lock irq_context: 0 sb_writers#5 &type->i_mutex_dir_key#5 &simple_offset_xa_lock &pcp->lock &zone->lock &____s->seqcount irq_context: 0 sb_writers#4 tomoyo_ss quarantine_lock irq_context: softirq (&ndev->rs_timer) rcu_read_lock rcu_read_lock &tbl->lock &n->lock irq_context: softirq (&ndev->rs_timer) rcu_read_lock rcu_read_lock &tbl->lock pool_lock#2 irq_context: softirq (&ndev->rs_timer) rcu_read_lock rcu_read_lock &tbl->lock &c->lock irq_context: softirq (&ndev->rs_timer) rcu_read_lock rcu_read_lock &tbl->lock &n->lock irq_context: softirq (&ndev->rs_timer) rcu_read_lock rcu_read_lock &tbl->lock &n->lock &____s->seqcount#9 irq_context: softirq (&ndev->rs_timer) rcu_read_lock rcu_read_lock &tbl->lock nl_table_lock irq_context: softirq (&ndev->rs_timer) rcu_read_lock rcu_read_lock &tbl->lock &obj_hash[i].lock irq_context: softirq (&ndev->rs_timer) rcu_read_lock rcu_read_lock &tbl->lock nl_table_wait.lock irq_context: softirq (&ndev->rs_timer) rcu_read_lock rcu_read_lock &tbl->lock rcu_read_lock lock#8 irq_context: softirq (&ndev->rs_timer) rcu_read_lock rcu_read_lock &tbl->lock rcu_read_lock id_table_lock irq_context: softirq (&ndev->rs_timer) rcu_read_lock rcu_read_lock &tbl->lock &dir->lock#2 irq_context: softirq (&ndev->rs_timer) rcu_read_lock rcu_read_lock &tbl->lock krc.lock irq_context: softirq (&ndev->rs_timer) rcu_read_lock rcu_read_lock &tbl->lock &obj_hash[i].lock pool_lock irq_context: softirq (&ndev->rs_timer) rcu_read_lock rcu_read_lock &tbl->lock &n->list_lock irq_context: softirq (&ndev->rs_timer) rcu_read_lock rcu_read_lock &tbl->lock &n->list_lock &c->lock irq_context: softirq (&ndev->rs_timer) rcu_read_lock rcu_read_lock &tbl->lock krc.lock &obj_hash[i].lock irq_context: softirq (&ndev->rs_timer) rcu_read_lock rcu_read_lock &tbl->lock krc.lock &base->lock irq_context: softirq (&ndev->rs_timer) rcu_read_lock rcu_read_lock &tbl->lock krc.lock &base->lock &obj_hash[i].lock irq_context: softirq (&ndev->rs_timer) rcu_read_lock rcu_read_lock &tbl->lock &____s->seqcount#2 irq_context: softirq (&ndev->rs_timer) rcu_read_lock rcu_read_lock &tbl->lock &____s->seqcount irq_context: 0 rtnl_mutex rcu_read_lock &entry->crc_lock irq_context: 0 rtnl_mutex rcu_read_lock tk_core.seq.seqcount irq_context: 0 rtnl_mutex rcu_read_lock &list->lock#5 irq_context: 0 rtnl_mutex key#19 irq_context: 0 (wq_completion)events (work_completion)(&(&krcp->monitor_work)->work) rcu_callback stock_lock irq_context: 0 rtnl_mutex &bat_priv->forw_bcast_list_lock irq_context: 0 rtnl_mutex &bat_priv->forw_bat_list_lock irq_context: 0 rtnl_mutex (work_completion)(&(&forw_packet_aggr->delayed_work)->work) irq_context: 0 namespace_sem &rq->__lock irq_context: 0 namespace_sem &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 &sb->s_type->i_mutex_key#8 &sbi->s_writepages_rwsem jbd2_handle &ei->i_data_sem &lg->lg_mutex &rq->__lock irq_context: 0 pernet_ops_rwsem k-sk_lock-AF_NETLINK rcu_read_lock &rcu_state.gp_wq irq_context: 0 &sb->s_type->i_mutex_key#8 &sbi->s_writepages_rwsem jbd2_handle &ei->i_data_sem &lg->lg_mutex &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 pernet_ops_rwsem k-sk_lock-AF_NETLINK rcu_read_lock &rcu_state.gp_wq &p->pi_lock irq_context: 0 pernet_ops_rwsem k-sk_lock-AF_NETLINK rcu_read_lock &rcu_state.gp_wq &p->pi_lock &rq->__lock irq_context: 0 pernet_ops_rwsem k-sk_lock-AF_NETLINK rcu_read_lock &rcu_state.gp_wq &p->pi_lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 &sb->s_type->i_mutex_key#8 &sbi->s_writepages_rwsem jbd2_handle lock#4 &lruvec->lru_lock irq_context: 0 pernet_ops_rwsem lock kernfs_idr_lock pool_lock#2 irq_context: 0 pernet_ops_rwsem &root->kernfs_rwsem &root->kernfs_iattr_rwsem &rq->__lock irq_context: 0 pernet_ops_rwsem &root->kernfs_rwsem &root->kernfs_iattr_rwsem &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 pernet_ops_rwsem &root->kernfs_rwsem &root->kernfs_iattr_rwsem &cfs_rq->removed.lock irq_context: 0 pernet_ops_rwsem &root->kernfs_rwsem &root->kernfs_iattr_rwsem &obj_hash[i].lock irq_context: 0 pernet_ops_rwsem &rq->__lock &cfs_rq->removed.lock irq_context: 0 &f->f_pos_lock sb_writers#4 &sb->s_type->i_mutex_key#8 &rcu_state.expedited_wq irq_context: 0 &f->f_pos_lock sb_writers#4 &sb->s_type->i_mutex_key#8 &rcu_state.expedited_wq &p->pi_lock irq_context: 0 rtnl_mutex &dev->tx_global_lock &batadv_netdev_xmit_lock_key irq_context: 0 rtnl_mutex &batadv_netdev_addr_lock_key/1 &obj_hash[i].lock irq_context: 0 rtnl_mutex &batadv_netdev_addr_lock_key/1 krc.lock irq_context: 0 rtnl_mutex rcu_read_lock &tb->tb6_lock rcu_read_lock &c->lock irq_context: 0 &f->f_pos_lock sb_writers#4 &sb->s_type->i_mutex_key#8 fs_reclaim &rq->__lock irq_context: 0 (wq_completion)bat_events (work_completion)(&(&bat_priv->mcast.work)->work) &bat_priv->mcast.mla_lock &obj_hash[i].lock irq_context: 0 &f->f_pos_lock sb_writers#4 &sb->s_type->i_mutex_key#8 batched_entropy_u8.lock irq_context: 0 &f->f_pos_lock sb_writers#4 &sb->s_type->i_mutex_key#8 kfence_freelist_lock irq_context: 0 rtnl_mutex &idev->mc_lock &batadv_netdev_addr_lock_key/1 irq_context: 0 rtnl_mutex &idev->mc_lock &batadv_netdev_addr_lock_key/1 &obj_hash[i].lock irq_context: 0 rtnl_mutex &idev->mc_lock &batadv_netdev_addr_lock_key/1 krc.lock irq_context: 0 &f->f_pos_lock sb_writers#4 &sb->s_type->i_mutex_key#8 fs_reclaim &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 &f->f_pos_lock sb_writers#4 &sb->s_type->i_mutex_key#8 &mm->mmap_lock &rq->__lock &cfs_rq->removed.lock irq_context: 0 (wq_completion)wg-crypt-wg0 (work_completion)(&peer->transmit_packet_work) &peer->endpoint_lock rcu_read_lock_bh batched_entropy_u32.lock crngs.lock irq_context: 0 nf_sockopt_mutex &rq->__lock irq_context: 0 nf_sockopt_mutex &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)rcu_gp (work_completion)(&rew->rew_work) rcu_state.exp_wake_mutex &rnp->exp_wq[3] &p->pi_lock &rq->__lock &cfs_rq->removed.lock irq_context: 0 cb_lock &rdev->wiphy.mtx &p->pi_lock &cfs_rq->removed.lock irq_context: 0 cb_lock &rdev->wiphy.mtx &rq->__lock irq_context: 0 cb_lock &rdev->wiphy.mtx &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 &f->f_pos_lock &type->i_mutex_dir_key#4 &root->kernfs_rwsem &sem->wait_lock irq_context: 0 &f->f_pos_lock &type->i_mutex_dir_key#4 &sem->wait_lock irq_context: 0 &f->f_pos_lock &type->i_mutex_dir_key#4 &p->pi_lock irq_context: 0 &f->f_pos_lock &type->i_mutex_dir_key#4 &p->pi_lock &rq->__lock irq_context: 0 &f->f_pos_lock &type->i_mutex_dir_key#4 &p->pi_lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 rtnl_mutex team->team_lock_key#2 pool_lock#2 irq_context: 0 rtnl_mutex team->team_lock_key#2 rcu_read_lock pool_lock#2 irq_context: softirq rcu_read_lock hwsim_radio_lock &____s->seqcount#2 irq_context: softirq rcu_read_lock hwsim_radio_lock &____s->seqcount irq_context: 0 &bat_priv->forw_bcast_list_lock irq_context: 0 &bat_priv->forw_bat_list_lock irq_context: 0 &bat_priv->gw.list_lock irq_context: 0 (work_completion)(&(&bat_priv->bat_v.ogm_wq)->work) irq_context: 0 &bat_priv->bat_v.ogm_buff_mutex irq_context: 0 &bat_priv->bat_v.ogm_buff_mutex &obj_hash[i].lock irq_context: 0 &bat_priv->bat_v.ogm_buff_mutex pool_lock#2 irq_context: 0 &bat_priv->tvlv.container_list_lock irq_context: 0 &bat_priv->tvlv.handler_list_lock irq_context: 0 (work_completion)(&(&bat_priv->nc.work)->work) irq_context: 0 key#17 irq_context: 0 key#18 irq_context: 0 &bat_priv->tvlv.container_list_lock &obj_hash[i].lock irq_context: 0 &bat_priv->tvlv.container_list_lock pool_lock#2 irq_context: 0 (work_completion)(&(&bat_priv->dat.work)->work) irq_context: 0 &hash->list_locks[i] irq_context: 0 (work_completion)(&(&bat_priv->bla.work)->work) irq_context: 0 key#20 irq_context: 0 key#20 &obj_hash[i].lock irq_context: 0 key#20 pool_lock#2 irq_context: 0 key#20 krc.lock irq_context: 0 (work_completion)(&(&bat_priv->mcast.work)->work) irq_context: 0 &bat_priv->tt.changes_list_lock irq_context: 0 (work_completion)(&(&bat_priv->tt.work)->work) irq_context: 0 key#16 irq_context: 0 key#16 &bat_priv->softif_vlan_list_lock irq_context: 0 key#16 &obj_hash[i].lock irq_context: 0 key#16 pool_lock#2 irq_context: 0 key#16 krc.lock irq_context: 0 key#21 irq_context: 0 &bat_priv->tt.req_list_lock irq_context: 0 &bat_priv->tt.changes_list_lock &obj_hash[i].lock irq_context: 0 &bat_priv->tt.changes_list_lock pool_lock#2 irq_context: 0 &bat_priv->tt.roam_list_lock irq_context: 0 (work_completion)(&(&bat_priv->orig_work)->work) irq_context: 0 key#19 irq_context: 0 &type->s_umount_key#23/1 &rq->__lock irq_context: 0 pernet_ops_rwsem pcpu_alloc_mutex &rq->__lock irq_context: 0 pernet_ops_rwsem pcpu_alloc_mutex &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 &mm->mmap_lock sb_pagefaults &rq->__lock irq_context: 0 &mm->mmap_lock sb_pagefaults &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 pernet_ops_rwsem fs_reclaim &rq->__lock irq_context: 0 pernet_ops_rwsem fs_reclaim &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 &mm->mmap_lock sb_pagefaults remove_cache_srcu irq_context: 0 &mm->mmap_lock sb_pagefaults remove_cache_srcu quarantine_lock irq_context: 0 &mm->mmap_lock sb_pagefaults remove_cache_srcu &c->lock irq_context: 0 &mm->mmap_lock sb_pagefaults remove_cache_srcu &n->list_lock irq_context: 0 &mm->mmap_lock sb_pagefaults remove_cache_srcu rcu_read_lock pool_lock#2 irq_context: 0 &mm->mmap_lock sb_pagefaults remove_cache_srcu &obj_hash[i].lock irq_context: 0 rtnl_mutex rcu_read_lock &tb->tb6_lock &obj_hash[i].lock pool_lock irq_context: 0 &mm->mmap_lock sb_pagefaults mapping.invalidate_lock rcu_read_lock rcu_node_0 irq_context: 0 &mm->mmap_lock sb_pagefaults mapping.invalidate_lock rcu_read_lock &rq->__lock irq_context: 0 &mm->mmap_lock sb_pagefaults mapping.invalidate_lock rcu_read_lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)events (work_completion)(&(&group->avgs_work)->work) &rq->__lock irq_context: 0 (wq_completion)events (work_completion)(&(&group->avgs_work)->work) &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 &mm->mmap_lock sb_pagefaults mapping.invalidate_lock rcu_read_lock &cfs_rq->removed.lock irq_context: 0 &mm->mmap_lock sb_pagefaults mapping.invalidate_lock rcu_read_lock &obj_hash[i].lock irq_context: 0 &mm->mmap_lock sb_pagefaults mapping.invalidate_lock rcu_read_lock pool_lock#2 irq_context: 0 &pipe->mutex/1 &mm->mmap_lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 file_rwsem &rq->__lock irq_context: 0 file_rwsem &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 sb_writers#4 &sb->s_type->i_mutex_key#8 ima_extend_list_mutex &c->lock irq_context: 0 sb_writers#4 &sb->s_type->i_mutex_key#8 kauditd_wait.lock &p->pi_lock irq_context: 0 sb_writers#4 &sb->s_type->i_mutex_key#8 kauditd_wait.lock &p->pi_lock &rq->__lock irq_context: 0 sb_writers#4 &sb->s_type->i_mutex_key#8 kauditd_wait.lock &p->pi_lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: softirq rcu_read_lock rcu_read_lock &base->lock irq_context: softirq rcu_read_lock rcu_read_lock &base->lock &obj_hash[i].lock irq_context: 0 sb_writers#8 kn->active#5 pool_lock#2 irq_context: 0 &p->lock &of->mutex kn->active#5 rcu_node_0 irq_context: softirq rcu_read_lock rcu_read_lock quarantine_lock irq_context: 0 &mm->mmap_lock sb_pagefaults mapping.invalidate_lock rcu_read_lock &rcu_state.gp_wq irq_context: 0 &mm->mmap_lock sb_pagefaults mapping.invalidate_lock rcu_read_lock &rcu_state.gp_wq &p->pi_lock irq_context: 0 &mm->mmap_lock sb_pagefaults mapping.invalidate_lock rcu_read_lock &rcu_state.gp_wq &p->pi_lock &rq->__lock irq_context: 0 &mm->mmap_lock sb_pagefaults mapping.invalidate_lock rcu_read_lock &rcu_state.gp_wq &p->pi_lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 &mm->mmap_lock sb_pagefaults mapping.invalidate_lock rcu_read_lock &rcu_state.expedited_wq irq_context: 0 &mm->mmap_lock sb_pagefaults mapping.invalidate_lock rcu_read_lock &rcu_state.expedited_wq &p->pi_lock irq_context: 0 &mm->mmap_lock sb_pagefaults mapping.invalidate_lock rcu_read_lock &rcu_state.expedited_wq &p->pi_lock &rq->__lock irq_context: 0 &mm->mmap_lock sb_pagefaults mapping.invalidate_lock rcu_read_lock &rcu_state.expedited_wq &p->pi_lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 rtnl_mutex team->team_lock_key#2 &root->kernfs_rwsem kernfs_idr_lock irq_context: 0 rtnl_mutex team->team_lock_key#2 &root->kernfs_rwsem &obj_hash[i].lock irq_context: 0 rtnl_mutex team->team_lock_key#2 &root->kernfs_rwsem pool_lock#2 irq_context: 0 rtnl_mutex team->team_lock_key#2 &root->kernfs_rwsem &rq->__lock irq_context: 0 rtnl_mutex team->team_lock_key#2 &root->kernfs_rwsem &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: softirq (&peer->timer_persistent_keepalive) &____s->seqcount#2 irq_context: softirq (&peer->timer_persistent_keepalive) &____s->seqcount irq_context: 0 (wq_completion)wg-crypt-wg0#4 (work_completion)(&peer->transmit_packet_work) &rq->__lock irq_context: 0 (wq_completion)wg-crypt-wg0#4 (work_completion)(&peer->transmit_packet_work) &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 rtnl_mutex team->team_lock_key#2 krc.lock irq_context: 0 rtnl_mutex team->team_lock_key#2 rcu_read_lock &pool->lock/1 pool_lock#2 irq_context: 0 rtnl_mutex team->team_lock_key#2 rcu_state.exp_mutex rcu_node_0 irq_context: 0 rtnl_mutex team->team_lock_key#2 rcu_state.exp_mutex &obj_hash[i].lock irq_context: 0 rtnl_mutex team->team_lock_key#2 rcu_state.exp_mutex rcu_read_lock &pool->lock irq_context: 0 rtnl_mutex team->team_lock_key#2 rcu_state.exp_mutex rcu_read_lock &pool->lock &obj_hash[i].lock irq_context: 0 rtnl_mutex team->team_lock_key#2 rcu_state.exp_mutex rcu_read_lock &pool->lock &p->pi_lock irq_context: 0 rtnl_mutex team->team_lock_key#2 rcu_state.exp_mutex rcu_read_lock &pool->lock &p->pi_lock &rq->__lock irq_context: 0 rtnl_mutex team->team_lock_key#2 rcu_state.exp_mutex rcu_read_lock &pool->lock &p->pi_lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 rtnl_mutex team->team_lock_key#2 rcu_state.exp_mutex &rq->__lock irq_context: 0 rtnl_mutex team->team_lock_key#2 rcu_state.exp_mutex &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)events_unbound &rq->__lock irq_context: 0 (wq_completion)events_unbound &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 rtnl_mutex team->team_lock_key#2 &dev_addr_list_lock_key#2/1 irq_context: 0 rtnl_mutex team->team_lock_key#2 &dev_addr_list_lock_key#2/1 _xmit_ETHER irq_context: 0 rtnl_mutex team->team_lock_key#2 &dev_addr_list_lock_key#2/1 _xmit_ETHER &obj_hash[i].lock irq_context: 0 rtnl_mutex team->team_lock_key#2 &dev_addr_list_lock_key#2/1 _xmit_ETHER pool_lock#2 irq_context: 0 rtnl_mutex team->team_lock_key#2 &dev_addr_list_lock_key#2/1 _xmit_ETHER krc.lock irq_context: 0 rtnl_mutex team->team_lock_key#2 &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 rtnl_mutex team->team_lock_key#2 &____s->seqcount#2 irq_context: 0 rtnl_mutex team->team_lock_key#2 &tbl->lock irq_context: 0 rtnl_mutex team->team_lock_key#2 &pn->hash_lock irq_context: 0 rtnl_mutex team->team_lock_key#2 &net->ipv6.fib6_gc_lock rcu_read_lock &tb->tb6_lock irq_context: 0 rtnl_mutex team->team_lock_key#2 &net->ipv6.fib6_gc_lock &obj_hash[i].lock irq_context: 0 rtnl_mutex uevent_sock_mutex &____s->seqcount#2 irq_context: softirq (&ndev->rs_timer) rcu_read_lock rcu_read_lock rcu_read_lock_bh _xmit_ETHER#2 rcu_read_lock &local->handle_wake_tx_queue_lock hwsim_radio_lock &c->lock irq_context: 0 pernet_ops_rwsem rtnl_mutex remove_cache_srcu pool_lock#2 irq_context: 0 pernet_ops_rwsem rtnl_mutex remove_cache_srcu &rq->__lock irq_context: 0 pernet_ops_rwsem rtnl_mutex remove_cache_srcu &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 pernet_ops_rwsem rtnl_mutex rcu_read_lock rcu_node_0 irq_context: 0 cb_lock &rdev->wiphy.mtx &____s->seqcount#2 irq_context: 0 pernet_ops_rwsem rtnl_mutex smc_ib_devices.mutex &rq->__lock irq_context: 0 pernet_ops_rwsem rtnl_mutex smc_ib_devices.mutex &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: softirq rcu_read_lock rcu_read_lock k-slock-AF_INET6 rcu_read_lock &pcp->lock &zone->lock irq_context: 0 sk_lock-AF_INET6 fs_reclaim &rq->__lock irq_context: 0 sk_lock-AF_INET6 fs_reclaim &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 sk_lock-AF_INET6 remove_cache_srcu pool_lock#2 irq_context: 0 sk_lock-AF_INET6 remove_cache_srcu rcu_read_lock rcu_node_0 irq_context: 0 sk_lock-AF_INET6 remove_cache_srcu rcu_read_lock &rq->__lock irq_context: 0 sk_lock-AF_INET6 remove_cache_srcu rcu_read_lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 pernet_ops_rwsem rtnl_mutex pcpu_alloc_mutex &rq->__lock irq_context: 0 pernet_ops_rwsem rtnl_mutex pcpu_alloc_mutex &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 sb_writers#4 &type->i_mutex_dir_key#3/1 sb_internal jbd2_handle &mapping->private_lock irq_context: 0 sb_writers#4 &type->i_mutex_dir_key#3/1 mmu_notifier_invalidate_range_start &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 &xt[i].mutex remove_cache_srcu &rq->__lock irq_context: 0 &xt[i].mutex remove_cache_srcu &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 pernet_ops_rwsem rtnl_mutex lock kernfs_idr_lock rcu_read_lock pool_lock#2 irq_context: 0 pernet_ops_rwsem rtnl_mutex lock kernfs_idr_lock &obj_hash[i].lock irq_context: 0 pernet_ops_rwsem rcu_state.exp_mutex &rnp->exp_wq[2] irq_context: 0 pernet_ops_rwsem rtnl_mutex fs_reclaim &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 sb_writers#4 &type->i_mutex_dir_key#3/1 &type->i_mutex_dir_key#3 &____s->seqcount#2 irq_context: 0 sb_writers#4 &type->i_mutex_dir_key#3/1 &type->i_mutex_dir_key#3 &____s->seqcount irq_context: 0 &p->lock remove_cache_srcu pool_lock#2 irq_context: 0 &vma->vm_lock->lock mmu_notifier_invalidate_range_start &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 sb_writers#4 &type->i_mutex_dir_key#3/1 rcu_read_lock rcu_read_lock rcu_node_0 irq_context: 0 sb_writers#4 &type->i_mutex_dir_key#3/1 rcu_read_lock rcu_read_lock &rq->__lock irq_context: 0 sb_writers#4 &type->i_mutex_dir_key#3/1 rcu_read_lock rcu_read_lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 sb_writers#3 &sb->s_type->i_mutex_key#9 rcu_node_0 irq_context: 0 &type->s_umount_key#23/1 &____s->seqcount#2 irq_context: 0 &type->s_umount_key#23/1 &xa->xa_lock#4 &c->lock irq_context: 0 &type->s_umount_key#23/1 &xa->xa_lock#4 &____s->seqcount#2 irq_context: 0 &type->s_umount_key#23/1 &xa->xa_lock#4 &____s->seqcount irq_context: 0 &type->s_umount_key#23/1 &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 pernet_ops_rwsem rtnl_mutex fs_reclaim mmu_notifier_invalidate_range_start &rq->__lock irq_context: 0 pernet_ops_rwsem uevent_sock_mutex &n->list_lock irq_context: 0 pernet_ops_rwsem uevent_sock_mutex &n->list_lock &c->lock irq_context: 0 dup_mmap_sem &mm->mmap_lock &mm->mmap_lock/1 fs_reclaim mmu_notifier_invalidate_range_start &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 pernet_ops_rwsem fs_reclaim mmu_notifier_invalidate_range_start &rq->__lock irq_context: 0 pernet_ops_rwsem fs_reclaim mmu_notifier_invalidate_range_start &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 pernet_ops_rwsem rtnl_mutex &root->kernfs_rwsem quarantine_lock irq_context: 0 &vma->vm_lock->lock fs_reclaim mmu_notifier_invalidate_range_start &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 tomoyo_ss remove_cache_srcu pool_lock#2 irq_context: 0 sk_lock-AF_INET6 fs_reclaim mmu_notifier_invalidate_range_start &rq->__lock irq_context: 0 sk_lock-AF_INET6 fs_reclaim mmu_notifier_invalidate_range_start &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 sb_writers#4 &type->i_mutex_dir_key#3/1 jbd2_handle &meta_group_info[i]->alloc_sem &rq->__lock irq_context: 0 sb_writers#4 &type->i_mutex_dir_key#3/1 jbd2_handle &meta_group_info[i]->alloc_sem &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 bpf_stats_enabled_mutex irq_context: 0 bpf_stats_enabled_mutex &newf->file_lock irq_context: 0 bpf_stats_enabled_mutex fs_reclaim irq_context: 0 bpf_stats_enabled_mutex fs_reclaim mmu_notifier_invalidate_range_start irq_context: 0 bpf_stats_enabled_mutex pool_lock#2 irq_context: 0 bpf_stats_enabled_mutex &xa->xa_lock#4 irq_context: 0 bpf_stats_enabled_mutex &rq->__lock irq_context: 0 bpf_stats_enabled_mutex &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 &f->f_pos_lock sb_writers#4 &sb->s_type->i_mutex_key#8 remove_cache_srcu pool_lock#2 irq_context: 0 bpf_stats_enabled_mutex &obj_hash[i].lock irq_context: 0 bpf_stats_enabled_mutex stock_lock irq_context: 0 bpf_stats_enabled_mutex &sb->s_type->i_lock_key#15 irq_context: 0 bpf_stats_enabled_mutex &sb->s_type->i_lock_key#15 &dentry->d_lock irq_context: 0 bpf_stats_enabled_mutex cpu_hotplug_lock irq_context: 0 bpf_stats_enabled_mutex cpu_hotplug_lock jump_label_mutex irq_context: 0 bpf_stats_enabled_mutex cpu_hotplug_lock jump_label_mutex text_mutex irq_context: 0 bpf_stats_enabled_mutex cpu_hotplug_lock jump_label_mutex text_mutex ptlock_ptr(page)#2 irq_context: 0 bpf_stats_enabled_mutex cpu_hotplug_lock jump_label_mutex text_mutex &rq->__lock irq_context: 0 bpf_stats_enabled_mutex cpu_hotplug_lock jump_label_mutex text_mutex &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 sk_lock-AF_INET6 &pcp->lock &zone->lock irq_context: 0 &mm->mmap_lock sb_pagefaults mapping.invalidate_lock &rq->__lock irq_context: 0 &mm->mmap_lock sb_pagefaults mapping.invalidate_lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 pernet_ops_rwsem rtnl_mutex &p->pi_lock &cfs_rq->removed.lock irq_context: 0 pernet_ops_rwsem rtnl_mutex sysctl_lock &obj_hash[i].lock pool_lock irq_context: 0 pernet_ops_rwsem rtnl_mutex &rnp->exp_wq[1] irq_context: 0 sb_writers#4 &sb->s_type->i_mutex_key#8 &n->list_lock irq_context: 0 sb_writers#4 &sb->s_type->i_mutex_key#8 &n->list_lock &c->lock irq_context: 0 (wq_completion)ext4-rsv-conversion (work_completion)(&ei->i_rsv_conversion_work) jbd2_handle &rq->__lock irq_context: 0 (wq_completion)ext4-rsv-conversion (work_completion)(&ei->i_rsv_conversion_work) jbd2_handle &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 sb_writers#4 &sb->s_type->i_mutex_key#8 rcu_read_lock rcu_read_lock rcu_node_0 irq_context: 0 sb_writers#4 &sb->s_type->i_mutex_key#8 rcu_read_lock rcu_read_lock &rq->__lock irq_context: 0 sb_writers#4 &sb->s_type->i_mutex_key#8 rcu_read_lock rcu_read_lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 pernet_ops_rwsem k-sk_lock-AF_INET6 &tcp_hashinfo.bhash[i].lock stock_lock irq_context: 0 pernet_ops_rwsem k-sk_lock-AF_INET6 &tcp_hashinfo.bhash[i].lock &tcp_hashinfo.bhash2[i].lock stock_lock irq_context: 0 pernet_ops_rwsem fill_pool_map-wait-type-override &rq->__lock irq_context: 0 pernet_ops_rwsem fill_pool_map-wait-type-override &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 pernet_ops_rwsem nf_ct_proto_mutex defrag4_mutex nf_hook_mutex stock_lock irq_context: 0 pernet_ops_rwsem nf_ct_proto_mutex nf_hook_mutex stock_lock irq_context: 0 pernet_ops_rwsem nf_ct_proto_mutex defrag6_mutex nf_hook_mutex stock_lock irq_context: 0 pernet_ops_rwsem k-sk_lock-AF_RXRPC &rxnet->local_mutex stock_lock irq_context: 0 pernet_ops_rwsem k-sk_lock-AF_RXRPC &rxnet->local_mutex &____s->seqcount#2 irq_context: 0 pernet_ops_rwsem k-sk_lock-AF_RXRPC &rxnet->local_mutex &____s->seqcount irq_context: softirq rcu_callback &rsp->gp_wait &obj_hash[i].lock irq_context: softirq rcu_callback &rsp->gp_wait pool_lock#2 irq_context: 0 sb_writers#4 &sb->s_type->i_mutex_key#8 mapping.invalidate_lock jbd2_handle &sbi->s_orphan_lock &rq->__lock irq_context: 0 sb_writers#4 &sb->s_type->i_mutex_key#8 mapping.invalidate_lock jbd2_handle &sbi->s_orphan_lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 sb_writers#4 &sb->s_type->i_mutex_key#8 &____s->seqcount#2 irq_context: 0 &sbi->s_writepages_rwsem jbd2_handle &ei->i_data_sem &ei->i_es_lock &____s->seqcount irq_context: 0 &sbi->s_writepages_rwsem jbd2_handle rcu_read_lock rcu_node_0 irq_context: 0 &sbi->s_writepages_rwsem jbd2_handle rcu_read_lock &rq->__lock irq_context: 0 &sbi->s_writepages_rwsem jbd2_handle rcu_read_lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 sk_lock-AF_INET6 rcu_read_lock &rq->__lock &cfs_rq->removed.lock irq_context: 0 pernet_ops_rwsem rtnl_mutex uevent_sock_mutex rcu_read_lock rcu_node_0 irq_context: 0 pernet_ops_rwsem rtnl_mutex uevent_sock_mutex rcu_read_lock &rq->__lock irq_context: 0 pernet_ops_rwsem rtnl_mutex uevent_sock_mutex rcu_read_lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 sk_lock-AF_INET6 &cfs_rq->removed.lock irq_context: 0 &sb->s_type->i_mutex_key#8 rcu_node_0 irq_context: 0 &sb->s_type->i_mutex_key#8 &rcu_state.expedited_wq irq_context: 0 &sb->s_type->i_mutex_key#8 &rcu_state.expedited_wq &p->pi_lock irq_context: 0 &sb->s_type->i_mutex_key#8 &rcu_state.expedited_wq &p->pi_lock &rq->__lock irq_context: 0 &sb->s_type->i_mutex_key#8 &rcu_state.expedited_wq &p->pi_lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 pernet_ops_rwsem rtnl_mutex dev_pm_qos_sysfs_mtx &rq->__lock irq_context: 0 pernet_ops_rwsem rtnl_mutex dev_pm_qos_sysfs_mtx &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: softirq rcu_callback prog_idr_lock irq_context: softirq rcu_callback bpf_lock irq_context: 0 sb_writers#8 &type->i_mutex_dir_key#4 &root->kernfs_rwsem &sem->wait_lock irq_context: 0 sb_writers#8 &type->i_mutex_dir_key#4 &sem->wait_lock irq_context: 0 sb_writers#8 &type->i_mutex_dir_key#4 &p->pi_lock irq_context: 0 sb_writers#4 &sb->s_type->i_mutex_key#8 &cfs_rq->removed.lock irq_context: 0 sb_writers#4 &sb->s_type->i_mutex_key#8 mapping.invalidate_lock remove_cache_srcu irq_context: 0 sb_writers#4 &sb->s_type->i_mutex_key#8 mapping.invalidate_lock remove_cache_srcu quarantine_lock irq_context: 0 sb_writers#4 &sb->s_type->i_mutex_key#8 mapping.invalidate_lock remove_cache_srcu &c->lock irq_context: 0 sb_writers#4 &sb->s_type->i_mutex_key#8 mapping.invalidate_lock remove_cache_srcu &n->list_lock irq_context: 0 sb_writers#4 &sb->s_type->i_mutex_key#8 mapping.invalidate_lock remove_cache_srcu rcu_read_lock pool_lock#2 irq_context: 0 sb_writers#4 &sb->s_type->i_mutex_key#8 mapping.invalidate_lock remove_cache_srcu &obj_hash[i].lock irq_context: 0 sb_writers#4 &sb->s_type->i_mutex_key#8 mapping.invalidate_lock jbd2_handle &ei->i_data_sem &ei->i_prealloc_lock &pa->pa_lock#2 irq_context: 0 &mm->mmap_lock &xa->xa_lock#9 &n->list_lock irq_context: 0 &rnp->exp_wq[2] irq_context: 0 sb_writers#4 &meta->lock irq_context: 0 sb_writers#4 kfence_freelist_lock irq_context: 0 (wq_completion)events free_ipc_work rcu_state.exp_mutex &rnp->exp_wq[3] irq_context: 0 (wq_completion)netns irq_context: 0 (wq_completion)netns net_cleanup_work irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem net_rwsem irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem &net->nsid_lock irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem rtnl_mutex irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem rtnl_mutex.wait_lock irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem &p->pi_lock irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem rcu_read_lock &tn->node_list_lock irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem ebt_mutex irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem &xt[i].mutex irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem &nft_net->commit_mutex irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem nf_ct_ecache_mutex irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem netns_bpf_mutex irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem &rnp->exp_lock irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem rcu_state.exp_mutex irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem rcu_state.exp_mutex rcu_state.exp_mutex.wait_lock irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem rcu_state.exp_mutex &rq->__lock irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem rcu_state.exp_mutex &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)events free_ipc_work rcu_state.exp_mutex.wait_lock irq_context: 0 (wq_completion)events free_ipc_work &p->pi_lock irq_context: 0 (wq_completion)events free_ipc_work &p->pi_lock &rq->__lock irq_context: 0 (wq_completion)events free_ipc_work &p->pi_lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem rcu_state.exp_mutex rcu_node_0 irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem rcu_state.exp_mutex &obj_hash[i].lock irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem rcu_state.exp_mutex rcu_read_lock &pool->lock irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem rcu_state.exp_mutex rcu_read_lock &pool->lock &obj_hash[i].lock irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem rcu_state.exp_mutex rcu_read_lock &pool->lock &p->pi_lock irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem rcu_state.exp_mutex rcu_read_lock &pool->lock &p->pi_lock &rq->__lock irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem rcu_state.exp_mutex rcu_read_lock &pool->lock &p->pi_lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem rcu_state.exp_mutex &rnp->exp_wq[0] irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem rcu_state.exp_mutex &rq->__lock &cfs_rq->removed.lock irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem rcu_state.exp_mutex &cfs_rq->removed.lock irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem rcu_state.exp_mutex pool_lock#2 irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem &obj_hash[i].lock irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem (&net->fs_probe_timer) irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem &base->lock irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem &net->cells_lock irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem (&net->cells_timer) irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem rcu_read_lock &pool->lock irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem rcu_read_lock &pool->lock &obj_hash[i].lock irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem rcu_read_lock &pool->lock pool_lock#2 irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem bit_wait_table + i irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem &rq->__lock irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)afs irq_context: 0 (wq_completion)afs (work_completion)(&net->cells_manager) irq_context: 0 (wq_completion)afs (work_completion)(&net->cells_manager) &net->cells_lock irq_context: 0 (wq_completion)afs (work_completion)(&net->cells_manager) bit_wait_table + i irq_context: 0 (wq_completion)afs (work_completion)(&net->cells_manager) bit_wait_table + i &p->pi_lock irq_context: 0 (wq_completion)afs (work_completion)(&net->cells_manager) bit_wait_table + i &p->pi_lock &rq->__lock irq_context: 0 (wq_completion)afs (work_completion)(&net->cells_manager) bit_wait_table + i &p->pi_lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem (&net->fs_timer) irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem rcu_read_lock &pool->lock &p->pi_lock irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem rcu_read_lock &pool->lock &p->pi_lock &rq->__lock irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem rcu_read_lock &pool->lock &p->pi_lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)afs (work_completion)(&net->fs_manager) irq_context: 0 (wq_completion)afs (work_completion)(&net->fs_manager) &(&net->fs_lock)->lock irq_context: 0 (wq_completion)afs (work_completion)(&net->fs_manager) bit_wait_table + i irq_context: 0 (wq_completion)afs (work_completion)(&net->fs_manager) bit_wait_table + i &p->pi_lock irq_context: 0 (wq_completion)afs (work_completion)(&net->fs_manager) bit_wait_table + i &p->pi_lock &rq->__lock irq_context: 0 (wq_completion)afs (work_completion)(&net->fs_manager) bit_wait_table + i &p->pi_lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem k-sk_lock-AF_RXRPC irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem k-sk_lock-AF_RXRPC k-slock-AF_RXRPC irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem k-sk_lock-AF_RXRPC &rx->incoming_lock irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem k-sk_lock-AF_RXRPC &obj_hash[i].lock irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem k-sk_lock-AF_RXRPC pool_lock#2 irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem k-sk_lock-AF_RXRPC &rq->__lock irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem k-sk_lock-AF_RXRPC &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem k-sk_lock-AF_RXRPC &rxnet->conn_lock irq_context: 0 (wq_completion)events (work_completion)(&s->destroy_work) quarantine_lock irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem k-sk_lock-AF_RXRPC &obj_hash[i].lock pool_lock irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem k-sk_lock-AF_RXRPC &call->waitq irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem k-sk_lock-AF_RXRPC rcu_read_lock &call->notify_lock irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem k-sk_lock-AF_RXRPC (rxrpc_call_limiter).lock irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem k-sk_lock-AF_RXRPC &rx->recvmsg_lock irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem k-sk_lock-AF_RXRPC &rx->call_lock irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem k-sk_lock-AF_RXRPC &rxnet->call_lock irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem k-sk_lock-AF_RXRPC (&call->timer) irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem k-sk_lock-AF_RXRPC &base->lock irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem k-sk_lock-AF_RXRPC &list->lock#22 irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem k-slock-AF_RXRPC irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem (wq_completion)kafsd irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem (wq_completion)kafsd &rq->__lock irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem (wq_completion)kafsd &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 sb_writers#4 rcu_read_lock &rcu_state.gp_wq irq_context: 0 sb_writers#4 rcu_read_lock &rcu_state.gp_wq &p->pi_lock irq_context: 0 sb_writers#4 rcu_read_lock &rcu_state.gp_wq &p->pi_lock &rq->__lock irq_context: 0 sb_writers#4 rcu_read_lock &rcu_state.gp_wq &p->pi_lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem &wq->mutex irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem &wq->mutex &pool->lock irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem &wq->mutex &x->wait#10 irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem pool_lock#2 irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem k-clock-AF_RXRPC irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem &local->services_lock irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem (wq_completion)krxrpcd irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem &wq->mutex &pool->lock/1 irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem rlock-AF_RXRPC irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem &p->pi_lock &rq->__lock irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem &p->pi_lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem &x->wait irq_context: 0 &rxnet->local_mutex irq_context: 0 (&local->client_conn_reap_timer) irq_context: 0 &rxnet->conn_lock irq_context: 0 &table->hash[i].lock irq_context: 0 &table->hash[i].lock &table->hash2[i].lock irq_context: 0 k-clock-AF_INET6 irq_context: 0 &list->lock#23 irq_context: 0 &p->alloc_lock &x->wait &p->pi_lock &rq->__lock irq_context: 0 &p->alloc_lock &x->wait &p->pi_lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem &sb->s_type->i_lock_key#8 irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem &xa->xa_lock#9 irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem &fsnotify_mark_srcu irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem proc_subdir_lock irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem &ent->pde_unload_lock irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem proc_inum_ida.xa_lock irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem ovs_mutex irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem ovs_mutex (work_completion)(&data->gc_work) irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem ovs_mutex nf_ct_proto_mutex irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem ovs_mutex nf_ct_proto_mutex nf_hook_mutex irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem ovs_mutex nf_ct_proto_mutex nf_hook_mutex cpu_hotplug_lock irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem ovs_mutex nf_ct_proto_mutex nf_hook_mutex fs_reclaim irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem ovs_mutex nf_ct_proto_mutex nf_hook_mutex fs_reclaim mmu_notifier_invalidate_range_start irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem ovs_mutex nf_ct_proto_mutex nf_hook_mutex pool_lock#2 irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem ovs_mutex nf_ct_proto_mutex &obj_hash[i].lock irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem ovs_mutex nf_ct_proto_mutex pool_lock#2 irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem ovs_mutex nf_ct_proto_mutex defrag4_mutex irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem ovs_mutex nf_ct_proto_mutex defrag4_mutex nf_hook_mutex irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem ovs_mutex nf_ct_proto_mutex defrag4_mutex nf_hook_mutex cpu_hotplug_lock irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem ovs_mutex nf_ct_proto_mutex defrag4_mutex &obj_hash[i].lock irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem ovs_mutex nf_ct_proto_mutex defrag4_mutex pool_lock#2 irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem ovs_mutex nf_ct_proto_mutex defrag6_mutex irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem ovs_mutex nf_ct_proto_mutex defrag6_mutex nf_hook_mutex irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem ovs_mutex nf_ct_proto_mutex defrag6_mutex nf_hook_mutex cpu_hotplug_lock irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem ovs_mutex nf_ct_proto_mutex defrag6_mutex &obj_hash[i].lock irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem ovs_mutex nf_ct_proto_mutex defrag6_mutex pool_lock#2 irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem ovs_mutex &obj_hash[i].lock irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem ovs_mutex pool_lock#2 irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem ovs_mutex nf_connlabels_lock irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem ovs_mutex net_rwsem irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem &base->lock &obj_hash[i].lock irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem (work_completion)(&(&ovs_net->masks_rebalance)->work) irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem (work_completion)(&ovs_net->dp_notify_work) irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem &srv->idr_lock irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem rcu_state.exp_mutex.wait_lock irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem wq_pool_mutex irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem wq_pool_mutex &wq->mutex irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem wq_pool_mutex &wq->mutex &pool->lock/1 irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem &pool->lock/1 irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem &pool->lock/1 rcu_read_lock &pool->lock irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem &pool->lock/1 rcu_read_lock &pool->lock &obj_hash[i].lock irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem &pool->lock/1 rcu_read_lock &pool->lock pool_lock#2 irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem &pool->lock/1 rcu_read_lock &pool->lock &p->pi_lock irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem &pool->lock/1 rcu_read_lock &pool->lock &p->pi_lock &rq->__lock irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem &pool->lock/1 rcu_read_lock &pool->lock &p->pi_lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)events (work_completion)(&pwq->unbound_release_work) rcu_state.exp_mutex &rnp->exp_wq[3] irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem k-sk_lock-AF_TIPC irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem k-sk_lock-AF_TIPC k-slock-AF_TIPC irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem k-sk_lock-AF_TIPC &tn->nametbl_lock irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem k-sk_lock-AF_TIPC &tn->nametbl_lock &service->lock irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem k-sk_lock-AF_TIPC &tn->nametbl_lock &service->lock &obj_hash[i].lock irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem k-sk_lock-AF_TIPC &tn->nametbl_lock &service->lock pool_lock#2 irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem k-sk_lock-AF_TIPC &tn->nametbl_lock &service->lock krc.lock irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem k-sk_lock-AF_TIPC &tn->nametbl_lock &nt->cluster_scope_lock irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem k-sk_lock-AF_TIPC &tn->nametbl_lock &obj_hash[i].lock irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem k-sk_lock-AF_TIPC &tn->nametbl_lock pool_lock#2 irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem k-sk_lock-AF_TIPC &tn->nametbl_lock krc.lock irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem k-sk_lock-AF_TIPC &obj_hash[i].lock irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem k-sk_lock-AF_TIPC rcu_read_lock rhashtable_bucket irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem k-sk_lock-AF_TIPC k-clock-AF_TIPC irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem k-slock-AF_TIPC irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem ptype_lock irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem rcu_state.exp_mutex &rnp->exp_wq[1] irq_context: 0 pernet_ops_rwsem rtnl_mutex &rnp->exp_wq[2] irq_context: 0 (wq_completion)events (work_completion)(&pwq->unbound_release_work) &rnp->exp_wq[2] irq_context: 0 (wq_completion)events (work_completion)(&pwq->unbound_release_work) &pool->lock irq_context: 0 (wq_completion)events (work_completion)(&pwq->unbound_release_work) &pool->lock &p->pi_lock irq_context: 0 (wq_completion)events (work_completion)(&pwq->unbound_release_work) &pool->lock &p->pi_lock &rq->__lock irq_context: 0 (wq_completion)events (work_completion)(&pwq->unbound_release_work) &pool->lock &p->pi_lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: softirq rcu_callback rlock-AF_RXRPC irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem &dir->lock#2 irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem (work_completion)(&tn->work) irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem &rnp->exp_wq[2] irq_context: 0 (wq_completion)events (work_completion)(&pwq->unbound_release_work) rcu_state.exp_mutex &rnp->exp_wq[2] irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem &tn->nametbl_lock irq_context: 0 (wq_completion)events (work_completion)(&(&nsim_dev->trap_data->trap_report_dw)->work) &devlink->lock_key rcu_node_0 irq_context: 0 sb_writers#4 rcu_read_lock &rcu_state.expedited_wq irq_context: 0 sb_writers#4 rcu_read_lock &rcu_state.expedited_wq &p->pi_lock irq_context: 0 sb_writers#4 rcu_read_lock &rcu_state.expedited_wq &p->pi_lock &rq->__lock irq_context: 0 sb_writers#4 rcu_read_lock &rcu_state.expedited_wq &p->pi_lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem rcu_read_lock pool_lock#2 irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem (work_completion)(&ht->run_work) irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem &ht->mutex irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem &ht->mutex &obj_hash[i].lock irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem &ht->mutex pool_lock#2 irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem (work_completion)(&(&c->work)->work) irq_context: 0 pernet_ops_rwsem rtnl_mutex &root->kernfs_rwsem &root->kernfs_iattr_rwsem &rq->__lock irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem pcpu_lock irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem k-sk_lock-AF_INET6 irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem k-sk_lock-AF_INET6 k-slock-AF_INET6 irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem k-sk_lock-AF_INET6 k-clock-AF_INET6 irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem k-slock-AF_INET6 irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem (wq_completion)krdsd irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem (work_completion)(&rtn->rds_tcp_accept_w) irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem k-sk_lock-AF_INET6 &h->lhash2[i].lock irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem k-sk_lock-AF_INET6 &queue->rskq_lock irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem k-sk_lock-AF_INET6 k-slock-AF_INET6 &obj_hash[i].lock irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem k-sk_lock-AF_INET6 k-slock-AF_INET6 &tcp_hashinfo.bhash[i].lock irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem k-sk_lock-AF_INET6 k-slock-AF_INET6 &tcp_hashinfo.bhash[i].lock stock_lock irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem k-sk_lock-AF_INET6 k-slock-AF_INET6 &tcp_hashinfo.bhash[i].lock &obj_hash[i].lock irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem k-sk_lock-AF_INET6 k-slock-AF_INET6 &tcp_hashinfo.bhash[i].lock pool_lock#2 irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem k-sk_lock-AF_INET6 k-slock-AF_INET6 &tcp_hashinfo.bhash[i].lock &tcp_hashinfo.bhash2[i].lock irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem k-sk_lock-AF_INET6 k-slock-AF_INET6 &tcp_hashinfo.bhash[i].lock &tcp_hashinfo.bhash2[i].lock stock_lock irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem k-sk_lock-AF_INET6 k-slock-AF_INET6 &tcp_hashinfo.bhash[i].lock &tcp_hashinfo.bhash2[i].lock &obj_hash[i].lock irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem k-sk_lock-AF_INET6 k-slock-AF_INET6 &tcp_hashinfo.bhash[i].lock &tcp_hashinfo.bhash2[i].lock pool_lock#2 irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem k-sk_lock-AF_INET6 k-slock-AF_INET6 elock-AF_INET6 irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem rds_tcp_conn_lock irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem sysctl_lock irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem sysctl_lock &obj_hash[i].lock irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem sysctl_lock pool_lock#2 irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem sysctl_lock krc.lock irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem loop_conns_lock irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem (wq_completion)l2tp irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem rcu_state.barrier_mutex irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem rcu_state.barrier_mutex rcu_state.barrier_lock irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem rcu_state.barrier_mutex rcu_state.barrier_lock &obj_hash[i].lock irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem rcu_state.barrier_mutex &x->wait#24 irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem rcu_state.barrier_mutex &rq->__lock irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem rcu_state.barrier_mutex &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 sb_writers#4 &type->i_mutex_dir_key#3/1 &type->i_mutex_dir_key#3 &rq->__lock &cfs_rq->removed.lock irq_context: 0 (wq_completion)events_unbound (work_completion)(&(&kfence_timer)->work) cpu_hotplug_lock jump_label_mutex jump_label_mutex.wait_lock irq_context: 0 (wq_completion)events_unbound (work_completion)(&(&kfence_timer)->work) cpu_hotplug_lock jump_label_mutex &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 bpf_stats_enabled_mutex cpu_hotplug_lock jump_label_mutex.wait_lock irq_context: 0 bpf_stats_enabled_mutex cpu_hotplug_lock &p->pi_lock irq_context: 0 bpf_stats_enabled_mutex cpu_hotplug_lock &p->pi_lock &rq->__lock irq_context: 0 bpf_stats_enabled_mutex cpu_hotplug_lock &p->pi_lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 sb_writers#4 &type->i_mutex_dir_key#3 jbd2_handle batched_entropy_u32.lock crngs.lock irq_context: 0 &xt[i].mutex rcu_node_0 irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem rcu_state.barrier_mutex.wait_lock irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem (&rxnet->peer_keepalive_timer) irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem (work_completion)(&rxnet->peer_keepalive_work) irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem rcu_read_lock &pool->lock/1 irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem (&rxnet->service_conn_reap_timer) irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem rcu_read_lock &pool->lock/1 &obj_hash[i].lock irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem rcu_read_lock &pool->lock/1 &p->pi_lock irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem rcu_read_lock &pool->lock/1 &p->pi_lock &rq->__lock irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem rcu_read_lock &pool->lock/1 &p->pi_lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem &x->wait#10 irq_context: 0 (wq_completion)krxrpcd (work_completion)(&rxnet->service_conn_reaper) irq_context: 0 (wq_completion)krxrpcd (work_completion)(&rxnet->service_conn_reaper) &rxnet->conn_lock irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem &rxnet->conn_lock irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem rtnl_mutex rtnl_mutex.wait_lock irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem rtnl_mutex &rq->__lock irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem rtnl_mutex &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 &xt[i].mutex remove_cache_srcu pool_lock#2 irq_context: 0 pernet_ops_rwsem rtnl_mutex &root->kernfs_rwsem &root->kernfs_iattr_rwsem &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 dup_mmap_sem &mm->mmap_lock &mm->mmap_lock/1 key#22 irq_context: 0 &f->f_pos_lock sb_writers#4 &sb->s_type->i_mutex_key#8 quarantine_lock irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem rtnl_mutex rcu_state.exp_mutex rcu_node_0 irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem rtnl_mutex rcu_state.exp_mutex &obj_hash[i].lock irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem rtnl_mutex rcu_state.exp_mutex rcu_read_lock &pool->lock irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem rtnl_mutex rcu_state.exp_mutex rcu_read_lock &pool->lock &obj_hash[i].lock irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem rtnl_mutex rcu_state.exp_mutex rcu_read_lock &pool->lock &p->pi_lock irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem rtnl_mutex rcu_state.exp_mutex rcu_read_lock &pool->lock &p->pi_lock &rq->__lock irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem rtnl_mutex rcu_state.exp_mutex rcu_read_lock &pool->lock &p->pi_lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem rtnl_mutex rcu_state.exp_mutex &rnp->exp_wq[0] irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem rtnl_mutex rcu_state.exp_mutex &rq->__lock irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem rtnl_mutex rcu_state.exp_mutex &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem rtnl_mutex &obj_hash[i].lock irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem rtnl_mutex dev_base_lock irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem rtnl_mutex dev_base_lock &xa->xa_lock#3 irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem rtnl_mutex cpu_hotplug_lock irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem rtnl_mutex cpu_hotplug_lock &list->lock#5 irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem rtnl_mutex rcu_state.exp_mutex &rnp->exp_wq[1] irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem rtnl_mutex bpf_devs_lock irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem rtnl_mutex net_rwsem irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem rtnl_mutex net_rwsem &list->lock#2 irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem rtnl_mutex &in_dev->mc_tomb_lock irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem rtnl_mutex sysctl_lock irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem rtnl_mutex sysctl_lock &obj_hash[i].lock irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem rtnl_mutex sysctl_lock pool_lock#2 irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem rtnl_mutex sysctl_lock krc.lock irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem rtnl_mutex stock_lock irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem rtnl_mutex pool_lock#2 irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem rtnl_mutex fs_reclaim irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem rtnl_mutex fs_reclaim mmu_notifier_invalidate_range_start irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem rtnl_mutex nl_table_lock irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem rtnl_mutex nl_table_wait.lock irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem rtnl_mutex &tbl->lock irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem rtnl_mutex &dir->lock#2 irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem rtnl_mutex class irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem rtnl_mutex (&tbl->proxy_timer) irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem rtnl_mutex &base->lock irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem rtnl_mutex &ul->lock irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem rtnl_mutex &net->xdp.lock irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem rtnl_mutex krc.lock irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem rtnl_mutex mirred_list_lock irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem rtnl_mutex &nft_net->commit_mutex irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem rtnl_mutex &tn->lock irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem rtnl_mutex rcu_read_lock &tb->tb6_lock irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem rtnl_mutex rcu_read_lock &tb->tb6_lock &net->ipv6.fib6_walker_lock irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem rtnl_mutex proc_subdir_lock irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem rtnl_mutex &ent->pde_unload_lock irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem rtnl_mutex proc_inum_ida.xa_lock irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem rtnl_mutex &net->ipv6.addrconf_hash_lock irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem rtnl_mutex &ndev->lock irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem rtnl_mutex &ndev->lock &obj_hash[i].lock irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem rtnl_mutex &idev->mc_lock irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem rtnl_mutex rcu_state.exp_mutex &rnp->exp_wq[2] irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem rtnl_mutex &idev->mc_query_lock irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem rtnl_mutex &idev->mc_query_lock &obj_hash[i].lock irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem rtnl_mutex (work_completion)(&(&idev->mc_report_work)->work) irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem rtnl_mutex &idev->mc_report_lock irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem rtnl_mutex &idev->mc_lock &obj_hash[i].lock irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem rtnl_mutex &idev->mc_lock pool_lock#2 irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem rtnl_mutex &idev->mc_lock krc.lock irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem rtnl_mutex &pnn->pndevs.lock irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem rtnl_mutex &pnn->routes.lock irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem rtnl_mutex &pnettable->lock irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem rtnl_mutex smc_ib_devices.mutex irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem rtnl_mutex target_list_lock irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem rtnl_mutex _xmit_NONE irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem rtnl_mutex &root->kernfs_rwsem irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem rtnl_mutex &root->kernfs_rwsem &root->kernfs_iattr_rwsem irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem rtnl_mutex &root->kernfs_rwsem kernfs_idr_lock irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem rtnl_mutex &root->kernfs_rwsem &obj_hash[i].lock irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem rtnl_mutex &root->kernfs_rwsem pool_lock#2 irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem rtnl_mutex &sem->wait_lock irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem rtnl_mutex &p->pi_lock irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem rtnl_mutex &p->pi_lock &rq->__lock irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem rtnl_mutex &p->pi_lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 sk_lock-AF_INET6 &rq->__lock &cfs_rq->removed.lock irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem rtnl_mutex sysfs_symlink_target_lock irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem rtnl_mutex kernfs_idr_lock irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem rtnl_mutex &k->list_lock irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem rtnl_mutex &root->kernfs_rwsem irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem rtnl_mutex dev_hotplug_mutex irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem rtnl_mutex dev_hotplug_mutex &dev->power.lock irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem rtnl_mutex dev_pm_qos_sysfs_mtx irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem rtnl_mutex dev_pm_qos_sysfs_mtx &root->kernfs_rwsem irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem rtnl_mutex dev_pm_qos_sysfs_mtx &root->kernfs_rwsem irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem rtnl_mutex dev_pm_qos_sysfs_mtx dev_pm_qos_mtx irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem rtnl_mutex subsys mutex#17 irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem rtnl_mutex subsys mutex#17 &k->k_lock irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem rtnl_mutex subsys mutex#17 &k->k_lock klist_remove_lock irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem rtnl_mutex &x->wait#9 irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem rtnl_mutex dpm_list_mtx irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem rtnl_mutex &dev->power.lock irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem rtnl_mutex deferred_probe_mutex irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem rtnl_mutex device_links_lock irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem rtnl_mutex gdp_mutex irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem rtnl_mutex cpu_hotplug_lock xps_map_mutex irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem rtnl_mutex rcu_state.exp_mutex &rnp->exp_wq[3] irq_context: 0 sk_lock-AF_INET6 clock-AF_INET6 irq_context: 0 sk_lock-AF_INET6 &sctp_port_hashtable[i].lock irq_context: 0 sk_lock-AF_INET6 &sctp_port_hashtable[i].lock &____s->seqcount#2 irq_context: 0 sk_lock-AF_INET6 &sctp_port_hashtable[i].lock &____s->seqcount irq_context: 0 sk_lock-AF_INET6 &sctp_port_hashtable[i].lock pool_lock#2 irq_context: 0 sk_lock-AF_INET6 &sctp_port_hashtable[i].lock &c->lock irq_context: 0 sk_lock-AF_INET6 crngs.lock irq_context: 0 sk_lock-AF_INET6 rcu_read_lock &n->list_lock irq_context: 0 sk_lock-AF_INET6 rcu_read_lock &n->list_lock &c->lock irq_context: 0 sk_lock-AF_INET6 rcu_read_lock &____s->seqcount#2 irq_context: 0 sk_lock-AF_INET6 rcu_read_lock batched_entropy_u8.lock irq_context: 0 sk_lock-AF_INET6 rcu_read_lock kfence_freelist_lock irq_context: softirq (&ndev->rs_timer) rcu_read_lock rcu_read_lock &obj_hash[i].lock pool_lock irq_context: 0 (wq_completion)events (work_completion)(&aux->work) pack_mutex text_mutex text_mutex.wait_lock irq_context: 0 (wq_completion)events (work_completion)(&aux->work) pack_mutex text_mutex &pool->lock irq_context: 0 (wq_completion)events (work_completion)(&aux->work) pack_mutex text_mutex &pool->lock &p->pi_lock irq_context: 0 (wq_completion)events (work_completion)(&aux->work) pack_mutex text_mutex &pool->lock &p->pi_lock &rq->__lock irq_context: 0 (wq_completion)events (work_completion)(&aux->work) pack_mutex text_mutex &pool->lock &p->pi_lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)events (work_completion)(&aux->work) pack_mutex text_mutex &rq->__lock irq_context: 0 (wq_completion)events (work_completion)(&aux->work) pack_mutex text_mutex &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 sk_lock-AF_INET6 rcu_read_lock rhashtable_bucket irq_context: 0 sk_lock-AF_INET6 &base->lock irq_context: 0 sk_lock-AF_INET6 &base->lock &obj_hash[i].lock irq_context: 0 sk_lock-AF_INET6 rcu_read_lock rcu_read_lock &ct->lock irq_context: 0 sk_lock-AF_INET6 rcu_read_lock rcu_read_lock rcu_read_lock &____s->seqcount#7 irq_context: 0 sk_lock-AF_INET6 rcu_read_lock rcu_read_lock &nf_nat_locks[i] irq_context: 0 sk_lock-AF_INET6 rcu_read_lock rcu_read_lock &nf_conntrack_locks[i] irq_context: 0 sk_lock-AF_INET6 rcu_read_lock rcu_read_lock &nf_conntrack_locks[i] batched_entropy_u8.lock irq_context: 0 sk_lock-AF_INET6 rcu_read_lock rcu_read_lock &nf_conntrack_locks[i] batched_entropy_u8.lock crngs.lock irq_context: 0 sk_lock-AF_INET6 rcu_read_lock rcu_read_lock &____s->seqcount#10 irq_context: 0 sk_lock-AF_INET6 rcu_read_lock rcu_read_lock rcu_read_lock_bh tk_core.seq.seqcount irq_context: 0 sk_lock-AF_INET6 rcu_read_lock rcu_read_lock rcu_read_lock_bh &list->lock#5 irq_context: softirq rcu_read_lock rcu_read_lock slock-AF_INET6 irq_context: 0 sk_lock-AF_INET6 &asoc->wait irq_context: 0 sk_lock-AF_INET6 tk_core.seq.seqcount irq_context: 0 sk_lock-AF_INET6 &obj_hash[i].lock pool_lock irq_context: 0 sk_lock-AF_INET6 rcu_read_lock rcu_read_lock rcu_read_lock stock_lock irq_context: 0 sk_lock-AF_INET6 rcu_read_lock rcu_read_lock rcu_node_0 irq_context: 0 sk_lock-AF_INET6 rcu_read_lock rcu_read_lock &rq->__lock irq_context: 0 sk_lock-AF_INET6 rcu_read_lock rcu_read_lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)events_unbound connector_reaper_work &rq->__lock &cfs_rq->removed.lock irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem dev_base_lock irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem lweventlist_lock irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem napi_hash_lock irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem fs_reclaim irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem fs_reclaim mmu_notifier_invalidate_range_start irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem stock_lock irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem krc.lock irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem &dir->lock#2 &obj_hash[i].lock irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem &dir->lock#2 pool_lock#2 irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem netdev_unregistering_wq.lock irq_context: 0 sk_lock-AF_INET6 slock-AF_INET6 &sk->sk_lock.wq irq_context: 0 sk_lock-AF_INET6 rcu_read_lock key#23 irq_context: 0 (wq_completion)events_unbound (work_completion)(&(&kfence_timer)->work) cpu_hotplug_lock jump_label_mutex text_mutex.wait_lock irq_context: 0 (wq_completion)events_unbound (work_completion)(&(&kfence_timer)->work) cpu_hotplug_lock jump_label_mutex &p->pi_lock irq_context: 0 (wq_completion)events_unbound (work_completion)(&(&kfence_timer)->work) cpu_hotplug_lock jump_label_mutex &p->pi_lock &rq->__lock irq_context: 0 (wq_completion)events_unbound (work_completion)(&(&kfence_timer)->work) cpu_hotplug_lock jump_label_mutex &p->pi_lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: softirq (&peer->timer_persistent_keepalive) rcu_read_lock_bh rcu_read_lock &pool->lock &p->pi_lock &rq->__lock &cfs_rq->removed.lock irq_context: 0 sk_lock-AF_INET6 fill_pool_map-wait-type-override pool_lock#2 irq_context: 0 sk_lock-AF_INET6 fill_pool_map-wait-type-override pool_lock irq_context: 0 sk_lock-AF_INET6 fill_pool_map-wait-type-override &c->lock irq_context: 0 sk_lock-AF_INET6 fill_pool_map-wait-type-override &____s->seqcount#2 irq_context: 0 sk_lock-AF_INET6 fill_pool_map-wait-type-override &____s->seqcount irq_context: 0 &sb->s_type->i_mutex_key#8 &sbi->s_writepages_rwsem jbd2_handle rcu_read_lock rcu_read_lock &q->queue_lock fill_pool_map-wait-type-override pool_lock#2 irq_context: 0 &sb->s_type->i_mutex_key#8 &sbi->s_writepages_rwsem jbd2_handle rcu_read_lock rcu_read_lock &q->queue_lock fill_pool_map-wait-type-override pool_lock irq_context: softirq &ei->i_completed_io_lock rcu_read_lock &pool->lock/1 fill_pool_map-wait-type-override pool_lock#2 irq_context: softirq &ei->i_completed_io_lock rcu_read_lock &pool->lock/1 fill_pool_map-wait-type-override &c->lock irq_context: softirq &ei->i_completed_io_lock rcu_read_lock &pool->lock/1 fill_pool_map-wait-type-override &____s->seqcount#2 irq_context: softirq &ei->i_completed_io_lock rcu_read_lock &pool->lock/1 fill_pool_map-wait-type-override &____s->seqcount irq_context: softirq &ei->i_completed_io_lock rcu_read_lock &pool->lock/1 fill_pool_map-wait-type-override pool_lock irq_context: 0 &sb->s_type->i_mutex_key#8 rcu_read_lock rcu_node_0 irq_context: 0 &sb->s_type->i_mutex_key#8 rcu_read_lock &rq->__lock irq_context: 0 &sb->s_type->i_mutex_key#8 rcu_read_lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 sk_lock-AF_INET6 rcu_read_lock rcu_read_lock rcu_read_lock &n->list_lock irq_context: 0 sk_lock-AF_INET6 rcu_read_lock rcu_read_lock rcu_read_lock &n->list_lock &c->lock irq_context: 0 sk_lock-AF_INET6 fill_pool_map-wait-type-override &rq->__lock irq_context: 0 sk_lock-AF_INET6 fill_pool_map-wait-type-override &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 sk_lock-AF_INET6 fill_pool_map-wait-type-override rcu_read_lock &rq->__lock irq_context: 0 sk_lock-AF_INET6 fill_pool_map-wait-type-override rcu_read_lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem rtnl_mutex &____s->seqcount irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem rtnl_mutex rcu_read_lock pool_lock#2 irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem rtnl_mutex _xmit_TUNNEL6 irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem rtnl_mutex &root->kernfs_rwsem kernfs_idr_lock &obj_hash[i].lock irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem rtnl_mutex &root->kernfs_rwsem kernfs_idr_lock pool_lock#2 irq_context: 0 sk_lock-AF_INET6 krc.lock irq_context: 0 pernet_ops_rwsem rtnl_mutex fill_pool_map-wait-type-override &rq->__lock irq_context: 0 pernet_ops_rwsem rtnl_mutex fill_pool_map-wait-type-override &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 &sb->s_type->i_mutex_key#8 &p->pi_lock irq_context: 0 &sb->s_type->i_mutex_key#8 &sbi->s_writepages_rwsem rcu_read_lock fill_pool_map-wait-type-override pool_lock#2 irq_context: 0 &sb->s_type->i_mutex_key#8 &sbi->s_writepages_rwsem rcu_read_lock fill_pool_map-wait-type-override pool_lock irq_context: 0 pernet_ops_rwsem cpu_hotplug_lock wq_pool_mutex fill_pool_map-wait-type-override pool_lock#2 irq_context: 0 pernet_ops_rwsem cpu_hotplug_lock wq_pool_mutex fill_pool_map-wait-type-override pool_lock irq_context: 0 &mm->mmap_lock fill_pool_map-wait-type-override &____s->seqcount#2 irq_context: 0 &sb->s_type->i_mutex_key#8 &sbi->s_writepages_rwsem jbd2_handle &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 syslog_lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 &mm->mmap_lock remove_cache_srcu pool_lock#2 irq_context: softirq rcu_callback key#23 irq_context: 0 sk_lock-AF_INET6 rcu_read_lock rcu_read_lock &pool->lock irq_context: 0 sk_lock-AF_INET6 rcu_read_lock rcu_read_lock &pool->lock &obj_hash[i].lock irq_context: 0 sk_lock-AF_INET6 rcu_read_lock rcu_read_lock &pool->lock &p->pi_lock irq_context: 0 &sb->s_type->i_mutex_key#8 &sbi->s_writepages_rwsem &mapping->i_mmap_rwsem mmu_notifier_invalidate_range_start irq_context: 0 &sb->s_type->i_mutex_key#8 &sbi->s_writepages_rwsem &mapping->i_mmap_rwsem mmu_notifier_invalidate_range_start &rq->__lock irq_context: 0 &sb->s_type->i_mutex_key#8 &sbi->s_writepages_rwsem &mapping->i_mmap_rwsem mmu_notifier_invalidate_range_start &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 &sb->s_type->i_mutex_key#8 &sbi->s_writepages_rwsem &mapping->i_mmap_rwsem ptlock_ptr(page)#2 irq_context: 0 &sb->s_type->i_mutex_key#8 &sbi->s_writepages_rwsem &mapping->private_lock irq_context: 0 &sb->s_type->i_mutex_key#8 &sbi->s_writepages_rwsem &mapping->private_lock rcu_read_lock &memcg->move_lock irq_context: 0 &sb->s_type->i_mutex_key#8 &sbi->s_writepages_rwsem rcu_read_lock &rq->__lock irq_context: 0 &sb->s_type->i_mutex_key#8 &sbi->s_writepages_rwsem rcu_read_lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 &sb->s_type->i_mutex_key#8 &sbi->s_writepages_rwsem rcu_read_lock &memcg->move_lock irq_context: 0 &sb->s_type->i_mutex_key#8 &sbi->s_writepages_rwsem rcu_read_lock &xa->xa_lock#9 irq_context: 0 &sb->s_type->i_mutex_key#8 &sbi->s_writepages_rwsem rcu_read_lock &xa->xa_lock#9 &s->s_inode_wblist_lock irq_context: 0 (wq_completion)events free_ipc_work rcu_state.exp_mutex &rnp->exp_wq[1] irq_context: 0 pernet_ops_rwsem rtnl_mutex rcu_state.exp_mutex irq_context: 0 pernet_ops_rwsem rtnl_mutex rcu_state.exp_mutex rcu_state.exp_mutex.wait_lock irq_context: 0 &mm->mmap_lock &mapping->i_mmap_rwsem &rcu_state.expedited_wq irq_context: 0 &mm->mmap_lock remove_cache_srcu pool_lock#2 irq_context: 0 sk_lock-AF_INET6 rcu_read_lock rcu_read_lock &pool->lock &p->pi_lock &rq->__lock irq_context: 0 sk_lock-AF_INET6 rcu_read_lock rcu_read_lock &pool->lock &p->pi_lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)events (work_completion)(&ht->run_work) &ht->mutex &n->list_lock irq_context: 0 (wq_completion)events (work_completion)(&ht->run_work) &ht->mutex &n->list_lock &c->lock irq_context: 0 sb_writers#5 &type->i_mutex_dir_key#5/1 tomoyo_ss rcu_read_lock rcu_read_lock rcu_node_0 irq_context: 0 sb_writers#5 &type->i_mutex_dir_key#5/1 tomoyo_ss rcu_read_lock rcu_read_lock &rq->__lock irq_context: 0 sb_writers#5 &type->i_mutex_dir_key#5/1 tomoyo_ss rcu_read_lock rcu_read_lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem rtnl_mutex &c->lock irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem rtnl_mutex _xmit_SIT irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem rtnl_mutex &pcp->lock &zone->lock irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem rcu_state.barrier_mutex rcu_state.barrier_mutex.wait_lock irq_context: 0 pernet_ops_rwsem quarantine_lock irq_context: 0 pernet_ops_rwsem rtnl_mutex &caifn->caifdevs.lock rcu_state.exp_mutex &rnp->exp_wq[1] irq_context: 0 (wq_completion)ext4-rsv-conversion (work_completion)(&ei->i_rsv_conversion_work) jbd2_handle &ei->i_data_sem remove_cache_srcu pool_lock#2 irq_context: 0 &sb->s_type->i_mutex_key#8 rcu_read_lock &rq->__lock &cfs_rq->removed.lock irq_context: 0 sk_lock-AF_INET6 sctp_assocs_id_lock irq_context: 0 sk_lock-AF_INET6 sctp_assocs_id_lock pool_lock#2 irq_context: 0 pernet_ops_rwsem rcu_read_lock &rcu_state.expedited_wq irq_context: 0 pernet_ops_rwsem rcu_read_lock &rcu_state.expedited_wq &p->pi_lock irq_context: 0 pernet_ops_rwsem rcu_read_lock &rcu_state.expedited_wq &p->pi_lock &rq->__lock irq_context: 0 pernet_ops_rwsem rcu_read_lock &rcu_state.expedited_wq &p->pi_lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 sk_lock-AF_INET6 batched_entropy_u32.lock crngs.lock irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem rtnl_mutex batched_entropy_u8.lock irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem rtnl_mutex kfence_freelist_lock irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem rtnl_mutex &meta->lock irq_context: 0 sk_lock-AF_INET6 &list->lock#24 irq_context: 0 sk_lock-AF_INET6 krc.lock &obj_hash[i].lock irq_context: 0 sk_lock-AF_INET6 krc.lock &base->lock irq_context: 0 sk_lock-AF_INET6 krc.lock &base->lock &obj_hash[i].lock irq_context: 0 slock-AF_INET6 &sk->sk_lock.wq irq_context: 0 slock-AF_INET6 &sk->sk_lock.wq &p->pi_lock irq_context: 0 slock-AF_INET6 &sk->sk_lock.wq &p->pi_lock &rq->__lock irq_context: 0 slock-AF_INET6 &sk->sk_lock.wq &p->pi_lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 &sb->s_type->i_mutex_key#8 &sbi->s_writepages_rwsem jbd2_handle &journal->j_wait_reserved irq_context: 0 &sb->s_type->i_mutex_key#10 sk_lock-AF_INET6/1 pool_lock#2 irq_context: 0 &sb->s_type->i_mutex_key#10 sk_lock-AF_INET6/1 &c->lock irq_context: 0 &sb->s_type->i_mutex_key#10 sk_lock-AF_INET6/1 &obj_hash[i].lock irq_context: 0 &sb->s_type->i_mutex_key#10 sk_lock-AF_INET6/1 &base->lock irq_context: 0 &sb->s_type->i_mutex_key#10 sk_lock-AF_INET6/1 &base->lock &obj_hash[i].lock irq_context: 0 &sb->s_type->i_mutex_key#10 sk_lock-AF_INET6/1 &rq->__lock irq_context: 0 &sb->s_type->i_mutex_key#10 sk_lock-AF_INET6/1 &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 &sb->s_type->i_mutex_key#10 sk_lock-AF_INET6/1 fill_pool_map-wait-type-override &rq->__lock irq_context: 0 &sb->s_type->i_mutex_key#10 sk_lock-AF_INET6/1 fill_pool_map-wait-type-override &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 &sb->s_type->i_mutex_key#10 sk_lock-AF_INET6/1 fs_reclaim irq_context: 0 &sb->s_type->i_mutex_key#10 sk_lock-AF_INET6/1 fs_reclaim mmu_notifier_invalidate_range_start irq_context: 0 &sb->s_type->i_mutex_key#10 sk_lock-AF_INET6/1 rcu_read_lock rcu_read_lock &____s->seqcount#7 irq_context: 0 &sb->s_type->i_mutex_key#10 sk_lock-AF_INET6/1 rcu_read_lock rcu_read_lock &ct->lock irq_context: 0 &sb->s_type->i_mutex_key#10 sk_lock-AF_INET6/1 rcu_read_lock rcu_read_lock &____s->seqcount#10 irq_context: 0 &sb->s_type->i_mutex_key#10 sk_lock-AF_INET6/1 rcu_read_lock rcu_read_lock rcu_read_lock_bh tk_core.seq.seqcount irq_context: 0 &sb->s_type->i_mutex_key#10 sk_lock-AF_INET6/1 rcu_read_lock rcu_read_lock rcu_read_lock_bh &list->lock#5 irq_context: 0 &sb->s_type->i_mutex_key#10 sk_lock-AF_INET6/1 tk_core.seq.seqcount irq_context: 0 &sb->s_type->i_mutex_key#10 sk_lock-AF_INET6/1 &list->lock#24 irq_context: 0 &sb->s_type->i_mutex_key#10 sk_lock-AF_INET6/1 &____s->seqcount irq_context: 0 &sb->s_type->i_mutex_key#10 sk_lock-AF_INET6/1 rcu_read_lock pool_lock#2 irq_context: 0 &sb->s_type->i_mutex_key#10 sk_lock-AF_INET6/1 &n->list_lock irq_context: 0 &sb->s_type->i_mutex_key#10 sk_lock-AF_INET6/1 &n->list_lock &c->lock irq_context: 0 &sb->s_type->i_mutex_key#10 sk_lock-AF_INET6/1 krc.lock irq_context: 0 &sb->s_type->i_mutex_key#10 sk_lock-AF_INET6/1 &obj_hash[i].lock pool_lock irq_context: 0 &sb->s_type->i_mutex_key#10 sk_lock-AF_INET6/1 krc.lock &obj_hash[i].lock irq_context: 0 &sb->s_type->i_mutex_key#10 sk_lock-AF_INET6/1 krc.lock &base->lock irq_context: 0 &sb->s_type->i_mutex_key#10 sk_lock-AF_INET6/1 krc.lock &base->lock &obj_hash[i].lock irq_context: 0 &sb->s_type->i_mutex_key#10 sk_lock-AF_INET6/1 fill_pool_map-wait-type-override pool_lock#2 irq_context: 0 &sb->s_type->i_mutex_key#10 sk_lock-AF_INET6/1 fill_pool_map-wait-type-override pool_lock irq_context: 0 &sb->s_type->i_mutex_key#8 rcu_read_lock fill_pool_map-wait-type-override pool_lock#2 irq_context: 0 &sb->s_type->i_mutex_key#8 rcu_read_lock fill_pool_map-wait-type-override pool_lock irq_context: 0 &sb->s_type->i_mutex_key#10 sk_lock-AF_INET6/1 fill_pool_map-wait-type-override &c->lock irq_context: 0 &sb->s_type->i_mutex_key#10 sk_lock-AF_INET6/1 fill_pool_map-wait-type-override &____s->seqcount#2 irq_context: 0 &sb->s_type->i_mutex_key#10 sk_lock-AF_INET6/1 fill_pool_map-wait-type-override &____s->seqcount irq_context: 0 &sb->s_type->i_mutex_key#8 fill_pool_map-wait-type-override pool_lock#2 irq_context: 0 &sb->s_type->i_mutex_key#8 fill_pool_map-wait-type-override pool_lock irq_context: 0 &sb->s_type->i_mutex_key#8 pool_lock irq_context: 0 &sb->s_type->i_mutex_key#10 sk_lock-AF_INET6/1 rcu_read_lock rhashtable_bucket irq_context: 0 &sb->s_type->i_mutex_key#8 fill_pool_map-wait-type-override rcu_read_lock rcu_node_0 irq_context: 0 &sb->s_type->i_mutex_key#8 fill_pool_map-wait-type-override rcu_read_lock &rq->__lock irq_context: 0 &sb->s_type->i_mutex_key#8 fill_pool_map-wait-type-override rcu_read_lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 &sb->s_type->i_mutex_key#8 fill_pool_map-wait-type-override &c->lock irq_context: 0 &sb->s_type->i_mutex_key#8 fill_pool_map-wait-type-override &____s->seqcount#2 irq_context: 0 &sb->s_type->i_mutex_key#8 fill_pool_map-wait-type-override &____s->seqcount irq_context: 0 &sb->s_type->i_mutex_key#10 sk_lock-AF_INET6/1 rcu_read_lock rcu_node_0 irq_context: 0 &sb->s_type->i_mutex_key#10 sk_lock-AF_INET6/1 rcu_read_lock &rq->__lock irq_context: 0 &sb->s_type->i_mutex_key#10 sk_lock-AF_INET6/1 rcu_read_lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 &sb->s_type->i_mutex_key#10 sk_lock-AF_INET6/1 rcu_read_lock rcu_read_lock &pool->lock irq_context: 0 &sb->s_type->i_mutex_key#10 sk_lock-AF_INET6/1 rcu_read_lock rcu_read_lock &pool->lock &obj_hash[i].lock irq_context: 0 &sb->s_type->i_mutex_key#10 sk_lock-AF_INET6/1 rcu_read_lock rcu_read_lock &pool->lock &p->pi_lock irq_context: 0 &sb->s_type->i_mutex_key#10 sk_lock-AF_INET6/1 rcu_read_lock rcu_read_lock &pool->lock &p->pi_lock &rq->__lock irq_context: 0 &sb->s_type->i_mutex_key#10 sk_lock-AF_INET6/1 rcu_read_lock rcu_read_lock &pool->lock &p->pi_lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem krc.lock &obj_hash[i].lock irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem krc.lock &base->lock irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem krc.lock &base->lock &obj_hash[i].lock irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem &ht->mutex rcu_read_lock pool_lock#2 irq_context: 0 &sb->s_type->i_mutex_key#10 sk_lock-AF_INET6/1 rcu_read_lock &rcu_state.expedited_wq irq_context: 0 &sb->s_type->i_mutex_key#10 sk_lock-AF_INET6/1 rcu_read_lock &rcu_state.expedited_wq &p->pi_lock irq_context: 0 &sb->s_type->i_mutex_key#10 sk_lock-AF_INET6/1 rcu_read_lock &rcu_state.expedited_wq &p->pi_lock &rq->__lock irq_context: 0 &sb->s_type->i_mutex_key#10 sk_lock-AF_INET6/1 rcu_read_lock &rcu_state.expedited_wq &p->pi_lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 &sb->s_type->i_mutex_key#8 &sbi->s_writepages_rwsem jbd2_handle &folio_wait_table[i] irq_context: 0 &sb->s_type->i_mutex_key#8 &sbi->s_writepages_rwsem jbd2_handle &folio_wait_table[i] &p->pi_lock irq_context: 0 &sb->s_type->i_mutex_key#8 &sbi->s_writepages_rwsem jbd2_handle &folio_wait_table[i] &p->pi_lock &rq->__lock irq_context: 0 &sb->s_type->i_mutex_key#8 &sbi->s_writepages_rwsem jbd2_handle &folio_wait_table[i] &p->pi_lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 &sb->s_type->i_mutex_key#10 sk_lock-AF_INET6/1 sctp_assocs_id_lock irq_context: 0 &sb->s_type->i_mutex_key#10 sk_lock-AF_INET6/1 sctp_assocs_id_lock &obj_hash[i].lock irq_context: 0 &sb->s_type->i_mutex_key#10 sk_lock-AF_INET6/1 sctp_assocs_id_lock pool_lock#2 irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem rtnl_mutex _xmit_TUNNEL irq_context: 0 &sb->s_type->i_mutex_key#10 &net->sctp.addr_wq_lock slock-AF_INET6/1 krc.lock irq_context: 0 &sb->s_type->i_mutex_key#10 &net->sctp.addr_wq_lock slock-AF_INET6/1 &sctp_port_hashtable[i].lock irq_context: 0 &sb->s_type->i_mutex_key#10 &net->sctp.addr_wq_lock slock-AF_INET6/1 &sctp_port_hashtable[i].lock &obj_hash[i].lock irq_context: 0 &sb->s_type->i_mutex_key#10 &net->sctp.addr_wq_lock slock-AF_INET6/1 &sctp_port_hashtable[i].lock pool_lock#2 irq_context: 0 (wq_completion)events (work_completion)(&(&krcp->monitor_work)->work) rcu_callback &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem rtnl_mutex &root->kernfs_rwsem &____s->seqcount irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem rtnl_mutex &root->kernfs_rwsem rcu_read_lock pool_lock#2 irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem rtnl_mutex kernfs_idr_lock &obj_hash[i].lock irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem rtnl_mutex kernfs_idr_lock pool_lock#2 irq_context: 0 pernet_ops_rwsem rcu_read_lock &rcu_state.gp_wq irq_context: 0 pernet_ops_rwsem rcu_read_lock &rcu_state.gp_wq &p->pi_lock irq_context: 0 pernet_ops_rwsem rcu_read_lock &rcu_state.gp_wq &p->pi_lock &rq->__lock irq_context: 0 pernet_ops_rwsem rcu_read_lock &rcu_state.gp_wq &p->pi_lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)events (work_completion)(&(&krcp->monitor_work)->work) rcu_callback per_cpu_ptr(&cgroup_rstat_cpu_lock, cpu) irq_context: 0 &sb->s_type->i_mutex_key#8 &sbi->s_writepages_rwsem jbd2_handle mmu_notifier_invalidate_range_start &rq->__lock irq_context: 0 &sb->s_type->i_mutex_key#8 &sbi->s_writepages_rwsem jbd2_handle mmu_notifier_invalidate_range_start &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 &sb->s_type->i_mutex_key#8 &____s->seqcount irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem rtnl_mutex rcu_read_lock &pool->lock/1 irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem rtnl_mutex rcu_read_lock &pool->lock/1 &obj_hash[i].lock irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem rtnl_mutex rcu_read_lock &pool->lock/1 pool_lock#2 irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem rtnl_mutex rcu_read_lock &pool->lock/1 &p->pi_lock irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem rtnl_mutex rcu_read_lock &pool->lock/1 &p->pi_lock &rq->__lock irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem rtnl_mutex rcu_read_lock &pool->lock/1 &p->pi_lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem rtnl_mutex &idev->mc_lock _xmit_ETHER irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem rtnl_mutex &idev->mc_lock _xmit_ETHER &obj_hash[i].lock irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem rtnl_mutex &idev->mc_lock _xmit_ETHER pool_lock#2 irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem rtnl_mutex &idev->mc_lock _xmit_ETHER krc.lock irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem rtnl_mutex _xmit_ETHER irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem rtnl_mutex &root->kernfs_rwsem rcu_read_lock rcu_node_0 irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem rtnl_mutex &root->kernfs_rwsem rcu_read_lock &rq->__lock irq_context: 0 &sb->s_type->i_mutex_key#8 &sbi->s_writepages_rwsem &folio_wait_table[i] irq_context: 0 sb_writers#5 &type->i_mutex_dir_key#5/1 tomoyo_ss remove_cache_srcu &rq->__lock irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem &c->lock irq_context: 0 sb_writers#5 &type->i_mutex_dir_key#5/1 tomoyo_ss remove_cache_srcu rcu_node_0 irq_context: 0 sb_writers#5 &type->i_mutex_dir_key#5/1 tomoyo_ss remove_cache_srcu &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 sb_writers#5 &type->i_mutex_dir_key#5/1 tomoyo_ss remove_cache_srcu rcu_read_lock rcu_node_0 irq_context: 0 sb_writers#5 &type->i_mutex_dir_key#5/1 tomoyo_ss remove_cache_srcu rcu_read_lock &rq->__lock irq_context: 0 sb_writers#5 &type->i_mutex_dir_key#5/1 tomoyo_ss remove_cache_srcu rcu_read_lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 sb_writers#5 &type->i_mutex_dir_key#5/1 tomoyo_ss remove_cache_srcu rcu_read_lock &rcu_state.expedited_wq irq_context: 0 sb_writers#5 &type->i_mutex_dir_key#5/1 tomoyo_ss remove_cache_srcu rcu_read_lock &rcu_state.expedited_wq &p->pi_lock irq_context: 0 sb_writers#5 &type->i_mutex_dir_key#5/1 tomoyo_ss remove_cache_srcu rcu_read_lock &rcu_state.expedited_wq &p->pi_lock &rq->__lock irq_context: 0 sb_writers#5 &type->i_mutex_dir_key#5/1 tomoyo_ss remove_cache_srcu rcu_read_lock &rcu_state.expedited_wq &p->pi_lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem rtnl_mutex &n->list_lock irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem rtnl_mutex &n->list_lock &c->lock irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem rtnl_mutex &root->kernfs_rwsem &rq->__lock irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem rtnl_mutex &root->kernfs_rwsem &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem rtnl_mutex _xmit_IPGRE irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem rcu_state.barrier_mutex rcu_read_lock &rq->__lock irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem rcu_state.barrier_mutex rcu_read_lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: softirq (&hsr->announce_timer) rcu_read_lock &hsr->seqnr_lock rcu_read_lock rcu_read_lock_bh &sch->q.lock irq_context: 0 &sb->s_type->i_mutex_key#8 &sbi->s_writepages_rwsem jbd2_handle &ei->i_data_sem &pcp->lock &zone->lock irq_context: 0 &sb->s_type->i_mutex_key#8 &sbi->s_writepages_rwsem jbd2_handle &ei->i_data_sem &pcp->lock &zone->lock &____s->seqcount irq_context: softirq (&ndev->rs_timer) rcu_read_lock rcu_read_lock &____s->seqcount irq_context: softirq (&peer->timer_send_keepalive) irq_context: softirq (&peer->timer_send_keepalive) pool_lock#2 irq_context: softirq (&peer->timer_send_keepalive) &list->lock#17 irq_context: softirq (&peer->timer_send_keepalive) tk_core.seq.seqcount irq_context: softirq (&peer->timer_send_keepalive) rcu_read_lock_bh &r->producer_lock#2 irq_context: softirq (&peer->timer_send_keepalive) rcu_read_lock_bh rcu_read_lock &pool->lock irq_context: softirq (&peer->timer_send_keepalive) rcu_read_lock_bh rcu_read_lock &pool->lock &obj_hash[i].lock irq_context: softirq (&peer->timer_send_keepalive) rcu_read_lock_bh rcu_read_lock &pool->lock &p->pi_lock irq_context: softirq (&peer->timer_send_keepalive) rcu_read_lock_bh rcu_read_lock &pool->lock &p->pi_lock &rq->__lock irq_context: softirq (&peer->timer_send_keepalive) rcu_read_lock_bh rcu_read_lock &pool->lock &p->pi_lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: softirq (&peer->timer_send_keepalive) rcu_read_lock_bh &obj_hash[i].lock irq_context: softirq (&peer->timer_send_keepalive) rcu_read_lock_bh &base->lock irq_context: 0 (wq_completion)wg-crypt-wg2 (work_completion)(&({ do { const void *__vpp_verify = (typeof((worker) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __asm__ ("" : "=r"(__ptr) : "0"((typeof(*((worker))) *)((worker)))); (typeof((typeof(*((worker))) *)((worker)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); })->work) rcu_read_lock &pool->lock &p->pi_lock &rq->__lock irq_context: softirq (&peer->timer_send_keepalive) rcu_read_lock_bh &base->lock &obj_hash[i].lock irq_context: 0 (wq_completion)wg-crypt-wg2 (work_completion)(&({ do { const void *__vpp_verify = (typeof((worker) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __asm__ ("" : "=r"(__ptr) : "0"((typeof(*((worker))) *)((worker)))); (typeof((typeof(*((worker))) *)((worker)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); })->work) rcu_read_lock &pool->lock &p->pi_lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem &fn->fou_lock irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem rtnl_mutex &ul->lock#2 irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem ipvs->sync_mutex irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem hwsim_radio_lock irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem hwsim_netgroup_ida.xa_lock irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem rdma_nets_rwsem irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem rdma_nets_rwsem rdma_nets.xa_lock irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem devices_rwsem irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem rcu_read_lock rhashtable_bucket irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem nl_table_lock irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem nl_table_wait.lock irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem k-clock-AF_NETLINK irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem &nlk->wait irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem wlock-AF_NETLINK irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem &dir->lock irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem remove_cache_srcu irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem remove_cache_srcu quarantine_lock irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem remove_cache_srcu &c->lock irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem remove_cache_srcu &n->list_lock irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem remove_cache_srcu rcu_read_lock pool_lock#2 irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem remove_cache_srcu &obj_hash[i].lock irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem remove_cache_srcu pool_lock#2 irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem rdma_nets.xa_lock irq_context: 0 (wq_completion)events (work_completion)(&(&nsim_dev->trap_data->trap_report_dw)->work) &devlink->lock_key#6 rcu_node_0 irq_context: 0 &sb->s_type->i_mutex_key#8 rcu_read_lock &rcu_state.expedited_wq irq_context: 0 &sb->s_type->i_mutex_key#8 rcu_read_lock &rcu_state.expedited_wq &p->pi_lock irq_context: 0 &sb->s_type->i_mutex_key#8 rcu_read_lock &rcu_state.expedited_wq &p->pi_lock &rq->__lock irq_context: 0 &sb->s_type->i_mutex_key#8 rcu_read_lock &rcu_state.expedited_wq &p->pi_lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem rtnl_mutex dev_base_lock &xa->xa_lock#3 &obj_hash[i].lock irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem rtnl_mutex dev_base_lock &xa->xa_lock#3 pool_lock#2 irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem rtnl_mutex _xmit_LOOPBACK irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem rtnl_mutex &(&net->nexthop.notifier_chain)->rwsem irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem &hn->hn_lock irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem rtnl_mutex &caifn->caifdevs.lock irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem rtnl_mutex &caifn->caifdevs.lock rcu_state.exp_mutex rcu_node_0 irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem rtnl_mutex &caifn->caifdevs.lock rcu_state.exp_mutex &obj_hash[i].lock irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem rtnl_mutex &caifn->caifdevs.lock rcu_state.exp_mutex rcu_read_lock &pool->lock irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem rtnl_mutex &caifn->caifdevs.lock rcu_state.exp_mutex rcu_read_lock &pool->lock &obj_hash[i].lock irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem rtnl_mutex &caifn->caifdevs.lock rcu_state.exp_mutex rcu_read_lock &pool->lock &p->pi_lock irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem rtnl_mutex &caifn->caifdevs.lock rcu_state.exp_mutex &rnp->exp_wq[3] irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem rtnl_mutex &caifn->caifdevs.lock rcu_state.exp_mutex &rq->__lock irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem rtnl_mutex &caifn->caifdevs.lock rcu_state.exp_mutex &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem rtnl_mutex &caifn->caifdevs.lock &obj_hash[i].lock irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem rtnl_mutex &caifn->caifdevs.lock pool_lock#2 irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem rtnl_mutex &caifn->caifdevs.lock &this->info_list_lock irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem &pnettable->lock irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem &pnetids_ndev->lock irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem k-sk_lock-AF_INET6/1 irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem k-sk_lock-AF_INET6/1 k-slock-AF_INET6 irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem k-sk_lock-AF_INET6/1 rlock-AF_INET6 irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem k-sk_lock-AF_INET6/1 &list->lock#21 irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem &net->sctp.addr_wq_lock irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem &net->sctp.addr_wq_lock k-slock-AF_INET6/1 irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem &net->sctp.addr_wq_lock k-slock-AF_INET6/1 &sctp_ep_hashtable[i].lock irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem &net->sctp.addr_wq_lock k-slock-AF_INET6/1 &obj_hash[i].lock irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem &net->sctp.addr_wq_lock k-slock-AF_INET6/1 pool_lock#2 irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem &net->sctp.addr_wq_lock k-slock-AF_INET6/1 k-clock-AF_INET6 irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem &net->sctp.addr_wq_lock &obj_hash[i].lock irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem k-sk_lock-AF_INET6 &obj_hash[i].lock irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem k-slock-AF_INET6 &obj_hash[i].lock irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem k-slock-AF_INET6 pool_lock#2 irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem k-slock-AF_INET6 elock-AF_INET6 irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem rcu_read_lock &rq->__lock irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem rcu_read_lock rcu_node_0 irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem rcu_read_lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem rcu_node_0 irq_context: 0 &sb->s_type->i_mutex_key#10 &net->sctp.addr_wq_lock slock-AF_INET/1 &obj_hash[i].lock pool_lock irq_context: 0 sb_writers#4 &type->i_mutex_dir_key#3/1 &sb->s_type->i_mutex_key#8 mmu_notifier_invalidate_range_start &rq->__lock irq_context: 0 sb_writers#4 &type->i_mutex_dir_key#3/1 &sb->s_type->i_mutex_key#8 mmu_notifier_invalidate_range_start &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 sb_writers#4 &sb->s_type->i_lock_key#22 &xa->xa_lock#9 &obj_hash[i].lock pool_lock irq_context: 0 &fsnotify_mark_srcu &cfs_rq->removed.lock irq_context: 0 &fsnotify_mark_srcu &obj_hash[i].lock irq_context: 0 kn->active#5 remove_cache_srcu pool_lock#2 irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem k-sk_lock-AF_INET irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem k-sk_lock-AF_INET k-slock-AF_INET irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem k-sk_lock-AF_INET &obj_hash[i].lock irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem k-sk_lock-AF_INET k-clock-AF_INET irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem k-slock-AF_INET irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem k-slock-AF_INET &obj_hash[i].lock irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem k-slock-AF_INET pool_lock#2 irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem k-slock-AF_INET elock-AF_INET irq_context: 0 pernet_ops_rwsem remove_cache_srcu &rq->__lock irq_context: 0 pernet_ops_rwsem remove_cache_srcu &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)events fqdir_free_work rcu_state.barrier_mutex &pool->lock &p->pi_lock irq_context: 0 (wq_completion)events fqdir_free_work rcu_state.barrier_mutex &pool->lock &p->pi_lock &rq->__lock irq_context: 0 (wq_completion)events fqdir_free_work rcu_state.barrier_mutex &pool->lock &p->pi_lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 pernet_ops_rwsem &meta->lock irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem rcu_read_lock &rcu_state.expedited_wq irq_context: 0 pernet_ops_rwsem rcu_state.exp_mutex rcu_read_lock &rq->__lock irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem rcu_read_lock &rcu_state.expedited_wq &p->pi_lock irq_context: 0 pernet_ops_rwsem rcu_state.exp_mutex rcu_read_lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 pernet_ops_rwsem kernfs_idr_lock &obj_hash[i].lock irq_context: 0 pernet_ops_rwsem kernfs_idr_lock pool_lock#2 irq_context: 0 pernet_ops_rwsem &fsnotify_mark_srcu &rq->__lock irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem rcu_read_lock &rcu_state.expedited_wq &p->pi_lock &rq->__lock irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem rcu_read_lock &rcu_state.expedited_wq &p->pi_lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 tracepoints_mutex rcu_node_0 irq_context: 0 sk_lock-AF_INET remove_cache_srcu &____s->seqcount irq_context: 0 sk_lock-AF_INET remove_cache_srcu pool_lock#2 irq_context: 0 tracepoints_mutex rcu_read_lock rcu_node_0 irq_context: 0 sb_writers#8 kn->active#5 &kernfs_locks->open_file_mutex[count] &n->list_lock &c->lock irq_context: 0 bpf_stats_enabled_mutex &c->lock irq_context: 0 &fp->aux->used_maps_mutex &rq->__lock irq_context: softirq (&peer->timer_send_keepalive) &c->lock irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem rcu_read_lock &rcu_state.gp_wq irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem rcu_read_lock &rcu_state.gp_wq &p->pi_lock irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem rcu_read_lock &rcu_state.gp_wq &p->pi_lock &rq->__lock irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem rcu_read_lock &rcu_state.gp_wq &p->pi_lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 sb_writers#4 &sb->s_type->i_mutex_key#8 &sbi->s_writepages_rwsem &rsp->gp_wait &p->pi_lock irq_context: 0 sb_writers#4 &sb->s_type->i_mutex_key#8 &sbi->s_writepages_rwsem &rsp->gp_wait &p->pi_lock &rq->__lock irq_context: 0 sb_writers#4 &sb->s_type->i_mutex_key#8 &sbi->s_writepages_rwsem &rsp->gp_wait &p->pi_lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: softirq rcu_read_lock hwsim_radio_lock &pcp->lock &zone->lock irq_context: 0 tracepoints_mutex cpu_hotplug_lock jump_label_mutex text_mutex &rq->__lock irq_context: 0 pernet_ops_rwsem &cfs_rq->removed.lock irq_context: 0 pernet_ops_rwsem rcu_read_lock &cfs_rq->removed.lock irq_context: 0 pernet_ops_rwsem rcu_read_lock &obj_hash[i].lock irq_context: 0 &mm->mmap_lock sb_writers#4 &journal->j_state_lock &journal->j_wait_transaction_locked irq_context: 0 &mm->mmap_lock sb_writers#4 jbd2_handle irq_context: 0 &f->f_pos_lock sb_writers#4 &sb->s_type->i_mutex_key#8 jbd2_handle &journal->j_wait_updates &p->pi_lock irq_context: 0 &f->f_pos_lock sb_writers#4 &sb->s_type->i_mutex_key#8 jbd2_handle &journal->j_wait_updates &p->pi_lock &rq->__lock irq_context: 0 &f->f_pos_lock sb_writers#4 &sb->s_type->i_mutex_key#8 jbd2_handle &journal->j_wait_updates &p->pi_lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 &mm->mmap_lock sb_writers#4 &journal->j_state_lock irq_context: 0 &mm->mmap_lock sb_writers#4 &journal->j_state_lock tk_core.seq.seqcount irq_context: 0 &mm->mmap_lock sb_writers#4 &journal->j_state_lock &obj_hash[i].lock irq_context: 0 &mm->mmap_lock sb_writers#4 &journal->j_state_lock &base->lock irq_context: 0 &mm->mmap_lock sb_writers#4 &journal->j_state_lock &base->lock &obj_hash[i].lock irq_context: 0 &mm->mmap_lock sb_writers#4 jbd2_handle &ret->b_state_lock irq_context: 0 &mm->mmap_lock sb_writers#4 jbd2_handle mmu_notifier_invalidate_range_start irq_context: 0 &mm->mmap_lock sb_writers#4 jbd2_handle &____s->seqcount irq_context: 0 &mm->mmap_lock sb_writers#4 jbd2_handle pool_lock#2 irq_context: 0 &mm->mmap_lock sb_writers#4 jbd2_handle bit_wait_table + i irq_context: 0 &mm->mmap_lock sb_writers#4 jbd2_handle &rq->__lock irq_context: 0 &mm->mmap_lock sb_writers#4 jbd2_handle &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 &f->f_pos_lock sb_writers#4 &sb->s_type->i_mutex_key#8 jbd2_handle &ret->b_state_lock irq_context: 0 &f->f_pos_lock sb_writers#4 &sb->s_type->i_mutex_key#8 jbd2_handle bit_wait_table + i irq_context: 0 &mm->mmap_lock sb_writers#4 jbd2_handle &ret->b_state_lock &journal->j_list_lock irq_context: 0 &mm->mmap_lock sb_writers#4 jbd2_handle rcu_read_lock pool_lock#2 irq_context: 0 &mm->mmap_lock sb_writers#4 jbd2_handle &obj_hash[i].lock irq_context: 0 sb_writers#5 &type->i_mutex_dir_key#5 remove_cache_srcu pool_lock#2 irq_context: 0 rtnl_mutex stock_lock irq_context: 0 rtnl_mutex &tb->tb6_lock rcu_read_lock rcu_read_lock &pool->lock irq_context: 0 rtnl_mutex &tb->tb6_lock rcu_read_lock rcu_read_lock &pool->lock &obj_hash[i].lock irq_context: 0 rtnl_mutex &tb->tb6_lock rcu_read_lock rcu_read_lock &pool->lock pool_lock#2 irq_context: 0 rtnl_mutex &tb->tb6_lock rcu_read_lock rcu_read_lock &pool->lock &p->pi_lock irq_context: 0 rtnl_mutex &tb->tb6_lock rcu_read_lock rcu_read_lock &pool->lock &p->pi_lock &rq->__lock irq_context: 0 rtnl_mutex &tb->tb6_lock rcu_read_lock rcu_read_lock &pool->lock &p->pi_lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 rtnl_mutex rcu_read_lock_bh &ifa->lock irq_context: 0 &mm->mmap_lock &anon_vma->rwsem &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem &cfs_rq->removed.lock irq_context: 0 sk_lock-AF_INET6 fill_pool_map-wait-type-override &n->list_lock irq_context: 0 sk_lock-AF_INET6 fill_pool_map-wait-type-override &n->list_lock &c->lock irq_context: 0 (wq_completion)events (work_completion)(&rdev->wiphy_work) &rdev->wiphy.mtx &wdev->mtx &rdev->bss_lock fill_pool_map-wait-type-override &c->lock irq_context: 0 (wq_completion)events (work_completion)(&rdev->wiphy_work) &rdev->wiphy.mtx &wdev->mtx fill_pool_map-wait-type-override &n->list_lock irq_context: 0 (wq_completion)events (work_completion)(&rdev->wiphy_work) &rdev->wiphy.mtx &wdev->mtx &rdev->bss_lock fill_pool_map-wait-type-override &n->list_lock irq_context: 0 (wq_completion)events (work_completion)(&rdev->wiphy_work) &rdev->wiphy.mtx &wdev->mtx fill_pool_map-wait-type-override &n->list_lock &c->lock irq_context: 0 (wq_completion)events (work_completion)(&rdev->wiphy_work) &rdev->wiphy.mtx &wdev->mtx &rdev->bss_lock fill_pool_map-wait-type-override &n->list_lock &c->lock irq_context: 0 (wq_completion)events (work_completion)(&rdev->wiphy_work) &rdev->wiphy.mtx &wdev->mtx &rdev->bss_lock fill_pool_map-wait-type-override pool_lock#2 irq_context: 0 (wq_completion)events (work_completion)(&rdev->wiphy_work) &rdev->wiphy.mtx &wdev->mtx &rdev->bss_lock fill_pool_map-wait-type-override pool_lock irq_context: 0 &root->kernfs_rwsem pgd_lock irq_context: 0 &root->kernfs_rwsem stock_lock irq_context: 0 &root->kernfs_rwsem rcu_read_lock pool_lock#2 irq_context: 0 &root->kernfs_rwsem key irq_context: 0 &root->kernfs_rwsem pcpu_lock irq_context: 0 &root->kernfs_rwsem percpu_counters_lock irq_context: 0 &root->kernfs_rwsem pcpu_lock stock_lock irq_context: 0 &root->kernfs_rwsem pool_lock#2 irq_context: hardirq hrtimer_bases.lock fill_pool_map-wait-type-override &n->list_lock irq_context: hardirq hrtimer_bases.lock fill_pool_map-wait-type-override &n->list_lock &c->lock irq_context: 0 sb_writers#4 &type->i_mutex_dir_key#3 &obj_hash[i].lock pool_lock irq_context: 0 sk_lock-AF_INET6 rcu_read_lock batched_entropy_u32.lock irq_context: 0 sk_lock-AF_INET6 rcu_read_lock batched_entropy_u32.lock crngs.lock irq_context: 0 (wq_completion)events (work_completion)(&ht->run_work) rcu_read_lock &pool->lock irq_context: 0 (wq_completion)events (work_completion)(&ht->run_work) rcu_read_lock &pool->lock &obj_hash[i].lock irq_context: 0 sb_writers#4 &type->i_mutex_dir_key#3/1 stock_lock rcu_read_lock per_cpu_ptr(&cgroup_rstat_cpu_lock, cpu) irq_context: 0 &mm->mmap_lock per_cpu_ptr(&cgroup_rstat_cpu_lock, cpu) irq_context: 0 &f->f_pos_lock sb_writers#4 &sb->s_type->i_mutex_key#8 rcu_read_lock &xa->xa_lock#9 rcu_read_lock per_cpu_ptr(&cgroup_rstat_cpu_lock, cpu) irq_context: 0 &f->f_pos_lock sb_writers#4 &sb->s_type->i_mutex_key#8 stock_lock rcu_read_lock per_cpu_ptr(&cgroup_rstat_cpu_lock, cpu) irq_context: 0 pernet_ops_rwsem rtnl_mutex per_cpu_ptr(&cgroup_rstat_cpu_lock, cpu) irq_context: 0 &mm->mmap_lock ptlock_ptr(page)#2 lock#4 &lruvec->lru_lock per_cpu_ptr(&cgroup_rstat_cpu_lock, cpu) irq_context: 0 remove_cache_srcu rcu_read_lock &pcp->lock &zone->lock irq_context: 0 &sb->s_type->i_mutex_key#8 ima_extend_list_mutex &n->list_lock irq_context: 0 &sb->s_type->i_mutex_key#8 ima_extend_list_mutex &n->list_lock &c->lock irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem &sn->gssp_lock irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem &cd->hash_lock irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem cache_list_lock irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem cache_list_lock &cd->hash_lock irq_context: 0 sk_lock-AF_INET6 &mm->mmap_lock ptlock_ptr(page)#2 irq_context: 0 sk_lock-AF_INET6 &mm->mmap_lock sb_pagefaults irq_context: 0 sk_lock-AF_INET6 &mm->mmap_lock sb_pagefaults tk_core.seq.seqcount irq_context: 0 sk_lock-AF_INET6 &mm->mmap_lock sb_pagefaults mmu_notifier_invalidate_range_start irq_context: 0 sk_lock-AF_INET6 &mm->mmap_lock sb_pagefaults &journal->j_state_lock irq_context: 0 sk_lock-AF_INET6 &mm->mmap_lock sb_pagefaults jbd2_handle irq_context: 0 sk_lock-AF_INET6 &mm->mmap_lock sb_pagefaults jbd2_handle &ei->i_raw_lock irq_context: 0 sk_lock-AF_INET6 &mm->mmap_lock sb_pagefaults jbd2_handle &journal->j_wait_updates irq_context: 0 sk_lock-AF_INET6 &mm->mmap_lock sb_pagefaults &obj_hash[i].lock irq_context: 0 sk_lock-AF_INET6 &mm->mmap_lock sb_pagefaults mapping.invalidate_lock irq_context: 0 sk_lock-AF_INET6 &mm->mmap_lock sb_pagefaults mapping.invalidate_lock &mapping->private_lock irq_context: 0 sk_lock-AF_INET6 &mm->mmap_lock sb_pagefaults mapping.invalidate_lock &mapping->private_lock rcu_read_lock &memcg->move_lock irq_context: 0 sk_lock-AF_INET6 &mm->mmap_lock &mapping->private_lock irq_context: 0 sk_lock-AF_INET6 &mm->mmap_lock &mapping->private_lock rcu_read_lock &memcg->move_lock irq_context: 0 sk_lock-AF_INET6 rcu_read_lock per_cpu_ptr(&cgroup_rstat_cpu_lock, cpu) irq_context: 0 &sb->s_type->i_mutex_key#10 sk_lock-AF_INET6/1 &____s->seqcount#2 irq_context: softirq (&net->sctp.addr_wq_timer) &net->sctp.addr_wq_lock &base->lock irq_context: softirq (&net->sctp.addr_wq_timer) &net->sctp.addr_wq_lock &base->lock &obj_hash[i].lock irq_context: 0 (wq_completion)events (work_completion)(&ht->run_work) &ht->mutex &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 &sbi->s_writepages_rwsem jbd2_handle &ei->i_data_sem &sbi->s_md_lock irq_context: 0 &sbi->s_writepages_rwsem jbd2_handle &ei->i_data_sem &bgl->locks[i].lock &sbi->s_md_lock irq_context: 0 &sbi->s_writepages_rwsem jbd2_handle &ei->i_data_sem &bgl->locks[i].lock &ei->i_prealloc_lock irq_context: 0 &sbi->s_writepages_rwsem jbd2_handle &ei->i_data_sem &rq->__lock irq_context: 0 &sbi->s_writepages_rwsem jbd2_handle &ei->i_data_sem &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 &sbi->s_writepages_rwsem jbd2_handle &ei->i_data_sem &ret->b_state_lock irq_context: 0 &sbi->s_writepages_rwsem jbd2_handle &ei->i_data_sem &ret->b_state_lock &journal->j_list_lock irq_context: 0 &sbi->s_writepages_rwsem jbd2_handle rcu_read_lock rcu_read_lock per_cpu_ptr(&cgroup_rstat_cpu_lock, cpu) irq_context: 0 &sbi->s_writepages_rwsem rcu_read_lock &base->lock irq_context: 0 &sbi->s_writepages_rwsem rcu_read_lock &base->lock &obj_hash[i].lock irq_context: 0 &sbi->s_writepages_rwsem rcu_read_lock rcu_node_0 irq_context: 0 &sbi->s_writepages_rwsem rcu_read_lock &rq->__lock irq_context: 0 &sbi->s_writepages_rwsem rcu_read_lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 rtnl_mutex rcu_read_lock rcu_read_lock &pool->lock/1 &p->pi_lock &cfs_rq->removed.lock irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem (&net->can.stattimer) irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem xfrm_state_gc_work irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem &net->xfrm.xfrm_state_lock irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem &hashinfo->lock#2 irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem k-clock-AF_INET6 irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem nl_table_lock irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem (work_completion)(&(&net->ipv6.addr_chk_work)->work) irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem &net->ipv6.ip6addrlbl_table.lock irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem &net->ipv6.ip6addrlbl_table.lock &obj_hash[i].lock irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem &net->ipv6.ip6addrlbl_table.lock pool_lock#2 irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem &net->ipv6.ip6addrlbl_table.lock krc.lock irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem ip6_fl_lock irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem rtnl_mutex &net->rules_mod_lock irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem percpu_counters_lock irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem (&net->ipv6.ip6_fib_timer) irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem rtnl_mutex (&mrt->ipmr_expire_timer) irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem rtnl_mutex (work_completion)(&ht->run_work) irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem rtnl_mutex &ht->mutex irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem rtnl_mutex &ht->mutex &obj_hash[i].lock irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem rtnl_mutex &ht->mutex pool_lock#2 irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem &____s->seqcount irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem __ip_vs_app_mutex irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem __ip_vs_app_mutex &obj_hash[i].lock irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem __ip_vs_app_mutex pool_lock#2 irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem __ip_vs_mutex irq_context: 0 pernet_ops_rwsem rcu_state.exp_mutex &rnp->exp_wq[3] irq_context: 0 dup_mmap_sem &mm->mmap_lock &mm->mmap_lock/1 ptlock_ptr(page)#2 ptlock_ptr(page)#2/1 rcu_read_lock &p->pi_lock irq_context: 0 dup_mmap_sem &mm->mmap_lock &mm->mmap_lock/1 ptlock_ptr(page)#2 ptlock_ptr(page)#2/1 rcu_read_lock &p->pi_lock &rq->__lock irq_context: 0 dup_mmap_sem &mm->mmap_lock &mm->mmap_lock/1 ptlock_ptr(page)#2 ptlock_ptr(page)#2/1 rcu_read_lock &p->pi_lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)inet_frag_wq (work_completion)(&fqdir->destroy_work) &ht->mutex &base->lock irq_context: 0 (wq_completion)inet_frag_wq (work_completion)(&fqdir->destroy_work) &ht->mutex &base->lock &obj_hash[i].lock irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem (&ipvs->dest_trash_timer) irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem (work_completion)(&(&ipvs->expire_nodest_conn_work)->work) irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem (work_completion)(&(&ipvs->defense_work)->work) irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem ipvs->est_mutex irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem ipvs->est_mutex pcpu_lock irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem ipvs->est_mutex &obj_hash[i].lock irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem ipvs->est_mutex pool_lock#2 irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem (work_completion)(&(&ipvs->est_reload_work)->work) irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem nfnl_subsys_ipset irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem recent_lock irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem hashlimit_mutex irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem trans_gc_work irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem nf_log_mutex irq_context: 0 pernet_ops_rwsem rcu_state.exp_mutex.wait_lock irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem (work_completion)(&(&cnet->ecache.dwork)->work) irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem &meta->lock irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem kfence_freelist_lock irq_context: 0 sb_writers#8 &type->i_mutex_dir_key#4 rcu_read_lock rcu_node_0 irq_context: 0 sb_writers#8 &type->i_mutex_dir_key#4 rcu_read_lock &rq->__lock irq_context: 0 sb_writers#8 &type->i_mutex_dir_key#4 rcu_read_lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem rtnl_mutex nf_connlabels_lock irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem nl_table_lock nl_table_wait.lock irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem &n->list_lock irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem &n->list_lock &c->lock irq_context: 0 sb_writers#4 &type->i_mutex_dir_key#3 jbd2_handle rcu_read_lock rcu_node_0 irq_context: 0 sb_writers#4 &type->i_mutex_dir_key#3 jbd2_handle rcu_read_lock &rq->__lock irq_context: 0 sb_writers#4 &type->i_mutex_dir_key#3 jbd2_handle rcu_read_lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem &k->list_lock irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem &root->kernfs_rwsem irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem &root->kernfs_rwsem &root->kernfs_iattr_rwsem irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem &root->kernfs_rwsem kernfs_idr_lock irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem &root->kernfs_rwsem &obj_hash[i].lock irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem &root->kernfs_rwsem pool_lock#2 irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem nf_hook_mutex irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem nf_hook_mutex cpu_hotplug_lock irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem &sn->pipefs_sb_lock irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem uevent_sock_mutex irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem uevent_sock_mutex fs_reclaim irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem uevent_sock_mutex fs_reclaim mmu_notifier_invalidate_range_start irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem uevent_sock_mutex pool_lock#2 irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem uevent_sock_mutex &____s->seqcount irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem uevent_sock_mutex rcu_read_lock pool_lock#2 irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem uevent_sock_mutex &obj_hash[i].lock irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem uevent_sock_mutex nl_table_lock irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem uevent_sock_mutex rlock-AF_NETLINK irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem uevent_sock_mutex rcu_read_lock &ei->socket.wq.wait irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem uevent_sock_mutex rcu_read_lock &ei->socket.wq.wait &ep->lock irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem uevent_sock_mutex nl_table_wait.lock irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem &root->kernfs_rwsem kernfs_idr_lock &obj_hash[i].lock irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem &root->kernfs_rwsem kernfs_idr_lock pool_lock#2 irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem sysfs_symlink_target_lock irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem kernfs_idr_lock irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem kernfs_idr_lock &obj_hash[i].lock irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem kernfs_idr_lock pool_lock#2 irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem tcp_metrics_lock irq_context: 0 sb_writers#4 &type->i_mutex_dir_key#3/1 &sb->s_type->i_mutex_key#8 &mapping->private_lock irq_context: 0 sb_writers#4 &type->i_mutex_dir_key#3/1 &type->i_mutex_dir_key#3 remove_cache_srcu irq_context: 0 sb_writers#4 &type->i_mutex_dir_key#3/1 &type->i_mutex_dir_key#3 remove_cache_srcu quarantine_lock irq_context: 0 sb_writers#4 &type->i_mutex_dir_key#3/1 &type->i_mutex_dir_key#3 remove_cache_srcu &c->lock irq_context: 0 sb_writers#4 &type->i_mutex_dir_key#3/1 &type->i_mutex_dir_key#3 remove_cache_srcu &n->list_lock irq_context: 0 sb_writers#4 &type->i_mutex_dir_key#3/1 &type->i_mutex_dir_key#3 remove_cache_srcu &obj_hash[i].lock irq_context: 0 sb_writers#4 &type->i_mutex_dir_key#3/1 &type->i_mutex_dir_key#3 remove_cache_srcu pool_lock#2 irq_context: 0 sb_writers#4 &type->i_mutex_dir_key#3/1 &type->i_mutex_dir_key#3 remove_cache_srcu rcu_read_lock pool_lock#2 irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem k-clock-AF_INET irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem (work_completion)(&net->xfrm.policy_hash_work) irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem &net->xfrm.xfrm_policy_lock irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem (work_completion)(&net->xfrm.state_hash_work) irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem rtnl_mutex &obj_hash[i].lock pool_lock irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem sysctl_lock &obj_hash[i].lock pool_lock irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem &list->lock#2 irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem &xa->xa_lock#3 irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem genl_sk_destructing_waitq.lock irq_context: 0 (wq_completion)netns net_cleanup_work rcu_state.barrier_mutex irq_context: 0 (wq_completion)netns net_cleanup_work rcu_state.barrier_mutex rcu_state.barrier_mutex.wait_lock irq_context: 0 (wq_completion)netns net_cleanup_work rcu_state.barrier_mutex &rq->__lock irq_context: 0 (wq_completion)netns net_cleanup_work rcu_state.barrier_mutex &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)netns net_cleanup_work &obj_hash[i].lock irq_context: 0 (wq_completion)netns net_cleanup_work pool_lock#2 irq_context: 0 (wq_completion)netns net_cleanup_work krc.lock irq_context: 0 (wq_completion)netns net_cleanup_work &dir->lock irq_context: 0 (wq_completion)netns net_cleanup_work &dir->lock &obj_hash[i].lock irq_context: 0 (wq_completion)netns net_cleanup_work &dir->lock pool_lock#2 irq_context: 0 (wq_completion)netns net_cleanup_work stock_lock irq_context: 0 pernet_ops_rwsem rtnl_mutex sysctl_lock fill_pool_map-wait-type-override pool_lock#2 irq_context: 0 pernet_ops_rwsem rtnl_mutex sysctl_lock fill_pool_map-wait-type-override pool_lock irq_context: 0 pernet_ops_rwsem rtnl_mutex &root->kernfs_rwsem kernfs_idr_lock fill_pool_map-wait-type-override &c->lock irq_context: 0 pernet_ops_rwsem rtnl_mutex &root->kernfs_rwsem kernfs_idr_lock fill_pool_map-wait-type-override &____s->seqcount irq_context: 0 pernet_ops_rwsem rtnl_mutex &root->kernfs_rwsem kernfs_idr_lock fill_pool_map-wait-type-override pool_lock#2 irq_context: 0 pernet_ops_rwsem rtnl_mutex &root->kernfs_rwsem kernfs_idr_lock fill_pool_map-wait-type-override pool_lock irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem rcu_state.exp_mutex &rnp->exp_wq[3] irq_context: 0 &p->lock &mm->mmap_lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem ovs_mutex nf_ct_proto_mutex &rq->__lock irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem ovs_mutex nf_ct_proto_mutex &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 vlan_ioctl_mutex rtnl_mutex lock kernfs_idr_lock &c->lock irq_context: 0 vlan_ioctl_mutex rtnl_mutex &root->kernfs_rwsem &root->kernfs_iattr_rwsem &rq->__lock irq_context: 0 vlan_ioctl_mutex rtnl_mutex &root->kernfs_rwsem &root->kernfs_iattr_rwsem &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 sk_lock-AF_INET6 rcu_read_lock rcu_read_lock rcu_read_lock &____s->seqcount#2 irq_context: 0 &sb->s_type->i_mutex_key#8 &sbi->s_writepages_rwsem jbd2_handle &mapping->i_mmap_rwsem &rq->__lock irq_context: 0 &sb->s_type->i_mutex_key#8 &sbi->s_writepages_rwsem jbd2_handle &mapping->i_mmap_rwsem &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 vlan_ioctl_mutex rtnl_mutex uevent_sock_mutex rcu_read_lock rcu_node_0 irq_context: 0 vlan_ioctl_mutex rtnl_mutex uevent_sock_mutex rcu_read_lock &rq->__lock irq_context: 0 vlan_ioctl_mutex rtnl_mutex uevent_sock_mutex rcu_read_lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 sk_lock-AF_INET6 rcu_read_lock rcu_read_lock &c->lock irq_context: 0 vlan_ioctl_mutex rtnl_mutex &idev->mc_lock &vlan_netdev_addr_lock_key &c->lock irq_context: 0 vlan_ioctl_mutex rtnl_mutex &idev->mc_lock &vlan_netdev_addr_lock_key &n->list_lock irq_context: 0 vlan_ioctl_mutex rtnl_mutex &idev->mc_lock &vlan_netdev_addr_lock_key &n->list_lock &c->lock irq_context: 0 vlan_ioctl_mutex rtnl_mutex &idev->mc_lock &rq->__lock irq_context: 0 vlan_ioctl_mutex rtnl_mutex &idev->mc_lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)events (work_completion)(&rdev->wiphy_work) &rdev->wiphy.mtx &wdev->mtx batched_entropy_u8.lock irq_context: 0 (wq_completion)events (work_completion)(&rdev->wiphy_work) &rdev->wiphy.mtx &wdev->mtx kfence_freelist_lock irq_context: 0 (wq_completion)events (work_completion)(&rdev->wiphy_work) &rdev->wiphy.mtx &wdev->mtx &meta->lock irq_context: 0 sk_lock-AF_INET6 sctp_assocs_id_lock &c->lock irq_context: 0 vlan_ioctl_mutex rtnl_mutex rcu_read_lock rcu_node_0 irq_context: 0 sk_lock-AF_INET6 fill_pool_map-wait-type-override &cfs_rq->removed.lock irq_context: 0 sk_lock-AF_INET6 fill_pool_map-wait-type-override &rq->__lock &cfs_rq->removed.lock irq_context: 0 vlan_ioctl_mutex rtnl_mutex rcu_read_lock &pool->lock irq_context: 0 vlan_ioctl_mutex rtnl_mutex rcu_read_lock &pool->lock &obj_hash[i].lock irq_context: 0 sk_lock-AF_INET6 &rcu_state.expedited_wq irq_context: 0 pernet_ops_rwsem rtnl_mutex fill_pool_map-wait-type-override &____s->seqcount#2 irq_context: 0 pernet_ops_rwsem rtnl_mutex &ndev->lock fill_pool_map-wait-type-override pool_lock#2 irq_context: 0 pernet_ops_rwsem rtnl_mutex &ndev->lock fill_pool_map-wait-type-override pool_lock irq_context: 0 (wq_completion)events (work_completion)(&pwq->unbound_release_work) rcu_state.exp_mutex.wait_lock irq_context: 0 (wq_completion)events (work_completion)(&pwq->unbound_release_work) &p->pi_lock irq_context: 0 (wq_completion)events (work_completion)(&pwq->unbound_release_work) &p->pi_lock &rq->__lock irq_context: 0 (wq_completion)events (work_completion)(&pwq->unbound_release_work) &p->pi_lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 sk_lock-AF_INET6 sctp_assocs_id_lock &____s->seqcount#2 irq_context: 0 sk_lock-AF_INET6 sctp_assocs_id_lock &____s->seqcount irq_context: 0 &sb->s_type->i_mutex_key#8 rcu_read_lock &obj_hash[i].lock pool_lock irq_context: 0 &sb->s_type->i_mutex_key#8 rcu_read_lock &rcu_state.gp_wq irq_context: 0 &sb->s_type->i_mutex_key#8 rcu_read_lock &rcu_state.gp_wq &p->pi_lock irq_context: 0 &sb->s_type->i_mutex_key#8 rcu_read_lock &rcu_state.gp_wq &p->pi_lock &rq->__lock irq_context: 0 &sb->s_type->i_mutex_key#8 rcu_read_lock &rcu_state.gp_wq &p->pi_lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 &sb->s_type->i_mutex_key#8 &sbi->s_writepages_rwsem rcu_read_lock &obj_hash[i].lock pool_lock irq_context: 0 &sb->s_type->i_mutex_key#8 &sbi->s_writepages_rwsem rcu_read_lock fill_pool_map-wait-type-override rcu_node_0 irq_context: 0 &sb->s_type->i_mutex_key#8 &sbi->s_writepages_rwsem rcu_read_lock fill_pool_map-wait-type-override &rq->__lock irq_context: 0 &sb->s_type->i_mutex_key#8 &sbi->s_writepages_rwsem rcu_read_lock fill_pool_map-wait-type-override &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 &sb->s_type->i_mutex_key#8 &sbi->s_writepages_rwsem rcu_read_lock &rcu_state.gp_wq irq_context: 0 &sb->s_type->i_mutex_key#8 &sbi->s_writepages_rwsem rcu_read_lock &rcu_state.gp_wq &p->pi_lock irq_context: 0 &sb->s_type->i_mutex_key#8 &sbi->s_writepages_rwsem rcu_read_lock &rcu_state.gp_wq &p->pi_lock &rq->__lock irq_context: 0 &sb->s_type->i_mutex_key#8 &sbi->s_writepages_rwsem rcu_read_lock &rcu_state.gp_wq &p->pi_lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 sk_lock-AF_INET6 &mm->mmap_lock sb_pagefaults mapping.invalidate_lock rcu_read_lock &memcg->move_lock irq_context: 0 sk_lock-AF_INET6 &mm->mmap_lock sb_pagefaults mapping.invalidate_lock rcu_read_lock &xa->xa_lock#9 irq_context: 0 &sb->s_type->i_mutex_key#8 &sbi->s_writepages_rwsem rcu_read_lock &rcu_state.expedited_wq &p->pi_lock &rq->__lock irq_context: 0 &sb->s_type->i_mutex_key#8 &sbi->s_writepages_rwsem rcu_read_lock &rcu_state.expedited_wq &p->pi_lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 &sb->s_type->i_mutex_key#8 rcu_read_lock fill_pool_map-wait-type-override &c->lock irq_context: 0 &sb->s_type->i_mutex_key#8 rcu_read_lock fill_pool_map-wait-type-override &____s->seqcount#2 irq_context: 0 &sb->s_type->i_mutex_key#8 rcu_read_lock fill_pool_map-wait-type-override &____s->seqcount irq_context: 0 &mm->mmap_lock &anon_vma->rwsem rcu_node_0 irq_context: 0 &mm->mmap_lock &anon_vma->rwsem &rcu_state.expedited_wq irq_context: 0 &mm->mmap_lock &anon_vma->rwsem &rcu_state.expedited_wq &p->pi_lock irq_context: 0 &mm->mmap_lock &anon_vma->rwsem &rcu_state.expedited_wq &p->pi_lock &rq->__lock irq_context: 0 &mm->mmap_lock &anon_vma->rwsem &rcu_state.expedited_wq &p->pi_lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 &sb->s_type->i_mutex_key#10 sk_lock-AF_INET6/1 rcu_read_lock rcu_read_lock rcu_node_0 irq_context: 0 &sb->s_type->i_mutex_key#10 sk_lock-AF_INET6/1 rcu_read_lock rcu_read_lock &rq->__lock irq_context: 0 &sb->s_type->i_mutex_key#10 sk_lock-AF_INET6/1 rcu_read_lock rcu_read_lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)ext4-rsv-conversion (work_completion)(&ei->i_rsv_conversion_work) rcu_read_lock &rcu_state.gp_wq irq_context: 0 (wq_completion)ext4-rsv-conversion (work_completion)(&ei->i_rsv_conversion_work) rcu_read_lock &rcu_state.gp_wq &p->pi_lock irq_context: 0 (wq_completion)ext4-rsv-conversion (work_completion)(&ei->i_rsv_conversion_work) rcu_read_lock &rcu_state.gp_wq &p->pi_lock &rq->__lock irq_context: 0 (wq_completion)ext4-rsv-conversion (work_completion)(&ei->i_rsv_conversion_work) rcu_read_lock &rcu_state.gp_wq &p->pi_lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)ext4-rsv-conversion (work_completion)(&ei->i_rsv_conversion_work) rcu_read_lock &rcu_state.expedited_wq irq_context: 0 (wq_completion)ext4-rsv-conversion (work_completion)(&ei->i_rsv_conversion_work) rcu_read_lock &rcu_state.expedited_wq &p->pi_lock irq_context: 0 (wq_completion)ext4-rsv-conversion (work_completion)(&ei->i_rsv_conversion_work) rcu_read_lock &rcu_state.expedited_wq &p->pi_lock &rq->__lock irq_context: 0 (wq_completion)ext4-rsv-conversion (work_completion)(&ei->i_rsv_conversion_work) rcu_read_lock &rcu_state.expedited_wq &p->pi_lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 &sb->s_type->i_mutex_key#8 &sbi->s_writepages_rwsem &folio_wait_table[i] &p->pi_lock irq_context: 0 &sb->s_type->i_mutex_key#8 &sbi->s_writepages_rwsem &folio_wait_table[i] &p->pi_lock &rq->__lock irq_context: 0 &sb->s_type->i_mutex_key#8 &sbi->s_writepages_rwsem &folio_wait_table[i] &p->pi_lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 &sb->s_type->i_mutex_key#8 &sbi->s_writepages_rwsem jbd2_handle &folio_wait_table[i] &p->pi_lock &cfs_rq->removed.lock irq_context: 0 pernet_ops_rwsem rtnl_mutex uevent_sock_mutex &cfs_rq->removed.lock irq_context: 0 &sb->s_type->i_mutex_key#10 sk_lock-AF_INET6/1 rcu_node_0 irq_context: 0 &sb->s_type->i_mutex_key#10 sk_lock-AF_INET6/1 &rcu_state.expedited_wq irq_context: 0 &sb->s_type->i_mutex_key#10 sk_lock-AF_INET6/1 &rcu_state.expedited_wq &p->pi_lock irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem rtnl_mutex sysctl_lock fill_pool_map-wait-type-override pool_lock#2 irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem rtnl_mutex sysctl_lock fill_pool_map-wait-type-override pool_lock irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem rtnl_mutex sysctl_lock fill_pool_map-wait-type-override &c->lock irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem rtnl_mutex sysctl_lock fill_pool_map-wait-type-override &____s->seqcount#2 irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem rtnl_mutex sysctl_lock fill_pool_map-wait-type-override &____s->seqcount irq_context: 0 &sbi->s_writepages_rwsem rcu_read_lock &memcg->move_lock irq_context: 0 &sbi->s_writepages_rwsem rcu_read_lock &xa->xa_lock#9 irq_context: 0 &sbi->s_writepages_rwsem rcu_read_lock &xa->xa_lock#9 &s->s_inode_wblist_lock irq_context: 0 &sbi->s_writepages_rwsem rcu_read_lock &rcu_state.expedited_wq irq_context: 0 &sbi->s_writepages_rwsem rcu_read_lock &rcu_state.expedited_wq &p->pi_lock irq_context: 0 &sbi->s_writepages_rwsem rcu_read_lock &obj_hash[i].lock pool_lock irq_context: 0 &sbi->s_writepages_rwsem rcu_read_lock &rcu_state.gp_wq irq_context: 0 &sbi->s_writepages_rwsem rcu_read_lock &rcu_state.gp_wq &p->pi_lock irq_context: 0 &sbi->s_writepages_rwsem rcu_read_lock &rcu_state.gp_wq &p->pi_lock &rq->__lock irq_context: 0 &sbi->s_writepages_rwsem rcu_read_lock &rcu_state.gp_wq &p->pi_lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 &sbi->s_writepages_rwsem rcu_read_lock &rcu_state.expedited_wq &p->pi_lock &rq->__lock irq_context: 0 &sbi->s_writepages_rwsem rcu_read_lock &rcu_state.expedited_wq &p->pi_lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)ext4-rsv-conversion (work_completion)(&ei->i_rsv_conversion_work) jbd2_handle rcu_read_lock rcu_node_0 irq_context: 0 (wq_completion)ext4-rsv-conversion (work_completion)(&ei->i_rsv_conversion_work) jbd2_handle rcu_read_lock &rq->__lock irq_context: 0 (wq_completion)ext4-rsv-conversion (work_completion)(&ei->i_rsv_conversion_work) jbd2_handle rcu_read_lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 sb_writers#4 &type->i_mutex_dir_key#3/1 jbd2_handle &ei->i_data_sem remove_cache_srcu irq_context: 0 sb_writers#4 &type->i_mutex_dir_key#3/1 jbd2_handle &ei->i_data_sem remove_cache_srcu quarantine_lock irq_context: 0 sb_writers#4 &type->i_mutex_dir_key#3/1 jbd2_handle &ei->i_data_sem remove_cache_srcu &c->lock irq_context: 0 sb_writers#4 &type->i_mutex_dir_key#3/1 jbd2_handle &ei->i_data_sem remove_cache_srcu &n->list_lock irq_context: 0 sb_writers#4 &type->i_mutex_dir_key#3/1 jbd2_handle &ei->i_data_sem remove_cache_srcu rcu_read_lock pool_lock#2 irq_context: 0 sb_writers#4 &type->i_mutex_dir_key#3/1 jbd2_handle &ei->i_data_sem remove_cache_srcu &obj_hash[i].lock irq_context: 0 sb_writers#4 &type->i_mutex_dir_key#3/1 jbd2_handle &ei->i_data_sem remove_cache_srcu pool_lock#2 irq_context: 0 remove_cache_srcu &rcu_state.expedited_wq irq_context: 0 remove_cache_srcu &rcu_state.expedited_wq &p->pi_lock irq_context: 0 remove_cache_srcu &rcu_state.expedited_wq &p->pi_lock &rq->__lock irq_context: 0 remove_cache_srcu &rcu_state.expedited_wq &p->pi_lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: softirq (&ndev->rs_timer) rcu_read_lock rcu_read_lock rcu_read_lock_bh &sch->q.lock irq_context: softirq (&ndev->rs_timer) rcu_read_lock rcu_read_lock rcu_read_lock_bh &hsr->seqnr_lock rcu_read_lock rcu_read_lock_bh &sch->q.lock irq_context: softirq (&ndev->rs_timer) rcu_read_lock rcu_read_lock rcu_read_lock_bh &hsr->seqnr_lock rcu_read_lock &n->list_lock irq_context: softirq (&ndev->rs_timer) rcu_read_lock rcu_read_lock rcu_read_lock_bh &hsr->seqnr_lock rcu_read_lock &n->list_lock &c->lock irq_context: 0 (wq_completion)wg-crypt-wg2 (work_completion)(&({ do { const void *__vpp_verify = (typeof((worker) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __asm__ ("" : "=r"(__ptr) : "0"((typeof(*((worker))) *)((worker)))); (typeof((typeof(*((worker))) *)((worker)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); })->work) &rq->__lock irq_context: 0 (wq_completion)ext4-rsv-conversion (work_completion)(&ei->i_rsv_conversion_work) &folio_wait_table[i] &p->pi_lock &rq->__lock &cfs_rq->removed.lock irq_context: 0 pernet_ops_rwsem rtnl_mutex &root->kernfs_rwsem kernfs_idr_lock &obj_hash[i].lock pool_lock irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem &p->pi_lock &cfs_rq->removed.lock irq_context: 0 (wq_completion)bat_events (work_completion)(&(&bat_priv->orig_work)->work) rcu_node_0 irq_context: 0 (wq_completion)bat_events (work_completion)(&(&bat_priv->orig_work)->work) &rcu_state.expedited_wq irq_context: 0 (wq_completion)bat_events (work_completion)(&(&bat_priv->orig_work)->work) &rcu_state.expedited_wq &p->pi_lock irq_context: 0 (wq_completion)bat_events (work_completion)(&(&bat_priv->orig_work)->work) &rcu_state.expedited_wq &p->pi_lock &rq->__lock irq_context: 0 (wq_completion)bat_events (work_completion)(&(&bat_priv->orig_work)->work) &rcu_state.expedited_wq &p->pi_lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem rtnl_mutex &idev->mc_lock &rq->__lock irq_context: 0 sb_writers#4 sb_internal jbd2_handle &ei->i_data_sem &____s->seqcount#2 irq_context: 0 sb_writers#4 sb_internal jbd2_handle &ei->i_data_sem &____s->seqcount irq_context: 0 &vma->vm_lock->lock &rcu_state.expedited_wq irq_context: 0 &vma->vm_lock->lock &rcu_state.expedited_wq &p->pi_lock irq_context: 0 &vma->vm_lock->lock &rcu_state.expedited_wq &p->pi_lock &rq->__lock irq_context: 0 &vma->vm_lock->lock &rcu_state.expedited_wq &p->pi_lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 pernet_ops_rwsem rtnl_mutex &caifn->caifdevs.lock rcu_state.exp_mutex &rnp->exp_wq[2] irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem rtnl_mutex sysctl_lock krc.lock &obj_hash[i].lock irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem rtnl_mutex sysctl_lock krc.lock &base->lock irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem rtnl_mutex sysctl_lock krc.lock &base->lock &obj_hash[i].lock irq_context: 0 (wq_completion)events (work_completion)(&(&krcp->monitor_work)->work) rcu_callback rcu_read_lock rcu_read_lock rcu_node_0 irq_context: 0 (wq_completion)events (work_completion)(&(&krcp->monitor_work)->work) rcu_callback rcu_read_lock rcu_read_lock &rq->__lock irq_context: 0 (wq_completion)events (work_completion)(&(&krcp->monitor_work)->work) rcu_callback rcu_read_lock rcu_node_0 irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem rtnl_mutex remove_cache_srcu irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem rtnl_mutex remove_cache_srcu quarantine_lock irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem rtnl_mutex remove_cache_srcu &c->lock irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem rtnl_mutex remove_cache_srcu &n->list_lock irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem rtnl_mutex remove_cache_srcu rcu_read_lock pool_lock#2 irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem rtnl_mutex remove_cache_srcu &obj_hash[i].lock irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem rtnl_mutex rcu_read_lock rcu_node_0 irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem rtnl_mutex rcu_read_lock &rq->__lock irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem rtnl_mutex rcu_read_lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem rtnl_mutex &caifn->caifdevs.lock rcu_state.exp_mutex &rnp->exp_wq[1] irq_context: softirq (&ndev->rs_timer) rcu_read_lock rcu_read_lock rcu_read_lock_bh rcu_read_lock &c->lock irq_context: softirq (&ndev->rs_timer) rcu_read_lock rcu_read_lock rcu_read_lock_bh rcu_read_lock &____s->seqcount#2 irq_context: softirq (&ndev->rs_timer) rcu_read_lock rcu_read_lock &pcp->lock &zone->lock irq_context: softirq (&ndev->rs_timer) rcu_read_lock rcu_read_lock rcu_read_lock_bh rcu_read_lock &____s->seqcount irq_context: softirq (&hsr->announce_timer) rcu_read_lock &hsr->seqnr_lock rcu_read_lock &____s->seqcount#2 irq_context: softirq (&hsr->announce_timer) rcu_read_lock &hsr->seqnr_lock rcu_read_lock &____s->seqcount irq_context: softirq (&p->forward_delay_timer) irq_context: softirq (&p->forward_delay_timer) &br->lock irq_context: softirq (&p->forward_delay_timer) &br->lock rcu_read_lock pool_lock#2 irq_context: softirq (&p->forward_delay_timer) &br->lock rcu_read_lock nl_table_lock irq_context: softirq (&p->forward_delay_timer) &br->lock rcu_read_lock &obj_hash[i].lock irq_context: softirq (&p->forward_delay_timer) &br->lock rcu_read_lock nl_table_wait.lock irq_context: 0 (wq_completion)events (work_completion)(&data->fib_event_work) &data->fib_lock remove_cache_srcu rcu_read_lock &____s->seqcount irq_context: 0 (wq_completion)wg-crypt-wg0#2 (work_completion)(&peer->transmit_packet_work) &peer->endpoint_lock rcu_read_lock_bh rcu_read_lock &obj_hash[i].lock irq_context: 0 (wq_completion)wg-crypt-wg0#2 (work_completion)(&peer->transmit_packet_work) &peer->endpoint_lock rcu_read_lock_bh rcu_read_lock &dir->lock#2 irq_context: 0 (wq_completion)wg-crypt-wg0#2 (work_completion)(&peer->transmit_packet_work) &peer->endpoint_lock rcu_read_lock_bh rcu_read_lock &tbl->lock irq_context: 0 (wq_completion)wg-crypt-wg0#2 (work_completion)(&peer->transmit_packet_work) &peer->endpoint_lock rcu_read_lock_bh rcu_read_lock &n->lock irq_context: 0 (wq_completion)wg-crypt-wg0#2 (work_completion)(&peer->transmit_packet_work) &peer->endpoint_lock rcu_read_lock_bh rcu_read_lock &____s->seqcount#9 irq_context: 0 (wq_completion)wg-crypt-wg0#4 (work_completion)(&({ do { const void *__vpp_verify = (typeof((worker) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __asm__ ("" : "=r"(__ptr) : "0"((typeof(*((worker))) *)((worker)))); (typeof((typeof(*((worker))) *)((worker)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); })->work) rcu_read_lock &pool->lock &p->pi_lock &rq->__lock irq_context: 0 (wq_completion)wg-crypt-wg0#4 (work_completion)(&({ do { const void *__vpp_verify = (typeof((worker) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __asm__ ("" : "=r"(__ptr) : "0"((typeof(*((worker))) *)((worker)))); (typeof((typeof(*((worker))) *)((worker)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); })->work) rcu_read_lock &pool->lock &p->pi_lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: softirq (&p->forward_delay_timer) &br->lock rcu_read_lock &c->lock irq_context: softirq (&p->forward_delay_timer) &br->lock rcu_read_lock quarantine_lock irq_context: softirq (&p->forward_delay_timer) &br->lock rcu_read_lock &n->list_lock irq_context: softirq (&p->forward_delay_timer) &br->lock rcu_read_lock &n->list_lock &c->lock irq_context: 0 pernet_ops_rwsem &xt[i].mutex &rq->__lock irq_context: 0 pernet_ops_rwsem remove_cache_srcu rcu_read_lock rcu_node_0 irq_context: 0 pernet_ops_rwsem remove_cache_srcu rcu_read_lock &rq->__lock irq_context: 0 pernet_ops_rwsem &net->ipv6.ip6addrlbl_table.lock &obj_hash[i].lock pool_lock irq_context: 0 pernet_ops_rwsem rcu_state.exp_mutex &obj_hash[i].lock pool_lock irq_context: 0 tracepoints_mutex cpu_hotplug_lock jump_label_mutex text_mutex &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)events free_ipc_work rcu_state.exp_mutex &rnp->exp_wq[0] irq_context: 0 dup_mmap_sem &mm->mmap_lock &mm->mmap_lock/1 remove_cache_srcu pool_lock#2 irq_context: 0 sb_writers#4 &type->i_mutex_dir_key#3 &ei->i_es_lock key#2 irq_context: 0 &f->f_pos_lock sb_writers#4 &sb->s_type->i_mutex_key#8 jbd2_handle &sbi->s_orphan_lock irq_context: 0 &f->f_pos_lock sb_writers#4 &sb->s_type->i_mutex_key#8 jbd2_handle &ei->i_es_lock irq_context: 0 &f->f_pos_lock sb_writers#4 &sb->s_type->i_mutex_key#8 jbd2_handle &ei->i_data_sem irq_context: 0 &f->f_pos_lock sb_writers#4 &sb->s_type->i_mutex_key#8 jbd2_handle &ei->i_data_sem mmu_notifier_invalidate_range_start irq_context: 0 &f->f_pos_lock sb_writers#4 &sb->s_type->i_mutex_key#8 jbd2_handle &ei->i_data_sem &c->lock irq_context: 0 &f->f_pos_lock sb_writers#4 &sb->s_type->i_mutex_key#8 jbd2_handle &ei->i_data_sem pool_lock#2 irq_context: 0 &f->f_pos_lock sb_writers#4 &sb->s_type->i_mutex_key#8 jbd2_handle &ei->i_data_sem &ei->i_es_lock irq_context: 0 &f->f_pos_lock sb_writers#4 &sb->s_type->i_mutex_key#8 jbd2_handle &ei->i_data_sem &ei->i_es_lock irq_context: 0 &f->f_pos_lock sb_writers#4 &sb->s_type->i_mutex_key#8 jbd2_handle &ei->i_data_sem &ei->i_es_lock pool_lock#2 irq_context: 0 &f->f_pos_lock sb_writers#4 &sb->s_type->i_mutex_key#8 jbd2_handle &ei->i_data_sem &ei->i_es_lock &sbi->s_es_lock irq_context: 0 &f->f_pos_lock sb_writers#4 &sb->s_type->i_mutex_key#8 jbd2_handle &ei->i_data_sem &obj_hash[i].lock irq_context: 0 &f->f_pos_lock sb_writers#4 &sb->s_type->i_mutex_key#8 jbd2_handle &ei->i_data_sem irq_context: 0 &f->f_pos_lock sb_writers#4 &sb->s_type->i_mutex_key#8 jbd2_handle &ei->i_data_sem mmu_notifier_invalidate_range_start irq_context: 0 &f->f_pos_lock sb_writers#4 &sb->s_type->i_mutex_key#8 jbd2_handle &ei->i_data_sem pool_lock#2 irq_context: 0 &f->f_pos_lock sb_writers#4 &sb->s_type->i_mutex_key#8 jbd2_handle &ei->i_data_sem &sb->s_type->i_lock_key#22 irq_context: 0 &f->f_pos_lock sb_writers#4 &sb->s_type->i_mutex_key#8 jbd2_handle &ei->i_data_sem &rq->__lock irq_context: 0 &f->f_pos_lock sb_writers#4 &sb->s_type->i_mutex_key#8 jbd2_handle &ei->i_data_sem &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 &f->f_pos_lock sb_writers#4 &sb->s_type->i_mutex_key#8 jbd2_handle &ei->i_data_sem &ei->i_raw_lock irq_context: 0 sk_lock-AF_CAIF irq_context: 0 &f->f_pos_lock sb_writers#4 &sb->s_type->i_mutex_key#8 jbd2_handle &ei->i_data_sem &lg->lg_mutex irq_context: 0 &f->f_pos_lock sb_writers#4 &sb->s_type->i_mutex_key#8 jbd2_handle &ei->i_data_sem &lg->lg_mutex &ei->i_prealloc_lock irq_context: 0 sk_lock-AF_CAIF &rq->__lock irq_context: 0 &f->f_pos_lock sb_writers#4 &sb->s_type->i_mutex_key#8 jbd2_handle &ei->i_data_sem &lg->lg_mutex rcu_read_lock &pa->pa_lock irq_context: 0 sk_lock-AF_CAIF &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 &f->f_pos_lock sb_writers#4 &sb->s_type->i_mutex_key#8 jbd2_handle &ei->i_data_sem &lg->lg_mutex &mapping->private_lock irq_context: 0 &f->f_pos_lock sb_writers#4 &sb->s_type->i_mutex_key#8 jbd2_handle &ei->i_data_sem &lg->lg_mutex &ret->b_state_lock irq_context: 0 &f->f_pos_lock sb_writers#4 &sb->s_type->i_mutex_key#8 jbd2_handle &ei->i_data_sem &lg->lg_mutex &ret->b_state_lock &journal->j_list_lock irq_context: 0 &f->f_pos_lock sb_writers#4 &sb->s_type->i_mutex_key#8 jbd2_handle &ei->i_data_sem &lg->lg_mutex &pa->pa_lock irq_context: 0 &f->f_pos_lock sb_writers#4 &sb->s_type->i_mutex_key#8 jbd2_handle &ei->i_data_sem &lg->lg_mutex &lg->lg_prealloc_lock irq_context: 0 &f->f_pos_lock sb_writers#4 &sb->s_type->i_mutex_key#8 jbd2_handle &ei->i_data_sem &obj_hash[i].lock irq_context: 0 &f->f_pos_lock sb_writers#4 &sb->s_type->i_mutex_key#8 jbd2_handle &ei->i_data_sem &ei->i_es_lock irq_context: 0 &f->f_pos_lock sb_writers#4 &sb->s_type->i_mutex_key#8 jbd2_handle &ei->i_data_sem &ei->i_es_lock irq_context: 0 &f->f_pos_lock sb_writers#4 &sb->s_type->i_mutex_key#8 jbd2_handle &ei->i_data_sem &ei->i_es_lock pool_lock#2 irq_context: 0 &f->f_pos_lock sb_writers#4 &sb->s_type->i_mutex_key#8 jbd2_handle &journal->j_list_lock irq_context: 0 &f->f_pos_lock sb_writers#4 &sb->s_type->i_mutex_key#8 &dd->lock irq_context: 0 &f->f_pos_lock sb_writers#4 &sb->s_type->i_mutex_key#8 rcu_read_lock &dd->lock irq_context: 0 &f->f_pos_lock sb_writers#4 &sb->s_type->i_mutex_key#8 rcu_read_lock &obj_hash[i].lock irq_context: 0 &f->f_pos_lock sb_writers#4 &sb->s_type->i_mutex_key#8 rcu_read_lock pool_lock#2 irq_context: 0 &f->f_pos_lock sb_writers#4 &sb->s_type->i_mutex_key#8 rcu_read_lock tk_core.seq.seqcount irq_context: 0 &f->f_pos_lock sb_writers#4 &sb->s_type->i_mutex_key#8 rcu_read_lock &virtscsi_vq->vq_lock irq_context: 0 &f->f_pos_lock sb_writers#4 &sb->s_type->i_mutex_key#8 &base->lock irq_context: 0 &f->f_pos_lock sb_writers#4 &sb->s_type->i_mutex_key#8 &base->lock &obj_hash[i].lock irq_context: 0 &f->f_pos_lock sb_writers#4 &sb->s_type->i_mutex_key#8 (&timer.timer) irq_context: 0 &f->f_pos_lock sb_writers#4 &sb->s_type->i_mutex_key#8 &journal->j_state_lock irq_context: 0 &f->f_pos_lock sb_writers#4 &sb->s_type->i_mutex_key#8 &journal->j_state_lock &journal->j_wait_commit irq_context: 0 &f->f_pos_lock sb_writers#4 &sb->s_type->i_mutex_key#8 &journal->j_state_lock &journal->j_wait_commit &p->pi_lock irq_context: 0 &f->f_pos_lock sb_writers#4 &sb->s_type->i_mutex_key#8 &journal->j_state_lock &journal->j_wait_commit &p->pi_lock &rq->__lock irq_context: 0 &f->f_pos_lock sb_writers#4 &sb->s_type->i_mutex_key#8 &journal->j_state_lock &journal->j_wait_commit &p->pi_lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 &f->f_pos_lock sb_writers#4 &sb->s_type->i_mutex_key#8 jbd2_handle irq_context: 0 &f->f_pos_lock sb_writers#4 &sb->s_type->i_mutex_key#8 &journal->j_wait_commit irq_context: 0 &f->f_pos_lock sb_writers#4 &sb->s_type->i_mutex_key#8 &journal->j_wait_done_commit irq_context: 0 &f->f_pos_lock sb_writers#4 &sb->s_type->i_mutex_key#8 &journal->j_state_lock tk_core.seq.seqcount irq_context: 0 &f->f_pos_lock sb_writers#4 &sb->s_type->i_mutex_key#8 &journal->j_state_lock &obj_hash[i].lock irq_context: 0 &f->f_pos_lock sb_writers#4 &sb->s_type->i_mutex_key#8 &journal->j_state_lock &base->lock irq_context: 0 &f->f_pos_lock sb_writers#4 &sb->s_type->i_mutex_key#8 &journal->j_state_lock &base->lock &obj_hash[i].lock irq_context: 0 &f->f_pos_lock sb_writers#4 &sb->s_type->i_mutex_key#8 jbd2_handle &ret->b_state_lock &journal->j_list_lock irq_context: 0 &f->f_pos_lock sb_writers#4 &sb->s_type->i_mutex_key#8 jbd2_handle &sbi->s_orphan_lock &ret->b_state_lock irq_context: 0 &f->f_pos_lock sb_writers#4 &sb->s_type->i_mutex_key#8 jbd2_handle &sbi->s_orphan_lock &ret->b_state_lock &journal->j_list_lock irq_context: 0 sb_writers#4 mapping.invalidate_lock irq_context: 0 sb_writers#4 mapping.invalidate_lock mmu_notifier_invalidate_range_start irq_context: 0 sb_writers#4 mapping.invalidate_lock &____s->seqcount irq_context: 0 sb_writers#4 mapping.invalidate_lock pool_lock#2 irq_context: 0 sb_writers#4 mapping.invalidate_lock stock_lock irq_context: 0 sb_writers#4 mapping.invalidate_lock &xa->xa_lock#9 irq_context: 0 sb_writers#4 mapping.invalidate_lock lock#4 irq_context: 0 sb_writers#4 mapping.invalidate_lock &xa->xa_lock#9 stock_lock irq_context: 0 sb_writers#4 mapping.invalidate_lock &xa->xa_lock#9 pool_lock#2 irq_context: 0 sb_writers#4 mapping.invalidate_lock &rq->__lock irq_context: 0 sb_writers#4 mapping.invalidate_lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 sk_lock-AF_CAIF slock-AF_CAIF irq_context: 0 sk_lock-AF_CAIF &obj_hash[i].lock irq_context: 0 slock-AF_CAIF irq_context: 0 &sb->s_type->i_mutex_key#10 rlock-AF_CAIF irq_context: 0 &sb->s_type->i_mutex_key#10 sk_lock-AF_CAIF irq_context: 0 &sb->s_type->i_mutex_key#10 sk_lock-AF_CAIF slock-AF_CAIF irq_context: 0 &sb->s_type->i_mutex_key#10 sk_lock-AF_CAIF &this->info_list_lock irq_context: 0 &sb->s_type->i_mutex_key#10 sk_lock-AF_CAIF (console_sem).lock irq_context: 0 &sb->s_type->i_mutex_key#10 sk_lock-AF_CAIF rcu_state.exp_mutex rcu_node_0 irq_context: 0 &sb->s_type->i_mutex_key#10 sk_lock-AF_CAIF rcu_state.exp_mutex &obj_hash[i].lock irq_context: 0 &sb->s_type->i_mutex_key#10 sk_lock-AF_CAIF rcu_state.exp_mutex rcu_read_lock &pool->lock irq_context: 0 &sb->s_type->i_mutex_key#10 sk_lock-AF_CAIF rcu_state.exp_mutex rcu_read_lock &pool->lock &obj_hash[i].lock irq_context: 0 &sb->s_type->i_mutex_key#10 sk_lock-AF_CAIF rcu_state.exp_mutex rcu_read_lock &pool->lock &p->pi_lock irq_context: 0 &sb->s_type->i_mutex_key#10 sk_lock-AF_CAIF rcu_state.exp_mutex rcu_read_lock &pool->lock &p->pi_lock &rq->__lock irq_context: 0 sb_writers#4 mapping.invalidate_lock &ei->i_es_lock irq_context: 0 &sb->s_type->i_mutex_key#10 sk_lock-AF_CAIF rcu_state.exp_mutex rcu_read_lock &pool->lock &p->pi_lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 sb_writers#4 mapping.invalidate_lock tk_core.seq.seqcount irq_context: 0 &sb->s_type->i_mutex_key#10 sk_lock-AF_CAIF rcu_state.exp_mutex &rq->__lock irq_context: 0 sb_writers#4 mapping.invalidate_lock &dd->lock irq_context: 0 &sb->s_type->i_mutex_key#10 sk_lock-AF_CAIF rcu_state.exp_mutex &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 sb_writers#4 mapping.invalidate_lock rcu_read_lock &dd->lock irq_context: 0 sb_writers#4 mapping.invalidate_lock rcu_read_lock &obj_hash[i].lock irq_context: 0 sb_writers#4 mapping.invalidate_lock rcu_read_lock pool_lock#2 irq_context: 0 sb_writers#4 mapping.invalidate_lock rcu_read_lock rcu_node_0 irq_context: 0 sb_writers#4 mapping.invalidate_lock rcu_read_lock &rq->__lock irq_context: 0 sb_writers#4 mapping.invalidate_lock rcu_read_lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 sb_writers#4 mapping.invalidate_lock rcu_read_lock tk_core.seq.seqcount irq_context: 0 sb_writers#4 mapping.invalidate_lock rcu_read_lock &virtscsi_vq->vq_lock irq_context: 0 sb_writers#4 &folio_wait_table[i] irq_context: 0 sb_writers#4 sb_writers#4 mount_lock irq_context: 0 sb_writers#4 sb_writers#4 tk_core.seq.seqcount irq_context: 0 sb_writers#4 sb_writers#4 mmu_notifier_invalidate_range_start irq_context: 0 sb_writers#4 sb_writers#4 pool_lock#2 irq_context: 0 sb_writers#4 sb_writers#4 &journal->j_state_lock irq_context: 0 sb_writers#4 sb_writers#4 jbd2_handle irq_context: 0 sb_writers#4 sb_writers#4 jbd2_handle &ei->i_raw_lock irq_context: 0 sb_writers#4 sb_writers#4 jbd2_handle &journal->j_wait_updates irq_context: 0 sb_writers#4 sb_writers#4 &obj_hash[i].lock irq_context: 0 sb_writers#4 &____s->seqcount#2 irq_context: 0 sb_writers#4 &sb->s_type->i_mutex_key#8 jbd2_handle &sbi->s_orphan_lock irq_context: 0 sb_writers#4 &sb->s_type->i_mutex_key#8 jbd2_handle &ei->i_es_lock irq_context: 0 sb_writers#4 &sb->s_type->i_mutex_key#8 jbd2_handle &ei->i_data_sem irq_context: 0 sb_writers#4 &sb->s_type->i_mutex_key#8 jbd2_handle &ei->i_data_sem mmu_notifier_invalidate_range_start irq_context: 0 sb_writers#4 &sb->s_type->i_mutex_key#8 jbd2_handle &ei->i_data_sem pool_lock#2 irq_context: 0 sb_writers#4 &sb->s_type->i_mutex_key#8 jbd2_handle &ei->i_data_sem &ei->i_es_lock irq_context: 0 sb_writers#4 &sb->s_type->i_mutex_key#8 jbd2_handle &ei->i_data_sem &sb->s_type->i_lock_key#22 irq_context: 0 sb_writers#4 &sb->s_type->i_mutex_key#8 jbd2_handle &ei->i_data_sem &ei->i_raw_lock irq_context: 0 sb_writers#4 &sb->s_type->i_mutex_key#8 jbd2_handle &ei->i_data_sem &ei->i_prealloc_lock irq_context: 0 sb_writers#4 &sb->s_type->i_mutex_key#8 jbd2_handle &ei->i_data_sem rcu_node_0 irq_context: 0 sb_writers#4 &sb->s_type->i_mutex_key#8 jbd2_handle &ei->i_data_sem &rq->__lock irq_context: 0 sb_writers#4 &sb->s_type->i_mutex_key#8 jbd2_handle &ei->i_data_sem &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 sb_writers#4 &sb->s_type->i_mutex_key#8 jbd2_handle &ei->i_data_sem rcu_read_lock rcu_node_0 irq_context: 0 sb_writers#4 &sb->s_type->i_mutex_key#8 jbd2_handle &ei->i_data_sem &sbi->s_md_lock irq_context: 0 sb_writers#4 &sb->s_type->i_mutex_key#8 jbd2_handle &ei->i_data_sem &bgl->locks[i].lock &sbi->s_md_lock irq_context: 0 sb_writers#4 &sb->s_type->i_mutex_key#8 jbd2_handle &ei->i_data_sem &bgl->locks[i].lock &ei->i_prealloc_lock irq_context: 0 sb_writers#4 &sb->s_type->i_mutex_key#8 jbd2_handle &ei->i_data_sem &mapping->private_lock irq_context: 0 sb_writers#4 &sb->s_type->i_mutex_key#8 jbd2_handle &ei->i_data_sem &ret->b_state_lock irq_context: 0 sb_writers#4 &sb->s_type->i_mutex_key#8 jbd2_handle &ei->i_data_sem &ret->b_state_lock &journal->j_list_lock irq_context: 0 sb_writers#4 &sb->s_type->i_mutex_key#8 jbd2_handle &ei->i_data_sem key#3 irq_context: 0 sb_writers#4 &sb->s_type->i_mutex_key#8 jbd2_handle &ei->i_data_sem &pa->pa_lock#2 irq_context: 0 sb_writers#4 &sb->s_type->i_mutex_key#8 jbd2_handle &ei->i_data_sem &obj_hash[i].lock irq_context: 0 sb_writers#4 &sb->s_type->i_mutex_key#8 jbd2_handle &ei->i_data_sem &ei->i_es_lock irq_context: 0 sb_writers#4 &sb->s_type->i_mutex_key#8 jbd2_handle &ei->i_data_sem &ei->i_es_lock pool_lock#2 irq_context: 0 &sb->s_type->i_mutex_key#10 sk_lock-AF_CAIF rcu_state.exp_mutex &rnp->exp_wq[2] irq_context: 0 sb_writers#4 &sb->s_type->i_mutex_key#8 jbd2_handle &journal->j_list_lock irq_context: 0 sb_writers#4 &sb->s_type->i_mutex_key#8 &dd->lock irq_context: 0 sb_writers#4 &sb->s_type->i_mutex_key#8 rcu_read_lock &dd->lock irq_context: 0 sb_writers#4 &sb->s_type->i_mutex_key#8 rcu_read_lock &obj_hash[i].lock irq_context: 0 sb_writers#4 &sb->s_type->i_mutex_key#8 rcu_read_lock pool_lock#2 irq_context: 0 sb_writers#4 &sb->s_type->i_mutex_key#8 rcu_read_lock tk_core.seq.seqcount irq_context: 0 sb_writers#4 &sb->s_type->i_mutex_key#8 rcu_read_lock &virtscsi_vq->vq_lock irq_context: 0 sb_writers#4 &sb->s_type->i_mutex_key#8 &base->lock irq_context: 0 sb_writers#4 &sb->s_type->i_mutex_key#8 &base->lock &obj_hash[i].lock irq_context: 0 &f->f_pos_lock sb_writers#4 &sb->s_type->i_mutex_key#8 &ei->i_data_sem &ei->i_es_lock &c->lock irq_context: 0 sb_writers#4 &sb->s_type->i_mutex_key#8 (&timer.timer) irq_context: 0 sb_writers#4 &sb->s_type->i_mutex_key#8 &journal->j_state_lock &journal->j_wait_commit irq_context: 0 sb_writers#4 &sb->s_type->i_mutex_key#8 &journal->j_state_lock &journal->j_wait_commit &p->pi_lock irq_context: 0 sb_writers#4 &sb->s_type->i_mutex_key#8 &journal->j_state_lock &journal->j_wait_commit &p->pi_lock &rq->__lock irq_context: 0 sb_writers#4 &sb->s_type->i_mutex_key#8 &journal->j_state_lock &journal->j_wait_commit &p->pi_lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 sb_writers#4 &sb->s_type->i_mutex_key#8 jbd2_handle irq_context: 0 sb_writers#4 &sb->s_type->i_mutex_key#8 &journal->j_wait_commit irq_context: 0 sb_writers#4 &sb->s_type->i_mutex_key#8 &journal->j_wait_done_commit irq_context: 0 &f->f_pos_lock sb_writers#4 &sb->s_type->i_mutex_key#8 &journal->j_state_lock &journal->j_wait_transaction_locked irq_context: 0 &sb->s_type->i_mutex_key#10 sk_lock-AF_CAIF &obj_hash[i].lock irq_context: 0 &sb->s_type->i_mutex_key#10 sk_lock-AF_CAIF &ei->socket.wq.wait irq_context: 0 &sb->s_type->i_mutex_key#10 sk_lock-AF_CAIF clock-AF_CAIF irq_context: 0 &sb->s_type->i_mutex_key#10 sk_lock-AF_CAIF elock-AF_CAIF irq_context: 0 &sb->s_type->i_mutex_key#10 slock-AF_CAIF irq_context: 0 &sb->s_type->i_mutex_key#10 elock-AF_CAIF irq_context: 0 sb_writers#4 &sb->s_type->i_mutex_key#8 jbd2_handle &sbi->s_orphan_lock &ei->i_raw_lock irq_context: 0 &f->f_pos_lock sb_writers#4 &sb->s_type->i_mutex_key#8 per_cpu_ptr(&cgroup_rstat_cpu_lock, cpu) irq_context: 0 sb_writers#4 mapping.invalidate_lock lock#4 &lruvec->lru_lock irq_context: 0 sb_writers#4 &sb->s_type->i_mutex_key#8 jbd2_handle &ei->i_data_sem &ei->i_prealloc_lock &pa->pa_lock#2 irq_context: 0 sb_writers#4 &sb->s_type->i_mutex_key#8 jbd2_handle &ei->i_data_sem &ei->i_prealloc_lock irq_context: 0 sb_writers#4 &sb->s_type->i_mutex_key#8 rcu_read_lock &c->lock irq_context: softirq &fq->mq_flush_lock bit_wait_table + i &p->pi_lock &rq->__lock &cfs_rq->removed.lock irq_context: 0 sb_writers#4 mapping.invalidate_lock rcu_read_lock &rcu_state.gp_wq irq_context: 0 sb_writers#4 mapping.invalidate_lock rcu_read_lock &rcu_state.gp_wq &p->pi_lock irq_context: 0 &f->f_pos_lock sb_writers#4 &sb->s_type->i_mutex_key#8 &meta->lock irq_context: 0 sb_writers#4 mapping.invalidate_lock rcu_read_lock &rcu_state.gp_wq &p->pi_lock &rq->__lock irq_context: 0 sb_writers#4 mapping.invalidate_lock rcu_read_lock &rcu_state.gp_wq &p->pi_lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 sb_writers#4 mapping.invalidate_lock rcu_read_lock &rcu_state.expedited_wq irq_context: 0 sb_writers#4 mapping.invalidate_lock rcu_read_lock &rcu_state.expedited_wq &p->pi_lock irq_context: 0 sb_writers#4 mapping.invalidate_lock rcu_read_lock &rcu_state.expedited_wq &p->pi_lock &rq->__lock irq_context: 0 sb_writers#4 mapping.invalidate_lock rcu_read_lock &rcu_state.expedited_wq &p->pi_lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 rtnl_mutex &block->lock fs_reclaim irq_context: 0 rtnl_mutex &block->lock fs_reclaim mmu_notifier_invalidate_range_start irq_context: 0 rtnl_mutex &block->lock fs_reclaim mmu_notifier_invalidate_range_start &rq->__lock irq_context: 0 rtnl_mutex &block->lock fs_reclaim mmu_notifier_invalidate_range_start &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 rtnl_mutex &block->lock pool_lock#2 irq_context: 0 rtnl_mutex &chain->filter_chain_lock irq_context: 0 rtnl_mutex cls_mod_lock irq_context: 0 rtnl_mutex &chain->filter_chain_lock &block->lock irq_context: 0 rtnl_mutex &chain->filter_chain_lock &block->proto_destroy_lock irq_context: 0 rtnl_mutex &block->proto_destroy_lock irq_context: 0 rtnl_mutex &block->lock &obj_hash[i].lock irq_context: 0 rtnl_mutex &block->lock &rq->__lock irq_context: 0 rtnl_mutex &block->lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 rtnl_mutex &block->lock nl_table_lock irq_context: 0 rtnl_mutex &block->lock nl_table_wait.lock irq_context: 0 sb_writers#4 &sb->s_type->i_mutex_key#8 irq_context: 0 sb_writers#4 &sb->s_type->i_mutex_key#8 &sem->wait_lock irq_context: 0 sb_writers#4 &sb->s_type->i_mutex_key#8 &rq->__lock irq_context: 0 sb_writers#4 &sb->s_type->i_mutex_key#8 &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 sb_writers#4 &sb->s_type->i_mutex_key#8 jbd2_handle &sbi->s_orphan_lock &ret->b_state_lock irq_context: 0 sb_writers#4 &sb->s_type->i_mutex_key#8 jbd2_handle &sbi->s_orphan_lock &ret->b_state_lock &journal->j_list_lock irq_context: 0 sb_writers#4 &sb->s_type->i_mutex_key#8 &ei->i_es_lock irq_context: 0 sb_writers#4 &sb->s_type->i_mutex_key#8 tk_core.seq.seqcount irq_context: 0 sb_writers#4 &sb->s_type->i_mutex_key#8 fs_reclaim irq_context: 0 sb_writers#4 &sb->s_type->i_mutex_key#8 fs_reclaim mmu_notifier_invalidate_range_start irq_context: 0 sb_writers#4 &rq->__lock &cfs_rq->removed.lock irq_context: 0 sb_writers#4 &sb->s_type->i_mutex_key#8 fs_reclaim mmu_notifier_invalidate_range_start &rq->__lock irq_context: 0 sb_writers#4 &sb->s_type->i_mutex_key#8 fs_reclaim mmu_notifier_invalidate_range_start &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 sb_writers#4 &sb->s_type->i_mutex_key#8 &sb->s_type->i_lock_key#22 irq_context: 0 sb_writers#4 &sb->s_type->i_mutex_key#8 &sb->s_type->i_lock_key#22 &xa->xa_lock#9 irq_context: 0 sb_writers#4 &sb->s_type->i_mutex_key#8 lock#4 irq_context: 0 sb_writers#4 &sb->s_type->i_mutex_key#8 lock#5 irq_context: 0 sb_writers#4 &sb->s_type->i_mutex_key#8 &lruvec->lru_lock irq_context: 0 sb_writers#4 &sb->s_type->i_mutex_key#8 rcu_read_lock pool_lock#2 irq_context: 0 sb_writers#4 &sb->s_type->i_mutex_key#8 &obj_hash[i].lock irq_context: 0 sb_writers#4 &sb->s_type->i_mutex_key#8 &journal->j_state_lock irq_context: 0 sb_writers#4 &sb->s_type->i_mutex_key#8 &dd->lock irq_context: 0 sb_writers#4 &sb->s_type->i_mutex_key#8 rcu_read_lock &dd->lock irq_context: 0 sb_writers#4 &sb->s_type->i_mutex_key#8 rcu_read_lock &obj_hash[i].lock irq_context: 0 sb_writers#4 &sb->s_type->i_mutex_key#8 rcu_read_lock tk_core.seq.seqcount irq_context: 0 sb_writers#4 &sb->s_type->i_mutex_key#8 rcu_read_lock &virtscsi_vq->vq_lock irq_context: 0 sb_writers#4 &sb->s_type->i_mutex_key#8 rcu_read_lock rcu_node_0 irq_context: 0 sb_writers#4 &sb->s_type->i_mutex_key#8 rcu_read_lock &rq->__lock irq_context: 0 sb_writers#4 &sb->s_type->i_mutex_key#8 rcu_read_lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 sk_lock-AF_INET6 stock_lock irq_context: 0 sk_lock-AF_INET6 &f->f_lock irq_context: 0 sk_lock-AF_INET6 &f->f_lock fasync_lock irq_context: 0 sb_writers#4 &sb->s_type->i_mutex_key#8 &base->lock irq_context: 0 sb_writers#4 &sb->s_type->i_mutex_key#8 &base->lock &obj_hash[i].lock irq_context: 0 sb_writers#4 &sb->s_type->i_mutex_key#8 (&timer.timer) irq_context: 0 sb_writers#4 &sb->s_type->i_mutex_key#8 &journal->j_state_lock irq_context: 0 sb_writers#4 &sb->s_type->i_mutex_key#8 &journal->j_state_lock &journal->j_wait_commit irq_context: 0 sb_writers#4 &sb->s_type->i_mutex_key#8 &journal->j_state_lock &journal->j_wait_commit &p->pi_lock irq_context: 0 sb_writers#4 &sb->s_type->i_mutex_key#8 &journal->j_state_lock &journal->j_wait_commit &p->pi_lock &rq->__lock irq_context: 0 sb_writers#4 &sb->s_type->i_mutex_key#8 &journal->j_state_lock &journal->j_wait_commit &p->pi_lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 sb_writers#4 &sb->s_type->i_mutex_key#8 jbd2_handle irq_context: 0 sb_writers#4 &sb->s_type->i_mutex_key#8 &journal->j_wait_commit irq_context: 0 sb_writers#4 &sb->s_type->i_mutex_key#8 &journal->j_wait_done_commit irq_context: 0 sb_writers#4 sb_writers#4 &journal->j_state_lock irq_context: 0 sb_writers#4 sb_writers#4 &journal->j_state_lock tk_core.seq.seqcount irq_context: 0 sb_writers#4 sb_writers#4 &journal->j_state_lock &obj_hash[i].lock irq_context: 0 sb_writers#4 sb_writers#4 &journal->j_state_lock &base->lock irq_context: 0 sb_writers#4 sb_writers#4 &journal->j_state_lock &base->lock &obj_hash[i].lock irq_context: 0 sb_writers#4 sb_writers#4 jbd2_handle &ret->b_state_lock irq_context: 0 sb_writers#4 sb_writers#4 jbd2_handle bit_wait_table + i irq_context: 0 sb_writers#4 sb_writers#4 jbd2_handle &rq->__lock irq_context: 0 sb_writers#4 sb_writers#4 jbd2_handle &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 sb_writers#4 &sb->s_type->i_mutex_key#8 mmu_notifier_invalidate_range_start irq_context: 0 sb_writers#4 &sb->s_type->i_mutex_key#8 pool_lock#2 irq_context: 0 sb_writers#4 &sb->s_type->i_mutex_key#8 jbd2_handle irq_context: 0 sb_writers#4 &sb->s_type->i_mutex_key#8 jbd2_handle &ret->b_state_lock irq_context: 0 sb_writers#4 &sb->s_type->i_mutex_key#8 jbd2_handle &ret->b_state_lock &journal->j_list_lock irq_context: 0 sb_writers#4 &sb->s_type->i_mutex_key#8 jbd2_handle &ei->i_raw_lock irq_context: 0 sb_writers#4 &sb->s_type->i_mutex_key#8 &ei->i_es_lock key#2 irq_context: 0 sb_writers#4 &sb->s_type->i_mutex_key#8 &fq->mq_flush_lock irq_context: 0 sb_writers#4 &sb->s_type->i_mutex_key#8 &fq->mq_flush_lock tk_core.seq.seqcount irq_context: 0 sb_writers#4 &sb->s_type->i_mutex_key#8 &fq->mq_flush_lock &q->requeue_lock irq_context: 0 sb_writers#4 &sb->s_type->i_mutex_key#8 &fq->mq_flush_lock &obj_hash[i].lock irq_context: 0 sb_writers#4 &sb->s_type->i_mutex_key#8 &fq->mq_flush_lock rcu_read_lock &pool->lock irq_context: 0 sb_writers#4 &sb->s_type->i_mutex_key#8 &fq->mq_flush_lock rcu_read_lock &pool->lock &obj_hash[i].lock irq_context: 0 sb_writers#4 &sb->s_type->i_mutex_key#8 &fq->mq_flush_lock rcu_read_lock &pool->lock pool_lock#2 irq_context: 0 sb_writers#4 &sb->s_type->i_mutex_key#8 &fq->mq_flush_lock rcu_read_lock &pool->lock &p->pi_lock irq_context: 0 sb_writers#4 &sb->s_type->i_mutex_key#8 &fq->mq_flush_lock rcu_read_lock &pool->lock &p->pi_lock &rq->__lock irq_context: 0 sb_writers#4 &sb->s_type->i_mutex_key#8 &fq->mq_flush_lock rcu_read_lock &pool->lock &p->pi_lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 sb_writers#4 &sb->s_type->i_mutex_key#8 &x->wait#26 irq_context: 0 sb_writers#4 mapping.invalidate_lock fs_reclaim irq_context: 0 sb_writers#4 mapping.invalidate_lock fs_reclaim mmu_notifier_invalidate_range_start irq_context: 0 sb_writers#4 mapping.invalidate_lock &obj_hash[i].lock irq_context: 0 sb_writers#4 mapping.invalidate_lock rcu_read_lock &pool->lock irq_context: 0 sb_writers#4 mapping.invalidate_lock rcu_read_lock &pool->lock &obj_hash[i].lock irq_context: 0 sb_writers#4 mapping.invalidate_lock rcu_read_lock &pool->lock pool_lock#2 irq_context: 0 sb_writers#4 mapping.invalidate_lock rcu_read_lock &pool->lock &p->pi_lock irq_context: 0 sb_writers#4 mapping.invalidate_lock rcu_read_lock &pool->lock &p->pi_lock &rq->__lock irq_context: 0 sb_writers#4 mapping.invalidate_lock rcu_read_lock &pool->lock &p->pi_lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 sk_lock-AF_INET6 &tcp_hashinfo.bhash[i].lock stock_lock irq_context: 0 sk_lock-AF_INET6 &tcp_hashinfo.bhash[i].lock pool_lock#2 irq_context: 0 sk_lock-AF_INET6 &tcp_hashinfo.bhash[i].lock &tcp_hashinfo.bhash2[i].lock stock_lock irq_context: 0 sk_lock-AF_INET6 &tcp_hashinfo.bhash[i].lock &tcp_hashinfo.bhash2[i].lock &obj_hash[i].lock irq_context: 0 sk_lock-AF_INET6 &hashinfo->ehash_locks[i] irq_context: softirq rcu_read_lock rcu_read_lock slock-AF_INET6/1 irq_context: 0 sk_lock-AF_INET6 rcu_read_lock tcp_metrics_lock irq_context: 0 sk_lock-AF_INET6 rcu_read_lock tcp_metrics_lock pool_lock#2 irq_context: 0 sk_lock-AF_INET6 rcu_read_lock rcu_read_lock &new->fa_lock irq_context: 0 sk_lock-AF_INET6 rcu_read_lock rcu_read_lock &new->fa_lock &f->f_owner.lock irq_context: 0 (wq_completion)events (work_completion)(&w->work) cpu_hotplug_lock jump_label_mutex text_mutex &rq->__lock irq_context: 0 (wq_completion)events (work_completion)(&w->work) cpu_hotplug_lock jump_label_mutex text_mutex &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 sb_writers#4 mapping.invalidate_lock rcu_read_lock &c->lock irq_context: 0 sb_writers#4 &sb->s_type->i_mutex_key#8 jbd2_handle &journal->j_wait_updates irq_context: 0 sb_writers#4 &sb->s_type->i_mutex_key#8 &c->lock irq_context: 0 sb_writers#4 &sb->s_type->i_mutex_key#8 lock#4 &lruvec->lru_lock irq_context: 0 rtnl_mutex per_cpu_ptr(&cgroup_rstat_cpu_lock, cpu) irq_context: softirq (&ndev->rs_timer) rcu_read_lock rcu_read_lock rcu_read_lock_bh rcu_read_lock_bh &sch->q.lock irq_context: softirq (&ndev->rs_timer) rcu_read_lock rcu_read_lock rcu_read_lock_bh rcu_read_lock_bh &sch->q.lock tk_core.seq.seqcount irq_context: softirq (&ndev->rs_timer) rcu_read_lock rcu_read_lock rcu_read_lock_bh rcu_read_lock_bh &obj_hash[i].lock pool_lock irq_context: softirq (&p->forward_delay_timer) &br->lock rcu_read_lock &____s->seqcount#2 irq_context: softirq (&p->forward_delay_timer) &br->lock rcu_read_lock &____s->seqcount irq_context: softirq &(&net->ipv6.addr_chk_work)->timer irq_context: softirq &(&net->ipv6.addr_chk_work)->timer rcu_read_lock &pool->lock irq_context: softirq &(&net->ipv6.addr_chk_work)->timer rcu_read_lock &pool->lock &obj_hash[i].lock irq_context: softirq &(&net->ipv6.addr_chk_work)->timer rcu_read_lock &pool->lock &p->pi_lock irq_context: softirq &(&net->ipv6.addr_chk_work)->timer rcu_read_lock &pool->lock &p->pi_lock &rq->__lock irq_context: softirq &(&net->ipv6.addr_chk_work)->timer rcu_read_lock &pool->lock &p->pi_lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: softirq (&ndev->rs_timer) rcu_read_lock rcu_read_lock rcu_read_lock_bh &sch->q.lock tk_core.seq.seqcount irq_context: 0 (wq_completion)ipv6_addrconf (work_completion)(&(&net->ipv6.addr_chk_work)->work) rtnl_mutex rtnl_mutex.wait_lock irq_context: 0 (wq_completion)ipv6_addrconf (work_completion)(&(&net->ipv6.addr_chk_work)->work) rtnl_mutex &pool->lock irq_context: 0 (wq_completion)ipv6_addrconf (work_completion)(&(&net->ipv6.addr_chk_work)->work) rtnl_mutex &pool->lock &p->pi_lock irq_context: 0 (wq_completion)ipv6_addrconf (work_completion)(&(&net->ipv6.addr_chk_work)->work) rtnl_mutex &pool->lock &p->pi_lock &rq->__lock irq_context: 0 (wq_completion)ipv6_addrconf (work_completion)(&(&net->ipv6.addr_chk_work)->work) rtnl_mutex &pool->lock &p->pi_lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)ipv6_addrconf (work_completion)(&(&net->ipv6.addr_chk_work)->work) rtnl_mutex &rq->__lock irq_context: 0 (wq_completion)ipv6_addrconf (work_completion)(&(&net->ipv6.addr_chk_work)->work) rtnl_mutex &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 &mm->mmap_lock sb_pagefaults &journal->j_state_lock irq_context: 0 &mm->mmap_lock sb_pagefaults &journal->j_state_lock tk_core.seq.seqcount irq_context: 0 &mm->mmap_lock sb_pagefaults &journal->j_state_lock &obj_hash[i].lock irq_context: 0 &mm->mmap_lock sb_pagefaults &journal->j_state_lock &base->lock irq_context: 0 &mm->mmap_lock sb_pagefaults &journal->j_state_lock &base->lock &obj_hash[i].lock irq_context: 0 &mm->mmap_lock sb_pagefaults jbd2_handle &ret->b_state_lock &journal->j_list_lock irq_context: 0 sb_writers#4 &p->pi_lock &cfs_rq->removed.lock irq_context: 0 &sb->s_type->i_mutex_key#10 sk_lock-AF_CAIF rcu_state.exp_mutex &rnp->exp_wq[0] irq_context: 0 sb_writers#4 &sb->s_type->i_mutex_key#8 jbd2_handle &ei->i_data_sem key#14 irq_context: 0 sb_writers#4 &sb->s_type->i_mutex_key#8 jbd2_handle &ei->i_data_sem &c->lock irq_context: 0 sb_writers#4 &sb->s_type->i_mutex_key#8 jbd2_handle &ei->i_data_sem &journal->j_revoke_lock irq_context: 0 sb_writers#4 &sb->s_type->i_mutex_key#8 jbd2_handle &ei->i_es_lock key#2 irq_context: 0 sb_writers#4 &sb->s_type->i_mutex_key#8 jbd2_handle &ei->i_data_sem &____s->seqcount#2 irq_context: 0 sb_writers#4 &sb->s_type->i_mutex_key#8 jbd2_handle &ei->i_data_sem &____s->seqcount irq_context: 0 sb_writers#4 &sb->s_type->i_mutex_key#8 jbd2_handle &ei->i_data_sem stock_lock irq_context: 0 sb_writers#4 &sb->s_type->i_mutex_key#8 jbd2_handle &ei->i_data_sem &xa->xa_lock#9 irq_context: 0 sb_writers#4 &sb->s_type->i_mutex_key#8 jbd2_handle &ei->i_data_sem lock#4 irq_context: 0 sb_writers#4 &sb->s_type->i_mutex_key#8 jbd2_handle &ei->i_data_sem &n->list_lock irq_context: 0 sb_writers#4 &sb->s_type->i_mutex_key#8 jbd2_handle &ei->i_data_sem &n->list_lock &c->lock irq_context: 0 sb_writers#4 &sb->s_type->i_mutex_key#8 jbd2_handle &ei->i_data_sem rcu_read_lock &rq->__lock irq_context: 0 sb_writers#4 &sb->s_type->i_mutex_key#8 jbd2_handle &ei->i_data_sem rcu_read_lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 &mm->mmap_lock &mapping->i_mmap_rwsem rcu_read_lock &rcu_state.gp_wq irq_context: 0 &mm->mmap_lock &mapping->i_mmap_rwsem rcu_read_lock &rcu_state.gp_wq &p->pi_lock irq_context: 0 &mm->mmap_lock &mapping->i_mmap_rwsem rcu_read_lock &rcu_state.gp_wq &p->pi_lock &rq->__lock irq_context: 0 &mm->mmap_lock &mapping->i_mmap_rwsem rcu_read_lock &rcu_state.gp_wq &p->pi_lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 sk_lock-AF_INET6 &mm->mmap_lock rcu_read_lock rcu_read_lock ptlock_ptr(page)#2 irq_context: 0 sk_lock-AF_INET6 &mm->mmap_lock rcu_read_lock rcu_read_lock ptlock_ptr(page)#2 key irq_context: 0 sk_lock-AF_INET6 &f->f_owner.lock irq_context: 0 sk_lock-AF_INET6 rcu_read_lock &ei->socket.wq.wait irq_context: 0 sk_lock-AF_INET6 rcu_read_lock &ei->socket.wq.wait &p->pi_lock irq_context: 0 sk_lock-AF_INET6 rcu_read_lock &ei->socket.wq.wait &p->pi_lock &rq->__lock irq_context: 0 sk_lock-AF_INET6 rcu_read_lock &ei->socket.wq.wait &p->pi_lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 sk_lock-AF_INET6 rcu_read_lock rcu_read_lock rcu_read_lock &new->fa_lock irq_context: 0 sk_lock-AF_INET6 rcu_read_lock rcu_read_lock rcu_read_lock &new->fa_lock &f->f_owner.lock irq_context: 0 &xt[i].mutex &rq->__lock &cfs_rq->removed.lock irq_context: 0 proto_tab_lock irq_context: 0 proto_tab_lock pool_lock#2 irq_context: 0 proto_tab_lock &dir->lock irq_context: 0 proto_tab_lock &obj_hash[i].lock irq_context: 0 proto_tab_lock raw_sk_list.lock irq_context: softirq &(&bat_priv->nc.work)->timer rcu_read_lock &pool->lock/1 &p->pi_lock &rq->__lock &obj_hash[i].lock irq_context: softirq &(&bat_priv->nc.work)->timer rcu_read_lock &pool->lock/1 &p->pi_lock &rq->__lock &base->lock irq_context: softirq &(&bat_priv->nc.work)->timer rcu_read_lock &pool->lock/1 &p->pi_lock &rq->__lock &base->lock &obj_hash[i].lock irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem &xt[i].mutex &lock->wait_lock irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem &xt[i].mutex &rq->__lock irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem &xt[i].mutex &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 cb_lock &rdev->wiphy.mtx &local->mtx irq_context: 0 cb_lock &rdev->wiphy.mtx &local->mtx &data->mutex irq_context: 0 cb_lock &rdev->wiphy.mtx &local->mtx &local->queue_stop_reason_lock irq_context: 0 cb_lock &rdev->wiphy.mtx &local->mtx &local->iflist_mtx irq_context: 0 cb_lock &rdev->wiphy.mtx &local->mtx &local->iflist_mtx hrtimer_bases.lock irq_context: 0 cb_lock &rdev->wiphy.mtx &local->mtx &local->iflist_mtx hrtimer_bases.lock &obj_hash[i].lock irq_context: 0 cb_lock &rdev->wiphy.mtx &local->mtx &local->filter_lock irq_context: 0 cb_lock &rdev->wiphy.mtx &local->mtx rcu_read_lock &pool->lock/1 irq_context: 0 cb_lock &rdev->wiphy.mtx &local->mtx rcu_read_lock &pool->lock/1 &obj_hash[i].lock irq_context: 0 cb_lock &rdev->wiphy.mtx &local->mtx rcu_read_lock &pool->lock/1 &p->pi_lock irq_context: 0 cb_lock &rdev->wiphy.mtx &local->mtx rcu_read_lock &pool->lock/1 &p->pi_lock &rq->__lock irq_context: 0 cb_lock &rdev->wiphy.mtx &local->mtx rcu_read_lock &pool->lock/1 &p->pi_lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 cb_lock &rdev->wiphy.mtx batched_entropy_u8.lock irq_context: 0 cb_lock &rdev->wiphy.mtx kfence_freelist_lock irq_context: 0 (wq_completion)phy4 (work_completion)(&(&local->scan_work)->work) irq_context: 0 (wq_completion)phy4 (work_completion)(&(&local->scan_work)->work) &local->mtx irq_context: 0 (wq_completion)phy4 (work_completion)(&(&local->scan_work)->work) &local->mtx &local->iflist_mtx irq_context: 0 cb_lock &rdev->wiphy.mtx nl_table_lock irq_context: 0 (wq_completion)phy4 (work_completion)(&(&local->scan_work)->work) &local->mtx &data->mutex irq_context: 0 cb_lock &rdev->wiphy.mtx &meta->lock irq_context: 0 (wq_completion)phy4 (work_completion)(&(&local->scan_work)->work) &local->mtx &obj_hash[i].lock irq_context: 0 cb_lock &rdev->wiphy.mtx nl_table_wait.lock irq_context: 0 (wq_completion)phy4 (work_completion)(&(&local->scan_work)->work) &local->mtx &base->lock irq_context: 0 (wq_completion)phy4 (work_completion)(&(&local->scan_work)->work) &local->mtx &base->lock &obj_hash[i].lock irq_context: 0 sb_writers#4 &sb->s_type->i_mutex_key#8 jbd2_handle &ei->i_data_sem rcu_read_lock &rcu_state.gp_wq irq_context: 0 sb_writers#4 &sb->s_type->i_mutex_key#8 jbd2_handle &ei->i_data_sem rcu_read_lock &rcu_state.gp_wq &p->pi_lock irq_context: 0 sb_writers#4 &sb->s_type->i_mutex_key#8 jbd2_handle &ei->i_data_sem rcu_read_lock &rcu_state.gp_wq &p->pi_lock &rq->__lock irq_context: 0 sb_writers#4 &sb->s_type->i_mutex_key#8 jbd2_handle &ei->i_data_sem rcu_read_lock &rcu_state.gp_wq &p->pi_lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 cb_lock &rdev->wiphy.mtx &local->mtx &____s->seqcount irq_context: 0 cb_lock &rdev->wiphy.mtx &local->mtx pool_lock#2 irq_context: 0 cb_lock &rdev->wiphy.mtx &local->mtx fs_reclaim irq_context: 0 cb_lock &rdev->wiphy.mtx &local->mtx fs_reclaim mmu_notifier_invalidate_range_start irq_context: 0 cb_lock &rdev->wiphy.mtx &local->mtx &local->ack_status_lock irq_context: 0 cb_lock &rdev->wiphy.mtx &local->mtx &local->ack_status_lock pool_lock#2 irq_context: 0 cb_lock &rdev->wiphy.mtx &local->mtx rcu_read_lock &local->queue_stop_reason_lock irq_context: 0 cb_lock &rdev->wiphy.mtx &local->mtx rcu_read_lock tk_core.seq.seqcount irq_context: 0 cb_lock &rdev->wiphy.mtx &local->mtx rcu_read_lock hwsim_radio_lock irq_context: 0 cb_lock &rdev->wiphy.mtx &local->mtx rcu_read_lock hwsim_radio_lock pool_lock#2 irq_context: 0 cb_lock &rdev->wiphy.mtx &local->mtx rcu_read_lock hwsim_radio_lock &list->lock#19 irq_context: 0 cb_lock &rdev->wiphy.mtx &local->mtx rcu_read_lock &list->lock#19 irq_context: softirq rcu_read_lock lock#6 &kcov->lock irq_context: softirq rcu_read_lock &local->ack_status_lock irq_context: softirq rcu_read_lock &local->ack_status_lock &obj_hash[i].lock irq_context: softirq rcu_read_lock &local->ack_status_lock pool_lock#2 irq_context: softirq rcu_read_lock rcu_read_lock nl_table_lock irq_context: softirq rcu_read_lock rcu_read_lock nl_table_wait.lock irq_context: 0 cb_lock &rdev->wiphy.mtx &local->mtx rcu_read_lock rcu_node_0 irq_context: 0 cb_lock &rdev->wiphy.mtx &local->mtx rcu_read_lock &rq->__lock irq_context: 0 cb_lock &rdev->wiphy.mtx &local->mtx rcu_read_lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)events (work_completion)(&rdev->wiphy_work) &rdev->wiphy.mtx lock#6 &kcov->lock irq_context: 0 sb_writers#4 &sb->s_type->i_mutex_key#8 jbd2_handle &ei->i_data_sem rcu_read_lock &cfs_rq->removed.lock irq_context: 0 sb_writers#4 &sb->s_type->i_mutex_key#8 jbd2_handle &ei->i_data_sem rcu_read_lock &obj_hash[i].lock irq_context: 0 sb_writers#4 &sb->s_type->i_mutex_key#8 jbd2_handle &ei->i_data_sem rcu_read_lock pool_lock#2 irq_context: 0 &sb->s_type->i_mutex_key#10 sk_lock-AF_CAIF &rnp->exp_lock irq_context: 0 &sb->s_type->i_mutex_key#10 sk_lock-AF_CAIF rcu_state.exp_mutex irq_context: 0 &sb->s_type->i_mutex_key#10 sk_lock-AF_CAIF rcu_state.exp_mutex rcu_state.exp_mutex.wait_lock irq_context: 0 sb_writers#4 &type->i_mutex_dir_key#3/1 jbd2_handle &ei->i_data_sem &n->list_lock irq_context: 0 sb_writers#4 &type->i_mutex_dir_key#3/1 jbd2_handle &ei->i_data_sem &n->list_lock &c->lock irq_context: 0 cb_lock &rdev->wiphy.mtx &local->mtx rcu_read_lock &rcu_state.gp_wq irq_context: 0 cb_lock &rdev->wiphy.mtx &local->mtx rcu_read_lock &rcu_state.gp_wq &p->pi_lock irq_context: 0 cb_lock &rdev->wiphy.mtx &local->mtx rcu_read_lock &rcu_state.gp_wq &p->pi_lock &rq->__lock irq_context: 0 cb_lock &rdev->wiphy.mtx &local->mtx rcu_read_lock &rcu_state.gp_wq &p->pi_lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 cb_lock &rdev->wiphy.mtx &local->mtx rcu_read_lock &rcu_state.expedited_wq irq_context: 0 cb_lock &rdev->wiphy.mtx &local->mtx rcu_read_lock &rcu_state.expedited_wq &p->pi_lock irq_context: 0 cb_lock &rdev->wiphy.mtx &local->mtx rcu_read_lock &rcu_state.expedited_wq &p->pi_lock &rq->__lock irq_context: 0 cb_lock &rdev->wiphy.mtx &local->mtx rcu_read_lock &rcu_state.expedited_wq &p->pi_lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 cb_lock &rdev->wiphy.mtx rcu_read_lock pool_lock#2 irq_context: 0 cb_lock &rdev->wiphy.mtx rcu_read_lock rcu_read_lock_bh &obj_hash[i].lock irq_context: 0 cb_lock &rdev->wiphy.mtx rcu_read_lock rcu_read_lock_bh pool_lock#2 irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem &obj_hash[i].lock pool_lock irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem &____s->seqcount#2 irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem uevent_sock_mutex &c->lock irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem uevent_sock_mutex &n->list_lock irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem uevent_sock_mutex &n->list_lock &c->lock irq_context: 0 cb_lock &rdev->wiphy.mtx &local->mtx &rq->__lock irq_context: 0 cb_lock &rdev->wiphy.mtx &local->mtx &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 cb_lock &rdev->wiphy.mtx &local->mtx rcu_read_lock hwsim_radio_lock &c->lock irq_context: 0 cb_lock &rdev->wiphy.mtx &local->mtx &c->lock irq_context: 0 cb_lock &rdev->wiphy.mtx &local->mtx &n->list_lock irq_context: 0 cb_lock &rdev->wiphy.mtx &local->mtx &n->list_lock &c->lock irq_context: softirq rcu_read_lock &rdev->bss_lock irq_context: softirq rcu_read_lock &rdev->bss_lock &obj_hash[i].lock irq_context: softirq rcu_read_lock &rdev->bss_lock pool_lock#2 irq_context: softirq rcu_read_lock &rdev->bss_lock krc.lock irq_context: 0 &sb->s_type->i_mutex_key#10 sk_lock-AF_CAIF rcu_state.exp_mutex &rnp->exp_wq[3] irq_context: 0 (wq_completion)ext4-rsv-conversion (work_completion)(&ei->i_rsv_conversion_work) jbd2_handle &ei->i_data_sem remove_cache_srcu rcu_node_0 irq_context: 0 (wq_completion)ext4-rsv-conversion (work_completion)(&ei->i_rsv_conversion_work) jbd2_handle &ei->i_data_sem remove_cache_srcu &rq->__lock irq_context: 0 (wq_completion)ext4-rsv-conversion (work_completion)(&ei->i_rsv_conversion_work) jbd2_handle &ei->i_data_sem remove_cache_srcu &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 &sb->s_type->i_mutex_key#8 &sbi->s_writepages_rwsem jbd2_handle &n->list_lock irq_context: 0 &sb->s_type->i_mutex_key#8 &sbi->s_writepages_rwsem jbd2_handle &n->list_lock &c->lock irq_context: 0 cb_lock &rdev->wiphy.mtx rcu_read_lock &c->lock irq_context: 0 cb_lock &rdev->wiphy.mtx rcu_read_lock &n->list_lock irq_context: 0 cb_lock &rdev->wiphy.mtx rcu_read_lock &n->list_lock &c->lock irq_context: 0 sb_writers#4 &sb->s_type->i_mutex_key#8 &sb->s_type->i_lock_key#22 &xa->xa_lock#9 irq_context: 0 sb_writers#4 &sb->s_type->i_mutex_key#8 lock#5 irq_context: 0 sb_writers#4 &sb->s_type->i_mutex_key#8 &lruvec->lru_lock irq_context: softirq &(&local->scan_work)->timer irq_context: softirq &(&local->scan_work)->timer rcu_read_lock &pool->lock/1 irq_context: softirq &(&local->scan_work)->timer rcu_read_lock &pool->lock/1 &obj_hash[i].lock irq_context: softirq &(&local->scan_work)->timer rcu_read_lock &pool->lock/1 &p->pi_lock irq_context: 0 sb_writers#3 &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 cb_lock &rdev->wiphy.mtx &local->mtx rcu_read_lock hwsim_radio_lock &n->list_lock irq_context: 0 cb_lock &rdev->wiphy.mtx &local->mtx rcu_read_lock hwsim_radio_lock &n->list_lock &c->lock irq_context: 0 (wq_completion)ext4-rsv-conversion (work_completion)(&ei->i_rsv_conversion_work) mmu_notifier_invalidate_range_start irq_context: 0 (wq_completion)ext4-rsv-conversion (work_completion)(&ei->i_rsv_conversion_work) &journal->j_state_lock irq_context: 0 (wq_completion)ext4-rsv-conversion (work_completion)(&ei->i_rsv_conversion_work) &journal->j_state_lock tk_core.seq.seqcount irq_context: 0 (wq_completion)ext4-rsv-conversion (work_completion)(&ei->i_rsv_conversion_work) &journal->j_state_lock &obj_hash[i].lock irq_context: 0 (wq_completion)ext4-rsv-conversion (work_completion)(&ei->i_rsv_conversion_work) &journal->j_state_lock &base->lock irq_context: 0 (wq_completion)ext4-rsv-conversion (work_completion)(&ei->i_rsv_conversion_work) &journal->j_state_lock &base->lock &obj_hash[i].lock irq_context: 0 (wq_completion)ext4-rsv-conversion (work_completion)(&ei->i_rsv_conversion_work) jbd2_handle &ei->i_data_sem &ret->b_state_lock irq_context: 0 (wq_completion)ext4-rsv-conversion (work_completion)(&ei->i_rsv_conversion_work) jbd2_handle &ei->i_data_sem bit_wait_table + i irq_context: 0 sb_writers#4 &type->i_mutex_dir_key#3/1 sb_internal jbd2_handle &ret->b_state_lock irq_context: 0 sb_writers#4 sb_writers#4 jbd2_handle &mapping->private_lock irq_context: 0 sb_writers#4 sb_writers#4 jbd2_handle &ret->b_state_lock &journal->j_list_lock irq_context: 0 dup_mmap_sem &mm->mmap_lock &mm->mmap_lock/1 &anon_vma->rwsem rcu_read_lock rcu_read_lock rcu_node_0 irq_context: 0 dup_mmap_sem &mm->mmap_lock &mm->mmap_lock/1 &anon_vma->rwsem rcu_read_lock rcu_read_lock &rq->__lock irq_context: 0 dup_mmap_sem &mm->mmap_lock &mm->mmap_lock/1 &anon_vma->rwsem rcu_read_lock rcu_read_lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)ext4-rsv-conversion (work_completion)(&ei->i_rsv_conversion_work) jbd2_handle &ei->i_data_sem &ret->b_state_lock &journal->j_list_lock irq_context: 0 sb_writers#4 &sb->s_type->i_mutex_key#8 mmu_notifier_invalidate_range_start &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 &sb->s_type->i_mutex_key#8 &sbi->s_writepages_rwsem rcu_read_lock pool_lock#2 irq_context: 0 sb_writers#5 &type->i_mutex_dir_key#5/1 &sb->s_type->i_mutex_key#12 &simple_offset_xa_lock &obj_hash[i].lock pool_lock irq_context: 0 lock pidmap_lock &____s->seqcount#2 irq_context: 0 (wq_completion)netns net_cleanup_work rcu_state.barrier_mutex rcu_state.barrier_lock irq_context: 0 (wq_completion)netns net_cleanup_work rcu_state.barrier_mutex rcu_state.barrier_lock &obj_hash[i].lock irq_context: 0 (wq_completion)netns net_cleanup_work rcu_state.barrier_mutex &x->wait#24 irq_context: 0 &mm->mmap_lock sb_writers#4 jbd2_handle rcu_read_lock &rq->__lock irq_context: 0 &mm->mmap_lock sb_writers#4 jbd2_handle rcu_read_lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 rtnl_mutex &dev->tx_global_lock &vlan_netdev_xmit_lock_key irq_context: 0 rtnl_mutex &block->lock &c->lock irq_context: 0 &sig->cred_guard_mutex &ei->xattr_sem &rq->__lock irq_context: 0 &sig->cred_guard_mutex &ei->xattr_sem &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 sb_writers#4 mapping.invalidate_lock &c->lock irq_context: 0 sb_writers#4 sb_writers#4 batched_entropy_u8.lock irq_context: 0 sb_writers#4 sb_writers#4 kfence_freelist_lock irq_context: 0 sb_writers#4 sb_writers#4 &meta->lock irq_context: 0 sb_writers#4 sb_writers#4 &rq->__lock irq_context: 0 sb_writers#4 sb_writers#4 &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: softirq rcu_read_lock &sta->rate_ctrl_lock irq_context: softirq rcu_read_lock &sta->rate_ctrl_lock pool_lock#2 irq_context: softirq rcu_read_lock &sta->rate_ctrl_lock &obj_hash[i].lock irq_context: softirq rcu_read_lock &sta->rate_ctrl_lock krc.lock irq_context: 0 sb_writers#4 mapping.invalidate_lock &____s->seqcount#2 irq_context: 0 sb_writers#4 sb_internal jbd2_handle &sbi->s_orphan_lock &rq->__lock irq_context: 0 sb_writers#4 sb_internal jbd2_handle &sbi->s_orphan_lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 sb_writers#4 &type->i_mutex_dir_key#3/1 sb_internal jbd2_handle &ret->b_state_lock &journal->j_list_lock irq_context: 0 &sb->s_type->i_mutex_key#8 &sbi->s_writepages_rwsem jbd2_handle &ei->i_data_sem bit_wait_table + i irq_context: 0 sb_writers#4 mapping.invalidate_lock &folio_wait_table[i] irq_context: 0 sb_writers#4 mapping.invalidate_lock &ei->i_es_lock key#2 irq_context: 0 sb_writers#4 sb_writers#4 &c->lock irq_context: 0 sb_writers#4 sb_writers#4 &journal->j_state_lock &journal->j_wait_transaction_locked irq_context: 0 sb_writers#4 sb_writers#4 jbd2_handle irq_context: 0 &sb->s_type->i_mutex_key#8 &sbi->s_writepages_rwsem &journal->j_state_lock &journal->j_wait_transaction_locked irq_context: 0 &sb->s_type->i_mutex_key#8 &sbi->s_writepages_rwsem jbd2_handle irq_context: 0 &sb->s_type->i_mutex_key#8 &sbi->s_writepages_rwsem jbd2_handle &journal->j_wait_updates &p->pi_lock irq_context: 0 &sb->s_type->i_mutex_key#8 &sbi->s_writepages_rwsem jbd2_handle &journal->j_wait_updates &p->pi_lock &rq->__lock irq_context: 0 &sb->s_type->i_mutex_key#8 &sbi->s_writepages_rwsem jbd2_handle &journal->j_wait_updates &p->pi_lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 &sb->s_type->i_mutex_key#8 &sbi->s_writepages_rwsem jbd2_handle &ei->i_data_sem rcu_read_lock rcu_node_0 irq_context: 0 sb_writers#4 mapping.invalidate_lock &xa->xa_lock#9 &c->lock irq_context: 0 (wq_completion)events (work_completion)(&rdev->wiphy_work) &rdev->wiphy.mtx &wdev->mtx &rdev->wiphy_work_lock irq_context: 0 (wq_completion)events (work_completion)(&rdev->wiphy_work) &rdev->wiphy.mtx &rdev->wiphy_work_lock rcu_read_lock &pool->lock irq_context: 0 (wq_completion)events (work_completion)(&rdev->wiphy_work) &rdev->wiphy.mtx &rdev->wiphy_work_lock rcu_read_lock &pool->lock &obj_hash[i].lock irq_context: 0 (wq_completion)events (work_completion)(&rdev->wiphy_work) &rdev->wiphy.mtx &wdev->mtx &local->sta_mtx &sta->ampdu_mlme.mtx irq_context: 0 (wq_completion)events (work_completion)(&rdev->wiphy_work) &rdev->wiphy.mtx &wdev->mtx &local->sta_mtx &sta->ampdu_mlme.mtx &rq->__lock irq_context: 0 (wq_completion)events (work_completion)(&rdev->wiphy_work) &rdev->wiphy.mtx &wdev->mtx &local->sta_mtx &sta->ampdu_mlme.mtx &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 sb_writers#4 &sb->s_type->i_mutex_key#8 rcu_read_lock &c->lock irq_context: 0 sb_writers#4 &sb->s_type->i_mutex_key#8 rcu_read_lock &____s->seqcount#2 irq_context: 0 (wq_completion)events (work_completion)(&rdev->wiphy_work) &rdev->wiphy.mtx &wdev->mtx &local->sta_mtx &sta->ampdu_mlme.mtx &sta->lock irq_context: 0 sb_writers#4 &sb->s_type->i_mutex_key#8 rcu_read_lock &____s->seqcount irq_context: 0 (wq_completion)events (work_completion)(&rdev->wiphy_work) &rdev->wiphy.mtx &wdev->mtx &local->sta_mtx (work_completion)(&sta->ampdu_mlme.work) irq_context: 0 (wq_completion)events (work_completion)(&rdev->wiphy_work) &rdev->wiphy.mtx &wdev->mtx &local->sta_mtx rcu_state.exp_mutex rcu_node_0 irq_context: 0 (wq_completion)events (work_completion)(&rdev->wiphy_work) &rdev->wiphy.mtx &wdev->mtx &local->sta_mtx rcu_state.exp_mutex &obj_hash[i].lock irq_context: 0 (wq_completion)events (work_completion)(&rdev->wiphy_work) &rdev->wiphy.mtx &wdev->mtx &local->sta_mtx rcu_state.exp_mutex rcu_read_lock &pool->lock irq_context: 0 (wq_completion)events (work_completion)(&rdev->wiphy_work) &rdev->wiphy.mtx &wdev->mtx &local->sta_mtx rcu_state.exp_mutex rcu_read_lock &pool->lock &obj_hash[i].lock irq_context: 0 (wq_completion)events (work_completion)(&rdev->wiphy_work) &rdev->wiphy.mtx &wdev->mtx &local->sta_mtx rcu_state.exp_mutex &rq->__lock irq_context: 0 (wq_completion)events (work_completion)(&rdev->wiphy_work) &rdev->wiphy.mtx &wdev->mtx &local->sta_mtx rcu_state.exp_mutex &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)events (work_completion)(&rdev->wiphy_work) &rdev->wiphy.mtx &wdev->mtx &local->sta_mtx rcu_state.exp_mutex &rnp->exp_wq[1] irq_context: 0 (wq_completion)events (work_completion)(&rdev->wiphy_work) &rdev->wiphy.mtx &wdev->mtx &local->sta_mtx rcu_state.exp_mutex &pool->lock irq_context: 0 (wq_completion)events (work_completion)(&rdev->wiphy_work) &rdev->wiphy.mtx &wdev->mtx &local->sta_mtx rcu_state.exp_mutex &pool->lock &p->pi_lock irq_context: 0 (wq_completion)events (work_completion)(&rdev->wiphy_work) &rdev->wiphy.mtx &wdev->mtx &local->sta_mtx rcu_state.exp_mutex &pool->lock &p->pi_lock &rq->__lock irq_context: 0 (wq_completion)events (work_completion)(&rdev->wiphy_work) &rdev->wiphy.mtx &wdev->mtx &local->sta_mtx rcu_state.exp_mutex &pool->lock &p->pi_lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 &sb->s_type->i_mutex_key#8 &sbi->s_writepages_rwsem jbd2_handle rcu_node_0 irq_context: 0 (wq_completion)events (work_completion)(&rdev->wiphy_work) &rdev->wiphy.mtx &wdev->mtx &local->sta_mtx &sta->lock irq_context: 0 (wq_completion)events (work_completion)(&rdev->wiphy_work) &rdev->wiphy.mtx &wdev->mtx &local->sta_mtx krc.lock irq_context: 0 (wq_completion)events (work_completion)(&rdev->wiphy_work) &rdev->wiphy.mtx &wdev->mtx &local->sta_mtx &local->key_mtx irq_context: 0 (wq_completion)events (work_completion)(&rdev->wiphy_work) &rdev->wiphy.mtx &wdev->mtx &local->sta_mtx &fq->lock irq_context: 0 (wq_completion)events (work_completion)(&rdev->wiphy_work) &rdev->wiphy.mtx &wdev->mtx &local->sta_mtx &dentry->d_lock irq_context: 0 (wq_completion)events (work_completion)(&rdev->wiphy_work) &rdev->wiphy.mtx &wdev->mtx &local->sta_mtx &sb->s_type->i_mutex_key#3 &dentry->d_lock &dentry->d_lock/1 irq_context: 0 (wq_completion)events (work_completion)(&rdev->wiphy_work) &rdev->wiphy.mtx &wdev->mtx &local->sta_mtx &sb->s_type->i_mutex_key#3 pin_fs_lock irq_context: 0 (wq_completion)events (work_completion)(&rdev->wiphy_work) &rdev->wiphy.mtx &wdev->mtx &local->sta_mtx &sb->s_type->i_mutex_key#3 rcu_read_lock mount_lock irq_context: 0 (wq_completion)events (work_completion)(&rdev->wiphy_work) &rdev->wiphy.mtx &wdev->mtx &local->sta_mtx &sb->s_type->i_mutex_key#3 rcu_read_lock mount_lock mount_lock.seqcount irq_context: 0 (wq_completion)events (work_completion)(&rdev->wiphy_work) &rdev->wiphy.mtx &wdev->mtx &local->sta_mtx &sb->s_type->i_mutex_key#3 mount_lock irq_context: 0 (wq_completion)events (work_completion)(&rdev->wiphy_work) &rdev->wiphy.mtx &wdev->mtx &local->sta_mtx &sb->s_type->i_mutex_key#3 mount_lock mount_lock.seqcount irq_context: 0 (wq_completion)events (work_completion)(&rdev->wiphy_work) &rdev->wiphy.mtx &wdev->mtx &local->sta_mtx &sb->s_type->i_mutex_key#3 rcu_read_lock &dentry->d_lock irq_context: 0 (wq_completion)events (work_completion)(&rdev->wiphy_work) &rdev->wiphy.mtx &wdev->mtx &local->sta_mtx &sb->s_type->i_mutex_key#3 &fsnotify_mark_srcu irq_context: 0 (wq_completion)events (work_completion)(&rdev->wiphy_work) &rdev->wiphy.mtx &wdev->mtx &local->sta_mtx &sb->s_type->i_mutex_key#3 &xa->xa_lock#9 irq_context: 0 (wq_completion)events (work_completion)(&rdev->wiphy_work) &rdev->wiphy.mtx &wdev->mtx &local->sta_mtx &sb->s_type->i_mutex_key#3 &rq->__lock irq_context: 0 (wq_completion)events (work_completion)(&rdev->wiphy_work) &rdev->wiphy.mtx &wdev->mtx &local->sta_mtx &sb->s_type->i_mutex_key#3 &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)events (work_completion)(&rdev->wiphy_work) &rdev->wiphy.mtx &wdev->mtx &local->sta_mtx &sb->s_type->i_mutex_key#3 &obj_hash[i].lock pool_lock irq_context: 0 (wq_completion)events (work_completion)(&rdev->wiphy_work) &rdev->wiphy.mtx &wdev->mtx &local->sta_mtx rcu_read_lock &dentry->d_lock irq_context: 0 (wq_completion)events (work_completion)(&rdev->wiphy_work) &rdev->wiphy.mtx &wdev->mtx &local->sta_mtx &rq->__lock irq_context: 0 (wq_completion)events (work_completion)(&rdev->wiphy_work) &rdev->wiphy.mtx &wdev->mtx &local->sta_mtx &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)events (work_completion)(&rdev->wiphy_work) &rdev->wiphy.mtx &wdev->mtx &local->sta_mtx &cfs_rq->removed.lock irq_context: 0 (wq_completion)events (work_completion)(&rdev->wiphy_work) &rdev->wiphy.mtx &wdev->mtx &local->sta_mtx &fsnotify_mark_srcu irq_context: 0 (wq_completion)events (work_completion)(&rdev->wiphy_work) &rdev->wiphy.mtx &wdev->mtx &local->sta_mtx &sb->s_type->i_lock_key#7 irq_context: 0 (wq_completion)events (work_completion)(&rdev->wiphy_work) &rdev->wiphy.mtx &wdev->mtx &local->sta_mtx &s->s_inode_list_lock irq_context: 0 (wq_completion)events (work_completion)(&rdev->wiphy_work) &rdev->wiphy.mtx &wdev->mtx &local->sta_mtx &xa->xa_lock#9 irq_context: 0 (wq_completion)events (work_completion)(&rdev->wiphy_work) &rdev->wiphy.mtx &wdev->mtx &local->sta_mtx &obj_hash[i].lock pool_lock irq_context: 0 (wq_completion)events (work_completion)(&rdev->wiphy_work) &rdev->wiphy.mtx &wdev->mtx &local->sta_mtx rcu_read_lock mount_lock irq_context: 0 (wq_completion)events (work_completion)(&rdev->wiphy_work) &rdev->wiphy.mtx &wdev->mtx &local->sta_mtx rcu_read_lock mount_lock mount_lock.seqcount irq_context: 0 (wq_completion)events (work_completion)(&rdev->wiphy_work) &rdev->wiphy.mtx &wdev->mtx &local->sta_mtx mount_lock irq_context: 0 (wq_completion)events (work_completion)(&rdev->wiphy_work) &rdev->wiphy.mtx &wdev->mtx &local->sta_mtx mount_lock mount_lock.seqcount irq_context: 0 (wq_completion)events (work_completion)(&rdev->wiphy_work) &rdev->wiphy.mtx &wdev->mtx &local->sta_mtx &local->active_txq_lock[i] irq_context: 0 (wq_completion)events (work_completion)(&rdev->wiphy_work) &rdev->wiphy.mtx &wdev->mtx &local->sta_mtx (work_completion)(&sta->drv_deliver_wk) irq_context: 0 (wq_completion)events (work_completion)(&rdev->wiphy_work) &rdev->wiphy.mtx &wdev->mtx &local->sta_mtx (work_completion)(&sta->drv_deliver_wk) &rq->__lock irq_context: 0 (wq_completion)events (work_completion)(&rdev->wiphy_work) &rdev->wiphy.mtx &wdev->mtx &local->sta_mtx (work_completion)(&sta->drv_deliver_wk) &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 cb_lock &rdev->wiphy.mtx &local->mtx &local->ack_status_lock &c->lock irq_context: 0 (wq_completion)events (work_completion)(&rdev->wiphy_work) &rdev->wiphy.mtx &wdev->mtx &local->sta_mtx rcu_read_lock pool_lock#2 irq_context: 0 (wq_completion)events (work_completion)(&rdev->wiphy_work) &rdev->wiphy.mtx &wdev->mtx lweventlist_lock irq_context: 0 (wq_completion)events (work_completion)(&rdev->wiphy_work) &rdev->wiphy.mtx &wdev->mtx lweventlist_lock &c->lock irq_context: 0 (wq_completion)events (work_completion)(&rdev->wiphy_work) &rdev->wiphy.mtx &wdev->mtx lweventlist_lock &n->list_lock irq_context: 0 (wq_completion)events (work_completion)(&rdev->wiphy_work) &rdev->wiphy.mtx &wdev->mtx lweventlist_lock &n->list_lock &c->lock irq_context: 0 (wq_completion)events (work_completion)(&rdev->wiphy_work) &rdev->wiphy.mtx &wdev->mtx lweventlist_lock pool_lock#2 irq_context: 0 (wq_completion)events (work_completion)(&rdev->wiphy_work) &rdev->wiphy.mtx &wdev->mtx lweventlist_lock &dir->lock#2 irq_context: 0 (wq_completion)events (work_completion)(&rdev->wiphy_work) &rdev->wiphy.mtx &wdev->mtx krc.lock irq_context: 0 (wq_completion)events (work_completion)(&rdev->wiphy_work) &rdev->wiphy.mtx &wdev->mtx &local->mtx &local->chanctx_mtx &local->queue_stop_reason_lock irq_context: 0 (wq_completion)events (work_completion)(&rdev->wiphy_work) &rdev->wiphy.mtx &wdev->mtx &local->mtx &local->chanctx_mtx rcu_read_lock &fq->lock irq_context: 0 (wq_completion)events (work_completion)(&rdev->wiphy_work) &rdev->wiphy.mtx &wdev->mtx &local->mtx &local->chanctx_mtx rcu_read_lock &local->queue_stop_reason_lock irq_context: 0 (wq_completion)events (work_completion)(&rdev->wiphy_work) &rdev->wiphy.mtx &wdev->mtx &local->mtx &local->chanctx_mtx &obj_hash[i].lock irq_context: 0 (wq_completion)events (work_completion)(&rdev->wiphy_work) &rdev->wiphy.mtx &wdev->mtx &local->mtx &local->chanctx_mtx krc.lock irq_context: 0 (wq_completion)events (work_completion)(&rdev->wiphy_work) &rdev->wiphy.mtx &wdev->mtx rcu_state.exp_mutex rcu_node_0 irq_context: 0 (wq_completion)events (work_completion)(&rdev->wiphy_work) &rdev->wiphy.mtx &wdev->mtx rcu_state.exp_mutex &obj_hash[i].lock irq_context: 0 (wq_completion)events (work_completion)(&rdev->wiphy_work) &rdev->wiphy.mtx &wdev->mtx rcu_state.exp_mutex rcu_read_lock &pool->lock irq_context: 0 (wq_completion)events (work_completion)(&rdev->wiphy_work) &rdev->wiphy.mtx &wdev->mtx rcu_state.exp_mutex rcu_read_lock &pool->lock &obj_hash[i].lock irq_context: 0 (wq_completion)events (work_completion)(&rdev->wiphy_work) &rdev->wiphy.mtx &wdev->mtx rcu_state.exp_mutex &rnp->exp_wq[2] irq_context: 0 (wq_completion)events (work_completion)(&rdev->wiphy_work) &rdev->wiphy.mtx &wdev->mtx rcu_state.exp_mutex &pool->lock irq_context: 0 (wq_completion)events (work_completion)(&rdev->wiphy_work) &rdev->wiphy.mtx &wdev->mtx rcu_state.exp_mutex &pool->lock &p->pi_lock irq_context: 0 (wq_completion)events (work_completion)(&rdev->wiphy_work) &rdev->wiphy.mtx &wdev->mtx rcu_state.exp_mutex &pool->lock &p->pi_lock &rq->__lock irq_context: 0 (wq_completion)events (work_completion)(&rdev->wiphy_work) &rdev->wiphy.mtx &wdev->mtx rcu_state.exp_mutex &pool->lock &p->pi_lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)events (work_completion)(&rdev->wiphy_work) &rdev->wiphy.mtx &wdev->mtx rcu_state.exp_mutex &rq->__lock irq_context: 0 (wq_completion)events (work_completion)(&rdev->wiphy_work) &rdev->wiphy.mtx &wdev->mtx rcu_state.exp_mutex &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)events (linkwatch_work).work rtnl_mutex &rnp->exp_lock irq_context: 0 (wq_completion)events (linkwatch_work).work rtnl_mutex rcu_state.exp_mutex irq_context: 0 (wq_completion)events (linkwatch_work).work rtnl_mutex rcu_state.exp_mutex rcu_state.exp_mutex.wait_lock irq_context: 0 &mm->mmap_lock remove_cache_srcu rcu_node_0 irq_context: 0 &mm->mmap_lock remove_cache_srcu &rcu_state.expedited_wq irq_context: 0 &mm->mmap_lock remove_cache_srcu &rcu_state.expedited_wq &p->pi_lock irq_context: 0 &mm->mmap_lock remove_cache_srcu &rcu_state.expedited_wq &p->pi_lock &rq->__lock irq_context: 0 &mm->mmap_lock remove_cache_srcu &rcu_state.expedited_wq &p->pi_lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)events (work_completion)(&rdev->wiphy_work) &rdev->wiphy.mtx &wdev->mtx rcu_state.exp_mutex.wait_lock irq_context: 0 (wq_completion)events (work_completion)(&rdev->wiphy_work) &rdev->wiphy.mtx &wdev->mtx &p->pi_lock irq_context: 0 (wq_completion)events (work_completion)(&rdev->wiphy_work) &rdev->wiphy.mtx &wdev->mtx &p->pi_lock &rq->__lock irq_context: 0 (wq_completion)events (work_completion)(&rdev->wiphy_work) &rdev->wiphy.mtx &wdev->mtx &p->pi_lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 sb_writers#4 sb_internal jbd2_handle &sbi->s_orphan_lock &ret->b_state_lock irq_context: 0 sb_writers#4 sb_internal jbd2_handle &sbi->s_orphan_lock &ret->b_state_lock &journal->j_list_lock irq_context: 0 sb_writers#4 &sb->s_type->i_mutex_key#8 jbd2_handle &sbi->s_orphan_lock &lock->wait_lock irq_context: 0 sb_writers#4 &sb->s_type->i_mutex_key#8 jbd2_handle &sbi->s_orphan_lock &rq->__lock irq_context: 0 sb_writers#4 &sb->s_type->i_mutex_key#8 jbd2_handle &sbi->s_orphan_lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 sb_writers#4 mapping.invalidate_lock rcu_read_lock rcu_read_lock rcu_read_lock rcu_node_0 irq_context: 0 sb_writers#4 mapping.invalidate_lock rcu_read_lock rcu_read_lock rcu_read_lock &rq->__lock irq_context: 0 sb_writers#4 mapping.invalidate_lock rcu_read_lock rcu_read_lock rcu_read_lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 &sb->s_type->i_mutex_key#8 &sbi->s_writepages_rwsem jbd2_handle rcu_read_lock &rcu_state.gp_wq irq_context: 0 &sb->s_type->i_mutex_key#8 &sbi->s_writepages_rwsem jbd2_handle rcu_read_lock &rcu_state.gp_wq &p->pi_lock irq_context: 0 &sb->s_type->i_mutex_key#8 &sbi->s_writepages_rwsem jbd2_handle rcu_read_lock &rcu_state.gp_wq &p->pi_lock &rq->__lock irq_context: 0 &sb->s_type->i_mutex_key#8 &sbi->s_writepages_rwsem jbd2_handle rcu_read_lock &rcu_state.gp_wq &p->pi_lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: softirq &(&bat_priv->bla.work)->timer rcu_read_lock &pool->lock/1 &p->pi_lock &rq->__lock irq_context: softirq &(&bat_priv->bla.work)->timer rcu_read_lock &pool->lock/1 &p->pi_lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: softirq rcu_read_lock &rdev->bss_lock &c->lock irq_context: 0 (wq_completion)events (work_completion)(&rdev->wiphy_work) &rdev->wiphy.mtx &wdev->mtx &list->lock#18 irq_context: 0 &sb->s_type->i_mutex_key#10 &rnp->exp_lock irq_context: 0 &sb->s_type->i_mutex_key#10 rcu_state.exp_mutex irq_context: 0 &sb->s_type->i_mutex_key#10 rcu_state.exp_mutex rcu_state.exp_mutex.wait_lock irq_context: 0 sb_writers#4 sb_internal jbd2_handle &lock->wait_lock irq_context: 0 sb_writers#4 sb_internal jbd2_handle &p->pi_lock irq_context: 0 sb_writers#4 sb_internal jbd2_handle &p->pi_lock &rq->__lock irq_context: 0 sb_writers#4 sb_internal jbd2_handle &p->pi_lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)events (linkwatch_work).work rtnl_mutex rcu_state.exp_mutex.wait_lock irq_context: 0 (wq_completion)events (linkwatch_work).work rtnl_mutex &p->pi_lock irq_context: 0 (wq_completion)events (linkwatch_work).work rtnl_mutex &p->pi_lock &rq->__lock irq_context: 0 sb_writers#4 &type->i_mutex_dir_key#3/1 &sb->s_type->i_mutex_key#8 jbd2_handle &lock->wait_lock irq_context: 0 (wq_completion)events (linkwatch_work).work rtnl_mutex &p->pi_lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 sb_writers#4 &type->i_mutex_dir_key#3/1 &sb->s_type->i_mutex_key#8 jbd2_handle &p->pi_lock irq_context: 0 sb_writers#4 &type->i_mutex_dir_key#3/1 &type->i_mutex_dir_key#3 jbd2_handle &lock->wait_lock irq_context: 0 (wq_completion)events (linkwatch_work).work rtnl_mutex &tbl->lock &n->lock irq_context: 0 sb_writers#4 &type->i_mutex_dir_key#3/1 &type->i_mutex_dir_key#3 jbd2_handle &p->pi_lock irq_context: 0 (wq_completion)events (linkwatch_work).work rtnl_mutex &tbl->lock pool_lock#2 irq_context: 0 (wq_completion)events (linkwatch_work).work rtnl_mutex &tbl->lock &n->lock irq_context: 0 (wq_completion)events (linkwatch_work).work rtnl_mutex &tbl->lock &n->lock &____s->seqcount#9 irq_context: 0 (wq_completion)events (linkwatch_work).work rtnl_mutex &tbl->lock nl_table_lock irq_context: 0 (wq_completion)events (linkwatch_work).work rtnl_mutex &tbl->lock &obj_hash[i].lock irq_context: 0 (wq_completion)events (linkwatch_work).work rtnl_mutex &tbl->lock nl_table_wait.lock irq_context: 0 (wq_completion)events (linkwatch_work).work rtnl_mutex &tbl->lock rcu_read_lock lock#8 irq_context: 0 (wq_completion)events (linkwatch_work).work rtnl_mutex &tbl->lock rcu_read_lock id_table_lock irq_context: 0 (wq_completion)events (linkwatch_work).work rtnl_mutex &tbl->lock &dir->lock#2 irq_context: 0 (wq_completion)events (linkwatch_work).work rtnl_mutex &tbl->lock krc.lock irq_context: 0 sb_writers#4 &type->i_mutex_dir_key#3/1 sb_internal jbd2_handle &sbi->s_orphan_lock &ei->i_raw_lock irq_context: 0 sb_writers#4 &sb->s_type->i_mutex_key#8 &cfs_rq->removed.lock irq_context: 0 &ei->i_data_sem &mapping->private_lock irq_context: 0 (wq_completion)bat_events (work_completion)(&(&forw_packet_aggr->delayed_work)->work) &hard_iface->bat_iv.ogm_buff_mutex rcu_read_lock &rcu_state.expedited_wq &p->pi_lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 sk_lock-AF_INET6 &f->f_lock fasync_lock &new->fa_lock irq_context: 0 sk_lock-AF_INET6 &f->f_lock fasync_lock &obj_hash[i].lock irq_context: 0 sk_lock-AF_INET6 &f->f_lock fasync_lock pool_lock#2 irq_context: 0 &sb->s_type->i_mutex_key#10 sk_lock-AF_INET6 rcu_read_lock pool_lock#2 irq_context: 0 &sb->s_type->i_mutex_key#10 sk_lock-AF_INET6 &obj_hash[i].lock irq_context: 0 &sb->s_type->i_mutex_key#10 sk_lock-AF_INET6 pool_lock#2 irq_context: 0 &sb->s_type->i_mutex_key#10 sk_lock-AF_INET6 slock-AF_INET6 &tcp_hashinfo.bhash[i].lock irq_context: 0 &sb->s_type->i_mutex_key#10 sk_lock-AF_INET6 slock-AF_INET6 &tcp_hashinfo.bhash[i].lock &tcp_hashinfo.bhash2[i].lock irq_context: 0 &sb->s_type->i_mutex_key#10 raw_sk_list.lock irq_context: 0 &sb->s_type->i_mutex_key#10 clock-AF_NFC irq_context: 0 &sb->s_type->i_mutex_key#10 rlock-AF_NFC irq_context: 0 (wq_completion)cfg80211 (work_completion)(&rdev->event_work) &rdev->wiphy.mtx &wdev->mtx &rdev->bss_lock &obj_hash[i].lock irq_context: 0 (wq_completion)cfg80211 (work_completion)(&rdev->event_work) &rdev->wiphy.mtx &wdev->mtx &rdev->bss_lock pool_lock#2 irq_context: 0 (wq_completion)cfg80211 (work_completion)(&rdev->event_work) &rdev->wiphy.mtx &wdev->mtx &rdev->bss_lock krc.lock irq_context: 0 sb_writers#4 sb_internal jbd2_handle &ei->i_data_sem &ret->b_state_lock irq_context: 0 sb_writers#4 sb_internal jbd2_handle &ei->i_data_sem &ret->b_state_lock &journal->j_list_lock irq_context: 0 sb_writers#4 sb_internal jbd2_handle &ei->i_data_sem &bgl->locks[i].lock &obj_hash[i].lock irq_context: 0 sb_writers#4 sb_internal jbd2_handle &ei->i_data_sem &bgl->locks[i].lock pool_lock#2 irq_context: 0 sb_writers#4 sb_internal jbd2_handle &ei->i_data_sem &journal->j_revoke_lock irq_context: 0 &iint->mutex ima_extend_list_mutex &rq->__lock irq_context: 0 &iint->mutex ima_extend_list_mutex &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)events (work_completion)(&rdev->wiphy_work) &rdev->wiphy.mtx &wdev->mtx remove_cache_srcu irq_context: 0 (wq_completion)events (work_completion)(&rdev->wiphy_work) &rdev->wiphy.mtx &wdev->mtx remove_cache_srcu quarantine_lock irq_context: 0 (wq_completion)events (work_completion)(&rdev->wiphy_work) &rdev->wiphy.mtx &wdev->mtx remove_cache_srcu &c->lock irq_context: 0 (wq_completion)events (work_completion)(&rdev->wiphy_work) &rdev->wiphy.mtx &wdev->mtx remove_cache_srcu &n->list_lock irq_context: 0 (wq_completion)events (work_completion)(&rdev->wiphy_work) &rdev->wiphy.mtx &wdev->mtx remove_cache_srcu rcu_read_lock pool_lock#2 irq_context: 0 (wq_completion)events (work_completion)(&rdev->wiphy_work) &rdev->wiphy.mtx &wdev->mtx remove_cache_srcu &obj_hash[i].lock irq_context: 0 (wq_completion)events (work_completion)(&rdev->wiphy_work) &rdev->wiphy.mtx &wdev->mtx &local->sta_mtx &sb->s_type->i_mutex_key#3 &n->list_lock irq_context: 0 (wq_completion)events (work_completion)(&rdev->wiphy_work) &rdev->wiphy.mtx &wdev->mtx &local->sta_mtx &sb->s_type->i_mutex_key#3 &n->list_lock &c->lock irq_context: 0 &disk->open_mutex &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 &mm->mmap_lock mmu_notifier_invalidate_range_start &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 &xt[i].mutex fs_reclaim mmu_notifier_invalidate_range_start &rq->__lock irq_context: 0 &xt[i].mutex fs_reclaim mmu_notifier_invalidate_range_start &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 &f->f_pos_lock sb_writers#4 &sb->s_type->i_mutex_key#8 &mm->mmap_lock rcu_read_lock rcu_read_lock ptlock_ptr(page)#2 key irq_context: 0 rtnl_mutex &dev->tx_global_lock &qdisc_xmit_lock_key irq_context: 0 &f->f_pos_lock rcu_read_lock rcu_node_0 irq_context: 0 &f->f_pos_lock rcu_read_lock &rq->__lock irq_context: 0 &f->f_pos_lock rcu_read_lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 &sb->s_type->i_mutex_key#10 (netlink_chain).rwsem &n->list_lock irq_context: 0 &sb->s_type->i_mutex_key#10 (netlink_chain).rwsem &n->list_lock &c->lock irq_context: 0 &f->f_pos_lock sb_writers#4 &sb->s_type->i_mutex_key#8 jbd2_handle &ei->i_data_sem key#14 irq_context: 0 sb_writers#4 mapping.invalidate_lock rcu_read_lock &____s->seqcount#2 irq_context: 0 sb_writers#4 mapping.invalidate_lock rcu_read_lock &____s->seqcount irq_context: 0 &mm->mmap_lock sb_pagefaults &c->lock irq_context: 0 &mm->mmap_lock sb_pagefaults &____s->seqcount#2 irq_context: 0 &mm->mmap_lock sb_pagefaults &____s->seqcount irq_context: 0 sk_lock-AF_INET6 &tcp_hashinfo.bhash[i].lock &tcp_hashinfo.bhash2[i].lock batched_entropy_u8.lock irq_context: 0 sk_lock-AF_INET6 &tcp_hashinfo.bhash[i].lock &tcp_hashinfo.bhash2[i].lock &hashinfo->ehash_locks[i] irq_context: 0 sk_lock-AF_INET6 rcu_read_lock fastopen_seqlock.seqcount irq_context: 0 sk_lock-AF_INET6 rcu_read_lock rcu_read_lock &nf_conntrack_locks[i] &nf_conntrack_locks[i]/1 irq_context: 0 sk_lock-AF_INET6 rcu_read_lock rcu_read_lock &nf_conntrack_locks[i] &nf_conntrack_locks[i]/1 batched_entropy_u8.lock irq_context: 0 sk_lock-AF_INET6 rcu_read_lock rcu_read_lock &nf_conntrack_locks[i]/1 irq_context: softirq rcu_read_lock rcu_read_lock rcu_read_lock rcu_read_lock rcu_read_lock_bh tk_core.seq.seqcount irq_context: softirq rcu_read_lock rcu_read_lock rcu_read_lock rcu_read_lock rcu_read_lock_bh &list->lock#5 irq_context: softirq rcu_read_lock rcu_read_lock rcu_read_lock rcu_read_lock &nf_nat_locks[i] irq_context: 0 sk_lock-AF_INET6 &tcp_hashinfo.bhash[i].lock &obj_hash[i].lock irq_context: 0 sk_lock-AF_INET6 rcu_read_lock &sighand->siglock irq_context: 0 sk_lock-AF_INET6 &sighand->siglock irq_context: 0 sk_lock-AF_INET6 &sighand->siglock stock_lock irq_context: 0 sk_lock-AF_INET6 &sighand->siglock pool_lock#2 irq_context: 0 sk_lock-AF_INET6 &sighand->siglock &p->pi_lock irq_context: 0 sb_writers#8 kn->active#5 fs_reclaim &rq->__lock irq_context: 0 sb_writers#8 kn->active#5 fs_reclaim &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 sb_writers#4 sb_internal jbd2_handle &journal->j_wait_updates &p->pi_lock irq_context: 0 sb_writers#4 sb_internal jbd2_handle &journal->j_wait_updates &p->pi_lock &rq->__lock irq_context: 0 sb_writers#4 sb_internal jbd2_handle &journal->j_wait_updates &p->pi_lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)bat_events (work_completion)(&(&forw_packet_aggr->delayed_work)->work) quarantine_lock irq_context: 0 &sb->s_type->i_lock_key#8 bit_wait_table + i irq_context: 0 &sb->s_type->i_mutex_key#10 fill_pool_map-wait-type-override pool_lock#2 irq_context: 0 &sb->s_type->i_mutex_key#10 fill_pool_map-wait-type-override pool_lock irq_context: 0 sb_writers#4 mapping.invalidate_lock rcu_read_lock &obj_hash[i].lock pool_lock irq_context: 0 (wq_completion)rcu_gp (work_completion)(&(&ssp->srcu_sup->work)->work) &ssp->srcu_sup->srcu_cb_mutex rcu_read_lock &pool->lock &p->pi_lock &rq->__lock &cfs_rq->removed.lock irq_context: softirq &(&local->scan_work)->timer rcu_read_lock &pool->lock/1 &p->pi_lock &rq->__lock irq_context: softirq &(&local->scan_work)->timer rcu_read_lock &pool->lock/1 &p->pi_lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 sb_writers#4 &sb->s_type->i_mutex_key#8 &folio_wait_table[i] irq_context: 0 tasklist_lock &sighand->siglock stock_lock irq_context: 0 tasklist_lock &sighand->siglock &obj_hash[i].lock irq_context: 0 sb_writers#4 &type->i_mutex_dir_key#3/1 &sb->s_type->i_mutex_key#8 remove_cache_srcu pool_lock#2 irq_context: 0 sb_writers#8 kn->active#5 batched_entropy_u8.lock irq_context: 0 sb_writers#8 kn->active#5 kfence_freelist_lock irq_context: 0 sb_writers#4 &sb->s_type->i_mutex_key#8 &fq->mq_flush_lock irq_context: 0 sb_writers#4 &sb->s_type->i_mutex_key#8 &fq->mq_flush_lock tk_core.seq.seqcount irq_context: 0 sb_writers#4 &sb->s_type->i_mutex_key#8 &fq->mq_flush_lock &q->requeue_lock irq_context: 0 sb_writers#4 &sb->s_type->i_mutex_key#8 &fq->mq_flush_lock &obj_hash[i].lock irq_context: 0 sb_writers#4 &sb->s_type->i_mutex_key#8 &fq->mq_flush_lock rcu_read_lock &pool->lock irq_context: 0 sb_writers#4 &sb->s_type->i_mutex_key#8 &fq->mq_flush_lock rcu_read_lock &pool->lock &obj_hash[i].lock irq_context: 0 sb_writers#4 &sb->s_type->i_mutex_key#8 &fq->mq_flush_lock rcu_read_lock &pool->lock &p->pi_lock irq_context: 0 sb_writers#4 &sb->s_type->i_mutex_key#8 &fq->mq_flush_lock rcu_read_lock &pool->lock &p->pi_lock &rq->__lock irq_context: 0 sb_writers#4 &sb->s_type->i_mutex_key#8 &fq->mq_flush_lock rcu_read_lock &pool->lock &p->pi_lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 sb_writers#4 &sb->s_type->i_mutex_key#8 &x->wait#26 irq_context: 0 sb_writers#4 &type->i_mutex_dir_key#3/1 &sb->s_type->i_mutex_key#8 jbd2_handle &ret->b_state_lock bit_wait_table + i irq_context: 0 sb_writers#4 &p->pi_lock &rq->__lock &cfs_rq->removed.lock irq_context: 0 &sb->s_type->i_mutex_key#8 &sbi->s_writepages_rwsem jbd2_handle &ei->i_data_sem &lg->lg_mutex bit_wait_table + i irq_context: 0 &sb->s_type->i_mutex_key#8 &sbi->s_writepages_rwsem jbd2_handle &ei->i_data_sem &lg->lg_mutex key#3 irq_context: 0 sb_writers#4 &type->i_mutex_dir_key#3 rcu_read_lock rcu_node_0 irq_context: 0 sb_writers#4 &type->i_mutex_dir_key#3 rcu_read_lock &rq->__lock irq_context: 0 sb_writers#4 &type->i_mutex_dir_key#3 rcu_read_lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 sb_writers#5 &type->i_mutex_dir_key#5 fs_reclaim &rq->__lock irq_context: 0 sb_writers#5 &type->i_mutex_dir_key#5 fs_reclaim &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 dup_mmap_sem &mm->mmap_lock &mm->mmap_lock/1 remove_cache_srcu &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 sk_lock-AF_INET6 &sighand->siglock &p->pi_lock &rq->__lock irq_context: 0 sk_lock-AF_INET6 &sighand->siglock &p->pi_lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 dup_mmap_sem &mm->mmap_lock &mm->mmap_lock/1 rcu_read_lock &cfs_rq->removed.lock irq_context: 0 dup_mmap_sem &mm->mmap_lock &mm->mmap_lock/1 rcu_read_lock &obj_hash[i].lock irq_context: 0 dup_mmap_sem &mm->mmap_lock &mm->mmap_lock/1 fs_reclaim &cfs_rq->removed.lock irq_context: 0 dup_mmap_sem &mm->mmap_lock &mm->mmap_lock/1 fs_reclaim &obj_hash[i].lock irq_context: 0 sb_writers#5 &type->i_mutex_dir_key#5/1 tomoyo_ss remove_cache_srcu pool_lock#2 irq_context: 0 sb_writers#4 sb_internal rcu_node_0 irq_context: 0 sb_writers#4 tomoyo_ss remove_cache_srcu irq_context: 0 sb_writers#4 tomoyo_ss remove_cache_srcu quarantine_lock irq_context: 0 sb_writers#4 tomoyo_ss remove_cache_srcu &c->lock irq_context: 0 sb_writers#4 tomoyo_ss remove_cache_srcu &n->list_lock irq_context: 0 sb_writers#4 tomoyo_ss remove_cache_srcu rcu_read_lock pool_lock#2 irq_context: 0 sb_writers#4 tomoyo_ss remove_cache_srcu &obj_hash[i].lock irq_context: 0 sb_writers#4 tomoyo_ss remove_cache_srcu rcu_read_lock rcu_node_0 irq_context: 0 sb_writers#4 tomoyo_ss remove_cache_srcu rcu_read_lock &rq->__lock irq_context: 0 sb_writers#4 tomoyo_ss remove_cache_srcu rcu_read_lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 sb_writers#4 tomoyo_ss remove_cache_srcu &rq->__lock irq_context: 0 sb_writers#4 tomoyo_ss remove_cache_srcu &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)rcu_gp (work_completion)(&rew->rew_work) rcu_state.exp_wake_mutex pool_lock#2 irq_context: softirq (&ndev->rs_timer) rcu_read_lock rcu_read_lock rcu_read_lock_bh rcu_read_lock batched_entropy_u32.lock crngs.lock irq_context: softirq (&ndev->rs_timer) rcu_read_lock rcu_read_lock rcu_read_lock_bh _xmit_ETHER#2 rcu_read_lock &obj_hash[i].lock irq_context: softirq (&ndev->rs_timer) rcu_read_lock rcu_read_lock rcu_read_lock_bh _xmit_ETHER#2 rcu_read_lock pool_lock#2 irq_context: 0 sb_writers#4 &type->i_mutex_dir_key#3/1 jbd2_handle &ei->i_data_sem remove_cache_srcu irq_context: 0 sb_writers#4 &type->i_mutex_dir_key#3/1 jbd2_handle &ei->i_data_sem remove_cache_srcu quarantine_lock irq_context: 0 sb_writers#4 &type->i_mutex_dir_key#3/1 jbd2_handle &ei->i_data_sem remove_cache_srcu &c->lock irq_context: 0 sb_writers#4 &type->i_mutex_dir_key#3/1 jbd2_handle &ei->i_data_sem remove_cache_srcu &n->list_lock irq_context: 0 sb_writers#4 &type->i_mutex_dir_key#3/1 jbd2_handle &ei->i_data_sem remove_cache_srcu rcu_read_lock pool_lock#2 irq_context: 0 sb_writers#4 &type->i_mutex_dir_key#3/1 jbd2_handle &ei->i_data_sem remove_cache_srcu &obj_hash[i].lock irq_context: 0 sb_writers#4 &type->i_mutex_dir_key#3/1 jbd2_handle &ei->i_data_sem remove_cache_srcu pool_lock#2 irq_context: 0 (wq_completion)wg-crypt-wg1#5 (work_completion)(&peer->transmit_packet_work) &peer->endpoint_lock rcu_read_lock_bh batched_entropy_u32.lock irq_context: 0 (wq_completion)wg-crypt-wg2#6 (work_completion)(&peer->transmit_packet_work) &peer->endpoint_lock rcu_read_lock_bh batched_entropy_u32.lock crngs.lock irq_context: 0 &f->f_pos_lock sb_writers#4 &sb->s_type->i_mutex_key#8 rcu_read_lock &obj_hash[i].lock pool_lock irq_context: 0 &f->f_pos_lock sb_writers#4 &sb->s_type->i_mutex_key#8 &journal->j_state_lock &journal->j_wait_commit &p->pi_lock &rq->__lock &cfs_rq->removed.lock irq_context: 0 sb_writers#4 &type->i_mutex_dir_key#3 tomoyo_ss rcu_read_lock rcu_read_lock rcu_node_0 irq_context: 0 sb_writers#4 &type->i_mutex_dir_key#3 tomoyo_ss rcu_read_lock rcu_read_lock &rq->__lock irq_context: 0 sb_writers#4 &type->i_mutex_dir_key#3 tomoyo_ss rcu_read_lock rcu_read_lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 sb_writers#4 &type->i_mutex_dir_key#3 tomoyo_ss rcu_read_lock rcu_node_0 irq_context: 0 sb_writers#4 &sb->s_type->i_mutex_key#8 mapping.invalidate_lock jbd2_handle bit_wait_table + i irq_context: 0 sb_writers#4 &sb->s_type->i_mutex_key#8 mapping.invalidate_lock jbd2_handle &rq->__lock irq_context: 0 sb_writers#4 &sb->s_type->i_mutex_key#8 mapping.invalidate_lock jbd2_handle &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 cgroup_threadgroup_rwsem freezer_mutex &rq->__lock irq_context: 0 cgroup_threadgroup_rwsem freezer_mutex &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 ppp_mutex irq_context: 0 ppp_mutex &mm->mmap_lock irq_context: 0 ppp_mutex &mm->mmap_lock ptlock_ptr(page)#2 irq_context: 0 ppp_mutex fs_reclaim irq_context: 0 ppp_mutex fs_reclaim mmu_notifier_invalidate_range_start irq_context: 0 ppp_mutex fs_reclaim mmu_notifier_invalidate_range_start &rq->__lock irq_context: 0 ppp_mutex fs_reclaim mmu_notifier_invalidate_range_start &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 ppp_mutex stock_lock irq_context: 0 ppp_mutex pool_lock#2 irq_context: 0 ppp_mutex stack_depot_init_mutex irq_context: 0 ppp_mutex rtnl_mutex irq_context: 0 ppp_mutex rtnl_mutex pcpu_alloc_mutex irq_context: 0 ppp_mutex rtnl_mutex pcpu_alloc_mutex pcpu_lock irq_context: 0 ppp_mutex rtnl_mutex &pn->all_ppp_mutex irq_context: 0 ppp_mutex rtnl_mutex &pn->all_ppp_mutex &rq->__lock irq_context: 0 ppp_mutex rtnl_mutex &pn->all_ppp_mutex &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 ppp_mutex rtnl_mutex &pn->all_ppp_mutex fs_reclaim irq_context: 0 ppp_mutex rtnl_mutex &pn->all_ppp_mutex fs_reclaim mmu_notifier_invalidate_range_start irq_context: 0 ppp_mutex rtnl_mutex &pn->all_ppp_mutex pool_lock#2 irq_context: 0 ppp_mutex rtnl_mutex fs_reclaim irq_context: 0 ppp_mutex rtnl_mutex fs_reclaim mmu_notifier_invalidate_range_start irq_context: 0 ppp_mutex rtnl_mutex pool_lock#2 irq_context: 0 ppp_mutex rtnl_mutex &xa->xa_lock#3 irq_context: 0 ppp_mutex rtnl_mutex net_rwsem irq_context: 0 ppp_mutex rtnl_mutex net_rwsem &list->lock#2 irq_context: 0 ppp_mutex rtnl_mutex &tn->lock irq_context: 0 ppp_mutex rtnl_mutex &x->wait#9 irq_context: 0 ppp_mutex rtnl_mutex &obj_hash[i].lock irq_context: 0 ppp_mutex rtnl_mutex &k->list_lock irq_context: 0 ppp_mutex rtnl_mutex gdp_mutex irq_context: 0 ppp_mutex rtnl_mutex gdp_mutex &k->list_lock irq_context: 0 ppp_mutex rtnl_mutex lock irq_context: 0 ppp_mutex rtnl_mutex lock kernfs_idr_lock irq_context: 0 ppp_mutex rtnl_mutex &rq->__lock irq_context: 0 ppp_mutex rtnl_mutex &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 ppp_mutex rtnl_mutex &root->kernfs_rwsem irq_context: 0 ppp_mutex rtnl_mutex &root->kernfs_rwsem &root->kernfs_iattr_rwsem irq_context: 0 ppp_mutex rtnl_mutex bus_type_sem irq_context: 0 ppp_mutex rtnl_mutex lock kernfs_idr_lock pool_lock#2 irq_context: 0 ppp_mutex rtnl_mutex sysfs_symlink_target_lock irq_context: 0 ppp_mutex rtnl_mutex &c->lock irq_context: 0 ppp_mutex rtnl_mutex &____s->seqcount#2 irq_context: 0 ppp_mutex rtnl_mutex &____s->seqcount irq_context: 0 sb_writers#4 &n->list_lock irq_context: 0 sb_writers#4 &n->list_lock &c->lock irq_context: 0 sb_writers#4 &sb->s_type->i_mutex_key#8 jbd2_handle bit_wait_table + i irq_context: 0 &f->f_pos_lock sb_writers#4 &sb->s_type->i_mutex_key#8 jbd2_handle &journal->j_wait_updates &p->pi_lock &cfs_rq->removed.lock irq_context: 0 &f->f_pos_lock sb_writers#4 &sb->s_type->i_mutex_key#8 &rq->__lock &cfs_rq->removed.lock irq_context: 0 &journal->j_state_lock &journal->j_wait_transaction_locked &p->pi_lock &rq->__lock &cfs_rq->removed.lock irq_context: 0 ppp_mutex rtnl_mutex &cfs_rq->removed.lock irq_context: 0 ppp_mutex rtnl_mutex &root->kernfs_rwsem irq_context: 0 ppp_mutex rtnl_mutex &root->kernfs_rwsem &rq->__lock irq_context: 0 ppp_mutex rtnl_mutex &root->kernfs_rwsem &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 ppp_mutex rtnl_mutex &dev->power.lock irq_context: 0 ppp_mutex rtnl_mutex dpm_list_mtx irq_context: 0 ppp_mutex rtnl_mutex uevent_sock_mutex irq_context: 0 ppp_mutex rtnl_mutex uevent_sock_mutex fs_reclaim irq_context: 0 ppp_mutex rtnl_mutex uevent_sock_mutex fs_reclaim mmu_notifier_invalidate_range_start irq_context: 0 ppp_mutex rtnl_mutex uevent_sock_mutex pool_lock#2 irq_context: 0 ppp_mutex rtnl_mutex uevent_sock_mutex nl_table_lock irq_context: 0 ppp_mutex rtnl_mutex uevent_sock_mutex &obj_hash[i].lock irq_context: 0 ppp_mutex rtnl_mutex uevent_sock_mutex nl_table_wait.lock irq_context: 0 ppp_mutex rtnl_mutex subsys mutex#17 irq_context: 0 ppp_mutex rtnl_mutex subsys mutex#17 &k->k_lock irq_context: 0 ppp_mutex rtnl_mutex &n->list_lock irq_context: 0 ppp_mutex rtnl_mutex &n->list_lock &c->lock irq_context: 0 ppp_mutex rtnl_mutex rcu_read_lock pool_lock#2 irq_context: 0 ppp_mutex rtnl_mutex &dir->lock#2 irq_context: 0 ppp_mutex rtnl_mutex uevent_sock_mutex &c->lock irq_context: 0 ppp_mutex rtnl_mutex uevent_sock_mutex &rq->__lock irq_context: 0 ppp_mutex rtnl_mutex uevent_sock_mutex &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 ppp_mutex rtnl_mutex dev_hotplug_mutex irq_context: 0 ppp_mutex rtnl_mutex dev_hotplug_mutex &dev->power.lock irq_context: 0 ppp_mutex rtnl_mutex dev_base_lock irq_context: 0 ppp_mutex rtnl_mutex input_pool.lock irq_context: 0 ppp_mutex rtnl_mutex batched_entropy_u32.lock irq_context: 0 ppp_mutex rtnl_mutex &tbl->lock irq_context: 0 ppp_mutex rtnl_mutex stock_lock irq_context: 0 ppp_mutex rtnl_mutex sysctl_lock irq_context: 0 ppp_mutex rtnl_mutex nl_table_lock irq_context: 0 ppp_mutex rtnl_mutex nl_table_wait.lock irq_context: 0 ppp_mutex rtnl_mutex proc_subdir_lock irq_context: 0 ppp_mutex rtnl_mutex proc_inum_ida.xa_lock irq_context: 0 ppp_mutex rtnl_mutex proc_subdir_lock irq_context: 0 ppp_mutex rtnl_mutex &idev->mc_lock irq_context: 0 ppp_mutex rtnl_mutex &idev->mc_lock fs_reclaim irq_context: 0 ppp_mutex rtnl_mutex &idev->mc_lock fs_reclaim mmu_notifier_invalidate_range_start irq_context: 0 ppp_mutex rtnl_mutex &idev->mc_lock pool_lock#2 irq_context: 0 ppp_mutex rtnl_mutex &idev->mc_lock &obj_hash[i].lock irq_context: 0 ppp_mutex rtnl_mutex &idev->mc_lock &c->lock irq_context: 0 ppp_mutex rtnl_mutex &pnettable->lock irq_context: 0 ppp_mutex rtnl_mutex smc_ib_devices.mutex irq_context: 0 ppp_mutex rtnl_mutex &ppp->rlock irq_context: 0 ppp_mutex rtnl_mutex &ppp->wlock irq_context: 0 ppp_mutex rtnl_mutex rcu_read_lock &ndev->lock irq_context: 0 ppp_mutex rtnl_mutex.wait_lock irq_context: 0 ppp_mutex &p->pi_lock irq_context: 0 ppp_mutex &p->pi_lock &rq->__lock irq_context: 0 ppp_mutex &p->pi_lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 ppp_mutex &rq->__lock irq_context: 0 ppp_mutex &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 ppp_mutex &mm->mmap_lock &sem->wait_lock irq_context: 0 ppp_mutex &mm->mmap_lock &rq->__lock irq_context: 0 ppp_mutex &mm->mmap_lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 ppp_mutex &mm->mmap_lock &rq->__lock &cfs_rq->removed.lock irq_context: 0 ppp_mutex &mm->mmap_lock fs_reclaim irq_context: 0 ppp_mutex &mm->mmap_lock fs_reclaim mmu_notifier_invalidate_range_start irq_context: 0 ppp_mutex &mm->mmap_lock &____s->seqcount irq_context: 0 ppp_mutex &mm->mmap_lock stock_lock irq_context: 0 ppp_mutex &mm->mmap_lock mmu_notifier_invalidate_range_start irq_context: 0 ppp_mutex &mm->mmap_lock ptlock_ptr(page)#2 lock#4 irq_context: 0 sb_writers#4 mapping.invalidate_lock &xa->xa_lock#9 &n->list_lock irq_context: 0 sb_writers#4 mapping.invalidate_lock &xa->xa_lock#9 &____s->seqcount irq_context: 0 sk_lock-AF_INET stock_lock irq_context: 0 sk_lock-AF_INET mmu_notifier_invalidate_range_start irq_context: 0 sk_lock-AF_INET &sb->s_type->i_lock_key#8 irq_context: 0 sk_lock-AF_INET &dir->lock irq_context: 0 sk_lock-AF_INET k-sk_lock-AF_INET/1 irq_context: 0 sk_lock-AF_INET k-sk_lock-AF_INET/1 k-slock-AF_INET irq_context: 0 sk_lock-AF_INET k-sk_lock-AF_INET/1 pool_lock#2 irq_context: 0 sk_lock-AF_INET k-sk_lock-AF_INET/1 &dir->lock irq_context: 0 sk_lock-AF_INET k-sk_lock-AF_INET/1 fs_reclaim irq_context: 0 sk_lock-AF_INET k-sk_lock-AF_INET/1 fs_reclaim mmu_notifier_invalidate_range_start irq_context: 0 sk_lock-AF_INET k-sk_lock-AF_INET/1 &c->lock irq_context: 0 sk_lock-AF_INET k-slock-AF_INET irq_context: 0 sk_lock-AF_INET k-clock-AF_INET irq_context: 0 sk_lock-AF_INET &xa->xa_lock#9 irq_context: 0 sk_lock-AF_INET &fsnotify_mark_srcu irq_context: 0 sb_writers#4 &sb->s_type->i_mutex_key#8 jbd2_handle rcu_read_lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 sk_lock-AF_INET k-sk_lock-AF_INET irq_context: 0 sk_lock-AF_INET k-sk_lock-AF_INET k-slock-AF_INET irq_context: 0 sk_lock-AF_INET k-sk_lock-AF_INET &tcp_hashinfo.bhash[i].lock irq_context: 0 sk_lock-AF_INET k-sk_lock-AF_INET &tcp_hashinfo.bhash[i].lock stock_lock irq_context: 0 sk_lock-AF_INET k-sk_lock-AF_INET &tcp_hashinfo.bhash[i].lock pool_lock#2 irq_context: 0 sk_lock-AF_INET k-sk_lock-AF_INET &tcp_hashinfo.bhash[i].lock k-clock-AF_INET irq_context: 0 sk_lock-AF_INET k-sk_lock-AF_INET &tcp_hashinfo.bhash[i].lock &tcp_hashinfo.bhash2[i].lock irq_context: 0 sk_lock-AF_INET k-sk_lock-AF_INET &tcp_hashinfo.bhash[i].lock &tcp_hashinfo.bhash2[i].lock stock_lock irq_context: 0 sk_lock-AF_INET k-sk_lock-AF_INET &tcp_hashinfo.bhash[i].lock &tcp_hashinfo.bhash2[i].lock pool_lock#2 irq_context: 0 sk_lock-AF_INET k-sk_lock-AF_INET &tcp_hashinfo.bhash[i].lock &tcp_hashinfo.bhash2[i].lock k-clock-AF_INET irq_context: 0 sk_lock-AF_INET crngs.lock irq_context: 0 sk_lock-AF_INET &token_hash[i].lock irq_context: 0 sk_lock-AF_INET k-sk_lock-AF_INET rcu_read_lock &____s->seqcount irq_context: 0 sk_lock-AF_INET k-sk_lock-AF_INET rcu_read_lock pool_lock#2 irq_context: 0 sk_lock-AF_INET k-sk_lock-AF_INET rcu_read_lock rcu_read_lock pool_lock#2 irq_context: 0 sk_lock-AF_INET k-sk_lock-AF_INET rcu_read_lock &obj_hash[i].lock irq_context: 0 sk_lock-AF_INET k-sk_lock-AF_INET rcu_read_lock &dir->lock#2 irq_context: 0 sk_lock-AF_INET k-sk_lock-AF_INET rcu_read_lock &ul->lock irq_context: 0 sk_lock-AF_INET k-sk_lock-AF_INET &obj_hash[i].lock irq_context: 0 sk_lock-AF_INET k-sk_lock-AF_INET pool_lock#2 irq_context: 0 sk_lock-AF_INET k-sk_lock-AF_INET rcu_read_lock &c->lock irq_context: 0 sk_lock-AF_INET k-sk_lock-AF_INET stock_lock irq_context: 0 sk_lock-AF_INET k-sk_lock-AF_INET &rq->__lock irq_context: 0 sk_lock-AF_INET k-sk_lock-AF_INET &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 sk_lock-AF_INET k-sk_lock-AF_INET per_cpu_ptr(&cgroup_rstat_cpu_lock, cpu) irq_context: 0 sk_lock-AF_INET k-sk_lock-AF_INET stock_lock rcu_read_lock per_cpu_ptr(&cgroup_rstat_cpu_lock, cpu) irq_context: 0 sk_lock-AF_INET k-sk_lock-AF_INET &tcp_hashinfo.bhash[i].lock &tcp_hashinfo.bhash2[i].lock &obj_hash[i].lock irq_context: 0 sk_lock-AF_INET k-sk_lock-AF_INET &hashinfo->ehash_locks[i] irq_context: 0 sk_lock-AF_INET k-sk_lock-AF_INET batched_entropy_u32.lock irq_context: 0 sk_lock-AF_INET k-sk_lock-AF_INET tk_core.seq.seqcount irq_context: 0 sk_lock-AF_INET k-sk_lock-AF_INET batched_entropy_u16.lock irq_context: 0 sk_lock-AF_INET k-sk_lock-AF_INET fs_reclaim irq_context: 0 sk_lock-AF_INET k-sk_lock-AF_INET fs_reclaim mmu_notifier_invalidate_range_start irq_context: 0 rtnl_mutex &ppp->rlock irq_context: 0 rtnl_mutex &ppp->wlock irq_context: 0 sk_lock-AF_INET k-sk_lock-AF_INET &c->lock irq_context: 0 sk_lock-AF_INET k-sk_lock-AF_INET rcu_read_lock rcu_read_lock &____s->seqcount#7 irq_context: 0 sk_lock-AF_INET k-sk_lock-AF_INET rcu_read_lock rcu_read_lock &ct->lock irq_context: 0 sk_lock-AF_INET k-sk_lock-AF_INET rcu_read_lock rcu_read_lock rcu_read_lock &____s->seqcount#7 irq_context: 0 sk_lock-AF_INET k-sk_lock-AF_INET rcu_read_lock rcu_read_lock &nf_nat_locks[i] irq_context: 0 sk_lock-AF_INET k-sk_lock-AF_INET rcu_read_lock rcu_read_lock rcu_node_0 irq_context: 0 sk_lock-AF_INET k-sk_lock-AF_INET rcu_read_lock rcu_read_lock &rq->__lock irq_context: 0 sk_lock-AF_INET k-sk_lock-AF_INET rcu_read_lock rcu_read_lock &nf_conntrack_locks[i] irq_context: 0 sk_lock-AF_INET k-sk_lock-AF_INET rcu_read_lock rcu_read_lock &nf_conntrack_locks[i] batched_entropy_u8.lock irq_context: 0 sk_lock-AF_INET k-sk_lock-AF_INET rcu_read_lock rcu_read_lock &____s->seqcount#10 irq_context: 0 sk_lock-AF_INET k-sk_lock-AF_INET rcu_read_lock rcu_read_lock rcu_read_lock_bh &list->lock#5 irq_context: softirq rcu_read_lock rcu_read_lock k-slock-AF_INET/1 irq_context: 0 sk_lock-AF_INET k-sk_lock-AF_INET rcu_read_lock rcu_node_0 irq_context: 0 sk_lock-AF_INET k-sk_lock-AF_INET &base->lock irq_context: 0 sk_lock-AF_INET k-sk_lock-AF_INET &base->lock &obj_hash[i].lock irq_context: 0 rtnl_mutex &dev_addr_list_lock_key#4 irq_context: 0 rtnl_mutex &ppp->wlock &ppp->rlock irq_context: 0 rtnl_mutex &pn->all_ppp_mutex irq_context: 0 rtnl_mutex &pn->all_ppp_mutex &obj_hash[i].lock irq_context: 0 rtnl_mutex &pn->all_ppp_mutex pool_lock#2 irq_context: 0 rtnl_mutex &pf->rwait irq_context: 0 sk_lock-AF_INET k-sk_lock-AF_INET rcu_read_lock tcp_metrics_lock irq_context: 0 sk_lock-AF_INET k-sk_lock-AF_INET rcu_read_lock tcp_metrics_lock pool_lock#2 irq_context: 0 sk_lock-AF_INET k-sk_lock-AF_INET slock-AF_INET irq_context: 0 sk_lock-AF_INET &ei->socket.wq.wait irq_context: 0 sk_lock-AF_INET slock-AF_INET rcu_read_lock &ei->socket.wq.wait irq_context: 0 rds_sock_lock irq_context: 0 rtnl_mutex &root->kernfs_rwsem &rq->__lock irq_context: 0 rtnl_mutex &root->kernfs_rwsem &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 sb_writers#4 mapping.invalidate_lock per_cpu_ptr(&cgroup_rstat_cpu_lock, cpu) irq_context: 0 sb_writers#8 &root->kernfs_iattr_rwsem &rq->__lock irq_context: 0 sb_writers#8 &root->kernfs_iattr_rwsem &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 sk_lock-AF_INET6 crypto_alg_sem irq_context: 0 sb_writers#4 &sb->s_type->i_mutex_key#8 &journal->j_state_lock &journal->j_wait_commit &p->pi_lock &cfs_rq->removed.lock irq_context: 0 sb_writers#4 &type->i_mutex_dir_key#3/1 &sb->s_type->i_mutex_key#8 jbd2_handle &journal->j_wait_updates &p->pi_lock irq_context: 0 sb_writers#4 &type->i_mutex_dir_key#3/1 &sb->s_type->i_mutex_key#8 jbd2_handle &journal->j_wait_updates &p->pi_lock &rq->__lock irq_context: 0 sb_writers#4 &type->i_mutex_dir_key#3/1 &sb->s_type->i_mutex_key#8 jbd2_handle &journal->j_wait_updates &p->pi_lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 sk_lock-AF_INET rcu_read_lock &ei->socket.wq.wait &p->pi_lock irq_context: 0 sk_lock-AF_INET rcu_read_lock &ei->socket.wq.wait &p->pi_lock &rq->__lock irq_context: 0 sk_lock-AF_INET rcu_read_lock &ei->socket.wq.wait &p->pi_lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 &sb->s_type->i_mutex_key#10 clock-AF_NETROM irq_context: 0 &sb->s_type->i_mutex_key#10 sk_lock-AF_NETROM irq_context: 0 &sb->s_type->i_mutex_key#10 sk_lock-AF_NETROM slock-AF_NETROM irq_context: 0 &sb->s_type->i_mutex_key#10 sk_lock-AF_NETROM &obj_hash[i].lock irq_context: 0 &sb->s_type->i_mutex_key#10 sk_lock-AF_NETROM wlock-AF_NETROM irq_context: 0 &sb->s_type->i_mutex_key#10 sk_lock-AF_NETROM &list->lock#25 irq_context: 0 &sb->s_type->i_mutex_key#10 sk_lock-AF_NETROM nr_list_lock irq_context: 0 &sb->s_type->i_mutex_key#10 sk_lock-AF_NETROM rlock-AF_NETROM irq_context: 0 &sb->s_type->i_mutex_key#10 slock-AF_NETROM irq_context: 0 &sb->s_type->i_mutex_key#10 sk_lock-AF_INET k-sk_lock-AF_INET irq_context: 0 &sb->s_type->i_mutex_key#10 sk_lock-AF_INET k-sk_lock-AF_INET k-slock-AF_INET irq_context: 0 &sb->s_type->i_mutex_key#10 sk_lock-AF_INET k-sk_lock-AF_INET fs_reclaim irq_context: 0 &sb->s_type->i_mutex_key#10 sk_lock-AF_INET k-sk_lock-AF_INET fs_reclaim mmu_notifier_invalidate_range_start irq_context: 0 &sb->s_type->i_mutex_key#10 sk_lock-AF_INET k-sk_lock-AF_INET batched_entropy_u8.lock irq_context: 0 &sb->s_type->i_mutex_key#10 sk_lock-AF_INET k-sk_lock-AF_INET kfence_freelist_lock irq_context: 0 &sb->s_type->i_mutex_key#10 sk_lock-AF_INET k-sk_lock-AF_INET pool_lock#2 irq_context: 0 &sb->s_type->i_mutex_key#10 sk_lock-AF_INET k-sk_lock-AF_INET tk_core.seq.seqcount irq_context: 0 &sb->s_type->i_mutex_key#10 sk_lock-AF_INET k-sk_lock-AF_INET rcu_read_lock rcu_read_lock &____s->seqcount#7 irq_context: 0 &sb->s_type->i_mutex_key#10 sk_lock-AF_INET k-sk_lock-AF_INET rcu_read_lock rcu_read_lock &ct->lock irq_context: 0 &sb->s_type->i_mutex_key#10 sk_lock-AF_INET k-sk_lock-AF_INET rcu_read_lock rcu_read_lock &____s->seqcount#10 irq_context: 0 &sb->s_type->i_mutex_key#10 sk_lock-AF_INET k-sk_lock-AF_INET rcu_read_lock rcu_read_lock rcu_read_lock_bh &list->lock#5 irq_context: 0 &sb->s_type->i_mutex_key#10 sk_lock-AF_INET k-sk_lock-AF_INET &obj_hash[i].lock irq_context: 0 &sb->s_type->i_mutex_key#10 sk_lock-AF_INET k-sk_lock-AF_INET &base->lock irq_context: 0 &sb->s_type->i_mutex_key#10 sk_lock-AF_INET k-sk_lock-AF_INET &base->lock &obj_hash[i].lock irq_context: 0 &sb->s_type->i_mutex_key#10 sk_lock-AF_INET k-sk_lock-AF_INET rcu_read_lock &pool->lock irq_context: 0 &sb->s_type->i_mutex_key#10 sk_lock-AF_INET k-sk_lock-AF_INET rcu_read_lock &pool->lock &obj_hash[i].lock irq_context: 0 &sb->s_type->i_mutex_key#10 sk_lock-AF_INET k-sk_lock-AF_INET rcu_read_lock &pool->lock pool_lock#2 irq_context: 0 &sb->s_type->i_mutex_key#10 sk_lock-AF_INET k-sk_lock-AF_INET rcu_read_lock &pool->lock &p->pi_lock irq_context: 0 &sb->s_type->i_mutex_key#10 sk_lock-AF_INET k-sk_lock-AF_INET slock-AF_INET irq_context: 0 &sb->s_type->i_mutex_key#10 sk_lock-AF_INET k-sk_lock-AF_INET &meta->lock irq_context: 0 &sb->s_type->i_mutex_key#10 sk_lock-AF_INET k-sk_lock-AF_INET stock_lock irq_context: 0 &sb->s_type->i_mutex_key#10 sk_lock-AF_INET k-sk_lock-AF_INET &____s->seqcount#2 irq_context: 0 &sb->s_type->i_mutex_key#10 sk_lock-AF_INET k-sk_lock-AF_INET &____s->seqcount irq_context: 0 &sb->s_type->i_mutex_key#10 sk_lock-AF_INET k-sk_lock-AF_INET &c->lock irq_context: 0 &sb->s_type->i_mutex_key#10 sk_lock-AF_INET k-sk_lock-AF_INET &tcp_hashinfo.bhash[i].lock irq_context: 0 &sb->s_type->i_mutex_key#10 sk_lock-AF_INET k-sk_lock-AF_INET &tcp_hashinfo.bhash[i].lock &tcp_hashinfo.bhash2[i].lock irq_context: 0 &sb->s_type->i_mutex_key#10 sk_lock-AF_INET k-sk_lock-AF_INET &hashinfo->ehash_locks[i] irq_context: 0 &sb->s_type->i_mutex_key#10 sk_lock-AF_INET k-slock-AF_INET irq_context: 0 &sb->s_type->i_mutex_key#10 sk_lock-AF_INET k-sk_lock-AF_INET/1 irq_context: 0 &sb->s_type->i_mutex_key#10 sk_lock-AF_INET k-sk_lock-AF_INET/1 k-slock-AF_INET irq_context: 0 &sb->s_type->i_mutex_key#10 sk_lock-AF_INET k-sk_lock-AF_INET/1 k-clock-AF_INET irq_context: 0 &sb->s_type->i_mutex_key#10 sk_lock-AF_INET k-sk_lock-AF_INET/1 k-slock-AF_INET &obj_hash[i].lock irq_context: 0 &sb->s_type->i_mutex_key#10 sk_lock-AF_INET k-sk_lock-AF_INET/1 k-slock-AF_INET pool_lock#2 irq_context: 0 &sb->s_type->i_mutex_key#10 sk_lock-AF_INET k-sk_lock-AF_INET/1 k-slock-AF_INET krc.lock irq_context: 0 &sb->s_type->i_mutex_key#10 sk_lock-AF_INET k-sk_lock-AF_INET/1 k-slock-AF_INET &tcp_hashinfo.bhash[i].lock irq_context: 0 &sb->s_type->i_mutex_key#10 sk_lock-AF_INET k-sk_lock-AF_INET/1 k-slock-AF_INET &tcp_hashinfo.bhash[i].lock &tcp_hashinfo.bhash2[i].lock irq_context: 0 &sb->s_type->i_mutex_key#10 sk_lock-AF_INET k-sk_lock-AF_INET/1 k-slock-AF_INET elock-AF_INET irq_context: 0 &sb->s_type->i_mutex_key#10 sk_lock-AF_INET &dir->lock irq_context: 0 &sb->s_type->i_mutex_key#10 sk_lock-AF_INET stock_lock irq_context: 0 &sb->s_type->i_mutex_key#10 sk_lock-AF_INET &token_hash[i].lock irq_context: 0 &sb->s_type->i_mutex_key#10 sk_lock-AF_INET &msk->pm.lock irq_context: 0 &sb->s_type->i_mutex_key#10 rcu_read_lock &pool->lock irq_context: 0 &sb->s_type->i_mutex_key#10 rcu_read_lock &pool->lock &obj_hash[i].lock irq_context: 0 &sb->s_type->i_mutex_key#10 (work_completion)(&msk->work) irq_context: 0 &sb->s_type->i_mutex_key#10 clock-AF_RDS irq_context: 0 &sb->s_type->i_mutex_key#10 &rs->rs_recv_lock irq_context: 0 &sb->s_type->i_mutex_key#10 rds_cong_monitor_lock irq_context: 0 &sb->s_type->i_mutex_key#10 rds_cong_lock irq_context: 0 &sb->s_type->i_mutex_key#10 &rs->rs_lock irq_context: 0 &sb->s_type->i_mutex_key#10 &rs->rs_rdma_lock irq_context: 0 &sb->s_type->i_mutex_key#10 &q->lock irq_context: 0 &sb->s_type->i_mutex_key#10 rds_sock_lock irq_context: 0 &ppp->wlock irq_context: 0 &ppp->wlock &ppp->rlock irq_context: 0 &list->lock#26 irq_context: 0 sb_writers#4 remove_cache_srcu irq_context: 0 sb_writers#4 remove_cache_srcu quarantine_lock irq_context: 0 sb_writers#4 remove_cache_srcu &c->lock irq_context: 0 sb_writers#4 remove_cache_srcu &n->list_lock irq_context: 0 sb_writers#4 remove_cache_srcu rcu_read_lock pool_lock#2 irq_context: 0 sb_writers#4 remove_cache_srcu &obj_hash[i].lock irq_context: 0 sb_writers#4 remove_cache_srcu pool_lock#2 irq_context: 0 sb_writers#4 &sb->s_type->i_mutex_key#8 jbd2_handle &ei->i_data_sem remove_cache_srcu irq_context: 0 sb_writers#4 &sb->s_type->i_mutex_key#8 jbd2_handle &ei->i_data_sem remove_cache_srcu quarantine_lock irq_context: 0 sb_writers#4 &type->i_mutex_dir_key#3/1 &journal->j_state_lock &journal->j_wait_transaction_locked irq_context: 0 sb_writers#4 &type->i_mutex_dir_key#3/1 jbd2_handle irq_context: 0 sb_writers#4 &type->i_mutex_dir_key#3 fs_reclaim &rq->__lock irq_context: 0 sb_writers#4 &type->i_mutex_dir_key#3 fs_reclaim &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 sb_writers#4 &type->i_mutex_dir_key#3 &journal->j_state_lock &journal->j_wait_transaction_locked irq_context: 0 sb_writers#4 &type->i_mutex_dir_key#3 jbd2_handle irq_context: 0 sb_writers#4 &type->i_mutex_dir_key#3/1 jbd2_handle &journal->j_wait_updates &p->pi_lock irq_context: 0 sb_writers#4 &type->i_mutex_dir_key#3/1 jbd2_handle &journal->j_wait_updates &p->pi_lock &rq->__lock irq_context: 0 sb_writers#4 &type->i_mutex_dir_key#3/1 jbd2_handle &journal->j_wait_updates &p->pi_lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 sb_writers#4 &type->i_mutex_dir_key#3 jbd2_handle bit_wait_table + i irq_context: 0 &f->f_pos_lock &type->i_mutex_dir_key#3 sb_writers#4 jbd2_handle bit_wait_table + i irq_context: 0 &xa->xa_lock#9 &____s->seqcount#2 irq_context: 0 rtnl_mutex &mm->mmap_lock &sem->wait_lock irq_context: 0 rtnl_mutex &mm->mmap_lock &rq->__lock irq_context: 0 rtnl_mutex &mm->mmap_lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 sb_writers#4 sb_writers#4 &n->list_lock irq_context: 0 sk_lock-AF_INET k-sk_lock-AF_INET (&tw->tw_timer) irq_context: 0 sk_lock-AF_INET6 crypto_alg_sem &rq->__lock irq_context: 0 &sb->s_type->i_mutex_key#10 sk_lock-AF_INET k-sk_lock-AF_INET rcu_read_lock &pool->lock &p->pi_lock &rq->__lock irq_context: 0 &sb->s_type->i_mutex_key#10 sk_lock-AF_INET k-sk_lock-AF_INET rcu_read_lock &pool->lock &p->pi_lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 &sb->s_type->i_mutex_key#10 sk_lock-AF_INET k-sk_lock-AF_INET &rq->__lock irq_context: 0 &sb->s_type->i_mutex_key#10 sk_lock-AF_INET k-sk_lock-AF_INET &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)events (work_completion)(&msk->work) irq_context: 0 (wq_completion)events (work_completion)(&msk->work) sk_lock-AF_INET irq_context: 0 (wq_completion)events (work_completion)(&msk->work) sk_lock-AF_INET slock-AF_INET irq_context: 0 (wq_completion)events (work_completion)(&msk->work) sk_lock-AF_INET slock-AF_INET &sk->sk_lock.wq irq_context: 0 (wq_completion)events (work_completion)(&msk->work) sk_lock-AF_INET &pool->lock irq_context: 0 (wq_completion)events (work_completion)(&msk->work) sk_lock-AF_INET &pool->lock &p->pi_lock irq_context: 0 (wq_completion)events (work_completion)(&msk->work) sk_lock-AF_INET &pool->lock &p->pi_lock &rq->__lock irq_context: 0 (wq_completion)events (work_completion)(&msk->work) sk_lock-AF_INET &pool->lock &p->pi_lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)events (work_completion)(&msk->work) sk_lock-AF_INET &rq->__lock irq_context: 0 (wq_completion)events (work_completion)(&msk->work) sk_lock-AF_INET &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 &sb->s_type->i_mutex_key#10 sk_lock-AF_INET k-sk_lock-AF_INET slock-AF_INET rcu_read_lock &pool->lock irq_context: 0 &sb->s_type->i_mutex_key#10 sk_lock-AF_INET k-sk_lock-AF_INET slock-AF_INET rcu_read_lock &pool->lock &obj_hash[i].lock irq_context: 0 &sb->s_type->i_mutex_key#10 sk_lock-AF_INET k-sk_lock-AF_INET slock-AF_INET rcu_read_lock &pool->lock pool_lock#2 irq_context: 0 &sb->s_type->i_mutex_key#10 sk_lock-AF_INET k-sk_lock-AF_INET slock-AF_INET rcu_read_lock &pool->lock &p->pi_lock irq_context: 0 &sb->s_type->i_mutex_key#10 sk_lock-AF_INET k-sk_lock-AF_INET slock-AF_INET rcu_read_lock &pool->lock &p->pi_lock &rq->__lock irq_context: 0 &sb->s_type->i_mutex_key#10 sk_lock-AF_INET k-sk_lock-AF_INET slock-AF_INET rcu_read_lock &pool->lock &p->pi_lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 &sb->s_type->i_mutex_key#10 slock-AF_INET &sk->sk_lock.wq irq_context: 0 &sb->s_type->i_mutex_key#10 slock-AF_INET &sk->sk_lock.wq &p->pi_lock irq_context: 0 &sb->s_type->i_mutex_key#10 slock-AF_INET &sk->sk_lock.wq &p->pi_lock &rq->__lock irq_context: 0 &sb->s_type->i_mutex_key#10 slock-AF_INET &sk->sk_lock.wq &p->pi_lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)events (work_completion)(&msk->work) slock-AF_INET irq_context: 0 (wq_completion)events (work_completion)(&msk->work) &rq->__lock irq_context: 0 (wq_completion)events (work_completion)(&msk->work) &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 remove_cache_srcu &base->lock irq_context: 0 remove_cache_srcu &base->lock &obj_hash[i].lock irq_context: 0 sb_writers#4 &sb->s_type->i_mutex_key#8 batched_entropy_u8.lock irq_context: 0 sb_writers#4 &sb->s_type->i_mutex_key#8 kfence_freelist_lock irq_context: 0 sb_writers#4 &sb->s_type->i_mutex_key#8 &meta->lock irq_context: 0 remove_cache_srcu &____s->seqcount irq_context: 0 sb_writers#4 sb_internal &journal->j_state_lock irq_context: 0 sb_writers#4 sb_internal &journal->j_state_lock tk_core.seq.seqcount irq_context: 0 sb_writers#4 sb_internal &journal->j_state_lock &obj_hash[i].lock irq_context: 0 sb_writers#4 sb_internal &journal->j_state_lock &base->lock irq_context: 0 sb_writers#4 sb_internal &journal->j_state_lock &base->lock &obj_hash[i].lock irq_context: 0 ppp_mutex &c->lock irq_context: 0 ppp_mutex rtnl_mutex &root->kernfs_rwsem &rq->__lock irq_context: 0 ppp_mutex rtnl_mutex &root->kernfs_rwsem &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 sb_writers#4 sb_internal jbd2_handle bit_wait_table + i irq_context: 0 ppp_mutex rtnl_mutex &root->kernfs_rwsem &cfs_rq->removed.lock irq_context: 0 ppp_mutex rtnl_mutex &root->kernfs_rwsem &obj_hash[i].lock irq_context: 0 sb_writers#4 &sb->s_type->i_mutex_key#8 jbd2_handle &ei->i_data_sem &ei->i_es_lock &obj_hash[i].lock irq_context: 0 sb_writers#4 &sb->s_type->i_mutex_key#8 &obj_hash[i].lock pool_lock irq_context: 0 sb_writers#4 &type->i_mutex_dir_key#3/1 tomoyo_ss remove_cache_srcu pool_lock#2 irq_context: 0 ppp_mutex ppp_mutex.wait_lock irq_context: 0 ppp_mutex.wait_lock irq_context: 0 ppp_mutex &cfs_rq->removed.lock irq_context: 0 ppp_mutex &obj_hash[i].lock irq_context: 0 sb_writers#4 sb_writers#4 remove_cache_srcu irq_context: 0 sb_writers#4 sb_writers#4 remove_cache_srcu quarantine_lock irq_context: 0 sb_writers#4 sb_writers#4 remove_cache_srcu &c->lock irq_context: 0 sb_writers#4 sb_writers#4 remove_cache_srcu &n->list_lock irq_context: 0 sb_writers#4 sb_writers#4 remove_cache_srcu rcu_read_lock pool_lock#2 irq_context: 0 sb_writers#4 sb_writers#4 remove_cache_srcu &obj_hash[i].lock irq_context: 0 sb_writers#4 sb_writers#4 remove_cache_srcu &rq->__lock irq_context: 0 sb_writers#4 sb_writers#4 remove_cache_srcu &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 sb_writers#4 sb_writers#4 remove_cache_srcu pool_lock#2 irq_context: 0 &mm->mmap_lock fs_reclaim &cfs_rq->removed.lock irq_context: 0 &mm->mmap_lock fs_reclaim &obj_hash[i].lock irq_context: 0 raw_lock irq_context: softirq &fq->mq_flush_lock rcu_read_lock &pool->lock &p->pi_lock &rq->__lock &cfs_rq->removed.lock irq_context: 0 rtnl_mutex &rnp->exp_wq[3] irq_context: 0 (wq_completion)rcu_gp (work_completion)(&rew->rew_work) rcu_state.exp_wake_mutex &rnp->exp_wq[3] &p->pi_lock &cfs_rq->removed.lock irq_context: 0 sb_writers#4 &sb->s_type->i_mutex_key#8 &sbi->s_writepages_rwsem &journal->j_state_lock irq_context: 0 sb_writers#4 &sb->s_type->i_mutex_key#8 &sbi->s_writepages_rwsem &journal->j_state_lock tk_core.seq.seqcount irq_context: 0 sb_writers#4 &sb->s_type->i_mutex_key#8 &sbi->s_writepages_rwsem &journal->j_state_lock &obj_hash[i].lock irq_context: 0 sb_writers#4 &sb->s_type->i_mutex_key#8 &sbi->s_writepages_rwsem &journal->j_state_lock &base->lock irq_context: 0 sb_writers#4 &sb->s_type->i_mutex_key#8 &sbi->s_writepages_rwsem &journal->j_state_lock &base->lock &obj_hash[i].lock irq_context: 0 sb_writers#4 &sb->s_type->i_mutex_key#8 &sbi->s_writepages_rwsem jbd2_handle &ei->i_data_sem &ret->b_state_lock irq_context: 0 sb_writers#4 remove_cache_srcu &rq->__lock irq_context: 0 sb_writers#4 remove_cache_srcu &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 sb_writers#4 mapping.invalidate_lock &cfs_rq->removed.lock irq_context: softirq (&ndev->rs_timer) rcu_read_lock rcu_read_lock rcu_read_lock_bh rcu_read_lock rcu_read_lock rcu_read_lock rcu_read_lock pool_lock#2 irq_context: softirq (&ndev->rs_timer) rcu_read_lock rcu_read_lock rcu_read_lock_bh rcu_read_lock rcu_read_lock rcu_read_lock rcu_read_lock &dir->lock#2 irq_context: softirq (&ndev->rs_timer) rcu_read_lock rcu_read_lock rcu_read_lock_bh rcu_read_lock rcu_read_lock &____s->seqcount#2 irq_context: softirq (&ndev->rs_timer) rcu_read_lock rcu_read_lock rcu_read_lock_bh rcu_read_lock rcu_read_lock &____s->seqcount irq_context: softirq (&ndev->rs_timer) rcu_read_lock rcu_read_lock rcu_read_lock_bh rcu_read_lock rcu_read_lock &obj_hash[i].lock irq_context: softirq (&ndev->rs_timer) rcu_read_lock rcu_read_lock rcu_read_lock_bh rcu_read_lock rcu_read_lock rcu_read_lock pool_lock#2 irq_context: softirq (&ndev->rs_timer) rcu_read_lock rcu_read_lock rcu_read_lock_bh rcu_read_lock rcu_read_lock &tbl->lock irq_context: softirq (&ndev->rs_timer) rcu_read_lock rcu_read_lock rcu_read_lock_bh rcu_read_lock rcu_read_lock &n->lock irq_context: softirq (&ndev->rs_timer) rcu_read_lock rcu_read_lock rcu_read_lock_bh rcu_read_lock rcu_read_lock &n->lock &obj_hash[i].lock irq_context: softirq (&ndev->rs_timer) rcu_read_lock rcu_read_lock rcu_read_lock_bh rcu_read_lock rcu_read_lock &n->lock &base->lock irq_context: softirq (&ndev->rs_timer) rcu_read_lock rcu_read_lock rcu_read_lock_bh rcu_read_lock rcu_read_lock &n->lock &base->lock &obj_hash[i].lock irq_context: softirq (&ndev->rs_timer) rcu_read_lock rcu_read_lock rcu_read_lock_bh rcu_read_lock rcu_read_lock &n->lock pool_lock#2 irq_context: softirq (&ndev->rs_timer) rcu_read_lock rcu_read_lock rcu_read_lock_bh rcu_read_lock rcu_read_lock rcu_read_lock &ndev->lock irq_context: softirq (&ndev->rs_timer) rcu_read_lock rcu_read_lock rcu_read_lock_bh rcu_read_lock rcu_read_lock &ul->lock#2 irq_context: softirq (&ndev->rs_timer) rcu_read_lock rcu_read_lock rcu_read_lock_bh rcu_read_lock rcu_read_lock rcu_read_lock rcu_read_lock &obj_hash[i].lock irq_context: softirq (&ndev->rs_timer) rcu_read_lock rcu_read_lock rcu_read_lock_bh rcu_read_lock rcu_read_lock rcu_read_lock rcu_read_lock &tbl->lock irq_context: softirq (&ndev->rs_timer) rcu_read_lock rcu_read_lock rcu_read_lock_bh rcu_read_lock rcu_read_lock rcu_read_lock rcu_read_lock &n->lock irq_context: softirq (&ndev->rs_timer) rcu_read_lock rcu_read_lock rcu_read_lock_bh rcu_read_lock rcu_read_lock rcu_read_lock rcu_read_lock &____s->seqcount#9 irq_context: softirq (&ndev->rs_timer) rcu_read_lock rcu_read_lock rcu_read_lock_bh rcu_read_lock rcu_read_lock rcu_read_lock rcu_read_lock rcu_read_lock_bh &obj_hash[i].lock irq_context: softirq (&ndev->rs_timer) rcu_read_lock rcu_read_lock rcu_read_lock_bh rcu_read_lock rcu_read_lock rcu_read_lock rcu_read_lock rcu_read_lock_bh pool_lock#2 irq_context: softirq (&ndev->rs_timer) rcu_read_lock rcu_read_lock rcu_read_lock_bh rcu_read_lock rcu_read_lock rcu_read_lock rcu_read_lock rcu_read_lock_bh rcu_read_lock tk_core.seq.seqcount irq_context: softirq (&ndev->rs_timer) rcu_read_lock rcu_read_lock rcu_read_lock_bh rcu_read_lock rcu_read_lock rcu_read_lock rcu_read_lock rcu_read_lock_bh rcu_read_lock &list->lock#5 irq_context: 0 (wq_completion)events (work_completion)(&(&group->avgs_work)->work) &group->avgs_lock &rq->__lock irq_context: 0 (wq_completion)events (work_completion)(&(&group->avgs_work)->work) &group->avgs_lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 rtnl_mutex &idev->mc_lock rcu_read_lock pool_lock#2 irq_context: 0 (wq_completion)wg-crypt-wg0#6 (work_completion)(&peer->transmit_packet_work) &rq->__lock irq_context: 0 (wq_completion)wg-crypt-wg0#6 (work_completion)(&peer->transmit_packet_work) &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 sb_writers#8 tomoyo_ss &base->lock irq_context: 0 sb_writers#8 tomoyo_ss &base->lock &obj_hash[i].lock irq_context: 0 sb_writers#4 &sb->s_type->i_mutex_key#8 rcu_read_lock &____s->seqcount#2 irq_context: 0 sb_writers#4 &sb->s_type->i_mutex_key#8 rcu_read_lock &____s->seqcount irq_context: 0 sb_writers#4 &sb->s_type->i_mutex_key#8 rcu_read_lock &n->list_lock irq_context: 0 &sb->s_type->i_mutex_key#10 sk_lock-AF_INET k-sk_lock-AF_INET &n->list_lock irq_context: 0 &sb->s_type->i_mutex_key#10 sk_lock-AF_INET &obj_hash[i].lock pool_lock irq_context: 0 sb_writers#4 &type->i_mutex_dir_key#3/1 sb_internal &journal->j_state_lock irq_context: 0 sb_writers#4 &type->i_mutex_dir_key#3/1 sb_internal &journal->j_state_lock tk_core.seq.seqcount irq_context: 0 sb_writers#4 &type->i_mutex_dir_key#3/1 sb_internal &journal->j_state_lock &obj_hash[i].lock irq_context: 0 sb_writers#4 &type->i_mutex_dir_key#3/1 sb_internal &journal->j_state_lock &base->lock irq_context: 0 sb_writers#4 &type->i_mutex_dir_key#3/1 sb_internal &journal->j_state_lock &base->lock &obj_hash[i].lock irq_context: 0 sb_writers#4 &type->i_mutex_dir_key#3/1 sb_internal &rq->__lock irq_context: 0 sb_writers#4 &type->i_mutex_dir_key#3/1 sb_internal &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 sb_writers#4 &type->i_mutex_dir_key#3/1 sb_internal jbd2_handle bit_wait_table + i irq_context: 0 sb_writers#4 &type->i_mutex_dir_key#3/1 sb_internal jbd2_handle &sbi->s_orphan_lock &ret->b_state_lock irq_context: 0 &sb->s_type->i_mutex_key#10 raw_lock irq_context: 0 &sb->s_type->i_mutex_key#10 clock-AF_IEEE802154 irq_context: 0 &sb->s_type->i_mutex_key#10 rlock-AF_IEEE802154 irq_context: 0 sb_writers#4 mapping.invalidate_lock lock#4 rcu_read_lock &____s->seqcount irq_context: 0 sb_writers#4 mapping.invalidate_lock lock#4 rcu_read_lock pool_lock#2 irq_context: 0 sb_writers#4 mapping.invalidate_lock lock#4 &obj_hash[i].lock irq_context: 0 &sb->s_type->i_mutex_key#13 &rq->__lock irq_context: 0 &sb->s_type->i_mutex_key#13 &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 sb_writers#4 mapping.invalidate_lock mmu_notifier_invalidate_range_start &rq->__lock irq_context: 0 sb_writers#4 mapping.invalidate_lock mmu_notifier_invalidate_range_start &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 sb_writers#4 &sb->s_type->i_mutex_key#8 &journal->j_state_lock &journal->j_wait_commit &p->pi_lock &cfs_rq->removed.lock irq_context: 0 sb_writers#4 &sb->s_type->i_mutex_key#8 &rq->__lock &cfs_rq->removed.lock irq_context: 0 (wq_completion)events (work_completion)(&aux->work) &obj_hash[i].lock pool_lock irq_context: 0 (wq_completion)events (work_completion)(&aux->work) &base->lock irq_context: 0 (wq_completion)events (work_completion)(&aux->work) &base->lock &obj_hash[i].lock irq_context: 0 sb_writers#4 &sb->s_type->i_mutex_key#8 rcu_read_lock &rq->__lock &cfs_rq->removed.lock irq_context: 0 sb_writers#4 &type->i_mutex_dir_key#3/1 jbd2_handle rcu_read_lock &rcu_state.gp_wq irq_context: 0 sb_writers#4 &type->i_mutex_dir_key#3/1 jbd2_handle rcu_read_lock &rcu_state.gp_wq &p->pi_lock irq_context: 0 sb_writers#4 &type->i_mutex_dir_key#3/1 jbd2_handle rcu_read_lock &rcu_state.gp_wq &p->pi_lock &rq->__lock irq_context: 0 sb_writers#4 &type->i_mutex_dir_key#3/1 jbd2_handle rcu_read_lock &rcu_state.gp_wq &p->pi_lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 ppp_mutex remove_cache_srcu irq_context: 0 ppp_mutex remove_cache_srcu quarantine_lock irq_context: 0 ppp_mutex rtnl_mutex &sem->wait_lock irq_context: 0 ppp_mutex rtnl_mutex &p->pi_lock irq_context: 0 ppp_mutex rtnl_mutex &p->pi_lock &rq->__lock irq_context: 0 ppp_mutex rtnl_mutex &p->pi_lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 ppp_mutex rtnl_mutex &root->kernfs_rwsem &sem->wait_lock irq_context: 0 pernet_ops_rwsem &root->kernfs_rwsem &sem->wait_lock irq_context: 0 pernet_ops_rwsem &root->kernfs_rwsem &rq->__lock irq_context: 0 pernet_ops_rwsem &root->kernfs_rwsem &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 ppp_mutex rtnl_mutex &root->kernfs_rwsem &p->pi_lock irq_context: 0 ppp_mutex rtnl_mutex lock kernfs_idr_lock &c->lock irq_context: 0 ppp_mutex rtnl_mutex &root->kernfs_rwsem &sem->wait_lock irq_context: 0 hashlimit_mutex irq_context: 0 hashlimit_mutex fs_reclaim irq_context: 0 hashlimit_mutex fs_reclaim mmu_notifier_invalidate_range_start irq_context: 0 hashlimit_mutex fs_reclaim mmu_notifier_invalidate_range_start &rq->__lock irq_context: 0 hashlimit_mutex fs_reclaim mmu_notifier_invalidate_range_start &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 ppp_mutex rtnl_mutex batched_entropy_u8.lock irq_context: 0 ppp_mutex rtnl_mutex kfence_freelist_lock irq_context: 0 ppp_mutex rtnl_mutex &idev->mc_lock &n->list_lock irq_context: 0 ppp_mutex rtnl_mutex &idev->mc_lock &n->list_lock &c->lock irq_context: 0 pernet_ops_rwsem &sem->wait_lock irq_context: 0 hashlimit_mutex pool_lock#2 irq_context: 0 hashlimit_mutex free_vmap_area_lock irq_context: 0 hashlimit_mutex vmap_area_lock irq_context: 0 hashlimit_mutex &____s->seqcount irq_context: 0 hashlimit_mutex init_mm.page_table_lock irq_context: 0 hashlimit_mutex &c->lock irq_context: 0 hashlimit_mutex proc_subdir_lock irq_context: 0 hashlimit_mutex proc_inum_ida.xa_lock irq_context: 0 hashlimit_mutex proc_subdir_lock irq_context: 0 hashlimit_mutex &obj_hash[i].lock irq_context: 0 hashlimit_mutex &base->lock irq_context: 0 hashlimit_mutex &base->lock &obj_hash[i].lock irq_context: 0 hashlimit_mutex &rq->__lock irq_context: 0 hashlimit_mutex &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 hashlimit_mutex &ent->pde_unload_lock irq_context: 0 (work_completion)(&(&hinfo->gc_work)->work) irq_context: 0 &hinfo->lock irq_context: 0 &group->mark_mutex fs_reclaim mmu_notifier_invalidate_range_start &rq->__lock irq_context: 0 &group->mark_mutex fs_reclaim mmu_notifier_invalidate_range_start &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 &f->f_pos_lock sb_writers#4 &sb->s_type->i_mutex_key#8 jbd2_handle rcu_node_0 irq_context: 0 tracepoints_mutex &c->lock irq_context: 0 sk_lock-AF_INET rcu_read_lock &sighand->siglock irq_context: 0 sk_lock-AF_INET &sighand->siglock irq_context: 0 sk_lock-AF_INET &sighand->siglock stock_lock irq_context: 0 sk_lock-AF_INET &sighand->siglock pool_lock#2 irq_context: 0 sk_lock-AF_INET &sighand->siglock &p->pi_lock irq_context: 0 sk_lock-AF_INET &sighand->siglock &p->pi_lock &rq->__lock irq_context: 0 sk_lock-AF_INET &sighand->siglock &p->pi_lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 sb_writers#4 &type->i_mutex_dir_key#3/1 tomoyo_ss &cfs_rq->removed.lock irq_context: 0 sb_writers#4 &type->i_mutex_dir_key#3/1 rcu_read_lock &rcu_state.expedited_wq irq_context: 0 sb_writers#4 &type->i_mutex_dir_key#3/1 rcu_read_lock &rcu_state.expedited_wq &p->pi_lock irq_context: 0 sb_writers#4 &type->i_mutex_dir_key#3/1 rcu_read_lock &rcu_state.expedited_wq &p->pi_lock &rq->__lock irq_context: 0 sb_writers#4 &type->i_mutex_dir_key#3/1 rcu_read_lock &rcu_state.expedited_wq &p->pi_lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 &f->f_pos_lock &type->i_mutex_dir_key#4 &root->kernfs_rwsem &p->pi_lock irq_context: 0 &f->f_pos_lock &type->i_mutex_dir_key#4 &root->kernfs_rwsem &p->pi_lock &rq->__lock irq_context: 0 &f->f_pos_lock &type->i_mutex_dir_key#4 &root->kernfs_rwsem &p->pi_lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 rtnl_mutex gdp_mutex &c->lock irq_context: 0 sb_writers#4 &type->i_mutex_dir_key#3/1 sb_internal jbd2_handle key#4 irq_context: 0 pernet_ops_rwsem rtnl_mutex rcu_read_lock rcu_read_lock rcu_node_0 irq_context: 0 pernet_ops_rwsem rtnl_mutex rcu_read_lock rcu_read_lock &rq->__lock irq_context: 0 pernet_ops_rwsem rtnl_mutex rcu_read_lock rcu_read_lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)bat_events (work_completion)(&(&bat_priv->nc.work)->work) rcu_read_lock &rq->__lock &cfs_rq->removed.lock irq_context: 0 &root->kernfs_rwsem &p->pi_lock &rq->__lock irq_context: 0 &root->kernfs_rwsem &p->pi_lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 sb_writers#5 &type->i_mutex_dir_key#5/1 &sb->s_type->i_mutex_key#12 &sb->s_type->i_mutex_key#12/4 &simple_offset_xa_lock &c->lock irq_context: 0 sk_lock-AF_INET &cfs_rq->removed.lock irq_context: 0 sb_writers#4 &type->i_mutex_dir_key#3/1 tomoyo_ss remove_cache_srcu &rq->__lock irq_context: 0 sb_writers#4 &type->i_mutex_dir_key#3/1 tomoyo_ss remove_cache_srcu &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 &f->f_pos_lock sb_writers#4 &sb->s_type->i_mutex_key#8 rcu_read_lock &cfs_rq->removed.lock irq_context: 0 pernet_ops_rwsem nf_ct_proto_mutex defrag4_mutex nf_hook_mutex rcu_read_lock rcu_read_lock rcu_node_0 irq_context: 0 pernet_ops_rwsem nf_ct_proto_mutex defrag4_mutex nf_hook_mutex rcu_read_lock rcu_read_lock &rq->__lock irq_context: 0 pernet_ops_rwsem nf_ct_proto_mutex defrag4_mutex nf_hook_mutex rcu_read_lock rcu_read_lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 pernet_ops_rwsem nf_ct_proto_mutex defrag4_mutex nf_hook_mutex rcu_read_lock rcu_node_0 irq_context: 0 lock pidmap_lock &pcp->lock &zone->lock &____s->seqcount irq_context: 0 (wq_completion)gid-cache-wq (work_completion)(&ndev_work->work) &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 rtnl_mutex rcu_read_lock rcu_read_lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 sk_lock-AF_INET6 &dccp_hashinfo.bhash[i].lock irq_context: 0 &sb->s_type->i_mutex_key#10 sk_lock-AF_INET6 &h->lhash2[i].lock irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem rcu_state.exp_mutex &rnp->exp_wq[2] irq_context: softirq &(&local->scan_work)->timer rcu_read_lock &pool->lock/1 &p->pi_lock &cfs_rq->removed.lock irq_context: 0 &sb->s_type->i_mutex_key#8 &sbi->s_writepages_rwsem rcu_node_0 irq_context: 0 sb_writers#5 &fsnotify_mark_srcu &rq->__lock irq_context: 0 sb_writers#5 &fsnotify_mark_srcu &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 sk_lock-AF_INET6 &dccp_hashinfo.bhash[i].lock &dccp_hashinfo.bhash2[i].lock irq_context: 0 sk_lock-AF_INET6 &dccp_hashinfo.bhash[i].lock &dccp_hashinfo.bhash2[i].lock stock_lock irq_context: 0 sk_lock-AF_INET6 &dccp_hashinfo.bhash[i].lock &dccp_hashinfo.bhash2[i].lock &____s->seqcount#2 irq_context: 0 sk_lock-AF_INET6 &dccp_hashinfo.bhash[i].lock &dccp_hashinfo.bhash2[i].lock &____s->seqcount irq_context: 0 sk_lock-AF_INET6 &dccp_hashinfo.bhash[i].lock &dccp_hashinfo.bhash2[i].lock pool_lock#2 irq_context: 0 sk_lock-AF_INET6 &dccp_hashinfo.bhash[i].lock &dccp_hashinfo.bhash2[i].lock &c->lock irq_context: 0 sk_lock-AF_INET6 &dccp_hashinfo.bhash[i].lock &dccp_hashinfo.bhash2[i].lock clock-AF_INET6 irq_context: 0 sk_lock-AF_INET &dccp_hashinfo.bhash[i].lock irq_context: 0 sk_lock-AF_INET &dccp_hashinfo.bhash[i].lock stock_lock irq_context: 0 sk_lock-AF_INET &dccp_hashinfo.bhash[i].lock pool_lock#2 irq_context: 0 sk_lock-AF_INET &dccp_hashinfo.bhash[i].lock &dccp_hashinfo.bhash2[i].lock irq_context: 0 sk_lock-AF_INET &dccp_hashinfo.bhash[i].lock &dccp_hashinfo.bhash2[i].lock stock_lock irq_context: 0 sk_lock-AF_INET &dccp_hashinfo.bhash[i].lock &dccp_hashinfo.bhash2[i].lock pool_lock#2 irq_context: 0 sk_lock-AF_INET &dccp_hashinfo.bhash[i].lock &dccp_hashinfo.bhash2[i].lock batched_entropy_u8.lock irq_context: 0 sk_lock-AF_INET &dccp_hashinfo.bhash[i].lock &dccp_hashinfo.bhash2[i].lock &hashinfo->ehash_locks[i] irq_context: 0 sk_lock-AF_INET rcu_read_lock rcu_read_lock rcu_read_lock &____s->seqcount#7 irq_context: 0 sk_lock-AF_INET rcu_read_lock rcu_read_lock &nf_nat_locks[i] irq_context: 0 sk_lock-AF_INET rcu_read_lock rcu_read_lock &nf_conntrack_locks[i]/1 irq_context: 0 sk_lock-AF_INET rcu_read_lock rcu_read_lock rcu_read_lock_bh tk_core.seq.seqcount irq_context: 0 sk_lock-AF_INET rcu_read_lock rcu_read_lock rcu_read_lock_bh &list->lock#5 irq_context: softirq rcu_read_lock rcu_read_lock k-slock-AF_INET irq_context: softirq rcu_read_lock rcu_read_lock k-slock-AF_INET rcu_read_lock &____s->seqcount#7 irq_context: softirq rcu_read_lock rcu_read_lock k-slock-AF_INET rcu_read_lock &nf_conntrack_locks[i] irq_context: softirq rcu_read_lock rcu_read_lock k-slock-AF_INET rcu_read_lock &nf_conntrack_locks[i] &nf_conntrack_locks[i]/1 irq_context: softirq rcu_read_lock rcu_read_lock k-slock-AF_INET rcu_read_lock &nf_conntrack_locks[i]/1 irq_context: 0 sk_lock-AF_INET &____s->seqcount#2 irq_context: 0 sk_lock-AF_INET &hashinfo->ehash_locks[i] irq_context: 0 sb_writers#8 tomoyo_ss remove_cache_srcu irq_context: 0 sb_writers#8 tomoyo_ss remove_cache_srcu &rq->__lock irq_context: 0 sb_writers#8 tomoyo_ss remove_cache_srcu &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 sb_writers#8 tomoyo_ss remove_cache_srcu quarantine_lock irq_context: 0 (wq_completion)events (work_completion)(&pwq->unbound_release_work) &rnp->exp_wq[3] irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem &rnp->exp_wq[3] irq_context: 0 (wq_completion)ext4-rsv-conversion (work_completion)(&ei->i_rsv_conversion_work) rcu_node_0 irq_context: 0 (wq_completion)ext4-rsv-conversion (work_completion)(&ei->i_rsv_conversion_work) &rcu_state.expedited_wq irq_context: 0 (wq_completion)ext4-rsv-conversion (work_completion)(&ei->i_rsv_conversion_work) &rcu_state.expedited_wq &p->pi_lock irq_context: 0 (wq_completion)ext4-rsv-conversion (work_completion)(&ei->i_rsv_conversion_work) &rcu_state.expedited_wq &p->pi_lock &rq->__lock irq_context: 0 (wq_completion)ext4-rsv-conversion (work_completion)(&ei->i_rsv_conversion_work) &rcu_state.expedited_wq &p->pi_lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 kn->active#5 rcu_read_lock &rcu_state.expedited_wq irq_context: 0 kn->active#5 rcu_read_lock &rcu_state.expedited_wq &p->pi_lock irq_context: 0 kn->active#5 rcu_read_lock &rcu_state.expedited_wq &p->pi_lock &rq->__lock irq_context: 0 kn->active#5 rcu_read_lock &rcu_state.expedited_wq &p->pi_lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 &sb->s_type->i_mutex_key#10 sk_lock-AF_INET6 &dccp_hashinfo.bhash[i].lock irq_context: 0 &sb->s_type->i_mutex_key#10 sk_lock-AF_INET6 &dccp_hashinfo.bhash[i].lock stock_lock irq_context: 0 &sb->s_type->i_mutex_key#10 sk_lock-AF_INET6 &dccp_hashinfo.bhash[i].lock &obj_hash[i].lock irq_context: 0 &sb->s_type->i_mutex_key#10 sk_lock-AF_INET6 &dccp_hashinfo.bhash[i].lock pool_lock#2 irq_context: 0 &sb->s_type->i_mutex_key#10 sk_lock-AF_INET6 &dccp_hashinfo.bhash[i].lock &dccp_hashinfo.bhash2[i].lock irq_context: 0 &sb->s_type->i_mutex_key#10 sk_lock-AF_INET6 &dccp_hashinfo.bhash[i].lock &dccp_hashinfo.bhash2[i].lock stock_lock irq_context: 0 &sb->s_type->i_mutex_key#10 sk_lock-AF_INET6 &dccp_hashinfo.bhash[i].lock &dccp_hashinfo.bhash2[i].lock &obj_hash[i].lock irq_context: 0 &sb->s_type->i_mutex_key#10 sk_lock-AF_INET6 &dccp_hashinfo.bhash[i].lock &dccp_hashinfo.bhash2[i].lock pool_lock#2 irq_context: 0 &sb->s_type->i_mutex_key#10 sk_lock-AF_INET6 &queue->rskq_lock irq_context: 0 &sb->s_type->i_mutex_key#10 slock-AF_INET6 &obj_hash[i].lock irq_context: 0 &sb->s_type->i_mutex_key#10 slock-AF_INET6 pool_lock#2 irq_context: 0 &sb->s_type->i_mutex_key#10 slock-AF_INET6 elock-AF_INET6 irq_context: 0 &sb->s_type->i_mutex_key#10 slock-AF_INET &obj_hash[i].lock irq_context: 0 &sb->s_type->i_mutex_key#10 slock-AF_INET pool_lock#2 irq_context: 0 &sb->s_type->i_mutex_key#10 slock-AF_INET elock-AF_INET irq_context: 0 tomoyo_ss rcu_read_lock &rcu_state.expedited_wq &p->pi_lock &rq->__lock &cfs_rq->removed.lock irq_context: 0 sb_writers#4 &type->i_mutex_dir_key#3/1 jbd2_handle &rcu_state.expedited_wq irq_context: 0 sb_writers#4 &type->i_mutex_dir_key#3/1 jbd2_handle &rcu_state.expedited_wq &p->pi_lock irq_context: 0 sb_writers#4 &type->i_mutex_dir_key#3/1 jbd2_handle &rcu_state.expedited_wq &p->pi_lock &rq->__lock irq_context: 0 sb_writers#4 &type->i_mutex_dir_key#3/1 jbd2_handle &rcu_state.expedited_wq &p->pi_lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 sb_writers#8 remove_cache_srcu pool_lock#2 irq_context: 0 &p->lock &rq->__lock &cfs_rq->removed.lock irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem rtnl_mutex &root->kernfs_rwsem &sem->wait_lock irq_context: 0 sb_writers#4 &type->i_mutex_dir_key#3/1 tomoyo_ss rcu_read_lock &rcu_state.gp_wq irq_context: 0 sb_writers#4 &type->i_mutex_dir_key#3/1 tomoyo_ss rcu_read_lock &rcu_state.gp_wq &p->pi_lock irq_context: 0 sb_writers#4 &type->i_mutex_dir_key#3/1 tomoyo_ss rcu_read_lock &rcu_state.gp_wq &p->pi_lock &rq->__lock irq_context: 0 sb_writers#4 &type->i_mutex_dir_key#3/1 tomoyo_ss rcu_read_lock &rcu_state.gp_wq &p->pi_lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 pernet_ops_rwsem nl_table_wait.lock &p->pi_lock irq_context: 0 pernet_ops_rwsem nl_table_wait.lock &p->pi_lock &rq->__lock irq_context: 0 pernet_ops_rwsem nl_table_wait.lock &p->pi_lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 sb_writers#8 tomoyo_ss rcu_read_lock &rcu_state.gp_wq irq_context: 0 sb_writers#8 tomoyo_ss rcu_read_lock &rcu_state.gp_wq &p->pi_lock irq_context: 0 sb_writers#8 tomoyo_ss rcu_read_lock &rcu_state.gp_wq &p->pi_lock &rq->__lock irq_context: 0 sb_writers#8 tomoyo_ss rcu_read_lock &rcu_state.gp_wq &p->pi_lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 sb_writers#4 &type->i_mutex_dir_key#3/1 &sb->s_type->i_mutex_key#8 &n->list_lock irq_context: 0 sb_writers#4 &type->i_mutex_dir_key#3/1 &sb->s_type->i_mutex_key#8 &n->list_lock &c->lock irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem quarantine_lock irq_context: 0 (wq_completion)events_unbound connector_reaper_work pool_lock irq_context: 0 &kernfs_locks->open_file_mutex[count] rcu_node_0 irq_context: 0 &kernfs_locks->open_file_mutex[count] &rcu_state.expedited_wq irq_context: 0 &kernfs_locks->open_file_mutex[count] &rcu_state.expedited_wq &p->pi_lock irq_context: 0 &kernfs_locks->open_file_mutex[count] &rcu_state.expedited_wq &p->pi_lock &rq->__lock irq_context: 0 &kernfs_locks->open_file_mutex[count] &rcu_state.expedited_wq &p->pi_lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 pernet_ops_rwsem rtnl_mutex rcu_read_lock &rcu_state.gp_wq irq_context: 0 pernet_ops_rwsem rtnl_mutex rcu_read_lock &rcu_state.gp_wq &p->pi_lock irq_context: 0 pernet_ops_rwsem rtnl_mutex rcu_read_lock &rcu_state.gp_wq &p->pi_lock &rq->__lock irq_context: 0 pernet_ops_rwsem rtnl_mutex rcu_read_lock &rcu_state.gp_wq &p->pi_lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: softirq (&n->timer) &n->lock pool_lock#2 irq_context: softirq (&n->timer) rcu_read_lock &ndev->lock irq_context: softirq (&n->timer) pool_lock#2 irq_context: softirq (&n->timer) &dir->lock#2 irq_context: softirq (&n->timer) &ul->lock#2 irq_context: softirq (&n->timer) rcu_read_lock rcu_read_lock &____s->seqcount#10 irq_context: softirq (&n->timer) rcu_read_lock rcu_read_lock rcu_read_lock_bh &obj_hash[i].lock irq_context: softirq (&n->timer) rcu_read_lock rcu_read_lock rcu_read_lock_bh pool_lock#2 irq_context: softirq (&n->timer) rcu_read_lock rcu_read_lock rcu_read_lock_bh rcu_read_lock tk_core.seq.seqcount irq_context: softirq (&n->timer) rcu_read_lock rcu_read_lock rcu_read_lock_bh rcu_read_lock &list->lock#5 irq_context: softirq (&n->timer) &obj_hash[i].lock irq_context: 0 pernet_ops_rwsem k-sk_lock-AF_TIPC &rq->__lock irq_context: 0 pernet_ops_rwsem k-sk_lock-AF_TIPC &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem rtnl_mutex rcu_state.exp_mutex rcu_read_lock &pool->lock &p->pi_lock &rq->__lock &cfs_rq->removed.lock irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem rtnl_mutex sysctl_lock &obj_hash[i].lock pool_lock irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem rtnl_mutex rcu_state.exp_mutex rcu_read_lock &rq->__lock irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem rtnl_mutex rcu_state.exp_mutex rcu_read_lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: softirq rcu_read_lock hwsim_radio_lock batched_entropy_u8.lock irq_context: softirq rcu_read_lock hwsim_radio_lock kfence_freelist_lock irq_context: 0 (wq_completion)events (work_completion)(&rdev->wiphy_work) &rdev->wiphy.mtx &meta->lock irq_context: 0 (wq_completion)events (work_completion)(&rdev->wiphy_work) &rdev->wiphy.mtx kfence_freelist_lock irq_context: 0 (wq_completion)events (work_completion)(&w->w) nfc_devlist_mutex &rq->__lock irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem rtnl_mutex &nft_net->commit_mutex &rq->__lock irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem rtnl_mutex net_rwsem &rq->__lock irq_context: 0 dup_mmap_sem &mm->mmap_lock &mm->mmap_lock/1 &rcu_state.expedited_wq &p->pi_lock &rq->__lock irq_context: 0 dup_mmap_sem &mm->mmap_lock &mm->mmap_lock/1 &rcu_state.expedited_wq &p->pi_lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 sb_writers#4 sb_internal remove_cache_srcu irq_context: 0 sb_writers#4 sb_internal remove_cache_srcu &rq->__lock irq_context: 0 sb_writers#4 sb_internal remove_cache_srcu &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 sb_writers#4 sb_internal remove_cache_srcu quarantine_lock irq_context: 0 sb_writers#4 sb_internal remove_cache_srcu &c->lock irq_context: 0 sb_writers#4 sb_internal remove_cache_srcu &n->list_lock irq_context: 0 sb_writers#4 sb_internal remove_cache_srcu rcu_read_lock pool_lock#2 irq_context: 0 sb_writers#4 sb_internal remove_cache_srcu &obj_hash[i].lock irq_context: 0 (wq_completion)bat_events (work_completion)(&(&forw_packet_aggr->delayed_work)->work) &hard_iface->bat_iv.ogm_buff_mutex &obj_hash[i].lock pool_lock irq_context: 0 sb_writers#5 &type->i_mutex_dir_key#5/1 &sb->s_type->i_mutex_key#12 &sb->s_type->i_mutex_key#12/4 &simple_offset_xa_lock &obj_hash[i].lock pool_lock irq_context: 0 sb_writers#5 fill_pool_map-wait-type-override rcu_read_lock rcu_node_0 irq_context: 0 sb_writers#5 fill_pool_map-wait-type-override rcu_read_lock &rq->__lock irq_context: 0 sb_writers#5 fill_pool_map-wait-type-override rcu_read_lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 &f->f_pos_lock sb_writers#4 &sb->s_type->i_mutex_key#8 &ei->i_es_lock key#5 irq_context: 0 (wq_completion)bat_events (work_completion)(&(&forw_packet_aggr->delayed_work)->work) &hard_iface->bat_iv.ogm_buff_mutex rcu_read_lock &cfs_rq->removed.lock irq_context: 0 pernet_ops_rwsem &rcu_state.expedited_wq irq_context: 0 pernet_ops_rwsem rtnl_mutex bus_type_sem &rq->__lock irq_context: 0 pernet_ops_rwsem rtnl_mutex bus_type_sem &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 &sb->s_type->i_mutex_key#8 &cfs_rq->removed.lock irq_context: 0 &mm->mmap_lock &rcu_state.expedited_wq irq_context: 0 &f->f_pos_lock &lock->wait_lock irq_context: 0 sb_writers#4 &sb->s_type->i_mutex_key#8 mapping.invalidate_lock &mapping->i_mmap_rwsem &lruvec->lru_lock irq_context: 0 sb_writers#4 &sb->s_type->i_mutex_key#8 mapping.invalidate_lock &mapping->private_lock rcu_read_lock &p->pi_lock irq_context: 0 sb_writers#4 &sb->s_type->i_mutex_key#8 mapping.invalidate_lock &mapping->private_lock rcu_read_lock &p->pi_lock &rq->__lock irq_context: 0 sb_writers#4 &sb->s_type->i_mutex_key#8 mapping.invalidate_lock &mapping->private_lock rcu_read_lock &p->pi_lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 sb_writers#5 &type->i_mutex_dir_key#5/1 &sb->s_type->i_mutex_key#12 &simple_offset_xa_lock fill_pool_map-wait-type-override pool_lock#2 irq_context: 0 sb_writers#5 &type->i_mutex_dir_key#5/1 &sb->s_type->i_mutex_key#12 &simple_offset_xa_lock fill_pool_map-wait-type-override &c->lock irq_context: 0 sb_writers#5 &type->i_mutex_dir_key#5/1 &sb->s_type->i_mutex_key#12 &simple_offset_xa_lock fill_pool_map-wait-type-override &____s->seqcount#2 irq_context: 0 sb_writers#5 &type->i_mutex_dir_key#5/1 &sb->s_type->i_mutex_key#12 &simple_offset_xa_lock fill_pool_map-wait-type-override &____s->seqcount irq_context: 0 sb_writers#5 &type->i_mutex_dir_key#5/1 &sb->s_type->i_mutex_key#12 &simple_offset_xa_lock fill_pool_map-wait-type-override pool_lock irq_context: 0 pernet_ops_rwsem rtnl_mutex rcu_read_lock &pool->lock/1 &p->pi_lock &cfs_rq->removed.lock irq_context: 0 sb_writers#8 tomoyo_ss quarantine_lock irq_context: 0 &f->f_pos_lock &cfs_rq->removed.lock irq_context: 0 &f->f_pos_lock &obj_hash[i].lock irq_context: 0 &f->f_pos_lock pool_lock#2 irq_context: 0 sb_writers#4 &sb->s_type->i_mutex_key#8 mapping.invalidate_lock rcu_read_lock &cfs_rq->removed.lock irq_context: 0 sb_writers#4 &sb->s_type->i_mutex_key#8 mapping.invalidate_lock rcu_read_lock &obj_hash[i].lock irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem rtnl_mutex rcu_state.exp_mutex &obj_hash[i].lock pool_lock irq_context: 0 remove_cache_srcu rcu_read_lock &rcu_state.expedited_wq irq_context: 0 remove_cache_srcu rcu_read_lock &rcu_state.expedited_wq &p->pi_lock irq_context: 0 remove_cache_srcu rcu_read_lock &rcu_state.expedited_wq &p->pi_lock &rq->__lock irq_context: 0 remove_cache_srcu rcu_read_lock &rcu_state.expedited_wq &p->pi_lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 kn->active#5 fs_reclaim mmu_notifier_invalidate_range_start &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)bat_events (work_completion)(&(&bat_priv->nc.work)->work) fill_pool_map-wait-type-override &____s->seqcount#2 irq_context: 0 tasklist_lock &sighand->siglock &(&sig->stats_lock)->lock &____s->seqcount#5 pidmap_lock &obj_hash[i].lock pool_lock irq_context: 0 dup_mmap_sem &mm->mmap_lock &mm->mmap_lock/1 &anon_vma->rwsem &____s->seqcount#2 irq_context: 0 dup_mmap_sem &mm->mmap_lock &mm->mmap_lock/1 &anon_vma->rwsem pool_lock#2 irq_context: 0 sb_writers#8 &of->mutex kn->active#5 fs_reclaim &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 sb_writers#8 &of->mutex kn->active#5 uevent_sock_mutex fs_reclaim &rq->__lock irq_context: 0 sb_writers#8 &of->mutex kn->active#5 uevent_sock_mutex fs_reclaim &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 pernet_ops_rwsem cpu_hotplug_lock wq_pool_mutex &n->list_lock irq_context: 0 pernet_ops_rwsem cpu_hotplug_lock wq_pool_mutex &n->list_lock &c->lock irq_context: 0 pernet_ops_rwsem cpu_hotplug_lock wq_pool_mutex &rq->__lock irq_context: 0 pernet_ops_rwsem cpu_hotplug_lock wq_pool_mutex &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: softirq rcu_read_lock &rdev->bss_lock &obj_hash[i].lock pool_lock irq_context: 0 pernet_ops_rwsem stock_lock rcu_read_lock per_cpu_ptr(&cgroup_rstat_cpu_lock, cpu) irq_context: 0 pernet_ops_rwsem k-sk_lock-AF_TIPC &tn->nametbl_lock &service->lock &c->lock irq_context: 0 pernet_ops_rwsem nf_ct_proto_mutex defrag4_mutex nf_hook_mutex &____s->seqcount#2 irq_context: 0 pernet_ops_rwsem nf_ct_proto_mutex nf_hook_mutex &____s->seqcount#2 irq_context: 0 pernet_ops_rwsem k-sk_lock-AF_RXRPC &rxnet->local_mutex &obj_hash[i].lock pool_lock irq_context: 0 sb_writers#4 sb_internal &n->list_lock irq_context: 0 sb_writers#4 sb_internal &n->list_lock &c->lock irq_context: 0 &f->f_pos_lock sb_writers#4 &sb->s_type->i_mutex_key#8 &mm->mmap_lock rcu_read_lock per_cpu_ptr(&cgroup_rstat_cpu_lock, cpu) irq_context: 0 &sb->s_type->i_mutex_key#10 sk_lock-AF_INET6 slock-AF_INET6 fill_pool_map-wait-type-override pool_lock#2 irq_context: 0 &sb->s_type->i_mutex_key#10 sk_lock-AF_INET6 slock-AF_INET6 fill_pool_map-wait-type-override pool_lock irq_context: 0 sb_writers#4 &type->i_mutex_dir_key#3/1 fill_pool_map-wait-type-override pool_lock#2 irq_context: 0 sb_writers#4 &type->i_mutex_dir_key#3/1 fill_pool_map-wait-type-override &rq->__lock irq_context: 0 sb_writers#4 &type->i_mutex_dir_key#3/1 fill_pool_map-wait-type-override &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 sb_writers#4 &type->i_mutex_dir_key#3/1 fill_pool_map-wait-type-override pool_lock irq_context: 0 sb_writers#4 &type->i_mutex_dir_key#3/1 &type->i_mutex_dir_key#3 fill_pool_map-wait-type-override pool_lock#2 irq_context: 0 sb_writers#4 &type->i_mutex_dir_key#3/1 &type->i_mutex_dir_key#3 fill_pool_map-wait-type-override &c->lock irq_context: 0 sb_writers#4 &type->i_mutex_dir_key#3/1 &type->i_mutex_dir_key#3 fill_pool_map-wait-type-override &n->list_lock irq_context: 0 sb_writers#4 &type->i_mutex_dir_key#3/1 &type->i_mutex_dir_key#3 fill_pool_map-wait-type-override &n->list_lock &c->lock irq_context: 0 &kernfs_locks->open_file_mutex[count] fill_pool_map-wait-type-override &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 sb_writers#4 &type->i_mutex_dir_key#3/1 &type->i_mutex_dir_key#3 fill_pool_map-wait-type-override pool_lock irq_context: 0 sb_writers#4 &type->i_mutex_dir_key#3/1 &type->i_mutex_dir_key#3 fill_pool_map-wait-type-override &rq->__lock irq_context: 0 sb_writers#4 &type->i_mutex_dir_key#3/1 &type->i_mutex_dir_key#3 fill_pool_map-wait-type-override &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem rcu_read_lock &cfs_rq->removed.lock irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem rcu_read_lock &obj_hash[i].lock irq_context: 0 &sb->s_type->i_mutex_key#10 sk_lock-AF_X25 irq_context: 0 &sb->s_type->i_mutex_key#10 sk_lock-AF_X25 slock-AF_X25 irq_context: 0 &sb->s_type->i_mutex_key#10 sk_lock-AF_X25 wlock-AF_X25 irq_context: 0 &sb->s_type->i_mutex_key#10 sk_lock-AF_X25 &list->lock#27 irq_context: 0 &sb->s_type->i_mutex_key#10 sk_lock-AF_X25 &rq->__lock irq_context: 0 &sb->s_type->i_mutex_key#10 sk_lock-AF_X25 &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 rtnl_mutex &br->hash_lock rcu_read_lock pool_lock#2 irq_context: 0 rtnl_mutex &bridge_netdev_addr_lock_key irq_context: 0 rtnl_mutex &bridge_netdev_addr_lock_key pool_lock#2 irq_context: 0 &sb->s_type->i_mutex_key#10 sk_lock-AF_X25 &obj_hash[i].lock irq_context: 0 &sb->s_type->i_mutex_key#10 sk_lock-AF_X25 x25_list_lock irq_context: 0 &sb->s_type->i_mutex_key#10 sk_lock-AF_X25 rlock-AF_X25 irq_context: 0 &sb->s_type->i_mutex_key#10 slock-AF_X25 irq_context: 0 sb_writers#5 &cfs_rq->removed.lock irq_context: softirq (&in_dev->mr_ifc_timer) rcu_read_lock rcu_read_lock &____s->seqcount#7 irq_context: softirq (&in_dev->mr_ifc_timer) rcu_read_lock &nf_nat_locks[i] irq_context: softirq (&in_dev->mr_ifc_timer) rcu_read_lock rcu_read_lock_bh rcu_read_lock &obj_hash[i].lock irq_context: 0 (wq_completion)mld (work_completion)(&(&idev->mc_ifc_work)->work) &idev->mc_lock rcu_read_lock rcu_read_lock &tbl->lock &n->lock irq_context: 0 (wq_completion)mld (work_completion)(&(&idev->mc_ifc_work)->work) &idev->mc_lock rcu_read_lock rcu_read_lock &tbl->lock pool_lock#2 irq_context: 0 (wq_completion)mld (work_completion)(&(&idev->mc_ifc_work)->work) &idev->mc_lock rcu_read_lock rcu_read_lock &tbl->lock &n->lock irq_context: 0 (wq_completion)mld (work_completion)(&(&idev->mc_ifc_work)->work) &idev->mc_lock rcu_read_lock rcu_read_lock &tbl->lock &n->lock &____s->seqcount#9 irq_context: 0 (wq_completion)mld (work_completion)(&(&idev->mc_ifc_work)->work) &idev->mc_lock rcu_read_lock rcu_read_lock &tbl->lock nl_table_lock irq_context: 0 (wq_completion)mld (work_completion)(&(&idev->mc_ifc_work)->work) &idev->mc_lock rcu_read_lock rcu_read_lock &tbl->lock &obj_hash[i].lock irq_context: 0 (wq_completion)mld (work_completion)(&(&idev->mc_ifc_work)->work) &idev->mc_lock rcu_read_lock rcu_read_lock &tbl->lock nl_table_wait.lock irq_context: 0 (wq_completion)mld (work_completion)(&(&idev->mc_ifc_work)->work) &idev->mc_lock rcu_read_lock rcu_read_lock &tbl->lock rcu_read_lock lock#8 irq_context: 0 (wq_completion)mld (work_completion)(&(&idev->mc_ifc_work)->work) &idev->mc_lock rcu_read_lock rcu_read_lock &tbl->lock rcu_read_lock id_table_lock irq_context: 0 (wq_completion)mld (work_completion)(&(&idev->mc_ifc_work)->work) &idev->mc_lock rcu_read_lock rcu_read_lock &tbl->lock &dir->lock#2 irq_context: 0 (wq_completion)mld (work_completion)(&(&idev->mc_ifc_work)->work) &idev->mc_lock rcu_read_lock rcu_read_lock &tbl->lock krc.lock irq_context: 0 (wq_completion)mld (work_completion)(&(&idev->mc_ifc_work)->work) &idev->mc_lock rcu_read_lock rcu_read_lock &tbl->lock &obj_hash[i].lock pool_lock irq_context: 0 (wq_completion)mld (work_completion)(&(&idev->mc_ifc_work)->work) &idev->mc_lock rcu_read_lock rcu_read_lock &tbl->lock &c->lock irq_context: 0 (wq_completion)mld (work_completion)(&(&idev->mc_ifc_work)->work) &idev->mc_lock rcu_read_lock rcu_read_lock rcu_read_lock_bh rcu_read_lock &____s->seqcount irq_context: 0 (wq_completion)ipv6_addrconf (work_completion)(&(&ifa->dad_work)->work) rtnl_mutex &idev->mc_lock &bridge_netdev_addr_lock_key irq_context: 0 (wq_completion)ipv6_addrconf (work_completion)(&(&ifa->dad_work)->work) rtnl_mutex &idev->mc_lock &bridge_netdev_addr_lock_key &c->lock irq_context: 0 &mm->mmap_lock &xa->xa_lock#4 irq_context: 0 &mm->mmap_lock &xa->xa_lock#4 pool_lock#2 irq_context: 0 &mm->mmap_lock &sb->s_type->i_lock_key irq_context: 0 &mm->mmap_lock &s->s_inode_list_lock irq_context: 0 &mm->mmap_lock batched_entropy_u32.lock irq_context: 0 &mm->mmap_lock &sb->s_type->i_lock_key &dentry->d_lock irq_context: 0 &mm->mmap_lock &xa->xa_lock#9 batched_entropy_u8.lock irq_context: 0 &mm->mmap_lock &xa->xa_lock#9 kfence_freelist_lock irq_context: 0 pernet_ops_rwsem ipvs->est_mutex &n->list_lock irq_context: 0 pernet_ops_rwsem ipvs->est_mutex &n->list_lock &c->lock irq_context: 0 rtnl_mutex lock kernfs_idr_lock &____s->seqcount#2 irq_context: softirq (&n->timer) &c->lock irq_context: 0 rtnl_mutex fill_pool_map-wait-type-override &n->list_lock irq_context: 0 rtnl_mutex fill_pool_map-wait-type-override &n->list_lock &c->lock irq_context: 0 &sb->s_type->i_mutex_key#10 sk_lock-AF_INET slock-AF_INET fill_pool_map-wait-type-override &c->lock irq_context: 0 rtnl_mutex &batadv_netdev_addr_lock_key irq_context: 0 rtnl_mutex &batadv_netdev_addr_lock_key pool_lock#2 irq_context: 0 rtnl_mutex &bond->mode_lock irq_context: 0 (wq_completion)wg-crypt-wg0#4 (work_completion)(&peer->transmit_packet_work) &peer->endpoint_lock rcu_read_lock_bh rcu_read_lock rcu_read_lock_bh rcu_read_lock pool_lock#2 irq_context: 0 (wq_completion)wg-crypt-wg0#4 (work_completion)(&peer->transmit_packet_work) &peer->endpoint_lock rcu_read_lock_bh rcu_read_lock rcu_read_lock_bh rcu_read_lock tk_core.seq.seqcount irq_context: 0 (wq_completion)wg-crypt-wg0#4 (work_completion)(&peer->transmit_packet_work) &peer->endpoint_lock rcu_read_lock_bh rcu_read_lock rcu_read_lock_bh rcu_read_lock rlock-AF_PACKET irq_context: 0 (wq_completion)wg-crypt-wg2#5 (work_completion)(&({ do { const void *__vpp_verify = (typeof((worker) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __asm__ ("" : "=r"(__ptr) : "0"((typeof(*((worker))) *)((worker)))); (typeof((typeof(*((worker))) *)((worker)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); })->work) pool_lock#2 irq_context: 0 &f->f_pos_lock sb_writers#4 &sb->s_type->i_mutex_key#8 &mm->mmap_lock ptlock_ptr(page) irq_context: 0 &f->f_pos_lock sb_writers#4 &sb->s_type->i_mutex_key#8 &journal->j_state_lock &journal->j_wait_commit &p->pi_lock &cfs_rq->removed.lock irq_context: 0 (wq_completion)phy4 (work_completion)(&(&local->scan_work)->work) &local->mtx rcu_state.exp_mutex rcu_node_0 irq_context: 0 (wq_completion)phy4 (work_completion)(&(&local->scan_work)->work) &local->mtx rcu_state.exp_mutex &obj_hash[i].lock irq_context: 0 (wq_completion)phy4 (work_completion)(&(&local->scan_work)->work) &local->mtx rcu_state.exp_mutex rcu_read_lock &pool->lock irq_context: 0 (wq_completion)phy4 (work_completion)(&(&local->scan_work)->work) &local->mtx rcu_state.exp_mutex rcu_read_lock &pool->lock &obj_hash[i].lock irq_context: 0 rtnl_mutex &tn->lock &rq->__lock irq_context: 0 rtnl_mutex &tn->lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 rtnl_mutex rcu_read_lock &bat_priv->tt.changes_list_lock irq_context: 0 rtnl_mutex rcu_read_lock &bat_priv->tt.changes_list_lock &obj_hash[i].lock irq_context: 0 rtnl_mutex rcu_read_lock &bat_priv->tt.changes_list_lock pool_lock#2 irq_context: 0 rtnl_mutex rcu_read_lock key#16 irq_context: 0 &sb->s_type->i_mutex_key#19 &rq->__lock irq_context: 0 rtnl_mutex &net->xfrm.xfrm_state_lock irq_context: 0 rtnl_mutex &idev->mc_lock &dev_addr_list_lock_key/1 irq_context: 0 rtnl_mutex &idev->mc_lock &dev_addr_list_lock_key/1 &obj_hash[i].lock irq_context: 0 rtnl_mutex &idev->mc_lock &dev_addr_list_lock_key/1 pool_lock#2 irq_context: 0 rtnl_mutex &idev->mc_lock &dev_addr_list_lock_key/1 krc.lock irq_context: 0 rtnl_mutex &bond->mode_lock pool_lock#2 irq_context: 0 rtnl_mutex &bond->mode_lock rcu_read_lock_bh &obj_hash[i].lock irq_context: 0 rtnl_mutex &bond->mode_lock rcu_read_lock_bh pool_lock#2 irq_context: 0 rtnl_mutex &batadv_netdev_addr_lock_key &obj_hash[i].lock irq_context: 0 rtnl_mutex &batadv_netdev_addr_lock_key krc.lock irq_context: 0 rtnl_mutex &bat_priv->tt.changes_list_lock &obj_hash[i].lock irq_context: 0 rtnl_mutex &bat_priv->tt.changes_list_lock pool_lock#2 irq_context: 0 rtnl_mutex (work_completion)(&(&slave->notify_work)->work) irq_context: 0 &sb->s_type->i_mutex_key#9 fs_reclaim &rq->__lock irq_context: 0 &sb->s_type->i_mutex_key#9 fs_reclaim &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 rtnl_mutex &bond->mode_lock rcu_read_lock_bh key#16 irq_context: 0 rtnl_mutex &bond->mode_lock rcu_read_lock_bh &bat_priv->tt.changes_list_lock irq_context: 0 &f->f_pos_lock sb_writers#4 &sb->s_type->i_mutex_key#8 &mm->mmap_lock rcu_read_lock &rcu_state.gp_wq irq_context: 0 &f->f_pos_lock sb_writers#4 &sb->s_type->i_mutex_key#8 &mm->mmap_lock rcu_read_lock &rcu_state.gp_wq &p->pi_lock irq_context: 0 &f->f_pos_lock sb_writers#4 &sb->s_type->i_mutex_key#8 &mm->mmap_lock rcu_read_lock &rcu_state.gp_wq &p->pi_lock &rq->__lock irq_context: 0 &f->f_pos_lock sb_writers#4 &sb->s_type->i_mutex_key#8 &mm->mmap_lock rcu_read_lock &rcu_state.gp_wq &p->pi_lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)events (work_completion)(&(&nsim_dev->trap_data->trap_report_dw)->work) &devlink->lock_key#6 &rcu_state.expedited_wq irq_context: 0 (wq_completion)events (work_completion)(&(&nsim_dev->trap_data->trap_report_dw)->work) &devlink->lock_key#6 &rcu_state.expedited_wq &p->pi_lock irq_context: 0 (wq_completion)events (work_completion)(&(&nsim_dev->trap_data->trap_report_dw)->work) &devlink->lock_key#6 &rcu_state.expedited_wq &p->pi_lock &rq->__lock irq_context: 0 (wq_completion)events (work_completion)(&(&nsim_dev->trap_data->trap_report_dw)->work) &devlink->lock_key#6 &rcu_state.expedited_wq &p->pi_lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 &sb->s_type->i_mutex_key#8 &sbi->s_writepages_rwsem rcu_read_lock &base->lock irq_context: 0 &sb->s_type->i_mutex_key#8 &sbi->s_writepages_rwsem rcu_read_lock &base->lock &obj_hash[i].lock irq_context: softirq (&in_dev->mr_ifc_timer) rcu_read_lock &im->lock &c->lock irq_context: 0 cb_lock nlk_cb_mutex-GENERIC rcu_read_lock pool_lock#2 irq_context: 0 cb_lock nlk_cb_mutex-GENERIC genl_mutex irq_context: 0 cb_lock nlk_cb_mutex-GENERIC genl_mutex &rq->__lock irq_context: 0 cb_lock nlk_cb_mutex-GENERIC genl_mutex &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 cb_lock nlk_cb_mutex-GENERIC genl_mutex &hash->list_locks[i] irq_context: 0 cb_lock nlk_cb_mutex-GENERIC rcu_read_lock &c->lock irq_context: 0 cb_lock nlk_cb_mutex-GENERIC rcu_read_lock &n->list_lock irq_context: 0 cb_lock nlk_cb_mutex-GENERIC rcu_read_lock &n->list_lock &c->lock irq_context: 0 cb_lock nlk_cb_mutex-GENERIC rcu_read_lock rcu_read_lock_bh &obj_hash[i].lock irq_context: 0 cb_lock nlk_cb_mutex-GENERIC rcu_read_lock rcu_read_lock_bh pool_lock#2 irq_context: 0 rtnl_mutex &root->kernfs_rwsem rcu_read_lock &rcu_state.gp_wq irq_context: 0 rtnl_mutex &root->kernfs_rwsem rcu_read_lock &rcu_state.gp_wq &p->pi_lock irq_context: 0 rtnl_mutex &root->kernfs_rwsem rcu_read_lock &rcu_state.gp_wq &p->pi_lock &rq->__lock irq_context: 0 rtnl_mutex &root->kernfs_rwsem rcu_read_lock &rcu_state.gp_wq &p->pi_lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 rcu_read_lock rcu_read_lock_bh rcu_read_lock pool_lock#2 irq_context: 0 rcu_read_lock rcu_read_lock_bh rcu_read_lock tk_core.seq.seqcount irq_context: 0 rcu_read_lock rcu_read_lock_bh rcu_read_lock &obj_hash[i].lock irq_context: 0 rcu_read_lock rcu_read_lock_bh rcu_read_lock rlock-AF_PACKET irq_context: 0 pernet_ops_rwsem rtnl_mutex fs_reclaim mmu_notifier_invalidate_range_start &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 rcu_read_lock rcu_read_lock_bh rcu_read_lock &c->lock irq_context: 0 (wq_completion)events (work_completion)(&(&nsim_dev->trap_data->trap_report_dw)->work) &devlink->lock_key#4 &nsim_trap_data->trap_lock batched_entropy_u8.lock irq_context: 0 (wq_completion)events (work_completion)(&(&nsim_dev->trap_data->trap_report_dw)->work) &devlink->lock_key#4 &nsim_trap_data->trap_lock kfence_freelist_lock irq_context: 0 (wq_completion)events (work_completion)(&(&nsim_dev->trap_data->trap_report_dw)->work) &devlink->lock_key#4 &nsim_trap_data->trap_lock &meta->lock irq_context: 0 (wq_completion)bat_events (work_completion)(&(&bat_priv->mcast.work)->work) &bat_priv->mcast.mla_lock &c->lock irq_context: 0 (wq_completion)bat_events (work_completion)(&(&bat_priv->mcast.work)->work) &bat_priv->mcast.mla_lock &n->list_lock irq_context: 0 (wq_completion)bat_events (work_completion)(&(&bat_priv->mcast.work)->work) &bat_priv->mcast.mla_lock &n->list_lock &c->lock irq_context: 0 (wq_completion)phy4 (work_completion)(&(&local->scan_work)->work) &local->mtx rcu_state.exp_mutex rcu_read_lock &pool->lock &p->pi_lock irq_context: 0 (wq_completion)phy4 (work_completion)(&(&local->scan_work)->work) &local->mtx rcu_state.exp_mutex rcu_read_lock &pool->lock &p->pi_lock &rq->__lock irq_context: 0 (wq_completion)phy4 (work_completion)(&(&local->scan_work)->work) &local->mtx rcu_state.exp_mutex rcu_read_lock &pool->lock &p->pi_lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)phy4 (work_completion)(&(&local->scan_work)->work) &local->mtx rcu_state.exp_mutex &rnp->exp_wq[1] irq_context: 0 (wq_completion)phy4 (work_completion)(&(&local->scan_work)->work) &local->mtx rcu_state.exp_mutex &rq->__lock irq_context: 0 (wq_completion)phy4 (work_completion)(&(&local->scan_work)->work) &local->mtx rcu_state.exp_mutex &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)bat_events (work_completion)(&(&forw_packet_aggr->delayed_work)->work) rcu_read_lock_bh rcu_read_lock pool_lock#2 irq_context: 0 (wq_completion)bat_events (work_completion)(&(&forw_packet_aggr->delayed_work)->work) rcu_read_lock_bh rcu_read_lock &obj_hash[i].lock irq_context: 0 (wq_completion)bat_events (work_completion)(&(&forw_packet_aggr->delayed_work)->work) rcu_read_lock_bh rcu_read_lock quarantine_lock irq_context: 0 rcu_read_lock rcu_read_lock_bh rcu_read_lock &n->list_lock irq_context: 0 rcu_read_lock rcu_read_lock_bh rcu_read_lock &n->list_lock &c->lock irq_context: 0 (wq_completion)phy4 (work_completion)(&(&local->scan_work)->work) &local->mtx &rdev->wiphy_work_lock irq_context: 0 (wq_completion)phy4 (work_completion)(&(&local->scan_work)->work) &local->mtx rcu_read_lock &pool->lock irq_context: 0 (wq_completion)phy4 (work_completion)(&(&local->scan_work)->work) &local->mtx rcu_read_lock &pool->lock &obj_hash[i].lock irq_context: 0 (wq_completion)phy4 (work_completion)(&(&local->scan_work)->work) &local->mtx rcu_read_lock &pool->lock &p->pi_lock irq_context: 0 (wq_completion)phy4 (work_completion)(&(&local->scan_work)->work) &local->mtx rcu_read_lock &pool->lock &p->pi_lock &rq->__lock irq_context: 0 (wq_completion)phy4 (work_completion)(&(&local->scan_work)->work) &local->mtx rcu_read_lock &pool->lock &p->pi_lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)phy4 (work_completion)(&(&local->scan_work)->work) &local->mtx rcu_node_0 irq_context: 0 (wq_completion)phy4 (work_completion)(&(&local->scan_work)->work) &local->mtx &rq->__lock irq_context: 0 (wq_completion)events (work_completion)(&rdev->wiphy_work) &rdev->wiphy.mtx fs_reclaim irq_context: 0 (wq_completion)events (work_completion)(&rdev->wiphy_work) &rdev->wiphy.mtx fs_reclaim mmu_notifier_invalidate_range_start irq_context: 0 (wq_completion)events (work_completion)(&rdev->wiphy_work) &rdev->wiphy.mtx &c->lock irq_context: 0 (wq_completion)events (work_completion)(&rdev->wiphy_work) &rdev->wiphy.mtx &n->list_lock irq_context: 0 (wq_completion)events (work_completion)(&rdev->wiphy_work) &rdev->wiphy.mtx &n->list_lock &c->lock irq_context: 0 (wq_completion)events (work_completion)(&rdev->wiphy_work) &rdev->wiphy.mtx &list->lock#2 irq_context: 0 (wq_completion)events (work_completion)(&rdev->wiphy_work) &rdev->wiphy.mtx rcu_read_lock &pool->lock irq_context: 0 (wq_completion)events (work_completion)(&rdev->wiphy_work) &rdev->wiphy.mtx rcu_read_lock &pool->lock &obj_hash[i].lock irq_context: 0 (wq_completion)events (work_completion)(&rdev->wiphy_work) &rdev->wiphy.mtx nl_table_lock irq_context: 0 (wq_completion)events (work_completion)(&rdev->wiphy_work) &rdev->wiphy.mtx nl_table_wait.lock irq_context: 0 (wq_completion)phy4 (work_completion)(&(&local->scan_work)->work) &local->mtx rcu_read_lock rcu_node_0 irq_context: 0 (wq_completion)phy4 (work_completion)(&(&local->scan_work)->work) &local->mtx &local->filter_lock irq_context: 0 (wq_completion)phy4 (work_completion)(&(&local->scan_work)->work) &local->mtx &local->iflist_mtx tk_core.seq.seqcount irq_context: 0 (wq_completion)phy4 (work_completion)(&(&local->scan_work)->work) &local->mtx &local->iflist_mtx hrtimer_bases.lock irq_context: 0 (wq_completion)phy4 (work_completion)(&(&local->scan_work)->work) &local->mtx &local->iflist_mtx hrtimer_bases.lock tk_core.seq.seqcount irq_context: 0 (wq_completion)phy4 (work_completion)(&(&local->scan_work)->work) &local->mtx &local->iflist_mtx hrtimer_bases.lock &obj_hash[i].lock irq_context: 0 (wq_completion)phy4 (work_completion)(&(&local->scan_work)->work) &local->mtx &local->queue_stop_reason_lock irq_context: 0 (wq_completion)phy4 (work_completion)(&(&local->scan_work)->work) &local->mtx rcu_read_lock &fq->lock irq_context: 0 (wq_completion)phy4 (work_completion)(&(&local->scan_work)->work) &local->mtx rcu_read_lock &local->queue_stop_reason_lock irq_context: 0 sb_writers#5 fill_pool_map-wait-type-override &c->lock irq_context: 0 sb_writers#5 fill_pool_map-wait-type-override &n->list_lock irq_context: 0 sb_writers#5 fill_pool_map-wait-type-override &n->list_lock &c->lock irq_context: 0 (wq_completion)bond1 irq_context: 0 (wq_completion)bond1 (work_completion)(&(&slave->notify_work)->work) irq_context: 0 (wq_completion)bond1 (work_completion)(&(&slave->notify_work)->work) &obj_hash[i].lock irq_context: 0 (wq_completion)bond1 (work_completion)(&(&slave->notify_work)->work) &base->lock irq_context: 0 (wq_completion)bond1 (work_completion)(&(&slave->notify_work)->work) &base->lock &obj_hash[i].lock irq_context: softirq (&in_dev->mr_ifc_timer) rcu_read_lock rcu_read_lock_bh rcu_read_lock &c->lock irq_context: softirq (&in_dev->mr_ifc_timer) rcu_read_lock rcu_read_lock_bh rcu_read_lock &n->list_lock irq_context: softirq (&in_dev->mr_ifc_timer) rcu_read_lock rcu_read_lock_bh rcu_read_lock &n->list_lock &c->lock irq_context: softirq (&in_dev->mr_ifc_timer) rcu_read_lock rcu_read_lock_bh rcu_read_lock tk_core.seq.seqcount irq_context: softirq (&in_dev->mr_ifc_timer) rcu_read_lock rcu_read_lock_bh rcu_read_lock rlock-AF_PACKET irq_context: 0 (wq_completion)mld (work_completion)(&(&idev->mc_ifc_work)->work) &idev->mc_lock rcu_read_lock rcu_read_lock rcu_read_lock_bh rcu_read_lock rlock-AF_PACKET irq_context: softirq (&hsr->announce_timer) rcu_read_lock &hsr->seqnr_lock rcu_read_lock rcu_read_lock_bh rcu_read_lock pool_lock#2 irq_context: softirq (&hsr->announce_timer) rcu_read_lock &hsr->seqnr_lock rcu_read_lock rcu_read_lock_bh rcu_read_lock &obj_hash[i].lock irq_context: softirq (&hsr->announce_timer) rcu_read_lock &hsr->seqnr_lock rcu_read_lock rcu_read_lock_bh rcu_read_lock &c->lock irq_context: 0 (wq_completion)bat_events (work_completion)(&(&forw_packet_aggr->delayed_work)->work) rcu_read_lock_bh rcu_read_lock rlock-AF_PACKET irq_context: 0 (wq_completion)wg-crypt-wg0#6 (work_completion)(&peer->transmit_packet_work) &peer->endpoint_lock rcu_read_lock_bh rcu_read_lock rcu_read_lock_bh rcu_read_lock tk_core.seq.seqcount irq_context: 0 (wq_completion)wg-crypt-wg0#6 (work_completion)(&peer->transmit_packet_work) &peer->endpoint_lock rcu_read_lock_bh rcu_read_lock rcu_read_lock_bh rcu_read_lock &obj_hash[i].lock irq_context: 0 (wq_completion)wg-crypt-wg0#6 (work_completion)(&peer->transmit_packet_work) &peer->endpoint_lock rcu_read_lock_bh rcu_read_lock rcu_read_lock_bh rcu_read_lock pool_lock#2 irq_context: 0 rcu_read_lock &dir->lock#2 irq_context: 0 sk_lock-AF_INET rcu_read_lock &c->lock irq_context: 0 sk_lock-AF_INET batched_entropy_u32.lock crngs.lock irq_context: 0 sk_lock-AF_INET rcu_read_lock &____s->seqcount#7 irq_context: 0 sk_lock-AF_INET rcu_read_lock &nf_nat_locks[i] irq_context: 0 sk_lock-AF_INET rcu_read_lock &nf_conntrack_locks[i] irq_context: 0 sk_lock-AF_INET rcu_read_lock &nf_conntrack_locks[i] batched_entropy_u8.lock irq_context: 0 sk_lock-AF_INET rcu_read_lock &____s->seqcount#10 irq_context: 0 sk_lock-AF_INET rcu_read_lock rcu_read_lock_bh rcu_read_lock pool_lock#2 irq_context: 0 sk_lock-AF_INET rcu_read_lock rcu_read_lock_bh rcu_read_lock tk_core.seq.seqcount irq_context: 0 sk_lock-AF_INET rcu_read_lock rcu_read_lock_bh rcu_read_lock &obj_hash[i].lock irq_context: 0 sk_lock-AF_INET rcu_read_lock rcu_read_lock_bh tk_core.seq.seqcount irq_context: 0 sk_lock-AF_INET rcu_read_lock rcu_read_lock_bh &list->lock#5 irq_context: softirq rcu_read_lock rcu_read_lock rlock-AF_INET irq_context: 0 &list->lock#28 irq_context: 0 &list->lock#28 rlock-AF_INET irq_context: 0 sb_writers#4 &sb->s_type->i_mutex_key#8 mapping.invalidate_lock remove_cache_srcu &rq->__lock irq_context: 0 sb_writers#4 &sb->s_type->i_mutex_key#8 mapping.invalidate_lock remove_cache_srcu &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: softirq (&n->timer) icmp_global.lock irq_context: softirq (&n->timer) icmp_global.lock batched_entropy_u8.lock irq_context: softirq (&n->timer) k-slock-AF_INET6 rcu_read_lock rcu_read_lock rcu_read_lock pool_lock#2 irq_context: softirq (&n->timer) k-slock-AF_INET6 rcu_read_lock rcu_read_lock rcu_read_lock &c->lock irq_context: softirq (&n->timer) k-slock-AF_INET6 rcu_read_lock rcu_read_lock rcu_read_lock &n->list_lock irq_context: softirq (&n->timer) k-slock-AF_INET6 rcu_read_lock rcu_read_lock rcu_read_lock &n->list_lock &c->lock irq_context: softirq (&n->timer) k-slock-AF_INET6 rcu_read_lock rcu_read_lock rcu_read_lock &dir->lock#2 irq_context: softirq (&n->timer) k-slock-AF_INET6 rcu_read_lock &c->lock irq_context: softirq (&n->timer) k-slock-AF_INET6 rcu_read_lock pool_lock#2 irq_context: softirq (&n->timer) k-slock-AF_INET6 rcu_read_lock rcu_read_lock pool_lock#2 irq_context: softirq (&n->timer) k-slock-AF_INET6 rcu_read_lock rcu_read_lock &obj_hash[i].lock irq_context: 0 sb_writers#4 &sb->s_type->i_mutex_key#8 mapping.invalidate_lock remove_cache_srcu rcu_read_lock rcu_node_0 irq_context: softirq (&n->timer) k-slock-AF_INET6 rcu_read_lock rcu_read_lock &c->lock irq_context: 0 sb_writers#4 &sb->s_type->i_mutex_key#8 mapping.invalidate_lock remove_cache_srcu rcu_read_lock &rq->__lock irq_context: softirq (&n->timer) k-slock-AF_INET6 rcu_read_lock rcu_read_lock &dir->lock#2 irq_context: 0 sb_writers#4 &sb->s_type->i_mutex_key#8 mapping.invalidate_lock remove_cache_srcu rcu_read_lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: softirq (&n->timer) k-slock-AF_INET6 rcu_read_lock rcu_read_lock &tbl->lock irq_context: softirq (&n->timer) k-slock-AF_INET6 rcu_read_lock rcu_read_lock &n->lock irq_context: softirq (&n->timer) k-slock-AF_INET6 rcu_read_lock rcu_read_lock &____s->seqcount#9 irq_context: softirq (&n->timer) k-slock-AF_INET6 rcu_read_lock rcu_read_lock rcu_read_lock_bh rcu_read_lock pool_lock#2 irq_context: softirq (&n->timer) k-slock-AF_INET6 rcu_read_lock rcu_read_lock rcu_read_lock_bh rcu_read_lock tk_core.seq.seqcount irq_context: softirq (&n->timer) k-slock-AF_INET6 rcu_read_lock rcu_read_lock rcu_read_lock_bh rcu_read_lock &obj_hash[i].lock irq_context: softirq (&n->timer) k-slock-AF_INET6 rcu_read_lock rcu_read_lock rcu_read_lock_bh tk_core.seq.seqcount irq_context: softirq (&n->timer) k-slock-AF_INET6 rcu_read_lock rcu_read_lock rcu_read_lock_bh &list->lock#5 irq_context: softirq (&n->timer) rcu_read_lock id_table_lock irq_context: softirq (&n->timer) &n->list_lock irq_context: softirq (&n->timer) &n->list_lock &c->lock irq_context: softirq (&n->timer) &n->lock irq_context: softirq (&n->timer) nl_table_lock irq_context: softirq (&n->timer) nl_table_wait.lock irq_context: 0 sb_writers#4 &sb->s_type->i_mutex_key#8 mapping.invalidate_lock remove_cache_srcu rcu_read_lock &cfs_rq->removed.lock irq_context: 0 sb_writers#4 &sb->s_type->i_mutex_key#8 mapping.invalidate_lock remove_cache_srcu rcu_read_lock &obj_hash[i].lock irq_context: 0 sk_lock-AF_NETLINK rcu_read_lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 pernet_ops_rwsem uevent_sock_mutex rcu_read_lock &ei->socket.wq.wait &ep->lock &ep->wq &p->pi_lock &cfs_rq->removed.lock irq_context: 0 rtnl_mutex fill_pool_map-wait-type-override rcu_read_lock rcu_node_0 irq_context: 0 rtnl_mutex fill_pool_map-wait-type-override rcu_read_lock &rq->__lock irq_context: 0 rtnl_mutex fill_pool_map-wait-type-override rcu_read_lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)events_power_efficient (work_completion)(&(&tbl->gc_work)->work) &rq->__lock irq_context: 0 (wq_completion)events_power_efficient (work_completion)(&(&tbl->gc_work)->work) &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)events_power_efficient (work_completion)(&(&tbl->gc_work)->work) &tbl->lock pool_lock#2 irq_context: 0 (wq_completion)events_power_efficient (work_completion)(&(&tbl->gc_work)->work) &tbl->lock &c->lock irq_context: 0 (wq_completion)events_power_efficient (work_completion)(&(&tbl->gc_work)->work) &tbl->lock &n->lock irq_context: 0 (wq_completion)events_power_efficient (work_completion)(&(&tbl->gc_work)->work) &tbl->lock nl_table_lock irq_context: 0 (wq_completion)events_power_efficient (work_completion)(&(&tbl->gc_work)->work) &tbl->lock nl_table_wait.lock irq_context: 0 (wq_completion)events_power_efficient (work_completion)(&(&tbl->gc_work)->work) &tbl->lock rcu_read_lock id_table_lock irq_context: 0 (wq_completion)events_power_efficient (work_completion)(&(&tbl->gc_work)->work) &tbl->lock &dir->lock#2 irq_context: 0 (wq_completion)events_power_efficient (work_completion)(&(&tbl->gc_work)->work) &tbl->lock krc.lock irq_context: 0 sb_writers#4 &sb->s_type->i_mutex_key#8 mapping.invalidate_lock rcu_read_lock rcu_read_lock rcu_node_0 irq_context: 0 sb_writers#4 &sb->s_type->i_mutex_key#8 mapping.invalidate_lock rcu_read_lock rcu_read_lock &rq->__lock irq_context: 0 sb_writers#4 &sb->s_type->i_mutex_key#8 mapping.invalidate_lock rcu_read_lock rcu_read_lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 rtnl_mutex &idev->mc_lock fill_pool_map-wait-type-override &rq->__lock irq_context: 0 rtnl_mutex &idev->mc_lock fill_pool_map-wait-type-override &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 sb_writers#4 &type->i_mutex_dir_key#3/1 &sb->s_type->i_mutex_key#8 remove_cache_srcu rcu_node_0 irq_context: 0 sb_writers#4 &type->i_mutex_dir_key#3/1 &sb->s_type->i_mutex_key#8 remove_cache_srcu &rcu_state.expedited_wq irq_context: 0 sb_writers#4 &type->i_mutex_dir_key#3/1 &sb->s_type->i_mutex_key#8 remove_cache_srcu &rcu_state.expedited_wq &p->pi_lock irq_context: 0 sb_writers#4 &type->i_mutex_dir_key#3/1 &sb->s_type->i_mutex_key#8 remove_cache_srcu &rcu_state.expedited_wq &p->pi_lock &rq->__lock irq_context: 0 sb_writers#4 &type->i_mutex_dir_key#3/1 &sb->s_type->i_mutex_key#8 remove_cache_srcu &rcu_state.expedited_wq &p->pi_lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 rtnl_mutex &batadv_netdev_addr_lock_key &c->lock irq_context: 0 pernet_ops_rwsem k-sk_lock-AF_INET6 &rq->__lock irq_context: 0 pernet_ops_rwsem k-sk_lock-AF_INET6 &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 pernet_ops_rwsem cpu_hotplug_lock wq_pool_mutex &____s->seqcount#2 irq_context: 0 pernet_ops_rwsem nf_ct_proto_mutex nf_hook_mutex fs_reclaim &rq->__lock irq_context: 0 pernet_ops_rwsem nf_ct_proto_mutex nf_hook_mutex fs_reclaim &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)wg-crypt-wg0 (work_completion)(&peer->transmit_packet_work) &peer->endpoint_lock rcu_read_lock_bh rcu_read_lock rcu_read_lock_bh rcu_read_lock pool_lock#2 irq_context: 0 (wq_completion)wg-crypt-wg0 (work_completion)(&peer->transmit_packet_work) &peer->endpoint_lock rcu_read_lock_bh rcu_read_lock rcu_read_lock_bh rcu_read_lock tk_core.seq.seqcount irq_context: 0 (wq_completion)wg-crypt-wg0 (work_completion)(&peer->transmit_packet_work) &peer->endpoint_lock rcu_read_lock_bh rcu_read_lock rcu_read_lock_bh rcu_read_lock &obj_hash[i].lock irq_context: 0 &sb->s_type->i_mutex_key#8 rcu_read_lock fill_pool_map-wait-type-override &n->list_lock irq_context: 0 &sb->s_type->i_mutex_key#8 rcu_read_lock fill_pool_map-wait-type-override &n->list_lock &c->lock irq_context: 0 &sbi->s_writepages_rwsem jbd2_handle &ei->i_data_sem &ei->i_es_lock &c->lock irq_context: 0 &sbi->s_writepages_rwsem jbd2_handle &mapping->i_mmap_rwsem mmu_notifier_invalidate_range_start irq_context: 0 &sbi->s_writepages_rwsem jbd2_handle &mapping->i_mmap_rwsem ptlock_ptr(page)#2 irq_context: 0 &sbi->s_writepages_rwsem jbd2_handle rcu_read_lock &rq->__lock &cfs_rq->removed.lock irq_context: 0 &sb->s_type->i_mutex_key#10 clock-AF_RXRPC irq_context: 0 &sb->s_type->i_mutex_key#10 (wq_completion)krxrpcd irq_context: 0 &sb->s_type->i_mutex_key#10 &wq->mutex irq_context: 0 &sb->s_type->i_mutex_key#10 &wq->mutex &pool->lock/1 irq_context: 0 &sb->s_type->i_mutex_key#10 &wq->mutex &x->wait#10 irq_context: 0 &sb->s_type->i_mutex_key#10 rlock-AF_RXRPC irq_context: 0 cb_lock genl_mutex rcu_read_lock rcu_read_lock_bh rcu_read_lock pool_lock#2 irq_context: 0 cb_lock genl_mutex rcu_read_lock rcu_read_lock_bh rcu_read_lock tk_core.seq.seqcount irq_context: 0 cb_lock genl_mutex rcu_read_lock rcu_read_lock_bh rcu_read_lock &obj_hash[i].lock irq_context: 0 cb_lock rcu_read_lock rcu_read_lock_bh rcu_read_lock pool_lock#2 irq_context: 0 cb_lock rcu_read_lock rcu_read_lock_bh rcu_read_lock tk_core.seq.seqcount irq_context: 0 cb_lock rcu_read_lock rcu_read_lock_bh rcu_read_lock &obj_hash[i].lock irq_context: 0 &sb->s_type->i_mutex_key#10 fill_pool_map-wait-type-override &c->lock irq_context: 0 sk_lock-AF_PHONET irq_context: 0 sk_lock-AF_PHONET slock-AF_PHONET irq_context: 0 sk_lock-AF_PHONET port_mutex#2 irq_context: 0 sk_lock-AF_PHONET port_mutex#2 local_port_range_lock.seqcount irq_context: 0 sk_lock-AF_PHONET port_mutex#2 &rq->__lock irq_context: 0 sk_lock-AF_PHONET port_mutex#2 &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 sk_lock-AF_PHONET port_mutex#2 &pnsocks.lock irq_context: 0 slock-AF_PHONET irq_context: 0 rtnl_mutex rcu_read_lock rcu_read_lock_bh rcu_read_lock pool_lock#2 irq_context: 0 rtnl_mutex rcu_read_lock rcu_read_lock_bh rcu_read_lock tk_core.seq.seqcount irq_context: 0 rtnl_mutex rcu_read_lock rcu_read_lock_bh rcu_read_lock rlock-AF_PACKET irq_context: 0 (wq_completion)bond1 (work_completion)(&(&slave->notify_work)->work) rtnl_mutex net_rwsem irq_context: 0 (wq_completion)bond1 (work_completion)(&(&slave->notify_work)->work) rtnl_mutex net_rwsem &list->lock#2 irq_context: 0 (wq_completion)bond1 (work_completion)(&(&slave->notify_work)->work) rtnl_mutex &tn->lock irq_context: 0 (wq_completion)bond1 (work_completion)(&(&slave->notify_work)->work) rtnl_mutex.wait_lock irq_context: 0 (wq_completion)bond1 (work_completion)(&(&slave->notify_work)->work) &p->pi_lock irq_context: 0 pernet_ops_rwsem k-sk_lock-AF_RXRPC &rxnet->local_mutex &n->list_lock irq_context: 0 pernet_ops_rwsem k-sk_lock-AF_RXRPC &rxnet->local_mutex &n->list_lock &c->lock irq_context: 0 &sb->s_type->i_mutex_key#10 sk_lock-AF_PHONET irq_context: 0 &sb->s_type->i_mutex_key#10 sk_lock-AF_PHONET slock-AF_PHONET irq_context: 0 &sb->s_type->i_mutex_key#10 sk_lock-AF_PHONET &pnsocks.lock irq_context: 0 rtnl_mutex &idev->mc_lock &batadv_netdev_addr_lock_key &c->lock irq_context: 0 rtnl_mutex rcu_state.exp_mutex &rq->__lock &cfs_rq->removed.lock irq_context: 0 (wq_completion)events (work_completion)(&(&nsim_dev->trap_data->trap_report_dw)->work) &devlink->lock_key#2 rcu_node_0 irq_context: 0 (wq_completion)events (work_completion)(&(&nsim_dev->trap_data->trap_report_dw)->work) &devlink->lock_key#2 &rcu_state.expedited_wq irq_context: 0 (wq_completion)events (work_completion)(&(&nsim_dev->trap_data->trap_report_dw)->work) &devlink->lock_key#2 &rcu_state.expedited_wq &p->pi_lock irq_context: 0 (wq_completion)events (work_completion)(&(&nsim_dev->trap_data->trap_report_dw)->work) &devlink->lock_key#2 &rcu_state.expedited_wq &p->pi_lock &rq->__lock irq_context: 0 (wq_completion)events (work_completion)(&(&nsim_dev->trap_data->trap_report_dw)->work) &devlink->lock_key#2 &rcu_state.expedited_wq &p->pi_lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 kn->active#5 fs_reclaim &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 rcu_read_lock &rcu_state.expedited_wq &p->pi_lock &rq->__lock &cfs_rq->removed.lock irq_context: 0 vlan_ioctl_mutex rtnl_mutex team->team_lock_key#3 irq_context: 0 vlan_ioctl_mutex rtnl_mutex team->team_lock_key#3 fs_reclaim irq_context: 0 vlan_ioctl_mutex rtnl_mutex team->team_lock_key#3 fs_reclaim mmu_notifier_invalidate_range_start irq_context: 0 vlan_ioctl_mutex rtnl_mutex team->team_lock_key#3 pool_lock#2 irq_context: 0 vlan_ioctl_mutex rtnl_mutex &dev_addr_list_lock_key#2/1 irq_context: 0 vlan_ioctl_mutex rtnl_mutex &dev_addr_list_lock_key#2/1 rcu_read_lock _xmit_ETHER irq_context: 0 vlan_ioctl_mutex rtnl_mutex &dev_addr_list_lock_key#2/1 rcu_read_lock _xmit_ETHER pool_lock#2 irq_context: 0 vlan_ioctl_mutex rtnl_mutex &idev->mc_lock &obj_hash[i].lock pool_lock irq_context: 0 vlan_ioctl_mutex &p->pi_lock &cfs_rq->removed.lock irq_context: 0 &sb->s_type->i_mutex_key#10 sk_lock-AF_PHONET resource_mutex irq_context: 0 &sb->s_type->i_mutex_key#10 sk_lock-AF_PHONET rcu_state.exp_mutex rcu_node_0 irq_context: 0 &sb->s_type->i_mutex_key#10 sk_lock-AF_PHONET rcu_state.exp_mutex &obj_hash[i].lock irq_context: 0 &sb->s_type->i_mutex_key#10 sk_lock-AF_PHONET rcu_state.exp_mutex rcu_read_lock &pool->lock irq_context: 0 &sb->s_type->i_mutex_key#10 sk_lock-AF_PHONET rcu_state.exp_mutex rcu_read_lock &pool->lock &obj_hash[i].lock irq_context: 0 &sb->s_type->i_mutex_key#10 sk_lock-AF_PHONET rcu_state.exp_mutex rcu_read_lock &pool->lock &p->pi_lock irq_context: 0 &sb->s_type->i_mutex_key#10 sk_lock-AF_PHONET rcu_state.exp_mutex rcu_read_lock &pool->lock &p->pi_lock &rq->__lock irq_context: 0 &sb->s_type->i_mutex_key#10 sk_lock-AF_PHONET rcu_state.exp_mutex rcu_read_lock &pool->lock &p->pi_lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 &sb->s_type->i_mutex_key#10 sk_lock-AF_PHONET rcu_state.exp_mutex &rq->__lock irq_context: 0 &sb->s_type->i_mutex_key#10 sk_lock-AF_PHONET rcu_state.exp_mutex &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 &sb->s_type->i_mutex_key#10 sk_lock-AF_PHONET rcu_state.exp_mutex &rnp->exp_wq[0] irq_context: 0 &sb->s_type->i_mutex_key#10 sk_lock-AF_PHONET rcu_state.exp_mutex rcu_read_lock &rq->__lock irq_context: 0 &sb->s_type->i_mutex_key#10 sk_lock-AF_PHONET rcu_state.exp_mutex rcu_read_lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 &sb->s_type->i_mutex_key#10 sk_lock-AF_PHONET &obj_hash[i].lock irq_context: 0 &sb->s_type->i_mutex_key#10 slock-AF_PHONET irq_context: 0 &sb->s_type->i_mutex_key#10 &list->lock#29 irq_context: 0 &mm->mmap_lock stock_lock rcu_read_lock per_cpu_ptr(&cgroup_rstat_cpu_lock, cpu) irq_context: 0 pernet_ops_rwsem rtnl_mutex kernfs_idr_lock &obj_hash[i].lock irq_context: 0 pernet_ops_rwsem rtnl_mutex kernfs_idr_lock pool_lock#2 irq_context: 0 sb_writers#4 &type->i_mutex_dir_key#3/1 sb_internal jbd2_handle &ei->i_data_sem &rq->__lock &cfs_rq->removed.lock irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem &rcu_state.expedited_wq irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem &rcu_state.expedited_wq &p->pi_lock irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem &rcu_state.expedited_wq &p->pi_lock &rq->__lock irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem &rcu_state.expedited_wq &p->pi_lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 pernet_ops_rwsem krc.lock &obj_hash[i].lock irq_context: 0 pernet_ops_rwsem krc.lock &base->lock irq_context: 0 pernet_ops_rwsem krc.lock &base->lock &obj_hash[i].lock irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem per_cpu_ptr(&cgroup_rstat_cpu_lock, cpu) irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem uevent_sock_mutex rcu_read_lock &ei->socket.wq.wait &ep->lock &ep->wq irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem uevent_sock_mutex rcu_read_lock &ei->socket.wq.wait &ep->lock &ep->wq &p->pi_lock irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem uevent_sock_mutex rcu_read_lock &ei->socket.wq.wait &ep->lock &ep->wq &p->pi_lock &rq->__lock irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem uevent_sock_mutex rcu_read_lock &ei->socket.wq.wait &ep->lock &ep->wq &p->pi_lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)netns net_cleanup_work &obj_hash[i].lock pool_lock irq_context: 0 &p->alloc_lock &x->wait &p->pi_lock &rq->__lock &cfs_rq->removed.lock irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem k-sk_lock-AF_RXRPC &base->lock &obj_hash[i].lock irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem ovs_mutex nf_ct_proto_mutex nf_hook_mutex batched_entropy_u8.lock irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem ovs_mutex nf_ct_proto_mutex nf_hook_mutex kfence_freelist_lock irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem ovs_mutex quarantine_lock irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem ovs_mutex nf_ct_proto_mutex nf_hook_mutex remove_cache_srcu irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem ovs_mutex nf_ct_proto_mutex nf_hook_mutex remove_cache_srcu quarantine_lock irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem ovs_mutex nf_ct_proto_mutex nf_hook_mutex remove_cache_srcu &c->lock irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem ovs_mutex nf_ct_proto_mutex nf_hook_mutex remove_cache_srcu &n->list_lock irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem ovs_mutex nf_ct_proto_mutex nf_hook_mutex remove_cache_srcu &rq->__lock irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem ovs_mutex nf_ct_proto_mutex nf_hook_mutex remove_cache_srcu &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem ovs_mutex nf_ct_proto_mutex nf_hook_mutex remove_cache_srcu rcu_read_lock pool_lock#2 irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem ovs_mutex nf_ct_proto_mutex nf_hook_mutex remove_cache_srcu &obj_hash[i].lock irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem ovs_mutex nf_ct_proto_mutex nf_hook_mutex &c->lock irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem ovs_mutex nf_ct_proto_mutex nf_hook_mutex &____s->seqcount#2 irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem ovs_mutex nf_ct_proto_mutex nf_hook_mutex &____s->seqcount irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem rtnl_mutex &____s->seqcount#2 irq_context: 0 (wq_completion)events (work_completion)(&rdev->wiphy_work) &rdev->wiphy.mtx &wdev->mtx rcu_read_lock rcu_read_lock rcu_node_0 irq_context: 0 (wq_completion)events (work_completion)(&rdev->wiphy_work) &rdev->wiphy.mtx &wdev->mtx rcu_read_lock rcu_read_lock &rq->__lock irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem rtnl_mutex quarantine_lock irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem rtnl_mutex remove_cache_srcu pool_lock#2 irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem rtnl_mutex &root->kernfs_rwsem &root->kernfs_iattr_rwsem &rq->__lock irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem rtnl_mutex &root->kernfs_rwsem &root->kernfs_iattr_rwsem &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem rtnl_mutex &pnn->pndevs.lock &rq->__lock irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem rtnl_mutex &pnn->pndevs.lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem rtnl_mutex &root->kernfs_rwsem kernfs_idr_lock &obj_hash[i].lock pool_lock irq_context: softirq &(&hwstats->traffic_dw)->timer rcu_read_lock &pool->lock &p->pi_lock &rq->__lock &cfs_rq->removed.lock irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem rtnl_mutex &idev->mc_lock &obj_hash[i].lock pool_lock irq_context: 0 (wq_completion)events fqdir_free_work rcu_state.barrier_mutex rcu_state.barrier_mutex.wait_lock irq_context: 0 (wq_completion)events fqdir_free_work rcu_state.barrier_mutex rcu_read_lock &rq->__lock irq_context: 0 (wq_completion)events fqdir_free_work rcu_state.barrier_mutex rcu_read_lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)events (work_completion)(&(&nsim_dev->trap_data->trap_report_dw)->work) &devlink->lock_key &rcu_state.expedited_wq irq_context: 0 (wq_completion)events free_ipc_work rcu_state.exp_mutex rcu_read_lock &rq->__lock irq_context: 0 (wq_completion)events (work_completion)(&(&nsim_dev->trap_data->trap_report_dw)->work) &devlink->lock_key &rcu_state.expedited_wq &p->pi_lock irq_context: 0 (wq_completion)events free_ipc_work rcu_state.exp_mutex rcu_read_lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem rtnl_mutex &caifn->caifdevs.lock &rnp->exp_lock irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem rtnl_mutex &caifn->caifdevs.lock &rnp->exp_wq[0] irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem rtnl_mutex &caifn->caifdevs.lock &rq->__lock irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem rtnl_mutex &caifn->caifdevs.lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem rtnl_mutex &caifn->caifdevs.lock rcu_state.exp_mutex rcu_read_lock &pool->lock &p->pi_lock &rq->__lock irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem rtnl_mutex &caifn->caifdevs.lock rcu_state.exp_mutex rcu_read_lock &pool->lock &p->pi_lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem rtnl_mutex &caifn->caifdevs.lock rcu_state.exp_mutex &rnp->exp_wq[2] irq_context: 0 &sb->s_type->i_lock_key &xa->xa_lock#9 irq_context: 0 &lruvec->lru_lock irq_context: 0 &sb->s_type->i_lock_key &xa->xa_lock#9 &obj_hash[i].lock irq_context: 0 &sb->s_type->i_lock_key &xa->xa_lock#9 pool_lock#2 irq_context: 0 &sb->s_type->i_lock_key#27 irq_context: 0 &f->f_pos_lock &sb->s_type->i_mutex_key#18 irq_context: 0 &f->f_pos_lock &sb->s_type->i_mutex_key#18 &dentry->d_lock irq_context: 0 &f->f_pos_lock &sb->s_type->i_mutex_key#18 tk_core.seq.seqcount irq_context: 0 &f->f_pos_lock &sb->s_type->i_mutex_key#18 sb_writers#13 mount_lock irq_context: 0 &f->f_pos_lock &sb->s_type->i_mutex_key#18 sb_writers#13 tk_core.seq.seqcount irq_context: 0 &f->f_pos_lock &sb->s_type->i_mutex_key#18 sb_writers#13 &sb->s_type->i_lock_key#27 irq_context: 0 &f->f_pos_lock &sb->s_type->i_mutex_key#18 sb_writers#13 &wb->list_lock irq_context: 0 &f->f_pos_lock &sb->s_type->i_mutex_key#18 sb_writers#13 &wb->list_lock &sb->s_type->i_lock_key#27 irq_context: 0 &info->lock irq_context: 0 rtnl_mutex &root->kernfs_rwsem &root->kernfs_iattr_rwsem &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 sb_writers#5 &type->i_mutex_dir_key#5/1 tomoyo_ss mmu_notifier_invalidate_range_start &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 sb_writers#5 &type->i_mutex_dir_key#5/1 &sb->s_type->i_mutex_key#12 rcu_read_lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 rtnl_mutex uevent_sock_mutex &cfs_rq->removed.lock irq_context: 0 rtnl_mutex pcpu_lock stock_lock irq_context: 0 &sbi->s_writepages_rwsem rcu_read_lock &____s->seqcount irq_context: 0 (wq_completion)events &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 rtnl_mutex &rnp->exp_wq[0] irq_context: 0 rtnl_mutex &idev->mc_lock &batadv_netdev_addr_lock_key &obj_hash[i].lock irq_context: 0 rtnl_mutex &wq->mutex &pool->lock/1 irq_context: 0 rtnl_mutex &wq->mutex &x->wait#10 irq_context: 0 rtnl_mutex &pool->lock/1 irq_context: 0 rtnl_mutex &pool->lock/1 rcu_read_lock &pool->lock irq_context: 0 rtnl_mutex &pool->lock/1 rcu_read_lock &pool->lock &obj_hash[i].lock irq_context: 0 rtnl_mutex &pool->lock/1 rcu_read_lock &pool->lock pool_lock#2 irq_context: 0 rtnl_mutex &pool->lock/1 rcu_read_lock &pool->lock &p->pi_lock irq_context: 0 rtnl_mutex &pool->lock/1 rcu_read_lock &pool->lock &p->pi_lock &rq->__lock irq_context: 0 rtnl_mutex &pool->lock/1 rcu_read_lock &pool->lock &p->pi_lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)events &rq->__lock irq_context: 0 rtnl_mutex &idev->mc_lock &batadv_netdev_addr_lock_key krc.lock irq_context: 0 rtnl_mutex uevent_sock_mutex rcu_node_0 irq_context: 0 (wq_completion)bond2 irq_context: 0 (wq_completion)bond2 (work_completion)(&(&slave->notify_work)->work) irq_context: 0 (wq_completion)bond2 (work_completion)(&(&slave->notify_work)->work) &obj_hash[i].lock irq_context: 0 (wq_completion)bond2 (work_completion)(&(&slave->notify_work)->work) &base->lock irq_context: 0 (wq_completion)bond2 (work_completion)(&(&slave->notify_work)->work) &base->lock &obj_hash[i].lock irq_context: 0 &kernfs_locks->open_file_mutex[count] rcu_read_lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: softirq (&in_dev->mr_ifc_timer) rcu_read_lock &n->list_lock irq_context: softirq (&in_dev->mr_ifc_timer) rcu_read_lock &n->list_lock &c->lock irq_context: softirq (&in_dev->mr_ifc_timer) rcu_read_lock &tbl->lock pool_lock#2 irq_context: softirq (&in_dev->mr_ifc_timer) rcu_read_lock &tbl->lock batched_entropy_u32.lock irq_context: softirq (&in_dev->mr_ifc_timer) rcu_read_lock &tbl->lock &obj_hash[i].lock irq_context: 0 (wq_completion)mld (work_completion)(&(&idev->mc_ifc_work)->work) &idev->mc_lock rcu_read_lock rcu_read_lock rcu_read_lock_bh rcu_read_lock &meta->lock irq_context: 0 (wq_completion)mld (work_completion)(&(&idev->mc_ifc_work)->work) &idev->mc_lock rcu_read_lock rcu_read_lock rcu_read_lock_bh rcu_read_lock kfence_freelist_lock irq_context: 0 sb_writers#5 &type->i_mutex_dir_key#5/1 batched_entropy_u8.lock irq_context: 0 sb_writers#5 &type->i_mutex_dir_key#5/1 kfence_freelist_lock irq_context: softirq (&peer->timer_new_handshake) irq_context: softirq (&peer->timer_new_handshake) &peer->endpoint_lock irq_context: softirq (&peer->timer_new_handshake) rcu_read_lock_bh tk_core.seq.seqcount irq_context: softirq (&peer->timer_new_handshake) rcu_read_lock_bh rcu_read_lock &pool->lock/1 irq_context: softirq (&peer->timer_new_handshake) rcu_read_lock_bh rcu_read_lock &pool->lock/1 &obj_hash[i].lock irq_context: softirq (&peer->timer_new_handshake) rcu_read_lock_bh rcu_read_lock &pool->lock/1 pool_lock#2 irq_context: softirq (&peer->timer_new_handshake) rcu_read_lock_bh rcu_read_lock &pool->lock/1 &p->pi_lock irq_context: softirq (&peer->timer_new_handshake) rcu_read_lock_bh rcu_read_lock &pool->lock/1 &p->pi_lock &rq->__lock irq_context: softirq (&peer->timer_new_handshake) rcu_read_lock_bh rcu_read_lock &pool->lock/1 &p->pi_lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)wg-kex-wg2#9 (work_completion)(&peer->transmit_handshake_work) &peer->endpoint_lock rcu_read_lock_bh rcu_read_lock &c->lock irq_context: 0 (wq_completion)wg-kex-wg2#9 (work_completion)(&peer->transmit_handshake_work) &peer->endpoint_lock rcu_read_lock_bh rcu_read_lock &ul->lock irq_context: 0 (wq_completion)wg-kex-wg2#9 (work_completion)(&peer->transmit_handshake_work) &peer->endpoint_lock rcu_read_lock_bh rcu_read_lock &obj_hash[i].lock irq_context: 0 &sbi->s_writepages_rwsem jbd2_handle &ei->i_data_sem &mapping->private_lock irq_context: 0 &sbi->s_writepages_rwsem jbd2_handle &rq->__lock irq_context: 0 &sbi->s_writepages_rwsem jbd2_handle &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 &sbi->s_writepages_rwsem jbd2_handle lock#4 &lruvec->lru_lock irq_context: 0 (wq_completion)bond2 (work_completion)(&(&slave->notify_work)->work) rtnl_mutex net_rwsem irq_context: 0 (wq_completion)bond2 (work_completion)(&(&slave->notify_work)->work) rtnl_mutex net_rwsem &list->lock#2 irq_context: 0 (wq_completion)bond2 (work_completion)(&(&slave->notify_work)->work) rtnl_mutex &tn->lock irq_context: 0 (wq_completion)bond2 (work_completion)(&(&slave->notify_work)->work) rtnl_mutex.wait_lock irq_context: 0 (wq_completion)bond2 (work_completion)(&(&slave->notify_work)->work) &p->pi_lock irq_context: 0 fill_pool_map-wait-type-override &rq->__lock &cfs_rq->removed.lock irq_context: 0 (wq_completion)wg-crypt-wg0#4 (work_completion)(&peer->transmit_packet_work) &peer->endpoint_lock rcu_read_lock_bh rcu_read_lock rcu_read_lock_bh rcu_read_lock &obj_hash[i].lock irq_context: 0 (wq_completion)bond1#2 irq_context: 0 (wq_completion)bond1#2 (work_completion)(&(&slave->notify_work)->work) irq_context: 0 (wq_completion)bond1#2 (work_completion)(&(&slave->notify_work)->work) &obj_hash[i].lock irq_context: 0 (wq_completion)bond1#2 (work_completion)(&(&slave->notify_work)->work) &base->lock irq_context: 0 (wq_completion)bond1#2 (work_completion)(&(&slave->notify_work)->work) &base->lock &obj_hash[i].lock irq_context: 0 lock pidmap_lock &n->list_lock irq_context: 0 sb_writers#11 &type->i_mutex_dir_key#7/1 cgroup_mutex &root->kernfs_rwsem &rq->__lock irq_context: 0 sb_writers#11 &type->i_mutex_dir_key#7/1 cgroup_mutex &root->kernfs_rwsem &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: softirq (&hsr->announce_timer) rcu_read_lock &hsr->seqnr_lock rcu_read_lock rcu_read_lock_bh rcu_read_lock rlock-AF_PACKET irq_context: 0 (wq_completion)bat_events (work_completion)(&(&forw_packet_aggr->delayed_work)->work) rcu_read_lock_bh rcu_read_lock &c->lock irq_context: 0 rcu_read_lock rcu_read_lock_bh rcu_read_lock_bh rcu_read_lock &pool->lock &p->pi_lock &rq->__lock irq_context: 0 cb_lock nlk_cb_mutex-GENERIC &rq->__lock irq_context: 0 cb_lock nlk_cb_mutex-GENERIC &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 cb_lock nlk_cb_mutex-GENERIC rcu_read_lock rcu_read_lock_bh rcu_read_lock pool_lock#2 irq_context: 0 cb_lock nlk_cb_mutex-GENERIC rcu_read_lock rcu_read_lock_bh rcu_read_lock tk_core.seq.seqcount irq_context: 0 cb_lock nlk_cb_mutex-GENERIC rcu_read_lock rcu_read_lock_bh rcu_read_lock &obj_hash[i].lock irq_context: 0 sb_writers#4 &type->i_mutex_dir_key#3 stock_lock irq_context: 0 sb_writers#4 &type->i_mutex_dir_key#3 &c->lock irq_context: 0 rcu_read_lock &ul->lock irq_context: 0 rcu_read_lock &____s->seqcount#7 irq_context: 0 rcu_read_lock &nf_nat_locks[i] irq_context: 0 rcu_read_lock &nf_conntrack_locks[i] irq_context: 0 rcu_read_lock &nf_conntrack_locks[i] &nf_conntrack_locks[i]/1 irq_context: 0 rcu_read_lock &nf_conntrack_locks[i] &nf_conntrack_locks[i]/1 batched_entropy_u8.lock irq_context: 0 rcu_read_lock &tbl->lock irq_context: 0 rcu_read_lock &____s->seqcount#9 irq_context: 0 rcu_read_lock rcu_read_lock_bh dev->qdisc_tx_busylock ?: &qdisc_tx_busylock rcu_read_lock pool_lock#2 irq_context: 0 rcu_read_lock rcu_read_lock_bh dev->qdisc_tx_busylock ?: &qdisc_tx_busylock rcu_read_lock tk_core.seq.seqcount irq_context: 0 rcu_read_lock rcu_read_lock_bh dev->qdisc_tx_busylock ?: &qdisc_tx_busylock rcu_read_lock &obj_hash[i].lock irq_context: 0 rcu_read_lock rcu_read_lock_bh dev->qdisc_tx_busylock ?: &qdisc_tx_busylock rcu_read_lock &r->producer_lock irq_context: 0 rcu_read_lock rcu_read_lock_bh &list->lock#17 irq_context: 0 rcu_read_lock rcu_read_lock_bh tk_core.seq.seqcount irq_context: 0 rcu_read_lock rcu_read_lock_bh rcu_read_lock_bh &r->producer_lock#2 irq_context: 0 rcu_read_lock rcu_read_lock_bh rcu_read_lock_bh rcu_read_lock &pool->lock irq_context: 0 rcu_read_lock rcu_read_lock_bh rcu_read_lock_bh rcu_read_lock &pool->lock &obj_hash[i].lock irq_context: 0 rcu_read_lock rcu_read_lock_bh rcu_read_lock_bh rcu_read_lock &pool->lock &p->pi_lock irq_context: 0 rcu_read_lock rcu_read_lock_bh rcu_read_lock_bh rcu_read_lock &pool->lock &p->pi_lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)wg-crypt-wg1#5 (work_completion)(&peer->transmit_packet_work) &peer->endpoint_lock rcu_read_lock_bh rcu_read_lock rcu_read_lock_bh rcu_read_lock tk_core.seq.seqcount irq_context: 0 (wq_completion)wg-crypt-wg1#5 (work_completion)(&peer->transmit_packet_work) &peer->endpoint_lock rcu_read_lock_bh rcu_read_lock rcu_read_lock_bh rcu_read_lock &obj_hash[i].lock irq_context: 0 (wq_completion)wg-crypt-wg1#5 (work_completion)(&peer->transmit_packet_work) &peer->endpoint_lock rcu_read_lock_bh rcu_read_lock rcu_read_lock_bh rcu_read_lock &c->lock irq_context: 0 (wq_completion)wg-crypt-wg1#5 (work_completion)(&peer->transmit_packet_work) &peer->endpoint_lock rcu_read_lock_bh rcu_read_lock rcu_read_lock_bh rcu_read_lock quarantine_lock irq_context: 0 rtnl_mutex (work_completion)(&(&br->gc_work)->work) irq_context: 0 (wq_completion)wg-crypt-wg1#5 (work_completion)(&peer->transmit_packet_work) &peer->endpoint_lock rcu_read_lock_bh rcu_read_lock rcu_read_lock_bh rcu_read_lock &n->list_lock irq_context: 0 (wq_completion)wg-crypt-wg1#5 (work_completion)(&peer->transmit_packet_work) &peer->endpoint_lock rcu_read_lock_bh rcu_read_lock rcu_read_lock_bh rcu_read_lock &n->list_lock &c->lock irq_context: 0 (wq_completion)wg-crypt-wg1#5 (work_completion)(&peer->transmit_packet_work) &rq->__lock irq_context: 0 (wq_completion)wg-crypt-wg1#5 (work_completion)(&peer->transmit_packet_work) &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)wg-crypt-wg2#4 (work_completion)(&({ do { const void *__vpp_verify = (typeof((worker) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __asm__ ("" : "=r"(__ptr) : "0"((typeof(*((worker))) *)((worker)))); (typeof((typeof(*((worker))) *)((worker)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); })->work) &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)wg-crypt-wg1#5 (work_completion)(&peer->transmit_packet_work) &peer->endpoint_lock rcu_read_lock_bh rcu_read_lock rcu_read_lock_bh rcu_read_lock &____s->seqcount#2 irq_context: 0 (wq_completion)wg-crypt-wg1#5 (work_completion)(&peer->transmit_packet_work) &peer->endpoint_lock rcu_read_lock_bh rcu_read_lock rcu_read_lock_bh rcu_read_lock &____s->seqcount irq_context: 0 (wq_completion)wg-crypt-wg1#5 (work_completion)(&peer->transmit_packet_work) &peer->endpoint_lock rcu_read_lock_bh rcu_read_lock rcu_read_lock_bh rcu_read_lock pool_lock#2 irq_context: 0 rcu_read_lock rcu_read_lock_bh rcu_read_lock &____s->seqcount#2 irq_context: 0 rcu_read_lock rcu_read_lock_bh rcu_read_lock &____s->seqcount irq_context: 0 key#23 irq_context: 0 rtnl_mutex &idev->mc_lock &bridge_netdev_addr_lock_key &obj_hash[i].lock irq_context: 0 rtnl_mutex &idev->mc_lock &bridge_netdev_addr_lock_key krc.lock irq_context: 0 &sb->s_type->i_mutex_key#8 &sbi->s_writepages_rwsem jbd2_handle &ei->i_data_sem rcu_read_lock &rq->__lock irq_context: 0 &sb->s_type->i_mutex_key#8 &sbi->s_writepages_rwsem jbd2_handle &ei->i_data_sem rcu_read_lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 rtnl_mutex (&brmctx->ip4_mc_router_timer) irq_context: 0 rtnl_mutex (&brmctx->ip4_other_query.timer) irq_context: 0 rtnl_mutex (&brmctx->ip4_own_query.timer) irq_context: 0 rtnl_mutex (&brmctx->ip6_mc_router_timer) irq_context: 0 rtnl_mutex (&brmctx->ip6_other_query.timer) irq_context: 0 rtnl_mutex (&brmctx->ip6_own_query.timer) irq_context: 0 (wq_completion)wg-crypt-wg1#5 (work_completion)(&peer->transmit_packet_work) fill_pool_map-wait-type-override &rq->__lock irq_context: 0 (wq_completion)wg-crypt-wg1#5 (work_completion)(&peer->transmit_packet_work) fill_pool_map-wait-type-override &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 rtnl_mutex (work_completion)(&br->mcast_gc_work) irq_context: 0 rtnl_mutex rcu_state.barrier_mutex irq_context: 0 rtnl_mutex rcu_state.barrier_mutex &rq->__lock irq_context: 0 rtnl_mutex rcu_state.barrier_mutex &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 rtnl_mutex rcu_state.barrier_mutex rcu_state.barrier_lock irq_context: 0 rtnl_mutex rcu_state.barrier_mutex rcu_state.barrier_lock &obj_hash[i].lock irq_context: 0 rtnl_mutex rcu_state.barrier_mutex &x->wait#24 irq_context: 0 (wq_completion)wg-crypt-wg1#5 (work_completion)(&peer->transmit_packet_work) &peer->endpoint_lock rcu_read_lock_bh batched_entropy_u32.lock crngs.lock irq_context: 0 &sb->s_type->i_mutex_key#8 quarantine_lock irq_context: 0 (wq_completion)bat_events (work_completion)(&(&bat_priv->nc.work)->work) &rq->__lock &cfs_rq->removed.lock irq_context: 0 (wq_completion)wg-crypt-wg1#5 (work_completion)(&peer->transmit_packet_work) rcu_node_0 irq_context: 0 (wq_completion)bond1#2 (work_completion)(&(&slave->notify_work)->work) rtnl_mutex net_rwsem irq_context: 0 (wq_completion)bond1#2 (work_completion)(&(&slave->notify_work)->work) rtnl_mutex net_rwsem &list->lock#2 irq_context: 0 (wq_completion)bond1#2 (work_completion)(&(&slave->notify_work)->work) rtnl_mutex &tn->lock irq_context: 0 (wq_completion)bond1#2 (work_completion)(&(&slave->notify_work)->work) rtnl_mutex.wait_lock irq_context: 0 (wq_completion)bond1#2 (work_completion)(&(&slave->notify_work)->work) &p->pi_lock irq_context: 0 &sb->s_type->i_mutex_key#8 &sbi->s_writepages_rwsem &n->list_lock irq_context: 0 &sb->s_type->i_mutex_key#8 &sbi->s_writepages_rwsem &n->list_lock &c->lock irq_context: softirq rcu_read_lock rcu_read_lock rcu_read_lock_bh rcu_read_lock &pool->lock &p->pi_lock &rq->__lock &cfs_rq->removed.lock irq_context: softirq (&in_dev->mr_ifc_timer) rcu_read_lock &im->lock rcu_read_lock &____s->seqcount#2 irq_context: softirq (&in_dev->mr_ifc_timer) rcu_read_lock &im->lock rcu_read_lock &____s->seqcount irq_context: softirq (&peer->timer_send_keepalive) &____s->seqcount#2 irq_context: softirq (&peer->timer_send_keepalive) &____s->seqcount irq_context: 0 (wq_completion)wg-crypt-wg2#2 (work_completion)(&peer->transmit_packet_work) &peer->endpoint_lock rcu_read_lock_bh rcu_read_lock rcu_read_lock_bh rcu_read_lock tk_core.seq.seqcount irq_context: 0 (wq_completion)wg-crypt-wg2#2 (work_completion)(&peer->transmit_packet_work) &peer->endpoint_lock rcu_read_lock_bh rcu_read_lock rcu_read_lock_bh rcu_read_lock pool_lock#2 irq_context: 0 (wq_completion)wg-crypt-wg2#2 (work_completion)(&peer->transmit_packet_work) &peer->endpoint_lock rcu_read_lock_bh rcu_read_lock rcu_read_lock_bh rcu_read_lock rlock-AF_PACKET irq_context: 0 (wq_completion)wg-crypt-wg2 (work_completion)(&peer->transmit_packet_work) &peer->endpoint_lock rcu_read_lock_bh rcu_read_lock rcu_read_lock_bh rcu_read_lock tk_core.seq.seqcount irq_context: 0 (wq_completion)wg-crypt-wg2 (work_completion)(&peer->transmit_packet_work) &peer->endpoint_lock rcu_read_lock_bh rcu_read_lock rcu_read_lock_bh rcu_read_lock &obj_hash[i].lock irq_context: 0 (wq_completion)wg-crypt-wg0#2 (work_completion)(&({ do { const void *__vpp_verify = (typeof((worker) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __asm__ ("" : "=r"(__ptr) : "0"((typeof(*((worker))) *)((worker)))); (typeof((typeof(*((worker))) *)((worker)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); })->work) &rq->__lock irq_context: 0 (wq_completion)wg-crypt-wg0#2 (work_completion)(&({ do { const void *__vpp_verify = (typeof((worker) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __asm__ ("" : "=r"(__ptr) : "0"((typeof(*((worker))) *)((worker)))); (typeof((typeof(*((worker))) *)((worker)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); })->work) &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 rtnl_mutex &root->kernfs_rwsem &root->kernfs_iattr_rwsem &sem->wait_lock irq_context: softirq (&in_dev->mr_ifc_timer) batched_entropy_u32.lock crngs.lock irq_context: 0 (wq_completion)rcu_gp (work_completion)(&sdp->work) &cfs_rq->removed.lock irq_context: 0 &f->f_pos_lock sb_writers#4 &sb->s_type->i_mutex_key#8 jbd2_handle &ei->i_data_sem &lg->lg_mutex key#3 irq_context: 0 rtnl_mutex &wg->device_update_lock stock_lock irq_context: 0 rtnl_mutex &wg->device_update_lock rcu_read_lock pool_lock#2 irq_context: 0 rtnl_mutex &wg->device_update_lock k-sk_lock-AF_INET &____s->seqcount#8 irq_context: 0 rtnl_mutex &wg->device_update_lock k-sk_lock-AF_INET batched_entropy_u32.lock irq_context: 0 rtnl_mutex &wg->device_update_lock fs_reclaim &rq->__lock irq_context: 0 rtnl_mutex &wg->device_update_lock fs_reclaim &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 rtnl_mutex &wg->device_update_lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 &sbi->s_writepages_rwsem &folio_wait_table[i] irq_context: 0 &sbi->s_writepages_rwsem &rq->__lock irq_context: 0 &sbi->s_writepages_rwsem &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 &xa->xa_lock#9 &n->list_lock irq_context: 0 &ret->b_state_lock &journal->j_list_lock rcu_read_lock &xa->xa_lock#9 rcu_read_lock per_cpu_ptr(&cgroup_rstat_cpu_lock, cpu) irq_context: 0 (wq_completion)wg-crypt-wg1#5 (work_completion)(&peer->transmit_packet_work) &peer->endpoint_lock rcu_read_lock_bh rcu_read_lock &c->lock irq_context: 0 (wq_completion)wg-crypt-wg1#5 (work_completion)(&peer->transmit_packet_work) &peer->endpoint_lock rcu_read_lock_bh rcu_read_lock &dir->lock#2 irq_context: 0 (wq_completion)wg-crypt-wg1#5 (work_completion)(&peer->transmit_packet_work) &peer->endpoint_lock rcu_read_lock_bh rcu_read_lock &ul->lock irq_context: 0 (wq_completion)ext4-rsv-conversion (work_completion)(&ei->i_rsv_conversion_work) jbd2_handle &journal->j_wait_updates &p->pi_lock irq_context: 0 (wq_completion)ext4-rsv-conversion (work_completion)(&ei->i_rsv_conversion_work) jbd2_handle &journal->j_wait_updates &p->pi_lock &rq->__lock irq_context: 0 (wq_completion)ext4-rsv-conversion (work_completion)(&ei->i_rsv_conversion_work) jbd2_handle &journal->j_wait_updates &p->pi_lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)ext4-rsv-conversion (work_completion)(&ei->i_rsv_conversion_work) jbd2_handle rcu_node_0 irq_context: 0 (wq_completion)ext4-rsv-conversion (work_completion)(&ei->i_rsv_conversion_work) remove_cache_srcu irq_context: 0 (wq_completion)ext4-rsv-conversion (work_completion)(&ei->i_rsv_conversion_work) remove_cache_srcu quarantine_lock irq_context: 0 (wq_completion)ext4-rsv-conversion (work_completion)(&ei->i_rsv_conversion_work) remove_cache_srcu &c->lock irq_context: 0 (wq_completion)ext4-rsv-conversion (work_completion)(&ei->i_rsv_conversion_work) remove_cache_srcu &n->list_lock irq_context: 0 (wq_completion)ext4-rsv-conversion (work_completion)(&ei->i_rsv_conversion_work) remove_cache_srcu rcu_read_lock pool_lock#2 irq_context: 0 (wq_completion)ext4-rsv-conversion (work_completion)(&ei->i_rsv_conversion_work) remove_cache_srcu &obj_hash[i].lock irq_context: 0 sb_writers#8 &of->mutex kn->active#5 uevent_sock_mutex &____s->seqcount#2 irq_context: 0 sb_writers#4 &sb->s_type->i_mutex_key#8 &____s->seqcount#2 irq_context: 0 sb_writers#4 &sb->s_type->i_mutex_key#8 &____s->seqcount irq_context: 0 (wq_completion)events (work_completion)(&(&hwstats->traffic_dw)->work) &hwstats->hwsdev_list_lock &rq->__lock irq_context: 0 sb_writers#5 &type->i_mutex_dir_key#5/1 rcu_read_lock &rcu_state.expedited_wq irq_context: 0 sb_writers#5 &type->i_mutex_dir_key#5/1 rcu_read_lock &rcu_state.expedited_wq &p->pi_lock irq_context: 0 sb_writers#5 &type->i_mutex_dir_key#5/1 rcu_read_lock &rcu_state.expedited_wq &p->pi_lock &rq->__lock irq_context: 0 sb_writers#5 &type->i_mutex_dir_key#5/1 rcu_read_lock &rcu_state.expedited_wq &p->pi_lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 sb_writers#4 mapping.invalidate_lock &sem->wait_lock irq_context: 0 sb_writers#4 &sb->s_type->i_mutex_key#8 mapping.invalidate_lock jbd2_handle &ei->i_data_sem &bgl->locks[i].lock &obj_hash[i].lock irq_context: 0 sb_writers#4 &sb->s_type->i_mutex_key#8 mapping.invalidate_lock jbd2_handle &ei->i_data_sem &bgl->locks[i].lock pool_lock#2 irq_context: 0 sb_writers#4 &sb->s_type->i_mutex_key#8 &p->pi_lock irq_context: 0 sb_writers#4 &sb->s_type->i_mutex_key#8 &p->pi_lock &cfs_rq->removed.lock irq_context: 0 sb_writers#4 &sb->s_type->i_mutex_key#8 &p->pi_lock &rq->__lock irq_context: 0 sb_writers#4 &sb->s_type->i_mutex_key#8 &p->pi_lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)ext4-rsv-conversion (work_completion)(&ei->i_rsv_conversion_work) jbd2_handle &ei->i_data_sem &ei->i_es_lock key#7 irq_context: 0 sb_writers#4 &sb->s_type->i_mutex_key#8 jbd2_handle &ei->i_data_sem irq_context: 0 sb_writers#4 &sb->s_type->i_mutex_key#8 jbd2_handle &ei->i_data_sem mmu_notifier_invalidate_range_start irq_context: 0 sb_writers#4 &sb->s_type->i_mutex_key#8 jbd2_handle &ei->i_data_sem pool_lock#2 irq_context: 0 sb_writers#4 &sb->s_type->i_mutex_key#8 jbd2_handle &ei->i_data_sem &ei->i_es_lock irq_context: 0 sb_writers#4 &sb->s_type->i_mutex_key#8 jbd2_handle &ei->i_data_sem &ei->i_es_lock irq_context: 0 sb_writers#4 &sb->s_type->i_mutex_key#8 jbd2_handle &ei->i_data_sem &ei->i_es_lock pool_lock#2 irq_context: 0 sb_writers#4 &sb->s_type->i_mutex_key#8 jbd2_handle &ei->i_data_sem &ei->i_es_lock &sbi->s_es_lock irq_context: 0 sb_writers#4 &sb->s_type->i_mutex_key#8 jbd2_handle &ei->i_data_sem &obj_hash[i].lock irq_context: 0 sb_writers#4 mapping.invalidate_lock batched_entropy_u8.lock irq_context: 0 sb_writers#4 mapping.invalidate_lock kfence_freelist_lock irq_context: 0 (wq_completion)ext4-rsv-conversion (work_completion)(&ei->i_rsv_conversion_work) jbd2_handle &ei->i_data_sem rcu_read_lock rcu_node_0 irq_context: 0 (wq_completion)ext4-rsv-conversion (work_completion)(&ei->i_rsv_conversion_work) jbd2_handle &ei->i_data_sem rcu_read_lock &rq->__lock irq_context: 0 (wq_completion)ext4-rsv-conversion (work_completion)(&ei->i_rsv_conversion_work) jbd2_handle &ei->i_data_sem rcu_read_lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 &f->f_pos_lock sb_writers#4 &sb->s_type->i_mutex_key#8 jbd2_handle mmu_notifier_invalidate_range_start irq_context: 0 &f->f_pos_lock sb_writers#4 &sb->s_type->i_mutex_key#8 jbd2_handle &____s->seqcount irq_context: 0 &f->f_pos_lock sb_writers#4 &sb->s_type->i_mutex_key#8 jbd2_handle pool_lock#2 irq_context: 0 (wq_completion)events (work_completion)(&(&nsim_dev->trap_data->trap_report_dw)->work) &devlink->lock_key#5 rcu_node_0 irq_context: 0 &f->f_pos_lock sb_writers#4 &sb->s_type->i_mutex_key#8 jbd2_handle rcu_read_lock pool_lock#2 irq_context: 0 &f->f_pos_lock sb_writers#4 &sb->s_type->i_mutex_key#8 jbd2_handle &obj_hash[i].lock irq_context: 0 sb_writers#4 sb_writers#4 &____s->seqcount#2 irq_context: 0 sb_writers#4 sb_writers#4 &____s->seqcount irq_context: 0 &f->f_pos_lock sb_writers#4 &sb->s_type->i_mutex_key#8 &ei->i_data_sem rcu_node_0 irq_context: 0 sb_writers#4 mapping.invalidate_lock &pcp->lock &zone->lock irq_context: 0 sb_writers#4 sb_writers#4 rcu_read_lock rcu_node_0 irq_context: 0 sb_writers#4 sb_writers#4 rcu_read_lock &rq->__lock irq_context: 0 sb_writers#4 sb_writers#4 rcu_read_lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 sb_writers#4 sb_writers#4 rcu_read_lock &rcu_state.expedited_wq irq_context: 0 sb_writers#4 sb_writers#4 rcu_read_lock &rcu_state.expedited_wq &p->pi_lock irq_context: 0 sb_writers#4 sb_writers#4 rcu_read_lock &rcu_state.expedited_wq &p->pi_lock &rq->__lock irq_context: 0 sb_writers#4 sb_writers#4 rcu_read_lock &rcu_state.expedited_wq &p->pi_lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 &sb->s_type->i_mutex_key#8 &sbi->s_writepages_rwsem jbd2_handle &journal->j_wait_updates &p->pi_lock &cfs_rq->removed.lock irq_context: 0 &sb->s_type->i_mutex_key#8 &sbi->s_writepages_rwsem jbd2_handle &ei->i_data_sem &ei->i_es_lock batched_entropy_u8.lock irq_context: 0 &sb->s_type->i_mutex_key#8 &sbi->s_writepages_rwsem jbd2_handle &ei->i_data_sem &ei->i_es_lock kfence_freelist_lock irq_context: 0 (wq_completion)ext4-rsv-conversion (work_completion)(&ei->i_rsv_conversion_work) jbd2_handle &ei->i_data_sem &ei->i_es_lock &meta->lock irq_context: 0 (wq_completion)ext4-rsv-conversion (work_completion)(&ei->i_rsv_conversion_work) jbd2_handle &ei->i_data_sem &ei->i_es_lock kfence_freelist_lock irq_context: 0 sb_writers#4 mapping.invalidate_lock pgd_lock irq_context: 0 sb_writers#4 mapping.invalidate_lock key irq_context: 0 sb_writers#4 mapping.invalidate_lock pcpu_lock irq_context: 0 sb_writers#4 mapping.invalidate_lock percpu_counters_lock irq_context: 0 sb_writers#4 mapping.invalidate_lock pcpu_lock stock_lock irq_context: 0 (wq_completion)rcu_gp (work_completion)(&(&ssp->srcu_sup->work)->work) tracepoint_srcu_srcu_usage.lock &ACCESS_PRIVATE(sdp, lock) irq_context: 0 &fsnotify_mark_srcu remove_cache_srcu irq_context: 0 &fsnotify_mark_srcu remove_cache_srcu quarantine_lock irq_context: 0 &fsnotify_mark_srcu remove_cache_srcu &c->lock irq_context: 0 &fsnotify_mark_srcu remove_cache_srcu &n->list_lock irq_context: 0 &fsnotify_mark_srcu remove_cache_srcu rcu_read_lock pool_lock#2 irq_context: 0 &fsnotify_mark_srcu remove_cache_srcu &obj_hash[i].lock irq_context: 0 sb_writers#8 remove_cache_srcu &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 rcu_read_lock &obj_hash[i].lock pool_lock irq_context: 0 sb_writers#5 &type->i_mutex_dir_key#5/1 fs_reclaim mmu_notifier_invalidate_range_start &rq->__lock irq_context: 0 &f->f_pos_lock sb_writers#4 &sb->s_type->i_mutex_key#8 &mm->mmap_lock rcu_read_lock &cfs_rq->removed.lock irq_context: 0 &f->f_pos_lock sb_writers#4 &sb->s_type->i_mutex_key#8 &mm->mmap_lock rcu_read_lock &obj_hash[i].lock irq_context: 0 sb_writers#4 &sb->s_type->i_mutex_key#8 mapping.invalidate_lock &c->lock irq_context: 0 sb_writers#4 &sb->s_type->i_mutex_key#8 mapping.invalidate_lock &journal->j_state_lock irq_context: 0 sb_writers#4 &sb->s_type->i_mutex_key#8 mapping.invalidate_lock &journal->j_state_lock tk_core.seq.seqcount irq_context: 0 sb_writers#4 &sb->s_type->i_mutex_key#8 mapping.invalidate_lock &journal->j_state_lock &obj_hash[i].lock irq_context: 0 sb_writers#4 &sb->s_type->i_mutex_key#8 mapping.invalidate_lock &journal->j_state_lock &base->lock irq_context: 0 sb_writers#4 &sb->s_type->i_mutex_key#8 mapping.invalidate_lock &journal->j_state_lock &base->lock &obj_hash[i].lock irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem &net->ipv6.ip6addrlbl_table.lock &obj_hash[i].lock pool_lock irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem &net->ipv6.ip6addrlbl_table.lock krc.lock &obj_hash[i].lock irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem &net->ipv6.ip6addrlbl_table.lock krc.lock &base->lock irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem &net->ipv6.ip6addrlbl_table.lock krc.lock &base->lock &obj_hash[i].lock irq_context: 0 sb_writers#4 &sb->s_type->i_mutex_key#8 mapping.invalidate_lock quarantine_lock irq_context: 0 sb_writers#4 &sb->s_type->i_mutex_key#8 mapping.invalidate_lock &meta->lock irq_context: 0 sb_writers#4 &sb->s_type->i_mutex_key#8 mapping.invalidate_lock kfence_freelist_lock irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem sysctl_lock krc.lock &obj_hash[i].lock irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem sysctl_lock krc.lock &base->lock irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem sysctl_lock krc.lock &base->lock &obj_hash[i].lock irq_context: 0 sb_writers#4 &sb->s_type->i_mutex_key#8 mapping.invalidate_lock &pcp->lock &zone->lock irq_context: 0 sb_writers#4 &sb->s_type->i_mutex_key#8 mapping.invalidate_lock &pcp->lock &zone->lock &____s->seqcount irq_context: 0 sb_writers#4 &sb->s_type->i_mutex_key#8 mapping.invalidate_lock remove_cache_srcu &pcp->lock &zone->lock irq_context: 0 sb_writers#4 &sb->s_type->i_mutex_key#8 mapping.invalidate_lock remove_cache_srcu &pcp->lock &zone->lock &____s->seqcount irq_context: 0 sb_writers#4 &sb->s_type->i_mutex_key#8 mapping.invalidate_lock remove_cache_srcu pool_lock#2 irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem fill_pool_map-wait-type-override pool_lock#2 irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem fill_pool_map-wait-type-override &c->lock irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem fill_pool_map-wait-type-override pool_lock irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem uevent_sock_mutex rcu_read_lock rcu_node_0 irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem uevent_sock_mutex rcu_read_lock &rq->__lock irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem &pcp->lock &zone->lock irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem &pcp->lock &zone->lock &____s->seqcount irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem remove_cache_srcu &rq->__lock irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem remove_cache_srcu &pcp->lock &zone->lock irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem remove_cache_srcu &pcp->lock &zone->lock &____s->seqcount irq_context: 0 sb_writers#4 sb_internal jbd2_handle key#4 irq_context: 0 sb_writers#5 &type->i_mutex_dir_key#5/1 &sb->s_type->i_mutex_key#12 rcu_node_0 irq_context: softirq rcu_read_lock hwsim_radio_lock init_task.mems_allowed_seq.seqcount irq_context: softirq rcu_read_lock rcu_read_lock init_task.mems_allowed_seq.seqcount irq_context: softirq (&hsr->announce_timer) rcu_read_lock &hsr->seqnr_lock rcu_read_lock init_task.mems_allowed_seq.seqcount irq_context: 0 (wq_completion)events free_ipc_work rcu_state.exp_mutex &cfs_rq->removed.lock irq_context: 0 (wq_completion)events free_ipc_work rcu_state.exp_mutex pool_lock#2 irq_context: 0 sb_writers#11 &type->i_mutex_dir_key#7/1 cgroup_mutex rcu_read_lock rcu_node_0 irq_context: 0 sb_writers#11 &type->i_mutex_dir_key#7/1 cgroup_mutex rcu_read_lock &rq->__lock irq_context: 0 sb_writers#11 &type->i_mutex_dir_key#7/1 cgroup_mutex rcu_read_lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 rcu_read_lock &nf_conntrack_locks[i]/1 irq_context: 0 lock link_idr_lock &c->lock irq_context: 0 lock link_idr_lock &n->list_lock irq_context: 0 lock link_idr_lock &____s->seqcount#2 irq_context: 0 lock link_idr_lock &pcp->lock &zone->lock irq_context: 0 lock link_idr_lock &pcp->lock &zone->lock &____s->seqcount irq_context: 0 lock link_idr_lock &____s->seqcount irq_context: 0 sb_writers#11 &type->i_mutex_dir_key#7/1 &dentry->d_lock &dentry->d_lock &lru->node[i].lock irq_context: 0 sb_writers#11 &type->i_mutex_dir_key#7/1 rcu_read_lock rcu_node_0 irq_context: 0 sb_writers#11 &type->i_mutex_dir_key#7/1 rcu_read_lock &rq->__lock irq_context: 0 sb_writers#11 &type->i_mutex_dir_key#7/1 rcu_read_lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 sb_writers#11 &type->i_mutex_dir_key#7/1 &root->kernfs_rwsem inode_hash_lock irq_context: 0 sb_writers#11 &type->i_mutex_dir_key#7/1 &root->kernfs_rwsem fs_reclaim irq_context: 0 sb_writers#11 &type->i_mutex_dir_key#7/1 &root->kernfs_rwsem fs_reclaim mmu_notifier_invalidate_range_start irq_context: 0 sb_writers#11 &type->i_mutex_dir_key#7/1 &root->kernfs_rwsem fs_reclaim mmu_notifier_invalidate_range_start &rq->__lock irq_context: 0 sb_writers#11 &type->i_mutex_dir_key#7/1 &root->kernfs_rwsem fs_reclaim mmu_notifier_invalidate_range_start &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 sb_writers#11 &type->i_mutex_dir_key#7/1 &root->kernfs_rwsem stock_lock irq_context: 0 sb_writers#11 &type->i_mutex_dir_key#7/1 &root->kernfs_rwsem pool_lock#2 irq_context: 0 sb_writers#11 &type->i_mutex_dir_key#7/1 &root->kernfs_rwsem mmu_notifier_invalidate_range_start irq_context: 0 sb_writers#11 &type->i_mutex_dir_key#7/1 &root->kernfs_rwsem inode_hash_lock &sb->s_type->i_lock_key#31 irq_context: 0 sb_writers#11 &type->i_mutex_dir_key#7/1 &root->kernfs_rwsem inode_hash_lock &s->s_inode_list_lock irq_context: 0 sb_writers#11 &type->i_mutex_dir_key#7/1 &root->kernfs_rwsem tk_core.seq.seqcount irq_context: 0 sb_writers#11 &type->i_mutex_dir_key#7/1 &root->kernfs_rwsem &sb->s_type->i_lock_key#31 irq_context: 0 sb_writers#11 &type->i_mutex_dir_key#7/1 &sb->s_type->i_lock_key#31 irq_context: 0 sb_writers#11 &type->i_mutex_dir_key#7/1 &sb->s_type->i_lock_key#31 &dentry->d_lock irq_context: 0 (wq_completion)wg-crypt-wg1#6 (work_completion)(&({ do { const void *__vpp_verify = (typeof((worker) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __asm__ ("" : "=r"(__ptr) : "0"((typeof(*((worker))) *)((worker)))); (typeof((typeof(*((worker))) *)((worker)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); })->work) &rq->__lock irq_context: 0 (wq_completion)wg-crypt-wg1#6 (work_completion)(&({ do { const void *__vpp_verify = (typeof((worker) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __asm__ ("" : "=r"(__ptr) : "0"((typeof(*((worker))) *)((worker)))); (typeof((typeof(*((worker))) *)((worker)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); })->work) &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)wg-crypt-wg1#6 (work_completion)(&peer->transmit_packet_work) &peer->endpoint_lock rcu_read_lock_bh batched_entropy_u32.lock crngs.lock irq_context: 0 sk_lock-AF_INET rcu_read_lock &n->list_lock irq_context: 0 sk_lock-AF_INET rcu_read_lock &n->list_lock &c->lock irq_context: 0 rcu_read_lock &____s->seqcount#10 irq_context: 0 (wq_completion)wg-crypt-wg1#5 (work_completion)(&peer->transmit_packet_work) &peer->endpoint_lock rcu_read_lock_bh rcu_read_lock &obj_hash[i].lock irq_context: 0 (wq_completion)wg-crypt-wg1#6 (work_completion)(&peer->transmit_packet_work) &rq->__lock irq_context: 0 (wq_completion)wg-crypt-wg1#5 (work_completion)(&({ do { const void *__vpp_verify = (typeof((worker) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __asm__ ("" : "=r"(__ptr) : "0"((typeof(*((worker))) *)((worker)))); (typeof((typeof(*((worker))) *)((worker)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); })->work) rcu_read_lock &pool->lock pool_lock#2 irq_context: 0 (wq_completion)wg-crypt-wg1#6 (work_completion)(&peer->transmit_packet_work) &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)wg-crypt-wg2#6 (work_completion)(&({ do { const void *__vpp_verify = (typeof((worker) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __asm__ ("" : "=r"(__ptr) : "0"((typeof(*((worker))) *)((worker)))); (typeof((typeof(*((worker))) *)((worker)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); })->work) &rq->__lock irq_context: 0 (wq_completion)wg-crypt-wg2#6 (work_completion)(&({ do { const void *__vpp_verify = (typeof((worker) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __asm__ ("" : "=r"(__ptr) : "0"((typeof(*((worker))) *)((worker)))); (typeof((typeof(*((worker))) *)((worker)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); })->work) &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 dup_mmap_sem &mm->mmap_lock &mm->mmap_lock/1 &rq->__lock &cfs_rq->removed.lock irq_context: 0 dup_mmap_sem &mm->mmap_lock &mm->mmap_lock/1 &anon_vma->rwsem rcu_read_lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 rtnl_mutex lock kernfs_idr_lock rcu_read_lock pool_lock#2 irq_context: 0 rtnl_mutex &bridge_netdev_addr_lock_key &c->lock irq_context: 0 rtnl_mutex &bridge_netdev_addr_lock_key &n->list_lock irq_context: 0 rtnl_mutex &bridge_netdev_addr_lock_key &n->list_lock &c->lock irq_context: 0 rtnl_mutex noop_qdisc.q.lock batched_entropy_u64.lock irq_context: 0 (wq_completion)mld (work_completion)(&(&idev->mc_ifc_work)->work) &idev->mc_lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)mld (work_completion)(&(&idev->mc_ifc_work)->work) &idev->mc_lock rcu_read_lock rcu_read_lock rcu_read_lock_bh rcu_read_lock &br->multicast_lock &____s->seqcount#2 irq_context: 0 (wq_completion)mld (work_completion)(&(&idev->mc_ifc_work)->work) &idev->mc_lock rcu_read_lock rcu_read_lock rcu_read_lock_bh rcu_read_lock &obj_hash[i].lock pool_lock irq_context: 0 (wq_completion)wg-crypt-wg1#6 (work_completion)(&peer->transmit_packet_work) &peer->endpoint_lock rcu_read_lock_bh rcu_read_lock &obj_hash[i].lock irq_context: 0 cgroup_threadgroup_rwsem freezer_mutex rcu_read_lock rcu_node_0 irq_context: 0 cgroup_threadgroup_rwsem freezer_mutex rcu_read_lock &rq->__lock irq_context: 0 cgroup_threadgroup_rwsem freezer_mutex rcu_read_lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 rtnl_mutex &sch->q.lock batched_entropy_u64.lock irq_context: 0 (wq_completion)events (linkwatch_work).work &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 &sb->s_type->i_mutex_key#8 &sbi->s_writepages_rwsem jbd2_handle rcu_read_lock &rcu_state.expedited_wq irq_context: 0 tomoyo_ss mount_lock irq_context: 0 tomoyo_ss mount_lock rcu_read_lock rename_lock.seqcount irq_context: 0 cb_lock nlk_cb_mutex-GENERIC rcu_read_lock rcu_node_0 irq_context: 0 cb_lock nlk_cb_mutex-GENERIC rcu_read_lock &rq->__lock irq_context: 0 cb_lock nlk_cb_mutex-GENERIC rcu_read_lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)events (work_completion)(&pwq->unbound_release_work) &cfs_rq->removed.lock irq_context: 0 (wq_completion)events (work_completion)(&pwq->unbound_release_work) rcu_state.exp_mutex &pool->lock &p->pi_lock &rq->__lock &cfs_rq->removed.lock irq_context: 0 &sb->s_type->i_mutex_key#8 &sbi->s_writepages_rwsem jbd2_handle rcu_read_lock &rcu_state.expedited_wq &p->pi_lock irq_context: 0 &sb->s_type->i_mutex_key#8 &sbi->s_writepages_rwsem jbd2_handle rcu_read_lock &rcu_state.expedited_wq &p->pi_lock &rq->__lock irq_context: 0 &sb->s_type->i_mutex_key#8 &sbi->s_writepages_rwsem jbd2_handle rcu_read_lock &rcu_state.expedited_wq &p->pi_lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 namespace_sem &n->list_lock irq_context: 0 namespace_sem &n->list_lock &c->lock irq_context: 0 (wq_completion)rcu_gp (work_completion)(&(&ssp->srcu_sup->work)->work) &ssp->srcu_sup->srcu_gp_mutex &cfs_rq->removed.lock irq_context: 0 (wq_completion)rcu_gp (work_completion)(&(&ssp->srcu_sup->work)->work) &ssp->srcu_sup->srcu_gp_mutex &obj_hash[i].lock irq_context: 0 sb_writers#4 &type->i_mutex_dir_key#3/1 tomoyo_ss remove_cache_srcu rcu_read_lock rcu_node_0 irq_context: 0 sb_writers#4 &type->i_mutex_dir_key#3/1 tomoyo_ss remove_cache_srcu rcu_read_lock &rq->__lock irq_context: 0 sb_writers#4 &type->i_mutex_dir_key#3/1 tomoyo_ss remove_cache_srcu rcu_read_lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 rtnl_mutex remove_cache_srcu pool_lock#2 irq_context: 0 &iint->mutex rcu_read_lock rcu_node_0 irq_context: 0 rtnl_mutex &block->cb_lock &rq->__lock irq_context: 0 rtnl_mutex &block->cb_lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 rcu_read_lock &dentry->d_lock &p->pi_lock irq_context: 0 rcu_read_lock &dentry->d_lock &p->pi_lock &rq->__lock irq_context: 0 rcu_read_lock &dentry->d_lock &p->pi_lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 &vma->vm_lock->lock rcu_read_lock rcu_read_lock rcu_node_0 irq_context: 0 &vma->vm_lock->lock rcu_read_lock rcu_read_lock &rq->__lock irq_context: 0 &vma->vm_lock->lock rcu_read_lock rcu_read_lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 &sb->s_type->i_mutex_key#12 &rq->__lock irq_context: 0 sb_writers#5 &type->i_mutex_dir_key#5/1 rcu_node_0 irq_context: 0 rtnl_mutex rcu_read_lock rcu_read_lock_bh rcu_read_lock &obj_hash[i].lock irq_context: 0 rtnl_mutex pcpu_alloc_mutex &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)bat_events (work_completion)(&(&forw_packet_aggr->delayed_work)->work) rcu_read_lock_bh rcu_read_lock &____s->seqcount#2 irq_context: 0 (wq_completion)bat_events (work_completion)(&(&forw_packet_aggr->delayed_work)->work) rcu_read_lock_bh rcu_read_lock &____s->seqcount irq_context: 0 (wq_completion)mld (work_completion)(&(&idev->mc_ifc_work)->work) &idev->mc_lock rcu_read_lock rcu_read_lock rcu_read_lock_bh &sch->q.lock irq_context: 0 (wq_completion)mld (work_completion)(&(&idev->mc_ifc_work)->work) &idev->mc_lock rcu_read_lock rcu_read_lock rcu_read_lock_bh &sch->q.lock tk_core.seq.seqcount irq_context: 0 &mm->mmap_lock rcu_read_lock &rq->__lock &cfs_rq->removed.lock irq_context: 0 (wq_completion)wg-crypt-wg1#5 &rq->__lock irq_context: 0 (wq_completion)wg-crypt-wg1#5 &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)wg-crypt-wg0#2 (work_completion)(&peer->transmit_packet_work) &peer->endpoint_lock rcu_read_lock_bh rcu_read_lock rcu_read_lock_bh rcu_read_lock tk_core.seq.seqcount irq_context: 0 (wq_completion)wg-crypt-wg0#2 (work_completion)(&peer->transmit_packet_work) &peer->endpoint_lock rcu_read_lock_bh rcu_read_lock rcu_read_lock_bh rcu_read_lock &obj_hash[i].lock irq_context: 0 (wq_completion)wg-crypt-wg1#5 (work_completion)(&({ do { const void *__vpp_verify = (typeof((worker) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __asm__ ("" : "=r"(__ptr) : "0"((typeof(*((worker))) *)((worker)))); (typeof((typeof(*((worker))) *)((worker)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); })->work) rcu_read_lock &pool->lock &p->pi_lock &rq->__lock &cfs_rq->removed.lock irq_context: 0 (wq_completion)wg-crypt-wg1#5 (work_completion)(&peer->transmit_packet_work) &peer->endpoint_lock rcu_read_lock_bh rcu_read_lock rcu_read_lock_bh rcu_read_lock &pcp->lock &zone->lock irq_context: 0 rtnl_mutex &idev->mc_lock _xmit_ETHER &____s->seqcount#2 irq_context: softirq &(&bat_priv->dat.work)->timer rcu_read_lock &pool->lock/1 &p->pi_lock &rq->__lock irq_context: softirq &(&bat_priv->dat.work)->timer rcu_read_lock &pool->lock/1 &p->pi_lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: softirq (&ndev->rs_timer) rcu_read_lock rcu_read_lock rcu_read_lock_bh rcu_read_lock rlock-AF_PACKET irq_context: softirq (&ndev->rs_timer) rcu_read_lock rcu_read_lock rcu_read_lock_bh dev->qdisc_tx_busylock ?: &qdisc_tx_busylock rcu_read_lock pool_lock#2 irq_context: softirq (&ndev->rs_timer) rcu_read_lock rcu_read_lock rcu_read_lock_bh dev->qdisc_tx_busylock ?: &qdisc_tx_busylock rcu_read_lock tk_core.seq.seqcount irq_context: softirq (&ndev->rs_timer) rcu_read_lock rcu_read_lock rcu_read_lock_bh dev->qdisc_tx_busylock ?: &qdisc_tx_busylock rcu_read_lock &obj_hash[i].lock irq_context: softirq (&ndev->rs_timer) rcu_read_lock rcu_read_lock rcu_read_lock_bh rcu_read_lock rcu_read_lock_bh rcu_read_lock pool_lock#2 irq_context: softirq (&ndev->rs_timer) rcu_read_lock rcu_read_lock rcu_read_lock_bh rcu_read_lock rcu_read_lock_bh rcu_read_lock &obj_hash[i].lock irq_context: softirq (&ndev->rs_timer) rcu_read_lock rcu_read_lock rcu_read_lock_bh dev->qdisc_tx_busylock ?: &qdisc_tx_busylock _xmit_ETHER#2 rcu_read_lock tk_core.seq.seqcount irq_context: softirq (&ndev->rs_timer) rcu_read_lock rcu_read_lock rcu_read_lock_bh dev->qdisc_tx_busylock ?: &qdisc_tx_busylock _xmit_ETHER#2 rcu_read_lock &obj_hash[i].lock irq_context: softirq (&ndev->rs_timer) rcu_read_lock rcu_read_lock rcu_read_lock_bh _xmit_TUNNEL6#2 rcu_read_lock tk_core.seq.seqcount irq_context: softirq (&ndev->rs_timer) rcu_read_lock rcu_read_lock rcu_read_lock_bh _xmit_TUNNEL6#2 rcu_read_lock &obj_hash[i].lock irq_context: softirq rcu_read_lock rcu_read_lock rcu_read_lock rcu_read_lock_bh rcu_read_lock &c->lock irq_context: softirq rcu_read_lock rcu_read_lock rcu_read_lock rcu_read_lock_bh rcu_read_lock pool_lock#2 irq_context: softirq rcu_read_lock rcu_read_lock rcu_read_lock rcu_read_lock_bh rcu_read_lock &obj_hash[i].lock irq_context: softirq (&in_dev->mr_ifc_timer) rcu_read_lock rcu_read_lock_bh &sch->q.lock irq_context: softirq (&in_dev->mr_ifc_timer) rcu_read_lock rcu_read_lock_bh &sch->q.lock tk_core.seq.seqcount irq_context: 0 (wq_completion)wg-crypt-wg2#3 (work_completion)(&peer->transmit_packet_work) &peer->endpoint_lock rcu_read_lock_bh rcu_read_lock rcu_read_lock_bh rcu_read_lock tk_core.seq.seqcount irq_context: 0 (wq_completion)wg-crypt-wg2#2 (work_completion)(&peer->transmit_packet_work) &peer->endpoint_lock rcu_read_lock_bh rcu_read_lock rcu_read_lock_bh rcu_read_lock &obj_hash[i].lock irq_context: 0 (wq_completion)wg-crypt-wg2#3 (work_completion)(&peer->transmit_packet_work) &peer->endpoint_lock rcu_read_lock_bh rcu_read_lock rcu_read_lock_bh rcu_read_lock rlock-AF_PACKET irq_context: 0 (wq_completion)wg-crypt-wg0#3 (work_completion)(&({ do { const void *__vpp_verify = (typeof((worker) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __asm__ ("" : "=r"(__ptr) : "0"((typeof(*((worker))) *)((worker)))); (typeof((typeof(*((worker))) *)((worker)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); })->work) &c->lock irq_context: 0 (wq_completion)wg-crypt-wg0#3 (work_completion)(&({ do { const void *__vpp_verify = (typeof((worker) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __asm__ ("" : "=r"(__ptr) : "0"((typeof(*((worker))) *)((worker)))); (typeof((typeof(*((worker))) *)((worker)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); })->work) &____s->seqcount#2 irq_context: 0 (wq_completion)wg-crypt-wg0#3 (work_completion)(&({ do { const void *__vpp_verify = (typeof((worker) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __asm__ ("" : "=r"(__ptr) : "0"((typeof(*((worker))) *)((worker)))); (typeof((typeof(*((worker))) *)((worker)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); })->work) &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)wg-crypt-wg0#3 (work_completion)(&({ do { const void *__vpp_verify = (typeof((worker) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __asm__ ("" : "=r"(__ptr) : "0"((typeof(*((worker))) *)((worker)))); (typeof((typeof(*((worker))) *)((worker)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); })->work) &____s->seqcount irq_context: 0 (wq_completion)wg-crypt-wg0#3 (work_completion)(&({ do { const void *__vpp_verify = (typeof((worker) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __asm__ ("" : "=r"(__ptr) : "0"((typeof(*((worker))) *)((worker)))); (typeof((typeof(*((worker))) *)((worker)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); })->work) pool_lock#2 irq_context: 0 sb_writers &sb->s_type->i_mutex_key#4 &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 sb_writers &sb->s_type->i_mutex_key#4 &cfs_rq->removed.lock irq_context: 0 sb_writers &sb->s_type->i_mutex_key#4 &obj_hash[i].lock irq_context: 0 (wq_completion)events (work_completion)(&(&nsim_dev->trap_data->trap_report_dw)->work) &devlink->lock_key#6 &nsim_trap_data->trap_lock batched_entropy_u8.lock crngs.lock irq_context: 0 (wq_completion)bat_events (work_completion)(&(&bat_priv->bla.work)->work) rcu_node_0 irq_context: 0 (wq_completion)bat_events (work_completion)(&(&bat_priv->bla.work)->work) rcu_read_lock rcu_node_0 irq_context: 0 (wq_completion)wg-crypt-wg1#5 (work_completion)(&peer->transmit_packet_work) &cfs_rq->removed.lock irq_context: 0 (wq_completion)wg-crypt-wg1#5 (work_completion)(&peer->transmit_packet_work) pool_lock#2 irq_context: 0 (wq_completion)bat_events (work_completion)(&(&bat_priv->dat.work)->work) &cfs_rq->removed.lock irq_context: 0 (wq_completion)bat_events (work_completion)(&(&bat_priv->dat.work)->work) pool_lock#2 irq_context: 0 (wq_completion)bat_events (work_completion)(&(&bat_priv->bla.work)->work) rcu_read_lock &rq->__lock irq_context: 0 (wq_completion)bat_events (work_completion)(&(&bat_priv->bla.work)->work) rcu_read_lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)wg-crypt-wg0#3 (work_completion)(&peer->transmit_packet_work) &peer->endpoint_lock rcu_read_lock_bh rcu_read_lock rcu_read_lock_bh rcu_read_lock tk_core.seq.seqcount irq_context: 0 (wq_completion)wg-crypt-wg0#3 (work_completion)(&peer->transmit_packet_work) &peer->endpoint_lock rcu_read_lock_bh rcu_read_lock rcu_read_lock_bh rcu_read_lock rlock-AF_PACKET irq_context: softirq rcu_read_lock rcu_read_lock k-slock-AF_INET6 rcu_read_lock rcu_read_lock rcu_read_lock_bh rcu_read_lock pool_lock#2 irq_context: softirq rcu_read_lock rcu_read_lock k-slock-AF_INET6 rcu_read_lock rcu_read_lock rcu_read_lock_bh rcu_read_lock tk_core.seq.seqcount irq_context: softirq rcu_read_lock rcu_read_lock k-slock-AF_INET6 rcu_read_lock rcu_read_lock rcu_read_lock_bh rcu_read_lock &obj_hash[i].lock irq_context: 0 rtnl_mutex _xmit_PHONET_PIPE irq_context: 0 &f->f_pos_lock &type->i_mutex_dir_key#4 &p->pi_lock &cfs_rq->removed.lock irq_context: 0 pernet_ops_rwsem &p->pi_lock &rq->__lock &cfs_rq->removed.lock irq_context: 0 (wq_completion)wg-crypt-wg0#5 (work_completion)(&peer->transmit_packet_work) &peer->endpoint_lock rcu_read_lock_bh rcu_read_lock rcu_read_lock_bh rcu_read_lock tk_core.seq.seqcount irq_context: 0 (wq_completion)wg-crypt-wg0#5 (work_completion)(&peer->transmit_packet_work) &peer->endpoint_lock rcu_read_lock_bh rcu_read_lock rcu_read_lock_bh rcu_read_lock &obj_hash[i].lock irq_context: 0 &f->f_pos_lock &type->i_mutex_dir_key#3 sb_writers#4 rcu_read_lock rcu_node_0 irq_context: 0 &f->f_pos_lock &type->i_mutex_dir_key#3 sb_writers#4 rcu_read_lock &rq->__lock irq_context: 0 &f->f_pos_lock &type->i_mutex_dir_key#3 sb_writers#4 rcu_read_lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 sb_writers#4 &type->i_mutex_dir_key#3/1 &type->i_mutex_dir_key#3 jbd2_handle rcu_read_lock rcu_node_0 irq_context: 0 sb_writers#4 &type->i_mutex_dir_key#3/1 &type->i_mutex_dir_key#3 jbd2_handle rcu_read_lock &rq->__lock irq_context: 0 sb_writers#4 &type->i_mutex_dir_key#3/1 &type->i_mutex_dir_key#3 jbd2_handle rcu_read_lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 &____s->seqcount#11 irq_context: 0 sk_lock-AF_INET6 &ping_table.lock irq_context: 0 sk_lock-AF_INET6 rcu_read_lock &nf_conntrack_locks[i] &nf_conntrack_locks[i]/1 irq_context: 0 sk_lock-AF_INET6 rcu_read_lock &nf_conntrack_locks[i] &nf_conntrack_locks[i]/1 batched_entropy_u8.lock irq_context: softirq rcu_read_lock rcu_read_lock k-slock-AF_INET6 &____s->seqcount irq_context: softirq rcu_read_lock rcu_read_lock k-slock-AF_INET6 pool_lock#2 irq_context: softirq rcu_read_lock rcu_read_lock k-slock-AF_INET6 &obj_hash[i].lock irq_context: softirq rcu_read_lock rcu_read_lock k-slock-AF_INET6 rcu_read_lock &____s->seqcount#7 irq_context: softirq rcu_read_lock rcu_read_lock k-slock-AF_INET6 rcu_read_lock &____s->seqcount#10 irq_context: softirq rcu_read_lock rcu_read_lock k-slock-AF_INET6 rcu_read_lock rcu_read_lock_bh tk_core.seq.seqcount irq_context: softirq rcu_read_lock rcu_read_lock k-slock-AF_INET6 rcu_read_lock rcu_read_lock_bh &list->lock#5 irq_context: softirq rcu_read_lock rcu_read_lock rlock-AF_INET6 irq_context: 0 &sb->s_type->i_mutex_key#10 &ping_table.lock irq_context: 0 sb_writers#5 &type->i_mutex_dir_key#5/1 tomoyo_ss batched_entropy_u8.lock irq_context: 0 sb_writers#5 &type->i_mutex_dir_key#5/1 tomoyo_ss kfence_freelist_lock irq_context: 0 sb_writers#5 &type->i_mutex_dir_key#5/1 tomoyo_ss &meta->lock irq_context: 0 pernet_ops_rwsem rtnl_mutex &root->kernfs_rwsem &cfs_rq->removed.lock irq_context: 0 pernet_ops_rwsem rtnl_mutex &root->kernfs_rwsem &rq->__lock &cfs_rq->removed.lock irq_context: 0 sb_writers#4 &type->i_mutex_dir_key#3/1 tomoyo_ss rcu_read_lock rename_lock irq_context: 0 (wq_completion)events (work_completion)(&p->wq) &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 cgroup_threadgroup_rwsem rcu_node_0 irq_context: 0 cgroup_threadgroup_rwsem &rcu_state.expedited_wq irq_context: 0 cgroup_threadgroup_rwsem &rcu_state.expedited_wq &p->pi_lock irq_context: 0 cgroup_threadgroup_rwsem &rcu_state.expedited_wq &p->pi_lock &rq->__lock irq_context: 0 cgroup_threadgroup_rwsem &rcu_state.expedited_wq &p->pi_lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: softirq rcu_callback put_task_map-wait-type-override per_cpu_ptr(&cgroup_rstat_cpu_lock, cpu) irq_context: 0 sk_lock-AF_ALG fs_reclaim irq_context: 0 sk_lock-AF_ALG fs_reclaim mmu_notifier_invalidate_range_start irq_context: 0 sk_lock-AF_ALG pool_lock#2 irq_context: 0 sk_lock-AF_ALG &mm->mmap_lock irq_context: 0 sk_lock-AF_ALG &obj_hash[i].lock irq_context: 0 sk_lock-AF_ALG &dir->lock irq_context: 0 sk_lock-AF_CAIF &obj_hash[i].lock pool_lock irq_context: 0 vlan_ioctl_mutex rtnl_mutex _xmit_ETHER irq_context: 0 vlan_ioctl_mutex rtnl_mutex _xmit_ETHER &local->filter_lock irq_context: 0 vlan_ioctl_mutex rtnl_mutex _xmit_ETHER &local->filter_lock pool_lock#2 irq_context: 0 vlan_ioctl_mutex rtnl_mutex _xmit_ETHER rcu_read_lock &pool->lock/1 irq_context: 0 vlan_ioctl_mutex rtnl_mutex _xmit_ETHER rcu_read_lock &pool->lock/1 &obj_hash[i].lock irq_context: 0 vlan_ioctl_mutex rtnl_mutex _xmit_ETHER rcu_read_lock &pool->lock/1 &p->pi_lock irq_context: 0 vlan_ioctl_mutex rtnl_mutex _xmit_ETHER rcu_read_lock &pool->lock/1 &p->pi_lock &rq->__lock irq_context: 0 vlan_ioctl_mutex rtnl_mutex _xmit_ETHER rcu_read_lock &pool->lock/1 &p->pi_lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 sk_lock-AF_ALG &ei->socket.wq.wait irq_context: 0 pernet_ops_rwsem rtnl_mutex subsys mutex#17 &rq->__lock irq_context: 0 pernet_ops_rwsem rtnl_mutex subsys mutex#17 &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 sk_lock-AF_ALG &mm->mmap_lock &sem->wait_lock irq_context: 0 sk_lock-AF_ALG &mm->mmap_lock rcu_read_lock rcu_node_0 irq_context: 0 (wq_completion)ext4-rsv-conversion (work_completion)(&ei->i_rsv_conversion_work) jbd2_handle &ei->i_data_sem &ei->i_es_lock &c->lock irq_context: 0 (wq_completion)ipv6_addrconf (work_completion)(&(&ifa->dad_work)->work) rtnl_mutex rcu_read_lock rcu_read_lock rcu_read_lock_bh &sch->q.lock irq_context: 0 (wq_completion)ipv6_addrconf (work_completion)(&(&ifa->dad_work)->work) rtnl_mutex rcu_read_lock rcu_read_lock rcu_read_lock_bh &sch->q.lock tk_core.seq.seqcount irq_context: 0 sk_lock-AF_ALG rcu_read_lock &ei->socket.wq.wait irq_context: 0 sk_lock-AF_ALG rcu_read_lock &ei->socket.wq.wait &p->pi_lock irq_context: 0 sk_lock-AF_ALG rcu_read_lock &ei->socket.wq.wait &p->pi_lock &rq->__lock irq_context: 0 sk_lock-AF_ALG rcu_read_lock &ei->socket.wq.wait &p->pi_lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 sk_lock-AF_ALG &mm->mmap_lock ptlock_ptr(page)#2 irq_context: 0 sk_lock-AF_ALG &mm->mmap_lock sb_pagefaults irq_context: 0 sk_lock-AF_ALG &mm->mmap_lock sb_pagefaults tk_core.seq.seqcount irq_context: 0 sk_lock-AF_ALG &mm->mmap_lock sb_pagefaults mmu_notifier_invalidate_range_start irq_context: 0 sk_lock-AF_ALG &mm->mmap_lock sb_pagefaults pool_lock#2 irq_context: 0 sk_lock-AF_ALG &mm->mmap_lock sb_pagefaults &journal->j_state_lock irq_context: 0 sk_lock-AF_ALG &mm->mmap_lock sb_pagefaults jbd2_handle irq_context: 0 sk_lock-AF_ALG &mm->mmap_lock sb_pagefaults jbd2_handle &ei->i_raw_lock irq_context: 0 sk_lock-AF_ALG &mm->mmap_lock sb_pagefaults jbd2_handle &journal->j_wait_updates irq_context: 0 sk_lock-AF_ALG &mm->mmap_lock sb_pagefaults &obj_hash[i].lock irq_context: 0 sk_lock-AF_ALG &mm->mmap_lock sb_pagefaults mapping.invalidate_lock irq_context: 0 sk_lock-AF_ALG &mm->mmap_lock sb_pagefaults mapping.invalidate_lock &mapping->private_lock irq_context: 0 sk_lock-AF_ALG &mm->mmap_lock sb_pagefaults mapping.invalidate_lock &mapping->private_lock rcu_read_lock &memcg->move_lock irq_context: 0 sk_lock-AF_ALG &mm->mmap_lock &mapping->private_lock irq_context: 0 sk_lock-AF_ALG &mm->mmap_lock &mapping->private_lock rcu_read_lock &memcg->move_lock irq_context: 0 sk_lock-AF_ALG &c->lock irq_context: 0 sk_lock-AF_ALG rcu_read_lock pool_lock#2 irq_context: 0 sk_lock-AF_ALG rcu_read_lock &ei->socket.wq.wait &p->pi_lock &cfs_rq->removed.lock irq_context: 0 sk_lock-AF_ALG &mm->mmap_lock sb_pagefaults mapping.invalidate_lock rcu_read_lock &memcg->move_lock irq_context: 0 sk_lock-AF_ALG &mm->mmap_lock sb_pagefaults mapping.invalidate_lock rcu_read_lock &xa->xa_lock#9 irq_context: 0 sk_lock-AF_ALG &mm->mmap_lock &folio_wait_table[i] irq_context: 0 (wq_completion)events (work_completion)(&pwq->unbound_release_work) &obj_hash[i].lock pool_lock irq_context: 0 sk_lock-AF_ALG &____s->seqcount#2 irq_context: 0 sk_lock-AF_ALG &____s->seqcount irq_context: 0 sk_lock-AF_ALG rcu_read_lock rcu_node_0 irq_context: 0 sk_lock-AF_ALG rcu_read_lock &rq->__lock irq_context: 0 sk_lock-AF_ALG rcu_read_lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 sk_lock-AF_ALG slock-AF_ALG &sk->sk_lock.wq irq_context: 0 sk_lock-AF_ALG &rq->__lock irq_context: 0 sk_lock-AF_ALG &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 sk_lock-AF_ALG rcu_read_lock &rcu_state.gp_wq irq_context: 0 sk_lock-AF_ALG rcu_read_lock &rcu_state.gp_wq &p->pi_lock irq_context: 0 sk_lock-AF_ALG rcu_read_lock &rcu_state.gp_wq &p->pi_lock &rq->__lock irq_context: 0 sk_lock-AF_ALG rcu_read_lock &rcu_state.gp_wq &p->pi_lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 slock-AF_ALG &sk->sk_lock.wq irq_context: 0 slock-AF_ALG &sk->sk_lock.wq &p->pi_lock irq_context: 0 slock-AF_ALG &sk->sk_lock.wq &p->pi_lock &rq->__lock irq_context: 0 slock-AF_ALG &sk->sk_lock.wq &p->pi_lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 sk_lock-AF_ALG &n->list_lock irq_context: 0 sk_lock-AF_ALG &n->list_lock &c->lock irq_context: 0 sk_lock-AF_ALG &mm->mmap_lock &rq->__lock irq_context: 0 sk_lock-AF_ALG &mm->mmap_lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 sk_lock-AF_ALG &mm->mmap_lock rcu_node_0 irq_context: 0 sk_lock-AF_ALG &mm->mmap_lock sb_pagefaults remove_cache_srcu irq_context: 0 sk_lock-AF_ALG &mm->mmap_lock sb_pagefaults remove_cache_srcu quarantine_lock irq_context: 0 sk_lock-AF_ALG &mm->mmap_lock sb_pagefaults remove_cache_srcu &c->lock irq_context: 0 sk_lock-AF_ALG &mm->mmap_lock sb_pagefaults remove_cache_srcu &n->list_lock irq_context: 0 sk_lock-AF_ALG &mm->mmap_lock sb_pagefaults remove_cache_srcu &rq->__lock irq_context: 0 sk_lock-AF_ALG &mm->mmap_lock sb_pagefaults remove_cache_srcu &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 sk_lock-AF_ALG &mm->mmap_lock sb_pagefaults remove_cache_srcu &obj_hash[i].lock irq_context: 0 sk_lock-AF_ALG &mm->mmap_lock sb_pagefaults remove_cache_srcu pool_lock#2 irq_context: 0 sk_lock-AF_ALG &mm->mmap_lock sb_pagefaults remove_cache_srcu rcu_read_lock pool_lock#2 irq_context: 0 sk_lock-AF_ALG &sem->wait_lock irq_context: 0 sk_lock-AF_ALG &p->pi_lock irq_context: 0 sk_lock-AF_ALG &p->pi_lock &cfs_rq->removed.lock irq_context: 0 sk_lock-AF_ALG &p->pi_lock &rq->__lock irq_context: 0 sk_lock-AF_ALG &p->pi_lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 &mm->mmap_lock sb_pagefaults mapping.invalidate_lock rcu_read_lock &xa->xa_lock#9 key#10 irq_context: 0 &ep->mtx stock_lock irq_context: 0 &ep->mtx &n->list_lock irq_context: 0 sk_lock-AF_ALG rcu_read_lock &ei->socket.wq.wait &ep->lock irq_context: 0 pernet_ops_rwsem rtnl_mutex rcu_state.exp_mutex rcu_read_lock &rq->__lock irq_context: 0 sk_lock-AF_ALG remove_cache_srcu irq_context: 0 sk_lock-AF_ALG remove_cache_srcu quarantine_lock irq_context: 0 sk_lock-AF_ALG remove_cache_srcu &c->lock irq_context: 0 sk_lock-AF_ALG remove_cache_srcu &n->list_lock irq_context: 0 sk_lock-AF_ALG remove_cache_srcu rcu_read_lock pool_lock#2 irq_context: 0 sk_lock-AF_ALG remove_cache_srcu &obj_hash[i].lock irq_context: 0 sk_lock-AF_ALG &mm->mmap_lock sb_pagefaults &rq->__lock irq_context: 0 sk_lock-AF_ALG &mm->mmap_lock sb_pagefaults &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 dgram_lock irq_context: 0 sk_lock-AF_ALG &mm->mmap_lock sb_pagefaults mapping.invalidate_lock rcu_read_lock &xa->xa_lock#9 key#10 irq_context: 0 sk_lock-AF_BLUETOOTH-BTPROTO_L2CAP &mm->mmap_lock irq_context: 0 bt_proto_lock &____s->seqcount#2 irq_context: 0 rtnl_mutex stack_depot_init_mutex &rq->__lock irq_context: 0 rtnl_mutex stack_depot_init_mutex &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 sk_lock-AF_ALG &mm->mmap_lock rcu_read_lock &rq->__lock irq_context: 0 sk_lock-AF_ALG &mm->mmap_lock rcu_read_lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 sk_lock-AF_ALG &mm->mmap_lock rcu_read_lock &rcu_state.gp_wq irq_context: 0 sk_lock-AF_ALG &mm->mmap_lock rcu_read_lock &rcu_state.gp_wq &p->pi_lock irq_context: 0 sk_lock-AF_ALG &mm->mmap_lock rcu_read_lock &rcu_state.gp_wq &p->pi_lock &rq->__lock irq_context: 0 sk_lock-AF_ALG &mm->mmap_lock rcu_read_lock &rcu_state.gp_wq &p->pi_lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 &sb->s_type->i_mutex_key#8 &sbi->s_writepages_rwsem jbd2_handle &ei->i_data_sem &rcu_state.expedited_wq &p->pi_lock &rq->__lock irq_context: 0 pernet_ops_rwsem rtnl_mutex rcu_state.exp_mutex rcu_read_lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)rcu_gp (work_completion)(&rew->rew_work) rcu_state.exp_wake_mutex &rnp->exp_wq[0] &p->pi_lock &cfs_rq->removed.lock irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem rtnl_mutex &rnp->exp_lock irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem rtnl_mutex rcu_state.exp_mutex irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem rtnl_mutex rcu_state.exp_mutex rcu_state.exp_mutex.wait_lock irq_context: 0 &wg->device_update_lock &rnp->exp_lock irq_context: 0 &wg->device_update_lock &rnp->exp_wq[0] irq_context: 0 &sb->s_type->i_mutex_key#8 &sbi->s_writepages_rwsem jbd2_handle &ei->i_data_sem &ei->i_es_lock &n->list_lock irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem rtnl_mutex &idev->mc_lock rcu_node_0 irq_context: 0 &wg->device_update_lock rcu_state.exp_mutex rcu_read_lock &rq->__lock irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem rtnl_mutex &idev->mc_lock &rcu_state.expedited_wq irq_context: 0 &wg->device_update_lock rcu_state.exp_mutex rcu_read_lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem rtnl_mutex &idev->mc_lock &rcu_state.expedited_wq &p->pi_lock irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem rtnl_mutex &idev->mc_lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem &p->pi_lock &rq->__lock &cfs_rq->removed.lock irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem rtnl_mutex &root->kernfs_rwsem &rq->__lock irq_context: 0 &sb->s_type->i_mutex_key#10 dgram_lock irq_context: 0 &sb->s_type->i_mutex_key#10 l2cap_sk_list.lock irq_context: 0 &sb->s_type->i_mutex_key#10 sk_lock-AF_BLUETOOTH-BTPROTO_L2CAP irq_context: 0 &wg->device_update_lock rcu_state.exp_mutex irq_context: 0 &wg->device_update_lock rcu_state.exp_mutex rcu_state.exp_mutex.wait_lock irq_context: 0 (wq_completion)events (work_completion)(&pwq->unbound_release_work) &rnp->exp_wq[1] irq_context: 0 (wq_completion)wg-crypt-wg2#6 (work_completion)(&({ do { const void *__vpp_verify = (typeof((worker) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __asm__ ("" : "=r"(__ptr) : "0"((typeof(*((worker))) *)((worker)))); (typeof((typeof(*((worker))) *)((worker)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); })->work) rcu_read_lock &pool->lock &p->pi_lock &rq->__lock irq_context: 0 (wq_completion)wg-crypt-wg2#6 (work_completion)(&({ do { const void *__vpp_verify = (typeof((worker) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __asm__ ("" : "=r"(__ptr) : "0"((typeof(*((worker))) *)((worker)))); (typeof((typeof(*((worker))) *)((worker)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); })->work) rcu_read_lock &pool->lock &p->pi_lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 &sb->s_type->i_mutex_key#8 &sbi->s_writepages_rwsem jbd2_handle &ei->i_data_sem remove_cache_srcu irq_context: 0 &sb->s_type->i_mutex_key#8 &sbi->s_writepages_rwsem jbd2_handle &ei->i_data_sem remove_cache_srcu quarantine_lock irq_context: 0 &sb->s_type->i_mutex_key#8 &sbi->s_writepages_rwsem jbd2_handle &ei->i_data_sem remove_cache_srcu &c->lock irq_context: 0 &sb->s_type->i_mutex_key#8 &sbi->s_writepages_rwsem jbd2_handle &ei->i_data_sem remove_cache_srcu &n->list_lock irq_context: 0 &sb->s_type->i_mutex_key#8 &sbi->s_writepages_rwsem jbd2_handle &ei->i_data_sem remove_cache_srcu rcu_read_lock pool_lock#2 irq_context: 0 &sb->s_type->i_mutex_key#8 &sbi->s_writepages_rwsem jbd2_handle &ei->i_data_sem remove_cache_srcu &obj_hash[i].lock irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem rtnl_mutex rcu_state.exp_mutex.wait_lock irq_context: 0 &sb->s_type->i_mutex_key#8 &sbi->s_writepages_rwsem jbd2_handle &ei->i_data_sem &ei->i_es_lock &n->list_lock &c->lock irq_context: 0 sk_lock-AF_ALG &mm->mmap_lock sb_pagefaults jbd2_handle &mapping->private_lock irq_context: 0 &sb->s_type->i_mutex_key#10 sk_lock-AF_BLUETOOTH-BTPROTO_L2CAP slock-AF_BLUETOOTH-BTPROTO_L2CAP irq_context: 0 &sb->s_type->i_mutex_key#10 slock-AF_BLUETOOTH-BTPROTO_L2CAP irq_context: 0 &sb->s_type->i_mutex_key#10 &chan->lock/1 irq_context: 0 &sb->s_type->i_mutex_key#10 &chan->lock/1 sk_lock-AF_BLUETOOTH-BTPROTO_L2CAP/1 irq_context: 0 &sb->s_type->i_mutex_key#10 &chan->lock/1 sk_lock-AF_BLUETOOTH-BTPROTO_L2CAP/1 slock-AF_BLUETOOTH-BTPROTO_L2CAP irq_context: 0 &sb->s_type->i_mutex_key#10 &chan->lock/1 slock-AF_BLUETOOTH-BTPROTO_L2CAP irq_context: 0 &sb->s_type->i_mutex_key#10 &chan->lock/1 clock-AF_BLUETOOTH irq_context: 0 &sb->s_type->i_mutex_key#10 &chan->lock/1 rlock-AF_BLUETOOTH irq_context: 0 &sb->s_type->i_mutex_key#10 &chan->lock/1 wlock-AF_BLUETOOTH irq_context: 0 &sb->s_type->i_mutex_key#10 &chan->lock/1 pool_lock#2 irq_context: 0 &sb->s_type->i_mutex_key#10 &chan->lock/1 &dir->lock irq_context: 0 &sb->s_type->i_mutex_key#10 &chan->lock/1 &obj_hash[i].lock irq_context: 0 &sb->s_type->i_mutex_key#10 chan_list_lock irq_context: 0 sk_lock-AF_ALG &mm->mmap_lock sb_pagefaults mapping.invalidate_lock &rq->__lock irq_context: 0 sk_lock-AF_ALG &mm->mmap_lock sb_pagefaults mapping.invalidate_lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)events_unbound (reaper_work).work rcu_node_0 irq_context: 0 (wq_completion)events_unbound (reaper_work).work &rcu_state.expedited_wq irq_context: 0 (wq_completion)events_unbound (reaper_work).work &rcu_state.expedited_wq &p->pi_lock irq_context: 0 sb_writers#4 &type->i_mutex_dir_key#3/1 &type->i_mutex_dir_key#3 jbd2_handle rcu_read_lock &rcu_state.expedited_wq irq_context: 0 sb_writers#4 &type->i_mutex_dir_key#3/1 &type->i_mutex_dir_key#3 jbd2_handle rcu_read_lock &rcu_state.expedited_wq &p->pi_lock irq_context: 0 sb_writers#4 &type->i_mutex_dir_key#3/1 &type->i_mutex_dir_key#3 jbd2_handle rcu_read_lock &rcu_state.expedited_wq &p->pi_lock &rq->__lock irq_context: 0 sb_writers#4 &type->i_mutex_dir_key#3/1 &type->i_mutex_dir_key#3 jbd2_handle rcu_read_lock &rcu_state.expedited_wq &p->pi_lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 pernet_ops_rwsem rtnl_mutex &caifn->caifdevs.lock rcu_state.exp_mutex &rnp->exp_wq[0] irq_context: 0 (wq_completion)events (work_completion)(&(&nsim_dev->trap_data->trap_report_dw)->work) &devlink->lock_key#5 &rcu_state.expedited_wq irq_context: 0 (wq_completion)events (work_completion)(&(&nsim_dev->trap_data->trap_report_dw)->work) &devlink->lock_key#5 &rcu_state.expedited_wq &p->pi_lock irq_context: 0 (wq_completion)events (work_completion)(&(&nsim_dev->trap_data->trap_report_dw)->work) &devlink->lock_key#5 &rcu_state.expedited_wq &p->pi_lock &rq->__lock irq_context: 0 (wq_completion)events (work_completion)(&(&nsim_dev->trap_data->trap_report_dw)->work) &devlink->lock_key#5 &rcu_state.expedited_wq &p->pi_lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem rtnl_mutex rcu_read_lock &pool->lock/1 &p->pi_lock &cfs_rq->removed.lock irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem rtnl_mutex &root->kernfs_rwsem rcu_read_lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem batched_entropy_u8.lock irq_context: 0 (wq_completion)ipv6_addrconf (work_completion)(&(&ifa->dad_work)->work) rtnl_mutex &idev->mc_lock remove_cache_srcu pool_lock#2 irq_context: 0 (wq_completion)ipv6_addrconf (work_completion)(&(&ifa->dad_work)->work) rtnl_mutex &idev->mc_lock rcu_read_lock rcu_read_lock rcu_read_lock_bh &sch->q.lock irq_context: 0 (wq_completion)ipv6_addrconf (work_completion)(&(&ifa->dad_work)->work) rtnl_mutex &idev->mc_lock rcu_read_lock rcu_read_lock rcu_read_lock_bh &sch->q.lock tk_core.seq.seqcount irq_context: 0 (wq_completion)ipv6_addrconf (work_completion)(&(&ifa->dad_work)->work) rtnl_mutex rcu_read_lock rcu_read_lock rcu_read_lock_bh &sch->q.lock hrtimer_bases.lock irq_context: 0 (wq_completion)ipv6_addrconf (work_completion)(&(&ifa->dad_work)->work) rtnl_mutex rcu_read_lock rcu_read_lock rcu_read_lock_bh &sch->q.lock hrtimer_bases.lock &obj_hash[i].lock irq_context: 0 (wq_completion)ipv6_addrconf (work_completion)(&(&ifa->dad_work)->work) rtnl_mutex rcu_read_lock rcu_read_lock rcu_read_lock_bh &sch->q.lock hrtimer_bases.lock tk_core.seq.seqcount irq_context: 0 (wq_completion)events (work_completion)(&data->fib_event_work) &data->fib_lock remove_cache_srcu pool_lock#2 irq_context: softirq rcu_read_lock &sch->q.lock irq_context: softirq rcu_read_lock &sch->q.lock tk_core.seq.seqcount irq_context: 0 (wq_completion)mld (work_completion)(&(&idev->mc_dad_work)->work) &idev->mc_lock remove_cache_srcu pool_lock#2 irq_context: 0 (wq_completion)mld (work_completion)(&(&idev->mc_dad_work)->work) &idev->mc_lock rcu_read_lock rcu_read_lock rcu_read_lock_bh &sch->q.lock irq_context: 0 (wq_completion)mld (work_completion)(&(&idev->mc_dad_work)->work) &idev->mc_lock rcu_read_lock rcu_read_lock rcu_read_lock_bh &sch->q.lock tk_core.seq.seqcount irq_context: 0 (wq_completion)mld (work_completion)(&(&idev->mc_dad_work)->work) &idev->mc_lock rcu_read_lock rcu_read_lock rcu_read_lock_bh &sch->q.lock hrtimer_bases.lock irq_context: 0 (wq_completion)mld (work_completion)(&(&idev->mc_dad_work)->work) &idev->mc_lock rcu_read_lock rcu_read_lock rcu_read_lock_bh &sch->q.lock hrtimer_bases.lock &obj_hash[i].lock irq_context: 0 (wq_completion)mld (work_completion)(&(&idev->mc_dad_work)->work) &idev->mc_lock rcu_read_lock rcu_read_lock rcu_read_lock_bh &sch->q.lock hrtimer_bases.lock tk_core.seq.seqcount irq_context: softirq (&ndev->rs_timer) rcu_read_lock rcu_read_lock rcu_read_lock_bh &hsr->seqnr_lock rcu_read_lock &____s->seqcount#2 irq_context: softirq (&ndev->rs_timer) rcu_read_lock rcu_read_lock rcu_read_lock_bh &hsr->seqnr_lock rcu_read_lock &____s->seqcount irq_context: 0 pernet_ops_rwsem sysctl_lock fill_pool_map-wait-type-override pool_lock#2 irq_context: 0 pernet_ops_rwsem sysctl_lock fill_pool_map-wait-type-override pool_lock irq_context: 0 sk_lock-AF_INET6 rcu_read_lock &nf_conntrack_locks[i]/1 irq_context: 0 sk_lock-AF_NETLINK &rq->__lock irq_context: 0 sk_lock-AF_NETLINK &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 &mm->mmap_lock stock_lock rcu_read_lock per_cpu_ptr(&cgroup_rstat_cpu_lock, cpu) irq_context: 0 lock pidmap_lock &n->list_lock &c->lock irq_context: 0 sb_writers#4 sb_internal jbd2_handle &sbi->s_orphan_lock rcu_read_lock rcu_node_0 irq_context: 0 sb_writers#4 sb_internal jbd2_handle &sbi->s_orphan_lock rcu_read_lock &rq->__lock irq_context: 0 sb_writers#4 sb_internal jbd2_handle &sbi->s_orphan_lock rcu_read_lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 pernet_ops_rwsem nf_hook_mutex &rq->__lock irq_context: 0 pernet_ops_rwsem nf_hook_mutex &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 &sighand->siglock per_cpu_ptr(&cgroup_rstat_cpu_lock, cpu) irq_context: 0 sb_writers#4 &fsnotify_mark_srcu &rq->__lock irq_context: 0 sb_writers#4 &fsnotify_mark_srcu &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 rtnl_mutex pcpu_alloc_mutex.wait_lock irq_context: 0 (wq_completion)ipv6_addrconf (work_completion)(&(&ifa->dad_work)->work) rtnl_mutex &idev->mc_lock fs_reclaim &rq->__lock irq_context: 0 (wq_completion)ipv6_addrconf (work_completion)(&(&ifa->dad_work)->work) rtnl_mutex &idev->mc_lock fs_reclaim &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)wg-crypt-wg2#2 (work_completion)(&({ do { const void *__vpp_verify = (typeof((worker) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __asm__ ("" : "=r"(__ptr) : "0"((typeof(*((worker))) *)((worker)))); (typeof((typeof(*((worker))) *)((worker)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); })->work) rcu_read_lock &pool->lock &p->pi_lock &rq->__lock irq_context: 0 (wq_completion)wg-crypt-wg2#2 (work_completion)(&({ do { const void *__vpp_verify = (typeof((worker) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __asm__ ("" : "=r"(__ptr) : "0"((typeof(*((worker))) *)((worker)))); (typeof((typeof(*((worker))) *)((worker)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); })->work) rcu_read_lock &pool->lock &p->pi_lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)wg-crypt-wg2#2 (work_completion)(&peer->transmit_packet_work) &peer->endpoint_lock rcu_read_lock_bh rcu_read_lock &ul->lock irq_context: 0 pernet_ops_rwsem pcpu_alloc_mutex pcpu_alloc_mutex.wait_lock irq_context: 0 (wq_completion)bat_events (work_completion)(&(&bat_priv->dat.work)->work) rcu_node_0 irq_context: 0 &sb->s_type->i_mutex_key#10 sk_lock-AF_CAIF rcu_state.exp_mutex &rnp->exp_wq[1] irq_context: 0 sb_writers#4 &type->i_mutex_dir_key#3/1 &sb->s_type->i_mutex_key#8 rcu_read_lock rcu_node_0 irq_context: 0 sb_writers#4 &type->i_mutex_dir_key#3/1 &sb->s_type->i_mutex_key#8 rcu_read_lock &rq->__lock irq_context: 0 sb_writers#4 &type->i_mutex_dir_key#3/1 &sb->s_type->i_mutex_key#8 rcu_read_lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 sb_writers#4 &type->i_mutex_dir_key#3/1 sb_internal jbd2_handle rcu_read_lock rcu_node_0 irq_context: 0 sb_writers#4 &type->i_mutex_dir_key#3/1 sb_internal jbd2_handle rcu_read_lock &rq->__lock irq_context: 0 sb_writers#4 &type->i_mutex_dir_key#3/1 sb_internal jbd2_handle rcu_read_lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 &sb->s_type->i_mutex_key#10 sk_lock-AF_BLUETOOTH-BTPROTO_L2CAP &rq->__lock irq_context: 0 &sb->s_type->i_mutex_key#10 sk_lock-AF_BLUETOOTH-BTPROTO_L2CAP &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 sk_lock-AF_ALG sk_lock-AF_ALG/1 irq_context: 0 sk_lock-AF_ALG sk_lock-AF_ALG/1 slock-AF_ALG irq_context: softirq (&in_dev->mr_ifc_timer) rcu_read_lock &im->lock &n->list_lock irq_context: softirq (&in_dev->mr_ifc_timer) rcu_read_lock &im->lock &n->list_lock &c->lock irq_context: 0 pernet_ops_rwsem remove_cache_srcu rcu_read_lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 pernet_ops_rwsem cpu_hotplug_lock wq_pool_mutex fill_pool_map-wait-type-override &rq->__lock irq_context: 0 pernet_ops_rwsem cpu_hotplug_lock wq_pool_mutex fill_pool_map-wait-type-override &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 pernet_ops_rwsem rtnl_mutex rcu_node_0 irq_context: 0 &f->f_pos_lock sb_writers#4 &sb->s_type->i_mutex_key#8 jbd2_handle &journal->j_state_lock irq_context: 0 &f->f_pos_lock sb_writers#4 &sb->s_type->i_mutex_key#8 jbd2_handle &journal->j_state_lock &journal->j_wait_commit irq_context: 0 &f->f_pos_lock sb_writers#4 &sb->s_type->i_mutex_key#8 jbd2_handle &journal->j_state_lock &journal->j_wait_commit &p->pi_lock irq_context: 0 &f->f_pos_lock sb_writers#4 &sb->s_type->i_mutex_key#8 jbd2_handle &journal->j_state_lock &journal->j_wait_commit &p->pi_lock &rq->__lock irq_context: 0 &f->f_pos_lock sb_writers#4 &sb->s_type->i_mutex_key#8 jbd2_handle &journal->j_state_lock &journal->j_wait_commit &p->pi_lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 &mm->mmap_lock sb_writers#4 jbd2_handle &mapping->private_lock irq_context: 0 pernet_ops_rwsem k-sk_lock-AF_RXRPC &rxnet->local_mutex &cfs_rq->removed.lock irq_context: 0 (wq_completion)mld (work_completion)(&(&idev->mc_ifc_work)->work) &idev->mc_lock rcu_read_lock &n->list_lock &c->lock irq_context: 0 &mm->mmap_lock &xa->xa_lock#9 &n->list_lock &c->lock irq_context: softirq (&in_dev->mr_ifc_timer) rcu_read_lock &im->lock &____s->seqcount#2 irq_context: softirq (&in_dev->mr_ifc_timer) rcu_read_lock &im->lock &____s->seqcount irq_context: softirq (&in_dev->mr_ifc_timer) rcu_read_lock &im->lock rcu_read_lock &n->list_lock irq_context: softirq (&in_dev->mr_ifc_timer) rcu_read_lock &im->lock rcu_read_lock &n->list_lock &c->lock irq_context: 0 sb_writers#4 sb_internal batched_entropy_u8.lock irq_context: 0 sb_writers#4 sb_internal kfence_freelist_lock irq_context: 0 sb_writers#4 sb_internal &meta->lock irq_context: 0 pernet_ops_rwsem rtnl_mutex &root->kernfs_rwsem rcu_read_lock &rq->__lock irq_context: 0 pernet_ops_rwsem rtnl_mutex &root->kernfs_rwsem rcu_read_lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)events_unbound (reaper_work).work &cfs_rq->removed.lock irq_context: 0 sb_writers#4 &type->i_mutex_dir_key#3/1 remove_cache_srcu pool_lock#2 irq_context: 0 (wq_completion)wg-crypt-wg0#3 (work_completion)(&peer->transmit_packet_work) &peer->endpoint_lock rcu_read_lock_bh rcu_read_lock &obj_hash[i].lock irq_context: 0 (wq_completion)wg-crypt-wg0#3 (work_completion)(&peer->transmit_packet_work) &peer->endpoint_lock rcu_read_lock_bh rcu_read_lock &dir->lock#2 irq_context: 0 (wq_completion)wg-crypt-wg0#3 (work_completion)(&peer->transmit_packet_work) &peer->endpoint_lock rcu_read_lock_bh rcu_read_lock &tbl->lock irq_context: 0 (wq_completion)wg-crypt-wg0#3 (work_completion)(&peer->transmit_packet_work) &peer->endpoint_lock rcu_read_lock_bh rcu_read_lock &n->lock irq_context: 0 (wq_completion)wg-crypt-wg0#3 (work_completion)(&peer->transmit_packet_work) &peer->endpoint_lock rcu_read_lock_bh rcu_read_lock &____s->seqcount#9 irq_context: 0 (wq_completion)wg-crypt-wg0#3 (work_completion)(&peer->transmit_packet_work) &peer->endpoint_lock rcu_read_lock_bh rcu_read_lock rcu_read_lock_bh rcu_read_lock pool_lock#2 irq_context: 0 (wq_completion)wg-crypt-wg0#3 (work_completion)(&peer->transmit_packet_work) &peer->endpoint_lock rcu_read_lock_bh rcu_read_lock rcu_read_lock_bh rcu_read_lock &obj_hash[i].lock irq_context: 0 (wq_completion)wg-crypt-wg0#3 (work_completion)(&peer->transmit_packet_work) &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 vlan_ioctl_mutex rtnl_mutex rcu_read_lock &pool->lock &p->pi_lock irq_context: 0 vlan_ioctl_mutex rtnl_mutex rcu_read_lock &pool->lock &p->pi_lock &rq->__lock irq_context: 0 vlan_ioctl_mutex rtnl_mutex rcu_read_lock &pool->lock &p->pi_lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)ipv6_addrconf (work_completion)(&(&ifa->dad_work)->work) rtnl_mutex rcu_read_lock rcu_read_lock rcu_read_lock_bh rcu_read_lock rlock-AF_PACKET irq_context: 0 pernet_ops_rwsem rtnl_mutex &idev->mc_lock &____s->seqcount#2 irq_context: 0 pernet_ops_rwsem cpu_hotplug_lock wq_pool_mutex fs_reclaim mmu_notifier_invalidate_range_start &rq->__lock irq_context: 0 pernet_ops_rwsem cpu_hotplug_lock wq_pool_mutex fs_reclaim mmu_notifier_invalidate_range_start &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 pernet_ops_rwsem k-sk_lock-AF_RXRPC &rxnet->local_mutex rcu_node_0 irq_context: 0 pernet_ops_rwsem k-sk_lock-AF_RXRPC &rxnet->local_mutex rcu_read_lock rcu_node_0 irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem uevent_sock_mutex rcu_read_lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem uevent_sock_mutex rcu_read_lock &rcu_state.gp_wq irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem uevent_sock_mutex rcu_read_lock &rcu_state.gp_wq &p->pi_lock irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem uevent_sock_mutex rcu_read_lock &rcu_state.gp_wq &p->pi_lock &rq->__lock irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem uevent_sock_mutex rcu_read_lock &rcu_state.gp_wq &p->pi_lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: softirq &(&kfence_timer)->timer rcu_read_lock &pool->lock/1 fill_pool_map-wait-type-override &c->lock irq_context: softirq &(&kfence_timer)->timer rcu_read_lock &pool->lock/1 fill_pool_map-wait-type-override &n->list_lock irq_context: softirq &(&kfence_timer)->timer rcu_read_lock &pool->lock/1 fill_pool_map-wait-type-override &n->list_lock &c->lock irq_context: softirq &(&kfence_timer)->timer rcu_read_lock &pool->lock/1 fill_pool_map-wait-type-override pool_lock irq_context: softirq (&ndev->rs_timer) rcu_read_lock rcu_read_lock rcu_read_lock_bh dev->qdisc_tx_busylock ?: &qdisc_tx_busylock#2 rcu_read_lock tk_core.seq.seqcount irq_context: softirq (&ndev->rs_timer) rcu_read_lock rcu_read_lock rcu_read_lock_bh dev->qdisc_tx_busylock ?: &qdisc_tx_busylock#2 rcu_read_lock &obj_hash[i].lock irq_context: softirq (&ndev->rs_timer) rcu_read_lock rcu_read_lock rcu_read_lock_bh dev->qdisc_tx_busylock ?: &qdisc_tx_busylock#2 rcu_read_lock_bh rcu_read_lock pool_lock#2 irq_context: softirq (&ndev->rs_timer) &ndev->lock fill_pool_map-wait-type-override &c->lock irq_context: softirq (&ndev->rs_timer) rcu_read_lock rcu_read_lock rcu_read_lock_bh dev->qdisc_tx_busylock ?: &qdisc_tx_busylock#2 rcu_read_lock_bh rcu_read_lock &obj_hash[i].lock irq_context: softirq (&ndev->rs_timer) &ndev->lock fill_pool_map-wait-type-override &n->list_lock irq_context: softirq (&ndev->rs_timer) &ndev->lock fill_pool_map-wait-type-override &n->list_lock &c->lock irq_context: softirq (&ndev->rs_timer) rcu_read_lock rcu_read_lock rcu_read_lock_bh rcu_read_lock rcu_read_lock_bh rcu_read_lock &c->lock irq_context: softirq (&ndev->rs_timer) rcu_read_lock rcu_read_lock rcu_read_lock_bh rcu_read_lock rcu_read_lock_bh rcu_read_lock &n->list_lock irq_context: softirq (&ndev->rs_timer) rcu_read_lock rcu_read_lock rcu_read_lock_bh rcu_read_lock rcu_read_lock_bh rcu_read_lock &n->list_lock &c->lock irq_context: softirq (&ndev->rs_timer) rcu_read_lock rcu_read_lock rcu_read_lock_bh rcu_read_lock_bh rcu_read_lock pool_lock#2 irq_context: softirq (&ndev->rs_timer) rcu_read_lock rcu_read_lock rcu_read_lock_bh rcu_read_lock rcu_read_lock_bh rcu_read_lock &____s->seqcount irq_context: softirq (&ndev->rs_timer) rcu_read_lock rcu_read_lock rcu_read_lock_bh rcu_read_lock_bh rcu_read_lock &obj_hash[i].lock irq_context: softirq (&ndev->rs_timer) rcu_read_lock rcu_read_lock rcu_read_lock_bh rcu_read_lock &n->list_lock irq_context: softirq (&ndev->rs_timer) rcu_read_lock rcu_read_lock rcu_read_lock_bh rcu_read_lock &n->list_lock &c->lock irq_context: softirq (&ndev->rs_timer) rcu_read_lock rcu_read_lock rcu_read_lock_bh rcu_read_lock_bh rcu_read_lock rlock-AF_PACKET irq_context: softirq (&ndev->rs_timer) rcu_read_lock rcu_read_lock rcu_read_lock_bh &hsr->seqnr_lock rcu_read_lock rcu_read_lock_bh rcu_read_lock pool_lock#2 irq_context: softirq (&ndev->rs_timer) rcu_read_lock rcu_read_lock rcu_read_lock_bh &hsr->seqnr_lock rcu_read_lock rcu_read_lock_bh rcu_read_lock &obj_hash[i].lock irq_context: softirq (&ndev->rs_timer) rcu_read_lock rcu_read_lock rcu_read_lock_bh _xmit_TUNNEL6#2 fill_pool_map-wait-type-override pool_lock#2 irq_context: softirq (&ndev->rs_timer) rcu_read_lock rcu_read_lock rcu_read_lock_bh _xmit_TUNNEL6#2 fill_pool_map-wait-type-override pool_lock irq_context: softirq (&ndev->rs_timer) rcu_read_lock rcu_read_lock rcu_read_lock_bh _xmit_TUNNEL6#2 &obj_hash[i].lock pool_lock irq_context: softirq (&ndev->rs_timer) rcu_read_lock rcu_read_lock rcu_read_lock_bh _xmit_ETHER#2 rcu_read_lock &c->lock irq_context: 0 (wq_completion)wg-crypt-wg1#6 (work_completion)(&peer->transmit_packet_work) &peer->endpoint_lock rcu_read_lock_bh rcu_read_lock rcu_read_lock_bh rcu_read_lock &c->lock irq_context: 0 (wq_completion)wg-crypt-wg1#6 (work_completion)(&peer->transmit_packet_work) &peer->endpoint_lock rcu_read_lock_bh rcu_read_lock rcu_read_lock_bh rcu_read_lock tk_core.seq.seqcount irq_context: 0 (wq_completion)wg-crypt-wg1#6 (work_completion)(&peer->transmit_packet_work) &peer->endpoint_lock rcu_read_lock_bh rcu_read_lock rcu_read_lock_bh rcu_read_lock &obj_hash[i].lock irq_context: 0 (wq_completion)wg-crypt-wg0 (work_completion)(&peer->transmit_packet_work) &peer->endpoint_lock rcu_read_lock_bh rcu_read_lock rcu_read_lock_bh rcu_read_lock &c->lock irq_context: 0 (wq_completion)wg-crypt-wg0#3 (work_completion)(&peer->transmit_packet_work) &peer->endpoint_lock rcu_read_lock_bh rcu_read_lock rcu_read_lock_bh rcu_read_lock &c->lock irq_context: 0 (wq_completion)wg-crypt-wg0#4 (work_completion)(&peer->transmit_packet_work) &peer->endpoint_lock rcu_read_lock_bh rcu_read_lock &obj_hash[i].lock irq_context: 0 (wq_completion)wg-crypt-wg0#4 (work_completion)(&peer->transmit_packet_work) &peer->endpoint_lock rcu_read_lock_bh rcu_read_lock &dir->lock#2 irq_context: 0 (wq_completion)wg-crypt-wg0#4 (work_completion)(&peer->transmit_packet_work) &peer->endpoint_lock rcu_read_lock_bh rcu_read_lock &tbl->lock irq_context: 0 (wq_completion)wg-crypt-wg0#4 (work_completion)(&peer->transmit_packet_work) &peer->endpoint_lock rcu_read_lock_bh rcu_read_lock &n->lock irq_context: 0 (wq_completion)wg-crypt-wg0#4 (work_completion)(&peer->transmit_packet_work) &peer->endpoint_lock rcu_read_lock_bh rcu_read_lock &____s->seqcount#9 irq_context: softirq rcu_read_lock rcu_read_lock k-slock-AF_INET6 rcu_read_lock rcu_read_lock rcu_read_lock_bh rcu_read_lock &c->lock irq_context: softirq rcu_read_lock rcu_read_lock k-slock-AF_INET6 rcu_read_lock rcu_read_lock rcu_read_lock_bh rcu_read_lock rlock-AF_PACKET irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem sysctl_lock fill_pool_map-wait-type-override pool_lock#2 irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem sysctl_lock fill_pool_map-wait-type-override pool_lock irq_context: 0 (wq_completion)wg-crypt-wg0#6 (work_completion)(&peer->transmit_packet_work) &peer->endpoint_lock rcu_read_lock_bh rcu_read_lock &obj_hash[i].lock irq_context: 0 (wq_completion)wg-crypt-wg0#6 (work_completion)(&peer->transmit_packet_work) &peer->endpoint_lock rcu_read_lock_bh rcu_read_lock &dir->lock#2 irq_context: 0 (wq_completion)wg-crypt-wg0#6 (work_completion)(&peer->transmit_packet_work) &peer->endpoint_lock rcu_read_lock_bh rcu_read_lock &tbl->lock irq_context: 0 (wq_completion)wg-crypt-wg0#6 (work_completion)(&peer->transmit_packet_work) &peer->endpoint_lock rcu_read_lock_bh rcu_read_lock &n->lock irq_context: 0 (wq_completion)wg-crypt-wg0#6 (work_completion)(&peer->transmit_packet_work) &peer->endpoint_lock rcu_read_lock_bh rcu_read_lock &____s->seqcount#9 irq_context: 0 (wq_completion)wg-crypt-wg0#6 (work_completion)(&peer->transmit_packet_work) &peer->endpoint_lock rcu_read_lock_bh rcu_read_lock rcu_read_lock_bh rcu_read_lock &c->lock irq_context: 0 (wq_completion)wg-crypt-wg0#5 (work_completion)(&peer->transmit_packet_work) &peer->endpoint_lock rcu_read_lock_bh rcu_read_lock &obj_hash[i].lock irq_context: 0 (wq_completion)wg-crypt-wg0#5 (work_completion)(&peer->transmit_packet_work) &peer->endpoint_lock rcu_read_lock_bh rcu_read_lock &dir->lock#2 irq_context: 0 (wq_completion)wg-crypt-wg0#5 (work_completion)(&peer->transmit_packet_work) &peer->endpoint_lock rcu_read_lock_bh rcu_read_lock &tbl->lock irq_context: 0 (wq_completion)wg-crypt-wg0#5 (work_completion)(&peer->transmit_packet_work) &peer->endpoint_lock rcu_read_lock_bh rcu_read_lock &n->lock irq_context: 0 (wq_completion)wg-crypt-wg0#5 (work_completion)(&peer->transmit_packet_work) &peer->endpoint_lock rcu_read_lock_bh rcu_read_lock &____s->seqcount#9 irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem rtnl_mutex remove_cache_srcu &rq->__lock irq_context: softirq rcu_callback &x->wait#24 &p->pi_lock &rq->__lock &cfs_rq->removed.lock irq_context: 0 (wq_completion)events (work_completion)(&pwq->unbound_release_work) &rnp->exp_wq[0] irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem &rnp->exp_wq[0] irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem rtnl_mutex batched_entropy_u8.lock crngs.lock irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem rtnl_mutex &root->kernfs_rwsem quarantine_lock irq_context: 0 (wq_completion)bat_events (work_completion)(&(&forw_packet_aggr->delayed_work)->work) &hard_iface->bat_iv.ogm_buff_mutex batched_entropy_u8.lock crngs.lock base_crng.lock irq_context: 0 (wq_completion)events (work_completion)(&rdev->wiphy_work) &rdev->wiphy.mtx &wdev->mtx rcu_node_0 irq_context: 0 (wq_completion)events (work_completion)(&rdev->wiphy_work) &rdev->wiphy.mtx &wdev->mtx &rcu_state.expedited_wq irq_context: 0 (wq_completion)events (work_completion)(&rdev->wiphy_work) &rdev->wiphy.mtx &wdev->mtx &rcu_state.expedited_wq &p->pi_lock irq_context: 0 (wq_completion)events (work_completion)(&rdev->wiphy_work) &rdev->wiphy.mtx &wdev->mtx &rcu_state.expedited_wq &p->pi_lock &rq->__lock irq_context: 0 (wq_completion)events (work_completion)(&rdev->wiphy_work) &rdev->wiphy.mtx &wdev->mtx &rcu_state.expedited_wq &p->pi_lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: softirq (&app->join_timer) &app->lock batched_entropy_u32.lock crngs.lock base_crng.lock irq_context: 0 (wq_completion)events fqdir_free_work &p->pi_lock &rq->__lock &cfs_rq->removed.lock irq_context: 0 &sb->s_type->i_mutex_key#10 sk_lock-AF_INET6 slock-AF_INET6 fill_pool_map-wait-type-override &c->lock irq_context: 0 pernet_ops_rwsem ipvs->est_mutex pcpu_alloc_mutex &rq->__lock irq_context: 0 pernet_ops_rwsem __ip_vs_app_mutex &____s->seqcount irq_context: 0 rtnl_mutex &rnp->exp_wq[2] irq_context: 0 sb_writers#4 &type->i_mutex_dir_key#3/1 sb_internal jbd2_handle &ei->i_data_sem &bgl->locks[i].lock rcu_read_lock &p->pi_lock irq_context: 0 sb_writers#4 &type->i_mutex_dir_key#3/1 sb_internal jbd2_handle &ei->i_data_sem &bgl->locks[i].lock rcu_read_lock &p->pi_lock &rq->__lock irq_context: 0 sb_writers#4 &type->i_mutex_dir_key#3/1 sb_internal jbd2_handle &ei->i_data_sem &bgl->locks[i].lock rcu_read_lock &p->pi_lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 sb_writers#4 &type->i_mutex_dir_key#3/1 sb_internal jbd2_handle &ei->i_data_sem rcu_node_0 irq_context: 0 sb_writers#4 &type->i_mutex_dir_key#3/1 sb_internal jbd2_handle &ei->i_data_sem &rcu_state.expedited_wq irq_context: 0 sb_writers#4 &type->i_mutex_dir_key#3/1 sb_internal jbd2_handle &ei->i_data_sem &rcu_state.expedited_wq &p->pi_lock irq_context: 0 sb_writers#4 &type->i_mutex_dir_key#3/1 sb_internal jbd2_handle &ei->i_data_sem &rcu_state.expedited_wq &p->pi_lock &rq->__lock irq_context: 0 sb_writers#4 &type->i_mutex_dir_key#3/1 sb_internal jbd2_handle &ei->i_data_sem &rcu_state.expedited_wq &p->pi_lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 sk_lock-AF_INET &obj_hash[i].lock pool_lock irq_context: 0 (wq_completion)wg-crypt-wg2#5 (work_completion)(&({ do { const void *__vpp_verify = (typeof((worker) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __asm__ ("" : "=r"(__ptr) : "0"((typeof(*((worker))) *)((worker)))); (typeof((typeof(*((worker))) *)((worker)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); })->work) &rq->__lock irq_context: 0 (wq_completion)wg-crypt-wg2#5 (work_completion)(&peer->transmit_packet_work) &peer->endpoint_lock rcu_read_lock_bh rcu_read_lock &c->lock irq_context: 0 (wq_completion)wg-crypt-wg2#5 (work_completion)(&peer->transmit_packet_work) &peer->endpoint_lock rcu_read_lock_bh rcu_read_lock &dir->lock#2 irq_context: 0 (wq_completion)wg-crypt-wg2#5 (work_completion)(&peer->transmit_packet_work) &peer->endpoint_lock rcu_read_lock_bh rcu_read_lock &ul->lock irq_context: 0 (wq_completion)wg-crypt-wg2#5 (work_completion)(&peer->transmit_packet_work) &peer->endpoint_lock rcu_read_lock_bh rcu_read_lock &obj_hash[i].lock irq_context: 0 rtnl_mutex &root->kernfs_rwsem &sem->wait_lock irq_context: 0 pernet_ops_rwsem ipvs->est_mutex &____s->seqcount#2 irq_context: 0 sb_writers#8 &of->mutex kn->active#5 uevent_sock_mutex uevent_sock_mutex.wait_lock irq_context: 0 pernet_ops_rwsem rtnl_mutex uevent_sock_mutex.wait_lock irq_context: 0 sb_writers#4 &type->i_mutex_dir_key#3/1 fill_pool_map-wait-type-override &c->lock irq_context: 0 rtnl_mutex &idev->mc_lock &bridge_netdev_addr_lock_key &n->list_lock irq_context: 0 rtnl_mutex &idev->mc_lock &bridge_netdev_addr_lock_key &n->list_lock &c->lock irq_context: 0 (wq_completion)events_unbound (reaper_work).work fill_pool_map-wait-type-override &c->lock irq_context: 0 (wq_completion)events_unbound (reaper_work).work fill_pool_map-wait-type-override pool_lock#2 irq_context: 0 (wq_completion)events_unbound (reaper_work).work fill_pool_map-wait-type-override pool_lock irq_context: 0 &type->i_mutex_dir_key#4 &p->pi_lock &cfs_rq->removed.lock irq_context: 0 fanout_mutex irq_context: 0 fanout_mutex fs_reclaim irq_context: 0 fanout_mutex fs_reclaim mmu_notifier_invalidate_range_start irq_context: 0 fanout_mutex fs_reclaim mmu_notifier_invalidate_range_start &rq->__lock irq_context: 0 fanout_mutex fs_reclaim mmu_notifier_invalidate_range_start &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 fanout_mutex pool_lock#2 irq_context: 0 fanout_mutex &po->bind_lock irq_context: 0 fanout_mutex &po->bind_lock ptype_lock irq_context: 0 fanout_mutex &po->bind_lock &match->lock irq_context: 0 fanout_mutex &po->bind_lock &match->lock ptype_lock irq_context: softirq (&in_dev->mr_ifc_timer) rcu_read_lock rcu_read_lock_bh rcu_read_lock &____s->seqcount irq_context: softirq (&in_dev->mr_ifc_timer) rcu_read_lock rcu_read_lock_bh rcu_read_lock rcu_read_lock pool_lock#2 irq_context: 0 dup_mmap_sem &mm->mmap_lock &rq->__lock irq_context: 0 dup_mmap_sem &mm->mmap_lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 &sb->s_type->i_mutex_key#8 remove_cache_srcu irq_context: 0 &sb->s_type->i_mutex_key#8 remove_cache_srcu quarantine_lock irq_context: 0 &sb->s_type->i_mutex_key#8 remove_cache_srcu &c->lock irq_context: 0 &sb->s_type->i_mutex_key#8 remove_cache_srcu &n->list_lock irq_context: 0 &sb->s_type->i_mutex_key#8 remove_cache_srcu rcu_read_lock pool_lock#2 irq_context: 0 &sb->s_type->i_mutex_key#8 remove_cache_srcu &obj_hash[i].lock irq_context: 0 &sb->s_type->i_mutex_key#8 remove_cache_srcu rcu_read_lock rcu_node_0 irq_context: 0 &sb->s_type->i_mutex_key#8 remove_cache_srcu rcu_read_lock &rq->__lock irq_context: 0 &sb->s_type->i_mutex_key#8 remove_cache_srcu rcu_read_lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)bat_events &rq->__lock irq_context: 0 (wq_completion)bat_events &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 pernet_ops_rwsem rtnl_mutex &idev->mc_lock fill_pool_map-wait-type-override &c->lock irq_context: 0 pernet_ops_rwsem rtnl_mutex &idev->mc_lock fill_pool_map-wait-type-override &n->list_lock irq_context: 0 pernet_ops_rwsem rtnl_mutex &idev->mc_lock fill_pool_map-wait-type-override &n->list_lock &c->lock irq_context: 0 pernet_ops_rwsem rtnl_mutex &idev->mc_lock fill_pool_map-wait-type-override &rq->__lock irq_context: 0 pernet_ops_rwsem rtnl_mutex &idev->mc_lock fill_pool_map-wait-type-override &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 &sb->s_type->i_mutex_key#8 &sbi->s_writepages_rwsem jbd2_handle &ei->i_data_sem remove_cache_srcu &rq->__lock irq_context: 0 &sb->s_type->i_mutex_key#8 &sbi->s_writepages_rwsem jbd2_handle &ei->i_data_sem remove_cache_srcu &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 &sb->s_type->i_mutex_key#8 &sbi->s_writepages_rwsem jbd2_handle &ei->i_data_sem remove_cache_srcu pool_lock#2 irq_context: 0 &group->notification_waitq &ep->lock &ep->wq &p->pi_lock &rq->__lock &cfs_rq->removed.lock irq_context: 0 (wq_completion)events (work_completion)(&s->destroy_work) &obj_hash[i].lock pool_lock irq_context: 0 pernet_ops_rwsem rtnl_mutex &idev->mc_lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 &f->f_pos_lock &type->i_mutex_dir_key#3 rcu_read_lock rcu_node_0 irq_context: 0 &f->f_pos_lock &type->i_mutex_dir_key#3 rcu_read_lock &rq->__lock irq_context: 0 &f->f_pos_lock &type->i_mutex_dir_key#3 rcu_read_lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)wg-crypt-wg0#4 (work_completion)(&peer->transmit_packet_work) &peer->endpoint_lock rcu_read_lock_bh rcu_read_lock rcu_read_lock_bh rcu_read_lock &c->lock irq_context: 0 sb_writers#4 &type->i_mutex_dir_key#3/1 sb_internal jbd2_handle rcu_node_0 irq_context: 0 (wq_completion)events_unbound (reaper_work).work &obj_hash[i].lock pool_lock irq_context: 0 pernet_ops_rwsem nf_ct_proto_mutex defrag4_mutex nf_hook_mutex &rq->__lock irq_context: 0 pernet_ops_rwsem nf_ct_proto_mutex defrag4_mutex nf_hook_mutex &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 pernet_ops_rwsem nf_ct_proto_mutex defrag6_mutex &rq->__lock irq_context: 0 pernet_ops_rwsem nf_ct_proto_mutex defrag6_mutex &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 rcu_read_lock batched_entropy_u8.lock irq_context: 0 rcu_read_lock rcu_read_lock_bh &meta->lock irq_context: 0 rcu_read_lock rcu_read_lock_bh kfence_freelist_lock irq_context: 0 sb_writers#5 &type->i_mutex_dir_key#5 &cfs_rq->removed.lock irq_context: 0 sb_writers#5 &type->i_mutex_dir_key#5 &obj_hash[i].lock irq_context: softirq (&ndev->rs_timer) rcu_read_lock rcu_read_lock rcu_read_lock_bh rcu_read_lock rcu_read_lock rcu_read_lock_bh rcu_read_lock pool_lock#2 irq_context: softirq (&ndev->rs_timer) rcu_read_lock rcu_read_lock rcu_read_lock_bh rcu_read_lock rcu_read_lock rcu_read_lock_bh rcu_read_lock tk_core.seq.seqcount irq_context: softirq (&ndev->rs_timer) rcu_read_lock rcu_read_lock rcu_read_lock_bh rcu_read_lock rcu_read_lock rcu_read_lock_bh rcu_read_lock &obj_hash[i].lock irq_context: softirq rcu_read_lock rcu_read_lock rcu_read_lock rcu_read_lock_bh rcu_read_lock &n->list_lock irq_context: softirq rcu_read_lock rcu_read_lock rcu_read_lock rcu_read_lock_bh rcu_read_lock &n->list_lock &c->lock irq_context: 0 (wq_completion)events (work_completion)(&(&krcp->monitor_work)->work) rcu_node_0 irq_context: 0 (wq_completion)events (work_completion)(&(&krcp->monitor_work)->work) &rcu_state.expedited_wq irq_context: 0 (wq_completion)events (work_completion)(&(&krcp->monitor_work)->work) &rcu_state.expedited_wq &p->pi_lock irq_context: 0 (wq_completion)events (work_completion)(&(&krcp->monitor_work)->work) &rcu_state.expedited_wq &p->pi_lock &rq->__lock irq_context: 0 (wq_completion)events (work_completion)(&(&krcp->monitor_work)->work) &rcu_state.expedited_wq &p->pi_lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)events (work_completion)(&(&krcp->monitor_work)->work) &rq->__lock irq_context: 0 (wq_completion)events (work_completion)(&(&krcp->monitor_work)->work) &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)wg-crypt-wg0 (work_completion)(&({ do { const void *__vpp_verify = (typeof((worker) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __asm__ ("" : "=r"(__ptr) : "0"((typeof(*((worker))) *)((worker)))); (typeof((typeof(*((worker))) *)((worker)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); })->work) &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 &sb->s_type->i_mutex_key#10 sk_lock-AF_PHONET rcu_state.exp_mutex &rnp->exp_wq[1] irq_context: 0 &sb->s_type->i_mutex_key#10 sk_lock-AF_PHONET rcu_state.exp_mutex &rnp->exp_wq[2] irq_context: 0 &sb->s_type->i_mutex_key#10 sk_lock-AF_PHONET rcu_state.exp_mutex.wait_lock irq_context: 0 &sb->s_type->i_mutex_key#10 sk_lock-AF_PHONET &p->pi_lock irq_context: 0 &sb->s_type->i_mutex_key#10 sk_lock-AF_PHONET &p->pi_lock &rq->__lock irq_context: 0 &sb->s_type->i_mutex_key#10 sk_lock-AF_PHONET &p->pi_lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 &sb->s_type->i_mutex_key#10 sk_lock-AF_PHONET &rq->__lock irq_context: 0 &sb->s_type->i_mutex_key#10 sk_lock-AF_PHONET &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 &sb->s_type->i_mutex_key#10 sk_lock-AF_PHONET &rnp->exp_lock irq_context: 0 &sb->s_type->i_mutex_key#10 sk_lock-AF_PHONET rcu_state.exp_mutex irq_context: 0 &sb->s_type->i_mutex_key#10 sk_lock-AF_PHONET rcu_state.exp_mutex rcu_state.exp_mutex.wait_lock irq_context: 0 &sb->s_type->i_mutex_key#10 sk_lock-AF_PHONET &rnp->exp_wq[1] irq_context: 0 sk_lock-AF_INET rcu_read_lock rcu_read_lock &n->list_lock irq_context: 0 sk_lock-AF_INET rcu_read_lock rcu_read_lock &n->list_lock &c->lock irq_context: 0 rtnl_mutex uevent_sock_mutex rcu_read_lock &ei->socket.wq.wait &ep->lock &ep->wq &p->pi_lock &cfs_rq->removed.lock irq_context: 0 &sb->s_type->i_mutex_key#10 sk_lock-AF_PHONET rcu_state.exp_mutex &rnp->exp_wq[3] irq_context: 0 &u->iolock rcu_node_0 irq_context: 0 rtnl_mutex rcu_read_lock &rcu_state.expedited_wq irq_context: 0 rtnl_mutex rcu_read_lock &rcu_state.expedited_wq &p->pi_lock irq_context: 0 rtnl_mutex rcu_read_lock &rcu_state.expedited_wq &p->pi_lock &rq->__lock irq_context: 0 rtnl_mutex rcu_read_lock &rcu_state.expedited_wq &p->pi_lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 &sb->s_type->i_mutex_key#10 &____s->seqcount irq_context: 0 (wq_completion)ext4-rsv-conversion (work_completion)(&ei->i_rsv_conversion_work) jbd2_handle &ei->i_data_sem &sem->wait_lock irq_context: 0 &sbi->s_writepages_rwsem jbd2_handle &sem->wait_lock irq_context: 0 &sbi->s_writepages_rwsem jbd2_handle &p->pi_lock irq_context: 0 &sbi->s_writepages_rwsem jbd2_handle &p->pi_lock &rq->__lock irq_context: 0 &sbi->s_writepages_rwsem jbd2_handle &p->pi_lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 &sb->s_type->i_mutex_key#10 &po->bind_lock &match->lock irq_context: 0 &sb->s_type->i_mutex_key#10 &po->bind_lock &match->lock ptype_lock irq_context: 0 &sb->s_type->i_mutex_key#10 rcu_state.exp_mutex.wait_lock irq_context: 0 &sb->s_type->i_mutex_key#10 &p->pi_lock irq_context: 0 &sb->s_type->i_mutex_key#10 &p->pi_lock &rq->__lock irq_context: 0 &sb->s_type->i_mutex_key#10 &p->pi_lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 &sb->s_type->i_mutex_key#10 &match->lock irq_context: 0 tomoyo_ss &rcu_state.expedited_wq irq_context: 0 tomoyo_ss &rcu_state.expedited_wq &p->pi_lock irq_context: 0 tomoyo_ss &rcu_state.expedited_wq &p->pi_lock &rq->__lock irq_context: 0 tomoyo_ss &rcu_state.expedited_wq &p->pi_lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 &sb->s_type->i_mutex_key#10 sk_lock-AF_PHONET rcu_state.exp_mutex &obj_hash[i].lock pool_lock irq_context: 0 br_ioctl_mutex rtnl_mutex irq_context: 0 br_ioctl_mutex rtnl_mutex &mm->mmap_lock irq_context: 0 sb_writers#4 &type->i_mutex_dir_key#3/1 sb_internal &cfs_rq->removed.lock irq_context: 0 tomoyo_ss rcu_read_lock &rq->__lock &cfs_rq->removed.lock irq_context: 0 sb_writers#4 &type->i_mutex_dir_key#3/1 sb_internal rcu_read_lock rcu_node_0 irq_context: 0 sb_writers#4 &type->i_mutex_dir_key#3/1 sb_internal rcu_read_lock &rq->__lock irq_context: 0 sb_writers#4 &type->i_mutex_dir_key#3/1 sb_internal rcu_read_lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 &xt[i].mutex rcu_read_lock rcu_read_lock &rq->__lock &cfs_rq->removed.lock irq_context: 0 &sb->s_type->i_mutex_key#9 remove_cache_srcu irq_context: 0 &sb->s_type->i_mutex_key#9 remove_cache_srcu quarantine_lock irq_context: 0 sk_lock-AF_BLUETOOTH-BTPROTO_HCI &mm->mmap_lock irq_context: 0 cb_lock genl_mutex rcu_read_lock rcu_read_lock_bh rcu_read_lock &c->lock irq_context: softirq (&ndev->rs_timer) rcu_read_lock rcu_read_lock rcu_read_lock_bh _xmit_ETHER#2 rcu_read_lock rlock-AF_PACKET irq_context: 0 (wq_completion)kblockd (work_completion)(&q->timeout_work) &tags->lock irq_context: 0 (wq_completion)kblockd (work_completion)(&q->timeout_work) &obj_hash[i].lock irq_context: 0 (wq_completion)kblockd (work_completion)(&q->timeout_work) &base->lock irq_context: 0 (wq_completion)kblockd (work_completion)(&q->timeout_work) &base->lock &obj_hash[i].lock irq_context: 0 sb_writers#4 &type->i_mutex_dir_key#3/1 jbd2_handle &ei->i_data_sem quarantine_lock irq_context: 0 &xt[i].mutex batched_entropy_u8.lock irq_context: 0 &xt[i].mutex kfence_freelist_lock irq_context: 0 &xt[i].mutex &meta->lock irq_context: softirq (&hsr->announce_timer) rcu_read_lock &hsr->seqnr_lock rcu_read_lock rcu_read_lock_bh rcu_read_lock &n->list_lock irq_context: softirq (&hsr->announce_timer) rcu_read_lock &hsr->seqnr_lock rcu_read_lock rcu_read_lock_bh rcu_read_lock &n->list_lock &c->lock irq_context: 0 sb_writers#4 &type->i_mutex_dir_key#3/1 jbd2_handle &ei->i_data_sem &ei->i_es_lock &n->list_lock irq_context: 0 sb_writers#4 &type->i_mutex_dir_key#3/1 jbd2_handle &ei->i_data_sem &ei->i_es_lock &n->list_lock &c->lock irq_context: 0 &group->mark_mutex &cfs_rq->removed.lock irq_context: 0 &group->mark_mutex &obj_hash[i].lock irq_context: 0 &sb->s_type->i_mutex_key#9 rcu_read_lock &rcu_state.gp_wq irq_context: 0 &sb->s_type->i_mutex_key#9 rcu_read_lock &rcu_state.gp_wq &p->pi_lock irq_context: 0 &sb->s_type->i_mutex_key#9 rcu_read_lock &rcu_state.gp_wq &p->pi_lock &rq->__lock irq_context: 0 &sb->s_type->i_mutex_key#9 rcu_read_lock &rcu_state.gp_wq &p->pi_lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 sb_writers#4 &type->i_mutex_dir_key#3/1 sb_internal jbd2_handle &ei->i_data_sem &ei->i_es_lock key#7 irq_context: 0 sb_writers#8 &type->i_mutex_dir_key#4 &rq->__lock irq_context: 0 sb_writers#8 &type->i_mutex_dir_key#4 &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 &rq->__lock cid_lock irq_context: 0 (wq_completion)wg-kex-wg2#9 (work_completion)(&peer->transmit_handshake_work) &wg->static_identity.lock &handshake->lock &rq->__lock irq_context: 0 (wq_completion)wg-kex-wg2#9 (work_completion)(&peer->transmit_handshake_work) &rq->__lock irq_context: 0 (wq_completion)wg-kex-wg2#9 (work_completion)(&peer->transmit_handshake_work) &c->lock irq_context: 0 (wq_completion)wg-kex-wg2#9 (work_completion)(&peer->transmit_handshake_work) &peer->endpoint_lock rcu_read_lock_bh rcu_read_lock &n->list_lock irq_context: 0 (wq_completion)wg-kex-wg2#9 (work_completion)(&peer->transmit_handshake_work) &peer->endpoint_lock rcu_read_lock_bh rcu_read_lock &n->list_lock &c->lock irq_context: 0 dup_mmap_sem &mm->mmap_lock &mm->mmap_lock/1 &base->lock irq_context: 0 dup_mmap_sem &mm->mmap_lock &mm->mmap_lock/1 &base->lock &obj_hash[i].lock irq_context: 0 cb_lock genl_mutex &ht->lock irq_context: 0 sb_writers#3 oom_adj_mutex &rq->__lock irq_context: 0 cb_lock rcu_read_lock rcu_read_lock_bh rcu_read_lock &c->lock irq_context: 0 cb_lock genl_mutex rtnl_mutex &____s->seqcount irq_context: 0 cb_lock genl_mutex rtnl_mutex rcu_read_lock pool_lock#2 irq_context: 0 cb_lock genl_mutex rtnl_mutex &c->lock irq_context: 0 cb_lock genl_mutex rtnl_mutex rlock-AF_NETLINK irq_context: 0 sb_writers#4 &type->i_mutex_dir_key#3/1 jbd2_handle &xa->xa_lock#9 stock_lock irq_context: 0 sb_writers#8 kn->active#5 remove_cache_srcu irq_context: 0 sb_writers#8 kn->active#5 remove_cache_srcu quarantine_lock irq_context: 0 (wq_completion)inet_frag_wq (work_completion)(&fqdir->destroy_work) &ht->mutex quarantine_lock irq_context: 0 &xt[i].mutex rcu_read_lock &pool->lock irq_context: 0 &xt[i].mutex rcu_read_lock &pool->lock &obj_hash[i].lock irq_context: 0 &xt[i].mutex rcu_read_lock &pool->lock &p->pi_lock irq_context: 0 &xt[i].mutex rcu_read_lock &pool->lock &p->pi_lock &rq->__lock irq_context: 0 &xt[i].mutex rcu_read_lock &pool->lock &p->pi_lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 &f->f_pos_lock sb_writers#4 &sb->s_type->i_mutex_key#8 &mm->mmap_lock rcu_read_lock rcu_read_lock ptlock_ptr(page)#2 rcu_read_lock per_cpu_ptr(&cgroup_rstat_cpu_lock, cpu) irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem &sem->wait_lock irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem &root->kernfs_rwsem &sem->wait_lock irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem &root->kernfs_rwsem &rq->__lock irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem &root->kernfs_rwsem &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)events drain_vmap_work vmap_purge_lock &rq->__lock &cfs_rq->removed.lock irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem k-sk_lock-AF_RXRPC &meta->lock irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem k-sk_lock-AF_RXRPC kfence_freelist_lock irq_context: 0 (wq_completion)wg-crypt-wg1#3 (work_completion)(&peer->transmit_packet_work) &peer->endpoint_lock rcu_read_lock_bh rcu_read_lock rcu_read_lock_bh rcu_read_lock &c->lock irq_context: softirq (&ndev->rs_timer) rcu_read_lock rcu_read_lock rcu_read_lock_bh dev->qdisc_tx_busylock ?: &qdisc_tx_busylock#2 rcu_read_lock_bh rcu_read_lock &c->lock irq_context: softirq (&ndev->rs_timer) rcu_read_lock rcu_read_lock rcu_read_lock_bh rcu_read_lock rcu_read_lock rcu_read_lock_bh rcu_read_lock &c->lock irq_context: softirq (&ndev->rs_timer) rcu_read_lock rcu_read_lock rcu_read_lock_bh rcu_read_lock rcu_read_lock rcu_read_lock_bh rcu_read_lock rlock-AF_PACKET irq_context: 0 (wq_completion)wg-crypt-wg2#3 (work_completion)(&peer->transmit_packet_work) &peer->endpoint_lock rcu_read_lock_bh rcu_read_lock rcu_read_lock_bh rcu_read_lock &obj_hash[i].lock irq_context: 0 (wq_completion)wg-crypt-wg2#3 (work_completion)(&peer->transmit_packet_work) &peer->endpoint_lock rcu_read_lock_bh rcu_read_lock rcu_read_lock_bh rcu_read_lock pool_lock#2 irq_context: 0 (wq_completion)events (work_completion)(&(&nsim_dev->trap_data->trap_report_dw)->work) &devlink->lock_key &rq->__lock &cfs_rq->removed.lock irq_context: 0 (wq_completion)wg-crypt-wg1#3 (work_completion)(&peer->transmit_packet_work) &peer->endpoint_lock rcu_read_lock_bh rcu_read_lock rcu_read_lock_bh rcu_read_lock &n->list_lock irq_context: 0 (wq_completion)wg-crypt-wg1#3 (work_completion)(&peer->transmit_packet_work) &peer->endpoint_lock rcu_read_lock_bh rcu_read_lock rcu_read_lock_bh rcu_read_lock &n->list_lock &c->lock irq_context: 0 (wq_completion)wg-crypt-wg2#4 (work_completion)(&peer->transmit_packet_work) &peer->endpoint_lock rcu_read_lock_bh &obj_hash[i].lock irq_context: 0 (wq_completion)wg-crypt-wg2#4 (work_completion)(&peer->transmit_packet_work) &peer->endpoint_lock rcu_read_lock_bh rcu_read_lock rcu_read_lock_bh rcu_read_lock tk_core.seq.seqcount irq_context: 0 (wq_completion)wg-crypt-wg2#4 (work_completion)(&peer->transmit_packet_work) &peer->endpoint_lock rcu_read_lock_bh rcu_read_lock rcu_read_lock_bh rcu_read_lock &obj_hash[i].lock irq_context: 0 (wq_completion)wg-crypt-wg1#3 (work_completion)(&peer->transmit_packet_work) &peer->endpoint_lock rcu_read_lock_bh rcu_read_lock rcu_read_lock_bh rcu_read_lock tk_core.seq.seqcount irq_context: 0 (wq_completion)wg-crypt-wg1#3 (work_completion)(&peer->transmit_packet_work) &peer->endpoint_lock rcu_read_lock_bh rcu_read_lock rcu_read_lock_bh rcu_read_lock &obj_hash[i].lock irq_context: 0 (wq_completion)mld (work_completion)(&(&idev->mc_dad_work)->work) &idev->mc_lock rcu_read_lock rcu_read_lock rcu_read_lock_bh rcu_read_lock &meta->lock irq_context: 0 (wq_completion)mld (work_completion)(&(&idev->mc_dad_work)->work) &idev->mc_lock rcu_read_lock rcu_read_lock rcu_read_lock_bh rcu_read_lock kfence_freelist_lock irq_context: 0 (wq_completion)wg-crypt-wg0 (work_completion)(&peer->transmit_packet_work) &peer->endpoint_lock rcu_read_lock_bh rcu_read_lock rcu_read_lock_bh rcu_read_lock rlock-AF_PACKET irq_context: 0 (wq_completion)wg-crypt-wg1 (work_completion)(&({ do { const void *__vpp_verify = (typeof((worker) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __asm__ ("" : "=r"(__ptr) : "0"((typeof(*((worker))) *)((worker)))); (typeof((typeof(*((worker))) *)((worker)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); })->work) &c->lock irq_context: 0 &f->f_pos_lock sb_writers#4 &sb->s_type->i_mutex_key#8 jbd2_handle rcu_read_lock &rcu_state.expedited_wq irq_context: 0 &f->f_pos_lock sb_writers#4 &sb->s_type->i_mutex_key#8 jbd2_handle rcu_read_lock &rcu_state.expedited_wq &p->pi_lock irq_context: 0 &f->f_pos_lock sb_writers#4 &sb->s_type->i_mutex_key#8 jbd2_handle rcu_read_lock &rcu_state.expedited_wq &p->pi_lock &rq->__lock irq_context: 0 &f->f_pos_lock sb_writers#4 &sb->s_type->i_mutex_key#8 jbd2_handle rcu_read_lock &rcu_state.expedited_wq &p->pi_lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 &f->f_pos_lock sb_writers#4 &sb->s_type->i_mutex_key#8 jbd2_handle &rcu_state.expedited_wq irq_context: 0 &f->f_pos_lock sb_writers#4 &sb->s_type->i_mutex_key#8 jbd2_handle &rcu_state.expedited_wq &p->pi_lock irq_context: 0 &f->f_pos_lock sb_writers#4 &sb->s_type->i_mutex_key#8 jbd2_handle &rcu_state.expedited_wq &p->pi_lock &rq->__lock irq_context: 0 &f->f_pos_lock sb_writers#4 &sb->s_type->i_mutex_key#8 jbd2_handle &rcu_state.expedited_wq &p->pi_lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 &sbi->s_writepages_rwsem jbd2_handle tk_core.seq.seqcount irq_context: 0 &sbi->s_writepages_rwsem jbd2_handle &dd->lock irq_context: 0 &sbi->s_writepages_rwsem jbd2_handle rcu_read_lock &dd->lock irq_context: 0 &sbi->s_writepages_rwsem jbd2_handle rcu_read_lock &obj_hash[i].lock irq_context: 0 &sbi->s_writepages_rwsem jbd2_handle rcu_read_lock &c->lock irq_context: 0 &sbi->s_writepages_rwsem jbd2_handle rcu_read_lock pool_lock#2 irq_context: 0 &sbi->s_writepages_rwsem jbd2_handle rcu_read_lock tk_core.seq.seqcount irq_context: 0 &sbi->s_writepages_rwsem jbd2_handle rcu_read_lock &virtscsi_vq->vq_lock irq_context: 0 &sbi->s_writepages_rwsem jbd2_handle rcu_read_lock &rcu_state.expedited_wq irq_context: 0 &sbi->s_writepages_rwsem jbd2_handle rcu_read_lock &rcu_state.expedited_wq &p->pi_lock irq_context: 0 &sbi->s_writepages_rwsem jbd2_handle rcu_read_lock &rcu_state.expedited_wq &p->pi_lock &rq->__lock irq_context: 0 &sbi->s_writepages_rwsem jbd2_handle rcu_read_lock &rcu_state.expedited_wq &p->pi_lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)wg-crypt-wg0#2 (work_completion)(&peer->transmit_packet_work) &peer->endpoint_lock rcu_read_lock_bh rcu_read_lock rcu_read_lock_bh rcu_read_lock &c->lock irq_context: 0 &f->f_pos_lock sb_writers#4 &sb->s_type->i_mutex_key#8 &ei->i_data_sem &sem->wait_lock irq_context: 0 &f->f_pos_lock sb_writers#4 &sb->s_type->i_mutex_key#8 &mm->mmap_lock rcu_read_lock &rcu_state.expedited_wq irq_context: 0 &f->f_pos_lock sb_writers#4 &sb->s_type->i_mutex_key#8 &mm->mmap_lock rcu_read_lock &rcu_state.expedited_wq &p->pi_lock irq_context: 0 &f->f_pos_lock sb_writers#4 &sb->s_type->i_mutex_key#8 &mm->mmap_lock rcu_read_lock &rcu_state.expedited_wq &p->pi_lock &rq->__lock irq_context: 0 &f->f_pos_lock sb_writers#4 &sb->s_type->i_mutex_key#8 &mm->mmap_lock rcu_read_lock &rcu_state.expedited_wq &p->pi_lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 &sbi->s_writepages_rwsem jbd2_handle rcu_read_lock &rcu_state.gp_wq irq_context: 0 &sbi->s_writepages_rwsem jbd2_handle rcu_read_lock &rcu_state.gp_wq &p->pi_lock irq_context: 0 &sbi->s_writepages_rwsem jbd2_handle rcu_read_lock &rcu_state.gp_wq &p->pi_lock &rq->__lock irq_context: 0 &sbi->s_writepages_rwsem jbd2_handle rcu_read_lock &rcu_state.gp_wq &p->pi_lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)events (work_completion)(&p->wq) rcu_read_lock &pool->lock &p->pi_lock irq_context: 0 (wq_completion)events drain_vmap_work vmap_purge_lock free_vmap_area_lock init_mm.page_table_lock &pcp->lock &zone->lock irq_context: 0 (wq_completion)events drain_vmap_work vmap_purge_lock free_vmap_area_lock init_mm.page_table_lock &pcp->lock &zone->lock &____s->seqcount irq_context: 0 sb_writers#4 &pcp->lock &zone->lock irq_context: 0 sb_writers#4 &pcp->lock &zone->lock &____s->seqcount irq_context: 0 (wq_completion)writeback (work_completion)(&(&wb->dwork)->work) &type->s_umount_key#43 rcu_read_lock &xa->xa_lock#9 &wb->work_lock irq_context: 0 (wq_completion)writeback (work_completion)(&(&wb->dwork)->work) &type->s_umount_key#43 rcu_read_lock &xa->xa_lock#9 key#13 irq_context: 0 (wq_completion)writeback (work_completion)(&(&wb->dwork)->work) &type->s_umount_key#43 rcu_read_lock &xa->xa_lock#9 key#12 irq_context: softirq (&ndev->rs_timer) rcu_read_lock rcu_read_lock rcu_read_lock_bh rcu_read_lock rcu_read_lock &tbl->lock &n->lock irq_context: softirq (&ndev->rs_timer) rcu_read_lock rcu_read_lock rcu_read_lock_bh rcu_read_lock rcu_read_lock &tbl->lock &c->lock irq_context: softirq (&ndev->rs_timer) rcu_read_lock rcu_read_lock rcu_read_lock_bh rcu_read_lock rcu_read_lock &tbl->lock pool_lock#2 irq_context: softirq (&ndev->rs_timer) rcu_read_lock rcu_read_lock rcu_read_lock_bh rcu_read_lock rcu_read_lock &tbl->lock &n->lock irq_context: softirq (&ndev->rs_timer) rcu_read_lock rcu_read_lock rcu_read_lock_bh rcu_read_lock rcu_read_lock &tbl->lock &n->lock &____s->seqcount#9 irq_context: softirq (&ndev->rs_timer) rcu_read_lock rcu_read_lock rcu_read_lock_bh rcu_read_lock rcu_read_lock &tbl->lock nl_table_lock irq_context: softirq (&ndev->rs_timer) rcu_read_lock rcu_read_lock rcu_read_lock_bh rcu_read_lock rcu_read_lock &tbl->lock &obj_hash[i].lock irq_context: softirq (&ndev->rs_timer) rcu_read_lock rcu_read_lock rcu_read_lock_bh rcu_read_lock rcu_read_lock &tbl->lock nl_table_wait.lock irq_context: softirq (&ndev->rs_timer) rcu_read_lock rcu_read_lock rcu_read_lock_bh rcu_read_lock rcu_read_lock &tbl->lock rcu_read_lock lock#8 irq_context: softirq (&ndev->rs_timer) rcu_read_lock rcu_read_lock rcu_read_lock_bh rcu_read_lock rcu_read_lock &tbl->lock rcu_read_lock id_table_lock irq_context: softirq (&ndev->rs_timer) rcu_read_lock rcu_read_lock rcu_read_lock_bh rcu_read_lock rcu_read_lock &tbl->lock &dir->lock#2 irq_context: softirq (&ndev->rs_timer) rcu_read_lock rcu_read_lock rcu_read_lock_bh rcu_read_lock rcu_read_lock &tbl->lock krc.lock irq_context: softirq (&ndev->rs_timer) rcu_read_lock rcu_read_lock rcu_read_lock_bh rcu_read_lock rcu_read_lock &tbl->lock &n->list_lock irq_context: softirq (&ndev->rs_timer) rcu_read_lock rcu_read_lock rcu_read_lock_bh rcu_read_lock rcu_read_lock &tbl->lock &n->list_lock &c->lock irq_context: 0 (wq_completion)wg-crypt-wg1#2 (work_completion)(&peer->transmit_packet_work) &peer->endpoint_lock rcu_read_lock_bh &obj_hash[i].lock irq_context: softirq (&ndev->rs_timer) rcu_read_lock rcu_read_lock rcu_read_lock_bh rcu_read_lock rcu_read_lock &n->lock &c->lock irq_context: softirq (&ndev->rs_timer) rcu_read_lock rcu_read_lock rcu_read_lock_bh rcu_read_lock rcu_read_lock rcu_read_lock rcu_read_lock &____s->seqcount#10 irq_context: 0 (wq_completion)events_power_efficient (gc_work).work rcu_node_0 irq_context: 0 (wq_completion)events_power_efficient (gc_work).work &rcu_state.expedited_wq irq_context: 0 (wq_completion)events_power_efficient (gc_work).work &rcu_state.expedited_wq &p->pi_lock irq_context: 0 (wq_completion)events_power_efficient (gc_work).work &rcu_state.expedited_wq &p->pi_lock &rq->__lock irq_context: 0 (wq_completion)events_power_efficient (gc_work).work &rcu_state.expedited_wq &p->pi_lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)bat_events (work_completion)(&(&bat_priv->tt.work)->work) rcu_node_0 irq_context: 0 (wq_completion)bat_events (work_completion)(&(&bat_priv->tt.work)->work) &rcu_state.expedited_wq irq_context: 0 (wq_completion)bat_events (work_completion)(&(&bat_priv->tt.work)->work) &rcu_state.expedited_wq &p->pi_lock irq_context: 0 (wq_completion)bat_events (work_completion)(&(&bat_priv->tt.work)->work) &rcu_state.expedited_wq &p->pi_lock &rq->__lock irq_context: 0 (wq_completion)bat_events (work_completion)(&(&bat_priv->tt.work)->work) &rcu_state.expedited_wq &p->pi_lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 sb_writers#4 &rcu_state.expedited_wq irq_context: 0 sb_writers#4 &rcu_state.expedited_wq &p->pi_lock irq_context: 0 sb_writers#4 &rcu_state.expedited_wq &p->pi_lock &rq->__lock irq_context: 0 sb_writers#4 &rcu_state.expedited_wq &p->pi_lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 &sec->lock irq_context: 0 rcu_read_lock_bh dev->qdisc_tx_busylock ?: &qdisc_tx_busylock _xmit_NETROM &rdev->wpan_phy.queue_lock irq_context: 0 rcu_read_lock_bh dev->qdisc_tx_busylock ?: &qdisc_tx_busylock _xmit_NETROM rcu_read_lock &c->lock irq_context: 0 rcu_read_lock_bh dev->qdisc_tx_busylock ?: &qdisc_tx_busylock _xmit_NETROM rcu_read_lock pool_lock#2 irq_context: 0 rcu_read_lock_bh dev->qdisc_tx_busylock ?: &qdisc_tx_busylock _xmit_NETROM rcu_read_lock &list->lock#30 irq_context: 0 rcu_read_lock_bh dev->qdisc_tx_busylock ?: &qdisc_tx_busylock _xmit_NETROM &rdev->wpan_phy.sync_txq irq_context: softirq &list->lock#30 irq_context: softirq rcu_read_lock rcu_read_lock raw_lock irq_context: 0 cb_lock genl_mutex stock_lock irq_context: 0 cb_lock genl_mutex &pernet->lock irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem rtnl_mutex kernfs_idr_lock &obj_hash[i].lock pool_lock irq_context: 0 sb_writers#4 &type->i_mutex_dir_key#3/1 sb_internal jbd2_handle &journal->j_state_lock &journal->j_wait_commit irq_context: 0 sb_writers#4 &type->i_mutex_dir_key#3/1 sb_internal jbd2_handle &journal->j_state_lock &journal->j_wait_commit &p->pi_lock irq_context: 0 sb_writers#4 &type->i_mutex_dir_key#3/1 sb_internal jbd2_handle &journal->j_state_lock &journal->j_wait_commit &p->pi_lock &rq->__lock irq_context: 0 sb_writers#4 &type->i_mutex_dir_key#3/1 sb_internal jbd2_handle &journal->j_state_lock &journal->j_wait_commit &p->pi_lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 sk_lock-AF_INET &tcp_hashinfo.bhash[i].lock stock_lock irq_context: 0 sk_lock-AF_INET &tcp_hashinfo.bhash[i].lock &tcp_hashinfo.bhash2[i].lock stock_lock irq_context: 0 sk_lock-AF_INET free_vmap_area_lock irq_context: 0 sk_lock-AF_INET vmap_area_lock irq_context: 0 sk_lock-AF_INET pcpu_alloc_mutex irq_context: 0 sk_lock-AF_INET pcpu_alloc_mutex pcpu_lock irq_context: 0 sk_lock-AF_INET pack_mutex irq_context: 0 sk_lock-AF_INET text_mutex irq_context: 0 sk_lock-AF_INET text_mutex ptlock_ptr(page)#2 irq_context: 0 sk_lock-AF_INET &fp->aux->used_maps_mutex irq_context: 0 sk_lock-AF_INET &tcp_hashinfo.bhash[i].lock &tcp_hashinfo.bhash2[i].lock &obj_hash[i].lock irq_context: 0 sk_lock-AF_INET rcu_read_lock rcu_read_lock &nf_conntrack_locks[i] batched_entropy_u8.lock irq_context: 0 &group->mark_mutex remove_cache_srcu pool_lock#2 irq_context: 0 sk_lock-AF_INET rcu_read_lock tcp_metrics_lock irq_context: 0 sk_lock-AF_INET rcu_read_lock tcp_metrics_lock &c->lock irq_context: 0 sk_lock-AF_INET rcu_read_lock tcp_metrics_lock pool_lock#2 irq_context: 0 sk_lock-AF_INET hrtimer_bases.lock irq_context: 0 sk_lock-AF_INET hrtimer_bases.lock &obj_hash[i].lock irq_context: 0 sk_lock-AF_INET hrtimer_bases.lock tk_core.seq.seqcount irq_context: 0 sk_lock-AF_INET &f->f_owner.lock irq_context: 0 sk_lock-AF_INET slock-AF_INET rcu_read_lock rcu_read_lock rcu_read_lock_bh &list->lock#5 irq_context: 0 sk_lock-AF_INET slock-AF_INET hrtimer_bases.lock irq_context: 0 sk_lock-AF_INET slock-AF_INET hrtimer_bases.lock &obj_hash[i].lock irq_context: 0 sk_lock-AF_INET slock-AF_INET hrtimer_bases.lock tk_core.seq.seqcount irq_context: softirq rcu_read_lock rcu_read_lock slock-AF_INET/1 rcu_read_lock rcu_read_lock rcu_read_lock_bh &list->lock#5 irq_context: 0 &sb->s_type->i_mutex_key#10 sk_lock-AF_INET &c->lock irq_context: 0 &sb->s_type->i_mutex_key#10 sk_lock-AF_INET rcu_read_lock rcu_read_lock rcu_read_lock_bh &list->lock#5 irq_context: 0 &sb->s_type->i_mutex_key#10 sk_lock-AF_INET slock-AF_INET &base->lock irq_context: 0 &sb->s_type->i_mutex_key#10 sk_lock-AF_INET slock-AF_INET &base->lock &obj_hash[i].lock irq_context: 0 &sb->s_type->i_mutex_key#10 sk_lock-AF_INET slock-AF_INET pool_lock#2 irq_context: 0 &sb->s_type->i_mutex_key#10 sk_lock-AF_INET slock-AF_INET &tcp_hashinfo.bhash[i].lock irq_context: 0 &sb->s_type->i_mutex_key#10 sk_lock-AF_INET slock-AF_INET &tcp_hashinfo.bhash[i].lock stock_lock irq_context: 0 &sb->s_type->i_mutex_key#10 sk_lock-AF_INET slock-AF_INET &tcp_hashinfo.bhash[i].lock &obj_hash[i].lock irq_context: 0 &sb->s_type->i_mutex_key#10 sk_lock-AF_INET slock-AF_INET &tcp_hashinfo.bhash[i].lock pool_lock#2 irq_context: 0 &sb->s_type->i_mutex_key#10 sk_lock-AF_INET slock-AF_INET &tcp_hashinfo.bhash[i].lock &tcp_hashinfo.bhash2[i].lock irq_context: 0 &sb->s_type->i_mutex_key#10 sk_lock-AF_INET slock-AF_INET &tcp_hashinfo.bhash[i].lock &tcp_hashinfo.bhash2[i].lock stock_lock irq_context: 0 &sb->s_type->i_mutex_key#10 sk_lock-AF_INET slock-AF_INET &tcp_hashinfo.bhash[i].lock &tcp_hashinfo.bhash2[i].lock &obj_hash[i].lock irq_context: 0 &sb->s_type->i_mutex_key#10 sk_lock-AF_INET slock-AF_INET &tcp_hashinfo.bhash[i].lock &tcp_hashinfo.bhash2[i].lock pool_lock#2 irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem rtnl_mutex &root->kernfs_rwsem &meta->lock irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem rtnl_mutex &root->kernfs_rwsem kfence_freelist_lock irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem rtnl_mutex &root->kernfs_rwsem &root->kernfs_iattr_rwsem &sem->wait_lock irq_context: 0 &sb->s_type->i_mutex_key#10 sk_lock-AF_CAIF rcu_state.exp_mutex.wait_lock irq_context: 0 &sb->s_type->i_mutex_key#10 sk_lock-AF_CAIF &p->pi_lock irq_context: 0 &sb->s_type->i_mutex_key#10 sk_lock-AF_CAIF &p->pi_lock &rq->__lock irq_context: 0 &sb->s_type->i_mutex_key#10 sk_lock-AF_CAIF &p->pi_lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 &sb->s_type->i_mutex_key#10 sk_lock-AF_CAIF &rq->__lock irq_context: 0 &sb->s_type->i_mutex_key#10 sk_lock-AF_CAIF &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 &sb->s_type->i_mutex_key#10 sk_lock-AF_CAIF &rnp->exp_wq[3] irq_context: 0 smc_v4_hashinfo.lock irq_context: 0 sk_lock-AF_SMC irq_context: 0 sk_lock-AF_SMC slock-AF_SMC irq_context: 0 sk_lock-AF_SMC &smc->clcsock_release_lock irq_context: 0 sk_lock-AF_SMC &smc->clcsock_release_lock &net->smc.mutex_fback_rsn irq_context: 0 sk_lock-AF_SMC &smc->clcsock_release_lock k-clock-AF_INET irq_context: 0 sk_lock-AF_SMC k-sk_lock-AF_INET irq_context: 0 sk_lock-AF_SMC k-sk_lock-AF_INET k-slock-AF_INET irq_context: 0 sk_lock-AF_SMC k-sk_lock-AF_INET fs_reclaim irq_context: 0 sk_lock-AF_SMC k-sk_lock-AF_INET fs_reclaim mmu_notifier_invalidate_range_start irq_context: 0 sk_lock-AF_SMC k-sk_lock-AF_INET pool_lock#2 irq_context: 0 sk_lock-AF_SMC k-sk_lock-AF_INET &____s->seqcount#8 irq_context: 0 sk_lock-AF_SMC k-sk_lock-AF_INET &tcp_hashinfo.bhash[i].lock irq_context: 0 sk_lock-AF_SMC k-sk_lock-AF_INET &tcp_hashinfo.bhash[i].lock stock_lock irq_context: 0 sk_lock-AF_SMC k-sk_lock-AF_INET &tcp_hashinfo.bhash[i].lock pool_lock#2 irq_context: 0 sk_lock-AF_SMC k-sk_lock-AF_INET &tcp_hashinfo.bhash[i].lock &tcp_hashinfo.bhash2[i].lock irq_context: 0 sk_lock-AF_SMC k-sk_lock-AF_INET &tcp_hashinfo.bhash[i].lock &tcp_hashinfo.bhash2[i].lock stock_lock irq_context: 0 sk_lock-AF_SMC k-sk_lock-AF_INET &tcp_hashinfo.bhash[i].lock &tcp_hashinfo.bhash2[i].lock pool_lock#2 irq_context: 0 sk_lock-AF_SMC k-sk_lock-AF_INET &tcp_hashinfo.bhash[i].lock &tcp_hashinfo.bhash2[i].lock batched_entropy_u8.lock irq_context: 0 sk_lock-AF_SMC k-sk_lock-AF_INET &tcp_hashinfo.bhash[i].lock &tcp_hashinfo.bhash2[i].lock &hashinfo->ehash_locks[i] irq_context: 0 sk_lock-AF_SMC k-sk_lock-AF_INET batched_entropy_u32.lock irq_context: 0 sk_lock-AF_SMC k-sk_lock-AF_INET tk_core.seq.seqcount irq_context: 0 sk_lock-AF_SMC k-sk_lock-AF_INET batched_entropy_u16.lock irq_context: 0 sk_lock-AF_SMC k-sk_lock-AF_INET &____s->seqcount irq_context: 0 sk_lock-AF_SMC k-sk_lock-AF_INET rcu_read_lock pool_lock#2 irq_context: 0 sk_lock-AF_SMC k-sk_lock-AF_INET &obj_hash[i].lock irq_context: 0 sk_lock-AF_SMC k-sk_lock-AF_INET rcu_read_lock rcu_read_lock &____s->seqcount#7 irq_context: 0 sk_lock-AF_SMC k-sk_lock-AF_INET rcu_read_lock rcu_read_lock &c->lock irq_context: 0 sk_lock-AF_SMC k-sk_lock-AF_INET rcu_read_lock rcu_read_lock &____s->seqcount#2 irq_context: 0 sk_lock-AF_SMC k-sk_lock-AF_INET rcu_read_lock rcu_read_lock &____s->seqcount irq_context: 0 sk_lock-AF_SMC k-sk_lock-AF_INET rcu_read_lock rcu_read_lock pool_lock#2 irq_context: 0 sk_lock-AF_SMC k-sk_lock-AF_INET rcu_read_lock rcu_read_lock &ct->lock irq_context: 0 sk_lock-AF_SMC k-sk_lock-AF_INET rcu_read_lock rcu_read_lock rcu_read_lock &____s->seqcount#7 irq_context: 0 sk_lock-AF_SMC k-sk_lock-AF_INET rcu_read_lock rcu_read_lock &nf_nat_locks[i] irq_context: 0 sk_lock-AF_SMC k-sk_lock-AF_INET rcu_read_lock rcu_read_lock &nf_conntrack_locks[i] irq_context: 0 sk_lock-AF_SMC k-sk_lock-AF_INET rcu_read_lock rcu_read_lock &nf_conntrack_locks[i] &nf_conntrack_locks[i]/1 irq_context: 0 sk_lock-AF_SMC k-sk_lock-AF_INET rcu_read_lock rcu_read_lock &nf_conntrack_locks[i] &nf_conntrack_locks[i]/1 batched_entropy_u8.lock irq_context: 0 sk_lock-AF_SMC k-sk_lock-AF_INET rcu_read_lock rcu_read_lock &____s->seqcount#10 irq_context: 0 sk_lock-AF_SMC k-sk_lock-AF_INET rcu_read_lock rcu_read_lock rcu_read_lock_bh &list->lock#5 irq_context: 0 sk_lock-AF_SMC k-sk_lock-AF_INET &base->lock irq_context: 0 sk_lock-AF_SMC k-sk_lock-AF_INET &base->lock &obj_hash[i].lock irq_context: 0 sk_lock-AF_SMC k-sk_lock-AF_INET &ei->socket.wq.wait irq_context: 0 sk_lock-AF_SMC k-sk_lock-AF_INET &hashinfo->ehash_locks[i] irq_context: 0 sk_lock-AF_SMC k-sk_lock-AF_INET &tcp_hashinfo.bhash[i].lock &obj_hash[i].lock irq_context: 0 sk_lock-AF_SMC k-sk_lock-AF_INET &tcp_hashinfo.bhash[i].lock &tcp_hashinfo.bhash2[i].lock &obj_hash[i].lock irq_context: 0 sk_lock-AF_SMC k-sk_lock-AF_INET k-clock-AF_INET irq_context: 0 sk_lock-AF_SMC k-sk_lock-AF_INET k-clock-AF_INET rcu_read_lock &ei->socket.wq.wait irq_context: 0 sk_lock-AF_SMC k-sk_lock-AF_INET k-clock-AF_INET rcu_read_lock rcu_read_lock &ei->socket.wq.wait irq_context: 0 sk_lock-AF_SMC k-slock-AF_INET irq_context: 0 slock-AF_SMC irq_context: 0 k-sk_lock-AF_INET fs_reclaim irq_context: 0 k-sk_lock-AF_INET fs_reclaim mmu_notifier_invalidate_range_start irq_context: 0 k-sk_lock-AF_INET fs_reclaim mmu_notifier_invalidate_range_start &rq->__lock irq_context: 0 k-sk_lock-AF_INET pool_lock#2 irq_context: 0 k-sk_lock-AF_INET &obj_hash[i].lock irq_context: 0 &sb->s_type->i_mutex_key#10 (work_completion)(&smc->connect_work) irq_context: 0 &sb->s_type->i_mutex_key#10 sk_lock-AF_SMC irq_context: 0 &sb->s_type->i_mutex_key#10 sk_lock-AF_SMC slock-AF_SMC irq_context: 0 &sb->s_type->i_mutex_key#10 sk_lock-AF_SMC k-clock-AF_INET irq_context: 0 &sb->s_type->i_mutex_key#10 sk_lock-AF_SMC smc_v4_hashinfo.lock irq_context: 0 &sb->s_type->i_mutex_key#10 slock-AF_SMC irq_context: 0 &sb->s_type->i_mutex_key#10 &smc->clcsock_release_lock irq_context: 0 &sb->s_type->i_mutex_key#10 &smc->clcsock_release_lock k-sk_lock-AF_INET irq_context: 0 &sb->s_type->i_mutex_key#10 &smc->clcsock_release_lock k-sk_lock-AF_INET k-slock-AF_INET irq_context: 0 &sb->s_type->i_mutex_key#10 &smc->clcsock_release_lock k-sk_lock-AF_INET k-clock-AF_INET irq_context: 0 &sb->s_type->i_mutex_key#10 &smc->clcsock_release_lock k-sk_lock-AF_INET k-slock-AF_INET &obj_hash[i].lock irq_context: 0 &sb->s_type->i_mutex_key#10 &smc->clcsock_release_lock k-sk_lock-AF_INET k-slock-AF_INET elock-AF_INET irq_context: 0 &sb->s_type->i_mutex_key#10 &smc->clcsock_release_lock k-slock-AF_INET irq_context: 0 &sb->s_type->i_mutex_key#10 &smc->clcsock_release_lock pool_lock#2 irq_context: 0 &sb->s_type->i_mutex_key#10 &smc->clcsock_release_lock &dir->lock irq_context: 0 &sb->s_type->i_mutex_key#10 &smc->clcsock_release_lock &obj_hash[i].lock irq_context: 0 &sb->s_type->i_mutex_key#10 &smc->clcsock_release_lock stock_lock irq_context: 0 &sb->s_type->i_mutex_key#10 &smc->clcsock_release_lock &sb->s_type->i_lock_key#8 irq_context: 0 &sb->s_type->i_mutex_key#10 &smc->clcsock_release_lock &xa->xa_lock#9 irq_context: 0 &sb->s_type->i_mutex_key#10 &smc->clcsock_release_lock &fsnotify_mark_srcu irq_context: 0 &sb->s_type->i_mutex_key#10 sk_lock-AF_SMC clock-AF_SMC irq_context: 0 rtnl_mutex sk_lock-AF_INET6 irq_context: 0 rtnl_mutex sk_lock-AF_INET6 slock-AF_INET6 irq_context: 0 rtnl_mutex sk_lock-AF_INET6 &mm->mmap_lock irq_context: 0 rtnl_mutex sk_lock-AF_INET6 fs_reclaim irq_context: 0 rtnl_mutex sk_lock-AF_INET6 fs_reclaim mmu_notifier_invalidate_range_start irq_context: 0 rtnl_mutex sk_lock-AF_INET6 &c->lock irq_context: 0 rtnl_mutex sk_lock-AF_INET6 &n->list_lock irq_context: 0 rtnl_mutex sk_lock-AF_INET6 &n->list_lock &c->lock irq_context: 0 rtnl_mutex sk_lock-AF_INET6 &rq->__lock irq_context: 0 rtnl_mutex sk_lock-AF_INET6 &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 rtnl_mutex sk_lock-AF_INET6 pool_lock#2 irq_context: 0 rtnl_mutex sk_lock-AF_INET6 rcu_read_lock stock_lock irq_context: 0 rtnl_mutex sk_lock-AF_INET6 rcu_read_lock pool_lock#2 irq_context: 0 rtnl_mutex sk_lock-AF_INET6 rcu_read_lock &dir->lock#2 irq_context: 0 rtnl_mutex sk_lock-AF_INET6 &obj_hash[i].lock irq_context: 0 rtnl_mutex sk_lock-AF_INET6 &ndev->lock irq_context: 0 rtnl_mutex sk_lock-AF_INET6 &ndev->lock pool_lock#2 irq_context: 0 rtnl_mutex sk_lock-AF_INET6 &ndev->lock &dir->lock#2 irq_context: 0 rtnl_mutex sk_lock-AF_INET6 &ndev->lock pcpu_lock irq_context: 0 rtnl_mutex sk_lock-AF_INET6 &ndev->lock &c->lock irq_context: 0 sk_lock-AF_SMC k-sk_lock-AF_INET rcu_read_lock &c->lock irq_context: 0 sk_lock-AF_SMC k-sk_lock-AF_INET rcu_read_lock &dir->lock#2 irq_context: 0 sk_lock-AF_SMC rcu_read_lock &pool->lock irq_context: 0 sk_lock-AF_SMC rcu_read_lock &pool->lock &obj_hash[i].lock irq_context: 0 sk_lock-AF_SMC rcu_read_lock &pool->lock pool_lock#2 irq_context: 0 sk_lock-AF_SMC rcu_read_lock &pool->lock &p->pi_lock irq_context: 0 sk_lock-AF_SMC rcu_read_lock &pool->lock &p->pi_lock &rq->__lock irq_context: 0 sk_lock-AF_SMC rcu_read_lock &pool->lock &p->pi_lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)smc_hs_wq irq_context: 0 (wq_completion)smc_hs_wq (work_completion)(&smc->connect_work) irq_context: 0 (wq_completion)smc_hs_wq (work_completion)(&smc->connect_work) k-sk_lock-AF_INET irq_context: 0 (wq_completion)smc_hs_wq (work_completion)(&smc->connect_work) k-sk_lock-AF_INET k-slock-AF_INET irq_context: 0 (wq_completion)smc_hs_wq (work_completion)(&smc->connect_work) k-slock-AF_INET irq_context: 0 (wq_completion)smc_hs_wq (work_completion)(&smc->connect_work) sk_lock-AF_SMC irq_context: 0 (wq_completion)smc_hs_wq (work_completion)(&smc->connect_work) sk_lock-AF_SMC slock-AF_SMC irq_context: 0 (wq_completion)smc_hs_wq (work_completion)(&smc->connect_work) slock-AF_SMC irq_context: 0 rtnl_mutex sk_lock-AF_INET6 acaddr_hash_lock irq_context: 0 rtnl_mutex sk_lock-AF_INET6 &tb->tb6_lock irq_context: 0 rtnl_mutex sk_lock-AF_INET6 &tb->tb6_lock stock_lock irq_context: 0 rtnl_mutex sk_lock-AF_INET6 &tb->tb6_lock pool_lock#2 irq_context: 0 rtnl_mutex sk_lock-AF_INET6 &tb->tb6_lock &c->lock irq_context: 0 rtnl_mutex sk_lock-AF_INET6 &tb->tb6_lock rcu_read_lock &c->lock irq_context: 0 rtnl_mutex sk_lock-AF_INET6 &tb->tb6_lock rcu_read_lock pool_lock#2 irq_context: 0 rtnl_mutex sk_lock-AF_INET6 &tb->tb6_lock rcu_read_lock &data->fib_event_queue_lock irq_context: 0 rtnl_mutex sk_lock-AF_INET6 &tb->tb6_lock rcu_read_lock rcu_read_lock &pool->lock irq_context: 0 rtnl_mutex sk_lock-AF_INET6 &tb->tb6_lock rcu_read_lock rcu_read_lock &pool->lock &obj_hash[i].lock irq_context: 0 rtnl_mutex sk_lock-AF_INET6 &tb->tb6_lock rcu_read_lock rcu_read_lock &pool->lock pool_lock#2 irq_context: 0 rtnl_mutex sk_lock-AF_INET6 &tb->tb6_lock rcu_read_lock rcu_read_lock &pool->lock &p->pi_lock irq_context: 0 rtnl_mutex sk_lock-AF_INET6 &tb->tb6_lock rcu_read_lock rcu_read_lock &pool->lock &p->pi_lock &rq->__lock irq_context: 0 rtnl_mutex sk_lock-AF_INET6 &tb->tb6_lock rcu_read_lock rcu_read_lock &pool->lock &p->pi_lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 rtnl_mutex sk_lock-AF_INET6 &tb->tb6_lock nl_table_lock irq_context: 0 rtnl_mutex sk_lock-AF_INET6 &tb->tb6_lock &obj_hash[i].lock irq_context: 0 rtnl_mutex sk_lock-AF_INET6 &tb->tb6_lock quarantine_lock irq_context: 0 rtnl_mutex sk_lock-AF_INET6 &tb->tb6_lock nl_table_wait.lock irq_context: 0 rtnl_mutex sk_lock-AF_INET6 &idev->mc_lock irq_context: 0 rtnl_mutex sk_lock-AF_INET6 &idev->mc_lock fs_reclaim irq_context: 0 rtnl_mutex sk_lock-AF_INET6 &idev->mc_lock fs_reclaim mmu_notifier_invalidate_range_start irq_context: 0 rtnl_mutex sk_lock-AF_INET6 &idev->mc_lock pool_lock#2 irq_context: 0 rtnl_mutex sk_lock-AF_INET6 &idev->mc_lock &obj_hash[i].lock irq_context: 0 rtnl_mutex sk_lock-AF_INET6 &idev->mc_lock _xmit_ETHER irq_context: 0 rtnl_mutex sk_lock-AF_INET6 &idev->mc_lock _xmit_ETHER pool_lock#2 irq_context: 0 rtnl_mutex sk_lock-AF_INET6 &idev->mc_lock &base->lock irq_context: 0 rtnl_mutex sk_lock-AF_INET6 &idev->mc_lock &base->lock &obj_hash[i].lock irq_context: 0 rtnl_mutex sk_lock-AF_INET6 &idev->mc_lock &rq->__lock irq_context: 0 rtnl_mutex slock-AF_INET6 irq_context: 0 &sb->s_type->i_mutex_key#10 rtnl_mutex irq_context: 0 &sb->s_type->i_mutex_key#10 rtnl_mutex &rq->__lock irq_context: 0 &sb->s_type->i_mutex_key#10 rtnl_mutex &ndev->lock irq_context: 0 &sb->s_type->i_mutex_key#10 rtnl_mutex acaddr_hash_lock irq_context: 0 &sb->s_type->i_mutex_key#10 rtnl_mutex &idev->mc_lock irq_context: 0 &sb->s_type->i_mutex_key#10 rtnl_mutex &idev->mc_lock _xmit_ETHER irq_context: 0 &sb->s_type->i_mutex_key#10 rtnl_mutex &idev->mc_lock _xmit_ETHER &obj_hash[i].lock irq_context: 0 &sb->s_type->i_mutex_key#10 rtnl_mutex &idev->mc_lock _xmit_ETHER pool_lock#2 irq_context: 0 &sb->s_type->i_mutex_key#10 rtnl_mutex &idev->mc_lock _xmit_ETHER krc.lock irq_context: 0 &sb->s_type->i_mutex_key#10 rtnl_mutex &idev->mc_lock fs_reclaim irq_context: 0 &sb->s_type->i_mutex_key#10 rtnl_mutex &idev->mc_lock fs_reclaim mmu_notifier_invalidate_range_start irq_context: 0 &sb->s_type->i_mutex_key#10 rtnl_mutex &idev->mc_lock pool_lock#2 irq_context: 0 &sb->s_type->i_mutex_key#10 rtnl_mutex &idev->mc_lock &obj_hash[i].lock irq_context: 0 &sb->s_type->i_mutex_key#10 rtnl_mutex &idev->mc_lock &base->lock irq_context: 0 &sb->s_type->i_mutex_key#10 rtnl_mutex &idev->mc_lock &base->lock &obj_hash[i].lock irq_context: 0 &sb->s_type->i_mutex_key#10 rtnl_mutex &obj_hash[i].lock irq_context: 0 &sb->s_type->i_mutex_key#10 rtnl_mutex pool_lock#2 irq_context: 0 &sb->s_type->i_mutex_key#10 rtnl_mutex krc.lock irq_context: 0 &sb->s_type->i_mutex_key#10 rtnl_mutex &tb->tb6_lock irq_context: 0 &sb->s_type->i_mutex_key#10 rtnl_mutex &tb->tb6_lock &net->ipv6.fib6_walker_lock irq_context: 0 &sb->s_type->i_mutex_key#10 rtnl_mutex &tb->tb6_lock &obj_hash[i].lock irq_context: 0 &sb->s_type->i_mutex_key#10 rtnl_mutex &tb->tb6_lock pool_lock#2 irq_context: 0 &sb->s_type->i_mutex_key#10 rtnl_mutex &tb->tb6_lock rt6_exception_lock irq_context: 0 &sb->s_type->i_mutex_key#10 rtnl_mutex &tb->tb6_lock rcu_read_lock pool_lock#2 irq_context: 0 &sb->s_type->i_mutex_key#10 rtnl_mutex &tb->tb6_lock rcu_read_lock &data->fib_event_queue_lock irq_context: 0 &sb->s_type->i_mutex_key#10 rtnl_mutex &tb->tb6_lock rcu_read_lock rcu_read_lock &pool->lock irq_context: 0 &sb->s_type->i_mutex_key#10 rtnl_mutex &tb->tb6_lock rcu_read_lock rcu_read_lock &pool->lock &obj_hash[i].lock irq_context: 0 &sb->s_type->i_mutex_key#10 rtnl_mutex &tb->tb6_lock rcu_read_lock rcu_read_lock &pool->lock pool_lock#2 irq_context: 0 &sb->s_type->i_mutex_key#10 rtnl_mutex &tb->tb6_lock rcu_read_lock rcu_read_lock &pool->lock &p->pi_lock irq_context: 0 &sb->s_type->i_mutex_key#10 rtnl_mutex &tb->tb6_lock rcu_read_lock rcu_read_lock &pool->lock &p->pi_lock &rq->__lock irq_context: 0 &sb->s_type->i_mutex_key#10 rtnl_mutex &tb->tb6_lock rcu_read_lock rcu_read_lock &pool->lock &p->pi_lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 &sb->s_type->i_mutex_key#10 rtnl_mutex &tb->tb6_lock nl_table_lock irq_context: 0 &sb->s_type->i_mutex_key#10 rtnl_mutex &tb->tb6_lock nl_table_wait.lock irq_context: 0 &sb->s_type->i_mutex_key#10 rtnl_mutex &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 sk_lock-AF_CAIF &mm->mmap_lock irq_context: 0 &xt[i].mutex rcu_read_lock &rcu_state.expedited_wq irq_context: 0 &xt[i].mutex rcu_read_lock &rcu_state.expedited_wq &p->pi_lock irq_context: 0 &xt[i].mutex rcu_read_lock &rcu_state.expedited_wq &p->pi_lock &rq->__lock irq_context: 0 &xt[i].mutex rcu_read_lock &rcu_state.expedited_wq &p->pi_lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 lock prog_idr_lock &c->lock irq_context: 0 tracepoints_mutex tasklist_lock irq_context: 0 tracepoints_mutex cpu_hotplug_lock static_call_mutex text_mutex &rq->__lock irq_context: 0 sb_writers#4 &type->i_mutex_dir_key#3/1 sb_internal &n->list_lock irq_context: 0 sb_writers#4 &type->i_mutex_dir_key#3/1 sb_internal &n->list_lock &c->lock irq_context: softirq (&ndev->rs_timer) batched_entropy_u8.lock irq_context: softirq (&ndev->rs_timer) kfence_freelist_lock irq_context: 0 (wq_completion)events (work_completion)(&aux->work) per_cpu_ptr(&cgroup_rstat_cpu_lock, cpu) irq_context: 0 (wq_completion)events (work_completion)(&aux->work) &rq->__lock irq_context: 0 (wq_completion)events (work_completion)(&aux->work) &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)events (work_completion)(&aux->work) rcu_read_lock rcu_node_0 irq_context: 0 (wq_completion)events (work_completion)(&aux->work) rcu_read_lock &rq->__lock irq_context: 0 (wq_completion)events (work_completion)(&aux->work) rcu_read_lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)events (work_completion)(&aux->work) rcu_read_lock rcu_read_lock rcu_node_0 irq_context: 0 (wq_completion)events (work_completion)(&aux->work) rcu_read_lock rcu_read_lock &rq->__lock irq_context: 0 (wq_completion)events (work_completion)(&aux->work) rcu_read_lock rcu_read_lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)events (work_completion)(&aux->work) rcu_read_lock &cfs_rq->removed.lock irq_context: 0 (wq_completion)events (work_completion)(&aux->work) rcu_read_lock &obj_hash[i].lock irq_context: 0 dup_mmap_sem &mm->mmap_lock &mm->mmap_lock/1 pcpu_lock stock_lock irq_context: 0 sb_writers#4 &type->i_mutex_dir_key#3/1 &sb->s_type->i_mutex_key#8 &cfs_rq->removed.lock irq_context: 0 sb_writers#4 &type->i_mutex_dir_key#3/1 &type->i_mutex_dir_key#3 &n->list_lock irq_context: 0 sb_writers#4 &type->i_mutex_dir_key#3/1 &type->i_mutex_dir_key#3 &n->list_lock &c->lock irq_context: 0 l2tp_ip6_lock irq_context: 0 rcu_read_lock rcu_read_lock rcu_read_lock stock_lock irq_context: 0 sk_lock-AF_INET6 &list->lock#5 irq_context: 0 &sb->s_type->i_mutex_key#10 l2tp_ip6_lock irq_context: 0 &sb->s_type->i_mutex_key#10 &cfs_rq->removed.lock irq_context: 0 misc_mtx remove_cache_srcu irq_context: 0 misc_mtx remove_cache_srcu quarantine_lock irq_context: 0 misc_mtx remove_cache_srcu &c->lock irq_context: 0 misc_mtx remove_cache_srcu &n->list_lock irq_context: 0 misc_mtx remove_cache_srcu rcu_read_lock pool_lock#2 irq_context: 0 misc_mtx remove_cache_srcu &obj_hash[i].lock irq_context: 0 misc_mtx remove_cache_srcu &rq->__lock irq_context: 0 misc_mtx remove_cache_srcu &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 &u->iolock stock_lock irq_context: 0 &mm->mmap_lock &base->lock irq_context: 0 &mm->mmap_lock &base->lock &obj_hash[i].lock irq_context: 0 sb_writers#4 &type->i_mutex_dir_key#3/1 jbd2_handle &ei->i_data_sem &cfs_rq->removed.lock irq_context: 0 sb_writers#4 &type->i_mutex_dir_key#3/1 jbd2_handle &ei->i_data_sem &rq->__lock &cfs_rq->removed.lock irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem &rq->__lock &cfs_rq->removed.lock irq_context: 0 &knet->mutex irq_context: 0 &mux->lock irq_context: 0 &mux->rx_lock irq_context: 0 &sb->s_type->i_mutex_key#10 sk_lock-AF_KCM irq_context: 0 &sb->s_type->i_mutex_key#10 sk_lock-AF_KCM slock-AF_KCM irq_context: 0 &sb->s_type->i_mutex_key#10 sk_lock-AF_KCM clock-AF_KCM irq_context: 0 &sb->s_type->i_mutex_key#10 slock-AF_KCM irq_context: 0 &sb->s_type->i_mutex_key#10 &mux->lock irq_context: 0 &sb->s_type->i_mutex_key#10 (work_completion)(&kcm->tx_work) irq_context: 0 &sb->s_type->i_mutex_key#10 &mux->rx_lock irq_context: 0 &sb->s_type->i_mutex_key#10 &mux->rx_lock rlock-AF_KCM irq_context: 0 &sb->s_type->i_mutex_key#10 &knet->mutex irq_context: 0 &f->f_lock fasync_lock irq_context: 0 sb_writers#4 &type->i_mutex_dir_key#3/1 sb_internal jbd2_handle &ei->i_data_sem rcu_read_lock &rcu_state.gp_wq irq_context: 0 sb_writers#4 &type->i_mutex_dir_key#3/1 sb_internal jbd2_handle &ei->i_data_sem rcu_read_lock &rcu_state.gp_wq &p->pi_lock irq_context: 0 sb_writers#4 &type->i_mutex_dir_key#3/1 sb_internal jbd2_handle &ei->i_data_sem rcu_read_lock &rcu_state.gp_wq &p->pi_lock &rq->__lock irq_context: 0 sb_writers#4 &type->i_mutex_dir_key#3/1 sb_internal jbd2_handle &ei->i_data_sem rcu_read_lock &rcu_state.gp_wq &p->pi_lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 &f->f_pos_lock &type->i_mutex_dir_key#4 &root->kernfs_rwsem &cfs_rq->removed.lock irq_context: 0 &f->f_pos_lock &type->i_mutex_dir_key#4 &root->kernfs_rwsem &obj_hash[i].lock irq_context: 0 rcu_read_lock rhashtable_bucket irq_context: 0 sk_lock-AF_TIPC irq_context: 0 sk_lock-AF_TIPC slock-AF_TIPC irq_context: 0 sk_lock-AF_TIPC &rq->__lock irq_context: 0 sk_lock-AF_TIPC &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 sk_lock-AF_TIPC fs_reclaim irq_context: 0 sk_lock-AF_TIPC fs_reclaim mmu_notifier_invalidate_range_start irq_context: 0 sk_lock-AF_TIPC pool_lock#2 irq_context: 0 sk_lock-AF_TIPC &mm->mmap_lock irq_context: 0 sk_lock-AF_TIPC &list->lock#31 irq_context: 0 sk_lock-AF_TIPC slock-AF_TIPC &list->lock#31 irq_context: 0 slock-AF_TIPC irq_context: 0 &sb->s_type->i_mutex_key#10 sk_lock-AF_TIPC irq_context: 0 &sb->s_type->i_mutex_key#10 sk_lock-AF_TIPC slock-AF_TIPC irq_context: 0 &sb->s_type->i_mutex_key#10 sk_lock-AF_TIPC &obj_hash[i].lock irq_context: 0 &sb->s_type->i_mutex_key#10 sk_lock-AF_TIPC pool_lock#2 irq_context: 0 &sb->s_type->i_mutex_key#10 sk_lock-AF_TIPC &base->lock irq_context: 0 &sb->s_type->i_mutex_key#10 sk_lock-AF_TIPC &base->lock &obj_hash[i].lock irq_context: 0 &sb->s_type->i_mutex_key#10 sk_lock-AF_TIPC rcu_read_lock rhashtable_bucket irq_context: 0 &sb->s_type->i_mutex_key#10 sk_lock-AF_TIPC clock-AF_TIPC irq_context: 0 &sb->s_type->i_mutex_key#10 slock-AF_TIPC irq_context: 0 sb_writers#5 &type->i_mutex_dir_key#5/1 tomoyo_ss remove_cache_srcu &cfs_rq->removed.lock irq_context: 0 sb_writers#4 &type->i_mutex_dir_key#3/1 jbd2_handle lock#4 &lruvec->lru_lock irq_context: 0 sb_writers#4 &type->i_mutex_dir_key#3/1 sb_internal jbd2_handle &ei->i_data_sem mmu_notifier_invalidate_range_start &rq->__lock irq_context: 0 sb_writers#4 &type->i_mutex_dir_key#3/1 sb_internal jbd2_handle &ei->i_data_sem mmu_notifier_invalidate_range_start &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 sb_writers#4 &type->i_mutex_dir_key#3/1 &cfs_rq->removed.lock irq_context: 0 lock btf_idr_lock irq_context: 0 lock btf_idr_lock pool_lock#2 irq_context: 0 sb_writers#4 &type->i_mutex_dir_key#3/1 sb_internal jbd2_handle &ei->i_data_sem &bgl->locks[i].lock irq_context: 0 rlock-AF_CAIF irq_context: 0 sk_lock-AF_CAIF &this->info_list_lock irq_context: 0 sk_lock-AF_CAIF (console_sem).lock irq_context: 0 sk_lock-AF_CAIF rcu_state.exp_mutex rcu_node_0 irq_context: 0 sk_lock-AF_CAIF rcu_state.exp_mutex &obj_hash[i].lock irq_context: 0 sk_lock-AF_CAIF rcu_state.exp_mutex &rq->__lock irq_context: 0 sk_lock-AF_CAIF rcu_state.exp_mutex &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 sk_lock-AF_CAIF rcu_state.exp_mutex rcu_read_lock &pool->lock irq_context: 0 sk_lock-AF_CAIF rcu_state.exp_mutex rcu_read_lock &pool->lock &obj_hash[i].lock irq_context: 0 sk_lock-AF_CAIF rcu_state.exp_mutex rcu_read_lock &pool->lock &p->pi_lock irq_context: 0 sk_lock-AF_CAIF rcu_state.exp_mutex rcu_read_lock &pool->lock &p->pi_lock &rq->__lock irq_context: 0 sk_lock-AF_CAIF rcu_state.exp_mutex rcu_read_lock &pool->lock &p->pi_lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 sk_lock-AF_CAIF &ei->socket.wq.wait irq_context: 0 sk_lock-AF_CAIF clock-AF_CAIF irq_context: 0 sk_lock-AF_CAIF elock-AF_CAIF irq_context: 0 elock-AF_CAIF irq_context: 0 sk_lock-AF_CAIF rcu_state.exp_mutex &rnp->exp_wq[3] irq_context: 0 sk_lock-AF_INET6 rcu_read_lock rcu_read_lock rcu_read_lock_bh rcu_read_lock &____s->seqcount irq_context: 0 sk_lock-AF_INET6 rcu_read_lock rcu_read_lock rcu_read_lock_bh rcu_read_lock pool_lock#2 irq_context: 0 sk_lock-AF_INET6 rcu_read_lock rcu_read_lock rcu_read_lock_bh rcu_read_lock rcu_read_lock pool_lock#2 irq_context: 0 sk_lock-AF_INET6 rcu_read_lock rcu_read_lock rcu_read_lock_bh rcu_read_lock &obj_hash[i].lock irq_context: 0 sk_lock-AF_INET6 rcu_read_lock rcu_read_lock rcu_read_lock_bh rcu_read_lock tk_core.seq.seqcount irq_context: 0 sk_lock-AF_INET6 rcu_read_lock rcu_read_lock rcu_read_lock_bh rcu_read_lock rlock-AF_PACKET irq_context: 0 (wq_completion)bat_events (work_completion)(&(&forw_packet_aggr->delayed_work)->work) &hard_iface->bat_iv.ogm_buff_mutex &bat_priv->tvlv.container_list_lock &____s->seqcount irq_context: 0 sk_lock-AF_INET6 rcu_read_lock rcu_read_lock rcu_read_lock_bh rcu_read_lock &c->lock irq_context: 0 sk_lock-AF_INET6 rcu_read_lock rcu_read_lock rcu_read_lock_bh rcu_read_lock &n->list_lock irq_context: 0 sk_lock-AF_INET6 rcu_read_lock rcu_read_lock rcu_read_lock_bh rcu_read_lock &n->list_lock &c->lock irq_context: 0 sb_writers#8 &of->mutex kn->active#5 fs_reclaim mmu_notifier_invalidate_range_start &rq->__lock irq_context: 0 sb_writers#8 &of->mutex kn->active#5 fs_reclaim mmu_notifier_invalidate_range_start &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 &rdma_nl_types[idx].sem irq_context: 0 &rdma_nl_types[idx].sem nlk_cb_mutex-RDMA irq_context: 0 &rdma_nl_types[idx].sem nlk_cb_mutex-RDMA fs_reclaim irq_context: 0 &rdma_nl_types[idx].sem nlk_cb_mutex-RDMA fs_reclaim mmu_notifier_invalidate_range_start irq_context: 0 &rdma_nl_types[idx].sem nlk_cb_mutex-RDMA pool_lock#2 irq_context: 0 &rdma_nl_types[idx].sem nlk_cb_mutex-RDMA devices_rwsem irq_context: 0 &rdma_nl_types[idx].sem nlk_cb_mutex-RDMA rlock-AF_NETLINK irq_context: 0 sk_lock-AF_PACKET &zone->lock irq_context: 0 sk_lock-AF_PACKET &po->pg_vec_lock irq_context: 0 sk_lock-AF_PACKET &po->pg_vec_lock wlock-AF_PACKET irq_context: 0 &mm->mmap_lock &po->pg_vec_lock irq_context: 0 &mm->mmap_lock &po->pg_vec_lock &vma->vm_lock->lock irq_context: 0 &mm->mmap_lock &po->pg_vec_lock fs_reclaim irq_context: 0 &mm->mmap_lock &po->pg_vec_lock fs_reclaim mmu_notifier_invalidate_range_start irq_context: 0 &mm->mmap_lock &po->pg_vec_lock &____s->seqcount irq_context: 0 &mm->mmap_lock &po->pg_vec_lock pool_lock#2 irq_context: 0 &mm->mmap_lock &po->pg_vec_lock stock_lock irq_context: 0 &mm->mmap_lock &po->pg_vec_lock ptlock_ptr(page) irq_context: 0 &mm->mmap_lock &po->pg_vec_lock ptlock_ptr(page)#2 irq_context: 0 &mm->mmap_lock &po->pg_vec_lock ptlock_ptr(page)#2 key irq_context: 0 &sb->s_type->i_mutex_key#10 sk_lock-AF_PACKET &po->bind_lock irq_context: 0 &sb->s_type->i_mutex_key#10 sk_lock-AF_PACKET rcu_state.exp_mutex rcu_node_0 irq_context: 0 &sb->s_type->i_mutex_key#10 sk_lock-AF_PACKET rcu_state.exp_mutex &obj_hash[i].lock irq_context: 0 &sb->s_type->i_mutex_key#10 sk_lock-AF_PACKET rcu_state.exp_mutex rcu_read_lock &pool->lock irq_context: 0 &sb->s_type->i_mutex_key#10 sk_lock-AF_PACKET rcu_state.exp_mutex rcu_read_lock &pool->lock &obj_hash[i].lock irq_context: 0 &sb->s_type->i_mutex_key#10 sk_lock-AF_PACKET rcu_state.exp_mutex rcu_read_lock &pool->lock &p->pi_lock irq_context: 0 &sb->s_type->i_mutex_key#10 sk_lock-AF_PACKET rcu_state.exp_mutex rcu_read_lock &pool->lock &p->pi_lock &rq->__lock irq_context: 0 &sb->s_type->i_mutex_key#10 sk_lock-AF_PACKET rcu_state.exp_mutex rcu_read_lock &pool->lock &p->pi_lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 &sb->s_type->i_mutex_key#10 sk_lock-AF_PACKET rcu_state.exp_mutex &rq->__lock irq_context: 0 &sb->s_type->i_mutex_key#10 sk_lock-AF_PACKET rcu_state.exp_mutex &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 &sb->s_type->i_mutex_key#10 sk_lock-AF_PACKET rcu_state.exp_mutex &rnp->exp_wq[1] irq_context: 0 &sb->s_type->i_mutex_key#10 sk_lock-AF_PACKET &obj_hash[i].lock irq_context: 0 &sb->s_type->i_mutex_key#10 sk_lock-AF_PACKET &po->pg_vec_lock irq_context: 0 &sb->s_type->i_mutex_key#10 sk_lock-AF_PACKET &po->pg_vec_lock wlock-AF_PACKET irq_context: 0 &sb->s_type->i_mutex_key#10 sk_lock-AF_PACKET rcu_read_lock &____s->seqcount irq_context: 0 &sb->s_type->i_mutex_key#10 sk_lock-AF_PACKET rcu_read_lock pool_lock#2 irq_context: 0 &sb->s_type->i_mutex_key#10 sk_lock-AF_PACKET &zone->lock irq_context: 0 &sb->s_type->i_mutex_key#10 sk_lock-AF_PACKET &zone->lock &____s->seqcount irq_context: 0 &sb->s_type->i_mutex_key#10 sk_lock-AF_PACKET pool_lock#2 irq_context: 0 &p->lock remove_cache_srcu rcu_read_lock rcu_node_0 irq_context: 0 &p->lock remove_cache_srcu rcu_read_lock &rq->__lock irq_context: 0 &p->lock remove_cache_srcu rcu_read_lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 &sb->s_type->i_mutex_key#10 sk_lock-AF_INET6/1 quarantine_lock irq_context: 0 sb_writers#4 &type->i_mutex_dir_key#3/1 jbd2_handle &ei->i_data_sem remove_cache_srcu &rq->__lock irq_context: 0 sb_writers#4 &type->i_mutex_dir_key#3/1 jbd2_handle &ei->i_data_sem remove_cache_srcu &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 &disk->open_mutex &lo->lo_mutex &rq->__lock irq_context: 0 &disk->open_mutex &lo->lo_mutex &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 sb_writers#8 kn->active#5 rcu_read_lock rcu_node_0 irq_context: 0 sb_writers#8 kn->active#5 rcu_read_lock &rq->__lock irq_context: 0 sb_writers#8 kn->active#5 rcu_read_lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 sb_writers#5 &type->i_mutex_dir_key#5 fs_reclaim mmu_notifier_invalidate_range_start &rq->__lock irq_context: 0 sb_writers#5 &type->i_mutex_dir_key#5 fs_reclaim mmu_notifier_invalidate_range_start &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)events_unbound connector_reaper_work &ACCESS_PRIVATE(ssp->srcu_sup, lock) rcu_read_lock &pool->lock &p->pi_lock &rq->__lock &cfs_rq->removed.lock irq_context: 0 (wq_completion)events_unbound (reaper_work).work &ACCESS_PRIVATE(ssp->srcu_sup, lock) irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem rtnl_mutex krc.lock &obj_hash[i].lock irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem rtnl_mutex krc.lock &base->lock irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem rtnl_mutex krc.lock &base->lock &obj_hash[i].lock irq_context: softirq (&n->timer) k-slock-AF_INET6 rcu_read_lock rcu_read_lock &____s->seqcount#10 irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem stock_lock rcu_read_lock per_cpu_ptr(&cgroup_rstat_cpu_lock, cpu) irq_context: 0 (wq_completion)events fqdir_free_work rcu_state.barrier_mutex &cfs_rq->removed.lock irq_context: 0 (wq_completion)events fqdir_free_work rcu_state.barrier_mutex &obj_hash[i].lock irq_context: 0 (wq_completion)events fqdir_free_work rcu_state.barrier_mutex pool_lock#2 irq_context: softirq &(&hinfo->gc_work)->timer irq_context: softirq &(&hinfo->gc_work)->timer rcu_read_lock &pool->lock irq_context: softirq &(&hinfo->gc_work)->timer rcu_read_lock &pool->lock &obj_hash[i].lock irq_context: softirq &(&hinfo->gc_work)->timer rcu_read_lock &pool->lock &p->pi_lock irq_context: 0 (wq_completion)events_power_efficient (work_completion)(&(&hinfo->gc_work)->work) irq_context: 0 (wq_completion)events_power_efficient (work_completion)(&(&hinfo->gc_work)->work) &hinfo->lock irq_context: 0 (wq_completion)events_power_efficient (work_completion)(&(&hinfo->gc_work)->work) rcu_node_0 irq_context: 0 (wq_completion)events_power_efficient (work_completion)(&(&hinfo->gc_work)->work) &rcu_state.expedited_wq irq_context: 0 (wq_completion)events_power_efficient (work_completion)(&(&hinfo->gc_work)->work) &rcu_state.expedited_wq &p->pi_lock irq_context: 0 (wq_completion)events_power_efficient (work_completion)(&(&hinfo->gc_work)->work) &rcu_state.expedited_wq &p->pi_lock &rq->__lock irq_context: 0 (wq_completion)events_power_efficient (work_completion)(&(&hinfo->gc_work)->work) &rcu_state.expedited_wq &p->pi_lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)events_power_efficient (work_completion)(&(&hinfo->gc_work)->work) &rq->__lock irq_context: 0 (wq_completion)events_power_efficient (work_completion)(&(&hinfo->gc_work)->work) &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)events_power_efficient (work_completion)(&(&hinfo->gc_work)->work) &cfs_rq->removed.lock irq_context: 0 (wq_completion)events_power_efficient (work_completion)(&(&hinfo->gc_work)->work) &obj_hash[i].lock irq_context: 0 (wq_completion)events_power_efficient (work_completion)(&(&hinfo->gc_work)->work) pool_lock#2 irq_context: 0 (wq_completion)events_power_efficient (work_completion)(&(&hinfo->gc_work)->work) &base->lock irq_context: 0 (wq_completion)events_power_efficient (work_completion)(&(&hinfo->gc_work)->work) &base->lock &obj_hash[i].lock irq_context: 0 &xt[i].mutex remove_cache_srcu &____s->seqcount irq_context: softirq &(&hinfo->gc_work)->timer rcu_read_lock &pool->lock &p->pi_lock &rq->__lock irq_context: softirq &(&hinfo->gc_work)->timer rcu_read_lock &pool->lock &p->pi_lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 &mm->mmap_lock rcu_read_lock &rq->__lock &cfs_rq->removed.lock irq_context: softirq (&hsr->announce_timer) rcu_read_lock batched_entropy_u8.lock irq_context: softirq (&hsr->announce_timer) rcu_read_lock kfence_freelist_lock irq_context: softirq (&hsr->announce_timer) rcu_read_lock &hsr->seqnr_lock &meta->lock irq_context: softirq (&hsr->announce_timer) rcu_read_lock &hsr->seqnr_lock kfence_freelist_lock irq_context: 0 fs_reclaim mmu_notifier_invalidate_range_start &rq->__lock &cfs_rq->removed.lock irq_context: 0 vsock_table_lock irq_context: 0 sb_writers#8 iattr_mutex &rq->__lock irq_context: 0 sb_writers#8 iattr_mutex &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 &sb->s_type->i_mutex_key#10 sk_lock-AF_VSOCK irq_context: 0 &sb->s_type->i_mutex_key#10 sk_lock-AF_VSOCK &rq->__lock irq_context: 0 &sb->s_type->i_mutex_key#10 sk_lock-AF_VSOCK &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 &sb->s_type->i_mutex_key#10 sk_lock-AF_VSOCK slock-AF_VSOCK irq_context: 0 &sb->s_type->i_mutex_key#10 sk_lock-AF_VSOCK vsock_table_lock irq_context: 0 &sb->s_type->i_mutex_key#10 sk_lock-AF_VSOCK clock-AF_VSOCK irq_context: 0 &sb->s_type->i_mutex_key#10 sk_lock-AF_VSOCK rlock-AF_VSOCK irq_context: 0 &sb->s_type->i_mutex_key#10 slock-AF_VSOCK irq_context: 0 sb_writers#8 iattr_mutex &cfs_rq->removed.lock irq_context: 0 &xt[i].mutex remove_cache_srcu rcu_read_lock rcu_node_0 irq_context: 0 &xt[i].mutex remove_cache_srcu rcu_read_lock &rq->__lock irq_context: 0 &xt[i].mutex remove_cache_srcu rcu_read_lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: softirq lib/debugobjects.c:101 rcu_read_lock &pool->lock &p->pi_lock &rq->__lock &cfs_rq->removed.lock irq_context: 0 &mm->mmap_lock &base->lock irq_context: 0 &mm->mmap_lock &base->lock &obj_hash[i].lock irq_context: 0 sb_writers#4 &type->i_mutex_dir_key#3/1 &type->i_mutex_dir_key#3 &cfs_rq->removed.lock irq_context: 0 (wq_completion)events (work_completion)(&rdev->wiphy_work) &rdev->wiphy.mtx &wdev->mtx rcu_read_lock rcu_read_lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)wg-crypt-wg2#2 (work_completion)(&({ do { const void *__vpp_verify = (typeof((worker) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __asm__ ("" : "=r"(__ptr) : "0"((typeof(*((worker))) *)((worker)))); (typeof((typeof(*((worker))) *)((worker)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); })->work) &rq->__lock irq_context: 0 (wq_completion)wg-crypt-wg1 (work_completion)(&({ do { const void *__vpp_verify = (typeof((worker) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __asm__ ("" : "=r"(__ptr) : "0"((typeof(*((worker))) *)((worker)))); (typeof((typeof(*((worker))) *)((worker)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); })->work) &rq->__lock irq_context: 0 (wq_completion)wg-crypt-wg1 (work_completion)(&peer->transmit_packet_work) &rq->__lock irq_context: 0 (wq_completion)wg-crypt-wg1 (work_completion)(&peer->transmit_packet_work) &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 sb_writers#4 &type->i_mutex_dir_key#3/1 jbd2_handle &ei->i_data_sem &ei->i_es_lock batched_entropy_u8.lock irq_context: 0 sb_writers#4 &type->i_mutex_dir_key#3/1 jbd2_handle &ei->i_data_sem &ei->i_es_lock kfence_freelist_lock irq_context: 0 (wq_completion)events_power_efficient (work_completion)(&barr->work) irq_context: 0 (wq_completion)events_power_efficient (work_completion)(&barr->work) &x->wait#10 irq_context: 0 (wq_completion)events_power_efficient (work_completion)(&barr->work) &x->wait#10 &p->pi_lock irq_context: 0 &sb->s_type->i_mutex_key#9 &rq->__lock irq_context: 0 &sb->s_type->i_mutex_key#9 &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 bcm_notifier_lock irq_context: 0 rlock-AF_CAN irq_context: 0 &nft_net->commit_mutex (console_sem).lock irq_context: 0 &nft_net->commit_mutex console_lock console_srcu console_owner_lock irq_context: 0 &nft_net->commit_mutex console_lock console_srcu console_owner irq_context: 0 &nft_net->commit_mutex console_lock console_srcu console_owner &port_lock_key irq_context: 0 &nft_net->commit_mutex console_lock console_srcu console_owner console_owner_lock irq_context: 0 sb_writers#5 &type->i_mutex_dir_key#5/1 tomoyo_ss &cfs_rq->removed.lock irq_context: 0 sb_writers#4 &type->i_mutex_dir_key#3/1 sb_internal jbd2_handle &ei->i_data_sem &ei->i_es_lock &meta->lock irq_context: 0 sb_writers#4 &type->i_mutex_dir_key#3/1 sb_internal jbd2_handle &ei->i_data_sem &ei->i_es_lock kfence_freelist_lock irq_context: 0 &nft_net->commit_mutex &base->lock irq_context: 0 &nft_net->commit_mutex &base->lock &obj_hash[i].lock irq_context: 0 &nft_net->commit_mutex (work_completion)(&(&priv->gc_work)->work) irq_context: 0 &nft_net->commit_mutex &ht->mutex &rq->__lock irq_context: 0 &nft_net->commit_mutex &ht->mutex &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 tomoyo_ss rcu_read_lock &cfs_rq->removed.lock irq_context: 0 &nft_net->commit_mutex &ht->mutex rcu_read_lock pool_lock#2 irq_context: 0 sb_writers#8 &of->mutex kn->active#5 uevent_sock_mutex nl_table_wait.lock &p->pi_lock irq_context: 0 sb_writers#8 &of->mutex kn->active#5 uevent_sock_mutex nl_table_wait.lock &p->pi_lock &rq->__lock irq_context: 0 sb_writers#8 &of->mutex kn->active#5 uevent_sock_mutex nl_table_wait.lock &p->pi_lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 dup_mmap_sem &mm->mmap_lock &mm->mmap_lock/1 remove_cache_srcu rcu_read_lock &rq->__lock irq_context: 0 dup_mmap_sem &mm->mmap_lock &mm->mmap_lock/1 remove_cache_srcu rcu_read_lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)rcu_gp (work_completion)(&(&ssp->srcu_sup->work)->work) &cfs_rq->removed.lock irq_context: 0 sb_writers#4 &type->i_mutex_dir_key#3/1 tomoyo_ss rcu_node_0 irq_context: 0 &u->iolock rcu_read_lock &rcu_state.gp_wq irq_context: 0 &u->iolock rcu_read_lock &rcu_state.gp_wq &p->pi_lock irq_context: 0 &u->iolock rcu_read_lock &rcu_state.gp_wq &p->pi_lock &rq->__lock irq_context: 0 &u->iolock rcu_read_lock &rcu_state.gp_wq &p->pi_lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 &nft_net->commit_mutex &ht->mutex &rq->__lock &cfs_rq->removed.lock irq_context: softirq &(&bat_priv->tt.work)->timer rcu_read_lock &pool->lock/1 &p->pi_lock &rq->__lock &cfs_rq->removed.lock irq_context: 0 (wq_completion)events_unbound (reaper_work).work &ACCESS_PRIVATE(ssp->srcu_sup, lock) rcu_read_lock &pool->lock &p->pi_lock &rq->__lock &cfs_rq->removed.lock irq_context: 0 (wq_completion)events (work_completion)(&(&nsim_dev->trap_data->trap_report_dw)->work) &devlink->lock_key &nsim_trap_data->trap_lock batched_entropy_u8.lock crngs.lock irq_context: softirq (&hsr->announce_timer) rcu_read_lock init_task.mems_allowed_seq.seqcount irq_context: softirq (&hsr->announce_timer) rcu_read_lock &____s->seqcount irq_context: 0 &sb->s_type->i_mutex_key#10 bcm_notifier_lock irq_context: 0 &sb->s_type->i_mutex_key#10 sk_lock-AF_CAN irq_context: 0 &sb->s_type->i_mutex_key#10 sk_lock-AF_CAN slock-AF_CAN irq_context: 0 &sb->s_type->i_mutex_key#10 sk_lock-AF_CAN rcu_state.exp_mutex rcu_node_0 irq_context: 0 &sb->s_type->i_mutex_key#10 sk_lock-AF_CAN rcu_state.exp_mutex &obj_hash[i].lock irq_context: 0 &sb->s_type->i_mutex_key#10 sk_lock-AF_CAN rcu_state.exp_mutex rcu_read_lock &pool->lock irq_context: 0 &sb->s_type->i_mutex_key#10 sk_lock-AF_CAN rcu_state.exp_mutex rcu_read_lock &pool->lock &obj_hash[i].lock irq_context: 0 &sb->s_type->i_mutex_key#10 sk_lock-AF_CAN rcu_state.exp_mutex rcu_read_lock &pool->lock &p->pi_lock irq_context: 0 &sb->s_type->i_mutex_key#10 sk_lock-AF_CAN rcu_state.exp_mutex rcu_read_lock &pool->lock &p->pi_lock &rq->__lock irq_context: 0 &sb->s_type->i_mutex_key#10 sk_lock-AF_CAN rcu_state.exp_mutex rcu_read_lock &pool->lock &p->pi_lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 &sb->s_type->i_mutex_key#10 sk_lock-AF_CAN rcu_state.exp_mutex &rq->__lock irq_context: 0 &sb->s_type->i_mutex_key#10 sk_lock-AF_CAN rcu_state.exp_mutex &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 &sb->s_type->i_mutex_key#10 sk_lock-AF_CAN &obj_hash[i].lock irq_context: 0 &sb->s_type->i_mutex_key#10 sk_lock-AF_CAN clock-AF_CAN irq_context: 0 &sb->s_type->i_mutex_key#10 slock-AF_CAN irq_context: 0 &sb->s_type->i_mutex_key#10 rlock-AF_CAN irq_context: 0 &sb->s_type->i_mutex_key#10 elock-AF_CAN irq_context: 0 pcpu_lock stock_lock rcu_read_lock per_cpu_ptr(&cgroup_rstat_cpu_lock, cpu) irq_context: 0 &f->f_pos_lock &type->i_mutex_dir_key#3 sb_writers#4 jbd2_handle &journal->j_state_lock &journal->j_wait_commit irq_context: 0 &f->f_pos_lock &type->i_mutex_dir_key#3 sb_writers#4 jbd2_handle &journal->j_state_lock &journal->j_wait_commit &p->pi_lock irq_context: 0 &f->f_pos_lock &type->i_mutex_dir_key#3 sb_writers#4 jbd2_handle &journal->j_state_lock &journal->j_wait_commit &p->pi_lock &rq->__lock irq_context: 0 &f->f_pos_lock &type->i_mutex_dir_key#3 sb_writers#4 jbd2_handle &journal->j_state_lock &journal->j_wait_commit &p->pi_lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 &f->f_pos_lock &type->i_mutex_dir_key#3 sb_writers#4 jbd2_handle &journal->j_wait_updates &p->pi_lock irq_context: 0 &f->f_pos_lock &type->i_mutex_dir_key#3 sb_writers#4 jbd2_handle &journal->j_wait_updates &p->pi_lock &rq->__lock irq_context: 0 &f->f_pos_lock &type->i_mutex_dir_key#3 sb_writers#4 jbd2_handle &journal->j_wait_updates &p->pi_lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)events (work_completion)(&(&nsim_dev->trap_data->trap_report_dw)->work) &devlink->lock_key#4 &nsim_trap_data->trap_lock batched_entropy_u8.lock crngs.lock irq_context: 0 dup_mmap_sem &mm->mmap_lock &mm->mmap_lock/1 &meta->lock irq_context: 0 tomoyo_ss rcu_read_lock pgd_lock irq_context: 0 tomoyo_ss rcu_read_lock stock_lock irq_context: 0 tomoyo_ss rcu_read_lock &obj_hash[i].lock irq_context: 0 tomoyo_ss rcu_read_lock key irq_context: 0 tomoyo_ss rcu_read_lock pcpu_lock irq_context: 0 tomoyo_ss rcu_read_lock percpu_counters_lock irq_context: 0 tomoyo_ss rcu_read_lock pcpu_lock stock_lock irq_context: 0 cb_lock &dir->lock#2 irq_context: 0 &f->f_pos_lock &type->i_mutex_dir_key#3 sb_writers#4 remove_cache_srcu irq_context: 0 &f->f_pos_lock &type->i_mutex_dir_key#3 sb_writers#4 remove_cache_srcu quarantine_lock irq_context: 0 sb_writers#4 &type->i_mutex_dir_key#3/1 sb_internal jbd2_handle rcu_read_lock &rcu_state.gp_wq irq_context: 0 sb_writers#4 &type->i_mutex_dir_key#3/1 sb_internal jbd2_handle rcu_read_lock &rcu_state.gp_wq &p->pi_lock irq_context: 0 sb_writers#4 &type->i_mutex_dir_key#3/1 sb_internal jbd2_handle rcu_read_lock &rcu_state.gp_wq &p->pi_lock &rq->__lock irq_context: 0 sb_writers#4 &type->i_mutex_dir_key#3/1 sb_internal jbd2_handle rcu_read_lock &rcu_state.gp_wq &p->pi_lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 sb_writers#4 &type->i_mutex_dir_key#3/1 rcu_read_lock &cfs_rq->removed.lock irq_context: 0 sb_writers#4 &type->i_mutex_dir_key#3/1 rcu_read_lock &obj_hash[i].lock irq_context: 0 misc_mtx &n->list_lock irq_context: 0 misc_mtx &n->list_lock &c->lock irq_context: 0 &type->i_mutex_dir_key#5 &cfs_rq->removed.lock irq_context: 0 sb_writers#4 &type->i_mutex_dir_key#3/1 &sb->s_type->i_mutex_key#8 jbd2_handle &cfs_rq->removed.lock irq_context: 0 sb_writers#4 &type->i_mutex_dir_key#3/1 &sb->s_type->i_mutex_key#8 jbd2_handle &obj_hash[i].lock irq_context: 0 &p->lock &of->mutex kn->active#5 &cfs_rq->removed.lock irq_context: softirq (&ndev->rs_timer) rcu_read_lock rcu_read_lock rcu_read_lock_bh rcu_read_lock &obj_hash[i].lock pool_lock irq_context: 0 &type->i_mutex_dir_key#7 &dentry->d_lock &wq irq_context: 0 misc_mtx &____s->seqcount#2 irq_context: 0 sb_writers#11 &type->i_mutex_dir_key#7/1 cgroup_mutex lock kernfs_idr_lock &c->lock irq_context: 0 sb_writers#11 &type->i_mutex_dir_key#7/1 cgroup_mutex stock_lock irq_context: 0 sb_writers#11 &type->i_mutex_dir_key#7/1 cgroup_mutex rcu_read_lock pool_lock#2 irq_context: 0 &f->f_pos_lock &type->i_mutex_dir_key#3 sb_writers#4 remove_cache_srcu &c->lock irq_context: 0 &f->f_pos_lock &type->i_mutex_dir_key#3 sb_writers#4 remove_cache_srcu &n->list_lock irq_context: 0 &f->f_pos_lock &type->i_mutex_dir_key#3 sb_writers#4 remove_cache_srcu rcu_read_lock pool_lock#2 irq_context: 0 &f->f_pos_lock &type->i_mutex_dir_key#3 sb_writers#4 remove_cache_srcu &obj_hash[i].lock irq_context: 0 &f->f_pos_lock &type->i_mutex_dir_key#3 sb_writers#4 remove_cache_srcu &rq->__lock irq_context: 0 &f->f_pos_lock &type->i_mutex_dir_key#3 sb_writers#4 remove_cache_srcu &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 sb_writers#4 &type->i_mutex_dir_key#3/1 jbd2_handle &cfs_rq->removed.lock irq_context: 0 &xt[i].mutex fs_reclaim &cfs_rq->removed.lock irq_context: 0 &xt[i].mutex fs_reclaim &obj_hash[i].lock irq_context: 0 sb_writers#11 &type->i_mutex_dir_key#7/1 cgroup_mutex fs_reclaim &rq->__lock irq_context: 0 sb_writers#11 &type->i_mutex_dir_key#7/1 cgroup_mutex fs_reclaim &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 sk_lock-AF_UNIX &mm->mmap_lock irq_context: 0 (wq_completion)bat_events (work_completion)(&(&bat_priv->orig_work)->work) &cfs_rq->removed.lock irq_context: 0 &sb->s_type->i_mutex_key#12 &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 &sb->s_type->i_mutex_key#10 rcu_read_lock &ei->socket.wq.wait &p->pi_lock &cfs_rq->removed.lock irq_context: 0 sb_writers#4 &type->i_mutex_dir_key#3/1 &sb->s_type->i_mutex_key#8 &rq->__lock &cfs_rq->removed.lock irq_context: 0 sb_writers#4 sb_internal &cfs_rq->removed.lock irq_context: 0 sb_writers#4 &type->i_mutex_dir_key#3/1 jbd2_handle &ei->i_data_sem &ei->i_es_lock &n->list_lock irq_context: 0 sb_writers#4 &type->i_mutex_dir_key#3/1 jbd2_handle &ei->i_data_sem &ei->i_es_lock &n->list_lock &c->lock irq_context: 0 sk_lock-AF_INET fs_reclaim mmu_notifier_invalidate_range_start &rq->__lock irq_context: 0 &qs->lock irq_context: 0 (wq_completion)events_unbound (work_completion)(&map->work) vmap_area_lock irq_context: 0 (wq_completion)events_unbound (work_completion)(&map->work) purge_vmap_area_lock irq_context: 0 (wq_completion)events_unbound (work_completion)(&map->work) rcu_read_lock pool_lock#2 irq_context: 0 sb_writers#4 &type->i_mutex_dir_key#3/1 batched_entropy_u8.lock irq_context: 0 sb_writers#4 &type->i_mutex_dir_key#3/1 kfence_freelist_lock irq_context: 0 sb_writers#4 sb_internal remove_cache_srcu pool_lock#2 irq_context: 0 rcu_read_lock stock_lock irq_context: 0 rcu_read_lock pcpu_lock stock_lock irq_context: 0 &sb->s_type->i_mutex_key#9 remove_cache_srcu &c->lock irq_context: 0 &sb->s_type->i_mutex_key#9 remove_cache_srcu &n->list_lock irq_context: 0 &sb->s_type->i_mutex_key#9 remove_cache_srcu rcu_read_lock pool_lock#2 irq_context: 0 &sb->s_type->i_mutex_key#9 remove_cache_srcu &obj_hash[i].lock irq_context: 0 &sb->s_type->i_mutex_key#9 batched_entropy_u8.lock irq_context: 0 &sb->s_type->i_mutex_key#9 kfence_freelist_lock irq_context: 0 (wq_completion)events (work_completion)(&(&krcp->monitor_work)->work) rcu_callback stock_lock rcu_read_lock per_cpu_ptr(&cgroup_rstat_cpu_lock, cpu) irq_context: 0 sb_writers#8 kn->active#5 remove_cache_srcu &c->lock irq_context: 0 sb_writers#8 kn->active#5 remove_cache_srcu &n->list_lock irq_context: 0 sb_writers#8 kn->active#5 remove_cache_srcu &obj_hash[i].lock irq_context: 0 &u->lock &f->f_owner.lock irq_context: 0 sb_writers#4 &type->i_mutex_dir_key#3/1 &type->i_mutex_dir_key#3 rcu_read_lock rcu_node_0 irq_context: 0 sb_writers#4 &type->i_mutex_dir_key#3/1 &type->i_mutex_dir_key#3 rcu_read_lock &rq->__lock irq_context: 0 sb_writers#4 &type->i_mutex_dir_key#3/1 &type->i_mutex_dir_key#3 rcu_read_lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 &sb->s_type->i_mutex_key#10 sk_lock-AF_INET &cfs_rq->removed.lock irq_context: 0 sb_writers#4 sb_internal jbd2_handle &sbi->s_orphan_lock &mapping->private_lock irq_context: 0 sb_writers#8 rcu_read_lock &cfs_rq->removed.lock irq_context: 0 sb_writers#8 rcu_read_lock &obj_hash[i].lock irq_context: 0 sb_writers#8 rcu_read_lock &rcu_state.gp_wq irq_context: 0 sb_writers#8 rcu_read_lock &rcu_state.gp_wq &p->pi_lock irq_context: 0 sb_writers#8 rcu_read_lock &rcu_state.gp_wq &p->pi_lock &rq->__lock irq_context: 0 sb_writers#8 rcu_read_lock &rcu_state.gp_wq &p->pi_lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 sb_writers#8 tomoyo_ss remove_cache_srcu &c->lock irq_context: 0 sb_writers#8 tomoyo_ss remove_cache_srcu &n->list_lock irq_context: 0 sb_writers#8 tomoyo_ss remove_cache_srcu &obj_hash[i].lock irq_context: 0 sb_writers#8 tomoyo_ss remove_cache_srcu pool_lock#2 irq_context: 0 &f->f_pos_lock rcu_read_lock &rcu_state.gp_wq irq_context: 0 &f->f_pos_lock rcu_read_lock &rcu_state.gp_wq &p->pi_lock irq_context: 0 &f->f_pos_lock rcu_read_lock &rcu_state.gp_wq &p->pi_lock &rq->__lock irq_context: 0 &f->f_pos_lock rcu_read_lock &rcu_state.gp_wq &p->pi_lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)bat_events (work_completion)(&(&forw_packet_aggr->delayed_work)->work) &hard_iface->bat_iv.ogm_buff_mutex kfence_freelist_lock irq_context: 0 sb_writers#4 &type->i_mutex_dir_key#3/1 sb_internal remove_cache_srcu irq_context: 0 sb_writers#4 &type->i_mutex_dir_key#3/1 sb_internal remove_cache_srcu quarantine_lock irq_context: 0 sb_writers#4 &type->i_mutex_dir_key#3/1 sb_internal remove_cache_srcu &c->lock irq_context: 0 sb_writers#4 &type->i_mutex_dir_key#3/1 sb_internal remove_cache_srcu &n->list_lock irq_context: 0 sb_writers#4 &type->i_mutex_dir_key#3/1 sb_internal remove_cache_srcu rcu_read_lock pool_lock#2 irq_context: 0 sb_writers#4 &type->i_mutex_dir_key#3/1 sb_internal remove_cache_srcu &obj_hash[i].lock irq_context: 0 sb_writers#4 &type->i_mutex_dir_key#3/1 sb_internal remove_cache_srcu &rq->__lock irq_context: 0 sb_writers#4 &type->i_mutex_dir_key#3/1 sb_internal remove_cache_srcu &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)events (work_completion)(&(&nsim_dev->trap_data->trap_report_dw)->work) &devlink->lock_key &cfs_rq->removed.lock irq_context: 0 (wq_completion)events (work_completion)(&(&nsim_dev->trap_data->trap_report_dw)->work) &devlink->lock_key &obj_hash[i].lock irq_context: 0 (wq_completion)events (work_completion)(&(&nsim_dev->trap_data->trap_report_dw)->work) &devlink->lock_key pool_lock#2 irq_context: 0 ebt_mutex &mm->mmap_lock &rq->__lock irq_context: 0 ebt_mutex &mm->mmap_lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 ebt_mutex ebt_mutex.wait_lock irq_context: 0 ebt_mutex.wait_lock irq_context: 0 &sb->s_type->i_mutex_key#10 stock_lock rcu_read_lock per_cpu_ptr(&cgroup_rstat_cpu_lock, cpu) irq_context: 0 sb_writers#4 &type->i_mutex_dir_key#3/1 &fsnotify_mark_srcu &rq->__lock irq_context: 0 sb_writers#4 &type->i_mutex_dir_key#3/1 &fsnotify_mark_srcu &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 dup_mmap_sem &mm->mmap_lock &mm->mmap_lock/1 rcu_read_lock rcu_read_lock &cfs_rq->removed.lock irq_context: 0 dup_mmap_sem &mm->mmap_lock &mm->mmap_lock/1 rcu_read_lock rcu_read_lock &obj_hash[i].lock irq_context: 0 sb_writers#5 &type->i_mutex_dir_key#5 remove_cache_srcu &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 sb_writers#4 &type->i_mutex_dir_key#3/1 &sb->s_type->i_mutex_key#8 jbd2_handle &sbi->s_orphan_lock &lock->wait_lock irq_context: 0 dup_mmap_sem &mm->mmap_lock &mm->mmap_lock/1 &anon_vma->rwsem rcu_node_0 irq_context: 0 sb_writers#4 sb_internal jbd2_handle &sbi->s_orphan_lock rcu_read_lock &rcu_state.gp_wq irq_context: 0 sb_writers#4 sb_internal jbd2_handle &sbi->s_orphan_lock rcu_read_lock &rcu_state.gp_wq &p->pi_lock irq_context: 0 sb_writers#4 sb_internal jbd2_handle &sbi->s_orphan_lock rcu_read_lock &rcu_state.gp_wq &p->pi_lock &rq->__lock irq_context: 0 sb_writers#4 sb_internal jbd2_handle &sbi->s_orphan_lock rcu_read_lock &rcu_state.gp_wq &p->pi_lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 sb_writers#4 &type->i_mutex_dir_key#3/1 sb_internal jbd2_handle &lock->wait_lock irq_context: 0 sb_writers#4 &type->i_mutex_dir_key#3/1 sb_internal jbd2_handle &p->pi_lock irq_context: 0 sb_writers#4 &type->i_mutex_dir_key#3/1 &sb->s_type->i_mutex_key#8 jbd2_handle &p->pi_lock &rq->__lock irq_context: 0 sb_writers#4 &type->i_mutex_dir_key#3/1 &sb->s_type->i_mutex_key#8 jbd2_handle &p->pi_lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 &f->f_pos_lock &type->i_mutex_dir_key#3 sb_writers#4 mmu_notifier_invalidate_range_start &rq->__lock irq_context: 0 dup_mmap_sem &mm->mmap_lock &mm->mmap_lock/1 &anon_vma->rwsem rcu_read_lock &rcu_state.gp_wq irq_context: 0 &f->f_pos_lock &type->i_mutex_dir_key#3 sb_writers#4 mmu_notifier_invalidate_range_start &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 dup_mmap_sem &mm->mmap_lock &mm->mmap_lock/1 &anon_vma->rwsem rcu_read_lock &rcu_state.gp_wq &p->pi_lock irq_context: 0 dup_mmap_sem &mm->mmap_lock &mm->mmap_lock/1 &anon_vma->rwsem rcu_read_lock &rcu_state.gp_wq &p->pi_lock &rq->__lock irq_context: 0 dup_mmap_sem &mm->mmap_lock &mm->mmap_lock/1 &anon_vma->rwsem rcu_read_lock &rcu_state.gp_wq &p->pi_lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 &sb->s_type->i_mutex_key#10 rcu_read_lock &rcu_state.gp_wq &p->pi_lock irq_context: 0 &sb->s_type->i_mutex_key#10 rcu_read_lock &rcu_state.gp_wq &p->pi_lock &rq->__lock irq_context: 0 &sb->s_type->i_mutex_key#10 rcu_read_lock &rcu_state.gp_wq &p->pi_lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 sb_writers#3 &sb->s_type->i_mutex_key#9 &cfs_rq->removed.lock irq_context: 0 &type->i_mutex_dir_key#7 &n->list_lock &c->lock irq_context: 0 &type->i_mutex_dir_key#7 &rq->__lock irq_context: 0 &type->i_mutex_dir_key#7 &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 sb_writers#4 &type->i_mutex_dir_key#3/1 sb_internal jbd2_handle &ei->i_data_sem &ei->i_es_lock &base->lock irq_context: 0 sb_writers#4 &type->i_mutex_dir_key#3/1 sb_internal jbd2_handle &ei->i_data_sem &ei->i_es_lock &base->lock &obj_hash[i].lock irq_context: 0 &sb->s_type->i_mutex_key#10 sk_lock-AF_PPPOX irq_context: 0 &sb->s_type->i_mutex_key#10 sk_lock-AF_PPPOX slock-AF_PPPOX irq_context: 0 &sb->s_type->i_mutex_key#10 sk_lock-AF_PPPOX &pn->hash_lock irq_context: 0 &sb->s_type->i_mutex_key#10 sk_lock-AF_PPPOX clock-AF_PPPOX irq_context: 0 &sb->s_type->i_mutex_key#10 sk_lock-AF_PPPOX rlock-AF_PPPOX irq_context: 0 &sb->s_type->i_mutex_key#10 slock-AF_PPPOX irq_context: 0 (wq_completion)events (work_completion)(&rdev->wiphy_work) &rdev->wiphy.mtx &wdev->mtx &rdev->bss_lock &base->lock irq_context: 0 (wq_completion)events (work_completion)(&rdev->wiphy_work) &rdev->wiphy.mtx &wdev->mtx &rdev->bss_lock &base->lock &obj_hash[i].lock irq_context: 0 sb_writers#4 sb_internal jbd2_handle pgd_lock irq_context: 0 sb_writers#4 sb_internal jbd2_handle stock_lock irq_context: 0 sb_writers#4 sb_internal jbd2_handle rcu_read_lock pool_lock#2 irq_context: 0 sb_writers#4 sb_internal jbd2_handle key irq_context: 0 sb_writers#4 sb_internal jbd2_handle pcpu_lock irq_context: 0 sb_writers#4 sb_internal jbd2_handle percpu_counters_lock irq_context: 0 sb_writers#4 sb_internal jbd2_handle pcpu_lock stock_lock irq_context: softirq (&peer->timer_send_keepalive) &n->list_lock irq_context: softirq (&peer->timer_send_keepalive) &n->list_lock &c->lock irq_context: 0 sb_writers#4 &type->i_mutex_dir_key#3/1 tomoyo_ss batched_entropy_u8.lock irq_context: 0 sb_writers#4 &type->i_mutex_dir_key#3/1 tomoyo_ss kfence_freelist_lock irq_context: 0 sb_writers#4 &type->i_mutex_dir_key#3/1 tomoyo_ss &meta->lock irq_context: 0 sb_writers#5 &type->i_mutex_dir_key#5/1 remove_cache_srcu pool_lock#2 irq_context: 0 sb_writers#3 oom_adj_mutex &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 &fsnotify_mark_srcu fs_reclaim mmu_notifier_invalidate_range_start &rq->__lock irq_context: 0 &fsnotify_mark_srcu fs_reclaim mmu_notifier_invalidate_range_start &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 sb_writers#4 &type->i_mutex_dir_key#3/1 remove_cache_srcu rcu_read_lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 kn->active#5 remove_cache_srcu rcu_read_lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 rcu_read_lock rcu_read_lock rcu_read_lock rcu_read_lock stock_lock irq_context: 0 rcu_read_lock rcu_read_lock rcu_read_lock rcu_read_lock pool_lock#2 irq_context: 0 rcu_read_lock rcu_read_lock rcu_read_lock rcu_read_lock &dir->lock#2 irq_context: 0 dup_mmap_sem &mm->mmap_lock &mm->mmap_lock/1 &anon_vma->rwsem &cfs_rq->removed.lock irq_context: 0 dup_mmap_sem &mm->mmap_lock &mm->mmap_lock/1 &anon_vma->rwsem &obj_hash[i].lock irq_context: 0 &p->lock &cfs_rq->removed.lock irq_context: 0 &p->lock &obj_hash[i].lock irq_context: 0 sb_writers#8 &of->mutex kn->active#5 uevent_sock_mutex &cfs_rq->removed.lock irq_context: 0 sk_lock-AF_PPPOX irq_context: 0 sk_lock-AF_PPPOX slock-AF_PPPOX irq_context: 0 slock-AF_PPPOX irq_context: 0 tasklist_lock per_cpu_ptr(&cgroup_rstat_cpu_lock, cpu) irq_context: 0 sb_writers#4 &type->i_mutex_dir_key#3/1 jbd2_handle remove_cache_srcu irq_context: 0 sb_writers#4 &type->i_mutex_dir_key#3/1 jbd2_handle remove_cache_srcu quarantine_lock irq_context: 0 sb_writers#4 &type->i_mutex_dir_key#3/1 jbd2_handle remove_cache_srcu &c->lock irq_context: 0 sb_writers#4 &type->i_mutex_dir_key#3/1 jbd2_handle remove_cache_srcu &n->list_lock irq_context: 0 sb_writers#4 &type->i_mutex_dir_key#3/1 jbd2_handle remove_cache_srcu rcu_read_lock pool_lock#2 irq_context: 0 sb_writers#4 &type->i_mutex_dir_key#3/1 jbd2_handle remove_cache_srcu &obj_hash[i].lock irq_context: 0 misc_mtx remove_cache_srcu pool_lock#2 irq_context: 0 sb_writers#5 &type->i_mutex_dir_key#5/1 &sb->s_type->i_mutex_key#12 &sb->s_type->i_mutex_key#12/4 fs_reclaim &rq->__lock irq_context: 0 sb_writers#5 &type->i_mutex_dir_key#5/1 &sb->s_type->i_mutex_key#12 &sb->s_type->i_mutex_key#12/4 fs_reclaim &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: softirq &(&bat_priv->dat.work)->timer rcu_read_lock &pool->lock/1 &p->pi_lock &cfs_rq->removed.lock irq_context: 0 sb_writers#3 &sb->s_type->i_mutex_key#9 &rcu_state.gp_wq irq_context: 0 sb_writers#3 &sb->s_type->i_mutex_key#9 &rcu_state.gp_wq &p->pi_lock irq_context: 0 sb_writers#3 &sb->s_type->i_mutex_key#9 &rcu_state.gp_wq &p->pi_lock &rq->__lock irq_context: 0 sb_writers#3 &sb->s_type->i_mutex_key#9 &rcu_state.gp_wq &p->pi_lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 sb_writers#8 fs_reclaim &rq->__lock irq_context: 0 sb_writers#8 fs_reclaim &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 sk_lock-AF_INET k-sk_lock-AF_INET crngs.lock irq_context: 0 sk_lock-AF_INET k-sk_lock-AF_INET &token_hash[i].lock irq_context: 0 sk_lock-AF_INET k-sk_lock-AF_INET rcu_read_lock rcu_read_lock &c->lock irq_context: 0 sk_lock-AF_INET k-sk_lock-AF_INET &____s->seqcount irq_context: 0 sk_lock-AF_INET k-sk_lock-AF_INET rcu_read_lock tcp_metrics_lock &c->lock irq_context: 0 sb_writers#4 sb_internal jbd2_handle &rq->__lock &cfs_rq->removed.lock irq_context: 0 cb_lock genl_mutex sk_lock-AF_INET irq_context: 0 cb_lock genl_mutex sk_lock-AF_INET slock-AF_INET irq_context: 0 cb_lock genl_mutex sk_lock-AF_INET &msk->pm.lock irq_context: 0 cb_lock genl_mutex sk_lock-AF_INET k-sk_lock-AF_INET irq_context: 0 cb_lock genl_mutex sk_lock-AF_INET k-sk_lock-AF_INET k-slock-AF_INET irq_context: 0 cb_lock genl_mutex sk_lock-AF_INET k-sk_lock-AF_INET k-slock-AF_INET &c->lock irq_context: 0 cb_lock genl_mutex sk_lock-AF_INET k-sk_lock-AF_INET k-slock-AF_INET pool_lock#2 irq_context: 0 cb_lock genl_mutex sk_lock-AF_INET k-sk_lock-AF_INET k-slock-AF_INET rcu_read_lock rcu_read_lock &____s->seqcount#7 irq_context: 0 cb_lock genl_mutex sk_lock-AF_INET k-sk_lock-AF_INET k-slock-AF_INET rcu_read_lock rcu_read_lock &ct->lock irq_context: 0 cb_lock genl_mutex sk_lock-AF_INET k-sk_lock-AF_INET k-slock-AF_INET rcu_read_lock rcu_read_lock &____s->seqcount#10 irq_context: 0 cb_lock genl_mutex sk_lock-AF_INET k-sk_lock-AF_INET k-slock-AF_INET rcu_read_lock rcu_read_lock rcu_read_lock_bh &list->lock#5 irq_context: 0 cb_lock genl_mutex sk_lock-AF_INET k-slock-AF_INET irq_context: softirq rcu_read_lock rcu_read_lock k-slock-AF_INET/1 tk_core.seq.seqcount irq_context: softirq rcu_read_lock rcu_read_lock k-slock-AF_INET/1 slock-AF_INET irq_context: softirq rcu_read_lock rcu_read_lock k-slock-AF_INET/1 &obj_hash[i].lock irq_context: 0 cb_lock genl_mutex sk_lock-AF_INET k-sk_lock-AF_INET fs_reclaim irq_context: 0 cb_lock genl_mutex sk_lock-AF_INET k-sk_lock-AF_INET fs_reclaim mmu_notifier_invalidate_range_start irq_context: 0 cb_lock genl_mutex sk_lock-AF_INET k-sk_lock-AF_INET pool_lock#2 irq_context: 0 cb_lock genl_mutex sk_lock-AF_INET k-sk_lock-AF_INET tk_core.seq.seqcount irq_context: 0 cb_lock genl_mutex sk_lock-AF_INET k-sk_lock-AF_INET rcu_read_lock rcu_read_lock &____s->seqcount#7 irq_context: 0 cb_lock genl_mutex sk_lock-AF_INET k-sk_lock-AF_INET rcu_read_lock rcu_read_lock &ct->lock irq_context: 0 cb_lock genl_mutex sk_lock-AF_INET k-sk_lock-AF_INET rcu_read_lock rcu_read_lock &____s->seqcount#10 irq_context: 0 cb_lock genl_mutex sk_lock-AF_INET k-sk_lock-AF_INET rcu_read_lock rcu_read_lock rcu_read_lock_bh &list->lock#5 irq_context: 0 cb_lock genl_mutex sk_lock-AF_INET k-sk_lock-AF_INET &obj_hash[i].lock irq_context: 0 cb_lock genl_mutex sk_lock-AF_INET k-sk_lock-AF_INET &base->lock irq_context: 0 &child->perf_event_mutex &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 cb_lock genl_mutex sk_lock-AF_INET k-sk_lock-AF_INET &base->lock &obj_hash[i].lock irq_context: 0 cb_lock genl_mutex sk_lock-AF_INET k-sk_lock-AF_INET rcu_read_lock &pool->lock irq_context: 0 cb_lock genl_mutex sk_lock-AF_INET k-sk_lock-AF_INET rcu_read_lock &pool->lock &obj_hash[i].lock irq_context: 0 cb_lock genl_mutex sk_lock-AF_INET k-sk_lock-AF_INET rcu_read_lock &pool->lock pool_lock#2 irq_context: 0 cb_lock genl_mutex sk_lock-AF_INET k-sk_lock-AF_INET rcu_read_lock &pool->lock &p->pi_lock irq_context: 0 cb_lock genl_mutex sk_lock-AF_INET k-sk_lock-AF_INET rcu_read_lock &pool->lock &p->pi_lock &rq->__lock irq_context: 0 cb_lock genl_mutex sk_lock-AF_INET k-sk_lock-AF_INET rcu_read_lock &pool->lock &p->pi_lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 cb_lock genl_mutex sk_lock-AF_INET k-sk_lock-AF_INET &rq->__lock irq_context: 0 cb_lock genl_mutex sk_lock-AF_INET k-sk_lock-AF_INET &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 cb_lock genl_mutex sk_lock-AF_INET k-sk_lock-AF_INET slock-AF_INET irq_context: 0 cb_lock genl_mutex sk_lock-AF_INET k-sk_lock-AF_INET stock_lock irq_context: 0 cb_lock genl_mutex sk_lock-AF_INET k-sk_lock-AF_INET &tcp_hashinfo.bhash[i].lock irq_context: 0 cb_lock genl_mutex sk_lock-AF_INET k-sk_lock-AF_INET &tcp_hashinfo.bhash[i].lock &tcp_hashinfo.bhash2[i].lock irq_context: 0 cb_lock genl_mutex sk_lock-AF_INET k-sk_lock-AF_INET &hashinfo->ehash_locks[i] irq_context: 0 cb_lock genl_mutex sk_lock-AF_INET k-sk_lock-AF_INET/1 irq_context: 0 cb_lock genl_mutex sk_lock-AF_INET k-sk_lock-AF_INET/1 k-slock-AF_INET irq_context: 0 cb_lock genl_mutex sk_lock-AF_INET k-sk_lock-AF_INET/1 &obj_hash[i].lock irq_context: 0 cb_lock genl_mutex sk_lock-AF_INET k-sk_lock-AF_INET/1 slock-AF_INET irq_context: 0 cb_lock genl_mutex slock-AF_INET irq_context: 0 cb_lock genl_mutex slock-AF_INET &sk->sk_lock.wq irq_context: 0 cb_lock genl_mutex slock-AF_INET &sk->sk_lock.wq &p->pi_lock irq_context: 0 cb_lock genl_mutex slock-AF_INET &sk->sk_lock.wq &p->pi_lock &rq->__lock irq_context: 0 cb_lock genl_mutex slock-AF_INET &sk->sk_lock.wq &p->pi_lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)events (work_completion)(&msk->work) sk_lock-AF_INET k-sk_lock-AF_INET/1 irq_context: 0 (wq_completion)events (work_completion)(&msk->work) sk_lock-AF_INET k-sk_lock-AF_INET/1 k-slock-AF_INET irq_context: 0 (wq_completion)events (work_completion)(&msk->work) sk_lock-AF_INET k-sk_lock-AF_INET/1 &obj_hash[i].lock irq_context: 0 (wq_completion)events (work_completion)(&msk->work) sk_lock-AF_INET k-sk_lock-AF_INET/1 slock-AF_INET irq_context: 0 (wq_completion)events (work_completion)(&msk->work) sk_lock-AF_INET k-slock-AF_INET irq_context: 0 tomoyo_ss &rq->__lock &cfs_rq->removed.lock irq_context: 0 &vma->vm_lock->lock &n->list_lock irq_context: 0 &vma->vm_lock->lock &n->list_lock &c->lock irq_context: 0 sk_lock-AF_INET6 &mm->mmap_lock rcu_read_lock rcu_node_0 irq_context: 0 sk_lock-AF_INET6 &mm->mmap_lock rcu_read_lock &rq->__lock irq_context: 0 sk_lock-AF_INET6 &mm->mmap_lock rcu_read_lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 __ip_vs_mutex irq_context: 0 __ip_vs_mutex (console_sem).lock irq_context: 0 __ip_vs_mutex console_lock console_srcu console_owner_lock irq_context: 0 __ip_vs_mutex console_lock console_srcu console_owner irq_context: 0 __ip_vs_mutex console_lock console_srcu console_owner &port_lock_key irq_context: 0 __ip_vs_mutex console_lock console_srcu console_owner console_owner_lock irq_context: 0 __ip_vs_mutex &rq->__lock irq_context: 0 __ip_vs_mutex &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 sk_lock-AF_INET6 &mm->mmap_lock rcu_read_lock &rcu_state.gp_wq irq_context: 0 sk_lock-AF_INET6 &mm->mmap_lock rcu_read_lock &rcu_state.gp_wq &p->pi_lock irq_context: 0 sk_lock-AF_INET6 &mm->mmap_lock rcu_read_lock &rcu_state.gp_wq &p->pi_lock &rq->__lock irq_context: softirq slock-AF_INET rcu_read_lock rcu_read_lock rcu_read_lock_bh &list->lock#5 irq_context: 0 sk_lock-AF_INET6 &mm->mmap_lock rcu_read_lock &rcu_state.gp_wq &p->pi_lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 sk_lock-AF_INET6 &mm->mmap_lock &rq->__lock irq_context: 0 sk_lock-AF_INET6 &mm->mmap_lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: softirq slock-AF_INET hrtimer_bases.lock irq_context: softirq slock-AF_INET hrtimer_bases.lock &obj_hash[i].lock irq_context: 0 &f->f_pos_lock &type->i_mutex_dir_key#3 remove_cache_srcu rcu_read_lock rcu_node_0 irq_context: 0 &f->f_pos_lock &type->i_mutex_dir_key#3 remove_cache_srcu rcu_read_lock &rq->__lock irq_context: 0 &f->f_pos_lock &type->i_mutex_dir_key#3 remove_cache_srcu rcu_read_lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: softirq (&icsk->icsk_retransmit_timer) slock-AF_INET tk_core.seq.seqcount irq_context: softirq (&icsk->icsk_retransmit_timer) slock-AF_INET rcu_read_lock rcu_read_lock &____s->seqcount#7 irq_context: softirq (&icsk->icsk_retransmit_timer) slock-AF_INET rcu_read_lock rcu_read_lock &ct->lock irq_context: softirq (&icsk->icsk_retransmit_timer) slock-AF_INET rcu_read_lock rcu_read_lock &____s->seqcount#10 irq_context: softirq (&icsk->icsk_retransmit_timer) slock-AF_INET rcu_read_lock rcu_read_lock rcu_read_lock_bh &list->lock#5 irq_context: softirq (&icsk->icsk_retransmit_timer) slock-AF_INET &obj_hash[i].lock irq_context: softirq (&icsk->icsk_retransmit_timer) slock-AF_INET &base->lock irq_context: softirq (&icsk->icsk_retransmit_timer) slock-AF_INET &base->lock &obj_hash[i].lock irq_context: 0 epnested_mutex &ep->mtx/1 irq_context: 0 epnested_mutex &ep->mtx stock_lock irq_context: 0 epnested_mutex &ep->mtx &ep->mtx/1 irq_context: 0 epnested_mutex &ep->mtx &ep->mtx/1 &f->f_lock irq_context: 0 epnested_mutex &ep->mtx &ep->mtx/1 &rq->__lock irq_context: 0 epnested_mutex &ep->mtx &ep->mtx/1 &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 epnested_mutex &ep->mtx &ep->poll_wait irq_context: 0 epnested_mutex &ep->mtx &ep->mtx/1 &ep->lock irq_context: 0 &ep->mtx rcu_read_lock &ep->poll_wait irq_context: 0 &sb->s_type->i_mutex_key#9 fs_reclaim mmu_notifier_invalidate_range_start &rq->__lock irq_context: 0 &sb->s_type->i_mutex_key#9 fs_reclaim mmu_notifier_invalidate_range_start &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: softirq rcu_read_lock rcu_read_lock slock-AF_INET/1 rcu_read_lock rcu_read_lock rcu_read_lock_bh dev->qdisc_tx_busylock ?: &qdisc_tx_busylock _xmit_ETHER#2 rcu_read_lock &____s->seqcount#2 irq_context: 0 sb_writers#4 &type->i_mutex_dir_key#3/1 jbd2_handle &ei->i_data_sem rcu_read_lock rcu_node_0 irq_context: 0 sb_writers#4 &type->i_mutex_dir_key#3/1 jbd2_handle &ei->i_data_sem rcu_read_lock &rq->__lock irq_context: 0 sb_writers#4 &type->i_mutex_dir_key#3/1 jbd2_handle &ei->i_data_sem rcu_read_lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 sb_writers#8 &of->mutex kn->active#5 uevent_sock_mutex fs_reclaim mmu_notifier_invalidate_range_start &rq->__lock irq_context: 0 sb_writers#8 &of->mutex kn->active#5 uevent_sock_mutex fs_reclaim mmu_notifier_invalidate_range_start &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 sb_writers#4 &type->i_mutex_dir_key#3/1 fs_reclaim &rq->__lock irq_context: 0 sb_writers#4 &type->i_mutex_dir_key#3/1 fs_reclaim &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)wg-crypt-wg1 (work_completion)(&peer->transmit_packet_work) &peer->endpoint_lock rcu_read_lock_bh batched_entropy_u32.lock crngs.lock irq_context: softirq (&icsk->icsk_retransmit_timer) slock-AF_INET batched_entropy_u32.lock irq_context: softirq rcu_read_lock rcu_read_lock slock-AF_INET/1 rcu_read_lock rcu_read_lock pool_lock#2 irq_context: softirq rcu_read_lock rcu_read_lock slock-AF_INET/1 rcu_read_lock rcu_read_lock &dir->lock#2 irq_context: softirq rcu_read_lock rcu_read_lock slock-AF_INET/1 rcu_read_lock rcu_read_lock &ul->lock irq_context: 0 &sb->s_type->i_mutex_key#10 sk_lock-AF_INET remove_cache_srcu irq_context: 0 &sb->s_type->i_mutex_key#10 sk_lock-AF_INET remove_cache_srcu quarantine_lock irq_context: 0 &sb->s_type->i_mutex_key#10 sk_lock-AF_INET remove_cache_srcu &c->lock irq_context: 0 &sb->s_type->i_mutex_key#10 sk_lock-AF_INET remove_cache_srcu &n->list_lock irq_context: 0 &sb->s_type->i_mutex_key#10 sk_lock-AF_INET remove_cache_srcu &obj_hash[i].lock irq_context: 0 &sb->s_type->i_mutex_key#10 sk_lock-AF_INET remove_cache_srcu pool_lock#2 irq_context: 0 &sb->s_type->i_mutex_key#10 sk_lock-AF_INET remove_cache_srcu rcu_read_lock pool_lock#2 irq_context: 0 &sb->s_type->i_mutex_key#10 sk_lock-AF_INET remove_cache_srcu &rq->__lock irq_context: 0 &sb->s_type->i_mutex_key#10 sk_lock-AF_INET remove_cache_srcu &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 &sb->s_type->i_mutex_key#10 sk_lock-AF_INET remove_cache_srcu &cfs_rq->removed.lock irq_context: 0 &sb->s_type->i_mutex_key#10 sk_lock-AF_INET remove_cache_srcu &rq->__lock &cfs_rq->removed.lock irq_context: 0 &sb->s_type->i_mutex_key#10 sk_lock-AF_INET slock-AF_INET &____s->seqcount irq_context: 0 &sb->s_type->i_mutex_key#10 sk_lock-AF_INET slock-AF_INET rcu_read_lock pool_lock#2 irq_context: 0 &f->f_pos_lock rcu_node_0 irq_context: 0 &net->xdp.lock irq_context: 0 &sb->s_type->i_mutex_key#10 &net->xdp.lock irq_context: 0 &sb->s_type->i_mutex_key#10 &xs->map_list_lock irq_context: 0 &sb->s_type->i_mutex_key#10 &xs->mutex irq_context: 0 &sb->s_type->i_mutex_key#10 clock-AF_XDP irq_context: 0 &p->lock batched_entropy_u8.lock irq_context: 0 &p->lock kfence_freelist_lock irq_context: 0 &sb->s_type->i_mutex_key#9 mmu_notifier_invalidate_range_start &rq->__lock irq_context: 0 &sb->s_type->i_mutex_key#9 mmu_notifier_invalidate_range_start &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 nf_sockopt_mutex nf_sockopt_mutex.wait_lock irq_context: 0 nf_sockopt_mutex.wait_lock irq_context: 0 sb_writers#8 kn->active#5 remove_cache_srcu &rq->__lock irq_context: 0 sb_writers#8 kn->active#5 remove_cache_srcu &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: softirq rcu_read_lock rcu_read_lock rcu_read_lock rlock-AF_INET irq_context: 0 (wq_completion)events (work_completion)(&rdev->wiphy_work) &rdev->wiphy.mtx &wdev->mtx &rdev->bss_lock batched_entropy_u8.lock irq_context: 0 (wq_completion)events (work_completion)(&rdev->wiphy_work) &rdev->wiphy.mtx &wdev->mtx &rdev->bss_lock kfence_freelist_lock irq_context: 0 (wq_completion)events (work_completion)(&rdev->wiphy_work) &rdev->wiphy.mtx &wdev->mtx &rdev->bss_lock &meta->lock irq_context: 0 sb_writers#8 &of->mutex kn->active#5 rcu_read_lock rcu_node_0 irq_context: 0 sb_writers#8 &of->mutex kn->active#5 rcu_read_lock &rq->__lock irq_context: 0 sb_writers#8 &of->mutex kn->active#5 rcu_read_lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 &p->lock rcu_read_lock &cfs_rq->removed.lock irq_context: 0 &p->lock rcu_read_lock &obj_hash[i].lock irq_context: 0 sb_writers#5 &type->i_mutex_dir_key#5/1 &sb->s_type->i_mutex_key#12 &sb->s_type->i_mutex_key#12/4 &simple_offset_xa_lock &n->list_lock irq_context: 0 sb_writers#5 &type->i_mutex_dir_key#5/1 &sb->s_type->i_mutex_key#12 &sb->s_type->i_mutex_key#12/4 &simple_offset_xa_lock &n->list_lock &c->lock irq_context: 0 rcu_read_lock &tbl->lock &n->lock irq_context: 0 rcu_read_lock &tbl->lock &c->lock irq_context: 0 rcu_read_lock &tbl->lock pool_lock#2 irq_context: 0 rcu_read_lock &tbl->lock &n->lock irq_context: 0 rcu_read_lock &tbl->lock &n->lock &____s->seqcount#9 irq_context: 0 rcu_read_lock &tbl->lock nl_table_lock irq_context: 0 rcu_read_lock &tbl->lock &obj_hash[i].lock irq_context: 0 rcu_read_lock &tbl->lock nl_table_wait.lock irq_context: 0 rcu_read_lock &tbl->lock rcu_read_lock lock#8 irq_context: 0 rcu_read_lock &tbl->lock rcu_read_lock id_table_lock irq_context: 0 rcu_read_lock &tbl->lock &dir->lock#2 irq_context: 0 rcu_read_lock &tbl->lock krc.lock irq_context: 0 rcu_read_lock &n->lock &c->lock irq_context: 0 rcu_read_lock &n->lock pool_lock#2 irq_context: 0 rcu_read_lock &ul->lock#2 irq_context: 0 rcu_read_lock rcu_read_lock rcu_read_lock &tbl->lock irq_context: 0 rcu_read_lock rcu_read_lock rcu_read_lock &n->lock irq_context: 0 rcu_read_lock rcu_read_lock rcu_read_lock &____s->seqcount#9 irq_context: 0 rcu_read_lock rcu_read_lock rcu_read_lock rcu_read_lock_bh &obj_hash[i].lock irq_context: 0 rcu_read_lock rcu_read_lock rcu_read_lock rcu_read_lock_bh pool_lock#2 irq_context: 0 rcu_read_lock rcu_read_lock rcu_read_lock rcu_read_lock_bh dev->qdisc_tx_busylock ?: &qdisc_tx_busylock rcu_read_lock &c->lock irq_context: 0 rcu_read_lock rcu_read_lock rcu_read_lock rcu_read_lock_bh dev->qdisc_tx_busylock ?: &qdisc_tx_busylock rcu_read_lock pool_lock#2 irq_context: 0 rcu_read_lock rcu_read_lock rcu_read_lock rcu_read_lock_bh dev->qdisc_tx_busylock ?: &qdisc_tx_busylock rcu_read_lock tk_core.seq.seqcount irq_context: 0 rcu_read_lock rcu_read_lock rcu_read_lock rcu_read_lock_bh dev->qdisc_tx_busylock ?: &qdisc_tx_busylock rcu_read_lock &obj_hash[i].lock irq_context: 0 rcu_read_lock rcu_read_lock rcu_read_lock rcu_read_lock_bh dev->qdisc_tx_busylock ?: &qdisc_tx_busylock rcu_read_lock &r->producer_lock irq_context: 0 sb_writers#4 &type->i_mutex_dir_key#3/1 &type->i_mutex_dir_key#3 jbd2_handle &cfs_rq->removed.lock irq_context: 0 sb_writers#4 &type->i_mutex_dir_key#3/1 &type->i_mutex_dir_key#3 jbd2_handle &obj_hash[i].lock irq_context: 0 sb_writers#4 &type->i_mutex_dir_key#3/1 sb_internal remove_cache_srcu pool_lock#2 irq_context: 0 &fsnotify_mark_srcu &rq->__lock &cfs_rq->removed.lock irq_context: 0 proto_tab_lock &c->lock irq_context: 0 proto_tab_lock &n->list_lock irq_context: 0 proto_tab_lock &n->list_lock &c->lock irq_context: 0 proto_tab_lock &obj_hash[i].lock pool_lock irq_context: 0 rcu_read_lock &n->lock &____s->seqcount#2 irq_context: 0 rcu_read_lock &n->lock &____s->seqcount irq_context: 0 sb_writers#4 &type->i_mutex_dir_key#3/1 tomoyo_ss rcu_read_lock rcu_read_lock &rq->__lock &cfs_rq->removed.lock irq_context: 0 sb_writers#4 &type->i_mutex_dir_key#3/1 &sb->s_type->i_mutex_key#8 jbd2_handle &journal->j_state_lock &journal->j_wait_commit irq_context: 0 sb_writers#4 &type->i_mutex_dir_key#3/1 &sb->s_type->i_mutex_key#8 jbd2_handle &journal->j_state_lock &journal->j_wait_commit &p->pi_lock irq_context: 0 sb_writers#4 &type->i_mutex_dir_key#3/1 &sb->s_type->i_mutex_key#8 jbd2_handle &journal->j_state_lock &journal->j_wait_commit &p->pi_lock &rq->__lock irq_context: 0 sb_writers#4 &type->i_mutex_dir_key#3/1 &sb->s_type->i_mutex_key#8 jbd2_handle &journal->j_state_lock &journal->j_wait_commit &p->pi_lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 &bgl->locks[i].lock irq_context: 0 dup_mmap_sem &mm->mmap_lock &mm->mmap_lock/1 stock_lock rcu_read_lock per_cpu_ptr(&cgroup_rstat_cpu_lock, cpu) irq_context: 0 (wq_completion)bat_events (work_completion)(&(&forw_packet_aggr->delayed_work)->work) &hard_iface->bat_iv.ogm_buff_mutex &bat_priv->tt.commit_lock &n->list_lock irq_context: 0 (wq_completion)bat_events (work_completion)(&(&forw_packet_aggr->delayed_work)->work) &hard_iface->bat_iv.ogm_buff_mutex &bat_priv->tt.commit_lock &n->list_lock &c->lock irq_context: 0 tomoyo_ss mmu_notifier_invalidate_range_start &cfs_rq->removed.lock irq_context: 0 tomoyo_ss mmu_notifier_invalidate_range_start &obj_hash[i].lock irq_context: 0 &sb->s_type->i_mutex_key#9 remove_cache_srcu pool_lock#2 irq_context: 0 sk_lock-AF_INET6 &zone->lock irq_context: 0 sk_lock-AF_INET6 &zone->lock &____s->seqcount irq_context: 0 &sb->s_type->i_mutex_key#10 sk_lock-AF_INET6/1 &zone->lock irq_context: 0 &sb->s_type->i_mutex_key#10 sk_lock-AF_INET6/1 &zone->lock &____s->seqcount irq_context: 0 sk_lock-AF_ALG &drbg->drbg_mutex irq_context: 0 sk_lock-AF_ALG &drbg->drbg_mutex crypto_alg_sem irq_context: 0 sk_lock-AF_ALG &drbg->drbg_mutex crypto_alg_sem &rq->__lock irq_context: 0 sk_lock-AF_ALG &drbg->drbg_mutex crypto_alg_sem &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 rcu_read_lock rcu_read_lock_bh &zone->lock irq_context: 0 rcu_read_lock rcu_read_lock_bh &zone->lock &____s->seqcount irq_context: 0 sk_lock-AF_ALG &drbg->drbg_mutex fs_reclaim irq_context: 0 sk_lock-AF_ALG &drbg->drbg_mutex fs_reclaim mmu_notifier_invalidate_range_start irq_context: 0 sk_lock-AF_ALG &drbg->drbg_mutex pool_lock#2 irq_context: 0 sk_lock-AF_ALG &drbg->drbg_mutex &c->lock irq_context: 0 sk_lock-AF_ALG &drbg->drbg_mutex &rq->__lock irq_context: 0 sk_lock-AF_ALG &drbg->drbg_mutex &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 rcu_read_lock &zone->lock &____s->seqcount irq_context: 0 sk_lock-AF_ALG &drbg->drbg_mutex crngs.lock irq_context: 0 sk_lock-AF_ALG &drbg->drbg_mutex &rng->jent_lock irq_context: softirq (&n->timer) rcu_read_lock rcu_read_lock rcu_read_lock_bh dev->qdisc_tx_busylock ?: &qdisc_tx_busylock rcu_read_lock &r->producer_lock irq_context: softirq &(&bat_priv->tt.work)->timer rcu_read_lock &pool->lock/1 &p->pi_lock &cfs_rq->removed.lock irq_context: 0 &sb->s_type->i_mutex_key#10 &rng->jent_lock irq_context: 0 &sb->s_type->i_mutex_key#10 &rng->jent_lock &obj_hash[i].lock irq_context: 0 &sb->s_type->i_mutex_key#10 &rng->jent_lock pool_lock#2 irq_context: 0 &x->wait#21 &p->pi_lock &rq->__lock &cfs_rq->removed.lock irq_context: 0 sb_writers#8 kn->active#5 fs_reclaim mmu_notifier_invalidate_range_start &rq->__lock irq_context: 0 sb_writers#8 kn->active#5 fs_reclaim mmu_notifier_invalidate_range_start &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 cb_lock &rdev->wiphy.mtx rcu_read_lock rcu_node_0 irq_context: 0 cb_lock &rdev->wiphy.mtx rcu_read_lock &rq->__lock irq_context: 0 cb_lock &rdev->wiphy.mtx rcu_read_lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 cb_lock &rdev->wiphy.mtx &local->mtx rcu_read_lock hwsim_radio_lock &____s->seqcount#2 irq_context: 0 cb_lock &rdev->wiphy.mtx &local->mtx rcu_read_lock hwsim_radio_lock &____s->seqcount irq_context: 0 &ei->i_data_sem &rq->__lock irq_context: 0 &ei->i_data_sem &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 &f->f_pos_lock &type->i_mutex_dir_key#3 sb_writers#4 batched_entropy_u8.lock irq_context: 0 &f->f_pos_lock &type->i_mutex_dir_key#3 sb_writers#4 kfence_freelist_lock irq_context: 0 &f->f_pos_lock &type->i_mutex_dir_key#3 sb_writers#4 &meta->lock irq_context: softirq (&timer) rcu_read_lock &n->list_lock irq_context: softirq (&timer) rcu_read_lock &n->list_lock &c->lock irq_context: softirq (&timer) rcu_read_lock &____s->seqcount#2 irq_context: 0 rcu_read_lock_bh &nr_netdev_xmit_lock_key irq_context: 0 rcu_read_lock_bh &nr_netdev_xmit_lock_key nr_node_list_lock irq_context: 0 rcu_read_lock_bh &nr_netdev_xmit_lock_key &obj_hash[i].lock irq_context: 0 rcu_read_lock_bh &nr_netdev_xmit_lock_key pool_lock#2 irq_context: 0 sk_lock-AF_INET rcu_read_lock &rcu_state.gp_wq irq_context: 0 sk_lock-AF_INET rcu_read_lock &rcu_state.gp_wq &p->pi_lock irq_context: 0 sk_lock-AF_INET rcu_read_lock &rcu_state.gp_wq &p->pi_lock &rq->__lock irq_context: 0 sk_lock-AF_INET rcu_read_lock &rcu_state.gp_wq &p->pi_lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 sb_writers#4 &type->i_mutex_dir_key#3 tomoyo_ss remove_cache_srcu irq_context: 0 sb_writers#4 &type->i_mutex_dir_key#3 tomoyo_ss remove_cache_srcu quarantine_lock irq_context: 0 sb_writers#8 &sb->s_type->i_mutex_key#13 &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)events (work_completion)(&rdev->wiphy_work) &rdev->wiphy.mtx &wdev->mtx rcu_read_lock rcu_read_lock &sta->rate_ctrl_lock &c->lock irq_context: softirq (&n->timer) &n->lock &c->lock irq_context: 0 (crypto_chain).rwsem &n->list_lock irq_context: 0 (crypto_chain).rwsem &n->list_lock &c->lock irq_context: 0 (wq_completion)events (work_completion)(&p->wq) &pcp->lock &zone->lock irq_context: 0 (wq_completion)events (work_completion)(&p->wq) &pcp->lock &zone->lock &____s->seqcount irq_context: softirq rcu_read_lock hwsim_radio_lock &zone->lock irq_context: softirq rcu_read_lock hwsim_radio_lock &pcp->lock &zone->lock &____s->seqcount irq_context: softirq rcu_read_lock rcu_read_lock &zone->lock irq_context: 0 &mm->mmap_lock pcpu_lock stock_lock irq_context: 0 &audit_cmd_mutex.lock irq_context: 0 &audit_cmd_mutex.lock fs_reclaim irq_context: 0 &audit_cmd_mutex.lock fs_reclaim mmu_notifier_invalidate_range_start irq_context: 0 &audit_cmd_mutex.lock pool_lock#2 irq_context: 0 &audit_cmd_mutex.lock rlock-AF_NETLINK irq_context: 0 (wq_completion)events (work_completion)(&(&nsim_dev->trap_data->trap_report_dw)->work) &devlink->lock_key#5 &cfs_rq->removed.lock irq_context: 0 (wq_completion)events (work_completion)(&(&nsim_dev->trap_data->trap_report_dw)->work) &devlink->lock_key#5 &obj_hash[i].lock irq_context: 0 (wq_completion)events (work_completion)(&(&nsim_dev->trap_data->trap_report_dw)->work) &devlink->lock_key#5 &rq->__lock &cfs_rq->removed.lock irq_context: 0 rtnl_mutex sk_lock-AF_INET irq_context: 0 rtnl_mutex sk_lock-AF_INET slock-AF_INET irq_context: 0 rtnl_mutex slock-AF_INET irq_context: 0 sb_writers#4 rcu_read_lock &cfs_rq->removed.lock irq_context: 0 sb_writers#4 rcu_read_lock &obj_hash[i].lock irq_context: 0 sb_writers#4 &type->i_mutex_dir_key#3/1 sb_internal jbd2_handle &bgl->locks[i].lock irq_context: 0 &fsnotify_mark_srcu &group->notification_waitq &ep->lock &ep->wq &p->pi_lock &rq->__lock &cfs_rq->removed.lock irq_context: 0 sb_writers#4 &type->i_mutex_dir_key#3/1 sb_internal jbd2_handle &ei->i_data_sem batched_entropy_u8.lock irq_context: 0 sb_writers#4 &type->i_mutex_dir_key#3/1 sb_internal jbd2_handle &ei->i_data_sem kfence_freelist_lock irq_context: 0 sb_writers#8 tomoyo_ss mmu_notifier_invalidate_range_start &rq->__lock irq_context: 0 sb_writers#8 tomoyo_ss mmu_notifier_invalidate_range_start &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 &f->f_pos_lock &type->i_mutex_dir_key#3 sb_writers#4 remove_cache_srcu pool_lock#2 irq_context: 0 sb_writers#4 &type->i_mutex_dir_key#3/1 sb_internal jbd2_handle &ei->i_data_sem &bgl->locks[i].lock &meta->lock irq_context: 0 sb_writers#4 &type->i_mutex_dir_key#3/1 sb_internal jbd2_handle &ei->i_data_sem &bgl->locks[i].lock kfence_freelist_lock irq_context: 0 rcu_read_lock rcu_read_lock_bh &list->lock#5 irq_context: softirq rcu_read_lock rcu_read_lock k-slock-AF_INET rcu_read_lock &c->lock irq_context: softirq rcu_read_lock rcu_read_lock k-slock-AF_INET rcu_read_lock pool_lock#2 irq_context: softirq rcu_read_lock rcu_read_lock k-slock-AF_INET rcu_read_lock &dir->lock#2 irq_context: softirq rcu_read_lock rcu_read_lock k-slock-AF_INET rcu_read_lock &ul->lock irq_context: softirq rcu_read_lock rcu_read_lock k-slock-AF_INET &n->list_lock irq_context: softirq rcu_read_lock rcu_read_lock k-slock-AF_INET &n->list_lock &c->lock irq_context: 0 (wq_completion)bat_events (work_completion)(&(&forw_packet_aggr->delayed_work)->work) &hard_iface->bat_iv.ogm_buff_mutex &cfs_rq->removed.lock irq_context: 0 (wq_completion)bat_events (work_completion)(&(&forw_packet_aggr->delayed_work)->work) &hard_iface->bat_iv.ogm_buff_mutex rcu_read_lock &rq->__lock &cfs_rq->removed.lock irq_context: 0 rtnl_mutex sk_lock-AF_INET &mm->mmap_lock irq_context: 0 sb_writers#4 &type->i_mutex_dir_key#3 jbd2_handle &ei->i_es_lock key#2 irq_context: 0 &f->f_pos_lock sb_writers#4 &sb->s_type->i_mutex_key#8 &ei->i_data_sem &n->list_lock irq_context: 0 &f->f_pos_lock sb_writers#4 &sb->s_type->i_mutex_key#8 &ei->i_data_sem &n->list_lock &c->lock irq_context: 0 sk_lock-AF_INET &net->xfrm.xfrm_policy_lock irq_context: 0 &sb->s_type->i_mutex_key#10 &net->xfrm.xfrm_policy_lock irq_context: 0 &sb->s_type->i_mutex_key#10 &policy->lock irq_context: 0 &sb->s_type->i_mutex_key#10 &list->lock#32 irq_context: 0 sb_writers#4 &type->i_mutex_dir_key#3/1 &meta->lock irq_context: softirq rcu_read_lock rcu_read_lock k-slock-AF_INET rcu_read_lock &obj_hash[i].lock irq_context: 0 &xt[i].mutex rcu_read_lock rcu_read_lock &cfs_rq->removed.lock irq_context: 0 &xt[i].mutex rcu_read_lock rcu_read_lock &obj_hash[i].lock irq_context: 0 &xt[i].mutex rcu_read_lock &rq->__lock &cfs_rq->removed.lock irq_context: 0 &xt[i].mutex rcu_read_lock &cfs_rq->removed.lock irq_context: 0 &xt[i].mutex rcu_read_lock &obj_hash[i].lock irq_context: 0 &mm->mmap_lock pgd_lock irq_context: 0 &mm->mmap_lock key irq_context: 0 &mm->mmap_lock pcpu_lock irq_context: 0 &mm->mmap_lock percpu_counters_lock irq_context: 0 &mm->mmap_lock pcpu_lock stock_lock irq_context: 0 sb_writers#4 &type->i_mutex_dir_key#3/1 &type->i_mutex_dir_key#3 jbd2_handle &sbi->s_orphan_lock &rq->__lock irq_context: 0 sb_writers#4 &type->i_mutex_dir_key#3/1 &type->i_mutex_dir_key#3 jbd2_handle &sbi->s_orphan_lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 &f->f_pos_lock sb_writers#4 &sb->s_type->i_mutex_key#8 &ei->i_data_sem &ei->i_es_lock &n->list_lock irq_context: 0 &f->f_pos_lock sb_writers#4 &sb->s_type->i_mutex_key#8 &ei->i_data_sem &ei->i_es_lock &n->list_lock &c->lock irq_context: 0 sb_writers#4 &type->i_mutex_dir_key#3/1 &type->i_mutex_dir_key#3 jbd2_handle &p->pi_lock &rq->__lock irq_context: 0 sb_writers#4 &type->i_mutex_dir_key#3/1 &type->i_mutex_dir_key#3 jbd2_handle &p->pi_lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 &f->f_pos_lock sb_writers#4 &sb->s_type->i_mutex_key#8 &mm->mmap_lock &c->lock irq_context: 0 (wq_completion)bat_events (work_completion)(&(&bat_priv->nc.work)->work) rcu_read_lock pool_lock#2 irq_context: 0 &sbi->s_writepages_rwsem &n->list_lock irq_context: 0 &sbi->s_writepages_rwsem &n->list_lock &c->lock irq_context: 0 (wq_completion)rcu_gp (work_completion)(&(&ssp->srcu_sup->work)->work) &rq->__lock &cfs_rq->removed.lock irq_context: 0 &audit_cmd_mutex.lock &c->lock irq_context: 0 (wq_completion)events (work_completion)(&w->w) &rq->__lock irq_context: 0 sb_writers#4 &type->i_mutex_dir_key#3/1 &sb->s_type->i_mutex_key#8 jbd2_handle rcu_read_lock &rcu_state.gp_wq irq_context: 0 sb_writers#4 &type->i_mutex_dir_key#3/1 &sb->s_type->i_mutex_key#8 jbd2_handle rcu_read_lock &rcu_state.gp_wq &p->pi_lock irq_context: 0 sb_writers#4 &type->i_mutex_dir_key#3/1 &sb->s_type->i_mutex_key#8 jbd2_handle rcu_read_lock &rcu_state.gp_wq &p->pi_lock &rq->__lock irq_context: 0 sb_writers#4 &type->i_mutex_dir_key#3/1 &sb->s_type->i_mutex_key#8 jbd2_handle rcu_read_lock &rcu_state.gp_wq &p->pi_lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 &f->f_pos_lock sb_writers#4 &sb->s_type->i_mutex_key#8 &folio_wait_table[i] irq_context: 0 &f->f_pos_lock sb_writers#4 &sb->s_type->i_mutex_key#8 &folio_wait_table[i] &p->pi_lock irq_context: 0 &f->f_pos_lock sb_writers#4 &sb->s_type->i_mutex_key#8 &folio_wait_table[i] &p->pi_lock &rq->__lock irq_context: 0 &f->f_pos_lock sb_writers#4 &sb->s_type->i_mutex_key#8 &folio_wait_table[i] &p->pi_lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 elock-AF_INET irq_context: 0 (wq_completion)events_unbound connector_reaper_work &obj_hash[i].lock pool_lock irq_context: 0 &iint->mutex fs_reclaim mmu_notifier_invalidate_range_start &rq->__lock irq_context: 0 &iint->mutex fs_reclaim mmu_notifier_invalidate_range_start &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 &iint->mutex rcu_read_lock rcu_read_lock rcu_node_0 irq_context: 0 &iint->mutex rcu_read_lock rcu_read_lock &rq->__lock irq_context: 0 &iint->mutex rcu_read_lock rcu_read_lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 &sb->s_type->i_mutex_key#10 rcu_read_lock &rq->__lock &cfs_rq->removed.lock irq_context: 0 (wq_completion)events_unbound connector_reaper_work &cfs_rq->removed.lock irq_context: 0 &f->f_owner.lock irq_context: 0 sb_writers#4 &type->i_mutex_dir_key#3/1 sb_internal jbd2_handle &ei->i_data_sem rcu_read_lock &rcu_state.expedited_wq irq_context: 0 sb_writers#4 &type->i_mutex_dir_key#3/1 sb_internal jbd2_handle &ei->i_data_sem rcu_read_lock &rcu_state.expedited_wq &p->pi_lock irq_context: 0 sb_writers#4 &type->i_mutex_dir_key#3/1 sb_internal jbd2_handle &ei->i_data_sem rcu_read_lock &rcu_state.expedited_wq &p->pi_lock &rq->__lock irq_context: 0 sb_writers#4 &type->i_mutex_dir_key#3/1 sb_internal jbd2_handle &ei->i_data_sem rcu_read_lock &rcu_state.expedited_wq &p->pi_lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 &f->f_pos_lock sb_writers#4 &sb->s_type->i_mutex_key#8 lock#4 &obj_hash[i].lock irq_context: 0 sb_writers#3 rcu_read_lock rcu_node_0 irq_context: 0 sb_writers#3 rcu_read_lock &rq->__lock irq_context: 0 sb_writers#3 rcu_read_lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 mapping.invalidate_lock irq_context: 0 &xt[i].mutex pgd_lock irq_context: 0 &xt[i].mutex stock_lock irq_context: 0 &xt[i].mutex key irq_context: 0 &xt[i].mutex pcpu_lock irq_context: 0 &xt[i].mutex percpu_counters_lock irq_context: 0 &xt[i].mutex pcpu_lock stock_lock irq_context: 0 sb_writers#4 &type->i_mutex_dir_key#3/1 sb_internal quarantine_lock irq_context: 0 key#22 irq_context: 0 dup_mmap_sem &mm->mmap_lock &mm->mmap_lock/1 remove_cache_srcu rcu_read_lock rcu_read_lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 &f->f_pos_lock &type->i_mutex_dir_key#3 remove_cache_srcu pool_lock#2 irq_context: 0 &fsnotify_mark_srcu remove_cache_srcu pool_lock#2 irq_context: 0 sb_writers#8 &of->mutex kn->active#5 &cfs_rq->removed.lock irq_context: 0 (wq_completion)events_unbound connector_reaper_work &ACCESS_PRIVATE(ssp->srcu_sup, lock) irq_context: 0 &f->f_pos_lock sb_writers#4 &sb->s_type->i_mutex_key#8 &mm->mmap_lock remove_cache_srcu irq_context: 0 &f->f_pos_lock sb_writers#4 &sb->s_type->i_mutex_key#8 &mm->mmap_lock remove_cache_srcu quarantine_lock irq_context: 0 &f->f_pos_lock sb_writers#4 &sb->s_type->i_mutex_key#8 &mm->mmap_lock remove_cache_srcu &c->lock irq_context: 0 &f->f_pos_lock sb_writers#4 &sb->s_type->i_mutex_key#8 &mm->mmap_lock remove_cache_srcu &n->list_lock irq_context: 0 &f->f_pos_lock sb_writers#4 &sb->s_type->i_mutex_key#8 &mm->mmap_lock remove_cache_srcu &obj_hash[i].lock irq_context: 0 (wq_completion)events_power_efficient (work_completion)(&(&gc_work->dwork)->work) rcu_read_lock &nf_conntrack_locks[i] irq_context: 0 (wq_completion)events_power_efficient (work_completion)(&(&gc_work->dwork)->work) rcu_read_lock &nf_conntrack_locks[i] &nf_conntrack_locks[i]/1 irq_context: 0 (wq_completion)events_power_efficient (work_completion)(&(&gc_work->dwork)->work) rcu_read_lock &nf_conntrack_locks[i]/1 irq_context: 0 (wq_completion)events_power_efficient (work_completion)(&(&gc_work->dwork)->work) rcu_read_lock rcu_read_lock &nf_nat_locks[i] irq_context: 0 (wq_completion)events_power_efficient (work_completion)(&(&gc_work->dwork)->work) rcu_read_lock &c->lock irq_context: 0 &mm->mmap_lock &mapping->i_mmap_rwsem fill_pool_map-wait-type-override &cfs_rq->removed.lock irq_context: 0 &mm->mmap_lock &mapping->i_mmap_rwsem fill_pool_map-wait-type-override &obj_hash[i].lock irq_context: 0 (wq_completion)events_power_efficient (work_completion)(&(&gc_work->dwork)->work) rcu_read_lock &rcu_state.expedited_wq irq_context: 0 (wq_completion)events_power_efficient (work_completion)(&(&gc_work->dwork)->work) rcu_read_lock &rcu_state.expedited_wq &p->pi_lock irq_context: 0 (wq_completion)events_power_efficient (work_completion)(&(&gc_work->dwork)->work) rcu_read_lock &rcu_state.expedited_wq &p->pi_lock &rq->__lock irq_context: 0 (wq_completion)events_power_efficient (work_completion)(&(&gc_work->dwork)->work) rcu_read_lock &rcu_state.expedited_wq &p->pi_lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)events_power_efficient (gc_work).work &cfs_rq->removed.lock irq_context: 0 (wq_completion)wg-crypt-wg0#2 (work_completion)(&peer->transmit_packet_work) &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)events (work_completion)(&(&nsim_dev->trap_data->trap_report_dw)->work) &devlink->lock_key#2 &cfs_rq->removed.lock irq_context: 0 (wq_completion)events (work_completion)(&(&nsim_dev->trap_data->trap_report_dw)->work) &devlink->lock_key#2 &obj_hash[i].lock irq_context: 0 (wq_completion)wg-crypt-wg1#2 (work_completion)(&({ do { const void *__vpp_verify = (typeof((worker) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __asm__ ("" : "=r"(__ptr) : "0"((typeof(*((worker))) *)((worker)))); (typeof((typeof(*((worker))) *)((worker)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); })->work) &rq->__lock irq_context: 0 (wq_completion)wg-crypt-wg1#2 (work_completion)(&({ do { const void *__vpp_verify = (typeof((worker) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __asm__ ("" : "=r"(__ptr) : "0"((typeof(*((worker))) *)((worker)))); (typeof((typeof(*((worker))) *)((worker)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); })->work) &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 &xt[i].mutex quarantine_lock irq_context: 0 sb_writers#5 &type->i_mutex_dir_key#5/1 &fsnotify_mark_srcu &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)events (work_completion)(&(&hwstats->traffic_dw)->work) &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)bat_events (work_completion)(&(&bat_priv->bla.work)->work) crngs.lock irq_context: 0 &sb->s_type->i_mutex_key#8 batched_entropy_u8.lock irq_context: 0 &sb->s_type->i_mutex_key#8 kfence_freelist_lock irq_context: 0 &sb->s_type->i_mutex_key#10 &chan->lock/1 &obj_hash[i].lock pool_lock irq_context: 0 &sb->s_type->i_mutex_key#10 &chan->lock/1 &rq->__lock irq_context: 0 &audit_cmd_mutex.lock tk_core.seq.seqcount irq_context: 0 &audit_cmd_mutex.lock &obj_hash[i].lock irq_context: 0 &audit_cmd_mutex.lock &list->lock irq_context: 0 &audit_cmd_mutex.lock kauditd_wait.lock irq_context: 0 &audit_cmd_mutex.lock kauditd_wait.lock &p->pi_lock irq_context: 0 &audit_cmd_mutex.lock kauditd_wait.lock &p->pi_lock &rq->__lock irq_context: 0 &audit_cmd_mutex.lock kauditd_wait.lock &p->pi_lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 &audit_cmd_mutex.lock &rq->__lock irq_context: 0 &audit_cmd_mutex.lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)events (work_completion)(&rdev->wiphy_work) &rdev->wiphy.mtx &pool->lock &p->pi_lock irq_context: 0 (wq_completion)events (work_completion)(&rdev->wiphy_work) &rdev->wiphy.mtx &pool->lock &p->pi_lock &rq->__lock irq_context: 0 (wq_completion)events (work_completion)(&rdev->wiphy_work) &rdev->wiphy.mtx &pool->lock &p->pi_lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)events (work_completion)(&rdev->wiphy_work) &rdev->wiphy.mtx &wdev->mtx rcu_read_lock rcu_read_lock &sta->rate_ctrl_lock &n->list_lock irq_context: 0 (wq_completion)events (work_completion)(&rdev->wiphy_work) &rdev->wiphy.mtx &wdev->mtx rcu_read_lock rcu_read_lock &sta->rate_ctrl_lock &n->list_lock &c->lock irq_context: 0 &sb->s_type->i_mutex_key#10 rcu_read_lock &cfs_rq->removed.lock irq_context: 0 (wq_completion)bat_events (work_completion)(&(&bat_priv->orig_work)->work) pool_lock#2 irq_context: 0 sb_writers#10 &type->i_mutex_dir_key#6/1 &xa->xa_lock#4 irq_context: 0 sb_writers#10 &type->i_mutex_dir_key#6/1 &obj_hash[i].lock irq_context: 0 sb_writers#10 &type->i_mutex_dir_key#6/1 stock_lock irq_context: 0 &fsnotify_mark_srcu &n->list_lock irq_context: 0 &fsnotify_mark_srcu &n->list_lock &c->lock irq_context: 0 sb_writers#3 &sb->s_type->i_mutex_key#9 rcu_read_lock rcu_read_lock rcu_node_0 irq_context: 0 sb_writers#3 &sb->s_type->i_mutex_key#9 rcu_read_lock rcu_read_lock &rq->__lock irq_context: 0 sb_writers#3 &sb->s_type->i_mutex_key#9 rcu_read_lock rcu_read_lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 &nft_net->commit_mutex rcu_read_lock &rq->__lock irq_context: 0 &nft_net->commit_mutex rcu_read_lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: softirq (&ndev->rs_timer) rcu_read_lock rcu_read_lock rcu_read_lock_bh _xmit_TUNNEL6#2 &meta->lock irq_context: softirq (&ndev->rs_timer) rcu_read_lock rcu_read_lock rcu_read_lock_bh _xmit_TUNNEL6#2 kfence_freelist_lock irq_context: 0 sk_lock-AF_INET6 &mm->mmap_lock rcu_node_0 irq_context: 0 dup_mmap_sem &mm->mmap_lock &mm->mmap_lock/1 mmu_notifier_invalidate_range_start &rq->__lock irq_context: 0 dup_mmap_sem &mm->mmap_lock &mm->mmap_lock/1 mmu_notifier_invalidate_range_start &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 sk_lock-AF_INET6 rcu_read_lock rcu_read_lock_bh dev->qdisc_tx_busylock ?: &qdisc_tx_busylock rcu_read_lock &r->producer_lock irq_context: 0 sb_writers#5 &type->i_mutex_dir_key#5/1 tomoyo_ss rcu_read_lock rcu_read_lock &cfs_rq->removed.lock irq_context: 0 sb_writers#5 &type->i_mutex_dir_key#5/1 tomoyo_ss rcu_read_lock rcu_read_lock &obj_hash[i].lock irq_context: 0 &f->f_pos_lock &type->i_mutex_dir_key#3 remove_cache_srcu &pcp->lock &zone->lock irq_context: 0 &f->f_pos_lock &type->i_mutex_dir_key#3 remove_cache_srcu &pcp->lock &zone->lock &____s->seqcount irq_context: 0 &f->f_pos_lock &type->i_mutex_dir_key#3 remove_cache_srcu &rq->__lock irq_context: 0 &f->f_pos_lock &type->i_mutex_dir_key#3 remove_cache_srcu &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 sk_lock-AF_INET6 rcu_read_lock &nf_conntrack_locks[i] &nf_conntrack_locks[i]/1 batched_entropy_u8.lock crngs.lock irq_context: 0 sk_lock-AF_INET6 rcu_read_lock &n->lock &obj_hash[i].lock irq_context: 0 sk_lock-AF_INET6 rcu_read_lock &n->lock &base->lock irq_context: 0 sk_lock-AF_INET6 rcu_read_lock &n->lock &base->lock &obj_hash[i].lock irq_context: 0 sk_lock-AF_INET6 rcu_read_lock &n->lock pool_lock#2 irq_context: 0 sk_lock-AF_INET6 rcu_read_lock stock_lock irq_context: 0 sk_lock-AF_INET6 rcu_read_lock &ul->lock#2 irq_context: 0 sk_lock-AF_INET6 rcu_read_lock rcu_read_lock rcu_read_lock &tbl->lock irq_context: 0 sk_lock-AF_INET6 rcu_read_lock rcu_read_lock rcu_read_lock &n->lock irq_context: 0 sk_lock-AF_INET6 rcu_read_lock rcu_read_lock rcu_read_lock &____s->seqcount#9 irq_context: 0 sk_lock-AF_INET6 rcu_read_lock rcu_read_lock rcu_read_lock rcu_read_lock_bh &obj_hash[i].lock irq_context: 0 sk_lock-AF_INET6 rcu_read_lock rcu_read_lock rcu_read_lock rcu_read_lock_bh pool_lock#2 irq_context: 0 sk_lock-AF_INET6 rcu_read_lock rcu_read_lock rcu_read_lock rcu_read_lock_bh dev->qdisc_tx_busylock ?: &qdisc_tx_busylock rcu_read_lock &r->producer_lock irq_context: 0 sb_writers#5 &type->i_mutex_dir_key#5/1 fs_reclaim mmu_notifier_invalidate_range_start &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 &f->f_pos_lock sb_writers#4 &sb->s_type->i_mutex_key#8 jbd2_handle rcu_read_lock &rcu_state.gp_wq irq_context: 0 &f->f_pos_lock sb_writers#4 &sb->s_type->i_mutex_key#8 jbd2_handle rcu_read_lock &rcu_state.gp_wq &p->pi_lock irq_context: 0 &f->f_pos_lock sb_writers#4 &sb->s_type->i_mutex_key#8 jbd2_handle rcu_read_lock &rcu_state.gp_wq &p->pi_lock &rq->__lock irq_context: 0 &f->f_pos_lock sb_writers#4 &sb->s_type->i_mutex_key#8 jbd2_handle rcu_read_lock &rcu_state.gp_wq &p->pi_lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 sk_lock-AF_INET &msk->pm.lock irq_context: 0 sk_lock-AF_INET elock-AF_INET irq_context: 0 (work_completion)(&msk->work) irq_context: 0 cb_lock nlk_cb_mutex-GENERIC genl_mutex rtnl_mutex irq_context: 0 &sb->s_type->i_mutex_key#10 rcu_state.exp_mutex &rq->__lock &cfs_rq->removed.lock irq_context: 0 defrag4_mutex irq_context: 0 sb_writers#8 tomoyo_ss batched_entropy_u8.lock irq_context: 0 sb_writers#8 tomoyo_ss batched_entropy_u8.lock crngs.lock irq_context: 0 sb_writers#8 tomoyo_ss kfence_freelist_lock irq_context: 0 sb_writers#8 tomoyo_ss &meta->lock irq_context: 0 sock_diag_mutex irq_context: 0 sock_diag_mutex fs_reclaim irq_context: 0 sock_diag_mutex fs_reclaim mmu_notifier_invalidate_range_start irq_context: 0 sock_diag_mutex pool_lock#2 irq_context: 0 sock_diag_mutex &c->lock irq_context: 0 sock_diag_mutex rcu_read_lock pool_lock#2 irq_context: 0 sock_diag_mutex rcu_read_lock rcu_read_lock_bh &obj_hash[i].lock irq_context: 0 sock_diag_mutex rcu_read_lock rcu_read_lock_bh pool_lock#2 irq_context: 0 sock_diag_mutex rcu_read_lock rcu_node_0 irq_context: 0 sock_diag_mutex rcu_read_lock &rq->__lock irq_context: 0 sock_diag_mutex rcu_read_lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 sock_diag_mutex rlock-AF_NETLINK irq_context: 0 &sbi->s_writepages_rwsem jbd2_handle rcu_read_lock &cfs_rq->removed.lock irq_context: 0 tomoyo_ss remove_cache_srcu rcu_read_lock rcu_read_lock rcu_node_0 irq_context: 0 tomoyo_ss remove_cache_srcu rcu_read_lock rcu_read_lock &rq->__lock irq_context: 0 tomoyo_ss remove_cache_srcu rcu_read_lock rcu_read_lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 &ep->poll_wait irq_context: 0 dup_mmap_sem &mm->mmap_lock &mm->mmap_lock/1 &mm->context.lock &rq->__lock irq_context: 0 dup_mmap_sem &mm->mmap_lock &mm->mmap_lock/1 &mm->context.lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 cb_lock &rdev->wiphy.mtx &local->mtx &local->ack_status_lock &n->list_lock irq_context: 0 cb_lock &rdev->wiphy.mtx &local->mtx &local->ack_status_lock &n->list_lock &c->lock irq_context: softirq rcu_read_lock &local->ack_status_lock &obj_hash[i].lock pool_lock irq_context: 0 sb_writers#5 &type->i_mutex_dir_key#5/1 &simple_offset_xa_lock &c->lock irq_context: 0 sb_writers#5 &type->i_mutex_dir_key#5/1 &simple_offset_xa_lock &n->list_lock irq_context: 0 sb_writers#5 &type->i_mutex_dir_key#5/1 &simple_offset_xa_lock &n->list_lock &c->lock irq_context: 0 &f->f_pos_lock &type->i_mutex_dir_key#3 sb_writers#4 &n->list_lock irq_context: 0 &f->f_pos_lock &type->i_mutex_dir_key#3 sb_writers#4 &n->list_lock &c->lock irq_context: 0 &f->f_pos_lock sb_writers#4 &sb->s_type->i_mutex_key#8 rcu_read_lock &rq->__lock &cfs_rq->removed.lock irq_context: 0 &f->f_pos_lock &type->i_mutex_dir_key#3 sb_writers#4 jbd2_handle rcu_read_lock rcu_node_0 irq_context: 0 &f->f_pos_lock &type->i_mutex_dir_key#3 sb_writers#4 jbd2_handle rcu_read_lock &rq->__lock irq_context: 0 &f->f_pos_lock &type->i_mutex_dir_key#3 sb_writers#4 jbd2_handle rcu_read_lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 sb_writers#4 tomoyo_ss rcu_read_lock rcu_node_0 irq_context: 0 sb_writers#4 tomoyo_ss rcu_read_lock &rq->__lock irq_context: 0 sb_writers#4 tomoyo_ss rcu_read_lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 sb_writers#4 &sb->s_type->i_mutex_key#8 mapping.invalidate_lock &____s->seqcount#2 irq_context: 0 sb_writers#4 &sb->s_type->i_mutex_key#8 mapping.invalidate_lock &____s->seqcount irq_context: 0 cb_lock &rdev->wiphy.mtx &local->mtx rcu_read_lock hwsim_radio_lock batched_entropy_u8.lock irq_context: 0 cb_lock &rdev->wiphy.mtx &local->mtx rcu_read_lock hwsim_radio_lock kfence_freelist_lock irq_context: 0 &mm->mmap_lock &anon_vma->rwsem &base->lock irq_context: 0 &mm->mmap_lock &anon_vma->rwsem &base->lock &obj_hash[i].lock irq_context: 0 sb_writers#4 sb_internal &journal->j_state_lock &journal->j_wait_transaction_locked irq_context: 0 sb_writers#4 sb_internal jbd2_handle irq_context: 0 sb_writers#5 &type->i_mutex_dir_key#5/1 &sb->s_type->i_mutex_key#12 &sb->s_type->i_mutex_key#12/4 rcu_read_lock rcu_node_0 irq_context: 0 sb_writers#5 &type->i_mutex_dir_key#5/1 &sb->s_type->i_mutex_key#12 &sb->s_type->i_mutex_key#12/4 rcu_read_lock &rq->__lock irq_context: 0 sb_writers#5 &type->i_mutex_dir_key#5/1 &sb->s_type->i_mutex_key#12 &sb->s_type->i_mutex_key#12/4 rcu_read_lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 &sig->cred_guard_mutex tomoyo_ss remove_cache_srcu rcu_read_lock rcu_node_0 irq_context: 0 &sig->cred_guard_mutex tomoyo_ss remove_cache_srcu rcu_read_lock &rq->__lock irq_context: 0 &sig->cred_guard_mutex tomoyo_ss remove_cache_srcu rcu_read_lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 sb_writers#5 &type->i_mutex_dir_key#5 tomoyo_ss rcu_read_lock rcu_read_lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 sock_diag_mutex sock_diag_table_mutex irq_context: 0 sock_diag_mutex &n->list_lock irq_context: 0 sock_diag_mutex &n->list_lock &c->lock irq_context: 0 &p->lock remove_cache_srcu rcu_read_lock rcu_read_lock rcu_node_0 irq_context: 0 &p->lock remove_cache_srcu rcu_read_lock rcu_read_lock &rq->__lock irq_context: 0 &p->lock remove_cache_srcu rcu_read_lock rcu_read_lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 sock_diag_mutex &rq->__lock irq_context: 0 sock_diag_mutex &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 sb_writers#5 &type->i_mutex_dir_key#5 tomoyo_ss &cfs_rq->removed.lock irq_context: 0 &pipe->mutex/1 &mm->mmap_lock rcu_read_lock rcu_node_0 irq_context: 0 &pipe->mutex/1 &mm->mmap_lock rcu_read_lock &rq->__lock irq_context: 0 &pipe->mutex/1 &mm->mmap_lock rcu_read_lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 sb_writers#4 mmu_notifier_invalidate_range_start &rq->__lock irq_context: 0 sb_writers#5 &type->i_mutex_dir_key#5 rcu_read_lock &cfs_rq->removed.lock irq_context: 0 sb_writers#5 &type->i_mutex_dir_key#5 rcu_read_lock &obj_hash[i].lock irq_context: 0 &xt[i].mutex purge_vmap_area_lock quarantine_lock irq_context: softirq (&ndev->rs_timer) init_task.mems_allowed_seq.seqcount irq_context: 0 cb_lock genl_mutex calipso_doi_list_lock irq_context: 0 sb_writers#5 &type->i_mutex_dir_key#5/1 &sb->s_type->i_mutex_key#12 &sb->s_type->i_mutex_key#12/4 fs_reclaim mmu_notifier_invalidate_range_start &rq->__lock irq_context: 0 sb_writers#5 &type->i_mutex_dir_key#5/1 &sb->s_type->i_mutex_key#12 &sb->s_type->i_mutex_key#12/4 fs_reclaim mmu_notifier_invalidate_range_start &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 &f->f_pos_lock sb_writers#4 &sb->s_type->i_mutex_key#8 remove_cache_srcu rcu_read_lock &rcu_state.gp_wq irq_context: 0 &f->f_pos_lock sb_writers#4 &sb->s_type->i_mutex_key#8 remove_cache_srcu rcu_read_lock &rcu_state.gp_wq &p->pi_lock irq_context: 0 &f->f_pos_lock sb_writers#4 &sb->s_type->i_mutex_key#8 remove_cache_srcu rcu_read_lock &rcu_state.gp_wq &p->pi_lock &rq->__lock irq_context: 0 &f->f_pos_lock sb_writers#4 &sb->s_type->i_mutex_key#8 remove_cache_srcu rcu_read_lock &rcu_state.gp_wq &p->pi_lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 &f->f_pos_lock &type->i_mutex_dir_key#4 &cfs_rq->removed.lock irq_context: 0 &f->f_pos_lock &type->i_mutex_dir_key#4 &obj_hash[i].lock irq_context: 0 sb_writers#4 lock#4 rcu_read_lock pool_lock#2 irq_context: 0 sb_writers#4 lock#4 &obj_hash[i].lock irq_context: 0 tomoyo_ss stock_lock irq_context: 0 tomoyo_ss pcpu_lock stock_lock irq_context: 0 &f->f_pos_lock &mm->mmap_lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 delayed_uprobe_lock &rq->__lock irq_context: 0 delayed_uprobe_lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: softirq (&n->timer) &dir->lock irq_context: softirq (&n->timer) stock_lock irq_context: 0 sb_writers#4 sb_internal mmu_notifier_invalidate_range_start &rq->__lock irq_context: 0 sb_writers#4 sb_internal mmu_notifier_invalidate_range_start &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 sb_writers#3 &sb->s_type->i_mutex_key#9 batched_entropy_u8.lock irq_context: 0 sb_writers#3 &sb->s_type->i_mutex_key#9 kfence_freelist_lock irq_context: 0 sb_writers#5 &sb->s_type->i_mutex_key#12 fs_reclaim mmu_notifier_invalidate_range_start &rq->__lock irq_context: 0 sb_writers#5 &sb->s_type->i_mutex_key#12 fs_reclaim mmu_notifier_invalidate_range_start &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 sk_lock-AF_RDS irq_context: 0 sk_lock-AF_RDS slock-AF_RDS irq_context: 0 slock-AF_RDS irq_context: 0 uevent_sock_mutex fs_reclaim irq_context: 0 uevent_sock_mutex fs_reclaim mmu_notifier_invalidate_range_start irq_context: 0 uevent_sock_mutex pool_lock#2 irq_context: 0 uevent_sock_mutex nl_table_lock irq_context: 0 uevent_sock_mutex nl_table_wait.lock irq_context: 0 sb_writers#4 sb_internal jbd2_handle &sbi->s_orphan_lock &rq->__lock &cfs_rq->removed.lock irq_context: 0 &sbi->s_writepages_rwsem jbd2_handle batched_entropy_u8.lock irq_context: 0 &sbi->s_writepages_rwsem jbd2_handle kfence_freelist_lock irq_context: 0 &sbi->s_writepages_rwsem rcu_read_lock &c->lock irq_context: 0 (wq_completion)bat_events (work_completion)(&(&bat_priv->nc.work)->work) rcu_read_lock &obj_hash[i].lock pool_lock irq_context: 0 (wq_completion)ext4-rsv-conversion (work_completion)(&ei->i_rsv_conversion_work) &meta->lock irq_context: 0 (wq_completion)ext4-rsv-conversion (work_completion)(&ei->i_rsv_conversion_work) kfence_freelist_lock irq_context: 0 &sb->s_type->i_mutex_key#10 sk_lock-AF_QIPCRTR irq_context: 0 &sb->s_type->i_mutex_key#10 sk_lock-AF_QIPCRTR slock-AF_QIPCRTR irq_context: 0 &sb->s_type->i_mutex_key#10 sk_lock-AF_QIPCRTR clock-AF_QIPCRTR irq_context: 0 &sb->s_type->i_mutex_key#10 sk_lock-AF_QIPCRTR rlock-AF_QIPCRTR irq_context: 0 &sb->s_type->i_mutex_key#10 slock-AF_QIPCRTR irq_context: 0 (wq_completion)wg-crypt-wg0 (work_completion)(&peer->transmit_packet_work) &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: softirq (&ndev->rs_timer) rcu_read_lock rcu_read_lock rcu_read_lock_bh rcu_read_lock rcu_read_lock rcu_read_lock &nf_nat_locks[i] irq_context: softirq (&ndev->rs_timer) rcu_read_lock rcu_read_lock &tbl->lock init_task.mems_allowed_seq.seqcount irq_context: 0 bt_proto_lock sco_sk_list.lock irq_context: 0 &sb->s_type->i_mutex_key#10 sk_lock-AF_BLUETOOTH-BTPROTO_SCO irq_context: 0 &sb->s_type->i_mutex_key#10 sk_lock-AF_BLUETOOTH-BTPROTO_SCO &rq->__lock irq_context: 0 &sb->s_type->i_mutex_key#10 sk_lock-AF_BLUETOOTH-BTPROTO_SCO slock-AF_BLUETOOTH-BTPROTO_SCO irq_context: 0 &sb->s_type->i_mutex_key#10 slock-AF_BLUETOOTH-BTPROTO_SCO irq_context: 0 &sb->s_type->i_mutex_key#10 clock-AF_BLUETOOTH irq_context: 0 &sb->s_type->i_mutex_key#10 sco_sk_list.lock irq_context: 0 &xt[i].mutex fs_reclaim mmu_notifier_invalidate_range_start &cfs_rq->removed.lock irq_context: 0 &xt[i].mutex fs_reclaim mmu_notifier_invalidate_range_start &obj_hash[i].lock irq_context: 0 sb_writers#4 &type->i_mutex_dir_key#3/1 jbd2_handle &obj_hash[i].lock irq_context: 0 &sbi->s_writepages_rwsem rcu_read_lock &____s->seqcount#2 irq_context: 0 &p->lock &of->mutex kn->active#5 fs_reclaim mmu_notifier_invalidate_range_start &rq->__lock irq_context: 0 &group->inotify_data.idr_lock &obj_hash[i].lock pool_lock irq_context: 0 &net->xfrm.xfrm_cfg_mutex irq_context: 0 &net->xfrm.xfrm_cfg_mutex fs_reclaim irq_context: 0 &net->xfrm.xfrm_cfg_mutex fs_reclaim mmu_notifier_invalidate_range_start irq_context: 0 &net->xfrm.xfrm_cfg_mutex &c->lock irq_context: 0 &net->xfrm.xfrm_cfg_mutex pool_lock#2 irq_context: 0 &net->xfrm.xfrm_cfg_mutex rcu_read_lock &c->lock irq_context: 0 &net->xfrm.xfrm_cfg_mutex rcu_read_lock pool_lock#2 irq_context: 0 &net->xfrm.xfrm_cfg_mutex rcu_read_lock rcu_read_lock_bh &obj_hash[i].lock irq_context: 0 &net->xfrm.xfrm_cfg_mutex rcu_read_lock rcu_read_lock_bh pool_lock#2 irq_context: 0 &net->xfrm.xfrm_cfg_mutex rlock-AF_NETLINK irq_context: 0 sb_writers#4 sb_internal jbd2_handle &sbi->s_orphan_lock &lock->wait_lock irq_context: 0 sb_writers#8 &of->mutex &rq->__lock irq_context: 0 sb_writers#8 &of->mutex &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 sb_writers#4 sb_internal jbd2_handle rcu_read_lock &cfs_rq->removed.lock irq_context: 0 sb_writers#4 sb_internal jbd2_handle rcu_read_lock &obj_hash[i].lock irq_context: 0 sb_writers#4 &type->i_mutex_dir_key#3/1 jbd2_handle &ei->i_data_sem remove_cache_srcu &rq->__lock irq_context: 0 sb_writers#4 &type->i_mutex_dir_key#3/1 jbd2_handle &ei->i_data_sem remove_cache_srcu &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 sb_writers#8 &of->mutex kn->active#5 uevent_sock_mutex batched_entropy_u8.lock irq_context: 0 sb_writers#8 &of->mutex kn->active#5 uevent_sock_mutex kfence_freelist_lock irq_context: 0 &f->f_pos_lock sb_writers#4 &sb->s_type->i_mutex_key#8 &mm->mmap_lock rcu_node_0 irq_context: 0 &sb->s_type->i_mutex_key#10 sk_lock-AF_CAN rcu_state.exp_mutex &rnp->exp_wq[2] irq_context: 0 &sb->s_type->i_mutex_key#10 sk_lock-AF_CAN rcu_state.exp_mutex &rnp->exp_wq[3] irq_context: 0 &sb->s_type->i_mutex_key#10 sk_lock-AF_CAN rcu_state.exp_mutex &rq->__lock &cfs_rq->removed.lock irq_context: 0 &sbi->s_writepages_rwsem jbd2_handle rcu_read_lock &____s->seqcount#2 irq_context: 0 &sbi->s_writepages_rwsem jbd2_handle rcu_read_lock &____s->seqcount irq_context: 0 (wq_completion)ext4-rsv-conversion (work_completion)(&ei->i_rsv_conversion_work) jbd2_handle &sem->wait_lock irq_context: 0 (wq_completion)ext4-rsv-conversion (work_completion)(&ei->i_rsv_conversion_work) jbd2_handle &p->pi_lock irq_context: 0 (wq_completion)ext4-rsv-conversion (work_completion)(&ei->i_rsv_conversion_work) jbd2_handle &p->pi_lock &rq->__lock irq_context: 0 (wq_completion)ext4-rsv-conversion (work_completion)(&ei->i_rsv_conversion_work) jbd2_handle &p->pi_lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 &sbi->s_writepages_rwsem jbd2_handle rcu_node_0 irq_context: 0 listen_lock irq_context: 0 &rs->rs_recv_lock irq_context: 0 &tsk->futex_exit_mutex &rq->__lock &cfs_rq->removed.lock irq_context: 0 sb_writers#4 &type->i_mutex_dir_key#3/1 &sb->s_type->i_mutex_key#8 jbd2_handle &sbi->s_orphan_lock &cfs_rq->removed.lock irq_context: 0 sb_writers#4 &type->i_mutex_dir_key#3/1 &sb->s_type->i_mutex_key#8 jbd2_handle &sbi->s_orphan_lock &obj_hash[i].lock irq_context: 0 sb_writers#4 &type->i_mutex_dir_key#3/1 sb_internal jbd2_handle &ei->i_data_sem &cfs_rq->removed.lock irq_context: 0 sb_writers#8 tomoyo_ss rcu_read_lock rcu_read_lock &cfs_rq->removed.lock irq_context: 0 sb_writers#8 tomoyo_ss rcu_node_0 irq_context: 0 mmu_notifier_invalidate_range_start &rq->__lock irq_context: 0 mmu_notifier_invalidate_range_start &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 sb_writers#4 sb_internal remove_cache_srcu &pcp->lock &zone->lock irq_context: 0 sb_writers#4 sb_internal remove_cache_srcu &pcp->lock &zone->lock &____s->seqcount irq_context: 0 &mm->mmap_lock &anon_vma->rwsem &rq->__lock &cfs_rq->removed.lock irq_context: 0 &f->f_pos_lock sb_writers#4 &sb->s_type->i_mutex_key#8 &xa->xa_lock#9 &pcp->lock &zone->lock irq_context: 0 &f->f_pos_lock sb_writers#4 &sb->s_type->i_mutex_key#8 &xa->xa_lock#9 &pcp->lock &zone->lock &____s->seqcount irq_context: 0 &f->f_pos_lock sb_writers#4 &sb->s_type->i_mutex_key#8 &xa->xa_lock#9 batched_entropy_u8.lock irq_context: 0 &f->f_pos_lock sb_writers#4 &sb->s_type->i_mutex_key#8 &xa->xa_lock#9 kfence_freelist_lock irq_context: 0 &sbi->s_writepages_rwsem rcu_read_lock rcu_read_lock rcu_node_0 irq_context: 0 &sbi->s_writepages_rwsem rcu_read_lock rcu_read_lock &rq->__lock irq_context: 0 sk_lock-AF_ROSE irq_context: 0 sk_lock-AF_ROSE slock-AF_ROSE irq_context: 0 sk_lock-AF_ROSE rose_node_list_lock irq_context: 0 slock-AF_ROSE irq_context: 0 sb_writers#8 tomoyo_ss &cfs_rq->removed.lock irq_context: 0 dup_mmap_sem &mm->mmap_lock &mm->mmap_lock/1 &vma->vm_lock->lock &cfs_rq->removed.lock irq_context: 0 dup_mmap_sem &mm->mmap_lock &mm->mmap_lock/1 &vma->vm_lock->lock &obj_hash[i].lock irq_context: 0 rcu_read_lock rcu_read_lock rcu_read_lock rcu_read_lock &c->lock irq_context: 0 sb_writers#4 rcu_read_lock rcu_read_lock &rq->__lock &cfs_rq->removed.lock irq_context: 0 &net->ipv4.ra_mutex irq_context: 0 clock-AF_INET irq_context: 0 sb_writers#4 sb_internal jbd2_handle &ei->i_data_sem batched_entropy_u8.lock irq_context: 0 sb_writers#4 sb_internal jbd2_handle &ei->i_data_sem kfence_freelist_lock irq_context: 0 sb_writers#4 sb_internal jbd2_handle &ei->i_data_sem remove_cache_srcu irq_context: 0 sb_writers#4 sb_internal jbd2_handle &ei->i_data_sem remove_cache_srcu quarantine_lock irq_context: 0 sb_writers#4 sb_internal jbd2_handle &ei->i_data_sem remove_cache_srcu &c->lock irq_context: 0 sb_writers#4 sb_internal jbd2_handle &ei->i_data_sem remove_cache_srcu &n->list_lock irq_context: 0 sb_writers#4 sb_internal jbd2_handle &ei->i_data_sem remove_cache_srcu rcu_read_lock pool_lock#2 irq_context: 0 sb_writers#4 sb_internal jbd2_handle &ei->i_data_sem remove_cache_srcu &obj_hash[i].lock irq_context: 0 sb_writers#4 sb_internal jbd2_handle &ei->i_data_sem remove_cache_srcu &pcp->lock &zone->lock irq_context: 0 sb_writers#4 sb_internal jbd2_handle &ei->i_data_sem remove_cache_srcu &pcp->lock &zone->lock &____s->seqcount irq_context: 0 sb_writers#4 sb_internal jbd2_handle &ei->i_data_sem remove_cache_srcu pool_lock#2 irq_context: 0 sb_writers#4 sb_internal jbd2_handle &ei->i_data_sem &meta->lock irq_context: 0 sb_writers#4 batched_entropy_u8.lock irq_context: 0 (wq_completion)events_power_efficient (work_completion)(&(&tbl->gc_work)->work) &tbl->lock &n->lock &____s->seqcount#9 irq_context: 0 (wq_completion)events_power_efficient (work_completion)(&(&tbl->gc_work)->work) &tbl->lock rcu_read_lock lock#8 irq_context: 0 (wq_completion)events_power_efficient (work_completion)(&(&tbl->gc_work)->work) &tbl->lock &____s->seqcount#2 irq_context: 0 (wq_completion)events_power_efficient (work_completion)(&(&tbl->gc_work)->work) &tbl->lock &____s->seqcount irq_context: 0 (wq_completion)events_power_efficient (work_completion)(&(&tbl->gc_work)->work) &tbl->lock &obj_hash[i].lock pool_lock irq_context: 0 (wq_completion)events_power_efficient (work_completion)(&(&tbl->gc_work)->work) &tbl->lock krc.lock &obj_hash[i].lock irq_context: 0 (wq_completion)events_power_efficient (work_completion)(&(&tbl->gc_work)->work) &tbl->lock krc.lock &base->lock irq_context: 0 (wq_completion)events_power_efficient (work_completion)(&(&tbl->gc_work)->work) &tbl->lock krc.lock &base->lock &obj_hash[i].lock irq_context: softirq (&peer->timer_persistent_keepalive) &pcp->lock &zone->lock irq_context: 0 &sbi->s_writepages_rwsem jbd2_handle rcu_read_lock &obj_hash[i].lock pool_lock irq_context: 0 &sbi->s_writepages_rwsem &____s->seqcount#2 irq_context: 0 &net->xfrm.xfrm_cfg_mutex &rq->__lock irq_context: 0 &net->xfrm.xfrm_cfg_mutex &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 &net->xfrm.xfrm_cfg_mutex &obj_hash[i].lock irq_context: 0 rlock-AF_KEY irq_context: 0 (wq_completion)ext4-rsv-conversion (work_completion)(&ei->i_rsv_conversion_work) rcu_read_lock &cfs_rq->removed.lock irq_context: 0 (wq_completion)ext4-rsv-conversion (work_completion)(&ei->i_rsv_conversion_work) rcu_read_lock &obj_hash[i].lock irq_context: 0 (wq_completion)ext4-rsv-conversion (work_completion)(&ei->i_rsv_conversion_work) rcu_read_lock pool_lock#2 irq_context: softirq (&ndev->rs_timer) rcu_read_lock rcu_read_lock rcu_read_lock_bh _xmit_ETHER#2 rcu_read_lock &local->handle_wake_tx_queue_lock hwsim_radio_lock &____s->seqcount#2 irq_context: softirq (&ndev->rs_timer) rcu_read_lock rcu_read_lock rcu_read_lock_bh _xmit_ETHER#2 rcu_read_lock &local->handle_wake_tx_queue_lock hwsim_radio_lock &____s->seqcount irq_context: 0 sk_lock-AF_INET rcu_read_lock rcu_read_lock rcu_read_lock_bh dev->qdisc_tx_busylock ?: &qdisc_tx_busylock _xmit_ETHER#2 rcu_read_lock &____s->seqcount#2 irq_context: 0 sb_writers#4 &type->i_mutex_dir_key#3/1 sb_internal jbd2_handle &ei->i_data_sem quarantine_lock irq_context: 0 &sighand->siglock &n->list_lock irq_context: 0 &sighand->siglock &n->list_lock &c->lock irq_context: 0 sk_lock-AF_KCM irq_context: 0 sk_lock-AF_KCM slock-AF_KCM irq_context: 0 sk_lock-AF_KCM fs_reclaim irq_context: 0 sk_lock-AF_KCM fs_reclaim mmu_notifier_invalidate_range_start irq_context: 0 sk_lock-AF_KCM pool_lock#2 irq_context: 0 sk_lock-AF_KCM &____s->seqcount irq_context: 0 sk_lock-AF_KCM rcu_read_lock pool_lock#2 irq_context: 0 sk_lock-AF_KCM &obj_hash[i].lock irq_context: 0 sk_lock-AF_KCM &mm->mmap_lock irq_context: 0 sk_lock-AF_KCM &mm->mmap_lock &rq->__lock irq_context: 0 sk_lock-AF_KCM &mm->mmap_lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 sk_lock-AF_KCM &mux->lock irq_context: 0 slock-AF_KCM irq_context: 0 &sb->s_type->i_mutex_key#10 sk_lock-AF_KCM &obj_hash[i].lock irq_context: 0 &sb->s_type->i_mutex_key#10 sk_lock-AF_KCM pool_lock#2 irq_context: 0 (wq_completion)wg-crypt-wg2#3 (work_completion)(&({ do { const void *__vpp_verify = (typeof((worker) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __asm__ ("" : "=r"(__ptr) : "0"((typeof(*((worker))) *)((worker)))); (typeof((typeof(*((worker))) *)((worker)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); })->work) &rq->__lock irq_context: 0 sb_writers#8 kn->active#5 &kernfs_locks->open_file_mutex[count] fs_reclaim mmu_notifier_invalidate_range_start &rq->__lock irq_context: 0 sb_writers#8 kn->active#5 &kernfs_locks->open_file_mutex[count] fs_reclaim mmu_notifier_invalidate_range_start &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 &sbi->s_writepages_rwsem remove_cache_srcu irq_context: 0 &sbi->s_writepages_rwsem remove_cache_srcu quarantine_lock irq_context: 0 &sbi->s_writepages_rwsem remove_cache_srcu &c->lock irq_context: 0 &sbi->s_writepages_rwsem remove_cache_srcu &n->list_lock irq_context: 0 &sbi->s_writepages_rwsem remove_cache_srcu rcu_read_lock pool_lock#2 irq_context: 0 &sbi->s_writepages_rwsem remove_cache_srcu &obj_hash[i].lock irq_context: softirq rcu_read_lock rcu_read_lock slock-AF_INET/1 &____s->seqcount#2 irq_context: 0 sk_lock-AF_TIPC &c->lock irq_context: 0 sk_lock-AF_TIPC &n->list_lock irq_context: 0 sk_lock-AF_TIPC &n->list_lock &c->lock irq_context: 0 sk_lock-AF_TIPC &obj_hash[i].lock irq_context: 0 (wq_completion)wg-crypt-wg1#3 (work_completion)(&peer->transmit_packet_work) &rq->__lock irq_context: 0 (wq_completion)writeback (work_completion)(&(&wb->dwork)->work) &type->s_umount_key#43 rcu_read_lock &xa->xa_lock#9 &wb->work_lock &obj_hash[i].lock irq_context: 0 (wq_completion)writeback (work_completion)(&(&wb->dwork)->work) &type->s_umount_key#43 rcu_read_lock &xa->xa_lock#9 &wb->work_lock &base->lock irq_context: 0 (wq_completion)writeback (work_completion)(&(&wb->dwork)->work) &type->s_umount_key#43 rcu_read_lock &xa->xa_lock#9 &wb->work_lock &base->lock &obj_hash[i].lock irq_context: 0 &sbi->s_writepages_rwsem jbd2_handle &n->list_lock irq_context: 0 &sbi->s_writepages_rwsem jbd2_handle &n->list_lock &c->lock irq_context: 0 &sbi->s_writepages_rwsem jbd2_handle rcu_read_lock batched_entropy_u8.lock irq_context: 0 &sbi->s_writepages_rwsem jbd2_handle rcu_read_lock kfence_freelist_lock irq_context: 0 &sbi->s_writepages_rwsem jbd2_handle &journal->j_state_lock irq_context: 0 &sbi->s_writepages_rwsem jbd2_handle &journal->j_state_lock &journal->j_wait_commit irq_context: 0 &sbi->s_writepages_rwsem jbd2_handle &journal->j_state_lock &journal->j_wait_commit &p->pi_lock irq_context: 0 &sbi->s_writepages_rwsem jbd2_handle &journal->j_state_lock &journal->j_wait_commit &p->pi_lock &rq->__lock irq_context: 0 &sbi->s_writepages_rwsem jbd2_handle &journal->j_state_lock &journal->j_wait_commit &p->pi_lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)ext4-rsv-conversion (work_completion)(&ei->i_rsv_conversion_work) jbd2_handle &journal->j_state_lock irq_context: 0 &sbi->s_writepages_rwsem jbd2_handle &journal->j_wait_updates &p->pi_lock irq_context: 0 &sbi->s_writepages_rwsem jbd2_handle &journal->j_wait_updates &p->pi_lock &rq->__lock irq_context: 0 &sbi->s_writepages_rwsem jbd2_handle &journal->j_wait_updates &p->pi_lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)ext4-rsv-conversion (work_completion)(&ei->i_rsv_conversion_work) &obj_hash[i].lock pool_lock irq_context: 0 sb_writers#4 &type->i_mutex_dir_key#3/1 sb_internal jbd2_handle &sbi->s_orphan_lock &rq->__lock irq_context: 0 sb_writers#4 &type->i_mutex_dir_key#3/1 sb_internal jbd2_handle &sbi->s_orphan_lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 sb_writers#8 &of->mutex kn->active#5 uevent_sock_mutex remove_cache_srcu pool_lock#2 irq_context: 0 &sb->s_type->i_mutex_key#10 sk_lock-AF_PHONET rcu_state.exp_mutex &cfs_rq->removed.lock irq_context: 0 &sb->s_type->i_mutex_key#10 sk_lock-AF_PHONET rcu_state.exp_mutex pool_lock#2 irq_context: 0 &sbi->s_writepages_rwsem jbd2_handle rcu_read_lock &pcp->lock &zone->lock irq_context: 0 rcu_read_lock &im->lock irq_context: 0 &sig->cred_guard_mutex remove_cache_srcu pool_lock#2 irq_context: 0 rcu_read_lock rcu_read_lock rcu_read_lock rcu_read_lock per_cpu_ptr(&cgroup_rstat_cpu_lock, cpu) irq_context: 0 &sb->s_type->i_mutex_key#10 sk_lock-AF_INET6 slock-AF_INET6 &tcp_hashinfo.bhash[i].lock stock_lock irq_context: 0 &sb->s_type->i_mutex_key#10 sk_lock-AF_INET6 slock-AF_INET6 &tcp_hashinfo.bhash[i].lock &obj_hash[i].lock irq_context: 0 &sb->s_type->i_mutex_key#10 sk_lock-AF_INET6 slock-AF_INET6 &tcp_hashinfo.bhash[i].lock pool_lock#2 irq_context: 0 &sb->s_type->i_mutex_key#10 sk_lock-AF_INET6 slock-AF_INET6 &tcp_hashinfo.bhash[i].lock &tcp_hashinfo.bhash2[i].lock stock_lock irq_context: 0 &sb->s_type->i_mutex_key#10 sk_lock-AF_INET6 slock-AF_INET6 &tcp_hashinfo.bhash[i].lock &tcp_hashinfo.bhash2[i].lock &obj_hash[i].lock irq_context: 0 &sb->s_type->i_mutex_key#10 sk_lock-AF_INET6 slock-AF_INET6 &tcp_hashinfo.bhash[i].lock &tcp_hashinfo.bhash2[i].lock pool_lock#2 irq_context: 0 sb_writers#4 &type->i_mutex_dir_key#3/1 tomoyo_ss remove_cache_srcu rcu_read_lock &rcu_state.gp_wq irq_context: 0 sb_writers#4 &type->i_mutex_dir_key#3/1 tomoyo_ss remove_cache_srcu rcu_read_lock &rcu_state.gp_wq &p->pi_lock irq_context: 0 sb_writers#4 &type->i_mutex_dir_key#3/1 tomoyo_ss remove_cache_srcu rcu_read_lock &rcu_state.gp_wq &p->pi_lock &rq->__lock irq_context: 0 sb_writers#4 &type->i_mutex_dir_key#3/1 tomoyo_ss remove_cache_srcu rcu_read_lock &rcu_state.gp_wq &p->pi_lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 &sb->s_type->i_mutex_key#10 &rnp->exp_wq[0] irq_context: softirq (&ndev->rs_timer) rcu_read_lock rcu_read_lock rcu_read_lock_bh dev->qdisc_tx_busylock ?: &qdisc_tx_busylock _xmit_ETHER#2 rcu_read_lock &c->lock irq_context: 0 dup_mmap_sem &mm->mmap_lock &mm->mmap_lock/1 remove_cache_srcu &pcp->lock &zone->lock irq_context: 0 dup_mmap_sem &mm->mmap_lock &mm->mmap_lock/1 remove_cache_srcu &pcp->lock &zone->lock &____s->seqcount irq_context: 0 sk_lock-AF_INET rcu_read_lock fastopen_seqlock.seqcount irq_context: 0 &net->xfrm.xfrm_cfg_mutex &n->list_lock irq_context: 0 &net->xfrm.xfrm_cfg_mutex &n->list_lock &c->lock irq_context: 0 &net->xfrm.xfrm_cfg_mutex pfkey_mutex irq_context: 0 &net->xfrm.xfrm_cfg_mutex pfkey_mutex crypto_alg_sem irq_context: 0 &net->xfrm.xfrm_cfg_mutex pfkey_mutex (kmod_concurrent_max).lock irq_context: 0 &net->xfrm.xfrm_cfg_mutex pfkey_mutex fs_reclaim irq_context: 0 &net->xfrm.xfrm_cfg_mutex pfkey_mutex fs_reclaim mmu_notifier_invalidate_range_start irq_context: 0 &net->xfrm.xfrm_cfg_mutex pfkey_mutex pool_lock#2 irq_context: 0 &net->xfrm.xfrm_cfg_mutex pfkey_mutex &c->lock irq_context: 0 &net->xfrm.xfrm_cfg_mutex pfkey_mutex &obj_hash[i].lock irq_context: 0 &net->xfrm.xfrm_cfg_mutex pfkey_mutex rcu_read_lock &pool->lock/1 irq_context: 0 &net->xfrm.xfrm_cfg_mutex pfkey_mutex rcu_read_lock &pool->lock/1 &obj_hash[i].lock irq_context: 0 &net->xfrm.xfrm_cfg_mutex pfkey_mutex rcu_read_lock &pool->lock/1 pool_lock#2 irq_context: 0 &net->xfrm.xfrm_cfg_mutex pfkey_mutex rcu_read_lock &pool->lock/1 &p->pi_lock irq_context: 0 &net->xfrm.xfrm_cfg_mutex pfkey_mutex &x->wait#17 irq_context: 0 &net->xfrm.xfrm_cfg_mutex pfkey_mutex &rq->__lock irq_context: 0 &net->xfrm.xfrm_cfg_mutex pfkey_mutex &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 bt_proto_lock rcu_read_lock pool_lock#2 irq_context: 0 bt_proto_lock rfcomm_sk_list.lock irq_context: 0 sk_lock-AF_BLUETOOTH-BTPROTO_RFCOMM irq_context: 0 sk_lock-AF_BLUETOOTH-BTPROTO_RFCOMM &rq->__lock irq_context: 0 sk_lock-AF_BLUETOOTH-BTPROTO_RFCOMM &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 sk_lock-AF_BLUETOOTH-BTPROTO_RFCOMM slock-AF_BLUETOOTH-BTPROTO_RFCOMM irq_context: 0 slock-AF_BLUETOOTH-BTPROTO_RFCOMM irq_context: 0 rfcomm_mutex fs_reclaim irq_context: 0 rfcomm_mutex fs_reclaim mmu_notifier_invalidate_range_start irq_context: 0 rfcomm_mutex stock_lock irq_context: 0 rfcomm_mutex pool_lock#2 irq_context: 0 rfcomm_mutex mmu_notifier_invalidate_range_start irq_context: 0 rfcomm_mutex &sb->s_type->i_lock_key#8 irq_context: 0 rfcomm_mutex bt_proto_lock irq_context: 0 rfcomm_mutex bt_proto_lock pool_lock#2 irq_context: 0 rfcomm_mutex bt_proto_lock &dir->lock irq_context: 0 rfcomm_mutex bt_proto_lock &obj_hash[i].lock irq_context: 0 rfcomm_mutex bt_proto_lock &c->lock irq_context: 0 rfcomm_mutex bt_proto_lock chan_list_lock irq_context: 0 rfcomm_mutex bt_proto_lock l2cap_sk_list.lock irq_context: 0 rfcomm_mutex sk_lock-AF_BLUETOOTH-BTPROTO_L2CAP irq_context: 0 rfcomm_mutex sk_lock-AF_BLUETOOTH-BTPROTO_L2CAP slock-AF_BLUETOOTH-BTPROTO_L2CAP irq_context: 0 rfcomm_mutex sk_lock-AF_BLUETOOTH-BTPROTO_L2CAP chan_list_lock irq_context: 0 rfcomm_mutex sk_lock-AF_BLUETOOTH-BTPROTO_L2CAP &rq->__lock irq_context: 0 rfcomm_mutex sk_lock-AF_BLUETOOTH-BTPROTO_L2CAP &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 rfcomm_mutex slock-AF_BLUETOOTH-BTPROTO_L2CAP irq_context: 0 rfcomm_mutex &c->lock irq_context: 0 rfcomm_mutex &obj_hash[i].lock irq_context: 0 rfcomm_mutex hci_dev_list_lock irq_context: 0 rfcomm_mutex &hdev->lock irq_context: 0 rfcomm_mutex &hdev->lock fs_reclaim irq_context: 0 rfcomm_mutex &hdev->lock fs_reclaim mmu_notifier_invalidate_range_start irq_context: 0 rfcomm_mutex &hdev->lock pool_lock#2 irq_context: 0 rfcomm_mutex &hdev->lock &obj_hash[i].lock irq_context: 0 rfcomm_mutex &hdev->lock &obj_hash[i].lock pool_lock irq_context: 0 rfcomm_mutex &hdev->lock &x->wait#9 irq_context: 0 rfcomm_mutex &hdev->lock &c->lock irq_context: 0 rfcomm_mutex &hdev->lock &list->lock#7 irq_context: 0 rfcomm_mutex &hdev->lock rcu_read_lock &pool->lock/1 irq_context: 0 rfcomm_mutex &hdev->lock rcu_read_lock &pool->lock/1 &obj_hash[i].lock irq_context: 0 rfcomm_mutex &hdev->lock rcu_read_lock &pool->lock/1 &p->pi_lock irq_context: 0 rfcomm_mutex &hdev->lock rcu_read_lock &pool->lock/1 &p->pi_lock &rq->__lock irq_context: 0 rfcomm_mutex &hdev->lock rcu_read_lock &pool->lock/1 &p->pi_lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 rfcomm_mutex &hdev->lock &rq->__lock irq_context: 0 rfcomm_mutex &hdev->lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 rfcomm_mutex &hdev->lock &conn->chan_lock irq_context: 0 rfcomm_mutex &hdev->lock &conn->chan_lock &chan->lock/1 irq_context: 0 rfcomm_mutex &hdev->lock &conn->chan_lock &chan->lock/1 &obj_hash[i].lock irq_context: 0 rfcomm_mutex &hdev->lock &conn->chan_lock &chan->lock/1 &base->lock irq_context: 0 rfcomm_mutex &hdev->lock &conn->chan_lock &chan->lock/1 &base->lock &obj_hash[i].lock irq_context: 0 rfcomm_mutex &hdev->lock &conn->chan_lock &chan->lock/1 chan_list_lock irq_context: 0 rfcomm_mutex sk_lock-AF_BLUETOOTH-BTPROTO_L2CAP &ei->socket.wq.wait irq_context: 0 rfcomm_mutex (&s->timer) irq_context: 0 rfcomm_mutex &rq->__lock irq_context: 0 rfcomm_mutex &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 rfcomm_mutex &base->lock irq_context: 0 rfcomm_mutex &base->lock &obj_hash[i].lock irq_context: 0 sk_lock-AF_BLUETOOTH-BTPROTO_RFCOMM &ei->socket.wq.wait irq_context: 0 &net->xfrm.xfrm_cfg_mutex &lock->wait_lock irq_context: 0 rfcomm_mutex &d->lock irq_context: 0 rfcomm_mutex &d->lock sk_lock-AF_BLUETOOTH-BTPROTO_RFCOMM irq_context: 0 rfcomm_mutex &d->lock sk_lock-AF_BLUETOOTH-BTPROTO_RFCOMM slock-AF_BLUETOOTH-BTPROTO_RFCOMM irq_context: 0 rfcomm_mutex &d->lock sk_lock-AF_BLUETOOTH-BTPROTO_RFCOMM rcu_read_lock &ei->socket.wq.wait irq_context: 0 rfcomm_mutex &d->lock sk_lock-AF_BLUETOOTH-BTPROTO_RFCOMM rcu_read_lock &ei->socket.wq.wait &p->pi_lock irq_context: 0 rfcomm_mutex &d->lock sk_lock-AF_BLUETOOTH-BTPROTO_RFCOMM rcu_read_lock &ei->socket.wq.wait &p->pi_lock &rq->__lock irq_context: 0 rfcomm_mutex &d->lock sk_lock-AF_BLUETOOTH-BTPROTO_RFCOMM rcu_read_lock &ei->socket.wq.wait &p->pi_lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 rfcomm_mutex &d->lock slock-AF_BLUETOOTH-BTPROTO_RFCOMM irq_context: 0 rfcomm_mutex &list->lock#33 irq_context: 0 &mm->mmap_lock &mapping->i_mmap_rwsem &anon_vma->rwsem rcu_read_lock &rcu_state.gp_wq irq_context: 0 &mm->mmap_lock &mapping->i_mmap_rwsem &anon_vma->rwsem rcu_read_lock &rcu_state.gp_wq &p->pi_lock irq_context: 0 &mm->mmap_lock &mapping->i_mmap_rwsem &anon_vma->rwsem rcu_read_lock &rcu_state.gp_wq &p->pi_lock &rq->__lock irq_context: 0 &mm->mmap_lock &mapping->i_mmap_rwsem &anon_vma->rwsem rcu_read_lock &rcu_state.gp_wq &p->pi_lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 &net->xfrm.xfrm_cfg_mutex pfkey_mutex running_helpers_waitq.lock irq_context: 0 &net->xfrm.xfrm_cfg_mutex pfkey_mutex rcu_read_lock &pool->lock/1 &p->pi_lock &rq->__lock irq_context: 0 &net->xfrm.xfrm_cfg_mutex pfkey_mutex rcu_read_lock &pool->lock/1 &p->pi_lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 sb_writers#4 &type->i_mutex_dir_key#3/1 &sb->s_type->i_mutex_key#8 remove_cache_srcu rcu_read_lock rcu_node_0 irq_context: 0 sb_writers#4 &type->i_mutex_dir_key#3/1 &sb->s_type->i_mutex_key#8 remove_cache_srcu rcu_read_lock &rq->__lock irq_context: 0 sb_writers#4 &type->i_mutex_dir_key#3/1 &sb->s_type->i_mutex_key#8 remove_cache_srcu rcu_read_lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)events (work_completion)(&(&nsim_dev->trap_data->trap_report_dw)->work) &devlink->lock_key#4 &cfs_rq->removed.lock irq_context: 0 &net->xfrm.xfrm_cfg_mutex pfkey_mutex crypto_alg_sem irq_context: 0 &net->xfrm.xfrm_cfg_mutex pfkey_mutex (crypto_chain).rwsem irq_context: 0 &net->xfrm.xfrm_cfg_mutex pfkey_mutex (crypto_chain).rwsem fs_reclaim irq_context: 0 &net->xfrm.xfrm_cfg_mutex pfkey_mutex (crypto_chain).rwsem fs_reclaim mmu_notifier_invalidate_range_start irq_context: 0 &net->xfrm.xfrm_cfg_mutex pfkey_mutex (crypto_chain).rwsem pool_lock#2 irq_context: 0 &net->xfrm.xfrm_cfg_mutex pfkey_mutex (crypto_chain).rwsem kthread_create_lock irq_context: 0 &net->xfrm.xfrm_cfg_mutex pfkey_mutex (crypto_chain).rwsem &p->pi_lock irq_context: 0 &net->xfrm.xfrm_cfg_mutex pfkey_mutex (crypto_chain).rwsem &p->pi_lock &rq->__lock irq_context: 0 &net->xfrm.xfrm_cfg_mutex pfkey_mutex (crypto_chain).rwsem &p->pi_lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 &net->xfrm.xfrm_cfg_mutex pfkey_mutex (crypto_chain).rwsem &rq->__lock irq_context: 0 &net->xfrm.xfrm_cfg_mutex pfkey_mutex (crypto_chain).rwsem &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 &net->xfrm.xfrm_cfg_mutex pfkey_mutex (crypto_chain).rwsem &x->wait irq_context: 0 &net->xfrm.xfrm_cfg_mutex pfkey_mutex (crypto_chain).rwsem &obj_hash[i].lock irq_context: 0 &net->xfrm.xfrm_cfg_mutex pfkey_mutex &x->wait#21 irq_context: 0 &net->xfrm.xfrm_cfg_mutex pfkey_mutex &base->lock irq_context: 0 &net->xfrm.xfrm_cfg_mutex pfkey_mutex &base->lock &obj_hash[i].lock irq_context: 0 &net->xfrm.xfrm_cfg_mutex pfkey_mutex (&timer.timer) irq_context: softirq (&icsk->icsk_retransmit_timer) slock-AF_INET pool_lock#2 irq_context: softirq (&icsk->icsk_retransmit_timer) slock-AF_INET &c->lock irq_context: softirq rcu_read_lock rcu_read_lock slock-AF_INET/1 &f->f_owner.lock irq_context: 0 &f->f_pos_lock &type->i_mutex_dir_key#3 &rq->__lock &cfs_rq->removed.lock irq_context: softirq (&icsk->icsk_delack_timer) slock-AF_INET rcu_read_lock rcu_read_lock rcu_read_lock_bh &list->lock#5 irq_context: softirq rcu_read_lock rcu_read_lock slock-AF_INET/1 hrtimer_bases.lock irq_context: softirq rcu_read_lock rcu_read_lock slock-AF_INET/1 hrtimer_bases.lock &obj_hash[i].lock irq_context: softirq rcu_read_lock rcu_read_lock slock-AF_INET/1 hrtimer_bases.lock tk_core.seq.seqcount irq_context: 0 &net->xfrm.xfrm_cfg_mutex pfkey_mutex remove_cache_srcu irq_context: 0 &net->xfrm.xfrm_cfg_mutex pfkey_mutex remove_cache_srcu quarantine_lock irq_context: 0 &net->xfrm.xfrm_cfg_mutex pfkey_mutex remove_cache_srcu &c->lock irq_context: 0 &net->xfrm.xfrm_cfg_mutex pfkey_mutex remove_cache_srcu &n->list_lock irq_context: 0 &net->xfrm.xfrm_cfg_mutex pfkey_mutex remove_cache_srcu rcu_read_lock pool_lock#2 irq_context: 0 &net->xfrm.xfrm_cfg_mutex pfkey_mutex remove_cache_srcu &obj_hash[i].lock irq_context: 0 &net->xfrm.xfrm_cfg_mutex pfkey_mutex remove_cache_srcu &rq->__lock irq_context: 0 &net->xfrm.xfrm_cfg_mutex pfkey_mutex remove_cache_srcu &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 &net->xfrm.xfrm_cfg_mutex pfkey_mutex (crypto_chain).rwsem &c->lock irq_context: 0 &net->xfrm.xfrm_cfg_mutex pfkey_mutex &n->list_lock irq_context: 0 &net->xfrm.xfrm_cfg_mutex pfkey_mutex &n->list_lock &c->lock irq_context: 0 &net->xfrm.xfrm_cfg_mutex pfkey_mutex (crypto_chain).rwsem rcu_read_lock rcu_node_0 irq_context: 0 &net->xfrm.xfrm_cfg_mutex pfkey_mutex (crypto_chain).rwsem rcu_read_lock &rq->__lock irq_context: 0 &net->xfrm.xfrm_cfg_mutex pfkey_mutex (crypto_chain).rwsem rcu_read_lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 &sig->cred_guard_mutex &sig->exec_update_lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 &net->xfrm.xfrm_cfg_mutex pfkey_mutex crypto_alg_sem &rq->__lock irq_context: 0 &net->xfrm.xfrm_cfg_mutex pfkey_mutex crypto_alg_sem &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 &net->xfrm.xfrm_cfg_mutex pfkey_mutex (crypto_chain).rwsem &n->list_lock irq_context: 0 &net->xfrm.xfrm_cfg_mutex pfkey_mutex (crypto_chain).rwsem &n->list_lock &c->lock irq_context: 0 &sb->s_type->i_mutex_key#10 sk_lock-AF_BLUETOOTH-BTPROTO_RFCOMM irq_context: 0 &sb->s_type->i_mutex_key#10 sk_lock-AF_BLUETOOTH-BTPROTO_RFCOMM slock-AF_BLUETOOTH-BTPROTO_RFCOMM irq_context: 0 &sb->s_type->i_mutex_key#10 slock-AF_BLUETOOTH-BTPROTO_RFCOMM irq_context: 0 &sb->s_type->i_mutex_key#10 rfcomm_sk_list.lock irq_context: 0 &sb->s_type->i_mutex_key#10 &d->lock irq_context: 0 &sb->s_type->i_mutex_key#10 &list->lock#33 irq_context: 0 &net->xfrm.xfrm_cfg_mutex pfkey_mutex (crypto_chain).rwsem &____s->seqcount irq_context: 0 &net->xfrm.xfrm_cfg_mutex pfkey_mutex (crypto_chain).rwsem rcu_read_lock pool_lock#2 irq_context: 0 &net->xfrm.xfrm_cfg_mutex pfkey_mutex &obj_hash[i].lock pool_lock irq_context: 0 &net->xfrm.xfrm_cfg_mutex pfkey_mutex quarantine_lock irq_context: 0 &net->xfrm.xfrm_cfg_mutex rlock-AF_KEY irq_context: 0 tasklist_lock &sighand->siglock &(&sig->stats_lock)->lock &____s->seqcount#5 fill_pool_map-wait-type-override &n->list_lock irq_context: 0 tasklist_lock &sighand->siglock &(&sig->stats_lock)->lock &____s->seqcount#5 fill_pool_map-wait-type-override &n->list_lock &c->lock irq_context: 0 &base->lock fill_pool_map-wait-type-override pool_lock#2 irq_context: 0 &base->lock fill_pool_map-wait-type-override &c->lock irq_context: 0 &base->lock fill_pool_map-wait-type-override pool_lock irq_context: 0 &mm->mmap_lock &mapping->i_mmap_rwsem fill_pool_map-wait-type-override &n->list_lock irq_context: 0 &mm->mmap_lock &mapping->i_mmap_rwsem fill_pool_map-wait-type-override &n->list_lock &c->lock irq_context: 0 &sb->s_type->i_mutex_key#10 rcu_state.exp_mutex &cfs_rq->removed.lock irq_context: 0 &sb->s_type->i_mutex_key#10 rcu_state.exp_mutex pool_lock#2 irq_context: 0 &sb->s_type->i_mutex_key#10 rcu_state.exp_mutex rcu_read_lock &pool->lock fill_pool_map-wait-type-override &c->lock irq_context: softirq (&pool->mayday_timer) &pool->lock/1 wq_mayday_lock &p->pi_lock irq_context: softirq (&pool->mayday_timer) &pool->lock/1 wq_mayday_lock &p->pi_lock &rq->__lock irq_context: softirq (&pool->mayday_timer) &pool->lock/1 wq_mayday_lock &p->pi_lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)wg-crypt-wg2#4 (work_completion)(&peer->transmit_packet_work) &peer->endpoint_lock rcu_read_lock_bh batched_entropy_u32.lock crngs.lock irq_context: 0 wq_pool_attach_mutex &p->pi_lock irq_context: 0 wq_pool_attach_mutex &p->pi_lock &rq->__lock irq_context: 0 wq_pool_attach_mutex &p->pi_lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 &pool->lock/1 wq_mayday_lock irq_context: 0 &mm->mmap_lock fill_pool_map-wait-type-override &n->list_lock irq_context: 0 &mm->mmap_lock fill_pool_map-wait-type-override &n->list_lock &c->lock irq_context: 0 &mm->mmap_lock &mapping->i_mmap_rwsem &anon_vma->rwsem fill_pool_map-wait-type-override &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)bat_events (work_completion)(&(&bat_priv->nc.work)->work) fill_pool_map-wait-type-override &n->list_lock irq_context: 0 (wq_completion)bat_events (work_completion)(&(&bat_priv->nc.work)->work) fill_pool_map-wait-type-override &n->list_lock &c->lock irq_context: 0 &f->f_pos_lock sb_writers#4 &sb->s_type->i_mutex_key#8 pgd_lock irq_context: 0 &f->f_pos_lock sb_writers#4 &sb->s_type->i_mutex_key#8 key irq_context: 0 &f->f_pos_lock sb_writers#4 &sb->s_type->i_mutex_key#8 pcpu_lock irq_context: 0 &f->f_pos_lock sb_writers#4 &sb->s_type->i_mutex_key#8 percpu_counters_lock irq_context: 0 &f->f_pos_lock sb_writers#4 &sb->s_type->i_mutex_key#8 pcpu_lock stock_lock irq_context: 0 tracepoints_mutex rcu_state.exp_mutex rcu_node_0 irq_context: 0 tracepoints_mutex rcu_state.exp_mutex &obj_hash[i].lock irq_context: 0 tracepoints_mutex rcu_state.exp_mutex rcu_read_lock &pool->lock irq_context: 0 tracepoints_mutex rcu_state.exp_mutex rcu_read_lock &pool->lock &obj_hash[i].lock irq_context: 0 tracepoints_mutex rcu_state.exp_mutex rcu_read_lock &pool->lock &p->pi_lock irq_context: 0 tracepoints_mutex rcu_state.exp_mutex &rnp->exp_wq[0] irq_context: 0 tracepoints_mutex rcu_state.exp_mutex &rq->__lock irq_context: 0 tracepoints_mutex rcu_state.exp_mutex &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 &f->f_pos_lock sb_writers#4 &sb->s_type->i_mutex_key#8 &ei->i_data_sem &ei->i_es_lock key#6 irq_context: 0 &f->f_pos_lock sb_writers#4 &sb->s_type->i_mutex_key#8 &ei->i_data_sem pgd_lock irq_context: 0 &f->f_pos_lock sb_writers#4 &sb->s_type->i_mutex_key#8 &ei->i_data_sem key irq_context: 0 &f->f_pos_lock sb_writers#4 &sb->s_type->i_mutex_key#8 &ei->i_data_sem pcpu_lock irq_context: 0 &f->f_pos_lock sb_writers#4 &sb->s_type->i_mutex_key#8 &ei->i_data_sem percpu_counters_lock irq_context: 0 &f->f_pos_lock sb_writers#4 &sb->s_type->i_mutex_key#8 &ei->i_data_sem &cfs_rq->removed.lock irq_context: 0 cb_lock remove_cache_srcu pool_lock#2 irq_context: 0 cb_lock remove_cache_srcu &rq->__lock irq_context: 0 cb_lock remove_cache_srcu &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 sb_writers#3 oom_adj_mutex rcu_read_lock rcu_node_0 irq_context: 0 sb_writers#3 oom_adj_mutex rcu_read_lock &rq->__lock irq_context: 0 sb_writers#3 oom_adj_mutex rcu_read_lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)bat_events (work_completion)(&(&forw_packet_aggr->delayed_work)->work) rcu_read_lock_bh rcu_read_lock &n->list_lock irq_context: 0 (wq_completion)bat_events (work_completion)(&(&forw_packet_aggr->delayed_work)->work) rcu_read_lock_bh rcu_read_lock &n->list_lock &c->lock irq_context: 0 (wq_completion)wg-crypt-wg0#5 (work_completion)(&peer->transmit_packet_work) &peer->endpoint_lock rcu_read_lock_bh rcu_read_lock rcu_read_lock_bh rcu_read_lock rlock-AF_PACKET irq_context: 0 (wq_completion)events_unbound (reaper_work).work pool_lock irq_context: 0 cb_lock &rdev->wiphy.mtx rcu_read_lock rcu_read_lock rcu_node_0 irq_context: 0 cb_lock &rdev->wiphy.mtx rcu_read_lock rcu_read_lock &rq->__lock irq_context: 0 cb_lock &rdev->wiphy.mtx rcu_read_lock rcu_read_lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)bat_events (work_completion)(&(&forw_packet_aggr->delayed_work)->work) &cfs_rq->removed.lock irq_context: 0 sb_writers#4 &type->i_mutex_dir_key#3/1 tomoyo_ss pgd_lock irq_context: 0 sb_writers#4 &type->i_mutex_dir_key#3/1 tomoyo_ss stock_lock irq_context: 0 sb_writers#4 &type->i_mutex_dir_key#3/1 tomoyo_ss key irq_context: 0 sb_writers#4 &type->i_mutex_dir_key#3/1 tomoyo_ss pcpu_lock irq_context: 0 sb_writers#4 &type->i_mutex_dir_key#3/1 tomoyo_ss percpu_counters_lock irq_context: 0 sb_writers#4 &type->i_mutex_dir_key#3/1 tomoyo_ss pcpu_lock stock_lock irq_context: 0 pernet_ops_rwsem nf_log_mutex &rq->__lock irq_context: 0 pernet_ops_rwsem nf_log_mutex &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 pernet_ops_rwsem __ip_vs_app_mutex &rq->__lock irq_context: 0 pernet_ops_rwsem __ip_vs_app_mutex &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 rtnl_mutex rcu_read_lock sysctl_lock irq_context: 0 rtnl_mutex rcu_read_lock &sb->s_type->i_lock_key#23 irq_context: 0 rtnl_mutex &sb->s_type->i_lock_key#23 irq_context: 0 rtnl_mutex &sb->s_type->i_lock_key#23 &dentry->d_lock irq_context: 0 rtnl_mutex &dentry->d_lock irq_context: 0 rtnl_mutex rename_lock.seqcount irq_context: 0 rtnl_mutex rcu_read_lock &dentry->d_lock irq_context: 0 rtnl_mutex &dentry->d_lock &sb->s_type->i_lock_key#23 &dentry->d_lock &lru->node[i].lock irq_context: 0 rtnl_mutex &dentry->d_lock &lru->node[i].lock irq_context: 0 rtnl_mutex &s->s_inode_list_lock irq_context: 0 rtnl_mutex rcu_read_lock &tb->tb6_lock rlock-AF_NETLINK irq_context: 0 &f->f_pos_lock &type->i_mutex_dir_key#3 remove_cache_srcu &cfs_rq->removed.lock irq_context: 0 sb_writers#4 tomoyo_ss rcu_node_0 irq_context: 0 sb_writers#4 &type->i_mutex_dir_key#3/1 jbd2_handle &xa->xa_lock#9 &c->lock irq_context: 0 &sig->cred_guard_mutex &sig->exec_update_lock &sem->wait_lock irq_context: 0 &sig->cred_guard_mutex &sig->exec_update_lock &p->pi_lock irq_context: 0 &sig->cred_guard_mutex &sig->exec_update_lock &p->pi_lock &rq->__lock irq_context: 0 &sig->cred_guard_mutex &sig->exec_update_lock &p->pi_lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 rtnl_mutex _xmit_IPGRE pool_lock#2 irq_context: 0 rtnl_mutex _xmit_IPGRE (console_sem).lock irq_context: 0 rtnl_mutex _xmit_IPGRE console_lock console_srcu console_owner_lock irq_context: 0 rtnl_mutex _xmit_IPGRE console_lock console_srcu console_owner irq_context: 0 rtnl_mutex _xmit_IPGRE console_lock console_srcu console_owner &port_lock_key irq_context: 0 rtnl_mutex _xmit_IPGRE console_lock console_srcu console_owner console_owner_lock irq_context: 0 rcu_read_lock rcu_read_lock rcu_read_lock rcu_node_0 irq_context: 0 rcu_read_lock rcu_read_lock rcu_read_lock &rq->__lock irq_context: 0 rcu_read_lock rcu_read_lock rcu_read_lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 rtnl_mutex &app->lock irq_context: 0 rtnl_mutex (&app->join_timer) irq_context: 0 rtnl_mutex (&app->periodic_timer) irq_context: 0 rtnl_mutex &list->lock#11 irq_context: 0 rtnl_mutex (&app->join_timer)#2 irq_context: 0 rtnl_mutex &app->lock#2 irq_context: 0 rtnl_mutex &list->lock#12 irq_context: 0 rtnl_mutex _xmit_IPGRE &obj_hash[i].lock irq_context: 0 rtnl_mutex _xmit_IPGRE krc.lock irq_context: 0 &f->f_pos_lock &type->i_mutex_dir_key#5 rcu_read_lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 &f->f_pos_lock &type->i_mutex_dir_key#5 rcu_read_lock &rcu_state.gp_wq irq_context: 0 &f->f_pos_lock &type->i_mutex_dir_key#5 rcu_read_lock &rcu_state.gp_wq &p->pi_lock irq_context: 0 &f->f_pos_lock &type->i_mutex_dir_key#5 rcu_read_lock &rcu_state.gp_wq &p->pi_lock &rq->__lock irq_context: 0 &f->f_pos_lock &type->i_mutex_dir_key#5 rcu_read_lock &rcu_state.gp_wq &p->pi_lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 sk_lock-AF_INET fs_reclaim mmu_notifier_invalidate_range_start &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 rtnl_mutex &vlan_netdev_addr_lock_key/1 &obj_hash[i].lock irq_context: 0 rtnl_mutex &vlan_netdev_addr_lock_key/1 krc.lock irq_context: 0 rtnl_mutex &idev->mc_lock &vlan_netdev_addr_lock_key/1 irq_context: 0 rtnl_mutex &idev->mc_lock &vlan_netdev_addr_lock_key/1 &obj_hash[i].lock irq_context: 0 rtnl_mutex &idev->mc_lock &vlan_netdev_addr_lock_key/1 pool_lock#2 irq_context: 0 rtnl_mutex &idev->mc_lock &vlan_netdev_addr_lock_key/1 krc.lock irq_context: 0 &p->lock fs_reclaim &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)ipv6_addrconf (work_completion)(&(&ifa->dad_work)->work) rtnl_mutex krc.lock irq_context: 0 krc.lock &obj_hash[i].lock irq_context: 0 krc.lock &base->lock irq_context: 0 krc.lock &base->lock &obj_hash[i].lock irq_context: 0 &sbi->s_writepages_rwsem jbd2_handle &mapping->private_lock irq_context: 0 &sbi->s_writepages_rwsem jbd2_handle &mapping->private_lock rcu_read_lock &memcg->move_lock irq_context: 0 pernet_ops_rwsem k-sk_lock-AF_TIPC &tn->nametbl_lock &n->list_lock irq_context: 0 pernet_ops_rwsem k-sk_lock-AF_TIPC &tn->nametbl_lock &n->list_lock &c->lock irq_context: 0 &f->f_pos_lock sb_writers#4 &sb->s_type->i_mutex_key#8 jbd2_handle &cfs_rq->removed.lock irq_context: 0 rtnl_mutex &block->lock &____s->seqcount#2 irq_context: 0 rtnl_mutex &block->lock &____s->seqcount irq_context: 0 (wq_completion)afs (work_completion)(&net->cells_manager) &rq->__lock irq_context: 0 (wq_completion)afs (work_completion)(&net->cells_manager) &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 &sbi->s_writepages_rwsem jbd2_handle &ei->i_data_sem rcu_read_lock rcu_node_0 irq_context: 0 &sbi->s_writepages_rwsem jbd2_handle &ei->i_data_sem rcu_read_lock &rq->__lock irq_context: 0 &sig->cred_guard_mutex rcu_read_lock rcu_read_lock rcu_node_0 irq_context: 0 &sig->cred_guard_mutex rcu_read_lock rcu_read_lock &rq->__lock irq_context: 0 &sig->cred_guard_mutex rcu_read_lock rcu_read_lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 &sbi->s_writepages_rwsem jbd2_handle &ei->i_data_sem rcu_read_lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 &sbi->s_writepages_rwsem jbd2_handle &ei->i_data_sem rcu_read_lock &rcu_state.gp_wq irq_context: 0 &sbi->s_writepages_rwsem jbd2_handle &ei->i_data_sem rcu_read_lock &rcu_state.gp_wq &p->pi_lock irq_context: 0 &sbi->s_writepages_rwsem jbd2_handle &ei->i_data_sem rcu_read_lock &rcu_state.gp_wq &p->pi_lock &rq->__lock irq_context: 0 &sbi->s_writepages_rwsem jbd2_handle &ei->i_data_sem rcu_read_lock &rcu_state.gp_wq &p->pi_lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 &sbi->s_writepages_rwsem jbd2_handle &ei->i_data_sem rcu_read_lock &rcu_state.expedited_wq irq_context: 0 &sbi->s_writepages_rwsem jbd2_handle &ei->i_data_sem rcu_read_lock &rcu_state.expedited_wq &p->pi_lock irq_context: 0 &sbi->s_writepages_rwsem jbd2_handle &ei->i_data_sem rcu_read_lock &rcu_state.expedited_wq &p->pi_lock &rq->__lock irq_context: 0 &sbi->s_writepages_rwsem jbd2_handle &ei->i_data_sem rcu_read_lock &rcu_state.expedited_wq &p->pi_lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 pernet_ops_rwsem &rcu_state.expedited_wq &p->pi_lock irq_context: 0 sb_writers &type->i_mutex_dir_key#2 rcu_read_lock rcu_node_0 irq_context: 0 sb_writers &type->i_mutex_dir_key#2 rcu_read_lock &rcu_state.expedited_wq irq_context: 0 sb_writers &type->i_mutex_dir_key#2 rcu_read_lock &rcu_state.expedited_wq &p->pi_lock irq_context: 0 sb_writers &type->i_mutex_dir_key#2 rcu_read_lock &rcu_state.expedited_wq &p->pi_lock &rq->__lock irq_context: 0 sb_writers &type->i_mutex_dir_key#2 rcu_read_lock &rcu_state.expedited_wq &p->pi_lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 sb_writers &type->i_mutex_dir_key#2 rcu_read_lock &rq->__lock irq_context: 0 sb_writers &type->i_mutex_dir_key#2 rcu_read_lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: softirq &x->wait#26 &p->pi_lock &rq->__lock &cfs_rq->removed.lock irq_context: softirq (&s->timer) irq_context: softirq (&s->timer) rfcomm_wq.lock irq_context: softirq (&s->timer) rfcomm_wq.lock &p->pi_lock irq_context: softirq (&s->timer) rfcomm_wq.lock &p->pi_lock &rq->__lock irq_context: softirq (&s->timer) rfcomm_wq.lock &p->pi_lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 &sig->cred_guard_mutex &sig->exec_update_lock &mm->mmap_lock ptlock_ptr(page)#2 rcu_read_lock &p->pi_lock irq_context: 0 &sig->cred_guard_mutex &sig->exec_update_lock &mm->mmap_lock ptlock_ptr(page)#2 rcu_read_lock &p->pi_lock &rq->__lock irq_context: 0 &sig->cred_guard_mutex &sig->exec_update_lock &mm->mmap_lock ptlock_ptr(page)#2 rcu_read_lock &p->pi_lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 &sig->cred_guard_mutex &sig->exec_update_lock &mm->mmap_lock rcu_node_0 irq_context: 0 &sig->cred_guard_mutex &sig->exec_update_lock &mm->mmap_lock &rcu_state.expedited_wq irq_context: 0 &sig->cred_guard_mutex &sig->exec_update_lock &mm->mmap_lock &rcu_state.expedited_wq &p->pi_lock irq_context: 0 &sig->cred_guard_mutex &sig->exec_update_lock &mm->mmap_lock &rcu_state.expedited_wq &p->pi_lock &rq->__lock irq_context: 0 &sig->cred_guard_mutex &sig->exec_update_lock &mm->mmap_lock &rcu_state.expedited_wq &p->pi_lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 dup_mmap_sem &mm->mmap_lock &mm->mmap_lock/1 rcu_read_lock &rq->__lock &cfs_rq->removed.lock irq_context: 0 rtnl_mutex &idev->mc_lock rcu_read_lock &rq->__lock irq_context: 0 rtnl_mutex &idev->mc_lock rcu_read_lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 &f->f_pos_lock sb_writers#4 &sb->s_type->i_mutex_key#8 &ei->i_data_sem &rcu_state.expedited_wq irq_context: 0 pernet_ops_rwsem rtnl_mutex krc.lock &obj_hash[i].lock irq_context: 0 pernet_ops_rwsem rtnl_mutex krc.lock &base->lock irq_context: 0 pernet_ops_rwsem rtnl_mutex krc.lock &base->lock &obj_hash[i].lock irq_context: 0 pernet_ops_rwsem rtnl_mutex _xmit_SIT irq_context: 0 pernet_ops_rwsem rtnl_mutex &root->kernfs_rwsem &____s->seqcount irq_context: 0 &sig->cred_guard_mutex tomoyo_ss rcu_read_lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)events (work_completion)(&(&nsim_dev->trap_data->trap_report_dw)->work) &devlink->lock_key#3 &rq->__lock &cfs_rq->removed.lock irq_context: 0 sb_writers#4 &sb->s_type->i_mutex_key#8 mapping.invalidate_lock &cfs_rq->removed.lock irq_context: 0 sb_writers#4 &sb->s_type->i_mutex_key#8 mapping.invalidate_lock rcu_read_lock &rq->__lock &cfs_rq->removed.lock irq_context: 0 &sig->cred_guard_mutex &sig->exec_update_lock &mm->mmap_lock rcu_read_lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 rtnl_mutex &idev->mc_lock &vlan_netdev_addr_lock_key/1 &obj_hash[i].lock pool_lock irq_context: 0 kn->active#5 rcu_read_lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 &sbi->s_writepages_rwsem jbd2_handle &ei->i_data_sem &ei->i_es_lock key#2 irq_context: 0 sb_writers#11 &type->i_mutex_dir_key#7/1 &c->lock irq_context: 0 sb_writers#11 &type->i_mutex_dir_key#7/1 cgroup_mutex fs_reclaim mmu_notifier_invalidate_range_start &rq->__lock irq_context: 0 sb_writers#11 &type->i_mutex_dir_key#7/1 cgroup_mutex fs_reclaim mmu_notifier_invalidate_range_start &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 sb_writers#11 &rq->__lock irq_context: 0 sb_writers#11 &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 pernet_ops_rwsem rtnl_mutex rcu_state.exp_mutex &obj_hash[i].lock pool_lock irq_context: 0 pernet_ops_rwsem rtnl_mutex &root->kernfs_rwsem rcu_read_lock rcu_node_0 irq_context: 0 &sbi->s_writepages_rwsem jbd2_handle &mapping->i_mmap_rwsem &rq->__lock irq_context: 0 tracepoints_mutex rcu_state.exp_mutex rcu_read_lock &pool->lock &p->pi_lock &rq->__lock irq_context: 0 tracepoints_mutex rcu_state.exp_mutex rcu_read_lock &pool->lock &p->pi_lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 tracepoints_mutex rcu_state.exp_mutex &rnp->exp_wq[3] irq_context: 0 tracepoints_mutex rcu_state.exp_mutex.wait_lock irq_context: 0 tracepoints_mutex &p->pi_lock irq_context: 0 tracepoints_mutex &p->pi_lock &rq->__lock irq_context: 0 tracepoints_mutex &p->pi_lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 tracepoints_mutex &ACCESS_PRIVATE(sdp, lock) irq_context: 0 tracepoints_mutex tracepoint_srcu irq_context: 0 tracepoints_mutex &x->wait#2 irq_context: 0 &f->f_pos_lock &type->i_mutex_dir_key#3 sb_writers#4 jbd2_handle rcu_node_0 irq_context: 0 &f->f_pos_lock &type->i_mutex_dir_key#3 sb_writers#4 jbd2_handle &rcu_state.expedited_wq irq_context: 0 &f->f_pos_lock &type->i_mutex_dir_key#3 sb_writers#4 jbd2_handle &rcu_state.expedited_wq &p->pi_lock irq_context: 0 &f->f_pos_lock &type->i_mutex_dir_key#3 sb_writers#4 jbd2_handle &rcu_state.expedited_wq &p->pi_lock &rq->__lock irq_context: 0 &f->f_pos_lock &type->i_mutex_dir_key#3 sb_writers#4 jbd2_handle &rcu_state.expedited_wq &p->pi_lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)wg-kex-wg2#9 (work_completion)(&peer->transmit_handshake_work) &wg->static_identity.lock &handshake->lock rcu_read_lock_bh batched_entropy_u32.lock crngs.lock irq_context: 0 (wq_completion)wg-kex-wg2#9 (work_completion)(&peer->transmit_handshake_work) &____s->seqcount#2 irq_context: 0 (wq_completion)wg-kex-wg2#9 (work_completion)(&peer->transmit_handshake_work) &____s->seqcount irq_context: 0 (wq_completion)wg-kex-wg2#9 (work_completion)(&peer->transmit_handshake_work) pool_lock#2 irq_context: 0 &mm->mmap_lock &rcu_state.expedited_wq &p->pi_lock irq_context: 0 pernet_ops_rwsem rtnl_mutex &root->kernfs_rwsem &sem->wait_lock irq_context: 0 pernet_ops_rwsem rtnl_mutex dev_pm_qos_sysfs_mtx &root->kernfs_rwsem &sem->wait_lock irq_context: 0 pernet_ops_rwsem rtnl_mutex dev_pm_qos_sysfs_mtx &sem->wait_lock irq_context: 0 pernet_ops_rwsem rtnl_mutex dev_pm_qos_sysfs_mtx &p->pi_lock irq_context: 0 pernet_ops_rwsem uevent_sock_mutex uevent_sock_mutex.wait_lock irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem rtnl_mutex &pnettable->lock &rq->__lock irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem rtnl_mutex &pnettable->lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem rtnl_mutex &root->kernfs_rwsem &p->pi_lock irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem rtnl_mutex &root->kernfs_rwsem &p->pi_lock &rq->__lock irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem rtnl_mutex &root->kernfs_rwsem &p->pi_lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 sb_writers#4 jbd2_handle rcu_read_lock rcu_node_0 irq_context: 0 sb_writers#4 jbd2_handle rcu_read_lock &rq->__lock irq_context: 0 sb_writers#4 jbd2_handle rcu_read_lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 cb_lock rcu_read_lock &____s->seqcount#2 irq_context: 0 cb_lock rcu_read_lock rcu_read_lock rcu_node_0 irq_context: 0 cb_lock rcu_read_lock rcu_read_lock &rq->__lock irq_context: 0 cb_lock rcu_read_lock rcu_read_lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 sb_writers#4 jbd2_handle &rq->__lock irq_context: 0 sb_writers#4 jbd2_handle &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 pernet_ops_rwsem rtnl_mutex remove_cache_srcu rcu_read_lock &____s->seqcount irq_context: 0 pernet_ops_rwsem rtnl_mutex uevent_sock_mutex batched_entropy_u8.lock irq_context: 0 pernet_ops_rwsem rtnl_mutex uevent_sock_mutex kfence_freelist_lock irq_context: 0 pernet_ops_rwsem rtnl_mutex uevent_sock_mutex &meta->lock irq_context: 0 &sbi->s_writepages_rwsem jbd2_handle &mapping->i_mmap_rwsem &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem rtnl_mutex &caifn->caifdevs.lock rcu_state.exp_mutex &rnp->exp_wq[0] irq_context: 0 sb_writers#4 &type->i_mutex_dir_key#3/1 sb_internal jbd2_handle &sbi->s_orphan_lock rcu_read_lock rcu_node_0 irq_context: 0 sb_writers#4 &type->i_mutex_dir_key#3/1 sb_internal jbd2_handle &sbi->s_orphan_lock rcu_read_lock &rq->__lock irq_context: 0 sb_writers#4 &type->i_mutex_dir_key#3/1 sb_internal jbd2_handle &sbi->s_orphan_lock rcu_read_lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 sb_writers#4 &type->i_mutex_dir_key#3/1 sb_internal jbd2_handle &sbi->s_orphan_lock rcu_read_lock &rcu_state.gp_wq irq_context: 0 sb_writers#4 &type->i_mutex_dir_key#3/1 sb_internal jbd2_handle &sbi->s_orphan_lock rcu_read_lock &rcu_state.gp_wq &p->pi_lock irq_context: 0 sb_writers#4 &type->i_mutex_dir_key#3/1 sb_internal jbd2_handle &sbi->s_orphan_lock rcu_read_lock &rcu_state.gp_wq &p->pi_lock &rq->__lock irq_context: 0 sb_writers#4 &type->i_mutex_dir_key#3/1 sb_internal jbd2_handle &sbi->s_orphan_lock rcu_read_lock &rcu_state.gp_wq &p->pi_lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 sb_writers#4 &type->i_mutex_dir_key#3/1 sb_internal jbd2_handle &sbi->s_orphan_lock rcu_read_lock &rcu_state.expedited_wq irq_context: 0 sb_writers#4 &type->i_mutex_dir_key#3/1 sb_internal jbd2_handle &sbi->s_orphan_lock rcu_read_lock &rcu_state.expedited_wq &p->pi_lock irq_context: 0 sb_writers#4 &type->i_mutex_dir_key#3/1 sb_internal jbd2_handle &sbi->s_orphan_lock rcu_read_lock &rcu_state.expedited_wq &p->pi_lock &rq->__lock irq_context: 0 sb_writers#4 &type->i_mutex_dir_key#3/1 sb_internal jbd2_handle &sbi->s_orphan_lock rcu_read_lock &rcu_state.expedited_wq &p->pi_lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: softirq (&tw->tw_timer) irq_context: softirq (&tw->tw_timer) &hashinfo->ehash_locks[i] irq_context: softirq (&tw->tw_timer) &tcp_hashinfo.bhash[i].lock irq_context: softirq (&tw->tw_timer) &tcp_hashinfo.bhash[i].lock &tcp_hashinfo.bhash2[i].lock irq_context: softirq (&tw->tw_timer) &tcp_hashinfo.bhash[i].lock &tcp_hashinfo.bhash2[i].lock stock_lock irq_context: softirq (&tw->tw_timer) &tcp_hashinfo.bhash[i].lock &tcp_hashinfo.bhash2[i].lock &obj_hash[i].lock irq_context: softirq (&tw->tw_timer) &tcp_hashinfo.bhash[i].lock &tcp_hashinfo.bhash2[i].lock pool_lock#2 irq_context: softirq (&tw->tw_timer) stock_lock irq_context: softirq (&tw->tw_timer) &obj_hash[i].lock irq_context: 0 (wq_completion)bat_events (work_completion)(&(&bat_priv->bla.work)->work) rcu_read_lock &____s->seqcount#2 irq_context: 0 (wq_completion)bat_events (work_completion)(&(&bat_priv->bla.work)->work) rcu_read_lock &____s->seqcount irq_context: 0 (wq_completion)bat_events (work_completion)(&(&bat_priv->bla.work)->work) rcu_read_lock batched_entropy_u8.lock irq_context: 0 (wq_completion)bat_events (work_completion)(&(&bat_priv->bla.work)->work) rcu_read_lock kfence_freelist_lock irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem remove_cache_srcu &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem rcu_state.exp_mutex &obj_hash[i].lock pool_lock irq_context: 0 (wq_completion)netns net_cleanup_work rcu_state.barrier_mutex rcu_read_lock &rq->__lock irq_context: 0 (wq_completion)netns net_cleanup_work rcu_state.barrier_mutex rcu_read_lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 pernet_ops_rwsem rcu_state.exp_mutex &rcu_state.expedited_wq irq_context: 0 pernet_ops_rwsem rcu_state.exp_mutex &rcu_state.expedited_wq &p->pi_lock irq_context: 0 pernet_ops_rwsem rcu_read_lock &pool->lock &p->pi_lock &rq->__lock &cfs_rq->removed.lock irq_context: 0 pernet_ops_rwsem rtnl_mutex &ht->mutex &rq->__lock irq_context: 0 pernet_ops_rwsem rtnl_mutex &ht->mutex &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)netns net_cleanup_work rcu_state.barrier_mutex.wait_lock irq_context: 0 (wq_completion)netns net_cleanup_work &p->pi_lock irq_context: 0 (wq_completion)netns net_cleanup_work &p->pi_lock &rq->__lock irq_context: 0 (wq_completion)netns net_cleanup_work &p->pi_lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)netns net_cleanup_work &rq->__lock irq_context: 0 pernet_ops_rwsem &rcu_state.expedited_wq &p->pi_lock &rq->__lock irq_context: 0 hashlimit_mutex &pcp->lock &zone->lock irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem &pool->lock/1 rcu_read_lock &pool->lock &p->pi_lock &rq->__lock &cfs_rq->removed.lock irq_context: 0 &f->f_pos_lock sb_writers#4 &sb->s_type->i_mutex_key#8 &obj_hash[i].lock pool_lock irq_context: 0 sb_writers#4 &sb->s_type->i_mutex_key#8 jbd2_handle &ei->i_data_sem &wb->list_lock irq_context: 0 (wq_completion)events (work_completion)(&pwq->unbound_release_work) rcu_state.exp_mutex rcu_read_lock &rq->__lock irq_context: 0 (wq_completion)events (work_completion)(&pwq->unbound_release_work) rcu_state.exp_mutex rcu_read_lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 sb_writers#4 &sb->s_type->i_mutex_key#8 jbd2_handle &ei->i_data_sem &wb->list_lock &sb->s_type->i_lock_key#22 irq_context: 0 sb_writers#4 &sb->s_type->i_mutex_key#8 jbd2_handle &ei->i_data_sem mmu_notifier_invalidate_range_start &rq->__lock irq_context: 0 sb_writers#4 &sb->s_type->i_mutex_key#8 jbd2_handle &ei->i_data_sem mmu_notifier_invalidate_range_start &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 &mm->mmap_lock sb_pagefaults &journal->j_state_lock &journal->j_wait_transaction_locked irq_context: 0 &mm->mmap_lock sb_pagefaults jbd2_handle irq_context: 0 sb_writers#4 &sb->s_type->i_mutex_key#8 jbd2_handle &journal->j_wait_updates &p->pi_lock irq_context: 0 sb_writers#4 &sb->s_type->i_mutex_key#8 jbd2_handle &journal->j_wait_updates &p->pi_lock &rq->__lock irq_context: 0 sb_writers#4 &sb->s_type->i_mutex_key#8 jbd2_handle &journal->j_wait_updates &p->pi_lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 sb_writers#4 &sb->s_type->i_mutex_key#8 jbd2_handle &ei->i_data_sem bit_wait_table + i irq_context: 0 sb_writers#4 &journal->j_state_lock &journal->j_wait_transaction_locked irq_context: 0 sb_writers#4 jbd2_handle irq_context: 0 sb_writers#4 jbd2_handle bit_wait_table + i irq_context: 0 &sb->s_type->i_mutex_key#8 remove_cache_srcu irq_context: 0 &sb->s_type->i_mutex_key#8 remove_cache_srcu quarantine_lock irq_context: 0 &sb->s_type->i_mutex_key#8 remove_cache_srcu &c->lock irq_context: 0 &sb->s_type->i_mutex_key#8 remove_cache_srcu &n->list_lock irq_context: 0 &sb->s_type->i_mutex_key#8 remove_cache_srcu rcu_read_lock pool_lock#2 irq_context: 0 &sb->s_type->i_mutex_key#8 remove_cache_srcu &obj_hash[i].lock irq_context: 0 &sb->s_type->i_mutex_key#8 remove_cache_srcu &rq->__lock irq_context: 0 &sb->s_type->i_mutex_key#8 remove_cache_srcu &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: softirq (&hsr->announce_timer) rcu_read_lock &____s->seqcount#2 irq_context: 0 sb_writers#4 &sb->s_type->i_mutex_key#8 jbd2_handle &ei->i_data_sem &xa->xa_lock#9 stock_lock irq_context: 0 sb_writers#4 &sb->s_type->i_mutex_key#8 jbd2_handle &ei->i_data_sem &xa->xa_lock#9 &c->lock irq_context: 0 sb_writers#4 &sb->s_type->i_mutex_key#8 jbd2_handle &ei->i_data_sem &xa->xa_lock#9 pool_lock#2 irq_context: 0 (wq_completion)events (work_completion)(&(&nsim_dev->trap_data->trap_report_dw)->work) &devlink->lock_key#2 &nsim_trap_data->trap_lock batched_entropy_u8.lock crngs.lock irq_context: 0 &pipe->mutex/1 sk_lock-AF_NETLINK irq_context: 0 &pipe->mutex/1 sk_lock-AF_NETLINK slock-AF_NETLINK irq_context: 0 &pipe->mutex/1 sk_lock-AF_NETLINK rcu_read_lock rhashtable_bucket irq_context: 0 &pipe->mutex/1 slock-AF_NETLINK irq_context: 0 &pipe->mutex/1 free_vmap_area_lock irq_context: 0 &pipe->mutex/1 vmap_area_lock irq_context: 0 &pipe->mutex/1 init_mm.page_table_lock irq_context: 0 &pipe->mutex/1 &c->lock irq_context: 0 &pipe->mutex/1 rcu_read_lock &zone->lock irq_context: 0 &pipe->mutex/1 rcu_read_lock &____s->seqcount irq_context: 0 &pipe->mutex/1 rcu_read_lock rcu_read_lock_bh rcu_read_lock pool_lock#2 irq_context: 0 &pipe->mutex/1 rcu_read_lock rcu_read_lock_bh &obj_hash[i].lock irq_context: 0 &pipe->mutex/1 rcu_read_lock rcu_read_lock_bh &zone->lock irq_context: 0 &pipe->mutex/1 rcu_read_lock rcu_read_lock_bh &zone->lock &____s->seqcount irq_context: 0 &pipe->mutex/1 rcu_read_lock rcu_read_lock_bh pool_lock#2 irq_context: 0 &pipe->mutex/1 nfnl_subsys_ctnetlink irq_context: 0 &pipe->mutex/1 nfnl_subsys_ctnetlink rcu_read_lock &____s->seqcount#7 irq_context: 0 &pipe->mutex/1 nfnl_subsys_ctnetlink pool_lock#2 irq_context: 0 &pipe->mutex/1 nfnl_subsys_ctnetlink rcu_read_lock pool_lock#2 irq_context: 0 &pipe->mutex/1 nfnl_subsys_ctnetlink rcu_read_lock &ct->lock irq_context: 0 &pipe->mutex/1 nfnl_subsys_ctnetlink rcu_read_lock &nf_conntrack_locks[i] irq_context: 0 &pipe->mutex/1 nfnl_subsys_ctnetlink rcu_read_lock &nf_conntrack_locks[i] &nf_conntrack_locks[i]/1 irq_context: 0 &pipe->mutex/1 nfnl_subsys_ctnetlink rcu_read_lock &nf_conntrack_locks[i] &nf_conntrack_locks[i]/1 batched_entropy_u8.lock irq_context: 0 &pipe->mutex/1 nfnl_subsys_ctnetlink rcu_read_lock &nf_conntrack_locks[i]/1 irq_context: 0 &pipe->mutex/1 nfnl_subsys_ctnetlink rcu_read_lock &obj_hash[i].lock irq_context: 0 &pipe->mutex/1 nfnl_subsys_ctnetlink rcu_read_lock rcu_node_0 irq_context: 0 &pipe->mutex/1 nfnl_subsys_ctnetlink rcu_read_lock &rq->__lock irq_context: 0 &pipe->mutex/1 nfnl_subsys_ctnetlink rcu_read_lock &c->lock irq_context: 0 &pipe->mutex/1 nfnl_subsys_ctnetlink rcu_read_lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 &pipe->mutex/1 nfnl_subsys_ctnetlink rcu_read_lock &rcu_state.expedited_wq irq_context: 0 &pipe->mutex/1 nfnl_subsys_ctnetlink rcu_read_lock &rcu_state.expedited_wq &p->pi_lock irq_context: 0 &pipe->mutex/1 nfnl_subsys_ctnetlink rcu_read_lock &rcu_state.expedited_wq &p->pi_lock &rq->__lock irq_context: 0 &pipe->mutex/1 nfnl_subsys_ctnetlink rcu_read_lock &rcu_state.expedited_wq &p->pi_lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 sb_writers#4 &sb->s_type->i_mutex_key#8 mapping.invalidate_lock rcu_read_lock &rcu_state.expedited_wq irq_context: 0 sb_writers#4 &sb->s_type->i_mutex_key#8 mapping.invalidate_lock rcu_read_lock &rcu_state.expedited_wq &p->pi_lock irq_context: 0 sb_writers#4 &sb->s_type->i_mutex_key#8 mapping.invalidate_lock rcu_read_lock &rcu_state.expedited_wq &p->pi_lock &rq->__lock irq_context: 0 sb_writers#4 &sb->s_type->i_mutex_key#8 mapping.invalidate_lock rcu_read_lock &rcu_state.expedited_wq &p->pi_lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 &pipe->mutex/1 nfnl_subsys_ctnetlink rcu_read_lock &n->list_lock irq_context: 0 &pipe->mutex/1 nfnl_subsys_ctnetlink rcu_read_lock &n->list_lock &c->lock irq_context: 0 sb_writers#4 &sb->s_type->i_mutex_key#8 mapping.invalidate_lock jbd2_handle &ei->i_data_sem &journal->j_state_lock irq_context: 0 sb_writers#4 &sb->s_type->i_mutex_key#8 mapping.invalidate_lock jbd2_handle &sbi->s_orphan_lock &ei->i_raw_lock irq_context: 0 &pipe->mutex/1 nfnl_subsys_ctnetlink rcu_read_lock &nf_conntrack_locks[i] &nf_conntrack_locks[i]/1 batched_entropy_u8.lock crngs.lock irq_context: 0 &pipe->mutex/1 nfnl_subsys_ctnetlink &rq->__lock irq_context: 0 &pipe->mutex/1 nfnl_subsys_ctnetlink &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 &pipe->mutex/1 nfnl_subsys_ctnetlink rcu_node_0 irq_context: softirq rcu_read_lock rcu_read_lock &zone->lock &____s->seqcount irq_context: 0 sb_writers#4 &sb->s_type->i_mutex_key#8 remove_cache_srcu pool_lock#2 irq_context: 0 sb_writers#4 &sb->s_type->i_mutex_key#8 remove_cache_srcu &rq->__lock irq_context: 0 sb_writers#4 &sb->s_type->i_mutex_key#8 remove_cache_srcu &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 &po->bind_lock irq_context: 0 &po->bind_lock ptype_lock irq_context: 0 clock-AF_PACKET irq_context: 0 elock-AF_PACKET irq_context: 0 &pipe->mutex/1 nfnl_subsys_ctnetlink rcu_read_lock &rcu_state.gp_wq irq_context: 0 &pipe->mutex/1 nfnl_subsys_ctnetlink rcu_read_lock &rcu_state.gp_wq &p->pi_lock irq_context: 0 &pipe->mutex/1 nfnl_subsys_ctnetlink rcu_read_lock &rcu_state.gp_wq &p->pi_lock &rq->__lock irq_context: 0 &pipe->mutex/1 nfnl_subsys_ctnetlink rcu_read_lock &rcu_state.gp_wq &p->pi_lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 &pipe->mutex/1 nfnl_subsys_ctnetlink batched_entropy_u8.lock irq_context: 0 &pipe->mutex/1 nfnl_subsys_ctnetlink kfence_freelist_lock irq_context: 0 &pipe->mutex/1 purge_vmap_area_lock irq_context: 0 sb_writers#4 &sb->s_type->i_mutex_key#8 jbd2_handle &ei->i_data_sem &c->lock irq_context: 0 sb_writers#4 &sb->s_type->i_mutex_key#8 &journal->j_state_lock &journal->j_wait_transaction_locked irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem rtnl_mutex &rnp->exp_wq[1] irq_context: 0 rtnl_mutex &br->lock &br->multicast_lock &c->lock irq_context: 0 rtnl_mutex &br->lock &br->multicast_lock &____s->seqcount#2 irq_context: 0 rtnl_mutex &br->lock &br->multicast_lock &____s->seqcount irq_context: 0 sb_writers#4 &sb->s_type->i_mutex_key#8 jbd2_handle &ei->i_data_sem quarantine_lock irq_context: 0 sb_writers#4 &sb->s_type->i_mutex_key#8 remove_cache_srcu rcu_read_lock rcu_node_0 irq_context: 0 sb_writers#4 &sb->s_type->i_mutex_key#8 remove_cache_srcu rcu_read_lock &rq->__lock irq_context: 0 sb_writers#4 &sb->s_type->i_mutex_key#8 remove_cache_srcu rcu_read_lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 rtnl_mutex (&br->hello_timer) irq_context: 0 rtnl_mutex (&br->topology_change_timer) irq_context: 0 rtnl_mutex (&br->tcn_timer) irq_context: 0 rtnl_mutex &bridge_netdev_addr_lock_key/1 &obj_hash[i].lock irq_context: 0 rtnl_mutex &bridge_netdev_addr_lock_key/1 krc.lock irq_context: 0 rtnl_mutex &im->lock &obj_hash[i].lock irq_context: 0 rtnl_mutex &idev->mc_lock &bridge_netdev_addr_lock_key/1 &obj_hash[i].lock irq_context: 0 rtnl_mutex &idev->mc_lock &bridge_netdev_addr_lock_key/1 krc.lock irq_context: 0 rtnl_mutex rcu_read_lock rcu_read_lock pool_lock#2 irq_context: 0 (wq_completion)events (work_completion)(&data->fib_event_work) &rq->__lock irq_context: 0 (wq_completion)events (work_completion)(&data->fib_event_work) &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: softirq rcu_callback &c->lock irq_context: softirq rcu_callback &n->list_lock irq_context: softirq rcu_callback &n->list_lock &c->lock irq_context: 0 sb_writers#4 sb_internal jbd2_handle mmu_notifier_invalidate_range_start irq_context: 0 sb_writers#4 sb_internal jbd2_handle &____s->seqcount irq_context: 0 &sbi->s_writepages_rwsem &journal->j_state_lock irq_context: 0 &sbi->s_writepages_rwsem &journal->j_state_lock tk_core.seq.seqcount irq_context: 0 &sbi->s_writepages_rwsem &journal->j_state_lock &obj_hash[i].lock irq_context: 0 &sbi->s_writepages_rwsem &journal->j_state_lock &base->lock irq_context: 0 &sbi->s_writepages_rwsem &journal->j_state_lock &base->lock &obj_hash[i].lock irq_context: 0 &journal->j_wait_done_commit &p->pi_lock &rq->__lock &cfs_rq->removed.lock irq_context: 0 (wq_completion)ext4-rsv-conversion (work_completion)(&ei->i_rsv_conversion_work) &c->lock irq_context: 0 sb_writers#4 &type->i_mutex_dir_key#3/1 jbd2_handle &rq->__lock &cfs_rq->removed.lock irq_context: 0 (wq_completion)kblockd (work_completion)(&(&q->requeue_work)->work) rcu_read_lock rcu_node_0 irq_context: 0 (wq_completion)kblockd (work_completion)(&(&q->requeue_work)->work) rcu_read_lock &rq->__lock irq_context: 0 (wq_completion)kblockd (work_completion)(&(&q->requeue_work)->work) rcu_read_lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 sb_writers#4 &sb->s_type->i_mutex_key#8 rcu_node_0 irq_context: 0 sb_writers#4 mapping.invalidate_lock rcu_read_lock batched_entropy_u8.lock irq_context: 0 sb_writers#4 mapping.invalidate_lock rcu_read_lock kfence_freelist_lock irq_context: 0 sb_writers#4 bit_wait_table + i irq_context: 0 &f->f_pos_lock &type->i_mutex_dir_key#3 sb_writers#4 jbd2_handle &____s->seqcount irq_context: 0 sb_writers#4 &type->i_mutex_dir_key#3/1 &type->i_mutex_dir_key#3 jbd2_handle bit_wait_table + i irq_context: 0 (wq_completion)events (work_completion)(&(&nsim_dev->trap_data->trap_report_dw)->work) &devlink->lock_key#4 rcu_node_0 irq_context: 0 &pipe->mutex/1 &pipe->wr_wait &p->pi_lock irq_context: 0 &pipe->mutex/1 &pipe->wr_wait &p->pi_lock &rq->__lock irq_context: 0 &pipe->mutex/1 &pipe->wr_wait &p->pi_lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 pernet_ops_rwsem rtnl_mutex uevent_sock_mutex &rq->__lock &cfs_rq->removed.lock irq_context: 0 &pipe->mutex/1 rcu_read_lock &sighand->siglock irq_context: 0 &pipe->mutex/1 &sighand->siglock irq_context: 0 pernet_ops_rwsem rtnl_mutex uevent_sock_mutex nl_table_wait.lock &p->pi_lock irq_context: 0 pernet_ops_rwsem rtnl_mutex uevent_sock_mutex nl_table_wait.lock &p->pi_lock &cfs_rq->removed.lock irq_context: 0 pernet_ops_rwsem rtnl_mutex uevent_sock_mutex nl_table_wait.lock &p->pi_lock &rq->__lock irq_context: 0 pernet_ops_rwsem rtnl_mutex uevent_sock_mutex nl_table_wait.lock &p->pi_lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 &f->f_pos_lock sb_writers#4 &p->pi_lock &rq->__lock &cfs_rq->removed.lock irq_context: 0 (wq_completion)wg-crypt-wg1#3 (work_completion)(&({ do { const void *__vpp_verify = (typeof((worker) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __asm__ ("" : "=r"(__ptr) : "0"((typeof(*((worker))) *)((worker)))); (typeof((typeof(*((worker))) *)((worker)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); })->work) rcu_read_lock rcu_node_0 irq_context: 0 (wq_completion)wg-crypt-wg1#3 (work_completion)(&({ do { const void *__vpp_verify = (typeof((worker) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __asm__ ("" : "=r"(__ptr) : "0"((typeof(*((worker))) *)((worker)))); (typeof((typeof(*((worker))) *)((worker)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); })->work) rcu_read_lock &rq->__lock irq_context: 0 &mm->mmap_lock sb_pagefaults batched_entropy_u8.lock irq_context: 0 &mm->mmap_lock sb_pagefaults kfence_freelist_lock irq_context: 0 &mm->mmap_lock sb_pagefaults &meta->lock irq_context: 0 pernet_ops_rwsem rtnl_mutex &caifn->caifdevs.lock rcu_state.exp_mutex pgd_lock irq_context: 0 pernet_ops_rwsem rtnl_mutex &caifn->caifdevs.lock rcu_state.exp_mutex stock_lock irq_context: 0 pernet_ops_rwsem rtnl_mutex &caifn->caifdevs.lock rcu_state.exp_mutex rcu_read_lock pool_lock#2 irq_context: 0 pernet_ops_rwsem rtnl_mutex &caifn->caifdevs.lock rcu_state.exp_mutex key irq_context: 0 pernet_ops_rwsem rtnl_mutex &caifn->caifdevs.lock rcu_state.exp_mutex pcpu_lock irq_context: 0 pernet_ops_rwsem rtnl_mutex &caifn->caifdevs.lock rcu_state.exp_mutex percpu_counters_lock irq_context: 0 pernet_ops_rwsem rtnl_mutex &caifn->caifdevs.lock rcu_state.exp_mutex pcpu_lock stock_lock irq_context: 0 pernet_ops_rwsem rtnl_mutex &caifn->caifdevs.lock rcu_state.exp_mutex pool_lock#2 irq_context: 0 (wq_completion)events_power_efficient (work_completion)(&barr->work) &x->wait#10 &p->pi_lock &rq->__lock irq_context: 0 (wq_completion)events_power_efficient (work_completion)(&barr->work) &x->wait#10 &p->pi_lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 sb_writers#4 sb_internal jbd2_handle &ei->i_data_sem &rq->__lock irq_context: 0 sb_writers#4 sb_internal jbd2_handle &ei->i_data_sem &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: softirq (&peer->timer_send_keepalive) batched_entropy_u8.lock irq_context: softirq (&peer->timer_send_keepalive) batched_entropy_u8.lock crngs.lock irq_context: softirq (&peer->timer_send_keepalive) kfence_freelist_lock irq_context: 0 pernet_ops_rwsem &net->ipv6.ip6addrlbl_table.lock krc.lock &obj_hash[i].lock irq_context: 0 pernet_ops_rwsem &net->ipv6.ip6addrlbl_table.lock krc.lock &base->lock irq_context: 0 pernet_ops_rwsem &net->ipv6.ip6addrlbl_table.lock krc.lock &base->lock &obj_hash[i].lock irq_context: 0 pernet_ops_rwsem sysctl_lock krc.lock &obj_hash[i].lock irq_context: 0 pernet_ops_rwsem sysctl_lock krc.lock &base->lock irq_context: 0 pernet_ops_rwsem sysctl_lock krc.lock &base->lock &obj_hash[i].lock irq_context: 0 sb_writers#4 &sb->s_type->i_mutex_key#8 mapping.invalidate_lock jbd2_handle &sbi->s_orphan_lock rcu_read_lock rcu_node_0 irq_context: 0 sb_writers#4 &sb->s_type->i_mutex_key#8 mapping.invalidate_lock jbd2_handle &sbi->s_orphan_lock rcu_read_lock &rq->__lock irq_context: 0 sb_writers#4 &sb->s_type->i_mutex_key#8 mapping.invalidate_lock jbd2_handle &cfs_rq->removed.lock irq_context: 0 &sig->cred_guard_mutex &sig->exec_update_lock &mm->mmap_lock lock#4 &lruvec->lru_lock per_cpu_ptr(&cgroup_rstat_cpu_lock, cpu) irq_context: 0 &f->f_pos_lock sb_writers#4 &sb->s_type->i_mutex_key#8 rcu_read_lock &base->lock irq_context: 0 &f->f_pos_lock sb_writers#4 &sb->s_type->i_mutex_key#8 rcu_read_lock &base->lock &obj_hash[i].lock irq_context: 0 bpf_module_mutex irq_context: 0 bpf_module_mutex &rq->__lock irq_context: 0 bpf_module_mutex &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 sb_writers#4 tomoyo_ss &cfs_rq->removed.lock irq_context: 0 sb_writers#4 &sb->s_type->i_mutex_key#8 jbd2_handle &cfs_rq->removed.lock irq_context: 0 sb_writers#4 &sb->s_type->i_mutex_key#8 jbd2_handle &obj_hash[i].lock irq_context: 0 (wq_completion)events_power_efficient (work_completion)(&(&hinfo->gc_work)->work) &rq->__lock &cfs_rq->removed.lock irq_context: 0 (wq_completion)events (work_completion)(&(&nsim_dev->trap_data->trap_report_dw)->work) &devlink->lock_key#2 &nsim_trap_data->trap_lock &base->lock irq_context: 0 (wq_completion)events (work_completion)(&(&nsim_dev->trap_data->trap_report_dw)->work) &devlink->lock_key#2 &nsim_trap_data->trap_lock &base->lock &obj_hash[i].lock irq_context: 0 &f->f_pos_lock sb_writers#4 &sb->s_type->i_mutex_key#8 &ei->i_data_sem &(ei->i_block_reservation_lock) key#3 irq_context: 0 sb_writers#4 &sb->s_type->i_mutex_key#8 jbd2_handle &ei->i_data_sem &ei->i_es_lock &c->lock irq_context: 0 cb_lock quarantine_lock irq_context: 0 cb_lock remove_cache_srcu rcu_read_lock rcu_node_0 irq_context: 0 cb_lock remove_cache_srcu rcu_read_lock &rq->__lock irq_context: 0 cb_lock remove_cache_srcu rcu_read_lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 &pipe->mutex/1 &n->list_lock irq_context: 0 &pipe->mutex/1 &n->list_lock &c->lock irq_context: 0 &pipe->mutex/1 rcu_read_lock &c->lock irq_context: 0 &pipe->mutex/1 nfnl_subsys_ctnetlink &c->lock irq_context: 0 sb_writers#4 &sb->s_type->i_mutex_key#8 key#3 irq_context: 0 sb_writers#4 &sb->s_type->i_mutex_key#8 key#14 irq_context: 0 &f->f_pos_lock sb_writers#4 &sb->s_type->i_mutex_key#8 jbd2_handle &ei->i_data_sem key#3 irq_context: 0 &f->f_pos_lock sb_writers#4 &sb->s_type->i_mutex_key#8 key#3 irq_context: 0 &f->f_pos_lock sb_writers#4 &sb->s_type->i_mutex_key#8 key#14 irq_context: 0 &f->f_pos_lock sb_writers#4 &sb->s_type->i_mutex_key#8 mapping.invalidate_lock irq_context: 0 &f->f_pos_lock sb_writers#4 &sb->s_type->i_mutex_key#8 mapping.invalidate_lock mmu_notifier_invalidate_range_start irq_context: 0 &f->f_pos_lock sb_writers#4 &sb->s_type->i_mutex_key#8 mapping.invalidate_lock pool_lock#2 irq_context: 0 &f->f_pos_lock sb_writers#4 &sb->s_type->i_mutex_key#8 mapping.invalidate_lock &journal->j_state_lock irq_context: 0 &f->f_pos_lock sb_writers#4 &sb->s_type->i_mutex_key#8 mapping.invalidate_lock jbd2_handle irq_context: 0 &f->f_pos_lock sb_writers#4 &sb->s_type->i_mutex_key#8 mapping.invalidate_lock jbd2_handle &ei->i_data_sem irq_context: 0 &f->f_pos_lock sb_writers#4 &sb->s_type->i_mutex_key#8 mapping.invalidate_lock jbd2_handle &ei->i_data_sem &ei->i_prealloc_lock irq_context: 0 &f->f_pos_lock sb_writers#4 &sb->s_type->i_mutex_key#8 mapping.invalidate_lock jbd2_handle &ei->i_data_sem &ei->i_raw_lock irq_context: 0 &f->f_pos_lock sb_writers#4 &sb->s_type->i_mutex_key#8 mapping.invalidate_lock jbd2_handle &ei->i_data_sem &ei->i_es_lock irq_context: 0 &f->f_pos_lock sb_writers#4 &sb->s_type->i_mutex_key#8 mapping.invalidate_lock jbd2_handle &ei->i_data_sem &ei->i_es_lock &sbi->s_es_lock irq_context: 0 &f->f_pos_lock sb_writers#4 &sb->s_type->i_mutex_key#8 mapping.invalidate_lock jbd2_handle &ei->i_data_sem &ei->i_es_lock &obj_hash[i].lock irq_context: 0 &f->f_pos_lock sb_writers#4 &sb->s_type->i_mutex_key#8 mapping.invalidate_lock jbd2_handle &ei->i_data_sem &ei->i_es_lock pool_lock#2 irq_context: 0 &f->f_pos_lock sb_writers#4 &sb->s_type->i_mutex_key#8 mapping.invalidate_lock jbd2_handle &ei->i_data_sem mmu_notifier_invalidate_range_start irq_context: 0 &f->f_pos_lock sb_writers#4 &sb->s_type->i_mutex_key#8 mapping.invalidate_lock jbd2_handle &ei->i_data_sem pool_lock#2 irq_context: 0 &f->f_pos_lock sb_writers#4 &sb->s_type->i_mutex_key#8 mapping.invalidate_lock jbd2_handle &ei->i_data_sem &obj_hash[i].lock irq_context: 0 &f->f_pos_lock sb_writers#4 &sb->s_type->i_mutex_key#8 mapping.invalidate_lock jbd2_handle &sbi->s_orphan_lock irq_context: 0 &f->f_pos_lock sb_writers#4 &sb->s_type->i_mutex_key#8 mapping.invalidate_lock jbd2_handle &ei->i_raw_lock irq_context: 0 &f->f_pos_lock sb_writers#4 &sb->s_type->i_mutex_key#8 mapping.invalidate_lock jbd2_handle tk_core.seq.seqcount irq_context: 0 &f->f_pos_lock sb_writers#4 &sb->s_type->i_mutex_key#8 mapping.invalidate_lock &obj_hash[i].lock irq_context: 0 text_mutex &rq->__lock irq_context: 0 text_mutex &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 sb_writers#4 &type->i_mutex_dir_key#3/1 jbd2_handle &ei->i_data_sem key#3 irq_context: 0 &f->f_pos_lock sb_writers#4 &sb->s_type->i_mutex_key#8 jbd2_handle &sbi->s_orphan_lock &rq->__lock irq_context: 0 &f->f_pos_lock sb_writers#4 &sb->s_type->i_mutex_key#8 jbd2_handle &sbi->s_orphan_lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 sb_writers#4 &sb->s_type->i_mutex_key#8 mapping.invalidate_lock jbd2_handle &journal->j_wait_updates &p->pi_lock irq_context: 0 sb_writers#4 &sb->s_type->i_mutex_key#8 mapping.invalidate_lock jbd2_handle &journal->j_wait_updates &p->pi_lock &cfs_rq->removed.lock irq_context: 0 sb_writers#4 &sb->s_type->i_mutex_key#8 mapping.invalidate_lock jbd2_handle &journal->j_wait_updates &p->pi_lock &rq->__lock irq_context: 0 sb_writers#4 &sb->s_type->i_mutex_key#8 mapping.invalidate_lock jbd2_handle &journal->j_wait_updates &p->pi_lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 sb_writers#4 &sb->s_type->i_mutex_key#8 mapping.invalidate_lock jbd2_handle &sbi->s_orphan_lock &ret->b_state_lock irq_context: 0 &journal->j_wait_done_commit &p->pi_lock &cfs_rq->removed.lock irq_context: 0 &pipe->mutex/1 nfnl_subsys_ctnetlink rcu_read_lock &____s->seqcount#2 irq_context: 0 &pipe->mutex/1 nfnl_subsys_ctnetlink rcu_read_lock &____s->seqcount irq_context: 0 rtnl_mutex &br->lock &br->multicast_lock &n->list_lock irq_context: 0 rtnl_mutex &br->lock &br->multicast_lock &n->list_lock &c->lock irq_context: 0 &rnp->exp_wq[0] irq_context: 0 rtnl_mutex &idev->mc_lock fs_reclaim mmu_notifier_invalidate_range_start &rq->__lock irq_context: 0 rtnl_mutex &idev->mc_lock fs_reclaim mmu_notifier_invalidate_range_start &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 &journal->j_state_lock &journal->j_wait_transaction_locked &p->pi_lock &cfs_rq->removed.lock irq_context: 0 rcu_read_lock (console_sem).lock irq_context: 0 rcu_read_lock kernfs_pr_cont_lock irq_context: 0 rcu_read_lock kernfs_pr_cont_lock kernfs_rename_lock irq_context: 0 rcu_read_lock kernfs_pr_cont_lock (console_sem).lock irq_context: 0 sb_writers#4 &sb->s_type->i_mutex_key#8 &n->list_lock irq_context: 0 sb_writers#4 &sb->s_type->i_mutex_key#8 &n->list_lock &c->lock irq_context: 0 rtnl_mutex rcu_read_lock &tb->tb6_lock &____s->seqcount#2 irq_context: softirq (&tw->tw_timer) &dccp_hashinfo.bhash[i].lock irq_context: softirq (&tw->tw_timer) &dccp_hashinfo.bhash[i].lock &dccp_hashinfo.bhash2[i].lock irq_context: softirq (&tw->tw_timer) &dccp_hashinfo.bhash[i].lock &dccp_hashinfo.bhash2[i].lock stock_lock irq_context: softirq (&tw->tw_timer) &dccp_hashinfo.bhash[i].lock &dccp_hashinfo.bhash2[i].lock &obj_hash[i].lock irq_context: 0 rtnl_mutex &idev->mc_lock rcu_read_lock &pool->lock irq_context: 0 rtnl_mutex &idev->mc_lock rcu_read_lock &pool->lock &obj_hash[i].lock irq_context: 0 (wq_completion)events_power_efficient (work_completion)(&(&tbl->gc_work)->work) &tbl->lock &n->list_lock irq_context: 0 (wq_completion)events_power_efficient (work_completion)(&(&tbl->gc_work)->work) &tbl->lock &n->list_lock &c->lock irq_context: 0 &f->f_pos_lock sb_writers#4 &sb->s_type->i_mutex_key#8 mapping.invalidate_lock &mapping->private_lock irq_context: 0 &f->f_pos_lock sb_writers#4 &sb->s_type->i_mutex_key#8 mapping.invalidate_lock stock_lock irq_context: 0 &f->f_pos_lock sb_writers#4 &sb->s_type->i_mutex_key#8 mapping.invalidate_lock &sb->s_type->i_lock_key#22 irq_context: 0 &f->f_pos_lock sb_writers#4 &sb->s_type->i_mutex_key#8 mapping.invalidate_lock &sb->s_type->i_lock_key#22 &xa->xa_lock#9 irq_context: 0 &f->f_pos_lock sb_writers#4 &sb->s_type->i_mutex_key#8 mapping.invalidate_lock lock#4 irq_context: 0 &f->f_pos_lock sb_writers#4 &sb->s_type->i_mutex_key#8 mapping.invalidate_lock lock#4 &lruvec->lru_lock irq_context: 0 &f->f_pos_lock sb_writers#4 &sb->s_type->i_mutex_key#8 mapping.invalidate_lock lock#5 irq_context: 0 &f->f_pos_lock sb_writers#4 &sb->s_type->i_mutex_key#8 mapping.invalidate_lock &lruvec->lru_lock irq_context: 0 &f->f_pos_lock sb_writers#4 &sb->s_type->i_mutex_key#8 mapping.invalidate_lock rcu_read_lock pool_lock#2 irq_context: 0 &f->f_pos_lock sb_writers#4 &sb->s_type->i_mutex_key#8 mapping.invalidate_lock jbd2_handle &ret->b_state_lock irq_context: 0 &f->f_pos_lock sb_writers#4 &sb->s_type->i_mutex_key#8 mapping.invalidate_lock jbd2_handle &ei->i_data_sem &rq->__lock irq_context: 0 &f->f_pos_lock sb_writers#4 &sb->s_type->i_mutex_key#8 mapping.invalidate_lock jbd2_handle &ei->i_data_sem &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 &f->f_pos_lock sb_writers#4 &sb->s_type->i_mutex_key#8 mapping.invalidate_lock jbd2_handle &journal->j_wait_updates irq_context: 0 &f->f_pos_lock sb_writers#4 &sb->s_type->i_mutex_key#8 mapping.invalidate_lock jbd2_handle &ret->b_state_lock &journal->j_list_lock irq_context: 0 sb_writers#4 &sb->s_type->i_mutex_key#8 mapping.invalidate_lock jbd2_handle &ei->i_data_sem &ei->i_es_lock &____s->seqcount irq_context: 0 sb_writers#4 &sb->s_type->i_mutex_key#8 mapping.invalidate_lock jbd2_handle &ei->i_data_sem &ei->i_es_lock rcu_read_lock pool_lock#2 irq_context: 0 (wq_completion)events_power_efficient (work_completion)(&(&tbl->gc_work)->work) &rq->__lock &cfs_rq->removed.lock irq_context: 0 tracepoints_mutex rcu_state.exp_mutex &rnp->exp_wq[2] irq_context: 0 (wq_completion)events_power_efficient (work_completion)(&(&tbl->gc_work)->work) rcu_node_0 irq_context: 0 (wq_completion)events_power_efficient (work_completion)(&(&tbl->gc_work)->work) &rcu_state.expedited_wq irq_context: 0 (wq_completion)events_power_efficient (work_completion)(&(&tbl->gc_work)->work) &rcu_state.expedited_wq &p->pi_lock irq_context: 0 (wq_completion)events_power_efficient (work_completion)(&(&tbl->gc_work)->work) &rcu_state.expedited_wq &p->pi_lock &rq->__lock irq_context: 0 (wq_completion)events_power_efficient (work_completion)(&(&tbl->gc_work)->work) &rcu_state.expedited_wq &p->pi_lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)events_power_efficient (work_completion)(&(&tbl->gc_work)->work) &tbl->lock batched_entropy_u8.lock irq_context: 0 (wq_completion)events_power_efficient (work_completion)(&(&tbl->gc_work)->work) &tbl->lock kfence_freelist_lock irq_context: 0 (wq_completion)events_power_efficient (work_completion)(&(&tbl->gc_work)->work) &tbl->lock &meta->lock irq_context: 0 &f->f_pos_lock sb_writers#4 &sb->s_type->i_mutex_key#8 mapping.invalidate_lock &journal->j_state_lock irq_context: 0 &f->f_pos_lock sb_writers#4 &sb->s_type->i_mutex_key#8 mapping.invalidate_lock &journal->j_state_lock tk_core.seq.seqcount irq_context: 0 &f->f_pos_lock sb_writers#4 &sb->s_type->i_mutex_key#8 mapping.invalidate_lock &journal->j_state_lock &obj_hash[i].lock irq_context: 0 &f->f_pos_lock sb_writers#4 &sb->s_type->i_mutex_key#8 mapping.invalidate_lock &journal->j_state_lock &base->lock irq_context: 0 &f->f_pos_lock sb_writers#4 &sb->s_type->i_mutex_key#8 mapping.invalidate_lock &journal->j_state_lock &base->lock &obj_hash[i].lock irq_context: 0 &f->f_pos_lock sb_writers#4 &sb->s_type->i_mutex_key#8 mapping.invalidate_lock &rq->__lock irq_context: 0 &f->f_pos_lock sb_writers#4 &sb->s_type->i_mutex_key#8 mapping.invalidate_lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 &mm->mmap_lock &anon_vma->rwsem mmu_notifier_invalidate_range_start &rq->__lock irq_context: 0 &mm->mmap_lock &anon_vma->rwsem mmu_notifier_invalidate_range_start &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 link_idr_lock &obj_hash[i].lock pool_lock irq_context: 0 sk_lock-AF_PACKET rcu_state.exp_mutex &obj_hash[i].lock pool_lock irq_context: 0 (wq_completion)rcu_gp (work_completion)(&sdp->work) &obj_hash[i].lock pool_lock irq_context: 0 sb_writers#4 &type->i_mutex_dir_key#3/1 &sb->s_type->i_mutex_key#8 remove_cache_srcu &pcp->lock &zone->lock irq_context: 0 sb_writers#4 &type->i_mutex_dir_key#3/1 &sb->s_type->i_mutex_key#8 remove_cache_srcu &pcp->lock &zone->lock &____s->seqcount irq_context: 0 hashlimit_mutex free_vmap_area_lock &obj_hash[i].lock irq_context: 0 hashlimit_mutex free_vmap_area_lock pool_lock#2 irq_context: 0 &sb->s_type->i_mutex_key#8 mmu_notifier_invalidate_range_start &rq->__lock irq_context: 0 &sb->s_type->i_mutex_key#8 mmu_notifier_invalidate_range_start &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 sk_lock-AF_INET6 &ei->socket.wq.wait irq_context: 0 sk_lock-AF_INET6 &mm->mmap_lock fs_reclaim irq_context: 0 sk_lock-AF_INET6 &mm->mmap_lock fs_reclaim mmu_notifier_invalidate_range_start irq_context: 0 sk_lock-AF_INET6 &mm->mmap_lock &____s->seqcount irq_context: 0 sk_lock-AF_INET6 &mm->mmap_lock stock_lock irq_context: 0 sk_lock-AF_INET6 &mm->mmap_lock ptlock_ptr(page) irq_context: 0 pernet_ops_rwsem rtnl_mutex &idev->mc_lock remove_cache_srcu irq_context: 0 pernet_ops_rwsem rtnl_mutex &idev->mc_lock remove_cache_srcu quarantine_lock irq_context: 0 pernet_ops_rwsem rtnl_mutex &idev->mc_lock remove_cache_srcu &n->list_lock irq_context: 0 pernet_ops_rwsem rtnl_mutex &idev->mc_lock remove_cache_srcu rcu_read_lock pool_lock#2 irq_context: 0 pernet_ops_rwsem rtnl_mutex &idev->mc_lock remove_cache_srcu &obj_hash[i].lock irq_context: 0 pernet_ops_rwsem rtnl_mutex &idev->mc_lock remove_cache_srcu &c->lock irq_context: 0 pernet_ops_rwsem rtnl_mutex &idev->mc_lock remove_cache_srcu &rq->__lock irq_context: 0 pernet_ops_rwsem rtnl_mutex &idev->mc_lock remove_cache_srcu &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: softirq (&icsk->icsk_delack_timer) slock-AF_INET6 irq_context: 0 &f->f_pos_lock sb_writers#4 &sb->s_type->i_mutex_key#8 &type->s_umount_key#32 rcu_read_lock pool_lock#2 irq_context: 0 &f->f_pos_lock sb_writers#4 &sb->s_type->i_mutex_key#8 &type->s_umount_key#32 rcu_read_lock &wb->work_lock irq_context: 0 &f->f_pos_lock sb_writers#4 &sb->s_type->i_mutex_key#8 &type->s_umount_key#32 rcu_read_lock &wb->work_lock &obj_hash[i].lock irq_context: 0 &f->f_pos_lock sb_writers#4 &sb->s_type->i_mutex_key#8 &type->s_umount_key#32 rcu_read_lock &wb->work_lock &base->lock irq_context: 0 &f->f_pos_lock sb_writers#4 &sb->s_type->i_mutex_key#8 &type->s_umount_key#32 rcu_read_lock &wb->work_lock &base->lock &obj_hash[i].lock irq_context: 0 &f->f_pos_lock sb_writers#4 &sb->s_type->i_mutex_key#8 &type->s_umount_key#32 rcu_read_lock &wb->work_lock rcu_read_lock &pool->lock/1 irq_context: 0 &f->f_pos_lock sb_writers#4 &sb->s_type->i_mutex_key#8 &type->s_umount_key#32 rcu_read_lock &wb->work_lock rcu_read_lock &pool->lock/1 &obj_hash[i].lock irq_context: 0 &f->f_pos_lock sb_writers#4 &sb->s_type->i_mutex_key#8 &type->s_umount_key#32 rcu_read_lock &wb->work_lock rcu_read_lock &pool->lock/1 pool_lock#2 irq_context: 0 &f->f_pos_lock sb_writers#4 &sb->s_type->i_mutex_key#8 &type->s_umount_key#32 rcu_read_lock &wb->work_lock rcu_read_lock &pool->lock/1 &p->pi_lock irq_context: 0 &f->f_pos_lock sb_writers#4 &sb->s_type->i_mutex_key#8 &type->s_umount_key#32 rcu_read_lock &wb->work_lock rcu_read_lock &pool->lock/1 &p->pi_lock &rq->__lock irq_context: 0 &f->f_pos_lock sb_writers#4 &sb->s_type->i_mutex_key#8 &type->s_umount_key#32 rcu_read_lock &wb->work_lock rcu_read_lock &pool->lock/1 &p->pi_lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 &f->f_pos_lock sb_writers#4 &sb->s_type->i_mutex_key#8 &type->s_umount_key#32 rcu_read_lock rcu_node_0 irq_context: 0 &f->f_pos_lock sb_writers#4 &sb->s_type->i_mutex_key#8 &type->s_umount_key#32 rcu_read_lock &rq->__lock irq_context: 0 &f->f_pos_lock sb_writers#4 &sb->s_type->i_mutex_key#8 &type->s_umount_key#32 rcu_read_lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)writeback (work_completion)(&(&wb->dwork)->work) &sb->s_type->i_lock_key#22 irq_context: 0 (wq_completion)writeback (work_completion)(&(&wb->dwork)->work) &sbi->s_writepages_rwsem irq_context: 0 (wq_completion)writeback (work_completion)(&(&wb->dwork)->work) &sbi->s_writepages_rwsem &xa->xa_lock#9 irq_context: 0 (wq_completion)writeback (work_completion)(&(&wb->dwork)->work) &sbi->s_writepages_rwsem mmu_notifier_invalidate_range_start irq_context: 0 (wq_completion)writeback (work_completion)(&(&wb->dwork)->work) &sbi->s_writepages_rwsem pool_lock#2 irq_context: 0 (wq_completion)writeback (work_completion)(&(&wb->dwork)->work) &sbi->s_writepages_rwsem lock#4 irq_context: 0 (wq_completion)writeback (work_completion)(&(&wb->dwork)->work) &sbi->s_writepages_rwsem lock#4 &lruvec->lru_lock irq_context: 0 (wq_completion)writeback (work_completion)(&(&wb->dwork)->work) &sbi->s_writepages_rwsem lock#5 irq_context: 0 (wq_completion)writeback (work_completion)(&(&wb->dwork)->work) &sbi->s_writepages_rwsem &obj_hash[i].lock irq_context: 0 (wq_completion)writeback (work_completion)(&(&wb->dwork)->work) &sbi->s_writepages_rwsem &journal->j_state_lock irq_context: 0 (wq_completion)writeback (work_completion)(&(&wb->dwork)->work) &sbi->s_writepages_rwsem jbd2_handle irq_context: 0 (wq_completion)writeback (work_completion)(&(&wb->dwork)->work) &sbi->s_writepages_rwsem jbd2_handle lock#4 irq_context: 0 (wq_completion)writeback (work_completion)(&(&wb->dwork)->work) &sbi->s_writepages_rwsem jbd2_handle lock#5 irq_context: 0 (wq_completion)writeback (work_completion)(&(&wb->dwork)->work) &sbi->s_writepages_rwsem jbd2_handle mmu_notifier_invalidate_range_start irq_context: 0 (wq_completion)writeback (work_completion)(&(&wb->dwork)->work) &sbi->s_writepages_rwsem jbd2_handle pool_lock#2 irq_context: 0 (wq_completion)writeback (work_completion)(&(&wb->dwork)->work) &sbi->s_writepages_rwsem jbd2_handle &ei->i_es_lock irq_context: 0 (wq_completion)writeback (work_completion)(&(&wb->dwork)->work) &sbi->s_writepages_rwsem jbd2_handle &ei->i_es_lock key#2 irq_context: 0 (wq_completion)writeback (work_completion)(&(&wb->dwork)->work) &sbi->s_writepages_rwsem jbd2_handle &ei->i_data_sem irq_context: 0 (wq_completion)writeback (work_completion)(&(&wb->dwork)->work) &sbi->s_writepages_rwsem jbd2_handle &ei->i_data_sem mmu_notifier_invalidate_range_start irq_context: 0 (wq_completion)writeback (work_completion)(&(&wb->dwork)->work) &sbi->s_writepages_rwsem jbd2_handle &ei->i_data_sem &c->lock irq_context: 0 (wq_completion)writeback (work_completion)(&(&wb->dwork)->work) &sbi->s_writepages_rwsem jbd2_handle &ei->i_data_sem pool_lock#2 irq_context: 0 (wq_completion)writeback (work_completion)(&(&wb->dwork)->work) &sbi->s_writepages_rwsem jbd2_handle &ei->i_data_sem &ei->i_prealloc_lock irq_context: 0 (wq_completion)writeback (work_completion)(&(&wb->dwork)->work) &sbi->s_writepages_rwsem jbd2_handle &ei->i_data_sem &rq->__lock irq_context: 0 (wq_completion)writeback (work_completion)(&(&wb->dwork)->work) &sbi->s_writepages_rwsem jbd2_handle &ei->i_data_sem &sbi->s_md_lock irq_context: 0 (wq_completion)writeback (work_completion)(&(&wb->dwork)->work) &sbi->s_writepages_rwsem jbd2_handle &ei->i_data_sem &bgl->locks[i].lock &sbi->s_md_lock irq_context: 0 (wq_completion)writeback (work_completion)(&(&wb->dwork)->work) &sbi->s_writepages_rwsem jbd2_handle &ei->i_data_sem &bgl->locks[i].lock &ei->i_prealloc_lock irq_context: 0 (wq_completion)writeback (work_completion)(&(&wb->dwork)->work) &sbi->s_writepages_rwsem jbd2_handle &ei->i_data_sem &mapping->private_lock irq_context: 0 (wq_completion)writeback (work_completion)(&(&wb->dwork)->work) &sbi->s_writepages_rwsem jbd2_handle &ei->i_data_sem key#3 irq_context: 0 (wq_completion)writeback (work_completion)(&(&wb->dwork)->work) &sbi->s_writepages_rwsem jbd2_handle &ei->i_data_sem &pa->pa_lock#2 irq_context: 0 (wq_completion)writeback (work_completion)(&(&wb->dwork)->work) &sbi->s_writepages_rwsem jbd2_handle &ei->i_data_sem &obj_hash[i].lock irq_context: 0 (wq_completion)writeback (work_completion)(&(&wb->dwork)->work) &sbi->s_writepages_rwsem jbd2_handle &ei->i_data_sem &ei->i_raw_lock irq_context: 0 (wq_completion)writeback (work_completion)(&(&wb->dwork)->work) &sbi->s_writepages_rwsem jbd2_handle &ei->i_data_sem &(ei->i_block_reservation_lock) irq_context: 0 (wq_completion)writeback (work_completion)(&(&wb->dwork)->work) &sbi->s_writepages_rwsem jbd2_handle &ei->i_data_sem &(ei->i_block_reservation_lock) key#14 irq_context: 0 (wq_completion)writeback (work_completion)(&(&wb->dwork)->work) &sbi->s_writepages_rwsem jbd2_handle &ei->i_data_sem &sb->s_type->i_lock_key#22 irq_context: 0 (wq_completion)writeback (work_completion)(&(&wb->dwork)->work) &sbi->s_writepages_rwsem jbd2_handle &ei->i_data_sem &ei->i_es_lock irq_context: 0 (wq_completion)writeback (work_completion)(&(&wb->dwork)->work) &sbi->s_writepages_rwsem jbd2_handle &ei->i_data_sem &ei->i_es_lock irq_context: 0 (wq_completion)writeback (work_completion)(&(&wb->dwork)->work) &sbi->s_writepages_rwsem jbd2_handle &ei->i_data_sem &ei->i_es_lock &obj_hash[i].lock irq_context: 0 (wq_completion)writeback (work_completion)(&(&wb->dwork)->work) &sbi->s_writepages_rwsem jbd2_handle &ei->i_data_sem &ei->i_es_lock pool_lock#2 irq_context: 0 (wq_completion)writeback (work_completion)(&(&wb->dwork)->work) &sbi->s_writepages_rwsem jbd2_handle &mapping->i_mmap_rwsem mmu_notifier_invalidate_range_start irq_context: 0 (wq_completion)writeback (work_completion)(&(&wb->dwork)->work) &sbi->s_writepages_rwsem jbd2_handle &mapping->i_mmap_rwsem ptlock_ptr(page)#2 irq_context: 0 (wq_completion)writeback (work_completion)(&(&wb->dwork)->work) &sbi->s_writepages_rwsem jbd2_handle &mapping->private_lock irq_context: 0 (wq_completion)writeback (work_completion)(&(&wb->dwork)->work) &sbi->s_writepages_rwsem jbd2_handle &mapping->private_lock rcu_read_lock &memcg->move_lock irq_context: 0 (wq_completion)writeback (work_completion)(&(&wb->dwork)->work) &sbi->s_writepages_rwsem jbd2_handle rcu_read_lock &memcg->move_lock irq_context: 0 (wq_completion)writeback (work_completion)(&(&wb->dwork)->work) &sbi->s_writepages_rwsem jbd2_handle rcu_read_lock &xa->xa_lock#9 irq_context: 0 (wq_completion)writeback (work_completion)(&(&wb->dwork)->work) &sbi->s_writepages_rwsem jbd2_handle rcu_read_lock &xa->xa_lock#9 &s->s_inode_wblist_lock irq_context: 0 (wq_completion)writeback (work_completion)(&(&wb->dwork)->work) &sbi->s_writepages_rwsem jbd2_handle rcu_read_lock &xa->xa_lock#9 key#10 irq_context: 0 (wq_completion)writeback (work_completion)(&(&wb->dwork)->work) &sbi->s_writepages_rwsem jbd2_handle rcu_read_lock key#10 irq_context: 0 (wq_completion)writeback (work_completion)(&(&wb->dwork)->work) &sbi->s_writepages_rwsem jbd2_handle &rq->__lock irq_context: 0 (wq_completion)writeback (work_completion)(&(&wb->dwork)->work) &sbi->s_writepages_rwsem jbd2_handle rcu_node_0 irq_context: 0 (wq_completion)writeback (work_completion)(&(&wb->dwork)->work) &sbi->s_writepages_rwsem jbd2_handle &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)writeback (work_completion)(&(&wb->dwork)->work) &sbi->s_writepages_rwsem jbd2_handle &rq->__lock &cfs_rq->removed.lock irq_context: 0 (wq_completion)writeback (work_completion)(&(&wb->dwork)->work) &sbi->s_writepages_rwsem jbd2_handle rcu_read_lock rcu_node_0 irq_context: 0 (wq_completion)writeback (work_completion)(&(&wb->dwork)->work) &sbi->s_writepages_rwsem jbd2_handle &ei->i_raw_lock irq_context: 0 (wq_completion)writeback (work_completion)(&(&wb->dwork)->work) &sbi->s_writepages_rwsem tk_core.seq.seqcount irq_context: 0 (wq_completion)writeback (work_completion)(&(&wb->dwork)->work) &sbi->s_writepages_rwsem &dd->lock irq_context: 0 (wq_completion)writeback (work_completion)(&(&wb->dwork)->work) &sbi->s_writepages_rwsem rcu_read_lock &dd->lock irq_context: 0 (wq_completion)writeback (work_completion)(&(&wb->dwork)->work) &sbi->s_writepages_rwsem rcu_read_lock &obj_hash[i].lock irq_context: 0 (wq_completion)writeback (work_completion)(&(&wb->dwork)->work) &sbi->s_writepages_rwsem rcu_read_lock pool_lock#2 irq_context: 0 (wq_completion)writeback (work_completion)(&(&wb->dwork)->work) &sbi->s_writepages_rwsem rcu_read_lock tk_core.seq.seqcount irq_context: 0 (wq_completion)writeback (work_completion)(&(&wb->dwork)->work) &sbi->s_writepages_rwsem rcu_read_lock &virtscsi_vq->vq_lock irq_context: 0 (wq_completion)writeback (work_completion)(&(&wb->dwork)->work) &obj_hash[i].lock irq_context: 0 (wq_completion)writeback (work_completion)(&(&wb->dwork)->work) pool_lock#2 irq_context: 0 sb_writers#4 &sb->s_type->i_mutex_key#8 jbd2_handle &ei->i_data_sem &pcp->lock &zone->lock irq_context: 0 pernet_ops_rwsem rtnl_mutex &idev->mc_lock _xmit_ETHER &____s->seqcount#2 irq_context: 0 pernet_ops_rwsem rtnl_mutex &idev->mc_lock _xmit_ETHER &____s->seqcount irq_context: 0 sb_writers#4 &sb->s_type->i_mutex_key#8 jbd2_handle &ei->i_data_sem &ei->i_es_lock key#6 irq_context: 0 slock-AF_INET6 tk_core.seq.seqcount irq_context: 0 slock-AF_INET6 pool_lock#2 irq_context: 0 slock-AF_INET6 rcu_read_lock rcu_read_lock &____s->seqcount#7 irq_context: 0 slock-AF_INET6 rcu_read_lock rcu_read_lock &ct->lock irq_context: 0 slock-AF_INET6 rcu_read_lock rcu_read_lock &____s->seqcount#10 irq_context: 0 slock-AF_INET6 rcu_read_lock rcu_read_lock rcu_read_lock_bh &list->lock#5 irq_context: softirq rcu_read_lock rcu_read_lock slock-AF_INET6/1 tk_core.seq.seqcount irq_context: softirq rcu_read_lock rcu_read_lock slock-AF_INET6/1 &obj_hash[i].lock irq_context: softirq rcu_read_lock rcu_read_lock slock-AF_INET6/1 pool_lock#2 irq_context: softirq rcu_read_lock rcu_read_lock slock-AF_INET6/1 rcu_read_lock rcu_read_lock &____s->seqcount#7 irq_context: softirq rcu_read_lock rcu_read_lock slock-AF_INET6/1 rcu_read_lock rcu_read_lock &ct->lock irq_context: softirq rcu_read_lock rcu_read_lock slock-AF_INET6/1 rcu_read_lock rcu_read_lock &____s->seqcount#10 irq_context: softirq rcu_read_lock rcu_read_lock slock-AF_INET6/1 rcu_read_lock rcu_read_lock rcu_read_lock_bh &list->lock#5 irq_context: softirq rcu_read_lock rcu_read_lock slock-AF_INET6/1 &base->lock irq_context: softirq rcu_read_lock rcu_read_lock slock-AF_INET6/1 &base->lock &obj_hash[i].lock irq_context: softirq rcu_read_lock rcu_read_lock slock-AF_INET6/1 rcu_read_lock &ei->socket.wq.wait irq_context: 0 &f->f_pos_lock sb_writers#4 &sb->s_type->i_mutex_key#8 &type->s_umount_key#32 &bdi->wb_waitq irq_context: 0 &f->f_pos_lock sb_writers#4 &sb->s_type->i_mutex_key#8 &type->s_umount_key#32 &rq->__lock irq_context: 0 (wq_completion)writeback (work_completion)(&(&wb->dwork)->work) &sbi->s_writepages_rwsem jbd2_handle &ei->i_data_sem &____s->seqcount#2 irq_context: 0 &f->f_pos_lock sb_writers#4 &sb->s_type->i_mutex_key#8 &type->s_umount_key#32 &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)writeback (work_completion)(&(&wb->dwork)->work) &sbi->s_writepages_rwsem jbd2_handle &ei->i_data_sem &____s->seqcount irq_context: 0 (wq_completion)writeback (work_completion)(&(&wb->dwork)->work) &sbi->s_writepages_rwsem jbd2_handle &ei->i_data_sem &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 sk_lock-AF_INET6 &mm->mmap_lock mmu_notifier_invalidate_range_start irq_context: 0 sk_lock-AF_INET6 &mm->mmap_lock ptlock_ptr(page)#2 lock#4 irq_context: 0 (wq_completion)writeback (work_completion)(&(&wb->dwork)->work) &sbi->s_writepages_rwsem jbd2_handle &c->lock irq_context: 0 sk_lock-AF_INET6 &mm->mmap_lock ptlock_ptr(page)#2 lock#4 &lruvec->lru_lock irq_context: 0 sk_lock-AF_INET6 &mm->mmap_lock ptlock_ptr(page)#2 key irq_context: 0 (wq_completion)writeback (work_completion)(&(&wb->dwork)->work) &bdi->wb_waitq irq_context: 0 (wq_completion)writeback (work_completion)(&(&wb->dwork)->work) &bdi->wb_waitq &p->pi_lock irq_context: 0 (wq_completion)writeback (work_completion)(&(&wb->dwork)->work) &bdi->wb_waitq &p->pi_lock &rq->__lock irq_context: 0 (wq_completion)writeback (work_completion)(&(&wb->dwork)->work) &bdi->wb_waitq &p->pi_lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 sb_writers#4 &sb->s_type->i_mutex_key#8 jbd2_handle &ei->i_data_sem &bgl->locks[i].lock &pa->pa_lock#2 irq_context: 0 sb_writers#4 &sb->s_type->i_mutex_key#8 jbd2_handle &ei->i_data_sem &bgl->locks[i].lock &pa->pa_lock irq_context: 0 sb_writers#4 &sb->s_type->i_mutex_key#8 jbd2_handle &ei->i_data_sem &bgl->locks[i].lock &lg->lg_prealloc_lock irq_context: 0 sb_writers#4 &sb->s_type->i_mutex_key#8 jbd2_handle &ei->i_data_sem &bgl->locks[i].lock &obj_hash[i].lock irq_context: 0 sb_writers#4 &sb->s_type->i_mutex_key#8 jbd2_handle &ei->i_data_sem &bgl->locks[i].lock pool_lock#2 irq_context: 0 &f->f_pos_lock sb_writers#4 &sb->s_type->i_mutex_key#8 jbd2_handle &ei->i_data_sem &wb->list_lock irq_context: 0 &f->f_pos_lock sb_writers#4 &sb->s_type->i_mutex_key#8 jbd2_handle &ei->i_data_sem &wb->list_lock &sb->s_type->i_lock_key#22 irq_context: 0 &f->f_pos_lock sb_writers#4 &sb->s_type->i_mutex_key#8 jbd2_handle &ei->i_data_sem &ei->i_prealloc_lock irq_context: 0 &f->f_pos_lock sb_writers#4 &sb->s_type->i_mutex_key#8 jbd2_handle &ei->i_data_sem &ei->i_prealloc_lock &pa->pa_lock#2 irq_context: 0 &f->f_pos_lock sb_writers#4 &sb->s_type->i_mutex_key#8 jbd2_handle &ei->i_data_sem &pa->pa_lock#2 irq_context: 0 &f->f_pos_lock sb_writers#4 &sb->s_type->i_mutex_key#8 jbd2_handle rcu_read_lock &memcg->move_lock irq_context: 0 &f->f_pos_lock sb_writers#4 &sb->s_type->i_mutex_key#8 jbd2_handle rcu_read_lock &xa->xa_lock#9 irq_context: 0 &f->f_pos_lock sb_writers#4 &sb->s_type->i_mutex_key#8 jbd2_handle &ei->i_data_sem &bgl->locks[i].lock &sbi->s_md_lock irq_context: 0 &f->f_pos_lock sb_writers#4 &sb->s_type->i_mutex_key#8 jbd2_handle &ei->i_data_sem &bgl->locks[i].lock &ei->i_prealloc_lock irq_context: 0 (wq_completion)writeback (work_completion)(&(&wb->dwork)->work) &sbi->s_writepages_rwsem &c->lock irq_context: 0 &f->f_pos_lock sb_writers#4 &sb->s_type->i_mutex_key#8 jbd2_handle &ei->i_data_sem &c->lock irq_context: 0 &f->f_pos_lock sb_writers#4 &sb->s_type->i_mutex_key#8 jbd2_handle &ei->i_data_sem &sbi->s_md_lock irq_context: 0 &f->f_pos_lock sb_writers#4 &sb->s_type->i_mutex_key#8 jbd2_handle &ei->i_data_sem rcu_read_lock rcu_node_0 irq_context: 0 &f->f_pos_lock sb_writers#4 &sb->s_type->i_mutex_key#8 jbd2_handle &ei->i_data_sem rcu_read_lock &rq->__lock irq_context: 0 &f->f_pos_lock sb_writers#4 &sb->s_type->i_mutex_key#8 mapping.invalidate_lock jbd2_handle &ei->i_data_sem &c->lock irq_context: 0 &f->f_pos_lock sb_writers#4 &sb->s_type->i_mutex_key#8 mapping.invalidate_lock jbd2_handle &ei->i_data_sem &n->list_lock irq_context: 0 &f->f_pos_lock sb_writers#4 &sb->s_type->i_mutex_key#8 mapping.invalidate_lock jbd2_handle &ei->i_data_sem &n->list_lock &c->lock irq_context: 0 sb_writers#4 &sb->s_type->i_mutex_key#8 jbd2_handle &journal->j_wait_updates &p->pi_lock &cfs_rq->removed.lock irq_context: 0 &f->f_pos_lock sb_writers#4 &sb->s_type->i_mutex_key#8 jbd2_handle &lruvec->lru_lock irq_context: 0 &f->f_pos_lock sb_writers#4 &sb->s_type->i_mutex_key#8 jbd2_handle &ei->i_data_sem &ret->b_state_lock irq_context: 0 &f->f_pos_lock sb_writers#4 &sb->s_type->i_mutex_key#8 jbd2_handle &ei->i_data_sem &ret->b_state_lock &journal->j_list_lock irq_context: 0 rtnl_mutex sk_lock-AF_INET fs_reclaim irq_context: 0 rtnl_mutex sk_lock-AF_INET fs_reclaim mmu_notifier_invalidate_range_start irq_context: 0 rtnl_mutex sk_lock-AF_INET &c->lock irq_context: 0 rtnl_mutex sk_lock-AF_INET pool_lock#2 irq_context: 0 rtnl_mutex sk_lock-AF_INET &obj_hash[i].lock irq_context: 0 rtnl_mutex sk_lock-AF_INET &in_dev->mc_tomb_lock irq_context: 0 rtnl_mutex sk_lock-AF_INET &im->lock irq_context: 0 rtnl_mutex sk_lock-AF_INET &base->lock irq_context: 0 rtnl_mutex sk_lock-AF_INET &base->lock &obj_hash[i].lock irq_context: 0 rtnl_mutex sk_lock-AF_INET &rq->__lock irq_context: 0 rtnl_mutex sk_lock-AF_INET &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: softirq (&in_dev->mr_ifc_timer) rcu_read_lock &im->lock &zone->lock irq_context: softirq (&in_dev->mr_ifc_timer) rcu_read_lock rcu_read_lock_bh tk_core.seq.seqcount irq_context: softirq (&in_dev->mr_ifc_timer) rcu_read_lock rcu_read_lock_bh &list->lock#5 irq_context: 0 &f->f_pos_lock sb_writers#4 &sb->s_type->i_mutex_key#8 mapping.invalidate_lock jbd2_handle &ei->i_data_sem &ei->i_prealloc_lock &pa->pa_lock#2 irq_context: softirq (&icsk->icsk_retransmit_timer) slock-AF_INET6 irq_context: 0 sk_lock-AF_INET6 &sem->wait_lock irq_context: 0 sk_lock-AF_INET6 &p->pi_lock irq_context: 0 sk_lock-AF_INET6 &p->pi_lock &cfs_rq->removed.lock irq_context: 0 sk_lock-AF_INET6 &p->pi_lock &rq->__lock irq_context: 0 sk_lock-AF_INET6 &p->pi_lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 sk_lock-AF_INET6 &mm->mmap_lock &sem->wait_lock irq_context: 0 sb_writers#4 sb_internal jbd2_handle &ei->i_data_sem &wb->list_lock irq_context: 0 sb_writers#4 sb_internal jbd2_handle &ei->i_data_sem &wb->list_lock &sb->s_type->i_lock_key#22 irq_context: 0 rtnl_mutex rcu_read_lock &im->lock irq_context: 0 rtnl_mutex (inetaddr_chain).rwsem rcu_read_lock rcu_read_lock pool_lock#2 irq_context: 0 rtnl_mutex (inetaddr_chain).rwsem rcu_read_lock &obj_hash[i].lock irq_context: 0 rtnl_mutex (inetaddr_chain).rwsem rcu_read_lock &rcu_state.gp_wq irq_context: 0 rtnl_mutex (inetaddr_chain).rwsem rcu_read_lock &rcu_state.gp_wq &p->pi_lock irq_context: 0 rtnl_mutex (inetaddr_chain).rwsem rcu_read_lock &rcu_state.gp_wq &p->pi_lock &rq->__lock irq_context: 0 rtnl_mutex (inetaddr_chain).rwsem rcu_read_lock &rcu_state.gp_wq &p->pi_lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 rtnl_mutex (inetaddr_chain).rwsem &tbl->lock &n->lock irq_context: 0 rtnl_mutex (inetaddr_chain).rwsem &tbl->lock pool_lock#2 irq_context: 0 rtnl_mutex (inetaddr_chain).rwsem &tbl->lock &c->lock irq_context: 0 rtnl_mutex (inetaddr_chain).rwsem &tbl->lock &n->lock irq_context: 0 rtnl_mutex (inetaddr_chain).rwsem &tbl->lock &n->lock &____s->seqcount#9 irq_context: 0 rtnl_mutex (inetaddr_chain).rwsem &tbl->lock nl_table_lock irq_context: 0 rtnl_mutex (inetaddr_chain).rwsem &tbl->lock &obj_hash[i].lock irq_context: 0 rtnl_mutex (inetaddr_chain).rwsem &tbl->lock nl_table_wait.lock irq_context: 0 rtnl_mutex (inetaddr_chain).rwsem &tbl->lock rcu_read_lock lock#8 irq_context: 0 rtnl_mutex (inetaddr_chain).rwsem &tbl->lock rcu_read_lock id_table_lock irq_context: 0 rtnl_mutex (inetaddr_chain).rwsem &tbl->lock &dir->lock#2 irq_context: 0 rtnl_mutex (inetaddr_chain).rwsem &tbl->lock krc.lock irq_context: 0 rtnl_mutex &tbl->lock &____s->seqcount#2 irq_context: 0 rtnl_mutex &tbl->lock rcu_read_lock pool_lock#2 irq_context: 0 (wq_completion)events drain_vmap_work vmap_purge_lock rcu_node_0 irq_context: 0 (wq_completion)events drain_vmap_work vmap_purge_lock &rcu_state.expedited_wq irq_context: 0 (wq_completion)events drain_vmap_work vmap_purge_lock &rcu_state.expedited_wq &p->pi_lock irq_context: 0 (wq_completion)events drain_vmap_work vmap_purge_lock &rcu_state.expedited_wq &p->pi_lock &rq->__lock irq_context: 0 (wq_completion)events drain_vmap_work vmap_purge_lock &rcu_state.expedited_wq &p->pi_lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 sk_lock-AF_PACKET rcu_state.exp_mutex.wait_lock irq_context: 0 sk_lock-AF_PACKET &p->pi_lock irq_context: 0 sk_lock-AF_PACKET &p->pi_lock &rq->__lock irq_context: 0 sk_lock-AF_PACKET &p->pi_lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 &sb->s_type->i_mutex_key#10 &rnp->exp_wq[1] irq_context: 0 slock-AF_INET6 rcu_read_lock rcu_read_lock rcu_read_lock stock_lock irq_context: 0 slock-AF_INET6 rcu_read_lock rcu_read_lock rcu_read_lock pool_lock#2 irq_context: 0 slock-AF_INET6 rcu_read_lock rcu_read_lock rcu_read_lock &dir->lock#2 irq_context: 0 slock-AF_INET6 rcu_read_lock rcu_read_lock pool_lock#2 irq_context: 0 slock-AF_INET6 rcu_read_lock rcu_read_lock &obj_hash[i].lock irq_context: 0 slock-AF_INET6 rcu_read_lock rcu_read_lock &c->lock irq_context: 0 slock-AF_INET6 rcu_read_lock rcu_read_lock &dir->lock#2 irq_context: 0 slock-AF_INET6 rcu_read_lock rcu_read_lock &tbl->lock irq_context: 0 slock-AF_INET6 rcu_read_lock rcu_read_lock &____s->seqcount#9 irq_context: 0 slock-AF_INET6 &obj_hash[i].lock irq_context: 0 slock-AF_INET6 &base->lock irq_context: 0 slock-AF_INET6 &base->lock &obj_hash[i].lock irq_context: 0 sk_lock-AF_PACKET &rnp->exp_lock irq_context: 0 sk_lock-AF_PACKET &rnp->exp_wq[2] irq_context: 0 sk_lock-AF_PACKET &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 sk_lock-AF_INET6 &mm->mmap_lock ptlock_ptr(page)#2 lock#4 &lruvec->lru_lock per_cpu_ptr(&cgroup_rstat_cpu_lock, cpu) irq_context: 0 sk_lock-AF_INET6 &sd->defer_lock irq_context: 0 sk_lock-AF_INET6 &mm->mmap_lock fs_reclaim &rq->__lock irq_context: 0 sk_lock-AF_INET6 &mm->mmap_lock fs_reclaim &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 rtnl_mutex &dev->tx_global_lock _xmit_LOOPBACK#2 irq_context: 0 sk_lock-AF_INET6 &mm->mmap_lock &pcp->lock &zone->lock irq_context: 0 sb_writers#4 &sb->s_type->i_mutex_key#8 &sbi->s_writepages_rwsem pool_lock irq_context: 0 &journal->j_checkpoint_mutex &fq->mq_flush_lock irq_context: 0 rtnl_mutex sk_lock-AF_INET krc.lock irq_context: 0 &f->f_pos_lock sb_writers#4 &sb->s_type->i_mutex_key#8 &mm->mmap_lock &cfs_rq->removed.lock irq_context: 0 &f->f_pos_lock sb_writers#4 &sb->s_type->i_mutex_key#8 &mm->mmap_lock pool_lock#2 irq_context: 0 sb_writers#4 &type->i_mutex_dir_key#3/1 jbd2_handle &sbi->s_orphan_lock irq_context: 0 sb_writers#4 &type->i_mutex_dir_key#3/1 sb_internal jbd2_handle &ei->i_es_lock &sbi->s_es_lock irq_context: 0 sb_writers#4 &type->i_mutex_dir_key#3/1 sb_internal jbd2_handle &ei->i_es_lock &obj_hash[i].lock irq_context: 0 sb_writers#4 &type->i_mutex_dir_key#3/1 sb_internal jbd2_handle &ei->i_es_lock pool_lock#2 irq_context: 0 sb_writers#4 &type->i_mutex_dir_key#3/1 &journal->j_state_lock &journal->j_wait_commit irq_context: 0 sb_writers#4 &type->i_mutex_dir_key#3/1 &journal->j_state_lock &journal->j_wait_commit &p->pi_lock irq_context: 0 sb_writers#4 &type->i_mutex_dir_key#3/1 &journal->j_state_lock &journal->j_wait_commit &p->pi_lock &rq->__lock irq_context: 0 sb_writers#4 &type->i_mutex_dir_key#3/1 &journal->j_state_lock &journal->j_wait_commit &p->pi_lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 sb_writers#4 &type->i_mutex_dir_key#3/1 &journal->j_wait_commit irq_context: 0 sb_writers#4 &type->i_mutex_dir_key#3/1 &journal->j_wait_done_commit irq_context: 0 (wq_completion)events (work_completion)(&w->work)#2 &rq->__lock irq_context: 0 (wq_completion)events (work_completion)(&w->work)#2 &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 sk_lock-AF_INET6 &mm->mmap_lock &p->pi_lock irq_context: 0 &f->f_pos_lock &type->i_mutex_dir_key#3 sb_writers#4 &journal->j_state_lock &journal->j_wait_transaction_locked irq_context: 0 &f->f_pos_lock &type->i_mutex_dir_key#3 sb_writers#4 jbd2_handle irq_context: 0 &mm->mmap_lock sb_pagefaults jbd2_handle mmu_notifier_invalidate_range_start irq_context: 0 &mm->mmap_lock sb_pagefaults jbd2_handle &____s->seqcount irq_context: 0 &mm->mmap_lock sb_pagefaults jbd2_handle pool_lock#2 irq_context: 0 &mm->mmap_lock sb_pagefaults jbd2_handle rcu_read_lock pool_lock#2 irq_context: 0 &mm->mmap_lock sb_pagefaults jbd2_handle &obj_hash[i].lock irq_context: 0 &xt[i].mutex &pcp->lock &zone->lock &____s->seqcount irq_context: 0 sk_lock-AF_INET6 &mm->mmap_lock &cfs_rq->removed.lock irq_context: 0 sk_lock-AF_INET6 &mm->mmap_lock &obj_hash[i].lock irq_context: 0 sk_lock-AF_INET6 &mm->mmap_lock pool_lock#2 irq_context: 0 sk_lock-AF_INET6 &mm->mmap_lock &obj_hash[i].lock pool_lock irq_context: 0 (wq_completion)afs (work_completion)(&net->fs_manager) &rq->__lock irq_context: 0 sk_lock-AF_INET6 &mm->mmap_lock &pcp->lock &zone->lock &____s->seqcount irq_context: 0 (wq_completion)events drain_vmap_work vmap_purge_lock free_vmap_area_lock init_mm.page_table_lock rcu_read_lock &p->pi_lock irq_context: 0 (wq_completion)events_power_efficient (work_completion)(&barr->work) &rq->__lock irq_context: 0 (wq_completion)events_power_efficient (work_completion)(&barr->work) &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem &rnp->exp_wq[1] irq_context: 0 (wq_completion)events drain_vmap_work vmap_purge_lock free_vmap_area_lock init_mm.page_table_lock rcu_read_lock &p->pi_lock &rq->__lock irq_context: 0 (wq_completion)events drain_vmap_work vmap_purge_lock free_vmap_area_lock init_mm.page_table_lock rcu_read_lock &p->pi_lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)wg-crypt-wg0#4 (work_completion)(&peer->transmit_packet_work) &peer->endpoint_lock rcu_read_lock_bh rcu_read_lock rcu_read_lock rcu_read_lock pool_lock#2 irq_context: 0 (wq_completion)wg-crypt-wg0#4 (work_completion)(&peer->transmit_packet_work) &peer->endpoint_lock rcu_read_lock_bh rcu_read_lock rcu_read_lock rcu_read_lock &c->lock irq_context: 0 (wq_completion)wg-crypt-wg0#4 (work_completion)(&peer->transmit_packet_work) &peer->endpoint_lock rcu_read_lock_bh rcu_read_lock rcu_read_lock rcu_read_lock &dir->lock#2 irq_context: 0 (wq_completion)wg-crypt-wg0#4 (work_completion)(&peer->transmit_packet_work) &peer->endpoint_lock rcu_read_lock_bh rcu_read_lock pool_lock#2 irq_context: 0 (wq_completion)wg-crypt-wg0#4 (work_completion)(&peer->transmit_packet_work) &peer->endpoint_lock rcu_read_lock_bh rcu_read_lock rcu_read_lock &____s->seqcount#7 irq_context: 0 (wq_completion)wg-crypt-wg0#4 (work_completion)(&peer->transmit_packet_work) &peer->endpoint_lock rcu_read_lock_bh rcu_read_lock &nf_nat_locks[i] irq_context: 0 (wq_completion)wg-crypt-wg0#4 (work_completion)(&peer->transmit_packet_work) &peer->endpoint_lock rcu_read_lock_bh rcu_read_lock &nf_conntrack_locks[i] irq_context: 0 (wq_completion)wg-crypt-wg0#4 (work_completion)(&peer->transmit_packet_work) &peer->endpoint_lock rcu_read_lock_bh rcu_read_lock &nf_conntrack_locks[i] &nf_conntrack_locks[i]/1 irq_context: 0 (wq_completion)wg-crypt-wg0#4 (work_completion)(&peer->transmit_packet_work) &peer->endpoint_lock rcu_read_lock_bh rcu_read_lock &nf_conntrack_locks[i] &nf_conntrack_locks[i]/1 batched_entropy_u8.lock irq_context: 0 (wq_completion)wg-crypt-wg0#4 (work_completion)(&peer->transmit_packet_work) &peer->endpoint_lock rcu_read_lock_bh pool_lock#2 irq_context: 0 (wq_completion)wg-crypt-wg0#4 (work_completion)(&peer->transmit_packet_work) &peer->endpoint_lock rcu_read_lock_bh rcu_read_lock rcu_read_lock_bh &obj_hash[i].lock irq_context: 0 (wq_completion)wg-crypt-wg0#4 (work_completion)(&peer->transmit_packet_work) &peer->endpoint_lock rcu_read_lock_bh rcu_read_lock rcu_read_lock_bh pool_lock#2 irq_context: 0 &mm->mmap_lock sb_pagefaults &n->list_lock irq_context: 0 sk_lock-AF_INET6 rcu_read_lock rcu_read_lock &____s->seqcount#9 irq_context: 0 sk_lock-AF_INET6 rcu_read_lock rcu_read_lock rcu_read_lock_bh &obj_hash[i].lock irq_context: 0 sk_lock-AF_INET6 rcu_read_lock rcu_read_lock rcu_read_lock_bh pool_lock#2 irq_context: 0 slock-AF_INET6 batched_entropy_u32.lock irq_context: 0 slock-AF_INET6 &____s->seqcount irq_context: 0 slock-AF_INET6 rcu_read_lock pool_lock#2 irq_context: 0 &sb->s_type->i_mutex_key#10 sk_lock-AF_INET6 tk_core.seq.seqcount irq_context: 0 &mm->mmap_lock sb_pagefaults &n->list_lock &c->lock irq_context: 0 (wq_completion)wg-crypt-wg0#5 (work_completion)(&({ do { const void *__vpp_verify = (typeof((worker) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __asm__ ("" : "=r"(__ptr) : "0"((typeof(*((worker))) *)((worker)))); (typeof((typeof(*((worker))) *)((worker)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); })->work) &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 tomoyo_ss remove_cache_srcu &pcp->lock &zone->lock irq_context: 0 tomoyo_ss remove_cache_srcu &pcp->lock &zone->lock &____s->seqcount irq_context: 0 sb_writers#4 rcu_read_lock rcu_read_lock rcu_read_lock rcu_node_0 irq_context: 0 sb_writers#4 rcu_read_lock rcu_read_lock rcu_read_lock &rq->__lock irq_context: 0 sb_writers#4 rcu_read_lock rcu_read_lock rcu_read_lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 sb_writers#4 sb_internal jbd2_handle rcu_read_lock &rcu_state.expedited_wq irq_context: 0 sb_writers#4 sb_internal jbd2_handle rcu_read_lock &rcu_state.expedited_wq &p->pi_lock irq_context: 0 sb_writers#4 sb_internal jbd2_handle rcu_read_lock &rcu_state.expedited_wq &p->pi_lock &rq->__lock irq_context: 0 sb_writers#4 sb_internal jbd2_handle rcu_read_lock &rcu_state.expedited_wq &p->pi_lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 sk_lock-AF_PACKET rcu_state.exp_mutex irq_context: 0 sk_lock-AF_PACKET rcu_state.exp_mutex rcu_state.exp_mutex.wait_lock irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem rtnl_mutex &rnp->exp_wq[2] irq_context: 0 &sb->s_type->i_mutex_key#10 &rnp->exp_wq[3] irq_context: 0 &mm->mmap_lock &mapping->i_mmap_rwsem &rq->__lock irq_context: 0 &mm->mmap_lock &mapping->i_mmap_rwsem &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 &mm->mmap_lock &mapping->i_mmap_rwsem &cfs_rq->removed.lock irq_context: 0 &mm->mmap_lock &mapping->i_mmap_rwsem &obj_hash[i].lock irq_context: 0 &mm->mmap_lock &mapping->i_mmap_rwsem pool_lock#2 irq_context: 0 sb_writers#4 sb_internal jbd2_handle &ei->i_data_sem &ei->i_es_lock key#6 irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem rtnl_mutex pool_lock irq_context: 0 (wq_completion)events (work_completion)(&(&nsim_dev->trap_data->trap_report_dw)->work) fill_pool_map-wait-type-override &c->lock irq_context: 0 rtnl_mutex &br->hash_lock &____s->seqcount#2 irq_context: 0 rtnl_mutex &br->hash_lock &pcp->lock &zone->lock irq_context: softirq (&icsk->icsk_retransmit_timer) slock-AF_INET6 tk_core.seq.seqcount irq_context: softirq (&icsk->icsk_retransmit_timer) slock-AF_INET6 batched_entropy_u32.lock irq_context: softirq (&icsk->icsk_retransmit_timer) slock-AF_INET6 rcu_read_lock rcu_read_lock &____s->seqcount#7 irq_context: softirq (&icsk->icsk_retransmit_timer) slock-AF_INET6 rcu_read_lock rcu_read_lock &ct->lock irq_context: softirq (&icsk->icsk_retransmit_timer) slock-AF_INET6 rcu_read_lock rcu_read_lock &____s->seqcount#9 irq_context: softirq (&icsk->icsk_retransmit_timer) slock-AF_INET6 &obj_hash[i].lock irq_context: softirq (&icsk->icsk_retransmit_timer) slock-AF_INET6 &base->lock irq_context: softirq (&icsk->icsk_retransmit_timer) slock-AF_INET6 &base->lock &obj_hash[i].lock irq_context: 0 rtnl_mutex uevent_sock_mutex fs_reclaim mmu_notifier_invalidate_range_start &rq->__lock irq_context: 0 rtnl_mutex uevent_sock_mutex fs_reclaim mmu_notifier_invalidate_range_start &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 rtnl_mutex &idev->mc_lock &rq->__lock &cfs_rq->removed.lock irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem rtnl_mutex &cfs_rq->removed.lock irq_context: 0 cb_lock rcu_read_lock &rcu_state.expedited_wq irq_context: 0 cb_lock rcu_read_lock &rcu_state.expedited_wq &p->pi_lock irq_context: 0 cb_lock rcu_read_lock &rcu_state.expedited_wq &p->pi_lock &rq->__lock irq_context: 0 cb_lock rcu_read_lock &rcu_state.expedited_wq &p->pi_lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 &sb->s_type->i_mutex_key#10 (netlink_chain).rwsem &cfs_rq->removed.lock irq_context: 0 (wq_completion)wg-crypt-wg0#2 (work_completion)(&peer->transmit_packet_work) &peer->endpoint_lock rcu_read_lock_bh rcu_read_lock rcu_read_lock_bh rcu_read_lock rlock-AF_PACKET irq_context: 0 (wq_completion)wg-crypt-wg1#2 (work_completion)(&({ do { const void *__vpp_verify = (typeof((worker) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __asm__ ("" : "=r"(__ptr) : "0"((typeof(*((worker))) *)((worker)))); (typeof((typeof(*((worker))) *)((worker)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); })->work) &c->lock irq_context: 0 rtnl_mutex rcu_read_lock_bh rcu_read_lock pool_lock#2 irq_context: 0 rtnl_mutex rcu_read_lock_bh rcu_read_lock tk_core.seq.seqcount irq_context: 0 rtnl_mutex rcu_read_lock_bh rcu_read_lock rlock-AF_PACKET irq_context: 0 rtnl_mutex rcu_read_lock_bh pool_lock#2 irq_context: 0 rtnl_mutex rcu_read_lock &bond->ipsec_lock irq_context: 0 (wq_completion)events (linkwatch_work).work rtnl_mutex krc.lock irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem rtnl_mutex fill_pool_map-wait-type-override &rq->__lock irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem rtnl_mutex fill_pool_map-wait-type-override &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 rtnl_mutex dev_base_lock &xa->xa_lock#3 &obj_hash[i].lock irq_context: 0 rtnl_mutex dev_base_lock &xa->xa_lock#3 pool_lock#2 irq_context: 0 rtnl_mutex fill_pool_map-wait-type-override &rq->__lock &cfs_rq->removed.lock irq_context: 0 rtnl_mutex &block->cb_lock flow_indr_block_lock &cfs_rq->removed.lock irq_context: 0 cb_lock batched_entropy_u8.lock irq_context: 0 cb_lock batched_entropy_u8.lock crngs.lock irq_context: 0 cb_lock kfence_freelist_lock irq_context: 0 rtnl_mutex &idev->mc_lock &bridge_netdev_addr_lock_key &____s->seqcount#2 irq_context: 0 rtnl_mutex &idev->mc_lock &bridge_netdev_addr_lock_key &____s->seqcount irq_context: 0 cb_lock rcu_read_lock rcu_read_lock_bh rcu_read_lock &____s->seqcount#2 irq_context: 0 cb_lock rcu_read_lock rcu_read_lock_bh rcu_read_lock &____s->seqcount irq_context: 0 (wq_completion)wg-crypt-wg0#4 (work_completion)(&peer->transmit_packet_work) &peer->endpoint_lock rcu_read_lock_bh &c->lock irq_context: 0 (wq_completion)events (work_completion)(&(&nsim_dev->trap_data->trap_report_dw)->work) &devlink->lock_key &nsim_trap_data->trap_lock &base->lock irq_context: 0 (wq_completion)events (work_completion)(&(&nsim_dev->trap_data->trap_report_dw)->work) &devlink->lock_key &nsim_trap_data->trap_lock &base->lock &obj_hash[i].lock irq_context: 0 rtnl_mutex &idev->mc_lock rcu_read_lock rcu_node_0 irq_context: 0 sb_writers#4 &type->i_mutex_dir_key#3/1 &u->bindlock irq_context: 0 sb_writers#4 &type->i_mutex_dir_key#3/1 &u->bindlock &net->unx.table.locks[i] irq_context: 0 sb_writers#4 &type->i_mutex_dir_key#3/1 &u->bindlock &net->unx.table.locks[i] &net->unx.table.locks[i]/1 irq_context: 0 sb_writers#4 &type->i_mutex_dir_key#3/1 &u->bindlock &net->unx.table.locks[i] &net->unx.table.locks[i]/1 &dentry->d_lock irq_context: 0 sb_writers#4 &type->i_mutex_dir_key#3/1 &u->bindlock &net->unx.table.locks[i]/1 irq_context: 0 sb_writers#4 &type->i_mutex_dir_key#3/1 &u->bindlock &bsd_socket_locks[i] irq_context: 0 &sb->s_type->i_mutex_key#10 &bsd_socket_locks[i] irq_context: 0 &sb->s_type->i_mutex_key#10 &dentry->d_lock irq_context: 0 &sb->s_type->i_mutex_key#10 &dentry->d_lock &lru->node[i].lock irq_context: 0 sb_writers#5 &type->i_mutex_dir_key#5/1 &rcu_state.expedited_wq irq_context: 0 sb_writers#5 &type->i_mutex_dir_key#5/1 &rcu_state.expedited_wq &p->pi_lock irq_context: 0 sb_writers#5 &type->i_mutex_dir_key#5/1 &rcu_state.expedited_wq &p->pi_lock &rq->__lock irq_context: 0 sb_writers#5 &type->i_mutex_dir_key#5/1 &rcu_state.expedited_wq &p->pi_lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 rtnl_mutex uevent_sock_mutex batched_entropy_u8.lock irq_context: 0 rtnl_mutex uevent_sock_mutex kfence_freelist_lock irq_context: 0 rtnl_mutex uevent_sock_mutex &meta->lock irq_context: 0 (wq_completion)wg-crypt-wg0#4 (work_completion)(&peer->transmit_packet_work) &peer->endpoint_lock rcu_read_lock_bh rcu_read_lock &c->lock irq_context: 0 (wq_completion)wg-crypt-wg0#4 (work_completion)(&peer->transmit_packet_work) &peer->endpoint_lock rcu_read_lock_bh rcu_read_lock &nf_conntrack_locks[i]/1 irq_context: 0 &sb->s_type->i_mutex_key#8 &sbi->s_writepages_rwsem jbd2_handle rcu_read_lock rcu_read_lock per_cpu_ptr(&cgroup_rstat_cpu_lock, cpu) irq_context: 0 cb_lock &meta->lock irq_context: 0 (wq_completion)mld (work_completion)(&(&idev->mc_ifc_work)->work) &idev->mc_lock remove_cache_srcu pool_lock#2 irq_context: 0 pernet_ops_rwsem rcu_read_lock &obj_hash[i].lock pool_lock irq_context: 0 (wq_completion)bat_events (work_completion)(&(&bat_priv->tt.work)->work) &cfs_rq->removed.lock irq_context: 0 (wq_completion)bat_events (work_completion)(&(&bat_priv->tt.work)->work) pool_lock#2 irq_context: softirq (&peer->timer_retransmit_handshake) &peer->endpoint_lock &obj_hash[i].lock irq_context: softirq (&peer->timer_retransmit_handshake) &peer->endpoint_lock pool_lock#2 irq_context: 0 (wq_completion)wg-kex-wg2#9 (work_completion)(&peer->transmit_handshake_work) &peer->endpoint_lock rcu_read_lock_bh &obj_hash[i].lock irq_context: 0 (wq_completion)wg-kex-wg2#9 (work_completion)(&peer->transmit_handshake_work) &peer->endpoint_lock rcu_read_lock_bh pool_lock#2 irq_context: 0 tracepoints_mutex cpu_hotplug_lock static_call_mutex text_mutex &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 sb_writers#4 &sb->s_type->i_mutex_key#8 mapping.invalidate_lock jbd2_handle rcu_read_lock &rq->__lock irq_context: 0 sb_writers#4 &sb->s_type->i_mutex_key#8 mapping.invalidate_lock jbd2_handle rcu_read_lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 rtnl_mutex &rnp->exp_wq[1] irq_context: 0 &f->f_pos_lock sb_writers#4 &sb->s_type->i_mutex_key#8 &mm->mmap_lock &n->list_lock irq_context: 0 &f->f_pos_lock sb_writers#4 &sb->s_type->i_mutex_key#8 &mm->mmap_lock &n->list_lock &c->lock irq_context: 0 pernet_ops_rwsem k-sk_lock-AF_NETLINK &rq->__lock irq_context: 0 pernet_ops_rwsem k-sk_lock-AF_NETLINK &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 &mm->mmap_lock remove_cache_srcu rcu_read_lock rcu_read_lock rcu_node_0 irq_context: 0 &mm->mmap_lock remove_cache_srcu rcu_read_lock rcu_read_lock &rq->__lock irq_context: 0 &mm->mmap_lock remove_cache_srcu rcu_read_lock rcu_read_lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem rcu_read_lock &rq->__lock &cfs_rq->removed.lock irq_context: 0 &f->f_pos_lock sb_writers#4 &sb->s_type->i_mutex_key#8 jbd2_handle &rq->__lock &cfs_rq->removed.lock irq_context: 0 lock prog_idr_lock &n->list_lock irq_context: 0 lock prog_idr_lock &n->list_lock &c->lock irq_context: 0 pernet_ops_rwsem k-sk_lock-AF_RXRPC &rxnet->local_mutex &table->hash[i].lock irq_context: 0 pernet_ops_rwsem k-sk_lock-AF_RXRPC &rxnet->local_mutex &table->hash[i].lock &table->hash2[i].lock irq_context: 0 pernet_ops_rwsem k-sk_lock-AF_RXRPC &rxnet->local_mutex k-clock-AF_INET6 irq_context: 0 pernet_ops_rwsem k-sk_lock-AF_RXRPC &rxnet->local_mutex &xa->xa_lock#9 irq_context: 0 pernet_ops_rwsem k-sk_lock-AF_RXRPC &rxnet->local_mutex &fsnotify_mark_srcu irq_context: 0 pernet_ops_rwsem k-sk_lock-AF_RXRPC &obj_hash[i].lock irq_context: 0 pernet_ops_rwsem k-clock-AF_RXRPC irq_context: 0 pernet_ops_rwsem (wq_completion)krxrpcd irq_context: 0 pernet_ops_rwsem &wq->mutex irq_context: 0 pernet_ops_rwsem &wq->mutex &pool->lock/1 irq_context: 0 pernet_ops_rwsem &wq->mutex &x->wait#10 irq_context: 0 pernet_ops_rwsem rlock-AF_RXRPC irq_context: 0 pernet_ops_rwsem (&net->fs_probe_timer) irq_context: 0 pernet_ops_rwsem &net->cells_lock irq_context: 0 pernet_ops_rwsem (&net->cells_timer) irq_context: 0 pernet_ops_rwsem bit_wait_table + i irq_context: 0 pernet_ops_rwsem (&net->fs_timer) irq_context: 0 &f->f_pos_lock sb_writers#4 &sb->s_type->i_mutex_key#8 rcu_read_lock rcu_read_lock &cfs_rq->removed.lock irq_context: 0 &f->f_pos_lock sb_writers#4 &sb->s_type->i_mutex_key#8 rcu_read_lock rcu_read_lock &obj_hash[i].lock irq_context: 0 (wq_completion)events fqdir_free_work rcu_state.barrier_mutex &pool->lock &p->pi_lock &rq->__lock &cfs_rq->removed.lock irq_context: 0 pernet_ops_rwsem ovs_mutex irq_context: 0 pernet_ops_rwsem ovs_mutex (work_completion)(&data->gc_work) irq_context: 0 pernet_ops_rwsem ovs_mutex nf_ct_proto_mutex irq_context: 0 pernet_ops_rwsem ovs_mutex nf_ct_proto_mutex nf_hook_mutex irq_context: 0 pernet_ops_rwsem ovs_mutex nf_ct_proto_mutex nf_hook_mutex cpu_hotplug_lock irq_context: 0 pernet_ops_rwsem ovs_mutex nf_ct_proto_mutex nf_hook_mutex fs_reclaim irq_context: 0 pernet_ops_rwsem ovs_mutex nf_ct_proto_mutex nf_hook_mutex fs_reclaim mmu_notifier_invalidate_range_start irq_context: 0 pernet_ops_rwsem ovs_mutex nf_ct_proto_mutex nf_hook_mutex stock_lock irq_context: 0 pernet_ops_rwsem ovs_mutex nf_ct_proto_mutex nf_hook_mutex &c->lock irq_context: 0 pernet_ops_rwsem ovs_mutex nf_ct_proto_mutex nf_hook_mutex pool_lock#2 irq_context: 0 pernet_ops_rwsem ovs_mutex nf_ct_proto_mutex &obj_hash[i].lock irq_context: 0 pernet_ops_rwsem ovs_mutex nf_ct_proto_mutex pool_lock#2 irq_context: 0 pernet_ops_rwsem ovs_mutex nf_ct_proto_mutex defrag4_mutex irq_context: 0 pernet_ops_rwsem ovs_mutex nf_ct_proto_mutex defrag4_mutex nf_hook_mutex irq_context: 0 pernet_ops_rwsem ovs_mutex nf_ct_proto_mutex defrag4_mutex nf_hook_mutex cpu_hotplug_lock irq_context: 0 pernet_ops_rwsem ovs_mutex nf_ct_proto_mutex defrag4_mutex &obj_hash[i].lock irq_context: 0 pernet_ops_rwsem ovs_mutex nf_ct_proto_mutex defrag4_mutex pool_lock#2 irq_context: 0 pernet_ops_rwsem ovs_mutex nf_ct_proto_mutex defrag6_mutex irq_context: 0 pernet_ops_rwsem ovs_mutex nf_ct_proto_mutex defrag6_mutex nf_hook_mutex irq_context: 0 pernet_ops_rwsem ovs_mutex nf_ct_proto_mutex defrag6_mutex nf_hook_mutex cpu_hotplug_lock irq_context: 0 pernet_ops_rwsem ovs_mutex nf_ct_proto_mutex defrag6_mutex &obj_hash[i].lock irq_context: 0 pernet_ops_rwsem ovs_mutex nf_ct_proto_mutex defrag6_mutex pool_lock#2 irq_context: 0 pernet_ops_rwsem ovs_mutex &obj_hash[i].lock irq_context: 0 pernet_ops_rwsem ovs_mutex pool_lock#2 irq_context: 0 pernet_ops_rwsem ovs_mutex nf_connlabels_lock irq_context: 0 pernet_ops_rwsem ovs_mutex net_rwsem irq_context: 0 pernet_ops_rwsem (work_completion)(&(&ovs_net->masks_rebalance)->work) irq_context: 0 pernet_ops_rwsem (work_completion)(&ovs_net->dp_notify_work) irq_context: 0 pernet_ops_rwsem &srv->idr_lock irq_context: 0 pernet_ops_rwsem (wq_completion)krdsd irq_context: 0 pernet_ops_rwsem wq_pool_mutex &wq->mutex &pool->lock/1 irq_context: 0 pernet_ops_rwsem &pool->lock/1 irq_context: 0 pernet_ops_rwsem &pool->lock/1 rcu_read_lock &pool->lock irq_context: 0 pernet_ops_rwsem &pool->lock/1 rcu_read_lock &pool->lock &obj_hash[i].lock irq_context: 0 pernet_ops_rwsem &pool->lock/1 rcu_read_lock &pool->lock pool_lock#2 irq_context: 0 pernet_ops_rwsem &pool->lock/1 rcu_read_lock &pool->lock &p->pi_lock irq_context: 0 pernet_ops_rwsem &pool->lock/1 rcu_read_lock &pool->lock &p->pi_lock &rq->__lock irq_context: 0 pernet_ops_rwsem &pool->lock/1 rcu_read_lock &pool->lock &p->pi_lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 pernet_ops_rwsem (&rxnet->peer_keepalive_timer) irq_context: 0 pernet_ops_rwsem k-sk_lock-AF_TIPC &tn->nametbl_lock &service->lock &obj_hash[i].lock irq_context: 0 pernet_ops_rwsem k-sk_lock-AF_TIPC &tn->nametbl_lock &service->lock krc.lock irq_context: 0 pernet_ops_rwsem k-sk_lock-AF_TIPC &tn->nametbl_lock &nt->cluster_scope_lock irq_context: 0 pernet_ops_rwsem k-sk_lock-AF_TIPC &tn->nametbl_lock &obj_hash[i].lock irq_context: 0 pernet_ops_rwsem k-sk_lock-AF_TIPC &tn->nametbl_lock krc.lock irq_context: 0 pernet_ops_rwsem k-sk_lock-AF_TIPC &obj_hash[i].lock irq_context: 0 pernet_ops_rwsem k-sk_lock-AF_TIPC rcu_read_lock rhashtable_bucket irq_context: 0 pernet_ops_rwsem k-sk_lock-AF_TIPC k-clock-AF_TIPC irq_context: 0 pernet_ops_rwsem &rnp->exp_lock irq_context: 0 pernet_ops_rwsem rcu_state.exp_mutex irq_context: 0 pernet_ops_rwsem rcu_state.exp_mutex rcu_state.exp_mutex.wait_lock irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem fill_pool_map-wait-type-override &____s->seqcount irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem fill_pool_map-wait-type-override rcu_read_lock pool_lock#2 irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem fill_pool_map-wait-type-override &obj_hash[i].lock irq_context: 0 pernet_ops_rwsem (work_completion)(&tn->work) irq_context: 0 pernet_ops_rwsem &tn->nametbl_lock irq_context: 0 pernet_ops_rwsem &ht->mutex quarantine_lock irq_context: 0 pernet_ops_rwsem (work_completion)(&(&c->work)->work) irq_context: 0 pernet_ops_rwsem rtnl_mutex pool_lock irq_context: 0 pernet_ops_rwsem (work_completion)(&rtn->rds_tcp_accept_w) irq_context: 0 pernet_ops_rwsem k-sk_lock-AF_INET6 &queue->rskq_lock irq_context: 0 pernet_ops_rwsem k-sk_lock-AF_INET6 k-slock-AF_INET6 &obj_hash[i].lock irq_context: 0 pernet_ops_rwsem k-sk_lock-AF_INET6 k-slock-AF_INET6 &tcp_hashinfo.bhash[i].lock irq_context: 0 pernet_ops_rwsem k-sk_lock-AF_INET6 k-slock-AF_INET6 &tcp_hashinfo.bhash[i].lock stock_lock irq_context: 0 pernet_ops_rwsem k-sk_lock-AF_INET6 k-slock-AF_INET6 &tcp_hashinfo.bhash[i].lock &obj_hash[i].lock irq_context: 0 pernet_ops_rwsem k-sk_lock-AF_INET6 k-slock-AF_INET6 &tcp_hashinfo.bhash[i].lock pool_lock#2 irq_context: 0 pernet_ops_rwsem k-sk_lock-AF_INET6 k-slock-AF_INET6 &tcp_hashinfo.bhash[i].lock &tcp_hashinfo.bhash2[i].lock irq_context: 0 pernet_ops_rwsem k-sk_lock-AF_INET6 k-slock-AF_INET6 &tcp_hashinfo.bhash[i].lock &tcp_hashinfo.bhash2[i].lock stock_lock irq_context: 0 pernet_ops_rwsem k-sk_lock-AF_INET6 k-slock-AF_INET6 &tcp_hashinfo.bhash[i].lock &tcp_hashinfo.bhash2[i].lock &obj_hash[i].lock irq_context: 0 pernet_ops_rwsem k-sk_lock-AF_INET6 k-slock-AF_INET6 &tcp_hashinfo.bhash[i].lock &tcp_hashinfo.bhash2[i].lock pool_lock#2 irq_context: 0 pernet_ops_rwsem k-sk_lock-AF_INET6 k-slock-AF_INET6 elock-AF_INET6 irq_context: 0 pernet_ops_rwsem rds_tcp_conn_lock irq_context: 0 pernet_ops_rwsem loop_conns_lock irq_context: 0 pernet_ops_rwsem (wq_completion)l2tp irq_context: 0 &mm->mmap_lock fill_pool_map-wait-type-override pool_lock#2 irq_context: 0 &mm->mmap_lock fill_pool_map-wait-type-override pool_lock irq_context: 0 sb_writers#4 fill_pool_map-wait-type-override &c->lock irq_context: 0 pernet_ops_rwsem (work_completion)(&rxnet->peer_keepalive_work) irq_context: 0 pernet_ops_rwsem rcu_read_lock &pool->lock/1 irq_context: 0 pernet_ops_rwsem (&rxnet->service_conn_reap_timer) irq_context: 0 pernet_ops_rwsem rcu_read_lock &pool->lock/1 &obj_hash[i].lock irq_context: 0 pernet_ops_rwsem rcu_read_lock &pool->lock/1 &p->pi_lock irq_context: 0 pernet_ops_rwsem rcu_read_lock &pool->lock/1 &p->pi_lock &rq->__lock irq_context: 0 pernet_ops_rwsem rcu_read_lock &pool->lock/1 &p->pi_lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 pernet_ops_rwsem &x->wait#10 irq_context: 0 (wq_completion)wg-crypt-wg0#2 (work_completion)(&peer->transmit_packet_work) rcu_node_0 irq_context: 0 pernet_ops_rwsem rtnl_mutex _xmit_NONE irq_context: 0 pernet_ops_rwsem &dir->lock#2 quarantine_lock irq_context: 0 (wq_completion)events (work_completion)(&p->wq) rcu_read_lock &rcu_state.expedited_wq &p->pi_lock irq_context: 0 (wq_completion)events (work_completion)(&p->wq) rcu_read_lock &rcu_state.expedited_wq &p->pi_lock &rq->__lock irq_context: 0 (wq_completion)events (work_completion)(&p->wq) rcu_read_lock &rcu_state.expedited_wq &p->pi_lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 sb_writers#4 &type->i_mutex_dir_key#3/1 jbd2_handle remove_cache_srcu pool_lock#2 irq_context: 0 sb_writers#4 &type->i_mutex_dir_key#3/1 jbd2_handle remove_cache_srcu &rq->__lock irq_context: 0 sb_writers#4 &type->i_mutex_dir_key#3/1 jbd2_handle remove_cache_srcu &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)bat_events (work_completion)(&(&forw_packet_aggr->delayed_work)->work) &rcu_state.expedited_wq &p->pi_lock &rq->__lock irq_context: 0 (wq_completion)mld (work_completion)(&(&idev->mc_dad_work)->work) &idev->mc_lock &rq->__lock irq_context: 0 (wq_completion)wg-crypt-wg0#4 (work_completion)(&peer->transmit_packet_work) &peer->endpoint_lock rcu_read_lock_bh rcu_read_lock rcu_read_lock_bh quarantine_lock irq_context: 0 (wq_completion)wg-crypt-wg0#4 (work_completion)(&peer->transmit_packet_work) &peer->endpoint_lock rcu_read_lock_bh &meta->lock irq_context: 0 (wq_completion)wg-crypt-wg0#4 (work_completion)(&peer->transmit_packet_work) &peer->endpoint_lock rcu_read_lock_bh kfence_freelist_lock irq_context: 0 (wq_completion)inet_frag_wq (work_completion)(&fqdir->destroy_work) rcu_read_lock &pool->lock &p->pi_lock &rq->__lock irq_context: 0 (wq_completion)inet_frag_wq (work_completion)(&fqdir->destroy_work) rcu_read_lock &pool->lock &p->pi_lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 sb_writers#10 &type->i_mutex_dir_key#6/1 tomoyo_ss &n->list_lock irq_context: 0 sb_writers#10 &type->i_mutex_dir_key#6/1 tomoyo_ss &n->list_lock &c->lock irq_context: 0 sb_writers#10 &type->i_mutex_dir_key#6/1 tomoyo_ss &rq->__lock irq_context: 0 sb_writers#10 &type->i_mutex_dir_key#6/1 tomoyo_ss &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 &type->i_mutex_dir_key#6 &c->lock irq_context: 0 &type->i_mutex_dir_key#6 &rq->__lock irq_context: 0 &type->i_mutex_dir_key#6 &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 &type->i_mutex_dir_key#6 &root->kernfs_rwsem pool_lock#2 irq_context: 0 &type->i_mutex_dir_key#6 &root->kernfs_rwsem &rq->__lock irq_context: 0 &type->i_mutex_dir_key#6 &root->kernfs_rwsem &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 &type->i_mutex_dir_key#6 &root->kernfs_rwsem &xa->xa_lock#4 irq_context: 0 &type->i_mutex_dir_key#6 &root->kernfs_rwsem &xa->xa_lock#4 pool_lock#2 irq_context: 0 &type->i_mutex_dir_key#6 &root->kernfs_rwsem &obj_hash[i].lock irq_context: 0 &type->i_mutex_dir_key#6 &root->kernfs_rwsem stock_lock irq_context: 0 rcu_read_lock rcu_read_lock_bh &c->lock irq_context: 0 rtnl_mutex gdp_mutex &rq->__lock irq_context: 0 rtnl_mutex gdp_mutex &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 sb_writers#4 &sb->s_type->i_mutex_key#8 ima_extend_list_mutex &rq->__lock irq_context: 0 sb_writers#4 &sb->s_type->i_mutex_key#8 ima_extend_list_mutex &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)gid-cache-wq (work_completion)(&ndev_work->work) &obj_hash[i].lock pool_lock irq_context: 0 &sb->s_type->i_mutex_key#10 sk_lock-AF_CAN &rnp->exp_lock irq_context: 0 &sb->s_type->i_mutex_key#10 sk_lock-AF_CAN rcu_state.exp_mutex irq_context: 0 &sb->s_type->i_mutex_key#10 sk_lock-AF_CAN rcu_state.exp_mutex rcu_state.exp_mutex.wait_lock irq_context: 0 &sighand->siglock &____s->seqcount#2 irq_context: 0 &sighand->siglock &____s->seqcount irq_context: 0 cb_lock rcu_read_lock rcu_read_lock_bh rcu_read_lock &n->list_lock irq_context: 0 cb_lock rcu_read_lock rcu_read_lock_bh rcu_read_lock &n->list_lock &c->lock irq_context: 0 (wq_completion)wg-crypt-wg1 (work_completion)(&peer->transmit_packet_work) &peer->endpoint_lock rcu_read_lock_bh rcu_read_lock &obj_hash[i].lock irq_context: 0 (wq_completion)wg-crypt-wg1 (work_completion)(&peer->transmit_packet_work) &peer->endpoint_lock rcu_read_lock_bh rcu_read_lock rcu_read_lock_bh rcu_read_lock tk_core.seq.seqcount irq_context: 0 (wq_completion)wg-crypt-wg1 (work_completion)(&peer->transmit_packet_work) &peer->endpoint_lock rcu_read_lock_bh rcu_read_lock rcu_read_lock_bh rcu_read_lock &obj_hash[i].lock irq_context: 0 &sighand->siglock batched_entropy_u8.lock irq_context: 0 rcu_read_lock rcu_read_lock_bh &obj_hash[i].lock pool_lock irq_context: 0 sb_writers#3 &sb->s_type->i_mutex_key#9 remove_cache_srcu irq_context: 0 sb_writers#3 &sb->s_type->i_mutex_key#9 remove_cache_srcu quarantine_lock irq_context: 0 sb_writers#3 &sb->s_type->i_mutex_key#9 remove_cache_srcu &c->lock irq_context: 0 sb_writers#3 &sb->s_type->i_mutex_key#9 remove_cache_srcu &n->list_lock irq_context: 0 sb_writers#3 &sb->s_type->i_mutex_key#9 remove_cache_srcu rcu_read_lock pool_lock#2 irq_context: 0 sb_writers#3 &sb->s_type->i_mutex_key#9 remove_cache_srcu &obj_hash[i].lock irq_context: 0 sb_writers#3 &sb->s_type->i_mutex_key#9 remove_cache_srcu rcu_read_lock rcu_node_0 irq_context: 0 sb_writers#3 &sb->s_type->i_mutex_key#9 remove_cache_srcu rcu_read_lock &rq->__lock irq_context: 0 sb_writers#3 &sb->s_type->i_mutex_key#9 remove_cache_srcu rcu_read_lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 &mm->mmap_lock rcu_read_lock per_cpu_ptr(&cgroup_rstat_cpu_lock, cpu) irq_context: 0 &mm->mmap_lock ptlock_ptr(page)#2 rcu_read_lock pool_lock#2 irq_context: 0 &mm->mmap_lock ptlock_ptr(page)#2 &obj_hash[i].lock irq_context: 0 rlock-AF_INET irq_context: 0 sk_lock-AF_INET rcu_read_lock &____s->seqcount#2 irq_context: 0 sk_lock-AF_INET rcu_read_lock &____s->seqcount irq_context: softirq rcu_read_lock rcu_read_lock rcu_read_lock rcu_read_lock &ei->socket.wq.wait irq_context: softirq rcu_read_lock rcu_read_lock rcu_read_lock rcu_read_lock &ei->socket.wq.wait &p->pi_lock irq_context: softirq rcu_read_lock rcu_read_lock rcu_read_lock rcu_read_lock &ei->socket.wq.wait &p->pi_lock &rq->__lock irq_context: softirq rcu_read_lock rcu_read_lock rcu_read_lock rcu_read_lock &ei->socket.wq.wait &p->pi_lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: softirq rcu_read_lock rcu_read_lock k-slock-AF_INET rcu_read_lock rcu_read_lock_bh rcu_read_lock pool_lock#2 irq_context: softirq rcu_read_lock rcu_read_lock k-slock-AF_INET rcu_read_lock rcu_read_lock_bh rcu_read_lock tk_core.seq.seqcount irq_context: softirq rcu_read_lock rcu_read_lock k-slock-AF_INET rcu_read_lock rcu_read_lock_bh rcu_read_lock rlock-AF_PACKET irq_context: softirq rcu_read_lock rcu_read_lock k-slock-AF_INET rcu_read_lock rcu_read_lock_bh rcu_read_lock &c->lock irq_context: softirq (&icsk->icsk_delack_timer) slock-AF_INET &n->list_lock irq_context: softirq (&icsk->icsk_delack_timer) slock-AF_INET &n->list_lock &c->lock irq_context: 0 &sb->s_type->i_mutex_key#10 sk_lock-AF_PACKET &rq->__lock irq_context: 0 &sb->s_type->i_mutex_key#10 sk_lock-AF_PACKET &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 &sb->s_type->i_mutex_key#10 sk_lock-AF_CAN &rq->__lock irq_context: 0 sb_writers#10 &type->i_mutex_dir_key#6/1 rcu_read_lock rcu_node_0 irq_context: 0 sb_writers#10 &type->i_mutex_dir_key#6/1 rcu_read_lock &rq->__lock irq_context: 0 sb_writers#10 &type->i_mutex_dir_key#6/1 rcu_read_lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 sb_writers#10 &rq->__lock irq_context: 0 sb_writers#10 &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 &rdma_nl_types[idx].sem link_ops_rwsem irq_context: 0 &rdma_nl_types[idx].sem link_ops_rwsem fs_reclaim irq_context: 0 &rdma_nl_types[idx].sem link_ops_rwsem fs_reclaim mmu_notifier_invalidate_range_start irq_context: 0 &rdma_nl_types[idx].sem link_ops_rwsem &c->lock irq_context: 0 &rdma_nl_types[idx].sem link_ops_rwsem pool_lock#2 irq_context: 0 &rdma_nl_types[idx].sem link_ops_rwsem &x->wait#9 irq_context: 0 &rdma_nl_types[idx].sem link_ops_rwsem &obj_hash[i].lock irq_context: 0 &rdma_nl_types[idx].sem link_ops_rwsem (console_sem).lock irq_context: 0 &rdma_nl_types[idx].sem link_ops_rwsem &pdata->netdev_lock irq_context: 0 &rdma_nl_types[idx].sem link_ops_rwsem &pdata->netdev_lock pool_lock#2 irq_context: 0 &rdma_nl_types[idx].sem link_ops_rwsem &pdata->netdev_lock &dir->lock#2 irq_context: 0 &rdma_nl_types[idx].sem link_ops_rwsem ndev_hash_lock irq_context: 0 &rdma_nl_types[idx].sem link_ops_rwsem crypto_alg_sem irq_context: 0 &rdma_nl_types[idx].sem link_ops_rwsem devices_rwsem irq_context: 0 &rdma_nl_types[idx].sem link_ops_rwsem devices_rwsem fs_reclaim irq_context: 0 &rdma_nl_types[idx].sem link_ops_rwsem devices_rwsem fs_reclaim mmu_notifier_invalidate_range_start irq_context: 0 &rdma_nl_types[idx].sem link_ops_rwsem devices_rwsem pool_lock#2 irq_context: 0 &rdma_nl_types[idx].sem link_ops_rwsem devices_rwsem devices.xa_lock irq_context: 0 &rdma_nl_types[idx].sem link_ops_rwsem &rxe->usdev_lock irq_context: 0 &rdma_nl_types[idx].sem link_ops_rwsem &rxe->usdev_lock &rq->__lock irq_context: 0 &rdma_nl_types[idx].sem link_ops_rwsem &rxe->usdev_lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 &rdma_nl_types[idx].sem link_ops_rwsem &rxe->usdev_lock &pdata->netdev_lock irq_context: 0 &rdma_nl_types[idx].sem link_ops_rwsem &rxe->usdev_lock rtnl_mutex irq_context: 0 &rdma_nl_types[idx].sem link_ops_rwsem rtnl_mutex irq_context: 0 &rdma_nl_types[idx].sem link_ops_rwsem rtnl_mutex net_rwsem irq_context: 0 &rdma_nl_types[idx].sem link_ops_rwsem rtnl_mutex net_rwsem &rq->__lock irq_context: 0 &rdma_nl_types[idx].sem link_ops_rwsem rtnl_mutex net_rwsem &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 &rdma_nl_types[idx].sem link_ops_rwsem rtnl_mutex net_rwsem &table->lock#4 irq_context: 0 &rdma_nl_types[idx].sem link_ops_rwsem rtnl_mutex net_rwsem &table->lock#4 fs_reclaim irq_context: 0 &rdma_nl_types[idx].sem link_ops_rwsem rtnl_mutex net_rwsem &table->lock#4 fs_reclaim mmu_notifier_invalidate_range_start irq_context: 0 &rdma_nl_types[idx].sem link_ops_rwsem rtnl_mutex net_rwsem &table->lock#4 fs_reclaim mmu_notifier_invalidate_range_start &rq->__lock irq_context: 0 &rdma_nl_types[idx].sem link_ops_rwsem rtnl_mutex net_rwsem &table->lock#4 fs_reclaim mmu_notifier_invalidate_range_start &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 &rdma_nl_types[idx].sem link_ops_rwsem rtnl_mutex net_rwsem &table->lock#4 pool_lock#2 irq_context: 0 &rdma_nl_types[idx].sem link_ops_rwsem rtnl_mutex net_rwsem &table->lock#4 &obj_hash[i].lock irq_context: 0 &rdma_nl_types[idx].sem link_ops_rwsem rtnl_mutex net_rwsem &table->lock#4 &table->rwlock irq_context: 0 &rdma_nl_types[idx].sem link_ops_rwsem rtnl_mutex net_rwsem &table->lock#4 &device->event_handler_rwsem irq_context: 0 &rdma_nl_types[idx].sem link_ops_rwsem rtnl_mutex net_rwsem rcu_read_lock pool_lock#2 irq_context: 0 &rdma_nl_types[idx].sem link_ops_rwsem rtnl_mutex net_rwsem &obj_hash[i].lock irq_context: 0 &rdma_nl_types[idx].sem link_ops_rwsem rtnl_mutex net_rwsem pool_lock#2 irq_context: 0 &rdma_nl_types[idx].sem link_ops_rwsem rtnl_mutex net_rwsem &ndev->lock irq_context: 0 &rdma_nl_types[idx].sem link_ops_rwsem rtnl_mutex net_rwsem &ndev->lock pool_lock#2 irq_context: 0 &rdma_nl_types[idx].sem link_ops_rwsem rtnl_mutex net_rwsem &ndev->lock &c->lock irq_context: 0 &rdma_nl_types[idx].sem link_ops_rwsem rtnl_mutex net_rwsem rcu_read_lock rcu_node_0 irq_context: 0 &rdma_nl_types[idx].sem link_ops_rwsem rtnl_mutex net_rwsem rcu_read_lock &rq->__lock irq_context: 0 &rdma_nl_types[idx].sem link_ops_rwsem rtnl_mutex net_rwsem rcu_read_lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 &rdma_nl_types[idx].sem link_ops_rwsem rtnl_mutex net_rwsem rcu_read_lock &rcu_state.gp_wq irq_context: 0 &rdma_nl_types[idx].sem link_ops_rwsem rtnl_mutex net_rwsem rcu_read_lock &rcu_state.gp_wq &p->pi_lock irq_context: 0 &rdma_nl_types[idx].sem link_ops_rwsem rtnl_mutex net_rwsem rcu_read_lock &rcu_state.gp_wq &p->pi_lock &rq->__lock irq_context: 0 &rdma_nl_types[idx].sem link_ops_rwsem rtnl_mutex net_rwsem rcu_read_lock &rcu_state.gp_wq &p->pi_lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 &rdma_nl_types[idx].sem link_ops_rwsem &device->cache_lock irq_context: 0 &rdma_nl_types[idx].sem link_ops_rwsem rdmacg_mutex irq_context: 0 &rdma_nl_types[idx].sem link_ops_rwsem &k->list_lock irq_context: 0 &rdma_nl_types[idx].sem link_ops_rwsem gdp_mutex irq_context: 0 &rdma_nl_types[idx].sem link_ops_rwsem gdp_mutex &k->list_lock irq_context: 0 &rdma_nl_types[idx].sem link_ops_rwsem gdp_mutex fs_reclaim irq_context: 0 &rdma_nl_types[idx].sem link_ops_rwsem gdp_mutex fs_reclaim mmu_notifier_invalidate_range_start irq_context: 0 &rdma_nl_types[idx].sem link_ops_rwsem gdp_mutex &c->lock irq_context: 0 &rdma_nl_types[idx].sem link_ops_rwsem gdp_mutex pool_lock#2 irq_context: 0 &rdma_nl_types[idx].sem link_ops_rwsem gdp_mutex lock irq_context: 0 &rdma_nl_types[idx].sem link_ops_rwsem gdp_mutex lock kernfs_idr_lock irq_context: 0 &rdma_nl_types[idx].sem link_ops_rwsem gdp_mutex &root->kernfs_rwsem irq_context: 0 &rdma_nl_types[idx].sem link_ops_rwsem gdp_mutex &root->kernfs_rwsem &root->kernfs_iattr_rwsem irq_context: 0 &rdma_nl_types[idx].sem link_ops_rwsem gdp_mutex kobj_ns_type_lock irq_context: 0 &rdma_nl_types[idx].sem link_ops_rwsem lock irq_context: 0 &rdma_nl_types[idx].sem link_ops_rwsem lock kernfs_idr_lock irq_context: 0 &rdma_nl_types[idx].sem link_ops_rwsem &root->kernfs_rwsem irq_context: 0 &rdma_nl_types[idx].sem link_ops_rwsem &root->kernfs_rwsem &root->kernfs_iattr_rwsem irq_context: 0 &rdma_nl_types[idx].sem link_ops_rwsem bus_type_sem irq_context: 0 &rdma_nl_types[idx].sem link_ops_rwsem sysfs_symlink_target_lock irq_context: 0 &rdma_nl_types[idx].sem link_ops_rwsem &root->kernfs_rwsem irq_context: 0 &rdma_nl_types[idx].sem link_ops_rwsem &dev->power.lock irq_context: 0 &rdma_nl_types[idx].sem link_ops_rwsem dpm_list_mtx irq_context: 0 &rdma_nl_types[idx].sem link_ops_rwsem subsys mutex#84 irq_context: 0 &rdma_nl_types[idx].sem link_ops_rwsem subsys mutex#84 &k->k_lock irq_context: 0 &rdma_nl_types[idx].sem link_ops_rwsem &zone->lock irq_context: 0 &rdma_nl_types[idx].sem link_ops_rwsem &____s->seqcount irq_context: 0 &rdma_nl_types[idx].sem link_ops_rwsem rcu_read_lock pool_lock#2 irq_context: 0 &rdma_nl_types[idx].sem link_ops_rwsem lock kernfs_idr_lock &c->lock irq_context: 0 &rdma_nl_types[idx].sem link_ops_rwsem lock kernfs_idr_lock pool_lock#2 irq_context: 0 &rdma_nl_types[idx].sem link_ops_rwsem &n->list_lock irq_context: 0 &rdma_nl_types[idx].sem link_ops_rwsem &n->list_lock &c->lock irq_context: 0 &rdma_nl_types[idx].sem link_ops_rwsem &rq->__lock irq_context: 0 &rdma_nl_types[idx].sem link_ops_rwsem &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 &rdma_nl_types[idx].sem link_ops_rwsem &root->kernfs_rwsem &rq->__lock irq_context: 0 &rdma_nl_types[idx].sem link_ops_rwsem &root->kernfs_rwsem &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 &rdma_nl_types[idx].sem link_ops_rwsem fs_reclaim &rq->__lock irq_context: 0 &rdma_nl_types[idx].sem link_ops_rwsem fs_reclaim &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 &rdma_nl_types[idx].sem link_ops_rwsem rcu_read_lock rcu_node_0 irq_context: 0 &rdma_nl_types[idx].sem link_ops_rwsem rcu_read_lock &rq->__lock irq_context: 0 &rdma_nl_types[idx].sem link_ops_rwsem lock kernfs_idr_lock &n->list_lock irq_context: 0 &rdma_nl_types[idx].sem link_ops_rwsem lock kernfs_idr_lock &n->list_lock &c->lock irq_context: 0 &f->f_pos_lock sb_writers#4 &cfs_rq->removed.lock irq_context: 0 &f->f_pos_lock sb_writers#4 &obj_hash[i].lock irq_context: 0 &f->f_pos_lock sb_writers#4 pool_lock#2 irq_context: 0 &rdma_nl_types[idx].sem link_ops_rwsem &____s->seqcount#2 irq_context: 0 &rdma_nl_types[idx].sem link_ops_rwsem &sem->wait_lock irq_context: 0 &rdma_nl_types[idx].sem link_ops_rwsem &p->pi_lock irq_context: 0 &rdma_nl_types[idx].sem link_ops_rwsem &p->pi_lock &rq->__lock irq_context: 0 &rdma_nl_types[idx].sem link_ops_rwsem &p->pi_lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 &rdma_nl_types[idx].sem link_ops_rwsem fs_reclaim mmu_notifier_invalidate_range_start &rq->__lock irq_context: 0 &rdma_nl_types[idx].sem link_ops_rwsem &root->kernfs_rwsem &root->kernfs_iattr_rwsem &rq->__lock irq_context: 0 &rdma_nl_types[idx].sem link_ops_rwsem &root->kernfs_rwsem &root->kernfs_iattr_rwsem &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 &rdma_nl_types[idx].sem link_ops_rwsem &root->kernfs_rwsem &sem->wait_lock irq_context: 0 &rdma_nl_types[idx].sem link_ops_rwsem &pcp->lock &zone->lock irq_context: 0 &iint->mutex &p->alloc_lock irq_context: 0 &iint->mutex &list->lock irq_context: 0 &iint->mutex kauditd_wait.lock irq_context: 0 &f->f_pos_lock sb_writers#4 &sb->s_type->i_mutex_key#8 &ei->i_data_sem pool_lock#2 irq_context: 0 sb_writers#4 &sb->s_type->i_mutex_key#8 jbd2_handle &ei->i_data_sem &lg->lg_mutex irq_context: 0 sb_writers#4 &sb->s_type->i_mutex_key#8 jbd2_handle &ei->i_data_sem &lg->lg_mutex &ei->i_prealloc_lock irq_context: 0 sb_writers#4 &sb->s_type->i_mutex_key#8 jbd2_handle &ei->i_data_sem &lg->lg_mutex mmu_notifier_invalidate_range_start irq_context: 0 sb_writers#4 &sb->s_type->i_mutex_key#8 jbd2_handle &ei->i_data_sem &lg->lg_mutex pool_lock#2 irq_context: 0 sb_writers#4 &sb->s_type->i_mutex_key#8 jbd2_handle &ei->i_data_sem &lg->lg_mutex rcu_read_lock rcu_node_0 irq_context: 0 sb_writers#4 &sb->s_type->i_mutex_key#8 jbd2_handle &ei->i_data_sem &lg->lg_mutex rcu_read_lock &rq->__lock irq_context: 0 sb_writers#4 &sb->s_type->i_mutex_key#8 jbd2_handle &ei->i_data_sem &lg->lg_mutex &mapping->private_lock irq_context: 0 sb_writers#4 &sb->s_type->i_mutex_key#8 jbd2_handle &ei->i_data_sem &lg->lg_mutex &ret->b_state_lock irq_context: 0 sb_writers#4 &sb->s_type->i_mutex_key#8 jbd2_handle &ei->i_data_sem &lg->lg_mutex &ret->b_state_lock &journal->j_list_lock irq_context: 0 sb_writers#4 &sb->s_type->i_mutex_key#8 jbd2_handle &ei->i_data_sem &lg->lg_mutex &pa->pa_lock irq_context: 0 sb_writers#4 &sb->s_type->i_mutex_key#8 jbd2_handle &ei->i_data_sem &lg->lg_mutex &lg->lg_prealloc_lock irq_context: 0 sb_writers#4 &sb->s_type->i_mutex_key#8 jbd2_handle &ei->i_data_sem &ei->i_es_lock &sbi->s_es_lock irq_context: 0 sb_writers#4 &sb->s_type->i_mutex_key#8 &mm->mmap_lock irq_context: 0 sb_writers#4 &sb->s_type->i_mutex_key#8 &mm->mmap_lock ptlock_ptr(page)#2 irq_context: 0 sb_writers#4 &sb->s_type->i_mutex_key#8 jbd2_handle &ei->i_data_sem &lg->lg_mutex rcu_read_lock &pa->pa_lock irq_context: 0 sb_writers#4 &sb->s_type->i_mutex_key#8 jbd2_handle &ei->i_data_sem &lg->lg_mutex rcu_read_lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 sb_writers#4 &sb->s_type->i_mutex_key#8 jbd2_handle &ei->i_data_sem &lg->lg_mutex rcu_read_lock &rcu_state.gp_wq irq_context: 0 sb_writers#4 &sb->s_type->i_mutex_key#8 jbd2_handle &ei->i_data_sem &lg->lg_mutex rcu_read_lock &rcu_state.gp_wq &p->pi_lock irq_context: 0 sb_writers#4 &sb->s_type->i_mutex_key#8 jbd2_handle &ei->i_data_sem &lg->lg_mutex rcu_read_lock &rcu_state.gp_wq &p->pi_lock &rq->__lock irq_context: 0 sb_writers#4 &sb->s_type->i_mutex_key#8 jbd2_handle &ei->i_data_sem &lg->lg_mutex rcu_read_lock &rcu_state.gp_wq &p->pi_lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 sb_writers#4 &sb->s_type->i_mutex_key#8 &mm->mmap_lock &rq->__lock irq_context: 0 sb_writers#4 &sb->s_type->i_mutex_key#8 &mm->mmap_lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 &f->f_pos_lock sb_writers#4 &sb->s_type->i_lock_key#22 irq_context: 0 &f->f_pos_lock sb_writers#4 &sbi->s_writepages_rwsem irq_context: 0 &f->f_pos_lock sb_writers#4 &sbi->s_writepages_rwsem &rq->__lock irq_context: 0 &f->f_pos_lock sb_writers#4 &sbi->s_writepages_rwsem &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 &f->f_pos_lock sb_writers#4 &sbi->s_writepages_rwsem &xa->xa_lock#9 irq_context: 0 &f->f_pos_lock sb_writers#4 &sbi->s_writepages_rwsem mmu_notifier_invalidate_range_start irq_context: 0 &f->f_pos_lock sb_writers#4 &sbi->s_writepages_rwsem &c->lock irq_context: 0 &f->f_pos_lock sb_writers#4 &sbi->s_writepages_rwsem pool_lock#2 irq_context: 0 &f->f_pos_lock sb_writers#4 &sbi->s_writepages_rwsem rcu_read_lock rcu_node_0 irq_context: 0 &f->f_pos_lock sb_writers#4 &sbi->s_writepages_rwsem rcu_read_lock &rq->__lock irq_context: 0 &f->f_pos_lock sb_writers#4 &sbi->s_writepages_rwsem rcu_read_lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 &f->f_pos_lock sb_writers#4 &sbi->s_writepages_rwsem lock#4 irq_context: 0 &f->f_pos_lock sb_writers#4 &sbi->s_writepages_rwsem lock#4 &lruvec->lru_lock irq_context: 0 &f->f_pos_lock sb_writers#4 &sbi->s_writepages_rwsem lock#5 irq_context: 0 &f->f_pos_lock sb_writers#4 &sbi->s_writepages_rwsem &obj_hash[i].lock irq_context: 0 &f->f_pos_lock sb_writers#4 &sbi->s_writepages_rwsem &journal->j_state_lock irq_context: 0 &f->f_pos_lock sb_writers#4 &sbi->s_writepages_rwsem jbd2_handle irq_context: 0 &f->f_pos_lock sb_writers#4 &sbi->s_writepages_rwsem jbd2_handle &rq->__lock irq_context: 0 &f->f_pos_lock sb_writers#4 &sbi->s_writepages_rwsem jbd2_handle &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 &f->f_pos_lock sb_writers#4 &sbi->s_writepages_rwsem jbd2_handle lock#4 irq_context: 0 &f->f_pos_lock sb_writers#4 &sbi->s_writepages_rwsem jbd2_handle lock#5 irq_context: 0 &f->f_pos_lock sb_writers#4 &sbi->s_writepages_rwsem jbd2_handle rcu_read_lock rcu_node_0 irq_context: 0 &f->f_pos_lock sb_writers#4 &sbi->s_writepages_rwsem jbd2_handle rcu_read_lock &rq->__lock irq_context: 0 &f->f_pos_lock sb_writers#4 &sbi->s_writepages_rwsem jbd2_handle rcu_read_lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 &f->f_pos_lock sb_writers#4 &sbi->s_writepages_rwsem jbd2_handle mmu_notifier_invalidate_range_start irq_context: 0 &f->f_pos_lock sb_writers#4 &sbi->s_writepages_rwsem jbd2_handle mmu_notifier_invalidate_range_start &rq->__lock irq_context: 0 &f->f_pos_lock sb_writers#4 &sbi->s_writepages_rwsem jbd2_handle mmu_notifier_invalidate_range_start &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 &f->f_pos_lock sb_writers#4 &sbi->s_writepages_rwsem jbd2_handle pool_lock#2 irq_context: 0 &f->f_pos_lock sb_writers#4 &sbi->s_writepages_rwsem jbd2_handle &ei->i_es_lock irq_context: 0 &f->f_pos_lock sb_writers#4 &sbi->s_writepages_rwsem jbd2_handle &ei->i_data_sem irq_context: 0 &f->f_pos_lock sb_writers#4 &sbi->s_writepages_rwsem jbd2_handle &ei->i_data_sem mmu_notifier_invalidate_range_start irq_context: 0 &f->f_pos_lock sb_writers#4 &sbi->s_writepages_rwsem jbd2_handle &ei->i_data_sem remove_cache_srcu irq_context: 0 &f->f_pos_lock sb_writers#4 &sbi->s_writepages_rwsem jbd2_handle &ei->i_data_sem remove_cache_srcu quarantine_lock irq_context: 0 &f->f_pos_lock sb_writers#4 &sbi->s_writepages_rwsem jbd2_handle &ei->i_data_sem remove_cache_srcu &c->lock irq_context: 0 &f->f_pos_lock sb_writers#4 &sbi->s_writepages_rwsem jbd2_handle &ei->i_data_sem remove_cache_srcu &n->list_lock irq_context: 0 &f->f_pos_lock sb_writers#4 &sbi->s_writepages_rwsem jbd2_handle &ei->i_data_sem remove_cache_srcu rcu_read_lock pool_lock#2 irq_context: 0 &f->f_pos_lock sb_writers#4 &sbi->s_writepages_rwsem jbd2_handle &ei->i_data_sem remove_cache_srcu &obj_hash[i].lock irq_context: 0 &f->f_pos_lock sb_writers#4 &sbi->s_writepages_rwsem jbd2_handle &ei->i_data_sem pool_lock#2 irq_context: 0 &f->f_pos_lock sb_writers#4 &sbi->s_writepages_rwsem jbd2_handle &ei->i_data_sem &ei->i_prealloc_lock irq_context: 0 &f->f_pos_lock sb_writers#4 &sbi->s_writepages_rwsem jbd2_handle &ei->i_data_sem &sbi->s_md_lock irq_context: 0 &f->f_pos_lock sb_writers#4 &sbi->s_writepages_rwsem jbd2_handle &ei->i_data_sem &bgl->locks[i].lock &sbi->s_md_lock irq_context: 0 &f->f_pos_lock sb_writers#4 &sbi->s_writepages_rwsem jbd2_handle &ei->i_data_sem &bgl->locks[i].lock &ei->i_prealloc_lock irq_context: 0 &f->f_pos_lock sb_writers#4 &sbi->s_writepages_rwsem jbd2_handle &ei->i_data_sem key#3 irq_context: 0 &f->f_pos_lock sb_writers#4 &sbi->s_writepages_rwsem jbd2_handle &ei->i_data_sem &pa->pa_lock#2 irq_context: 0 &f->f_pos_lock sb_writers#4 &sbi->s_writepages_rwsem jbd2_handle &ei->i_data_sem &obj_hash[i].lock irq_context: 0 &f->f_pos_lock sb_writers#4 &sbi->s_writepages_rwsem jbd2_handle &ei->i_data_sem &____s->seqcount irq_context: 0 &f->f_pos_lock sb_writers#4 &sbi->s_writepages_rwsem jbd2_handle &ei->i_data_sem rcu_read_lock pool_lock#2 irq_context: 0 &f->f_pos_lock sb_writers#4 &sbi->s_writepages_rwsem jbd2_handle &ei->i_data_sem &ei->i_raw_lock irq_context: 0 &f->f_pos_lock sb_writers#4 &sbi->s_writepages_rwsem jbd2_handle &ei->i_data_sem &(ei->i_block_reservation_lock) irq_context: 0 &f->f_pos_lock sb_writers#4 &sbi->s_writepages_rwsem jbd2_handle &ei->i_data_sem &(ei->i_block_reservation_lock) key#14 irq_context: 0 &f->f_pos_lock sb_writers#4 &sbi->s_writepages_rwsem jbd2_handle &ei->i_data_sem &sb->s_type->i_lock_key#22 irq_context: 0 &f->f_pos_lock sb_writers#4 &sbi->s_writepages_rwsem jbd2_handle &ei->i_data_sem &ei->i_es_lock irq_context: 0 &f->f_pos_lock sb_writers#4 &sbi->s_writepages_rwsem jbd2_handle &ei->i_data_sem &ei->i_es_lock irq_context: 0 &f->f_pos_lock sb_writers#4 &sbi->s_writepages_rwsem jbd2_handle &ei->i_data_sem &ei->i_es_lock &obj_hash[i].lock irq_context: 0 &f->f_pos_lock sb_writers#4 &sbi->s_writepages_rwsem jbd2_handle &ei->i_data_sem &ei->i_es_lock pool_lock#2 irq_context: 0 &f->f_pos_lock sb_writers#4 &sbi->s_writepages_rwsem jbd2_handle rcu_read_lock &memcg->move_lock irq_context: 0 &f->f_pos_lock sb_writers#4 &sbi->s_writepages_rwsem jbd2_handle rcu_read_lock &xa->xa_lock#9 irq_context: 0 &f->f_pos_lock sb_writers#4 &sbi->s_writepages_rwsem jbd2_handle rcu_read_lock &xa->xa_lock#9 &s->s_inode_wblist_lock irq_context: 0 &f->f_pos_lock sb_writers#4 &sbi->s_writepages_rwsem jbd2_handle rcu_read_lock key#10 irq_context: 0 &f->f_pos_lock sb_writers#4 &sbi->s_writepages_rwsem jbd2_handle rcu_read_lock &xa->xa_lock#9 key#10 irq_context: 0 &f->f_pos_lock sb_writers#4 &sbi->s_writepages_rwsem jbd2_handle &ei->i_raw_lock irq_context: 0 &f->f_pos_lock sb_writers#4 &sbi->s_writepages_rwsem jbd2_handle &journal->j_wait_updates irq_context: 0 &f->f_pos_lock sb_writers#4 &sbi->s_writepages_rwsem tk_core.seq.seqcount irq_context: 0 &f->f_pos_lock sb_writers#4 &sbi->s_writepages_rwsem &dd->lock irq_context: 0 &f->f_pos_lock sb_writers#4 &sbi->s_writepages_rwsem rcu_read_lock &dd->lock irq_context: 0 &f->f_pos_lock sb_writers#4 &sbi->s_writepages_rwsem rcu_read_lock &obj_hash[i].lock irq_context: 0 &f->f_pos_lock sb_writers#4 &sbi->s_writepages_rwsem rcu_read_lock pool_lock#2 irq_context: 0 &f->f_pos_lock sb_writers#4 &sbi->s_writepages_rwsem rcu_read_lock tk_core.seq.seqcount irq_context: 0 &f->f_pos_lock sb_writers#4 &sbi->s_writepages_rwsem rcu_read_lock &virtscsi_vq->vq_lock irq_context: 0 &f->f_pos_lock sb_writers#4 &wb->list_lock irq_context: 0 &f->f_pos_lock sb_writers#4 &folio_wait_table[i] irq_context: 0 sb_writers#4 &sb->s_type->i_mutex_key#8 &mm->mmap_lock rcu_read_lock rcu_node_0 irq_context: 0 sb_writers#4 &sb->s_type->i_mutex_key#8 &mm->mmap_lock rcu_read_lock &rq->__lock irq_context: 0 sb_writers#4 &sb->s_type->i_mutex_key#8 &mm->mmap_lock rcu_read_lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 sb_writers#4 &sb->s_type->i_mutex_key#8 &mm->mmap_lock fs_reclaim irq_context: 0 sb_writers#4 &sb->s_type->i_mutex_key#8 &mm->mmap_lock fs_reclaim mmu_notifier_invalidate_range_start irq_context: 0 sb_writers#4 &sb->s_type->i_mutex_key#8 &mm->mmap_lock &____s->seqcount irq_context: 0 sb_writers#4 &sb->s_type->i_mutex_key#8 &mm->mmap_lock pool_lock#2 irq_context: 0 sb_writers#4 &sb->s_type->i_mutex_key#8 &mm->mmap_lock stock_lock irq_context: 0 sb_writers#4 &sb->s_type->i_mutex_key#8 &mm->mmap_lock ptlock_ptr(page) irq_context: 0 sb_writers#4 &sb->s_type->i_mutex_key#8 &mm->mmap_lock &sem->wait_lock irq_context: 0 sb_writers#4 &sb->s_type->i_mutex_key#8 &mm->mmap_lock &dd->lock irq_context: 0 sb_writers#4 &sb->s_type->i_mutex_key#8 &mm->mmap_lock &obj_hash[i].lock irq_context: 0 sb_writers#4 &sb->s_type->i_mutex_key#8 &mm->mmap_lock rcu_read_lock &pool->lock irq_context: 0 sb_writers#4 &sb->s_type->i_mutex_key#8 &mm->mmap_lock rcu_read_lock &pool->lock &obj_hash[i].lock irq_context: 0 sb_writers#4 &sb->s_type->i_mutex_key#8 &mm->mmap_lock rcu_read_lock &pool->lock pool_lock#2 irq_context: 0 sb_writers#4 &sb->s_type->i_mutex_key#8 &mm->mmap_lock rcu_read_lock &pool->lock &p->pi_lock irq_context: 0 sb_writers#4 &sb->s_type->i_mutex_key#8 &mm->mmap_lock rcu_read_lock &pool->lock &p->pi_lock &rq->__lock irq_context: 0 sb_writers#4 &sb->s_type->i_mutex_key#8 &mm->mmap_lock rcu_read_lock &pool->lock &p->pi_lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 sb_writers#4 sb_internal jbd2_handle &ei->i_data_sem rcu_read_lock rcu_node_0 irq_context: 0 sb_writers#4 sb_internal jbd2_handle &ei->i_data_sem rcu_read_lock &rq->__lock irq_context: 0 sb_writers#4 sb_internal jbd2_handle &ei->i_data_sem rcu_read_lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 &f->f_pos_lock sb_writers#4 lock#4 irq_context: 0 &f->f_pos_lock sb_writers#4 lock#5 irq_context: 0 &f->f_pos_lock sb_writers#4 &journal->j_state_lock irq_context: 0 &f->f_pos_lock sb_writers#4 &journal->j_state_lock irq_context: 0 &f->f_pos_lock sb_writers#4 &journal->j_state_lock &journal->j_wait_commit irq_context: 0 &f->f_pos_lock sb_writers#4 &journal->j_state_lock &journal->j_wait_commit &p->pi_lock irq_context: 0 &f->f_pos_lock sb_writers#4 &journal->j_state_lock &journal->j_wait_commit &p->pi_lock &rq->__lock irq_context: 0 &f->f_pos_lock sb_writers#4 &journal->j_state_lock &journal->j_wait_commit &p->pi_lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 &f->f_pos_lock sb_writers#4 &journal->j_wait_commit irq_context: 0 &f->f_pos_lock sb_writers#4 &journal->j_wait_done_commit irq_context: 0 sb_writers#4 &sb->s_type->i_mutex_key#8 rcu_read_lock &obj_hash[i].lock pool_lock irq_context: 0 sb_writers#4 &sb->s_type->i_mutex_key#8 &mm->mmap_lock rcu_node_0 irq_context: 0 &rdma_nl_types[idx].sem link_ops_rwsem rcu_read_lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 sb_writers#4 &sb->s_type->i_mutex_key#8 rcu_read_lock &rcu_state.gp_wq irq_context: 0 sb_writers#4 &sb->s_type->i_mutex_key#8 rcu_read_lock &rcu_state.gp_wq &p->pi_lock irq_context: 0 sb_writers#4 &sb->s_type->i_mutex_key#8 rcu_read_lock &rcu_state.gp_wq &p->pi_lock &rq->__lock irq_context: 0 sb_writers#4 &sb->s_type->i_mutex_key#8 rcu_read_lock &rcu_state.gp_wq &p->pi_lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 &f->f_pos_lock sb_writers#4 &sbi->s_writepages_rwsem jbd2_handle &c->lock irq_context: 0 &f->f_pos_lock sb_writers#4 jbd2_handle irq_context: 0 &rdma_nl_types[idx].sem link_ops_rwsem fs_reclaim mmu_notifier_invalidate_range_start &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 sb_writers#4 &sb->s_type->i_mutex_key#8 rcu_read_lock rcu_read_lock rcu_read_lock rcu_node_0 irq_context: 0 sb_writers#4 &sb->s_type->i_mutex_key#8 rcu_read_lock rcu_read_lock rcu_read_lock &rq->__lock irq_context: 0 sb_writers#4 &sb->s_type->i_mutex_key#8 rcu_read_lock rcu_read_lock rcu_read_lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 sb_writers#4 &sb->s_type->i_mutex_key#8 jbd2_handle &ei->i_data_sem &mapping->private_lock irq_context: 0 &rdma_nl_types[idx].sem link_ops_rwsem batched_entropy_u8.lock irq_context: 0 &rdma_nl_types[idx].sem link_ops_rwsem kfence_freelist_lock irq_context: 0 &rdma_nl_types[idx].sem link_ops_rwsem devices_rwsem irq_context: 0 &rdma_nl_types[idx].sem link_ops_rwsem devices_rwsem pool_lock#2 irq_context: 0 &rdma_nl_types[idx].sem link_ops_rwsem devices_rwsem &rq->__lock irq_context: 0 &rdma_nl_types[idx].sem link_ops_rwsem devices_rwsem &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 &rdma_nl_types[idx].sem link_ops_rwsem devices_rwsem &obj_hash[i].lock irq_context: 0 &rdma_nl_types[idx].sem link_ops_rwsem devices_rwsem rcu_read_lock &pool->lock irq_context: 0 &rdma_nl_types[idx].sem link_ops_rwsem devices_rwsem rcu_read_lock &pool->lock &obj_hash[i].lock irq_context: 0 &rdma_nl_types[idx].sem link_ops_rwsem devices_rwsem rcu_read_lock &pool->lock pool_lock#2 irq_context: 0 &rdma_nl_types[idx].sem link_ops_rwsem devices_rwsem rcu_read_lock &pool->lock &p->pi_lock irq_context: 0 &rdma_nl_types[idx].sem link_ops_rwsem devices_rwsem rcu_read_lock &pool->lock &p->pi_lock &rq->__lock irq_context: 0 &rdma_nl_types[idx].sem link_ops_rwsem devices_rwsem rcu_read_lock &pool->lock &p->pi_lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)infiniband irq_context: 0 (wq_completion)infiniband (work_completion)(&work->work)#2 irq_context: 0 (wq_completion)infiniband (work_completion)(&work->work)#2 fs_reclaim irq_context: 0 (wq_completion)infiniband (work_completion)(&work->work)#2 fs_reclaim mmu_notifier_invalidate_range_start irq_context: 0 (wq_completion)infiniband (work_completion)(&work->work)#2 pool_lock#2 irq_context: 0 (wq_completion)infiniband (work_completion)(&work->work)#2 &rxe->usdev_lock irq_context: 0 (wq_completion)infiniband (work_completion)(&work->work)#2 &rxe->usdev_lock &pdata->netdev_lock irq_context: 0 (wq_completion)infiniband (work_completion)(&work->work)#2 &rxe->usdev_lock rtnl_mutex irq_context: 0 (wq_completion)infiniband (work_completion)(&work->work)#2 &device->cache_lock irq_context: 0 (wq_completion)infiniband (work_completion)(&work->work)#2 &obj_hash[i].lock irq_context: 0 (wq_completion)infiniband (work_completion)(&work->work)#2 &device->event_handler_rwsem irq_context: 0 &rdma_nl_types[idx].sem link_ops_rwsem devices_rwsem (console_sem).lock irq_context: 0 &rdma_nl_types[idx].sem link_ops_rwsem devices_rwsem console_lock console_srcu console_owner_lock irq_context: 0 &rdma_nl_types[idx].sem link_ops_rwsem devices_rwsem console_lock console_srcu console_owner irq_context: 0 &rdma_nl_types[idx].sem link_ops_rwsem devices_rwsem console_lock console_srcu console_owner &port_lock_key irq_context: 0 &rdma_nl_types[idx].sem link_ops_rwsem devices_rwsem console_lock console_srcu console_owner console_owner_lock irq_context: 0 &f->f_pos_lock sb_writers#4 &sbi->s_writepages_rwsem jbd2_handle &ei->i_data_sem &mapping->private_lock irq_context: 0 &f->f_pos_lock sb_writers#4 &sbi->s_writepages_rwsem jbd2_handle &ei->i_data_sem &rq->__lock irq_context: 0 &f->f_pos_lock sb_writers#4 &sbi->s_writepages_rwsem jbd2_handle &ei->i_data_sem &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 &f->f_pos_lock sb_writers#4 &sbi->s_writepages_rwsem jbd2_handle &____s->seqcount#2 irq_context: 0 &f->f_pos_lock sb_writers#4 &sbi->s_writepages_rwsem jbd2_handle &____s->seqcount irq_context: 0 &f->f_pos_lock sb_writers#4 &sbi->s_writepages_rwsem rcu_read_lock &obj_hash[i].lock pool_lock irq_context: 0 &f->f_pos_lock sb_writers#4 &sbi->s_writepages_rwsem rcu_read_lock &rcu_state.gp_wq irq_context: 0 &f->f_pos_lock sb_writers#4 &sbi->s_writepages_rwsem rcu_read_lock &rcu_state.gp_wq &p->pi_lock irq_context: 0 &f->f_pos_lock sb_writers#4 &sbi->s_writepages_rwsem rcu_read_lock &rcu_state.gp_wq &p->pi_lock &rq->__lock irq_context: 0 &f->f_pos_lock sb_writers#4 &sbi->s_writepages_rwsem rcu_read_lock &rcu_state.gp_wq &p->pi_lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 &rdma_nl_types[idx].sem link_ops_rwsem devices_rwsem clients_rwsem irq_context: 0 &rdma_nl_types[idx].sem link_ops_rwsem devices_rwsem clients_rwsem &device->client_data_rwsem irq_context: 0 &rdma_nl_types[idx].sem link_ops_rwsem devices_rwsem clients_rwsem &device->client_data_rwsem &xa->xa_lock#17 irq_context: 0 &rdma_nl_types[idx].sem link_ops_rwsem devices_rwsem clients_rwsem &device->client_data_rwsem &xa->xa_lock#17 pool_lock#2 irq_context: 0 &rdma_nl_types[idx].sem link_ops_rwsem devices_rwsem clients_rwsem &device->client_data_rwsem irq_context: 0 &rdma_nl_types[idx].sem link_ops_rwsem devices_rwsem clients_rwsem &device->client_data_rwsem fs_reclaim irq_context: 0 &rdma_nl_types[idx].sem link_ops_rwsem devices_rwsem clients_rwsem &device->client_data_rwsem fs_reclaim mmu_notifier_invalidate_range_start irq_context: 0 &rdma_nl_types[idx].sem link_ops_rwsem devices_rwsem clients_rwsem &device->client_data_rwsem pool_lock#2 irq_context: 0 &rdma_nl_types[idx].sem link_ops_rwsem devices_rwsem clients_rwsem &device->client_data_rwsem &c->lock irq_context: 0 &rdma_nl_types[idx].sem link_ops_rwsem devices_rwsem clients_rwsem &device->client_data_rwsem rcu_node_0 irq_context: 0 &rdma_nl_types[idx].sem link_ops_rwsem devices_rwsem clients_rwsem &device->client_data_rwsem &rq->__lock irq_context: 0 &rdma_nl_types[idx].sem link_ops_rwsem devices_rwsem clients_rwsem &device->client_data_rwsem &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 &rdma_nl_types[idx].sem link_ops_rwsem devices_rwsem clients_rwsem &device->client_data_rwsem &cfs_rq->removed.lock irq_context: 0 &rdma_nl_types[idx].sem link_ops_rwsem devices_rwsem clients_rwsem &device->client_data_rwsem &obj_hash[i].lock irq_context: 0 &rdma_nl_types[idx].sem link_ops_rwsem devices_rwsem clients_rwsem &device->client_data_rwsem rcu_read_lock rcu_node_0 irq_context: 0 &rdma_nl_types[idx].sem link_ops_rwsem devices_rwsem clients_rwsem &device->client_data_rwsem &xa->xa_lock#18 irq_context: 0 &rdma_nl_types[idx].sem link_ops_rwsem devices_rwsem clients_rwsem &device->client_data_rwsem &xa->xa_lock#18 pool_lock#2 irq_context: 0 &rdma_nl_types[idx].sem link_ops_rwsem devices_rwsem clients_rwsem &device->client_data_rwsem &xa->xa_lock#19 irq_context: 0 &rdma_nl_types[idx].sem link_ops_rwsem devices_rwsem clients_rwsem &device->client_data_rwsem &xa->xa_lock#18 &c->lock irq_context: 0 &rdma_nl_types[idx].sem link_ops_rwsem devices_rwsem clients_rwsem &device->client_data_rwsem crngs.lock irq_context: 0 &rdma_nl_types[idx].sem link_ops_rwsem devices_rwsem clients_rwsem &device->client_data_rwsem &xa->xa_lock#18 &n->list_lock irq_context: 0 &rdma_nl_types[idx].sem link_ops_rwsem devices_rwsem clients_rwsem &device->client_data_rwsem &xa->xa_lock#18 &n->list_lock &c->lock irq_context: 0 &rdma_nl_types[idx].sem link_ops_rwsem devices_rwsem clients_rwsem &device->client_data_rwsem free_vmap_area_lock irq_context: 0 &rdma_nl_types[idx].sem link_ops_rwsem devices_rwsem clients_rwsem &device->client_data_rwsem vmap_area_lock irq_context: 0 &rdma_nl_types[idx].sem link_ops_rwsem devices_rwsem clients_rwsem &device->client_data_rwsem &____s->seqcount irq_context: 0 &rdma_nl_types[idx].sem link_ops_rwsem devices_rwsem clients_rwsem &device->client_data_rwsem init_mm.page_table_lock irq_context: 0 &rdma_nl_types[idx].sem link_ops_rwsem devices_rwsem clients_rwsem &device->client_data_rwsem purge_vmap_area_lock irq_context: 0 &rdma_nl_types[idx].sem link_ops_rwsem devices_rwsem clients_rwsem &device->client_data_rwsem rcu_read_lock pool_lock#2 irq_context: 0 &rdma_nl_types[idx].sem link_ops_rwsem devices_rwsem clients_rwsem &device->client_data_rwsem &xa->xa_lock#18 &obj_hash[i].lock irq_context: 0 &rdma_nl_types[idx].sem link_ops_rwsem devices_rwsem clients_rwsem &device->client_data_rwsem &x->wait#27 irq_context: 0 &rdma_nl_types[idx].sem link_ops_rwsem devices_rwsem clients_rwsem &device->client_data_rwsem (console_sem).lock irq_context: 0 &rdma_nl_types[idx].sem link_ops_rwsem devices_rwsem clients_rwsem &device->client_data_rwsem &x->wait#28 irq_context: 0 &rdma_nl_types[idx].sem link_ops_rwsem devices_rwsem clients_rwsem &device->client_data_rwsem console_lock console_srcu console_owner_lock irq_context: 0 &rdma_nl_types[idx].sem link_ops_rwsem devices_rwsem clients_rwsem &device->client_data_rwsem console_lock console_srcu console_owner irq_context: 0 &rdma_nl_types[idx].sem link_ops_rwsem devices_rwsem clients_rwsem &device->client_data_rwsem console_lock console_srcu console_owner &port_lock_key irq_context: 0 &rdma_nl_types[idx].sem link_ops_rwsem devices_rwsem clients_rwsem &device->client_data_rwsem console_lock console_srcu console_owner console_owner_lock irq_context: 0 &rdma_nl_types[idx].sem link_ops_rwsem devices_rwsem clients_rwsem &device->client_data_rwsem krc.lock irq_context: 0 &rdma_nl_types[idx].sem link_ops_rwsem devices_rwsem clients_rwsem &device->client_data_rwsem &xa->xa_lock#17 irq_context: 0 &rdma_nl_types[idx].sem link_ops_rwsem devices_rwsem clients_rwsem &device->client_data_rwsem &xa->xa_lock#17 &obj_hash[i].lock irq_context: 0 &rdma_nl_types[idx].sem link_ops_rwsem devices_rwsem clients_rwsem &device->client_data_rwsem &xa->xa_lock#17 pool_lock#2 irq_context: 0 &rdma_nl_types[idx].sem link_ops_rwsem devices_rwsem clients_rwsem &device->client_data_rwsem lock irq_context: 0 &rdma_nl_types[idx].sem link_ops_rwsem devices_rwsem clients_rwsem &device->client_data_rwsem lock kernfs_idr_lock irq_context: 0 &rdma_nl_types[idx].sem link_ops_rwsem devices_rwsem clients_rwsem &device->client_data_rwsem &root->kernfs_rwsem irq_context: 0 &rdma_nl_types[idx].sem link_ops_rwsem devices_rwsem clients_rwsem &device->client_data_rwsem &root->kernfs_rwsem &root->kernfs_iattr_rwsem irq_context: 0 &rdma_nl_types[idx].sem link_ops_rwsem devices_rwsem clients_rwsem &device->client_data_rwsem &____s->seqcount#2 irq_context: 0 &rdma_nl_types[idx].sem link_ops_rwsem devices_rwsem clients_rwsem &device->client_data_rwsem lock kernfs_idr_lock pool_lock#2 irq_context: 0 &rdma_nl_types[idx].sem link_ops_rwsem devices_rwsem clients_rwsem &device->client_data_rwsem ib_mad_port_list_lock irq_context: 0 &rdma_nl_types[idx].sem link_ops_rwsem devices_rwsem clients_rwsem &device->client_data_rwsem &root->kernfs_rwsem irq_context: 0 &rdma_nl_types[idx].sem link_ops_rwsem devices_rwsem clients_rwsem &device->client_data_rwsem &root->kernfs_rwsem kernfs_idr_lock irq_context: 0 &rdma_nl_types[idx].sem link_ops_rwsem devices_rwsem clients_rwsem &device->client_data_rwsem &root->kernfs_rwsem &obj_hash[i].lock irq_context: 0 &rdma_nl_types[idx].sem link_ops_rwsem devices_rwsem clients_rwsem &device->client_data_rwsem &root->kernfs_rwsem pool_lock#2 irq_context: 0 &rdma_nl_types[idx].sem link_ops_rwsem devices_rwsem clients_rwsem &device->client_data_rwsem kernfs_idr_lock irq_context: 0 &rdma_nl_types[idx].sem link_ops_rwsem devices_rwsem clients_rwsem &device->client_data_rwsem &root->kernfs_rwsem kernfs_idr_lock &obj_hash[i].lock irq_context: 0 &rdma_nl_types[idx].sem link_ops_rwsem devices_rwsem clients_rwsem &device->client_data_rwsem &root->kernfs_rwsem kernfs_idr_lock pool_lock#2 irq_context: 0 &rdma_nl_types[idx].sem link_ops_rwsem devices_rwsem clients_rwsem &device->client_data_rwsem lock#7 irq_context: 0 &rdma_nl_types[idx].sem link_ops_rwsem devices_rwsem clients_rwsem &device->client_data_rwsem lock#7 fs_reclaim irq_context: 0 &rdma_nl_types[idx].sem link_ops_rwsem devices_rwsem clients_rwsem &device->client_data_rwsem lock#7 fs_reclaim mmu_notifier_invalidate_range_start irq_context: 0 &rdma_nl_types[idx].sem link_ops_rwsem devices_rwsem clients_rwsem &device->client_data_rwsem lock#7 pool_lock#2 irq_context: 0 &rdma_nl_types[idx].sem link_ops_rwsem devices_rwsem clients_rwsem &device->client_data_rwsem lock#7 crngs.lock irq_context: 0 &rdma_nl_types[idx].sem link_ops_rwsem devices_rwsem clients_rwsem &device->client_data_rwsem lock#7 &xa->xa_lock#19 irq_context: 0 &rdma_nl_types[idx].sem link_ops_rwsem devices_rwsem clients_rwsem &device->client_data_rwsem lock#7 &id_priv->qp_mutex irq_context: 0 &rdma_nl_types[idx].sem link_ops_rwsem devices_rwsem clients_rwsem &device->client_data_rwsem lock#7 &id_priv->lock irq_context: 0 &rdma_nl_types[idx].sem link_ops_rwsem devices_rwsem clients_rwsem &device->client_data_rwsem lock#7 &xa->xa_lock#20 irq_context: 0 &rdma_nl_types[idx].sem link_ops_rwsem devices_rwsem clients_rwsem &device->client_data_rwsem lock#7 &xa->xa_lock#20 pool_lock#2 irq_context: 0 &rdma_nl_types[idx].sem link_ops_rwsem devices_rwsem clients_rwsem &device->client_data_rwsem lock#7 &cm_id_priv->lock irq_context: 0 &rdma_nl_types[idx].sem link_ops_rwsem devices_rwsem clients_rwsem &device->client_data_rwsem lock#7 &cm_id_priv->lock &cm.lock irq_context: 0 &rdma_nl_types[idx].sem link_ops_rwsem devices_rwsem clients_rwsem &device->client_data_rwsem lock#7 &xa->xa_lock#19 pool_lock#2 irq_context: 0 &rdma_nl_types[idx].sem link_ops_rwsem devices_rwsem clients_rwsem &device->client_data_rwsem lock#7 &c->lock irq_context: 0 &rdma_nl_types[idx].sem link_ops_rwsem devices_rwsem clients_rwsem &device->client_data_rwsem umad_ida.xa_lock irq_context: 0 &rdma_nl_types[idx].sem link_ops_rwsem devices_rwsem clients_rwsem &device->client_data_rwsem &x->wait#9 irq_context: 0 &rdma_nl_types[idx].sem link_ops_rwsem devices_rwsem clients_rwsem &device->client_data_rwsem &n->list_lock irq_context: 0 &rdma_nl_types[idx].sem link_ops_rwsem devices_rwsem clients_rwsem &device->client_data_rwsem &n->list_lock &c->lock irq_context: 0 sb_writers#4 tomoyo_ss &pcp->lock &zone->lock irq_context: 0 tracepoints_mutex &n->list_lock irq_context: 0 tracepoints_mutex &n->list_lock &c->lock irq_context: 0 &rdma_nl_types[idx].sem link_ops_rwsem devices_rwsem clients_rwsem &device->client_data_rwsem chrdevs_lock irq_context: 0 &rdma_nl_types[idx].sem link_ops_rwsem devices_rwsem clients_rwsem &device->client_data_rwsem &k->list_lock irq_context: 0 &rdma_nl_types[idx].sem link_ops_rwsem devices_rwsem clients_rwsem &device->client_data_rwsem gdp_mutex irq_context: 0 &rdma_nl_types[idx].sem link_ops_rwsem devices_rwsem clients_rwsem &device->client_data_rwsem gdp_mutex &k->list_lock irq_context: 0 &rdma_nl_types[idx].sem link_ops_rwsem devices_rwsem clients_rwsem &device->client_data_rwsem gdp_mutex fs_reclaim irq_context: 0 &rdma_nl_types[idx].sem link_ops_rwsem devices_rwsem clients_rwsem &device->client_data_rwsem gdp_mutex fs_reclaim mmu_notifier_invalidate_range_start irq_context: 0 &rdma_nl_types[idx].sem link_ops_rwsem devices_rwsem clients_rwsem &device->client_data_rwsem gdp_mutex pool_lock#2 irq_context: 0 &rdma_nl_types[idx].sem link_ops_rwsem devices_rwsem clients_rwsem &device->client_data_rwsem gdp_mutex lock irq_context: 0 &rdma_nl_types[idx].sem link_ops_rwsem devices_rwsem clients_rwsem &device->client_data_rwsem gdp_mutex lock kernfs_idr_lock irq_context: 0 &rdma_nl_types[idx].sem link_ops_rwsem devices_rwsem clients_rwsem &device->client_data_rwsem gdp_mutex lock kernfs_idr_lock &c->lock irq_context: 0 &rdma_nl_types[idx].sem link_ops_rwsem devices_rwsem clients_rwsem &device->client_data_rwsem gdp_mutex lock kernfs_idr_lock pool_lock#2 irq_context: 0 &rdma_nl_types[idx].sem link_ops_rwsem devices_rwsem clients_rwsem &device->client_data_rwsem gdp_mutex &root->kernfs_rwsem irq_context: 0 &rdma_nl_types[idx].sem link_ops_rwsem devices_rwsem clients_rwsem &device->client_data_rwsem gdp_mutex &root->kernfs_rwsem &root->kernfs_iattr_rwsem irq_context: 0 &rdma_nl_types[idx].sem link_ops_rwsem devices_rwsem clients_rwsem &device->client_data_rwsem bus_type_sem irq_context: 0 &rdma_nl_types[idx].sem link_ops_rwsem devices_rwsem clients_rwsem &device->client_data_rwsem sysfs_symlink_target_lock irq_context: 0 &rdma_nl_types[idx].sem link_ops_rwsem devices_rwsem clients_rwsem &device->client_data_rwsem &dev->power.lock irq_context: 0 &rdma_nl_types[idx].sem link_ops_rwsem devices_rwsem clients_rwsem &device->client_data_rwsem dpm_list_mtx irq_context: 0 &rdma_nl_types[idx].sem link_ops_rwsem devices_rwsem clients_rwsem &device->client_data_rwsem req_lock irq_context: 0 &rdma_nl_types[idx].sem link_ops_rwsem devices_rwsem clients_rwsem &device->client_data_rwsem &p->pi_lock irq_context: 0 &rdma_nl_types[idx].sem link_ops_rwsem devices_rwsem clients_rwsem &device->client_data_rwsem &p->pi_lock &rq->__lock irq_context: 0 &rdma_nl_types[idx].sem link_ops_rwsem devices_rwsem clients_rwsem &device->client_data_rwsem &p->pi_lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 &rdma_nl_types[idx].sem link_ops_rwsem devices_rwsem clients_rwsem &device->client_data_rwsem &x->wait#11 irq_context: 0 &rdma_nl_types[idx].sem link_ops_rwsem devices_rwsem clients_rwsem &device->client_data_rwsem uevent_sock_mutex irq_context: 0 &rdma_nl_types[idx].sem link_ops_rwsem devices_rwsem clients_rwsem &device->client_data_rwsem uevent_sock_mutex fs_reclaim irq_context: 0 &rdma_nl_types[idx].sem link_ops_rwsem devices_rwsem clients_rwsem &device->client_data_rwsem uevent_sock_mutex fs_reclaim mmu_notifier_invalidate_range_start irq_context: 0 &rdma_nl_types[idx].sem link_ops_rwsem devices_rwsem clients_rwsem &device->client_data_rwsem uevent_sock_mutex pool_lock#2 irq_context: 0 &rdma_nl_types[idx].sem link_ops_rwsem devices_rwsem clients_rwsem &device->client_data_rwsem uevent_sock_mutex nl_table_lock irq_context: 0 &rdma_nl_types[idx].sem link_ops_rwsem devices_rwsem clients_rwsem &device->client_data_rwsem uevent_sock_mutex rlock-AF_NETLINK irq_context: 0 &rdma_nl_types[idx].sem link_ops_rwsem devices_rwsem clients_rwsem &device->client_data_rwsem uevent_sock_mutex rcu_read_lock &ei->socket.wq.wait irq_context: 0 &rdma_nl_types[idx].sem link_ops_rwsem devices_rwsem clients_rwsem &device->client_data_rwsem uevent_sock_mutex rcu_read_lock &ei->socket.wq.wait &ep->lock irq_context: 0 &rdma_nl_types[idx].sem link_ops_rwsem devices_rwsem clients_rwsem &device->client_data_rwsem uevent_sock_mutex rcu_read_lock &ei->socket.wq.wait &ep->lock &ep->wq irq_context: 0 &rdma_nl_types[idx].sem link_ops_rwsem devices_rwsem clients_rwsem &device->client_data_rwsem uevent_sock_mutex rcu_read_lock &ei->socket.wq.wait &ep->lock &ep->wq &p->pi_lock irq_context: 0 &rdma_nl_types[idx].sem link_ops_rwsem devices_rwsem clients_rwsem &device->client_data_rwsem uevent_sock_mutex rcu_read_lock &ei->socket.wq.wait &ep->lock &ep->wq &p->pi_lock &rq->__lock irq_context: 0 &rdma_nl_types[idx].sem link_ops_rwsem devices_rwsem clients_rwsem &device->client_data_rwsem uevent_sock_mutex rcu_read_lock &ei->socket.wq.wait &ep->lock &ep->wq &p->pi_lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 &rdma_nl_types[idx].sem link_ops_rwsem devices_rwsem clients_rwsem &device->client_data_rwsem uevent_sock_mutex nl_table_wait.lock irq_context: 0 &rdma_nl_types[idx].sem link_ops_rwsem devices_rwsem clients_rwsem &device->client_data_rwsem uevent_sock_mutex &obj_hash[i].lock irq_context: 0 &rdma_nl_types[idx].sem link_ops_rwsem devices_rwsem clients_rwsem &device->client_data_rwsem subsys mutex#85 irq_context: 0 &rdma_nl_types[idx].sem link_ops_rwsem devices_rwsem clients_rwsem &device->client_data_rwsem subsys mutex#85 &k->k_lock irq_context: 0 &rdma_nl_types[idx].sem link_ops_rwsem devices_rwsem clients_rwsem &device->client_data_rwsem pcpu_alloc_mutex irq_context: 0 &rdma_nl_types[idx].sem link_ops_rwsem devices_rwsem clients_rwsem &device->client_data_rwsem pcpu_alloc_mutex pcpu_lock irq_context: 0 &rdma_nl_types[idx].sem link_ops_rwsem devices_rwsem clients_rwsem &device->client_data_rwsem uverbs_ida.xa_lock irq_context: 0 &rdma_nl_types[idx].sem link_ops_rwsem devices_rwsem clients_rwsem &device->client_data_rwsem &pcp->lock &zone->lock irq_context: 0 &rdma_nl_types[idx].sem link_ops_rwsem devices_rwsem clients_rwsem &device->client_data_rwsem gdp_mutex &c->lock irq_context: 0 &rdma_nl_types[idx].sem link_ops_rwsem devices_rwsem clients_rwsem &device->client_data_rwsem subsys mutex#86 irq_context: 0 &rdma_nl_types[idx].sem link_ops_rwsem devices_rwsem clients_rwsem &device->client_data_rwsem subsys mutex#86 &k->k_lock irq_context: 0 &rdma_nl_types[idx].sem link_ops_rwsem devices_rwsem clients_rwsem &device->client_data_rwsem &xa->xa_lock#19 pool_lock#2 irq_context: 0 &rdma_nl_types[idx].sem link_ops_rwsem devices_rwsem clients_rwsem &device->client_data_rwsem gdp_mutex &rq->__lock irq_context: 0 &rdma_nl_types[idx].sem link_ops_rwsem devices_rwsem clients_rwsem &device->client_data_rwsem gdp_mutex &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 &rdma_nl_types[idx].sem link_ops_rwsem devices_rwsem clients_rwsem &device->client_data_rwsem uevent_sock_mutex &c->lock irq_context: 0 &rdma_nl_types[idx].sem link_ops_rwsem devices_rwsem clients_rwsem &device->client_data_rwsem subsys mutex#87 irq_context: 0 &rdma_nl_types[idx].sem link_ops_rwsem devices_rwsem clients_rwsem &device->client_data_rwsem subsys mutex#87 &k->k_lock irq_context: 0 &rdma_nl_types[idx].sem link_ops_rwsem devices_rwsem clients_rwsem &device->client_data_rwsem fs_reclaim &rq->__lock irq_context: 0 &rdma_nl_types[idx].sem link_ops_rwsem devices_rwsem clients_rwsem &device->client_data_rwsem fs_reclaim &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 &f->f_pos_lock sb_writers#4 lock#4 &lruvec->lru_lock irq_context: 0 &rdma_nl_types[idx].sem link_ops_rwsem devices_rwsem clients_rwsem &device->client_data_rwsem rds_ib_devices_lock irq_context: 0 &rdma_nl_types[idx].sem link_ops_rwsem devices_rwsem clients_rwsem &device->client_data_rwsem ib_nodev_conns_lock irq_context: 0 &rdma_nl_types[idx].sem link_ops_rwsem devices_rwsem clients_rwsem &device->client_data_rwsem smc_ib_devices.mutex irq_context: 0 &rdma_nl_types[idx].sem link_ops_rwsem devices_rwsem clients_rwsem &device->client_data_rwsem &device->event_handler_rwsem irq_context: 0 &sbi->s_writepages_rwsem jbd2_handle &ei->i_data_sem bit_wait_table + i irq_context: 0 &sbi->s_writepages_rwsem jbd2_handle &ei->i_data_sem &cfs_rq->removed.lock irq_context: 0 &rdma_nl_types[idx].sem link_ops_rwsem devices_rwsem clients_rwsem &device->client_data_rwsem &pnettable->lock irq_context: 0 &rdma_nl_types[idx].sem link_ops_rwsem devices_rwsem clients_rwsem &device->client_data_rwsem rcu_read_lock &pool->lock irq_context: 0 &rdma_nl_types[idx].sem link_ops_rwsem devices_rwsem clients_rwsem &device->client_data_rwsem rcu_read_lock &pool->lock &obj_hash[i].lock irq_context: 0 &rdma_nl_types[idx].sem link_ops_rwsem devices_rwsem clients_rwsem &device->client_data_rwsem rcu_read_lock &pool->lock pool_lock#2 irq_context: 0 &rdma_nl_types[idx].sem link_ops_rwsem devices_rwsem clients_rwsem &device->client_data_rwsem rcu_read_lock &pool->lock &p->pi_lock irq_context: 0 &rdma_nl_types[idx].sem link_ops_rwsem devices_rwsem rdma_nets_rwsem irq_context: 0 &rdma_nl_types[idx].sem link_ops_rwsem devices_rwsem rdma_nets_rwsem &device->compat_devs_mutex irq_context: 0 &rdma_nl_types[idx].sem link_ops_rwsem devices_rwsem rdma_nets_rwsem &device->compat_devs_mutex fs_reclaim irq_context: 0 &rdma_nl_types[idx].sem link_ops_rwsem devices_rwsem rdma_nets_rwsem &device->compat_devs_mutex fs_reclaim mmu_notifier_invalidate_range_start irq_context: 0 &rdma_nl_types[idx].sem link_ops_rwsem devices_rwsem rdma_nets_rwsem &device->compat_devs_mutex &xa->xa_lock#17 irq_context: 0 &rdma_nl_types[idx].sem link_ops_rwsem devices_rwsem rdma_nets_rwsem &device->compat_devs_mutex &xa->xa_lock#17 pool_lock#2 irq_context: 0 &rdma_nl_types[idx].sem link_ops_rwsem devices_rwsem rdma_nets_rwsem &device->compat_devs_mutex pool_lock#2 irq_context: 0 &rdma_nl_types[idx].sem link_ops_rwsem devices_rwsem rdma_nets_rwsem &device->compat_devs_mutex &x->wait#9 irq_context: 0 &rdma_nl_types[idx].sem link_ops_rwsem devices_rwsem rdma_nets_rwsem &device->compat_devs_mutex &obj_hash[i].lock irq_context: 0 &rdma_nl_types[idx].sem link_ops_rwsem devices_rwsem rdma_nets_rwsem &device->compat_devs_mutex &k->list_lock irq_context: 0 &rdma_nl_types[idx].sem link_ops_rwsem devices_rwsem rdma_nets_rwsem &device->compat_devs_mutex gdp_mutex irq_context: 0 &rdma_nl_types[idx].sem link_ops_rwsem devices_rwsem rdma_nets_rwsem &device->compat_devs_mutex gdp_mutex &k->list_lock irq_context: 0 &rdma_nl_types[idx].sem link_ops_rwsem devices_rwsem rdma_nets_rwsem &device->compat_devs_mutex &c->lock irq_context: 0 &rdma_nl_types[idx].sem link_ops_rwsem devices_rwsem rdma_nets_rwsem &device->compat_devs_mutex lock irq_context: 0 &rdma_nl_types[idx].sem link_ops_rwsem devices_rwsem rdma_nets_rwsem &device->compat_devs_mutex lock kernfs_idr_lock irq_context: 0 &rdma_nl_types[idx].sem link_ops_rwsem devices_rwsem rdma_nets_rwsem &device->compat_devs_mutex &root->kernfs_rwsem irq_context: 0 &rdma_nl_types[idx].sem link_ops_rwsem devices_rwsem rdma_nets_rwsem &device->compat_devs_mutex &root->kernfs_rwsem &root->kernfs_iattr_rwsem irq_context: 0 &rdma_nl_types[idx].sem link_ops_rwsem devices_rwsem rdma_nets_rwsem &device->compat_devs_mutex bus_type_sem irq_context: 0 &rdma_nl_types[idx].sem link_ops_rwsem devices_rwsem rdma_nets_rwsem &device->compat_devs_mutex sysfs_symlink_target_lock irq_context: 0 &rdma_nl_types[idx].sem link_ops_rwsem devices_rwsem rdma_nets_rwsem &device->compat_devs_mutex lock kernfs_idr_lock pool_lock#2 irq_context: 0 &rdma_nl_types[idx].sem link_ops_rwsem devices_rwsem rdma_nets_rwsem &device->compat_devs_mutex &root->kernfs_rwsem irq_context: 0 &rdma_nl_types[idx].sem link_ops_rwsem devices_rwsem rdma_nets_rwsem &device->compat_devs_mutex &____s->seqcount#2 irq_context: 0 &rdma_nl_types[idx].sem link_ops_rwsem devices_rwsem rdma_nets_rwsem &device->compat_devs_mutex &____s->seqcount irq_context: 0 &rdma_nl_types[idx].sem link_ops_rwsem devices_rwsem rdma_nets_rwsem &device->compat_devs_mutex &dev->power.lock irq_context: 0 &rdma_nl_types[idx].sem link_ops_rwsem devices_rwsem rdma_nets_rwsem &device->compat_devs_mutex dpm_list_mtx irq_context: 0 &rdma_nl_types[idx].sem link_ops_rwsem devices_rwsem rdma_nets_rwsem &device->compat_devs_mutex rcu_read_lock pool_lock#2 irq_context: 0 &rdma_nl_types[idx].sem link_ops_rwsem devices_rwsem rdma_nets_rwsem &device->compat_devs_mutex uevent_sock_mutex irq_context: 0 &rdma_nl_types[idx].sem link_ops_rwsem devices_rwsem rdma_nets_rwsem &device->compat_devs_mutex uevent_sock_mutex fs_reclaim irq_context: 0 &rdma_nl_types[idx].sem link_ops_rwsem devices_rwsem rdma_nets_rwsem &device->compat_devs_mutex uevent_sock_mutex fs_reclaim mmu_notifier_invalidate_range_start irq_context: 0 &rdma_nl_types[idx].sem link_ops_rwsem devices_rwsem rdma_nets_rwsem &device->compat_devs_mutex uevent_sock_mutex &c->lock irq_context: 0 &rdma_nl_types[idx].sem link_ops_rwsem devices_rwsem rdma_nets_rwsem &device->compat_devs_mutex uevent_sock_mutex &n->list_lock irq_context: 0 &rdma_nl_types[idx].sem link_ops_rwsem devices_rwsem rdma_nets_rwsem &device->compat_devs_mutex uevent_sock_mutex &n->list_lock &c->lock irq_context: 0 &rdma_nl_types[idx].sem link_ops_rwsem devices_rwsem rdma_nets_rwsem &device->compat_devs_mutex uevent_sock_mutex pool_lock#2 irq_context: 0 &rdma_nl_types[idx].sem link_ops_rwsem devices_rwsem rdma_nets_rwsem &device->compat_devs_mutex uevent_sock_mutex nl_table_lock irq_context: 0 &rdma_nl_types[idx].sem link_ops_rwsem devices_rwsem rdma_nets_rwsem &device->compat_devs_mutex uevent_sock_mutex &obj_hash[i].lock irq_context: 0 &rdma_nl_types[idx].sem link_ops_rwsem devices_rwsem rdma_nets_rwsem &device->compat_devs_mutex uevent_sock_mutex nl_table_wait.lock irq_context: 0 &rdma_nl_types[idx].sem link_ops_rwsem devices_rwsem rdma_nets_rwsem &device->compat_devs_mutex subsys mutex#84 irq_context: 0 &rdma_nl_types[idx].sem link_ops_rwsem devices_rwsem rdma_nets_rwsem &device->compat_devs_mutex subsys mutex#84 &k->k_lock irq_context: 0 &rdma_nl_types[idx].sem link_ops_rwsem devices_rwsem rdma_nets_rwsem &device->compat_devs_mutex &n->list_lock irq_context: 0 &rdma_nl_types[idx].sem link_ops_rwsem devices_rwsem rdma_nets_rwsem &device->compat_devs_mutex &n->list_lock &c->lock irq_context: 0 &rdma_nl_types[idx].sem link_ops_rwsem devices_rwsem rdma_nets_rwsem &device->compat_devs_mutex &rxe->usdev_lock irq_context: 0 &rdma_nl_types[idx].sem link_ops_rwsem devices_rwsem rdma_nets_rwsem &device->compat_devs_mutex &rxe->usdev_lock &pdata->netdev_lock irq_context: 0 &rdma_nl_types[idx].sem link_ops_rwsem devices_rwsem rdma_nets_rwsem &device->compat_devs_mutex &rxe->usdev_lock rtnl_mutex irq_context: 0 &rdma_nl_types[idx].sem link_ops_rwsem devices_rwsem rdma_nets_rwsem &device->compat_devs_mutex &zone->lock irq_context: 0 &rdma_nl_types[idx].sem link_ops_rwsem devices_rwsem rdma_nets_rwsem &device->compat_devs_mutex &rq->__lock irq_context: 0 &rdma_nl_types[idx].sem link_ops_rwsem devices_rwsem rdma_nets_rwsem &device->compat_devs_mutex &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)events (work_completion)(&smcibdev->port_event_work) irq_context: 0 (wq_completion)events (work_completion)(&smcibdev->port_event_work) &rxe->usdev_lock irq_context: 0 (wq_completion)events (work_completion)(&smcibdev->port_event_work) &rxe->usdev_lock &pdata->netdev_lock irq_context: 0 (wq_completion)events (work_completion)(&smcibdev->port_event_work) &rxe->usdev_lock rtnl_mutex irq_context: 0 (wq_completion)events (work_completion)(&smcibdev->port_event_work) &table->rwlock irq_context: 0 (wq_completion)events (work_completion)(&smcibdev->port_event_work) smc_lgr_list.lock irq_context: 0 &rdma_nl_types[idx].sem link_ops_rwsem devices_rwsem rdma_nets_rwsem &device->compat_devs_mutex fs_reclaim &rq->__lock irq_context: 0 &rdma_nl_types[idx].sem link_ops_rwsem devices_rwsem rdma_nets_rwsem &device->compat_devs_mutex fs_reclaim &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 &rdma_nl_types[idx].sem link_ops_rwsem devices_rwsem rdma_nets_rwsem &device->compat_devs_mutex fs_reclaim pgd_lock irq_context: 0 &rdma_nl_types[idx].sem link_ops_rwsem devices_rwsem rdma_nets_rwsem &device->compat_devs_mutex fs_reclaim stock_lock irq_context: 0 &rdma_nl_types[idx].sem link_ops_rwsem devices_rwsem rdma_nets_rwsem &device->compat_devs_mutex fs_reclaim &obj_hash[i].lock irq_context: 0 &rdma_nl_types[idx].sem link_ops_rwsem devices_rwsem rdma_nets_rwsem &device->compat_devs_mutex fs_reclaim key irq_context: 0 &rdma_nl_types[idx].sem link_ops_rwsem devices_rwsem rdma_nets_rwsem &device->compat_devs_mutex fs_reclaim pcpu_lock irq_context: 0 &rdma_nl_types[idx].sem link_ops_rwsem devices_rwsem rdma_nets_rwsem &device->compat_devs_mutex fs_reclaim percpu_counters_lock irq_context: 0 &rdma_nl_types[idx].sem link_ops_rwsem devices_rwsem rdma_nets_rwsem &device->compat_devs_mutex fs_reclaim pcpu_lock stock_lock irq_context: 0 &rdma_nl_types[idx].sem link_ops_rwsem devices_rwsem rdma_nets_rwsem &device->compat_devs_mutex fs_reclaim &cfs_rq->removed.lock irq_context: 0 &rdma_nl_types[idx].sem link_ops_rwsem devices_rwsem rdma_nets_rwsem &device->compat_devs_mutex &sem->wait_lock irq_context: 0 &rdma_nl_types[idx].sem link_ops_rwsem devices_rwsem rdma_nets_rwsem &device->compat_devs_mutex &p->pi_lock irq_context: 0 &rdma_nl_types[idx].sem link_ops_rwsem devices_rwsem rdma_nets_rwsem &device->compat_devs_mutex &p->pi_lock &rq->__lock irq_context: 0 &rdma_nl_types[idx].sem link_ops_rwsem devices_rwsem rdma_nets_rwsem &device->compat_devs_mutex &p->pi_lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 &rdma_nl_types[idx].sem link_ops_rwsem devices_rwsem rdma_nets_rwsem &device->compat_devs_mutex &root->kernfs_rwsem &sem->wait_lock irq_context: 0 &rdma_nl_types[idx].sem link_ops_rwsem devices_rwsem rdma_nets_rwsem &device->compat_devs_mutex &root->kernfs_rwsem &rq->__lock irq_context: 0 &rdma_nl_types[idx].sem link_ops_rwsem devices_rwsem rdma_nets_rwsem &device->compat_devs_mutex &root->kernfs_rwsem &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 &rdma_nl_types[idx].sem link_ops_rwsem devices_rwsem rdma_nets_rwsem &device->compat_devs_mutex &root->kernfs_rwsem rcu_read_lock &rq->__lock irq_context: 0 &rdma_nl_types[idx].sem link_ops_rwsem devices_rwsem rdma_nets_rwsem &device->compat_devs_mutex &root->kernfs_rwsem rcu_read_lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 &rdma_nl_types[idx].sem link_ops_rwsem devices_rwsem rdma_nets_rwsem &device->compat_devs_mutex remove_cache_srcu irq_context: 0 &rdma_nl_types[idx].sem link_ops_rwsem devices_rwsem rdma_nets_rwsem &device->compat_devs_mutex remove_cache_srcu quarantine_lock irq_context: 0 &rdma_nl_types[idx].sem link_ops_rwsem devices_rwsem rdma_nets_rwsem &device->compat_devs_mutex &root->kernfs_rwsem &root->kernfs_iattr_rwsem &rq->__lock irq_context: 0 &rdma_nl_types[idx].sem link_ops_rwsem devices_rwsem rdma_nets_rwsem &device->compat_devs_mutex rcu_read_lock rcu_node_0 irq_context: 0 &rdma_nl_types[idx].sem link_ops_rwsem devices_rwsem rdma_nets_rwsem &device->compat_devs_mutex rcu_read_lock &rq->__lock irq_context: 0 &rdma_nl_types[idx].sem link_ops_rwsem devices_rwsem rdma_nets_rwsem &device->compat_devs_mutex rcu_read_lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 &rdma_nl_types[idx].sem link_ops_rwsem devices_rwsem rdma_nets_rwsem &device->compat_devs_mutex lock kernfs_idr_lock &c->lock irq_context: 0 &rdma_nl_types[idx].sem link_ops_rwsem devices_rwsem rdma_nets_rwsem &device->compat_devs_mutex remove_cache_srcu &c->lock irq_context: 0 &rdma_nl_types[idx].sem link_ops_rwsem devices_rwsem rdma_nets_rwsem &device->compat_devs_mutex remove_cache_srcu &n->list_lock irq_context: 0 &rdma_nl_types[idx].sem link_ops_rwsem devices_rwsem rdma_nets_rwsem &device->compat_devs_mutex remove_cache_srcu rcu_read_lock pool_lock#2 irq_context: 0 &rdma_nl_types[idx].sem link_ops_rwsem devices_rwsem rdma_nets_rwsem &device->compat_devs_mutex remove_cache_srcu &obj_hash[i].lock irq_context: 0 &rdma_nl_types[idx].sem link_ops_rwsem devices_rwsem rdma_nets_rwsem &device->compat_devs_mutex lock kernfs_idr_lock &____s->seqcount#2 irq_context: 0 &rdma_nl_types[idx].sem link_ops_rwsem devices_rwsem rdma_nets_rwsem &device->compat_devs_mutex lock kernfs_idr_lock &____s->seqcount irq_context: 0 &rdma_nl_types[idx].sem link_ops_rwsem devices_rwsem rdma_nets_rwsem &device->compat_devs_mutex rcu_node_0 irq_context: 0 &rdma_nl_types[idx].sem link_ops_rwsem devices_rwsem rdma_nets_rwsem &device->compat_devs_mutex &root->kernfs_rwsem &root->kernfs_iattr_rwsem &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 &rdma_nl_types[idx].sem link_ops_rwsem devices_rwsem rdma_nets_rwsem &device->compat_devs_mutex fs_reclaim mmu_notifier_invalidate_range_start &rq->__lock irq_context: 0 &rdma_nl_types[idx].sem link_ops_rwsem devices_rwsem rdma_nets_rwsem &device->compat_devs_mutex fs_reclaim mmu_notifier_invalidate_range_start &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 &rdma_nl_types[idx].sem link_ops_rwsem devices_rwsem rdma_nets_rwsem &device->compat_devs_mutex batched_entropy_u8.lock irq_context: 0 &rdma_nl_types[idx].sem link_ops_rwsem devices_rwsem rdma_nets_rwsem &device->compat_devs_mutex kfence_freelist_lock irq_context: 0 &rdma_nl_types[idx].sem link_ops_rwsem devices_rwsem rdma_nets_rwsem &device->compat_devs_mutex remove_cache_srcu pool_lock#2 irq_context: 0 &rdma_nl_types[idx].sem link_ops_rwsem devices_rwsem rdma_nets_rwsem &device->compat_devs_mutex &pcp->lock &zone->lock irq_context: 0 &rdma_nl_types[idx].sem link_ops_rwsem uevent_sock_mutex irq_context: 0 &rdma_nl_types[idx].sem link_ops_rwsem uevent_sock_mutex fs_reclaim irq_context: 0 &rdma_nl_types[idx].sem link_ops_rwsem uevent_sock_mutex fs_reclaim mmu_notifier_invalidate_range_start irq_context: 0 &rdma_nl_types[idx].sem link_ops_rwsem uevent_sock_mutex pool_lock#2 irq_context: 0 &rdma_nl_types[idx].sem link_ops_rwsem uevent_sock_mutex nl_table_lock irq_context: 0 &rdma_nl_types[idx].sem link_ops_rwsem uevent_sock_mutex rlock-AF_NETLINK irq_context: 0 &rdma_nl_types[idx].sem link_ops_rwsem uevent_sock_mutex rcu_read_lock &ei->socket.wq.wait irq_context: 0 &rdma_nl_types[idx].sem link_ops_rwsem uevent_sock_mutex rcu_read_lock &ei->socket.wq.wait &ep->lock irq_context: 0 &rdma_nl_types[idx].sem link_ops_rwsem uevent_sock_mutex rcu_read_lock &ei->socket.wq.wait &ep->lock &ep->wq irq_context: 0 &rdma_nl_types[idx].sem link_ops_rwsem uevent_sock_mutex rcu_read_lock &ei->socket.wq.wait &ep->lock &ep->wq &p->pi_lock irq_context: 0 &rdma_nl_types[idx].sem link_ops_rwsem uevent_sock_mutex rcu_read_lock &ei->socket.wq.wait &ep->lock &ep->wq &p->pi_lock &rq->__lock irq_context: 0 &rdma_nl_types[idx].sem link_ops_rwsem uevent_sock_mutex rcu_read_lock &ei->socket.wq.wait &ep->lock &ep->wq &p->pi_lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 &rdma_nl_types[idx].sem link_ops_rwsem uevent_sock_mutex rcu_read_lock &ei->socket.wq.wait &ep->lock &ep->wq &p->pi_lock &rq->__lock &cfs_rq->removed.lock irq_context: 0 &rdma_nl_types[idx].sem link_ops_rwsem uevent_sock_mutex nl_table_wait.lock irq_context: 0 &mm->mmap_lock lock#4 &lruvec->lru_lock per_cpu_ptr(&cgroup_rstat_cpu_lock, cpu) irq_context: 0 tasklist_lock stock_lock rcu_read_lock per_cpu_ptr(&cgroup_rstat_cpu_lock, cpu) irq_context: 0 (wq_completion)events (work_completion)(&(&nsim_dev->trap_data->trap_report_dw)->work) &devlink->lock_key#4 &nsim_trap_data->trap_lock crngs.lock base_crng.lock irq_context: 0 (wq_completion)wg-kex-wg2#9 (work_completion)(&peer->transmit_handshake_work) &wg->static_identity.lock &handshake->lock crngs.lock base_crng.lock irq_context: 0 (wq_completion)wg-kex-wg2#9 (work_completion)(&peer->transmit_handshake_work) batched_entropy_u8.lock crngs.lock irq_context: 0 sb_writers#4 &type->i_mutex_dir_key#3/1 jbd2_handle &ei->i_data_sem batched_entropy_u8.lock irq_context: 0 sb_writers#4 &type->i_mutex_dir_key#3/1 jbd2_handle &ei->i_data_sem kfence_freelist_lock irq_context: 0 sb_writers#4 &type->i_mutex_dir_key#3/1 jbd2_handle &ei->i_data_sem &meta->lock irq_context: 0 (wq_completion)rcu_gp (work_completion)(&sdp->work) &rcu_state.expedited_wq &p->pi_lock irq_context: 0 (wq_completion)rcu_gp (work_completion)(&sdp->work) &rcu_state.expedited_wq &p->pi_lock &rq->__lock irq_context: 0 (wq_completion)rcu_gp (work_completion)(&sdp->work) &rcu_state.expedited_wq &p->pi_lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 sb_writers#4 &sb->s_type->i_mutex_key#8 jbd2_handle &ei->i_data_sem &lg->lg_mutex &c->lock irq_context: 0 sb_writers#4 &sb->s_type->i_mutex_key#8 jbd2_handle &ei->i_es_lock key#5 irq_context: 0 sb_writers#4 &sb->s_type->i_mutex_key#8 &mm->mmap_lock rcu_read_lock &rcu_state.gp_wq irq_context: 0 sb_writers#4 &sb->s_type->i_mutex_key#8 &mm->mmap_lock rcu_read_lock &rcu_state.gp_wq &p->pi_lock irq_context: 0 sb_writers#4 &sb->s_type->i_mutex_key#8 &mm->mmap_lock rcu_read_lock &rcu_state.gp_wq &p->pi_lock &rq->__lock irq_context: 0 sb_writers#4 &sb->s_type->i_mutex_key#8 &mm->mmap_lock rcu_read_lock &rcu_state.gp_wq &p->pi_lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 &nft_net->commit_mutex rcu_node_0 irq_context: 0 &nft_net->commit_mutex rcu_read_lock rcu_read_lock &pool->lock irq_context: 0 &nft_net->commit_mutex rcu_read_lock rcu_read_lock &pool->lock &obj_hash[i].lock irq_context: 0 &nft_net->commit_mutex rcu_read_lock rcu_read_lock &pool->lock pool_lock#2 irq_context: 0 &nft_net->commit_mutex rcu_read_lock rcu_read_lock &pool->lock &p->pi_lock irq_context: 0 &nft_net->commit_mutex rcu_read_lock rcu_read_lock &pool->lock &p->pi_lock &rq->__lock irq_context: 0 &nft_net->commit_mutex rcu_read_lock rcu_read_lock &pool->lock &p->pi_lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 &nft_net->commit_mutex rcu_read_lock &rcu_state.gp_wq irq_context: 0 &nft_net->commit_mutex rcu_read_lock &rcu_state.gp_wq &p->pi_lock irq_context: 0 &nft_net->commit_mutex rcu_read_lock &rcu_state.gp_wq &p->pi_lock &rq->__lock irq_context: 0 &nft_net->commit_mutex rcu_read_lock &rcu_state.gp_wq &p->pi_lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 sb_writers#4 &type->i_mutex_dir_key#3 &sem->wait_lock irq_context: 0 sb_writers#4 &iint->mutex ima_extend_list_mutex fs_reclaim irq_context: 0 sb_writers#4 &iint->mutex ima_extend_list_mutex fs_reclaim mmu_notifier_invalidate_range_start irq_context: 0 &nft_net->commit_mutex rcu_state.exp_mutex.wait_lock irq_context: 0 &nft_net->commit_mutex &p->pi_lock irq_context: 0 &nft_net->commit_mutex &p->pi_lock &rq->__lock irq_context: 0 &nft_net->commit_mutex &p->pi_lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 &nft_net->commit_mutex rcu_read_lock &pool->lock irq_context: 0 &nft_net->commit_mutex &rnp->exp_lock irq_context: 0 &nft_net->commit_mutex rcu_state.exp_mutex irq_context: 0 &nft_net->commit_mutex rcu_state.exp_mutex rcu_state.exp_mutex.wait_lock irq_context: 0 rtnl_mutex &tun->lock irq_context: 0 rtnl_mutex wlock-AF_UNSPEC irq_context: 0 rtnl_mutex elock-AF_UNSPEC irq_context: 0 rtnl_mutex uevent_sock_mutex fs_reclaim &rq->__lock irq_context: 0 rtnl_mutex uevent_sock_mutex fs_reclaim &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 rtnl_mutex rcu_state.exp_mutex &cfs_rq->removed.lock irq_context: 0 rtnl_mutex rcu_state.exp_mutex pool_lock#2 irq_context: 0 (wq_completion)gid-cache-wq (work_completion)(&work->work) devices_rwsem &pdata->netdev_lock irq_context: 0 (wq_completion)events (work_completion)(&w->work)#2 nf_conntrack_mutex &rq->__lock &cfs_rq->removed.lock irq_context: 0 &rdma_nl_types[idx].sem link_ops_rwsem console_lock console_srcu console_owner_lock irq_context: 0 &rdma_nl_types[idx].sem link_ops_rwsem console_lock console_srcu console_owner irq_context: 0 &rdma_nl_types[idx].sem link_ops_rwsem console_lock console_srcu console_owner &port_lock_key irq_context: 0 &rdma_nl_types[idx].sem link_ops_rwsem console_lock console_srcu console_owner console_owner_lock irq_context: 0 rtnl_mutex &dev->tx_global_lock _xmit_NETROM irq_context: 0 (wq_completion)rcu_gp (work_completion)(&rew->rew_work) rcu_state.exp_wake_mutex &rnp->exp_wq[1] &p->pi_lock &rq->__lock &cfs_rq->removed.lock irq_context: 0 rtnl_mutex dev->qdisc_tx_busylock ?: &qdisc_tx_busylock irq_context: 0 rtnl_mutex dev->qdisc_tx_busylock ?: &qdisc_tx_busylock &sch->q.lock irq_context: 0 rtnl_mutex sysctl_lock krc.lock &obj_hash[i].lock irq_context: 0 rtnl_mutex sysctl_lock krc.lock &base->lock irq_context: 0 rtnl_mutex sysctl_lock krc.lock &base->lock &obj_hash[i].lock irq_context: 0 sb_writers#4 &sb->s_type->i_mutex_key#8 &mm->mmap_lock rcu_read_lock &rcu_state.expedited_wq irq_context: 0 sb_writers#4 &sb->s_type->i_mutex_key#8 &mm->mmap_lock rcu_read_lock &rcu_state.expedited_wq &p->pi_lock irq_context: 0 sb_writers#4 &sb->s_type->i_mutex_key#8 &mm->mmap_lock rcu_read_lock &rcu_state.expedited_wq &p->pi_lock &rq->__lock irq_context: 0 sb_writers#4 &sb->s_type->i_mutex_key#8 &mm->mmap_lock rcu_read_lock &rcu_state.expedited_wq &p->pi_lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 &rdma_nl_types[idx].sem &rq->__lock irq_context: 0 &tun->lock irq_context: 0 &rdma_nl_types[idx].sem link_ops_rwsem &pdata->netdev_lock ndev_hash_lock irq_context: 0 &rdma_nl_types[idx].sem link_ops_rwsem &xa->xa_lock#19 irq_context: 0 &rdma_nl_types[idx].sem link_ops_rwsem krc.lock irq_context: 0 &rdma_nl_types[idx].sem link_ops_rwsem &xa->xa_lock#17 irq_context: 0 sb_writers#4 &journal->j_list_lock irq_context: 0 &rdma_nl_types[idx].sem link_ops_rwsem &rq->__lock &cfs_rq->removed.lock irq_context: 0 sb_writers#4 sb_internal jbd2_handle &ei->i_data_sem rcu_read_lock &rcu_state.gp_wq irq_context: 0 sb_writers#4 sb_internal jbd2_handle &ei->i_data_sem rcu_read_lock &rcu_state.gp_wq &p->pi_lock irq_context: 0 sb_writers#4 sb_internal jbd2_handle &ei->i_data_sem rcu_read_lock &rcu_state.gp_wq &p->pi_lock &rq->__lock irq_context: 0 sb_writers#4 sb_internal jbd2_handle &ei->i_data_sem rcu_read_lock &rcu_state.gp_wq &p->pi_lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)gid-cache-wq (work_completion)(&ndev_work->work) devices_rwsem &pdata->netdev_lock irq_context: 0 rtnl_mutex &app->lock#2 pool_lock#2 irq_context: 0 rtnl_mutex &app->lock pool_lock#2 irq_context: 0 rtnl_mutex &ndev->lock &ifa->lock rcu_read_lock &pool->lock irq_context: 0 rtnl_mutex &ndev->lock &ifa->lock rcu_read_lock &pool->lock &obj_hash[i].lock irq_context: 0 sb_writers#4 &type->i_mutex_dir_key#3/1 sb_internal jbd2_handle &p->pi_lock &rq->__lock irq_context: 0 sb_writers#4 &type->i_mutex_dir_key#3/1 sb_internal jbd2_handle &p->pi_lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: softirq (&app->join_timer) &app->lock &pcp->lock &zone->lock irq_context: softirq (&app->join_timer) &app->lock &____s->seqcount irq_context: softirq (&app->join_timer) &app->lock pool_lock#2 irq_context: softirq (&app->join_timer) &app->lock rcu_read_lock pool_lock#2 irq_context: softirq (&app->join_timer) &app->lock &list->lock#11 irq_context: softirq (&app->join_timer) rcu_read_lock_bh rcu_read_lock tk_core.seq.seqcount irq_context: softirq (&app->join_timer) rcu_read_lock_bh rcu_read_lock &list->lock#5 irq_context: softirq (&app->join_timer)#2 &app->lock#2 pool_lock#2 irq_context: softirq (&app->join_timer)#2 &app->lock#2 &c->lock irq_context: softirq (&app->join_timer)#2 &app->lock#2 &list->lock#12 irq_context: softirq (&app->join_timer)#2 rcu_read_lock_bh rcu_read_lock tk_core.seq.seqcount irq_context: softirq (&app->join_timer)#2 rcu_read_lock_bh rcu_read_lock &list->lock#5 irq_context: softirq rcu_read_lock &sap->sk_lock irq_context: softirq (&app->join_timer) &app->lock &c->lock irq_context: 0 &mm->mmap_lock sb_writers#4 jbd2_handle rcu_read_lock rcu_node_0 irq_context: 0 sb_writers#4 &type->i_mutex_dir_key#3 jbd2_handle &bgl->locks[i].lock irq_context: 0 sb_writers#3 &sb->s_type->i_mutex_key#9 fs_reclaim &rq->__lock irq_context: 0 sb_writers#3 &sb->s_type->i_mutex_key#9 fs_reclaim &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 rtnl_mutex _xmit_IPGRE &c->lock irq_context: 0 &nft_net->commit_mutex batched_entropy_u32.lock crngs.lock irq_context: 0 &sb->s_type->i_mutex_key#10 (netlink_chain).rwsem fill_pool_map-wait-type-override &rq->__lock irq_context: 0 &sb->s_type->i_mutex_key#10 (netlink_chain).rwsem fill_pool_map-wait-type-override &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: softirq (&app->join_timer) rcu_read_lock_bh &obj_hash[i].lock irq_context: softirq (&app->join_timer) rcu_read_lock_bh pool_lock#2 irq_context: 0 (wq_completion)bat_events (work_completion)(&(&bat_priv->nc.work)->work) rcu_read_lock &rcu_state.gp_wq &p->pi_lock &rq->__lock &cfs_rq->removed.lock irq_context: 0 sb_writers#4 &type->i_mutex_dir_key#3/1 &type->i_mutex_dir_key#3 rcu_read_lock &rcu_state.gp_wq irq_context: 0 sb_writers#4 &type->i_mutex_dir_key#3/1 &type->i_mutex_dir_key#3 rcu_read_lock &rcu_state.gp_wq &p->pi_lock irq_context: 0 sb_writers#4 &type->i_mutex_dir_key#3/1 &type->i_mutex_dir_key#3 rcu_read_lock &rcu_state.gp_wq &p->pi_lock &rq->__lock irq_context: 0 sb_writers#4 &type->i_mutex_dir_key#3/1 &type->i_mutex_dir_key#3 rcu_read_lock &rcu_state.gp_wq &p->pi_lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 rcu_read_lock rcu_read_lock &rq->__lock &cfs_rq->removed.lock irq_context: softirq (&app->join_timer)#2 rcu_read_lock_bh &obj_hash[i].lock irq_context: softirq (&app->join_timer)#2 rcu_read_lock_bh pool_lock#2 irq_context: 0 sb_writers#4 sb_internal jbd2_handle &ei->i_data_sem &n->list_lock irq_context: 0 sb_writers#4 sb_internal jbd2_handle &ei->i_data_sem &n->list_lock &c->lock irq_context: 0 sb_writers#4 &type->i_mutex_dir_key#3 rcu_read_lock rcu_read_lock rcu_node_0 irq_context: 0 sb_writers#4 &type->i_mutex_dir_key#3 rcu_read_lock rcu_read_lock &rq->__lock irq_context: 0 sb_writers#4 &type->i_mutex_dir_key#3 rcu_read_lock rcu_read_lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 br_ioctl_mutex rtnl_mutex fs_reclaim irq_context: 0 br_ioctl_mutex rtnl_mutex fs_reclaim mmu_notifier_invalidate_range_start irq_context: 0 br_ioctl_mutex rtnl_mutex stock_lock irq_context: 0 br_ioctl_mutex rtnl_mutex &c->lock irq_context: 0 br_ioctl_mutex rtnl_mutex pool_lock#2 irq_context: 0 br_ioctl_mutex rtnl_mutex stack_depot_init_mutex irq_context: 0 br_ioctl_mutex rtnl_mutex crngs.lock irq_context: 0 br_ioctl_mutex rtnl_mutex &obj_hash[i].lock irq_context: 0 br_ioctl_mutex rtnl_mutex &obj_hash[i].lock pool_lock irq_context: 0 br_ioctl_mutex rtnl_mutex &rq->__lock irq_context: 0 br_ioctl_mutex rtnl_mutex &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 br_ioctl_mutex rtnl_mutex pcpu_alloc_mutex irq_context: 0 br_ioctl_mutex rtnl_mutex pcpu_alloc_mutex pcpu_lock irq_context: 0 br_ioctl_mutex rtnl_mutex pcpu_alloc_mutex &rq->__lock irq_context: 0 br_ioctl_mutex rtnl_mutex pcpu_alloc_mutex &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 br_ioctl_mutex rtnl_mutex batched_entropy_u32.lock irq_context: 0 br_ioctl_mutex rtnl_mutex &xa->xa_lock#3 irq_context: 0 br_ioctl_mutex rtnl_mutex net_rwsem irq_context: 0 br_ioctl_mutex rtnl_mutex net_rwsem &list->lock#2 irq_context: 0 br_ioctl_mutex rtnl_mutex &tn->lock irq_context: 0 br_ioctl_mutex rtnl_mutex &x->wait#9 irq_context: 0 br_ioctl_mutex rtnl_mutex &k->list_lock irq_context: 0 br_ioctl_mutex rtnl_mutex gdp_mutex irq_context: 0 br_ioctl_mutex rtnl_mutex gdp_mutex &k->list_lock irq_context: 0 br_ioctl_mutex rtnl_mutex lock irq_context: 0 br_ioctl_mutex rtnl_mutex lock kernfs_idr_lock irq_context: 0 br_ioctl_mutex rtnl_mutex lock kernfs_idr_lock &c->lock irq_context: 0 br_ioctl_mutex rtnl_mutex lock kernfs_idr_lock pool_lock#2 irq_context: 0 br_ioctl_mutex rtnl_mutex &root->kernfs_rwsem irq_context: 0 br_ioctl_mutex rtnl_mutex &root->kernfs_rwsem &root->kernfs_iattr_rwsem irq_context: 0 br_ioctl_mutex rtnl_mutex bus_type_sem irq_context: 0 br_ioctl_mutex rtnl_mutex sysfs_symlink_target_lock irq_context: 0 br_ioctl_mutex rtnl_mutex lock kernfs_idr_lock &____s->seqcount irq_context: 0 br_ioctl_mutex rtnl_mutex &root->kernfs_rwsem irq_context: 0 br_ioctl_mutex rtnl_mutex &dev->power.lock irq_context: 0 br_ioctl_mutex rtnl_mutex dpm_list_mtx irq_context: 0 br_ioctl_mutex rtnl_mutex uevent_sock_mutex irq_context: 0 br_ioctl_mutex rtnl_mutex uevent_sock_mutex fs_reclaim irq_context: 0 br_ioctl_mutex rtnl_mutex uevent_sock_mutex fs_reclaim mmu_notifier_invalidate_range_start irq_context: 0 br_ioctl_mutex rtnl_mutex uevent_sock_mutex fs_reclaim mmu_notifier_invalidate_range_start &rq->__lock irq_context: 0 br_ioctl_mutex rtnl_mutex uevent_sock_mutex fs_reclaim mmu_notifier_invalidate_range_start &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 br_ioctl_mutex rtnl_mutex uevent_sock_mutex pool_lock#2 irq_context: 0 br_ioctl_mutex rtnl_mutex uevent_sock_mutex nl_table_lock irq_context: 0 br_ioctl_mutex rtnl_mutex uevent_sock_mutex &obj_hash[i].lock irq_context: 0 br_ioctl_mutex rtnl_mutex uevent_sock_mutex nl_table_wait.lock irq_context: 0 br_ioctl_mutex rtnl_mutex subsys mutex#17 irq_context: 0 br_ioctl_mutex rtnl_mutex subsys mutex#17 &k->k_lock irq_context: 0 br_ioctl_mutex rtnl_mutex &dir->lock#2 irq_context: 0 br_ioctl_mutex rtnl_mutex dev_hotplug_mutex irq_context: 0 br_ioctl_mutex rtnl_mutex dev_hotplug_mutex &rq->__lock irq_context: 0 br_ioctl_mutex rtnl_mutex dev_hotplug_mutex &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 br_ioctl_mutex rtnl_mutex dev_hotplug_mutex &dev->power.lock irq_context: 0 br_ioctl_mutex rtnl_mutex dev_base_lock irq_context: 0 br_ioctl_mutex rtnl_mutex input_pool.lock irq_context: 0 br_ioctl_mutex rtnl_mutex rcu_read_lock &pool->lock/1 irq_context: 0 br_ioctl_mutex rtnl_mutex rcu_read_lock &pool->lock/1 &obj_hash[i].lock irq_context: 0 br_ioctl_mutex rtnl_mutex rcu_read_lock &pool->lock/1 pool_lock#2 irq_context: 0 br_ioctl_mutex rtnl_mutex rcu_read_lock &pool->lock/1 &p->pi_lock irq_context: 0 br_ioctl_mutex rtnl_mutex rcu_read_lock &pool->lock/1 &p->pi_lock &rq->__lock irq_context: 0 br_ioctl_mutex rtnl_mutex rcu_read_lock &pool->lock/1 &p->pi_lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 br_ioctl_mutex rtnl_mutex &tbl->lock irq_context: 0 br_ioctl_mutex rtnl_mutex sysctl_lock irq_context: 0 br_ioctl_mutex rtnl_mutex nl_table_lock irq_context: 0 br_ioctl_mutex rtnl_mutex nl_table_wait.lock irq_context: 0 br_ioctl_mutex rtnl_mutex failover_lock irq_context: 0 br_ioctl_mutex rtnl_mutex &n->list_lock irq_context: 0 br_ioctl_mutex rtnl_mutex &n->list_lock &c->lock irq_context: 0 br_ioctl_mutex rtnl_mutex proc_subdir_lock irq_context: 0 br_ioctl_mutex rtnl_mutex proc_inum_ida.xa_lock irq_context: 0 br_ioctl_mutex rtnl_mutex proc_subdir_lock irq_context: 0 br_ioctl_mutex rtnl_mutex &idev->mc_lock irq_context: 0 br_ioctl_mutex rtnl_mutex &idev->mc_lock fs_reclaim irq_context: 0 br_ioctl_mutex rtnl_mutex &idev->mc_lock fs_reclaim mmu_notifier_invalidate_range_start irq_context: 0 br_ioctl_mutex rtnl_mutex &idev->mc_lock fs_reclaim mmu_notifier_invalidate_range_start &rq->__lock irq_context: 0 br_ioctl_mutex rtnl_mutex &idev->mc_lock fs_reclaim mmu_notifier_invalidate_range_start &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 br_ioctl_mutex rtnl_mutex &idev->mc_lock pool_lock#2 irq_context: 0 br_ioctl_mutex rtnl_mutex &idev->mc_lock &obj_hash[i].lock irq_context: 0 br_ioctl_mutex rtnl_mutex &idev->mc_lock &bridge_netdev_addr_lock_key irq_context: 0 br_ioctl_mutex rtnl_mutex &idev->mc_lock &bridge_netdev_addr_lock_key &c->lock irq_context: 0 br_ioctl_mutex rtnl_mutex &idev->mc_lock &bridge_netdev_addr_lock_key pool_lock#2 irq_context: 0 br_ioctl_mutex rtnl_mutex remove_cache_srcu irq_context: 0 br_ioctl_mutex rtnl_mutex remove_cache_srcu quarantine_lock irq_context: 0 br_ioctl_mutex rtnl_mutex (switchdev_blocking_notif_chain).rwsem irq_context: 0 br_ioctl_mutex rtnl_mutex &br->hash_lock irq_context: 0 br_ioctl_mutex rtnl_mutex &br->hash_lock pool_lock#2 irq_context: 0 br_ioctl_mutex rtnl_mutex &br->hash_lock rcu_read_lock rhashtable_bucket irq_context: 0 br_ioctl_mutex rtnl_mutex &br->hash_lock &c->lock irq_context: 0 br_ioctl_mutex rtnl_mutex &br->hash_lock nl_table_lock irq_context: 0 br_ioctl_mutex rtnl_mutex &br->hash_lock &obj_hash[i].lock irq_context: 0 br_ioctl_mutex rtnl_mutex &br->hash_lock nl_table_wait.lock irq_context: 0 br_ioctl_mutex rtnl_mutex rcu_read_lock rhashtable_bucket irq_context: 0 br_ioctl_mutex rtnl_mutex &____s->seqcount irq_context: 0 br_ioctl_mutex rtnl_mutex rcu_read_lock pool_lock#2 irq_context: 0 br_ioctl_mutex rtnl_mutex rcu_read_lock rcu_node_0 irq_context: 0 br_ioctl_mutex rtnl_mutex rcu_read_lock &rq->__lock irq_context: 0 br_ioctl_mutex rtnl_mutex rcu_read_lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 br_ioctl_mutex rtnl_mutex &root->kernfs_rwsem &rq->__lock irq_context: 0 br_ioctl_mutex rtnl_mutex &root->kernfs_rwsem &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 &sig->cred_guard_mutex fs_reclaim mmu_notifier_invalidate_range_start &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 br_ioctl_mutex rtnl_mutex &pnettable->lock irq_context: 0 br_ioctl_mutex rtnl_mutex smc_ib_devices.mutex irq_context: 0 br_ioctl_mutex rtnl_mutex rcu_read_lock &ndev->lock irq_context: 0 br_ioctl_mutex rtnl_mutex.wait_lock irq_context: 0 br_ioctl_mutex &p->pi_lock irq_context: 0 br_ioctl_mutex &p->pi_lock &rq->__lock irq_context: 0 br_ioctl_mutex &p->pi_lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 br_ioctl_mutex &rq->__lock irq_context: 0 br_ioctl_mutex &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 purge_vmap_area_lock quarantine_lock irq_context: 0 &f->f_pos_lock &obj_hash[i].lock pool_lock irq_context: 0 sk_lock-AF_ALG &mm->mmap_lock sb_pagefaults &c->lock irq_context: 0 &mm->mmap_lock &rcu_state.expedited_wq &p->pi_lock &rq->__lock irq_context: 0 &mm->mmap_lock &rcu_state.expedited_wq &p->pi_lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 &xt[i].mutex &rcu_state.expedited_wq irq_context: 0 &xt[i].mutex &rcu_state.expedited_wq &p->pi_lock irq_context: 0 &xt[i].mutex &rcu_state.expedited_wq &p->pi_lock &rq->__lock irq_context: 0 &xt[i].mutex &rcu_state.expedited_wq &p->pi_lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 rtnl_mutex &newf->file_lock irq_context: 0 rtnl_mutex lock link_idr_lock irq_context: 0 rtnl_mutex lock link_idr_lock pool_lock#2 irq_context: 0 rtnl_mutex &sb->s_type->i_lock_key#15 irq_context: 0 rtnl_mutex &sb->s_type->i_lock_key#15 &dentry->d_lock irq_context: 0 rtnl_mutex bpf_dispatcher_xdp.mutex irq_context: 0 rtnl_mutex bpf_dispatcher_xdp.mutex pack_mutex irq_context: 0 rtnl_mutex bpf_dispatcher_xdp.mutex fs_reclaim irq_context: 0 rtnl_mutex bpf_dispatcher_xdp.mutex fs_reclaim mmu_notifier_invalidate_range_start irq_context: 0 rtnl_mutex bpf_dispatcher_xdp.mutex fs_reclaim mmu_notifier_invalidate_range_start &rq->__lock irq_context: 0 rtnl_mutex bpf_dispatcher_xdp.mutex fs_reclaim mmu_notifier_invalidate_range_start &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 rtnl_mutex bpf_dispatcher_xdp.mutex pool_lock#2 irq_context: 0 rtnl_mutex bpf_dispatcher_xdp.mutex free_vmap_area_lock irq_context: 0 rtnl_mutex bpf_dispatcher_xdp.mutex vmap_area_lock irq_context: 0 rtnl_mutex bpf_dispatcher_xdp.mutex &____s->seqcount irq_context: 0 rtnl_mutex bpf_dispatcher_xdp.mutex init_mm.page_table_lock irq_context: 0 rtnl_mutex bpf_dispatcher_xdp.mutex bpf_lock irq_context: 0 rtnl_mutex bpf_dispatcher_xdp.mutex text_mutex irq_context: 0 rtnl_mutex bpf_dispatcher_xdp.mutex text_mutex ptlock_ptr(page)#2 irq_context: 0 rtnl_mutex bpf_dispatcher_xdp.mutex cpu_hotplug_lock irq_context: 0 rtnl_mutex bpf_dispatcher_xdp.mutex cpu_hotplug_lock static_call_mutex irq_context: 0 rtnl_mutex bpf_dispatcher_xdp.mutex cpu_hotplug_lock static_call_mutex text_mutex irq_context: 0 rtnl_mutex bpf_dispatcher_xdp.mutex cpu_hotplug_lock static_call_mutex text_mutex ptlock_ptr(page)#2 irq_context: 0 rtnl_mutex bpf_dispatcher_xdp.mutex &rnp->exp_lock irq_context: 0 rtnl_mutex bpf_dispatcher_xdp.mutex rcu_state.exp_mutex irq_context: 0 rtnl_mutex bpf_dispatcher_xdp.mutex rcu_state.exp_mutex rcu_state.exp_mutex.wait_lock irq_context: 0 rtnl_mutex bpf_dispatcher_xdp.mutex rcu_state.exp_mutex &rq->__lock irq_context: 0 rtnl_mutex bpf_dispatcher_xdp.mutex rcu_state.exp_mutex &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 rtnl_mutex bpf_dispatcher_xdp.mutex rcu_state.exp_mutex rcu_node_0 irq_context: 0 rtnl_mutex bpf_dispatcher_xdp.mutex rcu_state.exp_mutex &obj_hash[i].lock irq_context: 0 rtnl_mutex bpf_dispatcher_xdp.mutex rcu_state.exp_mutex rcu_read_lock &pool->lock irq_context: 0 rtnl_mutex bpf_dispatcher_xdp.mutex rcu_state.exp_mutex rcu_read_lock &pool->lock &obj_hash[i].lock irq_context: 0 rtnl_mutex bpf_dispatcher_xdp.mutex rcu_state.exp_mutex rcu_read_lock &pool->lock &p->pi_lock irq_context: 0 rtnl_mutex bpf_dispatcher_xdp.mutex rcu_state.exp_mutex rcu_read_lock &pool->lock &p->pi_lock &rq->__lock irq_context: 0 rtnl_mutex bpf_dispatcher_xdp.mutex rcu_state.exp_mutex rcu_read_lock &pool->lock &p->pi_lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 rtnl_mutex bpf_dispatcher_xdp.mutex rcu_state.exp_mutex &rnp->exp_wq[1] irq_context: 0 sk_lock-AF_ALG remove_cache_srcu rcu_read_lock rcu_node_0 irq_context: 0 sk_lock-AF_ALG remove_cache_srcu rcu_read_lock &rq->__lock irq_context: 0 sk_lock-AF_ALG remove_cache_srcu rcu_read_lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 sk_lock-AF_ALG remove_cache_srcu &rq->__lock irq_context: 0 sk_lock-AF_ALG remove_cache_srcu &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 rtnl_mutex bpf_dispatcher_xdp.mutex &obj_hash[i].lock irq_context: 0 br_ioctl_mutex rtnl_mutex &br->hash_lock rcu_read_lock rcu_read_lock &pool->lock irq_context: 0 br_ioctl_mutex rtnl_mutex &br->hash_lock rcu_read_lock rcu_read_lock &pool->lock &obj_hash[i].lock irq_context: 0 br_ioctl_mutex rtnl_mutex &br->hash_lock rcu_read_lock rcu_read_lock &pool->lock pool_lock#2 irq_context: 0 br_ioctl_mutex rtnl_mutex &br->hash_lock rcu_read_lock rcu_read_lock &pool->lock &p->pi_lock irq_context: 0 br_ioctl_mutex rtnl_mutex &br->hash_lock rcu_read_lock rcu_read_lock &pool->lock &p->pi_lock &rq->__lock irq_context: 0 br_ioctl_mutex rtnl_mutex &br->hash_lock rcu_read_lock rcu_read_lock &pool->lock &p->pi_lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 br_ioctl_mutex rtnl_mutex (work_completion)(&(&br->gc_work)->work) irq_context: 0 br_ioctl_mutex rtnl_mutex kernfs_idr_lock irq_context: 0 br_ioctl_mutex rtnl_mutex &root->kernfs_rwsem kernfs_idr_lock irq_context: 0 br_ioctl_mutex rtnl_mutex &root->kernfs_rwsem &obj_hash[i].lock irq_context: 0 br_ioctl_mutex rtnl_mutex &root->kernfs_rwsem pool_lock#2 irq_context: 0 br_ioctl_mutex rtnl_mutex rcu_state.exp_mutex rcu_node_0 irq_context: 0 br_ioctl_mutex rtnl_mutex rcu_state.exp_mutex &obj_hash[i].lock irq_context: 0 br_ioctl_mutex rtnl_mutex rcu_state.exp_mutex rcu_read_lock &pool->lock irq_context: 0 br_ioctl_mutex rtnl_mutex rcu_state.exp_mutex rcu_read_lock &pool->lock &obj_hash[i].lock irq_context: 0 br_ioctl_mutex rtnl_mutex rcu_state.exp_mutex rcu_read_lock &pool->lock &p->pi_lock irq_context: 0 br_ioctl_mutex rtnl_mutex rcu_state.exp_mutex &rnp->exp_wq[2] irq_context: 0 br_ioctl_mutex rtnl_mutex rcu_state.exp_mutex &rq->__lock irq_context: 0 br_ioctl_mutex rtnl_mutex rcu_state.exp_mutex &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 br_ioctl_mutex rtnl_mutex dev_base_lock &xa->xa_lock#3 irq_context: 0 br_ioctl_mutex rtnl_mutex cpu_hotplug_lock irq_context: 0 br_ioctl_mutex rtnl_mutex cpu_hotplug_lock &list->lock#5 irq_context: 0 br_ioctl_mutex rtnl_mutex rcu_state.exp_mutex rcu_read_lock &pool->lock &p->pi_lock &rq->__lock irq_context: 0 br_ioctl_mutex rtnl_mutex rcu_state.exp_mutex rcu_read_lock &pool->lock &p->pi_lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 br_ioctl_mutex rtnl_mutex rcu_state.exp_mutex &rnp->exp_wq[3] irq_context: 0 br_ioctl_mutex rtnl_mutex cpu_hotplug_lock jump_label_mutex irq_context: 0 br_ioctl_mutex rtnl_mutex cpu_hotplug_lock jump_label_mutex text_mutex irq_context: 0 br_ioctl_mutex rtnl_mutex cpu_hotplug_lock jump_label_mutex text_mutex ptlock_ptr(page)#2 irq_context: 0 br_ioctl_mutex rtnl_mutex bpf_dispatcher_xdp.mutex irq_context: 0 br_ioctl_mutex rtnl_mutex bpf_dispatcher_xdp.mutex cpu_hotplug_lock irq_context: 0 br_ioctl_mutex rtnl_mutex bpf_dispatcher_xdp.mutex cpu_hotplug_lock static_call_mutex irq_context: 0 br_ioctl_mutex rtnl_mutex bpf_dispatcher_xdp.mutex cpu_hotplug_lock static_call_mutex text_mutex irq_context: 0 br_ioctl_mutex rtnl_mutex bpf_dispatcher_xdp.mutex cpu_hotplug_lock static_call_mutex text_mutex ptlock_ptr(page)#2 irq_context: 0 br_ioctl_mutex rtnl_mutex bpf_dispatcher_xdp.mutex rcu_state.exp_mutex rcu_node_0 irq_context: 0 br_ioctl_mutex rtnl_mutex bpf_dispatcher_xdp.mutex rcu_state.exp_mutex &obj_hash[i].lock irq_context: 0 br_ioctl_mutex rtnl_mutex bpf_dispatcher_xdp.mutex rcu_state.exp_mutex &rq->__lock irq_context: 0 br_ioctl_mutex rtnl_mutex bpf_dispatcher_xdp.mutex rcu_state.exp_mutex &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 br_ioctl_mutex rtnl_mutex bpf_dispatcher_xdp.mutex rcu_state.exp_mutex rcu_read_lock &pool->lock irq_context: 0 br_ioctl_mutex rtnl_mutex bpf_dispatcher_xdp.mutex rcu_state.exp_mutex rcu_read_lock &pool->lock &obj_hash[i].lock irq_context: 0 br_ioctl_mutex rtnl_mutex bpf_dispatcher_xdp.mutex rcu_state.exp_mutex rcu_read_lock &pool->lock &p->pi_lock irq_context: 0 br_ioctl_mutex rtnl_mutex bpf_dispatcher_xdp.mutex rcu_state.exp_mutex rcu_read_lock &pool->lock &p->pi_lock &rq->__lock irq_context: 0 br_ioctl_mutex rtnl_mutex bpf_dispatcher_xdp.mutex rcu_state.exp_mutex rcu_read_lock &pool->lock &p->pi_lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 br_ioctl_mutex rtnl_mutex bpf_dispatcher_xdp.mutex rcu_state.exp_mutex &rnp->exp_wq[0] irq_context: 0 br_ioctl_mutex rtnl_mutex bpf_dispatcher_xdp.mutex &obj_hash[i].lock irq_context: 0 br_ioctl_mutex rtnl_mutex bpf_devs_lock irq_context: 0 br_ioctl_mutex rtnl_mutex bpf_devs_lock &rq->__lock irq_context: 0 br_ioctl_mutex rtnl_mutex bpf_devs_lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 br_ioctl_mutex rtnl_mutex &hwstats->hwsdev_list_lock irq_context: 0 br_ioctl_mutex rtnl_mutex &in_dev->mc_tomb_lock irq_context: 0 br_ioctl_mutex rtnl_mutex sysctl_lock &obj_hash[i].lock irq_context: 0 br_ioctl_mutex rtnl_mutex sysctl_lock pool_lock#2 irq_context: 0 br_ioctl_mutex rtnl_mutex sysctl_lock krc.lock irq_context: 0 (wq_completion)gid-cache-wq (work_completion)(&ndev_work->work) devices_rwsem &table->lock#4 irq_context: 0 br_ioctl_mutex rtnl_mutex class irq_context: 0 br_ioctl_mutex rtnl_mutex (&tbl->proxy_timer) irq_context: 0 br_ioctl_mutex rtnl_mutex &base->lock irq_context: 0 br_ioctl_mutex rtnl_mutex &ul->lock irq_context: 0 br_ioctl_mutex rtnl_mutex &net->xdp.lock irq_context: 0 br_ioctl_mutex rtnl_mutex krc.lock irq_context: 0 br_ioctl_mutex rtnl_mutex mirred_list_lock irq_context: 0 br_ioctl_mutex rtnl_mutex &nft_net->commit_mutex irq_context: 0 br_ioctl_mutex rtnl_mutex rcu_read_lock &tb->tb6_lock irq_context: 0 br_ioctl_mutex rtnl_mutex rcu_read_lock &tb->tb6_lock &net->ipv6.fib6_walker_lock irq_context: 0 br_ioctl_mutex rtnl_mutex &ent->pde_unload_lock irq_context: 0 br_ioctl_mutex rtnl_mutex &net->ipv6.addrconf_hash_lock irq_context: 0 br_ioctl_mutex rtnl_mutex &ndev->lock irq_context: 0 br_ioctl_mutex rtnl_mutex &ndev->lock &obj_hash[i].lock irq_context: 0 br_ioctl_mutex rtnl_mutex &idev->mc_lock &bridge_netdev_addr_lock_key &obj_hash[i].lock irq_context: 0 br_ioctl_mutex rtnl_mutex &idev->mc_lock &bridge_netdev_addr_lock_key krc.lock irq_context: 0 br_ioctl_mutex rtnl_mutex rcu_state.exp_mutex &rnp->exp_wq[1] irq_context: 0 br_ioctl_mutex rtnl_mutex &idev->mc_query_lock irq_context: 0 br_ioctl_mutex rtnl_mutex &idev->mc_query_lock &obj_hash[i].lock irq_context: 0 br_ioctl_mutex rtnl_mutex (work_completion)(&(&idev->mc_report_work)->work) irq_context: 0 br_ioctl_mutex rtnl_mutex &idev->mc_report_lock irq_context: 0 br_ioctl_mutex rtnl_mutex &idev->mc_lock krc.lock irq_context: 0 br_ioctl_mutex rtnl_mutex &rq->__lock &cfs_rq->removed.lock irq_context: 0 br_ioctl_mutex rtnl_mutex (&brmctx->ip4_mc_router_timer) irq_context: 0 br_ioctl_mutex rtnl_mutex (&brmctx->ip4_other_query.timer) irq_context: 0 br_ioctl_mutex rtnl_mutex (&brmctx->ip4_own_query.timer) irq_context: 0 br_ioctl_mutex rtnl_mutex (&brmctx->ip6_mc_router_timer) irq_context: 0 br_ioctl_mutex rtnl_mutex (&brmctx->ip6_other_query.timer) irq_context: 0 br_ioctl_mutex rtnl_mutex (&brmctx->ip6_own_query.timer) irq_context: 0 br_ioctl_mutex rtnl_mutex &pnn->pndevs.lock irq_context: 0 br_ioctl_mutex rtnl_mutex &pnn->routes.lock irq_context: 0 br_ioctl_mutex rtnl_mutex target_list_lock irq_context: 0 br_ioctl_mutex rtnl_mutex &bridge_netdev_addr_lock_key irq_context: 0 br_ioctl_mutex rtnl_mutex net_rwsem &rq->__lock irq_context: 0 br_ioctl_mutex rtnl_mutex net_rwsem &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 br_ioctl_mutex rtnl_mutex &br->multicast_lock irq_context: 0 br_ioctl_mutex rtnl_mutex (work_completion)(&br->mcast_gc_work) irq_context: 0 br_ioctl_mutex rtnl_mutex rcu_state.barrier_mutex irq_context: 0 br_ioctl_mutex rtnl_mutex rcu_state.barrier_mutex rcu_state.barrier_lock irq_context: 0 br_ioctl_mutex rtnl_mutex rcu_state.barrier_mutex rcu_state.barrier_lock &obj_hash[i].lock irq_context: 0 br_ioctl_mutex rtnl_mutex rcu_state.barrier_mutex &x->wait#24 irq_context: 0 br_ioctl_mutex rtnl_mutex rcu_state.barrier_mutex &rq->__lock irq_context: 0 br_ioctl_mutex rtnl_mutex rcu_state.barrier_mutex &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 br_ioctl_mutex rtnl_mutex pcpu_lock irq_context: 0 br_ioctl_mutex rtnl_mutex (work_completion)(&ht->run_work) irq_context: 0 br_ioctl_mutex rtnl_mutex &ht->mutex irq_context: 0 br_ioctl_mutex rtnl_mutex &ht->mutex &obj_hash[i].lock irq_context: 0 br_ioctl_mutex rtnl_mutex &ht->mutex pool_lock#2 irq_context: 0 br_ioctl_mutex rtnl_mutex &ht->mutex &rq->__lock irq_context: 0 br_ioctl_mutex rtnl_mutex &ht->mutex &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 br_ioctl_mutex rtnl_mutex rcu_read_lock &pool->lock irq_context: 0 br_ioctl_mutex rtnl_mutex dev_pm_qos_sysfs_mtx irq_context: 0 br_ioctl_mutex rtnl_mutex dev_pm_qos_sysfs_mtx &root->kernfs_rwsem irq_context: 0 br_ioctl_mutex rtnl_mutex dev_pm_qos_sysfs_mtx &root->kernfs_rwsem irq_context: 0 br_ioctl_mutex rtnl_mutex dev_pm_qos_sysfs_mtx dev_pm_qos_mtx irq_context: 0 br_ioctl_mutex rtnl_mutex subsys mutex#17 &k->k_lock klist_remove_lock irq_context: 0 br_ioctl_mutex rtnl_mutex &root->kernfs_rwsem kernfs_idr_lock &obj_hash[i].lock irq_context: 0 br_ioctl_mutex rtnl_mutex &root->kernfs_rwsem kernfs_idr_lock pool_lock#2 irq_context: 0 br_ioctl_mutex rtnl_mutex deferred_probe_mutex irq_context: 0 br_ioctl_mutex rtnl_mutex deferred_probe_mutex &rq->__lock irq_context: 0 br_ioctl_mutex rtnl_mutex deferred_probe_mutex &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 br_ioctl_mutex rtnl_mutex device_links_lock irq_context: 0 br_ioctl_mutex rtnl_mutex mmu_notifier_invalidate_range_start irq_context: 0 br_ioctl_mutex rtnl_mutex uevent_sock_mutex mmu_notifier_invalidate_range_start irq_context: 0 br_ioctl_mutex rtnl_mutex uevent_sock_mutex &c->lock irq_context: 0 br_ioctl_mutex rtnl_mutex kernfs_idr_lock &obj_hash[i].lock irq_context: 0 br_ioctl_mutex rtnl_mutex kernfs_idr_lock pool_lock#2 irq_context: 0 br_ioctl_mutex rtnl_mutex cpu_hotplug_lock xps_map_mutex irq_context: 0 br_ioctl_mutex rcu_state.barrier_mutex irq_context: 0 br_ioctl_mutex rcu_state.barrier_mutex &rq->__lock irq_context: 0 br_ioctl_mutex rcu_state.barrier_mutex &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 br_ioctl_mutex rcu_state.barrier_mutex rcu_state.barrier_lock irq_context: 0 br_ioctl_mutex rcu_state.barrier_mutex rcu_state.barrier_lock &obj_hash[i].lock irq_context: 0 br_ioctl_mutex rcu_state.barrier_mutex &x->wait#24 irq_context: 0 br_ioctl_mutex dev_base_lock irq_context: 0 br_ioctl_mutex lweventlist_lock irq_context: 0 br_ioctl_mutex stock_lock irq_context: 0 br_ioctl_mutex &obj_hash[i].lock irq_context: 0 br_ioctl_mutex pool_lock#2 irq_context: 0 br_ioctl_mutex krc.lock irq_context: 0 br_ioctl_mutex &dir->lock#2 irq_context: 0 br_ioctl_mutex &dir->lock#2 &obj_hash[i].lock irq_context: 0 br_ioctl_mutex &dir->lock#2 pool_lock#2 irq_context: 0 br_ioctl_mutex netdev_unregistering_wq.lock irq_context: 0 &sb->s_type->i_mutex_key#8 rcu_read_lock fill_pool_map-wait-type-override rcu_node_0 irq_context: 0 &sb->s_type->i_mutex_key#8 rcu_read_lock fill_pool_map-wait-type-override &rq->__lock irq_context: 0 &sb->s_type->i_mutex_key#8 rcu_read_lock fill_pool_map-wait-type-override &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 sk_lock-AF_ALG &mm->mmap_lock sb_pagefaults jbd2_handle &rq->__lock irq_context: 0 (wq_completion)events (work_completion)(&(&nsim_dev->trap_data->trap_report_dw)->work) &devlink->lock_key#6 &cfs_rq->removed.lock irq_context: 0 (wq_completion)events (work_completion)(&(&nsim_dev->trap_data->trap_report_dw)->work) &devlink->lock_key#6 &obj_hash[i].lock irq_context: softirq (&app->join_timer) &app->lock &n->list_lock irq_context: softirq (&app->join_timer) &app->lock &n->list_lock &c->lock irq_context: softirq (&app->join_timer) rcu_read_lock_bh rcu_read_lock pool_lock#2 irq_context: softirq (&app->join_timer) rcu_read_lock_bh rcu_read_lock rlock-AF_PACKET irq_context: 0 sk_lock-AF_ALG rcu_read_lock &ei->socket.wq.wait &p->pi_lock &rq->__lock &cfs_rq->removed.lock irq_context: 0 sk_lock-AF_ALG rcu_read_lock per_cpu_ptr(&cgroup_rstat_cpu_lock, cpu) irq_context: 0 sb_writers#4 stock_lock rcu_read_lock per_cpu_ptr(&cgroup_rstat_cpu_lock, cpu) irq_context: 0 sk_lock-AF_ALG batched_entropy_u8.lock irq_context: 0 sk_lock-AF_ALG kfence_freelist_lock irq_context: 0 sk_lock-AF_ALG &meta->lock irq_context: softirq (&app->join_timer) rcu_read_lock_bh rcu_read_lock &obj_hash[i].lock irq_context: 0 sk_lock-AF_ALG quarantine_lock irq_context: softirq &(&ssp->srcu_sup->work)->timer rcu_read_lock &pool->lock &p->pi_lock &rq->__lock &cfs_rq->removed.lock irq_context: 0 sb_writers#4 &type->i_mutex_dir_key#3 fs_reclaim mmu_notifier_invalidate_range_start &rq->__lock irq_context: 0 sb_writers#4 &type->i_mutex_dir_key#3 fs_reclaim mmu_notifier_invalidate_range_start &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 &sb->s_type->i_mutex_key#8 rcu_read_lock &p->pi_lock irq_context: 0 &sb->s_type->i_mutex_key#8 rcu_read_lock &p->pi_lock &rq->__lock irq_context: 0 &sb->s_type->i_mutex_key#8 rcu_read_lock &p->pi_lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 &sb->s_type->i_mutex_key#8 &sbi->s_writepages_rwsem jbd2_handle &ei->i_data_sem mmu_notifier_invalidate_range_start &rq->__lock irq_context: 0 sb_writers#8 kn->active#5 remove_cache_srcu pool_lock#2 irq_context: 0 br_ioctl_mutex rtnl_mutex uevent_sock_mutex &n->list_lock irq_context: 0 br_ioctl_mutex rtnl_mutex uevent_sock_mutex &n->list_lock &c->lock irq_context: 0 sb_writers#4 &type->i_mutex_dir_key#3/1 jbd2_handle &ei->i_data_sem &cfs_rq->removed.lock irq_context: 0 br_ioctl_mutex rtnl_mutex pcpu_alloc_mutex pcpu_alloc_mutex.wait_lock irq_context: 0 pernet_ops_rwsem pcpu_alloc_mutex.wait_lock irq_context: 0 br_ioctl_mutex rtnl_mutex pcpu_alloc_mutex.wait_lock irq_context: 0 br_ioctl_mutex rtnl_mutex &p->pi_lock irq_context: 0 br_ioctl_mutex rtnl_mutex &p->pi_lock &rq->__lock irq_context: 0 br_ioctl_mutex rtnl_mutex &p->pi_lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 br_ioctl_mutex rtnl_mutex &sem->wait_lock irq_context: 0 br_ioctl_mutex br_ioctl_mutex.wait_lock irq_context: 0 br_ioctl_mutex.wait_lock irq_context: 0 br_ioctl_mutex rtnl_mutex rtnl_mutex.wait_lock irq_context: 0 sk_lock-AF_INET6 (console_sem).lock irq_context: 0 sk_lock-AF_INET6 console_lock console_srcu console_owner_lock irq_context: 0 sk_lock-AF_INET6 console_lock console_srcu console_owner irq_context: 0 sk_lock-AF_INET6 console_lock console_srcu console_owner &port_lock_key irq_context: 0 sk_lock-AF_INET6 console_lock console_srcu console_owner console_owner_lock irq_context: 0 sk_lock-AF_ALG &obj_hash[i].lock pool_lock irq_context: 0 sk_lock-AF_ALG &cfs_rq->removed.lock irq_context: 0 pernet_ops_rwsem devices_rwsem rdma_nets_rwsem irq_context: 0 pernet_ops_rwsem devices_rwsem rdma_nets_rwsem &rq->__lock irq_context: 0 pernet_ops_rwsem devices_rwsem rdma_nets_rwsem &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 pernet_ops_rwsem devices_rwsem rdma_nets_rwsem &device->compat_devs_mutex irq_context: 0 pernet_ops_rwsem devices_rwsem rdma_nets_rwsem &device->compat_devs_mutex &rq->__lock irq_context: 0 pernet_ops_rwsem devices_rwsem rdma_nets_rwsem &device->compat_devs_mutex &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 pernet_ops_rwsem devices_rwsem rdma_nets_rwsem &device->compat_devs_mutex fs_reclaim irq_context: 0 pernet_ops_rwsem devices_rwsem rdma_nets_rwsem &device->compat_devs_mutex fs_reclaim mmu_notifier_invalidate_range_start irq_context: 0 pernet_ops_rwsem devices_rwsem rdma_nets_rwsem &device->compat_devs_mutex fs_reclaim mmu_notifier_invalidate_range_start &rq->__lock irq_context: 0 pernet_ops_rwsem devices_rwsem rdma_nets_rwsem &device->compat_devs_mutex fs_reclaim mmu_notifier_invalidate_range_start &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 pernet_ops_rwsem devices_rwsem rdma_nets_rwsem &device->compat_devs_mutex &xa->xa_lock#17 irq_context: 0 pernet_ops_rwsem devices_rwsem rdma_nets_rwsem &device->compat_devs_mutex pool_lock#2 irq_context: 0 pernet_ops_rwsem devices_rwsem rdma_nets_rwsem &device->compat_devs_mutex &x->wait#9 irq_context: 0 pernet_ops_rwsem devices_rwsem rdma_nets_rwsem &device->compat_devs_mutex &obj_hash[i].lock irq_context: 0 pernet_ops_rwsem devices_rwsem rdma_nets_rwsem &device->compat_devs_mutex &k->list_lock irq_context: 0 pernet_ops_rwsem devices_rwsem rdma_nets_rwsem &device->compat_devs_mutex gdp_mutex irq_context: 0 pernet_ops_rwsem devices_rwsem rdma_nets_rwsem &device->compat_devs_mutex gdp_mutex &k->list_lock irq_context: 0 pernet_ops_rwsem devices_rwsem rdma_nets_rwsem &device->compat_devs_mutex lock irq_context: 0 pernet_ops_rwsem devices_rwsem rdma_nets_rwsem &device->compat_devs_mutex lock kernfs_idr_lock irq_context: 0 pernet_ops_rwsem devices_rwsem rdma_nets_rwsem &device->compat_devs_mutex &root->kernfs_rwsem irq_context: 0 pernet_ops_rwsem devices_rwsem rdma_nets_rwsem &device->compat_devs_mutex &root->kernfs_rwsem &root->kernfs_iattr_rwsem irq_context: 0 pernet_ops_rwsem devices_rwsem rdma_nets_rwsem &device->compat_devs_mutex bus_type_sem irq_context: 0 pernet_ops_rwsem devices_rwsem rdma_nets_rwsem &device->compat_devs_mutex sysfs_symlink_target_lock irq_context: 0 pernet_ops_rwsem devices_rwsem rdma_nets_rwsem &device->compat_devs_mutex &c->lock irq_context: 0 pernet_ops_rwsem devices_rwsem rdma_nets_rwsem &device->compat_devs_mutex &____s->seqcount#2 irq_context: 0 pernet_ops_rwsem devices_rwsem rdma_nets_rwsem &device->compat_devs_mutex &____s->seqcount irq_context: 0 pernet_ops_rwsem devices_rwsem rdma_nets_rwsem &device->compat_devs_mutex &n->list_lock irq_context: 0 pernet_ops_rwsem devices_rwsem rdma_nets_rwsem &device->compat_devs_mutex &root->kernfs_rwsem irq_context: 0 pernet_ops_rwsem devices_rwsem rdma_nets_rwsem &device->compat_devs_mutex &dev->power.lock irq_context: 0 pernet_ops_rwsem devices_rwsem rdma_nets_rwsem &device->compat_devs_mutex dpm_list_mtx irq_context: 0 pernet_ops_rwsem devices_rwsem rdma_nets_rwsem &device->compat_devs_mutex uevent_sock_mutex irq_context: 0 pernet_ops_rwsem devices_rwsem rdma_nets_rwsem &device->compat_devs_mutex uevent_sock_mutex fs_reclaim irq_context: 0 pernet_ops_rwsem devices_rwsem rdma_nets_rwsem &device->compat_devs_mutex uevent_sock_mutex fs_reclaim mmu_notifier_invalidate_range_start irq_context: 0 pernet_ops_rwsem devices_rwsem rdma_nets_rwsem &device->compat_devs_mutex uevent_sock_mutex fs_reclaim mmu_notifier_invalidate_range_start &rq->__lock irq_context: 0 pernet_ops_rwsem devices_rwsem rdma_nets_rwsem &device->compat_devs_mutex uevent_sock_mutex fs_reclaim mmu_notifier_invalidate_range_start &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 rtnl_mutex bpf_dispatcher_xdp.mutex rcu_state.exp_mutex &rnp->exp_wq[2] irq_context: 0 pernet_ops_rwsem devices_rwsem rdma_nets_rwsem &device->compat_devs_mutex uevent_sock_mutex pool_lock#2 irq_context: 0 pernet_ops_rwsem devices_rwsem rdma_nets_rwsem &device->compat_devs_mutex uevent_sock_mutex nl_table_lock irq_context: 0 pernet_ops_rwsem devices_rwsem rdma_nets_rwsem &device->compat_devs_mutex uevent_sock_mutex &obj_hash[i].lock irq_context: 0 pernet_ops_rwsem devices_rwsem rdma_nets_rwsem &device->compat_devs_mutex uevent_sock_mutex nl_table_wait.lock irq_context: 0 pernet_ops_rwsem devices_rwsem rdma_nets_rwsem &device->compat_devs_mutex subsys mutex#84 irq_context: 0 pernet_ops_rwsem devices_rwsem rdma_nets_rwsem &device->compat_devs_mutex subsys mutex#84 &k->k_lock irq_context: 0 pernet_ops_rwsem devices_rwsem rdma_nets_rwsem &device->compat_devs_mutex subsys mutex#84 &rq->__lock irq_context: 0 pernet_ops_rwsem devices_rwsem rdma_nets_rwsem &device->compat_devs_mutex subsys mutex#84 &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 pernet_ops_rwsem devices_rwsem rdma_nets_rwsem &device->compat_devs_mutex &rxe->usdev_lock irq_context: 0 pernet_ops_rwsem devices_rwsem rdma_nets_rwsem &device->compat_devs_mutex &rxe->usdev_lock &pdata->netdev_lock irq_context: 0 pernet_ops_rwsem devices_rwsem rdma_nets_rwsem &device->compat_devs_mutex &rxe->usdev_lock rtnl_mutex irq_context: 0 pernet_ops_rwsem devices_rwsem rdma_nets_rwsem &device->compat_devs_mutex &rxe->usdev_lock rtnl_mutex rtnl_mutex.wait_lock irq_context: 0 pernet_ops_rwsem devices_rwsem rdma_nets_rwsem &device->compat_devs_mutex &rxe->usdev_lock rtnl_mutex &rq->__lock irq_context: 0 pernet_ops_rwsem devices_rwsem rdma_nets_rwsem &device->compat_devs_mutex &rxe->usdev_lock rtnl_mutex &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 pernet_ops_rwsem devices_rwsem rdma_nets_rwsem &device->compat_devs_mutex &lock->wait_lock irq_context: 0 rtnl_mutex bpf_dispatcher_xdp.mutex rcu_state.exp_mutex.wait_lock irq_context: 0 rtnl_mutex bpf_dispatcher_xdp.mutex &p->pi_lock irq_context: 0 rtnl_mutex bpf_dispatcher_xdp.mutex &p->pi_lock &rq->__lock irq_context: 0 rtnl_mutex bpf_dispatcher_xdp.mutex &p->pi_lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 br_ioctl_mutex rtnl_mutex &br->hash_lock &n->list_lock irq_context: 0 br_ioctl_mutex rtnl_mutex &br->hash_lock &n->list_lock &c->lock irq_context: 0 br_ioctl_mutex rtnl_mutex bpf_dispatcher_xdp.mutex rcu_state.exp_mutex rcu_read_lock &pool->lock &p->pi_lock &rq->__lock &cfs_rq->removed.lock irq_context: 0 br_ioctl_mutex rtnl_mutex bpf_dispatcher_xdp.mutex rcu_state.exp_mutex &rnp->exp_wq[3] irq_context: 0 br_ioctl_mutex rtnl_mutex &ul->lock#2 irq_context: 0 br_ioctl_mutex rtnl_mutex rcu_state.exp_mutex &rnp->exp_wq[0] irq_context: 0 br_ioctl_mutex rtnl_mutex sysctl_lock &obj_hash[i].lock pool_lock irq_context: 0 br_ioctl_mutex rtnl_mutex fill_pool_map-wait-type-override &rq->__lock irq_context: 0 br_ioctl_mutex rtnl_mutex fill_pool_map-wait-type-override &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 br_ioctl_mutex rtnl_mutex remove_cache_srcu &c->lock irq_context: 0 br_ioctl_mutex rtnl_mutex remove_cache_srcu &rq->__lock irq_context: 0 br_ioctl_mutex rtnl_mutex remove_cache_srcu &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 br_ioctl_mutex rtnl_mutex remove_cache_srcu &n->list_lock irq_context: 0 br_ioctl_mutex rtnl_mutex remove_cache_srcu rcu_read_lock pool_lock#2 irq_context: 0 br_ioctl_mutex rtnl_mutex remove_cache_srcu &obj_hash[i].lock irq_context: 0 sk_lock-AF_ALG &mm->mmap_lock sb_pagefaults jbd2_handle &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 br_ioctl_mutex rtnl_mutex rcu_state.exp_mutex &cfs_rq->removed.lock irq_context: 0 br_ioctl_mutex rtnl_mutex rcu_state.exp_mutex pool_lock#2 irq_context: 0 pernet_ops_rwsem devices_rwsem rdma_nets_rwsem &device->compat_devs_mutex &rxe->usdev_lock rtnl_mutex.wait_lock irq_context: 0 pernet_ops_rwsem devices_rwsem rdma_nets_rwsem &device->compat_devs_mutex &rxe->usdev_lock &rq->__lock irq_context: 0 pernet_ops_rwsem devices_rwsem rdma_nets_rwsem &device->compat_devs_mutex &rxe->usdev_lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 pernet_ops_rwsem devices_rwsem rdma_nets_rwsem &device->compat_devs_mutex &rxe->usdev_lock &p->pi_lock irq_context: 0 pernet_ops_rwsem devices_rwsem rdma_nets_rwsem &device->compat_devs_mutex &rxe->usdev_lock &p->pi_lock &rq->__lock irq_context: 0 pernet_ops_rwsem devices_rwsem rdma_nets_rwsem &device->compat_devs_mutex &rxe->usdev_lock &p->pi_lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 pernet_ops_rwsem devices_rwsem rdma_nets_rwsem &device->compat_devs_mutex &zone->lock irq_context: 0 pernet_ops_rwsem devices_rwsem rdma_nets_rwsem &device->compat_devs_mutex &n->list_lock &c->lock irq_context: 0 pernet_ops_rwsem devices_rwsem rdma_nets_rwsem &device->compat_devs_mutex &root->kernfs_rwsem &rq->__lock irq_context: 0 pernet_ops_rwsem devices_rwsem rdma_nets_rwsem &device->compat_devs_mutex &root->kernfs_rwsem &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 pernet_ops_rwsem devices_rwsem rdma_nets_rwsem &device->compat_devs_mutex lock kernfs_idr_lock pool_lock#2 irq_context: 0 pernet_ops_rwsem devices_rwsem rdma_nets_rwsem &device->compat_devs_mutex rcu_read_lock rcu_node_0 irq_context: 0 pernet_ops_rwsem devices_rwsem rdma_nets_rwsem &device->compat_devs_mutex rcu_read_lock &rq->__lock irq_context: 0 pernet_ops_rwsem devices_rwsem rdma_nets_rwsem &device->compat_devs_mutex rcu_read_lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 pernet_ops_rwsem devices_rwsem rdma_nets_rwsem &device->compat_devs_mutex rcu_read_lock &rcu_state.gp_wq irq_context: 0 pernet_ops_rwsem devices_rwsem rdma_nets_rwsem &device->compat_devs_mutex rcu_read_lock &rcu_state.gp_wq &p->pi_lock irq_context: 0 pernet_ops_rwsem devices_rwsem rdma_nets_rwsem &device->compat_devs_mutex rcu_read_lock &rcu_state.gp_wq &p->pi_lock &rq->__lock irq_context: 0 pernet_ops_rwsem devices_rwsem rdma_nets_rwsem &device->compat_devs_mutex rcu_read_lock &rcu_state.gp_wq &p->pi_lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 pernet_ops_rwsem devices_rwsem rdma_nets_rwsem &device->compat_devs_mutex lock kernfs_idr_lock &c->lock irq_context: 0 pernet_ops_rwsem devices_rwsem rdma_nets_rwsem &device->compat_devs_mutex batched_entropy_u8.lock irq_context: 0 pernet_ops_rwsem devices_rwsem rdma_nets_rwsem &device->compat_devs_mutex kfence_freelist_lock irq_context: 0 pernet_ops_rwsem devices_rwsem rdma_nets_rwsem &device->compat_devs_mutex &root->kernfs_rwsem &sem->wait_lock irq_context: 0 pernet_ops_rwsem devices_rwsem rdma_nets_rwsem &device->compat_devs_mutex &root->kernfs_rwsem &rq->__lock &cfs_rq->removed.lock irq_context: 0 pernet_ops_rwsem devices_rwsem rdma_nets_rwsem &device->compat_devs_mutex &root->kernfs_rwsem rcu_read_lock &rq->__lock irq_context: 0 pernet_ops_rwsem devices_rwsem rdma_nets_rwsem &device->compat_devs_mutex &root->kernfs_rwsem rcu_read_lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 pernet_ops_rwsem devices_rwsem rdma_nets_rwsem &device->compat_devs_mutex &sem->wait_lock irq_context: 0 pernet_ops_rwsem devices_rwsem rdma_nets_rwsem &device->compat_devs_mutex &p->pi_lock irq_context: 0 pernet_ops_rwsem devices_rwsem rdma_nets_rwsem &device->compat_devs_mutex &p->pi_lock &rq->__lock irq_context: 0 pernet_ops_rwsem devices_rwsem rdma_nets_rwsem &device->compat_devs_mutex &p->pi_lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 pernet_ops_rwsem devices_rwsem rdma_nets_rwsem &device->compat_devs_mutex rcu_node_0 irq_context: 0 pernet_ops_rwsem devices_rwsem rdma_nets_rwsem &device->compat_devs_mutex lock kernfs_idr_lock &____s->seqcount#2 irq_context: 0 pernet_ops_rwsem devices_rwsem rdma_nets_rwsem &device->compat_devs_mutex lock kernfs_idr_lock &____s->seqcount irq_context: 0 sk_lock-AF_ALG &mm->mmap_lock sb_pagefaults &n->list_lock irq_context: 0 sk_lock-AF_ALG &mm->mmap_lock sb_pagefaults &n->list_lock &c->lock irq_context: 0 pernet_ops_rwsem devices_rwsem rdma_nets_rwsem &device->compat_devs_mutex fs_reclaim &rq->__lock irq_context: 0 pernet_ops_rwsem devices_rwsem rdma_nets_rwsem &device->compat_devs_mutex &root->kernfs_rwsem &root->kernfs_iattr_rwsem &rq->__lock irq_context: 0 pernet_ops_rwsem devices_rwsem rdma_nets_rwsem &device->compat_devs_mutex &root->kernfs_rwsem &root->kernfs_iattr_rwsem &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 sk_lock-AF_ALG &mm->mmap_lock sb_pagefaults mapping.invalidate_lock rcu_read_lock rcu_node_0 irq_context: 0 sk_lock-AF_ALG &mm->mmap_lock sb_pagefaults mapping.invalidate_lock rcu_read_lock &rq->__lock irq_context: 0 pernet_ops_rwsem devices_rwsem rdma_nets_rwsem &device->compat_devs_mutex &rcu_state.expedited_wq irq_context: 0 kn->active#5 &kernfs_locks->open_file_mutex[count] rcu_node_0 irq_context: 0 &p->lock rcu_read_lock &rcu_state.gp_wq irq_context: 0 &p->lock rcu_read_lock &rcu_state.gp_wq &p->pi_lock irq_context: 0 &p->lock rcu_read_lock &rcu_state.gp_wq &p->pi_lock &rq->__lock irq_context: 0 &p->lock rcu_read_lock &rcu_state.gp_wq &p->pi_lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 pernet_ops_rwsem devices_rwsem rdma_nets_rwsem &device->compat_devs_mutex fs_reclaim &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 br_ioctl_mutex rtnl_mutex &____s->seqcount#2 irq_context: 0 br_ioctl_mutex rtnl_mutex &root->kernfs_rwsem &root->kernfs_iattr_rwsem &rq->__lock irq_context: 0 br_ioctl_mutex rtnl_mutex &root->kernfs_rwsem &root->kernfs_iattr_rwsem &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 br_ioctl_mutex rtnl_mutex uevent_sock_mutex &rq->__lock irq_context: 0 br_ioctl_mutex rtnl_mutex uevent_sock_mutex &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 pernet_ops_rwsem devices_rwsem rdma_nets_rwsem &device->compat_devs_mutex remove_cache_srcu irq_context: 0 pernet_ops_rwsem devices_rwsem rdma_nets_rwsem &device->compat_devs_mutex remove_cache_srcu quarantine_lock irq_context: 0 pernet_ops_rwsem devices_rwsem rdma_nets_rwsem &device->compat_devs_mutex remove_cache_srcu &c->lock irq_context: 0 pernet_ops_rwsem devices_rwsem rdma_nets_rwsem &device->compat_devs_mutex remove_cache_srcu &n->list_lock irq_context: 0 pernet_ops_rwsem devices_rwsem rdma_nets_rwsem &device->compat_devs_mutex remove_cache_srcu &obj_hash[i].lock irq_context: 0 pernet_ops_rwsem devices_rwsem rdma_nets_rwsem &device->compat_devs_mutex remove_cache_srcu &rq->__lock irq_context: 0 pernet_ops_rwsem devices_rwsem rdma_nets_rwsem &device->compat_devs_mutex remove_cache_srcu &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 br_ioctl_mutex rtnl_mutex &br->hash_lock &____s->seqcount#2 irq_context: 0 br_ioctl_mutex rtnl_mutex &br->hash_lock &____s->seqcount irq_context: 0 sk_lock-AF_ALG rcu_read_lock rcu_read_lock rcu_node_0 irq_context: 0 sk_lock-AF_ALG rcu_read_lock rcu_read_lock &rq->__lock irq_context: 0 sk_lock-AF_ALG rcu_read_lock rcu_read_lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 pernet_ops_rwsem devices_rwsem rdma_nets_rwsem &device->compat_devs_mutex &p->pi_lock &cfs_rq->removed.lock irq_context: 0 sk_lock-AF_ALG &mm->mmap_lock sb_pagefaults &____s->seqcount#2 irq_context: 0 sk_lock-AF_ALG &mm->mmap_lock sb_pagefaults &____s->seqcount irq_context: 0 fill_pool_map-wait-type-override batched_entropy_u8.lock irq_context: 0 fill_pool_map-wait-type-override kfence_freelist_lock irq_context: 0 cb_lock genl_mutex __ip_vs_mutex irq_context: 0 cb_lock genl_mutex __ip_vs_mutex &s->lock irq_context: 0 sb_writers#4 &iint->mutex rcu_read_lock rcu_node_0 irq_context: 0 sb_writers#4 &iint->mutex rcu_read_lock &rq->__lock irq_context: 0 sb_writers#4 &iint->mutex rcu_read_lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: softirq (&app->join_timer) rcu_read_lock_bh rcu_read_lock &c->lock irq_context: 0 sk_lock-AF_ALG fs_reclaim mmu_notifier_invalidate_range_start &rq->__lock irq_context: 0 sk_lock-AF_ALG fs_reclaim mmu_notifier_invalidate_range_start &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 pernet_ops_rwsem devices_rwsem rdma_nets_rwsem &device->compat_devs_mutex &cfs_rq->removed.lock irq_context: 0 pernet_ops_rwsem devices_rwsem rdma_nets_rwsem &device->compat_devs_mutex &rq->__lock &cfs_rq->removed.lock irq_context: 0 &sb->s_type->i_mutex_key#8 &sbi->s_writepages_rwsem jbd2_handle &ei->i_data_sem &ei->i_es_lock &sbi->s_es_lock irq_context: 0 &sb->s_type->i_mutex_key#8 &sbi->s_writepages_rwsem jbd2_handle &journal->j_list_lock irq_context: 0 pernet_ops_rwsem devices_rwsem rdma_nets_rwsem &lock->wait_lock irq_context: 0 pernet_ops_rwsem devices_rwsem rdma_nets_rwsem &p->pi_lock irq_context: 0 pernet_ops_rwsem devices_rwsem rdma_nets_rwsem &p->pi_lock &rq->__lock irq_context: 0 pernet_ops_rwsem devices_rwsem rdma_nets_rwsem &p->pi_lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)bat_events (work_completion)(&(&bat_priv->orig_work)->work) &rq->__lock &cfs_rq->removed.lock irq_context: 0 sk_lock-AF_ALG &pcp->lock &zone->lock irq_context: 0 &sb->s_type->i_mutex_key#8 &sbi->s_writepages_rwsem jbd2_handle &ei->i_data_sem stock_lock irq_context: 0 &sb->s_type->i_mutex_key#8 &sbi->s_writepages_rwsem jbd2_handle &ei->i_data_sem &xa->xa_lock#9 irq_context: 0 &sb->s_type->i_mutex_key#8 &sbi->s_writepages_rwsem jbd2_handle &ei->i_data_sem &xa->xa_lock#9 stock_lock irq_context: 0 &sb->s_type->i_mutex_key#8 &sbi->s_writepages_rwsem jbd2_handle &ei->i_data_sem &xa->xa_lock#9 pool_lock#2 irq_context: 0 &sb->s_type->i_mutex_key#8 &sbi->s_writepages_rwsem jbd2_handle &ei->i_data_sem &xa->xa_lock#9 &c->lock irq_context: 0 &sb->s_type->i_mutex_key#8 &sbi->s_writepages_rwsem jbd2_handle &ei->i_data_sem lock#4 irq_context: 0 &sb->s_type->i_mutex_key#8 &sbi->s_writepages_rwsem jbd2_handle &ei->i_data_sem &journal->j_revoke_lock irq_context: 0 sb_writers#11 &type->i_mutex_dir_key#7/1 cgroup_mutex blkcg_pol_mutex &c->lock irq_context: 0 (wq_completion)ipv6_addrconf (work_completion)(&(&ifa->dad_work)->work) rtnl_mutex &idev->mc_lock rcu_read_lock rcu_read_lock rcu_read_lock_bh rcu_read_lock &obj_hash[i].lock pool_lock irq_context: 0 sk_lock-AF_ALG &mm->mmap_lock sb_pagefaults mapping.invalidate_lock &folio_wait_table[i] irq_context: 0 (wq_completion)ext4-rsv-conversion (work_completion)(&ei->i_rsv_conversion_work) jbd2_handle &ei->i_data_sem &cfs_rq->removed.lock irq_context: 0 &sb->s_type->i_mutex_key#8 fs_reclaim &rq->__lock irq_context: 0 &sb->s_type->i_mutex_key#8 fs_reclaim &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 &sb->s_type->i_mutex_key#8 &sbi->s_writepages_rwsem quarantine_lock irq_context: 0 rtnl_mutex &tb->tb6_lock stock_lock irq_context: 0 &sb->s_type->i_mutex_key#8 &sbi->s_writepages_rwsem remove_cache_srcu pool_lock#2 irq_context: softirq rcu_read_lock &xa->xa_lock#9 &pl->lock irq_context: softirq rcu_read_lock &xa->xa_lock#9 &pl->lock key#12 irq_context: 0 pernet_ops_rwsem devices_rwsem rdma_nets_rwsem &device->compat_devs_mutex &pcp->lock &zone->lock irq_context: 0 rtnl_mutex &tb->tb6_lock rcu_read_lock &____s->seqcount#2 irq_context: 0 rtnl_mutex &tb->tb6_lock rcu_read_lock &n->list_lock irq_context: 0 rtnl_mutex &tb->tb6_lock rcu_read_lock &n->list_lock &c->lock irq_context: 0 &f->f_pos_lock sb_writers#4 &sb->s_type->i_mutex_key#8 &mm->mmap_lock &pcp->lock &zone->lock irq_context: 0 pernet_ops_rwsem devices_rwsem rdma_nets_rwsem &device->compat_devs_mutex remove_cache_srcu rcu_read_lock pool_lock#2 irq_context: 0 &sbi->s_writepages_rwsem jbd2_handle &____s->seqcount#2 irq_context: 0 sk_lock-AF_ALG rcu_read_lock &rcu_state.expedited_wq irq_context: 0 pernet_ops_rwsem cpu_hotplug_lock wq_pool_mutex &obj_hash[i].lock pool_lock irq_context: 0 pernet_ops_rwsem nf_ct_proto_mutex defrag6_mutex nf_hook_mutex &c->lock irq_context: 0 pernet_ops_rwsem nf_ct_proto_mutex defrag6_mutex nf_hook_mutex &n->list_lock irq_context: 0 pernet_ops_rwsem nf_ct_proto_mutex defrag6_mutex nf_hook_mutex &n->list_lock &c->lock irq_context: 0 pernet_ops_rwsem k-sk_lock-AF_RXRPC rcu_read_lock rcu_node_0 irq_context: 0 pernet_ops_rwsem k-sk_lock-AF_RXRPC rcu_read_lock &rq->__lock irq_context: 0 (wq_completion)afs (work_completion)(&net->fs_manager) &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)events (work_completion)(&(&nsim_dev->trap_data->trap_report_dw)->work) &devlink->lock_key#4 &rcu_state.expedited_wq irq_context: 0 (wq_completion)events (work_completion)(&(&nsim_dev->trap_data->trap_report_dw)->work) &devlink->lock_key#4 &rcu_state.expedited_wq &p->pi_lock irq_context: 0 pernet_ops_rwsem rcu_state.exp_mutex &rq->__lock &cfs_rq->removed.lock irq_context: 0 &sb->s_type->i_mutex_key#8 &sbi->s_writepages_rwsem jbd2_handle &ei->i_data_sem &cfs_rq->removed.lock irq_context: 0 pernet_ops_rwsem ovs_mutex nf_ct_proto_mutex nf_hook_mutex &rq->__lock irq_context: 0 (wq_completion)events (work_completion)(&(&nsim_dev->trap_data->trap_report_dw)->work) &devlink->lock_key#3 &rcu_state.expedited_wq &p->pi_lock &rq->__lock irq_context: 0 pernet_ops_rwsem rcu_read_lock &pool->lock/1 &p->pi_lock &cfs_rq->removed.lock irq_context: 0 pernet_ops_rwsem &rnp->exp_wq[0] irq_context: 0 pernet_ops_rwsem &device->compat_devs_mutex irq_context: 0 (wq_completion)events (work_completion)(&(&nsim_dev->trap_data->trap_report_dw)->work) &devlink->lock_key#3 &rcu_state.expedited_wq &p->pi_lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 sk_lock-AF_ALG remove_cache_srcu pool_lock#2 irq_context: 0 &sbi->s_writepages_rwsem &mapping->i_mmap_rwsem mmu_notifier_invalidate_range_start irq_context: 0 &sbi->s_writepages_rwsem &mapping->i_mmap_rwsem mmu_notifier_invalidate_range_start &rq->__lock irq_context: 0 &sbi->s_writepages_rwsem &mapping->i_mmap_rwsem mmu_notifier_invalidate_range_start &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 &sbi->s_writepages_rwsem &mapping->i_mmap_rwsem ptlock_ptr(page)#2 irq_context: 0 &sbi->s_writepages_rwsem &mapping->private_lock irq_context: 0 &sbi->s_writepages_rwsem &mapping->private_lock rcu_read_lock &memcg->move_lock irq_context: 0 &sbi->s_writepages_rwsem rcu_read_lock key#10 irq_context: 0 &sbi->s_writepages_rwsem rcu_read_lock &xa->xa_lock#9 key#10 irq_context: 0 &sbi->s_writepages_rwsem jbd2_handle &journal->j_list_lock irq_context: 0 pernet_ops_rwsem &device->compat_devs_mutex &xa->xa_lock#17 irq_context: 0 pernet_ops_rwsem &root->kernfs_rwsem irq_context: 0 pernet_ops_rwsem &root->kernfs_rwsem &____s->seqcount irq_context: 0 pernet_ops_rwsem &root->kernfs_rwsem rcu_read_lock pool_lock#2 irq_context: 0 pernet_ops_rwsem &root->kernfs_rwsem kernfs_idr_lock &obj_hash[i].lock irq_context: 0 pernet_ops_rwsem &root->kernfs_rwsem kernfs_idr_lock pool_lock#2 irq_context: 0 pernet_ops_rwsem &root->kernfs_rwsem rcu_read_lock &rq->__lock irq_context: 0 pernet_ops_rwsem &root->kernfs_rwsem &meta->lock irq_context: 0 pernet_ops_rwsem &root->kernfs_rwsem kfence_freelist_lock irq_context: 0 pernet_ops_rwsem &zone->lock irq_context: 0 pernet_ops_rwsem &zone->lock &____s->seqcount irq_context: 0 pernet_ops_rwsem &root->kernfs_rwsem rcu_read_lock rcu_node_0 irq_context: 0 pernet_ops_rwsem &root->kernfs_rwsem rcu_read_lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 pernet_ops_rwsem dev_pm_qos_sysfs_mtx irq_context: 0 pernet_ops_rwsem dev_pm_qos_sysfs_mtx &root->kernfs_rwsem irq_context: 0 pernet_ops_rwsem dev_pm_qos_sysfs_mtx &root->kernfs_rwsem irq_context: 0 pernet_ops_rwsem dev_pm_qos_sysfs_mtx dev_pm_qos_mtx irq_context: 0 pernet_ops_rwsem subsys mutex#84 irq_context: 0 pernet_ops_rwsem subsys mutex#84 &k->k_lock irq_context: 0 pernet_ops_rwsem subsys mutex#84 &k->k_lock klist_remove_lock irq_context: 0 pernet_ops_rwsem &x->wait#9 irq_context: 0 pernet_ops_rwsem dpm_list_mtx irq_context: 0 pernet_ops_rwsem &dev->power.lock irq_context: 0 pernet_ops_rwsem deferred_probe_mutex irq_context: 0 pernet_ops_rwsem device_links_lock irq_context: 0 pernet_ops_rwsem uevent_sock_mutex mmu_notifier_invalidate_range_start irq_context: 0 pernet_ops_rwsem gdp_mutex irq_context: 0 pernet_ops_rwsem &device->unregistration_lock irq_context: 0 pernet_ops_rwsem &root->kernfs_rwsem quarantine_lock irq_context: 0 (wq_completion)wg-crypt-wg2#5 (work_completion)(&peer->transmit_packet_work) &peer->endpoint_lock rcu_read_lock_bh rcu_read_lock rcu_read_lock &____s->seqcount#7 irq_context: 0 (wq_completion)wg-crypt-wg2#5 (work_completion)(&peer->transmit_packet_work) &peer->endpoint_lock rcu_read_lock_bh rcu_read_lock &nf_nat_locks[i] irq_context: 0 (wq_completion)wg-crypt-wg2#5 (work_completion)(&peer->transmit_packet_work) &peer->endpoint_lock rcu_read_lock_bh rcu_read_lock &nf_conntrack_locks[i] irq_context: 0 (wq_completion)wg-crypt-wg2#5 (work_completion)(&peer->transmit_packet_work) &peer->endpoint_lock rcu_read_lock_bh rcu_read_lock &nf_conntrack_locks[i] &nf_conntrack_locks[i]/1 irq_context: 0 (wq_completion)wg-crypt-wg2#5 (work_completion)(&peer->transmit_packet_work) &peer->endpoint_lock rcu_read_lock_bh rcu_read_lock &nf_conntrack_locks[i] &nf_conntrack_locks[i]/1 batched_entropy_u8.lock irq_context: 0 (wq_completion)wg-crypt-wg2#5 (work_completion)(&peer->transmit_packet_work) &peer->endpoint_lock rcu_read_lock_bh rcu_read_lock &nf_conntrack_locks[i]/1 irq_context: 0 (wq_completion)wg-crypt-wg2#5 (work_completion)(&peer->transmit_packet_work) &peer->endpoint_lock rcu_read_lock_bh rcu_read_lock &____s->seqcount#9 irq_context: 0 (wq_completion)wg-crypt-wg2#5 (work_completion)(&peer->transmit_packet_work) &peer->endpoint_lock rcu_read_lock_bh rcu_read_lock rcu_read_lock_bh &obj_hash[i].lock irq_context: 0 (wq_completion)events_power_efficient (work_completion)(&(&tbl->gc_work)->work) &tbl->lock quarantine_lock irq_context: 0 (wq_completion)wg-crypt-wg0#6 (work_completion)(&peer->transmit_packet_work) &peer->endpoint_lock rcu_read_lock_bh rcu_read_lock &c->lock irq_context: 0 &pipe->rd_wait &p->pi_lock &rq->__lock &obj_hash[i].lock irq_context: 0 &pipe->rd_wait &p->pi_lock &rq->__lock &base->lock irq_context: 0 &pipe->rd_wait &p->pi_lock &rq->__lock &base->lock &obj_hash[i].lock irq_context: 0 sb_writers#11 &type->i_mutex_dir_key#7/1 &root->kernfs_rwsem &c->lock irq_context: 0 &pipe->mutex/1 sk_lock-AF_INET6 irq_context: 0 &pipe->mutex/1 sk_lock-AF_INET6 slock-AF_INET6 irq_context: 0 &pipe->mutex/1 sk_lock-AF_INET6 fs_reclaim irq_context: 0 &pipe->mutex/1 sk_lock-AF_INET6 fs_reclaim mmu_notifier_invalidate_range_start irq_context: 0 &pipe->mutex/1 sk_lock-AF_INET6 pool_lock#2 irq_context: 0 &pipe->mutex/1 sk_lock-AF_INET6 rcu_read_lock rcu_read_lock &____s->seqcount#2 irq_context: 0 &pipe->mutex/1 sk_lock-AF_INET6 rcu_read_lock rcu_read_lock &____s->seqcount irq_context: 0 &pipe->mutex/1 sk_lock-AF_INET6 rcu_read_lock rcu_read_lock rcu_node_0 irq_context: 0 &pipe->mutex/1 sk_lock-AF_INET6 rcu_read_lock rcu_read_lock &rq->__lock irq_context: 0 &pipe->mutex/1 sk_lock-AF_INET6 rcu_read_lock rcu_read_lock pool_lock#2 irq_context: 0 &pipe->mutex/1 sk_lock-AF_INET6 rcu_read_lock rcu_read_lock &c->lock irq_context: 0 &pipe->mutex/1 sk_lock-AF_INET6 rcu_read_lock rcu_read_lock &obj_hash[i].lock irq_context: 0 &pipe->mutex/1 sk_lock-AF_INET6 rcu_read_lock rcu_read_lock &base->lock irq_context: 0 &pipe->mutex/1 sk_lock-AF_INET6 rcu_read_lock rcu_read_lock &base->lock &obj_hash[i].lock irq_context: 0 &pipe->mutex/1 sk_lock-AF_INET6 rcu_read_lock rcu_read_lock rcu_read_lock rhashtable_bucket irq_context: 0 &pipe->mutex/1 sk_lock-AF_INET6 rcu_read_lock &q->lock#2 irq_context: 0 &pipe->mutex/1 sk_lock-AF_INET6 rcu_read_lock &q->lock#2 &obj_hash[i].lock irq_context: 0 &pipe->mutex/1 sk_lock-AF_INET6 rcu_read_lock &q->lock#2 &base->lock irq_context: 0 &pipe->mutex/1 sk_lock-AF_INET6 rcu_read_lock &q->lock#2 &base->lock &obj_hash[i].lock irq_context: 0 &pipe->mutex/1 sk_lock-AF_INET6 rcu_read_lock &q->lock#2 rcu_read_lock rcu_read_lock rhashtable_bucket irq_context: 0 &pipe->mutex/1 sk_lock-AF_INET6 rcu_read_lock &q->lock#2 rcu_read_lock rcu_read_lock rcu_read_lock &pool->lock irq_context: 0 &pipe->mutex/1 sk_lock-AF_INET6 rcu_read_lock &q->lock#2 rcu_read_lock rcu_read_lock rcu_read_lock &pool->lock &obj_hash[i].lock irq_context: 0 &pipe->mutex/1 sk_lock-AF_INET6 rcu_read_lock &q->lock#2 rcu_read_lock rcu_read_lock rcu_read_lock &pool->lock pool_lock#2 irq_context: 0 &pipe->mutex/1 sk_lock-AF_INET6 rcu_read_lock &q->lock#2 rcu_read_lock rcu_read_lock rcu_read_lock &pool->lock &p->pi_lock irq_context: 0 &pipe->mutex/1 sk_lock-AF_INET6 rcu_read_lock &rq->__lock irq_context: 0 &pipe->mutex/1 sk_lock-AF_INET6 rcu_read_lock &obj_hash[i].lock irq_context: 0 &pipe->mutex/1 sk_lock-AF_INET6 rcu_read_lock pool_lock#2 irq_context: 0 &pipe->mutex/1 sk_lock-AF_INET6 rcu_read_lock &____s->seqcount#7 irq_context: 0 &pipe->mutex/1 sk_lock-AF_INET6 rcu_read_lock rcu_read_lock &____s->seqcount#7 irq_context: 0 &pipe->mutex/1 sk_lock-AF_INET6 rcu_read_lock rcu_node_0 irq_context: 0 &pipe->mutex/1 sk_lock-AF_INET6 rcu_read_lock &nf_nat_locks[i] irq_context: 0 &pipe->mutex/1 sk_lock-AF_INET6 rcu_read_lock &nf_conntrack_locks[i] irq_context: 0 &pipe->mutex/1 sk_lock-AF_INET6 rcu_read_lock &nf_conntrack_locks[i] batched_entropy_u8.lock irq_context: 0 &pipe->mutex/1 sk_lock-AF_INET6 rcu_read_lock &____s->seqcount#10 irq_context: 0 &pipe->mutex/1 sk_lock-AF_INET6 rcu_read_lock rcu_read_lock_bh tk_core.seq.seqcount irq_context: 0 &pipe->mutex/1 sk_lock-AF_INET6 rcu_read_lock rcu_read_lock_bh &list->lock#5 irq_context: 0 &pipe->mutex/1 slock-AF_INET6 irq_context: 0 rtnl_mutex &br->lock &____s->seqcount irq_context: 0 rtnl_mutex &br->lock rcu_read_lock pool_lock#2 irq_context: 0 cgroup_threadgroup_rwsem freezer_mutex freezer_mutex.wait_lock irq_context: softirq (&mp->timer) &br->multicast_lock rcu_read_lock &pool->lock irq_context: softirq (&mp->timer) &br->multicast_lock rcu_read_lock &pool->lock &obj_hash[i].lock irq_context: softirq (&mp->timer) &br->multicast_lock rcu_read_lock &pool->lock pool_lock#2 irq_context: softirq (&mp->timer) &br->multicast_lock rcu_read_lock &pool->lock &p->pi_lock irq_context: softirq (&mp->timer) &br->multicast_lock rcu_read_lock &pool->lock &p->pi_lock &rq->__lock irq_context: softirq (&mp->timer) &br->multicast_lock rcu_read_lock &pool->lock &p->pi_lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 cgroup_threadgroup_rwsem freezer_mutex.wait_lock irq_context: 0 rtnl_mutex &root->kernfs_rwsem rcu_node_0 irq_context: 0 sb_writers#4 &sb->s_type->i_mutex_key#8 mapping.invalidate_lock jbd2_handle &ei->i_data_sem &cfs_rq->removed.lock irq_context: 0 (wq_completion)events (linkwatch_work).work rtnl_mutex &tbl->lock &c->lock irq_context: 0 &mm->mmap_lock &mapping->i_mmap_rwsem &rcu_state.expedited_wq &p->pi_lock irq_context: 0 pernet_ops_rwsem rtnl_mutex nl_table_wait.lock &p->pi_lock irq_context: 0 pernet_ops_rwsem rtnl_mutex nl_table_wait.lock &p->pi_lock &rq->__lock irq_context: 0 pernet_ops_rwsem rtnl_mutex nl_table_wait.lock &p->pi_lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 &sb->s_type->i_mutex_key#10 nl_table_wait.lock &p->pi_lock irq_context: 0 &sb->s_type->i_mutex_key#10 nl_table_wait.lock &p->pi_lock &rq->__lock irq_context: 0 &sb->s_type->i_mutex_key#10 nl_table_wait.lock &p->pi_lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 pernet_ops_rwsem rtnl_mutex &rq->__lock &cfs_rq->removed.lock irq_context: 0 pernet_ops_rwsem devices_rwsem rdma_nets_rwsem &device->compat_devs_mutex uevent_sock_mutex &c->lock irq_context: 0 pernet_ops_rwsem devices_rwsem rdma_nets_rwsem &device->compat_devs_mutex lock kernfs_idr_lock &n->list_lock irq_context: 0 pernet_ops_rwsem devices_rwsem rdma_nets_rwsem &device->compat_devs_mutex lock kernfs_idr_lock &n->list_lock &c->lock irq_context: 0 lock#4 &lruvec->lru_lock per_cpu_ptr(&cgroup_rstat_cpu_lock, cpu) irq_context: 0 sb_writers#4 &type->i_mutex_dir_key#3 per_cpu_ptr(&cgroup_rstat_cpu_lock, cpu) irq_context: 0 sb_writers#4 &sb->s_type->i_mutex_key#8 &sbi->s_writepages_rwsem &c->lock irq_context: 0 &sb->s_type->i_mutex_key#10 fill_pool_map-wait-type-override &n->list_lock irq_context: 0 &sb->s_type->i_mutex_key#10 fill_pool_map-wait-type-override &n->list_lock &c->lock irq_context: 0 pernet_ops_rwsem devices_rwsem rdma_nets_rwsem &device->compat_devs_mutex &root->kernfs_rwsem &root->kernfs_iattr_rwsem &sem->wait_lock irq_context: 0 pernet_ops_rwsem devices_rwsem rdma_nets_rwsem &device->compat_devs_mutex remove_cache_srcu pool_lock#2 irq_context: 0 (wq_completion)rcu_gp (work_completion)(&rew->rew_work) rcu_state.exp_wake_mutex &rnp->exp_wq[0] &p->pi_lock &rq->__lock &cfs_rq->removed.lock irq_context: 0 pernet_ops_rwsem nf_ct_proto_mutex defrag6_mutex nf_hook_mutex &rq->__lock irq_context: 0 pernet_ops_rwsem ovs_mutex nf_ct_proto_mutex rcu_read_lock rcu_node_0 irq_context: 0 pernet_ops_rwsem ovs_mutex nf_ct_proto_mutex rcu_read_lock &rq->__lock irq_context: 0 pernet_ops_rwsem k-sk_lock-AF_TIPC &tn->nametbl_lock &service->lock &obj_hash[i].lock pool_lock irq_context: 0 pernet_ops_rwsem &rnp->exp_wq[3] irq_context: 0 cb_lock rcu_node_0 irq_context: 0 cb_lock &rcu_state.expedited_wq irq_context: 0 cb_lock &rcu_state.expedited_wq &p->pi_lock irq_context: 0 cb_lock &rcu_state.expedited_wq &p->pi_lock &rq->__lock irq_context: 0 cb_lock &rcu_state.expedited_wq &p->pi_lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 pernet_ops_rwsem ovs_mutex nf_ct_proto_mutex &obj_hash[i].lock pool_lock irq_context: 0 pernet_ops_rwsem rtnl_mutex rcu_state.exp_mutex rcu_read_lock &pool->lock &p->pi_lock &rq->__lock &cfs_rq->removed.lock irq_context: 0 pernet_ops_rwsem rtnl_mutex rcu_state.exp_mutex &rq->__lock &cfs_rq->removed.lock irq_context: 0 pernet_ops_rwsem &root->kernfs_rwsem rcu_node_0 irq_context: 0 pernet_ops_rwsem rtnl_mutex &rnp->exp_wq[3] irq_context: 0 pernet_ops_rwsem rcu_state.barrier_mutex rcu_node_0 irq_context: 0 pernet_ops_rwsem rcu_state.barrier_mutex &rcu_state.expedited_wq irq_context: 0 pernet_ops_rwsem rcu_state.barrier_mutex &rcu_state.expedited_wq &p->pi_lock irq_context: 0 pernet_ops_rwsem rcu_state.barrier_mutex &rcu_state.expedited_wq &p->pi_lock &rq->__lock irq_context: 0 pernet_ops_rwsem rcu_state.barrier_mutex &rcu_state.expedited_wq &p->pi_lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 pernet_ops_rwsem &root->kernfs_rwsem &rcu_state.expedited_wq irq_context: 0 pernet_ops_rwsem &root->kernfs_rwsem &rcu_state.expedited_wq &p->pi_lock irq_context: 0 pernet_ops_rwsem rtnl_mutex &p->pi_lock &rq->__lock &cfs_rq->removed.lock irq_context: 0 (wq_completion)wg-crypt-wg1 (work_completion)(&peer->transmit_packet_work) &peer->endpoint_lock rcu_read_lock_bh rcu_read_lock &c->lock irq_context: 0 (wq_completion)wg-crypt-wg1#3 (work_completion)(&peer->transmit_packet_work) &peer->endpoint_lock rcu_read_lock_bh rcu_read_lock &obj_hash[i].lock irq_context: 0 pernet_ops_rwsem &root->kernfs_rwsem kernfs_idr_lock &obj_hash[i].lock pool_lock irq_context: 0 pernet_ops_rwsem rtnl_mutex dev_pm_qos_sysfs_mtx &root->kernfs_rwsem &rq->__lock irq_context: 0 pernet_ops_rwsem rtnl_mutex dev_pm_qos_sysfs_mtx &root->kernfs_rwsem &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 pernet_ops_rwsem &root->kernfs_rwsem &rcu_state.expedited_wq &p->pi_lock &rq->__lock irq_context: 0 (wq_completion)events (work_completion)(&(&krcp->monitor_work)->work) rcu_callback &meta->lock irq_context: 0 (wq_completion)events (work_completion)(&(&krcp->monitor_work)->work) rcu_callback kfence_freelist_lock irq_context: softirq (&app->join_timer) &app->lock &____s->seqcount#2 irq_context: 0 (wq_completion)events (work_completion)(&(&nsim_dev->trap_data->trap_report_dw)->work) &devlink->lock_key#4 &rcu_state.expedited_wq &p->pi_lock &rq->__lock irq_context: 0 (wq_completion)events (work_completion)(&(&nsim_dev->trap_data->trap_report_dw)->work) &devlink->lock_key#4 &rcu_state.expedited_wq &p->pi_lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 &tsk->futex_exit_mutex &mm->mmap_lock ptlock_ptr(page)#2 irq_context: 0 &tsk->futex_exit_mutex &mm->mmap_lock fs_reclaim irq_context: 0 &tsk->futex_exit_mutex &mm->mmap_lock fs_reclaim mmu_notifier_invalidate_range_start irq_context: 0 &tsk->futex_exit_mutex &mm->mmap_lock fs_reclaim mmu_notifier_invalidate_range_start &rq->__lock irq_context: 0 (wq_completion)events fqdir_free_work &rq->__lock irq_context: 0 (wq_completion)events fqdir_free_work &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 &tsk->futex_exit_mutex &mm->mmap_lock remove_cache_srcu irq_context: 0 &tsk->futex_exit_mutex &mm->mmap_lock remove_cache_srcu quarantine_lock irq_context: 0 &tsk->futex_exit_mutex &mm->mmap_lock remove_cache_srcu &c->lock irq_context: 0 &tsk->futex_exit_mutex &mm->mmap_lock remove_cache_srcu &n->list_lock irq_context: 0 &tsk->futex_exit_mutex &mm->mmap_lock remove_cache_srcu rcu_read_lock pool_lock#2 irq_context: 0 &tsk->futex_exit_mutex &mm->mmap_lock remove_cache_srcu &obj_hash[i].lock irq_context: 0 pernet_ops_rwsem uevent_sock_mutex rcu_read_lock &rcu_state.gp_wq irq_context: 0 pernet_ops_rwsem uevent_sock_mutex rcu_read_lock &rcu_state.gp_wq &p->pi_lock irq_context: 0 pernet_ops_rwsem uevent_sock_mutex.wait_lock irq_context: 0 rtnl_mutex &br->lock &n->list_lock irq_context: 0 rtnl_mutex &br->lock &n->list_lock &c->lock irq_context: 0 &mm->mmap_lock remove_cache_srcu &____s->seqcount irq_context: 0 &pipe->mutex/1 sk_lock-AF_INET6 &c->lock irq_context: 0 &pipe->mutex/1 sk_lock-AF_INET6 rcu_read_lock &q->lock#2 rcu_read_lock rcu_read_lock rcu_read_lock &pool->lock &p->pi_lock &rq->__lock irq_context: 0 &pipe->mutex/1 sk_lock-AF_INET6 rcu_read_lock &q->lock#2 rcu_read_lock rcu_read_lock rcu_read_lock &pool->lock &p->pi_lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 &pipe->mutex/1 sk_lock-AF_INET6 rcu_read_lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 pernet_ops_rwsem devices_rwsem rdma_nets_rwsem &device->compat_devs_mutex &root->kernfs_rwsem &sem->wait_lock irq_context: 0 sb_writers#4 fill_pool_map-wait-type-override &rq->__lock irq_context: 0 sb_writers#4 fill_pool_map-wait-type-override &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 elock-AF_INET6 irq_context: 0 pernet_ops_rwsem __ip_vs_app_mutex &____s->seqcount#2 irq_context: 0 &list->lock#34 irq_context: 0 pernet_ops_rwsem devices_rwsem rdma_nets_rwsem &device->compat_devs_mutex &pcp->lock &zone->lock &____s->seqcount irq_context: 0 (wq_completion)events (work_completion)(&rdev->wiphy_work) &rdev->wiphy.mtx &wdev->mtx fill_pool_map-wait-type-override &rq->__lock irq_context: 0 (wq_completion)events (work_completion)(&rdev->wiphy_work) &rdev->wiphy.mtx &wdev->mtx fill_pool_map-wait-type-override &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)bat_events (work_completion)(&(&bat_priv->mcast.work)->work) fill_pool_map-wait-type-override pool_lock#2 irq_context: 0 (wq_completion)bat_events (work_completion)(&(&bat_priv->mcast.work)->work) fill_pool_map-wait-type-override pool_lock irq_context: 0 (wq_completion)events_unbound (work_completion)(&sub_info->work) tasklist_lock &sighand->siglock &(&sig->stats_lock)->lock &____s->seqcount#5 fill_pool_map-wait-type-override pool_lock#2 irq_context: 0 (wq_completion)events_unbound (work_completion)(&sub_info->work) tasklist_lock &sighand->siglock &(&sig->stats_lock)->lock &____s->seqcount#5 fill_pool_map-wait-type-override &c->lock irq_context: 0 (wq_completion)events_unbound (work_completion)(&sub_info->work) tasklist_lock &sighand->siglock &(&sig->stats_lock)->lock &____s->seqcount#5 fill_pool_map-wait-type-override &n->list_lock irq_context: 0 (wq_completion)events_unbound (work_completion)(&sub_info->work) tasklist_lock &sighand->siglock &(&sig->stats_lock)->lock &____s->seqcount#5 fill_pool_map-wait-type-override &n->list_lock &c->lock irq_context: 0 (wq_completion)events_unbound (work_completion)(&sub_info->work) tasklist_lock &sighand->siglock &(&sig->stats_lock)->lock &____s->seqcount#5 fill_pool_map-wait-type-override pool_lock irq_context: 0 &f->f_pos_lock &type->i_mutex_dir_key#4 &root->kernfs_rwsem rcu_read_lock &rq->__lock irq_context: 0 &f->f_pos_lock &type->i_mutex_dir_key#4 &root->kernfs_rwsem rcu_read_lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 pernet_ops_rwsem devices_rwsem rdma_nets_rwsem &device->compat_devs_mutex uevent_sock_mutex &____s->seqcount#2 irq_context: 0 pernet_ops_rwsem devices_rwsem rdma_nets_rwsem &device->compat_devs_mutex uevent_sock_mutex &____s->seqcount irq_context: 0 pernet_ops_rwsem devices_rwsem rdma_nets_rwsem &device->compat_devs_mutex uevent_sock_mutex &rq->__lock irq_context: 0 pernet_ops_rwsem devices_rwsem rdma_nets_rwsem &device->compat_devs_mutex uevent_sock_mutex &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 pernet_ops_rwsem devices_rwsem rdma_nets_rwsem &device->compat_devs_mutex uevent_sock_mutex.wait_lock irq_context: 0 pernet_ops_rwsem devices_rwsem rdma_nets_rwsem &device->compat_devs_mutex lock kernfs_idr_lock &pcp->lock &zone->lock irq_context: 0 (wq_completion)bat_events (work_completion)(&(&forw_packet_aggr->delayed_work)->work) &hard_iface->bat_iv.ogm_buff_mutex &rcu_state.expedited_wq &p->pi_lock &rq->__lock irq_context: 0 (wq_completion)bat_events (work_completion)(&(&forw_packet_aggr->delayed_work)->work) &hard_iface->bat_iv.ogm_buff_mutex &rcu_state.expedited_wq &p->pi_lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 pernet_ops_rwsem rtnl_mutex rcu_read_lock &rcu_state.expedited_wq irq_context: 0 &sbi->s_writepages_rwsem rcu_read_lock fill_pool_map-wait-type-override pool_lock#2 irq_context: 0 &sbi->s_writepages_rwsem rcu_read_lock fill_pool_map-wait-type-override pool_lock irq_context: 0 &sbi->s_writepages_rwsem rcu_read_lock fill_pool_map-wait-type-override rcu_node_0 irq_context: 0 pernet_ops_rwsem rtnl_mutex rcu_read_lock &rcu_state.expedited_wq &p->pi_lock irq_context: 0 pernet_ops_rwsem rtnl_mutex rcu_read_lock &rcu_state.expedited_wq &p->pi_lock &rq->__lock irq_context: 0 pernet_ops_rwsem rtnl_mutex rcu_read_lock &rcu_state.expedited_wq &p->pi_lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 pernet_ops_rwsem ovs_mutex nf_ct_proto_mutex defrag6_mutex &obj_hash[i].lock pool_lock irq_context: 0 &sbi->s_writepages_rwsem rcu_read_lock fill_pool_map-wait-type-override &rq->__lock irq_context: 0 &sbi->s_writepages_rwsem rcu_read_lock fill_pool_map-wait-type-override &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 pernet_ops_rwsem ovs_mutex nf_ct_proto_mutex defrag6_mutex &rq->__lock irq_context: 0 pernet_ops_rwsem ovs_mutex nf_ct_proto_mutex defrag6_mutex &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem rtnl_mutex dev_pm_qos_sysfs_mtx &root->kernfs_rwsem &sem->wait_lock irq_context: 0 &f->f_pos_lock &type->i_mutex_dir_key#3 sb_writers#4 &pcp->lock &zone->lock irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem rtnl_mutex &pnn->routes.lock &rq->__lock irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem rtnl_mutex &pnn->routes.lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 pernet_ops_rwsem devices_rwsem rdma_nets_rwsem &device->compat_devs_mutex &rcu_state.expedited_wq &p->pi_lock irq_context: 0 pernet_ops_rwsem devices_rwsem rdma_nets_rwsem &device->compat_devs_mutex &rcu_state.expedited_wq &p->pi_lock &rq->__lock irq_context: 0 pernet_ops_rwsem devices_rwsem rdma_nets_rwsem &device->compat_devs_mutex &rcu_state.expedited_wq &p->pi_lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 pernet_ops_rwsem devices_rwsem rdma_nets_rwsem &device->compat_devs_mutex &root->kernfs_rwsem rcu_node_0 irq_context: 0 pernet_ops_rwsem &rnp->exp_wq[1] irq_context: 0 pernet_ops_rwsem devices_rwsem rdma_nets_rwsem &device->compat_devs_mutex rcu_read_lock &rcu_state.expedited_wq irq_context: 0 pernet_ops_rwsem devices_rwsem rdma_nets_rwsem &device->compat_devs_mutex rcu_read_lock &rcu_state.expedited_wq &p->pi_lock irq_context: 0 pernet_ops_rwsem devices_rwsem rdma_nets_rwsem &device->compat_devs_mutex rcu_read_lock &rcu_state.expedited_wq &p->pi_lock &rq->__lock irq_context: 0 pernet_ops_rwsem devices_rwsem rdma_nets_rwsem &device->compat_devs_mutex rcu_read_lock &rcu_state.expedited_wq &p->pi_lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 pernet_ops_rwsem rtnl_mutex dev_pm_qos_sysfs_mtx &p->pi_lock &rq->__lock irq_context: 0 pernet_ops_rwsem rtnl_mutex dev_pm_qos_sysfs_mtx &p->pi_lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 pernet_ops_rwsem stack_depot_init_mutex rcu_node_0 irq_context: 0 pernet_ops_rwsem stack_depot_init_mutex &rcu_state.expedited_wq irq_context: 0 pernet_ops_rwsem stack_depot_init_mutex &rcu_state.expedited_wq &p->pi_lock irq_context: 0 pernet_ops_rwsem stack_depot_init_mutex &rq->__lock irq_context: 0 pernet_ops_rwsem stack_depot_init_mutex &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 pernet_ops_rwsem stack_depot_init_mutex &rcu_state.expedited_wq &p->pi_lock &rq->__lock irq_context: 0 pernet_ops_rwsem stack_depot_init_mutex &rcu_state.expedited_wq &p->pi_lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 pernet_ops_rwsem rtnl_mutex rcu_state.exp_mutex &rcu_state.expedited_wq irq_context: 0 pernet_ops_rwsem rtnl_mutex rcu_state.exp_mutex &rcu_state.expedited_wq &p->pi_lock irq_context: 0 pernet_ops_rwsem rcu_state.barrier_mutex &rq->__lock &cfs_rq->removed.lock irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem rtnl_mutex fill_pool_map-wait-type-override pool_lock#2 irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem rtnl_mutex fill_pool_map-wait-type-override &c->lock irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem rtnl_mutex fill_pool_map-wait-type-override pool_lock irq_context: 0 pernet_ops_rwsem remove_cache_srcu rcu_read_lock &rcu_state.expedited_wq irq_context: 0 pernet_ops_rwsem remove_cache_srcu rcu_read_lock &rcu_state.expedited_wq &p->pi_lock irq_context: 0 (wq_completion)events (work_completion)(&rdev->wiphy_work) &rdev->wiphy.mtx &wdev->mtx rcu_read_lock &rcu_state.expedited_wq irq_context: 0 (wq_completion)events (work_completion)(&rdev->wiphy_work) &rdev->wiphy.mtx &wdev->mtx rcu_read_lock &rcu_state.expedited_wq &p->pi_lock irq_context: 0 pernet_ops_rwsem rtnl_mutex &root->kernfs_rwsem &meta->lock irq_context: 0 pernet_ops_rwsem rtnl_mutex &root->kernfs_rwsem kfence_freelist_lock irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem rtnl_mutex &root->kernfs_rwsem &sem->wait_lock irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem rtnl_mutex dev_pm_qos_sysfs_mtx &root->kernfs_rwsem &rq->__lock irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem rtnl_mutex dev_pm_qos_sysfs_mtx &root->kernfs_rwsem &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem rtnl_mutex dev_pm_qos_sysfs_mtx &sem->wait_lock irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem rtnl_mutex dev_pm_qos_sysfs_mtx &p->pi_lock irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem rtnl_mutex dev_pm_qos_sysfs_mtx &p->pi_lock &rq->__lock irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem rtnl_mutex dev_pm_qos_sysfs_mtx &p->pi_lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem rtnl_mutex &root->kernfs_rwsem &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 pernet_ops_rwsem &root->kernfs_rwsem &sem->wait_lock irq_context: 0 pernet_ops_rwsem &root->kernfs_rwsem &rq->__lock &cfs_rq->removed.lock irq_context: 0 pernet_ops_rwsem dev_pm_qos_sysfs_mtx &root->kernfs_rwsem &sem->wait_lock irq_context: 0 pernet_ops_rwsem dev_pm_qos_sysfs_mtx &sem->wait_lock irq_context: 0 pernet_ops_rwsem dev_pm_qos_sysfs_mtx &p->pi_lock irq_context: 0 pernet_ops_rwsem dev_pm_qos_sysfs_mtx &rq->__lock irq_context: 0 pernet_ops_rwsem dev_pm_qos_sysfs_mtx &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem &device->compat_devs_mutex irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem &device->compat_devs_mutex rcu_node_0 irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem &device->compat_devs_mutex &rcu_state.expedited_wq irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem &device->compat_devs_mutex &rcu_state.expedited_wq &p->pi_lock irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem &device->compat_devs_mutex &rq->__lock irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem &device->compat_devs_mutex &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem &device->compat_devs_mutex &xa->xa_lock#17 irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem &root->kernfs_rwsem irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem &root->kernfs_rwsem &root->kernfs_iattr_rwsem &rq->__lock irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem &root->kernfs_rwsem quarantine_lock irq_context: 0 pernet_ops_rwsem &root->kernfs_rwsem &rq->__lock irq_context: 0 pernet_ops_rwsem &root->kernfs_rwsem &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem &zone->lock irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem &zone->lock &____s->seqcount irq_context: 0 pernet_ops_rwsem rtnl_mutex uevent_sock_mutex remove_cache_srcu irq_context: 0 pernet_ops_rwsem rtnl_mutex uevent_sock_mutex remove_cache_srcu quarantine_lock irq_context: 0 pernet_ops_rwsem rtnl_mutex uevent_sock_mutex remove_cache_srcu &c->lock irq_context: 0 pernet_ops_rwsem rtnl_mutex uevent_sock_mutex remove_cache_srcu &n->list_lock irq_context: 0 pernet_ops_rwsem rtnl_mutex uevent_sock_mutex remove_cache_srcu rcu_read_lock pool_lock#2 irq_context: 0 pernet_ops_rwsem rtnl_mutex uevent_sock_mutex remove_cache_srcu &obj_hash[i].lock irq_context: 0 pernet_ops_rwsem rtnl_mutex uevent_sock_mutex remove_cache_srcu pool_lock#2 irq_context: 0 pernet_ops_rwsem rtnl_mutex uevent_sock_mutex remove_cache_srcu &rq->__lock irq_context: 0 pernet_ops_rwsem rtnl_mutex uevent_sock_mutex remove_cache_srcu &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem &root->kernfs_rwsem &sem->wait_lock irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem &root->kernfs_rwsem &rq->__lock irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem &root->kernfs_rwsem &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 pernet_ops_rwsem rtnl_mutex dev_pm_qos_sysfs_mtx &p->pi_lock &cfs_rq->removed.lock irq_context: 0 pernet_ops_rwsem rtnl_mutex dev_pm_qos_sysfs_mtx &root->kernfs_rwsem &sem->wait_lock irq_context: 0 pernet_ops_rwsem rtnl_mutex dev_pm_qos_sysfs_mtx &root->kernfs_rwsem &rq->__lock irq_context: 0 pernet_ops_rwsem rtnl_mutex dev_pm_qos_sysfs_mtx &root->kernfs_rwsem &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem rcu_read_lock &____s->seqcount irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem dev_pm_qos_sysfs_mtx irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem dev_pm_qos_sysfs_mtx dev_pm_qos_sysfs_mtx.wait_lock irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem dev_pm_qos_sysfs_mtx &rq->__lock irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem dev_pm_qos_sysfs_mtx &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 pernet_ops_rwsem &root->kernfs_rwsem &p->pi_lock irq_context: 0 pernet_ops_rwsem rtnl_mutex dev_pm_qos_sysfs_mtx.wait_lock irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem dev_pm_qos_sysfs_mtx &root->kernfs_rwsem irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem dev_pm_qos_sysfs_mtx &root->kernfs_rwsem &sem->wait_lock irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem dev_pm_qos_sysfs_mtx &sem->wait_lock irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem dev_pm_qos_sysfs_mtx &p->pi_lock irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem dev_pm_qos_sysfs_mtx &p->pi_lock &rq->__lock irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem dev_pm_qos_sysfs_mtx &p->pi_lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem dev_pm_qos_sysfs_mtx &root->kernfs_rwsem irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem dev_pm_qos_sysfs_mtx &root->kernfs_rwsem &rq->__lock irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem dev_pm_qos_sysfs_mtx &root->kernfs_rwsem &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem dev_pm_qos_sysfs_mtx &p->pi_lock &cfs_rq->removed.lock irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem dev_pm_qos_sysfs_mtx &root->kernfs_rwsem &sem->wait_lock irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem dev_pm_qos_sysfs_mtx &root->kernfs_rwsem &rq->__lock irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem dev_pm_qos_sysfs_mtx &root->kernfs_rwsem &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem dev_pm_qos_sysfs_mtx dev_pm_qos_mtx irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem subsys mutex#84 irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem subsys mutex#84 &k->k_lock irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem subsys mutex#84 &k->k_lock klist_remove_lock irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem &x->wait#9 irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem dpm_list_mtx irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem &dev->power.lock irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem deferred_probe_mutex irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem device_links_lock irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem mmu_notifier_invalidate_range_start irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem uevent_sock_mutex mmu_notifier_invalidate_range_start irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem gdp_mutex irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem &device->unregistration_lock irq_context: softirq (&ndev->rs_timer) rcu_read_lock rcu_read_lock rcu_read_lock_bh &meta->lock irq_context: softirq (&ndev->rs_timer) rcu_read_lock rcu_read_lock rcu_read_lock_bh kfence_freelist_lock irq_context: 0 pernet_ops_rwsem &root->kernfs_rwsem rcu_read_lock &rcu_state.gp_wq irq_context: 0 pernet_ops_rwsem &root->kernfs_rwsem rcu_read_lock &rcu_state.gp_wq &p->pi_lock irq_context: 0 pernet_ops_rwsem &root->kernfs_rwsem rcu_read_lock &rcu_state.gp_wq &p->pi_lock &rq->__lock irq_context: 0 pernet_ops_rwsem &root->kernfs_rwsem rcu_read_lock &rcu_state.gp_wq &p->pi_lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)writeback (work_completion)(&(&wb->dwork)->work) &type->s_umount_key#43 &____s->seqcount#2 irq_context: 0 (wq_completion)writeback (work_completion)(&(&wb->dwork)->work) rcu_read_lock &c->lock irq_context: 0 (wq_completion)wg-crypt-wg0#4 (work_completion)(&peer->transmit_packet_work) &peer->endpoint_lock rcu_read_lock_bh &____s->seqcount#2 irq_context: 0 (wq_completion)wg-crypt-wg0#4 (work_completion)(&peer->transmit_packet_work) &peer->endpoint_lock rcu_read_lock_bh &____s->seqcount irq_context: 0 &mm->mmap_lock &anon_vma->rwsem &rq->__lock &cfs_rq->removed.lock irq_context: 0 &mm->mmap_lock &anon_vma->rwsem &cfs_rq->removed.lock irq_context: softirq (&app->join_timer) &app->lock batched_entropy_u8.lock irq_context: softirq (&app->join_timer) &app->lock kfence_freelist_lock irq_context: 0 pernet_ops_rwsem remove_cache_srcu rcu_node_0 irq_context: 0 pernet_ops_rwsem pgd_lock irq_context: 0 pernet_ops_rwsem key irq_context: 0 pernet_ops_rwsem pcpu_lock stock_lock irq_context: 0 pernet_ops_rwsem pool_lock irq_context: 0 &tsk->futex_exit_mutex &mm->mmap_lock &c->lock irq_context: 0 &tsk->futex_exit_mutex &mm->mmap_lock remove_cache_srcu pool_lock#2 irq_context: 0 &pipe->mutex/1 sk_lock-AF_INET6 &____s->seqcount#2 irq_context: 0 &pipe->mutex/1 sk_lock-AF_INET6 &____s->seqcount irq_context: 0 &pipe->mutex/1 sk_lock-AF_INET6 &n->list_lock irq_context: 0 &pipe->mutex/1 sk_lock-AF_INET6 &n->list_lock &c->lock irq_context: 0 &pipe->mutex/1 sk_lock-AF_INET6 &rq->__lock irq_context: 0 &pipe->mutex/1 sk_lock-AF_INET6 &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 pernet_ops_rwsem rtnl_mutex &idev->mc_lock remove_cache_srcu pool_lock#2 irq_context: 0 cb_lock genl_mutex &pn->l2tp_tunnel_idr_lock irq_context: 0 cb_lock genl_mutex &pn->l2tp_tunnel_idr_lock pool_lock#2 irq_context: 0 cb_lock genl_mutex mmu_notifier_invalidate_range_start irq_context: 0 cb_lock genl_mutex &sb->s_type->i_lock_key#8 irq_context: 0 cb_lock genl_mutex &dir->lock irq_context: 0 cb_lock genl_mutex l2tp_ip6_lock irq_context: 0 cb_lock genl_mutex k-sk_lock-AF_INET6 irq_context: 0 cb_lock genl_mutex k-sk_lock-AF_INET6 k-slock-AF_INET6 irq_context: 0 cb_lock genl_mutex k-slock-AF_INET6 irq_context: 0 cb_lock genl_mutex k-clock-AF_INET6 irq_context: 0 cb_lock genl_mutex &xa->xa_lock#9 irq_context: 0 cb_lock genl_mutex &fsnotify_mark_srcu irq_context: 0 cb_lock genl_mutex &pn->l2tp_tunnel_idr_lock &obj_hash[i].lock irq_context: 0 (wq_completion)events free_ipc_work &obj_hash[i].lock pool_lock irq_context: 0 cb_lock genl_mutex fs_reclaim &rq->__lock irq_context: 0 sb_writers#3 tomoyo_ss &____s->seqcount#2 irq_context: 0 &sb->s_type->i_mutex_key#10 rcu_read_lock rcu_read_lock rcu_node_0 irq_context: 0 &sb->s_type->i_mutex_key#10 rcu_read_lock rcu_read_lock &rq->__lock irq_context: 0 &sb->s_type->i_mutex_key#10 rcu_read_lock rcu_read_lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 &f->f_pos_lock &type->i_mutex_dir_key#4 &mm->mmap_lock ptlock_ptr(page)#2 lock#4 &lruvec->lru_lock irq_context: 0 (wq_completion)wg-crypt-wg0#4 (work_completion)(&peer->transmit_packet_work) &peer->endpoint_lock rcu_read_lock_bh &n->list_lock irq_context: 0 (wq_completion)wg-crypt-wg0#4 (work_completion)(&peer->transmit_packet_work) &peer->endpoint_lock rcu_read_lock_bh &n->list_lock &c->lock irq_context: 0 &pipe->mutex/1 sk_lock-AF_INET6 rcu_read_lock &q->lock#2 pool_lock#2 irq_context: 0 &pipe->mutex/1 sk_lock-AF_INET6 rcu_read_lock &q->lock#2 &____s->seqcount irq_context: 0 &pipe->mutex/1 sk_lock-AF_INET6 rcu_read_lock &q->lock#2 rcu_read_lock pool_lock#2 irq_context: 0 sb_writers#4 sb_internal jbd2_handle &sbi->s_orphan_lock rcu_node_0 irq_context: 0 pernet_ops_rwsem rtnl_mutex &idev->mc_lock fs_reclaim &rq->__lock irq_context: 0 pernet_ops_rwsem rtnl_mutex &idev->mc_lock fs_reclaim &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)events_unbound (work_completion)(&sub_info->work) fs_reclaim &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 pernet_ops_rwsem cpu_hotplug_lock wq_pool_mutex batched_entropy_u8.lock irq_context: 0 pernet_ops_rwsem cpu_hotplug_lock wq_pool_mutex kfence_freelist_lock irq_context: 0 pernet_ops_rwsem cpu_hotplug_lock wq_pool_mutex &meta->lock irq_context: 0 pernet_ops_rwsem nf_ct_proto_mutex &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem &wq->mutex &rq->__lock irq_context: 0 (wq_completion)events (work_completion)(&rdev->wiphy_work) &rdev->wiphy.mtx &wdev->mtx rcu_read_lock &rcu_state.expedited_wq &p->pi_lock &rq->__lock irq_context: 0 (wq_completion)events (work_completion)(&rdev->wiphy_work) &rdev->wiphy.mtx &wdev->mtx rcu_read_lock &rcu_state.expedited_wq &p->pi_lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 pernet_ops_rwsem lock kernfs_idr_lock &c->lock irq_context: 0 pernet_ops_rwsem fill_pool_map-wait-type-override &rq->__lock &cfs_rq->removed.lock irq_context: 0 &f->f_pos_lock &type->i_mutex_dir_key#4 &p->pi_lock &rq->__lock &cfs_rq->removed.lock irq_context: 0 sb_writers#5 &type->i_mutex_dir_key#5/1 fill_pool_map-wait-type-override &n->list_lock irq_context: 0 sb_writers#5 &type->i_mutex_dir_key#5/1 fill_pool_map-wait-type-override &n->list_lock &c->lock irq_context: 0 &vma->vm_lock->lock rcu_read_lock &rcu_state.expedited_wq &p->pi_lock irq_context: 0 &vma->vm_lock->lock rcu_read_lock &rcu_state.expedited_wq &p->pi_lock &rq->__lock irq_context: 0 &vma->vm_lock->lock rcu_read_lock &rcu_state.expedited_wq &p->pi_lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 sb_writers#3 tomoyo_ss &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem rcu_state.barrier_mutex &cfs_rq->removed.lock irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem rcu_state.barrier_mutex &obj_hash[i].lock irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem rcu_state.barrier_mutex pool_lock#2 irq_context: 0 pernet_ops_rwsem devices_rwsem rdma_nets_rwsem &device->compat_devs_mutex remove_cache_srcu rcu_read_lock rcu_node_0 irq_context: 0 pernet_ops_rwsem devices_rwsem rdma_nets_rwsem &device->compat_devs_mutex remove_cache_srcu rcu_read_lock &rq->__lock irq_context: 0 pernet_ops_rwsem devices_rwsem rdma_nets_rwsem &device->compat_devs_mutex remove_cache_srcu rcu_read_lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 &sig->cred_guard_mutex tomoyo_ss rcu_node_0 irq_context: 0 pernet_ops_rwsem devices_rwsem rdma_nets_rwsem &device->compat_devs_mutex &root->kernfs_rwsem &p->pi_lock irq_context: 0 &sbi->s_writepages_rwsem mmu_notifier_invalidate_range_start &rq->__lock irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem rtnl_mutex &root->kernfs_rwsem rcu_node_0 irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem rdma_nets_rwsem rdma_nets_rwsem.wait_lock irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem rdma_nets_rwsem &rq->__lock irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem rdma_nets_rwsem &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 &sig->cred_guard_mutex remove_cache_srcu &rq->__lock irq_context: 0 &sig->cred_guard_mutex remove_cache_srcu &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 pernet_ops_rwsem rtnl_mutex &tn->lock &rq->__lock irq_context: 0 pernet_ops_rwsem rtnl_mutex &tn->lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 pernet_ops_rwsem rtnl_mutex uevent_sock_mutex fs_reclaim mmu_notifier_invalidate_range_start &rq->__lock irq_context: 0 pernet_ops_rwsem rtnl_mutex uevent_sock_mutex fs_reclaim mmu_notifier_invalidate_range_start &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)bat_events (work_completion)(&(&bat_priv->bla.work)->work) rcu_read_lock &n->list_lock irq_context: 0 (wq_completion)bat_events (work_completion)(&(&bat_priv->bla.work)->work) rcu_read_lock &n->list_lock &c->lock irq_context: 0 pernet_ops_rwsem k-sk_lock-AF_RXRPC &rxnet->local_mutex k-sk_lock-AF_INET6 &rq->__lock irq_context: 0 pernet_ops_rwsem k-sk_lock-AF_RXRPC &rxnet->local_mutex k-sk_lock-AF_INET6 &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 pernet_ops_rwsem ovs_mutex nf_ct_proto_mutex defrag6_mutex nf_hook_mutex &rq->__lock irq_context: 0 pernet_ops_rwsem ovs_mutex nf_ct_proto_mutex defrag6_mutex nf_hook_mutex &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)events_unbound connector_reaper_work rcu_node_0 irq_context: 0 (wq_completion)events_unbound connector_reaper_work &rcu_state.expedited_wq irq_context: 0 (wq_completion)events_unbound connector_reaper_work &rcu_state.expedited_wq &p->pi_lock irq_context: 0 (wq_completion)events_unbound connector_reaper_work &rcu_state.expedited_wq &p->pi_lock &rq->__lock irq_context: 0 (wq_completion)events_unbound connector_reaper_work &rcu_state.expedited_wq &p->pi_lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 pernet_ops_rwsem devices_rwsem rdma_nets_rwsem.wait_lock irq_context: 0 pernet_ops_rwsem devices_rwsem &p->pi_lock irq_context: 0 pernet_ops_rwsem devices_rwsem &p->pi_lock &rq->__lock irq_context: 0 pernet_ops_rwsem devices_rwsem &p->pi_lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 pernet_ops_rwsem devices_rwsem &rq->__lock irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem &root->kernfs_rwsem rcu_read_lock &rq->__lock irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem &root->kernfs_rwsem rcu_read_lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem &root->kernfs_rwsem &root->kernfs_iattr_rwsem &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 pernet_ops_rwsem rtnl_mutex stock_lock rcu_read_lock per_cpu_ptr(&cgroup_rstat_cpu_lock, cpu) irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem &root->kernfs_rwsem kernfs_idr_lock &obj_hash[i].lock pool_lock irq_context: 0 (wq_completion)bat_events (work_completion)(&(&bat_priv->mcast.work)->work) rcu_read_lock &rcu_state.expedited_wq irq_context: 0 (wq_completion)bat_events (work_completion)(&(&bat_priv->mcast.work)->work) rcu_read_lock &rcu_state.expedited_wq &p->pi_lock irq_context: 0 (wq_completion)bat_events (work_completion)(&(&bat_priv->mcast.work)->work) rcu_read_lock &rcu_state.expedited_wq &p->pi_lock &rq->__lock irq_context: 0 (wq_completion)bat_events (work_completion)(&(&bat_priv->mcast.work)->work) rcu_read_lock &rcu_state.expedited_wq &p->pi_lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem &root->kernfs_rwsem rcu_node_0 irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem &root->kernfs_rwsem &rcu_state.expedited_wq irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem &root->kernfs_rwsem &rcu_state.expedited_wq &p->pi_lock irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem &root->kernfs_rwsem &rcu_state.expedited_wq &p->pi_lock &rq->__lock irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem &root->kernfs_rwsem &rcu_state.expedited_wq &p->pi_lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)wg-crypt-wg1#3 (work_completion)(&peer->transmit_packet_work) &peer->endpoint_lock rcu_read_lock_bh batched_entropy_u32.lock crngs.lock irq_context: 0 pernet_ops_rwsem rtnl_mutex sysctl_lock krc.lock &obj_hash[i].lock irq_context: 0 pernet_ops_rwsem rtnl_mutex sysctl_lock krc.lock &base->lock irq_context: 0 pernet_ops_rwsem rtnl_mutex sysctl_lock krc.lock &base->lock &obj_hash[i].lock irq_context: 0 (wq_completion)inet_frag_wq (work_completion)(&fqdir->destroy_work) &ht->mutex &rq->__lock irq_context: 0 &pool->lock &p->pi_lock &rq->__lock &obj_hash[i].lock irq_context: 0 &pool->lock &p->pi_lock &rq->__lock &base->lock irq_context: 0 &pool->lock &p->pi_lock &rq->__lock &base->lock &obj_hash[i].lock irq_context: 0 pernet_ops_rwsem rcu_state.barrier_mutex &cfs_rq->removed.lock irq_context: 0 pernet_ops_rwsem rcu_state.barrier_mutex &obj_hash[i].lock irq_context: 0 pernet_ops_rwsem rcu_state.barrier_mutex pool_lock#2 irq_context: 0 vlan_ioctl_mutex rtnl_mutex _xmit_NONE irq_context: 0 vlan_ioctl_mutex rtnl_mutex &root->kernfs_rwsem &rq->__lock irq_context: 0 vlan_ioctl_mutex rtnl_mutex &root->kernfs_rwsem &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 vlan_ioctl_mutex rtnl_mutex &sem->wait_lock irq_context: 0 vlan_ioctl_mutex rtnl_mutex &p->pi_lock irq_context: 0 vlan_ioctl_mutex rtnl_mutex &p->pi_lock &rq->__lock irq_context: 0 vlan_ioctl_mutex rtnl_mutex &p->pi_lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 vlan_ioctl_mutex rtnl_mutex uevent_sock_mutex &n->list_lock irq_context: 0 vlan_ioctl_mutex rtnl_mutex uevent_sock_mutex &n->list_lock &c->lock irq_context: 0 vlan_ioctl_mutex rtnl_mutex &idev->mc_lock fs_reclaim &rq->__lock irq_context: 0 vlan_ioctl_mutex rtnl_mutex &idev->mc_lock fs_reclaim &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 rtnl_mutex &bridge_netdev_addr_lock_key &obj_hash[i].lock irq_context: 0 rtnl_mutex &bridge_netdev_addr_lock_key krc.lock irq_context: 0 rtnl_mutex &pmc->lock irq_context: 0 (wq_completion)events (work_completion)(&rdev->wiphy_work) &rdev->wiphy.mtx rcu_node_0 irq_context: 0 &vma->vm_lock->lock fs_reclaim mmu_notifier_invalidate_range_start &cfs_rq->removed.lock irq_context: 0 &vma->vm_lock->lock fs_reclaim mmu_notifier_invalidate_range_start &obj_hash[i].lock irq_context: 0 rtnl_mutex &pnettable->lock &rq->__lock irq_context: 0 rtnl_mutex &pnettable->lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 &mm->mmap_lock &mapping->i_mmap_rwsem &anon_vma->rwsem rcu_node_0 irq_context: 0 &mm->mmap_lock &mapping->i_mmap_rwsem &anon_vma->rwsem rcu_read_lock &rcu_state.expedited_wq irq_context: 0 &mm->mmap_lock &mapping->i_mmap_rwsem &anon_vma->rwsem rcu_read_lock &rcu_state.expedited_wq &p->pi_lock irq_context: 0 &mm->mmap_lock &mapping->i_mmap_rwsem &anon_vma->rwsem rcu_read_lock &rcu_state.expedited_wq &p->pi_lock &rq->__lock irq_context: 0 &mm->mmap_lock &mapping->i_mmap_rwsem &anon_vma->rwsem rcu_read_lock &rcu_state.expedited_wq &p->pi_lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 sk_lock-AF_INET fill_pool_map-wait-type-override pool_lock#2 irq_context: 0 sk_lock-AF_INET fill_pool_map-wait-type-override &c->lock irq_context: 0 sk_lock-AF_INET fill_pool_map-wait-type-override pool_lock irq_context: 0 &sig->cred_guard_mutex remove_cache_srcu rcu_read_lock rcu_node_0 irq_context: 0 &sig->cred_guard_mutex remove_cache_srcu rcu_read_lock &rq->__lock irq_context: 0 &sig->cred_guard_mutex remove_cache_srcu rcu_read_lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 sb_writers#4 &type->i_mutex_dir_key#3/1 &sb->s_type->i_mutex_key#8 rcu_node_0 irq_context: 0 &group->mark_mutex rcu_read_lock rcu_node_0 irq_context: 0 &group->mark_mutex rcu_read_lock &rq->__lock irq_context: 0 &group->mark_mutex rcu_read_lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 &sb->s_type->i_mutex_key#10 sk_lock-AF_INET6 &hashinfo->ehash_locks[i] irq_context: 0 &sb->s_type->i_mutex_key#10 sk_lock-AF_INET6 slock-AF_INET6 &base->lock irq_context: 0 &sb->s_type->i_mutex_key#10 sk_lock-AF_INET6 slock-AF_INET6 &base->lock &obj_hash[i].lock irq_context: 0 &sb->s_type->i_mutex_key#10 sk_lock-AF_INET6 slock-AF_INET6 pool_lock#2 irq_context: 0 (wq_completion)events_unbound (work_completion)(&sub_info->work) tasklist_lock &n->list_lock irq_context: 0 sb_writers#4 &type->i_mutex_dir_key#3/1 &sb->s_type->i_mutex_key#8 &obj_hash[i].lock pool_lock irq_context: 0 rtnl_mutex &bridge_netdev_addr_lock_key &obj_hash[i].lock pool_lock irq_context: 0 (wq_completion)mld (work_completion)(&(&idev->mc_ifc_work)->work) &idev->mc_lock rcu_read_lock rcu_read_lock rcu_read_lock_bh &meta->lock irq_context: 0 (wq_completion)mld (work_completion)(&(&idev->mc_ifc_work)->work) &idev->mc_lock rcu_read_lock rcu_read_lock rcu_read_lock_bh kfence_freelist_lock irq_context: 0 rtnl_mutex &ul->lock#2 pool_lock#2 irq_context: 0 rtnl_mutex &ul->lock#2 &dir->lock#2 irq_context: 0 (wq_completion)events_unbound (work_completion)(&(&kfence_timer)->work) cpu_hotplug_lock &rq->__lock irq_context: softirq &(&cache_cleaner)->timer rcu_read_lock &pool->lock &p->pi_lock &rq->__lock &cfs_rq->removed.lock irq_context: softirq &(&chan->chan_timer)->timer irq_context: softirq &(&chan->chan_timer)->timer rcu_read_lock &pool->lock irq_context: softirq &(&chan->chan_timer)->timer rcu_read_lock &pool->lock &obj_hash[i].lock irq_context: softirq &(&chan->chan_timer)->timer rcu_read_lock &pool->lock &p->pi_lock irq_context: softirq &(&chan->chan_timer)->timer rcu_read_lock &pool->lock &p->pi_lock &rq->__lock irq_context: softirq &(&chan->chan_timer)->timer rcu_read_lock &pool->lock &p->pi_lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: softirq &(&chan->chan_timer)->timer rcu_read_lock &pool->lock &p->pi_lock &rq->__lock &cfs_rq->removed.lock irq_context: 0 (wq_completion)events (work_completion)(&(&chan->chan_timer)->work) irq_context: 0 (wq_completion)events (work_completion)(&(&chan->chan_timer)->work) &conn->chan_lock irq_context: 0 (wq_completion)events (work_completion)(&(&chan->chan_timer)->work) &conn->chan_lock &chan->lock/1 irq_context: 0 (wq_completion)events (work_completion)(&(&chan->chan_timer)->work) &conn->chan_lock &chan->lock/1 &obj_hash[i].lock irq_context: 0 (wq_completion)events (work_completion)(&(&chan->chan_timer)->work) &conn->chan_lock &chan->lock/1 sk_lock-AF_BLUETOOTH-BTPROTO_L2CAP/1 irq_context: 0 (wq_completion)events (work_completion)(&(&chan->chan_timer)->work) &conn->chan_lock &chan->lock/1 sk_lock-AF_BLUETOOTH-BTPROTO_L2CAP/1 slock-AF_BLUETOOTH-BTPROTO_L2CAP irq_context: 0 (wq_completion)events (work_completion)(&(&chan->chan_timer)->work) &conn->chan_lock &chan->lock/1 sk_lock-AF_BLUETOOTH-BTPROTO_L2CAP/1 rfcomm_wq.lock irq_context: 0 (wq_completion)events (work_completion)(&(&chan->chan_timer)->work) &conn->chan_lock &chan->lock/1 sk_lock-AF_BLUETOOTH-BTPROTO_L2CAP/1 rfcomm_wq.lock &p->pi_lock irq_context: 0 (wq_completion)events (work_completion)(&(&chan->chan_timer)->work) &conn->chan_lock &chan->lock/1 sk_lock-AF_BLUETOOTH-BTPROTO_L2CAP/1 rfcomm_wq.lock &p->pi_lock &rq->__lock irq_context: 0 (wq_completion)events (work_completion)(&(&chan->chan_timer)->work) &conn->chan_lock &chan->lock/1 sk_lock-AF_BLUETOOTH-BTPROTO_L2CAP/1 rfcomm_wq.lock &p->pi_lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)events (work_completion)(&(&chan->chan_timer)->work) &conn->chan_lock &chan->lock/1 slock-AF_BLUETOOTH-BTPROTO_L2CAP irq_context: 0 (wq_completion)events (work_completion)(&(&chan->chan_timer)->work) &conn->chan_lock &chan->lock/1 rcu_read_lock &pool->lock/1 irq_context: 0 (wq_completion)events (work_completion)(&(&chan->chan_timer)->work) &conn->chan_lock &chan->lock/1 rcu_read_lock &pool->lock/1 &obj_hash[i].lock irq_context: 0 (wq_completion)events (work_completion)(&(&chan->chan_timer)->work) &conn->chan_lock &chan->lock/1 rcu_read_lock &pool->lock/1 pool_lock#2 irq_context: 0 (wq_completion)events (work_completion)(&(&chan->chan_timer)->work) &conn->chan_lock &chan->lock/1 rcu_read_lock &pool->lock/1 &p->pi_lock irq_context: 0 (wq_completion)events (work_completion)(&(&chan->chan_timer)->work) &conn->chan_lock &chan->lock/1 rcu_read_lock &pool->lock/1 &p->pi_lock &rq->__lock irq_context: 0 (wq_completion)events (work_completion)(&(&chan->chan_timer)->work) &conn->chan_lock &chan->lock/1 rcu_read_lock &pool->lock/1 &p->pi_lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)events (work_completion)(&(&chan->chan_timer)->work) &conn->chan_lock &chan->lock/1 &rq->__lock irq_context: 0 (wq_completion)hci5#2 (work_completion)(&(&conn->disc_work)->work) irq_context: 0 (wq_completion)hci5#2 (work_completion)(&(&conn->disc_work)->work) &hdev->unregister_lock irq_context: 0 (wq_completion)hci5#2 (work_completion)(&(&conn->disc_work)->work) &hdev->unregister_lock fs_reclaim irq_context: 0 (wq_completion)hci5#2 (work_completion)(&(&conn->disc_work)->work) &hdev->unregister_lock fs_reclaim mmu_notifier_invalidate_range_start irq_context: 0 (wq_completion)hci5#2 (work_completion)(&(&conn->disc_work)->work) &hdev->unregister_lock &c->lock irq_context: 0 (wq_completion)hci5#2 (work_completion)(&(&conn->disc_work)->work) &hdev->unregister_lock remove_cache_srcu irq_context: 0 (wq_completion)hci5#2 (work_completion)(&(&conn->disc_work)->work) &hdev->unregister_lock remove_cache_srcu quarantine_lock irq_context: 0 (wq_completion)hci5#2 (work_completion)(&(&conn->disc_work)->work) &hdev->unregister_lock remove_cache_srcu &c->lock irq_context: 0 (wq_completion)hci5#2 (work_completion)(&(&conn->disc_work)->work) &hdev->unregister_lock remove_cache_srcu &n->list_lock irq_context: 0 (wq_completion)hci5#2 (work_completion)(&(&conn->disc_work)->work) &hdev->unregister_lock remove_cache_srcu rcu_read_lock pool_lock#2 irq_context: 0 (wq_completion)hci5#2 (work_completion)(&(&conn->disc_work)->work) &hdev->unregister_lock remove_cache_srcu &obj_hash[i].lock irq_context: 0 (wq_completion)hci5#2 (work_completion)(&(&conn->disc_work)->work) &hdev->unregister_lock pool_lock#2 irq_context: 0 (wq_completion)hci5#2 (work_completion)(&(&conn->disc_work)->work) &hdev->unregister_lock &hdev->cmd_sync_work_lock irq_context: 0 (wq_completion)hci5#2 (work_completion)(&(&conn->disc_work)->work) &hdev->unregister_lock rcu_read_lock &pool->lock/1 irq_context: 0 (wq_completion)hci5#2 (work_completion)(&(&conn->disc_work)->work) &hdev->unregister_lock rcu_read_lock &pool->lock/1 &obj_hash[i].lock irq_context: 0 (wq_completion)hci5#2 (work_completion)(&(&conn->disc_work)->work) &hdev->unregister_lock rcu_read_lock &pool->lock/1 &p->pi_lock irq_context: 0 (wq_completion)hci5#2 (work_completion)(&(&conn->disc_work)->work) &hdev->unregister_lock rcu_read_lock &pool->lock/1 &p->pi_lock &rq->__lock irq_context: 0 (wq_completion)hci5#2 (work_completion)(&(&conn->disc_work)->work) &hdev->unregister_lock rcu_read_lock &pool->lock/1 &p->pi_lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)hci5#2 (work_completion)(&(&conn->disc_work)->work) &hdev->unregister_lock &rq->__lock irq_context: 0 (wq_completion)hci5 (work_completion)(&hdev->cmd_sync_work) &hdev->req_lock &hdev->lock irq_context: 0 (wq_completion)hci5 (work_completion)(&hdev->cmd_sync_work) &hdev->req_lock &hdev->lock fs_reclaim irq_context: 0 (wq_completion)hci5 (work_completion)(&hdev->cmd_sync_work) &hdev->req_lock &hdev->lock fs_reclaim mmu_notifier_invalidate_range_start irq_context: 0 (wq_completion)hci5 (work_completion)(&hdev->cmd_sync_work) &hdev->req_lock &hdev->lock pool_lock#2 irq_context: 0 (wq_completion)hci5 (work_completion)(&hdev->cmd_sync_work) &hdev->req_lock &hdev->lock tk_core.seq.seqcount irq_context: 0 (wq_completion)hci5 (work_completion)(&hdev->cmd_sync_work) &hdev->req_lock &hdev->lock hci_sk_list.lock irq_context: 0 (wq_completion)hci5 (work_completion)(&hdev->cmd_sync_work) &hdev->req_lock &hdev->lock &obj_hash[i].lock irq_context: 0 (wq_completion)hci5 (work_completion)(&hdev->cmd_sync_work) &hdev->req_lock &hdev->lock hci_cb_list_lock irq_context: 0 (wq_completion)hci5 (work_completion)(&hdev->cmd_sync_work) &hdev->req_lock &hdev->lock hci_cb_list_lock &list->lock#10 irq_context: 0 (wq_completion)hci5 (work_completion)(&hdev->cmd_sync_work) &hdev->req_lock &hdev->lock hci_cb_list_lock &obj_hash[i].lock irq_context: 0 (wq_completion)hci5 (work_completion)(&hdev->cmd_sync_work) &hdev->req_lock &hdev->lock hci_cb_list_lock (work_completion)(&(&conn->id_addr_timer)->work) irq_context: 0 (wq_completion)hci5 (work_completion)(&hdev->cmd_sync_work) &hdev->req_lock &hdev->lock hci_cb_list_lock &conn->chan_lock irq_context: 0 (wq_completion)hci5 (work_completion)(&hdev->cmd_sync_work) &hdev->req_lock &hdev->lock hci_cb_list_lock &conn->chan_lock &lock->wait_lock irq_context: 0 (wq_completion)hci5 (work_completion)(&hdev->cmd_sync_work) &hdev->req_lock &hdev->lock hci_cb_list_lock &conn->chan_lock &rq->__lock irq_context: 0 (wq_completion)hci5 (work_completion)(&hdev->cmd_sync_work) &hdev->req_lock &hdev->lock hci_cb_list_lock &conn->chan_lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)events (work_completion)(&(&chan->chan_timer)->work) &lock->wait_lock irq_context: 0 (wq_completion)events (work_completion)(&(&chan->chan_timer)->work) &p->pi_lock irq_context: 0 (wq_completion)events (work_completion)(&(&chan->chan_timer)->work) &p->pi_lock &rq->__lock irq_context: 0 (wq_completion)events (work_completion)(&(&chan->chan_timer)->work) &p->pi_lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)events (work_completion)(&(&chan->chan_timer)->work) &rq->__lock irq_context: 0 (wq_completion)hci5 (work_completion)(&hdev->cmd_sync_work) &hdev->req_lock &hdev->lock hci_cb_list_lock rcu_state.exp_mutex rcu_node_0 irq_context: 0 (wq_completion)hci5 (work_completion)(&hdev->cmd_sync_work) &hdev->req_lock &hdev->lock hci_cb_list_lock rcu_state.exp_mutex &obj_hash[i].lock irq_context: 0 (wq_completion)hci5 (work_completion)(&hdev->cmd_sync_work) &hdev->req_lock &hdev->lock hci_cb_list_lock rcu_state.exp_mutex rcu_read_lock &pool->lock irq_context: 0 (wq_completion)hci5 (work_completion)(&hdev->cmd_sync_work) &hdev->req_lock &hdev->lock hci_cb_list_lock rcu_state.exp_mutex rcu_read_lock &pool->lock &obj_hash[i].lock irq_context: 0 (wq_completion)hci5 (work_completion)(&hdev->cmd_sync_work) &hdev->req_lock &hdev->lock hci_cb_list_lock rcu_state.exp_mutex &rnp->exp_wq[0] irq_context: 0 (wq_completion)hci5 (work_completion)(&hdev->cmd_sync_work) &hdev->req_lock &hdev->lock hci_cb_list_lock rcu_state.exp_mutex &rq->__lock irq_context: 0 (wq_completion)hci5 (work_completion)(&hdev->cmd_sync_work) &hdev->req_lock &hdev->lock hci_cb_list_lock rcu_state.exp_mutex &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 rfcomm_mutex rlock-AF_BLUETOOTH irq_context: 0 (wq_completion)hci5 (work_completion)(&hdev->cmd_sync_work) &hdev->req_lock &hdev->lock hci_cb_list_lock &list->lock#9 irq_context: 0 rfcomm_mutex l2cap_sk_list.lock irq_context: 0 (wq_completion)hci5 (work_completion)(&hdev->cmd_sync_work) &hdev->req_lock &hdev->lock hci_cb_list_lock pool_lock#2 irq_context: 0 rfcomm_mutex &chan->lock/1 irq_context: 0 (wq_completion)hci5 (work_completion)(&hdev->cmd_sync_work) &hdev->req_lock &hdev->lock (work_completion)(&(&conn->disc_work)->work) irq_context: 0 rfcomm_mutex &chan->lock/1 sk_lock-AF_BLUETOOTH-BTPROTO_L2CAP/1 irq_context: 0 (wq_completion)hci5 (work_completion)(&hdev->cmd_sync_work) &hdev->req_lock &hdev->lock rcu_read_lock &pool->lock/1 irq_context: 0 rfcomm_mutex &chan->lock/1 sk_lock-AF_BLUETOOTH-BTPROTO_L2CAP/1 slock-AF_BLUETOOTH-BTPROTO_L2CAP irq_context: 0 rfcomm_mutex &chan->lock/1 slock-AF_BLUETOOTH-BTPROTO_L2CAP irq_context: 0 (wq_completion)hci5 (work_completion)(&hdev->cmd_sync_work) &hdev->req_lock &hdev->lock (work_completion)(&(&conn->auto_accept_work)->work) irq_context: 0 rfcomm_mutex &chan->lock/1 k-clock-AF_BLUETOOTH irq_context: 0 rfcomm_mutex &chan->lock/1 rlock-AF_BLUETOOTH irq_context: 0 rfcomm_mutex &chan->lock/1 wlock-AF_BLUETOOTH irq_context: 0 (wq_completion)hci5 (work_completion)(&hdev->cmd_sync_work) &hdev->req_lock &hdev->lock (work_completion)(&(&conn->idle_work)->work) irq_context: 0 rfcomm_mutex &chan->lock/1 pool_lock#2 irq_context: 0 (wq_completion)hci5 (work_completion)(&hdev->cmd_sync_work) &hdev->req_lock &hdev->lock &list->lock#9 irq_context: 0 rfcomm_mutex &chan->lock/1 &dir->lock irq_context: 0 (wq_completion)hci5 (work_completion)(&hdev->cmd_sync_work) &hdev->req_lock &hdev->lock rcu_state.exp_mutex rcu_node_0 irq_context: 0 rfcomm_mutex &chan->lock/1 &obj_hash[i].lock irq_context: 0 (wq_completion)hci5 (work_completion)(&hdev->cmd_sync_work) &hdev->req_lock &hdev->lock rcu_state.exp_mutex &obj_hash[i].lock irq_context: 0 rfcomm_mutex chan_list_lock irq_context: 0 (wq_completion)hci5 (work_completion)(&hdev->cmd_sync_work) &hdev->req_lock &hdev->lock rcu_state.exp_mutex rcu_read_lock &pool->lock irq_context: 0 rfcomm_mutex &xa->xa_lock#9 irq_context: 0 (wq_completion)hci5 (work_completion)(&hdev->cmd_sync_work) &hdev->req_lock &hdev->lock rcu_state.exp_mutex rcu_read_lock &pool->lock &obj_hash[i].lock irq_context: 0 rfcomm_mutex &fsnotify_mark_srcu irq_context: 0 (wq_completion)hci5 (work_completion)(&hdev->cmd_sync_work) &hdev->req_lock &hdev->lock rcu_state.exp_mutex &rnp->exp_wq[1] irq_context: 0 (wq_completion)hci5 (work_completion)(&hdev->cmd_sync_work) &hdev->req_lock &hdev->lock rcu_state.exp_mutex &rq->__lock irq_context: 0 (wq_completion)hci5 (work_completion)(&hdev->cmd_sync_work) &hdev->req_lock &hdev->lock rcu_state.exp_mutex &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)wg-crypt-wg0#2 (work_completion)(&peer->transmit_packet_work) &rcu_state.expedited_wq irq_context: 0 (wq_completion)wg-crypt-wg0#2 (work_completion)(&peer->transmit_packet_work) &rcu_state.expedited_wq &p->pi_lock irq_context: 0 (wq_completion)wg-crypt-wg0#2 (work_completion)(&peer->transmit_packet_work) &rcu_state.expedited_wq &p->pi_lock &rq->__lock irq_context: 0 (wq_completion)wg-crypt-wg0#2 (work_completion)(&peer->transmit_packet_work) &rcu_state.expedited_wq &p->pi_lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)events (work_completion)(&(&nsim_dev->trap_data->trap_report_dw)->work) &devlink->lock_key#4 &nsim_trap_data->trap_lock &base->lock irq_context: 0 (wq_completion)events (work_completion)(&(&nsim_dev->trap_data->trap_report_dw)->work) &devlink->lock_key#4 &nsim_trap_data->trap_lock &base->lock &obj_hash[i].lock irq_context: 0 pernet_ops_rwsem rcu_state.exp_mutex rcu_read_lock &pool->lock &p->pi_lock &rq->__lock &cfs_rq->removed.lock irq_context: 0 &sig->cred_guard_mutex tomoyo_ss mmu_notifier_invalidate_range_start &rq->__lock irq_context: 0 &sig->cred_guard_mutex &tsk->futex_exit_mutex &rq->__lock irq_context: 0 &sig->cred_guard_mutex &tsk->futex_exit_mutex &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 &f->f_pos_lock sb_writers#4 &sb->s_type->i_mutex_key#8 fs_reclaim &cfs_rq->removed.lock irq_context: 0 pernet_ops_rwsem remove_cache_srcu &rq->__lock &cfs_rq->removed.lock irq_context: 0 &pipe->mutex/1 sk_lock-AF_INET6 rcu_read_lock rcu_read_lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 remove_cache_srcu stock_lock irq_context: 0 remove_cache_srcu pcpu_lock stock_lock irq_context: 0 (wq_completion)bat_events (work_completion)(&(&bat_priv->mcast.work)->work) rcu_node_0 irq_context: 0 pernet_ops_rwsem devices_rwsem rdma_nets_rwsem &device->compat_devs_mutex remove_cache_srcu rcu_read_lock &rcu_state.gp_wq irq_context: 0 pernet_ops_rwsem devices_rwsem rdma_nets_rwsem &device->compat_devs_mutex remove_cache_srcu rcu_read_lock &rcu_state.gp_wq &p->pi_lock irq_context: 0 pernet_ops_rwsem devices_rwsem rdma_nets_rwsem &device->compat_devs_mutex remove_cache_srcu rcu_read_lock &rcu_state.gp_wq &p->pi_lock &rq->__lock irq_context: 0 pernet_ops_rwsem devices_rwsem rdma_nets_rwsem &device->compat_devs_mutex remove_cache_srcu rcu_read_lock &rcu_state.gp_wq &p->pi_lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 pernet_ops_rwsem rtnl_mutex remove_cache_srcu rcu_read_lock &rcu_state.gp_wq irq_context: 0 pernet_ops_rwsem rtnl_mutex remove_cache_srcu rcu_read_lock &rcu_state.gp_wq &p->pi_lock irq_context: 0 pernet_ops_rwsem rtnl_mutex remove_cache_srcu rcu_read_lock &rcu_state.gp_wq &p->pi_lock &rq->__lock irq_context: 0 pernet_ops_rwsem rtnl_mutex remove_cache_srcu rcu_read_lock &rcu_state.gp_wq &p->pi_lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 pernet_ops_rwsem rtnl_mutex dpm_list_mtx &rq->__lock irq_context: 0 pernet_ops_rwsem rtnl_mutex dpm_list_mtx &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)ext4-rsv-conversion (work_completion)(&ei->i_rsv_conversion_work) jbd2_handle &ei->i_data_sem remove_cache_srcu rcu_read_lock rcu_node_0 irq_context: 0 (wq_completion)ext4-rsv-conversion (work_completion)(&ei->i_rsv_conversion_work) jbd2_handle &ei->i_data_sem remove_cache_srcu rcu_read_lock &rcu_state.expedited_wq irq_context: 0 (wq_completion)ext4-rsv-conversion (work_completion)(&ei->i_rsv_conversion_work) jbd2_handle &ei->i_data_sem remove_cache_srcu rcu_read_lock &rcu_state.expedited_wq &p->pi_lock irq_context: 0 pernet_ops_rwsem rtnl_mutex &root->kernfs_rwsem rcu_node_0 irq_context: 0 cb_lock &rdev->wiphy.mtx &local->mtx &____s->seqcount#2 irq_context: 0 &sb->s_type->i_mutex_key#8 rcu_read_lock rcu_read_lock rcu_read_lock rcu_node_0 irq_context: 0 &sb->s_type->i_mutex_key#8 rcu_read_lock rcu_read_lock rcu_read_lock &rq->__lock irq_context: 0 &sb->s_type->i_mutex_key#8 rcu_read_lock rcu_read_lock rcu_read_lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 cb_lock &rdev->wiphy.mtx &local->mtx fs_reclaim mmu_notifier_invalidate_range_start &rq->__lock irq_context: 0 cb_lock &rdev->wiphy.mtx &local->mtx fs_reclaim mmu_notifier_invalidate_range_start &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)events (work_completion)(&rdev->wiphy_work) &rdev->wiphy.mtx &wdev->mtx rcu_read_lock rcu_read_lock &sta->rate_ctrl_lock &obj_hash[i].lock pool_lock irq_context: 0 (wq_completion)bat_events (work_completion)(&(&forw_packet_aggr->delayed_work)->work) &hard_iface->bat_iv.ogm_buff_mutex rcu_read_lock fill_pool_map-wait-type-override rcu_node_0 irq_context: 0 (wq_completion)bat_events (work_completion)(&(&forw_packet_aggr->delayed_work)->work) &hard_iface->bat_iv.ogm_buff_mutex rcu_read_lock fill_pool_map-wait-type-override &rq->__lock irq_context: 0 &sb->s_type->i_mutex_key#8 kauditd_wait.lock &p->pi_lock &rq->__lock &cfs_rq->removed.lock irq_context: 0 sk_lock-AF_INET rcu_read_lock &tbl->lock irq_context: 0 sk_lock-AF_INET rcu_read_lock &n->lock irq_context: 0 sk_lock-AF_INET rcu_read_lock &n->lock &obj_hash[i].lock irq_context: 0 sk_lock-AF_INET rcu_read_lock &n->lock &base->lock irq_context: 0 sk_lock-AF_INET rcu_read_lock &n->lock &base->lock &obj_hash[i].lock irq_context: 0 sk_lock-AF_INET rcu_read_lock &n->lock pool_lock#2 irq_context: 0 sk_lock-AF_INET rcu_read_lock rcu_read_lock_bh rcu_read_lock &list->lock#5 irq_context: 0 &sbi->s_writepages_rwsem jbd2_handle &ei->i_es_lock key#2 irq_context: 0 (wq_completion)events (work_completion)(&nlk->work) irq_context: 0 (wq_completion)events (work_completion)(&nlk->work) &obj_hash[i].lock irq_context: 0 (wq_completion)events (work_completion)(&nlk->work) pool_lock#2 irq_context: 0 (wq_completion)events (work_completion)(&nlk->work) rlock-AF_NETLINK irq_context: 0 (wq_completion)events (work_completion)(&nlk->work) &dir->lock irq_context: 0 &sig->cred_guard_mutex tomoyo_ss mmu_notifier_invalidate_range_start &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 cb_lock nlk_cb_mutex-GENERIC &____s->seqcount#2 irq_context: 0 sb_writers#4 sb_internal jbd2_handle &journal->j_state_lock &journal->j_wait_commit irq_context: 0 sb_writers#4 sb_internal jbd2_handle &journal->j_state_lock &journal->j_wait_commit &p->pi_lock irq_context: 0 sb_writers#4 sb_internal jbd2_handle &journal->j_state_lock &journal->j_wait_commit &p->pi_lock &rq->__lock irq_context: 0 sb_writers#4 sb_internal jbd2_handle &journal->j_state_lock &journal->j_wait_commit &p->pi_lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 cb_lock nlk_cb_mutex-GENERIC genl_mutex rtnl_mutex rtnl_mutex.wait_lock irq_context: 0 cb_lock nlk_cb_mutex-GENERIC genl_mutex rtnl_mutex &rq->__lock irq_context: 0 cb_lock nlk_cb_mutex-GENERIC genl_mutex rtnl_mutex &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 vlan_ioctl_mutex rtnl_mutex &root->kernfs_rwsem &obj_hash[i].lock irq_context: 0 vlan_ioctl_mutex rtnl_mutex &root->kernfs_rwsem pool_lock#2 irq_context: 0 vlan_ioctl_mutex rtnl_mutex &root->kernfs_rwsem kernfs_idr_lock irq_context: 0 vlan_ioctl_mutex rtnl_mutex krc.lock irq_context: 0 dup_mmap_sem &mm->mmap_lock &mm->mmap_lock/1 &obj_hash[i].lock pool_lock irq_context: 0 vlan_ioctl_mutex rtnl_mutex netpoll_srcu irq_context: 0 vlan_ioctl_mutex rtnl_mutex &pn->hash_lock irq_context: 0 vlan_ioctl_mutex rtnl_mutex &dev->tx_global_lock irq_context: 0 vlan_ioctl_mutex rtnl_mutex &dev->tx_global_lock &vlan_netdev_xmit_lock_key irq_context: 0 vlan_ioctl_mutex rtnl_mutex &dev->tx_global_lock &obj_hash[i].lock irq_context: 0 vlan_ioctl_mutex rtnl_mutex rcu_state.exp_mutex rcu_node_0 irq_context: 0 vlan_ioctl_mutex rtnl_mutex rcu_state.exp_mutex &obj_hash[i].lock irq_context: 0 vlan_ioctl_mutex rtnl_mutex rcu_state.exp_mutex rcu_read_lock &pool->lock irq_context: 0 vlan_ioctl_mutex rtnl_mutex rcu_state.exp_mutex rcu_read_lock &pool->lock &obj_hash[i].lock irq_context: 0 vlan_ioctl_mutex rtnl_mutex rcu_state.exp_mutex rcu_read_lock &pool->lock &p->pi_lock irq_context: 0 vlan_ioctl_mutex rtnl_mutex rcu_state.exp_mutex rcu_read_lock &pool->lock &p->pi_lock &rq->__lock irq_context: 0 vlan_ioctl_mutex rtnl_mutex rcu_state.exp_mutex rcu_read_lock &pool->lock &p->pi_lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 vlan_ioctl_mutex rtnl_mutex rcu_state.exp_mutex &rq->__lock irq_context: 0 vlan_ioctl_mutex rtnl_mutex rcu_state.exp_mutex &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 vlan_ioctl_mutex rtnl_mutex &sch->q.lock irq_context: 0 vlan_ioctl_mutex rtnl_mutex &vlan_netdev_addr_lock_key/1 irq_context: 0 vlan_ioctl_mutex rtnl_mutex &vlan_netdev_addr_lock_key/1 _xmit_ETHER irq_context: 0 vlan_ioctl_mutex rtnl_mutex &vlan_netdev_addr_lock_key/1 _xmit_ETHER &obj_hash[i].lock irq_context: 0 vlan_ioctl_mutex rtnl_mutex &vlan_netdev_addr_lock_key/1 _xmit_ETHER pool_lock#2 irq_context: 0 vlan_ioctl_mutex rtnl_mutex &vlan_netdev_addr_lock_key/1 _xmit_ETHER krc.lock irq_context: 0 vlan_ioctl_mutex rtnl_mutex _xmit_ETHER &obj_hash[i].lock irq_context: 0 vlan_ioctl_mutex rtnl_mutex _xmit_ETHER pool_lock#2 irq_context: 0 vlan_ioctl_mutex rtnl_mutex _xmit_ETHER krc.lock irq_context: 0 vlan_ioctl_mutex rtnl_mutex __ip_vs_mutex irq_context: 0 vlan_ioctl_mutex rtnl_mutex __ip_vs_mutex &ipvs->dest_trash_lock irq_context: 0 vlan_ioctl_mutex rtnl_mutex &vlan_netdev_addr_lock_key/1 &obj_hash[i].lock irq_context: 0 vlan_ioctl_mutex rtnl_mutex &vlan_netdev_addr_lock_key/1 pool_lock#2 irq_context: 0 vlan_ioctl_mutex rtnl_mutex &vlan_netdev_addr_lock_key/1 krc.lock irq_context: 0 vlan_ioctl_mutex rtnl_mutex &im->lock irq_context: 0 vlan_ioctl_mutex rtnl_mutex fib_info_lock irq_context: 0 vlan_ioctl_mutex rtnl_mutex fib_info_lock &obj_hash[i].lock irq_context: 0 vlan_ioctl_mutex rtnl_mutex fib_info_lock pool_lock#2 irq_context: 0 vlan_ioctl_mutex rtnl_mutex class irq_context: 0 vlan_ioctl_mutex rtnl_mutex (&tbl->proxy_timer) irq_context: 0 vlan_ioctl_mutex rtnl_mutex flowtable_lock irq_context: 0 vlan_ioctl_mutex rtnl_mutex &dir->lock irq_context: 0 vlan_ioctl_mutex rtnl_mutex rcu_read_lock &pool->lock pool_lock#2 irq_context: 0 vlan_ioctl_mutex rtnl_mutex rcu_read_lock &tb->tb6_lock irq_context: 0 vlan_ioctl_mutex rtnl_mutex rcu_read_lock &tb->tb6_lock &net->ipv6.fib6_walker_lock irq_context: 0 vlan_ioctl_mutex rtnl_mutex rcu_read_lock &tb->tb6_lock &net->ipv6.fib6_walker_lock irq_context: 0 vlan_ioctl_mutex rtnl_mutex rcu_read_lock &tb->tb6_lock rt6_exception_lock irq_context: 0 vlan_ioctl_mutex rtnl_mutex rcu_read_lock &tb->tb6_lock &c->lock irq_context: 0 vlan_ioctl_mutex rtnl_mutex rcu_read_lock &tb->tb6_lock pool_lock#2 irq_context: 0 vlan_ioctl_mutex rtnl_mutex rcu_read_lock &tb->tb6_lock nl_table_lock irq_context: 0 &mm->mmap_lock &vma->vm_lock->lock &cfs_rq->removed.lock irq_context: 0 vlan_ioctl_mutex rtnl_mutex rcu_read_lock &tb->tb6_lock &obj_hash[i].lock irq_context: 0 &mm->mmap_lock &vma->vm_lock->lock &obj_hash[i].lock irq_context: 0 vlan_ioctl_mutex rtnl_mutex rcu_read_lock &tb->tb6_lock nl_table_wait.lock irq_context: 0 &mm->mmap_lock &vma->vm_lock->lock pool_lock#2 irq_context: 0 vlan_ioctl_mutex rtnl_mutex rcu_read_lock &tb->tb6_lock &obj_hash[i].lock pool_lock irq_context: 0 vlan_ioctl_mutex rtnl_mutex rcu_read_lock &tb->tb6_lock rcu_read_lock pool_lock#2 irq_context: 0 vlan_ioctl_mutex rtnl_mutex rcu_read_lock &tb->tb6_lock rcu_read_lock &data->fib_event_queue_lock irq_context: 0 vlan_ioctl_mutex rtnl_mutex rcu_read_lock &tb->tb6_lock rcu_read_lock rcu_read_lock &pool->lock irq_context: 0 vlan_ioctl_mutex rtnl_mutex rcu_read_lock &tb->tb6_lock rcu_read_lock rcu_read_lock &pool->lock &obj_hash[i].lock irq_context: 0 vlan_ioctl_mutex rtnl_mutex rcu_read_lock &tb->tb6_lock rcu_read_lock rcu_read_lock &pool->lock pool_lock#2 irq_context: 0 vlan_ioctl_mutex rtnl_mutex rcu_read_lock &tb->tb6_lock rcu_read_lock rcu_read_lock &pool->lock &p->pi_lock irq_context: 0 vlan_ioctl_mutex rtnl_mutex rcu_read_lock &tb->tb6_lock rcu_read_lock rcu_read_lock &pool->lock &p->pi_lock &rq->__lock irq_context: 0 vlan_ioctl_mutex rtnl_mutex rcu_read_lock &tb->tb6_lock rcu_read_lock rcu_read_lock &pool->lock &p->pi_lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 vlan_ioctl_mutex rtnl_mutex rcu_read_lock &tb->tb6_lock &n->list_lock irq_context: 0 vlan_ioctl_mutex rtnl_mutex rcu_read_lock &tb->tb6_lock &n->list_lock &c->lock irq_context: 0 vlan_ioctl_mutex rtnl_mutex &tbl->lock &n->lock irq_context: 0 vlan_ioctl_mutex rtnl_mutex &tbl->lock &c->lock irq_context: 0 vlan_ioctl_mutex rtnl_mutex &tbl->lock pool_lock#2 irq_context: 0 vlan_ioctl_mutex rtnl_mutex &tbl->lock &n->lock irq_context: 0 vlan_ioctl_mutex rtnl_mutex &tbl->lock &n->lock &____s->seqcount#9 irq_context: 0 vlan_ioctl_mutex rtnl_mutex &tbl->lock nl_table_lock irq_context: 0 vlan_ioctl_mutex rtnl_mutex &tbl->lock &obj_hash[i].lock irq_context: 0 vlan_ioctl_mutex rtnl_mutex &tbl->lock nl_table_wait.lock irq_context: 0 vlan_ioctl_mutex rtnl_mutex &tbl->lock rcu_read_lock lock#8 irq_context: 0 vlan_ioctl_mutex rtnl_mutex &tbl->lock rcu_read_lock id_table_lock irq_context: 0 vlan_ioctl_mutex rtnl_mutex &tbl->lock &dir->lock#2 irq_context: 0 vlan_ioctl_mutex rtnl_mutex &tbl->lock krc.lock irq_context: 0 vlan_ioctl_mutex rtnl_mutex &net->ipv6.addrconf_hash_lock irq_context: 0 vlan_ioctl_mutex rtnl_mutex &net->ipv6.addrconf_hash_lock &obj_hash[i].lock irq_context: 0 vlan_ioctl_mutex rtnl_mutex &ndev->lock irq_context: 0 vlan_ioctl_mutex rtnl_mutex &ndev->lock &obj_hash[i].lock irq_context: 0 vlan_ioctl_mutex rtnl_mutex &ndev->lock &base->lock irq_context: 0 vlan_ioctl_mutex rtnl_mutex &ndev->lock &base->lock &obj_hash[i].lock irq_context: 0 vlan_ioctl_mutex rtnl_mutex &ifa->lock irq_context: 0 vlan_ioctl_mutex rtnl_mutex &idev->mc_lock &vlan_netdev_addr_lock_key/1 irq_context: 0 vlan_ioctl_mutex rtnl_mutex &idev->mc_lock &vlan_netdev_addr_lock_key/1 &obj_hash[i].lock irq_context: 0 vlan_ioctl_mutex rtnl_mutex &idev->mc_lock &vlan_netdev_addr_lock_key/1 pool_lock#2 irq_context: 0 vlan_ioctl_mutex rtnl_mutex &idev->mc_lock &vlan_netdev_addr_lock_key/1 krc.lock irq_context: 0 vlan_ioctl_mutex rtnl_mutex &idev->mc_lock &base->lock irq_context: 0 vlan_ioctl_mutex rtnl_mutex &idev->mc_lock &base->lock &obj_hash[i].lock irq_context: 0 vlan_ioctl_mutex rtnl_mutex &tb->tb6_lock irq_context: 0 vlan_ioctl_mutex rtnl_mutex rcu_read_lock &c->lock irq_context: 0 vlan_ioctl_mutex rtnl_mutex rcu_read_lock &obj_hash[i].lock irq_context: 0 vlan_ioctl_mutex rtnl_mutex rcu_read_lock rcu_read_lock &pool->lock/1 irq_context: 0 vlan_ioctl_mutex rtnl_mutex rcu_read_lock rcu_read_lock &pool->lock/1 &obj_hash[i].lock irq_context: 0 vlan_ioctl_mutex rtnl_mutex rcu_read_lock rcu_read_lock &pool->lock/1 pool_lock#2 irq_context: 0 vlan_ioctl_mutex rtnl_mutex rcu_read_lock rcu_read_lock &pool->lock/1 &p->pi_lock irq_context: 0 vlan_ioctl_mutex rtnl_mutex rcu_read_lock rcu_read_lock &pool->lock/1 &p->pi_lock &rq->__lock irq_context: 0 vlan_ioctl_mutex rtnl_mutex rcu_read_lock rcu_read_lock &pool->lock/1 &p->pi_lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 vlan_ioctl_mutex rtnl_mutex rcu_read_lock batched_entropy_u8.lock irq_context: 0 vlan_ioctl_mutex rtnl_mutex rcu_read_lock kfence_freelist_lock irq_context: 0 vlan_ioctl_mutex rtnl_mutex rcu_read_lock &dir->lock irq_context: 0 vlan_ioctl_mutex rtnl_mutex rcu_read_lock rcu_read_lock &pool->lock irq_context: 0 vlan_ioctl_mutex rtnl_mutex rcu_read_lock rcu_read_lock &pool->lock &obj_hash[i].lock irq_context: 0 vlan_ioctl_mutex rtnl_mutex rcu_read_lock rcu_read_lock &pool->lock &p->pi_lock irq_context: 0 vlan_ioctl_mutex rtnl_mutex rcu_read_lock rcu_read_lock &pool->lock &p->pi_lock &rq->__lock irq_context: 0 vlan_ioctl_mutex rtnl_mutex rcu_read_lock rcu_read_lock &pool->lock &p->pi_lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 vlan_ioctl_mutex rtnl_mutex rcu_read_lock &net->sctp.local_addr_lock irq_context: 0 vlan_ioctl_mutex rtnl_mutex rcu_read_lock &net->sctp.local_addr_lock &net->sctp.addr_wq_lock irq_context: 0 vlan_ioctl_mutex rtnl_mutex rcu_read_lock &net->sctp.local_addr_lock &net->sctp.addr_wq_lock pool_lock#2 irq_context: 0 vlan_ioctl_mutex rtnl_mutex rcu_read_lock &net->sctp.local_addr_lock &net->sctp.addr_wq_lock &obj_hash[i].lock irq_context: 0 vlan_ioctl_mutex rtnl_mutex rcu_read_lock &net->sctp.local_addr_lock &net->sctp.addr_wq_lock &base->lock irq_context: 0 vlan_ioctl_mutex rtnl_mutex rcu_read_lock &net->sctp.local_addr_lock &net->sctp.addr_wq_lock &base->lock &obj_hash[i].lock irq_context: 0 vlan_ioctl_mutex rtnl_mutex rcu_read_lock krc.lock irq_context: 0 vlan_ioctl_mutex rtnl_mutex rcu_read_lock &rcu_state.gp_wq irq_context: 0 vlan_ioctl_mutex rtnl_mutex rcu_read_lock &rcu_state.gp_wq &p->pi_lock irq_context: 0 vlan_ioctl_mutex rtnl_mutex rcu_read_lock &rcu_state.gp_wq &p->pi_lock &rq->__lock irq_context: 0 vlan_ioctl_mutex rtnl_mutex rcu_read_lock &rcu_state.gp_wq &p->pi_lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 vlan_ioctl_mutex rtnl_mutex rcu_read_lock &n->list_lock irq_context: 0 vlan_ioctl_mutex rtnl_mutex rcu_read_lock &n->list_lock &c->lock irq_context: 0 vlan_ioctl_mutex rtnl_mutex rcu_read_lock rcu_read_lock &pool->lock pool_lock#2 irq_context: 0 vlan_ioctl_mutex rtnl_mutex rcu_state.exp_mutex &rnp->exp_wq[0] irq_context: 0 (wq_completion)events (work_completion)(&nlk->work) &rq->__lock irq_context: 0 (wq_completion)events (work_completion)(&nlk->work) &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 vlan_ioctl_mutex vlan_ioctl_mutex.wait_lock irq_context: 0 vlan_ioctl_mutex rtnl_mutex &idev->mc_query_lock irq_context: 0 vlan_ioctl_mutex rtnl_mutex &idev->mc_query_lock &obj_hash[i].lock irq_context: 0 vlan_ioctl_mutex rtnl_mutex (work_completion)(&(&idev->mc_report_work)->work) irq_context: 0 vlan_ioctl_mutex rtnl_mutex &net->ipv6.fib6_gc_lock rcu_read_lock &tb->tb6_lock irq_context: 0 vlan_ioctl_mutex rtnl_mutex &net->ipv6.fib6_gc_lock &obj_hash[i].lock irq_context: 0 vlan_ioctl_mutex rtnl_mutex dev_base_lock &xa->xa_lock#3 irq_context: 0 vlan_ioctl_mutex rtnl_mutex cpu_hotplug_lock irq_context: 0 vlan_ioctl_mutex rtnl_mutex cpu_hotplug_lock &list->lock#5 irq_context: 0 vlan_ioctl_mutex rtnl_mutex rcu_state.exp_mutex &rnp->exp_wq[1] irq_context: 0 sb_writers#4 &type->i_mutex_dir_key#3/1 tomoyo_ss &rcu_state.expedited_wq irq_context: 0 sb_writers#4 &type->i_mutex_dir_key#3/1 tomoyo_ss &rcu_state.expedited_wq &p->pi_lock irq_context: 0 sb_writers#4 &type->i_mutex_dir_key#3/1 tomoyo_ss &rcu_state.expedited_wq &p->pi_lock &rq->__lock irq_context: 0 sb_writers#4 &type->i_mutex_dir_key#3/1 tomoyo_ss &rcu_state.expedited_wq &p->pi_lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 vlan_ioctl_mutex rtnl_mutex bpf_devs_lock irq_context: 0 vlan_ioctl_mutex rtnl_mutex &hwstats->hwsdev_list_lock irq_context: 0 vlan_ioctl_mutex rtnl_mutex &in_dev->mc_tomb_lock irq_context: 0 vlan_ioctl_mutex rtnl_mutex (inetaddr_chain).rwsem irq_context: 0 vlan_ioctl_mutex rtnl_mutex (inetaddr_chain).rwsem pool_lock#2 irq_context: 0 vlan_ioctl_mutex rtnl_mutex (inetaddr_chain).rwsem &obj_hash[i].lock irq_context: 0 vlan_ioctl_mutex rtnl_mutex (inetaddr_chain).rwsem rcu_read_lock &pool->lock/1 irq_context: 0 vlan_ioctl_mutex rtnl_mutex (inetaddr_chain).rwsem rcu_read_lock &pool->lock/1 &obj_hash[i].lock irq_context: 0 vlan_ioctl_mutex rtnl_mutex (inetaddr_chain).rwsem rcu_read_lock &pool->lock/1 pool_lock#2 irq_context: 0 vlan_ioctl_mutex rtnl_mutex (inetaddr_chain).rwsem rcu_read_lock &pool->lock/1 &p->pi_lock irq_context: 0 vlan_ioctl_mutex rtnl_mutex (inetaddr_chain).rwsem rcu_read_lock &pool->lock/1 &p->pi_lock &rq->__lock irq_context: 0 vlan_ioctl_mutex rtnl_mutex (inetaddr_chain).rwsem rcu_read_lock &pool->lock/1 &p->pi_lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 vlan_ioctl_mutex rtnl_mutex (inetaddr_chain).rwsem &rq->__lock irq_context: 0 vlan_ioctl_mutex rtnl_mutex (inetaddr_chain).rwsem &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 vlan_ioctl_mutex rtnl_mutex (inetaddr_chain).rwsem rcu_read_lock &c->lock irq_context: 0 vlan_ioctl_mutex rtnl_mutex (inetaddr_chain).rwsem rcu_read_lock pool_lock#2 irq_context: 0 vlan_ioctl_mutex rtnl_mutex (inetaddr_chain).rwsem rcu_read_lock &data->fib_event_queue_lock irq_context: 0 vlan_ioctl_mutex rtnl_mutex (inetaddr_chain).rwsem rcu_read_lock rcu_read_lock &pool->lock irq_context: 0 vlan_ioctl_mutex rtnl_mutex (inetaddr_chain).rwsem rcu_read_lock rcu_read_lock &pool->lock &obj_hash[i].lock irq_context: 0 vlan_ioctl_mutex rtnl_mutex (inetaddr_chain).rwsem rcu_read_lock rcu_read_lock &pool->lock pool_lock#2 irq_context: 0 vlan_ioctl_mutex rtnl_mutex (inetaddr_chain).rwsem rcu_read_lock rcu_read_lock &pool->lock &p->pi_lock irq_context: 0 vlan_ioctl_mutex rtnl_mutex (inetaddr_chain).rwsem rcu_read_lock rcu_read_lock &pool->lock &p->pi_lock &rq->__lock irq_context: 0 vlan_ioctl_mutex rtnl_mutex (inetaddr_chain).rwsem rcu_read_lock rcu_read_lock &pool->lock &p->pi_lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 vlan_ioctl_mutex rtnl_mutex (inetaddr_chain).rwsem rcu_read_lock rcu_node_0 irq_context: 0 vlan_ioctl_mutex rtnl_mutex (inetaddr_chain).rwsem rcu_read_lock &rq->__lock irq_context: 0 vlan_ioctl_mutex rtnl_mutex (inetaddr_chain).rwsem rcu_read_lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 &iint->mutex ima_extend_list_mutex rcu_node_0 irq_context: 0 &iint->mutex ima_extend_list_mutex rcu_read_lock rcu_node_0 irq_context: 0 vlan_ioctl_mutex rtnl_mutex (inetaddr_chain).rwsem rcu_read_lock &rcu_state.gp_wq irq_context: 0 vlan_ioctl_mutex rtnl_mutex (inetaddr_chain).rwsem rcu_read_lock &rcu_state.gp_wq &p->pi_lock irq_context: 0 vlan_ioctl_mutex rtnl_mutex (inetaddr_chain).rwsem fs_reclaim irq_context: 0 vlan_ioctl_mutex rtnl_mutex (inetaddr_chain).rwsem fs_reclaim mmu_notifier_invalidate_range_start irq_context: 0 vlan_ioctl_mutex rtnl_mutex (inetaddr_chain).rwsem &c->lock irq_context: 0 vlan_ioctl_mutex rtnl_mutex (inetaddr_chain).rwsem &n->list_lock irq_context: 0 vlan_ioctl_mutex rtnl_mutex (inetaddr_chain).rwsem &n->list_lock &c->lock irq_context: 0 vlan_ioctl_mutex rtnl_mutex (inetaddr_chain).rwsem &____s->seqcount irq_context: 0 vlan_ioctl_mutex rtnl_mutex (inetaddr_chain).rwsem nl_table_lock irq_context: 0 vlan_ioctl_mutex rtnl_mutex (inetaddr_chain).rwsem nl_table_wait.lock irq_context: 0 vlan_ioctl_mutex rtnl_mutex (inetaddr_chain).rwsem fib_info_lock irq_context: 0 vlan_ioctl_mutex rtnl_mutex (inetaddr_chain).rwsem fib_info_lock &obj_hash[i].lock irq_context: 0 vlan_ioctl_mutex rtnl_mutex (inetaddr_chain).rwsem fib_info_lock pool_lock#2 irq_context: 0 vlan_ioctl_mutex rtnl_mutex (inetaddr_chain).rwsem &tbl->lock irq_context: 0 vlan_ioctl_mutex rtnl_mutex (inetaddr_chain).rwsem class irq_context: 0 vlan_ioctl_mutex rtnl_mutex (inetaddr_chain).rwsem (&tbl->proxy_timer) irq_context: 0 vlan_ioctl_mutex rtnl_mutex (inetaddr_chain).rwsem &base->lock irq_context: 0 vlan_ioctl_mutex rtnl_mutex (inetaddr_chain).rwsem &net->sctp.local_addr_lock irq_context: 0 vlan_ioctl_mutex rtnl_mutex (inetaddr_chain).rwsem &net->sctp.local_addr_lock &net->sctp.addr_wq_lock irq_context: 0 vlan_ioctl_mutex rtnl_mutex (inetaddr_chain).rwsem &net->sctp.local_addr_lock &net->sctp.addr_wq_lock pool_lock#2 irq_context: 0 vlan_ioctl_mutex rtnl_mutex (inetaddr_chain).rwsem krc.lock irq_context: 0 vlan_ioctl_mutex rtnl_mutex sysctl_lock &obj_hash[i].lock irq_context: 0 vlan_ioctl_mutex rtnl_mutex sysctl_lock pool_lock#2 irq_context: 0 vlan_ioctl_mutex rtnl_mutex sysctl_lock krc.lock irq_context: 0 vlan_ioctl_mutex rtnl_mutex sysctl_lock krc.lock &obj_hash[i].lock irq_context: 0 vlan_ioctl_mutex rtnl_mutex sysctl_lock krc.lock &base->lock irq_context: 0 vlan_ioctl_mutex rtnl_mutex sysctl_lock krc.lock &base->lock &obj_hash[i].lock irq_context: 0 vlan_ioctl_mutex rtnl_mutex &ul->lock irq_context: 0 vlan_ioctl_mutex rtnl_mutex &net->xdp.lock irq_context: 0 vlan_ioctl_mutex rtnl_mutex mirred_list_lock irq_context: 0 vlan_ioctl_mutex rtnl_mutex &nft_net->commit_mutex irq_context: 0 vlan_ioctl_mutex rtnl_mutex &ent->pde_unload_lock irq_context: 0 vlan_ioctl_mutex rtnl_mutex rcu_state.exp_mutex &rnp->exp_wq[2] irq_context: 0 vlan_ioctl_mutex rtnl_mutex &idev->mc_lock krc.lock irq_context: 0 vlan_ioctl_mutex rtnl_mutex &idev->mc_report_lock irq_context: 0 vlan_ioctl_mutex rtnl_mutex batched_entropy_u8.lock irq_context: 0 vlan_ioctl_mutex rtnl_mutex kfence_freelist_lock irq_context: 0 vlan_ioctl_mutex rtnl_mutex &meta->lock irq_context: 0 vlan_ioctl_mutex rtnl_mutex &pnn->pndevs.lock irq_context: 0 vlan_ioctl_mutex rtnl_mutex &pnn->routes.lock irq_context: 0 vlan_ioctl_mutex rtnl_mutex target_list_lock irq_context: 0 vlan_ioctl_mutex rtnl_mutex kernfs_idr_lock irq_context: 0 vlan_ioctl_mutex rtnl_mutex dev_pm_qos_sysfs_mtx irq_context: 0 vlan_ioctl_mutex rtnl_mutex dev_pm_qos_sysfs_mtx &root->kernfs_rwsem irq_context: 0 vlan_ioctl_mutex rtnl_mutex dev_pm_qos_sysfs_mtx &root->kernfs_rwsem irq_context: 0 vlan_ioctl_mutex rtnl_mutex dev_pm_qos_sysfs_mtx dev_pm_qos_mtx irq_context: 0 vlan_ioctl_mutex rtnl_mutex subsys mutex#17 &k->k_lock klist_remove_lock irq_context: 0 vlan_ioctl_mutex rtnl_mutex subsys mutex#17 &rq->__lock irq_context: 0 vlan_ioctl_mutex rtnl_mutex subsys mutex#17 &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 vlan_ioctl_mutex rtnl_mutex &root->kernfs_rwsem kernfs_idr_lock &obj_hash[i].lock irq_context: 0 vlan_ioctl_mutex rtnl_mutex &root->kernfs_rwsem kernfs_idr_lock pool_lock#2 irq_context: 0 vlan_ioctl_mutex rtnl_mutex deferred_probe_mutex irq_context: 0 vlan_ioctl_mutex rtnl_mutex device_links_lock irq_context: 0 vlan_ioctl_mutex rtnl_mutex mmu_notifier_invalidate_range_start irq_context: 0 vlan_ioctl_mutex rtnl_mutex uevent_sock_mutex mmu_notifier_invalidate_range_start irq_context: 0 vlan_ioctl_mutex rtnl_mutex cpu_hotplug_lock xps_map_mutex irq_context: 0 vlan_ioctl_mutex rtnl_mutex rcu_state.exp_mutex &rnp->exp_wq[3] irq_context: 0 vlan_ioctl_mutex rcu_state.barrier_mutex irq_context: 0 vlan_ioctl_mutex rcu_state.barrier_mutex rcu_state.barrier_lock irq_context: 0 vlan_ioctl_mutex rcu_state.barrier_mutex rcu_state.barrier_lock &obj_hash[i].lock irq_context: 0 vlan_ioctl_mutex rcu_state.barrier_mutex &x->wait#24 irq_context: 0 vlan_ioctl_mutex rcu_state.barrier_mutex &rq->__lock irq_context: 0 vlan_ioctl_mutex rcu_state.barrier_mutex &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 rtnl_mutex sk_lock-AF_UNSPEC irq_context: 0 rtnl_mutex sk_lock-AF_UNSPEC slock-AF_UNSPEC irq_context: 0 rtnl_mutex sk_lock-AF_UNSPEC fs_reclaim irq_context: 0 rtnl_mutex sk_lock-AF_UNSPEC fs_reclaim mmu_notifier_invalidate_range_start irq_context: 0 rtnl_mutex sk_lock-AF_UNSPEC pool_lock#2 irq_context: 0 rtnl_mutex sk_lock-AF_UNSPEC free_vmap_area_lock irq_context: 0 rtnl_mutex sk_lock-AF_UNSPEC vmap_area_lock irq_context: 0 rtnl_mutex sk_lock-AF_UNSPEC &____s->seqcount irq_context: 0 rtnl_mutex sk_lock-AF_UNSPEC stock_lock irq_context: 0 rtnl_mutex sk_lock-AF_UNSPEC &c->lock irq_context: 0 rtnl_mutex sk_lock-AF_UNSPEC pcpu_alloc_mutex irq_context: 0 rtnl_mutex sk_lock-AF_UNSPEC pcpu_alloc_mutex pcpu_lock irq_context: 0 rtnl_mutex sk_lock-AF_UNSPEC &mm->mmap_lock irq_context: 0 rtnl_mutex sk_lock-AF_UNSPEC &obj_hash[i].lock irq_context: 0 rtnl_mutex sk_lock-AF_UNSPEC pack_mutex irq_context: 0 rtnl_mutex sk_lock-AF_UNSPEC batched_entropy_u32.lock irq_context: 0 rtnl_mutex sk_lock-AF_UNSPEC text_mutex irq_context: 0 rtnl_mutex sk_lock-AF_UNSPEC text_mutex ptlock_ptr(page)#2 irq_context: 0 rtnl_mutex sk_lock-AF_UNSPEC &fp->aux->used_maps_mutex irq_context: 0 rtnl_mutex slock-AF_UNSPEC irq_context: 0 vlan_ioctl_mutex dev_base_lock irq_context: 0 vlan_ioctl_mutex lweventlist_lock irq_context: 0 vlan_ioctl_mutex pcpu_lock irq_context: 0 vlan_ioctl_mutex pool_lock#2 irq_context: 0 vlan_ioctl_mutex &dir->lock#2 irq_context: 0 vlan_ioctl_mutex &obj_hash[i].lock irq_context: 0 vlan_ioctl_mutex krc.lock irq_context: 0 vlan_ioctl_mutex &dir->lock#2 &obj_hash[i].lock irq_context: 0 vlan_ioctl_mutex &dir->lock#2 pool_lock#2 irq_context: 0 vlan_ioctl_mutex netdev_unregistering_wq.lock irq_context: 0 vlan_ioctl_mutex.wait_lock irq_context: 0 vlan_ioctl_mutex rtnl_mutex _xmit_ETHER (console_sem).lock irq_context: 0 vlan_ioctl_mutex rtnl_mutex _xmit_ETHER console_lock console_srcu console_owner_lock irq_context: 0 vlan_ioctl_mutex rtnl_mutex _xmit_ETHER console_lock console_srcu console_owner irq_context: 0 vlan_ioctl_mutex rtnl_mutex _xmit_ETHER console_lock console_srcu console_owner &port_lock_key irq_context: 0 vlan_ioctl_mutex rtnl_mutex _xmit_ETHER console_lock console_srcu console_owner console_owner_lock irq_context: 0 vlan_ioctl_mutex rtnl_mutex &app->lock irq_context: 0 vlan_ioctl_mutex rtnl_mutex (&app->join_timer) irq_context: 0 vlan_ioctl_mutex rtnl_mutex (&app->periodic_timer) irq_context: 0 vlan_ioctl_mutex rtnl_mutex &list->lock#11 irq_context: 0 vlan_ioctl_mutex rtnl_mutex (&app->join_timer)#2 irq_context: 0 vlan_ioctl_mutex rtnl_mutex &app->lock#2 irq_context: 0 vlan_ioctl_mutex rtnl_mutex &list->lock#12 irq_context: 0 dup_mmap_sem &mm->mmap_lock &mm->mmap_lock/1 fs_reclaim mmu_notifier_invalidate_range_start &cfs_rq->removed.lock irq_context: 0 dup_mmap_sem &mm->mmap_lock &mm->mmap_lock/1 fs_reclaim mmu_notifier_invalidate_range_start &obj_hash[i].lock irq_context: 0 vlan_ioctl_mutex lweventlist_lock pool_lock#2 irq_context: 0 vlan_ioctl_mutex lweventlist_lock &dir->lock#2 irq_context: 0 pcpu_alloc_mutex &rq->__lock &cfs_rq->removed.lock irq_context: 0 (wq_completion)ext4-rsv-conversion (work_completion)(&ei->i_rsv_conversion_work) jbd2_handle &ei->i_data_sem remove_cache_srcu rcu_read_lock &rq->__lock irq_context: 0 (wq_completion)ext4-rsv-conversion (work_completion)(&ei->i_rsv_conversion_work) jbd2_handle &ei->i_data_sem remove_cache_srcu rcu_read_lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: softirq (&n->timer) rcu_read_lock_bh rcu_read_lock tk_core.seq.seqcount irq_context: softirq (&n->timer) rcu_read_lock_bh rcu_read_lock &list->lock#5 irq_context: 0 sb_writers#4 &sb->s_type->i_mutex_key#8 mapping.invalidate_lock &mapping->private_lock rcu_read_lock rcu_read_lock rcu_read_lock per_cpu_ptr(&cgroup_rstat_cpu_lock, cpu) irq_context: 0 sb_writers#8 tomoyo_ss remove_cache_srcu rcu_read_lock rcu_node_0 irq_context: 0 sb_writers#8 tomoyo_ss remove_cache_srcu rcu_read_lock &rq->__lock irq_context: 0 sb_writers#8 tomoyo_ss remove_cache_srcu rcu_read_lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 sb_writers#4 &type->i_mutex_dir_key#3/1 jbd2_handle &ei->i_data_sem rcu_read_lock &cfs_rq->removed.lock irq_context: 0 sb_writers#4 &type->i_mutex_dir_key#3/1 jbd2_handle &ei->i_data_sem rcu_read_lock &obj_hash[i].lock irq_context: 0 sb_writers#4 &type->i_mutex_dir_key#3/1 jbd2_handle &ei->i_data_sem rcu_read_lock &rcu_state.gp_wq irq_context: 0 sb_writers#4 &type->i_mutex_dir_key#3/1 jbd2_handle &ei->i_data_sem rcu_read_lock &rcu_state.gp_wq &p->pi_lock irq_context: 0 sb_writers#4 &type->i_mutex_dir_key#3/1 jbd2_handle &ei->i_data_sem rcu_read_lock &rcu_state.gp_wq &p->pi_lock &rq->__lock irq_context: 0 sb_writers#4 &type->i_mutex_dir_key#3/1 jbd2_handle &ei->i_data_sem rcu_read_lock &rcu_state.gp_wq &p->pi_lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 sb_writers#4 &type->i_mutex_dir_key#3 &ei->i_es_lock key#2 irq_context: 0 sb_writers#4 &type->i_mutex_dir_key#3/1 jbd2_handle &ei->i_data_sem rcu_read_lock &rcu_state.expedited_wq irq_context: 0 sb_writers#4 &type->i_mutex_dir_key#3/1 jbd2_handle &ei->i_data_sem rcu_read_lock &rcu_state.expedited_wq &p->pi_lock irq_context: 0 sb_writers#4 &type->i_mutex_dir_key#3/1 jbd2_handle &ei->i_data_sem rcu_read_lock &rcu_state.expedited_wq &p->pi_lock &rq->__lock irq_context: 0 sb_writers#4 &type->i_mutex_dir_key#3/1 jbd2_handle &ei->i_data_sem rcu_read_lock &rcu_state.expedited_wq &p->pi_lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 sk_lock-AF_INET &rcu_state.expedited_wq &p->pi_lock irq_context: 0 &f->f_pos_lock sb_writers#4 &sb->s_type->i_mutex_key#8 &mm->mmap_lock &rcu_state.expedited_wq irq_context: 0 &f->f_pos_lock sb_writers#4 &sb->s_type->i_mutex_key#8 &mm->mmap_lock &rcu_state.expedited_wq &p->pi_lock irq_context: 0 &f->f_pos_lock sb_writers#4 &sb->s_type->i_mutex_key#8 &mm->mmap_lock &rcu_state.expedited_wq &p->pi_lock &rq->__lock irq_context: 0 &f->f_pos_lock sb_writers#4 &sb->s_type->i_mutex_key#8 &mm->mmap_lock &rcu_state.expedited_wq &p->pi_lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 nf_sockopt_mutex rcu_node_0 irq_context: 0 dup_mmap_sem &mm->mmap_lock &mm->mmap_lock/1 fs_reclaim mmu_notifier_invalidate_range_start &rq->__lock &cfs_rq->removed.lock irq_context: 0 &ei->i_data_sem rcu_read_lock rcu_node_0 irq_context: 0 &ei->i_data_sem rcu_read_lock &rq->__lock irq_context: 0 &ei->i_data_sem rcu_read_lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 &mm->mmap_lock sb_pagefaults mapping.invalidate_lock &cfs_rq->removed.lock irq_context: 0 sb_writers#4 tomoyo_ss rcu_read_lock rcu_read_lock rcu_node_0 irq_context: 0 sb_writers#4 tomoyo_ss rcu_read_lock rcu_read_lock &rq->__lock irq_context: 0 sb_writers#4 tomoyo_ss rcu_read_lock rcu_read_lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 sb_writers#4 &sb->s_type->i_mutex_key#8 mapping.invalidate_lock &mapping->i_mmap_rwsem &rq->__lock &cfs_rq->removed.lock irq_context: 0 tomoyo_ss &obj_hash[i].lock pool_lock irq_context: softirq &(&conn->disc_work)->timer irq_context: softirq &(&conn->disc_work)->timer rcu_read_lock &pool->lock/1 irq_context: softirq &(&conn->disc_work)->timer rcu_read_lock &pool->lock/1 &obj_hash[i].lock irq_context: softirq &(&conn->disc_work)->timer rcu_read_lock &pool->lock/1 &p->pi_lock irq_context: softirq &(&conn->disc_work)->timer rcu_read_lock &pool->lock/1 &p->pi_lock &rq->__lock irq_context: softirq &(&conn->disc_work)->timer rcu_read_lock &pool->lock/1 &p->pi_lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)hci3#2 (work_completion)(&(&conn->disc_work)->work) irq_context: 0 (wq_completion)hci3#2 (work_completion)(&(&conn->disc_work)->work) &hdev->unregister_lock irq_context: 0 (wq_completion)hci3#2 (work_completion)(&(&conn->disc_work)->work) &hdev->unregister_lock fs_reclaim irq_context: 0 (wq_completion)hci3#2 (work_completion)(&(&conn->disc_work)->work) &hdev->unregister_lock fs_reclaim mmu_notifier_invalidate_range_start irq_context: 0 (wq_completion)hci3#2 (work_completion)(&(&conn->disc_work)->work) &hdev->unregister_lock &hdev->cmd_sync_work_lock irq_context: 0 (wq_completion)hci3#2 (work_completion)(&(&conn->disc_work)->work) &hdev->unregister_lock rcu_read_lock &pool->lock/1 irq_context: 0 (wq_completion)hci3#2 (work_completion)(&(&conn->disc_work)->work) &hdev->unregister_lock rcu_read_lock &pool->lock/1 &obj_hash[i].lock irq_context: 0 (wq_completion)hci3#2 (work_completion)(&(&conn->disc_work)->work) &hdev->unregister_lock rcu_read_lock &pool->lock/1 &p->pi_lock irq_context: 0 (wq_completion)hci1#2 (work_completion)(&(&conn->disc_work)->work) irq_context: 0 (wq_completion)hci1#2 (work_completion)(&(&conn->disc_work)->work) &hdev->unregister_lock irq_context: 0 (wq_completion)hci1#2 (work_completion)(&(&conn->disc_work)->work) &hdev->unregister_lock fs_reclaim irq_context: 0 (wq_completion)hci1#2 (work_completion)(&(&conn->disc_work)->work) &hdev->unregister_lock fs_reclaim mmu_notifier_invalidate_range_start irq_context: 0 (wq_completion)hci1#2 (work_completion)(&(&conn->disc_work)->work) &hdev->unregister_lock fs_reclaim mmu_notifier_invalidate_range_start &rq->__lock irq_context: 0 (wq_completion)hci1#2 (work_completion)(&(&conn->disc_work)->work) &hdev->unregister_lock &hdev->cmd_sync_work_lock irq_context: 0 (wq_completion)hci1#2 (work_completion)(&(&conn->disc_work)->work) &hdev->unregister_lock rcu_read_lock &pool->lock/1 irq_context: 0 (wq_completion)hci1#2 (work_completion)(&(&conn->disc_work)->work) &hdev->unregister_lock rcu_read_lock &pool->lock/1 &obj_hash[i].lock irq_context: 0 (wq_completion)hci1#2 (work_completion)(&(&conn->disc_work)->work) &hdev->unregister_lock rcu_read_lock &pool->lock/1 &p->pi_lock irq_context: 0 (wq_completion)hci2#2 (work_completion)(&(&conn->disc_work)->work) irq_context: 0 (wq_completion)hci2#2 (work_completion)(&(&conn->disc_work)->work) &hdev->unregister_lock irq_context: 0 (wq_completion)hci2#2 (work_completion)(&(&conn->disc_work)->work) &hdev->unregister_lock fs_reclaim irq_context: 0 (wq_completion)hci2#2 (work_completion)(&(&conn->disc_work)->work) &hdev->unregister_lock fs_reclaim mmu_notifier_invalidate_range_start irq_context: 0 (wq_completion)hci2#2 (work_completion)(&(&conn->disc_work)->work) &hdev->unregister_lock &hdev->cmd_sync_work_lock irq_context: 0 (wq_completion)hci2#2 (work_completion)(&(&conn->disc_work)->work) &hdev->unregister_lock rcu_read_lock &pool->lock/1 irq_context: 0 (wq_completion)hci2#2 (work_completion)(&(&conn->disc_work)->work) &hdev->unregister_lock rcu_read_lock &pool->lock/1 &obj_hash[i].lock irq_context: 0 (wq_completion)hci2#2 (work_completion)(&(&conn->disc_work)->work) &hdev->unregister_lock rcu_read_lock &pool->lock/1 &p->pi_lock irq_context: 0 (wq_completion)hci0#2 (work_completion)(&(&conn->disc_work)->work) irq_context: 0 (wq_completion)hci0#2 (work_completion)(&(&conn->disc_work)->work) &hdev->unregister_lock irq_context: 0 (wq_completion)hci0#2 (work_completion)(&(&conn->disc_work)->work) &hdev->unregister_lock fs_reclaim irq_context: 0 (wq_completion)hci0#2 (work_completion)(&(&conn->disc_work)->work) &hdev->unregister_lock fs_reclaim mmu_notifier_invalidate_range_start irq_context: 0 (wq_completion)hci0#2 (work_completion)(&(&conn->disc_work)->work) &hdev->unregister_lock &hdev->cmd_sync_work_lock irq_context: 0 (wq_completion)hci0#2 (work_completion)(&(&conn->disc_work)->work) &hdev->unregister_lock rcu_read_lock &pool->lock/1 irq_context: 0 (wq_completion)hci0#2 (work_completion)(&(&conn->disc_work)->work) &hdev->unregister_lock rcu_read_lock &pool->lock/1 &obj_hash[i].lock irq_context: 0 (wq_completion)hci0#2 (work_completion)(&(&conn->disc_work)->work) &hdev->unregister_lock rcu_read_lock &pool->lock/1 &p->pi_lock irq_context: 0 (wq_completion)hci3 (work_completion)(&hdev->cmd_sync_work) &hdev->req_lock pool_lock#2 irq_context: 0 (wq_completion)hci3 (work_completion)(&hdev->cmd_sync_work) &hdev->req_lock &list->lock#8 irq_context: 0 (wq_completion)hci3 (work_completion)(&hdev->cmd_sync_work) &hdev->req_lock &list->lock#7 irq_context: 0 (wq_completion)hci3 (work_completion)(&hdev->cmd_sync_work) &hdev->req_lock rcu_read_lock &pool->lock/1 irq_context: 0 (wq_completion)hci3 (work_completion)(&hdev->cmd_sync_work) &hdev->req_lock rcu_read_lock &pool->lock/1 &obj_hash[i].lock irq_context: 0 (wq_completion)hci3 (work_completion)(&hdev->cmd_sync_work) &hdev->req_lock rcu_read_lock &pool->lock/1 &p->pi_lock irq_context: 0 (wq_completion)hci3 (work_completion)(&hdev->cmd_sync_work) &hdev->req_lock &hdev->req_wait_q irq_context: 0 (wq_completion)hci3 (work_completion)(&hdev->cmd_sync_work) &hdev->req_lock &obj_hash[i].lock irq_context: 0 (wq_completion)hci3 (work_completion)(&hdev->cmd_sync_work) &hdev->req_lock &base->lock irq_context: 0 (wq_completion)hci3 (work_completion)(&hdev->cmd_sync_work) &hdev->req_lock &base->lock &obj_hash[i].lock irq_context: 0 (wq_completion)hci3 (work_completion)(&hdev->cmd_sync_work) &hdev->req_lock &rq->__lock irq_context: 0 (wq_completion)hci3 (work_completion)(&hdev->cmd_sync_work) &hdev->req_lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: softirq (&ndev->rs_timer) rcu_read_lock rcu_read_lock rcu_read_lock_bh noop_qdisc.q.lock irq_context: 0 (wq_completion)hci5 (work_completion)(&hdev->cmd_sync_work) &hdev->req_lock &list->lock#8 irq_context: 0 (wq_completion)hci5 (work_completion)(&hdev->cmd_sync_work) &hdev->req_lock &list->lock#7 irq_context: 0 (wq_completion)hci5 (work_completion)(&hdev->cmd_sync_work) &hdev->req_lock rcu_read_lock &pool->lock/1 irq_context: 0 (wq_completion)hci5 (work_completion)(&hdev->cmd_sync_work) &hdev->req_lock rcu_read_lock &pool->lock/1 &obj_hash[i].lock irq_context: 0 (wq_completion)hci5 (work_completion)(&hdev->cmd_sync_work) &hdev->req_lock rcu_read_lock &pool->lock/1 &p->pi_lock irq_context: 0 (wq_completion)hci5 (work_completion)(&hdev->cmd_sync_work) &hdev->req_lock &hdev->req_wait_q irq_context: 0 (wq_completion)hci5 (work_completion)(&hdev->cmd_sync_work) &hdev->req_lock &obj_hash[i].lock irq_context: 0 (wq_completion)hci5 (work_completion)(&hdev->cmd_sync_work) &hdev->req_lock &base->lock irq_context: 0 (wq_completion)hci5 (work_completion)(&hdev->cmd_sync_work) &hdev->req_lock &base->lock &obj_hash[i].lock irq_context: 0 (wq_completion)hci5 (work_completion)(&hdev->cmd_sync_work) &hdev->req_lock &rq->__lock irq_context: 0 (wq_completion)hci5 (work_completion)(&hdev->cmd_sync_work) &hdev->req_lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)hci1 (work_completion)(&hdev->cmd_sync_work) &hdev->req_lock &c->lock irq_context: 0 (wq_completion)hci1 (work_completion)(&hdev->cmd_sync_work) &hdev->req_lock &n->list_lock irq_context: 0 (wq_completion)hci1 (work_completion)(&hdev->cmd_sync_work) &hdev->req_lock &n->list_lock &c->lock irq_context: 0 (wq_completion)hci1 (work_completion)(&hdev->cmd_sync_work) &hdev->req_lock &list->lock#8 irq_context: 0 (wq_completion)hci1 (work_completion)(&hdev->cmd_sync_work) &hdev->req_lock &list->lock#7 irq_context: 0 (wq_completion)hci1 (work_completion)(&hdev->cmd_sync_work) &hdev->req_lock rcu_read_lock &pool->lock/1 irq_context: 0 (wq_completion)hci1 (work_completion)(&hdev->cmd_sync_work) &hdev->req_lock rcu_read_lock &pool->lock/1 &obj_hash[i].lock irq_context: 0 (wq_completion)hci1 (work_completion)(&hdev->cmd_sync_work) &hdev->req_lock rcu_read_lock &pool->lock/1 &p->pi_lock irq_context: 0 (wq_completion)hci1 (work_completion)(&hdev->cmd_sync_work) &hdev->req_lock &hdev->req_wait_q irq_context: 0 (wq_completion)hci1 (work_completion)(&hdev->cmd_sync_work) &hdev->req_lock &obj_hash[i].lock irq_context: 0 (wq_completion)hci1 (work_completion)(&hdev->cmd_sync_work) &hdev->req_lock &base->lock irq_context: 0 (wq_completion)hci1 (work_completion)(&hdev->cmd_sync_work) &hdev->req_lock &base->lock &obj_hash[i].lock irq_context: 0 (wq_completion)hci1 (work_completion)(&hdev->cmd_sync_work) &hdev->req_lock &rq->__lock irq_context: 0 (wq_completion)hci1 (work_completion)(&hdev->cmd_sync_work) &hdev->req_lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)hci2 (work_completion)(&hdev->cmd_sync_work) &hdev->req_lock &c->lock irq_context: 0 (wq_completion)hci2 (work_completion)(&hdev->cmd_sync_work) &hdev->req_lock &n->list_lock irq_context: 0 (wq_completion)hci2 (work_completion)(&hdev->cmd_sync_work) &hdev->req_lock &n->list_lock &c->lock irq_context: 0 (wq_completion)hci4#2 (work_completion)(&(&conn->disc_work)->work) irq_context: 0 (wq_completion)hci2 (work_completion)(&hdev->cmd_sync_work) &hdev->req_lock &list->lock#8 irq_context: 0 (wq_completion)hci4#2 (work_completion)(&(&conn->disc_work)->work) &hdev->unregister_lock irq_context: 0 (wq_completion)hci2 (work_completion)(&hdev->cmd_sync_work) &hdev->req_lock &list->lock#7 irq_context: 0 (wq_completion)hci4#2 (work_completion)(&(&conn->disc_work)->work) &hdev->unregister_lock fs_reclaim irq_context: 0 (wq_completion)hci2 (work_completion)(&hdev->cmd_sync_work) &hdev->req_lock rcu_read_lock &pool->lock/1 irq_context: 0 (wq_completion)hci4#2 (work_completion)(&(&conn->disc_work)->work) &hdev->unregister_lock fs_reclaim mmu_notifier_invalidate_range_start irq_context: 0 (wq_completion)hci2 (work_completion)(&hdev->cmd_sync_work) &hdev->req_lock rcu_read_lock &pool->lock/1 &obj_hash[i].lock irq_context: 0 (wq_completion)hci4#2 (work_completion)(&(&conn->disc_work)->work) &hdev->unregister_lock &c->lock irq_context: 0 (wq_completion)hci2 (work_completion)(&hdev->cmd_sync_work) &hdev->req_lock rcu_read_lock &pool->lock/1 &p->pi_lock irq_context: 0 (wq_completion)hci4#2 (work_completion)(&(&conn->disc_work)->work) &hdev->unregister_lock &hdev->cmd_sync_work_lock irq_context: 0 (wq_completion)hci2 (work_completion)(&hdev->cmd_sync_work) &hdev->req_lock &hdev->req_wait_q irq_context: 0 (wq_completion)hci4#2 (work_completion)(&(&conn->disc_work)->work) &hdev->unregister_lock rcu_read_lock &pool->lock/1 irq_context: 0 (wq_completion)hci2 (work_completion)(&hdev->cmd_sync_work) &hdev->req_lock &obj_hash[i].lock irq_context: 0 (wq_completion)hci4#2 (work_completion)(&(&conn->disc_work)->work) &hdev->unregister_lock rcu_read_lock &pool->lock/1 &obj_hash[i].lock irq_context: 0 (wq_completion)hci2 (work_completion)(&hdev->cmd_sync_work) &hdev->req_lock &base->lock irq_context: 0 (wq_completion)hci4#2 (work_completion)(&(&conn->disc_work)->work) &hdev->unregister_lock rcu_read_lock &pool->lock/1 &p->pi_lock irq_context: 0 (wq_completion)hci2 (work_completion)(&hdev->cmd_sync_work) &hdev->req_lock &base->lock &obj_hash[i].lock irq_context: 0 (wq_completion)hci4 (work_completion)(&hdev->cmd_sync_work) &hdev->req_lock &c->lock irq_context: 0 (wq_completion)hci2 (work_completion)(&hdev->cmd_sync_work) &hdev->req_lock &rq->__lock irq_context: 0 (wq_completion)hci4 (work_completion)(&hdev->cmd_sync_work) &hdev->req_lock &list->lock#8 irq_context: 0 (wq_completion)hci2 (work_completion)(&hdev->cmd_sync_work) &hdev->req_lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)hci4 (work_completion)(&hdev->cmd_sync_work) &hdev->req_lock &list->lock#7 irq_context: 0 (wq_completion)hci4 (work_completion)(&hdev->cmd_sync_work) &hdev->req_lock rcu_read_lock &pool->lock/1 irq_context: 0 (wq_completion)hci4 (work_completion)(&hdev->cmd_sync_work) &hdev->req_lock rcu_read_lock &pool->lock/1 &obj_hash[i].lock irq_context: 0 (wq_completion)hci4 (work_completion)(&hdev->cmd_sync_work) &hdev->req_lock rcu_read_lock &pool->lock/1 &p->pi_lock irq_context: 0 (wq_completion)hci4 (work_completion)(&hdev->cmd_sync_work) &hdev->req_lock rcu_read_lock &pool->lock/1 &p->pi_lock &rq->__lock irq_context: 0 (wq_completion)hci4 (work_completion)(&hdev->cmd_sync_work) &hdev->req_lock rcu_read_lock &pool->lock/1 &p->pi_lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)hci4 (work_completion)(&hdev->cmd_sync_work) &hdev->req_lock &hdev->req_wait_q irq_context: 0 (wq_completion)hci4 (work_completion)(&hdev->cmd_sync_work) &hdev->req_lock &obj_hash[i].lock irq_context: 0 (wq_completion)hci4 (work_completion)(&hdev->cmd_sync_work) &hdev->req_lock &base->lock irq_context: 0 (wq_completion)hci4 (work_completion)(&hdev->cmd_sync_work) &hdev->req_lock &base->lock &obj_hash[i].lock irq_context: 0 (wq_completion)hci4 (work_completion)(&hdev->cmd_sync_work) &hdev->req_lock &rq->__lock irq_context: 0 (wq_completion)hci4 (work_completion)(&hdev->cmd_sync_work) &hdev->req_lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)hci0 (work_completion)(&hdev->cmd_sync_work) &hdev->req_lock &list->lock#8 irq_context: 0 (wq_completion)hci0 (work_completion)(&hdev->cmd_sync_work) &hdev->req_lock &list->lock#7 irq_context: 0 (wq_completion)hci0 (work_completion)(&hdev->cmd_sync_work) &hdev->req_lock rcu_read_lock &pool->lock/1 irq_context: 0 (wq_completion)hci0 (work_completion)(&hdev->cmd_sync_work) &hdev->req_lock rcu_read_lock &pool->lock/1 &obj_hash[i].lock irq_context: 0 (wq_completion)hci0 (work_completion)(&hdev->cmd_sync_work) &hdev->req_lock rcu_read_lock &pool->lock/1 &p->pi_lock irq_context: 0 (wq_completion)hci0 (work_completion)(&hdev->cmd_sync_work) &hdev->req_lock rcu_read_lock &pool->lock/1 &p->pi_lock &rq->__lock irq_context: 0 (wq_completion)hci0 (work_completion)(&hdev->cmd_sync_work) &hdev->req_lock rcu_read_lock &pool->lock/1 &p->pi_lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)hci0 (work_completion)(&hdev->cmd_sync_work) &hdev->req_lock &hdev->req_wait_q irq_context: 0 (wq_completion)hci0 (work_completion)(&hdev->cmd_sync_work) &hdev->req_lock &obj_hash[i].lock irq_context: 0 (wq_completion)hci0 (work_completion)(&hdev->cmd_sync_work) &hdev->req_lock &base->lock irq_context: 0 (wq_completion)hci0 (work_completion)(&hdev->cmd_sync_work) &hdev->req_lock &base->lock &obj_hash[i].lock irq_context: 0 (wq_completion)hci0 (work_completion)(&hdev->cmd_sync_work) &hdev->req_lock &rq->__lock irq_context: 0 (wq_completion)hci0 (work_completion)(&hdev->cmd_sync_work) &hdev->req_lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 &f->f_pos_lock sb_writers#4 &sb->s_type->i_mutex_key#8 batched_entropy_u8.lock crngs.lock irq_context: 0 rtnl_mutex &net->ipv6.addrconf_hash_lock &base->lock irq_context: 0 rtnl_mutex &net->ipv6.addrconf_hash_lock &base->lock &obj_hash[i].lock irq_context: 0 sb_writers#5 &type->i_mutex_dir_key#5 remove_cache_srcu rcu_read_lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 &ep->mtx wakeup_ida.xa_lock irq_context: 0 &ep->mtx &x->wait#9 irq_context: 0 &ep->mtx &k->list_lock irq_context: 0 &ep->mtx gdp_mutex irq_context: 0 &ep->mtx gdp_mutex &k->list_lock irq_context: 0 &ep->mtx gdp_mutex &rq->__lock irq_context: 0 &ep->mtx gdp_mutex &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 &ep->mtx gdp_mutex fs_reclaim irq_context: 0 &ep->mtx gdp_mutex fs_reclaim mmu_notifier_invalidate_range_start irq_context: 0 &ep->mtx gdp_mutex pool_lock#2 irq_context: 0 &ep->mtx gdp_mutex lock irq_context: 0 &ep->mtx gdp_mutex lock kernfs_idr_lock irq_context: 0 &ep->mtx gdp_mutex lock kernfs_idr_lock &c->lock irq_context: 0 &ep->mtx gdp_mutex lock kernfs_idr_lock pool_lock#2 irq_context: 0 &ep->mtx gdp_mutex &root->kernfs_rwsem irq_context: 0 &ep->mtx gdp_mutex &root->kernfs_rwsem &root->kernfs_iattr_rwsem irq_context: 0 &ep->mtx lock irq_context: 0 &ep->mtx lock kernfs_idr_lock irq_context: 0 &ep->mtx &root->kernfs_rwsem irq_context: 0 &ep->mtx &root->kernfs_rwsem &root->kernfs_iattr_rwsem irq_context: 0 &ep->mtx bus_type_sem irq_context: 0 &ep->mtx sysfs_symlink_target_lock irq_context: 0 &ep->mtx uevent_sock_mutex irq_context: 0 &ep->mtx uevent_sock_mutex fs_reclaim irq_context: 0 &ep->mtx uevent_sock_mutex fs_reclaim mmu_notifier_invalidate_range_start irq_context: 0 &ep->mtx uevent_sock_mutex pool_lock#2 irq_context: 0 &ep->mtx uevent_sock_mutex nl_table_lock irq_context: 0 &ep->mtx uevent_sock_mutex rlock-AF_NETLINK irq_context: 0 &ep->mtx uevent_sock_mutex rcu_read_lock &ei->socket.wq.wait irq_context: 0 &ep->mtx uevent_sock_mutex rcu_read_lock &ei->socket.wq.wait &ep->lock irq_context: 0 &ep->mtx uevent_sock_mutex rcu_read_lock &ei->socket.wq.wait &ep->lock &ep->wq irq_context: 0 &ep->mtx uevent_sock_mutex rcu_read_lock &ei->socket.wq.wait &ep->lock &ep->wq &p->pi_lock irq_context: 0 &ep->mtx uevent_sock_mutex rcu_read_lock &ei->socket.wq.wait &ep->lock &ep->wq &p->pi_lock &rq->__lock irq_context: 0 &ep->mtx uevent_sock_mutex rcu_read_lock &ei->socket.wq.wait &ep->lock &ep->wq &p->pi_lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 &ep->mtx uevent_sock_mutex rcu_read_lock rcu_node_0 irq_context: 0 &ep->mtx uevent_sock_mutex rcu_read_lock &rq->__lock irq_context: 0 &ep->mtx uevent_sock_mutex rcu_read_lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 &ep->mtx uevent_sock_mutex rcu_read_lock &rcu_state.gp_wq irq_context: 0 &ep->mtx uevent_sock_mutex rcu_read_lock &rcu_state.gp_wq &p->pi_lock irq_context: 0 &ep->mtx uevent_sock_mutex rcu_read_lock &rcu_state.gp_wq &p->pi_lock &rq->__lock irq_context: 0 &ep->mtx uevent_sock_mutex rcu_read_lock &rcu_state.gp_wq &p->pi_lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 &ep->mtx uevent_sock_mutex nl_table_wait.lock irq_context: 0 &ep->mtx uevent_sock_mutex &obj_hash[i].lock irq_context: 0 &ep->mtx uevent_sock_mutex &rq->__lock irq_context: 0 &ep->mtx uevent_sock_mutex &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 &ep->mtx subsys mutex#15 irq_context: 0 &ep->mtx subsys mutex#15 &k->k_lock irq_context: 0 &ep->mtx events_lock irq_context: 0 &ep->mtx &dentry->d_lock irq_context: 0 &ep->mtx &u->lock irq_context: 0 &ep->mtx &u->lock &u->peer_wait irq_context: 0 &u->iolock &u->peer_wait &ei->socket.wq.wait irq_context: 0 &u->iolock &u->peer_wait &ei->socket.wq.wait &ep->lock irq_context: 0 &u->iolock &u->peer_wait &ei->socket.wq.wait &ep->lock rcu_read_lock &ws->lock irq_context: 0 &u->iolock &u->peer_wait &ei->socket.wq.wait &ep->lock rcu_read_lock &ws->lock tk_core.seq.seqcount irq_context: 0 &u->iolock &u->peer_wait &ei->socket.wq.wait &ep->lock rcu_read_lock &ws->lock &obj_hash[i].lock irq_context: 0 &ep->mtx &ws->lock irq_context: 0 &ep->mtx &ws->lock tk_core.seq.seqcount irq_context: 0 &ep->mtx &ws->lock &obj_hash[i].lock irq_context: 0 &ep->mtx &ep->lock &ws->lock irq_context: 0 &ep->mtx &ep->lock &ws->lock tk_core.seq.seqcount irq_context: 0 &ep->mtx &ep->lock &ws->lock &obj_hash[i].lock irq_context: 0 &u->iolock &u->peer_wait &ei->socket.wq.wait &ep->lock &ep->wq irq_context: 0 &u->iolock &u->peer_wait &ei->socket.wq.wait &ep->lock &ep->wq &p->pi_lock irq_context: 0 &u->iolock &u->peer_wait &ei->socket.wq.wait &ep->lock &ep->wq &p->pi_lock &rq->__lock irq_context: 0 &u->iolock &u->peer_wait &ei->socket.wq.wait &ep->lock &ep->wq &p->pi_lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: softirq &(&bat_priv->bla.work)->timer rcu_read_lock &pool->lock/1 &p->pi_lock &cfs_rq->removed.lock irq_context: 0 (wq_completion)ipv6_addrconf (work_completion)(&(&ifa->dad_work)->work) rtnl_mutex &base->lock irq_context: 0 (wq_completion)ipv6_addrconf (work_completion)(&(&ifa->dad_work)->work) rtnl_mutex &base->lock &obj_hash[i].lock irq_context: 0 (wq_completion)ipv6_addrconf (work_completion)(&(&ifa->dad_work)->work) rtnl_mutex &idev->mc_lock rcu_read_lock rcu_read_lock &obj_hash[i].lock pool_lock irq_context: 0 (wq_completion)ipv6_addrconf (work_completion)(&(&ifa->dad_work)->work) rtnl_mutex rcu_read_lock rcu_read_lock rcu_read_lock_bh dev->qdisc_tx_busylock ?: &qdisc_tx_busylock rcu_read_lock &r->producer_lock irq_context: 0 (wq_completion)bat_events (work_completion)(&(&bat_priv->bla.work)->work) rcu_read_lock &rcu_state.gp_wq irq_context: 0 (wq_completion)bat_events (work_completion)(&(&bat_priv->bla.work)->work) rcu_read_lock &rcu_state.gp_wq &p->pi_lock irq_context: 0 (wq_completion)bat_events (work_completion)(&(&bat_priv->bla.work)->work) rcu_read_lock &rcu_state.gp_wq &p->pi_lock &rq->__lock irq_context: 0 (wq_completion)bat_events (work_completion)(&(&bat_priv->bla.work)->work) rcu_read_lock &rcu_state.gp_wq &p->pi_lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 &sb->s_type->i_mutex_key#10 rcu_read_lock &ei->socket.wq.wait &ep->lock irq_context: 0 &u->lock &ei->socket.wq.wait &ep->lock irq_context: 0 rtnl_mutex &ul->lock#2 &____s->seqcount irq_context: 0 rtnl_mutex &ul->lock#2 rcu_read_lock pool_lock#2 irq_context: 0 rtnl_mutex &ul->lock#2 &obj_hash[i].lock irq_context: 0 &ep->mtx &ACCESS_PRIVATE(sdp, lock) irq_context: 0 &ep->mtx wakeup_srcu irq_context: 0 &ep->mtx wakeup_srcu_srcu_usage.lock &ACCESS_PRIVATE(sdp, lock) irq_context: 0 &ep->mtx wakeup_srcu_srcu_usage.lock rcu_read_lock &pool->lock irq_context: 0 &ep->mtx wakeup_srcu_srcu_usage.lock rcu_read_lock &pool->lock &obj_hash[i].lock irq_context: 0 &ep->mtx &x->wait#2 irq_context: 0 &ep->mtx (&ws->timer) irq_context: 0 &ep->mtx &base->lock irq_context: 0 &ep->mtx &root->kernfs_rwsem kernfs_idr_lock irq_context: 0 &ep->mtx &root->kernfs_rwsem &obj_hash[i].lock irq_context: 0 &ep->mtx &root->kernfs_rwsem pool_lock#2 irq_context: 0 &ep->mtx subsys mutex#15 &k->k_lock klist_remove_lock irq_context: 0 &ep->mtx deferred_probe_mutex irq_context: 0 &ep->mtx device_links_lock irq_context: 0 &ep->mtx mmu_notifier_invalidate_range_start irq_context: 0 &ep->mtx uevent_sock_mutex mmu_notifier_invalidate_range_start irq_context: 0 &ep->mtx uevent_sock_mutex &c->lock irq_context: 0 &ep->mtx deleted_ws.lock irq_context: 0 events_lock irq_context: 0 wakeup_srcu irq_context: 0 wakeup_srcu_srcu_usage.lock &ACCESS_PRIVATE(sdp, lock) irq_context: 0 wakeup_srcu_srcu_usage.lock rcu_read_lock &pool->lock irq_context: 0 wakeup_srcu_srcu_usage.lock rcu_read_lock &pool->lock &obj_hash[i].lock irq_context: 0 (&ws->timer) irq_context: 0 subsys mutex#15 irq_context: 0 subsys mutex#15 &k->k_lock irq_context: 0 subsys mutex#15 &k->k_lock klist_remove_lock irq_context: 0 uevent_sock_mutex mmu_notifier_invalidate_range_start irq_context: 0 uevent_sock_mutex &c->lock irq_context: 0 uevent_sock_mutex rlock-AF_NETLINK irq_context: 0 uevent_sock_mutex rcu_read_lock &ei->socket.wq.wait irq_context: 0 uevent_sock_mutex rcu_read_lock &ei->socket.wq.wait &ep->lock irq_context: 0 uevent_sock_mutex rcu_read_lock &ei->socket.wq.wait &ep->lock &ep->wq irq_context: 0 uevent_sock_mutex rcu_read_lock &ei->socket.wq.wait &ep->lock &ep->wq &p->pi_lock irq_context: 0 uevent_sock_mutex rcu_read_lock &ei->socket.wq.wait &ep->lock &ep->wq &p->pi_lock &cfs_rq->removed.lock irq_context: 0 uevent_sock_mutex rcu_read_lock &ei->socket.wq.wait &ep->lock &ep->wq &p->pi_lock &rq->__lock irq_context: 0 uevent_sock_mutex rcu_read_lock &ei->socket.wq.wait &ep->lock &ep->wq &p->pi_lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 uevent_sock_mutex rcu_read_lock rcu_node_0 irq_context: 0 uevent_sock_mutex rcu_read_lock &rq->__lock irq_context: 0 uevent_sock_mutex rcu_read_lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 gdp_mutex sysfs_symlink_target_lock irq_context: 0 gdp_mutex &obj_hash[i].lock irq_context: 0 &ws->lock irq_context: 0 deleted_ws.lock irq_context: 0 wakeup_ida.xa_lock irq_context: 0 sb_writers#5 &type->i_mutex_dir_key#5 tomoyo_ss batched_entropy_u8.lock irq_context: 0 sb_writers#5 &type->i_mutex_dir_key#5 tomoyo_ss kfence_freelist_lock irq_context: 0 sb_writers#5 &type->i_mutex_dir_key#5 tomoyo_ss &meta->lock irq_context: 0 (wq_completion)events (work_completion)(&(&hwstats->traffic_dw)->work) &hwstats->hwsdev_list_lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)wg-crypt-wg1 &rq->__lock irq_context: 0 (wq_completion)wg-crypt-wg1 &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)wg-crypt-wg2#2 (work_completion)(&({ do { const void *__vpp_verify = (typeof((worker) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __asm__ ("" : "=r"(__ptr) : "0"((typeof(*((worker))) *)((worker)))); (typeof((typeof(*((worker))) *)((worker)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); })->work) &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: softirq (&n->timer) &meta->lock irq_context: softirq (&n->timer) kfence_freelist_lock irq_context: 0 tracepoints_mutex rcu_read_lock &rq->__lock irq_context: 0 tracepoints_mutex rcu_read_lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)events_freezable (work_completion)(&vb->update_balloon_stats_work) &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)events (work_completion)(&p->wq) &cfs_rq->removed.lock irq_context: 0 &pipe->mutex/1 sk_lock-AF_INET6 rcu_read_lock &nf_conntrack_locks[i] &nf_conntrack_locks[i]/1 irq_context: 0 &pipe->mutex/1 sk_lock-AF_INET6 rcu_read_lock &nf_conntrack_locks[i] &nf_conntrack_locks[i]/1 batched_entropy_u8.lock irq_context: 0 &pipe->mutex/1 sk_lock-AF_INET6 k-slock-AF_INET6 rcu_read_lock rcu_read_lock rcu_read_lock stock_lock irq_context: 0 &pipe->mutex/1 sk_lock-AF_INET6 k-slock-AF_INET6 rcu_read_lock rcu_read_lock rcu_read_lock pool_lock#2 irq_context: 0 &pipe->mutex/1 sk_lock-AF_INET6 k-slock-AF_INET6 rcu_read_lock rcu_read_lock rcu_read_lock &dir->lock#2 irq_context: 0 &pipe->mutex/1 sk_lock-AF_INET6 k-slock-AF_INET6 rcu_read_lock pool_lock#2 irq_context: 0 &pipe->mutex/1 sk_lock-AF_INET6 k-slock-AF_INET6 rcu_read_lock rcu_read_lock pool_lock#2 irq_context: 0 &pipe->mutex/1 sk_lock-AF_INET6 k-slock-AF_INET6 rcu_read_lock rcu_read_lock &obj_hash[i].lock irq_context: 0 &pipe->mutex/1 sk_lock-AF_INET6 k-slock-AF_INET6 rcu_read_lock rcu_read_lock &dir->lock#2 irq_context: 0 &pipe->mutex/1 sk_lock-AF_INET6 k-slock-AF_INET6 rcu_read_lock rcu_read_lock &tbl->lock irq_context: 0 &pipe->mutex/1 sk_lock-AF_INET6 k-slock-AF_INET6 rcu_read_lock rcu_read_lock &n->lock irq_context: 0 &pipe->mutex/1 sk_lock-AF_INET6 k-slock-AF_INET6 rcu_read_lock rcu_read_lock &____s->seqcount#9 irq_context: 0 &pipe->mutex/1 sk_lock-AF_INET6 k-slock-AF_INET6 rcu_read_lock rcu_read_lock rcu_read_lock_bh noop_qdisc.q.lock irq_context: 0 &pipe->mutex/1 sk_lock-AF_INET6 k-slock-AF_INET6 rcu_read_lock rcu_read_lock rcu_read_lock_bh &obj_hash[i].lock irq_context: 0 &pipe->mutex/1 sk_lock-AF_INET6 k-slock-AF_INET6 rcu_read_lock rcu_read_lock rcu_read_lock_bh pool_lock#2 irq_context: 0 &pipe->mutex/1 sk_lock-AF_INET6 &obj_hash[i].lock irq_context: 0 (wq_completion)events (work_completion)(&w->w) nfc_devlist_mutex &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)events (work_completion)(&w->w) &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)events (work_completion)(&(&nsim_dev->trap_data->trap_report_dw)->work) &devlink->lock_key#3 &cfs_rq->removed.lock irq_context: 0 &xt[i].mutex purge_vmap_area_lock &meta->lock irq_context: 0 &xt[i].mutex purge_vmap_area_lock kfence_freelist_lock irq_context: 0 (wq_completion)hci5 (work_completion)(&hdev->cmd_sync_work) &hdev->req_lock (&timer.timer) irq_context: 0 (wq_completion)hci5 (work_completion)(&hdev->cmd_sync_work) &hdev->req_lock &hdev->lock hci_cb_list_lock rcu_state.exp_mutex rcu_read_lock &pool->lock &p->pi_lock irq_context: 0 (wq_completion)hci2 (work_completion)(&hdev->cmd_sync_work) &hdev->req_lock (&timer.timer) irq_context: 0 (wq_completion)hci5 (work_completion)(&hdev->cmd_sync_work) &hdev->req_lock &hdev->lock hci_cb_list_lock rcu_state.exp_mutex rcu_read_lock &pool->lock &p->pi_lock &rq->__lock irq_context: 0 (wq_completion)hci2 (work_completion)(&hdev->cmd_sync_work) &hdev->req_lock &hdev->lock irq_context: 0 (wq_completion)hci5 (work_completion)(&hdev->cmd_sync_work) &hdev->req_lock &hdev->lock hci_cb_list_lock rcu_state.exp_mutex rcu_read_lock &pool->lock &p->pi_lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)hci2 (work_completion)(&hdev->cmd_sync_work) &hdev->req_lock &hdev->lock fs_reclaim irq_context: 0 (wq_completion)hci2 (work_completion)(&hdev->cmd_sync_work) &hdev->req_lock &hdev->lock fs_reclaim mmu_notifier_invalidate_range_start irq_context: 0 (wq_completion)hci2 (work_completion)(&hdev->cmd_sync_work) &hdev->req_lock &hdev->lock tk_core.seq.seqcount irq_context: 0 (wq_completion)hci2 (work_completion)(&hdev->cmd_sync_work) &hdev->req_lock &hdev->lock hci_sk_list.lock irq_context: 0 (wq_completion)hci2 (work_completion)(&hdev->cmd_sync_work) &hdev->req_lock &hdev->lock &obj_hash[i].lock irq_context: 0 (wq_completion)hci2 (work_completion)(&hdev->cmd_sync_work) &hdev->req_lock &hdev->lock hci_cb_list_lock irq_context: 0 (wq_completion)hci2 (work_completion)(&hdev->cmd_sync_work) &hdev->req_lock &hdev->lock hci_cb_list_lock hci_cb_list_lock.wait_lock irq_context: 0 (wq_completion)hci2 (work_completion)(&hdev->cmd_sync_work) &hdev->req_lock &hdev->lock hci_cb_list_lock &rq->__lock irq_context: 0 (wq_completion)hci2 (work_completion)(&hdev->cmd_sync_work) &hdev->req_lock &hdev->lock hci_cb_list_lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)hci1 (work_completion)(&hdev->cmd_sync_work) &hdev->req_lock (&timer.timer) irq_context: 0 (wq_completion)hci1 (work_completion)(&hdev->cmd_sync_work) &hdev->req_lock &hdev->lock irq_context: 0 (wq_completion)hci1 (work_completion)(&hdev->cmd_sync_work) &hdev->req_lock &hdev->lock fs_reclaim irq_context: 0 (wq_completion)hci1 (work_completion)(&hdev->cmd_sync_work) &hdev->req_lock &hdev->lock fs_reclaim mmu_notifier_invalidate_range_start irq_context: 0 (wq_completion)hci1 (work_completion)(&hdev->cmd_sync_work) &hdev->req_lock &hdev->lock &c->lock irq_context: 0 (wq_completion)hci1 (work_completion)(&hdev->cmd_sync_work) &hdev->req_lock &hdev->lock tk_core.seq.seqcount irq_context: 0 (wq_completion)hci1 (work_completion)(&hdev->cmd_sync_work) &hdev->req_lock &hdev->lock hci_sk_list.lock irq_context: 0 (wq_completion)hci1 (work_completion)(&hdev->cmd_sync_work) &hdev->req_lock &hdev->lock &obj_hash[i].lock irq_context: 0 (wq_completion)hci1 (work_completion)(&hdev->cmd_sync_work) &hdev->req_lock &hdev->lock hci_cb_list_lock irq_context: 0 (wq_completion)hci1 (work_completion)(&hdev->cmd_sync_work) &hdev->req_lock &hdev->lock hci_cb_list_lock hci_cb_list_lock.wait_lock irq_context: 0 (wq_completion)hci1 (work_completion)(&hdev->cmd_sync_work) &hdev->req_lock &hdev->lock hci_cb_list_lock &rq->__lock irq_context: 0 (wq_completion)hci1 (work_completion)(&hdev->cmd_sync_work) &hdev->req_lock &hdev->lock hci_cb_list_lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)hci4 (work_completion)(&hdev->cmd_sync_work) &hdev->req_lock (&timer.timer) irq_context: 0 (wq_completion)hci4 (work_completion)(&hdev->cmd_sync_work) &hdev->req_lock &hdev->lock irq_context: 0 (wq_completion)hci4 (work_completion)(&hdev->cmd_sync_work) &hdev->req_lock &hdev->lock fs_reclaim irq_context: 0 (wq_completion)hci4 (work_completion)(&hdev->cmd_sync_work) &hdev->req_lock &hdev->lock fs_reclaim mmu_notifier_invalidate_range_start irq_context: 0 (wq_completion)hci4 (work_completion)(&hdev->cmd_sync_work) &hdev->req_lock &hdev->lock tk_core.seq.seqcount irq_context: 0 (wq_completion)hci4 (work_completion)(&hdev->cmd_sync_work) &hdev->req_lock &hdev->lock hci_sk_list.lock irq_context: 0 (wq_completion)hci4 (work_completion)(&hdev->cmd_sync_work) &hdev->req_lock &hdev->lock &obj_hash[i].lock irq_context: 0 (wq_completion)hci4 (work_completion)(&hdev->cmd_sync_work) &hdev->req_lock &hdev->lock hci_cb_list_lock irq_context: 0 (wq_completion)hci4 (work_completion)(&hdev->cmd_sync_work) &hdev->req_lock &hdev->lock hci_cb_list_lock hci_cb_list_lock.wait_lock irq_context: 0 (wq_completion)hci4 (work_completion)(&hdev->cmd_sync_work) &hdev->req_lock &hdev->lock hci_cb_list_lock &rq->__lock irq_context: 0 (wq_completion)hci4 (work_completion)(&hdev->cmd_sync_work) &hdev->req_lock &hdev->lock hci_cb_list_lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)hci3 (work_completion)(&hdev->cmd_sync_work) &hdev->req_lock (&timer.timer) irq_context: 0 (wq_completion)hci3 (work_completion)(&hdev->cmd_sync_work) &hdev->req_lock &hdev->lock irq_context: 0 (wq_completion)hci3 (work_completion)(&hdev->cmd_sync_work) &hdev->req_lock &hdev->lock fs_reclaim irq_context: 0 (wq_completion)hci3 (work_completion)(&hdev->cmd_sync_work) &hdev->req_lock &hdev->lock fs_reclaim mmu_notifier_invalidate_range_start irq_context: 0 (wq_completion)hci3 (work_completion)(&hdev->cmd_sync_work) &hdev->req_lock &hdev->lock &c->lock irq_context: 0 (wq_completion)hci3 (work_completion)(&hdev->cmd_sync_work) &hdev->req_lock &hdev->lock tk_core.seq.seqcount irq_context: 0 (wq_completion)hci3 (work_completion)(&hdev->cmd_sync_work) &hdev->req_lock &hdev->lock hci_sk_list.lock irq_context: 0 (wq_completion)hci3 (work_completion)(&hdev->cmd_sync_work) &hdev->req_lock &hdev->lock &obj_hash[i].lock irq_context: 0 (wq_completion)hci3 (work_completion)(&hdev->cmd_sync_work) &hdev->req_lock &hdev->lock hci_cb_list_lock irq_context: 0 (wq_completion)hci3 (work_completion)(&hdev->cmd_sync_work) &hdev->req_lock &hdev->lock hci_cb_list_lock hci_cb_list_lock.wait_lock irq_context: 0 (wq_completion)hci3 (work_completion)(&hdev->cmd_sync_work) &hdev->req_lock &hdev->lock hci_cb_list_lock &rq->__lock irq_context: 0 (wq_completion)hci3 (work_completion)(&hdev->cmd_sync_work) &hdev->req_lock &hdev->lock hci_cb_list_lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)hci5#2 (work_completion)(&(&hdev->cmd_timer)->work) &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)hci3#2 (work_completion)(&(&hdev->cmd_timer)->work) &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)hci0 (work_completion)(&hdev->cmd_sync_work) &hdev->req_lock (&timer.timer) irq_context: 0 (wq_completion)hci0 (work_completion)(&hdev->cmd_sync_work) &hdev->req_lock &hdev->lock irq_context: 0 (wq_completion)hci0 (work_completion)(&hdev->cmd_sync_work) &hdev->req_lock &hdev->lock fs_reclaim irq_context: 0 (wq_completion)hci0 (work_completion)(&hdev->cmd_sync_work) &hdev->req_lock &hdev->lock fs_reclaim mmu_notifier_invalidate_range_start irq_context: 0 (wq_completion)hci0 (work_completion)(&hdev->cmd_sync_work) &hdev->req_lock &hdev->lock tk_core.seq.seqcount irq_context: 0 (wq_completion)hci0 (work_completion)(&hdev->cmd_sync_work) &hdev->req_lock &hdev->lock hci_sk_list.lock irq_context: 0 (wq_completion)hci0 (work_completion)(&hdev->cmd_sync_work) &hdev->req_lock &hdev->lock &obj_hash[i].lock irq_context: 0 (wq_completion)hci0 (work_completion)(&hdev->cmd_sync_work) &hdev->req_lock &hdev->lock hci_cb_list_lock irq_context: 0 (wq_completion)hci0 (work_completion)(&hdev->cmd_sync_work) &hdev->req_lock &hdev->lock hci_cb_list_lock hci_cb_list_lock.wait_lock irq_context: 0 (wq_completion)hci0 (work_completion)(&hdev->cmd_sync_work) &hdev->req_lock &hdev->lock hci_cb_list_lock &rq->__lock irq_context: 0 (wq_completion)hci0 (work_completion)(&hdev->cmd_sync_work) &hdev->req_lock &hdev->lock hci_cb_list_lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)hci5 (work_completion)(&hdev->cmd_sync_work) &hdev->req_lock &hdev->lock hci_cb_list_lock (work_completion)(&(&conn->info_timer)->work) irq_context: 0 (wq_completion)hci5 (work_completion)(&hdev->cmd_sync_work) &hdev->req_lock &hdev->lock hci_cb_list_lock rcu_read_lock &pool->lock irq_context: 0 (wq_completion)hci5 (work_completion)(&hdev->cmd_sync_work) &hdev->req_lock &hdev->lock hci_cb_list_lock.wait_lock irq_context: 0 (wq_completion)hci5 (work_completion)(&hdev->cmd_sync_work) &hdev->req_lock &hdev->lock &p->pi_lock irq_context: 0 (wq_completion)hci5 (work_completion)(&hdev->cmd_sync_work) &hdev->req_lock &hdev->lock &p->pi_lock &rq->__lock irq_context: 0 (wq_completion)hci5 (work_completion)(&hdev->cmd_sync_work) &hdev->req_lock &hdev->lock &p->pi_lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)hci2 (work_completion)(&hdev->cmd_sync_work) &hdev->req_lock &hdev->lock hci_cb_list_lock &list->lock#10 irq_context: 0 (wq_completion)hci2 (work_completion)(&hdev->cmd_sync_work) &hdev->req_lock &hdev->lock hci_cb_list_lock &obj_hash[i].lock irq_context: 0 (wq_completion)hci5 (work_completion)(&hdev->cmd_sync_work) &hdev->req_lock &hdev->lock rcu_state.exp_mutex rcu_read_lock &pool->lock &p->pi_lock irq_context: 0 (wq_completion)hci2 (work_completion)(&hdev->cmd_sync_work) &hdev->req_lock &hdev->lock hci_cb_list_lock (work_completion)(&(&conn->id_addr_timer)->work) irq_context: 0 (wq_completion)hci5 (work_completion)(&hdev->cmd_sync_work) &hdev->req_lock &hdev->lock rcu_state.exp_mutex rcu_read_lock &pool->lock &p->pi_lock &rq->__lock irq_context: 0 (wq_completion)hci2 (work_completion)(&hdev->cmd_sync_work) &hdev->req_lock &hdev->lock hci_cb_list_lock &conn->chan_lock irq_context: 0 (wq_completion)hci5 (work_completion)(&hdev->cmd_sync_work) &hdev->req_lock &hdev->lock rcu_state.exp_mutex rcu_read_lock &pool->lock &p->pi_lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)hci2 (work_completion)(&hdev->cmd_sync_work) &hdev->req_lock &hdev->lock hci_cb_list_lock &rnp->exp_lock irq_context: 0 (wq_completion)hci2 (work_completion)(&hdev->cmd_sync_work) &hdev->req_lock &hdev->lock hci_cb_list_lock rcu_node_0 irq_context: 0 (wq_completion)hci2 (work_completion)(&hdev->cmd_sync_work) &hdev->req_lock &hdev->lock hci_cb_list_lock rcu_state.exp_mutex irq_context: 0 (wq_completion)hci2 (work_completion)(&hdev->cmd_sync_work) &hdev->req_lock &hdev->lock hci_cb_list_lock rcu_state.exp_mutex rcu_state.exp_mutex.wait_lock irq_context: 0 (wq_completion)hci2 (work_completion)(&hdev->cmd_sync_work) &hdev->req_lock &hdev->lock hci_cb_list_lock rcu_state.exp_mutex &rq->__lock irq_context: 0 (wq_completion)hci2 (work_completion)(&hdev->cmd_sync_work) &hdev->req_lock &hdev->lock hci_cb_list_lock rcu_state.exp_mutex &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)hci5 (work_completion)(&hdev->cmd_sync_work) &hdev->req_lock &hdev->lock rcu_state.exp_mutex.wait_lock irq_context: 0 (wq_completion)hci5 (work_completion)(&hdev->cmd_sync_work) &hdev->req_lock &hdev->lock &k->k_lock irq_context: 0 (wq_completion)hci2 (work_completion)(&hdev->cmd_sync_work) &hdev->req_lock &hdev->lock hci_cb_list_lock rcu_state.exp_mutex rcu_node_0 irq_context: 0 (wq_completion)hci5 (work_completion)(&hdev->cmd_sync_work) &hdev->req_lock &hdev->lock &root->kernfs_rwsem irq_context: 0 (wq_completion)hci2 (work_completion)(&hdev->cmd_sync_work) &hdev->req_lock &hdev->lock hci_cb_list_lock rcu_state.exp_mutex &obj_hash[i].lock irq_context: 0 (wq_completion)hci5 (work_completion)(&hdev->cmd_sync_work) &hdev->req_lock &hdev->lock &root->kernfs_rwsem irq_context: 0 (wq_completion)hci2 (work_completion)(&hdev->cmd_sync_work) &hdev->req_lock &hdev->lock hci_cb_list_lock rcu_state.exp_mutex rcu_read_lock &pool->lock irq_context: 0 (wq_completion)hci5 (work_completion)(&hdev->cmd_sync_work) &hdev->req_lock &hdev->lock dev_pm_qos_sysfs_mtx irq_context: 0 (wq_completion)hci2 (work_completion)(&hdev->cmd_sync_work) &hdev->req_lock &hdev->lock hci_cb_list_lock rcu_state.exp_mutex rcu_read_lock &pool->lock &obj_hash[i].lock irq_context: 0 (wq_completion)hci5 (work_completion)(&hdev->cmd_sync_work) &hdev->req_lock &hdev->lock dev_pm_qos_sysfs_mtx &root->kernfs_rwsem irq_context: 0 (wq_completion)hci2 (work_completion)(&hdev->cmd_sync_work) &hdev->req_lock &hdev->lock hci_cb_list_lock rcu_state.exp_mutex rcu_read_lock &pool->lock &p->pi_lock irq_context: 0 (wq_completion)hci5 (work_completion)(&hdev->cmd_sync_work) &hdev->req_lock &hdev->lock dev_pm_qos_sysfs_mtx &root->kernfs_rwsem irq_context: 0 (wq_completion)hci2 (work_completion)(&hdev->cmd_sync_work) &hdev->req_lock &hdev->lock hci_cb_list_lock rcu_state.exp_mutex rcu_read_lock &pool->lock &p->pi_lock &rq->__lock irq_context: 0 (wq_completion)hci5 (work_completion)(&hdev->cmd_sync_work) &hdev->req_lock &hdev->lock dev_pm_qos_sysfs_mtx dev_pm_qos_mtx irq_context: 0 (wq_completion)hci2 (work_completion)(&hdev->cmd_sync_work) &hdev->req_lock &hdev->lock hci_cb_list_lock rcu_state.exp_mutex rcu_read_lock &pool->lock &p->pi_lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)hci5 (work_completion)(&hdev->cmd_sync_work) &hdev->req_lock &hdev->lock &root->kernfs_rwsem &root->kernfs_iattr_rwsem irq_context: 0 (wq_completion)hci5 (work_completion)(&hdev->cmd_sync_work) &hdev->req_lock &hdev->lock &root->kernfs_rwsem kernfs_idr_lock irq_context: 0 (wq_completion)hci5 (work_completion)(&hdev->cmd_sync_work) &hdev->req_lock &hdev->lock &root->kernfs_rwsem &obj_hash[i].lock irq_context: 0 (wq_completion)hci5 (work_completion)(&hdev->cmd_sync_work) &hdev->req_lock &hdev->lock &root->kernfs_rwsem pool_lock#2 irq_context: 0 (wq_completion)hci5 (work_completion)(&hdev->cmd_sync_work) &hdev->req_lock &hdev->lock &root->kernfs_rwsem rcu_node_0 irq_context: 0 (wq_completion)hci2 (work_completion)(&hdev->cmd_sync_work) &hdev->req_lock &hdev->lock hci_cb_list_lock rcu_state.exp_mutex &rnp->exp_wq[2] irq_context: 0 (wq_completion)hci5 (work_completion)(&hdev->cmd_sync_work) &hdev->req_lock &hdev->lock &root->kernfs_rwsem &rq->__lock irq_context: 0 (wq_completion)hci5 (work_completion)(&hdev->cmd_sync_work) &hdev->req_lock &hdev->lock &root->kernfs_rwsem &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)hci2 (work_completion)(&hdev->cmd_sync_work) &hdev->req_lock &hdev->lock hci_cb_list_lock &list->lock#9 irq_context: 0 (wq_completion)hci2 (work_completion)(&hdev->cmd_sync_work) &hdev->req_lock &hdev->lock hci_cb_list_lock (work_completion)(&(&conn->info_timer)->work) irq_context: 0 (wq_completion)hci2 (work_completion)(&hdev->cmd_sync_work) &hdev->req_lock &hdev->lock hci_cb_list_lock rcu_read_lock &pool->lock irq_context: 0 (wq_completion)hci2 (work_completion)(&hdev->cmd_sync_work) &hdev->req_lock &hdev->lock hci_cb_list_lock.wait_lock irq_context: 0 (wq_completion)hci5 (work_completion)(&hdev->cmd_sync_work) &hdev->req_lock &hdev->lock kernfs_idr_lock irq_context: 0 (wq_completion)hci2 (work_completion)(&hdev->cmd_sync_work) &hdev->req_lock &hdev->lock &p->pi_lock irq_context: 0 (wq_completion)hci5 (work_completion)(&hdev->cmd_sync_work) &hdev->req_lock &hdev->lock &k->k_lock klist_remove_lock irq_context: 0 (wq_completion)hci2 (work_completion)(&hdev->cmd_sync_work) &hdev->req_lock &hdev->lock &p->pi_lock &rq->__lock irq_context: 0 (wq_completion)hci5 (work_completion)(&hdev->cmd_sync_work) &hdev->req_lock &hdev->lock &k->list_lock irq_context: 0 (wq_completion)hci2 (work_completion)(&hdev->cmd_sync_work) &hdev->req_lock &hdev->lock &p->pi_lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)hci5 (work_completion)(&hdev->cmd_sync_work) &hdev->req_lock &hdev->lock sysfs_symlink_target_lock irq_context: 0 (wq_completion)hci2 (work_completion)(&hdev->cmd_sync_work) &hdev->req_lock &hdev->lock &rq->__lock irq_context: 0 (wq_completion)hci5 (work_completion)(&hdev->cmd_sync_work) &hdev->req_lock &hdev->lock subsys mutex#81 irq_context: 0 (wq_completion)hci5 (work_completion)(&hdev->cmd_sync_work) &hdev->req_lock &hdev->lock subsys mutex#81 &k->k_lock irq_context: 0 (wq_completion)hci5 (work_completion)(&hdev->cmd_sync_work) &hdev->req_lock &hdev->lock subsys mutex#81 &k->k_lock klist_remove_lock irq_context: 0 (wq_completion)hci1 (work_completion)(&hdev->cmd_sync_work) &hdev->req_lock &hdev->lock hci_cb_list_lock &list->lock#10 irq_context: 0 (wq_completion)hci5 (work_completion)(&hdev->cmd_sync_work) &hdev->req_lock &hdev->lock &x->wait#9 irq_context: 0 (wq_completion)hci5 (work_completion)(&hdev->cmd_sync_work) &hdev->req_lock &hdev->lock dpm_list_mtx irq_context: 0 (wq_completion)hci1 (work_completion)(&hdev->cmd_sync_work) &hdev->req_lock &hdev->lock hci_cb_list_lock &obj_hash[i].lock irq_context: 0 (wq_completion)hci5 (work_completion)(&hdev->cmd_sync_work) &hdev->req_lock &hdev->lock &dev->power.lock irq_context: 0 (wq_completion)hci1 (work_completion)(&hdev->cmd_sync_work) &hdev->req_lock &hdev->lock hci_cb_list_lock (work_completion)(&(&conn->id_addr_timer)->work) irq_context: 0 (wq_completion)hci5 (work_completion)(&hdev->cmd_sync_work) &hdev->req_lock &hdev->lock deferred_probe_mutex irq_context: 0 (wq_completion)hci1 (work_completion)(&hdev->cmd_sync_work) &hdev->req_lock &hdev->lock hci_cb_list_lock &conn->chan_lock irq_context: 0 (wq_completion)hci5 (work_completion)(&hdev->cmd_sync_work) &hdev->req_lock &hdev->lock device_links_lock irq_context: 0 (wq_completion)hci1 (work_completion)(&hdev->cmd_sync_work) &hdev->req_lock &hdev->lock hci_cb_list_lock rcu_state.exp_mutex rcu_node_0 irq_context: 0 (wq_completion)hci5 (work_completion)(&hdev->cmd_sync_work) &hdev->req_lock &hdev->lock mmu_notifier_invalidate_range_start irq_context: 0 (wq_completion)hci1 (work_completion)(&hdev->cmd_sync_work) &hdev->req_lock &hdev->lock hci_cb_list_lock rcu_state.exp_mutex &obj_hash[i].lock irq_context: 0 (wq_completion)hci5 (work_completion)(&hdev->cmd_sync_work) &hdev->req_lock &hdev->lock uevent_sock_mutex irq_context: 0 (wq_completion)hci1 (work_completion)(&hdev->cmd_sync_work) &hdev->req_lock &hdev->lock hci_cb_list_lock rcu_state.exp_mutex rcu_read_lock &pool->lock irq_context: 0 (wq_completion)hci5 (work_completion)(&hdev->cmd_sync_work) &hdev->req_lock &hdev->lock uevent_sock_mutex mmu_notifier_invalidate_range_start irq_context: 0 (wq_completion)hci1 (work_completion)(&hdev->cmd_sync_work) &hdev->req_lock &hdev->lock hci_cb_list_lock rcu_state.exp_mutex rcu_read_lock &pool->lock &obj_hash[i].lock irq_context: 0 (wq_completion)hci5 (work_completion)(&hdev->cmd_sync_work) &hdev->req_lock &hdev->lock uevent_sock_mutex pool_lock#2 irq_context: 0 (wq_completion)hci1 (work_completion)(&hdev->cmd_sync_work) &hdev->req_lock &hdev->lock hci_cb_list_lock rcu_state.exp_mutex rcu_read_lock &pool->lock &p->pi_lock irq_context: 0 (wq_completion)hci5 (work_completion)(&hdev->cmd_sync_work) &hdev->req_lock &hdev->lock uevent_sock_mutex nl_table_lock irq_context: 0 (wq_completion)hci1 (work_completion)(&hdev->cmd_sync_work) &hdev->req_lock &hdev->lock hci_cb_list_lock rcu_state.exp_mutex rcu_read_lock &pool->lock &p->pi_lock &rq->__lock irq_context: 0 (wq_completion)hci5 (work_completion)(&hdev->cmd_sync_work) &hdev->req_lock &hdev->lock uevent_sock_mutex rlock-AF_NETLINK irq_context: 0 (wq_completion)hci1 (work_completion)(&hdev->cmd_sync_work) &hdev->req_lock &hdev->lock hci_cb_list_lock rcu_state.exp_mutex rcu_read_lock &pool->lock &p->pi_lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)hci5 (work_completion)(&hdev->cmd_sync_work) &hdev->req_lock &hdev->lock uevent_sock_mutex rcu_read_lock &ei->socket.wq.wait irq_context: 0 (wq_completion)hci1 (work_completion)(&hdev->cmd_sync_work) &hdev->req_lock &hdev->lock hci_cb_list_lock rcu_state.exp_mutex &rnp->exp_wq[3] irq_context: 0 (wq_completion)hci5 (work_completion)(&hdev->cmd_sync_work) &hdev->req_lock &hdev->lock uevent_sock_mutex rcu_read_lock &ei->socket.wq.wait &ep->lock irq_context: 0 (wq_completion)hci1 (work_completion)(&hdev->cmd_sync_work) &hdev->req_lock &hdev->lock hci_cb_list_lock rcu_state.exp_mutex &rq->__lock irq_context: 0 (wq_completion)hci5 (work_completion)(&hdev->cmd_sync_work) &hdev->req_lock &hdev->lock uevent_sock_mutex nl_table_wait.lock irq_context: 0 (wq_completion)hci1 (work_completion)(&hdev->cmd_sync_work) &hdev->req_lock &hdev->lock hci_cb_list_lock rcu_state.exp_mutex &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)hci5 (work_completion)(&hdev->cmd_sync_work) &hdev->req_lock &hdev->lock uevent_sock_mutex &rq->__lock irq_context: 0 (wq_completion)hci5 (work_completion)(&hdev->cmd_sync_work) &hdev->req_lock &hdev->lock uevent_sock_mutex &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)hci2 (work_completion)(&hdev->cmd_sync_work) &hdev->req_lock &hdev->lock (work_completion)(&(&conn->disc_work)->work) irq_context: 0 (wq_completion)hci2 (work_completion)(&hdev->cmd_sync_work) &hdev->req_lock &hdev->lock (work_completion)(&(&conn->disc_work)->work) &rq->__lock irq_context: 0 (wq_completion)hci1 (work_completion)(&hdev->cmd_sync_work) &hdev->req_lock &hdev->lock hci_cb_list_lock &list->lock#9 irq_context: 0 (wq_completion)hci1 (work_completion)(&hdev->cmd_sync_work) &hdev->req_lock &hdev->lock hci_cb_list_lock (work_completion)(&(&conn->info_timer)->work) irq_context: 0 (wq_completion)hci1 (work_completion)(&hdev->cmd_sync_work) &hdev->req_lock &hdev->lock hci_cb_list_lock rcu_read_lock &pool->lock irq_context: 0 (wq_completion)hci5 (work_completion)(&hdev->cmd_sync_work) &hdev->req_lock &hdev->lock uevent_sock_mutex &obj_hash[i].lock irq_context: 0 (wq_completion)hci1 (work_completion)(&hdev->cmd_sync_work) &hdev->req_lock &hdev->lock hci_cb_list_lock.wait_lock irq_context: 0 (wq_completion)hci1 (work_completion)(&hdev->cmd_sync_work) &hdev->req_lock &hdev->lock &p->pi_lock irq_context: 0 (wq_completion)hci1 (work_completion)(&hdev->cmd_sync_work) &hdev->req_lock &hdev->lock &p->pi_lock &rq->__lock irq_context: 0 (wq_completion)hci1 (work_completion)(&hdev->cmd_sync_work) &hdev->req_lock &hdev->lock &p->pi_lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)hci1 (work_completion)(&hdev->cmd_sync_work) &hdev->req_lock &hdev->lock &rq->__lock irq_context: 0 (wq_completion)hci4 (work_completion)(&hdev->cmd_sync_work) &hdev->req_lock &hdev->lock hci_cb_list_lock &list->lock#10 irq_context: 0 (wq_completion)hci4 (work_completion)(&hdev->cmd_sync_work) &hdev->req_lock &hdev->lock hci_cb_list_lock &obj_hash[i].lock irq_context: 0 (wq_completion)hci4 (work_completion)(&hdev->cmd_sync_work) &hdev->req_lock &hdev->lock hci_cb_list_lock (work_completion)(&(&conn->id_addr_timer)->work) irq_context: 0 (wq_completion)hci4 (work_completion)(&hdev->cmd_sync_work) &hdev->req_lock &hdev->lock hci_cb_list_lock &conn->chan_lock irq_context: 0 (wq_completion)hci4 (work_completion)(&hdev->cmd_sync_work) &hdev->req_lock &hdev->lock hci_cb_list_lock rcu_state.exp_mutex rcu_node_0 irq_context: 0 (wq_completion)hci4 (work_completion)(&hdev->cmd_sync_work) &hdev->req_lock &hdev->lock hci_cb_list_lock rcu_state.exp_mutex &obj_hash[i].lock irq_context: 0 (wq_completion)hci4 (work_completion)(&hdev->cmd_sync_work) &hdev->req_lock &hdev->lock hci_cb_list_lock rcu_state.exp_mutex rcu_read_lock &pool->lock irq_context: 0 (wq_completion)hci4 (work_completion)(&hdev->cmd_sync_work) &hdev->req_lock &hdev->lock hci_cb_list_lock rcu_state.exp_mutex rcu_read_lock &pool->lock &obj_hash[i].lock irq_context: 0 (wq_completion)hci4 (work_completion)(&hdev->cmd_sync_work) &hdev->req_lock &hdev->lock hci_cb_list_lock rcu_state.exp_mutex rcu_read_lock &pool->lock &p->pi_lock irq_context: 0 (wq_completion)hci4 (work_completion)(&hdev->cmd_sync_work) &hdev->req_lock &hdev->lock hci_cb_list_lock rcu_state.exp_mutex rcu_read_lock &pool->lock &p->pi_lock &rq->__lock irq_context: 0 (wq_completion)hci4 (work_completion)(&hdev->cmd_sync_work) &hdev->req_lock &hdev->lock hci_cb_list_lock rcu_state.exp_mutex rcu_read_lock &pool->lock &p->pi_lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)hci4 (work_completion)(&hdev->cmd_sync_work) &hdev->req_lock &hdev->lock hci_cb_list_lock rcu_state.exp_mutex &rnp->exp_wq[0] irq_context: 0 (wq_completion)hci4 (work_completion)(&hdev->cmd_sync_work) &hdev->req_lock &hdev->lock hci_cb_list_lock rcu_state.exp_mutex &rq->__lock irq_context: 0 (wq_completion)hci4 (work_completion)(&hdev->cmd_sync_work) &hdev->req_lock &hdev->lock hci_cb_list_lock rcu_state.exp_mutex &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)hci1 (work_completion)(&hdev->cmd_sync_work) &hdev->req_lock &hdev->lock (work_completion)(&(&conn->disc_work)->work) irq_context: 0 (wq_completion)hci1 (work_completion)(&hdev->cmd_sync_work) &hdev->req_lock &hdev->lock rcu_read_lock &pool->lock/1 irq_context: 0 (wq_completion)hci1 (work_completion)(&hdev->cmd_sync_work) &hdev->req_lock &hdev->lock (work_completion)(&(&conn->auto_accept_work)->work) irq_context: 0 (wq_completion)hci1 (work_completion)(&hdev->cmd_sync_work) &hdev->req_lock &hdev->lock (work_completion)(&(&conn->idle_work)->work) irq_context: 0 (wq_completion)hci1 (work_completion)(&hdev->cmd_sync_work) &hdev->req_lock &hdev->lock &list->lock#9 irq_context: 0 (wq_completion)hci1 (work_completion)(&hdev->cmd_sync_work) &hdev->req_lock &hdev->lock &rnp->exp_lock irq_context: 0 (wq_completion)hci1 (work_completion)(&hdev->cmd_sync_work) &hdev->req_lock &hdev->lock rcu_state.exp_mutex irq_context: 0 (wq_completion)hci1 (work_completion)(&hdev->cmd_sync_work) &hdev->req_lock &hdev->lock rcu_state.exp_mutex rcu_state.exp_mutex.wait_lock irq_context: 0 (wq_completion)hci1 (work_completion)(&hdev->cmd_sync_work) &hdev->req_lock &hdev->lock rcu_state.exp_mutex &rq->__lock irq_context: 0 (wq_completion)hci1 (work_completion)(&hdev->cmd_sync_work) &hdev->req_lock &hdev->lock rcu_state.exp_mutex &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)hci4 (work_completion)(&hdev->cmd_sync_work) &hdev->req_lock &hdev->lock hci_cb_list_lock rcu_state.exp_mutex.wait_lock irq_context: 0 (wq_completion)hci4 (work_completion)(&hdev->cmd_sync_work) &hdev->req_lock &hdev->lock hci_cb_list_lock &p->pi_lock irq_context: 0 (wq_completion)hci4 (work_completion)(&hdev->cmd_sync_work) &hdev->req_lock &hdev->lock hci_cb_list_lock &p->pi_lock &rq->__lock irq_context: 0 (wq_completion)hci4 (work_completion)(&hdev->cmd_sync_work) &hdev->req_lock &hdev->lock hci_cb_list_lock &p->pi_lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)hci4 (work_completion)(&hdev->cmd_sync_work) &hdev->req_lock &hdev->lock hci_cb_list_lock &list->lock#9 irq_context: 0 (wq_completion)hci4 (work_completion)(&hdev->cmd_sync_work) &hdev->req_lock &hdev->lock hci_cb_list_lock (work_completion)(&(&conn->info_timer)->work) irq_context: 0 (wq_completion)hci4 (work_completion)(&hdev->cmd_sync_work) &hdev->req_lock &hdev->lock hci_cb_list_lock rcu_read_lock &pool->lock irq_context: 0 (wq_completion)hci4 (work_completion)(&hdev->cmd_sync_work) &hdev->req_lock &hdev->lock hci_cb_list_lock.wait_lock irq_context: 0 (wq_completion)hci4 (work_completion)(&hdev->cmd_sync_work) &hdev->req_lock &hdev->lock &p->pi_lock irq_context: 0 (wq_completion)hci4 (work_completion)(&hdev->cmd_sync_work) &hdev->req_lock &hdev->lock &p->pi_lock &rq->__lock irq_context: 0 (wq_completion)hci4 (work_completion)(&hdev->cmd_sync_work) &hdev->req_lock &hdev->lock &p->pi_lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)hci4 (work_completion)(&hdev->cmd_sync_work) &hdev->req_lock &hdev->lock (work_completion)(&(&conn->disc_work)->work) irq_context: 0 (wq_completion)hci4 (work_completion)(&hdev->cmd_sync_work) &hdev->req_lock &hdev->lock rcu_read_lock &pool->lock/1 irq_context: 0 (wq_completion)hci4 (work_completion)(&hdev->cmd_sync_work) &hdev->req_lock &hdev->lock (work_completion)(&(&conn->auto_accept_work)->work) irq_context: 0 (wq_completion)hci4 (work_completion)(&hdev->cmd_sync_work) &hdev->req_lock &hdev->lock (work_completion)(&(&conn->idle_work)->work) irq_context: 0 (wq_completion)hci4 (work_completion)(&hdev->cmd_sync_work) &hdev->req_lock &hdev->lock &list->lock#9 irq_context: 0 (wq_completion)hci4 (work_completion)(&hdev->cmd_sync_work) &hdev->req_lock &hdev->lock &rnp->exp_lock irq_context: 0 (wq_completion)hci4 (work_completion)(&hdev->cmd_sync_work) &hdev->req_lock &hdev->lock &rnp->exp_wq[1] irq_context: 0 (wq_completion)hci4 (work_completion)(&hdev->cmd_sync_work) &hdev->req_lock &hdev->lock &rq->__lock irq_context: 0 (wq_completion)hci4 (work_completion)(&hdev->cmd_sync_work) &hdev->req_lock &hdev->lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)hci3 (work_completion)(&hdev->cmd_sync_work) &hdev->req_lock &hdev->lock hci_cb_list_lock &list->lock#10 irq_context: 0 (wq_completion)hci3 (work_completion)(&hdev->cmd_sync_work) &hdev->req_lock &hdev->lock hci_cb_list_lock &obj_hash[i].lock irq_context: 0 (wq_completion)hci3 (work_completion)(&hdev->cmd_sync_work) &hdev->req_lock &hdev->lock hci_cb_list_lock (work_completion)(&(&conn->id_addr_timer)->work) irq_context: 0 (wq_completion)hci3 (work_completion)(&hdev->cmd_sync_work) &hdev->req_lock &hdev->lock hci_cb_list_lock &conn->chan_lock irq_context: 0 (wq_completion)hci3 (work_completion)(&hdev->cmd_sync_work) &hdev->req_lock &hdev->lock hci_cb_list_lock &rnp->exp_lock irq_context: 0 (wq_completion)hci3 (work_completion)(&hdev->cmd_sync_work) &hdev->req_lock &hdev->lock hci_cb_list_lock &rnp->exp_wq[1] irq_context: 0 (wq_completion)hci1 (work_completion)(&hdev->cmd_sync_work) &hdev->req_lock &hdev->lock rcu_state.exp_mutex rcu_node_0 irq_context: 0 (wq_completion)hci1 (work_completion)(&hdev->cmd_sync_work) &hdev->req_lock &hdev->lock rcu_state.exp_mutex &obj_hash[i].lock irq_context: 0 (wq_completion)hci1 (work_completion)(&hdev->cmd_sync_work) &hdev->req_lock &hdev->lock rcu_state.exp_mutex rcu_read_lock &pool->lock irq_context: 0 (wq_completion)hci1 (work_completion)(&hdev->cmd_sync_work) &hdev->req_lock &hdev->lock rcu_state.exp_mutex rcu_read_lock &pool->lock &obj_hash[i].lock irq_context: 0 (wq_completion)hci1 (work_completion)(&hdev->cmd_sync_work) &hdev->req_lock &hdev->lock rcu_state.exp_mutex rcu_read_lock &pool->lock &p->pi_lock irq_context: 0 (wq_completion)hci1 (work_completion)(&hdev->cmd_sync_work) &hdev->req_lock &hdev->lock rcu_state.exp_mutex rcu_read_lock &pool->lock &p->pi_lock &rq->__lock irq_context: 0 (wq_completion)hci1 (work_completion)(&hdev->cmd_sync_work) &hdev->req_lock &hdev->lock rcu_state.exp_mutex rcu_read_lock &pool->lock &p->pi_lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)hci1 (work_completion)(&hdev->cmd_sync_work) &hdev->req_lock &hdev->lock rcu_state.exp_mutex &rnp->exp_wq[1] irq_context: 0 (wq_completion)hci2 (work_completion)(&hdev->cmd_sync_work) &hdev->req_lock &hdev->lock rcu_read_lock &pool->lock/1 irq_context: 0 (wq_completion)hci2 (work_completion)(&hdev->cmd_sync_work) &hdev->req_lock &hdev->lock rcu_read_lock rcu_node_0 irq_context: 0 (wq_completion)hci2 (work_completion)(&hdev->cmd_sync_work) &hdev->req_lock &hdev->lock rcu_read_lock &rq->__lock irq_context: 0 (wq_completion)hci2 (work_completion)(&hdev->cmd_sync_work) &hdev->req_lock &hdev->lock rcu_read_lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)hci2 (work_completion)(&hdev->cmd_sync_work) &hdev->req_lock &hdev->lock (work_completion)(&(&conn->auto_accept_work)->work) irq_context: 0 (wq_completion)hci2 (work_completion)(&hdev->cmd_sync_work) &hdev->req_lock &hdev->lock (work_completion)(&(&conn->idle_work)->work) irq_context: 0 (wq_completion)hci2 (work_completion)(&hdev->cmd_sync_work) &hdev->req_lock &hdev->lock &list->lock#9 irq_context: 0 (wq_completion)hci2 (work_completion)(&hdev->cmd_sync_work) &hdev->req_lock &hdev->lock &rnp->exp_lock irq_context: 0 (wq_completion)hci2 (work_completion)(&hdev->cmd_sync_work) &hdev->req_lock &hdev->lock rcu_state.exp_mutex irq_context: 0 (wq_completion)hci2 (work_completion)(&hdev->cmd_sync_work) &hdev->req_lock &hdev->lock rcu_state.exp_mutex rcu_state.exp_mutex.wait_lock irq_context: 0 (wq_completion)hci2 (work_completion)(&hdev->cmd_sync_work) &hdev->req_lock &hdev->lock rcu_state.exp_mutex &rq->__lock irq_context: 0 (wq_completion)hci2 (work_completion)(&hdev->cmd_sync_work) &hdev->req_lock &hdev->lock rcu_state.exp_mutex &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 &u->iolock &rcu_state.expedited_wq irq_context: 0 (wq_completion)hci3 (work_completion)(&hdev->cmd_sync_work) &hdev->req_lock &hdev->lock hci_cb_list_lock &list->lock#9 irq_context: 0 (wq_completion)hci3 (work_completion)(&hdev->cmd_sync_work) &hdev->req_lock &hdev->lock hci_cb_list_lock (work_completion)(&(&conn->info_timer)->work) irq_context: 0 (wq_completion)hci3 (work_completion)(&hdev->cmd_sync_work) &hdev->req_lock &hdev->lock hci_cb_list_lock rcu_read_lock &pool->lock irq_context: 0 (wq_completion)hci3 (work_completion)(&hdev->cmd_sync_work) &hdev->req_lock &hdev->lock hci_cb_list_lock.wait_lock irq_context: 0 (wq_completion)hci3 (work_completion)(&hdev->cmd_sync_work) &hdev->req_lock &hdev->lock &p->pi_lock irq_context: 0 (wq_completion)hci3 (work_completion)(&hdev->cmd_sync_work) &hdev->req_lock &hdev->lock &p->pi_lock &rq->__lock irq_context: 0 (wq_completion)hci3 (work_completion)(&hdev->cmd_sync_work) &hdev->req_lock &hdev->lock &p->pi_lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)hci3 (work_completion)(&hdev->cmd_sync_work) &hdev->req_lock &hdev->lock (work_completion)(&(&conn->disc_work)->work) irq_context: 0 (wq_completion)hci3 (work_completion)(&hdev->cmd_sync_work) &hdev->req_lock &hdev->lock rcu_read_lock &pool->lock/1 irq_context: 0 (wq_completion)hci3 (work_completion)(&hdev->cmd_sync_work) &hdev->req_lock &hdev->lock (work_completion)(&(&conn->auto_accept_work)->work) irq_context: 0 (wq_completion)hci3 (work_completion)(&hdev->cmd_sync_work) &hdev->req_lock &hdev->lock (work_completion)(&(&conn->idle_work)->work) irq_context: 0 (wq_completion)hci3 (work_completion)(&hdev->cmd_sync_work) &hdev->req_lock &hdev->lock &list->lock#9 irq_context: 0 (wq_completion)hci3 (work_completion)(&hdev->cmd_sync_work) &hdev->req_lock &hdev->lock &rnp->exp_lock irq_context: 0 (wq_completion)hci3 (work_completion)(&hdev->cmd_sync_work) &hdev->req_lock &hdev->lock &rnp->exp_wq[2] irq_context: 0 (wq_completion)hci3 (work_completion)(&hdev->cmd_sync_work) &hdev->req_lock &hdev->lock &rq->__lock irq_context: 0 (wq_completion)hci3 (work_completion)(&hdev->cmd_sync_work) &hdev->req_lock &hdev->lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)hci0 (work_completion)(&hdev->cmd_sync_work) &hdev->req_lock &hdev->lock hci_cb_list_lock &list->lock#10 irq_context: 0 (wq_completion)hci0 (work_completion)(&hdev->cmd_sync_work) &hdev->req_lock &hdev->lock hci_cb_list_lock &obj_hash[i].lock irq_context: 0 (wq_completion)hci0 (work_completion)(&hdev->cmd_sync_work) &hdev->req_lock &hdev->lock hci_cb_list_lock (work_completion)(&(&conn->id_addr_timer)->work) irq_context: 0 (wq_completion)hci0 (work_completion)(&hdev->cmd_sync_work) &hdev->req_lock &hdev->lock hci_cb_list_lock &conn->chan_lock irq_context: 0 (wq_completion)hci0 (work_completion)(&hdev->cmd_sync_work) &hdev->req_lock &hdev->lock hci_cb_list_lock &rnp->exp_lock irq_context: 0 (wq_completion)hci0 (work_completion)(&hdev->cmd_sync_work) &hdev->req_lock &hdev->lock hci_cb_list_lock &rnp->exp_wq[2] irq_context: 0 (wq_completion)hci4 (work_completion)(&hdev->cmd_sync_work) &hdev->req_lock &hdev->lock &k->k_lock irq_context: 0 (wq_completion)hci4 (work_completion)(&hdev->cmd_sync_work) &hdev->req_lock &hdev->lock &root->kernfs_rwsem irq_context: 0 (wq_completion)hci4 (work_completion)(&hdev->cmd_sync_work) &hdev->req_lock &hdev->lock &root->kernfs_rwsem irq_context: 0 (wq_completion)hci4 (work_completion)(&hdev->cmd_sync_work) &hdev->req_lock &hdev->lock dev_pm_qos_sysfs_mtx irq_context: 0 (wq_completion)hci4 (work_completion)(&hdev->cmd_sync_work) &hdev->req_lock &hdev->lock dev_pm_qos_sysfs_mtx &root->kernfs_rwsem irq_context: 0 (wq_completion)hci4 (work_completion)(&hdev->cmd_sync_work) &hdev->req_lock &hdev->lock dev_pm_qos_sysfs_mtx &root->kernfs_rwsem irq_context: 0 (wq_completion)hci4 (work_completion)(&hdev->cmd_sync_work) &hdev->req_lock &hdev->lock dev_pm_qos_sysfs_mtx dev_pm_qos_mtx irq_context: 0 (wq_completion)hci4 (work_completion)(&hdev->cmd_sync_work) &hdev->req_lock &hdev->lock &root->kernfs_rwsem &root->kernfs_iattr_rwsem irq_context: 0 (wq_completion)hci4 (work_completion)(&hdev->cmd_sync_work) &hdev->req_lock &hdev->lock &root->kernfs_rwsem kernfs_idr_lock irq_context: 0 (wq_completion)hci4 (work_completion)(&hdev->cmd_sync_work) &hdev->req_lock &hdev->lock &root->kernfs_rwsem &obj_hash[i].lock irq_context: 0 (wq_completion)hci4 (work_completion)(&hdev->cmd_sync_work) &hdev->req_lock &hdev->lock kernfs_idr_lock irq_context: 0 (wq_completion)hci4 (work_completion)(&hdev->cmd_sync_work) &hdev->req_lock &hdev->lock &k->k_lock klist_remove_lock irq_context: 0 (wq_completion)hci4 (work_completion)(&hdev->cmd_sync_work) &hdev->req_lock &hdev->lock &k->list_lock irq_context: 0 (wq_completion)hci4 (work_completion)(&hdev->cmd_sync_work) &hdev->req_lock &hdev->lock sysfs_symlink_target_lock irq_context: 0 (wq_completion)hci4 (work_completion)(&hdev->cmd_sync_work) &hdev->req_lock &hdev->lock subsys mutex#81 irq_context: 0 (wq_completion)hci4 (work_completion)(&hdev->cmd_sync_work) &hdev->req_lock &hdev->lock subsys mutex#81 &k->k_lock irq_context: 0 (wq_completion)hci4 (work_completion)(&hdev->cmd_sync_work) &hdev->req_lock &hdev->lock subsys mutex#81 &k->k_lock klist_remove_lock irq_context: 0 (wq_completion)hci4 (work_completion)(&hdev->cmd_sync_work) &hdev->req_lock &hdev->lock &x->wait#9 irq_context: 0 (wq_completion)hci4 (work_completion)(&hdev->cmd_sync_work) &hdev->req_lock &hdev->lock dpm_list_mtx irq_context: 0 (wq_completion)hci4 (work_completion)(&hdev->cmd_sync_work) &hdev->req_lock &hdev->lock &dev->power.lock irq_context: 0 (wq_completion)hci4 (work_completion)(&hdev->cmd_sync_work) &hdev->req_lock &hdev->lock deferred_probe_mutex irq_context: 0 (wq_completion)hci4 (work_completion)(&hdev->cmd_sync_work) &hdev->req_lock &hdev->lock device_links_lock irq_context: 0 (wq_completion)hci4 (work_completion)(&hdev->cmd_sync_work) &hdev->req_lock &hdev->lock mmu_notifier_invalidate_range_start irq_context: 0 (wq_completion)hci4 (work_completion)(&hdev->cmd_sync_work) &hdev->req_lock &hdev->lock uevent_sock_mutex irq_context: 0 (wq_completion)hci4 (work_completion)(&hdev->cmd_sync_work) &hdev->req_lock &hdev->lock uevent_sock_mutex mmu_notifier_invalidate_range_start irq_context: 0 (wq_completion)hci4 (work_completion)(&hdev->cmd_sync_work) &hdev->req_lock &hdev->lock uevent_sock_mutex nl_table_lock irq_context: 0 (wq_completion)hci4 (work_completion)(&hdev->cmd_sync_work) &hdev->req_lock &hdev->lock uevent_sock_mutex rlock-AF_NETLINK irq_context: 0 (wq_completion)hci4 (work_completion)(&hdev->cmd_sync_work) &hdev->req_lock &hdev->lock uevent_sock_mutex rcu_read_lock &ei->socket.wq.wait irq_context: 0 (wq_completion)hci4 (work_completion)(&hdev->cmd_sync_work) &hdev->req_lock &hdev->lock uevent_sock_mutex rcu_read_lock &ei->socket.wq.wait &ep->lock irq_context: 0 (wq_completion)hci4 (work_completion)(&hdev->cmd_sync_work) &hdev->req_lock &hdev->lock uevent_sock_mutex nl_table_wait.lock irq_context: 0 (wq_completion)hci4 (work_completion)(&hdev->cmd_sync_work) &hdev->req_lock &hdev->lock uevent_sock_mutex &obj_hash[i].lock irq_context: 0 (wq_completion)hci1 (work_completion)(&hdev->cmd_sync_work) &hdev->req_lock &hdev->lock rcu_state.exp_mutex.wait_lock irq_context: 0 (wq_completion)hci1 (work_completion)(&hdev->cmd_sync_work) &hdev->req_lock &hdev->lock &k->k_lock irq_context: 0 (wq_completion)hci1 (work_completion)(&hdev->cmd_sync_work) &hdev->req_lock &hdev->lock &root->kernfs_rwsem irq_context: 0 (wq_completion)hci1 (work_completion)(&hdev->cmd_sync_work) &hdev->req_lock &hdev->lock &root->kernfs_rwsem irq_context: 0 (wq_completion)hci1 (work_completion)(&hdev->cmd_sync_work) &hdev->req_lock &hdev->lock dev_pm_qos_sysfs_mtx irq_context: 0 (wq_completion)hci1 (work_completion)(&hdev->cmd_sync_work) &hdev->req_lock &hdev->lock dev_pm_qos_sysfs_mtx &root->kernfs_rwsem irq_context: 0 (wq_completion)hci1 (work_completion)(&hdev->cmd_sync_work) &hdev->req_lock &hdev->lock dev_pm_qos_sysfs_mtx &root->kernfs_rwsem irq_context: 0 (wq_completion)hci1 (work_completion)(&hdev->cmd_sync_work) &hdev->req_lock &hdev->lock dev_pm_qos_sysfs_mtx dev_pm_qos_mtx irq_context: 0 (wq_completion)hci1 (work_completion)(&hdev->cmd_sync_work) &hdev->req_lock &hdev->lock &root->kernfs_rwsem &root->kernfs_iattr_rwsem irq_context: 0 (wq_completion)hci1 (work_completion)(&hdev->cmd_sync_work) &hdev->req_lock &hdev->lock &root->kernfs_rwsem kernfs_idr_lock irq_context: 0 (wq_completion)hci1 (work_completion)(&hdev->cmd_sync_work) &hdev->req_lock &hdev->lock &root->kernfs_rwsem &obj_hash[i].lock irq_context: 0 (wq_completion)hci1 (work_completion)(&hdev->cmd_sync_work) &hdev->req_lock &hdev->lock kernfs_idr_lock irq_context: 0 (wq_completion)hci1 (work_completion)(&hdev->cmd_sync_work) &hdev->req_lock &hdev->lock &k->k_lock klist_remove_lock irq_context: 0 (wq_completion)hci1 (work_completion)(&hdev->cmd_sync_work) &hdev->req_lock &hdev->lock &k->list_lock irq_context: 0 (wq_completion)hci1 (work_completion)(&hdev->cmd_sync_work) &hdev->req_lock &hdev->lock sysfs_symlink_target_lock irq_context: 0 (wq_completion)hci1 (work_completion)(&hdev->cmd_sync_work) &hdev->req_lock &hdev->lock subsys mutex#81 irq_context: 0 (wq_completion)hci1 (work_completion)(&hdev->cmd_sync_work) &hdev->req_lock &hdev->lock subsys mutex#81 &k->k_lock irq_context: 0 (wq_completion)hci1 (work_completion)(&hdev->cmd_sync_work) &hdev->req_lock &hdev->lock subsys mutex#81 &k->k_lock klist_remove_lock irq_context: 0 (wq_completion)hci1 (work_completion)(&hdev->cmd_sync_work) &hdev->req_lock &hdev->lock &x->wait#9 irq_context: 0 (wq_completion)hci1 (work_completion)(&hdev->cmd_sync_work) &hdev->req_lock &hdev->lock dpm_list_mtx irq_context: 0 (wq_completion)hci1 (work_completion)(&hdev->cmd_sync_work) &hdev->req_lock &hdev->lock &dev->power.lock irq_context: 0 (wq_completion)hci1 (work_completion)(&hdev->cmd_sync_work) &hdev->req_lock &hdev->lock deferred_probe_mutex irq_context: 0 (wq_completion)hci1 (work_completion)(&hdev->cmd_sync_work) &hdev->req_lock &hdev->lock device_links_lock irq_context: 0 (wq_completion)hci1 (work_completion)(&hdev->cmd_sync_work) &hdev->req_lock &hdev->lock mmu_notifier_invalidate_range_start irq_context: 0 (wq_completion)hci1 (work_completion)(&hdev->cmd_sync_work) &hdev->req_lock &hdev->lock uevent_sock_mutex irq_context: 0 (wq_completion)hci1 (work_completion)(&hdev->cmd_sync_work) &hdev->req_lock &hdev->lock uevent_sock_mutex mmu_notifier_invalidate_range_start irq_context: 0 (wq_completion)hci1 (work_completion)(&hdev->cmd_sync_work) &hdev->req_lock &hdev->lock uevent_sock_mutex nl_table_lock irq_context: 0 (wq_completion)hci1 (work_completion)(&hdev->cmd_sync_work) &hdev->req_lock &hdev->lock uevent_sock_mutex rlock-AF_NETLINK irq_context: 0 (wq_completion)hci1 (work_completion)(&hdev->cmd_sync_work) &hdev->req_lock &hdev->lock uevent_sock_mutex rcu_read_lock &ei->socket.wq.wait irq_context: 0 (wq_completion)hci1 (work_completion)(&hdev->cmd_sync_work) &hdev->req_lock &hdev->lock uevent_sock_mutex rcu_read_lock &ei->socket.wq.wait &ep->lock irq_context: 0 (wq_completion)hci1 (work_completion)(&hdev->cmd_sync_work) &hdev->req_lock &hdev->lock uevent_sock_mutex nl_table_wait.lock irq_context: 0 (wq_completion)hci1 (work_completion)(&hdev->cmd_sync_work) &hdev->req_lock &hdev->lock uevent_sock_mutex &obj_hash[i].lock irq_context: 0 (wq_completion)hci2 (work_completion)(&hdev->cmd_sync_work) &hdev->req_lock &hdev->lock rcu_state.exp_mutex rcu_node_0 irq_context: 0 (wq_completion)hci2 (work_completion)(&hdev->cmd_sync_work) &hdev->req_lock &hdev->lock rcu_state.exp_mutex &obj_hash[i].lock irq_context: 0 (wq_completion)hci2 (work_completion)(&hdev->cmd_sync_work) &hdev->req_lock &hdev->lock rcu_state.exp_mutex rcu_read_lock &pool->lock irq_context: 0 (wq_completion)hci2 (work_completion)(&hdev->cmd_sync_work) &hdev->req_lock &hdev->lock rcu_state.exp_mutex rcu_read_lock &pool->lock &obj_hash[i].lock irq_context: 0 (wq_completion)hci2 (work_completion)(&hdev->cmd_sync_work) &hdev->req_lock &hdev->lock rcu_state.exp_mutex rcu_read_lock &pool->lock &p->pi_lock irq_context: 0 (wq_completion)hci2 (work_completion)(&hdev->cmd_sync_work) &hdev->req_lock &hdev->lock rcu_state.exp_mutex rcu_read_lock &pool->lock &p->pi_lock &rq->__lock irq_context: 0 (wq_completion)hci2 (work_completion)(&hdev->cmd_sync_work) &hdev->req_lock &hdev->lock rcu_state.exp_mutex rcu_read_lock &pool->lock &p->pi_lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)hci2 (work_completion)(&hdev->cmd_sync_work) &hdev->req_lock &hdev->lock rcu_state.exp_mutex &rnp->exp_wq[2] irq_context: 0 (wq_completion)hci3 (work_completion)(&hdev->cmd_sync_work) &hdev->req_lock &hdev->lock &k->k_lock irq_context: 0 (wq_completion)hci3 (work_completion)(&hdev->cmd_sync_work) &hdev->req_lock &hdev->lock &root->kernfs_rwsem irq_context: 0 (wq_completion)hci3 (work_completion)(&hdev->cmd_sync_work) &hdev->req_lock &hdev->lock &root->kernfs_rwsem irq_context: 0 (wq_completion)hci3 (work_completion)(&hdev->cmd_sync_work) &hdev->req_lock &hdev->lock dev_pm_qos_sysfs_mtx irq_context: 0 (wq_completion)hci3 (work_completion)(&hdev->cmd_sync_work) &hdev->req_lock &hdev->lock dev_pm_qos_sysfs_mtx &root->kernfs_rwsem irq_context: 0 (wq_completion)hci3 (work_completion)(&hdev->cmd_sync_work) &hdev->req_lock &hdev->lock dev_pm_qos_sysfs_mtx &root->kernfs_rwsem irq_context: 0 (wq_completion)hci3 (work_completion)(&hdev->cmd_sync_work) &hdev->req_lock &hdev->lock dev_pm_qos_sysfs_mtx dev_pm_qos_mtx irq_context: 0 (wq_completion)hci3 (work_completion)(&hdev->cmd_sync_work) &hdev->req_lock &hdev->lock &root->kernfs_rwsem &root->kernfs_iattr_rwsem irq_context: 0 (wq_completion)hci3 (work_completion)(&hdev->cmd_sync_work) &hdev->req_lock &hdev->lock &root->kernfs_rwsem kernfs_idr_lock irq_context: 0 (wq_completion)hci3 (work_completion)(&hdev->cmd_sync_work) &hdev->req_lock &hdev->lock &root->kernfs_rwsem &obj_hash[i].lock irq_context: 0 (wq_completion)hci3 (work_completion)(&hdev->cmd_sync_work) &hdev->req_lock &hdev->lock kernfs_idr_lock irq_context: 0 (wq_completion)hci3 (work_completion)(&hdev->cmd_sync_work) &hdev->req_lock &hdev->lock &k->k_lock klist_remove_lock irq_context: 0 (wq_completion)hci3 (work_completion)(&hdev->cmd_sync_work) &hdev->req_lock &hdev->lock &k->list_lock irq_context: 0 (wq_completion)hci3 (work_completion)(&hdev->cmd_sync_work) &hdev->req_lock &hdev->lock sysfs_symlink_target_lock irq_context: 0 (wq_completion)hci3 (work_completion)(&hdev->cmd_sync_work) &hdev->req_lock &hdev->lock subsys mutex#81 irq_context: 0 (wq_completion)hci3 (work_completion)(&hdev->cmd_sync_work) &hdev->req_lock &hdev->lock subsys mutex#81 &k->k_lock irq_context: 0 (wq_completion)hci3 (work_completion)(&hdev->cmd_sync_work) &hdev->req_lock &hdev->lock subsys mutex#81 &k->k_lock klist_remove_lock irq_context: 0 (wq_completion)hci3 (work_completion)(&hdev->cmd_sync_work) &hdev->req_lock &hdev->lock &x->wait#9 irq_context: 0 (wq_completion)hci3 (work_completion)(&hdev->cmd_sync_work) &hdev->req_lock &hdev->lock dpm_list_mtx irq_context: 0 (wq_completion)hci3 (work_completion)(&hdev->cmd_sync_work) &hdev->req_lock &hdev->lock &dev->power.lock irq_context: 0 (wq_completion)hci3 (work_completion)(&hdev->cmd_sync_work) &hdev->req_lock &hdev->lock deferred_probe_mutex irq_context: 0 (wq_completion)hci3 (work_completion)(&hdev->cmd_sync_work) &hdev->req_lock &hdev->lock device_links_lock irq_context: 0 (wq_completion)hci3 (work_completion)(&hdev->cmd_sync_work) &hdev->req_lock &hdev->lock mmu_notifier_invalidate_range_start irq_context: 0 (wq_completion)hci3 (work_completion)(&hdev->cmd_sync_work) &hdev->req_lock &hdev->lock uevent_sock_mutex irq_context: 0 (wq_completion)hci3 (work_completion)(&hdev->cmd_sync_work) &hdev->req_lock &hdev->lock uevent_sock_mutex mmu_notifier_invalidate_range_start irq_context: 0 (wq_completion)hci3 (work_completion)(&hdev->cmd_sync_work) &hdev->req_lock &hdev->lock uevent_sock_mutex nl_table_lock irq_context: 0 (wq_completion)hci3 (work_completion)(&hdev->cmd_sync_work) &hdev->req_lock &hdev->lock uevent_sock_mutex rlock-AF_NETLINK irq_context: 0 (wq_completion)hci3 (work_completion)(&hdev->cmd_sync_work) &hdev->req_lock &hdev->lock uevent_sock_mutex rcu_read_lock &ei->socket.wq.wait irq_context: 0 (wq_completion)hci3 (work_completion)(&hdev->cmd_sync_work) &hdev->req_lock &hdev->lock uevent_sock_mutex rcu_read_lock &ei->socket.wq.wait &ep->lock irq_context: 0 (wq_completion)hci3 (work_completion)(&hdev->cmd_sync_work) &hdev->req_lock &hdev->lock uevent_sock_mutex nl_table_wait.lock irq_context: 0 (wq_completion)hci3 (work_completion)(&hdev->cmd_sync_work) &hdev->req_lock &hdev->lock uevent_sock_mutex &obj_hash[i].lock irq_context: 0 (wq_completion)hci0 (work_completion)(&hdev->cmd_sync_work) &hdev->req_lock &hdev->lock hci_cb_list_lock &list->lock#9 irq_context: 0 (wq_completion)hci0 (work_completion)(&hdev->cmd_sync_work) &hdev->req_lock &hdev->lock hci_cb_list_lock (work_completion)(&(&conn->info_timer)->work) irq_context: 0 (wq_completion)hci0 (work_completion)(&hdev->cmd_sync_work) &hdev->req_lock &hdev->lock hci_cb_list_lock rcu_read_lock &pool->lock irq_context: 0 (wq_completion)hci0 (work_completion)(&hdev->cmd_sync_work) &hdev->req_lock &hdev->lock (work_completion)(&(&conn->disc_work)->work) irq_context: 0 (wq_completion)hci0 (work_completion)(&hdev->cmd_sync_work) &hdev->req_lock &hdev->lock rcu_read_lock &pool->lock/1 irq_context: 0 (wq_completion)hci0 (work_completion)(&hdev->cmd_sync_work) &hdev->req_lock &hdev->lock (work_completion)(&(&conn->auto_accept_work)->work) irq_context: 0 (wq_completion)hci0 (work_completion)(&hdev->cmd_sync_work) &hdev->req_lock &hdev->lock (work_completion)(&(&conn->idle_work)->work) irq_context: 0 (wq_completion)hci0 (work_completion)(&hdev->cmd_sync_work) &hdev->req_lock &hdev->lock &list->lock#9 irq_context: 0 (wq_completion)hci0 (work_completion)(&hdev->cmd_sync_work) &hdev->req_lock &hdev->lock &rnp->exp_lock irq_context: 0 (wq_completion)hci0 (work_completion)(&hdev->cmd_sync_work) &hdev->req_lock &hdev->lock rcu_state.exp_mutex irq_context: 0 (wq_completion)hci0 (work_completion)(&hdev->cmd_sync_work) &hdev->req_lock &hdev->lock rcu_state.exp_mutex rcu_state.exp_mutex.wait_lock irq_context: 0 (wq_completion)hci0 (work_completion)(&hdev->cmd_sync_work) &hdev->req_lock &hdev->lock rcu_state.exp_mutex &rq->__lock irq_context: 0 (wq_completion)hci0 (work_completion)(&hdev->cmd_sync_work) &hdev->req_lock &hdev->lock rcu_state.exp_mutex &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)hci2 (work_completion)(&hdev->cmd_sync_work) &hdev->req_lock &hdev->lock rcu_state.exp_mutex.wait_lock irq_context: 0 (wq_completion)hci0 (work_completion)(&hdev->cmd_sync_work) &hdev->req_lock &hdev->lock rcu_state.exp_mutex rcu_node_0 irq_context: 0 (wq_completion)hci0 (work_completion)(&hdev->cmd_sync_work) &hdev->req_lock &hdev->lock rcu_state.exp_mutex &obj_hash[i].lock irq_context: 0 (wq_completion)hci0 (work_completion)(&hdev->cmd_sync_work) &hdev->req_lock &hdev->lock rcu_state.exp_mutex rcu_read_lock &pool->lock irq_context: 0 (wq_completion)hci0 (work_completion)(&hdev->cmd_sync_work) &hdev->req_lock &hdev->lock rcu_state.exp_mutex rcu_read_lock &pool->lock &obj_hash[i].lock irq_context: 0 (wq_completion)hci0 (work_completion)(&hdev->cmd_sync_work) &hdev->req_lock &hdev->lock rcu_state.exp_mutex rcu_read_lock &pool->lock &p->pi_lock irq_context: 0 (wq_completion)hci0 (work_completion)(&hdev->cmd_sync_work) &hdev->req_lock &hdev->lock rcu_state.exp_mutex rcu_read_lock &pool->lock &p->pi_lock &rq->__lock irq_context: 0 (wq_completion)hci0 (work_completion)(&hdev->cmd_sync_work) &hdev->req_lock &hdev->lock rcu_state.exp_mutex rcu_read_lock &pool->lock &p->pi_lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)hci0 (work_completion)(&hdev->cmd_sync_work) &hdev->req_lock &hdev->lock rcu_state.exp_mutex &rnp->exp_wq[3] irq_context: 0 (wq_completion)hci2 (work_completion)(&hdev->cmd_sync_work) &hdev->req_lock &hdev->lock &k->k_lock irq_context: 0 (wq_completion)hci2 (work_completion)(&hdev->cmd_sync_work) &hdev->req_lock &hdev->lock &root->kernfs_rwsem irq_context: 0 (wq_completion)hci2 (work_completion)(&hdev->cmd_sync_work) &hdev->req_lock &hdev->lock &root->kernfs_rwsem irq_context: 0 (wq_completion)hci2 (work_completion)(&hdev->cmd_sync_work) &hdev->req_lock &hdev->lock dev_pm_qos_sysfs_mtx irq_context: 0 (wq_completion)hci2 (work_completion)(&hdev->cmd_sync_work) &hdev->req_lock &hdev->lock dev_pm_qos_sysfs_mtx &root->kernfs_rwsem irq_context: 0 (wq_completion)hci2 (work_completion)(&hdev->cmd_sync_work) &hdev->req_lock &hdev->lock dev_pm_qos_sysfs_mtx &root->kernfs_rwsem irq_context: 0 (wq_completion)hci2 (work_completion)(&hdev->cmd_sync_work) &hdev->req_lock &hdev->lock dev_pm_qos_sysfs_mtx dev_pm_qos_mtx irq_context: 0 (wq_completion)hci2 (work_completion)(&hdev->cmd_sync_work) &hdev->req_lock &hdev->lock &root->kernfs_rwsem &root->kernfs_iattr_rwsem irq_context: 0 (wq_completion)hci2 (work_completion)(&hdev->cmd_sync_work) &hdev->req_lock &hdev->lock &root->kernfs_rwsem kernfs_idr_lock irq_context: 0 (wq_completion)hci2 (work_completion)(&hdev->cmd_sync_work) &hdev->req_lock &hdev->lock &root->kernfs_rwsem &obj_hash[i].lock irq_context: 0 (wq_completion)hci2 (work_completion)(&hdev->cmd_sync_work) &hdev->req_lock &hdev->lock kernfs_idr_lock irq_context: 0 (wq_completion)hci2 (work_completion)(&hdev->cmd_sync_work) &hdev->req_lock &hdev->lock &k->k_lock klist_remove_lock irq_context: 0 (wq_completion)hci2 (work_completion)(&hdev->cmd_sync_work) &hdev->req_lock &hdev->lock &k->list_lock irq_context: 0 (wq_completion)hci2 (work_completion)(&hdev->cmd_sync_work) &hdev->req_lock &hdev->lock sysfs_symlink_target_lock irq_context: 0 (wq_completion)hci2 (work_completion)(&hdev->cmd_sync_work) &hdev->req_lock &hdev->lock subsys mutex#81 irq_context: 0 (wq_completion)hci2 (work_completion)(&hdev->cmd_sync_work) &hdev->req_lock &hdev->lock subsys mutex#81 &k->k_lock irq_context: 0 (wq_completion)hci2 (work_completion)(&hdev->cmd_sync_work) &hdev->req_lock &hdev->lock subsys mutex#81 &k->k_lock klist_remove_lock irq_context: 0 (wq_completion)hci2 (work_completion)(&hdev->cmd_sync_work) &hdev->req_lock &hdev->lock &x->wait#9 irq_context: 0 (wq_completion)hci2 (work_completion)(&hdev->cmd_sync_work) &hdev->req_lock &hdev->lock dpm_list_mtx irq_context: 0 (wq_completion)hci2 (work_completion)(&hdev->cmd_sync_work) &hdev->req_lock &hdev->lock &dev->power.lock irq_context: 0 (wq_completion)hci2 (work_completion)(&hdev->cmd_sync_work) &hdev->req_lock &hdev->lock deferred_probe_mutex irq_context: 0 (wq_completion)hci2 (work_completion)(&hdev->cmd_sync_work) &hdev->req_lock &hdev->lock device_links_lock irq_context: 0 (wq_completion)hci2 (work_completion)(&hdev->cmd_sync_work) &hdev->req_lock &hdev->lock mmu_notifier_invalidate_range_start irq_context: 0 (wq_completion)hci2 (work_completion)(&hdev->cmd_sync_work) &hdev->req_lock &hdev->lock &c->lock irq_context: 0 (wq_completion)hci2 (work_completion)(&hdev->cmd_sync_work) &hdev->req_lock &hdev->lock uevent_sock_mutex irq_context: 0 (wq_completion)hci2 (work_completion)(&hdev->cmd_sync_work) &hdev->req_lock &hdev->lock uevent_sock_mutex &rq->__lock irq_context: 0 (wq_completion)hci2 (work_completion)(&hdev->cmd_sync_work) &hdev->req_lock &hdev->lock uevent_sock_mutex &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 &f->f_pos_lock sb_writers#4 &sb->s_type->i_mutex_key#8 &rcu_state.expedited_wq &p->pi_lock &rq->__lock irq_context: 0 (wq_completion)hci2 (work_completion)(&hdev->cmd_sync_work) &hdev->req_lock &hdev->lock uevent_sock_mutex mmu_notifier_invalidate_range_start irq_context: 0 &f->f_pos_lock sb_writers#4 &sb->s_type->i_mutex_key#8 &rcu_state.expedited_wq &p->pi_lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)hci2 (work_completion)(&hdev->cmd_sync_work) &hdev->req_lock &hdev->lock uevent_sock_mutex nl_table_lock irq_context: 0 (wq_completion)hci0 (work_completion)(&hdev->cmd_sync_work) &hdev->req_lock &hdev->lock &k->k_lock irq_context: 0 (wq_completion)hci0 (work_completion)(&hdev->cmd_sync_work) &hdev->req_lock &hdev->lock &root->kernfs_rwsem irq_context: 0 (wq_completion)hci0 (work_completion)(&hdev->cmd_sync_work) &hdev->req_lock &hdev->lock &root->kernfs_rwsem irq_context: 0 (wq_completion)hci0 (work_completion)(&hdev->cmd_sync_work) &hdev->req_lock &hdev->lock dev_pm_qos_sysfs_mtx irq_context: 0 (wq_completion)hci0 (work_completion)(&hdev->cmd_sync_work) &hdev->req_lock &hdev->lock dev_pm_qos_sysfs_mtx &root->kernfs_rwsem irq_context: 0 (wq_completion)hci0 (work_completion)(&hdev->cmd_sync_work) &hdev->req_lock &hdev->lock dev_pm_qos_sysfs_mtx &root->kernfs_rwsem irq_context: 0 (wq_completion)hci0 (work_completion)(&hdev->cmd_sync_work) &hdev->req_lock &hdev->lock dev_pm_qos_sysfs_mtx dev_pm_qos_mtx irq_context: 0 (wq_completion)hci0 (work_completion)(&hdev->cmd_sync_work) &hdev->req_lock &hdev->lock &root->kernfs_rwsem &root->kernfs_iattr_rwsem irq_context: 0 (wq_completion)hci0 (work_completion)(&hdev->cmd_sync_work) &hdev->req_lock &hdev->lock &root->kernfs_rwsem kernfs_idr_lock irq_context: 0 (wq_completion)hci0 (work_completion)(&hdev->cmd_sync_work) &hdev->req_lock &hdev->lock &root->kernfs_rwsem &obj_hash[i].lock irq_context: 0 (wq_completion)hci0 (work_completion)(&hdev->cmd_sync_work) &hdev->req_lock &hdev->lock kernfs_idr_lock irq_context: 0 (wq_completion)hci0 (work_completion)(&hdev->cmd_sync_work) &hdev->req_lock &hdev->lock &k->k_lock klist_remove_lock irq_context: 0 (wq_completion)hci0 (work_completion)(&hdev->cmd_sync_work) &hdev->req_lock &hdev->lock &k->list_lock irq_context: 0 (wq_completion)hci0 (work_completion)(&hdev->cmd_sync_work) &hdev->req_lock &hdev->lock sysfs_symlink_target_lock irq_context: 0 (wq_completion)hci0 (work_completion)(&hdev->cmd_sync_work) &hdev->req_lock &hdev->lock subsys mutex#81 irq_context: 0 (wq_completion)hci0 (work_completion)(&hdev->cmd_sync_work) &hdev->req_lock &hdev->lock subsys mutex#81 &k->k_lock irq_context: 0 (wq_completion)hci0 (work_completion)(&hdev->cmd_sync_work) &hdev->req_lock &hdev->lock subsys mutex#81 &k->k_lock klist_remove_lock irq_context: 0 (wq_completion)hci0 (work_completion)(&hdev->cmd_sync_work) &hdev->req_lock &hdev->lock &x->wait#9 irq_context: 0 (wq_completion)hci0 (work_completion)(&hdev->cmd_sync_work) &hdev->req_lock &hdev->lock dpm_list_mtx irq_context: 0 (wq_completion)hci0 (work_completion)(&hdev->cmd_sync_work) &hdev->req_lock &hdev->lock &dev->power.lock irq_context: 0 (wq_completion)hci0 (work_completion)(&hdev->cmd_sync_work) &hdev->req_lock &hdev->lock deferred_probe_mutex irq_context: 0 (wq_completion)hci0 (work_completion)(&hdev->cmd_sync_work) &hdev->req_lock &hdev->lock device_links_lock irq_context: 0 (wq_completion)hci0 (work_completion)(&hdev->cmd_sync_work) &hdev->req_lock &hdev->lock mmu_notifier_invalidate_range_start irq_context: 0 (wq_completion)hci0 (work_completion)(&hdev->cmd_sync_work) &hdev->req_lock &hdev->lock &c->lock irq_context: 0 (wq_completion)hci0 (work_completion)(&hdev->cmd_sync_work) &hdev->req_lock &hdev->lock uevent_sock_mutex irq_context: 0 (wq_completion)hci0 (work_completion)(&hdev->cmd_sync_work) &hdev->req_lock &hdev->lock uevent_sock_mutex uevent_sock_mutex.wait_lock irq_context: 0 (wq_completion)hci0 (work_completion)(&hdev->cmd_sync_work) &hdev->req_lock &hdev->lock uevent_sock_mutex &rq->__lock irq_context: 0 (wq_completion)hci0 (work_completion)(&hdev->cmd_sync_work) &hdev->req_lock &hdev->lock uevent_sock_mutex &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)hci2 (work_completion)(&hdev->cmd_sync_work) &hdev->req_lock &hdev->lock uevent_sock_mutex &c->lock irq_context: 0 (wq_completion)hci2 (work_completion)(&hdev->cmd_sync_work) &hdev->req_lock &hdev->lock uevent_sock_mutex rlock-AF_NETLINK irq_context: 0 (wq_completion)hci2 (work_completion)(&hdev->cmd_sync_work) &hdev->req_lock &hdev->lock uevent_sock_mutex rcu_read_lock &ei->socket.wq.wait irq_context: 0 (wq_completion)hci2 (work_completion)(&hdev->cmd_sync_work) &hdev->req_lock &hdev->lock uevent_sock_mutex rcu_read_lock &ei->socket.wq.wait &ep->lock irq_context: 0 (wq_completion)hci2 (work_completion)(&hdev->cmd_sync_work) &hdev->req_lock &hdev->lock uevent_sock_mutex nl_table_wait.lock irq_context: 0 (wq_completion)hci2 (work_completion)(&hdev->cmd_sync_work) &hdev->req_lock &hdev->lock uevent_sock_mutex &obj_hash[i].lock irq_context: 0 (wq_completion)hci2 (work_completion)(&hdev->cmd_sync_work) &hdev->req_lock &hdev->lock uevent_sock_mutex.wait_lock irq_context: 0 (wq_completion)hci0 (work_completion)(&hdev->cmd_sync_work) &hdev->req_lock &hdev->lock uevent_sock_mutex mmu_notifier_invalidate_range_start irq_context: 0 (wq_completion)hci0 (work_completion)(&hdev->cmd_sync_work) &hdev->req_lock &hdev->lock uevent_sock_mutex nl_table_lock irq_context: 0 (wq_completion)hci0 (work_completion)(&hdev->cmd_sync_work) &hdev->req_lock &hdev->lock uevent_sock_mutex rlock-AF_NETLINK irq_context: 0 (wq_completion)hci0 (work_completion)(&hdev->cmd_sync_work) &hdev->req_lock &hdev->lock uevent_sock_mutex rcu_read_lock &ei->socket.wq.wait irq_context: 0 (wq_completion)hci0 (work_completion)(&hdev->cmd_sync_work) &hdev->req_lock &hdev->lock uevent_sock_mutex rcu_read_lock &ei->socket.wq.wait &ep->lock irq_context: 0 (wq_completion)hci0 (work_completion)(&hdev->cmd_sync_work) &hdev->req_lock &hdev->lock uevent_sock_mutex nl_table_wait.lock irq_context: 0 (wq_completion)hci0 (work_completion)(&hdev->cmd_sync_work) &hdev->req_lock &hdev->lock uevent_sock_mutex &obj_hash[i].lock irq_context: 0 (wq_completion)hci2 (work_completion)(&hdev->cmd_sync_work) &hdev->req_lock &hdev->lock &obj_hash[i].lock pool_lock irq_context: 0 &sb->s_type->i_mutex_key#10 &pcp->lock &zone->lock irq_context: 0 fs_reclaim pgd_lock irq_context: 0 fs_reclaim stock_lock irq_context: 0 fs_reclaim key irq_context: 0 fs_reclaim pcpu_lock irq_context: 0 fs_reclaim percpu_counters_lock irq_context: 0 fs_reclaim pcpu_lock stock_lock irq_context: 0 &ep->mtx &____s->seqcount#2 irq_context: 0 &ep->mtx &n->list_lock &c->lock irq_context: 0 &ep->mtx rcu_read_lock rcu_node_0 irq_context: 0 &ep->mtx rcu_read_lock &rq->__lock irq_context: 0 &ep->mtx rcu_read_lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)ext4-rsv-conversion (work_completion)(&ei->i_rsv_conversion_work) jbd2_handle &ei->i_data_sem &ei->i_es_lock &____s->seqcount#2 irq_context: 0 (wq_completion)ext4-rsv-conversion (work_completion)(&ei->i_rsv_conversion_work) jbd2_handle &ei->i_data_sem &ei->i_es_lock &____s->seqcount irq_context: 0 &u->iolock &cfs_rq->removed.lock irq_context: 0 &u->iolock &rq->__lock &cfs_rq->removed.lock irq_context: 0 sb_writers#4 sb_internal jbd2_handle &ei->i_data_sem mmu_notifier_invalidate_range_start &rq->__lock irq_context: 0 sb_writers#4 sb_internal jbd2_handle &ei->i_data_sem mmu_notifier_invalidate_range_start &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)wg-crypt-wg0 (work_completion)(&peer->transmit_packet_work) rcu_read_lock_bh rcu_read_lock &pool->lock/1 irq_context: 0 (wq_completion)wg-crypt-wg0 (work_completion)(&peer->transmit_packet_work) rcu_read_lock_bh rcu_read_lock &pool->lock/1 &obj_hash[i].lock irq_context: 0 (wq_completion)wg-crypt-wg0 (work_completion)(&peer->transmit_packet_work) rcu_read_lock_bh rcu_read_lock &pool->lock/1 pool_lock#2 irq_context: 0 (wq_completion)wg-crypt-wg0 (work_completion)(&peer->transmit_packet_work) rcu_read_lock_bh rcu_read_lock &pool->lock/1 &p->pi_lock irq_context: 0 (wq_completion)wg-kex-wg1 (work_completion)(&({ do { const void *__vpp_verify = (typeof((worker) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __asm__ ("" : "=r"(__ptr) : "0"((typeof(*((worker))) *)((worker)))); (typeof((typeof(*((worker))) *)((worker)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); })->work) &handshake->lock &c->lock irq_context: 0 sb_writers#4 &type->i_mutex_dir_key#3/1 sb_internal mmu_notifier_invalidate_range_start &rq->__lock irq_context: 0 sb_writers#4 &type->i_mutex_dir_key#3/1 sb_internal mmu_notifier_invalidate_range_start &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 &mm->mmap_lock sb_pagefaults remove_cache_srcu rcu_read_lock rcu_node_0 irq_context: 0 &mm->mmap_lock sb_pagefaults remove_cache_srcu rcu_read_lock &rq->__lock irq_context: 0 &mm->mmap_lock sb_pagefaults remove_cache_srcu rcu_read_lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)wg-kex-wg2#9 (work_completion)(&peer->transmit_handshake_work) &wg->static_identity.lock &handshake->lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 sb_writers#4 sb_internal jbd2_handle &sbi->s_orphan_lock mmu_notifier_invalidate_range_start irq_context: 0 sb_writers#4 sb_internal jbd2_handle &sbi->s_orphan_lock &____s->seqcount irq_context: 0 sb_writers#4 sb_internal jbd2_handle &sbi->s_orphan_lock pool_lock#2 irq_context: 0 sb_writers#4 lock#4 &lruvec->lru_lock per_cpu_ptr(&cgroup_rstat_cpu_lock, cpu) irq_context: 0 &ep->mtx wakeup_srcu_srcu_usage.lock rcu_read_lock &pool->lock &p->pi_lock irq_context: 0 &ep->mtx wakeup_srcu_srcu_usage.lock rcu_read_lock &pool->lock &p->pi_lock &rq->__lock irq_context: 0 &ep->mtx wakeup_srcu_srcu_usage.lock rcu_read_lock &pool->lock &p->pi_lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 wakeup_srcu_srcu_usage.lock rcu_read_lock &pool->lock &p->pi_lock irq_context: 0 wakeup_srcu_srcu_usage.lock rcu_read_lock &pool->lock &p->pi_lock &rq->__lock irq_context: 0 wakeup_srcu_srcu_usage.lock rcu_read_lock &pool->lock &p->pi_lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 sb_writers#4 &type->i_mutex_dir_key#3 tomoyo_ss rcu_read_lock &rq->__lock irq_context: 0 sb_writers#4 &type->i_mutex_dir_key#3 tomoyo_ss rcu_read_lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 sb_writers#4 &type->i_mutex_dir_key#3/1 jbd2_handle &ei->i_data_sem &ei->i_es_lock &____s->seqcount#2 irq_context: 0 sb_writers#4 &type->i_mutex_dir_key#3/1 jbd2_handle &ei->i_data_sem &ei->i_es_lock &____s->seqcount irq_context: 0 &p->lock fs_reclaim mmu_notifier_invalidate_range_start &rq->__lock irq_context: 0 &p->lock fs_reclaim mmu_notifier_invalidate_range_start &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 sb_writers#4 &type->i_mutex_dir_key#3 jbd2_handle key#4 irq_context: 0 cb_lock fs_reclaim &rq->__lock irq_context: 0 cb_lock fs_reclaim &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 &xt[i].mutex remove_cache_srcu &pcp->lock &zone->lock irq_context: 0 &xt[i].mutex remove_cache_srcu &pcp->lock &zone->lock &____s->seqcount irq_context: 0 (wq_completion)wg-crypt-wg2#2 (work_completion)(&peer->transmit_packet_work) rcu_read_lock_bh rcu_read_lock &pool->lock/1 irq_context: 0 (wq_completion)wg-crypt-wg2#2 (work_completion)(&peer->transmit_packet_work) rcu_read_lock_bh rcu_read_lock &pool->lock/1 &obj_hash[i].lock irq_context: 0 (wq_completion)wg-crypt-wg2#2 (work_completion)(&peer->transmit_packet_work) rcu_read_lock_bh rcu_read_lock &pool->lock/1 &p->pi_lock irq_context: softirq &(&net->ipv6.addr_chk_work)->timer rcu_read_lock &pool->lock pool_lock#2 irq_context: 0 (wq_completion)wg-kex-wg2#4 (work_completion)(&({ do { const void *__vpp_verify = (typeof((worker) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __asm__ ("" : "=r"(__ptr) : "0"((typeof(*((worker))) *)((worker)))); (typeof((typeof(*((worker))) *)((worker)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); })->work) &c->lock irq_context: 0 &sb->s_type->i_mutex_key#10 sk_lock-AF_AX25 irq_context: 0 &sb->s_type->i_mutex_key#10 sk_lock-AF_AX25 slock-AF_AX25 irq_context: 0 &sb->s_type->i_mutex_key#10 sk_lock-AF_AX25 clock-AF_AX25 irq_context: 0 &sb->s_type->i_mutex_key#10 sk_lock-AF_AX25 ax25_list_lock irq_context: 0 &sb->s_type->i_mutex_key#10 sk_lock-AF_AX25 &rq->__lock irq_context: 0 &sb->s_type->i_mutex_key#10 sk_lock-AF_AX25 &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 &sb->s_type->i_mutex_key#10 sk_lock-AF_AX25 &obj_hash[i].lock irq_context: 0 &sb->s_type->i_mutex_key#10 sk_lock-AF_AX25 &list->lock#35 irq_context: 0 &sb->s_type->i_mutex_key#10 sk_lock-AF_AX25 rlock-AF_AX25 irq_context: 0 &sb->s_type->i_mutex_key#10 sk_lock-AF_AX25 wlock-AF_AX25 irq_context: 0 &sb->s_type->i_mutex_key#10 slock-AF_AX25 irq_context: softirq rcu_read_lock rcu_read_lock rcu_read_lock &n->list_lock irq_context: softirq rcu_read_lock rcu_read_lock rcu_read_lock &n->list_lock &c->lock irq_context: 0 (wq_completion)events (work_completion)(&(&nsim_dev->trap_data->trap_report_dw)->work) &devlink->lock_key#3 &obj_hash[i].lock irq_context: 0 &lo->lo_mutex &rq->__lock irq_context: 0 &iint->mutex kauditd_wait.lock &p->pi_lock irq_context: 0 &iint->mutex kauditd_wait.lock &p->pi_lock &rq->__lock irq_context: 0 &iint->mutex kauditd_wait.lock &p->pi_lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 &iint->mutex &cfs_rq->removed.lock irq_context: 0 &iint->mutex &lock->wait_lock irq_context: 0 &iint->mutex rcu_read_lock pool_lock#2 irq_context: 0 &sig->cred_guard_mutex tomoyo_ss remove_cache_srcu pool_lock#2 irq_context: 0 &fsnotify_mark_srcu pgd_lock irq_context: 0 &fsnotify_mark_srcu stock_lock irq_context: 0 &fsnotify_mark_srcu rcu_read_lock pool_lock#2 irq_context: 0 &fsnotify_mark_srcu key irq_context: 0 &fsnotify_mark_srcu pcpu_lock irq_context: 0 &fsnotify_mark_srcu percpu_counters_lock irq_context: 0 &fsnotify_mark_srcu pcpu_lock stock_lock irq_context: 0 rtnl_mutex team->team_lock_key &____s->seqcount#2 irq_context: 0 (wq_completion)events (work_completion)(&w->work)#2 nf_conntrack_mutex &obj_hash[i].lock irq_context: 0 (wq_completion)mld (work_completion)(&(&idev->mc_ifc_work)->work) &idev->mc_lock rcu_read_lock &rq->__lock irq_context: 0 (wq_completion)mld (work_completion)(&(&idev->mc_ifc_work)->work) &idev->mc_lock rcu_read_lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)mld (work_completion)(&(&idev->mc_ifc_work)->work) &idev->mc_lock rcu_read_lock &rcu_state.gp_wq &p->pi_lock &rq->__lock irq_context: 0 (wq_completion)mld (work_completion)(&(&idev->mc_ifc_work)->work) &idev->mc_lock rcu_read_lock &rcu_state.gp_wq &p->pi_lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)events (linkwatch_work).work rtnl_mutex team->team_lock_key &n->list_lock irq_context: 0 (wq_completion)events (linkwatch_work).work rtnl_mutex team->team_lock_key &n->list_lock &c->lock irq_context: 0 (wq_completion)events (linkwatch_work).work rtnl_mutex &idev->mc_lock pool_lock#2 irq_context: 0 (wq_completion)events (linkwatch_work).work rtnl_mutex &idev->mc_lock krc.lock irq_context: 0 (wq_completion)events (linkwatch_work).work rtnl_mutex &idev->mc_lock _xmit_ETHER irq_context: 0 (wq_completion)events (linkwatch_work).work rtnl_mutex &idev->mc_lock _xmit_ETHER pool_lock#2 irq_context: 0 (wq_completion)events (linkwatch_work).work rtnl_mutex &idev->mc_lock &rq->__lock irq_context: 0 (wq_completion)events (linkwatch_work).work rtnl_mutex &idev->mc_lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 &sb->s_type->i_mutex_key#10 sk_lock-AF_INET slock-AF_INET fill_pool_map-wait-type-override &n->list_lock irq_context: 0 &sb->s_type->i_mutex_key#10 sk_lock-AF_INET slock-AF_INET fill_pool_map-wait-type-override &n->list_lock &c->lock irq_context: 0 (wq_completion)wg-crypt-wg2#5 (work_completion)(&peer->transmit_packet_work) rcu_read_lock_bh rcu_read_lock &pool->lock/1 irq_context: 0 (wq_completion)wg-crypt-wg2#5 (work_completion)(&peer->transmit_packet_work) rcu_read_lock_bh rcu_read_lock &pool->lock/1 &obj_hash[i].lock irq_context: 0 (wq_completion)wg-crypt-wg2#5 (work_completion)(&peer->transmit_packet_work) rcu_read_lock_bh rcu_read_lock &pool->lock/1 &p->pi_lock irq_context: 0 (wq_completion)wg-crypt-wg2#5 (work_completion)(&peer->transmit_packet_work) rcu_read_lock_bh rcu_read_lock &pool->lock/1 &p->pi_lock &rq->__lock irq_context: 0 (wq_completion)wg-crypt-wg2#5 (work_completion)(&peer->transmit_packet_work) rcu_read_lock_bh rcu_read_lock &pool->lock/1 &p->pi_lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)wg-kex-wg2#9 (work_completion)(&peer->transmit_handshake_work) &peer->endpoint_lock rcu_read_lock_bh &c->lock irq_context: 0 (wq_completion)wg-kex-wg2#9 (work_completion)(&peer->transmit_handshake_work) &peer->endpoint_lock rcu_read_lock_bh rcu_read_lock &____s->seqcount#9 irq_context: 0 (wq_completion)wg-kex-wg2#9 (work_completion)(&peer->transmit_handshake_work) &peer->endpoint_lock rcu_read_lock_bh rcu_read_lock rcu_read_lock_bh &obj_hash[i].lock irq_context: 0 (wq_completion)wg-kex-wg2#9 (work_completion)(&peer->transmit_handshake_work) &peer->endpoint_lock rcu_read_lock_bh rcu_read_lock rcu_read_lock_bh pool_lock#2 irq_context: 0 sb_writers#4 &type->i_mutex_dir_key#3 remove_cache_srcu irq_context: 0 sb_writers#4 &type->i_mutex_dir_key#3 remove_cache_srcu quarantine_lock irq_context: 0 &group->mark_mutex remove_cache_srcu rcu_read_lock rcu_node_0 irq_context: 0 &group->mark_mutex remove_cache_srcu rcu_read_lock &rq->__lock irq_context: 0 &group->mark_mutex remove_cache_srcu rcu_read_lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 &knet->mutex &rq->__lock irq_context: 0 &knet->mutex &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 &sig->cred_guard_mutex &sb->s_type->i_mutex_key#8 &rq->__lock irq_context: 0 &group->mark_mutex &n->list_lock irq_context: 0 &group->mark_mutex &n->list_lock &c->lock irq_context: 0 sb_writers#4 &type->i_mutex_dir_key#3/1 jbd2_handle &ei->i_data_sem rcu_read_lock &cfs_rq->removed.lock irq_context: 0 sb_writers#4 &type->i_mutex_dir_key#3/1 jbd2_handle &ei->i_data_sem rcu_read_lock &obj_hash[i].lock irq_context: 0 &sb->s_type->i_mutex_key#10 sk_lock-AF_INET &rq->__lock &cfs_rq->removed.lock irq_context: 0 sb_writers#4 jbd2_handle &journal->j_state_lock irq_context: 0 sb_writers#4 jbd2_handle &journal->j_state_lock &journal->j_wait_commit irq_context: 0 sb_writers#4 jbd2_handle &journal->j_state_lock &journal->j_wait_commit &p->pi_lock irq_context: 0 sb_writers#4 jbd2_handle &journal->j_state_lock &journal->j_wait_commit &p->pi_lock &rq->__lock irq_context: 0 sb_writers#4 jbd2_handle &journal->j_state_lock &journal->j_wait_commit &p->pi_lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 sb_writers#4 &sb->s_type->i_mutex_key#8 mapping.invalidate_lock jbd2_handle &journal->j_state_lock irq_context: 0 sb_writers#4 sb_internal remove_cache_srcu &cfs_rq->removed.lock irq_context: 0 sb_writers#4 &type->i_mutex_dir_key#3/1 &sb->s_type->i_mutex_key#8 jbd2_handle &rq->__lock &cfs_rq->removed.lock irq_context: 0 pernet_ops_rwsem ipvs->est_mutex pcpu_alloc_mutex &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 &iint->mutex ima_extend_list_mutex fs_reclaim mmu_notifier_invalidate_range_start &rq->__lock irq_context: 0 &iint->mutex ima_extend_list_mutex fs_reclaim mmu_notifier_invalidate_range_start &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 sb_writers#10 &type->i_mutex_dir_key#6/1 cgroup_mutex rcu_node_0 irq_context: 0 sb_writers#10 &type->i_mutex_dir_key#6/1 cgroup_mutex rcu_read_lock rcu_node_0 irq_context: 0 sb_writers#10 &type->i_mutex_dir_key#6/1 cgroup_mutex fill_pool_map-wait-type-override &n->list_lock irq_context: 0 sb_writers#10 &type->i_mutex_dir_key#6/1 cgroup_mutex fill_pool_map-wait-type-override &n->list_lock &c->lock irq_context: 0 sb_writers#10 &type->i_mutex_dir_key#6/1 cgroup_mutex fill_pool_map-wait-type-override &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 sb_writers#10 &type->i_mutex_dir_key#6 rename_lock.seqcount irq_context: 0 sb_writers#10 &type->i_mutex_dir_key#6 fs_reclaim irq_context: 0 sb_writers#10 &type->i_mutex_dir_key#6 fs_reclaim mmu_notifier_invalidate_range_start irq_context: 0 sb_writers#10 &type->i_mutex_dir_key#6 stock_lock irq_context: 0 sb_writers#10 &type->i_mutex_dir_key#6 &dentry->d_lock irq_context: 0 sb_writers#10 &type->i_mutex_dir_key#6 rcu_read_lock rename_lock.seqcount irq_context: 0 sb_writers#10 &type->i_mutex_dir_key#6 tomoyo_ss &n->list_lock irq_context: 0 sb_writers#10 &type->i_mutex_dir_key#6 tomoyo_ss &n->list_lock &c->lock irq_context: 0 sb_writers#10 &type->i_mutex_dir_key#6 tomoyo_ss &rq->__lock irq_context: 0 sb_writers#10 &type->i_mutex_dir_key#6 tomoyo_ss &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 sb_writers#10 &type->i_mutex_dir_key#6 &root->kernfs_iattr_rwsem irq_context: 0 sb_writers#10 &type->i_mutex_dir_key#6 &root->kernfs_rwsem irq_context: 0 sb_writers#10 &type->i_mutex_dir_key#6 &root->kernfs_rwsem inode_hash_lock irq_context: 0 sb_writers#10 &type->i_mutex_dir_key#6 &root->kernfs_rwsem fs_reclaim irq_context: 0 sb_writers#10 &type->i_mutex_dir_key#6 &root->kernfs_rwsem fs_reclaim mmu_notifier_invalidate_range_start irq_context: 0 sb_writers#10 &type->i_mutex_dir_key#6 &root->kernfs_rwsem stock_lock irq_context: 0 sb_writers#10 &type->i_mutex_dir_key#6 &root->kernfs_rwsem mmu_notifier_invalidate_range_start irq_context: 0 sb_writers#10 &type->i_mutex_dir_key#6 &root->kernfs_rwsem inode_hash_lock &sb->s_type->i_lock_key#30 irq_context: 0 sb_writers#10 &type->i_mutex_dir_key#6 &root->kernfs_rwsem inode_hash_lock &s->s_inode_list_lock irq_context: 0 sb_writers#10 &type->i_mutex_dir_key#6 &root->kernfs_rwsem tk_core.seq.seqcount irq_context: 0 sb_writers#10 &type->i_mutex_dir_key#6 &root->kernfs_rwsem &sb->s_type->i_lock_key#30 irq_context: 0 sb_writers#10 &type->i_mutex_dir_key#6 &sb->s_type->i_lock_key#30 irq_context: 0 sb_writers#10 &type->i_mutex_dir_key#6 &sb->s_type->i_lock_key#30 &dentry->d_lock irq_context: 0 sb_writers#10 &type->i_mutex_dir_key#6 &sb->s_type->i_lock_key#30 &dentry->d_lock &wq#2 irq_context: 0 sb_writers#10 &root->kernfs_iattr_rwsem irq_context: 0 sb_writers#10 &dentry->d_lock irq_context: 0 sb_writers#10 tomoyo_ss irq_context: 0 sb_writers#10 tomoyo_ss mmu_notifier_invalidate_range_start irq_context: 0 sb_writers#10 tomoyo_ss &c->lock irq_context: 0 sb_writers#10 tomoyo_ss rcu_read_lock mount_lock.seqcount irq_context: 0 sb_writers#10 tomoyo_ss rcu_read_lock rcu_read_lock rename_lock.seqcount irq_context: 0 sb_writers#10 tomoyo_ss &obj_hash[i].lock irq_context: 0 sb_writers#10 kn->active#58 &rq->__lock irq_context: 0 sb_writers#10 kn->active#58 &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 sb_writers#10 kn->active#58 fs_reclaim irq_context: 0 sb_writers#10 kn->active#58 fs_reclaim mmu_notifier_invalidate_range_start irq_context: 0 sb_writers#10 kn->active#58 stock_lock irq_context: 0 sb_writers#10 kn->active#58 &kernfs_locks->open_file_mutex[count] irq_context: 0 sb_writers#10 kn->active#58 &kernfs_locks->open_file_mutex[count] fs_reclaim irq_context: 0 sb_writers#10 kn->active#58 &kernfs_locks->open_file_mutex[count] fs_reclaim mmu_notifier_invalidate_range_start irq_context: 0 sb_writers#10 &sb->s_type->i_mutex_key#15 irq_context: 0 sb_writers#10 iattr_mutex irq_context: 0 sb_writers#10 iattr_mutex &rq->__lock irq_context: 0 sb_writers#10 iattr_mutex &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 sb_writers#10 &sb->s_type->i_mutex_key#15 tk_core.seq.seqcount irq_context: 0 sb_writers#10 &sb->s_type->i_mutex_key#15 &root->kernfs_iattr_rwsem irq_context: 0 sb_writers#10 &sb->s_type->i_mutex_key#15 &root->kernfs_iattr_rwsem iattr_mutex irq_context: 0 sb_writers#10 &sb->s_type->i_mutex_key#15 &root->kernfs_iattr_rwsem iattr_mutex fs_reclaim irq_context: 0 sb_writers#10 &sb->s_type->i_mutex_key#15 &root->kernfs_iattr_rwsem iattr_mutex fs_reclaim mmu_notifier_invalidate_range_start irq_context: 0 sb_writers#10 &sb->s_type->i_mutex_key#15 &root->kernfs_iattr_rwsem iattr_mutex tk_core.seq.seqcount irq_context: 0 &journal->j_barrier irq_context: 0 &journal->j_barrier &journal->j_state_lock irq_context: 0 &journal->j_barrier &journal->j_state_lock &journal->j_wait_commit irq_context: 0 &journal->j_barrier &journal->j_state_lock &journal->j_wait_commit &p->pi_lock irq_context: 0 &journal->j_barrier &journal->j_state_lock &journal->j_wait_commit &p->pi_lock &rq->__lock irq_context: 0 &journal->j_barrier &journal->j_state_lock &journal->j_wait_commit &p->pi_lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 &journal->j_barrier &rq->__lock irq_context: 0 &journal->j_barrier &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 sb_writers#4 &type->i_mutex_dir_key#3/1 &journal->j_wait_transaction_locked irq_context: 0 sb_writers#4 &sb->s_type->i_mutex_key#8 mapping.invalidate_lock &journal->j_wait_transaction_locked irq_context: 0 &journal->j_barrier &journal->j_state_lock irq_context: 0 &journal->j_barrier &journal->j_list_lock irq_context: 0 &journal->j_barrier &journal->j_checkpoint_mutex irq_context: 0 &journal->j_barrier &journal->j_checkpoint_mutex &journal->j_state_lock irq_context: 0 &journal->j_barrier &journal->j_checkpoint_mutex &journal->j_state_lock &journal->j_list_lock irq_context: 0 &journal->j_barrier &journal->j_checkpoint_mutex tk_core.seq.seqcount irq_context: 0 &journal->j_barrier &journal->j_checkpoint_mutex &fq->mq_flush_lock irq_context: 0 &journal->j_barrier &journal->j_checkpoint_mutex &fq->mq_flush_lock tk_core.seq.seqcount irq_context: 0 &journal->j_barrier &journal->j_checkpoint_mutex &fq->mq_flush_lock &q->requeue_lock irq_context: 0 &journal->j_barrier &journal->j_checkpoint_mutex &fq->mq_flush_lock &obj_hash[i].lock irq_context: 0 &journal->j_barrier &journal->j_checkpoint_mutex &fq->mq_flush_lock rcu_read_lock &pool->lock irq_context: 0 &journal->j_barrier &journal->j_checkpoint_mutex &fq->mq_flush_lock rcu_read_lock &pool->lock &obj_hash[i].lock irq_context: 0 &journal->j_barrier &journal->j_checkpoint_mutex &fq->mq_flush_lock rcu_read_lock &pool->lock pool_lock#2 irq_context: 0 &journal->j_barrier &journal->j_checkpoint_mutex &fq->mq_flush_lock rcu_read_lock &pool->lock &p->pi_lock irq_context: 0 &journal->j_barrier &journal->j_checkpoint_mutex &fq->mq_flush_lock rcu_read_lock &pool->lock &p->pi_lock &rq->__lock irq_context: 0 &journal->j_barrier &journal->j_checkpoint_mutex &fq->mq_flush_lock rcu_read_lock &pool->lock &p->pi_lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 &journal->j_barrier &journal->j_checkpoint_mutex &rq->__lock irq_context: 0 &journal->j_barrier &journal->j_checkpoint_mutex &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 &journal->j_barrier &journal->j_checkpoint_mutex &x->wait#26 irq_context: 0 &journal->j_barrier &journal->j_checkpoint_mutex mmu_notifier_invalidate_range_start irq_context: 0 &journal->j_barrier &journal->j_checkpoint_mutex pool_lock#2 irq_context: 0 &journal->j_barrier &journal->j_checkpoint_mutex &dd->lock irq_context: 0 &journal->j_barrier &journal->j_checkpoint_mutex &obj_hash[i].lock irq_context: 0 &journal->j_barrier &journal->j_checkpoint_mutex rcu_read_lock &pool->lock irq_context: 0 &journal->j_barrier &journal->j_checkpoint_mutex rcu_read_lock &pool->lock &obj_hash[i].lock irq_context: 0 &journal->j_barrier &journal->j_checkpoint_mutex rcu_read_lock &pool->lock pool_lock#2 irq_context: 0 &journal->j_barrier &journal->j_checkpoint_mutex rcu_read_lock &pool->lock &p->pi_lock irq_context: 0 &journal->j_barrier &journal->j_checkpoint_mutex rcu_read_lock &pool->lock &p->pi_lock &rq->__lock irq_context: 0 &journal->j_barrier &journal->j_checkpoint_mutex rcu_read_lock &pool->lock &p->pi_lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 &journal->j_barrier &journal->j_checkpoint_mutex bit_wait_table + i irq_context: 0 &journal->j_barrier &journal->j_checkpoint_mutex rcu_read_lock &rq->__lock irq_context: 0 &journal->j_barrier &journal->j_checkpoint_mutex rcu_read_lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 &journal->j_barrier &journal->j_checkpoint_mutex &journal->j_state_lock irq_context: 0 &journal->j_barrier &journal->j_checkpoint_mutex &journal->j_list_lock irq_context: 0 &journal->j_barrier &journal->j_checkpoint_mutex rcu_read_lock &dd->lock irq_context: 0 &journal->j_barrier &journal->j_checkpoint_mutex rcu_read_lock &obj_hash[i].lock irq_context: 0 &journal->j_barrier &journal->j_checkpoint_mutex rcu_read_lock tk_core.seq.seqcount irq_context: 0 &journal->j_barrier &journal->j_checkpoint_mutex rcu_read_lock &virtscsi_vq->vq_lock irq_context: 0 &journal->j_barrier &journal->j_checkpoint_mutex &journal->j_list_lock &obj_hash[i].lock irq_context: 0 &journal->j_barrier &journal->j_checkpoint_mutex &journal->j_list_lock pool_lock#2 irq_context: 0 &journal->j_barrier &journal->j_checkpoint_mutex &c->lock irq_context: 0 &journal->j_barrier &journal->j_checkpoint_mutex rcu_read_lock rcu_node_0 irq_context: 0 &journal->j_barrier &journal->j_checkpoint_mutex rcu_read_lock pool_lock#2 irq_context: 0 &journal->j_barrier &journal->j_checkpoint_mutex rcu_read_lock &obj_hash[i].lock pool_lock irq_context: 0 sb_writers#4 &journal->j_wait_transaction_locked irq_context: 0 &f->f_pos_lock sb_writers#10 irq_context: 0 &f->f_pos_lock sb_writers#10 fs_reclaim irq_context: 0 &f->f_pos_lock sb_writers#10 fs_reclaim &rq->__lock irq_context: 0 &f->f_pos_lock sb_writers#10 fs_reclaim &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 &f->f_pos_lock sb_writers#10 fs_reclaim mmu_notifier_invalidate_range_start irq_context: 0 &f->f_pos_lock sb_writers#10 &mm->mmap_lock irq_context: 0 &f->f_pos_lock sb_writers#10 &of->mutex irq_context: 0 &f->f_pos_lock sb_writers#10 &of->mutex cgroup_mutex irq_context: 0 &f->f_pos_lock sb_writers#10 &of->mutex cgroup_mutex css_set_lock irq_context: 0 &f->f_pos_lock sb_writers#10 &of->mutex cgroup_mutex css_set_lock cgroup_file_kn_lock irq_context: 0 &f->f_pos_lock sb_writers#10 &of->mutex cgroup_mutex css_set_lock cgroup_file_kn_lock kernfs_notify_lock irq_context: 0 &f->f_pos_lock sb_writers#10 &of->mutex cgroup_mutex css_set_lock cgroup_file_kn_lock kernfs_notify_lock rcu_read_lock &pool->lock irq_context: 0 &f->f_pos_lock sb_writers#10 &of->mutex cgroup_mutex css_set_lock cgroup_file_kn_lock kernfs_notify_lock rcu_read_lock &pool->lock &obj_hash[i].lock irq_context: 0 &f->f_pos_lock sb_writers#10 &of->mutex cgroup_mutex css_set_lock cgroup_file_kn_lock kernfs_notify_lock rcu_read_lock &pool->lock &p->pi_lock irq_context: 0 &f->f_pos_lock sb_writers#10 &of->mutex cgroup_mutex css_set_lock cgroup_file_kn_lock kernfs_notify_lock rcu_read_lock &pool->lock &p->pi_lock &rq->__lock irq_context: 0 &f->f_pos_lock sb_writers#10 &of->mutex cgroup_mutex css_set_lock cgroup_file_kn_lock kernfs_notify_lock rcu_read_lock &pool->lock &p->pi_lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 &f->f_pos_lock sb_writers#10 &of->mutex cgroup_mutex &rq->__lock irq_context: 0 &f->f_pos_lock sb_writers#10 &of->mutex cgroup_mutex &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 &f->f_pos_lock sb_writers#10 &obj_hash[i].lock irq_context: 0 sb_writers#10 &type->i_mutex_dir_key#6/1 &c->lock irq_context: 0 &f->f_pos_lock sb_writers#4 &sb->s_type->i_mutex_key#8 &journal->j_wait_transaction_locked irq_context: 0 &journal->j_barrier &journal->j_checkpoint_mutex &base->lock irq_context: 0 &journal->j_barrier &journal->j_checkpoint_mutex &base->lock &obj_hash[i].lock irq_context: 0 &journal->j_barrier &journal->j_checkpoint_mutex (&timer.timer) irq_context: 0 &journal->j_barrier &journal->j_checkpoint_mutex &ei->i_es_lock irq_context: 0 &journal->j_barrier &journal->j_checkpoint_mutex &ei->i_es_lock key#2 irq_context: 0 &journal->j_barrier &journal->j_checkpoint_mutex &mapping->private_lock irq_context: 0 &journal->j_barrier &journal->j_checkpoint_mutex &meta->lock irq_context: 0 &journal->j_barrier &journal->j_checkpoint_mutex kfence_freelist_lock irq_context: 0 &journal->j_barrier &journal->j_checkpoint_mutex &sb->s_type->i_lock_key#3 irq_context: 0 &journal->j_barrier &journal->j_checkpoint_mutex &sb->s_type->i_lock_key#3 &xa->xa_lock#9 irq_context: 0 &journal->j_barrier &journal->j_checkpoint_mutex lock#4 irq_context: 0 &journal->j_barrier &journal->j_checkpoint_mutex lock#4 &lruvec->lru_lock irq_context: 0 &journal->j_barrier &journal->j_checkpoint_mutex lock#5 irq_context: 0 &journal->j_barrier &journal->j_checkpoint_mutex &lruvec->lru_lock irq_context: 0 &journal->j_barrier &journal->j_checkpoint_mutex &sb->s_type->i_lock_key#3 &xa->xa_lock#9 &obj_hash[i].lock irq_context: 0 &journal->j_barrier &journal->j_checkpoint_mutex &sb->s_type->i_lock_key#3 &xa->xa_lock#9 pool_lock#2 irq_context: 0 &journal->j_barrier &journal->j_checkpoint_mutex rcu_read_lock rcu_read_lock rcu_node_0 irq_context: 0 &journal->j_barrier &journal->j_checkpoint_mutex rcu_read_lock rcu_read_lock &rq->__lock irq_context: 0 &journal->j_barrier &journal->j_checkpoint_mutex rcu_read_lock rcu_read_lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 &journal->j_barrier &journal->j_checkpoint_mutex &sb->s_type->i_lock_key#3 &xa->xa_lock#9 &obj_hash[i].lock pool_lock irq_context: 0 &journal->j_barrier &journal->j_checkpoint_mutex rcu_node_0 irq_context: 0 pernet_ops_rwsem rtnl_mutex uevent_sock_mutex quarantine_lock irq_context: 0 &journal->j_wait_transaction_locked irq_context: 0 &journal->j_wait_transaction_locked &p->pi_lock irq_context: 0 &journal->j_wait_transaction_locked &p->pi_lock &rq->__lock irq_context: 0 &journal->j_wait_transaction_locked &p->pi_lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 sb_writers#4 &sb->s_type->i_mutex_key#8 mapping.invalidate_lock jbd2_handle &obj_hash[i].lock irq_context: 0 sb_writers#4 &type->i_mutex_dir_key#3/1 &type->i_mutex_dir_key#3 jbd2_handle mmu_notifier_invalidate_range_start irq_context: 0 sb_writers#4 &type->i_mutex_dir_key#3/1 &type->i_mutex_dir_key#3 jbd2_handle pool_lock#2 irq_context: 0 pernet_ops_rwsem rtnl_mutex &root->kernfs_rwsem &root->kernfs_iattr_rwsem &sem->wait_lock irq_context: 0 pernet_ops_rwsem nf_ct_proto_mutex defrag4_mutex cpu_hotplug_lock &rq->__lock irq_context: 0 pernet_ops_rwsem nf_ct_proto_mutex defrag4_mutex cpu_hotplug_lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 pernet_ops_rwsem k-sk_lock-AF_RXRPC &rxnet->local_mutex rcu_read_lock &rq->__lock irq_context: 0 pernet_ops_rwsem k-sk_lock-AF_RXRPC &rxnet->local_mutex rcu_read_lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 sb_writers#4 &type->i_mutex_dir_key#3/1 &sb->s_type->i_mutex_key#8 jbd2_handle rcu_read_lock &cfs_rq->removed.lock irq_context: 0 sb_writers#4 &type->i_mutex_dir_key#3/1 &sb->s_type->i_mutex_key#8 jbd2_handle rcu_read_lock &obj_hash[i].lock irq_context: 0 rcu_state.exp_mutex &rq->__lock &cfs_rq->removed.lock irq_context: 0 sb_writers#5 &type->i_mutex_dir_key#5 remove_cache_srcu rcu_node_0 irq_context: 0 sk_lock-AF_BLUETOOTH-BTPROTO_SCO irq_context: 0 sk_lock-AF_BLUETOOTH-BTPROTO_SCO slock-AF_BLUETOOTH-BTPROTO_SCO irq_context: 0 sk_lock-AF_BLUETOOTH-BTPROTO_SCO &mm->mmap_lock irq_context: 0 &type->s_umount_key#23/1 pcpu_alloc_mutex &rq->__lock irq_context: 0 slock-AF_BLUETOOTH-BTPROTO_SCO irq_context: 0 &type->s_umount_key#23/1 pcpu_alloc_mutex &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 sb_writers#5 rcu_node_0 irq_context: 0 sb_writers#4 &type->i_mutex_dir_key#3/1 &type->i_mutex_dir_key#3 jbd2_handle &c->lock irq_context: 0 sb_writers#4 &type->i_mutex_dir_key#3/1 &type->i_mutex_dir_key#3 jbd2_handle rcu_node_0 irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem rtnl_mutex rcu_node_0 irq_context: 0 sb_writers#4 &type->i_mutex_dir_key#3 rcu_node_0 irq_context: 0 sb_writers#10 &type->i_mutex_dir_key#6/1 cgroup_mutex fs_reclaim &rq->__lock irq_context: 0 sb_writers#10 &type->i_mutex_dir_key#6/1 cgroup_mutex fs_reclaim &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 &type->i_mutex_dir_key#7 rename_lock.seqcount irq_context: 0 &type->i_mutex_dir_key#7 &dentry->d_lock &wq#2 irq_context: 0 sb_writers#10 kn->active#58 rcu_read_lock rcu_node_0 irq_context: 0 sb_writers#10 kn->active#58 rcu_read_lock &rq->__lock irq_context: 0 sb_writers#10 kn->active#58 rcu_read_lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 sb_writers#10 kn->active#58 rcu_read_lock &rcu_state.expedited_wq irq_context: 0 sb_writers#10 kn->active#58 rcu_read_lock &rcu_state.expedited_wq &p->pi_lock irq_context: 0 sb_writers#10 kn->active#58 rcu_read_lock &rcu_state.expedited_wq &p->pi_lock &rq->__lock irq_context: 0 sb_writers#10 kn->active#58 rcu_read_lock &rcu_state.expedited_wq &p->pi_lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 sb_writers#10 kn->active#58 &c->lock irq_context: 0 sb_writers#10 kn->active#58 &n->list_lock irq_context: 0 sb_writers#10 kn->active#58 &n->list_lock &c->lock irq_context: 0 sb_writers#10 tomoyo_ss &____s->seqcount#2 irq_context: 0 sb_writers#10 tomoyo_ss &____s->seqcount irq_context: 0 sb_writers#10 tomoyo_ss &rq->__lock irq_context: 0 sb_writers#10 tomoyo_ss &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 sb_writers#10 tomoyo_ss &n->list_lock irq_context: 0 sb_writers#10 tomoyo_ss &n->list_lock &c->lock irq_context: 0 &journal->j_barrier jbd2_handle irq_context: 0 &journal->j_barrier &journal->j_wait_commit irq_context: 0 &journal->j_barrier &journal->j_wait_done_commit irq_context: 0 sb_writers#10 &type->i_mutex_dir_key#6/1 cgroup_mutex pcpu_alloc_mutex &rq->__lock irq_context: 0 sb_writers#10 &type->i_mutex_dir_key#6/1 cgroup_mutex pcpu_alloc_mutex &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 pernet_ops_rwsem rtnl_mutex pcpu_alloc_mutex pcpu_alloc_mutex.wait_lock irq_context: 0 pernet_ops_rwsem rdma_nets_rwsem rdma_nets_rwsem.wait_lock irq_context: 0 pernet_ops_rwsem rdma_nets_rwsem &rq->__lock irq_context: 0 pernet_ops_rwsem rdma_nets_rwsem &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 pernet_ops_rwsem devices_rwsem &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem rdma_nets_rwsem.wait_lock irq_context: 0 &f->f_pos_lock sb_writers#4 &sb->s_type->i_mutex_key#8 &ei->xattr_sem &rq->__lock irq_context: 0 &f->f_pos_lock sb_writers#4 &sb->s_type->i_mutex_key#8 &ei->xattr_sem &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem &root->kernfs_rwsem rcu_read_lock rcu_node_0 irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem &root->kernfs_rwsem &meta->lock irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem &root->kernfs_rwsem kfence_freelist_lock irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem &root->kernfs_rwsem &p->pi_lock irq_context: 0 &sbi->s_writepages_rwsem remove_cache_srcu pool_lock#2 irq_context: 0 (wq_completion)events free_ipc_work &xa->xa_lock#4 &obj_hash[i].lock pool_lock irq_context: 0 (wq_completion)events free_ipc_work &rq->__lock irq_context: 0 (wq_completion)events free_ipc_work &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: softirq (&ndev->rs_timer) rcu_read_lock rcu_read_lock rcu_read_lock_bh rcu_read_lock_bh rcu_read_lock &pool->lock fill_pool_map-wait-type-override &c->lock irq_context: softirq (&ndev->rs_timer) rcu_read_lock rcu_read_lock fill_pool_map-wait-type-override pool_lock#2 irq_context: softirq (&ndev->rs_timer) rcu_read_lock rcu_read_lock rcu_read_lock_bh rcu_read_lock_bh rcu_read_lock &pool->lock fill_pool_map-wait-type-override pool_lock#2 irq_context: softirq (&ndev->rs_timer) rcu_read_lock rcu_read_lock fill_pool_map-wait-type-override pool_lock irq_context: softirq (&ndev->rs_timer) rcu_read_lock rcu_read_lock rcu_read_lock_bh rcu_read_lock_bh rcu_read_lock &pool->lock fill_pool_map-wait-type-override &n->list_lock irq_context: softirq (&ndev->rs_timer) rcu_read_lock rcu_read_lock rcu_read_lock_bh rcu_read_lock_bh rcu_read_lock &pool->lock fill_pool_map-wait-type-override &n->list_lock &c->lock irq_context: softirq (&ndev->rs_timer) rcu_read_lock rcu_read_lock rcu_read_lock_bh rcu_read_lock_bh rcu_read_lock &pool->lock fill_pool_map-wait-type-override pool_lock irq_context: 0 (wq_completion)wg-kex-wg0#7 (work_completion)(&peer->transmit_handshake_work) &peer->endpoint_lock rcu_read_lock_bh &obj_hash[i].lock irq_context: 0 (wq_completion)wg-kex-wg0#7 (work_completion)(&peer->transmit_handshake_work) &peer->endpoint_lock rcu_read_lock_bh rcu_read_lock rcu_read_lock_bh &obj_hash[i].lock irq_context: 0 (wq_completion)wg-crypt-wg2#5 (work_completion)(&peer->transmit_packet_work) &peer->endpoint_lock rcu_read_lock_bh pool_lock#2 irq_context: 0 (wq_completion)wg-crypt-wg2 (work_completion)(&peer->transmit_packet_work) rcu_read_lock_bh rcu_read_lock &pool->lock/1 irq_context: 0 (wq_completion)wg-crypt-wg2 (work_completion)(&peer->transmit_packet_work) rcu_read_lock_bh rcu_read_lock &pool->lock/1 &obj_hash[i].lock irq_context: 0 (wq_completion)wg-crypt-wg2 (work_completion)(&peer->transmit_packet_work) rcu_read_lock_bh rcu_read_lock &pool->lock/1 &p->pi_lock irq_context: 0 (wq_completion)wg-kex-wg1 (work_completion)(&({ do { const void *__vpp_verify = (typeof((worker) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __asm__ ("" : "=r"(__ptr) : "0"((typeof(*((worker))) *)((worker)))); (typeof((typeof(*((worker))) *)((worker)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); })->work) &base->lock irq_context: 0 (wq_completion)wg-kex-wg1 (work_completion)(&({ do { const void *__vpp_verify = (typeof((worker) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __asm__ ("" : "=r"(__ptr) : "0"((typeof(*((worker))) *)((worker)))); (typeof((typeof(*((worker))) *)((worker)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); })->work) &base->lock &obj_hash[i].lock irq_context: 0 (wq_completion)wg-kex-wg2#2 (work_completion)(&({ do { const void *__vpp_verify = (typeof((worker) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __asm__ ("" : "=r"(__ptr) : "0"((typeof(*((worker))) *)((worker)))); (typeof((typeof(*((worker))) *)((worker)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); })->work) rcu_read_lock_bh rcu_read_lock &pool->lock &p->pi_lock &rq->__lock irq_context: 0 (wq_completion)wg-kex-wg2#2 (work_completion)(&({ do { const void *__vpp_verify = (typeof((worker) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __asm__ ("" : "=r"(__ptr) : "0"((typeof(*((worker))) *)((worker)))); (typeof((typeof(*((worker))) *)((worker)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); })->work) rcu_read_lock_bh rcu_read_lock &pool->lock &p->pi_lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 cb_lock genl_mutex rcu_read_lock &____s->seqcount#2 irq_context: 0 cb_lock genl_mutex rcu_read_lock &____s->seqcount irq_context: 0 (wq_completion)wg-crypt-wg0#4 (work_completion)(&peer->transmit_packet_work) &peer->endpoint_lock rcu_read_lock_bh batched_entropy_u8.lock irq_context: 0 (wq_completion)wg-crypt-wg0#4 (work_completion)(&peer->transmit_packet_work) &peer->endpoint_lock rcu_read_lock_bh batched_entropy_u8.lock crngs.lock irq_context: 0 (wq_completion)wg-crypt-wg0#4 (work_completion)(&peer->transmit_packet_work) &peer->endpoint_lock rcu_read_lock_bh rcu_read_lock rcu_read_lock_bh &meta->lock irq_context: 0 (wq_completion)wg-crypt-wg0#4 (work_completion)(&peer->transmit_packet_work) &peer->endpoint_lock rcu_read_lock_bh rcu_read_lock rcu_read_lock_bh kfence_freelist_lock irq_context: 0 &mm->mmap_lock sb_pagefaults remove_cache_srcu &____s->seqcount irq_context: 0 &mm->mmap_lock sb_pagefaults remove_cache_srcu pool_lock#2 irq_context: 0 &mm->mmap_lock sb_pagefaults remove_cache_srcu &rq->__lock irq_context: 0 &mm->mmap_lock sb_pagefaults remove_cache_srcu &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)wg-crypt-wg0#2 (work_completion)(&peer->transmit_packet_work) &peer->endpoint_lock rcu_read_lock_bh rcu_read_lock &c->lock irq_context: 0 (wq_completion)wg-crypt-wg0#2 (work_completion)(&peer->transmit_packet_work) &peer->endpoint_lock rcu_read_lock_bh rcu_read_lock &n->list_lock irq_context: 0 (wq_completion)wg-crypt-wg0#2 (work_completion)(&peer->transmit_packet_work) &peer->endpoint_lock rcu_read_lock_bh rcu_read_lock &n->list_lock &c->lock irq_context: 0 sb_writers#4 rcu_read_lock &rq->__lock &cfs_rq->removed.lock irq_context: 0 (wq_completion)bat_events (work_completion)(&(&bat_priv->bla.work)->work) rcu_read_lock &pcp->lock &zone->lock irq_context: 0 pernet_ops_rwsem sysctl_lock fill_pool_map-wait-type-override &c->lock irq_context: 0 pernet_ops_rwsem sysctl_lock fill_pool_map-wait-type-override &n->list_lock irq_context: 0 pernet_ops_rwsem sysctl_lock fill_pool_map-wait-type-override &n->list_lock &c->lock irq_context: 0 sb_writers#10 &type->i_mutex_dir_key#6 rcu_read_lock &dentry->d_lock irq_context: 0 sb_writers#10 &xattrs->lock irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem &net->ipv6.ip6addrlbl_table.lock fill_pool_map-wait-type-override pool_lock#2 irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem &net->ipv6.ip6addrlbl_table.lock fill_pool_map-wait-type-override &c->lock irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem &net->ipv6.ip6addrlbl_table.lock fill_pool_map-wait-type-override pool_lock irq_context: 0 &journal->j_barrier &cfs_rq->removed.lock irq_context: 0 &journal->j_barrier &obj_hash[i].lock irq_context: 0 &journal->j_barrier pool_lock#2 irq_context: 0 &journal->j_barrier &journal->j_checkpoint_mutex rcu_read_lock fill_pool_map-wait-type-override pool_lock#2 irq_context: 0 &journal->j_barrier &journal->j_checkpoint_mutex rcu_read_lock fill_pool_map-wait-type-override &c->lock irq_context: 0 &journal->j_barrier &journal->j_checkpoint_mutex rcu_read_lock fill_pool_map-wait-type-override &n->list_lock irq_context: 0 &journal->j_barrier &journal->j_checkpoint_mutex rcu_read_lock fill_pool_map-wait-type-override &n->list_lock &c->lock irq_context: 0 &journal->j_barrier &journal->j_checkpoint_mutex rcu_read_lock fill_pool_map-wait-type-override rcu_node_0 irq_context: 0 &journal->j_barrier &journal->j_checkpoint_mutex rcu_read_lock fill_pool_map-wait-type-override &rq->__lock irq_context: 0 &journal->j_barrier &journal->j_checkpoint_mutex rcu_read_lock fill_pool_map-wait-type-override &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 &journal->j_barrier &journal->j_checkpoint_mutex rcu_read_lock fill_pool_map-wait-type-override pool_lock irq_context: 0 &journal->j_barrier &journal->j_checkpoint_mutex rcu_read_lock &rcu_state.expedited_wq irq_context: 0 &journal->j_barrier &journal->j_checkpoint_mutex rcu_read_lock &rcu_state.expedited_wq &p->pi_lock irq_context: 0 &journal->j_barrier &journal->j_checkpoint_mutex rcu_read_lock &rcu_state.expedited_wq &p->pi_lock &rq->__lock irq_context: 0 &journal->j_barrier &journal->j_checkpoint_mutex rcu_read_lock &rcu_state.expedited_wq &p->pi_lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem uevent_sock_mutex &rq->__lock irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem uevent_sock_mutex &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 &mm->mmap_lock sb_pagefaults &journal->j_wait_transaction_locked irq_context: 0 &f->f_pos_lock sb_writers#10 remove_cache_srcu irq_context: 0 &f->f_pos_lock sb_writers#10 remove_cache_srcu quarantine_lock irq_context: 0 &f->f_pos_lock sb_writers#10 remove_cache_srcu &c->lock irq_context: 0 &f->f_pos_lock sb_writers#10 remove_cache_srcu &n->list_lock irq_context: 0 &f->f_pos_lock sb_writers#10 remove_cache_srcu &obj_hash[i].lock irq_context: 0 &f->f_pos_lock sb_writers#10 remove_cache_srcu &rq->__lock irq_context: 0 &f->f_pos_lock sb_writers#10 remove_cache_srcu &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 sb_writers#10 &type->i_mutex_dir_key#6/1 &dentry->d_lock &dentry->d_lock &lru->node[i].lock irq_context: 0 sb_writers#10 &type->i_mutex_dir_key#6/1 &root->kernfs_rwsem inode_hash_lock irq_context: 0 sb_writers#10 &type->i_mutex_dir_key#6/1 &root->kernfs_rwsem fs_reclaim irq_context: 0 sb_writers#10 &type->i_mutex_dir_key#6/1 &root->kernfs_rwsem fs_reclaim mmu_notifier_invalidate_range_start irq_context: 0 sb_writers#10 &type->i_mutex_dir_key#6/1 &root->kernfs_rwsem stock_lock irq_context: 0 sb_writers#10 &type->i_mutex_dir_key#6/1 &root->kernfs_rwsem &c->lock irq_context: 0 sb_writers#10 &type->i_mutex_dir_key#6/1 &root->kernfs_rwsem mmu_notifier_invalidate_range_start irq_context: 0 sb_writers#10 &type->i_mutex_dir_key#6/1 &root->kernfs_rwsem inode_hash_lock &sb->s_type->i_lock_key#30 irq_context: 0 sb_writers#10 &type->i_mutex_dir_key#6/1 &root->kernfs_rwsem inode_hash_lock &s->s_inode_list_lock irq_context: 0 sb_writers#10 &type->i_mutex_dir_key#6/1 &root->kernfs_rwsem &rq->__lock irq_context: 0 sb_writers#10 &type->i_mutex_dir_key#6/1 &root->kernfs_rwsem &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 sb_writers#10 &type->i_mutex_dir_key#6/1 &root->kernfs_rwsem tk_core.seq.seqcount irq_context: 0 sb_writers#10 &type->i_mutex_dir_key#6/1 &root->kernfs_rwsem &sb->s_type->i_lock_key#30 irq_context: 0 sb_writers#10 &type->i_mutex_dir_key#6/1 &sb->s_type->i_lock_key#30 irq_context: 0 sb_writers#10 &type->i_mutex_dir_key#6/1 &sb->s_type->i_lock_key#30 &dentry->d_lock irq_context: 0 &f->f_pos_lock &type->i_mutex_dir_key#3 sb_writers#4 &journal->j_wait_transaction_locked irq_context: 0 (wq_completion)netns net_cleanup_work rcu_state.barrier_mutex &rq->__lock &cfs_rq->removed.lock irq_context: 0 &mm->mmap_lock &mapping->i_mmap_rwsem &rcu_state.expedited_wq &p->pi_lock &rq->__lock irq_context: 0 &mm->mmap_lock &mapping->i_mmap_rwsem &rcu_state.expedited_wq &p->pi_lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 pernet_ops_rwsem devices_rwsem rdma_nets_rwsem &p->pi_lock &cfs_rq->removed.lock irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem &ht->mutex &rq->__lock irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem &ht->mutex &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 rcu_state.exp_mutex rcu_read_lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)bat_events (work_completion)(&(&bat_priv->mcast.work)->work) rcu_read_lock &rcu_state.gp_wq irq_context: 0 (wq_completion)bat_events (work_completion)(&(&bat_priv->mcast.work)->work) rcu_read_lock &rcu_state.gp_wq &p->pi_lock irq_context: 0 (wq_completion)bat_events (work_completion)(&(&bat_priv->mcast.work)->work) rcu_read_lock &rcu_state.gp_wq &p->pi_lock &rq->__lock irq_context: 0 (wq_completion)bat_events (work_completion)(&(&bat_priv->mcast.work)->work) rcu_read_lock &rcu_state.gp_wq &p->pi_lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)wg-crypt-wg1#5 (work_completion)(&({ do { const void *__vpp_verify = (typeof((worker) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __asm__ ("" : "=r"(__ptr) : "0"((typeof(*((worker))) *)((worker)))); (typeof((typeof(*((worker))) *)((worker)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); })->work) rcu_read_lock rcu_node_0 irq_context: 0 (wq_completion)wg-crypt-wg1#5 (work_completion)(&({ do { const void *__vpp_verify = (typeof((worker) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __asm__ ("" : "=r"(__ptr) : "0"((typeof(*((worker))) *)((worker)))); (typeof((typeof(*((worker))) *)((worker)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); })->work) rcu_read_lock &rq->__lock irq_context: 0 (wq_completion)wg-crypt-wg1#5 (work_completion)(&({ do { const void *__vpp_verify = (typeof((worker) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __asm__ ("" : "=r"(__ptr) : "0"((typeof(*((worker))) *)((worker)))); (typeof((typeof(*((worker))) *)((worker)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); })->work) rcu_read_lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 cb_lock &p->pi_lock &rq->__lock &cfs_rq->removed.lock irq_context: 0 &sbi->s_writepages_rwsem mmu_notifier_invalidate_range_start &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 sb_writers#4 sb_internal &journal->j_wait_transaction_locked irq_context: 0 &f->f_pos_lock sb_writers#10 &c->lock irq_context: 0 &f->f_pos_lock sb_writers#10 &n->list_lock irq_context: 0 &f->f_pos_lock sb_writers#10 &n->list_lock &c->lock irq_context: 0 &f->f_pos_lock &type->i_mutex_dir_key#3 sb_writers#4 jbd2_handle &obj_hash[i].lock irq_context: 0 pernet_ops_rwsem rtnl_mutex &root->kernfs_rwsem &p->pi_lock irq_context: 0 &mm->mmap_lock &mapping->i_mmap_rwsem rcu_read_lock &rcu_state.expedited_wq irq_context: 0 &mm->mmap_lock &mapping->i_mmap_rwsem rcu_read_lock &rcu_state.expedited_wq &p->pi_lock irq_context: 0 &mm->mmap_lock &mapping->i_mmap_rwsem rcu_read_lock &rcu_state.expedited_wq &p->pi_lock &rq->__lock irq_context: 0 &mm->mmap_lock &mapping->i_mmap_rwsem rcu_read_lock &rcu_state.expedited_wq &p->pi_lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 pernet_ops_rwsem devices_rwsem rdma_nets_rwsem &device->compat_devs_mutex &root->kernfs_rwsem &cfs_rq->removed.lock irq_context: 0 pernet_ops_rwsem devices_rwsem rdma_nets_rwsem &device->compat_devs_mutex &root->kernfs_rwsem &obj_hash[i].lock irq_context: 0 (wq_completion)bat_events (work_completion)(&(&forw_packet_aggr->delayed_work)->work) &hard_iface->bat_iv.ogm_buff_mutex rcu_read_lock fill_pool_map-wait-type-override &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 pernet_ops_rwsem key#24 irq_context: 0 (wq_completion)events free_ipc_work sysctl_lock fill_pool_map-wait-type-override pool_lock#2 irq_context: 0 (wq_completion)events free_ipc_work sysctl_lock fill_pool_map-wait-type-override &c->lock irq_context: 0 (wq_completion)events free_ipc_work sysctl_lock fill_pool_map-wait-type-override pool_lock irq_context: 0 pernet_ops_rwsem rtnl_mutex sysctl_lock fill_pool_map-wait-type-override &c->lock irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem remove_cache_srcu rcu_read_lock &rq->__lock irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem remove_cache_srcu rcu_read_lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem rcu_state.exp_mutex rcu_read_lock &rq->__lock irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem rcu_state.exp_mutex rcu_read_lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem rcu_read_lock &pool->lock &p->pi_lock &rq->__lock &cfs_rq->removed.lock irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem ovs_mutex nf_ct_proto_mutex nf_hook_mutex remove_cache_srcu pool_lock#2 irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem rcu_read_lock &pool->lock/1 &p->pi_lock &rq->__lock &cfs_rq->removed.lock irq_context: 0 (wq_completion)events free_ipc_work rcu_node_0 irq_context: 0 (wq_completion)events free_ipc_work &rcu_state.expedited_wq irq_context: 0 (wq_completion)events free_ipc_work &rcu_state.expedited_wq &p->pi_lock irq_context: 0 (wq_completion)events free_ipc_work &rcu_state.expedited_wq &p->pi_lock &rq->__lock irq_context: 0 (wq_completion)events free_ipc_work &rcu_state.expedited_wq &p->pi_lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 &f->f_pos_lock sb_writers#4 &sbi->s_writepages_rwsem jbd2_handle &ei->i_data_sem &c->lock irq_context: 0 &f->f_pos_lock sb_writers#4 &sbi->s_writepages_rwsem jbd2_handle &ei->i_data_sem &lg->lg_mutex irq_context: 0 &f->f_pos_lock sb_writers#4 &sbi->s_writepages_rwsem jbd2_handle &ei->i_data_sem &lg->lg_mutex &ei->i_prealloc_lock irq_context: 0 &f->f_pos_lock sb_writers#4 &sbi->s_writepages_rwsem jbd2_handle &ei->i_data_sem &lg->lg_mutex rcu_read_lock &pa->pa_lock irq_context: 0 &f->f_pos_lock sb_writers#4 &sbi->s_writepages_rwsem jbd2_handle &ei->i_data_sem &lg->lg_mutex &mapping->private_lock irq_context: 0 &f->f_pos_lock sb_writers#4 &sbi->s_writepages_rwsem jbd2_handle &ei->i_data_sem &lg->lg_mutex mmu_notifier_invalidate_range_start irq_context: 0 &f->f_pos_lock sb_writers#4 &sbi->s_writepages_rwsem jbd2_handle &ei->i_data_sem &lg->lg_mutex &c->lock irq_context: 0 &f->f_pos_lock sb_writers#4 &sbi->s_writepages_rwsem jbd2_handle &ei->i_data_sem &lg->lg_mutex pool_lock#2 irq_context: 0 &f->f_pos_lock sb_writers#4 &sbi->s_writepages_rwsem jbd2_handle &ei->i_data_sem &lg->lg_mutex &ret->b_state_lock irq_context: 0 &f->f_pos_lock sb_writers#4 &sbi->s_writepages_rwsem jbd2_handle &ei->i_data_sem &lg->lg_mutex &ret->b_state_lock &journal->j_list_lock irq_context: 0 &f->f_pos_lock sb_writers#4 &sbi->s_writepages_rwsem jbd2_handle &ei->i_data_sem &lg->lg_mutex &rq->__lock irq_context: 0 &f->f_pos_lock sb_writers#4 &sbi->s_writepages_rwsem jbd2_handle &ei->i_data_sem &lg->lg_mutex &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 &f->f_pos_lock sb_writers#4 &sbi->s_writepages_rwsem jbd2_handle &ei->i_data_sem &lg->lg_mutex &pa->pa_lock irq_context: 0 &f->f_pos_lock sb_writers#4 &sbi->s_writepages_rwsem jbd2_handle &ei->i_data_sem &lg->lg_mutex &lg->lg_prealloc_lock irq_context: 0 &f->f_pos_lock sb_writers#4 &sbi->s_writepages_rwsem jbd2_handle lock#4 &lruvec->lru_lock irq_context: 0 &f->f_pos_lock sb_writers#4 &sbi->s_writepages_rwsem &base->lock irq_context: 0 &f->f_pos_lock sb_writers#4 &sbi->s_writepages_rwsem &base->lock &obj_hash[i].lock irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem rtnl_mutex bpf_devs_lock rcu_read_lock rcu_node_0 irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem rtnl_mutex bpf_devs_lock rcu_read_lock &rq->__lock irq_context: 0 sb_writers#4 &type->i_mutex_dir_key#3/1 sb_internal jbd2_handle &journal->j_wait_updates &p->pi_lock irq_context: 0 sb_writers#4 &type->i_mutex_dir_key#3/1 sb_internal jbd2_handle &journal->j_wait_updates &p->pi_lock &rq->__lock irq_context: 0 sb_writers#4 &type->i_mutex_dir_key#3/1 sb_internal jbd2_handle &journal->j_wait_updates &p->pi_lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 &f->f_pos_lock &type->i_mutex_dir_key#3 sb_writers#4 &cfs_rq->removed.lock irq_context: 0 sb_writers#4 sb_internal &rq->__lock &cfs_rq->removed.lock irq_context: 0 bpf_stats_enabled_mutex &n->list_lock irq_context: 0 bpf_stats_enabled_mutex &n->list_lock &c->lock irq_context: 0 &f->f_pos_lock sb_writers#4 &sb->s_type->i_mutex_key#8 &sb->s_type->i_lock_key#22 &xa->xa_lock#9 irq_context: 0 &f->f_pos_lock sb_writers#4 &sb->s_type->i_mutex_key#8 lock#5 irq_context: 0 &f->f_pos_lock sb_writers#4 &sb->s_type->i_mutex_key#8 &lruvec->lru_lock irq_context: 0 &f->f_pos_lock sb_writers#4 &sb->s_type->i_mutex_key#8 jbd2_handle &ei->i_data_sem &mapping->private_lock irq_context: 0 cgroup_threadgroup_rwsem freezer_mutex rcu_read_lock &rcu_state.gp_wq irq_context: 0 cgroup_threadgroup_rwsem freezer_mutex rcu_read_lock &rcu_state.gp_wq &p->pi_lock irq_context: 0 cgroup_threadgroup_rwsem freezer_mutex rcu_read_lock &rcu_state.gp_wq &p->pi_lock &rq->__lock irq_context: 0 cgroup_threadgroup_rwsem freezer_mutex rcu_read_lock &rcu_state.gp_wq &p->pi_lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 &sb->s_type->i_mutex_key#10 console_lock console_srcu console_owner_lock irq_context: 0 &sb->s_type->i_mutex_key#10 console_lock console_srcu console_owner irq_context: 0 &sb->s_type->i_mutex_key#10 console_lock console_srcu console_owner &port_lock_key irq_context: 0 &sb->s_type->i_mutex_key#10 console_lock console_srcu console_owner console_owner_lock irq_context: 0 sb_writers#4 sb_writers#4 &sb->s_type->i_lock_key#22 irq_context: 0 sb_writers#4 sb_writers#4 &wb->list_lock irq_context: 0 sb_writers#4 sb_writers#4 &wb->list_lock &sb->s_type->i_lock_key#22 irq_context: 0 sb_writers#4 &sb->s_type->i_mutex_key#8 jbd2_handle rcu_node_0 irq_context: 0 vlan_ioctl_mutex rtnl_mutex rcu_read_lock (console_sem).lock irq_context: 0 vlan_ioctl_mutex rtnl_mutex rcu_read_lock console_lock console_srcu console_owner_lock irq_context: 0 vlan_ioctl_mutex rtnl_mutex rcu_read_lock console_lock console_srcu console_owner irq_context: 0 vlan_ioctl_mutex rtnl_mutex rcu_read_lock console_lock console_srcu console_owner &port_lock_key irq_context: 0 vlan_ioctl_mutex rtnl_mutex rcu_read_lock console_lock console_srcu console_owner console_owner_lock irq_context: 0 &f->f_pos_lock sb_writers#4 &sb->s_type->i_mutex_key#8 &ei->i_data_sem irq_context: 0 &f->f_pos_lock sb_writers#4 &sb->s_type->i_mutex_key#8 &ei->i_data_sem &rq->__lock irq_context: 0 &f->f_pos_lock sb_writers#4 &sb->s_type->i_mutex_key#8 &ei->i_data_sem &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 &f->f_pos_lock sb_writers#4 &sbi->s_writepages_rwsem rcu_read_lock &memcg->move_lock irq_context: 0 &f->f_pos_lock sb_writers#4 &sbi->s_writepages_rwsem rcu_read_lock &xa->xa_lock#9 irq_context: 0 &f->f_pos_lock sb_writers#4 &sbi->s_writepages_rwsem rcu_read_lock &xa->xa_lock#9 &s->s_inode_wblist_lock irq_context: 0 sb_writers#4 &sb->s_type->i_mutex_key#8 jbd2_handle &sbi->s_orphan_lock bit_wait_table + i irq_context: 0 sb_writers#4 &sb->s_type->i_mutex_key#8 jbd2_handle &lock->wait_lock irq_context: 0 sb_writers#4 &sb->s_type->i_mutex_key#8 jbd2_handle &p->pi_lock irq_context: 0 sb_writers#4 &sb->s_type->i_mutex_key#8 jbd2_handle &p->pi_lock &rq->__lock irq_context: 0 sb_writers#4 &sb->s_type->i_mutex_key#8 jbd2_handle &p->pi_lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 &f->f_pos_lock sb_writers#4 rcu_read_lock &rq->__lock irq_context: 0 &f->f_pos_lock sb_writers#4 rcu_read_lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)wg-crypt-wg0#6 &rq->__lock irq_context: 0 (wq_completion)wg-crypt-wg0#6 &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 &f->f_pos_lock sb_writers#4 &journal->j_state_lock &journal->j_wait_commit &p->pi_lock &cfs_rq->removed.lock irq_context: 0 sb_writers#4 &type->i_mutex_dir_key#3/1 sb_internal &journal->j_state_lock &journal->j_wait_transaction_locked irq_context: 0 sb_writers#4 &type->i_mutex_dir_key#3/1 sb_internal jbd2_handle irq_context: 0 tracepoints_mutex cpu_hotplug_lock static_call_mutex text_mutex text_mutex.wait_lock irq_context: 0 bpf_stats_enabled_mutex cpu_hotplug_lock jump_label_mutex text_mutex.wait_lock irq_context: 0 bpf_stats_enabled_mutex cpu_hotplug_lock jump_label_mutex &p->pi_lock irq_context: 0 bpf_stats_enabled_mutex cpu_hotplug_lock jump_label_mutex &p->pi_lock &rq->__lock irq_context: 0 bpf_stats_enabled_mutex cpu_hotplug_lock jump_label_mutex &p->pi_lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 bpf_stats_enabled_mutex cpu_hotplug_lock jump_label_mutex &rq->__lock irq_context: 0 bpf_stats_enabled_mutex cpu_hotplug_lock jump_label_mutex &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 tracepoints_mutex cpu_hotplug_lock jump_label_mutex jump_label_mutex.wait_lock irq_context: 0 tracepoints_mutex cpu_hotplug_lock jump_label_mutex &rq->__lock irq_context: 0 tracepoints_mutex cpu_hotplug_lock jump_label_mutex &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 tracepoints_mutex cpu_hotplug_lock jump_label_mutex &cfs_rq->removed.lock irq_context: 0 tracepoints_mutex cpu_hotplug_lock jump_label_mutex &obj_hash[i].lock irq_context: 0 tracepoints_mutex cpu_hotplug_lock jump_label_mutex pool_lock#2 irq_context: 0 &mm->mmap_lock rcu_read_lock rcu_read_lock ptlock_ptr(page)#2 rcu_read_lock per_cpu_ptr(&cgroup_rstat_cpu_lock, cpu) irq_context: 0 vlan_ioctl_mutex rtnl_mutex per_cpu_ptr(&cgroup_rstat_cpu_lock, cpu) irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem rtnl_mutex per_cpu_ptr(&cgroup_rstat_cpu_lock, cpu) irq_context: 0 (wq_completion)events (work_completion)(&aux->work) rcu_node_0 irq_context: 0 (wq_completion)events (work_completion)(&aux->work) rcu_read_lock &rcu_state.expedited_wq irq_context: 0 (wq_completion)events (work_completion)(&aux->work) rcu_read_lock &rcu_state.expedited_wq &p->pi_lock irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem rtnl_mutex dev_hotplug_mutex &rq->__lock irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem rtnl_mutex dev_hotplug_mutex &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 &dentry->d_lock &dentry->d_lock/1 &lru->node[i].lock rcu_read_lock &p->pi_lock irq_context: 0 &dentry->d_lock &dentry->d_lock/1 &lru->node[i].lock rcu_read_lock &p->pi_lock &rq->__lock irq_context: 0 &dentry->d_lock &dentry->d_lock/1 &lru->node[i].lock rcu_read_lock &p->pi_lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 loop_validate_mutex &rq->__lock irq_context: 0 loop_validate_mutex &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)wg-kex-wg0#7 (work_completion)(&peer->transmit_handshake_work) &c->lock irq_context: 0 (wq_completion)wg-kex-wg0#7 (work_completion)(&peer->transmit_handshake_work) &peer->endpoint_lock rcu_read_lock_bh &c->lock irq_context: 0 sb_writers#4 &sb->s_type->i_mutex_key#8 mapping.invalidate_lock &n->list_lock irq_context: 0 sb_writers#4 &sb->s_type->i_mutex_key#8 mapping.invalidate_lock &n->list_lock &c->lock irq_context: 0 bpf_stats_enabled_mutex cpu_hotplug_lock &rq->__lock irq_context: 0 bpf_stats_enabled_mutex cpu_hotplug_lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)wg-crypt-wg0#5 (work_completion)(&peer->transmit_packet_work) rcu_node_0 irq_context: 0 &f->f_pos_lock sb_writers#4 &sb->s_type->i_mutex_key#8 jbd2_handle &sbi->s_orphan_lock &ei->i_raw_lock irq_context: 0 &f->f_pos_lock sb_writers#4 &sb->s_type->i_mutex_key#8 &fq->mq_flush_lock irq_context: 0 &f->f_pos_lock sb_writers#4 &sb->s_type->i_mutex_key#8 &fq->mq_flush_lock tk_core.seq.seqcount irq_context: 0 &f->f_pos_lock sb_writers#4 &sb->s_type->i_mutex_key#8 &fq->mq_flush_lock &q->requeue_lock irq_context: 0 &f->f_pos_lock sb_writers#4 &sb->s_type->i_mutex_key#8 &fq->mq_flush_lock &obj_hash[i].lock irq_context: 0 &f->f_pos_lock sb_writers#4 &sb->s_type->i_mutex_key#8 &fq->mq_flush_lock rcu_read_lock &pool->lock irq_context: 0 &f->f_pos_lock sb_writers#4 &sb->s_type->i_mutex_key#8 &fq->mq_flush_lock rcu_read_lock &pool->lock &obj_hash[i].lock irq_context: 0 &f->f_pos_lock sb_writers#4 &sb->s_type->i_mutex_key#8 &fq->mq_flush_lock rcu_read_lock &pool->lock pool_lock#2 irq_context: 0 &f->f_pos_lock sb_writers#4 &sb->s_type->i_mutex_key#8 &fq->mq_flush_lock rcu_read_lock &pool->lock &p->pi_lock irq_context: 0 &f->f_pos_lock sb_writers#4 &sb->s_type->i_mutex_key#8 &fq->mq_flush_lock rcu_read_lock &pool->lock &p->pi_lock &rq->__lock irq_context: 0 &f->f_pos_lock sb_writers#4 &sb->s_type->i_mutex_key#8 &fq->mq_flush_lock rcu_read_lock &pool->lock &p->pi_lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 &f->f_pos_lock sb_writers#4 &sb->s_type->i_mutex_key#8 &x->wait#26 irq_context: 0 &f->f_pos_lock &type->i_mutex_dir_key#4 &root->kernfs_rwsem pool_lock#2 irq_context: 0 remove_cache_srcu &meta->lock irq_context: 0 remove_cache_srcu kfence_freelist_lock irq_context: 0 sb_writers#4 &sb->s_type->i_mutex_key#8 jbd2_handle &ei->i_data_sem &cfs_rq->removed.lock irq_context: 0 &sig->cred_guard_mutex tomoyo_ss &cfs_rq->removed.lock irq_context: 0 sb_writers#4 &type->i_mutex_dir_key#3/1 &type->i_mutex_dir_key#3 &journal->j_state_lock irq_context: 0 sb_writers#4 &type->i_mutex_dir_key#3/1 &type->i_mutex_dir_key#3 &journal->j_state_lock tk_core.seq.seqcount irq_context: 0 sb_writers#4 &type->i_mutex_dir_key#3/1 &type->i_mutex_dir_key#3 &journal->j_state_lock &obj_hash[i].lock irq_context: 0 sb_writers#4 &type->i_mutex_dir_key#3/1 &type->i_mutex_dir_key#3 &journal->j_state_lock &base->lock irq_context: 0 sb_writers#4 &type->i_mutex_dir_key#3/1 &type->i_mutex_dir_key#3 &journal->j_state_lock &base->lock &obj_hash[i].lock irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem &root->kernfs_rwsem &cfs_rq->removed.lock irq_context: 0 &xt[i].mutex &mm->mmap_lock &rq->__lock &cfs_rq->removed.lock irq_context: 0 &f->f_pos_lock sb_writers#4 &sbi->s_writepages_rwsem jbd2_handle &ei->i_data_sem &ei->i_es_lock &c->lock irq_context: 0 &f->f_pos_lock sb_writers#4 &sbi->s_writepages_rwsem rcu_read_lock &base->lock irq_context: 0 &f->f_pos_lock sb_writers#4 &sbi->s_writepages_rwsem rcu_read_lock &base->lock &obj_hash[i].lock irq_context: 0 &sbi->s_writepages_rwsem batched_entropy_u8.lock irq_context: 0 &sbi->s_writepages_rwsem kfence_freelist_lock irq_context: 0 &pipe->mutex/1 sk_lock-AF_INET6 remove_cache_srcu irq_context: 0 &pipe->mutex/1 sk_lock-AF_INET6 remove_cache_srcu quarantine_lock irq_context: 0 &pipe->mutex/1 sk_lock-AF_INET6 remove_cache_srcu &c->lock irq_context: 0 &pipe->mutex/1 sk_lock-AF_INET6 remove_cache_srcu &n->list_lock irq_context: 0 &pipe->mutex/1 sk_lock-AF_INET6 remove_cache_srcu rcu_read_lock pool_lock#2 irq_context: 0 &pipe->mutex/1 sk_lock-AF_INET6 remove_cache_srcu &obj_hash[i].lock irq_context: 0 &pipe->mutex/1 sk_lock-AF_INET6 remove_cache_srcu pool_lock#2 irq_context: 0 &pipe->mutex/1 sk_lock-AF_INET6 remove_cache_srcu &rq->__lock irq_context: 0 &pipe->mutex/1 sk_lock-AF_INET6 remove_cache_srcu &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 sb_writers#4 &type->i_mutex_dir_key#3/1 sb_internal jbd2_handle &ei->i_data_sem remove_cache_srcu irq_context: 0 sb_writers#4 &type->i_mutex_dir_key#3/1 sb_internal jbd2_handle &ei->i_data_sem remove_cache_srcu quarantine_lock irq_context: 0 sb_writers#4 &type->i_mutex_dir_key#3/1 sb_internal jbd2_handle &ei->i_data_sem remove_cache_srcu &c->lock irq_context: 0 sb_writers#4 &type->i_mutex_dir_key#3/1 sb_internal jbd2_handle &ei->i_data_sem remove_cache_srcu &n->list_lock irq_context: 0 sb_writers#4 &type->i_mutex_dir_key#3/1 sb_internal jbd2_handle &ei->i_data_sem remove_cache_srcu rcu_read_lock pool_lock#2 irq_context: 0 sb_writers#4 &type->i_mutex_dir_key#3/1 sb_internal jbd2_handle &ei->i_data_sem remove_cache_srcu &obj_hash[i].lock irq_context: 0 sb_writers#4 &type->i_mutex_dir_key#3/1 sb_internal jbd2_handle &ei->i_data_sem remove_cache_srcu &rq->__lock irq_context: 0 sb_writers#4 &type->i_mutex_dir_key#3/1 sb_internal jbd2_handle &ei->i_data_sem remove_cache_srcu &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 sb_writers#4 &type->i_mutex_dir_key#3/1 sb_internal jbd2_handle &ei->i_data_sem remove_cache_srcu pgd_lock irq_context: 0 sb_writers#4 &type->i_mutex_dir_key#3/1 sb_internal jbd2_handle &ei->i_data_sem remove_cache_srcu stock_lock irq_context: 0 sb_writers#4 &type->i_mutex_dir_key#3/1 sb_internal jbd2_handle &ei->i_data_sem remove_cache_srcu key irq_context: 0 sb_writers#4 &type->i_mutex_dir_key#3/1 sb_internal jbd2_handle &ei->i_data_sem remove_cache_srcu pcpu_lock irq_context: 0 sb_writers#4 &type->i_mutex_dir_key#3/1 sb_internal jbd2_handle &ei->i_data_sem remove_cache_srcu percpu_counters_lock irq_context: 0 sb_writers#4 &type->i_mutex_dir_key#3/1 sb_internal jbd2_handle &ei->i_data_sem remove_cache_srcu pcpu_lock stock_lock irq_context: 0 sb_writers#4 &type->i_mutex_dir_key#3/1 sb_internal jbd2_handle &ei->i_data_sem remove_cache_srcu pool_lock#2 irq_context: 0 vlan_ioctl_mutex rtnl_mutex rcu_node_0 irq_context: 0 vlan_ioctl_mutex rtnl_mutex &ul->lock#2 irq_context: 0 sk_lock-AF_TIPC rcu_read_lock &service->lock irq_context: 0 sk_lock-AF_TIPC k-slock-AF_TIPC &list->lock#31 irq_context: 0 sk_lock-AF_TIPC k-slock-AF_TIPC k-clock-AF_TIPC irq_context: 0 sk_lock-AF_TIPC k-slock-AF_TIPC k-clock-AF_TIPC rcu_read_lock &pool->lock/1 irq_context: 0 sk_lock-AF_TIPC k-slock-AF_TIPC k-clock-AF_TIPC rcu_read_lock &pool->lock/1 &obj_hash[i].lock irq_context: 0 sk_lock-AF_TIPC k-slock-AF_TIPC k-clock-AF_TIPC rcu_read_lock &pool->lock/1 pool_lock#2 irq_context: 0 sk_lock-AF_TIPC k-slock-AF_TIPC k-clock-AF_TIPC rcu_read_lock &pool->lock/1 &p->pi_lock irq_context: 0 sk_lock-AF_TIPC k-slock-AF_TIPC k-clock-AF_TIPC rcu_read_lock &pool->lock/1 &p->pi_lock &rq->__lock irq_context: 0 sk_lock-AF_TIPC k-slock-AF_TIPC k-clock-AF_TIPC rcu_read_lock &pool->lock/1 &p->pi_lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 sk_lock-AF_TIPC &ei->socket.wq.wait irq_context: 0 (wq_completion)mld (work_completion)(&(&idev->mc_ifc_work)->work) &idev->mc_lock rcu_read_lock rcu_read_lock &n->list_lock &c->lock irq_context: 0 (wq_completion)tipc_rcv irq_context: 0 (wq_completion)tipc_rcv (work_completion)(&srv->awork) irq_context: 0 (wq_completion)tipc_rcv (work_completion)(&srv->awork) &srv->idr_lock irq_context: 0 (wq_completion)tipc_rcv (work_completion)(&srv->awork) fs_reclaim irq_context: 0 (wq_completion)tipc_rcv (work_completion)(&srv->awork) fs_reclaim mmu_notifier_invalidate_range_start irq_context: 0 (wq_completion)tipc_rcv (work_completion)(&srv->awork) pool_lock#2 irq_context: 0 (wq_completion)tipc_rcv (work_completion)(&srv->awork) mmu_notifier_invalidate_range_start irq_context: 0 (wq_completion)tipc_rcv (work_completion)(&srv->awork) &sb->s_type->i_lock_key#8 irq_context: 0 (wq_completion)tipc_rcv (work_completion)(&srv->awork) k-sk_lock-AF_TIPC irq_context: 0 (wq_completion)tipc_rcv (work_completion)(&srv->awork) k-sk_lock-AF_TIPC k-slock-AF_TIPC irq_context: 0 (wq_completion)tipc_rcv (work_completion)(&srv->awork) k-sk_lock-AF_TIPC fs_reclaim irq_context: 0 (wq_completion)tipc_rcv (work_completion)(&srv->awork) k-sk_lock-AF_TIPC fs_reclaim mmu_notifier_invalidate_range_start irq_context: 0 (wq_completion)tipc_rcv (work_completion)(&srv->awork) k-sk_lock-AF_TIPC pool_lock#2 irq_context: 0 (wq_completion)tipc_rcv (work_completion)(&srv->awork) k-sk_lock-AF_TIPC &dir->lock irq_context: 0 (wq_completion)tipc_rcv (work_completion)(&srv->awork) k-sk_lock-AF_TIPC &obj_hash[i].lock irq_context: 0 (wq_completion)tipc_rcv (work_completion)(&srv->awork) k-sk_lock-AF_TIPC batched_entropy_u32.lock irq_context: 0 (wq_completion)tipc_rcv (work_completion)(&srv->awork) k-sk_lock-AF_TIPC rcu_read_lock rhashtable_bucket irq_context: 0 (wq_completion)tipc_rcv (work_completion)(&srv->awork) k-sk_lock-AF_TIPC k-sk_lock-AF_TIPC/1 irq_context: 0 (wq_completion)tipc_rcv (work_completion)(&srv->awork) k-sk_lock-AF_TIPC k-sk_lock-AF_TIPC/1 k-slock-AF_TIPC irq_context: 0 (wq_completion)tipc_rcv (work_completion)(&srv->awork) k-sk_lock-AF_TIPC k-sk_lock-AF_TIPC/1 &obj_hash[i].lock irq_context: 0 (wq_completion)tipc_rcv (work_completion)(&srv->awork) k-sk_lock-AF_TIPC k-sk_lock-AF_TIPC/1 &base->lock irq_context: 0 (wq_completion)tipc_rcv (work_completion)(&srv->awork) k-sk_lock-AF_TIPC k-sk_lock-AF_TIPC/1 &base->lock &obj_hash[i].lock irq_context: 0 (wq_completion)tipc_rcv (work_completion)(&srv->awork) k-sk_lock-AF_TIPC k-sk_lock-AF_TIPC/1 fs_reclaim irq_context: 0 (wq_completion)tipc_rcv (work_completion)(&srv->awork) k-sk_lock-AF_TIPC k-sk_lock-AF_TIPC/1 fs_reclaim mmu_notifier_invalidate_range_start irq_context: 0 (wq_completion)tipc_rcv (work_completion)(&srv->awork) k-sk_lock-AF_TIPC k-sk_lock-AF_TIPC/1 pool_lock#2 irq_context: 0 (wq_completion)tipc_rcv (work_completion)(&srv->awork) k-sk_lock-AF_TIPC k-sk_lock-AF_TIPC/1 &list->lock#31 irq_context: 0 (wq_completion)tipc_rcv (work_completion)(&srv->awork) k-sk_lock-AF_TIPC k-sk_lock-AF_TIPC/1 slock-AF_TIPC &list->lock#31 irq_context: 0 (wq_completion)tipc_rcv (work_completion)(&srv->awork) k-sk_lock-AF_TIPC k-sk_lock-AF_TIPC/1 slock-AF_TIPC &obj_hash[i].lock irq_context: 0 (wq_completion)tipc_rcv (work_completion)(&srv->awork) k-sk_lock-AF_TIPC k-sk_lock-AF_TIPC/1 slock-AF_TIPC &base->lock irq_context: 0 (wq_completion)tipc_rcv (work_completion)(&srv->awork) k-sk_lock-AF_TIPC k-sk_lock-AF_TIPC/1 slock-AF_TIPC &base->lock &obj_hash[i].lock irq_context: 0 (wq_completion)tipc_rcv (work_completion)(&srv->awork) k-sk_lock-AF_TIPC k-sk_lock-AF_TIPC/1 slock-AF_TIPC pool_lock#2 irq_context: 0 (wq_completion)tipc_rcv (work_completion)(&srv->awork) k-sk_lock-AF_TIPC k-sk_lock-AF_TIPC/1 slock-AF_TIPC rcu_read_lock &ei->socket.wq.wait irq_context: 0 (wq_completion)tipc_rcv (work_completion)(&srv->awork) k-sk_lock-AF_TIPC k-sk_lock-AF_TIPC/1 slock-AF_TIPC rcu_read_lock &ei->socket.wq.wait &p->pi_lock irq_context: 0 (wq_completion)tipc_rcv (work_completion)(&srv->awork) k-sk_lock-AF_TIPC k-sk_lock-AF_TIPC/1 slock-AF_TIPC rcu_read_lock &ei->socket.wq.wait &p->pi_lock &rq->__lock irq_context: 0 (wq_completion)tipc_rcv (work_completion)(&srv->awork) k-sk_lock-AF_TIPC k-sk_lock-AF_TIPC/1 slock-AF_TIPC rcu_read_lock &ei->socket.wq.wait &p->pi_lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)tipc_rcv (work_completion)(&srv->awork) k-slock-AF_TIPC irq_context: 0 (wq_completion)tipc_rcv (work_completion)(&srv->awork) &obj_hash[i].lock irq_context: 0 (wq_completion)tipc_rcv (work_completion)(&srv->awork) &srv->idr_lock pool_lock#2 irq_context: 0 (wq_completion)tipc_rcv (work_completion)(&srv->awork) k-clock-AF_TIPC irq_context: 0 (wq_completion)tipc_rcv (work_completion)(&srv->awork) k-clock-AF_TIPC irq_context: 0 (wq_completion)tipc_rcv (work_completion)(&srv->awork) k-clock-AF_TIPC rcu_read_lock &pool->lock/1 irq_context: 0 (wq_completion)tipc_rcv (work_completion)(&srv->awork) k-clock-AF_TIPC rcu_read_lock &pool->lock/1 &obj_hash[i].lock irq_context: 0 (wq_completion)tipc_rcv (work_completion)(&srv->awork) k-clock-AF_TIPC rcu_read_lock &pool->lock/1 pool_lock#2 irq_context: 0 (wq_completion)tipc_rcv (work_completion)(&srv->awork) k-clock-AF_TIPC rcu_read_lock &pool->lock/1 &p->pi_lock irq_context: 0 (wq_completion)tipc_rcv (work_completion)(&srv->awork) k-clock-AF_TIPC rcu_read_lock &pool->lock/1 &p->pi_lock &rq->__lock irq_context: 0 (wq_completion)tipc_rcv (work_completion)(&srv->awork) k-clock-AF_TIPC rcu_read_lock &pool->lock/1 &p->pi_lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)tipc_rcv (work_completion)(&srv->awork) &rq->__lock irq_context: 0 sk_lock-AF_TIPC k-slock-AF_TIPC pool_lock#2 irq_context: 0 sk_lock-AF_TIPC &____s->seqcount#2 irq_context: 0 sk_lock-AF_TIPC &____s->seqcount irq_context: 0 sk_lock-AF_TIPC &mm->mmap_lock &rq->__lock irq_context: 0 sk_lock-AF_TIPC &mm->mmap_lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 sk_lock-AF_TIPC rcu_read_lock rcu_node_0 irq_context: 0 sk_lock-AF_TIPC rcu_read_lock &rq->__lock irq_context: 0 sk_lock-AF_TIPC rcu_read_lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 sk_lock-AF_TIPC fs_reclaim &rq->__lock irq_context: 0 sk_lock-AF_TIPC fs_reclaim &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)tipc_rcv (work_completion)(&srv->awork) &xa->xa_lock#9 irq_context: 0 (wq_completion)tipc_rcv (work_completion)(&srv->awork) &fsnotify_mark_srcu irq_context: 0 (wq_completion)tipc_rcv (work_completion)(&con->rwork) irq_context: 0 (wq_completion)tipc_rcv (work_completion)(&con->rwork) &rq->__lock irq_context: 0 (wq_completion)tipc_rcv (work_completion)(&con->rwork) k-sk_lock-AF_TIPC irq_context: 0 (wq_completion)tipc_rcv (work_completion)(&con->rwork) k-sk_lock-AF_TIPC k-slock-AF_TIPC irq_context: 0 (wq_completion)tipc_rcv (work_completion)(&con->rwork) k-sk_lock-AF_TIPC &obj_hash[i].lock irq_context: 0 (wq_completion)tipc_rcv (work_completion)(&con->rwork) k-sk_lock-AF_TIPC pool_lock#2 irq_context: 0 (wq_completion)tipc_rcv (work_completion)(&con->rwork) k-sk_lock-AF_TIPC quarantine_lock irq_context: 0 (wq_completion)tipc_rcv (work_completion)(&con->rwork) k-slock-AF_TIPC irq_context: 0 sk_lock-AF_TIPC rcu_read_lock &rcu_state.gp_wq irq_context: 0 sk_lock-AF_TIPC rcu_read_lock &rcu_state.gp_wq &p->pi_lock irq_context: 0 sk_lock-AF_TIPC &pcp->lock &zone->lock irq_context: 0 (wq_completion)tipc_rcv (work_completion)(&con->rwork) k-sk_lock-AF_TIPC k-clock-AF_TIPC irq_context: 0 vlan_ioctl_mutex rtnl_mutex &block->lock irq_context: 0 vlan_ioctl_mutex rtnl_mutex &block->cb_lock irq_context: 0 vlan_ioctl_mutex rtnl_mutex &block->cb_lock flow_indr_block_lock irq_context: 0 vlan_ioctl_mutex rtnl_mutex &block->cb_lock flow_indr_block_lock &obj_hash[i].lock irq_context: 0 vlan_ioctl_mutex rtnl_mutex &block->cb_lock flow_indr_block_lock pool_lock#2 irq_context: 0 vlan_ioctl_mutex rtnl_mutex &root->kernfs_rwsem &sem->wait_lock irq_context: 0 (wq_completion)events free_ipc_work sb_lock &obj_hash[i].lock pool_lock irq_context: 0 (wq_completion)events free_ipc_work quarantine_lock irq_context: 0 &sb->s_type->i_lock_key &xa->xa_lock#9 &obj_hash[i].lock pool_lock irq_context: 0 dup_mmap_sem &rq->__lock irq_context: 0 dup_mmap_sem &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)events (work_completion)(&(&nsim_dev->trap_data->trap_report_dw)->work) fill_pool_map-wait-type-override &n->list_lock irq_context: 0 (wq_completion)events (work_completion)(&(&nsim_dev->trap_data->trap_report_dw)->work) fill_pool_map-wait-type-override &n->list_lock &c->lock irq_context: 0 &sb->s_type->i_mutex_key#10 sk_lock-AF_TIPC &list->lock#31 irq_context: 0 &sb->s_type->i_mutex_key#10 sk_lock-AF_TIPC k-slock-AF_TIPC &list->lock#31 irq_context: 0 &sb->s_type->i_mutex_key#10 sk_lock-AF_TIPC k-slock-AF_TIPC k-clock-AF_TIPC irq_context: 0 &sb->s_type->i_mutex_key#10 sk_lock-AF_TIPC k-slock-AF_TIPC k-clock-AF_TIPC rcu_read_lock &pool->lock/1 irq_context: 0 &sb->s_type->i_mutex_key#10 sk_lock-AF_TIPC k-slock-AF_TIPC k-clock-AF_TIPC rcu_read_lock &pool->lock/1 &obj_hash[i].lock irq_context: 0 &sb->s_type->i_mutex_key#10 sk_lock-AF_TIPC k-slock-AF_TIPC k-clock-AF_TIPC rcu_read_lock &pool->lock/1 pool_lock#2 irq_context: 0 &sb->s_type->i_mutex_key#10 sk_lock-AF_TIPC k-slock-AF_TIPC k-clock-AF_TIPC rcu_read_lock &pool->lock/1 &p->pi_lock irq_context: 0 &sb->s_type->i_mutex_key#10 sk_lock-AF_TIPC k-slock-AF_TIPC k-clock-AF_TIPC rcu_read_lock &pool->lock/1 &p->pi_lock &rq->__lock irq_context: 0 &sb->s_type->i_mutex_key#10 sk_lock-AF_TIPC k-slock-AF_TIPC k-clock-AF_TIPC rcu_read_lock &pool->lock/1 &p->pi_lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)tipc_rcv (work_completion)(&con->rwork) k-clock-AF_TIPC irq_context: 0 (wq_completion)tipc_rcv (work_completion)(&con->rwork) k-clock-AF_TIPC &con->sub_lock irq_context: 0 (wq_completion)tipc_rcv (work_completion)(&con->rwork) &srv->idr_lock irq_context: 0 (wq_completion)tipc_rcv (work_completion)(&con->rwork) &srv->idr_lock &obj_hash[i].lock irq_context: 0 (wq_completion)tipc_rcv (work_completion)(&con->rwork) &srv->idr_lock pool_lock#2 irq_context: 0 (wq_completion)tipc_rcv (work_completion)(&con->rwork) k-sk_lock-AF_TIPC &base->lock irq_context: 0 (wq_completion)tipc_rcv (work_completion)(&con->rwork) k-sk_lock-AF_TIPC &base->lock &obj_hash[i].lock irq_context: 0 (wq_completion)tipc_rcv (work_completion)(&con->rwork) k-sk_lock-AF_TIPC rcu_read_lock rhashtable_bucket irq_context: 0 (wq_completion)tipc_rcv (work_completion)(&con->rwork) k-sk_lock-AF_TIPC k-clock-AF_TIPC irq_context: 0 (wq_completion)tipc_rcv (work_completion)(&con->rwork) &obj_hash[i].lock irq_context: 0 (wq_completion)tipc_rcv (work_completion)(&con->rwork) pool_lock#2 irq_context: 0 (wq_completion)tipc_rcv (work_completion)(&con->rwork) &sb->s_type->i_lock_key#8 irq_context: 0 (wq_completion)tipc_rcv (work_completion)(&con->rwork) &xa->xa_lock#9 irq_context: 0 (wq_completion)tipc_rcv (work_completion)(&con->rwork) &fsnotify_mark_srcu irq_context: 0 (wq_completion)tipc_rcv (work_completion)(&con->rwork) &con->outqueue_lock irq_context: softirq (&n->timer) k-slock-AF_INET6 rcu_read_lock rcu_read_lock &n->list_lock irq_context: softirq (&n->timer) k-slock-AF_INET6 rcu_read_lock rcu_read_lock &n->list_lock &c->lock irq_context: 0 (wq_completion)events_unbound (reaper_work).work fill_pool_map-wait-type-override &____s->seqcount irq_context: 0 (wq_completion)events_unbound (reaper_work).work fill_pool_map-wait-type-override rcu_read_lock pool_lock#2 irq_context: 0 (wq_completion)events_unbound (reaper_work).work fill_pool_map-wait-type-override &obj_hash[i].lock irq_context: 0 (wq_completion)events free_ipc_work sysctl_lock krc.lock &obj_hash[i].lock irq_context: 0 (wq_completion)events free_ipc_work sysctl_lock krc.lock &base->lock irq_context: 0 (wq_completion)events free_ipc_work sysctl_lock krc.lock &base->lock &obj_hash[i].lock irq_context: 0 (wq_completion)events free_ipc_work sysctl_lock &obj_hash[i].lock pool_lock irq_context: softirq rcu_callback &dir->lock &obj_hash[i].lock irq_context: softirq rcu_callback &dir->lock pool_lock#2 irq_context: 0 pernet_ops_rwsem devices_rwsem rdma_nets_rwsem &device->compat_devs_mutex uevent_sock_mutex &n->list_lock irq_context: 0 (wq_completion)tipc_rcv (work_completion)(&con->rwork) k-sk_lock-AF_TIPC k-clock-AF_TIPC rcu_read_lock &pool->lock/1 irq_context: 0 (wq_completion)events (work_completion)(&pwq->unbound_release_work) &wq->mutex &rq->__lock irq_context: 0 sk_lock-AF_TIPC &base->lock irq_context: 0 namespace_sem fs_reclaim mmu_notifier_invalidate_range_start &rq->__lock irq_context: 0 namespace_sem fs_reclaim mmu_notifier_invalidate_range_start &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 pernet_ops_rwsem devices_rwsem rdma_nets_rwsem &device->compat_devs_mutex uevent_sock_mutex &n->list_lock &c->lock irq_context: 0 (wq_completion)events (work_completion)(&pwq->unbound_release_work) &wq->mutex &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 sk_lock-AF_TIPC &base->lock &obj_hash[i].lock irq_context: 0 sk_lock-AF_TIPC rcu_read_lock &ei->socket.wq.wait irq_context: 0 sk_lock-AF_TIPC k-slock-AF_TIPC k-clock-AF_TIPC rcu_read_lock &pool->lock/1 &p->pi_lock &cfs_rq->removed.lock irq_context: 0 (wq_completion)tipc_rcv (work_completion)(&con->rwork) k-sk_lock-AF_TIPC k-clock-AF_TIPC rcu_read_lock &pool->lock/1 &obj_hash[i].lock irq_context: 0 (wq_completion)tipc_rcv (work_completion)(&con->rwork) k-sk_lock-AF_TIPC k-clock-AF_TIPC rcu_read_lock &pool->lock/1 pool_lock#2 irq_context: 0 (wq_completion)tipc_rcv (work_completion)(&con->rwork) k-sk_lock-AF_TIPC k-clock-AF_TIPC rcu_read_lock &pool->lock/1 &p->pi_lock irq_context: 0 (wq_completion)tipc_rcv (work_completion)(&con->rwork) k-sk_lock-AF_TIPC k-clock-AF_TIPC rcu_read_lock &pool->lock/1 &p->pi_lock &rq->__lock irq_context: 0 (wq_completion)tipc_rcv (work_completion)(&con->rwork) k-sk_lock-AF_TIPC k-clock-AF_TIPC rcu_read_lock &pool->lock/1 &p->pi_lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)tipc_rcv (work_completion)(&con->rwork) k-sk_lock-AF_TIPC &rq->__lock irq_context: 0 (wq_completion)tipc_rcv (work_completion)(&con->rwork) k-sk_lock-AF_TIPC k-clock-AF_TIPC rcu_read_lock &pool->lock/1 &p->pi_lock &cfs_rq->removed.lock irq_context: 0 (wq_completion)tipc_rcv (work_completion)(&con->rwork) k-sk_lock-AF_TIPC k-clock-AF_TIPC rcu_read_lock &pool->lock/1 &p->pi_lock &rq->__lock &cfs_rq->removed.lock irq_context: 0 (wq_completion)tipc_rcv (work_completion)(&con->rwork) k-sk_lock-AF_TIPC &c->lock irq_context: 0 (wq_completion)tipc_rcv (work_completion)(&con->rwork) k-sk_lock-AF_TIPC &list->lock#31 irq_context: 0 (wq_completion)tipc_rcv (work_completion)(&con->rwork) k-sk_lock-AF_TIPC slock-AF_TIPC &list->lock#31 irq_context: 0 sk_lock-AF_TIPC remove_cache_srcu irq_context: 0 sk_lock-AF_TIPC k-slock-AF_TIPC &c->lock irq_context: 0 sk_lock-AF_TIPC k-slock-AF_TIPC &____s->seqcount#2 irq_context: 0 sk_lock-AF_TIPC k-slock-AF_TIPC &____s->seqcount irq_context: 0 (wq_completion)tipc_rcv (work_completion)(&con->rwork) k-sk_lock-AF_TIPC &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)tipc_rcv (work_completion)(&con->rwork) k-sk_lock-AF_TIPC slock-AF_TIPC k-slock-AF_TIPC &list->lock#31 irq_context: 0 (wq_completion)tipc_rcv (work_completion)(&con->rwork) k-sk_lock-AF_TIPC slock-AF_TIPC &obj_hash[i].lock irq_context: 0 (wq_completion)tipc_rcv (work_completion)(&con->rwork) k-sk_lock-AF_TIPC slock-AF_TIPC pool_lock#2 irq_context: 0 sk_lock-AF_TIPC remove_cache_srcu quarantine_lock irq_context: 0 sk_lock-AF_TIPC remove_cache_srcu &c->lock irq_context: 0 sk_lock-AF_TIPC remove_cache_srcu &n->list_lock irq_context: 0 sk_lock-AF_TIPC remove_cache_srcu rcu_read_lock pool_lock#2 irq_context: 0 sk_lock-AF_TIPC remove_cache_srcu &obj_hash[i].lock irq_context: 0 sk_lock-AF_TIPC remove_cache_srcu &rq->__lock irq_context: 0 sk_lock-AF_TIPC remove_cache_srcu &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)ext4-rsv-conversion (work_completion)(&ei->i_rsv_conversion_work) jbd2_handle &ei->i_data_sem &ei->i_es_lock &n->list_lock irq_context: 0 (wq_completion)ext4-rsv-conversion (work_completion)(&ei->i_rsv_conversion_work) jbd2_handle &ei->i_data_sem &ei->i_es_lock &n->list_lock &c->lock irq_context: 0 sb_writers#4 &type->i_mutex_dir_key#3/1 jbd2_handle &ei->i_data_sem &ei->i_es_lock rcu_read_lock &p->pi_lock irq_context: 0 sb_writers#4 &type->i_mutex_dir_key#3/1 jbd2_handle &ei->i_data_sem &ei->i_es_lock rcu_read_lock &p->pi_lock &rq->__lock irq_context: 0 sb_writers#4 &type->i_mutex_dir_key#3/1 jbd2_handle &ei->i_data_sem &ei->i_es_lock rcu_read_lock &p->pi_lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 pernet_ops_rwsem &root->kernfs_rwsem &cfs_rq->removed.lock irq_context: 0 (wq_completion)wg-crypt-wg2#6 (work_completion)(&peer->transmit_packet_work) rcu_read_lock_bh rcu_read_lock &pool->lock/1 irq_context: 0 (wq_completion)wg-crypt-wg2#6 (work_completion)(&peer->transmit_packet_work) rcu_read_lock_bh rcu_read_lock &pool->lock/1 &obj_hash[i].lock irq_context: 0 (wq_completion)wg-crypt-wg2#6 (work_completion)(&peer->transmit_packet_work) rcu_read_lock_bh rcu_read_lock &pool->lock/1 &p->pi_lock irq_context: 0 (wq_completion)wg-crypt-wg2#6 (work_completion)(&peer->transmit_packet_work) rcu_read_lock_bh rcu_read_lock &pool->lock/1 &p->pi_lock &rq->__lock irq_context: 0 (wq_completion)wg-crypt-wg2#6 (work_completion)(&peer->transmit_packet_work) rcu_read_lock_bh rcu_read_lock &pool->lock/1 &p->pi_lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)wg-kex-wg2#11 (work_completion)(&peer->transmit_handshake_work) &c->lock irq_context: 0 (wq_completion)wg-kex-wg2#11 (work_completion)(&peer->transmit_handshake_work) &peer->endpoint_lock rcu_read_lock_bh rcu_read_lock &ul->lock irq_context: 0 (wq_completion)wg-kex-wg2#11 (work_completion)(&peer->transmit_handshake_work) &peer->endpoint_lock rcu_read_lock_bh rcu_read_lock &obj_hash[i].lock irq_context: 0 (wq_completion)wg-kex-wg2#12 (work_completion)(&({ do { const void *__vpp_verify = (typeof((worker) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __asm__ ("" : "=r"(__ptr) : "0"((typeof(*((worker))) *)((worker)))); (typeof((typeof(*((worker))) *)((worker)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); })->work) &handshake->lock &c->lock irq_context: 0 (wq_completion)wg-kex-wg2#12 (work_completion)(&({ do { const void *__vpp_verify = (typeof((worker) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __asm__ ("" : "=r"(__ptr) : "0"((typeof(*((worker))) *)((worker)))); (typeof((typeof(*((worker))) *)((worker)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); })->work) &c->lock irq_context: 0 (wq_completion)ipv6_addrconf (work_completion)(&(&net->ipv6.addr_chk_work)->work) rtnl_mutex rcu_read_lock_bh &ifa->lock irq_context: 0 (wq_completion)ipv6_addrconf (work_completion)(&(&net->ipv6.addr_chk_work)->work) rtnl_mutex.wait_lock irq_context: 0 (wq_completion)ipv6_addrconf (work_completion)(&(&net->ipv6.addr_chk_work)->work) &p->pi_lock irq_context: 0 (wq_completion)ipv6_addrconf (work_completion)(&(&net->ipv6.addr_chk_work)->work) &p->pi_lock &rq->__lock irq_context: 0 (wq_completion)ipv6_addrconf (work_completion)(&(&net->ipv6.addr_chk_work)->work) &p->pi_lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 pernet_ops_rwsem rtnl_mutex &root->kernfs_rwsem rcu_read_lock &rq->__lock irq_context: 0 pernet_ops_rwsem rtnl_mutex &root->kernfs_rwsem &rq->__lock &cfs_rq->removed.lock irq_context: 0 pernet_ops_rwsem rtnl_mutex &root->kernfs_rwsem rcu_read_lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem rcu_state.barrier_mutex &rq->__lock &cfs_rq->removed.lock irq_context: 0 (wq_completion)events_unbound (work_completion)(&(&kfence_timer)->work) cpu_hotplug_lock jump_label_mutex text_mutex rcu_node_0 irq_context: 0 (wq_completion)wg-kex-wg0#7 (work_completion)(&peer->transmit_handshake_work) &rq->__lock irq_context: 0 pernet_ops_rwsem &rnp->exp_wq[2] irq_context: 0 (wq_completion)wg-kex-wg0#7 (work_completion)(&peer->transmit_handshake_work) kfence_freelist_lock irq_context: 0 (wq_completion)wg-kex-wg0#7 (work_completion)(&peer->transmit_handshake_work) &peer->endpoint_lock rcu_read_lock_bh &meta->lock irq_context: 0 (wq_completion)wg-kex-wg0#7 (work_completion)(&peer->transmit_handshake_work) &peer->endpoint_lock rcu_read_lock_bh kfence_freelist_lock irq_context: 0 (wq_completion)netns net_cleanup_work krc.lock &obj_hash[i].lock irq_context: 0 (wq_completion)netns net_cleanup_work krc.lock &base->lock irq_context: 0 (wq_completion)netns net_cleanup_work krc.lock &base->lock &obj_hash[i].lock irq_context: softirq rcu_callback rcu_read_lock &pool->lock/1 irq_context: softirq rcu_callback rcu_read_lock &pool->lock/1 &obj_hash[i].lock irq_context: softirq rcu_callback rcu_read_lock &pool->lock/1 &p->pi_lock irq_context: softirq rcu_callback rcu_read_lock &pool->lock/1 &p->pi_lock &rq->__lock irq_context: softirq rcu_callback rcu_read_lock &pool->lock/1 &p->pi_lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 dup_mmap_sem &mm->mmap_lock &mm->mmap_lock/1 &anon_vma->rwsem rcu_read_lock &rcu_state.expedited_wq irq_context: 0 dup_mmap_sem &mm->mmap_lock &mm->mmap_lock/1 &anon_vma->rwsem rcu_read_lock &rcu_state.expedited_wq &p->pi_lock irq_context: 0 dup_mmap_sem &mm->mmap_lock &mm->mmap_lock/1 &anon_vma->rwsem rcu_read_lock &rcu_state.expedited_wq &p->pi_lock &rq->__lock irq_context: 0 dup_mmap_sem &mm->mmap_lock &mm->mmap_lock/1 &anon_vma->rwsem rcu_read_lock &rcu_state.expedited_wq &p->pi_lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)tipc_rcv#2 irq_context: 0 (wq_completion)tipc_rcv#2 (work_completion)(&srv->awork) irq_context: 0 (wq_completion)dm_bufio_cache (work_completion)(&(&dm_bufio_cleanup_old_work)->work) dm_bufio_clients_lock &rq->__lock irq_context: 0 (wq_completion)dm_bufio_cache (work_completion)(&(&dm_bufio_cleanup_old_work)->work) dm_bufio_clients_lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 sk_lock-AF_TIPC rcu_node_0 irq_context: 0 sk_lock-AF_TIPC &rcu_state.expedited_wq irq_context: 0 sk_lock-AF_TIPC &rcu_state.expedited_wq &p->pi_lock irq_context: 0 (wq_completion)tipc_rcv#2 (work_completion)(&srv->awork) &srv->idr_lock irq_context: 0 (wq_completion)tipc_rcv#2 (work_completion)(&srv->awork) fs_reclaim irq_context: 0 (wq_completion)tipc_rcv#2 (work_completion)(&srv->awork) fs_reclaim mmu_notifier_invalidate_range_start irq_context: 0 (wq_completion)tipc_rcv#2 (work_completion)(&srv->awork) mmu_notifier_invalidate_range_start irq_context: 0 (wq_completion)tipc_rcv#2 (work_completion)(&srv->awork) &sb->s_type->i_lock_key#8 irq_context: 0 (wq_completion)tipc_rcv#2 (work_completion)(&srv->awork) k-sk_lock-AF_TIPC irq_context: 0 (wq_completion)tipc_rcv#2 (work_completion)(&srv->awork) k-sk_lock-AF_TIPC k-slock-AF_TIPC irq_context: 0 (wq_completion)tipc_rcv#2 (work_completion)(&srv->awork) k-sk_lock-AF_TIPC fs_reclaim irq_context: 0 (wq_completion)tipc_rcv#2 (work_completion)(&srv->awork) k-sk_lock-AF_TIPC fs_reclaim mmu_notifier_invalidate_range_start irq_context: 0 (wq_completion)tipc_rcv#2 (work_completion)(&srv->awork) k-sk_lock-AF_TIPC &c->lock irq_context: 0 (wq_completion)tipc_rcv#2 (work_completion)(&srv->awork) k-sk_lock-AF_TIPC &dir->lock irq_context: 0 (wq_completion)tipc_rcv#2 (work_completion)(&srv->awork) k-sk_lock-AF_TIPC &obj_hash[i].lock irq_context: 0 (wq_completion)tipc_rcv#2 (work_completion)(&srv->awork) k-sk_lock-AF_TIPC batched_entropy_u32.lock irq_context: 0 (wq_completion)tipc_rcv#2 (work_completion)(&srv->awork) k-sk_lock-AF_TIPC rcu_read_lock rhashtable_bucket irq_context: 0 (wq_completion)tipc_rcv#2 (work_completion)(&srv->awork) k-sk_lock-AF_TIPC k-sk_lock-AF_TIPC/1 irq_context: 0 (wq_completion)tipc_rcv#2 (work_completion)(&srv->awork) k-sk_lock-AF_TIPC k-sk_lock-AF_TIPC/1 k-slock-AF_TIPC irq_context: 0 (wq_completion)tipc_rcv#2 (work_completion)(&srv->awork) k-sk_lock-AF_TIPC k-sk_lock-AF_TIPC/1 &obj_hash[i].lock irq_context: 0 (wq_completion)tipc_rcv#2 (work_completion)(&srv->awork) k-sk_lock-AF_TIPC k-sk_lock-AF_TIPC/1 &base->lock irq_context: 0 (wq_completion)tipc_rcv#2 (work_completion)(&srv->awork) k-sk_lock-AF_TIPC k-sk_lock-AF_TIPC/1 &base->lock &obj_hash[i].lock irq_context: 0 (wq_completion)tipc_rcv#2 (work_completion)(&srv->awork) k-sk_lock-AF_TIPC k-sk_lock-AF_TIPC/1 fs_reclaim irq_context: 0 (wq_completion)tipc_rcv#2 (work_completion)(&srv->awork) k-sk_lock-AF_TIPC k-sk_lock-AF_TIPC/1 fs_reclaim mmu_notifier_invalidate_range_start irq_context: 0 (wq_completion)tipc_rcv#2 (work_completion)(&srv->awork) k-sk_lock-AF_TIPC k-sk_lock-AF_TIPC/1 &c->lock irq_context: 0 (wq_completion)tipc_rcv#2 (work_completion)(&srv->awork) k-sk_lock-AF_TIPC k-sk_lock-AF_TIPC/1 &list->lock#31 irq_context: 0 (wq_completion)tipc_rcv#2 (work_completion)(&srv->awork) k-sk_lock-AF_TIPC k-sk_lock-AF_TIPC/1 slock-AF_TIPC &list->lock#31 irq_context: 0 (wq_completion)tipc_rcv#2 (work_completion)(&srv->awork) k-sk_lock-AF_TIPC k-sk_lock-AF_TIPC/1 slock-AF_TIPC &obj_hash[i].lock irq_context: 0 (wq_completion)tipc_rcv#2 (work_completion)(&srv->awork) k-sk_lock-AF_TIPC k-sk_lock-AF_TIPC/1 slock-AF_TIPC &base->lock irq_context: 0 (wq_completion)tipc_rcv#2 (work_completion)(&srv->awork) k-sk_lock-AF_TIPC k-sk_lock-AF_TIPC/1 slock-AF_TIPC &base->lock &obj_hash[i].lock irq_context: 0 (wq_completion)tipc_rcv#2 (work_completion)(&srv->awork) k-sk_lock-AF_TIPC k-sk_lock-AF_TIPC/1 slock-AF_TIPC rcu_read_lock &ei->socket.wq.wait irq_context: 0 (wq_completion)tipc_rcv#2 (work_completion)(&srv->awork) k-sk_lock-AF_TIPC k-sk_lock-AF_TIPC/1 slock-AF_TIPC rcu_read_lock &ei->socket.wq.wait &p->pi_lock irq_context: 0 (wq_completion)tipc_rcv#2 (work_completion)(&srv->awork) k-sk_lock-AF_TIPC k-sk_lock-AF_TIPC/1 slock-AF_TIPC rcu_read_lock &ei->socket.wq.wait &p->pi_lock &cfs_rq->removed.lock irq_context: 0 (wq_completion)tipc_rcv#2 (work_completion)(&srv->awork) k-sk_lock-AF_TIPC k-sk_lock-AF_TIPC/1 slock-AF_TIPC rcu_read_lock &ei->socket.wq.wait &p->pi_lock &rq->__lock irq_context: 0 (wq_completion)tipc_rcv#2 (work_completion)(&srv->awork) k-sk_lock-AF_TIPC k-sk_lock-AF_TIPC/1 slock-AF_TIPC rcu_read_lock &ei->socket.wq.wait &p->pi_lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)tipc_rcv#2 (work_completion)(&srv->awork) k-sk_lock-AF_TIPC k-sk_lock-AF_TIPC/1 &rq->__lock irq_context: 0 (wq_completion)tipc_rcv#2 (work_completion)(&srv->awork) k-sk_lock-AF_TIPC k-sk_lock-AF_TIPC/1 &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)tipc_rcv#2 (work_completion)(&srv->awork) k-slock-AF_TIPC irq_context: 0 (wq_completion)tipc_rcv#2 (work_completion)(&srv->awork) &obj_hash[i].lock irq_context: 0 (wq_completion)tipc_rcv#2 (work_completion)(&srv->awork) &obj_hash[i].lock pool_lock irq_context: 0 (wq_completion)tipc_rcv#2 (work_completion)(&srv->awork) k-clock-AF_TIPC irq_context: 0 (wq_completion)tipc_rcv#2 (work_completion)(&srv->awork) k-clock-AF_TIPC irq_context: 0 (wq_completion)tipc_rcv#2 (work_completion)(&srv->awork) k-clock-AF_TIPC rcu_read_lock &pool->lock/1 irq_context: 0 (wq_completion)tipc_rcv#2 (work_completion)(&srv->awork) k-clock-AF_TIPC rcu_read_lock &pool->lock/1 &obj_hash[i].lock irq_context: 0 (wq_completion)tipc_rcv#2 (work_completion)(&srv->awork) k-clock-AF_TIPC rcu_read_lock &pool->lock/1 &p->pi_lock irq_context: 0 (wq_completion)tipc_rcv#2 (work_completion)(&srv->awork) k-clock-AF_TIPC rcu_read_lock &pool->lock/1 &p->pi_lock &rq->__lock irq_context: 0 (wq_completion)tipc_rcv#2 (work_completion)(&srv->awork) k-clock-AF_TIPC rcu_read_lock &pool->lock/1 &p->pi_lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)tipc_rcv#2 (work_completion)(&srv->awork) &xa->xa_lock#9 irq_context: 0 (wq_completion)tipc_rcv#2 (work_completion)(&srv->awork) &fsnotify_mark_srcu irq_context: 0 (wq_completion)tipc_rcv#2 (work_completion)(&con->rwork) irq_context: 0 (wq_completion)tipc_rcv#2 (work_completion)(&con->rwork) k-sk_lock-AF_TIPC irq_context: 0 (wq_completion)tipc_rcv#2 (work_completion)(&con->rwork) k-sk_lock-AF_TIPC k-slock-AF_TIPC irq_context: 0 (wq_completion)tipc_rcv#2 (work_completion)(&con->rwork) k-slock-AF_TIPC irq_context: 0 (wq_completion)tipc_rcv#2 (work_completion)(&con->rwork) k-sk_lock-AF_TIPC &obj_hash[i].lock irq_context: 0 (wq_completion)tipc_rcv#2 (work_completion)(&con->rwork) k-sk_lock-AF_TIPC k-clock-AF_TIPC irq_context: 0 sk_lock-AF_TIPC fs_reclaim mmu_notifier_invalidate_range_start &rq->__lock irq_context: 0 sk_lock-AF_TIPC fs_reclaim mmu_notifier_invalidate_range_start &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)tipc_rcv#2 &rq->__lock irq_context: 0 (wq_completion)tipc_rcv#2 &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)tipc_rcv#2 (work_completion)(&con->rwork) k-sk_lock-AF_TIPC k-clock-AF_TIPC rcu_read_lock &pool->lock/1 irq_context: 0 (wq_completion)tipc_rcv#2 (work_completion)(&con->rwork) k-sk_lock-AF_TIPC k-clock-AF_TIPC rcu_read_lock &pool->lock/1 &obj_hash[i].lock irq_context: 0 (wq_completion)tipc_rcv#2 (work_completion)(&con->rwork) k-sk_lock-AF_TIPC k-clock-AF_TIPC rcu_read_lock &pool->lock/1 &p->pi_lock irq_context: 0 (wq_completion)tipc_rcv#2 (work_completion)(&con->rwork) k-sk_lock-AF_TIPC k-clock-AF_TIPC rcu_read_lock &pool->lock/1 &p->pi_lock &rq->__lock irq_context: 0 (wq_completion)tipc_rcv#2 (work_completion)(&con->rwork) k-sk_lock-AF_TIPC k-clock-AF_TIPC rcu_read_lock &pool->lock/1 &p->pi_lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)tipc_rcv#2 (work_completion)(&con->rwork) k-sk_lock-AF_TIPC &rq->__lock irq_context: 0 (wq_completion)tipc_rcv#2 (work_completion)(&con->rwork) k-sk_lock-AF_TIPC &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)tipc_rcv#2 (work_completion)(&con->rwork) k-sk_lock-AF_TIPC k-clock-AF_TIPC rcu_read_lock &pool->lock/1 &p->pi_lock &cfs_rq->removed.lock irq_context: 0 sk_lock-AF_TIPC &rcu_state.expedited_wq &p->pi_lock &rq->__lock irq_context: 0 sk_lock-AF_TIPC &rcu_state.expedited_wq &p->pi_lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 sk_lock-AF_TIPC remove_cache_srcu pool_lock#2 irq_context: 0 sk_lock-AF_TIPC remove_cache_srcu rcu_node_0 irq_context: 0 sk_lock-AF_TIPC remove_cache_srcu &rcu_state.expedited_wq irq_context: 0 sk_lock-AF_TIPC remove_cache_srcu &rcu_state.expedited_wq &p->pi_lock irq_context: 0 sk_lock-AF_TIPC remove_cache_srcu &rcu_state.expedited_wq &p->pi_lock &rq->__lock irq_context: 0 sk_lock-AF_TIPC remove_cache_srcu &rcu_state.expedited_wq &p->pi_lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 sb_writers#4 &type->i_mutex_dir_key#3/1 tomoyo_ss rcu_read_lock &cfs_rq->removed.lock irq_context: 0 sb_writers#4 &type->i_mutex_dir_key#3/1 tomoyo_ss rcu_read_lock &obj_hash[i].lock irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem rtnl_mutex rcu_state.exp_mutex &rq->__lock &cfs_rq->removed.lock irq_context: 0 &sbi->s_writepages_rwsem lock#4 rcu_read_lock pool_lock#2 irq_context: 0 &sbi->s_writepages_rwsem lock#4 &obj_hash[i].lock irq_context: 0 sb_writers#4 sb_internal jbd2_handle &rcu_state.expedited_wq irq_context: 0 sb_writers#4 sb_internal jbd2_handle &rcu_state.expedited_wq &p->pi_lock irq_context: 0 sb_writers#4 sb_internal jbd2_handle &rcu_state.expedited_wq &p->pi_lock &rq->__lock irq_context: 0 sb_writers#4 sb_internal jbd2_handle &rcu_state.expedited_wq &p->pi_lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 sb_writers#4 &type->i_mutex_dir_key#3/1 sb_internal remove_cache_srcu rcu_read_lock rcu_node_0 irq_context: 0 sb_writers#4 &type->i_mutex_dir_key#3/1 sb_internal remove_cache_srcu rcu_read_lock &rq->__lock irq_context: 0 sb_writers#4 &type->i_mutex_dir_key#3/1 sb_internal remove_cache_srcu rcu_read_lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 dup_mmap_sem &mm->mmap_lock &mm->mmap_lock/1 &anon_vma->rwsem &rcu_state.expedited_wq irq_context: 0 dup_mmap_sem &mm->mmap_lock &mm->mmap_lock/1 &anon_vma->rwsem &rcu_state.expedited_wq &p->pi_lock irq_context: 0 dup_mmap_sem &mm->mmap_lock &mm->mmap_lock/1 &anon_vma->rwsem &rcu_state.expedited_wq &p->pi_lock &rq->__lock irq_context: 0 dup_mmap_sem &mm->mmap_lock &mm->mmap_lock/1 &anon_vma->rwsem &rcu_state.expedited_wq &p->pi_lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 sb_writers#4 &type->i_mutex_dir_key#3/1 tomoyo_ss rcu_read_lock &rcu_state.expedited_wq irq_context: 0 sb_writers#4 &type->i_mutex_dir_key#3/1 tomoyo_ss rcu_read_lock &rcu_state.expedited_wq &p->pi_lock irq_context: 0 sb_writers#4 &type->i_mutex_dir_key#3/1 tomoyo_ss rcu_read_lock &rcu_state.expedited_wq &p->pi_lock &rq->__lock irq_context: 0 sb_writers#4 &type->i_mutex_dir_key#3/1 tomoyo_ss rcu_read_lock &rcu_state.expedited_wq &p->pi_lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem rtnl_mutex rcu_state.exp_mutex &cfs_rq->removed.lock irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem rtnl_mutex rcu_state.exp_mutex pool_lock#2 irq_context: 0 &pipe->mutex/1 sk_lock-AF_INET6 batched_entropy_u8.lock irq_context: 0 &pipe->mutex/1 sk_lock-AF_INET6 kfence_freelist_lock irq_context: 0 &pipe->mutex/1 sk_lock-AF_INET6 rcu_read_lock &meta->lock irq_context: 0 &pipe->mutex/1 sk_lock-AF_INET6 rcu_read_lock kfence_freelist_lock irq_context: 0 &f->f_pos_lock sb_writers#4 &sb->s_type->i_mutex_key#8 &folio_wait_table[i] &p->pi_lock &cfs_rq->removed.lock irq_context: 0 (wq_completion)tipc_rcv#2 (work_completion)(&con->rwork) k-clock-AF_TIPC irq_context: 0 (wq_completion)tipc_rcv#2 (work_completion)(&con->rwork) k-clock-AF_TIPC &con->sub_lock irq_context: 0 (wq_completion)tipc_rcv#2 (work_completion)(&con->rwork) &srv->idr_lock irq_context: 0 (wq_completion)tipc_rcv#2 (work_completion)(&con->rwork) &srv->idr_lock &obj_hash[i].lock irq_context: 0 (wq_completion)tipc_rcv#2 (work_completion)(&con->rwork) k-sk_lock-AF_TIPC &base->lock irq_context: 0 (wq_completion)tipc_rcv#2 (work_completion)(&con->rwork) k-sk_lock-AF_TIPC &base->lock &obj_hash[i].lock irq_context: 0 (wq_completion)tipc_rcv#2 (work_completion)(&con->rwork) k-sk_lock-AF_TIPC rcu_read_lock rhashtable_bucket irq_context: 0 (wq_completion)tipc_rcv#2 (work_completion)(&con->rwork) k-sk_lock-AF_TIPC k-clock-AF_TIPC irq_context: 0 (wq_completion)tipc_rcv#2 (work_completion)(&con->rwork) &obj_hash[i].lock irq_context: 0 (wq_completion)tipc_rcv#2 (work_completion)(&con->rwork) &sb->s_type->i_lock_key#8 irq_context: 0 (wq_completion)tipc_rcv#2 (work_completion)(&con->rwork) &xa->xa_lock#9 irq_context: 0 (wq_completion)tipc_rcv#2 (work_completion)(&con->rwork) &fsnotify_mark_srcu irq_context: 0 (wq_completion)tipc_rcv#2 (work_completion)(&con->rwork) &con->outqueue_lock irq_context: 0 sb_writers#4 sb_internal remove_cache_srcu rcu_read_lock rcu_node_0 irq_context: 0 sb_writers#4 sb_internal remove_cache_srcu rcu_read_lock &rq->__lock irq_context: 0 sb_writers#4 sb_internal remove_cache_srcu rcu_read_lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 sb_writers#4 &type->i_mutex_dir_key#3/1 &type->i_mutex_dir_key#3 batched_entropy_u8.lock irq_context: 0 sb_writers#4 &type->i_mutex_dir_key#3/1 &type->i_mutex_dir_key#3 kfence_freelist_lock irq_context: 0 sb_writers#4 &type->i_mutex_dir_key#3/1 &type->i_mutex_dir_key#3 &meta->lock irq_context: 0 sb_writers#3 &mm->mmap_lock &rq->__lock irq_context: 0 sb_writers#3 &mm->mmap_lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 pernet_ops_rwsem nf_ct_proto_mutex nf_hook_mutex rcu_read_lock rcu_node_0 irq_context: 0 pernet_ops_rwsem nf_ct_proto_mutex nf_hook_mutex rcu_read_lock &rq->__lock irq_context: 0 pernet_ops_rwsem rtnl_mutex fill_pool_map-wait-type-override rcu_read_lock pool_lock#2 irq_context: 0 pernet_ops_rwsem rtnl_mutex fill_pool_map-wait-type-override &obj_hash[i].lock irq_context: 0 (wq_completion)bat_events (work_completion)(&(&bat_priv->nc.work)->work) &base->lock fill_pool_map-wait-type-override &c->lock irq_context: 0 pernet_ops_rwsem cpu_hotplug_lock wq_pool_mutex fill_pool_map-wait-type-override &c->lock irq_context: softirq (&ndev->rs_timer) &pcp->lock &zone->lock irq_context: 0 (wq_completion)wg-crypt-wg2 (work_completion)(&peer->transmit_packet_work) rcu_read_lock_bh rcu_read_lock &pool->lock/1 &p->pi_lock &rq->__lock irq_context: 0 (wq_completion)wg-crypt-wg2 (work_completion)(&peer->transmit_packet_work) rcu_read_lock_bh rcu_read_lock &pool->lock/1 &p->pi_lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)wg-crypt-wg2 (work_completion)(&peer->transmit_packet_work) &rq->__lock irq_context: 0 (wq_completion)wg-kex-wg2 (work_completion)(&peer->transmit_handshake_work) &peer->endpoint_lock rcu_read_lock_bh rcu_read_lock &c->lock irq_context: 0 (wq_completion)wg-kex-wg2 (work_completion)(&peer->transmit_handshake_work) &peer->endpoint_lock rcu_read_lock_bh rcu_read_lock &obj_hash[i].lock irq_context: 0 (wq_completion)events_power_efficient (work_completion)(&(&tbl->gc_work)->work) &tbl->lock fill_pool_map-wait-type-override pool_lock#2 irq_context: 0 (wq_completion)wg-kex-wg2 (work_completion)(&peer->transmit_handshake_work) &peer->endpoint_lock rcu_read_lock_bh rcu_read_lock &tbl->lock irq_context: 0 (wq_completion)events_power_efficient (work_completion)(&(&tbl->gc_work)->work) &tbl->lock fill_pool_map-wait-type-override pool_lock irq_context: 0 (wq_completion)wg-kex-wg2 (work_completion)(&peer->transmit_handshake_work) &peer->endpoint_lock rcu_read_lock_bh rcu_read_lock &n->lock irq_context: 0 (wq_completion)wg-kex-wg2 (work_completion)(&peer->transmit_handshake_work) &peer->endpoint_lock rcu_read_lock_bh rcu_read_lock &____s->seqcount#9 irq_context: 0 (wq_completion)wg-kex-wg0#2 (work_completion)(&({ do { const void *__vpp_verify = (typeof((worker) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __asm__ ("" : "=r"(__ptr) : "0"((typeof(*((worker))) *)((worker)))); (typeof((typeof(*((worker))) *)((worker)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); })->work) &handshake->lock &n->list_lock irq_context: 0 (wq_completion)wg-kex-wg0#2 (work_completion)(&({ do { const void *__vpp_verify = (typeof((worker) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __asm__ ("" : "=r"(__ptr) : "0"((typeof(*((worker))) *)((worker)))); (typeof((typeof(*((worker))) *)((worker)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); })->work) &handshake->lock &n->list_lock &c->lock irq_context: 0 (wq_completion)wg-kex-wg2#2 (work_completion)(&({ do { const void *__vpp_verify = (typeof((worker) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __asm__ ("" : "=r"(__ptr) : "0"((typeof(*((worker))) *)((worker)))); (typeof((typeof(*((worker))) *)((worker)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); })->work) &c->lock irq_context: 0 (wq_completion)wg-crypt-wg2#3 (work_completion)(&peer->transmit_packet_work) rcu_read_lock_bh rcu_read_lock &pool->lock/1 irq_context: 0 (wq_completion)wg-kex-wg2#2 (work_completion)(&({ do { const void *__vpp_verify = (typeof((worker) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __asm__ ("" : "=r"(__ptr) : "0"((typeof(*((worker))) *)((worker)))); (typeof((typeof(*((worker))) *)((worker)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); })->work) &____s->seqcount#2 irq_context: 0 (wq_completion)wg-crypt-wg2#3 (work_completion)(&peer->transmit_packet_work) rcu_read_lock_bh rcu_read_lock &pool->lock/1 &obj_hash[i].lock irq_context: 0 (wq_completion)wg-kex-wg2#2 (work_completion)(&({ do { const void *__vpp_verify = (typeof((worker) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __asm__ ("" : "=r"(__ptr) : "0"((typeof(*((worker))) *)((worker)))); (typeof((typeof(*((worker))) *)((worker)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); })->work) &____s->seqcount irq_context: 0 (wq_completion)wg-crypt-wg2#3 (work_completion)(&peer->transmit_packet_work) rcu_read_lock_bh rcu_read_lock &pool->lock/1 &p->pi_lock irq_context: 0 (wq_completion)wg-kex-wg2#2 (work_completion)(&({ do { const void *__vpp_verify = (typeof((worker) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __asm__ ("" : "=r"(__ptr) : "0"((typeof(*((worker))) *)((worker)))); (typeof((typeof(*((worker))) *)((worker)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); })->work) pool_lock#2 irq_context: 0 (wq_completion)wg-crypt-wg2#3 (work_completion)(&peer->transmit_packet_work) rcu_read_lock_bh rcu_read_lock &pool->lock/1 &p->pi_lock &rq->__lock irq_context: 0 (wq_completion)wg-crypt-wg2#3 (work_completion)(&peer->transmit_packet_work) rcu_read_lock_bh rcu_read_lock &pool->lock/1 &p->pi_lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)wg-kex-wg2#5 (work_completion)(&peer->transmit_handshake_work) &peer->endpoint_lock rcu_read_lock_bh rcu_read_lock &c->lock irq_context: 0 (wq_completion)wg-kex-wg2#5 (work_completion)(&peer->transmit_handshake_work) batched_entropy_u8.lock crngs.lock irq_context: 0 (wq_completion)wg-kex-wg1#6 (work_completion)(&({ do { const void *__vpp_verify = (typeof((worker) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __asm__ ("" : "=r"(__ptr) : "0"((typeof(*((worker))) *)((worker)))); (typeof((typeof(*((worker))) *)((worker)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); })->work) &handshake->lock &c->lock irq_context: 0 &fsnotify_mark_srcu remove_cache_srcu &rq->__lock irq_context: 0 &fsnotify_mark_srcu remove_cache_srcu &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 pernet_ops_rwsem devices_rwsem rdma_nets_rwsem &device->compat_devs_mutex fs_reclaim &cfs_rq->removed.lock irq_context: 0 pernet_ops_rwsem devices_rwsem rdma_nets_rwsem &device->compat_devs_mutex fs_reclaim &obj_hash[i].lock irq_context: 0 &vma->vm_lock->lock ptlock_ptr(page)#2 lock#4 &lruvec->lru_lock per_cpu_ptr(&cgroup_rstat_cpu_lock, cpu) irq_context: 0 pernet_ops_rwsem rtnl_mutex net_rwsem rcu_node_0 irq_context: 0 pernet_ops_rwsem &wq->mutex &x->wait#10 &p->pi_lock irq_context: 0 (wq_completion)events_power_efficient (work_completion)(&(&cache_cleaner)->work) &rq->__lock irq_context: 0 pernet_ops_rwsem rtnl_mutex uevent_sock_mutex rcu_node_0 irq_context: 0 pernet_ops_rwsem &wq->mutex &x->wait#10 &p->pi_lock &rq->__lock irq_context: 0 pernet_ops_rwsem &wq->mutex &x->wait#10 &p->pi_lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 pernet_ops_rwsem &wq->mutex rcu_node_0 irq_context: 0 pernet_ops_rwsem &wq->mutex &rcu_state.expedited_wq irq_context: 0 pernet_ops_rwsem &wq->mutex &rcu_state.expedited_wq &p->pi_lock irq_context: 0 pernet_ops_rwsem &wq->mutex &rcu_state.expedited_wq &p->pi_lock &rq->__lock irq_context: 0 pernet_ops_rwsem &wq->mutex &rcu_state.expedited_wq &p->pi_lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 pernet_ops_rwsem &wq->mutex &rq->__lock irq_context: 0 pernet_ops_rwsem &wq->mutex &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 pernet_ops_rwsem remove_cache_srcu rcu_read_lock &rcu_state.expedited_wq &p->pi_lock &rq->__lock irq_context: 0 pernet_ops_rwsem remove_cache_srcu rcu_read_lock &rcu_state.expedited_wq &p->pi_lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)events_unbound (work_completion)(&(&kfence_timer)->work) rcu_read_lock &rq->__lock &cfs_rq->removed.lock irq_context: 0 (wq_completion)events_unbound (work_completion)(&(&kfence_timer)->work) cpu_hotplug_lock jump_label_mutex text_mutex &rcu_state.expedited_wq irq_context: 0 (wq_completion)events_unbound (work_completion)(&(&kfence_timer)->work) cpu_hotplug_lock jump_label_mutex text_mutex &rcu_state.expedited_wq &p->pi_lock irq_context: 0 pernet_ops_rwsem remove_cache_srcu &rcu_state.expedited_wq irq_context: 0 pernet_ops_rwsem remove_cache_srcu &rcu_state.expedited_wq &p->pi_lock irq_context: 0 pernet_ops_rwsem remove_cache_srcu &rcu_state.expedited_wq &p->pi_lock &rq->__lock irq_context: 0 pernet_ops_rwsem remove_cache_srcu &rcu_state.expedited_wq &p->pi_lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 pernet_ops_rwsem rcu_state.barrier_mutex rcu_read_lock &rq->__lock &cfs_rq->removed.lock irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem rtnl_mutex &idev->mc_lock _xmit_ETHER &obj_hash[i].lock pool_lock irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem rtnl_mutex remove_cache_srcu &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem rtnl_mutex &root->kernfs_rwsem &rq->__lock &cfs_rq->removed.lock irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem rtnl_mutex &root->kernfs_rwsem rcu_read_lock &rq->__lock irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem rtnl_mutex &root->kernfs_rwsem rcu_read_lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 pernet_ops_rwsem kernfs_idr_lock &obj_hash[i].lock pool_lock irq_context: softirq (&hsr->announce_timer) rcu_read_lock batched_entropy_u8.lock crngs.lock irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem uevent_sock_mutex rcu_read_lock &ei->socket.wq.wait &ep->lock &ep->wq &p->pi_lock &cfs_rq->removed.lock irq_context: softirq (&peer->timer_persistent_keepalive) init_task.mems_allowed_seq.seqcount irq_context: 0 pernet_ops_rwsem __ip_vs_app_mutex &obj_hash[i].lock pool_lock irq_context: 0 (wq_completion)events free_ipc_work &ht->mutex &rq->__lock irq_context: 0 (wq_completion)events (work_completion)(&(&nsim_dev->trap_data->trap_report_dw)->work) &devlink->lock_key#6 &nsim_trap_data->trap_lock &base->lock irq_context: 0 (wq_completion)events (work_completion)(&(&nsim_dev->trap_data->trap_report_dw)->work) &devlink->lock_key#6 &nsim_trap_data->trap_lock &base->lock &obj_hash[i].lock irq_context: 0 pernet_ops_rwsem rtnl_mutex nl_table_wait.lock &p->pi_lock &cfs_rq->removed.lock irq_context: 0 pernet_ops_rwsem devices_rwsem rdma_nets_rwsem &device->compat_devs_mutex &root->kernfs_rwsem pool_lock#2 irq_context: 0 &lo->lo_mutex &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 pernet_ops_rwsem devices_rwsem rdma_nets_rwsem &device->compat_devs_mutex rcu_read_lock &cfs_rq->removed.lock irq_context: 0 pernet_ops_rwsem devices_rwsem rdma_nets_rwsem &device->compat_devs_mutex rcu_read_lock &obj_hash[i].lock irq_context: 0 &fsnotify_mark_srcu batched_entropy_u8.lock irq_context: 0 &fsnotify_mark_srcu kfence_freelist_lock irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem rcu_state.barrier_mutex rcu_node_0 irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem rcu_state.barrier_mutex &rcu_state.expedited_wq irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem rcu_state.barrier_mutex &rcu_state.expedited_wq &p->pi_lock irq_context: 0 (wq_completion)wg-crypt-wg2#4 (work_completion)(&peer->transmit_packet_work) rcu_read_lock_bh rcu_read_lock &pool->lock/1 irq_context: 0 (wq_completion)wg-crypt-wg2#4 (work_completion)(&peer->transmit_packet_work) rcu_read_lock_bh rcu_read_lock &pool->lock/1 &obj_hash[i].lock irq_context: 0 (wq_completion)wg-crypt-wg2#4 (work_completion)(&peer->transmit_packet_work) rcu_read_lock_bh rcu_read_lock &pool->lock/1 &p->pi_lock irq_context: 0 (wq_completion)wg-kex-wg0#12 (work_completion)(&({ do { const void *__vpp_verify = (typeof((worker) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __asm__ ("" : "=r"(__ptr) : "0"((typeof(*((worker))) *)((worker)))); (typeof((typeof(*((worker))) *)((worker)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); })->work) &handshake->lock &c->lock irq_context: 0 (wq_completion)wg-kex-wg0#12 (work_completion)(&({ do { const void *__vpp_verify = (typeof((worker) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __asm__ ("" : "=r"(__ptr) : "0"((typeof(*((worker))) *)((worker)))); (typeof((typeof(*((worker))) *)((worker)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); })->work) &handshake->lock rcu_read_lock_bh &peer->keypairs.keypair_update_lock &table->lock#2 irq_context: 0 (wq_completion)wg-kex-wg0#12 (work_completion)(&({ do { const void *__vpp_verify = (typeof((worker) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __asm__ ("" : "=r"(__ptr) : "0"((typeof(*((worker))) *)((worker)))); (typeof((typeof(*((worker))) *)((worker)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); })->work) &handshake->lock rcu_read_lock_bh &peer->keypairs.keypair_update_lock &obj_hash[i].lock irq_context: 0 (wq_completion)wg-crypt-wg2#4 (work_completion)(&peer->transmit_packet_work) rcu_read_lock_bh rcu_read_lock &pool->lock/1 &p->pi_lock &rq->__lock irq_context: 0 (wq_completion)wg-kex-wg0#12 (work_completion)(&({ do { const void *__vpp_verify = (typeof((worker) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __asm__ ("" : "=r"(__ptr) : "0"((typeof(*((worker))) *)((worker)))); (typeof((typeof(*((worker))) *)((worker)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); })->work) &handshake->lock rcu_read_lock_bh &peer->keypairs.keypair_update_lock pool_lock#2 irq_context: 0 (wq_completion)wg-crypt-wg2#4 (work_completion)(&peer->transmit_packet_work) rcu_read_lock_bh rcu_read_lock &pool->lock/1 &p->pi_lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)wg-kex-wg0#6 (work_completion)(&({ do { const void *__vpp_verify = (typeof((worker) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __asm__ ("" : "=r"(__ptr) : "0"((typeof(*((worker))) *)((worker)))); (typeof((typeof(*((worker))) *)((worker)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); })->work) &c->lock irq_context: 0 (wq_completion)wg-kex-wg2#6 (work_completion)(&({ do { const void *__vpp_verify = (typeof((worker) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __asm__ ("" : "=r"(__ptr) : "0"((typeof(*((worker))) *)((worker)))); (typeof((typeof(*((worker))) *)((worker)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); })->work) &handshake->lock &c->lock irq_context: 0 (wq_completion)wg-kex-wg2#8 (work_completion)(&({ do { const void *__vpp_verify = (typeof((worker) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __asm__ ("" : "=r"(__ptr) : "0"((typeof(*((worker))) *)((worker)))); (typeof((typeof(*((worker))) *)((worker)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); })->work) &handshake->lock rcu_read_lock_bh &peer->keypairs.keypair_update_lock &table->lock#2 irq_context: 0 (wq_completion)wg-kex-wg2#6 (work_completion)(&({ do { const void *__vpp_verify = (typeof((worker) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __asm__ ("" : "=r"(__ptr) : "0"((typeof(*((worker))) *)((worker)))); (typeof((typeof(*((worker))) *)((worker)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); })->work) &handshake->lock &n->list_lock irq_context: 0 (wq_completion)wg-kex-wg2#8 (work_completion)(&({ do { const void *__vpp_verify = (typeof((worker) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __asm__ ("" : "=r"(__ptr) : "0"((typeof(*((worker))) *)((worker)))); (typeof((typeof(*((worker))) *)((worker)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); })->work) &handshake->lock rcu_read_lock_bh &peer->keypairs.keypair_update_lock &obj_hash[i].lock irq_context: 0 (wq_completion)wg-kex-wg2#6 (work_completion)(&({ do { const void *__vpp_verify = (typeof((worker) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __asm__ ("" : "=r"(__ptr) : "0"((typeof(*((worker))) *)((worker)))); (typeof((typeof(*((worker))) *)((worker)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); })->work) &handshake->lock &n->list_lock &c->lock irq_context: 0 (wq_completion)wg-kex-wg2#8 (work_completion)(&({ do { const void *__vpp_verify = (typeof((worker) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __asm__ ("" : "=r"(__ptr) : "0"((typeof(*((worker))) *)((worker)))); (typeof((typeof(*((worker))) *)((worker)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); })->work) &handshake->lock rcu_read_lock_bh &peer->keypairs.keypair_update_lock pool_lock#2 irq_context: 0 (wq_completion)wg-kex-wg0#11 (work_completion)(&({ do { const void *__vpp_verify = (typeof((worker) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __asm__ ("" : "=r"(__ptr) : "0"((typeof(*((worker))) *)((worker)))); (typeof((typeof(*((worker))) *)((worker)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); })->work) &handshake->lock &c->lock irq_context: 0 (wq_completion)wg-kex-wg0#11 (work_completion)(&({ do { const void *__vpp_verify = (typeof((worker) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __asm__ ("" : "=r"(__ptr) : "0"((typeof(*((worker))) *)((worker)))); (typeof((typeof(*((worker))) *)((worker)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); })->work) &handshake->lock &n->list_lock irq_context: 0 (wq_completion)wg-kex-wg0#11 (work_completion)(&({ do { const void *__vpp_verify = (typeof((worker) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __asm__ ("" : "=r"(__ptr) : "0"((typeof(*((worker))) *)((worker)))); (typeof((typeof(*((worker))) *)((worker)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); })->work) &handshake->lock &n->list_lock &c->lock irq_context: 0 (wq_completion)wg-kex-wg2#12 (work_completion)(&({ do { const void *__vpp_verify = (typeof((worker) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __asm__ ("" : "=r"(__ptr) : "0"((typeof(*((worker))) *)((worker)))); (typeof((typeof(*((worker))) *)((worker)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); })->work) rcu_read_lock_bh rcu_read_lock &pool->lock &p->pi_lock &rq->__lock irq_context: 0 (wq_completion)wg-crypt-wg1#2 (work_completion)(&peer->transmit_packet_work) &rq->__lock irq_context: 0 (wq_completion)wg-kex-wg2#12 (work_completion)(&({ do { const void *__vpp_verify = (typeof((worker) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __asm__ ("" : "=r"(__ptr) : "0"((typeof(*((worker))) *)((worker)))); (typeof((typeof(*((worker))) *)((worker)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); })->work) rcu_read_lock_bh rcu_read_lock &pool->lock &p->pi_lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem remove_cache_srcu &____s->seqcount irq_context: 0 (wq_completion)rcu_gp (work_completion)(&rew->rew_work) rcu_state.exp_wake_mutex rcu_read_lock &rq->__lock irq_context: 0 (wq_completion)rcu_gp (work_completion)(&rew->rew_work) rcu_state.exp_wake_mutex rcu_read_lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: softirq (&app->join_timer) rcu_read_lock_bh &meta->lock irq_context: softirq (&app->join_timer) rcu_read_lock_bh kfence_freelist_lock irq_context: 0 sb_writers#4 &type->i_mutex_dir_key#3/1 &type->i_mutex_dir_key#3 remove_cache_srcu rcu_read_lock rcu_node_0 irq_context: 0 sb_writers#4 &type->i_mutex_dir_key#3/1 &type->i_mutex_dir_key#3 remove_cache_srcu rcu_read_lock &rq->__lock irq_context: 0 sb_writers#4 &type->i_mutex_dir_key#3/1 &type->i_mutex_dir_key#3 remove_cache_srcu rcu_read_lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 rcu_read_lock &sighand->siglock &sighand->signalfd_wqh irq_context: 0 rcu_read_lock &sighand->siglock &sighand->signalfd_wqh &ep->lock irq_context: 0 rcu_read_lock &sighand->siglock stock_lock irq_context: 0 rcu_read_lock &sighand->siglock &p->pi_lock &cfs_rq->removed.lock irq_context: 0 (wq_completion)bat_events (work_completion)(&(&forw_packet_aggr->delayed_work)->work) &hard_iface->bat_iv.ogm_buff_mutex &rq->__lock &cfs_rq->removed.lock irq_context: 0 rcu_read_lock &sighand->siglock &n->list_lock irq_context: 0 rcu_read_lock &sighand->siglock &n->list_lock &c->lock irq_context: 0 sb_writers#4 &type->i_mutex_dir_key#3/1 &rcu_state.expedited_wq irq_context: 0 rcu_read_lock &sighand->siglock &____s->seqcount#2 irq_context: 0 &ei->i_data_sem quarantine_lock irq_context: 0 &sighand->siglock quarantine_lock irq_context: 0 rcu_read_lock &sighand->siglock &pcp->lock &zone->lock irq_context: 0 &journal->j_state_lock &journal->j_wait_commit &p->pi_lock &cfs_rq->removed.lock irq_context: 0 delayed_uprobe_lock &rq->__lock &cfs_rq->removed.lock irq_context: 0 (wq_completion)events (work_completion)(&rdev->wiphy_work) &rdev->wiphy.mtx &rcu_state.expedited_wq irq_context: 0 &group->mark_mutex &fsnotify_mark_srcu &rq->__lock irq_context: 0 &group->mark_mutex &fsnotify_mark_srcu &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 sb_writers#8 &sb->s_type->i_mutex_key#13 &root->kernfs_iattr_rwsem &rq->__lock irq_context: 0 sb_writers#8 &sb->s_type->i_mutex_key#13 &root->kernfs_iattr_rwsem &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: softirq (&icsk->icsk_delack_timer) slock-AF_INET rcu_read_lock rcu_read_lock rcu_read_lock_bh dev->qdisc_tx_busylock ?: &qdisc_tx_busylock _xmit_ETHER#2 rcu_read_lock &n->list_lock irq_context: softirq (&icsk->icsk_delack_timer) slock-AF_INET rcu_read_lock rcu_read_lock rcu_read_lock_bh dev->qdisc_tx_busylock ?: &qdisc_tx_busylock _xmit_ETHER#2 rcu_read_lock &n->list_lock &c->lock irq_context: 0 rcu_read_lock &sighand->siglock per_cpu_ptr(&cgroup_rstat_cpu_lock, cpu) irq_context: 0 &ei->i_data_sem irq_context: 0 &sbi->s_writepages_rwsem jbd2_handle &ei->i_data_sem &lg->lg_mutex irq_context: 0 &sbi->s_writepages_rwsem jbd2_handle &ei->i_data_sem &lg->lg_mutex &ei->i_prealloc_lock irq_context: 0 &sbi->s_writepages_rwsem jbd2_handle &ei->i_data_sem &lg->lg_mutex rcu_read_lock &pa->pa_lock irq_context: 0 &sbi->s_writepages_rwsem jbd2_handle &ei->i_data_sem &lg->lg_mutex &mapping->private_lock irq_context: 0 &sbi->s_writepages_rwsem jbd2_handle &ei->i_data_sem &lg->lg_mutex mmu_notifier_invalidate_range_start irq_context: 0 &sbi->s_writepages_rwsem jbd2_handle &ei->i_data_sem &lg->lg_mutex pool_lock#2 irq_context: 0 &sbi->s_writepages_rwsem jbd2_handle &ei->i_data_sem &lg->lg_mutex &ret->b_state_lock irq_context: 0 &sbi->s_writepages_rwsem jbd2_handle &ei->i_data_sem &lg->lg_mutex &ret->b_state_lock &journal->j_list_lock irq_context: 0 &sbi->s_writepages_rwsem jbd2_handle &ei->i_data_sem &lg->lg_mutex &pa->pa_lock irq_context: 0 &sbi->s_writepages_rwsem jbd2_handle &ei->i_data_sem &lg->lg_mutex &lg->lg_prealloc_lock irq_context: 0 sb_writers#4 &sb->s_type->i_mutex_key#8 &sb->s_type->i_mutex_key#8/4 irq_context: 0 sb_writers#4 &sb->s_type->i_mutex_key#8 &sb->s_type->i_mutex_key#8/4 &sem->wait_lock irq_context: 0 sb_writers#4 &sb->s_type->i_mutex_key#8 &sb->s_type->i_mutex_key#8/4 &rq->__lock irq_context: 0 sb_writers#4 &sb->s_type->i_mutex_key#8 &sb->s_type->i_mutex_key#8/4 &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 sb_writers#4 &sbi->s_writepages_rwsem irq_context: 0 sb_writers#4 &sbi->s_writepages_rwsem mmu_notifier_invalidate_range_start irq_context: 0 sb_writers#4 &sbi->s_writepages_rwsem batched_entropy_u8.lock irq_context: 0 sb_writers#4 &sbi->s_writepages_rwsem kfence_freelist_lock irq_context: 0 sb_writers#4 &sbi->s_writepages_rwsem lock#4 irq_context: 0 sb_writers#4 &sbi->s_writepages_rwsem lock#5 irq_context: 0 sb_writers#4 &sbi->s_writepages_rwsem &obj_hash[i].lock irq_context: 0 sb_writers#4 &sbi->s_writepages_rwsem &meta->lock irq_context: 0 sb_writers#4 &sbi->s_writepages_rwsem pool_lock#2 irq_context: 0 sb_writers#4 &sbi->s_writepages_rwsem &rq->__lock irq_context: 0 sb_writers#4 &sbi->s_writepages_rwsem &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 sb_writers#4 &sbi->s_writepages_rwsem &journal->j_state_lock irq_context: 0 sb_writers#4 &sbi->s_writepages_rwsem jbd2_handle irq_context: 0 sb_writers#4 &sbi->s_writepages_rwsem jbd2_handle lock#4 irq_context: 0 sb_writers#4 &sbi->s_writepages_rwsem jbd2_handle lock#4 &lruvec->lru_lock irq_context: 0 sb_writers#4 &sbi->s_writepages_rwsem jbd2_handle &rq->__lock irq_context: 0 sb_writers#4 &sbi->s_writepages_rwsem jbd2_handle &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 sb_writers#4 &sbi->s_writepages_rwsem jbd2_handle lock#5 irq_context: 0 sb_writers#4 &sbi->s_writepages_rwsem jbd2_handle rcu_read_lock rcu_node_0 irq_context: 0 sb_writers#4 &sbi->s_writepages_rwsem jbd2_handle rcu_read_lock &rq->__lock irq_context: 0 sb_writers#4 &sbi->s_writepages_rwsem jbd2_handle rcu_read_lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 sb_writers#4 &sbi->s_writepages_rwsem jbd2_handle rcu_read_lock &rcu_state.gp_wq irq_context: 0 sb_writers#4 &sbi->s_writepages_rwsem jbd2_handle rcu_read_lock &rcu_state.gp_wq &p->pi_lock irq_context: 0 sb_writers#4 &sbi->s_writepages_rwsem jbd2_handle rcu_read_lock &rcu_state.gp_wq &p->pi_lock &rq->__lock irq_context: 0 sb_writers#4 &sbi->s_writepages_rwsem jbd2_handle rcu_read_lock &rcu_state.gp_wq &p->pi_lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 sb_writers#4 &sbi->s_writepages_rwsem jbd2_handle mmu_notifier_invalidate_range_start irq_context: 0 sb_writers#4 &sbi->s_writepages_rwsem jbd2_handle pool_lock#2 irq_context: 0 sb_writers#4 &sbi->s_writepages_rwsem jbd2_handle &ei->i_es_lock irq_context: 0 sb_writers#4 &sbi->s_writepages_rwsem jbd2_handle &ei->i_data_sem irq_context: 0 sb_writers#4 &sbi->s_writepages_rwsem jbd2_handle &ei->i_data_sem mmu_notifier_invalidate_range_start irq_context: 0 sb_writers#4 &sbi->s_writepages_rwsem jbd2_handle &ei->i_data_sem pool_lock#2 irq_context: 0 sb_writers#4 &sbi->s_writepages_rwsem jbd2_handle &ei->i_data_sem &ei->i_prealloc_lock irq_context: 0 sb_writers#4 &sbi->s_writepages_rwsem jbd2_handle &ei->i_data_sem &sbi->s_md_lock irq_context: 0 sb_writers#4 &sbi->s_writepages_rwsem jbd2_handle &ei->i_data_sem &bgl->locks[i].lock &sbi->s_md_lock irq_context: 0 sb_writers#4 &sbi->s_writepages_rwsem jbd2_handle &ei->i_data_sem &bgl->locks[i].lock &ei->i_prealloc_lock irq_context: 0 sb_writers#4 &sbi->s_writepages_rwsem jbd2_handle &ei->i_data_sem &mapping->private_lock irq_context: 0 sb_writers#4 &sbi->s_writepages_rwsem jbd2_handle &ei->i_data_sem key#3 irq_context: 0 sb_writers#4 &sbi->s_writepages_rwsem jbd2_handle &ei->i_data_sem &pa->pa_lock#2 irq_context: 0 sb_writers#4 &sbi->s_writepages_rwsem jbd2_handle &ei->i_data_sem &obj_hash[i].lock irq_context: 0 sb_writers#4 &sbi->s_writepages_rwsem jbd2_handle &ei->i_data_sem &ei->i_raw_lock irq_context: 0 sb_writers#4 &sbi->s_writepages_rwsem jbd2_handle &ei->i_data_sem &(ei->i_block_reservation_lock) irq_context: 0 sb_writers#4 &sbi->s_writepages_rwsem jbd2_handle &ei->i_data_sem &(ei->i_block_reservation_lock) key#14 irq_context: 0 sb_writers#4 &sbi->s_writepages_rwsem jbd2_handle &ei->i_data_sem &sb->s_type->i_lock_key#22 irq_context: 0 sb_writers#4 &sbi->s_writepages_rwsem jbd2_handle &ei->i_data_sem &ei->i_es_lock irq_context: 0 sb_writers#4 &sbi->s_writepages_rwsem jbd2_handle &ei->i_data_sem &ei->i_es_lock irq_context: 0 sb_writers#4 &sbi->s_writepages_rwsem jbd2_handle &ei->i_data_sem &ei->i_es_lock pool_lock#2 irq_context: 0 sb_writers#4 &sbi->s_writepages_rwsem jbd2_handle rcu_read_lock &memcg->move_lock irq_context: 0 sb_writers#4 &sbi->s_writepages_rwsem jbd2_handle rcu_read_lock &xa->xa_lock#9 irq_context: 0 sb_writers#4 &sbi->s_writepages_rwsem jbd2_handle rcu_read_lock &xa->xa_lock#9 &s->s_inode_wblist_lock irq_context: 0 sb_writers#4 &sbi->s_writepages_rwsem jbd2_handle rcu_read_lock key#10 irq_context: 0 sb_writers#4 &sbi->s_writepages_rwsem jbd2_handle rcu_read_lock &xa->xa_lock#9 key#10 irq_context: 0 sb_writers#4 &sbi->s_writepages_rwsem jbd2_handle tk_core.seq.seqcount irq_context: 0 sb_writers#4 &sbi->s_writepages_rwsem jbd2_handle &dd->lock irq_context: 0 sb_writers#4 &sbi->s_writepages_rwsem jbd2_handle rcu_read_lock &dd->lock irq_context: 0 sb_writers#4 &sbi->s_writepages_rwsem jbd2_handle rcu_read_lock &obj_hash[i].lock irq_context: 0 sb_writers#4 &sbi->s_writepages_rwsem jbd2_handle rcu_read_lock pool_lock#2 irq_context: 0 sb_writers#4 &sbi->s_writepages_rwsem jbd2_handle rcu_read_lock tk_core.seq.seqcount irq_context: 0 sb_writers#4 &sbi->s_writepages_rwsem jbd2_handle rcu_read_lock &virtscsi_vq->vq_lock irq_context: 0 sb_writers#4 &sbi->s_writepages_rwsem jbd2_handle &ei->i_raw_lock irq_context: 0 sb_writers#4 &sbi->s_writepages_rwsem jbd2_handle &journal->j_wait_updates irq_context: 0 sb_writers#4 &sbi->s_writepages_rwsem tk_core.seq.seqcount irq_context: 0 sb_writers#4 &sbi->s_writepages_rwsem &dd->lock irq_context: 0 sb_writers#4 &sbi->s_writepages_rwsem rcu_read_lock rcu_node_0 irq_context: 0 sb_writers#4 &sbi->s_writepages_rwsem rcu_read_lock &rq->__lock irq_context: 0 sb_writers#4 &sbi->s_writepages_rwsem rcu_read_lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 sb_writers#4 &sbi->s_writepages_rwsem rcu_read_lock &dd->lock irq_context: 0 sb_writers#4 &sbi->s_writepages_rwsem rcu_read_lock &obj_hash[i].lock irq_context: 0 sb_writers#4 &sbi->s_writepages_rwsem rcu_read_lock tk_core.seq.seqcount irq_context: 0 sb_writers#4 &sbi->s_writepages_rwsem rcu_read_lock &virtscsi_vq->vq_lock irq_context: 0 sb_writers#4 &sbi->s_writepages_rwsem rcu_read_lock pool_lock#2 irq_context: 0 sb_writers#4 &sb->s_type->i_mutex_key#8 &sb->s_type->i_mutex_key#8/4 &ei->i_data_sem irq_context: 0 sb_writers#4 &sb->s_type->i_mutex_key#8 &sb->s_type->i_mutex_key#8/4 &ei->i_data_sem &ei->i_data_sem/1 irq_context: 0 sb_writers#4 &sb->s_type->i_mutex_key#8 &sb->s_type->i_mutex_key#8/4 &ei->i_data_sem &ei->i_data_sem/1 mmu_notifier_invalidate_range_start irq_context: 0 sb_writers#4 &sb->s_type->i_mutex_key#8 &sb->s_type->i_mutex_key#8/4 &ei->i_data_sem &ei->i_data_sem/1 pool_lock#2 irq_context: 0 sb_writers#4 &sb->s_type->i_mutex_key#8 &sb->s_type->i_mutex_key#8/4 &ei->i_data_sem/1 irq_context: 0 sb_writers#4 &sb->s_type->i_mutex_key#8 &sb->s_type->i_mutex_key#8/4 mmu_notifier_invalidate_range_start irq_context: 0 sb_writers#4 &sb->s_type->i_mutex_key#8 &sb->s_type->i_mutex_key#8/4 mmu_notifier_invalidate_range_start &rq->__lock irq_context: 0 sb_writers#4 &sb->s_type->i_mutex_key#8 &sb->s_type->i_mutex_key#8/4 mmu_notifier_invalidate_range_start &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 sb_writers#4 &sb->s_type->i_mutex_key#8 &sb->s_type->i_mutex_key#8/4 pool_lock#2 irq_context: 0 sb_writers#4 &sb->s_type->i_mutex_key#8 &sb->s_type->i_mutex_key#8/4 &journal->j_state_lock irq_context: 0 sb_writers#4 &sb->s_type->i_mutex_key#8 &sb->s_type->i_mutex_key#8/4 jbd2_handle irq_context: 0 sb_writers#4 &sb->s_type->i_mutex_key#8 &sb->s_type->i_mutex_key#8/4 jbd2_handle &mapping->private_lock irq_context: 0 sb_writers#4 &sb->s_type->i_mutex_key#8 &sb->s_type->i_mutex_key#8/4 jbd2_handle stock_lock irq_context: 0 sb_writers#4 &sb->s_type->i_mutex_key#8 &sb->s_type->i_mutex_key#8/4 jbd2_handle &obj_hash[i].lock irq_context: 0 sb_writers#4 &sb->s_type->i_mutex_key#8 &sb->s_type->i_mutex_key#8/4 jbd2_handle pool_lock#2 irq_context: 0 sb_writers#4 &sb->s_type->i_mutex_key#8 &sb->s_type->i_mutex_key#8/4 jbd2_handle &ei->i_data_sem irq_context: 0 sb_writers#4 &sb->s_type->i_mutex_key#8 &sb->s_type->i_mutex_key#8/4 jbd2_handle &ei->i_data_sem &ei->i_data_sem/1 irq_context: 0 sb_writers#4 &sb->s_type->i_mutex_key#8 &sb->s_type->i_mutex_key#8/4 jbd2_handle &ei->i_data_sem &ei->i_data_sem/1 &ei->i_es_lock irq_context: 0 sb_writers#4 &sb->s_type->i_mutex_key#8 &sb->s_type->i_mutex_key#8/4 jbd2_handle &ei->i_data_sem &ei->i_data_sem/1 &ei->i_es_lock &obj_hash[i].lock irq_context: 0 sb_writers#4 &sb->s_type->i_mutex_key#8 &sb->s_type->i_mutex_key#8/4 jbd2_handle &ei->i_data_sem &ei->i_data_sem/1 &ei->i_es_lock pool_lock#2 irq_context: 0 sb_writers#4 &sb->s_type->i_mutex_key#8 &sb->s_type->i_mutex_key#8/4 jbd2_handle &ei->i_data_sem &ei->i_data_sem/1 mmu_notifier_invalidate_range_start irq_context: 0 sb_writers#4 &sb->s_type->i_mutex_key#8 &sb->s_type->i_mutex_key#8/4 jbd2_handle &ei->i_data_sem &ei->i_data_sem/1 mmu_notifier_invalidate_range_start &rq->__lock irq_context: 0 sb_writers#4 &sb->s_type->i_mutex_key#8 &sb->s_type->i_mutex_key#8/4 jbd2_handle &ei->i_data_sem &ei->i_data_sem/1 pool_lock#2 irq_context: 0 sb_writers#4 &sb->s_type->i_mutex_key#8 &sb->s_type->i_mutex_key#8/4 jbd2_handle &ei->i_data_sem &ei->i_data_sem/1 &rq->__lock irq_context: 0 sb_writers#4 &sb->s_type->i_mutex_key#8 &sb->s_type->i_mutex_key#8/4 jbd2_handle &ei->i_data_sem &ei->i_data_sem/1 &c->lock irq_context: 0 sb_writers#4 &sb->s_type->i_mutex_key#8 &sb->s_type->i_mutex_key#8/4 jbd2_handle &ei->i_data_sem &ei->i_data_sem/1 &ei->i_raw_lock irq_context: 0 sb_writers#4 &sb->s_type->i_mutex_key#8 &sb->s_type->i_mutex_key#8/4 jbd2_handle &ei->i_data_sem &ei->i_data_sem/1 &obj_hash[i].lock irq_context: 0 sb_writers#4 &sb->s_type->i_mutex_key#8 &sb->s_type->i_mutex_key#8/4 jbd2_handle &ei->i_data_sem/1 irq_context: 0 sb_writers#4 &sb->s_type->i_mutex_key#8 &sb->s_type->i_mutex_key#8/4 jbd2_handle &ei->i_data_sem/1 &rq->__lock irq_context: 0 sb_writers#4 &sb->s_type->i_mutex_key#8 &sb->s_type->i_mutex_key#8/4 jbd2_handle &ei->i_data_sem/1 &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 sb_writers#4 &sb->s_type->i_mutex_key#8 &sb->s_type->i_mutex_key#8/4 jbd2_handle mmu_notifier_invalidate_range_start irq_context: 0 sb_writers#4 &sb->s_type->i_mutex_key#8 &sb->s_type->i_mutex_key#8/4 jbd2_handle mmu_notifier_invalidate_range_start &rq->__lock irq_context: 0 sb_writers#4 &sb->s_type->i_mutex_key#8 &sb->s_type->i_mutex_key#8/4 jbd2_handle mmu_notifier_invalidate_range_start &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 sb_writers#4 &sb->s_type->i_mutex_key#8 &sb->s_type->i_mutex_key#8/4 jbd2_handle &c->lock irq_context: 0 sb_writers#4 &sb->s_type->i_mutex_key#8 &sb->s_type->i_mutex_key#8/4 jbd2_handle &ei->i_es_lock irq_context: 0 sb_writers#4 &sb->s_type->i_mutex_key#8 &sb->s_type->i_mutex_key#8/4 jbd2_handle &ei->i_data_sem irq_context: 0 sb_writers#4 &sb->s_type->i_mutex_key#8 &sb->s_type->i_mutex_key#8/4 jbd2_handle &ei->i_data_sem mmu_notifier_invalidate_range_start irq_context: 0 sb_writers#4 &sb->s_type->i_mutex_key#8 &sb->s_type->i_mutex_key#8/4 jbd2_handle &ei->i_data_sem pool_lock#2 irq_context: 0 sb_writers#4 &sb->s_type->i_mutex_key#8 &sb->s_type->i_mutex_key#8/4 jbd2_handle &ei->i_data_sem &ei->i_es_lock irq_context: 0 sb_writers#4 &sb->s_type->i_mutex_key#8 &sb->s_type->i_mutex_key#8/4 jbd2_handle &ei->i_data_sem &ei->i_es_lock &____s->seqcount irq_context: 0 sb_writers#4 &sb->s_type->i_mutex_key#8 &sb->s_type->i_mutex_key#8/4 jbd2_handle &ei->i_data_sem &ei->i_es_lock pool_lock#2 irq_context: 0 sb_writers#4 &sb->s_type->i_mutex_key#8 &sb->s_type->i_mutex_key#8/4 jbd2_handle &ei->i_data_sem &ei->i_es_lock rcu_read_lock pool_lock#2 irq_context: 0 sb_writers#4 &sb->s_type->i_mutex_key#8 &sb->s_type->i_mutex_key#8/4 jbd2_handle &ei->i_data_sem &ei->i_es_lock &obj_hash[i].lock irq_context: 0 sb_writers#4 &sb->s_type->i_mutex_key#8 &sb->s_type->i_mutex_key#8/4 jbd2_handle &ei->i_data_sem &obj_hash[i].lock irq_context: 0 sb_writers#4 &sb->s_type->i_mutex_key#8 &sb->s_type->i_mutex_key#8/4 jbd2_handle rcu_read_lock &memcg->move_lock irq_context: 0 sb_writers#4 &sb->s_type->i_mutex_key#8 &sb->s_type->i_mutex_key#8/4 jbd2_handle rcu_read_lock &xa->xa_lock#9 irq_context: 0 sb_writers#4 &sb->s_type->i_mutex_key#8 &sb->s_type->i_mutex_key#8/4 jbd2_handle &journal->j_list_lock irq_context: 0 sb_writers#4 &sb->s_type->i_mutex_key#8 &sb->s_type->i_mutex_key#8/4 jbd2_handle &journal->j_wait_updates irq_context: 0 sb_writers#4 &sb->s_type->i_mutex_key#8 &sb->s_type->i_mutex_key#8/4 &obj_hash[i].lock irq_context: 0 sb_writers#4 &sb->s_type->i_mutex_key#8 &sb->s_type->i_mutex_key#8/4 &ei->i_data_sem &ei->i_data_sem/1 &ei->i_prealloc_lock irq_context: 0 sb_writers#4 &sb->s_type->i_mutex_key#8 &sb->s_type->i_mutex_key#8/4 &ei->i_data_sem &ei->i_data_sem/1 &ei->i_prealloc_lock &pa->pa_lock#2 irq_context: 0 sb_writers#4 &sb->s_type->i_mutex_key#8 &sb->s_type->i_mutex_key#8/4 &ei->i_data_sem &ei->i_data_sem/1 &obj_hash[i].lock irq_context: 0 sb_writers#4 &sb->s_type->i_mutex_key#8 &sb->s_type->i_mutex_key#8/4 &ei->i_data_sem &ei->i_data_sem/1 &rq->__lock irq_context: 0 sb_writers#4 &sb->s_type->i_mutex_key#8 &sb->s_type->i_mutex_key#8/4 &ei->i_data_sem &ei->i_data_sem/1 &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 sb_writers#4 &sb->s_type->i_mutex_key#8/4 irq_context: 0 &mm->mmap_lock sb_pagefaults mapping.invalidate_lock mmu_notifier_invalidate_range_start irq_context: 0 &mm->mmap_lock sb_pagefaults mapping.invalidate_lock stock_lock irq_context: 0 &mm->mmap_lock sb_pagefaults mapping.invalidate_lock &c->lock irq_context: 0 &mm->mmap_lock sb_pagefaults mapping.invalidate_lock pool_lock#2 irq_context: 0 &mm->mmap_lock sb_pagefaults mapping.invalidate_lock &ei->i_es_lock irq_context: 0 &mm->mmap_lock sb_pagefaults mapping.invalidate_lock &ei->i_data_sem irq_context: 0 &mm->mmap_lock sb_pagefaults mapping.invalidate_lock &ei->i_data_sem mmu_notifier_invalidate_range_start irq_context: 0 &mm->mmap_lock sb_pagefaults mapping.invalidate_lock &ei->i_data_sem &c->lock irq_context: 0 &mm->mmap_lock sb_pagefaults mapping.invalidate_lock &ei->i_data_sem pool_lock#2 irq_context: 0 &mm->mmap_lock sb_pagefaults mapping.invalidate_lock &ei->i_data_sem &ei->i_es_lock irq_context: 0 &mm->mmap_lock sb_pagefaults mapping.invalidate_lock &ei->i_data_sem &ei->i_es_lock pool_lock#2 irq_context: 0 &mm->mmap_lock sb_pagefaults mapping.invalidate_lock &ei->i_data_sem &obj_hash[i].lock irq_context: 0 &mm->mmap_lock sb_pagefaults mapping.invalidate_lock &ei->i_data_sem &ei->i_es_lock &obj_hash[i].lock irq_context: 0 &mm->mmap_lock remove_cache_srcu &cfs_rq->removed.lock irq_context: 0 &pipe->mutex/1 sk_lock-AF_INET6 &cfs_rq->removed.lock irq_context: 0 (wq_completion)events (work_completion)(&rdev->wiphy_work) &rdev->wiphy.mtx &wdev->mtx batched_entropy_u8.lock crngs.lock irq_context: 0 tomoyo_ss rcu_read_lock rcu_read_lock pgd_lock irq_context: 0 tomoyo_ss rcu_read_lock rcu_read_lock stock_lock irq_context: 0 tomoyo_ss rcu_read_lock rcu_read_lock key irq_context: 0 tomoyo_ss rcu_read_lock rcu_read_lock pcpu_lock irq_context: 0 tomoyo_ss rcu_read_lock rcu_read_lock percpu_counters_lock irq_context: 0 tomoyo_ss rcu_read_lock rcu_read_lock pcpu_lock stock_lock irq_context: 0 sb_writers#4 &sb->s_type->i_mutex_key#8 &sb->s_type->i_mutex_key#8/4 &journal->j_state_lock irq_context: 0 sb_writers#4 &sb->s_type->i_mutex_key#8 &sb->s_type->i_mutex_key#8/4 &journal->j_state_lock &journal->j_wait_commit irq_context: 0 sb_writers#4 &sb->s_type->i_mutex_key#8 &sb->s_type->i_mutex_key#8/4 &journal->j_state_lock &journal->j_wait_commit &p->pi_lock irq_context: 0 sb_writers#4 &sb->s_type->i_mutex_key#8 &sb->s_type->i_mutex_key#8/4 &journal->j_state_lock &journal->j_wait_commit &p->pi_lock &rq->__lock irq_context: 0 sb_writers#4 &sb->s_type->i_mutex_key#8 &sb->s_type->i_mutex_key#8/4 &journal->j_state_lock &journal->j_wait_commit &p->pi_lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 sb_writers#4 &sb->s_type->i_mutex_key#8 &sb->s_type->i_mutex_key#8/4 &cfs_rq->removed.lock irq_context: 0 sb_writers#4 &sb->s_type->i_mutex_key#8 &sb->s_type->i_mutex_key#8/4 &journal->j_wait_commit irq_context: 0 sb_writers#4 &sb->s_type->i_mutex_key#8 &sb->s_type->i_mutex_key#8/4 &journal->j_wait_done_commit irq_context: 0 sb_writers#4 &sb->s_type->i_mutex_key#8 &sb->s_type->i_mutex_key#8/4 &rq->__lock &cfs_rq->removed.lock irq_context: 0 sb_writers#4 &sb->s_type->i_mutex_key#8 &sb->s_type->i_mutex_key#8/4 jbd2_handle irq_context: 0 &sb->s_type->i_mutex_key#9 rcu_read_lock &cfs_rq->removed.lock irq_context: 0 &sb->s_type->i_mutex_key#9 rcu_read_lock &obj_hash[i].lock irq_context: 0 &mm->mmap_lock &mapping->i_mmap_rwsem stock_lock irq_context: 0 &mm->mmap_lock &mapping->i_mmap_rwsem pcpu_lock stock_lock irq_context: 0 sb_writers#4 &type->i_mutex_dir_key#3/1 jbd2_handle &ei->i_data_sem bit_wait_table + i irq_context: 0 &sbi->s_writepages_rwsem jbd2_handle remove_cache_srcu irq_context: 0 &sbi->s_writepages_rwsem jbd2_handle remove_cache_srcu quarantine_lock irq_context: 0 &sbi->s_writepages_rwsem jbd2_handle remove_cache_srcu &c->lock irq_context: 0 &sbi->s_writepages_rwsem jbd2_handle remove_cache_srcu &n->list_lock irq_context: 0 &sbi->s_writepages_rwsem jbd2_handle remove_cache_srcu &obj_hash[i].lock irq_context: 0 &sbi->s_writepages_rwsem jbd2_handle remove_cache_srcu pool_lock#2 irq_context: 0 &sbi->s_writepages_rwsem jbd2_handle remove_cache_srcu rcu_read_lock pool_lock#2 irq_context: 0 sb_writers#4 &type->i_mutex_dir_key#3/1 &sb->s_type->i_mutex_key#8 rcu_read_lock &cfs_rq->removed.lock irq_context: 0 sb_writers#4 &type->i_mutex_dir_key#3/1 &sb->s_type->i_mutex_key#8 rcu_read_lock &obj_hash[i].lock irq_context: softirq (&icsk->icsk_retransmit_timer) slock-AF_INET6 rcu_read_lock rcu_read_lock &nf_conntrack_locks[i] irq_context: softirq (&icsk->icsk_retransmit_timer) slock-AF_INET6 rcu_read_lock rcu_read_lock rcu_read_lock &nf_nat_locks[i] irq_context: softirq (&icsk->icsk_retransmit_timer) slock-AF_INET6 rcu_read_lock rcu_read_lock &obj_hash[i].lock irq_context: softirq (&icsk->icsk_retransmit_timer) slock-AF_INET6 rcu_read_lock rcu_read_lock pool_lock#2 irq_context: softirq (&icsk->icsk_retransmit_timer) slock-AF_INET6 rcu_read_lock rcu_read_lock &c->lock irq_context: softirq (&icsk->icsk_retransmit_timer) slock-AF_INET6 rcu_read_lock rcu_read_lock &n->list_lock irq_context: softirq (&icsk->icsk_retransmit_timer) slock-AF_INET6 rcu_read_lock rcu_read_lock &n->list_lock &c->lock irq_context: softirq (&icsk->icsk_retransmit_timer) slock-AF_INET6 rcu_read_lock rcu_read_lock rcu_read_lock &____s->seqcount#7 irq_context: softirq (&icsk->icsk_retransmit_timer) slock-AF_INET6 rcu_read_lock rcu_read_lock &nf_nat_locks[i] irq_context: softirq (&icsk->icsk_retransmit_timer) slock-AF_INET6 rcu_read_lock rcu_read_lock &nf_conntrack_locks[i] batched_entropy_u8.lock irq_context: 0 &f->f_pos_lock sb_writers#4 &sb->s_type->i_mutex_key#8 &mm->mmap_lock rcu_read_lock &rcu_state.gp_wq &p->pi_lock &rq->__lock &cfs_rq->removed.lock irq_context: 0 sb_writers#4 &type->i_mutex_dir_key#3/1 &sb->s_type->i_mutex_key#8 &pcp->lock &zone->lock irq_context: 0 &sbi->s_writepages_rwsem jbd2_handle rcu_read_lock &n->list_lock irq_context: 0 &sbi->s_writepages_rwsem jbd2_handle rcu_read_lock &n->list_lock &c->lock irq_context: 0 &sbi->s_writepages_rwsem jbd2_handle rcu_read_lock &base->lock irq_context: 0 &sbi->s_writepages_rwsem jbd2_handle rcu_read_lock &base->lock &obj_hash[i].lock irq_context: 0 sb_writers#4 inode_hash_lock &s->s_inode_list_lock irq_context: 0 sb_writers#4 (console_sem).lock irq_context: 0 sb_writers#4 console_lock console_srcu console_owner_lock irq_context: 0 sb_writers#4 console_lock console_srcu console_owner irq_context: 0 sb_writers#4 console_lock console_srcu console_owner &port_lock_key irq_context: 0 sb_writers#4 console_lock console_srcu console_owner console_owner_lock irq_context: 0 sb_writers#4 &sb->s_type->i_lock_key#22 &lru->node[i].lock irq_context: 0 sb_writers#4 &journal->j_state_lock &journal->j_wait_updates irq_context: 0 sb_writers#4 &journal->j_barrier irq_context: 0 sb_writers#4 &journal->j_barrier &journal->j_state_lock irq_context: 0 sb_writers#4 &journal->j_barrier &journal->j_state_lock &journal->j_wait_commit irq_context: 0 sb_writers#4 &journal->j_barrier &journal->j_state_lock &journal->j_wait_commit &p->pi_lock irq_context: 0 sb_writers#4 &journal->j_barrier &journal->j_state_lock &journal->j_wait_commit &p->pi_lock &rq->__lock irq_context: 0 sb_writers#4 &journal->j_barrier &journal->j_state_lock &journal->j_wait_commit &p->pi_lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 sb_writers#4 &journal->j_barrier &rq->__lock irq_context: 0 sb_writers#4 &journal->j_barrier &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 sb_writers#4 &journal->j_barrier &journal->j_state_lock irq_context: 0 sb_writers#4 &journal->j_barrier &journal->j_wait_commit irq_context: 0 sb_writers#4 &journal->j_barrier &journal->j_wait_done_commit irq_context: 0 sb_writers#4 &journal->j_barrier &journal->j_list_lock irq_context: 0 sb_writers#4 &journal->j_barrier &journal->j_checkpoint_mutex irq_context: 0 sb_writers#4 &journal->j_barrier &journal->j_checkpoint_mutex &journal->j_state_lock irq_context: 0 sb_writers#4 &journal->j_barrier &journal->j_checkpoint_mutex &journal->j_state_lock &journal->j_list_lock irq_context: 0 sb_writers#4 &journal->j_barrier &journal->j_checkpoint_mutex tk_core.seq.seqcount irq_context: 0 sb_writers#4 &journal->j_barrier &journal->j_checkpoint_mutex &fq->mq_flush_lock irq_context: 0 sb_writers#4 &journal->j_barrier &journal->j_checkpoint_mutex &fq->mq_flush_lock tk_core.seq.seqcount irq_context: 0 sb_writers#4 &journal->j_barrier &journal->j_checkpoint_mutex &fq->mq_flush_lock &q->requeue_lock irq_context: 0 sb_writers#4 &journal->j_barrier &journal->j_checkpoint_mutex &fq->mq_flush_lock &obj_hash[i].lock irq_context: 0 sb_writers#4 &journal->j_barrier &journal->j_checkpoint_mutex &fq->mq_flush_lock rcu_read_lock &pool->lock irq_context: 0 sb_writers#4 &journal->j_barrier &journal->j_checkpoint_mutex &fq->mq_flush_lock rcu_read_lock &pool->lock &obj_hash[i].lock irq_context: 0 sb_writers#4 &journal->j_barrier &journal->j_checkpoint_mutex &fq->mq_flush_lock rcu_read_lock &pool->lock pool_lock#2 irq_context: 0 sb_writers#4 &journal->j_barrier &journal->j_checkpoint_mutex &fq->mq_flush_lock rcu_read_lock &pool->lock &p->pi_lock irq_context: 0 sb_writers#4 &journal->j_barrier &journal->j_checkpoint_mutex &fq->mq_flush_lock rcu_read_lock &pool->lock &p->pi_lock &rq->__lock irq_context: 0 sb_writers#4 &journal->j_barrier &journal->j_checkpoint_mutex &fq->mq_flush_lock rcu_read_lock &pool->lock &p->pi_lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 sb_writers#4 &journal->j_barrier &journal->j_checkpoint_mutex &rq->__lock irq_context: 0 sb_writers#4 &journal->j_barrier &journal->j_checkpoint_mutex &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 sb_writers#4 &journal->j_barrier &journal->j_checkpoint_mutex &x->wait#26 irq_context: 0 sb_writers#4 &journal->j_barrier &journal->j_checkpoint_mutex mmu_notifier_invalidate_range_start irq_context: 0 sb_writers#4 &journal->j_barrier &journal->j_checkpoint_mutex pool_lock#2 irq_context: 0 sb_writers#4 &journal->j_barrier &journal->j_checkpoint_mutex &dd->lock irq_context: 0 sb_writers#4 &journal->j_barrier &journal->j_checkpoint_mutex &obj_hash[i].lock irq_context: 0 sb_writers#4 &journal->j_barrier &journal->j_checkpoint_mutex rcu_read_lock &pool->lock irq_context: 0 sb_writers#4 &journal->j_barrier &journal->j_checkpoint_mutex rcu_read_lock &pool->lock &obj_hash[i].lock irq_context: 0 sb_writers#4 &journal->j_barrier &journal->j_checkpoint_mutex rcu_read_lock &pool->lock pool_lock#2 irq_context: 0 sb_writers#4 &journal->j_barrier &journal->j_checkpoint_mutex rcu_read_lock &pool->lock &p->pi_lock irq_context: 0 sb_writers#4 &journal->j_barrier &journal->j_checkpoint_mutex rcu_read_lock &pool->lock &p->pi_lock &rq->__lock irq_context: 0 sb_writers#4 &journal->j_barrier &journal->j_checkpoint_mutex rcu_read_lock &pool->lock &p->pi_lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 sb_writers#4 &journal->j_barrier &journal->j_checkpoint_mutex bit_wait_table + i irq_context: 0 sb_writers#4 &journal->j_barrier &journal->j_checkpoint_mutex &journal->j_state_lock irq_context: 0 sb_writers#4 &journal->j_barrier &journal->j_checkpoint_mutex &journal->j_list_lock irq_context: 0 sb_writers#4 &journal->j_barrier &journal->j_checkpoint_mutex &c->lock irq_context: 0 sb_writers#4 &journal->j_barrier &journal->j_checkpoint_mutex &____s->seqcount#2 irq_context: 0 sb_writers#4 &journal->j_barrier &journal->j_checkpoint_mutex &____s->seqcount irq_context: 0 sb_writers#4 &journal->j_barrier &journal->j_checkpoint_mutex rcu_read_lock &dd->lock irq_context: 0 sb_writers#4 &journal->j_barrier &journal->j_checkpoint_mutex rcu_read_lock &obj_hash[i].lock irq_context: 0 sb_writers#4 &journal->j_barrier &journal->j_checkpoint_mutex rcu_read_lock tk_core.seq.seqcount irq_context: 0 sb_writers#4 &journal->j_barrier &journal->j_checkpoint_mutex rcu_read_lock &virtscsi_vq->vq_lock irq_context: 0 sb_writers#4 &journal->j_barrier &journal->j_checkpoint_mutex rcu_read_lock pool_lock#2 irq_context: 0 sb_writers#4 &journal->j_barrier &journal->j_checkpoint_mutex &journal->j_list_lock &obj_hash[i].lock irq_context: 0 sb_writers#4 &journal->j_barrier &journal->j_checkpoint_mutex &journal->j_list_lock pool_lock#2 irq_context: 0 sb_writers#4 &journal->j_wait_transaction_locked &p->pi_lock irq_context: 0 sb_writers#4 &journal->j_wait_transaction_locked &p->pi_lock &rq->__lock irq_context: 0 sb_writers#4 &journal->j_wait_transaction_locked &p->pi_lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 sb_writers#3 &sb->s_type->i_mutex_key#9 &pcp->lock &zone->lock &____s->seqcount irq_context: softirq (&peer->timer_retransmit_handshake) rcu_read_lock_bh rcu_read_lock &pool->lock/1 &p->pi_lock &cfs_rq->removed.lock irq_context: 0 sb_writers#4 &journal->j_barrier jbd2_handle irq_context: 0 &journal->j_checkpoint_mutex &c->lock irq_context: 0 sb_writers#4 &journal->j_barrier &journal->j_checkpoint_mutex &journal->j_list_lock key#15 irq_context: 0 sb_writers#4 &journal->j_barrier &journal->j_checkpoint_mutex &base->lock irq_context: 0 sb_writers#4 &journal->j_barrier &journal->j_checkpoint_mutex &base->lock &obj_hash[i].lock irq_context: 0 sb_writers#4 &journal->j_barrier &journal->j_checkpoint_mutex (&timer.timer) irq_context: 0 &xt[i].mutex &obj_hash[i].lock pool_lock irq_context: 0 sb_writers#4 &journal->j_barrier &journal->j_checkpoint_mutex rcu_read_lock rcu_node_0 irq_context: 0 sb_writers#4 &journal->j_barrier &journal->j_checkpoint_mutex rcu_read_lock &rq->__lock irq_context: 0 sb_writers#4 &journal->j_barrier &journal->j_checkpoint_mutex rcu_read_lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 sb_writers#4 &journal->j_barrier &journal->j_checkpoint_mutex &n->list_lock irq_context: 0 sb_writers#4 &journal->j_barrier &journal->j_checkpoint_mutex &n->list_lock &c->lock irq_context: 0 sb_writers#4 &journal->j_barrier &journal->j_checkpoint_mutex &journal->j_list_lock &c->lock irq_context: 0 sb_writers#4 &journal->j_barrier &journal->j_checkpoint_mutex batched_entropy_u8.lock irq_context: 0 sb_writers#4 &journal->j_barrier &journal->j_checkpoint_mutex kfence_freelist_lock irq_context: 0 sb_writers#4 &type->i_mutex_dir_key#3/1 tomoyo_ss remove_cache_srcu rcu_node_0 irq_context: 0 sb_writers#4 &type->i_mutex_dir_key#3/1 &sb->s_type->i_mutex_key#8 jbd2_handle &c->lock irq_context: 0 &pipe->mutex/1 sk_lock-AF_INET6 rcu_read_lock rcu_read_lock &obj_hash[i].lock pool_lock irq_context: 0 dup_mmap_sem &mm->mmap_lock &mm->mmap_lock/1 remove_cache_srcu rcu_read_lock &rcu_state.gp_wq irq_context: 0 dup_mmap_sem &mm->mmap_lock &mm->mmap_lock/1 remove_cache_srcu rcu_read_lock &rcu_state.gp_wq &p->pi_lock irq_context: 0 dup_mmap_sem &mm->mmap_lock &mm->mmap_lock/1 remove_cache_srcu rcu_read_lock &rcu_state.gp_wq &p->pi_lock &rq->__lock irq_context: 0 dup_mmap_sem &mm->mmap_lock &mm->mmap_lock/1 remove_cache_srcu rcu_read_lock &rcu_state.gp_wq &p->pi_lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 sb_writers#8 remove_cache_srcu rcu_read_lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)wg-crypt-wg2 (work_completion)(&({ do { const void *__vpp_verify = (typeof((worker) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __asm__ ("" : "=r"(__ptr) : "0"((typeof(*((worker))) *)((worker)))); (typeof((typeof(*((worker))) *)((worker)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); })->work) &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 sb_writers#8 kn->active#5 remove_cache_srcu rcu_read_lock rcu_node_0 irq_context: 0 sb_writers#8 kn->active#5 remove_cache_srcu rcu_read_lock &rq->__lock irq_context: 0 sb_writers#8 kn->active#5 remove_cache_srcu rcu_read_lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 sb_writers#5 &type->i_mutex_dir_key#5/1 pgd_lock irq_context: 0 sb_writers#5 &type->i_mutex_dir_key#5/1 stock_lock irq_context: 0 sb_writers#5 &type->i_mutex_dir_key#5/1 key irq_context: 0 sb_writers#5 &type->i_mutex_dir_key#5/1 pcpu_lock irq_context: 0 sb_writers#5 &type->i_mutex_dir_key#5/1 percpu_counters_lock irq_context: 0 sb_writers#5 &type->i_mutex_dir_key#5/1 pcpu_lock stock_lock irq_context: 0 (wq_completion)wg-crypt-wg2#2 (work_completion)(&peer->transmit_packet_work) rcu_read_lock_bh rcu_read_lock &pool->lock/1 &p->pi_lock &rq->__lock irq_context: 0 (wq_completion)wg-crypt-wg2#2 (work_completion)(&peer->transmit_packet_work) rcu_read_lock_bh rcu_read_lock &pool->lock/1 &p->pi_lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)wg-kex-wg1#4 (work_completion)(&({ do { const void *__vpp_verify = (typeof((worker) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __asm__ ("" : "=r"(__ptr) : "0"((typeof(*((worker))) *)((worker)))); (typeof((typeof(*((worker))) *)((worker)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); })->work) &c->lock irq_context: 0 (wq_completion)wg-kex-wg1#4 (work_completion)(&({ do { const void *__vpp_verify = (typeof((worker) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __asm__ ("" : "=r"(__ptr) : "0"((typeof(*((worker))) *)((worker)))); (typeof((typeof(*((worker))) *)((worker)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); })->work) &n->list_lock irq_context: 0 (wq_completion)wg-kex-wg1#4 (work_completion)(&({ do { const void *__vpp_verify = (typeof((worker) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __asm__ ("" : "=r"(__ptr) : "0"((typeof(*((worker))) *)((worker)))); (typeof((typeof(*((worker))) *)((worker)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); })->work) &n->list_lock &c->lock irq_context: 0 (wq_completion)wg-kex-wg2#4 (work_completion)(&({ do { const void *__vpp_verify = (typeof((worker) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __asm__ ("" : "=r"(__ptr) : "0"((typeof(*((worker))) *)((worker)))); (typeof((typeof(*((worker))) *)((worker)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); })->work) &n->list_lock irq_context: 0 (wq_completion)wg-kex-wg2#4 (work_completion)(&({ do { const void *__vpp_verify = (typeof((worker) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __asm__ ("" : "=r"(__ptr) : "0"((typeof(*((worker))) *)((worker)))); (typeof((typeof(*((worker))) *)((worker)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); })->work) &n->list_lock &c->lock irq_context: 0 (wq_completion)wg-crypt-wg2#2 (work_completion)(&peer->transmit_packet_work) &peer->endpoint_lock rcu_read_lock_bh batched_entropy_u32.lock crngs.lock irq_context: 0 loop_validate_mutex &lo->lo_mutex &cfs_rq->removed.lock irq_context: 0 sb_writers#4 tomoyo_ss remove_cache_srcu pool_lock#2 irq_context: 0 sb_writers#4 tomoyo_ss rcu_read_lock rename_lock irq_context: 0 sb_writers#4 tomoyo_ss rcu_read_lock &rcu_state.gp_wq irq_context: 0 sb_writers#4 tomoyo_ss rcu_read_lock &rcu_state.gp_wq &p->pi_lock irq_context: 0 sb_writers#4 tomoyo_ss rcu_read_lock &rcu_state.gp_wq &p->pi_lock &rq->__lock irq_context: 0 sb_writers#4 tomoyo_ss rcu_read_lock &rcu_state.gp_wq &p->pi_lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 sb_writers#4 tomoyo_ss batched_entropy_u8.lock irq_context: 0 sb_writers#4 tomoyo_ss kfence_freelist_lock irq_context: 0 sb_writers#4 tomoyo_ss &meta->lock irq_context: 0 sb_writers#4 &sb->s_type->i_mutex_key#8 mapping.invalidate_lock jbd2_handle &ei->i_data_sem &____s->seqcount#2 irq_context: 0 sb_writers#4 &type->i_mutex_dir_key#3/1 &type->i_mutex_dir_key#3 remove_cache_srcu &rq->__lock irq_context: 0 sb_writers#4 &type->i_mutex_dir_key#3/1 &type->i_mutex_dir_key#3 remove_cache_srcu &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 &f->f_pos_lock rcu_read_lock &cfs_rq->removed.lock irq_context: 0 &f->f_pos_lock rcu_read_lock &obj_hash[i].lock irq_context: 0 &f->f_pos_lock rcu_read_lock &obj_hash[i].lock pool_lock irq_context: 0 &sbi->s_writepages_rwsem jbd2_handle &ei->i_data_sem &sem->wait_lock irq_context: 0 &sbi->s_writepages_rwsem jbd2_handle &ei->i_data_sem &dd->lock irq_context: 0 &sbi->s_writepages_rwsem jbd2_handle &ei->i_data_sem rcu_read_lock &pool->lock irq_context: 0 &sbi->s_writepages_rwsem jbd2_handle &ei->i_data_sem rcu_read_lock &pool->lock &obj_hash[i].lock irq_context: 0 &sbi->s_writepages_rwsem jbd2_handle &ei->i_data_sem rcu_read_lock &pool->lock pool_lock#2 irq_context: 0 &sbi->s_writepages_rwsem jbd2_handle &ei->i_data_sem rcu_read_lock &pool->lock &p->pi_lock irq_context: 0 &sbi->s_writepages_rwsem jbd2_handle &ei->i_data_sem rcu_read_lock &pool->lock &p->pi_lock &rq->__lock irq_context: 0 &sbi->s_writepages_rwsem jbd2_handle &ei->i_data_sem rcu_read_lock &pool->lock &p->pi_lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 lock#3 &cfs_rq->removed.lock irq_context: 0 lock#3 pool_lock#2 irq_context: 0 sb_writers#4 rcu_read_lock pgd_lock irq_context: 0 sb_writers#4 rcu_read_lock stock_lock irq_context: 0 sb_writers#4 rcu_read_lock key irq_context: 0 sb_writers#4 rcu_read_lock pcpu_lock irq_context: 0 sb_writers#4 rcu_read_lock percpu_counters_lock irq_context: 0 sb_writers#4 rcu_read_lock pcpu_lock stock_lock irq_context: 0 sb_writers#5 &type->i_mutex_dir_key#5 rcu_read_lock &rcu_state.gp_wq irq_context: 0 sb_writers#5 &type->i_mutex_dir_key#5 rcu_read_lock &rcu_state.gp_wq &p->pi_lock irq_context: 0 sb_writers#5 &type->i_mutex_dir_key#5 rcu_read_lock &rcu_state.gp_wq &p->pi_lock &rq->__lock irq_context: 0 sb_writers#5 &type->i_mutex_dir_key#5 rcu_read_lock &rcu_state.gp_wq &p->pi_lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 rcu_read_lock &trie->lock irq_context: 0 rcu_read_lock &trie->lock stock_lock irq_context: 0 rcu_read_lock &trie->lock &c->lock irq_context: 0 rcu_read_lock &trie->lock pool_lock#2 irq_context: 0 rcu_read_lock &trie->lock &obj_hash[i].lock irq_context: 0 rcu_read_lock &trie->lock krc.lock irq_context: 0 rcu_read_lock &trie->lock &obj_hash[i].lock pool_lock irq_context: 0 rcu_read_lock &trie->lock &____s->seqcount#2 irq_context: 0 rcu_read_lock &trie->lock &____s->seqcount irq_context: 0 rcu_read_lock &trie->lock krc.lock &obj_hash[i].lock irq_context: 0 rcu_read_lock &trie->lock krc.lock &base->lock irq_context: 0 rcu_read_lock &trie->lock krc.lock &base->lock &obj_hash[i].lock irq_context: 0 rcu_read_lock &trie->lock &n->list_lock irq_context: 0 rcu_read_lock &trie->lock &n->list_lock &c->lock irq_context: 0 (wq_completion)ext4-rsv-conversion (work_completion)(&ei->i_rsv_conversion_work) &base->lock irq_context: 0 (wq_completion)ext4-rsv-conversion (work_completion)(&ei->i_rsv_conversion_work) &base->lock &obj_hash[i].lock irq_context: 0 sb_writers#4 &type->i_mutex_dir_key#3/1 jbd2_handle &ei->i_data_sem rcu_read_lock pool_lock#2 irq_context: 0 (wq_completion)events (work_completion)(&(&krcp->krw_arr[i].rcu_work)->work) rcu_callback rcu_read_lock rcu_node_0 irq_context: 0 (wq_completion)events (work_completion)(&(&krcp->krw_arr[i].rcu_work)->work) rcu_callback rcu_read_lock &rq->__lock irq_context: 0 (wq_completion)events (work_completion)(&(&krcp->krw_arr[i].rcu_work)->work) rcu_callback rcu_read_lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 sb_writers#4 &mapping->private_lock rcu_read_lock rcu_read_lock rcu_read_lock per_cpu_ptr(&cgroup_rstat_cpu_lock, cpu) irq_context: softirq (&pool->idle_timer) irq_context: softirq (&pool->idle_timer) &pool->lock/1 irq_context: softirq (&pool->idle_timer) &pool->lock/1 &obj_hash[i].lock irq_context: softirq (&pool->idle_timer) &pool->lock/1 &base->lock irq_context: softirq (&pool->idle_timer) &pool->lock/1 &base->lock &obj_hash[i].lock irq_context: 0 &sbi->s_writepages_rwsem jbd2_handle &rq->__lock &cfs_rq->removed.lock irq_context: 0 remove_cache_srcu rcu_read_lock &cfs_rq->removed.lock irq_context: 0 sb_writers#4 &sb->s_type->i_mutex_key#8 mapping.invalidate_lock jbd2_handle &ei->i_data_sem rcu_read_lock rcu_node_0 irq_context: 0 sb_writers#4 &sb->s_type->i_mutex_key#8 mapping.invalidate_lock jbd2_handle &ei->i_data_sem rcu_read_lock &rq->__lock irq_context: 0 sb_writers#4 &sb->s_type->i_mutex_key#8 mapping.invalidate_lock jbd2_handle &ei->i_data_sem rcu_read_lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 sb_writers#4 &sb->s_type->i_mutex_key#8 mapping.invalidate_lock jbd2_handle &ei->i_data_sem &bgl->locks[i].lock &ei->i_prealloc_lock irq_context: 0 sb_writers#4 &sb->s_type->i_mutex_key#8 mapping.invalidate_lock jbd2_handle &ei->i_data_sem &pa->pa_lock#2 irq_context: 0 sb_writers#4 &sb->s_type->i_mutex_key#8 mapping.invalidate_lock jbd2_handle &ei->i_data_sem &ei->i_prealloc_lock &pa->pa_lock#2 irq_context: 0 sb_writers#4 &sb->s_type->i_mutex_key#8 mapping.invalidate_lock jbd2_handle &ei->i_es_lock key#2 irq_context: 0 sb_writers#4 &sb->s_type->i_mutex_key#8 jbd2_handle &ei->i_data_sem &ei->i_es_lock key#7 irq_context: 0 sb_writers#4 &sb->s_type->i_mutex_key#8 stock_lock irq_context: 0 sb_writers#4 &sb->s_type->i_mutex_key#8 &sb->s_type->i_mutex_key#8/4 jbd2_handle &____s->seqcount irq_context: 0 sb_writers#4 &sb->s_type->i_mutex_key#8 &sb->s_type->i_mutex_key#8/4 jbd2_handle &xa->xa_lock#9 irq_context: 0 sb_writers#4 &sb->s_type->i_mutex_key#8 &sb->s_type->i_mutex_key#8/4 jbd2_handle &xa->xa_lock#9 stock_lock irq_context: 0 sb_writers#4 &sb->s_type->i_mutex_key#8 &sb->s_type->i_mutex_key#8/4 jbd2_handle &xa->xa_lock#9 pool_lock#2 irq_context: 0 sb_writers#4 &sb->s_type->i_mutex_key#8 &sb->s_type->i_mutex_key#8/4 jbd2_handle lock#4 irq_context: 0 sb_writers#4 &sb->s_type->i_mutex_key#8 &sb->s_type->i_mutex_key#8/4 jbd2_handle &xa->xa_lock#9 &c->lock irq_context: 0 sb_writers#4 &sb->s_type->i_mutex_key#8 &sb->s_type->i_mutex_key#8/4 jbd2_handle &ei->i_data_sem &ei->i_data_sem/1 &ret->b_state_lock irq_context: 0 sb_writers#4 &sb->s_type->i_mutex_key#8 &sb->s_type->i_mutex_key#8/4 jbd2_handle &ei->i_data_sem &ei->i_data_sem/1 &ret->b_state_lock &journal->j_list_lock irq_context: 0 sb_writers#4 &sb->s_type->i_mutex_key#8 &sb->s_type->i_mutex_key#8/4 jbd2_handle &ei->i_data_sem &ei->i_data_sem/1 &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 sb_writers#4 &sb->s_type->i_mutex_key#8 &sb->s_type->i_mutex_key#8/4 jbd2_handle &ei->i_data_sem &ei->i_data_sem/1 key#3 irq_context: 0 sb_writers#4 &sb->s_type->i_mutex_key#8 &sb->s_type->i_mutex_key#8/4 jbd2_handle &ei->i_data_sem &ei->i_data_sem/1 key#14 irq_context: 0 sb_writers#4 &sb->s_type->i_mutex_key#8 &sb->s_type->i_mutex_key#8/4 remove_cache_srcu irq_context: 0 sb_writers#4 &sb->s_type->i_mutex_key#8 &sb->s_type->i_mutex_key#8/4 remove_cache_srcu quarantine_lock irq_context: 0 sb_writers#4 &sb->s_type->i_mutex_key#8 &sb->s_type->i_mutex_key#8/4 remove_cache_srcu &c->lock irq_context: 0 sb_writers#4 &sb->s_type->i_mutex_key#8 &sb->s_type->i_mutex_key#8/4 remove_cache_srcu &n->list_lock irq_context: 0 sb_writers#4 &sb->s_type->i_mutex_key#8 &sb->s_type->i_mutex_key#8/4 remove_cache_srcu rcu_read_lock pool_lock#2 irq_context: 0 sb_writers#4 &sb->s_type->i_mutex_key#8 &sb->s_type->i_mutex_key#8/4 remove_cache_srcu &obj_hash[i].lock irq_context: 0 sb_writers#4 &sb->s_type->i_mutex_key#8 &sb->s_type->i_mutex_key#8/4 remove_cache_srcu &rq->__lock irq_context: 0 sb_writers#4 &sb->s_type->i_mutex_key#8 &sb->s_type->i_mutex_key#8/4 remove_cache_srcu &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 sb_writers#4 &sb->s_type->i_mutex_key#8 &sb->s_type->i_mutex_key#8/4 remove_cache_srcu pool_lock#2 irq_context: 0 sb_writers#4 &sb->s_type->i_mutex_key#8 &sb->s_type->i_mutex_key#8/4 key#3 irq_context: 0 sb_writers#4 &sb->s_type->i_mutex_key#8 &sb->s_type->i_mutex_key#8/4 key#14 irq_context: 0 sb_writers#4 &sb->s_type->i_mutex_key#8 &sb->s_type->i_mutex_key#8/4 &ei->i_data_sem &ei->i_data_sem/1 &mapping->private_lock irq_context: 0 sb_writers#4 &sb->s_type->i_lock_key#22 &xa->xa_lock#9 rcu_read_lock per_cpu_ptr(&cgroup_rstat_cpu_lock, cpu) irq_context: 0 &f->f_pos_lock sb_writers#4 &sb->s_type->i_mutex_key#8 &ei->i_data_sem rcu_read_lock &rq->__lock irq_context: 0 &f->f_pos_lock sb_writers#4 &sb->s_type->i_mutex_key#8 &ei->i_data_sem rcu_read_lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)ext4-rsv-conversion (work_completion)(&ei->i_rsv_conversion_work) rcu_read_lock &rq->__lock &cfs_rq->removed.lock irq_context: 0 sb_writers#4 &type->i_mutex_dir_key#3/1 &sb->s_type->i_mutex_key#8 remove_cache_srcu &cfs_rq->removed.lock irq_context: 0 sb_writers#4 &sb->s_type->i_mutex_key#8 mapping.invalidate_lock jbd2_handle &ei->i_data_sem &ei->i_es_lock &c->lock irq_context: 0 sb_writers#4 &sb->s_type->i_mutex_key#8 mapping.invalidate_lock jbd2_handle &ei->i_data_sem &bgl->locks[i].lock &pa->pa_lock irq_context: 0 sb_writers#4 &sb->s_type->i_mutex_key#8 mapping.invalidate_lock jbd2_handle &ei->i_data_sem &bgl->locks[i].lock &lg->lg_prealloc_lock irq_context: 0 sb_writers#4 &sb->s_type->i_mutex_key#8 mapping.invalidate_lock &journal->j_state_lock &journal->j_wait_commit irq_context: 0 sb_writers#4 &sb->s_type->i_mutex_key#8 mapping.invalidate_lock &journal->j_state_lock &journal->j_wait_commit &p->pi_lock irq_context: 0 sb_writers#4 &sb->s_type->i_mutex_key#8 mapping.invalidate_lock &journal->j_state_lock &journal->j_wait_commit &p->pi_lock &rq->__lock irq_context: 0 sb_writers#4 &sb->s_type->i_mutex_key#8 mapping.invalidate_lock &journal->j_state_lock &journal->j_wait_commit &p->pi_lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 sb_writers#4 &sb->s_type->i_mutex_key#8 mapping.invalidate_lock jbd2_handle irq_context: 0 sb_writers#4 &sb->s_type->i_mutex_key#8 mapping.invalidate_lock &journal->j_wait_commit irq_context: 0 sb_writers#4 &sb->s_type->i_mutex_key#8 mapping.invalidate_lock &journal->j_wait_done_commit irq_context: 0 sb_writers#4 &sb->s_type->i_mutex_key#8 &sb->s_type->i_mutex_key#8/4 jbd2_handle &ei->i_data_sem &ei->i_data_sem/1 &____s->seqcount#2 irq_context: 0 sb_writers#4 &sb->s_type->i_mutex_key#8 &sb->s_type->i_mutex_key#8/4 jbd2_handle &ei->i_data_sem &ei->i_data_sem/1 &n->list_lock irq_context: 0 sb_writers#4 &sb->s_type->i_mutex_key#8 &sb->s_type->i_mutex_key#8/4 jbd2_handle &ei->i_data_sem &ei->i_data_sem/1 &n->list_lock &c->lock irq_context: 0 sb_writers#4 &sb->s_type->i_mutex_key#8 &sb->s_type->i_mutex_key#8/4 jbd2_handle rcu_read_lock rcu_node_0 irq_context: 0 sb_writers#4 &sb->s_type->i_mutex_key#8 &sb->s_type->i_mutex_key#8/4 jbd2_handle rcu_read_lock &rq->__lock irq_context: 0 sb_writers#4 &sb->s_type->i_mutex_key#8 &sb->s_type->i_mutex_key#8/4 jbd2_handle rcu_read_lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 sb_writers#4 &sb->s_type->i_mutex_key#8 &sb->s_type->i_mutex_key#8/4 jbd2_handle lock#4 &lruvec->lru_lock irq_context: 0 sb_writers#4 &sb->s_type->i_mutex_key#8 &sb->s_type->i_mutex_key#8/4 jbd2_handle &ei->i_data_sem &ei->i_data_sem/1 rcu_read_lock rcu_node_0 irq_context: 0 sb_writers#4 &sb->s_type->i_mutex_key#8 &sb->s_type->i_mutex_key#8/4 jbd2_handle &ei->i_data_sem &ei->i_data_sem/1 rcu_read_lock &rq->__lock irq_context: 0 sb_writers#4 &sb->s_type->i_mutex_key#8 &sb->s_type->i_mutex_key#8/4 jbd2_handle &ei->i_data_sem &ei->i_data_sem/1 rcu_read_lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 sb_writers#4 &sb->s_type->i_mutex_key#8 &sb->s_type->i_mutex_key#8/4 &c->lock irq_context: 0 sb_writers#4 &sb->s_type->i_mutex_key#8 &sb->s_type->i_mutex_key#8/4 &____s->seqcount#2 irq_context: 0 sb_writers#4 &sb->s_type->i_mutex_key#8 &sb->s_type->i_mutex_key#8/4 &____s->seqcount irq_context: 0 sb_writers#4 &sb->s_type->i_mutex_key#8 &sb->s_type->i_mutex_key#8/4 jbd2_handle &rq->__lock irq_context: 0 sb_writers#4 &sb->s_type->i_mutex_key#8 &sb->s_type->i_mutex_key#8/4 jbd2_handle &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 sb_writers#4 &sb->s_type->i_mutex_key#8 &sb->s_type->i_mutex_key#8/4 jbd2_handle &ei->i_data_sem &ei->i_data_sem/1 batched_entropy_u8.lock irq_context: 0 sb_writers#4 &sb->s_type->i_mutex_key#8 &sb->s_type->i_mutex_key#8/4 jbd2_handle &ei->i_data_sem &ei->i_data_sem/1 kfence_freelist_lock irq_context: 0 sb_writers#4 &sb->s_type->i_mutex_key#8 &sb->s_type->i_mutex_key#8/4 jbd2_handle &ei->i_data_sem &ei->i_data_sem/1 &meta->lock irq_context: 0 sb_writers#4 &sb->s_type->i_mutex_key#8 &sb->s_type->i_mutex_key#8/4 jbd2_handle &ei->i_data_sem &ei->i_data_sem/1 &cfs_rq->removed.lock irq_context: 0 sb_writers#4 &sb->s_type->i_mutex_key#8 &sb->s_type->i_mutex_key#8/4 &n->list_lock irq_context: 0 sb_writers#4 &sb->s_type->i_mutex_key#8 &sb->s_type->i_mutex_key#8/4 &n->list_lock &c->lock irq_context: 0 sb_writers#4 &sb->s_type->i_mutex_key#8 &sb->s_type->i_mutex_key#8/4 jbd2_handle &ei->i_data_sem &ei->i_data_sem/1 &mapping->private_lock irq_context: 0 sb_writers#4 &sb->s_type->i_mutex_key#8 &sb->s_type->i_mutex_key#8/4 &ei->i_data_sem &rq->__lock irq_context: 0 sb_writers#4 &sb->s_type->i_mutex_key#8 &sb->s_type->i_mutex_key#8/4 &ei->i_data_sem &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 sb_writers#4 &sb->s_type->i_mutex_key#8 &sb->s_type->i_mutex_key#8/4 rcu_read_lock &rq->__lock irq_context: 0 sb_writers#4 &sb->s_type->i_mutex_key#8 &sb->s_type->i_mutex_key#8/4 rcu_read_lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 sb_writers#4 &sb->s_type->i_mutex_key#8 &sb->s_type->i_mutex_key#8/4 &journal->j_state_lock tk_core.seq.seqcount irq_context: 0 sb_writers#4 &sb->s_type->i_mutex_key#8 &sb->s_type->i_mutex_key#8/4 &journal->j_state_lock &obj_hash[i].lock irq_context: 0 sb_writers#4 &sb->s_type->i_mutex_key#8 &sb->s_type->i_mutex_key#8/4 &journal->j_state_lock &base->lock irq_context: 0 sb_writers#4 &sb->s_type->i_mutex_key#8 &sb->s_type->i_mutex_key#8/4 &journal->j_state_lock &base->lock &obj_hash[i].lock irq_context: 0 &journal->j_state_lock &journal->j_list_lock &obj_hash[i].lock irq_context: 0 &journal->j_state_lock &journal->j_list_lock pool_lock#2 irq_context: 0 &p->lock remove_cache_srcu &cfs_rq->removed.lock irq_context: softirq rcu_read_lock_bh rcu_read_lock &pool->lock/1 irq_context: softirq rcu_read_lock_bh rcu_read_lock &pool->lock/1 &obj_hash[i].lock irq_context: softirq rcu_read_lock_bh rcu_read_lock &pool->lock/1 pool_lock#2 irq_context: softirq rcu_read_lock_bh rcu_read_lock &pool->lock/1 &p->pi_lock irq_context: 0 (wq_completion)wg-kex-wg0#4 (work_completion)(&({ do { const void *__vpp_verify = (typeof((worker) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __asm__ ("" : "=r"(__ptr) : "0"((typeof(*((worker))) *)((worker)))); (typeof((typeof(*((worker))) *)((worker)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); })->work) &c->lock irq_context: 0 (wq_completion)wg-kex-wg1#4 (work_completion)(&({ do { const void *__vpp_verify = (typeof((worker) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __asm__ ("" : "=r"(__ptr) : "0"((typeof(*((worker))) *)((worker)))); (typeof((typeof(*((worker))) *)((worker)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); })->work) rcu_read_lock_bh rcu_read_lock &pool->lock &p->pi_lock &rq->__lock irq_context: 0 (wq_completion)wg-kex-wg1#4 (work_completion)(&({ do { const void *__vpp_verify = (typeof((worker) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __asm__ ("" : "=r"(__ptr) : "0"((typeof(*((worker))) *)((worker)))); (typeof((typeof(*((worker))) *)((worker)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); })->work) rcu_read_lock_bh rcu_read_lock &pool->lock &p->pi_lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 &f->f_pos_lock &type->i_mutex_dir_key#3 sb_writers#4 rcu_node_0 irq_context: 0 &f->f_pos_lock sb_writers#3 irq_context: 0 &f->f_pos_lock sb_writers#3 &mm->mmap_lock irq_context: 0 &pipe->mutex/1 sk_lock-AF_INET6 (console_sem).lock irq_context: 0 &pipe->mutex/1 sk_lock-AF_INET6 console_lock console_srcu console_owner_lock irq_context: 0 &pipe->mutex/1 sk_lock-AF_INET6 console_lock console_srcu console_owner irq_context: 0 &pipe->mutex/1 sk_lock-AF_INET6 console_lock console_srcu console_owner &port_lock_key irq_context: 0 &pipe->mutex/1 sk_lock-AF_INET6 console_lock console_srcu console_owner console_owner_lock irq_context: 0 sb_writers#4 &sb->s_type->i_mutex_key#8 mapping.invalidate_lock &sb->s_type->i_lock_key#22 &xa->xa_lock#9 &obj_hash[i].lock pool_lock irq_context: 0 sb_writers#4 &sb->s_type->i_mutex_key#8 mapping.invalidate_lock jbd2_handle &ei->i_data_sem stock_lock irq_context: 0 sb_writers#4 &sb->s_type->i_mutex_key#8 mapping.invalidate_lock jbd2_handle &ei->i_data_sem &xa->xa_lock#9 irq_context: 0 sb_writers#4 &sb->s_type->i_mutex_key#8 mapping.invalidate_lock jbd2_handle &ei->i_data_sem lock#4 irq_context: 0 &f->f_pos_lock (console_sem).lock irq_context: 0 &f->f_pos_lock console_owner_lock irq_context: 0 &f->f_pos_lock console_owner irq_context: 0 &f->f_pos_lock console_lock console_srcu console_owner_lock irq_context: 0 &pipe->mutex/1 sk_lock-AF_INET6 console_owner_lock irq_context: 0 &f->f_pos_lock console_lock console_srcu console_owner irq_context: 0 &pipe->mutex/1 sk_lock-AF_INET6 console_owner irq_context: 0 &f->f_pos_lock console_lock console_srcu console_owner &port_lock_key irq_context: 0 &f->f_pos_lock console_lock console_srcu console_owner console_owner_lock irq_context: 0 sb_writers#4 &sb->s_type->i_mutex_key#8 mapping.invalidate_lock jbd2_handle &ei->i_data_sem &ei->i_es_lock key#6 irq_context: 0 (wq_completion)wg-kex-wg1#9 (work_completion)(&peer->transmit_handshake_work) &rq->__lock irq_context: 0 (wq_completion)wg-kex-wg1#9 (work_completion)(&peer->transmit_handshake_work) &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 sb_writers#4 pgd_lock irq_context: 0 sb_writers#4 key irq_context: 0 sb_writers#4 pcpu_lock irq_context: 0 sb_writers#4 percpu_counters_lock irq_context: 0 (wq_completion)wg-kex-wg1#11 (work_completion)(&({ do { const void *__vpp_verify = (typeof((worker) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __asm__ ("" : "=r"(__ptr) : "0"((typeof(*((worker))) *)((worker)))); (typeof((typeof(*((worker))) *)((worker)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); })->work) rcu_read_lock_bh rcu_read_lock &pool->lock &p->pi_lock &rq->__lock irq_context: 0 (wq_completion)wg-kex-wg1#11 (work_completion)(&({ do { const void *__vpp_verify = (typeof((worker) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __asm__ ("" : "=r"(__ptr) : "0"((typeof(*((worker))) *)((worker)))); (typeof((typeof(*((worker))) *)((worker)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); })->work) rcu_read_lock_bh rcu_read_lock &pool->lock &p->pi_lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)events_unbound (reaper_work).work rcu_read_lock &rq->__lock &cfs_rq->removed.lock irq_context: 0 sb_writers#4 &sb->s_type->i_mutex_key#8 &sb->s_type->i_mutex_key#8/4 jbd2_handle &xa->xa_lock#9 &n->list_lock irq_context: 0 sb_writers#4 &sb->s_type->i_mutex_key#8 &sb->s_type->i_mutex_key#8/4 jbd2_handle &xa->xa_lock#9 &n->list_lock &c->lock irq_context: 0 sb_writers#4 &sb->s_type->i_mutex_key#8 &sb->s_type->i_mutex_key#8/4 jbd2_handle &ei->i_data_sem &ei->i_data_sem/1 remove_cache_srcu irq_context: 0 sb_writers#4 &sb->s_type->i_mutex_key#8 &sb->s_type->i_mutex_key#8/4 jbd2_handle &ei->i_data_sem &ei->i_data_sem/1 remove_cache_srcu quarantine_lock irq_context: 0 sb_writers#4 &sb->s_type->i_mutex_key#8 &sb->s_type->i_mutex_key#8/4 jbd2_handle &ei->i_data_sem &ei->i_data_sem/1 remove_cache_srcu &c->lock irq_context: 0 sb_writers#4 &sb->s_type->i_mutex_key#8 &sb->s_type->i_mutex_key#8/4 jbd2_handle &ei->i_data_sem &ei->i_data_sem/1 remove_cache_srcu &n->list_lock irq_context: 0 sb_writers#4 &sb->s_type->i_mutex_key#8 &sb->s_type->i_mutex_key#8/4 jbd2_handle &ei->i_data_sem &ei->i_data_sem/1 remove_cache_srcu rcu_read_lock pool_lock#2 irq_context: 0 sb_writers#4 &sb->s_type->i_mutex_key#8 &sb->s_type->i_mutex_key#8/4 jbd2_handle &ei->i_data_sem &ei->i_data_sem/1 remove_cache_srcu rcu_read_lock rcu_node_0 irq_context: 0 sb_writers#4 &sb->s_type->i_mutex_key#8 &sb->s_type->i_mutex_key#8/4 jbd2_handle &ei->i_data_sem &ei->i_data_sem/1 remove_cache_srcu rcu_read_lock &rq->__lock irq_context: 0 sb_writers#4 &sb->s_type->i_mutex_key#8 &sb->s_type->i_mutex_key#8/4 jbd2_handle &ei->i_data_sem &ei->i_data_sem/1 remove_cache_srcu rcu_read_lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 sb_writers#4 &sb->s_type->i_mutex_key#8 &sb->s_type->i_mutex_key#8/4 jbd2_handle &ei->i_data_sem &ei->i_data_sem/1 remove_cache_srcu &obj_hash[i].lock irq_context: 0 sb_writers#4 &sb->s_type->i_mutex_key#8 &sb->s_type->i_mutex_key#8/4 jbd2_handle &ei->i_data_sem &ei->i_data_sem/1 remove_cache_srcu pool_lock#2 irq_context: 0 sb_writers#4 &sb->s_type->i_mutex_key#8 &sb->s_type->i_mutex_key#8/4 jbd2_handle &ei->i_data_sem &ei->i_data_sem/1 remove_cache_srcu &rq->__lock irq_context: 0 sb_writers#4 &sb->s_type->i_mutex_key#8 &sb->s_type->i_mutex_key#8/4 jbd2_handle &ei->i_data_sem &ei->i_data_sem/1 remove_cache_srcu &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 sb_writers#4 &sb->s_type->i_mutex_key#8 &sb->s_type->i_mutex_key#8/4 jbd2_handle &ei->i_data_sem &ei->i_data_sem/1 mmu_notifier_invalidate_range_start &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 sb_writers#4 &sb->s_type->i_mutex_key#8 &sb->s_type->i_mutex_key#8/4 jbd2_handle &ei->i_data_sem &ei->i_data_sem/1 &____s->seqcount irq_context: 0 sb_writers#4 &sb->s_type->i_mutex_key#8 &sb->s_type->i_mutex_key#8/4 jbd2_handle &ei->i_data_sem &ei->i_data_sem/1 rcu_node_0 irq_context: 0 sb_writers#4 &sb->s_type->i_mutex_key#8 &sb->s_type->i_mutex_key#8/4 jbd2_handle rcu_read_lock rcu_read_lock rcu_node_0 irq_context: 0 sb_writers#4 &sb->s_type->i_mutex_key#8 &sb->s_type->i_mutex_key#8/4 jbd2_handle rcu_read_lock rcu_read_lock &rq->__lock irq_context: 0 sb_writers#4 &sb->s_type->i_mutex_key#8 &sb->s_type->i_mutex_key#8/4 jbd2_handle rcu_read_lock rcu_read_lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 &pipe->mutex/1 sk_lock-AF_INET6 rcu_read_lock &____s->seqcount#9 irq_context: 0 &pipe->mutex/1 sk_lock-AF_INET6 rcu_read_lock rcu_read_lock_bh &obj_hash[i].lock irq_context: 0 &pipe->mutex/1 sk_lock-AF_INET6 rcu_read_lock rcu_read_lock_bh pool_lock#2 irq_context: 0 sb_writers#4 sb_internal jbd2_handle &journal->j_wait_updates &p->pi_lock &cfs_rq->removed.lock irq_context: 0 sb_writers#4 &type->i_mutex_dir_key#3/1 jbd2_handle &ei->i_data_sem &rq->__lock &cfs_rq->removed.lock irq_context: 0 fs_reclaim &rq->__lock &cfs_rq->removed.lock irq_context: softirq (&q->timer) irq_context: softirq (&q->timer) rcu_read_lock &q->lock#2 irq_context: softirq (&q->timer) rcu_read_lock &q->lock#2 &obj_hash[i].lock irq_context: softirq (&q->timer) rcu_read_lock &q->lock#2 rcu_read_lock rcu_read_lock rhashtable_bucket irq_context: softirq (&q->timer) rcu_read_lock &q->lock#2 rcu_read_lock rcu_read_lock rcu_read_lock &pool->lock irq_context: softirq (&q->timer) &obj_hash[i].lock irq_context: softirq (&q->timer) rcu_read_lock &q->lock#2 rcu_read_lock rcu_read_lock rcu_read_lock &pool->lock &obj_hash[i].lock irq_context: softirq (&q->timer) pool_lock#2 irq_context: softirq (&q->timer) rcu_read_lock &q->lock#2 rcu_read_lock rcu_read_lock rcu_read_lock &pool->lock pool_lock#2 irq_context: softirq (&q->timer) rcu_read_lock &q->lock#2 rcu_read_lock rcu_read_lock rcu_read_lock &pool->lock &p->pi_lock irq_context: softirq (&pool->idle_timer) &pool->lock irq_context: softirq (&q->timer) rcu_read_lock &q->lock#2 rcu_read_lock rcu_read_lock rcu_read_lock &pool->lock &p->pi_lock &rq->__lock irq_context: softirq (&pool->idle_timer) &pool->lock &obj_hash[i].lock irq_context: softirq (&q->timer) rcu_read_lock &q->lock#2 rcu_read_lock rcu_read_lock rcu_read_lock &pool->lock &p->pi_lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: softirq (&pool->idle_timer) &pool->lock &base->lock irq_context: softirq (&pool->idle_timer) &pool->lock &base->lock &obj_hash[i].lock irq_context: 0 remove_cache_srcu rcu_read_lock &obj_hash[i].lock irq_context: softirq rcu_read_lock rcu_read_lock rcu_read_lock rlock-AF_INET6 irq_context: 0 hashlimit_mutex fs_reclaim &rq->__lock irq_context: 0 hashlimit_mutex fs_reclaim &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 hashlimit_mutex rcu_read_lock rcu_node_0 irq_context: 0 hashlimit_mutex rcu_read_lock &rq->__lock irq_context: 0 hashlimit_mutex rcu_read_lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: softirq &(&hinfo->gc_work)->timer rcu_read_lock &pool->lock &p->pi_lock &rq->__lock &cfs_rq->removed.lock irq_context: 0 &child->perf_event_mutex &rq->__lock &cfs_rq->removed.lock irq_context: 0 purge_vmap_area_lock &____s->seqcount irq_context: 0 purge_vmap_area_lock rcu_read_lock pool_lock#2 irq_context: 0 &ei->i_data_sem &sem->wait_lock irq_context: 0 (wq_completion)wg-kex-wg1#9 (work_completion)(&peer->transmit_handshake_work) &n->list_lock irq_context: 0 (wq_completion)wg-kex-wg1#9 (work_completion)(&peer->transmit_handshake_work) &n->list_lock &c->lock irq_context: 0 (wq_completion)wg-kex-wg0#11 (work_completion)(&({ do { const void *__vpp_verify = (typeof((worker) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __asm__ ("" : "=r"(__ptr) : "0"((typeof(*((worker))) *)((worker)))); (typeof((typeof(*((worker))) *)((worker)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); })->work) &handshake->lock rcu_read_lock_bh &peer->keypairs.keypair_update_lock &table->lock#2 irq_context: 0 (wq_completion)wg-kex-wg0#11 (work_completion)(&({ do { const void *__vpp_verify = (typeof((worker) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __asm__ ("" : "=r"(__ptr) : "0"((typeof(*((worker))) *)((worker)))); (typeof((typeof(*((worker))) *)((worker)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); })->work) &handshake->lock rcu_read_lock_bh &peer->keypairs.keypair_update_lock &obj_hash[i].lock irq_context: 0 (wq_completion)wg-kex-wg1#11 (work_completion)(&({ do { const void *__vpp_verify = (typeof((worker) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __asm__ ("" : "=r"(__ptr) : "0"((typeof(*((worker))) *)((worker)))); (typeof((typeof(*((worker))) *)((worker)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); })->work) &handshake->lock rcu_read_lock_bh &peer->keypairs.keypair_update_lock &table->lock#2 irq_context: 0 (wq_completion)wg-kex-wg1#11 (work_completion)(&({ do { const void *__vpp_verify = (typeof((worker) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __asm__ ("" : "=r"(__ptr) : "0"((typeof(*((worker))) *)((worker)))); (typeof((typeof(*((worker))) *)((worker)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); })->work) &handshake->lock rcu_read_lock_bh &peer->keypairs.keypair_update_lock &obj_hash[i].lock irq_context: 0 (wq_completion)wg-kex-wg1#11 (work_completion)(&({ do { const void *__vpp_verify = (typeof((worker) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __asm__ ("" : "=r"(__ptr) : "0"((typeof(*((worker))) *)((worker)))); (typeof((typeof(*((worker))) *)((worker)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); })->work) &c->lock irq_context: 0 sb_writers#8 kn->active#5 remove_cache_srcu rcu_read_lock &rcu_state.gp_wq irq_context: 0 sb_writers#8 kn->active#5 remove_cache_srcu rcu_read_lock &rcu_state.gp_wq &p->pi_lock irq_context: 0 sb_writers#8 kn->active#5 remove_cache_srcu rcu_read_lock &rcu_state.gp_wq &p->pi_lock &rq->__lock irq_context: 0 sb_writers#8 kn->active#5 remove_cache_srcu rcu_read_lock &rcu_state.gp_wq &p->pi_lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 &pipe->mutex/1 rcu_read_lock rcu_read_lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 sb_writers#4 &type->i_mutex_dir_key#3/1 jbd2_handle &ei->i_data_sem mmu_notifier_invalidate_range_start &rq->__lock irq_context: 0 sb_writers#4 &type->i_mutex_dir_key#3/1 jbd2_handle &ei->i_data_sem mmu_notifier_invalidate_range_start &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: softirq (t) irq_context: softirq (t) cmci_poll_lock irq_context: softirq (t) &obj_hash[i].lock irq_context: softirq (t) &base->lock irq_context: softirq (t) &base->lock &obj_hash[i].lock irq_context: softirq security/integrity/ima/ima_queue_keys.c:35 irq_context: softirq security/integrity/ima/ima_queue_keys.c:35 rcu_read_lock &pool->lock irq_context: softirq security/integrity/ima/ima_queue_keys.c:35 rcu_read_lock &pool->lock &obj_hash[i].lock irq_context: softirq security/integrity/ima/ima_queue_keys.c:35 rcu_read_lock &pool->lock &p->pi_lock irq_context: softirq security/integrity/ima/ima_queue_keys.c:35 rcu_read_lock &pool->lock &p->pi_lock &rq->__lock irq_context: softirq security/integrity/ima/ima_queue_keys.c:35 rcu_read_lock &pool->lock &p->pi_lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: softirq security/integrity/ima/ima_queue_keys.c:35 rcu_read_lock &pool->lock &p->pi_lock &rq->__lock &cfs_rq->removed.lock irq_context: 0 (wq_completion)events (ima_keys_delayed_work).work irq_context: 0 (wq_completion)events (ima_keys_delayed_work).work ima_keys_lock irq_context: 0 (wq_completion)events (ima_keys_delayed_work).work &obj_hash[i].lock irq_context: 0 (wq_completion)events (ima_keys_delayed_work).work pool_lock#2 irq_context: 0 sb_writers#8 kn->active#5 &kernfs_locks->open_file_mutex[count] fs_reclaim &rq->__lock irq_context: 0 sb_writers#8 kn->active#5 &kernfs_locks->open_file_mutex[count] fs_reclaim &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 kn->active#59 &rq->__lock irq_context: 0 kn->active#59 &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 kn->active#59 fs_reclaim irq_context: 0 kn->active#59 fs_reclaim mmu_notifier_invalidate_range_start irq_context: 0 kn->active#59 stock_lock irq_context: 0 kn->active#59 &kernfs_locks->open_file_mutex[count] irq_context: 0 kn->active#59 &kernfs_locks->open_file_mutex[count] fs_reclaim irq_context: 0 kn->active#59 &kernfs_locks->open_file_mutex[count] fs_reclaim mmu_notifier_invalidate_range_start irq_context: 0 kn->active#59 &c->lock irq_context: 0 sb_writers#4 &type->i_mutex_dir_key#3/1 sb_internal batched_entropy_u8.lock irq_context: 0 sb_writers#4 &type->i_mutex_dir_key#3/1 sb_internal kfence_freelist_lock irq_context: 0 sb_writers#4 &type->i_mutex_dir_key#3/1 sb_internal &meta->lock irq_context: 0 (wq_completion)ext4-rsv-conversion &rq->__lock irq_context: 0 purge_vmap_area_lock &meta->lock irq_context: 0 purge_vmap_area_lock kfence_freelist_lock irq_context: 0 &pipe->mutex/1 sk_lock-AF_INET6 fs_reclaim &rq->__lock irq_context: 0 &pipe->mutex/1 sk_lock-AF_INET6 fs_reclaim &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 lock link_idr_lock &n->list_lock &c->lock irq_context: 0 sb_writers#4 &type->i_mutex_dir_key#3/1 &sb->s_type->i_mutex_key#8 remove_cache_srcu &rq->__lock &cfs_rq->removed.lock irq_context: 0 kn->active#59 &____s->seqcount#2 irq_context: 0 kn->active#59 &____s->seqcount irq_context: 0 prog_idr_lock &obj_hash[i].lock pool_lock irq_context: 0 (wq_completion)events_unbound (reaper_work).work &rq->__lock &cfs_rq->removed.lock irq_context: 0 &f->f_pos_lock sb_writers#4 &sb->s_type->i_mutex_key#8 &mm->mmap_lock remove_cache_srcu pool_lock#2 irq_context: 0 &f->f_pos_lock sb_writers#4 &sb->s_type->i_mutex_key#8 &mm->mmap_lock remove_cache_srcu &rq->__lock irq_context: 0 &f->f_pos_lock sb_writers#4 &sb->s_type->i_mutex_key#8 &mm->mmap_lock remove_cache_srcu &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 &f->f_pos_lock sb_writers#4 &sb->s_type->i_mutex_key#8 &mm->mmap_lock remove_cache_srcu rcu_read_lock rcu_node_0 irq_context: 0 &f->f_pos_lock sb_writers#4 &sb->s_type->i_mutex_key#8 &mm->mmap_lock remove_cache_srcu rcu_read_lock &rq->__lock irq_context: 0 &f->f_pos_lock sb_writers#4 &sb->s_type->i_mutex_key#8 &mm->mmap_lock remove_cache_srcu rcu_read_lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 sb_writers#4 tomoyo_ss mmu_notifier_invalidate_range_start &rq->__lock irq_context: 0 sb_writers#4 tomoyo_ss mmu_notifier_invalidate_range_start &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 &sb->s_type->i_mutex_key#8 &sbi->s_writepages_rwsem lock#4 rcu_read_lock pool_lock#2 irq_context: 0 &sb->s_type->i_mutex_key#8 &sbi->s_writepages_rwsem lock#4 &obj_hash[i].lock irq_context: 0 sb_writers#4 &type->i_mutex_dir_key#3/1 &type->i_mutex_dir_key#3 jbd2_handle &rq->__lock &cfs_rq->removed.lock irq_context: 0 sb_writers#5 &sb->s_type->i_mutex_key#12 rcu_read_lock &rcu_state.gp_wq irq_context: 0 sb_writers#5 &sb->s_type->i_mutex_key#12 rcu_read_lock &rcu_state.gp_wq &p->pi_lock irq_context: 0 sb_writers#5 &sb->s_type->i_mutex_key#12 rcu_read_lock &rcu_state.gp_wq &p->pi_lock &rq->__lock irq_context: 0 sb_writers#5 &sb->s_type->i_mutex_key#12 rcu_read_lock &rcu_state.gp_wq &p->pi_lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 sb_writers#5 &type->i_mutex_dir_key#5/1 &sb->s_type->i_mutex_key#12 &sb->s_type->i_mutex_key#12/4 rcu_read_lock &rcu_state.gp_wq irq_context: 0 sb_writers#5 &type->i_mutex_dir_key#5/1 &sb->s_type->i_mutex_key#12 &sb->s_type->i_mutex_key#12/4 rcu_read_lock &rcu_state.gp_wq &p->pi_lock irq_context: 0 sb_writers#5 &type->i_mutex_dir_key#5/1 &sb->s_type->i_mutex_key#12 &sb->s_type->i_mutex_key#12/4 rcu_read_lock &rcu_state.gp_wq &p->pi_lock &rq->__lock irq_context: 0 sb_writers#5 &type->i_mutex_dir_key#5/1 &sb->s_type->i_mutex_key#12 &sb->s_type->i_mutex_key#12/4 rcu_read_lock &rcu_state.gp_wq &p->pi_lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)rcu_gp (work_completion)(&(&ssp->srcu_sup->work)->work) &ssp->srcu_sup->srcu_gp_mutex pool_lock#2 irq_context: 0 &sb->s_type->i_mutex_key#8 &sbi->s_writepages_rwsem &rq->__lock &cfs_rq->removed.lock irq_context: 0 (wq_completion)events (work_completion)(&rdev->wiphy_work) &rdev->wiphy.mtx &wdev->mtx rcu_read_lock &rcu_state.gp_wq irq_context: 0 (wq_completion)events (work_completion)(&rdev->wiphy_work) &rdev->wiphy.mtx &wdev->mtx rcu_read_lock &rcu_state.gp_wq &p->pi_lock irq_context: 0 (wq_completion)events (work_completion)(&rdev->wiphy_work) &rdev->wiphy.mtx &wdev->mtx rcu_read_lock &rcu_state.gp_wq &p->pi_lock &rq->__lock irq_context: 0 (wq_completion)events (work_completion)(&rdev->wiphy_work) &rdev->wiphy.mtx &wdev->mtx rcu_read_lock &rcu_state.gp_wq &p->pi_lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)bat_events (work_completion)(&(&forw_packet_aggr->delayed_work)->work) &hard_iface->bat_iv.ogm_buff_mutex &bat_priv->tt.commit_lock &____s->seqcount#2 irq_context: 0 (wq_completion)bat_events (work_completion)(&(&forw_packet_aggr->delayed_work)->work) &hard_iface->bat_iv.ogm_buff_mutex &bat_priv->tt.commit_lock &____s->seqcount irq_context: softirq rcu_read_lock &rq->__lock &obj_hash[i].lock irq_context: softirq rcu_read_lock &rq->__lock &base->lock irq_context: softirq rcu_read_lock &rq->__lock &base->lock &obj_hash[i].lock irq_context: 0 sb_writers#5 &type->i_mutex_dir_key#5/1 remove_cache_srcu &rq->__lock irq_context: 0 sb_writers#5 &type->i_mutex_dir_key#5/1 remove_cache_srcu &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 &sbi->s_writepages_rwsem jbd2_handle &ei->i_data_sem &ei->i_es_lock key#6 irq_context: 0 sb_writers#5 &type->i_mutex_dir_key#5 &simple_offset_xa_lock batched_entropy_u8.lock irq_context: 0 sb_writers#5 &type->i_mutex_dir_key#5 &simple_offset_xa_lock kfence_freelist_lock irq_context: 0 &tsk->futex_exit_mutex &mm->mmap_lock &rq->__lock &cfs_rq->removed.lock irq_context: 0 sb_writers#4 &type->i_mutex_dir_key#3/1 sb_internal jbd2_handle &ei->i_data_sem &bgl->locks[i].lock quarantine_lock irq_context: 0 &mm->mmap_lock remove_cache_srcu &cfs_rq->removed.lock irq_context: 0 &f->f_pos_lock sb_writers#4 &sb->s_type->i_mutex_key#8 &ei->i_data_sem &rq->__lock &cfs_rq->removed.lock irq_context: 0 &f->f_pos_lock sb_writers#4 &sb->s_type->i_mutex_key#8 lock#4 &lruvec->lru_lock per_cpu_ptr(&cgroup_rstat_cpu_lock, cpu) irq_context: 0 rcu_read_lock rcu_read_lock &cfs_rq->removed.lock irq_context: 0 (wq_completion)ext4-rsv-conversion (work_completion)(&ei->i_rsv_conversion_work) jbd2_handle &ei->i_data_sem &ei->i_es_lock batched_entropy_u8.lock irq_context: 0 sb_writers#4 sb_internal jbd2_handle &ei->i_data_sem &ei->i_es_lock &meta->lock irq_context: 0 sb_writers#4 sb_internal jbd2_handle &ei->i_data_sem &ei->i_es_lock kfence_freelist_lock irq_context: 0 &sb->s_type->i_mutex_key#10 rcu_read_lock &obj_hash[i].lock irq_context: 0 pernet_ops_rwsem lock kernfs_idr_lock &n->list_lock irq_context: 0 pernet_ops_rwsem lock kernfs_idr_lock &n->list_lock &c->lock irq_context: 0 &iint->mutex fs_reclaim &rq->__lock irq_context: 0 &iint->mutex fs_reclaim &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 cb_lock &rdev->wiphy.mtx remove_cache_srcu irq_context: 0 cb_lock &rdev->wiphy.mtx remove_cache_srcu quarantine_lock irq_context: 0 cb_lock &rdev->wiphy.mtx remove_cache_srcu &c->lock irq_context: 0 cb_lock &rdev->wiphy.mtx remove_cache_srcu &n->list_lock irq_context: 0 cb_lock &rdev->wiphy.mtx remove_cache_srcu rcu_read_lock pool_lock#2 irq_context: 0 cb_lock &rdev->wiphy.mtx remove_cache_srcu &obj_hash[i].lock irq_context: 0 cb_lock &rdev->wiphy.mtx remove_cache_srcu &rq->__lock irq_context: 0 cb_lock &rdev->wiphy.mtx remove_cache_srcu &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 sb_writers#5 &type->i_mutex_dir_key#5/1 tomoyo_ss remove_cache_srcu rcu_read_lock rcu_read_lock rcu_node_0 irq_context: 0 sb_writers#5 &type->i_mutex_dir_key#5/1 tomoyo_ss remove_cache_srcu rcu_read_lock rcu_read_lock &rq->__lock irq_context: 0 sb_writers#5 &type->i_mutex_dir_key#5/1 tomoyo_ss remove_cache_srcu rcu_read_lock rcu_read_lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 pernet_ops_rwsem nf_ct_proto_mutex nf_hook_mutex &rq->__lock irq_context: 0 pernet_ops_rwsem nf_ct_proto_mutex nf_hook_mutex &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 pernet_ops_rwsem k-sk_lock-AF_RXRPC &rxnet->local_mutex cpu_hotplug_lock &rq->__lock irq_context: 0 pernet_ops_rwsem k-sk_lock-AF_RXRPC &rxnet->local_mutex cpu_hotplug_lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 cb_lock &rdev->wiphy.mtx rcu_read_lock &____s->seqcount#2 irq_context: 0 &mm->mmap_lock remove_cache_srcu rcu_node_0 irq_context: 0 (wq_completion)events (work_completion)(&(&nsim_dev->trap_data->trap_report_dw)->work) &devlink->lock_key &rcu_state.expedited_wq &p->pi_lock &rq->__lock irq_context: 0 cb_lock &rdev->wiphy.mtx rcu_read_lock &____s->seqcount irq_context: 0 (wq_completion)events (work_completion)(&(&nsim_dev->trap_data->trap_report_dw)->work) &devlink->lock_key &rcu_state.expedited_wq &p->pi_lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 rcu_read_lock &vma->vm_lock->lock &rcu_state.expedited_wq irq_context: 0 rcu_read_lock &vma->vm_lock->lock &rcu_state.expedited_wq &p->pi_lock irq_context: 0 rcu_read_lock &vma->vm_lock->lock &rcu_state.expedited_wq &p->pi_lock &rq->__lock irq_context: 0 rcu_read_lock &vma->vm_lock->lock &rcu_state.expedited_wq &p->pi_lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 dup_mmap_sem &mm->mmap_lock &mm->mmap_lock/1 ptlock_ptr(page)#2 ptlock_ptr(page)#2/1 rcu_read_lock &p->pi_lock &rq->__lock &cfs_rq->removed.lock irq_context: 0 sb_writers#4 &type->i_mutex_dir_key#3 tomoyo_ss quarantine_lock irq_context: 0 sb_writers#4 &type->i_mutex_dir_key#3 remove_cache_srcu &c->lock irq_context: 0 sb_writers#4 &type->i_mutex_dir_key#3 remove_cache_srcu &n->list_lock irq_context: 0 sb_writers#4 &type->i_mutex_dir_key#3 remove_cache_srcu &obj_hash[i].lock irq_context: 0 sb_writers#4 &type->i_mutex_dir_key#3 remove_cache_srcu pool_lock#2 irq_context: 0 sb_writers#4 &type->i_mutex_dir_key#3 remove_cache_srcu rcu_read_lock pool_lock#2 irq_context: 0 &sig->cred_guard_mutex tomoyo_ss rcu_read_lock rcu_read_lock rcu_node_0 irq_context: 0 &sig->cred_guard_mutex tomoyo_ss rcu_read_lock rcu_read_lock &rq->__lock irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem rtnl_mutex &root->kernfs_rwsem &cfs_rq->removed.lock irq_context: 0 sb_writers#8 &type->i_mutex_dir_key#4 &root->kernfs_rwsem &rq->__lock irq_context: 0 sb_writers#8 &type->i_mutex_dir_key#4 &root->kernfs_rwsem &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 sb_writers#8 &type->i_mutex_dir_key#4 &p->pi_lock &rq->__lock irq_context: 0 sb_writers#8 &type->i_mutex_dir_key#4 &p->pi_lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)wg-kex-wg2#9 (work_completion)(&peer->transmit_handshake_work) &n->list_lock irq_context: 0 (wq_completion)wg-kex-wg2#9 (work_completion)(&peer->transmit_handshake_work) &n->list_lock &c->lock irq_context: 0 pernet_ops_rwsem fill_pool_map-wait-type-override rcu_read_lock pool_lock#2 irq_context: 0 pernet_ops_rwsem fill_pool_map-wait-type-override &obj_hash[i].lock irq_context: 0 sb_writers#8 &of->mutex kn->active#5 uevent_sock_mutex remove_cache_srcu rcu_read_lock rcu_node_0 irq_context: 0 sb_writers#8 &of->mutex kn->active#5 uevent_sock_mutex remove_cache_srcu rcu_read_lock &rq->__lock irq_context: 0 sb_writers#8 &of->mutex kn->active#5 uevent_sock_mutex remove_cache_srcu rcu_read_lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem dev_pm_qos_sysfs_mtx dev_pm_qos_mtx &rq->__lock irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem dev_pm_qos_sysfs_mtx dev_pm_qos_mtx &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 pernet_ops_rwsem devices_rwsem rdma_nets_rwsem &device->compat_devs_mutex &p->pi_lock &rq->__lock &cfs_rq->removed.lock irq_context: 0 &type->i_mutex_dir_key#5 rcu_read_lock &p->pi_lock irq_context: 0 &type->i_mutex_dir_key#5 rcu_read_lock &p->pi_lock &rq->__lock irq_context: 0 &type->i_mutex_dir_key#5 rcu_read_lock &p->pi_lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 sb_writers#4 &type->i_mutex_dir_key#3/1 &type->i_mutex_dir_key#3 rcu_node_0 irq_context: 0 sb_writers#4 &type->i_mutex_dir_key#3/1 &type->i_mutex_dir_key#3 rcu_read_lock &rcu_state.expedited_wq irq_context: 0 sb_writers#4 &type->i_mutex_dir_key#3/1 &type->i_mutex_dir_key#3 rcu_read_lock &rcu_state.expedited_wq &p->pi_lock irq_context: 0 sb_writers#4 &type->i_mutex_dir_key#3/1 &type->i_mutex_dir_key#3 rcu_read_lock &rcu_state.expedited_wq &p->pi_lock &rq->__lock irq_context: 0 sb_writers#4 &type->i_mutex_dir_key#3/1 &type->i_mutex_dir_key#3 rcu_read_lock &rcu_state.expedited_wq &p->pi_lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)events (work_completion)(&(&ovs_net->masks_rebalance)->work) &rq->__lock irq_context: 0 (wq_completion)events (work_completion)(&(&ovs_net->masks_rebalance)->work) &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 &sig->cred_guard_mutex tomoyo_ss rcu_read_lock rcu_read_lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 &pipe->mutex/1 slock-AF_INET6 &sk->sk_lock.wq irq_context: 0 &pipe->mutex/1 slock-AF_INET6 &sk->sk_lock.wq &p->pi_lock irq_context: 0 &pipe->mutex/1 slock-AF_INET6 &sk->sk_lock.wq &p->pi_lock &rq->__lock irq_context: 0 &pipe->mutex/1 slock-AF_INET6 &sk->sk_lock.wq &p->pi_lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 sk_lock-AF_INET6 rcu_read_lock rcu_read_lock rcu_read_lock rcu_node_0 irq_context: 0 sk_lock-AF_INET6 rcu_read_lock rcu_read_lock rcu_read_lock &rq->__lock irq_context: 0 sk_lock-AF_INET6 rcu_read_lock rcu_read_lock rcu_read_lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 nf_sockopt_mutex &cfs_rq->removed.lock irq_context: 0 nf_sockopt_mutex &obj_hash[i].lock irq_context: softirq (&peer->timer_persistent_keepalive) rcu_read_lock_bh tk_core.seq.seqcount irq_context: softirq (&peer->timer_persistent_keepalive) rcu_read_lock_bh rcu_read_lock &pool->lock/1 irq_context: softirq (&peer->timer_persistent_keepalive) rcu_read_lock_bh rcu_read_lock &pool->lock/1 &obj_hash[i].lock irq_context: softirq (&peer->timer_persistent_keepalive) rcu_read_lock_bh rcu_read_lock &pool->lock/1 pool_lock#2 irq_context: softirq (&peer->timer_persistent_keepalive) rcu_read_lock_bh rcu_read_lock &pool->lock/1 &p->pi_lock irq_context: softirq (&peer->timer_persistent_keepalive) rcu_read_lock_bh rcu_read_lock &pool->lock/1 &p->pi_lock &rq->__lock irq_context: softirq (&peer->timer_persistent_keepalive) rcu_read_lock_bh rcu_read_lock &pool->lock/1 &p->pi_lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)wg-kex-wg0#7 (work_completion)(&peer->transmit_handshake_work) &peer->endpoint_lock rcu_read_lock_bh &n->list_lock irq_context: 0 (wq_completion)wg-kex-wg0#7 (work_completion)(&peer->transmit_handshake_work) &peer->endpoint_lock rcu_read_lock_bh &n->list_lock &c->lock irq_context: 0 &pipe->mutex/1 sk_lock-AF_INET6 rcu_read_lock &rcu_state.gp_wq irq_context: 0 &pipe->mutex/1 sk_lock-AF_INET6 rcu_read_lock &rcu_state.gp_wq &p->pi_lock irq_context: 0 &pipe->mutex/1 sk_lock-AF_INET6 rcu_read_lock &rcu_state.gp_wq &p->pi_lock &rq->__lock irq_context: 0 &pipe->mutex/1 sk_lock-AF_INET6 rcu_read_lock &rcu_state.gp_wq &p->pi_lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 ebt_mutex &mm->mmap_lock &cfs_rq->removed.lock irq_context: 0 ebt_mutex &mm->mmap_lock &obj_hash[i].lock irq_context: 0 sb_writers#5 &type->i_mutex_dir_key#5/1 &sb->s_type->i_lock_key &xa->xa_lock#9 &obj_hash[i].lock irq_context: 0 sb_writers#5 &type->i_mutex_dir_key#5/1 &sb->s_type->i_lock_key &xa->xa_lock#9 pool_lock#2 irq_context: 0 sb_writers#5 &type->i_mutex_dir_key#5/1 key#9 irq_context: 0 &pipe->mutex/1 sk_lock-AF_INET6 remove_cache_srcu rcu_read_lock rcu_node_0 irq_context: 0 &pipe->mutex/1 sk_lock-AF_INET6 remove_cache_srcu rcu_read_lock &rq->__lock irq_context: 0 &pipe->mutex/1 sk_lock-AF_INET6 remove_cache_srcu rcu_read_lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 &sb->s_type->i_mutex_key#9 &cfs_rq->removed.lock irq_context: 0 (wq_completion)events fqdir_free_work quarantine_lock irq_context: 0 sb_writers#3 &sb->s_type->i_mutex_key#9 rcu_read_lock &rcu_state.gp_wq irq_context: 0 sb_writers#3 &sb->s_type->i_mutex_key#9 rcu_read_lock &rcu_state.gp_wq &p->pi_lock irq_context: 0 sb_writers#3 &sb->s_type->i_mutex_key#9 rcu_read_lock &rcu_state.gp_wq &p->pi_lock &rq->__lock irq_context: 0 sb_writers#3 &sb->s_type->i_mutex_key#9 rcu_read_lock &rcu_state.gp_wq &p->pi_lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 sb_writers#4 &type->i_mutex_dir_key#3 tomoyo_ss &cfs_rq->removed.lock irq_context: 0 sb_writers#4 &type->i_mutex_dir_key#3 tomoyo_ss &obj_hash[i].lock pool_lock irq_context: 0 &mm->mmap_lock key#22 irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem rtnl_mutex remove_cache_srcu &pcp->lock &zone->lock irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem rtnl_mutex remove_cache_srcu &pcp->lock &zone->lock &____s->seqcount irq_context: 0 sb_writers#3 rcu_node_0 irq_context: 0 sb_writers#3 &rcu_state.expedited_wq irq_context: 0 sb_writers#3 &rcu_state.expedited_wq &p->pi_lock irq_context: 0 sb_writers#3 &rcu_state.expedited_wq &p->pi_lock &rq->__lock irq_context: 0 &mm->mmap_lock sb_pagefaults mmu_notifier_invalidate_range_start &rq->__lock irq_context: 0 sb_writers#3 &rcu_state.expedited_wq &p->pi_lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 sb_writers#4 &type->i_mutex_dir_key#3/1 sb_internal jbd2_handle &cfs_rq->removed.lock irq_context: 0 sb_writers#4 &type->i_mutex_dir_key#3/1 sb_internal jbd2_handle &obj_hash[i].lock irq_context: 0 sb_writers#4 &type->i_mutex_dir_key#3/1 jbd2_handle &ei->i_data_sem rcu_read_lock &rq->__lock &cfs_rq->removed.lock irq_context: 0 (wq_completion)gid-cache-wq (work_completion)(&ndev_work->work) &base->lock irq_context: 0 (wq_completion)gid-cache-wq (work_completion)(&ndev_work->work) &base->lock &obj_hash[i].lock irq_context: 0 &iint->mutex ima_extend_list_mutex rcu_read_lock &rq->__lock irq_context: 0 &iint->mutex ima_extend_list_mutex rcu_read_lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 pernet_ops_rwsem rtnl_mutex &root->kernfs_rwsem pgd_lock irq_context: 0 pernet_ops_rwsem rtnl_mutex &root->kernfs_rwsem stock_lock irq_context: 0 pernet_ops_rwsem rtnl_mutex &root->kernfs_rwsem rcu_read_lock pool_lock#2 irq_context: 0 pernet_ops_rwsem rtnl_mutex &root->kernfs_rwsem key irq_context: 0 pernet_ops_rwsem rtnl_mutex &root->kernfs_rwsem pcpu_lock irq_context: 0 pernet_ops_rwsem rtnl_mutex &root->kernfs_rwsem percpu_counters_lock irq_context: 0 pernet_ops_rwsem rtnl_mutex &root->kernfs_rwsem pcpu_lock stock_lock irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem &root->kernfs_rwsem rcu_read_lock &rcu_state.gp_wq irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem &root->kernfs_rwsem rcu_read_lock &rcu_state.gp_wq &p->pi_lock irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem &root->kernfs_rwsem rcu_read_lock &rcu_state.gp_wq &p->pi_lock &rq->__lock irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem &root->kernfs_rwsem rcu_read_lock &rcu_state.gp_wq &p->pi_lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem &root->kernfs_rwsem &base->lock irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem &root->kernfs_rwsem &base->lock &obj_hash[i].lock irq_context: 0 &f->f_pos_lock &type->i_mutex_dir_key#3 fs_reclaim &cfs_rq->removed.lock irq_context: 0 &f->f_pos_lock &type->i_mutex_dir_key#3 fs_reclaim &obj_hash[i].lock irq_context: 0 &sbi->s_writepages_rwsem jbd2_handle rcu_read_lock rcu_read_lock rcu_node_0 irq_context: 0 &sbi->s_writepages_rwsem jbd2_handle rcu_read_lock rcu_read_lock &rq->__lock irq_context: 0 &sbi->s_writepages_rwsem jbd2_handle rcu_read_lock rcu_read_lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem sysctl_lock fill_pool_map-wait-type-override &c->lock irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem __ip_vs_app_mutex &rq->__lock irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem __ip_vs_app_mutex &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 sb_writers#4 fill_pool_map-wait-type-override &cfs_rq->removed.lock irq_context: 0 sb_writers#4 fill_pool_map-wait-type-override &obj_hash[i].lock irq_context: 0 sb_writers#4 fill_pool_map-wait-type-override pool_lock#2 irq_context: 0 rcu_read_lock console_lock console_srcu console_owner_lock irq_context: 0 rcu_read_lock console_lock console_srcu console_owner irq_context: 0 rcu_read_lock console_lock console_srcu console_owner &port_lock_key irq_context: 0 rcu_read_lock console_lock console_srcu console_owner console_owner_lock irq_context: 0 pernet_ops_rwsem nf_ct_proto_mutex defrag4_mutex nf_hook_mutex rcu_read_lock &rq->__lock irq_context: 0 pernet_ops_rwsem nf_ct_proto_mutex defrag4_mutex nf_hook_mutex rcu_read_lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)events_power_efficient (work_completion)(&(&tbl->gc_work)->work) &tbl->lock &pcp->lock &zone->lock irq_context: 0 &mm->mmap_lock &xa->xa_lock#9 &pcp->lock &zone->lock irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem remove_cache_srcu rcu_read_lock rcu_node_0 irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem rtnl_mutex &root->kernfs_rwsem &rcu_state.expedited_wq irq_context: 0 (wq_completion)events (work_completion)(&rdev->wiphy_work) &rdev->wiphy.mtx &rcu_state.expedited_wq &p->pi_lock irq_context: 0 &mm->mmap_lock sb_pagefaults mmu_notifier_invalidate_range_start &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 &p->lock &of->mutex kn->active#5 remove_cache_srcu rcu_read_lock rcu_node_0 irq_context: 0 &p->lock &of->mutex kn->active#5 remove_cache_srcu rcu_read_lock &rq->__lock irq_context: 0 &p->lock &of->mutex kn->active#5 remove_cache_srcu rcu_read_lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 &f->f_pos_lock &type->i_mutex_dir_key#3 sb_writers#4 remove_cache_srcu rcu_read_lock rcu_node_0 irq_context: 0 &f->f_pos_lock &type->i_mutex_dir_key#3 sb_writers#4 remove_cache_srcu rcu_read_lock &rq->__lock irq_context: 0 &f->f_pos_lock &type->i_mutex_dir_key#3 sb_writers#4 remove_cache_srcu rcu_read_lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 sb_writers#4 &type->i_mutex_dir_key#3/1 tomoyo_ss &base->lock irq_context: 0 sb_writers#4 &type->i_mutex_dir_key#3/1 tomoyo_ss &base->lock &obj_hash[i].lock irq_context: 0 sb_writers &type->i_mutex_dir_key#2 fs_reclaim irq_context: 0 sb_writers &type->i_mutex_dir_key#2 fs_reclaim mmu_notifier_invalidate_range_start irq_context: 0 sb_writers &type->i_mutex_dir_key#2 &xa->xa_lock#4 irq_context: 0 sb_writers &type->i_mutex_dir_key#2 &obj_hash[i].lock irq_context: 0 sb_writers &type->i_mutex_dir_key#2 stock_lock irq_context: 0 sb_writers &type->i_mutex_dir_key#2 &c->lock irq_context: 0 sb_writers &type->i_mutex_dir_key#2 &dentry->d_lock irq_context: 0 sb_writers &type->i_mutex_dir_key#2 rcu_read_lock rename_lock.seqcount irq_context: 0 sb_writers &type->i_mutex_dir_key#2 tomoyo_ss irq_context: 0 sb_writers &type->i_mutex_dir_key#2 tomoyo_ss &rq->__lock irq_context: 0 sb_writers &type->i_mutex_dir_key#2 tomoyo_ss &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 sb_writers &type->i_mutex_dir_key#2 tomoyo_ss mmu_notifier_invalidate_range_start irq_context: 0 sb_writers &type->i_mutex_dir_key#2 tomoyo_ss rcu_read_lock mount_lock.seqcount irq_context: 0 sb_writers &type->i_mutex_dir_key#2 tomoyo_ss rcu_read_lock rcu_read_lock rename_lock.seqcount irq_context: 0 sb_writers &type->i_mutex_dir_key#2 tomoyo_ss &obj_hash[i].lock irq_context: 0 sb_writers &type->i_mutex_dir_key#2 &dentry->d_lock &wq#2 irq_context: 0 sb_writers &type->i_mutex_dir_key#2 &sbinfo->stat_lock irq_context: 0 sb_writers &type->i_mutex_dir_key#2 pool_lock#2 irq_context: 0 sb_writers &type->i_mutex_dir_key#2 &xa->xa_lock#4 pool_lock#2 irq_context: 0 sb_writers &type->i_mutex_dir_key#2 mmu_notifier_invalidate_range_start irq_context: 0 sb_writers &type->i_mutex_dir_key#2 &sb->s_type->i_lock_key#5 irq_context: 0 sb_writers &type->i_mutex_dir_key#2 &s->s_inode_list_lock irq_context: 0 sb_writers &type->i_mutex_dir_key#2 tk_core.seq.seqcount irq_context: 0 sb_writers &type->i_mutex_dir_key#2 batched_entropy_u32.lock irq_context: 0 sb_writers &type->i_mutex_dir_key#2 &simple_offset_xa_lock irq_context: 0 sb_writers &type->i_mutex_dir_key#2 &sb->s_type->i_lock_key#5 &dentry->d_lock irq_context: 0 &sb->s_type->i_mutex_key#4 irq_context: 0 &sb->s_type->i_mutex_key#4 &rq->__lock irq_context: 0 &sb->s_type->i_mutex_key#4 &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 &sb->s_type->i_mutex_key#4 &cfs_rq->removed.lock irq_context: 0 &sb->s_type->i_mutex_key#4 &obj_hash[i].lock irq_context: 0 sb_writers &sb->s_type->i_mutex_key#4 &sb->s_type->i_lock_key#5 irq_context: 0 sb_writers &sb->s_type->i_mutex_key#4 fs_reclaim irq_context: 0 sb_writers &sb->s_type->i_mutex_key#4 fs_reclaim &rq->__lock irq_context: 0 sb_writers &sb->s_type->i_mutex_key#4 fs_reclaim &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 sb_writers &sb->s_type->i_mutex_key#4 fs_reclaim mmu_notifier_invalidate_range_start irq_context: 0 sb_writers &sb->s_type->i_mutex_key#4 &____s->seqcount irq_context: 0 sb_writers &sb->s_type->i_mutex_key#4 pool_lock#2 irq_context: 0 sb_writers &sb->s_type->i_mutex_key#4 stock_lock irq_context: 0 sb_writers &sb->s_type->i_mutex_key#4 &xa->xa_lock#9 irq_context: 0 sb_writers &sb->s_type->i_mutex_key#4 lock#4 irq_context: 0 sb_writers &sb->s_type->i_mutex_key#4 &info->lock irq_context: 0 sb_writers &sb->s_type->i_mutex_key#4 &xa->xa_lock#9 stock_lock irq_context: 0 sb_writers &sb->s_type->i_mutex_key#4 &xa->xa_lock#9 &c->lock irq_context: 0 sb_writers &sb->s_type->i_mutex_key#4 &xa->xa_lock#9 pool_lock#2 irq_context: 0 sb_writers &sb->s_type->i_mutex_key#4 lock#4 &lruvec->lru_lock irq_context: 0 sb_writers &sb->s_type->i_mutex_key#4 key#9 irq_context: 0 sb_writers &sb->s_type->i_mutex_key#4 &xa->xa_lock#9 &n->list_lock irq_context: 0 sb_writers &sb->s_type->i_mutex_key#4 &xa->xa_lock#9 &n->list_lock &c->lock irq_context: 0 sb_writers &sb->s_type->i_mutex_key#4 rcu_read_lock rcu_node_0 irq_context: 0 sb_writers &sb->s_type->i_mutex_key#4 rcu_read_lock &rq->__lock irq_context: 0 sb_writers &sb->s_type->i_mutex_key#4 rcu_read_lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 sb_writers &sb->s_type->i_mutex_key#4 &sem->wait_lock irq_context: 0 sb_writers &sb->s_type->i_mutex_key#4 rcu_read_lock &rcu_state.gp_wq irq_context: 0 sb_writers &sb->s_type->i_mutex_key#4 rcu_read_lock &rcu_state.gp_wq &p->pi_lock irq_context: 0 sb_writers &sb->s_type->i_mutex_key#4 rcu_read_lock &rcu_state.gp_wq &p->pi_lock &rq->__lock irq_context: 0 sb_writers &sb->s_type->i_mutex_key#4 rcu_read_lock &rcu_state.gp_wq &p->pi_lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)events drain_vmap_work vmap_purge_lock free_vmap_area_lock &meta->lock irq_context: 0 (wq_completion)events drain_vmap_work vmap_purge_lock free_vmap_area_lock kfence_freelist_lock irq_context: 0 (wq_completion)wg-kex-wg2#9 (work_completion)(&peer->transmit_handshake_work) &peer->endpoint_lock rcu_read_lock_bh &n->list_lock irq_context: 0 (wq_completion)wg-kex-wg2#9 (work_completion)(&peer->transmit_handshake_work) &peer->endpoint_lock rcu_read_lock_bh &n->list_lock &c->lock irq_context: 0 (wq_completion)wg-kex-wg2#9 (work_completion)(&peer->transmit_handshake_work) &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 sb_writers &sb->s_type->i_mutex_key#4 rcu_node_0 irq_context: 0 sb_writers &sb->s_type->i_mutex_key#4 &wb->list_lock irq_context: 0 sb_writers &sb->s_type->i_mutex_key#4 &wb->list_lock &sb->s_type->i_lock_key#5 irq_context: 0 pernet_ops_rwsem ovs_mutex nf_ct_proto_mutex nf_hook_mutex batched_entropy_u8.lock irq_context: 0 pernet_ops_rwsem ovs_mutex nf_ct_proto_mutex nf_hook_mutex kfence_freelist_lock irq_context: 0 (wq_completion)netns net_cleanup_work &p->pi_lock &cfs_rq->removed.lock irq_context: 0 (wq_completion)netns net_cleanup_work &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem ovs_mutex nf_ct_proto_mutex nf_hook_mutex remove_cache_srcu rcu_read_lock rcu_node_0 irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem ovs_mutex nf_ct_proto_mutex nf_hook_mutex remove_cache_srcu rcu_read_lock &rq->__lock irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem ovs_mutex nf_ct_proto_mutex nf_hook_mutex remove_cache_srcu rcu_read_lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 pernet_ops_rwsem rtnl_mutex cpu_hotplug_lock &rq->__lock irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem rtnl_mutex remove_cache_srcu rcu_read_lock rcu_node_0 irq_context: 0 (wq_completion)events (work_completion)(&pwq->unbound_release_work) rcu_node_0 irq_context: 0 (wq_completion)events (work_completion)(&pwq->unbound_release_work) &rcu_state.expedited_wq irq_context: 0 (wq_completion)events (work_completion)(&pwq->unbound_release_work) &rcu_state.expedited_wq &p->pi_lock irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem rtnl_mutex remove_cache_srcu rcu_read_lock &rq->__lock irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem rtnl_mutex remove_cache_srcu rcu_read_lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem rtnl_mutex net_rwsem &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 &mm->mmap_lock &mapping->i_mmap_rwsem rcu_read_lock &cfs_rq->removed.lock irq_context: 0 &mm->mmap_lock &mapping->i_mmap_rwsem rcu_read_lock &obj_hash[i].lock irq_context: 0 pernet_ops_rwsem nf_ct_proto_mutex defrag6_mutex nf_hook_mutex rcu_read_lock rcu_node_0 irq_context: 0 pernet_ops_rwsem nf_ct_proto_mutex defrag6_mutex nf_hook_mutex rcu_read_lock &rq->__lock irq_context: 0 pernet_ops_rwsem nf_ct_proto_mutex defrag6_mutex nf_hook_mutex rcu_read_lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 pernet_ops_rwsem k-sk_lock-AF_RXRPC &obj_hash[i].lock pool_lock irq_context: 0 pernet_ops_rwsem k-sk_lock-AF_RXRPC &rq->__lock irq_context: 0 pernet_ops_rwsem k-sk_lock-AF_RXRPC &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)events (work_completion)(&p->wq) rcu_node_0 irq_context: 0 (wq_completion)events (work_completion)(&(&krcp->monitor_work)->work) rcu_callback rcu_node_0 irq_context: 0 (wq_completion)events (work_completion)(&(&krcp->monitor_work)->work) rcu_callback &rcu_state.expedited_wq irq_context: 0 (wq_completion)events (work_completion)(&(&krcp->monitor_work)->work) rcu_callback &rcu_state.expedited_wq &p->pi_lock irq_context: 0 (wq_completion)events (work_completion)(&(&krcp->monitor_work)->work) rcu_callback &rcu_state.expedited_wq &p->pi_lock &rq->__lock irq_context: 0 (wq_completion)events (work_completion)(&(&krcp->monitor_work)->work) rcu_callback &rcu_state.expedited_wq &p->pi_lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 sb_writers#4 &type->i_mutex_dir_key#3 jbd2_handle &meta_group_info[i]->alloc_sem &bgl->locks[i].lock irq_context: 0 &f->f_pos_lock sb_writers#4 &sb->s_type->i_mutex_key#8 &ei->i_data_sem batched_entropy_u8.lock irq_context: 0 &f->f_pos_lock sb_writers#4 &sb->s_type->i_mutex_key#8 &ei->i_data_sem kfence_freelist_lock irq_context: 0 &f->f_pos_lock sb_writers#4 &sb->s_type->i_mutex_key#8 &ei->i_data_sem &meta->lock irq_context: 0 &sb->s_type->i_mutex_key#9 remove_cache_srcu &rq->__lock irq_context: 0 &sb->s_type->i_mutex_key#9 remove_cache_srcu &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 &sb->s_type->i_mutex_key#9 remove_cache_srcu rcu_read_lock rcu_node_0 irq_context: 0 &sb->s_type->i_mutex_key#9 remove_cache_srcu rcu_read_lock &rq->__lock irq_context: 0 &sb->s_type->i_mutex_key#9 remove_cache_srcu rcu_read_lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 pernet_ops_rwsem nf_ct_proto_mutex defrag6_mutex nf_hook_mutex &____s->seqcount#2 irq_context: 0 pernet_ops_rwsem nf_ct_proto_mutex defrag6_mutex nf_hook_mutex &____s->seqcount irq_context: 0 (wq_completion)bat_events (work_completion)(&(&bat_priv->nc.work)->work) &rcu_state.expedited_wq &p->pi_lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 &sbi->s_writepages_rwsem remove_cache_srcu &rq->__lock irq_context: 0 &sbi->s_writepages_rwsem remove_cache_srcu &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 sb_writers#4 &type->i_mutex_dir_key#3/1 &type->i_mutex_dir_key#3 &rcu_state.expedited_wq irq_context: 0 sb_writers#4 &type->i_mutex_dir_key#3/1 &type->i_mutex_dir_key#3 &rcu_state.expedited_wq &p->pi_lock irq_context: 0 sb_writers &dentry->d_lock irq_context: 0 sb_writers tomoyo_ss irq_context: 0 sb_writers tomoyo_ss mmu_notifier_invalidate_range_start irq_context: 0 sb_writers tomoyo_ss rcu_read_lock mount_lock.seqcount irq_context: 0 sb_writers tomoyo_ss rcu_read_lock rcu_read_lock rename_lock.seqcount irq_context: 0 sb_writers tomoyo_ss &obj_hash[i].lock irq_context: 0 sb_writers tomoyo_ss &c->lock irq_context: 0 sb_writers tomoyo_ss &n->list_lock irq_context: 0 sb_writers tomoyo_ss &n->list_lock &c->lock irq_context: 0 sb_writers &sb->s_type->i_mutex_key#4 &sb->s_type->i_lock_key#5 &xa->xa_lock#9 irq_context: 0 sb_writers &sb->s_type->i_mutex_key#4 lock#5 irq_context: 0 sb_writers &sb->s_type->i_mutex_key#4 &lruvec->lru_lock irq_context: 0 sb_writers &sb->s_type->i_mutex_key#4 rcu_read_lock pool_lock#2 irq_context: 0 sb_writers &sb->s_type->i_mutex_key#4 &sb->s_type->i_lock_key#5 &xa->xa_lock#9 &obj_hash[i].lock irq_context: 0 sb_writers &sb->s_type->i_mutex_key#4 &sb->s_type->i_lock_key#5 &xa->xa_lock#9 pool_lock#2 irq_context: 0 sb_writers &sb->s_type->i_mutex_key#4 &sb->s_type->i_lock_key#5 &xa->xa_lock#9 rcu_read_lock per_cpu_ptr(&cgroup_rstat_cpu_lock, cpu) irq_context: 0 sb_writers &sb->s_type->i_mutex_key#4 rcu_read_lock rcu_read_lock rcu_node_0 irq_context: 0 sb_writers &sb->s_type->i_mutex_key#4 rcu_read_lock rcu_read_lock &rq->__lock irq_context: 0 sb_writers &sb->s_type->i_mutex_key#4 rcu_read_lock rcu_read_lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 pernet_ops_rwsem devices_rwsem rdma_nets_rwsem &device->compat_devs_mutex fs_reclaim mmu_notifier_invalidate_range_start &cfs_rq->removed.lock irq_context: 0 pernet_ops_rwsem devices_rwsem rdma_nets_rwsem &device->compat_devs_mutex fs_reclaim mmu_notifier_invalidate_range_start &obj_hash[i].lock irq_context: 0 sk_lock-AF_INET6 mmu_notifier_invalidate_range_start irq_context: 0 sk_lock-AF_INET6 &sb->s_type->i_lock_key#8 irq_context: 0 sk_lock-AF_INET6 &dir->lock irq_context: 0 sk_lock-AF_INET6 k-sk_lock-AF_INET6/1 irq_context: 0 sk_lock-AF_INET6 k-sk_lock-AF_INET6/1 k-slock-AF_INET6 irq_context: 0 sk_lock-AF_INET6 k-sk_lock-AF_INET6/1 &dir->lock irq_context: 0 sk_lock-AF_INET6 k-sk_lock-AF_INET6/1 fs_reclaim irq_context: 0 sk_lock-AF_INET6 k-sk_lock-AF_INET6/1 fs_reclaim &rq->__lock irq_context: 0 sk_lock-AF_INET6 k-sk_lock-AF_INET6/1 fs_reclaim &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 sk_lock-AF_INET6 k-sk_lock-AF_INET6/1 fs_reclaim mmu_notifier_invalidate_range_start irq_context: 0 sk_lock-AF_INET6 k-slock-AF_INET6 irq_context: 0 sk_lock-AF_INET6 k-clock-AF_INET6 irq_context: 0 sk_lock-AF_INET6 &xa->xa_lock#9 irq_context: 0 sk_lock-AF_INET6 &fsnotify_mark_srcu irq_context: 0 &sb->s_type->i_mutex_key#10 sk_lock-AF_INET6 k-sk_lock-AF_INET6/1 irq_context: 0 &sb->s_type->i_mutex_key#10 sk_lock-AF_INET6 k-sk_lock-AF_INET6/1 k-slock-AF_INET6 irq_context: 0 &sb->s_type->i_mutex_key#10 sk_lock-AF_INET6 k-slock-AF_INET6 irq_context: 0 &sb->s_type->i_mutex_key#10 sk_lock-AF_INET6 k-sk_lock-AF_INET6/1 k-clock-AF_INET6 irq_context: 0 &sb->s_type->i_mutex_key#10 sk_lock-AF_INET6 k-sk_lock-AF_INET6/1 k-slock-AF_INET6 &obj_hash[i].lock irq_context: 0 &sb->s_type->i_mutex_key#10 sk_lock-AF_INET6 k-sk_lock-AF_INET6/1 k-slock-AF_INET6 pool_lock#2 irq_context: 0 &sb->s_type->i_mutex_key#10 sk_lock-AF_INET6 k-sk_lock-AF_INET6/1 k-slock-AF_INET6 krc.lock irq_context: 0 &sb->s_type->i_mutex_key#10 sk_lock-AF_INET6 k-sk_lock-AF_INET6/1 k-slock-AF_INET6 elock-AF_INET6 irq_context: 0 &sb->s_type->i_mutex_key#10 sk_lock-AF_INET6 &dir->lock irq_context: 0 &sb->s_type->i_mutex_key#10 sk_lock-AF_INET6 &____s->seqcount irq_context: 0 &sb->s_type->i_mutex_key#10 sk_lock-AF_INET6 stock_lock irq_context: 0 &sb->s_type->i_mutex_key#10 sk_lock-AF_INET6 &msk->pm.lock irq_context: 0 &sb->s_type->i_mutex_key#10 sk_lock-AF_INET6 elock-AF_INET6 irq_context: 0 fill_pool_map-wait-type-override rcu_read_lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 &sbi->s_writepages_rwsem jbd2_handle &ei->i_data_sem mmu_notifier_invalidate_range_start &rq->__lock irq_context: 0 &sbi->s_writepages_rwsem jbd2_handle &ei->i_data_sem mmu_notifier_invalidate_range_start &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 &mm->mmap_lock &anon_vma->rwsem &mm->page_table_lock quarantine_lock irq_context: 0 sk_lock-AF_INET6 k-sk_lock-AF_INET6/1 &c->lock irq_context: 0 (wq_completion)ext4-rsv-conversion (work_completion)(&ei->i_rsv_conversion_work) jbd2_handle &journal->j_state_lock &journal->j_wait_commit irq_context: 0 (wq_completion)ext4-rsv-conversion (work_completion)(&ei->i_rsv_conversion_work) jbd2_handle &journal->j_state_lock &journal->j_wait_commit &p->pi_lock irq_context: 0 (wq_completion)ext4-rsv-conversion (work_completion)(&ei->i_rsv_conversion_work) jbd2_handle &journal->j_state_lock &journal->j_wait_commit &p->pi_lock &rq->__lock irq_context: 0 (wq_completion)ext4-rsv-conversion (work_completion)(&ei->i_rsv_conversion_work) jbd2_handle &journal->j_state_lock &journal->j_wait_commit &p->pi_lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 sb_writers tomoyo_ss quarantine_lock irq_context: 0 &sb->s_type->i_mutex_key#10 krc.lock irq_context: 0 sb_writers &sb->s_type->i_mutex_key#4 rcu_read_lock &rcu_state.expedited_wq irq_context: 0 sb_writers &sb->s_type->i_mutex_key#4 rcu_read_lock &rcu_state.expedited_wq &p->pi_lock irq_context: 0 sb_writers &sb->s_type->i_mutex_key#4 rcu_read_lock &rcu_state.expedited_wq &p->pi_lock &rq->__lock irq_context: 0 sb_writers &sb->s_type->i_mutex_key#4 rcu_read_lock &rcu_state.expedited_wq &p->pi_lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 sb_writers &sb->s_type->i_mutex_key#4 &pcp->lock &zone->lock irq_context: 0 sb_writers &sb->s_type->i_mutex_key#4 &pcp->lock &zone->lock &____s->seqcount irq_context: 0 &mm->mmap_lock remove_cache_srcu rcu_read_lock rcu_read_lock rcu_node_0 irq_context: 0 &mm->mmap_lock remove_cache_srcu rcu_read_lock rcu_read_lock &rq->__lock irq_context: 0 &mm->mmap_lock remove_cache_srcu rcu_read_lock rcu_read_lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 sb_writers &sb->s_type->i_mutex_key#4 fs_reclaim mmu_notifier_invalidate_range_start &rq->__lock irq_context: 0 sb_writers &sb->s_type->i_mutex_key#4 fs_reclaim mmu_notifier_invalidate_range_start &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 pernet_ops_rwsem rtnl_mutex gdp_mutex &rq->__lock irq_context: 0 pernet_ops_rwsem rtnl_mutex gdp_mutex &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 pernet_ops_rwsem cpu_hotplug_lock &rq->__lock irq_context: 0 pernet_ops_rwsem cpu_hotplug_lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 &sb->s_type->i_mutex_key#10 sk_lock-AF_INET6 k-sk_lock-AF_INET6/1 &rq->__lock irq_context: 0 &sb->s_type->i_mutex_key#10 sk_lock-AF_INET6 k-sk_lock-AF_INET6/1 &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 sb_writers#4 &type->i_mutex_dir_key#3/1 sb_internal jbd2_handle &rq->__lock &cfs_rq->removed.lock irq_context: 0 sb_writers#4 &type->i_mutex_dir_key#3/1 jbd2_handle batched_entropy_u8.lock irq_context: 0 sb_writers#4 &type->i_mutex_dir_key#3/1 jbd2_handle kfence_freelist_lock irq_context: 0 &map->freeze_mutex irq_context: 0 &sbi->s_writepages_rwsem jbd2_handle &obj_hash[i].lock irq_context: 0 &sbi->s_writepages_rwsem jbd2_handle &base->lock irq_context: 0 &sbi->s_writepages_rwsem jbd2_handle &base->lock &obj_hash[i].lock irq_context: 0 sb_writers#4 rcu_read_lock &obj_hash[i].lock pool_lock irq_context: 0 sb_writers#4 &sb->s_type->i_mutex_key#8 &sbi->s_writepages_rwsem mmu_notifier_invalidate_range_start &rq->__lock irq_context: 0 sb_writers#4 &sb->s_type->i_mutex_key#8 &sbi->s_writepages_rwsem mmu_notifier_invalidate_range_start &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)events (work_completion)(&aux->work) quarantine_lock irq_context: 0 &type->i_mutex_dir_key#3 mmu_notifier_invalidate_range_start &rq->__lock irq_context: 0 &type->i_mutex_dir_key#3 mmu_notifier_invalidate_range_start &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 sb_writers#4 &type->i_mutex_dir_key#3 tomoyo_ss rcu_read_lock rename_lock irq_context: 0 &f->f_pos_lock sb_writers#4 &sb->s_type->i_mutex_key#8 jbd2_handle &ei->i_data_sem &lg->lg_mutex mmu_notifier_invalidate_range_start irq_context: 0 (wq_completion)events_unbound (work_completion)(&sub_info->work) &rq->__lock &cfs_rq->removed.lock irq_context: softirq &fq->mq_flush_lock fill_pool_map-wait-type-override pool_lock#2 irq_context: softirq &fq->mq_flush_lock fill_pool_map-wait-type-override &c->lock irq_context: softirq &fq->mq_flush_lock fill_pool_map-wait-type-override &n->list_lock irq_context: softirq &fq->mq_flush_lock fill_pool_map-wait-type-override &n->list_lock &c->lock irq_context: softirq &fq->mq_flush_lock fill_pool_map-wait-type-override pool_lock irq_context: 0 sb_writers#4 sb_writers#4 jbd2_handle mmu_notifier_invalidate_range_start irq_context: 0 sb_writers#4 sb_writers#4 jbd2_handle &____s->seqcount irq_context: 0 sb_writers#4 sb_writers#4 jbd2_handle pool_lock#2 irq_context: 0 sb_writers#4 sb_writers#4 jbd2_handle rcu_read_lock pool_lock#2 irq_context: 0 sb_writers#4 sb_writers#4 jbd2_handle &obj_hash[i].lock irq_context: 0 &mm->mmap_lock &mapping->i_mmap_rwsem &anon_vma->rwsem fill_pool_map-wait-type-override &n->list_lock irq_context: 0 &mm->mmap_lock &mapping->i_mmap_rwsem &anon_vma->rwsem fill_pool_map-wait-type-override &n->list_lock &c->lock irq_context: 0 sb_writers#4 &sb->s_type->i_mutex_key#8 rcu_read_lock &obj_hash[i].lock pool_lock irq_context: 0 sb_writers#4 &type->i_mutex_dir_key#3/1 jbd2_handle rcu_read_lock pool_lock#2 irq_context: 0 sb_writers#4 &type->i_mutex_dir_key#3 &n->list_lock irq_context: 0 sb_writers#4 &type->i_mutex_dir_key#3 &n->list_lock &c->lock irq_context: 0 sb_writers#4 &type->i_mutex_dir_key#3 &rq->__lock irq_context: 0 sb_writers#4 &type->i_mutex_dir_key#3 &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: softirq &folio_wait_table[i] &p->pi_lock &rq->__lock &cfs_rq->removed.lock irq_context: 0 &f->f_pos_lock sb_writers#4 &sb->s_type->i_mutex_key#8 jbd2_handle &ret->b_state_lock bit_wait_table + i irq_context: 0 sb_writers#4 &sb->s_type->i_mutex_key#8 &sb->s_type->i_lock_key#22 &xa->xa_lock#9 rcu_read_lock per_cpu_ptr(&cgroup_rstat_cpu_lock, cpu) irq_context: 0 sb_writers#4 &sb->s_type->i_mutex_key#8 &obj_hash[i].lock pool_lock irq_context: 0 (wq_completion)ext4-rsv-conversion (work_completion)(&ei->i_rsv_conversion_work) jbd2_handle &p->pi_lock &cfs_rq->removed.lock irq_context: 0 sb_writers#4 &sb->s_type->i_mutex_key#8 rcu_read_lock &rcu_state.gp_wq irq_context: 0 sb_writers#4 &sb->s_type->i_mutex_key#8 rcu_read_lock &rcu_state.gp_wq &p->pi_lock irq_context: 0 sb_writers#4 &sb->s_type->i_mutex_key#8 rcu_read_lock &rcu_state.gp_wq &p->pi_lock &rq->__lock irq_context: 0 sb_writers#4 &sb->s_type->i_mutex_key#8 rcu_read_lock &rcu_state.gp_wq &p->pi_lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 sb_writers#4 mapping.invalidate_lock rcu_read_lock rcu_read_lock rcu_node_0 irq_context: 0 sb_writers#4 mapping.invalidate_lock rcu_read_lock rcu_read_lock &rq->__lock irq_context: 0 sb_writers#4 mapping.invalidate_lock rcu_node_0 irq_context: 0 sb_writers#4 mapping.invalidate_lock &pcp->lock &zone->lock &____s->seqcount irq_context: 0 sb_writers#4 mapping.invalidate_lock rcu_read_lock rcu_read_lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 cb_lock genl_mutex rcu_read_lock &cfs_rq->removed.lock irq_context: 0 cb_lock genl_mutex rcu_read_lock &obj_hash[i].lock irq_context: 0 &sbi->s_writepages_rwsem &meta->lock irq_context: 0 &f->f_pos_lock &type->i_mutex_dir_key#3 sb_writers#4 jbd2_handle &rq->__lock &cfs_rq->removed.lock irq_context: 0 &pipe->mutex/1 sk_lock-AF_INET6 rcu_read_lock &q->lock#2 quarantine_lock irq_context: 0 dup_mmap_sem &mm->mmap_lock &mm->mmap_lock/1 remove_cache_srcu &cfs_rq->removed.lock irq_context: 0 text_mutex text_mutex.wait_lock irq_context: 0 tracepoints_mutex cpu_hotplug_lock jump_label_mutex text_mutex.wait_lock irq_context: 0 tracepoints_mutex cpu_hotplug_lock jump_label_mutex &p->pi_lock irq_context: 0 tracepoints_mutex cpu_hotplug_lock jump_label_mutex &p->pi_lock &rq->__lock irq_context: 0 tracepoints_mutex cpu_hotplug_lock jump_label_mutex &p->pi_lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 sb_writers#4 &type->i_mutex_dir_key#3/1 &type->i_mutex_dir_key#3 jbd2_handle &ret->b_state_lock bit_wait_table + i irq_context: 0 sb_writers#4 sb_internal jbd2_handle &ei->i_data_sem rcu_node_0 irq_context: 0 sb_writers#4 sb_internal jbd2_handle &ei->i_data_sem &cfs_rq->removed.lock irq_context: 0 sk_lock-AF_INET slock-AF_INET rcu_read_lock rcu_read_lock rcu_read_lock_bh dev->qdisc_tx_busylock ?: &qdisc_tx_busylock _xmit_ETHER#2 rcu_read_lock &c->lock irq_context: 0 (wq_completion)rcu_gp (work_completion)(&rew->rew_work) &pool->lock &p->pi_lock &rq->__lock &cfs_rq->removed.lock irq_context: 0 key#24 irq_context: 0 &pipe->mutex/1 sk_lock-AF_INET6 rcu_node_0 irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem rtnl_mutex &rnp->exp_wq[3] irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem rtnl_mutex &rnp->exp_wq[0] irq_context: softirq net/ipv4/devinet.c:474 irq_context: softirq net/ipv4/devinet.c:474 rcu_read_lock &pool->lock irq_context: softirq net/ipv4/devinet.c:474 rcu_read_lock &pool->lock &obj_hash[i].lock irq_context: softirq net/ipv4/devinet.c:474 rcu_read_lock &pool->lock &p->pi_lock irq_context: softirq net/ipv4/devinet.c:474 rcu_read_lock &pool->lock &p->pi_lock &rq->__lock irq_context: softirq net/ipv4/devinet.c:474 rcu_read_lock &pool->lock &p->pi_lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 &mm->mmap_lock sb_pagefaults remove_cache_srcu rcu_node_0 irq_context: 0 &mm->mmap_lock sb_pagefaults remove_cache_srcu &rcu_state.expedited_wq irq_context: 0 &mm->mmap_lock sb_pagefaults remove_cache_srcu &rcu_state.expedited_wq &p->pi_lock irq_context: 0 &mm->mmap_lock sb_pagefaults remove_cache_srcu &rcu_state.expedited_wq &p->pi_lock &rq->__lock irq_context: 0 &mm->mmap_lock sb_pagefaults remove_cache_srcu &rcu_state.expedited_wq &p->pi_lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)events (work_completion)(&p->wq) &rcu_state.expedited_wq irq_context: 0 (wq_completion)events (work_completion)(&p->wq) &rcu_state.expedited_wq &p->pi_lock irq_context: 0 (wq_completion)events (work_completion)(&p->wq) &rcu_state.expedited_wq &p->pi_lock &rq->__lock irq_context: 0 (wq_completion)events (work_completion)(&p->wq) &rcu_state.expedited_wq &p->pi_lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 &sb->s_type->i_mutex_key#10 &c->lock irq_context: 0 &sb->s_type->i_mutex_key#10 sk_lock-AF_INET6 slock-AF_INET6 key#24 irq_context: 0 &sbi->s_writepages_rwsem rcu_read_lock &n->list_lock irq_context: 0 &sbi->s_writepages_rwsem rcu_read_lock &n->list_lock &c->lock irq_context: 0 cb_lock genl_mutex nbd_index_mutex irq_context: 0 cb_lock genl_mutex &nbd->config_lock irq_context: 0 cb_lock genl_mutex &nbd->config_lock mmu_notifier_invalidate_range_start irq_context: 0 cb_lock genl_mutex &nbd->config_lock pool_lock#2 irq_context: 0 cb_lock genl_mutex &nbd->config_lock &bdev->bd_size_lock irq_context: 0 cb_lock genl_mutex &nbd->config_lock &q->queue_lock irq_context: 0 cb_lock genl_mutex &nbd->config_lock &ACCESS_PRIVATE(sdp, lock) irq_context: 0 cb_lock genl_mutex &nbd->config_lock set->srcu irq_context: 0 cb_lock genl_mutex &nbd->config_lock &obj_hash[i].lock irq_context: 0 cb_lock genl_mutex &nbd->config_lock &ACCESS_PRIVATE(ssp->srcu_sup, lock) &ACCESS_PRIVATE(sdp, lock) irq_context: 0 cb_lock genl_mutex &nbd->config_lock &ACCESS_PRIVATE(ssp->srcu_sup, lock) rcu_read_lock &pool->lock irq_context: 0 cb_lock genl_mutex &nbd->config_lock &ACCESS_PRIVATE(ssp->srcu_sup, lock) rcu_read_lock &pool->lock &obj_hash[i].lock irq_context: 0 cb_lock genl_mutex &nbd->config_lock &ACCESS_PRIVATE(ssp->srcu_sup, lock) rcu_read_lock &pool->lock pool_lock#2 irq_context: 0 cb_lock genl_mutex &nbd->config_lock &rq->__lock irq_context: 0 cb_lock genl_mutex &nbd->config_lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 cb_lock genl_mutex &nbd->config_lock &x->wait#2 irq_context: 0 cb_lock genl_mutex &nbd->config_lock set->srcu irq_context: 0 sb_writers#4 &sb->s_type->i_mutex_key#8 mapping.invalidate_lock rcu_read_lock rcu_read_lock &rq->__lock &cfs_rq->removed.lock irq_context: softirq rcu_callback krc.lock irq_context: 0 cb_lock genl_mutex quarantine_lock irq_context: 0 (wq_completion)bat_events (work_completion)(&(&bat_priv->nc.work)->work) &obj_hash[i].lock pool_lock irq_context: 0 sb_writers#4 &sb->s_type->i_mutex_key#8 mapping.invalidate_lock jbd2_handle &lock->wait_lock irq_context: 0 sb_writers#4 &sb->s_type->i_mutex_key#8 mapping.invalidate_lock jbd2_handle &p->pi_lock irq_context: 0 sb_writers#4 &sb->s_type->i_mutex_key#8 mapping.invalidate_lock jbd2_handle &p->pi_lock &rq->__lock irq_context: 0 sb_writers#4 &sb->s_type->i_mutex_key#8 mapping.invalidate_lock jbd2_handle &p->pi_lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 sk_lock-AF_NETLINK rcu_read_lock &cfs_rq->removed.lock irq_context: 0 sk_lock-AF_NETLINK rcu_read_lock &obj_hash[i].lock irq_context: 0 rtnl_mutex _xmit_IPGRE console_owner_lock irq_context: 0 rtnl_mutex _xmit_IPGRE console_owner irq_context: 0 rtnl_mutex &app->lock#2 &c->lock irq_context: 0 &sb->s_type->i_mutex_key#8 &sbi->s_writepages_rwsem jbd2_handle &mapping->i_mmap_rwsem mmu_notifier_invalidate_range_start &rq->__lock irq_context: 0 &sb->s_type->i_mutex_key#8 &sbi->s_writepages_rwsem jbd2_handle &mapping->i_mmap_rwsem mmu_notifier_invalidate_range_start &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 rtnl_mutex &app->lock &c->lock irq_context: 0 rtnl_mutex &app->lock &obj_hash[i].lock irq_context: 0 rtnl_mutex &app->lock &list->lock#11 irq_context: 0 rtnl_mutex &app->lock#2 &obj_hash[i].lock irq_context: 0 sk_lock-AF_INET sctp_assocs_id_lock irq_context: 0 &sbi->s_writepages_rwsem jbd2_handle &journal->j_wait_reserved irq_context: 0 &f->f_pos_lock sb_writers#4 &sb->s_type->i_mutex_key#8 fs_reclaim mmu_notifier_invalidate_range_start &cfs_rq->removed.lock irq_context: 0 &f->f_pos_lock sb_writers#4 &sb->s_type->i_mutex_key#8 fs_reclaim mmu_notifier_invalidate_range_start &obj_hash[i].lock irq_context: 0 nf_sockopt_mutex pool_lock#2 irq_context: 0 sb_writers#4 &iint->mutex &lock->wait_lock irq_context: 0 sb_writers#4 jbd2_handle &journal->j_wait_updates &p->pi_lock irq_context: 0 sb_writers#4 jbd2_handle &journal->j_wait_updates &p->pi_lock &rq->__lock irq_context: 0 sb_writers#4 jbd2_handle &journal->j_wait_updates &p->pi_lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 sb_writers#4 &type->i_mutex_dir_key#3/1 tomoyo_ss rcu_read_lock rcu_read_lock pgd_lock irq_context: 0 sb_writers#4 &type->i_mutex_dir_key#3/1 tomoyo_ss rcu_read_lock rcu_read_lock stock_lock irq_context: 0 sb_writers#4 &type->i_mutex_dir_key#3/1 tomoyo_ss rcu_read_lock rcu_read_lock &obj_hash[i].lock irq_context: 0 sb_writers#4 &type->i_mutex_dir_key#3/1 tomoyo_ss rcu_read_lock rcu_read_lock key irq_context: 0 sb_writers#4 &type->i_mutex_dir_key#3/1 tomoyo_ss rcu_read_lock rcu_read_lock pcpu_lock irq_context: 0 sb_writers#4 &type->i_mutex_dir_key#3/1 tomoyo_ss rcu_read_lock rcu_read_lock percpu_counters_lock irq_context: 0 sb_writers#4 &type->i_mutex_dir_key#3/1 tomoyo_ss rcu_read_lock rcu_read_lock pcpu_lock stock_lock irq_context: 0 &pipe->mutex/1 sk_lock-AF_INET6 rcu_read_lock &c->lock irq_context: 0 sb_writers#4 &type->i_mutex_dir_key#3/1 jbd2_handle &ei->i_data_sem pgd_lock irq_context: 0 sb_writers#4 &type->i_mutex_dir_key#3/1 jbd2_handle &ei->i_data_sem key irq_context: 0 sb_writers#4 &type->i_mutex_dir_key#3/1 jbd2_handle &ei->i_data_sem pcpu_lock irq_context: 0 sb_writers#4 &type->i_mutex_dir_key#3/1 jbd2_handle &ei->i_data_sem percpu_counters_lock irq_context: 0 sb_writers#3 &sb->s_type->i_mutex_key#9 fs_reclaim mmu_notifier_invalidate_range_start &rq->__lock irq_context: 0 sb_writers#3 &sb->s_type->i_mutex_key#9 fs_reclaim mmu_notifier_invalidate_range_start &rq->__lock &cfs_rq->removed.lock irq_context: 0 sb_writers#3 &sb->s_type->i_mutex_key#9 fs_reclaim mmu_notifier_invalidate_range_start &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 &group->mark_mutex lock &group->inotify_data.idr_lock &n->list_lock irq_context: 0 &group->mark_mutex lock &group->inotify_data.idr_lock &n->list_lock &c->lock irq_context: 0 &disk->open_mutex rcu_node_0 irq_context: 0 &f->f_pos_lock &type->i_mutex_dir_key#3 remove_cache_srcu rcu_read_lock rcu_read_lock rcu_node_0 irq_context: 0 &f->f_pos_lock &type->i_mutex_dir_key#3 remove_cache_srcu rcu_read_lock rcu_read_lock &rq->__lock irq_context: 0 &f->f_pos_lock &type->i_mutex_dir_key#3 remove_cache_srcu rcu_read_lock rcu_read_lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)wg-kex-wg2#9 (work_completion)(&peer->transmit_handshake_work) kfence_freelist_lock irq_context: 0 (wq_completion)wg-kex-wg2#9 (work_completion)(&peer->transmit_handshake_work) &peer->endpoint_lock rcu_read_lock_bh &meta->lock irq_context: 0 (wq_completion)wg-kex-wg2#9 (work_completion)(&peer->transmit_handshake_work) &peer->endpoint_lock rcu_read_lock_bh kfence_freelist_lock irq_context: 0 &iint->mutex ima_extend_list_mutex ima_extend_list_mutex.wait_lock irq_context: 0 &iint->mutex ima_extend_list_mutex.wait_lock irq_context: 0 &iint->mutex &p->pi_lock irq_context: 0 &iint->mutex &p->pi_lock &rq->__lock irq_context: 0 &iint->mutex &p->pi_lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 uts_sem &rq->__lock irq_context: 0 uts_sem &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: softirq rcu_read_lock rcu_read_lock slock-AF_INET/1 rcu_read_lock rcu_read_lock rcu_read_lock_bh dev->qdisc_tx_busylock ?: &qdisc_tx_busylock _xmit_ETHER#2 rcu_read_lock init_task.mems_allowed_seq.seqcount irq_context: 0 sk_lock-AF_PACKET &po->bind_lock rcu_read_lock &n->list_lock irq_context: 0 sk_lock-AF_PACKET &po->bind_lock rcu_read_lock &n->list_lock &c->lock irq_context: 0 &xs->mutex irq_context: 0 &xs->mutex fs_reclaim irq_context: 0 &xs->mutex fs_reclaim mmu_notifier_invalidate_range_start irq_context: 0 &xs->mutex &c->lock irq_context: 0 &xs->mutex pool_lock#2 irq_context: 0 &xs->mutex free_vmap_area_lock irq_context: 0 &xs->mutex &rq->__lock irq_context: 0 &xs->mutex vmap_area_lock irq_context: 0 &xs->mutex &____s->seqcount irq_context: 0 &xs->mutex init_mm.page_table_lock irq_context: 0 &xs->mutex rcu_read_lock rcu_node_0 irq_context: 0 &xs->mutex rcu_read_lock &rq->__lock irq_context: 0 &xs->mutex rcu_read_lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 &xs->mutex &pcp->lock &zone->lock irq_context: 0 &xs->mutex fs_reclaim mmu_notifier_invalidate_range_start &rq->__lock irq_context: 0 &xs->mutex fs_reclaim mmu_notifier_invalidate_range_start &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 &xs->mutex rcu_read_lock &rcu_state.gp_wq irq_context: 0 &xs->mutex rcu_read_lock &rcu_state.gp_wq &p->pi_lock irq_context: 0 &xs->mutex fs_reclaim &rq->__lock irq_context: 0 &xs->mutex &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 &xs->mutex &pcp->lock &zone->lock &____s->seqcount irq_context: 0 &xs->mutex &lock->wait_lock irq_context: 0 &xs->mutex fs_reclaim &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 sb_writers#4 &sb->s_type->i_mutex_key#8 mapping.invalidate_lock &rq->__lock &cfs_rq->removed.lock irq_context: 0 &sbi->s_writepages_rwsem jbd2_handle &ei->i_data_sem rcu_node_0 irq_context: 0 (wq_completion)wg-crypt-wg2#3 &rq->__lock irq_context: 0 (wq_completion)wg-crypt-wg2#3 &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)wg-crypt-wg2#3 (work_completion)(&({ do { const void *__vpp_verify = (typeof((worker) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __asm__ ("" : "=r"(__ptr) : "0"((typeof(*((worker))) *)((worker)))); (typeof((typeof(*((worker))) *)((worker)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); })->work) &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 &sb->s_type->i_mutex_key#10 vmap_area_lock irq_context: 0 &sb->s_type->i_mutex_key#10 purge_vmap_area_lock irq_context: 0 &sb->s_type->i_mutex_key#10 purge_vmap_area_lock &obj_hash[i].lock irq_context: 0 &sb->s_type->i_mutex_key#10 purge_vmap_area_lock pool_lock#2 irq_context: 0 sb_writers#4 &ei->xattr_sem &rq->__lock irq_context: 0 sb_writers#4 &ei->xattr_sem &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 &xs->mutex &obj_hash[i].lock irq_context: 0 &xs->mutex purge_vmap_area_lock irq_context: 0 &xs->mutex rcu_read_lock &pool->lock irq_context: 0 &xs->mutex rcu_read_lock &pool->lock &obj_hash[i].lock irq_context: 0 &xs->mutex rcu_read_lock &pool->lock &p->pi_lock irq_context: 0 &xs->mutex rcu_read_lock &pool->lock &p->pi_lock &rq->__lock irq_context: 0 &xs->mutex rcu_read_lock &pool->lock &p->pi_lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 &xs->mutex rcu_read_lock pool_lock#2 irq_context: 0 &xs->mutex rcu_read_lock &rcu_state.gp_wq &p->pi_lock &rq->__lock irq_context: 0 &xs->mutex rcu_read_lock &rcu_state.gp_wq &p->pi_lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 &xs->mutex rcu_read_lock &rcu_state.expedited_wq irq_context: 0 &xs->mutex rcu_read_lock &rcu_state.expedited_wq &p->pi_lock irq_context: 0 &xs->mutex rcu_read_lock &rcu_state.expedited_wq &p->pi_lock &rq->__lock irq_context: 0 &xs->mutex rcu_read_lock &rcu_state.expedited_wq &p->pi_lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 &xs->mutex (console_sem).lock irq_context: 0 &xs->mutex console_lock console_srcu console_owner_lock irq_context: 0 &xs->mutex console_lock console_srcu console_owner irq_context: 0 &xs->mutex console_lock console_srcu console_owner &port_lock_key irq_context: 0 &pipe->mutex/1 sk_lock-AF_INET6 rcu_read_lock rcu_read_lock rcu_read_lock rcu_read_lock &pool->lock irq_context: 0 &pipe->mutex/1 sk_lock-AF_INET6 rcu_read_lock rcu_read_lock rcu_read_lock rcu_read_lock &pool->lock &obj_hash[i].lock irq_context: 0 &pipe->mutex/1 sk_lock-AF_INET6 rcu_read_lock rcu_read_lock rcu_read_lock rcu_read_lock &pool->lock pool_lock#2 irq_context: 0 &pipe->mutex/1 sk_lock-AF_INET6 rcu_read_lock rcu_read_lock rcu_read_lock rcu_read_lock &pool->lock &p->pi_lock irq_context: 0 &pipe->mutex/1 sk_lock-AF_INET6 rcu_read_lock rcu_read_lock rcu_read_lock rcu_read_lock &pool->lock &p->pi_lock &rq->__lock irq_context: 0 &pipe->mutex/1 sk_lock-AF_INET6 rcu_read_lock rcu_read_lock rcu_read_lock rcu_read_lock &pool->lock &p->pi_lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 &xs->mutex console_lock console_srcu console_owner console_owner_lock irq_context: 0 &xs->mutex rcu_read_lock (console_sem).lock irq_context: 0 &xs->mutex rcu_read_lock kernfs_pr_cont_lock irq_context: 0 &xs->mutex rcu_read_lock kernfs_pr_cont_lock kernfs_rename_lock irq_context: 0 &xs->mutex rcu_read_lock kernfs_pr_cont_lock (console_sem).lock irq_context: 0 &xs->mutex &cfs_rq->removed.lock irq_context: 0 &sb->s_type->i_mutex_key#10 &pcp->lock &zone->lock &____s->seqcount irq_context: 0 &sb->s_type->i_mutex_key#10 rcu_read_lock &rcu_state.expedited_wq irq_context: 0 &sb->s_type->i_mutex_key#10 rcu_read_lock &rcu_state.expedited_wq &p->pi_lock irq_context: 0 &sb->s_type->i_mutex_key#10 rcu_read_lock &rcu_state.expedited_wq &p->pi_lock &rq->__lock irq_context: 0 &sb->s_type->i_mutex_key#10 rcu_read_lock &rcu_state.expedited_wq &p->pi_lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 &xs->mutex rcu_node_0 irq_context: 0 &xs->mutex &rcu_state.expedited_wq irq_context: 0 &xs->mutex &rcu_state.expedited_wq &p->pi_lock irq_context: 0 &xs->mutex &rcu_state.expedited_wq &p->pi_lock &rq->__lock irq_context: 0 &xs->mutex &rcu_state.expedited_wq &p->pi_lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)events_unbound (work_completion)(&sub_info->work) batched_entropy_u8.lock crngs.lock irq_context: 0 &xs->mutex &zone->lock irq_context: 0 (wq_completion)wg-kex-wg0#7 (work_completion)(&peer->transmit_handshake_work) &wg->static_identity.lock &handshake->lock &rq->__lock irq_context: 0 &xs->mutex &n->list_lock irq_context: 0 &xs->mutex &n->list_lock &c->lock irq_context: 0 &xs->mutex purge_vmap_area_lock &obj_hash[i].lock irq_context: 0 &xs->mutex purge_vmap_area_lock pool_lock#2 irq_context: 0 &sb->s_type->i_mutex_key#10 fill_pool_map-wait-type-override batched_entropy_u8.lock irq_context: 0 &sb->s_type->i_mutex_key#10 fill_pool_map-wait-type-override kfence_freelist_lock irq_context: 0 &xt[i].mutex remove_cache_srcu &cfs_rq->removed.lock irq_context: 0 sk_lock-AF_PACKET stock_lock irq_context: 0 sk_lock-AF_PACKET rcu_read_lock &pool->lock irq_context: 0 sk_lock-AF_PACKET rcu_read_lock &pool->lock &obj_hash[i].lock irq_context: 0 sk_lock-AF_PACKET rcu_read_lock &pool->lock pool_lock#2 irq_context: 0 sk_lock-AF_PACKET rcu_read_lock &pool->lock &p->pi_lock irq_context: 0 sk_lock-AF_PACKET rcu_read_lock &pool->lock &p->pi_lock &rq->__lock irq_context: 0 sk_lock-AF_PACKET rcu_read_lock &pool->lock &p->pi_lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 sock_diag_mutex sock_diag_table_mutex nlk_cb_mutex-SOCK_DIAG irq_context: 0 sock_diag_mutex sock_diag_table_mutex nlk_cb_mutex-SOCK_DIAG fs_reclaim irq_context: 0 sock_diag_mutex sock_diag_table_mutex nlk_cb_mutex-SOCK_DIAG fs_reclaim mmu_notifier_invalidate_range_start irq_context: 0 sock_diag_mutex sock_diag_table_mutex nlk_cb_mutex-SOCK_DIAG fs_reclaim mmu_notifier_invalidate_range_start &rq->__lock irq_context: 0 sock_diag_mutex sock_diag_table_mutex nlk_cb_mutex-SOCK_DIAG fs_reclaim mmu_notifier_invalidate_range_start &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 sock_diag_mutex sock_diag_table_mutex nlk_cb_mutex-SOCK_DIAG pool_lock#2 irq_context: 0 sock_diag_mutex sock_diag_table_mutex nlk_cb_mutex-SOCK_DIAG &c->lock irq_context: 0 sock_diag_mutex sock_diag_table_mutex nlk_cb_mutex-SOCK_DIAG &n->list_lock irq_context: 0 sock_diag_mutex sock_diag_table_mutex nlk_cb_mutex-SOCK_DIAG &n->list_lock &c->lock irq_context: 0 sock_diag_mutex sock_diag_table_mutex nlk_cb_mutex-SOCK_DIAG &net->packet.sklist_lock irq_context: 0 sock_diag_mutex sock_diag_table_mutex nlk_cb_mutex-SOCK_DIAG &net->packet.sklist_lock clock-AF_PACKET irq_context: 0 sock_diag_mutex sock_diag_table_mutex nlk_cb_mutex-SOCK_DIAG &net->packet.sklist_lock rtnl_mutex irq_context: 0 sock_diag_mutex sock_diag_table_mutex nlk_cb_mutex-SOCK_DIAG &net->packet.sklist_lock rtnl_mutex &rq->__lock irq_context: 0 sock_diag_mutex sock_diag_table_mutex nlk_cb_mutex-SOCK_DIAG &net->packet.sklist_lock rtnl_mutex &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 sock_diag_mutex sock_diag_table_mutex nlk_cb_mutex-SOCK_DIAG &net->packet.sklist_lock &po->pg_vec_lock irq_context: 0 sock_diag_mutex sock_diag_table_mutex nlk_cb_mutex-SOCK_DIAG &net->packet.sklist_lock fanout_mutex irq_context: 0 sock_diag_mutex sock_diag_table_mutex rcu_read_lock &c->lock irq_context: 0 sock_diag_mutex sock_diag_table_mutex rcu_read_lock pool_lock#2 irq_context: 0 sock_diag_mutex sock_diag_table_mutex rcu_read_lock rcu_read_lock_bh rcu_read_lock pool_lock#2 irq_context: 0 sock_diag_mutex sock_diag_table_mutex rcu_read_lock rcu_read_lock_bh rcu_read_lock tk_core.seq.seqcount irq_context: 0 sock_diag_mutex sock_diag_table_mutex rcu_read_lock rcu_read_lock_bh rcu_read_lock rlock-AF_PACKET irq_context: 0 sock_diag_mutex sock_diag_table_mutex rcu_read_lock rcu_read_lock_bh &obj_hash[i].lock irq_context: 0 sock_diag_mutex sock_diag_table_mutex rcu_read_lock rcu_read_lock_bh pool_lock#2 irq_context: 0 sock_diag_mutex sock_diag_table_mutex rlock-AF_NETLINK irq_context: 0 sock_diag_mutex rcu_read_lock rcu_read_lock_bh rcu_read_lock pool_lock#2 irq_context: 0 sock_diag_mutex rcu_read_lock rcu_read_lock_bh rcu_read_lock tk_core.seq.seqcount irq_context: 0 sock_diag_mutex rcu_read_lock rcu_read_lock_bh rcu_read_lock rlock-AF_PACKET irq_context: 0 sock_diag_mutex &____s->seqcount#2 irq_context: 0 sock_diag_mutex &____s->seqcount irq_context: 0 sb_writers#4 &type->i_mutex_dir_key#3/1 sb_internal jbd2_handle &ei->i_data_sem &base->lock irq_context: 0 sb_writers#4 &type->i_mutex_dir_key#3/1 sb_internal jbd2_handle &ei->i_data_sem &base->lock &obj_hash[i].lock irq_context: 0 &pipe->mutex/1 sk_lock-AF_INET6 rcu_read_lock rcu_read_lock_bh rcu_read_lock pool_lock#2 irq_context: 0 &pipe->mutex/1 sk_lock-AF_INET6 rcu_read_lock rcu_read_lock_bh rcu_read_lock tk_core.seq.seqcount irq_context: 0 &pipe->mutex/1 sk_lock-AF_INET6 rcu_read_lock rcu_read_lock_bh rcu_read_lock &obj_hash[i].lock irq_context: softirq rcu_read_lock rcu_read_lock k-slock-AF_INET6 rcu_read_lock rcu_read_lock rcu_read_lock_bh rcu_read_lock &____s->seqcount irq_context: 0 cb_lock &rdev->wiphy.mtx rcu_read_lock rcu_read_lock_bh rcu_read_lock &c->lock irq_context: 0 cb_lock &rdev->wiphy.mtx rcu_read_lock rcu_read_lock_bh rcu_read_lock pool_lock#2 irq_context: 0 cb_lock &rdev->wiphy.mtx rcu_read_lock rcu_read_lock_bh rcu_read_lock tk_core.seq.seqcount irq_context: 0 cb_lock &rdev->wiphy.mtx rcu_read_lock rcu_read_lock_bh rcu_read_lock &obj_hash[i].lock irq_context: 0 &f->f_pos_lock &type->i_mutex_dir_key#3 &cfs_rq->removed.lock irq_context: 0 &iint->mutex mmu_notifier_invalidate_range_start &rq->__lock irq_context: 0 &iint->mutex mmu_notifier_invalidate_range_start &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 &f->f_pos_lock &type->i_mutex_dir_key#3 remove_cache_srcu &base->lock irq_context: 0 &f->f_pos_lock &type->i_mutex_dir_key#3 remove_cache_srcu &base->lock &obj_hash[i].lock irq_context: 0 &mm->mmap_lock remove_cache_srcu &meta->lock irq_context: 0 &mm->mmap_lock remove_cache_srcu kfence_freelist_lock irq_context: 0 (wq_completion)wg-crypt-wg1 (work_completion)(&peer->transmit_packet_work) &peer->endpoint_lock rcu_read_lock_bh rcu_read_lock rcu_read_lock_bh rcu_read_lock &c->lock irq_context: 0 sb_writers#4 &type->i_mutex_dir_key#3 remove_cache_srcu &rq->__lock irq_context: 0 sb_writers#4 &type->i_mutex_dir_key#3 remove_cache_srcu rcu_read_lock rcu_node_0 irq_context: 0 sb_writers#4 &type->i_mutex_dir_key#3 remove_cache_srcu rcu_read_lock &rq->__lock irq_context: 0 sb_writers#4 &type->i_mutex_dir_key#3 remove_cache_srcu rcu_read_lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 sb_writers#4 &sb->s_type->i_mutex_key#8 &sbi->s_writepages_rwsem rcu_state.exp_mutex rcu_read_lock &rq->__lock irq_context: 0 sb_writers#4 &sb->s_type->i_mutex_key#8 &sbi->s_writepages_rwsem rcu_state.exp_mutex rcu_read_lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 &pipe->mutex/1 &sighand->siglock stock_lock irq_context: 0 &pipe->mutex/1 &sighand->siglock pool_lock#2 irq_context: 0 (wq_completion)wg-crypt-wg0#5 (work_completion)(&peer->transmit_packet_work) &peer->endpoint_lock rcu_read_lock_bh rcu_read_lock rcu_read_lock_bh rcu_read_lock &c->lock irq_context: 0 &pipe->mutex/1 &sighand->siglock &p->pi_lock irq_context: 0 &pipe->mutex/1 &sighand->siglock &p->pi_lock &rq->__lock irq_context: 0 &pipe->mutex/1 &sighand->siglock &p->pi_lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 sk_lock-AF_INET6 batched_entropy_u16.lock crngs.lock irq_context: 0 &xs->mutex rcu_read_lock &cfs_rq->removed.lock irq_context: 0 &xs->mutex rcu_read_lock &obj_hash[i].lock irq_context: 0 sb_writers#4 &type->i_mutex_dir_key#3 remove_cache_srcu &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)events (work_completion)(&(&nsim_dev->trap_data->trap_report_dw)->work) &devlink->lock_key#6 &nsim_trap_data->trap_lock crngs.lock base_crng.lock irq_context: softirq (&app->join_timer)#2 batched_entropy_u32.lock crngs.lock base_crng.lock irq_context: 0 cb_lock &rdev->wiphy.mtx fs_reclaim &rq->__lock irq_context: 0 cb_lock &rdev->wiphy.mtx fs_reclaim &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 cb_lock &rdev->wiphy.mtx rcu_node_0 irq_context: 0 &pipe->mutex/1 &mm->mmap_lock &cfs_rq->removed.lock irq_context: 0 &pipe->mutex/1 &mm->mmap_lock &obj_hash[i].lock irq_context: 0 &pipe->mutex/1 &mm->mmap_lock pool_lock#2 irq_context: 0 &mm->mmap_lock fs_reclaim mmu_notifier_invalidate_range_start &cfs_rq->removed.lock irq_context: 0 &mm->mmap_lock fs_reclaim mmu_notifier_invalidate_range_start &obj_hash[i].lock irq_context: 0 vlan_ioctl_mutex rtnl_mutex rcu_read_lock &pool->lock/1 &p->pi_lock &cfs_rq->removed.lock irq_context: 0 vlan_ioctl_mutex rtnl_mutex sysctl_lock &obj_hash[i].lock pool_lock irq_context: 0 vlan_ioctl_mutex rtnl_mutex remove_cache_srcu &rq->__lock irq_context: 0 vlan_ioctl_mutex rtnl_mutex remove_cache_srcu &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 &xs->mutex fs_reclaim &cfs_rq->removed.lock irq_context: 0 &xs->mutex fs_reclaim &obj_hash[i].lock irq_context: 0 &xs->mutex fs_reclaim rcu_node_0 irq_context: 0 &xs->mutex fs_reclaim &rcu_state.expedited_wq irq_context: 0 &xs->mutex fs_reclaim &rcu_state.expedited_wq &p->pi_lock irq_context: 0 &xs->mutex fs_reclaim &rcu_state.expedited_wq &p->pi_lock &rq->__lock irq_context: 0 &xs->mutex fs_reclaim &rcu_state.expedited_wq &p->pi_lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 &xs->mutex fs_reclaim &rq->__lock &cfs_rq->removed.lock irq_context: 0 vlan_ioctl_mutex rtnl_mutex &ul->lock#2 pool_lock#2 irq_context: 0 vlan_ioctl_mutex rtnl_mutex &ul->lock#2 &dir->lock#2 irq_context: 0 vlan_ioctl_mutex rtnl_mutex &ul->lock#2 &c->lock irq_context: 0 vlan_ioctl_mutex rtnl_mutex &root->kernfs_rwsem &rq->__lock irq_context: 0 vlan_ioctl_mutex rtnl_mutex &root->kernfs_rwsem &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 &sb->s_type->i_mutex_key#10 rcu_read_lock &pool->lock &p->pi_lock irq_context: 0 &sb->s_type->i_mutex_key#10 rcu_read_lock &pool->lock &p->pi_lock &rq->__lock irq_context: 0 &sb->s_type->i_mutex_key#10 rcu_read_lock &pool->lock &p->pi_lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 sk_lock-AF_INET fill_pool_map-wait-type-override &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 &sb->s_type->i_mutex_key#10 &net->xdp.lock &rq->__lock irq_context: 0 &sb->s_type->i_mutex_key#10 &net->xdp.lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 &pipe->mutex/1 sk_lock-AF_INET6 rcu_read_lock rcu_read_lock &ei->socket.wq.wait irq_context: 0 &sbi->s_writepages_rwsem jbd2_handle &ei->i_data_sem &ei->i_es_lock &sbi->s_es_lock irq_context: 0 &sbi->s_writepages_rwsem jbd2_handle &ei->i_data_sem &ei->i_prealloc_lock irq_context: 0 &sbi->s_writepages_rwsem jbd2_handle &ei->i_data_sem &ei->i_prealloc_lock &pa->pa_lock#2 irq_context: 0 lock#4 rcu_read_lock pool_lock#2 irq_context: 0 lock#4 &obj_hash[i].lock irq_context: 0 &sb->s_type->i_lock_key#22 &xa->xa_lock#9 irq_context: 0 &sb->s_type->i_lock_key#22 &xa->xa_lock#9 &obj_hash[i].lock irq_context: 0 &sb->s_type->i_lock_key#22 &xa->xa_lock#9 pool_lock#2 irq_context: 0 &ei->i_es_lock &sbi->s_es_lock irq_context: 0 &ei->i_es_lock &obj_hash[i].lock irq_context: 0 &ei->i_es_lock pool_lock#2 irq_context: 0 cgroup_threadgroup_rwsem rcu_read_lock &rcu_state.gp_wq irq_context: 0 cgroup_threadgroup_rwsem rcu_read_lock &rcu_state.gp_wq &p->pi_lock irq_context: 0 cgroup_threadgroup_rwsem rcu_read_lock &rcu_state.gp_wq &p->pi_lock &rq->__lock irq_context: 0 cgroup_threadgroup_rwsem rcu_read_lock &rcu_state.gp_wq &p->pi_lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 sb_writers#4 &type->i_mutex_dir_key#3 mmu_notifier_invalidate_range_start &rq->__lock irq_context: 0 sb_writers#4 &type->i_mutex_dir_key#3 mmu_notifier_invalidate_range_start &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 cb_lock &rdev->wiphy.mtx &local->mtx rcu_node_0 irq_context: 0 cb_lock &rdev->wiphy.mtx &local->mtx &local->ack_status_lock batched_entropy_u8.lock irq_context: 0 cb_lock &rdev->wiphy.mtx &local->mtx &local->ack_status_lock kfence_freelist_lock irq_context: 0 &sig->cred_guard_mutex &sig->exec_update_lock rcu_read_lock &____s->seqcount irq_context: 0 &pipe->mutex/1 sk_lock-AF_INET6 slock-AF_INET6 &sk->sk_lock.wq irq_context: 0 slock-AF_INET6 &sk->sk_lock.wq &p->pi_lock &cfs_rq->removed.lock irq_context: 0 (wq_completion)events (work_completion)(&p->wq) quarantine_lock irq_context: 0 &sbi->s_writepages_rwsem rcu_read_lock rcu_read_lock rcu_read_lock rcu_node_0 irq_context: 0 &sbi->s_writepages_rwsem rcu_read_lock rcu_read_lock rcu_read_lock &rq->__lock irq_context: 0 &sbi->s_writepages_rwsem rcu_read_lock rcu_read_lock rcu_read_lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 lock#3 &rq->__lock &cfs_rq->removed.lock irq_context: 0 sb_writers#4 &type->i_mutex_dir_key#3/1 quarantine_lock irq_context: 0 sb_writers#7 irq_context: 0 &xs->mutex rcu_read_lock &rq->__lock &cfs_rq->removed.lock irq_context: 0 (wq_completion)events_unbound (work_completion)(&sub_info->work) lock pidmap_lock &n->list_lock irq_context: 0 (wq_completion)events_unbound (work_completion)(&sub_info->work) lock pidmap_lock &n->list_lock &c->lock irq_context: 0 &mm->mmap_lock sb_pagefaults jbd2_handle &cfs_rq->removed.lock irq_context: 0 sb_writers#3 oom_adj_mutex oom_adj_mutex.wait_lock irq_context: 0 sb_writers#3 oom_adj_mutex.wait_lock irq_context: 0 sb_writers#4 &type->i_mutex_dir_key#3/1 &type->i_mutex_dir_key#3 mmu_notifier_invalidate_range_start &rq->__lock irq_context: 0 sb_writers#4 &type->i_mutex_dir_key#3/1 &type->i_mutex_dir_key#3 mmu_notifier_invalidate_range_start &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 sk_lock-AF_INET6 rcu_read_lock rcu_read_lock &obj_hash[i].lock irq_context: 0 sk_lock-AF_INET6 rcu_read_lock rcu_read_lock &dir->lock#2 irq_context: 0 sk_lock-AF_INET6 rcu_read_lock rcu_read_lock &tbl->lock irq_context: 0 sk_lock-AF_INET6 rcu_read_lock rcu_read_lock &n->lock irq_context: 0 (wq_completion)events (work_completion)(&(&nsim_dev->trap_data->trap_report_dw)->work) &devlink->lock_key#3 &nsim_trap_data->trap_lock &base->lock irq_context: 0 (wq_completion)events (work_completion)(&(&nsim_dev->trap_data->trap_report_dw)->work) &devlink->lock_key#3 &nsim_trap_data->trap_lock &base->lock &obj_hash[i].lock irq_context: 0 &iint->mutex remove_cache_srcu &rq->__lock irq_context: 0 &iint->mutex remove_cache_srcu &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 &iint->mutex remove_cache_srcu pool_lock#2 irq_context: 0 &sbi->s_writepages_rwsem jbd2_handle remove_cache_srcu &rq->__lock irq_context: 0 &sbi->s_writepages_rwsem jbd2_handle remove_cache_srcu &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 sb_writers#4 per_cpu_ptr(&cgroup_rstat_cpu_lock, cpu) irq_context: 0 &sbi->s_writepages_rwsem rcu_node_0 irq_context: softirq (&icsk->icsk_retransmit_timer) slock-AF_INET6 rcu_read_lock pool_lock#2 irq_context: softirq (&icsk->icsk_retransmit_timer) slock-AF_INET6 pool_lock#2 irq_context: softirq (&ndev->rs_timer) rcu_read_lock rcu_read_lock rcu_read_lock_bh rcu_read_lock_bh tk_core.seq.seqcount irq_context: softirq (&icsk->icsk_retransmit_timer) slock-AF_INET6 &pcp->lock &zone->lock irq_context: softirq (&icsk->icsk_retransmit_timer) slock-AF_INET6 &pcp->lock &zone->lock &____s->seqcount irq_context: softirq (&icsk->icsk_retransmit_timer) slock-AF_INET6 &hashinfo->ehash_locks[i] irq_context: softirq (&icsk->icsk_retransmit_timer) slock-AF_INET6 &tcp_hashinfo.bhash[i].lock irq_context: softirq (&icsk->icsk_retransmit_timer) slock-AF_INET6 &tcp_hashinfo.bhash[i].lock stock_lock irq_context: softirq (&icsk->icsk_retransmit_timer) slock-AF_INET6 &tcp_hashinfo.bhash[i].lock &obj_hash[i].lock irq_context: softirq (&icsk->icsk_retransmit_timer) slock-AF_INET6 &tcp_hashinfo.bhash[i].lock pool_lock#2 irq_context: softirq (&icsk->icsk_retransmit_timer) slock-AF_INET6 &tcp_hashinfo.bhash[i].lock &tcp_hashinfo.bhash2[i].lock irq_context: softirq (&icsk->icsk_retransmit_timer) slock-AF_INET6 &tcp_hashinfo.bhash[i].lock &tcp_hashinfo.bhash2[i].lock stock_lock irq_context: softirq (&icsk->icsk_retransmit_timer) slock-AF_INET6 &tcp_hashinfo.bhash[i].lock &tcp_hashinfo.bhash2[i].lock &obj_hash[i].lock irq_context: softirq (&icsk->icsk_retransmit_timer) slock-AF_INET6 &tcp_hashinfo.bhash[i].lock &tcp_hashinfo.bhash2[i].lock pool_lock#2 irq_context: softirq (&icsk->icsk_retransmit_timer) slock-AF_INET6 elock-AF_INET6 irq_context: softirq (&icsk->icsk_retransmit_timer) &obj_hash[i].lock irq_context: softirq (&icsk->icsk_retransmit_timer) pool_lock#2 irq_context: softirq (&icsk->icsk_retransmit_timer) &dir->lock irq_context: softirq (&icsk->icsk_retransmit_timer) &pcp->lock &zone->lock irq_context: softirq (&icsk->icsk_retransmit_timer) &____s->seqcount irq_context: softirq (&icsk->icsk_retransmit_timer) stock_lock irq_context: 0 &mm->mmap_lock fs_reclaim &cfs_rq->removed.lock irq_context: 0 &mm->mmap_lock fs_reclaim &obj_hash[i].lock irq_context: 0 (wq_completion)wg-crypt-wg1#2 (work_completion)(&peer->transmit_packet_work) &peer->endpoint_lock rcu_read_lock_bh rcu_read_lock &obj_hash[i].lock irq_context: 0 &xt[i].mutex rcu_read_lock &obj_hash[i].lock pool_lock irq_context: 0 sb_writers#3 &cfs_rq->removed.lock irq_context: 0 sb_writers#3 &sb->s_type->i_mutex_key#9 remove_cache_srcu pool_lock#2 irq_context: 0 sb_writers#3 &sb->s_type->i_mutex_key#9 remove_cache_srcu &rq->__lock irq_context: 0 sb_writers#3 &sb->s_type->i_mutex_key#9 remove_cache_srcu &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 sb_writers#11 &type->i_mutex_dir_key#7 rename_lock.seqcount irq_context: 0 sb_writers#11 &type->i_mutex_dir_key#7 fs_reclaim irq_context: 0 sb_writers#11 &type->i_mutex_dir_key#7 fs_reclaim mmu_notifier_invalidate_range_start irq_context: 0 sb_writers#11 &type->i_mutex_dir_key#7 fs_reclaim mmu_notifier_invalidate_range_start &rq->__lock irq_context: 0 sb_writers#11 &type->i_mutex_dir_key#7 fs_reclaim mmu_notifier_invalidate_range_start &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 sb_writers#11 &type->i_mutex_dir_key#7 stock_lock irq_context: 0 sb_writers#11 &type->i_mutex_dir_key#7 &dentry->d_lock irq_context: 0 sb_writers#11 &type->i_mutex_dir_key#7 rcu_read_lock rename_lock.seqcount irq_context: 0 sb_writers#11 &type->i_mutex_dir_key#7 &root->kernfs_iattr_rwsem irq_context: 0 sb_writers#11 &type->i_mutex_dir_key#7 &root->kernfs_rwsem irq_context: 0 sb_writers#11 &type->i_mutex_dir_key#7 &dentry->d_lock &wq#2 irq_context: 0 sb_writers#11 &type->i_mutex_dir_key#7 rcu_read_lock &dentry->d_lock irq_context: 0 sb_writers#11 &type->i_mutex_dir_key#7 &dentry->d_lock &lru->node[i].lock irq_context: 0 key#25 irq_context: 0 &sb->s_type->i_mutex_key#10 &net->sctp.addr_wq_lock slock-AF_INET/1 key#25 irq_context: 0 sb_writers#4 fill_pool_map-wait-type-override &n->list_lock irq_context: 0 sb_writers#4 fill_pool_map-wait-type-override &n->list_lock &c->lock irq_context: softirq (&icsk->icsk_retransmit_timer) slock-AF_INET rcu_read_lock rcu_read_lock rcu_read_lock_bh dev->qdisc_tx_busylock ?: &qdisc_tx_busylock _xmit_ETHER#2 irq_context: softirq (&icsk->icsk_retransmit_timer) slock-AF_INET rcu_read_lock rcu_read_lock rcu_read_lock_bh dev->qdisc_tx_busylock ?: &qdisc_tx_busylock _xmit_ETHER#2 rcu_read_lock pool_lock#2 irq_context: softirq (&icsk->icsk_retransmit_timer) slock-AF_INET rcu_read_lock rcu_read_lock rcu_read_lock_bh dev->qdisc_tx_busylock ?: &qdisc_tx_busylock _xmit_ETHER#2 rcu_read_lock tk_core.seq.seqcount irq_context: softirq (&icsk->icsk_retransmit_timer) slock-AF_INET rcu_read_lock rcu_read_lock rcu_read_lock_bh dev->qdisc_tx_busylock ?: &qdisc_tx_busylock _xmit_ETHER#2 rcu_read_lock &obj_hash[i].lock irq_context: 0 (wq_completion)events (work_completion)(&w->w) nfc_devlist_mutex &genl_data->genl_data_mutex &rq->__lock irq_context: 0 (wq_completion)events (work_completion)(&w->w) nfc_devlist_mutex &genl_data->genl_data_mutex &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 &tsk->futex_exit_mutex pool_lock#2 irq_context: 0 (wq_completion)events (debug_obj_work).work &meta->lock irq_context: 0 (wq_completion)events (debug_obj_work).work kfence_freelist_lock irq_context: 0 &pipe->mutex/1 sk_lock-AF_INET6 rcu_read_lock quarantine_lock irq_context: 0 sk_lock-AF_INET6 rcu_read_lock rcu_read_lock &base->lock irq_context: 0 sk_lock-AF_INET6 rcu_read_lock rcu_read_lock &base->lock &obj_hash[i].lock irq_context: 0 sk_lock-AF_INET6 rcu_read_lock rcu_read_lock rcu_read_lock rhashtable_bucket irq_context: 0 sk_lock-AF_INET6 rcu_read_lock &q->lock#2 irq_context: 0 sk_lock-AF_INET6 rcu_read_lock &q->lock#2 &obj_hash[i].lock irq_context: 0 sk_lock-AF_INET6 rcu_read_lock &q->lock#2 &base->lock irq_context: 0 sk_lock-AF_INET6 rcu_read_lock &q->lock#2 &base->lock &obj_hash[i].lock irq_context: 0 sk_lock-AF_INET6 rcu_read_lock &q->lock#2 rcu_read_lock rcu_read_lock rhashtable_bucket irq_context: softirq rcu_read_lock rcu_read_lock rcu_read_lock rcu_read_lock rhashtable_bucket irq_context: softirq rcu_read_lock rcu_read_lock &q->lock#2 irq_context: softirq rcu_read_lock rcu_read_lock &q->lock#2 &obj_hash[i].lock irq_context: softirq rcu_read_lock rcu_read_lock &q->lock#2 &base->lock irq_context: softirq rcu_read_lock rcu_read_lock &q->lock#2 &base->lock &obj_hash[i].lock irq_context: softirq rcu_read_lock rcu_read_lock &q->lock#2 rcu_read_lock rcu_read_lock rhashtable_bucket irq_context: 0 cb_lock &rdev->wiphy.mtx fs_reclaim mmu_notifier_invalidate_range_start &rq->__lock irq_context: 0 cb_lock &rdev->wiphy.mtx fs_reclaim mmu_notifier_invalidate_range_start &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 sk_lock-AF_INET6 rcu_read_lock rcu_read_lock &ndev->lock irq_context: 0 sk_lock-AF_INET6 rcu_read_lock k-slock-AF_INET6 rcu_read_lock pool_lock#2 irq_context: 0 sk_lock-AF_INET6 rcu_read_lock k-slock-AF_INET6 rcu_read_lock rcu_read_lock pool_lock#2 irq_context: 0 sk_lock-AF_INET6 rcu_read_lock k-slock-AF_INET6 rcu_read_lock rcu_read_lock &obj_hash[i].lock irq_context: 0 sk_lock-AF_INET6 rcu_read_lock k-slock-AF_INET6 rcu_read_lock rcu_read_lock &dir->lock#2 irq_context: 0 sk_lock-AF_INET6 rcu_read_lock k-slock-AF_INET6 rcu_read_lock rcu_read_lock &tbl->lock irq_context: 0 sk_lock-AF_INET6 rcu_read_lock k-slock-AF_INET6 rcu_read_lock rcu_read_lock &n->lock irq_context: 0 sk_lock-AF_INET6 rcu_read_lock k-slock-AF_INET6 rcu_read_lock rcu_read_lock &____s->seqcount#9 irq_context: 0 sk_lock-AF_INET6 rcu_read_lock k-slock-AF_INET6 rcu_read_lock rcu_read_lock rcu_read_lock_bh tk_core.seq.seqcount irq_context: 0 sk_lock-AF_INET6 rcu_read_lock k-slock-AF_INET6 rcu_read_lock rcu_read_lock rcu_read_lock_bh &list->lock#5 irq_context: 0 sb_writers#4 &sb->s_type->i_mutex_key#8 mapping.invalidate_lock jbd2_handle &ei->i_data_sem remove_cache_srcu irq_context: 0 sb_writers#4 &sb->s_type->i_mutex_key#8 mapping.invalidate_lock jbd2_handle &ei->i_data_sem remove_cache_srcu quarantine_lock irq_context: 0 sb_writers#4 &sb->s_type->i_mutex_key#8 mapping.invalidate_lock jbd2_handle &ei->i_data_sem remove_cache_srcu &c->lock irq_context: 0 sb_writers#4 &sb->s_type->i_mutex_key#8 mapping.invalidate_lock jbd2_handle &ei->i_data_sem remove_cache_srcu &n->list_lock irq_context: 0 sb_writers#4 &sb->s_type->i_mutex_key#8 mapping.invalidate_lock jbd2_handle &ei->i_data_sem remove_cache_srcu rcu_read_lock pool_lock#2 irq_context: 0 sb_writers#4 &sb->s_type->i_mutex_key#8 mapping.invalidate_lock jbd2_handle &ei->i_data_sem remove_cache_srcu &obj_hash[i].lock irq_context: 0 sb_writers#4 &sb->s_type->i_mutex_key#8 mapping.invalidate_lock jbd2_handle &ei->i_data_sem remove_cache_srcu rcu_read_lock rcu_node_0 irq_context: 0 sb_writers#4 &sb->s_type->i_mutex_key#8 mapping.invalidate_lock jbd2_handle &ei->i_data_sem remove_cache_srcu rcu_read_lock &rq->__lock irq_context: 0 sb_writers#4 &sb->s_type->i_mutex_key#8 mapping.invalidate_lock jbd2_handle &ei->i_data_sem remove_cache_srcu rcu_read_lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 sb_writers#4 tomoyo_ss pool_lock#2 irq_context: 0 &mm->mmap_lock rcu_read_lock &obj_hash[i].lock pool_lock irq_context: 0 &sig->cred_guard_mutex &sb->s_type->i_mutex_key#8 &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 rcu_read_lock &vma->vm_lock->lock rcu_read_lock &cfs_rq->removed.lock irq_context: 0 rcu_read_lock &vma->vm_lock->lock rcu_read_lock &obj_hash[i].lock irq_context: 0 sb_writers#4 &type->i_mutex_dir_key#3/1 &rcu_state.gp_wq irq_context: 0 sb_writers#4 &type->i_mutex_dir_key#3/1 &rcu_state.gp_wq &p->pi_lock irq_context: 0 sb_writers#4 &type->i_mutex_dir_key#3/1 &rcu_state.gp_wq &p->pi_lock &rq->__lock irq_context: 0 sb_writers#4 &type->i_mutex_dir_key#3/1 &rcu_state.gp_wq &p->pi_lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 sb_writers#4 &sb->s_type->i_mutex_key#8 mapping.invalidate_lock remove_cache_srcu &rq->__lock &cfs_rq->removed.lock irq_context: 0 &sb->s_type->i_mutex_key#10 sk_lock-AF_INET6 &tcp_hashinfo.bhash[i].lock irq_context: 0 &sb->s_type->i_mutex_key#10 sk_lock-AF_INET6 &tcp_hashinfo.bhash[i].lock stock_lock irq_context: 0 &sb->s_type->i_mutex_key#10 sk_lock-AF_INET6 &tcp_hashinfo.bhash[i].lock &obj_hash[i].lock irq_context: 0 &sb->s_type->i_mutex_key#10 sk_lock-AF_INET6 &tcp_hashinfo.bhash[i].lock pool_lock#2 irq_context: 0 &sb->s_type->i_mutex_key#10 sk_lock-AF_INET6 &tcp_hashinfo.bhash[i].lock &tcp_hashinfo.bhash2[i].lock irq_context: 0 &sb->s_type->i_mutex_key#10 sk_lock-AF_INET6 &tcp_hashinfo.bhash[i].lock &tcp_hashinfo.bhash2[i].lock stock_lock irq_context: 0 &sb->s_type->i_mutex_key#10 sk_lock-AF_INET6 &tcp_hashinfo.bhash[i].lock &tcp_hashinfo.bhash2[i].lock &obj_hash[i].lock irq_context: 0 &sb->s_type->i_mutex_key#10 sk_lock-AF_INET6 &tcp_hashinfo.bhash[i].lock &tcp_hashinfo.bhash2[i].lock pool_lock#2 irq_context: 0 sk_lock-AF_INET6 rcu_read_lock k-slock-AF_INET6 rcu_read_lock rcu_read_lock &____s->seqcount#10 irq_context: 0 (wq_completion)events (work_completion)(&w->w) &obj_hash[i].lock pool_lock irq_context: 0 sk_lock-AF_INET6 &dccp_hashinfo.bhash[i].lock stock_lock irq_context: 0 sk_lock-AF_INET6 &dccp_hashinfo.bhash[i].lock &____s->seqcount#2 irq_context: 0 sk_lock-AF_INET6 &dccp_hashinfo.bhash[i].lock &____s->seqcount irq_context: 0 sk_lock-AF_INET6 &dccp_hashinfo.bhash[i].lock pool_lock#2 irq_context: 0 sk_lock-AF_INET6 &dccp_hashinfo.bhash[i].lock &c->lock irq_context: 0 sk_lock-AF_INET6 &dccp_hashinfo.bhash[i].lock &dccp_hashinfo.bhash2[i].lock &pcp->lock &zone->lock irq_context: 0 sk_lock-AF_INET6 &dccp_hashinfo.bhash[i].lock &dccp_hashinfo.bhash2[i].lock rcu_read_lock pool_lock#2 irq_context: 0 sk_lock-AF_INET6 &dccp_hashinfo.bhash[i].lock &dccp_hashinfo.bhash2[i].lock &obj_hash[i].lock irq_context: 0 sk_lock-AF_INET6 &dccp_hashinfo.bhash[i].lock clock-AF_INET6 irq_context: softirq rcu_read_lock rcu_read_lock slock-AF_INET6/1 sk_lock-AF_INET6 rcu_read_lock &____s->seqcount#2 irq_context: softirq rcu_read_lock rcu_read_lock slock-AF_INET6/1 sk_lock-AF_INET6 rcu_read_lock &____s->seqcount irq_context: softirq rcu_read_lock rcu_read_lock slock-AF_INET6/1 sk_lock-AF_INET6 rcu_read_lock pool_lock#2 irq_context: softirq rcu_read_lock rcu_read_lock slock-AF_INET6/1 sk_lock-AF_INET6 rcu_read_lock &c->lock irq_context: softirq rcu_read_lock rcu_read_lock slock-AF_INET6/1 sk_lock-AF_INET6 rcu_read_lock tk_core.seq.seqcount irq_context: softirq rcu_read_lock rcu_read_lock slock-AF_INET6/1 sk_lock-AF_INET6 rcu_read_lock rcu_read_lock rcu_read_lock &____s->seqcount#7 irq_context: softirq rcu_read_lock rcu_read_lock slock-AF_INET6/1 sk_lock-AF_INET6 rcu_read_lock rcu_read_lock rcu_read_lock &ct->lock irq_context: softirq rcu_read_lock rcu_read_lock slock-AF_INET6/1 sk_lock-AF_INET6 rcu_read_lock rcu_read_lock rcu_read_lock &____s->seqcount#10 irq_context: softirq rcu_read_lock rcu_read_lock slock-AF_INET6/1 sk_lock-AF_INET6 rcu_read_lock rcu_read_lock rcu_read_lock rcu_read_lock_bh tk_core.seq.seqcount irq_context: softirq rcu_read_lock rcu_read_lock slock-AF_INET6/1 sk_lock-AF_INET6 rcu_read_lock rcu_read_lock rcu_read_lock rcu_read_lock_bh &list->lock#5 irq_context: softirq rcu_read_lock rcu_read_lock slock-AF_INET6/1 sk_lock-AF_INET6 rcu_read_lock &obj_hash[i].lock irq_context: softirq rcu_read_lock rcu_read_lock slock-AF_INET6/1 sk_lock-AF_INET6 rcu_read_lock &base->lock irq_context: softirq rcu_read_lock rcu_read_lock slock-AF_INET6/1 sk_lock-AF_INET6 rcu_read_lock &base->lock &obj_hash[i].lock irq_context: softirq rcu_read_lock rcu_read_lock slock-AF_INET6/1 sk_lock-AF_INET6 rcu_read_lock &hashinfo->ehash_locks[i] irq_context: softirq rcu_read_lock rcu_read_lock slock-AF_INET6/1 sk_lock-AF_INET6 &obj_hash[i].lock irq_context: softirq rcu_read_lock rcu_read_lock slock-AF_INET6/1 sk_lock-AF_INET6 pool_lock#2 irq_context: 0 sk_lock-AF_INET rcu_read_lock rcu_read_lock rcu_read_lock_bh rcu_read_lock &ei->socket.wq.wait irq_context: softirq rcu_read_lock rcu_read_lock &dreq->dreq_lock irq_context: softirq rcu_read_lock rcu_read_lock &dreq->dreq_lock pool_lock#2 irq_context: softirq rcu_read_lock rcu_read_lock &dreq->dreq_lock &dir->lock irq_context: softirq rcu_read_lock rcu_read_lock &dreq->dreq_lock slock-AF_INET6 irq_context: softirq rcu_read_lock rcu_read_lock &dreq->dreq_lock slock-AF_INET6 pool_lock#2 irq_context: softirq rcu_read_lock rcu_read_lock &dreq->dreq_lock slock-AF_INET6 &____s->seqcount irq_context: softirq rcu_read_lock rcu_read_lock &dreq->dreq_lock slock-AF_INET6 &obj_hash[i].lock irq_context: softirq rcu_read_lock rcu_read_lock &dreq->dreq_lock slock-AF_INET6 batched_entropy_u16.lock irq_context: softirq rcu_read_lock rcu_read_lock &dreq->dreq_lock slock-AF_INET6 &dccp_hashinfo.bhash[i].lock irq_context: softirq rcu_read_lock rcu_read_lock &dreq->dreq_lock slock-AF_INET6 &dccp_hashinfo.bhash[i].lock &dccp_hashinfo.bhash2[i].lock irq_context: softirq rcu_read_lock rcu_read_lock &dreq->dreq_lock slock-AF_INET6 &dccp_hashinfo.bhash[i].lock &dccp_hashinfo.bhash2[i].lock pool_lock#2 irq_context: softirq rcu_read_lock rcu_read_lock &dreq->dreq_lock slock-AF_INET6 &hashinfo->ehash_locks[i] irq_context: softirq rcu_read_lock rcu_read_lock &dreq->dreq_lock slock-AF_INET6 (&req->rsk_timer) irq_context: softirq rcu_read_lock rcu_read_lock &dreq->dreq_lock slock-AF_INET6 &base->lock irq_context: softirq rcu_read_lock rcu_read_lock &dreq->dreq_lock slock-AF_INET6 &base->lock &obj_hash[i].lock irq_context: softirq rcu_read_lock rcu_read_lock &dreq->dreq_lock slock-AF_INET6 &queue->rskq_lock irq_context: softirq rcu_read_lock rcu_read_lock slock-AF_INET6 tk_core.seq.seqcount irq_context: softirq rcu_read_lock rcu_read_lock slock-AF_INET6 &obj_hash[i].lock irq_context: softirq rcu_read_lock rcu_read_lock slock-AF_INET6 pool_lock#2 irq_context: 0 sk_lock-AF_INET wlock-AF_INET irq_context: softirq rcu_read_lock rcu_read_lock slock-AF_INET6/1 sk_lock-AF_INET6 &____s->seqcount#2 irq_context: softirq rcu_read_lock rcu_read_lock slock-AF_INET6/1 sk_lock-AF_INET6 &____s->seqcount irq_context: softirq rcu_read_lock rcu_read_lock slock-AF_INET6/1 sk_lock-AF_INET6 &c->lock irq_context: softirq rcu_read_lock rcu_read_lock slock-AF_INET6/1 sk_lock-AF_INET6 rcu_read_lock rcu_read_lock &____s->seqcount#7 irq_context: softirq rcu_read_lock rcu_read_lock slock-AF_INET6/1 sk_lock-AF_INET6 rcu_read_lock rcu_read_lock &ct->lock irq_context: softirq rcu_read_lock rcu_read_lock slock-AF_INET6/1 sk_lock-AF_INET6 rcu_read_lock rcu_read_lock &____s->seqcount#10 irq_context: softirq rcu_read_lock rcu_read_lock slock-AF_INET6/1 sk_lock-AF_INET6 rcu_read_lock rcu_read_lock rcu_read_lock_bh tk_core.seq.seqcount irq_context: softirq rcu_read_lock rcu_read_lock slock-AF_INET6/1 sk_lock-AF_INET6 rcu_read_lock rcu_read_lock rcu_read_lock_bh &list->lock#5 irq_context: 0 sk_lock-AF_INET &p->pi_lock irq_context: 0 sk_lock-AF_INET &p->pi_lock &rq->__lock irq_context: 0 sk_lock-AF_INET &p->pi_lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 &sb->s_type->i_mutex_key#10 sk_lock-AF_INET6 slock-AF_INET6 &hashinfo->ehash_locks[i] irq_context: 0 &sb->s_type->i_mutex_key#10 sk_lock-AF_INET6 slock-AF_INET6 &dccp_hashinfo.bhash[i].lock irq_context: 0 &sb->s_type->i_mutex_key#10 sk_lock-AF_INET6 slock-AF_INET6 &dccp_hashinfo.bhash[i].lock &dccp_hashinfo.bhash2[i].lock irq_context: 0 &sb->s_type->i_mutex_key#10 sk_lock-AF_INET6 slock-AF_INET6 &dccp_hashinfo.bhash[i].lock &dccp_hashinfo.bhash2[i].lock &obj_hash[i].lock irq_context: 0 &sb->s_type->i_mutex_key#10 sk_lock-AF_INET6 slock-AF_INET6 &dccp_hashinfo.bhash[i].lock &dccp_hashinfo.bhash2[i].lock pool_lock#2 irq_context: 0 &sb->s_type->i_mutex_key#10 sk_lock-AF_INET6 slock-AF_INET6 rcu_read_lock rcu_read_lock &____s->seqcount#7 irq_context: 0 &sb->s_type->i_mutex_key#10 sk_lock-AF_INET6 slock-AF_INET6 rcu_read_lock rcu_read_lock &ct->lock irq_context: 0 &sb->s_type->i_mutex_key#10 sk_lock-AF_INET6 slock-AF_INET6 rcu_read_lock rcu_read_lock &____s->seqcount#10 irq_context: 0 &sb->s_type->i_mutex_key#10 sk_lock-AF_INET6 slock-AF_INET6 rcu_read_lock rcu_read_lock rcu_read_lock_bh tk_core.seq.seqcount irq_context: 0 &sb->s_type->i_mutex_key#10 sk_lock-AF_INET6 slock-AF_INET6 rcu_read_lock rcu_read_lock rcu_read_lock_bh &list->lock#5 irq_context: 0 &sb->s_type->i_mutex_key#10 sk_lock-AF_INET6 slock-AF_INET6 clock-AF_INET6 irq_context: softirq rcu_read_lock rcu_read_lock slock-AF_INET/1 sk_lock-AF_INET pool_lock#2 irq_context: softirq rcu_read_lock rcu_read_lock slock-AF_INET/1 sk_lock-AF_INET &obj_hash[i].lock irq_context: softirq rcu_read_lock rcu_read_lock slock-AF_INET/1 sk_lock-AF_INET &____s->seqcount#2 irq_context: softirq rcu_read_lock rcu_read_lock slock-AF_INET/1 sk_lock-AF_INET &____s->seqcount irq_context: softirq rcu_read_lock rcu_read_lock slock-AF_INET/1 sk_lock-AF_INET &c->lock irq_context: softirq rcu_read_lock rcu_read_lock slock-AF_INET/1 sk_lock-AF_INET &base->lock irq_context: softirq rcu_read_lock rcu_read_lock slock-AF_INET/1 sk_lock-AF_INET &base->lock &obj_hash[i].lock irq_context: softirq rcu_read_lock rcu_read_lock slock-AF_INET/1 sk_lock-AF_INET &dccp_hashinfo.bhash[i].lock irq_context: softirq rcu_read_lock rcu_read_lock slock-AF_INET/1 sk_lock-AF_INET &dccp_hashinfo.bhash[i].lock &dccp_hashinfo.bhash2[i].lock irq_context: softirq rcu_read_lock rcu_read_lock slock-AF_INET/1 sk_lock-AF_INET &hashinfo->ehash_locks[i] irq_context: 0 &sb->s_type->i_mutex_key#10 slock-AF_INET6 &dccp_hashinfo.bhash[i].lock irq_context: 0 &sb->s_type->i_mutex_key#10 slock-AF_INET6 &dccp_hashinfo.bhash[i].lock stock_lock irq_context: 0 &sb->s_type->i_mutex_key#10 slock-AF_INET6 &dccp_hashinfo.bhash[i].lock &obj_hash[i].lock irq_context: 0 &sb->s_type->i_mutex_key#10 slock-AF_INET6 &dccp_hashinfo.bhash[i].lock pool_lock#2 irq_context: 0 &sb->s_type->i_mutex_key#10 slock-AF_INET6 &dccp_hashinfo.bhash[i].lock &dccp_hashinfo.bhash2[i].lock irq_context: 0 &sb->s_type->i_mutex_key#10 slock-AF_INET6 &dccp_hashinfo.bhash[i].lock &dccp_hashinfo.bhash2[i].lock stock_lock irq_context: 0 &sb->s_type->i_mutex_key#10 slock-AF_INET6 &dccp_hashinfo.bhash[i].lock &dccp_hashinfo.bhash2[i].lock &obj_hash[i].lock irq_context: 0 &sb->s_type->i_mutex_key#10 slock-AF_INET6 &dccp_hashinfo.bhash[i].lock &dccp_hashinfo.bhash2[i].lock pool_lock#2 irq_context: 0 &sb->s_type->i_mutex_key#10 slock-AF_INET &____s->seqcount irq_context: 0 &sb->s_type->i_mutex_key#10 slock-AF_INET rcu_read_lock pool_lock#2 irq_context: softirq (&peer->timer_retransmit_handshake) &obj_hash[i].lock irq_context: softirq (&peer->timer_retransmit_handshake) &list->lock#17 irq_context: 0 sk_lock-AF_INET rcu_read_lock rcu_read_lock &nf_conntrack_locks[i] &nf_conntrack_locks[i]/1 batched_entropy_u8.lock crngs.lock irq_context: softirq rcu_read_lock rcu_read_lock slock-AF_INET6/1 sk_lock-AF_INET6 rcu_read_lock &n->list_lock irq_context: softirq rcu_read_lock rcu_read_lock slock-AF_INET6/1 sk_lock-AF_INET6 rcu_read_lock &n->list_lock &c->lock irq_context: 0 &pipe->mutex/1 &pipe->wr_wait &p->pi_lock &cfs_rq->removed.lock irq_context: 0 &pipe->mutex/1 &sighand->siglock per_cpu_ptr(&cgroup_rstat_cpu_lock, cpu) irq_context: softirq rcu_read_lock rcu_read_lock &dreq->dreq_lock &c->lock irq_context: softirq rcu_read_lock rcu_read_lock &dreq->dreq_lock slock-AF_INET6 &obj_hash[i].lock pool_lock irq_context: 0 &sb->s_type->i_mutex_key#10 sk_lock-AF_INET6 slock-AF_INET6 &c->lock irq_context: 0 sk_lock-AF_INET6 &tcp_hashinfo.bhash[i].lock &tcp_hashinfo.bhash2[i].lock &c->lock irq_context: 0 sk_lock-AF_INET6 rcu_read_lock k-slock-AF_INET6 rcu_read_lock &c->lock irq_context: 0 &xs->mutex fs_reclaim mmu_notifier_invalidate_range_start &cfs_rq->removed.lock irq_context: 0 &xs->mutex fs_reclaim mmu_notifier_invalidate_range_start &obj_hash[i].lock irq_context: softirq slock-AF_INET rcu_read_lock rcu_read_lock rcu_read_lock_bh dev->qdisc_tx_busylock ?: &qdisc_tx_busylock _xmit_ETHER#2 rcu_read_lock &n->list_lock irq_context: softirq slock-AF_INET rcu_read_lock rcu_read_lock rcu_read_lock_bh dev->qdisc_tx_busylock ?: &qdisc_tx_busylock _xmit_ETHER#2 rcu_read_lock &n->list_lock &c->lock irq_context: 0 (wq_completion)wg-crypt-wg0#6 (work_completion)(&({ do { const void *__vpp_verify = (typeof((worker) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __asm__ ("" : "=r"(__ptr) : "0"((typeof(*((worker))) *)((worker)))); (typeof((typeof(*((worker))) *)((worker)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); })->work) &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 &nft_net->commit_mutex rcu_state.exp_mutex &obj_hash[i].lock pool_lock irq_context: softirq rcu_read_lock rcu_read_lock rcu_read_lock &obj_hash[i].lock pool_lock irq_context: 0 sb_writers#4 &fsnotify_mark_srcu &cfs_rq->removed.lock irq_context: 0 sb_writers#4 &fsnotify_mark_srcu &obj_hash[i].lock irq_context: 0 sb_writers#4 &type->i_mutex_dir_key#3/1 &sb->s_type->i_mutex_key#8 jbd2_handle rcu_read_lock &rq->__lock &cfs_rq->removed.lock irq_context: 0 (wq_completion)wg-crypt-wg0#5 (work_completion)(&peer->transmit_packet_work) &cfs_rq->removed.lock irq_context: 0 &pipe->mutex/1 sk_lock-AF_INET6 pgd_lock irq_context: 0 &pipe->mutex/1 sk_lock-AF_INET6 stock_lock irq_context: 0 &pipe->mutex/1 sk_lock-AF_INET6 key irq_context: 0 &pipe->mutex/1 sk_lock-AF_INET6 pcpu_lock irq_context: 0 &pipe->mutex/1 sk_lock-AF_INET6 percpu_counters_lock irq_context: 0 &pipe->mutex/1 sk_lock-AF_INET6 pcpu_lock stock_lock irq_context: 0 &mm->mmap_lock remove_cache_srcu &pcp->lock &zone->lock irq_context: 0 &mm->mmap_lock remove_cache_srcu &pcp->lock &zone->lock &____s->seqcount irq_context: 0 &f->f_pos_lock sb_writers#4 &sb->s_type->i_mutex_key#8 jbd2_handle rcu_read_lock &cfs_rq->removed.lock irq_context: 0 &f->f_pos_lock sb_writers#4 &sb->s_type->i_mutex_key#8 jbd2_handle rcu_read_lock &obj_hash[i].lock irq_context: 0 (wq_completion)wg-crypt-wg1#3 (work_completion)(&peer->transmit_packet_work) &peer->endpoint_lock rcu_read_lock_bh rcu_read_lock batched_entropy_u8.lock irq_context: 0 (wq_completion)wg-crypt-wg1#3 (work_completion)(&peer->transmit_packet_work) &peer->endpoint_lock rcu_read_lock_bh rcu_read_lock kfence_freelist_lock irq_context: 0 (wq_completion)wg-crypt-wg1#3 (work_completion)(&peer->transmit_packet_work) &peer->endpoint_lock rcu_read_lock_bh rcu_read_lock &c->lock irq_context: 0 &sbi->s_writepages_rwsem jbd2_handle &rq_wait->wait irq_context: 0 &sbi->s_writepages_rwsem jbd2_handle rcu_read_lock &pool->lock irq_context: 0 &sbi->s_writepages_rwsem jbd2_handle rcu_read_lock &pool->lock &obj_hash[i].lock irq_context: 0 &sbi->s_writepages_rwsem jbd2_handle rcu_read_lock &pool->lock pool_lock#2 irq_context: 0 &sbi->s_writepages_rwsem jbd2_handle rcu_read_lock &pool->lock &p->pi_lock irq_context: 0 &sbi->s_writepages_rwsem jbd2_handle rcu_read_lock &pool->lock &p->pi_lock &rq->__lock irq_context: 0 &sbi->s_writepages_rwsem jbd2_handle rcu_read_lock &pool->lock &p->pi_lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 &sbi->s_writepages_rwsem &rq_wait->wait irq_context: 0 &sbi->s_writepages_rwsem rcu_read_lock &pool->lock irq_context: 0 &sbi->s_writepages_rwsem rcu_read_lock &pool->lock &obj_hash[i].lock irq_context: 0 &sbi->s_writepages_rwsem rcu_read_lock &pool->lock pool_lock#2 irq_context: 0 &sbi->s_writepages_rwsem rcu_read_lock &pool->lock &p->pi_lock irq_context: 0 &sbi->s_writepages_rwsem rcu_read_lock &pool->lock &p->pi_lock &rq->__lock irq_context: 0 &sbi->s_writepages_rwsem rcu_read_lock &pool->lock &p->pi_lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)kblockd (work_completion)(&(&hctx->run_work)->work) rcu_read_lock &c->lock irq_context: 0 sk_lock-AF_INET6 rcu_read_lock rcu_read_lock &____s->seqcount#2 irq_context: 0 sk_lock-AF_INET6 rcu_read_lock rcu_read_lock &____s->seqcount irq_context: 0 delayed_uprobe_lock delayed_uprobe_lock.wait_lock irq_context: 0 delayed_uprobe_lock &cfs_rq->removed.lock irq_context: 0 delayed_uprobe_lock &obj_hash[i].lock irq_context: 0 delayed_uprobe_lock pool_lock#2 irq_context: 0 delayed_uprobe_lock.wait_lock irq_context: 0 sb_writers#5 &type->i_mutex_dir_key#5/1 &sb->s_type->i_mutex_key#12 &sb->s_type->i_mutex_key#12/4 &cfs_rq->removed.lock irq_context: 0 sb_writers#5 &type->i_mutex_dir_key#5/1 &sb->s_type->i_mutex_key#12 &sb->s_type->i_mutex_key#12/4 &obj_hash[i].lock irq_context: 0 sb_writers#5 &type->i_mutex_dir_key#5/1 &sb->s_type->i_mutex_key#12 &sb->s_type->i_mutex_key#12/4 pool_lock#2 irq_context: 0 &p->lock &of->mutex kn->active#5 remove_cache_srcu &____s->seqcount irq_context: 0 &p->lock &of->mutex kn->active#5 remove_cache_srcu pool_lock#2 irq_context: 0 sk_lock-AF_INET6 rcu_read_lock &obj_hash[i].lock pool_lock irq_context: 0 sb_writers#4 &sb->s_type->i_lock_key#22 &xa->xa_lock#9 fill_pool_map-wait-type-override pool_lock#2 irq_context: 0 sb_writers#4 &sb->s_type->i_lock_key#22 &xa->xa_lock#9 fill_pool_map-wait-type-override pool_lock irq_context: 0 (wq_completion)wg-crypt-wg0 (work_completion)(&peer->transmit_packet_work) rcu_read_lock_bh rcu_read_lock &pool->lock/1 &p->pi_lock &rq->__lock irq_context: 0 (wq_completion)wg-crypt-wg0 (work_completion)(&peer->transmit_packet_work) rcu_read_lock_bh rcu_read_lock &pool->lock/1 &p->pi_lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)wg-kex-wg1 (work_completion)(&({ do { const void *__vpp_verify = (typeof((worker) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __asm__ ("" : "=r"(__ptr) : "0"((typeof(*((worker))) *)((worker)))); (typeof((typeof(*((worker))) *)((worker)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); })->work) &handshake->lock rcu_read_lock_bh &peer->keypairs.keypair_update_lock &table->lock#2 irq_context: 0 (wq_completion)wg-kex-wg1 (work_completion)(&({ do { const void *__vpp_verify = (typeof((worker) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __asm__ ("" : "=r"(__ptr) : "0"((typeof(*((worker))) *)((worker)))); (typeof((typeof(*((worker))) *)((worker)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); })->work) &handshake->lock rcu_read_lock_bh &peer->keypairs.keypair_update_lock &obj_hash[i].lock irq_context: 0 (wq_completion)wg-kex-wg0#2 (work_completion)(&({ do { const void *__vpp_verify = (typeof((worker) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __asm__ ("" : "=r"(__ptr) : "0"((typeof(*((worker))) *)((worker)))); (typeof((typeof(*((worker))) *)((worker)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); })->work) &handshake->lock rcu_read_lock_bh &peer->keypairs.keypair_update_lock &table->lock#2 irq_context: 0 (wq_completion)wg-kex-wg0#2 (work_completion)(&({ do { const void *__vpp_verify = (typeof((worker) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __asm__ ("" : "=r"(__ptr) : "0"((typeof(*((worker))) *)((worker)))); (typeof((typeof(*((worker))) *)((worker)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); })->work) &handshake->lock rcu_read_lock_bh &peer->keypairs.keypair_update_lock &obj_hash[i].lock irq_context: 0 sb_writers#4 &type->i_mutex_dir_key#3 jbd2_handle &journal->j_state_lock irq_context: 0 sb_writers#4 &type->i_mutex_dir_key#3 jbd2_handle &journal->j_state_lock &journal->j_wait_commit irq_context: 0 sb_writers#4 &type->i_mutex_dir_key#3 jbd2_handle &journal->j_state_lock &journal->j_wait_commit &p->pi_lock irq_context: 0 sb_writers#4 &type->i_mutex_dir_key#3 jbd2_handle &journal->j_state_lock &journal->j_wait_commit &p->pi_lock &rq->__lock irq_context: 0 sb_writers#4 &type->i_mutex_dir_key#3 jbd2_handle &journal->j_state_lock &journal->j_wait_commit &p->pi_lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 sb_writers#4 &type->i_mutex_dir_key#3 jbd2_handle &journal->j_wait_updates &p->pi_lock irq_context: 0 sb_writers#4 &type->i_mutex_dir_key#3 jbd2_handle &journal->j_wait_updates &p->pi_lock &rq->__lock irq_context: 0 sb_writers#4 &type->i_mutex_dir_key#3 jbd2_handle &journal->j_wait_updates &p->pi_lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)bat_events (work_completion)(&(&bat_priv->bla.work)->work) rcu_read_lock &cfs_rq->removed.lock irq_context: 0 (wq_completion)bat_events (work_completion)(&(&bat_priv->bla.work)->work) rcu_read_lock &obj_hash[i].lock irq_context: 0 cb_lock genl_mutex fs_reclaim mmu_notifier_invalidate_range_start &rq->__lock irq_context: 0 cb_lock genl_mutex fs_reclaim mmu_notifier_invalidate_range_start &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 sb_writers#8 &mm->mmap_lock &rq->__lock irq_context: 0 sb_writers#4 &type->i_mutex_dir_key#3/1 batched_entropy_u8.lock crngs.lock irq_context: 0 sk_lock-AF_INET slock-AF_INET &sk->sk_lock.wq irq_context: 0 slock-AF_INET &sk->sk_lock.wq irq_context: 0 slock-AF_INET &sk->sk_lock.wq &p->pi_lock irq_context: 0 slock-AF_INET &sk->sk_lock.wq &p->pi_lock &rq->__lock irq_context: 0 slock-AF_INET &sk->sk_lock.wq &p->pi_lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 &pipe->mutex/1 sk_lock-AF_INET irq_context: 0 &pipe->mutex/1 sk_lock-AF_INET slock-AF_INET irq_context: 0 &pipe->mutex/1 sk_lock-AF_INET rcu_read_lock &sighand->siglock irq_context: 0 &pipe->mutex/1 sk_lock-AF_INET &sighand->siglock irq_context: 0 &pipe->mutex/1 sk_lock-AF_INET &sighand->siglock stock_lock irq_context: 0 &pipe->mutex/1 sk_lock-AF_INET &sighand->siglock pool_lock#2 irq_context: 0 &pipe->mutex/1 sk_lock-AF_INET &sighand->siglock &p->pi_lock irq_context: 0 &pipe->mutex/1 sk_lock-AF_INET &sighand->siglock &p->pi_lock &rq->__lock irq_context: 0 &pipe->mutex/1 sk_lock-AF_INET &sighand->siglock &p->pi_lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 &pipe->mutex/1 sk_lock-AF_INET &rq->__lock irq_context: 0 &pipe->mutex/1 sk_lock-AF_INET &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 &pipe->mutex/1 sk_lock-AF_INET &cfs_rq->removed.lock irq_context: 0 &pipe->mutex/1 sk_lock-AF_INET &obj_hash[i].lock irq_context: 0 &pipe->mutex/1 slock-AF_INET irq_context: softirq (&in_dev->mr_ifc_timer) rcu_read_lock &tbl->lock &c->lock irq_context: 0 raw_notifier_lock irq_context: 0 sk_lock-AF_INET6 &dccp_hashinfo.bhash[i].lock &dccp_hashinfo.bhash2[i].lock batched_entropy_u8.lock irq_context: 0 sk_lock-AF_INET6 &dccp_hashinfo.bhash[i].lock &dccp_hashinfo.bhash2[i].lock &hashinfo->ehash_locks[i] irq_context: 0 &sb->s_type->i_mutex_key#10 raw_notifier_lock irq_context: 0 &sb->s_type->i_mutex_key#10 rtnl_mutex sk_lock-AF_CAN irq_context: 0 &sb->s_type->i_mutex_key#10 rtnl_mutex sk_lock-AF_CAN slock-AF_CAN irq_context: 0 &sb->s_type->i_mutex_key#10 rtnl_mutex sk_lock-AF_CAN pcpu_lock irq_context: 0 &sb->s_type->i_mutex_key#10 rtnl_mutex sk_lock-AF_CAN clock-AF_CAN irq_context: 0 &sb->s_type->i_mutex_key#10 rtnl_mutex slock-AF_CAN irq_context: 0 sb_writers#3 oom_adj_mutex &cfs_rq->removed.lock irq_context: 0 sb_writers#3 oom_adj_mutex &obj_hash[i].lock irq_context: 0 sb_writers#3 oom_adj_mutex pool_lock#2 irq_context: 0 (wq_completion)events (work_completion)(&rdev->wiphy_work) &rdev->wiphy.mtx &wdev->mtx rcu_read_lock &cfs_rq->removed.lock irq_context: 0 (wq_completion)events (work_completion)(&rdev->wiphy_work) &rdev->wiphy.mtx &wdev->mtx rcu_read_lock &obj_hash[i].lock irq_context: 0 (wq_completion)wg-kex-wg0#7 (work_completion)(&peer->transmit_handshake_work) &wg->static_identity.lock &handshake->lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: softirq (&ndev->rs_timer) rcu_read_lock rcu_read_lock rcu_read_lock_bh rcu_read_lock rcu_read_lock &n->list_lock irq_context: softirq (&ndev->rs_timer) rcu_read_lock rcu_read_lock rcu_read_lock_bh rcu_read_lock rcu_read_lock &n->list_lock &c->lock irq_context: 0 (wq_completion)wg-crypt-wg2 (work_completion)(&peer->transmit_packet_work) &peer->endpoint_lock rcu_read_lock_bh batched_entropy_u32.lock crngs.lock irq_context: 0 &sb->s_type->i_mutex_key#10 sk_lock-AF_INET6 &dccp_hashinfo.bhash[i].lock stock_lock rcu_read_lock per_cpu_ptr(&cgroup_rstat_cpu_lock, cpu) irq_context: 0 &f->f_pos_lock sb_writers#4 &sb->s_type->i_mutex_key#8 remove_cache_srcu rcu_node_0 irq_context: 0 remove_cache_srcu rcu_read_lock &rq->__lock &cfs_rq->removed.lock irq_context: 0 (wq_completion)bat_events (work_completion)(&(&forw_packet_aggr->delayed_work)->work) batched_entropy_u8.lock crngs.lock irq_context: 0 &sb->s_type->i_mutex_key#10 sk_lock-AF_INET6 &pcp->lock &zone->lock irq_context: 0 sb_writers#4 &type->i_mutex_dir_key#3/1 tomoyo_ss rcu_read_lock &rq->__lock &cfs_rq->removed.lock irq_context: 0 sb_writers#8 &of->mutex kn->active#5 uevent_sock_mutex rcu_read_lock &cfs_rq->removed.lock irq_context: 0 sb_writers#8 &of->mutex kn->active#5 uevent_sock_mutex rcu_read_lock &obj_hash[i].lock irq_context: 0 cb_lock &rdev->wiphy.mtx rcu_read_lock &rcu_state.gp_wq irq_context: 0 cb_lock &rdev->wiphy.mtx rcu_read_lock &rcu_state.gp_wq &p->pi_lock irq_context: 0 cb_lock &rdev->wiphy.mtx rcu_read_lock &rcu_state.gp_wq &p->pi_lock &rq->__lock irq_context: 0 cb_lock &rdev->wiphy.mtx rcu_read_lock &rcu_state.gp_wq &p->pi_lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 sb_writers#4 &sb->s_type->i_lock_key#22 &xa->xa_lock#9 fill_pool_map-wait-type-override &c->lock irq_context: 0 &iint->mutex ima_extend_list_mutex remove_cache_srcu irq_context: 0 &iint->mutex ima_extend_list_mutex remove_cache_srcu &rq->__lock irq_context: 0 &iint->mutex ima_extend_list_mutex remove_cache_srcu &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 &iint->mutex ima_extend_list_mutex remove_cache_srcu quarantine_lock irq_context: 0 &iint->mutex ima_extend_list_mutex remove_cache_srcu &c->lock irq_context: 0 &iint->mutex ima_extend_list_mutex remove_cache_srcu &n->list_lock irq_context: 0 &iint->mutex ima_extend_list_mutex remove_cache_srcu rcu_read_lock pool_lock#2 irq_context: 0 &iint->mutex ima_extend_list_mutex remove_cache_srcu &obj_hash[i].lock irq_context: 0 &pipe->mutex/1 &list->lock#34 irq_context: 0 &pipe->mutex/1 &ei->socket.wq.wait irq_context: 0 &mm->mmap_lock rcu_read_lock pgd_lock irq_context: 0 &mm->mmap_lock rcu_read_lock stock_lock irq_context: 0 &mm->mmap_lock rcu_read_lock key irq_context: 0 &mm->mmap_lock rcu_read_lock pcpu_lock irq_context: 0 &mm->mmap_lock rcu_read_lock percpu_counters_lock irq_context: 0 &mm->mmap_lock rcu_read_lock pcpu_lock stock_lock irq_context: 0 (wq_completion)writeback (work_completion)(&(&wb->dwork)->work) &type->s_umount_key#43 rcu_read_lock &rq->__lock irq_context: 0 (wq_completion)writeback (work_completion)(&(&wb->dwork)->work) &type->s_umount_key#43 rcu_read_lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 sb_writers#4 &type->i_mutex_dir_key#3/1 &sb->s_type->i_mutex_key#8 jbd2_handle &sbi->s_orphan_lock pool_lock#2 irq_context: 0 sk_lock-AF_INET6 rcu_read_lock &q->lock#2 pool_lock#2 irq_context: 0 cb_lock &rdev->wiphy.mtx &wdev->mtx &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 sb_writers#4 &type->i_mutex_dir_key#3/1 &sb->s_type->i_mutex_key#8 jbd2_handle pgd_lock irq_context: 0 sb_writers#4 &type->i_mutex_dir_key#3/1 &sb->s_type->i_mutex_key#8 jbd2_handle rcu_read_lock pool_lock#2 irq_context: 0 sb_writers#4 &type->i_mutex_dir_key#3/1 &sb->s_type->i_mutex_key#8 jbd2_handle key irq_context: 0 sb_writers#4 &type->i_mutex_dir_key#3/1 &sb->s_type->i_mutex_key#8 jbd2_handle pcpu_lock irq_context: 0 sb_writers#4 &type->i_mutex_dir_key#3/1 &sb->s_type->i_mutex_key#8 jbd2_handle percpu_counters_lock irq_context: 0 (wq_completion)wg-crypt-wg0 (work_completion)(&peer->transmit_packet_work) rcu_node_0 irq_context: 0 (wq_completion)wg-crypt-wg0 (work_completion)(&peer->transmit_packet_work) &rcu_state.expedited_wq irq_context: 0 (wq_completion)wg-crypt-wg0 (work_completion)(&peer->transmit_packet_work) &rcu_state.expedited_wq &p->pi_lock irq_context: 0 (wq_completion)wg-crypt-wg0 (work_completion)(&peer->transmit_packet_work) &rcu_state.expedited_wq &p->pi_lock &rq->__lock irq_context: 0 (wq_completion)wg-crypt-wg0 (work_completion)(&peer->transmit_packet_work) &rcu_state.expedited_wq &p->pi_lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 cb_lock rtnl_mutex &rdev->wiphy.mtx &lock->wait_lock irq_context: softirq (&q->timer) &obj_hash[i].lock pool_lock irq_context: 0 (wq_completion)wg-kex-wg0#7 (work_completion)(&peer->transmit_handshake_work) &peer->endpoint_lock rcu_read_lock_bh rcu_read_lock &c->lock irq_context: 0 (wq_completion)wg-crypt-wg0#5 (work_completion)(&peer->transmit_packet_work) &peer->endpoint_lock rcu_read_lock_bh rcu_read_lock &c->lock irq_context: 0 &f->f_pos_lock sb_writers#4 &sb->s_type->i_mutex_key#8 remove_cache_srcu &cfs_rq->removed.lock irq_context: 0 cb_lock &rdev->wiphy.mtx &wdev->mtx &local->sta_mtx irq_context: 0 sb_writers#4 &sb->s_type->i_mutex_key#8 &sbi->s_writepages_rwsem remove_cache_srcu irq_context: 0 sb_writers#4 &sb->s_type->i_mutex_key#8 &sbi->s_writepages_rwsem remove_cache_srcu quarantine_lock irq_context: 0 sb_writers#4 &sb->s_type->i_mutex_key#8 &sbi->s_writepages_rwsem remove_cache_srcu &c->lock irq_context: 0 sb_writers#4 &sb->s_type->i_mutex_key#8 &sbi->s_writepages_rwsem remove_cache_srcu &n->list_lock irq_context: 0 sb_writers#4 &sb->s_type->i_mutex_key#8 &sbi->s_writepages_rwsem remove_cache_srcu rcu_read_lock pool_lock#2 irq_context: 0 sb_writers#4 &sb->s_type->i_mutex_key#8 &sbi->s_writepages_rwsem remove_cache_srcu &obj_hash[i].lock irq_context: 0 cb_lock &rdev->wiphy.mtx &wdev->mtx &local->sta_mtx rcu_read_lock rcu_node_0 irq_context: 0 cb_lock &rdev->wiphy.mtx &wdev->mtx &local->sta_mtx rcu_read_lock &rq->__lock irq_context: 0 cb_lock &rdev->wiphy.mtx &wdev->mtx &local->sta_mtx rcu_read_lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 &pipe->mutex/1 rcu_read_lock rcu_read_lock stock_lock irq_context: 0 &pipe->mutex/1 rcu_read_lock rcu_read_lock pool_lock#2 irq_context: 0 &pipe->mutex/1 rcu_read_lock rcu_read_lock &dir->lock#2 irq_context: 0 &pipe->mutex/1 rcu_read_lock rcu_read_lock &ul->lock#2 irq_context: 0 &pipe->mutex/1 rcu_read_lock rcu_read_lock rcu_read_lock pool_lock#2 irq_context: 0 &pipe->mutex/1 rcu_read_lock rcu_read_lock rcu_read_lock &obj_hash[i].lock irq_context: 0 &pipe->mutex/1 rcu_read_lock rcu_read_lock rcu_read_lock &base->lock irq_context: 0 &pipe->mutex/1 rcu_read_lock rcu_read_lock rcu_read_lock &base->lock &obj_hash[i].lock irq_context: 0 &pipe->mutex/1 rcu_read_lock rcu_read_lock rcu_read_lock rcu_node_0 irq_context: 0 &pipe->mutex/1 rcu_read_lock rcu_read_lock rcu_read_lock &rq->__lock irq_context: 0 &pipe->mutex/1 rcu_read_lock rcu_read_lock rcu_read_lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 &pipe->mutex/1 rcu_read_lock rcu_read_lock rcu_read_lock rcu_read_lock rhashtable_bucket irq_context: 0 &pipe->mutex/1 rcu_read_lock rcu_read_lock &q->lock#2 irq_context: 0 &pipe->mutex/1 rcu_read_lock rcu_read_lock &obj_hash[i].lock irq_context: 0 sb_writers#4 &type->i_mutex_dir_key#3 jbd2_handle rcu_node_0 irq_context: 0 (wq_completion)events (work_completion)(&p->wq) purge_vmap_area_lock &meta->lock irq_context: 0 (wq_completion)events (work_completion)(&p->wq) purge_vmap_area_lock kfence_freelist_lock irq_context: 0 sb_writers#4 sb_internal jbd2_handle &sbi->s_orphan_lock &cfs_rq->removed.lock irq_context: 0 sb_writers#4 sb_internal jbd2_handle &sbi->s_orphan_lock &obj_hash[i].lock irq_context: 0 sb_writers#4 &type->i_mutex_dir_key#3/1 &type->i_mutex_dir_key#3 jbd2_handle &____s->seqcount irq_context: 0 sb_writers#4 &sb->s_type->i_mutex_key#8 &sbi->s_writepages_rwsem jbd2_handle &ei->i_data_sem &rq->__lock irq_context: 0 sb_writers#4 &sb->s_type->i_mutex_key#8 &sbi->s_writepages_rwsem jbd2_handle &ei->i_data_sem &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 cb_lock rtnl_mutex fs_reclaim irq_context: 0 cb_lock rtnl_mutex fs_reclaim mmu_notifier_invalidate_range_start irq_context: 0 cb_lock rtnl_mutex &c->lock irq_context: 0 cb_lock rtnl_mutex pool_lock#2 irq_context: 0 cb_lock rtnl_mutex rcu_read_lock pool_lock#2 irq_context: 0 cb_lock rtnl_mutex rcu_read_lock rcu_read_lock_bh &obj_hash[i].lock irq_context: 0 cb_lock rtnl_mutex rcu_read_lock rcu_read_lock_bh pool_lock#2 irq_context: 0 cb_lock rtnl_mutex rlock-AF_NETLINK irq_context: 0 cb_lock rtnl_mutex rcu_read_lock &ndev->lock irq_context: 0 cb_lock rtnl_mutex &obj_hash[i].lock irq_context: 0 cb_lock rtnl_mutex nl_table_lock irq_context: 0 cb_lock rtnl_mutex nl_table_wait.lock irq_context: 0 cb_lock rtnl_mutex net_rwsem irq_context: 0 cb_lock rtnl_mutex net_rwsem &list->lock#2 irq_context: 0 cb_lock rtnl_mutex &tn->lock irq_context: 0 (wq_completion)events (work_completion)(&w->w) &base->lock irq_context: 0 (wq_completion)events (work_completion)(&w->w) &base->lock &obj_hash[i].lock irq_context: 0 cb_lock rtnl_mutex &n->list_lock irq_context: 0 cb_lock rtnl_mutex &n->list_lock &c->lock irq_context: 0 &f->f_pos_lock sb_writers#4 &sb->s_type->i_mutex_key#8 &mm->mmap_lock batched_entropy_u8.lock irq_context: 0 &f->f_pos_lock sb_writers#4 &sb->s_type->i_mutex_key#8 &mm->mmap_lock kfence_freelist_lock irq_context: softirq (&ndev->rs_timer) rcu_read_lock rcu_read_lock rcu_read_lock_bh rcu_read_lock rcu_read_lock rcu_read_lock rcu_read_lock &c->lock irq_context: 0 sb_writers#4 &type->i_mutex_dir_key#3/1 jbd2_handle &ei->i_data_sem remove_cache_srcu &cfs_rq->removed.lock irq_context: 0 &sb->s_type->i_mutex_key#10 &smc->clcsock_release_lock &rq->__lock irq_context: 0 (wq_completion)events_power_efficient &rq->__lock irq_context: 0 (wq_completion)events_power_efficient &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 dup_mmap_sem &mm->mmap_lock &mm->mmap_lock/1 &anon_vma->rwsem batched_entropy_u8.lock irq_context: 0 dup_mmap_sem &mm->mmap_lock &mm->mmap_lock/1 &anon_vma->rwsem kfence_freelist_lock irq_context: 0 &mm->mmap_lock rcu_read_lock rcu_read_lock pgd_lock irq_context: 0 &mm->mmap_lock rcu_read_lock rcu_read_lock stock_lock irq_context: 0 &mm->mmap_lock rcu_read_lock rcu_read_lock rcu_read_lock pool_lock#2 irq_context: 0 &mm->mmap_lock rcu_read_lock rcu_read_lock key irq_context: 0 &mm->mmap_lock rcu_read_lock rcu_read_lock pcpu_lock irq_context: 0 &mm->mmap_lock rcu_read_lock rcu_read_lock percpu_counters_lock irq_context: 0 &mm->mmap_lock rcu_read_lock rcu_read_lock pcpu_lock stock_lock irq_context: 0 (wq_completion)events (work_completion)(&rdev->wiphy_work) &rdev->wiphy.mtx &wdev->mtx &cfs_rq->removed.lock irq_context: 0 (wq_completion)events (work_completion)(&rdev->wiphy_work) &rdev->wiphy.mtx &wdev->mtx &rq->__lock &cfs_rq->removed.lock irq_context: softirq rcu_read_lock rcu_read_lock &pcp->lock &zone->lock &____s->seqcount irq_context: 0 &sb->s_type->i_mutex_key#10 &smc->clcsock_release_lock &c->lock irq_context: 0 &sb->s_type->i_mutex_key#10 &smc->clcsock_release_lock k-sk_lock-AF_INET k-slock-AF_INET key#24 irq_context: 0 &sb->s_type->i_mutex_key#10 &smc->clcsock_release_lock &obj_hash[i].lock pool_lock irq_context: 0 &sb->s_type->i_mutex_key#10 &smc->clcsock_release_lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 &sb->s_type->i_mutex_key#10 &smc->clcsock_release_lock &n->list_lock irq_context: softirq rcu_callback put_task_map-wait-type-override stock_lock rcu_read_lock per_cpu_ptr(&cgroup_rstat_cpu_lock, cpu) irq_context: 0 &sb->s_type->i_mutex_key#10 sk_lock-AF_INET/1 &rq->__lock irq_context: 0 &sb->s_type->i_mutex_key#10 sk_lock-AF_INET/1 &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: softirq (&peer->timer_persistent_keepalive) &zone->lock irq_context: softirq (&p->timer) irq_context: softirq (&p->timer) &br->multicast_lock irq_context: softirq (&p->timer) &br->multicast_lock pool_lock#2 irq_context: softirq (&p->timer) &br->multicast_lock &c->lock irq_context: softirq (&p->timer) &br->multicast_lock &n->list_lock irq_context: softirq (&p->timer) &br->multicast_lock &n->list_lock &c->lock irq_context: softirq (&p->timer) &br->multicast_lock &dir->lock#2 irq_context: softirq (&p->timer) &br->multicast_lock deferred_lock irq_context: softirq (&p->timer) &br->multicast_lock rcu_read_lock &pool->lock irq_context: softirq (&p->timer) &br->multicast_lock rcu_read_lock &pool->lock &obj_hash[i].lock irq_context: softirq (&p->timer) &br->multicast_lock rcu_read_lock &pool->lock &p->pi_lock irq_context: softirq (&p->timer) &br->multicast_lock nl_table_lock irq_context: softirq (&p->timer) &br->multicast_lock &obj_hash[i].lock irq_context: softirq (&p->timer) &br->multicast_lock nl_table_wait.lock irq_context: softirq (&p->timer) &br->multicast_lock rcu_read_lock &pool->lock pool_lock#2 irq_context: softirq (&p->timer) &br->multicast_lock &base->lock irq_context: softirq (&p->timer) &br->multicast_lock &base->lock &obj_hash[i].lock irq_context: 0 (wq_completion)events_long (work_completion)(&br->mcast_gc_work) &obj_hash[i].lock pool_lock irq_context: softirq (&mp->timer) &br->multicast_lock rcu_read_lock rcu_read_lock &pool->lock &p->pi_lock &rq->__lock irq_context: softirq (&mp->timer) &br->multicast_lock rcu_read_lock rcu_read_lock &pool->lock &p->pi_lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)wg-kex-wg1#11 (work_completion)(&({ do { const void *__vpp_verify = (typeof((worker) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __asm__ ("" : "=r"(__ptr) : "0"((typeof(*((worker))) *)((worker)))); (typeof((typeof(*((worker))) *)((worker)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); })->work) &peer->endpoint_lock rcu_read_lock_bh &obj_hash[i].lock irq_context: 0 (wq_completion)wg-kex-wg2#12 (work_completion)(&({ do { const void *__vpp_verify = (typeof((worker) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __asm__ ("" : "=r"(__ptr) : "0"((typeof(*((worker))) *)((worker)))); (typeof((typeof(*((worker))) *)((worker)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); })->work) &handshake->lock rcu_read_lock_bh &peer->keypairs.keypair_update_lock &table->lock#2 irq_context: 0 (wq_completion)wg-kex-wg2#12 (work_completion)(&({ do { const void *__vpp_verify = (typeof((worker) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __asm__ ("" : "=r"(__ptr) : "0"((typeof(*((worker))) *)((worker)))); (typeof((typeof(*((worker))) *)((worker)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); })->work) &handshake->lock rcu_read_lock_bh &peer->keypairs.keypair_update_lock &obj_hash[i].lock irq_context: 0 cb_lock &rdev->wiphy.mtx &local->mtx remove_cache_srcu irq_context: 0 cb_lock &rdev->wiphy.mtx &local->mtx remove_cache_srcu quarantine_lock irq_context: 0 cb_lock &rdev->wiphy.mtx &local->mtx remove_cache_srcu &c->lock irq_context: 0 cb_lock &rdev->wiphy.mtx &local->mtx remove_cache_srcu &n->list_lock irq_context: 0 cb_lock &rdev->wiphy.mtx &local->mtx remove_cache_srcu rcu_read_lock pool_lock#2 irq_context: 0 cb_lock &rdev->wiphy.mtx &local->mtx remove_cache_srcu &obj_hash[i].lock irq_context: 0 cb_lock &rdev->wiphy.mtx &local->mtx remove_cache_srcu &rq->__lock irq_context: 0 rcu_read_lock &f->f_owner.lock irq_context: 0 &pipe->mutex/1 sk_lock-AF_INET6 rcu_read_lock rcu_read_lock rcu_read_lock rcu_node_0 irq_context: 0 &pipe->mutex/1 sk_lock-AF_INET6 rcu_read_lock rcu_read_lock rcu_read_lock &rq->__lock irq_context: 0 &pipe->mutex/1 sk_lock-AF_INET6 rcu_read_lock rcu_read_lock rcu_read_lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 sb_writers#3 &sb->s_type->i_mutex_key#9 mmu_notifier_invalidate_range_start &rq->__lock irq_context: 0 &pipe->mutex/1 &sighand->siglock batched_entropy_u8.lock irq_context: 0 &pipe->mutex/1 &sighand->siglock kfence_freelist_lock irq_context: 0 tasklist_lock &sighand->siglock &meta->lock irq_context: 0 vlan_ioctl_mutex rtnl_mutex rcu_read_lock &tb->tb6_lock rcu_read_lock &c->lock irq_context: 0 &pipe->mutex/1 &pipe->mutex#2/2 irq_context: 0 &pipe->mutex#2/2 irq_context: 0 vlan_ioctl_mutex rtnl_mutex rcu_read_lock &obj_hash[i].lock pool_lock irq_context: 0 vlan_ioctl_mutex rtnl_mutex rcu_read_lock &net->sctp.local_addr_lock &net->sctp.addr_wq_lock &c->lock irq_context: 0 vlan_ioctl_mutex rtnl_mutex rcu_read_lock &net->sctp.local_addr_lock &net->sctp.addr_wq_lock &n->list_lock irq_context: 0 vlan_ioctl_mutex rtnl_mutex rcu_read_lock &net->sctp.local_addr_lock &net->sctp.addr_wq_lock &n->list_lock &c->lock irq_context: 0 vlan_ioctl_mutex rtnl_mutex lweventlist_lock batched_entropy_u8.lock irq_context: 0 vlan_ioctl_mutex rtnl_mutex lweventlist_lock kfence_freelist_lock irq_context: 0 vlan_ioctl_mutex rtnl_mutex remove_cache_srcu pool_lock#2 irq_context: 0 &f->f_pos_lock sb_writers#4 &sb->s_type->i_mutex_key#8 jbd2_handle rcu_read_lock &rq->__lock &cfs_rq->removed.lock irq_context: 0 vlan_ioctl_mutex &dir->lock#2 &meta->lock irq_context: 0 vlan_ioctl_mutex &dir->lock#2 kfence_freelist_lock irq_context: 0 sb_writers#4 sb_internal jbd2_handle &ei->i_data_sem quarantine_lock irq_context: 0 (wq_completion)events (work_completion)(&rdev->wiphy_work) &rdev->wiphy.mtx &wdev->mtx rcu_read_lock rcu_read_lock &sta->rate_ctrl_lock batched_entropy_u8.lock irq_context: 0 (wq_completion)events (work_completion)(&rdev->wiphy_work) &rdev->wiphy.mtx &wdev->mtx rcu_read_lock rcu_read_lock &sta->rate_ctrl_lock kfence_freelist_lock irq_context: 0 cb_lock &rdev->wiphy.mtx &cfs_rq->removed.lock irq_context: 0 sb_writers#4 &sb->s_type->i_mutex_key#8 mapping.invalidate_lock jbd2_handle &ei->i_data_sem remove_cache_srcu &rq->__lock irq_context: 0 sb_writers#4 &sb->s_type->i_mutex_key#8 mapping.invalidate_lock jbd2_handle &ei->i_data_sem remove_cache_srcu &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 &f->f_pos_lock sb_writers#4 &sb->s_type->i_mutex_key#8 jbd2_handle pgd_lock irq_context: 0 &f->f_pos_lock sb_writers#4 &sb->s_type->i_mutex_key#8 jbd2_handle key irq_context: 0 &f->f_pos_lock sb_writers#4 &sb->s_type->i_mutex_key#8 jbd2_handle pcpu_lock irq_context: 0 &f->f_pos_lock sb_writers#4 &sb->s_type->i_mutex_key#8 jbd2_handle percpu_counters_lock irq_context: 0 (wq_completion)events (work_completion)(&rdev->wiphy_work) &rdev->wiphy.mtx &wdev->mtx rcu_read_lock rcu_read_lock &sta->rate_ctrl_lock krc.lock &obj_hash[i].lock irq_context: 0 (wq_completion)events (work_completion)(&rdev->wiphy_work) &rdev->wiphy.mtx &wdev->mtx rcu_read_lock rcu_read_lock &sta->rate_ctrl_lock krc.lock &base->lock irq_context: 0 (wq_completion)events (work_completion)(&rdev->wiphy_work) &rdev->wiphy.mtx &wdev->mtx rcu_read_lock rcu_read_lock &sta->rate_ctrl_lock krc.lock &base->lock &obj_hash[i].lock irq_context: 0 cb_lock &rdev->wiphy.mtx &pcp->lock &zone->lock irq_context: 0 &f->f_pos_lock sb_writers#3 (console_sem).lock irq_context: 0 &f->f_pos_lock sb_writers#3 console_lock console_srcu console_owner_lock irq_context: 0 &f->f_pos_lock sb_writers#3 console_lock console_srcu console_owner irq_context: 0 &f->f_pos_lock sb_writers#3 console_lock console_srcu console_owner &port_lock_key irq_context: 0 (wq_completion)events_unbound (work_completion)(&sub_info->work) tasklist_lock &sighand->siglock &(&sig->stats_lock)->lock &____s->seqcount#5 pidmap_lock &obj_hash[i].lock irq_context: 0 (wq_completion)events_unbound (work_completion)(&sub_info->work) tasklist_lock &sighand->siglock &(&sig->stats_lock)->lock &____s->seqcount#5 pidmap_lock pool_lock#2 irq_context: 0 &f->f_pos_lock sb_writers#3 console_lock console_srcu console_owner console_owner_lock irq_context: 0 &f->f_pos_lock sb_writers#3 &rq->__lock irq_context: 0 &f->f_pos_lock sb_writers#3 &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: softirq rcu_read_lock rcu_read_lock batched_entropy_u8.lock crngs.lock irq_context: 0 kn->active#60 fs_reclaim irq_context: 0 kn->active#60 fs_reclaim mmu_notifier_invalidate_range_start irq_context: 0 kn->active#60 stock_lock irq_context: 0 kn->active#60 &kernfs_locks->open_file_mutex[count] irq_context: 0 kn->active#60 &kernfs_locks->open_file_mutex[count] fs_reclaim irq_context: 0 kn->active#60 &kernfs_locks->open_file_mutex[count] fs_reclaim mmu_notifier_invalidate_range_start irq_context: 0 kn->active#60 &c->lock irq_context: 0 tomoyo_ss rcu_read_lock rcu_read_lock &rq->__lock &cfs_rq->removed.lock irq_context: 0 sb_writers#5 &type->i_mutex_dir_key#5/1 &rcu_state.gp_wq irq_context: 0 sb_writers#5 &type->i_mutex_dir_key#5/1 &rcu_state.gp_wq &p->pi_lock irq_context: 0 sb_writers#5 &type->i_mutex_dir_key#5/1 &rcu_state.gp_wq &p->pi_lock &rq->__lock irq_context: 0 sb_writers#5 &type->i_mutex_dir_key#5/1 &rcu_state.gp_wq &p->pi_lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 kn->active#60 &rq->__lock irq_context: 0 kn->active#60 &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 kn->active#60 &kernfs_locks->open_file_mutex[count] &c->lock irq_context: softirq (&mp->timer) &br->multicast_lock pool_lock#2 irq_context: softirq (&mp->timer) &br->multicast_lock &dir->lock#2 irq_context: softirq (&mp->timer) &br->multicast_lock deferred_lock irq_context: softirq (&mp->timer) &br->multicast_lock &c->lock irq_context: softirq (&mp->timer) &br->multicast_lock &n->list_lock irq_context: softirq (&mp->timer) &br->multicast_lock &n->list_lock &c->lock irq_context: softirq (&mp->timer) &br->multicast_lock nl_table_lock irq_context: softirq (&mp->timer) &br->multicast_lock &obj_hash[i].lock irq_context: softirq (&mp->timer) &br->multicast_lock nl_table_wait.lock irq_context: 0 (wq_completion)wg-kex-wg1 (work_completion)(&({ do { const void *__vpp_verify = (typeof((worker) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __asm__ ("" : "=r"(__ptr) : "0"((typeof(*((worker))) *)((worker)))); (typeof((typeof(*((worker))) *)((worker)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); })->work) &c->lock irq_context: 0 (wq_completion)wg-kex-wg2#2 (work_completion)(&({ do { const void *__vpp_verify = (typeof((worker) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __asm__ ("" : "=r"(__ptr) : "0"((typeof(*((worker))) *)((worker)))); (typeof((typeof(*((worker))) *)((worker)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); })->work) &wg->static_identity.lock &rq->__lock irq_context: 0 (wq_completion)wg-kex-wg2#2 (work_completion)(&({ do { const void *__vpp_verify = (typeof((worker) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __asm__ ("" : "=r"(__ptr) : "0"((typeof(*((worker))) *)((worker)))); (typeof((typeof(*((worker))) *)((worker)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); })->work) &wg->static_identity.lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)wg-kex-wg2#2 (work_completion)(&({ do { const void *__vpp_verify = (typeof((worker) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __asm__ ("" : "=r"(__ptr) : "0"((typeof(*((worker))) *)((worker)))); (typeof((typeof(*((worker))) *)((worker)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); })->work) &handshake->lock rcu_read_lock_bh &peer->keypairs.keypair_update_lock &table->lock#2 irq_context: 0 (wq_completion)wg-kex-wg2#2 (work_completion)(&({ do { const void *__vpp_verify = (typeof((worker) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __asm__ ("" : "=r"(__ptr) : "0"((typeof(*((worker))) *)((worker)))); (typeof((typeof(*((worker))) *)((worker)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); })->work) &handshake->lock rcu_read_lock_bh &peer->keypairs.keypair_update_lock &obj_hash[i].lock irq_context: 0 &sb->s_type->i_mutex_key#10 &net->sctp.addr_wq_lock slock-AF_INET6/1 rcu_read_lock pool_lock#2 irq_context: 0 &sbi->s_writepages_rwsem jbd2_handle &mapping->i_mmap_rwsem mmu_notifier_invalidate_range_start &rq->__lock irq_context: 0 &sbi->s_writepages_rwsem jbd2_handle &mapping->i_mmap_rwsem mmu_notifier_invalidate_range_start &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 &sbi->s_writepages_rwsem jbd2_handle &cfs_rq->removed.lock irq_context: 0 (wq_completion)wg-kex-wg0#7 (work_completion)(&peer->transmit_handshake_work) &wg->static_identity.lock &handshake->lock rcu_read_lock_bh batched_entropy_u32.lock crngs.lock irq_context: 0 (wq_completion)wg-kex-wg0#7 (work_completion)(&peer->transmit_handshake_work) &n->list_lock irq_context: 0 (wq_completion)wg-kex-wg0#7 (work_completion)(&peer->transmit_handshake_work) &n->list_lock &c->lock irq_context: 0 (wq_completion)wg-kex-wg0#7 (work_completion)(&peer->transmit_handshake_work) &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)bat_events (work_completion)(&(&bat_priv->tt.work)->work) &rq->__lock &cfs_rq->removed.lock irq_context: 0 &pipe->mutex/1 &sighand->siglock &p->pi_lock &cfs_rq->removed.lock irq_context: 0 sb_writers#4 &sb->s_type->i_mutex_key#8 &sbi->s_writepages_rwsem batched_entropy_u8.lock irq_context: 0 sb_writers#4 &sb->s_type->i_mutex_key#8 &sbi->s_writepages_rwsem kfence_freelist_lock irq_context: 0 sb_writers#4 &sb->s_type->i_mutex_key#8 &sbi->s_writepages_rwsem &ei->i_data_sem &meta->lock irq_context: 0 sb_writers#4 &sb->s_type->i_mutex_key#8 &sbi->s_writepages_rwsem &ei->i_data_sem kfence_freelist_lock irq_context: 0 sk_lock-AF_INET6 rcu_read_lock rcu_read_lock rcu_read_lock rcu_read_lock &pool->lock irq_context: 0 sk_lock-AF_INET6 rcu_read_lock rcu_read_lock rcu_read_lock rcu_read_lock &pool->lock &obj_hash[i].lock irq_context: 0 sk_lock-AF_INET6 rcu_read_lock rcu_read_lock rcu_read_lock rcu_read_lock &pool->lock pool_lock#2 irq_context: 0 sk_lock-AF_INET6 rcu_read_lock rcu_read_lock rcu_read_lock rcu_read_lock &pool->lock &p->pi_lock irq_context: 0 sk_lock-AF_INET6 rcu_read_lock rcu_read_lock rcu_read_lock rcu_read_lock &pool->lock &p->pi_lock &rq->__lock irq_context: 0 sk_lock-AF_INET6 rcu_read_lock rcu_read_lock rcu_read_lock rcu_read_lock &pool->lock &p->pi_lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 sb_writers#4 &sb->s_type->i_mutex_key#8 mapping.invalidate_lock jbd2_handle &sbi->s_orphan_lock rcu_read_lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 sb_writers#4 &sb->s_type->i_mutex_key#8 mapping.invalidate_lock jbd2_handle &sbi->s_orphan_lock &lock->wait_lock irq_context: 0 sb_writers#4 &sb->s_type->i_mutex_key#8 mapping.invalidate_lock jbd2_handle &sbi->s_orphan_lock rcu_read_lock &rcu_state.gp_wq irq_context: 0 sb_writers#4 &sb->s_type->i_mutex_key#8 mapping.invalidate_lock jbd2_handle &sbi->s_orphan_lock rcu_read_lock &rcu_state.gp_wq &p->pi_lock irq_context: 0 sb_writers#4 &sb->s_type->i_mutex_key#8 mapping.invalidate_lock jbd2_handle &sbi->s_orphan_lock rcu_read_lock &rcu_state.gp_wq &p->pi_lock &rq->__lock irq_context: 0 sb_writers#4 &sb->s_type->i_mutex_key#8 mapping.invalidate_lock jbd2_handle &sbi->s_orphan_lock rcu_read_lock &rcu_state.gp_wq &p->pi_lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 sk_lock-AF_INET6 rcu_read_lock rcu_read_lock fill_pool_map-wait-type-override rcu_node_0 irq_context: 0 sk_lock-AF_INET6 rcu_read_lock rcu_read_lock fill_pool_map-wait-type-override &rq->__lock irq_context: 0 sk_lock-AF_INET6 rcu_read_lock rcu_read_lock fill_pool_map-wait-type-override &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 sb_writers#4 &sb->s_type->i_mutex_key#8 mapping.invalidate_lock jbd2_handle &ei->i_data_sem remove_cache_srcu pool_lock#2 irq_context: 0 &sig->cred_guard_mutex tomoyo_ss &mm->mmap_lock rcu_read_lock rcu_node_0 irq_context: 0 &sig->cred_guard_mutex tomoyo_ss &mm->mmap_lock rcu_read_lock &rq->__lock irq_context: 0 &sig->cred_guard_mutex tomoyo_ss &mm->mmap_lock rcu_read_lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 sb_writers#4 &sb->s_type->i_mutex_key#8 mapping.invalidate_lock &mapping->i_mmap_rwsem &cfs_rq->removed.lock irq_context: 0 sb_writers#4 &sb->s_type->i_mutex_key#8 mapping.invalidate_lock &mapping->i_mmap_rwsem pool_lock#2 irq_context: 0 &iint->mutex ima_extend_list_mutex &____s->seqcount#2 irq_context: 0 sb_writers#4 &sb->s_type->i_mutex_key#8 mapping.invalidate_lock &sem->wait_lock irq_context: 0 sb_writers#4 &sb->s_type->i_mutex_key#8 mapping.invalidate_lock &p->pi_lock irq_context: 0 sb_writers#4 &sb->s_type->i_mutex_key#8 mapping.invalidate_lock &p->pi_lock &cfs_rq->removed.lock irq_context: 0 sb_writers#4 &sb->s_type->i_mutex_key#8 mapping.invalidate_lock &p->pi_lock &rq->__lock irq_context: 0 sb_writers#4 &sb->s_type->i_mutex_key#8 mapping.invalidate_lock &p->pi_lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 &vma->vm_lock->lock rcu_read_lock &rq->__lock &cfs_rq->removed.lock irq_context: 0 kn->active#5 &cfs_rq->removed.lock irq_context: 0 kn->active#5 &obj_hash[i].lock irq_context: 0 sb_writers#4 &type->i_mutex_dir_key#3/1 remove_cache_srcu rcu_read_lock rcu_read_lock rcu_node_0 irq_context: 0 sb_writers#4 &type->i_mutex_dir_key#3/1 remove_cache_srcu rcu_read_lock rcu_read_lock &rq->__lock irq_context: 0 sb_writers#4 &type->i_mutex_dir_key#3/1 remove_cache_srcu rcu_read_lock rcu_read_lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 sk_lock-AF_INET6 rcu_read_lock &meta->lock irq_context: 0 sb_writers#4 &sb->s_type->i_mutex_key#8 mapping.invalidate_lock jbd2_handle rcu_read_lock rcu_node_0 irq_context: 0 sb_writers#4 &type->i_mutex_dir_key#3/1 sb_internal jbd2_handle &sbi->s_orphan_lock &lock->wait_lock irq_context: 0 sb_writers#4 &type->i_mutex_dir_key#3/1 sb_internal jbd2_handle pool_lock#2 irq_context: 0 &pipe->mutex/1 rcu_read_lock &rcu_state.gp_wq irq_context: 0 &pipe->mutex/1 rcu_read_lock &rcu_state.gp_wq &p->pi_lock irq_context: 0 &pipe->mutex/1 rcu_read_lock &rcu_state.gp_wq &p->pi_lock &rq->__lock irq_context: 0 &pipe->mutex/1 rcu_read_lock &rcu_state.gp_wq &p->pi_lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 cb_lock &rdev->wiphy.mtx &local->mtx rcu_read_lock &cfs_rq->removed.lock irq_context: 0 cb_lock &rdev->wiphy.mtx &local->mtx rcu_read_lock &obj_hash[i].lock irq_context: 0 cb_lock &rdev->wiphy.mtx &local->mtx rcu_read_lock pool_lock#2 irq_context: 0 &mm->mmap_lock &anon_vma->rwsem stock_lock rcu_read_lock per_cpu_ptr(&cgroup_rstat_cpu_lock, cpu) irq_context: 0 sb_writers#4 &type->i_mutex_dir_key#3 &journal->j_wait_transaction_locked irq_context: 0 &f->f_pos_lock sb_writers#10 &rq->__lock irq_context: 0 &f->f_pos_lock sb_writers#10 &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 sb_writers#4 &type->i_mutex_dir_key#3 jbd2_handle &obj_hash[i].lock irq_context: 0 (wq_completion)rcu_gp (work_completion)(&rew->rew_work) rcu_read_lock &rq->__lock &cfs_rq->removed.lock irq_context: 0 sb_writers#8 &root->kernfs_iattr_rwsem &sem->wait_lock irq_context: 0 (wq_completion)wg-crypt-wg0#3 (work_completion)(&peer->transmit_packet_work) rcu_node_0 irq_context: 0 sb_writers#4 &sb->s_type->i_mutex_key#8 &sbi->s_writepages_rwsem &rnp->exp_wq[0] irq_context: softirq (&ndev->rs_timer) &ndev->lock batched_entropy_u32.lock crngs.lock base_crng.lock irq_context: 0 sb_writers#4 &type->i_mutex_dir_key#3/1 jbd2_handle batched_entropy_u32.lock crngs.lock base_crng.lock irq_context: 0 (wq_completion)wg-crypt-wg0#5 (work_completion)(&peer->transmit_packet_work) batched_entropy_u8.lock crngs.lock irq_context: 0 (wq_completion)bat_events (work_completion)(&(&forw_packet_aggr->delayed_work)->work) &rcu_state.expedited_wq &p->pi_lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 pernet_ops_rwsem rtnl_mutex mmu_notifier_invalidate_range_start &rq->__lock irq_context: 0 pernet_ops_rwsem rtnl_mutex mmu_notifier_invalidate_range_start &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 pernet_ops_rwsem rtnl_mutex cpu_hotplug_lock rcu_read_lock &pool->lock irq_context: 0 pernet_ops_rwsem rtnl_mutex cpu_hotplug_lock rcu_read_lock &pool->lock &obj_hash[i].lock irq_context: 0 pernet_ops_rwsem rtnl_mutex cpu_hotplug_lock rcu_read_lock &pool->lock &p->pi_lock irq_context: 0 pernet_ops_rwsem rtnl_mutex cpu_hotplug_lock rcu_read_lock &pool->lock &p->pi_lock &rq->__lock irq_context: 0 pernet_ops_rwsem rtnl_mutex cpu_hotplug_lock rcu_read_lock &pool->lock &p->pi_lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 pernet_ops_rwsem rtnl_mutex cpu_hotplug_lock (work_completion)(flush) irq_context: 0 pernet_ops_rwsem rtnl_mutex cpu_hotplug_lock &x->wait#10 irq_context: 0 pernet_ops_rwsem rtnl_mutex cpu_hotplug_lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)events_highpri (work_completion)(flush) irq_context: 0 (wq_completion)events_highpri (work_completion)(flush) &list->lock#5 irq_context: 0 (wq_completion)events_highpri (work_completion)(&barr->work) irq_context: 0 (wq_completion)events_highpri (work_completion)(&barr->work) &x->wait#10 irq_context: 0 (wq_completion)events_highpri (work_completion)(&barr->work) &x->wait#10 &p->pi_lock irq_context: 0 (wq_completion)events_highpri (work_completion)(&barr->work) &x->wait#10 &p->pi_lock &rq->__lock irq_context: 0 (wq_completion)events_highpri (work_completion)(&barr->work) &x->wait#10 &p->pi_lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 pernet_ops_rwsem rtnl_mutex cpu_hotplug_lock &obj_hash[i].lock irq_context: 0 (wq_completion)events (shepherd).work &rq->__lock irq_context: 0 (wq_completion)events (shepherd).work &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: softirq (&icsk->icsk_delack_timer) slock-AF_INET rcu_read_lock rcu_read_lock rcu_read_lock_bh dev->qdisc_tx_busylock ?: &qdisc_tx_busylock _xmit_ETHER#2 rcu_read_lock &____s->seqcount#2 irq_context: 0 pernet_ops_rwsem nf_ct_proto_mutex defrag6_mutex nf_hook_mutex &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 pernet_ops_rwsem &ht->mutex &rq->__lock irq_context: 0 pernet_ops_rwsem &ht->mutex &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)gid-cache-wq (work_completion)(&ndev_work->work) devices_rwsem rcu_read_lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 pernet_ops_rwsem &fn->fou_lock &rq->__lock irq_context: 0 pernet_ops_rwsem &fn->fou_lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 &f->f_pos_lock &type->i_mutex_dir_key#3 sb_writers#4 jbd2_handle &journal->j_state_lock &journal->j_wait_commit &p->pi_lock &rq->__lock &cfs_rq->removed.lock irq_context: 0 (wq_completion)wg-crypt-wg2#3 (work_completion)(&peer->transmit_packet_work) &peer->endpoint_lock rcu_read_lock_bh batched_entropy_u32.lock crngs.lock irq_context: 0 (wq_completion)wg-kex-wg2#5 (work_completion)(&peer->transmit_handshake_work) &base->lock irq_context: 0 (wq_completion)wg-kex-wg2#5 (work_completion)(&peer->transmit_handshake_work) &base->lock &obj_hash[i].lock irq_context: 0 (wq_completion)wg-kex-wg1#6 (work_completion)(&({ do { const void *__vpp_verify = (typeof((worker) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __asm__ ("" : "=r"(__ptr) : "0"((typeof(*((worker))) *)((worker)))); (typeof((typeof(*((worker))) *)((worker)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); })->work) &handshake->lock rcu_read_lock_bh &peer->keypairs.keypair_update_lock &table->lock#2 irq_context: 0 (wq_completion)wg-kex-wg1#6 (work_completion)(&({ do { const void *__vpp_verify = (typeof((worker) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __asm__ ("" : "=r"(__ptr) : "0"((typeof(*((worker))) *)((worker)))); (typeof((typeof(*((worker))) *)((worker)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); })->work) &handshake->lock rcu_read_lock_bh &peer->keypairs.keypair_update_lock &obj_hash[i].lock irq_context: 0 (wq_completion)wg-kex-wg1#6 (work_completion)(&({ do { const void *__vpp_verify = (typeof((worker) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __asm__ ("" : "=r"(__ptr) : "0"((typeof(*((worker))) *)((worker)))); (typeof((typeof(*((worker))) *)((worker)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); })->work) &handshake->lock rcu_read_lock_bh &peer->keypairs.keypair_update_lock &obj_hash[i].lock pool_lock irq_context: 0 (wq_completion)wg-kex-wg1#6 (work_completion)(&({ do { const void *__vpp_verify = (typeof((worker) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __asm__ ("" : "=r"(__ptr) : "0"((typeof(*((worker))) *)((worker)))); (typeof((typeof(*((worker))) *)((worker)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); })->work) &handshake->lock &rq->__lock irq_context: 0 (wq_completion)wg-kex-wg1#6 (work_completion)(&({ do { const void *__vpp_verify = (typeof((worker) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __asm__ ("" : "=r"(__ptr) : "0"((typeof(*((worker))) *)((worker)))); (typeof((typeof(*((worker))) *)((worker)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); })->work) &handshake->lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)wg-kex-wg2#6 (work_completion)(&({ do { const void *__vpp_verify = (typeof((worker) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __asm__ ("" : "=r"(__ptr) : "0"((typeof(*((worker))) *)((worker)))); (typeof((typeof(*((worker))) *)((worker)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); })->work) &handshake->lock rcu_read_lock_bh &peer->keypairs.keypair_update_lock &table->lock#2 irq_context: 0 (wq_completion)wg-kex-wg2#6 (work_completion)(&({ do { const void *__vpp_verify = (typeof((worker) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __asm__ ("" : "=r"(__ptr) : "0"((typeof(*((worker))) *)((worker)))); (typeof((typeof(*((worker))) *)((worker)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); })->work) &handshake->lock rcu_read_lock_bh &peer->keypairs.keypair_update_lock &obj_hash[i].lock irq_context: 0 (wq_completion)wg-kex-wg2#6 (work_completion)(&({ do { const void *__vpp_verify = (typeof((worker) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __asm__ ("" : "=r"(__ptr) : "0"((typeof(*((worker))) *)((worker)))); (typeof((typeof(*((worker))) *)((worker)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); })->work) &handshake->lock &rq->__lock irq_context: 0 (wq_completion)wg-kex-wg2#6 (work_completion)(&({ do { const void *__vpp_verify = (typeof((worker) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __asm__ ("" : "=r"(__ptr) : "0"((typeof(*((worker))) *)((worker)))); (typeof((typeof(*((worker))) *)((worker)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); })->work) &c->lock irq_context: 0 (wq_completion)wg-crypt-wg2 (work_completion)(&peer->transmit_packet_work) &peer->endpoint_lock rcu_read_lock_bh rcu_read_lock &obj_hash[i].lock irq_context: 0 (wq_completion)wg-crypt-wg2 (work_completion)(&peer->transmit_packet_work) &peer->endpoint_lock rcu_read_lock_bh rcu_read_lock &c->lock irq_context: 0 (wq_completion)wg-crypt-wg2 (work_completion)(&peer->transmit_packet_work) &peer->endpoint_lock rcu_read_lock_bh rcu_read_lock &tbl->lock irq_context: 0 (wq_completion)wg-crypt-wg2 (work_completion)(&peer->transmit_packet_work) &peer->endpoint_lock rcu_read_lock_bh rcu_read_lock &n->lock irq_context: 0 (wq_completion)wg-crypt-wg2 (work_completion)(&peer->transmit_packet_work) &peer->endpoint_lock rcu_read_lock_bh rcu_read_lock &____s->seqcount#9 irq_context: 0 (wq_completion)wg-kex-wg0#2 (work_completion)(&({ do { const void *__vpp_verify = (typeof((worker) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __asm__ ("" : "=r"(__ptr) : "0"((typeof(*((worker))) *)((worker)))); (typeof((typeof(*((worker))) *)((worker)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); })->work) &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 &f->f_pos_lock sb_writers#4 &sb->s_type->i_mutex_key#8 &ei->i_data_sem mmu_notifier_invalidate_range_start &rq->__lock irq_context: 0 &f->f_pos_lock sb_writers#4 &sb->s_type->i_mutex_key#8 &ei->i_data_sem mmu_notifier_invalidate_range_start &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)wg-kex-wg2#9 &rq->__lock irq_context: 0 sb_writers#4 &type->i_mutex_dir_key#3/1 remove_cache_srcu &rq->__lock &cfs_rq->removed.lock irq_context: 0 (wq_completion)events_power_efficient (gc_work).work &rq->__lock &cfs_rq->removed.lock irq_context: 0 sk_lock-AF_X25 irq_context: 0 sk_lock-AF_X25 slock-AF_X25 irq_context: 0 slock-AF_X25 irq_context: 0 (wq_completion)wg-crypt-wg1 (work_completion)(&({ do { const void *__vpp_verify = (typeof((worker) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __asm__ ("" : "=r"(__ptr) : "0"((typeof(*((worker))) *)((worker)))); (typeof((typeof(*((worker))) *)((worker)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); })->work) &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 &f->f_pos_lock sb_writers#4 &sb->s_type->i_mutex_key#8 jbd2_handle rcu_read_lock pgd_lock irq_context: 0 &f->f_pos_lock sb_writers#4 &sb->s_type->i_mutex_key#8 jbd2_handle rcu_read_lock stock_lock irq_context: 0 &f->f_pos_lock sb_writers#4 &sb->s_type->i_mutex_key#8 jbd2_handle rcu_read_lock key irq_context: 0 &f->f_pos_lock sb_writers#4 &sb->s_type->i_mutex_key#8 jbd2_handle rcu_read_lock pcpu_lock irq_context: 0 &f->f_pos_lock sb_writers#4 &sb->s_type->i_mutex_key#8 jbd2_handle rcu_read_lock percpu_counters_lock irq_context: 0 &f->f_pos_lock sb_writers#4 &sb->s_type->i_mutex_key#8 jbd2_handle rcu_read_lock pcpu_lock stock_lock irq_context: 0 sb_writers#4 &type->i_mutex_dir_key#3/1 &sb->s_type->i_mutex_key#8 &pcp->lock &zone->lock &____s->seqcount irq_context: 0 (wq_completion)bat_events (work_completion)(&(&forw_packet_aggr->delayed_work)->work) &hard_iface->bat_iv.ogm_buff_mutex &bat_priv->tt.commit_lock &bat_priv->softif_vlan_list_lock batched_entropy_u8.lock irq_context: 0 (wq_completion)bat_events (work_completion)(&(&forw_packet_aggr->delayed_work)->work) &hard_iface->bat_iv.ogm_buff_mutex &bat_priv->tt.commit_lock &bat_priv->softif_vlan_list_lock kfence_freelist_lock irq_context: 0 (wq_completion)bat_events (work_completion)(&(&forw_packet_aggr->delayed_work)->work) &hard_iface->bat_iv.ogm_buff_mutex &bat_priv->tt.commit_lock &meta->lock irq_context: 0 (wq_completion)bat_events (work_completion)(&(&forw_packet_aggr->delayed_work)->work) &hard_iface->bat_iv.ogm_buff_mutex &bat_priv->tt.commit_lock kfence_freelist_lock irq_context: softirq rcu_callback rlock-AF_CAN irq_context: softirq rcu_callback elock-AF_CAN irq_context: 0 hashlimit_mutex &____s->seqcount#2 irq_context: 0 sk_lock-AF_INET6 rcu_read_lock rcu_read_lock &obj_hash[i].lock pool_lock irq_context: 0 sb_writers#11 &type->i_mutex_dir_key#7 &c->lock irq_context: 0 sb_writers#11 &type->i_mutex_dir_key#7 &n->list_lock irq_context: 0 sb_writers#11 &type->i_mutex_dir_key#7 &n->list_lock &c->lock irq_context: 0 sb_writers#11 &type->i_mutex_dir_key#7 &rq->__lock irq_context: 0 sb_writers#11 &type->i_mutex_dir_key#7 &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 hashlimit_mutex &n->list_lock irq_context: 0 hashlimit_mutex &n->list_lock &c->lock irq_context: softirq (&icsk->icsk_retransmit_timer) slock-AF_INET rcu_read_lock rcu_read_lock rcu_read_lock_bh dev->qdisc_tx_busylock ?: &qdisc_tx_busylock _xmit_ETHER#2 rcu_read_lock &c->lock irq_context: 0 (wq_completion)events_unbound (work_completion)(&(&kfence_timer)->work) cpu_hotplug_lock jump_label_mutex text_mutex pool_lock#2 irq_context: 0 (wq_completion)events (work_completion)(&nlk->work) &obj_hash[i].lock pool_lock irq_context: 0 (wq_completion)events (work_completion)(&nlk->work) &base->lock irq_context: 0 (wq_completion)events (work_completion)(&nlk->work) &base->lock &obj_hash[i].lock irq_context: 0 (wq_completion)wg-kex-wg2#7 (work_completion)(&peer->transmit_handshake_work) &c->lock irq_context: 0 (wq_completion)wg-kex-wg2#7 (work_completion)(&peer->transmit_handshake_work) &n->list_lock irq_context: 0 (wq_completion)wg-kex-wg2#7 (work_completion)(&peer->transmit_handshake_work) &n->list_lock &c->lock irq_context: 0 (wq_completion)wg-kex-wg0#12 (work_completion)(&({ do { const void *__vpp_verify = (typeof((worker) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __asm__ ("" : "=r"(__ptr) : "0"((typeof(*((worker))) *)((worker)))); (typeof((typeof(*((worker))) *)((worker)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); })->work) &c->lock irq_context: 0 (wq_completion)wg-kex-wg0#12 (work_completion)(&({ do { const void *__vpp_verify = (typeof((worker) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __asm__ ("" : "=r"(__ptr) : "0"((typeof(*((worker))) *)((worker)))); (typeof((typeof(*((worker))) *)((worker)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); })->work) &n->list_lock irq_context: 0 (wq_completion)wg-kex-wg0#12 (work_completion)(&({ do { const void *__vpp_verify = (typeof((worker) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __asm__ ("" : "=r"(__ptr) : "0"((typeof(*((worker))) *)((worker)))); (typeof((typeof(*((worker))) *)((worker)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); })->work) &n->list_lock &c->lock irq_context: 0 (wq_completion)wg-kex-wg2#8 (work_completion)(&({ do { const void *__vpp_verify = (typeof((worker) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __asm__ ("" : "=r"(__ptr) : "0"((typeof(*((worker))) *)((worker)))); (typeof((typeof(*((worker))) *)((worker)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); })->work) &handshake->lock &c->lock irq_context: 0 (wq_completion)wg-kex-wg2#8 (work_completion)(&({ do { const void *__vpp_verify = (typeof((worker) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __asm__ ("" : "=r"(__ptr) : "0"((typeof(*((worker))) *)((worker)))); (typeof((typeof(*((worker))) *)((worker)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); })->work) &c->lock irq_context: 0 &ei->i_data_sem rcu_node_0 irq_context: 0 (wq_completion)wg-kex-wg0#7 (work_completion)(&peer->transmit_handshake_work) &peer->endpoint_lock rcu_read_lock_bh quarantine_lock irq_context: 0 (wq_completion)wg-kex-wg0#7 (work_completion)(&peer->transmit_handshake_work) &____s->seqcount#2 irq_context: 0 (wq_completion)wg-kex-wg0#7 (work_completion)(&peer->transmit_handshake_work) &____s->seqcount irq_context: 0 (wq_completion)wg-kex-wg0#7 (work_completion)(&peer->transmit_handshake_work) pool_lock#2 irq_context: softirq rcu_read_lock_bh rcu_read_lock &pool->lock/1 &p->pi_lock &rq->__lock irq_context: softirq rcu_read_lock_bh rcu_read_lock &pool->lock/1 &p->pi_lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 cb_lock nlk_cb_mutex-GENERIC &n->list_lock irq_context: 0 cb_lock nlk_cb_mutex-GENERIC &n->list_lock &c->lock irq_context: 0 (wq_completion)wg-kex-wg0#4 (work_completion)(&({ do { const void *__vpp_verify = (typeof((worker) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __asm__ ("" : "=r"(__ptr) : "0"((typeof(*((worker))) *)((worker)))); (typeof((typeof(*((worker))) *)((worker)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); })->work) &handshake->lock rcu_read_lock_bh &peer->keypairs.keypair_update_lock &table->lock#2 irq_context: 0 (wq_completion)wg-kex-wg0#4 (work_completion)(&({ do { const void *__vpp_verify = (typeof((worker) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __asm__ ("" : "=r"(__ptr) : "0"((typeof(*((worker))) *)((worker)))); (typeof((typeof(*((worker))) *)((worker)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); })->work) &handshake->lock rcu_read_lock_bh &peer->keypairs.keypair_update_lock &obj_hash[i].lock irq_context: 0 (wq_completion)wg-kex-wg2#4 (work_completion)(&({ do { const void *__vpp_verify = (typeof((worker) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __asm__ ("" : "=r"(__ptr) : "0"((typeof(*((worker))) *)((worker)))); (typeof((typeof(*((worker))) *)((worker)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); })->work) &handshake->lock rcu_read_lock_bh &peer->keypairs.keypair_update_lock &table->lock#2 irq_context: 0 (wq_completion)wg-kex-wg2#4 (work_completion)(&({ do { const void *__vpp_verify = (typeof((worker) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __asm__ ("" : "=r"(__ptr) : "0"((typeof(*((worker))) *)((worker)))); (typeof((typeof(*((worker))) *)((worker)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); })->work) &handshake->lock rcu_read_lock_bh &peer->keypairs.keypair_update_lock &obj_hash[i].lock irq_context: 0 (wq_completion)wg-kex-wg2#4 (work_completion)(&({ do { const void *__vpp_verify = (typeof((worker) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __asm__ ("" : "=r"(__ptr) : "0"((typeof(*((worker))) *)((worker)))); (typeof((typeof(*((worker))) *)((worker)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); })->work) rcu_read_lock_bh rcu_read_lock &pool->lock &p->pi_lock &rq->__lock irq_context: 0 (wq_completion)wg-kex-wg2#4 (work_completion)(&({ do { const void *__vpp_verify = (typeof((worker) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __asm__ ("" : "=r"(__ptr) : "0"((typeof(*((worker))) *)((worker)))); (typeof((typeof(*((worker))) *)((worker)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); })->work) rcu_read_lock_bh rcu_read_lock &pool->lock &p->pi_lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 sb_writers#4 &type->i_mutex_dir_key#3/1 jbd2_handle &ei->i_data_sem quarantine_lock irq_context: 0 sk_lock-AF_INET (console_sem).lock irq_context: 0 sk_lock-AF_INET console_lock console_srcu console_owner_lock irq_context: 0 sk_lock-AF_INET console_lock console_srcu console_owner irq_context: 0 sk_lock-AF_INET console_lock console_srcu console_owner &port_lock_key irq_context: 0 sk_lock-AF_INET console_lock console_srcu console_owner console_owner_lock irq_context: 0 &sb->s_type->i_mutex_key#10 &meta->lock irq_context: 0 &sb->s_type->i_mutex_key#10 kfence_freelist_lock irq_context: 0 rcu_read_lock &vma->vm_lock->lock &cfs_rq->removed.lock irq_context: 0 rcu_read_lock &vma->vm_lock->lock &obj_hash[i].lock irq_context: 0 sk_lock-AF_INET6 rcu_read_lock tcp_metrics_lock &c->lock irq_context: 0 sk_lock-AF_INET6 rcu_read_lock &ei->socket.wq.wait &p->pi_lock &cfs_rq->removed.lock irq_context: 0 slock-AF_INET6 &sk->sk_lock.wq &p->pi_lock &rq->__lock &cfs_rq->removed.lock irq_context: 0 sk_lock-AF_INET6 &mm->mmap_lock &c->lock irq_context: 0 sk_lock-AF_INET6 &mm->mmap_lock &n->list_lock irq_context: 0 sk_lock-AF_INET6 rcu_read_lock rcu_read_lock &cfs_rq->removed.lock irq_context: 0 elock-AF_RDS irq_context: 0 cb_lock &rdev->wiphy.mtx &local->mtx &cfs_rq->removed.lock irq_context: 0 cb_lock &rdev->wiphy.mtx &local->mtx &obj_hash[i].lock irq_context: 0 sb_writers#4 &type->i_mutex_dir_key#3/1 sb_internal jbd2_handle &ei->i_data_sem &meta->lock irq_context: 0 &pipe->mutex/1 sk_lock-AF_INET6 fs_reclaim mmu_notifier_invalidate_range_start &rq->__lock irq_context: 0 &pipe->mutex/1 sk_lock-AF_INET6 fs_reclaim mmu_notifier_invalidate_range_start &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 &sb->s_type->i_mutex_key#10 (netlink_chain).rwsem rcu_node_0 irq_context: 0 &mm->mmap_lock sb_pagefaults mapping.invalidate_lock rcu_read_lock &rcu_state.gp_wq &p->pi_lock &rq->__lock &cfs_rq->removed.lock irq_context: 0 &sbi->s_writepages_rwsem jbd2_handle &ei->i_data_sem &____s->seqcount irq_context: 0 &sbi->s_writepages_rwsem jbd2_handle &ei->i_data_sem stock_lock irq_context: 0 &sbi->s_writepages_rwsem jbd2_handle &ei->i_data_sem &xa->xa_lock#9 irq_context: 0 &sbi->s_writepages_rwsem jbd2_handle &ei->i_data_sem lock#4 irq_context: 0 &sbi->s_writepages_rwsem jbd2_handle &ei->i_data_sem &journal->j_revoke_lock irq_context: 0 (wq_completion)wg-kex-wg0#6 (work_completion)(&({ do { const void *__vpp_verify = (typeof((worker) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __asm__ ("" : "=r"(__ptr) : "0"((typeof(*((worker))) *)((worker)))); (typeof((typeof(*((worker))) *)((worker)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); })->work) &handshake->lock rcu_read_lock_bh &peer->keypairs.keypair_update_lock &table->lock#2 irq_context: 0 (wq_completion)wg-kex-wg0#6 (work_completion)(&({ do { const void *__vpp_verify = (typeof((worker) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __asm__ ("" : "=r"(__ptr) : "0"((typeof(*((worker))) *)((worker)))); (typeof((typeof(*((worker))) *)((worker)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); })->work) &handshake->lock rcu_read_lock_bh &peer->keypairs.keypair_update_lock &obj_hash[i].lock irq_context: 0 (wq_completion)wg-kex-wg1#12 (work_completion)(&({ do { const void *__vpp_verify = (typeof((worker) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __asm__ ("" : "=r"(__ptr) : "0"((typeof(*((worker))) *)((worker)))); (typeof((typeof(*((worker))) *)((worker)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); })->work) &wg->static_identity.lock &rq->__lock irq_context: 0 (wq_completion)wg-kex-wg1#12 (work_completion)(&({ do { const void *__vpp_verify = (typeof((worker) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __asm__ ("" : "=r"(__ptr) : "0"((typeof(*((worker))) *)((worker)))); (typeof((typeof(*((worker))) *)((worker)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); })->work) &handshake->lock rcu_read_lock_bh &peer->keypairs.keypair_update_lock &table->lock#2 irq_context: 0 (wq_completion)wg-kex-wg1#12 (work_completion)(&({ do { const void *__vpp_verify = (typeof((worker) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __asm__ ("" : "=r"(__ptr) : "0"((typeof(*((worker))) *)((worker)))); (typeof((typeof(*((worker))) *)((worker)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); })->work) &handshake->lock rcu_read_lock_bh &peer->keypairs.keypair_update_lock &obj_hash[i].lock irq_context: 0 (wq_completion)wg-kex-wg1#12 (work_completion)(&({ do { const void *__vpp_verify = (typeof((worker) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __asm__ ("" : "=r"(__ptr) : "0"((typeof(*((worker))) *)((worker)))); (typeof((typeof(*((worker))) *)((worker)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); })->work) &handshake->lock &rq->__lock irq_context: 0 (wq_completion)wg-kex-wg1#12 (work_completion)(&({ do { const void *__vpp_verify = (typeof((worker) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __asm__ ("" : "=r"(__ptr) : "0"((typeof(*((worker))) *)((worker)))); (typeof((typeof(*((worker))) *)((worker)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); })->work) &c->lock irq_context: 0 (wq_completion)wg-kex-wg1#12 (work_completion)(&({ do { const void *__vpp_verify = (typeof((worker) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __asm__ ("" : "=r"(__ptr) : "0"((typeof(*((worker))) *)((worker)))); (typeof((typeof(*((worker))) *)((worker)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); })->work) &peer->endpoint_lock rcu_read_lock_bh batched_entropy_u32.lock irq_context: 0 cb_lock genl_mutex remove_cache_srcu irq_context: 0 cb_lock genl_mutex remove_cache_srcu quarantine_lock irq_context: 0 sb_writers#4 &type->i_mutex_dir_key#3/1 tomoyo_ss mmu_notifier_invalidate_range_start &rq->__lock &cfs_rq->removed.lock irq_context: 0 &f->f_pos_lock &type->i_mutex_dir_key#3 sb_writers#4 quarantine_lock irq_context: 0 &pipe->mutex/1 fs_reclaim mmu_notifier_invalidate_range_start &rq->__lock irq_context: 0 sb_writers#4 &type->i_mutex_dir_key#3/1 jbd2_handle rcu_read_lock &cfs_rq->removed.lock irq_context: 0 sb_writers#4 &type->i_mutex_dir_key#3/1 jbd2_handle rcu_read_lock &obj_hash[i].lock irq_context: 0 sb_writers#4 &type->i_mutex_dir_key#3/1 sb_internal jbd2_handle rcu_read_lock &rq->__lock &cfs_rq->removed.lock irq_context: 0 &f->f_pos_lock &type->i_mutex_dir_key#3 rcu_node_0 irq_context: 0 (wq_completion)mm_percpu_wq (work_completion)(&barr->work) &rq->__lock irq_context: 0 &f->f_pos_lock sb_writers#4 &sb->s_type->i_mutex_key#8 rcu_read_lock &rcu_state.gp_wq &p->pi_lock &rq->__lock &cfs_rq->removed.lock irq_context: 0 sb_writers#4 &type->i_mutex_dir_key#3/1 sb_internal &pcp->lock &zone->lock irq_context: 0 sb_writers#4 &type->i_mutex_dir_key#3/1 jbd2_handle &ei->i_data_sem &obj_hash[i].lock pool_lock irq_context: 0 cb_lock rcu_read_lock batched_entropy_u8.lock irq_context: 0 cb_lock rcu_read_lock kfence_freelist_lock irq_context: 0 cb_lock rcu_read_lock rcu_read_lock_bh &meta->lock irq_context: 0 cb_lock rcu_read_lock rcu_read_lock_bh kfence_freelist_lock irq_context: 0 cb_lock rcu_read_lock &rq->__lock &cfs_rq->removed.lock irq_context: 0 cb_lock rcu_read_lock &rcu_state.gp_wq irq_context: 0 cb_lock rcu_read_lock &rcu_state.gp_wq &p->pi_lock irq_context: 0 cb_lock rcu_read_lock &rcu_state.gp_wq &p->pi_lock &rq->__lock irq_context: 0 cb_lock rcu_read_lock &rcu_state.gp_wq &p->pi_lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: softirq (&icsk->icsk_delack_timer) slock-AF_INET6 tk_core.seq.seqcount irq_context: softirq (&icsk->icsk_delack_timer) slock-AF_INET6 &c->lock irq_context: softirq (&icsk->icsk_delack_timer) slock-AF_INET6 &____s->seqcount#2 irq_context: softirq (&icsk->icsk_delack_timer) slock-AF_INET6 &____s->seqcount irq_context: softirq (&icsk->icsk_delack_timer) slock-AF_INET6 rcu_read_lock rcu_read_lock &____s->seqcount#7 irq_context: softirq (&icsk->icsk_delack_timer) slock-AF_INET6 rcu_read_lock rcu_read_lock &ct->lock irq_context: softirq (&icsk->icsk_delack_timer) slock-AF_INET6 rcu_read_lock rcu_read_lock &____s->seqcount#10 irq_context: softirq (&icsk->icsk_delack_timer) slock-AF_INET6 rcu_read_lock rcu_read_lock rcu_read_lock_bh &list->lock#5 irq_context: softirq rcu_read_lock rcu_read_lock slock-AF_INET6/1 &f->f_owner.lock irq_context: softirq (&icsk->icsk_retransmit_timer) slock-AF_INET6 rcu_read_lock rcu_read_lock &____s->seqcount#10 irq_context: softirq (&icsk->icsk_retransmit_timer) slock-AF_INET6 rcu_read_lock rcu_read_lock rcu_read_lock_bh &list->lock#5 irq_context: softirq rcu_read_lock rcu_read_lock slock-AF_INET6/1 rcu_read_lock &ei->socket.wq.wait &p->pi_lock irq_context: softirq rcu_read_lock rcu_read_lock slock-AF_INET6/1 rcu_read_lock &ei->socket.wq.wait &p->pi_lock &rq->__lock irq_context: softirq rcu_read_lock rcu_read_lock slock-AF_INET6/1 rcu_read_lock &ei->socket.wq.wait &p->pi_lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: softirq rcu_read_lock rcu_read_lock slock-AF_INET6/1 rcu_read_lock rcu_read_lock rcu_read_lock &new->fa_lock irq_context: softirq rcu_read_lock rcu_read_lock slock-AF_INET6/1 rcu_read_lock rcu_read_lock rcu_read_lock &new->fa_lock &f->f_owner.lock irq_context: softirq rcu_read_lock rcu_read_lock slock-AF_INET6/1 &c->lock irq_context: 0 &sb->s_type->i_mutex_key#10 sk_lock-AF_INET6 &pcp->lock &zone->lock &____s->seqcount irq_context: 0 &sb->s_type->i_mutex_key#10 sk_lock-AF_INET6 rcu_read_lock rcu_node_0 irq_context: 0 &sb->s_type->i_mutex_key#10 sk_lock-AF_INET6 rcu_read_lock &rq->__lock irq_context: 0 &sb->s_type->i_mutex_key#10 sk_lock-AF_INET6 fs_reclaim irq_context: 0 &sb->s_type->i_mutex_key#10 sk_lock-AF_INET6 fs_reclaim mmu_notifier_invalidate_range_start irq_context: 0 &sb->s_type->i_mutex_key#10 sk_lock-AF_INET6 fs_reclaim mmu_notifier_invalidate_range_start &rq->__lock irq_context: 0 &sb->s_type->i_mutex_key#10 sk_lock-AF_INET6 rcu_read_lock rcu_read_lock &____s->seqcount#7 irq_context: 0 &sb->s_type->i_mutex_key#10 sk_lock-AF_INET6 rcu_read_lock rcu_read_lock &____s->seqcount#10 irq_context: 0 &sb->s_type->i_mutex_key#10 sk_lock-AF_INET6 rcu_read_lock rcu_read_lock rcu_read_lock_bh &list->lock#5 irq_context: 0 &sb->s_type->i_mutex_key#10 sk_lock-AF_INET6 slock-AF_INET6 rcu_read_lock pool_lock#2 irq_context: 0 &sb->s_type->i_mutex_key#10 sk_lock-AF_INET6 slock-AF_INET6 &pcp->lock &zone->lock irq_context: 0 &sb->s_type->i_mutex_key#10 sk_lock-AF_INET6 slock-AF_INET6 &pcp->lock &zone->lock &____s->seqcount irq_context: 0 &sb->s_type->i_mutex_key#10 sk_lock-AF_INET6 rcu_read_lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 dup_mmap_sem &mm->mmap_lock &mm->mmap_lock/1 &anon_vma->rwsem pgd_lock irq_context: 0 dup_mmap_sem &mm->mmap_lock &mm->mmap_lock/1 &anon_vma->rwsem key irq_context: 0 dup_mmap_sem &mm->mmap_lock &mm->mmap_lock/1 &anon_vma->rwsem pcpu_lock irq_context: 0 dup_mmap_sem &mm->mmap_lock &mm->mmap_lock/1 &anon_vma->rwsem percpu_counters_lock irq_context: 0 kn->active#5 &kernfs_locks->open_file_mutex[count] fs_reclaim &rq->__lock irq_context: 0 &sb->s_type->i_mutex_key#10 sk_lock-AF_INET slock-AF_INET key#24 irq_context: 0 &pipe->mutex/1 sk_lock-AF_INET6 rcu_read_lock &cfs_rq->removed.lock irq_context: 0 &mm->mmap_lock &lruvec->lru_lock per_cpu_ptr(&cgroup_rstat_cpu_lock, cpu) irq_context: 0 sb_writers#4 &sb->s_type->i_mutex_key#8 &sbi->s_writepages_rwsem rcu_state.exp_mutex &rq->__lock &cfs_rq->removed.lock irq_context: 0 (wq_completion)wg-kex-wg2#3 (work_completion)(&peer->transmit_handshake_work) &c->lock irq_context: 0 (wq_completion)wg-kex-wg2#3 (work_completion)(&peer->transmit_handshake_work) &n->list_lock irq_context: 0 (wq_completion)wg-kex-wg2#3 (work_completion)(&peer->transmit_handshake_work) &n->list_lock &c->lock irq_context: 0 (wq_completion)wg-kex-wg1#4 (work_completion)(&({ do { const void *__vpp_verify = (typeof((worker) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __asm__ ("" : "=r"(__ptr) : "0"((typeof(*((worker))) *)((worker)))); (typeof((typeof(*((worker))) *)((worker)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); })->work) &handshake->lock rcu_read_lock_bh &peer->keypairs.keypair_update_lock &table->lock#2 irq_context: 0 (wq_completion)wg-kex-wg1#4 (work_completion)(&({ do { const void *__vpp_verify = (typeof((worker) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __asm__ ("" : "=r"(__ptr) : "0"((typeof(*((worker))) *)((worker)))); (typeof((typeof(*((worker))) *)((worker)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); })->work) &handshake->lock rcu_read_lock_bh &peer->keypairs.keypair_update_lock &obj_hash[i].lock irq_context: 0 sb_writers#4 &type->i_mutex_dir_key#3/1 jbd2_handle &ei->i_data_sem rcu_read_lock &rcu_state.gp_wq &p->pi_lock &rq->__lock &cfs_rq->removed.lock irq_context: 0 sk_lock-AF_INET6 free_vmap_area_lock irq_context: 0 sk_lock-AF_INET6 vmap_area_lock irq_context: 0 sk_lock-AF_INET6 init_mm.page_table_lock irq_context: 0 sk_lock-AF_INET6 pcpu_alloc_mutex irq_context: 0 sk_lock-AF_INET6 pcpu_alloc_mutex pcpu_lock irq_context: 0 tomoyo_ss remove_cache_srcu &cfs_rq->removed.lock irq_context: 0 &sb->s_type->i_mutex_key#10 rlock-AF_PPPOX irq_context: 0 &sb->s_type->i_mutex_key#10 wlock-AF_PPPOX irq_context: 0 &pipe->mutex/1 &mm->mmap_lock rcu_node_0 irq_context: 0 sk_lock-AF_INET6 rcu_read_lock &pool->lock &p->pi_lock &rq->__lock &cfs_rq->removed.lock irq_context: 0 sk_lock-AF_INET6 pcpu_alloc_mutex pcpu_alloc_mutex.wait_lock irq_context: 0 sk_lock-AF_INET6 pcpu_alloc_mutex &rq->__lock irq_context: 0 sk_lock-AF_INET6 pcpu_alloc_mutex &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 &pipe->mutex/1 sk_lock-AF_INET6 remove_cache_srcu &pcp->lock &zone->lock irq_context: 0 &pipe->mutex/1 sk_lock-AF_INET6 remove_cache_srcu &pcp->lock &zone->lock &____s->seqcount irq_context: 0 &mm->mmap_lock sb_pagefaults mapping.invalidate_lock &folio_wait_table[i] irq_context: softirq (&pool->idle_timer) rcu_read_lock &pool->lock/1 irq_context: softirq (&pool->idle_timer) rcu_read_lock &pool->lock/1 &obj_hash[i].lock irq_context: softirq (&pool->idle_timer) rcu_read_lock &pool->lock/1 &p->pi_lock irq_context: 0 (wq_completion)events_unbound (work_completion)(&pool->idle_cull_work) irq_context: 0 (wq_completion)events_unbound (work_completion)(&pool->idle_cull_work) wq_pool_attach_mutex irq_context: 0 (wq_completion)events_unbound (work_completion)(&pool->idle_cull_work) wq_pool_attach_mutex &pool->lock irq_context: 0 (wq_completion)events_unbound (work_completion)(&pool->idle_cull_work) wq_pool_attach_mutex &pool->lock &obj_hash[i].lock irq_context: 0 (wq_completion)events_unbound (work_completion)(&pool->idle_cull_work) wq_pool_attach_mutex &pool->lock &base->lock irq_context: 0 (wq_completion)events_unbound (work_completion)(&pool->idle_cull_work) wq_pool_attach_mutex &pool->lock &base->lock &obj_hash[i].lock irq_context: 0 (wq_completion)events_unbound (work_completion)(&pool->idle_cull_work) wq_pool_attach_mutex &p->pi_lock irq_context: 0 (wq_completion)events_unbound (work_completion)(&pool->idle_cull_work) wq_pool_attach_mutex &p->pi_lock &rq->__lock irq_context: 0 (wq_completion)events_unbound (work_completion)(&pool->idle_cull_work) wq_pool_attach_mutex &p->pi_lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 &xs->mutex &rq->__lock &cfs_rq->removed.lock irq_context: 0 sk_lock-AF_INET6 rcu_read_lock rcu_read_lock (console_sem).lock irq_context: 0 sk_lock-AF_INET6 rcu_read_lock rcu_read_lock console_lock console_srcu console_owner_lock irq_context: 0 sk_lock-AF_INET6 rcu_read_lock rcu_read_lock console_lock console_srcu console_owner irq_context: 0 sk_lock-AF_INET6 rcu_read_lock rcu_read_lock console_lock console_srcu console_owner &port_lock_key irq_context: 0 sk_lock-AF_INET6 rcu_read_lock rcu_read_lock console_lock console_srcu console_owner console_owner_lock irq_context: 0 sk_lock-AF_INET6 rcu_read_lock rcu_read_lock console_owner_lock irq_context: 0 sk_lock-AF_INET6 rcu_read_lock rcu_read_lock console_owner irq_context: 0 cb_lock &rdev->wiphy.mtx rcu_read_lock rcu_read_lock_bh rcu_read_lock rlock-AF_PACKET irq_context: 0 kn->active#5 remove_cache_srcu &cfs_rq->removed.lock irq_context: 0 &sb->s_type->i_mutex_key#10 &zone->lock irq_context: 0 &sb->s_type->i_mutex_key#10 &zone->lock &____s->seqcount irq_context: 0 cb_lock &rdev->wiphy.mtx rcu_read_lock rcu_read_lock_bh rcu_read_lock &n->list_lock irq_context: 0 cb_lock &rdev->wiphy.mtx rcu_read_lock rcu_read_lock_bh rcu_read_lock &n->list_lock &c->lock irq_context: 0 sk_lock-AF_INET6 cpu_hotplug_lock irq_context: 0 sk_lock-AF_INET6 cpu_hotplug_lock jump_label_mutex irq_context: 0 sk_lock-AF_INET6 cpu_hotplug_lock jump_label_mutex text_mutex irq_context: 0 sk_lock-AF_INET6 cpu_hotplug_lock jump_label_mutex text_mutex ptlock_ptr(page)#2 irq_context: 0 rtnl_mutex &block->lock &n->list_lock irq_context: 0 rtnl_mutex &block->lock &n->list_lock &c->lock irq_context: 0 sb_writers#4 &sb->s_type->i_mutex_key#8 mapping.invalidate_lock jbd2_handle &ei->i_data_sem batched_entropy_u8.lock irq_context: 0 sb_writers#4 &sb->s_type->i_mutex_key#8 mapping.invalidate_lock jbd2_handle &ei->i_data_sem kfence_freelist_lock irq_context: 0 sb_writers#4 &sb->s_type->i_mutex_key#8 mapping.invalidate_lock jbd2_handle &ei->i_data_sem &meta->lock irq_context: 0 &xs->mutex rcu_read_lock rcu_read_lock rcu_node_0 irq_context: 0 &xs->mutex rcu_read_lock rcu_read_lock &rq->__lock irq_context: 0 &xs->mutex rcu_read_lock rcu_read_lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 sb_writers#4 &type->i_mutex_dir_key#3/1 sb_internal remove_cache_srcu rcu_read_lock &rcu_state.gp_wq irq_context: 0 sb_writers#4 &type->i_mutex_dir_key#3/1 sb_internal remove_cache_srcu rcu_read_lock &rcu_state.gp_wq &p->pi_lock irq_context: 0 sb_writers#4 &type->i_mutex_dir_key#3/1 sb_internal remove_cache_srcu rcu_read_lock &rcu_state.gp_wq &p->pi_lock &rq->__lock irq_context: 0 sb_writers#4 &type->i_mutex_dir_key#3/1 sb_internal remove_cache_srcu rcu_read_lock &rcu_state.gp_wq &p->pi_lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 cb_lock &rdev->wiphy.mtx rcu_read_lock &rq->__lock &cfs_rq->removed.lock irq_context: 0 cb_lock &rdev->wiphy.mtx rcu_read_lock &cfs_rq->removed.lock irq_context: 0 sb_writers#4 sb_writers#4 quarantine_lock irq_context: 0 sb_writers#4 remove_cache_srcu rcu_read_lock rcu_node_0 irq_context: 0 sb_writers#4 remove_cache_srcu rcu_read_lock &rq->__lock irq_context: 0 sb_writers#4 remove_cache_srcu rcu_read_lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: softirq &fq->mq_flush_lock &x->wait#26 &p->pi_lock &cfs_rq->removed.lock irq_context: 0 sb_writers#4 &sb->s_type->i_mutex_key#8 &folio_wait_table[i] &p->pi_lock irq_context: 0 sb_writers#4 &sb->s_type->i_mutex_key#8 &folio_wait_table[i] &p->pi_lock &rq->__lock irq_context: 0 sb_writers#4 &sb->s_type->i_mutex_key#8 &folio_wait_table[i] &p->pi_lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 sb_writers#4 sb_writers#4 jbd2_handle &journal->j_wait_updates &p->pi_lock irq_context: 0 sb_writers#4 sb_writers#4 jbd2_handle &journal->j_wait_updates &p->pi_lock &rq->__lock irq_context: 0 sb_writers#4 sb_writers#4 jbd2_handle &journal->j_wait_updates &p->pi_lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 sb_writers#4 &sb->s_type->i_mutex_key#8 &folio_wait_table[i] irq_context: 0 sb_writers#4 &sb->s_type->i_mutex_key#8 &folio_wait_table[i] &p->pi_lock irq_context: 0 sb_writers#4 &sb->s_type->i_mutex_key#8 &folio_wait_table[i] &p->pi_lock &rq->__lock irq_context: 0 sb_writers#4 &sb->s_type->i_mutex_key#8 &folio_wait_table[i] &p->pi_lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 sb_writers#4 mapping.invalidate_lock &n->list_lock irq_context: 0 sb_writers#4 mapping.invalidate_lock &n->list_lock &c->lock irq_context: 0 sb_writers#4 &sb->s_type->i_mutex_key#8 &sb->s_type->i_lock_key#22 &xa->xa_lock#9 rcu_read_lock per_cpu_ptr(&cgroup_rstat_cpu_lock, cpu) irq_context: 0 sb_writers#4 &sb->s_type->i_mutex_key#8 rcu_read_lock fill_pool_map-wait-type-override rcu_node_0 irq_context: 0 sb_writers#4 &sb->s_type->i_mutex_key#8 rcu_read_lock fill_pool_map-wait-type-override &rq->__lock irq_context: 0 sb_writers#4 &sb->s_type->i_mutex_key#8 rcu_read_lock fill_pool_map-wait-type-override &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 sb_writers#4 &sb->s_type->i_mutex_key#8 &sb->s_type->i_lock_key#22 &xa->xa_lock#9 &obj_hash[i].lock irq_context: 0 sb_writers#4 &sb->s_type->i_mutex_key#8 &sb->s_type->i_lock_key#22 &xa->xa_lock#9 pool_lock#2 irq_context: 0 sb_writers#4 mapping.invalidate_lock rcu_read_lock rcu_read_lock &pool->lock irq_context: 0 sb_writers#4 mapping.invalidate_lock rcu_read_lock rcu_read_lock &pool->lock &obj_hash[i].lock irq_context: 0 sb_writers#4 mapping.invalidate_lock rcu_read_lock rcu_read_lock &pool->lock pool_lock#2 irq_context: 0 sb_writers#4 mapping.invalidate_lock rcu_read_lock rcu_read_lock &pool->lock &p->pi_lock irq_context: 0 sb_writers#4 mapping.invalidate_lock rcu_read_lock rcu_read_lock &pool->lock &p->pi_lock &rq->__lock irq_context: 0 sb_writers#4 mapping.invalidate_lock rcu_read_lock rcu_read_lock &pool->lock &p->pi_lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)wg-crypt-wg0#5 (work_completion)(&peer->transmit_packet_work) &peer->endpoint_lock rcu_read_lock_bh rcu_read_lock &n->list_lock irq_context: 0 (wq_completion)wg-crypt-wg0#5 (work_completion)(&peer->transmit_packet_work) &peer->endpoint_lock rcu_read_lock_bh rcu_read_lock &n->list_lock &c->lock irq_context: 0 sb_writers#4 &sb->s_type->i_mutex_key#8 &folio_wait_table[i] &p->pi_lock &cfs_rq->removed.lock irq_context: 0 sb_writers#5 &type->i_mutex_dir_key#5/1 rcu_read_lock &cfs_rq->removed.lock irq_context: 0 sb_writers#5 &type->i_mutex_dir_key#5/1 rcu_read_lock &obj_hash[i].lock irq_context: 0 sb_writers#4 &sb->s_type->i_mutex_key#8 &sbi->s_writepages_rwsem rcu_state.exp_mutex &obj_hash[i].lock pool_lock irq_context: 0 sb_writers#5 fill_pool_map-wait-type-override &rq->__lock irq_context: 0 sb_writers#5 fill_pool_map-wait-type-override &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)wg-kex-wg0#7 (work_completion)(&peer->transmit_handshake_work) &peer->endpoint_lock rcu_read_lock_bh rcu_read_lock &obj_hash[i].lock pool_lock irq_context: 0 (wq_completion)wg-kex-wg0#7 (work_completion)(&peer->transmit_handshake_work) &peer->endpoint_lock rcu_read_lock_bh rcu_read_lock batched_entropy_u8.lock irq_context: 0 (wq_completion)wg-kex-wg0#7 (work_completion)(&peer->transmit_handshake_work) &peer->endpoint_lock rcu_read_lock_bh rcu_read_lock kfence_freelist_lock irq_context: 0 &xs->mutex rcu_read_lock &rcu_state.gp_wq &p->pi_lock &rq->__lock &cfs_rq->removed.lock irq_context: 0 sb_writers#4 &type->i_mutex_dir_key#3/1 &sb->s_type->i_mutex_key#8 quarantine_lock irq_context: 0 &f->f_pos_lock sb_writers#4 &sb->s_type->i_mutex_key#8 remove_cache_srcu &pcp->lock &zone->lock irq_context: 0 &f->f_pos_lock sb_writers#4 &sb->s_type->i_mutex_key#8 remove_cache_srcu &pcp->lock &zone->lock &____s->seqcount irq_context: softirq rcu_read_lock rcu_read_lock rcu_read_lock rcu_read_lock rcu_read_lock &pool->lock irq_context: softirq rcu_read_lock rcu_read_lock rcu_read_lock rcu_read_lock rcu_read_lock &pool->lock &obj_hash[i].lock irq_context: softirq rcu_read_lock rcu_read_lock rcu_read_lock rcu_read_lock rcu_read_lock &pool->lock pool_lock#2 irq_context: softirq rcu_read_lock rcu_read_lock rcu_read_lock rcu_read_lock rcu_read_lock &pool->lock &p->pi_lock irq_context: softirq rcu_read_lock rcu_read_lock rcu_read_lock rcu_read_lock rcu_read_lock &pool->lock &p->pi_lock &rq->__lock irq_context: softirq rcu_read_lock rcu_read_lock rcu_read_lock rcu_read_lock rcu_read_lock &pool->lock &p->pi_lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 &f->f_pos_lock sb_writers#4 &sb->s_type->i_mutex_key#8 rcu_read_lock pgd_lock irq_context: 0 &f->f_pos_lock sb_writers#4 &sb->s_type->i_mutex_key#8 rcu_read_lock stock_lock irq_context: 0 &f->f_pos_lock sb_writers#4 &sb->s_type->i_mutex_key#8 rcu_read_lock key irq_context: 0 &f->f_pos_lock sb_writers#4 &sb->s_type->i_mutex_key#8 rcu_read_lock pcpu_lock irq_context: 0 &f->f_pos_lock sb_writers#4 &sb->s_type->i_mutex_key#8 rcu_read_lock percpu_counters_lock irq_context: 0 &f->f_pos_lock sb_writers#4 &sb->s_type->i_mutex_key#8 rcu_read_lock pcpu_lock stock_lock irq_context: 0 (wq_completion)wg-crypt-wg0#2 (work_completion)(&({ do { const void *__vpp_verify = (typeof((worker) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __asm__ ("" : "=r"(__ptr) : "0"((typeof(*((worker))) *)((worker)))); (typeof((typeof(*((worker))) *)((worker)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); })->work) rcu_read_lock &pool->lock &p->pi_lock &rq->__lock &cfs_rq->removed.lock irq_context: 0 sk_lock-AF_INET rcu_read_lock &____s->seqcount#9 irq_context: 0 sk_lock-AF_INET rcu_read_lock rcu_read_lock_bh &obj_hash[i].lock irq_context: 0 sk_lock-AF_INET rcu_read_lock rcu_read_lock_bh pool_lock#2 irq_context: 0 sk_lock-AF_INET rcu_read_lock rcu_read_lock_bh noop_qdisc.q.lock irq_context: softirq fs/notify/mark.c:89 rcu_read_lock &pool->lock/1 &p->pi_lock &rq->__lock &cfs_rq->removed.lock irq_context: 0 hashlimit_mutex &obj_hash[i].lock pool_lock irq_context: 0 cb_lock genl_mutex remove_cache_srcu &c->lock irq_context: 0 cb_lock genl_mutex remove_cache_srcu &rq->__lock irq_context: 0 cb_lock genl_mutex remove_cache_srcu &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 cb_lock genl_mutex remove_cache_srcu &n->list_lock irq_context: 0 cb_lock genl_mutex remove_cache_srcu rcu_read_lock pool_lock#2 irq_context: 0 cb_lock genl_mutex remove_cache_srcu &obj_hash[i].lock irq_context: 0 &f->f_pos_lock sb_writers#4 &sb->s_type->i_mutex_key#8 jbd2_handle &journal->j_state_lock &journal->j_wait_commit &p->pi_lock &cfs_rq->removed.lock irq_context: 0 &group->mark_mutex fs_reclaim &rq->__lock irq_context: 0 &group->mark_mutex fs_reclaim &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 &xs->mutex remove_cache_srcu irq_context: 0 &xs->mutex remove_cache_srcu quarantine_lock irq_context: 0 &xs->mutex remove_cache_srcu &c->lock irq_context: 0 &xs->mutex remove_cache_srcu &n->list_lock irq_context: 0 &xs->mutex remove_cache_srcu rcu_read_lock pool_lock#2 irq_context: 0 &xs->mutex remove_cache_srcu &obj_hash[i].lock irq_context: 0 &xs->mutex remove_cache_srcu &rq->__lock irq_context: 0 &xs->mutex remove_cache_srcu pool_lock#2 irq_context: 0 &sb->s_type->i_mutex_key#10 rcu_state.exp_mutex rcu_read_lock &pool->lock fill_pool_map-wait-type-override &n->list_lock irq_context: 0 &sb->s_type->i_mutex_key#10 rcu_state.exp_mutex rcu_read_lock &pool->lock fill_pool_map-wait-type-override &n->list_lock &c->lock irq_context: 0 (wq_completion)wg-crypt-wg1#2 (work_completion)(&peer->transmit_packet_work) &peer->endpoint_lock rcu_read_lock_bh rcu_read_lock &c->lock irq_context: 0 sk_lock-AF_INET6 remove_cache_srcu rcu_read_lock &rcu_state.gp_wq irq_context: 0 sk_lock-AF_INET6 remove_cache_srcu rcu_read_lock &rcu_state.gp_wq &p->pi_lock irq_context: 0 sk_lock-AF_INET6 remove_cache_srcu rcu_read_lock &rcu_state.gp_wq &p->pi_lock &rq->__lock irq_context: 0 sk_lock-AF_INET6 remove_cache_srcu rcu_read_lock &rcu_state.gp_wq &p->pi_lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 cb_lock &rdev->wiphy.mtx &local->mtx rcu_read_lock &rq->__lock &cfs_rq->removed.lock irq_context: softirq rcu_callback stock_lock per_cpu_ptr(&cgroup_rstat_cpu_lock, cpu) irq_context: 0 sb_writers#4 &type->i_mutex_dir_key#3 rcu_read_lock &cfs_rq->removed.lock irq_context: 0 sb_writers#4 &type->i_mutex_dir_key#3 rcu_read_lock &obj_hash[i].lock irq_context: 0 rcu_read_lock &n->lock &n->list_lock irq_context: 0 rcu_read_lock &n->lock &n->list_lock &c->lock irq_context: 0 rcu_read_lock &n->lock rcu_read_lock pool_lock#2 irq_context: 0 rcu_read_lock rcu_read_lock rcu_read_lock &____s->seqcount#10 irq_context: 0 sb_writers#4 &type->i_mutex_dir_key#3 jbd2_handle &meta_group_info[i]->alloc_sem &rq->__lock irq_context: 0 sb_writers#4 &type->i_mutex_dir_key#3 jbd2_handle &meta_group_info[i]->alloc_sem &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 &f->f_pos_lock sb_writers#4 &sb->s_type->i_mutex_key#8 &ei->i_data_sem rcu_read_lock rcu_node_0 irq_context: 0 nf_sockopt_mutex &rq->__lock &cfs_rq->removed.lock irq_context: 0 &sb->s_type->i_mutex_key#10 sk_lock-AF_CAIF rcu_state.exp_mutex &obj_hash[i].lock pool_lock irq_context: 0 sb_writers#4 &type->i_mutex_dir_key#3/1 &type->i_mutex_dir_key#3 rcu_read_lock &cfs_rq->removed.lock irq_context: 0 sb_writers#4 &type->i_mutex_dir_key#3/1 &type->i_mutex_dir_key#3 rcu_read_lock &obj_hash[i].lock irq_context: 0 sb_writers#4 sb_internal rcu_read_lock &cfs_rq->removed.lock irq_context: 0 sb_writers#4 sb_internal rcu_read_lock &obj_hash[i].lock irq_context: 0 &mm->mmap_lock &anon_vma->rwsem stock_lock irq_context: 0 &mm->mmap_lock &anon_vma->rwsem pcpu_lock stock_lock irq_context: 0 rtnl_mutex &xs->mutex irq_context: 0 rtnl_mutex &xs->mutex &rq->__lock irq_context: 0 rtnl_mutex &xs->mutex &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 cb_lock rtnl_mutex &rdev->wiphy.mtx rcu_read_lock &rq->__lock irq_context: 0 cb_lock rtnl_mutex &rdev->wiphy.mtx rcu_read_lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 cb_lock &rdev->wiphy.mtx remove_cache_srcu pool_lock#2 irq_context: 0 clock-AF_LLC irq_context: 0 &sb->s_type->i_mutex_key#10 sk_lock-AF_LLC irq_context: 0 &sb->s_type->i_mutex_key#10 sk_lock-AF_LLC slock-AF_LLC irq_context: 0 &sb->s_type->i_mutex_key#10 slock-AF_LLC irq_context: 0 &sb->s_type->i_mutex_key#10 (&llc->pf_cycle_timer.timer) irq_context: 0 &sb->s_type->i_mutex_key#10 &base->lock irq_context: 0 &sb->s_type->i_mutex_key#10 (&llc->ack_timer.timer) irq_context: 0 &sb->s_type->i_mutex_key#10 (&llc->rej_sent_timer.timer) irq_context: 0 &sb->s_type->i_mutex_key#10 (&llc->busy_state_timer.timer) irq_context: 0 &sb->s_type->i_mutex_key#10 rlock-AF_LLC irq_context: 0 &sb->s_type->i_mutex_key#10 wlock-AF_LLC irq_context: 0 &sb->s_type->i_mutex_key#10 &list->lock#36 irq_context: 0 sb_writers#4 &type->i_mutex_dir_key#3/1 rcu_read_lock &rq->__lock &cfs_rq->removed.lock irq_context: 0 &f->f_pos_lock sb_writers#4 &sb->s_type->i_mutex_key#8 remove_cache_srcu &rq->__lock &cfs_rq->removed.lock irq_context: 0 &xs->mutex &____s->seqcount#2 irq_context: 0 &f->f_pos_lock sb_writers#4 &sb->s_type->i_mutex_key#8 jbd2_handle &ei->i_data_sem &ei->i_es_lock &c->lock irq_context: 0 sb_writers#4 &iint->mutex ima_extend_list_mutex &c->lock irq_context: 0 sb_writers#4 sb_writers#4 jbd2_handle &rq->__lock &cfs_rq->removed.lock irq_context: 0 sb_writers#4 mapping.invalidate_lock &xa->xa_lock#9 &n->list_lock &c->lock irq_context: 0 &sb->s_type->i_mutex_key#10 rcu_read_lock &____s->seqcount irq_context: 0 sb_writers#10 &type->i_mutex_dir_key#6 &root->kernfs_rwsem &c->lock irq_context: 0 cb_lock &rdev->wiphy.mtx &local->mtx &local->ack_status_lock &____s->seqcount#2 irq_context: 0 cb_lock &rdev->wiphy.mtx &local->mtx &local->ack_status_lock &____s->seqcount irq_context: 0 &sb->s_type->i_mutex_key#10 rcu_read_lock rcu_read_lock &cfs_rq->removed.lock irq_context: 0 &sb->s_type->i_mutex_key#10 rcu_read_lock rcu_read_lock &obj_hash[i].lock irq_context: 0 sb_writers#4 tomoyo_ss rcu_read_lock rcu_read_lock &cfs_rq->removed.lock irq_context: 0 sb_writers#4 tomoyo_ss rcu_read_lock rcu_read_lock &obj_hash[i].lock irq_context: 0 sb_writers#4 tomoyo_ss rcu_read_lock &rq->__lock &cfs_rq->removed.lock irq_context: 0 &sb->s_type->i_mutex_key#10 sk_lock-AF_CAIF rcu_state.exp_mutex &cfs_rq->removed.lock irq_context: 0 &sb->s_type->i_mutex_key#10 sk_lock-AF_CAIF rcu_state.exp_mutex pool_lock#2 irq_context: 0 sb_writers#4 &type->i_mutex_dir_key#3 &cfs_rq->removed.lock irq_context: 0 sk_lock-AF_INET6 &net->sctp.addr_wq_lock irq_context: 0 (wq_completion)wg-kex-wg1#3 (work_completion)(&peer->transmit_handshake_work) &c->lock irq_context: 0 (wq_completion)wg-kex-wg1#4 (work_completion)(&({ do { const void *__vpp_verify = (typeof((worker) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __asm__ ("" : "=r"(__ptr) : "0"((typeof(*((worker))) *)((worker)))); (typeof((typeof(*((worker))) *)((worker)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); })->work) &handshake->lock &c->lock irq_context: 0 &sb->s_type->i_mutex_key#10 &net->sctp.addr_wq_lock slock-AF_INET6/1 &obj_hash[i].lock pool_lock irq_context: 0 pernet_ops_rwsem devices_rwsem rdma_nets_rwsem &device->compat_devs_mutex batched_entropy_u8.lock crngs.lock irq_context: 0 (wq_completion)wg-kex-wg0#6 (work_completion)(&({ do { const void *__vpp_verify = (typeof((worker) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __asm__ ("" : "=r"(__ptr) : "0"((typeof(*((worker))) *)((worker)))); (typeof((typeof(*((worker))) *)((worker)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); })->work) &handshake->lock &c->lock irq_context: 0 (wq_completion)wg-kex-wg0#6 (work_completion)(&({ do { const void *__vpp_verify = (typeof((worker) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __asm__ ("" : "=r"(__ptr) : "0"((typeof(*((worker))) *)((worker)))); (typeof((typeof(*((worker))) *)((worker)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); })->work) &handshake->lock &n->list_lock irq_context: 0 (wq_completion)wg-kex-wg0#6 (work_completion)(&({ do { const void *__vpp_verify = (typeof((worker) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __asm__ ("" : "=r"(__ptr) : "0"((typeof(*((worker))) *)((worker)))); (typeof((typeof(*((worker))) *)((worker)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); })->work) &handshake->lock &n->list_lock &c->lock irq_context: 0 (wq_completion)wg-kex-wg0#12 (work_completion)(&({ do { const void *__vpp_verify = (typeof((worker) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __asm__ ("" : "=r"(__ptr) : "0"((typeof(*((worker))) *)((worker)))); (typeof((typeof(*((worker))) *)((worker)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); })->work) &handshake->lock rcu_read_lock_bh &peer->keypairs.keypair_update_lock fill_pool_map-wait-type-override pool_lock#2 irq_context: 0 (wq_completion)wg-kex-wg0#12 (work_completion)(&({ do { const void *__vpp_verify = (typeof((worker) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __asm__ ("" : "=r"(__ptr) : "0"((typeof(*((worker))) *)((worker)))); (typeof((typeof(*((worker))) *)((worker)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); })->work) &handshake->lock rcu_read_lock_bh &peer->keypairs.keypair_update_lock fill_pool_map-wait-type-override &c->lock irq_context: 0 (wq_completion)wg-kex-wg0#12 (work_completion)(&({ do { const void *__vpp_verify = (typeof((worker) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __asm__ ("" : "=r"(__ptr) : "0"((typeof(*((worker))) *)((worker)))); (typeof((typeof(*((worker))) *)((worker)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); })->work) &handshake->lock rcu_read_lock_bh &peer->keypairs.keypair_update_lock fill_pool_map-wait-type-override pool_lock irq_context: 0 (wq_completion)wg-kex-wg1#12 (work_completion)(&({ do { const void *__vpp_verify = (typeof((worker) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __asm__ ("" : "=r"(__ptr) : "0"((typeof(*((worker))) *)((worker)))); (typeof((typeof(*((worker))) *)((worker)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); })->work) rcu_read_lock_bh rcu_read_lock &pool->lock &p->pi_lock &rq->__lock irq_context: 0 (wq_completion)wg-kex-wg1#12 (work_completion)(&({ do { const void *__vpp_verify = (typeof((worker) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __asm__ ("" : "=r"(__ptr) : "0"((typeof(*((worker))) *)((worker)))); (typeof((typeof(*((worker))) *)((worker)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); })->work) rcu_read_lock_bh rcu_read_lock &pool->lock &p->pi_lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 sk_lock-AF_INET6 quarantine_lock irq_context: 0 cb_lock &rdev->wiphy.mtx &local->mtx &local->ack_status_lock &pcp->lock &zone->lock irq_context: 0 (wq_completion)events_unbound (work_completion)(&sub_info->work) remove_cache_srcu pool_lock#2 irq_context: 0 pernet_ops_rwsem uevent_sock_mutex nl_table_wait.lock &p->pi_lock &cfs_rq->removed.lock irq_context: 0 sb_writers#4 &sb->s_type->i_mutex_key#8 &sbi->s_writepages_rwsem rcu_state.exp_mutex pgd_lock irq_context: 0 sb_writers#4 &sb->s_type->i_mutex_key#8 &sbi->s_writepages_rwsem rcu_state.exp_mutex rcu_read_lock pool_lock#2 irq_context: 0 sb_writers#4 &sb->s_type->i_mutex_key#8 &sbi->s_writepages_rwsem rcu_state.exp_mutex key irq_context: 0 sb_writers#4 &sb->s_type->i_mutex_key#8 &sbi->s_writepages_rwsem rcu_state.exp_mutex pcpu_lock irq_context: 0 sb_writers#4 &sb->s_type->i_mutex_key#8 &sbi->s_writepages_rwsem rcu_state.exp_mutex percpu_counters_lock irq_context: 0 sb_writers#4 &sb->s_type->i_mutex_key#8 &sbi->s_writepages_rwsem rcu_state.exp_mutex pool_lock#2 irq_context: 0 (wq_completion)wg-kex-wg0#7 (work_completion)(&peer->transmit_handshake_work) &peer->endpoint_lock rcu_read_lock_bh &____s->seqcount#2 irq_context: 0 (wq_completion)wg-kex-wg0#7 (work_completion)(&peer->transmit_handshake_work) &peer->endpoint_lock rcu_read_lock_bh &____s->seqcount irq_context: 0 pernet_ops_rwsem devices_rwsem rdma_nets_rwsem &device->compat_devs_mutex pgd_lock irq_context: 0 pernet_ops_rwsem devices_rwsem rdma_nets_rwsem &device->compat_devs_mutex key irq_context: 0 pernet_ops_rwsem devices_rwsem rdma_nets_rwsem &device->compat_devs_mutex pcpu_lock irq_context: 0 pernet_ops_rwsem devices_rwsem rdma_nets_rwsem &device->compat_devs_mutex percpu_counters_lock irq_context: 0 rlock-AF_BLUETOOTH irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem rtnl_mutex &root->kernfs_rwsem rcu_read_lock &rcu_state.gp_wq irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem rtnl_mutex &root->kernfs_rwsem rcu_read_lock &rcu_state.gp_wq &p->pi_lock irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem rtnl_mutex &root->kernfs_rwsem rcu_read_lock &rcu_state.gp_wq &p->pi_lock &rq->__lock irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem rtnl_mutex &root->kernfs_rwsem rcu_read_lock &rcu_state.gp_wq &p->pi_lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 &sb->s_type->i_mutex_key#9 &rq->__lock &cfs_rq->removed.lock irq_context: 0 crngs.lock base_crng.lock irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem rtnl_mutex &net->xdp.lock &rq->__lock irq_context: 0 &nft_net->commit_mutex &obj_hash[i].lock pool_lock irq_context: 0 &mm->mmap_lock rcu_read_lock &rcu_state.expedited_wq &p->pi_lock &rq->__lock &cfs_rq->removed.lock irq_context: 0 nfnl_subsys_nftables &rq->__lock irq_context: 0 nfnl_subsys_nftables &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)wg-kex-wg1#9 (work_completion)(&peer->transmit_handshake_work) &wg->static_identity.lock &handshake->lock rcu_read_lock_bh batched_entropy_u32.lock crngs.lock irq_context: 0 rcu_read_lock rcu_read_lock rcu_read_lock rcu_read_lock_bh dev->qdisc_tx_busylock ?: &qdisc_tx_busylock rcu_read_lock &ndev->lock irq_context: 0 rcu_read_lock rcu_read_lock rcu_read_lock rcu_read_lock_bh dev->qdisc_tx_busylock ?: &qdisc_tx_busylock &obj_hash[i].lock irq_context: 0 rcu_read_lock rcu_read_lock rcu_read_lock rcu_read_lock_bh dev->qdisc_tx_busylock ?: &qdisc_tx_busylock pool_lock#2 irq_context: 0 sb_writers#4 &sb->s_type->i_mutex_key#8 mapping.invalidate_lock jbd2_handle &rq->__lock &cfs_rq->removed.lock irq_context: 0 &mm->mmap_lock &mapping->i_mmap_rwsem &sem->wait_lock irq_context: 0 (wq_completion)bat_events (work_completion)(&(&bat_priv->nc.work)->work) rcu_read_lock &rcu_state.expedited_wq &p->pi_lock &rq->__lock &cfs_rq->removed.lock irq_context: 0 sb_writers#4 &sb->s_type->i_mutex_key#8 &sbi->s_writepages_rwsem irq_context: 0 sb_writers#4 &sb->s_type->i_mutex_key#8 &sbi->s_writepages_rwsem &xa->xa_lock#9 irq_context: 0 sb_writers#4 &sb->s_type->i_mutex_key#8 &sbi->s_writepages_rwsem mmu_notifier_invalidate_range_start irq_context: 0 sb_writers#4 &sb->s_type->i_mutex_key#8 &sbi->s_writepages_rwsem pool_lock#2 irq_context: 0 sb_writers#4 &sb->s_type->i_mutex_key#8 &sbi->s_writepages_rwsem &obj_hash[i].lock irq_context: 0 &xs->mutex umem_ida.xa_lock irq_context: 0 &xs->mutex &mm->mmap_lock irq_context: 0 sb_writers#4 &sb->s_type->i_mutex_key#8 jbd2_handle mmu_notifier_invalidate_range_start irq_context: 0 sb_writers#4 &sb->s_type->i_mutex_key#8 jbd2_handle &____s->seqcount irq_context: 0 sb_writers#4 &sb->s_type->i_mutex_key#8 jbd2_handle pool_lock#2 irq_context: 0 sb_writers#4 &sb->s_type->i_mutex_key#8 jbd2_handle rcu_read_lock pool_lock#2 irq_context: softirq (&n->timer) rcu_read_lock rcu_read_lock rcu_read_lock_bh dev->qdisc_tx_busylock ?: &qdisc_tx_busylock rcu_read_lock &ndev->lock irq_context: softirq (&n->timer) rcu_read_lock rcu_read_lock rcu_read_lock_bh dev->qdisc_tx_busylock ?: &qdisc_tx_busylock &obj_hash[i].lock irq_context: softirq (&n->timer) rcu_read_lock rcu_read_lock rcu_read_lock_bh dev->qdisc_tx_busylock ?: &qdisc_tx_busylock pool_lock#2 irq_context: 0 &f->f_pos_lock sb_writers#4 &sb->s_type->i_mutex_key#8 jbd2_handle &ei->i_data_sem &lg->lg_mutex &rq->__lock irq_context: 0 &f->f_pos_lock sb_writers#4 &sb->s_type->i_mutex_key#8 jbd2_handle &ei->i_data_sem &lg->lg_mutex &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 mapping.invalidate_lock mmu_notifier_invalidate_range_start irq_context: 0 mapping.invalidate_lock &____s->seqcount irq_context: 0 mapping.invalidate_lock pool_lock#2 irq_context: 0 mapping.invalidate_lock stock_lock irq_context: 0 mapping.invalidate_lock &xa->xa_lock#9 irq_context: 0 mapping.invalidate_lock lock#4 irq_context: 0 mapping.invalidate_lock &xa->xa_lock#9 stock_lock irq_context: 0 mapping.invalidate_lock &xa->xa_lock#9 pool_lock#2 irq_context: 0 mapping.invalidate_lock lock#4 &lruvec->lru_lock irq_context: 0 mapping.invalidate_lock &ei->i_es_lock irq_context: 0 mapping.invalidate_lock tk_core.seq.seqcount irq_context: 0 mapping.invalidate_lock &rq->__lock irq_context: 0 mapping.invalidate_lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 mapping.invalidate_lock &dd->lock irq_context: 0 mapping.invalidate_lock rcu_read_lock &dd->lock irq_context: 0 mapping.invalidate_lock rcu_read_lock &obj_hash[i].lock irq_context: 0 mapping.invalidate_lock rcu_read_lock pool_lock#2 irq_context: 0 mapping.invalidate_lock rcu_read_lock tk_core.seq.seqcount irq_context: 0 mapping.invalidate_lock rcu_read_lock &virtscsi_vq->vq_lock irq_context: 0 mapping.invalidate_lock &folio_wait_table[i] irq_context: 0 cb_lock genl_mutex remove_cache_srcu &base->lock irq_context: 0 cb_lock genl_mutex remove_cache_srcu &base->lock &obj_hash[i].lock irq_context: 0 sb_writers#4 &type->i_mutex_dir_key#3/1 &sb->s_type->i_mutex_key#8 rcu_read_lock &rcu_state.gp_wq irq_context: 0 sb_writers#4 &type->i_mutex_dir_key#3/1 &sb->s_type->i_mutex_key#8 rcu_read_lock &rcu_state.gp_wq &p->pi_lock irq_context: 0 sb_writers#4 &type->i_mutex_dir_key#3/1 &sb->s_type->i_mutex_key#8 rcu_read_lock &rcu_state.gp_wq &p->pi_lock &rq->__lock irq_context: 0 sb_writers#4 &type->i_mutex_dir_key#3/1 &sb->s_type->i_mutex_key#8 rcu_read_lock &rcu_state.gp_wq &p->pi_lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 sk_lock-AF_IEEE802154 irq_context: 0 sk_lock-AF_IEEE802154 slock-AF_IEEE802154 irq_context: 0 slock-AF_IEEE802154 irq_context: softirq (&n->timer) rcu_read_lock rcu_read_lock rcu_read_lock_bh dev->qdisc_tx_busylock ?: &qdisc_tx_busylock quarantine_lock irq_context: 0 &sbi->s_writepages_rwsem jbd2_handle rcu_read_lock rcu_read_lock rcu_read_lock rcu_node_0 irq_context: 0 &sbi->s_writepages_rwsem jbd2_handle rcu_read_lock rcu_read_lock rcu_read_lock &rq->__lock irq_context: 0 sb_writers#4 &sb->s_type->i_mutex_key#8 mapping.invalidate_lock &folio_wait_table[i] irq_context: 0 &nft_net->commit_mutex nf_hook_mutex irq_context: 0 &nft_net->commit_mutex nf_hook_mutex fs_reclaim irq_context: 0 &nft_net->commit_mutex nf_hook_mutex fs_reclaim mmu_notifier_invalidate_range_start irq_context: 0 &nft_net->commit_mutex nf_hook_mutex stock_lock irq_context: 0 &nft_net->commit_mutex nf_hook_mutex pool_lock#2 irq_context: 0 &nft_net->commit_mutex cpu_hotplug_lock irq_context: 0 &nft_net->commit_mutex cpu_hotplug_lock &rq->__lock irq_context: 0 &nft_net->commit_mutex cpu_hotplug_lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 &nft_net->commit_mutex rcu_read_lock &c->lock irq_context: 0 &nft_net->commit_mutex nf_hook_mutex cpu_hotplug_lock irq_context: 0 &nft_net->commit_mutex nf_hook_mutex batched_entropy_u8.lock irq_context: 0 &nft_net->commit_mutex nf_hook_mutex kfence_freelist_lock irq_context: 0 &nft_net->commit_mutex nf_tables_destroy_list_lock irq_context: 0 &nft_net->commit_mutex rcu_read_lock &pool->lock &obj_hash[i].lock irq_context: 0 &nft_net->commit_mutex rcu_read_lock &pool->lock &p->pi_lock irq_context: 0 &nft_net->commit_mutex rcu_read_lock &pool->lock &p->pi_lock &rq->__lock irq_context: 0 &nft_net->commit_mutex rcu_read_lock &pool->lock &p->pi_lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)events trans_destroy_work irq_context: 0 (wq_completion)events trans_destroy_work nf_tables_destroy_list_lock irq_context: 0 (wq_completion)events trans_destroy_work rcu_state.exp_mutex rcu_node_0 irq_context: 0 (wq_completion)events trans_destroy_work rcu_state.exp_mutex &obj_hash[i].lock irq_context: 0 (wq_completion)events trans_destroy_work rcu_state.exp_mutex rcu_read_lock &pool->lock irq_context: 0 (wq_completion)events trans_destroy_work rcu_state.exp_mutex rcu_read_lock &pool->lock &obj_hash[i].lock irq_context: 0 (wq_completion)events trans_destroy_work rcu_state.exp_mutex &rnp->exp_wq[1] irq_context: 0 (wq_completion)events trans_destroy_work rcu_state.exp_mutex &pool->lock irq_context: 0 (wq_completion)events trans_destroy_work rcu_state.exp_mutex &pool->lock &p->pi_lock irq_context: 0 (wq_completion)events trans_destroy_work rcu_state.exp_mutex &pool->lock &p->pi_lock &rq->__lock irq_context: 0 (wq_completion)events trans_destroy_work rcu_state.exp_mutex &pool->lock &p->pi_lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)events trans_destroy_work rcu_state.exp_mutex &rq->__lock irq_context: 0 (wq_completion)events trans_destroy_work rcu_state.exp_mutex &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)events trans_destroy_work &obj_hash[i].lock irq_context: 0 (wq_completion)events trans_destroy_work stock_lock irq_context: 0 (wq_completion)events trans_destroy_work pool_lock#2 irq_context: 0 (wq_completion)events trans_destroy_work (work_completion)(&ht->run_work) irq_context: 0 (wq_completion)events trans_destroy_work rcu_read_lock &pool->lock irq_context: 0 (wq_completion)events trans_destroy_work &ht->mutex irq_context: 0 (wq_completion)events trans_destroy_work &ht->mutex &obj_hash[i].lock irq_context: 0 (wq_completion)events trans_destroy_work &ht->mutex pool_lock#2 irq_context: 0 &nft_net->commit_mutex nf_hook_mutex &c->lock irq_context: 0 &nft_net->commit_mutex rcu_read_lock &n->list_lock irq_context: 0 &nft_net->commit_mutex rcu_read_lock &n->list_lock &c->lock irq_context: 0 &nft_net->commit_mutex nf_hook_mutex &n->list_lock irq_context: 0 &nft_net->commit_mutex nf_hook_mutex &n->list_lock &c->lock irq_context: 0 &nft_net->commit_mutex rcu_state.exp_mutex &cfs_rq->removed.lock irq_context: 0 &nft_net->commit_mutex rcu_state.exp_mutex pool_lock#2 irq_context: 0 (wq_completion)events trans_destroy_work rcu_read_lock rcu_node_0 irq_context: 0 (wq_completion)events trans_destroy_work rcu_read_lock &rq->__lock irq_context: 0 (wq_completion)events trans_destroy_work rcu_read_lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 sb_writers#4 sb_internal rcu_read_lock &rcu_state.gp_wq irq_context: 0 sb_writers#4 sb_internal rcu_read_lock &rcu_state.gp_wq &p->pi_lock irq_context: 0 sb_writers#4 sb_internal rcu_read_lock &rcu_state.gp_wq &p->pi_lock &rq->__lock irq_context: 0 sb_writers#4 sb_internal rcu_read_lock &rcu_state.gp_wq &p->pi_lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 &f->f_pos_lock sb_writers#4 &sb->s_type->i_mutex_key#8 &mm->mmap_lock pgd_lock irq_context: 0 &f->f_pos_lock sb_writers#4 &sb->s_type->i_mutex_key#8 &mm->mmap_lock rcu_read_lock pool_lock#2 irq_context: 0 &f->f_pos_lock sb_writers#4 &sb->s_type->i_mutex_key#8 &mm->mmap_lock key irq_context: 0 &f->f_pos_lock sb_writers#4 &sb->s_type->i_mutex_key#8 &mm->mmap_lock pcpu_lock irq_context: 0 &f->f_pos_lock sb_writers#4 &sb->s_type->i_mutex_key#8 &mm->mmap_lock percpu_counters_lock irq_context: 0 &f->f_pos_lock sb_writers#4 &sb->s_type->i_mutex_key#8 &mm->mmap_lock pcpu_lock stock_lock irq_context: 0 sb_writers#4 &sb->s_type->i_mutex_key#8 mapping.invalidate_lock jbd2_handle &ei->i_data_sem &rq->__lock &cfs_rq->removed.lock irq_context: 0 (wq_completion)netns net_cleanup_work rcu_state.barrier_mutex &cfs_rq->removed.lock irq_context: 0 (wq_completion)netns net_cleanup_work rcu_state.barrier_mutex &obj_hash[i].lock irq_context: 0 (wq_completion)netns net_cleanup_work rcu_state.barrier_mutex pool_lock#2 irq_context: 0 sk_lock-AF_INET6 &tcp_hashinfo.bhash[i].lock &c->lock irq_context: 0 rlock-AF_INET6 irq_context: 0 rtnl_mutex &xs->mutex &lock->wait_lock irq_context: 0 &mm->mmap_lock sb_pagefaults jbd2_handle &journal->j_state_lock irq_context: 0 &mm->mmap_lock sb_pagefaults jbd2_handle &journal->j_state_lock &journal->j_wait_commit irq_context: 0 &mm->mmap_lock sb_pagefaults jbd2_handle &journal->j_state_lock &journal->j_wait_commit &p->pi_lock irq_context: 0 &mm->mmap_lock sb_pagefaults jbd2_handle &journal->j_state_lock &journal->j_wait_commit &p->pi_lock &rq->__lock irq_context: 0 &mm->mmap_lock sb_pagefaults jbd2_handle &journal->j_state_lock &journal->j_wait_commit &p->pi_lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 &mm->mmap_lock sb_pagefaults jbd2_handle &journal->j_wait_updates &p->pi_lock irq_context: 0 &mm->mmap_lock sb_pagefaults jbd2_handle &journal->j_wait_updates &p->pi_lock &rq->__lock irq_context: 0 &mm->mmap_lock sb_pagefaults jbd2_handle &journal->j_wait_updates &p->pi_lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)wg-crypt-wg0 (work_completion)(&peer->transmit_packet_work) &peer->endpoint_lock rcu_read_lock_bh rcu_read_lock &n->list_lock irq_context: 0 (wq_completion)wg-crypt-wg0 (work_completion)(&peer->transmit_packet_work) &peer->endpoint_lock rcu_read_lock_bh rcu_read_lock &n->list_lock &c->lock irq_context: 0 dup_mmap_sem &mm->mmap_lock &mm->mmap_lock/1 &anon_vma->rwsem rcu_read_lock pgd_lock irq_context: 0 dup_mmap_sem &mm->mmap_lock &mm->mmap_lock/1 &anon_vma->rwsem rcu_read_lock stock_lock irq_context: 0 dup_mmap_sem &mm->mmap_lock &mm->mmap_lock/1 &anon_vma->rwsem rcu_read_lock &obj_hash[i].lock irq_context: 0 dup_mmap_sem &mm->mmap_lock &mm->mmap_lock/1 &anon_vma->rwsem rcu_read_lock key irq_context: 0 dup_mmap_sem &mm->mmap_lock &mm->mmap_lock/1 &anon_vma->rwsem rcu_read_lock pcpu_lock irq_context: 0 dup_mmap_sem &mm->mmap_lock &mm->mmap_lock/1 &anon_vma->rwsem rcu_read_lock percpu_counters_lock irq_context: 0 dup_mmap_sem &mm->mmap_lock &mm->mmap_lock/1 &anon_vma->rwsem rcu_read_lock pcpu_lock stock_lock irq_context: 0 rcu_read_lock rcu_read_lock_bh rcu_read_lock_bh rcu_read_lock tk_core.seq.seqcount irq_context: 0 rcu_read_lock rcu_read_lock_bh rcu_read_lock_bh rcu_read_lock &list->lock#5 irq_context: 0 sk_lock-AF_INET6 rcu_read_lock rcu_read_lock &ei->socket.wq.wait irq_context: 0 &xs->mutex remove_cache_srcu rcu_read_lock &____s->seqcount irq_context: 0 (wq_completion)events_unbound (work_completion)(&port->bc_work) rcu_read_lock &c->lock irq_context: 0 &u->iolock &u->peer_wait &p->pi_lock &rq->__lock &cfs_rq->removed.lock irq_context: 0 (wq_completion)events_long &rq->__lock irq_context: 0 rtnl_mutex &xs->mutex &cfs_rq->removed.lock irq_context: 0 rtnl_mutex &xs->mutex &obj_hash[i].lock irq_context: 0 rtnl_mutex &xs->mutex pool_lock#2 irq_context: 0 &u->iolock &lock->wait_lock irq_context: 0 &iint->mutex rcu_node_0 irq_context: 0 sb_writers#4 &type->i_mutex_dir_key#3/1 &rcu_state.expedited_wq &p->pi_lock irq_context: 0 sb_writers#4 &type->i_mutex_dir_key#3/1 &rcu_state.expedited_wq &p->pi_lock &rq->__lock irq_context: 0 sb_writers#4 &type->i_mutex_dir_key#3/1 &rcu_state.expedited_wq &p->pi_lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 &sig->cred_guard_mutex tomoyo_ss rcu_read_lock &rcu_state.expedited_wq irq_context: 0 &sig->cred_guard_mutex tomoyo_ss rcu_read_lock &rcu_state.expedited_wq &p->pi_lock irq_context: 0 &sig->cred_guard_mutex tomoyo_ss rcu_read_lock &rcu_state.expedited_wq &p->pi_lock &rq->__lock irq_context: 0 &sig->cred_guard_mutex tomoyo_ss rcu_read_lock &rcu_state.expedited_wq &p->pi_lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 &u->iolock &rcu_state.expedited_wq &p->pi_lock irq_context: 0 &vma->vm_lock->lock ptlock_ptr(page)#2 rcu_read_lock &p->pi_lock irq_context: 0 &vma->vm_lock->lock ptlock_ptr(page)#2 rcu_read_lock &p->pi_lock &rq->__lock irq_context: 0 &vma->vm_lock->lock ptlock_ptr(page)#2 rcu_read_lock &p->pi_lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 &xs->mutex free_vmap_area_lock &obj_hash[i].lock irq_context: 0 &xs->mutex free_vmap_area_lock pool_lock#2 irq_context: 0 &sig->cred_guard_mutex &ei->xattr_sem rcu_read_lock rcu_node_0 irq_context: 0 &sig->cred_guard_mutex &ei->xattr_sem rcu_read_lock &rq->__lock irq_context: 0 cb_lock genl_mutex &nbd->config_lock (console_sem).lock irq_context: 0 cb_lock genl_mutex &nbd->config_lock console_lock console_srcu console_owner_lock irq_context: 0 cb_lock genl_mutex &nbd->config_lock console_lock console_srcu console_owner irq_context: 0 cb_lock genl_mutex &nbd->config_lock console_lock console_srcu console_owner &port_lock_key irq_context: 0 cb_lock genl_mutex &nbd->config_lock console_lock console_srcu console_owner console_owner_lock irq_context: 0 cb_lock genl_mutex &nbd->config_lock fs_reclaim irq_context: 0 cb_lock genl_mutex &nbd->config_lock fs_reclaim mmu_notifier_invalidate_range_start irq_context: 0 cb_lock genl_mutex &nbd->config_lock uevent_sock_mutex irq_context: 0 cb_lock genl_mutex &nbd->config_lock uevent_sock_mutex fs_reclaim irq_context: 0 cb_lock genl_mutex &nbd->config_lock uevent_sock_mutex fs_reclaim mmu_notifier_invalidate_range_start irq_context: 0 cb_lock genl_mutex &nbd->config_lock uevent_sock_mutex fs_reclaim mmu_notifier_invalidate_range_start &rq->__lock irq_context: 0 cb_lock genl_mutex &nbd->config_lock uevent_sock_mutex fs_reclaim mmu_notifier_invalidate_range_start &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 cb_lock genl_mutex &nbd->config_lock uevent_sock_mutex pool_lock#2 irq_context: 0 cb_lock genl_mutex &nbd->config_lock uevent_sock_mutex nl_table_lock irq_context: 0 cb_lock genl_mutex &nbd->config_lock uevent_sock_mutex &c->lock irq_context: 0 cb_lock genl_mutex &nbd->config_lock uevent_sock_mutex rlock-AF_NETLINK irq_context: 0 cb_lock genl_mutex &nbd->config_lock uevent_sock_mutex &rq->__lock irq_context: 0 cb_lock genl_mutex &nbd->config_lock uevent_sock_mutex &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 cb_lock genl_mutex &nbd->config_lock uevent_sock_mutex rcu_read_lock &ei->socket.wq.wait irq_context: 0 cb_lock genl_mutex &nbd->config_lock uevent_sock_mutex rcu_read_lock &ei->socket.wq.wait &ep->lock irq_context: 0 cb_lock genl_mutex &nbd->config_lock uevent_sock_mutex rcu_read_lock &ei->socket.wq.wait &ep->lock &ep->wq irq_context: 0 cb_lock genl_mutex &nbd->config_lock uevent_sock_mutex rcu_read_lock &ei->socket.wq.wait &ep->lock &ep->wq &p->pi_lock irq_context: 0 cb_lock genl_mutex &nbd->config_lock uevent_sock_mutex rcu_read_lock &ei->socket.wq.wait &ep->lock &ep->wq &p->pi_lock &rq->__lock irq_context: 0 cb_lock genl_mutex &nbd->config_lock uevent_sock_mutex rcu_read_lock &ei->socket.wq.wait &ep->lock &ep->wq &p->pi_lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 cb_lock genl_mutex &nbd->config_lock uevent_sock_mutex rcu_read_lock rcu_node_0 irq_context: 0 cb_lock genl_mutex &nbd->config_lock uevent_sock_mutex rcu_read_lock &rq->__lock irq_context: 0 cb_lock genl_mutex &nbd->config_lock uevent_sock_mutex nl_table_wait.lock irq_context: 0 cb_lock genl_mutex &nbd->config_lock uevent_sock_mutex &obj_hash[i].lock irq_context: 0 cb_lock genl_mutex &nbd->config_lock &ACCESS_PRIVATE(ssp->srcu_sup, lock) rcu_read_lock &pool->lock &p->pi_lock irq_context: 0 cb_lock genl_mutex &nbd->config_lock &ACCESS_PRIVATE(ssp->srcu_sup, lock) rcu_read_lock &pool->lock &p->pi_lock &rq->__lock irq_context: 0 cb_lock genl_mutex &nbd->config_lock &ACCESS_PRIVATE(ssp->srcu_sup, lock) rcu_read_lock &pool->lock &p->pi_lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 &disk->open_mutex nbd_index_mutex &nbd->config_lock &lock->wait_lock irq_context: 0 &disk->open_mutex nbd_index_mutex &nbd->config_lock &rq->__lock irq_context: 0 &disk->open_mutex nbd_index_mutex &nbd->config_lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 cb_lock genl_mutex &lock->wait_lock irq_context: 0 cb_lock genl_mutex &nbd->config_lock &c->lock irq_context: 0 cb_lock genl_mutex &nbd->config_lock &n->list_lock irq_context: 0 cb_lock genl_mutex &nbd->config_lock &n->list_lock &c->lock irq_context: 0 cb_lock genl_mutex &nbd->config_lock uevent_sock_mutex rcu_read_lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 sb_writers#4 &sb->s_type->i_mutex_key#8 mapping.invalidate_lock jbd2_handle &journal->j_state_lock irq_context: 0 sb_writers#4 &sb->s_type->i_mutex_key#8 mapping.invalidate_lock jbd2_handle &journal->j_state_lock &journal->j_wait_commit irq_context: 0 sb_writers#4 &sb->s_type->i_mutex_key#8 mapping.invalidate_lock jbd2_handle &journal->j_state_lock &journal->j_wait_commit &p->pi_lock irq_context: 0 sb_writers#4 &sb->s_type->i_mutex_key#8 mapping.invalidate_lock jbd2_handle &journal->j_state_lock &journal->j_wait_commit &p->pi_lock &rq->__lock irq_context: 0 sb_writers#4 &sb->s_type->i_mutex_key#8 mapping.invalidate_lock jbd2_handle &journal->j_state_lock &journal->j_wait_commit &p->pi_lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 &vma->vm_lock->lock ptlock_ptr(page)#2 rcu_read_lock per_cpu_ptr(&cgroup_rstat_cpu_lock, cpu) irq_context: 0 rtnl_mutex &macsec_netdev_addr_lock_key/1 _xmit_ETHER &local->filter_lock irq_context: 0 rtnl_mutex &macsec_netdev_addr_lock_key/1 _xmit_ETHER rcu_read_lock &pool->lock/1 irq_context: 0 rtnl_mutex &macsec_netdev_addr_lock_key/1 _xmit_ETHER rcu_read_lock &pool->lock/1 &obj_hash[i].lock irq_context: 0 rtnl_mutex &macsec_netdev_addr_lock_key/1 _xmit_ETHER rcu_read_lock &pool->lock/1 &p->pi_lock irq_context: 0 rtnl_mutex &macsec_netdev_addr_lock_key/1 _xmit_ETHER rcu_read_lock &pool->lock/1 &p->pi_lock &rq->__lock irq_context: 0 rtnl_mutex &macsec_netdev_addr_lock_key/1 _xmit_ETHER rcu_read_lock &pool->lock/1 &p->pi_lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)ipv6_addrconf (work_completion)(&(&ifa->dad_work)->work) rtnl_mutex &idev->mc_lock &macsec_netdev_addr_lock_key/1 _xmit_ETHER &local->filter_lock irq_context: 0 (wq_completion)ipv6_addrconf (work_completion)(&(&ifa->dad_work)->work) rtnl_mutex &idev->mc_lock &macsec_netdev_addr_lock_key/1 _xmit_ETHER rcu_read_lock &pool->lock/1 irq_context: 0 (wq_completion)ipv6_addrconf (work_completion)(&(&ifa->dad_work)->work) rtnl_mutex &idev->mc_lock &macsec_netdev_addr_lock_key/1 _xmit_ETHER rcu_read_lock &pool->lock/1 &obj_hash[i].lock irq_context: 0 (wq_completion)ipv6_addrconf (work_completion)(&(&ifa->dad_work)->work) rtnl_mutex &idev->mc_lock &macsec_netdev_addr_lock_key/1 _xmit_ETHER rcu_read_lock &pool->lock/1 &p->pi_lock irq_context: 0 sb_writers#4 sb_internal jbd2_handle &journal->j_state_lock irq_context: 0 sb_writers#4 sb_internal &journal->j_wait_commit irq_context: 0 sb_writers#4 sb_internal &journal->j_wait_done_commit irq_context: 0 sb_writers#3 oom_adj_mutex rcu_read_lock &rcu_state.gp_wq irq_context: 0 sb_writers#3 oom_adj_mutex rcu_read_lock &rcu_state.gp_wq &p->pi_lock irq_context: 0 sb_writers#3 oom_adj_mutex rcu_read_lock &rcu_state.gp_wq &p->pi_lock &rq->__lock irq_context: 0 sb_writers#3 oom_adj_mutex rcu_read_lock &rcu_state.gp_wq &p->pi_lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 &f->f_pos_lock sb_writers#4 &sbi->s_writepages_rwsem jbd2_handle &ei->i_data_sem &lg->lg_mutex key#3 irq_context: 0 &f->f_pos_lock sb_writers#4 &sbi->s_writepages_rwsem jbd2_handle &ei->i_data_sem &ei->i_es_lock key#2 irq_context: 0 &f->f_pos_lock sb_writers#4 &sbi->s_writepages_rwsem remove_cache_srcu irq_context: 0 &f->f_pos_lock sb_writers#4 &sbi->s_writepages_rwsem remove_cache_srcu quarantine_lock irq_context: 0 &f->f_pos_lock sb_writers#4 &sbi->s_writepages_rwsem remove_cache_srcu &c->lock irq_context: 0 &f->f_pos_lock sb_writers#4 &sbi->s_writepages_rwsem remove_cache_srcu &n->list_lock irq_context: 0 &f->f_pos_lock sb_writers#4 &sbi->s_writepages_rwsem remove_cache_srcu rcu_read_lock pool_lock#2 irq_context: 0 &f->f_pos_lock sb_writers#4 &sbi->s_writepages_rwsem remove_cache_srcu &obj_hash[i].lock irq_context: 0 rtnl_mutex &idev->mc_lock &macsec_netdev_addr_lock_key/1 irq_context: 0 rtnl_mutex &idev->mc_lock &macsec_netdev_addr_lock_key/1 &obj_hash[i].lock irq_context: 0 rtnl_mutex &idev->mc_lock &macsec_netdev_addr_lock_key/1 pool_lock#2 irq_context: 0 rtnl_mutex &idev->mc_lock &macsec_netdev_addr_lock_key/1 krc.lock irq_context: 0 sb_writers#4 &sb->s_type->i_mutex_key#8 mapping.invalidate_lock &journal->j_state_lock &journal->j_wait_transaction_locked irq_context: 0 sb_writers#4 &sb->s_type->i_mutex_key#8 mapping.invalidate_lock jbd2_handle &ei->i_data_sem rcu_node_0 irq_context: 0 (wq_completion)events_long (work_completion)(&(&ipvs->defense_work)->work) &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 sb_writers#4 tomoyo_ss rcu_read_lock &rcu_state.expedited_wq irq_context: 0 sb_writers#4 tomoyo_ss rcu_read_lock &rcu_state.expedited_wq &p->pi_lock irq_context: 0 sb_writers#4 tomoyo_ss rcu_read_lock &rcu_state.expedited_wq &p->pi_lock &rq->__lock irq_context: 0 sb_writers#4 tomoyo_ss rcu_read_lock &rcu_state.expedited_wq &p->pi_lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 sb_writers#4 &sbi->s_writepages_rwsem lock#4 &lruvec->lru_lock irq_context: 0 sb_writers#4 &sbi->s_writepages_rwsem jbd2_handle &ei->i_data_sem &c->lock irq_context: 0 sb_writers#4 &sbi->s_writepages_rwsem jbd2_handle &ei->i_data_sem &ret->b_state_lock irq_context: 0 sb_writers#4 &sbi->s_writepages_rwsem jbd2_handle &ei->i_data_sem &ret->b_state_lock &journal->j_list_lock irq_context: 0 sb_writers#4 &sbi->s_writepages_rwsem jbd2_handle &ei->i_data_sem &ei->i_es_lock &obj_hash[i].lock irq_context: 0 sb_writers#4 tomoyo_ss rcu_read_lock &cfs_rq->removed.lock irq_context: 0 sb_writers#4 tomoyo_ss rcu_read_lock &obj_hash[i].lock irq_context: 0 sb_writers#4 tomoyo_ss rcu_read_lock pool_lock#2 irq_context: 0 sb_writers#4 &sbi->s_writepages_rwsem &c->lock irq_context: 0 sb_writers#4 &sbi->s_writepages_rwsem jbd2_handle rcu_read_lock &cfs_rq->removed.lock irq_context: 0 sb_writers#4 &type->i_mutex_dir_key#3 batched_entropy_u8.lock irq_context: 0 sb_writers#4 &type->i_mutex_dir_key#3 kfence_freelist_lock irq_context: 0 &xs->mutex remove_cache_srcu &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 &xs->mutex remove_cache_srcu rcu_read_lock &rq->__lock irq_context: 0 &xs->mutex remove_cache_srcu rcu_read_lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 &xs->mutex remove_cache_srcu rcu_read_lock rcu_node_0 irq_context: 0 lock map_idr_lock &c->lock irq_context: 0 &head->lock irq_context: 0 rcu_read_lock rcu_read_lock_bh rcu_read_lock &dir->lock#2 irq_context: 0 rcu_read_lock rcu_read_lock_bh rcu_read_lock &ul->lock irq_context: 0 rcu_read_lock rcu_read_lock_bh batched_entropy_u32.lock irq_context: 0 rcu_read_lock rcu_read_lock_bh rcu_read_lock &____s->seqcount#7 irq_context: 0 rcu_read_lock rcu_read_lock_bh rcu_read_lock &tbl->lock irq_context: 0 rcu_read_lock rcu_read_lock_bh rcu_read_lock rcu_read_lock_bh &obj_hash[i].lock irq_context: 0 rcu_read_lock rcu_read_lock_bh rcu_read_lock rcu_read_lock_bh pool_lock#2 irq_context: 0 rcu_read_lock rcu_read_lock_bh &n->list_lock irq_context: 0 rcu_read_lock rcu_read_lock_bh &n->list_lock &c->lock irq_context: 0 rcu_read_lock rcu_read_lock_bh &____s->seqcount#2 irq_context: 0 rcu_read_lock rcu_read_lock_bh &____s->seqcount irq_context: 0 rcu_read_lock rcu_read_lock_bh rcu_read_lock rcu_read_lock_bh &obj_hash[i].lock pool_lock irq_context: 0 rcu_read_lock rcu_read_lock_bh rcu_read_lock rcu_read_lock_bh quarantine_lock irq_context: 0 rcu_read_lock rcu_read_lock_bh rcu_read_lock rcu_read_lock_bh &meta->lock irq_context: 0 rcu_read_lock rcu_read_lock_bh rcu_read_lock rcu_read_lock_bh kfence_freelist_lock irq_context: 0 rcu_read_lock rcu_read_lock_bh rcu_read_lock key#23 irq_context: 0 rcu_read_lock rcu_read_lock_bh &pcp->lock &zone->lock irq_context: softirq &(&bat_priv->nc.work)->timer rcu_read_lock &pool->lock/1 fill_pool_map-wait-type-override &n->list_lock irq_context: softirq &(&bat_priv->nc.work)->timer rcu_read_lock &pool->lock/1 fill_pool_map-wait-type-override &n->list_lock &c->lock irq_context: 0 rcu_read_lock rcu_read_lock_bh rcu_read_lock rcu_read_lock_bh fill_pool_map-wait-type-override pool_lock#2 irq_context: 0 rcu_read_lock rcu_read_lock_bh rcu_read_lock rcu_read_lock_bh fill_pool_map-wait-type-override pool_lock irq_context: 0 cb_lock &fw_cache.lock irq_context: 0 cb_lock &fw_cache.lock pool_lock#2 irq_context: 0 cb_lock stock_lock irq_context: 0 cb_lock tk_core.seq.seqcount irq_context: 0 cb_lock async_lock irq_context: 0 cb_lock init_task.alloc_lock irq_context: 0 cb_lock init_task.alloc_lock init_fs.lock irq_context: 0 cb_lock init_task.alloc_lock init_fs.lock &dentry->d_lock irq_context: 0 cb_lock rcu_read_lock &____s->seqcount#4 irq_context: 0 cb_lock rcu_read_lock &dentry->d_lock irq_context: 0 cb_lock &type->i_mutex_dir_key#3 irq_context: 0 cb_lock &type->i_mutex_dir_key#3 fs_reclaim irq_context: 0 cb_lock &type->i_mutex_dir_key#3 fs_reclaim mmu_notifier_invalidate_range_start irq_context: 0 cb_lock &type->i_mutex_dir_key#3 stock_lock irq_context: 0 cb_lock &type->i_mutex_dir_key#3 pool_lock#2 irq_context: 0 cb_lock &type->i_mutex_dir_key#3 &dentry->d_lock irq_context: 0 cb_lock &type->i_mutex_dir_key#3 rcu_read_lock rename_lock.seqcount irq_context: 0 cb_lock &type->i_mutex_dir_key#3 &ei->i_es_lock irq_context: 0 cb_lock &type->i_mutex_dir_key#3 &mapping->private_lock irq_context: 0 cb_lock &type->i_mutex_dir_key#3 &rq->__lock irq_context: 0 cb_lock &type->i_mutex_dir_key#3 &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 rcu_read_lock rcu_read_lock_bh rcu_read_lock rcu_read_lock_bh fill_pool_map-wait-type-override &c->lock irq_context: 0 cb_lock &type->i_mutex_dir_key#3 inode_hash_lock irq_context: 0 cb_lock &type->i_mutex_dir_key#3 mmu_notifier_invalidate_range_start irq_context: 0 cb_lock &type->i_mutex_dir_key#3 &c->lock irq_context: 0 cb_lock &type->i_mutex_dir_key#3 &obj_hash[i].lock irq_context: 0 cb_lock &type->i_mutex_dir_key#3 inode_hash_lock &sb->s_type->i_lock_key#22 irq_context: 0 cb_lock &type->i_mutex_dir_key#3 inode_hash_lock &s->s_inode_list_lock irq_context: 0 cb_lock &type->i_mutex_dir_key#3 &journal->j_state_lock irq_context: 0 cb_lock &type->i_mutex_dir_key#3 &sb->s_type->i_lock_key#22 irq_context: 0 cb_lock &type->i_mutex_dir_key#3 &sb->s_type->i_lock_key#22 &dentry->d_lock irq_context: 0 cb_lock &type->i_mutex_dir_key#3 &sb->s_type->i_lock_key#22 &dentry->d_lock &wq irq_context: 0 cb_lock &sb->s_type->i_lock_key#22 irq_context: 0 cb_lock &type->i_mutex_dir_key#3 &ei->i_data_sem irq_context: 0 cb_lock &type->i_mutex_dir_key#3 &ei->i_data_sem mmu_notifier_invalidate_range_start irq_context: 0 cb_lock &type->i_mutex_dir_key#3 &ei->i_data_sem pool_lock#2 irq_context: 0 cb_lock &type->i_mutex_dir_key#3 &ei->i_data_sem &ei->i_es_lock irq_context: 0 cb_lock &type->i_mutex_dir_key#3 &ei->i_data_sem &ei->i_es_lock pool_lock#2 irq_context: 0 cb_lock &type->i_mutex_dir_key#3 &ei->i_data_sem &ei->i_es_lock &sbi->s_es_lock irq_context: 0 cb_lock &type->i_mutex_dir_key#3 &ei->i_data_sem &obj_hash[i].lock irq_context: 0 cb_lock &type->i_mutex_dir_key#3 &ei->i_data_sem &ei->i_es_lock &obj_hash[i].lock irq_context: 0 cb_lock &type->i_mutex_dir_key#3 &____s->seqcount irq_context: 0 cb_lock &type->i_mutex_dir_key#3 &xa->xa_lock#9 irq_context: 0 cb_lock &type->i_mutex_dir_key#3 lock#4 irq_context: 0 cb_lock &type->i_mutex_dir_key#3 tk_core.seq.seqcount irq_context: 0 cb_lock &type->i_mutex_dir_key#3 &dd->lock irq_context: 0 cb_lock &type->i_mutex_dir_key#3 rcu_read_lock &pool->lock irq_context: 0 cb_lock &type->i_mutex_dir_key#3 rcu_read_lock &pool->lock &obj_hash[i].lock irq_context: 0 cb_lock &type->i_mutex_dir_key#3 rcu_read_lock &pool->lock pool_lock#2 irq_context: 0 cb_lock &type->i_mutex_dir_key#3 rcu_read_lock &pool->lock &p->pi_lock irq_context: 0 cb_lock &type->i_mutex_dir_key#3 rcu_read_lock &pool->lock &p->pi_lock &rq->__lock irq_context: 0 cb_lock &type->i_mutex_dir_key#3 rcu_read_lock &pool->lock &p->pi_lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 cb_lock &type->i_mutex_dir_key#3 bit_wait_table + i irq_context: 0 cb_lock &type->i_mutex_dir_key#3 &dentry->d_lock &wq irq_context: 0 cb_lock &dentry->d_lock irq_context: 0 cb_lock &dentry->d_lock &lru->node[i].lock irq_context: 0 cb_lock &type->i_mutex_dir_key#3 rename_lock.seqcount irq_context: 0 cb_lock &type->i_mutex_dir_key#3 &dentry->d_lock &wq#2 irq_context: 0 nlk_cb_mutex-GENERIC irq_context: 0 nlk_cb_mutex-GENERIC fs_reclaim irq_context: 0 nlk_cb_mutex-GENERIC fs_reclaim mmu_notifier_invalidate_range_start irq_context: 0 nlk_cb_mutex-GENERIC fs_reclaim mmu_notifier_invalidate_range_start &rq->__lock irq_context: 0 nlk_cb_mutex-GENERIC fs_reclaim mmu_notifier_invalidate_range_start &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 nlk_cb_mutex-GENERIC pool_lock#2 irq_context: 0 nlk_cb_mutex-GENERIC genl_mutex irq_context: 0 nlk_cb_mutex-GENERIC genl_mutex rtnl_mutex irq_context: 0 nlk_cb_mutex-GENERIC genl_mutex rtnl_mutex.wait_lock irq_context: 0 nlk_cb_mutex-GENERIC genl_mutex &p->pi_lock irq_context: 0 nlk_cb_mutex-GENERIC rcu_read_lock pool_lock#2 irq_context: 0 nlk_cb_mutex-GENERIC rcu_read_lock rcu_read_lock_bh &obj_hash[i].lock irq_context: 0 nlk_cb_mutex-GENERIC rcu_read_lock rcu_read_lock_bh pool_lock#2 irq_context: 0 nlk_cb_mutex-GENERIC rlock-AF_NETLINK irq_context: 0 nlk_cb_mutex-GENERIC &obj_hash[i].lock irq_context: 0 nlk_cb_mutex-GENERIC &rq->__lock irq_context: 0 nlk_cb_mutex-GENERIC &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 rcu_read_lock rcu_read_lock_bh batched_entropy_u32.lock crngs.lock irq_context: 0 (wq_completion)mld (work_completion)(&(&idev->mc_ifc_work)->work) &idev->mc_lock rcu_read_lock rcu_read_lock rcu_read_lock_bh _xmit_ETHER#2 rcu_read_lock &local->handle_wake_tx_queue_lock hwsim_radio_lock &n->list_lock irq_context: 0 (wq_completion)mld (work_completion)(&(&idev->mc_ifc_work)->work) &idev->mc_lock rcu_read_lock rcu_read_lock rcu_read_lock_bh _xmit_ETHER#2 rcu_read_lock &local->handle_wake_tx_queue_lock hwsim_radio_lock &n->list_lock &c->lock irq_context: 0 cb_lock umhelper_sem irq_context: 0 cb_lock umhelper_sem usermodehelper_disabled_waitq.lock irq_context: 0 cb_lock umhelper_sem fs_reclaim irq_context: 0 cb_lock umhelper_sem fs_reclaim mmu_notifier_invalidate_range_start irq_context: 0 cb_lock umhelper_sem fs_reclaim mmu_notifier_invalidate_range_start &rq->__lock irq_context: 0 cb_lock umhelper_sem fs_reclaim mmu_notifier_invalidate_range_start &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 cb_lock umhelper_sem &c->lock irq_context: 0 cb_lock umhelper_sem pool_lock#2 irq_context: 0 cb_lock umhelper_sem &x->wait#9 irq_context: 0 cb_lock umhelper_sem &obj_hash[i].lock irq_context: 0 cb_lock umhelper_sem &obj_hash[i].lock pool_lock irq_context: 0 cb_lock umhelper_sem &k->list_lock irq_context: 0 cb_lock umhelper_sem &rq->__lock irq_context: 0 cb_lock umhelper_sem &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 cb_lock umhelper_sem gdp_mutex irq_context: 0 cb_lock umhelper_sem gdp_mutex &k->list_lock irq_context: 0 cb_lock umhelper_sem gdp_mutex fs_reclaim irq_context: 0 cb_lock umhelper_sem gdp_mutex fs_reclaim mmu_notifier_invalidate_range_start irq_context: 0 cb_lock umhelper_sem gdp_mutex &c->lock irq_context: 0 cb_lock umhelper_sem gdp_mutex pool_lock#2 irq_context: 0 cb_lock umhelper_sem gdp_mutex lock irq_context: 0 cb_lock umhelper_sem gdp_mutex lock kernfs_idr_lock irq_context: 0 cb_lock umhelper_sem gdp_mutex lock kernfs_idr_lock &c->lock irq_context: 0 cb_lock umhelper_sem gdp_mutex lock kernfs_idr_lock pool_lock#2 irq_context: 0 cb_lock umhelper_sem gdp_mutex &root->kernfs_rwsem irq_context: 0 cb_lock umhelper_sem gdp_mutex &root->kernfs_rwsem &root->kernfs_iattr_rwsem irq_context: 0 cb_lock umhelper_sem lock irq_context: 0 cb_lock umhelper_sem lock kernfs_idr_lock irq_context: 0 cb_lock umhelper_sem &root->kernfs_rwsem irq_context: 0 cb_lock umhelper_sem &root->kernfs_rwsem &root->kernfs_iattr_rwsem irq_context: 0 cb_lock umhelper_sem bus_type_sem irq_context: 0 cb_lock umhelper_sem sysfs_symlink_target_lock irq_context: 0 rcu_read_lock rcu_read_lock_bh rcu_read_lock rcu_read_lock_bh fill_pool_map-wait-type-override &n->list_lock irq_context: 0 rcu_read_lock rcu_read_lock_bh rcu_read_lock rcu_read_lock_bh fill_pool_map-wait-type-override &n->list_lock &c->lock irq_context: 0 cb_lock umhelper_sem &root->kernfs_rwsem irq_context: 0 cb_lock umhelper_sem &dev->power.lock irq_context: 0 cb_lock umhelper_sem dpm_list_mtx irq_context: 0 cb_lock umhelper_sem &k->k_lock irq_context: 0 cb_lock umhelper_sem subsys mutex#80 irq_context: 0 cb_lock umhelper_sem subsys mutex#80 &k->k_lock irq_context: 0 cb_lock umhelper_sem fw_lock irq_context: 0 cb_lock umhelper_sem uevent_sock_mutex irq_context: 0 cb_lock umhelper_sem uevent_sock_mutex fs_reclaim irq_context: 0 cb_lock umhelper_sem uevent_sock_mutex fs_reclaim mmu_notifier_invalidate_range_start irq_context: 0 cb_lock umhelper_sem uevent_sock_mutex pool_lock#2 irq_context: 0 cb_lock umhelper_sem uevent_sock_mutex &rq->__lock irq_context: 0 cb_lock umhelper_sem uevent_sock_mutex &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)events_unbound (work_completion)(&map->work) per_cpu_ptr(&cgroup_rstat_cpu_lock, cpu) irq_context: 0 (wq_completion)events_unbound (work_completion)(&map->work) pcpu_lock irq_context: 0 (wq_completion)events_unbound (work_completion)(&map->work) rcu_state.exp_mutex rcu_node_0 irq_context: 0 (wq_completion)events_unbound (work_completion)(&map->work) rcu_state.exp_mutex fill_pool_map-wait-type-override pool_lock#2 irq_context: 0 (wq_completion)events_unbound (work_completion)(&map->work) rcu_state.exp_mutex fill_pool_map-wait-type-override &c->lock irq_context: 0 (wq_completion)events_unbound (work_completion)(&map->work) rcu_state.exp_mutex fill_pool_map-wait-type-override pool_lock irq_context: 0 (wq_completion)events_unbound (work_completion)(&map->work) rcu_state.exp_mutex &obj_hash[i].lock irq_context: 0 (wq_completion)events_unbound (work_completion)(&map->work) rcu_state.exp_mutex rcu_read_lock &pool->lock irq_context: 0 (wq_completion)events_unbound (work_completion)(&map->work) rcu_state.exp_mutex rcu_read_lock &pool->lock &obj_hash[i].lock irq_context: 0 (wq_completion)events_unbound (work_completion)(&map->work) rcu_state.exp_mutex rcu_read_lock &pool->lock &p->pi_lock irq_context: 0 (wq_completion)events_unbound (work_completion)(&map->work) rcu_state.exp_mutex rcu_read_lock &pool->lock &p->pi_lock &rq->__lock irq_context: 0 (wq_completion)events_unbound (work_completion)(&map->work) rcu_state.exp_mutex rcu_read_lock &pool->lock &p->pi_lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 cb_lock umhelper_sem uevent_sock_mutex nl_table_lock irq_context: 0 cb_lock umhelper_sem uevent_sock_mutex rlock-AF_NETLINK irq_context: 0 cb_lock umhelper_sem uevent_sock_mutex rcu_read_lock &ei->socket.wq.wait irq_context: 0 cb_lock umhelper_sem uevent_sock_mutex rcu_read_lock &ei->socket.wq.wait &ep->lock irq_context: 0 cb_lock umhelper_sem uevent_sock_mutex rcu_read_lock &ei->socket.wq.wait &ep->lock &ep->wq irq_context: 0 cb_lock umhelper_sem uevent_sock_mutex rcu_read_lock &ei->socket.wq.wait &ep->lock &ep->wq &p->pi_lock irq_context: 0 (wq_completion)events_unbound (work_completion)(&map->work) rcu_state.exp_mutex &rq->__lock irq_context: 0 cb_lock umhelper_sem uevent_sock_mutex rcu_read_lock &ei->socket.wq.wait &ep->lock &ep->wq &p->pi_lock &rq->__lock irq_context: 0 (wq_completion)events_unbound (work_completion)(&map->work) rcu_state.exp_mutex &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 cb_lock umhelper_sem uevent_sock_mutex rcu_read_lock &ei->socket.wq.wait &ep->lock &ep->wq &p->pi_lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 cb_lock umhelper_sem uevent_sock_mutex rcu_read_lock rcu_node_0 irq_context: 0 cb_lock umhelper_sem uevent_sock_mutex rcu_read_lock &rq->__lock irq_context: 0 cb_lock umhelper_sem uevent_sock_mutex rcu_read_lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)events_unbound (work_completion)(&map->work) rcu_state.exp_mutex &rnp->exp_wq[0] irq_context: 0 cb_lock umhelper_sem uevent_sock_mutex nl_table_wait.lock irq_context: 0 cb_lock umhelper_sem uevent_sock_mutex &obj_hash[i].lock irq_context: 0 cb_lock umhelper_sem &x->wait#23 irq_context: 0 cb_lock umhelper_sem &base->lock irq_context: 0 cb_lock umhelper_sem &base->lock &obj_hash[i].lock irq_context: 0 rcu_read_lock rcu_read_lock_bh rcu_read_lock rcu_read_lock_bh &____s->seqcount irq_context: 0 rcu_read_lock rcu_read_lock_bh rcu_read_lock rcu_read_lock_bh rcu_read_lock pool_lock#2 irq_context: softirq (&peer->timer_retransmit_handshake) &list->lock#17 &obj_hash[i].lock irq_context: softirq (&peer->timer_retransmit_handshake) &list->lock#17 pool_lock#2 irq_context: 0 (wq_completion)events_unbound (work_completion)(&map->work) rcu_state.exp_mutex &rnp->exp_wq[2] irq_context: 0 (wq_completion)events_unbound (work_completion)(&map->work) &rnp->exp_lock irq_context: 0 (wq_completion)events_unbound (work_completion)(&map->work) rcu_state.exp_mutex irq_context: 0 (wq_completion)events_unbound (work_completion)(&map->work) rcu_state.exp_mutex rcu_state.exp_mutex.wait_lock irq_context: 0 (wq_completion)events_unbound (work_completion)(&map->work) rcu_state.exp_mutex.wait_lock irq_context: 0 (wq_completion)events_unbound (work_completion)(&map->work) &p->pi_lock irq_context: 0 (wq_completion)events_unbound (work_completion)(&map->work) &p->pi_lock &rq->__lock irq_context: 0 (wq_completion)events_unbound (work_completion)(&map->work) &p->pi_lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)events_unbound (work_completion)(&map->work) rcu_state.exp_mutex &rnp->exp_wq[3] irq_context: 0 (wq_completion)events_unbound (work_completion)(&map->work) &obj_hash[i].lock pool_lock irq_context: 0 &xt[i].mutex rcu_read_lock pgd_lock irq_context: 0 &xt[i].mutex rcu_read_lock key irq_context: 0 &xt[i].mutex rcu_read_lock pcpu_lock irq_context: 0 &xt[i].mutex rcu_read_lock percpu_counters_lock irq_context: 0 (wq_completion)events_unbound (work_completion)(&map->work) rcu_state.exp_mutex &rnp->exp_wq[1] irq_context: 0 cb_lock umhelper_sem &cfs_rq->removed.lock irq_context: 0 cb_lock umhelper_sem (&timer.timer) irq_context: 0 cb_lock umhelper_sem fw_lock &x->wait#23 irq_context: 0 cb_lock umhelper_sem dev_pm_qos_sysfs_mtx irq_context: 0 cb_lock umhelper_sem dev_pm_qos_sysfs_mtx &root->kernfs_rwsem irq_context: 0 cb_lock umhelper_sem dev_pm_qos_sysfs_mtx &root->kernfs_rwsem irq_context: 0 cb_lock umhelper_sem dev_pm_qos_sysfs_mtx dev_pm_qos_mtx irq_context: 0 cb_lock umhelper_sem &root->kernfs_rwsem kernfs_idr_lock irq_context: 0 cb_lock umhelper_sem &root->kernfs_rwsem &obj_hash[i].lock irq_context: 0 cb_lock umhelper_sem &root->kernfs_rwsem pool_lock#2 irq_context: 0 cb_lock umhelper_sem kernfs_idr_lock irq_context: 0 cb_lock umhelper_sem &k->k_lock klist_remove_lock irq_context: 0 cb_lock umhelper_sem &root->kernfs_rwsem &rq->__lock irq_context: 0 cb_lock umhelper_sem subsys mutex#80 &k->k_lock klist_remove_lock irq_context: 0 cb_lock umhelper_sem deferred_probe_mutex irq_context: 0 cb_lock umhelper_sem device_links_lock irq_context: 0 cb_lock umhelper_sem mmu_notifier_invalidate_range_start irq_context: 0 cb_lock umhelper_sem uevent_sock_mutex mmu_notifier_invalidate_range_start irq_context: 0 cb_lock umhelper_sem gdp_mutex sysfs_symlink_target_lock irq_context: 0 cb_lock umhelper_sem gdp_mutex &obj_hash[i].lock irq_context: 0 cb_lock fw_lock irq_context: 0 cb_lock console_owner_lock irq_context: 0 cb_lock console_owner irq_context: 0 &xs->mutex console_owner_lock irq_context: 0 &xs->mutex console_owner irq_context: 0 &sig->cred_guard_mutex &sig->exec_update_lock stock_lock irq_context: 0 &sig->cred_guard_mutex &sig->exec_update_lock pcpu_lock stock_lock irq_context: 0 cb_lock umhelper_sem &n->list_lock irq_context: 0 cb_lock umhelper_sem &n->list_lock &c->lock irq_context: 0 cb_lock umhelper_sem gdp_mutex &rq->__lock irq_context: 0 cb_lock umhelper_sem gdp_mutex &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 cb_lock umhelper_sem &root->kernfs_rwsem &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 cb_lock umhelper_sem &root->kernfs_rwsem &rq->__lock irq_context: 0 cb_lock umhelper_sem &root->kernfs_rwsem &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 cb_lock umhelper_sem &____s->seqcount#2 irq_context: 0 cb_lock umhelper_sem &____s->seqcount irq_context: 0 cb_lock umhelper_sem uevent_sock_mutex &c->lock irq_context: 0 cb_lock umhelper_sem gdp_mutex kernfs_idr_lock irq_context: 0 cb_lock umhelper_sem gdp_mutex kernfs_idr_lock &obj_hash[i].lock irq_context: 0 cb_lock umhelper_sem gdp_mutex kernfs_idr_lock pool_lock#2 irq_context: 0 cb_lock &x->wait#23 irq_context: 0 (wq_completion)writeback (work_completion)(&(&wb->dwork)->work) rcu_read_lock &obj_hash[i].lock pool_lock irq_context: 0 nlk_cb_mutex-GENERIC &c->lock irq_context: 0 nlk_cb_mutex-GENERIC rcu_read_lock &c->lock irq_context: 0 cb_lock umhelper_sem fw_lock &x->wait#23 &p->pi_lock irq_context: 0 cb_lock umhelper_sem fw_lock &x->wait#23 &p->pi_lock &cfs_rq->removed.lock irq_context: 0 cb_lock umhelper_sem fw_lock &x->wait#23 &p->pi_lock &rq->__lock irq_context: 0 cb_lock umhelper_sem fw_lock &x->wait#23 &p->pi_lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 cb_lock umhelper_sem fw_lock &rq->__lock irq_context: 0 cb_lock umhelper_sem fw_lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 cb_lock fw_lock fw_lock.wait_lock irq_context: 0 cb_lock fw_lock &rq->__lock irq_context: 0 cb_lock fw_lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 cb_lock umhelper_sem fw_lock.wait_lock irq_context: 0 cb_lock umhelper_sem &p->pi_lock irq_context: 0 cb_lock umhelper_sem &p->pi_lock &rq->__lock irq_context: 0 cb_lock umhelper_sem &p->pi_lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 &dentry->d_lock &sb->s_type->i_lock_key#24 &dentry->d_lock &lru->node[i].lock irq_context: 0 sb_writers#4 &sb->s_type->i_mutex_key#8 mapping.invalidate_lock jbd2_handle &ei->i_data_sem mmu_notifier_invalidate_range_start &rq->__lock irq_context: 0 sb_writers#4 &sb->s_type->i_mutex_key#8 mapping.invalidate_lock jbd2_handle &ei->i_data_sem mmu_notifier_invalidate_range_start &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 sb_writers#4 &sb->s_type->i_mutex_key#8 mapping.invalidate_lock jbd2_handle &ei->i_data_sem rcu_read_lock &rcu_state.gp_wq irq_context: 0 sb_writers#4 &sb->s_type->i_mutex_key#8 mapping.invalidate_lock jbd2_handle &ei->i_data_sem rcu_read_lock &rcu_state.gp_wq &p->pi_lock irq_context: 0 sb_writers#4 &sb->s_type->i_mutex_key#8 mapping.invalidate_lock jbd2_handle &ei->i_data_sem rcu_read_lock &rcu_state.gp_wq &p->pi_lock &rq->__lock irq_context: 0 sb_writers#4 &sb->s_type->i_mutex_key#8 mapping.invalidate_lock jbd2_handle &ei->i_data_sem rcu_read_lock &rcu_state.gp_wq &p->pi_lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 rtnl_mutex &idev->mc_lock &bridge_netdev_addr_lock_key &obj_hash[i].lock pool_lock irq_context: 0 (wq_completion)mld (work_completion)(&(&idev->mc_ifc_work)->work) &idev->mc_lock remove_cache_srcu rcu_read_lock &____s->seqcount irq_context: 0 rtnl_mutex &ul->lock#2 &c->lock irq_context: 0 sb_writers#4 &type->i_mutex_dir_key#3/1 &sb->s_type->i_mutex_key#8 jbd2_handle &rcu_state.expedited_wq irq_context: 0 sb_writers#4 &type->i_mutex_dir_key#3/1 &sb->s_type->i_mutex_key#8 jbd2_handle &rcu_state.expedited_wq &p->pi_lock irq_context: 0 sb_writers#4 &type->i_mutex_dir_key#3/1 &sb->s_type->i_mutex_key#8 jbd2_handle &rcu_state.expedited_wq &p->pi_lock &rq->__lock irq_context: 0 sb_writers#4 &type->i_mutex_dir_key#3/1 &sb->s_type->i_mutex_key#8 jbd2_handle &rcu_state.expedited_wq &p->pi_lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)ipv6_addrconf (work_completion)(&(&ifa->dad_work)->work) rtnl_mutex &idev->mc_lock rcu_read_lock rcu_read_lock rcu_read_lock_bh rcu_read_lock &br->multicast_lock pool_lock#2 irq_context: 0 (wq_completion)ipv6_addrconf (work_completion)(&(&ifa->dad_work)->work) rtnl_mutex &idev->mc_lock rcu_read_lock rcu_read_lock rcu_read_lock_bh rcu_read_lock &br->multicast_lock rcu_read_lock rhashtable_bucket irq_context: 0 (wq_completion)ipv6_addrconf (work_completion)(&(&ifa->dad_work)->work) rtnl_mutex &idev->mc_lock rcu_read_lock rcu_read_lock rcu_read_lock_bh rcu_read_lock &br->multicast_lock nl_table_lock irq_context: 0 (wq_completion)ipv6_addrconf (work_completion)(&(&ifa->dad_work)->work) rtnl_mutex &idev->mc_lock rcu_read_lock rcu_read_lock rcu_read_lock_bh rcu_read_lock &br->multicast_lock nl_table_wait.lock irq_context: 0 &xs->mutex &base->lock irq_context: 0 &xs->mutex &base->lock &obj_hash[i].lock irq_context: 0 sk_lock-AF_INET6 &meta->lock irq_context: 0 &f->f_pos_lock sb_writers#4 &sb->s_type->i_mutex_key#8 &mm->mmap_lock &p->pi_lock irq_context: 0 sb_writers#4 &type->i_mutex_dir_key#3/1 sb_internal jbd2_handle &sbi->s_orphan_lock &ret->b_state_lock &journal->j_list_lock irq_context: 0 &pipe->mutex/1 sk_lock-AF_INET6 rcu_read_lock &nf_conntrack_locks[i] batched_entropy_u8.lock crngs.lock irq_context: 0 &f->f_pos_lock sb_writers#4 &sb->s_type->i_mutex_key#8 irq_context: 0 &f->f_pos_lock sb_writers#4 &sb->s_type->i_mutex_key#8 &sem->wait_lock irq_context: 0 &f->f_pos_lock sb_writers#4 &sb->s_type->i_mutex_key#8 &rq->__lock irq_context: 0 &f->f_pos_lock sb_writers#4 &sb->s_type->i_mutex_key#8 &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 sb_writers#4 &sb->s_type->i_mutex_key#8 &sbi->s_writepages_rwsem rcu_state.exp_mutex &cfs_rq->removed.lock irq_context: 0 sb_writers#4 &sb->s_type->i_mutex_key#8 &sbi->s_writepages_rwsem jbd2_handle &ei->i_data_sem &ret->b_state_lock &journal->j_list_lock irq_context: 0 &f->f_pos_lock sb_writers#4 &sb->s_type->i_mutex_key#8 &ei->i_es_lock irq_context: 0 &sb->s_type->i_mutex_key#10 rcu_read_lock &rcu_state.gp_wq &p->pi_lock &rq->__lock &cfs_rq->removed.lock irq_context: 0 &xa->xa_lock#9 &n->list_lock &c->lock irq_context: 0 &p->lock remove_cache_srcu rcu_read_lock &cfs_rq->removed.lock irq_context: 0 &p->lock remove_cache_srcu rcu_read_lock &obj_hash[i].lock irq_context: 0 sb_writers#4 &type->i_mutex_dir_key#3/1 jbd2_handle &journal->j_wait_updates &p->pi_lock &cfs_rq->removed.lock irq_context: 0 &f->f_pos_lock sb_writers#4 &sb->s_type->i_mutex_key#8 tk_core.seq.seqcount irq_context: 0 &f->f_pos_lock sb_writers#4 &sb->s_type->i_mutex_key#8 mmu_notifier_invalidate_range_start irq_context: 0 &f->f_pos_lock sb_writers#4 &sb->s_type->i_mutex_key#8 pool_lock#2 irq_context: 0 &f->f_pos_lock sb_writers#4 &sb->s_type->i_mutex_key#8 &journal->j_state_lock irq_context: 0 &f->f_pos_lock sb_writers#4 &sb->s_type->i_mutex_key#8 jbd2_handle irq_context: 0 &f->f_pos_lock sb_writers#4 &sb->s_type->i_mutex_key#8 jbd2_handle &ei->i_raw_lock irq_context: 0 &f->f_pos_lock sb_writers#4 &sb->s_type->i_mutex_key#8 jbd2_handle &journal->j_wait_updates irq_context: 0 &f->f_pos_lock sb_writers#4 &sb->s_type->i_mutex_key#8 &obj_hash[i].lock irq_context: 0 &f->f_pos_lock sb_writers#4 &sb->s_type->i_mutex_key#8 fs_reclaim irq_context: 0 &f->f_pos_lock sb_writers#4 &sb->s_type->i_mutex_key#8 fs_reclaim mmu_notifier_invalidate_range_start irq_context: 0 &f->f_pos_lock sb_writers#4 &sb->s_type->i_mutex_key#8 fs_reclaim mmu_notifier_invalidate_range_start &rq->__lock irq_context: 0 &f->f_pos_lock sb_writers#4 &sb->s_type->i_mutex_key#8 fs_reclaim mmu_notifier_invalidate_range_start &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 &f->f_pos_lock sb_writers#4 &sb->s_type->i_mutex_key#8 &sb->s_type->i_lock_key#22 irq_context: 0 &f->f_pos_lock sb_writers#4 &sb->s_type->i_mutex_key#8 &sb->s_type->i_lock_key#22 &xa->xa_lock#9 irq_context: 0 &f->f_pos_lock sb_writers#4 &sb->s_type->i_mutex_key#8 lock#4 irq_context: 0 &f->f_pos_lock sb_writers#4 &sb->s_type->i_mutex_key#8 lock#4 &lruvec->lru_lock irq_context: 0 &f->f_pos_lock sb_writers#4 &sb->s_type->i_mutex_key#8 lock#5 irq_context: 0 &f->f_pos_lock sb_writers#4 &sb->s_type->i_mutex_key#8 &lruvec->lru_lock irq_context: 0 &f->f_pos_lock sb_writers#4 &sb->s_type->i_mutex_key#8 &mm->mmap_lock irq_context: 0 rtnl_mutex cpu_hotplug_lock wq_pool_mutex &rq->__lock irq_context: 0 rtnl_mutex cpu_hotplug_lock wq_pool_mutex &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 rtnl_mutex &bat_priv->softif_vlan_list_lock &n->list_lock irq_context: 0 rtnl_mutex &bat_priv->softif_vlan_list_lock &n->list_lock &c->lock irq_context: 0 (wq_completion)bond1#3 irq_context: 0 sb_writers#4 tomoyo_ss mmu_notifier_invalidate_range_start &cfs_rq->removed.lock irq_context: 0 rtnl_mutex &bond->mode_lock &c->lock irq_context: 0 &child->perf_event_mutex &cfs_rq->removed.lock irq_context: 0 &child->perf_event_mutex &obj_hash[i].lock irq_context: 0 &child->perf_event_mutex pool_lock#2 irq_context: 0 (wq_completion)bond1#3 (work_completion)(&(&slave->notify_work)->work) irq_context: 0 (wq_completion)bond1#3 (work_completion)(&(&slave->notify_work)->work) &obj_hash[i].lock irq_context: 0 (wq_completion)bond1#3 (work_completion)(&(&slave->notify_work)->work) &base->lock irq_context: 0 (wq_completion)bond1#3 (work_completion)(&(&slave->notify_work)->work) &base->lock &obj_hash[i].lock irq_context: 0 rtnl_mutex &idev->mc_lock &batadv_netdev_addr_lock_key &n->list_lock irq_context: 0 rtnl_mutex &idev->mc_lock &batadv_netdev_addr_lock_key &n->list_lock &c->lock irq_context: 0 (wq_completion)bond1#3 (work_completion)(&(&slave->notify_work)->work) rtnl_mutex net_rwsem irq_context: 0 (wq_completion)bond1#3 (work_completion)(&(&slave->notify_work)->work) rtnl_mutex net_rwsem &list->lock#2 irq_context: 0 (wq_completion)bond1#3 (work_completion)(&(&slave->notify_work)->work) rtnl_mutex &tn->lock irq_context: 0 (wq_completion)bond1#3 (work_completion)(&(&slave->notify_work)->work) rtnl_mutex.wait_lock irq_context: 0 (wq_completion)bond1#3 (work_completion)(&(&slave->notify_work)->work) &p->pi_lock irq_context: 0 &sb->s_type->i_mutex_key#8 &sbi->s_writepages_rwsem jbd2_handle &ei->i_data_sem rcu_read_lock &rcu_state.gp_wq irq_context: 0 &sb->s_type->i_mutex_key#8 &sbi->s_writepages_rwsem jbd2_handle &ei->i_data_sem rcu_read_lock &rcu_state.gp_wq &p->pi_lock irq_context: 0 &sb->s_type->i_mutex_key#8 &sbi->s_writepages_rwsem jbd2_handle &ei->i_data_sem rcu_read_lock &rcu_state.gp_wq &p->pi_lock &rq->__lock irq_context: 0 &sb->s_type->i_mutex_key#8 &sbi->s_writepages_rwsem jbd2_handle &ei->i_data_sem rcu_read_lock &rcu_state.gp_wq &p->pi_lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)wg-crypt-wg0#6 (work_completion)(&peer->transmit_packet_work) &peer->endpoint_lock rcu_read_lock_bh rcu_read_lock rcu_read_lock_bh rcu_read_lock &____s->seqcount#2 irq_context: 0 (wq_completion)wg-crypt-wg0#6 (work_completion)(&peer->transmit_packet_work) &peer->endpoint_lock rcu_read_lock_bh rcu_read_lock rcu_read_lock_bh rcu_read_lock &____s->seqcount irq_context: 0 sb_writers#5 &type->i_mutex_dir_key#5 remove_cache_srcu &pcp->lock &zone->lock irq_context: 0 clock-AF_RDS irq_context: 0 &rs->rs_recv_lock irq_context: 0 rds_cong_monitor_lock irq_context: 0 rds_cong_lock irq_context: 0 &rs->rs_lock irq_context: 0 &rs->rs_rdma_lock irq_context: 0 &q->lock irq_context: 0 clock-AF_NETLINK irq_context: 0 genl_sk_destructing_waitq.lock irq_context: 0 wlock-AF_NETLINK irq_context: 0 sb_writers#5 &type->i_mutex_dir_key#5 remove_cache_srcu &pcp->lock &zone->lock &____s->seqcount irq_context: 0 sb_writers#5 &type->i_mutex_dir_key#5/1 mmu_notifier_invalidate_range_start &rq->__lock irq_context: 0 sb_writers#5 &type->i_mutex_dir_key#5/1 mmu_notifier_invalidate_range_start &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 cb_lock &rdev->wiphy.mtx rcu_read_lock rcu_read_lock_bh rcu_read_lock &____s->seqcount#2 irq_context: 0 cb_lock &rdev->wiphy.mtx rcu_read_lock rcu_read_lock_bh rcu_read_lock &____s->seqcount irq_context: 0 rtnl_mutex &batadv_netdev_addr_lock_key &n->list_lock irq_context: 0 rtnl_mutex &batadv_netdev_addr_lock_key &n->list_lock &c->lock irq_context: 0 (wq_completion)bat_events (work_completion)(&(&bat_priv->mcast.work)->work) &bat_priv->mcast.mla_lock &____s->seqcount#2 irq_context: 0 (wq_completion)bat_events (work_completion)(&(&bat_priv->mcast.work)->work) &bat_priv->mcast.mla_lock &____s->seqcount irq_context: 0 rtnl_mutex &idev->mc_lock &dev_addr_list_lock_key/1 &obj_hash[i].lock pool_lock irq_context: 0 rtnl_mutex remove_cache_srcu &cfs_rq->removed.lock irq_context: 0 cb_lock &rdev->wiphy.mtx &local->mtx &pcp->lock &zone->lock irq_context: 0 cb_lock &rdev->wiphy.mtx &local->mtx rcu_read_lock pgd_lock irq_context: 0 cb_lock &rdev->wiphy.mtx &local->mtx rcu_read_lock stock_lock irq_context: 0 cb_lock &rdev->wiphy.mtx &local->mtx rcu_read_lock rcu_read_lock pool_lock#2 irq_context: 0 cb_lock &rdev->wiphy.mtx &local->mtx rcu_read_lock key irq_context: 0 rtnl_mutex &pool->lock/1 rcu_read_lock &pool->lock &p->pi_lock &rq->__lock &cfs_rq->removed.lock irq_context: 0 rtnl_mutex &bat_priv->tvlv.handler_list_lock &n->list_lock irq_context: 0 rtnl_mutex &bat_priv->tvlv.handler_list_lock &n->list_lock &c->lock irq_context: 0 rtnl_mutex &pnn->pndevs.lock &rq->__lock irq_context: 0 rtnl_mutex &pnn->pndevs.lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 rcu_state.barrier_mutex rcu_read_lock &rq->__lock irq_context: 0 rcu_state.barrier_mutex rcu_read_lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 rcu_state.barrier_mutex &cfs_rq->removed.lock irq_context: 0 rcu_state.barrier_mutex &obj_hash[i].lock irq_context: 0 rcu_state.barrier_mutex pool_lock#2 irq_context: 0 rcu_state.barrier_mutex &rq->__lock &cfs_rq->removed.lock irq_context: 0 cb_lock &rdev->wiphy.mtx &local->mtx rcu_read_lock pcpu_lock irq_context: 0 cb_lock &rdev->wiphy.mtx &local->mtx rcu_read_lock percpu_counters_lock irq_context: 0 cb_lock &rdev->wiphy.mtx &local->mtx rcu_read_lock pcpu_lock stock_lock irq_context: 0 sk_lock-AF_INET6 rcu_read_lock rcu_read_lock &n->list_lock irq_context: 0 sk_lock-AF_INET6 rcu_read_lock rcu_read_lock &n->list_lock &c->lock irq_context: 0 rtnl_mutex &block->cb_lock flow_indr_block_lock &____s->seqcount#2 irq_context: 0 rtnl_mutex &block->cb_lock flow_indr_block_lock &____s->seqcount irq_context: 0 sb_writers#3 &sb->s_type->i_mutex_key#9 &rq->__lock &cfs_rq->removed.lock irq_context: 0 &disk->open_mutex &rq->__lock &cfs_rq->removed.lock irq_context: 0 fill_pool_map-wait-type-override pgd_lock irq_context: 0 fill_pool_map-wait-type-override stock_lock irq_context: 0 fill_pool_map-wait-type-override key irq_context: 0 fill_pool_map-wait-type-override pcpu_lock irq_context: 0 fill_pool_map-wait-type-override percpu_counters_lock irq_context: 0 fill_pool_map-wait-type-override pcpu_lock stock_lock irq_context: 0 &sb->s_type->i_mutex_key#8 pgd_lock irq_context: 0 &sb->s_type->i_mutex_key#8 stock_lock irq_context: 0 &sb->s_type->i_mutex_key#8 rcu_read_lock pool_lock#2 irq_context: 0 &sb->s_type->i_mutex_key#8 key irq_context: 0 &sb->s_type->i_mutex_key#8 pcpu_lock irq_context: 0 &sb->s_type->i_mutex_key#8 percpu_counters_lock irq_context: 0 &sb->s_type->i_mutex_key#8 pcpu_lock stock_lock irq_context: 0 rtnl_mutex &dev_addr_list_lock_key#3/2 irq_context: 0 rtnl_mutex &dev_addr_list_lock_key#3/2 &macvlan_netdev_addr_lock_key/1 irq_context: 0 rtnl_mutex &dev_addr_list_lock_key#3/2 &macvlan_netdev_addr_lock_key/1 _xmit_ETHER irq_context: 0 rtnl_mutex rcu_read_lock &ipvlan->addrs_lock &c->lock irq_context: 0 rtnl_mutex rcu_read_lock &ipvlan->addrs_lock &____s->seqcount#2 irq_context: 0 rtnl_mutex rcu_read_lock &ipvlan->addrs_lock &pcp->lock &zone->lock irq_context: 0 rtnl_mutex rcu_read_lock &ipvlan->addrs_lock &pcp->lock &zone->lock &____s->seqcount irq_context: 0 rtnl_mutex rcu_read_lock &ipvlan->addrs_lock &____s->seqcount irq_context: 0 rtnl_mutex &ipvlan->addrs_lock irq_context: 0 rtnl_mutex &ipvlan->addrs_lock &obj_hash[i].lock irq_context: 0 rtnl_mutex &ipvlan->addrs_lock pool_lock#2 irq_context: 0 rtnl_mutex &ipvlan->addrs_lock krc.lock irq_context: 0 rtnl_mutex &dev->tx_global_lock &qdisc_xmit_lock_key#2 irq_context: 0 rtnl_mutex &dev_addr_list_lock_key#3/2 &obj_hash[i].lock irq_context: 0 rtnl_mutex &dev_addr_list_lock_key#3/2 krc.lock irq_context: 0 rtnl_mutex &idev->mc_lock &dev_addr_list_lock_key#3/2 irq_context: 0 rtnl_mutex &idev->mc_lock &dev_addr_list_lock_key#3/2 &obj_hash[i].lock irq_context: 0 rtnl_mutex &idev->mc_lock &dev_addr_list_lock_key#3/2 krc.lock irq_context: 0 rtnl_mutex (work_completion)(&port->wq) irq_context: 0 &f->f_pos_lock sb_writers#4 &sb->s_type->i_mutex_key#8 mmu_notifier_invalidate_range_start &cfs_rq->removed.lock irq_context: 0 &f->f_pos_lock sb_writers#4 &sb->s_type->i_mutex_key#8 mmu_notifier_invalidate_range_start &obj_hash[i].lock irq_context: 0 &f->f_pos_lock sb_writers#4 &sb->s_type->i_mutex_key#8 mmu_notifier_invalidate_range_start pool_lock#2 irq_context: 0 (wq_completion)wg-kex-wg2#9 (work_completion)(&peer->transmit_handshake_work) &peer->endpoint_lock rcu_read_lock_bh rcu_read_lock rcu_read_lock_bh rcu_read_lock pool_lock#2 irq_context: 0 (wq_completion)wg-kex-wg2#9 (work_completion)(&peer->transmit_handshake_work) &peer->endpoint_lock rcu_read_lock_bh rcu_read_lock rcu_read_lock_bh rcu_read_lock tk_core.seq.seqcount irq_context: 0 (wq_completion)wg-kex-wg2#9 (work_completion)(&peer->transmit_handshake_work) &peer->endpoint_lock rcu_read_lock_bh rcu_read_lock rcu_read_lock_bh rcu_read_lock &obj_hash[i].lock irq_context: 0 sb_writers#4 &type->i_mutex_dir_key#3/1 jbd2_handle rcu_read_lock &rcu_state.expedited_wq irq_context: 0 sb_writers#4 &type->i_mutex_dir_key#3/1 jbd2_handle rcu_read_lock &rcu_state.expedited_wq &p->pi_lock irq_context: 0 sb_writers#4 &type->i_mutex_dir_key#3/1 jbd2_handle rcu_read_lock &rcu_state.expedited_wq &p->pi_lock &rq->__lock irq_context: 0 sb_writers#4 &type->i_mutex_dir_key#3/1 jbd2_handle rcu_read_lock &rcu_state.expedited_wq &p->pi_lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)wg-kex-wg0#7 (work_completion)(&peer->transmit_handshake_work) &peer->endpoint_lock rcu_read_lock_bh rcu_read_lock rcu_read_lock_bh rcu_read_lock tk_core.seq.seqcount irq_context: 0 (wq_completion)wg-kex-wg0#7 (work_completion)(&peer->transmit_handshake_work) &peer->endpoint_lock rcu_read_lock_bh rcu_read_lock rcu_read_lock_bh rcu_read_lock &obj_hash[i].lock irq_context: 0 (wq_completion)bond3 irq_context: 0 (wq_completion)bond3 (work_completion)(&(&bond->alb_work)->work) irq_context: 0 (wq_completion)bond3 (work_completion)(&(&bond->alb_work)->work) &obj_hash[i].lock irq_context: 0 (wq_completion)bond3 (work_completion)(&(&bond->alb_work)->work) &base->lock irq_context: 0 (wq_completion)bond3 (work_completion)(&(&bond->alb_work)->work) &base->lock &obj_hash[i].lock irq_context: 0 (wq_completion)bond3 (work_completion)(&(&bond->alb_work)->work) &rq->__lock irq_context: 0 (wq_completion)bond3 (work_completion)(&(&bond->alb_work)->work) &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)bond3 (work_completion)(&(&bond->mii_work)->work) irq_context: 0 (wq_completion)bond3 (work_completion)(&(&bond->mii_work)->work) &obj_hash[i].lock irq_context: 0 (wq_completion)bond3 (work_completion)(&(&bond->mii_work)->work) &base->lock irq_context: 0 (wq_completion)bond3 (work_completion)(&(&bond->mii_work)->work) &base->lock &obj_hash[i].lock irq_context: 0 (wq_completion)events (linkwatch_work).work rtnl_mutex &dev->tx_global_lock &qdisc_xmit_lock_key#3 irq_context: 0 (wq_completion)events (linkwatch_work).work rtnl_mutex rcu_read_lock &bond->stats_lock irq_context: softirq &(&bond->alb_work)->timer irq_context: softirq &(&bond->alb_work)->timer rcu_read_lock &pool->lock/1 irq_context: softirq &(&bond->alb_work)->timer rcu_read_lock &pool->lock/1 &obj_hash[i].lock irq_context: softirq &(&bond->alb_work)->timer rcu_read_lock &pool->lock/1 &p->pi_lock irq_context: softirq &(&bond->mii_work)->timer irq_context: softirq &(&bond->mii_work)->timer rcu_read_lock &pool->lock/1 irq_context: softirq &(&bond->mii_work)->timer rcu_read_lock &pool->lock/1 &obj_hash[i].lock irq_context: softirq &(&bond->mii_work)->timer rcu_read_lock &pool->lock/1 &p->pi_lock irq_context: softirq &(&bond->mii_work)->timer rcu_read_lock &pool->lock/1 &p->pi_lock &rq->__lock irq_context: softirq &(&bond->mii_work)->timer rcu_read_lock &pool->lock/1 &p->pi_lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)bond4 irq_context: 0 (wq_completion)bond4 (work_completion)(&(&bond->alb_work)->work) irq_context: 0 (wq_completion)bond4 (work_completion)(&(&bond->alb_work)->work) &obj_hash[i].lock irq_context: 0 (wq_completion)bond4 (work_completion)(&(&bond->alb_work)->work) &base->lock irq_context: 0 (wq_completion)bond4 (work_completion)(&(&bond->alb_work)->work) &base->lock &obj_hash[i].lock irq_context: 0 (wq_completion)bond4 (work_completion)(&(&bond->mii_work)->work) irq_context: 0 (wq_completion)bond4 (work_completion)(&(&bond->mii_work)->work) &obj_hash[i].lock irq_context: 0 (wq_completion)bond4 (work_completion)(&(&bond->mii_work)->work) &base->lock irq_context: 0 (wq_completion)bond4 (work_completion)(&(&bond->mii_work)->work) &base->lock &obj_hash[i].lock irq_context: 0 sb_writers#4 &type->i_mutex_dir_key#3/1 &type->i_mutex_dir_key#3 jbd2_handle &journal->j_wait_updates &p->pi_lock irq_context: 0 sb_writers#4 &type->i_mutex_dir_key#3/1 &type->i_mutex_dir_key#3 jbd2_handle &journal->j_wait_updates &p->pi_lock &rq->__lock irq_context: 0 sb_writers#4 &type->i_mutex_dir_key#3/1 &type->i_mutex_dir_key#3 jbd2_handle &journal->j_wait_updates &p->pi_lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: softirq &(&bond->alb_work)->timer rcu_read_lock &pool->lock/1 &p->pi_lock &rq->__lock irq_context: softirq &(&bond->alb_work)->timer rcu_read_lock &pool->lock/1 &p->pi_lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)bond5 irq_context: 0 (wq_completion)bond5 (work_completion)(&(&bond->alb_work)->work) irq_context: 0 (wq_completion)bond5 (work_completion)(&(&bond->alb_work)->work) &obj_hash[i].lock irq_context: 0 (wq_completion)bond5 (work_completion)(&(&bond->alb_work)->work) &base->lock irq_context: 0 (wq_completion)bond5 (work_completion)(&(&bond->alb_work)->work) &base->lock &obj_hash[i].lock irq_context: 0 (wq_completion)bond5 (work_completion)(&(&bond->mii_work)->work) irq_context: 0 (wq_completion)bond5 (work_completion)(&(&bond->mii_work)->work) &obj_hash[i].lock irq_context: 0 (wq_completion)bond5 (work_completion)(&(&bond->mii_work)->work) &base->lock irq_context: 0 (wq_completion)bond5 (work_completion)(&(&bond->mii_work)->work) &base->lock &obj_hash[i].lock irq_context: 0 rtnl_mutex &dev_addr_list_lock_key &c->lock irq_context: 0 ppp_mutex rtnl_mutex rtnl_mutex.wait_lock irq_context: 0 ppp_mutex rtnl_mutex &pn->all_ppp_mutex &c->lock irq_context: 0 ppp_mutex rtnl_mutex uevent_sock_mutex &____s->seqcount#2 irq_context: 0 ppp_mutex rtnl_mutex uevent_sock_mutex &____s->seqcount irq_context: 0 ppp_mutex rtnl_mutex &obj_hash[i].lock pool_lock irq_context: 0 ppp_mutex &n->list_lock irq_context: 0 ppp_mutex &n->list_lock &c->lock irq_context: 0 ppp_mutex &ppp->wlock irq_context: 0 ppp_mutex &ppp->wlock &ppp->rlock irq_context: 0 (wq_completion)events (linkwatch_work).work rtnl_mutex &cfs_rq->removed.lock irq_context: 0 ppp_mutex &____s->seqcount#2 irq_context: 0 ppp_mutex &____s->seqcount irq_context: 0 ppp_mutex &ppp->wlock &ppp->rlock &obj_hash[i].lock irq_context: 0 ppp_mutex &ppp->wlock &ppp->rlock pool_lock#2 irq_context: softirq &(&bond->alb_work)->timer rcu_read_lock &pool->lock/1 &p->pi_lock &cfs_rq->removed.lock irq_context: 0 &sb->s_type->i_mutex_key#8 &sbi->s_writepages_rwsem jbd2_handle &cfs_rq->removed.lock irq_context: 0 &sb->s_type->i_mutex_key#8 &sbi->s_writepages_rwsem jbd2_handle &obj_hash[i].lock irq_context: 0 (wq_completion)events_unbound (work_completion)(&sub_info->work) fs_reclaim mmu_notifier_invalidate_range_start &rq->__lock irq_context: 0 (wq_completion)events_unbound (work_completion)(&sub_info->work) fs_reclaim mmu_notifier_invalidate_range_start &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 cb_lock nlk_cb_mutex-GENERIC genl_mutex rcu_read_lock rcu_node_0 irq_context: 0 cb_lock nlk_cb_mutex-GENERIC genl_mutex rcu_read_lock &rq->__lock irq_context: 0 cb_lock nlk_cb_mutex-GENERIC genl_mutex rcu_read_lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 cb_lock nlk_cb_mutex-GENERIC genl_mutex rcu_read_lock &rcu_state.gp_wq irq_context: 0 cb_lock nlk_cb_mutex-GENERIC genl_mutex rcu_read_lock &rcu_state.gp_wq &p->pi_lock irq_context: 0 cb_lock nlk_cb_mutex-GENERIC genl_mutex rcu_read_lock &rcu_state.gp_wq &p->pi_lock &rq->__lock irq_context: 0 cb_lock nlk_cb_mutex-GENERIC genl_mutex rcu_read_lock &rcu_state.gp_wq &p->pi_lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 sk_lock-AF_INET6 tcp_md5sig_mutex irq_context: 0 sk_lock-AF_INET6 tcp_md5sig_mutex crypto_alg_sem irq_context: 0 sk_lock-AF_INET6 tcp_md5sig_mutex fs_reclaim irq_context: 0 sk_lock-AF_INET6 tcp_md5sig_mutex fs_reclaim mmu_notifier_invalidate_range_start irq_context: 0 sk_lock-AF_INET6 tcp_md5sig_mutex pool_lock#2 irq_context: 0 sk_lock-AF_INET6 tcp_md5sig_mutex &c->lock irq_context: 0 &sb->s_type->i_mutex_key#10 sk_lock-AF_INET6 slock-AF_INET6 krc.lock irq_context: 0 cb_lock nlk_cb_mutex-GENERIC genl_mutex rcu_node_0 irq_context: 0 cb_lock nlk_cb_mutex-GENERIC genl_mutex.wait_lock irq_context: 0 cb_lock nlk_cb_mutex-GENERIC &p->pi_lock irq_context: 0 cb_lock nlk_cb_mutex-GENERIC &p->pi_lock &cfs_rq->removed.lock irq_context: 0 cb_lock nlk_cb_mutex-GENERIC &p->pi_lock &rq->__lock irq_context: 0 cb_lock nlk_cb_mutex-GENERIC &p->pi_lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 cb_lock nlk_cb_mutex-GENERIC &cfs_rq->removed.lock irq_context: 0 &sig->cred_guard_mutex tomoyo_ss fs_reclaim mmu_notifier_invalidate_range_start &rq->__lock irq_context: 0 &sig->cred_guard_mutex tomoyo_ss fs_reclaim mmu_notifier_invalidate_range_start &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: softirq (&hsr->announce_timer) rcu_read_lock &hsr->seqnr_lock rcu_read_lock batched_entropy_u8.lock irq_context: softirq (&hsr->announce_timer) rcu_read_lock &hsr->seqnr_lock rcu_read_lock batched_entropy_u8.lock crngs.lock irq_context: softirq (&hsr->announce_timer) rcu_read_lock &hsr->seqnr_lock rcu_read_lock kfence_freelist_lock irq_context: 0 sb_writers#4 &sb->s_type->i_mutex_key#8 ima_extend_list_mutex rcu_read_lock rcu_node_0 irq_context: 0 sb_writers#4 &sb->s_type->i_mutex_key#8 ima_extend_list_mutex rcu_read_lock &rq->__lock irq_context: softirq &(&bond->mii_work)->timer rcu_read_lock &pool->lock/1 &p->pi_lock &cfs_rq->removed.lock irq_context: 0 (wq_completion)bat_events (work_completion)(&(&bat_priv->orig_work)->work) &obj_hash[i].lock pool_lock irq_context: 0 sk_lock-AF_INET6 rcu_read_lock rcu_read_lock rcu_read_lock_bh rcu_read_lock &ei->socket.wq.wait irq_context: softirq rcu_read_lock rcu_read_lock slock-AF_INET6 &____s->seqcount irq_context: softirq rcu_read_lock rcu_read_lock slock-AF_INET6 rcu_read_lock pool_lock#2 irq_context: 0 sk_lock-AF_INET6 &queue->rskq_lock irq_context: 0 sk_lock-AF_INET6 clock-AF_INET6 irq_context: 0 sk_lock-AF_INET6 wlock-AF_INET6 irq_context: softirq rcu_read_lock rcu_read_lock slock-AF_INET6/1 sk_lock-AF_INET6 &n->list_lock irq_context: softirq rcu_read_lock rcu_read_lock slock-AF_INET6/1 sk_lock-AF_INET6 &n->list_lock &c->lock irq_context: softirq (&hc->tx_rtotimer) irq_context: softirq (&hc->tx_rtotimer) slock-AF_INET6 irq_context: softirq (&hc->tx_rtotimer) slock-AF_INET6 &c->lock irq_context: softirq (&hc->tx_rtotimer) slock-AF_INET6 pool_lock#2 irq_context: softirq (&hc->tx_rtotimer) slock-AF_INET6 &obj_hash[i].lock irq_context: softirq (&hc->tx_rtotimer) slock-AF_INET6 &base->lock irq_context: softirq (&hc->tx_rtotimer) slock-AF_INET6 &base->lock &obj_hash[i].lock irq_context: softirq slock-AF_INET6 wlock-AF_INET6 irq_context: softirq slock-AF_INET6 rcu_read_lock rcu_read_lock rcu_read_lock_bh tk_core.seq.seqcount irq_context: 0 &f->f_pos_lock &type->i_mutex_dir_key#4 &root->kernfs_rwsem &rq->__lock &cfs_rq->removed.lock irq_context: softirq net/ipv4/tcp_ipv4.c:1063 irq_context: softirq net/ipv4/tcp_ipv4.c:1063 rcu_read_lock &pool->lock irq_context: softirq net/ipv4/tcp_ipv4.c:1063 rcu_read_lock &pool->lock &obj_hash[i].lock irq_context: softirq net/ipv4/tcp_ipv4.c:1063 rcu_read_lock &pool->lock &p->pi_lock irq_context: softirq net/ipv4/tcp_ipv4.c:1063 rcu_read_lock &pool->lock &p->pi_lock &rq->__lock irq_context: softirq net/ipv4/tcp_ipv4.c:1063 rcu_read_lock &pool->lock &p->pi_lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)events ((tcp_md5_needed).work).work irq_context: 0 (wq_completion)events ((tcp_md5_needed).work).work cpu_hotplug_lock irq_context: 0 (wq_completion)events ((tcp_md5_needed).work).work cpu_hotplug_lock jump_label_mutex irq_context: 0 (wq_completion)events ((tcp_md5_needed).work).work cpu_hotplug_lock jump_label_mutex text_mutex irq_context: 0 (wq_completion)events ((tcp_md5_needed).work).work cpu_hotplug_lock jump_label_mutex text_mutex ptlock_ptr(page)#2 irq_context: 0 (wq_completion)events ((tcp_md5_needed).work).work cpu_hotplug_lock jump_label_mutex text_mutex &rq->__lock irq_context: 0 (wq_completion)events ((tcp_md5_needed).work).work cpu_hotplug_lock jump_label_mutex text_mutex &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 &sb->s_type->i_mutex_key#10 sk_lock-AF_INET6 quarantine_lock irq_context: 0 &sb->s_type->i_mutex_key#10 sk_lock-AF_INET6 (console_sem).lock irq_context: 0 &sb->s_type->i_mutex_key#10 sk_lock-AF_INET6 console_lock console_srcu console_owner_lock irq_context: 0 &sb->s_type->i_mutex_key#10 sk_lock-AF_INET6 console_lock console_srcu console_owner irq_context: 0 &sb->s_type->i_mutex_key#10 sk_lock-AF_INET6 console_lock console_srcu console_owner &port_lock_key irq_context: 0 pernet_ops_rwsem devices_rwsem rdma_nets_rwsem &device->compat_devs_mutex remove_cache_srcu &pcp->lock &zone->lock irq_context: 0 pernet_ops_rwsem devices_rwsem rdma_nets_rwsem &device->compat_devs_mutex remove_cache_srcu &pcp->lock &zone->lock &____s->seqcount irq_context: 0 &sb->s_type->i_mutex_key#10 sk_lock-AF_INET6 console_lock console_srcu console_owner console_owner_lock irq_context: 0 &sb->s_type->i_mutex_key#10 sk_lock-AF_INET6 &c->lock irq_context: 0 &sb->s_type->i_mutex_key#10 sk_lock-AF_INET6 &____s->seqcount#2 irq_context: 0 &sb->s_type->i_mutex_key#10 sk_lock-AF_INET6 rcu_read_lock rcu_read_lock &ct->lock irq_context: 0 &sb->s_type->i_mutex_key#10 sk_lock-AF_INET6 rcu_read_lock rcu_read_lock rcu_read_lock_bh tk_core.seq.seqcount irq_context: softirq rcu_read_lock rcu_read_lock slock-AF_INET6/1 sk_lock-AF_INET6 &base->lock irq_context: softirq rcu_read_lock rcu_read_lock slock-AF_INET6/1 sk_lock-AF_INET6 &base->lock &obj_hash[i].lock irq_context: softirq rcu_read_lock rcu_read_lock slock-AF_INET6/1 sk_lock-AF_INET6 &dccp_hashinfo.bhash[i].lock irq_context: softirq rcu_read_lock rcu_read_lock slock-AF_INET6/1 sk_lock-AF_INET6 &dccp_hashinfo.bhash[i].lock &dccp_hashinfo.bhash2[i].lock irq_context: softirq rcu_read_lock rcu_read_lock slock-AF_INET6/1 sk_lock-AF_INET6 &hashinfo->ehash_locks[i] irq_context: 0 &sb->s_type->i_mutex_key#10 sk_lock-AF_INET6 rcu_read_lock rcu_read_lock rcu_node_0 irq_context: 0 &sb->s_type->i_mutex_key#10 sk_lock-AF_INET6 rcu_read_lock rcu_read_lock &rq->__lock irq_context: 0 &sb->s_type->i_mutex_key#10 sk_lock-AF_INET6 rcu_read_lock rcu_read_lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 pernet_ops_rwsem rtnl_mutex &pnettable->lock &rq->__lock irq_context: 0 pernet_ops_rwsem rtnl_mutex &pnettable->lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem k-sk_lock-AF_RXRPC rcu_read_lock rcu_node_0 irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem k-sk_lock-AF_RXRPC rcu_read_lock &rq->__lock irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem k-sk_lock-AF_RXRPC rcu_read_lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem ovs_mutex nf_ct_proto_mutex defrag4_mutex &obj_hash[i].lock pool_lock irq_context: softirq rcu_read_lock rcu_read_lock rcu_read_lock rcu_read_lock_bh tk_core.seq.seqcount irq_context: 0 sb_writers#4 &type->i_mutex_dir_key#3/1 &sb->s_type->i_mutex_key#8 jbd2_handle stock_lock irq_context: 0 sb_writers#4 &type->i_mutex_dir_key#3/1 &sb->s_type->i_mutex_key#8 jbd2_handle pcpu_lock stock_lock irq_context: softirq rcu_read_lock rcu_read_lock rcu_read_lock rcu_read_lock_bh &list->lock#5 irq_context: softirq (&hc->tx_rtotimer) &obj_hash[i].lock irq_context: 0 (wq_completion)events (work_completion)(&pwq->unbound_release_work) &rq->__lock &cfs_rq->removed.lock irq_context: softirq (&hc->tx_rtotimer) rcu_read_lock pool_lock#2 irq_context: softirq (&hc->tx_rtotimer) pool_lock#2 irq_context: softirq (&hc->tx_rtotimer) &dir->lock irq_context: 0 rtnl_mutex cpu_hotplug_lock &rq->__lock irq_context: 0 rtnl_mutex cpu_hotplug_lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 pernet_ops_rwsem mmu_notifier_invalidate_range_start &rq->__lock irq_context: 0 pernet_ops_rwsem mmu_notifier_invalidate_range_start &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 sk_lock-AF_INET6 &rcu_state.expedited_wq &p->pi_lock irq_context: 0 sk_lock-AF_INET6 &rcu_state.expedited_wq &p->pi_lock &rq->__lock irq_context: 0 sk_lock-AF_INET6 &rcu_state.expedited_wq &p->pi_lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem rtnl_mutex gdp_mutex &rq->__lock irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem rtnl_mutex gdp_mutex &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)wg-crypt-wg1#6 (work_completion)(&peer->transmit_packet_work) &peer->endpoint_lock rcu_read_lock_bh rcu_read_lock &c->lock irq_context: 0 (wq_completion)wg-crypt-wg1#6 (work_completion)(&peer->transmit_packet_work) &peer->endpoint_lock rcu_read_lock_bh rcu_read_lock &____s->seqcount irq_context: 0 (wq_completion)wg-crypt-wg1#3 (work_completion)(&peer->transmit_packet_work) &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 &sb->s_type->i_mutex_key#9 rcu_read_lock &rcu_state.expedited_wq irq_context: 0 &sb->s_type->i_mutex_key#9 rcu_read_lock &rcu_state.expedited_wq &p->pi_lock irq_context: 0 &sb->s_type->i_mutex_key#9 rcu_read_lock &rcu_state.expedited_wq &p->pi_lock &rq->__lock irq_context: 0 &sb->s_type->i_mutex_key#9 rcu_read_lock &rcu_state.expedited_wq &p->pi_lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 pernet_ops_rwsem &xt[i].mutex &lock->wait_lock irq_context: 0 pernet_ops_rwsem &xt[i].mutex &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 pernet_ops_rwsem rtnl_mutex remove_cache_srcu &pcp->lock &zone->lock irq_context: 0 pernet_ops_rwsem rtnl_mutex remove_cache_srcu &pcp->lock &zone->lock &____s->seqcount irq_context: 0 pernet_ops_rwsem rtnl_mutex &idev->mc_lock rcu_read_lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 pernet_ops_rwsem rtnl_mutex uevent_sock_mutex rcu_read_lock &rcu_state.gp_wq irq_context: 0 pernet_ops_rwsem rtnl_mutex uevent_sock_mutex rcu_read_lock &rcu_state.gp_wq &p->pi_lock irq_context: 0 pernet_ops_rwsem rtnl_mutex uevent_sock_mutex rcu_read_lock &rcu_state.gp_wq &p->pi_lock &rq->__lock irq_context: 0 pernet_ops_rwsem rtnl_mutex uevent_sock_mutex rcu_read_lock &rcu_state.gp_wq &p->pi_lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)events (work_completion)(&(&nsim_dev->trap_data->trap_report_dw)->work) &devlink->lock_key#4 &rq->__lock &cfs_rq->removed.lock irq_context: softirq (&icsk->icsk_retransmit_timer) rcu_read_lock pool_lock#2 irq_context: 0 &f->f_pos_lock &type->i_mutex_dir_key#3 sb_writers#4 jbd2_handle &journal->j_state_lock &journal->j_wait_commit &p->pi_lock &cfs_rq->removed.lock irq_context: 0 (wq_completion)wg-crypt-wg1 (work_completion)(&peer->transmit_packet_work) &peer->endpoint_lock rcu_read_lock_bh rcu_read_lock &n->list_lock irq_context: 0 (wq_completion)wg-crypt-wg1 (work_completion)(&peer->transmit_packet_work) &peer->endpoint_lock rcu_read_lock_bh rcu_read_lock &n->list_lock &c->lock irq_context: 0 pernet_ops_rwsem rcu_read_lock pgd_lock irq_context: 0 pernet_ops_rwsem rcu_read_lock stock_lock irq_context: 0 pernet_ops_rwsem rcu_read_lock key irq_context: 0 pernet_ops_rwsem rcu_read_lock pcpu_lock irq_context: 0 pernet_ops_rwsem rcu_read_lock percpu_counters_lock irq_context: 0 pernet_ops_rwsem rcu_read_lock pcpu_lock stock_lock irq_context: 0 pernet_ops_rwsem uevent_sock_mutex &____s->seqcount#2 irq_context: 0 pernet_ops_rwsem uevent_sock_mutex &____s->seqcount irq_context: 0 (wq_completion)events (work_completion)(&(&nsim_dev->trap_data->trap_report_dw)->work) &devlink->lock_key#5 &nsim_trap_data->trap_lock rcu_read_lock &p->pi_lock irq_context: 0 &sb->s_type->i_mutex_key#10 &list->lock#37 irq_context: 0 &mm->mmap_lock &anon_vma->rwsem rcu_node_0 irq_context: 0 (wq_completion)events (work_completion)(&(&nsim_dev->trap_data->trap_report_dw)->work) &devlink->lock_key#5 &nsim_trap_data->trap_lock rcu_read_lock &p->pi_lock &rq->__lock irq_context: 0 (wq_completion)events (work_completion)(&(&nsim_dev->trap_data->trap_report_dw)->work) &devlink->lock_key#5 &nsim_trap_data->trap_lock rcu_read_lock &p->pi_lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem rtnl_mutex rcu_state.exp_mutex rcu_read_lock &pool->lock fill_pool_map-wait-type-override &____s->seqcount irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem rtnl_mutex rcu_state.exp_mutex rcu_read_lock &pool->lock fill_pool_map-wait-type-override pool_lock#2 irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem rtnl_mutex rcu_state.exp_mutex rcu_read_lock &pool->lock fill_pool_map-wait-type-override rcu_read_lock pool_lock#2 irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem rtnl_mutex rcu_state.exp_mutex rcu_read_lock &pool->lock fill_pool_map-wait-type-override &obj_hash[i].lock irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem rtnl_mutex rcu_state.exp_mutex rcu_read_lock &pool->lock fill_pool_map-wait-type-override &c->lock irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem rtnl_mutex rcu_state.exp_mutex rcu_read_lock &pool->lock fill_pool_map-wait-type-override &n->list_lock irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem rtnl_mutex rcu_state.exp_mutex rcu_read_lock &pool->lock fill_pool_map-wait-type-override &n->list_lock &c->lock irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem rtnl_mutex rcu_state.exp_mutex rcu_read_lock &pool->lock fill_pool_map-wait-type-override pool_lock irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem rtnl_mutex fs_reclaim mmu_notifier_invalidate_range_start &rq->__lock irq_context: 0 &xt[i].mutex free_vmap_area_lock quarantine_lock irq_context: 0 pernet_ops_rwsem rtnl_mutex uevent_sock_mutex fs_reclaim &rq->__lock irq_context: 0 pernet_ops_rwsem rtnl_mutex uevent_sock_mutex fs_reclaim &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: softirq (&pool->idle_timer) rcu_read_lock &pool->lock/1 pool_lock#2 irq_context: softirq (&pool->idle_timer) rcu_read_lock &pool->lock/1 &p->pi_lock &rq->__lock irq_context: 0 pernet_ops_rwsem k-sk_lock-AF_TIPC &tn->nametbl_lock &service->lock krc.lock &obj_hash[i].lock irq_context: 0 pernet_ops_rwsem k-sk_lock-AF_TIPC &tn->nametbl_lock &service->lock krc.lock &base->lock irq_context: 0 pernet_ops_rwsem k-sk_lock-AF_TIPC &tn->nametbl_lock &service->lock krc.lock &base->lock &obj_hash[i].lock irq_context: softirq (&pool->idle_timer) rcu_read_lock &pool->lock/1 &p->pi_lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)events_unbound (work_completion)(&pool->idle_cull_work) wq_pool_attach_mutex &pool->lock/1 irq_context: 0 (wq_completion)events_unbound (work_completion)(&pool->idle_cull_work) wq_pool_attach_mutex &pool->lock/1 &obj_hash[i].lock irq_context: 0 (wq_completion)events_unbound (work_completion)(&pool->idle_cull_work) wq_pool_attach_mutex &pool->lock/1 &base->lock irq_context: 0 (wq_completion)events_unbound (work_completion)(&pool->idle_cull_work) wq_pool_attach_mutex &pool->lock/1 &base->lock &obj_hash[i].lock irq_context: 0 (wq_completion)events_unbound (work_completion)(&pool->idle_cull_work) wq_pool_attach_mutex &rq->__lock irq_context: 0 wq_pool_attach_mutex wq_pool_attach_mutex.wait_lock irq_context: 0 wq_pool_attach_mutex &rq->__lock irq_context: 0 wq_pool_attach_mutex &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)events_unbound (work_completion)(&pool->idle_cull_work) wq_pool_attach_mutex.wait_lock irq_context: 0 (wq_completion)events_unbound (work_completion)(&pool->idle_cull_work) &p->pi_lock irq_context: 0 (wq_completion)events_unbound (work_completion)(&pool->idle_cull_work) &p->pi_lock &rq->__lock irq_context: 0 (wq_completion)events_unbound (work_completion)(&pool->idle_cull_work) &p->pi_lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 wq_pool_attach_mutex.wait_lock irq_context: 0 wq_pool_attach_mutex &cfs_rq->removed.lock irq_context: 0 wq_pool_attach_mutex &obj_hash[i].lock irq_context: 0 wq_pool_attach_mutex pool_lock#2 irq_context: 0 (wq_completion)bat_events (work_completion)(&(&forw_packet_aggr->delayed_work)->work) &hard_iface->bat_iv.ogm_buff_mutex rcu_read_lock &pcp->lock &zone->lock irq_context: 0 misc_mtx remove_cache_srcu rcu_read_lock rcu_node_0 irq_context: 0 misc_mtx remove_cache_srcu rcu_read_lock &rq->__lock irq_context: 0 misc_mtx remove_cache_srcu rcu_read_lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem ovs_mutex &rq->__lock irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem ovs_mutex &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)bond3 (work_completion)(&(&bond->mii_work)->work) &rq->__lock irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem rtnl_mutex cpu_hotplug_lock xps_map_mutex &rq->__lock irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem rtnl_mutex cpu_hotplug_lock xps_map_mutex &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)wg-crypt-wg0#2 (work_completion)(&peer->transmit_packet_work) &cfs_rq->removed.lock irq_context: 0 &pipe->mutex/1 &rcu_state.expedited_wq irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem rcu_read_lock &pool->lock/1 &p->pi_lock &cfs_rq->removed.lock irq_context: 0 &pipe->mutex/1 &rcu_state.expedited_wq &p->pi_lock irq_context: 0 &pipe->mutex/1 &rcu_state.expedited_wq &p->pi_lock &rq->__lock irq_context: 0 &pipe->mutex/1 &rcu_state.expedited_wq &p->pi_lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 &sb->s_type->i_mutex_key#10 sk_lock-AF_INET6 &cfs_rq->removed.lock irq_context: 0 dup_mmap_sem &mm->mmap_lock &mm->mmap_lock/1 remove_cache_srcu rcu_node_0 irq_context: 0 vlan_ioctl_mutex rtnl_mutex fill_pool_map-wait-type-override &rq->__lock irq_context: 0 vlan_ioctl_mutex rtnl_mutex fill_pool_map-wait-type-override &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 vlan_ioctl_mutex rtnl_mutex rcu_read_lock rcu_read_lock &pool->lock/1 &p->pi_lock &cfs_rq->removed.lock irq_context: 0 (wq_completion)events (work_completion)(&data->fib_event_work) &data->fib_lock rcu_read_lock rcu_node_0 irq_context: 0 (wq_completion)events (work_completion)(&data->fib_event_work) &data->fib_lock rcu_read_lock &rcu_state.gp_wq irq_context: 0 (wq_completion)events (work_completion)(&data->fib_event_work) &data->fib_lock rcu_read_lock &rcu_state.gp_wq &p->pi_lock irq_context: 0 (wq_completion)events (work_completion)(&data->fib_event_work) &data->fib_lock rcu_read_lock &rcu_state.gp_wq &p->pi_lock &rq->__lock irq_context: 0 (wq_completion)events (work_completion)(&data->fib_event_work) &data->fib_lock rcu_read_lock &rcu_state.gp_wq &p->pi_lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 vlan_ioctl_mutex rtnl_mutex uevent_sock_mutex rcu_node_0 irq_context: 0 vlan_ioctl_mutex rtnl_mutex kernfs_idr_lock &obj_hash[i].lock irq_context: 0 vlan_ioctl_mutex rtnl_mutex kernfs_idr_lock pool_lock#2 irq_context: 0 vlan_ioctl_mutex krc.lock &obj_hash[i].lock irq_context: 0 vlan_ioctl_mutex krc.lock &base->lock irq_context: 0 vlan_ioctl_mutex krc.lock &base->lock &obj_hash[i].lock irq_context: 0 vlan_ioctl_mutex rtnl_mutex &root->kernfs_rwsem &sem->wait_lock irq_context: 0 vlan_ioctl_mutex rtnl_mutex &idev->mc_lock remove_cache_srcu irq_context: 0 vlan_ioctl_mutex rtnl_mutex &idev->mc_lock remove_cache_srcu quarantine_lock irq_context: 0 vlan_ioctl_mutex rtnl_mutex &idev->mc_lock remove_cache_srcu &c->lock irq_context: 0 vlan_ioctl_mutex rtnl_mutex &idev->mc_lock remove_cache_srcu &n->list_lock irq_context: 0 vlan_ioctl_mutex rtnl_mutex &idev->mc_lock remove_cache_srcu &obj_hash[i].lock irq_context: 0 vlan_ioctl_mutex rtnl_mutex &idev->mc_lock remove_cache_srcu &rq->__lock irq_context: 0 vlan_ioctl_mutex rtnl_mutex &idev->mc_lock remove_cache_srcu &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 vlan_ioctl_mutex rtnl_mutex &idev->mc_lock remove_cache_srcu pool_lock#2 irq_context: 0 vlan_ioctl_mutex rtnl_mutex &idev->mc_lock remove_cache_srcu rcu_read_lock pool_lock#2 irq_context: 0 vlan_ioctl_mutex rtnl_mutex &root->kernfs_rwsem rcu_read_lock rcu_node_0 irq_context: 0 vlan_ioctl_mutex rtnl_mutex &root->kernfs_rwsem rcu_read_lock &rq->__lock irq_context: 0 vlan_ioctl_mutex rtnl_mutex &root->kernfs_rwsem rcu_read_lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 vlan_ioctl_mutex rcu_state.barrier_mutex.wait_lock irq_context: 0 &sb->s_type->i_mutex_key#10 sk_lock-AF_CAN rcu_state.exp_mutex &rnp->exp_wq[1] irq_context: 0 &xt[i].mutex &mm->mmap_lock rcu_node_0 irq_context: 0 fs_reclaim mmu_notifier_invalidate_range_start pgd_lock irq_context: 0 fs_reclaim mmu_notifier_invalidate_range_start stock_lock irq_context: 0 fs_reclaim mmu_notifier_invalidate_range_start key irq_context: 0 fs_reclaim mmu_notifier_invalidate_range_start pcpu_lock irq_context: 0 fs_reclaim mmu_notifier_invalidate_range_start percpu_counters_lock irq_context: 0 fs_reclaim mmu_notifier_invalidate_range_start pcpu_lock stock_lock irq_context: 0 sb_writers#4 &type->i_mutex_dir_key#3 inode_hash_lock &sb->s_type->i_lock_key#22 bit_wait_table + i irq_context: 0 &sb->s_type->i_lock_key#22 bit_wait_table + i irq_context: 0 &sb->s_type->i_lock_key#22 bit_wait_table + i &p->pi_lock irq_context: 0 &sb->s_type->i_lock_key#22 bit_wait_table + i &p->pi_lock &rq->__lock irq_context: 0 &sb->s_type->i_lock_key#22 bit_wait_table + i &p->pi_lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 sb_writers#4 sb_internal jbd2_handle &ei->i_es_lock quarantine_lock irq_context: 0 sb_writers#4 &journal->j_barrier &journal->j_checkpoint_mutex &rq->__lock &cfs_rq->removed.lock irq_context: 0 &f->f_pos_lock sb_writers#4 &sb->s_type->i_mutex_key#8 jbd2_handle &c->lock irq_context: 0 sb_writers#4 &type->i_mutex_dir_key#3/1 &type->i_mutex_dir_key#3 jbd2_handle rcu_read_lock &rcu_state.gp_wq irq_context: 0 sb_writers#4 &type->i_mutex_dir_key#3/1 &type->i_mutex_dir_key#3 jbd2_handle rcu_read_lock &rcu_state.gp_wq &p->pi_lock irq_context: 0 sb_writers#4 &type->i_mutex_dir_key#3/1 &type->i_mutex_dir_key#3 jbd2_handle rcu_read_lock &rcu_state.gp_wq &p->pi_lock &rq->__lock irq_context: 0 sb_writers#4 &type->i_mutex_dir_key#3/1 &type->i_mutex_dir_key#3 jbd2_handle rcu_read_lock &rcu_state.gp_wq &p->pi_lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 sk_lock-AF_INET remove_cache_srcu rcu_read_lock &cfs_rq->removed.lock irq_context: 0 sk_lock-AF_INET remove_cache_srcu rcu_read_lock &obj_hash[i].lock irq_context: softirq rcu_read_lock rcu_read_lock rcu_read_lock &pcp->lock &zone->lock &____s->seqcount irq_context: 0 &f->f_pos_lock sb_writers#4 &sbi->s_writepages_rwsem &____s->seqcount#2 irq_context: 0 &f->f_pos_lock sb_writers#4 &sbi->s_writepages_rwsem &____s->seqcount irq_context: 0 &f->f_pos_lock sb_writers#4 &sbi->s_writepages_rwsem jbd2_handle &ei->i_data_sem &ei->i_es_lock key#7 irq_context: 0 &f->f_pos_lock sb_writers#4 &sbi->s_writepages_rwsem jbd2_handle &mapping->i_mmap_rwsem mmu_notifier_invalidate_range_start irq_context: 0 &f->f_pos_lock sb_writers#4 &sbi->s_writepages_rwsem jbd2_handle &mapping->i_mmap_rwsem ptlock_ptr(page)#2 irq_context: 0 &f->f_pos_lock sb_writers#4 &sbi->s_writepages_rwsem jbd2_handle &folio_wait_table[i] irq_context: 0 &f->f_pos_lock sb_writers#4 &sbi->s_writepages_rwsem jbd2_handle &folio_wait_table[i] &p->pi_lock irq_context: 0 &f->f_pos_lock sb_writers#4 &sbi->s_writepages_rwsem jbd2_handle &folio_wait_table[i] &p->pi_lock &rq->__lock irq_context: 0 &f->f_pos_lock sb_writers#4 &sbi->s_writepages_rwsem jbd2_handle &folio_wait_table[i] &p->pi_lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 &f->f_pos_lock sb_writers#4 &sbi->s_writepages_rwsem jbd2_handle tk_core.seq.seqcount irq_context: 0 &f->f_pos_lock sb_writers#4 &sbi->s_writepages_rwsem jbd2_handle &dd->lock irq_context: 0 &f->f_pos_lock sb_writers#4 &sbi->s_writepages_rwsem jbd2_handle rcu_read_lock &dd->lock irq_context: 0 &f->f_pos_lock sb_writers#4 &sbi->s_writepages_rwsem jbd2_handle rcu_read_lock &obj_hash[i].lock irq_context: 0 &f->f_pos_lock sb_writers#4 &sbi->s_writepages_rwsem jbd2_handle rcu_read_lock pool_lock#2 irq_context: 0 &f->f_pos_lock sb_writers#4 &sbi->s_writepages_rwsem jbd2_handle rcu_read_lock tk_core.seq.seqcount irq_context: 0 &f->f_pos_lock sb_writers#4 &sbi->s_writepages_rwsem jbd2_handle rcu_read_lock &virtscsi_vq->vq_lock irq_context: 0 &f->f_pos_lock sb_writers#4 &sbi->s_writepages_rwsem jbd2_handle &journal->j_wait_updates &p->pi_lock irq_context: 0 &f->f_pos_lock sb_writers#4 &sbi->s_writepages_rwsem jbd2_handle &journal->j_wait_updates &p->pi_lock &rq->__lock irq_context: 0 &f->f_pos_lock sb_writers#4 &sbi->s_writepages_rwsem jbd2_handle &journal->j_wait_updates &p->pi_lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 &f->f_pos_lock sb_writers#4 &sbi->s_writepages_rwsem &folio_wait_table[i] irq_context: 0 &f->f_pos_lock sb_writers#4 &sbi->s_writepages_rwsem &folio_wait_table[i] &p->pi_lock irq_context: 0 &f->f_pos_lock sb_writers#4 &sbi->s_writepages_rwsem &folio_wait_table[i] &p->pi_lock &rq->__lock irq_context: 0 &f->f_pos_lock sb_writers#4 &sbi->s_writepages_rwsem &folio_wait_table[i] &p->pi_lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 sb_writers#4 &sb->s_type->i_mutex_key#8 &sb->s_type->i_mutex_key#8/4 jbd2_handle &folio_wait_table[i] irq_context: 0 &f->f_pos_lock sb_writers#4 &sbi->s_writepages_rwsem jbd2_handle &ei->i_data_sem &ret->b_state_lock irq_context: 0 &f->f_pos_lock sb_writers#4 &sbi->s_writepages_rwsem jbd2_handle &ei->i_data_sem &ret->b_state_lock &journal->j_list_lock irq_context: 0 &f->f_pos_lock sb_writers#4 &sbi->s_writepages_rwsem jbd2_handle &mapping->i_mmap_rwsem &rq->__lock irq_context: 0 &f->f_pos_lock sb_writers#4 &sbi->s_writepages_rwsem jbd2_handle &mapping->i_mmap_rwsem &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 sb_writers#4 &sb->s_type->i_mutex_key#8 &sb->s_type->i_mutex_key#8/4 jbd2_handle &ei->i_data_sem &ei->i_data_sem/1 &ei->i_es_lock &c->lock irq_context: 0 sb_writers#4 &sb->s_type->i_mutex_key#8 &sb->s_type->i_mutex_key#8/4 jbd2_handle batched_entropy_u8.lock irq_context: 0 sb_writers#4 &sb->s_type->i_mutex_key#8 &sb->s_type->i_mutex_key#8/4 jbd2_handle kfence_freelist_lock irq_context: 0 sb_writers#4 &sb->s_type->i_mutex_key#8 &sb->s_type->i_mutex_key#8/4 jbd2_handle rcu_read_lock &xa->xa_lock#9 key#10 irq_context: 0 sb_writers#4 &sb->s_type->i_mutex_key#8 &sb->s_type->i_mutex_key#8/4 jbd2_handle &ei->i_data_sem &c->lock irq_context: 0 sb_writers#4 &sb->s_type->i_mutex_key#8 &sb->s_type->i_mutex_key#8/4 jbd2_handle &ei->i_es_lock key#5 irq_context: 0 &f->f_pos_lock sb_writers#4 lock#4 rcu_read_lock pool_lock#2 irq_context: 0 &f->f_pos_lock sb_writers#4 lock#4 &obj_hash[i].lock irq_context: 0 sb_writers#4 &sbi->s_writepages_rwsem jbd2_handle &ei->i_data_sem &lg->lg_mutex irq_context: 0 sb_writers#4 &sbi->s_writepages_rwsem jbd2_handle &ei->i_data_sem &lg->lg_mutex &ei->i_prealloc_lock irq_context: 0 sb_writers#4 &sbi->s_writepages_rwsem jbd2_handle &ei->i_data_sem &lg->lg_mutex rcu_read_lock &pa->pa_lock irq_context: 0 sb_writers#4 &sbi->s_writepages_rwsem jbd2_handle &ei->i_data_sem &lg->lg_mutex &mapping->private_lock irq_context: 0 sb_writers#4 &sbi->s_writepages_rwsem jbd2_handle &ei->i_data_sem &lg->lg_mutex mmu_notifier_invalidate_range_start irq_context: 0 sb_writers#4 &sbi->s_writepages_rwsem jbd2_handle &ei->i_data_sem &lg->lg_mutex &c->lock irq_context: 0 sb_writers#4 &sbi->s_writepages_rwsem jbd2_handle &ei->i_data_sem &lg->lg_mutex &ret->b_state_lock irq_context: 0 sb_writers#4 &sbi->s_writepages_rwsem jbd2_handle &ei->i_data_sem &lg->lg_mutex &ret->b_state_lock &journal->j_list_lock irq_context: 0 sb_writers#4 &sbi->s_writepages_rwsem jbd2_handle &ei->i_data_sem &lg->lg_mutex &pa->pa_lock irq_context: 0 sb_writers#4 &sbi->s_writepages_rwsem jbd2_handle &ei->i_data_sem &lg->lg_mutex &lg->lg_prealloc_lock irq_context: 0 sk_lock-AF_INET rcu_read_lock &rq->__lock &cfs_rq->removed.lock irq_context: 0 sk_lock-AF_INET rcu_read_lock &cfs_rq->removed.lock irq_context: 0 sb_writers#4 sb_internal pgd_lock irq_context: 0 sb_writers#4 sb_internal rcu_read_lock pool_lock#2 irq_context: 0 sb_writers#4 sb_internal key irq_context: 0 sb_writers#4 sb_internal pcpu_lock irq_context: 0 sb_writers#4 sb_internal percpu_counters_lock irq_context: 0 &f->f_pos_lock &type->i_mutex_dir_key#3 rcu_read_lock &cfs_rq->removed.lock irq_context: 0 &f->f_pos_lock &type->i_mutex_dir_key#3 rcu_read_lock &obj_hash[i].lock irq_context: 0 sb_writers#4 &sbi->s_writepages_rwsem jbd2_handle &ei->i_data_sem rcu_node_0 irq_context: 0 sb_writers#4 &sbi->s_writepages_rwsem jbd2_handle &ei->i_data_sem &rq->__lock irq_context: 0 sb_writers#4 &sbi->s_writepages_rwsem jbd2_handle &ei->i_data_sem rcu_read_lock rcu_node_0 irq_context: 0 sb_writers#4 &sbi->s_writepages_rwsem jbd2_handle &ei->i_data_sem &ei->i_es_lock key#2 irq_context: 0 rtnl_mutex sk_lock-AF_INET slock-AF_INET &sk->sk_lock.wq irq_context: 0 &sb->s_type->i_mutex_key#10 (netlink_chain).rwsem rcu_read_lock &cfs_rq->removed.lock irq_context: 0 &sb->s_type->i_mutex_key#10 (netlink_chain).rwsem rcu_read_lock &obj_hash[i].lock irq_context: 0 &sb->s_type->i_mutex_key#10 (netlink_chain).rwsem rcu_read_lock pool_lock#2 irq_context: 0 &sb->s_type->i_mutex_key#10 (netlink_chain).rwsem rcu_read_lock &rq->__lock &cfs_rq->removed.lock irq_context: 0 &f->f_pos_lock sb_writers#4 &sbi->s_writepages_rwsem &journal->j_state_lock irq_context: 0 &f->f_pos_lock sb_writers#4 &sbi->s_writepages_rwsem &journal->j_state_lock tk_core.seq.seqcount irq_context: 0 &f->f_pos_lock sb_writers#4 &sbi->s_writepages_rwsem &journal->j_state_lock &obj_hash[i].lock irq_context: 0 &f->f_pos_lock sb_writers#4 &sbi->s_writepages_rwsem &journal->j_state_lock &base->lock irq_context: 0 &f->f_pos_lock sb_writers#4 &sbi->s_writepages_rwsem &journal->j_state_lock &base->lock &obj_hash[i].lock irq_context: 0 &f->f_pos_lock sb_writers#4 &sbi->s_writepages_rwsem &rq->__lock &cfs_rq->removed.lock irq_context: 0 &f->f_pos_lock sb_writers#4 &sbi->s_writepages_rwsem jbd2_handle &mapping->private_lock irq_context: 0 &f->f_pos_lock sb_writers#4 &sbi->s_writepages_rwsem jbd2_handle &mapping->private_lock rcu_read_lock &memcg->move_lock irq_context: 0 sb_writers#4 &sb->s_type->i_mutex_key#8 mapping.invalidate_lock jbd2_handle &____s->seqcount irq_context: 0 rcu_read_lock key#10 irq_context: 0 sk_lock-AF_INET l2tp_ip_lock irq_context: 0 pernet_ops_rwsem devices_rwsem rdma_nets_rwsem &device->compat_devs_mutex stock_lock irq_context: 0 pernet_ops_rwsem devices_rwsem rdma_nets_rwsem &device->compat_devs_mutex pcpu_lock stock_lock irq_context: softirq rcu_read_lock rcu_read_lock nl_table_wait.lock &p->pi_lock irq_context: 0 rtnl_mutex &p->alloc_lock irq_context: 0 rtnl_mutex &rdev->wiphy.mtx &wdev->mtx &rdev->bss_lock irq_context: 0 rtnl_mutex &rdev->wiphy.mtx &wdev->mtx &local->sta_mtx irq_context: 0 rtnl_mutex &rdev->wiphy.mtx &wdev->mtx &local->sta_mtx &sta->ampdu_mlme.mtx irq_context: 0 rtnl_mutex &rdev->wiphy.mtx &wdev->mtx &local->sta_mtx &sta->ampdu_mlme.mtx &sta->lock irq_context: 0 rtnl_mutex &rdev->wiphy.mtx &wdev->mtx &local->sta_mtx (work_completion)(&sta->ampdu_mlme.work) irq_context: 0 rtnl_mutex &rdev->wiphy.mtx &wdev->mtx &local->sta_mtx rcu_read_lock rhashtable_bucket irq_context: 0 rtnl_mutex &rdev->wiphy.mtx &wdev->mtx &local->sta_mtx rcu_state.exp_mutex rcu_node_0 irq_context: 0 rtnl_mutex &rdev->wiphy.mtx &wdev->mtx &local->sta_mtx rcu_state.exp_mutex &obj_hash[i].lock irq_context: 0 rtnl_mutex &rdev->wiphy.mtx &wdev->mtx &local->sta_mtx rcu_state.exp_mutex rcu_read_lock &pool->lock irq_context: 0 rtnl_mutex &rdev->wiphy.mtx &wdev->mtx &local->sta_mtx rcu_state.exp_mutex rcu_read_lock &pool->lock &obj_hash[i].lock irq_context: 0 rtnl_mutex &rdev->wiphy.mtx &wdev->mtx &local->sta_mtx rcu_state.exp_mutex rcu_read_lock &pool->lock &p->pi_lock irq_context: 0 rtnl_mutex &rdev->wiphy.mtx &wdev->mtx &local->sta_mtx rcu_state.exp_mutex rcu_read_lock &pool->lock &p->pi_lock &rq->__lock irq_context: 0 rtnl_mutex &rdev->wiphy.mtx &wdev->mtx &local->sta_mtx rcu_state.exp_mutex rcu_read_lock &pool->lock &p->pi_lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 rtnl_mutex &rdev->wiphy.mtx &wdev->mtx &local->sta_mtx rcu_state.exp_mutex &rq->__lock irq_context: 0 rtnl_mutex &rdev->wiphy.mtx &wdev->mtx &local->sta_mtx rcu_state.exp_mutex &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 rtnl_mutex &rdev->wiphy.mtx &wdev->mtx &local->sta_mtx &obj_hash[i].lock irq_context: 0 rtnl_mutex &rdev->wiphy.mtx &wdev->mtx &local->sta_mtx &sta->lock irq_context: 0 rtnl_mutex &rdev->wiphy.mtx &wdev->mtx &local->sta_mtx pool_lock#2 irq_context: 0 rtnl_mutex &rdev->wiphy.mtx &wdev->mtx &local->sta_mtx krc.lock irq_context: 0 rtnl_mutex &rdev->wiphy.mtx &wdev->mtx &local->sta_mtx &local->key_mtx irq_context: 0 rtnl_mutex &rdev->wiphy.mtx &wdev->mtx &local->sta_mtx fs_reclaim irq_context: 0 rtnl_mutex &rdev->wiphy.mtx &wdev->mtx &local->sta_mtx fs_reclaim mmu_notifier_invalidate_range_start irq_context: 0 rtnl_mutex &rdev->wiphy.mtx &wdev->mtx &local->sta_mtx &fq->lock irq_context: 0 rtnl_mutex &rdev->wiphy.mtx &wdev->mtx &local->sta_mtx &c->lock irq_context: 0 rtnl_mutex &rdev->wiphy.mtx &wdev->mtx &local->sta_mtx &____s->seqcount#2 irq_context: 0 rtnl_mutex &rdev->wiphy.mtx &wdev->mtx &local->sta_mtx &____s->seqcount irq_context: 0 rtnl_mutex &rdev->wiphy.mtx &wdev->mtx &local->sta_mtx nl_table_lock irq_context: 0 rtnl_mutex &rdev->wiphy.mtx &wdev->mtx &local->sta_mtx &rq->__lock irq_context: 0 rtnl_mutex &rdev->wiphy.mtx &wdev->mtx &local->sta_mtx &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 rtnl_mutex &rdev->wiphy.mtx &wdev->mtx &local->sta_mtx nl_table_wait.lock irq_context: 0 rtnl_mutex &rdev->wiphy.mtx &wdev->mtx &local->sta_mtx pin_fs_lock irq_context: 0 rtnl_mutex &rdev->wiphy.mtx &wdev->mtx &local->sta_mtx &dentry->d_lock irq_context: 0 rtnl_mutex &rdev->wiphy.mtx &wdev->mtx &local->sta_mtx &sb->s_type->i_mutex_key#3 irq_context: 0 rtnl_mutex &rdev->wiphy.mtx &wdev->mtx &local->sta_mtx &sb->s_type->i_mutex_key#3 &dentry->d_lock irq_context: 0 rtnl_mutex &rdev->wiphy.mtx &wdev->mtx &local->sta_mtx &sb->s_type->i_mutex_key#3 &dentry->d_lock &dentry->d_lock/1 irq_context: 0 rtnl_mutex &rdev->wiphy.mtx &wdev->mtx &local->sta_mtx &sb->s_type->i_mutex_key#3 tk_core.seq.seqcount irq_context: 0 rtnl_mutex &rdev->wiphy.mtx &wdev->mtx &local->sta_mtx &sb->s_type->i_mutex_key#3 rename_lock.seqcount irq_context: 0 rtnl_mutex &rdev->wiphy.mtx &wdev->mtx &local->sta_mtx &sb->s_type->i_mutex_key#3 pin_fs_lock irq_context: 0 rtnl_mutex &rdev->wiphy.mtx &wdev->mtx &local->sta_mtx &sb->s_type->i_mutex_key#3 rcu_read_lock mount_lock irq_context: 0 rtnl_mutex &rdev->wiphy.mtx &wdev->mtx &local->sta_mtx &sb->s_type->i_mutex_key#3 rcu_read_lock mount_lock mount_lock.seqcount irq_context: 0 rtnl_mutex &rdev->wiphy.mtx &wdev->mtx &local->sta_mtx &sb->s_type->i_mutex_key#3 mount_lock irq_context: 0 rtnl_mutex &rdev->wiphy.mtx &wdev->mtx &local->sta_mtx &sb->s_type->i_mutex_key#3 mount_lock mount_lock.seqcount irq_context: 0 rtnl_mutex &rdev->wiphy.mtx &wdev->mtx &local->sta_mtx &sb->s_type->i_mutex_key#3 rcu_read_lock &dentry->d_lock irq_context: 0 rtnl_mutex &rdev->wiphy.mtx &wdev->mtx &local->sta_mtx &sb->s_type->i_mutex_key#3 &fsnotify_mark_srcu irq_context: 0 rtnl_mutex &rdev->wiphy.mtx &wdev->mtx &local->sta_mtx &sb->s_type->i_mutex_key#3 &sb->s_type->i_lock_key#7 irq_context: 0 rtnl_mutex &rdev->wiphy.mtx &wdev->mtx &local->sta_mtx &sb->s_type->i_mutex_key#3 &rq->__lock irq_context: 0 rtnl_mutex &rdev->wiphy.mtx &wdev->mtx &local->sta_mtx &sb->s_type->i_mutex_key#3 &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 rtnl_mutex &rdev->wiphy.mtx &wdev->mtx &local->sta_mtx &sb->s_type->i_mutex_key#3 &s->s_inode_list_lock irq_context: 0 rtnl_mutex &rdev->wiphy.mtx &wdev->mtx &local->sta_mtx &sb->s_type->i_mutex_key#3 &xa->xa_lock#9 irq_context: 0 rtnl_mutex &rdev->wiphy.mtx &wdev->mtx &local->sta_mtx &sb->s_type->i_mutex_key#3 &obj_hash[i].lock irq_context: 0 rtnl_mutex &rdev->wiphy.mtx &wdev->mtx &local->sta_mtx &sb->s_type->i_mutex_key#3 pool_lock#2 irq_context: 0 rtnl_mutex &rdev->wiphy.mtx &wdev->mtx &local->sta_mtx rcu_read_lock &dentry->d_lock irq_context: 0 rtnl_mutex &rdev->wiphy.mtx &wdev->mtx &local->sta_mtx &fsnotify_mark_srcu irq_context: 0 rtnl_mutex &rdev->wiphy.mtx &wdev->mtx &local->sta_mtx &sb->s_type->i_lock_key#7 irq_context: 0 rtnl_mutex &rdev->wiphy.mtx &wdev->mtx &local->sta_mtx &s->s_inode_list_lock irq_context: 0 rtnl_mutex &rdev->wiphy.mtx &wdev->mtx &local->sta_mtx &xa->xa_lock#9 irq_context: 0 rtnl_mutex &rdev->wiphy.mtx &wdev->mtx &local->sta_mtx rcu_read_lock mount_lock irq_context: 0 rtnl_mutex &rdev->wiphy.mtx &wdev->mtx &local->sta_mtx rcu_read_lock mount_lock mount_lock.seqcount irq_context: 0 rtnl_mutex &rdev->wiphy.mtx &wdev->mtx &local->sta_mtx mount_lock irq_context: 0 rtnl_mutex &rdev->wiphy.mtx &wdev->mtx &local->sta_mtx mount_lock mount_lock.seqcount irq_context: 0 rtnl_mutex &rdev->wiphy.mtx &wdev->mtx &local->sta_mtx &local->active_txq_lock[i] irq_context: 0 rtnl_mutex &rdev->wiphy.mtx &wdev->mtx &local->sta_mtx (work_completion)(&sta->drv_deliver_wk) irq_context: 0 rtnl_mutex &rdev->wiphy.mtx &wdev->mtx &local->sta_mtx rcu_read_lock pool_lock#2 irq_context: 0 rtnl_mutex &rdev->wiphy.mtx &wdev->mtx &local->sta_mtx &local->chanctx_mtx irq_context: 0 rtnl_mutex &rdev->wiphy.mtx &wdev->mtx &ifibss->incomplete_lock irq_context: 0 rtnl_mutex &rdev->wiphy.mtx &wdev->mtx lweventlist_lock irq_context: 0 rtnl_mutex &rdev->wiphy.mtx &wdev->mtx lweventlist_lock pool_lock#2 irq_context: 0 rtnl_mutex &rdev->wiphy.mtx &wdev->mtx lweventlist_lock &dir->lock#2 irq_context: 0 rtnl_mutex &rdev->wiphy.mtx &wdev->mtx rcu_read_lock &pool->lock irq_context: 0 rtnl_mutex &rdev->wiphy.mtx &wdev->mtx rcu_read_lock &pool->lock &obj_hash[i].lock irq_context: 0 rtnl_mutex &rdev->wiphy.mtx &wdev->mtx rcu_read_lock &pool->lock &p->pi_lock irq_context: 0 rtnl_mutex &rdev->wiphy.mtx &wdev->mtx rcu_read_lock &pool->lock &p->pi_lock &rq->__lock irq_context: 0 rtnl_mutex &rdev->wiphy.mtx &wdev->mtx rcu_read_lock &pool->lock &p->pi_lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 rtnl_mutex &rdev->wiphy.mtx &wdev->mtx &rq->__lock irq_context: 0 rtnl_mutex &rdev->wiphy.mtx &wdev->mtx &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 rtnl_mutex &rdev->wiphy.mtx &wdev->mtx &obj_hash[i].lock irq_context: 0 rtnl_mutex &rdev->wiphy.mtx &wdev->mtx pool_lock#2 irq_context: 0 rtnl_mutex &rdev->wiphy.mtx &wdev->mtx krc.lock irq_context: 0 rtnl_mutex &rdev->wiphy.mtx &wdev->mtx hrtimer_bases.lock irq_context: 0 rtnl_mutex &rdev->wiphy.mtx &wdev->mtx hrtimer_bases.lock &obj_hash[i].lock irq_context: 0 rtnl_mutex &rdev->wiphy.mtx &wdev->mtx &local->mtx irq_context: 0 rtnl_mutex &rdev->wiphy.mtx &wdev->mtx &local->mtx &local->chanctx_mtx irq_context: 0 rtnl_mutex &rdev->wiphy.mtx &wdev->mtx &local->mtx &local->chanctx_mtx &data->mutex irq_context: 0 rtnl_mutex &rdev->wiphy.mtx &wdev->mtx &local->mtx &local->chanctx_mtx &local->queue_stop_reason_lock irq_context: 0 rtnl_mutex &rdev->wiphy.mtx &wdev->mtx &local->mtx &local->chanctx_mtx rcu_read_lock &fq->lock irq_context: 0 rtnl_mutex &rdev->wiphy.mtx &wdev->mtx &local->mtx &local->chanctx_mtx rcu_read_lock &local->queue_stop_reason_lock irq_context: 0 rtnl_mutex &rdev->wiphy.mtx &wdev->mtx &local->mtx &local->chanctx_mtx &obj_hash[i].lock irq_context: 0 rtnl_mutex &rdev->wiphy.mtx &wdev->mtx &local->mtx &local->chanctx_mtx pool_lock#2 irq_context: 0 rtnl_mutex &rdev->wiphy.mtx &wdev->mtx &local->mtx &local->chanctx_mtx krc.lock irq_context: 0 rtnl_mutex &rdev->wiphy.mtx &wdev->mtx &local->mtx &local->chanctx_mtx &rq->__lock irq_context: 0 rtnl_mutex &rdev->wiphy.mtx &wdev->mtx &local->mtx &local->chanctx_mtx &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 rtnl_mutex &rdev->wiphy.mtx &wdev->mtx rcu_state.exp_mutex rcu_node_0 irq_context: 0 rtnl_mutex &rdev->wiphy.mtx &wdev->mtx rcu_state.exp_mutex &obj_hash[i].lock irq_context: 0 rtnl_mutex &rdev->wiphy.mtx &wdev->mtx rcu_state.exp_mutex rcu_read_lock &pool->lock irq_context: 0 rtnl_mutex &rdev->wiphy.mtx &wdev->mtx rcu_state.exp_mutex rcu_read_lock &pool->lock &obj_hash[i].lock irq_context: 0 rtnl_mutex &rdev->wiphy.mtx &wdev->mtx rcu_state.exp_mutex rcu_read_lock &pool->lock &p->pi_lock irq_context: 0 rtnl_mutex &rdev->wiphy.mtx &wdev->mtx rcu_state.exp_mutex rcu_read_lock &pool->lock &p->pi_lock &rq->__lock irq_context: 0 rtnl_mutex &rdev->wiphy.mtx &wdev->mtx rcu_state.exp_mutex rcu_read_lock &pool->lock &p->pi_lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 rtnl_mutex &rdev->wiphy.mtx &wdev->mtx rcu_state.exp_mutex &rnp->exp_wq[1] irq_context: 0 rtnl_mutex &rdev->wiphy.mtx &wdev->mtx rcu_state.exp_mutex &rq->__lock irq_context: 0 rtnl_mutex &rdev->wiphy.mtx &wdev->mtx rcu_state.exp_mutex &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 rtnl_mutex &rdev->wiphy.mtx &wdev->mtx &list->lock#18 irq_context: 0 rtnl_mutex &rdev->wiphy.mtx &wdev->mtx (&ifibss->timer) irq_context: 0 rtnl_mutex &rdev->wiphy.mtx &wdev->mtx &base->lock irq_context: 0 rtnl_mutex &rdev->wiphy.mtx &wdev->mtx &base->lock &obj_hash[i].lock irq_context: 0 rtnl_mutex &rdev->wiphy.mtx &wdev->mtx &rdev->bss_lock &obj_hash[i].lock irq_context: 0 rtnl_mutex &rdev->wiphy.mtx &wdev->mtx &rdev->bss_lock pool_lock#2 irq_context: 0 rtnl_mutex &rdev->wiphy.mtx &wdev->mtx &rdev->bss_lock krc.lock irq_context: 0 rtnl_mutex &rdev->wiphy.mtx &wdev->mtx rcu_read_lock &pool->lock/1 irq_context: 0 rtnl_mutex &rdev->wiphy.mtx &wdev->mtx rcu_read_lock &pool->lock/1 &obj_hash[i].lock irq_context: 0 rtnl_mutex &rdev->wiphy.mtx &wdev->mtx rcu_read_lock &pool->lock/1 &p->pi_lock irq_context: 0 rtnl_mutex &rdev->wiphy.mtx &wdev->mtx rcu_read_lock &pool->lock/1 &p->pi_lock &rq->__lock irq_context: 0 rtnl_mutex &rdev->wiphy.mtx &wdev->mtx rcu_read_lock &pool->lock/1 &p->pi_lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)cfg80211 (work_completion)(&(&rdev->dfs_update_channels_wk)->work) irq_context: 0 (wq_completion)cfg80211 (work_completion)(&(&rdev->dfs_update_channels_wk)->work) rtnl_mutex irq_context: 0 (wq_completion)cfg80211 (work_completion)(&(&rdev->dfs_update_channels_wk)->work) rtnl_mutex rtnl_mutex.wait_lock irq_context: 0 (wq_completion)cfg80211 (work_completion)(&(&rdev->dfs_update_channels_wk)->work) rtnl_mutex &rq->__lock irq_context: 0 (wq_completion)cfg80211 (work_completion)(&(&rdev->dfs_update_channels_wk)->work) rtnl_mutex &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 rtnl_mutex &list->lock#2 irq_context: 0 (wq_completion)cfg80211 (work_completion)(&(&rdev->dfs_update_channels_wk)->work) rtnl_mutex.wait_lock irq_context: 0 (wq_completion)cfg80211 (work_completion)(&(&rdev->dfs_update_channels_wk)->work) &p->pi_lock irq_context: 0 (wq_completion)cfg80211 (work_completion)(&(&rdev->dfs_update_channels_wk)->work) &p->pi_lock &rq->__lock irq_context: 0 (wq_completion)cfg80211 (work_completion)(&(&rdev->dfs_update_channels_wk)->work) &p->pi_lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 &p->lock &rcu_state.expedited_wq irq_context: 0 &p->lock &rcu_state.expedited_wq &p->pi_lock irq_context: 0 &p->lock &rcu_state.expedited_wq &p->pi_lock &rq->__lock irq_context: 0 &p->lock &rcu_state.expedited_wq &p->pi_lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 &sb->s_type->i_mutex_key#8 &sbi->s_writepages_rwsem jbd2_handle rcu_read_lock rcu_read_lock rcu_read_lock rcu_node_0 irq_context: 0 &sb->s_type->i_mutex_key#8 &sbi->s_writepages_rwsem jbd2_handle rcu_read_lock rcu_read_lock rcu_read_lock &rq->__lock irq_context: 0 &sb->s_type->i_mutex_key#8 &sbi->s_writepages_rwsem jbd2_handle rcu_read_lock rcu_read_lock rcu_read_lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem rcu_state.exp_mutex rcu_read_lock &pool->lock &p->pi_lock &rq->__lock &cfs_rq->removed.lock irq_context: 0 (wq_completion)events (work_completion)(&(&nsim_dev->trap_data->trap_report_dw)->work) &devlink->lock_key#5 &nsim_trap_data->trap_lock &base->lock irq_context: 0 (wq_completion)events (work_completion)(&(&nsim_dev->trap_data->trap_report_dw)->work) &devlink->lock_key#5 &nsim_trap_data->trap_lock &base->lock &obj_hash[i].lock irq_context: 0 cb_lock genl_mutex &rq->__lock &cfs_rq->removed.lock irq_context: 0 pernet_ops_rwsem devices_rwsem rdma_nets_rwsem &device->compat_devs_mutex remove_cache_srcu &base->lock irq_context: 0 pernet_ops_rwsem devices_rwsem rdma_nets_rwsem &device->compat_devs_mutex remove_cache_srcu &base->lock &obj_hash[i].lock irq_context: 0 &sbi->s_writepages_rwsem jbd2_handle &ei->i_data_sem &lg->lg_mutex key#3 irq_context: 0 rcu_read_lock rcu_read_lock rcu_read_lock_bh &obj_hash[i].lock irq_context: 0 rcu_read_lock rcu_read_lock rcu_read_lock_bh pool_lock#2 irq_context: 0 sb_writers#4 sb_internal &pcp->lock &zone->lock irq_context: 0 ppp_mutex rtnl_mutex &pn->all_ppp_mutex &n->list_lock irq_context: 0 ppp_mutex rtnl_mutex &pn->all_ppp_mutex &n->list_lock &c->lock irq_context: 0 ppp_mutex compressor_list_lock irq_context: 0 ppp_mutex rtnl_mutex &pn->all_ppp_mutex remove_cache_srcu irq_context: 0 ppp_mutex rtnl_mutex &pn->all_ppp_mutex remove_cache_srcu quarantine_lock irq_context: 0 ppp_mutex rtnl_mutex &pn->all_ppp_mutex remove_cache_srcu &c->lock irq_context: 0 ppp_mutex rtnl_mutex &pn->all_ppp_mutex remove_cache_srcu &n->list_lock irq_context: 0 ppp_mutex rtnl_mutex &pn->all_ppp_mutex remove_cache_srcu rcu_read_lock pool_lock#2 irq_context: 0 ppp_mutex rtnl_mutex &pn->all_ppp_mutex remove_cache_srcu &obj_hash[i].lock irq_context: 0 sb_writers#11 &type->i_mutex_dir_key#7 tomoyo_ss &rq->__lock irq_context: 0 sb_writers#11 &type->i_mutex_dir_key#7 tomoyo_ss &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 sb_writers#11 &type->i_mutex_dir_key#7 tomoyo_ss &n->list_lock irq_context: 0 sb_writers#11 &type->i_mutex_dir_key#7 tomoyo_ss &n->list_lock &c->lock irq_context: 0 pernet_ops_rwsem cpu_hotplug_lock wq_pool_mutex rcu_read_lock rcu_node_0 irq_context: 0 pernet_ops_rwsem cpu_hotplug_lock wq_pool_mutex rcu_read_lock &rq->__lock irq_context: 0 pernet_ops_rwsem cpu_hotplug_lock wq_pool_mutex rcu_read_lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem &root->kernfs_rwsem &rq->__lock &cfs_rq->removed.lock irq_context: 0 __ip_vs_mutex &mm->mmap_lock irq_context: 0 lock prog_idr_lock &____s->seqcount#2 irq_context: 0 lock prog_idr_lock &____s->seqcount irq_context: 0 ppp_mutex rtnl_mutex rcu_read_lock rcu_node_0 irq_context: 0 ppp_mutex rtnl_mutex rcu_read_lock &rq->__lock irq_context: 0 ppp_mutex rtnl_mutex rcu_read_lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem rtnl_mutex &caifn->caifdevs.lock quarantine_lock irq_context: 0 cb_lock genl_mutex rcu_read_lock rcu_read_lock_bh rcu_read_lock rlock-AF_PACKET irq_context: 0 cb_lock rcu_read_lock rcu_read_lock_bh rcu_read_lock rlock-AF_PACKET irq_context: 0 rcu_read_lock key#23 irq_context: 0 &sig->cred_guard_mutex tomoyo_ss remove_cache_srcu &pcp->lock &zone->lock irq_context: 0 &sig->cred_guard_mutex tomoyo_ss remove_cache_srcu &pcp->lock &zone->lock &____s->seqcount irq_context: 0 cb_lock rtnl_mutex &rdev->wiphy.mtx &local->iflist_mtx irq_context: 0 cb_lock rtnl_mutex &rdev->wiphy.mtx rcu_read_lock pool_lock#2 irq_context: 0 cb_lock rtnl_mutex &rdev->wiphy.mtx rcu_read_lock rcu_read_lock_bh rcu_read_lock pool_lock#2 irq_context: 0 cb_lock rtnl_mutex &rdev->wiphy.mtx rcu_read_lock rcu_read_lock_bh rcu_read_lock tk_core.seq.seqcount irq_context: 0 cb_lock rtnl_mutex &rdev->wiphy.mtx rcu_read_lock rcu_read_lock_bh rcu_read_lock rlock-AF_PACKET irq_context: 0 cb_lock rtnl_mutex &rdev->wiphy.mtx rcu_read_lock rcu_read_lock_bh &obj_hash[i].lock irq_context: 0 cb_lock rtnl_mutex &rdev->wiphy.mtx rcu_read_lock rcu_read_lock_bh pool_lock#2 irq_context: 0 cb_lock rtnl_mutex &rdev->wiphy.mtx rlock-AF_NETLINK irq_context: 0 cb_lock genl_mutex k-sk_lock-AF_INET irq_context: 0 cb_lock genl_mutex k-sk_lock-AF_INET &rq->__lock irq_context: 0 cb_lock genl_mutex k-sk_lock-AF_INET &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 cb_lock genl_mutex k-sk_lock-AF_INET k-slock-AF_INET irq_context: 0 cb_lock genl_mutex k-sk_lock-AF_INET &____s->seqcount#8 irq_context: 0 cb_lock genl_mutex k-sk_lock-AF_INET batched_entropy_u32.lock irq_context: 0 cb_lock genl_mutex k-sk_lock-AF_INET &table->hash[i].lock irq_context: 0 cb_lock genl_mutex k-sk_lock-AF_INET &table->hash[i].lock k-clock-AF_INET irq_context: 0 cb_lock genl_mutex k-sk_lock-AF_INET &table->hash[i].lock &table->hash2[i].lock irq_context: 0 cb_lock genl_mutex k-slock-AF_INET irq_context: 0 cb_lock genl_mutex &table->hash[i].lock irq_context: 0 cb_lock genl_mutex &table->hash[i].lock &table->hash2[i].lock irq_context: 0 cb_lock genl_mutex k-clock-AF_INET irq_context: 0 sk_lock-AF_CAN irq_context: 0 sk_lock-AF_CAN slock-AF_CAN irq_context: 0 sk_lock-AF_CAN clock-AF_CAN irq_context: 0 sk_lock-AF_CAN proc_subdir_lock irq_context: 0 sk_lock-AF_CAN fs_reclaim irq_context: 0 sk_lock-AF_CAN fs_reclaim mmu_notifier_invalidate_range_start irq_context: 0 sk_lock-AF_CAN pool_lock#2 irq_context: 0 sk_lock-AF_CAN proc_inum_ida.xa_lock irq_context: 0 sk_lock-AF_CAN proc_subdir_lock irq_context: 0 slock-AF_CAN irq_context: 0 &sb->s_type->i_mutex_key#10 sk_lock-AF_CAN proc_subdir_lock irq_context: 0 &sb->s_type->i_mutex_key#10 sk_lock-AF_CAN &ent->pde_unload_lock irq_context: 0 &sb->s_type->i_mutex_key#10 sk_lock-AF_CAN proc_inum_ida.xa_lock irq_context: 0 &sb->s_type->i_mutex_key#10 sk_lock-AF_CAN pool_lock#2 irq_context: 0 &sb->s_type->i_mutex_key#10 sk_lock-AF_CAN rcu_state.exp_mutex &rnp->exp_wq[0] irq_context: 0 &sb->s_type->i_mutex_key#10 (netlink_chain).rwsem rcu_read_lock rcu_read_lock &pool->lock irq_context: 0 &sb->s_type->i_mutex_key#10 (netlink_chain).rwsem rcu_read_lock rcu_read_lock &pool->lock &obj_hash[i].lock irq_context: 0 &sb->s_type->i_mutex_key#10 (netlink_chain).rwsem rcu_read_lock rcu_read_lock &pool->lock &p->pi_lock irq_context: 0 &sb->s_type->i_mutex_key#10 (netlink_chain).rwsem &rq->__lock &cfs_rq->removed.lock irq_context: 0 (wq_completion)events (work_completion)(&rdev->destroy_work) irq_context: 0 (wq_completion)events (work_completion)(&rdev->destroy_work) rtnl_mutex irq_context: 0 (wq_completion)events (work_completion)(&rdev->destroy_work) rtnl_mutex &rdev->wiphy.mtx irq_context: 0 (wq_completion)events (work_completion)(&rdev->destroy_work) rtnl_mutex &rdev->wiphy.mtx &wdev->mtx irq_context: 0 (wq_completion)events (work_completion)(&rdev->destroy_work) rtnl_mutex &rdev->wiphy.mtx &wdev->mtx &wdev->pmsr_lock irq_context: 0 (wq_completion)events (work_completion)(&rdev->destroy_work) rtnl_mutex &rdev->wiphy.mtx &local->iflist_mtx irq_context: 0 (wq_completion)events (work_completion)(&rdev->destroy_work) rtnl_mutex &rdev->wiphy.mtx rcu_state.exp_mutex rcu_node_0 irq_context: 0 (wq_completion)events (work_completion)(&rdev->destroy_work) rtnl_mutex &rdev->wiphy.mtx rcu_state.exp_mutex &obj_hash[i].lock irq_context: 0 (wq_completion)events (work_completion)(&rdev->destroy_work) rtnl_mutex &rdev->wiphy.mtx rcu_state.exp_mutex rcu_read_lock &pool->lock irq_context: 0 (wq_completion)events (work_completion)(&rdev->destroy_work) rtnl_mutex &rdev->wiphy.mtx rcu_state.exp_mutex rcu_read_lock &pool->lock &obj_hash[i].lock irq_context: 0 (wq_completion)events (work_completion)(&rdev->destroy_work) rtnl_mutex &rdev->wiphy.mtx rcu_state.exp_mutex &rnp->exp_wq[3] irq_context: 0 (wq_completion)events (work_completion)(&rdev->destroy_work) rtnl_mutex &rdev->wiphy.mtx rcu_state.exp_mutex &pool->lock irq_context: 0 (wq_completion)events (work_completion)(&rdev->destroy_work) rtnl_mutex &rdev->wiphy.mtx rcu_state.exp_mutex &pool->lock &p->pi_lock irq_context: 0 (wq_completion)events (work_completion)(&rdev->destroy_work) rtnl_mutex &rdev->wiphy.mtx rcu_state.exp_mutex &pool->lock &p->pi_lock &rq->__lock irq_context: 0 (wq_completion)events (work_completion)(&rdev->destroy_work) rtnl_mutex &rdev->wiphy.mtx rcu_state.exp_mutex &pool->lock &p->pi_lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)events (work_completion)(&rdev->destroy_work) rtnl_mutex &rdev->wiphy.mtx rcu_state.exp_mutex &rq->__lock irq_context: 0 (wq_completion)events (work_completion)(&rdev->destroy_work) rtnl_mutex &rdev->wiphy.mtx rcu_state.exp_mutex &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)events (work_completion)(&rdev->destroy_work) rtnl_mutex &rdev->wiphy.mtx &obj_hash[i].lock irq_context: 0 (wq_completion)events (work_completion)(&rdev->destroy_work) rtnl_mutex &rdev->wiphy.mtx fs_reclaim irq_context: 0 (wq_completion)events (work_completion)(&rdev->destroy_work) rtnl_mutex &rdev->wiphy.mtx fs_reclaim mmu_notifier_invalidate_range_start irq_context: 0 (wq_completion)events (work_completion)(&rdev->destroy_work) rtnl_mutex &rdev->wiphy.mtx pool_lock#2 irq_context: 0 (wq_completion)events (work_completion)(&rdev->destroy_work) rtnl_mutex &rdev->wiphy.mtx &fq->lock irq_context: 0 (wq_completion)events (work_completion)(&rdev->destroy_work) rtnl_mutex &rdev->wiphy.mtx nl_table_lock irq_context: 0 (wq_completion)events (work_completion)(&rdev->destroy_work) rtnl_mutex &rdev->wiphy.mtx nl_table_wait.lock irq_context: 0 (wq_completion)events (work_completion)(&rdev->destroy_work) rtnl_mutex &rdev->wiphy.mtx rcu_state.exp_mutex &rnp->exp_wq[0] irq_context: 0 (wq_completion)events (work_completion)(&rdev->destroy_work) rtnl_mutex &rdev->wiphy.mtx &rdev->mgmt_registrations_lock irq_context: 0 (wq_completion)events (work_completion)(&rdev->destroy_work) rtnl_mutex &rdev->wiphy.mtx &wdev->event_lock irq_context: 0 (wq_completion)events (work_completion)(&rdev->destroy_work) rtnl_mutex &rdev->wiphy.mtx (work_completion)(&(&sdata->dec_tailroom_needed_wk)->work) irq_context: 0 (wq_completion)events (work_completion)(&rdev->destroy_work) rtnl_mutex &rdev->wiphy.mtx &local->key_mtx irq_context: 0 (wq_completion)events (work_completion)(&rdev->destroy_work) rtnl_mutex &rdev->wiphy.mtx (work_completion)(&(&link->color_collision_detect_work)->work) irq_context: 0 (wq_completion)events (work_completion)(&rdev->destroy_work) rtnl_mutex &rdev->wiphy.mtx &local->chanctx_mtx irq_context: 0 (wq_completion)events (work_completion)(&rdev->destroy_work) rtnl_mutex &rdev->wiphy.mtx rcu_read_lock pool_lock#2 irq_context: 0 &sb->s_type->i_mutex_key#10 sk_lock-AF_CAN rcu_state.exp_mutex &obj_hash[i].lock pool_lock irq_context: 0 &sb->s_type->i_mutex_key#10 sk_lock-AF_CAN pool_lock irq_context: 0 sb_writers#4 &type->i_mutex_dir_key#3/1 jbd2_handle &ei->i_data_sem &ei->i_es_lock &pcp->lock &zone->lock irq_context: 0 sb_writers#3 &sb->s_type->i_mutex_key#9 rcu_read_lock &p->pi_lock irq_context: 0 sb_writers#3 &sb->s_type->i_mutex_key#9 rcu_read_lock &p->pi_lock &rq->__lock irq_context: 0 sb_writers#3 &sb->s_type->i_mutex_key#9 rcu_read_lock &p->pi_lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 sb_writers#3 &sb->s_type->i_mutex_key#9 &rcu_state.expedited_wq irq_context: 0 (wq_completion)wg-crypt-wg0#6 (work_completion)(&peer->transmit_packet_work) &peer->endpoint_lock rcu_read_lock_bh rcu_read_lock rcu_read_lock_bh rcu_read_lock rlock-AF_PACKET irq_context: 0 &sb->s_type->i_mutex_key#10 &p->pi_lock &cfs_rq->removed.lock irq_context: 0 (wq_completion)gid-cache-wq (work_completion)(&ndev_work->work) devices_rwsem &rq->__lock irq_context: 0 &smc->clcsock_release_lock irq_context: 0 &smc->clcsock_release_lock k-sk_lock-AF_INET irq_context: 0 &smc->clcsock_release_lock k-sk_lock-AF_INET k-slock-AF_INET irq_context: 0 &smc->clcsock_release_lock k-slock-AF_INET irq_context: 0 &smc->clcsock_release_lock nf_sockopt_mutex irq_context: 0 &smc->clcsock_release_lock &mm->mmap_lock irq_context: 0 &smc->clcsock_release_lock rtnl_mutex irq_context: 0 &smc->clcsock_release_lock rtnl_mutex ipvs->sync_mutex fs_reclaim irq_context: 0 &smc->clcsock_release_lock rtnl_mutex ipvs->sync_mutex fs_reclaim mmu_notifier_invalidate_range_start irq_context: 0 &smc->clcsock_release_lock rtnl_mutex ipvs->sync_mutex pool_lock#2 irq_context: 0 &smc->clcsock_release_lock rtnl_mutex ipvs->sync_mutex &c->lock irq_context: 0 &smc->clcsock_release_lock rtnl_mutex ipvs->sync_mutex stock_lock irq_context: 0 &smc->clcsock_release_lock rtnl_mutex ipvs->sync_mutex mmu_notifier_invalidate_range_start irq_context: 0 &smc->clcsock_release_lock rtnl_mutex ipvs->sync_mutex &sb->s_type->i_lock_key#8 irq_context: 0 &smc->clcsock_release_lock rtnl_mutex ipvs->sync_mutex &dir->lock irq_context: 0 &smc->clcsock_release_lock rtnl_mutex ipvs->sync_mutex &obj_hash[i].lock irq_context: 0 &smc->clcsock_release_lock rtnl_mutex ipvs->sync_mutex k-sk_lock-AF_INET irq_context: 0 &smc->clcsock_release_lock rtnl_mutex ipvs->sync_mutex k-sk_lock-AF_INET k-slock-AF_INET irq_context: 0 &smc->clcsock_release_lock rtnl_mutex ipvs->sync_mutex k-sk_lock-AF_INET &table->hash[i].lock irq_context: 0 &smc->clcsock_release_lock rtnl_mutex ipvs->sync_mutex k-sk_lock-AF_INET &table->hash[i].lock k-clock-AF_INET irq_context: 0 &smc->clcsock_release_lock rtnl_mutex ipvs->sync_mutex k-sk_lock-AF_INET &table->hash[i].lock &table->hash2[i].lock irq_context: 0 &smc->clcsock_release_lock rtnl_mutex ipvs->sync_mutex k-slock-AF_INET irq_context: 0 &smc->clcsock_release_lock rtnl_mutex ipvs->sync_mutex k-sk_lock-AF_INET fs_reclaim irq_context: 0 &smc->clcsock_release_lock rtnl_mutex ipvs->sync_mutex k-sk_lock-AF_INET fs_reclaim mmu_notifier_invalidate_range_start irq_context: 0 &smc->clcsock_release_lock rtnl_mutex ipvs->sync_mutex k-sk_lock-AF_INET pool_lock#2 irq_context: 0 &smc->clcsock_release_lock rtnl_mutex ipvs->sync_mutex k-sk_lock-AF_INET &obj_hash[i].lock irq_context: 0 &smc->clcsock_release_lock rtnl_mutex ipvs->sync_mutex k-sk_lock-AF_INET &in_dev->mc_tomb_lock irq_context: 0 &smc->clcsock_release_lock rtnl_mutex ipvs->sync_mutex k-sk_lock-AF_INET &im->lock irq_context: 0 &smc->clcsock_release_lock rtnl_mutex ipvs->sync_mutex k-sk_lock-AF_INET _xmit_ETHER irq_context: 0 &smc->clcsock_release_lock rtnl_mutex ipvs->sync_mutex k-sk_lock-AF_INET _xmit_ETHER &c->lock irq_context: 0 &smc->clcsock_release_lock rtnl_mutex ipvs->sync_mutex k-sk_lock-AF_INET _xmit_ETHER pool_lock#2 irq_context: 0 &smc->clcsock_release_lock rtnl_mutex ipvs->sync_mutex k-sk_lock-AF_INET &base->lock irq_context: 0 &smc->clcsock_release_lock rtnl_mutex ipvs->sync_mutex k-sk_lock-AF_INET &base->lock &obj_hash[i].lock irq_context: 0 &smc->clcsock_release_lock rtnl_mutex ipvs->sync_mutex kthread_create_lock irq_context: 0 &smc->clcsock_release_lock rtnl_mutex ipvs->sync_mutex &p->pi_lock irq_context: 0 &smc->clcsock_release_lock rtnl_mutex ipvs->sync_mutex &p->pi_lock &rq->__lock irq_context: 0 &smc->clcsock_release_lock rtnl_mutex ipvs->sync_mutex &p->pi_lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 &smc->clcsock_release_lock rtnl_mutex ipvs->sync_mutex &x->wait irq_context: 0 &smc->clcsock_release_lock rtnl_mutex ipvs->sync_mutex &rq->__lock irq_context: 0 &smc->clcsock_release_lock rtnl_mutex ipvs->sync_mutex &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: softirq (&in_dev->mr_ifc_timer) rcu_read_lock rcu_read_lock_bh dev->qdisc_tx_busylock ?: &qdisc_tx_busylock _xmit_ETHER#2 irq_context: 0 &smc->clcsock_release_lock rtnl_mutex ipvs->sync_mutex &ipvs->sync_buff_lock irq_context: softirq (&in_dev->mr_ifc_timer) rcu_read_lock rcu_read_lock_bh dev->qdisc_tx_busylock ?: &qdisc_tx_busylock _xmit_ETHER#2 &obj_hash[i].lock irq_context: softirq (&in_dev->mr_ifc_timer) rcu_read_lock rcu_read_lock_bh dev->qdisc_tx_busylock ?: &qdisc_tx_busylock _xmit_ETHER#2 pool_lock#2 irq_context: 0 &smc->clcsock_release_lock rtnl_mutex.wait_lock irq_context: 0 &smc->clcsock_release_lock &p->pi_lock irq_context: 0 &smc->clcsock_release_lock &p->pi_lock &rq->__lock irq_context: 0 &smc->clcsock_release_lock &p->pi_lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 &smc->clcsock_release_lock &rq->__lock irq_context: 0 &smc->clcsock_release_lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 &smc->clcsock_release_lock pgd_lock irq_context: 0 &smc->clcsock_release_lock stock_lock irq_context: 0 &smc->clcsock_release_lock rcu_read_lock pool_lock#2 irq_context: 0 &smc->clcsock_release_lock &obj_hash[i].lock irq_context: 0 &smc->clcsock_release_lock key irq_context: 0 &smc->clcsock_release_lock pcpu_lock irq_context: 0 &smc->clcsock_release_lock percpu_counters_lock irq_context: 0 &smc->clcsock_release_lock pcpu_lock stock_lock irq_context: 0 &smc->clcsock_release_lock pool_lock#2 irq_context: 0 &smc->clcsock_release_lock ipvs->sync_mutex irq_context: 0 &smc->clcsock_release_lock ipvs->sync_mutex (console_sem).lock irq_context: 0 &smc->clcsock_release_lock ipvs->sync_mutex console_lock console_srcu console_owner_lock irq_context: 0 &smc->clcsock_release_lock ipvs->sync_mutex console_lock console_srcu console_owner irq_context: 0 &smc->clcsock_release_lock ipvs->sync_mutex console_lock console_srcu console_owner &port_lock_key irq_context: 0 &smc->clcsock_release_lock ipvs->sync_mutex console_lock console_srcu console_owner console_owner_lock irq_context: 0 &smc->clcsock_release_lock ipvs->sync_mutex &rq->__lock irq_context: 0 &smc->clcsock_release_lock ipvs->sync_mutex &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 &smc->clcsock_release_lock ipvs->sync_mutex &p->pi_lock irq_context: 0 &smc->clcsock_release_lock ipvs->sync_mutex &p->pi_lock &rq->__lock irq_context: 0 &smc->clcsock_release_lock ipvs->sync_mutex &p->pi_lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 &smc->clcsock_release_lock ipvs->sync_mutex &x->wait irq_context: 0 &smc->clcsock_release_lock ipvs->sync_mutex &cfs_rq->removed.lock irq_context: 0 &smc->clcsock_release_lock ipvs->sync_mutex &obj_hash[i].lock irq_context: 0 &smc->clcsock_release_lock ipvs->sync_mutex pool_lock#2 irq_context: 0 &smc->clcsock_release_lock rtnl_mutex rtnl_mutex.wait_lock irq_context: 0 &smc->clcsock_release_lock rtnl_mutex &rq->__lock irq_context: 0 &smc->clcsock_release_lock rtnl_mutex &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 &smc->clcsock_release_lock rtnl_mutex &rq->__lock &cfs_rq->removed.lock irq_context: 0 &smc->clcsock_release_lock rtnl_mutex rcu_read_lock &im->lock irq_context: 0 &smc->clcsock_release_lock rtnl_mutex &im->lock irq_context: 0 &smc->clcsock_release_lock rtnl_mutex _xmit_ETHER irq_context: 0 &smc->clcsock_release_lock rtnl_mutex _xmit_ETHER &obj_hash[i].lock irq_context: 0 &smc->clcsock_release_lock rtnl_mutex _xmit_ETHER pool_lock#2 irq_context: 0 &smc->clcsock_release_lock rtnl_mutex _xmit_ETHER krc.lock irq_context: 0 &smc->clcsock_release_lock rtnl_mutex &im->lock &obj_hash[i].lock irq_context: 0 &smc->clcsock_release_lock rtnl_mutex fs_reclaim irq_context: 0 &smc->clcsock_release_lock rtnl_mutex fs_reclaim mmu_notifier_invalidate_range_start irq_context: 0 &smc->clcsock_release_lock rtnl_mutex pool_lock#2 irq_context: 0 &smc->clcsock_release_lock rtnl_mutex &in_dev->mc_tomb_lock irq_context: 0 &smc->clcsock_release_lock rtnl_mutex &obj_hash[i].lock irq_context: 0 &smc->clcsock_release_lock rtnl_mutex &base->lock irq_context: 0 &smc->clcsock_release_lock rtnl_mutex &base->lock &obj_hash[i].lock irq_context: 0 &smc->clcsock_release_lock rtnl_mutex krc.lock irq_context: 0 &smc->clcsock_release_lock &table->hash[i].lock irq_context: 0 &smc->clcsock_release_lock &table->hash[i].lock &table->hash2[i].lock irq_context: 0 &smc->clcsock_release_lock k-clock-AF_INET irq_context: 0 &smc->clcsock_release_lock &sb->s_type->i_lock_key#8 irq_context: 0 &smc->clcsock_release_lock &xa->xa_lock#9 irq_context: 0 &smc->clcsock_release_lock &fsnotify_mark_srcu irq_context: softirq (&in_dev->mr_ifc_timer) rcu_read_lock &in_dev->mc_tomb_lock pool_lock#2 irq_context: softirq (&in_dev->mr_ifc_timer) rcu_read_lock &in_dev->mc_tomb_lock rcu_read_lock pool_lock#2 irq_context: softirq (&in_dev->mr_ifc_timer) rcu_read_lock &in_dev->mc_tomb_lock rcu_read_lock &dir->lock#2 irq_context: softirq (&in_dev->mr_ifc_timer) rcu_read_lock &in_dev->mc_tomb_lock rcu_read_lock &ul->lock irq_context: 0 sk_lock-AF_INET6 lock irq_context: 0 sk_lock-AF_INET6 lock sctp_assocs_id_lock irq_context: 0 sk_lock-AF_INET6 lock sctp_assocs_id_lock &____s->seqcount irq_context: 0 sk_lock-AF_INET6 lock sctp_assocs_id_lock pool_lock#2 irq_context: 0 sk_lock-AF_INET6 lock sctp_assocs_id_lock rcu_read_lock pool_lock#2 irq_context: 0 sk_lock-AF_INET6 lock sctp_assocs_id_lock &obj_hash[i].lock irq_context: 0 &dir->lock#2 quarantine_lock irq_context: 0 pernet_ops_rwsem rtnl_mutex pgd_lock irq_context: 0 pernet_ops_rwsem rtnl_mutex key irq_context: 0 pernet_ops_rwsem rtnl_mutex pcpu_lock irq_context: 0 pernet_ops_rwsem rtnl_mutex percpu_counters_lock irq_context: 0 pernet_ops_rwsem rtnl_mutex pcpu_lock stock_lock irq_context: 0 (wq_completion)bond3 (work_completion)(&(&bond->mii_work)->work) &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)bat_events (work_completion)(&(&forw_packet_aggr->delayed_work)->work) &hard_iface->bat_iv.ogm_buff_mutex &bat_priv->tt.commit_lock &bat_priv->softif_vlan_list_lock &____s->seqcount irq_context: 0 &mm->mmap_lock ptlock_ptr(page)#2 lock#5 &lruvec->lru_lock irq_context: softirq (&in_dev->mr_ifc_timer) rcu_read_lock &in_dev->mc_tomb_lock &obj_hash[i].lock irq_context: softirq (&in_dev->mr_ifc_timer) rcu_read_lock rcu_read_lock_bh dev->qdisc_tx_busylock ?: &qdisc_tx_busylock _xmit_ETHER#2 &obj_hash[i].lock pool_lock irq_context: 0 (wq_completion)mm_percpu_wq (work_completion)(work) lock#5 &lruvec->lru_lock irq_context: 0 &mm->mmap_lock lock#5 &lruvec->lru_lock irq_context: 0 &xt[i].mutex &base->lock irq_context: 0 &xt[i].mutex &base->lock &obj_hash[i].lock irq_context: 0 &smc->clcsock_release_lock rtnl_mutex ipvs->sync_mutex k-sk_lock-AF_INET &c->lock irq_context: 0 &type->s_umount_key#23/1 &xa->xa_lock#4 &n->list_lock irq_context: 0 &type->s_umount_key#23/1 &xa->xa_lock#4 &n->list_lock &c->lock irq_context: 0 sk_lock-AF_INET6 rcu_read_lock rcu_read_lock rcu_read_lock &cfs_rq->removed.lock irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem rcu_read_lock rcu_read_lock rcu_node_0 irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem rcu_read_lock rcu_read_lock &rq->__lock irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem rcu_read_lock rcu_read_lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 &smc->clcsock_release_lock ipvs->sync_mutex &rq->__lock &cfs_rq->removed.lock irq_context: softirq (&in_dev->mr_ifc_timer) rcu_read_lock &in_dev->mc_tomb_lock &c->lock irq_context: softirq (&in_dev->mr_ifc_timer) rcu_read_lock &in_dev->mc_tomb_lock &n->list_lock irq_context: softirq (&in_dev->mr_ifc_timer) rcu_read_lock &in_dev->mc_tomb_lock &n->list_lock &c->lock irq_context: softirq (&in_dev->mr_ifc_timer) rcu_read_lock &in_dev->mc_tomb_lock rcu_read_lock &c->lock irq_context: softirq (&in_dev->mr_ifc_timer) rcu_read_lock &in_dev->mc_tomb_lock rcu_read_lock &n->list_lock irq_context: softirq (&in_dev->mr_ifc_timer) rcu_read_lock &in_dev->mc_tomb_lock rcu_read_lock &n->list_lock &c->lock irq_context: 0 sk_lock-AF_INET6 lock sctp_assocs_id_lock &c->lock irq_context: 0 &smc->clcsock_release_lock rtnl_mutex ipvs->sync_mutex k-sk_lock-AF_INET &rq->__lock irq_context: 0 &smc->clcsock_release_lock rtnl_mutex ipvs->sync_mutex k-sk_lock-AF_INET &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 &smc->clcsock_release_lock ipvs->sync_mutex console_owner_lock irq_context: 0 &smc->clcsock_release_lock ipvs->sync_mutex console_owner irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem rtnl_mutex fs_reclaim mmu_notifier_invalidate_range_start &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 &smc->clcsock_release_lock &lock->wait_lock irq_context: 0 &mm->mmap_lock &mapping->i_mmap_rwsem &anon_vma->rwsem &rcu_state.expedited_wq irq_context: 0 sb_writers#8 tomoyo_ss remove_cache_srcu rcu_read_lock pool_lock#2 irq_context: 0 sb_writers#4 &type->i_mutex_dir_key#3/1 &sb->s_type->i_mutex_key#8 jbd2_handle &sbi->s_orphan_lock &rq->__lock &cfs_rq->removed.lock irq_context: 0 sb_writers#5 &type->i_mutex_dir_key#5/1 lock#5 &lruvec->lru_lock irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem devices_rwsem &rq->__lock irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem devices_rwsem &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 &mm->mmap_lock ptlock_ptr(page)#2 lock#5 irq_context: 0 &mm->mmap_lock ptlock_ptr(page)#2 lock#5 &lruvec->lru_lock irq_context: 0 &mm->mmap_lock lock#5 &lruvec->lru_lock irq_context: 0 &mm->mmap_lock lock#5 rcu_read_lock pool_lock#2 irq_context: 0 &mm->mmap_lock lock#5 &obj_hash[i].lock irq_context: 0 &sb->s_type->i_mutex_key#10 &n->list_lock irq_context: 0 &type->s_umount_key#23/1 batched_entropy_u8.lock irq_context: 0 &type->s_umount_key#23/1 kfence_freelist_lock irq_context: softirq rcu_read_lock rcu_read_lock &sctp_ep_hashtable[i].lock irq_context: softirq rcu_read_lock rcu_read_lock k-slock-AF_INET6 irq_context: softirq rcu_read_lock rcu_read_lock k-slock-AF_INET6 &c->lock irq_context: softirq rcu_read_lock rcu_read_lock k-slock-AF_INET6 &obj_hash[i].lock pool_lock irq_context: softirq rcu_read_lock rcu_read_lock k-slock-AF_INET6 crngs.lock irq_context: softirq rcu_read_lock rcu_read_lock k-slock-AF_INET6 rcu_read_lock rcu_read_lock &____s->seqcount#7 irq_context: softirq rcu_read_lock rcu_read_lock k-slock-AF_INET6 rcu_read_lock rcu_read_lock &ct->lock irq_context: 0 sk_lock-AF_INET6 fill_pool_map-wait-type-override rcu_read_lock rcu_node_0 irq_context: 0 &f->f_pos_lock &type->i_mutex_dir_key#3 pgd_lock irq_context: 0 &f->f_pos_lock &type->i_mutex_dir_key#3 stock_lock irq_context: 0 &f->f_pos_lock &type->i_mutex_dir_key#3 key irq_context: 0 &f->f_pos_lock &type->i_mutex_dir_key#3 pcpu_lock irq_context: 0 &f->f_pos_lock &type->i_mutex_dir_key#3 percpu_counters_lock irq_context: 0 &f->f_pos_lock &type->i_mutex_dir_key#3 pcpu_lock stock_lock irq_context: 0 sk_lock-AF_INET6 rcu_read_lock rcu_read_lock rcu_read_lock per_cpu_ptr(&cgroup_rstat_cpu_lock, cpu) irq_context: softirq rcu_read_lock rcu_read_lock &dreq->dreq_lock &____s->seqcount#2 irq_context: softirq rcu_read_lock rcu_read_lock &dreq->dreq_lock &____s->seqcount irq_context: softirq rcu_read_lock rcu_read_lock &dreq->dreq_lock &n->list_lock irq_context: softirq rcu_read_lock rcu_read_lock &dreq->dreq_lock &n->list_lock &c->lock irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem rcu_read_lock &obj_hash[i].lock pool_lock irq_context: 0 pernet_ops_rwsem rtnl_mutex batched_entropy_u8.lock crngs.lock irq_context: 0 pernet_ops_rwsem uevent_sock_mutex batched_entropy_u8.lock irq_context: 0 pernet_ops_rwsem uevent_sock_mutex kfence_freelist_lock irq_context: 0 pernet_ops_rwsem uevent_sock_mutex &meta->lock irq_context: 0 rtnl_mutex &idev->mc_lock &macvlan_netdev_addr_lock_key/1 irq_context: 0 rtnl_mutex &idev->mc_lock &macvlan_netdev_addr_lock_key/1 &obj_hash[i].lock irq_context: 0 rtnl_mutex &idev->mc_lock &macvlan_netdev_addr_lock_key/1 krc.lock irq_context: 0 rtnl_mutex (work_completion)(&port->bc_work) irq_context: softirq rcu_read_lock rcu_read_lock rcu_read_lock_bh rcu_read_lock &pool->lock fill_pool_map-wait-type-override pool_lock#2 irq_context: softirq rcu_read_lock rcu_read_lock rcu_read_lock_bh rcu_read_lock &pool->lock fill_pool_map-wait-type-override &c->lock irq_context: softirq rcu_read_lock rcu_read_lock rcu_read_lock_bh rcu_read_lock &pool->lock fill_pool_map-wait-type-override pool_lock irq_context: 0 pernet_ops_rwsem rtnl_mutex rcu_read_lock &cfs_rq->removed.lock irq_context: 0 pernet_ops_rwsem rtnl_mutex rcu_read_lock &obj_hash[i].lock irq_context: 0 &sb->s_type->i_mutex_key#10 sk_lock-AF_INET6/1 rcu_read_lock rcu_read_lock rcu_read_lock_bh rcu_read_lock pool_lock#2 irq_context: 0 &sb->s_type->i_mutex_key#10 sk_lock-AF_INET6/1 rcu_read_lock rcu_read_lock rcu_read_lock_bh rcu_read_lock tk_core.seq.seqcount irq_context: 0 &sb->s_type->i_mutex_key#10 sk_lock-AF_INET6/1 rcu_read_lock rcu_read_lock rcu_read_lock_bh rcu_read_lock &obj_hash[i].lock irq_context: 0 pernet_ops_rwsem fill_pool_map-wait-type-override rcu_read_lock &rq->__lock irq_context: 0 pernet_ops_rwsem fill_pool_map-wait-type-override rcu_read_lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 pernet_ops_rwsem devices_rwsem rdma_nets_rwsem &device->compat_devs_mutex &obj_hash[i].lock pool_lock irq_context: softirq rcu_callback prog_idr_lock &obj_hash[i].lock irq_context: softirq rcu_callback prog_idr_lock pool_lock#2 irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem fill_pool_map-wait-type-override &n->list_lock irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem fill_pool_map-wait-type-override &n->list_lock &c->lock irq_context: 0 (wq_completion)bat_events (work_completion)(&(&bat_priv->orig_work)->work) fill_pool_map-wait-type-override &c->lock irq_context: 0 (wq_completion)bat_events (work_completion)(&(&bat_priv->orig_work)->work) fill_pool_map-wait-type-override &n->list_lock irq_context: 0 (wq_completion)bat_events (work_completion)(&(&bat_priv->orig_work)->work) fill_pool_map-wait-type-override &n->list_lock &c->lock irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem uevent_sock_mutex rcu_read_lock &rcu_state.expedited_wq irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem uevent_sock_mutex rcu_read_lock &rcu_state.expedited_wq &p->pi_lock irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem uevent_sock_mutex rcu_read_lock &rcu_state.expedited_wq &p->pi_lock &rq->__lock irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem uevent_sock_mutex rcu_read_lock &rcu_state.expedited_wq &p->pi_lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 &type->s_umount_key#23/1 rcu_read_lock rcu_node_0 irq_context: 0 &type->s_umount_key#23/1 rcu_read_lock &rq->__lock irq_context: 0 (wq_completion)events free_ipc_work sysctl_lock fill_pool_map-wait-type-override &n->list_lock irq_context: 0 (wq_completion)events free_ipc_work sysctl_lock fill_pool_map-wait-type-override &n->list_lock &c->lock irq_context: 0 (wq_completion)events free_ipc_work &xa->xa_lock#4 fill_pool_map-wait-type-override pool_lock#2 irq_context: 0 (wq_completion)events free_ipc_work &xa->xa_lock#4 fill_pool_map-wait-type-override &c->lock irq_context: 0 (wq_completion)events free_ipc_work &xa->xa_lock#4 fill_pool_map-wait-type-override &n->list_lock irq_context: 0 (wq_completion)events free_ipc_work &xa->xa_lock#4 fill_pool_map-wait-type-override &n->list_lock &c->lock irq_context: 0 (wq_completion)events free_ipc_work &xa->xa_lock#4 fill_pool_map-wait-type-override pool_lock irq_context: 0 namespace_sem fs_reclaim &rq->__lock irq_context: 0 &type->s_umount_key#23/1 shrinker_rwsem &rq->__lock irq_context: 0 (wq_completion)events free_ipc_work &type->s_umount_key#51 shrinker_rwsem shrinker_rwsem.wait_lock irq_context: 0 (wq_completion)events free_ipc_work &type->s_umount_key#51 shrinker_rwsem &pool->lock irq_context: 0 (wq_completion)events free_ipc_work &type->s_umount_key#51 shrinker_rwsem &rq->__lock irq_context: 0 (wq_completion)events free_ipc_work &type->s_umount_key#51 shrinker_rwsem &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 &type->s_umount_key#23/1 shrinker_rwsem.wait_lock irq_context: 0 &type->s_umount_key#23/1 &p->pi_lock irq_context: 0 &type->s_umount_key#23/1 &p->pi_lock &rq->__lock irq_context: 0 &type->s_umount_key#23/1 &p->pi_lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)events free_ipc_work &type->s_umount_key#51 &obj_hash[i].lock pool_lock irq_context: 0 (wq_completion)events free_ipc_work &type->s_umount_key#51 &rq->__lock irq_context: 0 (wq_completion)events free_ipc_work &type->s_umount_key#51 &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)events free_ipc_work &ht->mutex quarantine_lock irq_context: 0 (wq_completion)events free_ipc_work &rnp->exp_wq[1] irq_context: 0 (wq_completion)events free_ipc_work &pool->lock irq_context: 0 (wq_completion)events (work_completion)(&s->destroy_work) &base->lock irq_context: 0 (wq_completion)events (work_completion)(&s->destroy_work) &base->lock &obj_hash[i].lock irq_context: 0 &type->s_umount_key#23/1 &xa->xa_lock#4 &pcp->lock &zone->lock irq_context: 0 (wq_completion)events free_ipc_work &cfs_rq->removed.lock irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem ovs_mutex nf_ct_proto_mutex rcu_node_0 irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem ovs_mutex nf_ct_proto_mutex &rcu_state.expedited_wq irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem ovs_mutex nf_ct_proto_mutex &rcu_state.expedited_wq &p->pi_lock irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem ovs_mutex nf_ct_proto_mutex &rcu_state.expedited_wq &p->pi_lock &rq->__lock irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem ovs_mutex nf_ct_proto_mutex &rcu_state.expedited_wq &p->pi_lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)events (work_completion)(&pwq->unbound_release_work) &p->pi_lock &rq->__lock &cfs_rq->removed.lock irq_context: 0 &type->s_umount_key#23/1 pcpu_alloc_mutex pcpu_alloc_mutex.wait_lock irq_context: 0 (wq_completion)events (work_completion)(&pwq->unbound_release_work) rcu_read_lock &rq->__lock irq_context: 0 (wq_completion)events (work_completion)(&pwq->unbound_release_work) rcu_read_lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 &sb->s_type->i_mutex_key#10 sk_lock-AF_PHONET &rnp->exp_wq[0] irq_context: 0 pernet_ops_rwsem ipvs->est_mutex pcpu_alloc_mutex.wait_lock irq_context: 0 pernet_ops_rwsem ipvs->est_mutex &p->pi_lock irq_context: 0 sb_writers#3 &sb->s_type->i_mutex_key#9 tomoyo_ss &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 sb_writers#3 &sb->s_type->i_mutex_key#9 rcu_read_lock &rq->__lock irq_context: 0 sb_writers#3 &sb->s_type->i_mutex_key#9 rcu_read_lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 &pipe->mutex/1 remove_cache_srcu irq_context: 0 &pipe->mutex/1 remove_cache_srcu quarantine_lock irq_context: 0 &pipe->mutex/1 remove_cache_srcu &c->lock irq_context: 0 &pipe->mutex/1 remove_cache_srcu &n->list_lock irq_context: 0 &pipe->mutex/1 remove_cache_srcu rcu_read_lock pool_lock#2 irq_context: 0 &pipe->mutex/1 remove_cache_srcu &obj_hash[i].lock irq_context: 0 &pipe->mutex/1 remove_cache_srcu rcu_read_lock rcu_node_0 irq_context: 0 &pipe->mutex/1 remove_cache_srcu rcu_read_lock &rq->__lock irq_context: 0 &pipe->mutex/1 remove_cache_srcu rcu_read_lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 &pipe->mutex/1 rtnl_mutex irq_context: 0 &pipe->mutex/1 rtnl_mutex rtnl_mutex.wait_lock irq_context: 0 &pipe->mutex/1 rtnl_mutex &rq->__lock irq_context: 0 &pipe->mutex/1 rtnl_mutex &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 &pipe->mutex/1 rtnl_mutex (console_sem).lock irq_context: 0 &pipe->mutex/1 rtnl_mutex console_lock console_srcu console_owner_lock irq_context: 0 &pipe->mutex/1 rtnl_mutex console_lock console_srcu console_owner irq_context: 0 &pipe->mutex/1 rtnl_mutex console_lock console_srcu console_owner &port_lock_key irq_context: 0 &pipe->mutex/1 rtnl_mutex console_lock console_srcu console_owner console_owner_lock irq_context: 0 &pipe->mutex/1 rtnl_mutex fs_reclaim irq_context: 0 &pipe->mutex/1 rtnl_mutex fs_reclaim mmu_notifier_invalidate_range_start irq_context: 0 &pipe->mutex/1 rtnl_mutex stock_lock irq_context: 0 &pipe->mutex/1 rtnl_mutex pool_lock#2 irq_context: 0 &pipe->mutex/1 rtnl_mutex (inetaddr_validator_chain).rwsem irq_context: 0 &pipe->mutex/1 rtnl_mutex &obj_hash[i].lock irq_context: 0 &pipe->mutex/1 rtnl_mutex &base->lock irq_context: 0 &pipe->mutex/1 rtnl_mutex &base->lock &obj_hash[i].lock irq_context: 0 &pipe->mutex/1 rtnl_mutex rcu_read_lock &pool->lock irq_context: 0 &pipe->mutex/1 rtnl_mutex rcu_read_lock &pool->lock &obj_hash[i].lock irq_context: 0 &pipe->mutex/1 rtnl_mutex rcu_read_lock &pool->lock &p->pi_lock irq_context: 0 &pipe->mutex/1 rtnl_mutex rcu_read_lock &pool->lock &p->pi_lock &rq->__lock irq_context: 0 &pipe->mutex/1 rtnl_mutex rcu_read_lock &pool->lock &p->pi_lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 &pipe->mutex/1 rtnl_mutex nl_table_lock irq_context: 0 &pipe->mutex/1 rtnl_mutex nl_table_wait.lock irq_context: 0 &pipe->mutex/1 rtnl_mutex (inetaddr_chain).rwsem irq_context: 0 &pipe->mutex/1 rtnl_mutex (inetaddr_chain).rwsem fs_reclaim irq_context: 0 &pipe->mutex/1 rtnl_mutex (inetaddr_chain).rwsem fs_reclaim mmu_notifier_invalidate_range_start irq_context: 0 &pipe->mutex/1 rtnl_mutex (inetaddr_chain).rwsem pool_lock#2 irq_context: 0 &pipe->mutex/1 rtnl_mutex (inetaddr_chain).rwsem pcpu_alloc_mutex irq_context: 0 &pipe->mutex/1 rtnl_mutex (inetaddr_chain).rwsem pcpu_alloc_mutex pcpu_lock irq_context: 0 &pipe->mutex/1 rtnl_mutex (inetaddr_chain).rwsem &dir->lock#2 irq_context: 0 &pipe->mutex/1 rtnl_mutex (inetaddr_chain).rwsem fib_info_lock irq_context: 0 &pipe->mutex/1 rtnl_mutex (inetaddr_chain).rwsem stock_lock irq_context: 0 &pipe->mutex/1 rtnl_mutex (inetaddr_chain).rwsem &obj_hash[i].lock irq_context: 0 &pipe->mutex/1 rtnl_mutex (inetaddr_chain).rwsem rcu_read_lock pool_lock#2 irq_context: 0 &pipe->mutex/1 rtnl_mutex (inetaddr_chain).rwsem rcu_read_lock &data->fib_event_queue_lock irq_context: 0 &pipe->mutex/1 rtnl_mutex (inetaddr_chain).rwsem rcu_read_lock rcu_read_lock &pool->lock irq_context: 0 &pipe->mutex/1 rtnl_mutex (inetaddr_chain).rwsem rcu_read_lock rcu_read_lock &pool->lock &obj_hash[i].lock irq_context: 0 &pipe->mutex/1 rtnl_mutex (inetaddr_chain).rwsem rcu_read_lock rcu_read_lock &pool->lock pool_lock#2 irq_context: 0 &pipe->mutex/1 rtnl_mutex (inetaddr_chain).rwsem rcu_read_lock rcu_read_lock &pool->lock &p->pi_lock irq_context: 0 &pipe->mutex/1 rtnl_mutex (inetaddr_chain).rwsem rcu_read_lock rcu_read_lock &pool->lock &p->pi_lock &rq->__lock irq_context: 0 &pipe->mutex/1 rtnl_mutex (inetaddr_chain).rwsem rcu_read_lock rcu_read_lock &pool->lock &p->pi_lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 &pipe->mutex/1 rtnl_mutex (inetaddr_chain).rwsem rcu_read_lock rcu_node_0 irq_context: 0 &pipe->mutex/1 rtnl_mutex (inetaddr_chain).rwsem rcu_read_lock &rq->__lock irq_context: 0 &pipe->mutex/1 rtnl_mutex (inetaddr_chain).rwsem rcu_read_lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 &pipe->mutex/1 rtnl_mutex (inetaddr_chain).rwsem rcu_read_lock &rcu_state.gp_wq irq_context: 0 &pipe->mutex/1 rtnl_mutex (inetaddr_chain).rwsem rcu_read_lock &rcu_state.gp_wq &p->pi_lock irq_context: 0 &pipe->mutex/1 rtnl_mutex (inetaddr_chain).rwsem rcu_read_lock &rcu_state.gp_wq &p->pi_lock &rq->__lock irq_context: 0 &pipe->mutex/1 rtnl_mutex (inetaddr_chain).rwsem rcu_read_lock &rcu_state.gp_wq &p->pi_lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 &pipe->mutex/1 rtnl_mutex (inetaddr_chain).rwsem nl_table_lock irq_context: 0 &pipe->mutex/1 rtnl_mutex (inetaddr_chain).rwsem nl_table_wait.lock irq_context: 0 &pipe->mutex/1 rtnl_mutex (inetaddr_chain).rwsem rcu_read_lock &dir->lock#2 irq_context: 0 &pipe->mutex/1 rtnl_mutex (inetaddr_chain).rwsem &net->sctp.local_addr_lock irq_context: 0 &pipe->mutex/1 rtnl_mutex (inetaddr_chain).rwsem &net->sctp.local_addr_lock &net->sctp.addr_wq_lock irq_context: 0 &pipe->mutex/1 rtnl_mutex (inetaddr_chain).rwsem &net->sctp.local_addr_lock &net->sctp.addr_wq_lock pool_lock#2 irq_context: 0 &pipe->mutex/1 rtnl_mutex (inetaddr_chain).rwsem &net->sctp.local_addr_lock &net->sctp.addr_wq_lock &obj_hash[i].lock irq_context: 0 &pipe->mutex/1 rtnl_mutex (inetaddr_chain).rwsem &net->sctp.local_addr_lock &net->sctp.addr_wq_lock &base->lock irq_context: 0 &pipe->mutex/1 rtnl_mutex (inetaddr_chain).rwsem &net->sctp.local_addr_lock &net->sctp.addr_wq_lock &base->lock &obj_hash[i].lock irq_context: 0 &pipe->mutex/1 rtnl_mutex (inetaddr_chain).rwsem &rq->__lock irq_context: 0 &pipe->mutex/1 rtnl_mutex (inetaddr_chain).rwsem &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 &pipe->mutex/1 rtnl_mutex.wait_lock irq_context: 0 &pipe->mutex/1 &p->pi_lock irq_context: 0 &pipe->mutex/1 &p->pi_lock &rq->__lock irq_context: 0 &pipe->mutex/1 &p->pi_lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 &pipe->mutex/1 rlock-AF_NETLINK irq_context: 0 &pipe->mutex/1 purge_vmap_area_lock &obj_hash[i].lock irq_context: 0 &pipe->mutex/1 purge_vmap_area_lock pool_lock#2 irq_context: 0 &pipe->mutex/1 purge_vmap_area_lock &____s->seqcount irq_context: 0 &pipe->mutex/1 purge_vmap_area_lock rcu_read_lock pool_lock#2 irq_context: 0 rtnl_mutex (inetaddr_chain).rwsem rcu_read_lock &n->list_lock irq_context: 0 rtnl_mutex (inetaddr_chain).rwsem rcu_read_lock &n->list_lock &c->lock irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem rtnl_mutex dev_pm_qos_sysfs_mtx &rq->__lock irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem rtnl_mutex dev_pm_qos_sysfs_mtx &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 &type->i_mutex_dir_key#5 remove_cache_srcu pool_lock#2 irq_context: 0 pernet_ops_rwsem k-sk_lock-AF_RXRPC &rxnet->local_mutex &rq->__lock &cfs_rq->removed.lock irq_context: 0 &f->f_pos_lock sb_writers#4 &sb->s_type->i_mutex_key#8 jbd2_handle &ei->i_es_lock key#5 irq_context: 0 &f->f_pos_lock sb_writers#4 &sb->s_type->i_mutex_key#8 jbd2_handle &ei->i_data_sem &rq->__lock irq_context: 0 &f->f_pos_lock sb_writers#4 &sb->s_type->i_mutex_key#8 jbd2_handle &ei->i_data_sem &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 &iint->mutex mapping.invalidate_lock stock_lock irq_context: 0 &iint->mutex mapping.invalidate_lock &xa->xa_lock#9 stock_lock irq_context: 0 &iint->mutex mapping.invalidate_lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)wg-crypt-wg0#6 (work_completion)(&peer->transmit_packet_work) &peer->endpoint_lock rcu_read_lock_bh rcu_read_lock rcu_read_lock rcu_read_lock &dir->lock#2 irq_context: 0 (wq_completion)wg-crypt-wg0#6 (work_completion)(&peer->transmit_packet_work) &peer->endpoint_lock rcu_read_lock_bh rcu_read_lock rcu_read_lock &____s->seqcount#7 irq_context: 0 (wq_completion)wg-crypt-wg0#6 (work_completion)(&peer->transmit_packet_work) &peer->endpoint_lock rcu_read_lock_bh rcu_read_lock &nf_nat_locks[i] irq_context: 0 (wq_completion)wg-crypt-wg0#6 (work_completion)(&peer->transmit_packet_work) &peer->endpoint_lock rcu_read_lock_bh rcu_read_lock &nf_conntrack_locks[i] irq_context: 0 (wq_completion)wg-crypt-wg0#6 (work_completion)(&peer->transmit_packet_work) &peer->endpoint_lock rcu_read_lock_bh rcu_read_lock &nf_conntrack_locks[i] &nf_conntrack_locks[i]/1 irq_context: 0 (wq_completion)wg-crypt-wg0#6 (work_completion)(&peer->transmit_packet_work) &peer->endpoint_lock rcu_read_lock_bh rcu_read_lock &nf_conntrack_locks[i] &nf_conntrack_locks[i]/1 batched_entropy_u8.lock irq_context: 0 (wq_completion)wg-crypt-wg0#6 (work_completion)(&peer->transmit_packet_work) &peer->endpoint_lock rcu_read_lock_bh &c->lock irq_context: 0 (wq_completion)wg-crypt-wg0#6 (work_completion)(&peer->transmit_packet_work) &peer->endpoint_lock rcu_read_lock_bh rcu_read_lock &n->list_lock irq_context: 0 (wq_completion)wg-crypt-wg0#6 (work_completion)(&peer->transmit_packet_work) &peer->endpoint_lock rcu_read_lock_bh rcu_read_lock &n->list_lock &c->lock irq_context: 0 (wq_completion)wg-crypt-wg0#6 (work_completion)(&peer->transmit_packet_work) &peer->endpoint_lock rcu_read_lock_bh rcu_read_lock rcu_read_lock_bh &obj_hash[i].lock irq_context: 0 &iint->mutex mapping.invalidate_lock per_cpu_ptr(&cgroup_rstat_cpu_lock, cpu) irq_context: 0 &iint->mutex mapping.invalidate_lock rcu_read_lock rcu_read_lock rcu_node_0 irq_context: 0 &iint->mutex mapping.invalidate_lock rcu_read_lock rcu_read_lock &rq->__lock irq_context: 0 &iint->mutex mapping.invalidate_lock rcu_read_lock rcu_read_lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 &iint->mutex mapping.invalidate_lock rcu_read_lock &rcu_state.gp_wq irq_context: 0 &iint->mutex mapping.invalidate_lock rcu_read_lock &rcu_state.gp_wq &p->pi_lock irq_context: 0 &iint->mutex mapping.invalidate_lock rcu_read_lock &rcu_state.gp_wq &p->pi_lock &rq->__lock irq_context: 0 &iint->mutex mapping.invalidate_lock rcu_read_lock &rcu_state.gp_wq &p->pi_lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 &iint->mutex mapping.invalidate_lock rcu_read_lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 pernet_ops_rwsem devices_rwsem &p->pi_lock &cfs_rq->removed.lock irq_context: 0 pernet_ops_rwsem &root->kernfs_rwsem rcu_read_lock &rcu_state.expedited_wq irq_context: 0 pernet_ops_rwsem &root->kernfs_rwsem rcu_read_lock &rcu_state.expedited_wq &p->pi_lock irq_context: 0 pernet_ops_rwsem &root->kernfs_rwsem rcu_read_lock &rcu_state.expedited_wq &p->pi_lock &rq->__lock irq_context: 0 pernet_ops_rwsem &root->kernfs_rwsem rcu_read_lock &rcu_state.expedited_wq &p->pi_lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 pernet_ops_rwsem k-sk_lock-AF_TIPC &tn->nametbl_lock &obj_hash[i].lock pool_lock irq_context: 0 (wq_completion)wg-crypt-wg1#6 (work_completion)(&peer->transmit_packet_work) &peer->endpoint_lock rcu_read_lock_bh &obj_hash[i].lock irq_context: 0 (wq_completion)wg-crypt-wg0#6 (work_completion)(&peer->transmit_packet_work) &peer->endpoint_lock rcu_read_lock_bh rcu_read_lock &nf_conntrack_locks[i]/1 irq_context: 0 pernet_ops_rwsem k-sk_lock-AF_INET &rq->__lock irq_context: 0 pernet_ops_rwsem proc_inum_ida.xa_lock &obj_hash[i].lock irq_context: 0 pernet_ops_rwsem proc_inum_ida.xa_lock &____s->seqcount irq_context: 0 pernet_ops_rwsem proc_inum_ida.xa_lock rcu_read_lock pool_lock#2 irq_context: 0 (wq_completion)events free_ipc_work &rnp->exp_wq[3] irq_context: 0 (wq_completion)events free_ipc_work &pool->lock &p->pi_lock irq_context: 0 (wq_completion)events free_ipc_work &pool->lock &p->pi_lock &rq->__lock irq_context: 0 (wq_completion)events free_ipc_work &pool->lock &p->pi_lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)wg-crypt-wg0#6 (work_completion)(&peer->transmit_packet_work) &peer->endpoint_lock rcu_read_lock_bh batched_entropy_u8.lock irq_context: 0 (wq_completion)wg-crypt-wg0#6 (work_completion)(&peer->transmit_packet_work) &peer->endpoint_lock rcu_read_lock_bh kfence_freelist_lock irq_context: 0 (wq_completion)wg-crypt-wg0#6 (work_completion)(&peer->transmit_packet_work) &peer->endpoint_lock rcu_read_lock_bh rcu_read_lock rcu_read_lock_bh &meta->lock irq_context: 0 (wq_completion)wg-crypt-wg0#6 (work_completion)(&peer->transmit_packet_work) &peer->endpoint_lock rcu_read_lock_bh rcu_read_lock rcu_read_lock_bh kfence_freelist_lock irq_context: 0 &f->f_pos_lock sb_writers#4 &sb->s_type->i_mutex_key#8 rcu_read_lock &c->lock irq_context: 0 pernet_ops_rwsem uevent_sock_mutex rcu_read_lock &rcu_state.gp_wq &p->pi_lock &rq->__lock irq_context: 0 pernet_ops_rwsem uevent_sock_mutex rcu_read_lock &rcu_state.gp_wq &p->pi_lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 &xs->mutex batched_entropy_u8.lock irq_context: 0 &xs->mutex kfence_freelist_lock irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem rtnl_mutex cpu_hotplug_lock xps_map_mutex &cfs_rq->removed.lock irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem rtnl_mutex cpu_hotplug_lock xps_map_mutex &obj_hash[i].lock irq_context: 0 pernet_ops_rwsem devices_rwsem rdma_nets_rwsem rdma_nets_rwsem.wait_lock irq_context: 0 pernet_ops_rwsem ipvs->est_mutex &____s->seqcount irq_context: 0 &sb->s_type->i_mutex_key#10 &base->lock &obj_hash[i].lock irq_context: 0 sb_writers#4 &type->i_mutex_dir_key#3/1 tomoyo_ss remove_cache_srcu &pcp->lock &zone->lock irq_context: 0 sb_writers#4 &type->i_mutex_dir_key#3/1 tomoyo_ss remove_cache_srcu &pcp->lock &zone->lock &____s->seqcount irq_context: 0 &f->f_pos_lock sb_writers#4 &sb->s_type->i_mutex_key#8 rcu_read_lock rcu_read_lock rcu_read_lock rcu_node_0 irq_context: 0 &f->f_pos_lock sb_writers#4 &sb->s_type->i_mutex_key#8 rcu_read_lock rcu_read_lock rcu_read_lock &rq->__lock irq_context: 0 &f->f_pos_lock sb_writers#4 &sb->s_type->i_mutex_key#8 rcu_read_lock rcu_read_lock rcu_read_lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 pernet_ops_rwsem proc_inum_ida.xa_lock &c->lock irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem &device->compat_devs_mutex &lock->wait_lock irq_context: 0 pernet_ops_rwsem rtnl_mutex batched_entropy_u8.lock crngs.lock base_crng.lock irq_context: 0 &journal->j_barrier &journal->j_checkpoint_mutex rcu_read_lock &rcu_state.gp_wq irq_context: 0 &journal->j_barrier &journal->j_checkpoint_mutex rcu_read_lock &rcu_state.gp_wq &p->pi_lock irq_context: 0 &journal->j_barrier &journal->j_checkpoint_mutex rcu_read_lock &rcu_state.gp_wq &p->pi_lock &rq->__lock irq_context: 0 &journal->j_barrier &journal->j_checkpoint_mutex rcu_read_lock &rcu_state.gp_wq &p->pi_lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 pernet_ops_rwsem ovs_mutex nf_ct_proto_mutex nf_hook_mutex &____s->seqcount#2 irq_context: 0 pernet_ops_rwsem ovs_mutex nf_ct_proto_mutex nf_hook_mutex &____s->seqcount irq_context: 0 (wq_completion)wg-crypt-wg1#5 (work_completion)(&({ do { const void *__vpp_verify = (typeof((worker) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __asm__ ("" : "=r"(__ptr) : "0"((typeof(*((worker))) *)((worker)))); (typeof((typeof(*((worker))) *)((worker)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); })->work) rcu_node_0 irq_context: 0 &f->f_pos_lock sb_writers#4 &sb->s_type->i_mutex_key#8 &ei->i_data_sem &rcu_state.expedited_wq &p->pi_lock irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem &lock->wait_lock irq_context: 0 (wq_completion)wg-crypt-wg1#5 (work_completion)(&({ do { const void *__vpp_verify = (typeof((worker) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __asm__ ("" : "=r"(__ptr) : "0"((typeof(*((worker))) *)((worker)))); (typeof((typeof(*((worker))) *)((worker)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); })->work) &rcu_state.expedited_wq irq_context: 0 (wq_completion)wg-crypt-wg1#5 (work_completion)(&({ do { const void *__vpp_verify = (typeof((worker) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __asm__ ("" : "=r"(__ptr) : "0"((typeof(*((worker))) *)((worker)))); (typeof((typeof(*((worker))) *)((worker)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); })->work) &rcu_state.expedited_wq &p->pi_lock irq_context: 0 (wq_completion)wg-crypt-wg1#5 (work_completion)(&({ do { const void *__vpp_verify = (typeof((worker) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __asm__ ("" : "=r"(__ptr) : "0"((typeof(*((worker))) *)((worker)))); (typeof((typeof(*((worker))) *)((worker)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); })->work) &rcu_state.expedited_wq &p->pi_lock &rq->__lock irq_context: 0 (wq_completion)wg-crypt-wg1#5 (work_completion)(&({ do { const void *__vpp_verify = (typeof((worker) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __asm__ ("" : "=r"(__ptr) : "0"((typeof(*((worker))) *)((worker)))); (typeof((typeof(*((worker))) *)((worker)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); })->work) &rcu_state.expedited_wq &p->pi_lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 &f->f_pos_lock sb_writers#4 &sb->s_type->i_mutex_key#8 remove_cache_srcu rcu_read_lock &rcu_state.expedited_wq irq_context: 0 &f->f_pos_lock sb_writers#4 &sb->s_type->i_mutex_key#8 remove_cache_srcu rcu_read_lock &rcu_state.expedited_wq &p->pi_lock irq_context: 0 &f->f_pos_lock sb_writers#4 &sb->s_type->i_mutex_key#8 remove_cache_srcu rcu_read_lock &rcu_state.expedited_wq &p->pi_lock &rq->__lock irq_context: 0 &f->f_pos_lock sb_writers#4 &sb->s_type->i_mutex_key#8 remove_cache_srcu rcu_read_lock &rcu_state.expedited_wq &p->pi_lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem &root->kernfs_rwsem rcu_read_lock &rcu_state.expedited_wq irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem &root->kernfs_rwsem rcu_read_lock &rcu_state.expedited_wq &p->pi_lock irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem &root->kernfs_rwsem rcu_read_lock &rcu_state.expedited_wq &p->pi_lock &rq->__lock irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem &root->kernfs_rwsem rcu_read_lock &rcu_state.expedited_wq &p->pi_lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)wg-crypt-wg1 (work_completion)(&peer->transmit_packet_work) &peer->endpoint_lock rcu_read_lock_bh rcu_read_lock batched_entropy_u8.lock irq_context: 0 (wq_completion)wg-crypt-wg1 (work_completion)(&peer->transmit_packet_work) &peer->endpoint_lock rcu_read_lock_bh rcu_read_lock kfence_freelist_lock irq_context: 0 (wq_completion)wg-crypt-wg1#6 (work_completion)(&peer->transmit_packet_work) &peer->endpoint_lock rcu_read_lock_bh rcu_read_lock rcu_read_lock &____s->seqcount#7 irq_context: 0 (wq_completion)wg-crypt-wg1#6 (work_completion)(&peer->transmit_packet_work) &peer->endpoint_lock rcu_read_lock_bh rcu_read_lock &nf_nat_locks[i] irq_context: 0 (wq_completion)wg-crypt-wg1#6 (work_completion)(&peer->transmit_packet_work) &peer->endpoint_lock rcu_read_lock_bh rcu_read_lock &nf_conntrack_locks[i] irq_context: 0 (wq_completion)wg-crypt-wg1#6 (work_completion)(&peer->transmit_packet_work) &peer->endpoint_lock rcu_read_lock_bh rcu_read_lock &nf_conntrack_locks[i] &nf_conntrack_locks[i]/1 irq_context: 0 (wq_completion)wg-crypt-wg1#6 (work_completion)(&peer->transmit_packet_work) &peer->endpoint_lock rcu_read_lock_bh rcu_read_lock &nf_conntrack_locks[i] &nf_conntrack_locks[i]/1 batched_entropy_u8.lock irq_context: 0 (wq_completion)wg-crypt-wg1#6 (work_completion)(&peer->transmit_packet_work) &peer->endpoint_lock rcu_read_lock_bh &c->lock irq_context: 0 (wq_completion)wg-crypt-wg1#6 (work_completion)(&peer->transmit_packet_work) &peer->endpoint_lock rcu_read_lock_bh rcu_read_lock &____s->seqcount#9 irq_context: 0 (wq_completion)wg-crypt-wg1#6 (work_completion)(&peer->transmit_packet_work) &peer->endpoint_lock rcu_read_lock_bh rcu_read_lock rcu_read_lock_bh &obj_hash[i].lock irq_context: 0 (wq_completion)events free_ipc_work &rnp->exp_wq[2] irq_context: 0 pernet_ops_rwsem rcu_read_lock rcu_read_lock pool_lock#2 irq_context: 0 pernet_ops_rwsem rcu_read_lock &rcu_state.gp_wq &p->pi_lock &rq->__lock &cfs_rq->removed.lock irq_context: 0 loop_validate_mutex &lo->lo_mutex &obj_hash[i].lock irq_context: 0 loop_validate_mutex &lo->lo_mutex pool_lock#2 irq_context: 0 &pipe->mutex/1 rtnl_mutex rcu_node_0 irq_context: 0 sk_lock-AF_INET6 rcu_read_lock &ndev->lock irq_context: 0 sk_lock-AF_INET6 k-slock-AF_INET6 rcu_read_lock rcu_read_lock rcu_read_lock stock_lock irq_context: 0 sk_lock-AF_INET6 k-slock-AF_INET6 rcu_read_lock rcu_read_lock rcu_read_lock pool_lock#2 irq_context: 0 sk_lock-AF_INET6 k-slock-AF_INET6 rcu_read_lock rcu_read_lock rcu_read_lock &c->lock irq_context: 0 sk_lock-AF_INET6 k-slock-AF_INET6 rcu_read_lock rcu_read_lock rcu_read_lock &n->list_lock irq_context: 0 sk_lock-AF_INET6 k-slock-AF_INET6 rcu_read_lock rcu_read_lock rcu_read_lock &n->list_lock &c->lock irq_context: 0 sk_lock-AF_INET6 k-slock-AF_INET6 rcu_read_lock rcu_read_lock rcu_read_lock &dir->lock#2 irq_context: 0 sk_lock-AF_INET6 k-slock-AF_INET6 rcu_read_lock pool_lock#2 irq_context: 0 sk_lock-AF_INET6 k-slock-AF_INET6 rcu_read_lock rcu_read_lock pool_lock#2 irq_context: 0 sk_lock-AF_INET6 k-slock-AF_INET6 rcu_read_lock rcu_read_lock &obj_hash[i].lock irq_context: 0 sk_lock-AF_INET6 k-slock-AF_INET6 rcu_read_lock rcu_read_lock &dir->lock#2 irq_context: 0 sk_lock-AF_INET6 k-slock-AF_INET6 rcu_read_lock rcu_read_lock &tbl->lock irq_context: 0 sk_lock-AF_INET6 k-slock-AF_INET6 rcu_read_lock rcu_read_lock &n->lock irq_context: 0 sk_lock-AF_INET6 k-slock-AF_INET6 rcu_read_lock rcu_read_lock &____s->seqcount#9 irq_context: 0 sk_lock-AF_INET6 k-slock-AF_INET6 rcu_read_lock rcu_read_lock rcu_read_lock_bh tk_core.seq.seqcount irq_context: 0 sk_lock-AF_INET6 k-slock-AF_INET6 rcu_read_lock rcu_read_lock rcu_read_lock_bh &list->lock#5 irq_context: softirq (&q->timer) rcu_read_lock &q->lock#2 rcu_read_lock rcu_read_lock rcu_read_lock &pool->lock &p->pi_lock &rq->__lock &cfs_rq->removed.lock irq_context: 0 pernet_ops_rwsem k-sk_lock-AF_INET6 &tcp_hashinfo.bhash[i].lock &c->lock irq_context: 0 sk_lock-AF_INET batched_entropy_u16.lock crngs.lock irq_context: 0 sk_lock-AF_INET6 k-slock-AF_INET6 rcu_read_lock &c->lock irq_context: 0 sk_lock-AF_INET6 k-slock-AF_INET6 rcu_read_lock rcu_read_lock &____s->seqcount#10 irq_context: 0 &f->f_pos_lock sb_writers#4 &sb->s_type->i_mutex_key#8 jbd2_handle &ei->i_data_sem &lg->lg_mutex pool_lock#2 irq_context: 0 slock-AF_INET &c->lock irq_context: 0 slock-AF_INET pool_lock#2 irq_context: 0 sb_writers#4 &type->i_mutex_dir_key#3/1 sb_internal jbd2_handle mmu_notifier_invalidate_range_start irq_context: 0 sb_writers#4 &type->i_mutex_dir_key#3/1 sb_internal jbd2_handle &____s->seqcount irq_context: 0 sb_writers#4 &type->i_mutex_dir_key#3/1 sb_internal jbd2_handle rcu_read_lock pool_lock#2 irq_context: 0 sb_writers#4 &sb->s_type->i_mutex_key#8 fs_reclaim mmu_notifier_invalidate_range_start &rq->__lock irq_context: 0 sb_writers#4 &sb->s_type->i_mutex_key#8 fs_reclaim mmu_notifier_invalidate_range_start &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: softirq (&icsk->icsk_retransmit_timer) slock-AF_INET batched_entropy_u32.lock crngs.lock irq_context: 0 sb_writers#4 jbd2_handle &cfs_rq->removed.lock irq_context: 0 sb_writers#4 jbd2_handle &obj_hash[i].lock irq_context: 0 &pipe->mutex/1 rcu_read_lock (console_sem).lock irq_context: 0 &pipe->mutex/1 rcu_read_lock console_lock console_srcu console_owner_lock irq_context: 0 &pipe->mutex/1 rcu_read_lock console_lock console_srcu console_owner irq_context: 0 &pipe->mutex/1 rcu_read_lock console_lock console_srcu console_owner &port_lock_key irq_context: 0 &pipe->mutex/1 rcu_read_lock console_lock console_srcu console_owner console_owner_lock irq_context: 0 sb_writers#4 &sb->s_type->i_mutex_key#8 &journal->j_state_lock &journal->j_wait_commit &p->pi_lock &rq->__lock &cfs_rq->removed.lock irq_context: 0 &pipe->mutex/1 nfnl_subsys_ctnetlink rcu_read_lock &ct->lock (console_sem).lock irq_context: 0 &pipe->mutex/1 nfnl_subsys_ctnetlink rcu_read_lock &ct->lock console_lock console_srcu console_owner_lock irq_context: 0 &pipe->mutex/1 nfnl_subsys_ctnetlink rcu_read_lock &ct->lock console_lock console_srcu console_owner irq_context: 0 &pipe->mutex/1 nfnl_subsys_ctnetlink rcu_read_lock &ct->lock console_lock console_srcu console_owner &port_lock_key irq_context: 0 &pipe->mutex/1 nfnl_subsys_ctnetlink rcu_read_lock &ct->lock console_lock console_srcu console_owner console_owner_lock irq_context: 0 &pipe->mutex/1 nfnl_subsys_ctnetlink &obj_hash[i].lock irq_context: 0 (wq_completion)kblockd (work_completion)(&(&q->requeue_work)->work) &rq->__lock irq_context: 0 (wq_completion)kblockd (work_completion)(&(&q->requeue_work)->work) &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)kblockd (work_completion)(&(&q->requeue_work)->work) rcu_read_lock fill_pool_map-wait-type-override rcu_node_0 irq_context: 0 (wq_completion)kblockd (work_completion)(&(&q->requeue_work)->work) rcu_read_lock fill_pool_map-wait-type-override &rq->__lock irq_context: 0 (wq_completion)kblockd (work_completion)(&(&q->requeue_work)->work) rcu_read_lock fill_pool_map-wait-type-override &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem proc_inum_ida.xa_lock &obj_hash[i].lock irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem proc_inum_ida.xa_lock pool_lock#2 irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem kernfs_idr_lock &obj_hash[i].lock pool_lock irq_context: 0 (wq_completion)wg-kex-wg2#11 (work_completion)(&peer->transmit_handshake_work) &peer->endpoint_lock rcu_read_lock_bh &obj_hash[i].lock irq_context: 0 pernet_ops_rwsem ipvs->est_mutex &rq->__lock irq_context: 0 pernet_ops_rwsem ipvs->est_mutex &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 sb_writers#4 &sb->s_type->i_mutex_key#8 &sbi->s_writepages_rwsem remove_cache_srcu pool_lock#2 irq_context: 0 sb_writers#4 &sb->s_type->i_mutex_key#8 &sbi->s_writepages_rwsem remove_cache_srcu rcu_read_lock rcu_node_0 irq_context: 0 sb_writers#4 &sb->s_type->i_mutex_key#8 &sbi->s_writepages_rwsem remove_cache_srcu rcu_read_lock &rq->__lock irq_context: 0 sb_writers#4 &sb->s_type->i_mutex_key#8 &sbi->s_writepages_rwsem remove_cache_srcu rcu_read_lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 sb_writers#4 &journal->j_barrier &journal->j_checkpoint_mutex rcu_read_lock &c->lock irq_context: 0 &journal->j_barrier &journal->j_checkpoint_mutex &cfs_rq->removed.lock irq_context: 0 sb_writers#4 &type->i_mutex_dir_key#3 jbd2_handle &c->lock irq_context: 0 sb_writers#4 &sb->s_type->i_mutex_key#8 mapping.invalidate_lock jbd2_handle rcu_read_lock &rcu_state.gp_wq irq_context: 0 sb_writers#4 &sb->s_type->i_mutex_key#8 mapping.invalidate_lock jbd2_handle rcu_read_lock &rcu_state.gp_wq &p->pi_lock irq_context: 0 sb_writers#4 &sb->s_type->i_mutex_key#8 mapping.invalidate_lock jbd2_handle rcu_read_lock &rcu_state.gp_wq &p->pi_lock &rq->__lock irq_context: 0 sb_writers#4 &sb->s_type->i_mutex_key#8 mapping.invalidate_lock jbd2_handle rcu_read_lock &rcu_state.gp_wq &p->pi_lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 slock-AF_INET &____s->seqcount#2 irq_context: 0 slock-AF_INET &____s->seqcount irq_context: 0 (wq_completion)wg-crypt-wg0#3 (work_completion)(&({ do { const void *__vpp_verify = (typeof((worker) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __asm__ ("" : "=r"(__ptr) : "0"((typeof(*((worker))) *)((worker)))); (typeof((typeof(*((worker))) *)((worker)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); })->work) &cfs_rq->removed.lock irq_context: 0 (wq_completion)wg-crypt-wg0#3 (work_completion)(&({ do { const void *__vpp_verify = (typeof((worker) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __asm__ ("" : "=r"(__ptr) : "0"((typeof(*((worker))) *)((worker)))); (typeof((typeof(*((worker))) *)((worker)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); })->work) &obj_hash[i].lock irq_context: 0 pernet_ops_rwsem rtnl_mutex dev_hotplug_mutex &rq->__lock irq_context: 0 pernet_ops_rwsem rtnl_mutex dev_hotplug_mutex &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)wg-crypt-wg2#6 (work_completion)(&peer->transmit_packet_work) &peer->endpoint_lock rcu_read_lock_bh rcu_read_lock rcu_read_lock &____s->seqcount#7 irq_context: 0 (wq_completion)wg-crypt-wg2#6 (work_completion)(&peer->transmit_packet_work) &peer->endpoint_lock rcu_read_lock_bh rcu_read_lock &nf_nat_locks[i] irq_context: 0 (wq_completion)wg-crypt-wg2#6 (work_completion)(&peer->transmit_packet_work) &peer->endpoint_lock rcu_read_lock_bh rcu_read_lock &nf_conntrack_locks[i] irq_context: 0 (wq_completion)wg-crypt-wg2#6 (work_completion)(&peer->transmit_packet_work) &peer->endpoint_lock rcu_read_lock_bh rcu_read_lock &nf_conntrack_locks[i] &nf_conntrack_locks[i]/1 irq_context: 0 (wq_completion)wg-crypt-wg2#6 (work_completion)(&peer->transmit_packet_work) &peer->endpoint_lock rcu_read_lock_bh rcu_read_lock &nf_conntrack_locks[i] &nf_conntrack_locks[i]/1 batched_entropy_u8.lock irq_context: 0 (wq_completion)wg-crypt-wg2#6 (work_completion)(&peer->transmit_packet_work) &peer->endpoint_lock rcu_read_lock_bh rcu_read_lock &nf_conntrack_locks[i]/1 irq_context: 0 (wq_completion)wg-crypt-wg2#6 (work_completion)(&peer->transmit_packet_work) &peer->endpoint_lock rcu_read_lock_bh rcu_read_lock &____s->seqcount#9 irq_context: 0 (wq_completion)wg-crypt-wg2#6 (work_completion)(&peer->transmit_packet_work) &peer->endpoint_lock rcu_read_lock_bh rcu_read_lock rcu_read_lock_bh &obj_hash[i].lock irq_context: 0 (wq_completion)wg-crypt-wg2#6 (work_completion)(&peer->transmit_packet_work) rcu_read_lock_bh rcu_read_lock &pool->lock/1 &p->pi_lock &cfs_rq->removed.lock irq_context: 0 (wq_completion)wg-kex-wg2#11 (work_completion)(&peer->transmit_handshake_work) &peer->endpoint_lock rcu_read_lock_bh rcu_read_lock &____s->seqcount#9 irq_context: 0 (wq_completion)wg-kex-wg2#11 (work_completion)(&peer->transmit_handshake_work) &peer->endpoint_lock rcu_read_lock_bh rcu_read_lock rcu_read_lock_bh &obj_hash[i].lock irq_context: 0 sb_writers#4 sb_writers#4 remove_cache_srcu rcu_read_lock rcu_node_0 irq_context: 0 sb_writers#4 sb_writers#4 remove_cache_srcu rcu_read_lock &rq->__lock irq_context: 0 sb_writers#4 sb_writers#4 remove_cache_srcu rcu_read_lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 slock-AF_INET hrtimer_bases.lock tk_core.seq.seqcount irq_context: 0 sk_lock-AF_INET &sighand->siglock &c->lock irq_context: 0 &pipe->mutex/1 nfnl_subsys_ctnetlink rcu_read_lock rcu_read_lock rcu_node_0 irq_context: 0 &pipe->mutex/1 nfnl_subsys_ctnetlink rcu_read_lock rcu_read_lock &rq->__lock irq_context: 0 &pipe->mutex/1 nfnl_subsys_ctnetlink rcu_read_lock rcu_read_lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 &pipe->mutex/1 nfnl_subsys_ctnetlink rcu_read_lock &rq->__lock &cfs_rq->removed.lock irq_context: 0 (wq_completion)mm_percpu_wq (work_completion)(&(({ do { const void *__vpp_verify = (typeof((&vmstat_work) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __asm__ ("" : "=r"(__ptr) : "0"((typeof(*((&vmstat_work))) *)((&vmstat_work)))); (typeof((typeof(*((&vmstat_work))) *)((&vmstat_work)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); }))->work) &rq->__lock irq_context: 0 (wq_completion)mm_percpu_wq (work_completion)(&(({ do { const void *__vpp_verify = (typeof((&vmstat_work) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __asm__ ("" : "=r"(__ptr) : "0"((typeof(*((&vmstat_work))) *)((&vmstat_work)))); (typeof((typeof(*((&vmstat_work))) *)((&vmstat_work)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); }))->work) &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 &sb->s_type->i_mutex_key#10 sk_lock-AF_INET slock-AF_INET quarantine_lock irq_context: 0 rtnl_mutex &idev->mc_lock remove_cache_srcu irq_context: 0 rtnl_mutex &idev->mc_lock remove_cache_srcu quarantine_lock irq_context: 0 rtnl_mutex &idev->mc_lock remove_cache_srcu &c->lock irq_context: 0 rtnl_mutex &idev->mc_lock remove_cache_srcu &n->list_lock irq_context: 0 rtnl_mutex &idev->mc_lock remove_cache_srcu rcu_read_lock pool_lock#2 irq_context: 0 rtnl_mutex &idev->mc_lock remove_cache_srcu &obj_hash[i].lock irq_context: 0 rtnl_mutex &idev->mc_lock remove_cache_srcu pool_lock#2 irq_context: 0 cb_lock &rdev->wiphy.mtx remove_cache_srcu &pcp->lock &zone->lock irq_context: 0 cb_lock &rdev->wiphy.mtx remove_cache_srcu &pcp->lock &zone->lock &____s->seqcount irq_context: 0 &sb->s_type->i_mutex_key#8 &sbi->s_writepages_rwsem batched_entropy_u8.lock irq_context: 0 &sb->s_type->i_mutex_key#8 &sbi->s_writepages_rwsem kfence_freelist_lock irq_context: 0 sb_writers#4 sb_internal &pcp->lock &zone->lock &____s->seqcount irq_context: 0 rtnl_mutex &tbl->lock irq_context: 0 &sb->s_type->i_mutex_key#8 rcu_read_lock &stopper->lock irq_context: 0 &sb->s_type->i_mutex_key#8 rcu_read_lock &stop_pi_lock irq_context: 0 &sb->s_type->i_mutex_key#8 rcu_read_lock &stop_pi_lock &rq->__lock irq_context: 0 &sb->s_type->i_mutex_key#8 rcu_read_lock &stop_pi_lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: softirq (&ifibss->timer) irq_context: softirq (&ifibss->timer) &rdev->wiphy_work_lock irq_context: softirq (&ifibss->timer) rcu_read_lock &pool->lock irq_context: softirq (&ifibss->timer) rcu_read_lock &pool->lock &obj_hash[i].lock irq_context: softirq (&ifibss->timer) rcu_read_lock &pool->lock &p->pi_lock irq_context: 0 (wq_completion)bat_events (work_completion)(&(&forw_packet_aggr->delayed_work)->work) &hard_iface->bat_iv.ogm_buff_mutex &bat_priv->tt.commit_lock key#16 &bat_priv->softif_vlan_list_lock irq_context: 0 (wq_completion)bat_events (work_completion)(&(&forw_packet_aggr->delayed_work)->work) &hard_iface->bat_iv.ogm_buff_mutex &bat_priv->tt.commit_lock key#16 &obj_hash[i].lock irq_context: 0 (wq_completion)bat_events (work_completion)(&(&forw_packet_aggr->delayed_work)->work) &hard_iface->bat_iv.ogm_buff_mutex &bat_priv->tt.commit_lock key#16 pool_lock#2 irq_context: 0 (wq_completion)bat_events (work_completion)(&(&forw_packet_aggr->delayed_work)->work) &hard_iface->bat_iv.ogm_buff_mutex &bat_priv->tt.commit_lock key#16 krc.lock irq_context: 0 (wq_completion)bat_events (work_completion)(&(&forw_packet_aggr->delayed_work)->work) &hard_iface->bat_iv.ogm_buff_mutex &bat_priv->tt.commit_lock &bat_priv->tt.last_changeset_lock &obj_hash[i].lock irq_context: 0 &sb->s_type->i_mutex_key#10 (netlink_chain).rwsem fill_pool_map-wait-type-override &n->list_lock irq_context: 0 &sb->s_type->i_mutex_key#10 (netlink_chain).rwsem fill_pool_map-wait-type-override &n->list_lock &c->lock irq_context: 0 cb_lock genl_mutex rcu_node_0 irq_context: 0 sk_lock-AF_ALG &mm->mmap_lock sb_pagefaults jbd2_handle rcu_read_lock rcu_node_0 irq_context: 0 sk_lock-AF_ALG &mm->mmap_lock sb_pagefaults jbd2_handle rcu_read_lock &rq->__lock irq_context: 0 sk_lock-AF_ALG &mm->mmap_lock sb_pagefaults jbd2_handle rcu_read_lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 sk_lock-AF_LLC irq_context: 0 sk_lock-AF_LLC &rq->__lock irq_context: 0 sk_lock-AF_LLC &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 sk_lock-AF_LLC slock-AF_LLC irq_context: 0 sk_lock-AF_LLC llc_sap_list_lock irq_context: 0 sk_lock-AF_LLC llc_sap_list_lock &c->lock irq_context: 0 sk_lock-AF_LLC llc_sap_list_lock &n->list_lock irq_context: 0 sk_lock-AF_LLC llc_sap_list_lock &n->list_lock &c->lock irq_context: 0 sk_lock-AF_LLC llc_sap_list_lock pool_lock#2 irq_context: 0 sk_lock-AF_LLC fs_reclaim irq_context: 0 sk_lock-AF_LLC fs_reclaim mmu_notifier_invalidate_range_start irq_context: 0 sk_lock-AF_LLC &c->lock irq_context: 0 sk_lock-AF_LLC pool_lock#2 irq_context: 0 sk_lock-AF_LLC &dir->lock#2 irq_context: 0 sk_lock-AF_LLC &sap->sk_lock irq_context: 0 sk_lock-AF_LLC wlock-AF_LLC irq_context: 0 sk_lock-AF_LLC rcu_read_lock_bh noop_qdisc.q.lock irq_context: 0 sk_lock-AF_LLC rcu_read_lock_bh &obj_hash[i].lock irq_context: 0 sk_lock-AF_LLC rcu_read_lock_bh pool_lock#2 irq_context: 0 sk_lock-AF_LLC &obj_hash[i].lock irq_context: 0 (wq_completion)wg-kex-wg2#11 (work_completion)(&peer->transmit_handshake_work) &wg->static_identity.lock &handshake->lock &rq->__lock irq_context: 0 (wq_completion)wg-kex-wg2#11 (work_completion)(&peer->transmit_handshake_work) &wg->static_identity.lock &handshake->lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 sk_lock-AF_LLC &base->lock irq_context: 0 sk_lock-AF_LLC &base->lock &obj_hash[i].lock irq_context: 0 slock-AF_ALG &sk->sk_lock.wq &p->pi_lock &cfs_rq->removed.lock irq_context: 0 sk_lock-AF_LLC &ei->socket.wq.wait irq_context: 0 slock-AF_LLC irq_context: 0 sk_lock-AF_ALG &mm->mmap_lock sb_pagefaults mapping.invalidate_lock rcu_read_lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 sk_lock-AF_LLC rcu_read_lock &ei->socket.wq.wait irq_context: 0 sk_lock-AF_LLC rcu_read_lock &ei->socket.wq.wait &p->pi_lock irq_context: 0 sk_lock-AF_LLC rcu_read_lock &ei->socket.wq.wait &p->pi_lock &rq->__lock irq_context: 0 sk_lock-AF_LLC rcu_read_lock &ei->socket.wq.wait &p->pi_lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 sk_lock-AF_LLC rcu_read_lock rcu_node_0 irq_context: 0 sk_lock-AF_LLC rcu_read_lock &rq->__lock irq_context: 0 sk_lock-AF_LLC rcu_read_lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 sk_lock-AF_LLC slock-AF_LLC &sk->sk_lock.wq irq_context: 0 slock-AF_LLC &sk->sk_lock.wq irq_context: 0 slock-AF_LLC &sk->sk_lock.wq &p->pi_lock irq_context: 0 slock-AF_LLC &sk->sk_lock.wq &p->pi_lock &rq->__lock irq_context: 0 slock-AF_LLC &sk->sk_lock.wq &p->pi_lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 sk_lock-AF_LLC quarantine_lock irq_context: 0 sk_lock-AF_LLC rcu_read_lock &ei->socket.wq.wait &p->pi_lock &cfs_rq->removed.lock irq_context: 0 sk_lock-AF_LLC &cfs_rq->removed.lock irq_context: 0 sb_writers#4 &base->lock irq_context: 0 sb_writers#4 &base->lock &obj_hash[i].lock irq_context: 0 &tn->lock irq_context: 0 xt_led_mutex irq_context: 0 xt_led_mutex fs_reclaim irq_context: 0 xt_led_mutex fs_reclaim mmu_notifier_invalidate_range_start irq_context: 0 xt_led_mutex fs_reclaim mmu_notifier_invalidate_range_start &rq->__lock irq_context: 0 xt_led_mutex fs_reclaim mmu_notifier_invalidate_range_start &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 xt_led_mutex pool_lock#2 irq_context: 0 xt_led_mutex triggers_list_lock irq_context: 0 xt_led_mutex leds_list_lock irq_context: 0 xt_led_mutex leds_list_lock &led_cdev->trigger_lock irq_context: 0 xt_led_mutex &obj_hash[i].lock irq_context: 0 cpu_hotplug_lock jump_label_mutex text_mutex &rq->__lock irq_context: 0 (wq_completion)events (work_completion)(&aux->work) pack_mutex text_mutex &rq->__lock &cfs_rq->removed.lock irq_context: 0 cpu_hotplug_lock jump_label_mutex text_mutex.wait_lock irq_context: 0 cpu_hotplug_lock jump_label_mutex &p->pi_lock irq_context: 0 cpu_hotplug_lock jump_label_mutex &p->pi_lock &rq->__lock irq_context: 0 cpu_hotplug_lock jump_label_mutex &p->pi_lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 cpu_hotplug_lock jump_label_mutex &rq->__lock irq_context: 0 cpu_hotplug_lock jump_label_mutex &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 xt_led_mutex (&ledinternal->timer) irq_context: 0 xt_led_mutex &base->lock irq_context: 0 sk_lock-AF_LLC rcu_read_lock &rcu_state.gp_wq irq_context: 0 sk_lock-AF_LLC rcu_read_lock &rcu_state.gp_wq &p->pi_lock irq_context: 0 sk_lock-AF_LLC rcu_read_lock &rcu_state.gp_wq &p->pi_lock &rq->__lock irq_context: 0 sk_lock-AF_LLC rcu_read_lock &rcu_state.gp_wq &p->pi_lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 sk_lock-AF_LLC &rq->__lock &cfs_rq->removed.lock irq_context: 0 sk_lock-AF_LLC &meta->lock irq_context: 0 sk_lock-AF_LLC kfence_freelist_lock irq_context: 0 sb_writers#4 &sb->s_type->i_mutex_key#8 mapping.invalidate_lock stock_lock rcu_read_lock per_cpu_ptr(&cgroup_rstat_cpu_lock, cpu) irq_context: 0 sb_writers#4 &sb->s_type->i_mutex_key#8 mapping.invalidate_lock per_cpu_ptr(&cgroup_rstat_cpu_lock, cpu) irq_context: 0 &sb->s_type->i_mutex_key#10 sk_lock-AF_LLC &sap->sk_lock irq_context: 0 &sb->s_type->i_mutex_key#10 &dir->lock#2 irq_context: 0 xt_led_mutex &c->lock irq_context: 0 xt_led_mutex &rq->__lock irq_context: 0 xt_led_mutex &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 rtnl_mutex &dev_addr_list_lock_key &n->list_lock irq_context: 0 rtnl_mutex &dev_addr_list_lock_key &n->list_lock &c->lock irq_context: 0 rtnl_mutex &dev_addr_list_lock_key &obj_hash[i].lock irq_context: 0 rtnl_mutex &dev_addr_list_lock_key krc.lock irq_context: 0 rtnl_mutex &idev->mc_lock &dev_addr_list_lock_key &obj_hash[i].lock irq_context: 0 rtnl_mutex &idev->mc_lock &dev_addr_list_lock_key krc.lock irq_context: 0 rtnl_mutex &dev->tx_global_lock &qdisc_xmit_lock_key#3 irq_context: 0 rtnl_mutex (work_completion)(&(&bond->mii_work)->work) irq_context: 0 rtnl_mutex (work_completion)(&(&bond->mii_work)->work) &rq->__lock irq_context: 0 rtnl_mutex (work_completion)(&(&bond->mii_work)->work) &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 rtnl_mutex (work_completion)(&(&bond->arp_work)->work) irq_context: 0 rtnl_mutex (work_completion)(&(&bond->arp_work)->work) &rq->__lock irq_context: 0 rtnl_mutex (work_completion)(&(&bond->arp_work)->work) &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 rtnl_mutex (work_completion)(&(&bond->alb_work)->work) irq_context: 0 rtnl_mutex (work_completion)(&(&bond->ad_work)->work) irq_context: 0 rtnl_mutex (work_completion)(&(&bond->mcast_work)->work) irq_context: 0 rtnl_mutex (work_completion)(&(&bond->slave_arr_work)->work) irq_context: 0 &sb->s_type->i_mutex_key#8 fs_reclaim mmu_notifier_invalidate_range_start &rq->__lock irq_context: 0 &sb->s_type->i_mutex_key#8 fs_reclaim mmu_notifier_invalidate_range_start &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 sk_lock-AF_INET6 rcu_read_lock rcu_read_lock rcu_read_lock_bh noop_qdisc.q.lock irq_context: 0 rtnl_mutex kernfs_idr_lock &obj_hash[i].lock pool_lock irq_context: 0 napi_hash_lock irq_context: 0 dup_mmap_sem &mm->mmap_lock &mm->mmap_lock/1 batched_entropy_u8.lock crngs.lock irq_context: 0 &sbi->s_writepages_rwsem jbd2_handle &folio_wait_table[i] irq_context: 0 &sbi->s_writepages_rwsem jbd2_handle &folio_wait_table[i] &p->pi_lock irq_context: 0 &sbi->s_writepages_rwsem jbd2_handle &folio_wait_table[i] &p->pi_lock &rq->__lock irq_context: 0 &sbi->s_writepages_rwsem jbd2_handle &folio_wait_table[i] &p->pi_lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 sb_writers#4 &type->i_mutex_dir_key#3 &pcp->lock &zone->lock irq_context: 0 &f->f_pos_lock &rq->__lock &cfs_rq->removed.lock irq_context: 0 &f->f_pos_lock sb_writers#4 &sb->s_type->i_mutex_key#8 &ei->i_data_sem &ei->i_es_lock &____s->seqcount#2 irq_context: 0 &f->f_pos_lock sb_writers#4 &sb->s_type->i_mutex_key#8 &ei->i_data_sem &ei->i_es_lock &____s->seqcount irq_context: 0 &f->f_pos_lock sb_writers#4 &sb->s_type->i_mutex_key#8 &ei->i_data_sem &ei->i_es_lock pool_lock#2 irq_context: 0 &sb->s_type->i_mutex_key#8 &pcp->lock &zone->lock irq_context: 0 free_vmap_area_lock &meta->lock irq_context: 0 free_vmap_area_lock kfence_freelist_lock irq_context: 0 kn->active#59 &kernfs_locks->open_file_mutex[count] &c->lock irq_context: 0 kn->active#59 &kernfs_locks->open_file_mutex[count] &n->list_lock irq_context: 0 kn->active#59 &kernfs_locks->open_file_mutex[count] &n->list_lock &c->lock irq_context: 0 kn->active#59 &kernfs_locks->open_file_mutex[count] &rq->__lock irq_context: 0 kn->active#59 &kernfs_locks->open_file_mutex[count] &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 &f->f_pos_lock sb_writers#4 &sb->s_type->i_mutex_key#8 &sbi->s_writepages_rwsem irq_context: 0 &f->f_pos_lock sb_writers#4 &sb->s_type->i_mutex_key#8 &sbi->s_writepages_rwsem &xa->xa_lock#9 irq_context: 0 &f->f_pos_lock sb_writers#4 &sb->s_type->i_mutex_key#8 &sbi->s_writepages_rwsem mmu_notifier_invalidate_range_start irq_context: 0 &f->f_pos_lock sb_writers#4 &sb->s_type->i_mutex_key#8 &sbi->s_writepages_rwsem pool_lock#2 irq_context: 0 &f->f_pos_lock sb_writers#4 &sb->s_type->i_mutex_key#8 &sbi->s_writepages_rwsem lock#4 irq_context: 0 &f->f_pos_lock sb_writers#4 &sb->s_type->i_mutex_key#8 &sbi->s_writepages_rwsem lock#4 &lruvec->lru_lock irq_context: 0 &f->f_pos_lock sb_writers#4 &sb->s_type->i_mutex_key#8 &sbi->s_writepages_rwsem lock#5 irq_context: 0 &f->f_pos_lock sb_writers#4 &sb->s_type->i_mutex_key#8 &sbi->s_writepages_rwsem &obj_hash[i].lock irq_context: 0 &f->f_pos_lock sb_writers#4 &sb->s_type->i_mutex_key#8 &sbi->s_writepages_rwsem &journal->j_state_lock irq_context: 0 &f->f_pos_lock sb_writers#4 &sb->s_type->i_mutex_key#8 &sbi->s_writepages_rwsem jbd2_handle irq_context: 0 &f->f_pos_lock sb_writers#4 &sb->s_type->i_mutex_key#8 &sbi->s_writepages_rwsem jbd2_handle lock#4 irq_context: 0 &f->f_pos_lock sb_writers#4 &sb->s_type->i_mutex_key#8 &sbi->s_writepages_rwsem jbd2_handle lock#5 irq_context: 0 &f->f_pos_lock sb_writers#4 &sb->s_type->i_mutex_key#8 &sbi->s_writepages_rwsem jbd2_handle mmu_notifier_invalidate_range_start irq_context: 0 &f->f_pos_lock sb_writers#4 &sb->s_type->i_mutex_key#8 &sbi->s_writepages_rwsem jbd2_handle pool_lock#2 irq_context: 0 &f->f_pos_lock sb_writers#4 &sb->s_type->i_mutex_key#8 &sbi->s_writepages_rwsem jbd2_handle &ei->i_es_lock irq_context: 0 &f->f_pos_lock sb_writers#4 &sb->s_type->i_mutex_key#8 &sbi->s_writepages_rwsem jbd2_handle &ei->i_data_sem irq_context: 0 &f->f_pos_lock sb_writers#4 &sb->s_type->i_mutex_key#8 &sbi->s_writepages_rwsem jbd2_handle &ei->i_data_sem mmu_notifier_invalidate_range_start irq_context: 0 &f->f_pos_lock sb_writers#4 &sb->s_type->i_mutex_key#8 &sbi->s_writepages_rwsem jbd2_handle &ei->i_data_sem pool_lock#2 irq_context: 0 &f->f_pos_lock sb_writers#4 &sb->s_type->i_mutex_key#8 &sbi->s_writepages_rwsem jbd2_handle &ei->i_data_sem &rq->__lock irq_context: 0 &f->f_pos_lock sb_writers#4 &sb->s_type->i_mutex_key#8 &sbi->s_writepages_rwsem jbd2_handle &ei->i_data_sem &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 &f->f_pos_lock sb_writers#4 &sb->s_type->i_mutex_key#8 &sbi->s_writepages_rwsem jbd2_handle &ei->i_data_sem &lg->lg_mutex irq_context: 0 &f->f_pos_lock sb_writers#4 &sb->s_type->i_mutex_key#8 &sbi->s_writepages_rwsem jbd2_handle &ei->i_data_sem &lg->lg_mutex &ei->i_prealloc_lock irq_context: 0 &f->f_pos_lock sb_writers#4 &sb->s_type->i_mutex_key#8 &sbi->s_writepages_rwsem jbd2_handle &ei->i_data_sem &lg->lg_mutex rcu_read_lock &pa->pa_lock irq_context: 0 &f->f_pos_lock sb_writers#4 &sb->s_type->i_mutex_key#8 &sbi->s_writepages_rwsem jbd2_handle &ei->i_data_sem &lg->lg_mutex &mapping->private_lock irq_context: 0 &f->f_pos_lock sb_writers#4 &sb->s_type->i_mutex_key#8 &sbi->s_writepages_rwsem jbd2_handle &ei->i_data_sem &lg->lg_mutex &ret->b_state_lock irq_context: 0 &f->f_pos_lock sb_writers#4 &sb->s_type->i_mutex_key#8 &sbi->s_writepages_rwsem jbd2_handle &ei->i_data_sem &lg->lg_mutex &ret->b_state_lock &journal->j_list_lock irq_context: 0 &f->f_pos_lock sb_writers#4 &sb->s_type->i_mutex_key#8 &sbi->s_writepages_rwsem jbd2_handle &ei->i_data_sem &lg->lg_mutex &pa->pa_lock irq_context: 0 &f->f_pos_lock sb_writers#4 &sb->s_type->i_mutex_key#8 &sbi->s_writepages_rwsem jbd2_handle &ei->i_data_sem &lg->lg_mutex &lg->lg_prealloc_lock irq_context: 0 &f->f_pos_lock sb_writers#4 &sb->s_type->i_mutex_key#8 &sbi->s_writepages_rwsem jbd2_handle &ei->i_data_sem &lg->lg_mutex &rq->__lock irq_context: 0 &f->f_pos_lock sb_writers#4 &sb->s_type->i_mutex_key#8 &sbi->s_writepages_rwsem jbd2_handle &ei->i_data_sem &lg->lg_mutex &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 &f->f_pos_lock sb_writers#4 &sb->s_type->i_mutex_key#8 &sbi->s_writepages_rwsem jbd2_handle &ei->i_data_sem &obj_hash[i].lock irq_context: 0 &f->f_pos_lock sb_writers#4 &sb->s_type->i_mutex_key#8 &sbi->s_writepages_rwsem jbd2_handle &ei->i_data_sem &ei->i_raw_lock irq_context: 0 &f->f_pos_lock sb_writers#4 &sb->s_type->i_mutex_key#8 &sbi->s_writepages_rwsem jbd2_handle &ei->i_data_sem &(ei->i_block_reservation_lock) irq_context: 0 &f->f_pos_lock sb_writers#4 &sb->s_type->i_mutex_key#8 &sbi->s_writepages_rwsem jbd2_handle &ei->i_data_sem &sb->s_type->i_lock_key#22 irq_context: 0 &f->f_pos_lock sb_writers#4 &sb->s_type->i_mutex_key#8 &sbi->s_writepages_rwsem jbd2_handle &ei->i_data_sem &ei->i_es_lock irq_context: 0 &f->f_pos_lock sb_writers#4 &sb->s_type->i_mutex_key#8 &sbi->s_writepages_rwsem jbd2_handle &ei->i_data_sem &ei->i_es_lock irq_context: 0 &f->f_pos_lock sb_writers#4 &sb->s_type->i_mutex_key#8 &sbi->s_writepages_rwsem jbd2_handle &ei->i_data_sem &ei->i_es_lock pool_lock#2 irq_context: 0 &f->f_pos_lock sb_writers#4 &sb->s_type->i_mutex_key#8 &sbi->s_writepages_rwsem jbd2_handle rcu_read_lock &memcg->move_lock irq_context: 0 &f->f_pos_lock sb_writers#4 &sb->s_type->i_mutex_key#8 &sbi->s_writepages_rwsem jbd2_handle rcu_read_lock &xa->xa_lock#9 irq_context: 0 &f->f_pos_lock sb_writers#4 &sb->s_type->i_mutex_key#8 &sbi->s_writepages_rwsem jbd2_handle rcu_read_lock &xa->xa_lock#9 &s->s_inode_wblist_lock irq_context: 0 &f->f_pos_lock sb_writers#4 &sb->s_type->i_mutex_key#8 &sbi->s_writepages_rwsem jbd2_handle lock#4 &lruvec->lru_lock irq_context: 0 &f->f_pos_lock sb_writers#4 &sb->s_type->i_mutex_key#8 &sbi->s_writepages_rwsem jbd2_handle &ei->i_raw_lock irq_context: 0 &f->f_pos_lock sb_writers#4 &sb->s_type->i_mutex_key#8 &sbi->s_writepages_rwsem tk_core.seq.seqcount irq_context: 0 &f->f_pos_lock sb_writers#4 &sb->s_type->i_mutex_key#8 &sbi->s_writepages_rwsem &dd->lock irq_context: 0 &f->f_pos_lock sb_writers#4 &sb->s_type->i_mutex_key#8 &sbi->s_writepages_rwsem rcu_read_lock &dd->lock irq_context: 0 &f->f_pos_lock sb_writers#4 &sb->s_type->i_mutex_key#8 &sbi->s_writepages_rwsem rcu_read_lock &obj_hash[i].lock irq_context: 0 &f->f_pos_lock sb_writers#4 &sb->s_type->i_mutex_key#8 &sbi->s_writepages_rwsem rcu_read_lock tk_core.seq.seqcount irq_context: 0 &f->f_pos_lock sb_writers#4 &sb->s_type->i_mutex_key#8 &sbi->s_writepages_rwsem rcu_read_lock &virtscsi_vq->vq_lock irq_context: 0 sb_writers#4 &sb->s_type->i_mutex_key#8 jbd2_handle &ei->i_data_sem &(ei->i_block_reservation_lock) irq_context: 0 &f->f_pos_lock sb_writers#4 &sb->s_type->i_mutex_key#8 &sbi->s_writepages_rwsem jbd2_handle &ei->i_data_sem &ei->i_es_lock &obj_hash[i].lock irq_context: 0 &f->f_pos_lock sb_writers#4 &sb->s_type->i_mutex_key#8 &sbi->s_writepages_rwsem jbd2_handle &c->lock irq_context: 0 &f->f_pos_lock sb_writers#4 &sb->s_type->i_mutex_key#8 &sbi->s_writepages_rwsem jbd2_handle &____s->seqcount#2 irq_context: 0 &f->f_pos_lock sb_writers#4 &sb->s_type->i_mutex_key#8 &sbi->s_writepages_rwsem jbd2_handle &____s->seqcount irq_context: 0 &f->f_pos_lock sb_writers#4 &sb->s_type->i_mutex_key#8 &sbi->s_writepages_rwsem jbd2_handle &journal->j_wait_updates irq_context: 0 &f->f_pos_lock sb_writers#4 &sb->s_type->i_mutex_key#8 &sbi->s_writepages_rwsem rcu_read_lock rcu_node_0 irq_context: 0 &f->f_pos_lock sb_writers#4 &sb->s_type->i_mutex_key#8 &sbi->s_writepages_rwsem rcu_read_lock &rq->__lock irq_context: 0 &f->f_pos_lock sb_writers#4 &sb->s_type->i_mutex_key#8 &sbi->s_writepages_rwsem rcu_read_lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 &f->f_pos_lock sb_writers#4 &sb->s_type->i_mutex_key#8 &sb->s_type->i_lock_key#22 &xa->xa_lock#9 &obj_hash[i].lock irq_context: 0 &f->f_pos_lock sb_writers#4 &sb->s_type->i_mutex_key#8 &sb->s_type->i_lock_key#22 &xa->xa_lock#9 pool_lock#2 irq_context: 0 &f->f_pos_lock sb_writers#4 &sb->s_type->i_mutex_key#8 mapping.invalidate_lock jbd2_handle mmu_notifier_invalidate_range_start irq_context: 0 &f->f_pos_lock sb_writers#4 &sb->s_type->i_mutex_key#8 mapping.invalidate_lock jbd2_handle &____s->seqcount irq_context: 0 &f->f_pos_lock sb_writers#4 &sb->s_type->i_mutex_key#8 mapping.invalidate_lock jbd2_handle pool_lock#2 irq_context: 0 &f->f_pos_lock sb_writers#4 &sb->s_type->i_mutex_key#8 mapping.invalidate_lock jbd2_handle stock_lock irq_context: 0 &f->f_pos_lock sb_writers#4 &sb->s_type->i_mutex_key#8 mapping.invalidate_lock jbd2_handle &xa->xa_lock#9 irq_context: 0 &f->f_pos_lock sb_writers#4 &sb->s_type->i_mutex_key#8 mapping.invalidate_lock jbd2_handle &xa->xa_lock#9 stock_lock irq_context: 0 &f->f_pos_lock sb_writers#4 &sb->s_type->i_mutex_key#8 mapping.invalidate_lock jbd2_handle &xa->xa_lock#9 pool_lock#2 irq_context: 0 &f->f_pos_lock sb_writers#4 &sb->s_type->i_mutex_key#8 mapping.invalidate_lock jbd2_handle lock#4 irq_context: 0 &f->f_pos_lock sb_writers#4 &sb->s_type->i_mutex_key#8 mapping.invalidate_lock jbd2_handle &mapping->private_lock irq_context: 0 &f->f_pos_lock sb_writers#4 &sb->s_type->i_mutex_key#8 mapping.invalidate_lock jbd2_handle &ei->i_es_lock irq_context: 0 &f->f_pos_lock sb_writers#4 &sb->s_type->i_mutex_key#8 mapping.invalidate_lock jbd2_handle &dd->lock irq_context: 0 &f->f_pos_lock sb_writers#4 &sb->s_type->i_mutex_key#8 mapping.invalidate_lock jbd2_handle &obj_hash[i].lock irq_context: 0 &f->f_pos_lock sb_writers#4 &sb->s_type->i_mutex_key#8 mapping.invalidate_lock jbd2_handle rcu_read_lock &pool->lock irq_context: 0 &f->f_pos_lock sb_writers#4 &sb->s_type->i_mutex_key#8 mapping.invalidate_lock jbd2_handle rcu_read_lock &pool->lock &obj_hash[i].lock irq_context: 0 &f->f_pos_lock sb_writers#4 &sb->s_type->i_mutex_key#8 mapping.invalidate_lock jbd2_handle rcu_read_lock &pool->lock pool_lock#2 irq_context: 0 &f->f_pos_lock sb_writers#4 &sb->s_type->i_mutex_key#8 mapping.invalidate_lock jbd2_handle rcu_read_lock &pool->lock &p->pi_lock irq_context: 0 &f->f_pos_lock sb_writers#4 &sb->s_type->i_mutex_key#8 mapping.invalidate_lock jbd2_handle rcu_read_lock &pool->lock &p->pi_lock &rq->__lock irq_context: 0 &f->f_pos_lock sb_writers#4 &sb->s_type->i_mutex_key#8 mapping.invalidate_lock jbd2_handle rcu_read_lock &pool->lock &p->pi_lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 &f->f_pos_lock sb_writers#4 &sb->s_type->i_mutex_key#8 mapping.invalidate_lock jbd2_handle &rq->__lock irq_context: 0 &f->f_pos_lock sb_writers#4 &sb->s_type->i_mutex_key#8 mapping.invalidate_lock jbd2_handle &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 &f->f_pos_lock sb_writers#4 &sb->s_type->i_mutex_key#8 mapping.invalidate_lock jbd2_handle rcu_read_lock &memcg->move_lock irq_context: 0 &f->f_pos_lock sb_writers#4 &sb->s_type->i_mutex_key#8 mapping.invalidate_lock jbd2_handle rcu_read_lock &xa->xa_lock#9 irq_context: 0 &f->f_pos_lock sb_writers#4 &sb->s_type->i_mutex_key#8 mapping.invalidate_lock jbd2_handle &journal->j_list_lock irq_context: 0 (wq_completion)wg-kex-wg2#11 (work_completion)(&peer->transmit_handshake_work) &peer->endpoint_lock rcu_read_lock_bh &c->lock irq_context: 0 (wq_completion)wg-kex-wg0#7 (work_completion)(&peer->transmit_handshake_work) &peer->endpoint_lock rcu_read_lock_bh rcu_read_lock rcu_read_lock_bh quarantine_lock irq_context: 0 &nft_net->commit_mutex remove_cache_srcu irq_context: 0 &nft_net->commit_mutex remove_cache_srcu quarantine_lock irq_context: 0 &nft_net->commit_mutex remove_cache_srcu &c->lock irq_context: 0 &nft_net->commit_mutex remove_cache_srcu &n->list_lock irq_context: 0 &nft_net->commit_mutex remove_cache_srcu rcu_read_lock pool_lock#2 irq_context: 0 &nft_net->commit_mutex remove_cache_srcu &obj_hash[i].lock irq_context: 0 &nft_net->commit_mutex remove_cache_srcu pool_lock#2 irq_context: 0 &nft_net->commit_mutex remove_cache_srcu &rq->__lock irq_context: 0 &nft_net->commit_mutex remove_cache_srcu &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 nfnl_subsys_ipset fs_reclaim irq_context: 0 nfnl_subsys_ipset fs_reclaim mmu_notifier_invalidate_range_start irq_context: 0 nfnl_subsys_ipset pool_lock#2 irq_context: 0 nfnl_subsys_ipset &obj_hash[i].lock irq_context: 0 &f->f_pos_lock sb_writers#4 &sb->s_type->i_mutex_key#8 jbd2_handle &ei->i_data_sem &____s->seqcount irq_context: 0 &f->f_pos_lock sb_writers#4 &sb->s_type->i_mutex_key#8 mapping.invalidate_lock jbd2_handle &ei->i_data_sem &bgl->locks[i].lock &sbi->s_md_lock irq_context: 0 &f->f_pos_lock sb_writers#4 &sb->s_type->i_mutex_key#8 mapping.invalidate_lock jbd2_handle &ei->i_data_sem &sb->s_type->i_lock_key#22 irq_context: 0 &f->f_pos_lock sb_writers#4 &sb->s_type->i_mutex_key#8 &sbi->s_writepages_rwsem jbd2_handle &ei->i_data_sem &c->lock irq_context: 0 &f->f_pos_lock sb_writers#4 &sb->s_type->i_mutex_key#8 &sbi->s_writepages_rwsem jbd2_handle &ei->i_data_sem &ei->i_es_lock key#2 irq_context: 0 &f->f_pos_lock sb_writers#4 &sb->s_type->i_mutex_key#8 &sbi->s_writepages_rwsem jbd2_handle &ei->i_data_sem rcu_read_lock rcu_node_0 irq_context: 0 &f->f_pos_lock sb_writers#4 &sb->s_type->i_mutex_key#8 &sbi->s_writepages_rwsem jbd2_handle &ei->i_data_sem rcu_read_lock &rq->__lock irq_context: 0 &f->f_pos_lock sb_writers#4 &sb->s_type->i_mutex_key#8 &mm->mmap_lock rcu_read_lock rcu_read_lock &cfs_rq->removed.lock irq_context: 0 &f->f_pos_lock sb_writers#4 &sb->s_type->i_mutex_key#8 mapping.invalidate_lock jbd2_handle bit_wait_table + i irq_context: 0 &f->f_pos_lock sb_writers#4 &sb->s_type->i_mutex_key#8 mapping.invalidate_lock jbd2_handle &sbi->s_orphan_lock &lock->wait_lock irq_context: 0 &f->f_pos_lock sb_writers#4 &sb->s_type->i_mutex_key#8 mapping.invalidate_lock jbd2_handle &sbi->s_orphan_lock &rq->__lock irq_context: 0 &f->f_pos_lock sb_writers#4 &sb->s_type->i_mutex_key#8 mapping.invalidate_lock jbd2_handle &sbi->s_orphan_lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 &f->f_pos_lock sb_writers#4 &sb->s_type->i_mutex_key#8 &sbi->s_writepages_rwsem irq_context: 0 &f->f_pos_lock sb_writers#4 &sb->s_type->i_mutex_key#8 &sbi->s_writepages_rwsem &xa->xa_lock#9 irq_context: 0 &f->f_pos_lock sb_writers#4 &sb->s_type->i_mutex_key#8 &sbi->s_writepages_rwsem mmu_notifier_invalidate_range_start irq_context: 0 &f->f_pos_lock sb_writers#4 &sb->s_type->i_mutex_key#8 &sbi->s_writepages_rwsem &obj_hash[i].lock irq_context: 0 fs_reclaim mmu_notifier_invalidate_range_start &obj_hash[i].lock pool_lock irq_context: 0 (wq_completion)bat_events (work_completion)(&(&bat_priv->bla.work)->work) rcu_read_lock pcpu_lock irq_context: 0 &f->f_pos_lock sb_writers#4 &sb->s_type->i_mutex_key#8 &mm->mmap_lock &____s->seqcount#2 irq_context: 0 sk_lock-AF_INET6 tcpv6_prot_mutex irq_context: 0 sk_lock-AF_INET6 device_spinlock irq_context: 0 sk_lock-AF_INET6 clock-AF_INET6 pool_lock#2 irq_context: 0 sk_lock-AF_INET6 (kmod_concurrent_max).lock irq_context: 0 sk_lock-AF_INET6 rcu_read_lock &pool->lock/1 irq_context: 0 sk_lock-AF_INET6 rcu_read_lock &pool->lock/1 &obj_hash[i].lock irq_context: 0 sk_lock-AF_INET6 rcu_read_lock &pool->lock/1 pool_lock#2 irq_context: 0 sk_lock-AF_INET6 rcu_read_lock &pool->lock/1 &p->pi_lock irq_context: 0 sk_lock-AF_INET6 rcu_read_lock &pool->lock/1 &p->pi_lock &rq->__lock irq_context: 0 sk_lock-AF_INET6 rcu_read_lock &pool->lock/1 &p->pi_lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 sk_lock-AF_INET6 &x->wait#17 irq_context: 0 sk_lock-AF_INET6 k-sk_lock-AF_INET6 irq_context: 0 sk_lock-AF_INET6 k-sk_lock-AF_INET6 k-slock-AF_INET6 irq_context: softirq (&icsk->icsk_retransmit_timer) &n->list_lock irq_context: 0 sk_lock-AF_INET6 running_helpers_waitq.lock irq_context: 0 &pipe->mutex/1 sk_lock-AF_INET6 tk_core.seq.seqcount irq_context: 0 &pipe->mutex/1 sk_lock-AF_INET6 rcu_read_lock rcu_read_lock rcu_read_lock stock_lock irq_context: 0 &pipe->mutex/1 sk_lock-AF_INET6 rcu_read_lock rcu_read_lock rcu_read_lock pool_lock#2 irq_context: 0 &pipe->mutex/1 sk_lock-AF_INET6 rcu_read_lock rcu_read_lock rcu_read_lock &dir->lock#2 irq_context: 0 &pipe->mutex/1 sk_lock-AF_INET6 &base->lock irq_context: 0 &pipe->mutex/1 sk_lock-AF_INET6 &base->lock &obj_hash[i].lock irq_context: 0 &sb->s_type->i_mutex_key#8 rcu_read_lock &rcu_state.gp_wq irq_context: 0 &sb->s_type->i_mutex_key#8 rcu_read_lock &rcu_state.gp_wq &p->pi_lock irq_context: 0 &sb->s_type->i_mutex_key#8 rcu_read_lock &rcu_state.gp_wq &p->pi_lock &rq->__lock irq_context: 0 &sb->s_type->i_mutex_key#8 rcu_read_lock &rcu_state.gp_wq &p->pi_lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 sk_lock-AF_INET6 crypto_alg_sem irq_context: 0 sk_lock-AF_INET6 (crypto_chain).rwsem irq_context: 0 sk_lock-AF_INET6 (crypto_chain).rwsem fs_reclaim irq_context: 0 sk_lock-AF_INET6 (crypto_chain).rwsem fs_reclaim mmu_notifier_invalidate_range_start irq_context: 0 sk_lock-AF_INET6 (crypto_chain).rwsem &c->lock irq_context: 0 sk_lock-AF_INET6 (crypto_chain).rwsem pool_lock#2 irq_context: 0 sk_lock-AF_INET6 (crypto_chain).rwsem kthread_create_lock irq_context: 0 sk_lock-AF_INET6 (crypto_chain).rwsem &p->pi_lock irq_context: 0 sk_lock-AF_INET6 (crypto_chain).rwsem &p->pi_lock &rq->__lock irq_context: 0 sk_lock-AF_INET6 (crypto_chain).rwsem &p->pi_lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 sk_lock-AF_INET6 (crypto_chain).rwsem &rq->__lock irq_context: 0 sk_lock-AF_INET6 (crypto_chain).rwsem &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 sk_lock-AF_INET6 (crypto_chain).rwsem &x->wait irq_context: 0 sk_lock-AF_INET6 (crypto_chain).rwsem &obj_hash[i].lock irq_context: 0 sk_lock-AF_INET6 &x->wait#21 irq_context: 0 (wq_completion)wg-crypt-wg1#3 (work_completion)(&peer->transmit_packet_work) &peer->endpoint_lock rcu_read_lock_bh rcu_read_lock &n->list_lock irq_context: 0 (wq_completion)wg-crypt-wg1#3 (work_completion)(&peer->transmit_packet_work) &peer->endpoint_lock rcu_read_lock_bh rcu_read_lock &n->list_lock &c->lock irq_context: softirq (&icsk->icsk_retransmit_timer) slock-AF_INET6 rcu_read_lock rcu_read_lock &dir->lock#2 irq_context: softirq (&icsk->icsk_retransmit_timer) slock-AF_INET6 rcu_read_lock rcu_read_lock &tbl->lock irq_context: 0 sk_lock-AF_INET6 (&timer.timer) irq_context: 0 sk_lock-AF_INET6 rcu_read_lock &pool->lock/1 &p->pi_lock &cfs_rq->removed.lock irq_context: 0 &sb->s_type->i_mutex_key#8 fill_pool_map-wait-type-override &rq->__lock irq_context: 0 &sb->s_type->i_mutex_key#8 fill_pool_map-wait-type-override &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 sk_lock-AF_INET6 (crypto_chain).rwsem &____s->seqcount#2 irq_context: 0 sk_lock-AF_INET6 (crypto_chain).rwsem &____s->seqcount irq_context: 0 pernet_ops_rwsem nf_ct_proto_mutex nf_hook_mutex rcu_read_lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 pernet_ops_rwsem k-sk_lock-AF_RXRPC &rxnet->local_mutex &p->pi_lock &rq->__lock &cfs_rq->removed.lock irq_context: 0 sk_lock-AF_INET6 &mm->mmap_lock rcu_read_lock per_cpu_ptr(&cgroup_rstat_cpu_lock, cpu) irq_context: 0 &ctx->tx_lock irq_context: 0 &ctx->tx_lock sk_lock-AF_INET6 irq_context: 0 &ctx->tx_lock sk_lock-AF_INET6 slock-AF_INET6 irq_context: 0 &sb->s_type->i_mutex_key#10 (work_completion)(&(&sw_ctx_tx->tx_work.work)->work) irq_context: 0 &ctx->tx_lock sk_lock-AF_INET6 fs_reclaim irq_context: 0 &sb->s_type->i_mutex_key#10 sk_lock-AF_INET6 &sw_ctx_tx->encrypt_compl_lock irq_context: 0 &ctx->tx_lock sk_lock-AF_INET6 fs_reclaim mmu_notifier_invalidate_range_start irq_context: 0 &ctx->tx_lock sk_lock-AF_INET6 fs_reclaim mmu_notifier_invalidate_range_start &rq->__lock irq_context: 0 &ctx->tx_lock &lock->wait_lock irq_context: 0 &ctx->tx_lock &rq->__lock irq_context: 0 &ctx->tx_lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 &ctx->tx_lock sk_lock-AF_INET6 pool_lock#2 irq_context: 0 &ctx->tx_lock sk_lock-AF_INET6 &pcp->lock &zone->lock irq_context: 0 &ctx->tx_lock sk_lock-AF_INET6 &____s->seqcount irq_context: 0 &ctx->tx_lock sk_lock-AF_INET6 &rq->__lock irq_context: 0 &ctx->tx_lock sk_lock-AF_INET6 &mm->mmap_lock irq_context: 0 &ctx->tx_lock sk_lock-AF_INET6 &mm->mmap_lock ptlock_ptr(page)#2 irq_context: 0 &ctx->tx_lock sk_lock-AF_INET6 &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 &ctx->tx_lock sk_lock-AF_INET6 &c->lock irq_context: 0 &ctx->tx_lock sk_lock-AF_INET6 &n->list_lock irq_context: 0 &ctx->tx_lock sk_lock-AF_INET6 &n->list_lock &c->lock irq_context: 0 &ctx->tx_lock sk_lock-AF_INET6 &obj_hash[i].lock irq_context: 0 &ctx->tx_lock sk_lock-AF_INET6 &____s->seqcount#2 irq_context: 0 &ctx->tx_lock sk_lock-AF_INET6 &mm->mmap_lock &rq->__lock irq_context: 0 &ctx->tx_lock sk_lock-AF_INET6 &mm->mmap_lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 &ctx->tx_lock sk_lock-AF_INET6 remove_cache_srcu irq_context: 0 &ctx->tx_lock sk_lock-AF_INET6 remove_cache_srcu quarantine_lock irq_context: 0 &ctx->tx_lock sk_lock-AF_INET6 remove_cache_srcu &c->lock irq_context: 0 &ctx->tx_lock sk_lock-AF_INET6 remove_cache_srcu &n->list_lock irq_context: 0 &ctx->tx_lock sk_lock-AF_INET6 remove_cache_srcu rcu_read_lock pool_lock#2 irq_context: 0 &ctx->tx_lock sk_lock-AF_INET6 remove_cache_srcu &obj_hash[i].lock irq_context: 0 &ctx->tx_lock sk_lock-AF_INET6 &cfs_rq->removed.lock irq_context: 0 &ctx->tx_lock sk_lock-AF_INET6 &mm->mmap_lock fs_reclaim irq_context: 0 &ctx->tx_lock sk_lock-AF_INET6 &mm->mmap_lock fs_reclaim mmu_notifier_invalidate_range_start irq_context: 0 &ctx->tx_lock sk_lock-AF_INET6 &mm->mmap_lock &____s->seqcount irq_context: 0 &ctx->tx_lock sk_lock-AF_INET6 &mm->mmap_lock pool_lock#2 irq_context: 0 &ctx->tx_lock sk_lock-AF_INET6 &mm->mmap_lock stock_lock irq_context: 0 &ctx->tx_lock sk_lock-AF_INET6 &mm->mmap_lock ptlock_ptr(page) irq_context: 0 &ctx->tx_lock sk_lock-AF_INET6 rcu_node_0 irq_context: 0 &ctx->tx_lock sk_lock-AF_INET6 batched_entropy_u8.lock irq_context: 0 &ctx->tx_lock sk_lock-AF_INET6 kfence_freelist_lock irq_context: 0 &ctx->tx_lock sk_lock-AF_INET6 &meta->lock irq_context: 0 &ctx->tx_lock sk_lock-AF_INET6 quarantine_lock irq_context: 0 &ctx->tx_lock sk_lock-AF_INET6 remove_cache_srcu pool_lock#2 irq_context: 0 &ctx->tx_lock sk_lock-AF_INET6 &mm->mmap_lock rcu_read_lock &rq->__lock irq_context: 0 &ctx->tx_lock sk_lock-AF_INET6 &mm->mmap_lock rcu_read_lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 &ctx->tx_lock sk_lock-AF_INET6 pgd_lock irq_context: 0 &ctx->tx_lock sk_lock-AF_INET6 stock_lock irq_context: 0 &ctx->tx_lock sk_lock-AF_INET6 key irq_context: 0 &ctx->tx_lock sk_lock-AF_INET6 pcpu_lock irq_context: 0 &ctx->tx_lock sk_lock-AF_INET6 percpu_counters_lock irq_context: 0 &ctx->tx_lock sk_lock-AF_INET6 pcpu_lock stock_lock irq_context: 0 &ctx->tx_lock sk_lock-AF_INET6 &mm->mmap_lock rcu_read_lock rcu_node_0 irq_context: 0 &ctx->tx_lock sk_lock-AF_INET6 &mm->mmap_lock rcu_read_lock per_cpu_ptr(&cgroup_rstat_cpu_lock, cpu) irq_context: 0 &ctx->tx_lock sk_lock-AF_INET6 &rcu_state.expedited_wq irq_context: 0 &ctx->tx_lock sk_lock-AF_INET6 &rcu_state.expedited_wq &p->pi_lock irq_context: 0 &ctx->tx_lock sk_lock-AF_INET6 &rcu_state.expedited_wq &p->pi_lock &rq->__lock irq_context: 0 &ctx->tx_lock sk_lock-AF_INET6 &rcu_state.expedited_wq &p->pi_lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 &ctx->tx_lock sk_lock-AF_INET6 rcu_read_lock &rq->__lock irq_context: 0 &pool->lock/1 rcu_read_lock &pool->lock &p->pi_lock &rq->__lock &cfs_rq->removed.lock irq_context: 0 cgroup_threadgroup_rwsem freezer_mutex rcu_read_lock &cfs_rq->removed.lock irq_context: 0 cgroup_threadgroup_rwsem freezer_mutex rcu_read_lock &obj_hash[i].lock irq_context: 0 &ctx->tx_lock sk_lock-AF_INET6 rcu_read_lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 &ctx->tx_lock sk_lock-AF_INET6 fs_reclaim &rq->__lock irq_context: 0 &ctx->tx_lock sk_lock-AF_INET6 fs_reclaim &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)wg-crypt-wg0 (work_completion)(&peer->transmit_packet_work) &peer->endpoint_lock rcu_read_lock_bh rcu_read_lock rcu_read_lock &____s->seqcount#7 irq_context: 0 (wq_completion)wg-crypt-wg0 (work_completion)(&peer->transmit_packet_work) &peer->endpoint_lock rcu_read_lock_bh rcu_read_lock &nf_nat_locks[i] irq_context: 0 (wq_completion)wg-crypt-wg0 (work_completion)(&peer->transmit_packet_work) &peer->endpoint_lock rcu_read_lock_bh rcu_read_lock &nf_conntrack_locks[i] irq_context: 0 (wq_completion)wg-crypt-wg0 (work_completion)(&peer->transmit_packet_work) &peer->endpoint_lock rcu_read_lock_bh rcu_read_lock &nf_conntrack_locks[i] &nf_conntrack_locks[i]/1 irq_context: 0 (wq_completion)wg-crypt-wg0 (work_completion)(&peer->transmit_packet_work) &peer->endpoint_lock rcu_read_lock_bh rcu_read_lock &nf_conntrack_locks[i] &nf_conntrack_locks[i]/1 batched_entropy_u8.lock irq_context: 0 (wq_completion)wg-crypt-wg0 (work_completion)(&peer->transmit_packet_work) &peer->endpoint_lock rcu_read_lock_bh rcu_read_lock &nf_conntrack_locks[i]/1 irq_context: 0 (wq_completion)wg-crypt-wg0 (work_completion)(&peer->transmit_packet_work) &peer->endpoint_lock rcu_read_lock_bh rcu_read_lock rcu_read_lock_bh &obj_hash[i].lock irq_context: 0 &ctx->tx_lock sk_lock-AF_INET6 remove_cache_srcu rcu_read_lock rcu_node_0 irq_context: 0 &ctx->tx_lock sk_lock-AF_INET6 remove_cache_srcu rcu_read_lock &rq->__lock irq_context: 0 &ctx->tx_lock sk_lock-AF_INET6 remove_cache_srcu rcu_read_lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 sb_writers#4 &sb->s_type->i_mutex_key#8 mapping.invalidate_lock &mapping->i_mmap_rwsem rcu_read_lock rcu_node_0 irq_context: 0 sb_writers#4 &sb->s_type->i_mutex_key#8 mapping.invalidate_lock &mapping->i_mmap_rwsem rcu_read_lock &rq->__lock irq_context: 0 sb_writers#4 &sb->s_type->i_mutex_key#8 mapping.invalidate_lock &mapping->i_mmap_rwsem rcu_read_lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 &sb->s_type->i_mutex_key#10 sk_lock-AF_INET6 per_cpu_ptr(&cgroup_rstat_cpu_lock, cpu) irq_context: 0 &ctx->tx_lock sk_lock-AF_INET6 rcu_read_lock pool_lock#2 irq_context: 0 &ctx->tx_lock slock-AF_INET6 irq_context: 0 &ctx->tx_lock sk_lock-AF_INET6 &mm->mmap_lock &cfs_rq->removed.lock irq_context: 0 pernet_ops_rwsem devices_rwsem rdma_nets_rwsem &device->compat_devs_mutex quarantine_lock irq_context: 0 sk_lock-AF_INET6 clock-AF_INET6 &c->lock irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem k-sk_lock-AF_INET6 &rq->__lock irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem k-sk_lock-AF_INET6 &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 &ctx->tx_lock sk_lock-AF_INET6 &mm->mmap_lock &sem->wait_lock irq_context: 0 &sb->s_type->i_mutex_key#8 &sbi->s_writepages_rwsem jbd2_handle &ei->i_data_sem mmu_notifier_invalidate_range_start &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 &ctx->tx_lock sk_lock-AF_INET6 &sem->wait_lock irq_context: 0 &ctx->tx_lock sk_lock-AF_INET6 &p->pi_lock irq_context: 0 &ctx->tx_lock sk_lock-AF_INET6 &p->pi_lock &rq->__lock irq_context: 0 &ctx->tx_lock sk_lock-AF_INET6 &p->pi_lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 &ctx->tx_lock sk_lock-AF_INET6 &mm->mmap_lock &obj_hash[i].lock irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem rtnl_mutex &rq->__lock &cfs_rq->removed.lock irq_context: 0 &ctx->tx_lock sk_lock-AF_INET6 remove_cache_srcu &rq->__lock irq_context: 0 &ctx->tx_lock sk_lock-AF_INET6 remove_cache_srcu &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 &ctx->tx_lock sk_lock-AF_INET6 fs_reclaim mmu_notifier_invalidate_range_start &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 &rcu_state.expedited_wq &p->pi_lock &rq->__lock &cfs_rq->removed.lock irq_context: 0 &mm->mmap_lock ptlock_ptr(page)#2 rcu_read_lock &p->pi_lock irq_context: 0 &mm->mmap_lock ptlock_ptr(page)#2 rcu_read_lock &p->pi_lock &rq->__lock irq_context: 0 &mm->mmap_lock ptlock_ptr(page)#2 rcu_read_lock &p->pi_lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)rcu_gp (work_completion)(&rew->rew_work) &p->pi_lock &rq->__lock irq_context: 0 (wq_completion)rcu_gp (work_completion)(&rew->rew_work) &p->pi_lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 sb_writers#4 &sb->s_type->i_mutex_key#8 ima_extend_list_mutex &n->list_lock irq_context: 0 sb_writers#4 &sb->s_type->i_mutex_key#8 ima_extend_list_mutex &n->list_lock &c->lock irq_context: softirq (&timer) rcu_read_lock batched_entropy_u8.lock irq_context: softirq (&timer) rcu_read_lock kfence_freelist_lock irq_context: 0 (wq_completion)wg-crypt-wg1 (work_completion)(&peer->transmit_packet_work) &peer->endpoint_lock rcu_read_lock_bh &obj_hash[i].lock irq_context: 0 rcu_read_lock_bh &meta->lock irq_context: 0 rcu_read_lock_bh kfence_freelist_lock irq_context: 0 &ctx->tx_lock sk_lock-AF_INET6 &mm->mmap_lock &c->lock irq_context: 0 &ctx->tx_lock sk_lock-AF_INET6 &mm->mmap_lock rcu_node_0 irq_context: 0 &ctx->tx_lock sk_lock-AF_INET6 rcu_read_lock rcu_node_0 irq_context: 0 &ctx->tx_lock sk_lock-AF_INET6 remove_cache_srcu rcu_node_0 irq_context: 0 pcpu_alloc_mutex rcu_node_0 irq_context: 0 sb_writers#4 &sbi->s_writepages_rwsem jbd2_handle &ei->i_data_sem &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 sb_writers#4 &sb->s_type->i_mutex_key#8 mapping.invalidate_lock &folio_wait_table[i] &p->pi_lock irq_context: 0 sb_writers#4 &sb->s_type->i_mutex_key#8 mapping.invalidate_lock &folio_wait_table[i] &p->pi_lock &rq->__lock irq_context: 0 sb_writers#4 &sb->s_type->i_mutex_key#8 mapping.invalidate_lock &folio_wait_table[i] &p->pi_lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 &f->f_pos_lock sb_writers#4 &sbi->s_writepages_rwsem &cfs_rq->removed.lock irq_context: 0 &f->f_pos_lock sb_writers#4 &sbi->s_writepages_rwsem jbd2_handle &journal->j_wait_reserved irq_context: 0 &ctx->tx_lock slock-AF_INET6 &sk->sk_lock.wq irq_context: 0 &ctx->tx_lock slock-AF_INET6 &sk->sk_lock.wq &p->pi_lock irq_context: 0 &ctx->tx_lock slock-AF_INET6 &sk->sk_lock.wq &p->pi_lock &rq->__lock irq_context: 0 &ctx->tx_lock slock-AF_INET6 &sk->sk_lock.wq &p->pi_lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 &ctx->tx_lock &cfs_rq->removed.lock irq_context: 0 &ctx->tx_lock &obj_hash[i].lock irq_context: 0 &ctx->tx_lock pool_lock#2 irq_context: 0 sb_writers#4 &sb->s_type->i_mutex_key#8 mapping.invalidate_lock jbd2_handle &sbi->s_orphan_lock &ret->b_state_lock &journal->j_list_lock irq_context: 0 sb_writers#4 &sb->s_type->i_mutex_key#8 mapping.invalidate_lock jbd2_handle hrtimer_bases.lock irq_context: 0 sb_writers#4 &sb->s_type->i_mutex_key#8 mapping.invalidate_lock jbd2_handle hrtimer_bases.lock &obj_hash[i].lock irq_context: 0 &sb->s_type->i_mutex_key#10 sk_lock-AF_INET6 rcu_node_0 irq_context: 0 &xs->mutex rcu_read_lock &rcu_state.gp_wq &p->pi_lock &cfs_rq->removed.lock irq_context: 0 pernet_ops_rwsem nf_ct_proto_mutex fill_pool_map-wait-type-override &c->lock irq_context: 0 pernet_ops_rwsem nf_ct_proto_mutex fill_pool_map-wait-type-override &n->list_lock irq_context: 0 pernet_ops_rwsem nf_ct_proto_mutex fill_pool_map-wait-type-override &n->list_lock &c->lock irq_context: 0 pernet_ops_rwsem nf_ct_proto_mutex fill_pool_map-wait-type-override &rq->__lock irq_context: 0 pernet_ops_rwsem nf_ct_proto_mutex fill_pool_map-wait-type-override &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 pernet_ops_rwsem rtnl_mutex rcu_state.exp_mutex fill_pool_map-wait-type-override pool_lock#2 irq_context: 0 pernet_ops_rwsem rtnl_mutex rcu_state.exp_mutex fill_pool_map-wait-type-override &c->lock irq_context: 0 pernet_ops_rwsem rtnl_mutex rcu_state.exp_mutex fill_pool_map-wait-type-override pool_lock irq_context: 0 (wq_completion)wg-crypt-wg0#6 (work_completion)(&peer->transmit_packet_work) &peer->endpoint_lock rcu_read_lock_bh quarantine_lock irq_context: 0 pernet_ops_rwsem rtnl_mutex uevent_sock_mutex &base->lock irq_context: 0 pernet_ops_rwsem rtnl_mutex uevent_sock_mutex &base->lock &obj_hash[i].lock irq_context: 0 &pipe->mutex/1 per_cpu_ptr(&cgroup_rstat_cpu_lock, cpu) irq_context: 0 &sb->s_type->i_mutex_key#9 &sb->s_type->i_lock_key#23 &dentry->d_lock &wq &p->pi_lock irq_context: 0 &sb->s_type->i_mutex_key#9 &sb->s_type->i_lock_key#23 &dentry->d_lock &wq &p->pi_lock &rq->__lock irq_context: 0 &sb->s_type->i_mutex_key#9 &sb->s_type->i_lock_key#23 &dentry->d_lock &wq &p->pi_lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 sb_writers#4 &sb->s_type->i_mutex_key#8 mapping.invalidate_lock jbd2_handle &ei->i_data_sem &sem->wait_lock irq_context: 0 &f->f_pos_lock sb_writers#4 &sbi->s_writepages_rwsem jbd2_handle &sem->wait_lock irq_context: 0 &f->f_pos_lock sb_writers#4 &sbi->s_writepages_rwsem jbd2_handle &p->pi_lock irq_context: 0 &f->f_pos_lock sb_writers#4 &sbi->s_writepages_rwsem jbd2_handle &p->pi_lock &cfs_rq->removed.lock irq_context: 0 &f->f_pos_lock sb_writers#4 &sbi->s_writepages_rwsem jbd2_handle &p->pi_lock &rq->__lock irq_context: 0 &f->f_pos_lock sb_writers#4 &sbi->s_writepages_rwsem jbd2_handle &p->pi_lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 &f->f_pos_lock sb_writers#4 &sbi->s_writepages_rwsem jbd2_handle &rq->__lock &cfs_rq->removed.lock irq_context: 0 &f->f_pos_lock sb_writers#4 &sbi->s_writepages_rwsem &journal->j_state_lock &journal->j_wait_reserved irq_context: 0 &f->f_pos_lock sb_writers#4 &sbi->s_writepages_rwsem &ext4__ioend_wq[i] irq_context: 0 sb_writers#4 &sb->s_type->i_mutex_key#8 jbd2_handle &journal->j_state_lock irq_context: 0 &f->f_pos_lock sb_writers#4 &sbi->s_writepages_rwsem batched_entropy_u8.lock irq_context: 0 &f->f_pos_lock sb_writers#4 &sbi->s_writepages_rwsem kfence_freelist_lock irq_context: 0 &f->f_pos_lock sb_writers#4 &sbi->s_writepages_rwsem &meta->lock irq_context: 0 &f->f_pos_lock sb_writers#4 &sbi->s_writepages_rwsem jbd2_handle remove_cache_srcu irq_context: 0 &f->f_pos_lock sb_writers#4 &sbi->s_writepages_rwsem jbd2_handle remove_cache_srcu quarantine_lock irq_context: 0 &f->f_pos_lock sb_writers#4 &sbi->s_writepages_rwsem jbd2_handle remove_cache_srcu &c->lock irq_context: 0 &f->f_pos_lock sb_writers#4 &sbi->s_writepages_rwsem jbd2_handle remove_cache_srcu &n->list_lock irq_context: 0 &f->f_pos_lock sb_writers#4 &sbi->s_writepages_rwsem jbd2_handle remove_cache_srcu rcu_read_lock pool_lock#2 irq_context: 0 &f->f_pos_lock sb_writers#4 &sbi->s_writepages_rwsem jbd2_handle remove_cache_srcu rcu_read_lock rcu_node_0 irq_context: 0 &f->f_pos_lock sb_writers#4 &sbi->s_writepages_rwsem jbd2_handle remove_cache_srcu rcu_read_lock &rq->__lock irq_context: 0 &f->f_pos_lock sb_writers#4 &sbi->s_writepages_rwsem jbd2_handle remove_cache_srcu rcu_read_lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 &f->f_pos_lock sb_writers#4 &sbi->s_writepages_rwsem jbd2_handle remove_cache_srcu rcu_read_lock &rcu_state.gp_wq irq_context: 0 &f->f_pos_lock sb_writers#4 &sbi->s_writepages_rwsem jbd2_handle remove_cache_srcu rcu_read_lock &rcu_state.gp_wq &p->pi_lock irq_context: 0 &f->f_pos_lock sb_writers#4 &sbi->s_writepages_rwsem jbd2_handle remove_cache_srcu rcu_read_lock &rcu_state.gp_wq &p->pi_lock &rq->__lock irq_context: 0 &f->f_pos_lock sb_writers#4 &sbi->s_writepages_rwsem jbd2_handle remove_cache_srcu rcu_read_lock &rcu_state.gp_wq &p->pi_lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 &f->f_pos_lock sb_writers#4 &sbi->s_writepages_rwsem jbd2_handle remove_cache_srcu &obj_hash[i].lock irq_context: 0 &f->f_pos_lock sb_writers#4 &sbi->s_writepages_rwsem jbd2_handle remove_cache_srcu pool_lock#2 irq_context: 0 tracepoints_mutex rcu_state.exp_mutex &rnp->exp_wq[1] irq_context: 0 tracepoints_mutex tracepoints_mutex.wait_lock irq_context: 0 tracepoints_mutex.wait_lock irq_context: 0 tracepoints_mutex &cfs_rq->removed.lock irq_context: 0 rtnl_mutex _xmit_ETHER rcu_read_lock &pool->lock/1 &p->pi_lock &cfs_rq->removed.lock irq_context: 0 rtnl_mutex &rdev->wiphy.mtx &wdev->mtx &wdev->pmsr_lock irq_context: 0 rtnl_mutex &rdev->wiphy.mtx &wdev->mtx &local->sta_mtx rcu_state.exp_mutex &rnp->exp_wq[2] irq_context: 0 rtnl_mutex &rdev->wiphy.mtx &wdev->mtx &local->sta_mtx &sb->s_type->i_mutex_key#3 &obj_hash[i].lock pool_lock irq_context: 0 rtnl_mutex &rdev->wiphy.mtx &wdev->mtx rcu_state.exp_mutex &rnp->exp_wq[3] irq_context: 0 rtnl_mutex (work_completion)(&wdev->disconnect_wk) irq_context: 0 rtnl_mutex (work_completion)(&wdev->pmsr_free_wk) irq_context: 0 rtnl_mutex (work_completion)(&sdata->activate_links_work) irq_context: 0 rtnl_mutex &rdev->wiphy.mtx rcu_state.exp_mutex rcu_node_0 irq_context: 0 rtnl_mutex &rdev->wiphy.mtx rcu_state.exp_mutex &obj_hash[i].lock irq_context: 0 rtnl_mutex &rdev->wiphy.mtx rcu_state.exp_mutex rcu_read_lock &pool->lock irq_context: 0 rtnl_mutex &rdev->wiphy.mtx rcu_state.exp_mutex rcu_read_lock &pool->lock &obj_hash[i].lock irq_context: 0 rtnl_mutex &rdev->wiphy.mtx rcu_state.exp_mutex rcu_read_lock &pool->lock &p->pi_lock irq_context: 0 rtnl_mutex &rdev->wiphy.mtx rcu_state.exp_mutex rcu_read_lock &pool->lock &p->pi_lock &rq->__lock irq_context: 0 rtnl_mutex &rdev->wiphy.mtx rcu_state.exp_mutex rcu_read_lock &pool->lock &p->pi_lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 rtnl_mutex &rdev->wiphy.mtx rcu_state.exp_mutex &rq->__lock irq_context: 0 rtnl_mutex &rdev->wiphy.mtx rcu_state.exp_mutex &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 rtnl_mutex &rdev->wiphy.mtx &local->mtx irq_context: 0 rtnl_mutex &rdev->wiphy.mtx &rdev->wiphy_work_lock irq_context: 0 rtnl_mutex &rdev->wiphy.mtx &local->sta_mtx irq_context: 0 rtnl_mutex &rdev->wiphy.mtx _xmit_ETHER irq_context: 0 rtnl_mutex &rdev->wiphy.mtx _xmit_ETHER &local->filter_lock irq_context: 0 rtnl_mutex &rdev->wiphy.mtx _xmit_ETHER &local->filter_lock &obj_hash[i].lock irq_context: 0 rtnl_mutex &rdev->wiphy.mtx _xmit_ETHER &local->filter_lock pool_lock#2 irq_context: 0 rtnl_mutex &rdev->wiphy.mtx _xmit_ETHER &local->filter_lock krc.lock irq_context: 0 rtnl_mutex &rdev->wiphy.mtx &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 rtnl_mutex &rdev->wiphy.mtx (&local->dynamic_ps_timer) irq_context: 0 rtnl_mutex &rdev->wiphy.mtx (work_completion)(&local->dynamic_ps_enable_work) irq_context: 0 rtnl_mutex &rdev->wiphy.mtx (work_completion)(&sdata->recalc_smps) irq_context: 0 rtnl_mutex &rdev->wiphy.mtx (work_completion)(&link->csa_finalize_work) irq_context: 0 rtnl_mutex &rdev->wiphy.mtx (work_completion)(&link->color_change_finalize_work) irq_context: 0 rtnl_mutex &rdev->wiphy.mtx (work_completion)(&(&link->dfs_cac_timer_work)->work) irq_context: 0 rtnl_mutex &rdev->wiphy.mtx (work_completion)(&(&sdata->dec_tailroom_needed_wk)->work) irq_context: 0 rtnl_mutex &rdev->wiphy.mtx &local->key_mtx irq_context: 0 rtnl_mutex &rdev->wiphy.mtx &local->key_mtx &rq->__lock irq_context: 0 rtnl_mutex &rdev->wiphy.mtx &local->key_mtx &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 rtnl_mutex &rdev->wiphy.mtx &local->key_mtx rcu_state.exp_mutex rcu_node_0 irq_context: 0 rtnl_mutex &rdev->wiphy.mtx &local->key_mtx rcu_state.exp_mutex &obj_hash[i].lock irq_context: 0 rtnl_mutex &rdev->wiphy.mtx &local->key_mtx rcu_state.exp_mutex rcu_read_lock &pool->lock irq_context: 0 rtnl_mutex &rdev->wiphy.mtx &local->key_mtx rcu_state.exp_mutex rcu_read_lock &pool->lock &obj_hash[i].lock irq_context: 0 rtnl_mutex &rdev->wiphy.mtx &local->key_mtx rcu_state.exp_mutex rcu_read_lock &pool->lock &p->pi_lock irq_context: 0 rtnl_mutex &rdev->wiphy.mtx &local->key_mtx rcu_state.exp_mutex rcu_read_lock &pool->lock &p->pi_lock &rq->__lock irq_context: 0 rtnl_mutex &rdev->wiphy.mtx &local->key_mtx rcu_state.exp_mutex rcu_read_lock &pool->lock &p->pi_lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 rtnl_mutex &rdev->wiphy.mtx &local->key_mtx rcu_state.exp_mutex &rnp->exp_wq[2] irq_context: 0 rtnl_mutex &rdev->wiphy.mtx &local->key_mtx rcu_state.exp_mutex &rq->__lock irq_context: 0 rtnl_mutex &rdev->wiphy.mtx &local->key_mtx rcu_state.exp_mutex &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 rtnl_mutex &rdev->wiphy.mtx &local->key_mtx &obj_hash[i].lock irq_context: 0 rtnl_mutex &rdev->wiphy.mtx &list->lock#18 irq_context: 0 rtnl_mutex &rdev->wiphy.mtx &local->queue_stop_reason_lock irq_context: 0 rtnl_mutex &rdev->wiphy.mtx &list->lock#19 irq_context: 0 rtnl_mutex &rdev->wiphy.mtx (work_completion)(&local->reconfig_filter) irq_context: 0 rtnl_mutex &rdev->wiphy.mtx (wq_completion)phy8 irq_context: 0 rtnl_mutex &rdev->wiphy.mtx &wq->mutex irq_context: 0 rtnl_mutex &rdev->wiphy.mtx &wq->mutex &pool->lock/1 irq_context: 0 rtnl_mutex &rdev->wiphy.mtx &wq->mutex &x->wait#10 irq_context: 0 rtnl_mutex __ip_vs_mutex &rq->__lock irq_context: 0 rtnl_mutex __ip_vs_mutex &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 rtnl_mutex &rdev->dev_wait irq_context: 0 (wq_completion)wg-crypt-wg0 (work_completion)(&peer->transmit_packet_work) &peer->endpoint_lock rcu_read_lock_bh &c->lock irq_context: 0 (wq_completion)wg-crypt-wg0#6 (work_completion)(&peer->transmit_packet_work) &peer->endpoint_lock rcu_read_lock_bh &n->list_lock irq_context: 0 (wq_completion)wg-crypt-wg0#6 (work_completion)(&peer->transmit_packet_work) &peer->endpoint_lock rcu_read_lock_bh &n->list_lock &c->lock irq_context: 0 (wq_completion)events_power_efficient (work_completion)(&(&gc_work->dwork)->work) rcu_read_lock &n->list_lock irq_context: 0 rtnl_mutex &rdev->wiphy.mtx rcu_state.exp_mutex &rnp->exp_wq[1] irq_context: 0 rtnl_mutex &rdev->wiphy.mtx &local->key_mtx rcu_state.exp_mutex &obj_hash[i].lock pool_lock irq_context: 0 rtnl_mutex &rdev->wiphy.mtx (wq_completion)phy4 irq_context: 0 sb_writers#4 &type->i_mutex_dir_key#3/1 &sb->s_type->i_mutex_key#8 remove_cache_srcu pgd_lock irq_context: 0 sb_writers#4 &type->i_mutex_dir_key#3/1 &sb->s_type->i_mutex_key#8 remove_cache_srcu stock_lock irq_context: 0 sb_writers#4 &type->i_mutex_dir_key#3/1 &sb->s_type->i_mutex_key#8 remove_cache_srcu key irq_context: 0 sb_writers#4 &type->i_mutex_dir_key#3/1 &sb->s_type->i_mutex_key#8 remove_cache_srcu pcpu_lock irq_context: 0 sb_writers#4 &type->i_mutex_dir_key#3/1 &sb->s_type->i_mutex_key#8 remove_cache_srcu percpu_counters_lock irq_context: 0 sb_writers#4 &type->i_mutex_dir_key#3/1 &sb->s_type->i_mutex_key#8 remove_cache_srcu pcpu_lock stock_lock irq_context: 0 (wq_completion)wg-crypt-wg1#5 (work_completion)(&({ do { const void *__vpp_verify = (typeof((worker) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __asm__ ("" : "=r"(__ptr) : "0"((typeof(*((worker))) *)((worker)))); (typeof((typeof(*((worker))) *)((worker)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); })->work) &c->lock irq_context: 0 &sb->s_type->i_mutex_key#10 &rnp->exp_wq[2] irq_context: 0 cb_lock genl_mutex rcu_read_lock &rcu_state.expedited_wq irq_context: 0 cb_lock genl_mutex rcu_read_lock &rcu_state.expedited_wq &p->pi_lock irq_context: 0 cb_lock genl_mutex rcu_read_lock &rcu_state.expedited_wq &p->pi_lock &rq->__lock irq_context: 0 cb_lock genl_mutex rcu_read_lock &rcu_state.expedited_wq &p->pi_lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 &fsnotify_mark_srcu rcu_read_lock rcu_node_0 irq_context: 0 &fsnotify_mark_srcu rcu_read_lock &rq->__lock irq_context: 0 &fsnotify_mark_srcu rcu_read_lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 &fsnotify_mark_srcu rcu_read_lock &rcu_state.expedited_wq irq_context: 0 &fsnotify_mark_srcu rcu_read_lock &rcu_state.expedited_wq &p->pi_lock irq_context: 0 &fsnotify_mark_srcu rcu_read_lock &rcu_state.expedited_wq &p->pi_lock &rq->__lock irq_context: 0 &fsnotify_mark_srcu rcu_read_lock &rcu_state.expedited_wq &p->pi_lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)events (linkwatch_work).work rtnl_mutex &rq->__lock &cfs_rq->removed.lock irq_context: 0 (wq_completion)wg-crypt-wg1#2 (work_completion)(&peer->transmit_packet_work) &peer->endpoint_lock rcu_read_lock_bh rcu_read_lock rcu_read_lock_bh rcu_read_lock tk_core.seq.seqcount irq_context: 0 (wq_completion)wg-crypt-wg1#2 (work_completion)(&peer->transmit_packet_work) &peer->endpoint_lock rcu_read_lock_bh rcu_read_lock rcu_read_lock_bh rcu_read_lock &obj_hash[i].lock irq_context: 0 &sb->s_type->i_mutex_key#10 rtnl_mutex rtnl_mutex.wait_lock irq_context: 0 pernet_ops_rwsem rtnl_mutex pcpu_alloc_mutex &cfs_rq->removed.lock irq_context: 0 pernet_ops_rwsem rtnl_mutex pcpu_alloc_mutex &obj_hash[i].lock irq_context: 0 pernet_ops_rwsem rtnl_mutex pcpu_alloc_mutex pool_lock#2 irq_context: 0 pernet_ops_rwsem rtnl_mutex fs_reclaim &cfs_rq->removed.lock irq_context: 0 pernet_ops_rwsem rtnl_mutex fs_reclaim &obj_hash[i].lock irq_context: 0 &sb->s_type->i_mutex_key#10 rtnl_mutex.wait_lock irq_context: 0 &sb->s_type->i_mutex_key#10 sk_lock-AF_PPPOX chan_lock irq_context: 0 &sb->s_type->i_mutex_key#10 sk_lock-AF_PPPOX rcu_state.exp_mutex rcu_node_0 irq_context: 0 &sb->s_type->i_mutex_key#10 sk_lock-AF_PPPOX rcu_state.exp_mutex &obj_hash[i].lock irq_context: 0 &sb->s_type->i_mutex_key#10 sk_lock-AF_PPPOX rcu_state.exp_mutex rcu_read_lock &pool->lock irq_context: 0 &sb->s_type->i_mutex_key#10 sk_lock-AF_PPPOX rcu_state.exp_mutex rcu_read_lock &pool->lock &obj_hash[i].lock irq_context: 0 &sb->s_type->i_mutex_key#10 sk_lock-AF_PPPOX rcu_state.exp_mutex rcu_read_lock &pool->lock &p->pi_lock irq_context: 0 &sb->s_type->i_mutex_key#10 sk_lock-AF_PPPOX rcu_state.exp_mutex rcu_read_lock &pool->lock &p->pi_lock &rq->__lock irq_context: 0 &sb->s_type->i_mutex_key#10 sk_lock-AF_PPPOX rcu_state.exp_mutex rcu_read_lock &pool->lock &p->pi_lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 &sb->s_type->i_mutex_key#10 sk_lock-AF_PPPOX rcu_state.exp_mutex &rq->__lock irq_context: 0 &sb->s_type->i_mutex_key#10 sk_lock-AF_PPPOX rcu_state.exp_mutex &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 &sb->s_type->i_mutex_key#10 sk_lock-AF_PPPOX &obj_hash[i].lock irq_context: 0 &sb->s_type->i_mutex_key#10 sk_lock-AF_PPPOX rcu_state.exp_mutex &rnp->exp_wq[1] irq_context: 0 &sb->s_type->i_mutex_key#10 sk_lock-AF_PPPOX rcu_state.exp_mutex rcu_read_lock &rq->__lock irq_context: 0 &sb->s_type->i_mutex_key#10 sk_lock-AF_PPPOX rcu_state.exp_mutex rcu_read_lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 &sb->s_type->i_mutex_key#10 sk_lock-AF_PPPOX rcu_state.exp_mutex.wait_lock irq_context: 0 &sb->s_type->i_mutex_key#10 sk_lock-AF_PPPOX &p->pi_lock irq_context: 0 &sb->s_type->i_mutex_key#10 sk_lock-AF_PPPOX &p->pi_lock &rq->__lock irq_context: 0 &sb->s_type->i_mutex_key#10 sk_lock-AF_PPPOX &p->pi_lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 &sb->s_type->i_mutex_key#10 sk_lock-AF_PPPOX &rq->__lock irq_context: 0 &sb->s_type->i_mutex_key#10 sk_lock-AF_PPPOX &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 nfnl_subsys_ipset &c->lock irq_context: 0 nfnl_subsys_ipset rcu_read_lock &c->lock irq_context: 0 nfnl_subsys_ipset rcu_read_lock pool_lock#2 irq_context: 0 nfnl_subsys_ipset rcu_read_lock rcu_read_lock_bh &obj_hash[i].lock irq_context: 0 nfnl_subsys_ipset rcu_read_lock rcu_read_lock_bh pool_lock#2 irq_context: 0 nfnl_subsys_ipset rlock-AF_NETLINK irq_context: 0 &sb->s_type->i_mutex_key#10 sk_lock-AF_PPPOX &rnp->exp_lock irq_context: 0 &sb->s_type->i_mutex_key#10 sk_lock-AF_PPPOX rcu_state.exp_mutex irq_context: 0 &sb->s_type->i_mutex_key#10 sk_lock-AF_PPPOX rcu_state.exp_mutex rcu_state.exp_mutex.wait_lock irq_context: 0 &sb->s_type->i_mutex_key#10 sk_lock-AF_PPPOX rcu_state.exp_mutex &rnp->exp_wq[3] irq_context: 0 cb_lock genl_mutex &sdata->sec_mtx irq_context: 0 nfnl_subsys_ipset &n->list_lock irq_context: 0 nfnl_subsys_ipset &n->list_lock &c->lock irq_context: 0 nfnl_subsys_ipset &rq->__lock irq_context: 0 (wq_completion)wg-kex-wg2 (work_completion)(&peer->transmit_handshake_work) &peer->endpoint_lock rcu_read_lock_bh &obj_hash[i].lock irq_context: 0 rtnl_mutex uevent_sock_mutex.wait_lock irq_context: 0 (wq_completion)wg-crypt-wg2 (work_completion)(&peer->transmit_packet_work) &peer->endpoint_lock rcu_read_lock_bh rcu_read_lock rcu_read_lock &____s->seqcount#7 irq_context: 0 (wq_completion)wg-crypt-wg2 (work_completion)(&peer->transmit_packet_work) &peer->endpoint_lock rcu_read_lock_bh rcu_read_lock &nf_nat_locks[i] irq_context: 0 (wq_completion)wg-crypt-wg2 (work_completion)(&peer->transmit_packet_work) &peer->endpoint_lock rcu_read_lock_bh rcu_read_lock &nf_conntrack_locks[i] irq_context: 0 (wq_completion)wg-crypt-wg2 (work_completion)(&peer->transmit_packet_work) &peer->endpoint_lock rcu_read_lock_bh rcu_read_lock &nf_conntrack_locks[i] &nf_conntrack_locks[i]/1 irq_context: 0 (wq_completion)wg-crypt-wg2 (work_completion)(&peer->transmit_packet_work) &peer->endpoint_lock rcu_read_lock_bh rcu_read_lock &nf_conntrack_locks[i] &nf_conntrack_locks[i]/1 batched_entropy_u8.lock irq_context: 0 (wq_completion)wg-crypt-wg2 (work_completion)(&peer->transmit_packet_work) &peer->endpoint_lock rcu_read_lock_bh rcu_read_lock rcu_read_lock_bh &obj_hash[i].lock irq_context: 0 (wq_completion)wg-kex-wg2 (work_completion)(&peer->transmit_handshake_work) &peer->endpoint_lock rcu_read_lock_bh pool_lock#2 irq_context: 0 (wq_completion)wg-kex-wg2 (work_completion)(&peer->transmit_handshake_work) &peer->endpoint_lock rcu_read_lock_bh &c->lock irq_context: 0 (wq_completion)wg-kex-wg2 (work_completion)(&peer->transmit_handshake_work) &peer->endpoint_lock rcu_read_lock_bh &n->list_lock irq_context: 0 (wq_completion)wg-kex-wg2 (work_completion)(&peer->transmit_handshake_work) &peer->endpoint_lock rcu_read_lock_bh &n->list_lock &c->lock irq_context: 0 (wq_completion)wg-kex-wg2 (work_completion)(&peer->transmit_handshake_work) &peer->endpoint_lock rcu_read_lock_bh rcu_read_lock rcu_read_lock_bh &obj_hash[i].lock irq_context: 0 sb_writers#8 kn->active#5 rcu_read_lock &cfs_rq->removed.lock irq_context: 0 sb_writers#8 kn->active#5 rcu_read_lock &obj_hash[i].lock irq_context: 0 &sb->s_type->i_mutex_key#10 sk_lock-AF_INET6/1 fill_pool_map-wait-type-override &n->list_lock irq_context: 0 &sb->s_type->i_mutex_key#10 sk_lock-AF_INET6/1 fill_pool_map-wait-type-override &n->list_lock &c->lock irq_context: 0 &sb->s_type->i_mutex_key#10 &net->sctp.addr_wq_lock slock-AF_INET6/1 fill_pool_map-wait-type-override &____s->seqcount irq_context: 0 &sb->s_type->i_mutex_key#10 &net->sctp.addr_wq_lock slock-AF_INET6/1 fill_pool_map-wait-type-override pool_lock#2 irq_context: 0 &sb->s_type->i_mutex_key#10 &net->sctp.addr_wq_lock slock-AF_INET6/1 fill_pool_map-wait-type-override rcu_read_lock pool_lock#2 irq_context: 0 &sb->s_type->i_mutex_key#10 &net->sctp.addr_wq_lock slock-AF_INET6/1 fill_pool_map-wait-type-override &obj_hash[i].lock irq_context: 0 &sb->s_type->i_mutex_key#10 &net->sctp.addr_wq_lock slock-AF_INET6/1 fill_pool_map-wait-type-override &c->lock irq_context: 0 &sb->s_type->i_mutex_key#10 &net->sctp.addr_wq_lock slock-AF_INET6/1 fill_pool_map-wait-type-override &n->list_lock irq_context: 0 &sb->s_type->i_mutex_key#10 &net->sctp.addr_wq_lock slock-AF_INET6/1 fill_pool_map-wait-type-override &n->list_lock &c->lock irq_context: 0 &sb->s_type->i_mutex_key#10 &net->sctp.addr_wq_lock slock-AF_INET6/1 fill_pool_map-wait-type-override pool_lock irq_context: 0 &sb->s_type->i_mutex_key#10 &net->sctp.addr_wq_lock slock-AF_INET6/1 krc.lock &obj_hash[i].lock irq_context: 0 &sb->s_type->i_mutex_key#10 &net->sctp.addr_wq_lock slock-AF_INET6/1 krc.lock &base->lock irq_context: 0 &sb->s_type->i_mutex_key#10 &net->sctp.addr_wq_lock slock-AF_INET6/1 krc.lock &base->lock &obj_hash[i].lock irq_context: softirq rcu_read_lock rcu_read_lock rcu_read_lock rcu_read_lock &ul->lock irq_context: 0 sk_lock-AF_KCM &mm->mmap_lock ptlock_ptr(page)#2 irq_context: 0 sk_lock-AF_KCM &rq->__lock irq_context: 0 sk_lock-AF_KCM &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 sk_lock-AF_KCM &c->lock irq_context: 0 sk_lock-AF_KCM &mm->mmap_lock rcu_read_lock rcu_node_0 irq_context: 0 sk_lock-AF_KCM &mm->mmap_lock rcu_read_lock &rq->__lock irq_context: 0 sk_lock-AF_KCM &mm->mmap_lock rcu_read_lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)events (work_completion)(&(&krcp->monitor_work)->work) rcu_callback rcu_read_lock &rq->__lock irq_context: 0 (wq_completion)events (work_completion)(&(&krcp->monitor_work)->work) rcu_callback rcu_read_lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 sk_lock-AF_KCM remove_cache_srcu irq_context: 0 sk_lock-AF_KCM remove_cache_srcu quarantine_lock irq_context: 0 sk_lock-AF_KCM remove_cache_srcu &c->lock irq_context: 0 sk_lock-AF_KCM remove_cache_srcu &n->list_lock irq_context: 0 sk_lock-AF_KCM remove_cache_srcu rcu_read_lock pool_lock#2 irq_context: 0 sk_lock-AF_KCM remove_cache_srcu &obj_hash[i].lock irq_context: 0 sk_lock-AF_KCM remove_cache_srcu rcu_read_lock rcu_node_0 irq_context: 0 sk_lock-AF_KCM remove_cache_srcu rcu_read_lock &rq->__lock irq_context: 0 sk_lock-AF_KCM remove_cache_srcu rcu_read_lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 sk_lock-AF_KCM rcu_node_0 irq_context: 0 sk_lock-AF_KCM rcu_read_lock rcu_node_0 irq_context: 0 sk_lock-AF_KCM rcu_read_lock &rq->__lock irq_context: 0 sk_lock-AF_KCM rcu_read_lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 sk_lock-AF_KCM &mm->mmap_lock fs_reclaim irq_context: 0 sk_lock-AF_KCM &mm->mmap_lock fs_reclaim mmu_notifier_invalidate_range_start irq_context: 0 sk_lock-AF_KCM &mm->mmap_lock &____s->seqcount irq_context: 0 sk_lock-AF_KCM &mm->mmap_lock stock_lock irq_context: 0 sk_lock-AF_KCM &mm->mmap_lock ptlock_ptr(page) irq_context: 0 sk_lock-AF_KCM &sem->wait_lock irq_context: 0 sk_lock-AF_KCM &p->pi_lock irq_context: 0 sk_lock-AF_KCM &p->pi_lock &rq->__lock irq_context: 0 sk_lock-AF_KCM &p->pi_lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 rlock-AF_KCM irq_context: 0 sk_lock-AF_KCM &pcp->lock &zone->lock irq_context: 0 sk_lock-AF_KCM &mm->mmap_lock &sem->wait_lock irq_context: 0 sk_lock-AF_KCM slock-AF_KCM &sk->sk_lock.wq irq_context: 0 sk_lock-AF_KCM &n->list_lock irq_context: 0 sk_lock-AF_KCM &n->list_lock &c->lock irq_context: 0 sk_lock-AF_KCM &cfs_rq->removed.lock irq_context: 0 sk_lock-AF_KCM rcu_read_lock &rq->__lock &cfs_rq->removed.lock irq_context: 0 sk_lock-AF_KCM &mm->mmap_lock rcu_node_0 irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem rtnl_mutex dpm_list_mtx &rq->__lock irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem rtnl_mutex dpm_list_mtx &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 sk_lock-AF_KCM &pcp->lock &zone->lock &____s->seqcount irq_context: 0 slock-AF_KCM &sk->sk_lock.wq irq_context: 0 slock-AF_KCM &sk->sk_lock.wq &p->pi_lock irq_context: 0 slock-AF_KCM &sk->sk_lock.wq &p->pi_lock &rq->__lock irq_context: 0 slock-AF_KCM &sk->sk_lock.wq &p->pi_lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 &sb->s_type->i_mutex_key#10 sk_lock-AF_KCM rcu_read_lock pool_lock#2 irq_context: 0 &sb->s_type->i_mutex_key#10 sk_lock-AF_KCM &rq->__lock irq_context: 0 &sb->s_type->i_mutex_key#10 sk_lock-AF_KCM &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 &sb->s_type->i_mutex_key#10 sk_lock-AF_KCM &pcp->lock &zone->lock irq_context: 0 &sb->s_type->i_mutex_key#10 sk_lock-AF_KCM &pcp->lock &zone->lock &____s->seqcount irq_context: softirq rcu_read_lock rcu_read_lock slock-AF_INET/1 &obj_hash[i].lock pool_lock irq_context: softirq rcu_read_lock rcu_read_lock rcu_read_lock rcu_read_lock &n->list_lock irq_context: softirq rcu_read_lock rcu_read_lock rcu_read_lock rcu_read_lock &n->list_lock &c->lock irq_context: 0 slock-AF_INET &sk->sk_lock.wq &p->pi_lock &cfs_rq->removed.lock irq_context: softirq rcu_read_lock rcu_read_lock rcu_read_lock rcu_read_lock key#23 irq_context: 0 (wq_completion)wg-crypt-wg0 (work_completion)(&peer->transmit_packet_work) &peer->endpoint_lock rcu_read_lock_bh quarantine_lock irq_context: 0 (wq_completion)wg-crypt-wg0#6 (work_completion)(&peer->transmit_packet_work) &peer->endpoint_lock rcu_read_lock_bh &meta->lock irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem rtnl_mutex &tn->lock &rq->__lock irq_context: 0 pernet_ops_rwsem rtnl_mutex rcu_read_lock pgd_lock irq_context: 0 pernet_ops_rwsem rtnl_mutex rcu_read_lock stock_lock irq_context: 0 pernet_ops_rwsem rtnl_mutex rcu_read_lock key irq_context: 0 pernet_ops_rwsem rtnl_mutex rcu_read_lock pcpu_lock irq_context: 0 pernet_ops_rwsem rtnl_mutex rcu_read_lock percpu_counters_lock irq_context: 0 pernet_ops_rwsem rtnl_mutex rcu_read_lock pcpu_lock stock_lock irq_context: 0 pernet_ops_rwsem cpu_hotplug_lock wq_pool_mutex &wq->mutex &rq->__lock irq_context: 0 pernet_ops_rwsem cpu_hotplug_lock wq_pool_mutex &wq->mutex &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 &sb->s_type->i_mutex_key#10 (work_completion)(&strp->work) irq_context: 0 sb_writers#4 &sbi->s_writepages_rwsem jbd2_handle &ei->i_data_sem &ei->i_es_lock &c->lock irq_context: 0 &sb->s_type->i_mutex_key#10 sk_lock-AF_KCM rcu_read_lock rcu_node_0 irq_context: 0 &sb->s_type->i_mutex_key#10 sk_lock-AF_KCM rcu_read_lock &rq->__lock irq_context: 0 &sb->s_type->i_mutex_key#10 sk_lock-AF_KCM rcu_read_lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 sb_writers#4 &sbi->s_writepages_rwsem jbd2_handle &ei->i_data_sem &sem->wait_lock irq_context: 0 prog_idr_lock fill_pool_map-wait-type-override &c->lock irq_context: 0 prog_idr_lock fill_pool_map-wait-type-override pool_lock#2 irq_context: 0 prog_idr_lock fill_pool_map-wait-type-override &n->list_lock irq_context: 0 prog_idr_lock fill_pool_map-wait-type-override &n->list_lock &c->lock irq_context: 0 prog_idr_lock fill_pool_map-wait-type-override pool_lock irq_context: 0 sb_writers#4 &type->i_mutex_dir_key#3/1 fill_pool_map-wait-type-override &n->list_lock irq_context: 0 sb_writers#4 &type->i_mutex_dir_key#3/1 fill_pool_map-wait-type-override &n->list_lock &c->lock irq_context: 0 sb_writers#4 &sbi->s_writepages_rwsem jbd2_handle &c->lock irq_context: 0 &mm->mmap_lock &anon_vma->rwsem rcu_read_lock rcu_read_lock rcu_node_0 irq_context: 0 &mm->mmap_lock &anon_vma->rwsem rcu_read_lock rcu_read_lock &rq->__lock irq_context: 0 &mm->mmap_lock &anon_vma->rwsem rcu_read_lock rcu_read_lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 &mm->mmap_lock sb_pagefaults mapping.invalidate_lock &ei->i_es_lock key#5 irq_context: 0 &mm->mmap_lock sb_pagefaults mapping.invalidate_lock &ei->i_data_sem &ei->i_es_lock &c->lock irq_context: 0 ppp_mutex rtnl_mutex fs_reclaim mmu_notifier_invalidate_range_start &rq->__lock irq_context: 0 ppp_mutex rtnl_mutex fs_reclaim mmu_notifier_invalidate_range_start &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 rtnl_mutex remove_cache_srcu &rq->__lock &cfs_rq->removed.lock irq_context: 0 (wq_completion)bond5 &rq->__lock irq_context: 0 (wq_completion)bond5 &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)mld (work_completion)(&(&idev->mc_ifc_work)->work) &idev->mc_lock rcu_read_lock rcu_read_lock rcu_read_lock_bh rcu_read_lock &br->multicast_lock &obj_hash[i].lock pool_lock irq_context: 0 sb_writers#4 &type->i_mutex_dir_key#3/1 fill_pool_map-wait-type-override &____s->seqcount#2 irq_context: 0 sb_writers#4 &type->i_mutex_dir_key#3/1 fill_pool_map-wait-type-override &____s->seqcount irq_context: 0 tasklist_lock &sighand->siglock &(&sig->stats_lock)->lock &____s->seqcount#5 fill_pool_map-wait-type-override &____s->seqcount#2 irq_context: 0 ppp_mutex rtnl_mutex fs_reclaim &rq->__lock irq_context: 0 pernet_ops_rwsem rtnl_mutex &idev->mc_lock fill_pool_map-wait-type-override &____s->seqcount#2 irq_context: 0 pernet_ops_rwsem rtnl_mutex &idev->mc_lock fill_pool_map-wait-type-override &____s->seqcount irq_context: 0 lock map_idr_lock &n->list_lock irq_context: 0 lock map_idr_lock &n->list_lock &c->lock irq_context: 0 sb_writers#4 &type->i_mutex_dir_key#3/1 &type->i_mutex_dir_key#3 fill_pool_map-wait-type-override &____s->seqcount#2 irq_context: 0 sb_writers#4 &type->i_mutex_dir_key#3/1 &type->i_mutex_dir_key#3 fill_pool_map-wait-type-override &____s->seqcount irq_context: 0 pernet_ops_rwsem pcpu_alloc_mutex &cfs_rq->removed.lock irq_context: 0 pernet_ops_rwsem pcpu_alloc_mutex &obj_hash[i].lock irq_context: 0 pernet_ops_rwsem proc_inum_ida.xa_lock &n->list_lock irq_context: 0 pernet_ops_rwsem proc_inum_ida.xa_lock &n->list_lock &c->lock irq_context: 0 rtnl_mutex pcpu_alloc_mutex rcu_node_0 irq_context: 0 sb_writers#4 &type->i_mutex_dir_key#3 jbd2_handle &cfs_rq->removed.lock irq_context: 0 sb_writers#4 &sb->s_type->i_mutex_key#8 &sbi->s_writepages_rwsem &rsp->gp_wait &p->pi_lock &cfs_rq->removed.lock irq_context: 0 (wq_completion)bat_events (work_completion)(&(&bat_priv->nc.work)->work) fill_pool_map-wait-type-override rcu_read_lock pool_lock#2 irq_context: 0 (wq_completion)bat_events (work_completion)(&(&bat_priv->nc.work)->work) fill_pool_map-wait-type-override &obj_hash[i].lock irq_context: 0 ppp_mutex rtnl_mutex &idev->mc_lock fill_pool_map-wait-type-override pool_lock#2 irq_context: 0 ppp_mutex rtnl_mutex &idev->mc_lock fill_pool_map-wait-type-override pool_lock irq_context: 0 sb_writers#4 &type->i_mutex_dir_key#3 fill_pool_map-wait-type-override &rq->__lock irq_context: 0 cb_lock genl_mutex rcu_read_lock netlbl_domhsh_lock irq_context: 0 pernet_ops_rwsem rtnl_mutex fill_pool_map-wait-type-override rcu_read_lock rcu_node_0 irq_context: 0 pernet_ops_rwsem rtnl_mutex fill_pool_map-wait-type-override rcu_read_lock &rq->__lock irq_context: 0 pernet_ops_rwsem rtnl_mutex fill_pool_map-wait-type-override rcu_read_lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem rcu_state.barrier_mutex &rcu_state.expedited_wq &p->pi_lock &rq->__lock irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem rcu_state.barrier_mutex &rcu_state.expedited_wq &p->pi_lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)events (work_completion)(&pwq->unbound_release_work) &rcu_state.expedited_wq &p->pi_lock &rq->__lock irq_context: 0 (wq_completion)events (work_completion)(&pwq->unbound_release_work) &rcu_state.expedited_wq &p->pi_lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 pernet_ops_rwsem rtnl_mutex &rcu_state.expedited_wq irq_context: 0 pernet_ops_rwsem rtnl_mutex dev_pm_qos_sysfs_mtx &root->kernfs_rwsem rcu_read_lock &rq->__lock irq_context: 0 pernet_ops_rwsem rtnl_mutex dev_pm_qos_sysfs_mtx &root->kernfs_rwsem rcu_read_lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 pernet_ops_rwsem rtnl_mutex &rcu_state.expedited_wq &p->pi_lock irq_context: 0 pernet_ops_rwsem rtnl_mutex &rcu_state.expedited_wq &p->pi_lock &rq->__lock irq_context: 0 pernet_ops_rwsem rtnl_mutex &rcu_state.expedited_wq &p->pi_lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)events_unbound (work_completion)(&(&kfence_timer)->work) cpu_hotplug_lock jump_label_mutex rcu_node_0 irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem rtnl_mutex rcu_read_lock rcu_read_lock rcu_node_0 irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem rtnl_mutex rcu_read_lock rcu_read_lock &rq->__lock irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem rtnl_mutex rcu_read_lock rcu_read_lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: softirq (&tw->tw_timer) per_cpu_ptr(&cgroup_rstat_cpu_lock, cpu) irq_context: 0 (wq_completion)wg-kex-wg2 (work_completion)(&peer->transmit_handshake_work) &c->lock irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem rtnl_mutex dev_pm_qos_sysfs_mtx &p->pi_lock &cfs_rq->removed.lock irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem rtnl_mutex dev_pm_qos_sysfs_mtx &root->kernfs_rwsem &sem->wait_lock irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem rtnl_mutex dev_pm_qos_sysfs_mtx &root->kernfs_rwsem &rq->__lock irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem rtnl_mutex dev_pm_qos_sysfs_mtx &root->kernfs_rwsem &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 pernet_ops_rwsem dev_pm_qos_sysfs_mtx &root->kernfs_rwsem &rq->__lock irq_context: 0 pernet_ops_rwsem dev_pm_qos_sysfs_mtx &root->kernfs_rwsem &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 pernet_ops_rwsem dev_pm_qos_sysfs_mtx &p->pi_lock &rq->__lock irq_context: 0 pernet_ops_rwsem dev_pm_qos_sysfs_mtx &p->pi_lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem rcu_read_lock rcu_read_lock &pool->lock irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem rcu_read_lock rcu_read_lock &pool->lock &obj_hash[i].lock irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem rcu_read_lock rcu_read_lock &pool->lock &p->pi_lock irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem rcu_read_lock rcu_read_lock &pool->lock &p->pi_lock &rq->__lock irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem rcu_read_lock rcu_read_lock &pool->lock &p->pi_lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem __ip_vs_app_mutex &obj_hash[i].lock pool_lock irq_context: 0 ppp_mutex rtnl_mutex remove_cache_srcu irq_context: 0 ppp_mutex rtnl_mutex remove_cache_srcu quarantine_lock irq_context: 0 ppp_mutex rtnl_mutex remove_cache_srcu &c->lock irq_context: 0 ppp_mutex rtnl_mutex remove_cache_srcu &n->list_lock irq_context: 0 ppp_mutex rtnl_mutex remove_cache_srcu rcu_read_lock pool_lock#2 irq_context: 0 ppp_mutex rtnl_mutex remove_cache_srcu &obj_hash[i].lock irq_context: 0 cb_lock genl_mutex fs_reclaim &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 &smc->clcsock_release_lock &xt[i].mutex irq_context: 0 &smc->clcsock_release_lock &xt[i].mutex &mm->mmap_lock irq_context: 0 &smc->clcsock_release_lock &xt[i].mutex &rq->__lock irq_context: 0 &smc->clcsock_release_lock &xt[i].mutex &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)bond3 &rq->__lock irq_context: 0 (wq_completion)bond3 &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)rcu_gp (work_completion)(&rew->rew_work) rcu_state.exp_wake_mutex &rcu_state.expedited_wq irq_context: 0 (wq_completion)rcu_gp (work_completion)(&rew->rew_work) rcu_state.exp_wake_mutex &rcu_state.expedited_wq &p->pi_lock irq_context: 0 (wq_completion)rcu_gp (work_completion)(&rew->rew_work) rcu_state.exp_wake_mutex &rcu_state.expedited_wq &p->pi_lock &rq->__lock irq_context: 0 (wq_completion)rcu_gp (work_completion)(&rew->rew_work) rcu_state.exp_wake_mutex &rcu_state.expedited_wq &p->pi_lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 pernet_ops_rwsem ovs_mutex nf_ct_proto_mutex nf_hook_mutex fs_reclaim mmu_notifier_invalidate_range_start &rq->__lock irq_context: 0 pernet_ops_rwsem ovs_mutex nf_ct_proto_mutex nf_hook_mutex fs_reclaim mmu_notifier_invalidate_range_start &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 pernet_ops_rwsem rtnl_mutex sysctl_lock fill_pool_map-wait-type-override &n->list_lock irq_context: 0 pernet_ops_rwsem devices_rwsem rdma_nets_rwsem &device->compat_devs_mutex &root->kernfs_rwsem &p->pi_lock &rq->__lock irq_context: 0 (wq_completion)events (work_completion)(&pwq->unbound_release_work) rcu_state.exp_mutex &obj_hash[i].lock pool_lock irq_context: 0 pernet_ops_rwsem devices_rwsem rdma_nets_rwsem &device->compat_devs_mutex &root->kernfs_rwsem &p->pi_lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 pernet_ops_rwsem rtnl_mutex sysctl_lock fill_pool_map-wait-type-override &n->list_lock &c->lock irq_context: softirq rcu_callback rcu_read_lock &pool->lock fill_pool_map-wait-type-override pool_lock#2 irq_context: 0 (wq_completion)wg-crypt-wg0 (work_completion)(&({ do { const void *__vpp_verify = (typeof((worker) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __asm__ ("" : "=r"(__ptr) : "0"((typeof(*((worker))) *)((worker)))); (typeof((typeof(*((worker))) *)((worker)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); })->work) rcu_read_lock &pool->lock fill_pool_map-wait-type-override pool_lock#2 irq_context: softirq rcu_callback rcu_read_lock &pool->lock fill_pool_map-wait-type-override &c->lock irq_context: 0 (wq_completion)wg-crypt-wg0 (work_completion)(&({ do { const void *__vpp_verify = (typeof((worker) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __asm__ ("" : "=r"(__ptr) : "0"((typeof(*((worker))) *)((worker)))); (typeof((typeof(*((worker))) *)((worker)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); })->work) rcu_read_lock &pool->lock fill_pool_map-wait-type-override &c->lock irq_context: softirq rcu_callback rcu_read_lock &pool->lock fill_pool_map-wait-type-override &n->list_lock irq_context: 0 (wq_completion)wg-crypt-wg0 (work_completion)(&({ do { const void *__vpp_verify = (typeof((worker) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __asm__ ("" : "=r"(__ptr) : "0"((typeof(*((worker))) *)((worker)))); (typeof((typeof(*((worker))) *)((worker)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); })->work) rcu_read_lock &pool->lock fill_pool_map-wait-type-override &n->list_lock irq_context: softirq rcu_callback rcu_read_lock &pool->lock fill_pool_map-wait-type-override &n->list_lock &c->lock irq_context: 0 (wq_completion)wg-crypt-wg0 (work_completion)(&({ do { const void *__vpp_verify = (typeof((worker) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __asm__ ("" : "=r"(__ptr) : "0"((typeof(*((worker))) *)((worker)))); (typeof((typeof(*((worker))) *)((worker)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); })->work) rcu_read_lock &pool->lock fill_pool_map-wait-type-override &n->list_lock &c->lock irq_context: softirq rcu_callback rcu_read_lock &pool->lock fill_pool_map-wait-type-override pool_lock irq_context: 0 (wq_completion)wg-crypt-wg0 (work_completion)(&({ do { const void *__vpp_verify = (typeof((worker) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __asm__ ("" : "=r"(__ptr) : "0"((typeof(*((worker))) *)((worker)))); (typeof((typeof(*((worker))) *)((worker)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); })->work) rcu_read_lock &pool->lock fill_pool_map-wait-type-override pool_lock irq_context: 0 rtnl_mutex batched_entropy_u8.lock crngs.lock irq_context: 0 pernet_ops_rwsem rtnl_mutex uevent_sock_mutex mmu_notifier_invalidate_range_start &rq->__lock irq_context: 0 pernet_ops_rwsem rtnl_mutex uevent_sock_mutex mmu_notifier_invalidate_range_start &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem rtnl_mutex dev_pm_qos_sysfs_mtx &root->kernfs_rwsem rcu_read_lock &rq->__lock irq_context: 0 pernet_ops_rwsem rtnl_mutex bpf_devs_lock &rq->__lock irq_context: 0 pernet_ops_rwsem rtnl_mutex bpf_devs_lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem rtnl_mutex dev_pm_qos_sysfs_mtx &root->kernfs_rwsem rcu_read_lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 pernet_ops_rwsem &root->kernfs_rwsem rcu_read_lock &rq->__lock irq_context: 0 pernet_ops_rwsem &root->kernfs_rwsem rcu_read_lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 pernet_ops_rwsem rtnl_mutex &root->kernfs_rwsem rcu_read_lock &rcu_state.gp_wq irq_context: 0 pernet_ops_rwsem rtnl_mutex &root->kernfs_rwsem rcu_read_lock &rcu_state.gp_wq &p->pi_lock irq_context: 0 pernet_ops_rwsem rtnl_mutex &root->kernfs_rwsem rcu_read_lock &rcu_state.gp_wq &p->pi_lock &rq->__lock irq_context: 0 pernet_ops_rwsem rtnl_mutex &root->kernfs_rwsem rcu_read_lock &rcu_state.gp_wq &p->pi_lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: softirq (&ifibss->timer) rcu_read_lock &pool->lock &p->pi_lock &rq->__lock irq_context: softirq (&ifibss->timer) rcu_read_lock &pool->lock &p->pi_lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)events (work_completion)(&rdev->wiphy_work) &rdev->wiphy.mtx &wdev->mtx &local->sta_mtx batched_entropy_u8.lock irq_context: 0 (wq_completion)events (work_completion)(&rdev->wiphy_work) &rdev->wiphy.mtx &wdev->mtx &local->sta_mtx kfence_freelist_lock irq_context: 0 (wq_completion)events (work_completion)(&rdev->wiphy_work) &rdev->wiphy.mtx &wdev->mtx &local->sta_mtx rcu_read_lock &local->queue_stop_reason_lock irq_context: 0 (wq_completion)events (work_completion)(&rdev->wiphy_work) &rdev->wiphy.mtx &wdev->mtx &local->sta_mtx rcu_read_lock tk_core.seq.seqcount irq_context: 0 (wq_completion)events (work_completion)(&rdev->wiphy_work) &rdev->wiphy.mtx &wdev->mtx &local->sta_mtx rcu_read_lock hwsim_radio_lock irq_context: 0 (wq_completion)events (work_completion)(&rdev->wiphy_work) &rdev->wiphy.mtx &wdev->mtx &local->sta_mtx rcu_read_lock &list->lock#19 irq_context: 0 (wq_completion)events (work_completion)(&rdev->wiphy_work) &rdev->wiphy.mtx &wdev->mtx &local->sta_mtx rcu_state.exp_mutex &rnp->exp_wq[0] irq_context: 0 (wq_completion)events (work_completion)(&rdev->wiphy_work) &rdev->wiphy.mtx &wdev->mtx &local->sta_mtx remove_cache_srcu irq_context: 0 (wq_completion)events (work_completion)(&rdev->wiphy_work) &rdev->wiphy.mtx &wdev->mtx &local->sta_mtx remove_cache_srcu quarantine_lock irq_context: 0 (wq_completion)events (work_completion)(&rdev->wiphy_work) &rdev->wiphy.mtx &wdev->mtx &local->sta_mtx remove_cache_srcu &c->lock irq_context: 0 (wq_completion)events (work_completion)(&rdev->wiphy_work) &rdev->wiphy.mtx &wdev->mtx &local->sta_mtx remove_cache_srcu &n->list_lock irq_context: 0 (wq_completion)events (work_completion)(&rdev->wiphy_work) &rdev->wiphy.mtx &wdev->mtx &local->sta_mtx remove_cache_srcu rcu_read_lock pool_lock#2 irq_context: 0 (wq_completion)events (work_completion)(&rdev->wiphy_work) &rdev->wiphy.mtx &wdev->mtx &local->sta_mtx remove_cache_srcu &obj_hash[i].lock irq_context: 0 (wq_completion)events (work_completion)(&rdev->wiphy_work) &rdev->wiphy.mtx &wdev->mtx &local->sta_mtx remove_cache_srcu pool_lock#2 irq_context: 0 (wq_completion)events (work_completion)(&rdev->wiphy_work) &rdev->wiphy.mtx &wdev->mtx &local->sta_mtx remove_cache_srcu rcu_read_lock rcu_node_0 irq_context: 0 (wq_completion)events (work_completion)(&rdev->wiphy_work) &rdev->wiphy.mtx &wdev->mtx &local->sta_mtx remove_cache_srcu rcu_read_lock &rq->__lock irq_context: 0 (wq_completion)events (work_completion)(&rdev->wiphy_work) &rdev->wiphy.mtx &wdev->mtx &local->sta_mtx remove_cache_srcu rcu_read_lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)ipv6_addrconf (work_completion)(&(&ifa->dad_work)->work) rtnl_mutex &idev->mc_lock rcu_read_lock rcu_read_lock rcu_read_lock_bh rcu_read_lock &br->multicast_lock &c->lock irq_context: 0 rtnl_mutex &tb->tb6_lock (console_sem).lock irq_context: 0 rtnl_mutex &tb->tb6_lock console_lock console_srcu console_owner_lock irq_context: 0 rtnl_mutex &tb->tb6_lock console_lock console_srcu console_owner irq_context: 0 rtnl_mutex &tb->tb6_lock console_lock console_srcu console_owner &port_lock_key irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem nl_table_wait.lock &p->pi_lock irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem nl_table_wait.lock &p->pi_lock &rq->__lock irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem nl_table_wait.lock &p->pi_lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 rtnl_mutex &tb->tb6_lock console_lock console_srcu console_owner console_owner_lock irq_context: 0 sb_writers#4 &type->i_mutex_dir_key#3/1 sb_internal jbd2_handle &ei->i_data_sem remove_cache_srcu rcu_read_lock rcu_node_0 irq_context: 0 sb_writers#4 &type->i_mutex_dir_key#3/1 sb_internal jbd2_handle &ei->i_data_sem remove_cache_srcu rcu_read_lock &rq->__lock irq_context: 0 sb_writers#4 &type->i_mutex_dir_key#3/1 sb_internal jbd2_handle &ei->i_data_sem remove_cache_srcu rcu_read_lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 rtnl_mutex rcu_read_lock rcu_read_lock &pool->lock &p->pi_lock &rq->__lock &cfs_rq->removed.lock irq_context: 0 &mm->mmap_lock fill_pool_map-wait-type-override rcu_read_lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)wg-kex-wg2#11 (work_completion)(&peer->transmit_handshake_work) &wg->static_identity.lock &handshake->lock rcu_read_lock_bh batched_entropy_u32.lock crngs.lock irq_context: 0 (wq_completion)wg-kex-wg2#11 (work_completion)(&peer->transmit_handshake_work) &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)mld (work_completion)(&(&idev->mc_ifc_work)->work) &idev->mc_lock remove_cache_srcu &rq->__lock irq_context: 0 (wq_completion)mld (work_completion)(&(&idev->mc_ifc_work)->work) &idev->mc_lock remove_cache_srcu &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 rtnl_mutex &idev->mc_lock &____s->seqcount#2 irq_context: 0 lock prog_idr_lock batched_entropy_u8.lock irq_context: 0 lock prog_idr_lock kfence_freelist_lock irq_context: 0 pernet_ops_rwsem ovs_mutex nf_ct_proto_mutex defrag4_mutex fill_pool_map-wait-type-override pool_lock#2 irq_context: 0 (wq_completion)mld (work_completion)(&(&idev->mc_ifc_work)->work) &idev->mc_lock rcu_read_lock rcu_read_lock rcu_read_lock_bh rcu_read_lock &br->multicast_lock fill_pool_map-wait-type-override pool_lock#2 irq_context: 0 pernet_ops_rwsem ovs_mutex nf_ct_proto_mutex defrag4_mutex fill_pool_map-wait-type-override &c->lock irq_context: 0 (wq_completion)mld (work_completion)(&(&idev->mc_ifc_work)->work) &idev->mc_lock rcu_read_lock rcu_read_lock rcu_read_lock_bh rcu_read_lock &br->multicast_lock fill_pool_map-wait-type-override &c->lock irq_context: 0 pernet_ops_rwsem ovs_mutex nf_ct_proto_mutex defrag4_mutex fill_pool_map-wait-type-override pool_lock irq_context: 0 (wq_completion)mld (work_completion)(&(&idev->mc_ifc_work)->work) &idev->mc_lock rcu_read_lock rcu_read_lock rcu_read_lock_bh rcu_read_lock &br->multicast_lock fill_pool_map-wait-type-override &n->list_lock irq_context: 0 (wq_completion)mld (work_completion)(&(&idev->mc_ifc_work)->work) &idev->mc_lock rcu_read_lock rcu_read_lock rcu_read_lock_bh rcu_read_lock &br->multicast_lock fill_pool_map-wait-type-override &n->list_lock &c->lock irq_context: 0 (wq_completion)mld (work_completion)(&(&idev->mc_ifc_work)->work) &idev->mc_lock rcu_read_lock rcu_read_lock rcu_read_lock_bh rcu_read_lock &br->multicast_lock fill_pool_map-wait-type-override pool_lock irq_context: 0 &mm->mmap_lock remove_cache_srcu rcu_read_lock &rcu_state.expedited_wq irq_context: 0 &mm->mmap_lock remove_cache_srcu rcu_read_lock &rcu_state.expedited_wq &p->pi_lock irq_context: 0 &mm->mmap_lock remove_cache_srcu rcu_read_lock &rcu_state.expedited_wq &p->pi_lock &rq->__lock irq_context: 0 &mm->mmap_lock remove_cache_srcu rcu_read_lock &rcu_state.expedited_wq &p->pi_lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 pernet_ops_rwsem rtnl_mutex &idev->mc_lock _xmit_ETHER &obj_hash[i].lock pool_lock irq_context: 0 (wq_completion)wg-crypt-wg1#3 (work_completion)(&({ do { const void *__vpp_verify = (typeof((worker) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __asm__ ("" : "=r"(__ptr) : "0"((typeof(*((worker))) *)((worker)))); (typeof((typeof(*((worker))) *)((worker)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); })->work) &rcu_state.expedited_wq irq_context: 0 (wq_completion)wg-crypt-wg1#3 (work_completion)(&({ do { const void *__vpp_verify = (typeof((worker) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __asm__ ("" : "=r"(__ptr) : "0"((typeof(*((worker))) *)((worker)))); (typeof((typeof(*((worker))) *)((worker)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); })->work) &rcu_state.expedited_wq &p->pi_lock irq_context: 0 (wq_completion)wg-crypt-wg1#3 (work_completion)(&({ do { const void *__vpp_verify = (typeof((worker) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __asm__ ("" : "=r"(__ptr) : "0"((typeof(*((worker))) *)((worker)))); (typeof((typeof(*((worker))) *)((worker)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); })->work) &rcu_state.expedited_wq &p->pi_lock &rq->__lock irq_context: 0 (wq_completion)wg-crypt-wg1#3 (work_completion)(&({ do { const void *__vpp_verify = (typeof((worker) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __asm__ ("" : "=r"(__ptr) : "0"((typeof(*((worker))) *)((worker)))); (typeof((typeof(*((worker))) *)((worker)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); })->work) &rcu_state.expedited_wq &p->pi_lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)wg-kex-wg1#9 (work_completion)(&peer->transmit_handshake_work) &peer->endpoint_lock rcu_read_lock_bh &obj_hash[i].lock irq_context: 0 (wq_completion)wg-crypt-wg0 (work_completion)(&peer->transmit_packet_work) &peer->endpoint_lock rcu_read_lock_bh &____s->seqcount#2 irq_context: 0 (wq_completion)wg-crypt-wg0 (work_completion)(&peer->transmit_packet_work) &peer->endpoint_lock rcu_read_lock_bh &____s->seqcount irq_context: 0 sb_writers#5 &type->i_mutex_dir_key#5 pgd_lock irq_context: 0 sb_writers#5 &type->i_mutex_dir_key#5 stock_lock irq_context: 0 sb_writers#5 &type->i_mutex_dir_key#5 rcu_read_lock pool_lock#2 irq_context: 0 sb_writers#5 &type->i_mutex_dir_key#5 key irq_context: 0 sb_writers#5 &type->i_mutex_dir_key#5 pcpu_lock irq_context: 0 sb_writers#5 &type->i_mutex_dir_key#5 percpu_counters_lock irq_context: 0 sb_writers#5 &type->i_mutex_dir_key#5 pcpu_lock stock_lock irq_context: 0 (wq_completion)bat_events (work_completion)(&(&bat_priv->bla.work)->work) rcu_read_lock rcu_read_lock rcu_node_0 irq_context: 0 (wq_completion)bat_events (work_completion)(&(&bat_priv->bla.work)->work) rcu_read_lock rcu_read_lock &rq->__lock irq_context: 0 pernet_ops_rwsem devices_rwsem rdma_nets_rwsem &device->compat_devs_mutex uevent_sock_mutex nl_table_wait.lock &p->pi_lock irq_context: 0 cb_lock nlk_cb_mutex-GENERIC genl_mutex rtnl_mutex fs_reclaim irq_context: 0 cb_lock nlk_cb_mutex-GENERIC genl_mutex rtnl_mutex fs_reclaim mmu_notifier_invalidate_range_start irq_context: 0 cb_lock nlk_cb_mutex-GENERIC genl_mutex rtnl_mutex pool_lock#2 irq_context: 0 rcu_read_lock rcu_read_lock_bh dev->qdisc_tx_busylock ?: &qdisc_tx_busylock _xmit_ETHER#2 irq_context: 0 rcu_read_lock rcu_read_lock_bh dev->qdisc_tx_busylock ?: &qdisc_tx_busylock _xmit_ETHER#2 &obj_hash[i].lock irq_context: 0 rcu_read_lock rcu_read_lock_bh dev->qdisc_tx_busylock ?: &qdisc_tx_busylock _xmit_ETHER#2 pool_lock#2 irq_context: 0 (wq_completion)events (work_completion)(&nlk->work) genl_mutex irq_context: 0 (wq_completion)events (work_completion)(&nlk->work) genl_mutex &obj_hash[i].lock irq_context: 0 (wq_completion)events (work_completion)(&nlk->work) genl_mutex pool_lock#2 irq_context: 0 (wq_completion)ipv6_addrconf (work_completion)(&(&ifa->dad_work)->work) rtnl_mutex rcu_read_lock rcu_read_lock rcu_read_lock_bh dev->qdisc_tx_busylock ?: &qdisc_tx_busylock _xmit_ETHER#2 &____s->seqcount irq_context: 0 (wq_completion)ipv6_addrconf (work_completion)(&(&ifa->dad_work)->work) rtnl_mutex rcu_read_lock rcu_read_lock rcu_read_lock_bh dev->qdisc_tx_busylock ?: &qdisc_tx_busylock _xmit_ETHER#2 rcu_read_lock pool_lock#2 irq_context: 0 rtnl_mutex _xmit_PIMREG irq_context: 0 rtnl_mutex mrt_lock irq_context: 0 rtnl_mutex mrt_lock &c->lock irq_context: 0 rtnl_mutex mrt_lock pool_lock#2 irq_context: 0 rtnl_mutex mrt_lock &dir->lock#2 irq_context: 0 rtnl_mutex &dev->tx_global_lock _xmit_PIMREG#2 irq_context: 0 &sb->s_type->i_mutex_key#10 fill_pool_map-wait-type-override &____s->seqcount#2 irq_context: 0 &sb->s_type->i_mutex_key#10 fill_pool_map-wait-type-override &____s->seqcount irq_context: 0 sb_writers#4 &type->i_mutex_dir_key#3/1 rcu_read_lock &dentry->d_lock &p->pi_lock irq_context: 0 sb_writers#4 &type->i_mutex_dir_key#3/1 rcu_read_lock &dentry->d_lock &p->pi_lock &rq->__lock irq_context: 0 sb_writers#4 &type->i_mutex_dir_key#3/1 rcu_read_lock &dentry->d_lock &p->pi_lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 pernet_ops_rwsem ovs_mutex nf_ct_proto_mutex nf_hook_mutex per_cpu_ptr(&cgroup_rstat_cpu_lock, cpu) irq_context: 0 pernet_ops_rwsem ovs_mutex nf_ct_proto_mutex nf_hook_mutex stock_lock rcu_read_lock per_cpu_ptr(&cgroup_rstat_cpu_lock, cpu) irq_context: 0 (wq_completion)wg-kex-wg1#2 (work_completion)(&peer->transmit_handshake_work) &peer->endpoint_lock rcu_read_lock_bh &obj_hash[i].lock irq_context: 0 &sb->s_type->i_lock_key#16 irq_context: 0 &sb->s_type->i_lock_key#16 &dentry->d_lock irq_context: 0 &mm->mmap_lock &sb->s_type->i_mutex_key#21 irq_context: 0 &mm->mmap_lock &sb->s_type->i_mutex_key#21 &rq->__lock irq_context: 0 &mm->mmap_lock &sb->s_type->i_mutex_key#21 &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 &mm->mmap_lock &sb->s_type->i_mutex_key#21 tk_core.seq.seqcount irq_context: 0 &mm->mmap_lock &sb->s_type->i_mutex_key#21 fs_reclaim irq_context: 0 &mm->mmap_lock &sb->s_type->i_mutex_key#21 fs_reclaim mmu_notifier_invalidate_range_start irq_context: 0 &mm->mmap_lock &sb->s_type->i_mutex_key#21 &c->lock irq_context: 0 &mm->mmap_lock &sb->s_type->i_mutex_key#21 pool_lock#2 irq_context: 0 &mm->mmap_lock &sb->s_type->i_mutex_key#21 &resv_map->lock irq_context: 0 &mm->mmap_lock &sb->s_type->i_mutex_key#21 hugetlb_lock irq_context: 0 &mm->mmap_lock &hugetlbfs_i_mmap_rwsem_key irq_context: 0 &mm->mmap_lock &hugetlbfs_i_mmap_rwsem_key &obj_hash[i].lock irq_context: 0 &mm->mmap_lock &hugetlbfs_i_mmap_rwsem_key &rq->__lock irq_context: 0 &mm->mmap_lock &hugetlbfs_i_mmap_rwsem_key &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 &mm->mmap_lock &hugetlbfs_i_mmap_rwsem_key pool_lock#2 irq_context: 0 &mm->mmap_lock &vma_lock->rw_sema irq_context: 0 &mm->mmap_lock &vma_lock->rw_sema ptlock_ptr(page) irq_context: 0 &mm->mmap_lock &vma_lock->rw_sema &rq->__lock irq_context: 0 &mm->mmap_lock &vma_lock->rw_sema &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 &mm->mmap_lock &hugetlb_fault_mutex_table[i] irq_context: 0 &mm->mmap_lock &hugetlb_fault_mutex_table[i] &vma_lock->rw_sema irq_context: 0 &mm->mmap_lock &hugetlb_fault_mutex_table[i] &vma_lock->rw_sema &resv_map->lock irq_context: 0 &mm->mmap_lock &hugetlb_fault_mutex_table[i] &vma_lock->rw_sema fs_reclaim irq_context: 0 &mm->mmap_lock &hugetlb_fault_mutex_table[i] &vma_lock->rw_sema fs_reclaim mmu_notifier_invalidate_range_start irq_context: 0 &mm->mmap_lock &hugetlb_fault_mutex_table[i] &vma_lock->rw_sema pool_lock#2 irq_context: 0 &mm->mmap_lock &hugetlb_fault_mutex_table[i] &vma_lock->rw_sema hugetlb_lock irq_context: 0 &mm->mmap_lock &hugetlb_fault_mutex_table[i] &vma_lock->rw_sema hugetlb_lock &____s->seqcount#2 irq_context: 0 &mm->mmap_lock &hugetlb_fault_mutex_table[i] &vma_lock->rw_sema &rq->__lock irq_context: 0 &mm->mmap_lock &hugetlb_fault_mutex_table[i] &vma_lock->rw_sema &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 &mm->mmap_lock &hugetlb_fault_mutex_table[i] &vma_lock->rw_sema &xa->xa_lock#9 irq_context: 0 &mm->mmap_lock &hugetlb_fault_mutex_table[i] &vma_lock->rw_sema &sb->s_type->i_lock_key#16 irq_context: 0 &mm->mmap_lock &hugetlb_fault_mutex_table[i] &vma_lock->rw_sema ptlock_ptr(page) irq_context: 0 &mm->mmap_lock &vma_lock->rw_sema irq_context: 0 &mm->mmap_lock &vma_lock->rw_sema &hugetlbfs_i_mmap_rwsem_key irq_context: 0 &mm->mmap_lock &vma_lock->rw_sema &hugetlbfs_i_mmap_rwsem_key ptlock_ptr(page) irq_context: 0 &mm->mmap_lock &vma_lock->rw_sema &hugetlbfs_i_mmap_rwsem_key &rq->__lock irq_context: 0 &mm->mmap_lock &vma_lock->rw_sema &hugetlbfs_i_mmap_rwsem_key &rq->__lock &cfs_rq->removed.lock irq_context: 0 &mm->mmap_lock &vma_lock->rw_sema &hugetlbfs_i_mmap_rwsem_key &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 &mm->mmap_lock &hugetlbfs_i_mmap_rwsem_key irq_context: 0 &mm->mmap_lock &hugetlbfs_i_mmap_rwsem_key &obj_hash[i].lock irq_context: 0 &mm->mmap_lock &hugetlbfs_i_mmap_rwsem_key pool_lock#2 irq_context: 0 &hugetlb_fault_mutex_table[i] irq_context: 0 &hugetlb_fault_mutex_table[i] &sb->s_type->i_lock_key#16 irq_context: 0 &hugetlb_fault_mutex_table[i] &sb->s_type->i_lock_key#16 &xa->xa_lock#9 irq_context: 0 hugetlb_lock irq_context: 0 &resv_map->lock irq_context: 0 &resv_map->lock &obj_hash[i].lock irq_context: 0 &resv_map->lock pool_lock#2 irq_context: 0 rtnl_mutex pcpu_alloc_mutex pcpu_alloc_mutex.wait_lock irq_context: 0 rtnl_mutex &ndev->lock &ifa->lock batched_entropy_u32.lock crngs.lock irq_context: 0 &f->f_pos_lock sb_writers#4 &sb->s_type->i_mutex_key#8 jbd2_handle &ei->i_data_sem &lg->lg_mutex rcu_read_lock rcu_node_0 irq_context: 0 &f->f_pos_lock sb_writers#4 &sb->s_type->i_mutex_key#8 jbd2_handle &ei->i_data_sem &lg->lg_mutex rcu_read_lock &rq->__lock irq_context: 0 &f->f_pos_lock sb_writers#4 &sb->s_type->i_mutex_key#8 jbd2_handle &ei->i_data_sem &lg->lg_mutex rcu_read_lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 sb_writers#4 &type->i_mutex_dir_key#3/1 jbd2_handle &ei->i_data_sem mmu_notifier_invalidate_range_start rcu_node_0 irq_context: 0 sb_writers#4 &type->i_mutex_dir_key#3/1 jbd2_handle &ei->i_data_sem mmu_notifier_invalidate_range_start &rcu_state.expedited_wq irq_context: 0 sb_writers#4 &type->i_mutex_dir_key#3/1 jbd2_handle &ei->i_data_sem mmu_notifier_invalidate_range_start &rcu_state.expedited_wq &p->pi_lock irq_context: 0 sb_writers#4 &type->i_mutex_dir_key#3/1 jbd2_handle &ei->i_data_sem mmu_notifier_invalidate_range_start &rcu_state.expedited_wq &p->pi_lock &rq->__lock irq_context: 0 sb_writers#4 &type->i_mutex_dir_key#3/1 jbd2_handle &ei->i_data_sem mmu_notifier_invalidate_range_start &rcu_state.expedited_wq &p->pi_lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 rtnl_mutex rcu_read_lock &ul->lock irq_context: 0 &mm->mmap_lock &hugetlb_fault_mutex_table[i] &vma_lock->rw_sema &c->lock irq_context: 0 (wq_completion)ipv6_addrconf (work_completion)(&(&ifa->dad_work)->work) rtnl_mutex quarantine_lock irq_context: 0 cb_lock &cfs_rq->removed.lock irq_context: 0 rtnl_mutex fs_reclaim &cfs_rq->removed.lock irq_context: 0 rtnl_mutex fs_reclaim &obj_hash[i].lock irq_context: softirq (&ndev->rs_timer) rcu_read_lock rcu_read_lock rcu_read_lock_bh dev->qdisc_tx_busylock ?: &qdisc_tx_busylock#2 rcu_read_lock pool_lock#2 irq_context: 0 (wq_completion)rcu_gp (work_completion)(&rew->rew_work) fill_pool_map-wait-type-override &rq->__lock irq_context: 0 (wq_completion)rcu_gp (work_completion)(&rew->rew_work) fill_pool_map-wait-type-override &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 rcu_read_lock_bh &sch->q.lock irq_context: 0 rcu_read_lock_bh &sch->q.lock pool_lock#2 irq_context: 0 fanout_mutex &c->lock irq_context: 0 (wq_completion)wg-crypt-wg1#5 (work_completion)(&peer->transmit_packet_work) &peer->endpoint_lock rcu_read_lock_bh rcu_read_lock &n->list_lock irq_context: 0 (wq_completion)wg-crypt-wg1#5 (work_completion)(&peer->transmit_packet_work) &peer->endpoint_lock rcu_read_lock_bh rcu_read_lock &n->list_lock &c->lock irq_context: 0 sb_writers#3 &sb->s_type->i_mutex_key#9 rcu_read_lock &rcu_state.expedited_wq irq_context: 0 sb_writers#3 &sb->s_type->i_mutex_key#9 rcu_read_lock &rcu_state.expedited_wq &p->pi_lock irq_context: 0 sb_writers#3 &sb->s_type->i_mutex_key#9 rcu_read_lock &rcu_state.expedited_wq &p->pi_lock &rq->__lock irq_context: 0 sb_writers#3 &sb->s_type->i_mutex_key#9 rcu_read_lock &rcu_state.expedited_wq &p->pi_lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 rtnl_mutex &dev->tx_global_lock _xmit_IPGRE#2 irq_context: 0 rtnl_mutex fib_info_lock &obj_hash[i].lock pool_lock irq_context: 0 rtnl_mutex rcu_read_lock &net->sctp.local_addr_lock &net->sctp.addr_wq_lock rcu_read_lock pool_lock#2 irq_context: 0 rtnl_mutex rcu_read_lock rcu_read_lock &pool->lock fill_pool_map-wait-type-override pool_lock#2 irq_context: 0 rtnl_mutex rcu_read_lock rcu_read_lock &pool->lock fill_pool_map-wait-type-override pool_lock irq_context: 0 pernet_ops_rwsem fill_pool_map-wait-type-override rcu_read_lock rcu_node_0 irq_context: 0 (wq_completion)events_power_efficient (work_completion)(&(&hinfo->gc_work)->work) &obj_hash[i].lock pool_lock irq_context: 0 &nft_net->commit_mutex &rnp->exp_wq[1] irq_context: 0 (wq_completion)events (work_completion)(&(&hwstats->traffic_dw)->work) &cfs_rq->removed.lock irq_context: 0 (wq_completion)ipv6_addrconf &rq->__lock irq_context: 0 (wq_completion)ipv6_addrconf &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 fanout_mutex &n->list_lock irq_context: 0 fanout_mutex &n->list_lock &c->lock irq_context: 0 fanout_mutex &rq->__lock irq_context: 0 (wq_completion)events (work_completion)(&aux->work) rcu_read_lock &pool->lock irq_context: 0 (wq_completion)events (work_completion)(&aux->work) rcu_read_lock &pool->lock &obj_hash[i].lock irq_context: 0 (wq_completion)events_power_efficient (gc_work).work fill_pool_map-wait-type-override pool_lock#2 irq_context: 0 (wq_completion)events_power_efficient (gc_work).work fill_pool_map-wait-type-override pool_lock irq_context: 0 (wq_completion)wg-crypt-wg0#5 (work_completion)(&peer->transmit_packet_work) &peer->endpoint_lock rcu_read_lock_bh rcu_read_lock rcu_read_lock_bh rcu_read_lock &n->list_lock irq_context: 0 (wq_completion)wg-crypt-wg0#5 (work_completion)(&peer->transmit_packet_work) &peer->endpoint_lock rcu_read_lock_bh rcu_read_lock rcu_read_lock_bh rcu_read_lock &n->list_lock &c->lock irq_context: 0 (wq_completion)wg-crypt-wg0 (work_completion)(&peer->transmit_packet_work) &peer->endpoint_lock rcu_read_lock_bh &meta->lock irq_context: 0 (wq_completion)wg-crypt-wg0 (work_completion)(&peer->transmit_packet_work) &peer->endpoint_lock rcu_read_lock_bh kfence_freelist_lock irq_context: 0 sk_lock-AF_BLUETOOTH-BTPROTO_RFCOMM &mm->mmap_lock irq_context: 0 sk_lock-AF_BLUETOOTH-BTPROTO_RFCOMM fs_reclaim irq_context: 0 sk_lock-AF_BLUETOOTH-BTPROTO_RFCOMM fs_reclaim mmu_notifier_invalidate_range_start irq_context: 0 sk_lock-AF_BLUETOOTH-BTPROTO_RFCOMM fs_reclaim mmu_notifier_invalidate_range_start &rq->__lock irq_context: 0 sk_lock-AF_BLUETOOTH-BTPROTO_RFCOMM fs_reclaim mmu_notifier_invalidate_range_start &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 sk_lock-AF_BLUETOOTH-BTPROTO_RFCOMM batched_entropy_u8.lock irq_context: 0 sk_lock-AF_BLUETOOTH-BTPROTO_RFCOMM kfence_freelist_lock irq_context: 0 sk_lock-AF_BLUETOOTH-BTPROTO_RFCOMM rfcomm_dev_lock irq_context: 0 sk_lock-AF_BLUETOOTH-BTPROTO_RFCOMM &obj_hash[i].lock irq_context: 0 sk_lock-AF_BLUETOOTH-BTPROTO_RFCOMM &meta->lock irq_context: 0 rtnl_mutex &xa->xa_lock#3 &n->list_lock irq_context: 0 rtnl_mutex &xa->xa_lock#3 &n->list_lock &c->lock irq_context: 0 rtnl_mutex uevent_sock_mutex nl_table_wait.lock &p->pi_lock &cfs_rq->removed.lock irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem k-sk_lock-AF_INET &rq->__lock irq_context: 0 (wq_completion)wg-kex-wg2#8 (work_completion)(&({ do { const void *__vpp_verify = (typeof((worker) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __asm__ ("" : "=r"(__ptr) : "0"((typeof(*((worker))) *)((worker)))); (typeof((typeof(*((worker))) *)((worker)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); })->work) &handshake->lock remove_cache_srcu irq_context: 0 (wq_completion)wg-kex-wg2#8 (work_completion)(&({ do { const void *__vpp_verify = (typeof((worker) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __asm__ ("" : "=r"(__ptr) : "0"((typeof(*((worker))) *)((worker)))); (typeof((typeof(*((worker))) *)((worker)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); })->work) &handshake->lock remove_cache_srcu quarantine_lock irq_context: 0 (wq_completion)wg-kex-wg2#8 (work_completion)(&({ do { const void *__vpp_verify = (typeof((worker) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __asm__ ("" : "=r"(__ptr) : "0"((typeof(*((worker))) *)((worker)))); (typeof((typeof(*((worker))) *)((worker)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); })->work) &handshake->lock remove_cache_srcu &c->lock irq_context: 0 (wq_completion)wg-kex-wg2#8 (work_completion)(&({ do { const void *__vpp_verify = (typeof((worker) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __asm__ ("" : "=r"(__ptr) : "0"((typeof(*((worker))) *)((worker)))); (typeof((typeof(*((worker))) *)((worker)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); })->work) &handshake->lock remove_cache_srcu &n->list_lock irq_context: 0 (wq_completion)wg-kex-wg2#8 (work_completion)(&({ do { const void *__vpp_verify = (typeof((worker) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __asm__ ("" : "=r"(__ptr) : "0"((typeof(*((worker))) *)((worker)))); (typeof((typeof(*((worker))) *)((worker)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); })->work) &handshake->lock remove_cache_srcu rcu_read_lock pool_lock#2 irq_context: 0 (wq_completion)wg-kex-wg2#8 (work_completion)(&({ do { const void *__vpp_verify = (typeof((worker) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __asm__ ("" : "=r"(__ptr) : "0"((typeof(*((worker))) *)((worker)))); (typeof((typeof(*((worker))) *)((worker)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); })->work) &handshake->lock remove_cache_srcu &obj_hash[i].lock irq_context: 0 (wq_completion)wg-kex-wg2#8 (work_completion)(&({ do { const void *__vpp_verify = (typeof((worker) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __asm__ ("" : "=r"(__ptr) : "0"((typeof(*((worker))) *)((worker)))); (typeof((typeof(*((worker))) *)((worker)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); })->work) &handshake->lock remove_cache_srcu pool_lock#2 irq_context: 0 (wq_completion)wg-kex-wg2#9 (work_completion)(&peer->transmit_handshake_work) &peer->endpoint_lock rcu_read_lock_bh &____s->seqcount#2 irq_context: 0 (wq_completion)wg-kex-wg2#9 (work_completion)(&peer->transmit_handshake_work) &peer->endpoint_lock rcu_read_lock_bh &____s->seqcount irq_context: 0 (wq_completion)wg-kex-wg1#2 (work_completion)(&peer->transmit_handshake_work) &n->list_lock irq_context: 0 (wq_completion)wg-kex-wg1#2 (work_completion)(&peer->transmit_handshake_work) &n->list_lock &c->lock irq_context: 0 (wq_completion)wg-kex-wg1#6 (work_completion)(&({ do { const void *__vpp_verify = (typeof((worker) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __asm__ ("" : "=r"(__ptr) : "0"((typeof(*((worker))) *)((worker)))); (typeof((typeof(*((worker))) *)((worker)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); })->work) &handshake->lock &n->list_lock irq_context: 0 (wq_completion)wg-kex-wg1#6 (work_completion)(&({ do { const void *__vpp_verify = (typeof((worker) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __asm__ ("" : "=r"(__ptr) : "0"((typeof(*((worker))) *)((worker)))); (typeof((typeof(*((worker))) *)((worker)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); })->work) &handshake->lock &n->list_lock &c->lock irq_context: 0 (wq_completion)wg-kex-wg1#6 (work_completion)(&({ do { const void *__vpp_verify = (typeof((worker) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __asm__ ("" : "=r"(__ptr) : "0"((typeof(*((worker))) *)((worker)))); (typeof((typeof(*((worker))) *)((worker)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); })->work) &handshake->lock remove_cache_srcu irq_context: 0 (wq_completion)wg-kex-wg1#6 (work_completion)(&({ do { const void *__vpp_verify = (typeof((worker) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __asm__ ("" : "=r"(__ptr) : "0"((typeof(*((worker))) *)((worker)))); (typeof((typeof(*((worker))) *)((worker)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); })->work) &handshake->lock remove_cache_srcu quarantine_lock irq_context: 0 (wq_completion)wg-kex-wg1#6 (work_completion)(&({ do { const void *__vpp_verify = (typeof((worker) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __asm__ ("" : "=r"(__ptr) : "0"((typeof(*((worker))) *)((worker)))); (typeof((typeof(*((worker))) *)((worker)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); })->work) &handshake->lock remove_cache_srcu &c->lock irq_context: 0 (wq_completion)wg-kex-wg1#6 (work_completion)(&({ do { const void *__vpp_verify = (typeof((worker) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __asm__ ("" : "=r"(__ptr) : "0"((typeof(*((worker))) *)((worker)))); (typeof((typeof(*((worker))) *)((worker)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); })->work) &handshake->lock remove_cache_srcu &n->list_lock irq_context: 0 (wq_completion)wg-kex-wg1#6 (work_completion)(&({ do { const void *__vpp_verify = (typeof((worker) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __asm__ ("" : "=r"(__ptr) : "0"((typeof(*((worker))) *)((worker)))); (typeof((typeof(*((worker))) *)((worker)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); })->work) &handshake->lock remove_cache_srcu &obj_hash[i].lock irq_context: 0 sb_writers#5 &type->i_mutex_dir_key#5 tomoyo_ss pgd_lock irq_context: 0 sb_writers#5 &type->i_mutex_dir_key#5 tomoyo_ss stock_lock irq_context: 0 sb_writers#5 &type->i_mutex_dir_key#5 tomoyo_ss key irq_context: 0 sb_writers#5 &type->i_mutex_dir_key#5 tomoyo_ss pcpu_lock irq_context: 0 sb_writers#5 &type->i_mutex_dir_key#5 tomoyo_ss percpu_counters_lock irq_context: 0 sb_writers#5 &type->i_mutex_dir_key#5 tomoyo_ss pcpu_lock stock_lock irq_context: 0 sk_lock-AF_BLUETOOTH-BTPROTO_RFCOMM &c->lock irq_context: 0 sk_lock-AF_BLUETOOTH-BTPROTO_RFCOMM &n->list_lock irq_context: 0 sk_lock-AF_BLUETOOTH-BTPROTO_RFCOMM &n->list_lock &c->lock irq_context: 0 sk_lock-AF_BLUETOOTH-BTPROTO_RFCOMM pool_lock#2 irq_context: 0 sk_lock-AF_INET clock-AF_INET irq_context: 0 sk_lock-AF_INET &sctp_port_hashtable[i].lock irq_context: 0 sk_lock-AF_INET &sctp_port_hashtable[i].lock pool_lock#2 irq_context: 0 sk_lock-AF_INET rcu_read_lock rcu_read_lock rcu_read_lock rcu_node_0 irq_context: 0 sk_lock-AF_INET rcu_read_lock rcu_read_lock rcu_read_lock &rq->__lock irq_context: 0 sk_lock-AF_INET rcu_read_lock rcu_read_lock rcu_read_lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 sk_lock-AF_INET rcu_read_lock rhashtable_bucket irq_context: 0 sk_lock-AF_INET lock irq_context: 0 sk_lock-AF_INET lock sctp_assocs_id_lock irq_context: 0 sk_lock-AF_INET lock sctp_assocs_id_lock pool_lock#2 irq_context: 0 sk_lock-AF_INET &asoc->wait irq_context: 0 bt_proto_lock fill_pool_map-wait-type-override pool_lock#2 irq_context: 0 bt_proto_lock fill_pool_map-wait-type-override &c->lock irq_context: 0 bt_proto_lock fill_pool_map-wait-type-override &n->list_lock irq_context: 0 bt_proto_lock fill_pool_map-wait-type-override &n->list_lock &c->lock irq_context: 0 bt_proto_lock fill_pool_map-wait-type-override pool_lock irq_context: 0 tomoyo_ss remove_cache_srcu rcu_read_lock &rq->__lock &cfs_rq->removed.lock irq_context: 0 sk_lock-AF_INET6 &asoc->wait &p->pi_lock irq_context: 0 sk_lock-AF_INET6 &asoc->wait &p->pi_lock &rq->__lock irq_context: 0 sk_lock-AF_INET6 &asoc->wait &p->pi_lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)gid-cache-wq (work_completion)(&ndev_work->work) devices_rwsem &table->lock#4 &rq->__lock irq_context: 0 (wq_completion)gid-cache-wq (work_completion)(&ndev_work->work) devices_rwsem &table->lock#4 &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 &sb->s_type->i_mutex_key#10 sk_lock-AF_INET/1 &obj_hash[i].lock irq_context: 0 &sb->s_type->i_mutex_key#10 sk_lock-AF_INET/1 &base->lock irq_context: 0 &sb->s_type->i_mutex_key#10 sk_lock-AF_INET/1 &base->lock &obj_hash[i].lock irq_context: 0 &sb->s_type->i_mutex_key#10 sk_lock-AF_INET/1 rcu_read_lock pool_lock#2 irq_context: 0 &sb->s_type->i_mutex_key#10 sk_lock-AF_INET/1 pool_lock#2 irq_context: 0 &sb->s_type->i_mutex_key#10 sk_lock-AF_INET/1 krc.lock irq_context: 0 &sb->s_type->i_mutex_key#10 sk_lock-AF_INET/1 &obj_hash[i].lock pool_lock irq_context: 0 &sb->s_type->i_mutex_key#10 sk_lock-AF_INET/1 rcu_read_lock rhashtable_bucket irq_context: 0 &sb->s_type->i_mutex_key#10 sk_lock-AF_INET/1 sctp_assocs_id_lock irq_context: 0 &sb->s_type->i_mutex_key#10 sk_lock-AF_INET/1 sctp_assocs_id_lock &obj_hash[i].lock irq_context: 0 &sb->s_type->i_mutex_key#10 sk_lock-AF_INET/1 sctp_assocs_id_lock pool_lock#2 irq_context: 0 &sb->s_type->i_mutex_key#10 &net->sctp.addr_wq_lock slock-AF_INET/1 krc.lock irq_context: 0 &sb->s_type->i_mutex_key#10 &net->sctp.addr_wq_lock slock-AF_INET/1 &sctp_port_hashtable[i].lock irq_context: 0 &sb->s_type->i_mutex_key#10 &net->sctp.addr_wq_lock slock-AF_INET/1 &sctp_port_hashtable[i].lock &obj_hash[i].lock irq_context: 0 &sb->s_type->i_mutex_key#10 &net->sctp.addr_wq_lock slock-AF_INET/1 &sctp_port_hashtable[i].lock pool_lock#2 irq_context: 0 pernet_ops_rwsem &root->kernfs_rwsem &base->lock irq_context: 0 pernet_ops_rwsem &root->kernfs_rwsem &base->lock &obj_hash[i].lock irq_context: 0 (wq_completion)writeback (work_completion)(&(&wb->dwork)->work) &rq->__lock irq_context: 0 (wq_completion)bond4 (work_completion)(&(&bond->alb_work)->work) &rq->__lock irq_context: 0 (wq_completion)bond4 (work_completion)(&(&bond->alb_work)->work) &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)wg-kex-wg2#11 (work_completion)(&peer->transmit_handshake_work) &____s->seqcount#2 irq_context: 0 (wq_completion)wg-kex-wg2#11 (work_completion)(&peer->transmit_handshake_work) &____s->seqcount irq_context: 0 sb_writers &sb->s_type->i_mutex_key#4 &sb->s_type->i_lock_key#5 &xa->xa_lock#9 &obj_hash[i].lock pool_lock irq_context: 0 &xt[i].mutex free_vmap_area_lock &meta->lock irq_context: 0 &xt[i].mutex free_vmap_area_lock kfence_freelist_lock irq_context: 0 sk_lock-AF_INET rcu_read_lock &rcu_state.expedited_wq &p->pi_lock &rq->__lock irq_context: 0 sk_lock-AF_INET rcu_read_lock &rcu_state.expedited_wq &p->pi_lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 sk_lock-AF_INET lock sctp_assocs_id_lock &c->lock irq_context: 0 sk_lock-AF_INET lock sctp_assocs_id_lock &n->list_lock irq_context: 0 sk_lock-AF_INET lock sctp_assocs_id_lock &n->list_lock &c->lock irq_context: 0 sb_writers &sb->s_type->i_mutex_key#4 rcu_read_lock &cfs_rq->removed.lock irq_context: 0 sb_writers &sb->s_type->i_mutex_key#4 rcu_read_lock &obj_hash[i].lock irq_context: 0 sb_writers &sb->s_type->i_mutex_key#4 &xa->xa_lock#9 batched_entropy_u8.lock irq_context: 0 sb_writers &sb->s_type->i_mutex_key#4 &xa->xa_lock#9 kfence_freelist_lock irq_context: 0 rtnl_mutex subsys mutex#17 &rq->__lock irq_context: 0 rtnl_mutex subsys mutex#17 &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 sk_lock-AF_XDP irq_context: 0 sk_lock-AF_XDP slock-AF_XDP irq_context: 0 slock-AF_XDP irq_context: 0 rtnl_mutex rcu_state.exp_mutex fill_pool_map-wait-type-override &rq->__lock irq_context: 0 rtnl_mutex rcu_state.exp_mutex fill_pool_map-wait-type-override &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 sb_writers#4 &sb->s_type->i_mutex_key#8 jbd2_handle &ei->i_data_sem &ei->i_es_lock &n->list_lock irq_context: 0 sb_writers#4 &sb->s_type->i_mutex_key#8 jbd2_handle &ei->i_data_sem &ei->i_es_lock &n->list_lock &c->lock irq_context: 0 &vma->vm_lock->lock remove_cache_srcu &rq->__lock irq_context: 0 &vma->vm_lock->lock remove_cache_srcu &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)mld (work_completion)(&(&idev->mc_ifc_work)->work) &idev->mc_lock rcu_read_lock rcu_read_lock rcu_read_lock_bh rcu_read_lock &n->list_lock irq_context: 0 (wq_completion)mld (work_completion)(&(&idev->mc_ifc_work)->work) &idev->mc_lock rcu_read_lock rcu_read_lock rcu_read_lock_bh rcu_read_lock &n->list_lock &c->lock irq_context: 0 &sighand->siglock stock_lock rcu_read_lock per_cpu_ptr(&cgroup_rstat_cpu_lock, cpu) irq_context: 0 &f->f_pos_lock sb_writers#4 &sb->s_type->i_mutex_key#8 &xa->xa_lock#9 rcu_read_lock per_cpu_ptr(&cgroup_rstat_cpu_lock, cpu) irq_context: 0 (wq_completion)wg-crypt-wg0#3 (work_completion)(&peer->transmit_packet_work) &peer->endpoint_lock rcu_read_lock_bh rcu_read_lock rcu_read_lock rcu_read_lock &dir->lock#2 irq_context: 0 (wq_completion)wg-crypt-wg0#3 (work_completion)(&peer->transmit_packet_work) &peer->endpoint_lock rcu_read_lock_bh rcu_read_lock rcu_read_lock &____s->seqcount#7 irq_context: 0 (wq_completion)wg-crypt-wg0#3 (work_completion)(&peer->transmit_packet_work) &peer->endpoint_lock rcu_read_lock_bh rcu_read_lock &nf_nat_locks[i] irq_context: 0 (wq_completion)wg-crypt-wg0#3 (work_completion)(&peer->transmit_packet_work) &peer->endpoint_lock rcu_read_lock_bh rcu_read_lock &nf_conntrack_locks[i] irq_context: 0 (wq_completion)wg-crypt-wg0#3 (work_completion)(&peer->transmit_packet_work) &peer->endpoint_lock rcu_read_lock_bh rcu_read_lock &nf_conntrack_locks[i] &nf_conntrack_locks[i]/1 irq_context: 0 (wq_completion)wg-crypt-wg0#3 (work_completion)(&peer->transmit_packet_work) &peer->endpoint_lock rcu_read_lock_bh rcu_read_lock &nf_conntrack_locks[i] &nf_conntrack_locks[i]/1 batched_entropy_u8.lock irq_context: 0 (wq_completion)wg-crypt-wg0#3 (work_completion)(&peer->transmit_packet_work) &peer->endpoint_lock rcu_read_lock_bh &c->lock irq_context: 0 (wq_completion)wg-crypt-wg0#3 (work_completion)(&peer->transmit_packet_work) &peer->endpoint_lock rcu_read_lock_bh &obj_hash[i].lock irq_context: 0 (wq_completion)wg-crypt-wg0#3 (work_completion)(&peer->transmit_packet_work) &peer->endpoint_lock rcu_read_lock_bh rcu_read_lock &c->lock irq_context: 0 (wq_completion)wg-crypt-wg0#3 (work_completion)(&peer->transmit_packet_work) &peer->endpoint_lock rcu_read_lock_bh rcu_read_lock rcu_read_lock_bh &obj_hash[i].lock irq_context: 0 (wq_completion)wg-crypt-wg0#5 (work_completion)(&peer->transmit_packet_work) &peer->endpoint_lock rcu_read_lock_bh rcu_read_lock rcu_read_lock_bh rcu_read_lock pool_lock#2 irq_context: 0 rtnl_mutex &dev->tx_global_lock _xmit_TUNNEL#2 irq_context: 0 rtnl_mutex rcu_read_lock &tb->tb6_lock rcu_read_lock &n->list_lock irq_context: 0 rtnl_mutex rcu_read_lock &tb->tb6_lock rcu_read_lock &n->list_lock &c->lock irq_context: 0 sk_lock-AF_BLUETOOTH-BTPROTO_HCI mgmt_chan_list_lock irq_context: 0 sk_lock-AF_BLUETOOTH-BTPROTO_HCI &rq->__lock irq_context: 0 sk_lock-AF_BLUETOOTH-BTPROTO_HCI &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 &sb->s_type->i_mutex_key#10 sk_lock-AF_BLUETOOTH-BTPROTO_HCI &c->lock irq_context: 0 (wq_completion)ext4-rsv-conversion (work_completion)(&ei->i_rsv_conversion_work) jbd2_handle &ei->i_data_sem &journal->j_state_lock irq_context: 0 (wq_completion)ext4-rsv-conversion (work_completion)(&ei->i_rsv_conversion_work) jbd2_handle &ei->i_data_sem &journal->j_revoke_lock irq_context: 0 (wq_completion)ext4-rsv-conversion (work_completion)(&ei->i_rsv_conversion_work) jbd2_handle &ei->i_data_sem &bgl->locks[i].lock &sbi->s_md_lock irq_context: 0 (wq_completion)ext4-rsv-conversion (work_completion)(&ei->i_rsv_conversion_work) jbd2_handle &ei->i_data_sem &bgl->locks[i].lock &obj_hash[i].lock irq_context: 0 (wq_completion)ext4-rsv-conversion (work_completion)(&ei->i_rsv_conversion_work) jbd2_handle &ei->i_data_sem &bgl->locks[i].lock pool_lock#2 irq_context: 0 (wq_completion)ext4-rsv-conversion (work_completion)(&ei->i_rsv_conversion_work) jbd2_handle &ei->i_data_sem &sb->s_type->i_lock_key#22 irq_context: 0 &sbi->s_writepages_rwsem jbd2_handle &ei->i_data_sem &____s->seqcount#2 irq_context: 0 (wq_completion)wg-crypt-wg1#6 (work_completion)(&peer->transmit_packet_work) rcu_read_lock_bh rcu_read_lock &pool->lock/1 irq_context: 0 (wq_completion)wg-crypt-wg1#6 (work_completion)(&peer->transmit_packet_work) rcu_read_lock_bh rcu_read_lock &pool->lock/1 &obj_hash[i].lock irq_context: 0 (wq_completion)wg-crypt-wg1#6 (work_completion)(&peer->transmit_packet_work) rcu_read_lock_bh rcu_read_lock &pool->lock/1 &p->pi_lock irq_context: 0 (wq_completion)wg-crypt-wg1#6 (work_completion)(&peer->transmit_packet_work) rcu_read_lock_bh rcu_read_lock &pool->lock/1 &p->pi_lock &rq->__lock irq_context: 0 (wq_completion)wg-crypt-wg1#6 (work_completion)(&peer->transmit_packet_work) rcu_read_lock_bh rcu_read_lock &pool->lock/1 &p->pi_lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)wg-kex-wg1#9 (work_completion)(&peer->transmit_handshake_work) &wg->static_identity.lock &handshake->lock &rq->__lock irq_context: 0 (wq_completion)wg-kex-wg1#9 (work_completion)(&peer->transmit_handshake_work) &wg->static_identity.lock &handshake->lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)wg-kex-wg1#9 (work_completion)(&peer->transmit_handshake_work) &peer->endpoint_lock rcu_read_lock_bh rcu_read_lock &____s->seqcount#9 irq_context: 0 (wq_completion)wg-kex-wg1#9 (work_completion)(&peer->transmit_handshake_work) &peer->endpoint_lock rcu_read_lock_bh rcu_read_lock rcu_read_lock_bh &obj_hash[i].lock irq_context: 0 sb_writers#4 &sb->s_type->i_mutex_key#8 jbd2_handle &ei->i_data_sem &bgl->locks[i].lock irq_context: 0 sb_writers#4 &sb->s_type->i_mutex_key#8 jbd2_handle &ei->i_data_sem &ei->i_es_lock &____s->seqcount#2 irq_context: 0 sb_writers#4 &sb->s_type->i_mutex_key#8 jbd2_handle &ei->i_data_sem &ei->i_es_lock &____s->seqcount irq_context: 0 rtnl_mutex &dev->tx_global_lock _xmit_TUNNEL6#2 irq_context: 0 rtnl_mutex &dev->tx_global_lock _xmit_SIT#2 irq_context: 0 (wq_completion)wg-crypt-wg0#3 (work_completion)(&peer->transmit_packet_work) &peer->endpoint_lock rcu_read_lock_bh rcu_read_lock rcu_read_lock rcu_read_lock &c->lock irq_context: softirq rcu_callback batched_entropy_u8.lock irq_context: 0 (wq_completion)events (work_completion)(&w->work)#2 nf_conntrack_mutex nf_conntrack_mutex.wait_lock irq_context: 0 (wq_completion)events (work_completion)(&w->work)#2 nf_conntrack_mutex &pool->lock irq_context: 0 (wq_completion)events (work_completion)(&w->work)#2 nf_conntrack_mutex &pool->lock &p->pi_lock irq_context: 0 (wq_completion)events (work_completion)(&w->work)#2 nf_conntrack_mutex &pool->lock &p->pi_lock &rq->__lock irq_context: 0 (wq_completion)events (work_completion)(&w->work)#2 nf_conntrack_mutex &pool->lock &p->pi_lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)events (work_completion)(&w->work)#2 nf_conntrack_mutex.wait_lock irq_context: 0 (wq_completion)events (work_completion)(&w->work)#2 &p->pi_lock irq_context: 0 (wq_completion)events (work_completion)(&w->work)#2 &p->pi_lock &rq->__lock irq_context: 0 (wq_completion)events (work_completion)(&w->work)#2 &p->pi_lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)wg-crypt-wg0#3 (work_completion)(&peer->transmit_packet_work) &peer->endpoint_lock rcu_read_lock_bh rcu_read_lock rcu_read_lock rcu_read_lock pool_lock#2 irq_context: 0 rtnl_mutex rcu_read_lock fill_pool_map-wait-type-override rcu_node_0 irq_context: 0 rtnl_mutex rcu_read_lock fill_pool_map-wait-type-override &rq->__lock irq_context: 0 rtnl_mutex rcu_read_lock fill_pool_map-wait-type-override &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)events (work_completion)(&data->fib_event_work) &data->fib_lock &meta->lock irq_context: 0 rtnl_mutex &dev_addr_list_lock_key/1 _xmit_ETHER irq_context: 0 rtnl_mutex &dev_addr_list_lock_key/1 _xmit_ETHER &obj_hash[i].lock irq_context: 0 rtnl_mutex &dev_addr_list_lock_key/1 _xmit_ETHER pool_lock#2 irq_context: 0 rtnl_mutex &dev_addr_list_lock_key/1 _xmit_ETHER krc.lock irq_context: 0 rtnl_mutex &dev_addr_list_lock_key/1 &obj_hash[i].lock irq_context: 0 rtnl_mutex &dev_addr_list_lock_key/1 krc.lock irq_context: 0 rtnl_mutex &dev->tx_global_lock &qdisc_xmit_lock_key#4 irq_context: 0 rtnl_mutex &dev_addr_list_lock_key#2/1 _xmit_ETHER irq_context: 0 rtnl_mutex &dev_addr_list_lock_key#2/1 _xmit_ETHER &obj_hash[i].lock irq_context: 0 rtnl_mutex &dev_addr_list_lock_key#2/1 _xmit_ETHER pool_lock#2 irq_context: 0 rtnl_mutex &dev_addr_list_lock_key#2/1 _xmit_ETHER krc.lock irq_context: 0 rtnl_mutex &dev_addr_list_lock_key#2/1 &obj_hash[i].lock irq_context: 0 rtnl_mutex &dev_addr_list_lock_key#2/1 krc.lock irq_context: 0 rtnl_mutex &idev->mc_lock &dev_addr_list_lock_key#2/1 irq_context: 0 rtnl_mutex &idev->mc_lock &dev_addr_list_lock_key#2/1 &obj_hash[i].lock irq_context: 0 rtnl_mutex &idev->mc_lock &dev_addr_list_lock_key#2/1 krc.lock irq_context: 0 rtnl_mutex _xmit_NETROM#2 irq_context: 0 rtnl_mutex &idev->mc_lock _xmit_ETHER &obj_hash[i].lock pool_lock irq_context: 0 rtnl_mutex team->team_lock_key#2 lweventlist_lock &c->lock irq_context: 0 rtnl_mutex team->team_lock_key#2 lweventlist_lock pool_lock#2 irq_context: 0 rtnl_mutex &idev->mc_lock remove_cache_srcu rcu_read_lock rcu_node_0 irq_context: 0 rtnl_mutex &idev->mc_lock remove_cache_srcu rcu_read_lock &rq->__lock irq_context: 0 rtnl_mutex &idev->mc_lock remove_cache_srcu rcu_read_lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 rtnl_mutex _xmit_ETHER/1 &obj_hash[i].lock irq_context: 0 rtnl_mutex _xmit_ETHER/1 krc.lock irq_context: 0 rtnl_mutex &idev->mc_lock _xmit_ETHER/1 irq_context: 0 rtnl_mutex &idev->mc_lock _xmit_ETHER/1 &obj_hash[i].lock irq_context: 0 rtnl_mutex &idev->mc_lock _xmit_ETHER/1 krc.lock irq_context: 0 rtnl_mutex &idev->mc_lock remove_cache_srcu &rq->__lock irq_context: 0 rtnl_mutex &idev->mc_lock remove_cache_srcu &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 rtnl_mutex &macvlan_netdev_addr_lock_key/1 _xmit_ETHER &obj_hash[i].lock irq_context: 0 rtnl_mutex &macvlan_netdev_addr_lock_key/1 _xmit_ETHER pool_lock#2 irq_context: 0 rtnl_mutex &macvlan_netdev_addr_lock_key/1 _xmit_ETHER krc.lock irq_context: 0 rtnl_mutex &macvlan_netdev_addr_lock_key/1 &obj_hash[i].lock irq_context: 0 rtnl_mutex &macvlan_netdev_addr_lock_key/1 krc.lock irq_context: 0 rtnl_mutex rcu_read_lock &tb->tb6_lock quarantine_lock irq_context: 0 rtnl_mutex dev->qdisc_tx_busylock ?: &qdisc_tx_busylock#2 irq_context: 0 rtnl_mutex dev->qdisc_tx_busylock ?: &qdisc_tx_busylock#2 &sch->q.lock irq_context: 0 (wq_completion)events (work_completion)(&(&nsim_dev->trap_data->trap_report_dw)->work) &devlink->lock_key &nsim_trap_data->trap_lock crngs.lock base_crng.lock irq_context: 0 (wq_completion)bat_events (work_completion)(&(&forw_packet_aggr->delayed_work)->work) batched_entropy_u8.lock crngs.lock base_crng.lock irq_context: 0 (wq_completion)wg-kex-wg2#7 (work_completion)(&peer->transmit_handshake_work) &wg->static_identity.lock &handshake->lock rcu_read_lock_bh batched_entropy_u32.lock crngs.lock irq_context: 0 (wq_completion)wg-kex-wg2#7 (work_completion)(&peer->transmit_handshake_work) &wg->static_identity.lock &handshake->lock rcu_node_0 irq_context: 0 (wq_completion)wg-kex-wg2#7 (work_completion)(&peer->transmit_handshake_work) &wg->static_identity.lock &handshake->lock &rcu_state.expedited_wq irq_context: 0 (wq_completion)wg-kex-wg2#7 (work_completion)(&peer->transmit_handshake_work) &wg->static_identity.lock &handshake->lock &rcu_state.expedited_wq &p->pi_lock irq_context: 0 rtnl_mutex &dev->tx_global_lock &qdisc_xmit_lock_key#5 irq_context: 0 (wq_completion)wg-kex-wg1#12 (work_completion)(&({ do { const void *__vpp_verify = (typeof((worker) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __asm__ ("" : "=r"(__ptr) : "0"((typeof(*((worker))) *)((worker)))); (typeof((typeof(*((worker))) *)((worker)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); })->work) &wg->static_identity.lock &handshake->lock &rq->__lock irq_context: 0 (wq_completion)wg-kex-wg1#12 (work_completion)(&({ do { const void *__vpp_verify = (typeof((worker) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __asm__ ("" : "=r"(__ptr) : "0"((typeof(*((worker))) *)((worker)))); (typeof((typeof(*((worker))) *)((worker)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); })->work) &____s->seqcount#2 irq_context: 0 (wq_completion)wg-kex-wg1#12 (work_completion)(&({ do { const void *__vpp_verify = (typeof((worker) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __asm__ ("" : "=r"(__ptr) : "0"((typeof(*((worker))) *)((worker)))); (typeof((typeof(*((worker))) *)((worker)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); })->work) &____s->seqcount irq_context: 0 (wq_completion)wg-kex-wg1#12 (work_completion)(&({ do { const void *__vpp_verify = (typeof((worker) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __asm__ ("" : "=r"(__ptr) : "0"((typeof(*((worker))) *)((worker)))); (typeof((typeof(*((worker))) *)((worker)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); })->work) pool_lock#2 irq_context: 0 (wq_completion)wg-kex-wg2#8 (work_completion)(&({ do { const void *__vpp_verify = (typeof((worker) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __asm__ ("" : "=r"(__ptr) : "0"((typeof(*((worker))) *)((worker)))); (typeof((typeof(*((worker))) *)((worker)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); })->work) &wg->static_identity.lock &rq->__lock irq_context: 0 (wq_completion)wg-kex-wg2#8 (work_completion)(&({ do { const void *__vpp_verify = (typeof((worker) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __asm__ ("" : "=r"(__ptr) : "0"((typeof(*((worker))) *)((worker)))); (typeof((typeof(*((worker))) *)((worker)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); })->work) &wg->static_identity.lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 rtnl_mutex &macsec_netdev_addr_lock_key/1 _xmit_ETHER &obj_hash[i].lock irq_context: 0 rtnl_mutex &macsec_netdev_addr_lock_key/1 _xmit_ETHER pool_lock#2 irq_context: 0 rtnl_mutex &macsec_netdev_addr_lock_key/1 _xmit_ETHER krc.lock irq_context: 0 rtnl_mutex _xmit_ETHER krc.lock &obj_hash[i].lock irq_context: 0 rtnl_mutex _xmit_ETHER krc.lock &base->lock irq_context: 0 rtnl_mutex _xmit_ETHER krc.lock &base->lock &obj_hash[i].lock irq_context: 0 rtnl_mutex &macsec_netdev_addr_lock_key/1 &obj_hash[i].lock irq_context: 0 rtnl_mutex &macsec_netdev_addr_lock_key/1 krc.lock irq_context: 0 rtnl_mutex &tbl->lock &obj_hash[i].lock pool_lock irq_context: 0 (wq_completion)bat_events (work_completion)(&(&forw_packet_aggr->delayed_work)->work) &hard_iface->bat_iv.ogm_buff_mutex &bat_priv->tt.commit_lock &bat_priv->tt.last_changeset_lock &c->lock irq_context: 0 rtnl_mutex &net->xfrm.xfrm_policy_lock irq_context: 0 rtnl_mutex &rdev->wiphy.mtx &wdev->mtx &local->sta_mtx remove_cache_srcu irq_context: 0 rtnl_mutex &rdev->wiphy.mtx &wdev->mtx &local->sta_mtx remove_cache_srcu quarantine_lock irq_context: 0 rtnl_mutex &rdev->wiphy.mtx &wdev->mtx &local->sta_mtx remove_cache_srcu &c->lock irq_context: 0 rtnl_mutex &rdev->wiphy.mtx &wdev->mtx &local->sta_mtx remove_cache_srcu &rq->__lock irq_context: 0 rtnl_mutex &rdev->wiphy.mtx &wdev->mtx &local->sta_mtx remove_cache_srcu &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 rtnl_mutex &rdev->wiphy.mtx &wdev->mtx &local->sta_mtx remove_cache_srcu &n->list_lock irq_context: 0 rtnl_mutex &rdev->wiphy.mtx &wdev->mtx &local->sta_mtx remove_cache_srcu rcu_read_lock pool_lock#2 irq_context: 0 rtnl_mutex &rdev->wiphy.mtx &wdev->mtx &local->sta_mtx remove_cache_srcu &obj_hash[i].lock irq_context: 0 rtnl_mutex &rdev->wiphy.mtx &wdev->mtx &local->sta_mtx &sb->s_type->i_mutex_key#3 rcu_read_lock rcu_node_0 irq_context: 0 rtnl_mutex &rdev->wiphy.mtx &wdev->mtx &local->sta_mtx &sb->s_type->i_mutex_key#3 rcu_read_lock &rq->__lock irq_context: 0 rtnl_mutex &rdev->wiphy.mtx &wdev->mtx &local->sta_mtx &sb->s_type->i_mutex_key#3 rcu_read_lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 rtnl_mutex &rdev->wiphy.mtx (wq_completion)phy7 irq_context: 0 rtnl_mutex &rdev->wiphy.mtx (wq_completion)phy7 &rq->__lock irq_context: 0 rtnl_mutex &rdev->wiphy.mtx (wq_completion)phy7 &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 rtnl_mutex &vlan_netdev_addr_lock_key/2 irq_context: 0 rtnl_mutex &tbl->lock &n->list_lock irq_context: 0 rtnl_mutex &tbl->lock &n->list_lock &c->lock irq_context: 0 (wq_completion)wg-crypt-wg0 (work_completion)(&peer->transmit_packet_work) &peer->endpoint_lock rcu_read_lock_bh batched_entropy_u8.lock irq_context: 0 (wq_completion)wg-crypt-wg0 (work_completion)(&peer->transmit_packet_work) &peer->endpoint_lock rcu_read_lock_bh rcu_read_lock rcu_read_lock_bh &meta->lock irq_context: 0 (wq_completion)wg-crypt-wg0 (work_completion)(&peer->transmit_packet_work) &peer->endpoint_lock rcu_read_lock_bh rcu_read_lock rcu_read_lock_bh kfence_freelist_lock irq_context: 0 (wq_completion)wg-crypt-wg0 (work_completion)(&peer->transmit_packet_work) &peer->endpoint_lock rcu_read_lock_bh &base->lock irq_context: 0 (wq_completion)wg-crypt-wg0 (work_completion)(&peer->transmit_packet_work) &peer->endpoint_lock rcu_read_lock_bh &base->lock &obj_hash[i].lock irq_context: 0 (wq_completion)ipv6_addrconf (work_completion)(&(&ifa->dad_work)->work) rtnl_mutex &idev->mc_lock rcu_read_lock rcu_read_lock rcu_read_lock_bh rcu_read_lock quarantine_lock irq_context: 0 cb_lock genl_mutex rtnl_mutex &sdata->sec_mtx irq_context: 0 sb_writers#4 &type->i_mutex_dir_key#3/1 jbd2_handle &ei->i_data_sem &pcp->lock &zone->lock irq_context: 0 sb_writers#4 sb_internal jbd2_handle &meta->lock irq_context: 0 sb_writers#4 sb_internal jbd2_handle kfence_freelist_lock irq_context: 0 (wq_completion)wg-kex-wg2 (work_completion)(&peer->transmit_handshake_work) &peer->endpoint_lock rcu_read_lock_bh rcu_read_lock rcu_read_lock_bh rcu_read_lock &c->lock irq_context: 0 (wq_completion)wg-kex-wg2 (work_completion)(&peer->transmit_handshake_work) &peer->endpoint_lock rcu_read_lock_bh rcu_read_lock rcu_read_lock_bh rcu_read_lock tk_core.seq.seqcount irq_context: 0 (wq_completion)wg-kex-wg2 (work_completion)(&peer->transmit_handshake_work) &peer->endpoint_lock rcu_read_lock_bh rcu_read_lock rcu_read_lock_bh rcu_read_lock &obj_hash[i].lock irq_context: 0 (wq_completion)wg-kex-wg2 (work_completion)(&peer->transmit_handshake_work) &peer->endpoint_lock rcu_read_lock_bh rcu_read_lock rcu_read_lock_bh rcu_read_lock pool_lock#2 irq_context: 0 sk_lock-AF_QIPCRTR irq_context: 0 sk_lock-AF_QIPCRTR slock-AF_QIPCRTR irq_context: 0 slock-AF_QIPCRTR irq_context: 0 &mm->mmap_lock sb_pagefaults mapping.invalidate_lock &obj_hash[i].lock irq_context: 0 remove_cache_srcu rcu_read_lock rcu_read_lock &cfs_rq->removed.lock irq_context: 0 remove_cache_srcu rcu_read_lock rcu_read_lock &obj_hash[i].lock irq_context: 0 rtnl_mutex rcu_read_lock &macvlan_netdev_addr_lock_key/1 irq_context: 0 rtnl_mutex &idev->mc_lock &macvlan_netdev_addr_lock_key/1 pool_lock#2 irq_context: 0 callchain_mutex irq_context: 0 callchain_mutex fs_reclaim irq_context: 0 callchain_mutex fs_reclaim mmu_notifier_invalidate_range_start irq_context: 0 callchain_mutex pool_lock#2 irq_context: 0 callchain_mutex &c->lock irq_context: 0 sb_writers#4 &sb->s_type->i_mutex_key#8 jbd2_handle &ei->i_data_sem &rq->__lock &cfs_rq->removed.lock irq_context: 0 sb_writers#4 &sb->s_type->i_mutex_key#8 mmu_notifier_invalidate_range_start &cfs_rq->removed.lock irq_context: 0 sb_writers#4 &sb->s_type->i_mutex_key#8 mmu_notifier_invalidate_range_start &obj_hash[i].lock irq_context: 0 sb_writers#4 &sb->s_type->i_mutex_key#8 mmu_notifier_invalidate_range_start pool_lock#2 irq_context: 0 &mm->mmap_lock rcu_read_lock rcu_read_lock ptlock_ptr(page)#2 &folio_wait_table[i] irq_context: 0 &mm->mmap_lock rcu_read_lock rcu_read_lock ptlock_ptr(page)#2 &folio_wait_table[i] &p->pi_lock irq_context: 0 &mm->mmap_lock rcu_read_lock rcu_read_lock ptlock_ptr(page)#2 &folio_wait_table[i] &p->pi_lock &rq->__lock irq_context: 0 &mm->mmap_lock rcu_read_lock rcu_read_lock ptlock_ptr(page)#2 &folio_wait_table[i] &p->pi_lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 sb_writers#4 &sb->s_type->i_mutex_key#8 jbd2_handle &ei->i_data_sem rcu_read_lock &rcu_state.expedited_wq irq_context: 0 sb_writers#4 &sb->s_type->i_mutex_key#8 jbd2_handle &ei->i_data_sem rcu_read_lock &rcu_state.expedited_wq &p->pi_lock irq_context: 0 sb_writers#4 &sb->s_type->i_mutex_key#8 jbd2_handle &ei->i_data_sem rcu_read_lock &rcu_state.expedited_wq &p->pi_lock &rq->__lock irq_context: 0 sb_writers#4 &sb->s_type->i_mutex_key#8 jbd2_handle &ei->i_data_sem rcu_read_lock &rcu_state.expedited_wq &p->pi_lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 &sb->s_type->i_mutex_key#10 &u->peer_wait &p->pi_lock irq_context: 0 &sb->s_type->i_mutex_key#10 &u->peer_wait &p->pi_lock &rq->__lock irq_context: 0 &sb->s_type->i_mutex_key#10 &u->peer_wait &p->pi_lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)events_unbound (work_completion)(&map->work) callchain_mutex irq_context: 0 (wq_completion)events_unbound (work_completion)(&map->work) callchain_mutex &obj_hash[i].lock irq_context: 0 (wq_completion)events_unbound (work_completion)(&map->work) callchain_mutex pool_lock#2 irq_context: 0 rtnl_mutex dev_hotplug_mutex &rq->__lock irq_context: 0 rtnl_mutex dev_hotplug_mutex &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)mld (work_completion)(&(&idev->mc_dad_work)->work) &idev->mc_lock rcu_read_lock rcu_read_lock rcu_read_lock_bh rcu_read_lock batched_entropy_u8.lock irq_context: 0 (wq_completion)mld (work_completion)(&(&idev->mc_dad_work)->work) &idev->mc_lock rcu_read_lock rcu_read_lock rcu_read_lock_bh rcu_read_lock rlock-AF_PACKET irq_context: 0 (wq_completion)wg-kex-wg1#9 (work_completion)(&peer->transmit_handshake_work) &peer->endpoint_lock rcu_read_lock_bh rcu_read_lock rcu_read_lock_bh rcu_read_lock tk_core.seq.seqcount irq_context: 0 (wq_completion)wg-kex-wg1#9 (work_completion)(&peer->transmit_handshake_work) &peer->endpoint_lock rcu_read_lock_bh rcu_read_lock rcu_read_lock_bh rcu_read_lock &obj_hash[i].lock irq_context: 0 rtnl_mutex device_links_lock &rq->__lock irq_context: 0 rtnl_mutex device_links_lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)events (work_completion)(&(&krcp->monitor_work)->work) rcu_callback rcu_read_lock pool_lock#2 irq_context: 0 rtnl_mutex rcu_read_lock rcu_read_lock_bh rcu_read_lock &____s->seqcount irq_context: 0 rtnl_mutex rcu_read_lock rcu_read_lock_bh rcu_read_lock &c->lock irq_context: 0 rtnl_mutex &root->kernfs_rwsem rcu_read_lock &rcu_state.expedited_wq irq_context: 0 rtnl_mutex &root->kernfs_rwsem rcu_read_lock &rcu_state.expedited_wq &p->pi_lock irq_context: 0 rtnl_mutex &root->kernfs_rwsem rcu_read_lock &rcu_state.expedited_wq &p->pi_lock &rq->__lock irq_context: 0 rtnl_mutex &root->kernfs_rwsem rcu_read_lock &rcu_state.expedited_wq &p->pi_lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)wg-kex-wg0#9 (work_completion)(&peer->transmit_handshake_work) &peer->endpoint_lock rcu_read_lock_bh &obj_hash[i].lock irq_context: 0 (wq_completion)wg-kex-wg0#9 (work_completion)(&peer->transmit_handshake_work) &peer->endpoint_lock rcu_read_lock_bh &c->lock irq_context: 0 (wq_completion)wg-kex-wg0#9 (work_completion)(&peer->transmit_handshake_work) &peer->endpoint_lock rcu_read_lock_bh rcu_read_lock rcu_read_lock_bh &obj_hash[i].lock irq_context: 0 (wq_completion)wg-kex-wg0 (work_completion)(&peer->transmit_handshake_work) &peer->endpoint_lock rcu_read_lock_bh &c->lock irq_context: 0 (wq_completion)wg-kex-wg0 (work_completion)(&peer->transmit_handshake_work) &peer->endpoint_lock rcu_read_lock_bh &n->list_lock irq_context: 0 (wq_completion)wg-kex-wg0 (work_completion)(&peer->transmit_handshake_work) &peer->endpoint_lock rcu_read_lock_bh &n->list_lock &c->lock irq_context: 0 (wq_completion)wg-kex-wg0 (work_completion)(&peer->transmit_handshake_work) &peer->endpoint_lock rcu_read_lock_bh &obj_hash[i].lock irq_context: 0 (wq_completion)wg-kex-wg0 (work_completion)(&peer->transmit_handshake_work) &peer->endpoint_lock rcu_read_lock_bh rcu_read_lock rcu_read_lock_bh &obj_hash[i].lock irq_context: 0 pernet_ops_rwsem dev_pm_qos_sysfs_mtx &root->kernfs_rwsem &rq->__lock irq_context: 0 pernet_ops_rwsem dev_pm_qos_sysfs_mtx &root->kernfs_rwsem &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 callchain_mutex &n->list_lock irq_context: 0 callchain_mutex &n->list_lock &c->lock irq_context: 0 callchain_mutex &rq->__lock irq_context: 0 sock_diag_mutex rcu_read_lock &c->lock irq_context: 0 sock_diag_mutex rcu_read_lock rcu_read_lock_bh rcu_read_lock &c->lock irq_context: 0 cb_lock nlk_cb_mutex-GENERIC rcu_read_lock rcu_read_lock_bh rcu_read_lock &c->lock irq_context: 0 &sb->s_type->i_mutex_key#10 purge_vmap_area_lock &meta->lock irq_context: 0 &sb->s_type->i_mutex_key#10 purge_vmap_area_lock kfence_freelist_lock irq_context: 0 sb_writers#10 &type->i_mutex_dir_key#6 stock_lock irq_context: 0 sb_writers#10 &type->i_mutex_dir_key#6 &dentry->d_lock &wq#2 irq_context: 0 sb_writers#10 &type->i_mutex_dir_key#6 &rq->__lock irq_context: 0 sb_writers#10 &type->i_mutex_dir_key#6 &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: softirq (&ndev->rs_timer) rcu_read_lock rcu_read_lock rcu_read_lock_bh &hsr->seqnr_lock rcu_read_lock rcu_read_lock_bh rcu_read_lock &c->lock irq_context: softirq (&ndev->rs_timer) rcu_read_lock rcu_read_lock rcu_read_lock_bh _xmit_ETHER#2 rcu_read_lock &n->list_lock irq_context: softirq (&ndev->rs_timer) rcu_read_lock rcu_read_lock rcu_read_lock_bh _xmit_ETHER#2 rcu_read_lock &n->list_lock &c->lock irq_context: 0 rtnl_mutex smc_ib_devices.mutex &rq->__lock irq_context: 0 rtnl_mutex smc_ib_devices.mutex &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 rtnl_mutex (inetaddr_chain).rwsem stock_lock irq_context: 0 &f->f_pos_lock sb_writers#11 irq_context: 0 &f->f_pos_lock sb_writers#11 &rq->__lock irq_context: 0 &f->f_pos_lock sb_writers#11 &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 &f->f_pos_lock sb_writers#11 fs_reclaim irq_context: 0 &f->f_pos_lock sb_writers#11 fs_reclaim mmu_notifier_invalidate_range_start irq_context: 0 &f->f_pos_lock sb_writers#11 &c->lock irq_context: 0 &f->f_pos_lock sb_writers#11 &mm->mmap_lock irq_context: 0 &f->f_pos_lock sb_writers#11 &mm->mmap_lock fs_reclaim irq_context: 0 &f->f_pos_lock sb_writers#11 &mm->mmap_lock fs_reclaim mmu_notifier_invalidate_range_start irq_context: 0 &f->f_pos_lock sb_writers#11 &mm->mmap_lock &____s->seqcount irq_context: 0 &f->f_pos_lock sb_writers#11 &mm->mmap_lock stock_lock irq_context: 0 &f->f_pos_lock sb_writers#11 &mm->mmap_lock &obj_hash[i].lock irq_context: 0 &f->f_pos_lock sb_writers#11 &obj_hash[i].lock irq_context: 0 pernet_ops_rwsem k-sk_lock-AF_TIPC &tn->nametbl_lock &service->lock &n->list_lock irq_context: 0 pernet_ops_rwsem k-sk_lock-AF_TIPC &tn->nametbl_lock &service->lock &n->list_lock &c->lock irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem ovs_mutex nf_ct_proto_mutex defrag6_mutex &obj_hash[i].lock pool_lock irq_context: 0 callchain_mutex &____s->seqcount irq_context: 0 &sb->s_type->i_mutex_key#10 rcu_state.exp_mutex fill_pool_map-wait-type-override &rq->__lock irq_context: 0 &sb->s_type->i_mutex_key#10 rcu_state.exp_mutex fill_pool_map-wait-type-override &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 kn->active#60 &n->list_lock irq_context: 0 kn->active#60 &n->list_lock &c->lock irq_context: 0 &u->iolock &rcu_state.expedited_wq &p->pi_lock &rq->__lock irq_context: 0 &u->iolock &rcu_state.expedited_wq &p->pi_lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 tracepoints_mutex cpu_hotplug_lock static_call_mutex text_mutex.wait_lock irq_context: 0 tracepoints_mutex cpu_hotplug_lock static_call_mutex &p->pi_lock irq_context: 0 tracepoints_mutex cpu_hotplug_lock static_call_mutex &p->pi_lock &rq->__lock irq_context: 0 tracepoints_mutex cpu_hotplug_lock static_call_mutex &p->pi_lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 tracepoints_mutex cpu_hotplug_lock static_call_mutex &rq->__lock irq_context: 0 tracepoints_mutex cpu_hotplug_lock static_call_mutex &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 rtnl_mutex uevent_sock_mutex rcu_read_lock &rcu_state.gp_wq irq_context: 0 rtnl_mutex uevent_sock_mutex rcu_read_lock &rcu_state.gp_wq &p->pi_lock irq_context: 0 rtnl_mutex uevent_sock_mutex rcu_read_lock &rcu_state.gp_wq &p->pi_lock &rq->__lock irq_context: 0 rtnl_mutex uevent_sock_mutex rcu_read_lock &rcu_state.gp_wq &p->pi_lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)bat_events (work_completion)(&(&bat_priv->dat.work)->work) &rcu_state.expedited_wq irq_context: 0 (wq_completion)bat_events (work_completion)(&(&bat_priv->dat.work)->work) &rcu_state.expedited_wq &p->pi_lock irq_context: 0 (wq_completion)bat_events (work_completion)(&(&bat_priv->dat.work)->work) &rcu_state.expedited_wq &p->pi_lock &rq->__lock irq_context: 0 (wq_completion)bat_events (work_completion)(&(&bat_priv->dat.work)->work) &rcu_state.expedited_wq &p->pi_lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 sk_lock-AF_PACKET &po->bind_lock rcu_read_lock &____s->seqcount#2 irq_context: 0 rcu_read_lock_bh rcu_read_lock rcu_read_lock pool_lock#2 irq_context: 0 &mm->mmap_lock fill_pool_map-wait-type-override &c->lock irq_context: 0 &sb->s_type->i_mutex_key#9 rcu_read_lock &p->pi_lock irq_context: 0 &sb->s_type->i_mutex_key#9 rcu_read_lock &p->pi_lock &rq->__lock irq_context: 0 &sb->s_type->i_mutex_key#9 rcu_read_lock &p->pi_lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 &sb->s_type->i_mutex_key#9 &rcu_state.expedited_wq irq_context: 0 &sb->s_type->i_mutex_key#9 &rcu_state.expedited_wq &p->pi_lock irq_context: 0 &sb->s_type->i_mutex_key#9 &rcu_state.expedited_wq &p->pi_lock &rq->__lock irq_context: 0 &sb->s_type->i_mutex_key#9 &rcu_state.expedited_wq &p->pi_lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 rtnl_mutex &lock->wait_lock irq_context: 0 sb_writers#4 &type->i_mutex_dir_key#3/1 jbd2_handle &ei->i_data_sem batched_entropy_u8.lock irq_context: 0 sb_writers#4 &type->i_mutex_dir_key#3/1 jbd2_handle &ei->i_data_sem kfence_freelist_lock irq_context: 0 sb_writers#4 &type->i_mutex_dir_key#3/1 jbd2_handle &ei->i_data_sem &meta->lock irq_context: 0 (wq_completion)bat_events (work_completion)(&(&bat_priv->bla.work)->work) rcu_read_lock rcu_read_lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 sb_writers#11 &type->i_mutex_dir_key#7 &rq->__lock irq_context: 0 sb_writers#11 &type->i_mutex_dir_key#7 &n->list_lock &c->lock irq_context: 0 sb_writers#11 &type->i_mutex_dir_key#7 &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem kernfs_idr_lock fill_pool_map-wait-type-override pool_lock#2 irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem kernfs_idr_lock fill_pool_map-wait-type-override &c->lock irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem kernfs_idr_lock fill_pool_map-wait-type-override pool_lock irq_context: 0 (wq_completion)bat_events (work_completion)(&(&forw_packet_aggr->delayed_work)->work) &hard_iface->bat_iv.ogm_buff_mutex rcu_read_lock fill_pool_map-wait-type-override &c->lock irq_context: 0 (wq_completion)bat_events (work_completion)(&(&forw_packet_aggr->delayed_work)->work) &hard_iface->bat_iv.ogm_buff_mutex rcu_read_lock fill_pool_map-wait-type-override &n->list_lock irq_context: 0 (wq_completion)bat_events (work_completion)(&(&forw_packet_aggr->delayed_work)->work) &hard_iface->bat_iv.ogm_buff_mutex rcu_read_lock fill_pool_map-wait-type-override &n->list_lock &c->lock irq_context: 0 &f->f_pos_lock sb_writers#11 &n->list_lock irq_context: 0 &f->f_pos_lock sb_writers#11 &n->list_lock &c->lock irq_context: 0 (wq_completion)wg-kex-wg0#7 (work_completion)(&peer->transmit_handshake_work) &peer->endpoint_lock rcu_read_lock_bh rcu_read_lock rcu_read_lock_bh rcu_read_lock &c->lock irq_context: 0 (wq_completion)wg-kex-wg0#7 (work_completion)(&peer->transmit_handshake_work) &peer->endpoint_lock rcu_read_lock_bh rcu_read_lock rcu_read_lock_bh rcu_read_lock rlock-AF_PACKET irq_context: softirq (&in_dev->mr_ifc_timer) rcu_read_lock &im->lock batched_entropy_u8.lock irq_context: softirq (&in_dev->mr_ifc_timer) rcu_read_lock &im->lock kfence_freelist_lock irq_context: softirq (&in_dev->mr_ifc_timer) rcu_read_lock rcu_read_lock_bh rcu_read_lock &meta->lock irq_context: softirq (&in_dev->mr_ifc_timer) rcu_read_lock rcu_read_lock_bh rcu_read_lock kfence_freelist_lock irq_context: 0 rtnl_mutex dpm_list_mtx &rq->__lock irq_context: 0 rtnl_mutex dpm_list_mtx &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 &f->f_pos_lock sb_writers#4 &sb->s_type->i_mutex_key#8 jbd2_handle &ei->i_data_sem &ei->i_es_lock &c->lock irq_context: 0 &f->f_pos_lock sb_writers#4 &sb->s_type->i_mutex_key#8 jbd2_handle &ei->i_data_sem &ei->i_es_lock &____s->seqcount#2 irq_context: 0 &f->f_pos_lock sb_writers#4 &sb->s_type->i_mutex_key#8 jbd2_handle &ei->i_data_sem &ei->i_es_lock &____s->seqcount irq_context: 0 &f->f_pos_lock sb_writers#4 &sb->s_type->i_mutex_key#8 &ei->i_es_lock key#2 irq_context: 0 sb_writers#11 &type->i_mutex_dir_key#7/1 cgroup_mutex &root->kernfs_rwsem &root->kernfs_iattr_rwsem &rq->__lock irq_context: 0 &sb->s_type->i_mutex_key#8 &rq->__lock &cfs_rq->removed.lock irq_context: 0 &sb->s_type->i_mutex_key#8 bit_wait_table + i irq_context: 0 (wq_completion)ext4-rsv-conversion (work_completion)(&ei->i_rsv_conversion_work) jbd2_handle &ei->i_data_sem rcu_node_0 irq_context: 0 sb_writers#4 &sb->s_type->i_mutex_key#8 mapping.invalidate_lock jbd2_handle rcu_read_lock &cfs_rq->removed.lock irq_context: 0 sb_writers#4 &sb->s_type->i_mutex_key#8 mapping.invalidate_lock jbd2_handle rcu_read_lock &obj_hash[i].lock irq_context: 0 sb_writers#4 &sb->s_type->i_mutex_key#8 mapping.invalidate_lock jbd2_handle rcu_read_lock &rq->__lock &cfs_rq->removed.lock irq_context: 0 &mm->mmap_lock quarantine_lock irq_context: 0 sb_writers#4 &type->i_mutex_dir_key#3/1 &dentry->d_lock &lru->node[i].lock rcu_read_lock &p->pi_lock irq_context: 0 rtnl_mutex nl_table_wait.lock &p->pi_lock &rq->__lock irq_context: 0 rtnl_mutex nl_table_wait.lock &p->pi_lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)wg-kex-wg1#9 (work_completion)(&peer->transmit_handshake_work) &peer->endpoint_lock rcu_read_lock_bh &c->lock irq_context: 0 (wq_completion)wg-kex-wg0#9 (work_completion)(&peer->transmit_handshake_work) &peer->endpoint_lock rcu_read_lock_bh &n->list_lock irq_context: 0 (wq_completion)wg-kex-wg0#9 (work_completion)(&peer->transmit_handshake_work) &peer->endpoint_lock rcu_read_lock_bh &n->list_lock &c->lock irq_context: 0 &sb->s_type->i_mutex_key#8 ima_extend_list_mutex &____s->seqcount#2 irq_context: 0 &sb->s_type->i_mutex_key#8 ima_extend_list_mutex &____s->seqcount irq_context: 0 cb_lock nlk_cb_mutex-GENERIC genl_mutex smcd_dev_list.mutex irq_context: 0 &sb->s_type->i_mutex_key#8 &sbi->s_writepages_rwsem remove_cache_srcu rcu_read_lock rcu_node_0 irq_context: 0 &sb->s_type->i_mutex_key#8 &sbi->s_writepages_rwsem remove_cache_srcu rcu_read_lock &rq->__lock irq_context: 0 (wq_completion)ext4-rsv-conversion (work_completion)(&ei->i_rsv_conversion_work) &journal->j_state_lock &journal->j_wait_transaction_locked irq_context: 0 &sb->s_type->i_mutex_key#8 &sbi->s_writepages_rwsem jbd2_handle &ei->i_data_sem &bgl->locks[i].lock irq_context: 0 sb_writers#4 &lock->wait_lock irq_context: 0 (wq_completion)events_unbound (stats_flush_dwork).work &rq->__lock irq_context: 0 sb_writers#11 &type->i_mutex_dir_key#7 &dentry->d_lock &wq#2 irq_context: 0 &pipe->mutex/1 rcu_read_lock &cfs_rq->removed.lock irq_context: 0 &pipe->mutex/1 rcu_read_lock &obj_hash[i].lock irq_context: 0 &sb->s_type->i_mutex_key#8 &sbi->s_writepages_rwsem &cfs_rq->removed.lock irq_context: 0 rtnl_mutex &macsec_netdev_addr_lock_key/1 pool_lock#2 irq_context: 0 rtnl_mutex &macsec_netdev_addr_lock_key/1 _xmit_ETHER &c->lock irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem rcu_read_lock &rcu_state.gp_wq &p->pi_lock &rq->__lock &cfs_rq->removed.lock irq_context: 0 &sb->s_type->i_mutex_key#8 &sbi->s_writepages_rwsem &pcp->lock &zone->lock irq_context: 0 &sb->s_type->i_mutex_key#8 &sbi->s_writepages_rwsem &pcp->lock &zone->lock &____s->seqcount irq_context: 0 &sb->s_type->i_mutex_key#8 rcu_read_lock &cfs_rq->removed.lock irq_context: 0 (wq_completion)wg-kex-wg2 (work_completion)(&peer->transmit_handshake_work) &wg->static_identity.lock &handshake->lock rcu_read_lock_bh batched_entropy_u32.lock crngs.lock irq_context: 0 sb_writers#4 sb_internal jbd2_handle &sbi->s_orphan_lock rcu_read_lock &rq->__lock &cfs_rq->removed.lock irq_context: 0 sb_writers#8 &of->mutex kn->active#5 uevent_sock_mutex rcu_read_lock &rcu_state.expedited_wq irq_context: 0 sb_writers#8 &of->mutex kn->active#5 uevent_sock_mutex rcu_read_lock &rcu_state.expedited_wq &p->pi_lock irq_context: 0 sb_writers#8 &of->mutex kn->active#5 uevent_sock_mutex rcu_read_lock &rcu_state.expedited_wq &p->pi_lock &rq->__lock irq_context: 0 sb_writers#8 &of->mutex kn->active#5 uevent_sock_mutex rcu_read_lock &rcu_state.expedited_wq &p->pi_lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 &type->i_mutex_dir_key#4 remove_cache_srcu pool_lock#2 irq_context: 0 rtnl_mutex &root->kernfs_rwsem &obj_hash[i].lock pool_lock irq_context: 0 &mm->mmap_lock &xa->xa_lock#9 stock_lock rcu_read_lock per_cpu_ptr(&cgroup_rstat_cpu_lock, cpu) irq_context: 0 (wq_completion)wg-kex-wg2#11 (work_completion)(&peer->transmit_handshake_work) &n->list_lock irq_context: 0 (wq_completion)wg-kex-wg2#11 (work_completion)(&peer->transmit_handshake_work) &n->list_lock &c->lock irq_context: 0 &type->s_umount_key#23/1 rcu_node_0 irq_context: 0 &type->s_umount_key#23/1 rcu_read_lock &rcu_state.expedited_wq irq_context: 0 &type->s_umount_key#23/1 rcu_read_lock &rcu_state.expedited_wq &p->pi_lock irq_context: 0 &type->s_umount_key#23/1 rcu_read_lock &rcu_state.expedited_wq &p->pi_lock &rq->__lock irq_context: 0 &type->s_umount_key#23/1 rcu_read_lock &rcu_state.expedited_wq &p->pi_lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 &type->s_umount_key#23/1 rcu_read_lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 stack_depot_init_mutex rcu_node_0 irq_context: 0 stack_depot_init_mutex &rq->__lock irq_context: 0 stack_depot_init_mutex &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 tracepoints_mutex &p->pi_lock &cfs_rq->removed.lock irq_context: 0 misc_mtx fs_reclaim &rq->__lock irq_context: 0 misc_mtx fs_reclaim &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem rtnl_mutex fs_reclaim &rq->__lock irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem rtnl_mutex fs_reclaim &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 sb_writers#4 &sb->s_type->i_mutex_key#8 &ei->i_data_sem irq_context: 0 sb_writers#4 &sbi->s_writepages_rwsem &xa->xa_lock#9 irq_context: 0 sb_writers#4 &sbi->s_writepages_rwsem rcu_read_lock &memcg->move_lock irq_context: 0 sb_writers#4 &sbi->s_writepages_rwsem rcu_read_lock &xa->xa_lock#9 irq_context: 0 sb_writers#4 &sbi->s_writepages_rwsem rcu_read_lock &xa->xa_lock#9 &s->s_inode_wblist_lock irq_context: 0 sb_writers#4 &journal->j_state_lock &journal->j_wait_commit irq_context: 0 sb_writers#4 &journal->j_state_lock &journal->j_wait_commit &p->pi_lock irq_context: 0 sb_writers#4 &journal->j_state_lock &journal->j_wait_commit &p->pi_lock &rq->__lock irq_context: 0 sb_writers#4 &journal->j_state_lock &journal->j_wait_commit &p->pi_lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 sb_writers#4 &journal->j_wait_commit irq_context: 0 sb_writers#4 &journal->j_wait_done_commit irq_context: 0 sb_writers#4 &sbi->s_writepages_rwsem &____s->seqcount#2 irq_context: 0 sb_writers#4 &sbi->s_writepages_rwsem &____s->seqcount irq_context: 0 sb_writers#4 &sbi->s_writepages_rwsem &folio_wait_table[i] irq_context: 0 sb_writers#4 &sbi->s_writepages_rwsem rcu_read_lock fill_pool_map-wait-type-override pool_lock#2 irq_context: 0 sb_writers#4 &sbi->s_writepages_rwsem rcu_read_lock fill_pool_map-wait-type-override &c->lock irq_context: 0 sb_writers#4 &sbi->s_writepages_rwsem rcu_read_lock fill_pool_map-wait-type-override &n->list_lock irq_context: 0 sb_writers#4 &sbi->s_writepages_rwsem rcu_read_lock fill_pool_map-wait-type-override &n->list_lock &c->lock irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem rtnl_mutex fill_pool_map-wait-type-override &n->list_lock irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem rtnl_mutex fill_pool_map-wait-type-override &n->list_lock &c->lock irq_context: 0 sb_writers#4 &sbi->s_writepages_rwsem rcu_read_lock fill_pool_map-wait-type-override pool_lock irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem rtnl_mutex &p->pi_lock &cfs_rq->removed.lock irq_context: 0 sb_writers#11 &type->i_mutex_dir_key#7/1 tomoyo_ss &rq->__lock irq_context: 0 sb_writers#11 &type->i_mutex_dir_key#7/1 tomoyo_ss &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 sb_writers#4 &type->i_mutex_dir_key#3/1 &sb->s_type->i_mutex_key#8 jbd2_handle &journal->j_wait_updates &p->pi_lock &cfs_rq->removed.lock irq_context: 0 sb_writers#11 &type->i_mutex_dir_key#7/1 cgroup_mutex per_cpu_ptr(&cgroup_rstat_cpu_lock, cpu) irq_context: 0 sb_writers#11 &type->i_mutex_dir_key#7/1 cgroup_mutex &root->kernfs_rwsem &root->kernfs_iattr_rwsem &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 sb_writers#11 &type->i_mutex_dir_key#7/1 cgroup_mutex remove_cache_srcu pool_lock#2 irq_context: 0 sb_writers#4 &journal->j_state_lock &journal->j_wait_commit &p->pi_lock &cfs_rq->removed.lock irq_context: 0 &iint->mutex &ei->xattr_sem &rq->__lock irq_context: 0 &iint->mutex &ei->xattr_sem &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 &group->mark_mutex lock &group->inotify_data.idr_lock &____s->seqcount#2 irq_context: 0 sb_writers#4 &fq->mq_flush_lock irq_context: 0 sb_writers#4 &fq->mq_flush_lock tk_core.seq.seqcount irq_context: 0 sb_writers#4 &fq->mq_flush_lock &q->requeue_lock irq_context: 0 sb_writers#4 &fq->mq_flush_lock &obj_hash[i].lock irq_context: 0 sb_writers#4 &fq->mq_flush_lock rcu_read_lock &pool->lock irq_context: 0 sb_writers#4 &fq->mq_flush_lock rcu_read_lock &pool->lock &obj_hash[i].lock irq_context: 0 sb_writers#4 &fq->mq_flush_lock rcu_read_lock &pool->lock pool_lock#2 irq_context: 0 sb_writers#4 &fq->mq_flush_lock rcu_read_lock &pool->lock &p->pi_lock irq_context: 0 sb_writers#4 &fq->mq_flush_lock rcu_read_lock &pool->lock &p->pi_lock &rq->__lock irq_context: 0 sb_writers#4 &fq->mq_flush_lock rcu_read_lock &pool->lock &p->pi_lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 sb_writers#4 &x->wait#26 irq_context: 0 sb_writers#4 (&timer.timer) irq_context: 0 sb_writers#4 &sbi->s_writepages_rwsem &folio_wait_table[i] &p->pi_lock irq_context: 0 sb_writers#4 &sbi->s_writepages_rwsem &folio_wait_table[i] &p->pi_lock &rq->__lock irq_context: 0 sb_writers#4 &sbi->s_writepages_rwsem &folio_wait_table[i] &p->pi_lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 &sb->s_type->i_mutex_key#8 &sbi->s_writepages_rwsem jbd2_handle &ei->i_data_sem rcu_read_lock pool_lock#2 irq_context: 0 &journal->j_state_lock &journal->j_wait_done_commit &p->pi_lock irq_context: 0 sb_writers#4 &sbi->s_writepages_rwsem &folio_wait_table[i] &p->pi_lock &cfs_rq->removed.lock irq_context: 0 sb_writers#4 &sbi->s_writepages_rwsem &rq->__lock &cfs_rq->removed.lock irq_context: 0 (wq_completion)ext4-rsv-conversion (work_completion)(&ei->i_rsv_conversion_work) jbd2_handle &ei->i_data_sem &rq->__lock &cfs_rq->removed.lock irq_context: 0 (wq_completion)kblockd (work_completion)(&(&q->requeue_work)->work) rcu_read_lock &base->lock irq_context: 0 (wq_completion)kblockd (work_completion)(&(&q->requeue_work)->work) rcu_read_lock &base->lock &obj_hash[i].lock irq_context: 0 sb_writers#4 &sbi->s_writepages_rwsem rcu_read_lock &xa->xa_lock#9 key#10 irq_context: 0 sb_writers#4 &sbi->s_writepages_rwsem &cfs_rq->removed.lock irq_context: 0 sb_writers#4 &sbi->s_writepages_rwsem rcu_read_lock &obj_hash[i].lock pool_lock irq_context: 0 link_idr_lock fill_pool_map-wait-type-override pool_lock#2 irq_context: 0 link_idr_lock fill_pool_map-wait-type-override &c->lock irq_context: 0 link_idr_lock fill_pool_map-wait-type-override pool_lock irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem rtnl_mutex rcu_state.exp_mutex fill_pool_map-wait-type-override &c->lock irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem rtnl_mutex rcu_state.exp_mutex fill_pool_map-wait-type-override pool_lock#2 irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem rtnl_mutex rcu_state.exp_mutex fill_pool_map-wait-type-override &n->list_lock irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem rtnl_mutex rcu_state.exp_mutex fill_pool_map-wait-type-override &n->list_lock &c->lock irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem rtnl_mutex rcu_state.exp_mutex fill_pool_map-wait-type-override pool_lock irq_context: 0 sb_writers#4 &type->i_mutex_dir_key#3/1 sb_internal remove_cache_srcu &pcp->lock &zone->lock irq_context: 0 sb_writers#4 &type->i_mutex_dir_key#3/1 sb_internal remove_cache_srcu &pcp->lock &zone->lock &____s->seqcount irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem &net->sctp.addr_wq_lock k-slock-AF_INET6/1 key#25 irq_context: 0 pernet_ops_rwsem rtnl_mutex &net->xdp.lock &rq->__lock irq_context: 0 pernet_ops_rwsem rtnl_mutex cpu_hotplug_lock xps_map_mutex &rq->__lock irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem rcu_state.exp_mutex fill_pool_map-wait-type-override pool_lock#2 irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem rcu_state.exp_mutex fill_pool_map-wait-type-override &c->lock irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem rcu_state.exp_mutex fill_pool_map-wait-type-override pool_lock irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem sysctl_lock fill_pool_map-wait-type-override &n->list_lock irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem sysctl_lock fill_pool_map-wait-type-override &n->list_lock &c->lock irq_context: 0 &f->f_pos_lock sb_writers#4 &sbi->s_writepages_rwsem rcu_read_lock &xa->xa_lock#9 key#10 irq_context: 0 sb_writers#4 &sbi->s_writepages_rwsem &n->list_lock irq_context: 0 sb_writers#4 &sbi->s_writepages_rwsem &n->list_lock &c->lock irq_context: 0 (wq_completion)wg-kex-wg2 (work_completion)(&peer->transmit_handshake_work) &rq->__lock irq_context: 0 (wq_completion)wg-kex-wg2 (work_completion)(&peer->transmit_handshake_work) &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 &journal->j_state_lock &journal->j_wait_done_commit &p->pi_lock &rq->__lock irq_context: 0 &journal->j_state_lock &journal->j_wait_done_commit &p->pi_lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 sb_writers#4 &sbi->s_writepages_rwsem &pcp->lock &zone->lock irq_context: 0 rtnl_mutex rcu_read_lock &tb->tb6_lock rcu_read_lock &____s->seqcount irq_context: 0 rtnl_mutex rcu_read_lock &tb->tb6_lock rcu_read_lock rcu_read_lock pool_lock#2 irq_context: 0 rtnl_mutex rcu_read_lock &tb->tb6_lock rcu_read_lock &obj_hash[i].lock irq_context: 0 (wq_completion)ext4-rsv-conversion (work_completion)(&ei->i_rsv_conversion_work) batched_entropy_u8.lock irq_context: 0 sb_writers#4 &sbi->s_writepages_rwsem rcu_read_lock &base->lock irq_context: 0 sb_writers#4 &sbi->s_writepages_rwsem rcu_read_lock &base->lock &obj_hash[i].lock irq_context: 0 &ret->b_state_lock &journal->j_list_lock &meta->lock irq_context: 0 &ret->b_state_lock &journal->j_list_lock kfence_freelist_lock irq_context: 0 sb_writers#4 &sbi->s_writepages_rwsem &pcp->lock &zone->lock &____s->seqcount irq_context: 0 &sb->s_type->i_mutex_key#8 &sbi->s_writepages_rwsem jbd2_handle &ei->i_data_sem &ret->b_state_lock bit_wait_table + i irq_context: 0 tracepoints_mutex fill_pool_map-wait-type-override &rq->__lock irq_context: 0 tracepoints_mutex fill_pool_map-wait-type-override &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 &sb->s_type->i_mutex_key#8 &sbi->s_writepages_rwsem rcu_read_lock &hctx->lock irq_context: 0 sb_writers#4 sb_writers#4 &pcp->lock &zone->lock irq_context: 0 sb_writers#4 sb_writers#4 &pcp->lock &zone->lock &____s->seqcount irq_context: 0 sb_writers#4 &sbi->s_writepages_rwsem rcu_node_0 irq_context: 0 sb_writers#11 &type->i_mutex_dir_key#7/1 cgroup_mutex remove_cache_srcu &pcp->lock &zone->lock irq_context: 0 sb_writers#11 &type->i_mutex_dir_key#7/1 cgroup_mutex remove_cache_srcu &pcp->lock &zone->lock &____s->seqcount irq_context: 0 sb_writers#11 &type->i_mutex_dir_key#7/1 cgroup_mutex remove_cache_srcu rcu_read_lock rcu_node_0 irq_context: 0 sb_writers#11 &type->i_mutex_dir_key#7/1 cgroup_mutex remove_cache_srcu rcu_read_lock &rq->__lock irq_context: 0 rcu_read_lock fill_pool_map-wait-type-override rcu_node_0 irq_context: 0 rcu_read_lock fill_pool_map-wait-type-override &rq->__lock irq_context: 0 rcu_read_lock fill_pool_map-wait-type-override &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 sb_writers#4 sb_writers#4 jbd2_handle &ret->b_state_lock bit_wait_table + i irq_context: 0 sb_writers#4 sb_writers#4 jbd2_handle &ret->b_state_lock bit_wait_table + i &p->pi_lock irq_context: 0 sb_writers#4 sb_writers#4 jbd2_handle &ret->b_state_lock bit_wait_table + i &p->pi_lock &rq->__lock irq_context: 0 sb_writers#4 sb_writers#4 jbd2_handle &ret->b_state_lock bit_wait_table + i &p->pi_lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)bat_events (work_completion)(&(&forw_packet_aggr->delayed_work)->work) &hard_iface->bat_iv.ogm_buff_mutex fill_pool_map-wait-type-override &rq->__lock irq_context: 0 &sb->s_type->i_mutex_key#8 &sbi->s_writepages_rwsem jbd2_handle &ei->i_data_sem &lg->lg_mutex &____s->seqcount irq_context: 0 &sb->s_type->i_mutex_key#8 rcu_read_lock &hctx->lock irq_context: 0 &sb->s_type->i_mutex_key#8 &sbi->s_writepages_rwsem jbd2_handle &ei->i_data_sem &lg->lg_mutex rcu_read_lock pool_lock#2 irq_context: 0 &sb->s_type->i_mutex_key#8 &sbi->s_writepages_rwsem jbd2_handle &ei->i_data_sem &lg->lg_mutex rcu_read_lock rcu_node_0 irq_context: 0 &sb->s_type->i_mutex_key#8 &sbi->s_writepages_rwsem jbd2_handle &ei->i_data_sem &lg->lg_mutex rcu_read_lock &rq->__lock irq_context: 0 &sb->s_type->i_mutex_key#8 &sbi->s_writepages_rwsem jbd2_handle &ei->i_data_sem &lg->lg_mutex rcu_read_lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 &sb->s_type->i_mutex_key#8 &sbi->s_writepages_rwsem jbd2_handle &ei->i_data_sem &lg->lg_mutex &obj_hash[i].lock irq_context: 0 &sb->s_type->i_lock_key#23 bit_wait_table + i irq_context: 0 tracepoints_mutex &rnp->exp_lock irq_context: 0 tracepoints_mutex rcu_state.exp_mutex irq_context: 0 tracepoints_mutex rcu_state.exp_mutex rcu_state.exp_mutex.wait_lock irq_context: 0 sb_writers#4 &sb->s_type->i_mutex_key#8 &xa->xa_lock#9 stock_lock irq_context: 0 sb_writers#4 &sbi->s_writepages_rwsem jbd2_handle &ei->i_data_sem &____s->seqcount#2 irq_context: 0 sb_writers#4 &sbi->s_writepages_rwsem jbd2_handle &ei->i_data_sem &____s->seqcount irq_context: 0 sb_writers#4 &sbi->s_writepages_rwsem jbd2_handle &____s->seqcount#2 irq_context: 0 sb_writers#4 &sbi->s_writepages_rwsem jbd2_handle &____s->seqcount irq_context: 0 &sb->s_type->i_mutex_key#8 lock#4 rcu_read_lock pool_lock#2 irq_context: 0 &sb->s_type->i_mutex_key#8 lock#4 &obj_hash[i].lock irq_context: 0 (wq_completion)wg-kex-wg0#9 (work_completion)(&peer->transmit_handshake_work) &c->lock irq_context: 0 (wq_completion)wg-kex-wg0#9 (work_completion)(&peer->transmit_handshake_work) &n->list_lock irq_context: 0 (wq_completion)wg-kex-wg0#9 (work_completion)(&peer->transmit_handshake_work) &n->list_lock &c->lock irq_context: 0 pernet_ops_rwsem cpu_hotplug_lock wq_pool_mutex fill_pool_map-wait-type-override &n->list_lock irq_context: 0 pernet_ops_rwsem cpu_hotplug_lock wq_pool_mutex fill_pool_map-wait-type-override &n->list_lock &c->lock irq_context: 0 (wq_completion)events (work_completion)(&data->fib_event_work) &data->fib_lock remove_cache_srcu &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 rcu_read_lock_bh rcu_read_lock &____s->seqcount#2 irq_context: 0 pernet_ops_rwsem rtnl_mutex remove_cache_srcu rcu_node_0 irq_context: 0 pernet_ops_rwsem rtnl_mutex remove_cache_srcu &rcu_state.expedited_wq irq_context: 0 pernet_ops_rwsem rtnl_mutex remove_cache_srcu &rcu_state.expedited_wq &p->pi_lock irq_context: 0 pernet_ops_rwsem rtnl_mutex remove_cache_srcu &rcu_state.expedited_wq &p->pi_lock &rq->__lock irq_context: 0 pernet_ops_rwsem rtnl_mutex remove_cache_srcu &rcu_state.expedited_wq &p->pi_lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 pernet_ops_rwsem rtnl_mutex &root->kernfs_rwsem rcu_node_0 irq_context: 0 pernet_ops_rwsem rtnl_mutex &root->kernfs_rwsem &rcu_state.expedited_wq irq_context: 0 pernet_ops_rwsem rtnl_mutex &root->kernfs_rwsem &rcu_state.expedited_wq &p->pi_lock irq_context: 0 pernet_ops_rwsem rtnl_mutex &root->kernfs_rwsem &rcu_state.expedited_wq &p->pi_lock &rq->__lock irq_context: 0 pernet_ops_rwsem rtnl_mutex &root->kernfs_rwsem &rcu_state.expedited_wq &p->pi_lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)bat_events (work_completion)(&(&bat_priv->mcast.work)->work) quarantine_lock irq_context: 0 &sb->s_type->i_mutex_key#8 &sbi->s_writepages_rwsem rcu_read_lock &p->pi_lock irq_context: 0 &sb->s_type->i_mutex_key#8 &sbi->s_writepages_rwsem &sem->waiters irq_context: 0 &sb->s_type->i_mutex_key#8 &sbi->s_writepages_rwsem &sem->waiters rcu_read_lock &p->pi_lock irq_context: 0 sb_writers#4 &sb->s_type->i_mutex_key#8 &sem->waiters &p->pi_lock irq_context: 0 sb_writers#4 &sb->s_type->i_mutex_key#8 &sem->waiters &p->pi_lock &rq->__lock irq_context: 0 sb_writers#4 &sb->s_type->i_mutex_key#8 &sem->waiters &p->pi_lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 pernet_ops_rwsem rtnl_mutex stack_depot_init_mutex &rq->__lock irq_context: 0 pernet_ops_rwsem rtnl_mutex stack_depot_init_mutex &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)ext4-rsv-conversion (work_completion)(&ei->i_rsv_conversion_work) jbd2_handle &ei->i_data_sem quarantine_lock irq_context: 0 &sb->s_type->i_mutex_key#10 &x->wait#10 irq_context: 0 pernet_ops_rwsem fs_reclaim mmu_notifier_invalidate_range_start &cfs_rq->removed.lock irq_context: 0 pernet_ops_rwsem fs_reclaim mmu_notifier_invalidate_range_start &obj_hash[i].lock irq_context: 0 pernet_ops_rwsem devices_rwsem rdma_nets_rwsem &device->compat_devs_mutex lock kernfs_idr_lock &pcp->lock &zone->lock &____s->seqcount irq_context: 0 sb_writers#4 &sb->s_type->i_mutex_key#8 rcu_read_lock &rq->__lock &cfs_rq->removed.lock irq_context: 0 sb_writers#4 &sb->s_type->i_mutex_key#8 rcu_read_lock &cfs_rq->removed.lock irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem rtnl_mutex &root->kernfs_rwsem &rq->__lock &cfs_rq->removed.lock irq_context: 0 vlan_ioctl_mutex rtnl_mutex fill_pool_map-wait-type-override pool_lock#2 irq_context: 0 vlan_ioctl_mutex rtnl_mutex fill_pool_map-wait-type-override &c->lock irq_context: 0 vlan_ioctl_mutex rtnl_mutex fill_pool_map-wait-type-override pool_lock irq_context: 0 &f->f_pos_lock sb_writers#4 &sb->s_type->i_mutex_key#8 remove_cache_srcu &rcu_state.expedited_wq irq_context: 0 &f->f_pos_lock sb_writers#4 &sb->s_type->i_mutex_key#8 remove_cache_srcu &rcu_state.expedited_wq &p->pi_lock irq_context: 0 &f->f_pos_lock sb_writers#4 &sb->s_type->i_mutex_key#8 remove_cache_srcu &rcu_state.expedited_wq &p->pi_lock &rq->__lock irq_context: 0 &f->f_pos_lock sb_writers#4 &sb->s_type->i_mutex_key#8 remove_cache_srcu &rcu_state.expedited_wq &p->pi_lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 &f->f_pos_lock sb_writers#4 &sb->s_type->i_mutex_key#8 &ei->i_data_sem &rcu_state.expedited_wq &p->pi_lock &rq->__lock irq_context: 0 &f->f_pos_lock sb_writers#4 &sb->s_type->i_mutex_key#8 &ei->i_data_sem &rcu_state.expedited_wq &p->pi_lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem &root->kernfs_rwsem &root->kernfs_iattr_rwsem &rq->__lock &cfs_rq->removed.lock irq_context: 0 (wq_completion)wg-kex-wg2#11 (work_completion)(&peer->transmit_handshake_work) &peer->endpoint_lock rcu_read_lock_bh &n->list_lock irq_context: 0 (wq_completion)wg-kex-wg2#11 (work_completion)(&peer->transmit_handshake_work) &peer->endpoint_lock rcu_read_lock_bh &n->list_lock &c->lock irq_context: softirq (&tsc_sync_check_timer) irq_context: softirq (&tsc_sync_check_timer) &obj_hash[i].lock irq_context: softirq (&tsc_sync_check_timer) &base->lock irq_context: softirq (&tsc_sync_check_timer) &base->lock &obj_hash[i].lock irq_context: 0 (wq_completion)events (work_completion)(&rdev->wiphy_work) &rdev->wiphy.mtx &wdev->mtx &local->sta_mtx &rnp->exp_lock irq_context: 0 (wq_completion)events (work_completion)(&rdev->wiphy_work) &rdev->wiphy.mtx &wdev->mtx &local->sta_mtx rcu_state.exp_mutex irq_context: 0 (wq_completion)events (work_completion)(&rdev->wiphy_work) &rdev->wiphy.mtx &wdev->mtx &local->sta_mtx rcu_state.exp_mutex rcu_state.exp_mutex.wait_lock irq_context: 0 (wq_completion)events (work_completion)(&rdev->wiphy_work) &rdev->wiphy.mtx &wdev->mtx &local->sta_mtx &meta->lock irq_context: 0 (wq_completion)wg-kex-wg2 (work_completion)(&peer->transmit_handshake_work) &peer->endpoint_lock rcu_read_lock_bh &____s->seqcount#2 irq_context: 0 (wq_completion)wg-kex-wg2 (work_completion)(&peer->transmit_handshake_work) &peer->endpoint_lock rcu_read_lock_bh &____s->seqcount irq_context: 0 sk_lock-AF_TIPC &srv->idr_lock irq_context: 0 sk_lock-AF_TIPC &srv->idr_lock pool_lock#2 irq_context: 0 sk_lock-AF_TIPC &tn->nametbl_lock irq_context: 0 sk_lock-AF_TIPC &tn->nametbl_lock &c->lock irq_context: 0 sk_lock-AF_TIPC &tn->nametbl_lock pool_lock#2 irq_context: 0 sk_lock-AF_TIPC &tn->nametbl_lock &service->lock irq_context: 0 sk_lock-AF_TIPC &con->sub_lock irq_context: 0 sk_lock-AF_TIPC &tn->nametbl_lock &service->lock pool_lock#2 irq_context: 0 sk_lock-AF_TIPC &tn->nametbl_lock &service->lock &sub->lock irq_context: 0 sk_lock-AF_TIPC &tn->nametbl_lock &service->lock &sub->lock &srv->idr_lock irq_context: 0 sk_lock-AF_TIPC &tn->nametbl_lock &service->lock &sub->lock pool_lock#2 irq_context: 0 sk_lock-AF_TIPC &tn->nametbl_lock &service->lock &sub->lock &con->outqueue_lock irq_context: 0 sk_lock-AF_TIPC &tn->nametbl_lock &service->lock &sub->lock rcu_read_lock &pool->lock/1 irq_context: 0 sk_lock-AF_TIPC &tn->nametbl_lock &service->lock &sub->lock rcu_read_lock &pool->lock/1 &obj_hash[i].lock irq_context: 0 sk_lock-AF_TIPC &tn->nametbl_lock &service->lock &sub->lock rcu_read_lock &pool->lock/1 pool_lock#2 irq_context: 0 sk_lock-AF_TIPC &tn->nametbl_lock &service->lock &sub->lock rcu_read_lock &pool->lock/1 &p->pi_lock irq_context: 0 sk_lock-AF_TIPC &tn->nametbl_lock &nt->cluster_scope_lock irq_context: 0 (wq_completion)tipc_send irq_context: 0 (wq_completion)tipc_send (work_completion)(&con->swork) irq_context: 0 (wq_completion)tipc_send (work_completion)(&con->swork) &con->outqueue_lock irq_context: 0 (wq_completion)tipc_send (work_completion)(&con->swork) pool_lock#2 irq_context: 0 (wq_completion)tipc_send (work_completion)(&con->swork) &list->lock#38 irq_context: 0 (wq_completion)tipc_send (work_completion)(&con->swork) slock-AF_TIPC &list->lock#38 irq_context: 0 (wq_completion)tipc_send (work_completion)(&con->swork) &con->outqueue_lock &obj_hash[i].lock irq_context: 0 (wq_completion)tipc_send (work_completion)(&con->swork) &con->outqueue_lock pool_lock#2 irq_context: 0 sk_lock-AF_TIPC &tipc_net(net)->bclock irq_context: 0 sk_lock-AF_TIPC rcu_read_lock &service->lock pool_lock#2 irq_context: 0 sk_lock-AF_TIPC &tn->nametbl_lock &service->lock &sub->lock rcu_read_lock &pool->lock/1 &p->pi_lock &rq->__lock irq_context: 0 sk_lock-AF_TIPC &tn->nametbl_lock &service->lock &sub->lock rcu_read_lock &pool->lock/1 &p->pi_lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)tipc_send (work_completion)(&con->swork) slock-AF_TIPC &c->lock irq_context: 0 (wq_completion)tipc_send (work_completion)(&con->swork) slock-AF_TIPC pool_lock#2 irq_context: 0 (wq_completion)tipc_send (work_completion)(&con->swork) slock-AF_TIPC &obj_hash[i].lock irq_context: 0 (wq_completion)tipc_send (work_completion)(&con->swork) &list->lock#31 irq_context: 0 (wq_completion)tipc_send (work_completion)(&con->swork) slock-AF_TIPC &list->lock#31 irq_context: 0 sk_lock-AF_TIPC slock-AF_TIPC pool_lock#2 irq_context: 0 sk_lock-AF_TIPC slock-AF_TIPC &____s->seqcount irq_context: 0 sk_lock-AF_TIPC slock-AF_TIPC rcu_read_lock pool_lock#2 irq_context: 0 sk_lock-AF_TIPC slock-AF_TIPC &obj_hash[i].lock irq_context: 0 cb_lock genl_mutex rtnl_mutex console_lock console_srcu console_owner_lock irq_context: 0 cb_lock genl_mutex rtnl_mutex console_lock console_srcu console_owner irq_context: 0 cb_lock genl_mutex rtnl_mutex console_lock console_srcu console_owner &port_lock_key irq_context: 0 cb_lock genl_mutex rtnl_mutex console_lock console_srcu console_owner console_owner_lock irq_context: 0 sk_lock-AF_TIPC rcu_read_lock &service->lock &c->lock irq_context: 0 sk_lock-AF_TIPC &tn->nametbl_lock &service->lock &sub->lock rcu_read_lock &pool->lock/1 &p->pi_lock &cfs_rq->removed.lock irq_context: 0 sk_lock-AF_TIPC &tn->nametbl_lock &service->lock &sub->lock &c->lock irq_context: 0 sk_lock-AF_TIPC &con->sub_lock &tn->nametbl_lock irq_context: 0 sk_lock-AF_TIPC &con->sub_lock &tn->nametbl_lock &service->lock irq_context: 0 sk_lock-AF_TIPC &con->sub_lock &obj_hash[i].lock irq_context: 0 sk_lock-AF_TIPC &con->sub_lock pool_lock#2 irq_context: 0 sk_lock-AF_TIPC &con->outqueue_lock irq_context: 0 sk_lock-AF_TIPC &tn->nametbl_lock &obj_hash[i].lock irq_context: 0 sk_lock-AF_TIPC &tn->nametbl_lock krc.lock irq_context: 0 cb_lock genl_mutex rtnl_mutex &tn->nametbl_lock irq_context: 0 cb_lock genl_mutex rtnl_mutex &ht->lock irq_context: 0 cb_lock genl_mutex rtnl_mutex rcu_read_lock &ht->lock irq_context: 0 cb_lock genl_mutex rtnl_mutex rcu_read_lock rcu_node_0 irq_context: 0 cb_lock genl_mutex rtnl_mutex sk_lock-AF_TIPC irq_context: 0 cb_lock genl_mutex rtnl_mutex sk_lock-AF_TIPC slock-AF_TIPC irq_context: 0 cb_lock genl_mutex rtnl_mutex slock-AF_TIPC irq_context: 0 cb_lock genl_mutex rtnl_mutex k-sk_lock-AF_TIPC irq_context: 0 cb_lock genl_mutex rtnl_mutex k-sk_lock-AF_TIPC k-slock-AF_TIPC irq_context: 0 cb_lock genl_mutex rtnl_mutex k-slock-AF_TIPC irq_context: 0 cb_lock genl_mutex rtnl_mutex &tn->nametbl_lock pool_lock#2 irq_context: 0 cb_lock genl_mutex rtnl_mutex &tn->nametbl_lock &service->lock irq_context: 0 cb_lock genl_mutex rtnl_mutex &tn->nametbl_lock &service->lock pool_lock#2 irq_context: 0 cb_lock genl_mutex rtnl_mutex &tn->nametbl_lock &nt->cluster_scope_lock irq_context: 0 cb_lock genl_mutex rtnl_mutex &tipc_net(net)->bclock irq_context: 0 &xt[i].mutex remove_cache_srcu rcu_node_0 irq_context: 0 rtnl_mutex uevent_sock_mutex &obj_hash[i].lock pool_lock irq_context: 0 &sb->s_type->i_mutex_key#8 &sbi->s_writepages_rwsem &journal->j_wait_commit irq_context: 0 &sb->s_type->i_mutex_key#8 &sbi->s_writepages_rwsem &journal->j_wait_done_commit irq_context: 0 &sb->s_type->i_mutex_key#8 &sbi->s_writepages_rwsem &journal->j_state_lock &journal->j_wait_commit irq_context: 0 &sb->s_type->i_mutex_key#10 sk_lock-AF_TIPC slock-AF_TIPC &list->lock#31 irq_context: 0 &sb->s_type->i_mutex_key#10 sk_lock-AF_TIPC slock-AF_TIPC &obj_hash[i].lock irq_context: 0 &sb->s_type->i_mutex_key#10 sk_lock-AF_TIPC slock-AF_TIPC pool_lock#2 irq_context: 0 &sb->s_type->i_mutex_key#10 sk_lock-AF_TIPC &srv->idr_lock irq_context: 0 &sb->s_type->i_mutex_key#10 sk_lock-AF_TIPC &con->sub_lock irq_context: 0 &sb->s_type->i_mutex_key#10 sk_lock-AF_TIPC &con->sub_lock &tn->nametbl_lock irq_context: 0 &sb->s_type->i_mutex_key#10 sk_lock-AF_TIPC &con->sub_lock &tn->nametbl_lock &service->lock irq_context: 0 &sb->s_type->i_mutex_key#10 sk_lock-AF_TIPC &con->sub_lock &obj_hash[i].lock irq_context: 0 &sb->s_type->i_mutex_key#10 sk_lock-AF_TIPC &con->sub_lock pool_lock#2 irq_context: 0 &sb->s_type->i_mutex_key#10 sk_lock-AF_TIPC &rq->__lock irq_context: 0 &sb->s_type->i_mutex_key#10 sk_lock-AF_TIPC &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 &sb->s_type->i_mutex_key#10 sk_lock-AF_TIPC &con->outqueue_lock irq_context: 0 &sb->s_type->i_mutex_key#10 sk_lock-AF_TIPC &tn->nametbl_lock irq_context: 0 &sb->s_type->i_mutex_key#10 sk_lock-AF_TIPC &tn->nametbl_lock &service->lock irq_context: 0 &sb->s_type->i_mutex_key#10 sk_lock-AF_TIPC &tn->nametbl_lock &service->lock &sub->lock irq_context: 0 &sb->s_type->i_mutex_key#10 sk_lock-AF_TIPC &tn->nametbl_lock &service->lock &sub->lock &srv->idr_lock irq_context: 0 &sb->s_type->i_mutex_key#10 sk_lock-AF_TIPC &tn->nametbl_lock &service->lock &sub->lock pool_lock#2 irq_context: 0 &sb->s_type->i_mutex_key#10 sk_lock-AF_TIPC &tn->nametbl_lock &service->lock &sub->lock &con->outqueue_lock irq_context: 0 &sb->s_type->i_mutex_key#10 sk_lock-AF_TIPC &tn->nametbl_lock &service->lock &sub->lock rcu_read_lock &pool->lock/1 irq_context: 0 &sb->s_type->i_mutex_key#10 sk_lock-AF_TIPC &tn->nametbl_lock &service->lock &sub->lock rcu_read_lock &pool->lock/1 &obj_hash[i].lock irq_context: 0 &sb->s_type->i_mutex_key#10 sk_lock-AF_TIPC &tn->nametbl_lock &service->lock &sub->lock rcu_read_lock &pool->lock/1 pool_lock#2 irq_context: 0 &sb->s_type->i_mutex_key#10 sk_lock-AF_TIPC &tn->nametbl_lock &service->lock &sub->lock rcu_read_lock &pool->lock/1 &p->pi_lock irq_context: 0 &sb->s_type->i_mutex_key#10 sk_lock-AF_TIPC &tn->nametbl_lock &service->lock &sub->lock rcu_read_lock &pool->lock/1 &p->pi_lock &rq->__lock irq_context: 0 &sb->s_type->i_mutex_key#10 sk_lock-AF_TIPC &tn->nametbl_lock &service->lock &sub->lock rcu_read_lock &pool->lock/1 &p->pi_lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 &sb->s_type->i_mutex_key#10 sk_lock-AF_TIPC &tn->nametbl_lock &nt->cluster_scope_lock irq_context: 0 &sb->s_type->i_mutex_key#10 sk_lock-AF_TIPC &tn->nametbl_lock pool_lock#2 irq_context: 0 &sb->s_type->i_mutex_key#10 sk_lock-AF_TIPC &tn->nametbl_lock &c->lock irq_context: 0 &sb->s_type->i_mutex_key#10 sk_lock-AF_TIPC &tn->nametbl_lock &____s->seqcount#2 irq_context: 0 &sb->s_type->i_mutex_key#10 sk_lock-AF_TIPC &tn->nametbl_lock &____s->seqcount irq_context: 0 &sb->s_type->i_mutex_key#10 sk_lock-AF_TIPC &tn->nametbl_lock &obj_hash[i].lock irq_context: 0 &sb->s_type->i_mutex_key#10 sk_lock-AF_TIPC &tn->nametbl_lock krc.lock irq_context: 0 &sb->s_type->i_mutex_key#10 sk_lock-AF_TIPC &tipc_net(net)->bclock irq_context: 0 &sb->s_type->i_mutex_key#10 sk_lock-AF_TIPC &srv->idr_lock &obj_hash[i].lock irq_context: 0 &sb->s_type->i_mutex_key#10 sk_lock-AF_TIPC &srv->idr_lock pool_lock#2 irq_context: 0 &sb->s_type->i_mutex_key#10 sk_lock-AF_TIPC &tn->nametbl_lock &service->lock &obj_hash[i].lock irq_context: 0 &sb->s_type->i_mutex_key#10 sk_lock-AF_TIPC &tn->nametbl_lock &service->lock pool_lock#2 irq_context: 0 &sb->s_type->i_mutex_key#10 sk_lock-AF_TIPC &tn->nametbl_lock &service->lock krc.lock irq_context: 0 cb_lock rtnl_mutex rcu_node_0 irq_context: 0 (wq_completion)rcu_gp (work_completion)(&rew->rew_work) rcu_state.exp_wake_mutex &rnp->exp_wq[2] &p->pi_lock &cfs_rq->removed.lock irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem k-sk_lock-AF_TIPC rcu_read_lock rcu_node_0 irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem k-sk_lock-AF_TIPC rcu_read_lock &rq->__lock irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem k-sk_lock-AF_TIPC rcu_read_lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem k-sk_lock-AF_TIPC rcu_read_lock &rcu_state.gp_wq irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem k-sk_lock-AF_TIPC rcu_read_lock &rcu_state.gp_wq &p->pi_lock irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem k-sk_lock-AF_TIPC rcu_read_lock &rcu_state.gp_wq &p->pi_lock &rq->__lock irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem k-sk_lock-AF_TIPC rcu_read_lock &rcu_state.gp_wq &p->pi_lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem rtnl_mutex &root->kernfs_rwsem &cfs_rq->removed.lock irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem rtnl_mutex &root->kernfs_rwsem &obj_hash[i].lock irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem rtnl_mutex &root->kernfs_rwsem pool_lock#2 irq_context: 0 pernet_ops_rwsem rtnl_mutex &pnn->pndevs.lock &rq->__lock irq_context: 0 (wq_completion)wg-kex-wg2#11 (work_completion)(&peer->transmit_handshake_work) kfence_freelist_lock irq_context: 0 (wq_completion)wg-kex-wg2#11 (work_completion)(&peer->transmit_handshake_work) &peer->endpoint_lock rcu_read_lock_bh &meta->lock irq_context: 0 (wq_completion)wg-kex-wg2#11 (work_completion)(&peer->transmit_handshake_work) &peer->endpoint_lock rcu_read_lock_bh kfence_freelist_lock irq_context: 0 sk_lock-AF_TIPC &tn->nametbl_lock &n->list_lock irq_context: 0 sk_lock-AF_TIPC &tn->nametbl_lock &n->list_lock &c->lock irq_context: 0 sk_lock-AF_TIPC &tn->nametbl_lock &service->lock &c->lock irq_context: 0 sk_lock-AF_TIPC &tn->nametbl_lock &service->lock &n->list_lock irq_context: 0 sk_lock-AF_TIPC &tn->nametbl_lock &service->lock &n->list_lock &c->lock irq_context: 0 sk_lock-AF_TIPC slock-AF_TIPC &c->lock irq_context: 0 sk_lock-AF_TIPC slock-AF_TIPC &____s->seqcount#2 irq_context: 0 (wq_completion)tipc_send (work_completion)(&con->swork) &c->lock irq_context: 0 &xs->mutex &obj_hash[i].lock pool_lock irq_context: 0 cb_lock genl_mutex rtnl_mutex &cfs_rq->removed.lock irq_context: 0 cb_lock genl_mutex &p->pi_lock &rq->__lock &cfs_rq->removed.lock irq_context: 0 sk_lock-AF_TIPC &tn->nametbl_lock &____s->seqcount#2 irq_context: 0 sk_lock-AF_TIPC &tn->nametbl_lock &____s->seqcount irq_context: 0 (wq_completion)tipc_send#2 irq_context: 0 (wq_completion)tipc_send#2 (work_completion)(&con->swork) irq_context: 0 (wq_completion)tipc_send#2 (work_completion)(&con->swork) &con->outqueue_lock irq_context: 0 (wq_completion)tipc_send#2 (work_completion)(&con->swork) &list->lock#38 irq_context: 0 (wq_completion)tipc_send#2 (work_completion)(&con->swork) slock-AF_TIPC &list->lock#38 irq_context: 0 (wq_completion)tipc_send#2 (work_completion)(&con->swork) slock-AF_TIPC &obj_hash[i].lock irq_context: 0 (wq_completion)tipc_send#2 (work_completion)(&con->swork) &con->outqueue_lock &obj_hash[i].lock irq_context: 0 (wq_completion)tipc_send#2 (work_completion)(&con->swork) slock-AF_TIPC pool_lock#2 irq_context: 0 (wq_completion)tipc_send#2 (work_completion)(&con->swork) &list->lock#31 irq_context: 0 (wq_completion)tipc_send#2 (work_completion)(&con->swork) slock-AF_TIPC &list->lock#31 irq_context: 0 (wq_completion)tipc_send#2 (work_completion)(&con->swork) slock-AF_TIPC &c->lock irq_context: 0 (wq_completion)tipc_send#2 (work_completion)(&con->swork) slock-AF_TIPC &n->list_lock irq_context: 0 (wq_completion)tipc_send#2 (work_completion)(&con->swork) slock-AF_TIPC &n->list_lock &c->lock irq_context: 0 cb_lock genl_mutex rtnl_mutex &tn->nametbl_lock &c->lock irq_context: 0 cb_lock genl_mutex rtnl_mutex &tn->nametbl_lock &n->list_lock irq_context: 0 cb_lock genl_mutex rtnl_mutex &tn->nametbl_lock &n->list_lock &c->lock irq_context: 0 (wq_completion)tipc_send#2 (work_completion)(&con->swork) &c->lock irq_context: 0 (wq_completion)tipc_send#2 (work_completion)(&con->swork) batched_entropy_u8.lock irq_context: 0 (wq_completion)tipc_send#2 (work_completion)(&con->swork) kfence_freelist_lock irq_context: 0 (wq_completion)tipc_send#2 (work_completion)(&con->swork) slock-AF_TIPC &meta->lock irq_context: 0 (wq_completion)tipc_send#2 (work_completion)(&con->swork) slock-AF_TIPC kfence_freelist_lock irq_context: 0 sb_writers#4 &sb->s_type->i_mutex_key#8 jbd2_handle &ei->i_data_sem remove_cache_srcu &c->lock irq_context: 0 sb_writers#4 &sb->s_type->i_mutex_key#8 jbd2_handle &ei->i_data_sem remove_cache_srcu &n->list_lock irq_context: 0 sb_writers#4 &sb->s_type->i_mutex_key#8 jbd2_handle &ei->i_data_sem remove_cache_srcu rcu_read_lock pool_lock#2 irq_context: 0 sb_writers#4 &sb->s_type->i_mutex_key#8 jbd2_handle &ei->i_data_sem remove_cache_srcu &obj_hash[i].lock irq_context: 0 sb_writers#4 &sb->s_type->i_mutex_key#8 jbd2_handle &ei->i_data_sem remove_cache_srcu &rq->__lock irq_context: 0 sb_writers#4 &sb->s_type->i_mutex_key#8 jbd2_handle &ei->i_data_sem remove_cache_srcu &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 sb_writers#4 &sb->s_type->i_mutex_key#8 jbd2_handle &ei->i_data_sem remove_cache_srcu rcu_read_lock rcu_node_0 irq_context: 0 sb_writers#4 &sb->s_type->i_mutex_key#8 jbd2_handle &ei->i_data_sem remove_cache_srcu rcu_read_lock &rq->__lock irq_context: 0 sb_writers#4 &sb->s_type->i_mutex_key#8 jbd2_handle &ei->i_data_sem remove_cache_srcu rcu_read_lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 sk_lock-AF_ALG &mm->mmap_lock sb_pagefaults jbd2_handle &ret->b_state_lock irq_context: 0 &sb->s_type->i_mutex_key#8 &sbi->s_writepages_rwsem &journal->j_state_lock &journal->j_wait_commit &p->pi_lock irq_context: 0 &sb->s_type->i_mutex_key#8 &sbi->s_writepages_rwsem &journal->j_state_lock &journal->j_wait_commit &p->pi_lock &rq->__lock irq_context: 0 &sb->s_type->i_mutex_key#8 &sbi->s_writepages_rwsem &journal->j_state_lock &journal->j_wait_commit &p->pi_lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 &sb->s_type->i_mutex_key#10 sk_lock-AF_TIPC &c->lock irq_context: 0 &sb->s_type->i_mutex_key#10 sk_lock-AF_TIPC rcu_node_0 irq_context: 0 &sb->s_type->i_mutex_key#10 sk_lock-AF_TIPC &rcu_state.expedited_wq irq_context: 0 &sb->s_type->i_mutex_key#10 sk_lock-AF_TIPC &rcu_state.expedited_wq &p->pi_lock irq_context: 0 &sb->s_type->i_mutex_key#10 sk_lock-AF_TIPC &rcu_state.expedited_wq &p->pi_lock &rq->__lock irq_context: 0 &sb->s_type->i_mutex_key#10 sk_lock-AF_TIPC &rcu_state.expedited_wq &p->pi_lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 &sb->s_type->i_mutex_key#10 sk_lock-AF_TIPC &tn->nametbl_lock &service->lock &sub->lock &c->lock irq_context: 0 &sb->s_type->i_mutex_key#10 sk_lock-AF_TIPC &tn->nametbl_lock &service->lock &sub->lock &n->list_lock irq_context: 0 &sb->s_type->i_mutex_key#10 sk_lock-AF_TIPC &tn->nametbl_lock &service->lock &sub->lock &n->list_lock &c->lock irq_context: 0 (wq_completion)tipc_send (work_completion)(&con->swork) &____s->seqcount#2 irq_context: 0 (wq_completion)tipc_send (work_completion)(&con->swork) &____s->seqcount irq_context: 0 &sb->s_type->i_mutex_key#8 &sbi->s_writepages_rwsem jbd2_handle &ei->i_data_sem &rcu_state.expedited_wq &p->pi_lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 vlan_ioctl_mutex rtnl_mutex rcu_state.exp_mutex &obj_hash[i].lock pool_lock irq_context: 0 (wq_completion)afs (work_completion)(&net->fs_manager) &cfs_rq->removed.lock irq_context: 0 (wq_completion)afs (work_completion)(&net->fs_manager) &obj_hash[i].lock irq_context: 0 (wq_completion)afs (work_completion)(&net->fs_manager) pool_lock#2 irq_context: 0 pernet_ops_rwsem ovs_mutex nf_ct_proto_mutex nf_hook_mutex &n->list_lock irq_context: 0 pernet_ops_rwsem ovs_mutex nf_ct_proto_mutex nf_hook_mutex &n->list_lock &c->lock irq_context: 0 pernet_ops_rwsem rtnl_mutex &root->kernfs_rwsem pgd_lock irq_context: 0 pernet_ops_rwsem rtnl_mutex &root->kernfs_rwsem stock_lock irq_context: 0 cb_lock nlk_cb_mutex-GENERIC rtnl_mutex fs_reclaim irq_context: 0 cb_lock nlk_cb_mutex-GENERIC rtnl_mutex fs_reclaim &rq->__lock irq_context: 0 cb_lock nlk_cb_mutex-GENERIC rtnl_mutex fs_reclaim &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 cb_lock nlk_cb_mutex-GENERIC rtnl_mutex fs_reclaim mmu_notifier_invalidate_range_start irq_context: 0 cb_lock nlk_cb_mutex-GENERIC rtnl_mutex fs_reclaim mmu_notifier_invalidate_range_start &rq->__lock irq_context: 0 cb_lock nlk_cb_mutex-GENERIC rtnl_mutex fs_reclaim mmu_notifier_invalidate_range_start &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 pernet_ops_rwsem rtnl_mutex &root->kernfs_rwsem rcu_read_lock pool_lock#2 irq_context: 0 cb_lock nlk_cb_mutex-GENERIC rtnl_mutex &c->lock irq_context: 0 cb_lock nlk_cb_mutex-GENERIC rtnl_mutex &n->list_lock irq_context: 0 cb_lock nlk_cb_mutex-GENERIC rtnl_mutex &n->list_lock &c->lock irq_context: 0 cb_lock nlk_cb_mutex-GENERIC rtnl_mutex &wdev->mtx irq_context: 0 cb_lock nlk_cb_mutex-GENERIC rtnl_mutex &fq->lock irq_context: 0 pernet_ops_rwsem rtnl_mutex &root->kernfs_rwsem &obj_hash[i].lock irq_context: 0 pernet_ops_rwsem rtnl_mutex &root->kernfs_rwsem key irq_context: 0 pernet_ops_rwsem rtnl_mutex &root->kernfs_rwsem pcpu_lock irq_context: 0 pernet_ops_rwsem rtnl_mutex &root->kernfs_rwsem percpu_counters_lock irq_context: 0 pernet_ops_rwsem rtnl_mutex &root->kernfs_rwsem pcpu_lock stock_lock irq_context: 0 pernet_ops_rwsem rtnl_mutex &root->kernfs_rwsem pool_lock#2 irq_context: 0 sk_lock-AF_TIPC &srv->idr_lock &c->lock irq_context: 0 sk_lock-AF_TIPC &srv->idr_lock &n->list_lock irq_context: 0 sk_lock-AF_TIPC &srv->idr_lock &n->list_lock &c->lock irq_context: 0 (wq_completion)tipc_send (work_completion)(&con->swork) &n->list_lock irq_context: 0 (wq_completion)tipc_send (work_completion)(&con->swork) &n->list_lock &c->lock irq_context: 0 sk_lock-AF_TIPC &obj_hash[i].lock pool_lock irq_context: 0 (wq_completion)tipc_send (work_completion)(&con->swork) slock-AF_TIPC &____s->seqcount#2 irq_context: 0 (wq_completion)tipc_send (work_completion)(&con->swork) slock-AF_TIPC &____s->seqcount irq_context: 0 (wq_completion)wg-crypt-wg0#6 (work_completion)(&peer->transmit_packet_work) &peer->endpoint_lock rcu_read_lock_bh &base->lock irq_context: 0 (wq_completion)wg-crypt-wg0#6 (work_completion)(&peer->transmit_packet_work) &peer->endpoint_lock rcu_read_lock_bh &base->lock &obj_hash[i].lock irq_context: 0 sk_lock-AF_INET6 rcu_read_lock rcu_read_lock &rq->__lock &cfs_rq->removed.lock irq_context: 0 &sb->s_type->i_mutex_key#10 sk_lock-AF_INET6/1 rcu_read_lock &rcu_state.gp_wq irq_context: 0 &sb->s_type->i_mutex_key#10 sk_lock-AF_INET6/1 rcu_read_lock &rcu_state.gp_wq &p->pi_lock irq_context: 0 &sb->s_type->i_mutex_key#10 sk_lock-AF_INET6/1 rcu_read_lock &rcu_state.gp_wq &p->pi_lock &rq->__lock irq_context: 0 &sb->s_type->i_mutex_key#10 sk_lock-AF_INET6/1 rcu_read_lock &rcu_state.gp_wq &p->pi_lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 &mm->mmap_lock sb_pagefaults jbd2_handle rcu_read_lock &rq->__lock irq_context: 0 &mm->mmap_lock sb_pagefaults jbd2_handle rcu_read_lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)events_unbound (work_completion)(&(&kfence_timer)->work) rcu_node_0 irq_context: 0 (wq_completion)events_unbound (work_completion)(&(&kfence_timer)->work) &rcu_state.expedited_wq irq_context: 0 &sb->s_type->i_mutex_key#8 &sbi->s_writepages_rwsem jbd2_handle &rcu_state.expedited_wq irq_context: 0 (wq_completion)events_unbound (work_completion)(&(&kfence_timer)->work) &rcu_state.expedited_wq &p->pi_lock irq_context: 0 (wq_completion)events_unbound (work_completion)(&(&kfence_timer)->work) &rcu_state.expedited_wq &p->pi_lock &rq->__lock irq_context: 0 (wq_completion)events_unbound (work_completion)(&(&kfence_timer)->work) &rcu_state.expedited_wq &p->pi_lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 &sb->s_type->i_mutex_key#8 kauditd_wait.lock &p->pi_lock &cfs_rq->removed.lock irq_context: 0 &xa->xa_lock#9 batched_entropy_u8.lock irq_context: 0 &xa->xa_lock#9 kfence_freelist_lock irq_context: 0 pernet_ops_rwsem rtnl_mutex &root->kernfs_rwsem rcu_read_lock &rq->__lock &cfs_rq->removed.lock irq_context: 0 cb_lock genl_mutex batched_entropy_u8.lock irq_context: 0 cb_lock genl_mutex kfence_freelist_lock irq_context: 0 cb_lock genl_mutex &meta->lock irq_context: softirq &(&hctx->run_work)->timer rcu_read_lock &pool->lock &p->pi_lock &rq->__lock &cfs_rq->removed.lock irq_context: 0 (wq_completion)ext4-rsv-conversion (work_completion)(&ei->i_rsv_conversion_work) jbd2_handle &journal->j_wait_updates &p->pi_lock &cfs_rq->removed.lock irq_context: 0 &sb->s_type->i_mutex_key#8 &sbi->s_writepages_rwsem jbd2_handle &folio_wait_table[i] &p->pi_lock &rq->__lock &cfs_rq->removed.lock irq_context: 0 (wq_completion)ext4-rsv-conversion (work_completion)(&ei->i_rsv_conversion_work) &n->list_lock irq_context: 0 (wq_completion)ext4-rsv-conversion (work_completion)(&ei->i_rsv_conversion_work) &n->list_lock &c->lock irq_context: 0 (wq_completion)wg-kex-wg1#3 (work_completion)(&peer->transmit_handshake_work) &____s->seqcount#2 irq_context: 0 (wq_completion)wg-kex-wg1#3 (work_completion)(&peer->transmit_handshake_work) &____s->seqcount irq_context: 0 (wq_completion)wg-kex-wg0#4 (work_completion)(&({ do { const void *__vpp_verify = (typeof((worker) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __asm__ ("" : "=r"(__ptr) : "0"((typeof(*((worker))) *)((worker)))); (typeof((typeof(*((worker))) *)((worker)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); })->work) &handshake->lock remove_cache_srcu irq_context: 0 (wq_completion)wg-kex-wg0#4 (work_completion)(&({ do { const void *__vpp_verify = (typeof((worker) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __asm__ ("" : "=r"(__ptr) : "0"((typeof(*((worker))) *)((worker)))); (typeof((typeof(*((worker))) *)((worker)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); })->work) &handshake->lock remove_cache_srcu quarantine_lock irq_context: 0 (wq_completion)wg-kex-wg0#4 (work_completion)(&({ do { const void *__vpp_verify = (typeof((worker) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __asm__ ("" : "=r"(__ptr) : "0"((typeof(*((worker))) *)((worker)))); (typeof((typeof(*((worker))) *)((worker)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); })->work) &handshake->lock remove_cache_srcu &c->lock irq_context: 0 (wq_completion)wg-kex-wg0#4 (work_completion)(&({ do { const void *__vpp_verify = (typeof((worker) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __asm__ ("" : "=r"(__ptr) : "0"((typeof(*((worker))) *)((worker)))); (typeof((typeof(*((worker))) *)((worker)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); })->work) &handshake->lock remove_cache_srcu &n->list_lock irq_context: 0 (wq_completion)wg-kex-wg0#4 (work_completion)(&({ do { const void *__vpp_verify = (typeof((worker) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __asm__ ("" : "=r"(__ptr) : "0"((typeof(*((worker))) *)((worker)))); (typeof((typeof(*((worker))) *)((worker)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); })->work) &handshake->lock remove_cache_srcu &obj_hash[i].lock irq_context: 0 (wq_completion)wg-kex-wg0#4 (work_completion)(&({ do { const void *__vpp_verify = (typeof((worker) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __asm__ ("" : "=r"(__ptr) : "0"((typeof(*((worker))) *)((worker)))); (typeof((typeof(*((worker))) *)((worker)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); })->work) &handshake->lock remove_cache_srcu &pcp->lock &zone->lock irq_context: 0 (wq_completion)wg-kex-wg0#4 (work_completion)(&({ do { const void *__vpp_verify = (typeof((worker) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __asm__ ("" : "=r"(__ptr) : "0"((typeof(*((worker))) *)((worker)))); (typeof((typeof(*((worker))) *)((worker)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); })->work) &handshake->lock remove_cache_srcu &pcp->lock &zone->lock &____s->seqcount irq_context: 0 pernet_ops_rwsem uevent_sock_mutex fs_reclaim mmu_notifier_invalidate_range_start &rq->__lock irq_context: 0 &sb->s_type->i_mutex_key#10 sk_lock-AF_TIPC slock-AF_TIPC &c->lock irq_context: 0 &mm->mmap_lock rcu_read_lock rcu_read_lock ptlock_ptr(page)#2 lock#5 irq_context: 0 &mm->mmap_lock rcu_read_lock rcu_read_lock ptlock_ptr(page)#2 lock#5 &lruvec->lru_lock irq_context: 0 &f->f_pos_lock sb_writers#4 &sb->s_type->i_mutex_key#8 &mm->mmap_lock rcu_read_lock rcu_read_lock ptlock_ptr(page)#2 lock#5 irq_context: 0 cb_lock genl_mutex remove_cache_srcu pool_lock#2 irq_context: 0 cb_lock genl_mutex remove_cache_srcu rcu_read_lock rcu_node_0 irq_context: 0 cb_lock genl_mutex remove_cache_srcu rcu_read_lock &rq->__lock irq_context: 0 cb_lock genl_mutex remove_cache_srcu rcu_read_lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 sb_writers#4 &sb->s_type->i_mutex_key#8 mapping.invalidate_lock jbd2_handle &ei->i_data_sem mmu_notifier_invalidate_range_start &cfs_rq->removed.lock irq_context: 0 sb_writers#4 &sb->s_type->i_mutex_key#8 mapping.invalidate_lock jbd2_handle &ei->i_data_sem mmu_notifier_invalidate_range_start &obj_hash[i].lock irq_context: 0 rtnl_mutex ematch_mod_lock irq_context: 0 (wq_completion)kblockd &rq->__lock irq_context: 0 (wq_completion)kblockd &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 &sb->s_type->i_mutex_key#8 &sbi->s_writepages_rwsem jbd2_handle batched_entropy_u8.lock irq_context: 0 &sb->s_type->i_mutex_key#8 &sbi->s_writepages_rwsem jbd2_handle kfence_freelist_lock irq_context: 0 &sb->s_type->i_mutex_key#8 remove_cache_srcu pool_lock#2 irq_context: 0 &sb->s_type->i_mutex_key#8 remove_cache_srcu &rq->__lock irq_context: 0 &sb->s_type->i_mutex_key#8 remove_cache_srcu &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 cb_lock remove_cache_srcu &cfs_rq->removed.lock irq_context: 0 (wq_completion)bond5 (work_completion)(&(&bond->alb_work)->work) &rq->__lock irq_context: 0 &mm->mmap_lock sb_pagefaults jbd2_handle rcu_read_lock rcu_node_0 irq_context: 0 sb_writers#4 &sb->s_type->i_mutex_key#8 mapping.invalidate_lock &mapping->i_mmap_rwsem ptlock_ptr(page)#2 lock#5 irq_context: 0 sb_writers#4 &sb->s_type->i_mutex_key#8 mapping.invalidate_lock &mapping->i_mmap_rwsem lock#5 &lruvec->lru_lock irq_context: 0 sb_writers#4 &sb->s_type->i_mutex_key#8 mapping.invalidate_lock &mapping->i_mmap_rwsem ptlock_ptr(page)#2 lock#5 &lruvec->lru_lock irq_context: 0 &f->f_pos_lock sb_writers#4 &sb->s_type->i_mutex_key#8 &mm->mmap_lock rcu_read_lock rcu_read_lock ptlock_ptr(page)#2 lock#5 &lruvec->lru_lock irq_context: 0 sk_lock-AF_VSOCK irq_context: 0 sk_lock-AF_VSOCK slock-AF_VSOCK irq_context: 0 sk_lock-AF_VSOCK fs_reclaim irq_context: 0 sk_lock-AF_VSOCK fs_reclaim mmu_notifier_invalidate_range_start irq_context: 0 sk_lock-AF_VSOCK &c->lock irq_context: 0 sk_lock-AF_VSOCK &n->list_lock irq_context: 0 sk_lock-AF_VSOCK &n->list_lock &c->lock irq_context: 0 sk_lock-AF_VSOCK pool_lock#2 irq_context: 0 sk_lock-AF_VSOCK vsock_table_lock irq_context: 0 sk_lock-AF_VSOCK vsock_table_lock batched_entropy_u32.lock irq_context: 0 sk_lock-AF_VSOCK &vvs->rx_lock irq_context: 0 sk_lock-AF_VSOCK rcu_read_lock &obj_hash[i].lock irq_context: 0 sk_lock-AF_VSOCK rcu_read_lock pool_lock#2 irq_context: 0 slock-AF_VSOCK irq_context: 0 pernet_ops_rwsem rtnl_mutex &root->kernfs_rwsem &p->pi_lock &rq->__lock irq_context: 0 pernet_ops_rwsem rtnl_mutex &root->kernfs_rwsem &p->pi_lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 pernet_ops_rwsem nf_ct_proto_mutex nf_hook_mutex &n->list_lock irq_context: 0 pernet_ops_rwsem nf_ct_proto_mutex nf_hook_mutex &n->list_lock &c->lock irq_context: 0 pernet_ops_rwsem k-sk_lock-AF_RXRPC &n->list_lock irq_context: 0 pernet_ops_rwsem k-sk_lock-AF_RXRPC &n->list_lock &c->lock irq_context: 0 sb_writers#4 &iint->mutex mapping.invalidate_lock irq_context: 0 sb_writers#4 &iint->mutex mapping.invalidate_lock mmu_notifier_invalidate_range_start irq_context: 0 sb_writers#4 &iint->mutex mapping.invalidate_lock &____s->seqcount irq_context: 0 sb_writers#4 &iint->mutex mapping.invalidate_lock stock_lock irq_context: 0 sb_writers#4 &iint->mutex mapping.invalidate_lock &xa->xa_lock#9 irq_context: 0 sb_writers#4 &iint->mutex mapping.invalidate_lock lock#4 irq_context: 0 sb_writers#4 &iint->mutex mapping.invalidate_lock &xa->xa_lock#9 stock_lock irq_context: 0 sb_writers#4 &iint->mutex mapping.invalidate_lock &ei->i_es_lock irq_context: 0 sb_writers#4 &iint->mutex mapping.invalidate_lock &c->lock irq_context: 0 sb_writers#4 &iint->mutex mapping.invalidate_lock tk_core.seq.seqcount irq_context: 0 sb_writers#4 &iint->mutex mapping.invalidate_lock &rq->__lock irq_context: 0 sb_writers#4 &iint->mutex mapping.invalidate_lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 sb_writers#4 &iint->mutex mapping.invalidate_lock &dd->lock irq_context: 0 sb_writers#4 &iint->mutex mapping.invalidate_lock rcu_read_lock &dd->lock irq_context: 0 sb_writers#4 &iint->mutex mapping.invalidate_lock rcu_read_lock &obj_hash[i].lock irq_context: 0 sb_writers#4 &iint->mutex mapping.invalidate_lock rcu_read_lock tk_core.seq.seqcount irq_context: 0 sb_writers#4 &iint->mutex mapping.invalidate_lock rcu_read_lock &virtscsi_vq->vq_lock irq_context: 0 sb_writers#4 &iint->mutex &folio_wait_table[i] irq_context: 0 sb_writers#4 &iint->mutex tk_core.seq.seqcount irq_context: 0 sb_writers#4 &iint->mutex sb_writers#4 mount_lock irq_context: 0 sb_writers#4 &iint->mutex sb_writers#4 tk_core.seq.seqcount irq_context: 0 sb_writers#4 &iint->mutex sb_writers#4 mmu_notifier_invalidate_range_start irq_context: 0 sb_writers#4 &iint->mutex sb_writers#4 &journal->j_state_lock irq_context: 0 sb_writers#4 &iint->mutex sb_writers#4 jbd2_handle irq_context: 0 sb_writers#4 &iint->mutex sb_writers#4 jbd2_handle &ei->i_raw_lock irq_context: 0 sb_writers#4 &iint->mutex sb_writers#4 jbd2_handle &journal->j_wait_updates irq_context: 0 sb_writers#4 &iint->mutex sb_writers#4 &obj_hash[i].lock irq_context: 0 sb_writers#4 &iint->mutex mapping.invalidate_lock &pcp->lock &zone->lock irq_context: 0 sb_writers#4 &iint->mutex mapping.invalidate_lock &pcp->lock &zone->lock &____s->seqcount irq_context: 0 sb_writers#4 &sb->s_type->i_mutex_key#8 mapping.invalidate_lock lock#5 &lruvec->lru_lock irq_context: 0 sk_lock-AF_VSOCK &mm->mmap_lock irq_context: 0 (wq_completion)events free_ipc_work rcu_read_lock rcu_node_0 irq_context: 0 (wq_completion)events free_ipc_work rcu_read_lock &rq->__lock irq_context: 0 (wq_completion)events free_ipc_work rcu_read_lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 sb_writers#4 &type->i_mutex_dir_key#3/1 sb_internal jbd2_handle pgd_lock irq_context: 0 sb_writers#4 &type->i_mutex_dir_key#3/1 sb_internal jbd2_handle stock_lock irq_context: softirq &ei->i_completed_io_lock rcu_read_lock &pool->lock/1 fill_pool_map-wait-type-override &n->list_lock irq_context: softirq &ei->i_completed_io_lock rcu_read_lock &pool->lock/1 fill_pool_map-wait-type-override &n->list_lock &c->lock irq_context: 0 sb_writers#4 &type->i_mutex_dir_key#3/1 sb_internal jbd2_handle key irq_context: 0 sb_writers#4 &type->i_mutex_dir_key#3/1 sb_internal jbd2_handle pcpu_lock irq_context: 0 sb_writers#4 &type->i_mutex_dir_key#3/1 sb_internal jbd2_handle percpu_counters_lock irq_context: 0 sb_writers#4 &type->i_mutex_dir_key#3/1 sb_internal jbd2_handle pcpu_lock stock_lock irq_context: 0 sb_writers#4 &iint->mutex mapping.invalidate_lock &xa->xa_lock#9 &c->lock irq_context: 0 sb_writers#4 &iint->mutex mapping.invalidate_lock &xa->xa_lock#9 &n->list_lock irq_context: 0 sb_writers#4 &iint->mutex mapping.invalidate_lock &xa->xa_lock#9 &n->list_lock &c->lock irq_context: 0 sb_writers#4 &iint->mutex mapping.invalidate_lock lock#4 &lruvec->lru_lock irq_context: 0 sb_writers#4 &iint->mutex mapping.invalidate_lock rcu_read_lock &obj_hash[i].lock pool_lock irq_context: 0 sb_writers#4 &iint->mutex mapping.invalidate_lock rcu_read_lock rcu_node_0 irq_context: 0 sb_writers#4 &iint->mutex mapping.invalidate_lock rcu_read_lock &rq->__lock irq_context: 0 sb_writers#4 &iint->mutex mapping.invalidate_lock rcu_read_lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 sb_writers#4 &iint->mutex mapping.invalidate_lock rcu_read_lock &rcu_state.gp_wq irq_context: 0 sb_writers#4 &iint->mutex mapping.invalidate_lock rcu_read_lock &rcu_state.gp_wq &p->pi_lock irq_context: 0 sb_writers#4 &iint->mutex mapping.invalidate_lock rcu_read_lock &rcu_state.gp_wq &p->pi_lock &rq->__lock irq_context: 0 sb_writers#4 &iint->mutex mapping.invalidate_lock rcu_read_lock &rcu_state.gp_wq &p->pi_lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 sb_writers#4 &iint->mutex mapping.invalidate_lock rcu_read_lock &c->lock irq_context: 0 sb_writers#4 &iint->mutex mapping.invalidate_lock rcu_read_lock &n->list_lock irq_context: 0 sb_writers#4 &iint->mutex mapping.invalidate_lock rcu_read_lock &n->list_lock &c->lock irq_context: 0 sb_writers#4 &iint->mutex mapping.invalidate_lock rcu_read_lock &rcu_state.expedited_wq irq_context: 0 sb_writers#4 &iint->mutex mapping.invalidate_lock rcu_read_lock &rcu_state.expedited_wq &p->pi_lock irq_context: 0 sb_writers#4 &iint->mutex mapping.invalidate_lock rcu_read_lock &rcu_state.expedited_wq &p->pi_lock &rq->__lock irq_context: 0 sb_writers#4 &iint->mutex mapping.invalidate_lock rcu_read_lock &rcu_state.expedited_wq &p->pi_lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 sb_writers#4 &iint->mutex mapping.invalidate_lock &sem->wait_lock irq_context: 0 sb_writers#4 &iint->mutex &lruvec->lru_lock irq_context: 0 sb_writers#4 &iint->mutex rcu_read_lock pool_lock#2 irq_context: 0 sb_writers#4 &iint->mutex &p->alloc_lock irq_context: 0 sb_writers#4 &iint->mutex &list->lock irq_context: 0 sb_writers#4 &iint->mutex kauditd_wait.lock irq_context: 0 sb_writers#4 &iint->mutex kauditd_wait.lock &p->pi_lock irq_context: 0 sb_writers#4 &iint->mutex kauditd_wait.lock &p->pi_lock &rq->__lock irq_context: 0 sb_writers#4 &iint->mutex kauditd_wait.lock &p->pi_lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 &sb->s_type->i_mutex_key#10 sk_lock-AF_INET6 slock-AF_INET6 &tcp_hashinfo.bhash[i].lock &tcp_hashinfo.bhash2[i].lock &____s->seqcount irq_context: 0 &sb->s_type->i_mutex_key#10 sk_lock-AF_INET6 slock-AF_INET6 &tcp_hashinfo.bhash[i].lock &tcp_hashinfo.bhash2[i].lock rcu_read_lock pool_lock#2 irq_context: softirq (&ndev->rs_timer) rcu_read_lock rcu_read_lock rcu_read_lock_bh dev->qdisc_tx_busylock ?: &qdisc_tx_busylock fill_pool_map-wait-type-override pool_lock#2 irq_context: softirq (&ndev->rs_timer) rcu_read_lock rcu_read_lock rcu_read_lock_bh dev->qdisc_tx_busylock ?: &qdisc_tx_busylock fill_pool_map-wait-type-override pool_lock irq_context: 0 sb_writers#4 &type->i_mutex_dir_key#3/1 &type->i_mutex_dir_key#3 fill_pool_map-wait-type-override rcu_read_lock rcu_node_0 irq_context: 0 sb_writers#4 &type->i_mutex_dir_key#3/1 &type->i_mutex_dir_key#3 fill_pool_map-wait-type-override rcu_read_lock &rq->__lock irq_context: 0 sb_writers#4 &type->i_mutex_dir_key#3/1 &type->i_mutex_dir_key#3 fill_pool_map-wait-type-override rcu_read_lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)ext4-rsv-conversion (work_completion)(&ei->i_rsv_conversion_work) &folio_wait_table[i] &p->pi_lock &rq->__lock &obj_hash[i].lock irq_context: 0 (wq_completion)ext4-rsv-conversion (work_completion)(&ei->i_rsv_conversion_work) &folio_wait_table[i] &p->pi_lock &rq->__lock &base->lock irq_context: 0 (wq_completion)ext4-rsv-conversion (work_completion)(&ei->i_rsv_conversion_work) &folio_wait_table[i] &p->pi_lock &rq->__lock &base->lock &obj_hash[i].lock irq_context: 0 (wq_completion)events_power_efficient (gc_work).work pool_lock#2 irq_context: 0 tracepoints_mutex cpu_hotplug_lock jump_label_mutex.wait_lock irq_context: 0 tracepoints_mutex cpu_hotplug_lock &p->pi_lock irq_context: 0 tracepoints_mutex cpu_hotplug_lock &p->pi_lock &rq->__lock irq_context: 0 tracepoints_mutex cpu_hotplug_lock &p->pi_lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 text_mutex &cfs_rq->removed.lock irq_context: 0 text_mutex &obj_hash[i].lock irq_context: 0 text_mutex pool_lock#2 irq_context: 0 tracepoints_mutex &____s->seqcount#2 irq_context: 0 &mm->mmap_lock ptlock_ptr(page)#2 rcu_read_lock per_cpu_ptr(&cgroup_rstat_cpu_lock, cpu) irq_context: 0 &mm->mmap_lock &xa->xa_lock#9 &pcp->lock &zone->lock &____s->seqcount irq_context: 0 sb_writers#4 &iint->mutex mapping.invalidate_lock rcu_node_0 irq_context: 0 sb_writers#4 &iint->mutex &sem->wait_lock irq_context: 0 sb_writers#4 &iint->mutex &p->pi_lock irq_context: 0 sb_writers#4 &iint->mutex &p->pi_lock &rq->__lock irq_context: 0 sb_writers#4 &iint->mutex &p->pi_lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 sb_writers#4 &iint->mutex batched_entropy_u8.lock irq_context: 0 sb_writers#4 &iint->mutex kfence_freelist_lock irq_context: 0 sb_writers#4 &iint->mutex ima_extend_list_mutex &____s->seqcount#2 irq_context: 0 sb_writers#4 &iint->mutex ima_extend_list_mutex &n->list_lock irq_context: 0 sb_writers#4 &iint->mutex ima_extend_list_mutex &n->list_lock &c->lock irq_context: 0 sb_writers#4 &iint->mutex mapping.invalidate_lock mmu_notifier_invalidate_range_start &rq->__lock irq_context: 0 sb_writers#4 &iint->mutex mapping.invalidate_lock mmu_notifier_invalidate_range_start &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: softirq rcu_read_lock &local->rx_path_lock rcu_read_lock &pool->lock fill_pool_map-wait-type-override pool_lock#2 irq_context: softirq rcu_read_lock &local->rx_path_lock rcu_read_lock &pool->lock fill_pool_map-wait-type-override pool_lock irq_context: 0 data_sockets.lock irq_context: 0 sk_lock-AF_ISDN irq_context: 0 sk_lock-AF_ISDN slock-AF_ISDN irq_context: 0 sk_lock-AF_ISDN clock-AF_ISDN irq_context: 0 sk_lock-AF_ISDN rlock-AF_ISDN irq_context: 0 slock-AF_ISDN irq_context: 0 pernet_ops_rwsem rtnl_mutex rcu_read_lock &pool->lock/1 fill_pool_map-wait-type-override pool_lock#2 irq_context: 0 pernet_ops_rwsem rtnl_mutex rcu_read_lock &pool->lock/1 fill_pool_map-wait-type-override pool_lock irq_context: 0 pernet_ops_rwsem rtnl_mutex fill_pool_map-wait-type-override rcu_read_lock &rcu_state.gp_wq irq_context: 0 slock-AF_LLC &sk->sk_lock.wq &p->pi_lock &cfs_rq->removed.lock irq_context: 0 pernet_ops_rwsem rtnl_mutex &xa->xa_lock#3 &____s->seqcount#2 irq_context: 0 pernet_ops_rwsem rtnl_mutex &xa->xa_lock#3 &____s->seqcount irq_context: 0 &mm->mmap_lock &xa->xa_lock#9 rcu_read_lock per_cpu_ptr(&cgroup_rstat_cpu_lock, cpu) irq_context: 0 slock-AF_LLC &sk->sk_lock.wq &p->pi_lock &rq->__lock &cfs_rq->removed.lock irq_context: 0 sk_lock-AF_LLC rcu_read_lock &rq->__lock &cfs_rq->removed.lock irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem &root->kernfs_rwsem kernfs_idr_lock fill_pool_map-wait-type-override pool_lock#2 irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem &root->kernfs_rwsem kernfs_idr_lock fill_pool_map-wait-type-override &c->lock irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem &root->kernfs_rwsem kernfs_idr_lock fill_pool_map-wait-type-override &____s->seqcount#2 irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem &root->kernfs_rwsem kernfs_idr_lock fill_pool_map-wait-type-override &____s->seqcount irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem &root->kernfs_rwsem kernfs_idr_lock fill_pool_map-wait-type-override pool_lock irq_context: 0 sk_lock-AF_LLC rcu_node_0 irq_context: 0 sk_lock-AF_LLC rcu_read_lock &rcu_state.expedited_wq irq_context: 0 sk_lock-AF_LLC rcu_read_lock &rcu_state.expedited_wq &p->pi_lock irq_context: 0 sk_lock-AF_LLC rcu_read_lock &rcu_state.expedited_wq &p->pi_lock &rq->__lock irq_context: 0 sk_lock-AF_LLC rcu_read_lock &rcu_state.expedited_wq &p->pi_lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)bat_events (work_completion)(&(&forw_packet_aggr->delayed_work)->work) &hard_iface->bat_iv.ogm_buff_mutex fill_pool_map-wait-type-override &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 pernet_ops_rwsem rtnl_mutex fill_pool_map-wait-type-override batched_entropy_u8.lock irq_context: 0 pernet_ops_rwsem rtnl_mutex fill_pool_map-wait-type-override kfence_freelist_lock irq_context: 0 pernet_ops_rwsem fill_pool_map-wait-type-override &____s->seqcount#2 irq_context: 0 pernet_ops_rwsem devices_rwsem rdma_nets_rwsem &device->compat_devs_mutex rcu_read_lock &rq->__lock &cfs_rq->removed.lock irq_context: 0 pernet_ops_rwsem rtnl_mutex &nft_net->commit_mutex &rq->__lock irq_context: 0 pernet_ops_rwsem rtnl_mutex &nft_net->commit_mutex &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)inet_frag_wq (work_completion)(&fqdir->destroy_work) &rq->__lock irq_context: 0 (wq_completion)inet_frag_wq (work_completion)(&fqdir->destroy_work) &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)wg-crypt-wg0#6 (work_completion)(&peer->transmit_packet_work) &peer->endpoint_lock rcu_read_lock_bh &____s->seqcount#2 irq_context: 0 (wq_completion)wg-crypt-wg0#6 (work_completion)(&peer->transmit_packet_work) &peer->endpoint_lock rcu_read_lock_bh &____s->seqcount irq_context: 0 (wq_completion)netns net_cleanup_work rcu_state.barrier_mutex rcu_state.barrier_lock fill_pool_map-wait-type-override pool_lock#2 irq_context: 0 (wq_completion)netns net_cleanup_work rcu_state.barrier_mutex rcu_state.barrier_lock fill_pool_map-wait-type-override pool_lock irq_context: 0 sb_writers#4 &sb->s_type->i_mutex_key#8 ima_extend_list_mutex &cfs_rq->removed.lock irq_context: 0 (wq_completion)wg-kex-wg0 (work_completion)(&peer->transmit_handshake_work) &n->list_lock irq_context: 0 (wq_completion)wg-kex-wg0 (work_completion)(&peer->transmit_handshake_work) &n->list_lock &c->lock irq_context: 0 sb_writers#4 &sb->s_type->i_mutex_key#8 ima_extend_list_mutex &obj_hash[i].lock irq_context: 0 pernet_ops_rwsem rcu_state.barrier_mutex pgd_lock irq_context: 0 pernet_ops_rwsem rcu_state.barrier_mutex stock_lock irq_context: 0 pernet_ops_rwsem rcu_state.barrier_mutex rcu_read_lock pool_lock#2 irq_context: 0 pernet_ops_rwsem rcu_state.barrier_mutex key irq_context: 0 pernet_ops_rwsem rcu_state.barrier_mutex pcpu_lock irq_context: 0 pernet_ops_rwsem rcu_state.barrier_mutex percpu_counters_lock irq_context: 0 pernet_ops_rwsem rcu_state.barrier_mutex pcpu_lock stock_lock irq_context: 0 &sb->s_type->i_mutex_key#8 &sbi->s_writepages_rwsem jbd2_handle &journal->j_state_lock irq_context: 0 sb_writers#4 &type->i_mutex_dir_key#3/1 &sb->s_type->i_mutex_key#8 remove_cache_srcu rcu_read_lock &rcu_state.gp_wq irq_context: 0 sb_writers#4 &type->i_mutex_dir_key#3/1 &sb->s_type->i_mutex_key#8 remove_cache_srcu rcu_read_lock &rcu_state.gp_wq &p->pi_lock irq_context: 0 sb_writers#4 &type->i_mutex_dir_key#3/1 &sb->s_type->i_mutex_key#8 remove_cache_srcu rcu_read_lock &rcu_state.gp_wq &p->pi_lock &rq->__lock irq_context: 0 sb_writers#4 &type->i_mutex_dir_key#3/1 &sb->s_type->i_mutex_key#8 remove_cache_srcu rcu_read_lock &rcu_state.gp_wq &p->pi_lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 sb_writers#4 &type->i_mutex_dir_key#3/1 jbd2_handle &ei->i_data_sem &obj_hash[i].lock pool_lock irq_context: 0 misc_mtx rcu_read_lock rcu_read_lock rcu_node_0 irq_context: 0 misc_mtx rcu_read_lock rcu_read_lock &rq->__lock irq_context: 0 misc_mtx rcu_read_lock rcu_read_lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 &pipe->mutex/1 &mm->mmap_lock fs_reclaim &rq->__lock irq_context: 0 &pipe->mutex/1 &mm->mmap_lock &pcp->lock &zone->lock irq_context: 0 &pipe->mutex/1 &mm->mmap_lock ptlock_ptr(page)#2 lock#4 &lruvec->lru_lock irq_context: 0 misc_mtx rcu_read_lock rcu_node_0 irq_context: 0 &pipe->mutex/1 &mm->mmap_lock ptlock_ptr(page)#2 key irq_context: 0 rtnl_mutex &macsec_netdev_addr_lock_key/1 _xmit_ETHER &n->list_lock irq_context: 0 rtnl_mutex &macsec_netdev_addr_lock_key/1 _xmit_ETHER &n->list_lock &c->lock irq_context: 0 &sb->s_type->i_mutex_key#10 sk_lock-AF_INET slock-AF_INET &tcp_hashinfo.bhash[i].lock &tcp_hashinfo.bhash2[i].lock quarantine_lock irq_context: 0 rtnl_mutex _xmit_ETHER &obj_hash[i].lock pool_lock irq_context: 0 &pipe->mutex/1 &mm->mmap_lock &sem->wait_lock irq_context: 0 &mm->mmap_lock &anon_vma->rwsem rcu_read_lock &rcu_state.expedited_wq irq_context: 0 &mm->mmap_lock &anon_vma->rwsem rcu_read_lock &rcu_state.expedited_wq &p->pi_lock irq_context: 0 &mm->mmap_lock &anon_vma->rwsem rcu_read_lock &rcu_state.expedited_wq &p->pi_lock &rq->__lock irq_context: 0 &mm->mmap_lock &anon_vma->rwsem rcu_read_lock &rcu_state.expedited_wq &p->pi_lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 cb_lock genl_mutex &rcu_state.expedited_wq irq_context: 0 &pipe->mutex/1 &mm->mmap_lock ptlock_ptr(page) irq_context: 0 sb_writers#4 &type->i_mutex_dir_key#3/1 mmu_notifier_invalidate_range_start &rq->__lock &cfs_rq->removed.lock irq_context: 0 &pipe->mutex/1 fs_reclaim mmu_notifier_invalidate_range_start &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 &pipe->mutex/1 fs_reclaim &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 &xs->mutex &mm->mmap_lock ptlock_ptr(page)#2 irq_context: 0 &xs->mutex &mm->mmap_lock fs_reclaim irq_context: 0 &xs->mutex &mm->mmap_lock fs_reclaim mmu_notifier_invalidate_range_start irq_context: 0 &xs->mutex &mm->mmap_lock &____s->seqcount irq_context: 0 &xs->mutex &mm->mmap_lock pool_lock#2 irq_context: 0 &xs->mutex &mm->mmap_lock stock_lock irq_context: 0 &xs->mutex &mm->mmap_lock ptlock_ptr(page)#2 lock#4 irq_context: 0 &xs->mutex &mm->mmap_lock ptlock_ptr(page)#2 lock#4 &lruvec->lru_lock irq_context: 0 &xs->mutex &mm->mmap_lock ptlock_ptr(page)#2 key irq_context: 0 &xs->mutex &mm->mmap_lock &rq->__lock irq_context: 0 &xs->mutex &mm->mmap_lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 &xs->mutex &mm->mmap_lock per_cpu_ptr(&cgroup_rstat_cpu_lock, cpu) irq_context: 0 &xs->mutex &mm->mmap_lock rcu_read_lock rcu_node_0 irq_context: 0 &xs->mutex &mm->mmap_lock rcu_read_lock &rq->__lock irq_context: 0 &xs->mutex &mm->mmap_lock rcu_read_lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 &xs->mutex &mm->mmap_lock rcu_read_lock rcu_read_lock rcu_node_0 irq_context: 0 &xs->mutex &mm->mmap_lock rcu_read_lock rcu_read_lock &rq->__lock irq_context: 0 &xs->mutex &mm->mmap_lock rcu_read_lock rcu_read_lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 &xs->mutex &mm->mmap_lock ptlock_ptr(page) irq_context: 0 &xs->mutex &mm->mmap_lock &pcp->lock &zone->lock irq_context: 0 &xs->mutex &mm->mmap_lock rcu_node_0 irq_context: 0 &xs->mutex &mm->mmap_lock &c->lock irq_context: 0 &xs->mutex &mm->mmap_lock &pcp->lock &zone->lock &____s->seqcount irq_context: 0 rtnl_mutex sk_lock-AF_UNSPEC &n->list_lock irq_context: 0 rtnl_mutex sk_lock-AF_UNSPEC &n->list_lock &c->lock irq_context: 0 br_ioctl_mutex rtnl_mutex batched_entropy_u32.lock crngs.lock irq_context: 0 br_ioctl_mutex rtnl_mutex &idev->mc_lock &c->lock irq_context: 0 br_ioctl_mutex rtnl_mutex &idev->mc_lock &n->list_lock irq_context: 0 br_ioctl_mutex rtnl_mutex &idev->mc_lock &n->list_lock &c->lock irq_context: 0 br_ioctl_mutex rtnl_mutex &idev->mc_lock &obj_hash[i].lock pool_lock irq_context: 0 br_ioctl_mutex rtnl_mutex &idev->mc_lock &rq->__lock irq_context: 0 br_ioctl_mutex rtnl_mutex &idev->mc_lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 &smc->clcsock_release_lock rtnl_mutex ipvs->sync_mutex &n->list_lock irq_context: 0 &smc->clcsock_release_lock rtnl_mutex ipvs->sync_mutex &n->list_lock &c->lock irq_context: softirq (&in_dev->mr_ifc_timer) rcu_read_lock &im->lock &obj_hash[i].lock irq_context: 0 &xs->mutex &mm->mmap_lock fs_reclaim &rq->__lock irq_context: 0 &xs->mutex &mm->mmap_lock fs_reclaim &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 sk_lock-AF_INET6 lock sctp_assocs_id_lock &n->list_lock irq_context: 0 sk_lock-AF_INET6 lock sctp_assocs_id_lock &n->list_lock &c->lock irq_context: 0 &xs->mutex &mm->mmap_lock rcu_read_lock &rcu_state.gp_wq irq_context: 0 &xs->mutex &mm->mmap_lock rcu_read_lock &rcu_state.gp_wq &p->pi_lock irq_context: 0 &xs->mutex &mm->mmap_lock rcu_read_lock &rcu_state.gp_wq &p->pi_lock &rq->__lock irq_context: 0 &xs->mutex &mm->mmap_lock rcu_read_lock &rcu_state.gp_wq &p->pi_lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 &xs->mutex &mm->mmap_lock fs_reclaim mmu_notifier_invalidate_range_start &rq->__lock irq_context: 0 &xs->mutex &mm->mmap_lock fs_reclaim mmu_notifier_invalidate_range_start &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 &xs->mutex &sem->wait_lock irq_context: 0 &xs->mutex &p->pi_lock irq_context: 0 &xs->mutex &p->pi_lock &cfs_rq->removed.lock irq_context: 0 &xs->mutex &p->pi_lock &rq->__lock irq_context: 0 &xs->mutex &p->pi_lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)bat_events (work_completion)(&(&bat_priv->mcast.work)->work) rcu_read_lock &cfs_rq->removed.lock irq_context: 0 (wq_completion)bat_events (work_completion)(&(&bat_priv->mcast.work)->work) rcu_read_lock &obj_hash[i].lock irq_context: 0 (wq_completion)events (work_completion)(&umem->work) irq_context: 0 (wq_completion)events (work_completion)(&umem->work) umem_ida.xa_lock irq_context: 0 (wq_completion)events (work_completion)(&umem->work) vmap_area_lock irq_context: 0 (wq_completion)events (work_completion)(&umem->work) &obj_hash[i].lock irq_context: 0 (wq_completion)events (work_completion)(&umem->work) purge_vmap_area_lock irq_context: 0 (wq_completion)events (work_completion)(&umem->work) purge_vmap_area_lock &obj_hash[i].lock irq_context: 0 (wq_completion)events (work_completion)(&umem->work) purge_vmap_area_lock pool_lock#2 irq_context: 0 (wq_completion)events (work_completion)(&umem->work) pool_lock#2 irq_context: 0 (wq_completion)events (work_completion)(&umem->work) &rq->__lock irq_context: 0 (wq_completion)events (work_completion)(&umem->work) rcu_read_lock pool_lock#2 irq_context: 0 &xs->mutex &mm->mmap_lock rcu_read_lock &rcu_state.expedited_wq irq_context: 0 &xs->mutex &mm->mmap_lock rcu_read_lock &rcu_state.expedited_wq &p->pi_lock irq_context: 0 &xs->mutex &mm->mmap_lock rcu_read_lock &rcu_state.expedited_wq &p->pi_lock &rq->__lock irq_context: 0 &xs->mutex &mm->mmap_lock rcu_read_lock &rcu_state.expedited_wq &p->pi_lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 rtnl_mutex &net->xdp.lock &xs->mutex irq_context: 0 rtnl_mutex &net->xdp.lock &xs->mutex &lock->wait_lock irq_context: 0 rtnl_mutex &net->xdp.lock &xs->mutex &rq->__lock irq_context: 0 rtnl_mutex &net->xdp.lock &xs->mutex &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)ext4-rsv-conversion (work_completion)(&ei->i_rsv_conversion_work) &rq->__lock &cfs_rq->removed.lock irq_context: 0 (wq_completion)wg-kex-wg2 (work_completion)(&peer->transmit_handshake_work) &n->list_lock irq_context: 0 (wq_completion)wg-kex-wg2 (work_completion)(&peer->transmit_handshake_work) &n->list_lock &c->lock irq_context: 0 rtnl_mutex &net->xdp.lock &lock->wait_lock irq_context: 0 rtnl_mutex &net->xdp.lock &p->pi_lock irq_context: 0 rtnl_mutex &net->xdp.lock &p->pi_lock &rq->__lock irq_context: 0 rtnl_mutex &net->xdp.lock &p->pi_lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 rtnl_mutex &dev->mutex &rq->__lock irq_context: 0 rtnl_mutex &dev->mutex &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 rtnl_mutex dev_pm_qos_sysfs_mtx &root->kernfs_rwsem &rq->__lock irq_context: 0 rtnl_mutex dev_pm_qos_sysfs_mtx &root->kernfs_rwsem &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 &smc->clcsock_release_lock ipvs->sync_mutex rtnl_mutex.wait_lock irq_context: softirq (&in_dev->mr_ifc_timer) rcu_read_lock &____s->seqcount#2 irq_context: 0 &sb->s_type->i_mutex_key#8 &sbi->s_writepages_rwsem &meta->lock irq_context: 0 &smc->clcsock_release_lock rtnl_mutex &c->lock irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem fill_pool_map-wait-type-override &rq->__lock irq_context: 0 (wq_completion)events (work_completion)(&umem->work) &lruvec->lru_lock irq_context: 0 (wq_completion)events (work_completion)(&umem->work) &lruvec->lru_lock per_cpu_ptr(&cgroup_rstat_cpu_lock, cpu) irq_context: 0 (wq_completion)events (work_completion)(&umem->work) rcu_read_lock rcu_node_0 irq_context: 0 (wq_completion)events (work_completion)(&umem->work) rcu_read_lock &rq->__lock irq_context: 0 (wq_completion)events (work_completion)(&umem->work) rcu_read_lock &rcu_state.gp_wq irq_context: 0 (wq_completion)events (work_completion)(&umem->work) rcu_read_lock &rcu_state.gp_wq &p->pi_lock irq_context: 0 (wq_completion)events (work_completion)(&umem->work) rcu_read_lock &rcu_state.gp_wq &p->pi_lock &rq->__lock irq_context: 0 (wq_completion)events (work_completion)(&umem->work) rcu_read_lock &rcu_state.gp_wq &p->pi_lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)events (work_completion)(&umem->work) rcu_read_lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)events (work_completion)(&umem->work) &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 rtnl_mutex bus_type_sem &rq->__lock irq_context: 0 rtnl_mutex bus_type_sem &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)events (work_completion)(&umem->work) rcu_node_0 irq_context: 0 &sb->s_type->i_mutex_key#8 &sbi->s_writepages_rwsem jbd2_handle &ei->i_data_sem &ei->i_prealloc_lock irq_context: 0 (wq_completion)events (work_completion)(&umem->work) rcu_read_lock &rcu_state.expedited_wq irq_context: 0 (wq_completion)events (work_completion)(&umem->work) rcu_read_lock &rcu_state.expedited_wq &p->pi_lock irq_context: 0 (wq_completion)events (work_completion)(&umem->work) rcu_read_lock &rcu_state.expedited_wq &p->pi_lock &rq->__lock irq_context: 0 (wq_completion)events (work_completion)(&umem->work) rcu_read_lock &rcu_state.expedited_wq &p->pi_lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)events (work_completion)(&umem->work) &pcp->lock &zone->lock irq_context: 0 (wq_completion)events (work_completion)(&umem->work) &pcp->lock &zone->lock &____s->seqcount irq_context: 0 (wq_completion)events (work_completion)(&umem->work) &cfs_rq->removed.lock irq_context: 0 &smc->clcsock_release_lock rtnl_mutex ipvs->sync_mutex k-sk_lock-AF_INET &im->lock &obj_hash[i].lock irq_context: 0 &smc->clcsock_release_lock rtnl_mutex ipvs->sync_mutex k-sk_lock-AF_INET &im->lock pool_lock#2 irq_context: 0 sk_lock-AF_INET6 k-sk_lock-AF_INET6 &____s->seqcount#8 irq_context: 0 sk_lock-AF_INET6 k-sk_lock-AF_INET6 batched_entropy_u32.lock irq_context: 0 sk_lock-AF_INET6 k-sk_lock-AF_INET6 &rq->__lock irq_context: 0 sk_lock-AF_INET6 k-sk_lock-AF_INET6 &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 sk_lock-AF_INET6 k-sk_lock-AF_INET6 &tcp_hashinfo.bhash[i].lock irq_context: 0 sk_lock-AF_INET6 k-sk_lock-AF_INET6 &tcp_hashinfo.bhash[i].lock &tcp_hashinfo.bhash2[i].lock irq_context: 0 sk_lock-AF_INET6 k-sk_lock-AF_INET6 &tcp_hashinfo.bhash[i].lock &tcp_hashinfo.bhash2[i].lock stock_lock irq_context: 0 sk_lock-AF_INET6 k-sk_lock-AF_INET6 &tcp_hashinfo.bhash[i].lock &tcp_hashinfo.bhash2[i].lock pool_lock#2 irq_context: 0 sk_lock-AF_INET6 k-sk_lock-AF_INET6 &tcp_hashinfo.bhash[i].lock &tcp_hashinfo.bhash2[i].lock k-clock-AF_INET6 irq_context: 0 sk_lock-AF_INET6 k-sk_lock-AF_INET6 &h->lhash2[i].lock irq_context: 0 &sb->s_type->i_mutex_key#10 sk_lock-AF_INET6 k-sk_lock-AF_INET6/1 &h->lhash2[i].lock irq_context: 0 &sb->s_type->i_mutex_key#10 sk_lock-AF_INET6 k-sk_lock-AF_INET6/1 &tcp_hashinfo.bhash[i].lock irq_context: 0 &sb->s_type->i_mutex_key#10 sk_lock-AF_INET6 k-sk_lock-AF_INET6/1 &tcp_hashinfo.bhash[i].lock stock_lock irq_context: 0 &sb->s_type->i_mutex_key#10 sk_lock-AF_INET6 k-sk_lock-AF_INET6/1 &tcp_hashinfo.bhash[i].lock &obj_hash[i].lock irq_context: 0 &sb->s_type->i_mutex_key#10 sk_lock-AF_INET6 k-sk_lock-AF_INET6/1 &tcp_hashinfo.bhash[i].lock pool_lock#2 irq_context: 0 &sb->s_type->i_mutex_key#10 sk_lock-AF_INET6 k-sk_lock-AF_INET6/1 &tcp_hashinfo.bhash[i].lock &tcp_hashinfo.bhash2[i].lock irq_context: 0 &sb->s_type->i_mutex_key#10 sk_lock-AF_INET6 k-sk_lock-AF_INET6/1 &tcp_hashinfo.bhash[i].lock &tcp_hashinfo.bhash2[i].lock stock_lock irq_context: 0 &sb->s_type->i_mutex_key#10 sk_lock-AF_INET6 k-sk_lock-AF_INET6/1 &tcp_hashinfo.bhash[i].lock &tcp_hashinfo.bhash2[i].lock &obj_hash[i].lock irq_context: 0 &sb->s_type->i_mutex_key#10 sk_lock-AF_INET6 k-sk_lock-AF_INET6/1 &tcp_hashinfo.bhash[i].lock &tcp_hashinfo.bhash2[i].lock pool_lock#2 irq_context: 0 &sb->s_type->i_mutex_key#10 sk_lock-AF_INET6 k-sk_lock-AF_INET6/1 &queue->rskq_lock irq_context: 0 sb_writers#4 &type->i_mutex_dir_key#3 rcu_read_lock &rcu_state.gp_wq irq_context: 0 sb_writers#4 &type->i_mutex_dir_key#3 rcu_read_lock &rcu_state.gp_wq &p->pi_lock irq_context: 0 sb_writers#4 &type->i_mutex_dir_key#3 rcu_read_lock &rcu_state.gp_wq &p->pi_lock &rq->__lock irq_context: 0 sb_writers#4 &type->i_mutex_dir_key#3 rcu_read_lock &rcu_state.gp_wq &p->pi_lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 crypto_alg_sem &rq->__lock irq_context: 0 crypto_alg_sem &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 &sb->s_type->i_mutex_key#8 &sbi->s_writepages_rwsem jbd2_handle rcu_read_lock &cfs_rq->removed.lock irq_context: 0 sk_lock-AF_ALG fs_reclaim &rq->__lock irq_context: 0 sk_lock-AF_ALG fs_reclaim &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 sk_lock-AF_ALG rcu_read_lock &rcu_state.expedited_wq &p->pi_lock irq_context: 0 sk_lock-AF_ALG rcu_read_lock &rcu_state.expedited_wq &p->pi_lock &rq->__lock irq_context: 0 sk_lock-AF_ALG rcu_read_lock &rcu_state.expedited_wq &p->pi_lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 sk_lock-AF_ALG &rq->__lock &cfs_rq->removed.lock irq_context: 0 sk_lock-AF_ALG &mm->mmap_lock sb_pagefaults jbd2_handle bit_wait_table + i irq_context: 0 sk_lock-AF_ALG &mm->mmap_lock sb_pagefaults jbd2_handle &rq->__lock &cfs_rq->removed.lock irq_context: 0 &mm->mmap_lock sb_pagefaults mapping.invalidate_lock &rq->__lock &cfs_rq->removed.lock irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem &root->kernfs_rwsem &root->kernfs_iattr_rwsem &cfs_rq->removed.lock irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem &root->kernfs_rwsem &root->kernfs_iattr_rwsem &obj_hash[i].lock irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem &root->kernfs_rwsem &p->pi_lock &rq->__lock irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem &root->kernfs_rwsem &p->pi_lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 sk_lock-AF_INET rcu_read_lock rcu_read_lock &tbl->lock irq_context: 0 sk_lock-AF_INET rcu_read_lock rcu_read_lock &n->lock irq_context: 0 sk_lock-AF_INET rcu_read_lock rcu_read_lock &____s->seqcount#9 irq_context: 0 sk_lock-AF_INET rcu_read_lock rcu_read_lock rcu_read_lock_bh noop_qdisc.q.lock irq_context: softirq (&icsk->icsk_retransmit_timer) slock-AF_INET rcu_read_lock rcu_read_lock rcu_read_lock_bh noop_qdisc.q.lock irq_context: 0 (wq_completion)ext4-rsv-conversion (work_completion)(&ei->i_rsv_conversion_work) jbd2_handle &ei->i_data_sem remove_cache_srcu &pcp->lock &zone->lock irq_context: 0 (wq_completion)ext4-rsv-conversion (work_completion)(&ei->i_rsv_conversion_work) jbd2_handle &ei->i_data_sem remove_cache_srcu &pcp->lock &zone->lock &____s->seqcount irq_context: 0 sb_writers#4 &sb->s_type->i_mutex_key#8 kauditd_wait.lock &p->pi_lock &cfs_rq->removed.lock irq_context: 0 &sb->s_type->i_mutex_key#10 sk_lock-AF_INET &tcp_hashinfo.bhash[i].lock stock_lock irq_context: 0 &sb->s_type->i_mutex_key#10 sk_lock-AF_INET &tcp_hashinfo.bhash[i].lock &obj_hash[i].lock irq_context: 0 &sb->s_type->i_mutex_key#10 sk_lock-AF_INET &tcp_hashinfo.bhash[i].lock pool_lock#2 irq_context: 0 &sb->s_type->i_mutex_key#10 sk_lock-AF_INET &tcp_hashinfo.bhash[i].lock &tcp_hashinfo.bhash2[i].lock stock_lock irq_context: 0 &sb->s_type->i_mutex_key#10 sk_lock-AF_INET &tcp_hashinfo.bhash[i].lock &tcp_hashinfo.bhash2[i].lock &obj_hash[i].lock irq_context: 0 &sb->s_type->i_mutex_key#10 sk_lock-AF_INET &tcp_hashinfo.bhash[i].lock &tcp_hashinfo.bhash2[i].lock pool_lock#2 irq_context: 0 cb_lock rcu_read_lock rcu_read_lock_bh quarantine_lock irq_context: 0 sk_lock-AF_INET6 rcu_read_lock rcu_read_lock &n->lock &obj_hash[i].lock irq_context: 0 sk_lock-AF_INET6 rcu_read_lock rcu_read_lock &n->lock &base->lock irq_context: 0 sk_lock-AF_INET6 rcu_read_lock rcu_read_lock &n->lock &base->lock &obj_hash[i].lock irq_context: 0 sk_lock-AF_INET6 rcu_read_lock rcu_read_lock &n->lock pool_lock#2 irq_context: 0 sk_lock-AF_INET6 rcu_read_lock rcu_read_lock stock_lock irq_context: 0 sk_lock-AF_INET6 rcu_read_lock rcu_read_lock &ul->lock#2 irq_context: 0 sk_lock-AF_INET6 rcu_read_lock rcu_read_lock rcu_read_lock rcu_read_lock &obj_hash[i].lock irq_context: 0 sk_lock-AF_INET6 rcu_read_lock rcu_read_lock rcu_read_lock rcu_read_lock &dir->lock#2 irq_context: 0 sk_lock-AF_INET6 rcu_read_lock rcu_read_lock rcu_read_lock rcu_read_lock &tbl->lock irq_context: 0 sk_lock-AF_INET6 rcu_read_lock rcu_read_lock rcu_read_lock rcu_read_lock &n->lock irq_context: 0 sk_lock-AF_INET6 rcu_read_lock rcu_read_lock rcu_read_lock rcu_read_lock &____s->seqcount#9 irq_context: 0 sk_lock-AF_INET6 rcu_read_lock rcu_read_lock rcu_read_lock rcu_read_lock rcu_read_lock_bh rcu_read_lock pool_lock#2 irq_context: 0 sk_lock-AF_INET6 rcu_read_lock rcu_read_lock rcu_read_lock rcu_read_lock rcu_read_lock_bh rcu_read_lock rcu_read_lock &table->lock#3 irq_context: 0 sk_lock-AF_INET6 rcu_read_lock rcu_read_lock rcu_read_lock rcu_read_lock rcu_read_lock_bh rcu_read_lock rcu_read_lock_bh rcu_read_lock tk_core.seq.seqcount irq_context: 0 sk_lock-AF_INET6 rcu_read_lock rcu_read_lock rcu_read_lock rcu_read_lock rcu_read_lock_bh rcu_read_lock rcu_read_lock_bh rcu_read_lock &list->lock#5 irq_context: 0 sk_lock-AF_INET6 rcu_read_lock rcu_read_lock rcu_read_lock rcu_read_lock rcu_read_lock_bh rcu_read_lock rcu_read_lock_bh &obj_hash[i].lock irq_context: 0 sk_lock-AF_INET6 rcu_read_lock rcu_read_lock rcu_read_lock rcu_read_lock rcu_read_lock_bh rcu_read_lock rcu_read_lock_bh pool_lock#2 irq_context: softirq rcu_read_lock rcu_read_lock &tbl->lock irq_context: softirq rcu_read_lock rcu_read_lock &n->lock &(&n->ha_lock)->lock irq_context: softirq rcu_read_lock rcu_read_lock &n->lock &(&n->ha_lock)->lock &____s->seqcount#9 irq_context: softirq rcu_read_lock rcu_read_lock rcu_read_lock lock#8 irq_context: softirq rcu_read_lock rcu_read_lock rcu_read_lock id_table_lock irq_context: softirq rcu_read_lock rcu_read_lock &n->lock irq_context: softirq rcu_read_lock rcu_read_lock &n->lock &____s->seqcount#9 irq_context: softirq rcu_read_lock rcu_read_lock &ul->lock#2 irq_context: softirq rcu_read_lock rcu_read_lock rcu_read_lock rcu_read_lock &n->lock irq_context: softirq rcu_read_lock rcu_read_lock rcu_read_lock rcu_read_lock &n->lock &obj_hash[i].lock irq_context: softirq rcu_read_lock rcu_read_lock rcu_read_lock rcu_read_lock &n->lock &base->lock irq_context: softirq rcu_read_lock rcu_read_lock rcu_read_lock rcu_read_lock &n->lock &base->lock &obj_hash[i].lock irq_context: softirq rcu_read_lock rcu_read_lock rcu_read_lock rcu_read_lock &____s->seqcount#9 irq_context: softirq rcu_read_lock rcu_read_lock rcu_read_lock rcu_read_lock rcu_read_lock_bh &obj_hash[i].lock irq_context: softirq rcu_read_lock rcu_read_lock rcu_read_lock rcu_read_lock rcu_read_lock_bh pool_lock#2 irq_context: softirq rcu_read_lock rcu_read_lock rcu_read_lock rcu_read_lock rcu_read_lock_bh rcu_read_lock tk_core.seq.seqcount irq_context: softirq rcu_read_lock rcu_read_lock rcu_read_lock rcu_read_lock rcu_read_lock_bh rcu_read_lock &list->lock#5 irq_context: softirq rcu_read_lock rcu_read_lock rcu_read_lock rcu_read_lock &n->lock &(&n->ha_lock)->lock irq_context: softirq rcu_read_lock rcu_read_lock rcu_read_lock rcu_read_lock &n->lock &(&n->ha_lock)->lock &____s->seqcount#9 irq_context: softirq rcu_read_lock rcu_read_lock rcu_read_lock rcu_read_lock rcu_read_lock &____s->seqcount#9 irq_context: softirq rcu_read_lock rcu_read_lock rcu_read_lock rcu_read_lock rcu_read_lock rcu_read_lock_bh rcu_read_lock rcu_read_lock &table->lock#3 irq_context: softirq rcu_read_lock rcu_read_lock rcu_read_lock rcu_read_lock rcu_read_lock rcu_read_lock_bh rcu_read_lock rcu_read_lock_bh rcu_read_lock &list->lock#5 irq_context: softirq rcu_read_lock rcu_read_lock rcu_read_lock rcu_read_lock rcu_read_lock lock#8 irq_context: softirq rcu_read_lock rcu_read_lock rcu_read_lock rcu_read_lock rcu_read_lock id_table_lock irq_context: softirq rcu_read_lock rcu_read_lock rcu_read_lock rcu_read_lock &n->lock irq_context: softirq rcu_read_lock rcu_read_lock rcu_read_lock rcu_read_lock &n->lock &____s->seqcount#9 irq_context: softirq rcu_read_lock rcu_read_lock rcu_read_lock rcu_read_lock nl_table_lock irq_context: softirq rcu_read_lock rcu_read_lock rcu_read_lock rcu_read_lock &obj_hash[i].lock irq_context: softirq rcu_read_lock rcu_read_lock rcu_read_lock rcu_read_lock nl_table_wait.lock irq_context: softirq rcu_read_lock rcu_read_lock rcu_read_lock rcu_read_lock rcu_read_lock rcu_read_lock pool_lock#2 irq_context: softirq rcu_read_lock rcu_read_lock rcu_read_lock rcu_read_lock rcu_read_lock rcu_read_lock &dir->lock#2 irq_context: softirq rcu_read_lock rcu_read_lock rcu_read_lock rcu_read_lock rcu_read_lock rcu_read_lock_bh rcu_read_lock &list->lock#5 irq_context: softirq rcu_read_lock rcu_read_lock rcu_read_lock rcu_read_lock slock-AF_INET6/1 irq_context: 0 sk_lock-AF_INET6 rcu_read_lock tcp_metrics_lock &n->list_lock irq_context: 0 sk_lock-AF_INET6 rcu_read_lock tcp_metrics_lock &n->list_lock &c->lock irq_context: 0 sk_lock-AF_INET6 rcu_read_lock rcu_read_lock rcu_read_lock_bh rcu_read_lock rcu_read_lock &table->lock#3 irq_context: 0 sk_lock-AF_INET6 rcu_read_lock rcu_read_lock rcu_read_lock_bh rcu_read_lock rcu_read_lock_bh rcu_read_lock &list->lock#5 irq_context: softirq rcu_read_lock rcu_read_lock k-slock-AF_INET6 &tcp_hashinfo.bhash[i].lock irq_context: softirq rcu_read_lock rcu_read_lock k-slock-AF_INET6 &tcp_hashinfo.bhash[i].lock &tcp_hashinfo.bhash2[i].lock irq_context: softirq rcu_read_lock rcu_read_lock k-slock-AF_INET6 &tcp_hashinfo.bhash[i].lock &tcp_hashinfo.bhash2[i].lock pool_lock#2 irq_context: softirq rcu_read_lock rcu_read_lock k-slock-AF_INET6 &hashinfo->ehash_locks[i] irq_context: softirq rcu_read_lock rcu_read_lock k-slock-AF_INET6 tk_core.seq.seqcount irq_context: softirq rcu_read_lock rcu_read_lock k-slock-AF_INET6 (&req->rsk_timer) irq_context: softirq rcu_read_lock rcu_read_lock k-slock-AF_INET6 &base->lock irq_context: softirq rcu_read_lock rcu_read_lock k-slock-AF_INET6 &base->lock &obj_hash[i].lock irq_context: softirq rcu_read_lock rcu_read_lock k-slock-AF_INET6 &queue->rskq_lock irq_context: softirq rcu_read_lock rcu_read_lock k-slock-AF_INET6 rcu_read_lock tcp_metrics_lock irq_context: softirq rcu_read_lock rcu_read_lock k-slock-AF_INET6 rcu_read_lock tcp_metrics_lock &c->lock irq_context: softirq rcu_read_lock rcu_read_lock k-slock-AF_INET6 rcu_read_lock tcp_metrics_lock pool_lock#2 irq_context: softirq rcu_read_lock rcu_read_lock k-slock-AF_INET6 k-clock-AF_INET6 irq_context: softirq rcu_read_lock rcu_read_lock k-slock-AF_INET6 k-clock-AF_INET6 rcu_read_lock &pool->lock/1 irq_context: softirq rcu_read_lock rcu_read_lock k-slock-AF_INET6 k-clock-AF_INET6 rcu_read_lock &pool->lock/1 &obj_hash[i].lock irq_context: softirq rcu_read_lock rcu_read_lock k-slock-AF_INET6 k-clock-AF_INET6 rcu_read_lock &pool->lock/1 pool_lock#2 irq_context: softirq rcu_read_lock rcu_read_lock k-slock-AF_INET6 k-clock-AF_INET6 rcu_read_lock &pool->lock/1 &p->pi_lock irq_context: 0 &sb->s_type->i_mutex_key#10 sk_lock-AF_INET6 rcu_read_lock rcu_read_lock rcu_read_lock_bh rcu_read_lock rcu_read_lock &table->lock#3 irq_context: 0 &sb->s_type->i_mutex_key#10 sk_lock-AF_INET6 rcu_read_lock rcu_read_lock rcu_read_lock_bh rcu_read_lock rcu_read_lock_bh rcu_read_lock &list->lock#5 irq_context: softirq rcu_read_lock rcu_read_lock k-slock-AF_INET6/1 irq_context: softirq rcu_read_lock rcu_read_lock k-slock-AF_INET6/1 tk_core.seq.seqcount irq_context: softirq rcu_read_lock rcu_read_lock k-slock-AF_INET6/1 clock-AF_INET6 irq_context: softirq rcu_read_lock rcu_read_lock k-slock-AF_INET6/1 &obj_hash[i].lock irq_context: softirq rcu_read_lock rcu_read_lock k-slock-AF_INET6/1 &base->lock irq_context: softirq rcu_read_lock rcu_read_lock k-slock-AF_INET6/1 &base->lock &obj_hash[i].lock irq_context: 0 &sb->s_type->i_mutex_key#10 sk_lock-AF_INET6 &base->lock irq_context: 0 &sb->s_type->i_mutex_key#10 sk_lock-AF_INET6 &base->lock &obj_hash[i].lock irq_context: 0 (wq_completion)krdsd irq_context: 0 (wq_completion)krdsd (work_completion)(&rtn->rds_tcp_accept_w) irq_context: 0 (wq_completion)krdsd (work_completion)(&rtn->rds_tcp_accept_w) fs_reclaim irq_context: 0 (wq_completion)krdsd (work_completion)(&rtn->rds_tcp_accept_w) fs_reclaim mmu_notifier_invalidate_range_start irq_context: 0 (wq_completion)krdsd (work_completion)(&rtn->rds_tcp_accept_w) pool_lock#2 irq_context: 0 (wq_completion)krdsd (work_completion)(&rtn->rds_tcp_accept_w) mmu_notifier_invalidate_range_start irq_context: 0 (wq_completion)krdsd (work_completion)(&rtn->rds_tcp_accept_w) &sb->s_type->i_lock_key#8 irq_context: 0 (wq_completion)krdsd (work_completion)(&rtn->rds_tcp_accept_w) k-sk_lock-AF_INET6 irq_context: 0 (wq_completion)krdsd (work_completion)(&rtn->rds_tcp_accept_w) k-sk_lock-AF_INET6 k-slock-AF_INET6 irq_context: 0 (wq_completion)krdsd (work_completion)(&rtn->rds_tcp_accept_w) k-sk_lock-AF_INET6 &queue->rskq_lock irq_context: 0 (wq_completion)krdsd (work_completion)(&rtn->rds_tcp_accept_w) k-slock-AF_INET6 irq_context: 0 (wq_completion)krdsd (work_completion)(&rtn->rds_tcp_accept_w) &obj_hash[i].lock irq_context: 0 (wq_completion)krdsd (work_completion)(&rtn->rds_tcp_accept_w) k-sk_lock-AF_INET6 clock-AF_INET6 irq_context: 0 (wq_completion)krdsd (work_completion)(&rtn->rds_tcp_accept_w) k-sk_lock-AF_INET6 &obj_hash[i].lock irq_context: 0 (wq_completion)krdsd (work_completion)(&rtn->rds_tcp_accept_w) k-sk_lock-AF_INET6 &base->lock irq_context: 0 (wq_completion)krdsd (work_completion)(&rtn->rds_tcp_accept_w) k-sk_lock-AF_INET6 &base->lock &obj_hash[i].lock irq_context: 0 (wq_completion)krdsd (work_completion)(&rtn->rds_tcp_accept_w) k-sk_lock-AF_INET6 pool_lock#2 irq_context: 0 (wq_completion)krdsd (work_completion)(&rtn->rds_tcp_accept_w) k-sk_lock-AF_INET6 &dir->lock irq_context: 0 (wq_completion)krdsd (work_completion)(&rtn->rds_tcp_accept_w) k-sk_lock-AF_INET6 fs_reclaim irq_context: 0 (wq_completion)krdsd (work_completion)(&rtn->rds_tcp_accept_w) k-sk_lock-AF_INET6 fs_reclaim mmu_notifier_invalidate_range_start irq_context: 0 (wq_completion)krdsd (work_completion)(&rtn->rds_tcp_accept_w) once_lock irq_context: 0 (wq_completion)krdsd (work_completion)(&rtn->rds_tcp_accept_w) once_lock crngs.lock irq_context: 0 (wq_completion)krdsd (work_completion)(&rtn->rds_tcp_accept_w) rcu_read_lock &pool->lock irq_context: 0 (wq_completion)krdsd (work_completion)(&rtn->rds_tcp_accept_w) rcu_read_lock &pool->lock &obj_hash[i].lock irq_context: 0 (wq_completion)krdsd (work_completion)(&rtn->rds_tcp_accept_w) rcu_read_lock &pool->lock pool_lock#2 irq_context: 0 (wq_completion)krdsd (work_completion)(&rtn->rds_tcp_accept_w) rcu_read_lock &pool->lock &p->pi_lock irq_context: 0 (wq_completion)krdsd (work_completion)(&rtn->rds_tcp_accept_w) &____s->seqcount#2 irq_context: 0 (wq_completion)krdsd (work_completion)(&rtn->rds_tcp_accept_w) &____s->seqcount irq_context: 0 (wq_completion)krdsd (work_completion)(&rtn->rds_tcp_accept_w) &c->lock irq_context: 0 (wq_completion)krdsd (work_completion)(&rtn->rds_tcp_accept_w) rds_cong_lock irq_context: 0 (wq_completion)krdsd (work_completion)(&rtn->rds_tcp_accept_w) &n->list_lock irq_context: 0 (wq_completion)krdsd (work_completion)(&rtn->rds_tcp_accept_w) &n->list_lock &c->lock irq_context: 0 (wq_completion)krdsd (work_completion)(&rtn->rds_tcp_accept_w) rds_trans_sem irq_context: 0 (wq_completion)krdsd (work_completion)(&rtn->rds_tcp_accept_w) rds_trans_sem fs_reclaim irq_context: 0 (wq_completion)krdsd (work_completion)(&rtn->rds_tcp_accept_w) rds_trans_sem fs_reclaim mmu_notifier_invalidate_range_start irq_context: 0 (wq_completion)krdsd (work_completion)(&rtn->rds_tcp_accept_w) rds_trans_sem &c->lock irq_context: 0 (wq_completion)krdsd (work_completion)(&rtn->rds_tcp_accept_w) rds_trans_sem pool_lock#2 irq_context: 0 (wq_completion)krdsd (work_completion)(&rtn->rds_tcp_accept_w) rds_trans_sem crngs.lock irq_context: 0 (wq_completion)krdsd (work_completion)(&rtn->rds_tcp_accept_w) rds_trans_sem &id_priv->handler_mutex irq_context: 0 (wq_completion)krdsd (work_completion)(&rtn->rds_tcp_accept_w) rds_trans_sem &id_priv->handler_mutex &id_priv->lock irq_context: 0 (wq_completion)krdsd (work_completion)(&rtn->rds_tcp_accept_w) rds_trans_sem id_table_lock irq_context: 0 (wq_completion)krdsd (work_completion)(&rtn->rds_tcp_accept_w) rds_trans_sem &x->wait#29 irq_context: 0 (wq_completion)krdsd (work_completion)(&rtn->rds_tcp_accept_w) rds_trans_sem &obj_hash[i].lock irq_context: 0 (wq_completion)krdsd (work_completion)(&rtn->rds_tcp_accept_w) &obj_hash[i].lock pool_lock irq_context: 0 (wq_completion)krdsd (work_completion)(&rtn->rds_tcp_accept_w) rcu_read_lock &____s->seqcount#2 irq_context: 0 (wq_completion)krdsd (work_completion)(&rtn->rds_tcp_accept_w) rcu_read_lock &____s->seqcount irq_context: 0 (wq_completion)krdsd (work_completion)(&rtn->rds_tcp_accept_w) rcu_read_lock pool_lock#2 irq_context: 0 (wq_completion)krdsd (work_completion)(&rtn->rds_tcp_accept_w) rcu_read_lock &c->lock irq_context: 0 (wq_completion)krdsd (work_completion)(&rtn->rds_tcp_accept_w) rcu_read_lock rds_tcp_conn_lock irq_context: 0 (wq_completion)krdsd (work_completion)(&rtn->rds_tcp_accept_w) rcu_read_lock rds_conn_lock irq_context: 0 (wq_completion)krdsd (work_completion)(&rtn->rds_tcp_accept_w) rcu_read_lock rds_conn_lock rds_cong_lock irq_context: 0 (wq_completion)krdsd (work_completion)(&rtn->rds_tcp_accept_w) &tc->t_conn_path_lock irq_context: 0 (wq_completion)krdsd (work_completion)(&rtn->rds_tcp_accept_w) &tc->t_conn_path_lock clock-AF_INET6 irq_context: 0 (wq_completion)krdsd (work_completion)(&rtn->rds_tcp_accept_w) &tc->t_conn_path_lock clock-AF_INET6 rds_tcp_tc_list_lock irq_context: 0 (wq_completion)krdsd (work_completion)(&rtn->rds_tcp_accept_w) &tc->t_conn_path_lock rcu_read_lock rcu_read_lock &pool->lock/1 irq_context: 0 (wq_completion)krdsd (work_completion)(&rtn->rds_tcp_accept_w) &tc->t_conn_path_lock rcu_read_lock rcu_read_lock &pool->lock/1 &obj_hash[i].lock irq_context: 0 (wq_completion)krdsd (work_completion)(&rtn->rds_tcp_accept_w) &tc->t_conn_path_lock rcu_read_lock rcu_read_lock &pool->lock/1 pool_lock#2 irq_context: 0 (wq_completion)krdsd (work_completion)(&rtn->rds_tcp_accept_w) &tc->t_conn_path_lock rcu_read_lock rcu_read_lock &pool->lock/1 &p->pi_lock irq_context: 0 (wq_completion)krdsd (work_completion)(&rtn->rds_tcp_accept_w) &tc->t_conn_path_lock &cp->cp_lock irq_context: 0 (wq_completion)krdsd (work_completion)(&rtn->rds_tcp_accept_w) &tc->t_conn_path_lock pool_lock#2 irq_context: 0 (wq_completion)krdsd (work_completion)(&rtn->rds_tcp_accept_w) &xa->xa_lock#9 irq_context: 0 (wq_completion)krdsd (work_completion)(&rtn->rds_tcp_accept_w) &fsnotify_mark_srcu irq_context: 0 (wq_completion)krdsd (work_completion)(&(&cp->cp_send_w)->work) irq_context: 0 (wq_completion)krdsd (work_completion)(&(&cp->cp_send_w)->work) k-sk_lock-AF_INET6 irq_context: 0 (wq_completion)krdsd (work_completion)(&(&cp->cp_send_w)->work) k-sk_lock-AF_INET6 k-slock-AF_INET6 irq_context: 0 (wq_completion)krdsd (work_completion)(&(&cp->cp_send_w)->work) k-slock-AF_INET6 irq_context: 0 (wq_completion)krdsd (work_completion)(&(&cp->cp_send_w)->work) pool_lock#2 irq_context: 0 (wq_completion)krdsd (work_completion)(&(&cp->cp_send_w)->work) k-sk_lock-AF_INET6 fs_reclaim irq_context: 0 (wq_completion)krdsd (work_completion)(&(&cp->cp_send_w)->work) k-sk_lock-AF_INET6 fs_reclaim mmu_notifier_invalidate_range_start irq_context: 0 (wq_completion)krdsd (work_completion)(&(&cp->cp_send_w)->work) k-sk_lock-AF_INET6 pool_lock#2 irq_context: 0 (wq_completion)krdsd (work_completion)(&(&cp->cp_send_w)->work) k-sk_lock-AF_INET6 &____s->seqcount irq_context: 0 (wq_completion)krdsd (work_completion)(&(&cp->cp_send_w)->work) k-sk_lock-AF_INET6 tk_core.seq.seqcount irq_context: 0 (wq_completion)krdsd (work_completion)(&(&cp->cp_send_w)->work) k-sk_lock-AF_INET6 &obj_hash[i].lock irq_context: 0 (wq_completion)krdsd (work_completion)(&(&cp->cp_send_w)->work) k-sk_lock-AF_INET6 &base->lock irq_context: 0 (wq_completion)krdsd (work_completion)(&(&cp->cp_send_w)->work) k-sk_lock-AF_INET6 &base->lock &obj_hash[i].lock irq_context: 0 (wq_completion)krdsd (work_completion)(&(&cp->cp_send_w)->work) k-sk_lock-AF_INET6 rcu_read_lock rcu_read_lock &____s->seqcount#7 irq_context: 0 (wq_completion)krdsd (work_completion)(&(&cp->cp_send_w)->work) k-sk_lock-AF_INET6 rcu_read_lock rcu_read_lock &ct->lock irq_context: 0 (wq_completion)krdsd (work_completion)(&(&cp->cp_send_w)->work) k-sk_lock-AF_INET6 rcu_read_lock rcu_read_lock &____s->seqcount#9 irq_context: 0 (wq_completion)krdsd (work_completion)(&(&cp->cp_send_w)->work) k-sk_lock-AF_INET6 rcu_read_lock rcu_read_lock rcu_read_lock_bh rcu_read_lock &list->lock#5 irq_context: softirq rcu_read_lock rcu_read_lock rcu_read_lock rcu_read_lock slock-AF_INET6/1 tk_core.seq.seqcount irq_context: softirq rcu_read_lock rcu_read_lock rcu_read_lock rcu_read_lock slock-AF_INET6/1 &hashinfo->ehash_locks[i] irq_context: softirq rcu_read_lock rcu_read_lock rcu_read_lock rcu_read_lock slock-AF_INET6/1 &tcp_hashinfo.bhash[i].lock irq_context: softirq rcu_read_lock rcu_read_lock rcu_read_lock rcu_read_lock slock-AF_INET6/1 &tcp_hashinfo.bhash[i].lock stock_lock irq_context: softirq rcu_read_lock rcu_read_lock rcu_read_lock rcu_read_lock slock-AF_INET6/1 &tcp_hashinfo.bhash[i].lock &obj_hash[i].lock irq_context: softirq rcu_read_lock rcu_read_lock rcu_read_lock rcu_read_lock slock-AF_INET6/1 &tcp_hashinfo.bhash[i].lock pool_lock#2 irq_context: softirq rcu_read_lock rcu_read_lock rcu_read_lock rcu_read_lock slock-AF_INET6/1 &tcp_hashinfo.bhash[i].lock &tcp_hashinfo.bhash2[i].lock irq_context: softirq rcu_read_lock rcu_read_lock rcu_read_lock rcu_read_lock slock-AF_INET6/1 &tcp_hashinfo.bhash[i].lock &tcp_hashinfo.bhash2[i].lock stock_lock irq_context: softirq rcu_read_lock rcu_read_lock rcu_read_lock rcu_read_lock slock-AF_INET6/1 &tcp_hashinfo.bhash[i].lock &tcp_hashinfo.bhash2[i].lock &obj_hash[i].lock irq_context: softirq rcu_read_lock rcu_read_lock rcu_read_lock rcu_read_lock slock-AF_INET6/1 &tcp_hashinfo.bhash[i].lock &tcp_hashinfo.bhash2[i].lock &____s->seqcount irq_context: softirq rcu_read_lock rcu_read_lock rcu_read_lock rcu_read_lock slock-AF_INET6/1 &tcp_hashinfo.bhash[i].lock &tcp_hashinfo.bhash2[i].lock pool_lock#2 irq_context: softirq rcu_read_lock rcu_read_lock rcu_read_lock rcu_read_lock slock-AF_INET6/1 &tcp_hashinfo.bhash[i].lock &tcp_hashinfo.bhash2[i].lock rcu_read_lock pool_lock#2 irq_context: softirq rcu_read_lock rcu_read_lock rcu_read_lock rcu_read_lock slock-AF_INET6/1 &obj_hash[i].lock irq_context: softirq rcu_read_lock rcu_read_lock rcu_read_lock rcu_read_lock slock-AF_INET6/1 &base->lock irq_context: softirq rcu_read_lock rcu_read_lock rcu_read_lock rcu_read_lock slock-AF_INET6/1 &base->lock &obj_hash[i].lock irq_context: softirq rcu_read_lock rcu_read_lock rcu_read_lock rcu_read_lock slock-AF_INET6/1 elock-AF_INET6 irq_context: softirq (&icsk->icsk_delack_timer) k-slock-AF_INET6 irq_context: softirq rcu_read_lock rcu_read_lock rcu_read_lock rcu_read_lock slock-AF_INET6/1 rcu_read_lock pool_lock#2 irq_context: softirq rcu_read_lock rcu_read_lock rcu_read_lock rcu_read_lock slock-AF_INET6/1 rcu_read_lock rcu_read_lock rcu_read_lock rcu_read_lock &c->lock irq_context: softirq rcu_read_lock rcu_read_lock rcu_read_lock rcu_read_lock slock-AF_INET6/1 rcu_read_lock rcu_read_lock rcu_read_lock rcu_read_lock &n->list_lock irq_context: softirq rcu_read_lock rcu_read_lock rcu_read_lock rcu_read_lock slock-AF_INET6/1 rcu_read_lock rcu_read_lock rcu_read_lock rcu_read_lock &n->list_lock &c->lock irq_context: softirq rcu_read_lock rcu_read_lock rcu_read_lock rcu_read_lock slock-AF_INET6/1 rcu_read_lock rcu_read_lock rcu_read_lock rcu_read_lock pool_lock#2 irq_context: softirq rcu_read_lock rcu_read_lock rcu_read_lock rcu_read_lock slock-AF_INET6/1 rcu_read_lock rcu_read_lock rcu_read_lock rcu_read_lock &dir->lock#2 irq_context: softirq rcu_read_lock rcu_read_lock rcu_read_lock rcu_read_lock slock-AF_INET6/1 rcu_read_lock rcu_read_lock &____s->seqcount#7 irq_context: softirq rcu_read_lock rcu_read_lock rcu_read_lock rcu_read_lock slock-AF_INET6/1 rcu_read_lock rcu_read_lock &ct->lock irq_context: softirq rcu_read_lock rcu_read_lock rcu_read_lock rcu_read_lock slock-AF_INET6/1 rcu_read_lock rcu_read_lock &____s->seqcount#10 irq_context: softirq rcu_read_lock rcu_read_lock rcu_read_lock rcu_read_lock slock-AF_INET6/1 rcu_read_lock rcu_read_lock rcu_read_lock_bh rcu_read_lock rcu_read_lock &table->lock#3 irq_context: softirq rcu_read_lock rcu_read_lock rcu_read_lock rcu_read_lock slock-AF_INET6/1 rcu_read_lock rcu_read_lock rcu_read_lock_bh rcu_read_lock rcu_read_lock_bh rcu_read_lock tk_core.seq.seqcount irq_context: softirq rcu_read_lock rcu_read_lock rcu_read_lock rcu_read_lock slock-AF_INET6/1 rcu_read_lock rcu_read_lock rcu_read_lock_bh rcu_read_lock rcu_read_lock_bh rcu_read_lock &list->lock#5 irq_context: softirq rcu_read_lock rcu_read_lock rcu_read_lock rcu_read_lock slock-AF_INET6/1 pool_lock#2 irq_context: softirq rcu_read_lock rcu_read_lock rcu_read_lock rcu_read_lock &dir->lock irq_context: 0 (wq_completion)krdsd (work_completion)(&(&cp->cp_send_w)->work) k-sk_lock-AF_INET6 rcu_read_lock rcu_read_lock rcu_node_0 irq_context: 0 (wq_completion)krdsd (work_completion)(&(&cp->cp_send_w)->work) k-sk_lock-AF_INET6 rcu_read_lock rcu_read_lock &rq->__lock irq_context: 0 (wq_completion)krdsd (work_completion)(&(&cp->cp_send_w)->work) k-sk_lock-AF_INET6 rcu_read_lock rcu_read_lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)krdsd (work_completion)(&(&cp->cp_send_w)->work) k-sk_lock-AF_INET6 rcu_read_lock rcu_node_0 irq_context: 0 (wq_completion)krdsd (work_completion)(&(&cp->cp_send_w)->work) k-sk_lock-AF_INET6 clock-AF_INET6 irq_context: 0 (wq_completion)krdsd (work_completion)(&(&cp->cp_send_w)->work) k-sk_lock-AF_INET6 clock-AF_INET6 &cp->cp_lock irq_context: 0 (wq_completion)krdsd (work_completion)(&(&cp->cp_send_w)->work) k-sk_lock-AF_INET6 clock-AF_INET6 rcu_read_lock rcu_read_lock &pool->lock/1 irq_context: 0 (wq_completion)krdsd (work_completion)(&(&cp->cp_send_w)->work) k-sk_lock-AF_INET6 clock-AF_INET6 rcu_read_lock rcu_read_lock &pool->lock/1 &obj_hash[i].lock irq_context: 0 (wq_completion)krdsd (work_completion)(&(&cp->cp_send_w)->work) k-sk_lock-AF_INET6 clock-AF_INET6 rcu_read_lock rcu_read_lock &pool->lock/1 pool_lock#2 irq_context: 0 (wq_completion)krdsd (work_completion)(&(&cp->cp_send_w)->work) k-sk_lock-AF_INET6 clock-AF_INET6 rcu_read_lock rcu_read_lock &pool->lock/1 &p->pi_lock irq_context: 0 (wq_completion)krdsd (work_completion)(&(&cp->cp_send_w)->work) k-sk_lock-AF_INET6 &hashinfo->ehash_locks[i] irq_context: 0 (wq_completion)krdsd (work_completion)(&(&cp->cp_send_w)->work) k-sk_lock-AF_INET6 &tcp_hashinfo.bhash[i].lock irq_context: 0 (wq_completion)krdsd (work_completion)(&(&cp->cp_send_w)->work) k-sk_lock-AF_INET6 &tcp_hashinfo.bhash[i].lock &tcp_hashinfo.bhash2[i].lock irq_context: 0 (wq_completion)krdsd (work_completion)(&(&cp->cp_send_w)->work) k-sk_lock-AF_INET6 &tcp_hashinfo.bhash[i].lock &tcp_hashinfo.bhash2[i].lock &obj_hash[i].lock irq_context: 0 (wq_completion)krdsd (work_completion)(&(&cp->cp_send_w)->work) k-sk_lock-AF_INET6 &tcp_hashinfo.bhash[i].lock &tcp_hashinfo.bhash2[i].lock pool_lock#2 irq_context: 0 (wq_completion)krdsd (work_completion)(&(&cp->cp_recv_w)->work) irq_context: 0 (wq_completion)krdsd (work_completion)(&cp->cp_down_w) irq_context: 0 (wq_completion)krdsd (work_completion)(&cp->cp_down_w) &cp->cp_cm_lock irq_context: 0 (wq_completion)krdsd (work_completion)(&cp->cp_down_w) k-sk_lock-AF_INET6 irq_context: 0 (wq_completion)krdsd (work_completion)(&cp->cp_down_w) k-sk_lock-AF_INET6 k-slock-AF_INET6 irq_context: 0 (wq_completion)krdsd (work_completion)(&cp->cp_down_w) k-sk_lock-AF_INET6 clock-AF_INET6 irq_context: 0 (wq_completion)krdsd (work_completion)(&cp->cp_down_w) k-sk_lock-AF_INET6 clock-AF_INET6 rds_tcp_tc_list_lock irq_context: 0 (wq_completion)krdsd (work_completion)(&cp->cp_down_w) k-slock-AF_INET6 irq_context: 0 (wq_completion)krdsd (work_completion)(&cp->cp_down_w) k-sk_lock-AF_INET6 &obj_hash[i].lock irq_context: 0 (wq_completion)krdsd (work_completion)(&cp->cp_down_w) k-sk_lock-AF_INET6 pool_lock#2 irq_context: 0 (wq_completion)krdsd (work_completion)(&cp->cp_down_w) k-sk_lock-AF_INET6 k-slock-AF_INET6 &obj_hash[i].lock irq_context: 0 (wq_completion)krdsd (work_completion)(&cp->cp_down_w) k-sk_lock-AF_INET6 k-slock-AF_INET6 elock-AF_INET6 irq_context: 0 (wq_completion)krdsd (work_completion)(&cp->cp_down_w) pool_lock#2 irq_context: 0 (wq_completion)krdsd (work_completion)(&cp->cp_down_w) &dir->lock irq_context: 0 (wq_completion)krdsd (work_completion)(&cp->cp_down_w) &obj_hash[i].lock irq_context: 0 (wq_completion)krdsd (work_completion)(&cp->cp_down_w) &sb->s_type->i_lock_key#8 irq_context: 0 (wq_completion)krdsd (work_completion)(&cp->cp_down_w) &xa->xa_lock#9 irq_context: 0 (wq_completion)krdsd (work_completion)(&cp->cp_down_w) &fsnotify_mark_srcu irq_context: 0 (wq_completion)krdsd (work_completion)(&cp->cp_down_w) &rm->m_flush_wait irq_context: 0 (wq_completion)krdsd (work_completion)(&cp->cp_down_w) &cp->cp_lock irq_context: 0 (wq_completion)krdsd (work_completion)(&cp->cp_down_w) (work_completion)(&(&cp->cp_conn_w)->work) irq_context: 0 (wq_completion)netns net_cleanup_work quarantine_lock irq_context: softirq rcu_read_lock rcu_read_lock rcu_read_lock rcu_read_lock rcu_read_lock rcu_read_lock &c->lock irq_context: 0 (wq_completion)krdsd (work_completion)(&rtn->rds_tcp_accept_w) &tc->t_conn_path_lock rcu_read_lock rcu_read_lock &pool->lock/1 &p->pi_lock &rq->__lock irq_context: 0 (wq_completion)krdsd (work_completion)(&rtn->rds_tcp_accept_w) &tc->t_conn_path_lock rcu_read_lock rcu_read_lock &pool->lock/1 &p->pi_lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)krdsd (work_completion)(&(&cp->cp_send_w)->work) k-sk_lock-AF_INET6 &c->lock irq_context: 0 (wq_completion)krdsd (work_completion)(&(&cp->cp_send_w)->work) &obj_hash[i].lock irq_context: 0 (wq_completion)krdsd (work_completion)(&(&cp->cp_send_w)->work) &cp->cp_lock irq_context: 0 (wq_completion)krdsd (work_completion)(&(&cp->cp_recv_w)->work) k-sk_lock-AF_INET6 irq_context: 0 (wq_completion)krdsd (work_completion)(&(&cp->cp_recv_w)->work) k-sk_lock-AF_INET6 k-slock-AF_INET6 irq_context: 0 (wq_completion)krdsd (work_completion)(&(&cp->cp_recv_w)->work) k-sk_lock-AF_INET6 tk_core.seq.seqcount irq_context: 0 (wq_completion)krdsd (work_completion)(&(&cp->cp_recv_w)->work) k-slock-AF_INET6 irq_context: softirq rcu_read_lock rcu_read_lock k-slock-AF_INET6/1 clock-AF_INET6 &cp->cp_lock irq_context: softirq rcu_read_lock rcu_read_lock k-slock-AF_INET6/1 clock-AF_INET6 rcu_read_lock rcu_read_lock &pool->lock/1 irq_context: softirq rcu_read_lock rcu_read_lock k-slock-AF_INET6/1 clock-AF_INET6 rcu_read_lock rcu_read_lock &pool->lock/1 &obj_hash[i].lock irq_context: softirq rcu_read_lock rcu_read_lock k-slock-AF_INET6/1 clock-AF_INET6 rcu_read_lock rcu_read_lock &pool->lock/1 pool_lock#2 irq_context: softirq rcu_read_lock rcu_read_lock k-slock-AF_INET6/1 clock-AF_INET6 rcu_read_lock rcu_read_lock &pool->lock/1 &p->pi_lock irq_context: softirq rcu_read_lock rcu_read_lock k-slock-AF_INET6/1 clock-AF_INET6 rcu_read_lock rcu_read_lock &pool->lock/1 &p->pi_lock &rq->__lock irq_context: softirq rcu_read_lock rcu_read_lock k-slock-AF_INET6/1 clock-AF_INET6 rcu_read_lock rcu_read_lock &pool->lock/1 &p->pi_lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: softirq rcu_read_lock rcu_read_lock k-slock-AF_INET6/1 pool_lock#2 irq_context: softirq rcu_read_lock rcu_read_lock k-slock-AF_INET6/1 clock-AF_INET6 &rm->m_rs_lock irq_context: softirq rcu_read_lock rcu_read_lock k-slock-AF_INET6/1 clock-AF_INET6 &obj_hash[i].lock irq_context: softirq rcu_read_lock rcu_read_lock k-slock-AF_INET6/1 clock-AF_INET6 pool_lock#2 irq_context: 0 &sb->s_type->i_mutex_key#10 sk_lock-AF_INET6 &n->list_lock irq_context: 0 &sb->s_type->i_mutex_key#10 sk_lock-AF_INET6 &n->list_lock &c->lock irq_context: softirq rcu_read_lock rcu_read_lock k-slock-AF_INET6/1 &hashinfo->ehash_locks[i] irq_context: softirq rcu_read_lock rcu_read_lock k-slock-AF_INET6/1 &tcp_hashinfo.bhash[i].lock irq_context: softirq rcu_read_lock rcu_read_lock k-slock-AF_INET6/1 &tcp_hashinfo.bhash[i].lock &tcp_hashinfo.bhash2[i].lock irq_context: softirq rcu_read_lock rcu_read_lock k-slock-AF_INET6/1 &tcp_hashinfo.bhash[i].lock &tcp_hashinfo.bhash2[i].lock &obj_hash[i].lock irq_context: softirq rcu_read_lock rcu_read_lock k-slock-AF_INET6/1 &tcp_hashinfo.bhash[i].lock &tcp_hashinfo.bhash2[i].lock pool_lock#2 irq_context: 0 rtnl_mutex &dev_addr_list_lock_key#2 irq_context: 0 sock_diag_mutex sock_diag_table_mutex nlk_cb_mutex-SOCK_DIAG &ht->lock irq_context: 0 sock_diag_mutex sock_diag_table_mutex nlk_cb_mutex-SOCK_DIAG rcu_read_lock &ht->lock irq_context: 0 sock_diag_mutex sock_diag_table_mutex nlk_cb_mutex-SOCK_DIAG rcu_read_lock k-clock-AF_NETLINK irq_context: 0 sock_diag_mutex sock_diag_table_mutex nlk_cb_mutex-SOCK_DIAG nl_table_lock irq_context: 0 sock_diag_mutex sock_diag_table_mutex nlk_cb_mutex-SOCK_DIAG rcu_read_lock clock-AF_NETLINK irq_context: 0 sock_diag_mutex sock_diag_table_mutex nlk_cb_mutex-SOCK_DIAG &rq->__lock irq_context: 0 sock_diag_mutex sock_diag_table_mutex nlk_cb_mutex-SOCK_DIAG &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: softirq rcu_read_lock rcu_read_lock k-slock-AF_INET6 k-clock-AF_INET6 rcu_read_lock &pool->lock/1 &p->pi_lock &rq->__lock irq_context: softirq rcu_read_lock rcu_read_lock k-slock-AF_INET6 k-clock-AF_INET6 rcu_read_lock &pool->lock/1 &p->pi_lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: softirq rcu_read_lock rcu_read_lock rcu_read_lock rcu_read_lock slock-AF_INET6/1 &c->lock irq_context: 0 (wq_completion)krdsd (work_completion)(&(&cp->cp_send_w)->work) k-sk_lock-AF_INET6 clock-AF_INET6 rcu_read_lock rcu_read_lock &pool->lock/1 &p->pi_lock &rq->__lock irq_context: 0 (wq_completion)krdsd (work_completion)(&(&cp->cp_send_w)->work) k-sk_lock-AF_INET6 clock-AF_INET6 rcu_read_lock rcu_read_lock &pool->lock/1 &p->pi_lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 sk_lock-AF_ALG &drbg->drbg_mutex &n->list_lock irq_context: 0 sk_lock-AF_ALG &drbg->drbg_mutex &n->list_lock &c->lock irq_context: 0 rtnl_mutex stock_lock rcu_read_lock per_cpu_ptr(&cgroup_rstat_cpu_lock, cpu) irq_context: 0 sk_lock-AF_ALG &drbg->drbg_mutex &cfs_rq->removed.lock irq_context: 0 sk_lock-AF_ALG &drbg->drbg_mutex &obj_hash[i].lock irq_context: 0 &sb->s_type->i_mutex_key#10 &rng->jent_lock &____s->seqcount irq_context: 0 &sb->s_type->i_mutex_key#10 &rng->jent_lock &pcp->lock &zone->lock irq_context: 0 &sb->s_type->i_mutex_key#10 &rng->jent_lock rcu_read_lock pool_lock#2 irq_context: 0 (wq_completion)events_unbound (work_completion)(&map->work) rcu_state.exp_mutex &obj_hash[i].lock pool_lock irq_context: 0 &sb->s_type->i_mutex_key#10 &rng->jent_lock quarantine_lock irq_context: 0 sk_lock-AF_INET6 &list->lock#39 irq_context: 0 &xs->mutex remove_cache_srcu rcu_read_lock &rcu_state.gp_wq irq_context: 0 &xs->mutex remove_cache_srcu rcu_read_lock &rcu_state.gp_wq &p->pi_lock irq_context: 0 &xs->mutex remove_cache_srcu rcu_read_lock &rcu_state.gp_wq &p->pi_lock &rq->__lock irq_context: 0 &xs->mutex remove_cache_srcu rcu_read_lock &rcu_state.gp_wq &p->pi_lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 &xs->mutex remove_cache_srcu rcu_read_lock &cfs_rq->removed.lock irq_context: 0 &xs->mutex remove_cache_srcu rcu_read_lock &obj_hash[i].lock irq_context: 0 sk_lock-AF_INET6 rcu_read_lock rcu_read_lock &nf_conntrack_locks[i] &nf_conntrack_locks[i]/1 batched_entropy_u8.lock crngs.lock irq_context: softirq (&asoc->timers[i]) irq_context: softirq (&asoc->timers[i]) slock-AF_INET6 irq_context: softirq (&asoc->timers[i]) slock-AF_INET6 pool_lock#2 irq_context: softirq (&asoc->timers[i]) slock-AF_INET6 &c->lock irq_context: softirq (&asoc->timers[i]) slock-AF_INET6 &obj_hash[i].lock irq_context: softirq (&asoc->timers[i]) slock-AF_INET6 rcu_read_lock rcu_read_lock &____s->seqcount#7 irq_context: softirq (&asoc->timers[i]) slock-AF_INET6 rcu_read_lock rcu_read_lock &ct->lock irq_context: softirq (&asoc->timers[i]) slock-AF_INET6 rcu_read_lock rcu_read_lock &____s->seqcount#10 irq_context: softirq (&asoc->timers[i]) slock-AF_INET6 rcu_read_lock rcu_read_lock rcu_read_lock_bh tk_core.seq.seqcount irq_context: softirq (&asoc->timers[i]) slock-AF_INET6 rcu_read_lock rcu_read_lock rcu_read_lock_bh &list->lock#5 irq_context: softirq rcu_read_lock rcu_read_lock slock-AF_INET6 &base->lock irq_context: softirq rcu_read_lock rcu_read_lock slock-AF_INET6 &base->lock &obj_hash[i].lock irq_context: softirq rcu_read_lock rcu_read_lock slock-AF_INET6 &c->lock irq_context: softirq rcu_read_lock rcu_read_lock slock-AF_INET6 &zone->lock irq_context: softirq rcu_read_lock rcu_read_lock slock-AF_INET6 rcu_read_lock rcu_read_lock &____s->seqcount#7 irq_context: softirq rcu_read_lock rcu_read_lock slock-AF_INET6 rcu_read_lock rcu_read_lock &ct->lock irq_context: softirq rcu_read_lock rcu_read_lock slock-AF_INET6 rcu_read_lock rcu_read_lock &____s->seqcount#10 irq_context: softirq rcu_read_lock rcu_read_lock slock-AF_INET6 rcu_read_lock rcu_read_lock rcu_read_lock_bh tk_core.seq.seqcount irq_context: softirq rcu_read_lock rcu_read_lock slock-AF_INET6 rcu_read_lock rcu_read_lock rcu_read_lock_bh &list->lock#5 irq_context: softirq rcu_read_lock rcu_read_lock slock-AF_INET6 &list->lock#39 irq_context: softirq rcu_read_lock rcu_read_lock slock-AF_INET6 &zone->lock &____s->seqcount irq_context: softirq rcu_read_lock rcu_read_lock slock-AF_INET6 rcu_read_lock &____s->seqcount irq_context: softirq rcu_read_lock rcu_read_lock slock-AF_INET6 &list->lock#24 irq_context: softirq rcu_read_lock rcu_read_lock slock-AF_INET6 krc.lock irq_context: softirq rcu_read_lock rcu_read_lock slock-AF_INET6 rcu_read_lock rhashtable_bucket irq_context: softirq rcu_read_lock rcu_read_lock krc.lock irq_context: softirq rcu_read_lock rcu_read_lock &sctp_port_hashtable[i].lock irq_context: softirq rcu_read_lock rcu_read_lock &sctp_port_hashtable[i].lock &obj_hash[i].lock irq_context: softirq rcu_read_lock rcu_read_lock &sctp_port_hashtable[i].lock pool_lock#2 irq_context: softirq rcu_read_lock rcu_read_lock sctp_assocs_id_lock irq_context: softirq rcu_read_lock rcu_read_lock sctp_assocs_id_lock &obj_hash[i].lock irq_context: softirq rcu_read_lock rcu_read_lock sctp_assocs_id_lock pool_lock#2 irq_context: 0 dup_mmap_sem &mm->mmap_lock &mm->mmap_lock/1 fs_reclaim mmu_notifier_invalidate_range_start pgd_lock irq_context: 0 dup_mmap_sem &mm->mmap_lock &mm->mmap_lock/1 fs_reclaim mmu_notifier_invalidate_range_start stock_lock irq_context: 0 dup_mmap_sem &mm->mmap_lock &mm->mmap_lock/1 fs_reclaim mmu_notifier_invalidate_range_start key irq_context: 0 dup_mmap_sem &mm->mmap_lock &mm->mmap_lock/1 fs_reclaim mmu_notifier_invalidate_range_start pcpu_lock irq_context: 0 dup_mmap_sem &mm->mmap_lock &mm->mmap_lock/1 fs_reclaim mmu_notifier_invalidate_range_start percpu_counters_lock irq_context: 0 dup_mmap_sem &mm->mmap_lock &mm->mmap_lock/1 fs_reclaim mmu_notifier_invalidate_range_start pcpu_lock stock_lock irq_context: 0 sb_writers#4 &type->i_mutex_dir_key#3/1 sb_internal &rq->__lock &cfs_rq->removed.lock irq_context: softirq rcu_read_lock rcu_read_lock slock-AF_INET6 &n->list_lock irq_context: softirq rcu_read_lock rcu_read_lock slock-AF_INET6 &n->list_lock &c->lock irq_context: 0 &sb->s_type->i_mutex_key#8 &rq->__lock &obj_hash[i].lock irq_context: 0 &sb->s_type->i_mutex_key#8 &rq->__lock &base->lock irq_context: 0 &sb->s_type->i_mutex_key#8 &rq->__lock &base->lock &obj_hash[i].lock irq_context: 0 &sb->s_type->i_mutex_key#8 rcu_read_lock &____s->seqcount irq_context: 0 sk_lock-AF_TIPC &list->lock#40 irq_context: 0 sk_lock-AF_TIPC rcu_read_lock &service->lock &n->list_lock irq_context: 0 sk_lock-AF_TIPC rcu_read_lock &service->lock &n->list_lock &c->lock irq_context: 0 sk_lock-AF_TIPC k-slock-AF_TIPC &list->lock#40 irq_context: 0 sk_lock-AF_TIPC k-slock-AF_TIPC &obj_hash[i].lock irq_context: 0 sk_lock-AF_TIPC quarantine_lock irq_context: 0 sk_lock-AF_TIPC rcu_read_lock &service->lock &____s->seqcount#2 irq_context: 0 sk_lock-AF_TIPC rcu_read_lock &service->lock &____s->seqcount irq_context: 0 sk_lock-AF_TIPC batched_entropy_u8.lock irq_context: 0 sk_lock-AF_TIPC kfence_freelist_lock irq_context: 0 sk_lock-AF_TIPC &meta->lock irq_context: 0 sk_lock-AF_TIPC k-slock-AF_TIPC quarantine_lock irq_context: 0 sk_lock-AF_TIPC rcu_read_lock &rcu_state.gp_wq &p->pi_lock &rq->__lock irq_context: 0 sk_lock-AF_TIPC rcu_read_lock &rcu_state.gp_wq &p->pi_lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 &pn->l2tp_tunnel_idr_lock irq_context: 0 &pn->l2tp_tunnel_idr_lock pool_lock#2 irq_context: 0 sk_lock-AF_PPPOX fs_reclaim irq_context: 0 sk_lock-AF_PPPOX fs_reclaim mmu_notifier_invalidate_range_start irq_context: 0 sk_lock-AF_PPPOX pool_lock#2 irq_context: 0 sk_lock-AF_PPPOX &ps->sk_lock irq_context: 0 sk_lock-AF_PPPOX &ps->sk_lock &rq->__lock irq_context: 0 sk_lock-AF_PPPOX &ps->sk_lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 sk_lock-AF_PPPOX &ps->sk_lock &tunnel->hlist_lock irq_context: 0 &sb->s_type->i_mutex_key#10 rcu_read_lock &pool->lock/1 irq_context: 0 &sb->s_type->i_mutex_key#10 rcu_read_lock &pool->lock/1 &obj_hash[i].lock irq_context: 0 &sb->s_type->i_mutex_key#10 rcu_read_lock &pool->lock/1 pool_lock#2 irq_context: 0 &sb->s_type->i_mutex_key#10 rcu_read_lock &pool->lock/1 &p->pi_lock irq_context: 0 &sb->s_type->i_mutex_key#10 rcu_read_lock &pool->lock/1 &p->pi_lock &cfs_rq->removed.lock irq_context: 0 &sb->s_type->i_mutex_key#10 rcu_read_lock &pool->lock/1 &p->pi_lock &rq->__lock irq_context: 0 &sb->s_type->i_mutex_key#10 rcu_read_lock &pool->lock/1 &p->pi_lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)l2tp irq_context: 0 (wq_completion)l2tp (work_completion)(&tunnel->del_work) irq_context: 0 (wq_completion)l2tp (work_completion)(&tunnel->del_work) &tunnel->hlist_lock irq_context: 0 (wq_completion)l2tp (work_completion)(&tunnel->del_work) rcu_state.exp_mutex rcu_node_0 irq_context: 0 (wq_completion)l2tp (work_completion)(&tunnel->del_work) rcu_state.exp_mutex &obj_hash[i].lock irq_context: 0 (wq_completion)l2tp (work_completion)(&tunnel->del_work) rcu_state.exp_mutex rcu_read_lock &pool->lock irq_context: 0 (wq_completion)l2tp (work_completion)(&tunnel->del_work) rcu_state.exp_mutex rcu_read_lock &pool->lock &obj_hash[i].lock irq_context: 0 (wq_completion)l2tp (work_completion)(&tunnel->del_work) rcu_state.exp_mutex rcu_read_lock &pool->lock &p->pi_lock irq_context: 0 (wq_completion)l2tp (work_completion)(&tunnel->del_work) rcu_state.exp_mutex rcu_read_lock &pool->lock &p->pi_lock &rq->__lock irq_context: 0 (wq_completion)l2tp (work_completion)(&tunnel->del_work) rcu_state.exp_mutex rcu_read_lock &pool->lock &p->pi_lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)l2tp (work_completion)(&tunnel->del_work) rcu_state.exp_mutex &rnp->exp_wq[2] irq_context: 0 (wq_completion)l2tp (work_completion)(&tunnel->del_work) rcu_state.exp_mutex &rq->__lock irq_context: 0 (wq_completion)l2tp (work_completion)(&tunnel->del_work) rcu_state.exp_mutex &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 &sb->s_type->i_mutex_key#10 cpu_hotplug_lock irq_context: 0 &sb->s_type->i_mutex_key#10 sk_lock-AF_PPPOX &ps->sk_lock irq_context: 0 &sb->s_type->i_mutex_key#10 sk_lock-AF_PPPOX pool_lock#2 irq_context: 0 (wq_completion)l2tp (work_completion)(&tunnel->del_work) &obj_hash[i].lock irq_context: 0 (wq_completion)l2tp (work_completion)(&tunnel->del_work) &list->lock#41 irq_context: 0 (wq_completion)l2tp (work_completion)(&tunnel->del_work) &pn->l2tp_tunnel_idr_lock irq_context: 0 (wq_completion)l2tp (work_completion)(&tunnel->del_work) &pn->l2tp_tunnel_idr_lock &obj_hash[i].lock irq_context: 0 (wq_completion)l2tp (work_completion)(&tunnel->del_work) &pn->l2tp_tunnel_idr_lock pool_lock#2 irq_context: 0 sk_lock-AF_TIPC rcu_read_lock rcu_read_lock rcu_node_0 irq_context: 0 sk_lock-AF_TIPC rcu_read_lock rcu_read_lock &rq->__lock irq_context: 0 sk_lock-AF_TIPC rcu_read_lock rcu_read_lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: softirq rcu_callback rlock-AF_PPPOX irq_context: softirq rcu_callback wlock-AF_PPPOX irq_context: softirq rcu_callback clock-AF_INET6 irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem k-sk_lock-AF_RXRPC quarantine_lock irq_context: 0 pernet_ops_rwsem ipvs->est_mutex rcu_read_lock rcu_node_0 irq_context: 0 (wq_completion)l2tp (work_completion)(&tunnel->del_work) &rnp->exp_lock irq_context: 0 (wq_completion)l2tp (work_completion)(&tunnel->del_work) rcu_state.exp_mutex irq_context: 0 (wq_completion)l2tp (work_completion)(&tunnel->del_work) rcu_state.exp_mutex rcu_state.exp_mutex.wait_lock irq_context: 0 (wq_completion)l2tp (work_completion)(&tunnel->del_work) rcu_state.exp_mutex &rnp->exp_wq[1] irq_context: 0 (wq_completion)l2tp (work_completion)(&tunnel->del_work) rcu_state.exp_mutex.wait_lock irq_context: 0 (wq_completion)l2tp (work_completion)(&tunnel->del_work) &p->pi_lock irq_context: 0 (wq_completion)l2tp (work_completion)(&tunnel->del_work) &p->pi_lock &rq->__lock irq_context: 0 (wq_completion)l2tp (work_completion)(&tunnel->del_work) &p->pi_lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 pernet_ops_rwsem ipvs->est_mutex rcu_read_lock &rq->__lock irq_context: 0 &pn->l2tp_tunnel_idr_lock &c->lock irq_context: 0 sk_lock-AF_PPPOX &c->lock irq_context: 0 &sb->s_type->i_mutex_key#10 sk_lock-AF_PPPOX &tunnel->hlist_lock irq_context: 0 &sb->s_type->i_mutex_key#10 sk_lock-AF_PPPOX rcu_state.exp_mutex &rnp->exp_wq[2] irq_context: 0 &sb->s_type->i_mutex_key#10 sk_lock-AF_PPPOX &list->lock#41 irq_context: 0 sk_lock-AF_TIPC rcu_read_lock &rcu_state.expedited_wq irq_context: 0 sk_lock-AF_TIPC rcu_read_lock &rcu_state.expedited_wq &p->pi_lock irq_context: 0 sk_lock-AF_TIPC rcu_read_lock &rcu_state.expedited_wq &p->pi_lock &rq->__lock irq_context: 0 sk_lock-AF_TIPC rcu_read_lock &rcu_state.expedited_wq &p->pi_lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 sb_writers#5 &type->i_mutex_dir_key#5/1 &fsnotify_mark_srcu rcu_node_0 irq_context: 0 sk_lock-AF_TIPC &cfs_rq->removed.lock irq_context: 0 pernet_ops_rwsem devices_rwsem rdma_nets_rwsem &device->compat_devs_mutex fs_reclaim mmu_notifier_invalidate_range_start &rq->__lock &cfs_rq->removed.lock irq_context: 0 sk_lock-AF_TIPC remove_cache_srcu rcu_read_lock rcu_node_0 irq_context: 0 sk_lock-AF_TIPC remove_cache_srcu rcu_read_lock &rq->__lock irq_context: 0 sk_lock-AF_TIPC remove_cache_srcu rcu_read_lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 rcu_read_lock mount_lock mount_lock.seqcount &p->pi_lock irq_context: 0 rcu_read_lock mount_lock mount_lock.seqcount &p->pi_lock &rq->__lock irq_context: 0 rcu_read_lock mount_lock mount_lock.seqcount &p->pi_lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: softirq (&n->timer) rcu_read_lock lock#8 irq_context: softirq (&n->timer) batched_entropy_u8.lock irq_context: softirq (&n->timer) &n->lock &____s->seqcount#9 irq_context: 0 (wq_completion)wg-kex-wg1#9 (work_completion)(&peer->transmit_handshake_work) &wg->static_identity.lock &handshake->lock crngs.lock base_crng.lock irq_context: 0 (wq_completion)wg-kex-wg0#7 (work_completion)(&peer->transmit_handshake_work) &wg->static_identity.lock &handshake->lock crngs.lock base_crng.lock irq_context: 0 (wq_completion)wg-kex-wg1#9 (work_completion)(&peer->transmit_handshake_work) batched_entropy_u8.lock crngs.lock irq_context: 0 (wq_completion)wg-kex-wg0#7 (work_completion)(&peer->transmit_handshake_work) batched_entropy_u8.lock crngs.lock irq_context: 0 (wq_completion)wg-kex-wg2 (work_completion)(&peer->transmit_handshake_work) &____s->seqcount#2 irq_context: 0 (wq_completion)wg-kex-wg2 (work_completion)(&peer->transmit_handshake_work) &____s->seqcount irq_context: 0 (wq_completion)wg-kex-wg0#9 (work_completion)(&peer->transmit_handshake_work) &wg->static_identity.lock &handshake->lock &rq->__lock irq_context: 0 pernet_ops_rwsem rtnl_mutex &idev->mc_lock batched_entropy_u8.lock irq_context: 0 pernet_ops_rwsem rtnl_mutex &idev->mc_lock kfence_freelist_lock irq_context: 0 (wq_completion)events free_ipc_work &rnp->exp_wq[0] irq_context: 0 &ep->mtx gdp_mutex &c->lock irq_context: 0 sb_writers#4 &sb->s_type->i_mutex_key#8 jbd2_handle &ei->i_data_sem &ei->i_es_lock &c->lock irq_context: 0 (wq_completion)events (work_completion)(&pwq->unbound_release_work) rcu_state.exp_mutex &cfs_rq->removed.lock irq_context: 0 (wq_completion)events (work_completion)(&pwq->unbound_release_work) rcu_state.exp_mutex pool_lock#2 irq_context: 0 sb_writers#4 &sb->s_type->i_mutex_key#8 jbd2_handle &ei->i_data_sem &____s->seqcount irq_context: 0 sb_writers#4 &sb->s_type->i_mutex_key#8 jbd2_handle &ei->i_data_sem rcu_read_lock pool_lock#2 irq_context: 0 sb_writers#4 &sb->s_type->i_mutex_key#8 jbd2_handle stock_lock irq_context: 0 sb_writers#4 &sb->s_type->i_mutex_key#8 jbd2_handle &xa->xa_lock#9 irq_context: 0 sb_writers#4 &sb->s_type->i_mutex_key#8 jbd2_handle &xa->xa_lock#9 stock_lock irq_context: 0 sb_writers#4 &sb->s_type->i_mutex_key#8 jbd2_handle &xa->xa_lock#9 pool_lock#2 irq_context: 0 sb_writers#4 &sb->s_type->i_mutex_key#8 jbd2_handle &xa->xa_lock#9 &c->lock irq_context: 0 sb_writers#4 &sb->s_type->i_mutex_key#8 jbd2_handle lock#4 irq_context: 0 &ep->mtx gdp_mutex &n->list_lock irq_context: 0 &ep->mtx gdp_mutex &n->list_lock &c->lock irq_context: 0 (wq_completion)ext4-rsv-conversion (work_completion)(&ei->i_rsv_conversion_work) jbd2_handle &ei->i_data_sem &rcu_state.expedited_wq irq_context: 0 (wq_completion)ext4-rsv-conversion (work_completion)(&ei->i_rsv_conversion_work) jbd2_handle &ei->i_data_sem &rcu_state.expedited_wq &p->pi_lock irq_context: 0 (wq_completion)ext4-rsv-conversion (work_completion)(&ei->i_rsv_conversion_work) jbd2_handle &ei->i_data_sem &rcu_state.expedited_wq &p->pi_lock &rq->__lock irq_context: 0 (wq_completion)ext4-rsv-conversion (work_completion)(&ei->i_rsv_conversion_work) jbd2_handle &ei->i_data_sem &rcu_state.expedited_wq &p->pi_lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 &u->iolock &u->peer_wait &ei->socket.wq.wait &ep->poll_wait/1 irq_context: 0 &u->iolock &u->peer_wait &ei->socket.wq.wait &ep->poll_wait/1 &p->pi_lock irq_context: 0 &u->iolock &u->peer_wait &ei->socket.wq.wait &ep->poll_wait/1 &p->pi_lock &rq->__lock irq_context: 0 &u->iolock &u->peer_wait &ei->socket.wq.wait &ep->poll_wait/1 &p->pi_lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 sb_writers#4 &sb->s_type->i_mutex_key#8 &sbi->s_writepages_rwsem &ei->i_data_sem rcu_read_lock rcu_node_0 irq_context: 0 sb_writers#4 &sb->s_type->i_mutex_key#8 &sbi->s_writepages_rwsem &ei->i_data_sem rcu_read_lock &rq->__lock irq_context: 0 sb_writers#4 &sb->s_type->i_mutex_key#8 &sbi->s_writepages_rwsem &ei->i_data_sem rcu_read_lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 &dentry->d_lock &sb->s_type->i_lock_key#22 &dentry->d_lock &lru->node[i].lock irq_context: 0 (wq_completion)wg-kex-wg0 (work_completion)(&peer->transmit_handshake_work) &wg->static_identity.lock &handshake->lock &rq->__lock irq_context: 0 (wq_completion)wg-kex-wg0 (work_completion)(&peer->transmit_handshake_work) &wg->static_identity.lock &handshake->lock rcu_read_lock_bh batched_entropy_u32.lock crngs.lock irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem rtnl_mutex &root->kernfs_rwsem rcu_read_lock &rq->__lock &cfs_rq->removed.lock irq_context: 0 rcu_read_lock &ei->socket.wq.wait &ep->lock rcu_read_lock &ws->lock irq_context: 0 rcu_read_lock &ei->socket.wq.wait &ep->lock rcu_read_lock &ws->lock tk_core.seq.seqcount irq_context: 0 rcu_read_lock &ei->socket.wq.wait &ep->lock rcu_read_lock &ws->lock &obj_hash[i].lock irq_context: 0 rcu_read_lock &ei->socket.wq.wait &ep->poll_wait/1 irq_context: 0 rcu_read_lock &ei->socket.wq.wait &ep->poll_wait/1 &p->pi_lock irq_context: 0 &ep->mtx remove_cache_srcu pool_lock#2 irq_context: 0 pernet_ops_rwsem nfnl_subsys_ipset &rq->__lock irq_context: 0 pernet_ops_rwsem nfnl_subsys_ipset &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 sb_writers#4 &type->i_mutex_dir_key#3/1 &base->lock irq_context: 0 sb_writers#4 &type->i_mutex_dir_key#3/1 &base->lock &obj_hash[i].lock irq_context: 0 &ep->mtx lock kernfs_idr_lock &c->lock irq_context: 0 &ep->mtx lock kernfs_idr_lock pool_lock#2 irq_context: 0 &ep->mtx uevent_sock_mutex rcu_read_lock &ei->socket.wq.wait &ep->lock &ep->wq &p->pi_lock &cfs_rq->removed.lock irq_context: 0 &ep->mtx uevent_sock_mutex rcu_read_lock &rcu_state.expedited_wq irq_context: 0 &ep->mtx uevent_sock_mutex rcu_read_lock &rcu_state.expedited_wq &p->pi_lock irq_context: 0 &ep->mtx uevent_sock_mutex rcu_read_lock &rcu_state.expedited_wq &p->pi_lock &rq->__lock irq_context: 0 &ep->mtx uevent_sock_mutex rcu_read_lock &rcu_state.expedited_wq &p->pi_lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 sb_writers#4 &sb->s_type->i_mutex_key#8 jbd2_handle &xa->xa_lock#9 &n->list_lock irq_context: 0 sb_writers#4 &sb->s_type->i_mutex_key#8 jbd2_handle &xa->xa_lock#9 &n->list_lock &c->lock irq_context: 0 &ep->mtx &root->kernfs_rwsem &sem->wait_lock irq_context: 0 rtnl_mutex &sch->q.lock &obj_hash[i].lock irq_context: 0 (wq_completion)mld (work_completion)(&(&idev->mc_ifc_work)->work) &idev->mc_lock rcu_read_lock rcu_read_lock rcu_read_lock_bh rcu_read_lock &pcp->lock &zone->lock irq_context: 0 (wq_completion)mld (work_completion)(&(&idev->mc_ifc_work)->work) &idev->mc_lock rcu_read_lock rcu_read_lock rcu_read_lock_bh rcu_read_lock &br->multicast_lock rcu_read_lock pool_lock#2 irq_context: 0 &sb->s_type->i_mutex_key#10 &u->peer_wait &ei->socket.wq.wait irq_context: 0 &sb->s_type->i_mutex_key#10 &u->peer_wait &ei->socket.wq.wait &ep->lock irq_context: 0 &sb->s_type->i_mutex_key#10 &u->peer_wait &ei->socket.wq.wait &ep->lock rcu_read_lock &ws->lock irq_context: 0 &sb->s_type->i_mutex_key#10 &u->peer_wait &ei->socket.wq.wait &ep->lock rcu_read_lock &ws->lock tk_core.seq.seqcount irq_context: 0 &sb->s_type->i_mutex_key#10 &u->peer_wait &ei->socket.wq.wait &ep->lock rcu_read_lock &ws->lock &obj_hash[i].lock irq_context: 0 pernet_ops_rwsem k-sk_lock-AF_RXRPC &rxnet->local_mutex fs_reclaim &rq->__lock irq_context: 0 pernet_ops_rwsem k-sk_lock-AF_RXRPC &rxnet->local_mutex fs_reclaim &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 remove_cache_srcu &obj_hash[i].lock pool_lock irq_context: 0 sb_writers#4 &type->i_mutex_dir_key#3 tomoyo_ss rcu_node_0 irq_context: 0 &f->f_pos_lock sb_writers#4 &obj_hash[i].lock pool_lock irq_context: 0 uevent_sock_mutex &n->list_lock irq_context: 0 uevent_sock_mutex &n->list_lock &c->lock irq_context: 0 uevent_sock_mutex &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)events (work_completion)(&pwq->unbound_release_work) rcu_state.exp_mutex &rq->__lock &cfs_rq->removed.lock irq_context: 0 &f->f_pos_lock sb_writers#4 fill_pool_map-wait-type-override pool_lock#2 irq_context: 0 &f->f_pos_lock sb_writers#4 fill_pool_map-wait-type-override pool_lock irq_context: 0 fill_pool_map-wait-type-override rcu_read_lock &rcu_state.expedited_wq irq_context: 0 fill_pool_map-wait-type-override rcu_read_lock &rcu_state.expedited_wq &p->pi_lock irq_context: 0 fill_pool_map-wait-type-override rcu_read_lock &rcu_state.expedited_wq &p->pi_lock &rq->__lock irq_context: 0 fill_pool_map-wait-type-override rcu_read_lock &rcu_state.expedited_wq &p->pi_lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 pernet_ops_rwsem rtnl_mutex rcu_state.exp_mutex rcu_read_lock &pool->lock fill_pool_map-wait-type-override pool_lock#2 irq_context: 0 pernet_ops_rwsem rtnl_mutex rcu_state.exp_mutex rcu_read_lock &pool->lock fill_pool_map-wait-type-override &c->lock irq_context: 0 pernet_ops_rwsem rtnl_mutex rcu_state.exp_mutex rcu_read_lock &pool->lock fill_pool_map-wait-type-override pool_lock irq_context: 0 rtnl_mutex &sch->q.lock crngs.lock irq_context: 0 sb_writers#4 &sb->s_type->i_mutex_key#8 &sbi->s_writepages_rwsem &c->lock irq_context: 0 &xs->mutex &zone->lock &____s->seqcount irq_context: 0 sb_writers#4 &sb->s_type->i_mutex_key#8 &sbi->s_writepages_rwsem remove_cache_srcu irq_context: 0 sb_writers#4 &sb->s_type->i_mutex_key#8 &sbi->s_writepages_rwsem remove_cache_srcu quarantine_lock irq_context: 0 sb_writers#4 &sb->s_type->i_mutex_key#8 &sbi->s_writepages_rwsem remove_cache_srcu &c->lock irq_context: 0 sb_writers#4 &sb->s_type->i_mutex_key#8 &sbi->s_writepages_rwsem remove_cache_srcu &n->list_lock irq_context: 0 sb_writers#4 &sb->s_type->i_mutex_key#8 &sbi->s_writepages_rwsem remove_cache_srcu rcu_read_lock pool_lock#2 irq_context: 0 sb_writers#4 &sb->s_type->i_mutex_key#8 &sbi->s_writepages_rwsem remove_cache_srcu &obj_hash[i].lock irq_context: 0 sb_writers#4 &sb->s_type->i_mutex_key#8 &sbi->s_writepages_rwsem remove_cache_srcu &rq->__lock irq_context: 0 sb_writers#4 &sb->s_type->i_mutex_key#8 &sbi->s_writepages_rwsem remove_cache_srcu &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: softirq (&app->join_timer) rcu_read_lock_bh &c->lock irq_context: 0 pernet_ops_rwsem rtnl_mutex &caifn->caifdevs.lock &rnp->exp_lock irq_context: 0 pernet_ops_rwsem rtnl_mutex &caifn->caifdevs.lock rcu_state.exp_mutex irq_context: 0 pernet_ops_rwsem rtnl_mutex &caifn->caifdevs.lock rcu_state.exp_mutex rcu_state.exp_mutex.wait_lock irq_context: 0 sb_writers#4 &sb->s_type->i_mutex_key#8 &sbi->s_writepages_rwsem &rq->__lock irq_context: 0 sb_writers#4 &sb->s_type->i_mutex_key#8 &sbi->s_writepages_rwsem &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 sb_writers#4 &sb->s_type->i_mutex_key#8 &sbi->s_writepages_rwsem &n->list_lock irq_context: 0 sb_writers#4 &sb->s_type->i_mutex_key#8 &sbi->s_writepages_rwsem &n->list_lock &c->lock irq_context: 0 sb_writers#4 &sb->s_type->i_mutex_key#8 fill_pool_map-wait-type-override &rq->__lock irq_context: 0 sb_writers#4 &sb->s_type->i_mutex_key#8 fill_pool_map-wait-type-override &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 sb_writers#4 &sb->s_type->i_mutex_key#8 &rcu_state.expedited_wq irq_context: softirq (&sdp->delay_work) rcu_read_lock &pool->lock &p->pi_lock &rq->__lock &cfs_rq->removed.lock irq_context: 0 (wq_completion)wg-kex-wg2#11 (work_completion)(&peer->transmit_handshake_work) &peer->endpoint_lock rcu_read_lock_bh rcu_read_lock rcu_read_lock_bh rcu_read_lock &c->lock irq_context: 0 (wq_completion)wg-kex-wg2#11 (work_completion)(&peer->transmit_handshake_work) &peer->endpoint_lock rcu_read_lock_bh rcu_read_lock rcu_read_lock_bh rcu_read_lock tk_core.seq.seqcount irq_context: 0 (wq_completion)wg-kex-wg2#11 (work_completion)(&peer->transmit_handshake_work) &peer->endpoint_lock rcu_read_lock_bh rcu_read_lock rcu_read_lock_bh rcu_read_lock &obj_hash[i].lock irq_context: 0 (wq_completion)wg-crypt-wg2#4 (work_completion)(&({ do { const void *__vpp_verify = (typeof((worker) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __asm__ ("" : "=r"(__ptr) : "0"((typeof(*((worker))) *)((worker)))); (typeof((typeof(*((worker))) *)((worker)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); })->work) &c->lock irq_context: softirq (&app->join_timer) rcu_read_lock_bh rcu_read_lock &____s->seqcount#2 irq_context: softirq (&app->join_timer) rcu_read_lock_bh rcu_read_lock &____s->seqcount irq_context: softirq (&app->join_timer) rcu_read_lock_bh &n->list_lock irq_context: softirq (&app->join_timer) rcu_read_lock_bh &n->list_lock &c->lock irq_context: 0 (wq_completion)wg-kex-wg0#7 (work_completion)(&peer->transmit_handshake_work) &peer->endpoint_lock rcu_read_lock_bh rcu_read_lock rcu_read_lock_bh rcu_read_lock &____s->seqcount#2 irq_context: 0 (wq_completion)wg-kex-wg0#7 (work_completion)(&peer->transmit_handshake_work) &peer->endpoint_lock rcu_read_lock_bh rcu_read_lock rcu_read_lock_bh rcu_read_lock &____s->seqcount irq_context: 0 &net->xfrm.xfrm_cfg_mutex rcu_read_lock rcu_read_lock_bh rcu_read_lock pool_lock#2 irq_context: 0 &net->xfrm.xfrm_cfg_mutex rcu_read_lock rcu_read_lock_bh rcu_read_lock tk_core.seq.seqcount irq_context: 0 &net->xfrm.xfrm_cfg_mutex rcu_read_lock rcu_read_lock_bh rcu_read_lock &obj_hash[i].lock irq_context: 0 &net->xfrm.xfrm_cfg_mutex rcu_read_lock rcu_read_lock_bh rcu_read_lock &____s->seqcount irq_context: 0 (wq_completion)wg-kex-wg0#9 (work_completion)(&peer->transmit_handshake_work) &peer->endpoint_lock rcu_read_lock_bh rcu_read_lock rcu_read_lock_bh rcu_read_lock tk_core.seq.seqcount irq_context: 0 (wq_completion)wg-kex-wg0#9 (work_completion)(&peer->transmit_handshake_work) &peer->endpoint_lock rcu_read_lock_bh rcu_read_lock rcu_read_lock_bh rcu_read_lock &obj_hash[i].lock irq_context: 0 (wq_completion)wg-kex-wg0 (work_completion)(&peer->transmit_handshake_work) &peer->endpoint_lock rcu_read_lock_bh rcu_read_lock rcu_read_lock_bh rcu_read_lock tk_core.seq.seqcount irq_context: 0 (wq_completion)wg-kex-wg0 (work_completion)(&peer->transmit_handshake_work) &peer->endpoint_lock rcu_read_lock_bh rcu_read_lock rcu_read_lock_bh rcu_read_lock &obj_hash[i].lock irq_context: 0 pernet_ops_rwsem ovs_mutex nf_ct_proto_mutex nf_hook_mutex cpu_hotplug_lock &rq->__lock irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem rtnl_mutex &rcu_state.expedited_wq irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem rtnl_mutex &rcu_state.expedited_wq &p->pi_lock irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem rtnl_mutex &rcu_state.expedited_wq &p->pi_lock &rq->__lock irq_context: 0 sk_lock-AF_BLUETOOTH-BTPROTO_HCI &n->list_lock irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem rtnl_mutex &rcu_state.expedited_wq &p->pi_lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 namespace_sem rcu_node_0 irq_context: 0 sk_lock-AF_BLUETOOTH-BTPROTO_HCI &n->list_lock &c->lock irq_context: 0 namespace_sem &rcu_state.expedited_wq irq_context: 0 namespace_sem &rcu_state.expedited_wq &p->pi_lock irq_context: 0 namespace_sem &rcu_state.expedited_wq &p->pi_lock &rq->__lock irq_context: 0 namespace_sem &rcu_state.expedited_wq &p->pi_lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)mld (work_completion)(&(&idev->mc_ifc_work)->work) &idev->mc_lock rcu_read_lock rcu_read_lock rcu_read_lock_bh rcu_read_lock &br->multicast_lock quarantine_lock irq_context: 0 (wq_completion)mld (work_completion)(&(&idev->mc_ifc_work)->work) &idev->mc_lock rcu_read_lock rcu_read_lock rcu_read_lock_bh rcu_read_lock &br->multicast_lock nl_table_wait.lock &p->pi_lock irq_context: 0 (wq_completion)mld (work_completion)(&(&idev->mc_ifc_work)->work) &idev->mc_lock rcu_read_lock rcu_read_lock rcu_read_lock_bh rcu_read_lock &br->multicast_lock nl_table_wait.lock &p->pi_lock &rq->__lock irq_context: 0 (wq_completion)mld (work_completion)(&(&idev->mc_ifc_work)->work) &idev->mc_lock rcu_read_lock rcu_read_lock rcu_read_lock_bh rcu_read_lock &br->multicast_lock nl_table_wait.lock &p->pi_lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 namespace_sem remove_cache_srcu irq_context: 0 namespace_sem remove_cache_srcu quarantine_lock irq_context: 0 rtnl_mutex rcu_read_lock rcu_read_lock_bh rcu_read_lock &n->list_lock irq_context: 0 rtnl_mutex rcu_read_lock rcu_read_lock_bh rcu_read_lock &n->list_lock &c->lock irq_context: 0 (wq_completion)mld (work_completion)(&(&idev->mc_ifc_work)->work) &idev->mc_lock rcu_read_lock rcu_read_lock rcu_read_lock_bh rcu_read_lock &br->multicast_lock &n->list_lock irq_context: 0 (wq_completion)mld (work_completion)(&(&idev->mc_ifc_work)->work) &idev->mc_lock rcu_read_lock rcu_read_lock rcu_read_lock_bh rcu_read_lock &br->multicast_lock &n->list_lock &c->lock irq_context: 0 sk_lock-AF_BLUETOOTH-BTPROTO_HCI quarantine_lock irq_context: 0 sk_lock-AF_INET k-sk_lock-AF_INET rcu_read_lock rcu_read_lock rcu_read_lock_bh noop_qdisc.q.lock irq_context: 0 sock_diag_mutex sock_diag_table_mutex nlk_cb_mutex-SOCK_DIAG inet_diag_table_mutex irq_context: 0 sock_diag_mutex sock_diag_table_mutex nlk_cb_mutex-SOCK_DIAG inet_diag_table_mutex &h->lhash2[i].lock irq_context: 0 sock_diag_mutex sock_diag_table_mutex nlk_cb_mutex-SOCK_DIAG inet_diag_table_mutex &rq->__lock irq_context: 0 sock_diag_mutex sock_diag_table_mutex nlk_cb_mutex-SOCK_DIAG inet_diag_table_mutex &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)wg-kex-wg0#9 (work_completion)(&peer->transmit_handshake_work) &peer->endpoint_lock rcu_read_lock_bh rcu_read_lock rcu_read_lock_bh rcu_read_lock &c->lock irq_context: 0 sock_diag_mutex sock_diag_table_mutex nlk_cb_mutex-SOCK_DIAG inet_diag_table_mutex &h->lhash2[i].lock k-clock-AF_INET6 irq_context: 0 sock_diag_mutex sock_diag_table_mutex nlk_cb_mutex-SOCK_DIAG inet_diag_table_mutex &hashinfo->ehash_locks[i] irq_context: 0 sock_diag_mutex sock_diag_table_mutex nlk_cb_mutex-SOCK_DIAG inet_diag_table_mutex k-clock-AF_INET irq_context: 0 sock_diag_mutex sock_diag_table_mutex nlk_cb_mutex-SOCK_DIAG inet_diag_table_mutex k-sk_lock-AF_INET irq_context: 0 sock_diag_mutex sock_diag_table_mutex nlk_cb_mutex-SOCK_DIAG inet_diag_table_mutex k-sk_lock-AF_INET k-slock-AF_INET irq_context: 0 sock_diag_mutex sock_diag_table_mutex nlk_cb_mutex-SOCK_DIAG inet_diag_table_mutex k-slock-AF_INET irq_context: 0 (wq_completion)wg-kex-wg0 (work_completion)(&peer->transmit_handshake_work) batched_entropy_u8.lock crngs.lock irq_context: softirq (&app->join_timer) rcu_read_lock_bh rcu_read_lock &n->list_lock irq_context: softirq (&app->join_timer) rcu_read_lock_bh rcu_read_lock &n->list_lock &c->lock irq_context: 0 pernet_ops_rwsem devices_rwsem rdma_nets_rwsem &p->pi_lock &rq->__lock &cfs_rq->removed.lock irq_context: 0 pernet_ops_rwsem devices_rwsem rdma_nets_rwsem &device->compat_devs_mutex &rxe->usdev_lock rtnl_mutex rcu_read_lock &rq->__lock irq_context: 0 pernet_ops_rwsem devices_rwsem rdma_nets_rwsem &device->compat_devs_mutex &rxe->usdev_lock rtnl_mutex rcu_read_lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: softirq (&in_dev->mr_ifc_timer) rcu_read_lock rcu_read_lock_bh rcu_read_lock &obj_hash[i].lock pool_lock irq_context: 0 pernet_ops_rwsem rtnl_mutex &idev->mc_lock fill_pool_map-wait-type-override &cfs_rq->removed.lock irq_context: 0 pernet_ops_rwsem rtnl_mutex &idev->mc_lock fill_pool_map-wait-type-override &obj_hash[i].lock irq_context: 0 &sb->s_type->i_mutex_key#10 sk_lock-AF_INET k-sk_lock-AF_INET/1 &hashinfo->ehash_locks[i] irq_context: 0 &sb->s_type->i_mutex_key#10 sk_lock-AF_INET k-sk_lock-AF_INET/1 k-slock-AF_INET &base->lock irq_context: 0 &sb->s_type->i_mutex_key#10 sk_lock-AF_INET k-sk_lock-AF_INET/1 k-slock-AF_INET &base->lock &obj_hash[i].lock irq_context: 0 &sb->s_type->i_mutex_key#10 sk_lock-AF_INET k-sk_lock-AF_INET/1 k-slock-AF_INET &tcp_hashinfo.bhash[i].lock stock_lock irq_context: 0 &sb->s_type->i_mutex_key#10 sk_lock-AF_INET k-sk_lock-AF_INET/1 k-slock-AF_INET &tcp_hashinfo.bhash[i].lock &obj_hash[i].lock irq_context: 0 &sb->s_type->i_mutex_key#10 sk_lock-AF_INET k-sk_lock-AF_INET/1 k-slock-AF_INET &tcp_hashinfo.bhash[i].lock pool_lock#2 irq_context: 0 &sb->s_type->i_mutex_key#10 sk_lock-AF_INET k-sk_lock-AF_INET/1 k-slock-AF_INET &tcp_hashinfo.bhash[i].lock &tcp_hashinfo.bhash2[i].lock stock_lock irq_context: 0 &sb->s_type->i_mutex_key#10 sk_lock-AF_INET k-sk_lock-AF_INET/1 k-slock-AF_INET &tcp_hashinfo.bhash[i].lock &tcp_hashinfo.bhash2[i].lock &obj_hash[i].lock irq_context: 0 &sb->s_type->i_mutex_key#10 sk_lock-AF_INET k-sk_lock-AF_INET/1 k-slock-AF_INET &tcp_hashinfo.bhash[i].lock &tcp_hashinfo.bhash2[i].lock pool_lock#2 irq_context: 0 pernet_ops_rwsem ovs_mutex nf_ct_proto_mutex nf_hook_mutex &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 &mm->mmap_lock fs_reclaim mmu_notifier_invalidate_range_start &rq->__lock &cfs_rq->removed.lock irq_context: 0 &sb->s_type->i_lock_key &xa->xa_lock#9 fill_pool_map-wait-type-override pool_lock#2 irq_context: 0 (wq_completion)events (work_completion)(&nlk->work) vmap_area_lock irq_context: 0 (wq_completion)events (work_completion)(&nlk->work) purge_vmap_area_lock irq_context: 0 (wq_completion)events (work_completion)(&nlk->work) purge_vmap_area_lock &obj_hash[i].lock irq_context: 0 (wq_completion)events (work_completion)(&nlk->work) purge_vmap_area_lock pool_lock#2 irq_context: 0 (wq_completion)events (work_completion)(&nlk->work) rcu_read_lock pool_lock#2 irq_context: 0 &mm->mmap_lock remove_cache_srcu rcu_read_lock &rcu_state.expedited_wq irq_context: 0 &mm->mmap_lock remove_cache_srcu rcu_read_lock &rcu_state.expedited_wq &p->pi_lock irq_context: 0 &mm->mmap_lock remove_cache_srcu rcu_read_lock &rcu_state.expedited_wq &p->pi_lock &rq->__lock irq_context: 0 &mm->mmap_lock remove_cache_srcu rcu_read_lock &rcu_state.expedited_wq &p->pi_lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 &sb->s_type->i_lock_key &xa->xa_lock#9 fill_pool_map-wait-type-override pool_lock irq_context: 0 pernet_ops_rwsem rdma_nets_rwsem.wait_lock irq_context: 0 pernet_ops_rwsem devices_rwsem rdma_nets_rwsem &device->compat_devs_mutex &root->kernfs_rwsem &rq->__lock irq_context: 0 pernet_ops_rwsem devices_rwsem rdma_nets_rwsem &device->compat_devs_mutex &root->kernfs_rwsem &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)bat_events (work_completion)(&(&bat_priv->mcast.work)->work) &rcu_state.expedited_wq irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem k-sk_lock-AF_INET6/1 &rq->__lock irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem k-sk_lock-AF_INET6/1 &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 pernet_ops_rwsem devices_rwsem rdma_nets_rwsem &device->compat_devs_mutex lock kernfs_idr_lock batched_entropy_u8.lock irq_context: 0 pernet_ops_rwsem devices_rwsem rdma_nets_rwsem &device->compat_devs_mutex lock kernfs_idr_lock kfence_freelist_lock irq_context: 0 pernet_ops_rwsem &root->kernfs_rwsem pgd_lock irq_context: 0 pernet_ops_rwsem &root->kernfs_rwsem stock_lock irq_context: 0 pernet_ops_rwsem &root->kernfs_rwsem rcu_read_lock &____s->seqcount irq_context: 0 pernet_ops_rwsem &root->kernfs_rwsem key irq_context: 0 pernet_ops_rwsem &root->kernfs_rwsem pcpu_lock irq_context: 0 pernet_ops_rwsem &root->kernfs_rwsem percpu_counters_lock irq_context: 0 pernet_ops_rwsem &root->kernfs_rwsem pcpu_lock stock_lock irq_context: softirq (&ndev->rs_timer) rcu_read_lock rcu_read_lock rcu_read_lock_bh rcu_read_lock_bh rcu_read_lock &c->lock irq_context: softirq (&ndev->rs_timer) rcu_read_lock rcu_read_lock rcu_read_lock_bh dev->qdisc_tx_busylock ?: &qdisc_tx_busylock rcu_read_lock &c->lock irq_context: 0 (wq_completion)wg-crypt-wg1#5 (work_completion)(&peer->transmit_packet_work) &peer->endpoint_lock rcu_read_lock_bh rcu_read_lock rcu_read_lock_bh rcu_read_lock rlock-AF_PACKET irq_context: softirq (&app->join_timer) rcu_read_lock_bh &____s->seqcount#2 irq_context: softirq (&app->join_timer) rcu_read_lock_bh &____s->seqcount irq_context: 0 (wq_completion)wg-crypt-wg0#5 (work_completion)(&peer->transmit_packet_work) &peer->endpoint_lock rcu_read_lock_bh rcu_read_lock rcu_read_lock_bh rcu_read_lock batched_entropy_u8.lock irq_context: 0 (wq_completion)wg-crypt-wg0#5 (work_completion)(&peer->transmit_packet_work) &peer->endpoint_lock rcu_read_lock_bh rcu_read_lock rcu_read_lock_bh rcu_read_lock kfence_freelist_lock irq_context: 0 (wq_completion)wg-crypt-wg0#5 (work_completion)(&peer->transmit_packet_work) &peer->endpoint_lock rcu_read_lock_bh rcu_read_lock rcu_read_lock_bh rcu_read_lock &____s->seqcount#2 irq_context: 0 (wq_completion)wg-crypt-wg0#5 (work_completion)(&peer->transmit_packet_work) &peer->endpoint_lock rcu_read_lock_bh rcu_read_lock rcu_read_lock_bh rcu_read_lock &____s->seqcount irq_context: softirq (&hsr->announce_timer) rcu_read_lock &hsr->seqnr_lock rcu_read_lock rcu_read_lock_bh rcu_read_lock &____s->seqcount#2 irq_context: softirq (&hsr->announce_timer) rcu_read_lock &hsr->seqnr_lock rcu_read_lock rcu_read_lock_bh rcu_read_lock &____s->seqcount irq_context: 0 (wq_completion)wg-crypt-wg2#4 (work_completion)(&({ do { const void *__vpp_verify = (typeof((worker) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __asm__ ("" : "=r"(__ptr) : "0"((typeof(*((worker))) *)((worker)))); (typeof((typeof(*((worker))) *)((worker)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); })->work) batched_entropy_u8.lock irq_context: 0 (wq_completion)wg-crypt-wg2#4 (work_completion)(&({ do { const void *__vpp_verify = (typeof((worker) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __asm__ ("" : "=r"(__ptr) : "0"((typeof(*((worker))) *)((worker)))); (typeof((typeof(*((worker))) *)((worker)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); })->work) kfence_freelist_lock irq_context: 0 pernet_ops_rwsem nf_ct_proto_mutex cpu_hotplug_lock &rq->__lock irq_context: 0 pernet_ops_rwsem nf_ct_proto_mutex cpu_hotplug_lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: softirq (&in_dev->mr_ifc_timer) rcu_read_lock &obj_hash[i].lock pool_lock irq_context: softirq (&in_dev->mr_ifc_timer) rcu_read_lock rcu_read_lock_bh &sch->q.lock batched_entropy_u16.lock irq_context: 0 &sb->s_type->i_mutex_key#8 &sbi->s_writepages_rwsem jbd2_handle &journal->j_state_lock &journal->j_wait_commit irq_context: 0 &sb->s_type->i_mutex_key#8 &sbi->s_writepages_rwsem jbd2_handle &journal->j_state_lock &journal->j_wait_commit &p->pi_lock irq_context: 0 &sb->s_type->i_mutex_key#8 &sbi->s_writepages_rwsem jbd2_handle &journal->j_state_lock &journal->j_wait_commit &p->pi_lock &rq->__lock irq_context: 0 &sb->s_type->i_mutex_key#8 &sbi->s_writepages_rwsem jbd2_handle &journal->j_state_lock &journal->j_wait_commit &p->pi_lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)mld (work_completion)(&(&idev->mc_ifc_work)->work) &idev->mc_lock rcu_read_lock rcu_read_lock rcu_read_lock_bh &sch->q.lock batched_entropy_u16.lock irq_context: 0 rcu_state.exp_mutex &cfs_rq->removed.lock irq_context: 0 rcu_state.exp_mutex pool_lock#2 irq_context: 0 &sb->s_type->i_lock_key &xa->xa_lock#9 rcu_read_lock per_cpu_ptr(&cgroup_rstat_cpu_lock, cpu) irq_context: 0 (wq_completion)ipv6_addrconf (work_completion)(&(&ifa->dad_work)->work) rtnl_mutex rcu_read_lock rcu_read_lock rcu_read_lock_bh &sch->q.lock batched_entropy_u16.lock irq_context: 0 &f->f_pos_lock sb_writers#11 &of->mutex irq_context: 0 sb_writers#11 &type->i_mutex_dir_key#7/1 tomoyo_ss &n->list_lock irq_context: 0 sb_writers#11 &type->i_mutex_dir_key#7/1 tomoyo_ss &n->list_lock &c->lock irq_context: 0 sb_writers#11 &type->i_mutex_dir_key#7/1 cgroup_mutex &cfs_rq->removed.lock irq_context: 0 &xt[i].mutex remove_cache_srcu &rcu_state.expedited_wq irq_context: 0 sb_writers#4 &type->i_mutex_dir_key#3 tomoyo_ss remove_cache_srcu &c->lock irq_context: 0 sb_writers#4 &type->i_mutex_dir_key#3 tomoyo_ss remove_cache_srcu &n->list_lock irq_context: 0 sb_writers#4 &type->i_mutex_dir_key#3 tomoyo_ss remove_cache_srcu &obj_hash[i].lock irq_context: 0 sb_writers#4 &type->i_mutex_dir_key#3 tomoyo_ss remove_cache_srcu pool_lock#2 irq_context: 0 sb_writers#4 &type->i_mutex_dir_key#3 tomoyo_ss remove_cache_srcu &pcp->lock &zone->lock irq_context: 0 sb_writers#4 &type->i_mutex_dir_key#3 tomoyo_ss remove_cache_srcu &pcp->lock &zone->lock &____s->seqcount irq_context: 0 (wq_completion)ipv6_addrconf (work_completion)(&(&ifa->dad_work)->work) rtnl_mutex &idev->mc_lock rcu_read_lock rcu_read_lock rcu_read_lock_bh &sch->q.lock batched_entropy_u16.lock irq_context: 0 (wq_completion)events_unbound (work_completion)(&port->bc_work) &meta->lock irq_context: 0 (wq_completion)events_unbound (work_completion)(&port->bc_work) kfence_freelist_lock irq_context: 0 (wq_completion)mld (work_completion)(&(&idev->mc_dad_work)->work) &idev->mc_lock batched_entropy_u8.lock crngs.lock irq_context: 0 (wq_completion)mld (work_completion)(&(&idev->mc_dad_work)->work) &idev->mc_lock rcu_read_lock rcu_read_lock rcu_read_lock_bh &sch->q.lock batched_entropy_u16.lock irq_context: 0 (wq_completion)wg-kex-wg0#7 (work_completion)(&peer->transmit_handshake_work) &peer->endpoint_lock rcu_read_lock_bh rcu_read_lock &n->list_lock irq_context: 0 (wq_completion)wg-kex-wg0#7 (work_completion)(&peer->transmit_handshake_work) &peer->endpoint_lock rcu_read_lock_bh rcu_read_lock &n->list_lock &c->lock irq_context: 0 sb_writers#4 tomoyo_ss remove_cache_srcu &rq->__lock &cfs_rq->removed.lock irq_context: 0 rtnl_mutex pcpu_alloc_mutex &cfs_rq->removed.lock irq_context: 0 rtnl_mutex pcpu_alloc_mutex &obj_hash[i].lock irq_context: 0 rtnl_mutex pcpu_alloc_mutex pool_lock#2 irq_context: 0 (wq_completion)ipv6_addrconf (work_completion)(&(&ifa->dad_work)->work) rtnl_mutex &idev->mc_lock &bridge_netdev_addr_lock_key &n->list_lock irq_context: 0 (wq_completion)ipv6_addrconf (work_completion)(&(&ifa->dad_work)->work) rtnl_mutex &idev->mc_lock &bridge_netdev_addr_lock_key &n->list_lock &c->lock irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem nfnl_subsys_ipset &rq->__lock irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem nfnl_subsys_ipset &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 cb_lock &rdev->wiphy.mtx &wdev->mtx &local->mtx irq_context: 0 cb_lock &rdev->wiphy.mtx &wdev->mtx &local->mtx &local->chanctx_mtx irq_context: 0 cb_lock &rdev->wiphy.mtx &wdev->mtx &local->mtx &local->chanctx_mtx fs_reclaim irq_context: 0 cb_lock &rdev->wiphy.mtx &wdev->mtx &local->mtx &local->chanctx_mtx fs_reclaim mmu_notifier_invalidate_range_start irq_context: 0 cb_lock &rdev->wiphy.mtx &wdev->mtx &local->mtx &local->chanctx_mtx pool_lock#2 irq_context: 0 cb_lock &rdev->wiphy.mtx &wdev->mtx &local->mtx &local->chanctx_mtx &obj_hash[i].lock irq_context: 0 cb_lock &rdev->wiphy.mtx &wdev->mtx &local->mtx &local->chanctx_mtx &c->lock irq_context: 0 cb_lock &rdev->wiphy.mtx &wdev->mtx &local->mtx &local->chanctx_mtx rcu_read_lock &local->queue_stop_reason_lock irq_context: 0 cb_lock &rdev->wiphy.mtx &wdev->mtx &local->mtx &local->chanctx_mtx rcu_read_lock tk_core.seq.seqcount irq_context: 0 cb_lock &rdev->wiphy.mtx &wdev->mtx &local->mtx &local->chanctx_mtx rcu_read_lock hwsim_radio_lock irq_context: 0 cb_lock &rdev->wiphy.mtx &wdev->mtx &local->mtx &local->chanctx_mtx rcu_read_lock hwsim_radio_lock pool_lock#2 irq_context: 0 cb_lock &rdev->wiphy.mtx &wdev->mtx &local->mtx &local->chanctx_mtx rcu_read_lock hwsim_radio_lock &list->lock#19 irq_context: 0 cb_lock &rdev->wiphy.mtx &wdev->mtx &local->mtx &local->chanctx_mtx rcu_read_lock &list->lock#19 irq_context: 0 cb_lock &rdev->wiphy.mtx &wdev->mtx &local->mtx &local->chanctx_mtx rcu_read_lock rcu_node_0 irq_context: 0 cb_lock &rdev->wiphy.mtx &wdev->mtx &local->mtx &local->chanctx_mtx rcu_read_lock &rq->__lock irq_context: 0 cb_lock &rdev->wiphy.mtx &wdev->mtx &local->mtx &local->chanctx_mtx rcu_read_lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)events (work_completion)(&rdev->wiphy_work) &rdev->wiphy.mtx &wdev->mtx &local->mtx &local->chanctx_mtx rcu_read_lock tk_core.seq.seqcount irq_context: 0 (wq_completion)events (work_completion)(&rdev->wiphy_work) &rdev->wiphy.mtx &wdev->mtx &local->mtx &local->chanctx_mtx rcu_read_lock hwsim_radio_lock irq_context: 0 (wq_completion)events (work_completion)(&rdev->wiphy_work) &rdev->wiphy.mtx &wdev->mtx &local->mtx &local->chanctx_mtx rcu_read_lock hwsim_radio_lock pool_lock#2 irq_context: 0 (wq_completion)events (work_completion)(&rdev->wiphy_work) &rdev->wiphy.mtx &wdev->mtx &local->mtx &local->chanctx_mtx rcu_read_lock hwsim_radio_lock &list->lock#19 irq_context: 0 (wq_completion)events (work_completion)(&rdev->wiphy_work) &rdev->wiphy.mtx &wdev->mtx &local->mtx &local->chanctx_mtx rcu_read_lock &list->lock#19 irq_context: 0 (wq_completion)events (work_completion)(&rdev->wiphy_work) &rdev->wiphy.mtx &wdev->mtx &local->mtx &local->chanctx_mtx nl_table_lock irq_context: 0 (wq_completion)events (work_completion)(&rdev->wiphy_work) &rdev->wiphy.mtx &wdev->mtx &local->mtx &local->chanctx_mtx nl_table_wait.lock irq_context: 0 (wq_completion)events (work_completion)(&rdev->wiphy_work) &rdev->wiphy.mtx &wdev->mtx &local->mtx &local->chanctx_mtx rcu_read_lock &sta->lock irq_context: 0 (wq_completion)events (work_completion)(&rdev->wiphy_work) &rdev->wiphy.mtx &wdev->mtx &local->mtx &local->chanctx_mtx rcu_read_lock &sta->lock &c->lock irq_context: 0 (wq_completion)events (work_completion)(&rdev->wiphy_work) &rdev->wiphy.mtx &wdev->mtx &local->mtx &local->chanctx_mtx rcu_read_lock &sta->lock pool_lock#2 irq_context: 0 (wq_completion)events (work_completion)(&rdev->wiphy_work) &rdev->wiphy.mtx &wdev->mtx &local->mtx &local->chanctx_mtx rcu_read_lock &sta->lock &obj_hash[i].lock irq_context: 0 (wq_completion)events (work_completion)(&rdev->wiphy_work) &rdev->wiphy.mtx &wdev->mtx &local->mtx &local->chanctx_mtx rcu_read_lock &sta->lock krc.lock irq_context: 0 (wq_completion)events (work_completion)(&rdev->wiphy_work) &rdev->wiphy.mtx &wdev->mtx &local->mtx &local->chanctx_mtx rcu_read_lock &pool->lock/1 irq_context: 0 (wq_completion)events (work_completion)(&rdev->wiphy_work) &rdev->wiphy.mtx &wdev->mtx &local->mtx &local->chanctx_mtx rcu_read_lock &pool->lock/1 &obj_hash[i].lock irq_context: 0 (wq_completion)events (work_completion)(&rdev->wiphy_work) &rdev->wiphy.mtx &wdev->mtx &local->mtx &local->chanctx_mtx rcu_read_lock &pool->lock/1 &p->pi_lock irq_context: 0 (wq_completion)events (work_completion)(&rdev->wiphy_work) &rdev->wiphy.mtx &wdev->mtx &local->mtx &local->chanctx_mtx rcu_read_lock &pool->lock/1 &p->pi_lock &cfs_rq->removed.lock irq_context: 0 (wq_completion)events (work_completion)(&rdev->wiphy_work) &rdev->wiphy.mtx &wdev->mtx &local->mtx &local->chanctx_mtx rcu_read_lock &pool->lock/1 &p->pi_lock &rq->__lock irq_context: 0 (wq_completion)events (work_completion)(&rdev->wiphy_work) &rdev->wiphy.mtx &wdev->mtx &local->mtx &local->chanctx_mtx rcu_read_lock &pool->lock/1 &p->pi_lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)events (work_completion)(&rdev->wiphy_work) &rdev->wiphy.mtx &wdev->mtx &local->mtx &local->chanctx_mtx &rq->__lock irq_context: 0 (wq_completion)phy13 (work_completion)(&link->csa_finalize_work) irq_context: 0 (wq_completion)phy13 (work_completion)(&link->csa_finalize_work) &wdev->mtx irq_context: 0 (wq_completion)phy13 (work_completion)(&link->csa_finalize_work) &wdev->mtx &lock->wait_lock irq_context: 0 (wq_completion)phy13 (work_completion)(&link->csa_finalize_work) &wdev->mtx &rq->__lock irq_context: 0 (wq_completion)phy13 (work_completion)(&link->csa_finalize_work) &wdev->mtx &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 cb_lock &rdev->wiphy.mtx &wdev->mtx &local->mtx &local->chanctx_mtx &rq->__lock irq_context: 0 cb_lock &rdev->wiphy.mtx &wdev->mtx &local->mtx &local->chanctx_mtx &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)events (work_completion)(&rdev->wiphy_work) &rdev->wiphy.mtx &p->pi_lock irq_context: 0 (wq_completion)events (work_completion)(&rdev->wiphy_work) &rdev->wiphy.mtx &p->pi_lock &rq->__lock irq_context: 0 (wq_completion)events (work_completion)(&rdev->wiphy_work) &rdev->wiphy.mtx &p->pi_lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)phy13 (work_completion)(&link->csa_finalize_work) &wdev->mtx &local->mtx irq_context: 0 (wq_completion)phy13 (work_completion)(&link->csa_finalize_work) &wdev->mtx &local->mtx &local->chanctx_mtx irq_context: 0 (wq_completion)phy13 (work_completion)(&link->csa_finalize_work) &wdev->mtx &local->mtx &local->chanctx_mtx &rdev->bss_lock irq_context: 0 (wq_completion)phy13 (work_completion)(&link->csa_finalize_work) &wdev->mtx &local->mtx &local->chanctx_mtx fs_reclaim irq_context: 0 (wq_completion)phy13 (work_completion)(&link->csa_finalize_work) &wdev->mtx &local->mtx &local->chanctx_mtx fs_reclaim mmu_notifier_invalidate_range_start irq_context: 0 (wq_completion)phy13 (work_completion)(&link->csa_finalize_work) &wdev->mtx &local->mtx &local->chanctx_mtx &c->lock irq_context: 0 (wq_completion)phy13 (work_completion)(&link->csa_finalize_work) &wdev->mtx &local->mtx &local->chanctx_mtx pool_lock#2 irq_context: 0 (wq_completion)phy13 (work_completion)(&link->csa_finalize_work) &wdev->mtx &local->mtx &local->chanctx_mtx &obj_hash[i].lock irq_context: 0 (wq_completion)phy13 (work_completion)(&link->csa_finalize_work) &wdev->mtx &local->mtx &local->chanctx_mtx krc.lock irq_context: 0 (wq_completion)phy13 (work_completion)(&link->csa_finalize_work) &wdev->mtx &local->mtx &local->chanctx_mtx rcu_read_lock &pool->lock/1 irq_context: 0 (wq_completion)phy13 (work_completion)(&link->csa_finalize_work) &wdev->mtx &local->mtx &local->chanctx_mtx rcu_read_lock &pool->lock/1 &obj_hash[i].lock irq_context: 0 (wq_completion)phy13 (work_completion)(&link->csa_finalize_work) &wdev->mtx &local->mtx &local->chanctx_mtx rcu_read_lock &pool->lock/1 &p->pi_lock irq_context: 0 (wq_completion)phy13 (work_completion)(&link->csa_finalize_work) &wdev->mtx &local->mtx &local->chanctx_mtx rcu_read_lock &pool->lock/1 &p->pi_lock &cfs_rq->removed.lock irq_context: 0 (wq_completion)phy13 (work_completion)(&link->csa_finalize_work) &wdev->mtx &local->mtx &local->chanctx_mtx rcu_read_lock &pool->lock/1 &p->pi_lock &rq->__lock irq_context: 0 (wq_completion)phy13 (work_completion)(&link->csa_finalize_work) &wdev->mtx &local->mtx &local->chanctx_mtx rcu_read_lock &pool->lock/1 &p->pi_lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)phy13 (work_completion)(&link->csa_finalize_work) &wdev->mtx &local->mtx &local->chanctx_mtx &____s->seqcount irq_context: 0 (wq_completion)phy13 (work_completion)(&link->csa_finalize_work) &wdev->mtx &local->mtx &local->chanctx_mtx rcu_read_lock pool_lock#2 irq_context: 0 (wq_completion)phy13 (work_completion)(&link->csa_finalize_work) &wdev->mtx &local->mtx &local->chanctx_mtx nl_table_lock irq_context: 0 (wq_completion)phy13 (work_completion)(&link->csa_finalize_work) &wdev->mtx &local->mtx &local->chanctx_mtx nl_table_wait.lock irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem uevent_sock_mutex rcu_read_lock &ei->socket.wq.wait &ep->lock &ep->wq &p->pi_lock &rq->__lock &cfs_rq->removed.lock irq_context: 0 cb_lock &rdev->wiphy.mtx &wdev->mtx &local->mtx &local->chanctx_mtx nl_table_lock irq_context: 0 cb_lock &rdev->wiphy.mtx &wdev->mtx &local->mtx &local->chanctx_mtx nl_table_wait.lock irq_context: 0 cb_lock &rdev->wiphy.mtx &wdev->mtx &local->mtx &local->chanctx_mtx nl_table_wait.lock &p->pi_lock irq_context: 0 cb_lock &rdev->wiphy.mtx &wdev->mtx &local->mtx &local->chanctx_mtx nl_table_wait.lock &p->pi_lock &rq->__lock irq_context: 0 cb_lock &rdev->wiphy.mtx &wdev->mtx &local->mtx &local->chanctx_mtx nl_table_wait.lock &p->pi_lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 cb_lock &rdev->wiphy.mtx &wdev->mtx &local->mtx &local->chanctx_mtx &data->mutex irq_context: 0 cb_lock &rdev->wiphy.mtx &wdev->mtx &local->mtx &local->chanctx_mtx rcu_read_lock &sta->lock irq_context: 0 cb_lock &rdev->wiphy.mtx &wdev->mtx &local->mtx &local->chanctx_mtx rcu_read_lock &sta->lock pool_lock#2 irq_context: 0 cb_lock &rdev->wiphy.mtx &wdev->mtx &local->mtx &local->chanctx_mtx rcu_read_lock &sta->lock &obj_hash[i].lock irq_context: 0 cb_lock &rdev->wiphy.mtx &wdev->mtx &local->mtx &local->chanctx_mtx rcu_read_lock &sta->lock krc.lock irq_context: 0 cb_lock &rdev->wiphy.mtx &wdev->mtx &local->mtx &local->chanctx_mtx rcu_read_lock &pool->lock/1 irq_context: 0 cb_lock &rdev->wiphy.mtx &wdev->mtx &local->mtx &local->chanctx_mtx rcu_read_lock &pool->lock/1 &obj_hash[i].lock irq_context: 0 cb_lock &rdev->wiphy.mtx &wdev->mtx &local->mtx &local->chanctx_mtx rcu_read_lock &pool->lock/1 &p->pi_lock irq_context: 0 cb_lock &rdev->wiphy.mtx &wdev->mtx &local->mtx &local->chanctx_mtx rcu_read_lock &pool->lock/1 &p->pi_lock &rq->__lock irq_context: 0 cb_lock &rdev->wiphy.mtx &wdev->mtx &local->mtx &local->chanctx_mtx rcu_read_lock &pool->lock/1 &p->pi_lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)phy10 (work_completion)(&link->csa_finalize_work) irq_context: 0 (wq_completion)phy10 (work_completion)(&link->csa_finalize_work) &wdev->mtx irq_context: 0 (wq_completion)phy10 (work_completion)(&link->csa_finalize_work) &wdev->mtx &lock->wait_lock irq_context: 0 (wq_completion)phy10 (work_completion)(&link->csa_finalize_work) &wdev->mtx &rq->__lock irq_context: 0 (wq_completion)phy10 (work_completion)(&link->csa_finalize_work) &wdev->mtx &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)phy10 (work_completion)(&link->csa_finalize_work) &wdev->mtx &rq->__lock &cfs_rq->removed.lock irq_context: 0 cb_lock &rdev->wiphy.mtx &wdev->mtx &local->mtx &local->chanctx_mtx krc.lock irq_context: 0 cb_lock &rdev->wiphy.mtx &lock->wait_lock irq_context: 0 (wq_completion)phy10 (work_completion)(&link->csa_finalize_work) &wdev->mtx &local->mtx irq_context: 0 (wq_completion)phy10 (work_completion)(&link->csa_finalize_work) &wdev->mtx &local->mtx &local->chanctx_mtx irq_context: 0 (wq_completion)phy10 (work_completion)(&link->csa_finalize_work) &wdev->mtx &local->mtx &local->chanctx_mtx &rdev->bss_lock irq_context: 0 (wq_completion)phy10 (work_completion)(&link->csa_finalize_work) &wdev->mtx &local->mtx &local->chanctx_mtx fs_reclaim irq_context: 0 (wq_completion)phy10 (work_completion)(&link->csa_finalize_work) &wdev->mtx &local->mtx &local->chanctx_mtx fs_reclaim mmu_notifier_invalidate_range_start irq_context: 0 (wq_completion)phy10 (work_completion)(&link->csa_finalize_work) &wdev->mtx &local->mtx &local->chanctx_mtx &obj_hash[i].lock irq_context: 0 (wq_completion)phy10 (work_completion)(&link->csa_finalize_work) &wdev->mtx &local->mtx &local->chanctx_mtx krc.lock irq_context: 0 (wq_completion)phy10 (work_completion)(&link->csa_finalize_work) &wdev->mtx &local->mtx &local->chanctx_mtx rcu_read_lock &pool->lock/1 irq_context: 0 (wq_completion)phy10 (work_completion)(&link->csa_finalize_work) &wdev->mtx &local->mtx &local->chanctx_mtx rcu_read_lock &pool->lock/1 &obj_hash[i].lock irq_context: 0 (wq_completion)phy10 (work_completion)(&link->csa_finalize_work) &wdev->mtx &local->mtx &local->chanctx_mtx rcu_read_lock &pool->lock/1 &p->pi_lock irq_context: 0 (wq_completion)phy10 (work_completion)(&link->csa_finalize_work) &wdev->mtx &local->mtx &local->chanctx_mtx rcu_read_lock &pool->lock/1 &p->pi_lock &cfs_rq->removed.lock irq_context: 0 (wq_completion)phy10 (work_completion)(&link->csa_finalize_work) &wdev->mtx &local->mtx &local->chanctx_mtx rcu_read_lock &pool->lock/1 &p->pi_lock &rq->__lock irq_context: 0 (wq_completion)phy10 (work_completion)(&link->csa_finalize_work) &wdev->mtx &local->mtx &local->chanctx_mtx rcu_read_lock &pool->lock/1 &p->pi_lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)phy10 (work_completion)(&link->csa_finalize_work) &wdev->mtx &local->mtx &local->chanctx_mtx &rq->__lock irq_context: 0 (wq_completion)phy10 (work_completion)(&link->csa_finalize_work) &wdev->mtx &local->mtx &local->chanctx_mtx &c->lock irq_context: 0 (wq_completion)phy10 (work_completion)(&link->csa_finalize_work) &wdev->mtx &local->mtx &local->chanctx_mtx nl_table_lock irq_context: 0 (wq_completion)phy10 (work_completion)(&link->csa_finalize_work) &wdev->mtx &local->mtx &local->chanctx_mtx nl_table_wait.lock irq_context: 0 sb_writers#11 &type->i_mutex_dir_key#7/1 cgroup_mutex shrinker_rwsem rcu_read_lock pool_lock#2 irq_context: 0 sb_writers#11 &type->i_mutex_dir_key#7/1 cgroup_mutex shrinker_rwsem &obj_hash[i].lock irq_context: 0 sb_writers#4 &sb->s_type->i_mutex_key#8 jbd2_handle &ei->i_data_sem remove_cache_srcu pool_lock#2 irq_context: 0 (wq_completion)writeback (work_completion)(&(&wb->dwork)->work) &sbi->s_writepages_rwsem &mapping->i_mmap_rwsem mmu_notifier_invalidate_range_start irq_context: 0 (wq_completion)writeback (work_completion)(&(&wb->dwork)->work) &sbi->s_writepages_rwsem &mapping->i_mmap_rwsem ptlock_ptr(page)#2 irq_context: 0 (wq_completion)writeback (work_completion)(&(&wb->dwork)->work) &sbi->s_writepages_rwsem &mapping->private_lock irq_context: 0 (wq_completion)writeback (work_completion)(&(&wb->dwork)->work) &sbi->s_writepages_rwsem &mapping->private_lock rcu_read_lock &memcg->move_lock irq_context: 0 (wq_completion)writeback (work_completion)(&(&wb->dwork)->work) &sbi->s_writepages_rwsem rcu_read_lock &memcg->move_lock irq_context: 0 (wq_completion)writeback (work_completion)(&(&wb->dwork)->work) &sbi->s_writepages_rwsem rcu_read_lock &xa->xa_lock#9 irq_context: 0 (wq_completion)writeback (work_completion)(&(&wb->dwork)->work) &sbi->s_writepages_rwsem rcu_read_lock &xa->xa_lock#9 &s->s_inode_wblist_lock irq_context: 0 (wq_completion)writeback (work_completion)(&(&wb->dwork)->work) &sbi->s_writepages_rwsem jbd2_handle &ei->i_data_sem &ei->i_prealloc_lock &pa->pa_lock#2 irq_context: 0 &hdev->lock irq_context: 0 &hdev->lock fs_reclaim irq_context: 0 &hdev->lock fs_reclaim mmu_notifier_invalidate_range_start irq_context: 0 (wq_completion)writeback (work_completion)(&(&wb->dwork)->work) &sbi->s_writepages_rwsem jbd2_handle &journal->j_wait_reserved irq_context: 0 &hdev->lock &c->lock irq_context: 0 &hdev->lock pool_lock#2 irq_context: 0 &hdev->lock &obj_hash[i].lock irq_context: 0 &hdev->lock &x->wait#9 irq_context: 0 &hdev->lock &rq->__lock irq_context: 0 &hdev->lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 &mm->mmap_lock sb_pagefaults &sb->s_type->i_lock_key#22 irq_context: 0 &hdev->lock &list->lock#7 irq_context: 0 &hdev->lock rcu_read_lock &pool->lock/1 irq_context: 0 &hdev->lock rcu_read_lock &pool->lock/1 &obj_hash[i].lock irq_context: 0 &hdev->lock rcu_read_lock &pool->lock/1 &p->pi_lock irq_context: 0 &hdev->lock rcu_read_lock &pool->lock/1 &p->pi_lock &rq->__lock irq_context: 0 &hdev->lock rcu_read_lock &pool->lock/1 &p->pi_lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 &hdev->lock sk_lock-AF_BLUETOOTH-BTPROTO_SCO irq_context: 0 &hdev->lock sk_lock-AF_BLUETOOTH-BTPROTO_SCO slock-AF_BLUETOOTH-BTPROTO_SCO irq_context: 0 &hdev->lock sk_lock-AF_BLUETOOTH-BTPROTO_SCO &conn->lock#2 irq_context: 0 &hdev->lock sk_lock-AF_BLUETOOTH-BTPROTO_SCO &obj_hash[i].lock irq_context: 0 &hdev->lock sk_lock-AF_BLUETOOTH-BTPROTO_SCO &base->lock irq_context: 0 &hdev->lock sk_lock-AF_BLUETOOTH-BTPROTO_SCO &base->lock &obj_hash[i].lock irq_context: 0 &hdev->lock sk_lock-AF_BLUETOOTH-BTPROTO_SCO &rq->__lock irq_context: 0 &hdev->lock sk_lock-AF_BLUETOOTH-BTPROTO_SCO &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 &hdev->lock slock-AF_BLUETOOTH-BTPROTO_SCO irq_context: 0 sk_lock-AF_BLUETOOTH-BTPROTO_SCO &ei->socket.wq.wait irq_context: 0 &sb->s_type->i_mutex_key#10 sk_lock-AF_BLUETOOTH-BTPROTO_SCO &obj_hash[i].lock irq_context: 0 &sb->s_type->i_mutex_key#10 sk_lock-AF_BLUETOOTH-BTPROTO_SCO &base->lock irq_context: 0 &sb->s_type->i_mutex_key#10 sk_lock-AF_BLUETOOTH-BTPROTO_SCO &base->lock &obj_hash[i].lock irq_context: 0 &sb->s_type->i_mutex_key#10 sk_lock-AF_BLUETOOTH-BTPROTO_SCO &conn->lock#2 irq_context: 0 &sb->s_type->i_mutex_key#10 sk_lock-AF_BLUETOOTH-BTPROTO_SCO rcu_read_lock &pool->lock/1 irq_context: 0 &sb->s_type->i_mutex_key#10 sk_lock-AF_BLUETOOTH-BTPROTO_SCO rcu_read_lock &pool->lock/1 &obj_hash[i].lock irq_context: 0 &sb->s_type->i_mutex_key#10 sk_lock-AF_BLUETOOTH-BTPROTO_SCO rcu_read_lock &pool->lock/1 pool_lock#2 irq_context: 0 &sb->s_type->i_mutex_key#10 sk_lock-AF_BLUETOOTH-BTPROTO_SCO rcu_read_lock &pool->lock/1 &p->pi_lock irq_context: 0 &sb->s_type->i_mutex_key#10 sk_lock-AF_BLUETOOTH-BTPROTO_SCO rcu_read_lock &pool->lock/1 &p->pi_lock &rq->__lock irq_context: 0 &sb->s_type->i_mutex_key#10 sk_lock-AF_BLUETOOTH-BTPROTO_SCO rcu_read_lock &pool->lock/1 &p->pi_lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 &sb->s_type->i_mutex_key#10 sk_lock-AF_BLUETOOTH-BTPROTO_SCO &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)hci5 (work_completion)(&hdev->cmd_sync_work) &hdev->req_lock &hdev->lock hci_cb_list_lock fs_reclaim irq_context: 0 (wq_completion)hci5 (work_completion)(&hdev->cmd_sync_work) &hdev->req_lock &hdev->lock hci_cb_list_lock fs_reclaim mmu_notifier_invalidate_range_start irq_context: 0 (wq_completion)hci5 (work_completion)(&hdev->cmd_sync_work) &hdev->req_lock &hdev->lock hci_cb_list_lock chan_list_lock irq_context: 0 (wq_completion)hci5 (work_completion)(&hdev->cmd_sync_work) &hdev->req_lock &hdev->lock hci_cb_list_lock &conn->ident_lock irq_context: 0 (wq_completion)hci5 (work_completion)(&hdev->cmd_sync_work) &hdev->req_lock &hdev->lock hci_cb_list_lock &base->lock irq_context: 0 (wq_completion)hci5 (work_completion)(&hdev->cmd_sync_work) &hdev->req_lock &hdev->lock hci_cb_list_lock &base->lock &obj_hash[i].lock irq_context: 0 (wq_completion)hci5 (work_completion)(&hdev->cmd_sync_work) &hdev->req_lock &hdev->lock hci_cb_list_lock rcu_read_lock &pool->lock/1 irq_context: 0 (wq_completion)hci5 (work_completion)(&hdev->cmd_sync_work) &hdev->req_lock &hdev->lock hci_cb_list_lock rcu_read_lock &pool->lock/1 &obj_hash[i].lock irq_context: 0 (wq_completion)hci5 (work_completion)(&hdev->cmd_sync_work) &hdev->req_lock &hdev->lock hci_cb_list_lock rcu_read_lock &pool->lock/1 &p->pi_lock irq_context: 0 (wq_completion)hci5 (work_completion)(&hdev->cmd_sync_work) &hdev->req_lock &hdev->lock hci_cb_list_lock rcu_read_lock &pool->lock/1 pool_lock#2 irq_context: 0 (wq_completion)hci5 (work_completion)(&hdev->cmd_sync_work) &hdev->req_lock &hdev->lock rcu_read_lock &pool->lock/1 &obj_hash[i].lock irq_context: 0 (wq_completion)hci5 (work_completion)(&hdev->cmd_sync_work) &hdev->req_lock &hdev->lock rcu_read_lock &pool->lock/1 pool_lock#2 irq_context: 0 (wq_completion)hci5 (work_completion)(&hdev->cmd_sync_work) &hdev->req_lock &hdev->lock rcu_read_lock &pool->lock/1 &p->pi_lock irq_context: 0 (wq_completion)hci5 (work_completion)(&hdev->cmd_sync_work) &hdev->req_lock &hdev->lock rcu_read_lock &pool->lock/1 &p->pi_lock &rq->__lock irq_context: 0 (wq_completion)hci5 (work_completion)(&hdev->cmd_sync_work) &hdev->req_lock &hdev->lock rcu_read_lock &pool->lock/1 &p->pi_lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)hci5 (work_completion)(&hdev->cmd_sync_work) &hdev->req_lock &hdev->lock hci_cb_list_lock &conn->lock#2 irq_context: 0 (wq_completion)hci5 (work_completion)(&hdev->cmd_sync_work) &hdev->req_lock &hdev->lock hci_cb_list_lock (work_completion)(&(&conn->timeout_work)->work) irq_context: 0 (wq_completion)hci5 (work_completion)(&hdev->cmd_sync_work) &hdev->req_lock &hdev->lock rcu_state.exp_mutex &rnp->exp_wq[2] irq_context: 0 (wq_completion)wg-kex-wg1#2 (work_completion)(&peer->transmit_handshake_work) &peer->endpoint_lock rcu_read_lock_bh pool_lock#2 irq_context: 0 (wq_completion)hci5 (work_completion)(&hdev->cmd_sync_work) &hdev->req_lock &hdev->lock rcu_state.exp_mutex &rnp->exp_wq[3] irq_context: 0 (wq_completion)hci5 (work_completion)(&hdev->cmd_sync_work) &hdev->req_lock &hdev->lock rcu_state.exp_mutex &rnp->exp_wq[0] irq_context: 0 sb_writers#4 &sb->s_type->i_mutex_key#8 jbd2_handle pgd_lock irq_context: 0 sb_writers#4 &sb->s_type->i_mutex_key#8 jbd2_handle key irq_context: 0 sb_writers#4 &sb->s_type->i_mutex_key#8 jbd2_handle pcpu_lock irq_context: 0 sb_writers#4 &sb->s_type->i_mutex_key#8 jbd2_handle percpu_counters_lock irq_context: 0 sb_writers#4 &sb->s_type->i_mutex_key#8 jbd2_handle pcpu_lock stock_lock irq_context: 0 &hdev->lock &lock->wait_lock irq_context: 0 (wq_completion)hci5 (work_completion)(&hdev->cmd_sync_work) &hdev->req_lock &hdev->lock rcu_state.exp_mutex irq_context: 0 (wq_completion)hci5 (work_completion)(&hdev->cmd_sync_work) &hdev->req_lock &hdev->lock rcu_state.exp_mutex rcu_state.exp_mutex.wait_lock irq_context: 0 (wq_completion)hci5 (work_completion)(&hdev->cmd_sync_work) &hdev->req_lock &hdev->lock &rnp->exp_wq[2] irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem k-sk_lock-AF_TIPC &tn->nametbl_lock &obj_hash[i].lock pool_lock irq_context: 0 &hdev->lock &n->list_lock irq_context: 0 &hdev->lock &n->list_lock &c->lock irq_context: 0 &hdev->lock rcu_node_0 irq_context: 0 &hdev->lock rcu_read_lock rcu_node_0 irq_context: 0 &hdev->lock &obj_hash[i].lock pool_lock irq_context: 0 &sb->s_type->i_mutex_key#10 sk_lock-AF_BLUETOOTH-BTPROTO_SCO rcu_node_0 irq_context: 0 (wq_completion)hci5 (work_completion)(&hdev->cmd_sync_work) &hdev->req_lock &hdev->lock hci_cb_list_lock &c->lock irq_context: 0 (wq_completion)hci5 (work_completion)(&hdev->cmd_sync_work) &hdev->req_lock &hdev->lock &rnp->exp_lock irq_context: 0 (wq_completion)hci5 (work_completion)(&hdev->cmd_sync_work) &hdev->req_lock &hdev->lock &rq->__lock irq_context: 0 (wq_completion)hci5 (work_completion)(&hdev->cmd_sync_work) &hdev->req_lock &hdev->lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)hci5 (work_completion)(&hdev->cmd_sync_work) &hdev->req_lock &hdev->lock &rnp->exp_wq[3] irq_context: 0 &sb->s_type->i_mutex_key#10 sk_lock-AF_BLUETOOTH-BTPROTO_SCO rcu_read_lock rcu_node_0 irq_context: 0 (wq_completion)hci5 (work_completion)(&hdev->cmd_sync_work) &hdev->req_lock &lock->wait_lock irq_context: 0 (wq_completion)hci5 (work_completion)(&hdev->cmd_sync_work) &hdev->req_lock &p->pi_lock irq_context: 0 (wq_completion)hci5 (work_completion)(&hdev->cmd_sync_work) &hdev->req_lock &p->pi_lock &rq->__lock irq_context: 0 (wq_completion)hci5 (work_completion)(&hdev->cmd_sync_work) &hdev->req_lock &p->pi_lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 sk_lock-AF_LLC llc_sap_list_lock batched_entropy_u8.lock irq_context: 0 (wq_completion)hci5 (work_completion)(&hdev->cmd_sync_work) &hdev->req_lock &hdev->lock hci_cb_list_lock rcu_read_lock &pool->lock/1 &p->pi_lock &rq->__lock irq_context: 0 (wq_completion)hci5 (work_completion)(&hdev->cmd_sync_work) &hdev->req_lock &hdev->lock hci_cb_list_lock rcu_read_lock &pool->lock/1 &p->pi_lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)hci5 (work_completion)(&hdev->cmd_sync_work) &hdev->req_lock &hdev->lock hci_cb_list_lock &rq->__lock irq_context: 0 sk_lock-AF_LLC &rcu_state.expedited_wq irq_context: 0 sk_lock-AF_LLC &rcu_state.expedited_wq &p->pi_lock irq_context: 0 sk_lock-AF_LLC &rcu_state.expedited_wq &p->pi_lock &rq->__lock irq_context: 0 sk_lock-AF_LLC &rcu_state.expedited_wq &p->pi_lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 rtnl_mutex &idev->mc_lock _xmit_ETHER/1 pool_lock#2 irq_context: 0 sk_lock-AF_LLC llc_sap_list_lock kfence_freelist_lock irq_context: 0 (wq_completion)wg-kex-wg2#11 (work_completion)(&peer->transmit_handshake_work) &peer->endpoint_lock rcu_read_lock_bh &____s->seqcount#2 irq_context: 0 (wq_completion)wg-kex-wg2#11 (work_completion)(&peer->transmit_handshake_work) &peer->endpoint_lock rcu_read_lock_bh &____s->seqcount irq_context: 0 (wq_completion)wg-kex-wg2#11 (work_completion)(&peer->transmit_handshake_work) &peer->endpoint_lock rcu_read_lock_bh rcu_read_lock &tbl->lock irq_context: 0 cb_lock genl_mutex rtnl_mutex &wg->device_update_lock &list->lock#17 irq_context: 0 cb_lock &rdev->wiphy.mtx &wdev->mtx &obj_hash[i].lock irq_context: 0 cb_lock &rdev->wiphy.mtx &wdev->mtx pool_lock#2 irq_context: 0 sb_writers#4 &journal->j_barrier &journal->j_state_lock &journal->j_wait_commit &p->pi_lock &cfs_rq->removed.lock irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem rtnl_mutex &dev->mutex &rq->__lock irq_context: 0 (wq_completion)events free_ipc_work &rq->__lock &cfs_rq->removed.lock irq_context: 0 cb_lock &rdev->wiphy.mtx &wdev->mtx &local->mtx &local->chanctx_mtx &n->list_lock irq_context: 0 cb_lock &rdev->wiphy.mtx &wdev->mtx &local->mtx &local->chanctx_mtx &n->list_lock &c->lock irq_context: 0 (wq_completion)phy3 (work_completion)(&link->csa_finalize_work) irq_context: 0 (wq_completion)phy3 (work_completion)(&link->csa_finalize_work) &wdev->mtx irq_context: 0 (wq_completion)phy3 (work_completion)(&link->csa_finalize_work) &wdev->mtx &local->mtx irq_context: 0 (wq_completion)phy3 (work_completion)(&link->csa_finalize_work) &wdev->mtx &local->mtx &local->chanctx_mtx irq_context: 0 (wq_completion)phy3 (work_completion)(&link->csa_finalize_work) &wdev->mtx &local->mtx &local->chanctx_mtx &rdev->bss_lock irq_context: 0 (wq_completion)phy3 (work_completion)(&link->csa_finalize_work) &wdev->mtx &local->mtx &local->chanctx_mtx fs_reclaim irq_context: 0 (wq_completion)phy3 (work_completion)(&link->csa_finalize_work) &wdev->mtx &local->mtx &local->chanctx_mtx fs_reclaim mmu_notifier_invalidate_range_start irq_context: 0 (wq_completion)phy3 (work_completion)(&link->csa_finalize_work) &wdev->mtx &local->mtx &local->chanctx_mtx &obj_hash[i].lock irq_context: 0 (wq_completion)phy3 (work_completion)(&link->csa_finalize_work) &wdev->mtx &local->mtx &local->chanctx_mtx krc.lock irq_context: 0 (wq_completion)phy3 (work_completion)(&link->csa_finalize_work) &wdev->mtx &local->mtx &local->chanctx_mtx rcu_read_lock &pool->lock/1 irq_context: 0 (wq_completion)phy3 (work_completion)(&link->csa_finalize_work) &wdev->mtx &local->mtx &local->chanctx_mtx rcu_read_lock &pool->lock/1 &obj_hash[i].lock irq_context: 0 (wq_completion)phy3 (work_completion)(&link->csa_finalize_work) &wdev->mtx &local->mtx &local->chanctx_mtx rcu_read_lock &pool->lock/1 &p->pi_lock irq_context: 0 (wq_completion)phy3 (work_completion)(&link->csa_finalize_work) &wdev->mtx &local->mtx &local->chanctx_mtx rcu_read_lock &pool->lock/1 &p->pi_lock &rq->__lock irq_context: 0 (wq_completion)phy3 (work_completion)(&link->csa_finalize_work) &wdev->mtx &local->mtx &local->chanctx_mtx rcu_read_lock &pool->lock/1 &p->pi_lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)phy3 (work_completion)(&link->csa_finalize_work) &wdev->mtx &local->mtx &local->chanctx_mtx nl_table_lock irq_context: 0 (wq_completion)phy3 (work_completion)(&link->csa_finalize_work) &wdev->mtx &local->mtx &local->chanctx_mtx nl_table_wait.lock irq_context: 0 &f->f_pos_lock sb_writers#4 &sb->s_type->i_mutex_key#8 &type->s_umount_key#32 rcu_read_lock &c->lock irq_context: 0 (wq_completion)writeback (work_completion)(&(&wb->dwork)->work) &sbi->s_writepages_rwsem rcu_read_lock rcu_read_lock per_cpu_ptr(&cgroup_rstat_cpu_lock, cpu) irq_context: 0 (wq_completion)writeback (work_completion)(&(&wb->dwork)->work) &sbi->s_writepages_rwsem jbd2_handle &ei->i_data_sem &ei->i_prealloc_lock irq_context: 0 (wq_completion)writeback (work_completion)(&(&wb->dwork)->work) &sbi->s_writepages_rwsem jbd2_handle rcu_read_lock rcu_read_lock per_cpu_ptr(&cgroup_rstat_cpu_lock, cpu) irq_context: 0 (wq_completion)writeback (work_completion)(&(&wb->dwork)->work) &sbi->s_writepages_rwsem jbd2_handle rcu_read_lock &rq->__lock irq_context: 0 (wq_completion)writeback (work_completion)(&(&wb->dwork)->work) &sbi->s_writepages_rwsem jbd2_handle lock#4 &lruvec->lru_lock irq_context: 0 (wq_completion)writeback (work_completion)(&(&wb->dwork)->work) &sbi->s_writepages_rwsem jbd2_handle rcu_read_lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)writeback (work_completion)(&(&wb->dwork)->work) &sbi->s_writepages_rwsem &rq->__lock irq_context: 0 (wq_completion)writeback (work_completion)(&(&wb->dwork)->work) &sbi->s_writepages_rwsem &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 &mm->mmap_lock sb_pagefaults &wb->list_lock irq_context: 0 &mm->mmap_lock sb_pagefaults &wb->list_lock &sb->s_type->i_lock_key#22 irq_context: 0 &mm->mmap_lock sb_pagefaults mapping.invalidate_lock &sb->s_type->i_lock_key#22 irq_context: 0 (wq_completion)writeback (work_completion)(&(&wb->dwork)->work) &sbi->s_writepages_rwsem &____s->seqcount#2 irq_context: 0 (wq_completion)writeback (work_completion)(&(&wb->dwork)->work) &sbi->s_writepages_rwsem &____s->seqcount irq_context: 0 (wq_completion)writeback (work_completion)(&(&wb->dwork)->work) &sbi->s_writepages_rwsem jbd2_handle &sem->wait_lock irq_context: 0 (wq_completion)writeback (work_completion)(&(&wb->dwork)->work) &sbi->s_writepages_rwsem jbd2_handle &p->pi_lock irq_context: 0 (wq_completion)writeback (work_completion)(&(&wb->dwork)->work) &sbi->s_writepages_rwsem jbd2_handle &p->pi_lock &rq->__lock irq_context: 0 (wq_completion)writeback (work_completion)(&(&wb->dwork)->work) &sbi->s_writepages_rwsem jbd2_handle &p->pi_lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)writeback (work_completion)(&(&wb->dwork)->work) &sbi->s_writepages_rwsem jbd2_handle &journal->j_wait_updates irq_context: 0 (wq_completion)writeback (work_completion)(&(&wb->dwork)->work) &sbi->s_writepages_rwsem rcu_read_lock &c->lock irq_context: 0 (wq_completion)writeback (work_completion)(&(&wb->dwork)->work) &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 rtnl_mutex bpf_dispatcher_xdp.mutex rcu_state.exp_mutex &rnp->exp_wq[0] irq_context: 0 (wq_completion)mld (work_completion)(&(&idev->mc_ifc_work)->work) &idev->mc_lock rcu_read_lock rcu_read_lock rcu_read_lock_bh rcu_read_lock rcu_read_lock &table->lock#3 irq_context: softirq (&ndev->rs_timer) rcu_read_lock rcu_read_lock rcu_read_lock_bh &sch->q.lock crngs.lock irq_context: softirq rcu_read_lock &br->multicast_lock rcu_read_lock rcu_read_lock &pool->lock irq_context: softirq rcu_read_lock &br->multicast_lock rcu_read_lock rcu_read_lock &pool->lock &obj_hash[i].lock irq_context: softirq rcu_read_lock &br->multicast_lock rcu_read_lock rcu_read_lock &pool->lock pool_lock#2 irq_context: 0 sk_lock-AF_LLC rcu_read_lock &cfs_rq->removed.lock irq_context: 0 sk_lock-AF_LLC rcu_read_lock &obj_hash[i].lock irq_context: 0 sk_lock-AF_LLC rcu_read_lock pool_lock#2 irq_context: 0 sb_writers#4 &sb->s_type->i_mutex_key#8 mapping.invalidate_lock mmu_notifier_invalidate_range_start &rq->__lock irq_context: 0 sb_writers#4 &sb->s_type->i_mutex_key#8 mapping.invalidate_lock mmu_notifier_invalidate_range_start &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)mld (work_completion)(&(&idev->mc_ifc_work)->work) &idev->mc_lock rcu_read_lock rcu_read_lock rcu_read_lock_bh dev->qdisc_tx_busylock ?: &qdisc_tx_busylock _xmit_ETHER#2 &meta->lock irq_context: 0 (wq_completion)mld (work_completion)(&(&idev->mc_ifc_work)->work) &idev->mc_lock rcu_read_lock rcu_read_lock rcu_read_lock_bh dev->qdisc_tx_busylock ?: &qdisc_tx_busylock _xmit_ETHER#2 kfence_freelist_lock irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem rtnl_mutex &caifn->caifdevs.lock rcu_state.exp_mutex irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem rtnl_mutex &caifn->caifdevs.lock rcu_state.exp_mutex rcu_state.exp_mutex.wait_lock irq_context: 0 rcu_read_lock rcu_read_lock_bh noop_qdisc.q.lock irq_context: 0 rtnl_mutex bpf_dispatcher_xdp.mutex rcu_state.exp_mutex &obj_hash[i].lock pool_lock irq_context: 0 rtnl_mutex bpf_dispatcher_xdp.mutex rcu_state.exp_mutex &rnp->exp_wq[3] irq_context: 0 rtnl_mutex prog_idr_lock irq_context: 0 rtnl_mutex bpf_lock irq_context: 0 pernet_ops_rwsem devices_rwsem rdma_nets_rwsem &device->compat_devs_mutex &rxe->usdev_lock rtnl_mutex &cfs_rq->removed.lock irq_context: 0 pernet_ops_rwsem devices_rwsem rdma_nets_rwsem &device->compat_devs_mutex &rxe->usdev_lock rtnl_mutex &obj_hash[i].lock irq_context: 0 pernet_ops_rwsem devices_rwsem rdma_nets_rwsem &device->compat_devs_mutex &rxe->usdev_lock rtnl_mutex pool_lock#2 irq_context: 0 pernet_ops_rwsem devices_rwsem rdma_nets_rwsem &device->compat_devs_mutex &rxe->usdev_lock &p->pi_lock &rq->__lock &cfs_rq->removed.lock irq_context: 0 sb_writers#4 &sb->s_type->i_mutex_key#8 mapping.invalidate_lock jbd2_handle &sbi->s_orphan_lock &mapping->private_lock irq_context: softirq (&llc->ack_timer.timer) irq_context: softirq (&llc->ack_timer.timer) pool_lock#2 irq_context: softirq (&llc->ack_timer.timer) slock-AF_LLC irq_context: softirq (&llc->ack_timer.timer) slock-AF_LLC pool_lock#2 irq_context: softirq (&llc->ack_timer.timer) slock-AF_LLC wlock-AF_LLC irq_context: softirq (&llc->ack_timer.timer) slock-AF_LLC rcu_read_lock_bh noop_qdisc.q.lock irq_context: softirq (&llc->ack_timer.timer) slock-AF_LLC rcu_read_lock_bh rcu_read_lock &ei->socket.wq.wait irq_context: softirq (&llc->ack_timer.timer) slock-AF_LLC rcu_read_lock_bh rcu_read_lock &ei->socket.wq.wait &p->pi_lock irq_context: softirq (&llc->ack_timer.timer) slock-AF_LLC rcu_read_lock_bh rcu_read_lock &ei->socket.wq.wait &p->pi_lock &rq->__lock irq_context: softirq (&llc->ack_timer.timer) slock-AF_LLC rcu_read_lock_bh rcu_read_lock &ei->socket.wq.wait &p->pi_lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: softirq (&llc->ack_timer.timer) slock-AF_LLC rcu_read_lock_bh &obj_hash[i].lock irq_context: softirq (&llc->ack_timer.timer) slock-AF_LLC rcu_read_lock_bh pool_lock#2 irq_context: softirq (&llc->ack_timer.timer) slock-AF_LLC &obj_hash[i].lock irq_context: softirq (&llc->ack_timer.timer) slock-AF_LLC &base->lock irq_context: softirq (&llc->ack_timer.timer) slock-AF_LLC &base->lock &obj_hash[i].lock irq_context: softirq (&ndev->rs_timer) rcu_read_lock rcu_read_lock rcu_read_lock_bh &sch->q.lock batched_entropy_u16.lock irq_context: 0 (wq_completion)events_unbound (work_completion)(&map->work) &rnp->exp_wq[1] irq_context: 0 (wq_completion)events_unbound (work_completion)(&map->work) &rq->__lock irq_context: 0 (wq_completion)events_unbound (work_completion)(&map->work) &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)events (work_completion)(&aux->work) &meta->lock irq_context: 0 (wq_completion)events (work_completion)(&aux->work) kfence_freelist_lock irq_context: 0 (wq_completion)wg-kex-wg0#9 (work_completion)(&peer->transmit_handshake_work) kfence_freelist_lock irq_context: 0 (wq_completion)wg-kex-wg0#9 (work_completion)(&peer->transmit_handshake_work) &peer->endpoint_lock rcu_read_lock_bh &meta->lock irq_context: 0 (wq_completion)wg-kex-wg0#9 (work_completion)(&peer->transmit_handshake_work) &peer->endpoint_lock rcu_read_lock_bh kfence_freelist_lock irq_context: 0 (wq_completion)hci5#2 (work_completion)(&(&hdev->cmd_timer)->work) rcu_node_0 irq_context: 0 (wq_completion)hci5#2 (work_completion)(&(&hdev->cmd_timer)->work) &rcu_state.expedited_wq irq_context: 0 (wq_completion)hci5#2 (work_completion)(&(&hdev->cmd_timer)->work) &rcu_state.expedited_wq &p->pi_lock irq_context: 0 (wq_completion)hci5#2 (work_completion)(&(&hdev->cmd_timer)->work) &rcu_state.expedited_wq &p->pi_lock &rq->__lock irq_context: 0 (wq_completion)hci5#2 (work_completion)(&(&hdev->cmd_timer)->work) &rcu_state.expedited_wq &p->pi_lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)wg-kex-wg0#7 (work_completion)(&peer->transmit_handshake_work) &peer->endpoint_lock rcu_read_lock_bh pool_lock#2 irq_context: 0 (wq_completion)events (debug_obj_work).work quarantine_lock irq_context: 0 cgroup_threadgroup_rwsem freezer_mutex rcu_node_0 irq_context: 0 sb_writers#4 &sb->s_type->i_mutex_key#8 mapping.invalidate_lock jbd2_handle rcu_read_lock &rcu_state.expedited_wq irq_context: 0 sb_writers#4 &sb->s_type->i_mutex_key#8 mapping.invalidate_lock jbd2_handle rcu_read_lock &rcu_state.expedited_wq &p->pi_lock irq_context: 0 sb_writers#4 &sb->s_type->i_mutex_key#8 mapping.invalidate_lock jbd2_handle rcu_read_lock &rcu_state.expedited_wq &p->pi_lock &rq->__lock irq_context: 0 sb_writers#4 &sb->s_type->i_mutex_key#8 mapping.invalidate_lock jbd2_handle rcu_read_lock &rcu_state.expedited_wq &p->pi_lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 &mm->mmap_lock sb_writers#4 rcu_read_lock rcu_node_0 irq_context: 0 &mm->mmap_lock sb_writers#4 rcu_read_lock &rq->__lock irq_context: 0 &mm->mmap_lock sb_writers#4 rcu_read_lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem &net->sctp.addr_wq_lock k-slock-AF_INET6/1 &obj_hash[i].lock pool_lock irq_context: 0 &sb->s_type->i_mutex_key#8 &sbi->s_writepages_rwsem jbd2_handle &ei->i_data_sem quarantine_lock irq_context: 0 (wq_completion)wg-kex-wg2 (work_completion)(&peer->transmit_handshake_work) kfence_freelist_lock irq_context: 0 (wq_completion)wg-kex-wg2 (work_completion)(&peer->transmit_handshake_work) &peer->endpoint_lock rcu_read_lock_bh &meta->lock irq_context: 0 (wq_completion)wg-crypt-wg1#2 (work_completion)(&peer->transmit_packet_work) &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)wg-kex-wg2 (work_completion)(&peer->transmit_handshake_work) &peer->endpoint_lock rcu_read_lock_bh kfence_freelist_lock irq_context: softirq (&ndev->rs_timer) rcu_read_lock rcu_read_lock rcu_read_lock_bh rcu_read_lock_bh rcu_read_lock &pool->lock/1 irq_context: softirq (&ndev->rs_timer) rcu_read_lock rcu_read_lock rcu_read_lock_bh rcu_read_lock_bh rcu_read_lock &pool->lock/1 &obj_hash[i].lock irq_context: softirq (&ndev->rs_timer) rcu_read_lock rcu_read_lock rcu_read_lock_bh rcu_read_lock_bh rcu_read_lock &pool->lock/1 pool_lock#2 irq_context: softirq (&ndev->rs_timer) rcu_read_lock rcu_read_lock rcu_read_lock_bh rcu_read_lock_bh rcu_read_lock &pool->lock/1 &p->pi_lock irq_context: softirq (&ndev->rs_timer) rcu_read_lock rcu_read_lock rcu_read_lock_bh rcu_read_lock_bh rcu_read_lock &pool->lock/1 &p->pi_lock &rq->__lock irq_context: softirq (&ndev->rs_timer) rcu_read_lock rcu_read_lock rcu_read_lock_bh rcu_read_lock_bh rcu_read_lock &pool->lock/1 &p->pi_lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 &nft_net->commit_mutex &cfs_rq->removed.lock irq_context: 0 &nft_net->commit_mutex rcu_state.exp_mutex &rq->__lock &cfs_rq->removed.lock irq_context: softirq &(&bond->mii_work)->timer rcu_read_lock &pool->lock/1 &p->pi_lock &rq->__lock &cfs_rq->removed.lock irq_context: 0 dup_mmap_sem &mm->mmap_lock &mm->mmap_lock/1 &anon_vma->rwsem rcu_read_lock &rq->__lock &cfs_rq->removed.lock irq_context: 0 &sbi->s_writepages_rwsem jbd2_handle &ei->i_data_sem &lg->lg_mutex &c->lock irq_context: 0 &sbi->s_writepages_rwsem jbd2_handle &ei->i_data_sem &lg->lg_mutex &rq->__lock irq_context: 0 &sbi->s_writepages_rwsem jbd2_handle &ei->i_data_sem &lg->lg_mutex &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 cb_lock genl_mutex rtnl_mutex &rdev->wiphy.mtx gdp_mutex remove_cache_srcu irq_context: 0 cb_lock genl_mutex rtnl_mutex &rdev->wiphy.mtx gdp_mutex remove_cache_srcu quarantine_lock irq_context: 0 cb_lock genl_mutex rtnl_mutex &rdev->wiphy.mtx gdp_mutex remove_cache_srcu &c->lock irq_context: 0 cb_lock genl_mutex rtnl_mutex &rdev->wiphy.mtx gdp_mutex remove_cache_srcu &n->list_lock irq_context: 0 cb_lock genl_mutex rtnl_mutex &rdev->wiphy.mtx gdp_mutex remove_cache_srcu &rq->__lock irq_context: 0 cb_lock genl_mutex rtnl_mutex &rdev->wiphy.mtx gdp_mutex remove_cache_srcu rcu_read_lock pool_lock#2 irq_context: 0 cb_lock genl_mutex rtnl_mutex &rdev->wiphy.mtx gdp_mutex remove_cache_srcu &obj_hash[i].lock irq_context: 0 cb_lock genl_mutex rtnl_mutex &rdev->wiphy.mtx gdp_mutex remove_cache_srcu rcu_read_lock rcu_node_0 irq_context: 0 cb_lock genl_mutex rtnl_mutex &rdev->wiphy.mtx gdp_mutex remove_cache_srcu rcu_read_lock &rq->__lock irq_context: 0 cb_lock genl_mutex rtnl_mutex &rdev->wiphy.mtx gdp_mutex remove_cache_srcu rcu_read_lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 cb_lock genl_mutex rtnl_mutex &rdev->wiphy.mtx &sb->s_type->i_mutex_key#3 &xa->xa_lock#4 irq_context: 0 cb_lock genl_mutex rtnl_mutex &rdev->wiphy.mtx &sb->s_type->i_mutex_key#3 &xa->xa_lock#4 &c->lock irq_context: 0 cb_lock genl_mutex rtnl_mutex &rdev->wiphy.mtx &sb->s_type->i_mutex_key#3 &xa->xa_lock#4 pool_lock#2 irq_context: 0 cb_lock genl_mutex rtnl_mutex &rdev->wiphy.mtx &sb->s_type->i_mutex_key#3 stock_lock irq_context: 0 cb_lock genl_mutex rtnl_mutex &rdev->wiphy.mtx &sb->s_type->i_mutex_key#3 fs_reclaim mmu_notifier_invalidate_range_start &rq->__lock irq_context: 0 cb_lock genl_mutex rtnl_mutex &rdev->wiphy.mtx &sb->s_type->i_mutex_key#3 fs_reclaim mmu_notifier_invalidate_range_start &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 cb_lock genl_mutex rtnl_mutex &rdev->wiphy.mtx &sb->s_type->i_mutex_key#3 &n->list_lock &c->lock irq_context: 0 cb_lock genl_mutex rfkill_global_mutex &sem->wait_lock irq_context: 0 cb_lock genl_mutex rfkill_global_mutex &p->pi_lock irq_context: 0 cb_lock genl_mutex rfkill_global_mutex &root->kernfs_rwsem &sem->wait_lock irq_context: 0 cb_lock genl_mutex &sb->s_type->i_mutex_key#3 stock_lock irq_context: 0 cb_lock genl_mutex &sb->s_type->i_mutex_key#3 &n->list_lock irq_context: 0 cb_lock genl_mutex &sb->s_type->i_mutex_key#3 &n->list_lock &c->lock irq_context: 0 cb_lock genl_mutex &sb->s_type->i_mutex_key#3 rcu_read_lock rcu_read_lock rcu_node_0 irq_context: 0 cb_lock genl_mutex &sb->s_type->i_mutex_key#3 rcu_read_lock rcu_read_lock &rq->__lock irq_context: 0 cb_lock genl_mutex &sb->s_type->i_mutex_key#3 rcu_read_lock rcu_read_lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 cb_lock genl_mutex &sb->s_type->i_mutex_key#3 rcu_read_lock rcu_node_0 irq_context: 0 cb_lock genl_mutex rtnl_mutex &rdev->wiphy.mtx stock_lock irq_context: 0 cb_lock genl_mutex rtnl_mutex &rdev->wiphy.mtx &sb->s_type->i_mutex_key#3 rcu_read_lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 cb_lock genl_mutex rtnl_mutex &rdev->wiphy.mtx &n->list_lock irq_context: 0 cb_lock genl_mutex rtnl_mutex &rdev->wiphy.mtx &n->list_lock &c->lock irq_context: 0 cb_lock genl_mutex rtnl_mutex &rdev->wiphy.mtx &root->kernfs_rwsem &sem->wait_lock irq_context: 0 cb_lock genl_mutex rtnl_mutex &rdev->wiphy.mtx &root->kernfs_rwsem &rq->__lock irq_context: 0 cb_lock genl_mutex rtnl_mutex &rdev->wiphy.mtx &root->kernfs_rwsem &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 cb_lock genl_mutex rtnl_mutex &rdev->wiphy.mtx uevent_sock_mutex &rq->__lock irq_context: 0 cb_lock genl_mutex rtnl_mutex &rdev->wiphy.mtx uevent_sock_mutex &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 cb_lock genl_mutex rtnl_mutex &rdev->wiphy.mtx &idev->mc_lock &n->list_lock irq_context: 0 cb_lock genl_mutex rtnl_mutex &rdev->wiphy.mtx &idev->mc_lock &n->list_lock &c->lock irq_context: 0 &mm->mmap_lock sb_pagefaults mapping.invalidate_lock &____s->seqcount irq_context: 0 &f->f_pos_lock sb_writers#4 &sb->s_type->i_mutex_key#8 &sbi->s_writepages_rwsem &mapping->i_mmap_rwsem mmu_notifier_invalidate_range_start irq_context: 0 &f->f_pos_lock sb_writers#4 &sb->s_type->i_mutex_key#8 &sbi->s_writepages_rwsem &mapping->i_mmap_rwsem mmu_notifier_invalidate_range_start &rq->__lock irq_context: 0 &f->f_pos_lock sb_writers#4 &sb->s_type->i_mutex_key#8 &sbi->s_writepages_rwsem &mapping->i_mmap_rwsem mmu_notifier_invalidate_range_start &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 &f->f_pos_lock sb_writers#4 &sb->s_type->i_mutex_key#8 &sbi->s_writepages_rwsem &mapping->i_mmap_rwsem ptlock_ptr(page)#2 irq_context: 0 &f->f_pos_lock sb_writers#4 &sb->s_type->i_mutex_key#8 &sbi->s_writepages_rwsem &mapping->private_lock irq_context: 0 &f->f_pos_lock sb_writers#4 &sb->s_type->i_mutex_key#8 &sbi->s_writepages_rwsem &mapping->private_lock rcu_read_lock &memcg->move_lock irq_context: 0 &f->f_pos_lock sb_writers#4 &sb->s_type->i_mutex_key#8 &sbi->s_writepages_rwsem rcu_read_lock &memcg->move_lock irq_context: 0 &f->f_pos_lock sb_writers#4 &sb->s_type->i_mutex_key#8 &sbi->s_writepages_rwsem rcu_read_lock &xa->xa_lock#9 irq_context: 0 &f->f_pos_lock sb_writers#4 &sb->s_type->i_mutex_key#8 &sbi->s_writepages_rwsem rcu_read_lock &xa->xa_lock#9 &s->s_inode_wblist_lock irq_context: 0 &f->f_pos_lock sb_writers#4 &sb->s_type->i_mutex_key#8 &mapping->i_mmap_rwsem irq_context: 0 &f->f_pos_lock sb_writers#4 &sb->s_type->i_mutex_key#8 &mapping->i_mmap_rwsem lock#4 irq_context: 0 &f->f_pos_lock sb_writers#4 &sb->s_type->i_mutex_key#8 &mapping->i_mmap_rwsem lock#5 irq_context: 0 &f->f_pos_lock sb_writers#4 &sb->s_type->i_mutex_key#8 &mapping->i_mmap_rwsem mmu_notifier_invalidate_range_start irq_context: 0 &f->f_pos_lock sb_writers#4 &sb->s_type->i_mutex_key#8 &mapping->i_mmap_rwsem ptlock_ptr(page)#2 irq_context: 0 cb_lock genl_mutex batched_entropy_u32.lock crngs.lock irq_context: 0 cb_lock genl_mutex &sem->wait_lock irq_context: 0 cb_lock genl_mutex &root->kernfs_rwsem &sem->wait_lock irq_context: 0 (wq_completion)ext4-rsv-conversion (work_completion)(&ei->i_rsv_conversion_work) jbd2_handle &ei->i_data_sem mmu_notifier_invalidate_range_start &rq->__lock irq_context: 0 (wq_completion)ext4-rsv-conversion (work_completion)(&ei->i_rsv_conversion_work) jbd2_handle &ei->i_data_sem mmu_notifier_invalidate_range_start &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 cb_lock genl_mutex rtnl_mutex &rdev->wiphy.mtx &sem->wait_lock irq_context: 0 cb_lock genl_mutex rtnl_mutex &rdev->wiphy.mtx &p->pi_lock irq_context: 0 cb_lock genl_mutex rtnl_mutex &rdev->wiphy.mtx &p->pi_lock &rq->__lock irq_context: 0 cb_lock genl_mutex rtnl_mutex &rdev->wiphy.mtx &p->pi_lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 &mm->mmap_lock sb_writers#4 &cfs_rq->removed.lock irq_context: 0 cb_lock genl_mutex rfkill_global_mutex &root->kernfs_rwsem &sem->wait_lock irq_context: 0 cb_lock genl_mutex rfkill_global_mutex &root->kernfs_rwsem &rq->__lock irq_context: 0 cb_lock genl_mutex rfkill_global_mutex &root->kernfs_rwsem &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 cb_lock genl_mutex rtnl_mutex &rdev->wiphy.mtx rcu_read_lock rcu_node_0 irq_context: 0 cb_lock genl_mutex &sb->s_type->i_mutex_key#3 &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)wg-kex-wg1#2 (work_completion)(&peer->transmit_handshake_work) &wg->static_identity.lock &handshake->lock rcu_read_lock_bh batched_entropy_u32.lock crngs.lock irq_context: 0 (wq_completion)bond4 (work_completion)(&(&bond->alb_work)->work) fill_pool_map-wait-type-override &c->lock irq_context: 0 (wq_completion)bond4 (work_completion)(&(&bond->alb_work)->work) fill_pool_map-wait-type-override pool_lock#2 irq_context: 0 (wq_completion)bond4 (work_completion)(&(&bond->alb_work)->work) fill_pool_map-wait-type-override &____s->seqcount#2 irq_context: 0 (wq_completion)bond4 (work_completion)(&(&bond->alb_work)->work) fill_pool_map-wait-type-override &____s->seqcount irq_context: 0 (wq_completion)bond4 (work_completion)(&(&bond->alb_work)->work) fill_pool_map-wait-type-override pool_lock irq_context: 0 (wq_completion)kblockd (work_completion)(&(&q->requeue_work)->work) rcu_read_lock fill_pool_map-wait-type-override pool_lock#2 irq_context: 0 (wq_completion)kblockd (work_completion)(&(&q->requeue_work)->work) rcu_read_lock fill_pool_map-wait-type-override &c->lock irq_context: 0 (wq_completion)kblockd (work_completion)(&(&q->requeue_work)->work) rcu_read_lock fill_pool_map-wait-type-override &____s->seqcount#2 irq_context: 0 (wq_completion)kblockd (work_completion)(&(&q->requeue_work)->work) rcu_read_lock fill_pool_map-wait-type-override &____s->seqcount irq_context: 0 (wq_completion)kblockd (work_completion)(&(&q->requeue_work)->work) rcu_read_lock fill_pool_map-wait-type-override pool_lock irq_context: 0 &f->f_pos_lock sb_writers#4 &sb->s_type->i_mutex_key#8 jbd2_handle &ei->i_data_sem &ei->i_prealloc_lock irq_context: 0 cb_lock genl_mutex uevent_sock_mutex &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 &f->f_pos_lock sb_writers#4 &sb->s_type->i_mutex_key#8 jbd2_handle &ei->i_data_sem &____s->seqcount#2 irq_context: 0 &mm->mmap_lock &anon_vma->rwsem fill_pool_map-wait-type-override pool_lock irq_context: 0 &mm->mmap_lock &anon_vma->rwsem fill_pool_map-wait-type-override &rq->__lock irq_context: 0 &mm->mmap_lock &anon_vma->rwsem fill_pool_map-wait-type-override &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 cb_lock genl_mutex rtnl_mutex &rdev->wiphy.mtx &sb->s_type->i_mutex_key#3 rcu_read_lock &rcu_state.gp_wq irq_context: 0 cb_lock genl_mutex rtnl_mutex &rdev->wiphy.mtx &sb->s_type->i_mutex_key#3 rcu_read_lock &rcu_state.gp_wq &p->pi_lock irq_context: 0 cb_lock genl_mutex rtnl_mutex &rdev->wiphy.mtx &sb->s_type->i_mutex_key#3 rcu_read_lock &rcu_state.gp_wq &p->pi_lock &rq->__lock irq_context: 0 cb_lock genl_mutex rtnl_mutex &rdev->wiphy.mtx &sb->s_type->i_mutex_key#3 rcu_read_lock &rcu_state.gp_wq &p->pi_lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 &f->f_pos_lock sb_writers#4 &sb->s_type->i_mutex_key#8 mapping.invalidate_lock &sem->wait_lock irq_context: 0 &mm->mmap_lock sb_pagefaults mapping.invalidate_lock &sem->wait_lock irq_context: 0 &mm->mmap_lock sb_pagefaults &sem->wait_lock irq_context: 0 &mm->mmap_lock sb_pagefaults &p->pi_lock irq_context: 0 cb_lock genl_mutex rfkill_global_mutex &n->list_lock irq_context: 0 cb_lock genl_mutex rfkill_global_mutex &n->list_lock &c->lock irq_context: 0 &f->f_pos_lock sb_writers#4 &sb->s_type->i_mutex_key#8 rcu_read_lock fill_pool_map-wait-type-override pool_lock#2 irq_context: 0 &f->f_pos_lock sb_writers#4 &sb->s_type->i_mutex_key#8 rcu_read_lock fill_pool_map-wait-type-override pool_lock irq_context: 0 sk_lock-AF_INET6 rcu_read_lock rcu_read_lock_bh quarantine_lock irq_context: 0 cb_lock genl_mutex rfkill_global_mutex uevent_sock_mutex &n->list_lock irq_context: 0 cb_lock genl_mutex rfkill_global_mutex uevent_sock_mutex &n->list_lock &c->lock irq_context: 0 cb_lock genl_mutex rtnl_mutex &rdev->wiphy.mtx &sb->s_type->i_mutex_key#3 rcu_read_lock rcu_read_lock rcu_node_0 irq_context: 0 cb_lock genl_mutex rtnl_mutex &rdev->wiphy.mtx &sb->s_type->i_mutex_key#3 rcu_read_lock rcu_read_lock &rq->__lock irq_context: 0 cb_lock genl_mutex rtnl_mutex &rdev->wiphy.mtx &sb->s_type->i_mutex_key#3 rcu_read_lock rcu_read_lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 sb_writers#5 fill_pool_map-wait-type-override &____s->seqcount#2 irq_context: 0 sb_writers#5 fill_pool_map-wait-type-override &____s->seqcount irq_context: 0 cb_lock genl_mutex rtnl_mutex &rdev->wiphy.mtx fs_reclaim &rq->__lock irq_context: 0 cb_lock genl_mutex rtnl_mutex &rdev->wiphy.mtx fs_reclaim &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 &sb->s_type->i_mutex_key#10 &smc->clcsock_release_lock fill_pool_map-wait-type-override pool_lock#2 irq_context: 0 &sb->s_type->i_mutex_key#10 &smc->clcsock_release_lock fill_pool_map-wait-type-override &c->lock irq_context: 0 &sb->s_type->i_mutex_key#10 &smc->clcsock_release_lock fill_pool_map-wait-type-override &____s->seqcount#2 irq_context: 0 &sb->s_type->i_mutex_key#10 &smc->clcsock_release_lock fill_pool_map-wait-type-override &____s->seqcount irq_context: 0 &sb->s_type->i_mutex_key#10 &smc->clcsock_release_lock fill_pool_map-wait-type-override pool_lock irq_context: 0 &mm->mmap_lock rcu_read_lock rcu_read_lock fill_pool_map-wait-type-override pool_lock#2 irq_context: 0 &mm->mmap_lock rcu_read_lock rcu_read_lock fill_pool_map-wait-type-override pool_lock irq_context: 0 (wq_completion)events_power_efficient (gc_work).work fill_pool_map-wait-type-override &c->lock irq_context: 0 (wq_completion)events_power_efficient (gc_work).work fill_pool_map-wait-type-override &____s->seqcount#2 irq_context: 0 (wq_completion)events_power_efficient (gc_work).work fill_pool_map-wait-type-override &____s->seqcount irq_context: 0 mapping.invalidate_lock &xa->xa_lock#9 &c->lock irq_context: 0 mapping.invalidate_lock &c->lock irq_context: 0 mapping.invalidate_lock rcu_read_lock rcu_node_0 irq_context: 0 mapping.invalidate_lock rcu_read_lock &rq->__lock irq_context: 0 mapping.invalidate_lock rcu_read_lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 &f->f_pos_lock sb_writers#4 &sb->s_type->i_mutex_key#8 &mapping->i_mmap_rwsem ptlock_ptr(page)#2 &____s->seqcount irq_context: 0 &f->f_pos_lock sb_writers#4 &sb->s_type->i_mutex_key#8 &mapping->i_mmap_rwsem ptlock_ptr(page)#2 pool_lock#2 irq_context: 0 &f->f_pos_lock sb_writers#4 &sb->s_type->i_mutex_key#8 &mapping->i_mmap_rwsem rcu_read_lock pool_lock#2 irq_context: 0 &f->f_pos_lock sb_writers#4 &sb->s_type->i_mutex_key#8 &mapping->i_mmap_rwsem rcu_read_lock rcu_node_0 irq_context: 0 &f->f_pos_lock sb_writers#4 &sb->s_type->i_mutex_key#8 &mapping->i_mmap_rwsem rcu_read_lock &rq->__lock irq_context: 0 &f->f_pos_lock sb_writers#4 &sb->s_type->i_mutex_key#8 &mapping->i_mmap_rwsem rcu_read_lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 &f->f_pos_lock sb_writers#4 &sb->s_type->i_mutex_key#8 &mapping->i_mmap_rwsem &obj_hash[i].lock irq_context: 0 &f->f_pos_lock sb_writers#4 &sb->s_type->i_mutex_key#8 jbd2_handle &ei->i_data_sem &ei->i_es_lock &____s->seqcount#2 irq_context: 0 &f->f_pos_lock sb_writers#4 &sb->s_type->i_mutex_key#8 jbd2_handle &ei->i_data_sem &ei->i_es_lock &____s->seqcount irq_context: 0 pernet_ops_rwsem rtnl_mutex fs_reclaim &rq->__lock &cfs_rq->removed.lock irq_context: 0 cb_lock genl_mutex uevent_sock_mutex rcu_read_lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 cb_lock genl_mutex &sb->s_type->i_mutex_key#3 rcu_read_lock &rq->__lock irq_context: 0 cb_lock genl_mutex &sb->s_type->i_mutex_key#3 rcu_read_lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 &mm->mmap_lock batched_entropy_u32.lock crngs.lock irq_context: 0 sb_writers#4 &type->i_mutex_dir_key#3 tomoyo_ss &rcu_state.expedited_wq irq_context: 0 (wq_completion)ext4-rsv-conversion (work_completion)(&ei->i_rsv_conversion_work) jbd2_handle &rcu_state.expedited_wq irq_context: 0 sb_writers#4 &type->i_mutex_dir_key#3 tomoyo_ss &rcu_state.expedited_wq &p->pi_lock irq_context: 0 sb_writers#4 &type->i_mutex_dir_key#3 tomoyo_ss &rcu_state.expedited_wq &p->pi_lock &rq->__lock irq_context: 0 sb_writers#4 &type->i_mutex_dir_key#3 tomoyo_ss &rcu_state.expedited_wq &p->pi_lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)ext4-rsv-conversion &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 kn->active#60 &kernfs_locks->open_file_mutex[count] &n->list_lock irq_context: 0 kn->active#60 &kernfs_locks->open_file_mutex[count] &n->list_lock &c->lock irq_context: 0 kn->active#60 &kernfs_locks->open_file_mutex[count] &rq->__lock irq_context: 0 kn->active#60 &kernfs_locks->open_file_mutex[count] &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 &sb->s_type->i_mutex_key#10 sk_lock-AF_INET rcu_read_lock rcu_read_lock rcu_node_0 irq_context: 0 &sb->s_type->i_mutex_key#10 sk_lock-AF_INET rcu_read_lock rcu_read_lock &rq->__lock irq_context: 0 &sb->s_type->i_mutex_key#10 sk_lock-AF_INET rcu_read_lock rcu_read_lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 &sb->s_type->i_mutex_key#10 sk_lock-AF_INET rcu_read_lock rcu_node_0 irq_context: 0 kn->active#60 &____s->seqcount#2 irq_context: 0 kn->active#60 &____s->seqcount irq_context: 0 pernet_ops_rwsem remove_cache_srcu rcu_read_lock &rcu_state.gp_wq irq_context: 0 pernet_ops_rwsem remove_cache_srcu rcu_read_lock &rcu_state.gp_wq &p->pi_lock irq_context: 0 pernet_ops_rwsem remove_cache_srcu rcu_read_lock &rcu_state.gp_wq &p->pi_lock &rq->__lock irq_context: 0 pernet_ops_rwsem remove_cache_srcu rcu_read_lock &rcu_state.gp_wq &p->pi_lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 &f->f_pos_lock sb_writers#11 &mm->mmap_lock &rq->__lock irq_context: 0 &f->f_pos_lock sb_writers#11 &mm->mmap_lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)wg-kex-wg1#2 (work_completion)(&peer->transmit_handshake_work) &wg->static_identity.lock &handshake->lock &rq->__lock irq_context: 0 (wq_completion)wg-kex-wg1#2 (work_completion)(&peer->transmit_handshake_work) &wg->static_identity.lock &handshake->lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: softirq (&icsk->icsk_retransmit_timer) slock-AF_INET rcu_read_lock rcu_read_lock rcu_read_lock_bh rcu_read_lock pool_lock#2 irq_context: softirq (&icsk->icsk_retransmit_timer) slock-AF_INET rcu_read_lock rcu_read_lock rcu_read_lock_bh rcu_read_lock tk_core.seq.seqcount irq_context: softirq (&icsk->icsk_retransmit_timer) slock-AF_INET rcu_read_lock rcu_read_lock rcu_read_lock_bh rcu_read_lock &obj_hash[i].lock irq_context: softirq rcu_read_lock rcu_read_lock slock-AF_INET/1 rcu_read_lock rcu_read_lock &c->lock irq_context: softirq rcu_read_lock rcu_read_lock slock-AF_INET/1 rcu_read_lock rcu_read_lock &n->list_lock irq_context: softirq rcu_read_lock rcu_read_lock slock-AF_INET/1 rcu_read_lock rcu_read_lock &n->list_lock &c->lock irq_context: softirq rcu_read_lock rcu_read_lock slock-AF_INET/1 rcu_read_lock rcu_read_lock rcu_read_lock_bh rcu_read_lock pool_lock#2 irq_context: softirq rcu_read_lock rcu_read_lock slock-AF_INET/1 rcu_read_lock rcu_read_lock rcu_read_lock_bh rcu_read_lock tk_core.seq.seqcount irq_context: softirq rcu_read_lock rcu_read_lock slock-AF_INET/1 rcu_read_lock rcu_read_lock rcu_read_lock_bh rcu_read_lock &obj_hash[i].lock irq_context: 0 (wq_completion)bat_events (work_completion)(&(&forw_packet_aggr->delayed_work)->work) &hard_iface->bat_iv.ogm_buff_mutex &bat_priv->forw_bat_list_lock fill_pool_map-wait-type-override pool_lock#2 irq_context: 0 (wq_completion)bat_events (work_completion)(&(&forw_packet_aggr->delayed_work)->work) &hard_iface->bat_iv.ogm_buff_mutex &bat_priv->forw_bat_list_lock fill_pool_map-wait-type-override &c->lock irq_context: 0 (wq_completion)bat_events (work_completion)(&(&forw_packet_aggr->delayed_work)->work) &hard_iface->bat_iv.ogm_buff_mutex &bat_priv->forw_bat_list_lock fill_pool_map-wait-type-override &n->list_lock irq_context: 0 (wq_completion)bat_events (work_completion)(&(&forw_packet_aggr->delayed_work)->work) &hard_iface->bat_iv.ogm_buff_mutex &bat_priv->forw_bat_list_lock fill_pool_map-wait-type-override &n->list_lock &c->lock irq_context: 0 (wq_completion)bat_events (work_completion)(&(&forw_packet_aggr->delayed_work)->work) &hard_iface->bat_iv.ogm_buff_mutex &bat_priv->forw_bat_list_lock fill_pool_map-wait-type-override &____s->seqcount#2 irq_context: 0 (wq_completion)bat_events (work_completion)(&(&forw_packet_aggr->delayed_work)->work) &hard_iface->bat_iv.ogm_buff_mutex &bat_priv->forw_bat_list_lock fill_pool_map-wait-type-override &____s->seqcount irq_context: 0 (wq_completion)bat_events (work_completion)(&(&forw_packet_aggr->delayed_work)->work) &hard_iface->bat_iv.ogm_buff_mutex &bat_priv->forw_bat_list_lock fill_pool_map-wait-type-override pool_lock irq_context: 0 (wq_completion)mld (work_completion)(&(&idev->mc_ifc_work)->work) &idev->mc_lock rcu_read_lock rcu_read_lock rcu_read_lock_bh dev->qdisc_tx_busylock ?: &qdisc_tx_busylock _xmit_ETHER#2 rcu_read_lock pool_lock#2 irq_context: 0 (wq_completion)mld (work_completion)(&(&idev->mc_ifc_work)->work) &idev->mc_lock rcu_read_lock rcu_read_lock rcu_read_lock_bh dev->qdisc_tx_busylock ?: &qdisc_tx_busylock _xmit_ETHER#2 rcu_read_lock tk_core.seq.seqcount irq_context: 0 (wq_completion)mld (work_completion)(&(&idev->mc_ifc_work)->work) &idev->mc_lock rcu_read_lock rcu_read_lock rcu_read_lock_bh dev->qdisc_tx_busylock ?: &qdisc_tx_busylock _xmit_ETHER#2 rcu_read_lock rlock-AF_PACKET irq_context: 0 (wq_completion)mld (work_completion)(&(&idev->mc_ifc_work)->work) &idev->mc_lock rcu_read_lock rcu_read_lock rcu_read_lock_bh dev->qdisc_tx_busylock ?: &qdisc_tx_busylock _xmit_ETHER#2 rcu_read_lock &obj_hash[i].lock irq_context: 0 namespace_sem namespace_sem.wait_lock irq_context: 0 namespace_sem.wait_lock irq_context: 0 namespace_sem pcpu_alloc_mutex.wait_lock irq_context: 0 namespace_sem &p->pi_lock irq_context: 0 namespace_sem &p->pi_lock &rq->__lock irq_context: 0 namespace_sem &p->pi_lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)events (work_completion)(&(&hwstats->traffic_dw)->work) fill_pool_map-wait-type-override pool_lock irq_context: 0 &sb->s_type->i_mutex_key#10 sk_lock-AF_INET slock-AF_INET &meta->lock irq_context: 0 &sb->s_type->i_mutex_key#10 sk_lock-AF_INET slock-AF_INET kfence_freelist_lock irq_context: 0 pernet_ops_rwsem devices_rwsem rdma_nets_rwsem &device->compat_devs_mutex uevent_sock_mutex remove_cache_srcu irq_context: 0 pernet_ops_rwsem devices_rwsem rdma_nets_rwsem &device->compat_devs_mutex uevent_sock_mutex remove_cache_srcu quarantine_lock irq_context: 0 pernet_ops_rwsem devices_rwsem rdma_nets_rwsem &device->compat_devs_mutex uevent_sock_mutex remove_cache_srcu &c->lock irq_context: 0 pernet_ops_rwsem devices_rwsem rdma_nets_rwsem &device->compat_devs_mutex uevent_sock_mutex remove_cache_srcu &n->list_lock irq_context: 0 pernet_ops_rwsem devices_rwsem rdma_nets_rwsem &device->compat_devs_mutex uevent_sock_mutex remove_cache_srcu rcu_read_lock pool_lock#2 irq_context: 0 pernet_ops_rwsem devices_rwsem rdma_nets_rwsem &device->compat_devs_mutex uevent_sock_mutex remove_cache_srcu &obj_hash[i].lock irq_context: 0 pernet_ops_rwsem devices_rwsem rdma_nets_rwsem &device->compat_devs_mutex uevent_sock_mutex remove_cache_srcu rcu_read_lock rcu_node_0 irq_context: 0 pernet_ops_rwsem devices_rwsem rdma_nets_rwsem &device->compat_devs_mutex uevent_sock_mutex remove_cache_srcu rcu_read_lock &rq->__lock irq_context: 0 pernet_ops_rwsem devices_rwsem rdma_nets_rwsem &device->compat_devs_mutex uevent_sock_mutex remove_cache_srcu rcu_read_lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 pernet_ops_rwsem devices_rwsem rdma_nets_rwsem &device->compat_devs_mutex uevent_sock_mutex remove_cache_srcu rcu_read_lock &rcu_state.gp_wq irq_context: 0 pernet_ops_rwsem devices_rwsem rdma_nets_rwsem &device->compat_devs_mutex uevent_sock_mutex remove_cache_srcu rcu_read_lock &rcu_state.gp_wq &p->pi_lock irq_context: 0 pernet_ops_rwsem devices_rwsem rdma_nets_rwsem &device->compat_devs_mutex uevent_sock_mutex remove_cache_srcu rcu_read_lock &rcu_state.gp_wq &p->pi_lock &rq->__lock irq_context: 0 pernet_ops_rwsem devices_rwsem rdma_nets_rwsem &device->compat_devs_mutex uevent_sock_mutex remove_cache_srcu rcu_read_lock &rcu_state.gp_wq &p->pi_lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 pernet_ops_rwsem devices_rwsem rdma_nets_rwsem &device->compat_devs_mutex uevent_sock_mutex remove_cache_srcu pool_lock#2 irq_context: 0 (wq_completion)events_unbound (work_completion)(&map->work) &htab->buckets[i].lock irq_context: 0 (wq_completion)events_unbound (work_completion)(&map->work) &rnp->exp_wq[0] irq_context: 0 (wq_completion)wg-crypt-wg2#2 (work_completion)(&({ do { const void *__vpp_verify = (typeof((worker) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __asm__ ("" : "=r"(__ptr) : "0"((typeof(*((worker))) *)((worker)))); (typeof((typeof(*((worker))) *)((worker)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); })->work) rcu_node_0 irq_context: 0 (wq_completion)events_unbound (work_completion)(&map->work) rcu_state.exp_mutex rcu_read_lock &pool->lock &p->pi_lock &rq->__lock &cfs_rq->removed.lock irq_context: 0 (wq_completion)mld (work_completion)(&(&idev->mc_ifc_work)->work) &idev->mc_lock rcu_read_lock rcu_read_lock rcu_read_lock_bh dev->qdisc_tx_busylock ?: &qdisc_tx_busylock _xmit_ETHER#2 &____s->seqcount irq_context: 0 pernet_ops_rwsem devices_rwsem rdma_nets_rwsem &device->compat_devs_mutex &root->kernfs_rwsem pgd_lock irq_context: 0 pernet_ops_rwsem devices_rwsem rdma_nets_rwsem &device->compat_devs_mutex &root->kernfs_rwsem stock_lock irq_context: 0 pernet_ops_rwsem devices_rwsem rdma_nets_rwsem &device->compat_devs_mutex &root->kernfs_rwsem key irq_context: 0 pernet_ops_rwsem devices_rwsem rdma_nets_rwsem &device->compat_devs_mutex &root->kernfs_rwsem pcpu_lock irq_context: 0 pernet_ops_rwsem devices_rwsem rdma_nets_rwsem &device->compat_devs_mutex &root->kernfs_rwsem percpu_counters_lock irq_context: 0 pernet_ops_rwsem devices_rwsem rdma_nets_rwsem &device->compat_devs_mutex &root->kernfs_rwsem pcpu_lock stock_lock irq_context: 0 pernet_ops_rwsem devices_rwsem rdma_nets_rwsem &device->compat_devs_mutex &root->kernfs_rwsem &rcu_state.expedited_wq irq_context: 0 pernet_ops_rwsem devices_rwsem rdma_nets_rwsem &device->compat_devs_mutex &root->kernfs_rwsem &rcu_state.expedited_wq &p->pi_lock irq_context: 0 pernet_ops_rwsem devices_rwsem rdma_nets_rwsem &device->compat_devs_mutex &root->kernfs_rwsem &rcu_state.expedited_wq &p->pi_lock &rq->__lock irq_context: 0 pernet_ops_rwsem devices_rwsem rdma_nets_rwsem &device->compat_devs_mutex &root->kernfs_rwsem &rcu_state.expedited_wq &p->pi_lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 pernet_ops_rwsem k-sk_lock-AF_RXRPC &rxnet->local_mutex quarantine_lock irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem rtnl_mutex netpoll_srcu irq_context: 0 pernet_ops_rwsem k-sk_lock-AF_RXRPC remove_cache_srcu irq_context: 0 pernet_ops_rwsem k-sk_lock-AF_RXRPC remove_cache_srcu quarantine_lock irq_context: 0 pernet_ops_rwsem k-sk_lock-AF_RXRPC remove_cache_srcu &c->lock irq_context: 0 pernet_ops_rwsem k-sk_lock-AF_RXRPC remove_cache_srcu &n->list_lock irq_context: 0 pernet_ops_rwsem k-sk_lock-AF_RXRPC remove_cache_srcu &obj_hash[i].lock irq_context: 0 pernet_ops_rwsem rtnl_mutex pcpu_alloc_mutex rcu_node_0 irq_context: 0 pernet_ops_rwsem rtnl_mutex pcpu_alloc_mutex &rcu_state.expedited_wq irq_context: 0 pernet_ops_rwsem rtnl_mutex pcpu_alloc_mutex &rcu_state.expedited_wq &p->pi_lock irq_context: 0 pernet_ops_rwsem rtnl_mutex pcpu_alloc_mutex &rcu_state.expedited_wq &p->pi_lock &rq->__lock irq_context: 0 pernet_ops_rwsem rtnl_mutex pcpu_alloc_mutex &rcu_state.expedited_wq &p->pi_lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem &wq->mutex &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 pernet_ops_rwsem k-sk_lock-AF_RXRPC remove_cache_srcu pool_lock#2 irq_context: 0 pernet_ops_rwsem k-sk_lock-AF_RXRPC remove_cache_srcu rcu_read_lock pool_lock#2 irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem rtnl_mutex &pn->hash_lock irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem rtnl_mutex &dev->tx_global_lock irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem rtnl_mutex &dev->tx_global_lock _xmit_ETHER#2 irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem rtnl_mutex &dev->tx_global_lock &obj_hash[i].lock irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem rtnl_mutex dev->qdisc_tx_busylock ?: &qdisc_tx_busylock irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem rtnl_mutex dev->qdisc_tx_busylock ?: &qdisc_tx_busylock &sch->q.lock irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem rtnl_mutex &sch->q.lock irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem rtnl_mutex rcu_read_lock &ndev->lock irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem rtnl_mutex __ip_vs_mutex irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem rtnl_mutex __ip_vs_mutex &ipvs->dest_trash_lock irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem rtnl_mutex _xmit_ETHER &obj_hash[i].lock irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem rtnl_mutex _xmit_ETHER pool_lock#2 irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem rtnl_mutex _xmit_ETHER krc.lock irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem rtnl_mutex &im->lock irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem rtnl_mutex flowtable_lock irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem rtnl_mutex rcu_read_lock &tb->tb6_lock &net->ipv6.fib6_walker_lock irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem rtnl_mutex rcu_read_lock &tb->tb6_lock &obj_hash[i].lock irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem rtnl_mutex rcu_read_lock &tb->tb6_lock pool_lock#2 irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem rtnl_mutex rcu_read_lock &tb->tb6_lock rt6_exception_lock irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem rtnl_mutex rcu_read_lock &tb->tb6_lock nl_table_lock irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem rtnl_mutex rcu_read_lock &tb->tb6_lock nl_table_wait.lock irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem rtnl_mutex &tbl->lock &n->lock irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem rtnl_mutex &tbl->lock pool_lock#2 irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem rtnl_mutex &tbl->lock &n->lock irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem rtnl_mutex &tbl->lock &n->lock &____s->seqcount#9 irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem rtnl_mutex &tbl->lock nl_table_lock irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem rtnl_mutex &tbl->lock &obj_hash[i].lock irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem rtnl_mutex &tbl->lock nl_table_wait.lock irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem rtnl_mutex &tbl->lock rcu_read_lock lock#8 irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem rtnl_mutex &tbl->lock rcu_read_lock id_table_lock irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem rtnl_mutex &tbl->lock &dir->lock#2 irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem rtnl_mutex &tbl->lock krc.lock irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem rtnl_mutex &tbl->lock &c->lock irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem rtnl_mutex &net->ipv6.addrconf_hash_lock &obj_hash[i].lock irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem rtnl_mutex &ndev->lock &base->lock irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem rtnl_mutex &ndev->lock &base->lock &obj_hash[i].lock irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem rtnl_mutex &ifa->lock irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem rtnl_mutex &idev->mc_lock fs_reclaim irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem rtnl_mutex &idev->mc_lock fs_reclaim mmu_notifier_invalidate_range_start irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem rtnl_mutex &idev->mc_lock &c->lock irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem rtnl_mutex &idev->mc_lock &____s->seqcount#2 irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem rtnl_mutex &idev->mc_lock &____s->seqcount irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem rtnl_mutex &idev->mc_lock &base->lock irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem rtnl_mutex &idev->mc_lock &base->lock &obj_hash[i].lock irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem rtnl_mutex &tb->tb6_lock irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem rtnl_mutex rcu_read_lock &obj_hash[i].lock irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem rtnl_mutex rcu_read_lock rcu_read_lock &pool->lock/1 irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem rtnl_mutex rcu_read_lock rcu_read_lock &pool->lock/1 &obj_hash[i].lock irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem rtnl_mutex rcu_read_lock rcu_read_lock &pool->lock/1 pool_lock#2 irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem rtnl_mutex rcu_read_lock rcu_read_lock &pool->lock/1 &p->pi_lock irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem rtnl_mutex rcu_read_lock rcu_read_lock &pool->lock/1 &p->pi_lock &rq->__lock irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem rtnl_mutex rcu_read_lock rcu_read_lock &pool->lock/1 &p->pi_lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem rtnl_mutex rcu_read_lock &net->sctp.local_addr_lock irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem rtnl_mutex rcu_read_lock &net->sctp.local_addr_lock &net->sctp.addr_wq_lock irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem rtnl_mutex rcu_read_lock &net->sctp.local_addr_lock &net->sctp.addr_wq_lock &c->lock irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem rtnl_mutex rcu_read_lock &net->sctp.local_addr_lock &net->sctp.addr_wq_lock pool_lock#2 irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem rtnl_mutex rcu_read_lock &net->sctp.local_addr_lock &net->sctp.addr_wq_lock &obj_hash[i].lock irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem rtnl_mutex rcu_read_lock &net->sctp.local_addr_lock &net->sctp.addr_wq_lock &base->lock irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem rtnl_mutex rcu_read_lock &net->sctp.local_addr_lock &net->sctp.addr_wq_lock &base->lock &obj_hash[i].lock irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem rtnl_mutex rcu_read_lock krc.lock irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem rtnl_mutex rcu_read_lock &rcu_state.gp_wq irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem rtnl_mutex rcu_read_lock &rcu_state.gp_wq &p->pi_lock irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem rtnl_mutex rcu_read_lock &rcu_state.gp_wq &p->pi_lock &rq->__lock irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem rtnl_mutex rcu_read_lock &rcu_state.gp_wq &p->pi_lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem rtnl_mutex &base->lock &obj_hash[i].lock irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem rtnl_mutex &net->ipv6.fib6_gc_lock rcu_read_lock &tb->tb6_lock irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem rtnl_mutex &net->ipv6.fib6_gc_lock &obj_hash[i].lock irq_context: 0 (wq_completion)events pcpu_balance_work pcpu_alloc_mutex free_vmap_area_lock &obj_hash[i].lock irq_context: 0 (wq_completion)events pcpu_balance_work pcpu_alloc_mutex free_vmap_area_lock pool_lock#2 irq_context: 0 (wq_completion)events pcpu_balance_work pcpu_alloc_mutex rcu_read_lock &rq->__lock irq_context: 0 pernet_ops_rwsem &rcu_state.expedited_wq &p->pi_lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 sb_writers#5 &type->i_mutex_dir_key#5/1 &sb->s_type->i_mutex_key#12 &rcu_state.expedited_wq irq_context: 0 sb_writers#5 &type->i_mutex_dir_key#5/1 &sb->s_type->i_mutex_key#12 &rcu_state.expedited_wq &p->pi_lock irq_context: 0 sb_writers#5 &type->i_mutex_dir_key#5/1 &sb->s_type->i_mutex_key#12 &rcu_state.expedited_wq &p->pi_lock &rq->__lock irq_context: 0 sb_writers#5 &type->i_mutex_dir_key#5/1 &sb->s_type->i_mutex_key#12 &rcu_state.expedited_wq &p->pi_lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)bond5 (work_completion)(&(&bond->mii_work)->work) &rq->__lock irq_context: 0 (wq_completion)bond5 (work_completion)(&(&bond->mii_work)->work) &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem &dir->lock#2 &meta->lock irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem &dir->lock#2 kfence_freelist_lock irq_context: 0 pernet_ops_rwsem devices_rwsem rdma_nets_rwsem &device->compat_devs_mutex remove_cache_srcu rcu_read_lock &rcu_state.expedited_wq irq_context: 0 pernet_ops_rwsem devices_rwsem rdma_nets_rwsem &device->compat_devs_mutex remove_cache_srcu rcu_read_lock &rcu_state.expedited_wq &p->pi_lock irq_context: 0 pernet_ops_rwsem devices_rwsem rdma_nets_rwsem &device->compat_devs_mutex remove_cache_srcu rcu_read_lock &rcu_state.expedited_wq &p->pi_lock &rq->__lock irq_context: 0 pernet_ops_rwsem devices_rwsem rdma_nets_rwsem &device->compat_devs_mutex remove_cache_srcu rcu_read_lock &rcu_state.expedited_wq &p->pi_lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 &sb->s_type->i_mutex_key#8 rcu_read_lock pool_lock#2 irq_context: 0 sock_diag_mutex sock_diag_table_mutex nlk_cb_mutex-SOCK_DIAG inet_diag_table_mutex rcu_read_lock rcu_node_0 irq_context: 0 sock_diag_mutex sock_diag_table_mutex nlk_cb_mutex-SOCK_DIAG inet_diag_table_mutex rcu_read_lock &rq->__lock irq_context: 0 sock_diag_mutex sock_diag_table_mutex nlk_cb_mutex-SOCK_DIAG inet_diag_table_mutex rcu_read_lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 sock_diag_mutex sock_diag_table_mutex nlk_cb_mutex-SOCK_DIAG rcu_read_lock pool_lock#2 irq_context: 0 sock_diag_mutex sock_diag_table_mutex nlk_cb_mutex-SOCK_DIAG rcu_read_lock rcu_read_lock_bh &obj_hash[i].lock irq_context: 0 sock_diag_mutex sock_diag_table_mutex nlk_cb_mutex-SOCK_DIAG rcu_read_lock rcu_read_lock_bh pool_lock#2 irq_context: 0 sock_diag_mutex sock_diag_table_mutex nlk_cb_mutex-SOCK_DIAG rlock-AF_NETLINK irq_context: 0 sock_diag_mutex sock_diag_table_mutex nlk_cb_mutex-SOCK_DIAG &obj_hash[i].lock irq_context: 0 (wq_completion)events (work_completion)(&(&krcp->krw_arr[i].rcu_work)->work) rcu_node_0 irq_context: 0 (wq_completion)events (work_completion)(&(&krcp->krw_arr[i].rcu_work)->work) &rcu_state.expedited_wq irq_context: 0 (wq_completion)events (work_completion)(&(&krcp->krw_arr[i].rcu_work)->work) &rcu_state.expedited_wq &p->pi_lock irq_context: 0 (wq_completion)events (work_completion)(&(&krcp->krw_arr[i].rcu_work)->work) &rcu_state.expedited_wq &p->pi_lock &rq->__lock irq_context: 0 (wq_completion)events (work_completion)(&(&krcp->krw_arr[i].rcu_work)->work) &rcu_state.expedited_wq &p->pi_lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 rtnl_mutex &br->multicast_lock rcu_read_lock rhashtable_bucket irq_context: 0 rtnl_mutex &br->multicast_lock rcu_read_lock rcu_read_lock &pool->lock irq_context: 0 rtnl_mutex &br->multicast_lock rcu_read_lock rcu_read_lock &pool->lock &obj_hash[i].lock irq_context: 0 rtnl_mutex &br->multicast_lock rcu_read_lock rcu_read_lock &pool->lock pool_lock#2 irq_context: 0 rtnl_mutex &br->multicast_lock rcu_read_lock rcu_read_lock &pool->lock &p->pi_lock irq_context: 0 rtnl_mutex &br->multicast_lock rcu_read_lock rcu_read_lock &pool->lock &p->pi_lock &rq->__lock irq_context: 0 rtnl_mutex &br->multicast_lock rcu_read_lock rcu_read_lock &pool->lock &p->pi_lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 rtnl_mutex &br->multicast_lock rcu_read_lock &pool->lock irq_context: 0 rtnl_mutex &br->multicast_lock rcu_read_lock &pool->lock &obj_hash[i].lock irq_context: 0 rtnl_mutex &br->multicast_lock rcu_read_lock &pool->lock pool_lock#2 irq_context: 0 rtnl_mutex &br->multicast_lock rcu_read_lock &pool->lock &p->pi_lock irq_context: 0 rtnl_mutex (&mp->timer) irq_context: 0 pernet_ops_rwsem rtnl_mutex uevent_sock_mutex remove_cache_srcu rcu_read_lock rcu_node_0 irq_context: 0 pernet_ops_rwsem rtnl_mutex uevent_sock_mutex remove_cache_srcu rcu_read_lock &rq->__lock irq_context: 0 pernet_ops_rwsem rtnl_mutex uevent_sock_mutex remove_cache_srcu rcu_read_lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 &sb->s_type->i_mutex_key#10 sk_lock-AF_INET6 &rcu_state.expedited_wq irq_context: 0 &sb->s_type->i_mutex_key#10 sk_lock-AF_INET6 &rcu_state.expedited_wq &p->pi_lock irq_context: 0 &sb->s_type->i_mutex_key#10 sk_lock-AF_INET6 &rcu_state.expedited_wq &p->pi_lock &rq->__lock irq_context: 0 &sb->s_type->i_mutex_key#10 sk_lock-AF_INET6 &rcu_state.expedited_wq &p->pi_lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem rtnl_mutex device_links_lock &rq->__lock irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem rtnl_mutex device_links_lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 sb_writers#4 &type->i_mutex_dir_key#3/1 &sb->s_type->i_mutex_key#8 jbd2_handle &sbi->s_orphan_lock rcu_node_0 irq_context: 0 sb_writers#4 &type->i_mutex_dir_key#3/1 &sb->s_type->i_mutex_key#8 jbd2_handle &sbi->s_orphan_lock &rcu_state.expedited_wq irq_context: 0 sb_writers#4 &type->i_mutex_dir_key#3/1 &sb->s_type->i_mutex_key#8 jbd2_handle &sbi->s_orphan_lock &rcu_state.expedited_wq &p->pi_lock irq_context: 0 sb_writers#4 &type->i_mutex_dir_key#3/1 &sb->s_type->i_mutex_key#8 jbd2_handle &sbi->s_orphan_lock &rcu_state.expedited_wq &p->pi_lock &rq->__lock irq_context: 0 sb_writers#4 &type->i_mutex_dir_key#3/1 &sb->s_type->i_mutex_key#8 jbd2_handle &sbi->s_orphan_lock &rcu_state.expedited_wq &p->pi_lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 sock_diag_mutex sock_diag_table_mutex nlk_cb_mutex-SOCK_DIAG rcu_read_lock &c->lock irq_context: 0 &f->f_pos_lock sb_writers#4 &sb->s_type->i_mutex_key#8 &mm->mmap_lock ptlock_ptr(page)#2 rcu_read_lock &p->pi_lock irq_context: 0 &f->f_pos_lock sb_writers#4 &sb->s_type->i_mutex_key#8 &mm->mmap_lock ptlock_ptr(page)#2 rcu_read_lock &p->pi_lock &rq->__lock irq_context: 0 &f->f_pos_lock sb_writers#4 &sb->s_type->i_mutex_key#8 &mm->mmap_lock ptlock_ptr(page)#2 rcu_read_lock &p->pi_lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 pernet_ops_rwsem rtnl_mutex net_rwsem &rcu_state.expedited_wq irq_context: softirq (&ndev->rs_timer) rcu_read_lock rcu_read_lock rcu_read_lock_bh rcu_read_lock rcu_read_lock &____s->seqcount#9 irq_context: 0 (wq_completion)wg-kex-wg0#9 (work_completion)(&peer->transmit_handshake_work) &peer->endpoint_lock rcu_read_lock_bh rcu_read_lock rcu_read_lock rcu_read_lock &c->lock irq_context: 0 (wq_completion)wg-kex-wg0#9 (work_completion)(&peer->transmit_handshake_work) &peer->endpoint_lock rcu_read_lock_bh rcu_read_lock &c->lock irq_context: 0 (wq_completion)wg-kex-wg0#9 (work_completion)(&peer->transmit_handshake_work) &peer->endpoint_lock rcu_read_lock_bh rcu_read_lock &n->list_lock irq_context: 0 (wq_completion)wg-kex-wg0#9 (work_completion)(&peer->transmit_handshake_work) &peer->endpoint_lock rcu_read_lock_bh rcu_read_lock &n->list_lock &c->lock irq_context: 0 (wq_completion)wg-kex-wg0#9 (work_completion)(&peer->transmit_handshake_work) &peer->endpoint_lock rcu_read_lock_bh rcu_read_lock rcu_read_lock &____s->seqcount#7 irq_context: 0 (wq_completion)wg-kex-wg0#9 (work_completion)(&peer->transmit_handshake_work) &peer->endpoint_lock rcu_read_lock_bh rcu_read_lock &nf_nat_locks[i] irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem rtnl_mutex dev_base_lock &xa->xa_lock#3 &obj_hash[i].lock pool_lock irq_context: 0 (wq_completion)netns net_cleanup_work per_cpu_ptr(&cgroup_rstat_cpu_lock, cpu) irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem ovs_mutex nf_ct_proto_mutex nf_hook_mutex &n->list_lock irq_context: 0 rtnl_mutex rcu_state.barrier_mutex rcu_state.barrier_mutex.wait_lock irq_context: 0 sb_writers#11 rcu_node_0 irq_context: softirq (&in_dev->mr_ifc_timer) rcu_read_lock rcu_read_lock_bh quarantine_lock irq_context: 0 sb_writers#11 &rcu_state.expedited_wq irq_context: 0 sb_writers#11 &rcu_state.expedited_wq &p->pi_lock irq_context: 0 sb_writers#11 &rcu_state.expedited_wq &p->pi_lock &rq->__lock irq_context: 0 sb_writers#11 &rcu_state.expedited_wq &p->pi_lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 sk_lock-AF_RDS batched_entropy_u16.lock irq_context: 0 sk_lock-AF_RDS once_lock irq_context: 0 sk_lock-AF_RDS once_lock crngs.lock irq_context: 0 sk_lock-AF_RDS &rq->__lock irq_context: 0 sk_lock-AF_RDS &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 sk_lock-AF_RDS pool_lock#2 irq_context: 0 sk_lock-AF_RDS &obj_hash[i].lock irq_context: 0 sk_lock-AF_RDS rcu_read_lock &pool->lock irq_context: 0 sk_lock-AF_RDS rcu_read_lock &pool->lock &obj_hash[i].lock irq_context: 0 sk_lock-AF_RDS rcu_read_lock &pool->lock pool_lock#2 irq_context: 0 sk_lock-AF_RDS rcu_read_lock &pool->lock &p->pi_lock irq_context: 0 sk_lock-AF_RDS rcu_read_lock rhashtable_bucket irq_context: 0 rds_trans_sem irq_context: 0 rds_trans_sem fs_reclaim irq_context: 0 rds_trans_sem fs_reclaim mmu_notifier_invalidate_range_start irq_context: 0 rds_trans_sem fs_reclaim mmu_notifier_invalidate_range_start &rq->__lock irq_context: 0 rds_trans_sem fs_reclaim mmu_notifier_invalidate_range_start &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 rds_trans_sem pool_lock#2 irq_context: 0 rds_trans_sem crngs.lock irq_context: 0 rds_trans_sem &id_priv->lock irq_context: 0 rds_trans_sem &id_priv->handler_mutex irq_context: 0 rds_trans_sem &id_priv->handler_mutex &id_priv->lock irq_context: 0 rds_trans_sem id_table_lock irq_context: 0 rds_trans_sem &x->wait#29 irq_context: 0 rds_trans_sem &obj_hash[i].lock irq_context: 0 rcu_read_lock loop_conns_lock irq_context: 0 rcu_read_lock rds_conn_lock irq_context: 0 rcu_read_lock rds_conn_lock rds_cong_lock irq_context: 0 &rm->m_rs_lock irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem rtnl_mutex &root->kernfs_rwsem &p->pi_lock irq_context: 0 &sb->s_type->i_mutex_key#8 rcu_node_0 irq_context: 0 &sb->s_type->i_mutex_key#8 &rcu_state.expedited_wq irq_context: 0 &sb->s_type->i_mutex_key#8 &rcu_state.expedited_wq &p->pi_lock irq_context: 0 &sb->s_type->i_mutex_key#8 &rcu_state.expedited_wq &p->pi_lock &rq->__lock irq_context: 0 &sb->s_type->i_mutex_key#8 &rcu_state.expedited_wq &p->pi_lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem rtnl_mutex dev_pm_qos_sysfs_mtx dev_pm_qos_mtx &rq->__lock irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem rtnl_mutex dev_pm_qos_sysfs_mtx dev_pm_qos_mtx &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 cb_lock genl_mutex &nbd->config_lock rcu_node_0 irq_context: 0 sb_writers#11 &type->i_mutex_dir_key#7/1 &n->list_lock irq_context: 0 sb_writers#11 &type->i_mutex_dir_key#7/1 &n->list_lock &c->lock irq_context: 0 (wq_completion)rcu_gp (work_completion)(&(&ssp->srcu_sup->work)->work) &ssp->srcu_sup->srcu_gp_mutex rcu_node_0 irq_context: 0 (wq_completion)rcu_gp (work_completion)(&(&ssp->srcu_sup->work)->work) &ssp->srcu_sup->srcu_gp_mutex &rcu_state.expedited_wq irq_context: 0 (wq_completion)rcu_gp (work_completion)(&(&ssp->srcu_sup->work)->work) &ssp->srcu_sup->srcu_gp_mutex &rcu_state.expedited_wq &p->pi_lock irq_context: 0 (wq_completion)rcu_gp (work_completion)(&(&ssp->srcu_sup->work)->work) &ssp->srcu_sup->srcu_gp_mutex &rcu_state.expedited_wq &p->pi_lock &rq->__lock irq_context: 0 (wq_completion)rcu_gp (work_completion)(&(&ssp->srcu_sup->work)->work) &ssp->srcu_sup->srcu_gp_mutex &rcu_state.expedited_wq &p->pi_lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 pernet_ops_rwsem ovs_mutex nf_ct_proto_mutex &rq->__lock irq_context: 0 pernet_ops_rwsem ovs_mutex nf_ct_proto_mutex &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 pernet_ops_rwsem ovs_mutex nf_ct_proto_mutex rcu_read_lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)wg-kex-wg1#9 (work_completion)(&peer->transmit_handshake_work) &peer->endpoint_lock rcu_read_lock_bh rcu_read_lock &c->lock irq_context: 0 (wq_completion)wg-kex-wg1#9 (work_completion)(&peer->transmit_handshake_work) &peer->endpoint_lock rcu_read_lock_bh rcu_read_lock rcu_read_lock &____s->seqcount#7 irq_context: 0 (wq_completion)wg-kex-wg1#9 (work_completion)(&peer->transmit_handshake_work) &peer->endpoint_lock rcu_read_lock_bh rcu_read_lock &nf_nat_locks[i] irq_context: 0 (wq_completion)mld (work_completion)(&(&idev->mc_ifc_work)->work) &idev->mc_lock remove_cache_srcu &pcp->lock &zone->lock irq_context: 0 (wq_completion)mld (work_completion)(&(&idev->mc_ifc_work)->work) &idev->mc_lock remove_cache_srcu &pcp->lock &zone->lock &____s->seqcount irq_context: 0 (wq_completion)wg-kex-wg0#9 (work_completion)(&peer->transmit_handshake_work) &peer->endpoint_lock rcu_read_lock_bh rcu_read_lock &____s->seqcount#10 irq_context: 0 (wq_completion)wg-kex-wg2#11 (work_completion)(&peer->transmit_handshake_work) &peer->endpoint_lock rcu_read_lock_bh rcu_read_lock &c->lock irq_context: softirq (&peer->timer_zero_key_material) irq_context: softirq (&peer->timer_zero_key_material) rcu_read_lock_bh rcu_read_lock &pool->lock/1 irq_context: softirq (&peer->timer_zero_key_material) rcu_read_lock_bh rcu_read_lock &pool->lock/1 &obj_hash[i].lock irq_context: softirq (&peer->timer_zero_key_material) rcu_read_lock_bh rcu_read_lock &pool->lock/1 pool_lock#2 irq_context: softirq (&peer->timer_zero_key_material) rcu_read_lock_bh rcu_read_lock &pool->lock/1 &p->pi_lock irq_context: softirq (&peer->timer_zero_key_material) rcu_read_lock_bh rcu_read_lock &pool->lock/1 &p->pi_lock &rq->__lock irq_context: softirq (&peer->timer_zero_key_material) rcu_read_lock_bh rcu_read_lock &pool->lock/1 &p->pi_lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)wg-kex-wg2#9 (work_completion)(&peer->clear_peer_work) irq_context: 0 (wq_completion)wg-kex-wg2#9 (work_completion)(&peer->clear_peer_work) &handshake->lock irq_context: 0 (wq_completion)wg-kex-wg2#9 (work_completion)(&peer->clear_peer_work) &handshake->lock &table->lock#2 irq_context: 0 (wq_completion)wg-kex-wg2#9 (work_completion)(&peer->clear_peer_work) &peer->keypairs.keypair_update_lock irq_context: 0 (wq_completion)wg-kex-wg2#9 (work_completion)(&peer->clear_peer_work) &peer->keypairs.keypair_update_lock &table->lock#2 irq_context: 0 (wq_completion)wg-kex-wg2#9 (work_completion)(&peer->clear_peer_work) &peer->keypairs.keypair_update_lock &obj_hash[i].lock irq_context: 0 (wq_completion)wg-kex-wg2#9 (work_completion)(&peer->clear_peer_work) &peer->keypairs.keypair_update_lock &obj_hash[i].lock pool_lock irq_context: 0 (wq_completion)wg-kex-wg2#9 (work_completion)(&peer->clear_peer_work) &peer->keypairs.keypair_update_lock pool_lock#2 irq_context: 0 (wq_completion)wg-kex-wg0#7 (work_completion)(&peer->clear_peer_work) irq_context: 0 (wq_completion)wg-kex-wg0#7 (work_completion)(&peer->clear_peer_work) &handshake->lock irq_context: 0 (wq_completion)wg-kex-wg0#7 (work_completion)(&peer->clear_peer_work) &handshake->lock &table->lock#2 irq_context: 0 (wq_completion)wg-kex-wg0#7 (work_completion)(&peer->clear_peer_work) &peer->keypairs.keypair_update_lock irq_context: 0 (wq_completion)wg-kex-wg0#7 (work_completion)(&peer->clear_peer_work) &peer->keypairs.keypair_update_lock &table->lock#2 irq_context: 0 (wq_completion)wg-kex-wg0#7 (work_completion)(&peer->clear_peer_work) &peer->keypairs.keypair_update_lock &obj_hash[i].lock irq_context: 0 (wq_completion)tipc_rcv (work_completion)(&srv->awork) k-sk_lock-AF_TIPC &obj_hash[i].lock pool_lock irq_context: 0 sk_lock-AF_TIPC &zone->lock irq_context: 0 sk_lock-AF_TIPC &mm->mmap_lock ptlock_ptr(page)#2 irq_context: 0 (wq_completion)tipc_rcv (work_completion)(&con->rwork) k-sk_lock-AF_TIPC rcu_read_lock pool_lock#2 irq_context: 0 (wq_completion)tipc_rcv (work_completion)(&con->rwork) k-sk_lock-AF_TIPC &zone->lock irq_context: 0 (wq_completion)tipc_rcv (work_completion)(&con->rwork) k-sk_lock-AF_TIPC &zone->lock &____s->seqcount irq_context: 0 (wq_completion)tipc_rcv (work_completion)(&con->rwork) k-clock-AF_TIPC irq_context: 0 (wq_completion)tipc_rcv (work_completion)(&con->rwork) k-clock-AF_TIPC pool_lock#2 irq_context: 0 (wq_completion)tipc_rcv (work_completion)(&con->rwork) k-clock-AF_TIPC &tn->nametbl_lock irq_context: 0 (wq_completion)tipc_rcv (work_completion)(&con->rwork) k-clock-AF_TIPC &tn->nametbl_lock pool_lock#2 irq_context: 0 (wq_completion)tipc_rcv (work_completion)(&con->rwork) k-clock-AF_TIPC &tn->nametbl_lock &service->lock irq_context: 0 (wq_completion)tipc_rcv (work_completion)(&con->rwork) k-clock-AF_TIPC &obj_hash[i].lock irq_context: 0 (wq_completion)tipc_rcv (work_completion)(&con->rwork) k-clock-AF_TIPC &base->lock irq_context: 0 (wq_completion)tipc_rcv (work_completion)(&con->rwork) k-clock-AF_TIPC &base->lock &obj_hash[i].lock irq_context: 0 (wq_completion)tipc_rcv (work_completion)(&con->rwork) k-clock-AF_TIPC &con->sub_lock irq_context: 0 (wq_completion)tipc_rcv (work_completion)(&con->rwork) k-clock-AF_TIPC &c->lock irq_context: softirq (&sub->timer) irq_context: softirq (&sub->timer) &sub->lock irq_context: softirq (&sub->timer) &sub->lock &srv->idr_lock irq_context: softirq (&sub->timer) &sub->lock pool_lock#2 irq_context: softirq (&sub->timer) &sub->lock &con->outqueue_lock irq_context: softirq (&sub->timer) &sub->lock rcu_read_lock &pool->lock/1 irq_context: softirq (&sub->timer) &sub->lock rcu_read_lock &pool->lock/1 &obj_hash[i].lock irq_context: softirq (&sub->timer) &sub->lock rcu_read_lock &pool->lock/1 pool_lock#2 irq_context: softirq (&sub->timer) &sub->lock rcu_read_lock &pool->lock/1 &p->pi_lock irq_context: softirq (&sub->timer) &sub->lock &c->lock irq_context: 0 sk_lock-AF_TIPC &mm->mmap_lock rcu_read_lock rcu_node_0 irq_context: 0 sk_lock-AF_TIPC &mm->mmap_lock rcu_read_lock &rq->__lock irq_context: 0 sk_lock-AF_TIPC &mm->mmap_lock rcu_read_lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)tipc_rcv (work_completion)(&con->rwork) k-clock-AF_TIPC &n->list_lock irq_context: 0 (wq_completion)tipc_rcv (work_completion)(&con->rwork) k-clock-AF_TIPC &n->list_lock &c->lock irq_context: 0 (wq_completion)tipc_send#3 irq_context: 0 (wq_completion)tipc_send#3 &rq->__lock irq_context: 0 (wq_completion)tipc_send#3 (work_completion)(&con->swork) irq_context: 0 (wq_completion)tipc_send#3 (work_completion)(&con->swork) &con->outqueue_lock irq_context: 0 (wq_completion)tipc_send#3 (work_completion)(&con->swork) &con->sub_lock irq_context: 0 (wq_completion)tipc_send#3 (work_completion)(&con->swork) &con->sub_lock &tn->nametbl_lock irq_context: 0 (wq_completion)tipc_send#3 (work_completion)(&con->swork) &con->sub_lock &tn->nametbl_lock &service->lock irq_context: 0 (wq_completion)tipc_send#3 (work_completion)(&con->swork) &con->sub_lock (&sub->timer) irq_context: 0 (wq_completion)tipc_send#3 (work_completion)(&con->swork) &con->sub_lock &obj_hash[i].lock irq_context: 0 (wq_completion)tipc_send#3 (work_completion)(&con->swork) &con->sub_lock &base->lock irq_context: 0 (wq_completion)tipc_send#3 (work_completion)(&con->swork) &con->sub_lock &base->lock &obj_hash[i].lock irq_context: 0 (wq_completion)tipc_send#3 (work_completion)(&con->swork) &con->sub_lock pool_lock#2 irq_context: 0 (wq_completion)tipc_send#3 (work_completion)(&con->swork) &rq->__lock irq_context: 0 (wq_completion)tipc_send#3 (work_completion)(&con->swork) &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)tipc_send#3 (work_completion)(&con->swork) k-sk_lock-AF_TIPC irq_context: 0 (wq_completion)tipc_send#3 (work_completion)(&con->swork) k-sk_lock-AF_TIPC k-slock-AF_TIPC irq_context: 0 (wq_completion)tipc_send#3 (work_completion)(&con->swork) k-sk_lock-AF_TIPC fs_reclaim irq_context: 0 (wq_completion)tipc_send#3 (work_completion)(&con->swork) k-sk_lock-AF_TIPC fs_reclaim mmu_notifier_invalidate_range_start irq_context: 0 (wq_completion)tipc_send#3 (work_completion)(&con->swork) k-sk_lock-AF_TIPC pool_lock#2 irq_context: 0 (wq_completion)tipc_send#3 (work_completion)(&con->swork) k-sk_lock-AF_TIPC &c->lock irq_context: 0 (wq_completion)tipc_send#3 (work_completion)(&con->swork) k-sk_lock-AF_TIPC &list->lock#31 irq_context: 0 (wq_completion)tipc_send#3 (work_completion)(&con->swork) k-sk_lock-AF_TIPC slock-AF_TIPC &list->lock#31 irq_context: 0 (wq_completion)tipc_send#3 (work_completion)(&con->swork) k-slock-AF_TIPC irq_context: 0 (wq_completion)tipc_send#3 (work_completion)(&con->swork) &con->outqueue_lock &obj_hash[i].lock irq_context: 0 (wq_completion)tipc_rcv (work_completion)(&con->rwork) k-clock-AF_TIPC &obj_hash[i].lock pool_lock irq_context: softirq (&sub->timer) &sub->lock rcu_read_lock &pool->lock/1 &p->pi_lock &rq->__lock irq_context: softirq (&sub->timer) &sub->lock rcu_read_lock &pool->lock/1 &p->pi_lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)tipc_send#3 (work_completion)(&con->swork) k-sk_lock-AF_TIPC k-slock-AF_TIPC &sk->sk_lock.wq#2 irq_context: 0 (wq_completion)tipc_send#3 (work_completion)(&con->swork) k-sk_lock-AF_TIPC &rq->__lock irq_context: 0 (wq_completion)tipc_send#3 (work_completion)(&con->swork) k-sk_lock-AF_TIPC &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)tipc_rcv (work_completion)(&con->rwork) k-slock-AF_TIPC &sk->sk_lock.wq#2 irq_context: 0 (wq_completion)tipc_rcv (work_completion)(&con->rwork) k-slock-AF_TIPC &sk->sk_lock.wq#2 &p->pi_lock irq_context: 0 (wq_completion)tipc_rcv (work_completion)(&con->rwork) k-slock-AF_TIPC &sk->sk_lock.wq#2 &p->pi_lock &rq->__lock irq_context: 0 (wq_completion)tipc_rcv (work_completion)(&con->rwork) k-slock-AF_TIPC &sk->sk_lock.wq#2 &p->pi_lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)tipc_rcv (work_completion)(&con->rwork) k-sk_lock-AF_TIPC k-slock-AF_TIPC &sk->sk_lock.wq#2 irq_context: 0 (wq_completion)tipc_send#3 (work_completion)(&con->swork) k-slock-AF_TIPC &sk->sk_lock.wq#2 irq_context: 0 (wq_completion)tipc_send#3 (work_completion)(&con->swork) k-slock-AF_TIPC &sk->sk_lock.wq#2 &p->pi_lock irq_context: 0 (wq_completion)tipc_send#3 (work_completion)(&con->swork) k-slock-AF_TIPC &sk->sk_lock.wq#2 &p->pi_lock &rq->__lock irq_context: 0 (wq_completion)tipc_send#3 (work_completion)(&con->swork) k-slock-AF_TIPC &sk->sk_lock.wq#2 &p->pi_lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)tipc_rcv (work_completion)(&con->rwork) k-sk_lock-AF_TIPC &____s->seqcount#2 irq_context: 0 (wq_completion)tipc_rcv (work_completion)(&con->rwork) k-sk_lock-AF_TIPC &____s->seqcount irq_context: 0 (wq_completion)tipc_rcv (work_completion)(&con->rwork) k-sk_lock-AF_TIPC slock-AF_TIPC rcu_read_lock &ei->socket.wq.wait irq_context: 0 (wq_completion)tipc_rcv (work_completion)(&con->rwork) k-sk_lock-AF_TIPC slock-AF_TIPC rcu_read_lock &ei->socket.wq.wait &p->pi_lock irq_context: 0 (wq_completion)tipc_rcv (work_completion)(&con->rwork) k-sk_lock-AF_TIPC slock-AF_TIPC rcu_read_lock &ei->socket.wq.wait &p->pi_lock &rq->__lock irq_context: 0 (wq_completion)tipc_rcv (work_completion)(&con->rwork) k-sk_lock-AF_TIPC slock-AF_TIPC rcu_read_lock &ei->socket.wq.wait &p->pi_lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)tipc_send#3 (work_completion)(&con->swork) k-sk_lock-AF_TIPC slock-AF_TIPC rcu_read_lock &ei->socket.wq.wait irq_context: 0 (wq_completion)tipc_send#3 (work_completion)(&con->swork) k-sk_lock-AF_TIPC slock-AF_TIPC rcu_read_lock &ei->socket.wq.wait &p->pi_lock irq_context: 0 sk_lock-AF_TIPC &mm->mmap_lock fs_reclaim irq_context: 0 sk_lock-AF_TIPC &mm->mmap_lock fs_reclaim mmu_notifier_invalidate_range_start irq_context: 0 sk_lock-AF_TIPC &mm->mmap_lock &____s->seqcount irq_context: 0 sk_lock-AF_TIPC &mm->mmap_lock stock_lock irq_context: 0 sk_lock-AF_TIPC &mm->mmap_lock &c->lock irq_context: 0 sk_lock-AF_TIPC &mm->mmap_lock ptlock_ptr(page) irq_context: 0 (wq_completion)tipc_rcv (work_completion)(&con->rwork) &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)tipc_rcv (work_completion)(&con->rwork) k-sk_lock-AF_TIPC &n->list_lock irq_context: 0 (wq_completion)tipc_rcv (work_completion)(&con->rwork) k-sk_lock-AF_TIPC &n->list_lock &c->lock irq_context: 0 sock_diag_mutex sock_diag_table_mutex nlk_cb_mutex-SOCK_DIAG k-sk_lock-AF_TIPC irq_context: 0 sock_diag_mutex sock_diag_table_mutex nlk_cb_mutex-SOCK_DIAG k-sk_lock-AF_TIPC k-slock-AF_TIPC irq_context: 0 sock_diag_mutex sock_diag_table_mutex nlk_cb_mutex-SOCK_DIAG k-sk_lock-AF_TIPC k-clock-AF_TIPC irq_context: 0 sock_diag_mutex sock_diag_table_mutex nlk_cb_mutex-SOCK_DIAG k-slock-AF_TIPC irq_context: 0 sock_diag_mutex sock_diag_table_mutex nlk_cb_mutex-SOCK_DIAG sk_lock-AF_TIPC irq_context: 0 sock_diag_mutex sock_diag_table_mutex nlk_cb_mutex-SOCK_DIAG sk_lock-AF_TIPC slock-AF_TIPC irq_context: 0 sock_diag_mutex sock_diag_table_mutex nlk_cb_mutex-SOCK_DIAG sk_lock-AF_TIPC clock-AF_TIPC irq_context: 0 sock_diag_mutex sock_diag_table_mutex nlk_cb_mutex-SOCK_DIAG slock-AF_TIPC irq_context: 0 (wq_completion)tipc_rcv (work_completion)(&con->rwork) k-clock-AF_TIPC (console_sem).lock irq_context: 0 (wq_completion)tipc_rcv (work_completion)(&con->rwork) k-clock-AF_TIPC console_lock console_srcu console_owner_lock irq_context: 0 (wq_completion)tipc_rcv (work_completion)(&con->rwork) k-clock-AF_TIPC console_lock console_srcu console_owner irq_context: 0 (wq_completion)tipc_send#3 (work_completion)(&con->swork) k-sk_lock-AF_TIPC &____s->seqcount#2 irq_context: 0 (wq_completion)tipc_rcv (work_completion)(&con->rwork) k-clock-AF_TIPC console_lock console_srcu console_owner &port_lock_key irq_context: 0 (wq_completion)tipc_send#3 (work_completion)(&con->swork) k-sk_lock-AF_TIPC &____s->seqcount irq_context: 0 (wq_completion)tipc_rcv (work_completion)(&con->rwork) k-clock-AF_TIPC console_lock console_srcu console_owner console_owner_lock irq_context: 0 (wq_completion)tipc_rcv (work_completion)(&con->rwork) k-clock-AF_TIPC &con->sub_lock &tn->nametbl_lock irq_context: 0 (wq_completion)tipc_rcv (work_completion)(&con->rwork) k-clock-AF_TIPC &con->sub_lock &tn->nametbl_lock &service->lock irq_context: 0 (wq_completion)tipc_rcv (work_completion)(&con->rwork) k-clock-AF_TIPC &con->sub_lock (&sub->timer) irq_context: 0 (wq_completion)tipc_rcv (work_completion)(&con->rwork) k-clock-AF_TIPC &con->sub_lock &obj_hash[i].lock irq_context: 0 (wq_completion)tipc_rcv (work_completion)(&con->rwork) k-clock-AF_TIPC &con->sub_lock &base->lock irq_context: 0 (wq_completion)tipc_rcv (work_completion)(&con->rwork) k-clock-AF_TIPC &con->sub_lock pool_lock#2 irq_context: 0 (wq_completion)tipc_rcv (work_completion)(&con->rwork) k-clock-AF_TIPC &con->sub_lock &tn->nametbl_lock &service->lock &obj_hash[i].lock irq_context: 0 (wq_completion)tipc_rcv (work_completion)(&con->rwork) k-clock-AF_TIPC &con->sub_lock &tn->nametbl_lock &service->lock pool_lock#2 irq_context: 0 (wq_completion)tipc_rcv (work_completion)(&con->rwork) k-clock-AF_TIPC &con->sub_lock &tn->nametbl_lock &service->lock krc.lock irq_context: 0 (wq_completion)events (work_completion)(&nlk->work) &ht->lock irq_context: 0 (wq_completion)tipc_rcv (work_completion)(&srv->awork) &c->lock irq_context: 0 (wq_completion)tipc_rcv#3 irq_context: 0 (wq_completion)tipc_rcv#3 (work_completion)(&srv->awork) irq_context: 0 (wq_completion)tipc_rcv#3 (work_completion)(&srv->awork) &srv->idr_lock irq_context: 0 (wq_completion)tipc_rcv#3 (work_completion)(&srv->awork) fs_reclaim irq_context: 0 (wq_completion)tipc_rcv#3 (work_completion)(&srv->awork) fs_reclaim mmu_notifier_invalidate_range_start irq_context: 0 (wq_completion)tipc_rcv#3 (work_completion)(&srv->awork) mmu_notifier_invalidate_range_start irq_context: 0 (wq_completion)tipc_rcv#3 (work_completion)(&srv->awork) &sb->s_type->i_lock_key#8 irq_context: 0 (wq_completion)tipc_rcv#3 (work_completion)(&srv->awork) k-sk_lock-AF_TIPC irq_context: 0 (wq_completion)tipc_rcv#3 (work_completion)(&srv->awork) k-sk_lock-AF_TIPC k-slock-AF_TIPC irq_context: 0 (wq_completion)tipc_rcv#3 (work_completion)(&srv->awork) k-sk_lock-AF_TIPC fs_reclaim irq_context: 0 (wq_completion)tipc_rcv#3 (work_completion)(&srv->awork) k-sk_lock-AF_TIPC fs_reclaim mmu_notifier_invalidate_range_start irq_context: 0 (wq_completion)tipc_rcv#3 (work_completion)(&srv->awork) k-sk_lock-AF_TIPC &dir->lock irq_context: 0 (wq_completion)tipc_rcv#3 (work_completion)(&srv->awork) k-sk_lock-AF_TIPC &obj_hash[i].lock irq_context: 0 (wq_completion)tipc_rcv#3 (work_completion)(&srv->awork) k-sk_lock-AF_TIPC batched_entropy_u32.lock irq_context: 0 (wq_completion)tipc_rcv#3 (work_completion)(&srv->awork) k-sk_lock-AF_TIPC rcu_read_lock rhashtable_bucket irq_context: 0 (wq_completion)tipc_rcv#3 (work_completion)(&srv->awork) k-sk_lock-AF_TIPC k-sk_lock-AF_TIPC/1 irq_context: 0 (wq_completion)tipc_rcv#3 (work_completion)(&srv->awork) k-sk_lock-AF_TIPC k-sk_lock-AF_TIPC/1 k-slock-AF_TIPC irq_context: 0 (wq_completion)tipc_rcv#3 (work_completion)(&srv->awork) k-sk_lock-AF_TIPC k-sk_lock-AF_TIPC/1 &obj_hash[i].lock irq_context: 0 (wq_completion)tipc_rcv#3 (work_completion)(&srv->awork) k-sk_lock-AF_TIPC k-sk_lock-AF_TIPC/1 &base->lock irq_context: 0 (wq_completion)tipc_rcv#3 (work_completion)(&srv->awork) k-sk_lock-AF_TIPC k-sk_lock-AF_TIPC/1 &base->lock &obj_hash[i].lock irq_context: 0 (wq_completion)tipc_rcv#3 (work_completion)(&srv->awork) k-sk_lock-AF_TIPC k-sk_lock-AF_TIPC/1 fs_reclaim irq_context: 0 (wq_completion)tipc_rcv#3 (work_completion)(&srv->awork) k-sk_lock-AF_TIPC k-sk_lock-AF_TIPC/1 fs_reclaim mmu_notifier_invalidate_range_start irq_context: 0 (wq_completion)tipc_rcv#3 (work_completion)(&srv->awork) k-sk_lock-AF_TIPC k-sk_lock-AF_TIPC/1 &list->lock#31 irq_context: 0 (wq_completion)tipc_rcv#3 (work_completion)(&srv->awork) k-sk_lock-AF_TIPC k-sk_lock-AF_TIPC/1 slock-AF_TIPC &list->lock#31 irq_context: 0 (wq_completion)tipc_rcv#3 (work_completion)(&srv->awork) k-slock-AF_TIPC irq_context: 0 (wq_completion)tipc_rcv#3 (work_completion)(&srv->awork) &c->lock irq_context: 0 (wq_completion)tipc_rcv#3 (work_completion)(&srv->awork) &obj_hash[i].lock irq_context: 0 (wq_completion)tipc_rcv#3 (work_completion)(&srv->awork) k-clock-AF_TIPC irq_context: 0 (wq_completion)tipc_rcv#3 (work_completion)(&srv->awork) k-clock-AF_TIPC irq_context: 0 (wq_completion)tipc_rcv#3 (work_completion)(&srv->awork) k-clock-AF_TIPC rcu_read_lock &pool->lock/1 irq_context: 0 (wq_completion)tipc_rcv#3 (work_completion)(&srv->awork) k-clock-AF_TIPC rcu_read_lock &pool->lock/1 &obj_hash[i].lock irq_context: 0 (wq_completion)tipc_rcv#3 (work_completion)(&srv->awork) k-clock-AF_TIPC rcu_read_lock &pool->lock/1 &p->pi_lock irq_context: 0 (wq_completion)tipc_rcv#3 (work_completion)(&srv->awork) &xa->xa_lock#9 irq_context: 0 (wq_completion)tipc_rcv#3 (work_completion)(&srv->awork) &fsnotify_mark_srcu irq_context: 0 (wq_completion)tipc_rcv#3 (work_completion)(&con->rwork) irq_context: 0 (wq_completion)tipc_rcv#3 (work_completion)(&con->rwork) k-sk_lock-AF_TIPC irq_context: 0 (wq_completion)tipc_rcv#3 (work_completion)(&con->rwork) k-sk_lock-AF_TIPC k-slock-AF_TIPC irq_context: 0 (wq_completion)tipc_rcv#3 (work_completion)(&con->rwork) k-slock-AF_TIPC irq_context: 0 (wq_completion)tipc_rcv (work_completion)(&con->rwork) k-clock-AF_TIPC &tn->nametbl_lock &c->lock irq_context: 0 (wq_completion)tipc_send#3 (work_completion)(&con->swork) k-sk_lock-AF_TIPC k-clock-AF_TIPC irq_context: 0 (wq_completion)tipc_send#3 (work_completion)(&con->swork) k-sk_lock-AF_TIPC k-clock-AF_TIPC rcu_read_lock &pool->lock/1 irq_context: 0 (wq_completion)tipc_send#3 (work_completion)(&con->swork) k-sk_lock-AF_TIPC k-clock-AF_TIPC rcu_read_lock &pool->lock/1 &obj_hash[i].lock irq_context: 0 (wq_completion)tipc_send#3 (work_completion)(&con->swork) k-sk_lock-AF_TIPC k-clock-AF_TIPC rcu_read_lock &pool->lock/1 pool_lock#2 irq_context: 0 (wq_completion)tipc_send#3 (work_completion)(&con->swork) k-sk_lock-AF_TIPC k-clock-AF_TIPC rcu_read_lock &pool->lock/1 &p->pi_lock irq_context: 0 (wq_completion)tipc_send#3 (work_completion)(&con->swork) k-sk_lock-AF_TIPC k-clock-AF_TIPC rcu_read_lock &pool->lock/1 &p->pi_lock &rq->__lock irq_context: 0 (wq_completion)tipc_send#3 (work_completion)(&con->swork) k-sk_lock-AF_TIPC k-clock-AF_TIPC rcu_read_lock &pool->lock/1 &p->pi_lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)tipc_send#3 (work_completion)(&con->swork) &con->sub_lock &tn->nametbl_lock &service->lock &obj_hash[i].lock irq_context: 0 (wq_completion)tipc_send#3 (work_completion)(&con->swork) &con->sub_lock &tn->nametbl_lock &service->lock pool_lock#2 irq_context: 0 (wq_completion)tipc_send#3 (work_completion)(&con->swork) &con->sub_lock &tn->nametbl_lock &service->lock krc.lock irq_context: 0 sk_lock-AF_TIPC &mm->mmap_lock rcu_node_0 irq_context: 0 (wq_completion)tipc_rcv#3 (work_completion)(&con->rwork) k-sk_lock-AF_TIPC &obj_hash[i].lock irq_context: 0 (wq_completion)tipc_rcv#3 (work_completion)(&con->rwork) k-sk_lock-AF_TIPC &zone->lock irq_context: 0 (wq_completion)tipc_rcv#3 (work_completion)(&con->rwork) k-sk_lock-AF_TIPC &zone->lock &____s->seqcount irq_context: 0 (wq_completion)tipc_rcv#3 (work_completion)(&con->rwork) k-sk_lock-AF_TIPC &list->lock#31 irq_context: 0 (wq_completion)tipc_rcv#3 (work_completion)(&con->rwork) k-sk_lock-AF_TIPC slock-AF_TIPC &list->lock#31 irq_context: 0 (wq_completion)tipc_rcv#3 (work_completion)(&con->rwork) k-sk_lock-AF_TIPC slock-AF_TIPC rcu_read_lock &ei->socket.wq.wait irq_context: 0 (wq_completion)tipc_rcv#3 (work_completion)(&con->rwork) k-sk_lock-AF_TIPC slock-AF_TIPC rcu_read_lock &ei->socket.wq.wait &p->pi_lock irq_context: 0 (wq_completion)tipc_rcv#3 (work_completion)(&con->rwork) k-sk_lock-AF_TIPC slock-AF_TIPC rcu_read_lock &ei->socket.wq.wait &p->pi_lock &cfs_rq->removed.lock irq_context: 0 (wq_completion)tipc_rcv#3 (work_completion)(&con->rwork) k-sk_lock-AF_TIPC slock-AF_TIPC rcu_read_lock &ei->socket.wq.wait &p->pi_lock &rq->__lock irq_context: 0 (wq_completion)tipc_rcv#3 (work_completion)(&con->rwork) k-sk_lock-AF_TIPC slock-AF_TIPC rcu_read_lock &ei->socket.wq.wait &p->pi_lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)tipc_rcv#3 (work_completion)(&con->rwork) k-sk_lock-AF_TIPC slock-AF_TIPC &obj_hash[i].lock irq_context: 0 (wq_completion)tipc_rcv#3 (work_completion)(&con->rwork) k-sk_lock-AF_TIPC &rq->__lock irq_context: 0 (wq_completion)tipc_rcv#3 (work_completion)(&con->rwork) k-sk_lock-AF_TIPC &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 sk_lock-AF_TIPC &mm->mmap_lock &sem->wait_lock irq_context: 0 (wq_completion)tipc_rcv#3 (work_completion)(&con->rwork) k-sk_lock-AF_TIPC k-clock-AF_TIPC irq_context: 0 (wq_completion)tipc_rcv#3 (work_completion)(&con->rwork) k-sk_lock-AF_TIPC k-clock-AF_TIPC rcu_read_lock &pool->lock/1 irq_context: 0 (wq_completion)tipc_rcv#3 (work_completion)(&con->rwork) k-sk_lock-AF_TIPC k-clock-AF_TIPC rcu_read_lock &pool->lock/1 &obj_hash[i].lock irq_context: 0 (wq_completion)tipc_rcv#3 (work_completion)(&con->rwork) k-sk_lock-AF_TIPC k-clock-AF_TIPC rcu_read_lock &pool->lock/1 &p->pi_lock irq_context: 0 (wq_completion)tipc_rcv#3 (work_completion)(&con->rwork) k-sk_lock-AF_TIPC k-clock-AF_TIPC rcu_read_lock &pool->lock/1 &p->pi_lock &rq->__lock irq_context: 0 (wq_completion)tipc_rcv#3 (work_completion)(&con->rwork) k-sk_lock-AF_TIPC k-clock-AF_TIPC rcu_read_lock &pool->lock/1 &p->pi_lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)tipc_rcv#3 (work_completion)(&con->rwork) k-clock-AF_TIPC irq_context: 0 (wq_completion)tipc_rcv#3 (work_completion)(&con->rwork) k-clock-AF_TIPC &tn->nametbl_lock irq_context: 0 (wq_completion)tipc_rcv#3 (work_completion)(&con->rwork) k-clock-AF_TIPC &tn->nametbl_lock &service->lock irq_context: 0 (wq_completion)tipc_rcv#3 (work_completion)(&con->rwork) k-clock-AF_TIPC &obj_hash[i].lock irq_context: 0 (wq_completion)tipc_rcv#3 (work_completion)(&con->rwork) k-clock-AF_TIPC &base->lock irq_context: 0 (wq_completion)tipc_rcv#3 (work_completion)(&con->rwork) k-clock-AF_TIPC &base->lock &obj_hash[i].lock irq_context: 0 (wq_completion)tipc_rcv#3 (work_completion)(&con->rwork) k-clock-AF_TIPC &con->sub_lock irq_context: 0 (wq_completion)tipc_rcv#3 (work_completion)(&con->rwork) &rq->__lock irq_context: 0 (wq_completion)tipc_rcv#3 (work_completion)(&con->rwork) k-clock-AF_TIPC &c->lock irq_context: 0 (wq_completion)tipc_rcv#3 (work_completion)(&con->rwork) k-clock-AF_TIPC &obj_hash[i].lock pool_lock irq_context: 0 (wq_completion)tipc_rcv#3 (work_completion)(&con->rwork) &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)tipc_send (work_completion)(&con->swork) &con->sub_lock irq_context: 0 (wq_completion)tipc_send (work_completion)(&con->swork) &con->sub_lock &tn->nametbl_lock irq_context: 0 (wq_completion)tipc_send (work_completion)(&con->swork) &con->sub_lock &tn->nametbl_lock &service->lock irq_context: 0 (wq_completion)tipc_send (work_completion)(&con->swork) &con->sub_lock (&sub->timer) irq_context: 0 (wq_completion)tipc_send (work_completion)(&con->swork) &con->sub_lock &obj_hash[i].lock irq_context: 0 (wq_completion)tipc_send (work_completion)(&con->swork) &con->sub_lock &base->lock irq_context: 0 (wq_completion)tipc_send (work_completion)(&con->swork) k-sk_lock-AF_TIPC irq_context: 0 (wq_completion)tipc_send (work_completion)(&con->swork) k-sk_lock-AF_TIPC k-slock-AF_TIPC irq_context: 0 (wq_completion)tipc_send (work_completion)(&con->swork) k-sk_lock-AF_TIPC fs_reclaim irq_context: 0 (wq_completion)tipc_send (work_completion)(&con->swork) k-sk_lock-AF_TIPC fs_reclaim mmu_notifier_invalidate_range_start irq_context: 0 (wq_completion)tipc_send (work_completion)(&con->swork) k-sk_lock-AF_TIPC &c->lock irq_context: 0 (wq_completion)tipc_send (work_completion)(&con->swork) k-sk_lock-AF_TIPC &list->lock#31 irq_context: 0 (wq_completion)tipc_send (work_completion)(&con->swork) k-sk_lock-AF_TIPC slock-AF_TIPC &list->lock#31 irq_context: 0 (wq_completion)tipc_send (work_completion)(&con->swork) k-sk_lock-AF_TIPC slock-AF_TIPC rcu_read_lock &ei->socket.wq.wait irq_context: 0 (wq_completion)tipc_send (work_completion)(&con->swork) k-sk_lock-AF_TIPC slock-AF_TIPC rcu_read_lock &ei->socket.wq.wait &p->pi_lock irq_context: 0 (wq_completion)tipc_send (work_completion)(&con->swork) k-slock-AF_TIPC irq_context: 0 (wq_completion)tipc_rcv#3 (work_completion)(&con->rwork) k-sk_lock-AF_TIPC &c->lock irq_context: 0 (wq_completion)tipc_rcv#3 (work_completion)(&con->rwork) k-sk_lock-AF_TIPC &n->list_lock irq_context: 0 (wq_completion)tipc_rcv#3 (work_completion)(&con->rwork) k-sk_lock-AF_TIPC &n->list_lock &c->lock irq_context: 0 (wq_completion)tipc_rcv#3 (work_completion)(&con->rwork) k-sk_lock-AF_TIPC &____s->seqcount#2 irq_context: 0 (wq_completion)tipc_rcv#3 (work_completion)(&con->rwork) k-sk_lock-AF_TIPC &pcp->lock &zone->lock irq_context: 0 (wq_completion)tipc_rcv#3 (work_completion)(&con->rwork) k-sk_lock-AF_TIPC &____s->seqcount irq_context: 0 (wq_completion)tipc_rcv#3 (work_completion)(&con->rwork) k-clock-AF_TIPC &n->list_lock irq_context: 0 (wq_completion)tipc_rcv#3 (work_completion)(&con->rwork) k-clock-AF_TIPC &n->list_lock &c->lock irq_context: 0 sk_lock-AF_TIPC rcu_read_lock pool_lock#2 irq_context: 0 &sb->s_type->i_mutex_key#10 sk_lock-AF_TIPC quarantine_lock irq_context: 0 (wq_completion)tipc_send (work_completion)(&con->swork) k-sk_lock-AF_TIPC k-slock-AF_TIPC &sk->sk_lock.wq#2 irq_context: 0 (wq_completion)tipc_rcv#3 (work_completion)(&con->rwork) k-slock-AF_TIPC &sk->sk_lock.wq#2 irq_context: 0 (wq_completion)tipc_send (work_completion)(&con->swork) k-sk_lock-AF_TIPC &rq->__lock irq_context: 0 (wq_completion)tipc_rcv#3 (work_completion)(&con->rwork) k-slock-AF_TIPC &sk->sk_lock.wq#2 &p->pi_lock irq_context: 0 (wq_completion)tipc_send (work_completion)(&con->swork) k-sk_lock-AF_TIPC &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)tipc_rcv#3 (work_completion)(&con->rwork) k-slock-AF_TIPC &sk->sk_lock.wq#2 &p->pi_lock &rq->__lock irq_context: 0 (wq_completion)tipc_rcv#3 (work_completion)(&con->rwork) k-slock-AF_TIPC &sk->sk_lock.wq#2 &p->pi_lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)tipc_rcv#3 (work_completion)(&con->rwork) k-clock-AF_TIPC batched_entropy_u8.lock irq_context: 0 (wq_completion)tipc_rcv#3 (work_completion)(&con->rwork) k-clock-AF_TIPC kfence_freelist_lock irq_context: 0 (wq_completion)tipc_send (work_completion)(&con->swork) &con->sub_lock &base->lock &obj_hash[i].lock irq_context: 0 (wq_completion)tipc_send (work_completion)(&con->swork) &con->sub_lock &meta->lock irq_context: 0 (wq_completion)tipc_send (work_completion)(&con->swork) &con->sub_lock kfence_freelist_lock irq_context: 0 (wq_completion)tipc_send (work_completion)(&con->swork) k-sk_lock-AF_TIPC &____s->seqcount#2 irq_context: 0 (wq_completion)tipc_send (work_completion)(&con->swork) k-sk_lock-AF_TIPC &____s->seqcount irq_context: 0 &sb->s_type->i_mutex_key#10 sk_lock-AF_TIPC &n->list_lock irq_context: 0 &sb->s_type->i_mutex_key#10 sk_lock-AF_TIPC &n->list_lock &c->lock irq_context: 0 &sb->s_type->i_mutex_key#10 sk_lock-AF_TIPC k-slock-AF_TIPC &obj_hash[i].lock irq_context: 0 &sb->s_type->i_mutex_key#10 sk_lock-AF_TIPC k-slock-AF_TIPC pool_lock#2 irq_context: 0 (wq_completion)tipc_rcv#3 (work_completion)(&con->rwork) k-sk_lock-AF_TIPC rcu_node_0 irq_context: 0 (wq_completion)tipc_send (work_completion)(&con->swork) k-sk_lock-AF_TIPC k-clock-AF_TIPC irq_context: 0 (wq_completion)tipc_rcv#3 (work_completion)(&con->rwork) k-sk_lock-AF_TIPC rcu_read_lock rcu_node_0 irq_context: 0 (wq_completion)tipc_rcv#3 (work_completion)(&con->rwork) k-sk_lock-AF_TIPC rcu_read_lock &rq->__lock irq_context: 0 (wq_completion)tipc_rcv#3 (work_completion)(&con->rwork) k-sk_lock-AF_TIPC rcu_read_lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)tipc_rcv#3 (work_completion)(&con->rwork) k-slock-AF_TIPC &sk->sk_lock.wq#2 &p->pi_lock &cfs_rq->removed.lock irq_context: 0 (wq_completion)tipc_send (work_completion)(&con->swork) k-sk_lock-AF_TIPC slock-AF_TIPC rcu_read_lock &ei->socket.wq.wait &p->pi_lock &rq->__lock irq_context: 0 (wq_completion)tipc_send (work_completion)(&con->swork) k-sk_lock-AF_TIPC slock-AF_TIPC rcu_read_lock &ei->socket.wq.wait &p->pi_lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 sk_lock-AF_TIPC &mm->mmap_lock rcu_read_lock &rcu_state.gp_wq irq_context: 0 sk_lock-AF_TIPC &mm->mmap_lock rcu_read_lock &rcu_state.gp_wq &p->pi_lock irq_context: 0 sk_lock-AF_TIPC &mm->mmap_lock rcu_read_lock &rcu_state.gp_wq &p->pi_lock &rq->__lock irq_context: 0 sk_lock-AF_TIPC &mm->mmap_lock rcu_read_lock &rcu_state.gp_wq &p->pi_lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)tipc_send (work_completion)(&con->swork) &con->sub_lock &obj_hash[i].lock pool_lock irq_context: 0 (wq_completion)tipc_rcv#3 (work_completion)(&con->rwork) k-sk_lock-AF_TIPC k-clock-AF_TIPC rcu_read_lock &pool->lock/1 &p->pi_lock &cfs_rq->removed.lock irq_context: 0 (wq_completion)tipc_rcv#3 (work_completion)(&con->rwork) k-sk_lock-AF_TIPC k-slock-AF_TIPC &sk->sk_lock.wq#2 irq_context: 0 (wq_completion)tipc_send (work_completion)(&con->swork) k-slock-AF_TIPC &sk->sk_lock.wq#2 irq_context: 0 (wq_completion)tipc_send (work_completion)(&con->swork) k-slock-AF_TIPC &sk->sk_lock.wq#2 &p->pi_lock irq_context: 0 (wq_completion)tipc_send (work_completion)(&con->swork) k-slock-AF_TIPC &sk->sk_lock.wq#2 &p->pi_lock &rq->__lock irq_context: 0 (wq_completion)tipc_send (work_completion)(&con->swork) k-slock-AF_TIPC &sk->sk_lock.wq#2 &p->pi_lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)tipc_send (work_completion)(&con->swork) k-sk_lock-AF_TIPC k-clock-AF_TIPC rcu_read_lock &pool->lock/1 irq_context: 0 (wq_completion)tipc_send (work_completion)(&con->swork) k-sk_lock-AF_TIPC k-clock-AF_TIPC rcu_read_lock &pool->lock/1 &obj_hash[i].lock irq_context: 0 (wq_completion)tipc_send (work_completion)(&con->swork) k-sk_lock-AF_TIPC k-clock-AF_TIPC rcu_read_lock &pool->lock/1 &p->pi_lock irq_context: 0 (wq_completion)tipc_send (work_completion)(&con->swork) k-sk_lock-AF_TIPC k-clock-AF_TIPC rcu_read_lock &pool->lock/1 &p->pi_lock &rq->__lock irq_context: 0 (wq_completion)tipc_send (work_completion)(&con->swork) k-sk_lock-AF_TIPC k-clock-AF_TIPC rcu_read_lock &pool->lock/1 &p->pi_lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)tipc_send (work_completion)(&con->swork) &con->outqueue_lock quarantine_lock irq_context: 0 (wq_completion)tipc_send (work_completion)(&con->swork) k-sk_lock-AF_TIPC remove_cache_srcu irq_context: 0 (wq_completion)tipc_send (work_completion)(&con->swork) k-sk_lock-AF_TIPC remove_cache_srcu quarantine_lock irq_context: 0 (wq_completion)tipc_send (work_completion)(&con->swork) k-sk_lock-AF_TIPC remove_cache_srcu &c->lock irq_context: 0 (wq_completion)tipc_send (work_completion)(&con->swork) k-sk_lock-AF_TIPC remove_cache_srcu &n->list_lock irq_context: 0 (wq_completion)tipc_send (work_completion)(&con->swork) k-sk_lock-AF_TIPC remove_cache_srcu rcu_read_lock pool_lock#2 irq_context: 0 (wq_completion)tipc_send (work_completion)(&con->swork) k-sk_lock-AF_TIPC remove_cache_srcu &obj_hash[i].lock irq_context: 0 (wq_completion)tipc_send (work_completion)(&con->swork) k-sk_lock-AF_TIPC remove_cache_srcu pool_lock#2 irq_context: 0 cb_lock genl_mutex &nbd->config_lock &cfs_rq->removed.lock irq_context: 0 (wq_completion)tipc_send (work_completion)(&con->swork) k-sk_lock-AF_TIPC &n->list_lock irq_context: 0 (wq_completion)tipc_send (work_completion)(&con->swork) k-sk_lock-AF_TIPC &n->list_lock &c->lock irq_context: softirq (&sub->timer) &sub->lock &n->list_lock irq_context: softirq (&sub->timer) &sub->lock &n->list_lock &c->lock irq_context: 0 cb_lock genl_mutex &nbd->config_lock &base->lock irq_context: 0 cb_lock genl_mutex &nbd->config_lock &base->lock &obj_hash[i].lock irq_context: 0 (wq_completion)tipc_rcv#3 (work_completion)(&con->rwork) k-sk_lock-AF_TIPC &meta->lock irq_context: 0 (wq_completion)tipc_rcv#3 (work_completion)(&con->rwork) k-sk_lock-AF_TIPC kfence_freelist_lock irq_context: 0 (wq_completion)tipc_rcv#3 (work_completion)(&con->rwork) k-clock-AF_TIPC &____s->seqcount#2 irq_context: 0 (wq_completion)tipc_rcv#3 (work_completion)(&con->rwork) k-clock-AF_TIPC &____s->seqcount irq_context: 0 (wq_completion)tipc_rcv#3 (work_completion)(&con->rwork) k-clock-AF_TIPC pool_lock#2 irq_context: 0 sk_lock-AF_TIPC &zone->lock &____s->seqcount irq_context: 0 (wq_completion)tipc_rcv#3 (work_completion)(&con->rwork) &srv->idr_lock irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem (wq_completion)krxrpcd &rq->__lock irq_context: 0 (wq_completion)tipc_rcv#3 (work_completion)(&con->rwork) k-clock-AF_TIPC irq_context: 0 (wq_completion)tipc_rcv#3 (work_completion)(&con->rwork) k-clock-AF_TIPC &con->sub_lock irq_context: 0 (wq_completion)tipc_rcv#3 (work_completion)(&con->rwork) k-clock-AF_TIPC &con->sub_lock &tn->nametbl_lock irq_context: 0 (wq_completion)tipc_rcv#3 (work_completion)(&con->rwork) k-clock-AF_TIPC &con->sub_lock &tn->nametbl_lock &service->lock irq_context: 0 (wq_completion)tipc_rcv#3 (work_completion)(&con->rwork) k-clock-AF_TIPC &con->sub_lock (&sub->timer) irq_context: 0 (wq_completion)tipc_rcv#3 (work_completion)(&con->rwork) k-clock-AF_TIPC &con->sub_lock &obj_hash[i].lock irq_context: 0 (wq_completion)tipc_rcv#3 (work_completion)(&con->rwork) k-clock-AF_TIPC &con->sub_lock &base->lock irq_context: 0 (wq_completion)tipc_rcv#3 (work_completion)(&con->rwork) k-clock-AF_TIPC &con->sub_lock &obj_hash[i].lock pool_lock irq_context: 0 (wq_completion)tipc_rcv#3 (work_completion)(&con->rwork) &srv->idr_lock &obj_hash[i].lock irq_context: 0 (wq_completion)tipc_rcv#3 (work_completion)(&con->rwork) k-sk_lock-AF_TIPC &base->lock irq_context: 0 (wq_completion)tipc_rcv#3 (work_completion)(&con->rwork) k-sk_lock-AF_TIPC &base->lock &obj_hash[i].lock irq_context: 0 (wq_completion)tipc_rcv#3 (work_completion)(&con->rwork) k-sk_lock-AF_TIPC rcu_read_lock rhashtable_bucket irq_context: 0 (wq_completion)tipc_rcv#3 (work_completion)(&con->rwork) k-sk_lock-AF_TIPC k-clock-AF_TIPC irq_context: 0 (wq_completion)tipc_rcv#3 (work_completion)(&con->rwork) &obj_hash[i].lock irq_context: 0 (wq_completion)tipc_rcv#3 (work_completion)(&con->rwork) &sb->s_type->i_lock_key#8 irq_context: 0 (wq_completion)tipc_rcv#3 (work_completion)(&con->rwork) &xa->xa_lock#9 irq_context: 0 (wq_completion)tipc_rcv#3 (work_completion)(&con->rwork) &fsnotify_mark_srcu irq_context: 0 (wq_completion)tipc_rcv#3 (work_completion)(&con->rwork) &con->outqueue_lock irq_context: 0 (wq_completion)tipc_rcv#3 (work_completion)(&con->rwork) &con->outqueue_lock &obj_hash[i].lock irq_context: 0 (wq_completion)tipc_rcv#3 (work_completion)(&con->rwork) &con->outqueue_lock pool_lock#2 irq_context: 0 (wq_completion)tipc_rcv (work_completion)(&con->rwork) k-sk_lock-AF_TIPC rcu_node_0 irq_context: 0 (wq_completion)tipc_rcv (work_completion)(&srv->awork) &srv->idr_lock &c->lock irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem (wq_completion)krxrpcd &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)tipc_rcv (work_completion)(&con->rwork) k-sk_lock-AF_TIPC &rcu_state.expedited_wq irq_context: 0 (wq_completion)tipc_rcv (work_completion)(&con->rwork) k-sk_lock-AF_TIPC &rcu_state.expedited_wq &p->pi_lock irq_context: 0 (wq_completion)tipc_rcv (work_completion)(&con->rwork) k-sk_lock-AF_TIPC &rcu_state.expedited_wq &p->pi_lock &rq->__lock irq_context: 0 (wq_completion)tipc_rcv (work_completion)(&con->rwork) k-sk_lock-AF_TIPC &rcu_state.expedited_wq &p->pi_lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 sk_lock-AF_TIPC &rq->__lock &cfs_rq->removed.lock irq_context: 0 mapping.invalidate_lock &obj_hash[i].lock irq_context: 0 &mm->mmap_lock mapping.invalidate_lock irq_context: 0 &f->f_pos_lock sb_writers#4 &sb->s_type->i_mutex_key#8 &sbi->s_writepages_rwsem &c->lock irq_context: 0 &f->f_pos_lock sb_writers#4 &sb->s_type->i_mutex_key#8 &sbi->s_writepages_rwsem rcu_read_lock &obj_hash[i].lock pool_lock irq_context: 0 &f->f_pos_lock sb_writers#4 &sb->s_type->i_mutex_key#8 &mapping->i_mmap_rwsem ptlock_ptr(page)#2 &mapping->private_lock irq_context: 0 &f->f_pos_lock sb_writers#4 &sb->s_type->i_mutex_key#8 &mapping->i_mmap_rwsem ptlock_ptr(page)#2 &mapping->private_lock rcu_read_lock &memcg->move_lock irq_context: 0 &f->f_pos_lock sb_writers#4 &sb->s_type->i_mutex_key#8 mapping.invalidate_lock irq_context: 0 &f->f_pos_lock sb_writers#4 &sb->s_type->i_mutex_key#8 mapping.invalidate_lock mmu_notifier_invalidate_range_start irq_context: 0 &f->f_pos_lock sb_writers#4 &sb->s_type->i_mutex_key#8 mapping.invalidate_lock &____s->seqcount irq_context: 0 &f->f_pos_lock sb_writers#4 &sb->s_type->i_mutex_key#8 mapping.invalidate_lock pool_lock#2 irq_context: 0 &f->f_pos_lock sb_writers#4 &sb->s_type->i_mutex_key#8 mapping.invalidate_lock stock_lock irq_context: 0 &f->f_pos_lock sb_writers#4 &sb->s_type->i_mutex_key#8 mapping.invalidate_lock &xa->xa_lock#9 irq_context: 0 &f->f_pos_lock sb_writers#4 &sb->s_type->i_mutex_key#8 mapping.invalidate_lock lock#4 irq_context: 0 &f->f_pos_lock sb_writers#4 &sb->s_type->i_mutex_key#8 mapping.invalidate_lock &ei->i_es_lock irq_context: 0 &f->f_pos_lock sb_writers#4 &sb->s_type->i_mutex_key#8 mapping.invalidate_lock tk_core.seq.seqcount irq_context: 0 &f->f_pos_lock sb_writers#4 &sb->s_type->i_mutex_key#8 mapping.invalidate_lock &dd->lock irq_context: 0 &f->f_pos_lock sb_writers#4 &sb->s_type->i_mutex_key#8 mapping.invalidate_lock rcu_read_lock &dd->lock irq_context: 0 &f->f_pos_lock sb_writers#4 &sb->s_type->i_mutex_key#8 mapping.invalidate_lock rcu_read_lock &obj_hash[i].lock irq_context: 0 &f->f_pos_lock sb_writers#4 &sb->s_type->i_mutex_key#8 mapping.invalidate_lock rcu_read_lock pool_lock#2 irq_context: 0 &f->f_pos_lock sb_writers#4 &sb->s_type->i_mutex_key#8 mapping.invalidate_lock rcu_read_lock rcu_node_0 irq_context: 0 &f->f_pos_lock sb_writers#4 &sb->s_type->i_mutex_key#8 mapping.invalidate_lock rcu_read_lock &rq->__lock irq_context: 0 &f->f_pos_lock sb_writers#4 &sb->s_type->i_mutex_key#8 mapping.invalidate_lock rcu_read_lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 &f->f_pos_lock sb_writers#4 &sb->s_type->i_mutex_key#8 mapping.invalidate_lock rcu_read_lock tk_core.seq.seqcount irq_context: 0 &f->f_pos_lock sb_writers#4 &sb->s_type->i_mutex_key#8 mapping.invalidate_lock rcu_read_lock &virtscsi_vq->vq_lock irq_context: 0 &f->f_pos_lock sb_writers#4 &sb->s_type->i_mutex_key#8 &mm->mmap_lock mapping.invalidate_lock irq_context: 0 &f->f_pos_lock sb_writers#4 &sb->s_type->i_mutex_key#8 mapping.invalidate_lock &folio_wait_table[i] irq_context: 0 &f->f_pos_lock sb_writers#4 &sb->s_type->i_mutex_key#8 mapping.invalidate_lock &rq->__lock irq_context: 0 &f->f_pos_lock sb_writers#4 &sb->s_type->i_mutex_key#8 mapping.invalidate_lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 &f->f_pos_lock sb_writers#4 &sbi->s_writepages_rwsem &mapping->i_mmap_rwsem mmu_notifier_invalidate_range_start irq_context: 0 &f->f_pos_lock sb_writers#4 &sbi->s_writepages_rwsem &mapping->i_mmap_rwsem mmu_notifier_invalidate_range_start &rq->__lock irq_context: 0 &f->f_pos_lock sb_writers#4 &sbi->s_writepages_rwsem &mapping->i_mmap_rwsem ptlock_ptr(page)#2 irq_context: 0 &f->f_pos_lock sb_writers#4 &sbi->s_writepages_rwsem &mapping->private_lock irq_context: 0 &f->f_pos_lock sb_writers#4 &sbi->s_writepages_rwsem &mapping->private_lock rcu_read_lock &memcg->move_lock irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem rtnl_mutex &root->kernfs_rwsem &base->lock irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem rtnl_mutex &root->kernfs_rwsem &base->lock &obj_hash[i].lock irq_context: 0 sk_lock-AF_INET6 slock-AF_INET6 &obj_hash[i].lock irq_context: 0 sk_lock-AF_INET6 slock-AF_INET6 elock-AF_INET6 irq_context: 0 sk_lock-AF_INET6 rcu_read_lock rcu_read_lock_bh &meta->lock irq_context: 0 sk_lock-AF_INET6 rcu_read_lock rcu_read_lock_bh kfence_freelist_lock irq_context: 0 mapping.invalidate_lock &lruvec->lru_lock irq_context: 0 mapping.invalidate_lock fs_reclaim irq_context: 0 mapping.invalidate_lock fs_reclaim mmu_notifier_invalidate_range_start irq_context: 0 mapping.invalidate_lock rcu_read_lock &pool->lock irq_context: 0 mapping.invalidate_lock rcu_read_lock &pool->lock &obj_hash[i].lock irq_context: 0 mapping.invalidate_lock rcu_read_lock &pool->lock pool_lock#2 irq_context: 0 mapping.invalidate_lock rcu_read_lock &pool->lock &p->pi_lock irq_context: 0 mapping.invalidate_lock rcu_read_lock &pool->lock &p->pi_lock &rq->__lock irq_context: 0 mapping.invalidate_lock rcu_read_lock &pool->lock &p->pi_lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 sb_writers#4 &sb->s_type->i_mutex_key#8 &journal->j_list_lock irq_context: 0 &f->f_pos_lock sb_writers#4 &sb->s_type->i_mutex_key#8 jbd2_handle &ei->i_data_sem &ei->i_es_lock &n->list_lock irq_context: 0 &f->f_pos_lock sb_writers#4 &sb->s_type->i_mutex_key#8 jbd2_handle &ei->i_data_sem &ei->i_es_lock &n->list_lock &c->lock irq_context: 0 &f->f_pos_lock sb_writers#4 &sb->s_type->i_mutex_key#8 mapping.invalidate_lock &c->lock irq_context: 0 &f->f_pos_lock sb_writers#4 &sb->s_type->i_mutex_key#8 mapping.invalidate_lock jbd2_handle &xa->xa_lock#9 &c->lock irq_context: 0 (wq_completion)events (work_completion)(&p->wq) purge_vmap_area_lock &base->lock irq_context: 0 (wq_completion)events (work_completion)(&p->wq) purge_vmap_area_lock &base->lock &obj_hash[i].lock irq_context: 0 &mm->mmap_lock sb_pagefaults mapping.invalidate_lock &ei->i_es_lock key#2 irq_context: 0 &vma->vm_lock->lock rcu_read_lock rcu_read_lock &cfs_rq->removed.lock irq_context: 0 &vma->vm_lock->lock rcu_read_lock rcu_read_lock &obj_hash[i].lock irq_context: 0 rtnl_mutex &tb->tb6_lock batched_entropy_u8.lock irq_context: 0 rtnl_mutex &tb->tb6_lock kfence_freelist_lock irq_context: 0 rtnl_mutex &tb->tb6_lock &meta->lock irq_context: 0 &tfile->napi_mutex irq_context: 0 &tfile->napi_mutex &mm->mmap_lock irq_context: 0 &tfile->napi_mutex pcpu_lock irq_context: 0 &tfile->napi_mutex &obj_hash[i].lock irq_context: 0 &tfile->napi_mutex pool_lock#2 irq_context: 0 rtnl_mutex rcu_read_lock rcu_read_lock rcu_read_lock rcu_node_0 irq_context: 0 rtnl_mutex rcu_read_lock rcu_read_lock rcu_read_lock &rq->__lock irq_context: 0 rtnl_mutex rcu_read_lock rcu_read_lock rcu_read_lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 rtnl_mutex &this->info_list_lock irq_context: 0 (wq_completion)mld (work_completion)(&(&idev->mc_ifc_work)->work) &idev->mc_lock rcu_read_lock rcu_read_lock rcu_read_lock_bh rcu_read_lock rcu_read_lock_bh rcu_read_lock pool_lock#2 irq_context: 0 (wq_completion)mld (work_completion)(&(&idev->mc_ifc_work)->work) &idev->mc_lock rcu_read_lock rcu_read_lock rcu_read_lock_bh rcu_read_lock rcu_read_lock_bh rcu_read_lock &obj_hash[i].lock irq_context: 0 (wq_completion)mld (work_completion)(&(&idev->mc_ifc_work)->work) &idev->mc_lock rcu_read_lock rcu_read_lock rcu_read_lock_bh rcu_read_lock &____s->seqcount#2 irq_context: 0 (wq_completion)mld (work_completion)(&(&idev->mc_ifc_work)->work) &idev->mc_lock rcu_read_lock rcu_read_lock rcu_read_lock_bh rcu_read_lock rcu_read_lock_bh dev->qdisc_tx_busylock ?: &qdisc_tx_busylock rcu_read_lock pool_lock#2 irq_context: 0 (wq_completion)mld (work_completion)(&(&idev->mc_ifc_work)->work) &idev->mc_lock rcu_read_lock rcu_read_lock rcu_read_lock_bh rcu_read_lock rcu_read_lock_bh dev->qdisc_tx_busylock ?: &qdisc_tx_busylock rcu_read_lock tk_core.seq.seqcount irq_context: 0 (wq_completion)mld (work_completion)(&(&idev->mc_ifc_work)->work) &idev->mc_lock rcu_read_lock rcu_read_lock rcu_read_lock_bh rcu_read_lock rcu_read_lock_bh dev->qdisc_tx_busylock ?: &qdisc_tx_busylock rcu_read_lock &obj_hash[i].lock irq_context: 0 (wq_completion)mld (work_completion)(&(&idev->mc_ifc_work)->work) &idev->mc_lock rcu_read_lock rcu_read_lock rcu_read_lock_bh rcu_read_lock rcu_read_lock_bh dev->qdisc_tx_busylock ?: &qdisc_tx_busylock rcu_read_lock &r->producer_lock irq_context: 0 (wq_completion)mld (work_completion)(&(&idev->mc_ifc_work)->work) &idev->mc_lock rcu_read_lock rcu_read_lock rcu_read_lock_bh dev->qdisc_tx_busylock ?: &qdisc_tx_busylock rcu_read_lock tk_core.seq.seqcount irq_context: 0 (wq_completion)mld (work_completion)(&(&idev->mc_ifc_work)->work) &idev->mc_lock rcu_read_lock rcu_read_lock rcu_read_lock_bh dev->qdisc_tx_busylock ?: &qdisc_tx_busylock rcu_read_lock &obj_hash[i].lock irq_context: 0 (wq_completion)ipv6_addrconf (work_completion)(&(&ifa->dad_work)->work) rtnl_mutex &idev->mc_lock krc.lock irq_context: 0 (wq_completion)ipv6_addrconf (work_completion)(&(&ifa->dad_work)->work) rtnl_mutex &idev->mc_lock rcu_read_lock rcu_read_lock rcu_read_lock_bh dev->qdisc_tx_busylock ?: &qdisc_tx_busylock rcu_read_lock pool_lock#2 irq_context: 0 (wq_completion)ipv6_addrconf (work_completion)(&(&ifa->dad_work)->work) rtnl_mutex &idev->mc_lock rcu_read_lock rcu_read_lock rcu_read_lock_bh dev->qdisc_tx_busylock ?: &qdisc_tx_busylock rcu_read_lock tk_core.seq.seqcount irq_context: 0 (wq_completion)ipv6_addrconf (work_completion)(&(&ifa->dad_work)->work) rtnl_mutex &idev->mc_lock rcu_read_lock rcu_read_lock rcu_read_lock_bh dev->qdisc_tx_busylock ?: &qdisc_tx_busylock rcu_read_lock &obj_hash[i].lock irq_context: 0 rtnl_mutex &br->multicast_lock &obj_hash[i].lock irq_context: 0 rtnl_mutex &br->multicast_lock &base->lock irq_context: 0 rtnl_mutex &br->multicast_lock &base->lock &obj_hash[i].lock irq_context: 0 rtnl_mutex rcu_read_lock &____s->seqcount#13 irq_context: 0 rtnl_mutex rcu_read_lock &____s->seqcount#14 irq_context: softirq (&brmctx->ip6_own_query.timer) &br->multicast_lock &c->lock irq_context: softirq (&brmctx->ip6_own_query.timer) &br->multicast_lock pool_lock#2 irq_context: softirq (&brmctx->ip6_own_query.timer) &br->multicast_lock batched_entropy_u8.lock irq_context: softirq (&brmctx->ip6_own_query.timer) &br->multicast_lock kfence_freelist_lock irq_context: softirq (&brmctx->ip6_own_query.timer) &br->multicast_lock &obj_hash[i].lock irq_context: softirq (&brmctx->ip6_own_query.timer) &br->multicast_lock &meta->lock irq_context: softirq (&brmctx->ip6_own_query.timer) &br->multicast_lock &base->lock irq_context: softirq (&brmctx->ip6_own_query.timer) &br->multicast_lock &base->lock &obj_hash[i].lock irq_context: softirq (&brmctx->ip4_own_query.timer) &br->multicast_lock pool_lock#2 irq_context: softirq (&brmctx->ip4_own_query.timer) &br->multicast_lock &c->lock irq_context: softirq (&brmctx->ip4_own_query.timer) &br->multicast_lock tk_core.seq.seqcount irq_context: softirq (&brmctx->ip4_own_query.timer) &br->multicast_lock &list->lock#5 irq_context: softirq (&brmctx->ip4_own_query.timer) &br->multicast_lock &obj_hash[i].lock irq_context: softirq (&brmctx->ip4_own_query.timer) &br->multicast_lock &base->lock irq_context: softirq (&brmctx->ip4_own_query.timer) &br->multicast_lock &base->lock &obj_hash[i].lock irq_context: softirq rcu_read_lock rcu_read_lock &nf_nat_locks[i] irq_context: softirq rcu_read_lock rcu_read_lock &in_dev->mc_tomb_lock irq_context: softirq rcu_read_lock rcu_read_lock rcu_read_lock &im->lock irq_context: softirq rcu_read_lock rcu_read_lock rcu_read_lock &im->lock &obj_hash[i].lock irq_context: softirq rcu_read_lock rcu_read_lock rcu_read_lock &im->lock batched_entropy_u32.lock irq_context: softirq rcu_read_lock rcu_read_lock rcu_read_lock &im->lock &base->lock irq_context: softirq rcu_read_lock rcu_read_lock rcu_read_lock &im->lock &base->lock &obj_hash[i].lock irq_context: 0 rtnl_mutex &im->lock &base->lock irq_context: 0 rtnl_mutex &im->lock &base->lock &obj_hash[i].lock irq_context: 0 (wq_completion)wg-kex-wg1#9 (work_completion)(&peer->transmit_handshake_work) &peer->endpoint_lock rcu_read_lock_bh rcu_read_lock &obj_hash[i].lock irq_context: 0 (wq_completion)wg-kex-wg1#9 (work_completion)(&peer->transmit_handshake_work) &peer->endpoint_lock rcu_read_lock_bh rcu_read_lock rcu_read_lock_bh rcu_read_lock pool_lock#2 irq_context: 0 sb_writers#4 &type->i_mutex_dir_key#3 rcu_read_lock &p->pi_lock irq_context: 0 sb_writers#4 &type->i_mutex_dir_key#3 rcu_read_lock &p->pi_lock &rq->__lock irq_context: 0 sb_writers#4 &type->i_mutex_dir_key#3 rcu_read_lock &p->pi_lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)mld (work_completion)(&(&idev->mc_ifc_work)->work) &idev->mc_lock rcu_read_lock rcu_read_lock &meta->lock irq_context: 0 (wq_completion)mld (work_completion)(&(&idev->mc_ifc_work)->work) &idev->mc_lock rcu_read_lock rcu_read_lock kfence_freelist_lock irq_context: 0 (wq_completion)mld (work_completion)(&(&idev->mc_ifc_work)->work) &idev->mc_lock rcu_read_lock rcu_read_lock rcu_read_lock_bh rcu_read_lock &ndev->lock irq_context: 0 sk_lock-AF_QIPCRTR clock-AF_QIPCRTR irq_context: 0 sk_lock-AF_QIPCRTR rlock-AF_QIPCRTR irq_context: 0 (wq_completion)mld (work_completion)(&(&idev->mc_ifc_work)->work) &idev->mc_lock rcu_read_lock rcu_read_lock rcu_read_lock_bh _xmit_TUNNEL6#2 irq_context: 0 (wq_completion)mld (work_completion)(&(&idev->mc_ifc_work)->work) &idev->mc_lock rcu_read_lock rcu_read_lock rcu_read_lock_bh _xmit_TUNNEL6#2 rcu_read_lock &ndev->lock irq_context: 0 (wq_completion)mld (work_completion)(&(&idev->mc_ifc_work)->work) &idev->mc_lock rcu_read_lock rcu_read_lock rcu_read_lock_bh _xmit_TUNNEL6#2 &obj_hash[i].lock irq_context: 0 (wq_completion)mld (work_completion)(&(&idev->mc_ifc_work)->work) &idev->mc_lock rcu_read_lock rcu_read_lock rcu_read_lock_bh _xmit_TUNNEL6#2 pool_lock#2 irq_context: 0 sk_lock-AF_QIPCRTR &rq->__lock irq_context: 0 sk_lock-AF_QIPCRTR &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 sk_lock-AF_QIPCRTR rcu_read_lock rcu_node_0 irq_context: 0 sk_lock-AF_QIPCRTR rcu_read_lock &rq->__lock irq_context: 0 sk_lock-AF_QIPCRTR rcu_read_lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 &tfile->napi_mutex rcu_read_lock pool_lock#2 irq_context: 0 &tfile->napi_mutex rcu_read_lock rcu_node_0 irq_context: 0 &tfile->napi_mutex rcu_read_lock &rq->__lock irq_context: 0 &tfile->napi_mutex rcu_read_lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 &tfile->napi_mutex &lock->wait_lock irq_context: 0 &tfile->napi_mutex &rq->__lock irq_context: 0 &tfile->napi_mutex &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 &tfile->napi_mutex &____s->seqcount irq_context: 0 &tfile->napi_mutex &c->lock irq_context: 0 &tfile->napi_mutex &n->list_lock irq_context: 0 &tfile->napi_mutex &n->list_lock &c->lock irq_context: 0 (wq_completion)mld (work_completion)(&(&idev->mc_ifc_work)->work) &idev->mc_lock rcu_read_lock rcu_read_lock rcu_read_lock_bh rcu_read_lock rcu_read_lock_bh dev->qdisc_tx_busylock ?: &qdisc_tx_busylock rcu_read_lock &c->lock irq_context: 0 (wq_completion)mld (work_completion)(&(&idev->mc_ifc_work)->work) &idev->mc_lock rcu_read_lock rcu_read_lock rcu_read_lock_bh dev->qdisc_tx_busylock ?: &qdisc_tx_busylock rcu_read_lock &c->lock irq_context: 0 (wq_completion)mld (work_completion)(&(&idev->mc_ifc_work)->work) &idev->mc_lock rcu_read_lock rcu_read_lock rcu_read_lock_bh _xmit_TUNNEL6#2 rcu_read_lock tk_core.seq.seqcount irq_context: 0 (wq_completion)mld (work_completion)(&(&idev->mc_ifc_work)->work) &idev->mc_lock rcu_read_lock rcu_read_lock rcu_read_lock_bh _xmit_TUNNEL6#2 rcu_read_lock &obj_hash[i].lock irq_context: 0 dup_mmap_sem &mm->mmap_lock &mm->mmap_lock/1 &rq->__lock &obj_hash[i].lock irq_context: 0 dup_mmap_sem &mm->mmap_lock &mm->mmap_lock/1 &rq->__lock &base->lock irq_context: 0 dup_mmap_sem &mm->mmap_lock &mm->mmap_lock/1 &rq->__lock &base->lock &obj_hash[i].lock irq_context: 0 (wq_completion)ipv6_addrconf (work_completion)(&(&ifa->dad_work)->work) rtnl_mutex &idev->mc_lock rcu_read_lock rcu_read_lock rcu_read_lock_bh rcu_read_lock &ndev->lock irq_context: 0 rtnl_mutex &pnn->routes.lock &rq->__lock irq_context: 0 rtnl_mutex &pnn->routes.lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)wg-kex-wg0#9 (work_completion)(&peer->transmit_handshake_work) &wg->static_identity.lock &handshake->lock rcu_read_lock_bh batched_entropy_u32.lock crngs.lock irq_context: 0 (wq_completion)wg-kex-wg2#11 (work_completion)(&peer->transmit_handshake_work) &peer->endpoint_lock rcu_read_lock_bh rcu_read_lock rcu_read_lock_bh rcu_read_lock pool_lock#2 irq_context: 0 (wq_completion)wg-kex-wg0#9 (work_completion)(&peer->transmit_handshake_work) &peer->endpoint_lock rcu_read_lock_bh rcu_read_lock rcu_read_lock_bh rcu_read_lock pool_lock#2 irq_context: 0 (wq_completion)mld (work_completion)(&(&idev->mc_dad_work)->work) &idev->mc_lock rcu_read_lock rcu_read_lock rcu_read_lock_bh dev->qdisc_tx_busylock ?: &qdisc_tx_busylock rcu_read_lock tk_core.seq.seqcount irq_context: 0 (wq_completion)mld (work_completion)(&(&idev->mc_dad_work)->work) &idev->mc_lock rcu_read_lock rcu_read_lock rcu_read_lock_bh dev->qdisc_tx_busylock ?: &qdisc_tx_busylock rcu_read_lock &obj_hash[i].lock irq_context: 0 (wq_completion)ipv6_addrconf (work_completion)(&(&ifa->dad_work)->work) rtnl_mutex &idev->mc_lock rcu_read_lock &pool->lock irq_context: 0 (wq_completion)ipv6_addrconf (work_completion)(&(&ifa->dad_work)->work) rtnl_mutex &idev->mc_lock rcu_read_lock &pool->lock &obj_hash[i].lock irq_context: 0 rcu_state.barrier_mutex pgd_lock irq_context: 0 rcu_state.barrier_mutex stock_lock irq_context: 0 rcu_state.barrier_mutex rcu_read_lock pool_lock#2 irq_context: 0 rcu_state.barrier_mutex key irq_context: 0 rcu_state.barrier_mutex pcpu_lock irq_context: 0 rcu_state.barrier_mutex percpu_counters_lock irq_context: 0 rcu_state.barrier_mutex pcpu_lock stock_lock irq_context: 0 (wq_completion)mld (work_completion)(&(&idev->mc_ifc_work)->work) &idev->mc_lock rcu_read_lock rcu_read_lock rcu_read_lock_bh _xmit_TUNNEL6#2 rcu_read_lock &c->lock irq_context: 0 (wq_completion)mld (work_completion)(&(&idev->mc_dad_work)->work) &idev->mc_lock rcu_read_lock rcu_read_lock rcu_read_lock_bh &meta->lock irq_context: 0 (wq_completion)mld (work_completion)(&(&idev->mc_dad_work)->work) &idev->mc_lock rcu_read_lock rcu_read_lock rcu_read_lock_bh kfence_freelist_lock irq_context: 0 (wq_completion)wg-kex-wg2#11 (work_completion)(&peer->transmit_handshake_work) &peer->endpoint_lock rcu_read_lock_bh batched_entropy_u32.lock crngs.lock irq_context: 0 rtnl_mutex uevent_sock_mutex &rq->__lock &cfs_rq->removed.lock irq_context: 0 (wq_completion)mld (work_completion)(&(&idev->mc_dad_work)->work) &idev->mc_lock rcu_read_lock rcu_read_lock rcu_read_lock_bh rcu_read_lock &ndev->lock irq_context: 0 sb_writers#4 mmu_notifier_invalidate_range_start &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 &u->iolock rcu_read_lock rcu_read_lock rcu_node_0 irq_context: 0 &u->iolock rcu_read_lock rcu_read_lock &rq->__lock irq_context: 0 &u->iolock rcu_read_lock rcu_read_lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 &u->iolock &mm->mmap_lock stock_lock irq_context: 0 &u->iolock &mm->mmap_lock &anon_vma->rwsem irq_context: 0 &u->iolock &mm->mmap_lock &anon_vma->rwsem &mm->page_table_lock irq_context: 0 &u->iolock &mm->mmap_lock &anon_vma->rwsem &rq->__lock irq_context: 0 &u->iolock &mm->mmap_lock &anon_vma->rwsem &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 &u->iolock &mm->mmap_lock rcu_node_0 irq_context: 0 &u->iolock &mm->mmap_lock &rq->__lock irq_context: 0 &u->iolock &mm->mmap_lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 &u->iolock &mm->mmap_lock rcu_read_lock rcu_node_0 irq_context: 0 &u->iolock &mm->mmap_lock &pcp->lock &zone->lock irq_context: 0 &u->iolock &mm->mmap_lock &pcp->lock &zone->lock &____s->seqcount irq_context: 0 &u->iolock &mm->mmap_lock ptlock_ptr(page)#2 key irq_context: 0 &u->iolock &mm->mmap_lock rcu_read_lock &rq->__lock irq_context: 0 &u->iolock &mm->mmap_lock rcu_read_lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 &u->iolock &mm->mmap_lock rcu_read_lock rcu_read_lock rcu_node_0 irq_context: 0 &u->iolock &mm->mmap_lock rcu_read_lock rcu_read_lock &rq->__lock irq_context: 0 &u->iolock &mm->mmap_lock rcu_read_lock rcu_read_lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)mld (work_completion)(&(&idev->mc_ifc_work)->work) &idev->mc_lock rcu_node_0 irq_context: 0 &u->iolock &mm->mmap_lock fs_reclaim &rq->__lock irq_context: 0 fs_reclaim mmu_notifier_invalidate_range_start &rcu_state.expedited_wq &p->pi_lock &rq->__lock irq_context: 0 fs_reclaim mmu_notifier_invalidate_range_start &rcu_state.expedited_wq &p->pi_lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 &u->iolock &sem->wait_lock irq_context: 0 &u->iolock &p->pi_lock irq_context: 0 &u->iolock &p->pi_lock &rq->__lock irq_context: 0 &u->iolock &p->pi_lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 &u->iolock &mm->mmap_lock &sem->wait_lock irq_context: 0 &u->iolock &mm->mmap_lock fs_reclaim &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 &u->iolock &mm->mmap_lock rcu_read_lock &rq->__lock &cfs_rq->removed.lock irq_context: 0 &u->iolock &mm->mmap_lock ptlock_ptr(page)#2 lock#4 &obj_hash[i].lock irq_context: 0 &u->iolock &mm->mmap_lock rcu_read_lock &rcu_state.gp_wq irq_context: 0 &u->iolock &mm->mmap_lock rcu_read_lock &rcu_state.gp_wq &p->pi_lock irq_context: 0 &u->iolock &mm->mmap_lock rcu_read_lock &rcu_state.gp_wq &p->pi_lock &rq->__lock irq_context: 0 &u->iolock &mm->mmap_lock rcu_read_lock &rcu_state.gp_wq &p->pi_lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 &u->iolock &mm->mmap_lock mmu_notifier_invalidate_range_start &rq->__lock irq_context: 0 pernet_ops_rwsem __ip_vs_mutex &rq->__lock irq_context: 0 &sb->s_type->i_mutex_key#8 &sbi->s_writepages_rwsem lock#5 &lruvec->lru_lock irq_context: 0 cb_lock rtnl_mutex &dev->power.lock irq_context: 0 &sb->s_type->i_mutex_key#8 lock#5 &lruvec->lru_lock irq_context: 0 &sb->s_type->i_mutex_key#8 &sbi->s_writepages_rwsem jbd2_handle lock#5 &lruvec->lru_lock irq_context: 0 &f->f_pos_lock sb_writers#11 &of->mutex &rq->__lock irq_context: 0 &f->f_pos_lock sb_writers#11 &of->mutex &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 pernet_ops_rwsem nf_ct_proto_mutex defrag4_mutex nf_hook_mutex fs_reclaim mmu_notifier_invalidate_range_start &rq->__lock irq_context: 0 pernet_ops_rwsem nf_ct_proto_mutex defrag4_mutex nf_hook_mutex fs_reclaim mmu_notifier_invalidate_range_start &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 namespace_sem remove_cache_srcu &c->lock irq_context: 0 namespace_sem remove_cache_srcu &n->list_lock irq_context: 0 namespace_sem remove_cache_srcu &rq->__lock irq_context: 0 pernet_ops_rwsem devices_rwsem rdma_nets_rwsem &device->compat_devs_mutex bus_type_sem &rq->__lock irq_context: 0 pernet_ops_rwsem devices_rwsem rdma_nets_rwsem &device->compat_devs_mutex bus_type_sem &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem wq_pool_mutex &wq->mutex &rq->__lock irq_context: 0 namespace_sem remove_cache_srcu &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 namespace_sem remove_cache_srcu rcu_read_lock pool_lock#2 irq_context: 0 namespace_sem remove_cache_srcu &obj_hash[i].lock irq_context: 0 sb_writers#4 lock#5 &lruvec->lru_lock irq_context: 0 rcu_read_lock rcu_read_lock_bh rcu_read_lock once_lock irq_context: 0 rcu_read_lock rcu_read_lock_bh rcu_read_lock once_lock crngs.lock irq_context: 0 rcu_read_lock rcu_read_lock_bh rcu_read_lock rcu_read_lock &pool->lock irq_context: 0 rcu_read_lock rcu_read_lock_bh rcu_read_lock rcu_read_lock &pool->lock &obj_hash[i].lock irq_context: 0 rcu_read_lock rcu_read_lock_bh rcu_read_lock rcu_read_lock &pool->lock pool_lock#2 irq_context: 0 rcu_read_lock rcu_read_lock_bh rcu_read_lock rcu_read_lock &pool->lock &p->pi_lock irq_context: 0 rcu_read_lock rcu_read_lock_bh rcu_read_lock rcu_read_lock &pool->lock &p->pi_lock &rq->__lock irq_context: 0 rcu_read_lock rcu_read_lock_bh rcu_read_lock rcu_read_lock &pool->lock &p->pi_lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 rcu_read_lock rcu_read_lock_bh rcu_read_lock fnhe_lock irq_context: 0 rcu_read_lock rcu_read_lock_bh rcu_read_lock fnhe_lock &____s->seqcount irq_context: 0 rcu_read_lock rcu_read_lock_bh rcu_read_lock fnhe_lock pool_lock#2 irq_context: 0 rcu_read_lock rcu_read_lock_bh rcu_read_lock fnhe_lock batched_entropy_u8.lock irq_context: 0 rcu_read_lock rcu_read_lock_bh rcu_read_lock rcu_read_lock &____s->seqcount#7 irq_context: 0 rcu_read_lock rcu_read_lock_bh rcu_read_lock &nf_nat_locks[i] irq_context: 0 rcu_read_lock rcu_read_lock_bh rcu_read_lock &nf_conntrack_locks[i] irq_context: 0 rcu_read_lock rcu_read_lock_bh rcu_read_lock &nf_conntrack_locks[i] &nf_conntrack_locks[i]/1 irq_context: 0 rcu_read_lock rcu_read_lock_bh rcu_read_lock &nf_conntrack_locks[i] &nf_conntrack_locks[i]/1 batched_entropy_u8.lock irq_context: 0 rcu_read_lock rcu_read_lock_bh rcu_read_lock &nf_conntrack_locks[i]/1 irq_context: 0 rcu_read_lock rcu_read_lock_bh rcu_read_lock &____s->seqcount#9 irq_context: 0 sk_lock-AF_INET6 k-sk_lock-AF_INET6/1 pool_lock#2 irq_context: 0 sk_lock-AF_INET6 &token_hash[i].lock irq_context: 0 sk_lock-AF_INET6 k-sk_lock-AF_INET6 rcu_read_lock rcu_read_lock rcu_node_0 irq_context: 0 sk_lock-AF_INET6 k-sk_lock-AF_INET6 rcu_read_lock rcu_read_lock &rq->__lock irq_context: 0 rcu_read_lock fnhe_lock irq_context: 0 sk_lock-AF_INET6 k-sk_lock-AF_INET6 rcu_read_lock rcu_node_0 irq_context: 0 sk_lock-AF_TIPC k-slock-AF_TIPC &meta->lock irq_context: 0 sk_lock-AF_TIPC k-slock-AF_TIPC kfence_freelist_lock irq_context: 0 &mm->mmap_lock rcu_read_lock rcu_read_lock &cfs_rq->removed.lock irq_context: 0 &mm->mmap_lock rcu_read_lock rcu_read_lock &obj_hash[i].lock irq_context: 0 &mm->mmap_lock rcu_read_lock rcu_read_lock pool_lock#2 irq_context: 0 (wq_completion)rcu_gp (work_completion)(&(&ssp->srcu_sup->work)->work) rcu_read_lock &pool->lock fill_pool_map-wait-type-override &c->lock irq_context: 0 (wq_completion)bond5 (work_completion)(&(&bond->alb_work)->work) &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 sk_lock-AF_TIPC rcu_read_lock &cfs_rq->removed.lock irq_context: 0 sk_lock-AF_TIPC rcu_read_lock &obj_hash[i].lock irq_context: 0 &u->peer_wait &p->pi_lock irq_context: 0 &u->peer_wait &p->pi_lock &rq->__lock irq_context: 0 &u->peer_wait &p->pi_lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)wg-kex-wg2#11 (work_completion)(&peer->transmit_handshake_work) batched_entropy_u8.lock crngs.lock irq_context: 0 &f->f_pos_lock sb_writers#4 &sb->s_type->i_mutex_key#8 rcu_read_lock rcu_read_lock &rq->__lock &cfs_rq->removed.lock irq_context: 0 &nft_net->commit_mutex defrag4_mutex irq_context: 0 kn->active#61 fs_reclaim irq_context: 0 kn->active#61 fs_reclaim mmu_notifier_invalidate_range_start irq_context: 0 kn->active#61 stock_lock irq_context: 0 kn->active#61 &kernfs_locks->open_file_mutex[count] irq_context: 0 kn->active#61 &kernfs_locks->open_file_mutex[count] fs_reclaim irq_context: 0 kn->active#61 &kernfs_locks->open_file_mutex[count] fs_reclaim mmu_notifier_invalidate_range_start irq_context: 0 kn->active#61 &kernfs_locks->open_file_mutex[count] fs_reclaim mmu_notifier_invalidate_range_start &rq->__lock irq_context: 0 kn->active#61 &kernfs_locks->open_file_mutex[count] fs_reclaim mmu_notifier_invalidate_range_start &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 kn->active#61 &kernfs_locks->open_file_mutex[count] &c->lock irq_context: 0 kn->active#62 &rq->__lock irq_context: 0 kn->active#62 &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 kn->active#62 fs_reclaim irq_context: 0 kn->active#62 fs_reclaim &rq->__lock irq_context: 0 kn->active#62 fs_reclaim &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 kn->active#62 fs_reclaim mmu_notifier_invalidate_range_start irq_context: 0 kn->active#62 fs_reclaim mmu_notifier_invalidate_range_start &rq->__lock irq_context: 0 kn->active#62 fs_reclaim mmu_notifier_invalidate_range_start &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 kn->active#62 &c->lock irq_context: 0 kn->active#62 stock_lock irq_context: 0 kn->active#62 &kernfs_locks->open_file_mutex[count] irq_context: 0 kn->active#62 &kernfs_locks->open_file_mutex[count] fs_reclaim irq_context: 0 kn->active#62 &kernfs_locks->open_file_mutex[count] fs_reclaim mmu_notifier_invalidate_range_start irq_context: 0 kn->active#62 &kernfs_locks->open_file_mutex[count] &c->lock irq_context: 0 sb_writers#11 stock_lock irq_context: 0 sb_writers#11 pool_lock#2 irq_context: 0 sb_writers#11 &____s->seqcount irq_context: 0 sb_writers#11 &p->lock irq_context: 0 sb_writers#11 &p->lock fs_reclaim irq_context: 0 sb_writers#11 &p->lock fs_reclaim &rq->__lock irq_context: 0 sb_writers#11 &p->lock fs_reclaim &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 sb_writers#11 &p->lock fs_reclaim mmu_notifier_invalidate_range_start irq_context: 0 sb_writers#11 &p->lock stock_lock irq_context: 0 sb_writers#11 &p->lock &c->lock irq_context: 0 sb_writers#11 &p->lock pool_lock#2 irq_context: 0 sb_writers#11 &p->lock &of->mutex irq_context: 0 sb_writers#11 &of->mutex cgroup_mutex cpu_hotplug_lock css_set_lock irq_context: 0 sb_writers#11 &of->mutex cgroup_mutex cpu_hotplug_lock fs_reclaim irq_context: 0 sb_writers#11 &of->mutex cgroup_mutex cpu_hotplug_lock fs_reclaim mmu_notifier_invalidate_range_start irq_context: 0 sb_writers#11 &of->mutex cgroup_mutex cpu_hotplug_lock fs_reclaim mmu_notifier_invalidate_range_start &rq->__lock irq_context: 0 sb_writers#11 &of->mutex cgroup_mutex cpu_hotplug_lock fs_reclaim mmu_notifier_invalidate_range_start &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 sb_writers#11 &of->mutex cgroup_mutex cpu_hotplug_lock pool_lock#2 irq_context: 0 sb_writers#11 &of->mutex cgroup_mutex cpu_hotplug_lock &c->lock irq_context: 0 sb_writers#11 &of->mutex cgroup_mutex cpu_hotplug_lock &____s->seqcount#2 irq_context: 0 sb_writers#11 &of->mutex cgroup_mutex cpu_hotplug_lock &rq->__lock irq_context: 0 sb_writers#11 &of->mutex cgroup_mutex cpu_hotplug_lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 sb_writers#11 &of->mutex cgroup_mutex cpu_hotplug_lock &____s->seqcount irq_context: 0 sb_writers#11 &of->mutex cgroup_mutex cpu_hotplug_lock cpuset_mutex irq_context: 0 sb_writers#11 &of->mutex cgroup_mutex cpu_hotplug_lock css_set_lock cgroup_file_kn_lock irq_context: 0 sb_writers#11 &of->mutex cgroup_mutex cpu_hotplug_lock css_set_lock &p->pi_lock irq_context: 0 sb_writers#11 &of->mutex cgroup_mutex cpu_hotplug_lock css_set_lock &p->pi_lock &rq->__lock irq_context: 0 sb_writers#11 &of->mutex cgroup_mutex cpu_hotplug_lock css_set_lock &p->pi_lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 sb_writers#11 &of->mutex cgroup_mutex cpu_hotplug_lock cpuset_mutex &p->pi_lock irq_context: 0 sb_writers#11 &of->mutex cgroup_mutex cpu_hotplug_lock cpuset_mutex &p->pi_lock &rq->__lock irq_context: 0 sb_writers#11 &of->mutex cgroup_mutex cpu_hotplug_lock cpuset_mutex &p->pi_lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 sb_writers#11 &of->mutex cgroup_mutex cpu_hotplug_lock cpuset_mutex &p->alloc_lock irq_context: 0 sb_writers#11 &of->mutex cgroup_mutex cpu_hotplug_lock cpuset_mutex &p->alloc_lock &____s->seqcount#2 irq_context: 0 sb_writers#11 &of->mutex cgroup_mutex cpu_hotplug_lock cpuset_mutex cpuset_attach_wq.lock irq_context: 0 sb_writers#11 &of->mutex cgroup_mutex &wq->mutex &rq->__lock irq_context: 0 sb_writers#11 &of->mutex cgroup_mutex &wq->mutex &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 sb_writers#11 rcu_read_lock pool_lock#2 irq_context: 0 sb_writers#11 tk_core.seq.seqcount irq_context: 0 sb_writers#11 sb_writers#11 mount_lock irq_context: 0 sb_writers#11 sb_writers#11 tk_core.seq.seqcount irq_context: 0 sb_writers#11 sb_writers#11 &sb->s_type->i_lock_key#31 irq_context: 0 sb_writers#11 sb_writers#11 &wb->list_lock irq_context: 0 sb_writers#11 sb_writers#11 &wb->list_lock &sb->s_type->i_lock_key#31 irq_context: 0 sb_writers#11 &p->lock per_cpu_ptr(&cgroup_rstat_cpu_lock, cpu) irq_context: 0 sb_writers#11 &p->lock &n->list_lock irq_context: 0 sb_writers#11 &p->lock &n->list_lock &c->lock irq_context: 0 sb_writers#11 &p->lock &rq->__lock irq_context: 0 sb_writers#11 &p->lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 sb_writers#11 &p->lock &of->mutex kn->active#62 &cgrp->pidlist_mutex irq_context: 0 sb_writers#11 &p->lock &of->mutex kn->active#62 &cgrp->pidlist_mutex &rq->__lock irq_context: 0 sb_writers#11 &p->lock &of->mutex kn->active#62 &cgrp->pidlist_mutex &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 sb_writers#11 &p->lock &of->mutex kn->active#62 &cgrp->pidlist_mutex css_set_lock irq_context: 0 sb_writers#11 &p->lock &of->mutex kn->active#62 &cgrp->pidlist_mutex fs_reclaim irq_context: 0 sb_writers#11 &p->lock &of->mutex kn->active#62 &cgrp->pidlist_mutex fs_reclaim mmu_notifier_invalidate_range_start irq_context: 0 sb_writers#11 &p->lock &of->mutex kn->active#62 &cgrp->pidlist_mutex pool_lock#2 irq_context: 0 sb_writers#11 &p->lock &of->mutex kn->active#62 &cgrp->pidlist_mutex &obj_hash[i].lock irq_context: 0 sb_writers#11 &p->lock &of->mutex kn->active#62 &cgrp->pidlist_mutex &base->lock irq_context: 0 sb_writers#11 &p->lock &of->mutex kn->active#62 &cgrp->pidlist_mutex &base->lock &obj_hash[i].lock irq_context: 0 sb_writers#11 &of->mutex kn->active#61 cpu_hotplug_lock irq_context: 0 sb_writers#11 &of->mutex kn->active#61 cpu_hotplug_lock cpuset_mutex irq_context: 0 sb_writers#11 &of->mutex kn->active#61 cpu_hotplug_lock cpuset_mutex fs_reclaim irq_context: 0 sb_writers#11 &of->mutex kn->active#61 cpu_hotplug_lock cpuset_mutex fs_reclaim mmu_notifier_invalidate_range_start irq_context: 0 sb_writers#11 &of->mutex kn->active#61 cpu_hotplug_lock cpuset_mutex pool_lock#2 irq_context: 0 sb_writers#11 &of->mutex kn->active#61 cpu_hotplug_lock cpuset_mutex callback_lock irq_context: 0 sb_writers#11 &of->mutex kn->active#61 cpu_hotplug_lock cpuset_mutex css_set_lock irq_context: 0 sb_writers#11 &of->mutex kn->active#61 cpu_hotplug_lock cpuset_mutex &obj_hash[i].lock irq_context: 0 sb_writers#11 &type->i_mutex_dir_key#7/1 &____s->seqcount#2 irq_context: 0 sb_writers#11 &type->i_mutex_dir_key#7/1 &____s->seqcount irq_context: 0 &sb->s_type->i_mutex_key#10 (netlink_chain).rwsem &nft_net->commit_mutex defrag4_mutex irq_context: 0 &sb->s_type->i_mutex_key#10 (netlink_chain).rwsem &nft_net->commit_mutex defrag4_mutex &rq->__lock irq_context: 0 &sb->s_type->i_mutex_key#10 (netlink_chain).rwsem &nft_net->commit_mutex defrag4_mutex &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 css_set_lock cgroup_file_kn_lock irq_context: softirq rcu_callback put_task_map-wait-type-override css_set_lock irq_context: softirq rcu_callback put_task_map-wait-type-override css_set_lock &obj_hash[i].lock irq_context: softirq rcu_callback put_task_map-wait-type-override css_set_lock pool_lock#2 irq_context: softirq rcu_callback put_task_map-wait-type-override css_set_lock krc.lock irq_context: 0 event_mutex irq_context: 0 event_mutex sched_register_mutex irq_context: 0 event_mutex sched_register_mutex tracepoints_mutex irq_context: 0 event_mutex sched_register_mutex tracepoints_mutex fs_reclaim irq_context: 0 event_mutex sched_register_mutex tracepoints_mutex fs_reclaim mmu_notifier_invalidate_range_start irq_context: 0 event_mutex sched_register_mutex tracepoints_mutex pool_lock#2 irq_context: 0 event_mutex sched_register_mutex tracepoints_mutex cpu_hotplug_lock irq_context: 0 event_mutex sched_register_mutex tracepoints_mutex cpu_hotplug_lock static_call_mutex irq_context: 0 event_mutex sched_register_mutex tracepoints_mutex cpu_hotplug_lock static_call_mutex text_mutex irq_context: 0 event_mutex sched_register_mutex tracepoints_mutex cpu_hotplug_lock static_call_mutex text_mutex ptlock_ptr(page)#2 irq_context: 0 event_mutex sched_register_mutex tracepoints_mutex cpu_hotplug_lock jump_label_mutex irq_context: 0 event_mutex sched_register_mutex tracepoints_mutex cpu_hotplug_lock jump_label_mutex text_mutex irq_context: 0 event_mutex sched_register_mutex tracepoints_mutex cpu_hotplug_lock jump_label_mutex text_mutex ptlock_ptr(page)#2 irq_context: 0 event_mutex tracepoints_mutex irq_context: 0 event_mutex tracepoints_mutex fs_reclaim irq_context: 0 event_mutex tracepoints_mutex fs_reclaim mmu_notifier_invalidate_range_start irq_context: 0 event_mutex tracepoints_mutex pool_lock#2 irq_context: 0 event_mutex tracepoints_mutex cpu_hotplug_lock irq_context: 0 event_mutex tracepoints_mutex cpu_hotplug_lock static_call_mutex irq_context: 0 event_mutex tracepoints_mutex cpu_hotplug_lock static_call_mutex text_mutex irq_context: 0 event_mutex tracepoints_mutex cpu_hotplug_lock static_call_mutex text_mutex ptlock_ptr(page)#2 irq_context: 0 event_mutex tracepoints_mutex cpu_hotplug_lock jump_label_mutex irq_context: 0 event_mutex tracepoints_mutex cpu_hotplug_lock jump_label_mutex text_mutex irq_context: 0 event_mutex tracepoints_mutex cpu_hotplug_lock jump_label_mutex text_mutex ptlock_ptr(page)#2 irq_context: 0 kn->active#61 &c->lock irq_context: 0 kn->active#61 &rq->__lock irq_context: 0 kn->active#61 &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 kn->active#61 rcu_read_lock rcu_node_0 irq_context: 0 kn->active#61 rcu_read_lock &rq->__lock irq_context: 0 kn->active#61 &n->list_lock irq_context: 0 kn->active#61 &n->list_lock &c->lock irq_context: 0 kn->active#61 rcu_read_lock rcu_read_lock rcu_node_0 irq_context: 0 kn->active#61 rcu_read_lock rcu_read_lock &rq->__lock irq_context: 0 kn->active#61 rcu_read_lock rcu_read_lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 sb_writers#4 &sb->s_type->i_mutex_key#8 jbd2_handle rcu_read_lock &rq->__lock &cfs_rq->removed.lock irq_context: 0 sb_writers#4 &sb->s_type->i_mutex_key#8 jbd2_handle rcu_read_lock &cfs_rq->removed.lock irq_context: 0 sb_writers#4 &sb->s_type->i_mutex_key#8 jbd2_handle rcu_read_lock &obj_hash[i].lock irq_context: 0 &sb->s_type->i_mutex_key#10 fill_pool_map-wait-type-override rcu_read_lock rcu_node_0 irq_context: 0 &sb->s_type->i_mutex_key#10 fill_pool_map-wait-type-override rcu_read_lock &rq->__lock irq_context: 0 &sb->s_type->i_mutex_key#10 fill_pool_map-wait-type-override rcu_read_lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 rtnl_mutex &idev->mc_lock &macvlan_netdev_addr_lock_key &n->list_lock irq_context: 0 rtnl_mutex &idev->mc_lock &macvlan_netdev_addr_lock_key &n->list_lock &c->lock irq_context: 0 kn->active#61 &____s->seqcount#2 irq_context: 0 kn->active#61 &____s->seqcount irq_context: 0 &sb->s_type->i_mutex_key#10 (netlink_chain).rwsem &nft_net->commit_mutex &rq->__lock irq_context: 0 &sb->s_type->i_mutex_key#10 (netlink_chain).rwsem &nft_net->commit_mutex &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 rtnl_mutex &idev->mc_lock fill_pool_map-wait-type-override &n->list_lock irq_context: 0 rtnl_mutex &idev->mc_lock fill_pool_map-wait-type-override &n->list_lock &c->lock irq_context: 0 rtnl_mutex rcu_read_lock &pool->lock/1 fill_pool_map-wait-type-override &n->list_lock irq_context: 0 rtnl_mutex rcu_read_lock &pool->lock/1 fill_pool_map-wait-type-override &n->list_lock &c->lock irq_context: 0 (wq_completion)mld (work_completion)(&(&idev->mc_ifc_work)->work) &idev->mc_lock rcu_read_lock rcu_read_lock rcu_read_lock_bh rcu_read_lock batched_entropy_u8.lock irq_context: 0 fill_pool_map-wait-type-override rcu_read_lock &rcu_state.gp_wq irq_context: 0 fill_pool_map-wait-type-override rcu_read_lock &rcu_state.gp_wq &p->pi_lock irq_context: 0 fill_pool_map-wait-type-override rcu_read_lock &rcu_state.gp_wq &p->pi_lock &rq->__lock irq_context: 0 fill_pool_map-wait-type-override rcu_read_lock &rcu_state.gp_wq &p->pi_lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 sk_lock-AF_INET6 rcu_read_lock rcu_read_lock_bh rcu_read_lock &n->list_lock irq_context: 0 sk_lock-AF_INET6 rcu_read_lock rcu_read_lock_bh rcu_read_lock &n->list_lock &c->lock irq_context: softirq &(&l->destroy_dwork)->timer irq_context: softirq &(&l->destroy_dwork)->timer rcu_read_lock &pool->lock irq_context: softirq &(&l->destroy_dwork)->timer rcu_read_lock &pool->lock &obj_hash[i].lock irq_context: softirq &(&l->destroy_dwork)->timer rcu_read_lock &pool->lock &p->pi_lock irq_context: softirq &(&l->destroy_dwork)->timer rcu_read_lock &pool->lock &p->pi_lock &rq->__lock irq_context: softirq &(&l->destroy_dwork)->timer rcu_read_lock &pool->lock &p->pi_lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)cgroup_pidlist_destroy irq_context: 0 (wq_completion)cgroup_pidlist_destroy (work_completion)(&(&l->destroy_dwork)->work) irq_context: 0 (wq_completion)cgroup_pidlist_destroy (work_completion)(&(&l->destroy_dwork)->work) &cgrp->pidlist_mutex irq_context: 0 (wq_completion)cgroup_pidlist_destroy (work_completion)(&(&l->destroy_dwork)->work) &cgrp->pidlist_mutex &obj_hash[i].lock irq_context: 0 (wq_completion)cgroup_pidlist_destroy (work_completion)(&(&l->destroy_dwork)->work) &cgrp->pidlist_mutex pool_lock#2 irq_context: 0 (wq_completion)cgroup_pidlist_destroy (work_completion)(&(&l->destroy_dwork)->work) &obj_hash[i].lock irq_context: 0 event_mutex &rq->__lock irq_context: 0 event_mutex &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 sk_lock-AF_INET6 rcu_read_lock rcu_read_lock_bh rcu_read_lock &____s->seqcount#2 irq_context: 0 sk_lock-AF_INET6 rcu_read_lock rcu_read_lock_bh rcu_read_lock &____s->seqcount irq_context: 0 &sb->s_type->i_mutex_key#10 pool_lock irq_context: 0 pernet_ops_rwsem rtnl_mutex rcu_state.exp_mutex &cfs_rq->removed.lock irq_context: 0 pernet_ops_rwsem rtnl_mutex rcu_state.exp_mutex pool_lock#2 irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem uevent_sock_mutex.wait_lock irq_context: 0 (wq_completion)writeback &rq->__lock irq_context: 0 pernet_ops_rwsem __ip_vs_app_mutex fs_reclaim mmu_notifier_invalidate_range_start &rq->__lock irq_context: 0 pernet_ops_rwsem __ip_vs_app_mutex fs_reclaim mmu_notifier_invalidate_range_start &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 &sb->s_type->i_mutex_key#8 &sbi->s_writepages_rwsem jbd2_handle &ei->i_data_sem &rq->__lock &cfs_rq->removed.lock irq_context: 0 sb_writers#4 &sb->s_type->i_mutex_key#8 rcu_read_lock &n->list_lock &c->lock irq_context: 0 sb_writers#4 sb_internal jbd2_handle &ret->b_state_lock bit_wait_table + i irq_context: 0 sb_writers#4 &sb->s_type->i_mutex_key#8 jbd2_handle &rq->__lock irq_context: 0 mapping.invalidate_lock rcu_read_lock &c->lock irq_context: 0 kn->active#60 remove_cache_srcu irq_context: 0 kn->active#60 remove_cache_srcu quarantine_lock irq_context: 0 kn->active#60 remove_cache_srcu &c->lock irq_context: 0 kn->active#60 remove_cache_srcu &rq->__lock irq_context: 0 kn->active#60 remove_cache_srcu &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 kn->active#60 remove_cache_srcu &n->list_lock irq_context: 0 kn->active#60 remove_cache_srcu &obj_hash[i].lock irq_context: 0 sk_lock-AF_INET6 remove_cache_srcu &cfs_rq->removed.lock irq_context: 0 sk_lock-AF_INET6 remove_cache_srcu &obj_hash[i].lock pool_lock irq_context: 0 sb_writers#4 &sb->s_type->i_mutex_key#8 pool_lock irq_context: 0 &mm->mmap_lock sb_pagefaults mapping.invalidate_lock per_cpu_ptr(&cgroup_rstat_cpu_lock, cpu) irq_context: 0 sb_writers#4 mapping.invalidate_lock &base->lock irq_context: 0 sb_writers#4 mapping.invalidate_lock &base->lock &obj_hash[i].lock irq_context: 0 &sb->s_type->i_mutex_key#8 &sbi->s_writepages_rwsem &rcu_state.expedited_wq irq_context: 0 &sb->s_type->i_mutex_key#8 &sbi->s_writepages_rwsem &rcu_state.expedited_wq &p->pi_lock irq_context: 0 &sb->s_type->i_mutex_key#8 &sbi->s_writepages_rwsem &rcu_state.expedited_wq &p->pi_lock &rq->__lock irq_context: 0 &sb->s_type->i_mutex_key#8 &sbi->s_writepages_rwsem &rcu_state.expedited_wq &p->pi_lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 sb_writers#4 mapping.invalidate_lock rcu_read_lock fill_pool_map-wait-type-override rcu_node_0 irq_context: 0 sb_writers#4 mapping.invalidate_lock rcu_read_lock fill_pool_map-wait-type-override &rq->__lock irq_context: 0 sb_writers#4 mapping.invalidate_lock rcu_read_lock fill_pool_map-wait-type-override &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: softirq (&ndev->rs_timer) rcu_read_lock rcu_read_lock rcu_read_lock_bh rcu_read_lock rcu_read_lock rcu_read_lock rcu_read_lock rcu_read_lock &pool->lock irq_context: softirq (&ndev->rs_timer) rcu_read_lock rcu_read_lock rcu_read_lock_bh rcu_read_lock rcu_read_lock rcu_read_lock rcu_read_lock rcu_read_lock &pool->lock &obj_hash[i].lock irq_context: softirq (&ndev->rs_timer) rcu_read_lock rcu_read_lock rcu_read_lock_bh rcu_read_lock rcu_read_lock rcu_read_lock rcu_read_lock rcu_read_lock &pool->lock pool_lock#2 irq_context: softirq (&ndev->rs_timer) rcu_read_lock rcu_read_lock rcu_read_lock_bh rcu_read_lock rcu_read_lock rcu_read_lock rcu_read_lock rcu_read_lock &pool->lock &p->pi_lock irq_context: softirq (&ndev->rs_timer) rcu_read_lock rcu_read_lock rcu_read_lock_bh rcu_read_lock rcu_read_lock rcu_read_lock rcu_read_lock rcu_read_lock &pool->lock &p->pi_lock &rq->__lock irq_context: softirq (&ndev->rs_timer) rcu_read_lock rcu_read_lock rcu_read_lock_bh rcu_read_lock rcu_read_lock rcu_read_lock rcu_read_lock rcu_read_lock &pool->lock &p->pi_lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)wg-kex-wg2#11 (work_completion)(&peer->transmit_handshake_work) &wg->static_identity.lock &handshake->lock crngs.lock base_crng.lock irq_context: 0 (wq_completion)events (work_completion)(&work->work)#3 irq_context: 0 (wq_completion)events (work_completion)(&work->work)#3 &rq->__lock irq_context: 0 (wq_completion)events (work_completion)(&work->work)#3 &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)events (work_completion)(&work->work)#3 rcu_read_lock &ndev->lock irq_context: 0 (wq_completion)events (work_completion)(&work->work)#3 pool_lock#2 irq_context: 0 (wq_completion)events (work_completion)(&work->work)#3 &c->lock irq_context: 0 (wq_completion)events (work_completion)(&work->work)#3 &dir->lock#2 irq_context: 0 (wq_completion)events (work_completion)(&work->work)#3 &ul->lock#2 irq_context: 0 (wq_completion)events (work_completion)(&work->work)#3 rcu_read_lock rcu_read_lock &____s->seqcount#10 irq_context: 0 (wq_completion)events (work_completion)(&work->work)#3 rcu_read_lock rcu_read_lock rcu_read_lock_bh &obj_hash[i].lock irq_context: 0 (wq_completion)events (work_completion)(&work->work)#3 rcu_read_lock rcu_read_lock rcu_read_lock_bh pool_lock#2 irq_context: 0 (wq_completion)events (work_completion)(&work->work)#3 &obj_hash[i].lock irq_context: 0 (wq_completion)wg-crypt-wg2#4 (work_completion)(&peer->transmit_packet_work) batched_entropy_u8.lock crngs.lock irq_context: 0 (wq_completion)wg-kex-wg2#7 (work_completion)(&peer->transmit_handshake_work) &peer->endpoint_lock rcu_read_lock_bh rcu_read_lock &c->lock irq_context: 0 (wq_completion)wg-kex-wg2#7 (work_completion)(&peer->transmit_handshake_work) &peer->endpoint_lock rcu_read_lock_bh rcu_read_lock &obj_hash[i].lock irq_context: 0 sk_lock-AF_CAN &mm->mmap_lock irq_context: 0 sk_lock-AF_CAN &mm->mmap_lock &rq->__lock irq_context: 0 sk_lock-AF_CAN &mm->mmap_lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 rtnl_mutex &tb->tb6_lock nl_table_wait.lock &p->pi_lock &rq->__lock irq_context: 0 rtnl_mutex &tb->tb6_lock nl_table_wait.lock &p->pi_lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 sb_writers#4 fs_reclaim mmu_notifier_invalidate_range_start &rq->__lock irq_context: 0 sb_writers#4 fs_reclaim mmu_notifier_invalidate_range_start &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)wg-kex-wg0#7 (work_completion)(&peer->transmit_handshake_work) &peer->endpoint_lock rcu_read_lock_bh rcu_read_lock rcu_read_lock rcu_read_lock &____s->seqcount irq_context: 0 (wq_completion)wg-kex-wg0#7 (work_completion)(&peer->transmit_handshake_work) &peer->endpoint_lock rcu_read_lock_bh rcu_read_lock rcu_read_lock rcu_read_lock pool_lock#2 irq_context: 0 (wq_completion)wg-kex-wg0#7 (work_completion)(&peer->transmit_handshake_work) &peer->endpoint_lock rcu_read_lock_bh rcu_read_lock rcu_read_lock rcu_read_lock rcu_read_lock pool_lock#2 irq_context: 0 (wq_completion)wg-kex-wg0#7 (work_completion)(&peer->transmit_handshake_work) &peer->endpoint_lock rcu_read_lock_bh rcu_read_lock rcu_read_lock rcu_read_lock &obj_hash[i].lock irq_context: 0 (wq_completion)wg-kex-wg0#7 (work_completion)(&peer->transmit_handshake_work) &peer->endpoint_lock rcu_read_lock_bh rcu_read_lock rcu_read_lock &____s->seqcount#7 irq_context: 0 (wq_completion)wg-kex-wg0#7 (work_completion)(&peer->transmit_handshake_work) &peer->endpoint_lock rcu_read_lock_bh rcu_read_lock &nf_nat_locks[i] irq_context: 0 (wq_completion)wg-kex-wg0#7 (work_completion)(&peer->transmit_handshake_work) &peer->endpoint_lock rcu_read_lock_bh rcu_read_lock &n->lock &obj_hash[i].lock irq_context: 0 (wq_completion)wg-kex-wg0#7 (work_completion)(&peer->transmit_handshake_work) &peer->endpoint_lock rcu_read_lock_bh rcu_read_lock &n->lock &base->lock irq_context: 0 (wq_completion)wg-kex-wg0#7 (work_completion)(&peer->transmit_handshake_work) &peer->endpoint_lock rcu_read_lock_bh rcu_read_lock &n->lock &base->lock &obj_hash[i].lock irq_context: 0 (wq_completion)wg-kex-wg0#7 (work_completion)(&peer->transmit_handshake_work) &peer->endpoint_lock rcu_read_lock_bh rcu_read_lock &n->lock pool_lock#2 irq_context: 0 (wq_completion)wg-kex-wg0#7 (work_completion)(&peer->transmit_handshake_work) &peer->endpoint_lock rcu_read_lock_bh rcu_read_lock pool_lock#2 irq_context: 0 (wq_completion)wg-kex-wg0#7 (work_completion)(&peer->transmit_handshake_work) &peer->endpoint_lock rcu_read_lock_bh rcu_read_lock &ul->lock#2 irq_context: 0 (wq_completion)wg-kex-wg0#7 (work_completion)(&peer->transmit_handshake_work) &peer->endpoint_lock rcu_read_lock_bh rcu_read_lock rcu_read_lock rcu_read_lock &tbl->lock irq_context: 0 (wq_completion)wg-kex-wg0#7 (work_completion)(&peer->transmit_handshake_work) &peer->endpoint_lock rcu_read_lock_bh rcu_read_lock rcu_read_lock rcu_read_lock &____s->seqcount#9 irq_context: 0 (wq_completion)wg-kex-wg0#7 (work_completion)(&peer->transmit_handshake_work) &peer->endpoint_lock rcu_read_lock_bh rcu_read_lock rcu_read_lock rcu_read_lock rcu_read_lock_bh &obj_hash[i].lock irq_context: 0 (wq_completion)wg-kex-wg0#7 (work_completion)(&peer->transmit_handshake_work) &peer->endpoint_lock rcu_read_lock_bh rcu_read_lock rcu_read_lock rcu_read_lock rcu_read_lock_bh pool_lock#2 irq_context: 0 sb_writers#4 &sb->s_type->i_mutex_key#8 rcu_read_lock &base->lock irq_context: 0 sb_writers#4 &sb->s_type->i_mutex_key#8 rcu_read_lock &base->lock &obj_hash[i].lock irq_context: 0 rtnl_mutex (inet6addr_validator_chain).rwsem &rq->__lock irq_context: 0 rtnl_mutex (inet6addr_validator_chain).rwsem &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 rtnl_mutex &ndev->lock &ifa->lock rcu_read_lock &pool->lock pool_lock#2 irq_context: 0 rtnl_mutex &ndev->lock &ifa->lock rcu_read_lock &pool->lock &p->pi_lock irq_context: 0 rtnl_mutex &ndev->lock &ifa->lock rcu_read_lock &pool->lock &p->pi_lock &rq->__lock irq_context: 0 rtnl_mutex &ndev->lock &ifa->lock rcu_read_lock &pool->lock &p->pi_lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)events deferred_process_work &p->pi_lock &cfs_rq->removed.lock irq_context: 0 (wq_completion)ipv6_addrconf (work_completion)(&(&ifa->dad_work)->work) rtnl_mutex &idev->mc_lock rcu_read_lock rcu_read_lock &n->list_lock irq_context: 0 (wq_completion)ipv6_addrconf (work_completion)(&(&ifa->dad_work)->work) rtnl_mutex &idev->mc_lock rcu_read_lock rcu_read_lock &n->list_lock &c->lock irq_context: 0 rtnl_mutex rcu_read_lock &rq->__lock &cfs_rq->removed.lock irq_context: 0 (wq_completion)ipv6_addrconf (work_completion)(&(&ifa->dad_work)->work) rtnl_mutex rcu_read_lock rcu_read_lock rcu_read_lock rcu_node_0 irq_context: 0 (wq_completion)ipv6_addrconf (work_completion)(&(&ifa->dad_work)->work) rtnl_mutex rcu_read_lock rcu_read_lock rcu_read_lock &rq->__lock irq_context: 0 sk_lock-AF_INET cpu_hotplug_lock irq_context: 0 sk_lock-AF_INET cpu_hotplug_lock jump_label_mutex irq_context: 0 sk_lock-AF_INET cpu_hotplug_lock jump_label_mutex text_mutex irq_context: 0 sk_lock-AF_INET cpu_hotplug_lock jump_label_mutex text_mutex ptlock_ptr(page)#2 irq_context: 0 (wq_completion)mld (work_completion)(&(&idev->mc_ifc_work)->work) &idev->mc_lock rcu_read_lock rcu_read_lock rcu_read_lock_bh rcu_read_lock rcu_read_lock_bh rcu_read_lock &c->lock irq_context: 0 &sb->s_type->i_mutex_key#10 sk_lock-AF_INET slock-AF_INET krc.lock irq_context: softirq (&n->timer) rcu_read_lock rcu_read_lock &____s->seqcount#9 irq_context: softirq (&n->timer) rcu_read_lock rcu_read_lock rcu_read_lock_bh rcu_read_lock pool_lock#2 irq_context: softirq (&n->timer) rcu_read_lock rcu_read_lock rcu_read_lock_bh rcu_read_lock &obj_hash[i].lock irq_context: 0 &f->f_pos_lock sb_writers#4 &sb->s_type->i_mutex_key#8 remove_cache_srcu &meta->lock irq_context: 0 &f->f_pos_lock sb_writers#4 &sb->s_type->i_mutex_key#8 remove_cache_srcu kfence_freelist_lock irq_context: 0 (wq_completion)mld (work_completion)(&(&idev->mc_dad_work)->work) &idev->mc_lock rcu_read_lock rcu_read_lock rcu_read_lock_bh quarantine_lock irq_context: 0 &sb->s_type->i_mutex_key#8 &meta->lock irq_context: 0 sb_writers#4 &type->i_mutex_dir_key#3 tomoyo_ss mmu_notifier_invalidate_range_start &rq->__lock irq_context: 0 sb_writers#4 &type->i_mutex_dir_key#3 tomoyo_ss mmu_notifier_invalidate_range_start &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 &sb->s_type->i_mutex_key#8 &sbi->s_writepages_rwsem rcu_read_lock &cfs_rq->removed.lock irq_context: 0 &f->f_pos_lock sb_writers#4 &sb->s_type->i_mutex_key#8 &ei->i_data_sem &ei->i_es_lock key#7 irq_context: 0 sb_writers#5 &sb->s_type->i_mutex_key#12 fs_reclaim &rq->__lock irq_context: 0 sb_writers#5 &sb->s_type->i_mutex_key#12 fs_reclaim &rq->__lock &cfs_rq->removed.lock irq_context: 0 sb_writers#5 &sb->s_type->i_mutex_key#12 fs_reclaim &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 sk_lock-AF_INET6 rcu_read_lock rcu_read_lock_bh rcu_read_lock rlock-AF_PACKET irq_context: 0 sock_diag_mutex sock_diag_table_mutex nlk_cb_mutex-SOCK_DIAG inet_diag_table_mutex &ht->lock irq_context: 0 sock_diag_mutex sock_diag_table_mutex nlk_cb_mutex-SOCK_DIAG inet_diag_table_mutex rcu_read_lock &ht->lock irq_context: 0 sock_diag_mutex sock_diag_table_mutex nlk_cb_mutex-SOCK_DIAG rcu_read_lock rcu_read_lock_bh rcu_read_lock &c->lock irq_context: 0 sock_diag_mutex sock_diag_table_mutex nlk_cb_mutex-SOCK_DIAG rcu_read_lock rcu_read_lock_bh rcu_read_lock pool_lock#2 irq_context: 0 sock_diag_mutex sock_diag_table_mutex nlk_cb_mutex-SOCK_DIAG rcu_read_lock rcu_read_lock_bh rcu_read_lock tk_core.seq.seqcount irq_context: 0 sock_diag_mutex sock_diag_table_mutex nlk_cb_mutex-SOCK_DIAG rcu_read_lock rcu_read_lock_bh rcu_read_lock &obj_hash[i].lock irq_context: 0 sock_diag_mutex sock_diag_table_mutex nlk_cb_mutex-SOCK_DIAG rcu_read_lock rcu_node_0 irq_context: 0 sock_diag_mutex sock_diag_table_mutex nlk_cb_mutex-SOCK_DIAG rcu_read_lock &rq->__lock irq_context: 0 sock_diag_mutex sock_diag_table_mutex nlk_cb_mutex-SOCK_DIAG rcu_read_lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 sock_diag_mutex sock_diag_table_mutex nlk_cb_mutex-SOCK_DIAG rcu_read_lock &ei->socket.wq.wait irq_context: 0 sock_diag_mutex sock_diag_table_mutex nlk_cb_mutex-SOCK_DIAG rcu_read_lock &ei->socket.wq.wait &p->pi_lock irq_context: 0 sock_diag_mutex sock_diag_table_mutex nlk_cb_mutex-SOCK_DIAG rcu_read_lock &ei->socket.wq.wait &p->pi_lock &rq->__lock irq_context: 0 sock_diag_mutex sock_diag_table_mutex nlk_cb_mutex-SOCK_DIAG rcu_read_lock &ei->socket.wq.wait &p->pi_lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 nlk_cb_mutex-SOCK_DIAG irq_context: 0 nlk_cb_mutex-SOCK_DIAG &lock->wait_lock irq_context: 0 nlk_cb_mutex-SOCK_DIAG &rq->__lock irq_context: 0 nlk_cb_mutex-SOCK_DIAG &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 sock_diag_mutex sock_diag_table_mutex &lock->wait_lock irq_context: 0 sock_diag_mutex sock_diag_table_mutex &p->pi_lock irq_context: 0 sock_diag_mutex sock_diag_table_mutex &p->pi_lock &rq->__lock irq_context: 0 sock_diag_mutex sock_diag_table_mutex &p->pi_lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 sock_diag_mutex sock_diag_table_mutex &rq->__lock irq_context: 0 sock_diag_mutex sock_diag_table_mutex &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 sk_lock-AF_INET rcu_read_lock &nf_conntrack_locks[i] &nf_conntrack_locks[i]/1 irq_context: 0 sk_lock-AF_INET rcu_read_lock &nf_conntrack_locks[i] &nf_conntrack_locks[i]/1 batched_entropy_u8.lock irq_context: 0 sk_lock-AF_INET rcu_read_lock rcu_read_lock_bh rcu_read_lock &c->lock irq_context: softirq rcu_read_lock rcu_read_lock k-slock-AF_INET rcu_read_lock rcu_read_lock_bh rcu_read_lock &obj_hash[i].lock irq_context: 0 sk_lock-AF_KCM batched_entropy_u8.lock irq_context: 0 sk_lock-AF_KCM kfence_freelist_lock irq_context: 0 &sb->s_type->i_mutex_key#10 sk_lock-AF_INET6/1 &rq->__lock &cfs_rq->removed.lock irq_context: softirq (&n->timer) k-slock-AF_INET6 rcu_read_lock rcu_read_lock rcu_read_lock_bh rcu_read_lock &c->lock irq_context: softirq (&n->timer) k-slock-AF_INET6 rcu_read_lock &n->list_lock irq_context: softirq (&n->timer) k-slock-AF_INET6 rcu_read_lock &n->list_lock &c->lock irq_context: softirq (&n->timer) k-slock-AF_INET6 rcu_read_lock rcu_read_lock rcu_read_lock_bh rcu_read_lock batched_entropy_u8.lock irq_context: softirq (&n->timer) k-slock-AF_INET6 rcu_read_lock rcu_read_lock rcu_read_lock_bh rcu_read_lock kfence_freelist_lock irq_context: softirq (&n->timer) k-slock-AF_INET6 rcu_read_lock rcu_read_lock rcu_read_lock_bh rcu_read_lock &meta->lock irq_context: 0 sb_writers#4 sb_internal jbd2_handle &ei->i_data_sem remove_cache_srcu rcu_read_lock rcu_node_0 irq_context: 0 sb_writers#4 sb_internal jbd2_handle &ei->i_data_sem remove_cache_srcu rcu_read_lock &rq->__lock irq_context: 0 sb_writers#4 sb_internal jbd2_handle &ei->i_data_sem remove_cache_srcu rcu_read_lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 sb_writers#4 sb_internal jbd2_handle &ei->i_data_sem remove_cache_srcu &rq->__lock irq_context: 0 sb_writers#4 sb_internal jbd2_handle &ei->i_data_sem remove_cache_srcu &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)wg-kex-wg2#9 (work_completion)(&peer->transmit_handshake_work) &peer->endpoint_lock rcu_read_lock_bh rcu_read_lock rcu_read_lock &____s->seqcount#7 irq_context: 0 (wq_completion)wg-kex-wg2#9 (work_completion)(&peer->transmit_handshake_work) &peer->endpoint_lock rcu_read_lock_bh rcu_read_lock &nf_nat_locks[i] irq_context: 0 (wq_completion)wg-kex-wg2#9 (work_completion)(&peer->transmit_handshake_work) &peer->endpoint_lock rcu_read_lock_bh rcu_read_lock &nf_conntrack_locks[i] irq_context: 0 (wq_completion)wg-kex-wg2#9 (work_completion)(&peer->transmit_handshake_work) &peer->endpoint_lock rcu_read_lock_bh rcu_read_lock &nf_conntrack_locks[i] &nf_conntrack_locks[i]/1 irq_context: 0 (wq_completion)wg-kex-wg2#9 (work_completion)(&peer->transmit_handshake_work) &peer->endpoint_lock rcu_read_lock_bh rcu_read_lock &nf_conntrack_locks[i] &nf_conntrack_locks[i]/1 batched_entropy_u8.lock irq_context: 0 (wq_completion)wg-kex-wg2#9 (work_completion)(&peer->transmit_handshake_work) &peer->endpoint_lock rcu_read_lock_bh rcu_read_lock &nf_conntrack_locks[i]/1 irq_context: 0 (wq_completion)wg-kex-wg2#9 (work_completion)(&peer->transmit_handshake_work) &peer->endpoint_lock rcu_read_lock_bh rcu_read_lock &n->lock irq_context: 0 (wq_completion)wg-kex-wg2#9 (work_completion)(&peer->transmit_handshake_work) &peer->endpoint_lock rcu_read_lock_bh rcu_read_lock &n->lock &obj_hash[i].lock irq_context: 0 (wq_completion)wg-kex-wg2#9 (work_completion)(&peer->transmit_handshake_work) &peer->endpoint_lock rcu_read_lock_bh rcu_read_lock &n->lock &base->lock irq_context: 0 (wq_completion)wg-kex-wg2#9 (work_completion)(&peer->transmit_handshake_work) &peer->endpoint_lock rcu_read_lock_bh rcu_read_lock &n->lock &base->lock &obj_hash[i].lock irq_context: 0 (wq_completion)wg-kex-wg2#9 (work_completion)(&peer->transmit_handshake_work) &peer->endpoint_lock rcu_read_lock_bh rcu_read_lock &n->lock &c->lock irq_context: 0 (wq_completion)wg-kex-wg2#9 (work_completion)(&peer->transmit_handshake_work) &peer->endpoint_lock rcu_read_lock_bh rcu_read_lock &n->lock &n->list_lock irq_context: 0 (wq_completion)wg-kex-wg2#9 (work_completion)(&peer->transmit_handshake_work) &peer->endpoint_lock rcu_read_lock_bh rcu_read_lock &n->lock &n->list_lock &c->lock irq_context: 0 (wq_completion)wg-kex-wg2#9 (work_completion)(&peer->transmit_handshake_work) &peer->endpoint_lock rcu_read_lock_bh rcu_read_lock &ul->lock#2 irq_context: 0 (wq_completion)wg-kex-wg2#9 (work_completion)(&peer->transmit_handshake_work) &peer->endpoint_lock rcu_read_lock_bh rcu_read_lock rcu_read_lock rcu_read_lock &____s->seqcount#9 irq_context: 0 (wq_completion)wg-kex-wg2#9 (work_completion)(&peer->transmit_handshake_work) &peer->endpoint_lock rcu_read_lock_bh rcu_read_lock rcu_read_lock rcu_read_lock rcu_read_lock_bh rcu_read_lock pool_lock#2 irq_context: 0 (wq_completion)wg-kex-wg2#9 (work_completion)(&peer->transmit_handshake_work) &peer->endpoint_lock rcu_read_lock_bh rcu_read_lock rcu_read_lock rcu_read_lock rcu_read_lock_bh rcu_read_lock tk_core.seq.seqcount irq_context: 0 (wq_completion)wg-kex-wg2#9 (work_completion)(&peer->transmit_handshake_work) &peer->endpoint_lock rcu_read_lock_bh rcu_read_lock rcu_read_lock rcu_read_lock rcu_read_lock_bh rcu_read_lock &obj_hash[i].lock irq_context: 0 (wq_completion)wg-kex-wg2#9 (work_completion)(&peer->transmit_handshake_work) &peer->endpoint_lock rcu_read_lock_bh rcu_read_lock rcu_read_lock rcu_read_lock rcu_read_lock_bh &obj_hash[i].lock irq_context: 0 (wq_completion)wg-kex-wg1#9 (work_completion)(&peer->transmit_handshake_work) &peer->endpoint_lock rcu_read_lock_bh rcu_read_lock &tbl->lock irq_context: 0 (wq_completion)wg-kex-wg1#9 (work_completion)(&peer->transmit_handshake_work) &peer->endpoint_lock rcu_read_lock_bh rcu_read_lock &n->lock irq_context: 0 (wq_completion)wg-kex-wg1#9 (work_completion)(&peer->transmit_handshake_work) &peer->endpoint_lock rcu_read_lock_bh rcu_read_lock rcu_read_lock_bh noop_qdisc.q.lock irq_context: 0 (wq_completion)wg-kex-wg1#9 (work_completion)(&peer->transmit_handshake_work) &peer->endpoint_lock rcu_read_lock_bh rcu_read_lock rcu_read_lock_bh pool_lock#2 irq_context: 0 &mm->mmap_lock ptlock_ptr(page)#2 rcu_read_lock &p->pi_lock &rq->__lock &cfs_rq->removed.lock irq_context: 0 (wq_completion)events (work_completion)(&w->work)#2 nf_conntrack_mutex rcu_node_0 irq_context: 0 sb_writers#8 &cfs_rq->removed.lock irq_context: 0 rtnl_mutex rcu_read_lock &tb->tb6_lock rcu_read_lock &____s->seqcount#2 irq_context: 0 &sb->s_type->i_mutex_key#10 sk_lock-AF_KCM &meta->lock irq_context: 0 &sb->s_type->i_mutex_key#10 sk_lock-AF_KCM kfence_freelist_lock irq_context: 0 (wq_completion)events (work_completion)(&w->work)#2 nf_conntrack_mutex &rcu_state.expedited_wq irq_context: 0 (wq_completion)events (work_completion)(&w->work)#2 nf_conntrack_mutex &rcu_state.expedited_wq &p->pi_lock irq_context: 0 (wq_completion)events (work_completion)(&w->work)#2 nf_conntrack_mutex &rcu_state.expedited_wq &p->pi_lock &rq->__lock irq_context: 0 (wq_completion)events (work_completion)(&w->work)#2 nf_conntrack_mutex &rcu_state.expedited_wq &p->pi_lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)wg-kex-wg2#3 (work_completion)(&peer->transmit_handshake_work) &peer->endpoint_lock rcu_read_lock_bh rcu_read_lock &c->lock irq_context: 0 (wq_completion)wg-kex-wg2#3 (work_completion)(&peer->transmit_handshake_work) &peer->endpoint_lock rcu_read_lock_bh rcu_read_lock &obj_hash[i].lock irq_context: 0 (wq_completion)wg-kex-wg1#4 (work_completion)(&({ do { const void *__vpp_verify = (typeof((worker) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __asm__ ("" : "=r"(__ptr) : "0"((typeof(*((worker))) *)((worker)))); (typeof((typeof(*((worker))) *)((worker)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); })->work) &wg->static_identity.lock &handshake->lock &rq->__lock irq_context: 0 (wq_completion)wg-kex-wg1#4 (work_completion)(&({ do { const void *__vpp_verify = (typeof((worker) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __asm__ ("" : "=r"(__ptr) : "0"((typeof(*((worker))) *)((worker)))); (typeof((typeof(*((worker))) *)((worker)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); })->work) &wg->static_identity.lock &handshake->lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)wg-kex-wg1#4 (work_completion)(&({ do { const void *__vpp_verify = (typeof((worker) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __asm__ ("" : "=r"(__ptr) : "0"((typeof(*((worker))) *)((worker)))); (typeof((typeof(*((worker))) *)((worker)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); })->work) &rq->__lock irq_context: 0 (wq_completion)wg-kex-wg2#4 (work_completion)(&({ do { const void *__vpp_verify = (typeof((worker) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __asm__ ("" : "=r"(__ptr) : "0"((typeof(*((worker))) *)((worker)))); (typeof((typeof(*((worker))) *)((worker)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); })->work) &handshake->lock rcu_read_lock_bh &peer->keypairs.keypair_update_lock &obj_hash[i].lock pool_lock irq_context: 0 (wq_completion)wg-kex-wg2#4 (work_completion)(&({ do { const void *__vpp_verify = (typeof((worker) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __asm__ ("" : "=r"(__ptr) : "0"((typeof(*((worker))) *)((worker)))); (typeof((typeof(*((worker))) *)((worker)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); })->work) &handshake->lock &rq->__lock irq_context: 0 (wq_completion)wg-kex-wg2#4 (work_completion)(&({ do { const void *__vpp_verify = (typeof((worker) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __asm__ ("" : "=r"(__ptr) : "0"((typeof(*((worker))) *)((worker)))); (typeof((typeof(*((worker))) *)((worker)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); })->work) &handshake->lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)mld &rq->__lock irq_context: 0 (wq_completion)mld &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 rtnl_mutex &wg->device_update_lock &list->lock#17 &obj_hash[i].lock irq_context: 0 rtnl_mutex &wg->device_update_lock &list->lock#17 pool_lock#2 irq_context: 0 rtnl_mutex &tbl->lock krc.lock &obj_hash[i].lock irq_context: 0 rtnl_mutex &tbl->lock krc.lock &base->lock irq_context: 0 rtnl_mutex &tbl->lock krc.lock &base->lock &obj_hash[i].lock irq_context: 0 rtnl_mutex team->team_lock_key#3 lweventlist_lock irq_context: 0 rtnl_mutex team->team_lock_key#3 lweventlist_lock &dir->lock#2 irq_context: 0 (wq_completion)bat_events (work_completion)(&(&forw_packet_aggr->delayed_work)->work) &base->lock irq_context: 0 (wq_completion)bat_events (work_completion)(&(&forw_packet_aggr->delayed_work)->work) &base->lock &obj_hash[i].lock irq_context: 0 (wq_completion)gid-cache-wq (work_completion)(&work->work) &rq->__lock irq_context: 0 (wq_completion)gid-cache-wq (work_completion)(&work->work) &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 rtnl_mutex &idev->mc_lock &macvlan_netdev_addr_lock_key/1 &obj_hash[i].lock pool_lock irq_context: 0 (wq_completion)bond4 &rq->__lock irq_context: 0 (wq_completion)bond4 &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 rtnl_mutex &rdev->wiphy.mtx &wdev->mtx lweventlist_lock &c->lock irq_context: 0 rtnl_mutex &rdev->wiphy.mtx rcu_state.exp_mutex &rnp->exp_wq[3] irq_context: 0 rtnl_mutex &rdev->wiphy.mtx &local->key_mtx rcu_state.exp_mutex &rnp->exp_wq[0] irq_context: 0 rtnl_mutex &rdev->wiphy.mtx (wq_completion)phy11 irq_context: 0 rtnl_mutex &rdev->wiphy.mtx (wq_completion)phy14 irq_context: 0 (wq_completion)wg-kex-wg2 (work_completion)(&peer->transmit_handshake_work) &wg->static_identity.lock &handshake->lock &rq->__lock irq_context: 0 rtnl_mutex rcu_read_lock &tb->tb6_lock &base->lock irq_context: 0 rtnl_mutex rcu_read_lock &tb->tb6_lock &base->lock &obj_hash[i].lock irq_context: 0 rtnl_mutex rcu_read_lock krc.lock &obj_hash[i].lock irq_context: 0 rtnl_mutex rcu_read_lock krc.lock &base->lock irq_context: 0 rtnl_mutex rcu_read_lock krc.lock &base->lock &obj_hash[i].lock irq_context: 0 (wq_completion)wg-kex-wg0 (work_completion)(&peer->transmit_handshake_work) &rq->__lock irq_context: 0 (wq_completion)wg-kex-wg0 (work_completion)(&peer->transmit_handshake_work) &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 rtnl_mutex &sb->s_type->i_mutex_key#3 stock_lock irq_context: 0 rtnl_mutex &sb->s_type->i_mutex_key#3 &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 rtnl_mutex &sb->s_type->i_mutex_key#3 &c->lock irq_context: 0 rtnl_mutex &sb->s_type->i_mutex_key#3 rcu_read_lock rcu_node_0 irq_context: 0 rtnl_mutex &sb->s_type->i_mutex_key#3 rcu_read_lock &rq->__lock irq_context: 0 rtnl_mutex &sb->s_type->i_mutex_key#3 rcu_read_lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 rtnl_mutex _xmit_ETHER/1 &c->lock irq_context: 0 rtnl_mutex _xmit_ETHER/1 &____s->seqcount#2 irq_context: 0 rtnl_mutex _xmit_ETHER/1 &____s->seqcount irq_context: 0 rtnl_mutex (&hsr->prune_timer) irq_context: 0 rtnl_mutex (&hsr->announce_timer) irq_context: 0 rtnl_mutex &sb->s_type->i_mutex_key#3 &dentry->d_lock &dentry->d_lock/1 irq_context: 0 rtnl_mutex &sb->s_type->i_mutex_key#3 pin_fs_lock irq_context: 0 rtnl_mutex &sb->s_type->i_mutex_key#3 rcu_read_lock mount_lock irq_context: 0 rtnl_mutex &sb->s_type->i_mutex_key#3 rcu_read_lock mount_lock mount_lock.seqcount irq_context: 0 rtnl_mutex &sb->s_type->i_mutex_key#3 mount_lock irq_context: 0 rtnl_mutex &sb->s_type->i_mutex_key#3 mount_lock mount_lock.seqcount irq_context: 0 rtnl_mutex &sb->s_type->i_mutex_key#3 &fsnotify_mark_srcu irq_context: 0 rtnl_mutex &sb->s_type->i_mutex_key#3 &xa->xa_lock#9 irq_context: 0 rtnl_mutex &sb->s_type->i_mutex_key#3 &obj_hash[i].lock irq_context: 0 rtnl_mutex &sb->s_type->i_lock_key#7 irq_context: 0 cb_lock genl_mutex rcu_read_lock rcu_read_lock_bh rcu_read_lock &____s->seqcount#2 irq_context: 0 cb_lock genl_mutex rcu_read_lock rcu_read_lock_bh rcu_read_lock &____s->seqcount irq_context: 0 cpu_hotplug_lock jump_label_mutex text_mutex &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 cpu_hotplug_lock jump_label_mutex.wait_lock irq_context: 0 (wq_completion)cfg80211 (work_completion)(&(&rdev->dfs_update_channels_wk)->work) &rq->__lock irq_context: 0 (wq_completion)cfg80211 (work_completion)(&(&rdev->dfs_update_channels_wk)->work) &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 rtnl_mutex act_mod_lock irq_context: 0 rtnl_mutex &tn->idrinfo->lock irq_context: 0 rtnl_mutex &tn->idrinfo->lock fs_reclaim irq_context: 0 rtnl_mutex &tn->idrinfo->lock fs_reclaim mmu_notifier_invalidate_range_start irq_context: 0 rtnl_mutex &tn->idrinfo->lock pool_lock#2 irq_context: 0 rtnl_mutex &p->tcfa_lock irq_context: 0 rcu_read_lock rcu_read_lock_bh rcu_read_lock quarantine_lock irq_context: 0 (wq_completion)kblockd (work_completion)(&q->timeout_work) rcu_read_lock &rq->__lock irq_context: 0 (wq_completion)kblockd (work_completion)(&q->timeout_work) rcu_read_lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 sk_lock-AF_INET6 rcu_read_lock rcu_read_lock rcu_read_lock_bh rcu_read_lock &____s->seqcount#2 irq_context: 0 rtnl_mutex &sb->s_type->i_mutex_key#3 &n->list_lock irq_context: 0 rtnl_mutex &sb->s_type->i_mutex_key#3 &n->list_lock &c->lock irq_context: 0 rtnl_mutex _xmit_ETHER/1 &n->list_lock irq_context: 0 rtnl_mutex _xmit_ETHER/1 &n->list_lock &c->lock irq_context: 0 sb_writers#4 &sb->s_type->i_mutex_key#8 &sbi->s_writepages_rwsem jbd2_handle &ei->i_data_sem rcu_read_lock rcu_node_0 irq_context: 0 sb_writers#4 &sb->s_type->i_mutex_key#8 &sbi->s_writepages_rwsem jbd2_handle &ei->i_data_sem rcu_read_lock &rq->__lock irq_context: 0 sb_writers#4 &sb->s_type->i_mutex_key#8 &sbi->s_writepages_rwsem jbd2_handle &ei->i_data_sem rcu_read_lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 &sb->s_type->i_mutex_key#10 sk_lock-AF_INET slock-AF_INET rcu_read_lock rcu_read_lock pool_lock#2 irq_context: 0 &sb->s_type->i_mutex_key#10 sk_lock-AF_INET slock-AF_INET rcu_read_lock rcu_read_lock &c->lock irq_context: 0 &sb->s_type->i_mutex_key#10 sk_lock-AF_INET slock-AF_INET rcu_read_lock rcu_read_lock &dir->lock#2 irq_context: 0 &sb->s_type->i_mutex_key#10 sk_lock-AF_INET slock-AF_INET rcu_read_lock rcu_read_lock &ul->lock irq_context: 0 &sb->s_type->i_mutex_key#10 sk_lock-AF_INET slock-AF_INET rcu_read_lock rcu_read_lock &obj_hash[i].lock irq_context: 0 sock_diag_mutex sock_diag_table_mutex nlk_cb_mutex-SOCK_DIAG rcu_read_lock &n->list_lock irq_context: 0 sock_diag_mutex sock_diag_table_mutex nlk_cb_mutex-SOCK_DIAG rcu_read_lock &n->list_lock &c->lock irq_context: 0 sock_diag_mutex sock_diag_table_mutex nlk_cb_mutex-SOCK_DIAG rcu_read_lock &ei->socket.wq.wait &p->pi_lock &cfs_rq->removed.lock irq_context: 0 sock_diag_mutex sock_diag_mutex.wait_lock irq_context: 0 sock_diag_mutex sock_diag_table_mutex nlk_cb_mutex-SOCK_DIAG rcu_read_lock &cfs_rq->removed.lock irq_context: 0 sock_diag_mutex sock_diag_table_mutex nlk_cb_mutex-SOCK_DIAG rcu_read_lock &obj_hash[i].lock irq_context: 0 sock_diag_mutex sock_diag_table_mutex &p->pi_lock &rq->__lock &cfs_rq->removed.lock irq_context: 0 sock_diag_mutex.wait_lock irq_context: 0 sock_diag_mutex &cfs_rq->removed.lock irq_context: 0 sock_diag_mutex &obj_hash[i].lock irq_context: 0 (wq_completion)wg-kex-wg0#7 (work_completion)(&peer->transmit_handshake_work) &peer->endpoint_lock rcu_read_lock_bh rcu_read_lock rcu_read_lock rcu_read_lock rcu_read_lock_bh rcu_read_lock tk_core.seq.seqcount irq_context: 0 (wq_completion)wg-kex-wg0#7 (work_completion)(&peer->transmit_handshake_work) &peer->endpoint_lock rcu_read_lock_bh rcu_read_lock rcu_read_lock rcu_read_lock rcu_read_lock_bh rcu_read_lock &obj_hash[i].lock irq_context: 0 sk_lock-AF_INET &ping_table.lock irq_context: 0 rtnl_mutex netpoll_srcu &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 sk_lock-AF_INET rcu_read_lock &nf_conntrack_locks[i]/1 irq_context: 0 file_rwsem rcu_node_0 irq_context: 0 (wq_completion)bond8 irq_context: 0 (wq_completion)bond8 (work_completion)(&(&slave->notify_work)->work) irq_context: 0 (wq_completion)bond8 (work_completion)(&(&slave->notify_work)->work) &obj_hash[i].lock irq_context: 0 (wq_completion)bond8 (work_completion)(&(&slave->notify_work)->work) &base->lock irq_context: 0 (wq_completion)bond8 (work_completion)(&(&slave->notify_work)->work) &base->lock &obj_hash[i].lock irq_context: 0 (wq_completion)bond8 (work_completion)(&(&slave->notify_work)->work) &rq->__lock irq_context: 0 (wq_completion)bond8 (work_completion)(&(&slave->notify_work)->work) &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 &root->kernfs_iattr_rwsem &cfs_rq->removed.lock irq_context: 0 &root->kernfs_iattr_rwsem &obj_hash[i].lock irq_context: 0 rtnl_mutex &bat_priv->tvlv.handler_list_lock &____s->seqcount#2 irq_context: 0 rtnl_mutex &bat_priv->tvlv.handler_list_lock &____s->seqcount irq_context: 0 (wq_completion)bond9 irq_context: 0 (wq_completion)bond9 (work_completion)(&(&slave->notify_work)->work) irq_context: 0 (wq_completion)bond9 (work_completion)(&(&slave->notify_work)->work) &obj_hash[i].lock irq_context: 0 (wq_completion)bond9 (work_completion)(&(&slave->notify_work)->work) &base->lock irq_context: 0 (wq_completion)bond9 (work_completion)(&(&slave->notify_work)->work) &base->lock &obj_hash[i].lock irq_context: 0 namespace_sem fs_reclaim &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)bond8 (work_completion)(&(&slave->notify_work)->work) rtnl_mutex net_rwsem irq_context: 0 (wq_completion)bond8 (work_completion)(&(&slave->notify_work)->work) rtnl_mutex net_rwsem &list->lock#2 irq_context: 0 (wq_completion)bond8 (work_completion)(&(&slave->notify_work)->work) rtnl_mutex &tn->lock irq_context: 0 &type->s_umount_key#23/1 shrinker_rwsem &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)bond9 (work_completion)(&(&slave->notify_work)->work) rtnl_mutex net_rwsem irq_context: 0 (wq_completion)bond9 (work_completion)(&(&slave->notify_work)->work) rtnl_mutex net_rwsem &list->lock#2 irq_context: 0 (wq_completion)bond9 (work_completion)(&(&slave->notify_work)->work) rtnl_mutex &tn->lock irq_context: 0 (wq_completion)bond9 (work_completion)(&(&slave->notify_work)->work) rtnl_mutex.wait_lock irq_context: 0 (wq_completion)bond9 (work_completion)(&(&slave->notify_work)->work) &p->pi_lock irq_context: 0 sb_writers#11 &type->i_mutex_dir_key#7/1 cgroup_mutex rcu_node_0 irq_context: 0 sk_lock-AF_INET k-sk_lock-AF_INET rcu_read_lock &n->list_lock irq_context: 0 sk_lock-AF_INET k-sk_lock-AF_INET rcu_read_lock &n->list_lock &c->lock irq_context: softirq (&peer->timer_retransmit_handshake) rcu_read_lock_bh &obj_hash[i].lock irq_context: softirq (&peer->timer_retransmit_handshake) rcu_read_lock_bh &base->lock irq_context: softirq (&peer->timer_retransmit_handshake) rcu_read_lock_bh &base->lock &obj_hash[i].lock irq_context: 0 pernet_ops_rwsem k-sk_lock-AF_TIPC &tn->nametbl_lock &____s->seqcount#2 irq_context: 0 pernet_ops_rwsem k-sk_lock-AF_TIPC &tn->nametbl_lock &____s->seqcount irq_context: softirq (&peer->timer_retransmit_handshake) rcu_read_lock_bh rcu_read_lock &pool->lock/1 &p->pi_lock &rq->__lock &cfs_rq->removed.lock irq_context: 0 sk_lock-AF_INET k-sk_lock-AF_INET rcu_read_lock &rq->__lock irq_context: 0 sk_lock-AF_INET k-sk_lock-AF_INET rcu_read_lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 sk_lock-AF_INET k-sk_lock-AF_INET rcu_read_lock rcu_read_lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 &mm->mmap_lock rcu_read_lock &obj_hash[i].lock pool_lock irq_context: 0 &mm->mmap_lock sb_writers#4 jbd2_handle &journal->j_state_lock irq_context: 0 sb_writers#4 &type->i_mutex_dir_key#3/1 tomoyo_ss mount_lock irq_context: 0 sb_writers#4 &type->i_mutex_dir_key#3/1 tomoyo_ss mount_lock rcu_read_lock rename_lock.seqcount irq_context: 0 pernet_ops_rwsem uevent_sock_mutex rcu_read_lock &ei->socket.wq.wait &ep->lock &ep->wq &p->pi_lock &rq->__lock &cfs_rq->removed.lock irq_context: 0 pernet_ops_rwsem uevent_sock_mutex rcu_read_lock &cfs_rq->removed.lock irq_context: 0 pernet_ops_rwsem uevent_sock_mutex rcu_read_lock &obj_hash[i].lock irq_context: 0 rtnl_mutex &idev->mc_lock rcu_node_0 irq_context: 0 rtnl_mutex &br->hash_lock nl_table_wait.lock &p->pi_lock irq_context: 0 &f->f_pos_lock sb_writers#4 &sb->s_type->i_mutex_key#8 &type->s_umount_key#32 rcu_read_lock &rcu_state.gp_wq irq_context: 0 &f->f_pos_lock sb_writers#4 &sb->s_type->i_mutex_key#8 &type->s_umount_key#32 rcu_read_lock &rcu_state.gp_wq &p->pi_lock irq_context: 0 &f->f_pos_lock sb_writers#4 &sb->s_type->i_mutex_key#8 &type->s_umount_key#32 rcu_read_lock &rcu_state.gp_wq &p->pi_lock &rq->__lock irq_context: 0 &f->f_pos_lock sb_writers#4 &sb->s_type->i_mutex_key#8 &type->s_umount_key#32 rcu_read_lock &rcu_state.gp_wq &p->pi_lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)writeback (work_completion)(&(&wb->dwork)->work) &sbi->s_writepages_rwsem rcu_read_lock rcu_node_0 irq_context: 0 (wq_completion)writeback (work_completion)(&(&wb->dwork)->work) &sbi->s_writepages_rwsem rcu_read_lock &rq->__lock irq_context: 0 (wq_completion)writeback (work_completion)(&(&wb->dwork)->work) &sbi->s_writepages_rwsem rcu_read_lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 &sb->s_type->i_mutex_key#10 (netlink_chain).rwsem &nft_net->commit_mutex &rnp->exp_lock irq_context: 0 &sb->s_type->i_mutex_key#10 (netlink_chain).rwsem &nft_net->commit_mutex rcu_state.exp_mutex irq_context: 0 &sb->s_type->i_mutex_key#10 (netlink_chain).rwsem &nft_net->commit_mutex rcu_state.exp_mutex rcu_state.exp_mutex.wait_lock irq_context: 0 &f->f_pos_lock sb_writers#11 &____s->seqcount#2 irq_context: 0 sb_writers#4 sb_internal jbd2_handle &ei->i_es_lock key#7 irq_context: 0 (wq_completion)ext4-rsv-conversion (work_completion)(&ei->i_rsv_conversion_work) jbd2_handle &ei->i_data_sem &ei->i_es_lock quarantine_lock irq_context: 0 sk_lock-AF_ALG &mm->mmap_lock sb_pagefaults mmu_notifier_invalidate_range_start &rq->__lock irq_context: 0 sk_lock-AF_ALG &mm->mmap_lock sb_pagefaults mmu_notifier_invalidate_range_start &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 sk_lock-AF_ALG rcu_node_0 irq_context: 0 kn->active#5 &kernfs_locks->open_file_mutex[count] pool_lock#2 irq_context: 0 &ep->mtx &sem->wait_lock irq_context: 0 &ep->mtx &p->pi_lock irq_context: 0 &ep->mtx &p->pi_lock &rq->__lock irq_context: 0 &ep->mtx &p->pi_lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 &ep->mtx &root->kernfs_rwsem &rq->__lock irq_context: 0 &ep->mtx &root->kernfs_rwsem &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 &ep->mtx &root->kernfs_rwsem &cfs_rq->removed.lock irq_context: softirq (&ndev->rs_timer) rcu_read_lock rcu_read_lock rcu_read_lock_bh &sch->q.lock hrtimer_bases.lock irq_context: softirq (&ndev->rs_timer) rcu_read_lock rcu_read_lock rcu_read_lock_bh &sch->q.lock hrtimer_bases.lock &obj_hash[i].lock irq_context: softirq (&ndev->rs_timer) rcu_read_lock rcu_read_lock rcu_read_lock_bh &sch->q.lock hrtimer_bases.lock tk_core.seq.seqcount irq_context: 0 sb_writers#3 &sb->s_type->i_mutex_key#9 &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 sb_writers#3 tomoyo_ss rcu_read_lock rcu_node_0 irq_context: 0 sb_writers#3 tomoyo_ss rcu_read_lock &rq->__lock irq_context: 0 sb_writers#3 tomoyo_ss rcu_read_lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 pernet_ops_rwsem ovs_mutex quarantine_lock irq_context: 0 pernet_ops_rwsem &dir->lock#2 &meta->lock irq_context: 0 pernet_ops_rwsem &dir->lock#2 kfence_freelist_lock irq_context: 0 &ep->mtx remove_cache_srcu &rq->__lock irq_context: 0 &ep->mtx remove_cache_srcu &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 &u->iolock &u->peer_wait &ei->socket.wq.wait &ep->poll_wait/1 &p->pi_lock &rq->__lock &cfs_rq->removed.lock irq_context: 0 pernet_ops_rwsem rtnl_mutex &caifn->caifdevs.lock &rq->__lock irq_context: 0 pernet_ops_rwsem k-slock-AF_INET6 &meta->lock irq_context: 0 pernet_ops_rwsem k-slock-AF_INET6 kfence_freelist_lock irq_context: 0 (wq_completion)wg-kex-wg0#7 (work_completion)(&peer->transmit_handshake_work) &peer->endpoint_lock rcu_read_lock_bh rcu_read_lock rcu_read_lock rcu_read_lock rcu_read_lock_bh rcu_read_lock rlock-AF_PACKET irq_context: softirq (&n->timer) &n->lock &n->list_lock irq_context: softirq (&n->timer) &n->lock &n->list_lock &c->lock irq_context: softirq (&n->timer) rcu_read_lock rcu_read_lock rcu_read_lock_bh rcu_read_lock rlock-AF_PACKET irq_context: 0 pernet_ops_rwsem ipvs->est_mutex quarantine_lock irq_context: 0 &tsk->futex_exit_mutex &mm->mmap_lock batched_entropy_u8.lock irq_context: 0 &tsk->futex_exit_mutex &mm->mmap_lock kfence_freelist_lock irq_context: 0 &ep->mtx subsys mutex#15 &rq->__lock irq_context: 0 &ep->mtx subsys mutex#15 &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 &type->i_mutex_dir_key#2 rename_lock.seqcount irq_context: 0 &type->i_mutex_dir_key#2 &xa->xa_lock#4 irq_context: 0 &type->i_mutex_dir_key#2 stock_lock irq_context: 0 &type->i_mutex_dir_key#2 &dentry->d_lock &wq#2 irq_context: 0 cb_lock umhelper_sem batched_entropy_u8.lock irq_context: 0 cb_lock umhelper_sem kfence_freelist_lock irq_context: 0 &type->i_mutex_dir_key#4 rcu_read_lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 cb_lock umhelper_sem &root->kernfs_rwsem &meta->lock irq_context: 0 cb_lock umhelper_sem &root->kernfs_rwsem kfence_freelist_lock irq_context: 0 cb_lock umhelper_sem &root->kernfs_rwsem &sem->wait_lock irq_context: 0 cb_lock umhelper_sem &root->kernfs_rwsem &sem->wait_lock irq_context: 0 &ep->mtx uevent_sock_mutex.wait_lock irq_context: 0 rcu_read_lock_bh dev->qdisc_tx_busylock ?: &qdisc_tx_busylock _xmit_ETHER#2 rcu_read_lock &n->list_lock irq_context: 0 rcu_read_lock_bh dev->qdisc_tx_busylock ?: &qdisc_tx_busylock _xmit_ETHER#2 rcu_read_lock &n->list_lock &c->lock irq_context: 0 pernet_ops_rwsem &dev->mutex &rq->__lock irq_context: 0 pernet_ops_rwsem &dev->mutex &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 &root->kernfs_rwsem &sem->wait_lock irq_context: 0 sb_writers#11 &type->i_mutex_dir_key#7/1 tomoyo_ss quarantine_lock irq_context: 0 pernet_ops_rwsem devices_rwsem rdma_nets_rwsem &device->compat_devs_mutex &root->kernfs_rwsem rcu_read_lock pool_lock#2 irq_context: 0 &f->f_pos_lock sb_writers#4 &sb->s_type->i_mutex_key#8 &ei->i_data_sem remove_cache_srcu irq_context: 0 &f->f_pos_lock sb_writers#4 &sb->s_type->i_mutex_key#8 &ei->i_data_sem remove_cache_srcu quarantine_lock irq_context: 0 &f->f_pos_lock sb_writers#4 &sb->s_type->i_mutex_key#8 &ei->i_data_sem remove_cache_srcu &c->lock irq_context: 0 &f->f_pos_lock sb_writers#4 &sb->s_type->i_mutex_key#8 &ei->i_data_sem remove_cache_srcu &n->list_lock irq_context: 0 &f->f_pos_lock sb_writers#4 &sb->s_type->i_mutex_key#8 &ei->i_data_sem remove_cache_srcu rcu_read_lock pool_lock#2 irq_context: 0 &f->f_pos_lock sb_writers#4 &sb->s_type->i_mutex_key#8 &ei->i_data_sem remove_cache_srcu &obj_hash[i].lock irq_context: 0 &f->f_pos_lock sb_writers#4 &sb->s_type->i_mutex_key#8 &ei->i_data_sem remove_cache_srcu &rq->__lock irq_context: 0 &f->f_pos_lock sb_writers#4 &sb->s_type->i_mutex_key#8 &ei->i_data_sem remove_cache_srcu &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 sk_lock-AF_BLUETOOTH-BTPROTO_HCI mgmt_chan_list_lock pool_lock#2 irq_context: 0 sk_lock-AF_BLUETOOTH-BTPROTO_HCI mgmt_chan_list_lock tk_core.seq.seqcount irq_context: 0 sk_lock-AF_BLUETOOTH-BTPROTO_HCI mgmt_chan_list_lock hci_sk_list.lock irq_context: 0 sk_lock-AF_BLUETOOTH-BTPROTO_HCI mgmt_chan_list_lock &obj_hash[i].lock irq_context: 0 sk_lock-AF_BLUETOOTH-BTPROTO_HCI mgmt_chan_list_lock hci_dev_list_lock irq_context: 0 sk_lock-AF_BLUETOOTH-BTPROTO_HCI mgmt_chan_list_lock (console_sem).lock irq_context: 0 sk_lock-AF_BLUETOOTH-BTPROTO_HCI mgmt_chan_list_lock console_lock console_srcu console_owner_lock irq_context: 0 sk_lock-AF_BLUETOOTH-BTPROTO_HCI mgmt_chan_list_lock console_lock console_srcu console_owner irq_context: 0 sk_lock-AF_BLUETOOTH-BTPROTO_HCI mgmt_chan_list_lock console_lock console_srcu console_owner &port_lock_key irq_context: 0 sk_lock-AF_BLUETOOTH-BTPROTO_HCI mgmt_chan_list_lock console_lock console_srcu console_owner console_owner_lock irq_context: 0 sk_lock-AF_BLUETOOTH-BTPROTO_HCI mgmt_chan_list_lock &rq->__lock irq_context: 0 sk_lock-AF_BLUETOOTH-BTPROTO_HCI mgmt_chan_list_lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 sk_lock-AF_BLUETOOTH-BTPROTO_HCI mgmt_chan_list_lock &hdev->lock irq_context: 0 sk_lock-AF_BLUETOOTH-BTPROTO_HCI mgmt_chan_list_lock &hdev->lock fs_reclaim irq_context: 0 sk_lock-AF_BLUETOOTH-BTPROTO_HCI mgmt_chan_list_lock &hdev->lock fs_reclaim mmu_notifier_invalidate_range_start irq_context: 0 sk_lock-AF_BLUETOOTH-BTPROTO_HCI mgmt_chan_list_lock &hdev->lock pool_lock#2 irq_context: 0 sk_lock-AF_BLUETOOTH-BTPROTO_HCI mgmt_chan_list_lock &hdev->lock tk_core.seq.seqcount irq_context: 0 sk_lock-AF_BLUETOOTH-BTPROTO_HCI mgmt_chan_list_lock &hdev->lock rlock-AF_BLUETOOTH irq_context: 0 sk_lock-AF_BLUETOOTH-BTPROTO_HCI mgmt_chan_list_lock &hdev->lock hci_sk_list.lock irq_context: 0 sk_lock-AF_BLUETOOTH-BTPROTO_HCI mgmt_chan_list_lock &hdev->lock &obj_hash[i].lock irq_context: 0 sb_writers#4 &mapping->private_lock rcu_read_lock &p->pi_lock irq_context: 0 sb_writers#4 &mapping->private_lock rcu_read_lock &p->pi_lock &rq->__lock irq_context: 0 sb_writers#4 &mapping->private_lock rcu_read_lock &p->pi_lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 sb_writers#4 &type->i_mutex_dir_key#3/1 &sb->s_type->i_mutex_key#8 rcu_read_lock &rq->__lock &cfs_rq->removed.lock irq_context: 0 sk_lock-AF_BLUETOOTH-BTPROTO_HCI mgmt_chan_list_lock &c->lock irq_context: 0 sk_lock-AF_BLUETOOTH-BTPROTO_HCI mgmt_chan_list_lock &hdev->lock &c->lock irq_context: 0 pcpu_lock rcu_read_lock per_cpu_ptr(&cgroup_rstat_cpu_lock, cpu) irq_context: 0 namespace_sem &pcp->lock &zone->lock &____s->seqcount irq_context: 0 namespace_sem &rq->__lock &cfs_rq->removed.lock irq_context: 0 sb_writers#4 sb_internal quarantine_lock irq_context: 0 sb_writers#4 &type->i_mutex_dir_key#3/1 &sb->s_type->i_mutex_key#8 remove_cache_srcu rcu_read_lock rcu_read_lock rcu_node_0 irq_context: 0 sb_writers#4 &type->i_mutex_dir_key#3/1 &sb->s_type->i_mutex_key#8 remove_cache_srcu rcu_read_lock rcu_read_lock &rq->__lock irq_context: 0 sb_writers#4 &type->i_mutex_dir_key#3/1 &sb->s_type->i_mutex_key#8 remove_cache_srcu rcu_read_lock rcu_read_lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 &mapping->private_lock rcu_read_lock &memcg->move_lock irq_context: 0 &ei->i_prealloc_lock &pa->pa_lock#2 irq_context: 0 &(ei->i_block_reservation_lock) irq_context: 0 &dentry->d_lock &lru->node[i].lock rcu_read_lock &p->pi_lock irq_context: 0 sb_writers#4 &type->i_mutex_dir_key#3/1 tomoyo_ss remove_cache_srcu rcu_read_lock &cfs_rq->removed.lock irq_context: 0 sb_writers#4 &type->i_mutex_dir_key#3/1 tomoyo_ss remove_cache_srcu rcu_read_lock &obj_hash[i].lock irq_context: 0 &sbi->s_writepages_rwsem jbd2_handle &ei->i_data_sem &xa->xa_lock#9 pool_lock#2 irq_context: 0 pernet_ops_rwsem rtnl_mutex remove_cache_srcu &____s->seqcount irq_context: 0 &mm->mmap_lock sb_writers#4 &n->list_lock irq_context: 0 &mm->mmap_lock sb_writers#4 &n->list_lock &c->lock irq_context: 0 &sb->s_type->i_mutex_key#10 (netlink_chain).rwsem rcu_read_lock &pool->lock fill_pool_map-wait-type-override pool_lock#2 irq_context: 0 &sb->s_type->i_mutex_key#10 (netlink_chain).rwsem rcu_read_lock &pool->lock fill_pool_map-wait-type-override &c->lock irq_context: 0 &sb->s_type->i_mutex_key#10 (netlink_chain).rwsem rcu_read_lock &pool->lock fill_pool_map-wait-type-override &n->list_lock irq_context: 0 &sb->s_type->i_mutex_key#10 (netlink_chain).rwsem rcu_read_lock &pool->lock fill_pool_map-wait-type-override &n->list_lock &c->lock irq_context: 0 &sb->s_type->i_mutex_key#10 (netlink_chain).rwsem rcu_read_lock &pool->lock fill_pool_map-wait-type-override pool_lock irq_context: 0 pernet_ops_rwsem nl_table_lock &meta->lock irq_context: 0 pernet_ops_rwsem nl_table_lock kfence_freelist_lock irq_context: 0 (wq_completion)bond6 irq_context: 0 (wq_completion)bond6 (work_completion)(&(&bond->alb_work)->work) irq_context: 0 (wq_completion)bond6 (work_completion)(&(&bond->alb_work)->work) &obj_hash[i].lock irq_context: 0 (wq_completion)bond6 (work_completion)(&(&bond->alb_work)->work) &base->lock irq_context: 0 (wq_completion)bond6 (work_completion)(&(&bond->alb_work)->work) &base->lock &obj_hash[i].lock irq_context: 0 (wq_completion)bond6 (work_completion)(&(&bond->mii_work)->work) irq_context: 0 (wq_completion)bond6 (work_completion)(&(&bond->mii_work)->work) &obj_hash[i].lock irq_context: 0 (wq_completion)bond6 (work_completion)(&(&bond->mii_work)->work) &base->lock irq_context: 0 (wq_completion)bond6 (work_completion)(&(&bond->mii_work)->work) &base->lock &obj_hash[i].lock irq_context: 0 sb_writers#4 &sb->s_type->i_mutex_key#8 &sbi->s_writepages_rwsem jbd2_handle &ei->i_data_sem bit_wait_table + i irq_context: 0 sb_writers#4 &sb->s_type->i_mutex_key#8 &sbi->s_writepages_rwsem jbd2_handle &ei->i_data_sem &journal->j_wait_updates &p->pi_lock irq_context: 0 sb_writers#4 &sb->s_type->i_mutex_key#8 &sbi->s_writepages_rwsem jbd2_handle &ei->i_data_sem &journal->j_wait_updates &p->pi_lock &rq->__lock irq_context: 0 sb_writers#4 &sb->s_type->i_mutex_key#8 &sbi->s_writepages_rwsem jbd2_handle &ei->i_data_sem &journal->j_wait_updates &p->pi_lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 &mm->mmap_lock sb_writers#4 jbd2_handle &journal->j_wait_updates &p->pi_lock irq_context: 0 &mm->mmap_lock sb_writers#4 jbd2_handle &journal->j_wait_updates &p->pi_lock &rq->__lock irq_context: 0 &mm->mmap_lock sb_writers#4 jbd2_handle &journal->j_wait_updates &p->pi_lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 &tsk->futex_exit_mutex &mm->mmap_lock &obj_hash[i].lock pool_lock irq_context: 0 (wq_completion)rcu_gp (work_completion)(&rew->rew_work) fill_pool_map-wait-type-override &____s->seqcount#2 irq_context: 0 sb_writers#4 &sb->s_type->i_mutex_key#8 &rcu_state.expedited_wq &p->pi_lock irq_context: 0 sb_writers#4 &type->i_mutex_dir_key#3/1 &type->i_mutex_dir_key#3 remove_cache_srcu &cfs_rq->removed.lock irq_context: 0 (wq_completion)wg-kex-wg1#2 (work_completion)(&peer->transmit_handshake_work) &____s->seqcount#2 irq_context: 0 (wq_completion)wg-kex-wg1#2 (work_completion)(&peer->transmit_handshake_work) &____s->seqcount irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem fill_pool_map-wait-type-override &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem &nft_net->commit_mutex &rq->__lock irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem &nft_net->commit_mutex &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem uevent_sock_mutex &____s->seqcount#2 irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem ovs_mutex nf_ct_proto_mutex defrag6_mutex rcu_read_lock rcu_node_0 irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem ovs_mutex nf_ct_proto_mutex defrag6_mutex rcu_read_lock &rq->__lock irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem ovs_mutex nf_ct_proto_mutex defrag6_mutex rcu_read_lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)wg-kex-wg2 (work_completion)(&peer->transmit_handshake_work) &peer->endpoint_lock rcu_read_lock_bh rcu_read_lock rcu_read_lock_bh rcu_read_lock rlock-AF_PACKET irq_context: 0 (wq_completion)wg-kex-wg2 (work_completion)(&peer->transmit_handshake_work) &peer->endpoint_lock rcu_read_lock_bh rcu_read_lock rcu_read_lock_bh rcu_read_lock &____s->seqcount#2 irq_context: 0 (wq_completion)wg-kex-wg2 (work_completion)(&peer->transmit_handshake_work) &peer->endpoint_lock rcu_read_lock_bh rcu_read_lock rcu_read_lock_bh rcu_read_lock &____s->seqcount irq_context: 0 (wq_completion)wg-kex-wg0 (work_completion)(&peer->transmit_handshake_work) &peer->endpoint_lock rcu_read_lock_bh rcu_read_lock rcu_read_lock_bh rcu_read_lock rlock-AF_PACKET irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem k-sk_lock-AF_INET6 k-slock-AF_INET6 &tcp_hashinfo.bhash[i].lock stock_lock rcu_read_lock per_cpu_ptr(&cgroup_rstat_cpu_lock, cpu) irq_context: 0 (wq_completion)wg-kex-wg0#7 (work_completion)(&peer->transmit_handshake_work) &peer->endpoint_lock rcu_read_lock_bh rcu_read_lock &n->lock &c->lock irq_context: 0 (wq_completion)wg-kex-wg0 (work_completion)(&peer->transmit_handshake_work) &peer->endpoint_lock rcu_read_lock_bh rcu_read_lock rcu_read_lock_bh rcu_read_lock &c->lock irq_context: 0 (wq_completion)wg-kex-wg0 (work_completion)(&peer->transmit_handshake_work) &peer->endpoint_lock rcu_read_lock_bh rcu_read_lock rcu_read_lock_bh rcu_read_lock &n->list_lock irq_context: 0 (wq_completion)wg-kex-wg0 (work_completion)(&peer->transmit_handshake_work) &peer->endpoint_lock rcu_read_lock_bh rcu_read_lock rcu_read_lock_bh rcu_read_lock &n->list_lock &c->lock irq_context: softirq (&n->timer) rcu_read_lock rcu_read_lock rcu_read_lock_bh rcu_read_lock &c->lock irq_context: 0 sb_writers#4 &pipe->mutex/1 irq_context: 0 sb_writers#4 &pipe->mutex/1 &rq->__lock irq_context: 0 sb_writers#4 &pipe->mutex/1 &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 sb_writers#4 &pipe->rd_wait irq_context: 0 &pipe->mutex/1 &nlk->wait irq_context: 0 sb_writers#4 &pipe->mutex/1 &sb->s_type->i_mutex_key#8 irq_context: 0 sb_writers#4 &pipe->mutex/1 &sb->s_type->i_mutex_key#8 &ei->xattr_sem irq_context: 0 sb_writers#4 &pipe->mutex/1 &sb->s_type->i_mutex_key#8 tk_core.seq.seqcount irq_context: 0 sb_writers#4 &pipe->mutex/1 &sb->s_type->i_mutex_key#8 mmu_notifier_invalidate_range_start irq_context: 0 sb_writers#4 &pipe->mutex/1 &sb->s_type->i_mutex_key#8 pool_lock#2 irq_context: 0 sb_writers#4 &pipe->mutex/1 &sb->s_type->i_mutex_key#8 &journal->j_state_lock irq_context: 0 sb_writers#4 &pipe->mutex/1 &sb->s_type->i_mutex_key#8 jbd2_handle irq_context: 0 sb_writers#4 &pipe->mutex/1 &sb->s_type->i_mutex_key#8 jbd2_handle &ret->b_state_lock irq_context: 0 sb_writers#4 &pipe->mutex/1 &sb->s_type->i_mutex_key#8 jbd2_handle &ret->b_state_lock &journal->j_list_lock irq_context: 0 sb_writers#4 &pipe->mutex/1 &sb->s_type->i_mutex_key#8 jbd2_handle &ei->i_raw_lock irq_context: 0 sb_writers#4 &pipe->mutex/1 &sb->s_type->i_mutex_key#8 jbd2_handle &journal->j_wait_updates irq_context: 0 sb_writers#4 &pipe->mutex/1 &sb->s_type->i_mutex_key#8 &obj_hash[i].lock irq_context: 0 sb_writers#4 &pipe->mutex/1 &sb->s_type->i_mutex_key#8 jbd2_handle &sbi->s_orphan_lock irq_context: 0 sb_writers#4 &pipe->mutex/1 &sb->s_type->i_mutex_key#8 fs_reclaim irq_context: 0 sb_writers#4 &pipe->mutex/1 &sb->s_type->i_mutex_key#8 fs_reclaim mmu_notifier_invalidate_range_start irq_context: 0 sb_writers#4 &pipe->mutex/1 &sb->s_type->i_mutex_key#8 &c->lock irq_context: 0 sb_writers#4 &pipe->mutex/1 &sb->s_type->i_mutex_key#8 jbd2_handle &ei->i_es_lock irq_context: 0 sb_writers#4 &pipe->mutex/1 &sb->s_type->i_mutex_key#8 jbd2_handle &ei->i_data_sem irq_context: 0 sb_writers#4 &pipe->mutex/1 &sb->s_type->i_mutex_key#8 jbd2_handle &ei->i_data_sem mmu_notifier_invalidate_range_start irq_context: 0 sb_writers#4 &pipe->mutex/1 &sb->s_type->i_mutex_key#8 jbd2_handle &ei->i_data_sem pool_lock#2 irq_context: 0 sb_writers#4 &pipe->mutex/1 &sb->s_type->i_mutex_key#8 jbd2_handle &ei->i_data_sem &ei->i_es_lock irq_context: 0 sb_writers#4 &pipe->mutex/1 &sb->s_type->i_mutex_key#8 jbd2_handle &ei->i_data_sem &ei->i_es_lock irq_context: 0 sb_writers#4 &pipe->mutex/1 &sb->s_type->i_mutex_key#8 jbd2_handle &ei->i_data_sem &ei->i_es_lock pool_lock#2 irq_context: 0 sb_writers#4 &pipe->mutex/1 &sb->s_type->i_mutex_key#8 jbd2_handle &ei->i_data_sem &ei->i_es_lock &sbi->s_es_lock irq_context: 0 sb_writers#4 &pipe->mutex/1 &sb->s_type->i_mutex_key#8 jbd2_handle &ei->i_data_sem &obj_hash[i].lock irq_context: 0 sb_writers#4 &pipe->mutex/1 &sb->s_type->i_mutex_key#8 jbd2_handle &ei->i_data_sem irq_context: 0 sb_writers#4 &pipe->mutex/1 &sb->s_type->i_mutex_key#8 jbd2_handle &ei->i_data_sem mmu_notifier_invalidate_range_start irq_context: 0 sb_writers#4 &pipe->mutex/1 &sb->s_type->i_mutex_key#8 jbd2_handle &ei->i_data_sem pool_lock#2 irq_context: 0 sb_writers#4 &pipe->mutex/1 &sb->s_type->i_mutex_key#8 jbd2_handle &ei->i_data_sem &sb->s_type->i_lock_key#22 irq_context: 0 sb_writers#4 &pipe->mutex/1 &sb->s_type->i_mutex_key#8 jbd2_handle &ei->i_data_sem &ei->i_raw_lock irq_context: 0 sb_writers#4 &pipe->mutex/1 &sb->s_type->i_mutex_key#8 jbd2_handle &ei->i_data_sem &lg->lg_mutex irq_context: 0 sb_writers#4 &pipe->mutex/1 &sb->s_type->i_mutex_key#8 jbd2_handle &ei->i_data_sem &lg->lg_mutex &ei->i_prealloc_lock irq_context: 0 sb_writers#4 &pipe->mutex/1 &sb->s_type->i_mutex_key#8 jbd2_handle &ei->i_data_sem &lg->lg_mutex rcu_read_lock &pa->pa_lock irq_context: 0 sb_writers#4 &pipe->mutex/1 &sb->s_type->i_mutex_key#8 jbd2_handle &ei->i_data_sem &lg->lg_mutex &mapping->private_lock irq_context: 0 sb_writers#4 &pipe->mutex/1 &sb->s_type->i_mutex_key#8 jbd2_handle &ei->i_data_sem &lg->lg_mutex &ret->b_state_lock irq_context: 0 sb_writers#4 &pipe->mutex/1 &sb->s_type->i_mutex_key#8 jbd2_handle &ei->i_data_sem &lg->lg_mutex &ret->b_state_lock &journal->j_list_lock irq_context: 0 sb_writers#4 &pipe->mutex/1 &sb->s_type->i_mutex_key#8 jbd2_handle &ei->i_data_sem &lg->lg_mutex &pa->pa_lock irq_context: 0 sb_writers#4 &pipe->mutex/1 &sb->s_type->i_mutex_key#8 jbd2_handle &ei->i_data_sem &lg->lg_mutex &lg->lg_prealloc_lock irq_context: 0 sb_writers#4 &pipe->mutex/1 &sb->s_type->i_mutex_key#8 jbd2_handle &ei->i_data_sem &obj_hash[i].lock irq_context: 0 sb_writers#4 &pipe->mutex/1 &sb->s_type->i_mutex_key#8 jbd2_handle &ei->i_data_sem &ei->i_es_lock irq_context: 0 sb_writers#4 &pipe->mutex/1 &sb->s_type->i_mutex_key#8 jbd2_handle &ei->i_data_sem &ei->i_es_lock irq_context: 0 sb_writers#4 &pipe->mutex/1 &sb->s_type->i_mutex_key#8 jbd2_handle &ei->i_data_sem &ei->i_es_lock pool_lock#2 irq_context: 0 sb_writers#4 &pipe->mutex/1 &sb->s_type->i_mutex_key#8 jbd2_handle &journal->j_list_lock irq_context: 0 sb_writers#4 &pipe->mutex/1 &sb->s_type->i_mutex_key#8 mapping.invalidate_lock irq_context: 0 sb_writers#4 &pipe->mutex/1 &sb->s_type->i_mutex_key#8 mapping.invalidate_lock mmu_notifier_invalidate_range_start irq_context: 0 sb_writers#4 &pipe->mutex/1 &sb->s_type->i_mutex_key#8 mapping.invalidate_lock pool_lock#2 irq_context: 0 sb_writers#4 &pipe->mutex/1 &sb->s_type->i_mutex_key#8 mapping.invalidate_lock &journal->j_state_lock irq_context: 0 sb_writers#4 &pipe->mutex/1 &sb->s_type->i_mutex_key#8 mapping.invalidate_lock jbd2_handle irq_context: 0 sb_writers#4 &pipe->mutex/1 &sb->s_type->i_mutex_key#8 mapping.invalidate_lock jbd2_handle &ei->i_data_sem irq_context: 0 sb_writers#4 &pipe->mutex/1 &sb->s_type->i_mutex_key#8 mapping.invalidate_lock jbd2_handle &ei->i_data_sem &ei->i_prealloc_lock irq_context: 0 sb_writers#4 &pipe->mutex/1 &sb->s_type->i_mutex_key#8 mapping.invalidate_lock jbd2_handle &ei->i_data_sem &ei->i_raw_lock irq_context: 0 sb_writers#4 &pipe->mutex/1 &sb->s_type->i_mutex_key#8 mapping.invalidate_lock jbd2_handle &ei->i_data_sem &ei->i_es_lock irq_context: 0 sb_writers#4 &pipe->mutex/1 &sb->s_type->i_mutex_key#8 mapping.invalidate_lock jbd2_handle &ei->i_data_sem &ei->i_es_lock &obj_hash[i].lock irq_context: 0 sb_writers#4 &pipe->mutex/1 &sb->s_type->i_mutex_key#8 mapping.invalidate_lock jbd2_handle &ei->i_data_sem &ei->i_es_lock pool_lock#2 irq_context: 0 sb_writers#4 &pipe->mutex/1 &sb->s_type->i_mutex_key#8 mapping.invalidate_lock jbd2_handle &ei->i_data_sem &ei->i_es_lock &sbi->s_es_lock irq_context: 0 sb_writers#4 &pipe->mutex/1 &sb->s_type->i_mutex_key#8 mapping.invalidate_lock jbd2_handle &ei->i_data_sem mmu_notifier_invalidate_range_start irq_context: 0 sb_writers#4 &pipe->mutex/1 &sb->s_type->i_mutex_key#8 mapping.invalidate_lock jbd2_handle &ei->i_data_sem pool_lock#2 irq_context: 0 sb_writers#4 &pipe->mutex/1 &sb->s_type->i_mutex_key#8 mapping.invalidate_lock jbd2_handle &ei->i_data_sem &bgl->locks[i].lock &sbi->s_md_lock irq_context: 0 sb_writers#4 &pipe->mutex/1 &sb->s_type->i_mutex_key#8 mapping.invalidate_lock jbd2_handle &ei->i_data_sem &sb->s_type->i_lock_key#22 irq_context: 0 sb_writers#4 &pipe->mutex/1 &sb->s_type->i_mutex_key#8 mapping.invalidate_lock jbd2_handle &ei->i_data_sem &obj_hash[i].lock irq_context: 0 sb_writers#4 &pipe->mutex/1 &sb->s_type->i_mutex_key#8 mapping.invalidate_lock jbd2_handle &sbi->s_orphan_lock irq_context: 0 sb_writers#4 &pipe->mutex/1 &sb->s_type->i_mutex_key#8 mapping.invalidate_lock jbd2_handle &ei->i_raw_lock irq_context: 0 sb_writers#4 &pipe->mutex/1 &sb->s_type->i_mutex_key#8 mapping.invalidate_lock jbd2_handle tk_core.seq.seqcount irq_context: 0 sb_writers#4 &pipe->mutex/1 &sb->s_type->i_mutex_key#8 mapping.invalidate_lock jbd2_handle &journal->j_wait_updates irq_context: 0 sb_writers#4 &pipe->mutex/1 &sb->s_type->i_mutex_key#8 mapping.invalidate_lock &obj_hash[i].lock irq_context: 0 sb_writers#4 &pipe->mutex/1 &obj_hash[i].lock irq_context: 0 sb_writers#4 &pipe->mutex/1 pool_lock#2 irq_context: 0 (wq_completion)bond7 irq_context: 0 (wq_completion)bond7 (work_completion)(&(&bond->alb_work)->work) irq_context: 0 (wq_completion)bond7 (work_completion)(&(&bond->alb_work)->work) &obj_hash[i].lock irq_context: 0 (wq_completion)bond7 (work_completion)(&(&bond->alb_work)->work) &base->lock irq_context: 0 (wq_completion)bond7 (work_completion)(&(&bond->alb_work)->work) &base->lock &obj_hash[i].lock irq_context: 0 (wq_completion)bond7 (work_completion)(&(&bond->mii_work)->work) irq_context: 0 (wq_completion)bond7 (work_completion)(&(&bond->mii_work)->work) &obj_hash[i].lock irq_context: 0 (wq_completion)bond7 (work_completion)(&(&bond->mii_work)->work) &base->lock irq_context: 0 (wq_completion)bond7 (work_completion)(&(&bond->mii_work)->work) &base->lock &obj_hash[i].lock irq_context: 0 pernet_ops_rwsem devices_rwsem rdma_nets_rwsem &device->compat_devs_mutex &rxe->usdev_lock &p->pi_lock &cfs_rq->removed.lock irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem k-sk_lock-AF_INET fill_pool_map-wait-type-override &rq->__lock irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem k-sk_lock-AF_INET fill_pool_map-wait-type-override &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 sb_writers#4 mapping.invalidate_lock rcu_read_lock &n->list_lock irq_context: 0 sb_writers#4 mapping.invalidate_lock rcu_read_lock &n->list_lock &c->lock irq_context: 0 (wq_completion)bat_events (work_completion)(&(&bat_priv->tt.work)->work) key#16 pool_lock#2 irq_context: 0 (wq_completion)bat_events (work_completion)(&(&bat_priv->tt.work)->work) key#16 &bat_priv->tt.changes_list_lock irq_context: 0 (wq_completion)bat_events (work_completion)(&(&bat_priv->tt.work)->work) key#16 &bat_priv->tt.changes_list_lock &obj_hash[i].lock irq_context: 0 (wq_completion)bat_events (work_completion)(&(&bat_priv->tt.work)->work) key#16 &bat_priv->tt.changes_list_lock pool_lock#2 irq_context: 0 pernet_ops_rwsem rtnl_mutex &idev->mc_lock fill_pool_map-wait-type-override rcu_read_lock rcu_node_0 irq_context: 0 &sb->s_type->i_mutex_key#8 &sbi->s_writepages_rwsem rcu_read_lock fill_pool_map-wait-type-override &c->lock irq_context: 0 pernet_ops_rwsem rtnl_mutex &idev->mc_lock fill_pool_map-wait-type-override rcu_read_lock &rq->__lock irq_context: 0 &sb->s_type->i_mutex_key#8 &sbi->s_writepages_rwsem rcu_read_lock fill_pool_map-wait-type-override &n->list_lock irq_context: 0 &sb->s_type->i_mutex_key#8 &sbi->s_writepages_rwsem rcu_read_lock fill_pool_map-wait-type-override &n->list_lock &c->lock irq_context: 0 pernet_ops_rwsem nf_ct_proto_mutex nf_hook_mutex batched_entropy_u8.lock irq_context: 0 pernet_ops_rwsem nf_ct_proto_mutex nf_hook_mutex kfence_freelist_lock irq_context: 0 pernet_ops_rwsem k-sk_lock-AF_RXRPC &rxnet->local_mutex fill_pool_map-wait-type-override pool_lock#2 irq_context: 0 pernet_ops_rwsem k-sk_lock-AF_RXRPC &rxnet->local_mutex fill_pool_map-wait-type-override &c->lock irq_context: 0 pernet_ops_rwsem k-sk_lock-AF_RXRPC &rxnet->local_mutex fill_pool_map-wait-type-override &____s->seqcount#2 irq_context: 0 pernet_ops_rwsem k-sk_lock-AF_RXRPC &rxnet->local_mutex fill_pool_map-wait-type-override &____s->seqcount irq_context: 0 pernet_ops_rwsem k-sk_lock-AF_RXRPC &rxnet->local_mutex fill_pool_map-wait-type-override pool_lock irq_context: 0 pernet_ops_rwsem ovs_mutex nf_ct_proto_mutex defrag6_mutex fill_pool_map-wait-type-override &____s->seqcount irq_context: 0 pernet_ops_rwsem ovs_mutex nf_ct_proto_mutex defrag6_mutex fill_pool_map-wait-type-override pool_lock#2 irq_context: 0 pernet_ops_rwsem ovs_mutex nf_ct_proto_mutex defrag6_mutex fill_pool_map-wait-type-override rcu_read_lock pool_lock#2 irq_context: 0 pernet_ops_rwsem ovs_mutex nf_ct_proto_mutex defrag6_mutex fill_pool_map-wait-type-override &obj_hash[i].lock irq_context: 0 pernet_ops_rwsem ovs_mutex nf_ct_proto_mutex defrag6_mutex fill_pool_map-wait-type-override pool_lock irq_context: 0 pernet_ops_rwsem devices_rwsem rdma_nets_rwsem &device->compat_devs_mutex uevent_sock_mutex uevent_sock_mutex.wait_lock irq_context: 0 &mm->mmap_lock ptlock_ptr(page)#2 &pcp->lock &zone->lock irq_context: 0 (wq_completion)events fqdir_free_work &p->pi_lock &cfs_rq->removed.lock irq_context: 0 pernet_ops_rwsem rtnl_mutex &base->lock &obj_hash[i].lock irq_context: 0 pernet_ops_rwsem &root->kernfs_rwsem rcu_read_lock &cfs_rq->removed.lock irq_context: 0 pernet_ops_rwsem &root->kernfs_rwsem rcu_read_lock &obj_hash[i].lock irq_context: 0 (wq_completion)wg-kex-wg1#2 (work_completion)(&peer->transmit_handshake_work) &peer->endpoint_lock rcu_read_lock_bh rcu_read_lock rcu_read_lock &____s->seqcount#7 irq_context: 0 (wq_completion)wg-kex-wg1#2 (work_completion)(&peer->transmit_handshake_work) &peer->endpoint_lock rcu_read_lock_bh rcu_read_lock &nf_nat_locks[i] irq_context: 0 (wq_completion)wg-kex-wg1#2 (work_completion)(&peer->transmit_handshake_work) &peer->endpoint_lock rcu_read_lock_bh rcu_read_lock &nf_conntrack_locks[i]/1 irq_context: 0 (wq_completion)wg-kex-wg1#2 (work_completion)(&peer->transmit_handshake_work) &peer->endpoint_lock rcu_read_lock_bh rcu_read_lock &c->lock irq_context: 0 (wq_completion)wg-kex-wg1#2 (work_completion)(&peer->transmit_handshake_work) &peer->endpoint_lock rcu_read_lock_bh rcu_read_lock &obj_hash[i].lock irq_context: 0 (wq_completion)wg-kex-wg1#2 (work_completion)(&peer->transmit_handshake_work) &peer->endpoint_lock rcu_read_lock_bh rcu_read_lock &tbl->lock irq_context: 0 (wq_completion)wg-kex-wg1#2 (work_completion)(&peer->transmit_handshake_work) &peer->endpoint_lock rcu_read_lock_bh rcu_read_lock &n->lock irq_context: 0 (wq_completion)wg-kex-wg1#2 (work_completion)(&peer->transmit_handshake_work) &peer->endpoint_lock rcu_read_lock_bh rcu_read_lock &____s->seqcount#9 irq_context: 0 (wq_completion)wg-kex-wg0#2 (work_completion)(&({ do { const void *__vpp_verify = (typeof((worker) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __asm__ ("" : "=r"(__ptr) : "0"((typeof(*((worker))) *)((worker)))); (typeof((typeof(*((worker))) *)((worker)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); })->work) &wg->static_identity.lock &handshake->lock rcu_read_lock_bh batched_entropy_u32.lock crngs.lock irq_context: 0 (wq_completion)wg-kex-wg0#2 (work_completion)(&({ do { const void *__vpp_verify = (typeof((worker) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __asm__ ("" : "=r"(__ptr) : "0"((typeof(*((worker))) *)((worker)))); (typeof((typeof(*((worker))) *)((worker)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); })->work) &c->lock irq_context: 0 (wq_completion)wg-kex-wg0#2 (work_completion)(&({ do { const void *__vpp_verify = (typeof((worker) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __asm__ ("" : "=r"(__ptr) : "0"((typeof(*((worker))) *)((worker)))); (typeof((typeof(*((worker))) *)((worker)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); })->work) &peer->endpoint_lock rcu_read_lock_bh batched_entropy_u32.lock irq_context: 0 (wq_completion)wg-kex-wg1 (work_completion)(&({ do { const void *__vpp_verify = (typeof((worker) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __asm__ ("" : "=r"(__ptr) : "0"((typeof(*((worker))) *)((worker)))); (typeof((typeof(*((worker))) *)((worker)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); })->work) &list->lock#17 irq_context: 0 (wq_completion)wg-kex-wg1 (work_completion)(&({ do { const void *__vpp_verify = (typeof((worker) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __asm__ ("" : "=r"(__ptr) : "0"((typeof(*((worker))) *)((worker)))); (typeof((typeof(*((worker))) *)((worker)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); })->work) rcu_read_lock_bh &r->producer_lock#2 irq_context: 0 (wq_completion)wg-kex-wg1 (work_completion)(&({ do { const void *__vpp_verify = (typeof((worker) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __asm__ ("" : "=r"(__ptr) : "0"((typeof(*((worker))) *)((worker)))); (typeof((typeof(*((worker))) *)((worker)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); })->work) rcu_read_lock_bh rcu_read_lock &pool->lock irq_context: 0 (wq_completion)wg-kex-wg1 (work_completion)(&({ do { const void *__vpp_verify = (typeof((worker) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __asm__ ("" : "=r"(__ptr) : "0"((typeof(*((worker))) *)((worker)))); (typeof((typeof(*((worker))) *)((worker)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); })->work) rcu_read_lock_bh rcu_read_lock &pool->lock &obj_hash[i].lock irq_context: 0 (wq_completion)wg-kex-wg1 (work_completion)(&({ do { const void *__vpp_verify = (typeof((worker) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __asm__ ("" : "=r"(__ptr) : "0"((typeof(*((worker))) *)((worker)))); (typeof((typeof(*((worker))) *)((worker)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); })->work) rcu_read_lock_bh rcu_read_lock &pool->lock &p->pi_lock irq_context: 0 (wq_completion)wg-crypt-wg1 (work_completion)(&peer->transmit_packet_work) rcu_read_lock_bh tk_core.seq.seqcount irq_context: 0 (wq_completion)wg-kex-wg2 (work_completion)(&peer->transmit_handshake_work) &peer->endpoint_lock rcu_read_lock_bh rcu_read_lock rcu_read_lock rcu_read_lock &c->lock irq_context: 0 (wq_completion)wg-kex-wg2 (work_completion)(&peer->transmit_handshake_work) &peer->endpoint_lock rcu_read_lock_bh rcu_read_lock rcu_read_lock rcu_read_lock &n->list_lock irq_context: 0 (wq_completion)wg-kex-wg2 (work_completion)(&peer->transmit_handshake_work) &peer->endpoint_lock rcu_read_lock_bh rcu_read_lock rcu_read_lock rcu_read_lock &n->list_lock &c->lock irq_context: 0 (wq_completion)wg-kex-wg2 (work_completion)(&peer->transmit_handshake_work) &peer->endpoint_lock rcu_read_lock_bh rcu_read_lock rcu_read_lock rcu_read_lock &dir->lock#2 irq_context: 0 (wq_completion)wg-kex-wg2 (work_completion)(&peer->transmit_handshake_work) &peer->endpoint_lock rcu_read_lock_bh rcu_read_lock rcu_read_lock &____s->seqcount#7 irq_context: 0 (wq_completion)wg-kex-wg2 (work_completion)(&peer->transmit_handshake_work) &peer->endpoint_lock rcu_read_lock_bh rcu_read_lock &nf_nat_locks[i] irq_context: 0 (wq_completion)wg-kex-wg2 (work_completion)(&peer->transmit_handshake_work) &peer->endpoint_lock rcu_read_lock_bh rcu_read_lock &nf_conntrack_locks[i] irq_context: 0 (wq_completion)wg-kex-wg2 (work_completion)(&peer->transmit_handshake_work) &peer->endpoint_lock rcu_read_lock_bh rcu_read_lock &nf_conntrack_locks[i] &nf_conntrack_locks[i]/1 irq_context: 0 (wq_completion)wg-kex-wg2 (work_completion)(&peer->transmit_handshake_work) &peer->endpoint_lock rcu_read_lock_bh rcu_read_lock &nf_conntrack_locks[i] &nf_conntrack_locks[i]/1 batched_entropy_u8.lock irq_context: 0 (wq_completion)wg-kex-wg0#2 (work_completion)(&({ do { const void *__vpp_verify = (typeof((worker) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __asm__ ("" : "=r"(__ptr) : "0"((typeof(*((worker))) *)((worker)))); (typeof((typeof(*((worker))) *)((worker)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); })->work) &handshake->lock remove_cache_srcu irq_context: 0 (wq_completion)wg-kex-wg0#2 (work_completion)(&({ do { const void *__vpp_verify = (typeof((worker) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __asm__ ("" : "=r"(__ptr) : "0"((typeof(*((worker))) *)((worker)))); (typeof((typeof(*((worker))) *)((worker)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); })->work) &handshake->lock remove_cache_srcu quarantine_lock irq_context: 0 (wq_completion)wg-kex-wg0#2 (work_completion)(&({ do { const void *__vpp_verify = (typeof((worker) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __asm__ ("" : "=r"(__ptr) : "0"((typeof(*((worker))) *)((worker)))); (typeof((typeof(*((worker))) *)((worker)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); })->work) &handshake->lock remove_cache_srcu &c->lock irq_context: 0 (wq_completion)wg-kex-wg0#2 (work_completion)(&({ do { const void *__vpp_verify = (typeof((worker) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __asm__ ("" : "=r"(__ptr) : "0"((typeof(*((worker))) *)((worker)))); (typeof((typeof(*((worker))) *)((worker)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); })->work) &handshake->lock remove_cache_srcu &n->list_lock irq_context: 0 (wq_completion)wg-kex-wg0#2 (work_completion)(&({ do { const void *__vpp_verify = (typeof((worker) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __asm__ ("" : "=r"(__ptr) : "0"((typeof(*((worker))) *)((worker)))); (typeof((typeof(*((worker))) *)((worker)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); })->work) &handshake->lock remove_cache_srcu &obj_hash[i].lock irq_context: 0 (wq_completion)wg-kex-wg0#2 (work_completion)(&({ do { const void *__vpp_verify = (typeof((worker) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __asm__ ("" : "=r"(__ptr) : "0"((typeof(*((worker))) *)((worker)))); (typeof((typeof(*((worker))) *)((worker)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); })->work) &peer->endpoint_lock rcu_read_lock_bh rcu_read_lock rcu_read_lock rcu_read_lock &dir->lock#2 irq_context: 0 (wq_completion)wg-kex-wg2#2 (work_completion)(&({ do { const void *__vpp_verify = (typeof((worker) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __asm__ ("" : "=r"(__ptr) : "0"((typeof(*((worker))) *)((worker)))); (typeof((typeof(*((worker))) *)((worker)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); })->work) rcu_read_lock_bh rcu_read_lock &pool->lock &p->pi_lock &rq->__lock &cfs_rq->removed.lock irq_context: 0 (wq_completion)wg-kex-wg2#2 (work_completion)(&({ do { const void *__vpp_verify = (typeof((worker) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __asm__ ("" : "=r"(__ptr) : "0"((typeof(*((worker))) *)((worker)))); (typeof((typeof(*((worker))) *)((worker)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); })->work) &wg->static_identity.lock &handshake->lock tk_core.seq.seqcount irq_context: 0 (wq_completion)wg-kex-wg2#2 (work_completion)(&({ do { const void *__vpp_verify = (typeof((worker) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __asm__ ("" : "=r"(__ptr) : "0"((typeof(*((worker))) *)((worker)))); (typeof((typeof(*((worker))) *)((worker)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); })->work) &wg->static_identity.lock &handshake->lock tk_core.seq.seqcount irq_context: 0 (wq_completion)wg-kex-wg2#2 (work_completion)(&({ do { const void *__vpp_verify = (typeof((worker) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __asm__ ("" : "=r"(__ptr) : "0"((typeof(*((worker))) *)((worker)))); (typeof((typeof(*((worker))) *)((worker)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); })->work) &wg->static_identity.lock &handshake->lock crngs.lock irq_context: 0 (wq_completion)wg-kex-wg2#2 (work_completion)(&({ do { const void *__vpp_verify = (typeof((worker) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __asm__ ("" : "=r"(__ptr) : "0"((typeof(*((worker))) *)((worker)))); (typeof((typeof(*((worker))) *)((worker)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); })->work) &wg->static_identity.lock &handshake->lock &table->lock#2 irq_context: 0 (wq_completion)wg-kex-wg2#2 (work_completion)(&({ do { const void *__vpp_verify = (typeof((worker) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __asm__ ("" : "=r"(__ptr) : "0"((typeof(*((worker))) *)((worker)))); (typeof((typeof(*((worker))) *)((worker)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); })->work) &wg->static_identity.lock &handshake->lock rcu_read_lock_bh batched_entropy_u32.lock irq_context: 0 (wq_completion)wg-kex-wg2#2 (work_completion)(&({ do { const void *__vpp_verify = (typeof((worker) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __asm__ ("" : "=r"(__ptr) : "0"((typeof(*((worker))) *)((worker)))); (typeof((typeof(*((worker))) *)((worker)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); })->work) &wg->static_identity.lock &handshake->lock rcu_read_lock_bh batched_entropy_u32.lock crngs.lock irq_context: 0 (wq_completion)wg-kex-wg2#2 (work_completion)(&({ do { const void *__vpp_verify = (typeof((worker) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __asm__ ("" : "=r"(__ptr) : "0"((typeof(*((worker))) *)((worker)))); (typeof((typeof(*((worker))) *)((worker)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); })->work) &wg->static_identity.lock &handshake->lock rcu_read_lock_bh &table->lock#2 irq_context: 0 (wq_completion)wg-kex-wg2#2 (work_completion)(&({ do { const void *__vpp_verify = (typeof((worker) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __asm__ ("" : "=r"(__ptr) : "0"((typeof(*((worker))) *)((worker)))); (typeof((typeof(*((worker))) *)((worker)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); })->work) &cookie->lock irq_context: 0 (wq_completion)wg-kex-wg2#2 (work_completion)(&({ do { const void *__vpp_verify = (typeof((worker) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __asm__ ("" : "=r"(__ptr) : "0"((typeof(*((worker))) *)((worker)))); (typeof((typeof(*((worker))) *)((worker)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); })->work) &cookie->lock irq_context: 0 (wq_completion)wg-kex-wg2#2 (work_completion)(&({ do { const void *__vpp_verify = (typeof((worker) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __asm__ ("" : "=r"(__ptr) : "0"((typeof(*((worker))) *)((worker)))); (typeof((typeof(*((worker))) *)((worker)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); })->work) &peer->endpoint_lock irq_context: 0 (wq_completion)wg-kex-wg2#2 (work_completion)(&({ do { const void *__vpp_verify = (typeof((worker) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __asm__ ("" : "=r"(__ptr) : "0"((typeof(*((worker))) *)((worker)))); (typeof((typeof(*((worker))) *)((worker)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); })->work) &peer->endpoint_lock rcu_read_lock_bh rcu_read_lock &____s->seqcount#7 irq_context: 0 (wq_completion)wg-kex-wg2#2 (work_completion)(&({ do { const void *__vpp_verify = (typeof((worker) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __asm__ ("" : "=r"(__ptr) : "0"((typeof(*((worker))) *)((worker)))); (typeof((typeof(*((worker))) *)((worker)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); })->work) &peer->endpoint_lock rcu_read_lock_bh rcu_read_lock &____s->seqcount#10 irq_context: 0 (wq_completion)wg-kex-wg2#2 (work_completion)(&({ do { const void *__vpp_verify = (typeof((worker) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __asm__ ("" : "=r"(__ptr) : "0"((typeof(*((worker))) *)((worker)))); (typeof((typeof(*((worker))) *)((worker)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); })->work) &peer->endpoint_lock rcu_read_lock_bh rcu_read_lock rcu_read_lock_bh tk_core.seq.seqcount irq_context: 0 (wq_completion)wg-kex-wg2#2 (work_completion)(&({ do { const void *__vpp_verify = (typeof((worker) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __asm__ ("" : "=r"(__ptr) : "0"((typeof(*((worker))) *)((worker)))); (typeof((typeof(*((worker))) *)((worker)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); })->work) &peer->endpoint_lock rcu_read_lock_bh rcu_read_lock rcu_read_lock_bh &list->lock#5 irq_context: 0 (wq_completion)wg-kex-wg1 (work_completion)(&({ do { const void *__vpp_verify = (typeof((worker) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __asm__ ("" : "=r"(__ptr) : "0"((typeof(*((worker))) *)((worker)))); (typeof((typeof(*((worker))) *)((worker)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); })->work) rcu_read_lock_bh rcu_read_lock &pool->lock &p->pi_lock &rq->__lock irq_context: 0 (wq_completion)wg-kex-wg1 (work_completion)(&({ do { const void *__vpp_verify = (typeof((worker) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __asm__ ("" : "=r"(__ptr) : "0"((typeof(*((worker))) *)((worker)))); (typeof((typeof(*((worker))) *)((worker)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); })->work) rcu_read_lock_bh rcu_read_lock &pool->lock &p->pi_lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 pernet_ops_rwsem uevent_sock_mutex rcu_read_lock &rcu_state.expedited_wq irq_context: 0 pernet_ops_rwsem uevent_sock_mutex rcu_read_lock &rcu_state.expedited_wq &p->pi_lock irq_context: 0 pernet_ops_rwsem uevent_sock_mutex rcu_read_lock &rcu_state.expedited_wq &p->pi_lock &rq->__lock irq_context: 0 pernet_ops_rwsem uevent_sock_mutex rcu_read_lock &rcu_state.expedited_wq &p->pi_lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 pernet_ops_rwsem devices_rwsem rdma_nets_rwsem &device->compat_devs_mutex &meta->lock irq_context: 0 pernet_ops_rwsem fs_reclaim &rq->__lock &cfs_rq->removed.lock irq_context: 0 pernet_ops_rwsem fs_reclaim &cfs_rq->removed.lock irq_context: 0 pernet_ops_rwsem fs_reclaim &obj_hash[i].lock irq_context: 0 (wq_completion)events (work_completion)(&(&nsim_dev->trap_data->trap_report_dw)->work) &devlink->lock_key#3 &nsim_trap_data->trap_lock batched_entropy_u8.lock crngs.lock irq_context: 0 pernet_ops_rwsem cpu_hotplug_lock wq_pool_mutex rcu_node_0 irq_context: 0 pernet_ops_rwsem ovs_mutex nf_ct_proto_mutex fill_pool_map-wait-type-override &c->lock irq_context: 0 pernet_ops_rwsem ovs_mutex nf_ct_proto_mutex fill_pool_map-wait-type-override pool_lock#2 irq_context: 0 pernet_ops_rwsem ovs_mutex nf_ct_proto_mutex fill_pool_map-wait-type-override &n->list_lock irq_context: 0 pernet_ops_rwsem ovs_mutex nf_ct_proto_mutex fill_pool_map-wait-type-override &n->list_lock &c->lock irq_context: 0 pernet_ops_rwsem ovs_mutex nf_ct_proto_mutex fill_pool_map-wait-type-override &rq->__lock irq_context: 0 pernet_ops_rwsem ovs_mutex nf_ct_proto_mutex fill_pool_map-wait-type-override pool_lock irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem rtnl_mutex rcu_state.exp_mutex &rcu_state.expedited_wq irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem rtnl_mutex rcu_state.exp_mutex &rcu_state.expedited_wq &p->pi_lock irq_context: 0 pernet_ops_rwsem k-sk_lock-AF_TIPC &tn->nametbl_lock krc.lock &obj_hash[i].lock irq_context: 0 pernet_ops_rwsem k-sk_lock-AF_TIPC &tn->nametbl_lock krc.lock &base->lock irq_context: 0 pernet_ops_rwsem k-sk_lock-AF_TIPC &tn->nametbl_lock krc.lock &base->lock &obj_hash[i].lock irq_context: 0 (wq_completion)events (work_completion)(&(&nsim_dev->trap_data->trap_report_dw)->work) &devlink->lock_key#6 &nsim_trap_data->trap_lock rcu_read_lock &p->pi_lock irq_context: 0 (wq_completion)events (work_completion)(&(&nsim_dev->trap_data->trap_report_dw)->work) &devlink->lock_key#6 &nsim_trap_data->trap_lock rcu_read_lock &p->pi_lock &rq->__lock irq_context: 0 (wq_completion)events (work_completion)(&(&nsim_dev->trap_data->trap_report_dw)->work) &devlink->lock_key#6 &nsim_trap_data->trap_lock rcu_read_lock &p->pi_lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 &type->s_umount_key#23/1 shrinker_rwsem fs_reclaim irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem k-sk_lock-AF_INET6 k-slock-AF_INET6 &tcp_hashinfo.bhash[i].lock per_cpu_ptr(&cgroup_rstat_cpu_lock, cpu) irq_context: softirq (&ndev->rs_timer) rcu_read_lock rcu_read_lock init_task.mems_allowed_seq.seqcount irq_context: 0 &type->s_umount_key#23/1 shrinker_rwsem fs_reclaim mmu_notifier_invalidate_range_start irq_context: 0 &type->s_umount_key#23/1 shrinker_rwsem pool_lock#2 irq_context: 0 &type->s_umount_key#23/1 shrinker_rwsem &c->lock irq_context: 0 &type->s_umount_key#23/1 shrinker_rwsem &n->list_lock irq_context: 0 &type->s_umount_key#23/1 shrinker_rwsem &n->list_lock &c->lock irq_context: 0 &type->s_umount_key#23/1 shrinker_rwsem &obj_hash[i].lock irq_context: 0 &type->s_umount_key#23/1 shrinker_rwsem &obj_hash[i].lock pool_lock irq_context: 0 &type->s_umount_key#23/1 shrinker_rwsem krc.lock irq_context: 0 &type->s_umount_key#23/1 shrinker_rwsem &____s->seqcount irq_context: 0 &type->s_umount_key#23/1 sb_lock unnamed_dev_ida.xa_lock pool_lock#2 irq_context: 0 &type->s_umount_key#23/1 fs_reclaim &rq->__lock irq_context: 0 sk_lock-AF_INET6 rlock-AF_INET6 irq_context: 0 (wq_completion)events free_ipc_work &type->s_umount_key#51 shrinker_rwsem &obj_hash[i].lock irq_context: 0 (wq_completion)events free_ipc_work &type->s_umount_key#51 shrinker_rwsem pool_lock#2 irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem nf_hook_mutex &rq->__lock irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem nf_hook_mutex &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 &sb->s_type->i_mutex_key#10 sk_lock-AF_INET6/1 &cfs_rq->removed.lock irq_context: 0 &sb->s_type->i_mutex_key#10 sk_lock-AF_INET6/1 remove_cache_srcu irq_context: 0 &sb->s_type->i_mutex_key#10 sk_lock-AF_INET6/1 remove_cache_srcu quarantine_lock irq_context: 0 &sb->s_type->i_mutex_key#10 sk_lock-AF_INET6/1 remove_cache_srcu &c->lock irq_context: 0 &sb->s_type->i_mutex_key#10 sk_lock-AF_INET6/1 remove_cache_srcu &n->list_lock irq_context: 0 &sb->s_type->i_mutex_key#10 sk_lock-AF_INET6/1 remove_cache_srcu rcu_read_lock pool_lock#2 irq_context: 0 &sb->s_type->i_mutex_key#10 sk_lock-AF_INET6/1 remove_cache_srcu &obj_hash[i].lock irq_context: 0 &sb->s_type->i_mutex_key#10 sk_lock-AF_INET6/1 remove_cache_srcu pool_lock#2 irq_context: 0 &sb->s_type->i_mutex_key#10 sk_lock-AF_INET6/1 remove_cache_srcu &rq->__lock irq_context: 0 &sb->s_type->i_mutex_key#10 sk_lock-AF_INET6/1 remove_cache_srcu &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 sk_lock-AF_INET6 &pcp->lock &zone->lock &____s->seqcount irq_context: 0 sk_lock-AF_ALG &mm->mmap_lock &cfs_rq->removed.lock irq_context: 0 sk_lock-AF_ALG &mm->mmap_lock &obj_hash[i].lock irq_context: 0 sk_lock-AF_ALG &mm->mmap_lock pool_lock#2 irq_context: 0 rtnl_mutex &idev->mc_lock fs_reclaim &rq->__lock irq_context: 0 rtnl_mutex &idev->mc_lock fs_reclaim &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 cb_lock rtnl_mutex &rdev->wiphy.mtx &k->list_lock irq_context: 0 cb_lock rtnl_mutex &rdev->wiphy.mtx &root->kernfs_rwsem irq_context: 0 cb_lock rtnl_mutex &rdev->wiphy.mtx &root->kernfs_rwsem irq_context: 0 cb_lock rtnl_mutex &rdev->wiphy.mtx &root->kernfs_rwsem fs_reclaim irq_context: 0 cb_lock rtnl_mutex &rdev->wiphy.mtx &root->kernfs_rwsem fs_reclaim mmu_notifier_invalidate_range_start irq_context: 0 cb_lock rtnl_mutex &rdev->wiphy.mtx &root->kernfs_rwsem pool_lock#2 irq_context: 0 cb_lock rtnl_mutex &rdev->wiphy.mtx &root->kernfs_rwsem kernfs_rename_lock irq_context: 0 cb_lock rtnl_mutex &rdev->wiphy.mtx &root->kernfs_rwsem &obj_hash[i].lock irq_context: 0 cb_lock rtnl_mutex &rdev->wiphy.mtx kernfs_rename_lock irq_context: 0 cb_lock rtnl_mutex &rdev->wiphy.mtx &n->list_lock irq_context: 0 cb_lock rtnl_mutex &rdev->wiphy.mtx &n->list_lock &c->lock irq_context: 0 cb_lock rtnl_mutex &rdev->wiphy.mtx uevent_sock_mutex irq_context: 0 cb_lock rtnl_mutex &rdev->wiphy.mtx uevent_sock_mutex fs_reclaim irq_context: 0 cb_lock rtnl_mutex &rdev->wiphy.mtx uevent_sock_mutex fs_reclaim mmu_notifier_invalidate_range_start irq_context: 0 cb_lock rtnl_mutex &rdev->wiphy.mtx uevent_sock_mutex pool_lock#2 irq_context: 0 cb_lock rtnl_mutex &rdev->wiphy.mtx uevent_sock_mutex nl_table_lock irq_context: 0 cb_lock rtnl_mutex &rdev->wiphy.mtx uevent_sock_mutex rlock-AF_NETLINK irq_context: 0 cb_lock rtnl_mutex &rdev->wiphy.mtx uevent_sock_mutex rcu_read_lock &ei->socket.wq.wait irq_context: 0 cb_lock rtnl_mutex &rdev->wiphy.mtx uevent_sock_mutex rcu_read_lock &ei->socket.wq.wait &ep->lock irq_context: 0 cb_lock rtnl_mutex &rdev->wiphy.mtx uevent_sock_mutex rcu_read_lock &ei->socket.wq.wait &ep->lock &ep->wq irq_context: 0 cb_lock rtnl_mutex &rdev->wiphy.mtx uevent_sock_mutex rcu_read_lock &ei->socket.wq.wait &ep->lock &ep->wq &p->pi_lock irq_context: 0 cb_lock rtnl_mutex &rdev->wiphy.mtx uevent_sock_mutex rcu_read_lock &ei->socket.wq.wait &ep->lock &ep->wq &p->pi_lock &rq->__lock irq_context: 0 cb_lock rtnl_mutex &rdev->wiphy.mtx uevent_sock_mutex rcu_read_lock &ei->socket.wq.wait &ep->lock &ep->wq &p->pi_lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 &type->i_mutex_dir_key#4 rename_lock irq_context: 0 &type->i_mutex_dir_key#4 rename_lock rename_lock.seqcount irq_context: 0 &type->i_mutex_dir_key#4 rename_lock rename_lock.seqcount &dentry->d_lock irq_context: 0 &type->i_mutex_dir_key#4 rename_lock rename_lock.seqcount &dentry->d_lock &dentry->d_lock/2 irq_context: 0 &type->i_mutex_dir_key#4 rename_lock rename_lock.seqcount &dentry->d_lock &dentry->d_lock/2 &dentry->d_lock/3 irq_context: 0 cb_lock rtnl_mutex &rdev->wiphy.mtx uevent_sock_mutex rcu_read_lock rcu_node_0 irq_context: 0 cb_lock rtnl_mutex &rdev->wiphy.mtx uevent_sock_mutex rcu_read_lock &rq->__lock irq_context: 0 &type->i_mutex_dir_key#4 rename_lock rename_lock.seqcount &dentry->d_lock &dentry->d_lock/2 &dentry->d_lock/3 &____s->seqcount#4 irq_context: 0 cb_lock rtnl_mutex &rdev->wiphy.mtx uevent_sock_mutex rcu_read_lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 &type->i_mutex_dir_key#4 rename_lock rename_lock.seqcount &dentry->d_lock &dentry->d_lock/2 &dentry->d_lock/3 &____s->seqcount#4 &____s->seqcount#4/1 irq_context: 0 &type->i_mutex_dir_key#4 rename_lock rename_lock.seqcount &dentry->d_lock &dentry->d_lock/2 &dentry->d_lock/3 &wq irq_context: 0 &type->i_mutex_dir_key#4 rename_lock rename_lock.seqcount &dentry->d_lock/2 irq_context: 0 &type->i_mutex_dir_key#4 rename_lock rename_lock.seqcount &dentry->d_lock/2 &dentry->d_lock/3 irq_context: 0 cb_lock rtnl_mutex &rdev->wiphy.mtx uevent_sock_mutex &obj_hash[i].lock irq_context: 0 cb_lock rtnl_mutex &rdev->wiphy.mtx uevent_sock_mutex nl_table_wait.lock irq_context: 0 cb_lock rtnl_mutex &rdev->wiphy.mtx &sb->s_type->i_mutex_key#3/1 irq_context: 0 cb_lock rtnl_mutex &rdev->wiphy.mtx &sb->s_type->i_mutex_key#3/1 rename_lock.seqcount irq_context: 0 cb_lock rtnl_mutex &rdev->wiphy.mtx &sb->s_type->i_mutex_key#3/1 fs_reclaim irq_context: 0 cb_lock rtnl_mutex &rdev->wiphy.mtx &sb->s_type->i_mutex_key#3/1 fs_reclaim mmu_notifier_invalidate_range_start irq_context: 0 cb_lock rtnl_mutex &rdev->wiphy.mtx &sb->s_type->i_mutex_key#3/1 stock_lock irq_context: 0 cb_lock rtnl_mutex &rdev->wiphy.mtx &sb->s_type->i_mutex_key#3/1 &c->lock irq_context: 0 cb_lock rtnl_mutex &rdev->wiphy.mtx &sb->s_type->i_mutex_key#3/1 pool_lock#2 irq_context: 0 cb_lock rtnl_mutex &rdev->wiphy.mtx &sb->s_type->i_mutex_key#3/1 &dentry->d_lock irq_context: 0 cb_lock rtnl_mutex &rdev->wiphy.mtx &sb->s_type->i_mutex_key#3/1 &rq->__lock irq_context: 0 cb_lock rtnl_mutex &rdev->wiphy.mtx &sb->s_type->i_mutex_key#3/1 &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 cb_lock rtnl_mutex &rdev->wiphy.mtx &sb->s_type->i_mutex_key#3/1 rcu_read_lock rename_lock.seqcount irq_context: 0 cb_lock rtnl_mutex &rdev->wiphy.mtx &sb->s_type->i_mutex_key#3/1 &dentry->d_lock &wq irq_context: 0 cb_lock rtnl_mutex &rdev->wiphy.mtx &sb->s_type->i_mutex_key#3/1 tk_core.seq.seqcount irq_context: 0 cb_lock rtnl_mutex &rdev->wiphy.mtx &sb->s_type->i_mutex_key#3/1 rename_lock irq_context: 0 cb_lock rtnl_mutex &rdev->wiphy.mtx &sb->s_type->i_mutex_key#3/1 rename_lock rename_lock.seqcount irq_context: 0 cb_lock rtnl_mutex &rdev->wiphy.mtx &sb->s_type->i_mutex_key#3/1 rename_lock rename_lock.seqcount &dentry->d_lock irq_context: 0 cb_lock rtnl_mutex &rdev->wiphy.mtx &sb->s_type->i_mutex_key#3/1 rename_lock rename_lock.seqcount &dentry->d_lock &dentry->d_lock/2 irq_context: 0 cb_lock rtnl_mutex &rdev->wiphy.mtx &sb->s_type->i_mutex_key#3/1 rename_lock rename_lock.seqcount &dentry->d_lock &dentry->d_lock/2 &dentry->d_lock/3 irq_context: 0 cb_lock rtnl_mutex &rdev->wiphy.mtx &sb->s_type->i_mutex_key#3/1 rename_lock rename_lock.seqcount &dentry->d_lock &dentry->d_lock/2 &dentry->d_lock/3 &____s->seqcount#4 irq_context: 0 cb_lock rtnl_mutex &rdev->wiphy.mtx &sb->s_type->i_mutex_key#3/1 rename_lock rename_lock.seqcount &dentry->d_lock &dentry->d_lock/2 &dentry->d_lock/3 &____s->seqcount#4 &____s->seqcount#4/1 irq_context: 0 cb_lock rtnl_mutex &rdev->wiphy.mtx &sb->s_type->i_mutex_key#3/1 rename_lock rename_lock.seqcount &dentry->d_lock/2 irq_context: 0 cb_lock rtnl_mutex &rdev->wiphy.mtx &sb->s_type->i_mutex_key#3/1 rename_lock rename_lock.seqcount &dentry->d_lock/2 &dentry->d_lock/3 irq_context: 0 cb_lock rtnl_mutex &rdev->wiphy.mtx remove_cache_srcu irq_context: 0 cb_lock rtnl_mutex &rdev->wiphy.mtx remove_cache_srcu quarantine_lock irq_context: 0 cb_lock rtnl_mutex &rdev->wiphy.mtx remove_cache_srcu &c->lock irq_context: 0 cb_lock rtnl_mutex &rdev->wiphy.mtx remove_cache_srcu &n->list_lock irq_context: 0 cb_lock rtnl_mutex &rdev->wiphy.mtx remove_cache_srcu rcu_read_lock pool_lock#2 irq_context: 0 cb_lock rtnl_mutex &rdev->wiphy.mtx remove_cache_srcu &obj_hash[i].lock irq_context: 0 cb_lock rtnl_mutex &rdev->wiphy.mtx remove_cache_srcu &rq->__lock irq_context: 0 cb_lock rtnl_mutex &rdev->wiphy.mtx remove_cache_srcu &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 cb_lock rtnl_mutex &rdev->wiphy.mtx remove_cache_srcu pool_lock#2 irq_context: 0 rcu_read_lock_bh &sch->q.lock tk_core.seq.seqcount irq_context: 0 (wq_completion)mld (work_completion)(&(&idev->mc_ifc_work)->work) &idev->mc_lock rcu_read_lock rcu_read_lock rcu_read_lock_bh &sch->q.lock hrtimer_bases.lock irq_context: 0 (wq_completion)mld (work_completion)(&(&idev->mc_ifc_work)->work) &idev->mc_lock rcu_read_lock rcu_read_lock rcu_read_lock_bh &sch->q.lock hrtimer_bases.lock &obj_hash[i].lock irq_context: 0 (wq_completion)mld (work_completion)(&(&idev->mc_ifc_work)->work) &idev->mc_lock rcu_read_lock rcu_read_lock rcu_read_lock_bh &sch->q.lock hrtimer_bases.lock tk_core.seq.seqcount irq_context: 0 (wq_completion)bat_events (work_completion)(&(&bat_priv->bla.work)->work) rcu_read_lock &meta->lock irq_context: 0 (wq_completion)bat_events (work_completion)(&(&bat_priv->bla.work)->work) rcu_read_lock &rcu_state.expedited_wq irq_context: 0 (wq_completion)bat_events (work_completion)(&(&bat_priv->bla.work)->work) fill_pool_map-wait-type-override &c->lock irq_context: 0 (wq_completion)bat_events (work_completion)(&(&bat_priv->bla.work)->work) fill_pool_map-wait-type-override &n->list_lock irq_context: 0 (wq_completion)bat_events (work_completion)(&(&bat_priv->bla.work)->work) fill_pool_map-wait-type-override &n->list_lock &c->lock irq_context: 0 (wq_completion)bat_events (work_completion)(&(&bat_priv->bla.work)->work) fill_pool_map-wait-type-override pool_lock irq_context: 0 &sb->s_type->i_mutex_key#10 rcu_state.exp_mutex fill_pool_map-wait-type-override &c->lock irq_context: 0 &sb->s_type->i_mutex_key#10 rcu_state.exp_mutex fill_pool_map-wait-type-override pool_lock#2 irq_context: 0 &sb->s_type->i_mutex_key#10 rcu_state.exp_mutex fill_pool_map-wait-type-override &n->list_lock irq_context: 0 &sb->s_type->i_mutex_key#10 rcu_state.exp_mutex fill_pool_map-wait-type-override &n->list_lock &c->lock irq_context: 0 &sb->s_type->i_mutex_key#10 rcu_state.exp_mutex fill_pool_map-wait-type-override pool_lock irq_context: 0 &sb->s_type->i_mutex_key#10 &net->sctp.addr_wq_lock slock-AF_INET6/1 key#25 irq_context: 0 rtnl_mutex sysctl_lock fill_pool_map-wait-type-override &____s->seqcount#2 irq_context: 0 sb_writers#4 &type->i_mutex_dir_key#3 tomoyo_ss rcu_read_lock &rcu_state.gp_wq irq_context: 0 sb_writers#4 &type->i_mutex_dir_key#3 tomoyo_ss rcu_read_lock &rcu_state.gp_wq &p->pi_lock irq_context: 0 sb_writers#4 &type->i_mutex_dir_key#3 tomoyo_ss rcu_read_lock &rcu_state.gp_wq &p->pi_lock &rq->__lock irq_context: 0 sb_writers#4 &type->i_mutex_dir_key#3 tomoyo_ss rcu_read_lock &rcu_state.gp_wq &p->pi_lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 sb_writers#4 &type->i_mutex_dir_key#3 tomoyo_ss rcu_read_lock &rcu_state.expedited_wq irq_context: 0 sb_writers#4 &type->i_mutex_dir_key#3 tomoyo_ss rcu_read_lock &rcu_state.expedited_wq &p->pi_lock irq_context: 0 sb_writers#4 &type->i_mutex_dir_key#3 tomoyo_ss rcu_read_lock &rcu_state.expedited_wq &p->pi_lock &rq->__lock irq_context: 0 sb_writers#4 &type->i_mutex_dir_key#3 tomoyo_ss rcu_read_lock &rcu_state.expedited_wq &p->pi_lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 fill_pool_map-wait-type-override rcu_node_0 irq_context: softirq (&n->timer) icmp_global.lock batched_entropy_u8.lock crngs.lock irq_context: 0 (wq_completion)phy12 (work_completion)(&(&local->roc_work)->work) irq_context: 0 (wq_completion)phy12 (work_completion)(&(&local->roc_work)->work) &local->mtx irq_context: 0 (wq_completion)phy12 (work_completion)(&(&local->roc_work)->work) &local->mtx &obj_hash[i].lock irq_context: 0 (wq_completion)phy12 (work_completion)(&(&local->roc_work)->work) &local->mtx &base->lock irq_context: 0 (wq_completion)phy12 (work_completion)(&(&local->roc_work)->work) &local->mtx &base->lock &obj_hash[i].lock irq_context: 0 (wq_completion)phy12 (work_completion)(&(&local->roc_work)->work) &local->mtx &rq->__lock irq_context: 0 (wq_completion)phy12 (work_completion)(&(&local->roc_work)->work) &local->mtx &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)phy12 (work_completion)(&(&local->roc_work)->work) &local->mtx rcu_read_lock &local->queue_stop_reason_lock irq_context: 0 (wq_completion)phy12 (work_completion)(&(&local->roc_work)->work) &local->mtx rcu_read_lock tk_core.seq.seqcount irq_context: 0 (wq_completion)phy12 (work_completion)(&(&local->roc_work)->work) &local->mtx rcu_read_lock hwsim_radio_lock irq_context: 0 (wq_completion)phy12 (work_completion)(&(&local->roc_work)->work) &local->mtx rcu_read_lock hwsim_radio_lock pool_lock#2 irq_context: 0 (wq_completion)phy12 (work_completion)(&(&local->roc_work)->work) &local->mtx rcu_read_lock hwsim_radio_lock &list->lock#19 irq_context: 0 (wq_completion)phy12 (work_completion)(&(&local->roc_work)->work) &local->mtx rcu_read_lock &list->lock#19 irq_context: 0 sk_lock-AF_BLUETOOTH-BTPROTO_RFCOMM rlock-AF_BLUETOOTH irq_context: 0 cb_lock genl_mutex rcu_read_lock rcu_read_lock rcu_node_0 irq_context: 0 cb_lock genl_mutex rcu_read_lock rcu_read_lock &rq->__lock irq_context: 0 cb_lock genl_mutex rcu_read_lock rcu_read_lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 sk_lock-AF_ALG &drbg->drbg_mutex rcu_node_0 irq_context: 0 sk_lock-AF_ALG &drbg->drbg_mutex &rq->__lock &cfs_rq->removed.lock irq_context: 0 &mm->mmap_lock sb_pagefaults rcu_read_lock rcu_node_0 irq_context: 0 &mm->mmap_lock sb_pagefaults rcu_read_lock &rq->__lock irq_context: 0 &mm->mmap_lock sb_pagefaults rcu_read_lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 &f->f_pos_lock sb_writers#4 &sb->s_type->i_mutex_key#8 &ei->i_data_sem stock_lock irq_context: 0 &f->f_pos_lock sb_writers#4 &sb->s_type->i_mutex_key#8 &ei->i_data_sem rcu_read_lock pool_lock#2 irq_context: 0 &f->f_pos_lock sb_writers#4 &sb->s_type->i_mutex_key#8 &ei->i_data_sem pcpu_lock stock_lock irq_context: 0 sk_lock-AF_BLUETOOTH-BTPROTO_RFCOMM rcu_node_0 irq_context: 0 sk_lock-AF_BLUETOOTH-BTPROTO_RFCOMM &rcu_state.expedited_wq irq_context: 0 sk_lock-AF_BLUETOOTH-BTPROTO_RFCOMM &rcu_state.expedited_wq &p->pi_lock irq_context: 0 sk_lock-AF_BLUETOOTH-BTPROTO_RFCOMM &rcu_state.expedited_wq &p->pi_lock &rq->__lock irq_context: 0 sk_lock-AF_BLUETOOTH-BTPROTO_RFCOMM &rcu_state.expedited_wq &p->pi_lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 sk_lock-AF_BLUETOOTH-BTPROTO_RFCOMM &cfs_rq->removed.lock irq_context: 0 &sb->s_type->i_mutex_key#10 sk_lock-AF_INET6/1 rcu_read_lock &cfs_rq->removed.lock irq_context: 0 &sb->s_type->i_mutex_key#10 sk_lock-AF_INET6/1 rcu_read_lock &obj_hash[i].lock irq_context: 0 sk_lock-AF_BLUETOOTH-BTPROTO_RFCOMM &rq->__lock &cfs_rq->removed.lock irq_context: 0 vlan_ioctl_mutex rtnl_mutex &macsec_netdev_addr_lock_key/1 irq_context: 0 vlan_ioctl_mutex rtnl_mutex &macsec_netdev_addr_lock_key/1 _xmit_ETHER irq_context: 0 vlan_ioctl_mutex rtnl_mutex &macsec_netdev_addr_lock_key/1 _xmit_ETHER &c->lock irq_context: 0 vlan_ioctl_mutex rtnl_mutex &macsec_netdev_addr_lock_key/1 _xmit_ETHER pool_lock#2 irq_context: 0 vlan_ioctl_mutex rtnl_mutex &xa->xa_lock#3 pool_lock#2 irq_context: 0 vlan_ioctl_mutex rtnl_mutex batched_entropy_u32.lock crngs.lock irq_context: softirq &(&local->roc_work)->timer irq_context: softirq &(&local->roc_work)->timer rcu_read_lock &pool->lock/1 irq_context: softirq &(&local->roc_work)->timer rcu_read_lock &pool->lock/1 &obj_hash[i].lock irq_context: softirq &(&local->roc_work)->timer rcu_read_lock &pool->lock/1 &p->pi_lock irq_context: softirq &(&local->roc_work)->timer rcu_read_lock &pool->lock/1 &p->pi_lock &rq->__lock irq_context: softirq &(&local->roc_work)->timer rcu_read_lock &pool->lock/1 &p->pi_lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)phy12 (work_completion)(&(&local->roc_work)->work) &local->mtx fs_reclaim irq_context: 0 (wq_completion)phy12 (work_completion)(&(&local->roc_work)->work) &local->mtx fs_reclaim mmu_notifier_invalidate_range_start irq_context: 0 (wq_completion)phy12 (work_completion)(&(&local->roc_work)->work) &local->mtx pool_lock#2 irq_context: 0 (wq_completion)phy12 (work_completion)(&(&local->roc_work)->work) &local->mtx &c->lock irq_context: 0 (wq_completion)phy12 (work_completion)(&(&local->roc_work)->work) &local->mtx nl_table_lock irq_context: 0 (wq_completion)phy12 (work_completion)(&(&local->roc_work)->work) &local->mtx nl_table_wait.lock irq_context: 0 &f->f_pos_lock sb_writers#4 &sb->s_type->i_mutex_key#8 &ei->i_data_sem &ei->i_es_lock batched_entropy_u8.lock irq_context: 0 &f->f_pos_lock sb_writers#4 &sb->s_type->i_mutex_key#8 &ei->i_data_sem &ei->i_es_lock kfence_freelist_lock irq_context: 0 key#26 irq_context: 0 sb_writers#3 &sb->s_type->i_mutex_key#9 per_cpu_ptr(&cgroup_rstat_cpu_lock, cpu) irq_context: 0 sb_writers#3 &sb->s_type->i_mutex_key#9 stock_lock rcu_read_lock per_cpu_ptr(&cgroup_rstat_cpu_lock, cpu) irq_context: 0 sk_lock-AF_INET6 k-sk_lock-AF_INET6 &tcp_hashinfo.bhash[i].lock &tcp_hashinfo.bhash2[i].lock per_cpu_ptr(&cgroup_rstat_cpu_lock, cpu) irq_context: 0 &sb->s_type->i_mutex_key#10 sk_lock-AF_INET6 key#26 irq_context: 0 sb_writers#4 &sb->s_type->i_mutex_key#8 &sbi->s_writepages_rwsem &n->list_lock irq_context: 0 sb_writers#4 &sb->s_type->i_mutex_key#8 &sbi->s_writepages_rwsem &n->list_lock &c->lock irq_context: 0 sb_writers#4 sb_internal jbd2_handle &ei->i_es_lock &meta->lock irq_context: 0 sb_writers#4 sb_internal jbd2_handle &ei->i_es_lock kfence_freelist_lock irq_context: 0 kn->active#60 fs_reclaim mmu_notifier_invalidate_range_start &rq->__lock irq_context: 0 kn->active#60 fs_reclaim mmu_notifier_invalidate_range_start &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 bpf_stats_enabled_mutex rcu_read_lock rcu_node_0 irq_context: 0 bpf_stats_enabled_mutex rcu_read_lock &rq->__lock irq_context: 0 bpf_stats_enabled_mutex rcu_read_lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 bpf_stats_enabled_mutex rcu_read_lock &rcu_state.gp_wq irq_context: 0 bpf_stats_enabled_mutex rcu_read_lock &rcu_state.gp_wq &p->pi_lock irq_context: 0 bpf_stats_enabled_mutex rcu_read_lock &rcu_state.gp_wq &p->pi_lock &rq->__lock irq_context: 0 bpf_stats_enabled_mutex rcu_read_lock &rcu_state.gp_wq &p->pi_lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 bpf_stats_enabled_mutex rcu_read_lock &rcu_state.expedited_wq irq_context: 0 bpf_stats_enabled_mutex rcu_read_lock &rcu_state.expedited_wq &p->pi_lock irq_context: 0 bpf_stats_enabled_mutex rcu_read_lock &rcu_state.expedited_wq &p->pi_lock &rq->__lock irq_context: 0 bpf_stats_enabled_mutex rcu_read_lock &rcu_state.expedited_wq &p->pi_lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 bpf_stats_enabled_mutex fs_reclaim &rq->__lock irq_context: 0 bpf_stats_enabled_mutex fs_reclaim &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 bpf_stats_enabled_mutex key#8 irq_context: 0 sb_writers#4 sb_internal jbd2_handle &sbi->s_orphan_lock rcu_read_lock &cfs_rq->removed.lock irq_context: 0 sb_writers#4 sb_internal jbd2_handle &sbi->s_orphan_lock rcu_read_lock &obj_hash[i].lock irq_context: 0 sb_writers#4 &type->i_mutex_dir_key#3 tomoyo_ss mount_lock irq_context: 0 sb_writers#4 &type->i_mutex_dir_key#3 tomoyo_ss mount_lock rcu_read_lock rename_lock.seqcount irq_context: 0 sk_lock-AF_ALG remove_cache_srcu rcu_read_lock &rcu_state.gp_wq irq_context: 0 sk_lock-AF_ALG remove_cache_srcu rcu_read_lock &rcu_state.gp_wq &p->pi_lock irq_context: 0 sk_lock-AF_ALG remove_cache_srcu rcu_read_lock &rcu_state.gp_wq &p->pi_lock &rq->__lock irq_context: 0 sk_lock-AF_ALG remove_cache_srcu rcu_read_lock &rcu_state.gp_wq &p->pi_lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 sk_lock-AF_ALG &mm->mmap_lock fs_reclaim irq_context: 0 sk_lock-AF_ALG &mm->mmap_lock fs_reclaim mmu_notifier_invalidate_range_start irq_context: 0 sk_lock-AF_ALG &mm->mmap_lock &____s->seqcount irq_context: 0 sk_lock-AF_ALG &mm->mmap_lock stock_lock irq_context: 0 sk_lock-AF_ALG &mm->mmap_lock rcu_read_lock pool_lock#2 irq_context: 0 &sb->s_type->i_mutex_key#10 sk_lock-AF_INET6 pgd_lock irq_context: 0 &sb->s_type->i_mutex_key#10 sk_lock-AF_INET6 key irq_context: 0 &sb->s_type->i_mutex_key#10 sk_lock-AF_INET6 pcpu_lock irq_context: 0 &sb->s_type->i_mutex_key#10 sk_lock-AF_INET6 percpu_counters_lock irq_context: 0 &sb->s_type->i_mutex_key#10 sk_lock-AF_INET6 pcpu_lock stock_lock irq_context: 0 &mm->mmap_lock fs_reclaim &rq->__lock &cfs_rq->removed.lock irq_context: 0 nfnl_subsys_cttimeout fs_reclaim irq_context: 0 nfnl_subsys_cttimeout fs_reclaim mmu_notifier_invalidate_range_start irq_context: 0 nfnl_subsys_cttimeout pool_lock#2 irq_context: 0 nfnl_subsys_cttimeout &obj_hash[i].lock irq_context: 0 nfnl_subsys_cttimeout krc.lock irq_context: 0 (wq_completion)events (work_completion)(&(&krcp->krw_arr[i].rcu_work)->work) rcu_callback rcu_node_0 irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem rtnl_mutex cpu_hotplug_lock &rq->__lock irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem rtnl_mutex cpu_hotplug_lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 rtnl_mutex gdp_mutex &root->kernfs_rwsem &sem->wait_lock irq_context: 0 rtnl_mutex gdp_mutex &root->kernfs_rwsem &rq->__lock irq_context: 0 rtnl_mutex gdp_mutex &root->kernfs_rwsem &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 rtnl_mutex team->team_lock_key#3 _xmit_ETHER &c->lock irq_context: 0 rtnl_mutex team->team_lock_key#3 _xmit_ETHER pool_lock#2 irq_context: 0 rtnl_mutex team->team_lock_key#3 &macvlan_netdev_addr_lock_key/1 irq_context: 0 rtnl_mutex team->team_lock_key#3 &macvlan_netdev_addr_lock_key/1 _xmit_ETHER irq_context: 0 rtnl_mutex team->team_lock_key#3 &macvlan_netdev_addr_lock_key/1 _xmit_ETHER pool_lock#2 irq_context: 0 rtnl_mutex team->team_lock_key#3 pcpu_alloc_mutex irq_context: 0 rtnl_mutex team->team_lock_key#3 pcpu_alloc_mutex pcpu_lock irq_context: 0 rtnl_mutex team->team_lock_key#3 &____s->seqcount#2 irq_context: 0 rtnl_mutex team->team_lock_key#3 rcu_read_lock rcu_read_lock rcu_node_0 irq_context: 0 rtnl_mutex team->team_lock_key#3 rcu_read_lock rcu_read_lock &rq->__lock irq_context: 0 rtnl_mutex team->team_lock_key#3 rcu_read_lock rcu_read_lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 rtnl_mutex team->team_lock_key#3 rcu_read_lock rcu_node_0 irq_context: 0 rtnl_mutex team->team_lock_key#3 rcu_read_lock &pool->lock/1 pool_lock#2 irq_context: 0 rtnl_mutex team->team_lock_key#3 rcu_read_lock pool_lock#2 irq_context: 0 rtnl_mutex team->team_lock_key#3 rcu_read_lock &rq->__lock irq_context: 0 rtnl_mutex team->team_lock_key#3 rcu_read_lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 rtnl_mutex team->team_lock_key#3 &macvlan_netdev_addr_lock_key/1 pool_lock#2 irq_context: 0 rtnl_mutex team->team_lock_key#3 &macvlan_netdev_addr_lock_key/1 _xmit_ETHER &c->lock irq_context: 0 rtnl_mutex team->team_lock_key#3 &macvlan_netdev_addr_lock_key/1 _xmit_ETHER &n->list_lock irq_context: 0 rtnl_mutex team->team_lock_key#3 &macvlan_netdev_addr_lock_key/1 _xmit_ETHER &n->list_lock &c->lock irq_context: 0 rtnl_mutex team->team_lock_key#3 &idev->mc_lock irq_context: 0 rtnl_mutex team->team_lock_key#3 &tb->tb6_lock irq_context: 0 rtnl_mutex team->team_lock_key#3 &tb->tb6_lock stock_lock irq_context: 0 rtnl_mutex team->team_lock_key#3 &tb->tb6_lock pool_lock#2 irq_context: 0 rtnl_mutex team->team_lock_key#3 &tb->tb6_lock &c->lock irq_context: 0 rtnl_mutex team->team_lock_key#3 &tb->tb6_lock rcu_read_lock pool_lock#2 irq_context: 0 rtnl_mutex team->team_lock_key#3 &tb->tb6_lock rcu_read_lock &data->fib_event_queue_lock irq_context: 0 rtnl_mutex team->team_lock_key#3 &tb->tb6_lock rcu_read_lock rcu_read_lock &pool->lock irq_context: 0 rtnl_mutex team->team_lock_key#3 &tb->tb6_lock rcu_read_lock rcu_read_lock &pool->lock &obj_hash[i].lock irq_context: 0 rtnl_mutex team->team_lock_key#3 &tb->tb6_lock rcu_read_lock rcu_read_lock &pool->lock pool_lock#2 irq_context: 0 rtnl_mutex team->team_lock_key#3 &tb->tb6_lock rcu_read_lock rcu_read_lock &pool->lock &p->pi_lock irq_context: 0 rtnl_mutex team->team_lock_key#3 &tb->tb6_lock rcu_read_lock rcu_read_lock &pool->lock &p->pi_lock &rq->__lock irq_context: 0 rtnl_mutex team->team_lock_key#3 &tb->tb6_lock rcu_read_lock rcu_read_lock &pool->lock &p->pi_lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 rtnl_mutex team->team_lock_key#3 &tb->tb6_lock nl_table_lock irq_context: 0 rtnl_mutex team->team_lock_key#3 &tb->tb6_lock &obj_hash[i].lock irq_context: 0 rtnl_mutex team->team_lock_key#3 &tb->tb6_lock nl_table_wait.lock irq_context: 0 rtnl_mutex team->team_lock_key#3 (inet6addr_validator_chain).rwsem irq_context: 0 rtnl_mutex team->team_lock_key#3 stock_lock irq_context: 0 rtnl_mutex team->team_lock_key#3 rcu_read_lock &net->ipv6.addrconf_hash_lock irq_context: 0 rtnl_mutex team->team_lock_key#3 rcu_read_lock &ndev->lock irq_context: 0 rtnl_mutex team->team_lock_key#3 rcu_read_lock &obj_hash[i].lock irq_context: 0 rtnl_mutex team->team_lock_key#3 rcu_read_lock &obj_hash[i].lock pool_lock irq_context: 0 rtnl_mutex team->team_lock_key#3 rcu_read_lock rcu_read_lock &pool->lock/1 irq_context: 0 rtnl_mutex team->team_lock_key#3 rcu_read_lock rcu_read_lock &pool->lock/1 &obj_hash[i].lock irq_context: 0 rtnl_mutex team->team_lock_key#3 rcu_read_lock rcu_read_lock &pool->lock/1 pool_lock#2 irq_context: 0 rtnl_mutex team->team_lock_key#3 rcu_read_lock rcu_read_lock &pool->lock/1 &p->pi_lock irq_context: 0 rtnl_mutex team->team_lock_key#3 rcu_read_lock rcu_read_lock &pool->lock/1 &p->pi_lock &rq->__lock irq_context: 0 rtnl_mutex team->team_lock_key#3 rcu_read_lock rcu_read_lock &pool->lock/1 &p->pi_lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 rtnl_mutex team->team_lock_key#3 rcu_read_lock &c->lock irq_context: 0 rtnl_mutex team->team_lock_key#3 rcu_read_lock &net->sctp.local_addr_lock irq_context: 0 rtnl_mutex team->team_lock_key#3 rcu_read_lock &net->sctp.local_addr_lock &net->sctp.addr_wq_lock irq_context: 0 rtnl_mutex team->team_lock_key#3 rcu_read_lock &net->sctp.local_addr_lock &net->sctp.addr_wq_lock &c->lock irq_context: 0 rtnl_mutex team->team_lock_key#3 rcu_read_lock &net->sctp.local_addr_lock &net->sctp.addr_wq_lock &n->list_lock irq_context: 0 rtnl_mutex team->team_lock_key#3 rcu_read_lock &net->sctp.local_addr_lock &net->sctp.addr_wq_lock &n->list_lock &c->lock irq_context: 0 rtnl_mutex team->team_lock_key#3 rcu_read_lock &net->sctp.local_addr_lock &net->sctp.addr_wq_lock pool_lock#2 irq_context: 0 rtnl_mutex team->team_lock_key#3 rcu_read_lock &net->sctp.local_addr_lock &net->sctp.addr_wq_lock &obj_hash[i].lock irq_context: 0 rtnl_mutex team->team_lock_key#3 rcu_read_lock &net->sctp.local_addr_lock &net->sctp.addr_wq_lock &base->lock irq_context: 0 rtnl_mutex team->team_lock_key#3 rcu_read_lock &net->sctp.local_addr_lock &net->sctp.addr_wq_lock &base->lock &obj_hash[i].lock irq_context: 0 rtnl_mutex team->team_lock_key#3 pcpu_lock irq_context: 0 rtnl_mutex team->team_lock_key#3 &ifa->lock irq_context: 0 rtnl_mutex team->team_lock_key#3 rcu_read_lock &pool->lock irq_context: 0 rtnl_mutex team->team_lock_key#3 rcu_read_lock &pool->lock &obj_hash[i].lock irq_context: 0 rtnl_mutex team->team_lock_key#3 rcu_read_lock &pool->lock pool_lock#2 irq_context: 0 rtnl_mutex team->team_lock_key#3 rcu_read_lock &pool->lock &p->pi_lock irq_context: 0 rtnl_mutex team->team_lock_key#3 &ndev->lock irq_context: 0 rtnl_mutex team->team_lock_key#3 &ndev->lock &ifa->lock irq_context: 0 rtnl_mutex team->team_lock_key#3 rcu_read_lock &tb->tb6_lock irq_context: 0 rtnl_mutex team->team_lock_key#3 rcu_read_lock &tb->tb6_lock &net->ipv6.fib6_walker_lock irq_context: 0 rtnl_mutex team->team_lock_key#3 &n->list_lock irq_context: 0 rtnl_mutex team->team_lock_key#3 &n->list_lock &c->lock irq_context: 0 sb_writers#11 &type->i_mutex_dir_key#7/1 &root->kernfs_rwsem &rq->__lock irq_context: 0 sb_writers#11 &type->i_mutex_dir_key#7/1 &root->kernfs_rwsem &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 rtnl_mutex team->team_lock_key#3 &pn->hash_lock irq_context: 0 rtnl_mutex team->team_lock_key#3 &dev->tx_global_lock irq_context: 0 rtnl_mutex team->team_lock_key#3 &dev->tx_global_lock &qdisc_xmit_lock_key irq_context: 0 rtnl_mutex team->team_lock_key#3 &dev->tx_global_lock &obj_hash[i].lock irq_context: 0 rtnl_mutex team->team_lock_key#3 rcu_state.exp_mutex rcu_node_0 irq_context: 0 rtnl_mutex team->team_lock_key#3 rcu_state.exp_mutex &obj_hash[i].lock irq_context: 0 rtnl_mutex team->team_lock_key#3 rcu_state.exp_mutex rcu_read_lock &pool->lock irq_context: 0 rtnl_mutex team->team_lock_key#3 rcu_state.exp_mutex rcu_read_lock &pool->lock &obj_hash[i].lock irq_context: 0 rtnl_mutex team->team_lock_key#3 rcu_state.exp_mutex rcu_read_lock &pool->lock &p->pi_lock irq_context: 0 rtnl_mutex team->team_lock_key#3 rcu_state.exp_mutex rcu_read_lock &pool->lock &p->pi_lock &rq->__lock irq_context: 0 rtnl_mutex team->team_lock_key#3 rcu_state.exp_mutex rcu_read_lock &pool->lock &p->pi_lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 rtnl_mutex team->team_lock_key#3 rcu_state.exp_mutex &rq->__lock irq_context: 0 rtnl_mutex team->team_lock_key#3 rcu_state.exp_mutex &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 rtnl_mutex team->team_lock_key#3 rcu_state.exp_mutex &rnp->exp_wq[0] irq_context: 0 rtnl_mutex team->team_lock_key#3 dev->qdisc_tx_busylock ?: &qdisc_tx_busylock#2 irq_context: 0 rtnl_mutex team->team_lock_key#3 dev->qdisc_tx_busylock ?: &qdisc_tx_busylock#2 &sch->q.lock irq_context: 0 rtnl_mutex team->team_lock_key#3 &sch->q.lock irq_context: 0 rtnl_mutex team->team_lock_key#3 &macvlan_netdev_addr_lock_key/1 _xmit_ETHER &obj_hash[i].lock irq_context: 0 rtnl_mutex team->team_lock_key#3 &macvlan_netdev_addr_lock_key/1 _xmit_ETHER krc.lock irq_context: 0 rtnl_mutex team->team_lock_key#3 _xmit_ETHER &obj_hash[i].lock irq_context: 0 rtnl_mutex team->team_lock_key#3 _xmit_ETHER krc.lock irq_context: 0 rtnl_mutex team->team_lock_key#3 rcu_state.exp_mutex &rnp->exp_wq[1] irq_context: 0 rtnl_mutex team->team_lock_key#3 __ip_vs_mutex irq_context: 0 rtnl_mutex team->team_lock_key#3 __ip_vs_mutex &ipvs->dest_trash_lock irq_context: 0 rtnl_mutex team->team_lock_key#3 &macvlan_netdev_addr_lock_key/1 &obj_hash[i].lock irq_context: 0 rtnl_mutex team->team_lock_key#3 &macvlan_netdev_addr_lock_key/1 krc.lock irq_context: 0 rtnl_mutex team->team_lock_key#3 krc.lock irq_context: 0 rtnl_mutex team->team_lock_key#3 &tbl->lock irq_context: 0 rtnl_mutex team->team_lock_key#3 class irq_context: 0 rtnl_mutex team->team_lock_key#3 (&tbl->proxy_timer) irq_context: 0 rtnl_mutex team->team_lock_key#3 &base->lock irq_context: 0 rtnl_mutex team->team_lock_key#3 flowtable_lock irq_context: 0 rtnl_mutex team->team_lock_key#3 &dir->lock irq_context: 0 rtnl_mutex team->team_lock_key#3 rcu_read_lock &pool->lock &p->pi_lock &rq->__lock irq_context: 0 rtnl_mutex team->team_lock_key#3 rcu_read_lock &pool->lock &p->pi_lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 rtnl_mutex team->team_lock_key#3 rcu_read_lock &tb->tb6_lock &net->ipv6.fib6_walker_lock irq_context: 0 rtnl_mutex team->team_lock_key#3 rcu_read_lock &tb->tb6_lock &obj_hash[i].lock irq_context: 0 rtnl_mutex team->team_lock_key#3 rcu_read_lock &tb->tb6_lock pool_lock#2 irq_context: 0 rtnl_mutex team->team_lock_key#3 rcu_read_lock &tb->tb6_lock rt6_exception_lock irq_context: 0 rtnl_mutex team->team_lock_key#3 rcu_read_lock &tb->tb6_lock rcu_read_lock &____s->seqcount irq_context: 0 rtnl_mutex team->team_lock_key#3 rcu_read_lock &tb->tb6_lock rcu_read_lock pool_lock#2 irq_context: 0 rtnl_mutex team->team_lock_key#3 rcu_read_lock &tb->tb6_lock rcu_read_lock &data->fib_event_queue_lock irq_context: 0 rtnl_mutex team->team_lock_key#3 rcu_read_lock &tb->tb6_lock rcu_read_lock rcu_read_lock &pool->lock irq_context: 0 rtnl_mutex team->team_lock_key#3 rcu_read_lock &tb->tb6_lock rcu_read_lock rcu_read_lock &pool->lock &obj_hash[i].lock irq_context: 0 rtnl_mutex team->team_lock_key#3 rcu_read_lock &tb->tb6_lock rcu_read_lock rcu_read_lock &pool->lock pool_lock#2 irq_context: 0 rtnl_mutex team->team_lock_key#3 rcu_read_lock &tb->tb6_lock nl_table_lock irq_context: 0 rtnl_mutex team->team_lock_key#3 rcu_read_lock &tb->tb6_lock nl_table_wait.lock irq_context: 0 rtnl_mutex team->team_lock_key#3 &net->ipv6.addrconf_hash_lock irq_context: 0 rtnl_mutex team->team_lock_key#3 &net->ipv6.addrconf_hash_lock &obj_hash[i].lock irq_context: 0 rtnl_mutex team->team_lock_key#3 &ndev->lock &obj_hash[i].lock irq_context: 0 rtnl_mutex team->team_lock_key#3 rcu_read_lock &dir->lock irq_context: 0 rtnl_mutex team->team_lock_key#3 rcu_read_lock rcu_read_lock &pool->lock irq_context: 0 rtnl_mutex team->team_lock_key#3 rcu_read_lock rcu_read_lock &pool->lock &obj_hash[i].lock irq_context: 0 rtnl_mutex team->team_lock_key#3 rcu_read_lock rcu_read_lock &pool->lock pool_lock#2 irq_context: 0 rtnl_mutex team->team_lock_key#3 rcu_read_lock krc.lock irq_context: 0 rtnl_mutex team->team_lock_key#3 &idev->mc_lock &macvlan_netdev_addr_lock_key/1 irq_context: 0 rtnl_mutex team->team_lock_key#3 &idev->mc_lock &macvlan_netdev_addr_lock_key/1 &obj_hash[i].lock irq_context: 0 rtnl_mutex team->team_lock_key#3 &idev->mc_lock &macvlan_netdev_addr_lock_key/1 pool_lock#2 irq_context: 0 rtnl_mutex team->team_lock_key#3 &idev->mc_lock &macvlan_netdev_addr_lock_key/1 krc.lock irq_context: 0 rtnl_mutex team->team_lock_key#3 rcu_state.exp_mutex &rnp->exp_wq[2] irq_context: 0 rtnl_mutex team->team_lock_key#3 &idev->mc_query_lock irq_context: 0 rtnl_mutex team->team_lock_key#3 &idev->mc_query_lock &obj_hash[i].lock irq_context: 0 rtnl_mutex team->team_lock_key#3 (work_completion)(&(&idev->mc_report_work)->work) irq_context: 0 rtnl_mutex team->team_lock_key#3 &net->ipv6.fib6_gc_lock rcu_read_lock &tb->tb6_lock irq_context: 0 rtnl_mutex team->team_lock_key#3 &net->ipv6.fib6_gc_lock &obj_hash[i].lock irq_context: 0 sb_writers#4 &sb->s_type->i_mutex_key#8 ima_extend_list_mutex &____s->seqcount#2 irq_context: 0 sb_writers#4 &sb->s_type->i_mutex_key#8 ima_extend_list_mutex &____s->seqcount irq_context: 0 rtnl_mutex &k->k_lock klist_remove_lock irq_context: 0 &type->i_mutex_dir_key/1 irq_context: 0 &type->i_mutex_dir_key/1 rename_lock.seqcount irq_context: 0 &type->i_mutex_dir_key/1 rcu_read_lock &dentry->d_lock irq_context: 0 &type->i_mutex_dir_key/1 &obj_hash[i].lock irq_context: 0 &type->i_mutex_dir_key/1 pool_lock#2 irq_context: 0 &type->i_mutex_dir_key/1 tomoyo_ss irq_context: 0 &type->i_mutex_dir_key/1 tomoyo_ss mmu_notifier_invalidate_range_start irq_context: 0 &type->i_mutex_dir_key/1 tomoyo_ss pool_lock#2 irq_context: 0 &type->i_mutex_dir_key/1 tomoyo_ss rcu_read_lock mount_lock.seqcount irq_context: 0 &type->i_mutex_dir_key/1 tomoyo_ss rcu_read_lock rcu_read_lock rename_lock.seqcount irq_context: 0 &type->i_mutex_dir_key/1 tomoyo_ss &obj_hash[i].lock irq_context: 0 &type->i_mutex_dir_key/1 tomoyo_ss rcu_read_lock &dentry->d_lock irq_context: 0 &type->i_mutex_dir_key/1 tomoyo_ss tomoyo_policy_lock irq_context: 0 &type->i_mutex_dir_key/1 tomoyo_ss tomoyo_policy_lock mmu_notifier_invalidate_range_start irq_context: 0 &type->i_mutex_dir_key/1 tomoyo_ss tomoyo_policy_lock pool_lock#2 irq_context: 0 &type->i_mutex_dir_key/1 &sb->s_type->i_mutex_key#4 irq_context: 0 &type->i_mutex_dir_key/1 &sb->s_type->i_mutex_key#4 tk_core.seq.seqcount irq_context: 0 &type->i_mutex_dir_key/1 &sb->s_type->i_mutex_key#4 &simple_offset_xa_lock irq_context: 0 &type->i_mutex_dir_key/1 &sb->s_type->i_mutex_key#4 rcu_read_lock &dentry->d_lock irq_context: 0 &type->i_mutex_dir_key/1 &sb->s_type->i_mutex_key#4 &dentry->d_lock irq_context: 0 &type->i_mutex_dir_key/1 &sb->s_type->i_lock_key#5 irq_context: 0 &type->i_mutex_dir_key/1 &sb->s_type->i_lock_key#5 &dentry->d_lock irq_context: 0 &type->i_mutex_dir_key/1 &fsnotify_mark_srcu irq_context: 0 &type->i_mutex_dir_key/1 &s->s_inode_list_lock irq_context: 0 &type->i_mutex_dir_key/1 &sbinfo->stat_lock irq_context: 0 &type->i_mutex_dir_key/1 &xa->xa_lock#9 irq_context: 0 &type->i_mutex_dir_key/1 &dentry->d_lock irq_context: 0 rtnl_mutex subsys mutex#82 &k->k_lock klist_remove_lock irq_context: 0 rtnl_mutex gdp_mutex sysfs_symlink_target_lock irq_context: 0 rtnl_mutex gdp_mutex kernfs_idr_lock irq_context: 0 rtnl_mutex gdp_mutex &obj_hash[i].lock irq_context: 0 pernet_ops_rwsem devices_rwsem rdma_nets_rwsem &device->compat_devs_mutex per_cpu_ptr(&cgroup_rstat_cpu_lock, cpu) irq_context: 0 pernet_ops_rwsem devices_rwsem rdma_nets_rwsem &device->compat_devs_mutex rcu_read_lock pool_lock#2 irq_context: softirq rcu_read_lock rcu_read_lock k-slock-AF_INET6 &n->list_lock irq_context: softirq rcu_read_lock rcu_read_lock k-slock-AF_INET6 &n->list_lock &c->lock irq_context: softirq rcu_read_lock rcu_read_lock k-slock-AF_INET6 &____s->seqcount#2 irq_context: 0 sk_lock-AF_INET &list->lock#24 irq_context: 0 sk_lock-AF_INET krc.lock irq_context: 0 cb_lock nlk_cb_mutex-GENERIC genl_mutex &bat_priv->gw.list_lock irq_context: 0 nfnl_subsys_cttimeout nf_conntrack_mutex irq_context: 0 nfnl_subsys_cttimeout nf_conntrack_mutex &rq->__lock irq_context: 0 nfnl_subsys_cttimeout nf_conntrack_mutex &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 nfnl_subsys_cttimeout nf_conntrack_mutex &nf_conntrack_locks[i] irq_context: 0 rtnl_mutex team->team_lock_key#3 pcpu_alloc_mutex &rq->__lock irq_context: 0 rtnl_mutex team->team_lock_key#3 pcpu_alloc_mutex &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 rtnl_mutex team->team_lock_key#3 rcu_state.exp_mutex &rnp->exp_wq[3] irq_context: 0 rtnl_mutex team->team_lock_key#3 rcu_read_lock &tb->tb6_lock rcu_read_lock rcu_read_lock &pool->lock &p->pi_lock irq_context: 0 rtnl_mutex team->team_lock_key#3 rcu_read_lock rcu_read_lock &pool->lock &p->pi_lock irq_context: 0 pcpu_alloc_mutex &rcu_state.expedited_wq irq_context: 0 pcpu_alloc_mutex &rcu_state.expedited_wq &p->pi_lock irq_context: 0 pcpu_alloc_mutex &rcu_state.expedited_wq &p->pi_lock &rq->__lock irq_context: 0 pcpu_alloc_mutex &rcu_state.expedited_wq &p->pi_lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 &type->i_mutex_dir_key/1 tomoyo_ss &c->lock irq_context: 0 &f->f_pos_lock sb_writers#4 &sb->s_type->i_mutex_key#8 rcu_read_lock fill_pool_map-wait-type-override rcu_node_0 irq_context: 0 &f->f_pos_lock sb_writers#4 &sb->s_type->i_mutex_key#8 rcu_read_lock fill_pool_map-wait-type-override &rq->__lock irq_context: 0 &f->f_pos_lock sb_writers#4 &sb->s_type->i_mutex_key#8 &mm->mmap_lock &dd->lock irq_context: 0 &f->f_pos_lock sb_writers#4 &sb->s_type->i_mutex_key#8 &mm->mmap_lock rcu_read_lock &pool->lock irq_context: 0 &f->f_pos_lock sb_writers#4 &sb->s_type->i_mutex_key#8 &mm->mmap_lock rcu_read_lock &pool->lock &obj_hash[i].lock irq_context: 0 &f->f_pos_lock sb_writers#4 &sb->s_type->i_mutex_key#8 &mm->mmap_lock rcu_read_lock &pool->lock pool_lock#2 irq_context: 0 &f->f_pos_lock sb_writers#4 &sb->s_type->i_mutex_key#8 &mm->mmap_lock rcu_read_lock &pool->lock &p->pi_lock irq_context: 0 &f->f_pos_lock sb_writers#4 &sb->s_type->i_mutex_key#8 &mm->mmap_lock rcu_read_lock &pool->lock &p->pi_lock &rq->__lock irq_context: 0 &f->f_pos_lock sb_writers#4 &sb->s_type->i_mutex_key#8 &mm->mmap_lock rcu_read_lock &pool->lock &p->pi_lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 rtnl_mutex team->team_lock_key#4 _xmit_ETHER &c->lock irq_context: 0 rtnl_mutex team->team_lock_key#4 _xmit_ETHER &n->list_lock irq_context: 0 rtnl_mutex team->team_lock_key#4 _xmit_ETHER &n->list_lock &c->lock irq_context: 0 rtnl_mutex team->team_lock_key#4 _xmit_ETHER (console_sem).lock irq_context: 0 rtnl_mutex team->team_lock_key#4 _xmit_ETHER console_lock console_srcu console_owner_lock irq_context: 0 rtnl_mutex team->team_lock_key#4 _xmit_ETHER console_lock console_srcu console_owner irq_context: 0 rtnl_mutex team->team_lock_key#4 _xmit_ETHER console_lock console_srcu console_owner &port_lock_key irq_context: 0 rtnl_mutex team->team_lock_key#4 _xmit_ETHER console_lock console_srcu console_owner console_owner_lock irq_context: 0 rtnl_mutex team->team_lock_key#4 &macvlan_netdev_addr_lock_key/1 irq_context: 0 rtnl_mutex team->team_lock_key#4 &macvlan_netdev_addr_lock_key/1 _xmit_ETHER irq_context: 0 rtnl_mutex team->team_lock_key#4 pcpu_alloc_mutex irq_context: 0 rtnl_mutex team->team_lock_key#4 pcpu_alloc_mutex pcpu_lock irq_context: 0 rtnl_mutex team->team_lock_key#4 &macvlan_netdev_addr_lock_key/1 &c->lock irq_context: 0 rtnl_mutex team->team_lock_key#4 &idev->mc_lock irq_context: 0 rtnl_mutex team->team_lock_key#4 &idev->mc_lock &rq->__lock irq_context: 0 rtnl_mutex team->team_lock_key#4 &tb->tb6_lock irq_context: 0 rtnl_mutex team->team_lock_key#4 &tb->tb6_lock nl_table_lock irq_context: 0 rtnl_mutex team->team_lock_key#4 &tb->tb6_lock &obj_hash[i].lock irq_context: 0 rtnl_mutex team->team_lock_key#4 &tb->tb6_lock nl_table_wait.lock irq_context: 0 rtnl_mutex team->team_lock_key#4 (inet6addr_validator_chain).rwsem irq_context: 0 rtnl_mutex team->team_lock_key#4 stock_lock irq_context: 0 rtnl_mutex team->team_lock_key#4 rcu_read_lock &net->ipv6.addrconf_hash_lock irq_context: 0 rtnl_mutex team->team_lock_key#4 rcu_read_lock &ndev->lock irq_context: 0 rtnl_mutex team->team_lock_key#4 rcu_read_lock &c->lock irq_context: 0 rtnl_mutex team->team_lock_key#4 rcu_read_lock &obj_hash[i].lock irq_context: 0 rtnl_mutex team->team_lock_key#4 rcu_read_lock rcu_read_lock &pool->lock/1 irq_context: 0 rtnl_mutex team->team_lock_key#4 rcu_read_lock rcu_read_lock &pool->lock/1 &obj_hash[i].lock irq_context: 0 rtnl_mutex team->team_lock_key#4 rcu_read_lock rcu_read_lock &pool->lock/1 &p->pi_lock irq_context: 0 rtnl_mutex team->team_lock_key#4 rcu_read_lock rcu_read_lock &pool->lock/1 &p->pi_lock &rq->__lock irq_context: 0 rtnl_mutex team->team_lock_key#4 rcu_read_lock rcu_read_lock &pool->lock/1 &p->pi_lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 rtnl_mutex team->team_lock_key#4 rcu_read_lock rcu_node_0 irq_context: 0 rtnl_mutex team->team_lock_key#4 rcu_read_lock &rq->__lock irq_context: 0 rtnl_mutex team->team_lock_key#4 rcu_read_lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 rtnl_mutex team->team_lock_key#4 rcu_read_lock &net->sctp.local_addr_lock irq_context: 0 rtnl_mutex team->team_lock_key#4 rcu_read_lock &net->sctp.local_addr_lock &net->sctp.addr_wq_lock irq_context: 0 rtnl_mutex team->team_lock_key#4 rcu_read_lock &net->sctp.local_addr_lock &net->sctp.addr_wq_lock &obj_hash[i].lock irq_context: 0 rtnl_mutex team->team_lock_key#4 rcu_read_lock &net->sctp.local_addr_lock &net->sctp.addr_wq_lock &base->lock irq_context: 0 rtnl_mutex team->team_lock_key#4 rcu_read_lock &net->sctp.local_addr_lock &net->sctp.addr_wq_lock &base->lock &obj_hash[i].lock irq_context: 0 rtnl_mutex team->team_lock_key#4 pcpu_lock irq_context: 0 rtnl_mutex team->team_lock_key#4 &tb->tb6_lock &c->lock irq_context: 0 rtnl_mutex team->team_lock_key#4 &ifa->lock irq_context: 0 rtnl_mutex team->team_lock_key#4 rcu_read_lock &pool->lock irq_context: 0 rtnl_mutex team->team_lock_key#4 rcu_read_lock &pool->lock &obj_hash[i].lock irq_context: 0 rtnl_mutex team->team_lock_key#4 rcu_read_lock &pool->lock &p->pi_lock irq_context: 0 rtnl_mutex team->team_lock_key#4 rcu_read_lock &pool->lock &p->pi_lock &rq->__lock irq_context: 0 rtnl_mutex team->team_lock_key#4 rcu_read_lock &pool->lock &p->pi_lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 rtnl_mutex team->team_lock_key#4 &ndev->lock irq_context: 0 rtnl_mutex team->team_lock_key#4 &ndev->lock &ifa->lock irq_context: 0 rtnl_mutex team->team_lock_key#4 rcu_read_lock &tb->tb6_lock irq_context: 0 rtnl_mutex team->team_lock_key#4 rcu_read_lock &tb->tb6_lock &net->ipv6.fib6_walker_lock irq_context: 0 rtnl_mutex team->team_lock_key#4 &pn->hash_lock irq_context: 0 rtnl_mutex team->team_lock_key#4 &dev->tx_global_lock irq_context: 0 rtnl_mutex team->team_lock_key#4 &dev->tx_global_lock &qdisc_xmit_lock_key irq_context: 0 rtnl_mutex team->team_lock_key#4 &dev->tx_global_lock &obj_hash[i].lock irq_context: 0 rtnl_mutex team->team_lock_key#4 rcu_state.exp_mutex rcu_node_0 irq_context: 0 rtnl_mutex team->team_lock_key#4 rcu_state.exp_mutex &obj_hash[i].lock irq_context: 0 rtnl_mutex team->team_lock_key#4 rcu_state.exp_mutex rcu_read_lock &pool->lock irq_context: 0 rtnl_mutex team->team_lock_key#4 rcu_state.exp_mutex rcu_read_lock &pool->lock &obj_hash[i].lock irq_context: 0 rtnl_mutex team->team_lock_key#4 rcu_state.exp_mutex rcu_read_lock &pool->lock &p->pi_lock irq_context: 0 rtnl_mutex team->team_lock_key#4 rcu_state.exp_mutex rcu_read_lock &pool->lock &p->pi_lock &rq->__lock irq_context: 0 rtnl_mutex team->team_lock_key#4 rcu_state.exp_mutex rcu_read_lock &pool->lock &p->pi_lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 rtnl_mutex team->team_lock_key#4 rcu_state.exp_mutex &rq->__lock irq_context: 0 rtnl_mutex team->team_lock_key#4 rcu_state.exp_mutex &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 rtnl_mutex team->team_lock_key#4 rcu_state.exp_mutex &rnp->exp_wq[3] irq_context: 0 &sb->s_type->i_mutex_key#8 remove_cache_srcu pool_lock#2 irq_context: 0 rtnl_mutex team->team_lock_key#4 dev->qdisc_tx_busylock ?: &qdisc_tx_busylock#2 irq_context: 0 rtnl_mutex team->team_lock_key#4 dev->qdisc_tx_busylock ?: &qdisc_tx_busylock#2 &sch->q.lock irq_context: 0 rtnl_mutex team->team_lock_key#4 &sch->q.lock irq_context: 0 rtnl_mutex team->team_lock_key#4 _xmit_ETHER &obj_hash[i].lock irq_context: 0 rtnl_mutex team->team_lock_key#4 _xmit_ETHER krc.lock irq_context: 0 rtnl_mutex team->team_lock_key#4 rcu_state.exp_mutex &rnp->exp_wq[0] irq_context: 0 rtnl_mutex team->team_lock_key#4 __ip_vs_mutex irq_context: 0 rtnl_mutex team->team_lock_key#4 __ip_vs_mutex &ipvs->dest_trash_lock irq_context: 0 rtnl_mutex team->team_lock_key#4 &macvlan_netdev_addr_lock_key/1 &obj_hash[i].lock irq_context: 0 rtnl_mutex team->team_lock_key#4 &macvlan_netdev_addr_lock_key/1 krc.lock irq_context: 0 rtnl_mutex team->team_lock_key#4 krc.lock irq_context: 0 rtnl_mutex team->team_lock_key#4 &tbl->lock irq_context: 0 rtnl_mutex team->team_lock_key#4 class irq_context: 0 rtnl_mutex team->team_lock_key#4 (&tbl->proxy_timer) irq_context: 0 rtnl_mutex team->team_lock_key#4 &base->lock irq_context: 0 rtnl_mutex team->team_lock_key#4 flowtable_lock irq_context: 0 rtnl_mutex team->team_lock_key#4 &dir->lock irq_context: 0 rtnl_mutex team->team_lock_key#4 rcu_read_lock &tb->tb6_lock &net->ipv6.fib6_walker_lock irq_context: 0 rtnl_mutex team->team_lock_key#4 rcu_read_lock &tb->tb6_lock rt6_exception_lock irq_context: 0 rtnl_mutex team->team_lock_key#4 rcu_read_lock &tb->tb6_lock nl_table_lock irq_context: 0 rtnl_mutex team->team_lock_key#4 rcu_read_lock &tb->tb6_lock &obj_hash[i].lock irq_context: 0 rtnl_mutex team->team_lock_key#4 rcu_read_lock &tb->tb6_lock nl_table_wait.lock irq_context: 0 rtnl_mutex team->team_lock_key#4 rcu_read_lock &tb->tb6_lock pool_lock#2 irq_context: 0 rtnl_mutex team->team_lock_key#4 rcu_read_lock &tb->tb6_lock &c->lock irq_context: 0 rtnl_mutex team->team_lock_key#4 &net->ipv6.addrconf_hash_lock irq_context: 0 rtnl_mutex team->team_lock_key#4 &net->ipv6.addrconf_hash_lock &obj_hash[i].lock irq_context: 0 rtnl_mutex team->team_lock_key#4 &ndev->lock &obj_hash[i].lock irq_context: 0 rtnl_mutex team->team_lock_key#4 rcu_read_lock &dir->lock irq_context: 0 rtnl_mutex team->team_lock_key#4 rcu_read_lock rcu_read_lock &pool->lock irq_context: 0 rtnl_mutex team->team_lock_key#4 rcu_read_lock rcu_read_lock &pool->lock &obj_hash[i].lock irq_context: 0 rtnl_mutex team->team_lock_key#4 rcu_read_lock rcu_read_lock &pool->lock &p->pi_lock irq_context: 0 rtnl_mutex team->team_lock_key#4 rcu_read_lock rcu_read_lock &pool->lock &p->pi_lock &rq->__lock irq_context: 0 rtnl_mutex team->team_lock_key#4 rcu_read_lock rcu_read_lock &pool->lock &p->pi_lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 rtnl_mutex team->team_lock_key#4 rcu_read_lock krc.lock irq_context: 0 rtnl_mutex team->team_lock_key#4 &idev->mc_lock &macvlan_netdev_addr_lock_key/1 irq_context: 0 rtnl_mutex team->team_lock_key#4 &idev->mc_lock &macvlan_netdev_addr_lock_key/1 &obj_hash[i].lock irq_context: 0 rtnl_mutex team->team_lock_key#4 &idev->mc_lock &macvlan_netdev_addr_lock_key/1 krc.lock irq_context: 0 rtnl_mutex team->team_lock_key#4 rcu_state.exp_mutex &rnp->exp_wq[1] irq_context: 0 rtnl_mutex team->team_lock_key#4 &idev->mc_query_lock irq_context: 0 rtnl_mutex team->team_lock_key#4 &idev->mc_query_lock &obj_hash[i].lock irq_context: 0 rtnl_mutex team->team_lock_key#4 (work_completion)(&(&idev->mc_report_work)->work) irq_context: 0 rtnl_mutex team->team_lock_key#4 &net->ipv6.fib6_gc_lock rcu_read_lock &tb->tb6_lock irq_context: 0 rtnl_mutex team->team_lock_key#4 &net->ipv6.fib6_gc_lock &obj_hash[i].lock irq_context: 0 &type->i_mutex_dir_key/1 tomoyo_ss &____s->seqcount#2 irq_context: 0 &type->i_mutex_dir_key/1 tomoyo_ss &____s->seqcount irq_context: 0 &type->i_mutex_dir_key/1 tomoyo_ss tomoyo_policy_lock &c->lock irq_context: 0 pernet_ops_rwsem k-sk_lock-AF_INET6 &tcp_hashinfo.bhash[i].lock &tcp_hashinfo.bhash2[i].lock &c->lock irq_context: 0 pernet_ops_rwsem &fsnotify_mark_srcu &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)mld (work_completion)(&(&idev->mc_dad_work)->work) &idev->mc_lock rcu_read_lock rcu_read_lock &meta->lock irq_context: 0 (wq_completion)mld (work_completion)(&(&idev->mc_dad_work)->work) &idev->mc_lock rcu_read_lock rcu_read_lock kfence_freelist_lock irq_context: 0 &kernfs_locks->open_file_mutex[count] quarantine_lock irq_context: 0 sk_lock-AF_ALG &mm->mmap_lock sb_pagefaults mapping.invalidate_lock rcu_read_lock &cfs_rq->removed.lock irq_context: 0 rcu_read_lock rcu_read_lock rcu_read_lock rcu_read_lock rcu_node_0 irq_context: 0 rcu_read_lock rcu_read_lock rcu_read_lock rcu_read_lock &rq->__lock irq_context: 0 (wq_completion)events_unbound (work_completion)(&(&kfence_timer)->work) cpu_hotplug_lock jump_label_mutex text_mutex &rcu_state.expedited_wq &p->pi_lock &rq->__lock irq_context: 0 (wq_completion)events_unbound (work_completion)(&(&kfence_timer)->work) cpu_hotplug_lock jump_label_mutex text_mutex &rcu_state.expedited_wq &p->pi_lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 rcu_read_lock rcu_read_lock rcu_read_lock rcu_read_lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem rtnl_mutex nl_table_wait.lock &p->pi_lock irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem rtnl_mutex sysctl_lock fill_pool_map-wait-type-override &n->list_lock irq_context: 0 tasklist_lock &sighand->siglock &(&sig->stats_lock)->lock &____s->seqcount#5 pidmap_lock fill_pool_map-wait-type-override pool_lock#2 irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem rtnl_mutex sysctl_lock fill_pool_map-wait-type-override &n->list_lock &c->lock irq_context: 0 tasklist_lock &sighand->siglock &(&sig->stats_lock)->lock &____s->seqcount#5 pidmap_lock fill_pool_map-wait-type-override &c->lock irq_context: 0 tasklist_lock &sighand->siglock &(&sig->stats_lock)->lock &____s->seqcount#5 pidmap_lock fill_pool_map-wait-type-override &n->list_lock irq_context: 0 tasklist_lock &sighand->siglock &(&sig->stats_lock)->lock &____s->seqcount#5 pidmap_lock fill_pool_map-wait-type-override &n->list_lock &c->lock irq_context: 0 tasklist_lock &sighand->siglock &(&sig->stats_lock)->lock &____s->seqcount#5 pidmap_lock fill_pool_map-wait-type-override pool_lock irq_context: 0 &f->f_pos_lock sb_writers#4 &sb->s_type->i_mutex_key#8 &mm->mmap_lock rcu_read_lock rcu_read_lock &obj_hash[i].lock irq_context: 0 &ping_table.lock irq_context: 0 pernet_ops_rwsem k-sk_lock-AF_NETLINK rcu_read_lock &cfs_rq->removed.lock irq_context: 0 defrag6_mutex irq_context: 0 pernet_ops_rwsem k-sk_lock-AF_NETLINK rcu_read_lock &obj_hash[i].lock irq_context: 0 pernet_ops_rwsem k-sk_lock-AF_NETLINK rcu_read_lock &obj_hash[i].lock pool_lock irq_context: 0 pernet_ops_rwsem k-sk_lock-AF_NETLINK rcu_read_lock pool_lock#2 irq_context: 0 pernet_ops_rwsem k-sk_lock-AF_NETLINK rcu_read_lock &rq->__lock &cfs_rq->removed.lock irq_context: 0 pernet_ops_rwsem devices_rwsem rdma_nets_rwsem &device->compat_devs_mutex fill_pool_map-wait-type-override pool_lock#2 irq_context: 0 pernet_ops_rwsem devices_rwsem rdma_nets_rwsem &device->compat_devs_mutex fill_pool_map-wait-type-override &c->lock irq_context: 0 pernet_ops_rwsem devices_rwsem rdma_nets_rwsem &device->compat_devs_mutex fill_pool_map-wait-type-override &n->list_lock irq_context: 0 pernet_ops_rwsem devices_rwsem rdma_nets_rwsem &device->compat_devs_mutex fill_pool_map-wait-type-override &n->list_lock &c->lock irq_context: 0 pernet_ops_rwsem devices_rwsem rdma_nets_rwsem &device->compat_devs_mutex fill_pool_map-wait-type-override pool_lock irq_context: 0 pernet_ops_rwsem k-sk_lock-AF_RXRPC fill_pool_map-wait-type-override pool_lock#2 irq_context: 0 pernet_ops_rwsem k-sk_lock-AF_RXRPC fill_pool_map-wait-type-override pool_lock irq_context: 0 pernet_ops_rwsem ovs_mutex nf_ct_proto_mutex defrag4_mutex &rq->__lock irq_context: softirq (&n->timer) &n->lock batched_entropy_u8.lock irq_context: softirq (&n->timer) &n->lock kfence_freelist_lock irq_context: 0 clock-AF_INET6 irq_context: 0 sk_lock-AF_INET6 k-sk_lock-AF_INET6 batched_entropy_u32.lock crngs.lock irq_context: 0 sb_writers#4 &type->i_mutex_dir_key#3/1 tomoyo_ss remove_cache_srcu &meta->lock irq_context: 0 sb_writers#4 &type->i_mutex_dir_key#3/1 tomoyo_ss remove_cache_srcu kfence_freelist_lock irq_context: 0 cb_lock genl_mutex k-sk_lock-AF_INET6 &____s->seqcount#8 irq_context: 0 cb_lock genl_mutex k-sk_lock-AF_INET6 batched_entropy_u32.lock irq_context: 0 cb_lock genl_mutex k-sk_lock-AF_INET6 &table->hash[i].lock irq_context: 0 cb_lock genl_mutex k-sk_lock-AF_INET6 &table->hash[i].lock k-clock-AF_INET6 irq_context: 0 cb_lock genl_mutex k-sk_lock-AF_INET6 &table->hash[i].lock &table->hash2[i].lock irq_context: 0 cb_lock genl_mutex &fn->fou_lock irq_context: 0 sb_writers#4 rcu_read_lock rcu_read_lock &cfs_rq->removed.lock irq_context: 0 sb_writers#4 rcu_read_lock rcu_read_lock &obj_hash[i].lock irq_context: 0 (wq_completion)wg-kex-wg2#9 (work_completion)(&peer->transmit_handshake_work) &peer->endpoint_lock rcu_read_lock_bh batched_entropy_u8.lock irq_context: 0 &pipe->mutex/1 &mm->mmap_lock rcu_read_lock &rcu_state.gp_wq irq_context: 0 &pipe->mutex/1 &mm->mmap_lock rcu_read_lock &rcu_state.gp_wq &p->pi_lock irq_context: 0 &pipe->mutex/1 &mm->mmap_lock rcu_read_lock &rcu_state.gp_wq &p->pi_lock &rq->__lock irq_context: 0 &pipe->mutex/1 &mm->mmap_lock rcu_read_lock &rcu_state.gp_wq &p->pi_lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 &sb->s_type->i_mutex_key#9 &dentry->d_lock &wq#2 irq_context: 0 sk_lock-AF_TIPC slock-AF_TIPC rcu_read_lock &ei->socket.wq.wait irq_context: 0 sk_lock-AF_TIPC slock-AF_TIPC rcu_read_lock &ei->socket.wq.wait &p->pi_lock irq_context: 0 sk_lock-AF_TIPC slock-AF_TIPC rcu_read_lock &ei->socket.wq.wait &p->pi_lock &rq->__lock irq_context: 0 sk_lock-AF_TIPC slock-AF_TIPC rcu_read_lock &ei->socket.wq.wait &p->pi_lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 sk_lock-AF_TIPC &mm->mmap_lock mmu_notifier_invalidate_range_start irq_context: 0 sk_lock-AF_TIPC &mm->mmap_lock ptlock_ptr(page)#2 lock#4 irq_context: 0 sk_lock-AF_TIPC &mm->mmap_lock ptlock_ptr(page)#2 lock#4 &lruvec->lru_lock irq_context: 0 sk_lock-AF_TIPC &mm->mmap_lock ptlock_ptr(page)#2 key irq_context: 0 unix_gc_lock irq_context: 0 &sb->s_type->i_mutex_key#10 unix_gc_lock irq_context: 0 &sb->s_type->i_mutex_key#10 unix_gc_lock unix_gc_wait.lock irq_context: 0 sk_lock-AF_TIPC &mm->mmap_lock &cfs_rq->removed.lock irq_context: 0 sk_lock-AF_TIPC &mm->mmap_lock fs_reclaim mmu_notifier_invalidate_range_start &rq->__lock irq_context: 0 sk_lock-AF_TIPC &mm->mmap_lock fs_reclaim mmu_notifier_invalidate_range_start &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 sk_lock-AF_TIPC &mm->mmap_lock fs_reclaim &rq->__lock irq_context: 0 sk_lock-AF_TIPC &mm->mmap_lock &pcp->lock &zone->lock irq_context: 0 sk_lock-AF_TIPC &mm->mmap_lock fs_reclaim &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 &sb->s_type->i_mutex_key#10 unix_gc_lock rlock-AF_UNIX irq_context: 0 sk_lock-AF_TIPC &mm->mmap_lock &rq->__lock &cfs_rq->removed.lock irq_context: 0 sk_lock-AF_TIPC &mm->mmap_lock &pcp->lock &zone->lock &____s->seqcount irq_context: 0 (wq_completion)events (work_completion)(&(&krcp->krw_arr[i].rcu_work)->work) rcu_callback rcu_read_lock &rcu_state.gp_wq irq_context: 0 (wq_completion)events (work_completion)(&(&krcp->krw_arr[i].rcu_work)->work) rcu_callback rcu_read_lock &rcu_state.gp_wq &p->pi_lock irq_context: 0 (wq_completion)events (work_completion)(&(&krcp->krw_arr[i].rcu_work)->work) rcu_callback rcu_read_lock &rcu_state.gp_wq &p->pi_lock &rq->__lock irq_context: 0 (wq_completion)events (work_completion)(&(&krcp->krw_arr[i].rcu_work)->work) rcu_callback rcu_read_lock &rcu_state.gp_wq &p->pi_lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)inet_frag_wq (work_completion)(&fqdir->destroy_work) rcu_read_lock rcu_node_0 irq_context: 0 (wq_completion)inet_frag_wq (work_completion)(&fqdir->destroy_work) rcu_read_lock &rq->__lock irq_context: 0 (wq_completion)inet_frag_wq (work_completion)(&fqdir->destroy_work) rcu_read_lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 sk_lock-AF_TIPC slock-AF_TIPC &meta->lock irq_context: 0 sk_lock-AF_TIPC slock-AF_TIPC kfence_freelist_lock irq_context: 0 sk_lock-AF_TIPC &mm->mmap_lock mmu_notifier_invalidate_range_start &rq->__lock irq_context: 0 sk_lock-AF_TIPC &mm->mmap_lock &obj_hash[i].lock irq_context: 0 sk_lock-AF_TIPC &mm->mmap_lock pool_lock#2 irq_context: softirq (&n->timer) k-slock-AF_INET6 rcu_read_lock batched_entropy_u8.lock irq_context: softirq (&n->timer) k-slock-AF_INET6 rcu_read_lock kfence_freelist_lock irq_context: 0 sk_lock-AF_INET k-sk_lock-AF_INET &n->list_lock irq_context: 0 sk_lock-AF_INET k-sk_lock-AF_INET &n->list_lock &c->lock irq_context: 0 &sb->s_type->i_mutex_key#10 sk_lock-AF_INET k-sk_lock-AF_INET/1 &rq->__lock irq_context: 0 &sb->s_type->i_mutex_key#10 sk_lock-AF_INET k-sk_lock-AF_INET/1 &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 sk_lock-AF_INET6 k-sk_lock-AF_INET6 &tcp_hashinfo.bhash[i].lock stock_lock irq_context: 0 sk_lock-AF_INET6 k-sk_lock-AF_INET6 &tcp_hashinfo.bhash[i].lock pool_lock#2 irq_context: 0 sk_lock-AF_INET6 k-sk_lock-AF_INET6 &tcp_hashinfo.bhash[i].lock &tcp_hashinfo.bhash2[i].lock batched_entropy_u8.lock irq_context: 0 sk_lock-AF_INET6 k-sk_lock-AF_INET6 &tcp_hashinfo.bhash[i].lock &tcp_hashinfo.bhash2[i].lock &hashinfo->ehash_locks[i] irq_context: 0 sk_lock-AF_INET6 k-sk_lock-AF_INET6 tk_core.seq.seqcount irq_context: 0 sk_lock-AF_INET6 k-sk_lock-AF_INET6 fs_reclaim irq_context: 0 sk_lock-AF_INET6 k-sk_lock-AF_INET6 fs_reclaim mmu_notifier_invalidate_range_start irq_context: 0 sk_lock-AF_INET6 k-sk_lock-AF_INET6 &c->lock irq_context: 0 sk_lock-AF_INET6 k-sk_lock-AF_INET6 pool_lock#2 irq_context: 0 sk_lock-AF_INET6 k-sk_lock-AF_INET6 rcu_read_lock rcu_read_lock &____s->seqcount#7 irq_context: 0 sk_lock-AF_INET6 k-sk_lock-AF_INET6 rcu_read_lock rcu_read_lock pool_lock#2 irq_context: 0 sk_lock-AF_INET6 k-sk_lock-AF_INET6 rcu_read_lock rcu_read_lock &ct->lock irq_context: 0 sk_lock-AF_INET6 k-sk_lock-AF_INET6 rcu_read_lock rcu_read_lock rcu_read_lock &____s->seqcount#7 irq_context: 0 sk_lock-AF_INET6 k-sk_lock-AF_INET6 rcu_read_lock rcu_read_lock &nf_nat_locks[i] irq_context: 0 sk_lock-AF_INET6 k-sk_lock-AF_INET6 rcu_read_lock rcu_read_lock &nf_conntrack_locks[i] irq_context: 0 sk_lock-AF_INET6 k-sk_lock-AF_INET6 rcu_read_lock rcu_read_lock &nf_conntrack_locks[i] &nf_conntrack_locks[i]/1 irq_context: 0 sk_lock-AF_INET6 k-sk_lock-AF_INET6 rcu_read_lock rcu_read_lock &nf_conntrack_locks[i] &nf_conntrack_locks[i]/1 batched_entropy_u8.lock irq_context: 0 sk_lock-AF_INET6 k-sk_lock-AF_INET6 rcu_read_lock rcu_read_lock &____s->seqcount#10 irq_context: 0 sk_lock-AF_INET6 k-sk_lock-AF_INET6 rcu_read_lock rcu_read_lock rcu_read_lock_bh &list->lock#5 irq_context: softirq rcu_read_lock rcu_read_lock rcu_read_lock rcu_read_lock rcu_read_lock rcu_read_lock_bh &list->lock#5 irq_context: 0 sk_lock-AF_INET6 k-sk_lock-AF_INET6 &obj_hash[i].lock irq_context: 0 sk_lock-AF_INET6 k-sk_lock-AF_INET6 &base->lock irq_context: 0 sk_lock-AF_INET6 k-sk_lock-AF_INET6 &base->lock &obj_hash[i].lock irq_context: 0 sk_lock-AF_INET6 k-sk_lock-AF_INET6 slock-AF_INET6 irq_context: softirq rcu_read_lock rcu_read_lock k-slock-AF_INET6 &tcp_hashinfo.bhash[i].lock &tcp_hashinfo.bhash2[i].lock &c->lock irq_context: 0 sk_lock-AF_INET6 slock-AF_INET6 rcu_read_lock &ei->socket.wq.wait irq_context: 0 (wq_completion)krdsd (work_completion)(&rtn->rds_tcp_accept_w) k-sk_lock-AF_INET6 tk_core.seq.seqcount irq_context: 0 (wq_completion)krdsd (work_completion)(&rtn->rds_tcp_accept_w) k-sk_lock-AF_INET6 rcu_read_lock rcu_read_lock &____s->seqcount#7 irq_context: 0 (wq_completion)krdsd (work_completion)(&rtn->rds_tcp_accept_w) k-sk_lock-AF_INET6 rcu_read_lock rcu_read_lock &ct->lock irq_context: 0 (wq_completion)krdsd (work_completion)(&rtn->rds_tcp_accept_w) k-sk_lock-AF_INET6 rcu_read_lock rcu_read_lock &____s->seqcount#10 irq_context: 0 (wq_completion)krdsd (work_completion)(&rtn->rds_tcp_accept_w) k-sk_lock-AF_INET6 rcu_read_lock rcu_read_lock rcu_read_lock_bh &list->lock#5 irq_context: softirq rcu_read_lock rcu_read_lock k-slock-AF_INET6/1 slock-AF_INET6 irq_context: softirq rcu_read_lock rcu_read_lock k-slock-AF_INET6/1 rcu_read_lock &pool->lock irq_context: softirq rcu_read_lock rcu_read_lock k-slock-AF_INET6/1 rcu_read_lock &pool->lock &obj_hash[i].lock irq_context: softirq rcu_read_lock rcu_read_lock k-slock-AF_INET6/1 rcu_read_lock &pool->lock pool_lock#2 irq_context: softirq rcu_read_lock rcu_read_lock k-slock-AF_INET6/1 rcu_read_lock &pool->lock &p->pi_lock irq_context: softirq rcu_read_lock rcu_read_lock k-slock-AF_INET6/1 rcu_read_lock &pool->lock &p->pi_lock &rq->__lock irq_context: softirq rcu_read_lock rcu_read_lock k-slock-AF_INET6/1 rcu_read_lock &pool->lock &p->pi_lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)events (work_completion)(&msk->work) sk_lock-AF_INET6 irq_context: 0 (wq_completion)events (work_completion)(&msk->work) sk_lock-AF_INET6 slock-AF_INET6 irq_context: 0 (wq_completion)events (work_completion)(&msk->work) slock-AF_INET6 irq_context: softirq (&icsk->icsk_delack_timer) k-slock-AF_INET6 tk_core.seq.seqcount irq_context: softirq (&icsk->icsk_delack_timer) k-slock-AF_INET6 rcu_read_lock rcu_read_lock &____s->seqcount#7 irq_context: softirq (&icsk->icsk_delack_timer) k-slock-AF_INET6 rcu_read_lock rcu_read_lock &ct->lock irq_context: softirq (&icsk->icsk_delack_timer) k-slock-AF_INET6 rcu_read_lock rcu_read_lock &____s->seqcount#10 irq_context: softirq (&icsk->icsk_delack_timer) k-slock-AF_INET6 rcu_read_lock rcu_read_lock rcu_read_lock_bh &list->lock#5 irq_context: softirq rcu_read_lock rcu_read_lock k-slock-AF_INET6/1 elock-AF_INET6 irq_context: 0 &sb->s_type->i_mutex_key#10 sk_lock-AF_INET6 k-sk_lock-AF_INET6 irq_context: 0 &sb->s_type->i_mutex_key#10 sk_lock-AF_INET6 k-sk_lock-AF_INET6 k-slock-AF_INET6 irq_context: 0 &sb->s_type->i_mutex_key#10 sk_lock-AF_INET6 k-sk_lock-AF_INET6 fs_reclaim irq_context: 0 &sb->s_type->i_mutex_key#10 sk_lock-AF_INET6 k-sk_lock-AF_INET6 fs_reclaim mmu_notifier_invalidate_range_start irq_context: 0 &sb->s_type->i_mutex_key#10 sk_lock-AF_INET6 k-sk_lock-AF_INET6 &c->lock irq_context: 0 &sb->s_type->i_mutex_key#10 sk_lock-AF_INET6 k-sk_lock-AF_INET6 pool_lock#2 irq_context: 0 &sb->s_type->i_mutex_key#10 sk_lock-AF_INET6 k-sk_lock-AF_INET6 &rq->__lock irq_context: 0 &sb->s_type->i_mutex_key#10 sk_lock-AF_INET6 k-sk_lock-AF_INET6 &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 &sb->s_type->i_mutex_key#10 sk_lock-AF_INET6 k-sk_lock-AF_INET6 tk_core.seq.seqcount irq_context: 0 &sb->s_type->i_mutex_key#10 sk_lock-AF_INET6 k-sk_lock-AF_INET6 rcu_read_lock rcu_read_lock &____s->seqcount#7 irq_context: 0 &sb->s_type->i_mutex_key#10 sk_lock-AF_INET6 k-sk_lock-AF_INET6 rcu_read_lock rcu_read_lock &ct->lock irq_context: 0 &sb->s_type->i_mutex_key#10 sk_lock-AF_INET6 k-sk_lock-AF_INET6 rcu_read_lock rcu_read_lock &____s->seqcount#10 irq_context: 0 &sb->s_type->i_mutex_key#10 sk_lock-AF_INET6 k-sk_lock-AF_INET6 rcu_read_lock rcu_read_lock rcu_read_lock_bh &list->lock#5 irq_context: softirq rcu_read_lock rcu_read_lock tk_core.seq.seqcount irq_context: softirq rcu_read_lock rcu_read_lock rcu_read_lock &ct->lock irq_context: 0 &sb->s_type->i_mutex_key#10 sk_lock-AF_INET6 k-sk_lock-AF_INET6 &obj_hash[i].lock irq_context: 0 &sb->s_type->i_mutex_key#10 sk_lock-AF_INET6 k-sk_lock-AF_INET6 &base->lock irq_context: 0 &sb->s_type->i_mutex_key#10 sk_lock-AF_INET6 k-sk_lock-AF_INET6 &base->lock &obj_hash[i].lock irq_context: 0 &sb->s_type->i_mutex_key#10 sk_lock-AF_INET6 k-sk_lock-AF_INET6 rcu_read_lock &pool->lock irq_context: 0 &sb->s_type->i_mutex_key#10 sk_lock-AF_INET6 k-sk_lock-AF_INET6 rcu_read_lock &pool->lock &obj_hash[i].lock irq_context: 0 &sb->s_type->i_mutex_key#10 sk_lock-AF_INET6 k-sk_lock-AF_INET6 rcu_read_lock &pool->lock pool_lock#2 irq_context: 0 &sb->s_type->i_mutex_key#10 sk_lock-AF_INET6 k-sk_lock-AF_INET6 rcu_read_lock &pool->lock &p->pi_lock irq_context: 0 &sb->s_type->i_mutex_key#10 sk_lock-AF_INET6 k-sk_lock-AF_INET6 rcu_read_lock &pool->lock &p->pi_lock &rq->__lock irq_context: 0 &sb->s_type->i_mutex_key#10 sk_lock-AF_INET6 k-sk_lock-AF_INET6 rcu_read_lock &pool->lock &p->pi_lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)events (work_completion)(&msk->work) sk_lock-AF_INET6 slock-AF_INET6 &sk->sk_lock.wq irq_context: 0 (wq_completion)events (work_completion)(&msk->work) sk_lock-AF_INET6 &pool->lock irq_context: 0 (wq_completion)events (work_completion)(&msk->work) sk_lock-AF_INET6 &rq->__lock irq_context: 0 (wq_completion)events (work_completion)(&msk->work) sk_lock-AF_INET6 &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 &sb->s_type->i_mutex_key#10 sk_lock-AF_INET6 k-sk_lock-AF_INET6 &____s->seqcount irq_context: 0 &sb->s_type->i_mutex_key#10 sk_lock-AF_INET6 k-sk_lock-AF_INET6 rcu_read_lock pool_lock#2 irq_context: 0 &sb->s_type->i_mutex_key#10 sk_lock-AF_INET6 k-sk_lock-AF_INET6 &hashinfo->ehash_locks[i] irq_context: 0 &sb->s_type->i_mutex_key#10 sk_lock-AF_INET6 k-sk_lock-AF_INET6 &tcp_hashinfo.bhash[i].lock irq_context: 0 &sb->s_type->i_mutex_key#10 sk_lock-AF_INET6 k-sk_lock-AF_INET6 &tcp_hashinfo.bhash[i].lock stock_lock irq_context: 0 &sb->s_type->i_mutex_key#10 sk_lock-AF_INET6 k-sk_lock-AF_INET6 &tcp_hashinfo.bhash[i].lock &obj_hash[i].lock irq_context: 0 &sb->s_type->i_mutex_key#10 sk_lock-AF_INET6 k-sk_lock-AF_INET6 &tcp_hashinfo.bhash[i].lock pool_lock#2 irq_context: 0 &sb->s_type->i_mutex_key#10 sk_lock-AF_INET6 k-sk_lock-AF_INET6 &tcp_hashinfo.bhash[i].lock &tcp_hashinfo.bhash2[i].lock irq_context: 0 &sb->s_type->i_mutex_key#10 sk_lock-AF_INET6 k-sk_lock-AF_INET6 &tcp_hashinfo.bhash[i].lock &tcp_hashinfo.bhash2[i].lock stock_lock irq_context: 0 &sb->s_type->i_mutex_key#10 sk_lock-AF_INET6 k-sk_lock-AF_INET6 &tcp_hashinfo.bhash[i].lock &tcp_hashinfo.bhash2[i].lock &obj_hash[i].lock irq_context: 0 &sb->s_type->i_mutex_key#10 sk_lock-AF_INET6 k-sk_lock-AF_INET6 &tcp_hashinfo.bhash[i].lock &tcp_hashinfo.bhash2[i].lock pool_lock#2 irq_context: 0 &sb->s_type->i_mutex_key#10 sk_lock-AF_INET6 &token_hash[i].lock irq_context: 0 &sb->s_type->i_mutex_key#10 slock-AF_INET6 &sk->sk_lock.wq irq_context: 0 &sb->s_type->i_mutex_key#10 slock-AF_INET6 &sk->sk_lock.wq &p->pi_lock irq_context: 0 &sb->s_type->i_mutex_key#10 slock-AF_INET6 &sk->sk_lock.wq &p->pi_lock &rq->__lock irq_context: 0 &sb->s_type->i_mutex_key#10 slock-AF_INET6 &sk->sk_lock.wq &p->pi_lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 sk_lock-AF_INET k-sk_lock-AF_INET rcu_read_lock fastopen_seqlock.seqcount irq_context: 0 sk_lock-AF_INET6 rcu_read_lock rcu_read_lock rcu_read_lock rcu_read_lock rcu_read_lock_bh &obj_hash[i].lock irq_context: 0 sk_lock-AF_INET6 rcu_read_lock rcu_read_lock rcu_read_lock rcu_read_lock rcu_read_lock_bh pool_lock#2 irq_context: 0 &sb->s_type->i_mutex_key#10 sk_lock-AF_INET rcu_read_lock rcu_read_lock pool_lock#2 irq_context: 0 &sb->s_type->i_mutex_key#10 sk_lock-AF_INET rcu_read_lock rcu_read_lock &dir->lock#2 irq_context: 0 &sb->s_type->i_mutex_key#10 sk_lock-AF_INET rcu_read_lock rcu_read_lock &ul->lock irq_context: 0 &sb->s_type->i_mutex_key#10 sk_lock-AF_INET rcu_read_lock rcu_read_lock rcu_read_lock_bh tk_core.seq.seqcount irq_context: 0 sk_lock-AF_INET6 rcu_read_lock rcu_read_lock rcu_read_lock &nf_nat_locks[i] irq_context: 0 rcu_read_lock rcu_read_lock rcu_read_lock rcu_read_lock_bh rcu_read_lock rcu_read_lock &table->lock#3 irq_context: 0 rcu_read_lock rcu_read_lock rcu_read_lock rcu_read_lock_bh rcu_read_lock rcu_read_lock_bh &obj_hash[i].lock irq_context: 0 rcu_read_lock rcu_read_lock rcu_read_lock rcu_read_lock_bh rcu_read_lock rcu_read_lock_bh pool_lock#2 irq_context: 0 rcu_read_lock rcu_read_lock rcu_read_lock rcu_read_lock_bh rcu_read_lock rcu_read_lock_bh rcu_read_lock tk_core.seq.seqcount irq_context: 0 rcu_read_lock rcu_read_lock rcu_read_lock rcu_read_lock_bh rcu_read_lock rcu_read_lock_bh rcu_read_lock &list->lock#5 irq_context: 0 (wq_completion)events (work_completion)(&(&nsim_dev->trap_data->trap_report_dw)->work) fill_pool_map-wait-type-override &____s->seqcount#2 irq_context: 0 (wq_completion)events (work_completion)(&(&nsim_dev->trap_data->trap_report_dw)->work) fill_pool_map-wait-type-override &____s->seqcount irq_context: 0 &drbg->drbg_mutex irq_context: 0 sb_writers#4 &type->i_mutex_dir_key#3/1 jbd2_handle &ei->i_data_sem remove_cache_srcu rcu_read_lock rcu_node_0 irq_context: 0 sb_writers#4 &type->i_mutex_dir_key#3/1 jbd2_handle &ei->i_data_sem remove_cache_srcu rcu_read_lock &rq->__lock irq_context: 0 sb_writers#4 &type->i_mutex_dir_key#3/1 jbd2_handle &ei->i_data_sem remove_cache_srcu rcu_read_lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 sk_lock-AF_INET6 rcu_read_lock k-slock-AF_INET6 rcu_read_lock rcu_read_lock rcu_read_lock stock_lock irq_context: 0 sk_lock-AF_INET6 rcu_read_lock k-slock-AF_INET6 rcu_read_lock rcu_read_lock rcu_read_lock pool_lock#2 irq_context: 0 sk_lock-AF_INET6 rcu_read_lock k-slock-AF_INET6 rcu_read_lock rcu_read_lock rcu_read_lock &dir->lock#2 irq_context: 0 &sb->s_type->i_mutex_key#10 sk_lock-AF_INET6/1 remove_cache_srcu &base->lock irq_context: 0 &sb->s_type->i_mutex_key#10 sk_lock-AF_INET6/1 remove_cache_srcu &base->lock &obj_hash[i].lock irq_context: 0 rcu_read_lock rcu_read_lock_bh rcu_read_lock rcu_read_lock &table->lock#3 irq_context: 0 rcu_read_lock rcu_read_lock_bh rcu_read_lock rcu_read_lock_bh rcu_read_lock tk_core.seq.seqcount irq_context: 0 rcu_read_lock rcu_read_lock_bh rcu_read_lock rcu_read_lock_bh rcu_read_lock &list->lock#5 irq_context: 0 sb_writers#4 fill_pool_map-wait-type-override &____s->seqcount#2 irq_context: 0 sb_writers#4 fill_pool_map-wait-type-override &____s->seqcount irq_context: softirq (&n->timer) rcu_read_lock rcu_read_lock rcu_read_lock_bh rcu_read_lock rcu_read_lock &table->lock#3 irq_context: softirq (&n->timer) rcu_read_lock rcu_read_lock rcu_read_lock_bh rcu_read_lock rcu_read_lock_bh fill_pool_map-wait-type-override pool_lock#2 irq_context: softirq (&n->timer) rcu_read_lock rcu_read_lock rcu_read_lock_bh rcu_read_lock rcu_read_lock_bh fill_pool_map-wait-type-override pool_lock irq_context: softirq (&n->timer) rcu_read_lock rcu_read_lock rcu_read_lock_bh rcu_read_lock rcu_read_lock_bh &obj_hash[i].lock irq_context: softirq (&n->timer) rcu_read_lock rcu_read_lock rcu_read_lock_bh rcu_read_lock rcu_read_lock_bh pool_lock#2 irq_context: softirq (&n->timer) rcu_read_lock rcu_read_lock rcu_read_lock_bh rcu_read_lock rcu_read_lock_bh rcu_read_lock tk_core.seq.seqcount irq_context: softirq (&n->timer) rcu_read_lock rcu_read_lock rcu_read_lock_bh rcu_read_lock rcu_read_lock_bh rcu_read_lock &list->lock#5 irq_context: 0 sk_lock-AF_INET6 rcu_read_lock rcu_read_lock &pool->lock fill_pool_map-wait-type-override pool_lock#2 irq_context: 0 sk_lock-AF_INET6 rcu_read_lock rcu_read_lock &pool->lock fill_pool_map-wait-type-override pool_lock irq_context: softirq &(&bat_priv->mcast.work)->timer rcu_read_lock &pool->lock/1 fill_pool_map-wait-type-override pool_lock irq_context: 0 sk_lock-AF_INET6 sctp_assocs_id_lock &n->list_lock irq_context: 0 sk_lock-AF_INET6 sctp_assocs_id_lock &n->list_lock &c->lock irq_context: softirq (&n->timer) rcu_read_lock_bh rcu_read_lock rcu_read_lock &table->lock#3 irq_context: softirq (&n->timer) rcu_read_lock_bh rcu_read_lock rcu_read_lock_bh rcu_read_lock tk_core.seq.seqcount irq_context: softirq (&n->timer) rcu_read_lock_bh rcu_read_lock rcu_read_lock_bh rcu_read_lock &list->lock#5 irq_context: 0 rcu_read_lock &ndev->lock irq_context: 0 k-slock-AF_INET6 rcu_read_lock pool_lock#2 irq_context: 0 k-slock-AF_INET6 rcu_read_lock &c->lock irq_context: 0 k-slock-AF_INET6 rcu_read_lock rcu_read_lock &____s->seqcount#10 irq_context: 0 k-slock-AF_INET6 rcu_read_lock rcu_read_lock rcu_read_lock_bh tk_core.seq.seqcount irq_context: 0 k-slock-AF_INET6 rcu_read_lock rcu_read_lock rcu_read_lock_bh &list->lock#5 irq_context: softirq (&app->join_timer) rcu_read_lock_bh &base->lock irq_context: softirq (&app->join_timer) rcu_read_lock_bh &base->lock &obj_hash[i].lock irq_context: 0 sb_writers#4 &type->i_mutex_dir_key#3/1 &type->i_mutex_dir_key#3 remove_cache_srcu rcu_read_lock &rcu_state.gp_wq irq_context: 0 sb_writers#4 &type->i_mutex_dir_key#3/1 &type->i_mutex_dir_key#3 remove_cache_srcu rcu_read_lock &rcu_state.gp_wq &p->pi_lock irq_context: 0 sb_writers#4 &type->i_mutex_dir_key#3/1 &type->i_mutex_dir_key#3 remove_cache_srcu rcu_read_lock &rcu_state.gp_wq &p->pi_lock &rq->__lock irq_context: 0 sb_writers#4 &type->i_mutex_dir_key#3/1 &type->i_mutex_dir_key#3 remove_cache_srcu rcu_read_lock &rcu_state.gp_wq &p->pi_lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 dup_mmap_sem &mm->mmap_lock &mm->mmap_lock/1 remove_cache_srcu pgd_lock irq_context: 0 dup_mmap_sem &mm->mmap_lock &mm->mmap_lock/1 remove_cache_srcu stock_lock irq_context: 0 dup_mmap_sem &mm->mmap_lock &mm->mmap_lock/1 remove_cache_srcu key irq_context: 0 dup_mmap_sem &mm->mmap_lock &mm->mmap_lock/1 remove_cache_srcu pcpu_lock irq_context: 0 dup_mmap_sem &mm->mmap_lock &mm->mmap_lock/1 remove_cache_srcu percpu_counters_lock irq_context: 0 dup_mmap_sem &mm->mmap_lock &mm->mmap_lock/1 remove_cache_srcu pcpu_lock stock_lock irq_context: 0 &kernfs_locks->open_file_mutex[count] rcu_read_lock &rcu_state.gp_wq irq_context: 0 &kernfs_locks->open_file_mutex[count] rcu_read_lock &rcu_state.gp_wq &p->pi_lock irq_context: 0 &kernfs_locks->open_file_mutex[count] rcu_read_lock &rcu_state.gp_wq &p->pi_lock &rq->__lock irq_context: 0 &kernfs_locks->open_file_mutex[count] rcu_read_lock &rcu_state.gp_wq &p->pi_lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 sb_writers#4 &type->i_mutex_dir_key#3/1 tomoyo_ss rcu_read_lock rcu_read_lock &cfs_rq->removed.lock irq_context: 0 rcu_read_lock &tbl->lock batched_entropy_u32.lock irq_context: 0 rcu_read_lock rcu_read_lock_bh rcu_read_lock &list->lock#5 irq_context: softirq rcu_read_lock rcu_read_lock &im->lock irq_context: 0 ppp_mutex rtnl_mutex &idev->mc_lock &rq->__lock irq_context: 0 ppp_mutex rtnl_mutex &idev->mc_lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 ppp_mutex &pn->all_ppp_mutex irq_context: 0 dup_mmap_sem &mm->mmap_lock &mm->mmap_lock/1 &anon_vma->rwsem rcu_read_lock &cfs_rq->removed.lock irq_context: 0 rcu_read_lock rcu_read_lock_bh rcu_read_lock &ei->socket.wq.wait irq_context: softirq rcu_read_lock rcu_read_lock k-slock-AF_INET rcu_read_lock &____s->seqcount#2 irq_context: softirq rcu_read_lock rcu_read_lock k-slock-AF_INET rcu_read_lock &____s->seqcount irq_context: 0 sb_writers#4 &type->i_mutex_dir_key#3/1 jbd2_handle &n->list_lock irq_context: 0 sb_writers#4 &type->i_mutex_dir_key#3/1 jbd2_handle &n->list_lock &c->lock irq_context: softirq (&n->timer) &obj_hash[i].lock pool_lock irq_context: softirq &(&group->avgs_work)->timer rcu_read_lock &pool->lock &p->pi_lock &rq->__lock &obj_hash[i].lock irq_context: softirq &(&group->avgs_work)->timer rcu_read_lock &pool->lock &p->pi_lock &rq->__lock &base->lock irq_context: softirq &(&group->avgs_work)->timer rcu_read_lock &pool->lock &p->pi_lock &rq->__lock &base->lock &obj_hash[i].lock irq_context: 0 sk_lock-AF_INET rcu_read_lock rcu_read_lock_bh rcu_read_lock rcu_read_lock &table->lock#3 irq_context: 0 sk_lock-AF_INET rcu_read_lock rcu_read_lock_bh rcu_read_lock rcu_read_lock_bh rcu_read_lock tk_core.seq.seqcount irq_context: 0 sk_lock-AF_INET rcu_read_lock rcu_read_lock_bh rcu_read_lock rcu_read_lock_bh rcu_read_lock &list->lock#5 irq_context: 0 &audit_cmd_mutex.lock &n->list_lock irq_context: 0 &audit_cmd_mutex.lock &n->list_lock &c->lock irq_context: softirq rcu_read_lock_bh rcu_read_lock &pool->lock/1 &p->pi_lock &cfs_rq->removed.lock irq_context: 0 (wq_completion)wg-kex-wg1#10 (work_completion)(&peer->transmit_handshake_work) &c->lock irq_context: softirq rcu_read_lock rcu_read_lock rcu_read_lock &pool->lock &p->pi_lock &rq->__lock &cfs_rq->removed.lock irq_context: 0 rtnl_mutex sk_lock-AF_INET _xmit_ETHER irq_context: 0 rtnl_mutex sk_lock-AF_INET _xmit_ETHER &c->lock irq_context: 0 rtnl_mutex sk_lock-AF_INET _xmit_ETHER pool_lock#2 irq_context: 0 &sb->s_type->i_mutex_key#10 rtnl_mutex rcu_read_lock &im->lock irq_context: 0 &sb->s_type->i_mutex_key#10 rtnl_mutex &im->lock irq_context: 0 &sb->s_type->i_mutex_key#10 rtnl_mutex _xmit_ETHER irq_context: 0 &sb->s_type->i_mutex_key#10 rtnl_mutex _xmit_ETHER &obj_hash[i].lock irq_context: 0 &sb->s_type->i_mutex_key#10 rtnl_mutex _xmit_ETHER pool_lock#2 irq_context: 0 &sb->s_type->i_mutex_key#10 rtnl_mutex _xmit_ETHER krc.lock irq_context: 0 &sb->s_type->i_mutex_key#10 rtnl_mutex &obj_hash[i].lock pool_lock irq_context: softirq (&n->timer) quarantine_lock irq_context: 0 rtnl_mutex &rdev->wiphy.mtx &wdev->event_lock irq_context: 0 rtnl_mutex &rdev->wiphy.mtx &rdev->mgmt_registrations_lock irq_context: 0 rtnl_mutex &rdev->wiphy.mtx &dentry->d_lock irq_context: 0 rtnl_mutex &rdev->wiphy.mtx &sb->s_type->i_mutex_key#3 &dentry->d_lock &dentry->d_lock/1 irq_context: 0 rtnl_mutex &rdev->wiphy.mtx &sb->s_type->i_mutex_key#3 &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 rtnl_mutex &rdev->wiphy.mtx &sb->s_type->i_mutex_key#3 pin_fs_lock irq_context: 0 rtnl_mutex &rdev->wiphy.mtx &sb->s_type->i_mutex_key#3 rcu_read_lock mount_lock irq_context: 0 rtnl_mutex &rdev->wiphy.mtx &sb->s_type->i_mutex_key#3 rcu_read_lock mount_lock mount_lock.seqcount irq_context: 0 rtnl_mutex &rdev->wiphy.mtx &sb->s_type->i_mutex_key#3 mount_lock irq_context: 0 rtnl_mutex &rdev->wiphy.mtx &sb->s_type->i_mutex_key#3 mount_lock mount_lock.seqcount irq_context: 0 rtnl_mutex &rdev->wiphy.mtx &sb->s_type->i_mutex_key#3 rcu_read_lock &dentry->d_lock irq_context: 0 rtnl_mutex &rdev->wiphy.mtx &sb->s_type->i_mutex_key#3 &fsnotify_mark_srcu irq_context: 0 rtnl_mutex &rdev->wiphy.mtx &sb->s_type->i_mutex_key#3 &xa->xa_lock#9 irq_context: 0 rtnl_mutex &rdev->wiphy.mtx &sb->s_type->i_mutex_key#3 &obj_hash[i].lock irq_context: 0 rtnl_mutex &rdev->wiphy.mtx &sb->s_type->i_mutex_key#3 &obj_hash[i].lock pool_lock irq_context: 0 rtnl_mutex &rdev->wiphy.mtx rcu_read_lock &dentry->d_lock irq_context: 0 rtnl_mutex &rdev->wiphy.mtx &fsnotify_mark_srcu irq_context: 0 rtnl_mutex &rdev->wiphy.mtx &sb->s_type->i_lock_key#7 irq_context: 0 rtnl_mutex &rdev->wiphy.mtx &s->s_inode_list_lock irq_context: 0 rtnl_mutex &rdev->wiphy.mtx &xa->xa_lock#9 irq_context: 0 rtnl_mutex &rdev->wiphy.mtx rcu_read_lock mount_lock irq_context: 0 rtnl_mutex &rdev->wiphy.mtx rcu_read_lock mount_lock mount_lock.seqcount irq_context: 0 rtnl_mutex &rdev->wiphy.mtx mount_lock irq_context: 0 rtnl_mutex &rdev->wiphy.mtx mount_lock mount_lock.seqcount irq_context: 0 rtnl_mutex &rdev->wiphy.mtx (work_completion)(&(&link->color_collision_detect_work)->work) irq_context: 0 rtnl_mutex &rdev->wiphy.mtx &local->chanctx_mtx irq_context: 0 rtnl_mutex &rdev->wiphy.mtx &sb->s_type->i_mutex_key#3 &xa->xa_lock#4 irq_context: 0 rtnl_mutex &rdev->wiphy.mtx &sb->s_type->i_mutex_key#3 stock_lock irq_context: 0 rtnl_mutex &rdev->wiphy.mtx &sb->s_type->i_mutex_key#3 &n->list_lock irq_context: 0 rtnl_mutex &rdev->wiphy.mtx &sb->s_type->i_mutex_key#3 &n->list_lock &c->lock irq_context: 0 rtnl_mutex &rdev->wiphy.mtx &sb->s_type->i_mutex_key#3 rcu_read_lock pool_lock#2 irq_context: 0 rtnl_mutex &rdev->wiphy.mtx &sb->s_type->i_mutex_key#3 &____s->seqcount#2 irq_context: 0 rtnl_mutex &rdev->wiphy.mtx &sb->s_type->i_mutex_key#3 rcu_read_lock rcu_node_0 irq_context: 0 rtnl_mutex &rdev->wiphy.mtx &sb->s_type->i_mutex_key#3 rcu_read_lock &rq->__lock irq_context: 0 rtnl_mutex &rdev->wiphy.mtx &sb->s_type->i_mutex_key#3 rcu_read_lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 sk_lock-AF_BLUETOOTH-BTPROTO_HCI mgmt_chan_list_lock &obj_hash[i].lock pool_lock irq_context: 0 sk_lock-AF_BLUETOOTH-BTPROTO_HCI mgmt_chan_list_lock &hdev->lock &hdev->unregister_lock irq_context: 0 sk_lock-AF_BLUETOOTH-BTPROTO_HCI mgmt_chan_list_lock &hdev->lock &hdev->unregister_lock fs_reclaim irq_context: 0 sk_lock-AF_BLUETOOTH-BTPROTO_HCI mgmt_chan_list_lock &hdev->lock &hdev->unregister_lock fs_reclaim mmu_notifier_invalidate_range_start irq_context: 0 sk_lock-AF_BLUETOOTH-BTPROTO_HCI mgmt_chan_list_lock &hdev->lock &hdev->unregister_lock pool_lock#2 irq_context: 0 sk_lock-AF_BLUETOOTH-BTPROTO_HCI mgmt_chan_list_lock &hdev->lock &hdev->unregister_lock &hdev->cmd_sync_work_lock irq_context: 0 sk_lock-AF_BLUETOOTH-BTPROTO_HCI mgmt_chan_list_lock &hdev->lock &hdev->unregister_lock rcu_read_lock &pool->lock/1 irq_context: 0 sk_lock-AF_BLUETOOTH-BTPROTO_HCI mgmt_chan_list_lock &hdev->lock &hdev->unregister_lock rcu_read_lock &pool->lock/1 &obj_hash[i].lock irq_context: 0 sk_lock-AF_BLUETOOTH-BTPROTO_HCI mgmt_chan_list_lock &hdev->lock &hdev->unregister_lock rcu_read_lock &pool->lock/1 &p->pi_lock irq_context: 0 sk_lock-AF_BLUETOOTH-BTPROTO_HCI mgmt_chan_list_lock &hdev->lock &hdev->unregister_lock rcu_read_lock &pool->lock/1 &p->pi_lock &rq->__lock irq_context: 0 sk_lock-AF_BLUETOOTH-BTPROTO_HCI mgmt_chan_list_lock &hdev->lock &hdev->unregister_lock rcu_read_lock &pool->lock/1 &p->pi_lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 sk_lock-AF_BLUETOOTH-BTPROTO_HCI mgmt_chan_list_lock &hdev->lock &hdev->unregister_lock &rq->__lock irq_context: 0 sk_lock-AF_BLUETOOTH-BTPROTO_HCI mgmt_chan_list_lock &hdev->lock &hdev->unregister_lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 sb_writers#4 &type->i_mutex_dir_key#3/1 sb_internal jbd2_handle &ei->i_data_sem &ei->i_es_lock quarantine_lock irq_context: 0 &sb->s_type->i_lock_key#23 rcu_read_lock &dentry->d_lock irq_context: 0 &sb->s_type->i_lock_key#23 &dentry->d_lock &dentry->d_lock/1 irq_context: 0 &sb->s_type->i_lock_key#23 &dentry->d_lock/1 irq_context: 0 sk_lock-AF_INET remove_cache_srcu &cfs_rq->removed.lock irq_context: 0 sb_writers#4 sb_internal jbd2_handle rcu_read_lock pgd_lock irq_context: 0 sb_writers#4 sb_internal jbd2_handle rcu_read_lock stock_lock irq_context: 0 sb_writers#4 sb_internal jbd2_handle rcu_read_lock key irq_context: 0 sb_writers#4 sb_internal jbd2_handle rcu_read_lock pcpu_lock irq_context: 0 sb_writers#4 sb_internal jbd2_handle rcu_read_lock percpu_counters_lock irq_context: 0 sb_writers#4 sb_internal jbd2_handle rcu_read_lock pcpu_lock stock_lock irq_context: 0 nfnl_subsys_ctnetlink nf_conntrack_mutex irq_context: 0 nfnl_subsys_ctnetlink nf_conntrack_mutex &nf_conntrack_locks[i] irq_context: 0 nfnl_subsys_ctnetlink nf_conntrack_mutex &____s->seqcount#7 irq_context: 0 nfnl_subsys_ctnetlink nf_conntrack_mutex rcu_read_lock &nf_nat_locks[i] irq_context: 0 nfnl_subsys_ctnetlink nf_conntrack_mutex &obj_hash[i].lock irq_context: 0 nfnl_subsys_ctnetlink nf_conntrack_mutex pool_lock#2 irq_context: 0 nfnl_subsys_ctnetlink nf_conntrack_mutex &c->lock irq_context: 0 nfnl_subsys_ctnetlink nf_conntrack_mutex &n->list_lock irq_context: 0 nfnl_subsys_ctnetlink nf_conntrack_mutex &rq->__lock irq_context: 0 nfnl_subsys_ctnetlink nf_conntrack_mutex &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 nfnl_subsys_ctnetlink nf_conntrack_mutex &nf_conntrack_locks[i] &nf_conntrack_locks[i]/1 irq_context: 0 nfnl_subsys_ctnetlink nf_conntrack_mutex &nf_conntrack_locks[i]/1 irq_context: 0 nfnl_subsys_ctnetlink nf_conntrack_mutex rcu_read_lock rcu_node_0 irq_context: 0 nfnl_subsys_ctnetlink nf_conntrack_mutex rcu_read_lock &rq->__lock irq_context: 0 nfnl_subsys_ctnetlink nf_conntrack_mutex rcu_read_lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 &sb->s_type->i_mutex_key#10 sk_lock-AF_INET &n->list_lock irq_context: 0 &sb->s_type->i_mutex_key#10 sk_lock-AF_INET &n->list_lock &c->lock irq_context: 0 cb_lock genl_mutex rtnl_mutex stock_lock irq_context: 0 cb_lock genl_mutex rtnl_mutex stack_depot_init_mutex irq_context: 0 cb_lock genl_mutex rtnl_mutex krc.lock irq_context: 0 cb_lock genl_mutex rtnl_mutex &dir->lock#2 irq_context: 0 (wq_completion)wg-crypt-wg0#5 (work_completion)(&peer->transmit_packet_work) &peer->endpoint_lock rcu_read_lock_bh rcu_read_lock rcu_read_lock &____s->seqcount#7 irq_context: 0 (wq_completion)wg-crypt-wg0#5 (work_completion)(&peer->transmit_packet_work) &peer->endpoint_lock rcu_read_lock_bh rcu_read_lock &nf_nat_locks[i] irq_context: 0 (wq_completion)wg-crypt-wg0#5 (work_completion)(&peer->transmit_packet_work) &peer->endpoint_lock rcu_read_lock_bh rcu_read_lock &nf_conntrack_locks[i] irq_context: 0 (wq_completion)wg-crypt-wg0#5 (work_completion)(&peer->transmit_packet_work) &peer->endpoint_lock rcu_read_lock_bh rcu_read_lock &nf_conntrack_locks[i] &nf_conntrack_locks[i]/1 irq_context: 0 (wq_completion)wg-crypt-wg0#5 (work_completion)(&peer->transmit_packet_work) &peer->endpoint_lock rcu_read_lock_bh rcu_read_lock &nf_conntrack_locks[i] &nf_conntrack_locks[i]/1 batched_entropy_u8.lock irq_context: 0 &sb->s_type->i_mutex_key#10 sk_lock-AF_INET6 &obj_hash[i].lock pool_lock irq_context: 0 rcu_read_lock &vma->vm_lock->lock &rq->__lock &cfs_rq->removed.lock irq_context: 0 rtnl_mutex mfc_unres_lock irq_context: 0 rtnl_mutex mfc_unres_lock &obj_hash[i].lock irq_context: 0 rcu_read_lock rcu_read_lock rcu_read_lock rcu_read_lock &____s->seqcount irq_context: 0 sk_lock-AF_INET6 k-sk_lock-AF_INET6 &tcp_hashinfo.bhash[i].lock &tcp_hashinfo.bhash2[i].lock &c->lock irq_context: 0 kn->active#63 fs_reclaim irq_context: 0 kn->active#63 fs_reclaim mmu_notifier_invalidate_range_start irq_context: 0 kn->active#63 stock_lock irq_context: 0 kn->active#63 &kernfs_locks->open_file_mutex[count] irq_context: 0 kn->active#63 &kernfs_locks->open_file_mutex[count] fs_reclaim irq_context: 0 kn->active#63 &kernfs_locks->open_file_mutex[count] fs_reclaim mmu_notifier_invalidate_range_start irq_context: 0 kn->active#63 &c->lock irq_context: 0 (crypto_chain).rwsem &p->pi_lock &cfs_rq->removed.lock irq_context: 0 crypto_default_null_skcipher_lock irq_context: 0 crypto_default_null_skcipher_lock crypto_alg_sem irq_context: 0 crypto_default_null_skcipher_lock fs_reclaim irq_context: 0 crypto_default_null_skcipher_lock fs_reclaim mmu_notifier_invalidate_range_start irq_context: 0 crypto_default_null_skcipher_lock pool_lock#2 irq_context: 0 &sb->s_type->i_mutex_key#10 crypto_default_null_skcipher_lock irq_context: 0 &sb->s_type->i_mutex_key#10 crypto_default_null_skcipher_lock &obj_hash[i].lock irq_context: 0 &sb->s_type->i_mutex_key#10 crypto_default_null_skcipher_lock pool_lock#2 irq_context: 0 &type->i_mutex_dir_key#7 &sb->s_type->i_lock_key#31 &dentry->d_lock &wq#2 irq_context: 0 kn->active#64 &rq->__lock irq_context: 0 kn->active#64 &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 kn->active#64 fs_reclaim irq_context: 0 kn->active#64 fs_reclaim mmu_notifier_invalidate_range_start irq_context: 0 kn->active#64 fs_reclaim mmu_notifier_invalidate_range_start &rq->__lock irq_context: 0 kn->active#64 fs_reclaim mmu_notifier_invalidate_range_start &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 kn->active#64 stock_lock irq_context: 0 kn->active#64 &kernfs_locks->open_file_mutex[count] irq_context: 0 kn->active#64 &kernfs_locks->open_file_mutex[count] fs_reclaim irq_context: 0 kn->active#64 &kernfs_locks->open_file_mutex[count] fs_reclaim mmu_notifier_invalidate_range_start irq_context: 0 sb_writers#4 &type->i_mutex_dir_key#3/1 &fsnotify_mark_srcu &cfs_rq->removed.lock irq_context: 0 (wq_completion)wg-crypt-wg1#5 (work_completion)(&peer->transmit_packet_work) &peer->endpoint_lock rcu_read_lock_bh rcu_read_lock pool_lock#2 irq_context: 0 (wq_completion)wg-crypt-wg1#5 (work_completion)(&peer->transmit_packet_work) &peer->endpoint_lock rcu_read_lock_bh rcu_read_lock rcu_read_lock &____s->seqcount#7 irq_context: 0 (wq_completion)wg-crypt-wg1#5 (work_completion)(&peer->transmit_packet_work) &peer->endpoint_lock rcu_read_lock_bh rcu_read_lock &nf_nat_locks[i] irq_context: 0 (wq_completion)wg-crypt-wg1#5 (work_completion)(&peer->transmit_packet_work) &peer->endpoint_lock rcu_read_lock_bh rcu_read_lock &nf_conntrack_locks[i] irq_context: 0 (wq_completion)wg-crypt-wg1#5 (work_completion)(&peer->transmit_packet_work) &peer->endpoint_lock rcu_read_lock_bh rcu_read_lock &nf_conntrack_locks[i] &nf_conntrack_locks[i]/1 irq_context: 0 (wq_completion)wg-crypt-wg1#5 (work_completion)(&peer->transmit_packet_work) &peer->endpoint_lock rcu_read_lock_bh rcu_read_lock &nf_conntrack_locks[i] &nf_conntrack_locks[i]/1 batched_entropy_u8.lock irq_context: 0 (wq_completion)wg-crypt-wg1#5 (work_completion)(&peer->transmit_packet_work) &peer->endpoint_lock rcu_read_lock_bh rcu_read_lock &nf_conntrack_locks[i]/1 irq_context: 0 (wq_completion)bat_events (work_completion)(&(&bat_priv->bla.work)->work) rcu_read_lock quarantine_lock irq_context: 0 sock_diag_mutex sock_diag_table_mutex inet_diag_table_mutex irq_context: 0 rtnl_mutex &mm->mmap_lock ptlock_ptr(page)#2 irq_context: 0 sb_writers#4 &type->i_mutex_dir_key#3/1 jbd2_handle &ei->i_data_sem &ei->i_es_lock key#7 irq_context: 0 devnet_rename_sem irq_context: 0 (wq_completion)events_unbound (work_completion)(&map->work) purge_vmap_area_lock &obj_hash[i].lock irq_context: 0 (wq_completion)events_unbound (work_completion)(&map->work) purge_vmap_area_lock pool_lock#2 irq_context: 0 rtnl_mutex (inetaddr_chain).rwsem &dir->lock irq_context: 0 rtnl_mutex (inetaddr_chain).rwsem rcu_read_lock &pool->lock irq_context: 0 rtnl_mutex (inetaddr_chain).rwsem rcu_read_lock &pool->lock &obj_hash[i].lock irq_context: 0 rtnl_mutex (inetaddr_chain).rwsem rcu_read_lock &pool->lock pool_lock#2 irq_context: 0 rtnl_mutex (inetaddr_chain).rwsem rcu_read_lock &pool->lock &p->pi_lock irq_context: 0 rtnl_mutex (inetaddr_chain).rwsem rcu_read_lock &pool->lock &p->pi_lock &rq->__lock irq_context: 0 rtnl_mutex (inetaddr_chain).rwsem rcu_read_lock &pool->lock &p->pi_lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 &f->f_pos_lock &type->i_mutex_dir_key#3 sb_writers#4 jbd2_handle &cfs_rq->removed.lock irq_context: 0 sb_writers#4 sb_internal jbd2_handle rcu_read_lock &rq->__lock &cfs_rq->removed.lock irq_context: softirq drivers/net/wireguard/ratelimiter.c:20 rcu_read_lock &pool->lock &p->pi_lock &rq->__lock &cfs_rq->removed.lock irq_context: 0 &xt[i].mutex &mm->mmap_lock pool_lock#2 irq_context: 0 sock_diag_mutex sock_diag_table_mutex (kmod_concurrent_max).lock irq_context: 0 sock_diag_mutex sock_diag_table_mutex fs_reclaim irq_context: 0 sock_diag_mutex sock_diag_table_mutex fs_reclaim mmu_notifier_invalidate_range_start irq_context: 0 sock_diag_mutex sock_diag_table_mutex &c->lock irq_context: 0 sock_diag_mutex sock_diag_table_mutex pool_lock#2 irq_context: 0 sock_diag_mutex sock_diag_table_mutex &obj_hash[i].lock irq_context: 0 sock_diag_mutex sock_diag_table_mutex rcu_read_lock &pool->lock/1 irq_context: 0 sock_diag_mutex sock_diag_table_mutex rcu_read_lock &pool->lock/1 &obj_hash[i].lock irq_context: 0 sock_diag_mutex sock_diag_table_mutex rcu_read_lock &pool->lock/1 pool_lock#2 irq_context: 0 sock_diag_mutex sock_diag_table_mutex rcu_read_lock &pool->lock/1 &p->pi_lock irq_context: 0 sock_diag_mutex sock_diag_table_mutex rcu_read_lock &pool->lock/1 &p->pi_lock &rq->__lock irq_context: 0 sock_diag_mutex sock_diag_table_mutex rcu_read_lock &pool->lock/1 &p->pi_lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 sock_diag_mutex sock_diag_table_mutex &x->wait#17 irq_context: 0 sock_diag_mutex sock_diag_table_mutex running_helpers_waitq.lock irq_context: 0 sb_writers#4 &type->i_mutex_dir_key#3/1 rcu_read_lock pool_lock#2 irq_context: 0 (wq_completion)bat_events (work_completion)(&(&bat_priv->tt.work)->work) &obj_hash[i].lock pool_lock irq_context: 0 sb_writers#4 &type->i_mutex_dir_key#3/1 &fsnotify_mark_srcu &rq->__lock &cfs_rq->removed.lock irq_context: 0 &mm->mmap_lock sb_writers#4 mmu_notifier_invalidate_range_start &rq->__lock irq_context: 0 &mm->mmap_lock sb_writers#4 mmu_notifier_invalidate_range_start &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 lock btf_idr_lock &c->lock irq_context: 0 lock btf_idr_lock &n->list_lock irq_context: 0 lock btf_idr_lock &n->list_lock &c->lock irq_context: softirq (&n->timer) k-slock-AF_INET6 rcu_read_lock rcu_read_lock rcu_read_lock_bh &sch->q.lock irq_context: softirq (&n->timer) k-slock-AF_INET6 rcu_read_lock rcu_read_lock rcu_read_lock_bh &sch->q.lock pool_lock#2 irq_context: softirq (&n->timer) k-slock-AF_INET6 rcu_read_lock rcu_read_lock rcu_read_lock_bh &sch->q.lock tk_core.seq.seqcount irq_context: softirq (&n->timer) k-slock-AF_INET6 rcu_read_lock rcu_read_lock rcu_read_lock_bh &sch->q.lock hrtimer_bases.lock irq_context: softirq (&n->timer) k-slock-AF_INET6 rcu_read_lock rcu_read_lock rcu_read_lock_bh &sch->q.lock hrtimer_bases.lock &obj_hash[i].lock irq_context: softirq (&n->timer) k-slock-AF_INET6 rcu_read_lock rcu_read_lock rcu_read_lock_bh &sch->q.lock hrtimer_bases.lock tk_core.seq.seqcount irq_context: softirq rcu_read_lock &list->lock#5 irq_context: softirq rcu_read_lock &sch->q.lock hrtimer_bases.lock irq_context: softirq rcu_read_lock &sch->q.lock hrtimer_bases.lock &obj_hash[i].lock irq_context: softirq rcu_read_lock &sch->q.lock hrtimer_bases.lock tk_core.seq.seqcount irq_context: 0 &sb->s_type->i_mutex_key#10 sk_lock-AF_TIPC slock-AF_TIPC rcu_read_lock &ei->socket.wq.wait irq_context: 0 &sb->s_type->i_mutex_key#10 sk_lock-AF_TIPC slock-AF_TIPC rcu_read_lock &ei->socket.wq.wait &p->pi_lock irq_context: 0 &sb->s_type->i_mutex_key#10 sk_lock-AF_TIPC slock-AF_TIPC rcu_read_lock &ei->socket.wq.wait &p->pi_lock &rq->__lock irq_context: 0 &sb->s_type->i_mutex_key#10 sk_lock-AF_TIPC slock-AF_TIPC rcu_read_lock &ei->socket.wq.wait &p->pi_lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 tasklist_lock &sighand->siglock batched_entropy_u8.lock crngs.lock irq_context: 0 sb_writers#4 &type->i_mutex_dir_key#3/1 tomoyo_ss remove_cache_srcu &cfs_rq->removed.lock irq_context: 0 &mm->mmap_lock &mm->page_table_lock &obj_hash[i].lock irq_context: 0 &mm->mmap_lock &mm->page_table_lock pool_lock#2 irq_context: 0 &mm->mmap_lock &mm->page_table_lock stock_lock irq_context: 0 &mm->mmap_lock &mm->page_table_lock rcu_read_lock &____s->seqcount irq_context: 0 &mm->mmap_lock &mm->page_table_lock rcu_read_lock pool_lock#2 irq_context: 0 (wq_completion)wg-kex-wg2#9 (work_completion)(&peer->transmit_handshake_work) &peer->endpoint_lock rcu_read_lock_bh rcu_read_lock &tbl->lock irq_context: 0 (wq_completion)wg-kex-wg2#9 (work_completion)(&peer->transmit_handshake_work) &peer->endpoint_lock rcu_read_lock_bh rcu_read_lock pool_lock#2 irq_context: 0 (wq_completion)wg-kex-wg2#9 (work_completion)(&peer->transmit_handshake_work) &peer->endpoint_lock rcu_read_lock_bh rcu_read_lock krc.lock irq_context: 0 &f->f_pos_lock sb_writers#11 &of->mutex kn->active#60 &rq->__lock irq_context: 0 &f->f_pos_lock sb_writers#11 &of->mutex kn->active#60 &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 rcu_read_lock rcu_read_lock_bh rcu_read_lock batched_entropy_u8.lock irq_context: 0 rcu_read_lock rcu_read_lock_bh rcu_read_lock kfence_freelist_lock irq_context: 0 &xs->map_list_lock irq_context: 0 clock-AF_XDP irq_context: 0 &u->iolock &u->lock rlock-AF_UNIX irq_context: 0 sk_lock-AF_KCM &____s->seqcount#2 irq_context: softirq (&n->timer) k-slock-AF_INET6 rcu_read_lock &____s->seqcount#2 irq_context: softirq (&n->timer) k-slock-AF_INET6 rcu_read_lock &____s->seqcount irq_context: 0 kn->active#60 rcu_read_lock rcu_node_0 irq_context: 0 kn->active#60 rcu_read_lock &rq->__lock irq_context: 0 kn->active#60 rcu_read_lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)events_unbound (work_completion)(&map->work) &cfs_rq->removed.lock irq_context: 0 (wq_completion)events_unbound (work_completion)(&map->work) rcu_read_lock rcu_node_0 irq_context: 0 (wq_completion)events_unbound (work_completion)(&map->work) rcu_read_lock &rq->__lock irq_context: 0 (wq_completion)events_unbound (work_completion)(&map->work) percpu_counters_lock irq_context: 0 sk_lock-AF_VSOCK clock-AF_VSOCK irq_context: 0 sk_lock-AF_VSOCK rlock-AF_VSOCK irq_context: 0 sb_writers#4 &type->i_mutex_dir_key#3 tomoyo_ss remove_cache_srcu &rq->__lock irq_context: 0 sb_writers#4 &type->i_mutex_dir_key#3 tomoyo_ss remove_cache_srcu &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 sk_lock-AF_INET (&tw->tw_timer) irq_context: 0 &sb->s_type->i_mutex_key#10 sk_lock-AF_INET6/1 &meta->lock irq_context: 0 &sb->s_type->i_mutex_key#10 sk_lock-AF_INET6/1 kfence_freelist_lock irq_context: softirq (&ndev->rs_timer) rcu_read_lock rcu_read_lock rcu_read_lock_bh rcu_read_lock rcu_read_lock &ei->socket.wq.wait irq_context: softirq (&ndev->rs_timer) rcu_read_lock rcu_read_lock rcu_read_lock_bh rcu_read_lock rcu_read_lock &ei->socket.wq.wait &p->pi_lock irq_context: softirq (&ndev->rs_timer) rcu_read_lock rcu_read_lock rcu_read_lock_bh rcu_read_lock rcu_read_lock &ei->socket.wq.wait &p->pi_lock &rq->__lock irq_context: softirq (&ndev->rs_timer) rcu_read_lock rcu_read_lock rcu_read_lock_bh rcu_read_lock rcu_read_lock &ei->socket.wq.wait &p->pi_lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 sb_writers#4 rcu_read_lock rcu_read_lock pgd_lock irq_context: 0 sb_writers#4 rcu_read_lock rcu_read_lock stock_lock irq_context: 0 sb_writers#4 rcu_read_lock rcu_read_lock key irq_context: 0 sb_writers#4 rcu_read_lock rcu_read_lock pcpu_lock irq_context: 0 sb_writers#4 rcu_read_lock rcu_read_lock percpu_counters_lock irq_context: 0 sb_writers#4 rcu_read_lock rcu_read_lock pcpu_lock stock_lock irq_context: 0 sb_writers#4 rcu_read_lock rcu_read_lock pool_lock irq_context: 0 sb_writers#4 rcu_read_lock rcu_read_lock pool_lock#2 irq_context: softirq (&n->timer) k-slock-AF_INET6 rcu_read_lock rcu_read_lock rcu_read_lock_bh &sch->q.lock &c->lock irq_context: softirq (&n->timer) k-slock-AF_INET6 rcu_read_lock rcu_read_lock rcu_read_lock_bh &sch->q.lock &n->list_lock irq_context: softirq (&n->timer) k-slock-AF_INET6 rcu_read_lock rcu_read_lock rcu_read_lock_bh &sch->q.lock &n->list_lock &c->lock irq_context: 0 &sb->s_type->i_mutex_key#10 sk_lock-AF_INET &h->lhash2[i].lock irq_context: 0 &sb->s_type->i_mutex_key#10 sk_lock-AF_INET &queue->rskq_lock irq_context: 0 (wq_completion)bat_events (work_completion)(&(&forw_packet_aggr->delayed_work)->work) rcu_read_lock_bh rcu_read_lock rcu_read_lock &ei->socket.wq.wait irq_context: 0 (wq_completion)bat_events (work_completion)(&(&forw_packet_aggr->delayed_work)->work) rcu_read_lock_bh rcu_read_lock rcu_read_lock &ei->socket.wq.wait &p->pi_lock irq_context: 0 (wq_completion)bat_events (work_completion)(&(&forw_packet_aggr->delayed_work)->work) rcu_read_lock_bh rcu_read_lock rcu_read_lock &ei->socket.wq.wait &p->pi_lock &rq->__lock irq_context: 0 (wq_completion)bat_events (work_completion)(&(&forw_packet_aggr->delayed_work)->work) rcu_read_lock_bh rcu_read_lock rcu_read_lock &ei->socket.wq.wait &p->pi_lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)wg-crypt-wg0#2 &rq->__lock irq_context: 0 &fp->aux->used_maps_mutex &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 sk_lock-AF_INET k-sk_lock-AF_INET &rq->__lock &cfs_rq->removed.lock irq_context: 0 sk_lock-AF_INET k-sk_lock-AF_INET slock-AF_INET rcu_read_lock &ei->socket.wq.wait irq_context: 0 sk_lock-AF_INET k-sk_lock-AF_INET slock-AF_INET rcu_read_lock &ei->socket.wq.wait &p->pi_lock irq_context: 0 sk_lock-AF_INET k-sk_lock-AF_INET slock-AF_INET rcu_read_lock &ei->socket.wq.wait &p->pi_lock &cfs_rq->removed.lock irq_context: 0 sk_lock-AF_INET k-sk_lock-AF_INET slock-AF_INET rcu_read_lock &ei->socket.wq.wait &p->pi_lock &rq->__lock irq_context: 0 sk_lock-AF_INET k-sk_lock-AF_INET slock-AF_INET rcu_read_lock &ei->socket.wq.wait &p->pi_lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 &sb->s_type->i_mutex_key#10 sk_lock-AF_INET k-sk_lock-AF_INET/1 fs_reclaim irq_context: 0 &sb->s_type->i_mutex_key#10 sk_lock-AF_INET k-sk_lock-AF_INET/1 fs_reclaim mmu_notifier_invalidate_range_start irq_context: 0 &sb->s_type->i_mutex_key#10 sk_lock-AF_INET k-sk_lock-AF_INET/1 fs_reclaim mmu_notifier_invalidate_range_start &rq->__lock irq_context: 0 &sb->s_type->i_mutex_key#10 sk_lock-AF_INET k-sk_lock-AF_INET/1 fs_reclaim mmu_notifier_invalidate_range_start &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 &sb->s_type->i_mutex_key#10 sk_lock-AF_INET k-sk_lock-AF_INET/1 &c->lock irq_context: 0 &sb->s_type->i_mutex_key#10 sk_lock-AF_INET k-sk_lock-AF_INET/1 pool_lock#2 irq_context: 0 &sb->s_type->i_mutex_key#10 sk_lock-AF_INET k-sk_lock-AF_INET/1 tk_core.seq.seqcount irq_context: 0 &sb->s_type->i_mutex_key#10 sk_lock-AF_INET k-sk_lock-AF_INET/1 rcu_read_lock rcu_read_lock &____s->seqcount#7 irq_context: 0 &sb->s_type->i_mutex_key#10 sk_lock-AF_INET k-sk_lock-AF_INET/1 rcu_read_lock rcu_read_lock &ct->lock irq_context: 0 &sb->s_type->i_mutex_key#10 sk_lock-AF_INET k-sk_lock-AF_INET/1 rcu_read_lock rcu_read_lock &____s->seqcount#10 irq_context: 0 &sb->s_type->i_mutex_key#10 sk_lock-AF_INET k-sk_lock-AF_INET/1 rcu_read_lock rcu_read_lock rcu_read_lock_bh &list->lock#5 irq_context: 0 &sb->s_type->i_mutex_key#10 sk_lock-AF_INET k-sk_lock-AF_INET/1 &obj_hash[i].lock irq_context: 0 &sb->s_type->i_mutex_key#10 sk_lock-AF_INET k-sk_lock-AF_INET/1 &base->lock irq_context: 0 &sb->s_type->i_mutex_key#10 sk_lock-AF_INET k-sk_lock-AF_INET/1 &base->lock &obj_hash[i].lock irq_context: 0 &sb->s_type->i_mutex_key#10 sk_lock-AF_INET k-sk_lock-AF_INET/1 slock-AF_INET irq_context: 0 nfnl_subsys_none irq_context: 0 rtnl_mutex lock kernfs_idr_lock batched_entropy_u8.lock irq_context: 0 rtnl_mutex lock kernfs_idr_lock kfence_freelist_lock irq_context: 0 &sb->s_type->i_mutex_key#10 sk_lock-AF_CAN &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 sk_lock-AF_ALG (console_sem).lock irq_context: 0 sk_lock-AF_ALG console_lock console_srcu console_owner_lock irq_context: 0 sk_lock-AF_ALG console_lock console_srcu console_owner irq_context: 0 sk_lock-AF_ALG console_lock console_srcu console_owner &port_lock_key irq_context: 0 sk_lock-AF_ALG console_lock console_srcu console_owner console_owner_lock irq_context: 0 sk_lock-AF_ALG &mm->mmap_lock mmu_notifier_invalidate_range_start irq_context: 0 sk_lock-AF_ALG &mm->mmap_lock ptlock_ptr(page)#2 lock#4 irq_context: 0 sk_lock-AF_ALG &mm->mmap_lock ptlock_ptr(page)#2 lock#4 &lruvec->lru_lock irq_context: 0 sk_lock-AF_ALG &mm->mmap_lock mmu_notifier_invalidate_range_start &rq->__lock irq_context: 0 sk_lock-AF_ALG &mm->mmap_lock mmu_notifier_invalidate_range_start &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 sk_lock-AF_BLUETOOTH-BTPROTO_HCI hci_dev_list_lock irq_context: 0 (wq_completion)bond3#2 irq_context: 0 rtnl_mutex &dev_addr_list_lock_key/1 &batadv_netdev_addr_lock_key irq_context: softirq &(&bond->mcast_work)->timer irq_context: softirq &(&bond->mcast_work)->timer rcu_read_lock &pool->lock/1 irq_context: softirq &(&bond->mcast_work)->timer rcu_read_lock &pool->lock/1 &obj_hash[i].lock irq_context: softirq &(&bond->mcast_work)->timer rcu_read_lock &pool->lock/1 &p->pi_lock irq_context: 0 (wq_completion)bond3#2 (work_completion)(&(&slave->notify_work)->work) irq_context: 0 (wq_completion)bond3#2 (work_completion)(&(&slave->notify_work)->work) &obj_hash[i].lock irq_context: 0 (wq_completion)bond3#2 (work_completion)(&(&slave->notify_work)->work) &base->lock irq_context: 0 (wq_completion)bond3#2 (work_completion)(&(&slave->notify_work)->work) &base->lock &obj_hash[i].lock irq_context: 0 sb_writers#4 &lruvec->lru_lock per_cpu_ptr(&cgroup_rstat_cpu_lock, cpu) irq_context: 0 rtnl_mutex &bond->mode_lock &obj_hash[i].lock irq_context: 0 rtnl_mutex &bond->mode_lock rcu_read_lock pool_lock#2 irq_context: 0 rtnl_mutex rcu_read_lock &dev_addr_list_lock_key/1 irq_context: 0 rtnl_mutex rcu_read_lock &dev_addr_list_lock_key/1 &batadv_netdev_addr_lock_key irq_context: 0 sb_writers#4 sb_writers#4 jbd2_handle rcu_read_lock rcu_node_0 irq_context: 0 sb_writers#4 sb_writers#4 jbd2_handle rcu_read_lock &rq->__lock irq_context: 0 sb_writers#4 sb_writers#4 jbd2_handle rcu_read_lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 &x->wait#8 &p->pi_lock irq_context: 0 sb_writers#4 sb_internal &rcu_state.expedited_wq irq_context: 0 rtnl_mutex &bond->mode_lock rcu_read_lock_bh rcu_read_lock &c->lock irq_context: 0 rtnl_mutex &bond->mode_lock rcu_read_lock_bh rcu_read_lock pool_lock#2 irq_context: 0 rtnl_mutex &bond->mode_lock rcu_read_lock_bh rcu_read_lock tk_core.seq.seqcount irq_context: 0 rtnl_mutex &bond->mode_lock rcu_read_lock_bh rcu_read_lock rlock-AF_PACKET irq_context: 0 sb_writers#4 &sb->s_type->i_mutex_key#8 rcu_read_lock batched_entropy_u8.lock irq_context: 0 sb_writers#4 &sb->s_type->i_mutex_key#8 rcu_read_lock kfence_freelist_lock irq_context: 0 sb_writers#4 &sb->s_type->i_mutex_key#8 jbd2_handle &sbi->s_orphan_lock &mapping->private_lock irq_context: 0 &x->wait#8 &p->pi_lock &rq->__lock irq_context: 0 &x->wait#8 &p->pi_lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)bond3#2 (work_completion)(&(&slave->notify_work)->work) rtnl_mutex net_rwsem irq_context: 0 (wq_completion)bond3#2 (work_completion)(&(&slave->notify_work)->work) rtnl_mutex net_rwsem &list->lock#2 irq_context: 0 (wq_completion)bond3#2 (work_completion)(&(&slave->notify_work)->work) rtnl_mutex &tn->lock irq_context: 0 (wq_completion)bond7 &rq->__lock irq_context: 0 (wq_completion)bond7 &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 sb_writers#4 &type->i_mutex_dir_key#3/1 rcu_read_lock rcu_read_lock &cfs_rq->removed.lock irq_context: 0 &f->f_pos_lock sb_writers#4 &sb->s_type->i_mutex_key#8 &mm->mmap_lock &obj_hash[i].lock pool_lock irq_context: softirq (&icsk->icsk_retransmit_timer) slock-AF_INET6 &dccp_hashinfo.bhash[i].lock irq_context: softirq (&icsk->icsk_retransmit_timer) slock-AF_INET6 &dccp_hashinfo.bhash[i].lock stock_lock irq_context: softirq (&icsk->icsk_retransmit_timer) slock-AF_INET6 &dccp_hashinfo.bhash[i].lock &obj_hash[i].lock irq_context: softirq (&icsk->icsk_retransmit_timer) slock-AF_INET6 &dccp_hashinfo.bhash[i].lock pool_lock#2 irq_context: softirq (&icsk->icsk_retransmit_timer) slock-AF_INET6 &dccp_hashinfo.bhash[i].lock &dccp_hashinfo.bhash2[i].lock irq_context: softirq (&icsk->icsk_retransmit_timer) slock-AF_INET6 &dccp_hashinfo.bhash[i].lock &dccp_hashinfo.bhash2[i].lock stock_lock irq_context: softirq (&icsk->icsk_retransmit_timer) slock-AF_INET6 &dccp_hashinfo.bhash[i].lock &dccp_hashinfo.bhash2[i].lock &obj_hash[i].lock irq_context: softirq (&icsk->icsk_retransmit_timer) slock-AF_INET6 &dccp_hashinfo.bhash[i].lock &dccp_hashinfo.bhash2[i].lock pool_lock#2 irq_context: 0 cb_lock nlk_cb_mutex-GENERIC rtnl_mutex rtnl_mutex.wait_lock irq_context: 0 cb_lock nlk_cb_mutex-GENERIC rtnl_mutex &rq->__lock irq_context: 0 cb_lock nlk_cb_mutex-GENERIC rtnl_mutex &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 nlk_cb_mutex-GENERIC &lock->wait_lock irq_context: 0 rtnl_mutex &br->lock lweventlist_lock &c->lock irq_context: 0 rtnl_mutex &ht->mutex &rq->__lock irq_context: 0 rtnl_mutex &ht->mutex &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 cb_lock nlk_cb_mutex-GENERIC rtnl_mutex.wait_lock irq_context: 0 nlk_cb_mutex-GENERIC rtnl_mutex irq_context: 0 nlk_cb_mutex-GENERIC rtnl_mutex &rq->__lock irq_context: 0 nlk_cb_mutex-GENERIC rtnl_mutex &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 nlk_cb_mutex-GENERIC rtnl_mutex.wait_lock irq_context: 0 nlk_cb_mutex-GENERIC &p->pi_lock irq_context: 0 nlk_cb_mutex-GENERIC rtnl_mutex rtnl_mutex.wait_lock irq_context: 0 rtnl_mutex &dev_addr_list_lock_key/1 rcu_read_lock &batadv_netdev_addr_lock_key irq_context: 0 nlk_cb_mutex-GENERIC &p->pi_lock &rq->__lock irq_context: 0 nlk_cb_mutex-GENERIC &p->pi_lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)events (work_completion)(&nlk->work) &____s->seqcount irq_context: 0 (wq_completion)events (work_completion)(&nlk->work) quarantine_lock irq_context: 0 tomoyo_ss console_owner_lock irq_context: 0 (wq_completion)events (linkwatch_work).work rtnl_mutex &tb->tb6_lock rcu_read_lock pool_lock#2 irq_context: 0 (wq_completion)events (linkwatch_work).work rtnl_mutex &tb->tb6_lock rcu_read_lock &data->fib_event_queue_lock irq_context: 0 (wq_completion)events (linkwatch_work).work rtnl_mutex &tb->tb6_lock rcu_read_lock rcu_read_lock &pool->lock irq_context: 0 (wq_completion)events (linkwatch_work).work rtnl_mutex &tb->tb6_lock rcu_read_lock rcu_read_lock &pool->lock &obj_hash[i].lock irq_context: 0 (wq_completion)events (linkwatch_work).work rtnl_mutex &tb->tb6_lock rcu_read_lock rcu_read_lock &pool->lock pool_lock#2 irq_context: 0 (wq_completion)events (linkwatch_work).work rtnl_mutex rcu_read_lock &bond->stats_lock/1 irq_context: 0 (wq_completion)ipv6_addrconf (work_completion)(&(&ifa->dad_work)->work) rtnl_mutex &idev->mc_lock &dev_addr_list_lock_key/1 rcu_read_lock &batadv_netdev_addr_lock_key irq_context: 0 tomoyo_ss console_owner irq_context: 0 &f->f_pos_lock &rcu_state.expedited_wq irq_context: 0 &f->f_pos_lock &rcu_state.expedited_wq &p->pi_lock irq_context: 0 &f->f_pos_lock &rcu_state.expedited_wq &p->pi_lock &rq->__lock irq_context: 0 &f->f_pos_lock &rcu_state.expedited_wq &p->pi_lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 &sb->s_type->i_mutex_key#9 per_cpu_ptr(&cgroup_rstat_cpu_lock, cpu) irq_context: 0 &sb->s_type->i_mutex_key#10 sk_lock-AF_ROSE &rq->__lock irq_context: 0 &sb->s_type->i_mutex_key#10 sk_lock-AF_ROSE &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 br_ioctl_mutex rtnl_mutex uevent_sock_mutex &____s->seqcount#2 irq_context: 0 br_ioctl_mutex rtnl_mutex uevent_sock_mutex &____s->seqcount irq_context: 0 br_ioctl_mutex rtnl_mutex lock kernfs_idr_lock &n->list_lock irq_context: 0 br_ioctl_mutex rtnl_mutex lock kernfs_idr_lock &n->list_lock &c->lock irq_context: 0 cb_lock genl_mutex rtnl_mutex &wg->device_update_lock &peer->endpoint_lock &obj_hash[i].lock irq_context: 0 rtnl_mutex &dev_addr_list_lock_key/1 &batadv_netdev_addr_lock_key &obj_hash[i].lock irq_context: 0 rtnl_mutex &dev_addr_list_lock_key/1 &batadv_netdev_addr_lock_key pool_lock#2 irq_context: 0 rtnl_mutex &dev_addr_list_lock_key/1 &batadv_netdev_addr_lock_key krc.lock irq_context: 0 rtnl_mutex &ul->lock#2 &n->list_lock irq_context: 0 rtnl_mutex &ul->lock#2 &n->list_lock &c->lock irq_context: 0 rtnl_mutex &bond->mode_lock rcu_read_lock_bh rcu_read_lock &____s->seqcount#2 irq_context: 0 rtnl_mutex &bond->mode_lock rcu_read_lock_bh rcu_read_lock &____s->seqcount irq_context: 0 rtnl_mutex &bond->mode_lock rcu_read_lock_bh rcu_read_lock &obj_hash[i].lock irq_context: 0 rtnl_mutex &bond->mode_lock rcu_read_lock lweventlist_lock irq_context: 0 rtnl_mutex &bond->mode_lock rcu_read_lock lweventlist_lock &c->lock irq_context: 0 rtnl_mutex &bond->mode_lock rcu_read_lock lweventlist_lock pool_lock#2 irq_context: 0 rtnl_mutex &bond->mode_lock rcu_read_lock lweventlist_lock &dir->lock#2 irq_context: 0 rtnl_mutex &bond->mode_lock rcu_read_lock &obj_hash[i].lock irq_context: 0 rtnl_mutex &bond->mode_lock rcu_read_lock &base->lock irq_context: 0 rtnl_mutex &bond->mode_lock rcu_read_lock &base->lock &obj_hash[i].lock irq_context: 0 rtnl_mutex &bond->mode_lock &____s->seqcount#2 irq_context: 0 rtnl_mutex &bond->mode_lock &____s->seqcount irq_context: 0 rtnl_mutex &bond->mode_lock (console_sem).lock irq_context: 0 rtnl_mutex &bond->mode_lock console_lock console_srcu console_owner_lock irq_context: 0 rtnl_mutex &bond->mode_lock console_lock console_srcu console_owner irq_context: 0 rtnl_mutex &bond->mode_lock console_lock console_srcu console_owner &port_lock_key irq_context: 0 rtnl_mutex &bond->mode_lock console_lock console_srcu console_owner console_owner_lock irq_context: 0 rtnl_mutex &bond->mode_lock rcu_read_lock (console_sem).lock irq_context: 0 rtnl_mutex &bond->mode_lock rcu_read_lock console_lock console_srcu console_owner_lock irq_context: 0 rtnl_mutex &bond->mode_lock rcu_read_lock console_lock console_srcu console_owner irq_context: 0 rtnl_mutex &bond->mode_lock rcu_read_lock console_lock console_srcu console_owner &port_lock_key irq_context: 0 rtnl_mutex &bond->mode_lock rcu_read_lock console_lock console_srcu console_owner console_owner_lock irq_context: softirq (&n->timer) rcu_read_lock pool_lock#2 irq_context: 0 cb_lock genl_mutex rtnl_mutex &wg->device_update_lock &peer->endpoint_lock pool_lock#2 irq_context: 0 sb_writers#4 &type->i_mutex_dir_key#3/1 tomoyo_ss &obj_hash[i].lock pool_lock irq_context: 0 sb_writers#4 &type->i_mutex_dir_key#3/1 sb_internal jbd2_handle &ei->i_data_sem mmu_notifier_invalidate_range_start &cfs_rq->removed.lock irq_context: 0 sk_lock-AF_ROSE fs_reclaim irq_context: 0 sk_lock-AF_ROSE fs_reclaim mmu_notifier_invalidate_range_start irq_context: 0 sk_lock-AF_ROSE stock_lock irq_context: 0 sk_lock-AF_ROSE &f->f_lock irq_context: 0 sk_lock-AF_ROSE &f->f_lock fasync_lock irq_context: 0 sk_lock-AF_ROSE &f->f_lock fasync_lock &new->fa_lock irq_context: 0 sk_lock-AF_ROSE &f->f_lock fasync_lock &obj_hash[i].lock irq_context: 0 &tfile->napi_mutex rcu_read_lock tk_core.seq.seqcount irq_context: 0 &tfile->napi_mutex rcu_read_lock rcu_read_lock &obj_hash[i].lock irq_context: 0 &tfile->napi_mutex rcu_read_lock rcu_read_lock pool_lock#2 irq_context: 0 rtnl_mutex team->team_lock_key#7 irq_context: 0 (wq_completion)mld (work_completion)(&(&idev->mc_ifc_work)->work) &idev->mc_lock rcu_read_lock rcu_read_lock rcu_read_lock_bh dev->qdisc_tx_busylock ?: &qdisc_tx_busylock rcu_read_lock pcpu_lock irq_context: 0 sb_writers#4 &sb->s_type->i_mutex_key#8 &rcu_state.expedited_wq &p->pi_lock &rq->__lock irq_context: 0 sb_writers#4 &sb->s_type->i_mutex_key#8 &rcu_state.expedited_wq &p->pi_lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)ipv6_addrconf (work_completion)(&(&ifa->dad_work)->work) rtnl_mutex rcu_read_lock rcu_read_lock rcu_read_lock_bh dev->qdisc_tx_busylock ?: &qdisc_tx_busylock rcu_read_lock &obj_hash[i].lock irq_context: 0 (wq_completion)ipv6_addrconf (work_completion)(&(&ifa->dad_work)->work) rtnl_mutex rcu_read_lock rcu_read_lock rcu_read_lock_bh dev->qdisc_tx_busylock ?: &qdisc_tx_busylock rcu_read_lock pool_lock#2 irq_context: 0 rtnl_mutex &idev->mc_lock &dev_addr_list_lock_key#2 &n->list_lock irq_context: 0 rtnl_mutex &idev->mc_lock &dev_addr_list_lock_key#2 &n->list_lock &c->lock irq_context: 0 sb_writers#4 &type->i_mutex_dir_key#3/1 jbd2_handle rcu_read_lock &rq->__lock &cfs_rq->removed.lock irq_context: 0 rtnl_mutex team->team_lock_key#8 irq_context: 0 (wq_completion)events (linkwatch_work).work rtnl_mutex &dev->tx_global_lock _xmit_NETROM irq_context: 0 (wq_completion)events (linkwatch_work).work rtnl_mutex dev->qdisc_tx_busylock ?: &qdisc_tx_busylock irq_context: 0 (wq_completion)events (linkwatch_work).work rtnl_mutex dev->qdisc_tx_busylock ?: &qdisc_tx_busylock &sch->q.lock irq_context: 0 (wq_completion)events (linkwatch_work).work rtnl_mutex &tbl->lock &n->list_lock irq_context: 0 (wq_completion)events (linkwatch_work).work rtnl_mutex &tbl->lock &n->list_lock &c->lock irq_context: 0 rtnl_mutex team->team_lock_key#9 irq_context: 0 (wq_completion)mld (work_completion)(&(&idev->mc_dad_work)->work) &idev->mc_lock rcu_read_lock rcu_read_lock rcu_read_lock_bh noop_qdisc.q.lock irq_context: 0 sb_writers#4 &sb->s_type->i_mutex_key#8 &sbi->s_writepages_rwsem &____s->seqcount#2 irq_context: 0 sb_writers#4 &sb->s_type->i_mutex_key#8 &sbi->s_writepages_rwsem &____s->seqcount irq_context: 0 br_ioctl_mutex rtnl_mutex uevent_sock_mutex rlock-AF_NETLINK irq_context: 0 br_ioctl_mutex rtnl_mutex uevent_sock_mutex rcu_read_lock &ei->socket.wq.wait irq_context: 0 br_ioctl_mutex rtnl_mutex uevent_sock_mutex rcu_read_lock &ei->socket.wq.wait &ep->lock irq_context: 0 br_ioctl_mutex rtnl_mutex uevent_sock_mutex rcu_read_lock &ei->socket.wq.wait &ep->lock &ep->wq irq_context: 0 br_ioctl_mutex rtnl_mutex uevent_sock_mutex rcu_read_lock &ei->socket.wq.wait &ep->lock &ep->wq &p->pi_lock irq_context: 0 br_ioctl_mutex rtnl_mutex uevent_sock_mutex rcu_read_lock &ei->socket.wq.wait &ep->lock &ep->wq &p->pi_lock &rq->__lock irq_context: 0 br_ioctl_mutex rtnl_mutex uevent_sock_mutex rcu_read_lock &ei->socket.wq.wait &ep->lock &ep->wq &p->pi_lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 br_ioctl_mutex rtnl_mutex uevent_sock_mutex rcu_read_lock rcu_node_0 irq_context: 0 br_ioctl_mutex rtnl_mutex uevent_sock_mutex rcu_read_lock &rq->__lock irq_context: 0 br_ioctl_mutex rtnl_mutex uevent_sock_mutex rcu_read_lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 br_ioctl_mutex rtnl_mutex &br->hash_lock rlock-AF_NETLINK irq_context: 0 br_ioctl_mutex rtnl_mutex &br->hash_lock rcu_read_lock &ei->socket.wq.wait irq_context: 0 br_ioctl_mutex rtnl_mutex &br->hash_lock rcu_read_lock &ei->socket.wq.wait &p->pi_lock irq_context: 0 br_ioctl_mutex rtnl_mutex &br->hash_lock rcu_read_lock &ei->socket.wq.wait &p->pi_lock &rq->__lock irq_context: 0 br_ioctl_mutex rtnl_mutex &br->hash_lock rcu_read_lock &ei->socket.wq.wait &p->pi_lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 br_ioctl_mutex rtnl_mutex nf_hook_mutex irq_context: 0 br_ioctl_mutex rtnl_mutex nf_hook_mutex &rq->__lock irq_context: 0 br_ioctl_mutex rtnl_mutex nf_hook_mutex fs_reclaim irq_context: 0 br_ioctl_mutex rtnl_mutex nf_hook_mutex fs_reclaim mmu_notifier_invalidate_range_start irq_context: 0 br_ioctl_mutex rtnl_mutex nf_hook_mutex stock_lock irq_context: 0 br_ioctl_mutex rtnl_mutex nf_hook_mutex pool_lock#2 irq_context: 0 br_ioctl_mutex rtnl_mutex nf_hook_mutex &c->lock irq_context: 0 br_ioctl_mutex rtnl_mutex rlock-AF_NETLINK irq_context: 0 br_ioctl_mutex rtnl_mutex rcu_read_lock &ei->socket.wq.wait irq_context: 0 br_ioctl_mutex rtnl_mutex rcu_read_lock &ei->socket.wq.wait &p->pi_lock irq_context: 0 br_ioctl_mutex rtnl_mutex rcu_read_lock &ei->socket.wq.wait &p->pi_lock &rq->__lock irq_context: 0 br_ioctl_mutex rtnl_mutex rcu_read_lock &ei->socket.wq.wait &p->pi_lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 kn->active#16 &kernfs_locks->open_file_mutex[count] &c->lock irq_context: 0 kn->active#65 fs_reclaim irq_context: 0 kn->active#65 fs_reclaim mmu_notifier_invalidate_range_start irq_context: 0 kn->active#65 &c->lock irq_context: 0 kn->active#65 &n->list_lock irq_context: 0 kn->active#65 &n->list_lock &c->lock irq_context: 0 kn->active#65 &kernfs_locks->open_file_mutex[count] irq_context: 0 kn->active#65 &kernfs_locks->open_file_mutex[count] fs_reclaim irq_context: 0 kn->active#65 &kernfs_locks->open_file_mutex[count] fs_reclaim mmu_notifier_invalidate_range_start irq_context: 0 sb_writers#4 &type->i_mutex_dir_key#3/1 &sb->s_type->i_mutex_key#8 jbd2_handle rcu_read_lock fill_pool_map-wait-type-override pool_lock irq_context: 0 br_ioctl_mutex rtnl_mutex &root->kernfs_rwsem rcu_read_lock rcu_node_0 irq_context: 0 br_ioctl_mutex rtnl_mutex &root->kernfs_rwsem rcu_read_lock &rq->__lock irq_context: 0 br_ioctl_mutex rtnl_mutex &root->kernfs_rwsem rcu_read_lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 br_ioctl_mutex rtnl_mutex sysctl_lock krc.lock &obj_hash[i].lock irq_context: 0 br_ioctl_mutex rtnl_mutex sysctl_lock krc.lock &base->lock irq_context: 0 br_ioctl_mutex rtnl_mutex sysctl_lock krc.lock &base->lock &obj_hash[i].lock irq_context: 0 br_ioctl_mutex rtnl_mutex pool_lock irq_context: 0 br_ioctl_mutex &obj_hash[i].lock pool_lock irq_context: 0 &sb->s_type->i_mutex_key#8 fill_pool_map-wait-type-override &n->list_lock irq_context: 0 &sb->s_type->i_mutex_key#8 fill_pool_map-wait-type-override &n->list_lock &c->lock irq_context: 0 sb_writers#4 &type->i_mutex_dir_key#3 jbd2_handle batched_entropy_u32.lock crngs.lock base_crng.lock irq_context: 0 sk_lock-AF_INET6 &mm->mmap_lock sb_pagefaults &c->lock irq_context: 0 sk_lock-AF_INET6 &mm->mmap_lock sb_pagefaults remove_cache_srcu irq_context: 0 sk_lock-AF_INET6 &mm->mmap_lock sb_pagefaults remove_cache_srcu quarantine_lock irq_context: 0 sk_lock-AF_INET6 &mm->mmap_lock sb_pagefaults remove_cache_srcu &c->lock irq_context: 0 sk_lock-AF_INET6 &mm->mmap_lock sb_pagefaults remove_cache_srcu &n->list_lock irq_context: 0 sk_lock-AF_INET6 &mm->mmap_lock sb_pagefaults remove_cache_srcu &obj_hash[i].lock irq_context: 0 rose_node_list_lock irq_context: 0 rose_node_list_lock rose_neigh_list_lock irq_context: 0 sb_writers#4 &sb->s_type->i_mutex_key#8 mapping.invalidate_lock jbd2_handle rcu_node_0 irq_context: 0 sk_lock-AF_INET6 rcu_read_lock &rcu_state.gp_wq &p->pi_lock &rq->__lock &cfs_rq->removed.lock irq_context: 0 (wq_completion)wg-crypt-wg0#2 (work_completion)(&peer->transmit_packet_work) rcu_read_lock_bh fill_pool_map-wait-type-override pool_lock#2 irq_context: 0 (wq_completion)wg-crypt-wg0#2 (work_completion)(&peer->transmit_packet_work) rcu_read_lock_bh fill_pool_map-wait-type-override &c->lock irq_context: 0 (wq_completion)wg-crypt-wg0#2 (work_completion)(&peer->transmit_packet_work) rcu_read_lock_bh fill_pool_map-wait-type-override &n->list_lock irq_context: 0 (wq_completion)wg-crypt-wg0#2 (work_completion)(&peer->transmit_packet_work) rcu_read_lock_bh fill_pool_map-wait-type-override &n->list_lock &c->lock irq_context: 0 (wq_completion)wg-crypt-wg0#2 (work_completion)(&peer->transmit_packet_work) rcu_read_lock_bh fill_pool_map-wait-type-override pool_lock irq_context: 0 sk_lock-AF_INET6 rcu_read_lock rcu_read_lock &pool->lock fill_pool_map-wait-type-override &c->lock irq_context: 0 &sb->s_type->i_mutex_key#8 fs_reclaim mmu_notifier_invalidate_range_start &rq->__lock irq_context: 0 &sb->s_type->i_mutex_key#8 fs_reclaim mmu_notifier_invalidate_range_start &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 &f->f_pos_lock sb_writers#4 &sb->s_type->i_mutex_key#8 &sbi->s_writepages_rwsem jbd2_handle &ei->i_data_sem &ei->i_prealloc_lock irq_context: 0 &f->f_pos_lock sb_writers#4 &sb->s_type->i_mutex_key#8 &sbi->s_writepages_rwsem jbd2_handle &ei->i_data_sem &ei->i_prealloc_lock &pa->pa_lock#2 irq_context: 0 &f->f_pos_lock sb_writers#4 &sb->s_type->i_mutex_key#8 &sbi->s_writepages_rwsem jbd2_handle &ei->i_data_sem key#3 irq_context: 0 &f->f_pos_lock sb_writers#4 &sb->s_type->i_mutex_key#8 &sbi->s_writepages_rwsem jbd2_handle &ei->i_data_sem &pa->pa_lock#2 irq_context: 0 &f->f_pos_lock sb_writers#4 &sb->s_type->i_mutex_key#8 &sbi->s_writepages_rwsem jbd2_handle &ei->i_data_sem &(ei->i_block_reservation_lock) key#14 irq_context: 0 &f->f_pos_lock sb_writers#4 &sb->s_type->i_mutex_key#8 &sbi->s_writepages_rwsem jbd2_handle rcu_read_lock rcu_node_0 irq_context: 0 &f->f_pos_lock sb_writers#4 &sb->s_type->i_mutex_key#8 &sbi->s_writepages_rwsem jbd2_handle rcu_read_lock &rq->__lock irq_context: 0 &f->f_pos_lock sb_writers#4 &sb->s_type->i_mutex_key#8 &sbi->s_writepages_rwsem jbd2_handle rcu_read_lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 &f->f_pos_lock sb_writers#4 &sb->s_type->i_mutex_key#8 &sbi->s_writepages_rwsem jbd2_handle rcu_read_lock key#10 irq_context: 0 &f->f_pos_lock sb_writers#4 &sb->s_type->i_mutex_key#8 &sbi->s_writepages_rwsem jbd2_handle rcu_read_lock &xa->xa_lock#9 key#10 irq_context: 0 &f->f_pos_lock sb_writers#4 &sb->s_type->i_mutex_key#8 &sbi->s_writepages_rwsem rcu_read_lock pool_lock#2 irq_context: 0 &f->f_pos_lock sb_writers#4 &sb->s_type->i_mutex_key#8 mapping.invalidate_lock jbd2_handle &ei->i_data_sem &mapping->private_lock irq_context: softirq &(&bat_priv->orig_work)->timer rcu_read_lock &pool->lock/1 &p->pi_lock &rq->__lock &obj_hash[i].lock irq_context: softirq &(&bat_priv->orig_work)->timer rcu_read_lock &pool->lock/1 &p->pi_lock &rq->__lock &base->lock irq_context: softirq &(&bat_priv->orig_work)->timer rcu_read_lock &pool->lock/1 &p->pi_lock &rq->__lock &base->lock &obj_hash[i].lock irq_context: 0 &f->f_pos_lock sb_writers#4 &sb->s_type->i_mutex_key#8 &sbi->s_writepages_rwsem &rq->__lock irq_context: 0 &f->f_pos_lock sb_writers#4 &sb->s_type->i_mutex_key#8 &sbi->s_writepages_rwsem jbd2_handle &rq->__lock irq_context: 0 &f->f_pos_lock sb_writers#4 &sb->s_type->i_mutex_key#8 &sbi->s_writepages_rwsem jbd2_handle &ei->i_data_sem &sbi->s_md_lock irq_context: 0 &f->f_pos_lock sb_writers#4 &sb->s_type->i_mutex_key#8 &sbi->s_writepages_rwsem jbd2_handle &ei->i_data_sem &bgl->locks[i].lock &sbi->s_md_lock irq_context: 0 &f->f_pos_lock sb_writers#4 &sb->s_type->i_mutex_key#8 &sbi->s_writepages_rwsem jbd2_handle tk_core.seq.seqcount irq_context: 0 &f->f_pos_lock sb_writers#4 &sb->s_type->i_mutex_key#8 &sbi->s_writepages_rwsem jbd2_handle &dd->lock irq_context: 0 &f->f_pos_lock sb_writers#4 &sb->s_type->i_mutex_key#8 &sbi->s_writepages_rwsem jbd2_handle rcu_read_lock &dd->lock irq_context: 0 &f->f_pos_lock sb_writers#4 &sb->s_type->i_mutex_key#8 &sbi->s_writepages_rwsem jbd2_handle rcu_read_lock &obj_hash[i].lock irq_context: 0 &f->f_pos_lock sb_writers#4 &sb->s_type->i_mutex_key#8 &sbi->s_writepages_rwsem jbd2_handle rcu_read_lock pool_lock#2 irq_context: 0 &f->f_pos_lock sb_writers#4 &sb->s_type->i_mutex_key#8 &sbi->s_writepages_rwsem jbd2_handle rcu_read_lock tk_core.seq.seqcount irq_context: 0 &f->f_pos_lock sb_writers#4 &sb->s_type->i_mutex_key#8 &sbi->s_writepages_rwsem jbd2_handle rcu_read_lock &virtscsi_vq->vq_lock irq_context: 0 &f->f_pos_lock sb_writers#4 &sb->s_type->i_mutex_key#8 &sbi->s_writepages_rwsem jbd2_handle rcu_read_lock &c->lock irq_context: 0 &f->f_pos_lock sb_writers#4 &sb->s_type->i_mutex_key#8 &sbi->s_writepages_rwsem jbd2_handle rcu_read_lock &n->list_lock irq_context: 0 &f->f_pos_lock sb_writers#4 &sb->s_type->i_mutex_key#8 &sbi->s_writepages_rwsem jbd2_handle rcu_read_lock &n->list_lock &c->lock irq_context: 0 &f->f_pos_lock sb_writers#4 &sb->s_type->i_mutex_key#8 &sbi->s_writepages_rwsem jbd2_handle &n->list_lock irq_context: 0 &f->f_pos_lock sb_writers#4 &sb->s_type->i_mutex_key#8 &sbi->s_writepages_rwsem jbd2_handle &n->list_lock &c->lock irq_context: 0 &f->f_pos_lock sb_writers#4 &sb->s_type->i_mutex_key#8 &sbi->s_writepages_rwsem jbd2_handle &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 &f->f_pos_lock sb_writers#4 &sb->s_type->i_mutex_key#8 &sbi->s_writepages_rwsem jbd2_handle batched_entropy_u8.lock irq_context: 0 &f->f_pos_lock sb_writers#4 &sb->s_type->i_mutex_key#8 &sbi->s_writepages_rwsem jbd2_handle kfence_freelist_lock irq_context: 0 &f->f_pos_lock sb_writers#4 &sb->s_type->i_mutex_key#8 &sb->s_type->i_lock_key#22 &xa->xa_lock#9 rcu_read_lock per_cpu_ptr(&cgroup_rstat_cpu_lock, cpu) irq_context: 0 &mm->mmap_lock &vma->vm_lock->lock &rq->__lock &cfs_rq->removed.lock irq_context: 0 (wq_completion)bond7 (work_completion)(&(&bond->alb_work)->work) &rq->__lock irq_context: 0 (wq_completion)bond7 (work_completion)(&(&bond->alb_work)->work) &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 &f->f_pos_lock sb_writers#4 &sb->s_type->i_mutex_key#8 &sbi->s_writepages_rwsem jbd2_handle &ei->i_data_sem &mapping->private_lock irq_context: 0 &f->f_pos_lock sb_writers#4 &sb->s_type->i_mutex_key#8 &sbi->s_writepages_rwsem jbd2_handle rcu_read_lock rcu_read_lock per_cpu_ptr(&cgroup_rstat_cpu_lock, cpu) irq_context: 0 &f->f_pos_lock sb_writers#4 &sb->s_type->i_mutex_key#8 &sbi->s_writepages_rwsem rcu_read_lock &c->lock irq_context: 0 &f->f_pos_lock sb_writers#4 &sb->s_type->i_mutex_key#8 &sbi->s_writepages_rwsem jbd2_handle &folio_wait_table[i] irq_context: 0 &f->f_pos_lock sb_writers#4 &sb->s_type->i_mutex_key#8 &sbi->s_writepages_rwsem jbd2_handle &obj_hash[i].lock irq_context: 0 &f->f_pos_lock sb_writers#4 &sb->s_type->i_mutex_key#8 &sbi->s_writepages_rwsem jbd2_handle rcu_read_lock &pool->lock irq_context: 0 &f->f_pos_lock sb_writers#4 &sb->s_type->i_mutex_key#8 &sbi->s_writepages_rwsem jbd2_handle rcu_read_lock &pool->lock &obj_hash[i].lock irq_context: 0 &f->f_pos_lock sb_writers#4 &sb->s_type->i_mutex_key#8 &sbi->s_writepages_rwsem jbd2_handle rcu_read_lock &pool->lock pool_lock#2 irq_context: 0 &f->f_pos_lock sb_writers#4 &sb->s_type->i_mutex_key#8 &sbi->s_writepages_rwsem jbd2_handle rcu_read_lock &pool->lock &p->pi_lock irq_context: 0 &f->f_pos_lock sb_writers#4 &sb->s_type->i_mutex_key#8 &sbi->s_writepages_rwsem jbd2_handle rcu_read_lock &pool->lock &p->pi_lock &rq->__lock irq_context: 0 &f->f_pos_lock sb_writers#4 &sb->s_type->i_mutex_key#8 &sbi->s_writepages_rwsem jbd2_handle rcu_read_lock &pool->lock &p->pi_lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 &f->f_pos_lock sb_writers#4 &sb->s_type->i_mutex_key#8 &sbi->s_writepages_rwsem jbd2_handle &journal->j_wait_reserved irq_context: 0 &f->f_pos_lock sb_writers#4 &sb->s_type->i_mutex_key#8 jbd2_handle &ei->i_es_lock key#2 irq_context: 0 &f->f_pos_lock sb_writers#4 &sb->s_type->i_mutex_key#8 mapping.invalidate_lock &journal->j_state_lock &journal->j_wait_transaction_locked irq_context: 0 &f->f_pos_lock sb_writers#4 &sb->s_type->i_mutex_key#8 mapping.invalidate_lock jbd2_handle irq_context: 0 &f->f_pos_lock sb_writers#4 &sb->s_type->i_mutex_key#8 mapping.invalidate_lock rcu_read_lock &rq->__lock irq_context: 0 &f->f_pos_lock sb_writers#4 &sb->s_type->i_mutex_key#8 mapping.invalidate_lock rcu_read_lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 &f->f_pos_lock sb_writers#4 &sb->s_type->i_mutex_key#8 mapping.invalidate_lock jbd2_handle &c->lock irq_context: 0 &f->f_pos_lock sb_writers#4 &sb->s_type->i_mutex_key#8 mapping.invalidate_lock jbd2_handle &____s->seqcount#2 irq_context: 0 &f->f_pos_lock sb_writers#4 &sb->s_type->i_mutex_key#8 mapping.invalidate_lock jbd2_handle &ei->i_data_sem &ret->b_state_lock irq_context: 0 &f->f_pos_lock sb_writers#4 &sb->s_type->i_mutex_key#8 mapping.invalidate_lock jbd2_handle &sbi->s_orphan_lock &ret->b_state_lock irq_context: 0 text_mutex.wait_lock irq_context: 0 (wq_completion)events_unbound (work_completion)(&(&kfence_timer)->work) cpu_hotplug_lock jump_label_mutex.wait_lock irq_context: 0 (wq_completion)events_unbound (work_completion)(&(&kfence_timer)->work) cpu_hotplug_lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)events_unbound (work_completion)(&(&kfence_timer)->work) cpu_hotplug_lock &p->pi_lock irq_context: 0 (wq_completion)events_unbound (work_completion)(&(&kfence_timer)->work) cpu_hotplug_lock &p->pi_lock &rq->__lock irq_context: 0 (wq_completion)events_unbound (work_completion)(&(&kfence_timer)->work) cpu_hotplug_lock &p->pi_lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 tracepoints_mutex cpu_hotplug_lock static_call_mutex text_mutex &cfs_rq->removed.lock irq_context: 0 tracepoints_mutex cpu_hotplug_lock static_call_mutex text_mutex &obj_hash[i].lock irq_context: 0 rtnl_mutex &tn->idrinfo->lock#2 irq_context: 0 rtnl_mutex &tn->idrinfo->lock#2 fs_reclaim irq_context: 0 rtnl_mutex &tn->idrinfo->lock#2 fs_reclaim mmu_notifier_invalidate_range_start irq_context: 0 rtnl_mutex &tn->idrinfo->lock#2 pool_lock#2 irq_context: 0 rtnl_mutex &p->tcfa_lock &c->lock irq_context: 0 rtnl_mutex &p->tcfa_lock &n->list_lock irq_context: 0 rtnl_mutex &p->tcfa_lock &n->list_lock &c->lock irq_context: 0 rtnl_mutex &p->tcfa_lock pool_lock#2 irq_context: 0 sk_lock-AF_INET6 sk_lock-AF_INET6/1 irq_context: 0 sk_lock-AF_INET6 sk_lock-AF_INET6/1 slock-AF_INET6 irq_context: 0 sk_lock-AF_INET6 &newf->file_lock irq_context: 0 sk_lock-AF_INET6 &sb->s_type->i_lock_key#8 &dentry->d_lock irq_context: 0 &sb->s_type->i_mutex_key#10 sk_lock-AF_INET6/1 &asoc->wait irq_context: 0 &sb->s_type->i_mutex_key#10 sk_lock-AF_INET6/1 &asoc->wait &p->pi_lock irq_context: 0 &sb->s_type->i_mutex_key#10 sk_lock-AF_INET6/1 &asoc->wait &p->pi_lock &cfs_rq->removed.lock irq_context: 0 &sb->s_type->i_mutex_key#10 sk_lock-AF_INET6/1 &asoc->wait &p->pi_lock &rq->__lock irq_context: 0 &sb->s_type->i_mutex_key#10 sk_lock-AF_INET6/1 &asoc->wait &p->pi_lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 sk_lock-AF_INET6 sctp_assocs_id_lock &obj_hash[i].lock irq_context: 0 sb_writers#4 rcu_read_lock &rcu_state.gp_wq &p->pi_lock &cfs_rq->removed.lock irq_context: 0 sb_writers#4 &sb->s_type->i_mutex_key#8 &sb->s_type->i_lock_key#22 &xa->xa_lock#9 &obj_hash[i].lock irq_context: 0 &f->f_pos_lock sb_writers#4 &sb->s_type->i_mutex_key#8 rcu_read_lock &rcu_state.gp_wq &p->pi_lock &cfs_rq->removed.lock irq_context: 0 &f->f_pos_lock sb_writers#4 &sb->s_type->i_mutex_key#8 rcu_read_lock &rcu_state.expedited_wq &p->pi_lock &rq->__lock &cfs_rq->removed.lock irq_context: 0 nfnl_subsys_nftables &nft_net->commit_mutex &cfs_rq->removed.lock irq_context: 0 nfnl_subsys_nftables &nft_net->commit_mutex &obj_hash[i].lock irq_context: 0 nfnl_subsys_nftables &nft_net->commit_mutex pool_lock#2 irq_context: 0 sb_writers#4 sb_internal jbd2_handle &sbi->s_orphan_lock bit_wait_table + i irq_context: 0 (wq_completion)wg-kex-wg0#7 (work_completion)(&peer->transmit_handshake_work) &peer->endpoint_lock rcu_read_lock_bh rcu_read_lock &n->lock &n->list_lock irq_context: 0 (wq_completion)wg-kex-wg0#7 (work_completion)(&peer->transmit_handshake_work) &peer->endpoint_lock rcu_read_lock_bh rcu_read_lock &n->lock &n->list_lock &c->lock irq_context: 0 &f->f_pos_lock sb_writers#11 fs_reclaim mmu_notifier_invalidate_range_start &rq->__lock irq_context: 0 &f->f_pos_lock sb_writers#11 fs_reclaim mmu_notifier_invalidate_range_start &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 nfnl_grp_active_lock irq_context: 0 &sb->s_type->i_mutex_key#10 pgd_lock irq_context: 0 &sb->s_type->i_mutex_key#10 key irq_context: 0 &sb->s_type->i_mutex_key#10 percpu_counters_lock irq_context: 0 &sb->s_type->i_mutex_key#10 nfnl_grp_active_lock irq_context: 0 &pipe->mutex/1 &mm->mmap_lock rcu_read_lock rcu_read_lock ptlock_ptr(page)#2 irq_context: 0 &pipe->mutex/1 &mm->mmap_lock rcu_read_lock rcu_read_lock rcu_node_0 irq_context: 0 &pipe->mutex/1 &mm->mmap_lock rcu_read_lock rcu_read_lock &rq->__lock irq_context: 0 &pipe->mutex/1 &mm->mmap_lock rcu_read_lock rcu_read_lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 rcu_read_lock &loc_l->lock irq_context: 0 rcu_read_lock &loc_l->lock &l->lock irq_context: 0 sk_lock-AF_ALG &mm->mmap_lock ptlock_ptr(page) irq_context: 0 sk_lock-AF_ALG &mm->mmap_lock rcu_read_lock &cfs_rq->removed.lock irq_context: 0 rcu_read_lock &l->lock irq_context: 0 (wq_completion)tipc_send#4 irq_context: 0 (wq_completion)tipc_send#4 (work_completion)(&con->swork) irq_context: 0 (wq_completion)tipc_send#4 (work_completion)(&con->swork) &con->outqueue_lock irq_context: 0 (wq_completion)tipc_send#4 (work_completion)(&con->swork) &list->lock#38 irq_context: 0 (wq_completion)tipc_send#4 (work_completion)(&con->swork) slock-AF_TIPC &list->lock#38 irq_context: 0 (wq_completion)tipc_send#4 (work_completion)(&con->swork) &con->outqueue_lock &obj_hash[i].lock irq_context: 0 (wq_completion)tipc_send#4 (work_completion)(&con->swork) &c->lock irq_context: 0 (wq_completion)tipc_send#4 (work_completion)(&con->swork) slock-AF_TIPC &c->lock irq_context: 0 (wq_completion)tipc_send#4 (work_completion)(&con->swork) slock-AF_TIPC &obj_hash[i].lock irq_context: 0 (wq_completion)tipc_send#4 (work_completion)(&con->swork) &list->lock#31 irq_context: 0 (wq_completion)tipc_send#4 (work_completion)(&con->swork) slock-AF_TIPC &list->lock#31 irq_context: 0 &xt[i].mutex remove_cache_srcu &meta->lock irq_context: 0 &xt[i].mutex remove_cache_srcu kfence_freelist_lock irq_context: 0 sk_lock-AF_ALG &mm->mmap_lock rcu_read_lock &obj_hash[i].lock irq_context: 0 sk_lock-AF_ALG &mm->mmap_lock fs_reclaim mmu_notifier_invalidate_range_start &rq->__lock irq_context: 0 sk_lock-AF_ALG &mm->mmap_lock fs_reclaim mmu_notifier_invalidate_range_start &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 sk_lock-AF_ALG &mm->mmap_lock &c->lock irq_context: 0 sk_lock-AF_ALG &mm->mmap_lock rcu_read_lock &rcu_state.expedited_wq irq_context: 0 sk_lock-AF_ALG &mm->mmap_lock rcu_read_lock &rcu_state.expedited_wq &p->pi_lock irq_context: 0 sk_lock-AF_ALG &mm->mmap_lock rcu_read_lock &rcu_state.expedited_wq &p->pi_lock &rq->__lock irq_context: 0 sk_lock-AF_ALG &mm->mmap_lock rcu_read_lock &rcu_state.expedited_wq &p->pi_lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 sb_writers#3 &sb->s_type->i_mutex_key#9 &rcu_state.expedited_wq &p->pi_lock irq_context: 0 sb_writers#3 &sb->s_type->i_mutex_key#9 &rcu_state.expedited_wq &p->pi_lock &rq->__lock irq_context: 0 sb_writers#3 &sb->s_type->i_mutex_key#9 &rcu_state.expedited_wq &p->pi_lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 sk_lock-AF_ALG &mm->mmap_lock &rcu_state.expedited_wq irq_context: 0 sk_lock-AF_ALG &mm->mmap_lock &rcu_state.expedited_wq &p->pi_lock irq_context: 0 sk_lock-AF_ALG &mm->mmap_lock &rcu_state.expedited_wq &p->pi_lock &rq->__lock irq_context: 0 sk_lock-AF_ALG &mm->mmap_lock &rcu_state.expedited_wq &p->pi_lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 sk_lock-AF_ALG &mm->mmap_lock batched_entropy_u8.lock irq_context: 0 sk_lock-AF_ALG &mm->mmap_lock kfence_freelist_lock irq_context: 0 sb_writers#4 &sb->s_type->i_mutex_key#8 mapping.invalidate_lock &sbi->s_writepages_rwsem irq_context: 0 sb_writers#4 &sb->s_type->i_mutex_key#8 mapping.invalidate_lock &sbi->s_writepages_rwsem &rsp->gp_wait irq_context: 0 sb_writers#4 &sb->s_type->i_mutex_key#8 mapping.invalidate_lock &sbi->s_writepages_rwsem rcu_state.exp_mutex rcu_node_0 irq_context: 0 sb_writers#4 &sb->s_type->i_mutex_key#8 mapping.invalidate_lock &sbi->s_writepages_rwsem rcu_state.exp_mutex &obj_hash[i].lock irq_context: 0 sb_writers#4 &sb->s_type->i_mutex_key#8 mapping.invalidate_lock &sbi->s_writepages_rwsem rcu_state.exp_mutex rcu_read_lock &pool->lock irq_context: 0 sb_writers#4 &sb->s_type->i_mutex_key#8 mapping.invalidate_lock &sbi->s_writepages_rwsem rcu_state.exp_mutex rcu_read_lock &pool->lock &obj_hash[i].lock irq_context: 0 sb_writers#4 &sb->s_type->i_mutex_key#8 mapping.invalidate_lock &sbi->s_writepages_rwsem rcu_state.exp_mutex rcu_read_lock &pool->lock &p->pi_lock irq_context: 0 sb_writers#4 &sb->s_type->i_mutex_key#8 mapping.invalidate_lock &sbi->s_writepages_rwsem rcu_state.exp_mutex rcu_read_lock &pool->lock &p->pi_lock &rq->__lock irq_context: 0 sb_writers#4 &sb->s_type->i_mutex_key#8 mapping.invalidate_lock &sbi->s_writepages_rwsem rcu_state.exp_mutex rcu_read_lock &pool->lock &p->pi_lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 sb_writers#4 &sb->s_type->i_mutex_key#8 mapping.invalidate_lock &sbi->s_writepages_rwsem rcu_state.exp_mutex &rq->__lock irq_context: 0 sb_writers#4 &sb->s_type->i_mutex_key#8 mapping.invalidate_lock &sbi->s_writepages_rwsem rcu_state.exp_mutex &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 sb_writers#4 &sb->s_type->i_mutex_key#8 mapping.invalidate_lock &sbi->s_writepages_rwsem rcu_state.exp_mutex &rnp->exp_wq[0] irq_context: 0 sb_writers#4 &sb->s_type->i_mutex_key#8 mapping.invalidate_lock &sbi->s_writepages_rwsem &obj_hash[i].lock irq_context: 0 sb_writers#4 &sb->s_type->i_mutex_key#8 mapping.invalidate_lock &sbi->s_writepages_rwsem jbd2_handle irq_context: 0 sb_writers#4 &sb->s_type->i_mutex_key#8 mapping.invalidate_lock &sbi->s_writepages_rwsem &journal->j_state_lock irq_context: 0 sb_writers#4 &sb->s_type->i_mutex_key#8 mapping.invalidate_lock &sbi->s_writepages_rwsem &journal->j_state_lock &journal->j_wait_updates irq_context: 0 sb_writers#4 &sb->s_type->i_mutex_key#8 mapping.invalidate_lock &sbi->s_writepages_rwsem &journal->j_barrier irq_context: 0 sb_writers#4 &sb->s_type->i_mutex_key#8 mapping.invalidate_lock &sbi->s_writepages_rwsem &journal->j_wait_transaction_locked irq_context: 0 sb_writers#4 &sb->s_type->i_mutex_key#8 mapping.invalidate_lock &sbi->s_writepages_rwsem &rq->__lock irq_context: 0 sb_writers#4 &sb->s_type->i_mutex_key#8 mapping.invalidate_lock &sbi->s_writepages_rwsem &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 sb_writers#4 &sb->s_type->i_mutex_key#8 mapping.invalidate_lock &sem->waiters irq_context: 0 sb_writers#4 &sb->s_type->i_mutex_key#8 mapping.invalidate_lock &rsp->gp_wait irq_context: 0 sb_writers#4 &sb->s_type->i_mutex_key#8 mapping.invalidate_lock &rsp->gp_wait &obj_hash[i].lock irq_context: 0 sb_writers#4 &sb->s_type->i_mutex_key#8 mapping.invalidate_lock &rsp->gp_wait pool_lock#2 irq_context: 0 namespace_sem batched_entropy_u8.lock irq_context: 0 namespace_sem kfence_freelist_lock irq_context: softirq (&n->timer) rcu_read_lock rcu_read_lock rcu_read_lock_bh &obj_hash[i].lock pool_lock irq_context: softirq &(&bat_priv->tt.work)->timer rcu_read_lock &pool->lock/1 fill_pool_map-wait-type-override &c->lock irq_context: softirq &(&bat_priv->tt.work)->timer rcu_read_lock &pool->lock/1 fill_pool_map-wait-type-override &n->list_lock irq_context: softirq &(&bat_priv->tt.work)->timer rcu_read_lock &pool->lock/1 fill_pool_map-wait-type-override &n->list_lock &c->lock irq_context: softirq &(&bat_priv->tt.work)->timer rcu_read_lock &pool->lock/1 fill_pool_map-wait-type-override pool_lock irq_context: 0 sb_writers#4 &sb->s_type->i_mutex_key#8 mapping.invalidate_lock &sbi->s_writepages_rwsem rcu_state.exp_mutex &rnp->exp_wq[1] irq_context: 0 pernet_ops_rwsem k-sk_lock-AF_RXRPC batched_entropy_u8.lock irq_context: 0 pernet_ops_rwsem k-sk_lock-AF_RXRPC kfence_freelist_lock irq_context: 0 sb_writers#4 &sb->s_type->i_mutex_key#8 mapping.invalidate_lock &sbi->s_writepages_rwsem rcu_state.exp_mutex &rnp->exp_wq[2] irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem &fsnotify_mark_srcu &rq->__lock irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem &fsnotify_mark_srcu &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 &f->f_pos_lock sb_writers#4 &sb->s_type->i_mutex_key#8 &mm->mmap_lock rcu_read_lock &rq->__lock &cfs_rq->removed.lock irq_context: 0 (wq_completion)wg-crypt-wg1#2 (work_completion)(&peer->transmit_packet_work) &peer->endpoint_lock rcu_read_lock_bh batched_entropy_u32.lock crngs.lock irq_context: 0 sb_writers#4 &sb->s_type->i_mutex_key#8 &sbi->s_writepages_rwsem jbd2_handle &ei->i_data_sem &mapping->private_lock irq_context: 0 &iint->mutex &rcu_state.expedited_wq irq_context: 0 &iint->mutex &rcu_state.expedited_wq &p->pi_lock irq_context: 0 &iint->mutex &rcu_state.expedited_wq &p->pi_lock &rq->__lock irq_context: 0 &iint->mutex &rcu_state.expedited_wq &p->pi_lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 &iint->mutex mapping.invalidate_lock rcu_read_lock &rcu_state.expedited_wq irq_context: 0 &iint->mutex mapping.invalidate_lock rcu_read_lock &rcu_state.expedited_wq &p->pi_lock irq_context: 0 &iint->mutex mapping.invalidate_lock rcu_read_lock &rcu_state.expedited_wq &p->pi_lock &rq->__lock irq_context: 0 &iint->mutex mapping.invalidate_lock rcu_read_lock &rcu_state.expedited_wq &p->pi_lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 sb_writers#4 &type->i_mutex_dir_key#3/1 &type->i_mutex_dir_key#3 &rcu_state.expedited_wq &p->pi_lock &rq->__lock irq_context: 0 sb_writers#4 &type->i_mutex_dir_key#3/1 &type->i_mutex_dir_key#3 &rcu_state.expedited_wq &p->pi_lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 &iint->mutex mapping.invalidate_lock mmu_notifier_invalidate_range_start &rq->__lock irq_context: 0 &iint->mutex mapping.invalidate_lock mmu_notifier_invalidate_range_start &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 &iint->mutex rcu_read_lock &rcu_state.gp_wq irq_context: 0 &iint->mutex rcu_read_lock &rcu_state.gp_wq &p->pi_lock irq_context: 0 &iint->mutex rcu_read_lock &rcu_state.gp_wq &p->pi_lock &rq->__lock irq_context: 0 &iint->mutex rcu_read_lock &rcu_state.gp_wq &p->pi_lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 &iint->mutex mapping.invalidate_lock rcu_read_lock &n->list_lock irq_context: 0 &iint->mutex mapping.invalidate_lock rcu_read_lock &n->list_lock &c->lock irq_context: 0 ebt_mutex &rq->__lock &cfs_rq->removed.lock irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem ovs_mutex nf_ct_proto_mutex defrag4_mutex &rq->__lock irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem ovs_mutex nf_ct_proto_mutex defrag4_mutex &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)events (work_completion)(&aux->work) &rcu_state.expedited_wq irq_context: 0 (wq_completion)events_power_efficient (work_completion)(&(&tbl->gc_work)->work) &tbl->lock fill_pool_map-wait-type-override &c->lock irq_context: softirq (&n->timer) rcu_read_lock rcu_read_lock pool_lock#2 irq_context: 0 (wq_completion)events (work_completion)(&aux->work) &rcu_state.expedited_wq &p->pi_lock irq_context: 0 (wq_completion)events (work_completion)(&aux->work) &rcu_state.expedited_wq &p->pi_lock &rq->__lock irq_context: 0 (wq_completion)events (work_completion)(&aux->work) &rcu_state.expedited_wq &p->pi_lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: softirq (&n->timer) rcu_read_lock rcu_read_lock &obj_hash[i].lock irq_context: softirq (&n->timer) rcu_read_lock rcu_read_lock &dir->lock#2 irq_context: softirq (&n->timer) rcu_read_lock rcu_read_lock &tbl->lock irq_context: softirq (&n->timer) k-slock-AF_INET6 rcu_read_lock rcu_read_lock rcu_read_lock_bh &sch->q.lock &____s->seqcount#2 irq_context: softirq (&n->timer) k-slock-AF_INET6 rcu_read_lock rcu_read_lock rcu_read_lock_bh &sch->q.lock &____s->seqcount irq_context: 0 &sb->s_type->i_mutex_key#10 (netlink_chain).rwsem pgd_lock irq_context: 0 &sb->s_type->i_mutex_key#10 (netlink_chain).rwsem stock_lock irq_context: 0 &sb->s_type->i_mutex_key#10 (netlink_chain).rwsem key irq_context: 0 &sb->s_type->i_mutex_key#10 (netlink_chain).rwsem pcpu_lock irq_context: 0 &sb->s_type->i_mutex_key#10 (netlink_chain).rwsem percpu_counters_lock irq_context: 0 &sb->s_type->i_mutex_key#10 (netlink_chain).rwsem pcpu_lock stock_lock irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem k-sk_lock-AF_RXRPC fill_pool_map-wait-type-override &rq->__lock irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem k-sk_lock-AF_RXRPC fill_pool_map-wait-type-override &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 sb_writers#3 oom_adj_mutex rcu_node_0 irq_context: 0 rtnl_mutex &idev->mc_lock _xmit_IPGRE irq_context: 0 rtnl_mutex &idev->mc_lock _xmit_IPGRE &c->lock irq_context: 0 rtnl_mutex &idev->mc_lock _xmit_IPGRE pool_lock#2 irq_context: 0 rtnl_mutex &idev->mc_lock _xmit_IPGRE &obj_hash[i].lock irq_context: 0 rtnl_mutex &idev->mc_lock _xmit_IPGRE krc.lock irq_context: 0 &sb->s_type->i_mutex_key#10 sk_lock-AF_INET6 remove_cache_srcu irq_context: 0 &sb->s_type->i_mutex_key#10 sk_lock-AF_INET6 remove_cache_srcu quarantine_lock irq_context: 0 sk_lock-AF_INET6 &tcp_hashinfo.bhash[i].lock &tcp_hashinfo.bhash2[i].lock &____s->seqcount#2 irq_context: 0 sk_lock-AF_INET6 &tcp_hashinfo.bhash[i].lock &tcp_hashinfo.bhash2[i].lock &____s->seqcount irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem rcu_state.barrier_mutex rcu_state.barrier_lock fill_pool_map-wait-type-override pool_lock#2 irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem rcu_state.barrier_mutex rcu_state.barrier_lock fill_pool_map-wait-type-override pool_lock irq_context: 0 fill_pool_map-wait-type-override &rcu_state.expedited_wq irq_context: 0 rtnl_mutex &im->lock pool_lock#2 irq_context: 0 pernet_ops_rwsem devices_rwsem rdma_nets_rwsem &device->compat_devs_mutex fs_reclaim &rq->__lock &cfs_rq->removed.lock irq_context: 0 &nft_net->commit_mutex (work_completion)(&(&priv->gc_work)->work)#2 irq_context: 0 &nft_net->commit_mutex rcu_state.barrier_mutex irq_context: 0 &nft_net->commit_mutex rcu_state.barrier_mutex rcu_state.barrier_lock irq_context: 0 &nft_net->commit_mutex rcu_state.barrier_mutex rcu_state.barrier_lock &obj_hash[i].lock irq_context: 0 &nft_net->commit_mutex rcu_state.barrier_mutex &x->wait#24 irq_context: 0 &nft_net->commit_mutex rcu_state.barrier_mutex &rq->__lock irq_context: 0 &nft_net->commit_mutex rcu_state.barrier_mutex &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 rtnl_mutex &net->ipv4.ra_mutex irq_context: 0 rtnl_mutex mfc_unres_lock#2 irq_context: 0 rtnl_mutex mfc_unres_lock#2 &obj_hash[i].lock irq_context: 0 &sb->s_type->i_mutex_key#10 rtnl_mutex fs_reclaim irq_context: 0 &sb->s_type->i_mutex_key#10 rtnl_mutex fs_reclaim mmu_notifier_invalidate_range_start irq_context: 0 &sb->s_type->i_mutex_key#10 rtnl_mutex &c->lock irq_context: 0 &sb->s_type->i_mutex_key#10 rtnl_mutex &____s->seqcount irq_context: 0 &sb->s_type->i_mutex_key#10 rtnl_mutex rcu_read_lock pool_lock#2 irq_context: 0 &sb->s_type->i_mutex_key#10 rtnl_mutex nl_table_lock irq_context: 0 &sb->s_type->i_mutex_key#10 rtnl_mutex nl_table_wait.lock irq_context: 0 &sb->s_type->i_mutex_key#10 rtnl_mutex rcu_read_lock rhashtable_bucket irq_context: 0 (wq_completion)events_power_efficient (work_completion)(&(&gc_work->dwork)->work) rcu_read_lock &rq->__lock &cfs_rq->removed.lock irq_context: 0 sb_writers#4 &type->i_mutex_dir_key#3/1 jbd2_handle remove_cache_srcu rcu_read_lock rcu_node_0 irq_context: 0 sb_writers#4 &type->i_mutex_dir_key#3/1 jbd2_handle remove_cache_srcu rcu_read_lock &rq->__lock irq_context: 0 sb_writers#4 &type->i_mutex_dir_key#3/1 jbd2_handle remove_cache_srcu rcu_read_lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 cb_lock genl_mutex sk_lock-AF_NETLINK irq_context: 0 cb_lock genl_mutex sk_lock-AF_NETLINK slock-AF_NETLINK irq_context: 0 cb_lock genl_mutex sk_lock-AF_NETLINK clock-AF_NETLINK irq_context: 0 cb_lock genl_mutex slock-AF_NETLINK irq_context: 0 pernet_ops_rwsem ovs_mutex nf_ct_proto_mutex defrag6_mutex rcu_read_lock rcu_node_0 irq_context: 0 pernet_ops_rwsem ovs_mutex nf_ct_proto_mutex defrag6_mutex rcu_read_lock &rq->__lock irq_context: 0 pernet_ops_rwsem devices_rwsem rdma_nets_rwsem &device->compat_devs_mutex remove_cache_srcu rcu_node_0 irq_context: 0 (wq_completion)writeback (work_completion)(&(&wb->dwork)->work) &type->s_umount_key#43 rcu_read_lock &rcu_state.expedited_wq irq_context: 0 (wq_completion)writeback (work_completion)(&(&wb->dwork)->work) &type->s_umount_key#43 rcu_read_lock &rcu_state.expedited_wq &p->pi_lock irq_context: 0 pernet_ops_rwsem devices_rwsem rdma_nets_rwsem &device->compat_devs_mutex remove_cache_srcu &rcu_state.expedited_wq irq_context: 0 pernet_ops_rwsem devices_rwsem rdma_nets_rwsem &device->compat_devs_mutex remove_cache_srcu &rcu_state.expedited_wq &p->pi_lock irq_context: 0 pernet_ops_rwsem devices_rwsem rdma_nets_rwsem &device->compat_devs_mutex remove_cache_srcu &rcu_state.expedited_wq &p->pi_lock &rq->__lock irq_context: 0 pernet_ops_rwsem devices_rwsem rdma_nets_rwsem &device->compat_devs_mutex remove_cache_srcu &rcu_state.expedited_wq &p->pi_lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)writeback (work_completion)(&(&wb->dwork)->work) &type->s_umount_key#43 rcu_read_lock &rcu_state.expedited_wq &p->pi_lock &rq->__lock irq_context: 0 (wq_completion)writeback (work_completion)(&(&wb->dwork)->work) &type->s_umount_key#43 rcu_read_lock &rcu_state.expedited_wq &p->pi_lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)events fqdir_free_work rcu_state.barrier_mutex rcu_node_0 irq_context: 0 (wq_completion)events fqdir_free_work rcu_state.barrier_mutex &rcu_state.expedited_wq irq_context: 0 (wq_completion)events fqdir_free_work rcu_state.barrier_mutex &rcu_state.expedited_wq &p->pi_lock irq_context: 0 (wq_completion)events fqdir_free_work rcu_state.barrier_mutex &rcu_state.expedited_wq &p->pi_lock &rq->__lock irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem &wq->mutex rcu_node_0 irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem &wq->mutex &rcu_state.expedited_wq irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem &wq->mutex &rcu_state.expedited_wq &p->pi_lock irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem &wq->mutex &rcu_state.expedited_wq &p->pi_lock &rq->__lock irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem &wq->mutex &rcu_state.expedited_wq &p->pi_lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)events fqdir_free_work rcu_state.barrier_mutex &rcu_state.expedited_wq &p->pi_lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem rtnl_mutex stock_lock rcu_read_lock per_cpu_ptr(&cgroup_rstat_cpu_lock, cpu) irq_context: 0 &tsk->futex_exit_mutex &mm->mmap_lock &n->list_lock irq_context: 0 &tsk->futex_exit_mutex &mm->mmap_lock &n->list_lock &c->lock irq_context: 0 (wq_completion)bond6 (work_completion)(&(&bond->alb_work)->work) &rq->__lock irq_context: 0 pernet_ops_rwsem nf_hook_mutex fs_reclaim &rq->__lock irq_context: 0 pernet_ops_rwsem nf_hook_mutex fs_reclaim &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: softirq rcu_read_lock rcu_read_lock &idev->mc_query_lock irq_context: softirq rcu_read_lock rcu_read_lock &idev->mc_query_lock &obj_hash[i].lock irq_context: softirq rcu_read_lock rcu_read_lock &idev->mc_query_lock rcu_read_lock &pool->lock irq_context: softirq rcu_read_lock rcu_read_lock &idev->mc_query_lock rcu_read_lock &pool->lock &obj_hash[i].lock irq_context: softirq rcu_read_lock rcu_read_lock &idev->mc_query_lock rcu_read_lock &pool->lock pool_lock#2 irq_context: softirq rcu_read_lock rcu_read_lock &idev->mc_query_lock rcu_read_lock &pool->lock &p->pi_lock irq_context: softirq rcu_read_lock rcu_read_lock &idev->mc_query_lock rcu_read_lock &pool->lock &p->pi_lock &rq->__lock irq_context: softirq rcu_read_lock rcu_read_lock &idev->mc_query_lock rcu_read_lock &pool->lock &p->pi_lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)mld (work_completion)(&(&idev->mc_query_work)->work) irq_context: 0 (wq_completion)mld (work_completion)(&(&idev->mc_query_work)->work) &idev->mc_query_lock irq_context: 0 (wq_completion)mld (work_completion)(&(&idev->mc_query_work)->work) &idev->mc_lock irq_context: 0 (wq_completion)mld (work_completion)(&(&idev->mc_query_work)->work) &idev->mc_lock &obj_hash[i].lock irq_context: 0 (wq_completion)mld (work_completion)(&(&idev->mc_query_work)->work) &idev->mc_lock pool_lock#2 irq_context: 0 sk_lock-AF_INET6 rcu_read_lock rcu_read_lock_bh rcu_read_lock &____s->seqcount#12 irq_context: 0 sk_lock-AF_INET6 rcu_read_lock rcu_read_lock_bh rcu_read_lock &ndev->lock irq_context: 0 pernet_ops_rwsem rtnl_mutex &idev->mc_lock fs_reclaim mmu_notifier_invalidate_range_start &rq->__lock irq_context: 0 &xt[i].mutex rcu_read_lock &rcu_state.gp_wq &p->pi_lock &cfs_rq->removed.lock irq_context: 0 sb_writers#4 &sb->s_type->i_mutex_key#8 &sbi->s_writepages_rwsem rcu_state.exp_mutex rcu_read_lock &pool->lock &p->pi_lock &rq->__lock &cfs_rq->removed.lock irq_context: 0 sb_writers#4 &type->i_mutex_dir_key#3 tomoyo_ss batched_entropy_u8.lock irq_context: 0 sb_writers#4 &type->i_mutex_dir_key#3 tomoyo_ss kfence_freelist_lock irq_context: 0 sb_writers#4 &type->i_mutex_dir_key#3 tomoyo_ss &meta->lock irq_context: 0 sb_writers#4 &sb->s_type->i_mutex_key#8 &sbi->s_writepages_rwsem &sem->waiters irq_context: 0 (wq_completion)bat_events (work_completion)(&(&forw_packet_aggr->delayed_work)->work) &hard_iface->bat_iv.ogm_buff_mutex fill_pool_map-wait-type-override &c->lock irq_context: 0 rcu_state.exp_mutex fill_pool_map-wait-type-override pool_lock#2 irq_context: 0 rcu_state.exp_mutex fill_pool_map-wait-type-override &c->lock irq_context: 0 rcu_state.exp_mutex fill_pool_map-wait-type-override pool_lock irq_context: 0 rcu_read_lock &pool->lock fill_pool_map-wait-type-override &c->lock irq_context: 0 rcu_read_lock &pool->lock fill_pool_map-wait-type-override &n->list_lock irq_context: 0 rcu_read_lock &pool->lock fill_pool_map-wait-type-override &n->list_lock &c->lock irq_context: 0 (wq_completion)ext4-rsv-conversion (work_completion)(&ei->i_rsv_conversion_work) jbd2_handle &rcu_state.expedited_wq &p->pi_lock irq_context: 0 (wq_completion)ext4-rsv-conversion (work_completion)(&ei->i_rsv_conversion_work) jbd2_handle &rcu_state.expedited_wq &p->pi_lock &rq->__lock irq_context: 0 (wq_completion)ext4-rsv-conversion (work_completion)(&ei->i_rsv_conversion_work) jbd2_handle &rcu_state.expedited_wq &p->pi_lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem ovs_mutex nf_ct_proto_mutex fill_pool_map-wait-type-override pool_lock#2 irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem ovs_mutex nf_ct_proto_mutex fill_pool_map-wait-type-override &c->lock irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem ovs_mutex nf_ct_proto_mutex fill_pool_map-wait-type-override pool_lock irq_context: 0 sk_lock-AF_INET6 &mm->mmap_lock sb_pagefaults mapping.invalidate_lock rcu_read_lock &xa->xa_lock#9 rcu_read_lock per_cpu_ptr(&cgroup_rstat_cpu_lock, cpu) irq_context: 0 sk_lock-AF_INET6 &mm->mmap_lock sb_pagefaults mapping.invalidate_lock rcu_read_lock &xa->xa_lock#9 key#10 irq_context: 0 (wq_completion)events (work_completion)(&ht->run_work) &ht->mutex &ht->lock fill_pool_map-wait-type-override pool_lock#2 irq_context: 0 (wq_completion)events (work_completion)(&ht->run_work) &ht->mutex &ht->lock fill_pool_map-wait-type-override pool_lock irq_context: 0 (wq_completion)events (work_completion)(&pwq->unbound_release_work) rcu_state.exp_mutex fill_pool_map-wait-type-override &c->lock irq_context: 0 (wq_completion)events (work_completion)(&pwq->unbound_release_work) rcu_state.exp_mutex fill_pool_map-wait-type-override pool_lock#2 irq_context: 0 (wq_completion)events (work_completion)(&pwq->unbound_release_work) rcu_state.exp_mutex fill_pool_map-wait-type-override &____s->seqcount#2 irq_context: 0 (wq_completion)events (work_completion)(&pwq->unbound_release_work) rcu_state.exp_mutex fill_pool_map-wait-type-override &rq->__lock irq_context: 0 (wq_completion)events (work_completion)(&pwq->unbound_release_work) rcu_state.exp_mutex fill_pool_map-wait-type-override &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)events (work_completion)(&pwq->unbound_release_work) rcu_state.exp_mutex fill_pool_map-wait-type-override &____s->seqcount irq_context: 0 (wq_completion)events (work_completion)(&pwq->unbound_release_work) rcu_state.exp_mutex fill_pool_map-wait-type-override pool_lock irq_context: 0 rtnl_mutex &idev->mc_lock &vlan_netdev_addr_lock_key/2 irq_context: 0 rtnl_mutex &idev->mc_lock &vlan_netdev_addr_lock_key/2 &obj_hash[i].lock irq_context: 0 pernet_ops_rwsem &pool->lock/1 rcu_read_lock &pool->lock fill_pool_map-wait-type-override pool_lock#2 irq_context: 0 pernet_ops_rwsem &pool->lock/1 rcu_read_lock &pool->lock fill_pool_map-wait-type-override pool_lock irq_context: 0 (wq_completion)events (work_completion)(&pwq->unbound_release_work) rcu_read_lock &rq->__lock &cfs_rq->removed.lock irq_context: 0 rtnl_mutex &idev->mc_lock &vlan_netdev_addr_lock_key/2 krc.lock irq_context: 0 &f->f_pos_lock &type->i_mutex_dir_key#3 sb_writers#4 remove_cache_srcu rcu_node_0 irq_context: 0 sk_lock-AF_INET rcu_read_lock rcu_read_lock_bh &obj_hash[i].lock pool_lock irq_context: 0 sk_lock-AF_INET6 fill_pool_map-wait-type-override &obj_hash[i].lock irq_context: 0 &sb->s_type->i_mutex_key#8 &sbi->s_writepages_rwsem jbd2_handle &ei->i_data_sem rcu_read_lock &rcu_state.expedited_wq irq_context: 0 &sb->s_type->i_mutex_key#8 &sbi->s_writepages_rwsem jbd2_handle &ei->i_data_sem rcu_read_lock &rcu_state.expedited_wq &p->pi_lock irq_context: 0 &sb->s_type->i_mutex_key#8 &sbi->s_writepages_rwsem jbd2_handle &ei->i_data_sem rcu_read_lock &rcu_state.expedited_wq &p->pi_lock &rq->__lock irq_context: 0 &sb->s_type->i_mutex_key#8 &sbi->s_writepages_rwsem jbd2_handle &ei->i_data_sem rcu_read_lock &rcu_state.expedited_wq &p->pi_lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 &sb->s_type->i_mutex_key#10 sk_lock-AF_INET6/1 fill_pool_map-wait-type-override rcu_read_lock rcu_node_0 irq_context: 0 &sb->s_type->i_mutex_key#10 sk_lock-AF_INET6/1 fill_pool_map-wait-type-override rcu_read_lock &rq->__lock irq_context: softirq rcu_read_lock rcu_read_lock slock-AF_INET/1 fill_pool_map-wait-type-override pool_lock#2 irq_context: softirq rcu_read_lock rcu_read_lock slock-AF_INET/1 fill_pool_map-wait-type-override pool_lock irq_context: 0 put_task_map-wait-type-override#2 &obj_hash[i].lock irq_context: 0 put_task_map-wait-type-override#2 pool_lock#2 irq_context: 0 put_task_map-wait-type-override#2 stock_lock irq_context: 0 put_task_map-wait-type-override#2 per_cpu_ptr(&cgroup_rstat_cpu_lock, cpu) irq_context: 0 (wq_completion)events_unbound (work_completion)(&sub_info->work) fill_pool_map-wait-type-override &rq->__lock irq_context: 0 (wq_completion)events_unbound (work_completion)(&sub_info->work) fill_pool_map-wait-type-override &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 rtnl_mutex mrt_lock#2 irq_context: 0 rtnl_mutex mrt_lock#2 pool_lock#2 irq_context: 0 rtnl_mutex mrt_lock#2 &dir->lock#2 irq_context: 0 rtnl_mutex deferred_probe_mutex &rq->__lock irq_context: 0 rtnl_mutex deferred_probe_mutex &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 sk_lock-AF_ROSE &mm->mmap_lock irq_context: 0 &sb->s_type->i_mutex_key#10 sk_lock-AF_ROSE fill_pool_map-wait-type-override &rq->__lock irq_context: 0 &sb->s_type->i_mutex_key#10 sk_lock-AF_ROSE fill_pool_map-wait-type-override &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)events_long (work_completion)(&(&ipvs->defense_work)->work) rcu_node_0 irq_context: 0 sb_writers#4 &sb->s_type->i_mutex_key#8 &sbi->s_writepages_rwsem &p->pi_lock &rq->__lock &cfs_rq->removed.lock irq_context: 0 (wq_completion)wg-crypt-wg1 (work_completion)(&peer->transmit_packet_work) rcu_read_lock_bh rcu_read_lock &pool->lock/1 irq_context: 0 (wq_completion)wg-crypt-wg1 (work_completion)(&peer->transmit_packet_work) rcu_read_lock_bh rcu_read_lock &pool->lock/1 &obj_hash[i].lock irq_context: 0 (wq_completion)wg-crypt-wg1 (work_completion)(&peer->transmit_packet_work) rcu_read_lock_bh rcu_read_lock &pool->lock/1 &p->pi_lock irq_context: 0 (wq_completion)wg-kex-wg2#2 (work_completion)(&({ do { const void *__vpp_verify = (typeof((worker) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __asm__ ("" : "=r"(__ptr) : "0"((typeof(*((worker))) *)((worker)))); (typeof((typeof(*((worker))) *)((worker)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); })->work) &peer->endpoint_lock rcu_read_lock_bh &obj_hash[i].lock irq_context: 0 (wq_completion)wg-kex-wg2#7 (work_completion)(&peer->transmit_handshake_work) &peer->endpoint_lock rcu_read_lock_bh &obj_hash[i].lock irq_context: 0 (wq_completion)wg-kex-wg1#12 (work_completion)(&({ do { const void *__vpp_verify = (typeof((worker) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __asm__ ("" : "=r"(__ptr) : "0"((typeof(*((worker))) *)((worker)))); (typeof((typeof(*((worker))) *)((worker)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); })->work) &handshake->lock rcu_read_lock_bh &peer->keypairs.keypair_update_lock &obj_hash[i].lock pool_lock irq_context: 0 (wq_completion)wg-kex-wg1#12 (work_completion)(&({ do { const void *__vpp_verify = (typeof((worker) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __asm__ ("" : "=r"(__ptr) : "0"((typeof(*((worker))) *)((worker)))); (typeof((typeof(*((worker))) *)((worker)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); })->work) &handshake->lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 vlan_ioctl_mutex &mm->mmap_lock &sem->wait_lock irq_context: 0 vlan_ioctl_mutex &mm->mmap_lock &rq->__lock irq_context: 0 vlan_ioctl_mutex &mm->mmap_lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 vlan_ioctl_mutex &mm->mmap_lock fs_reclaim irq_context: 0 vlan_ioctl_mutex &mm->mmap_lock fs_reclaim mmu_notifier_invalidate_range_start irq_context: 0 vlan_ioctl_mutex &mm->mmap_lock &____s->seqcount irq_context: 0 vlan_ioctl_mutex &mm->mmap_lock stock_lock irq_context: 0 vlan_ioctl_mutex &mm->mmap_lock &obj_hash[i].lock irq_context: 0 vlan_ioctl_mutex &mm->mmap_lock &c->lock irq_context: 0 rtnl_mutex &sch->q.lock tk_core.seq.seqcount irq_context: 0 (wq_completion)mld (work_completion)(&(&idev->mc_ifc_work)->work) &idev->mc_lock rcu_read_lock rcu_read_lock rcu_read_lock_bh rcu_read_lock quarantine_lock irq_context: softirq (&in_dev->mr_ifc_timer) rcu_read_lock batched_entropy_u8.lock irq_context: softirq (&in_dev->mr_ifc_timer) rcu_read_lock kfence_freelist_lock irq_context: 0 rtnl_mutex &rdev->wiphy.mtx &wdev->mtx &local->sta_mtx rcu_state.exp_mutex &rnp->exp_wq[3] irq_context: 0 rtnl_mutex &rdev->wiphy.mtx &wdev->mtx &local->sta_mtx &n->list_lock irq_context: 0 rtnl_mutex &rdev->wiphy.mtx &wdev->mtx &local->sta_mtx &n->list_lock &c->lock irq_context: 0 rtnl_mutex &rdev->wiphy.mtx &wdev->mtx &local->sta_mtx &obj_hash[i].lock pool_lock irq_context: 0 rtnl_mutex &rdev->wiphy.mtx &wdev->mtx rcu_state.exp_mutex &rnp->exp_wq[0] irq_context: 0 rtnl_mutex &rdev->wiphy.mtx &wdev->mtx &local->chanctx_mtx irq_context: 0 rtnl_mutex &rdev->wiphy.mtx &wdev->mtx &rdev->wiphy_work_lock irq_context: 0 (wq_completion)ipv6_addrconf (work_completion)(&(&ifa->dad_work)->work) rtnl_mutex rcu_read_lock rcu_read_lock rcu_read_lock_bh &sch->q.lock &____s->seqcount#2 irq_context: 0 (wq_completion)ipv6_addrconf (work_completion)(&(&ifa->dad_work)->work) rtnl_mutex rcu_read_lock rcu_read_lock rcu_read_lock_bh &sch->q.lock &____s->seqcount irq_context: 0 (wq_completion)ipv6_addrconf (work_completion)(&(&ifa->dad_work)->work) rtnl_mutex rcu_read_lock rcu_read_lock rcu_read_lock_bh &sch->q.lock pool_lock#2 irq_context: 0 (wq_completion)ipv6_addrconf (work_completion)(&(&ifa->dad_work)->work) rtnl_mutex rcu_read_lock rcu_read_lock rcu_read_lock_bh &sch->q.lock &c->lock irq_context: 0 (wq_completion)ipv6_addrconf (work_completion)(&(&ifa->dad_work)->work) rtnl_mutex rcu_read_lock rcu_read_lock rcu_read_lock_bh &sch->q.lock rcu_read_lock pool_lock#2 irq_context: 0 (wq_completion)ipv6_addrconf (work_completion)(&(&ifa->dad_work)->work) rtnl_mutex rcu_read_lock rcu_read_lock rcu_read_lock_bh &sch->q.lock &obj_hash[i].lock irq_context: 0 (wq_completion)ipv6_addrconf (work_completion)(&(&ifa->dad_work)->work) rtnl_mutex rcu_read_lock rcu_read_lock rcu_read_lock_bh &sch->q.lock batched_entropy_u8.lock irq_context: 0 (wq_completion)ipv6_addrconf (work_completion)(&(&ifa->dad_work)->work) rtnl_mutex rcu_read_lock rcu_read_lock rcu_read_lock_bh &sch->q.lock kfence_freelist_lock irq_context: 0 sb_writers#4 &sb->s_type->i_mutex_key#8 jbd2_handle &sbi->s_orphan_lock mmu_notifier_invalidate_range_start irq_context: 0 sb_writers#4 &sb->s_type->i_mutex_key#8 jbd2_handle &sbi->s_orphan_lock &____s->seqcount irq_context: 0 sb_writers#4 &sb->s_type->i_mutex_key#8 jbd2_handle &sbi->s_orphan_lock pool_lock#2 irq_context: 0 sb_writers#4 &sb->s_type->i_mutex_key#8 jbd2_handle &sbi->s_orphan_lock rcu_read_lock pool_lock#2 irq_context: 0 sb_writers#4 &sb->s_type->i_mutex_key#8 jbd2_handle &sbi->s_orphan_lock &obj_hash[i].lock irq_context: 0 sb_writers#4 &sb->s_type->i_mutex_key#8 mapping.invalidate_lock jbd2_handle &ei->i_data_sem bit_wait_table + i irq_context: 0 sb_writers#4 &type->i_mutex_dir_key#3/1 &fsnotify_mark_srcu &obj_hash[i].lock irq_context: 0 sb_writers#4 tomoyo_ss remove_cache_srcu rcu_read_lock &rcu_state.gp_wq irq_context: 0 sb_writers#4 tomoyo_ss remove_cache_srcu rcu_read_lock &rcu_state.gp_wq &p->pi_lock irq_context: 0 sb_writers#4 tomoyo_ss remove_cache_srcu rcu_read_lock &rcu_state.gp_wq &p->pi_lock &rq->__lock irq_context: 0 sb_writers#4 tomoyo_ss remove_cache_srcu rcu_read_lock &rcu_state.gp_wq &p->pi_lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 &f->f_pos_lock &type->i_mutex_dir_key#3 sb_writers#4 &rq->__lock &cfs_rq->removed.lock irq_context: 0 (wq_completion)events (work_completion)(&rdev->wiphy_work) &rdev->wiphy.mtx &wdev->mtx &local->mtx &data->mutex irq_context: 0 (wq_completion)events (work_completion)(&rdev->wiphy_work) &rdev->wiphy.mtx &wdev->mtx &local->mtx &local->queue_stop_reason_lock irq_context: 0 (wq_completion)events (work_completion)(&rdev->wiphy_work) &rdev->wiphy.mtx &wdev->mtx &local->mtx &local->iflist_mtx irq_context: 0 (wq_completion)events (work_completion)(&rdev->wiphy_work) &rdev->wiphy.mtx &wdev->mtx &local->mtx &local->filter_lock irq_context: 0 (wq_completion)events (work_completion)(&rdev->wiphy_work) &rdev->wiphy.mtx &wdev->mtx &local->mtx rcu_read_lock &pool->lock/1 irq_context: 0 (wq_completion)events (work_completion)(&rdev->wiphy_work) &rdev->wiphy.mtx &wdev->mtx &local->mtx rcu_read_lock &pool->lock/1 &obj_hash[i].lock irq_context: 0 (wq_completion)events (work_completion)(&rdev->wiphy_work) &rdev->wiphy.mtx &wdev->mtx &local->mtx rcu_read_lock &pool->lock/1 &p->pi_lock irq_context: 0 (wq_completion)events (work_completion)(&rdev->wiphy_work) &rdev->wiphy.mtx &wdev->mtx &local->mtx rcu_read_lock &pool->lock/1 &p->pi_lock &rq->__lock irq_context: 0 (wq_completion)events (work_completion)(&rdev->wiphy_work) &rdev->wiphy.mtx &wdev->mtx &local->mtx rcu_read_lock &pool->lock/1 &p->pi_lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)events (work_completion)(&rdev->wiphy_work) &rdev->wiphy.mtx &wdev->mtx &local->mtx &rq->__lock irq_context: 0 (wq_completion)events (work_completion)(&rdev->wiphy_work) &rdev->wiphy.mtx &wdev->mtx &local->mtx &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)phy12 (work_completion)(&(&local->scan_work)->work) irq_context: 0 (wq_completion)phy12 (work_completion)(&(&local->scan_work)->work) &rq->__lock irq_context: 0 (wq_completion)phy12 (work_completion)(&(&local->scan_work)->work) &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)phy12 (work_completion)(&(&local->scan_work)->work) &local->mtx irq_context: 0 (wq_completion)phy12 (work_completion)(&(&local->scan_work)->work) &local->mtx &lock->wait_lock irq_context: 0 (wq_completion)phy12 (work_completion)(&(&local->scan_work)->work) &local->mtx &rq->__lock irq_context: 0 (wq_completion)phy12 (work_completion)(&(&local->scan_work)->work) &local->mtx &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)events (work_completion)(&rdev->wiphy_work) &rdev->wiphy.mtx &wdev->mtx &lock->wait_lock irq_context: 0 (wq_completion)phy12 (work_completion)(&(&local->scan_work)->work) &local->mtx &local->iflist_mtx irq_context: 0 (wq_completion)phy12 (work_completion)(&(&local->scan_work)->work) &local->mtx &data->mutex irq_context: 0 (wq_completion)phy12 (work_completion)(&(&local->scan_work)->work) &local->mtx &obj_hash[i].lock irq_context: 0 (wq_completion)phy12 (work_completion)(&(&local->scan_work)->work) &local->mtx &base->lock irq_context: 0 (wq_completion)phy12 (work_completion)(&(&local->scan_work)->work) &local->mtx &base->lock &obj_hash[i].lock irq_context: 0 (wq_completion)phy12 (work_completion)(&(&local->scan_work)->work) &local->mtx pool_lock#2 irq_context: 0 (wq_completion)phy12 (work_completion)(&(&local->scan_work)->work) &local->mtx rcu_read_lock &local->queue_stop_reason_lock irq_context: 0 (wq_completion)phy12 (work_completion)(&(&local->scan_work)->work) &local->mtx rcu_read_lock tk_core.seq.seqcount irq_context: 0 (wq_completion)phy12 (work_completion)(&(&local->scan_work)->work) &local->mtx rcu_read_lock hwsim_radio_lock irq_context: 0 (wq_completion)phy12 (work_completion)(&(&local->scan_work)->work) &local->mtx rcu_read_lock hwsim_radio_lock pool_lock#2 irq_context: 0 (wq_completion)phy12 (work_completion)(&(&local->scan_work)->work) &local->mtx rcu_read_lock hwsim_radio_lock &list->lock#19 irq_context: 0 (wq_completion)phy12 (work_completion)(&(&local->scan_work)->work) &local->mtx rcu_read_lock &list->lock#19 irq_context: softirq rcu_read_lock &local->rx_path_lock &rdev->mgmt_registrations_lock irq_context: 0 (wq_completion)phy12 (work_completion)(&(&local->scan_work)->work) &local->mtx rcu_read_lock rcu_node_0 irq_context: 0 (wq_completion)phy12 (work_completion)(&(&local->scan_work)->work) &local->mtx rcu_read_lock &rq->__lock irq_context: 0 (wq_completion)phy12 (work_completion)(&(&local->scan_work)->work) &local->mtx &c->lock irq_context: 0 (wq_completion)phy12 (work_completion)(&(&local->scan_work)->work) &local->mtx rcu_state.exp_mutex rcu_node_0 irq_context: 0 (wq_completion)phy12 (work_completion)(&(&local->scan_work)->work) &local->mtx rcu_state.exp_mutex &obj_hash[i].lock irq_context: 0 (wq_completion)phy12 (work_completion)(&(&local->scan_work)->work) &local->mtx rcu_state.exp_mutex rcu_read_lock &pool->lock irq_context: 0 (wq_completion)phy12 (work_completion)(&(&local->scan_work)->work) &local->mtx rcu_state.exp_mutex rcu_read_lock &pool->lock &obj_hash[i].lock irq_context: 0 (wq_completion)phy12 (work_completion)(&(&local->scan_work)->work) &local->mtx rcu_state.exp_mutex rcu_read_lock &pool->lock &p->pi_lock irq_context: 0 (wq_completion)phy12 (work_completion)(&(&local->scan_work)->work) &local->mtx rcu_state.exp_mutex rcu_read_lock &pool->lock &p->pi_lock &rq->__lock irq_context: 0 (wq_completion)phy12 (work_completion)(&(&local->scan_work)->work) &local->mtx rcu_state.exp_mutex rcu_read_lock &pool->lock &p->pi_lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)phy12 (work_completion)(&(&local->scan_work)->work) &local->mtx rcu_state.exp_mutex &rq->__lock irq_context: 0 (wq_completion)phy12 (work_completion)(&(&local->scan_work)->work) &local->mtx rcu_state.exp_mutex &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)phy12 (work_completion)(&(&local->scan_work)->work) &local->mtx rcu_state.exp_mutex &rnp->exp_wq[3] irq_context: 0 (wq_completion)phy12 (work_completion)(&(&local->scan_work)->work) &local->mtx rcu_state.exp_mutex.wait_lock irq_context: 0 (wq_completion)phy12 (work_completion)(&(&local->scan_work)->work) &local->mtx &p->pi_lock irq_context: 0 (wq_completion)phy12 (work_completion)(&(&local->scan_work)->work) &local->mtx &p->pi_lock &rq->__lock irq_context: 0 (wq_completion)phy12 (work_completion)(&(&local->scan_work)->work) &local->mtx &p->pi_lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)phy12 (work_completion)(&(&local->scan_work)->work) &local->mtx &local->filter_lock irq_context: 0 (wq_completion)phy12 (work_completion)(&(&local->scan_work)->work) &local->mtx &local->queue_stop_reason_lock irq_context: 0 (wq_completion)phy12 (work_completion)(&(&local->scan_work)->work) &local->mtx rcu_read_lock &fq->lock irq_context: 0 (wq_completion)phy12 (work_completion)(&(&local->scan_work)->work) &local->mtx &rdev->wiphy_work_lock irq_context: 0 (wq_completion)phy12 (work_completion)(&(&local->scan_work)->work) &local->mtx rcu_read_lock &pool->lock irq_context: 0 (wq_completion)phy12 (work_completion)(&(&local->scan_work)->work) &local->mtx rcu_read_lock &pool->lock &obj_hash[i].lock irq_context: 0 (wq_completion)phy12 (work_completion)(&(&local->scan_work)->work) &local->mtx rcu_read_lock &pool->lock &p->pi_lock irq_context: 0 (wq_completion)phy12 (work_completion)(&(&local->scan_work)->work) &local->mtx rcu_read_lock &pool->lock &p->pi_lock &rq->__lock irq_context: 0 (wq_completion)phy12 (work_completion)(&(&local->scan_work)->work) &local->mtx rcu_read_lock &pool->lock &p->pi_lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)events_unbound (work_completion)(&map->work) map_idr_lock irq_context: 0 (wq_completion)events_unbound (work_completion)(&map->work) map_idr_lock &obj_hash[i].lock irq_context: 0 (wq_completion)events_unbound (work_completion)(&map->work) map_idr_lock pool_lock#2 irq_context: 0 (wq_completion)events_unbound (work_completion)(&map->work) rcu_read_lock &pool->lock/1 irq_context: 0 (wq_completion)events_unbound (work_completion)(&map->work) rcu_read_lock &pool->lock/1 &obj_hash[i].lock irq_context: 0 (wq_completion)events_unbound (work_completion)(&map->work) rcu_read_lock &pool->lock/1 pool_lock#2 irq_context: 0 (wq_completion)events_unbound (work_completion)(&map->work) rcu_read_lock &pool->lock/1 &p->pi_lock irq_context: 0 &sb->s_type->i_mutex_key#10 sk_lock-AF_CAN rcu_state.exp_mutex.wait_lock irq_context: 0 &sb->s_type->i_mutex_key#10 sk_lock-AF_CAN &p->pi_lock irq_context: 0 &sb->s_type->i_mutex_key#10 sk_lock-AF_CAN &p->pi_lock &rq->__lock irq_context: 0 &sb->s_type->i_mutex_key#10 sk_lock-AF_CAN &p->pi_lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 sk_lock-AF_NETLINK slock-AF_NETLINK &sk->sk_lock.wq irq_context: 0 slock-AF_NETLINK &sk->sk_lock.wq irq_context: 0 slock-AF_NETLINK &sk->sk_lock.wq &p->pi_lock irq_context: 0 slock-AF_NETLINK &sk->sk_lock.wq &p->pi_lock &rq->__lock irq_context: 0 slock-AF_NETLINK &sk->sk_lock.wq &p->pi_lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 sb_writers#4 &sb->s_type->i_mutex_key#8 &sbi->s_writepages_rwsem lock#4 irq_context: 0 sb_writers#4 &sb->s_type->i_mutex_key#8 &sbi->s_writepages_rwsem lock#4 &lruvec->lru_lock irq_context: 0 sb_writers#4 &sb->s_type->i_mutex_key#8 &sbi->s_writepages_rwsem lock#5 irq_context: 0 sb_writers#4 &sb->s_type->i_mutex_key#8 &sbi->s_writepages_rwsem &journal->j_state_lock irq_context: 0 sb_writers#4 &sb->s_type->i_mutex_key#8 &sbi->s_writepages_rwsem jbd2_handle irq_context: 0 sb_writers#4 &sb->s_type->i_mutex_key#8 &sbi->s_writepages_rwsem jbd2_handle lock#4 irq_context: 0 sb_writers#4 &sb->s_type->i_mutex_key#8 &sbi->s_writepages_rwsem jbd2_handle lock#5 irq_context: 0 sb_writers#4 &sb->s_type->i_mutex_key#8 &sbi->s_writepages_rwsem jbd2_handle mmu_notifier_invalidate_range_start irq_context: 0 sb_writers#4 &sb->s_type->i_mutex_key#8 &sbi->s_writepages_rwsem jbd2_handle pool_lock#2 irq_context: 0 sb_writers#4 &sb->s_type->i_mutex_key#8 &sbi->s_writepages_rwsem jbd2_handle &ei->i_es_lock irq_context: 0 sb_writers#4 &sb->s_type->i_mutex_key#8 &sbi->s_writepages_rwsem jbd2_handle &ei->i_es_lock key#2 irq_context: 0 (wq_completion)bat_events (work_completion)(&(&forw_packet_aggr->delayed_work)->work) &rq->__lock &cfs_rq->removed.lock irq_context: 0 sb_writers#4 &sb->s_type->i_mutex_key#8 &sbi->s_writepages_rwsem jbd2_handle &rq->__lock irq_context: 0 sb_writers#4 &sb->s_type->i_mutex_key#8 &sbi->s_writepages_rwsem jbd2_handle &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 sb_writers#4 &sb->s_type->i_mutex_key#8 &sbi->s_writepages_rwsem jbd2_handle &ei->i_data_sem irq_context: 0 sb_writers#4 &sb->s_type->i_mutex_key#8 &sbi->s_writepages_rwsem jbd2_handle &ei->i_data_sem mmu_notifier_invalidate_range_start irq_context: 0 sb_writers#4 &sb->s_type->i_mutex_key#8 &sbi->s_writepages_rwsem jbd2_handle &ei->i_data_sem &c->lock irq_context: 0 sb_writers#4 &sb->s_type->i_mutex_key#8 &sbi->s_writepages_rwsem jbd2_handle &ei->i_data_sem &rq->__lock irq_context: 0 sb_writers#4 &sb->s_type->i_mutex_key#8 &sbi->s_writepages_rwsem jbd2_handle &ei->i_data_sem &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 sb_writers#4 &sb->s_type->i_mutex_key#8 &sbi->s_writepages_rwsem jbd2_handle &ei->i_data_sem pool_lock#2 irq_context: 0 sb_writers#4 &sb->s_type->i_mutex_key#8 &sbi->s_writepages_rwsem jbd2_handle &ei->i_data_sem &ei->i_prealloc_lock irq_context: 0 sb_writers#4 &sb->s_type->i_mutex_key#8 &sbi->s_writepages_rwsem jbd2_handle &ei->i_data_sem &sbi->s_md_lock irq_context: 0 sb_writers#4 &sb->s_type->i_mutex_key#8 &sbi->s_writepages_rwsem jbd2_handle &ei->i_data_sem &bgl->locks[i].lock &sbi->s_md_lock irq_context: 0 sb_writers#4 &sb->s_type->i_mutex_key#8 &sbi->s_writepages_rwsem jbd2_handle &ei->i_data_sem &bgl->locks[i].lock &ei->i_prealloc_lock irq_context: 0 sb_writers#4 &sb->s_type->i_mutex_key#8 &sbi->s_writepages_rwsem jbd2_handle &ei->i_data_sem key#3 irq_context: 0 sb_writers#4 &sb->s_type->i_mutex_key#8 &sbi->s_writepages_rwsem jbd2_handle &ei->i_data_sem &pa->pa_lock#2 irq_context: 0 sb_writers#4 &sb->s_type->i_mutex_key#8 &sbi->s_writepages_rwsem jbd2_handle &ei->i_data_sem &obj_hash[i].lock irq_context: 0 sb_writers#4 &sb->s_type->i_mutex_key#8 &sbi->s_writepages_rwsem jbd2_handle &ei->i_data_sem &ei->i_raw_lock irq_context: 0 sb_writers#4 &sb->s_type->i_mutex_key#8 &sbi->s_writepages_rwsem jbd2_handle &ei->i_data_sem &(ei->i_block_reservation_lock) irq_context: 0 sb_writers#4 &sb->s_type->i_mutex_key#8 &sbi->s_writepages_rwsem jbd2_handle &ei->i_data_sem &sb->s_type->i_lock_key#22 irq_context: 0 sb_writers#4 &sb->s_type->i_mutex_key#8 &sbi->s_writepages_rwsem jbd2_handle &ei->i_data_sem &ei->i_es_lock irq_context: 0 sb_writers#4 &sb->s_type->i_mutex_key#8 &sbi->s_writepages_rwsem jbd2_handle &ei->i_data_sem &ei->i_es_lock irq_context: 0 sb_writers#4 &sb->s_type->i_mutex_key#8 &sbi->s_writepages_rwsem jbd2_handle &ei->i_data_sem &ei->i_es_lock pool_lock#2 irq_context: 0 sb_writers#4 &sb->s_type->i_mutex_key#8 &sbi->s_writepages_rwsem jbd2_handle rcu_read_lock &memcg->move_lock irq_context: 0 sb_writers#4 &sb->s_type->i_mutex_key#8 &sbi->s_writepages_rwsem jbd2_handle rcu_read_lock rcu_node_0 irq_context: 0 sb_writers#4 &sb->s_type->i_mutex_key#8 &sbi->s_writepages_rwsem jbd2_handle rcu_read_lock &rq->__lock irq_context: 0 sb_writers#4 &sb->s_type->i_mutex_key#8 &sbi->s_writepages_rwsem jbd2_handle rcu_read_lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 sb_writers#4 &sb->s_type->i_mutex_key#8 &sbi->s_writepages_rwsem jbd2_handle rcu_read_lock &xa->xa_lock#9 irq_context: 0 sb_writers#4 &sb->s_type->i_mutex_key#8 &sbi->s_writepages_rwsem jbd2_handle rcu_read_lock &xa->xa_lock#9 &s->s_inode_wblist_lock irq_context: 0 sb_writers#4 &sb->s_type->i_mutex_key#8 &sbi->s_writepages_rwsem jbd2_handle &c->lock irq_context: 0 sb_writers#4 &sb->s_type->i_mutex_key#8 &sbi->s_writepages_rwsem jbd2_handle rcu_read_lock key#10 irq_context: 0 sb_writers#4 &sb->s_type->i_mutex_key#8 &sbi->s_writepages_rwsem jbd2_handle rcu_read_lock &xa->xa_lock#9 key#10 irq_context: 0 sb_writers#4 &sb->s_type->i_mutex_key#8 &sbi->s_writepages_rwsem jbd2_handle &ei->i_raw_lock irq_context: 0 sb_writers#4 &sb->s_type->i_mutex_key#8 &sbi->s_writepages_rwsem jbd2_handle &journal->j_wait_updates irq_context: 0 sb_writers#4 &sb->s_type->i_mutex_key#8 &sbi->s_writepages_rwsem jbd2_handle &journal->j_wait_updates &p->pi_lock irq_context: 0 sb_writers#4 &sb->s_type->i_mutex_key#8 &sbi->s_writepages_rwsem jbd2_handle &journal->j_wait_updates &p->pi_lock &rq->__lock irq_context: 0 sb_writers#4 &sb->s_type->i_mutex_key#8 &sbi->s_writepages_rwsem jbd2_handle &journal->j_wait_updates &p->pi_lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 sb_writers#4 &sb->s_type->i_mutex_key#8 &sbi->s_writepages_rwsem tk_core.seq.seqcount irq_context: 0 sb_writers#4 &sb->s_type->i_mutex_key#8 &sbi->s_writepages_rwsem &dd->lock irq_context: 0 (wq_completion)kblockd (work_completion)(&(&q->requeue_work)->work) rcu_read_lock pool_lock#2 irq_context: 0 sb_writers#4 &sb->s_type->i_mutex_key#8 &sbi->s_writepages_rwsem jbd2_handle &ei->i_data_sem &ei->i_prealloc_lock &pa->pa_lock#2 irq_context: 0 sb_writers#4 &sb->s_type->i_mutex_key#8 &sbi->s_writepages_rwsem jbd2_handle &ei->i_data_sem &mapping->private_lock irq_context: 0 sb_writers#4 &sb->s_type->i_mutex_key#8 &sbi->s_writepages_rwsem jbd2_handle &ei->i_data_sem &ret->b_state_lock irq_context: 0 sb_writers#4 &sb->s_type->i_mutex_key#8 &sbi->s_writepages_rwsem jbd2_handle &ei->i_data_sem &ret->b_state_lock &journal->j_list_lock irq_context: 0 sb_writers#4 &sb->s_type->i_mutex_key#8 &sbi->s_writepages_rwsem rcu_read_lock &dd->lock irq_context: 0 sb_writers#4 &sb->s_type->i_mutex_key#8 &sbi->s_writepages_rwsem rcu_read_lock &obj_hash[i].lock irq_context: 0 sb_writers#4 &sb->s_type->i_mutex_key#8 &sbi->s_writepages_rwsem rcu_read_lock rcu_node_0 irq_context: 0 sb_writers#4 &sb->s_type->i_mutex_key#8 &sbi->s_writepages_rwsem rcu_read_lock &rq->__lock irq_context: 0 sb_writers#4 &sb->s_type->i_mutex_key#8 &sbi->s_writepages_rwsem rcu_read_lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 sb_writers#4 &sb->s_type->i_mutex_key#8 &sbi->s_writepages_rwsem rcu_read_lock pool_lock#2 irq_context: 0 sb_writers#4 &sb->s_type->i_mutex_key#8 &sbi->s_writepages_rwsem rcu_read_lock tk_core.seq.seqcount irq_context: 0 sb_writers#4 &sb->s_type->i_mutex_key#8 &sbi->s_writepages_rwsem rcu_read_lock &virtscsi_vq->vq_lock irq_context: 0 sb_writers#4 &sb->s_type->i_mutex_key#8 &sbi->s_writepages_rwsem jbd2_handle &ei->i_data_sem &ei->i_es_lock &c->lock irq_context: 0 sb_writers#4 &sb->s_type->i_mutex_key#8 &sbi->s_writepages_rwsem rcu_read_lock &c->lock irq_context: 0 sb_writers#4 &sb->s_type->i_mutex_key#8 &sbi->s_writepages_rwsem irq_context: 0 sb_writers#4 &sb->s_type->i_mutex_key#8 &sbi->s_writepages_rwsem &xa->xa_lock#9 irq_context: 0 sb_writers#4 &sb->s_type->i_mutex_key#8 &sbi->s_writepages_rwsem mmu_notifier_invalidate_range_start irq_context: 0 sb_writers#4 &sb->s_type->i_mutex_key#8 &sbi->s_writepages_rwsem &obj_hash[i].lock irq_context: 0 sb_writers#4 &sb->s_type->i_mutex_key#8 &sbi->s_writepages_rwsem jbd2_handle &ei->i_data_sem &(ei->i_block_reservation_lock) key#14 irq_context: 0 sb_writers#4 &sb->s_type->i_mutex_key#8 &sbi->s_writepages_rwsem jbd2_handle rcu_node_0 irq_context: 0 sb_writers#4 &sb->s_type->i_mutex_key#8 &sbi->s_writepages_rwsem jbd2_handle &ei->i_data_sem &____s->seqcount#2 irq_context: 0 sb_writers#4 &sb->s_type->i_mutex_key#8 &sbi->s_writepages_rwsem jbd2_handle &ei->i_data_sem &n->list_lock irq_context: 0 sb_writers#4 &sb->s_type->i_mutex_key#8 &sbi->s_writepages_rwsem jbd2_handle &ei->i_data_sem &n->list_lock &c->lock irq_context: 0 sb_writers#4 &sb->s_type->i_mutex_key#8 &sbi->s_writepages_rwsem jbd2_handle &ei->i_data_sem &ei->i_es_lock key#2 irq_context: 0 sb_writers#4 &sb->s_type->i_mutex_key#8 &sbi->s_writepages_rwsem jbd2_handle &n->list_lock irq_context: 0 sb_writers#4 &sb->s_type->i_mutex_key#8 &sbi->s_writepages_rwsem jbd2_handle &n->list_lock &c->lock irq_context: 0 sb_writers#4 &sb->s_type->i_mutex_key#8 &sbi->s_writepages_rwsem &c->lock irq_context: 0 sb_writers#4 &sb->s_type->i_mutex_key#8 &sbi->s_writepages_rwsem &____s->seqcount#2 irq_context: 0 sb_writers#4 &sb->s_type->i_mutex_key#8 &sbi->s_writepages_rwsem &____s->seqcount irq_context: 0 sb_writers#4 &sb->s_type->i_mutex_key#8 &sbi->s_writepages_rwsem pool_lock#2 irq_context: 0 sb_writers#4 &sb->s_type->i_mutex_key#8 &sbi->s_writepages_rwsem &____s->seqcount#2 irq_context: 0 sb_writers#4 &sb->s_type->i_mutex_key#8 &sbi->s_writepages_rwsem &____s->seqcount irq_context: 0 sb_writers#4 &sb->s_type->i_mutex_key#8 &sbi->s_writepages_rwsem rcu_read_lock &____s->seqcount#2 irq_context: 0 sb_writers#4 &sb->s_type->i_mutex_key#8 &sbi->s_writepages_rwsem rcu_read_lock &____s->seqcount irq_context: 0 sb_writers#4 &sb->s_type->i_mutex_key#8 &sbi->s_writepages_rwsem jbd2_handle &____s->seqcount#2 irq_context: 0 sb_writers#4 &sb->s_type->i_mutex_key#8 &sbi->s_writepages_rwsem jbd2_handle &____s->seqcount irq_context: 0 sb_writers#4 &sb->s_type->i_mutex_key#8 &sbi->s_writepages_rwsem jbd2_handle lock#4 &lruvec->lru_lock irq_context: 0 (wq_completion)ext4-rsv-conversion (work_completion)(&ei->i_rsv_conversion_work) jbd2_handle &ei->i_data_sem rcu_read_lock pool_lock#2 irq_context: 0 sb_writers#4 &sb->s_type->i_mutex_key#8 &wb->list_lock irq_context: 0 sb_writers#4 &sb->s_type->i_mutex_key#8 &sbi->s_writepages_rwsem jbd2_handle rcu_read_lock &cfs_rq->removed.lock irq_context: 0 sb_writers#4 &sb->s_type->i_mutex_key#8 &sbi->s_writepages_rwsem jbd2_handle rcu_read_lock &obj_hash[i].lock irq_context: 0 &mm->mmap_lock sb_pagefaults mapping.invalidate_lock rcu_read_lock &rq->__lock &cfs_rq->removed.lock irq_context: 0 &sbi->s_writepages_rwsem jbd2_handle &pcp->lock &zone->lock irq_context: 0 sb_writers#4 &sb->s_type->i_mutex_key#8 fs_reclaim &rq->__lock irq_context: 0 sb_writers#4 &sb->s_type->i_mutex_key#8 fs_reclaim &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 slock-AF_NETLINK &sk->sk_lock.wq &p->pi_lock &cfs_rq->removed.lock irq_context: 0 nfnl_subsys_nftables &lock->wait_lock irq_context: 0 &nft_net->commit_mutex &lock->wait_lock irq_context: 0 (wq_completion)phy12 (work_completion)(&(&local->scan_work)->work) &local->mtx rcu_read_lock hwsim_radio_lock &c->lock irq_context: 0 (wq_completion)phy12 (work_completion)(&(&local->scan_work)->work) &local->mtx rcu_state.exp_mutex &rnp->exp_wq[1] irq_context: 0 (wq_completion)events (work_completion)(&rdev->wiphy_work) &rdev->wiphy.mtx &wdev->mtx crngs.lock irq_context: 0 &sb->s_type->i_mutex_key#19 &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)wg-crypt-wg0#2 (work_completion)(&peer->transmit_packet_work) &peer->endpoint_lock rcu_read_lock_bh rcu_read_lock rcu_read_lock_bh rcu_read_lock batched_entropy_u8.lock irq_context: 0 (wq_completion)wg-crypt-wg0#2 (work_completion)(&peer->transmit_packet_work) &peer->endpoint_lock rcu_read_lock_bh rcu_read_lock rcu_read_lock_bh rcu_read_lock kfence_freelist_lock irq_context: 0 (wq_completion)wg-crypt-wg0#2 (work_completion)(&peer->transmit_packet_work) &peer->endpoint_lock rcu_read_lock_bh rcu_read_lock rcu_read_lock_bh rcu_read_lock &meta->lock irq_context: 0 (wq_completion)writeback (work_completion)(&(&wb->dwork)->work) &sbi->s_writepages_rwsem jbd2_handle &ei->i_data_sem &ret->b_state_lock irq_context: 0 (wq_completion)writeback (work_completion)(&(&wb->dwork)->work) &sbi->s_writepages_rwsem jbd2_handle &ei->i_data_sem &ret->b_state_lock &journal->j_list_lock irq_context: 0 (wq_completion)writeback (work_completion)(&(&wb->dwork)->work) &sbi->s_writepages_rwsem jbd2_handle &folio_wait_table[i] irq_context: 0 (wq_completion)writeback (work_completion)(&(&wb->dwork)->work) &sbi->s_writepages_rwsem jbd2_handle &folio_wait_table[i] &p->pi_lock irq_context: 0 (wq_completion)writeback (work_completion)(&(&wb->dwork)->work) &sbi->s_writepages_rwsem jbd2_handle &folio_wait_table[i] &p->pi_lock &rq->__lock irq_context: 0 (wq_completion)writeback (work_completion)(&(&wb->dwork)->work) &sbi->s_writepages_rwsem jbd2_handle &folio_wait_table[i] &p->pi_lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 sb_writers#4 &type->i_mutex_dir_key#3/1 sb_internal jbd2_handle &ei->i_data_sem &wb->list_lock irq_context: 0 sb_writers#4 &type->i_mutex_dir_key#3/1 sb_internal jbd2_handle &ei->i_data_sem &wb->list_lock &sb->s_type->i_lock_key#22 irq_context: 0 sb_writers#4 &type->i_mutex_dir_key#3/1 remove_cache_srcu &cfs_rq->removed.lock irq_context: 0 sb_writers#4 &type->i_mutex_dir_key#3/1 sb_internal jbd2_handle &ei->i_data_sem rcu_read_lock &cfs_rq->removed.lock irq_context: 0 sb_writers#4 &type->i_mutex_dir_key#3/1 sb_internal jbd2_handle &ei->i_data_sem rcu_read_lock &obj_hash[i].lock irq_context: 0 sb_writers#4 &type->i_mutex_dir_key#3/1 &type->i_mutex_dir_key#3 &journal->j_state_lock &journal->j_wait_transaction_locked irq_context: 0 sb_writers#4 &type->i_mutex_dir_key#3/1 &type->i_mutex_dir_key#3 jbd2_handle irq_context: 0 (wq_completion)writeback (work_completion)(&(&wb->dwork)->work) &sbi->s_writepages_rwsem jbd2_handle &ei->i_data_sem &n->list_lock irq_context: 0 (wq_completion)writeback (work_completion)(&(&wb->dwork)->work) &sbi->s_writepages_rwsem jbd2_handle &ei->i_data_sem &n->list_lock &c->lock irq_context: 0 (wq_completion)writeback (work_completion)(&(&wb->dwork)->work) &sbi->s_writepages_rwsem jbd2_handle &mapping->i_mmap_rwsem &rq->__lock irq_context: 0 (wq_completion)writeback (work_completion)(&(&wb->dwork)->work) &sbi->s_writepages_rwsem jbd2_handle &mapping->i_mmap_rwsem &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)writeback (work_completion)(&(&wb->dwork)->work) &sbi->s_writepages_rwsem jbd2_handle rcu_read_lock &cfs_rq->removed.lock irq_context: 0 (wq_completion)writeback (work_completion)(&(&wb->dwork)->work) &sbi->s_writepages_rwsem jbd2_handle rcu_read_lock &obj_hash[i].lock irq_context: 0 (wq_completion)writeback (work_completion)(&(&wb->dwork)->work) &sbi->s_writepages_rwsem jbd2_handle &journal->j_wait_updates &p->pi_lock irq_context: 0 (wq_completion)writeback (work_completion)(&(&wb->dwork)->work) &sbi->s_writepages_rwsem jbd2_handle &journal->j_wait_updates &p->pi_lock &rq->__lock irq_context: 0 (wq_completion)writeback (work_completion)(&(&wb->dwork)->work) &sbi->s_writepages_rwsem jbd2_handle &journal->j_wait_updates &p->pi_lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)writeback (work_completion)(&(&wb->dwork)->work) &sbi->s_writepages_rwsem jbd2_handle &ei->i_data_sem &lg->lg_mutex irq_context: 0 (wq_completion)writeback (work_completion)(&(&wb->dwork)->work) &sbi->s_writepages_rwsem jbd2_handle &ei->i_data_sem &lg->lg_mutex &ei->i_prealloc_lock irq_context: 0 (wq_completion)writeback (work_completion)(&(&wb->dwork)->work) &sbi->s_writepages_rwsem jbd2_handle &ei->i_data_sem &lg->lg_mutex mmu_notifier_invalidate_range_start irq_context: 0 (wq_completion)writeback (work_completion)(&(&wb->dwork)->work) &sbi->s_writepages_rwsem jbd2_handle &ei->i_data_sem &lg->lg_mutex &ret->b_state_lock irq_context: 0 (wq_completion)writeback (work_completion)(&(&wb->dwork)->work) &sbi->s_writepages_rwsem jbd2_handle &ei->i_data_sem &lg->lg_mutex &ret->b_state_lock &journal->j_list_lock irq_context: 0 (wq_completion)writeback (work_completion)(&(&wb->dwork)->work) &sbi->s_writepages_rwsem jbd2_handle &ei->i_data_sem &lg->lg_mutex &pa->pa_lock irq_context: 0 (wq_completion)writeback (work_completion)(&(&wb->dwork)->work) &sbi->s_writepages_rwsem jbd2_handle &ei->i_data_sem &lg->lg_mutex &lg->lg_prealloc_lock irq_context: 0 (wq_completion)writeback (work_completion)(&(&wb->dwork)->work) &sbi->s_writepages_rwsem jbd2_handle &ei->i_data_sem &lg->lg_mutex &rq->__lock irq_context: 0 (wq_completion)writeback (work_completion)(&(&wb->dwork)->work) &sbi->s_writepages_rwsem jbd2_handle &ei->i_data_sem &lg->lg_mutex &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 sb_writers#4 &sb->s_type->i_mutex_key#8 jbd2_handle &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)writeback (work_completion)(&(&wb->dwork)->work) &sbi->s_writepages_rwsem jbd2_handle &p->pi_lock &cfs_rq->removed.lock irq_context: 0 (wq_completion)writeback (work_completion)(&(&wb->dwork)->work) &sbi->s_writepages_rwsem jbd2_handle &ei->i_data_sem &sem->wait_lock irq_context: 0 (wq_completion)writeback (work_completion)(&(&wb->dwork)->work) &sbi->s_writepages_rwsem jbd2_handle &ei->i_data_sem &dd->lock irq_context: 0 (wq_completion)writeback (work_completion)(&(&wb->dwork)->work) &sbi->s_writepages_rwsem jbd2_handle &ei->i_data_sem rcu_read_lock &pool->lock irq_context: 0 (wq_completion)writeback (work_completion)(&(&wb->dwork)->work) &sbi->s_writepages_rwsem jbd2_handle &ei->i_data_sem rcu_read_lock &pool->lock &obj_hash[i].lock irq_context: 0 (wq_completion)writeback (work_completion)(&(&wb->dwork)->work) &sbi->s_writepages_rwsem jbd2_handle &ei->i_data_sem rcu_read_lock &pool->lock pool_lock#2 irq_context: 0 (wq_completion)writeback (work_completion)(&(&wb->dwork)->work) &sbi->s_writepages_rwsem jbd2_handle &ei->i_data_sem rcu_read_lock &pool->lock &p->pi_lock irq_context: 0 (wq_completion)writeback (work_completion)(&(&wb->dwork)->work) &sbi->s_writepages_rwsem jbd2_handle &ei->i_data_sem rcu_read_lock &pool->lock &p->pi_lock &rq->__lock irq_context: 0 (wq_completion)writeback (work_completion)(&(&wb->dwork)->work) &sbi->s_writepages_rwsem jbd2_handle &ei->i_data_sem rcu_read_lock &pool->lock &p->pi_lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 &mm->mmap_lock sb_writers#4 jbd2_handle &rq->__lock &cfs_rq->removed.lock irq_context: 0 sb_writers#4 &type->i_mutex_dir_key#3/1 sb_internal jbd2_handle &p->pi_lock &cfs_rq->removed.lock irq_context: 0 sb_writers#4 sb_internal jbd2_handle &p->pi_lock &cfs_rq->removed.lock irq_context: 0 &sb->s_type->i_mutex_key#10 rtnl_mutex sk_lock-AF_CAN &rq->__lock irq_context: 0 &sb->s_type->i_mutex_key#10 rtnl_mutex sk_lock-AF_CAN &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 sb_writers#4 &type->i_mutex_dir_key#3/1 jbd2_handle &ei->i_data_sem pgd_lock irq_context: 0 sb_writers#4 &type->i_mutex_dir_key#3/1 jbd2_handle &ei->i_data_sem stock_lock irq_context: 0 sb_writers#4 &type->i_mutex_dir_key#3/1 jbd2_handle &ei->i_data_sem rcu_read_lock pool_lock#2 irq_context: 0 sb_writers#4 &type->i_mutex_dir_key#3/1 jbd2_handle &ei->i_data_sem key irq_context: 0 sb_writers#4 &type->i_mutex_dir_key#3/1 jbd2_handle &ei->i_data_sem pcpu_lock irq_context: 0 sb_writers#4 &type->i_mutex_dir_key#3/1 jbd2_handle &ei->i_data_sem percpu_counters_lock irq_context: 0 sb_writers#4 &type->i_mutex_dir_key#3/1 jbd2_handle &ei->i_data_sem pcpu_lock stock_lock irq_context: 0 sb_writers#4 mapping.invalidate_lock &xa->xa_lock#9 stock_lock rcu_read_lock per_cpu_ptr(&cgroup_rstat_cpu_lock, cpu) irq_context: 0 &nft_net->commit_mutex batched_entropy_u8.lock irq_context: 0 &nft_net->commit_mutex kfence_freelist_lock irq_context: 0 nfnl_subsys_nftables &nft_net->commit_mutex rcu_node_0 irq_context: 0 pernet_ops_rwsem devices_rwsem rdma_nets_rwsem &device->compat_devs_mutex &root->kernfs_rwsem &root->kernfs_iattr_rwsem &rq->__lock &cfs_rq->removed.lock irq_context: 0 &type->s_umount_key#23/1 fs_reclaim &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 &nft_net->commit_mutex &rnp->exp_wq[0] irq_context: 0 pernet_ops_rwsem rtnl_mutex fs_reclaim mmu_notifier_invalidate_range_start &cfs_rq->removed.lock irq_context: 0 pernet_ops_rwsem rtnl_mutex fs_reclaim mmu_notifier_invalidate_range_start &obj_hash[i].lock irq_context: softirq (&app->join_timer) &app->lock fill_pool_map-wait-type-override pool_lock#2 irq_context: softirq (&app->join_timer) &app->lock fill_pool_map-wait-type-override pool_lock irq_context: 0 rcu_read_lock fill_pool_map-wait-type-override pool_lock#2 irq_context: 0 rcu_read_lock fill_pool_map-wait-type-override pool_lock irq_context: 0 sb_writers#4 &sb->s_type->i_mutex_key#8 rcu_read_lock &rcu_state.gp_wq &p->pi_lock &rq->__lock &cfs_rq->removed.lock irq_context: 0 sb_writers#4 &type->i_mutex_dir_key#3/1 sb_internal jbd2_handle &ei->i_data_sem bit_wait_table + i irq_context: 0 sb_writers#4 sb_writers#4 &n->list_lock &c->lock irq_context: 0 (wq_completion)wg-kex-wg2#3 (work_completion)(&peer->transmit_handshake_work) &____s->seqcount#2 irq_context: 0 (wq_completion)wg-kex-wg2#3 (work_completion)(&peer->transmit_handshake_work) &____s->seqcount irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem k-sk_lock-AF_INET &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: softirq &(&br->gc_work)->timer rcu_read_lock &pool->lock &p->pi_lock &rq->__lock &cfs_rq->removed.lock irq_context: 0 (wq_completion)events free_ipc_work rcu_state.exp_mutex &obj_hash[i].lock pool_lock irq_context: 0 &sb->s_type->i_mutex_key#8 &sbi->s_writepages_rwsem jbd2_handle &ei->i_data_sem &obj_hash[i].lock pool_lock irq_context: 0 &sbi->s_writepages_rwsem jbd2_handle &ei->i_data_sem &ei->i_es_lock &n->list_lock irq_context: 0 &sbi->s_writepages_rwsem jbd2_handle &ei->i_data_sem &ei->i_es_lock &n->list_lock &c->lock irq_context: 0 &mm->mmap_lock sb_pagefaults &rq->__lock &cfs_rq->removed.lock irq_context: 0 (wq_completion)events free_ipc_work rcu_state.exp_mutex &rq->__lock &cfs_rq->removed.lock irq_context: softirq (&n->timer) k-slock-AF_INET6 rcu_read_lock rcu_read_lock rcu_read_lock_bh &sch->q.lock batched_entropy_u8.lock irq_context: softirq (&n->timer) k-slock-AF_INET6 rcu_read_lock rcu_read_lock rcu_read_lock_bh &sch->q.lock kfence_freelist_lock irq_context: 0 pernet_ops_rwsem remove_cache_srcu &____s->seqcount irq_context: 0 ppp_mutex rtnl_mutex fill_pool_map-wait-type-override &rq->__lock irq_context: 0 ppp_mutex rtnl_mutex fill_pool_map-wait-type-override &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 &nft_net->commit_mutex &rnp->exp_wq[3] irq_context: 0 (wq_completion)wg-crypt-wg0#5 (work_completion)(&({ do { const void *__vpp_verify = (typeof((worker) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __asm__ ("" : "=r"(__ptr) : "0"((typeof(*((worker))) *)((worker)))); (typeof((typeof(*((worker))) *)((worker)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); })->work) rcu_read_lock rcu_node_0 irq_context: 0 (wq_completion)wg-crypt-wg0#5 (work_completion)(&({ do { const void *__vpp_verify = (typeof((worker) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __asm__ ("" : "=r"(__ptr) : "0"((typeof(*((worker))) *)((worker)))); (typeof((typeof(*((worker))) *)((worker)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); })->work) rcu_read_lock &rq->__lock irq_context: 0 ppp_mutex rtnl_mutex uevent_sock_mutex.wait_lock irq_context: 0 ppp_mutex rtnl_mutex rcu_node_0 irq_context: 0 ppp_mutex rtnl_mutex remove_cache_srcu pool_lock#2 irq_context: 0 ppp_mutex rtnl_mutex remove_cache_srcu rcu_read_lock rcu_node_0 irq_context: 0 ppp_mutex rtnl_mutex remove_cache_srcu rcu_read_lock &rq->__lock irq_context: 0 ppp_mutex rtnl_mutex remove_cache_srcu rcu_read_lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 sb_writers#4 &type->i_mutex_dir_key#3/1 &dentry->d_lock &lru->node[i].lock rcu_read_lock &p->pi_lock &rq->__lock irq_context: 0 sb_writers#4 &type->i_mutex_dir_key#3/1 &dentry->d_lock &lru->node[i].lock rcu_read_lock &p->pi_lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 sk_lock-AF_VSOCK &____s->seqcount#2 irq_context: 0 sk_lock-AF_VSOCK &____s->seqcount irq_context: 0 sk_lock-AF_VSOCK &obj_hash[i].lock irq_context: 0 sk_lock-AF_SMC k-sk_lock-AF_INET &c->lock irq_context: 0 &sb->s_type->i_mutex_key#10 sk_lock-AF_INET6 slock-AF_INET6 fill_pool_map-wait-type-override &n->list_lock irq_context: 0 &sb->s_type->i_mutex_key#10 sk_lock-AF_INET6 slock-AF_INET6 fill_pool_map-wait-type-override &n->list_lock &c->lock irq_context: 0 tasklist_lock &sighand->siglock &(&sig->stats_lock)->lock &____s->seqcount#5 fill_pool_map-wait-type-override batched_entropy_u8.lock irq_context: 0 tasklist_lock &sighand->siglock &(&sig->stats_lock)->lock &____s->seqcount#5 fill_pool_map-wait-type-override kfence_freelist_lock irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem fs_reclaim mmu_notifier_invalidate_range_start &rq->__lock irq_context: 0 sk_lock-AF_VSOCK &rq->__lock irq_context: 0 sk_lock-AF_VSOCK &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 &mm->mmap_lock &anon_vma->rwsem &obj_hash[i].lock pool_lock irq_context: 0 sb_writers#4 &type->i_mutex_dir_key#3/1 sb_internal jbd2_handle &rcu_state.expedited_wq irq_context: 0 sb_writers#4 &type->i_mutex_dir_key#3/1 sb_internal jbd2_handle &rcu_state.expedited_wq &p->pi_lock irq_context: 0 sb_writers#4 &type->i_mutex_dir_key#3/1 sb_internal jbd2_handle &rcu_state.expedited_wq &p->pi_lock &rq->__lock irq_context: 0 sb_writers#4 &type->i_mutex_dir_key#3/1 sb_internal jbd2_handle &rcu_state.expedited_wq &p->pi_lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)wg-kex-wg0#7 (work_completion)(&peer->transmit_handshake_work) &peer->endpoint_lock rcu_read_lock_bh rcu_read_lock &____s->seqcount#2 irq_context: 0 (wq_completion)wg-kex-wg0#7 (work_completion)(&peer->transmit_handshake_work) &peer->endpoint_lock rcu_read_lock_bh rcu_read_lock &____s->seqcount irq_context: 0 (wq_completion)inet_frag_wq (work_completion)(&fqdir->destroy_work) &ht->mutex &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)events (work_completion)(&ht->run_work) &ht->mutex remove_cache_srcu &rq->__lock irq_context: 0 (wq_completion)events (work_completion)(&ht->run_work) &ht->mutex remove_cache_srcu &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 sk_lock-AF_INET6 k-sk_lock-AF_INET6/1 &n->list_lock irq_context: 0 sk_lock-AF_INET6 k-sk_lock-AF_INET6/1 &n->list_lock &c->lock irq_context: 0 sk_lock-AF_INET6 k-sk_lock-AF_INET6 rcu_read_lock rcu_read_lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)events_unbound (work_completion)(&sub_info->work) rcu_read_lock rcu_read_lock rcu_node_0 irq_context: 0 (wq_completion)events_unbound (work_completion)(&sub_info->work) rcu_read_lock rcu_read_lock &rq->__lock irq_context: 0 (wq_completion)events_unbound (work_completion)(&sub_info->work) rcu_read_lock rcu_read_lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 &f->f_pos_lock &type->i_mutex_dir_key#3 remove_cache_srcu &rq->__lock &cfs_rq->removed.lock irq_context: 0 sk_lock-AF_INET &tcp_hashinfo.bhash[i].lock &obj_hash[i].lock irq_context: 0 &f->f_pos_lock sb_writers#4 &sb->s_type->i_mutex_key#8 fs_reclaim &rq->__lock &cfs_rq->removed.lock irq_context: 0 (wq_completion)wg-kex-wg0#2 (work_completion)(&({ do { const void *__vpp_verify = (typeof((worker) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __asm__ ("" : "=r"(__ptr) : "0"((typeof(*((worker))) *)((worker)))); (typeof((typeof(*((worker))) *)((worker)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); })->work) &wg->static_identity.lock &handshake->lock &rq->__lock irq_context: 0 (wq_completion)wg-kex-wg0#2 (work_completion)(&({ do { const void *__vpp_verify = (typeof((worker) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __asm__ ("" : "=r"(__ptr) : "0"((typeof(*((worker))) *)((worker)))); (typeof((typeof(*((worker))) *)((worker)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); })->work) &wg->static_identity.lock &handshake->lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)wg-kex-wg1 (work_completion)(&({ do { const void *__vpp_verify = (typeof((worker) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __asm__ ("" : "=r"(__ptr) : "0"((typeof(*((worker))) *)((worker)))); (typeof((typeof(*((worker))) *)((worker)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); })->work) &n->list_lock irq_context: 0 (wq_completion)wg-kex-wg1 (work_completion)(&({ do { const void *__vpp_verify = (typeof((worker) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __asm__ ("" : "=r"(__ptr) : "0"((typeof(*((worker))) *)((worker)))); (typeof((typeof(*((worker))) *)((worker)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); })->work) &n->list_lock &c->lock irq_context: 0 &sb->s_type->i_mutex_key#8 &sbi->s_writepages_rwsem jbd2_handle tk_core.seq.seqcount irq_context: 0 &sb->s_type->i_mutex_key#8 &sbi->s_writepages_rwsem jbd2_handle &base->lock irq_context: 0 &sb->s_type->i_mutex_key#8 &sbi->s_writepages_rwsem jbd2_handle &base->lock &obj_hash[i].lock irq_context: 0 &sb->s_type->i_mutex_key#8 &sbi->s_writepages_rwsem jbd2_handle &dd->lock irq_context: 0 &sb->s_type->i_mutex_key#8 &sbi->s_writepages_rwsem jbd2_handle rcu_read_lock &dd->lock irq_context: 0 &sb->s_type->i_mutex_key#8 &sbi->s_writepages_rwsem jbd2_handle rcu_read_lock &obj_hash[i].lock irq_context: 0 &sb->s_type->i_mutex_key#8 &sbi->s_writepages_rwsem jbd2_handle rcu_read_lock pool_lock#2 irq_context: 0 &sb->s_type->i_mutex_key#8 &sbi->s_writepages_rwsem jbd2_handle rcu_read_lock tk_core.seq.seqcount irq_context: 0 &sb->s_type->i_mutex_key#8 &sbi->s_writepages_rwsem jbd2_handle rcu_read_lock &base->lock irq_context: 0 &sb->s_type->i_mutex_key#8 &sbi->s_writepages_rwsem jbd2_handle rcu_read_lock &base->lock &obj_hash[i].lock irq_context: 0 &sb->s_type->i_mutex_key#8 &sbi->s_writepages_rwsem jbd2_handle rcu_read_lock &virtscsi_vq->vq_lock irq_context: 0 &sb->s_type->i_mutex_key#8 &sbi->s_writepages_rwsem jbd2_handle rcu_read_lock &c->lock irq_context: 0 &sb->s_type->i_mutex_key#8 &sbi->s_writepages_rwsem rcu_read_lock &c->lock irq_context: 0 (wq_completion)wg-kex-wg0#12 (work_completion)(&({ do { const void *__vpp_verify = (typeof((worker) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __asm__ ("" : "=r"(__ptr) : "0"((typeof(*((worker))) *)((worker)))); (typeof((typeof(*((worker))) *)((worker)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); })->work) &rq->__lock irq_context: 0 (wq_completion)events (work_completion)(&rdev->wiphy_work) &rdev->wiphy.mtx &wdev->mtx &local->mtx &local->iflist_mtx hrtimer_bases.lock irq_context: 0 (wq_completion)events (work_completion)(&rdev->wiphy_work) &rdev->wiphy.mtx &wdev->mtx &local->mtx &local->iflist_mtx hrtimer_bases.lock &obj_hash[i].lock irq_context: softirq rcu_read_lock &rdev->bss_lock &n->list_lock irq_context: softirq rcu_read_lock &rdev->bss_lock &n->list_lock &c->lock irq_context: 0 (wq_completion)wg-kex-wg0#2 (work_completion)(&({ do { const void *__vpp_verify = (typeof((worker) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __asm__ ("" : "=r"(__ptr) : "0"((typeof(*((worker))) *)((worker)))); (typeof((typeof(*((worker))) *)((worker)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); })->work) &n->list_lock irq_context: 0 (wq_completion)wg-kex-wg0#2 (work_completion)(&({ do { const void *__vpp_verify = (typeof((worker) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __asm__ ("" : "=r"(__ptr) : "0"((typeof(*((worker))) *)((worker)))); (typeof((typeof(*((worker))) *)((worker)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); })->work) &n->list_lock &c->lock irq_context: 0 sb_writers#4 pcpu_lock stock_lock irq_context: 0 &sb->s_type->i_mutex_key#8 &sbi->s_writepages_rwsem mmu_notifier_invalidate_range_start &rq->__lock &cfs_rq->removed.lock irq_context: softirq rcu_read_lock &rdev->bss_lock krc.lock &obj_hash[i].lock irq_context: softirq rcu_read_lock &rdev->bss_lock krc.lock &base->lock irq_context: softirq rcu_read_lock &rdev->bss_lock krc.lock &base->lock &obj_hash[i].lock irq_context: 0 (wq_completion)phy12 (work_completion)(&(&local->scan_work)->work) &local->mtx rcu_state.exp_mutex &rnp->exp_wq[0] irq_context: 0 (wq_completion)phy12 (work_completion)(&(&local->scan_work)->work) &local->mtx &local->iflist_mtx tk_core.seq.seqcount irq_context: 0 (wq_completion)phy12 (work_completion)(&(&local->scan_work)->work) &local->mtx &local->iflist_mtx hrtimer_bases.lock irq_context: 0 (wq_completion)phy12 (work_completion)(&(&local->scan_work)->work) &local->mtx &local->iflist_mtx hrtimer_bases.lock tk_core.seq.seqcount irq_context: 0 (wq_completion)phy12 (work_completion)(&(&local->scan_work)->work) &local->mtx &local->iflist_mtx hrtimer_bases.lock &obj_hash[i].lock irq_context: 0 sb_writers#4 &type->i_mutex_dir_key#3/1 remove_cache_srcu rcu_read_lock &rcu_state.gp_wq irq_context: 0 sb_writers#4 &type->i_mutex_dir_key#3/1 remove_cache_srcu rcu_read_lock &rcu_state.gp_wq &p->pi_lock irq_context: 0 sb_writers#4 &type->i_mutex_dir_key#3/1 remove_cache_srcu rcu_read_lock &rcu_state.gp_wq &p->pi_lock &rq->__lock irq_context: 0 sb_writers#4 &type->i_mutex_dir_key#3/1 remove_cache_srcu rcu_read_lock &rcu_state.gp_wq &p->pi_lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)bat_events (work_completion)(&(&bat_priv->mcast.work)->work) &rcu_state.expedited_wq &p->pi_lock irq_context: 0 pernet_ops_rwsem k-sk_lock-AF_NETLINK rcu_read_lock &rcu_state.expedited_wq irq_context: 0 pernet_ops_rwsem k-sk_lock-AF_NETLINK rcu_read_lock &rcu_state.expedited_wq &p->pi_lock irq_context: 0 pernet_ops_rwsem k-sk_lock-AF_NETLINK rcu_read_lock &rcu_state.expedited_wq &p->pi_lock &rq->__lock irq_context: 0 pernet_ops_rwsem k-sk_lock-AF_NETLINK rcu_read_lock &rcu_state.expedited_wq &p->pi_lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 &sb->s_type->i_mutex_key#10 sk_lock-AF_CAIF rcu_state.exp_mutex rcu_read_lock &rq->__lock irq_context: 0 &sb->s_type->i_mutex_key#10 sk_lock-AF_CAIF rcu_state.exp_mutex rcu_read_lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 sk_lock-AF_CAIF fill_pool_map-wait-type-override pool_lock#2 irq_context: 0 sk_lock-AF_CAIF fill_pool_map-wait-type-override &rq->__lock irq_context: 0 sk_lock-AF_CAIF fill_pool_map-wait-type-override &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 sk_lock-AF_CAIF fill_pool_map-wait-type-override &c->lock irq_context: 0 sk_lock-AF_CAIF fill_pool_map-wait-type-override &n->list_lock irq_context: 0 sk_lock-AF_CAIF fill_pool_map-wait-type-override &n->list_lock &c->lock irq_context: 0 sk_lock-AF_CAIF fill_pool_map-wait-type-override pool_lock irq_context: 0 cb_lock rtnl_mutex &rdev->wiphy.mtx stock_lock irq_context: 0 cb_lock rtnl_mutex &rdev->wiphy.mtx stack_depot_init_mutex irq_context: 0 cb_lock rtnl_mutex &rdev->wiphy.mtx pcpu_alloc_mutex irq_context: 0 cb_lock rtnl_mutex &rdev->wiphy.mtx pcpu_alloc_mutex pcpu_lock irq_context: 0 cb_lock rtnl_mutex &rdev->wiphy.mtx &sb->s_type->i_mutex_key#3 stock_lock irq_context: 0 cb_lock rtnl_mutex &rdev->wiphy.mtx &sb->s_type->i_mutex_key#3 &n->list_lock &c->lock irq_context: 0 cb_lock rtnl_mutex &rdev->wiphy.mtx &sb->s_type->i_mutex_key#3 rcu_read_lock rcu_read_lock rcu_node_0 irq_context: 0 cb_lock rtnl_mutex &rdev->wiphy.mtx &sb->s_type->i_mutex_key#3 rcu_read_lock rcu_read_lock &rq->__lock irq_context: 0 cb_lock rtnl_mutex &rdev->wiphy.mtx &sb->s_type->i_mutex_key#3 rcu_read_lock rcu_read_lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 cb_lock rtnl_mutex &rdev->wiphy.mtx &sb->s_type->i_mutex_key#3 rcu_read_lock rcu_node_0 irq_context: 0 cb_lock rtnl_mutex &rdev->wiphy.mtx &xa->xa_lock#3 irq_context: 0 cb_lock rtnl_mutex &rdev->wiphy.mtx net_rwsem irq_context: 0 cb_lock rtnl_mutex &rdev->wiphy.mtx net_rwsem &list->lock#2 irq_context: 0 cb_lock rtnl_mutex &rdev->wiphy.mtx &tn->lock irq_context: 0 cb_lock rtnl_mutex &rdev->wiphy.mtx &x->wait#9 irq_context: 0 cb_lock rtnl_mutex &rdev->wiphy.mtx gdp_mutex irq_context: 0 cb_lock rtnl_mutex &rdev->wiphy.mtx gdp_mutex &k->list_lock irq_context: 0 cb_lock rtnl_mutex &rdev->wiphy.mtx lock irq_context: 0 cb_lock rtnl_mutex &rdev->wiphy.mtx lock kernfs_idr_lock irq_context: 0 cb_lock rtnl_mutex &rdev->wiphy.mtx &root->kernfs_rwsem &root->kernfs_iattr_rwsem irq_context: 0 cb_lock rtnl_mutex &rdev->wiphy.mtx bus_type_sem irq_context: 0 cb_lock rtnl_mutex &rdev->wiphy.mtx sysfs_symlink_target_lock irq_context: 0 cb_lock rtnl_mutex &rdev->wiphy.mtx &dev->power.lock irq_context: 0 cb_lock rtnl_mutex &rdev->wiphy.mtx dpm_list_mtx irq_context: 0 cb_lock rtnl_mutex &rdev->wiphy.mtx &k->k_lock irq_context: 0 cb_lock rtnl_mutex &rdev->wiphy.mtx subsys mutex#17 irq_context: 0 cb_lock rtnl_mutex &rdev->wiphy.mtx subsys mutex#17 &k->k_lock irq_context: 0 cb_lock rtnl_mutex &rdev->wiphy.mtx &dir->lock#2 irq_context: 0 cb_lock rtnl_mutex &rdev->wiphy.mtx uevent_sock_mutex &c->lock irq_context: 0 cb_lock rtnl_mutex &rdev->wiphy.mtx uevent_sock_mutex &n->list_lock irq_context: 0 cb_lock rtnl_mutex &rdev->wiphy.mtx uevent_sock_mutex &n->list_lock &c->lock irq_context: 0 cb_lock rtnl_mutex &rdev->wiphy.mtx dev_hotplug_mutex irq_context: 0 cb_lock rtnl_mutex &rdev->wiphy.mtx dev_hotplug_mutex &dev->power.lock irq_context: 0 cb_lock rtnl_mutex &rdev->wiphy.mtx dev_base_lock irq_context: 0 cb_lock rtnl_mutex &rdev->wiphy.mtx input_pool.lock irq_context: 0 cb_lock rtnl_mutex &rdev->wiphy.mtx rcu_read_lock &pool->lock/1 irq_context: 0 cb_lock rtnl_mutex &rdev->wiphy.mtx rcu_read_lock &pool->lock/1 &obj_hash[i].lock irq_context: 0 cb_lock rtnl_mutex &rdev->wiphy.mtx rcu_read_lock &pool->lock/1 pool_lock#2 irq_context: 0 cb_lock rtnl_mutex &rdev->wiphy.mtx rcu_read_lock &pool->lock/1 &p->pi_lock irq_context: 0 cb_lock rtnl_mutex &rdev->wiphy.mtx rcu_read_lock &pool->lock/1 &p->pi_lock &rq->__lock irq_context: 0 cb_lock rtnl_mutex &rdev->wiphy.mtx rcu_read_lock &pool->lock/1 &p->pi_lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 cb_lock rtnl_mutex &rdev->wiphy.mtx batched_entropy_u32.lock irq_context: 0 cb_lock rtnl_mutex &rdev->wiphy.mtx &tbl->lock irq_context: 0 cb_lock rtnl_mutex &rdev->wiphy.mtx sysctl_lock irq_context: 0 cb_lock rtnl_mutex &rdev->wiphy.mtx failover_lock irq_context: 0 cb_lock rtnl_mutex &rdev->wiphy.mtx proc_subdir_lock irq_context: 0 cb_lock rtnl_mutex &rdev->wiphy.mtx proc_inum_ida.xa_lock irq_context: 0 cb_lock rtnl_mutex &rdev->wiphy.mtx proc_subdir_lock irq_context: 0 cb_lock rtnl_mutex &rdev->wiphy.mtx fill_pool_map-wait-type-override &rq->__lock irq_context: 0 cb_lock rtnl_mutex &rdev->wiphy.mtx fill_pool_map-wait-type-override &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 cb_lock rtnl_mutex &rdev->wiphy.mtx &idev->mc_lock irq_context: 0 cb_lock rtnl_mutex &rdev->wiphy.mtx &idev->mc_lock fs_reclaim irq_context: 0 cb_lock rtnl_mutex &rdev->wiphy.mtx &idev->mc_lock fs_reclaim mmu_notifier_invalidate_range_start irq_context: 0 cb_lock rtnl_mutex &rdev->wiphy.mtx &idev->mc_lock pool_lock#2 irq_context: 0 cb_lock rtnl_mutex &rdev->wiphy.mtx &idev->mc_lock &obj_hash[i].lock irq_context: 0 cb_lock rtnl_mutex &rdev->wiphy.mtx &idev->mc_lock _xmit_ETHER irq_context: 0 cb_lock rtnl_mutex &rdev->wiphy.mtx &idev->mc_lock _xmit_ETHER &c->lock irq_context: 0 cb_lock rtnl_mutex &rdev->wiphy.mtx &idev->mc_lock _xmit_ETHER pool_lock#2 irq_context: 0 cb_lock rtnl_mutex &rdev->wiphy.mtx &pnettable->lock irq_context: 0 cb_lock rtnl_mutex &rdev->wiphy.mtx smc_ib_devices.mutex irq_context: 0 cb_lock rtnl_mutex &rdev->wiphy.mtx rcu_read_lock &ndev->lock irq_context: 0 (wq_completion)bond4 (work_completion)(&(&bond->mii_work)->work) &rq->__lock irq_context: 0 (wq_completion)bond4 (work_completion)(&(&bond->mii_work)->work) &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 pernet_ops_rwsem rtnl_mutex rcu_state.exp_mutex fill_pool_map-wait-type-override &n->list_lock irq_context: 0 pernet_ops_rwsem rtnl_mutex rcu_state.exp_mutex fill_pool_map-wait-type-override &n->list_lock &c->lock irq_context: 0 pernet_ops_rwsem rtnl_mutex rcu_state.exp_mutex fill_pool_map-wait-type-override &rq->__lock irq_context: 0 pernet_ops_rwsem rtnl_mutex rcu_state.exp_mutex fill_pool_map-wait-type-override &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 pernet_ops_rwsem rtnl_mutex rcu_state.exp_mutex fill_pool_map-wait-type-override batched_entropy_u8.lock irq_context: 0 pernet_ops_rwsem rtnl_mutex rcu_state.exp_mutex fill_pool_map-wait-type-override kfence_freelist_lock irq_context: 0 &sb->s_type->i_mutex_key#10 (netlink_chain).rwsem rcu_read_lock rcu_read_lock &pool->lock &p->pi_lock &rq->__lock irq_context: 0 &sb->s_type->i_mutex_key#10 (netlink_chain).rwsem rcu_read_lock rcu_read_lock &pool->lock &p->pi_lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)events (work_completion)(&rdev->destroy_work) rtnl_mutex rtnl_mutex.wait_lock irq_context: 0 (wq_completion)events (work_completion)(&rdev->destroy_work) rtnl_mutex &pool->lock irq_context: 0 (wq_completion)events (work_completion)(&rdev->destroy_work) rtnl_mutex &rq->__lock irq_context: 0 (wq_completion)events (work_completion)(&rdev->destroy_work) rtnl_mutex &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)events (work_completion)(&rdev->destroy_work) rtnl_mutex &rdev->wiphy.mtx &root->kernfs_rwsem irq_context: 0 (wq_completion)events (work_completion)(&rdev->destroy_work) rtnl_mutex &rdev->wiphy.mtx &root->kernfs_rwsem &root->kernfs_iattr_rwsem irq_context: 0 (wq_completion)events (work_completion)(&rdev->destroy_work) rtnl_mutex &rdev->wiphy.mtx &root->kernfs_rwsem kernfs_idr_lock irq_context: 0 (wq_completion)events (work_completion)(&rdev->destroy_work) rtnl_mutex &rdev->wiphy.mtx &root->kernfs_rwsem &obj_hash[i].lock irq_context: 0 (wq_completion)events (work_completion)(&rdev->destroy_work) rtnl_mutex &rdev->wiphy.mtx &root->kernfs_rwsem pool_lock#2 irq_context: 0 (wq_completion)events (work_completion)(&rdev->destroy_work) rtnl_mutex &rdev->wiphy.mtx rcu_state.exp_mutex &rnp->exp_wq[1] irq_context: 0 (wq_completion)events (work_completion)(&rdev->destroy_work) rtnl_mutex &rdev->wiphy.mtx dev_base_lock irq_context: 0 (wq_completion)events (work_completion)(&rdev->destroy_work) rtnl_mutex &rdev->wiphy.mtx dev_base_lock &xa->xa_lock#3 irq_context: 0 (wq_completion)events (work_completion)(&rdev->destroy_work) rtnl_mutex &rdev->wiphy.mtx cpu_hotplug_lock irq_context: 0 (wq_completion)events (work_completion)(&rdev->destroy_work) rtnl_mutex &rdev->wiphy.mtx cpu_hotplug_lock &list->lock#5 irq_context: 0 (wq_completion)events (work_completion)(&rdev->destroy_work) rtnl_mutex &rdev->wiphy.mtx rcu_state.exp_mutex &rnp->exp_wq[2] irq_context: 0 (wq_completion)events (work_completion)(&rdev->destroy_work) rtnl_mutex &rdev->wiphy.mtx bpf_devs_lock irq_context: 0 (wq_completion)events (work_completion)(&rdev->destroy_work) rtnl_mutex &rdev->wiphy.mtx &hwstats->hwsdev_list_lock irq_context: 0 (wq_completion)events (work_completion)(&rdev->destroy_work) rtnl_mutex &rdev->wiphy.mtx net_rwsem irq_context: 0 (wq_completion)events (work_completion)(&rdev->destroy_work) rtnl_mutex &rdev->wiphy.mtx net_rwsem &list->lock#2 irq_context: 0 (wq_completion)events (work_completion)(&rdev->destroy_work) rtnl_mutex &rdev->wiphy.mtx rcu_read_lock &pool->lock/1 irq_context: 0 (wq_completion)events (work_completion)(&rdev->destroy_work) rtnl_mutex &rdev->wiphy.mtx rcu_read_lock &pool->lock/1 &obj_hash[i].lock irq_context: 0 (wq_completion)events (work_completion)(&rdev->destroy_work) rtnl_mutex &rdev->wiphy.mtx rcu_read_lock &pool->lock/1 pool_lock#2 irq_context: 0 (wq_completion)events (work_completion)(&rdev->destroy_work) rtnl_mutex &rdev->wiphy.mtx rcu_read_lock &pool->lock/1 &p->pi_lock irq_context: 0 (wq_completion)events (work_completion)(&rdev->destroy_work) rtnl_mutex &rdev->wiphy.mtx rcu_read_lock &pool->lock/1 &p->pi_lock &rq->__lock irq_context: 0 (wq_completion)events (work_completion)(&rdev->destroy_work) rtnl_mutex &rdev->wiphy.mtx rcu_read_lock &pool->lock/1 &p->pi_lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)events (work_completion)(&rdev->destroy_work) rtnl_mutex &rdev->wiphy.mtx &in_dev->mc_tomb_lock irq_context: 0 (wq_completion)events (work_completion)(&rdev->destroy_work) rtnl_mutex &rdev->wiphy.mtx sysctl_lock irq_context: 0 (wq_completion)events (work_completion)(&rdev->destroy_work) rtnl_mutex &rdev->wiphy.mtx sysctl_lock &obj_hash[i].lock irq_context: 0 (wq_completion)events (work_completion)(&rdev->destroy_work) rtnl_mutex &rdev->wiphy.mtx sysctl_lock pool_lock#2 irq_context: 0 (wq_completion)events (work_completion)(&rdev->destroy_work) rtnl_mutex &rdev->wiphy.mtx sysctl_lock krc.lock irq_context: 0 (wq_completion)events (work_completion)(&rdev->destroy_work) rtnl_mutex &rdev->wiphy.mtx stock_lock irq_context: 0 (wq_completion)events (work_completion)(&rdev->destroy_work) rtnl_mutex &rdev->wiphy.mtx &c->lock irq_context: 0 (wq_completion)events (work_completion)(&rdev->destroy_work) rtnl_mutex &rdev->wiphy.mtx &n->list_lock irq_context: 0 (wq_completion)events (work_completion)(&rdev->destroy_work) rtnl_mutex &rdev->wiphy.mtx &n->list_lock &c->lock irq_context: 0 (wq_completion)events (work_completion)(&rdev->destroy_work) rtnl_mutex &rdev->wiphy.mtx &rq->__lock irq_context: 0 (wq_completion)events (work_completion)(&rdev->destroy_work) rtnl_mutex &rdev->wiphy.mtx &____s->seqcount irq_context: 0 (wq_completion)events (work_completion)(&rdev->destroy_work) rtnl_mutex &rdev->wiphy.mtx &tbl->lock irq_context: 0 (wq_completion)events (work_completion)(&rdev->destroy_work) rtnl_mutex &rdev->wiphy.mtx &dir->lock#2 irq_context: 0 (wq_completion)events (work_completion)(&rdev->destroy_work) rtnl_mutex &rdev->wiphy.mtx class irq_context: 0 (wq_completion)events (work_completion)(&rdev->destroy_work) rtnl_mutex &rdev->wiphy.mtx (&tbl->proxy_timer) irq_context: 0 (wq_completion)events (work_completion)(&rdev->destroy_work) rtnl_mutex &rdev->wiphy.mtx &base->lock irq_context: 0 (wq_completion)events (work_completion)(&rdev->destroy_work) rtnl_mutex &rdev->wiphy.mtx &obj_hash[i].lock pool_lock irq_context: 0 (wq_completion)events (work_completion)(&rdev->destroy_work) rtnl_mutex &rdev->wiphy.mtx fill_pool_map-wait-type-override pool_lock#2 irq_context: 0 (wq_completion)events (work_completion)(&rdev->destroy_work) rtnl_mutex &rdev->wiphy.mtx fill_pool_map-wait-type-override &c->lock irq_context: 0 (wq_completion)events (work_completion)(&rdev->destroy_work) rtnl_mutex &rdev->wiphy.mtx fill_pool_map-wait-type-override pool_lock irq_context: 0 (wq_completion)events (work_completion)(&rdev->destroy_work) rtnl_mutex &rdev->wiphy.mtx &ul->lock irq_context: 0 (wq_completion)events (work_completion)(&rdev->destroy_work) rtnl_mutex &rdev->wiphy.mtx &net->xdp.lock irq_context: 0 (wq_completion)events (work_completion)(&rdev->destroy_work) rtnl_mutex &rdev->wiphy.mtx krc.lock irq_context: 0 (wq_completion)events (work_completion)(&rdev->destroy_work) rtnl_mutex &rdev->wiphy.mtx mirred_list_lock irq_context: 0 (wq_completion)events (work_completion)(&rdev->destroy_work) rtnl_mutex &rdev->wiphy.mtx &nft_net->commit_mutex irq_context: 0 (wq_completion)events (work_completion)(&rdev->destroy_work) rtnl_mutex &rdev->wiphy.mtx &tn->lock irq_context: 0 (wq_completion)events (work_completion)(&rdev->destroy_work) rtnl_mutex &rdev->wiphy.mtx rcu_read_lock &tb->tb6_lock irq_context: 0 (wq_completion)events (work_completion)(&rdev->destroy_work) rtnl_mutex &rdev->wiphy.mtx rcu_read_lock &tb->tb6_lock &net->ipv6.fib6_walker_lock irq_context: 0 (wq_completion)events (work_completion)(&rdev->destroy_work) rtnl_mutex &rdev->wiphy.mtx proc_subdir_lock irq_context: 0 (wq_completion)events (work_completion)(&rdev->destroy_work) rtnl_mutex &rdev->wiphy.mtx &ent->pde_unload_lock irq_context: 0 (wq_completion)events (work_completion)(&rdev->destroy_work) rtnl_mutex &rdev->wiphy.mtx proc_inum_ida.xa_lock irq_context: 0 (wq_completion)events (work_completion)(&rdev->destroy_work) rtnl_mutex &rdev->wiphy.mtx &net->ipv6.addrconf_hash_lock irq_context: 0 (wq_completion)events (work_completion)(&rdev->destroy_work) rtnl_mutex &rdev->wiphy.mtx &ndev->lock irq_context: 0 (wq_completion)events (work_completion)(&rdev->destroy_work) rtnl_mutex &rdev->wiphy.mtx &ndev->lock &obj_hash[i].lock irq_context: 0 (wq_completion)events (work_completion)(&rdev->destroy_work) rtnl_mutex &rdev->wiphy.mtx &idev->mc_lock irq_context: 0 (wq_completion)events (work_completion)(&rdev->destroy_work) rtnl_mutex &rdev->wiphy.mtx &idev->mc_lock _xmit_ETHER irq_context: 0 (wq_completion)events (work_completion)(&rdev->destroy_work) rtnl_mutex &rdev->wiphy.mtx &idev->mc_lock _xmit_ETHER &obj_hash[i].lock irq_context: 0 (wq_completion)events (work_completion)(&rdev->destroy_work) rtnl_mutex &rdev->wiphy.mtx &idev->mc_lock _xmit_ETHER pool_lock#2 irq_context: 0 (wq_completion)events (work_completion)(&rdev->destroy_work) rtnl_mutex &rdev->wiphy.mtx &idev->mc_lock _xmit_ETHER krc.lock irq_context: 0 (wq_completion)events (work_completion)(&rdev->destroy_work) rtnl_mutex &rdev->wiphy.mtx &idev->mc_query_lock irq_context: 0 (wq_completion)events (work_completion)(&rdev->destroy_work) rtnl_mutex &rdev->wiphy.mtx &idev->mc_query_lock &obj_hash[i].lock irq_context: 0 (wq_completion)events (work_completion)(&rdev->destroy_work) rtnl_mutex &rdev->wiphy.mtx (work_completion)(&(&idev->mc_report_work)->work) irq_context: 0 (wq_completion)events (work_completion)(&rdev->destroy_work) rtnl_mutex &rdev->wiphy.mtx &idev->mc_report_lock irq_context: 0 (wq_completion)events (work_completion)(&rdev->destroy_work) rtnl_mutex &rdev->wiphy.mtx &idev->mc_lock &obj_hash[i].lock irq_context: 0 (wq_completion)events (work_completion)(&rdev->destroy_work) rtnl_mutex &rdev->wiphy.mtx &idev->mc_lock pool_lock#2 irq_context: 0 (wq_completion)events (work_completion)(&rdev->destroy_work) rtnl_mutex &rdev->wiphy.mtx &idev->mc_lock krc.lock irq_context: 0 (wq_completion)events (work_completion)(&rdev->destroy_work) rtnl_mutex &rdev->wiphy.mtx &pnn->pndevs.lock irq_context: 0 (wq_completion)events (work_completion)(&rdev->destroy_work) rtnl_mutex &rdev->wiphy.mtx &pnn->routes.lock irq_context: 0 (wq_completion)events (work_completion)(&rdev->destroy_work) rtnl_mutex &rdev->wiphy.mtx &pnettable->lock irq_context: 0 (wq_completion)events (work_completion)(&rdev->destroy_work) rtnl_mutex &rdev->wiphy.mtx smc_ib_devices.mutex irq_context: 0 (wq_completion)events (work_completion)(&rdev->destroy_work) rtnl_mutex &rdev->wiphy.mtx smc_ib_devices.mutex &rq->__lock irq_context: 0 (wq_completion)events (work_completion)(&rdev->destroy_work) rtnl_mutex &rdev->wiphy.mtx smc_ib_devices.mutex &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)events (work_completion)(&rdev->destroy_work) rtnl_mutex &rdev->wiphy.mtx target_list_lock irq_context: 0 (wq_completion)events (work_completion)(&rdev->destroy_work) rtnl_mutex &rdev->wiphy.mtx _xmit_ETHER irq_context: 0 (wq_completion)events (work_completion)(&rdev->destroy_work) rtnl_mutex &rdev->wiphy.mtx pin_fs_lock irq_context: 0 (wq_completion)events (work_completion)(&rdev->destroy_work) rtnl_mutex &rdev->wiphy.mtx &dentry->d_lock irq_context: 0 (wq_completion)events (work_completion)(&rdev->destroy_work) rtnl_mutex &rdev->wiphy.mtx &sb->s_type->i_mutex_key#3 irq_context: 0 (wq_completion)events (work_completion)(&rdev->destroy_work) rtnl_mutex &rdev->wiphy.mtx &sb->s_type->i_mutex_key#3 &dentry->d_lock irq_context: 0 (wq_completion)events (work_completion)(&rdev->destroy_work) rtnl_mutex &rdev->wiphy.mtx &sb->s_type->i_mutex_key#3 &dentry->d_lock &dentry->d_lock/1 irq_context: 0 (wq_completion)events (work_completion)(&rdev->destroy_work) rtnl_mutex &rdev->wiphy.mtx &sb->s_type->i_mutex_key#3 tk_core.seq.seqcount irq_context: 0 (wq_completion)events (work_completion)(&rdev->destroy_work) rtnl_mutex &rdev->wiphy.mtx &sb->s_type->i_mutex_key#3 rename_lock.seqcount irq_context: 0 (wq_completion)events (work_completion)(&rdev->destroy_work) rtnl_mutex &rdev->wiphy.mtx &sb->s_type->i_mutex_key#3 pin_fs_lock irq_context: 0 (wq_completion)events (work_completion)(&rdev->destroy_work) rtnl_mutex &rdev->wiphy.mtx &sb->s_type->i_mutex_key#3 rcu_read_lock mount_lock irq_context: 0 (wq_completion)events (work_completion)(&rdev->destroy_work) rtnl_mutex &rdev->wiphy.mtx &sb->s_type->i_mutex_key#3 rcu_read_lock mount_lock mount_lock.seqcount irq_context: 0 (wq_completion)events (work_completion)(&rdev->destroy_work) rtnl_mutex &rdev->wiphy.mtx &sb->s_type->i_mutex_key#3 mount_lock irq_context: 0 (wq_completion)events (work_completion)(&rdev->destroy_work) rtnl_mutex &rdev->wiphy.mtx &sb->s_type->i_mutex_key#3 mount_lock mount_lock.seqcount irq_context: 0 (wq_completion)events (work_completion)(&rdev->destroy_work) rtnl_mutex &rdev->wiphy.mtx &sb->s_type->i_mutex_key#3 rcu_read_lock &dentry->d_lock irq_context: 0 (wq_completion)events (work_completion)(&rdev->destroy_work) rtnl_mutex &rdev->wiphy.mtx &sb->s_type->i_mutex_key#3 rcu_read_lock rcu_node_0 irq_context: 0 (wq_completion)events (work_completion)(&rdev->destroy_work) rtnl_mutex &rdev->wiphy.mtx &sb->s_type->i_mutex_key#3 rcu_read_lock &rq->__lock irq_context: 0 (wq_completion)events (work_completion)(&rdev->destroy_work) rtnl_mutex &rdev->wiphy.mtx &sb->s_type->i_mutex_key#3 rcu_read_lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)events (work_completion)(&rdev->destroy_work) rtnl_mutex &rdev->wiphy.mtx &sb->s_type->i_mutex_key#3 rcu_node_0 irq_context: 0 (wq_completion)events (work_completion)(&rdev->destroy_work) rtnl_mutex &rdev->wiphy.mtx &sb->s_type->i_mutex_key#3 &rq->__lock irq_context: 0 (wq_completion)events (work_completion)(&rdev->destroy_work) rtnl_mutex &rdev->wiphy.mtx &sb->s_type->i_mutex_key#3 &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)events (work_completion)(&rdev->destroy_work) rtnl_mutex &rdev->wiphy.mtx &sb->s_type->i_mutex_key#3 &fsnotify_mark_srcu irq_context: 0 (wq_completion)events (work_completion)(&rdev->destroy_work) rtnl_mutex &rdev->wiphy.mtx &sb->s_type->i_mutex_key#3 &sb->s_type->i_lock_key#7 irq_context: 0 (wq_completion)events (work_completion)(&rdev->destroy_work) rtnl_mutex &rdev->wiphy.mtx &sb->s_type->i_mutex_key#3 &s->s_inode_list_lock irq_context: 0 (wq_completion)events (work_completion)(&rdev->destroy_work) rtnl_mutex &rdev->wiphy.mtx &sb->s_type->i_mutex_key#3 &xa->xa_lock#9 irq_context: 0 (wq_completion)events (work_completion)(&rdev->destroy_work) rtnl_mutex &rdev->wiphy.mtx &sb->s_type->i_mutex_key#3 &obj_hash[i].lock irq_context: 0 (wq_completion)events (work_completion)(&rdev->destroy_work) rtnl_mutex &rdev->wiphy.mtx &sb->s_type->i_mutex_key#3 pool_lock#2 irq_context: 0 (wq_completion)events (work_completion)(&rdev->destroy_work) rtnl_mutex &rdev->wiphy.mtx &sb->s_type->i_mutex_key#3 &obj_hash[i].lock pool_lock irq_context: 0 (wq_completion)events (work_completion)(&rdev->destroy_work) rtnl_mutex &rdev->wiphy.mtx &sb->s_type->i_mutex_key#3 fill_pool_map-wait-type-override pool_lock#2 irq_context: 0 (wq_completion)events (work_completion)(&rdev->destroy_work) rtnl_mutex &rdev->wiphy.mtx &sb->s_type->i_mutex_key#3 fill_pool_map-wait-type-override &c->lock irq_context: 0 (wq_completion)events (work_completion)(&rdev->destroy_work) rtnl_mutex &rdev->wiphy.mtx &sb->s_type->i_mutex_key#3 fill_pool_map-wait-type-override &n->list_lock irq_context: 0 (wq_completion)events (work_completion)(&rdev->destroy_work) rtnl_mutex &rdev->wiphy.mtx &sb->s_type->i_mutex_key#3 fill_pool_map-wait-type-override &n->list_lock &c->lock irq_context: 0 (wq_completion)events (work_completion)(&rdev->destroy_work) rtnl_mutex &rdev->wiphy.mtx &sb->s_type->i_mutex_key#3 fill_pool_map-wait-type-override &rq->__lock irq_context: 0 (wq_completion)events (work_completion)(&rdev->destroy_work) rtnl_mutex &rdev->wiphy.mtx &sb->s_type->i_mutex_key#3 fill_pool_map-wait-type-override &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)events (work_completion)(&rdev->destroy_work) rtnl_mutex &rdev->wiphy.mtx &sb->s_type->i_mutex_key#3 fill_pool_map-wait-type-override pool_lock irq_context: 0 (wq_completion)events (work_completion)(&rdev->destroy_work) rtnl_mutex &rdev->wiphy.mtx rcu_read_lock &dentry->d_lock irq_context: 0 (wq_completion)events (work_completion)(&rdev->destroy_work) rtnl_mutex &rdev->wiphy.mtx &fsnotify_mark_srcu irq_context: 0 (wq_completion)events (work_completion)(&rdev->destroy_work) rtnl_mutex &rdev->wiphy.mtx &sb->s_type->i_lock_key#7 irq_context: 0 (wq_completion)events (work_completion)(&rdev->destroy_work) rtnl_mutex &rdev->wiphy.mtx &s->s_inode_list_lock irq_context: 0 (wq_completion)events (work_completion)(&rdev->destroy_work) rtnl_mutex &rdev->wiphy.mtx &xa->xa_lock#9 irq_context: 0 (wq_completion)events (work_completion)(&rdev->destroy_work) rtnl_mutex &rdev->wiphy.mtx rcu_read_lock mount_lock irq_context: 0 (wq_completion)events (work_completion)(&rdev->destroy_work) rtnl_mutex &rdev->wiphy.mtx rcu_read_lock mount_lock mount_lock.seqcount irq_context: 0 (wq_completion)events (work_completion)(&rdev->destroy_work) rtnl_mutex &rdev->wiphy.mtx mount_lock irq_context: 0 (wq_completion)events (work_completion)(&rdev->destroy_work) rtnl_mutex &rdev->wiphy.mtx mount_lock mount_lock.seqcount irq_context: 0 (wq_completion)events (work_completion)(&rdev->destroy_work) rtnl_mutex &rdev->wiphy.mtx uevent_sock_mutex irq_context: 0 (wq_completion)events (work_completion)(&rdev->destroy_work) rtnl_mutex &rdev->wiphy.mtx uevent_sock_mutex fs_reclaim irq_context: 0 (wq_completion)events (work_completion)(&rdev->destroy_work) rtnl_mutex &rdev->wiphy.mtx uevent_sock_mutex fs_reclaim mmu_notifier_invalidate_range_start irq_context: 0 (wq_completion)events (work_completion)(&rdev->destroy_work) rtnl_mutex &rdev->wiphy.mtx uevent_sock_mutex fs_reclaim mmu_notifier_invalidate_range_start &rq->__lock irq_context: 0 (wq_completion)events (work_completion)(&rdev->destroy_work) rtnl_mutex &rdev->wiphy.mtx uevent_sock_mutex fs_reclaim mmu_notifier_invalidate_range_start &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)events (work_completion)(&rdev->destroy_work) rtnl_mutex &rdev->wiphy.mtx uevent_sock_mutex pool_lock#2 irq_context: 0 (wq_completion)events (work_completion)(&rdev->destroy_work) rtnl_mutex &rdev->wiphy.mtx uevent_sock_mutex nl_table_lock irq_context: 0 (wq_completion)events (work_completion)(&rdev->destroy_work) rtnl_mutex &rdev->wiphy.mtx uevent_sock_mutex &obj_hash[i].lock irq_context: 0 (wq_completion)events (work_completion)(&rdev->destroy_work) rtnl_mutex &rdev->wiphy.mtx uevent_sock_mutex nl_table_wait.lock irq_context: 0 (wq_completion)events (work_completion)(&rdev->destroy_work) rtnl_mutex &rdev->wiphy.mtx sysfs_symlink_target_lock irq_context: 0 (wq_completion)events (work_completion)(&rdev->destroy_work) rtnl_mutex &rdev->wiphy.mtx kernfs_idr_lock irq_context: 0 (wq_completion)events (work_completion)(&rdev->destroy_work) rtnl_mutex &rdev->wiphy.mtx &k->list_lock irq_context: 0 (wq_completion)events (work_completion)(&rdev->destroy_work) rtnl_mutex &rdev->wiphy.mtx &root->kernfs_rwsem irq_context: 0 (wq_completion)events (work_completion)(&rdev->destroy_work) rtnl_mutex &rdev->wiphy.mtx dev_hotplug_mutex irq_context: 0 (wq_completion)events (work_completion)(&rdev->destroy_work) rtnl_mutex &rdev->wiphy.mtx dev_hotplug_mutex &dev->power.lock irq_context: 0 (wq_completion)events (work_completion)(&rdev->destroy_work) rtnl_mutex &rdev->wiphy.mtx dev_hotplug_mutex &k->k_lock irq_context: 0 (wq_completion)events (work_completion)(&rdev->destroy_work) rtnl_mutex &rdev->wiphy.mtx dev_hotplug_mutex &rq->__lock irq_context: 0 (wq_completion)events (work_completion)(&rdev->destroy_work) rtnl_mutex &rdev->wiphy.mtx dev_hotplug_mutex &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)events (work_completion)(&rdev->destroy_work) rtnl_mutex &rdev->wiphy.mtx dev_pm_qos_sysfs_mtx irq_context: 0 (wq_completion)events (work_completion)(&rdev->destroy_work) rtnl_mutex &rdev->wiphy.mtx dev_pm_qos_sysfs_mtx &root->kernfs_rwsem irq_context: 0 (wq_completion)events (work_completion)(&rdev->destroy_work) rtnl_mutex &rdev->wiphy.mtx dev_pm_qos_sysfs_mtx &root->kernfs_rwsem irq_context: 0 (wq_completion)events (work_completion)(&rdev->destroy_work) rtnl_mutex &rdev->wiphy.mtx dev_pm_qos_sysfs_mtx dev_pm_qos_mtx irq_context: 0 (wq_completion)events (work_completion)(&rdev->destroy_work) rtnl_mutex &rdev->wiphy.mtx &k->k_lock irq_context: 0 (wq_completion)events (work_completion)(&rdev->destroy_work) rtnl_mutex &rdev->wiphy.mtx &k->k_lock klist_remove_lock irq_context: 0 (wq_completion)events (work_completion)(&rdev->destroy_work) rtnl_mutex &rdev->wiphy.mtx subsys mutex#17 irq_context: 0 (wq_completion)events (work_completion)(&rdev->destroy_work) rtnl_mutex &rdev->wiphy.mtx subsys mutex#17 &k->k_lock irq_context: 0 (wq_completion)events (work_completion)(&rdev->destroy_work) rtnl_mutex &rdev->wiphy.mtx subsys mutex#17 &k->k_lock klist_remove_lock irq_context: 0 (wq_completion)events (work_completion)(&rdev->destroy_work) rtnl_mutex &rdev->wiphy.mtx &root->kernfs_rwsem &rq->__lock irq_context: 0 (wq_completion)events (work_completion)(&rdev->destroy_work) rtnl_mutex &rdev->wiphy.mtx &x->wait#9 irq_context: 0 (wq_completion)events (work_completion)(&rdev->destroy_work) rtnl_mutex &rdev->wiphy.mtx dpm_list_mtx irq_context: 0 (wq_completion)events (work_completion)(&rdev->destroy_work) rtnl_mutex &rdev->wiphy.mtx &dev->power.lock irq_context: 0 (wq_completion)events (work_completion)(&rdev->destroy_work) rtnl_mutex &rdev->wiphy.mtx deferred_probe_mutex irq_context: 0 (wq_completion)events (work_completion)(&rdev->destroy_work) rtnl_mutex &rdev->wiphy.mtx device_links_lock irq_context: 0 (wq_completion)events (work_completion)(&rdev->destroy_work) rtnl_mutex &rdev->wiphy.mtx mmu_notifier_invalidate_range_start irq_context: 0 (wq_completion)events (work_completion)(&rdev->destroy_work) rtnl_mutex &rdev->wiphy.mtx uevent_sock_mutex mmu_notifier_invalidate_range_start irq_context: 0 (wq_completion)events (work_completion)(&rdev->destroy_work) rtnl_mutex &rdev->wiphy.mtx gdp_mutex irq_context: 0 (wq_completion)events (work_completion)(&rdev->destroy_work) rtnl_mutex &rdev->wiphy.mtx cpu_hotplug_lock xps_map_mutex irq_context: 0 (wq_completion)events (work_completion)(&rdev->destroy_work) rtnl_mutex &lock->wait_lock irq_context: 0 (wq_completion)events (work_completion)(&rdev->destroy_work) rtnl_mutex &p->pi_lock irq_context: 0 (wq_completion)events (work_completion)(&rdev->destroy_work) rtnl_mutex &p->pi_lock &rq->__lock irq_context: 0 (wq_completion)events (work_completion)(&rdev->destroy_work) rtnl_mutex &p->pi_lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)events (work_completion)(&rdev->destroy_work) rtnl_mutex.wait_lock irq_context: 0 (wq_completion)events (work_completion)(&rdev->destroy_work) &p->pi_lock irq_context: 0 (wq_completion)events (work_completion)(&rdev->destroy_work) &p->pi_lock &rq->__lock irq_context: 0 (wq_completion)events (work_completion)(&rdev->destroy_work) &p->pi_lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)events (work_completion)(&rdev->destroy_work) &rq->__lock irq_context: 0 (wq_completion)events (work_completion)(&rdev->destroy_work) rcu_state.barrier_mutex irq_context: 0 (wq_completion)events (work_completion)(&rdev->destroy_work) rcu_state.barrier_mutex &rq->__lock irq_context: 0 (wq_completion)events (work_completion)(&rdev->destroy_work) rcu_state.barrier_mutex &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)events (work_completion)(&rdev->destroy_work) rcu_state.barrier_mutex rcu_state.barrier_mutex.wait_lock irq_context: 0 (wq_completion)events (work_completion)(&rdev->destroy_work) rcu_state.barrier_mutex &pool->lock irq_context: 0 (wq_completion)events (work_completion)(&rdev->destroy_work) rcu_state.barrier_mutex rcu_state.barrier_lock irq_context: 0 (wq_completion)events (work_completion)(&rdev->destroy_work) rcu_state.barrier_mutex rcu_state.barrier_lock &obj_hash[i].lock irq_context: 0 (wq_completion)events (work_completion)(&rdev->destroy_work) rcu_state.barrier_mutex &x->wait#24 irq_context: 0 (wq_completion)events (work_completion)(&rdev->destroy_work) rcu_state.barrier_mutex.wait_lock irq_context: 0 (wq_completion)events (work_completion)(&rdev->destroy_work) dev_base_lock irq_context: 0 (wq_completion)events (work_completion)(&rdev->destroy_work) lweventlist_lock irq_context: 0 (wq_completion)events (work_completion)(&rdev->destroy_work) pcpu_lock irq_context: 0 (wq_completion)events (work_completion)(&rdev->destroy_work) stock_lock irq_context: 0 (wq_completion)events (work_completion)(&rdev->destroy_work) &obj_hash[i].lock irq_context: 0 (wq_completion)events (work_completion)(&rdev->destroy_work) pool_lock#2 irq_context: 0 (wq_completion)events (work_completion)(&rdev->destroy_work) krc.lock irq_context: 0 (wq_completion)events (work_completion)(&rdev->destroy_work) &dir->lock#2 irq_context: 0 (wq_completion)events (work_completion)(&rdev->destroy_work) &dir->lock#2 &obj_hash[i].lock irq_context: 0 (wq_completion)events (work_completion)(&rdev->destroy_work) &dir->lock#2 pool_lock#2 irq_context: 0 (wq_completion)events (work_completion)(&rdev->destroy_work) netdev_unregistering_wq.lock irq_context: 0 (wq_completion)events (work_completion)(&rdev->destroy_work) rcu_read_lock pool_lock#2 irq_context: 0 (wq_completion)events_power_efficient (work_completion)(&(&gc_work->dwork)->work) rcu_read_lock &base->lock irq_context: 0 (wq_completion)events_power_efficient (work_completion)(&(&gc_work->dwork)->work) rcu_read_lock &base->lock &obj_hash[i].lock irq_context: 0 sk_lock-AF_ALG batched_entropy_u8.lock crngs.lock irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem &net->ipv6.ip6addrlbl_table.lock fill_pool_map-wait-type-override &n->list_lock irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem &net->ipv6.ip6addrlbl_table.lock fill_pool_map-wait-type-override &n->list_lock &c->lock irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem rtnl_mutex &ht->mutex &rq->__lock irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem rtnl_mutex &ht->mutex &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem __ip_vs_app_mutex fill_pool_map-wait-type-override pool_lock#2 irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem __ip_vs_app_mutex fill_pool_map-wait-type-override &c->lock irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem __ip_vs_app_mutex fill_pool_map-wait-type-override &n->list_lock irq_context: 0 pernet_ops_rwsem rtnl_mutex &caifn->caifdevs.lock &rnp->exp_wq[2] irq_context: 0 pernet_ops_rwsem rtnl_mutex &caifn->caifdevs.lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem __ip_vs_app_mutex fill_pool_map-wait-type-override &n->list_lock &c->lock irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem __ip_vs_app_mutex fill_pool_map-wait-type-override pool_lock irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem &root->kernfs_rwsem &root->kernfs_iattr_rwsem rcu_node_0 irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem &root->kernfs_rwsem &root->kernfs_iattr_rwsem &rcu_state.expedited_wq irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem &root->kernfs_rwsem &root->kernfs_iattr_rwsem &rcu_state.expedited_wq &p->pi_lock irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem &root->kernfs_rwsem &root->kernfs_iattr_rwsem &rcu_state.expedited_wq &p->pi_lock &rq->__lock irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem &root->kernfs_rwsem &root->kernfs_iattr_rwsem &rcu_state.expedited_wq &p->pi_lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 pernet_ops_rwsem devices_rwsem rcu_read_lock rcu_node_0 irq_context: 0 pernet_ops_rwsem devices_rwsem rcu_read_lock &rq->__lock irq_context: 0 pernet_ops_rwsem devices_rwsem rcu_read_lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 pernet_ops_rwsem rtnl_mutex kernfs_idr_lock &obj_hash[i].lock pool_lock irq_context: softirq (&mp->timer) &br->multicast_lock &____s->seqcount#2 irq_context: softirq (&mp->timer) &br->multicast_lock &____s->seqcount irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem remove_cache_srcu rcu_node_0 irq_context: 0 &mm->mmap_lock &mapping->i_mmap_rwsem &anon_vma->rwsem &base->lock irq_context: 0 &mm->mmap_lock &mapping->i_mmap_rwsem &anon_vma->rwsem &base->lock &obj_hash[i].lock irq_context: 0 sb_writers#4 &sb->s_type->i_mutex_key#8 &sbi->s_writepages_rwsem jbd2_handle &mapping->i_mmap_rwsem mmu_notifier_invalidate_range_start irq_context: 0 sb_writers#4 &sb->s_type->i_mutex_key#8 &sbi->s_writepages_rwsem jbd2_handle &mapping->i_mmap_rwsem mmu_notifier_invalidate_range_start &rq->__lock irq_context: 0 sb_writers#4 &sb->s_type->i_mutex_key#8 &sbi->s_writepages_rwsem jbd2_handle &mapping->i_mmap_rwsem mmu_notifier_invalidate_range_start &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 sb_writers#4 &sb->s_type->i_mutex_key#8 &sbi->s_writepages_rwsem jbd2_handle &mapping->i_mmap_rwsem ptlock_ptr(page)#2 irq_context: 0 sb_writers#4 &sb->s_type->i_mutex_key#8 &sbi->s_writepages_rwsem jbd2_handle &mapping->private_lock irq_context: 0 sb_writers#4 &sb->s_type->i_mutex_key#8 &sbi->s_writepages_rwsem jbd2_handle &mapping->private_lock rcu_read_lock &memcg->move_lock irq_context: 0 sb_writers#4 &sb->s_type->i_mutex_key#8 &sbi->s_writepages_rwsem jbd2_handle rcu_read_lock &rcu_state.gp_wq irq_context: 0 sb_writers#4 &sb->s_type->i_mutex_key#8 &sbi->s_writepages_rwsem jbd2_handle rcu_read_lock &rcu_state.gp_wq &p->pi_lock irq_context: 0 sb_writers#4 &sb->s_type->i_mutex_key#8 &sbi->s_writepages_rwsem jbd2_handle rcu_read_lock &rcu_state.gp_wq &p->pi_lock &rq->__lock irq_context: 0 sb_writers#4 &sb->s_type->i_mutex_key#8 &sbi->s_writepages_rwsem jbd2_handle rcu_read_lock &rcu_state.gp_wq &p->pi_lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 sb_writers#4 &sb->s_type->i_mutex_key#8 &sbi->s_writepages_rwsem &base->lock irq_context: 0 sb_writers#4 &sb->s_type->i_mutex_key#8 &sbi->s_writepages_rwsem &base->lock &obj_hash[i].lock irq_context: 0 sb_writers#4 &sb->s_type->i_mutex_key#8 &sbi->s_writepages_rwsem rcu_node_0 irq_context: 0 sb_writers#4 &sb->s_type->i_mutex_key#8 &sbi->s_writepages_rwsem rcu_read_lock &base->lock irq_context: 0 sb_writers#4 &sb->s_type->i_mutex_key#8 &sbi->s_writepages_rwsem rcu_read_lock &base->lock &obj_hash[i].lock irq_context: 0 sb_writers#4 &sb->s_type->i_mutex_key#8 &sbi->s_writepages_rwsem rcu_read_lock &rcu_state.gp_wq irq_context: 0 sb_writers#4 &sb->s_type->i_mutex_key#8 &sbi->s_writepages_rwsem rcu_read_lock &rcu_state.gp_wq &p->pi_lock irq_context: 0 sb_writers#4 &sb->s_type->i_mutex_key#8 &sbi->s_writepages_rwsem rcu_read_lock &rcu_state.gp_wq &p->pi_lock &rq->__lock irq_context: 0 sb_writers#4 &sb->s_type->i_mutex_key#8 &sbi->s_writepages_rwsem rcu_read_lock &rcu_state.gp_wq &p->pi_lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)bond6 (work_completion)(&(&bond->mii_work)->work) &rq->__lock irq_context: 0 sb_writers#4 &sb->s_type->i_mutex_key#8 &sbi->s_writepages_rwsem rcu_read_lock &memcg->move_lock irq_context: 0 sb_writers#4 &sb->s_type->i_mutex_key#8 &sbi->s_writepages_rwsem rcu_read_lock &xa->xa_lock#9 irq_context: 0 sb_writers#4 &sb->s_type->i_mutex_key#8 &sbi->s_writepages_rwsem rcu_read_lock &xa->xa_lock#9 &s->s_inode_wblist_lock irq_context: 0 sb_writers#4 &sb->s_type->i_mutex_key#8 &sbi->s_writepages_rwsem jbd2_handle &journal->j_wait_reserved irq_context: 0 sb_writers#4 &sb->s_type->i_mutex_key#8 &sbi->s_writepages_rwsem jbd2_handle &mapping->i_mmap_rwsem &rq->__lock irq_context: 0 sb_writers#4 &sb->s_type->i_mutex_key#8 &sbi->s_writepages_rwsem jbd2_handle &mapping->i_mmap_rwsem &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 sb_writers#4 &sb->s_type->i_mutex_key#8 &sbi->s_writepages_rwsem jbd2_handle &ei->i_data_sem &ei->i_prealloc_lock irq_context: 0 sb_writers#4 &sb->s_type->i_mutex_key#8 &sbi->s_writepages_rwsem jbd2_handle &ei->i_data_sem &ei->i_prealloc_lock &pa->pa_lock#2 irq_context: 0 sb_writers#4 &sb->s_type->i_mutex_key#8 &sbi->s_writepages_rwsem jbd2_handle &ei->i_data_sem &ei->i_es_lock &obj_hash[i].lock irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem &root->kernfs_rwsem rcu_read_lock &cfs_rq->removed.lock irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem &root->kernfs_rwsem rcu_read_lock &obj_hash[i].lock irq_context: 0 sb_writers#4 &sb->s_type->i_mutex_key#8 &sbi->s_writepages_rwsem jbd2_handle tk_core.seq.seqcount irq_context: 0 sb_writers#4 &sb->s_type->i_mutex_key#8 &sbi->s_writepages_rwsem jbd2_handle &dd->lock irq_context: 0 sb_writers#4 &sb->s_type->i_mutex_key#8 &sbi->s_writepages_rwsem jbd2_handle rcu_read_lock &dd->lock irq_context: 0 sb_writers#4 &sb->s_type->i_mutex_key#8 &sbi->s_writepages_rwsem jbd2_handle rcu_read_lock pool_lock#2 irq_context: 0 sb_writers#4 &sb->s_type->i_mutex_key#8 &sbi->s_writepages_rwsem jbd2_handle rcu_read_lock tk_core.seq.seqcount irq_context: 0 sb_writers#4 &sb->s_type->i_mutex_key#8 &sbi->s_writepages_rwsem jbd2_handle rcu_read_lock &virtscsi_vq->vq_lock irq_context: 0 sb_writers#4 &sb->s_type->i_mutex_key#8 &sbi->s_writepages_rwsem jbd2_handle rcu_read_lock &c->lock irq_context: 0 sb_writers#4 &sb->s_type->i_mutex_key#8 &sbi->s_writepages_rwsem rcu_read_lock &obj_hash[i].lock pool_lock irq_context: 0 cb_lock genl_mutex rcu_read_lock &ei->socket.wq.wait irq_context: 0 cb_lock genl_mutex rcu_read_lock &ei->socket.wq.wait &p->pi_lock irq_context: 0 cb_lock genl_mutex rcu_read_lock &ei->socket.wq.wait &p->pi_lock &rq->__lock irq_context: 0 cb_lock genl_mutex rcu_read_lock &ei->socket.wq.wait &p->pi_lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 cb_lock genl_mutex rtnl_mutex crngs.lock irq_context: 0 cb_lock genl_mutex rtnl_mutex &sdata->sec_mtx &sec->lock irq_context: 0 cb_lock genl_mutex rtnl_mutex &xa->xa_lock#3 irq_context: 0 cb_lock genl_mutex rtnl_mutex net_rwsem irq_context: 0 cb_lock genl_mutex rtnl_mutex net_rwsem &list->lock#2 irq_context: 0 cb_lock genl_mutex rtnl_mutex &tn->lock irq_context: 0 cb_lock genl_mutex rtnl_mutex &x->wait#9 irq_context: 0 cb_lock genl_mutex rtnl_mutex &k->list_lock irq_context: 0 cb_lock genl_mutex rtnl_mutex gdp_mutex irq_context: 0 cb_lock genl_mutex rtnl_mutex gdp_mutex &k->list_lock irq_context: 0 cb_lock genl_mutex rtnl_mutex lock irq_context: 0 cb_lock genl_mutex rtnl_mutex lock kernfs_idr_lock irq_context: 0 cb_lock genl_mutex rtnl_mutex lock kernfs_idr_lock pool_lock#2 irq_context: 0 cb_lock genl_mutex rtnl_mutex lock kernfs_idr_lock &c->lock irq_context: 0 cb_lock genl_mutex rtnl_mutex &root->kernfs_rwsem irq_context: 0 cb_lock genl_mutex rtnl_mutex &root->kernfs_rwsem &root->kernfs_iattr_rwsem irq_context: 0 cb_lock genl_mutex rtnl_mutex bus_type_sem irq_context: 0 cb_lock genl_mutex rtnl_mutex sysfs_symlink_target_lock irq_context: 0 cb_lock genl_mutex rtnl_mutex &root->kernfs_rwsem irq_context: 0 cb_lock genl_mutex rtnl_mutex &dev->power.lock irq_context: 0 cb_lock genl_mutex rtnl_mutex dpm_list_mtx irq_context: 0 cb_lock genl_mutex rtnl_mutex uevent_sock_mutex irq_context: 0 cb_lock genl_mutex rtnl_mutex uevent_sock_mutex fs_reclaim irq_context: 0 cb_lock genl_mutex rtnl_mutex uevent_sock_mutex fs_reclaim mmu_notifier_invalidate_range_start irq_context: 0 cb_lock genl_mutex rtnl_mutex uevent_sock_mutex fs_reclaim mmu_notifier_invalidate_range_start &rq->__lock irq_context: 0 cb_lock genl_mutex rtnl_mutex uevent_sock_mutex fs_reclaim mmu_notifier_invalidate_range_start &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 cb_lock genl_mutex rtnl_mutex uevent_sock_mutex pool_lock#2 irq_context: 0 cb_lock genl_mutex rtnl_mutex uevent_sock_mutex nl_table_lock irq_context: 0 cb_lock genl_mutex rtnl_mutex uevent_sock_mutex rlock-AF_NETLINK irq_context: 0 cb_lock genl_mutex rtnl_mutex uevent_sock_mutex rcu_read_lock &ei->socket.wq.wait irq_context: 0 cb_lock genl_mutex rtnl_mutex uevent_sock_mutex rcu_read_lock &ei->socket.wq.wait &ep->lock irq_context: 0 cb_lock genl_mutex rtnl_mutex uevent_sock_mutex rcu_read_lock &ei->socket.wq.wait &ep->lock &ep->wq irq_context: 0 cb_lock genl_mutex rtnl_mutex uevent_sock_mutex rcu_read_lock &ei->socket.wq.wait &ep->lock &ep->wq &p->pi_lock irq_context: 0 cb_lock genl_mutex rtnl_mutex uevent_sock_mutex rcu_read_lock &ei->socket.wq.wait &ep->lock &ep->wq &p->pi_lock &rq->__lock irq_context: 0 cb_lock genl_mutex rtnl_mutex uevent_sock_mutex rcu_read_lock &ei->socket.wq.wait &ep->lock &ep->wq &p->pi_lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 cb_lock genl_mutex rtnl_mutex uevent_sock_mutex rcu_read_lock rcu_node_0 irq_context: 0 cb_lock genl_mutex rtnl_mutex uevent_sock_mutex rcu_read_lock &rq->__lock irq_context: 0 cb_lock genl_mutex rtnl_mutex uevent_sock_mutex rcu_read_lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 cb_lock genl_mutex rtnl_mutex uevent_sock_mutex rcu_read_lock &rcu_state.gp_wq irq_context: 0 cb_lock genl_mutex rtnl_mutex uevent_sock_mutex rcu_read_lock &rcu_state.gp_wq &p->pi_lock irq_context: 0 cb_lock genl_mutex rtnl_mutex uevent_sock_mutex rcu_read_lock &rcu_state.gp_wq &p->pi_lock &rq->__lock irq_context: 0 cb_lock genl_mutex rtnl_mutex uevent_sock_mutex rcu_read_lock &rcu_state.gp_wq &p->pi_lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 cb_lock genl_mutex rtnl_mutex uevent_sock_mutex &obj_hash[i].lock irq_context: 0 cb_lock genl_mutex rtnl_mutex uevent_sock_mutex nl_table_wait.lock irq_context: 0 cb_lock genl_mutex rtnl_mutex &k->k_lock irq_context: 0 cb_lock genl_mutex rtnl_mutex subsys mutex#17 irq_context: 0 cb_lock genl_mutex rtnl_mutex subsys mutex#17 &k->k_lock irq_context: 0 cb_lock genl_mutex rtnl_mutex uevent_sock_mutex &c->lock irq_context: 0 cb_lock genl_mutex rtnl_mutex dev_hotplug_mutex irq_context: 0 cb_lock genl_mutex rtnl_mutex dev_hotplug_mutex &dev->power.lock irq_context: 0 cb_lock genl_mutex rtnl_mutex dev_base_lock irq_context: 0 cb_lock genl_mutex rtnl_mutex input_pool.lock irq_context: 0 cb_lock genl_mutex rtnl_mutex batched_entropy_u32.lock irq_context: 0 cb_lock genl_mutex rtnl_mutex &tbl->lock irq_context: 0 cb_lock genl_mutex rtnl_mutex sysctl_lock irq_context: 0 cb_lock genl_mutex rtnl_mutex nl_table_lock irq_context: 0 cb_lock genl_mutex rtnl_mutex nl_table_wait.lock irq_context: 0 cb_lock genl_mutex rtnl_mutex &pnettable->lock irq_context: 0 cb_lock genl_mutex rtnl_mutex smc_ib_devices.mutex irq_context: 0 cb_lock genl_mutex rtnl_mutex fs_reclaim mmu_notifier_invalidate_range_start &rq->__lock irq_context: 0 cb_lock genl_mutex rtnl_mutex fs_reclaim mmu_notifier_invalidate_range_start &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 cb_lock genl_mutex rtnl_mutex pcpu_alloc_mutex irq_context: 0 cb_lock genl_mutex rtnl_mutex pcpu_alloc_mutex pcpu_lock irq_context: 0 cb_lock genl_mutex rtnl_mutex rcu_read_lock &ei->socket.wq.wait irq_context: 0 cb_lock genl_mutex rtnl_mutex rcu_read_lock &ei->socket.wq.wait &p->pi_lock irq_context: 0 cb_lock genl_mutex rtnl_mutex rcu_read_lock &ei->socket.wq.wait &p->pi_lock &rq->__lock irq_context: 0 cb_lock genl_mutex rtnl_mutex rcu_read_lock &ei->socket.wq.wait &p->pi_lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 cb_lock genl_mutex rtnl_mutex &local->iflist_mtx#2 irq_context: 0 cb_lock genl_mutex rtnl_mutex rcu_state.exp_mutex rcu_node_0 irq_context: 0 cb_lock genl_mutex rtnl_mutex rcu_state.exp_mutex &obj_hash[i].lock irq_context: 0 cb_lock genl_mutex rtnl_mutex rcu_state.exp_mutex rcu_read_lock &pool->lock irq_context: 0 cb_lock genl_mutex rtnl_mutex rcu_state.exp_mutex rcu_read_lock &pool->lock &obj_hash[i].lock irq_context: 0 cb_lock genl_mutex rtnl_mutex rcu_state.exp_mutex rcu_read_lock &pool->lock &p->pi_lock irq_context: 0 cb_lock genl_mutex rtnl_mutex rcu_state.exp_mutex rcu_read_lock &pool->lock &p->pi_lock &rq->__lock irq_context: 0 cb_lock genl_mutex rtnl_mutex rcu_state.exp_mutex rcu_read_lock &pool->lock &p->pi_lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 cb_lock genl_mutex rtnl_mutex rcu_state.exp_mutex &rq->__lock irq_context: 0 cb_lock genl_mutex rtnl_mutex rcu_state.exp_mutex &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 cb_lock genl_mutex rtnl_mutex rcu_state.exp_mutex &rnp->exp_wq[0] irq_context: 0 cb_lock genl_mutex rtnl_mutex rcu_state.exp_mutex &rnp->exp_wq[1] irq_context: 0 sb_writers#4 &type->i_mutex_dir_key#3/1 sb_internal rcu_node_0 irq_context: 0 cb_lock genl_mutex rtnl_mutex dev_base_lock &xa->xa_lock#3 irq_context: 0 cb_lock genl_mutex rtnl_mutex cpu_hotplug_lock irq_context: 0 cb_lock genl_mutex rtnl_mutex cpu_hotplug_lock &list->lock#5 irq_context: 0 cb_lock genl_mutex rtnl_mutex rcu_state.exp_mutex &rnp->exp_wq[2] irq_context: 0 cb_lock genl_mutex rtnl_mutex bpf_devs_lock irq_context: 0 cb_lock genl_mutex rtnl_mutex rcu_state.exp_mutex &rnp->exp_wq[3] irq_context: 0 cb_lock genl_mutex rtnl_mutex &in_dev->mc_tomb_lock irq_context: 0 cb_lock genl_mutex rtnl_mutex sysctl_lock &obj_hash[i].lock irq_context: 0 cb_lock genl_mutex rtnl_mutex sysctl_lock pool_lock#2 irq_context: 0 cb_lock genl_mutex rtnl_mutex sysctl_lock krc.lock irq_context: 0 cb_lock genl_mutex rtnl_mutex class irq_context: 0 cb_lock genl_mutex rtnl_mutex (&tbl->proxy_timer) irq_context: 0 cb_lock genl_mutex rtnl_mutex &ul->lock irq_context: 0 cb_lock genl_mutex rtnl_mutex &net->xdp.lock irq_context: 0 cb_lock genl_mutex rtnl_mutex mirred_list_lock irq_context: 0 cb_lock genl_mutex rtnl_mutex &nft_net->commit_mutex irq_context: 0 cb_lock genl_mutex rtnl_mutex rcu_read_lock &tb->tb6_lock irq_context: 0 cb_lock genl_mutex rtnl_mutex rcu_read_lock &tb->tb6_lock &net->ipv6.fib6_walker_lock irq_context: 0 cb_lock genl_mutex rtnl_mutex &pnn->pndevs.lock irq_context: 0 cb_lock genl_mutex rtnl_mutex &pnn->routes.lock irq_context: 0 cb_lock genl_mutex rtnl_mutex target_list_lock irq_context: 0 cb_lock genl_mutex rtnl_mutex _xmit_IEEE802154 irq_context: 0 cb_lock genl_mutex rtnl_mutex &root->kernfs_rwsem kernfs_idr_lock irq_context: 0 cb_lock genl_mutex rtnl_mutex &root->kernfs_rwsem &obj_hash[i].lock irq_context: 0 cb_lock genl_mutex rtnl_mutex &root->kernfs_rwsem pool_lock#2 irq_context: 0 cb_lock genl_mutex rtnl_mutex kernfs_idr_lock irq_context: 0 cb_lock genl_mutex rtnl_mutex &root->kernfs_rwsem &rq->__lock irq_context: 0 cb_lock genl_mutex rtnl_mutex &n->list_lock irq_context: 0 cb_lock genl_mutex rtnl_mutex &n->list_lock &c->lock irq_context: 0 cb_lock genl_mutex rtnl_mutex uevent_sock_mutex &n->list_lock irq_context: 0 cb_lock genl_mutex rtnl_mutex uevent_sock_mutex &n->list_lock &c->lock irq_context: 0 cb_lock genl_mutex rtnl_mutex uevent_sock_mutex &rq->__lock irq_context: 0 cb_lock genl_mutex rtnl_mutex uevent_sock_mutex &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 cb_lock genl_mutex rtnl_mutex uevent_sock_mutex &____s->seqcount irq_context: 0 cb_lock genl_mutex rtnl_mutex uevent_sock_mutex rcu_read_lock pool_lock#2 irq_context: 0 cb_lock genl_mutex rtnl_mutex dev_hotplug_mutex &k->k_lock irq_context: 0 cb_lock genl_mutex rtnl_mutex dev_pm_qos_sysfs_mtx irq_context: 0 cb_lock genl_mutex rtnl_mutex dev_pm_qos_sysfs_mtx &root->kernfs_rwsem irq_context: 0 cb_lock genl_mutex rtnl_mutex dev_pm_qos_sysfs_mtx &root->kernfs_rwsem irq_context: 0 cb_lock genl_mutex rtnl_mutex dev_pm_qos_sysfs_mtx dev_pm_qos_mtx irq_context: 0 cb_lock genl_mutex rtnl_mutex &sem->wait_lock irq_context: 0 cb_lock genl_mutex rtnl_mutex &p->pi_lock irq_context: 0 cb_lock genl_mutex rtnl_mutex &p->pi_lock &rq->__lock irq_context: 0 cb_lock genl_mutex rtnl_mutex &p->pi_lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 cb_lock genl_mutex rtnl_mutex &k->k_lock klist_remove_lock irq_context: 0 cb_lock genl_mutex rtnl_mutex subsys mutex#17 &k->k_lock klist_remove_lock irq_context: 0 cb_lock genl_mutex rtnl_mutex &root->kernfs_rwsem &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 cb_lock genl_mutex rtnl_mutex &root->kernfs_rwsem &sem->wait_lock irq_context: 0 &type->i_mutex_dir_key#4 &dentry->d_lock &wq &p->pi_lock irq_context: 0 &type->i_mutex_dir_key#4 &dentry->d_lock &wq &p->pi_lock &rq->__lock irq_context: 0 cb_lock genl_mutex rtnl_mutex &root->kernfs_rwsem kernfs_idr_lock &obj_hash[i].lock irq_context: 0 &type->i_mutex_dir_key#4 &dentry->d_lock &wq &p->pi_lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 cb_lock genl_mutex rtnl_mutex &root->kernfs_rwsem kernfs_idr_lock pool_lock#2 irq_context: 0 cb_lock genl_mutex rtnl_mutex deferred_probe_mutex irq_context: 0 cb_lock genl_mutex rtnl_mutex device_links_lock irq_context: 0 cb_lock genl_mutex rtnl_mutex mmu_notifier_invalidate_range_start irq_context: 0 cb_lock genl_mutex rtnl_mutex uevent_sock_mutex mmu_notifier_invalidate_range_start irq_context: 0 cb_lock genl_mutex rtnl_mutex cpu_hotplug_lock xps_map_mutex irq_context: 0 cb_lock genl_mutex rtnl_mutex rcu_state.exp_mutex.wait_lock irq_context: 0 cb_lock genl_mutex rcu_state.barrier_mutex irq_context: 0 cb_lock genl_mutex rcu_state.barrier_mutex rcu_node_0 irq_context: 0 cb_lock genl_mutex rcu_state.barrier_mutex &rq->__lock irq_context: 0 cb_lock genl_mutex rcu_state.barrier_mutex &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 cb_lock genl_mutex rcu_state.barrier_mutex rcu_state.barrier_lock irq_context: 0 cb_lock genl_mutex rcu_state.barrier_mutex rcu_state.barrier_lock &obj_hash[i].lock irq_context: 0 cb_lock genl_mutex rcu_state.barrier_mutex &x->wait#24 irq_context: 0 cb_lock genl_mutex dev_base_lock irq_context: 0 cb_lock genl_mutex lweventlist_lock irq_context: 0 cb_lock genl_mutex krc.lock irq_context: 0 cb_lock genl_mutex &dir->lock#2 irq_context: 0 cb_lock genl_mutex &dir->lock#2 &obj_hash[i].lock irq_context: 0 cb_lock genl_mutex &dir->lock#2 pool_lock#2 irq_context: 0 cb_lock genl_mutex netdev_unregistering_wq.lock irq_context: 0 sb_writers#4 &type->i_mutex_dir_key#3/1 sb_internal rcu_read_lock pgd_lock irq_context: 0 sb_writers#4 &type->i_mutex_dir_key#3/1 sb_internal rcu_read_lock &obj_hash[i].lock irq_context: 0 sb_writers#4 &type->i_mutex_dir_key#3/1 sb_internal rcu_read_lock key irq_context: 0 sb_writers#4 &type->i_mutex_dir_key#3/1 sb_internal rcu_read_lock pcpu_lock irq_context: 0 sb_writers#4 &type->i_mutex_dir_key#3/1 sb_internal rcu_read_lock percpu_counters_lock irq_context: 0 cb_lock genl_mutex rtnl_mutex rcu_read_lock &ei->socket.wq.wait &p->pi_lock &cfs_rq->removed.lock irq_context: 0 cb_lock genl_mutex rtnl_mutex rcu_read_lock &rcu_state.gp_wq irq_context: 0 cb_lock genl_mutex rtnl_mutex rcu_read_lock &rcu_state.gp_wq &p->pi_lock irq_context: 0 cb_lock genl_mutex rtnl_mutex rcu_read_lock &rcu_state.gp_wq &p->pi_lock &rq->__lock irq_context: 0 cb_lock genl_mutex rtnl_mutex rcu_read_lock &rcu_state.gp_wq &p->pi_lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)tipc_send#5 irq_context: 0 (wq_completion)tipc_send#5 (work_completion)(&con->swork) irq_context: 0 (wq_completion)tipc_send#5 (work_completion)(&con->swork) &con->outqueue_lock irq_context: 0 (wq_completion)tipc_send#5 (work_completion)(&con->swork) &list->lock#38 irq_context: 0 (wq_completion)tipc_send#5 (work_completion)(&con->swork) slock-AF_TIPC &list->lock#38 irq_context: 0 (wq_completion)tipc_send#5 (work_completion)(&con->swork) &con->outqueue_lock &obj_hash[i].lock irq_context: 0 cb_lock genl_mutex rcu_read_lock &service->lock irq_context: 0 (wq_completion)tipc_send#5 (work_completion)(&con->swork) &c->lock irq_context: 0 (wq_completion)tipc_send#5 (work_completion)(&con->swork) slock-AF_TIPC &obj_hash[i].lock irq_context: 0 &f->f_pos_lock sb_writers#4 &sb->s_type->i_mutex_key#8 jbd2_handle &ei->i_data_sem &ei->i_es_lock &sbi->s_es_lock irq_context: 0 &f->f_pos_lock sb_writers#4 &sb->s_type->i_mutex_key#8 jbd2_handle &ei->i_data_sem &mapping->private_lock irq_context: 0 &f->f_pos_lock sb_writers#4 &sb->s_type->i_mutex_key#8 jbd2_handle &ei->i_data_sem rcu_node_0 irq_context: 0 sb_writers#5 &type->i_mutex_dir_key#5 rcu_read_lock &rcu_state.expedited_wq irq_context: 0 sb_writers#5 &type->i_mutex_dir_key#5 rcu_read_lock &rcu_state.expedited_wq &p->pi_lock irq_context: 0 sb_writers#5 &type->i_mutex_dir_key#5 rcu_read_lock &rcu_state.expedited_wq &p->pi_lock &rq->__lock irq_context: 0 sb_writers#5 &type->i_mutex_dir_key#5 rcu_read_lock &rcu_state.expedited_wq &p->pi_lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 cb_lock genl_mutex &p->pi_lock &cfs_rq->removed.lock irq_context: 0 &f->f_pos_lock sb_writers#4 &sb->s_type->i_mutex_key#8 jbd2_handle &ei->i_data_sem remove_cache_srcu irq_context: 0 &f->f_pos_lock sb_writers#4 &sb->s_type->i_mutex_key#8 jbd2_handle &ei->i_data_sem remove_cache_srcu quarantine_lock irq_context: 0 &f->f_pos_lock sb_writers#4 &sb->s_type->i_mutex_key#8 jbd2_handle &ei->i_data_sem remove_cache_srcu &rq->__lock irq_context: 0 &f->f_pos_lock sb_writers#4 &sb->s_type->i_mutex_key#8 jbd2_handle &ei->i_data_sem remove_cache_srcu &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 &f->f_pos_lock sb_writers#4 &sb->s_type->i_mutex_key#8 jbd2_handle &ei->i_data_sem remove_cache_srcu &c->lock irq_context: 0 &f->f_pos_lock sb_writers#4 &sb->s_type->i_mutex_key#8 jbd2_handle &ei->i_data_sem remove_cache_srcu &n->list_lock irq_context: 0 &f->f_pos_lock sb_writers#4 &sb->s_type->i_mutex_key#8 jbd2_handle &ei->i_data_sem remove_cache_srcu rcu_read_lock &____s->seqcount irq_context: 0 &f->f_pos_lock sb_writers#4 &sb->s_type->i_mutex_key#8 jbd2_handle &ei->i_data_sem remove_cache_srcu rcu_read_lock pool_lock#2 irq_context: 0 &f->f_pos_lock sb_writers#4 &sb->s_type->i_mutex_key#8 jbd2_handle &ei->i_data_sem remove_cache_srcu &obj_hash[i].lock irq_context: 0 &f->f_pos_lock sb_writers#4 &sb->s_type->i_mutex_key#8 jbd2_handle &ei->i_data_sem remove_cache_srcu pool_lock#2 irq_context: 0 &f->f_pos_lock sb_writers#4 &sb->s_type->i_mutex_key#8 jbd2_handle &ei->i_data_sem remove_cache_srcu rcu_read_lock rcu_node_0 irq_context: 0 &f->f_pos_lock sb_writers#4 &sb->s_type->i_mutex_key#8 jbd2_handle &ei->i_data_sem remove_cache_srcu rcu_read_lock &rq->__lock irq_context: 0 &f->f_pos_lock sb_writers#4 &sb->s_type->i_mutex_key#8 jbd2_handle &ei->i_data_sem remove_cache_srcu rcu_read_lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 &f->f_pos_lock sb_writers#4 &sb->s_type->i_mutex_key#8 jbd2_handle &ei->i_data_sem rcu_read_lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 &f->f_pos_lock sb_writers#4 &sb->s_type->i_mutex_key#8 jbd2_handle &ei->i_data_sem rcu_read_lock &rq->__lock &cfs_rq->removed.lock irq_context: 0 &f->f_pos_lock sb_writers#4 &sb->s_type->i_mutex_key#8 jbd2_handle &ei->i_data_sem mmu_notifier_invalidate_range_start &rq->__lock irq_context: 0 &f->f_pos_lock sb_writers#4 &sb->s_type->i_mutex_key#8 jbd2_handle &ei->i_data_sem mmu_notifier_invalidate_range_start &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 &f->f_pos_lock sb_writers#4 &sb->s_type->i_mutex_key#8 jbd2_handle &ei->i_data_sem rcu_read_lock &rcu_state.gp_wq irq_context: 0 &f->f_pos_lock sb_writers#4 &sb->s_type->i_mutex_key#8 jbd2_handle &ei->i_data_sem rcu_read_lock &rcu_state.gp_wq &p->pi_lock irq_context: 0 &f->f_pos_lock sb_writers#4 &sb->s_type->i_mutex_key#8 jbd2_handle &ei->i_data_sem rcu_read_lock &rcu_state.gp_wq &p->pi_lock &rq->__lock irq_context: 0 &f->f_pos_lock sb_writers#4 &sb->s_type->i_mutex_key#8 jbd2_handle &ei->i_data_sem rcu_read_lock &rcu_state.gp_wq &p->pi_lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 &f->f_pos_lock sb_writers#4 &sb->s_type->i_mutex_key#8 mapping.invalidate_lock jbd2_handle &ei->i_data_sem key#3 irq_context: 0 sb_writers#4 &type->i_mutex_dir_key#3/1 sb_internal jbd2_handle &sbi->s_orphan_lock &mapping->private_lock irq_context: 0 &f->f_pos_lock sb_writers#4 &sb->s_type->i_mutex_key#8 jbd2_handle &ei->i_data_sem rcu_read_lock &rcu_state.expedited_wq irq_context: 0 &f->f_pos_lock sb_writers#4 &sb->s_type->i_mutex_key#8 jbd2_handle &ei->i_data_sem rcu_read_lock &rcu_state.expedited_wq &p->pi_lock irq_context: 0 &f->f_pos_lock sb_writers#4 &sb->s_type->i_mutex_key#8 jbd2_handle &ei->i_data_sem rcu_read_lock &rcu_state.expedited_wq &p->pi_lock &rq->__lock irq_context: 0 &f->f_pos_lock sb_writers#4 &sb->s_type->i_mutex_key#8 jbd2_handle &ei->i_data_sem rcu_read_lock &rcu_state.expedited_wq &p->pi_lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 &f->f_pos_lock sb_writers#4 &sb->s_type->i_mutex_key#8 jbd2_handle &ei->i_data_sem quarantine_lock irq_context: 0 &f->f_pos_lock sb_writers#4 &sb->s_type->i_mutex_key#8 jbd2_handle &ei->i_data_sem stock_lock irq_context: 0 &f->f_pos_lock sb_writers#4 &sb->s_type->i_mutex_key#8 jbd2_handle &ei->i_data_sem &xa->xa_lock#9 irq_context: 0 &f->f_pos_lock sb_writers#4 &sb->s_type->i_mutex_key#8 jbd2_handle &ei->i_data_sem &xa->xa_lock#9 stock_lock irq_context: 0 &f->f_pos_lock sb_writers#4 &sb->s_type->i_mutex_key#8 jbd2_handle &ei->i_data_sem &xa->xa_lock#9 pool_lock#2 irq_context: 0 &f->f_pos_lock sb_writers#4 &sb->s_type->i_mutex_key#8 jbd2_handle &ei->i_data_sem &xa->xa_lock#9 &c->lock irq_context: 0 &f->f_pos_lock sb_writers#4 &sb->s_type->i_mutex_key#8 jbd2_handle &ei->i_data_sem lock#4 irq_context: 0 &f->f_pos_lock sb_writers#4 &sb->s_type->i_mutex_key#8 jbd2_handle &ei->i_data_sem &journal->j_revoke_lock irq_context: 0 &sb->s_type->i_mutex_key#10 sk_lock-AF_TIPC &____s->seqcount#2 irq_context: 0 &sb->s_type->i_mutex_key#10 sk_lock-AF_TIPC &____s->seqcount irq_context: 0 &f->f_pos_lock sb_writers#4 &sb->s_type->i_mutex_key#8 rcu_read_lock &pool->lock irq_context: 0 &f->f_pos_lock sb_writers#4 &sb->s_type->i_mutex_key#8 rcu_read_lock &pool->lock &obj_hash[i].lock irq_context: 0 &f->f_pos_lock sb_writers#4 &sb->s_type->i_mutex_key#8 rcu_read_lock &pool->lock pool_lock#2 irq_context: 0 &f->f_pos_lock sb_writers#4 &sb->s_type->i_mutex_key#8 rcu_read_lock &pool->lock &p->pi_lock irq_context: 0 &f->f_pos_lock sb_writers#4 &sb->s_type->i_mutex_key#8 rcu_read_lock &pool->lock &p->pi_lock &rq->__lock irq_context: 0 &f->f_pos_lock sb_writers#4 &sb->s_type->i_mutex_key#8 rcu_read_lock &pool->lock &p->pi_lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 &f->f_pos_lock sb_writers#4 &sb->s_type->i_mutex_key#8 jbd2_handle &ei->i_data_sem bit_wait_table + i irq_context: 0 &f->f_pos_lock sb_writers#4 &sb->s_type->i_mutex_key#8 jbd2_handle &ei->i_data_sem batched_entropy_u8.lock irq_context: 0 &f->f_pos_lock sb_writers#4 &sb->s_type->i_mutex_key#8 jbd2_handle &ei->i_data_sem kfence_freelist_lock irq_context: 0 &f->f_pos_lock sb_writers#4 &sb->s_type->i_mutex_key#8 jbd2_handle &ei->i_data_sem &meta->lock irq_context: 0 &f->f_pos_lock sb_writers#4 &sb->s_type->i_mutex_key#8 jbd2_handle &ei->i_data_sem &bgl->locks[i].lock irq_context: 0 &f->f_pos_lock sb_writers#4 &sb->s_type->i_mutex_key#8 rcu_read_lock &n->list_lock irq_context: 0 &f->f_pos_lock sb_writers#4 &sb->s_type->i_mutex_key#8 rcu_read_lock &n->list_lock &c->lock irq_context: 0 &f->f_pos_lock sb_writers#4 &sb->s_type->i_mutex_key#8 jbd2_handle &ei->i_data_sem &cfs_rq->removed.lock irq_context: 0 &f->f_pos_lock sb_writers#4 &sb->s_type->i_mutex_key#8 mapping.invalidate_lock &____s->seqcount#2 irq_context: 0 &f->f_pos_lock sb_writers#4 &sb->s_type->i_mutex_key#8 mapping.invalidate_lock &____s->seqcount irq_context: 0 &f->f_pos_lock sb_writers#4 &sb->s_type->i_mutex_key#8 mapping.invalidate_lock jbd2_handle &ei->i_data_sem &bgl->locks[i].lock irq_context: 0 &f->f_pos_lock sb_writers#4 &sb->s_type->i_mutex_key#8 mapping.invalidate_lock jbd2_handle &ei->i_data_sem &ret->b_state_lock &journal->j_list_lock irq_context: 0 &f->f_pos_lock sb_writers#4 &sb->s_type->i_mutex_key#8 mapping.invalidate_lock jbd2_handle &sbi->s_orphan_lock &ei->i_raw_lock irq_context: 0 &f->f_pos_lock sb_writers#4 &sb->s_type->i_mutex_key#8 rcu_read_lock &____s->seqcount#2 irq_context: 0 &f->f_pos_lock sb_writers#4 &sb->s_type->i_mutex_key#8 rcu_read_lock &____s->seqcount irq_context: 0 &f->f_pos_lock sb_writers#4 &sb->s_type->i_mutex_key#8 mapping.invalidate_lock jbd2_handle &ei->i_data_sem &bgl->locks[i].lock &obj_hash[i].lock irq_context: 0 &f->f_pos_lock sb_writers#4 &sb->s_type->i_mutex_key#8 mapping.invalidate_lock jbd2_handle &ei->i_data_sem &bgl->locks[i].lock pool_lock#2 irq_context: 0 &mm->mmap_lock fs_reclaim mmu_notifier_invalidate_range_start &cfs_rq->removed.lock irq_context: 0 &mm->mmap_lock fs_reclaim mmu_notifier_invalidate_range_start &obj_hash[i].lock irq_context: 0 &f->f_pos_lock sb_writers#4 &sb->s_type->i_mutex_key#8 &mapping->i_mmap_rwsem lock#4 &lruvec->lru_lock irq_context: 0 &f->f_pos_lock sb_writers#4 &sb->s_type->i_mutex_key#8 mapping.invalidate_lock jbd2_handle &ei->i_es_lock key#2 irq_context: 0 &f->f_pos_lock sb_writers#4 &sb->s_type->i_mutex_key#8 mapping.invalidate_lock jbd2_handle &ei->i_data_sem remove_cache_srcu irq_context: 0 &f->f_pos_lock sb_writers#4 &sb->s_type->i_mutex_key#8 mapping.invalidate_lock jbd2_handle &ei->i_data_sem remove_cache_srcu quarantine_lock irq_context: 0 &f->f_pos_lock sb_writers#4 &sb->s_type->i_mutex_key#8 mapping.invalidate_lock jbd2_handle &ei->i_data_sem remove_cache_srcu &c->lock irq_context: 0 &f->f_pos_lock sb_writers#4 &sb->s_type->i_mutex_key#8 mapping.invalidate_lock jbd2_handle &ei->i_data_sem remove_cache_srcu &n->list_lock irq_context: 0 &f->f_pos_lock sb_writers#4 &sb->s_type->i_mutex_key#8 mapping.invalidate_lock jbd2_handle &ei->i_data_sem remove_cache_srcu rcu_read_lock pool_lock#2 irq_context: 0 &f->f_pos_lock sb_writers#4 &sb->s_type->i_mutex_key#8 mapping.invalidate_lock jbd2_handle &ei->i_data_sem remove_cache_srcu &obj_hash[i].lock irq_context: 0 &f->f_pos_lock sb_writers#4 &sb->s_type->i_mutex_key#8 mapping.invalidate_lock jbd2_handle &ei->i_data_sem remove_cache_srcu pool_lock#2 irq_context: 0 &f->f_pos_lock sb_writers#4 &sb->s_type->i_mutex_key#8 mapping.invalidate_lock jbd2_handle &ei->i_data_sem remove_cache_srcu &rq->__lock irq_context: 0 &f->f_pos_lock sb_writers#4 &sb->s_type->i_mutex_key#8 mapping.invalidate_lock jbd2_handle &ei->i_data_sem remove_cache_srcu &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 sb_writers#4 &sb->s_type->i_mutex_key#8 mapping.invalidate_lock jbd2_handle &rcu_state.expedited_wq irq_context: 0 sb_writers#4 &sb->s_type->i_mutex_key#8 mapping.invalidate_lock jbd2_handle &rcu_state.expedited_wq &p->pi_lock irq_context: 0 sb_writers#4 &sb->s_type->i_mutex_key#8 mapping.invalidate_lock jbd2_handle &rcu_state.expedited_wq &p->pi_lock &rq->__lock irq_context: 0 sb_writers#4 &sb->s_type->i_mutex_key#8 mapping.invalidate_lock jbd2_handle &rcu_state.expedited_wq &p->pi_lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 &mm->mmap_lock &anon_vma->rwsem &rcu_state.expedited_wq irq_context: 0 &mm->mmap_lock &anon_vma->rwsem &rcu_state.expedited_wq &p->pi_lock irq_context: 0 &mm->mmap_lock &anon_vma->rwsem &rcu_state.expedited_wq &p->pi_lock &rq->__lock irq_context: 0 &mm->mmap_lock &anon_vma->rwsem &rcu_state.expedited_wq &p->pi_lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 sb_writers#4 &sb->s_type->i_mutex_key#8 &sbi->s_writepages_rwsem jbd2_handle mmu_notifier_invalidate_range_start irq_context: 0 sb_writers#4 &sb->s_type->i_mutex_key#8 &sbi->s_writepages_rwsem jbd2_handle stock_lock irq_context: 0 sb_writers#4 &sb->s_type->i_mutex_key#8 &sbi->s_writepages_rwsem jbd2_handle &____s->seqcount irq_context: 0 sb_writers#4 &sb->s_type->i_mutex_key#8 &sbi->s_writepages_rwsem jbd2_handle pool_lock#2 irq_context: 0 sb_writers#4 &sb->s_type->i_mutex_key#8 &sbi->s_writepages_rwsem jbd2_handle rcu_read_lock pool_lock#2 irq_context: 0 sb_writers#4 &sb->s_type->i_mutex_key#8 &sbi->s_writepages_rwsem jbd2_handle &obj_hash[i].lock irq_context: 0 sb_writers#4 &sb->s_type->i_mutex_key#8 &sbi->s_writepages_rwsem jbd2_handle &sb->s_type->i_lock_key#22 irq_context: 0 sb_writers#4 &sb->s_type->i_mutex_key#8 &sbi->s_writepages_rwsem jbd2_handle &s->s_inode_list_lock irq_context: 0 sb_writers#4 &sb->s_type->i_mutex_key#8 &sbi->s_writepages_rwsem jbd2_handle &mapping->private_lock irq_context: 0 sb_writers#4 &sb->s_type->i_mutex_key#8 &sbi->s_writepages_rwsem jbd2_handle &ret->b_state_lock irq_context: 0 sb_writers#4 &sb->s_type->i_mutex_key#8 &sbi->s_writepages_rwsem jbd2_handle &ret->b_state_lock &journal->j_list_lock irq_context: 0 sb_writers#4 &sb->s_type->i_mutex_key#8 &sbi->s_writepages_rwsem jbd2_handle &meta_group_info[i]->alloc_sem irq_context: 0 sb_writers#4 &sb->s_type->i_mutex_key#8 &sbi->s_writepages_rwsem jbd2_handle tk_core.seq.seqcount irq_context: 0 sb_writers#4 &sb->s_type->i_mutex_key#8 &sbi->s_writepages_rwsem jbd2_handle inode_hash_lock irq_context: 0 sb_writers#4 &sb->s_type->i_mutex_key#8 &sbi->s_writepages_rwsem jbd2_handle inode_hash_lock &sb->s_type->i_lock_key#22 irq_context: 0 sb_writers#4 &sb->s_type->i_mutex_key#8 &sbi->s_writepages_rwsem jbd2_handle batched_entropy_u32.lock irq_context: 0 sb_writers#4 &sb->s_type->i_mutex_key#8 &sbi->s_writepages_rwsem jbd2_handle &ei->i_raw_lock irq_context: 0 sb_writers#4 &sb->s_type->i_mutex_key#8 &sbi->s_writepages_rwsem jbd2_handle &journal->j_wait_updates irq_context: 0 sb_writers#4 &sb->s_type->i_mutex_key#8 &sbi->s_writepages_rwsem &ei->i_data_sem irq_context: 0 sb_writers#4 &sb->s_type->i_mutex_key#8 &sbi->s_writepages_rwsem jbd2_handle &ei->i_data_sem mmu_notifier_invalidate_range_start irq_context: 0 sb_writers#4 &sb->s_type->i_mutex_key#8 &sbi->s_writepages_rwsem jbd2_handle &ei->i_data_sem pool_lock#2 irq_context: 0 sb_writers#4 &sb->s_type->i_mutex_key#8 &sbi->s_writepages_rwsem jbd2_handle &ei->i_data_sem &journal->j_state_lock irq_context: 0 sb_writers#4 &sb->s_type->i_mutex_key#8 &sbi->s_writepages_rwsem jbd2_handle &ei->i_data_sem &ei->i_es_lock irq_context: 0 sb_writers#4 &sb->s_type->i_mutex_key#8 &sbi->s_writepages_rwsem jbd2_handle &ei->i_data_sem &ei->i_es_lock pool_lock#2 irq_context: 0 sb_writers#4 &sb->s_type->i_mutex_key#8 &sbi->s_writepages_rwsem jbd2_handle &ei->i_data_sem &ei->i_es_lock &sbi->s_es_lock irq_context: 0 sb_writers#4 &sb->s_type->i_mutex_key#8 &sbi->s_writepages_rwsem jbd2_handle &ei->i_data_sem &sb->s_type->i_lock_key#22 irq_context: 0 sb_writers#4 &sb->s_type->i_mutex_key#8 &sbi->s_writepages_rwsem jbd2_handle &journal->j_state_lock irq_context: 0 sb_writers#4 &sb->s_type->i_mutex_key#8 &sbi->s_writepages_rwsem jbd2_handle &journal->j_revoke_lock irq_context: 0 sb_writers#4 &sb->s_type->i_mutex_key#8 &sbi->s_writepages_rwsem jbd2_handle &bgl->locks[i].lock &sbi->s_md_lock irq_context: 0 sb_writers#4 &sb->s_type->i_mutex_key#8 &sbi->s_writepages_rwsem jbd2_handle &c->lock irq_context: 0 sb_writers#4 &sb->s_type->i_mutex_key#8 &sbi->s_writepages_rwsem jbd2_handle &bgl->locks[i].lock &obj_hash[i].lock irq_context: 0 sb_writers#4 &sb->s_type->i_mutex_key#8 &sbi->s_writepages_rwsem jbd2_handle &bgl->locks[i].lock pool_lock#2 irq_context: 0 sb_writers#4 &sb->s_type->i_mutex_key#8 &sbi->s_writepages_rwsem &sb->s_type->i_lock_key#22 irq_context: 0 sb_writers#4 &sb->s_type->i_mutex_key#8 &sbi->s_writepages_rwsem &s->s_inode_list_lock irq_context: 0 sb_writers#4 &sb->s_type->i_mutex_key#8 &sbi->s_writepages_rwsem sb_internal irq_context: 0 sb_writers#4 &sb->s_type->i_mutex_key#8 &sbi->s_writepages_rwsem sb_internal mmu_notifier_invalidate_range_start irq_context: 0 sb_writers#4 &sb->s_type->i_mutex_key#8 &sbi->s_writepages_rwsem sb_internal pool_lock#2 irq_context: 0 sb_writers#4 &sb->s_type->i_mutex_key#8 &sbi->s_writepages_rwsem sb_internal &journal->j_state_lock irq_context: 0 sb_writers#4 &sb->s_type->i_mutex_key#8 &sbi->s_writepages_rwsem sb_internal jbd2_handle irq_context: 0 sb_writers#4 &sb->s_type->i_mutex_key#8 &sbi->s_writepages_rwsem sb_internal jbd2_handle &ei->i_raw_lock irq_context: 0 sb_writers#4 &sb->s_type->i_mutex_key#8 &sbi->s_writepages_rwsem sb_internal jbd2_handle &xa->xa_lock#9 irq_context: 0 sb_writers#4 &sb->s_type->i_mutex_key#8 &sbi->s_writepages_rwsem sb_internal jbd2_handle &ei->i_prealloc_lock irq_context: 0 sb_writers#4 &sb->s_type->i_mutex_key#8 &sbi->s_writepages_rwsem sb_internal jbd2_handle &ei->i_es_lock irq_context: 0 sb_writers#4 &sb->s_type->i_mutex_key#8 &sbi->s_writepages_rwsem sb_internal jbd2_handle &ei->i_es_lock &obj_hash[i].lock irq_context: 0 sb_writers#4 &sb->s_type->i_mutex_key#8 &sbi->s_writepages_rwsem sb_internal jbd2_handle &ei->i_es_lock pool_lock#2 irq_context: 0 sb_writers#4 &sb->s_type->i_mutex_key#8 &sbi->s_writepages_rwsem sb_internal jbd2_handle &ei->i_es_lock &sbi->s_es_lock irq_context: 0 sb_writers#4 &sb->s_type->i_mutex_key#8 &sbi->s_writepages_rwsem sb_internal jbd2_handle &journal->j_wait_updates irq_context: 0 sb_writers#4 &sb->s_type->i_mutex_key#8 &sbi->s_writepages_rwsem sb_internal &obj_hash[i].lock irq_context: 0 sb_writers#4 &sb->s_type->i_mutex_key#8 &sbi->s_writepages_rwsem inode_hash_lock irq_context: 0 sb_writers#4 &sb->s_type->i_mutex_key#8 &sbi->s_writepages_rwsem inode_hash_lock &sb->s_type->i_lock_key#22 irq_context: 0 sb_writers#4 &sb->s_type->i_mutex_key#8 &sbi->s_writepages_rwsem &fsnotify_mark_srcu irq_context: 0 sb_writers#4 &type->i_mutex_dir_key#3/1 sb_internal jbd2_handle &sbi->s_orphan_lock &cfs_rq->removed.lock irq_context: 0 &sb->s_type->i_mutex_key#10 sk_lock-AF_PPPOX rcu_state.exp_mutex &rnp->exp_wq[0] irq_context: 0 rlock-AF_PHONET irq_context: 0 &disk->open_mutex &cfs_rq->removed.lock irq_context: 0 (wq_completion)bat_events (work_completion)(&(&forw_packet_aggr->delayed_work)->work) &hard_iface->bat_iv.ogm_buff_mutex &bat_priv->tt.commit_lock &bat_priv->softif_vlan_list_lock batched_entropy_u8.lock crngs.lock irq_context: 0 (wq_completion)events (work_completion)(&rdev->wiphy_work) &rdev->wiphy.mtx &wdev->mtx &local->sta_mtx rcu_read_lock hwsim_radio_lock pool_lock#2 irq_context: 0 (wq_completion)events (work_completion)(&rdev->wiphy_work) &rdev->wiphy.mtx &wdev->mtx &local->sta_mtx rcu_read_lock hwsim_radio_lock &c->lock irq_context: 0 (wq_completion)events (work_completion)(&rdev->wiphy_work) &rdev->wiphy.mtx &wdev->mtx &local->sta_mtx rcu_read_lock hwsim_radio_lock &n->list_lock irq_context: 0 (wq_completion)events (work_completion)(&rdev->wiphy_work) &rdev->wiphy.mtx &wdev->mtx &local->sta_mtx rcu_read_lock hwsim_radio_lock &n->list_lock &c->lock irq_context: 0 (wq_completion)events (work_completion)(&rdev->wiphy_work) &rdev->wiphy.mtx &wdev->mtx &local->sta_mtx rcu_read_lock hwsim_radio_lock &list->lock#19 irq_context: softirq rcu_read_lock &sta->rate_ctrl_lock &c->lock irq_context: 0 (wq_completion)events (work_completion)(&rdev->wiphy_work) &rdev->wiphy.mtx &wdev->mtx &local->sta_mtx rcu_read_lock rcu_node_0 irq_context: 0 (wq_completion)events (work_completion)(&rdev->wiphy_work) &rdev->wiphy.mtx &wdev->mtx &local->sta_mtx rcu_read_lock &rq->__lock irq_context: 0 (wq_completion)events (work_completion)(&rdev->wiphy_work) &rdev->wiphy.mtx &wdev->mtx &local->sta_mtx rcu_read_lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 rtnl_mutex &root->kernfs_rwsem &rq->__lock &cfs_rq->removed.lock irq_context: 0 hashlimit_mutex purge_vmap_area_lock irq_context: 0 hashlimit_mutex rcu_read_lock pool_lock#2 irq_context: 0 &ep->mtx uevent_sock_mutex nl_table_wait.lock &p->pi_lock irq_context: 0 &ep->mtx uevent_sock_mutex nl_table_wait.lock &p->pi_lock &cfs_rq->removed.lock irq_context: 0 &ep->mtx uevent_sock_mutex nl_table_wait.lock &p->pi_lock &rq->__lock irq_context: 0 &ep->mtx uevent_sock_mutex nl_table_wait.lock &p->pi_lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 &sb->s_type->i_mutex_key#10 &u->peer_wait &ei->socket.wq.wait &ep->lock &ep->wq irq_context: 0 &sb->s_type->i_mutex_key#10 &u->peer_wait &ei->socket.wq.wait &ep->lock &ep->wq &p->pi_lock irq_context: 0 &sb->s_type->i_mutex_key#10 &u->peer_wait &ei->socket.wq.wait &ep->lock &ep->wq &p->pi_lock &rq->__lock irq_context: 0 &sb->s_type->i_mutex_key#10 &u->peer_wait &ei->socket.wq.wait &ep->lock &ep->wq &p->pi_lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 &vma->vm_lock->lock fs_reclaim &cfs_rq->removed.lock irq_context: 0 &vma->vm_lock->lock fs_reclaim &obj_hash[i].lock irq_context: 0 rcu_read_lock rcu_read_lock_bh rcu_read_lock &____s->seqcount#12 irq_context: 0 rcu_read_lock rcu_read_lock_bh rcu_read_lock &ndev->lock irq_context: 0 &list->lock#34 rlock-AF_INET6 irq_context: 0 &f->f_pos_lock sb_writers#11 &mm->mmap_lock &c->lock irq_context: 0 &f->f_pos_lock sb_writers#11 &mm->mmap_lock &____s->seqcount#2 irq_context: 0 sb_writers#10 &type->i_mutex_dir_key#6 &root->kernfs_rwsem &xa->xa_lock#4 irq_context: 0 sb_writers#10 &type->i_mutex_dir_key#6 &root->kernfs_rwsem &obj_hash[i].lock irq_context: 0 sb_writers#10 &type->i_mutex_dir_key#6 &root->kernfs_rwsem stock_lock irq_context: 0 kn->active#48 stock_lock irq_context: 0 kn->active#48 &c->lock irq_context: 0 &type->i_mutex_dir_key#6 rename_lock.seqcount irq_context: 0 &type->i_mutex_dir_key#6 &dentry->d_lock &wq#2 irq_context: 0 cgroup_mutex krc.lock irq_context: 0 kn->active#48 &kernfs_locks->open_file_mutex[count] &c->lock irq_context: 0 kn->active#48 &kernfs_locks->open_file_mutex[count] &rq->__lock irq_context: 0 kn->active#48 &kernfs_locks->open_file_mutex[count] &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 &type->i_mutex_dir_key#6 &n->list_lock irq_context: 0 &type->i_mutex_dir_key#6 &n->list_lock &c->lock irq_context: 0 cgroup_mutex &obj_hash[i].lock pool_lock irq_context: 0 cgroup_mutex &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 cgroup_mutex cpu_hotplug_lock jump_label_mutex text_mutex &rq->__lock irq_context: 0 cgroup_mutex cpu_hotplug_lock jump_label_mutex text_mutex &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)ext4-rsv-conversion (work_completion)(&ei->i_rsv_conversion_work) jbd2_handle &ei->i_data_sem rcu_read_lock &cfs_rq->removed.lock irq_context: 0 (wq_completion)ext4-rsv-conversion (work_completion)(&ei->i_rsv_conversion_work) jbd2_handle &ei->i_data_sem rcu_read_lock &obj_hash[i].lock irq_context: 0 sb_writers#4 &sb->s_type->i_mutex_key#8 mapping.invalidate_lock batched_entropy_u8.lock irq_context: 0 sb_writers#4 &sb->s_type->i_mutex_key#8 mapping.invalidate_lock jbd2_handle &sbi->s_orphan_lock &cfs_rq->removed.lock irq_context: 0 sb_writers#4 &sb->s_type->i_mutex_key#8 mapping.invalidate_lock jbd2_handle &sbi->s_orphan_lock &obj_hash[i].lock irq_context: 0 tomoyo_ss mmu_notifier_invalidate_range_start &rq->__lock &cfs_rq->removed.lock irq_context: 0 kn->active#48 &kernfs_locks->open_file_mutex[count] &n->list_lock irq_context: 0 kn->active#48 &kernfs_locks->open_file_mutex[count] &n->list_lock &c->lock irq_context: 0 &kernfs_locks->open_file_mutex[count] &lock->wait_lock irq_context: 0 kn->active#48 &lock->wait_lock irq_context: 0 kn->active#48 &rq->__lock irq_context: 0 kn->active#48 &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 kn->active#48 &p->pi_lock irq_context: 0 kn->active#48 &p->pi_lock &rq->__lock irq_context: 0 kn->active#48 &p->pi_lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 vlan_ioctl_mutex &mm->mmap_lock &mm->page_table_lock irq_context: 0 vlan_ioctl_mutex &mm->mmap_lock ptlock_ptr(page) irq_context: 0 vlan_ioctl_mutex &mm->mmap_lock ptlock_ptr(page)#2 irq_context: 0 sb_writers#4 sb_internal jbd2_handle &obj_hash[i].lock pool_lock irq_context: 0 sb_writers#4 &type->i_mutex_dir_key#3 tomoyo_ss remove_cache_srcu rcu_read_lock pool_lock#2 irq_context: 0 sb_writers#7 rcu_node_0 irq_context: 0 (wq_completion)wg-crypt-wg1#2 (work_completion)(&({ do { const void *__vpp_verify = (typeof((worker) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __asm__ ("" : "=r"(__ptr) : "0"((typeof(*((worker))) *)((worker)))); (typeof((typeof(*((worker))) *)((worker)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); })->work) rcu_read_lock rcu_node_0 irq_context: 0 (wq_completion)wg-crypt-wg1#2 (work_completion)(&({ do { const void *__vpp_verify = (typeof((worker) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __asm__ ("" : "=r"(__ptr) : "0"((typeof(*((worker))) *)((worker)))); (typeof((typeof(*((worker))) *)((worker)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); })->work) rcu_read_lock &rq->__lock irq_context: 0 &u->iolock rcu_read_lock &ei->socket.wq.wait &p->pi_lock &rq->__lock &cfs_rq->removed.lock irq_context: 0 &sb->s_type->i_mutex_key#10 sk_lock-AF_TIPC &srv->idr_lock fill_pool_map-wait-type-override pool_lock#2 irq_context: 0 &sb->s_type->i_mutex_key#10 sk_lock-AF_TIPC &srv->idr_lock fill_pool_map-wait-type-override pool_lock irq_context: 0 sb_writers#4 &type->i_mutex_dir_key#3/1 sb_internal jbd2_handle &ei->i_data_sem rcu_read_lock pgd_lock irq_context: 0 sb_writers#4 &type->i_mutex_dir_key#3/1 sb_internal jbd2_handle &ei->i_data_sem rcu_read_lock stock_lock irq_context: 0 sb_writers#4 &type->i_mutex_dir_key#3/1 sb_internal jbd2_handle &ei->i_data_sem rcu_read_lock key irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem rtnl_mutex &tn->node_list_lock irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem (console_sem).lock irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem console_lock console_srcu console_owner_lock irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem console_lock console_srcu console_owner irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem console_lock console_srcu console_owner &port_lock_key irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem console_lock console_srcu console_owner console_owner_lock irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem &tn->nametbl_lock &service->lock irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem &tn->nametbl_lock &service->lock &obj_hash[i].lock irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem &tn->nametbl_lock &service->lock pool_lock#2 irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem &tn->nametbl_lock &service->lock krc.lock irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem &tn->nametbl_lock &obj_hash[i].lock irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem &tn->nametbl_lock pool_lock#2 irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem &tn->nametbl_lock krc.lock irq_context: 0 sb_writers#4 &type->i_mutex_dir_key#3/1 sb_internal jbd2_handle &ei->i_data_sem rcu_read_lock pcpu_lock irq_context: 0 sb_writers#4 &type->i_mutex_dir_key#3/1 sb_internal jbd2_handle &ei->i_data_sem rcu_read_lock percpu_counters_lock irq_context: 0 sb_writers#4 &type->i_mutex_dir_key#3/1 sb_internal jbd2_handle &ei->i_data_sem rcu_read_lock pcpu_lock stock_lock irq_context: 0 unix_gc_lock rlock-AF_UNIX irq_context: 0 unix_gc_lock unix_gc_wait.lock irq_context: 0 &u->peer_wait &p->pi_lock &rq->__lock &cfs_rq->removed.lock irq_context: 0 &f->f_pos_lock &type->i_mutex_dir_key#3 sb_writers#4 &rcu_state.expedited_wq irq_context: 0 kn->active#48 &n->list_lock irq_context: 0 kn->active#48 &n->list_lock &c->lock irq_context: 0 sk_lock-AF_ROSE &rq->__lock irq_context: 0 sk_lock-AF_ROSE &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem rtnl_mutex smc_ib_devices.mutex &rq->__lock irq_context: 0 &nft_net->commit_mutex &rnp->exp_wq[2] irq_context: 0 dev_map_lock irq_context: 0 sk_lock-AF_PACKET fs_reclaim mmu_notifier_invalidate_range_start &rq->__lock irq_context: 0 sk_lock-AF_PACKET fs_reclaim mmu_notifier_invalidate_range_start &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 sb_writers#4 &type->i_mutex_dir_key#3 jbd2_handle rcu_read_lock &rcu_state.expedited_wq irq_context: 0 sb_writers#4 &type->i_mutex_dir_key#3 jbd2_handle rcu_read_lock &rcu_state.expedited_wq &p->pi_lock irq_context: 0 sb_writers#4 &type->i_mutex_dir_key#3 jbd2_handle rcu_read_lock &rcu_state.expedited_wq &p->pi_lock &rq->__lock irq_context: 0 sb_writers#4 &type->i_mutex_dir_key#3 jbd2_handle rcu_read_lock &rcu_state.expedited_wq &p->pi_lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 sk_lock-AF_PACKET rcu_node_0 irq_context: 0 sk_lock-AF_PACKET rcu_read_lock rcu_node_0 irq_context: 0 sk_lock-AF_PACKET &pcp->lock &zone->lock irq_context: 0 sk_lock-AF_PACKET &pcp->lock &zone->lock &____s->seqcount irq_context: 0 sk_lock-AF_PACKET rcu_read_lock &rq->__lock irq_context: 0 sk_lock-AF_PACKET rcu_read_lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 sk_lock-AF_PACKET &po->pg_vec_lock rlock-AF_PACKET irq_context: 0 &mm->mmap_lock &po->pg_vec_lock &rq->__lock irq_context: 0 &mm->mmap_lock &po->pg_vec_lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 &mm->mmap_lock &po->pg_vec_lock rcu_read_lock rcu_node_0 irq_context: 0 &mm->mmap_lock &po->pg_vec_lock rcu_read_lock &rq->__lock irq_context: 0 &mm->mmap_lock &po->pg_vec_lock rcu_read_lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)wg-crypt-wg0 (work_completion)(&({ do { const void *__vpp_verify = (typeof((worker) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __asm__ ("" : "=r"(__ptr) : "0"((typeof(*((worker))) *)((worker)))); (typeof((typeof(*((worker))) *)((worker)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); })->work) rcu_read_lock rcu_node_0 irq_context: 0 (wq_completion)wg-crypt-wg0 (work_completion)(&({ do { const void *__vpp_verify = (typeof((worker) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __asm__ ("" : "=r"(__ptr) : "0"((typeof(*((worker))) *)((worker)))); (typeof((typeof(*((worker))) *)((worker)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); })->work) rcu_read_lock &rq->__lock irq_context: 0 sb_writers#4 sb_internal remove_cache_srcu rcu_read_lock &rcu_state.gp_wq irq_context: 0 sb_writers#4 sb_internal remove_cache_srcu rcu_read_lock &rcu_state.gp_wq &p->pi_lock irq_context: 0 sb_writers#4 sb_internal remove_cache_srcu rcu_read_lock &rcu_state.gp_wq &p->pi_lock &rq->__lock irq_context: 0 sb_writers#4 sb_internal remove_cache_srcu rcu_read_lock &rcu_state.gp_wq &p->pi_lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)wg-crypt-wg0#2 (work_completion)(&peer->transmit_packet_work) &peer->endpoint_lock rcu_read_lock_bh rcu_read_lock rcu_read_lock_bh rcu_read_lock pool_lock#2 irq_context: 0 &f->f_pos_lock sb_writers#10 (console_sem).lock irq_context: 0 &f->f_pos_lock sb_writers#10 console_lock console_srcu console_owner_lock irq_context: 0 &f->f_pos_lock sb_writers#10 console_lock console_srcu console_owner irq_context: 0 &f->f_pos_lock sb_writers#10 console_lock console_srcu console_owner &port_lock_key irq_context: 0 &f->f_pos_lock sb_writers#10 console_lock console_srcu console_owner console_owner_lock irq_context: 0 (wq_completion)events_unbound (work_completion)(&map->work) dev_map_lock irq_context: 0 (wq_completion)events_unbound (work_completion)(&map->work) rcu_state.barrier_mutex irq_context: 0 (wq_completion)events_unbound (work_completion)(&map->work) rcu_state.barrier_mutex rcu_state.barrier_lock irq_context: 0 (wq_completion)events_unbound (work_completion)(&map->work) rcu_state.barrier_mutex rcu_state.barrier_lock &obj_hash[i].lock irq_context: 0 (wq_completion)events_unbound (work_completion)(&map->work) rcu_state.barrier_mutex &x->wait#24 irq_context: 0 (wq_completion)events_unbound (work_completion)(&map->work) rcu_state.barrier_mutex &rq->__lock irq_context: 0 (wq_completion)events_unbound (work_completion)(&map->work) rcu_state.barrier_mutex &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 &sb->s_type->i_mutex_key#10 sk_lock-AF_PACKET rcu_state.exp_mutex &rnp->exp_wq[2] irq_context: 0 &f->f_pos_lock sb_writers#10 rcu_node_0 irq_context: 0 &f->f_pos_lock sb_writers#10 &rcu_state.expedited_wq irq_context: 0 &f->f_pos_lock sb_writers#10 &rcu_state.expedited_wq &p->pi_lock irq_context: 0 &f->f_pos_lock sb_writers#10 &rcu_state.expedited_wq &p->pi_lock &rq->__lock irq_context: 0 &f->f_pos_lock sb_writers#10 &rcu_state.expedited_wq &p->pi_lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 &sb->s_type->i_mutex_key#10 sk_lock-AF_PACKET &po->pg_vec_lock rlock-AF_PACKET irq_context: 0 &sb->s_type->i_mutex_key#10 sk_lock-AF_PACKET &pcp->lock &zone->lock irq_context: 0 &sb->s_type->i_mutex_key#10 sk_lock-AF_PACKET &pcp->lock &zone->lock &____s->seqcount irq_context: 0 (wq_completion)events_freezable &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 sk_lock-AF_PACKET rcu_read_lock &cfs_rq->removed.lock irq_context: 0 sk_lock-AF_PACKET rcu_read_lock &obj_hash[i].lock irq_context: 0 sk_lock-AF_PACKET &rq->__lock &cfs_rq->removed.lock irq_context: 0 sk_lock-AF_PACKET fs_reclaim &rq->__lock irq_context: 0 sk_lock-AF_PACKET fs_reclaim &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 sk_lock-AF_PACKET &rnp->exp_wq[3] irq_context: 0 &mm->mmap_lock &vma->vm_lock->lock rcu_node_0 irq_context: 0 &mm->mmap_lock &po->pg_vec_lock &c->lock irq_context: 0 &mm->mmap_lock &po->pg_vec_lock fs_reclaim mmu_notifier_invalidate_range_start &rq->__lock irq_context: 0 (wq_completion)wg-kex-wg0#7 (work_completion)(&peer->transmit_handshake_work) &peer->endpoint_lock rcu_read_lock_bh rcu_read_lock rcu_read_lock rcu_read_lock rcu_read_lock_bh rcu_read_lock pool_lock#2 irq_context: 0 &sb->s_type->i_mutex_key#10 sk_lock-AF_PACKET rcu_state.exp_mutex.wait_lock irq_context: 0 &sb->s_type->i_mutex_key#10 sk_lock-AF_PACKET &p->pi_lock irq_context: 0 &sb->s_type->i_mutex_key#10 sk_lock-AF_PACKET &p->pi_lock &rq->__lock irq_context: 0 &sb->s_type->i_mutex_key#10 sk_lock-AF_PACKET &p->pi_lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 &sb->s_type->i_mutex_key#10 sk_lock-AF_PACKET rcu_read_lock rcu_node_0 irq_context: 0 &sb->s_type->i_mutex_key#10 sk_lock-AF_PACKET rcu_read_lock &rcu_state.expedited_wq irq_context: 0 &sb->s_type->i_mutex_key#10 sk_lock-AF_PACKET rcu_read_lock &rcu_state.expedited_wq &p->pi_lock irq_context: 0 &sb->s_type->i_mutex_key#10 sk_lock-AF_PACKET rcu_read_lock &rcu_state.expedited_wq &p->pi_lock &rq->__lock irq_context: 0 &sb->s_type->i_mutex_key#10 sk_lock-AF_PACKET rcu_read_lock &rcu_state.expedited_wq &p->pi_lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 &sb->s_type->i_mutex_key#10 sk_lock-AF_PACKET rcu_read_lock &rq->__lock irq_context: 0 &sb->s_type->i_mutex_key#10 sk_lock-AF_PACKET rcu_read_lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 cb_lock rtnl_mutex &rdev->wiphy.mtx net_rwsem &rq->__lock irq_context: 0 cb_lock rtnl_mutex &rdev->wiphy.mtx net_rwsem &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 cb_lock rtnl_mutex &rdev->wiphy.mtx lock kernfs_idr_lock pool_lock#2 irq_context: 0 cb_lock rtnl_mutex &rdev->wiphy.mtx fs_reclaim &rq->__lock irq_context: 0 cb_lock rtnl_mutex &rdev->wiphy.mtx fs_reclaim &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 &sb->s_type->i_mutex_key#10 sk_lock-AF_PACKET &cfs_rq->removed.lock irq_context: 0 cb_lock rtnl_mutex &rdev->wiphy.mtx &idev->mc_lock &c->lock irq_context: 0 &sb->s_type->i_mutex_key#10 sk_lock-AF_PACKET rcu_read_lock &rcu_state.gp_wq irq_context: 0 &sb->s_type->i_mutex_key#10 sk_lock-AF_PACKET rcu_read_lock &rcu_state.gp_wq &p->pi_lock irq_context: 0 &sb->s_type->i_mutex_key#10 sk_lock-AF_PACKET rcu_read_lock &rcu_state.gp_wq &p->pi_lock &rq->__lock irq_context: 0 &sb->s_type->i_mutex_key#10 sk_lock-AF_PACKET rcu_read_lock &rcu_state.gp_wq &p->pi_lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)events (work_completion)(&rdev->destroy_work) rtnl_mutex &rdev->wiphy.mtx &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)events (work_completion)(&rdev->destroy_work) rtnl_mutex &rdev->wiphy.mtx uevent_sock_mutex &rq->__lock irq_context: 0 (wq_completion)events (work_completion)(&rdev->destroy_work) rtnl_mutex &rdev->wiphy.mtx uevent_sock_mutex nl_table_wait.lock &p->pi_lock irq_context: 0 (wq_completion)events (work_completion)(&rdev->destroy_work) rtnl_mutex &rdev->wiphy.mtx uevent_sock_mutex nl_table_wait.lock &p->pi_lock &rq->__lock irq_context: 0 (wq_completion)events (work_completion)(&rdev->destroy_work) rtnl_mutex &rdev->wiphy.mtx uevent_sock_mutex nl_table_wait.lock &p->pi_lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)events (work_completion)(&rdev->destroy_work) rtnl_mutex &rdev->wiphy.mtx &root->kernfs_rwsem &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)events (work_completion)(&rdev->destroy_work) rtnl_mutex &rdev->wiphy.mtx kernfs_idr_lock &obj_hash[i].lock irq_context: 0 (wq_completion)events (work_completion)(&rdev->destroy_work) rtnl_mutex &rdev->wiphy.mtx kernfs_idr_lock pool_lock#2 irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem fs_reclaim mmu_notifier_invalidate_range_start &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)netns net_cleanup_work &base->lock irq_context: 0 (wq_completion)netns net_cleanup_work &base->lock &obj_hash[i].lock irq_context: 0 cb_lock rtnl_mutex &rdev->wiphy.mtx &sb->s_type->i_mutex_key#3 &rq->__lock &cfs_rq->removed.lock irq_context: 0 cb_lock rtnl_mutex &rdev->wiphy.mtx &idev->mc_lock &n->list_lock irq_context: 0 cb_lock rtnl_mutex &rdev->wiphy.mtx &idev->mc_lock &n->list_lock &c->lock irq_context: 0 (wq_completion)events (work_completion)(&rdev->destroy_work) rtnl_mutex &rdev->wiphy.mtx rcu_read_lock rcu_node_0 irq_context: 0 (wq_completion)events (work_completion)(&rdev->destroy_work) rtnl_mutex &rdev->wiphy.mtx rcu_read_lock &rq->__lock irq_context: 0 (wq_completion)events (work_completion)(&rdev->destroy_work) rtnl_mutex &rdev->wiphy.mtx rcu_read_lock &rcu_state.gp_wq irq_context: 0 (wq_completion)events (work_completion)(&rdev->destroy_work) rtnl_mutex &rdev->wiphy.mtx rcu_read_lock &rcu_state.gp_wq &p->pi_lock irq_context: 0 (wq_completion)events (work_completion)(&rdev->destroy_work) rtnl_mutex &rdev->wiphy.mtx rcu_read_lock &rcu_state.gp_wq &p->pi_lock &rq->__lock irq_context: 0 (wq_completion)events (work_completion)(&rdev->destroy_work) rtnl_mutex &rdev->wiphy.mtx rcu_read_lock &rcu_state.gp_wq &p->pi_lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)events (work_completion)(&rdev->destroy_work) rtnl_mutex &rdev->wiphy.mtx rcu_read_lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)events (work_completion)(&rdev->destroy_work) rtnl_mutex &rdev->wiphy.mtx &____s->seqcount#2 irq_context: 0 (wq_completion)events (work_completion)(&rdev->destroy_work) rtnl_mutex &rdev->wiphy.mtx uevent_sock_mutex &c->lock irq_context: 0 (wq_completion)events (work_completion)(&rdev->destroy_work) rtnl_mutex &rdev->wiphy.mtx uevent_sock_mutex &n->list_lock irq_context: 0 (wq_completion)events (work_completion)(&rdev->destroy_work) rtnl_mutex &rdev->wiphy.mtx uevent_sock_mutex &n->list_lock &c->lock irq_context: 0 kn->active#54 stock_lock irq_context: 0 kn->active#54 &rq->__lock irq_context: 0 (wq_completion)events (work_completion)(&rdev->destroy_work) rtnl_mutex &rdev->wiphy.mtx &root->kernfs_rwsem &root->kernfs_iattr_rwsem &rq->__lock irq_context: 0 kn->active#54 &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)events (work_completion)(&rdev->destroy_work) rtnl_mutex &rdev->wiphy.mtx &root->kernfs_rwsem &root->kernfs_iattr_rwsem &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)events (work_completion)(&rdev->destroy_work) rtnl_mutex &rdev->wiphy.mtx &root->kernfs_rwsem kernfs_idr_lock &obj_hash[i].lock irq_context: 0 (wq_completion)events (work_completion)(&rdev->destroy_work) rtnl_mutex &rdev->wiphy.mtx &root->kernfs_rwsem kernfs_idr_lock pool_lock#2 irq_context: 0 sk_lock-AF_INET &tcp_hashinfo.bhash[i].lock per_cpu_ptr(&cgroup_rstat_cpu_lock, cpu) irq_context: 0 &sb->s_type->i_mutex_key#10 cpu_hotplug_lock &rq->__lock irq_context: 0 cb_lock rtnl_mutex &rdev->wiphy.mtx &root->kernfs_rwsem &rq->__lock irq_context: 0 cb_lock rtnl_mutex &rdev->wiphy.mtx &root->kernfs_rwsem &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 kn->active#54 &n->list_lock irq_context: 0 kn->active#54 &n->list_lock &c->lock irq_context: 0 kn->active#54 remove_cache_srcu irq_context: 0 kn->active#54 remove_cache_srcu quarantine_lock irq_context: 0 kn->active#54 remove_cache_srcu &c->lock irq_context: 0 kn->active#54 remove_cache_srcu &n->list_lock irq_context: 0 kn->active#54 remove_cache_srcu &obj_hash[i].lock irq_context: 0 &sb->s_type->i_mutex_key#10 &smc->clcsock_release_lock rcu_read_lock rcu_node_0 irq_context: 0 &sb->s_type->i_mutex_key#10 &smc->clcsock_release_lock rcu_read_lock &rq->__lock irq_context: 0 &sb->s_type->i_mutex_key#10 &smc->clcsock_release_lock rcu_read_lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)events (work_completion)(&rdev->destroy_work) rtnl_mutex &rdev->wiphy.mtx sysctl_lock &obj_hash[i].lock pool_lock irq_context: 0 (wq_completion)events (work_completion)(&rdev->destroy_work) rtnl_mutex &rdev->wiphy.mtx &idev->mc_lock &rq->__lock irq_context: 0 (wq_completion)events (work_completion)(&rdev->destroy_work) rtnl_mutex &rdev->wiphy.mtx &idev->mc_lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)events (work_completion)(&rdev->destroy_work) rtnl_mutex &rdev->wiphy.mtx &sb->s_type->i_mutex_key#3 &fsnotify_mark_srcu &rq->__lock irq_context: 0 (wq_completion)events (work_completion)(&rdev->destroy_work) rtnl_mutex &rdev->wiphy.mtx &sb->s_type->i_mutex_key#3 &fsnotify_mark_srcu &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)events (work_completion)(&rdev->destroy_work) rtnl_mutex &rdev->wiphy.mtx uevent_sock_mutex &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)events (work_completion)(&rdev->destroy_work) rtnl_mutex &rdev->wiphy.mtx &sem->wait_lock irq_context: 0 (wq_completion)events (work_completion)(&rdev->destroy_work) rtnl_mutex &rdev->wiphy.mtx &p->pi_lock irq_context: 0 (wq_completion)events (work_completion)(&rdev->destroy_work) rtnl_mutex &rdev->wiphy.mtx &p->pi_lock &rq->__lock irq_context: 0 (wq_completion)events (work_completion)(&rdev->destroy_work) rtnl_mutex &rdev->wiphy.mtx &p->pi_lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)events (work_completion)(&rdev->destroy_work) krc.lock &obj_hash[i].lock irq_context: 0 (wq_completion)events (work_completion)(&rdev->destroy_work) krc.lock &base->lock irq_context: 0 (wq_completion)events (work_completion)(&rdev->destroy_work) krc.lock &base->lock &obj_hash[i].lock irq_context: 0 &mm->mmap_lock &xa->xa_lock#9 batched_entropy_u8.lock crngs.lock irq_context: softirq (&n->timer) &____s->seqcount#2 irq_context: softirq (&n->timer) &____s->seqcount irq_context: 0 (wq_completion)wg-kex-wg0#12 (work_completion)(&({ do { const void *__vpp_verify = (typeof((worker) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __asm__ ("" : "=r"(__ptr) : "0"((typeof(*((worker))) *)((worker)))); (typeof((typeof(*((worker))) *)((worker)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); })->work) &handshake->lock &n->list_lock irq_context: 0 (wq_completion)wg-kex-wg0#12 (work_completion)(&({ do { const void *__vpp_verify = (typeof((worker) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __asm__ ("" : "=r"(__ptr) : "0"((typeof(*((worker))) *)((worker)))); (typeof((typeof(*((worker))) *)((worker)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); })->work) &handshake->lock &n->list_lock &c->lock irq_context: 0 (wq_completion)wg-kex-wg0#12 (work_completion)(&({ do { const void *__vpp_verify = (typeof((worker) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __asm__ ("" : "=r"(__ptr) : "0"((typeof(*((worker))) *)((worker)))); (typeof((typeof(*((worker))) *)((worker)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); })->work) &handshake->lock &rq->__lock irq_context: 0 (wq_completion)wg-kex-wg1#12 (work_completion)(&({ do { const void *__vpp_verify = (typeof((worker) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __asm__ ("" : "=r"(__ptr) : "0"((typeof(*((worker))) *)((worker)))); (typeof((typeof(*((worker))) *)((worker)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); })->work) &wg->static_identity.lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)wg-kex-wg0#4 (work_completion)(&({ do { const void *__vpp_verify = (typeof((worker) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __asm__ ("" : "=r"(__ptr) : "0"((typeof(*((worker))) *)((worker)))); (typeof((typeof(*((worker))) *)((worker)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); })->work) &wg->static_identity.lock &handshake->lock &rq->__lock irq_context: 0 (wq_completion)wg-kex-wg1#4 (work_completion)(&({ do { const void *__vpp_verify = (typeof((worker) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __asm__ ("" : "=r"(__ptr) : "0"((typeof(*((worker))) *)((worker)))); (typeof((typeof(*((worker))) *)((worker)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); })->work) &handshake->lock &n->list_lock irq_context: 0 (wq_completion)wg-kex-wg1#4 (work_completion)(&({ do { const void *__vpp_verify = (typeof((worker) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __asm__ ("" : "=r"(__ptr) : "0"((typeof(*((worker))) *)((worker)))); (typeof((typeof(*((worker))) *)((worker)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); })->work) &handshake->lock &n->list_lock &c->lock irq_context: 0 (wq_completion)wg-kex-wg1#4 (work_completion)(&({ do { const void *__vpp_verify = (typeof((worker) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __asm__ ("" : "=r"(__ptr) : "0"((typeof(*((worker))) *)((worker)))); (typeof((typeof(*((worker))) *)((worker)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); })->work) &handshake->lock &rq->__lock irq_context: 0 (wq_completion)wg-kex-wg1#4 (work_completion)(&({ do { const void *__vpp_verify = (typeof((worker) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __asm__ ("" : "=r"(__ptr) : "0"((typeof(*((worker))) *)((worker)))); (typeof((typeof(*((worker))) *)((worker)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); })->work) &handshake->lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 sb_writers#4 &type->i_mutex_dir_key#3/1 &type->i_mutex_dir_key#3 jbd2_handle &journal->j_wait_updates &p->pi_lock &cfs_rq->removed.lock irq_context: 0 pernet_ops_rwsem rtnl_mutex &idev->mc_lock &cfs_rq->removed.lock irq_context: 0 sb_writers#4 &type->i_mutex_dir_key#3/1 tomoyo_ss remove_cache_srcu &rcu_state.expedited_wq irq_context: 0 sb_writers#4 &type->i_mutex_dir_key#3/1 tomoyo_ss remove_cache_srcu &rcu_state.expedited_wq &p->pi_lock irq_context: 0 sb_writers#4 &type->i_mutex_dir_key#3/1 tomoyo_ss remove_cache_srcu &rcu_state.expedited_wq &p->pi_lock &rq->__lock irq_context: 0 sb_writers#4 &type->i_mutex_dir_key#3/1 tomoyo_ss remove_cache_srcu &rcu_state.expedited_wq &p->pi_lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 &f->f_pos_lock sb_writers#4 &sb->s_type->i_mutex_key#8 jbd2_handle &ei->i_data_sem &n->list_lock irq_context: 0 &f->f_pos_lock sb_writers#4 &sb->s_type->i_mutex_key#8 jbd2_handle &ei->i_data_sem &n->list_lock &c->lock irq_context: 0 (wq_completion)mld (work_completion)(&(&idev->mc_ifc_work)->work) &idev->mc_lock remove_cache_srcu rcu_node_0 irq_context: 0 (wq_completion)mld (work_completion)(&(&idev->mc_ifc_work)->work) &idev->mc_lock remove_cache_srcu &rcu_state.expedited_wq irq_context: 0 (wq_completion)mld (work_completion)(&(&idev->mc_ifc_work)->work) &idev->mc_lock remove_cache_srcu &rcu_state.expedited_wq &p->pi_lock irq_context: 0 (wq_completion)mld (work_completion)(&(&idev->mc_ifc_work)->work) &idev->mc_lock remove_cache_srcu &rcu_state.expedited_wq &p->pi_lock &rq->__lock irq_context: 0 (wq_completion)mld (work_completion)(&(&idev->mc_ifc_work)->work) &idev->mc_lock remove_cache_srcu &rcu_state.expedited_wq &p->pi_lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 &f->f_pos_lock sb_writers#4 &sb->s_type->i_mutex_key#8 jbd2_handle &ei->i_data_sem mmu_notifier_invalidate_range_start &cfs_rq->removed.lock irq_context: 0 &f->f_pos_lock sb_writers#4 &sb->s_type->i_mutex_key#8 jbd2_handle &ei->i_data_sem mmu_notifier_invalidate_range_start &obj_hash[i].lock irq_context: 0 sb_writers#4 sb_internal jbd2_handle &ei->i_data_sem &bgl->locks[i].lock &____s->seqcount irq_context: 0 sb_writers#4 sb_internal jbd2_handle &ei->i_data_sem &bgl->locks[i].lock rcu_read_lock pool_lock#2 irq_context: 0 sb_writers#4 &sb->s_type->i_mutex_key#8 &sbi->s_writepages_rwsem rcu_state.exp_mutex fill_pool_map-wait-type-override pool_lock#2 irq_context: 0 sb_writers#4 &sb->s_type->i_mutex_key#8 &sbi->s_writepages_rwsem rcu_state.exp_mutex fill_pool_map-wait-type-override &c->lock irq_context: 0 sb_writers#4 &sb->s_type->i_mutex_key#8 &sbi->s_writepages_rwsem rcu_state.exp_mutex fill_pool_map-wait-type-override &n->list_lock irq_context: 0 sb_writers#4 &sb->s_type->i_mutex_key#8 &sbi->s_writepages_rwsem rcu_state.exp_mutex fill_pool_map-wait-type-override &n->list_lock &c->lock irq_context: 0 sb_writers#4 &sb->s_type->i_mutex_key#8 &sbi->s_writepages_rwsem rcu_state.exp_mutex fill_pool_map-wait-type-override &rq->__lock irq_context: 0 sb_writers#4 &sb->s_type->i_mutex_key#8 &sbi->s_writepages_rwsem rcu_state.exp_mutex fill_pool_map-wait-type-override &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 sb_writers#4 &sb->s_type->i_mutex_key#8 &sbi->s_writepages_rwsem rcu_state.exp_mutex fill_pool_map-wait-type-override pool_lock irq_context: 0 &nft_net->commit_mutex rcu_state.exp_mutex rcu_read_lock &rq->__lock irq_context: 0 &nft_net->commit_mutex rcu_state.exp_mutex rcu_read_lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 &mm->mmap_lock sb_pagefaults jbd2_handle &journal->j_wait_updates &p->pi_lock &cfs_rq->removed.lock irq_context: 0 (wq_completion)kblockd rcu_node_0 irq_context: 0 sb_writers#4 &type->i_mutex_dir_key#3 fill_pool_map-wait-type-override &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 sb_writers#4 &type->i_mutex_dir_key#3/1 &sb->s_type->i_mutex_key#8 jbd2_handle rcu_read_lock &rcu_state.expedited_wq irq_context: 0 sb_writers#4 &type->i_mutex_dir_key#3/1 &sb->s_type->i_mutex_key#8 jbd2_handle rcu_read_lock &rcu_state.expedited_wq &p->pi_lock irq_context: 0 sb_writers#4 &type->i_mutex_dir_key#3/1 &sb->s_type->i_mutex_key#8 jbd2_handle rcu_read_lock &rcu_state.expedited_wq &p->pi_lock &rq->__lock irq_context: 0 sb_writers#4 &type->i_mutex_dir_key#3/1 &sb->s_type->i_mutex_key#8 jbd2_handle rcu_read_lock &rcu_state.expedited_wq &p->pi_lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 &type->s_umount_key#23/1 &cfs_rq->removed.lock irq_context: 0 rtnl_mutex nf_hook_mutex stock_lock irq_context: 0 sb_writers#4 &sb->s_type->i_lock_key#22 &xa->xa_lock#9 fill_pool_map-wait-type-override &n->list_lock irq_context: 0 sb_writers#4 &sb->s_type->i_lock_key#22 &xa->xa_lock#9 fill_pool_map-wait-type-override &n->list_lock &c->lock irq_context: 0 dup_mmap_sem &mm->mmap_lock &mm->mmap_lock/1 rcu_read_lock &rq->__lock &obj_hash[i].lock irq_context: 0 dup_mmap_sem &mm->mmap_lock &mm->mmap_lock/1 rcu_read_lock &rq->__lock &base->lock irq_context: 0 dup_mmap_sem &mm->mmap_lock &mm->mmap_lock/1 rcu_read_lock &rq->__lock &base->lock &obj_hash[i].lock irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem ovs_mutex nf_ct_proto_mutex defrag4_mutex nf_hook_mutex fs_reclaim irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem ovs_mutex nf_ct_proto_mutex defrag4_mutex nf_hook_mutex fs_reclaim mmu_notifier_invalidate_range_start irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem ovs_mutex nf_ct_proto_mutex defrag4_mutex nf_hook_mutex fs_reclaim mmu_notifier_invalidate_range_start &rq->__lock irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem ovs_mutex nf_ct_proto_mutex defrag4_mutex nf_hook_mutex fs_reclaim mmu_notifier_invalidate_range_start &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem ovs_mutex nf_ct_proto_mutex defrag4_mutex nf_hook_mutex pool_lock#2 irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem ovs_mutex nf_ct_proto_mutex defrag6_mutex nf_hook_mutex fs_reclaim irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem ovs_mutex nf_ct_proto_mutex defrag6_mutex nf_hook_mutex fs_reclaim mmu_notifier_invalidate_range_start irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem ovs_mutex nf_ct_proto_mutex defrag6_mutex nf_hook_mutex pool_lock#2 irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem ovs_mutex nf_ct_proto_mutex &obj_hash[i].lock pool_lock irq_context: 0 (wq_completion)wg-crypt-wg0#2 (work_completion)(&({ do { const void *__vpp_verify = (typeof((worker) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __asm__ ("" : "=r"(__ptr) : "0"((typeof(*((worker))) *)((worker)))); (typeof((typeof(*((worker))) *)((worker)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); })->work) rcu_read_lock rcu_node_0 irq_context: 0 sb_writers#4 sb_internal jbd2_handle &ret->b_state_lock bit_wait_table + i &p->pi_lock irq_context: 0 sb_writers#4 sb_internal jbd2_handle &ret->b_state_lock bit_wait_table + i &p->pi_lock &rq->__lock irq_context: 0 sb_writers#4 sb_internal jbd2_handle &ret->b_state_lock bit_wait_table + i &p->pi_lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)wg-crypt-wg0#2 (work_completion)(&({ do { const void *__vpp_verify = (typeof((worker) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __asm__ ("" : "=r"(__ptr) : "0"((typeof(*((worker))) *)((worker)))); (typeof((typeof(*((worker))) *)((worker)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); })->work) rcu_read_lock &rq->__lock irq_context: 0 &mm->mmap_lock rcu_read_lock &rcu_state.expedited_wq &p->pi_lock &rq->__lock &cfs_rq->removed.lock irq_context: 0 (wq_completion)events (work_completion)(&(&krcp->krw_arr[i].rcu_work)->work) rcu_callback &cfs_rq->removed.lock irq_context: 0 sb_writers#4 sb_internal &base->lock irq_context: 0 sb_writers#4 sb_internal &base->lock &obj_hash[i].lock irq_context: 0 (wq_completion)wg-crypt-wg1#5 (work_completion)(&peer->transmit_packet_work) &peer->endpoint_lock rcu_read_lock_bh rcu_read_lock batched_entropy_u8.lock irq_context: 0 (wq_completion)wg-crypt-wg1#5 (work_completion)(&peer->transmit_packet_work) &peer->endpoint_lock rcu_read_lock_bh rcu_read_lock kfence_freelist_lock irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem rtnl_mutex &br->hash_lock irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem rtnl_mutex &br->hash_lock rcu_read_lock rhashtable_bucket irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem rtnl_mutex &br->hash_lock rcu_read_lock rcu_read_lock &pool->lock irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem rtnl_mutex &br->hash_lock rcu_read_lock rcu_read_lock &pool->lock &obj_hash[i].lock irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem rtnl_mutex &br->hash_lock rcu_read_lock rcu_read_lock &pool->lock pool_lock#2 irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem rtnl_mutex &br->hash_lock rcu_read_lock rcu_read_lock &pool->lock &p->pi_lock irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem rtnl_mutex &br->hash_lock rcu_read_lock rcu_read_lock &pool->lock &p->pi_lock &rq->__lock irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem rtnl_mutex &br->hash_lock rcu_read_lock rcu_read_lock &pool->lock &p->pi_lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem rtnl_mutex &br->hash_lock pool_lock#2 irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem rtnl_mutex &br->hash_lock &c->lock irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem rtnl_mutex &br->hash_lock nl_table_lock irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem rtnl_mutex &br->hash_lock &obj_hash[i].lock irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem rtnl_mutex &br->hash_lock nl_table_wait.lock irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem rtnl_mutex (work_completion)(&(&br->gc_work)->work) irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem rtnl_mutex rcu_read_lock &pool->lock irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem rtnl_mutex &br->lock irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem rtnl_mutex (&br->hello_timer) irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem rtnl_mutex (&br->topology_change_timer) irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem rtnl_mutex (&br->tcn_timer) irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem rtnl_mutex (&brmctx->ip4_mc_router_timer) irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem rtnl_mutex (&brmctx->ip4_other_query.timer) irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem rtnl_mutex (&brmctx->ip4_own_query.timer) irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem rtnl_mutex (&brmctx->ip6_mc_router_timer) irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem rtnl_mutex (&brmctx->ip6_other_query.timer) irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem rtnl_mutex (&brmctx->ip6_own_query.timer) irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem rtnl_mutex &bridge_netdev_addr_lock_key irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem rtnl_mutex &bridge_netdev_addr_lock_key &obj_hash[i].lock irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem rtnl_mutex &bridge_netdev_addr_lock_key pool_lock#2 irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem rtnl_mutex &bridge_netdev_addr_lock_key krc.lock irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem rtnl_mutex &im->lock &obj_hash[i].lock irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem rtnl_mutex &idev->mc_lock &bridge_netdev_addr_lock_key irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem rtnl_mutex &idev->mc_lock &bridge_netdev_addr_lock_key &obj_hash[i].lock irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem rtnl_mutex &idev->mc_lock &bridge_netdev_addr_lock_key pool_lock#2 irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem rtnl_mutex &idev->mc_lock &bridge_netdev_addr_lock_key krc.lock irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem rtnl_mutex rcu_read_lock &tb->tb6_lock &c->lock irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem rtnl_mutex rcu_read_lock &tb->tb6_lock &obj_hash[i].lock pool_lock irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem rtnl_mutex rcu_read_lock &c->lock irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem rtnl_mutex rcu_read_lock &net->sctp.local_addr_lock &net->sctp.addr_wq_lock &____s->seqcount irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem rtnl_mutex &pmc->lock irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem rtnl_mutex (switchdev_blocking_notif_chain).rwsem irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem rtnl_mutex rcu_read_lock rhashtable_bucket irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem rtnl_mutex &br->multicast_lock irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem rtnl_mutex &br->multicast_lock rcu_read_lock rhashtable_bucket irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem rtnl_mutex &br->multicast_lock rcu_read_lock rcu_read_lock &pool->lock irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem rtnl_mutex &br->multicast_lock rcu_read_lock rcu_read_lock &pool->lock &obj_hash[i].lock irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem rtnl_mutex &br->multicast_lock rcu_read_lock rcu_read_lock &pool->lock pool_lock#2 irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem rtnl_mutex &br->multicast_lock rcu_read_lock rcu_read_lock &pool->lock &p->pi_lock irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem rtnl_mutex &br->multicast_lock rcu_read_lock rcu_read_lock &pool->lock &p->pi_lock &rq->__lock irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem rtnl_mutex &br->multicast_lock rcu_read_lock rcu_read_lock &pool->lock &p->pi_lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem rtnl_mutex &br->multicast_lock rcu_read_lock &pool->lock irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem rtnl_mutex &br->multicast_lock rcu_read_lock &pool->lock &obj_hash[i].lock irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem rtnl_mutex &br->multicast_lock rcu_read_lock &pool->lock pool_lock#2 irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem rtnl_mutex &br->multicast_lock rcu_read_lock &pool->lock &p->pi_lock irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem rtnl_mutex (&mp->timer) irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem rtnl_mutex (work_completion)(&br->mcast_gc_work) irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem rtnl_mutex rcu_state.barrier_mutex irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem rtnl_mutex rcu_state.barrier_mutex rcu_state.barrier_lock irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem rtnl_mutex rcu_state.barrier_mutex rcu_state.barrier_lock &obj_hash[i].lock irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem rtnl_mutex rcu_state.barrier_mutex &x->wait#24 irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem rtnl_mutex rcu_state.barrier_mutex &rq->__lock irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem rtnl_mutex rcu_state.barrier_mutex &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem rtnl_mutex pcpu_lock irq_context: 0 (work_completion)(&data->gc_work) irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem &net->sctp.addr_wq_lock &base->lock irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem &net->sctp.addr_wq_lock &base->lock &obj_hash[i].lock irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem &net->sctp.addr_wq_lock pool_lock#2 irq_context: 0 (wq_completion)events_unbound (work_completion)(&map->work) rcu_state.barrier_mutex rcu_state.barrier_mutex.wait_lock irq_context: 0 (wq_completion)events_unbound (stats_flush_dwork).work rcu_node_0 irq_context: 0 (wq_completion)events_unbound (stats_flush_dwork).work &rcu_state.expedited_wq irq_context: 0 (wq_completion)events_unbound (stats_flush_dwork).work &rcu_state.expedited_wq &p->pi_lock irq_context: 0 (wq_completion)events_unbound (stats_flush_dwork).work &rcu_state.expedited_wq &p->pi_lock &rq->__lock irq_context: 0 (wq_completion)events_unbound (stats_flush_dwork).work &rcu_state.expedited_wq &p->pi_lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)events_unbound (stats_flush_dwork).work &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem nf_hook_mutex fs_reclaim irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem nf_hook_mutex fs_reclaim mmu_notifier_invalidate_range_start irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem nf_hook_mutex pool_lock#2 irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem &ht->mutex quarantine_lock irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem nf_conntrack_mutex irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem nf_conntrack_mutex &nf_conntrack_locks[i] irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem nf_conntrack_mutex &rq->__lock irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem nf_conntrack_mutex &____s->seqcount#7 irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem nf_conntrack_mutex &nf_conntrack_locks[i] &nf_conntrack_locks[i]/1 irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem nf_conntrack_mutex &obj_hash[i].lock irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem nf_conntrack_mutex pool_lock#2 irq_context: 0 sk_lock-AF_INET cpu_hotplug_lock jump_label_mutex text_mutex &rq->__lock irq_context: 0 &net->xfrm.xfrm_cfg_mutex &____s->seqcount#2 irq_context: 0 &net->xfrm.xfrm_cfg_mutex &____s->seqcount irq_context: 0 &net->xfrm.xfrm_cfg_mutex rcu_read_lock &n->list_lock irq_context: 0 &net->xfrm.xfrm_cfg_mutex rcu_read_lock &n->list_lock &c->lock irq_context: 0 &net->xfrm.xfrm_cfg_mutex rcu_read_lock rcu_node_0 irq_context: 0 &net->xfrm.xfrm_cfg_mutex rcu_read_lock &rq->__lock irq_context: 0 &net->xfrm.xfrm_cfg_mutex rcu_read_lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 &net->xfrm.xfrm_cfg_mutex rcu_read_lock nl_table_lock irq_context: 0 &net->xfrm.xfrm_cfg_mutex rcu_read_lock &obj_hash[i].lock irq_context: 0 &net->xfrm.xfrm_cfg_mutex rcu_read_lock nl_table_wait.lock irq_context: 0 cb_lock rtnl_mutex &rdev->wiphy.mtx &sb->s_type->i_mutex_key#3 &xa->xa_lock#4 irq_context: 0 cb_lock rtnl_mutex &rdev->wiphy.mtx lock kernfs_idr_lock &c->lock irq_context: 0 &sb->s_type->i_mutex_key#10 sk_lock-AF_INET6 k-sk_lock-AF_INET6/1 k-slock-AF_INET6 krc.lock &obj_hash[i].lock irq_context: 0 &sb->s_type->i_mutex_key#10 sk_lock-AF_INET6 k-sk_lock-AF_INET6/1 k-slock-AF_INET6 krc.lock &base->lock irq_context: 0 &sb->s_type->i_mutex_key#10 sk_lock-AF_INET6 k-sk_lock-AF_INET6/1 k-slock-AF_INET6 krc.lock &base->lock &obj_hash[i].lock irq_context: 0 cb_lock rtnl_mutex &rdev->wiphy.mtx &idev->mc_lock &obj_hash[i].lock pool_lock irq_context: 0 cb_lock rtnl_mutex &rdev->wiphy.mtx &idev->mc_lock &rq->__lock irq_context: 0 cb_lock rtnl_mutex &rdev->wiphy.mtx &idev->mc_lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 cb_lock rtnl_mutex &rdev->wiphy.mtx rcu_read_lock rcu_node_0 irq_context: 0 &net->xfrm.xfrm_cfg_mutex quarantine_lock irq_context: 0 cb_lock rtnl_mutex &rdev->wiphy.mtx &sb->s_type->i_mutex_key#3 per_cpu_ptr(&cgroup_rstat_cpu_lock, cpu) irq_context: 0 cb_lock rtnl_mutex &rdev->wiphy.mtx &sb->s_type->i_mutex_key#3 rcu_read_lock &rq->__lock irq_context: 0 cb_lock rtnl_mutex &rdev->wiphy.mtx &sb->s_type->i_mutex_key#3 rcu_read_lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 rcu_read_lock rcu_read_lock_bh &base->lock irq_context: 0 rcu_read_lock rcu_read_lock_bh &base->lock &obj_hash[i].lock irq_context: 0 (wq_completion)events (work_completion)(&rdev->destroy_work) rtnl_mutex &pool->lock &p->pi_lock irq_context: 0 (wq_completion)events (work_completion)(&rdev->destroy_work) rtnl_mutex &pool->lock &p->pi_lock &rq->__lock irq_context: 0 (wq_completion)events (work_completion)(&rdev->destroy_work) rtnl_mutex &pool->lock &p->pi_lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)events (work_completion)(&rdev->destroy_work) rtnl_mutex &rdev->wiphy.mtx rcu_state.exp_mutex &rq->__lock &cfs_rq->removed.lock irq_context: 0 cb_lock rtnl_mutex &rdev->wiphy.mtx uevent_sock_mutex &____s->seqcount#2 irq_context: 0 cb_lock rtnl_mutex &rdev->wiphy.mtx uevent_sock_mutex &____s->seqcount irq_context: 0 (wq_completion)events (work_completion)(&rdev->destroy_work) &obj_hash[i].lock pool_lock irq_context: 0 &sb->s_type->i_mutex_key#10 &dir->lock &obj_hash[i].lock irq_context: 0 &sb->s_type->i_mutex_key#10 &dir->lock pool_lock#2 irq_context: 0 rtnl_mutex &xa->xa_lock#3 &____s->seqcount#2 irq_context: 0 rtnl_mutex &xa->xa_lock#3 &____s->seqcount irq_context: 0 &mm->mmap_lock lock#5 rcu_read_lock pool_lock#2 irq_context: 0 &mm->mmap_lock lock#5 &obj_hash[i].lock irq_context: 0 &f->f_pos_lock sb_writers#11 batched_entropy_u8.lock irq_context: 0 &f->f_pos_lock sb_writers#11 kfence_freelist_lock irq_context: 0 &f->f_pos_lock sb_writers#11 &meta->lock irq_context: 0 cb_lock genl_mutex rcu_read_lock rcu_read_lock_bh rcu_read_lock &n->list_lock irq_context: 0 cb_lock genl_mutex rcu_read_lock rcu_read_lock_bh rcu_read_lock &n->list_lock &c->lock irq_context: 0 cb_lock rtnl_mutex &rdev->wiphy.mtx &sb->s_type->i_mutex_key#3 fill_pool_map-wait-type-override &n->list_lock irq_context: 0 cb_lock rtnl_mutex &rdev->wiphy.mtx &sb->s_type->i_mutex_key#3 fill_pool_map-wait-type-override &n->list_lock &c->lock irq_context: 0 sb_writers#4 &sb->s_type->i_mutex_key#8 jbd2_handle &ei->i_data_sem &lg->lg_mutex key#3 irq_context: 0 pernet_ops_rwsem nf_hook_mutex rcu_read_lock rcu_node_0 irq_context: 0 pernet_ops_rwsem nf_hook_mutex rcu_read_lock &rq->__lock irq_context: 0 pernet_ops_rwsem nf_hook_mutex rcu_read_lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 cb_lock rtnl_mutex &rdev->wiphy.mtx &wdev->mtx &local->sta_mtx irq_context: 0 cb_lock rtnl_mutex &rdev->wiphy.mtx &wdev->mtx &ifibss->incomplete_lock irq_context: 0 cb_lock rtnl_mutex &rdev->wiphy.mtx &wdev->mtx &local->mtx irq_context: 0 cb_lock rtnl_mutex &rdev->wiphy.mtx &wdev->mtx &local->mtx &local->chanctx_mtx irq_context: 0 cb_lock rtnl_mutex &rdev->wiphy.mtx &wdev->mtx rcu_state.exp_mutex rcu_node_0 irq_context: 0 cb_lock rtnl_mutex &rdev->wiphy.mtx &wdev->mtx rcu_state.exp_mutex &obj_hash[i].lock irq_context: 0 cb_lock rtnl_mutex &rdev->wiphy.mtx &wdev->mtx rcu_state.exp_mutex rcu_read_lock &pool->lock irq_context: 0 cb_lock rtnl_mutex &rdev->wiphy.mtx &wdev->mtx rcu_state.exp_mutex rcu_read_lock &pool->lock &obj_hash[i].lock irq_context: 0 cb_lock rtnl_mutex &rdev->wiphy.mtx &wdev->mtx rcu_state.exp_mutex rcu_read_lock &pool->lock &p->pi_lock irq_context: 0 cb_lock rtnl_mutex &rdev->wiphy.mtx &wdev->mtx rcu_state.exp_mutex rcu_read_lock &pool->lock &p->pi_lock &rq->__lock irq_context: 0 cb_lock rtnl_mutex &rdev->wiphy.mtx &wdev->mtx rcu_state.exp_mutex rcu_read_lock &pool->lock &p->pi_lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 cb_lock rtnl_mutex &rdev->wiphy.mtx &wdev->mtx rcu_state.exp_mutex &rq->__lock irq_context: 0 cb_lock rtnl_mutex &rdev->wiphy.mtx &wdev->mtx rcu_state.exp_mutex &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 cb_lock rtnl_mutex &rdev->wiphy.mtx &wdev->mtx rcu_state.exp_mutex &rnp->exp_wq[3] irq_context: 0 cb_lock rtnl_mutex &rdev->wiphy.mtx &wdev->mtx &obj_hash[i].lock irq_context: 0 cb_lock rtnl_mutex &rdev->wiphy.mtx &wdev->mtx &list->lock#18 irq_context: 0 cb_lock rtnl_mutex &rdev->wiphy.mtx &wdev->mtx (&ifibss->timer) irq_context: 0 cb_lock rtnl_mutex &rdev->wiphy.mtx &wdev->mtx &base->lock irq_context: 0 cb_lock rtnl_mutex &rdev->wiphy.mtx &wdev->mtx &base->lock &obj_hash[i].lock irq_context: 0 cb_lock rtnl_mutex &rdev->wiphy.mtx &wdev->mtx &local->sta_mtx &local->key_mtx irq_context: 0 cb_lock rtnl_mutex &rdev->wiphy.mtx &wdev->mtx rcu_read_lock &pool->lock/1 irq_context: 0 cb_lock rtnl_mutex &rdev->wiphy.mtx &wdev->mtx rcu_read_lock &pool->lock/1 &obj_hash[i].lock irq_context: 0 cb_lock rtnl_mutex &rdev->wiphy.mtx &wdev->mtx rcu_read_lock &pool->lock/1 &p->pi_lock irq_context: 0 cb_lock rtnl_mutex &rdev->wiphy.mtx &wdev->mtx rcu_read_lock &pool->lock/1 &p->pi_lock &rq->__lock irq_context: 0 cb_lock rtnl_mutex &rdev->wiphy.mtx &wdev->mtx rcu_read_lock &pool->lock/1 &p->pi_lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 cb_lock rtnl_mutex &rdev->wiphy.mtx &wdev->mtx &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 sb_writers#4 &sb->s_type->i_mutex_key#8 jbd2_handle &ei->i_data_sem &(ei->i_block_reservation_lock) key#14 irq_context: 0 cb_lock pgd_lock irq_context: 0 cb_lock key irq_context: 0 cb_lock pcpu_lock irq_context: 0 cb_lock percpu_counters_lock irq_context: 0 cb_lock pcpu_lock stock_lock irq_context: 0 rtnl_mutex team->team_lock_key#2 &bond->mode_lock irq_context: 0 rtnl_mutex team->team_lock_key#2 &dev_addr_list_lock_key irq_context: 0 rtnl_mutex team->team_lock_key#2 &n->list_lock irq_context: 0 rtnl_mutex team->team_lock_key#2 &n->list_lock &c->lock irq_context: 0 (wq_completion)bond8#2 irq_context: 0 (wq_completion)bond8#2 (work_completion)(&(&bond->mii_work)->work) irq_context: 0 (wq_completion)bond8#2 (work_completion)(&(&bond->mii_work)->work) &obj_hash[i].lock irq_context: 0 (wq_completion)bond8#2 (work_completion)(&(&bond->mii_work)->work) &base->lock irq_context: 0 (wq_completion)bond8#2 (work_completion)(&(&bond->mii_work)->work) &base->lock &obj_hash[i].lock irq_context: 0 (wq_completion)bond8#2 (work_completion)(&(&bond->mii_work)->work) &rq->__lock irq_context: 0 (wq_completion)bond8#2 (work_completion)(&(&bond->mii_work)->work) &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)bond8#2 (work_completion)(&(&bond->ad_work)->work) irq_context: 0 (wq_completion)bond8#2 (work_completion)(&(&bond->ad_work)->work) &bond->mode_lock irq_context: 0 (wq_completion)bond8#2 (work_completion)(&(&bond->ad_work)->work) &obj_hash[i].lock irq_context: 0 (wq_completion)bond8#2 (work_completion)(&(&bond->ad_work)->work) &base->lock irq_context: 0 (wq_completion)bond8#2 (work_completion)(&(&bond->ad_work)->work) &base->lock &obj_hash[i].lock irq_context: 0 rtnl_mutex team->team_lock_key#2 rcu_read_lock &bond->stats_lock irq_context: 0 rtnl_mutex team->team_lock_key#2 &dev_addr_list_lock_key pool_lock#2 irq_context: 0 rtnl_mutex team->team_lock_key#2 console_owner_lock irq_context: 0 rtnl_mutex team->team_lock_key#2 console_owner irq_context: 0 rtnl_mutex team->team_lock_key#2 rcu_read_lock rcu_node_0 irq_context: 0 rtnl_mutex team->team_lock_key#2 rcu_read_lock &rq->__lock irq_context: 0 rtnl_mutex team->team_lock_key#2 rcu_read_lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: softirq &(&bond->ad_work)->timer irq_context: softirq &(&bond->ad_work)->timer rcu_read_lock &pool->lock/1 irq_context: softirq &(&bond->ad_work)->timer rcu_read_lock &pool->lock/1 &obj_hash[i].lock irq_context: softirq &(&bond->ad_work)->timer rcu_read_lock &pool->lock/1 &p->pi_lock irq_context: softirq &(&bond->ad_work)->timer rcu_read_lock &pool->lock/1 &p->pi_lock &rq->__lock irq_context: softirq &(&bond->ad_work)->timer rcu_read_lock &pool->lock/1 &p->pi_lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 rtnl_mutex uevent_sock_mutex rcu_read_lock &rcu_state.expedited_wq irq_context: 0 rtnl_mutex uevent_sock_mutex rcu_read_lock &rcu_state.expedited_wq &p->pi_lock irq_context: 0 rtnl_mutex uevent_sock_mutex rcu_read_lock &rcu_state.expedited_wq &p->pi_lock &rq->__lock irq_context: 0 rtnl_mutex uevent_sock_mutex rcu_read_lock &rcu_state.expedited_wq &p->pi_lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 rtnl_mutex &macvlan_netdev_addr_lock_key/1 _xmit_ETHER &c->lock irq_context: 0 rtnl_mutex &macvlan_netdev_addr_lock_key/1 _xmit_ETHER &n->list_lock irq_context: 0 rtnl_mutex &macvlan_netdev_addr_lock_key/1 _xmit_ETHER &n->list_lock &c->lock irq_context: 0 (wq_completion)bond8#2 (work_completion)(&(&slave->notify_work)->work) irq_context: 0 (wq_completion)bond8#2 (work_completion)(&(&slave->notify_work)->work) &obj_hash[i].lock irq_context: 0 (wq_completion)bond8#2 (work_completion)(&(&slave->notify_work)->work) &base->lock irq_context: 0 (wq_completion)bond8#2 (work_completion)(&(&slave->notify_work)->work) &base->lock &obj_hash[i].lock irq_context: 0 rtnl_mutex &dev_addr_list_lock_key/2 irq_context: 0 rtnl_mutex &dev_addr_list_lock_key/2 &macvlan_netdev_addr_lock_key/1 irq_context: 0 rtnl_mutex &dev_addr_list_lock_key/2 &macvlan_netdev_addr_lock_key/1 _xmit_ETHER irq_context: 0 rtnl_mutex &macvlan_netdev_addr_lock_key/1 pool_lock#2 irq_context: 0 rtnl_mutex rcu_read_lock &bond->stats_lock/2 irq_context: 0 (wq_completion)bond8#2 (work_completion)(&(&slave->notify_work)->work) &rq->__lock irq_context: 0 (wq_completion)events (linkwatch_work).work rtnl_mutex team->team_lock_key#2 rcu_read_lock &bond->stats_lock/2 irq_context: 0 (wq_completion)events (linkwatch_work).work rtnl_mutex rcu_read_lock &bond->stats_lock/2 irq_context: 0 (wq_completion)events (linkwatch_work).work rtnl_mutex &dev->tx_global_lock &qdisc_xmit_lock_key irq_context: 0 (wq_completion)events (linkwatch_work).work rtnl_mutex &rnp->exp_wq[1] irq_context: 0 (wq_completion)events (linkwatch_work).work rtnl_mutex &bond->mode_lock irq_context: 0 (wq_completion)bond8#2 (work_completion)(&(&slave->notify_work)->work) rtnl_mutex net_rwsem irq_context: 0 (wq_completion)bond8#2 (work_completion)(&(&slave->notify_work)->work) rtnl_mutex net_rwsem &list->lock#2 irq_context: 0 (wq_completion)bond8#2 (work_completion)(&(&slave->notify_work)->work) rtnl_mutex &tn->lock irq_context: 0 namespace_sem &cfs_rq->removed.lock irq_context: 0 namespace_sem &obj_hash[i].lock irq_context: 0 &journal->j_barrier &journal->j_checkpoint_mutex rcu_read_lock &c->lock irq_context: 0 pernet_ops_rwsem nf_hook_mutex per_cpu_ptr(&cgroup_rstat_cpu_lock, cpu) irq_context: 0 &mm->mmap_lock sb_writers#4 &journal->j_wait_transaction_locked irq_context: 0 &mm->mmap_lock sb_writers#4 &rq->__lock &cfs_rq->removed.lock irq_context: 0 sb_writers#10 &type->i_mutex_dir_key#6/1 cgroup_mutex &root->kernfs_rwsem &rq->__lock irq_context: 0 sb_writers#10 &type->i_mutex_dir_key#6/1 cgroup_mutex &root->kernfs_rwsem &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 &journal->j_barrier &journal->j_checkpoint_mutex &obj_hash[i].lock pool_lock irq_context: 0 mem_id_lock irq_context: softirq rcu_callback mem_id_pool.xa_lock irq_context: 0 sb_writers#4 &type->i_mutex_dir_key#3/1 &sb->s_type->i_mutex_key#8 jbd2_handle &obj_hash[i].lock pool_lock irq_context: 0 sb_writers#10 &type->i_mutex_dir_key#6 &c->lock irq_context: 0 mem_id_lock &rq->__lock irq_context: 0 mem_id_lock fs_reclaim irq_context: 0 mem_id_lock fs_reclaim mmu_notifier_invalidate_range_start irq_context: 0 mem_id_lock &c->lock irq_context: 0 mem_id_lock &____s->seqcount#2 irq_context: 0 mem_id_lock &____s->seqcount irq_context: 0 mem_id_lock pool_lock#2 irq_context: 0 mem_id_lock batched_entropy_u32.lock irq_context: 0 mem_id_lock &obj_hash[i].lock irq_context: 0 mem_id_lock mem_id_pool.xa_lock irq_context: 0 mem_id_lock rcu_read_lock rhashtable_bucket irq_context: 0 rcu_read_lock &r->producer_lock#3 irq_context: 0 &r->consumer_lock#3 irq_context: 0 mem_id_lock &ht->lock irq_context: 0 mem_id_lock rcu_read_lock &ht->lock irq_context: 0 mem_id_lock rcu_read_lock rcu_read_lock rhashtable_bucket irq_context: 0 mem_id_lock rcu_read_lock rcu_read_lock rcu_read_lock &pool->lock irq_context: 0 mem_id_lock rcu_read_lock rcu_read_lock rcu_read_lock &pool->lock &obj_hash[i].lock irq_context: 0 mem_id_lock rcu_read_lock rcu_read_lock rcu_read_lock &pool->lock pool_lock#2 irq_context: 0 mem_id_lock rcu_read_lock rcu_read_lock rcu_read_lock &pool->lock &p->pi_lock irq_context: 0 mem_id_lock rcu_read_lock rcu_read_lock rcu_read_lock &pool->lock &p->pi_lock &rq->__lock irq_context: 0 mem_id_lock rcu_read_lock rcu_read_lock rcu_read_lock &pool->lock &p->pi_lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 mem_id_lock rcu_read_lock rcu_read_lock rcu_node_0 irq_context: 0 mem_id_lock rcu_read_lock rcu_read_lock &rq->__lock irq_context: 0 mem_id_lock rcu_read_lock rcu_read_lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 mem_id_lock rcu_read_lock &obj_hash[i].lock irq_context: 0 mem_id_lock rcu_read_lock pool_lock#2 irq_context: 0 mem_id_lock rcu_read_lock rcu_node_0 irq_context: 0 mem_id_lock rcu_read_lock &rcu_state.expedited_wq irq_context: 0 mem_id_lock rcu_read_lock &rcu_state.expedited_wq &p->pi_lock irq_context: 0 mem_id_lock rcu_read_lock &rcu_state.expedited_wq &p->pi_lock &rq->__lock irq_context: 0 mem_id_lock rcu_read_lock &rcu_state.expedited_wq &p->pi_lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 mem_id_lock rcu_read_lock &rq->__lock irq_context: 0 mem_id_lock rcu_read_lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 sb_writers#10 &type->i_mutex_dir_key#6 &n->list_lock irq_context: 0 sb_writers#10 &type->i_mutex_dir_key#6 &n->list_lock &c->lock irq_context: 0 sb_writers#10 &type->i_mutex_dir_key#6 tomoyo_ss &____s->seqcount#2 irq_context: 0 sb_writers#10 &type->i_mutex_dir_key#6 &root->kernfs_rwsem rcu_read_lock rcu_node_0 irq_context: 0 sb_writers#10 &type->i_mutex_dir_key#6 &root->kernfs_rwsem rcu_read_lock &rq->__lock irq_context: 0 sb_writers#10 &type->i_mutex_dir_key#6 &root->kernfs_rwsem rcu_read_lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 sb_writers#10 &type->i_mutex_dir_key#6 &root->kernfs_rwsem rcu_read_lock &rcu_state.gp_wq irq_context: 0 sb_writers#10 &type->i_mutex_dir_key#6 &root->kernfs_rwsem rcu_read_lock &rcu_state.gp_wq &p->pi_lock irq_context: 0 sb_writers#10 &type->i_mutex_dir_key#6 &root->kernfs_rwsem rcu_read_lock &rcu_state.gp_wq &p->pi_lock &rq->__lock irq_context: 0 sb_writers#10 &type->i_mutex_dir_key#6 &root->kernfs_rwsem rcu_read_lock &rcu_state.gp_wq &p->pi_lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 pernet_ops_rwsem pcpu_alloc_mutex rcu_node_0 irq_context: 0 sb_writers#4 &sb->s_type->i_mutex_key#8 &journal->j_wait_transaction_locked irq_context: 0 sb_writers#4 sb_writers#4 &journal->j_wait_transaction_locked irq_context: 0 (wq_completion)rcu_gp (work_completion)(&rew->rew_work) fill_pool_map-wait-type-override &n->list_lock irq_context: 0 (wq_completion)rcu_gp (work_completion)(&rew->rew_work) fill_pool_map-wait-type-override &n->list_lock &c->lock irq_context: 0 &journal->j_barrier &journal->j_checkpoint_mutex key#27 irq_context: 0 &journal->j_wait_transaction_locked &p->pi_lock &cfs_rq->removed.lock irq_context: 0 sb_writers &sb->s_type->i_mutex_key#4 &rq->__lock &cfs_rq->removed.lock irq_context: 0 sb_writers#4 sb_internal rcu_read_lock &rcu_state.expedited_wq irq_context: 0 pernet_ops_rwsem ovs_mutex nf_ct_proto_mutex defrag6_mutex rcu_read_lock rcu_read_lock rcu_node_0 irq_context: 0 pernet_ops_rwsem ovs_mutex nf_ct_proto_mutex defrag6_mutex rcu_read_lock rcu_read_lock &rq->__lock irq_context: 0 (wq_completion)events fqdir_free_work rcu_node_0 irq_context: 0 (wq_completion)events fqdir_free_work rcu_state.barrier_mutex &rq->__lock &cfs_rq->removed.lock irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem ovs_mutex nf_ct_proto_mutex nf_hook_mutex &rq->__lock irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem ovs_mutex nf_ct_proto_mutex nf_hook_mutex &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)bond8#2 (work_completion)(&(&bond->ad_work)->work) rcu_node_0 irq_context: 0 dup_mmap_sem &mm->mmap_lock &mm->mmap_lock/1 &mm->context.lock rcu_node_0 irq_context: 0 dup_mmap_sem &mm->mmap_lock &mm->mmap_lock/1 &mm->context.lock &rcu_state.expedited_wq irq_context: 0 dup_mmap_sem &mm->mmap_lock &mm->mmap_lock/1 &mm->context.lock &rcu_state.expedited_wq &p->pi_lock irq_context: 0 dup_mmap_sem &mm->mmap_lock &mm->mmap_lock/1 &mm->context.lock &rcu_state.expedited_wq &p->pi_lock &rq->__lock irq_context: 0 (wq_completion)bond8#2 (work_completion)(&(&bond->ad_work)->work) &rcu_state.expedited_wq irq_context: 0 (wq_completion)bond8#2 (work_completion)(&(&bond->ad_work)->work) &rcu_state.expedited_wq &p->pi_lock irq_context: 0 (wq_completion)bond8#2 (work_completion)(&(&bond->ad_work)->work) &rcu_state.expedited_wq &p->pi_lock &rq->__lock irq_context: 0 (wq_completion)bond8#2 (work_completion)(&(&bond->ad_work)->work) &rcu_state.expedited_wq &p->pi_lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)bond8#2 (work_completion)(&(&bond->ad_work)->work) &rq->__lock irq_context: 0 (wq_completion)bond8#2 (work_completion)(&(&bond->ad_work)->work) &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 dup_mmap_sem &mm->mmap_lock &mm->mmap_lock/1 &mm->context.lock &rcu_state.expedited_wq &p->pi_lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 &journal->j_barrier &journal->j_checkpoint_mutex &____s->seqcount#2 irq_context: 0 &journal->j_barrier &journal->j_checkpoint_mutex &____s->seqcount irq_context: 0 &f->f_pos_lock sb_writers irq_context: 0 &f->f_pos_lock sb_writers &sb->s_type->i_mutex_key#4 irq_context: 0 &f->f_pos_lock sb_writers &sb->s_type->i_mutex_key#4 &sem->wait_lock irq_context: 0 &f->f_pos_lock sb_writers &sb->s_type->i_mutex_key#4 &rq->__lock irq_context: 0 &f->f_pos_lock sb_writers &sb->s_type->i_mutex_key#4 &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)phy12 (work_completion)(&(&local->scan_work)->work) &local->mtx batched_entropy_u8.lock irq_context: 0 (wq_completion)phy12 (work_completion)(&(&local->scan_work)->work) &local->mtx kfence_freelist_lock irq_context: 0 sb_writers#4 sb_internal jbd2_handle &ei->i_data_sem &rq->__lock &cfs_rq->removed.lock irq_context: softirq &(&bond->ad_work)->timer rcu_read_lock &pool->lock/1 &p->pi_lock &cfs_rq->removed.lock irq_context: 0 sk_lock-AF_PACKET rcu_state.exp_mutex &cfs_rq->removed.lock irq_context: 0 sk_lock-AF_PACKET rcu_state.exp_mutex pool_lock#2 irq_context: 0 sk_lock-AF_TIPC tk_core.seq.seqcount irq_context: 0 sk_lock-AF_TIPC &list->lock#5 irq_context: 0 sk_lock-AF_TIPC pcpu_lock irq_context: 0 sk_lock-AF_TIPC &sem->wait_lock irq_context: 0 sk_lock-AF_TIPC &p->pi_lock irq_context: 0 sk_lock-AF_TIPC &p->pi_lock &cfs_rq->removed.lock irq_context: 0 sk_lock-AF_TIPC &p->pi_lock &rq->__lock irq_context: 0 sk_lock-AF_TIPC &p->pi_lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 sk_lock-AF_TIPC slock-AF_TIPC rcu_read_lock &ei->socket.wq.wait &p->pi_lock &cfs_rq->removed.lock irq_context: 0 &sb->s_type->i_mutex_key#10 sk_lock-AF_INET6 fill_pool_map-wait-type-override &rq->__lock irq_context: 0 &sb->s_type->i_mutex_key#10 sk_lock-AF_INET6 fill_pool_map-wait-type-override &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 sk_lock-AF_TIPC &mm->mmap_lock rcu_read_lock rcu_read_lock rcu_node_0 irq_context: 0 sk_lock-AF_TIPC &mm->mmap_lock rcu_read_lock rcu_read_lock &rq->__lock irq_context: 0 sk_lock-AF_TIPC &mm->mmap_lock rcu_read_lock rcu_read_lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: softirq (&ifibss->timer) rcu_read_lock &pool->lock &p->pi_lock &rq->__lock &cfs_rq->removed.lock irq_context: 0 sk_lock-AF_TIPC &mm->mmap_lock mmu_notifier_invalidate_range_start &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 sk_lock-AF_TIPC &mm->mmap_lock per_cpu_ptr(&cgroup_rstat_cpu_lock, cpu) irq_context: 0 sk_lock-AF_TIPC &mm->mmap_lock ptlock_ptr(page)#2 lock#4 &lruvec->lru_lock per_cpu_ptr(&cgroup_rstat_cpu_lock, cpu) irq_context: 0 sk_lock-AF_TIPC &mm->mmap_lock rcu_read_lock &rcu_state.expedited_wq irq_context: 0 sk_lock-AF_TIPC &mm->mmap_lock rcu_read_lock &rcu_state.expedited_wq &p->pi_lock irq_context: 0 sk_lock-AF_TIPC &mm->mmap_lock rcu_read_lock &rcu_state.expedited_wq &p->pi_lock &rq->__lock irq_context: 0 sk_lock-AF_TIPC &mm->mmap_lock rcu_read_lock &rcu_state.expedited_wq &p->pi_lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)events_long (work_completion)(&(&br->gc_work)->work) rcu_read_lock &br->hash_lock irq_context: 0 (wq_completion)events_long (work_completion)(&(&br->gc_work)->work) rcu_read_lock &br->hash_lock rcu_read_lock rhashtable_bucket irq_context: 0 (wq_completion)events_long (work_completion)(&(&br->gc_work)->work) rcu_read_lock &br->hash_lock pool_lock#2 irq_context: 0 (wq_completion)events_long (work_completion)(&(&br->gc_work)->work) rcu_read_lock &br->hash_lock nl_table_lock irq_context: 0 (wq_completion)events_long (work_completion)(&(&br->gc_work)->work) rcu_read_lock &br->hash_lock &obj_hash[i].lock irq_context: 0 (wq_completion)events_long (work_completion)(&(&br->gc_work)->work) rcu_read_lock &br->hash_lock nl_table_wait.lock irq_context: 0 (wq_completion)mld (work_completion)(&(&idev->mc_dad_work)->work) &idev->mc_lock rcu_read_lock &rq->__lock irq_context: 0 sk_lock-AF_ALG &mm->mmap_lock rcu_read_lock rcu_read_lock ptlock_ptr(page)#2 irq_context: 0 sk_lock-AF_TIPC &mm->mmap_lock rcu_read_lock per_cpu_ptr(&cgroup_rstat_cpu_lock, cpu) irq_context: 0 sb_writers#4 &type->i_mutex_dir_key#3 jbd2_handle &____s->seqcount irq_context: 0 sb_writers#4 &type->i_mutex_dir_key#3 jbd2_handle rcu_read_lock pool_lock#2 irq_context: 0 sk_lock-AF_TIPC &mm->mmap_lock rcu_read_lock rcu_read_lock &cfs_rq->removed.lock irq_context: 0 sk_lock-AF_TIPC &mm->mmap_lock rcu_read_lock rcu_read_lock &obj_hash[i].lock irq_context: 0 &sb->s_type->i_mutex_key#10 sk_lock-AF_TIPC rcu_read_lock &rq->__lock irq_context: 0 &sb->s_type->i_mutex_key#10 sk_lock-AF_TIPC rcu_read_lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 sk_lock-AF_TIPC &mm->mmap_lock rcu_read_lock &cfs_rq->removed.lock irq_context: 0 sk_lock-AF_TIPC &mm->mmap_lock rcu_read_lock &obj_hash[i].lock irq_context: 0 (wq_completion)events (work_completion)(&(&krcp->krw_arr[i].rcu_work)->work) rcu_callback &rcu_state.expedited_wq irq_context: 0 (wq_completion)events (work_completion)(&(&krcp->krw_arr[i].rcu_work)->work) rcu_callback &rcu_state.expedited_wq &p->pi_lock irq_context: 0 (wq_completion)events (work_completion)(&(&krcp->krw_arr[i].rcu_work)->work) rcu_callback &rcu_state.expedited_wq &p->pi_lock &rq->__lock irq_context: 0 (wq_completion)events (work_completion)(&(&krcp->krw_arr[i].rcu_work)->work) rcu_callback &rcu_state.expedited_wq &p->pi_lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)bat_events (work_completion)(&(&bat_priv->mcast.work)->work) &rq->__lock &cfs_rq->removed.lock irq_context: 0 (wq_completion)wg-crypt-wg0#5 (work_completion)(&({ do { const void *__vpp_verify = (typeof((worker) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __asm__ ("" : "=r"(__ptr) : "0"((typeof(*((worker))) *)((worker)))); (typeof((typeof(*((worker))) *)((worker)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); })->work) rcu_read_lock &pool->lock &p->pi_lock &rq->__lock &cfs_rq->removed.lock irq_context: 0 rtnl_mutex sk_lock-AF_INET cpu_hotplug_lock irq_context: 0 rtnl_mutex sk_lock-AF_INET cpu_hotplug_lock &rq->__lock irq_context: 0 rtnl_mutex sk_lock-AF_INET cpu_hotplug_lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 &sb->s_type->i_mutex_key#10 rtnl_mutex sk_lock-AF_INET irq_context: 0 &sb->s_type->i_mutex_key#10 rtnl_mutex sk_lock-AF_INET slock-AF_INET irq_context: 0 &sb->s_type->i_mutex_key#10 rtnl_mutex slock-AF_INET irq_context: 0 sb_writers#4 &type->i_mutex_dir_key#3/1 jbd2_handle &ei->i_data_sem remove_cache_srcu rcu_read_lock rcu_node_0 irq_context: 0 sb_writers#4 &type->i_mutex_dir_key#3/1 jbd2_handle &ei->i_data_sem remove_cache_srcu rcu_read_lock &rq->__lock irq_context: 0 sb_writers#4 &type->i_mutex_dir_key#3/1 jbd2_handle &ei->i_data_sem remove_cache_srcu rcu_read_lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 sb_writers#4 tomoyo_ss &rq->__lock &cfs_rq->removed.lock irq_context: 0 kn->active#48 rcu_read_lock rcu_node_0 irq_context: 0 kn->active#48 rcu_read_lock &rq->__lock irq_context: 0 kn->active#48 rcu_read_lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 sk_lock-AF_BLUETOOTH-BTPROTO_SCO &obj_hash[i].lock irq_context: 0 sk_lock-AF_BLUETOOTH-BTPROTO_SCO &base->lock irq_context: 0 sk_lock-AF_BLUETOOTH-BTPROTO_SCO &base->lock &obj_hash[i].lock irq_context: 0 sk_lock-AF_BLUETOOTH-BTPROTO_SCO &conn->lock#2 irq_context: 0 sk_lock-AF_BLUETOOTH-BTPROTO_SCO rcu_read_lock &pool->lock/1 irq_context: 0 sk_lock-AF_BLUETOOTH-BTPROTO_SCO rcu_read_lock &pool->lock/1 &obj_hash[i].lock irq_context: 0 sk_lock-AF_BLUETOOTH-BTPROTO_SCO rcu_read_lock &pool->lock/1 pool_lock#2 irq_context: 0 sk_lock-AF_BLUETOOTH-BTPROTO_SCO rcu_read_lock &pool->lock/1 &p->pi_lock irq_context: 0 sk_lock-AF_BLUETOOTH-BTPROTO_SCO rcu_read_lock &pool->lock/1 &p->pi_lock &rq->__lock irq_context: 0 sk_lock-AF_BLUETOOTH-BTPROTO_SCO rcu_read_lock &pool->lock/1 &p->pi_lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 sk_lock-AF_BLUETOOTH-BTPROTO_SCO &rq->__lock irq_context: 0 sk_lock-AF_BLUETOOTH-BTPROTO_SCO &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)hci5 (work_completion)(&hdev->cmd_sync_work) &hdev->req_lock &hdev->lock &c->lock irq_context: 0 (wq_completion)hci5 (work_completion)(&hdev->cmd_sync_work) &hdev->req_lock &hdev->lock hci_cb_list_lock rcu_read_lock &pool->lock/1 &p->pi_lock &cfs_rq->removed.lock irq_context: 0 sk_lock-AF_BLUETOOTH-BTPROTO_SCO rcu_read_lock &ei->socket.wq.wait irq_context: 0 sk_lock-AF_BLUETOOTH-BTPROTO_SCO rcu_read_lock &ei->socket.wq.wait &p->pi_lock irq_context: 0 (wq_completion)hci5 (work_completion)(&hdev->cmd_sync_work) &hdev->req_lock &hdev->lock rcu_state.exp_mutex &rq->__lock &cfs_rq->removed.lock irq_context: 0 sk_lock-AF_BLUETOOTH-BTPROTO_SCO rcu_read_lock &ei->socket.wq.wait &p->pi_lock &rq->__lock irq_context: 0 (wq_completion)hci5 (work_completion)(&hdev->cmd_sync_work) &hdev->req_lock &hdev->lock rcu_state.exp_mutex rcu_read_lock &rq->__lock irq_context: 0 sk_lock-AF_BLUETOOTH-BTPROTO_SCO rcu_read_lock &ei->socket.wq.wait &p->pi_lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)hci5 (work_completion)(&hdev->cmd_sync_work) &hdev->req_lock &hdev->lock rcu_state.exp_mutex rcu_read_lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 sk_lock-AF_BLUETOOTH-BTPROTO_SCO rcu_read_lock rcu_node_0 irq_context: 0 sk_lock-AF_BLUETOOTH-BTPROTO_SCO rcu_read_lock &rcu_state.expedited_wq irq_context: 0 sk_lock-AF_BLUETOOTH-BTPROTO_SCO rcu_read_lock &rcu_state.expedited_wq &p->pi_lock irq_context: 0 sk_lock-AF_BLUETOOTH-BTPROTO_SCO rcu_read_lock &rcu_state.expedited_wq &p->pi_lock &rq->__lock irq_context: 0 sk_lock-AF_BLUETOOTH-BTPROTO_SCO rcu_read_lock &rcu_state.expedited_wq &p->pi_lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)hci5 (work_completion)(&hdev->cmd_sync_work) &hdev->req_lock &hdev->lock hci_cb_list_lock &obj_hash[i].lock pool_lock irq_context: 0 &sbi->s_writepages_rwsem jbd2_handle mmu_notifier_invalidate_range_start &rq->__lock irq_context: 0 &sbi->s_writepages_rwsem jbd2_handle mmu_notifier_invalidate_range_start &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)hci5#2 (work_completion)(&(&conn->disc_work)->work) &hdev->unregister_lock remove_cache_srcu pool_lock#2 irq_context: 0 (wq_completion)hci5#2 (work_completion)(&(&conn->disc_work)->work) &hdev->unregister_lock rcu_node_0 irq_context: 0 (wq_completion)hci5#2 (work_completion)(&(&conn->disc_work)->work) &hdev->unregister_lock &rcu_state.expedited_wq irq_context: 0 (wq_completion)hci5#2 (work_completion)(&(&conn->disc_work)->work) &hdev->unregister_lock &rcu_state.expedited_wq &p->pi_lock irq_context: 0 (wq_completion)hci5#2 (work_completion)(&(&conn->disc_work)->work) &hdev->unregister_lock &rcu_state.expedited_wq &p->pi_lock &rq->__lock irq_context: 0 (wq_completion)hci5#2 (work_completion)(&(&conn->disc_work)->work) &hdev->unregister_lock &rcu_state.expedited_wq &p->pi_lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: softirq (&app->join_timer)#2 rcu_read_lock_bh dev->qdisc_tx_busylock ?: &qdisc_tx_busylock &obj_hash[i].lock irq_context: softirq (&app->join_timer)#2 rcu_read_lock_bh dev->qdisc_tx_busylock ?: &qdisc_tx_busylock pool_lock#2 irq_context: softirq (&app->join_timer) rcu_read_lock_bh dev->qdisc_tx_busylock ?: &qdisc_tx_busylock &obj_hash[i].lock irq_context: softirq (&app->join_timer) rcu_read_lock_bh dev->qdisc_tx_busylock ?: &qdisc_tx_busylock pool_lock#2 irq_context: 0 (wq_completion)ipv6_addrconf (work_completion)(&(&ifa->dad_work)->work) rtnl_mutex rcu_read_lock rcu_read_lock rcu_read_lock_bh rcu_read_lock_bh dev->qdisc_tx_busylock ?: &qdisc_tx_busylock rcu_read_lock &c->lock irq_context: 0 (wq_completion)ipv6_addrconf (work_completion)(&(&ifa->dad_work)->work) rtnl_mutex rcu_read_lock rcu_read_lock rcu_read_lock_bh rcu_read_lock_bh dev->qdisc_tx_busylock ?: &qdisc_tx_busylock rcu_read_lock pool_lock#2 irq_context: 0 (wq_completion)ipv6_addrconf (work_completion)(&(&ifa->dad_work)->work) rtnl_mutex rcu_read_lock rcu_read_lock rcu_read_lock_bh rcu_read_lock_bh dev->qdisc_tx_busylock ?: &qdisc_tx_busylock rcu_read_lock tk_core.seq.seqcount irq_context: 0 (wq_completion)ipv6_addrconf (work_completion)(&(&ifa->dad_work)->work) rtnl_mutex rcu_read_lock rcu_read_lock rcu_read_lock_bh rcu_read_lock_bh dev->qdisc_tx_busylock ?: &qdisc_tx_busylock rcu_read_lock &obj_hash[i].lock irq_context: 0 (wq_completion)ipv6_addrconf (work_completion)(&(&ifa->dad_work)->work) rtnl_mutex rcu_read_lock rcu_read_lock rcu_read_lock_bh rcu_read_lock_bh dev->qdisc_tx_busylock ?: &qdisc_tx_busylock rcu_read_lock &ndev->lock irq_context: 0 (wq_completion)ipv6_addrconf (work_completion)(&(&ifa->dad_work)->work) rtnl_mutex rcu_read_lock rcu_read_lock rcu_read_lock_bh rcu_read_lock_bh dev->qdisc_tx_busylock ?: &qdisc_tx_busylock &obj_hash[i].lock irq_context: 0 (wq_completion)ipv6_addrconf (work_completion)(&(&ifa->dad_work)->work) rtnl_mutex rcu_read_lock rcu_read_lock rcu_read_lock_bh rcu_read_lock_bh dev->qdisc_tx_busylock ?: &qdisc_tx_busylock pool_lock#2 irq_context: 0 (wq_completion)wg-kex-wg0#7 (work_completion)(&peer->transmit_handshake_work) &peer->endpoint_lock rcu_read_lock_bh rcu_read_lock rcu_read_lock rcu_read_lock rcu_read_lock_bh rcu_read_lock &c->lock irq_context: 0 tracepoints_mutex batched_entropy_u8.lock irq_context: 0 tracepoints_mutex kfence_freelist_lock irq_context: 0 event_mutex event_mutex.wait_lock irq_context: 0 event_mutex.wait_lock irq_context: 0 sk_lock-AF_INET &sctp_port_hashtable[i].lock batched_entropy_u8.lock irq_context: 0 sk_lock-AF_INET &sctp_port_hashtable[i].lock kfence_freelist_lock irq_context: 0 &sb->s_type->i_mutex_key#10 &net->sctp.addr_wq_lock slock-AF_INET/1 &sctp_port_hashtable[i].lock &meta->lock irq_context: 0 &sb->s_type->i_mutex_key#10 &net->sctp.addr_wq_lock slock-AF_INET/1 &sctp_port_hashtable[i].lock kfence_freelist_lock irq_context: 0 (wq_completion)rcu_gp (work_completion)(&sdp->work) &meta->lock irq_context: 0 (wq_completion)rcu_gp (work_completion)(&sdp->work) kfence_freelist_lock irq_context: 0 sb_writers#4 &sb->s_type->i_mutex_key#8 mapping.invalidate_lock fill_pool_map-wait-type-override pool_lock#2 irq_context: 0 sb_writers#4 &sb->s_type->i_mutex_key#8 mapping.invalidate_lock fill_pool_map-wait-type-override &c->lock irq_context: 0 sb_writers#4 &sb->s_type->i_mutex_key#8 mapping.invalidate_lock fill_pool_map-wait-type-override pool_lock irq_context: 0 &f->f_pos_lock sb_writers#10 &of->mutex kn->active#48 &rq->__lock irq_context: 0 &f->f_pos_lock sb_writers#10 &of->mutex kn->active#48 &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 kn->active#48 &____s->seqcount#2 irq_context: 0 kn->active#48 &____s->seqcount irq_context: 0 &sb->s_type->i_mutex_key#10 &net->sctp.addr_wq_lock slock-AF_INET/1 quarantine_lock irq_context: 0 &pipe->mutex/1 rtnl_mutex &c->lock irq_context: 0 &pipe->mutex/1 &____s->seqcount#2 irq_context: 0 &pipe->mutex/1 batched_entropy_u8.lock irq_context: 0 &pipe->mutex/1 kfence_freelist_lock irq_context: 0 &pipe->mutex/1 &meta->lock irq_context: 0 cb_lock nlk_cb_mutex-GENERIC remove_cache_srcu irq_context: 0 cb_lock nlk_cb_mutex-GENERIC remove_cache_srcu quarantine_lock irq_context: 0 cb_lock nlk_cb_mutex-GENERIC remove_cache_srcu &c->lock irq_context: 0 cb_lock nlk_cb_mutex-GENERIC remove_cache_srcu &n->list_lock irq_context: 0 cb_lock nlk_cb_mutex-GENERIC remove_cache_srcu rcu_read_lock pool_lock#2 irq_context: 0 cb_lock nlk_cb_mutex-GENERIC remove_cache_srcu &obj_hash[i].lock irq_context: 0 cb_lock nlk_cb_mutex-GENERIC remove_cache_srcu pool_lock#2 irq_context: 0 &sb->s_type->i_mutex_key#8 &sbi->s_writepages_rwsem lock#4 &lruvec->lru_lock per_cpu_ptr(&cgroup_rstat_cpu_lock, cpu) irq_context: 0 pernet_ops_rwsem nf_ct_proto_mutex nf_hook_mutex rcu_read_lock rcu_read_lock rcu_node_0 irq_context: 0 pernet_ops_rwsem nf_ct_proto_mutex nf_hook_mutex rcu_read_lock rcu_read_lock &rq->__lock irq_context: 0 pernet_ops_rwsem nf_ct_proto_mutex nf_hook_mutex rcu_read_lock rcu_read_lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)rcu_gp (work_completion)(&rew->rew_work) pool_lock irq_context: 0 nf_sockopt_mutex &rcu_state.expedited_wq irq_context: 0 pernet_ops_rwsem uevent_sock_mutex quarantine_lock irq_context: softirq (&app->join_timer) rcu_read_lock_bh dev->qdisc_tx_busylock ?: &qdisc_tx_busylock rcu_read_lock pool_lock#2 irq_context: softirq (&app->join_timer) rcu_read_lock_bh dev->qdisc_tx_busylock ?: &qdisc_tx_busylock rcu_read_lock tk_core.seq.seqcount irq_context: softirq (&app->join_timer) rcu_read_lock_bh dev->qdisc_tx_busylock ?: &qdisc_tx_busylock rcu_read_lock &obj_hash[i].lock irq_context: 0 (wq_completion)wg-crypt-wg1 (work_completion)(&peer->transmit_packet_work) &peer->endpoint_lock rcu_read_lock_bh rcu_read_lock rcu_read_lock_bh rcu_read_lock rlock-AF_PACKET irq_context: 0 (wq_completion)wg-crypt-wg0 (work_completion)(&peer->transmit_packet_work) &peer->endpoint_lock rcu_read_lock_bh rcu_read_lock rcu_read_lock_bh rcu_read_lock &n->list_lock irq_context: 0 (wq_completion)wg-crypt-wg1 (work_completion)(&peer->transmit_packet_work) &peer->endpoint_lock rcu_read_lock_bh rcu_read_lock rcu_read_lock_bh rcu_read_lock &n->list_lock irq_context: 0 (wq_completion)wg-crypt-wg0 (work_completion)(&peer->transmit_packet_work) &peer->endpoint_lock rcu_read_lock_bh rcu_read_lock rcu_read_lock_bh rcu_read_lock &n->list_lock &c->lock irq_context: 0 (wq_completion)wg-crypt-wg1 (work_completion)(&peer->transmit_packet_work) &peer->endpoint_lock rcu_read_lock_bh rcu_read_lock rcu_read_lock_bh rcu_read_lock &____s->seqcount#2 irq_context: 0 (wq_completion)wg-crypt-wg1 (work_completion)(&peer->transmit_packet_work) &peer->endpoint_lock rcu_read_lock_bh rcu_read_lock rcu_read_lock_bh rcu_read_lock &____s->seqcount irq_context: 0 (wq_completion)wg-crypt-wg1 (work_completion)(&peer->transmit_packet_work) &peer->endpoint_lock rcu_read_lock_bh rcu_read_lock rcu_read_lock_bh rcu_read_lock pool_lock#2 irq_context: 0 (wq_completion)wg-crypt-wg0 (work_completion)(&peer->transmit_packet_work) &peer->endpoint_lock rcu_read_lock_bh rcu_read_lock rcu_read_lock_bh rcu_read_lock &____s->seqcount#2 irq_context: 0 (wq_completion)wg-crypt-wg0 (work_completion)(&peer->transmit_packet_work) &peer->endpoint_lock rcu_read_lock_bh rcu_read_lock rcu_read_lock_bh rcu_read_lock &____s->seqcount irq_context: 0 kn->active#60 pool_lock#2 irq_context: softirq rcu_read_lock rcu_read_lock rlock-AF_PACKET irq_context: softirq rcu_read_lock rcu_read_lock &br->multicast_lock irq_context: softirq rcu_read_lock rcu_read_lock &br->multicast_lock &obj_hash[i].lock irq_context: softirq rcu_read_lock rcu_read_lock &br->multicast_lock &base->lock irq_context: softirq (&in_dev->mr_ifc_timer) rcu_read_lock rcu_read_lock_bh rcu_read_lock &____s->seqcount#2 irq_context: softirq (&ndev->rs_timer) rcu_read_lock rcu_read_lock rcu_read_lock_bh rcu_read_lock_bh dev->qdisc_tx_busylock ?: &qdisc_tx_busylock rcu_read_lock &ndev->lock irq_context: softirq (&ndev->rs_timer) rcu_read_lock rcu_read_lock rcu_read_lock_bh rcu_read_lock_bh dev->qdisc_tx_busylock ?: &qdisc_tx_busylock &obj_hash[i].lock irq_context: softirq (&ndev->rs_timer) rcu_read_lock rcu_read_lock rcu_read_lock_bh rcu_read_lock_bh dev->qdisc_tx_busylock ?: &qdisc_tx_busylock pool_lock#2 irq_context: 0 (wq_completion)ipv6_addrconf (work_completion)(&(&ifa->dad_work)->work) rtnl_mutex &idev->mc_lock rcu_read_lock rcu_read_lock rcu_read_lock_bh &sch->q.lock hrtimer_bases.lock irq_context: 0 (wq_completion)ipv6_addrconf (work_completion)(&(&ifa->dad_work)->work) rtnl_mutex &idev->mc_lock rcu_read_lock rcu_read_lock rcu_read_lock_bh &sch->q.lock hrtimer_bases.lock &obj_hash[i].lock irq_context: softirq rcu_read_lock rcu_read_lock elock-AF_PACKET irq_context: softirq (&in_dev->mr_ifc_timer) rcu_read_lock rcu_read_lock_bh &sch->q.lock hrtimer_bases.lock irq_context: softirq (&in_dev->mr_ifc_timer) rcu_read_lock rcu_read_lock_bh &sch->q.lock hrtimer_bases.lock &obj_hash[i].lock irq_context: 0 rcu_read_lock_bh &sch->q.lock hrtimer_bases.lock irq_context: 0 rcu_read_lock_bh &sch->q.lock hrtimer_bases.lock &obj_hash[i].lock irq_context: softirq rcu_read_lock rcu_read_lock rcu_read_lock rcu_read_lock rcu_read_lock_bh rcu_read_lock pool_lock#2 irq_context: softirq rcu_read_lock rcu_read_lock rcu_read_lock rcu_read_lock rcu_read_lock_bh rcu_read_lock &obj_hash[i].lock irq_context: softirq (&ndev->rs_timer) rcu_read_lock rcu_read_lock rcu_read_lock_bh rcu_read_lock &meta->lock irq_context: softirq (&ndev->rs_timer) rcu_read_lock rcu_read_lock rcu_read_lock_bh rcu_read_lock kfence_freelist_lock irq_context: softirq rcu_read_lock rcu_read_lock &br->multicast_lock &base->lock &obj_hash[i].lock irq_context: 0 (wq_completion)ipv6_addrconf (work_completion)(&(&ifa->dad_work)->work) rtnl_mutex &idev->mc_lock remove_cache_srcu &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)events (work_completion)(&data->fib_event_work) &data->fib_lock remove_cache_srcu rcu_read_lock rcu_node_0 irq_context: 0 (wq_completion)events (work_completion)(&data->fib_event_work) &data->fib_lock remove_cache_srcu rcu_read_lock &rq->__lock irq_context: 0 (wq_completion)events (work_completion)(&data->fib_event_work) &data->fib_lock remove_cache_srcu rcu_read_lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 rtnl_mutex gdp_mutex &____s->seqcount#2 irq_context: 0 rtnl_mutex gdp_mutex &____s->seqcount irq_context: 0 sb_writers#4 &type->i_mutex_dir_key#3/1 jbd2_handle &ei->i_data_sem &rcu_state.expedited_wq irq_context: 0 sb_writers#4 &type->i_mutex_dir_key#3/1 jbd2_handle &ei->i_data_sem &rcu_state.expedited_wq &p->pi_lock irq_context: 0 sb_writers#4 &type->i_mutex_dir_key#3/1 jbd2_handle &ei->i_data_sem &rcu_state.expedited_wq &p->pi_lock &rq->__lock irq_context: 0 sb_writers#4 &type->i_mutex_dir_key#3/1 jbd2_handle &ei->i_data_sem &rcu_state.expedited_wq &p->pi_lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 &mm->mmap_lock sb_pagefaults mapping.invalidate_lock rcu_read_lock &rcu_state.gp_wq &p->pi_lock &cfs_rq->removed.lock irq_context: 0 sb_writers#4 sb_internal jbd2_handle &ei->i_es_lock key#6 irq_context: 0 (wq_completion)wg-crypt-wg1 (work_completion)(&peer->transmit_packet_work) rcu_read_lock_bh rcu_read_lock &pool->lock/1 &p->pi_lock &rq->__lock irq_context: 0 (wq_completion)wg-crypt-wg1 (work_completion)(&peer->transmit_packet_work) rcu_read_lock_bh rcu_read_lock &pool->lock/1 &p->pi_lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 &mm->mmap_lock sb_pagefaults quarantine_lock irq_context: 0 sk_lock-AF_VSOCK &list->lock#42 irq_context: 0 sk_lock-AF_VSOCK rcu_read_lock &pool->lock irq_context: 0 sk_lock-AF_VSOCK rcu_read_lock &pool->lock &obj_hash[i].lock irq_context: 0 sk_lock-AF_VSOCK rcu_read_lock &pool->lock &p->pi_lock irq_context: 0 sk_lock-AF_VSOCK rcu_read_lock &pool->lock &p->pi_lock &rq->__lock irq_context: 0 sk_lock-AF_VSOCK rcu_read_lock &pool->lock &p->pi_lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)vsock-loopback irq_context: 0 (wq_completion)vsock-loopback (work_completion)(&vsock->pkt_work) irq_context: 0 (wq_completion)vsock-loopback (work_completion)(&vsock->pkt_work) &list->lock#42 irq_context: 0 (wq_completion)vsock-loopback (work_completion)(&vsock->pkt_work) vsock_table_lock irq_context: 0 (wq_completion)vsock-loopback (work_completion)(&vsock->pkt_work) sk_lock-AF_VSOCK irq_context: 0 (wq_completion)vsock-loopback (work_completion)(&vsock->pkt_work) sk_lock-AF_VSOCK slock-AF_VSOCK irq_context: 0 (wq_completion)vsock-loopback (work_completion)(&vsock->pkt_work) sk_lock-AF_VSOCK fs_reclaim irq_context: 0 (wq_completion)vsock-loopback (work_completion)(&vsock->pkt_work) sk_lock-AF_VSOCK fs_reclaim mmu_notifier_invalidate_range_start irq_context: 0 (wq_completion)vsock-loopback (work_completion)(&vsock->pkt_work) sk_lock-AF_VSOCK pool_lock#2 irq_context: 0 (wq_completion)vsock-loopback (work_completion)(&vsock->pkt_work) sk_lock-AF_VSOCK &c->lock irq_context: 0 (wq_completion)vsock-loopback (work_completion)(&vsock->pkt_work) sk_lock-AF_VSOCK &dir->lock irq_context: 0 (wq_completion)vsock-loopback (work_completion)(&vsock->pkt_work) sk_lock-AF_VSOCK &rq->__lock irq_context: 0 (wq_completion)vsock-loopback (work_completion)(&vsock->pkt_work) sk_lock-AF_VSOCK &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)vsock-loopback (work_completion)(&vsock->pkt_work) sk_lock-AF_VSOCK &obj_hash[i].lock irq_context: 0 (wq_completion)vsock-loopback (work_completion)(&vsock->pkt_work) sk_lock-AF_VSOCK sk_lock-AF_VSOCK/1 irq_context: 0 (wq_completion)vsock-loopback (work_completion)(&vsock->pkt_work) sk_lock-AF_VSOCK sk_lock-AF_VSOCK/1 slock-AF_VSOCK irq_context: 0 (wq_completion)vsock-loopback (work_completion)(&vsock->pkt_work) sk_lock-AF_VSOCK sk_lock-AF_VSOCK/1 fs_reclaim irq_context: 0 (wq_completion)vsock-loopback (work_completion)(&vsock->pkt_work) sk_lock-AF_VSOCK sk_lock-AF_VSOCK/1 fs_reclaim mmu_notifier_invalidate_range_start irq_context: 0 (wq_completion)vsock-loopback (work_completion)(&vsock->pkt_work) sk_lock-AF_VSOCK sk_lock-AF_VSOCK/1 pool_lock#2 irq_context: 0 (wq_completion)vsock-loopback (work_completion)(&vsock->pkt_work) sk_lock-AF_VSOCK sk_lock-AF_VSOCK/1 &vvs->tx_lock irq_context: 0 (wq_completion)vsock-loopback (work_completion)(&vsock->pkt_work) sk_lock-AF_VSOCK sk_lock-AF_VSOCK/1 vsock_table_lock irq_context: 0 (wq_completion)vsock-loopback (work_completion)(&vsock->pkt_work) sk_lock-AF_VSOCK sk_lock-AF_VSOCK/1 &vvs->rx_lock irq_context: 0 (wq_completion)vsock-loopback (work_completion)(&vsock->pkt_work) sk_lock-AF_VSOCK sk_lock-AF_VSOCK/1 &list->lock#42 irq_context: 0 (wq_completion)vsock-loopback (work_completion)(&vsock->pkt_work) sk_lock-AF_VSOCK sk_lock-AF_VSOCK/1 rcu_read_lock &pool->lock irq_context: 0 (wq_completion)vsock-loopback (work_completion)(&vsock->pkt_work) sk_lock-AF_VSOCK sk_lock-AF_VSOCK/1 rcu_read_lock &pool->lock &obj_hash[i].lock irq_context: 0 (wq_completion)vsock-loopback (work_completion)(&vsock->pkt_work) slock-AF_VSOCK irq_context: 0 (wq_completion)vsock-loopback (work_completion)(&vsock->pkt_work) sk_lock-AF_VSOCK slock-AF_VSOCK &sk->sk_lock.wq irq_context: 0 (wq_completion)vsock-loopback (work_completion)(&vsock->pkt_work) sk_lock-AF_VSOCK &pool->lock irq_context: 0 sk_lock-AF_VSOCK &ei->socket.wq.wait irq_context: 0 slock-AF_VSOCK &sk->sk_lock.wq irq_context: 0 slock-AF_VSOCK &sk->sk_lock.wq &p->pi_lock irq_context: 0 slock-AF_VSOCK &sk->sk_lock.wq &p->pi_lock &rq->__lock irq_context: 0 slock-AF_VSOCK &sk->sk_lock.wq &p->pi_lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)vsock-loopback (work_completion)(&vsock->pkt_work) sk_lock-AF_VSOCK &vvs->tx_lock irq_context: 0 (wq_completion)vsock-loopback (work_completion)(&vsock->pkt_work) sk_lock-AF_VSOCK rcu_read_lock &ei->socket.wq.wait irq_context: 0 (wq_completion)vsock-loopback (work_completion)(&vsock->pkt_work) sk_lock-AF_VSOCK rcu_read_lock &ei->socket.wq.wait &p->pi_lock irq_context: 0 (wq_completion)vsock-loopback (work_completion)(&vsock->pkt_work) sk_lock-AF_VSOCK rcu_read_lock &ei->socket.wq.wait &p->pi_lock &rq->__lock irq_context: 0 (wq_completion)vsock-loopback (work_completion)(&vsock->pkt_work) sk_lock-AF_VSOCK rcu_read_lock &ei->socket.wq.wait &p->pi_lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)vsock-loopback (work_completion)(&vsock->pkt_work) sk_lock-AF_VSOCK vsock_table_lock irq_context: 0 sk_lock-AF_VSOCK &vvs->tx_lock irq_context: 0 sk_lock-AF_VSOCK &zone->lock irq_context: 0 sk_lock-AF_VSOCK &mm->mmap_lock ptlock_ptr(page)#2 irq_context: 0 sk_lock-AF_VSOCK &zone->lock &____s->seqcount irq_context: 0 &sb->s_type->i_mutex_key#10 sk_lock-AF_VSOCK sk_lock-AF_VSOCK/1 irq_context: 0 &sb->s_type->i_mutex_key#10 sk_lock-AF_VSOCK sk_lock-AF_VSOCK/1 slock-AF_VSOCK irq_context: 0 &sb->s_type->i_mutex_key#10 sk_lock-AF_VSOCK sk_lock-AF_VSOCK/1 fs_reclaim irq_context: 0 &sb->s_type->i_mutex_key#10 sk_lock-AF_VSOCK sk_lock-AF_VSOCK/1 fs_reclaim mmu_notifier_invalidate_range_start irq_context: 0 &sb->s_type->i_mutex_key#10 sk_lock-AF_VSOCK sk_lock-AF_VSOCK/1 pool_lock#2 irq_context: 0 &sb->s_type->i_mutex_key#10 sk_lock-AF_VSOCK sk_lock-AF_VSOCK/1 &c->lock irq_context: 0 &sb->s_type->i_mutex_key#10 sk_lock-AF_VSOCK sk_lock-AF_VSOCK/1 &vvs->rx_lock irq_context: 0 &sb->s_type->i_mutex_key#10 sk_lock-AF_VSOCK sk_lock-AF_VSOCK/1 &list->lock#42 irq_context: 0 &sb->s_type->i_mutex_key#10 sk_lock-AF_VSOCK sk_lock-AF_VSOCK/1 &obj_hash[i].lock irq_context: 0 &sb->s_type->i_mutex_key#10 sk_lock-AF_VSOCK sk_lock-AF_VSOCK/1 &base->lock irq_context: 0 &sb->s_type->i_mutex_key#10 sk_lock-AF_VSOCK sk_lock-AF_VSOCK/1 &base->lock &obj_hash[i].lock irq_context: 0 &sb->s_type->i_mutex_key#10 sk_lock-AF_VSOCK sk_lock-AF_VSOCK/1 clock-AF_VSOCK irq_context: 0 &sb->s_type->i_mutex_key#10 sk_lock-AF_VSOCK sk_lock-AF_VSOCK/1 rlock-AF_VSOCK irq_context: 0 &sb->s_type->i_mutex_key#10 sk_lock-AF_VSOCK fs_reclaim irq_context: 0 &sb->s_type->i_mutex_key#10 sk_lock-AF_VSOCK fs_reclaim mmu_notifier_invalidate_range_start irq_context: 0 &sb->s_type->i_mutex_key#10 sk_lock-AF_VSOCK pool_lock#2 irq_context: 0 &sb->s_type->i_mutex_key#10 sk_lock-AF_VSOCK &vvs->rx_lock irq_context: 0 &sb->s_type->i_mutex_key#10 sk_lock-AF_VSOCK &list->lock#42 irq_context: 0 &sb->s_type->i_mutex_key#10 sk_lock-AF_VSOCK &obj_hash[i].lock irq_context: 0 &sb->s_type->i_mutex_key#10 sk_lock-AF_VSOCK &base->lock irq_context: 0 &sb->s_type->i_mutex_key#10 sk_lock-AF_VSOCK &base->lock &obj_hash[i].lock irq_context: 0 vlan_ioctl_mutex rtnl_mutex &idev->mc_lock &vlan_netdev_addr_lock_key/1 &obj_hash[i].lock pool_lock irq_context: 0 (wq_completion)vsock-loopback (work_completion)(&vsock->pkt_work) sk_lock-AF_VSOCK &vvs->rx_lock irq_context: 0 (wq_completion)vsock-loopback (work_completion)(&vsock->pkt_work) sk_lock-AF_VSOCK &list->lock#42 irq_context: 0 (wq_completion)vsock-loopback (work_completion)(&vsock->pkt_work) sk_lock-AF_VSOCK rcu_read_lock &pool->lock irq_context: 0 (wq_completion)vsock-loopback (work_completion)(&vsock->pkt_work) sk_lock-AF_VSOCK rcu_read_lock &pool->lock &obj_hash[i].lock irq_context: 0 (wq_completion)vsock-loopback (work_completion)(&vsock->pkt_work) sk_lock-AF_VSOCK &base->lock irq_context: 0 (wq_completion)vsock-loopback (work_completion)(&vsock->pkt_work) sk_lock-AF_VSOCK &base->lock &obj_hash[i].lock irq_context: 0 (wq_completion)vsock-loopback (work_completion)(&vsock->pkt_work) &obj_hash[i].lock irq_context: 0 (wq_completion)vsock-loopback (work_completion)(&vsock->pkt_work) pool_lock#2 irq_context: 0 (wq_completion)vsock-loopback (work_completion)(&vsock->pkt_work) &dir->lock irq_context: 0 (wq_completion)vsock-loopback (work_completion)(&vsock->pkt_work) stock_lock irq_context: 0 (wq_completion)vsock-loopback (work_completion)(&vsock->pkt_work) sk_lock-AF_VSOCK rcu_read_lock pool_lock#2 irq_context: 0 (wq_completion)vsock-loopback (work_completion)(&vsock->pkt_work) sk_lock-AF_VSOCK &zone->lock irq_context: 0 (wq_completion)vsock-loopback (work_completion)(&vsock->pkt_work) sk_lock-AF_VSOCK &zone->lock &____s->seqcount irq_context: 0 vlan_ioctl_mutex rtnl_mutex &rnp->exp_lock irq_context: 0 vlan_ioctl_mutex rtnl_mutex rcu_state.exp_mutex irq_context: 0 vlan_ioctl_mutex rtnl_mutex rcu_state.exp_mutex rcu_state.exp_mutex.wait_lock irq_context: 0 map_idr_lock &obj_hash[i].lock pool_lock irq_context: 0 (wq_completion)vsock-loopback (work_completion)(&vsock->pkt_work) sk_lock-AF_VSOCK sk_lock-AF_VSOCK/1 &c->lock irq_context: 0 &sb->s_type->i_mutex_key#10 sk_lock-AF_VSOCK sk_lock-AF_VSOCK/1 rcu_read_lock &pool->lock irq_context: 0 &sb->s_type->i_mutex_key#10 sk_lock-AF_VSOCK sk_lock-AF_VSOCK/1 rcu_read_lock &pool->lock &obj_hash[i].lock irq_context: 0 &sb->s_type->i_mutex_key#10 sk_lock-AF_VSOCK sk_lock-AF_VSOCK/1 rcu_read_lock &pool->lock &p->pi_lock irq_context: 0 &sb->s_type->i_mutex_key#10 sk_lock-AF_VSOCK sk_lock-AF_VSOCK/1 rcu_read_lock &pool->lock &p->pi_lock &rq->__lock irq_context: 0 &sb->s_type->i_mutex_key#10 sk_lock-AF_VSOCK sk_lock-AF_VSOCK/1 rcu_read_lock &pool->lock &p->pi_lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 sb_writers#4 &type->i_mutex_dir_key#3/1 &type->i_mutex_dir_key#3 rcu_read_lock &rq->__lock &cfs_rq->removed.lock irq_context: 0 sb_writers#3 &sb->s_type->i_mutex_key#9 mmu_notifier_invalidate_range_start &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 sb_writers#4 &type->i_mutex_dir_key#3 jbd2_handle rcu_read_lock &rcu_state.gp_wq irq_context: 0 sb_writers#4 &type->i_mutex_dir_key#3 jbd2_handle rcu_read_lock &rcu_state.gp_wq &p->pi_lock irq_context: 0 sb_writers#4 &type->i_mutex_dir_key#3 jbd2_handle rcu_read_lock &rcu_state.gp_wq &p->pi_lock &rq->__lock irq_context: 0 sb_writers#4 &type->i_mutex_dir_key#3 jbd2_handle rcu_read_lock &rcu_state.gp_wq &p->pi_lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: softirq (&in_dev->mr_ifc_timer) rcu_read_lock rcu_read_lock_bh rcu_read_lock rcu_read_lock &table->lock#3 irq_context: softirq (&in_dev->mr_ifc_timer) rcu_read_lock rcu_read_lock_bh rcu_read_lock rcu_read_lock_bh rcu_read_lock pool_lock#2 irq_context: 0 (wq_completion)ipv6_addrconf (work_completion)(&(&ifa->dad_work)->work) rtnl_mutex rcu_read_lock rcu_read_lock rcu_read_lock_bh rcu_read_lock rcu_read_lock &table->lock#3 irq_context: 0 sk_lock-AF_INET6 rcu_read_lock rcu_read_lock rcu_read_lock &ndev->lock irq_context: 0 sk_lock-AF_INET6 rcu_read_lock rcu_read_lock rcu_read_lock rcu_read_lock &____s->seqcount#10 irq_context: 0 sk_lock-AF_INET6 rcu_read_lock rcu_read_lock rcu_read_lock rcu_read_lock rcu_read_lock_bh rcu_read_lock tk_core.seq.seqcount irq_context: 0 sk_lock-AF_INET6 rcu_read_lock rcu_read_lock rcu_read_lock rcu_read_lock rcu_read_lock_bh rcu_read_lock &list->lock#5 irq_context: 0 &type->i_mutex_dir_key#3 stock_lock rcu_read_lock per_cpu_ptr(&cgroup_rstat_cpu_lock, cpu) irq_context: 0 sk_lock-AF_INET6 rcu_read_lock rcu_read_lock rcu_read_lock_bh _xmit_TUNNEL6#2 irq_context: 0 sk_lock-AF_INET6 rcu_read_lock rcu_read_lock rcu_read_lock_bh _xmit_TUNNEL6#2 rcu_read_lock &ndev->lock irq_context: 0 sk_lock-AF_INET6 rcu_read_lock rcu_read_lock rcu_read_lock_bh _xmit_TUNNEL6#2 icmp_global.lock irq_context: 0 sk_lock-AF_INET6 rcu_read_lock rcu_read_lock rcu_read_lock_bh _xmit_TUNNEL6#2 icmp_global.lock batched_entropy_u8.lock irq_context: 0 sk_lock-AF_INET6 rcu_read_lock rcu_read_lock rcu_read_lock_bh _xmit_TUNNEL6#2 k-slock-AF_INET6 rcu_read_lock rcu_read_lock rcu_read_lock stock_lock irq_context: 0 sk_lock-AF_INET6 rcu_read_lock rcu_read_lock rcu_read_lock_bh _xmit_TUNNEL6#2 k-slock-AF_INET6 rcu_read_lock rcu_read_lock rcu_read_lock pool_lock#2 irq_context: 0 sk_lock-AF_INET6 rcu_read_lock rcu_read_lock rcu_read_lock_bh _xmit_TUNNEL6#2 k-slock-AF_INET6 rcu_read_lock rcu_read_lock rcu_read_lock &dir->lock#2 irq_context: 0 sk_lock-AF_INET6 rcu_read_lock rcu_read_lock rcu_read_lock_bh _xmit_TUNNEL6#2 k-slock-AF_INET6 rcu_read_lock &c->lock irq_context: 0 sk_lock-AF_INET6 rcu_read_lock rcu_read_lock rcu_read_lock_bh _xmit_TUNNEL6#2 k-slock-AF_INET6 rcu_read_lock pool_lock#2 irq_context: 0 sk_lock-AF_INET6 rcu_read_lock rcu_read_lock rcu_read_lock_bh _xmit_TUNNEL6#2 k-slock-AF_INET6 rcu_read_lock rcu_read_lock pool_lock#2 irq_context: 0 sk_lock-AF_INET6 rcu_read_lock rcu_read_lock rcu_read_lock_bh _xmit_TUNNEL6#2 k-slock-AF_INET6 rcu_read_lock rcu_read_lock &obj_hash[i].lock irq_context: 0 sk_lock-AF_INET6 rcu_read_lock rcu_read_lock rcu_read_lock_bh _xmit_TUNNEL6#2 k-slock-AF_INET6 rcu_read_lock rcu_read_lock &dir->lock#2 irq_context: 0 sk_lock-AF_INET6 rcu_read_lock rcu_read_lock rcu_read_lock_bh _xmit_TUNNEL6#2 k-slock-AF_INET6 rcu_read_lock rcu_read_lock &tbl->lock irq_context: 0 sk_lock-AF_INET6 rcu_read_lock rcu_read_lock rcu_read_lock_bh _xmit_TUNNEL6#2 k-slock-AF_INET6 rcu_read_lock rcu_read_lock &n->lock irq_context: 0 sk_lock-AF_INET6 rcu_read_lock rcu_read_lock rcu_read_lock_bh _xmit_TUNNEL6#2 k-slock-AF_INET6 rcu_read_lock rcu_read_lock &____s->seqcount#9 irq_context: 0 sk_lock-AF_INET6 rcu_read_lock rcu_read_lock rcu_read_lock_bh _xmit_TUNNEL6#2 k-slock-AF_INET6 rcu_read_lock rcu_read_lock rcu_read_lock_bh tk_core.seq.seqcount irq_context: 0 sk_lock-AF_INET6 rcu_read_lock rcu_read_lock rcu_read_lock_bh _xmit_TUNNEL6#2 k-slock-AF_INET6 rcu_read_lock rcu_read_lock rcu_read_lock_bh &list->lock#5 irq_context: 0 sk_lock-AF_INET6 rcu_read_lock rcu_read_lock rcu_read_lock_bh _xmit_TUNNEL6#2 &obj_hash[i].lock irq_context: 0 sk_lock-AF_INET6 rcu_read_lock rcu_read_lock rcu_read_lock_bh _xmit_TUNNEL6#2 pool_lock#2 irq_context: 0 sk_lock-AF_INET6 rcu_read_lock rcu_read_lock rcu_read_lock_bh _xmit_TUNNEL6#2 k-slock-AF_INET6 rcu_read_lock rcu_read_lock &____s->seqcount#10 irq_context: 0 sk_lock-AF_INET6 &sctp_ep_hashtable[i].lock irq_context: 0 rtnl_mutex cpu_hotplug_lock wq_pool_mutex rcu_read_lock rcu_node_0 irq_context: 0 rtnl_mutex cpu_hotplug_lock wq_pool_mutex rcu_read_lock &rq->__lock irq_context: 0 rtnl_mutex cpu_hotplug_lock wq_pool_mutex rcu_read_lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 &mm->mmap_lock sb_pagefaults jbd2_handle rcu_node_0 irq_context: 0 (wq_completion)wg-crypt-wg0#2 (work_completion)(&peer->transmit_packet_work) &peer->endpoint_lock rcu_read_lock_bh rcu_read_lock rcu_read_lock_bh rcu_read_lock &n->list_lock irq_context: 0 (wq_completion)wg-crypt-wg0#2 (work_completion)(&peer->transmit_packet_work) &peer->endpoint_lock rcu_read_lock_bh rcu_read_lock rcu_read_lock_bh rcu_read_lock &n->list_lock &c->lock irq_context: 0 (wq_completion)events_unbound (work_completion)(&map->work) &p->pi_lock &cfs_rq->removed.lock irq_context: softirq (&in_dev->mr_ifc_timer) rcu_read_lock rcu_read_lock_bh rcu_read_lock rcu_read_lock_bh rcu_read_lock &obj_hash[i].lock irq_context: 0 (wq_completion)ipv6_addrconf (work_completion)(&(&ifa->dad_work)->work) rtnl_mutex &idev->mc_lock rcu_read_lock rcu_read_lock rcu_read_lock_bh rcu_read_lock rcu_read_lock &table->lock#3 irq_context: 0 sk_lock-AF_INET6 rcu_read_lock rcu_read_lock rcu_read_lock_bh _xmit_TUNNEL6#2 k-slock-AF_INET6 rcu_read_lock rcu_read_lock rcu_read_lock &c->lock irq_context: 0 sk_lock-AF_INET6 rcu_read_lock rcu_read_lock rcu_read_lock_bh _xmit_TUNNEL6#2 k-slock-AF_INET6 rcu_read_lock &n->list_lock irq_context: 0 sk_lock-AF_INET6 rcu_read_lock rcu_read_lock rcu_read_lock_bh _xmit_TUNNEL6#2 k-slock-AF_INET6 rcu_read_lock &n->list_lock &c->lock irq_context: softirq (&n->timer) rcu_read_lock rcu_read_lock rcu_read_lock_bh rcu_read_lock &n->list_lock irq_context: softirq (&n->timer) rcu_read_lock rcu_read_lock rcu_read_lock_bh rcu_read_lock &n->list_lock &c->lock irq_context: 0 (wq_completion)mld (work_completion)(&(&idev->mc_dad_work)->work) &idev->mc_lock rcu_read_lock rcu_read_lock rcu_read_lock_bh rcu_read_lock rcu_read_lock &table->lock#3 irq_context: 0 (wq_completion)mld (work_completion)(&(&idev->mc_dad_work)->work) &idev->mc_lock rcu_read_lock rcu_read_lock rcu_read_lock_bh rcu_read_lock rcu_read_lock_bh rcu_read_lock pool_lock#2 irq_context: 0 (wq_completion)mld (work_completion)(&(&idev->mc_dad_work)->work) &idev->mc_lock rcu_read_lock rcu_read_lock rcu_read_lock_bh rcu_read_lock rcu_read_lock_bh rcu_read_lock &obj_hash[i].lock irq_context: 0 sk_lock-AF_INET6 rcu_read_lock rcu_read_lock rcu_read_lock_bh _xmit_TUNNEL6#2 rcu_read_lock &c->lock irq_context: 0 rtnl_mutex pcpu_alloc_mutex &rcu_state.expedited_wq irq_context: 0 rtnl_mutex pcpu_alloc_mutex &rcu_state.expedited_wq &p->pi_lock irq_context: 0 rtnl_mutex pcpu_alloc_mutex &rcu_state.expedited_wq &p->pi_lock &rq->__lock irq_context: 0 rtnl_mutex pcpu_alloc_mutex &rcu_state.expedited_wq &p->pi_lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 sk_lock-AF_INET6 rcu_read_lock rcu_read_lock rcu_read_lock_bh _xmit_TUNNEL6#2 rcu_read_lock pool_lock#2 irq_context: 0 sk_lock-AF_INET6 rcu_read_lock rcu_read_lock rcu_read_lock_bh _xmit_TUNNEL6#2 rcu_read_lock tk_core.seq.seqcount irq_context: 0 sk_lock-AF_INET6 rcu_read_lock rcu_read_lock rcu_read_lock_bh _xmit_TUNNEL6#2 rcu_read_lock &obj_hash[i].lock irq_context: 0 sk_lock-AF_INET6 rcu_read_lock rcu_read_lock rcu_read_lock_bh _xmit_TUNNEL6#2 k-slock-AF_INET6 rcu_read_lock rcu_read_lock rcu_read_lock_bh rcu_read_lock pool_lock#2 irq_context: 0 sk_lock-AF_INET6 rcu_read_lock rcu_read_lock rcu_read_lock_bh _xmit_TUNNEL6#2 k-slock-AF_INET6 rcu_read_lock rcu_read_lock rcu_read_lock_bh rcu_read_lock tk_core.seq.seqcount irq_context: 0 sk_lock-AF_INET6 rcu_read_lock rcu_read_lock rcu_read_lock_bh _xmit_TUNNEL6#2 k-slock-AF_INET6 rcu_read_lock rcu_read_lock rcu_read_lock_bh rcu_read_lock &obj_hash[i].lock irq_context: 0 sk_lock-AF_INET6 rcu_read_lock rcu_read_lock rcu_read_lock_bh _xmit_TUNNEL6#2 k-slock-AF_INET6 rcu_read_lock rcu_read_lock rcu_read_lock_bh rcu_read_lock &c->lock irq_context: 0 sk_lock-AF_INET6 rcu_read_lock rcu_read_lock rcu_read_lock_bh _xmit_TUNNEL6#2 k-slock-AF_INET6 rcu_read_lock rcu_read_lock rcu_read_lock_bh rcu_read_lock &n->list_lock irq_context: 0 sk_lock-AF_INET6 rcu_read_lock rcu_read_lock rcu_read_lock_bh _xmit_TUNNEL6#2 k-slock-AF_INET6 rcu_read_lock rcu_read_lock rcu_read_lock_bh rcu_read_lock &n->list_lock &c->lock irq_context: 0 sk_lock-AF_INET6 crypto_alg_sem &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 rtnl_mutex pool_lock irq_context: 0 rtnl_mutex &root->kernfs_rwsem &p->pi_lock irq_context: 0 sk_lock-AF_INET rcu_read_lock rcu_read_lock rcu_read_lock_bh dev->qdisc_tx_busylock ?: &qdisc_tx_busylock _xmit_ETHER#2 rcu_read_lock &base->lock irq_context: 0 sk_lock-AF_INET rcu_read_lock rcu_read_lock rcu_read_lock_bh dev->qdisc_tx_busylock ?: &qdisc_tx_busylock _xmit_ETHER#2 rcu_read_lock &base->lock &obj_hash[i].lock irq_context: 0 (wq_completion)wg-crypt-wg1#2 (work_completion)(&({ do { const void *__vpp_verify = (typeof((worker) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __asm__ ("" : "=r"(__ptr) : "0"((typeof(*((worker))) *)((worker)))); (typeof((typeof(*((worker))) *)((worker)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); })->work) rcu_node_0 irq_context: 0 sk_lock-AF_INET6 clock-AF_INET6 &n->list_lock irq_context: 0 sk_lock-AF_INET6 clock-AF_INET6 &n->list_lock &c->lock irq_context: 0 &sb->s_type->i_mutex_key#10 krc.lock &obj_hash[i].lock irq_context: 0 &sb->s_type->i_mutex_key#10 krc.lock &base->lock irq_context: 0 &sb->s_type->i_mutex_key#10 krc.lock &base->lock &obj_hash[i].lock irq_context: 0 (wq_completion)phy12 (work_completion)(&(&local->scan_work)->work) &local->mtx rcu_state.exp_mutex &rnp->exp_wq[2] irq_context: 0 &sb->s_type->i_mutex_key#10 sk_lock-AF_INET6/1 &rcu_state.expedited_wq &p->pi_lock &rq->__lock irq_context: 0 &sb->s_type->i_mutex_key#10 sk_lock-AF_INET6/1 &rcu_state.expedited_wq &p->pi_lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 kn->active#66 fs_reclaim irq_context: 0 kn->active#66 fs_reclaim mmu_notifier_invalidate_range_start irq_context: 0 kn->active#66 stock_lock irq_context: 0 kn->active#66 &kernfs_locks->open_file_mutex[count] irq_context: 0 kn->active#66 &kernfs_locks->open_file_mutex[count] &rq->__lock irq_context: 0 kn->active#66 &kernfs_locks->open_file_mutex[count] fs_reclaim irq_context: 0 kn->active#66 &kernfs_locks->open_file_mutex[count] fs_reclaim mmu_notifier_invalidate_range_start irq_context: 0 kn->active#66 &kernfs_locks->open_file_mutex[count] &c->lock irq_context: 0 rtnl_mutex sk_lock-AF_INET6 rcu_read_lock &c->lock irq_context: 0 rtnl_mutex sk_lock-AF_INET6 &idev->mc_lock &c->lock irq_context: 0 rtnl_mutex sk_lock-AF_INET6 &idev->mc_lock _xmit_ETHER &c->lock irq_context: 0 rtnl_mutex ifalias_mutex irq_context: 0 rtnl_mutex &sch->q.lock pool_lock#2 irq_context: 0 (wq_completion)bat_events (work_completion)(&(&bat_priv->mcast.work)->work) &bat_priv->mcast.mla_lock &bat_priv->tt.changes_list_lock &obj_hash[i].lock irq_context: 0 (wq_completion)bat_events (work_completion)(&(&bat_priv->mcast.work)->work) &bat_priv->mcast.mla_lock &bat_priv->tt.changes_list_lock pool_lock#2 irq_context: 0 (wq_completion)bond6 (work_completion)(&(&bond->alb_work)->work) &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 sb_writers#4 &sb->s_type->i_mutex_key#8 &sbi->s_writepages_rwsem &rnp->exp_wq[1] irq_context: 0 &f->f_pos_lock sb_writers#10 remove_cache_srcu rcu_read_lock rcu_node_0 irq_context: 0 &f->f_pos_lock sb_writers#10 remove_cache_srcu rcu_read_lock &rq->__lock irq_context: 0 &f->f_pos_lock sb_writers#10 remove_cache_srcu rcu_read_lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 sk_lock-AF_XDP slock-AF_XDP &sk->sk_lock.wq irq_context: 0 kn->active#48 &kernfs_locks->open_file_mutex[count] &____s->seqcount#2 irq_context: 0 kn->active#48 &kernfs_locks->open_file_mutex[count] &____s->seqcount irq_context: 0 kn->active#66 &c->lock irq_context: 0 kn->active#66 &n->list_lock irq_context: 0 kn->active#66 &n->list_lock &c->lock irq_context: 0 kn->active#66 &rq->__lock irq_context: 0 kn->active#66 &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 sk_lock-AF_XDP &rq->__lock irq_context: 0 kn->active#66 &____s->seqcount#2 irq_context: 0 kn->active#66 &____s->seqcount irq_context: 0 slock-AF_XDP &sk->sk_lock.wq irq_context: 0 sk_lock-AF_XDP &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 slock-AF_XDP &sk->sk_lock.wq &p->pi_lock irq_context: 0 slock-AF_XDP &sk->sk_lock.wq &p->pi_lock &rq->__lock irq_context: 0 slock-AF_XDP &sk->sk_lock.wq &p->pi_lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 rtnl_mutex &sch->q.lock batched_entropy_u64.lock crngs.lock irq_context: 0 rtnl_mutex &bridge_netdev_addr_lock_key &____s->seqcount#2 irq_context: 0 rtnl_mutex &bridge_netdev_addr_lock_key &____s->seqcount irq_context: 0 pernet_ops_rwsem &ht->mutex rcu_node_0 irq_context: 0 pernet_ops_rwsem &ht->mutex &rcu_state.expedited_wq irq_context: 0 pernet_ops_rwsem &ht->mutex &rcu_state.expedited_wq &p->pi_lock irq_context: 0 pernet_ops_rwsem &ht->mutex &rcu_state.expedited_wq &p->pi_lock &rq->__lock irq_context: 0 pernet_ops_rwsem &ht->mutex &rcu_state.expedited_wq &p->pi_lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)inet_frag_wq (work_completion)(&fqdir->destroy_work) &ht->mutex rcu_node_0 irq_context: 0 (wq_completion)inet_frag_wq (work_completion)(&fqdir->destroy_work) &ht->mutex &rcu_state.expedited_wq irq_context: 0 (wq_completion)inet_frag_wq (work_completion)(&fqdir->destroy_work) &ht->mutex &rcu_state.expedited_wq &p->pi_lock irq_context: 0 (wq_completion)inet_frag_wq (work_completion)(&fqdir->destroy_work) &ht->mutex &rcu_state.expedited_wq &p->pi_lock &rq->__lock irq_context: 0 (wq_completion)inet_frag_wq (work_completion)(&fqdir->destroy_work) &ht->mutex &rcu_state.expedited_wq &p->pi_lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 rtnl_mutex rcu_state.exp_mutex rcu_read_lock &pool->lock fill_pool_map-wait-type-override &c->lock irq_context: 0 rtnl_mutex rcu_state.exp_mutex rcu_read_lock &pool->lock fill_pool_map-wait-type-override &n->list_lock irq_context: 0 rtnl_mutex rcu_state.exp_mutex rcu_read_lock &pool->lock fill_pool_map-wait-type-override &n->list_lock &c->lock irq_context: 0 dup_mmap_sem &mm->mmap_lock &mm->mmap_lock/1 fill_pool_map-wait-type-override pool_lock#2 irq_context: 0 dup_mmap_sem &mm->mmap_lock &mm->mmap_lock/1 fill_pool_map-wait-type-override &c->lock irq_context: 0 dup_mmap_sem &mm->mmap_lock &mm->mmap_lock/1 fill_pool_map-wait-type-override pool_lock irq_context: 0 tomoyo_ss remove_cache_srcu rcu_read_lock &rcu_state.expedited_wq irq_context: 0 tomoyo_ss remove_cache_srcu rcu_read_lock &rcu_state.expedited_wq &p->pi_lock irq_context: 0 tomoyo_ss remove_cache_srcu rcu_read_lock &rcu_state.expedited_wq &p->pi_lock &rq->__lock irq_context: 0 tomoyo_ss remove_cache_srcu rcu_read_lock &rcu_state.expedited_wq &p->pi_lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 pernet_ops_rwsem &root->kernfs_rwsem &p->pi_lock irq_context: 0 pernet_ops_rwsem rtnl_mutex bpf_devs_lock rcu_read_lock rcu_node_0 irq_context: 0 pernet_ops_rwsem rtnl_mutex bpf_devs_lock rcu_read_lock &rq->__lock irq_context: 0 pernet_ops_rwsem rtnl_mutex bpf_devs_lock rcu_read_lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)bond8#2 &rq->__lock irq_context: 0 sk_lock-AF_AX25 irq_context: 0 sk_lock-AF_AX25 slock-AF_AX25 irq_context: 0 slock-AF_AX25 irq_context: 0 &ret->b_state_lock &journal->j_list_lock rcu_read_lock &p->pi_lock irq_context: 0 &ret->b_state_lock &journal->j_list_lock rcu_read_lock &p->pi_lock &rq->__lock irq_context: 0 &ret->b_state_lock &journal->j_list_lock rcu_read_lock &p->pi_lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 rtnl_mutex &block->cb_lock irq_context: 0 sb_writers#4 mapping.invalidate_lock rcu_read_lock fill_pool_map-wait-type-override pool_lock#2 irq_context: 0 sb_writers#4 mapping.invalidate_lock rcu_read_lock fill_pool_map-wait-type-override &c->lock irq_context: 0 sb_writers#4 mapping.invalidate_lock rcu_read_lock fill_pool_map-wait-type-override &n->list_lock irq_context: 0 sb_writers#4 mapping.invalidate_lock rcu_read_lock fill_pool_map-wait-type-override &n->list_lock &c->lock irq_context: 0 sb_writers#4 mapping.invalidate_lock rcu_read_lock fill_pool_map-wait-type-override pool_lock irq_context: 0 sb_writers#4 mapping.invalidate_lock &rq->__lock &cfs_rq->removed.lock irq_context: 0 (wq_completion)wg-crypt-wg0#5 (work_completion)(&peer->transmit_packet_work) &rq->__lock &cfs_rq->removed.lock irq_context: 0 pernet_ops_rwsem devices_rwsem rdma_nets_rwsem &device->compat_devs_mutex &rxe->usdev_lock &cfs_rq->removed.lock irq_context: 0 pernet_ops_rwsem devices_rwsem rdma_nets_rwsem &device->compat_devs_mutex &rxe->usdev_lock &obj_hash[i].lock irq_context: 0 pernet_ops_rwsem devices_rwsem rdma_nets_rwsem &device->compat_devs_mutex &rxe->usdev_lock pool_lock#2 irq_context: 0 rtnl_mutex sk_lock-AF_INET rcu_read_lock &im->lock irq_context: 0 rtnl_mutex sk_lock-AF_INET &im->lock pool_lock#2 irq_context: 0 rtnl_mutex sk_lock-AF_INET &im->lock &obj_hash[i].lock irq_context: 0 (wq_completion)events_power_efficient (work_completion)(&(&tbl->gc_work)->work) &tbl->lock batched_entropy_u32.lock irq_context: 0 (wq_completion)events_power_efficient (work_completion)(&(&tbl->gc_work)->work) &tbl->lock batched_entropy_u32.lock crngs.lock irq_context: 0 pernet_ops_rwsem nf_ct_proto_mutex defrag6_mutex nf_hook_mutex batched_entropy_u8.lock irq_context: 0 pernet_ops_rwsem nf_ct_proto_mutex defrag6_mutex nf_hook_mutex kfence_freelist_lock irq_context: 0 (wq_completion)wg-kex-wg2#4 (work_completion)(&({ do { const void *__vpp_verify = (typeof((worker) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __asm__ ("" : "=r"(__ptr) : "0"((typeof(*((worker))) *)((worker)))); (typeof((typeof(*((worker))) *)((worker)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); })->work) &handshake->lock &n->list_lock irq_context: 0 (wq_completion)wg-kex-wg2#4 (work_completion)(&({ do { const void *__vpp_verify = (typeof((worker) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __asm__ ("" : "=r"(__ptr) : "0"((typeof(*((worker))) *)((worker)))); (typeof((typeof(*((worker))) *)((worker)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); })->work) &handshake->lock &n->list_lock &c->lock irq_context: 0 &mm->mmap_lock ptlock_ptr(page)#2 &pcp->lock &zone->lock &____s->seqcount irq_context: 0 &tsk->futex_exit_mutex &mm->mmap_lock rcu_read_lock per_cpu_ptr(&cgroup_rstat_cpu_lock, cpu) irq_context: 0 sb_writers#4 mapping.invalidate_lock rcu_read_lock &base->lock irq_context: 0 &sb->s_type->i_mutex_key#8 &sbi->s_writepages_rwsem jbd2_handle &ei->i_data_sem &ei->i_prealloc_lock &pa->pa_lock#2 irq_context: 0 (wq_completion)bat_events (work_completion)(&(&bat_priv->mcast.work)->work) rcu_read_lock batched_entropy_u8.lock crngs.lock irq_context: 0 sb_writers#4 mapping.invalidate_lock rcu_read_lock &base->lock &obj_hash[i].lock irq_context: 0 cb_lock genl_mutex console_owner_lock irq_context: 0 cb_lock genl_mutex console_owner irq_context: 0 &xt[i].mutex remove_cache_srcu rcu_read_lock &rcu_state.gp_wq irq_context: 0 &xt[i].mutex remove_cache_srcu rcu_read_lock &rcu_state.gp_wq &p->pi_lock irq_context: 0 (wq_completion)bond5#2 irq_context: 0 (wq_completion)bond5#2 (work_completion)(&(&slave->notify_work)->work) irq_context: 0 (wq_completion)bond5#2 (work_completion)(&(&slave->notify_work)->work) &obj_hash[i].lock irq_context: 0 (wq_completion)bond5#2 (work_completion)(&(&slave->notify_work)->work) &base->lock irq_context: 0 (wq_completion)bond5#2 (work_completion)(&(&slave->notify_work)->work) &base->lock &obj_hash[i].lock irq_context: 0 &u->iolock fs_reclaim irq_context: 0 &u->iolock fs_reclaim mmu_notifier_invalidate_range_start irq_context: 0 &u->iolock fs_reclaim mmu_notifier_invalidate_range_start &rq->__lock irq_context: 0 &u->iolock unix_gc_lock irq_context: 0 &u->iolock &c->lock irq_context: 0 &u->iolock &____s->seqcount#2 irq_context: 0 &u->iolock &____s->seqcount irq_context: 0 &u->iolock &n->list_lock irq_context: 0 &u->iolock &n->list_lock &c->lock irq_context: 0 (wq_completion)events (work_completion)(&rdev->wiphy_work) &rdev->wiphy.mtx &wdev->mtx &local->sta_mtx rcu_state.exp_mutex &rnp->exp_wq[3] irq_context: 0 (wq_completion)events (work_completion)(&rdev->wiphy_work) &rdev->wiphy.mtx &wdev->mtx &local->sta_mtx &n->list_lock irq_context: 0 (wq_completion)events (work_completion)(&rdev->wiphy_work) &rdev->wiphy.mtx &wdev->mtx &local->sta_mtx &n->list_lock &c->lock irq_context: 0 &u->iolock batched_entropy_u8.lock irq_context: 0 rtnl_mutex rcu_read_lock &bond->stats_lock/1 rcu_read_lock &bond->stats_lock irq_context: 0 &u->iolock remove_cache_srcu irq_context: 0 &u->iolock remove_cache_srcu quarantine_lock irq_context: 0 (wq_completion)bond5#2 (work_completion)(&(&slave->notify_work)->work) rtnl_mutex net_rwsem irq_context: 0 (wq_completion)bond5#2 (work_completion)(&(&slave->notify_work)->work) rtnl_mutex net_rwsem &list->lock#2 irq_context: 0 (wq_completion)bond5#2 (work_completion)(&(&slave->notify_work)->work) rtnl_mutex &tn->lock irq_context: 0 sb_writers#11 &type->i_mutex_dir_key#7/1 cgroup_mutex blkcg_pol_mutex &n->list_lock irq_context: 0 sb_writers#11 &type->i_mutex_dir_key#7/1 cgroup_mutex blkcg_pol_mutex &n->list_lock &c->lock irq_context: 0 sb_writers#11 &type->i_mutex_dir_key#7/1 cgroup_mutex blkcg_pol_mutex &rq->__lock irq_context: 0 sb_writers#11 &type->i_mutex_dir_key#7/1 cgroup_mutex blkcg_pol_mutex &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 &f->f_pos_lock sb_writers#11 &of->mutex cgroup_mutex irq_context: 0 &f->f_pos_lock sb_writers#11 &of->mutex cgroup_mutex cpu_hotplug_lock irq_context: 0 &f->f_pos_lock sb_writers#11 &of->mutex cgroup_mutex cpu_hotplug_lock css_set_lock irq_context: 0 &f->f_pos_lock sb_writers#11 &of->mutex cgroup_mutex cpu_hotplug_lock &rq->__lock irq_context: 0 &f->f_pos_lock sb_writers#11 &of->mutex cgroup_mutex cpu_hotplug_lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 &f->f_pos_lock sb_writers#11 &of->mutex cgroup_mutex cpu_hotplug_lock fs_reclaim irq_context: 0 &f->f_pos_lock sb_writers#11 &of->mutex cgroup_mutex cpu_hotplug_lock fs_reclaim mmu_notifier_invalidate_range_start irq_context: 0 &f->f_pos_lock sb_writers#11 &of->mutex cgroup_mutex cpu_hotplug_lock &c->lock irq_context: 0 &f->f_pos_lock sb_writers#11 &of->mutex cgroup_mutex cpu_hotplug_lock &n->list_lock irq_context: 0 &f->f_pos_lock sb_writers#11 &of->mutex cgroup_mutex cpu_hotplug_lock &n->list_lock &c->lock irq_context: 0 &f->f_pos_lock sb_writers#11 &of->mutex cgroup_mutex cpu_hotplug_lock css_set_lock cgroup_file_kn_lock irq_context: 0 &f->f_pos_lock sb_writers#11 &of->mutex cgroup_mutex cpu_hotplug_lock css_set_lock &p->pi_lock irq_context: 0 &f->f_pos_lock sb_writers#11 &of->mutex cgroup_mutex cpu_hotplug_lock css_set_lock &p->pi_lock &rq->__lock irq_context: 0 &f->f_pos_lock sb_writers#11 &of->mutex cgroup_mutex cpu_hotplug_lock css_set_lock &p->pi_lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 &f->f_pos_lock sb_writers#11 &of->mutex cgroup_mutex cpu_hotplug_lock freezer_mutex irq_context: 0 &f->f_pos_lock sb_writers#11 &of->mutex cgroup_mutex cpu_hotplug_lock freezer_mutex freezer_lock irq_context: 0 &f->f_pos_lock sb_writers#11 &of->mutex cgroup_mutex cpu_hotplug_lock freezer_mutex freezer_lock rcu_read_lock &sighand->siglock irq_context: 0 &f->f_pos_lock sb_writers#11 &of->mutex cgroup_mutex cpu_hotplug_lock freezer_mutex freezer_lock &sighand->siglock irq_context: 0 &f->f_pos_lock sb_writers#11 &of->mutex cgroup_mutex cpu_hotplug_lock freezer_mutex freezer_lock &sighand->siglock &p->pi_lock irq_context: 0 &f->f_pos_lock sb_writers#11 &of->mutex cgroup_mutex cpu_hotplug_lock freezer_mutex freezer_lock &sighand->siglock &p->pi_lock &rq->__lock irq_context: 0 &f->f_pos_lock sb_writers#11 &of->mutex cgroup_mutex cpu_hotplug_lock &p->alloc_lock irq_context: 0 &f->f_pos_lock sb_writers#11 &of->mutex cgroup_mutex cpu_hotplug_lock &p->alloc_lock &newf->file_lock irq_context: 0 &f->f_pos_lock sb_writers#11 &of->mutex cgroup_mutex (wq_completion)cpuset_migrate_mm irq_context: 0 &f->f_pos_lock sb_writers#11 &of->mutex cgroup_mutex &wq->mutex irq_context: 0 &f->f_pos_lock sb_writers#11 &of->mutex cgroup_mutex &wq->mutex &pool->lock/1 irq_context: 0 &f->f_pos_lock sb_writers#11 &of->mutex cgroup_mutex &wq->mutex &x->wait#10 irq_context: 0 &f->f_pos_lock sb_writers#11 &of->mutex cgroup_mutex &wq->mutex &rq->__lock irq_context: 0 &f->f_pos_lock sb_writers#11 &of->mutex cgroup_mutex &wq->mutex &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 &u->iolock remove_cache_srcu &c->lock irq_context: 0 &u->iolock remove_cache_srcu &n->list_lock irq_context: 0 &u->iolock remove_cache_srcu &rq->__lock irq_context: 0 &u->iolock remove_cache_srcu &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 &f->f_pos_lock sb_writers#11 &of->mutex cgroup_mutex cpu_hotplug_lock cgroup_threadgroup_rwsem irq_context: 0 &f->f_pos_lock sb_writers#11 &of->mutex cgroup_mutex cpu_hotplug_lock cgroup_threadgroup_rwsem cgroup_threadgroup_rwsem.rss.gp_wait.lock irq_context: 0 &f->f_pos_lock sb_writers#11 &of->mutex cgroup_mutex cpu_hotplug_lock cgroup_threadgroup_rwsem rcu_state.exp_mutex rcu_node_0 irq_context: 0 &f->f_pos_lock sb_writers#11 &of->mutex cgroup_mutex cpu_hotplug_lock cgroup_threadgroup_rwsem rcu_state.exp_mutex &obj_hash[i].lock irq_context: 0 &f->f_pos_lock sb_writers#11 &of->mutex cgroup_mutex cpu_hotplug_lock cgroup_threadgroup_rwsem rcu_state.exp_mutex rcu_read_lock &pool->lock irq_context: 0 &f->f_pos_lock sb_writers#11 &of->mutex cgroup_mutex cpu_hotplug_lock cgroup_threadgroup_rwsem rcu_state.exp_mutex rcu_read_lock &pool->lock &obj_hash[i].lock irq_context: 0 &f->f_pos_lock sb_writers#11 &of->mutex cgroup_mutex cpu_hotplug_lock cgroup_threadgroup_rwsem rcu_state.exp_mutex rcu_read_lock &pool->lock &p->pi_lock irq_context: 0 &f->f_pos_lock sb_writers#11 &of->mutex cgroup_mutex cpu_hotplug_lock cgroup_threadgroup_rwsem rcu_state.exp_mutex rcu_read_lock &pool->lock &p->pi_lock &rq->__lock irq_context: 0 &u->iolock remove_cache_srcu rcu_read_lock pool_lock#2 irq_context: 0 &f->f_pos_lock sb_writers#11 &of->mutex cgroup_mutex cpu_hotplug_lock cgroup_threadgroup_rwsem rcu_state.exp_mutex rcu_read_lock &pool->lock &p->pi_lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 &u->iolock remove_cache_srcu &obj_hash[i].lock irq_context: 0 &f->f_pos_lock sb_writers#11 &of->mutex cgroup_mutex cpu_hotplug_lock cgroup_threadgroup_rwsem rcu_state.exp_mutex &rq->__lock irq_context: 0 &f->f_pos_lock sb_writers#11 &of->mutex cgroup_mutex cpu_hotplug_lock cgroup_threadgroup_rwsem rcu_state.exp_mutex &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 &f->f_pos_lock sb_writers#11 &of->mutex cgroup_mutex cpu_hotplug_lock cgroup_threadgroup_rwsem rcu_state.exp_mutex &rnp->exp_wq[0] irq_context: 0 &u->iolock remove_cache_srcu rcu_node_0 irq_context: 0 &u->iolock remove_cache_srcu rcu_read_lock rcu_node_0 irq_context: 0 &u->iolock remove_cache_srcu rcu_read_lock &rq->__lock irq_context: 0 &u->iolock remove_cache_srcu rcu_read_lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 &u->iolock remove_cache_srcu rcu_read_lock &rcu_state.gp_wq irq_context: 0 &u->iolock remove_cache_srcu rcu_read_lock &rcu_state.gp_wq &p->pi_lock irq_context: 0 &u->iolock remove_cache_srcu rcu_read_lock &rcu_state.gp_wq &p->pi_lock &rq->__lock irq_context: 0 &u->iolock remove_cache_srcu rcu_read_lock &rcu_state.gp_wq &p->pi_lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 &f->f_pos_lock sb_writers#11 &of->mutex cgroup_mutex cpu_hotplug_lock cgroup_threadgroup_rwsem &obj_hash[i].lock irq_context: 0 &f->f_pos_lock sb_writers#11 &of->mutex cgroup_mutex cpu_hotplug_lock cgroup_threadgroup_rwsem css_set_lock irq_context: 0 &f->f_pos_lock sb_writers#11 &of->mutex cgroup_mutex cpu_hotplug_lock cgroup_threadgroup_rwsem fs_reclaim irq_context: 0 &f->f_pos_lock sb_writers#11 &of->mutex cgroup_mutex cpu_hotplug_lock cgroup_threadgroup_rwsem fs_reclaim mmu_notifier_invalidate_range_start irq_context: 0 &f->f_pos_lock sb_writers#11 &of->mutex cgroup_mutex cpu_hotplug_lock cgroup_threadgroup_rwsem &c->lock irq_context: 0 &f->f_pos_lock sb_writers#11 &of->mutex cgroup_mutex cpu_hotplug_lock cgroup_threadgroup_rwsem css_set_lock cgroup_file_kn_lock irq_context: 0 &f->f_pos_lock sb_writers#11 &of->mutex cgroup_mutex cpu_hotplug_lock cgroup_threadgroup_rwsem css_set_lock &p->pi_lock irq_context: 0 &f->f_pos_lock sb_writers#11 &of->mutex cgroup_mutex cpu_hotplug_lock cgroup_threadgroup_rwsem css_set_lock &p->pi_lock &rq->__lock irq_context: 0 &f->f_pos_lock sb_writers#11 &of->mutex cgroup_mutex cpu_hotplug_lock cgroup_threadgroup_rwsem css_set_lock &p->pi_lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 &f->f_pos_lock sb_writers#11 &of->mutex cgroup_mutex cpu_hotplug_lock cgroup_threadgroup_rwsem freezer_mutex irq_context: 0 &f->f_pos_lock sb_writers#11 &of->mutex cgroup_mutex cpu_hotplug_lock cgroup_threadgroup_rwsem freezer_mutex freezer_lock irq_context: 0 &f->f_pos_lock sb_writers#11 &of->mutex cgroup_mutex cpu_hotplug_lock cgroup_threadgroup_rwsem freezer_mutex freezer_lock rcu_read_lock &sighand->siglock irq_context: 0 &f->f_pos_lock sb_writers#11 &of->mutex cgroup_mutex cpu_hotplug_lock cgroup_threadgroup_rwsem freezer_mutex freezer_lock &sighand->siglock irq_context: 0 &f->f_pos_lock sb_writers#11 &of->mutex cgroup_mutex cpu_hotplug_lock cgroup_threadgroup_rwsem freezer_mutex freezer_lock &sighand->siglock &p->pi_lock irq_context: 0 &f->f_pos_lock sb_writers#11 &of->mutex cgroup_mutex cpu_hotplug_lock cgroup_threadgroup_rwsem freezer_mutex freezer_lock &p->pi_lock irq_context: 0 &f->f_pos_lock sb_writers#11 &of->mutex cgroup_mutex cpu_hotplug_lock cgroup_threadgroup_rwsem freezer_mutex freezer_lock &sighand->siglock &p->pi_lock &rq->__lock irq_context: 0 &f->f_pos_lock sb_writers#11 &of->mutex cgroup_mutex cpu_hotplug_lock cgroup_threadgroup_rwsem &p->alloc_lock irq_context: 0 &f->f_pos_lock sb_writers#11 &of->mutex cgroup_mutex cpu_hotplug_lock cgroup_threadgroup_rwsem &p->alloc_lock &newf->file_lock irq_context: 0 &f->f_pos_lock sb_writers#11 &of->mutex cgroup_mutex cpu_hotplug_lock cgroup_threadgroup_rwsem css_set_lock &obj_hash[i].lock irq_context: 0 &f->f_pos_lock sb_writers#11 &of->mutex cgroup_mutex cpu_hotplug_lock cgroup_threadgroup_rwsem css_set_lock krc.lock irq_context: 0 &f->f_pos_lock sb_writers#11 &of->mutex cgroup_mutex cpu_hotplug_lock cgroup_threadgroup_rwsem css_set_lock krc.lock &obj_hash[i].lock irq_context: 0 &f->f_pos_lock sb_writers#11 &of->mutex cgroup_mutex cpu_hotplug_lock cgroup_threadgroup_rwsem css_set_lock krc.lock &base->lock irq_context: 0 &f->f_pos_lock sb_writers#11 &of->mutex cgroup_mutex cpu_hotplug_lock cgroup_threadgroup_rwsem css_set_lock krc.lock &base->lock &obj_hash[i].lock irq_context: 0 &f->f_pos_lock sb_writers#11 &of->mutex cgroup_mutex cpu_hotplug_lock cgroup_threadgroup_rwsem.waiters.lock irq_context: 0 &f->f_pos_lock sb_writers#11 &of->mutex cgroup_mutex cpu_hotplug_lock cgroup_threadgroup_rwsem.rss.gp_wait.lock irq_context: 0 &f->f_pos_lock sb_writers#11 &of->mutex cgroup_mutex cpu_hotplug_lock cgroup_threadgroup_rwsem.rss.gp_wait.lock &obj_hash[i].lock irq_context: 0 kn->active#55 &n->list_lock irq_context: 0 kn->active#55 &n->list_lock &c->lock irq_context: 0 &f->f_pos_lock sb_writers#11 &of->mutex cgroup_mutex cpu_hotplug_lock cgroup_threadgroup_rwsem rcu_state.exp_mutex &rnp->exp_wq[1] irq_context: 0 (wq_completion)writeback (work_completion)(&(&wb->dwork)->work) &rq->__lock &cfs_rq->removed.lock irq_context: 0 &u->iolock fs_reclaim &rq->__lock irq_context: 0 (wq_completion)bond7#2 irq_context: 0 (wq_completion)bond7#2 (work_completion)(&(&slave->notify_work)->work) irq_context: 0 (wq_completion)bond7#2 (work_completion)(&(&slave->notify_work)->work) &obj_hash[i].lock irq_context: 0 (wq_completion)bond7#2 (work_completion)(&(&slave->notify_work)->work) &base->lock irq_context: 0 (wq_completion)bond7#2 (work_completion)(&(&slave->notify_work)->work) &base->lock &obj_hash[i].lock irq_context: 0 &u->iolock fs_reclaim &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 &u->iolock fs_reclaim &cfs_rq->removed.lock irq_context: 0 &u->iolock fs_reclaim &obj_hash[i].lock irq_context: 0 &u->iolock fs_reclaim pool_lock#2 irq_context: 0 &f->f_pos_lock sb_writers#4 &sb->s_type->i_mutex_key#8 jbd2_handle &sbi->s_orphan_lock &mapping->private_lock irq_context: 0 (wq_completion)bond7#2 (work_completion)(&(&slave->notify_work)->work) rtnl_mutex net_rwsem irq_context: 0 (wq_completion)bond7#2 (work_completion)(&(&slave->notify_work)->work) rtnl_mutex net_rwsem &list->lock#2 irq_context: 0 (wq_completion)bond7#2 (work_completion)(&(&slave->notify_work)->work) rtnl_mutex &tn->lock irq_context: 0 (wq_completion)bond7#2 (work_completion)(&(&slave->notify_work)->work) rtnl_mutex.wait_lock irq_context: 0 (wq_completion)bond7#2 (work_completion)(&(&slave->notify_work)->work) &p->pi_lock irq_context: 0 &u->iolock fs_reclaim mmu_notifier_invalidate_range_start &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 sb_writers#4 &sb->s_type->i_mutex_key#8 &folio_wait_table[i] &p->pi_lock &cfs_rq->removed.lock irq_context: 0 sb_writers#4 &type->i_mutex_dir_key#3 jbd2_handle &rq->__lock &cfs_rq->removed.lock irq_context: 0 sb_writers#14 irq_context: 0 sb_writers#14 mount_lock irq_context: 0 sb_writers#4 mapping.invalidate_lock rcu_read_lock &pcp->lock &zone->lock irq_context: 0 sb_writers#4 mapping.invalidate_lock rcu_read_lock &pcp->lock &zone->lock &____s->seqcount irq_context: 0 &sb->s_type->i_mutex_key#8 &sbi->s_writepages_rwsem jbd2_handle remove_cache_srcu irq_context: 0 &sb->s_type->i_mutex_key#8 &sbi->s_writepages_rwsem jbd2_handle remove_cache_srcu quarantine_lock irq_context: 0 &sb->s_type->i_mutex_key#8 &sbi->s_writepages_rwsem jbd2_handle remove_cache_srcu &c->lock irq_context: 0 &sb->s_type->i_mutex_key#8 &sbi->s_writepages_rwsem jbd2_handle remove_cache_srcu &n->list_lock irq_context: 0 &sb->s_type->i_mutex_key#8 &sbi->s_writepages_rwsem jbd2_handle remove_cache_srcu rcu_read_lock pool_lock#2 irq_context: 0 &sb->s_type->i_mutex_key#8 &sbi->s_writepages_rwsem jbd2_handle remove_cache_srcu &obj_hash[i].lock irq_context: 0 rtnl_mutex &idev->mc_lock &macsec_netdev_addr_lock_key#2/2 irq_context: 0 rtnl_mutex &idev->mc_lock &macsec_netdev_addr_lock_key#2/2 &obj_hash[i].lock irq_context: 0 rtnl_mutex &idev->mc_lock &macsec_netdev_addr_lock_key#2/2 krc.lock irq_context: 0 rtnl_mutex &macsec_netdev_addr_lock_key#2/2 irq_context: 0 rcu_read_lock rcu_read_lock_bh icmp_global.lock irq_context: 0 rcu_read_lock rcu_read_lock_bh icmp_global.lock batched_entropy_u8.lock irq_context: 0 rcu_read_lock rcu_read_lock_bh k-slock-AF_INET6 rcu_read_lock rcu_read_lock rcu_read_lock stock_lock irq_context: 0 rcu_read_lock rcu_read_lock_bh k-slock-AF_INET6 rcu_read_lock rcu_read_lock rcu_read_lock pool_lock#2 irq_context: 0 rcu_read_lock rcu_read_lock_bh k-slock-AF_INET6 rcu_read_lock rcu_read_lock rcu_read_lock &dir->lock#2 irq_context: 0 rcu_read_lock rcu_read_lock_bh k-slock-AF_INET6 rcu_read_lock pool_lock#2 irq_context: 0 rcu_read_lock rcu_read_lock_bh k-slock-AF_INET6 rcu_read_lock rcu_read_lock pool_lock#2 irq_context: 0 rcu_read_lock rcu_read_lock_bh k-slock-AF_INET6 rcu_read_lock rcu_read_lock &obj_hash[i].lock irq_context: 0 rcu_read_lock rcu_read_lock_bh k-slock-AF_INET6 rcu_read_lock rcu_read_lock &dir->lock#2 irq_context: 0 rcu_read_lock rcu_read_lock_bh k-slock-AF_INET6 rcu_read_lock rcu_read_lock &tbl->lock irq_context: 0 rcu_read_lock rcu_read_lock_bh k-slock-AF_INET6 rcu_read_lock rcu_read_lock &n->lock irq_context: 0 rcu_read_lock rcu_read_lock_bh k-slock-AF_INET6 rcu_read_lock rcu_read_lock &____s->seqcount#9 irq_context: 0 rcu_read_lock rcu_read_lock_bh k-slock-AF_INET6 rcu_read_lock rcu_read_lock rcu_read_lock_bh &sch->q.lock irq_context: 0 rcu_read_lock rcu_read_lock_bh k-slock-AF_INET6 rcu_read_lock rcu_read_lock rcu_read_lock_bh &sch->q.lock pool_lock#2 irq_context: 0 rcu_read_lock rcu_read_lock_bh k-slock-AF_INET6 rcu_read_lock rcu_read_lock rcu_read_lock_bh &sch->q.lock tk_core.seq.seqcount irq_context: 0 rcu_read_lock rcu_read_lock_bh k-slock-AF_INET6 rcu_read_lock rcu_read_lock rcu_read_lock_bh tk_core.seq.seqcount irq_context: 0 rcu_read_lock rcu_read_lock_bh k-slock-AF_INET6 rcu_read_lock rcu_read_lock rcu_read_lock_bh &list->lock#5 irq_context: 0 rcu_read_lock rcu_read_lock_bh k-slock-AF_INET6 rcu_read_lock rcu_read_lock rcu_read_lock_bh &sch->q.lock hrtimer_bases.lock irq_context: 0 rcu_read_lock rcu_read_lock_bh k-slock-AF_INET6 rcu_read_lock rcu_read_lock rcu_read_lock_bh &sch->q.lock hrtimer_bases.lock &obj_hash[i].lock irq_context: 0 rcu_read_lock rcu_read_lock_bh k-slock-AF_INET6 rcu_read_lock rcu_read_lock rcu_read_lock_bh &sch->q.lock hrtimer_bases.lock tk_core.seq.seqcount irq_context: 0 &rdma_nl_types[idx].sem link_ops_rwsem &pdata->netdev_lock &c->lock irq_context: 0 &rdma_nl_types[idx].sem link_ops_rwsem devices_rwsem devices.xa_lock pool_lock#2 irq_context: 0 &rdma_nl_types[idx].sem link_ops_rwsem &rxe->usdev_lock rtnl_mutex rtnl_mutex.wait_lock irq_context: 0 &rdma_nl_types[idx].sem link_ops_rwsem &rxe->usdev_lock rtnl_mutex &rq->__lock irq_context: 0 &rdma_nl_types[idx].sem link_ops_rwsem &rxe->usdev_lock rtnl_mutex &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 &rdma_nl_types[idx].sem link_ops_rwsem &rxe->usdev_lock (console_sem).lock irq_context: 0 &rdma_nl_types[idx].sem link_ops_rwsem &rxe->usdev_lock console_lock console_srcu console_owner_lock irq_context: 0 &rdma_nl_types[idx].sem link_ops_rwsem &rxe->usdev_lock console_lock console_srcu console_owner irq_context: 0 &rdma_nl_types[idx].sem link_ops_rwsem &rxe->usdev_lock console_lock console_srcu console_owner &port_lock_key irq_context: 0 &rdma_nl_types[idx].sem link_ops_rwsem &rxe->usdev_lock console_lock console_srcu console_owner console_owner_lock irq_context: 0 &rdma_nl_types[idx].sem link_ops_rwsem rtnl_mutex net_rwsem &table->lock#4 &c->lock irq_context: 0 rcu_read_lock rcu_read_lock_bh k-slock-AF_INET6 rcu_read_lock rcu_read_lock &____s->seqcount#10 irq_context: 0 &rdma_nl_types[idx].sem link_ops_rwsem remove_cache_srcu irq_context: 0 &rdma_nl_types[idx].sem link_ops_rwsem remove_cache_srcu quarantine_lock irq_context: 0 &rdma_nl_types[idx].sem link_ops_rwsem remove_cache_srcu &c->lock irq_context: 0 &rdma_nl_types[idx].sem link_ops_rwsem remove_cache_srcu &n->list_lock irq_context: 0 &rdma_nl_types[idx].sem link_ops_rwsem remove_cache_srcu rcu_read_lock pool_lock#2 irq_context: 0 &rdma_nl_types[idx].sem link_ops_rwsem remove_cache_srcu &obj_hash[i].lock irq_context: 0 &rdma_nl_types[idx].sem link_ops_rwsem remove_cache_srcu &rq->__lock irq_context: 0 &rdma_nl_types[idx].sem link_ops_rwsem remove_cache_srcu &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 kn->active#48 rcu_node_0 irq_context: 0 kn->active#48 &rcu_state.expedited_wq irq_context: 0 kn->active#48 &rcu_state.expedited_wq &p->pi_lock irq_context: 0 kn->active#48 &rcu_state.expedited_wq &p->pi_lock &rq->__lock irq_context: 0 kn->active#48 &rcu_state.expedited_wq &p->pi_lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 &rdma_nl_types[idx].sem link_ops_rwsem remove_cache_srcu rcu_read_lock rcu_node_0 irq_context: 0 &rdma_nl_types[idx].sem link_ops_rwsem remove_cache_srcu rcu_read_lock &rq->__lock irq_context: 0 &rdma_nl_types[idx].sem link_ops_rwsem remove_cache_srcu rcu_read_lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 &rdma_nl_types[idx].sem link_ops_rwsem rcu_read_lock &rcu_state.gp_wq irq_context: 0 &rdma_nl_types[idx].sem link_ops_rwsem rcu_read_lock &rcu_state.gp_wq &p->pi_lock irq_context: 0 &rdma_nl_types[idx].sem link_ops_rwsem rcu_read_lock &rcu_state.gp_wq &p->pi_lock &rq->__lock irq_context: 0 &rdma_nl_types[idx].sem link_ops_rwsem rcu_read_lock &rcu_state.gp_wq &p->pi_lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 &rdma_nl_types[idx].sem link_ops_rwsem rcu_node_0 irq_context: 0 (wq_completion)infiniband (work_completion)(&work->work)#2 &rxe->usdev_lock rtnl_mutex rtnl_mutex.wait_lock irq_context: 0 (wq_completion)infiniband (work_completion)(&work->work)#2 &rxe->usdev_lock rtnl_mutex &pool->lock irq_context: 0 (wq_completion)infiniband (work_completion)(&work->work)#2 &rxe->usdev_lock rtnl_mutex rcu_node_0 irq_context: 0 (wq_completion)infiniband (work_completion)(&work->work)#2 &rxe->usdev_lock rtnl_mutex &rcu_state.expedited_wq irq_context: 0 (wq_completion)infiniband (work_completion)(&work->work)#2 &rxe->usdev_lock rtnl_mutex &rcu_state.expedited_wq &p->pi_lock irq_context: 0 (wq_completion)infiniband (work_completion)(&work->work)#2 &rxe->usdev_lock rtnl_mutex &rcu_state.expedited_wq &p->pi_lock &rq->__lock irq_context: 0 (wq_completion)infiniband (work_completion)(&work->work)#2 &rxe->usdev_lock rtnl_mutex &rcu_state.expedited_wq &p->pi_lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)infiniband (work_completion)(&work->work)#2 &rxe->usdev_lock rtnl_mutex &rq->__lock irq_context: 0 (wq_completion)infiniband (work_completion)(&work->work)#2 &rxe->usdev_lock rtnl_mutex &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)infiniband (work_completion)(&work->work)#2 &rxe->usdev_lock (console_sem).lock irq_context: 0 (wq_completion)infiniband (work_completion)(&work->work)#2 &rxe->usdev_lock console_lock console_srcu console_owner_lock irq_context: 0 (wq_completion)infiniband (work_completion)(&work->work)#2 &rxe->usdev_lock console_lock console_srcu console_owner irq_context: 0 (wq_completion)infiniband (work_completion)(&work->work)#2 &rxe->usdev_lock console_lock console_srcu console_owner &port_lock_key irq_context: 0 (wq_completion)infiniband (work_completion)(&work->work)#2 &rxe->usdev_lock console_lock console_srcu console_owner console_owner_lock irq_context: 0 (wq_completion)infiniband (work_completion)(&work->work)#2 &rxe->usdev_lock &rq->__lock irq_context: 0 (wq_completion)infiniband (work_completion)(&work->work)#2 &rxe->usdev_lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 &rdma_nl_types[idx].sem link_ops_rwsem devices_rwsem clients_rwsem &device->client_data_rwsem &cq->cq_lock irq_context: 0 &rdma_nl_types[idx].sem link_ops_rwsem devices_rwsem clients_rwsem &device->client_data_rwsem stock_lock irq_context: 0 &rdma_nl_types[idx].sem link_ops_rwsem devices_rwsem clients_rwsem &device->client_data_rwsem mmu_notifier_invalidate_range_start irq_context: 0 &rdma_nl_types[idx].sem link_ops_rwsem devices_rwsem clients_rwsem &device->client_data_rwsem &sb->s_type->i_lock_key#8 irq_context: 0 &rdma_nl_types[idx].sem link_ops_rwsem devices_rwsem clients_rwsem &device->client_data_rwsem &dir->lock irq_context: 0 &rdma_nl_types[idx].sem link_ops_rwsem devices_rwsem clients_rwsem &device->client_data_rwsem rcu_read_lock &rq->__lock irq_context: 0 &rdma_nl_types[idx].sem link_ops_rwsem devices_rwsem clients_rwsem &device->client_data_rwsem rcu_read_lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 &rdma_nl_types[idx].sem link_ops_rwsem devices_rwsem clients_rwsem &device->client_data_rwsem &qp->state_lock irq_context: 0 &rdma_nl_types[idx].sem link_ops_rwsem devices_rwsem clients_rwsem &device->client_data_rwsem &xa->xa_lock#19 &c->lock irq_context: 0 &rdma_nl_types[idx].sem link_ops_rwsem devices_rwsem clients_rwsem &device->client_data_rwsem cpu_hotplug_lock irq_context: 0 &rdma_nl_types[idx].sem link_ops_rwsem devices_rwsem clients_rwsem &device->client_data_rwsem cpu_hotplug_lock wq_pool_mutex irq_context: 0 &rdma_nl_types[idx].sem link_ops_rwsem devices_rwsem clients_rwsem &device->client_data_rwsem cpu_hotplug_lock wq_pool_mutex fs_reclaim irq_context: 0 &rdma_nl_types[idx].sem link_ops_rwsem devices_rwsem clients_rwsem &device->client_data_rwsem cpu_hotplug_lock wq_pool_mutex fs_reclaim mmu_notifier_invalidate_range_start irq_context: 0 &rdma_nl_types[idx].sem link_ops_rwsem devices_rwsem clients_rwsem &device->client_data_rwsem cpu_hotplug_lock wq_pool_mutex fs_reclaim mmu_notifier_invalidate_range_start &rq->__lock irq_context: 0 &rdma_nl_types[idx].sem link_ops_rwsem devices_rwsem clients_rwsem &device->client_data_rwsem cpu_hotplug_lock wq_pool_mutex fs_reclaim mmu_notifier_invalidate_range_start &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 &rdma_nl_types[idx].sem link_ops_rwsem devices_rwsem clients_rwsem &device->client_data_rwsem cpu_hotplug_lock wq_pool_mutex pool_lock#2 irq_context: 0 &rdma_nl_types[idx].sem link_ops_rwsem devices_rwsem clients_rwsem &device->client_data_rwsem cpu_hotplug_lock wq_pool_mutex &obj_hash[i].lock irq_context: 0 &rdma_nl_types[idx].sem link_ops_rwsem devices_rwsem clients_rwsem &device->client_data_rwsem cpu_hotplug_lock wq_pool_mutex &wq->mutex irq_context: 0 &rdma_nl_types[idx].sem link_ops_rwsem devices_rwsem clients_rwsem &device->client_data_rwsem cpu_hotplug_lock wq_pool_mutex &wq->mutex &pool->lock/1 irq_context: 0 &rdma_nl_types[idx].sem link_ops_rwsem devices_rwsem clients_rwsem &device->client_data_rwsem kthread_create_lock irq_context: 0 &rdma_nl_types[idx].sem link_ops_rwsem devices_rwsem clients_rwsem &device->client_data_rwsem &x->wait irq_context: 0 &rdma_nl_types[idx].sem link_ops_rwsem devices_rwsem clients_rwsem &device->client_data_rwsem wq_pool_mutex irq_context: 0 &rdma_nl_types[idx].sem link_ops_rwsem devices_rwsem clients_rwsem &device->client_data_rwsem wq_pool_mutex &wq->mutex irq_context: 0 &rdma_nl_types[idx].sem link_ops_rwsem devices_rwsem clients_rwsem &device->client_data_rwsem &mad_queue->lock irq_context: 0 &rdma_nl_types[idx].sem link_ops_rwsem devices_rwsem clients_rwsem &device->client_data_rwsem &qp->rq.producer_lock irq_context: 0 &rdma_nl_types[idx].sem link_ops_rwsem devices_rwsem clients_rwsem &device->client_data_rwsem ib_mad_clients.xa_lock irq_context: 0 &rdma_nl_types[idx].sem link_ops_rwsem devices_rwsem clients_rwsem &device->client_data_rwsem ib_mad_clients.xa_lock pool_lock#2 irq_context: 0 &rdma_nl_types[idx].sem link_ops_rwsem devices_rwsem clients_rwsem &device->client_data_rwsem &port_priv->reg_lock irq_context: 0 &rdma_nl_types[idx].sem link_ops_rwsem devices_rwsem clients_rwsem &device->client_data_rwsem ib_agent_port_list_lock irq_context: 0 &rdma_nl_types[idx].sem link_ops_rwsem devices_rwsem clients_rwsem &device->client_data_rwsem lock kernfs_idr_lock &c->lock irq_context: 0 &rdma_nl_types[idx].sem link_ops_rwsem devices_rwsem clients_rwsem &device->client_data_rwsem &root->kernfs_rwsem &rq->__lock irq_context: 0 &rdma_nl_types[idx].sem link_ops_rwsem devices_rwsem clients_rwsem &device->client_data_rwsem &root->kernfs_rwsem &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 &rdma_nl_types[idx].sem link_ops_rwsem devices_rwsem clients_rwsem &device->client_data_rwsem &port_priv->reg_lock pool_lock#2 irq_context: 0 &rdma_nl_types[idx].sem link_ops_rwsem devices_rwsem clients_rwsem &device->client_data_rwsem &cm.device_lock irq_context: 0 &rdma_nl_types[idx].sem link_ops_rwsem devices_rwsem clients_rwsem &device->client_data_rwsem lock#7 &rq->__lock irq_context: 0 &rdma_nl_types[idx].sem link_ops_rwsem devices_rwsem clients_rwsem &device->client_data_rwsem lock#7 &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 &rdma_nl_types[idx].sem link_ops_rwsem devices_rwsem clients_rwsem &device->client_data_rwsem uevent_sock_mutex rcu_read_lock &ei->socket.wq.wait &ep->lock &ep->wq &p->pi_lock &cfs_rq->removed.lock irq_context: 0 &rdma_nl_types[idx].sem link_ops_rwsem devices_rwsem clients_rwsem &device->client_data_rwsem uevent_sock_mutex &rq->__lock irq_context: 0 &rdma_nl_types[idx].sem link_ops_rwsem devices_rwsem clients_rwsem &device->client_data_rwsem uevent_sock_mutex &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 sb_writers#3 &sb->s_type->i_mutex_key#9 tomoyo_ss &____s->seqcount#2 irq_context: 0 sb_writers &type->i_mutex_dir_key/1 tomoyo_ss rcu_read_lock &rq->__lock irq_context: 0 &rdma_nl_types[idx].sem link_ops_rwsem devices_rwsem clients_rwsem &device->client_data_rwsem rcu_read_lock &pool->lock &p->pi_lock &rq->__lock irq_context: 0 &rdma_nl_types[idx].sem link_ops_rwsem devices_rwsem clients_rwsem &device->client_data_rwsem rcu_read_lock &pool->lock &p->pi_lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)events (work_completion)(&smcibdev->port_event_work) &rxe->usdev_lock (console_sem).lock irq_context: 0 (wq_completion)events (work_completion)(&smcibdev->port_event_work) &rxe->usdev_lock console_lock console_srcu console_owner_lock irq_context: 0 (wq_completion)events (work_completion)(&smcibdev->port_event_work) &rxe->usdev_lock console_lock console_srcu console_owner irq_context: 0 (wq_completion)events (work_completion)(&smcibdev->port_event_work) &rxe->usdev_lock console_lock console_srcu console_owner &port_lock_key irq_context: 0 (wq_completion)events (work_completion)(&smcibdev->port_event_work) &rxe->usdev_lock console_lock console_srcu console_owner console_owner_lock irq_context: 0 (wq_completion)events (work_completion)(&smcibdev->port_event_work) &rxe->usdev_lock &rq->__lock irq_context: 0 &rdma_nl_types[idx].sem link_ops_rwsem devices_rwsem rdma_nets_rwsem &device->compat_devs_mutex quarantine_lock irq_context: 0 &rdma_nl_types[idx].sem link_ops_rwsem devices_rwsem rdma_nets_rwsem &device->compat_devs_mutex remove_cache_srcu rcu_read_lock rcu_node_0 irq_context: 0 &rdma_nl_types[idx].sem link_ops_rwsem devices_rwsem rdma_nets_rwsem &device->compat_devs_mutex remove_cache_srcu rcu_read_lock &rq->__lock irq_context: 0 &rdma_nl_types[idx].sem link_ops_rwsem devices_rwsem rdma_nets_rwsem &device->compat_devs_mutex remove_cache_srcu rcu_read_lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 &rdma_nl_types[idx].sem link_ops_rwsem devices_rwsem rdma_nets_rwsem &device->compat_devs_mutex remove_cache_srcu &rq->__lock irq_context: 0 &rdma_nl_types[idx].sem link_ops_rwsem devices_rwsem rdma_nets_rwsem &device->compat_devs_mutex remove_cache_srcu &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 &rdma_nl_types[idx].sem link_ops_rwsem devices_rwsem rdma_nets_rwsem &device->compat_devs_mutex &rxe->usdev_lock (console_sem).lock irq_context: 0 &rdma_nl_types[idx].sem link_ops_rwsem devices_rwsem rdma_nets_rwsem &device->compat_devs_mutex &rxe->usdev_lock console_lock console_srcu console_owner_lock irq_context: 0 &rdma_nl_types[idx].sem link_ops_rwsem devices_rwsem rdma_nets_rwsem &device->compat_devs_mutex &rxe->usdev_lock console_lock console_srcu console_owner irq_context: 0 &rdma_nl_types[idx].sem link_ops_rwsem devices_rwsem rdma_nets_rwsem &device->compat_devs_mutex &rxe->usdev_lock console_lock console_srcu console_owner &port_lock_key irq_context: 0 &rdma_nl_types[idx].sem link_ops_rwsem devices_rwsem rdma_nets_rwsem &device->compat_devs_mutex &rxe->usdev_lock console_lock console_srcu console_owner console_owner_lock irq_context: 0 &rdma_nl_types[idx].sem link_ops_rwsem devices_rwsem rdma_nets_rwsem &device->compat_devs_mutex &rxe->usdev_lock &rq->__lock irq_context: 0 &rdma_nl_types[idx].sem link_ops_rwsem devices_rwsem rdma_nets_rwsem &device->compat_devs_mutex &rxe->usdev_lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 &rdma_nl_types[idx].sem link_ops_rwsem devices_rwsem rdma_nets_rwsem &device->compat_devs_mutex &cfs_rq->removed.lock irq_context: 0 &rdma_nl_types[idx].sem link_ops_rwsem devices_rwsem rdma_nets_rwsem &device->compat_devs_mutex fs_reclaim mmu_notifier_invalidate_range_start &cfs_rq->removed.lock irq_context: 0 &rdma_nl_types[idx].sem link_ops_rwsem devices_rwsem rdma_nets_rwsem &device->compat_devs_mutex &rq->__lock &cfs_rq->removed.lock irq_context: 0 &rdma_nl_types[idx].sem link_ops_rwsem devices_rwsem rdma_nets_rwsem &device->compat_devs_mutex rcu_read_lock &rcu_state.gp_wq irq_context: 0 &rdma_nl_types[idx].sem link_ops_rwsem devices_rwsem rdma_nets_rwsem &device->compat_devs_mutex rcu_read_lock &rcu_state.gp_wq &p->pi_lock irq_context: 0 &rdma_nl_types[idx].sem link_ops_rwsem devices_rwsem rdma_nets_rwsem &device->compat_devs_mutex rcu_read_lock &rcu_state.gp_wq &p->pi_lock &rq->__lock irq_context: 0 &rdma_nl_types[idx].sem link_ops_rwsem devices_rwsem rdma_nets_rwsem &device->compat_devs_mutex rcu_read_lock &rcu_state.gp_wq &p->pi_lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 &rdma_nl_types[idx].sem link_ops_rwsem devices_rwsem rdma_nets_rwsem &device->compat_devs_mutex lock kernfs_idr_lock &n->list_lock irq_context: 0 &rdma_nl_types[idx].sem link_ops_rwsem devices_rwsem rdma_nets_rwsem &device->compat_devs_mutex lock kernfs_idr_lock &n->list_lock &c->lock irq_context: 0 cb_lock genl_mutex rcu_read_lock &pool->lock/1 pool_lock#2 irq_context: 0 (wq_completion)nbd-del irq_context: 0 (wq_completion)nbd-del (work_completion)(&nbd->remove_work) irq_context: 0 (wq_completion)nbd-del (work_completion)(&nbd->remove_work) &disk->open_mutex irq_context: 0 (wq_completion)nbd-del (work_completion)(&nbd->remove_work) &disk->open_mutex inode_hash_lock irq_context: 0 (wq_completion)nbd-del (work_completion)(&nbd->remove_work) &disk->open_mutex inode_hash_lock &sb->s_type->i_lock_key#3 irq_context: 0 (wq_completion)nbd-del (work_completion)(&nbd->remove_work) &bdev->bd_holder_lock irq_context: 0 (wq_completion)nbd-del (work_completion)(&nbd->remove_work) &bdev->bd_size_lock irq_context: 0 (wq_completion)nbd-del (work_completion)(&nbd->remove_work) &q->mq_freeze_lock irq_context: 0 (wq_completion)nbd-del (work_completion)(&nbd->remove_work) &q->mq_freeze_lock percpu_ref_switch_lock irq_context: 0 (wq_completion)nbd-del (work_completion)(&nbd->remove_work) &q->mq_freeze_lock percpu_ref_switch_lock &obj_hash[i].lock irq_context: 0 (wq_completion)nbd-del (work_completion)(&nbd->remove_work) &q->mq_freeze_lock percpu_ref_switch_lock pool_lock#2 irq_context: 0 (wq_completion)nbd-del (work_completion)(&nbd->remove_work) set->srcu irq_context: 0 (wq_completion)nbd-del (work_completion)(&nbd->remove_work) &q->mq_freeze_wq irq_context: 0 (wq_completion)nbd-del (work_completion)(&nbd->remove_work) &root->kernfs_rwsem irq_context: 0 (wq_completion)nbd-del (work_completion)(&nbd->remove_work) &root->kernfs_rwsem &root->kernfs_iattr_rwsem irq_context: 0 (wq_completion)nbd-del (work_completion)(&nbd->remove_work) &root->kernfs_rwsem kernfs_idr_lock irq_context: 0 (wq_completion)nbd-del (work_completion)(&nbd->remove_work) &root->kernfs_rwsem &obj_hash[i].lock irq_context: 0 (wq_completion)nbd-del (work_completion)(&nbd->remove_work) &root->kernfs_rwsem pool_lock#2 irq_context: 0 (wq_completion)nbd-del (work_completion)(&nbd->remove_work) (&bdi->laptop_mode_wb_timer) irq_context: 0 (wq_completion)nbd-del (work_completion)(&nbd->remove_work) &obj_hash[i].lock irq_context: 0 (wq_completion)nbd-del (work_completion)(&nbd->remove_work) &base->lock irq_context: 0 (wq_completion)nbd-del (work_completion)(&nbd->remove_work) bdi_lock irq_context: 0 (wq_completion)nbd-del (work_completion)(&nbd->remove_work) rcu_state.exp_mutex rcu_node_0 irq_context: 0 (wq_completion)nbd-del (work_completion)(&nbd->remove_work) rcu_state.exp_mutex &obj_hash[i].lock irq_context: 0 (wq_completion)nbd-del (work_completion)(&nbd->remove_work) rcu_state.exp_mutex rcu_read_lock &pool->lock irq_context: 0 (wq_completion)nbd-del (work_completion)(&nbd->remove_work) rcu_state.exp_mutex rcu_read_lock &pool->lock &obj_hash[i].lock irq_context: 0 (wq_completion)nbd-del (work_completion)(&nbd->remove_work) rcu_state.exp_mutex rcu_read_lock &pool->lock &p->pi_lock irq_context: 0 (wq_completion)nbd-del (work_completion)(&nbd->remove_work) rcu_state.exp_mutex rcu_read_lock &pool->lock &p->pi_lock &rq->__lock irq_context: 0 (wq_completion)nbd-del (work_completion)(&nbd->remove_work) rcu_state.exp_mutex rcu_read_lock &pool->lock &p->pi_lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)nbd-del (work_completion)(&nbd->remove_work) rcu_state.exp_mutex &rq->__lock irq_context: 0 (wq_completion)nbd-del (work_completion)(&nbd->remove_work) rcu_state.exp_mutex.wait_lock irq_context: 0 (wq_completion)nbd-del (work_completion)(&nbd->remove_work) &p->pi_lock irq_context: 0 (wq_completion)nbd-del (work_completion)(&nbd->remove_work) &p->pi_lock &rq->__lock irq_context: 0 (wq_completion)nbd-del (work_completion)(&nbd->remove_work) &p->pi_lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)nbd-del (work_completion)(&nbd->remove_work) &rq->__lock irq_context: 0 (wq_completion)nbd-del (work_completion)(&nbd->remove_work) &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)nbd-del (work_completion)(&nbd->remove_work) &wb->work_lock irq_context: 0 (wq_completion)nbd-del (work_completion)(&nbd->remove_work) cgwb_lock irq_context: 0 (wq_completion)nbd-del (work_completion)(&nbd->remove_work) rcu_read_lock &pool->lock/1 irq_context: 0 (wq_completion)nbd-del (work_completion)(&nbd->remove_work) rcu_read_lock &pool->lock/1 &obj_hash[i].lock irq_context: 0 (wq_completion)nbd-del (work_completion)(&nbd->remove_work) rcu_read_lock &pool->lock/1 pool_lock#2 irq_context: 0 (wq_completion)nbd-del (work_completion)(&nbd->remove_work) rcu_read_lock &pool->lock/1 &p->pi_lock irq_context: 0 (wq_completion)nbd-del (work_completion)(&nbd->remove_work) rcu_read_lock &pool->lock/1 &p->pi_lock &rq->__lock irq_context: 0 (wq_completion)nbd-del (work_completion)(&nbd->remove_work) rcu_read_lock &pool->lock/1 &p->pi_lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)nbd-del (work_completion)(&nbd->remove_work) &(&wb->dwork)->timer irq_context: 0 (wq_completion)nbd-del (work_completion)(&nbd->remove_work) (work_completion)(&(&wb->dwork)->work) irq_context: 0 (wq_completion)nbd-del (work_completion)(&nbd->remove_work) &(&wb->bw_dwork)->timer irq_context: 0 (wq_completion)nbd-del (work_completion)(&nbd->remove_work) (work_completion)(&(&wb->bw_dwork)->work) irq_context: 0 (wq_completion)nbd-del (work_completion)(&nbd->remove_work) &bdi->cgwb_release_mutex irq_context: 0 (wq_completion)nbd-del (work_completion)(&nbd->remove_work) &bdi->cgwb_release_mutex cgwb_lock irq_context: 0 (wq_completion)nbd-del (work_completion)(&nbd->remove_work) pin_fs_lock irq_context: 0 (wq_completion)nbd-del (work_completion)(&nbd->remove_work) &dentry->d_lock irq_context: 0 (wq_completion)nbd-del (work_completion)(&nbd->remove_work) &sb->s_type->i_mutex_key#3 irq_context: 0 (wq_completion)nbd-del (work_completion)(&nbd->remove_work) &sb->s_type->i_mutex_key#3 &dentry->d_lock irq_context: 0 (wq_completion)nbd-del (work_completion)(&nbd->remove_work) &sb->s_type->i_mutex_key#3 &dentry->d_lock &dentry->d_lock/1 irq_context: 0 (wq_completion)nbd-del (work_completion)(&nbd->remove_work) &sb->s_type->i_mutex_key#3 tk_core.seq.seqcount irq_context: 0 (wq_completion)nbd-del (work_completion)(&nbd->remove_work) &sb->s_type->i_mutex_key#3 rename_lock.seqcount irq_context: 0 (wq_completion)nbd-del (work_completion)(&nbd->remove_work) &sb->s_type->i_mutex_key#3 pin_fs_lock irq_context: 0 (wq_completion)nbd-del (work_completion)(&nbd->remove_work) &sb->s_type->i_mutex_key#3 rcu_read_lock mount_lock irq_context: 0 (wq_completion)nbd-del (work_completion)(&nbd->remove_work) &sb->s_type->i_mutex_key#3 rcu_read_lock mount_lock mount_lock.seqcount irq_context: 0 (wq_completion)nbd-del (work_completion)(&nbd->remove_work) &sb->s_type->i_mutex_key#3 mount_lock irq_context: 0 (wq_completion)nbd-del (work_completion)(&nbd->remove_work) &sb->s_type->i_mutex_key#3 mount_lock mount_lock.seqcount irq_context: 0 (wq_completion)nbd-del (work_completion)(&nbd->remove_work) &sb->s_type->i_mutex_key#3 rcu_read_lock &dentry->d_lock irq_context: 0 (wq_completion)nbd-del (work_completion)(&nbd->remove_work) &sb->s_type->i_mutex_key#3 &fsnotify_mark_srcu irq_context: 0 (wq_completion)nbd-del (work_completion)(&nbd->remove_work) &sb->s_type->i_mutex_key#3 &sb->s_type->i_lock_key#7 irq_context: 0 (wq_completion)nbd-del (work_completion)(&nbd->remove_work) &sb->s_type->i_mutex_key#3 &s->s_inode_list_lock irq_context: 0 (wq_completion)nbd-del (work_completion)(&nbd->remove_work) &sb->s_type->i_mutex_key#3 &xa->xa_lock#9 irq_context: 0 (wq_completion)nbd-del (work_completion)(&nbd->remove_work) &sb->s_type->i_mutex_key#3 &obj_hash[i].lock irq_context: 0 (wq_completion)nbd-del (work_completion)(&nbd->remove_work) &sb->s_type->i_mutex_key#3 pool_lock#2 irq_context: 0 (wq_completion)nbd-del (work_completion)(&nbd->remove_work) rcu_read_lock &dentry->d_lock irq_context: 0 (wq_completion)nbd-del (work_completion)(&nbd->remove_work) &fsnotify_mark_srcu irq_context: 0 (wq_completion)nbd-del (work_completion)(&nbd->remove_work) &sb->s_type->i_lock_key#7 irq_context: 0 (wq_completion)nbd-del (work_completion)(&nbd->remove_work) &s->s_inode_list_lock irq_context: 0 (wq_completion)nbd-del (work_completion)(&nbd->remove_work) &xa->xa_lock#9 irq_context: 0 (wq_completion)nbd-del (work_completion)(&nbd->remove_work) pool_lock#2 irq_context: 0 (wq_completion)nbd-del (work_completion)(&nbd->remove_work) rcu_read_lock mount_lock irq_context: 0 (wq_completion)nbd-del (work_completion)(&nbd->remove_work) rcu_read_lock mount_lock mount_lock.seqcount irq_context: 0 (wq_completion)nbd-del (work_completion)(&nbd->remove_work) mount_lock irq_context: 0 (wq_completion)nbd-del (work_completion)(&nbd->remove_work) mount_lock mount_lock.seqcount irq_context: 0 (wq_completion)nbd-del (work_completion)(&nbd->remove_work) &root->kernfs_rwsem irq_context: 0 (wq_completion)nbd-del (work_completion)(&nbd->remove_work) dev_pm_qos_sysfs_mtx irq_context: 0 (wq_completion)nbd-del (work_completion)(&nbd->remove_work) dev_pm_qos_sysfs_mtx &root->kernfs_rwsem irq_context: 0 (wq_completion)nbd-del (work_completion)(&nbd->remove_work) dev_pm_qos_sysfs_mtx &root->kernfs_rwsem irq_context: 0 (wq_completion)nbd-del (work_completion)(&nbd->remove_work) dev_pm_qos_sysfs_mtx dev_pm_qos_mtx irq_context: 0 (wq_completion)nbd-del (work_completion)(&nbd->remove_work) kernfs_idr_lock irq_context: 0 (wq_completion)nbd-del (work_completion)(&nbd->remove_work) &k->list_lock irq_context: 0 (wq_completion)nbd-del (work_completion)(&nbd->remove_work) sysfs_symlink_target_lock irq_context: 0 (wq_completion)nbd-del (work_completion)(&nbd->remove_work) subsys mutex#38 irq_context: 0 (wq_completion)nbd-del (work_completion)(&nbd->remove_work) subsys mutex#38 &k->k_lock irq_context: 0 (wq_completion)nbd-del (work_completion)(&nbd->remove_work) subsys mutex#38 &k->k_lock klist_remove_lock irq_context: 0 (wq_completion)nbd-del (work_completion)(&nbd->remove_work) &x->wait#9 irq_context: 0 (wq_completion)nbd-del (work_completion)(&nbd->remove_work) dpm_list_mtx irq_context: 0 (wq_completion)nbd-del (work_completion)(&nbd->remove_work) &dev->power.lock irq_context: 0 (wq_completion)nbd-del (work_completion)(&nbd->remove_work) deferred_probe_mutex irq_context: 0 (wq_completion)nbd-del (work_completion)(&nbd->remove_work) device_links_lock irq_context: 0 (wq_completion)nbd-del (work_completion)(&nbd->remove_work) mmu_notifier_invalidate_range_start irq_context: 0 (wq_completion)nbd-del (work_completion)(&nbd->remove_work) uevent_sock_mutex irq_context: 0 (wq_completion)nbd-del (work_completion)(&nbd->remove_work) uevent_sock_mutex mmu_notifier_invalidate_range_start irq_context: 0 (wq_completion)nbd-del (work_completion)(&nbd->remove_work) uevent_sock_mutex pool_lock#2 irq_context: 0 (wq_completion)nbd-del (work_completion)(&nbd->remove_work) uevent_sock_mutex nl_table_lock irq_context: 0 (wq_completion)nbd-del (work_completion)(&nbd->remove_work) uevent_sock_mutex rlock-AF_NETLINK irq_context: 0 (wq_completion)nbd-del (work_completion)(&nbd->remove_work) uevent_sock_mutex rcu_read_lock &ei->socket.wq.wait irq_context: 0 (wq_completion)nbd-del (work_completion)(&nbd->remove_work) uevent_sock_mutex rcu_read_lock &ei->socket.wq.wait &ep->lock irq_context: 0 (wq_completion)nbd-del (work_completion)(&nbd->remove_work) uevent_sock_mutex rcu_read_lock &ei->socket.wq.wait &ep->lock &ep->wq irq_context: 0 (wq_completion)nbd-del (work_completion)(&nbd->remove_work) uevent_sock_mutex rcu_read_lock &ei->socket.wq.wait &ep->lock &ep->wq &p->pi_lock irq_context: 0 (wq_completion)nbd-del (work_completion)(&nbd->remove_work) uevent_sock_mutex rcu_read_lock &ei->socket.wq.wait &ep->lock &ep->wq &p->pi_lock &rq->__lock irq_context: 0 (wq_completion)nbd-del (work_completion)(&nbd->remove_work) uevent_sock_mutex rcu_read_lock &ei->socket.wq.wait &ep->lock &ep->wq &p->pi_lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)nbd-del (work_completion)(&nbd->remove_work) uevent_sock_mutex rcu_read_lock rcu_node_0 irq_context: 0 (wq_completion)nbd-del (work_completion)(&nbd->remove_work) uevent_sock_mutex rcu_read_lock &rq->__lock irq_context: 0 (wq_completion)nbd-del (work_completion)(&nbd->remove_work) uevent_sock_mutex rcu_read_lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 sb_writers#4 &iint->mutex ima_extend_list_mutex rcu_read_lock rcu_node_0 irq_context: 0 sb_writers#4 &iint->mutex ima_extend_list_mutex rcu_read_lock &rq->__lock irq_context: softirq rcu_callback rcu_read_lock &q->mq_freeze_wq irq_context: 0 (wq_completion)nbd-del (work_completion)(&nbd->remove_work) uevent_sock_mutex nl_table_wait.lock irq_context: 0 (wq_completion)nbd-del (work_completion)(&nbd->remove_work) uevent_sock_mutex &obj_hash[i].lock irq_context: 0 (wq_completion)nbd-del (work_completion)(&nbd->remove_work) uevent_sock_mutex.wait_lock irq_context: 0 (wq_completion)nbd-del (work_completion)(&nbd->remove_work) gdp_mutex irq_context: 0 (wq_completion)nbd-del (work_completion)(&nbd->remove_work) &q->sysfs_lock irq_context: 0 (wq_completion)nbd-del (work_completion)(&nbd->remove_work) &q->sysfs_dir_lock irq_context: 0 (wq_completion)nbd-del (work_completion)(&nbd->remove_work) &q->sysfs_dir_lock sysfs_symlink_target_lock irq_context: 0 (wq_completion)nbd-del (work_completion)(&nbd->remove_work) &q->sysfs_dir_lock &root->kernfs_rwsem irq_context: 0 (wq_completion)nbd-del (work_completion)(&nbd->remove_work) &q->sysfs_dir_lock &root->kernfs_rwsem &root->kernfs_iattr_rwsem irq_context: 0 (wq_completion)nbd-del (work_completion)(&nbd->remove_work) &q->sysfs_dir_lock &obj_hash[i].lock irq_context: 0 (wq_completion)nbd-del (work_completion)(&nbd->remove_work) &q->sysfs_dir_lock pool_lock#2 irq_context: 0 (wq_completion)nbd-del (work_completion)(&nbd->remove_work) &q->sysfs_dir_lock kernfs_idr_lock irq_context: 0 (wq_completion)nbd-del (work_completion)(&nbd->remove_work) &q->sysfs_dir_lock &root->kernfs_rwsem kernfs_idr_lock irq_context: 0 (wq_completion)nbd-del (work_completion)(&nbd->remove_work) &q->sysfs_dir_lock &root->kernfs_rwsem &obj_hash[i].lock irq_context: 0 (wq_completion)nbd-del (work_completion)(&nbd->remove_work) &q->sysfs_dir_lock &root->kernfs_rwsem pool_lock#2 irq_context: 0 (wq_completion)nbd-del (work_completion)(&nbd->remove_work) &q->sysfs_dir_lock &q->sysfs_lock irq_context: 0 (wq_completion)nbd-del (work_completion)(&nbd->remove_work) &q->sysfs_dir_lock &q->sysfs_lock sysfs_symlink_target_lock irq_context: 0 (wq_completion)nbd-del (work_completion)(&nbd->remove_work) &q->sysfs_dir_lock &q->sysfs_lock &root->kernfs_rwsem irq_context: 0 (wq_completion)nbd-del (work_completion)(&nbd->remove_work) &q->sysfs_dir_lock &q->sysfs_lock &root->kernfs_rwsem &root->kernfs_iattr_rwsem irq_context: 0 (wq_completion)nbd-del (work_completion)(&nbd->remove_work) &q->sysfs_dir_lock &q->sysfs_lock &root->kernfs_rwsem kernfs_idr_lock irq_context: 0 (wq_completion)nbd-del (work_completion)(&nbd->remove_work) &q->sysfs_dir_lock &q->sysfs_lock &root->kernfs_rwsem &obj_hash[i].lock irq_context: 0 (wq_completion)nbd-del (work_completion)(&nbd->remove_work) &q->sysfs_dir_lock &q->sysfs_lock &root->kernfs_rwsem pool_lock#2 irq_context: 0 (wq_completion)nbd-del (work_completion)(&nbd->remove_work) &q->sysfs_dir_lock &q->sysfs_lock kernfs_idr_lock irq_context: 0 (wq_completion)nbd-del (work_completion)(&nbd->remove_work) &q->sysfs_dir_lock &q->sysfs_lock &obj_hash[i].lock irq_context: 0 (wq_completion)nbd-del (work_completion)(&nbd->remove_work) &q->sysfs_dir_lock &q->sysfs_lock pool_lock#2 irq_context: 0 (wq_completion)nbd-del (work_completion)(&nbd->remove_work) &q->debugfs_mutex irq_context: 0 (wq_completion)nbd-del (work_completion)(&nbd->remove_work) &q->debugfs_mutex pin_fs_lock irq_context: 0 (wq_completion)nbd-del (work_completion)(&nbd->remove_work) &q->debugfs_mutex &dentry->d_lock irq_context: 0 (wq_completion)nbd-del (work_completion)(&nbd->remove_work) &q->debugfs_mutex &sb->s_type->i_mutex_key#3 irq_context: 0 (wq_completion)nbd-del (work_completion)(&nbd->remove_work) &q->debugfs_mutex &sb->s_type->i_mutex_key#3 &dentry->d_lock irq_context: 0 (wq_completion)nbd-del (work_completion)(&nbd->remove_work) &q->debugfs_mutex &sb->s_type->i_mutex_key#3 &dentry->d_lock &dentry->d_lock/1 irq_context: 0 (wq_completion)nbd-del (work_completion)(&nbd->remove_work) &q->debugfs_mutex &sb->s_type->i_mutex_key#3 tk_core.seq.seqcount irq_context: 0 (wq_completion)nbd-del (work_completion)(&nbd->remove_work) &q->debugfs_mutex &sb->s_type->i_mutex_key#3 rename_lock.seqcount irq_context: 0 (wq_completion)nbd-del (work_completion)(&nbd->remove_work) &q->debugfs_mutex &sb->s_type->i_mutex_key#3 pin_fs_lock irq_context: 0 (wq_completion)nbd-del (work_completion)(&nbd->remove_work) &q->debugfs_mutex &sb->s_type->i_mutex_key#3 rcu_read_lock mount_lock irq_context: 0 (wq_completion)nbd-del (work_completion)(&nbd->remove_work) &q->debugfs_mutex &sb->s_type->i_mutex_key#3 rcu_read_lock mount_lock mount_lock.seqcount irq_context: 0 (wq_completion)nbd-del (work_completion)(&nbd->remove_work) &q->debugfs_mutex &sb->s_type->i_mutex_key#3 mount_lock irq_context: 0 (wq_completion)nbd-del (work_completion)(&nbd->remove_work) &q->debugfs_mutex &sb->s_type->i_mutex_key#3 mount_lock mount_lock.seqcount irq_context: 0 (wq_completion)nbd-del (work_completion)(&nbd->remove_work) &q->debugfs_mutex &sb->s_type->i_mutex_key#3 rcu_read_lock &dentry->d_lock irq_context: 0 (wq_completion)nbd-del (work_completion)(&nbd->remove_work) &q->debugfs_mutex &sb->s_type->i_mutex_key#3 &fsnotify_mark_srcu irq_context: 0 (wq_completion)nbd-del (work_completion)(&nbd->remove_work) &q->debugfs_mutex &sb->s_type->i_mutex_key#3 &sb->s_type->i_lock_key#7 irq_context: 0 (wq_completion)nbd-del (work_completion)(&nbd->remove_work) &q->debugfs_mutex &sb->s_type->i_mutex_key#3 &s->s_inode_list_lock irq_context: 0 (wq_completion)nbd-del (work_completion)(&nbd->remove_work) &q->debugfs_mutex &sb->s_type->i_mutex_key#3 &xa->xa_lock#9 irq_context: 0 (wq_completion)nbd-del (work_completion)(&nbd->remove_work) &q->debugfs_mutex &sb->s_type->i_mutex_key#3 &obj_hash[i].lock irq_context: 0 (wq_completion)nbd-del (work_completion)(&nbd->remove_work) &q->debugfs_mutex &sb->s_type->i_mutex_key#3 pool_lock#2 irq_context: 0 (wq_completion)nbd-del (work_completion)(&nbd->remove_work) &q->debugfs_mutex &sb->s_type->i_mutex_key#3 &obj_hash[i].lock pool_lock irq_context: 0 (wq_completion)nbd-del (work_completion)(&nbd->remove_work) &q->debugfs_mutex &sb->s_type->i_mutex_key#3 fill_pool_map-wait-type-override pool_lock#2 irq_context: 0 (wq_completion)nbd-del (work_completion)(&nbd->remove_work) &q->debugfs_mutex &sb->s_type->i_mutex_key#3 fill_pool_map-wait-type-override &c->lock irq_context: 0 (wq_completion)nbd-del (work_completion)(&nbd->remove_work) &q->debugfs_mutex &sb->s_type->i_mutex_key#3 fill_pool_map-wait-type-override &n->list_lock irq_context: 0 (wq_completion)nbd-del (work_completion)(&nbd->remove_work) &q->debugfs_mutex &sb->s_type->i_mutex_key#3 fill_pool_map-wait-type-override &n->list_lock &c->lock irq_context: 0 (wq_completion)nbd-del (work_completion)(&nbd->remove_work) &q->debugfs_mutex &sb->s_type->i_mutex_key#3 fill_pool_map-wait-type-override &rq->__lock irq_context: 0 (wq_completion)nbd-del (work_completion)(&nbd->remove_work) &q->debugfs_mutex &sb->s_type->i_mutex_key#3 fill_pool_map-wait-type-override &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)nbd-del (work_completion)(&nbd->remove_work) &q->debugfs_mutex &sb->s_type->i_mutex_key#3 fill_pool_map-wait-type-override pool_lock irq_context: 0 (wq_completion)nbd-del (work_completion)(&nbd->remove_work) &q->debugfs_mutex rcu_read_lock &dentry->d_lock irq_context: 0 (wq_completion)nbd-del (work_completion)(&nbd->remove_work) &q->debugfs_mutex &fsnotify_mark_srcu irq_context: 0 (wq_completion)nbd-del (work_completion)(&nbd->remove_work) &q->debugfs_mutex &sb->s_type->i_lock_key#7 irq_context: 0 (wq_completion)nbd-del (work_completion)(&nbd->remove_work) &q->debugfs_mutex &s->s_inode_list_lock irq_context: 0 (wq_completion)nbd-del (work_completion)(&nbd->remove_work) &q->debugfs_mutex &xa->xa_lock#9 irq_context: 0 (wq_completion)nbd-del (work_completion)(&nbd->remove_work) &q->debugfs_mutex &rq->__lock irq_context: 0 (wq_completion)nbd-del (work_completion)(&nbd->remove_work) &q->debugfs_mutex &obj_hash[i].lock irq_context: 0 (wq_completion)nbd-del (work_completion)(&nbd->remove_work) &q->debugfs_mutex pool_lock#2 irq_context: 0 (wq_completion)nbd-del (work_completion)(&nbd->remove_work) &q->debugfs_mutex rcu_read_lock mount_lock irq_context: 0 (wq_completion)nbd-del (work_completion)(&nbd->remove_work) &q->debugfs_mutex rcu_read_lock mount_lock mount_lock.seqcount irq_context: 0 (wq_completion)nbd-del (work_completion)(&nbd->remove_work) &q->debugfs_mutex mount_lock irq_context: 0 (wq_completion)nbd-del (work_completion)(&nbd->remove_work) &q->debugfs_mutex mount_lock mount_lock.seqcount irq_context: 0 (wq_completion)nbd-del (work_completion)(&nbd->remove_work) dev_hotplug_mutex irq_context: 0 (wq_completion)nbd-del (work_completion)(&nbd->remove_work) dev_hotplug_mutex &dev->power.lock irq_context: 0 (wq_completion)nbd-del (work_completion)(&nbd->remove_work) req_lock irq_context: 0 (wq_completion)nbd-del (work_completion)(&nbd->remove_work) &x->wait#11 irq_context: 0 &type->i_mutex_dir_key/1 tomoyo_ss &n->list_lock irq_context: 0 &type->i_mutex_dir_key/1 tomoyo_ss &n->list_lock &c->lock irq_context: 0 &type->i_mutex_dir_key/1 &rq->__lock irq_context: 0 (wq_completion)nbd-del (work_completion)(&nbd->remove_work) &root->kernfs_rwsem &____s->seqcount irq_context: 0 (wq_completion)nbd-del (work_completion)(&nbd->remove_work) &root->kernfs_rwsem rcu_read_lock pool_lock#2 irq_context: 0 (wq_completion)nbd-del (work_completion)(&nbd->remove_work) subsys mutex#37 irq_context: 0 (wq_completion)nbd-del (work_completion)(&nbd->remove_work) subsys mutex#37 &k->k_lock irq_context: 0 (wq_completion)nbd-del (work_completion)(&nbd->remove_work) subsys mutex#37 &k->k_lock klist_remove_lock irq_context: 0 (wq_completion)nbd-del (work_completion)(&nbd->remove_work) percpu_ref_switch_lock irq_context: 0 (wq_completion)nbd-del (work_completion)(&nbd->remove_work) &q->queue_lock irq_context: 0 (wq_completion)nbd-del (work_completion)(&nbd->remove_work) (&q->timeout) irq_context: 0 (wq_completion)nbd-del (work_completion)(&nbd->remove_work) (work_completion)(&q->timeout_work) irq_context: 0 (wq_completion)nbd-del (work_completion)(&nbd->remove_work) (wq_completion)kintegrityd irq_context: 0 (wq_completion)nbd-del (work_completion)(&nbd->remove_work) &wq->mutex irq_context: 0 (wq_completion)nbd-del (work_completion)(&nbd->remove_work) &wq->mutex &pool->lock irq_context: 0 (wq_completion)nbd-del (work_completion)(&nbd->remove_work) &wq->mutex &x->wait#10 irq_context: 0 (wq_completion)nbd-del (work_completion)(&nbd->remove_work) (work_completion)(&(&q->requeue_work)->work) irq_context: 0 (wq_completion)nbd-del (work_completion)(&nbd->remove_work) (work_completion)(&(&q->requeue_work)->work) &rq->__lock irq_context: 0 (wq_completion)nbd-del (work_completion)(&nbd->remove_work) (work_completion)(&(&q->requeue_work)->work) &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)nbd-del (work_completion)(&nbd->remove_work) (work_completion)(&(&hctx->run_work)->work) irq_context: 0 (wq_completion)nbd-del (work_completion)(&nbd->remove_work) &ACCESS_PRIVATE(sdp, lock) irq_context: 0 (wq_completion)nbd-del (work_completion)(&nbd->remove_work) set->srcu irq_context: 0 (wq_completion)nbd-del (work_completion)(&nbd->remove_work) &ACCESS_PRIVATE(ssp->srcu_sup, lock) &ACCESS_PRIVATE(sdp, lock) irq_context: 0 (wq_completion)nbd-del (work_completion)(&nbd->remove_work) &ACCESS_PRIVATE(ssp->srcu_sup, lock) rcu_read_lock &pool->lock irq_context: 0 (wq_completion)nbd-del (work_completion)(&nbd->remove_work) &ACCESS_PRIVATE(ssp->srcu_sup, lock) rcu_read_lock &pool->lock &obj_hash[i].lock irq_context: 0 (wq_completion)nbd-del (work_completion)(&nbd->remove_work) &ACCESS_PRIVATE(ssp->srcu_sup, lock) rcu_read_lock &pool->lock pool_lock#2 irq_context: 0 (wq_completion)nbd-del (work_completion)(&nbd->remove_work) &ACCESS_PRIVATE(ssp->srcu_sup, lock) rcu_read_lock &pool->lock &p->pi_lock irq_context: 0 (wq_completion)nbd-del (work_completion)(&nbd->remove_work) &ACCESS_PRIVATE(ssp->srcu_sup, lock) rcu_read_lock &pool->lock &p->pi_lock &rq->__lock irq_context: 0 (wq_completion)nbd-del (work_completion)(&nbd->remove_work) &ACCESS_PRIVATE(ssp->srcu_sup, lock) rcu_read_lock &pool->lock &p->pi_lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)nbd-del (work_completion)(&nbd->remove_work) &x->wait#2 irq_context: 0 (wq_completion)nbd-del (work_completion)(&nbd->remove_work) &q->sysfs_lock &q->queue_lock irq_context: 0 (wq_completion)nbd-del (work_completion)(&nbd->remove_work) &q->sysfs_lock &tags->lock irq_context: 0 (wq_completion)nbd-del (work_completion)(&nbd->remove_work) &q->sysfs_lock rcu_read_lock pool_lock#2 irq_context: 0 (wq_completion)nbd-del (work_completion)(&nbd->remove_work) &q->sysfs_lock &obj_hash[i].lock irq_context: 0 (wq_completion)nbd-del (work_completion)(&nbd->remove_work) &q->sysfs_lock &zone->lock irq_context: 0 (wq_completion)nbd-del (work_completion)(&nbd->remove_work) &q->sysfs_lock &zone->lock &____s->seqcount irq_context: 0 (wq_completion)nbd-del (work_completion)(&nbd->remove_work) &q->sysfs_lock &eq->sysfs_lock irq_context: 0 (wq_completion)nbd-del (work_completion)(&nbd->remove_work) &q->sysfs_lock &eq->sysfs_lock &q->debugfs_mutex irq_context: 0 (wq_completion)nbd-del (work_completion)(&nbd->remove_work) &q->sysfs_lock &eq->sysfs_lock &dd->lock irq_context: 0 (wq_completion)nbd-del (work_completion)(&nbd->remove_work) &q->sysfs_lock &eq->sysfs_lock &obj_hash[i].lock irq_context: 0 (wq_completion)nbd-del (work_completion)(&nbd->remove_work) &q->sysfs_lock &eq->sysfs_lock pool_lock#2 irq_context: 0 (wq_completion)nbd-del (work_completion)(&nbd->remove_work) &q->sysfs_lock &eq->sysfs_lock pcpu_lock irq_context: 0 (wq_completion)nbd-del (work_completion)(&nbd->remove_work) &q->sysfs_lock pool_lock#2 irq_context: 0 (wq_completion)nbd-del (work_completion)(&nbd->remove_work) &q->rq_qos_mutex irq_context: 0 (wq_completion)nbd-del (work_completion)(&nbd->remove_work) &q->rq_qos_mutex &stats->lock irq_context: 0 (wq_completion)nbd-del (work_completion)(&nbd->remove_work) &q->rq_qos_mutex (&cb->timer) irq_context: 0 (wq_completion)nbd-del (work_completion)(&nbd->remove_work) &q->rq_qos_mutex &obj_hash[i].lock irq_context: 0 (wq_completion)nbd-del (work_completion)(&nbd->remove_work) &q->rq_qos_mutex &base->lock irq_context: 0 (wq_completion)nbd-del (work_completion)(&nbd->remove_work) &q->rq_qos_mutex pool_lock#2 irq_context: 0 (wq_completion)nbd-del (work_completion)(&nbd->remove_work) &tags->lock irq_context: 0 (wq_completion)nbd-del (work_completion)(&nbd->remove_work) cpu_hotplug_lock irq_context: 0 (wq_completion)nbd-del (work_completion)(&nbd->remove_work) cpu_hotplug_lock cpuhp_state_mutex irq_context: 0 (wq_completion)nbd-del (work_completion)(&nbd->remove_work) &xa->xa_lock#10 irq_context: 0 (wq_completion)nbd-del (work_completion)(&nbd->remove_work) &q->unused_hctx_lock irq_context: 0 (wq_completion)nbd-del (work_completion)(&nbd->remove_work) &set->tag_list_lock irq_context: 0 (wq_completion)nbd-del (work_completion)(&nbd->remove_work) &q->queue_lock &blkcg->lock irq_context: 0 (wq_completion)nbd-del (work_completion)(&nbd->remove_work) &q->queue_lock &blkcg->lock percpu_ref_switch_lock irq_context: 0 (wq_completion)nbd-del (work_completion)(&nbd->remove_work) &q->queue_lock &blkcg->lock percpu_ref_switch_lock &obj_hash[i].lock irq_context: 0 (wq_completion)nbd-del (work_completion)(&nbd->remove_work) &q->queue_lock &blkcg->lock percpu_ref_switch_lock pool_lock#2 irq_context: 0 (wq_completion)nbd-del (work_completion)(&nbd->remove_work) (&sq->pending_timer) irq_context: 0 (wq_completion)nbd-del (work_completion)(&nbd->remove_work) (work_completion)(&td->dispatch_work) irq_context: 0 (wq_completion)nbd-del (work_completion)(&nbd->remove_work) &q->blkcg_mutex irq_context: 0 (wq_completion)nbd-del (work_completion)(&nbd->remove_work) &q->blkcg_mutex &q->queue_lock irq_context: 0 (wq_completion)nbd-del (work_completion)(&nbd->remove_work) &q->blkcg_mutex &q->queue_lock &blkcg->lock irq_context: 0 (wq_completion)nbd-del (work_completion)(&nbd->remove_work) &q->blkcg_mutex &q->queue_lock &blkcg->lock (&sq->pending_timer) irq_context: 0 (wq_completion)nbd-del (work_completion)(&nbd->remove_work) &q->blkcg_mutex &q->queue_lock &blkcg->lock &obj_hash[i].lock irq_context: 0 (wq_completion)nbd-del (work_completion)(&nbd->remove_work) &q->blkcg_mutex &q->queue_lock &blkcg->lock &base->lock irq_context: 0 (wq_completion)nbd-del (work_completion)(&nbd->remove_work) &q->blkcg_mutex &q->queue_lock &blkcg->lock percpu_counters_lock irq_context: 0 (wq_completion)nbd-del (work_completion)(&nbd->remove_work) &q->blkcg_mutex &q->queue_lock &blkcg->lock pcpu_lock irq_context: 0 (wq_completion)nbd-del (work_completion)(&nbd->remove_work) &q->blkcg_mutex &q->queue_lock &blkcg->lock pool_lock#2 irq_context: 0 (wq_completion)nbd-del (work_completion)(&nbd->remove_work) pcpu_lock irq_context: 0 (wq_completion)nbd-del (work_completion)(&nbd->remove_work) &c->lock irq_context: 0 (wq_completion)nbd-del (work_completion)(&nbd->remove_work) bio_slab_lock irq_context: 0 (wq_completion)nbd-del (work_completion)(&nbd->remove_work) &xa->xa_lock#11 irq_context: 0 (wq_completion)nbd-del (work_completion)(&nbd->remove_work) &sb->s_type->i_lock_key#3 irq_context: 0 (wq_completion)nbd-del (work_completion)(&nbd->remove_work) rcu_read_lock pool_lock#2 irq_context: 0 (wq_completion)nbd-del (work_completion)(&nbd->remove_work) &zone->lock irq_context: 0 (wq_completion)nbd-del (work_completion)(&nbd->remove_work) &zone->lock &____s->seqcount irq_context: 0 (wq_completion)nbd-del (work_completion)(&nbd->remove_work) &(&ssp->srcu_sup->work)->timer irq_context: 0 (wq_completion)nbd-del (work_completion)(&nbd->remove_work) (work_completion)(&(&ssp->srcu_sup->work)->work) irq_context: 0 (wq_completion)nbd-del (work_completion)(&nbd->remove_work) rcu_read_lock &pool->lock irq_context: 0 (wq_completion)nbd-del (work_completion)(&nbd->remove_work) (&sdp->delay_work) irq_context: 0 (wq_completion)nbd-del (work_completion)(&nbd->remove_work) (work_completion)(&sdp->work) irq_context: 0 (wq_completion)nbd-del (work_completion)(&nbd->remove_work) nbd_index_mutex irq_context: 0 sb_writers &type->i_mutex_dir_key/1 &sb->s_type->i_mutex_key#4 &simple_offset_xa_lock irq_context: 0 (wq_completion)nbd-del (work_completion)(&nbd->remove_work) &wq->mutex &pool->lock/1 irq_context: 0 (wq_completion)nbd-del (work_completion)(&nbd->remove_work) wq_mayday_lock irq_context: 0 (wq_completion)nbd-del (work_completion)(&nbd->remove_work) &x->wait irq_context: 0 (wq_completion)nbd-del (work_completion)(&nbd->remove_work) wq_pool_mutex irq_context: 0 (wq_completion)nbd-del (work_completion)(&nbd->remove_work) wq_pool_mutex &wq->mutex irq_context: 0 (wq_completion)nbd-del (work_completion)(&nbd->remove_work) wq_pool_mutex &wq->mutex &pool->lock/1 irq_context: 0 (wq_completion)nbd-del (work_completion)(&nbd->remove_work) &pool->lock/1 irq_context: 0 (wq_completion)nbd-del (work_completion)(&nbd->remove_work) &pool->lock/1 rcu_read_lock &pool->lock irq_context: 0 (wq_completion)nbd-del (work_completion)(&nbd->remove_work) &pool->lock/1 rcu_read_lock &pool->lock &obj_hash[i].lock irq_context: 0 (wq_completion)nbd-del (work_completion)(&nbd->remove_work) &pool->lock/1 rcu_read_lock &pool->lock pool_lock#2 irq_context: 0 (wq_completion)nbd-del (work_completion)(&nbd->remove_work) &pool->lock/1 rcu_read_lock &pool->lock &p->pi_lock irq_context: 0 (wq_completion)nbd-del (work_completion)(&nbd->remove_work) &pool->lock/1 rcu_read_lock &pool->lock &p->pi_lock &rq->__lock irq_context: 0 (wq_completion)nbd-del (work_completion)(&nbd->remove_work) &pool->lock/1 rcu_read_lock &pool->lock &p->pi_lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 sb_writers &type->i_mutex_dir_key/1 &sb->s_type->i_mutex_key#4 rcu_read_lock &dentry->d_lock irq_context: 0 sb_writers &type->i_mutex_dir_key/1 &sb->s_type->i_mutex_key#4 &dentry->d_lock irq_context: 0 sb_writers &s->s_inode_list_lock irq_context: 0 sb_writers &sbinfo->stat_lock irq_context: 0 sb_writers &xa->xa_lock#9 irq_context: 0 sb_writers &obj_hash[i].lock irq_context: 0 sb_writers &fsnotify_mark_srcu irq_context: 0 cb_lock genl_mutex nbd_index_mutex nbd_index_mutex.wait_lock irq_context: 0 cb_lock genl_mutex nbd_index_mutex &rq->__lock irq_context: 0 cb_lock genl_mutex nbd_index_mutex &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 &disk->open_mutex nbd_index_mutex.wait_lock irq_context: 0 &disk->open_mutex &p->pi_lock irq_context: 0 &disk->open_mutex &p->pi_lock &rq->__lock irq_context: 0 &disk->open_mutex &p->pi_lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 cb_lock genl_mutex &nbd->config_lock uevent_sock_mutex remove_cache_srcu irq_context: 0 cb_lock genl_mutex &nbd->config_lock uevent_sock_mutex remove_cache_srcu quarantine_lock irq_context: 0 cb_lock genl_mutex &nbd->config_lock uevent_sock_mutex remove_cache_srcu &c->lock irq_context: 0 cb_lock genl_mutex &nbd->config_lock uevent_sock_mutex remove_cache_srcu &n->list_lock irq_context: 0 cb_lock genl_mutex &nbd->config_lock uevent_sock_mutex remove_cache_srcu &obj_hash[i].lock irq_context: 0 cb_lock genl_mutex &nbd->config_lock uevent_sock_mutex remove_cache_srcu pool_lock#2 irq_context: 0 cb_lock genl_mutex &nbd->config_lock uevent_sock_mutex remove_cache_srcu rcu_read_lock pool_lock#2 irq_context: 0 cb_lock genl_mutex &nbd->config_lock uevent_sock_mutex remove_cache_srcu &rq->__lock irq_context: 0 cb_lock genl_mutex &nbd->config_lock uevent_sock_mutex remove_cache_srcu &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 &disk->open_mutex rcu_read_lock &pool->lock/1 irq_context: 0 &disk->open_mutex rcu_read_lock &pool->lock/1 &obj_hash[i].lock irq_context: 0 &disk->open_mutex rcu_read_lock &pool->lock/1 pool_lock#2 irq_context: 0 &disk->open_mutex rcu_read_lock &pool->lock/1 &p->pi_lock irq_context: 0 &disk->open_mutex rcu_read_lock &pool->lock/1 &p->pi_lock &rq->__lock irq_context: 0 &disk->open_mutex rcu_read_lock &pool->lock/1 &p->pi_lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)nbd-del (work_completion)(&nbd->remove_work) &disk->open_mutex &lock->wait_lock irq_context: 0 (wq_completion)nbd-del (work_completion)(&nbd->remove_work) &disk->open_mutex &rq->__lock irq_context: 0 (wq_completion)nbd-del (work_completion)(&nbd->remove_work) &disk->open_mutex &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 cb_lock genl_mutex &nbd->config_lock uevent_sock_mutex rcu_read_lock &rcu_state.gp_wq irq_context: 0 cb_lock genl_mutex &nbd->config_lock uevent_sock_mutex rcu_read_lock &rcu_state.gp_wq &p->pi_lock irq_context: 0 cb_lock genl_mutex &nbd->config_lock uevent_sock_mutex rcu_read_lock &rcu_state.gp_wq &p->pi_lock &rq->__lock irq_context: 0 cb_lock genl_mutex &nbd->config_lock uevent_sock_mutex rcu_read_lock &rcu_state.gp_wq &p->pi_lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)nbd-del (work_completion)(&nbd->remove_work) rcu_state.exp_mutex &rnp->exp_wq[1] irq_context: 0 (wq_completion)nbd-del (work_completion)(&nbd->remove_work) rcu_state.exp_mutex &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)nbd-del (work_completion)(&nbd->remove_work) rcu_read_lock &pool->lock/1 &p->pi_lock &cfs_rq->removed.lock irq_context: 0 (wq_completion)nbd-del (work_completion)(&nbd->remove_work) rcu_read_lock (wq_completion)writeback irq_context: 0 (wq_completion)nbd-del (work_completion)(&nbd->remove_work) &x->wait#10 irq_context: 0 (wq_completion)writeback (work_completion)(&barr->work) irq_context: 0 (wq_completion)writeback (work_completion)(&barr->work) &x->wait#10 irq_context: 0 (wq_completion)writeback (work_completion)(&barr->work) &x->wait#10 &p->pi_lock irq_context: 0 (wq_completion)writeback (work_completion)(&barr->work) &x->wait#10 &p->pi_lock &rq->__lock irq_context: 0 (wq_completion)writeback (work_completion)(&barr->work) &x->wait#10 &p->pi_lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)nbd-del (work_completion)(&nbd->remove_work) &q->sysfs_dir_lock &q->sysfs_lock &root->kernfs_rwsem &rq->__lock irq_context: 0 (wq_completion)nbd-del (work_completion)(&nbd->remove_work) &q->debugfs_mutex &sb->s_type->i_mutex_key#3 &rq->__lock irq_context: 0 (wq_completion)nbd-del (work_completion)(&nbd->remove_work) &q->debugfs_mutex &sb->s_type->i_mutex_key#3 &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)nbd-del (work_completion)(&nbd->remove_work) uevent_sock_mutex &c->lock irq_context: softirq rcu_callback rcu_read_lock &q->mq_freeze_wq &p->pi_lock irq_context: softirq rcu_callback rcu_read_lock &q->mq_freeze_wq &p->pi_lock &rq->__lock irq_context: softirq rcu_callback rcu_read_lock &q->mq_freeze_wq &p->pi_lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)nbd-del (work_completion)(&nbd->remove_work) &q->sysfs_dir_lock &sem->wait_lock irq_context: 0 (wq_completion)nbd-del (work_completion)(&nbd->remove_work) &q->sysfs_dir_lock &p->pi_lock irq_context: 0 (wq_completion)nbd-del (work_completion)(&nbd->remove_work) &q->sysfs_dir_lock &p->pi_lock &rq->__lock irq_context: 0 (wq_completion)nbd-del (work_completion)(&nbd->remove_work) &q->sysfs_dir_lock &p->pi_lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)nbd-del (work_completion)(&nbd->remove_work) &q->sysfs_dir_lock &rq->__lock irq_context: 0 (wq_completion)nbd-del (work_completion)(&nbd->remove_work) &q->sysfs_dir_lock &root->kernfs_rwsem &sem->wait_lock irq_context: 0 (wq_completion)nbd-del (work_completion)(&nbd->remove_work) &p->pi_lock &rq->__lock &cfs_rq->removed.lock irq_context: 0 (wq_completion)nbd-del (work_completion)(&nbd->remove_work) &cfs_rq->removed.lock irq_context: 0 (wq_completion)nbd-del (work_completion)(&nbd->remove_work) &q->blkcg_mutex &q->queue_lock &blkcg->lock pool_lock irq_context: 0 sb_writers#4 &sb->s_type->i_mutex_key#8 fill_pool_map-wait-type-override pool_lock#2 irq_context: 0 sb_writers#4 &sb->s_type->i_mutex_key#8 fill_pool_map-wait-type-override &c->lock irq_context: 0 sb_writers#4 &sb->s_type->i_mutex_key#8 fill_pool_map-wait-type-override &n->list_lock irq_context: 0 sb_writers#4 &sb->s_type->i_mutex_key#8 fill_pool_map-wait-type-override &n->list_lock &c->lock irq_context: 0 sb_writers#4 &sb->s_type->i_mutex_key#8 fill_pool_map-wait-type-override rcu_node_0 irq_context: 0 sb_writers#4 &sb->s_type->i_mutex_key#8 fill_pool_map-wait-type-override pool_lock irq_context: 0 (wq_completion)nbd-del (work_completion)(&nbd->remove_work) rcu_state.exp_mutex &rnp->exp_wq[2] irq_context: 0 (wq_completion)nbd-del (work_completion)(&nbd->remove_work) &pool->lock/1 rcu_read_lock &pool->lock &p->pi_lock &rq->__lock &cfs_rq->removed.lock irq_context: 0 (wq_completion)nbd-del (work_completion)(&nbd->remove_work) &meta->lock irq_context: 0 (wq_completion)nbd-del (work_completion)(&nbd->remove_work) kfence_freelist_lock irq_context: 0 (wq_completion)nbd-del (work_completion)(&nbd->remove_work) uevent_sock_mutex &n->list_lock irq_context: 0 (wq_completion)nbd-del (work_completion)(&nbd->remove_work) uevent_sock_mutex &n->list_lock &c->lock irq_context: 0 (wq_completion)nbd-del (work_completion)(&nbd->remove_work) uevent_sock_mutex rcu_read_lock &ei->socket.wq.wait &ep->lock &ep->wq &p->pi_lock &cfs_rq->removed.lock irq_context: 0 sock_diag_mutex sock_diag_table_mutex nlk_cb_mutex-SOCK_DIAG rcu_read_lock rcu_read_lock_bh rcu_read_lock rlock-AF_PACKET irq_context: 0 sk_lock-AF_INET rcu_read_lock rcu_read_lock rcu_read_lock_bh rcu_read_lock pool_lock#2 irq_context: 0 sk_lock-AF_INET rcu_read_lock rcu_read_lock rcu_read_lock_bh rcu_read_lock tk_core.seq.seqcount irq_context: 0 sk_lock-AF_INET rcu_read_lock rcu_read_lock rcu_read_lock_bh rcu_read_lock rlock-AF_PACKET irq_context: 0 &f->f_pos_lock sb_writers#4 &sb->s_type->i_mutex_key#8 &sbi->s_writepages_rwsem &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 &f->f_pos_lock sb_writers#4 &sb->s_type->i_mutex_key#8 &sbi->s_writepages_rwsem &folio_wait_table[i] irq_context: 0 &sbi->s_writepages_rwsem jbd2_handle rcu_read_lock rcu_read_lock rcu_read_lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 &sbi->s_writepages_rwsem &folio_wait_table[i] &p->pi_lock irq_context: 0 &sbi->s_writepages_rwsem &folio_wait_table[i] &p->pi_lock &rq->__lock irq_context: 0 &sbi->s_writepages_rwsem &folio_wait_table[i] &p->pi_lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 &f->f_pos_lock sb_writers#4 &sb->s_type->i_mutex_key#8 &sbi->s_writepages_rwsem &folio_wait_table[i] &p->pi_lock irq_context: 0 &f->f_pos_lock sb_writers#4 &sb->s_type->i_mutex_key#8 &sbi->s_writepages_rwsem &folio_wait_table[i] &p->pi_lock &rq->__lock irq_context: 0 &f->f_pos_lock sb_writers#4 &sb->s_type->i_mutex_key#8 &sbi->s_writepages_rwsem &folio_wait_table[i] &p->pi_lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 &sbi->s_writepages_rwsem &journal->j_state_lock &journal->j_wait_transaction_locked irq_context: 0 &sbi->s_writepages_rwsem jbd2_handle irq_context: 0 &f->f_pos_lock sb_writers#4 &sb->s_type->i_mutex_key#8 &sbi->s_writepages_rwsem rcu_node_0 irq_context: 0 &f->f_pos_lock sb_writers#4 &sb->s_type->i_mutex_key#8 &sbi->s_writepages_rwsem &cfs_rq->removed.lock irq_context: 0 sb_writers#4 &sb->s_type->i_mutex_key#8 rcu_read_lock &pcp->lock &zone->lock irq_context: 0 sb_writers#4 &sb->s_type->i_mutex_key#8 rcu_read_lock &pcp->lock &zone->lock &____s->seqcount irq_context: 0 sb_writers#4 &type->i_mutex_dir_key#3/1 sb_internal jbd2_handle &ei->i_data_sem &ret->b_state_lock bit_wait_table + i irq_context: softirq (&ndev->rs_timer) rcu_read_lock rcu_read_lock rcu_read_lock_bh rcu_read_lock rcu_read_lock_bh rcu_read_lock rlock-AF_PACKET irq_context: 0 sb_writers#4 sb_writers#4 jbd2_handle rcu_node_0 irq_context: 0 sb_writers#4 &type->i_mutex_dir_key#3/1 sb_internal jbd2_handle &ei->i_data_sem key#3 irq_context: 0 &f->f_pos_lock sb_writers#4 &sb->s_type->i_mutex_key#8 &sbi->s_writepages_rwsem jbd2_handle &ei->i_data_sem &bgl->locks[i].lock &ei->i_prealloc_lock irq_context: 0 &f->f_pos_lock sb_writers#4 &sb->s_type->i_mutex_key#8 &sbi->s_writepages_rwsem jbd2_handle &folio_wait_table[i] &p->pi_lock irq_context: 0 &f->f_pos_lock sb_writers#4 &sb->s_type->i_mutex_key#8 &sbi->s_writepages_rwsem jbd2_handle &folio_wait_table[i] &p->pi_lock &rq->__lock irq_context: 0 &f->f_pos_lock sb_writers#4 &sb->s_type->i_mutex_key#8 &sbi->s_writepages_rwsem jbd2_handle &folio_wait_table[i] &p->pi_lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 &f->f_pos_lock sb_writers#4 &sb->s_type->i_mutex_key#8 &sbi->s_writepages_rwsem jbd2_handle &cfs_rq->removed.lock irq_context: 0 &f->f_pos_lock sb_writers#4 &sb->s_type->i_mutex_key#8 &sbi->s_writepages_rwsem rcu_read_lock &rq->__lock &cfs_rq->removed.lock irq_context: 0 sk_lock-AF_LLC &n->list_lock irq_context: 0 sk_lock-AF_LLC &n->list_lock &c->lock irq_context: 0 &f->f_pos_lock sb_writers#4 &sb->s_type->i_mutex_key#8 &sbi->s_writepages_rwsem &n->list_lock irq_context: 0 &f->f_pos_lock sb_writers#4 &sb->s_type->i_mutex_key#8 &sbi->s_writepages_rwsem &n->list_lock &c->lock irq_context: softirq rcu_read_lock rcu_read_lock &br->multicast_lock &c->lock irq_context: softirq rcu_read_lock rcu_read_lock &br->multicast_lock pool_lock#2 irq_context: softirq rcu_read_lock rcu_read_lock &br->multicast_lock rcu_read_lock rhashtable_bucket irq_context: softirq rcu_read_lock rcu_read_lock &br->multicast_lock nl_table_lock irq_context: softirq rcu_read_lock rcu_read_lock &br->multicast_lock nl_table_wait.lock irq_context: 0 &f->f_pos_lock sb_writers#4 &sb->s_type->i_mutex_key#8 &sbi->s_writepages_rwsem jbd2_handle rcu_read_lock &rcu_state.gp_wq irq_context: 0 &f->f_pos_lock sb_writers#4 &sb->s_type->i_mutex_key#8 &sbi->s_writepages_rwsem jbd2_handle rcu_read_lock &rcu_state.gp_wq &p->pi_lock irq_context: 0 &f->f_pos_lock sb_writers#4 &sb->s_type->i_mutex_key#8 &sbi->s_writepages_rwsem jbd2_handle rcu_read_lock &rcu_state.gp_wq &p->pi_lock &rq->__lock irq_context: 0 &f->f_pos_lock sb_writers#4 &sb->s_type->i_mutex_key#8 &sbi->s_writepages_rwsem jbd2_handle rcu_read_lock &rcu_state.gp_wq &p->pi_lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 sk_lock-AF_LLC rcu_read_lock &ei->socket.wq.wait &p->pi_lock &rq->__lock &cfs_rq->removed.lock irq_context: 0 &f->f_pos_lock sb_writers#4 &sb->s_type->i_mutex_key#8 &sbi->s_writepages_rwsem jbd2_handle &mapping->private_lock irq_context: 0 &f->f_pos_lock sb_writers#4 &sb->s_type->i_mutex_key#8 &sb->s_type->i_lock_key#22 &xa->xa_lock#9 &obj_hash[i].lock pool_lock irq_context: 0 rcu_read_lock rcu_read_lock &sighand->siglock &sighand->signalfd_wqh &ep->lock &ep->wq &p->pi_lock &cfs_rq->removed.lock irq_context: softirq (&app->join_timer) rcu_read_lock_bh dev->qdisc_tx_busylock ?: &qdisc_tx_busylock quarantine_lock irq_context: 0 &u->iolock &pcp->lock &zone->lock irq_context: 0 &u->iolock &pcp->lock &zone->lock &____s->seqcount irq_context: 0 &u->iolock &mm->mmap_lock &cfs_rq->removed.lock irq_context: 0 &u->iolock &mm->mmap_lock &obj_hash[i].lock irq_context: 0 &u->iolock &p->pi_lock &cfs_rq->removed.lock irq_context: 0 rtnl_mutex &idev->mc_lock _xmit_IPGRE &obj_hash[i].lock pool_lock irq_context: 0 &f->f_pos_lock sb_writers#10 batched_entropy_u8.lock irq_context: 0 &f->f_pos_lock sb_writers#10 kfence_freelist_lock irq_context: 0 &f->f_pos_lock sb_writers#10 &meta->lock irq_context: 0 rtnl_mutex &macsec_netdev_addr_lock_key/1 &c->lock irq_context: softirq (&app->join_timer) rcu_read_lock_bh quarantine_lock irq_context: 0 rtnl_mutex _xmit_ETHER &n->list_lock irq_context: 0 rtnl_mutex _xmit_ETHER &n->list_lock &c->lock irq_context: softirq (&app->join_timer) rcu_read_lock_bh dev->qdisc_tx_busylock ?: &qdisc_tx_busylock rcu_read_lock rlock-AF_PACKET irq_context: 0 (wq_completion)events_power_efficient (work_completion)(&(&gc_work->dwork)->work) rcu_read_lock rcu_read_lock rcu_node_0 irq_context: 0 (wq_completion)events_power_efficient (work_completion)(&(&gc_work->dwork)->work) rcu_read_lock rcu_read_lock &rq->__lock irq_context: 0 (wq_completion)events_power_efficient (work_completion)(&(&gc_work->dwork)->work) rcu_read_lock rcu_read_lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 &journal->j_barrier &journal->j_checkpoint_mutex rcu_read_lock batched_entropy_u8.lock irq_context: 0 &journal->j_barrier &journal->j_checkpoint_mutex rcu_read_lock kfence_freelist_lock irq_context: 0 &journal->j_barrier &journal->j_checkpoint_mutex &journal->j_list_lock key#15 irq_context: 0 &f->f_pos_lock sb_writers#11 &mm->mmap_lock &n->list_lock irq_context: 0 &f->f_pos_lock sb_writers#11 &mm->mmap_lock &n->list_lock &c->lock irq_context: softirq (&ndev->rs_timer) rcu_read_lock rcu_read_lock rcu_read_lock_bh rcu_read_lock_bh dev->qdisc_tx_busylock ?: &qdisc_tx_busylock rcu_read_lock pool_lock#2 irq_context: softirq (&ndev->rs_timer) rcu_read_lock rcu_read_lock rcu_read_lock_bh rcu_read_lock_bh dev->qdisc_tx_busylock ?: &qdisc_tx_busylock rcu_read_lock tk_core.seq.seqcount irq_context: softirq (&ndev->rs_timer) rcu_read_lock rcu_read_lock rcu_read_lock_bh rcu_read_lock_bh dev->qdisc_tx_busylock ?: &qdisc_tx_busylock rcu_read_lock rlock-AF_PACKET irq_context: 0 (wq_completion)wg-kex-wg2#9 (work_completion)(&peer->transmit_handshake_work) &peer->endpoint_lock rcu_read_lock_bh rcu_read_lock rcu_read_lock rcu_read_lock rcu_read_lock_bh rcu_read_lock rlock-AF_PACKET irq_context: 0 pernet_ops_rwsem devices_rwsem rdma_nets_rwsem &device->compat_devs_mutex &rxe->usdev_lock (console_sem).lock irq_context: 0 pernet_ops_rwsem devices_rwsem rdma_nets_rwsem &device->compat_devs_mutex &rxe->usdev_lock console_lock console_srcu console_owner_lock irq_context: 0 pernet_ops_rwsem devices_rwsem rdma_nets_rwsem &device->compat_devs_mutex &rxe->usdev_lock console_lock console_srcu console_owner irq_context: 0 pernet_ops_rwsem devices_rwsem rdma_nets_rwsem &device->compat_devs_mutex &rxe->usdev_lock console_lock console_srcu console_owner &port_lock_key irq_context: 0 pernet_ops_rwsem devices_rwsem rdma_nets_rwsem &device->compat_devs_mutex &rxe->usdev_lock console_lock console_srcu console_owner console_owner_lock irq_context: 0 (wq_completion)mld (work_completion)(&(&idev->mc_ifc_work)->work) &idev->mc_lock rcu_read_lock rcu_read_lock rcu_read_lock_bh rcu_read_lock batched_entropy_u8.lock crngs.lock irq_context: 0 pernet_ops_rwsem rtnl_mutex uevent_sock_mutex remove_cache_srcu rcu_read_lock &____s->seqcount irq_context: 0 pernet_ops_rwsem &net->sctp.addr_wq_lock k-slock-AF_INET6/1 key#25 irq_context: 0 tracepoints_mutex reg_lock irq_context: 0 tracepoints_mutex reg_lock &rq->__lock irq_context: 0 tracepoints_mutex reg_lock fs_reclaim irq_context: 0 tracepoints_mutex reg_lock fs_reclaim mmu_notifier_invalidate_range_start irq_context: 0 tracepoints_mutex reg_lock pool_lock#2 irq_context: 0 &f->f_pos_lock sb_writers#4 &sb->s_type->i_mutex_key#8 &mm->mmap_lock lock#10 irq_context: 0 &mm->mmap_lock lock#10 irq_context: 0 &f->f_pos_lock sb_writers#4 &sb->s_type->i_mutex_key#8 &mm->mmap_lock lock#10 rcu_read_lock kernfs_rename_lock irq_context: 0 &mm->mmap_lock lock#10 rcu_read_lock kernfs_rename_lock irq_context: 0 &pipe->mutex/1 &mm->mmap_lock lock#10 irq_context: 0 &pipe->mutex/1 &mm->mmap_lock lock#10 rcu_read_lock kernfs_rename_lock irq_context: 0 &mm->mmap_lock lock#10 irq_context: 0 &mm->mmap_lock lock#10 rcu_read_lock kernfs_rename_lock irq_context: 0 tracepoints_mutex reg_lock rcu_state.exp_mutex rcu_node_0 irq_context: 0 tracepoints_mutex reg_lock rcu_state.exp_mutex &obj_hash[i].lock irq_context: 0 tracepoints_mutex reg_lock rcu_state.exp_mutex rcu_read_lock &pool->lock irq_context: 0 tracepoints_mutex reg_lock rcu_state.exp_mutex rcu_read_lock &pool->lock &obj_hash[i].lock irq_context: 0 tracepoints_mutex reg_lock rcu_state.exp_mutex rcu_read_lock &pool->lock &p->pi_lock irq_context: 0 tracepoints_mutex reg_lock rcu_state.exp_mutex rcu_read_lock &pool->lock &p->pi_lock &rq->__lock irq_context: 0 tracepoints_mutex reg_lock rcu_state.exp_mutex rcu_read_lock &pool->lock &p->pi_lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 tracepoints_mutex reg_lock rcu_state.exp_mutex &rnp->exp_wq[0] irq_context: 0 tracepoints_mutex reg_lock rcu_state.exp_mutex &rq->__lock irq_context: 0 tracepoints_mutex reg_lock rcu_state.exp_mutex &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 tracepoints_mutex reg_lock &obj_hash[i].lock irq_context: 0 tracepoints_mutex reg_lock &c->lock irq_context: 0 tracepoints_mutex reg_lock &n->list_lock irq_context: 0 tracepoints_mutex reg_lock &n->list_lock &c->lock irq_context: 0 lock#10 irq_context: 0 lock#10 rcu_read_lock kernfs_rename_lock irq_context: 0 &f->f_pos_lock sb_writers#4 &sb->s_type->i_mutex_key#8 lock#10 irq_context: 0 &f->f_pos_lock sb_writers#4 &sb->s_type->i_mutex_key#8 lock#10 rcu_read_lock kernfs_rename_lock irq_context: 0 dup_mmap_sem &mm->mmap_lock lock#10 irq_context: 0 dup_mmap_sem &mm->mmap_lock lock#10 rcu_read_lock kernfs_rename_lock irq_context: 0 dup_mmap_sem &mm->mmap_lock &mm->mmap_lock/1 lock#10 irq_context: 0 dup_mmap_sem &mm->mmap_lock &mm->mmap_lock/1 lock#10 rcu_read_lock kernfs_rename_lock irq_context: 0 &pipe->mutex/1 lock#10 irq_context: 0 &pipe->mutex/1 lock#10 rcu_read_lock kernfs_rename_lock irq_context: 0 &pipe->mutex/1 &mm->mmap_lock &rq->__lock &cfs_rq->removed.lock irq_context: 0 (wq_completion)wg-kex-wg2#8 (work_completion)(&({ do { const void *__vpp_verify = (typeof((worker) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __asm__ ("" : "=r"(__ptr) : "0"((typeof(*((worker))) *)((worker)))); (typeof((typeof(*((worker))) *)((worker)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); })->work) fill_pool_map-wait-type-override pool_lock irq_context: 0 (wq_completion)wg-crypt-wg2#4 (work_completion)(&({ do { const void *__vpp_verify = (typeof((worker) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __asm__ ("" : "=r"(__ptr) : "0"((typeof(*((worker))) *)((worker)))); (typeof((typeof(*((worker))) *)((worker)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); })->work) rcu_read_lock &pool->lock fill_pool_map-wait-type-override pool_lock irq_context: 0 sb_writers#4 &type->i_mutex_dir_key#3/1 stock_lock per_cpu_ptr(&cgroup_rstat_cpu_lock, cpu) irq_context: 0 (wq_completion)wg-kex-wg2#9 (work_completion)(&peer->transmit_handshake_work) &peer->endpoint_lock rcu_read_lock_bh rcu_read_lock &n->lock &____s->seqcount#2 irq_context: 0 (wq_completion)wg-kex-wg2#9 (work_completion)(&peer->transmit_handshake_work) &peer->endpoint_lock rcu_read_lock_bh rcu_read_lock &n->lock &____s->seqcount irq_context: 0 tracepoints_mutex reg_lock fs_reclaim &rq->__lock irq_context: 0 pernet_ops_rwsem rcu_read_lock rcu_read_lock &cfs_rq->removed.lock irq_context: 0 pernet_ops_rwsem rcu_read_lock rcu_read_lock &obj_hash[i].lock irq_context: 0 tracepoints_mutex reg_lock rcu_state.exp_mutex &rnp->exp_wq[1] irq_context: 0 pernet_ops_rwsem wq_pool_mutex &wq->mutex &rq->__lock irq_context: 0 pernet_ops_rwsem wq_pool_mutex &wq->mutex &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 pernet_ops_rwsem ovs_mutex &rq->__lock irq_context: 0 pernet_ops_rwsem ovs_mutex &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem ipvs->est_mutex &rq->__lock irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem ipvs->est_mutex &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 nfnl_subsys_queue rcu_read_lock &q->instances_lock irq_context: 0 nfnl_subsys_queue rcu_read_lock &q->instances_lock pool_lock#2 irq_context: 0 nfnl_subsys_queue rcu_read_lock rcu_node_0 irq_context: 0 nfnl_subsys_queue rcu_read_lock &rq->__lock irq_context: 0 nfnl_subsys_queue rcu_read_lock &inst->lock irq_context: 0 &sb->s_type->i_mutex_key#10 (netlink_chain).rwsem &q->instances_lock &obj_hash[i].lock irq_context: 0 &sb->s_type->i_mutex_key#10 (netlink_chain).rwsem &q->instances_lock pool_lock#2 irq_context: softirq rcu_callback &inst->lock irq_context: 0 &vma->vm_lock->lock rcu_read_lock &rcu_state.gp_wq &p->pi_lock &cfs_rq->removed.lock irq_context: 0 &pipe->mutex/1 &mm->mmap_lock rcu_read_lock &cfs_rq->removed.lock irq_context: 0 &pipe->mutex/1 &mm->mmap_lock rcu_read_lock &obj_hash[i].lock irq_context: 0 &pipe->mutex/1 &mm->mmap_lock rcu_read_lock pool_lock#2 irq_context: 0 (wq_completion)wg-kex-wg2#3 (work_completion)(&peer->transmit_handshake_work) &cookie->lock &rq->__lock irq_context: 0 (wq_completion)wg-kex-wg2#3 (work_completion)(&peer->transmit_handshake_work) &cookie->lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 sb_writers#3 oom_adj_mutex rcu_read_lock &rcu_state.expedited_wq irq_context: 0 sb_writers#3 oom_adj_mutex rcu_read_lock &rcu_state.expedited_wq &p->pi_lock irq_context: 0 sb_writers#3 oom_adj_mutex rcu_read_lock &rcu_state.expedited_wq &p->pi_lock &rq->__lock irq_context: 0 pernet_ops_rwsem uevent_sock_mutex fs_reclaim &rq->__lock irq_context: 0 pernet_ops_rwsem uevent_sock_mutex fs_reclaim &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 sb_writers#3 oom_adj_mutex rcu_read_lock &rcu_state.expedited_wq &p->pi_lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: softirq rcu_read_lock rcu_read_lock slock-AF_INET6 rcu_read_lock &ei->socket.wq.wait irq_context: softirq rcu_read_lock rcu_read_lock slock-AF_INET6 rcu_read_lock &ei->socket.wq.wait &p->pi_lock irq_context: softirq rcu_read_lock rcu_read_lock slock-AF_INET6 rcu_read_lock &ei->socket.wq.wait &p->pi_lock &cfs_rq->removed.lock irq_context: softirq rcu_read_lock rcu_read_lock slock-AF_INET6 rcu_read_lock &ei->socket.wq.wait &p->pi_lock &rq->__lock irq_context: softirq rcu_read_lock rcu_read_lock slock-AF_INET6 rcu_read_lock &ei->socket.wq.wait &p->pi_lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: softirq rcu_read_lock rcu_read_lock slock-AF_INET6 &hashinfo->ehash_locks[i] irq_context: 0 pernet_ops_rwsem rtnl_mutex fs_reclaim mmu_notifier_invalidate_range_start &rq->__lock &cfs_rq->removed.lock irq_context: 0 sb_writers#4 &sb->s_type->i_mutex_key#8 mapping.invalidate_lock &rcu_state.expedited_wq irq_context: 0 sb_writers#4 &sb->s_type->i_mutex_key#8 mapping.invalidate_lock &rcu_state.expedited_wq &p->pi_lock irq_context: 0 sb_writers#4 &sb->s_type->i_mutex_key#8 mapping.invalidate_lock &rcu_state.expedited_wq &p->pi_lock &rq->__lock irq_context: 0 sb_writers#4 &sb->s_type->i_mutex_key#8 mapping.invalidate_lock &rcu_state.expedited_wq &p->pi_lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 pernet_ops_rwsem devices_rwsem rdma_nets_rwsem &device->compat_devs_mutex &rxe->usdev_lock console_owner_lock irq_context: 0 pernet_ops_rwsem devices_rwsem rdma_nets_rwsem &device->compat_devs_mutex &rxe->usdev_lock console_owner irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem rtnl_mutex rcu_read_lock &pool->lock/1 &p->pi_lock &rq->__lock &cfs_rq->removed.lock irq_context: 0 pernet_ops_rwsem dpm_list_mtx &rq->__lock irq_context: 0 pernet_ops_rwsem dpm_list_mtx &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)writeback (work_completion)(&(&wb->dwork)->work) &type->s_umount_key#43 lock#4 &lruvec->lru_lock per_cpu_ptr(&cgroup_rstat_cpu_lock, cpu) irq_context: 0 (wq_completion)writeback (work_completion)(&(&wb->dwork)->work) &type->s_umount_key#43 lock#4 rcu_read_lock pool_lock#2 irq_context: 0 (wq_completion)writeback (work_completion)(&(&wb->dwork)->work) &type->s_umount_key#43 lock#4 &obj_hash[i].lock irq_context: 0 (wq_completion)wg-kex-wg2#9 (work_completion)(&peer->transmit_handshake_work) &peer->endpoint_lock rcu_read_lock_bh rcu_read_lock &____s->seqcount#2 irq_context: 0 (wq_completion)wg-kex-wg2#9 (work_completion)(&peer->transmit_handshake_work) &peer->endpoint_lock rcu_read_lock_bh rcu_read_lock &____s->seqcount irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem uevent_sock_mutex batched_entropy_u8.lock irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem uevent_sock_mutex kfence_freelist_lock irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem uevent_sock_mutex &meta->lock irq_context: 0 pernet_ops_rwsem ipvs->est_mutex rcu_node_0 irq_context: 0 pernet_ops_rwsem devices_rwsem rdma_nets_rwsem &device->compat_devs_mutex rcu_read_lock pgd_lock irq_context: 0 pernet_ops_rwsem devices_rwsem rdma_nets_rwsem &device->compat_devs_mutex rcu_read_lock stock_lock irq_context: 0 pernet_ops_rwsem devices_rwsem rdma_nets_rwsem &device->compat_devs_mutex rcu_read_lock rcu_read_lock pool_lock#2 irq_context: 0 pernet_ops_rwsem devices_rwsem rdma_nets_rwsem &device->compat_devs_mutex rcu_read_lock key irq_context: 0 pernet_ops_rwsem devices_rwsem rdma_nets_rwsem &device->compat_devs_mutex rcu_read_lock pcpu_lock irq_context: 0 pernet_ops_rwsem devices_rwsem rdma_nets_rwsem &device->compat_devs_mutex rcu_read_lock percpu_counters_lock irq_context: 0 pernet_ops_rwsem devices_rwsem rdma_nets_rwsem &device->compat_devs_mutex rcu_read_lock pcpu_lock stock_lock irq_context: 0 &f->f_pos_lock &type->i_mutex_dir_key#3 sb_writers#4 rcu_read_lock &rq->__lock &cfs_rq->removed.lock irq_context: 0 dup_mmap_sem &mm->mmap_lock &mm->mmap_lock/1 remove_cache_srcu rcu_read_lock &rcu_state.expedited_wq irq_context: 0 dup_mmap_sem &mm->mmap_lock &mm->mmap_lock/1 remove_cache_srcu rcu_read_lock &rcu_state.expedited_wq &p->pi_lock irq_context: 0 dup_mmap_sem &mm->mmap_lock &mm->mmap_lock/1 remove_cache_srcu rcu_read_lock &rcu_state.expedited_wq &p->pi_lock &rq->__lock irq_context: 0 dup_mmap_sem &mm->mmap_lock &mm->mmap_lock/1 remove_cache_srcu rcu_read_lock &rcu_state.expedited_wq &p->pi_lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 &xt[i].mutex remove_cache_srcu rcu_read_lock &rcu_state.gp_wq &p->pi_lock &rq->__lock irq_context: 0 &xt[i].mutex remove_cache_srcu rcu_read_lock &rcu_state.gp_wq &p->pi_lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 &type->i_mutex_dir_key#6 &sb->s_type->i_lock_key#30 &dentry->d_lock &wq#2 irq_context: 0 kn->active#67 fs_reclaim irq_context: 0 kn->active#67 fs_reclaim mmu_notifier_invalidate_range_start irq_context: 0 &nft_net->commit_mutex flowtable_lock irq_context: 0 kn->active#67 stock_lock irq_context: 0 kn->active#67 &kernfs_locks->open_file_mutex[count] irq_context: 0 &nft_net->commit_mutex cpu_hotplug_lock jump_label_mutex irq_context: 0 kn->active#67 &kernfs_locks->open_file_mutex[count] fs_reclaim irq_context: 0 &nft_net->commit_mutex cpu_hotplug_lock jump_label_mutex text_mutex irq_context: 0 kn->active#67 &kernfs_locks->open_file_mutex[count] fs_reclaim mmu_notifier_invalidate_range_start irq_context: 0 &nft_net->commit_mutex cpu_hotplug_lock jump_label_mutex text_mutex ptlock_ptr(page)#2 irq_context: 0 &nft_net->commit_mutex nf_hook_mutex cpu_hotplug_lock jump_label_mutex irq_context: 0 &nft_net->commit_mutex nf_hook_mutex cpu_hotplug_lock jump_label_mutex text_mutex irq_context: 0 &nft_net->commit_mutex nf_hook_mutex cpu_hotplug_lock jump_label_mutex text_mutex ptlock_ptr(page)#2 irq_context: 0 &nft_net->commit_mutex krc.lock irq_context: 0 &nft_net->commit_mutex (work_completion)(&(&flowtable->gc_work)->work) irq_context: 0 &nft_net->commit_mutex &ht->lock irq_context: 0 &nft_net->commit_mutex rcu_read_lock &ht->lock irq_context: 0 kn->active#48 &kernfs_locks->open_file_mutex[count] fs_reclaim &rq->__lock irq_context: 0 kn->active#48 &kernfs_locks->open_file_mutex[count] fs_reclaim &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 kn->active#67 &c->lock irq_context: 0 kn->active#67 &rq->__lock irq_context: 0 kn->active#67 &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 kn->active#67 &n->list_lock irq_context: 0 kn->active#67 &n->list_lock &c->lock irq_context: 0 &nft_net->commit_mutex nf_hook_mutex rcu_read_lock rcu_node_0 irq_context: 0 &nft_net->commit_mutex nf_hook_mutex rcu_read_lock &rq->__lock irq_context: 0 &nft_net->commit_mutex nf_hook_mutex rcu_read_lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 &nft_net->commit_mutex &p->pi_lock &cfs_rq->removed.lock irq_context: 0 (wq_completion)bond6 &rq->__lock irq_context: 0 sk_lock-AF_RDS &mm->mmap_lock irq_context: 0 sk_lock-AF_RDS rds_trans_sem irq_context: 0 (wq_completion)phy12 (work_completion)(&(&local->scan_work)->work) &local->mtx rcu_read_lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)phy12 (work_completion)(&(&local->scan_work)->work) &local->mtx rcu_node_0 irq_context: 0 (wq_completion)phy12 (work_completion)(&(&local->scan_work)->work) &local->mtx &n->list_lock irq_context: 0 (wq_completion)phy12 (work_completion)(&(&local->scan_work)->work) &local->mtx &n->list_lock &c->lock irq_context: 0 &sb->s_type->i_mutex_key#8 &sbi->s_writepages_rwsem jbd2_handle &ei->i_data_sem batched_entropy_u8.lock irq_context: 0 &sb->s_type->i_mutex_key#8 &sbi->s_writepages_rwsem jbd2_handle &ei->i_data_sem kfence_freelist_lock irq_context: 0 &sb->s_type->i_mutex_key#8 &sbi->s_writepages_rwsem jbd2_handle &ei->i_data_sem &meta->lock irq_context: 0 (wq_completion)phy12 (work_completion)(&(&local->scan_work)->work) &local->mtx &rnp->exp_lock irq_context: 0 (wq_completion)phy12 (work_completion)(&(&local->scan_work)->work) &local->mtx rcu_state.exp_mutex irq_context: 0 (wq_completion)phy12 (work_completion)(&(&local->scan_work)->work) &local->mtx rcu_state.exp_mutex rcu_state.exp_mutex.wait_lock irq_context: 0 kn->active#68 fs_reclaim irq_context: 0 kn->active#68 fs_reclaim mmu_notifier_invalidate_range_start irq_context: 0 kn->active#68 stock_lock irq_context: 0 kn->active#68 &kernfs_locks->open_file_mutex[count] irq_context: 0 kn->active#68 &kernfs_locks->open_file_mutex[count] fs_reclaim irq_context: 0 kn->active#68 &kernfs_locks->open_file_mutex[count] fs_reclaim mmu_notifier_invalidate_range_start irq_context: 0 &f->f_pos_lock sb_writers#11 &of->mutex kn->active#68 devcgroup_mutex irq_context: 0 &f->f_pos_lock sb_writers#11 &of->mutex kn->active#68 devcgroup_mutex &rq->__lock irq_context: 0 &f->f_pos_lock sb_writers#11 &of->mutex kn->active#68 devcgroup_mutex &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 kn->active#48 remove_cache_srcu irq_context: 0 kn->active#48 remove_cache_srcu quarantine_lock irq_context: 0 (wq_completion)events (work_completion)(&data->fib_event_work) &data->fib_lock remove_cache_srcu &pcp->lock &zone->lock irq_context: 0 (wq_completion)events (work_completion)(&data->fib_event_work) &data->fib_lock remove_cache_srcu &pcp->lock &zone->lock &____s->seqcount irq_context: 0 kn->active#68 &kernfs_locks->open_file_mutex[count] &c->lock irq_context: 0 kn->active#68 &c->lock irq_context: 0 &tfile->napi_mutex rcu_read_lock rcu_read_lock rcu_read_lock &____s->seqcount#7 irq_context: 0 &tfile->napi_mutex rcu_read_lock rcu_read_lock rcu_read_lock pool_lock#2 irq_context: 0 &tfile->napi_mutex rcu_read_lock rcu_read_lock rcu_read_lock &obj_hash[i].lock irq_context: 0 &sb->s_type->i_mutex_key#10 sk_lock-AF_INET slock-AF_INET rcu_read_lock &obj_hash[i].lock irq_context: 0 sb_writers#4 sb_internal remove_cache_srcu rcu_read_lock &rcu_state.expedited_wq irq_context: 0 sb_writers#4 sb_internal remove_cache_srcu rcu_read_lock &rcu_state.expedited_wq &p->pi_lock irq_context: 0 sb_writers#4 sb_internal remove_cache_srcu rcu_read_lock &rcu_state.expedited_wq &p->pi_lock &rq->__lock irq_context: 0 sb_writers#4 sb_internal remove_cache_srcu rcu_read_lock &rcu_state.expedited_wq &p->pi_lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 &f->f_pos_lock sb_writers#4 &sb->s_type->i_mutex_key#8 jbd2_handle &sbi->s_orphan_lock &lock->wait_lock irq_context: 0 sk_lock-AF_BLUETOOTH-BTPROTO_HCI rlock-AF_BLUETOOTH irq_context: 0 sk_lock-AF_BLUETOOTH-BTPROTO_HCI hci_dev_list_lock pool_lock#2 irq_context: 0 sk_lock-AF_BLUETOOTH-BTPROTO_HCI hci_dev_list_lock tk_core.seq.seqcount irq_context: 0 sk_lock-AF_BLUETOOTH-BTPROTO_HCI hci_dev_list_lock rlock-AF_BLUETOOTH irq_context: 0 sk_lock-AF_BLUETOOTH-BTPROTO_HCI hci_dev_list_lock &____s->seqcount irq_context: 0 sk_lock-AF_BLUETOOTH-BTPROTO_HCI hci_dev_list_lock rcu_read_lock pool_lock#2 irq_context: 0 sk_lock-AF_BLUETOOTH-BTPROTO_HCI hci_dev_list_lock &obj_hash[i].lock irq_context: 0 sk_lock-AF_BLUETOOTH-BTPROTO_HCI hci_dev_list_lock &c->lock irq_context: 0 sk_lock-AF_BLUETOOTH-BTPROTO_HCI hci_dev_list_lock &n->list_lock irq_context: 0 sk_lock-AF_BLUETOOTH-BTPROTO_HCI hci_dev_list_lock &n->list_lock &c->lock irq_context: 0 rtnl_mutex sk_lock-AF_INET6 &idev->mc_lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 &sb->s_type->i_mutex_key#10 rtnl_mutex sk_lock-AF_INET6 irq_context: 0 &sb->s_type->i_mutex_key#10 rtnl_mutex sk_lock-AF_INET6 slock-AF_INET6 irq_context: 0 &sb->s_type->i_mutex_key#10 rtnl_mutex sk_lock-AF_INET6 &idev->mc_lock irq_context: 0 &sb->s_type->i_mutex_key#10 rtnl_mutex sk_lock-AF_INET6 &idev->mc_lock &obj_hash[i].lock irq_context: 0 &sb->s_type->i_mutex_key#10 rtnl_mutex sk_lock-AF_INET6 &idev->mc_lock pool_lock#2 irq_context: 0 &sb->s_type->i_mutex_key#10 rtnl_mutex sk_lock-AF_INET6 &idev->mc_lock krc.lock irq_context: 0 &sb->s_type->i_mutex_key#10 rtnl_mutex sk_lock-AF_INET6 &obj_hash[i].lock irq_context: 0 &sb->s_type->i_mutex_key#10 rtnl_mutex sk_lock-AF_INET6 &rq->__lock irq_context: 0 &sb->s_type->i_mutex_key#10 rtnl_mutex sk_lock-AF_INET6 &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 &sb->s_type->i_mutex_key#10 rtnl_mutex sk_lock-AF_INET6 pool_lock#2 irq_context: 0 &sb->s_type->i_mutex_key#10 rtnl_mutex sk_lock-AF_INET6 krc.lock irq_context: 0 &sb->s_type->i_mutex_key#10 rtnl_mutex slock-AF_INET6 irq_context: 0 &sb->s_type->i_mutex_key#8 &sbi->s_writepages_rwsem &journal->j_wait_transaction_locked irq_context: 0 &journal->j_barrier &journal->j_checkpoint_mutex batched_entropy_u8.lock irq_context: 0 &mm->mmap_lock &po->pg_vec_lock &rq->__lock &cfs_rq->removed.lock irq_context: 0 sb_writers#4 &type->i_mutex_dir_key#3/1 &type->i_mutex_dir_key#3 rcu_read_lock &rcu_state.gp_wq &p->pi_lock &cfs_rq->removed.lock irq_context: 0 tracepoints_mutex &obj_hash[i].lock pool_lock irq_context: 0 &sb->s_type->i_mutex_key#10 sk_lock-AF_PACKET rcu_state.exp_mutex &rnp->exp_wq[0] irq_context: 0 &journal->j_barrier &journal->j_state_lock &journal->j_wait_commit &p->pi_lock &cfs_rq->removed.lock irq_context: 0 &journal->j_checkpoint_mutex batched_entropy_u8.lock irq_context: 0 &journal->j_checkpoint_mutex kfence_freelist_lock irq_context: 0 &journal->j_barrier &lock->wait_lock irq_context: 0 &f->f_pos_lock sb_writers#4 &sb->s_type->i_mutex_key#8 jbd2_handle &sbi->s_orphan_lock mmu_notifier_invalidate_range_start irq_context: 0 &f->f_pos_lock sb_writers#4 &sb->s_type->i_mutex_key#8 jbd2_handle &sbi->s_orphan_lock pool_lock#2 irq_context: 0 rtnl_mutex noop_qdisc.q.lock batched_entropy_u64.lock crngs.lock irq_context: 0 rtnl_mutex team->team_lock_key#3 net_rwsem &rq->__lock irq_context: 0 rtnl_mutex team->team_lock_key#3 net_rwsem &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 rtnl_mutex team->team_lock_key#3 &batadv_netdev_addr_lock_key irq_context: 0 rtnl_mutex team->team_lock_key#3 rcu_read_lock &pool->lock/1 &p->pi_lock &cfs_rq->removed.lock irq_context: 0 rtnl_mutex team->team_lock_key#3 &bat_priv->softif_vlan_list_lock irq_context: 0 rtnl_mutex team->team_lock_key#3 &bat_priv->softif_vlan_list_lock &c->lock irq_context: 0 rtnl_mutex team->team_lock_key#3 &bat_priv->softif_vlan_list_lock pool_lock#2 irq_context: 0 rtnl_mutex team->team_lock_key#3 key#16 irq_context: 0 rtnl_mutex team->team_lock_key#3 &bat_priv->tt.changes_list_lock irq_context: 0 rtnl_mutex team->team_lock_key#3 lock kernfs_idr_lock pool_lock#2 irq_context: 0 rtnl_mutex team->team_lock_key#3 rcu_read_lock &rcu_state.expedited_wq irq_context: 0 rtnl_mutex team->team_lock_key#3 rcu_read_lock &rcu_state.expedited_wq &p->pi_lock irq_context: 0 rtnl_mutex team->team_lock_key#3 rcu_read_lock &rcu_state.expedited_wq &p->pi_lock &rq->__lock irq_context: 0 rtnl_mutex team->team_lock_key#3 rcu_read_lock &rcu_state.expedited_wq &p->pi_lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 rtnl_mutex team->team_lock_key#3 lweventlist_lock pool_lock#2 irq_context: 0 (wq_completion)ipv6_addrconf (work_completion)(&(&ifa->dad_work)->work) rtnl_mutex &idev->mc_lock &batadv_netdev_addr_lock_key irq_context: 0 (wq_completion)wg-kex-wg1#2 (work_completion)(&peer->transmit_handshake_work) &peer->endpoint_lock rcu_read_lock_bh rcu_read_lock rcu_read_lock_bh rcu_read_lock tk_core.seq.seqcount irq_context: 0 (wq_completion)wg-kex-wg1#2 (work_completion)(&peer->transmit_handshake_work) &peer->endpoint_lock rcu_read_lock_bh rcu_read_lock rcu_read_lock_bh rcu_read_lock &obj_hash[i].lock irq_context: 0 (wq_completion)wg-kex-wg0#2 (work_completion)(&({ do { const void *__vpp_verify = (typeof((worker) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __asm__ ("" : "=r"(__ptr) : "0"((typeof(*((worker))) *)((worker)))); (typeof((typeof(*((worker))) *)((worker)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); })->work) &peer->endpoint_lock rcu_read_lock_bh rcu_read_lock rcu_read_lock_bh rcu_read_lock pool_lock#2 irq_context: 0 (wq_completion)wg-kex-wg0#2 (work_completion)(&({ do { const void *__vpp_verify = (typeof((worker) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __asm__ ("" : "=r"(__ptr) : "0"((typeof(*((worker))) *)((worker)))); (typeof((typeof(*((worker))) *)((worker)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); })->work) &peer->endpoint_lock rcu_read_lock_bh rcu_read_lock rcu_read_lock_bh rcu_read_lock tk_core.seq.seqcount irq_context: 0 (wq_completion)wg-kex-wg0#2 (work_completion)(&({ do { const void *__vpp_verify = (typeof((worker) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __asm__ ("" : "=r"(__ptr) : "0"((typeof(*((worker))) *)((worker)))); (typeof((typeof(*((worker))) *)((worker)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); })->work) &peer->endpoint_lock rcu_read_lock_bh rcu_read_lock rcu_read_lock_bh rcu_read_lock &obj_hash[i].lock irq_context: 0 rtnl_mutex team->team_lock_key#3 &tb->tb6_lock &n->list_lock irq_context: 0 rtnl_mutex team->team_lock_key#3 &tb->tb6_lock &n->list_lock &c->lock irq_context: 0 rtnl_mutex team->team_lock_key#3 &cfs_rq->removed.lock irq_context: 0 (wq_completion)ipv6_addrconf (work_completion)(&(&ifa->dad_work)->work) rtnl_mutex &idev->mc_lock &batadv_netdev_addr_lock_key &c->lock irq_context: 0 (wq_completion)kblockd &rcu_state.expedited_wq irq_context: 0 (wq_completion)kblockd &rcu_state.expedited_wq &p->pi_lock irq_context: 0 (wq_completion)kblockd &rcu_state.expedited_wq &p->pi_lock &rq->__lock irq_context: 0 (wq_completion)kblockd &rcu_state.expedited_wq &p->pi_lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 rtnl_mutex team->team_lock_key#3 &root->kernfs_rwsem kernfs_idr_lock irq_context: 0 rtnl_mutex team->team_lock_key#3 &root->kernfs_rwsem &obj_hash[i].lock irq_context: 0 rtnl_mutex team->team_lock_key#3 &bat_priv->tt.changes_list_lock &obj_hash[i].lock irq_context: 0 rtnl_mutex team->team_lock_key#3 &bat_priv->tt.changes_list_lock pool_lock#2 irq_context: 0 &sb->s_type->i_mutex_key#8 &sbi->s_writepages_rwsem jbd2_handle &rcu_state.expedited_wq &p->pi_lock irq_context: 0 &journal->j_list_lock bit_wait_table + i irq_context: 0 &bat_priv->bat_v.ogm_buff_mutex &rq->__lock irq_context: 0 &bat_priv->bat_v.ogm_buff_mutex &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 sb_writers#4 &sb->s_type->i_mutex_key#8 &sb->s_type->i_lock_key#22 &xa->xa_lock#9 &obj_hash[i].lock pool_lock irq_context: 0 sb_writers#4 &type->i_mutex_dir_key#3/1 jbd2_handle &ei->i_data_sem mmu_notifier_invalidate_range_start &cfs_rq->removed.lock irq_context: 0 sb_writers#4 &type->i_mutex_dir_key#3/1 jbd2_handle &ei->i_data_sem mmu_notifier_invalidate_range_start &obj_hash[i].lock irq_context: 0 sb_writers#4 &sb->s_type->i_mutex_key#8 jbd2_handle &rq->__lock irq_context: 0 sb_writers#4 &sb->s_type->i_mutex_key#8 jbd2_handle &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 sb_writers#4 &type->i_mutex_dir_key#3/1 sb_internal jbd2_handle &sbi->s_orphan_lock bit_wait_table + i irq_context: 0 &f->f_pos_lock sb_writers#4 &sb->s_type->i_mutex_key#8 &ei->i_data_sem &____s->seqcount#2 irq_context: 0 &f->f_pos_lock sb_writers#4 &sb->s_type->i_mutex_key#8 &ei->i_data_sem &____s->seqcount irq_context: 0 br_ioctl_mutex rtnl_mutex &cfs_rq->removed.lock irq_context: 0 cb_lock genl_mutex uevent_sock_mutex rcu_read_lock &rcu_state.gp_wq irq_context: 0 cb_lock genl_mutex uevent_sock_mutex rcu_read_lock &rcu_state.gp_wq &p->pi_lock irq_context: 0 cb_lock genl_mutex uevent_sock_mutex rcu_read_lock &rcu_state.gp_wq &p->pi_lock &rq->__lock irq_context: 0 cb_lock genl_mutex uevent_sock_mutex rcu_read_lock &rcu_state.gp_wq &p->pi_lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 cb_lock genl_mutex cpu_hotplug_lock wq_pool_mutex &n->list_lock irq_context: 0 cb_lock genl_mutex cpu_hotplug_lock wq_pool_mutex &n->list_lock &c->lock irq_context: 0 cb_lock genl_mutex cpu_hotplug_lock wq_pool_mutex &rq->__lock irq_context: 0 cb_lock genl_mutex cpu_hotplug_lock wq_pool_mutex &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 cb_lock genl_mutex rtnl_mutex &rdev->wiphy.mtx uevent_sock_mutex rlock-AF_NETLINK irq_context: 0 cb_lock genl_mutex rtnl_mutex &rdev->wiphy.mtx uevent_sock_mutex rcu_read_lock &ei->socket.wq.wait irq_context: 0 cb_lock genl_mutex rtnl_mutex &rdev->wiphy.mtx uevent_sock_mutex rcu_read_lock &ei->socket.wq.wait &ep->lock irq_context: 0 cb_lock genl_mutex rfkill_global_mutex uevent_sock_mutex rcu_read_lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 cb_lock genl_mutex &sb->s_type->i_mutex_key#3 remove_cache_srcu irq_context: 0 cb_lock genl_mutex &sb->s_type->i_mutex_key#3 remove_cache_srcu quarantine_lock irq_context: 0 cb_lock genl_mutex &sb->s_type->i_mutex_key#3 remove_cache_srcu &c->lock irq_context: 0 cb_lock genl_mutex &sb->s_type->i_mutex_key#3 remove_cache_srcu &n->list_lock irq_context: 0 cb_lock genl_mutex &sb->s_type->i_mutex_key#3 remove_cache_srcu rcu_read_lock pool_lock#2 irq_context: 0 cb_lock genl_mutex &sb->s_type->i_mutex_key#3 remove_cache_srcu &obj_hash[i].lock irq_context: 0 cb_lock genl_mutex &sb->s_type->i_mutex_key#3 remove_cache_srcu &rq->__lock irq_context: 0 cb_lock genl_mutex &sb->s_type->i_mutex_key#3 remove_cache_srcu &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 cb_lock genl_mutex rtnl_mutex &rdev->wiphy.mtx uevent_sock_mutex rcu_read_lock &ei->socket.wq.wait &ep->lock &ep->wq irq_context: 0 cb_lock genl_mutex rtnl_mutex &rdev->wiphy.mtx uevent_sock_mutex rcu_read_lock &ei->socket.wq.wait &ep->lock &ep->wq &p->pi_lock irq_context: 0 cb_lock genl_mutex rtnl_mutex &rdev->wiphy.mtx uevent_sock_mutex rcu_read_lock &ei->socket.wq.wait &ep->lock &ep->wq &p->pi_lock &rq->__lock irq_context: 0 cb_lock genl_mutex rtnl_mutex &rdev->wiphy.mtx uevent_sock_mutex rcu_read_lock &ei->socket.wq.wait &ep->lock &ep->wq &p->pi_lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 cb_lock genl_mutex rtnl_mutex &rdev->wiphy.mtx uevent_sock_mutex rcu_read_lock rcu_node_0 irq_context: 0 cb_lock genl_mutex rtnl_mutex &rdev->wiphy.mtx uevent_sock_mutex rcu_read_lock &rq->__lock irq_context: 0 cb_lock genl_mutex rtnl_mutex &rdev->wiphy.mtx uevent_sock_mutex rcu_read_lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 cb_lock genl_mutex rtnl_mutex &rdev->wiphy.mtx uevent_sock_mutex rcu_read_lock &rcu_state.gp_wq irq_context: 0 cb_lock genl_mutex rtnl_mutex &rdev->wiphy.mtx uevent_sock_mutex rcu_read_lock &rcu_state.gp_wq &p->pi_lock irq_context: 0 cb_lock genl_mutex rtnl_mutex &rdev->wiphy.mtx uevent_sock_mutex rcu_read_lock &rcu_state.gp_wq &p->pi_lock &rq->__lock irq_context: 0 cb_lock genl_mutex rtnl_mutex &rdev->wiphy.mtx uevent_sock_mutex rcu_read_lock &rcu_state.gp_wq &p->pi_lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 kn->active#20 &kernfs_locks->open_file_mutex[count] &c->lock irq_context: 0 cb_lock genl_mutex rtnl_mutex &rdev->wiphy.mtx &idev->mc_lock &obj_hash[i].lock pool_lock irq_context: 0 cb_lock genl_mutex rtnl_mutex &rdev->wiphy.mtx &idev->mc_lock &rq->__lock irq_context: 0 cb_lock genl_mutex rtnl_mutex &rdev->wiphy.mtx &idev->mc_lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 cb_lock genl_mutex rtnl_mutex &rdev->wiphy.mtx &idev->mc_lock _xmit_ETHER &n->list_lock irq_context: 0 cb_lock genl_mutex rtnl_mutex &rdev->wiphy.mtx &idev->mc_lock _xmit_ETHER &n->list_lock &c->lock irq_context: 0 cb_lock genl_mutex rtnl_mutex &rdev->wiphy.mtx rlock-AF_NETLINK irq_context: 0 cb_lock genl_mutex &sb->s_type->i_mutex_key#3 rcu_read_lock &p->pi_lock irq_context: 0 cb_lock genl_mutex &sb->s_type->i_mutex_key#3 rcu_read_lock &p->pi_lock &rq->__lock irq_context: 0 cb_lock genl_mutex &sb->s_type->i_mutex_key#3 rcu_read_lock &p->pi_lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 cb_lock genl_mutex &sb->s_type->i_mutex_key#3 rcu_node_0 irq_context: 0 sb_writers#4 &sb->s_type->i_mutex_key#8 rcu_read_lock &cfs_rq->removed.lock irq_context: 0 &sig->cred_guard_mutex &ei->xattr_sem rcu_read_lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 &mm->mmap_lock &mapping->i_mmap_rwsem &anon_vma->rwsem &rcu_state.expedited_wq &p->pi_lock irq_context: 0 br_ioctl_mutex rtnl_mutex fill_pool_map-wait-type-override pool_lock#2 irq_context: 0 br_ioctl_mutex rtnl_mutex fill_pool_map-wait-type-override &c->lock irq_context: 0 br_ioctl_mutex rtnl_mutex fill_pool_map-wait-type-override &n->list_lock irq_context: 0 br_ioctl_mutex rtnl_mutex fill_pool_map-wait-type-override &n->list_lock &c->lock irq_context: 0 br_ioctl_mutex rtnl_mutex fill_pool_map-wait-type-override pool_lock irq_context: 0 cb_lock genl_mutex rtnl_mutex rcu_read_lock &rq->__lock &cfs_rq->removed.lock irq_context: 0 cb_lock genl_mutex rtnl_mutex &rdev->wiphy.mtx &sb->s_type->i_mutex_key#3 rcu_node_0 irq_context: 0 cb_lock genl_mutex rtnl_mutex &rdev->wiphy.mtx &root->kernfs_rwsem &root->kernfs_iattr_rwsem &rq->__lock irq_context: 0 cb_lock genl_mutex rtnl_mutex &rdev->wiphy.mtx &root->kernfs_rwsem &root->kernfs_iattr_rwsem &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 kn->active#16 &n->list_lock irq_context: 0 kn->active#16 &n->list_lock &c->lock irq_context: 0 cb_lock genl_mutex rtnl_mutex &rdev->wiphy.mtx fill_pool_map-wait-type-override &n->list_lock irq_context: 0 cb_lock genl_mutex rtnl_mutex &rdev->wiphy.mtx fill_pool_map-wait-type-override &n->list_lock &c->lock irq_context: 0 cb_lock genl_mutex rtnl_mutex &rdev->wiphy.mtx fill_pool_map-wait-type-override &rq->__lock irq_context: 0 cb_lock genl_mutex rtnl_mutex &rdev->wiphy.mtx fill_pool_map-wait-type-override &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 cb_lock genl_mutex rtnl_mutex &rdev->wiphy.mtx remove_cache_srcu &rq->__lock irq_context: 0 (wq_completion)wg-kex-wg0#2 (work_completion)(&({ do { const void *__vpp_verify = (typeof((worker) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __asm__ ("" : "=r"(__ptr) : "0"((typeof(*((worker))) *)((worker)))); (typeof((typeof(*((worker))) *)((worker)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); })->work) &handshake->lock rcu_read_lock_bh &peer->keypairs.keypair_update_lock fill_pool_map-wait-type-override &c->lock irq_context: 0 (wq_completion)wg-kex-wg0#2 (work_completion)(&({ do { const void *__vpp_verify = (typeof((worker) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __asm__ ("" : "=r"(__ptr) : "0"((typeof(*((worker))) *)((worker)))); (typeof((typeof(*((worker))) *)((worker)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); })->work) &handshake->lock rcu_read_lock_bh &peer->keypairs.keypair_update_lock fill_pool_map-wait-type-override pool_lock irq_context: 0 (wq_completion)wg-crypt-wg2 (work_completion)(&({ do { const void *__vpp_verify = (typeof((worker) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __asm__ ("" : "=r"(__ptr) : "0"((typeof(*((worker))) *)((worker)))); (typeof((typeof(*((worker))) *)((worker)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); })->work) rcu_read_lock &pool->lock &p->pi_lock &rq->__lock &cfs_rq->removed.lock irq_context: 0 &iint->mutex mapping.invalidate_lock rcu_read_lock &____s->seqcount#2 irq_context: 0 (wq_completion)phy19 irq_context: 0 (wq_completion)phy19 (work_completion)(&local->reconfig_filter) irq_context: 0 (wq_completion)phy19 (work_completion)(&local->reconfig_filter) &local->filter_lock irq_context: softirq (&icsk->icsk_delack_timer) slock-AF_INET rcu_read_lock rcu_read_lock pool_lock#2 irq_context: softirq (&icsk->icsk_delack_timer) slock-AF_INET rcu_read_lock rcu_read_lock &dir->lock#2 irq_context: softirq (&icsk->icsk_delack_timer) slock-AF_INET rcu_read_lock rcu_read_lock &ul->lock irq_context: softirq (&icsk->icsk_delack_timer) slock-AF_INET rcu_read_lock rcu_read_lock &obj_hash[i].lock irq_context: 0 &iint->mutex mapping.invalidate_lock &n->list_lock irq_context: 0 &iint->mutex mapping.invalidate_lock &n->list_lock &c->lock irq_context: 0 cb_lock genl_mutex rfkill_global_mutex &rq->__lock &cfs_rq->removed.lock irq_context: 0 cb_lock genl_mutex rtnl_mutex &rdev->wiphy.mtx pcpu_alloc_mutex &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 kn->active#18 rcu_read_lock rcu_node_0 irq_context: 0 kn->active#18 rcu_read_lock &rq->__lock irq_context: 0 kn->active#18 rcu_read_lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 &type->i_mutex_dir_key#4 remove_cache_srcu &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 sb_writers#4 &sb->s_type->i_mutex_key#8 mapping.invalidate_lock &sb->s_type->i_lock_key#22 &xa->xa_lock#9 rcu_read_lock per_cpu_ptr(&cgroup_rstat_cpu_lock, cpu) irq_context: 0 &sig->cred_guard_mutex &sig->exec_update_lock &mm->mmap_lock pool_lock#2 irq_context: 0 rtnl_mutex mrt_lock#2 &c->lock irq_context: softirq rcu_read_lock rcu_read_lock &br->multicast_lock fill_pool_map-wait-type-override &c->lock irq_context: softirq rcu_read_lock rcu_read_lock &br->multicast_lock fill_pool_map-wait-type-override pool_lock#2 irq_context: softirq rcu_read_lock rcu_read_lock &br->multicast_lock fill_pool_map-wait-type-override pool_lock irq_context: 0 &sb->s_type->i_mutex_key#8 rcu_read_lock rcu_read_lock rcu_node_0 irq_context: 0 &sb->s_type->i_mutex_key#8 rcu_read_lock rcu_read_lock &rq->__lock irq_context: 0 &sb->s_type->i_mutex_key#8 rcu_read_lock rcu_read_lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)phy20 irq_context: 0 (wq_completion)phy20 (work_completion)(&local->reconfig_filter) irq_context: 0 (wq_completion)phy20 (work_completion)(&local->reconfig_filter) &local->filter_lock irq_context: 0 (wq_completion)phy20 (work_completion)(&local->reconfig_filter) &rq->__lock irq_context: 0 (wq_completion)phy20 (work_completion)(&local->reconfig_filter) &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 kn->active#15 &____s->seqcount#2 irq_context: 0 (wq_completion)bat_events (work_completion)(&(&forw_packet_aggr->delayed_work)->work) &hard_iface->bat_iv.ogm_buff_mutex rcu_read_lock &bat_priv->forw_bat_list_lock fill_pool_map-wait-type-override &c->lock irq_context: 0 (wq_completion)bat_events (work_completion)(&(&forw_packet_aggr->delayed_work)->work) &hard_iface->bat_iv.ogm_buff_mutex rcu_read_lock &bat_priv->forw_bat_list_lock fill_pool_map-wait-type-override &n->list_lock irq_context: 0 (wq_completion)bat_events (work_completion)(&(&forw_packet_aggr->delayed_work)->work) &hard_iface->bat_iv.ogm_buff_mutex rcu_read_lock &bat_priv->forw_bat_list_lock fill_pool_map-wait-type-override &n->list_lock &c->lock irq_context: 0 kn->active#60 &kernfs_locks->open_file_mutex[count] &____s->seqcount#2 irq_context: 0 kn->active#60 &kernfs_locks->open_file_mutex[count] &____s->seqcount irq_context: 0 (wq_completion)bat_events (work_completion)(&(&bat_priv->mcast.work)->work) rcu_read_lock &rcu_state.gp_wq &p->pi_lock &rq->__lock &cfs_rq->removed.lock irq_context: 0 (wq_completion)events (work_completion)(&work->work)#3 rcu_read_lock rcu_read_lock pool_lock#2 irq_context: 0 (wq_completion)events (work_completion)(&work->work)#3 rcu_read_lock rcu_read_lock rcu_node_0 irq_context: 0 (wq_completion)events (work_completion)(&work->work)#3 rcu_read_lock rcu_read_lock &rq->__lock irq_context: 0 (wq_completion)events (work_completion)(&work->work)#3 rcu_read_lock rcu_read_lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)events (work_completion)(&work->work)#3 rcu_read_lock rcu_read_lock &obj_hash[i].lock irq_context: 0 (wq_completion)events (work_completion)(&work->work)#3 rcu_read_lock rcu_read_lock &c->lock irq_context: 0 (wq_completion)events (work_completion)(&work->work)#3 rcu_read_lock rcu_read_lock &dir->lock#2 irq_context: 0 (wq_completion)events (work_completion)(&work->work)#3 rcu_read_lock rcu_read_lock &tbl->lock irq_context: 0 (wq_completion)events (work_completion)(&work->work)#3 rcu_read_lock rcu_read_lock &n->lock irq_context: 0 (wq_completion)events (work_completion)(&work->work)#3 rcu_read_lock rcu_read_lock &____s->seqcount#9 irq_context: 0 (wq_completion)events (work_completion)(&work->work)#3 rcu_read_lock rcu_read_lock rcu_read_lock_bh rcu_read_lock &obj_hash[i].lock irq_context: 0 (wq_completion)events (work_completion)(&work->work)#3 rcu_read_lock rcu_read_lock rcu_read_lock_bh rcu_read_lock pool_lock#2 irq_context: 0 (wq_completion)events (work_completion)(&work->work)#3 rcu_read_lock rcu_node_0 irq_context: 0 (wq_completion)events (work_completion)(&work->work)#3 rcu_read_lock &rcu_state.gp_wq irq_context: 0 (wq_completion)events (work_completion)(&work->work)#3 rcu_read_lock &rcu_state.gp_wq &p->pi_lock irq_context: 0 (wq_completion)events (work_completion)(&work->work)#3 rcu_read_lock &rcu_state.gp_wq &p->pi_lock &rq->__lock irq_context: 0 (wq_completion)events (work_completion)(&work->work)#3 rcu_read_lock &rcu_state.gp_wq &p->pi_lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)events (work_completion)(&work->work)#3 rcu_read_lock &rq->__lock irq_context: 0 sk_lock-AF_INET6 &net->xfrm.xfrm_policy_lock irq_context: 0 pcpu_alloc_mutex pgd_lock irq_context: 0 pcpu_alloc_mutex rcu_read_lock pool_lock#2 irq_context: 0 pcpu_alloc_mutex key irq_context: 0 pcpu_alloc_mutex percpu_counters_lock irq_context: 0 sb_writers#4 &sb->s_type->i_mutex_key#8 mapping.invalidate_lock &mapping->i_mmap_rwsem &obj_hash[i].lock pool_lock irq_context: 0 rtnl_mutex _xmit_ETHER rcu_read_lock &pool->lock/1 &p->pi_lock &rq->__lock &cfs_rq->removed.lock irq_context: 0 (wq_completion)phy21 irq_context: 0 (wq_completion)phy21 (work_completion)(&local->reconfig_filter) irq_context: 0 (wq_completion)phy21 (work_completion)(&local->reconfig_filter) &local->filter_lock irq_context: 0 (wq_completion)writeback (work_completion)(&(&wb->bw_dwork)->work) &rq->__lock irq_context: 0 &mm->mmap_lock sb_writers#4 pgd_lock irq_context: 0 &mm->mmap_lock sb_writers#4 rcu_read_lock pool_lock#2 irq_context: 0 &mm->mmap_lock sb_writers#4 key irq_context: 0 &mm->mmap_lock sb_writers#4 pcpu_lock irq_context: 0 &mm->mmap_lock sb_writers#4 percpu_counters_lock irq_context: 0 sb_writers#4 &type->i_mutex_dir_key#3/1 sb_internal jbd2_handle &ei->i_data_sem rcu_read_lock &rq->__lock &cfs_rq->removed.lock irq_context: 0 cb_lock rtnl_mutex &rdev->wiphy.mtx &wdev->mtx rcu_state.exp_mutex &rnp->exp_wq[1] irq_context: 0 cb_lock rtnl_mutex &rdev->wiphy.mtx &local->queue_stop_reason_lock irq_context: 0 cb_lock rtnl_mutex &rdev->wiphy.mtx rcu_state.exp_mutex rcu_node_0 irq_context: 0 cb_lock rtnl_mutex &rdev->wiphy.mtx rcu_state.exp_mutex &obj_hash[i].lock irq_context: 0 cb_lock rtnl_mutex &rdev->wiphy.mtx rcu_state.exp_mutex rcu_read_lock &pool->lock irq_context: 0 cb_lock rtnl_mutex &rdev->wiphy.mtx rcu_state.exp_mutex rcu_read_lock &pool->lock &obj_hash[i].lock irq_context: 0 cb_lock rtnl_mutex &rdev->wiphy.mtx rcu_state.exp_mutex rcu_read_lock &pool->lock &p->pi_lock irq_context: 0 cb_lock rtnl_mutex &rdev->wiphy.mtx rcu_state.exp_mutex rcu_read_lock &pool->lock &p->pi_lock &rq->__lock irq_context: 0 cb_lock rtnl_mutex &rdev->wiphy.mtx rcu_state.exp_mutex rcu_read_lock &pool->lock &p->pi_lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 cb_lock rtnl_mutex &rdev->wiphy.mtx rcu_state.exp_mutex &rq->__lock irq_context: 0 cb_lock rtnl_mutex &rdev->wiphy.mtx rcu_state.exp_mutex &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 cb_lock rtnl_mutex &rdev->wiphy.mtx &local->mtx irq_context: 0 cb_lock rtnl_mutex &rdev->wiphy.mtx &local->sta_mtx irq_context: 0 cb_lock rtnl_mutex &rdev->wiphy.mtx _xmit_ETHER irq_context: 0 cb_lock rtnl_mutex &rdev->wiphy.mtx _xmit_ETHER &local->filter_lock irq_context: 0 cb_lock rtnl_mutex &rdev->wiphy.mtx _xmit_ETHER &local->filter_lock &obj_hash[i].lock irq_context: 0 cb_lock rtnl_mutex &rdev->wiphy.mtx _xmit_ETHER &local->filter_lock pool_lock#2 irq_context: 0 cb_lock rtnl_mutex &rdev->wiphy.mtx _xmit_ETHER &local->filter_lock krc.lock irq_context: 0 cb_lock rtnl_mutex &rdev->wiphy.mtx (&local->dynamic_ps_timer) irq_context: 0 cb_lock rtnl_mutex &rdev->wiphy.mtx (work_completion)(&local->dynamic_ps_enable_work) irq_context: 0 cb_lock rtnl_mutex &rdev->wiphy.mtx (work_completion)(&sdata->recalc_smps) irq_context: 0 cb_lock rtnl_mutex &rdev->wiphy.mtx (work_completion)(&link->csa_finalize_work) irq_context: 0 cb_lock rtnl_mutex &rdev->wiphy.mtx (work_completion)(&link->color_change_finalize_work) irq_context: 0 cb_lock rtnl_mutex &rdev->wiphy.mtx (work_completion)(&(&link->dfs_cac_timer_work)->work) irq_context: 0 cb_lock rtnl_mutex &rdev->wiphy.mtx &local->key_mtx rcu_state.exp_mutex rcu_node_0 irq_context: 0 cb_lock rtnl_mutex &rdev->wiphy.mtx &local->key_mtx rcu_state.exp_mutex &rq->__lock irq_context: 0 cb_lock rtnl_mutex &rdev->wiphy.mtx &local->key_mtx rcu_state.exp_mutex &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 cb_lock rtnl_mutex &rdev->wiphy.mtx &local->key_mtx rcu_state.exp_mutex &obj_hash[i].lock irq_context: 0 cb_lock rtnl_mutex &rdev->wiphy.mtx &local->key_mtx rcu_state.exp_mutex rcu_read_lock &pool->lock irq_context: 0 cb_lock rtnl_mutex &rdev->wiphy.mtx &local->key_mtx rcu_state.exp_mutex rcu_read_lock &pool->lock &obj_hash[i].lock irq_context: 0 cb_lock rtnl_mutex &rdev->wiphy.mtx &local->key_mtx rcu_state.exp_mutex rcu_read_lock &pool->lock &p->pi_lock irq_context: 0 cb_lock rtnl_mutex &rdev->wiphy.mtx &local->key_mtx rcu_state.exp_mutex rcu_read_lock &pool->lock &p->pi_lock &rq->__lock irq_context: 0 cb_lock rtnl_mutex &rdev->wiphy.mtx &local->key_mtx rcu_state.exp_mutex rcu_read_lock &pool->lock &p->pi_lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 cb_lock rtnl_mutex &rdev->wiphy.mtx &local->key_mtx rcu_state.exp_mutex &rnp->exp_wq[3] irq_context: 0 cb_lock rtnl_mutex &rdev->wiphy.mtx &local->key_mtx &obj_hash[i].lock irq_context: 0 cb_lock rtnl_mutex &rdev->wiphy.mtx &list->lock#18 irq_context: 0 cb_lock rtnl_mutex &rdev->wiphy.mtx &local->filter_lock irq_context: 0 cb_lock rtnl_mutex &rdev->wiphy.mtx &sb->s_type->i_mutex_key#3 &pcp->lock &zone->lock irq_context: 0 cb_lock rtnl_mutex &rdev->wiphy.mtx rcu_read_lock &fq->lock irq_context: 0 cb_lock rtnl_mutex &rdev->wiphy.mtx rcu_read_lock &local->queue_stop_reason_lock irq_context: 0 kn->active#69 fs_reclaim irq_context: 0 kn->active#69 fs_reclaim mmu_notifier_invalidate_range_start irq_context: 0 kn->active#69 &c->lock irq_context: 0 kn->active#69 stock_lock irq_context: 0 kn->active#69 &kernfs_locks->open_file_mutex[count] irq_context: 0 kn->active#69 &kernfs_locks->open_file_mutex[count] fs_reclaim irq_context: 0 kn->active#69 &kernfs_locks->open_file_mutex[count] fs_reclaim &rq->__lock irq_context: 0 kn->active#69 &kernfs_locks->open_file_mutex[count] fs_reclaim &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 kn->active#69 &kernfs_locks->open_file_mutex[count] fs_reclaim mmu_notifier_invalidate_range_start irq_context: 0 sb_writers#11 &n->list_lock irq_context: 0 sb_writers#11 &n->list_lock &c->lock irq_context: 0 sb_writers#11 &p->lock &of->mutex kn->active#62 &cgrp->pidlist_mutex &c->lock irq_context: 0 &type->i_mutex_dir_key#5 fs_reclaim mmu_notifier_invalidate_range_start &rq->__lock irq_context: 0 &f->f_pos_lock &type->i_mutex_dir_key#5 rcu_node_0 irq_context: 0 kn->active#69 &kernfs_locks->open_file_mutex[count] &c->lock irq_context: 0 &sig->cred_guard_mutex &ei->xattr_sem &cfs_rq->removed.lock irq_context: 0 &sig->cred_guard_mutex &ei->xattr_sem &obj_hash[i].lock irq_context: 0 sb_writers#11 &p->lock &____s->seqcount#2 irq_context: 0 sb_writers#11 &p->lock &____s->seqcount irq_context: 0 sb_writers#11 &of->mutex cgroup_mutex cpu_hotplug_lock cgroup_threadgroup_rwsem rcu_state.exp_mutex &rnp->exp_wq[2] irq_context: 0 sb_writers#11 &of->mutex cgroup_mutex cpu_hotplug_lock cgroup_threadgroup_rwsem &c->lock irq_context: 0 rtnl_mutex bpf_devs_lock fs_reclaim irq_context: 0 rtnl_mutex bpf_devs_lock fs_reclaim mmu_notifier_invalidate_range_start irq_context: 0 rtnl_mutex bpf_devs_lock stock_lock irq_context: 0 rtnl_mutex bpf_devs_lock pool_lock#2 irq_context: 0 rtnl_mutex bpf_devs_lock &obj_hash[i].lock irq_context: 0 sb_writers#11 rcu_read_lock rcu_node_0 irq_context: 0 sb_writers#11 rcu_read_lock &rq->__lock irq_context: 0 sb_writers#11 &of->mutex cgroup_mutex cpu_hotplug_lock cgroup_threadgroup_rwsem rcu_state.exp_mutex &rnp->exp_wq[3] irq_context: 0 sb_writers#11 &type->i_mutex_dir_key#7/1 &root->kernfs_rwsem &xa->xa_lock#4 irq_context: 0 sb_writers#11 &type->i_mutex_dir_key#7/1 &root->kernfs_rwsem &obj_hash[i].lock irq_context: 0 sb_writers#4 &sb->s_type->i_mutex_key#8 jbd2_handle rcu_read_lock &rcu_state.gp_wq irq_context: 0 sb_writers#4 &sb->s_type->i_mutex_key#8 jbd2_handle rcu_read_lock &rcu_state.gp_wq &p->pi_lock irq_context: 0 sb_writers#4 &sb->s_type->i_mutex_key#8 jbd2_handle rcu_read_lock &rcu_state.gp_wq &p->pi_lock &rq->__lock irq_context: 0 sb_writers#4 &sb->s_type->i_mutex_key#8 jbd2_handle rcu_read_lock &rcu_state.gp_wq &p->pi_lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)events_unbound (work_completion)(&sub_info->work) quarantine_lock irq_context: 0 &sig->cred_guard_mutex &sig->exec_update_lock &mm->mmap_lock &anon_vma->rwsem rcu_read_lock rcu_node_0 irq_context: 0 &sig->cred_guard_mutex &sig->exec_update_lock &mm->mmap_lock &anon_vma->rwsem rcu_read_lock &rq->__lock irq_context: 0 &sig->cred_guard_mutex &sig->exec_update_lock &mm->mmap_lock &anon_vma->rwsem rcu_read_lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 &f->f_pos_lock sb_writers#4 &sb->s_type->i_mutex_key#8 batched_entropy_u32.lock irq_context: 0 sb_writers#4 &sb->s_type->i_mutex_key#8 remove_cache_srcu &cfs_rq->removed.lock irq_context: 0 sb_writers#4 &sb->s_type->i_mutex_key#8 remove_cache_srcu &rq->__lock &cfs_rq->removed.lock irq_context: 0 (wq_completion)events (work_completion)(&rdev->wiphy_work) &rdev->wiphy.mtx &wdev->mtx rcu_read_lock &sta->rate_ctrl_lock &c->lock irq_context: 0 (wq_completion)events (work_completion)(&rdev->wiphy_work) &rdev->wiphy.mtx &wdev->mtx rcu_read_lock &sta->rate_ctrl_lock &n->list_lock irq_context: 0 (wq_completion)events (work_completion)(&rdev->wiphy_work) &rdev->wiphy.mtx &wdev->mtx rcu_read_lock &sta->rate_ctrl_lock &n->list_lock &c->lock irq_context: 0 &f->f_pos_lock &type->i_mutex_dir_key#5 rcu_read_lock &cfs_rq->removed.lock irq_context: 0 &f->f_pos_lock &type->i_mutex_dir_key#5 rcu_read_lock &obj_hash[i].lock irq_context: 0 &f->f_pos_lock &type->i_mutex_dir_key#5 rcu_read_lock pool_lock#2 irq_context: 0 cb_lock &rdev->wiphy.mtx &local->mtx fs_reclaim &rq->__lock irq_context: 0 cb_lock &rdev->wiphy.mtx &local->mtx fs_reclaim &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 cb_lock genl_mutex lock kernfs_idr_lock &c->lock irq_context: 0 &type->i_mutex_dir_key#5 fs_reclaim mmu_notifier_invalidate_range_start &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 cb_lock genl_mutex rtnl_mutex &rdev->wiphy.mtx rcu_read_lock &pool->lock/1 &p->pi_lock &cfs_rq->removed.lock irq_context: 0 kn->active#16 &kernfs_locks->open_file_mutex[count] &rq->__lock irq_context: 0 kn->active#16 &kernfs_locks->open_file_mutex[count] &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 kn->active#18 &kernfs_locks->open_file_mutex[count] &rq->__lock irq_context: 0 kn->active#18 &kernfs_locks->open_file_mutex[count] &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 cb_lock genl_mutex rtnl_mutex &rdev->wiphy.mtx gdp_mutex rcu_read_lock rcu_node_0 irq_context: 0 cb_lock genl_mutex rtnl_mutex &rdev->wiphy.mtx gdp_mutex rcu_read_lock &rq->__lock irq_context: 0 cb_lock genl_mutex rtnl_mutex &rdev->wiphy.mtx gdp_mutex rcu_read_lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 cb_lock genl_mutex rtnl_mutex &rdev->wiphy.mtx gdp_mutex &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 cb_lock genl_mutex rtnl_mutex &rdev->wiphy.mtx lock kernfs_idr_lock &c->lock irq_context: 0 kn->active#16 &kernfs_locks->open_file_mutex[count] &n->list_lock irq_context: 0 kn->active#16 &kernfs_locks->open_file_mutex[count] &n->list_lock &c->lock irq_context: 0 &pipe->mutex/1 nfnl_subsys_ctnetlink rcu_read_lock once_lock irq_context: 0 &pipe->mutex/1 nfnl_subsys_ctnetlink rcu_read_lock once_lock crngs.lock irq_context: 0 &pipe->mutex/1 nfnl_subsys_ctnetlink rcu_read_lock rcu_read_lock &pool->lock irq_context: 0 &pipe->mutex/1 nfnl_subsys_ctnetlink rcu_read_lock rcu_read_lock &pool->lock &obj_hash[i].lock irq_context: 0 &pipe->mutex/1 nfnl_subsys_ctnetlink rcu_read_lock rcu_read_lock &pool->lock pool_lock#2 irq_context: 0 &pipe->mutex/1 nfnl_subsys_ctnetlink rcu_read_lock rcu_read_lock &pool->lock &p->pi_lock irq_context: 0 &pipe->mutex/1 nfnl_subsys_ctnetlink rcu_read_lock rcu_read_lock &pool->lock &p->pi_lock &rq->__lock irq_context: 0 &pipe->mutex/1 nfnl_subsys_ctnetlink rcu_read_lock rcu_read_lock &pool->lock &p->pi_lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 &pipe->mutex/1 nfnl_subsys_ctnetlink rcu_read_lock nl_table_lock irq_context: 0 &pipe->mutex/1 nfnl_subsys_ctnetlink rcu_read_lock nl_table_wait.lock irq_context: 0 &pipe->mutex/1 nfnl_subsys_ctnetlink rcu_read_lock rlock-AF_NETLINK irq_context: 0 &pipe->mutex/1 nfnl_subsys_ctnetlink rcu_read_lock rcu_read_lock pool_lock#2 irq_context: 0 &pipe->mutex/1 nfnl_subsys_ctnetlink rcu_read_lock rcu_read_lock nl_table_lock irq_context: 0 &pipe->mutex/1 nfnl_subsys_ctnetlink rcu_read_lock rcu_read_lock rlock-AF_NETLINK irq_context: 0 &pipe->mutex/1 nfnl_subsys_ctnetlink rcu_read_lock rcu_read_lock nl_table_wait.lock irq_context: 0 &pipe->mutex/1 nfnl_subsys_ctnetlink rcu_read_lock rcu_read_lock &c->lock irq_context: 0 &pipe->mutex/1 nfnl_subsys_ctnetlink rcu_read_lock rcu_read_lock &n->list_lock irq_context: 0 &pipe->mutex/1 nfnl_subsys_ctnetlink rcu_read_lock rcu_read_lock &n->list_lock &c->lock irq_context: 0 &pipe->mutex/1 nfnl_subsys_ctnetlink rcu_read_lock rcu_read_lock rcu_read_lock rcu_node_0 irq_context: 0 &pipe->mutex/1 nfnl_subsys_ctnetlink rcu_read_lock rcu_read_lock rcu_read_lock &rq->__lock irq_context: 0 &pipe->mutex/1 nfnl_subsys_ctnetlink rcu_read_lock rcu_read_lock rcu_read_lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 &pipe->mutex/1 nfnl_subsys_ctnetlink rcu_read_lock rcu_read_lock &obj_hash[i].lock irq_context: 0 &pipe->mutex/1 nfnl_subsys_ctnetlink rcu_read_lock &cnet->ecache.dying_lock irq_context: 0 &pipe->mutex/1 nfnl_subsys_ctnetlink rcu_read_lock &base->lock irq_context: 0 &pipe->mutex/1 nfnl_subsys_ctnetlink rcu_read_lock &base->lock &obj_hash[i].lock irq_context: 0 &pipe->mutex/1 nfnl_subsys_ctnetlink &n->list_lock irq_context: 0 &pipe->mutex/1 nfnl_subsys_ctnetlink &n->list_lock &c->lock irq_context: 0 cb_lock genl_mutex rfkill_global_mutex triggers_list_lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 &type->i_mutex_dir_key#4 fs_reclaim &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 kn->active#17 &rq->__lock irq_context: 0 kn->active#17 &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 cb_lock genl_mutex rtnl_mutex &rdev->wiphy.mtx rcu_read_lock &rcu_state.gp_wq irq_context: 0 cb_lock genl_mutex rtnl_mutex &rdev->wiphy.mtx rcu_read_lock &rcu_state.gp_wq &p->pi_lock irq_context: 0 cb_lock genl_mutex rtnl_mutex &rdev->wiphy.mtx rcu_read_lock &rcu_state.gp_wq &p->pi_lock &rq->__lock irq_context: 0 cb_lock genl_mutex rtnl_mutex &rdev->wiphy.mtx rcu_read_lock &rcu_state.gp_wq &p->pi_lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 &pipe->mutex/1 nfnl_subsys_ctnetlink &____s->seqcount#2 irq_context: 0 &pipe->mutex/1 nfnl_subsys_ctnetlink &____s->seqcount irq_context: 0 kn->active#17 &____s->seqcount#2 irq_context: 0 kn->active#17 &____s->seqcount irq_context: 0 &pipe->mutex/1 nfnl_subsys_ctnetlink rcu_read_lock rcu_read_lock quarantine_lock irq_context: 0 &pipe->mutex/1 nfnl_subsys_ctnetlink rcu_read_lock rcu_read_lock batched_entropy_u8.lock irq_context: 0 &pipe->mutex/1 nfnl_subsys_ctnetlink rcu_read_lock rcu_read_lock kfence_freelist_lock irq_context: 0 &pipe->mutex/1 nfnl_subsys_ctnetlink rcu_read_lock rcu_read_lock &meta->lock irq_context: 0 namespace_sem remove_cache_srcu pool_lock#2 irq_context: 0 cb_lock genl_mutex uevent_sock_mutex nl_table_wait.lock &p->pi_lock irq_context: 0 cb_lock genl_mutex uevent_sock_mutex nl_table_wait.lock &p->pi_lock &cfs_rq->removed.lock irq_context: 0 cb_lock genl_mutex uevent_sock_mutex nl_table_wait.lock &p->pi_lock &rq->__lock irq_context: 0 cb_lock genl_mutex uevent_sock_mutex nl_table_wait.lock &p->pi_lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 &pipe->mutex/1 nfnl_subsys_ctnetlink rcu_read_lock rcu_read_lock nl_table_wait.lock &p->pi_lock irq_context: 0 cb_lock genl_mutex triggers_list_lock &rq->__lock irq_context: 0 cb_lock genl_mutex triggers_list_lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 &pipe->mutex/1 nfnl_subsys_ctnetlink rcu_read_lock rcu_read_lock &____s->seqcount#2 irq_context: 0 &pipe->mutex/1 nfnl_subsys_ctnetlink rcu_read_lock rcu_read_lock &____s->seqcount irq_context: 0 cb_lock genl_mutex &sb->s_type->i_mutex_key#3 remove_cache_srcu pool_lock#2 irq_context: 0 cb_lock genl_mutex rtnl_mutex &rdev->wiphy.mtx &root->kernfs_rwsem &rq->__lock irq_context: 0 cb_lock genl_mutex rtnl_mutex &rdev->wiphy.mtx &root->kernfs_rwsem &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 cb_lock genl_mutex rtnl_mutex &rdev->wiphy.mtx net_rwsem &rq->__lock irq_context: 0 cb_lock genl_mutex rtnl_mutex &rdev->wiphy.mtx net_rwsem &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 &type->i_mutex_dir_key#4 fs_reclaim mmu_notifier_invalidate_range_start &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 sb_writers#4 &sb->s_type->i_mutex_key#8 quarantine_lock irq_context: 0 cb_lock genl_mutex uevent_sock_mutex &____s->seqcount#2 irq_context: 0 (wq_completion)events_unbound (work_completion)(&map->work) quarantine_lock irq_context: 0 cb_lock genl_mutex rtnl_mutex &rdev->wiphy.mtx &root->kernfs_rwsem &sem->wait_lock irq_context: 0 cb_lock genl_mutex &sb->s_type->i_mutex_key#3 fs_reclaim mmu_notifier_invalidate_range_start &rq->__lock irq_context: 0 cb_lock genl_mutex &sb->s_type->i_mutex_key#3 fs_reclaim mmu_notifier_invalidate_range_start &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)events (work_completion)(&(&cnet->ecache.dwork)->work) irq_context: 0 (wq_completion)events (work_completion)(&(&cnet->ecache.dwork)->work) &cnet->ecache.dying_lock irq_context: 0 (wq_completion)events (work_completion)(&(&cnet->ecache.dwork)->work) &cnet->ecache.dying_lock rcu_read_lock pool_lock#2 irq_context: 0 (wq_completion)events (work_completion)(&(&cnet->ecache.dwork)->work) &cnet->ecache.dying_lock rcu_read_lock &c->lock irq_context: 0 (wq_completion)events (work_completion)(&(&cnet->ecache.dwork)->work) &cnet->ecache.dying_lock rcu_read_lock nl_table_lock irq_context: 0 (wq_completion)events (work_completion)(&(&cnet->ecache.dwork)->work) &cnet->ecache.dying_lock rcu_read_lock rlock-AF_NETLINK irq_context: 0 (wq_completion)events (work_completion)(&(&cnet->ecache.dwork)->work) &cnet->ecache.dying_lock rcu_read_lock nl_table_wait.lock irq_context: 0 (wq_completion)events (work_completion)(&(&cnet->ecache.dwork)->work) &obj_hash[i].lock irq_context: 0 (wq_completion)events (work_completion)(&(&cnet->ecache.dwork)->work) pool_lock#2 irq_context: 0 (wq_completion)events (work_completion)(&(&cnet->ecache.dwork)->work) rcu_read_lock &pool->lock irq_context: 0 (wq_completion)events (work_completion)(&(&cnet->ecache.dwork)->work) rcu_read_lock &pool->lock &obj_hash[i].lock irq_context: 0 (wq_completion)events (work_completion)(&(&cnet->ecache.dwork)->work) rcu_read_lock &pool->lock pool_lock#2 irq_context: 0 (wq_completion)events (work_completion)(&(&cnet->ecache.dwork)->work) &cnet->ecache.dying_lock rcu_read_lock &____s->seqcount#2 irq_context: 0 (wq_completion)events (work_completion)(&(&cnet->ecache.dwork)->work) &cnet->ecache.dying_lock rcu_read_lock &____s->seqcount irq_context: 0 (wq_completion)events (work_completion)(&(&cnet->ecache.dwork)->work) &c->lock irq_context: 0 (wq_completion)events (work_completion)(&(&cnet->ecache.dwork)->work) &cnet->ecache.dying_lock rcu_read_lock &n->list_lock irq_context: 0 (wq_completion)events (work_completion)(&(&cnet->ecache.dwork)->work) &cnet->ecache.dying_lock rcu_read_lock &n->list_lock &c->lock irq_context: 0 (wq_completion)events (work_completion)(&(&cnet->ecache.dwork)->work) &rq->__lock irq_context: 0 (wq_completion)events (work_completion)(&(&cnet->ecache.dwork)->work) &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 kn->active#15 &kernfs_locks->open_file_mutex[count] &____s->seqcount#2 irq_context: 0 kn->active#15 &kernfs_locks->open_file_mutex[count] &____s->seqcount irq_context: 0 (wq_completion)events (work_completion)(&(&cnet->ecache.dwork)->work) &cnet->ecache.dying_lock rcu_read_lock &obj_hash[i].lock irq_context: 0 (wq_completion)events (work_completion)(&(&cnet->ecache.dwork)->work) &cnet->ecache.dying_lock rcu_read_lock rcu_read_lock pool_lock#2 irq_context: 0 (wq_completion)events (work_completion)(&(&cnet->ecache.dwork)->work) &base->lock irq_context: 0 (wq_completion)events (work_completion)(&(&cnet->ecache.dwork)->work) &base->lock &obj_hash[i].lock irq_context: softirq &(&cnet->ecache.dwork)->timer irq_context: softirq &(&cnet->ecache.dwork)->timer rcu_read_lock &pool->lock irq_context: softirq &(&cnet->ecache.dwork)->timer rcu_read_lock &pool->lock &obj_hash[i].lock irq_context: softirq &(&cnet->ecache.dwork)->timer rcu_read_lock &pool->lock &p->pi_lock irq_context: softirq &(&cnet->ecache.dwork)->timer rcu_read_lock &pool->lock &p->pi_lock &rq->__lock irq_context: softirq &(&cnet->ecache.dwork)->timer rcu_read_lock &pool->lock &p->pi_lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 sb_writers#4 &iint->mutex &meta->lock irq_context: 0 &pipe->mutex/1 nfnl_subsys_ctnetlink rcu_read_lock rcu_read_lock nl_table_wait.lock &p->pi_lock &rq->__lock irq_context: 0 &pipe->mutex/1 nfnl_subsys_ctnetlink rcu_read_lock rcu_read_lock nl_table_wait.lock &p->pi_lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)events (work_completion)(&(&cnet->ecache.dwork)->work) &cnet->ecache.dying_lock rcu_read_lock batched_entropy_u8.lock irq_context: 0 (wq_completion)events (work_completion)(&(&cnet->ecache.dwork)->work) &cnet->ecache.dying_lock rcu_read_lock kfence_freelist_lock irq_context: 0 (wq_completion)events (work_completion)(&(&cnet->ecache.dwork)->work) &cnet->ecache.dying_lock rcu_read_lock &meta->lock irq_context: 0 &sig->cred_guard_mutex &sig->exec_update_lock &mm->mmap_lock &rq->__lock &cfs_rq->removed.lock irq_context: 0 &nft_net->commit_mutex rcu_state.exp_mutex rcu_read_lock &pool->lock &p->pi_lock &rq->__lock &cfs_rq->removed.lock irq_context: 0 cb_lock genl_mutex rfkill_global_mutex uevent_sock_mutex rcu_read_lock &rcu_state.expedited_wq irq_context: 0 cb_lock genl_mutex rfkill_global_mutex uevent_sock_mutex rcu_read_lock &rcu_state.expedited_wq &p->pi_lock irq_context: 0 cb_lock genl_mutex rfkill_global_mutex uevent_sock_mutex rcu_read_lock &rcu_state.expedited_wq &p->pi_lock &rq->__lock irq_context: 0 cb_lock genl_mutex rfkill_global_mutex uevent_sock_mutex rcu_read_lock &rcu_state.expedited_wq &p->pi_lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 cb_lock genl_mutex rfkill_global_mutex uevent_sock_mutex nl_table_wait.lock &p->pi_lock irq_context: 0 cb_lock genl_mutex rfkill_global_mutex uevent_sock_mutex nl_table_wait.lock &p->pi_lock &rq->__lock irq_context: 0 cb_lock genl_mutex rfkill_global_mutex uevent_sock_mutex nl_table_wait.lock &p->pi_lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)events (work_completion)(&(&cnet->ecache.dwork)->work) &n->list_lock irq_context: 0 kn->active#20 &kernfs_locks->open_file_mutex[count] &____s->seqcount#2 irq_context: 0 kn->active#20 &kernfs_locks->open_file_mutex[count] &____s->seqcount irq_context: 0 cb_lock genl_mutex rtnl_mutex &rdev->wiphy.mtx remove_cache_srcu pool_lock#2 irq_context: 0 &pipe->mutex/1 nfnl_subsys_ctnetlink rcu_read_lock rcu_read_lock rcu_read_lock_bh &obj_hash[i].lock irq_context: 0 &pipe->mutex/1 nfnl_subsys_ctnetlink rcu_read_lock rcu_read_lock rcu_read_lock_bh pool_lock#2 irq_context: 0 &pipe->mutex/1 nfnl_subsys_ctnetlink rcu_read_lock rcu_read_lock rcu_read_lock pool_lock#2 irq_context: 0 &pipe->mutex/1 nfnl_subsys_ctnetlink rcu_read_lock rcu_read_lock rcu_read_lock rcu_read_lock_bh &obj_hash[i].lock irq_context: 0 &pipe->mutex/1 nfnl_subsys_ctnetlink rcu_read_lock rcu_read_lock rcu_read_lock rcu_read_lock_bh pool_lock#2 irq_context: 0 cb_lock genl_mutex remove_cache_srcu rcu_node_0 irq_context: 0 &pipe->mutex/1 nfnl_subsys_ctnetlink rcu_read_lock &pcp->lock &zone->lock irq_context: 0 &pipe->mutex/1 nfnl_subsys_ctnetlink rcu_read_lock rcu_read_lock rcu_read_lock &c->lock irq_context: 0 &pipe->mutex/1 nfnl_subsys_ctnetlink rcu_read_lock rcu_read_lock rcu_read_lock &____s->seqcount#2 irq_context: 0 &pipe->mutex/1 nfnl_subsys_ctnetlink rcu_read_lock rcu_read_lock rcu_read_lock &____s->seqcount irq_context: 0 (wq_completion)phy22 irq_context: 0 (wq_completion)phy22 (work_completion)(&local->reconfig_filter) irq_context: 0 (wq_completion)phy22 (work_completion)(&local->reconfig_filter) &local->filter_lock irq_context: 0 &type->i_mutex_dir_key#4 &root->kernfs_rwsem &p->pi_lock irq_context: 0 pernet_ops_rwsem rcu_state.barrier_mutex pool_lock irq_context: 0 (wq_completion)events (work_completion)(&(&cnet->ecache.dwork)->work) quarantine_lock irq_context: 0 (wq_completion)phy23 irq_context: 0 (wq_completion)phy23 &rq->__lock irq_context: 0 (wq_completion)phy23 (work_completion)(&local->reconfig_filter) irq_context: 0 (wq_completion)phy23 (work_completion)(&local->reconfig_filter) &local->filter_lock irq_context: 0 (wq_completion)phy24 irq_context: 0 (wq_completion)phy24 (work_completion)(&local->reconfig_filter) irq_context: 0 (wq_completion)phy24 (work_completion)(&local->reconfig_filter) &local->filter_lock irq_context: 0 rtnl_mutex &idev->mc_lock _xmit_ETHER rcu_read_lock &pool->lock/1 &p->pi_lock &cfs_rq->removed.lock irq_context: 0 rtnl_mutex rcu_read_lock &net->sctp.local_addr_lock &net->sctp.addr_wq_lock batched_entropy_u8.lock irq_context: 0 rtnl_mutex rcu_read_lock &net->sctp.local_addr_lock &net->sctp.addr_wq_lock kfence_freelist_lock irq_context: softirq (&net->sctp.addr_wq_timer) &net->sctp.addr_wq_lock &meta->lock irq_context: softirq (&net->sctp.addr_wq_timer) &net->sctp.addr_wq_lock kfence_freelist_lock irq_context: 0 cb_lock nlk_cb_mutex-GENERIC &rdev->wiphy.mtx rtnl_mutex.wait_lock irq_context: 0 cb_lock nlk_cb_mutex-GENERIC &rdev->wiphy.mtx &p->pi_lock irq_context: 0 sb_writers#4 &sb->s_type->i_mutex_key#8 mapping.invalidate_lock jbd2_handle &ei->i_data_sem mmu_notifier_invalidate_range_start &rq->__lock &cfs_rq->removed.lock irq_context: 0 sk_lock-AF_INET6 &sighand->siglock &p->pi_lock &cfs_rq->removed.lock irq_context: 0 sb_writers#4 &sb->s_type->i_mutex_key#8 &sb->s_type->i_mutex_key#8/4 jbd2_handle rcu_read_lock &xa->xa_lock#9 rcu_read_lock per_cpu_ptr(&cgroup_rstat_cpu_lock, cpu) irq_context: 0 sb_writers#4 &sb->s_type->i_mutex_key#8 &sb->s_type->i_mutex_key#8/4 jbd2_handle &folio_wait_table[i] &p->pi_lock irq_context: 0 sb_writers#4 &sb->s_type->i_mutex_key#8 &sb->s_type->i_mutex_key#8/4 jbd2_handle &folio_wait_table[i] &p->pi_lock &rq->__lock irq_context: 0 sb_writers#4 &sb->s_type->i_mutex_key#8 &sb->s_type->i_mutex_key#8/4 jbd2_handle &folio_wait_table[i] &p->pi_lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 &f->f_pos_lock sb_writers#4 lock#4 &lruvec->lru_lock per_cpu_ptr(&cgroup_rstat_cpu_lock, cpu) irq_context: 0 &mm->mmap_lock sb_pagefaults mapping.invalidate_lock &ei->i_data_sem &sem->wait_lock irq_context: 0 &mm->mmap_lock sb_pagefaults mapping.invalidate_lock &ei->i_data_sem &rq->__lock irq_context: 0 &mm->mmap_lock sb_pagefaults mapping.invalidate_lock &ei->i_data_sem &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 sb_writers#4 &sb->s_type->i_mutex_key#8 &sb->s_type->i_mutex_key#8/4 jbd2_handle &sem->wait_lock irq_context: 0 sb_writers#4 &sb->s_type->i_mutex_key#8 &sb->s_type->i_mutex_key#8/4 jbd2_handle &p->pi_lock irq_context: 0 sb_writers#4 &sb->s_type->i_mutex_key#8 &sb->s_type->i_mutex_key#8/4 jbd2_handle &p->pi_lock &rq->__lock irq_context: 0 sb_writers#4 &sb->s_type->i_mutex_key#8 &sb->s_type->i_mutex_key#8/4 jbd2_handle &p->pi_lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 sb_writers#4 &sb->s_type->i_mutex_key#8 &sb->s_type->i_mutex_key#8/4 jbd2_handle &journal->j_wait_updates &p->pi_lock irq_context: 0 sb_writers#4 &sb->s_type->i_mutex_key#8 &sb->s_type->i_mutex_key#8/4 jbd2_handle &journal->j_wait_updates &p->pi_lock &rq->__lock irq_context: 0 sb_writers#4 &sb->s_type->i_mutex_key#8 &sb->s_type->i_mutex_key#8/4 jbd2_handle &journal->j_wait_updates &p->pi_lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 sb_writers#4 &sb->s_type->i_mutex_key#8 &sb->s_type->i_mutex_key#8/4 jbd2_handle &ei->i_data_sem &ei->i_es_lock irq_context: 0 sb_writers#4 &sb->s_type->i_mutex_key#8 &sb->s_type->i_mutex_key#8/4 jbd2_handle &ei->i_data_sem &ei->i_data_sem/1 &sb->s_type->i_lock_key#22 irq_context: 0 sb_writers#4 &sb->s_type->i_mutex_key#8 &sb->s_type->i_mutex_key#8/4 jbd2_handle &ei->i_data_sem &ei->i_data_sem/1 &journal->j_state_lock irq_context: 0 sb_writers#4 &sb->s_type->i_mutex_key#8 &sb->s_type->i_mutex_key#8/4 jbd2_handle &ei->i_data_sem &ei->i_data_sem/1 rcu_read_lock pool_lock#2 irq_context: 0 sb_writers#4 &sb->s_type->i_mutex_key#8 &sb->s_type->i_mutex_key#8/4 jbd2_handle &ei->i_data_sem &ei->i_data_sem/1 &journal->j_revoke_lock irq_context: 0 sb_writers#4 &sb->s_type->i_mutex_key#8 &sb->s_type->i_mutex_key#8/4 jbd2_handle &ei->i_data_sem &ei->i_data_sem/1 &bgl->locks[i].lock &sbi->s_md_lock irq_context: 0 &mapping->i_mmap_rwsem mmu_notifier_invalidate_range_start irq_context: 0 &mapping->i_mmap_rwsem ptlock_ptr(page)#2 irq_context: 0 sb_writers#4 &sb->s_type->i_mutex_key#8 &sb->s_type->i_mutex_key#8/4 jbd2_handle &ei->i_data_sem &ei->i_data_sem/1 &bgl->locks[i].lock &obj_hash[i].lock irq_context: 0 sb_writers#4 &sb->s_type->i_mutex_key#8 &sb->s_type->i_mutex_key#8/4 jbd2_handle &ei->i_data_sem &ei->i_data_sem/1 &bgl->locks[i].lock pool_lock#2 irq_context: 0 sb_writers#4 &sb->s_type->i_mutex_key#8 &sb->s_type->i_mutex_key#8/4 jbd2_handle &ei->i_data_sem &n->list_lock irq_context: 0 sb_writers#4 &sb->s_type->i_mutex_key#8 &sb->s_type->i_mutex_key#8/4 jbd2_handle &ei->i_data_sem &n->list_lock &c->lock irq_context: 0 cb_lock rtnl_mutex &cfs_rq->removed.lock irq_context: 0 rtnl_mutex &rdev->wiphy.mtx lweventlist_lock &c->lock irq_context: 0 (wq_completion)phy25 irq_context: 0 (wq_completion)phy25 (work_completion)(&local->reconfig_filter) irq_context: 0 (wq_completion)phy25 (work_completion)(&local->reconfig_filter) &local->filter_lock irq_context: 0 rcu_read_lock_bh &nr_netdev_xmit_lock_key &meta->lock irq_context: 0 rcu_read_lock_bh &nr_netdev_xmit_lock_key kfence_freelist_lock irq_context: 0 sb_writers#4 &sb->s_type->i_mutex_key#8 &sb->s_type->i_mutex_key#8/4 jbd2_handle &rq->__lock &cfs_rq->removed.lock irq_context: 0 sb_writers#4 &sb->s_type->i_mutex_key#8/4 &sem->wait_lock irq_context: 0 sb_writers#4 &sb->s_type->i_mutex_key#8/4 &p->pi_lock irq_context: 0 sb_writers#4 &sb->s_type->i_mutex_key#8/4 &p->pi_lock &rq->__lock irq_context: 0 sb_writers#4 &sb->s_type->i_mutex_key#8/4 &p->pi_lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 sb_writers#4 &sb->s_type->i_mutex_key#8/4 &rq->__lock irq_context: 0 sb_writers#4 &sb->s_type->i_mutex_key#8/4 &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 br_ioctl_mutex rtnl_mutex &root->kernfs_rwsem &sem->wait_lock irq_context: 0 rcu_read_lock &dtab->index_lock irq_context: 0 rcu_read_lock &dtab->index_lock stock_lock irq_context: 0 rcu_read_lock &dtab->index_lock pool_lock#2 irq_context: 0 br_ioctl_mutex rtnl_mutex &root->kernfs_rwsem &sem->wait_lock irq_context: 0 br_ioctl_mutex rtnl_mutex rcu_read_lock &dtab->index_lock irq_context: 0 br_ioctl_mutex rtnl_mutex rcu_read_lock &dtab->index_lock &obj_hash[i].lock irq_context: 0 br_ioctl_mutex rtnl_mutex rcu_read_lock &dtab->index_lock pool_lock#2 irq_context: 0 sb_writers#4 &sbi->s_writepages_rwsem jbd2_handle &ei->i_data_sem bit_wait_table + i irq_context: 0 (wq_completion)events_unbound (work_completion)(&map->work) rcu_state.exp_mutex &cfs_rq->removed.lock irq_context: 0 (wq_completion)events_unbound (work_completion)(&map->work) rcu_state.barrier_mutex &rq->__lock &cfs_rq->removed.lock irq_context: 0 &f->f_pos_lock sb_writers#4 &sbi->s_writepages_rwsem rcu_read_lock &c->lock irq_context: 0 sb_writers#4 &sbi->s_writepages_rwsem jbd2_handle &ei->i_data_sem rcu_read_lock &rq->__lock irq_context: 0 sb_writers#4 &sbi->s_writepages_rwsem jbd2_handle &ei->i_data_sem rcu_read_lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 sb_writers#4 &sb->s_type->i_mutex_key#8 &sb->s_type->i_mutex_key#8/4 &ei->i_data_sem &ei->i_data_sem/1 &c->lock irq_context: 0 &f->f_pos_lock sb_writers#4 &sbi->s_writepages_rwsem jbd2_handle rcu_read_lock &c->lock irq_context: 0 &f->f_pos_lock &type->i_mutex_dir_key#5 rcu_read_lock &rcu_state.expedited_wq irq_context: 0 &f->f_pos_lock &type->i_mutex_dir_key#5 rcu_read_lock &rcu_state.expedited_wq &p->pi_lock irq_context: 0 &f->f_pos_lock &type->i_mutex_dir_key#5 rcu_read_lock &rcu_state.expedited_wq &p->pi_lock &rq->__lock irq_context: 0 &f->f_pos_lock &type->i_mutex_dir_key#5 rcu_read_lock &rcu_state.expedited_wq &p->pi_lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 sb_writers#4 &sb->s_type->i_mutex_key#8 &sb->s_type->i_mutex_key#8/4 jbd2_handle &ei->i_data_sem &rq->__lock irq_context: 0 sb_writers#4 &sb->s_type->i_mutex_key#8 &sb->s_type->i_mutex_key#8/4 jbd2_handle &ei->i_data_sem &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 sb_writers#4 &type->i_mutex_dir_key#3/1 jbd2_handle &ei->i_data_sem rcu_node_0 irq_context: 0 sb_writers#4 &type->i_mutex_dir_key#3/1 jbd2_handle &ei->i_data_sem &rcu_state.expedited_wq irq_context: 0 sb_writers#4 &type->i_mutex_dir_key#3/1 jbd2_handle &ei->i_data_sem &rcu_state.expedited_wq &p->pi_lock irq_context: 0 sb_writers#4 &type->i_mutex_dir_key#3/1 jbd2_handle &ei->i_data_sem &rcu_state.expedited_wq &p->pi_lock &rq->__lock irq_context: 0 sb_writers#4 &type->i_mutex_dir_key#3/1 jbd2_handle &ei->i_data_sem &rcu_state.expedited_wq &p->pi_lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 &sig->cred_guard_mutex &sig->exec_update_lock &mm->mmap_lock &cfs_rq->removed.lock irq_context: 0 br_ioctl_mutex rtnl_mutex rcu_state.exp_mutex &rq->__lock &cfs_rq->removed.lock irq_context: 0 br_ioctl_mutex rtnl_mutex fs_reclaim mmu_notifier_invalidate_range_start &rq->__lock irq_context: 0 br_ioctl_mutex rtnl_mutex fs_reclaim mmu_notifier_invalidate_range_start &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)events_unbound (work_completion)(&map->work) rcu_state.exp_mutex pool_lock#2 irq_context: 0 (wq_completion)events_unbound (work_completion)(&map->work) rcu_node_0 irq_context: 0 (wq_completion)events_unbound (work_completion)(&map->work) &rcu_state.expedited_wq irq_context: 0 (wq_completion)events_unbound (work_completion)(&map->work) &rcu_state.expedited_wq &p->pi_lock irq_context: 0 (wq_completion)events_unbound (work_completion)(&map->work) &rcu_state.expedited_wq &p->pi_lock &rq->__lock irq_context: 0 (wq_completion)events_unbound (work_completion)(&map->work) &rcu_state.expedited_wq &p->pi_lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 tracepoints_mutex remove_cache_srcu irq_context: 0 tracepoints_mutex remove_cache_srcu quarantine_lock irq_context: 0 tracepoints_mutex remove_cache_srcu &c->lock irq_context: 0 tracepoints_mutex remove_cache_srcu &rq->__lock irq_context: 0 tracepoints_mutex remove_cache_srcu &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 tracepoints_mutex remove_cache_srcu &n->list_lock irq_context: 0 tracepoints_mutex remove_cache_srcu rcu_read_lock pool_lock#2 irq_context: 0 tracepoints_mutex remove_cache_srcu &obj_hash[i].lock irq_context: 0 dup_mmap_sem &mm->mmap_lock &mm->mmap_lock/1 &vma->vm_lock->lock &rq->__lock &cfs_rq->removed.lock irq_context: 0 &mm->mmap_lock sb_pagefaults mapping.invalidate_lock &type->s_umount_key#32 rcu_read_lock &c->lock irq_context: 0 &mm->mmap_lock sb_pagefaults mapping.invalidate_lock &type->s_umount_key#32 rcu_read_lock pool_lock#2 irq_context: 0 &mm->mmap_lock sb_pagefaults mapping.invalidate_lock &type->s_umount_key#32 rcu_read_lock &wb->work_lock irq_context: 0 &mm->mmap_lock sb_pagefaults mapping.invalidate_lock &type->s_umount_key#32 rcu_read_lock &wb->work_lock &obj_hash[i].lock irq_context: 0 &mm->mmap_lock sb_pagefaults mapping.invalidate_lock &type->s_umount_key#32 rcu_read_lock &wb->work_lock &base->lock irq_context: 0 &mm->mmap_lock sb_pagefaults mapping.invalidate_lock &type->s_umount_key#32 rcu_read_lock &wb->work_lock &base->lock &obj_hash[i].lock irq_context: 0 &mm->mmap_lock sb_pagefaults mapping.invalidate_lock &type->s_umount_key#32 rcu_read_lock &wb->work_lock rcu_read_lock &pool->lock/1 irq_context: 0 &mm->mmap_lock sb_pagefaults mapping.invalidate_lock &type->s_umount_key#32 rcu_read_lock &wb->work_lock rcu_read_lock &pool->lock/1 &obj_hash[i].lock irq_context: 0 &mm->mmap_lock sb_pagefaults mapping.invalidate_lock &type->s_umount_key#32 rcu_read_lock &wb->work_lock rcu_read_lock &pool->lock/1 pool_lock#2 irq_context: 0 &mm->mmap_lock sb_pagefaults mapping.invalidate_lock &type->s_umount_key#32 rcu_read_lock &wb->work_lock rcu_read_lock &pool->lock/1 &p->pi_lock irq_context: 0 &mm->mmap_lock sb_pagefaults mapping.invalidate_lock &type->s_umount_key#32 rcu_read_lock &wb->work_lock rcu_read_lock &pool->lock/1 &p->pi_lock &rq->__lock irq_context: 0 &mm->mmap_lock sb_pagefaults mapping.invalidate_lock &type->s_umount_key#32 rcu_read_lock &wb->work_lock rcu_read_lock &pool->lock/1 &p->pi_lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 &mm->mmap_lock sb_pagefaults mapping.invalidate_lock &type->s_umount_key#32 rcu_read_lock rcu_node_0 irq_context: 0 &mm->mmap_lock sb_pagefaults mapping.invalidate_lock &type->s_umount_key#32 rcu_read_lock &rq->__lock irq_context: 0 &mm->mmap_lock sb_pagefaults mapping.invalidate_lock &type->s_umount_key#32 rcu_read_lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 &mm->mmap_lock sb_pagefaults mapping.invalidate_lock &type->s_umount_key#32 &bdi->wb_waitq irq_context: 0 &mm->mmap_lock sb_pagefaults mapping.invalidate_lock &type->s_umount_key#32 &rq->__lock irq_context: 0 &mm->mmap_lock sb_pagefaults mapping.invalidate_lock &type->s_umount_key#32 &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)writeback (work_completion)(&(&wb->dwork)->work) &sbi->s_writepages_rwsem jbd2_handle remove_cache_srcu irq_context: 0 (wq_completion)writeback (work_completion)(&(&wb->dwork)->work) &sbi->s_writepages_rwsem jbd2_handle remove_cache_srcu quarantine_lock irq_context: 0 (wq_completion)writeback (work_completion)(&(&wb->dwork)->work) &sbi->s_writepages_rwsem jbd2_handle remove_cache_srcu &c->lock irq_context: 0 (wq_completion)writeback (work_completion)(&(&wb->dwork)->work) &sbi->s_writepages_rwsem jbd2_handle remove_cache_srcu &n->list_lock irq_context: 0 (wq_completion)writeback (work_completion)(&(&wb->dwork)->work) &sbi->s_writepages_rwsem jbd2_handle remove_cache_srcu &rq->__lock irq_context: 0 (wq_completion)writeback (work_completion)(&(&wb->dwork)->work) &sbi->s_writepages_rwsem jbd2_handle remove_cache_srcu rcu_read_lock pool_lock#2 irq_context: 0 (wq_completion)writeback (work_completion)(&(&wb->dwork)->work) &sbi->s_writepages_rwsem jbd2_handle remove_cache_srcu &obj_hash[i].lock irq_context: 0 (wq_completion)writeback (work_completion)(&(&wb->dwork)->work) &sbi->s_writepages_rwsem jbd2_handle remove_cache_srcu pool_lock#2 irq_context: 0 (wq_completion)writeback (work_completion)(&(&wb->dwork)->work) &sbi->s_writepages_rwsem jbd2_handle rcu_read_lock &rq->__lock &cfs_rq->removed.lock irq_context: 0 &sb->s_type->i_mutex_key#10 sk_lock-AF_SMC &rq->__lock irq_context: 0 &sb->s_type->i_mutex_key#10 sk_lock-AF_SMC &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)writeback (work_completion)(&(&wb->dwork)->work) &sbi->s_writepages_rwsem jbd2_handle rcu_read_lock &rcu_state.gp_wq irq_context: 0 (wq_completion)writeback (work_completion)(&(&wb->dwork)->work) &sbi->s_writepages_rwsem jbd2_handle rcu_read_lock &rcu_state.gp_wq &p->pi_lock irq_context: 0 (wq_completion)writeback (work_completion)(&(&wb->dwork)->work) &sbi->s_writepages_rwsem jbd2_handle rcu_read_lock &rcu_state.gp_wq &p->pi_lock &rq->__lock irq_context: 0 (wq_completion)writeback (work_completion)(&(&wb->dwork)->work) &sbi->s_writepages_rwsem jbd2_handle rcu_read_lock &rcu_state.gp_wq &p->pi_lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 rtnl_mutex &rdev->wiphy.mtx rcu_read_lock &pool->lock &p->pi_lock &rq->__lock &cfs_rq->removed.lock irq_context: 0 (wq_completion)phy26 irq_context: 0 (wq_completion)phy26 (work_completion)(&local->reconfig_filter) irq_context: 0 (wq_completion)phy26 (work_completion)(&local->reconfig_filter) &local->filter_lock irq_context: 0 &type->i_mutex_dir_key#4 &root->kernfs_rwsem batched_entropy_u8.lock irq_context: 0 &type->i_mutex_dir_key#4 &root->kernfs_rwsem kfence_freelist_lock irq_context: 0 sk_lock-AF_UNIX fs_reclaim irq_context: 0 sk_lock-AF_UNIX fs_reclaim mmu_notifier_invalidate_range_start irq_context: 0 sk_lock-AF_UNIX &obj_hash[i].lock irq_context: 0 sk_lock-AF_UNIX pool_lock#2 irq_context: 0 sb_writers &type->i_mutex_dir_key#2 &rq->__lock irq_context: 0 &mm->mmap_lock &anon_vma->rwsem pgd_lock irq_context: 0 &mm->mmap_lock &anon_vma->rwsem key irq_context: 0 &mm->mmap_lock &anon_vma->rwsem pcpu_lock irq_context: 0 &mm->mmap_lock &anon_vma->rwsem percpu_counters_lock irq_context: 0 &mm->mmap_lock &anon_vma->rwsem pcpu_lock stock_lock irq_context: 0 (wq_completion)phy27 irq_context: 0 (wq_completion)phy27 (work_completion)(&local->reconfig_filter) irq_context: 0 (wq_completion)phy27 (work_completion)(&local->reconfig_filter) &local->filter_lock irq_context: 0 sb_writers#4 mapping.invalidate_lock batched_entropy_u32.lock irq_context: softirq &fq->mq_flush_lock &x->wait#26 &p->pi_lock &rq->__lock &cfs_rq->removed.lock irq_context: 0 &f->f_pos_lock sb_writers#4 &sb->s_type->i_mutex_key#8 jbd2_handle &sbi->s_orphan_lock &____s->seqcount irq_context: 0 &f->f_pos_lock sb_writers#4 &sb->s_type->i_mutex_key#8 jbd2_handle &sbi->s_orphan_lock bit_wait_table + i irq_context: 0 &f->f_pos_lock sb_writers#4 &sb->s_type->i_mutex_key#8 jbd2_handle &sbi->s_orphan_lock rcu_read_lock pool_lock#2 irq_context: 0 &f->f_pos_lock sb_writers#4 &sb->s_type->i_mutex_key#8 jbd2_handle &sbi->s_orphan_lock &obj_hash[i].lock irq_context: 0 sb_writers#4 &sb->s_type->i_mutex_key#8 rcu_read_lock &n->list_lock irq_context: 0 sb_writers#4 &sb->s_type->i_mutex_key#8 rcu_read_lock &n->list_lock &c->lock irq_context: 0 sb_writers#4 &sb->s_type->i_mutex_key#8 pgd_lock irq_context: 0 sb_writers#4 &sb->s_type->i_mutex_key#8 key irq_context: 0 sb_writers#4 &sb->s_type->i_mutex_key#8 pcpu_lock irq_context: 0 sb_writers#4 &sb->s_type->i_mutex_key#8 percpu_counters_lock irq_context: 0 &sig->cred_guard_mutex &sig->exec_update_lock delayed_uprobe_lock &rq->__lock irq_context: 0 &sig->cred_guard_mutex &sig->exec_update_lock delayed_uprobe_lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)ipv6_addrconf (work_completion)(&(&ifa->dad_work)->work) rtnl_mutex &idev->mc_lock _xmit_ETHER &local->filter_lock &c->lock irq_context: 0 &sig->cred_guard_mutex &mm->mmap_lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 kn->active#60 remove_cache_srcu rcu_read_lock rcu_node_0 irq_context: 0 kn->active#60 remove_cache_srcu rcu_read_lock &rq->__lock irq_context: 0 kn->active#60 remove_cache_srcu rcu_read_lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 rtnl_mutex &macvlan_netdev_addr_lock_key/1 &n->list_lock irq_context: 0 rtnl_mutex &macvlan_netdev_addr_lock_key/1 &n->list_lock &c->lock irq_context: 0 sb_writers &type->i_mutex_dir_key#2 &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 pernet_ops_rwsem wq_pool_mutex &rq->__lock irq_context: 0 pernet_ops_rwsem wq_pool_mutex &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 pernet_ops_rwsem ovs_mutex nf_ct_proto_mutex fill_pool_map-wait-type-override &____s->seqcount#2 irq_context: 0 pernet_ops_rwsem ovs_mutex nf_ct_proto_mutex fill_pool_map-wait-type-override &____s->seqcount irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem rcu_read_lock &rcu_state.expedited_wq &p->pi_lock &rq->__lock &cfs_rq->removed.lock irq_context: 0 pernet_ops_rwsem stock_lock per_cpu_ptr(&cgroup_rstat_cpu_lock, cpu) irq_context: 0 rtnl_mutex &lapb->lock &n->list_lock irq_context: 0 rtnl_mutex &lapb->lock &n->list_lock &c->lock irq_context: 0 &type->s_umount_key#23/1 remove_cache_srcu irq_context: 0 &type->s_umount_key#23/1 remove_cache_srcu quarantine_lock irq_context: 0 &type->s_umount_key#23/1 remove_cache_srcu &c->lock irq_context: 0 &type->s_umount_key#23/1 remove_cache_srcu &n->list_lock irq_context: 0 &type->s_umount_key#23/1 remove_cache_srcu &obj_hash[i].lock irq_context: 0 &type->s_umount_key#23/1 remove_cache_srcu pool_lock#2 irq_context: 0 &type->s_umount_key#23/1 remove_cache_srcu rcu_read_lock pool_lock#2 irq_context: 0 sb_writers#4 &type->i_mutex_dir_key#3 fill_pool_map-wait-type-override pool_lock#2 irq_context: 0 sb_writers#4 &type->i_mutex_dir_key#3 fill_pool_map-wait-type-override &c->lock irq_context: 0 sb_writers#4 &type->i_mutex_dir_key#3 fill_pool_map-wait-type-override &____s->seqcount#2 irq_context: 0 sb_writers#4 &type->i_mutex_dir_key#3 fill_pool_map-wait-type-override &____s->seqcount irq_context: 0 sb_writers#4 &type->i_mutex_dir_key#3 fill_pool_map-wait-type-override pool_lock irq_context: 0 pernet_ops_rwsem __ip_vs_app_mutex fill_pool_map-wait-type-override pool_lock#2 irq_context: 0 pernet_ops_rwsem __ip_vs_app_mutex fill_pool_map-wait-type-override pool_lock irq_context: softirq &(&forw_packet_aggr->delayed_work)->timer rcu_read_lock &pool->lock/1 &p->pi_lock &rq->__lock &obj_hash[i].lock irq_context: softirq &(&forw_packet_aggr->delayed_work)->timer rcu_read_lock &pool->lock/1 &p->pi_lock &rq->__lock &base->lock irq_context: softirq &(&forw_packet_aggr->delayed_work)->timer rcu_read_lock &pool->lock/1 &p->pi_lock &rq->__lock &base->lock &obj_hash[i].lock irq_context: 0 tasklist_lock &sighand->siglock &n->list_lock irq_context: 0 tasklist_lock &sighand->siglock &n->list_lock &c->lock irq_context: 0 &type->i_mutex_dir_key#5 fs_reclaim &rq->__lock irq_context: 0 &type->i_mutex_dir_key#5 fs_reclaim &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)events free_ipc_work &type->s_umount_key#51 rcu_read_lock rcu_node_0 irq_context: 0 (wq_completion)events free_ipc_work &type->s_umount_key#51 rcu_read_lock &rcu_state.expedited_wq irq_context: 0 (wq_completion)events free_ipc_work &type->s_umount_key#51 rcu_read_lock &rcu_state.expedited_wq &p->pi_lock irq_context: 0 (wq_completion)events free_ipc_work &type->s_umount_key#51 rcu_read_lock &rcu_state.expedited_wq &p->pi_lock &rq->__lock irq_context: 0 (wq_completion)events free_ipc_work &type->s_umount_key#51 rcu_read_lock &rcu_state.expedited_wq &p->pi_lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 pernet_ops_rwsem devices_rwsem rdma_nets_rwsem &device->compat_devs_mutex remove_cache_srcu &cfs_rq->removed.lock irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem dev_pm_qos_sysfs_mtx &root->kernfs_rwsem rcu_read_lock &rq->__lock irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem dev_pm_qos_sysfs_mtx &root->kernfs_rwsem rcu_read_lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem wq_pool_mutex &rq->__lock irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem wq_pool_mutex &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 sb_writers#4 &type->i_mutex_dir_key#3/1 jbd2_handle &journal->j_state_lock &journal->j_wait_commit &p->pi_lock &cfs_rq->removed.lock irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem fill_pool_map-wait-type-override &____s->seqcount#2 irq_context: 0 pernet_ops_rwsem sysctl_lock fill_pool_map-wait-type-override &____s->seqcount#2 irq_context: 0 pernet_ops_rwsem sysctl_lock fill_pool_map-wait-type-override &____s->seqcount irq_context: 0 &type->s_umount_key#23/1 list_lrus_mutex &rq->__lock irq_context: 0 sk_lock-AF_INET k-sk_lock-AF_INET &____s->seqcount#8 irq_context: 0 sk_lock-AF_INET k-sk_lock-AF_INET &tcp_hashinfo.bhash[i].lock &tcp_hashinfo.bhash2[i].lock batched_entropy_u8.lock irq_context: 0 sk_lock-AF_INET k-sk_lock-AF_INET &tcp_hashinfo.bhash[i].lock &tcp_hashinfo.bhash2[i].lock &hashinfo->ehash_locks[i] irq_context: 0 sk_lock-AF_INET k-sk_lock-AF_INET rcu_read_lock rcu_read_lock &nf_conntrack_locks[i] &nf_conntrack_locks[i]/1 irq_context: 0 sk_lock-AF_INET k-sk_lock-AF_INET rcu_read_lock rcu_read_lock &nf_conntrack_locks[i] &nf_conntrack_locks[i]/1 batched_entropy_u8.lock irq_context: 0 sk_lock-AF_INET k-sk_lock-AF_INET rcu_read_lock rcu_read_lock &nf_conntrack_locks[i]/1 irq_context: 0 sk_lock-AF_INET k-sk_lock-AF_INET &tcp_hashinfo.bhash[i].lock &obj_hash[i].lock irq_context: 0 sk_lock-AF_INET k-sk_lock-AF_INET rcu_read_lock &pool->lock irq_context: 0 sk_lock-AF_INET k-sk_lock-AF_INET rcu_read_lock &pool->lock &obj_hash[i].lock irq_context: 0 sk_lock-AF_INET k-sk_lock-AF_INET rcu_read_lock &pool->lock pool_lock#2 irq_context: 0 sk_lock-AF_INET k-sk_lock-AF_INET rcu_read_lock &pool->lock &p->pi_lock irq_context: 0 sk_lock-AF_INET k-sk_lock-AF_INET rcu_read_lock &pool->lock &p->pi_lock &rq->__lock irq_context: 0 sk_lock-AF_INET k-sk_lock-AF_INET rcu_read_lock &pool->lock &p->pi_lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 sk_lock-AF_INET k-sk_lock-AF_INET/1 &obj_hash[i].lock irq_context: 0 sk_lock-AF_INET k-sk_lock-AF_INET/1 slock-AF_INET irq_context: softirq _xmit_ETHER#2 &base->lock irq_context: softirq _xmit_ETHER#2 &base->lock &obj_hash[i].lock irq_context: 0 pernet_ops_rwsem nf_ct_proto_mutex defrag6_mutex nf_hook_mutex fs_reclaim mmu_notifier_invalidate_range_start &rq->__lock irq_context: 0 pernet_ops_rwsem nf_ct_proto_mutex defrag6_mutex nf_hook_mutex fs_reclaim mmu_notifier_invalidate_range_start &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 sk_lock-AF_PACKET rcu_state.exp_mutex &rq->__lock &cfs_rq->removed.lock irq_context: 0 pernet_ops_rwsem devices_rwsem rdma_nets_rwsem &device->compat_devs_mutex &rxe->usdev_lock rcu_node_0 irq_context: 0 pernet_ops_rwsem devices_rwsem rdma_nets_rwsem &device->compat_devs_mutex &rxe->usdev_lock &rcu_state.expedited_wq irq_context: 0 pernet_ops_rwsem devices_rwsem rdma_nets_rwsem &device->compat_devs_mutex &rxe->usdev_lock &rcu_state.expedited_wq &p->pi_lock irq_context: 0 pernet_ops_rwsem devices_rwsem rdma_nets_rwsem &device->compat_devs_mutex &rxe->usdev_lock &rcu_state.expedited_wq &p->pi_lock &rq->__lock irq_context: 0 pernet_ops_rwsem devices_rwsem rdma_nets_rwsem &device->compat_devs_mutex &rxe->usdev_lock &rcu_state.expedited_wq &p->pi_lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 &sig->cred_guard_mutex rcu_node_0 irq_context: softirq rcu_callback pcpu_lock rcu_read_lock per_cpu_ptr(&cgroup_rstat_cpu_lock, cpu) irq_context: 0 tasklist_lock &sighand->siglock &____s->seqcount#2 irq_context: 0 &sig->cred_guard_mutex &sig->exec_update_lock &mm->mmap_lock rcu_read_lock rcu_read_lock rcu_node_0 irq_context: 0 &sig->cred_guard_mutex &sig->exec_update_lock &mm->mmap_lock rcu_read_lock rcu_read_lock &rq->__lock irq_context: 0 &sig->cred_guard_mutex &sig->exec_update_lock &mm->mmap_lock rcu_read_lock rcu_read_lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 &mm->mmap_lock &anon_vma->rwsem rcu_read_lock &cfs_rq->removed.lock irq_context: 0 &mm->mmap_lock &anon_vma->rwsem rcu_read_lock &obj_hash[i].lock irq_context: 0 &f->f_pos_lock sb_writers#4 &sb->s_type->i_mutex_key#8 &ei->i_data_sem quarantine_lock irq_context: 0 sk_lock-AF_INET k-sk_lock-AF_INET/1 &rq->__lock irq_context: 0 sk_lock-AF_INET k-sk_lock-AF_INET/1 &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 &mm->mmap_lock pool_lock irq_context: 0 &sb->s_type->i_mutex_key#10 sk_lock-AF_INET k-sk_lock-AF_INET/1 &tcp_hashinfo.bhash[i].lock irq_context: 0 &sb->s_type->i_mutex_key#10 sk_lock-AF_INET k-sk_lock-AF_INET/1 &tcp_hashinfo.bhash[i].lock stock_lock irq_context: 0 &sb->s_type->i_mutex_key#10 sk_lock-AF_INET k-sk_lock-AF_INET/1 &tcp_hashinfo.bhash[i].lock &obj_hash[i].lock irq_context: 0 &sb->s_type->i_mutex_key#10 sk_lock-AF_INET k-sk_lock-AF_INET/1 &tcp_hashinfo.bhash[i].lock pool_lock#2 irq_context: 0 &sb->s_type->i_mutex_key#10 sk_lock-AF_INET k-sk_lock-AF_INET/1 &tcp_hashinfo.bhash[i].lock &tcp_hashinfo.bhash2[i].lock irq_context: 0 &sb->s_type->i_mutex_key#10 sk_lock-AF_INET k-sk_lock-AF_INET/1 &tcp_hashinfo.bhash[i].lock &tcp_hashinfo.bhash2[i].lock stock_lock irq_context: 0 &sb->s_type->i_mutex_key#10 sk_lock-AF_INET k-sk_lock-AF_INET/1 &tcp_hashinfo.bhash[i].lock &tcp_hashinfo.bhash2[i].lock &obj_hash[i].lock irq_context: 0 &sb->s_type->i_mutex_key#10 sk_lock-AF_INET k-sk_lock-AF_INET/1 &tcp_hashinfo.bhash[i].lock &tcp_hashinfo.bhash2[i].lock &____s->seqcount irq_context: 0 &sb->s_type->i_mutex_key#10 sk_lock-AF_INET k-sk_lock-AF_INET/1 &tcp_hashinfo.bhash[i].lock &tcp_hashinfo.bhash2[i].lock pool_lock#2 irq_context: 0 &sb->s_type->i_mutex_key#10 sk_lock-AF_INET k-sk_lock-AF_INET/1 &tcp_hashinfo.bhash[i].lock &tcp_hashinfo.bhash2[i].lock rcu_read_lock pool_lock#2 irq_context: 0 &f->f_pos_lock &p->lock stock_lock irq_context: 0 &f->f_pos_lock &p->lock cgroup_mutex irq_context: 0 &f->f_pos_lock &p->lock cgroup_mutex css_set_lock irq_context: 0 &f->f_pos_lock &p->lock cgroup_mutex css_set_lock kernfs_rename_lock irq_context: 0 &f->f_pos_lock &p->lock &obj_hash[i].lock irq_context: 0 sk_lock-AF_PACKET &po->bind_lock rcu_read_lock batched_entropy_u8.lock irq_context: 0 sk_lock-AF_PACKET &po->bind_lock rcu_read_lock kfence_freelist_lock irq_context: 0 sb_writers &sb->s_type->i_mutex_key#4 &rcu_state.expedited_wq irq_context: 0 sb_writers &sb->s_type->i_mutex_key#4 &rcu_state.expedited_wq &p->pi_lock irq_context: 0 sb_writers &sb->s_type->i_mutex_key#4 &rcu_state.expedited_wq &p->pi_lock &rq->__lock irq_context: 0 sb_writers &sb->s_type->i_mutex_key#4 &rcu_state.expedited_wq &p->pi_lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 sb_writers#4 rcu_read_lock &rcu_state.gp_wq &p->pi_lock &rq->__lock &cfs_rq->removed.lock irq_context: 0 sb_writers#4 &sb->s_type->i_mutex_key#8 jbd2_handle &ei->i_data_sem batched_entropy_u8.lock irq_context: 0 sb_writers#4 &sb->s_type->i_mutex_key#8 jbd2_handle &ei->i_data_sem kfence_freelist_lock irq_context: 0 sb_writers#4 &sb->s_type->i_mutex_key#8 jbd2_handle &ei->i_data_sem &meta->lock irq_context: 0 (wq_completion)wg-kex-wg1#10 (work_completion)(&peer->transmit_handshake_work) &n->list_lock irq_context: 0 (wq_completion)wg-kex-wg1#10 (work_completion)(&peer->transmit_handshake_work) &n->list_lock &c->lock irq_context: 0 (wq_completion)wg-kex-wg1#12 (work_completion)(&({ do { const void *__vpp_verify = (typeof((worker) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __asm__ ("" : "=r"(__ptr) : "0"((typeof(*((worker))) *)((worker)))); (typeof((typeof(*((worker))) *)((worker)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); })->work) &rq->__lock irq_context: 0 (wq_completion)wg-kex-wg2#2 (work_completion)(&({ do { const void *__vpp_verify = (typeof((worker) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __asm__ ("" : "=r"(__ptr) : "0"((typeof(*((worker))) *)((worker)))); (typeof((typeof(*((worker))) *)((worker)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); })->work) &peer->endpoint_lock rcu_read_lock_bh rcu_read_lock rcu_read_lock_bh rcu_read_lock tk_core.seq.seqcount irq_context: 0 (wq_completion)wg-kex-wg2#2 (work_completion)(&({ do { const void *__vpp_verify = (typeof((worker) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __asm__ ("" : "=r"(__ptr) : "0"((typeof(*((worker))) *)((worker)))); (typeof((typeof(*((worker))) *)((worker)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); })->work) &peer->endpoint_lock rcu_read_lock_bh rcu_read_lock rcu_read_lock_bh rcu_read_lock &obj_hash[i].lock irq_context: 0 (wq_completion)events_power_efficient (check_lifetime_work).work rcu_read_lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 dup_mmap_sem &mm->mmap_lock &mm->mmap_lock/1 rcu_read_lock &rcu_state.expedited_wq &p->pi_lock &rq->__lock &cfs_rq->removed.lock irq_context: 0 (wq_completion)wg-kex-wg1#3 (work_completion)(&peer->transmit_handshake_work) &peer->endpoint_lock rcu_read_lock_bh rcu_read_lock rcu_read_lock_bh rcu_read_lock tk_core.seq.seqcount irq_context: 0 (wq_completion)wg-kex-wg1#3 (work_completion)(&peer->transmit_handshake_work) &peer->endpoint_lock rcu_read_lock_bh rcu_read_lock rcu_read_lock_bh rcu_read_lock &obj_hash[i].lock irq_context: 0 (wq_completion)wg-kex-wg0#4 (work_completion)(&({ do { const void *__vpp_verify = (typeof((worker) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __asm__ ("" : "=r"(__ptr) : "0"((typeof(*((worker))) *)((worker)))); (typeof((typeof(*((worker))) *)((worker)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); })->work) &peer->endpoint_lock rcu_read_lock_bh rcu_read_lock rcu_read_lock_bh rcu_read_lock pool_lock#2 irq_context: 0 (wq_completion)wg-kex-wg0#4 (work_completion)(&({ do { const void *__vpp_verify = (typeof((worker) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __asm__ ("" : "=r"(__ptr) : "0"((typeof(*((worker))) *)((worker)))); (typeof((typeof(*((worker))) *)((worker)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); })->work) &peer->endpoint_lock rcu_read_lock_bh rcu_read_lock rcu_read_lock_bh rcu_read_lock tk_core.seq.seqcount irq_context: 0 (wq_completion)wg-kex-wg0#4 (work_completion)(&({ do { const void *__vpp_verify = (typeof((worker) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __asm__ ("" : "=r"(__ptr) : "0"((typeof(*((worker))) *)((worker)))); (typeof((typeof(*((worker))) *)((worker)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); })->work) &peer->endpoint_lock rcu_read_lock_bh rcu_read_lock rcu_read_lock_bh rcu_read_lock &obj_hash[i].lock irq_context: 0 (wq_completion)wg-crypt-wg1#2 (work_completion)(&peer->transmit_packet_work) &peer->endpoint_lock rcu_read_lock_bh rcu_read_lock rcu_read_lock_bh rcu_read_lock &c->lock irq_context: 0 pernet_ops_rwsem ovs_mutex nf_ct_proto_mutex fill_pool_map-wait-type-override &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: softirq (&app->join_timer) rcu_read_lock_bh dev->qdisc_tx_busylock ?: &qdisc_tx_busylock &meta->lock irq_context: softirq (&app->join_timer) rcu_read_lock_bh dev->qdisc_tx_busylock ?: &qdisc_tx_busylock kfence_freelist_lock irq_context: 0 &f->f_pos_lock &sb->s_type->i_mutex_key#18 &rq->__lock irq_context: 0 &f->f_pos_lock &sb->s_type->i_mutex_key#18 &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 pernet_ops_rwsem devices_rwsem rdma_nets_rwsem &device->compat_devs_mutex rcu_read_lock &rcu_state.expedited_wq &p->pi_lock &rq->__lock &cfs_rq->removed.lock irq_context: 0 pernet_ops_rwsem rtnl_mutex &caifn->caifdevs.lock rcu_state.exp_mutex rcu_read_lock &rq->__lock irq_context: 0 pernet_ops_rwsem rtnl_mutex &caifn->caifdevs.lock rcu_state.exp_mutex rcu_read_lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: softirq (&n->timer) k-slock-AF_INET6 rcu_read_lock rcu_read_lock rcu_read_lock_bh rcu_read_lock rlock-AF_PACKET irq_context: 0 pernet_ops_rwsem nf_ct_proto_mutex defrag4_mutex nf_hook_mutex fs_reclaim &rq->__lock irq_context: 0 pernet_ops_rwsem k-sk_lock-AF_RXRPC &rxnet->local_mutex mmu_notifier_invalidate_range_start &rq->__lock irq_context: 0 pernet_ops_rwsem k-sk_lock-AF_RXRPC &rxnet->local_mutex mmu_notifier_invalidate_range_start &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 rcu_read_lock_bh rcu_read_lock rcu_read_lock &dir->lock#2 irq_context: 0 rcu_read_lock_bh rcu_read_lock rcu_read_lock &ul->lock irq_context: 0 rcu_read_lock_bh rcu_read_lock crngs.lock irq_context: 0 rcu_read_lock_bh rcu_read_lock batched_entropy_u32.lock irq_context: 0 rcu_read_lock_bh rcu_read_lock rcu_read_lock &____s->seqcount#7 irq_context: 0 rcu_read_lock_bh rcu_read_lock rcu_read_lock &nf_conntrack_locks[i] irq_context: 0 rcu_read_lock_bh rcu_read_lock rcu_read_lock &nf_conntrack_locks[i] &nf_conntrack_locks[i]/1 irq_context: 0 rcu_read_lock_bh rcu_read_lock rcu_read_lock &nf_conntrack_locks[i] &nf_conntrack_locks[i]/1 batched_entropy_u8.lock irq_context: 0 rcu_read_lock_bh rcu_read_lock rcu_read_lock &obj_hash[i].lock irq_context: 0 rcu_read_lock_bh rcu_read_lock rcu_read_lock &tbl->lock irq_context: 0 rcu_read_lock_bh rcu_read_lock rcu_read_lock &n->lock irq_context: 0 rcu_read_lock_bh rcu_read_lock rcu_read_lock &____s->seqcount#9 irq_context: 0 rcu_read_lock_bh rcu_read_lock rcu_read_lock rcu_read_lock_bh noop_qdisc.q.lock irq_context: 0 rcu_read_lock_bh rcu_read_lock rcu_read_lock rcu_read_lock_bh &obj_hash[i].lock irq_context: 0 rcu_read_lock_bh rcu_read_lock rcu_read_lock rcu_read_lock_bh pool_lock#2 irq_context: 0 (wq_completion)mld (work_completion)(&(&idev->mc_ifc_work)->work) &idev->mc_lock rcu_read_lock rcu_read_lock rcu_read_lock_bh rcu_read_lock rcu_read_lock rcu_read_lock_bh noop_qdisc.q.lock irq_context: 0 (wq_completion)mld (work_completion)(&(&idev->mc_ifc_work)->work) &idev->mc_lock rcu_read_lock rcu_read_lock rcu_read_lock_bh rcu_read_lock rcu_read_lock rcu_read_lock_bh &obj_hash[i].lock irq_context: 0 (wq_completion)mld (work_completion)(&(&idev->mc_ifc_work)->work) &idev->mc_lock rcu_read_lock rcu_read_lock rcu_read_lock_bh rcu_read_lock rcu_read_lock rcu_read_lock_bh pool_lock#2 irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem k-sk_lock-AF_RXRPC rcu_node_0 irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem k-sk_lock-AF_RXRPC &rcu_state.expedited_wq irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem k-sk_lock-AF_RXRPC &rcu_state.expedited_wq &p->pi_lock irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem k-sk_lock-AF_RXRPC &rcu_state.expedited_wq &p->pi_lock &rq->__lock irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem k-sk_lock-AF_RXRPC &rcu_state.expedited_wq &p->pi_lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)ipv6_addrconf (work_completion)(&(&ifa->dad_work)->work) rtnl_mutex rcu_read_lock rcu_read_lock rcu_read_lock_bh rcu_read_lock rcu_read_lock rcu_read_lock_bh noop_qdisc.q.lock irq_context: 0 (wq_completion)ipv6_addrconf (work_completion)(&(&ifa->dad_work)->work) rtnl_mutex rcu_read_lock rcu_read_lock rcu_read_lock_bh rcu_read_lock rcu_read_lock rcu_read_lock_bh &obj_hash[i].lock irq_context: 0 (wq_completion)ipv6_addrconf (work_completion)(&(&ifa->dad_work)->work) rtnl_mutex rcu_read_lock rcu_read_lock rcu_read_lock_bh rcu_read_lock rcu_read_lock rcu_read_lock_bh pool_lock#2 irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem rtnl_mutex k-sk_lock-AF_INET irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem rtnl_mutex k-sk_lock-AF_INET k-slock-AF_INET irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem rtnl_mutex k-slock-AF_INET irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem rtnl_mutex &table->hash[i].lock irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem rtnl_mutex &table->hash[i].lock &table->hash2[i].lock irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem rtnl_mutex k-clock-AF_INET irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem rtnl_mutex &sb->s_type->i_lock_key#8 irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem rtnl_mutex &xa->xa_lock#9 irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem rtnl_mutex &fsnotify_mark_srcu irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem rtnl_mutex &net->ipv6.addrconf_hash_lock &base->lock irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem rtnl_mutex &net->ipv6.addrconf_hash_lock &base->lock &obj_hash[i].lock irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem rtnl_mutex napi_hash_lock irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem rtnl_mutex &ul->lock pool_lock#2 irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem rtnl_mutex &ul->lock &dir->lock#2 irq_context: 0 sk_lock-AF_ALG &mm->mmap_lock ptlock_ptr(page)#2 key irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem nf_conntrack_mutex &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem nf_conntrack_mutex &nf_conntrack_locks[i]/1 irq_context: 0 &sig->cred_guard_mutex &sig->exec_update_lock &mm->mmap_lock &rcu_state.expedited_wq irq_context: 0 &sig->cred_guard_mutex &sig->exec_update_lock &mm->mmap_lock &rcu_state.expedited_wq &p->pi_lock irq_context: 0 &sig->cred_guard_mutex &sig->exec_update_lock &mm->mmap_lock &rcu_state.expedited_wq &p->pi_lock &rq->__lock irq_context: 0 &sig->cred_guard_mutex &sig->exec_update_lock &mm->mmap_lock &rcu_state.expedited_wq &p->pi_lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 pernet_ops_rwsem &ht->mutex rcu_read_lock rcu_node_0 irq_context: 0 pernet_ops_rwsem &ht->mutex rcu_read_lock &rq->__lock irq_context: 0 pernet_ops_rwsem &ht->mutex rcu_read_lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 sb_writers#5 &rcu_state.expedited_wq irq_context: 0 sb_writers#5 &rcu_state.expedited_wq &p->pi_lock irq_context: 0 sb_writers#5 &rcu_state.expedited_wq &p->pi_lock &rq->__lock irq_context: 0 sb_writers#5 &rcu_state.expedited_wq &p->pi_lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)events_power_efficient (check_lifetime_work).work &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 sk_lock-AF_PACKET &rnp->exp_wq[0] irq_context: 0 sk_lock-AF_PACKET rcu_state.exp_mutex pgd_lock irq_context: 0 sk_lock-AF_PACKET rcu_state.exp_mutex rcu_read_lock pool_lock#2 irq_context: 0 sk_lock-AF_PACKET rcu_state.exp_mutex key irq_context: 0 sk_lock-AF_PACKET rcu_state.exp_mutex pcpu_lock irq_context: 0 sk_lock-AF_PACKET rcu_state.exp_mutex percpu_counters_lock irq_context: 0 sk_lock-AF_PACKET pgd_lock irq_context: 0 sk_lock-AF_PACKET rcu_read_lock pool_lock#2 irq_context: 0 sk_lock-AF_PACKET key irq_context: 0 sk_lock-AF_PACKET pcpu_lock irq_context: 0 sk_lock-AF_PACKET percpu_counters_lock irq_context: 0 nfnl_subsys_ipset &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 &pipe->mutex/1 pgd_lock irq_context: 0 &pipe->mutex/1 key irq_context: 0 &pipe->mutex/1 pcpu_lock irq_context: 0 &pipe->mutex/1 percpu_counters_lock irq_context: 0 &pipe->mutex/1 pcpu_lock stock_lock irq_context: 0 cb_lock &rdev->wiphy.mtx &local->mtx batched_entropy_u8.lock irq_context: 0 cb_lock &rdev->wiphy.mtx &local->mtx kfence_freelist_lock irq_context: 0 (wq_completion)phy10 (work_completion)(&(&local->roc_work)->work) irq_context: 0 (wq_completion)phy10 (work_completion)(&(&local->roc_work)->work) &local->mtx irq_context: 0 (wq_completion)phy10 (work_completion)(&(&local->roc_work)->work) &local->mtx &local->queue_stop_reason_lock irq_context: 0 (wq_completion)phy10 (work_completion)(&(&local->roc_work)->work) &local->mtx &local->iflist_mtx irq_context: 0 (wq_completion)phy10 (work_completion)(&(&local->roc_work)->work) &local->mtx &local->iflist_mtx hrtimer_bases.lock irq_context: 0 (wq_completion)phy10 (work_completion)(&(&local->roc_work)->work) &local->mtx &local->iflist_mtx hrtimer_bases.lock &obj_hash[i].lock irq_context: 0 (wq_completion)phy10 (work_completion)(&(&local->roc_work)->work) &local->mtx &data->mutex irq_context: 0 (wq_completion)phy10 (work_completion)(&(&local->roc_work)->work) &local->mtx &data->mutex &rq->__lock irq_context: 0 (wq_completion)phy10 (work_completion)(&(&local->roc_work)->work) &local->mtx &data->mutex &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)phy10 (work_completion)(&(&local->roc_work)->work) &local->mtx &obj_hash[i].lock irq_context: 0 (wq_completion)phy10 (work_completion)(&(&local->roc_work)->work) &local->mtx &base->lock irq_context: 0 (wq_completion)phy10 (work_completion)(&(&local->roc_work)->work) &local->mtx &base->lock &obj_hash[i].lock irq_context: 0 (wq_completion)phy10 (work_completion)(&(&local->roc_work)->work) &local->mtx rcu_read_lock &local->queue_stop_reason_lock irq_context: 0 (wq_completion)phy10 (work_completion)(&(&local->roc_work)->work) &local->mtx rcu_read_lock tk_core.seq.seqcount irq_context: 0 (wq_completion)phy10 (work_completion)(&(&local->roc_work)->work) &local->mtx rcu_read_lock hwsim_radio_lock irq_context: 0 (wq_completion)phy10 (work_completion)(&(&local->roc_work)->work) &local->mtx rcu_read_lock &list->lock#19 irq_context: 0 cb_lock &rdev->wiphy.mtx &local->mtx &base->lock irq_context: 0 cb_lock &rdev->wiphy.mtx &local->mtx &base->lock &obj_hash[i].lock irq_context: 0 pernet_ops_rwsem &root->kernfs_rwsem rcu_read_lock &rcu_state.gp_wq &p->pi_lock &rq->__lock &cfs_rq->removed.lock irq_context: 0 (wq_completion)phy10 (work_completion)(&(&local->roc_work)->work) &local->mtx fs_reclaim irq_context: 0 (wq_completion)phy10 (work_completion)(&(&local->roc_work)->work) &local->mtx fs_reclaim mmu_notifier_invalidate_range_start irq_context: 0 (wq_completion)phy10 (work_completion)(&(&local->roc_work)->work) &local->mtx &c->lock irq_context: 0 (wq_completion)phy10 (work_completion)(&(&local->roc_work)->work) &local->mtx nl_table_lock irq_context: 0 (wq_completion)phy10 (work_completion)(&(&local->roc_work)->work) &local->mtx nl_table_wait.lock irq_context: 0 (wq_completion)phy10 (work_completion)(&(&local->roc_work)->work) &local->mtx &n->list_lock irq_context: 0 (wq_completion)phy10 (work_completion)(&(&local->roc_work)->work) &local->mtx &n->list_lock &c->lock irq_context: 0 (wq_completion)phy10 (work_completion)(&(&local->roc_work)->work) &local->mtx &local->iflist_mtx tk_core.seq.seqcount irq_context: 0 (wq_completion)phy10 (work_completion)(&(&local->roc_work)->work) &local->mtx &local->iflist_mtx hrtimer_bases.lock tk_core.seq.seqcount irq_context: 0 (wq_completion)phy10 (work_completion)(&(&local->roc_work)->work) &local->mtx rcu_read_lock &fq->lock irq_context: 0 (wq_completion)phy10 (work_completion)(&(&local->roc_work)->work) &local->mtx &lock->wait_lock irq_context: 0 (wq_completion)phy10 (work_completion)(&(&local->roc_work)->work) &local->mtx &rq->__lock irq_context: 0 (wq_completion)phy10 (work_completion)(&(&local->roc_work)->work) &local->mtx &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 &f->f_pos_lock sb_writers#4 &sb->s_type->i_mutex_key#8 fs_reclaim &obj_hash[i].lock irq_context: 0 &f->f_pos_lock sb_writers#4 &sb->s_type->i_mutex_key#8 &mm->mmap_lock per_cpu_ptr(&cgroup_rstat_cpu_lock, cpu) irq_context: 0 &sb->s_type->i_mutex_key#10 rcu_read_lock pgd_lock irq_context: 0 &sb->s_type->i_mutex_key#10 rcu_read_lock stock_lock irq_context: 0 &sb->s_type->i_mutex_key#10 rcu_read_lock key irq_context: 0 &sb->s_type->i_mutex_key#10 rcu_read_lock pcpu_lock irq_context: 0 &sb->s_type->i_mutex_key#10 rcu_read_lock percpu_counters_lock irq_context: 0 &sb->s_type->i_mutex_key#10 rcu_read_lock pcpu_lock stock_lock irq_context: 0 pernet_ops_rwsem devices_rwsem rdma_nets_rwsem &device->compat_devs_mutex &zone->lock &____s->seqcount irq_context: 0 sb_writers#4 &sb->s_type->i_mutex_key#8 mapping.invalidate_lock jbd2_handle &ei->i_data_sem rcu_read_lock &rcu_state.expedited_wq irq_context: 0 sb_writers#4 &sb->s_type->i_mutex_key#8 mapping.invalidate_lock jbd2_handle &ei->i_data_sem rcu_read_lock &rcu_state.expedited_wq &p->pi_lock irq_context: 0 sb_writers#4 &sb->s_type->i_mutex_key#8 mapping.invalidate_lock jbd2_handle &ei->i_data_sem rcu_read_lock &rcu_state.expedited_wq &p->pi_lock &rq->__lock irq_context: 0 sb_writers#4 &sb->s_type->i_mutex_key#8 mapping.invalidate_lock jbd2_handle &ei->i_data_sem rcu_read_lock &rcu_state.expedited_wq &p->pi_lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)writeback (work_completion)(&(&wb->dwork)->work) &sbi->s_writepages_rwsem jbd2_handle &ei->i_data_sem &lg->lg_mutex &c->lock irq_context: 0 (wq_completion)writeback (work_completion)(&(&wb->dwork)->work) &sbi->s_writepages_rwsem jbd2_handle &ei->i_data_sem &lg->lg_mutex &mapping->private_lock irq_context: 0 (wq_completion)bat_events &rq->__lock &cfs_rq->removed.lock irq_context: 0 (wq_completion)events_long (work_completion)(&br->mcast_gc_work) krc.lock &obj_hash[i].lock irq_context: 0 (wq_completion)events_long (work_completion)(&br->mcast_gc_work) krc.lock &base->lock irq_context: 0 (wq_completion)events_long (work_completion)(&br->mcast_gc_work) krc.lock &base->lock &obj_hash[i].lock irq_context: 0 (wq_completion)phy10 (work_completion)(&(&local->roc_work)->work) &local->mtx remove_cache_srcu irq_context: 0 (wq_completion)phy10 (work_completion)(&(&local->roc_work)->work) &local->mtx remove_cache_srcu quarantine_lock irq_context: 0 (wq_completion)phy10 (work_completion)(&(&local->roc_work)->work) &local->mtx remove_cache_srcu &n->list_lock irq_context: 0 (wq_completion)phy10 (work_completion)(&(&local->roc_work)->work) &local->mtx remove_cache_srcu rcu_read_lock pool_lock#2 irq_context: 0 (wq_completion)phy10 (work_completion)(&(&local->roc_work)->work) &local->mtx remove_cache_srcu &obj_hash[i].lock irq_context: 0 (wq_completion)phy10 (work_completion)(&(&local->roc_work)->work) &local->mtx remove_cache_srcu &c->lock irq_context: 0 (wq_completion)phy10 (work_completion)(&(&local->roc_work)->work) &local->mtx remove_cache_srcu pool_lock#2 irq_context: 0 (wq_completion)phy10 (work_completion)(&(&local->roc_work)->work) &local->mtx &____s->seqcount#2 irq_context: 0 (wq_completion)phy10 (work_completion)(&(&local->roc_work)->work) &local->mtx &____s->seqcount irq_context: 0 (wq_completion)phy10 (work_completion)(&(&local->roc_work)->work) &local->mtx pool_lock#2 irq_context: 0 (wq_completion)writeback (work_completion)(&(&wb->dwork)->work) &sbi->s_writepages_rwsem jbd2_handle &____s->seqcount#2 irq_context: 0 (wq_completion)writeback (work_completion)(&(&wb->dwork)->work) &sbi->s_writepages_rwsem jbd2_handle &____s->seqcount irq_context: 0 &sb->s_type->i_mutex_key#8 fs_reclaim &rq->__lock irq_context: 0 &sb->s_type->i_mutex_key#8 fs_reclaim &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 &sb->s_type->i_mutex_key#8 fs_reclaim &cfs_rq->removed.lock irq_context: 0 &sb->s_type->i_mutex_key#8 fs_reclaim &obj_hash[i].lock irq_context: 0 pernet_ops_rwsem rtnl_mutex rcu_read_lock &rq->__lock &cfs_rq->removed.lock irq_context: 0 sk_lock-AF_INET6 &vma->vm_lock->lock lock#4 irq_context: 0 sk_lock-AF_INET6 &vma->vm_lock->lock lock#5 irq_context: 0 sk_lock-AF_INET6 &vma->vm_lock->lock mmu_notifier_invalidate_range_start irq_context: 0 sk_lock-AF_INET6 &vma->vm_lock->lock ptlock_ptr(page)#2 irq_context: 0 (wq_completion)events (work_completion)(&(&hwstats->traffic_dw)->work) &hwstats->hwsdev_list_lock rcu_node_0 irq_context: 0 (wq_completion)events free_ipc_work fill_pool_map-wait-type-override pool_lock#2 irq_context: 0 (wq_completion)events free_ipc_work fill_pool_map-wait-type-override &c->lock irq_context: 0 (wq_completion)events free_ipc_work fill_pool_map-wait-type-override &n->list_lock irq_context: 0 (wq_completion)events free_ipc_work fill_pool_map-wait-type-override &n->list_lock &c->lock irq_context: 0 (wq_completion)events free_ipc_work fill_pool_map-wait-type-override pool_lock irq_context: softirq (&lapb->t1timer) &lapb->lock &____s->seqcount#2 irq_context: 0 (wq_completion)events_power_efficient (work_completion)(&(&tbl->gc_work)->work) &tbl->lock rlock-AF_NETLINK irq_context: 0 (wq_completion)events_power_efficient (work_completion)(&(&tbl->gc_work)->work) &tbl->lock rcu_read_lock &ei->socket.wq.wait irq_context: 0 (wq_completion)events_power_efficient (work_completion)(&(&tbl->gc_work)->work) &tbl->lock rcu_read_lock &ei->socket.wq.wait &p->pi_lock irq_context: 0 (wq_completion)events_power_efficient (work_completion)(&(&tbl->gc_work)->work) &tbl->lock rcu_read_lock &ei->socket.wq.wait &p->pi_lock &rq->__lock irq_context: 0 (wq_completion)events_power_efficient (work_completion)(&(&tbl->gc_work)->work) &tbl->lock rcu_read_lock &ei->socket.wq.wait &p->pi_lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: softirq (&n->timer) &n->lock &____s->seqcount#2 irq_context: softirq (&n->timer) &n->lock &____s->seqcount irq_context: 0 (wq_completion)bat_events (work_completion)(&(&forw_packet_aggr->delayed_work)->work) &hard_iface->bat_iv.ogm_buff_mutex rcu_read_lock &pcp->lock &zone->lock &____s->seqcount irq_context: 0 lock#5 &lruvec->lru_lock irq_context: 0 lock#5 rcu_read_lock pool_lock#2 irq_context: 0 lock#5 &obj_hash[i].lock irq_context: 0 rtnl_mutex remove_cache_srcu rcu_read_lock &rq->__lock &cfs_rq->removed.lock irq_context: 0 sk_lock-AF_TIPC rcu_read_lock &rq->__lock &cfs_rq->removed.lock irq_context: 0 rtnl_mutex &tn->idrinfo->lock#3 irq_context: 0 rtnl_mutex &tn->idrinfo->lock#3 fs_reclaim irq_context: 0 rtnl_mutex &tn->idrinfo->lock#3 fs_reclaim mmu_notifier_invalidate_range_start irq_context: 0 rtnl_mutex &tn->idrinfo->lock#3 pool_lock#2 irq_context: 0 rtnl_mutex &p->tcfa_lock &(to_police(*a)->tcfp_lock) irq_context: 0 &sb->s_type->i_mutex_key#10 sk_lock-AF_TIPC tk_core.seq.seqcount irq_context: 0 &sb->s_type->i_mutex_key#10 sk_lock-AF_TIPC &list->lock#5 irq_context: 0 (wq_completion)bond8#2 (work_completion)(&(&bond->ad_work)->work) fill_pool_map-wait-type-override pool_lock#2 irq_context: 0 (wq_completion)bond8#2 (work_completion)(&(&bond->ad_work)->work) fill_pool_map-wait-type-override pool_lock irq_context: softirq (&app->join_timer) rcu_read_lock_bh dev->qdisc_tx_busylock ?: &qdisc_tx_busylock &c->lock irq_context: 0 &sb->s_type->i_mutex_key#10 sk_lock-AF_TIPC rcu_read_lock pool_lock#2 irq_context: 0 &sb->s_type->i_mutex_key#10 sk_lock-AF_TIPC &zone->lock irq_context: 0 &sb->s_type->i_mutex_key#10 sk_lock-AF_TIPC &zone->lock &____s->seqcount irq_context: 0 rtnl_mutex rcu_read_lock rcu_read_lock_bh rcu_read_lock &____s->seqcount#2 irq_context: 0 rtnl_mutex &p->tcfa_lock &(to_police(*a)->tcfp_lock) tk_core.seq.seqcount irq_context: 0 rtnl_mutex &p->tcfa_lock rcu_read_lock pool_lock#2 irq_context: 0 (wq_completion)bond8#3 irq_context: 0 (wq_completion)bond8#3 (work_completion)(&(&slave->notify_work)->work) irq_context: 0 (wq_completion)bond8#3 (work_completion)(&(&slave->notify_work)->work) &obj_hash[i].lock irq_context: 0 (wq_completion)bond8#3 (work_completion)(&(&slave->notify_work)->work) &base->lock irq_context: 0 (wq_completion)bond8#3 (work_completion)(&(&slave->notify_work)->work) &base->lock &obj_hash[i].lock irq_context: 0 (wq_completion)bond8#3 (work_completion)(&(&slave->notify_work)->work) &rq->__lock irq_context: 0 (wq_completion)bond8#3 (work_completion)(&(&slave->notify_work)->work) &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)bond8#3 (work_completion)(&(&slave->notify_work)->work) rtnl_mutex net_rwsem irq_context: 0 (wq_completion)bond8#3 (work_completion)(&(&slave->notify_work)->work) rtnl_mutex net_rwsem &list->lock#2 irq_context: 0 (wq_completion)bond8#3 (work_completion)(&(&slave->notify_work)->work) rtnl_mutex &tn->lock irq_context: 0 &sb->s_type->i_mutex_key#10 rcu_state.exp_mutex fill_pool_map-wait-type-override &____s->seqcount#2 irq_context: 0 &sb->s_type->i_mutex_key#10 rcu_state.exp_mutex fill_pool_map-wait-type-override &____s->seqcount irq_context: 0 tasklist_lock &sighand->siglock &(&sig->stats_lock)->lock &____s->seqcount#5 pidmap_lock fill_pool_map-wait-type-override &____s->seqcount#2 irq_context: 0 tasklist_lock &sighand->siglock &(&sig->stats_lock)->lock &____s->seqcount#5 pidmap_lock fill_pool_map-wait-type-override &____s->seqcount irq_context: 0 rtnl_mutex &p->tcfa_lock rcu_read_lock &c->lock irq_context: 0 sb_writers#4 &sb->s_type->i_mutex_key#8 jbd2_handle &ei->i_data_sem &ei->i_es_lock &____s->seqcount#2 irq_context: 0 sb_writers#4 &sb->s_type->i_mutex_key#8 jbd2_handle &ei->i_data_sem &ei->i_es_lock &pcp->lock &zone->lock irq_context: 0 sb_writers#4 &sb->s_type->i_mutex_key#8 jbd2_handle &ei->i_data_sem &ei->i_es_lock &pcp->lock &zone->lock &____s->seqcount irq_context: 0 sb_writers#4 &sb->s_type->i_mutex_key#8 jbd2_handle &ei->i_data_sem &ei->i_es_lock &____s->seqcount irq_context: 0 kn->active#70 &rq->__lock irq_context: 0 kn->active#70 &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 kn->active#70 fs_reclaim irq_context: 0 kn->active#70 fs_reclaim mmu_notifier_invalidate_range_start irq_context: 0 kn->active#70 stock_lock irq_context: 0 kn->active#70 &c->lock irq_context: 0 kn->active#70 &kernfs_locks->open_file_mutex[count] irq_context: 0 kn->active#70 &kernfs_locks->open_file_mutex[count] fs_reclaim irq_context: 0 kn->active#70 &kernfs_locks->open_file_mutex[count] fs_reclaim mmu_notifier_invalidate_range_start irq_context: 0 kn->active#70 &kernfs_locks->open_file_mutex[count] &c->lock irq_context: 0 &type->i_mutex_dir_key#7 &____s->seqcount#2 irq_context: 0 kn->active#71 fs_reclaim irq_context: 0 kn->active#71 fs_reclaim mmu_notifier_invalidate_range_start irq_context: 0 kn->active#71 stock_lock irq_context: 0 kn->active#71 &kernfs_locks->open_file_mutex[count] irq_context: 0 kn->active#71 &kernfs_locks->open_file_mutex[count] fs_reclaim irq_context: 0 kn->active#71 &kernfs_locks->open_file_mutex[count] fs_reclaim mmu_notifier_invalidate_range_start irq_context: 0 kn->active#71 &kernfs_locks->open_file_mutex[count] fs_reclaim mmu_notifier_invalidate_range_start &rq->__lock irq_context: 0 kn->active#71 &kernfs_locks->open_file_mutex[count] fs_reclaim mmu_notifier_invalidate_range_start &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 kn->active#71 &kernfs_locks->open_file_mutex[count] &c->lock irq_context: 0 kn->active#72 fs_reclaim irq_context: 0 kn->active#72 fs_reclaim mmu_notifier_invalidate_range_start irq_context: 0 kn->active#72 fs_reclaim mmu_notifier_invalidate_range_start &rq->__lock irq_context: 0 kn->active#72 fs_reclaim mmu_notifier_invalidate_range_start &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 kn->active#72 stock_lock irq_context: 0 kn->active#72 &kernfs_locks->open_file_mutex[count] irq_context: 0 kn->active#72 &kernfs_locks->open_file_mutex[count] fs_reclaim irq_context: 0 kn->active#72 &kernfs_locks->open_file_mutex[count] fs_reclaim mmu_notifier_invalidate_range_start irq_context: 0 sb_writers#11 &p->lock &of->mutex kn->active#71 &prev->lock#2 irq_context: 0 sb_writers#11 &p->lock &of->mutex kn->active#71 &rq->__lock irq_context: 0 (wq_completion)bond9#2 irq_context: 0 (wq_completion)bond9#2 (work_completion)(&(&slave->notify_work)->work) irq_context: 0 (wq_completion)bond9#2 (work_completion)(&(&slave->notify_work)->work) &obj_hash[i].lock irq_context: 0 (wq_completion)bond9#2 (work_completion)(&(&slave->notify_work)->work) &base->lock irq_context: 0 (wq_completion)bond9#2 (work_completion)(&(&slave->notify_work)->work) &base->lock &obj_hash[i].lock irq_context: 0 kn->active#70 &kernfs_locks->open_file_mutex[count] &____s->seqcount#2 irq_context: 0 kn->active#70 &kernfs_locks->open_file_mutex[count] &____s->seqcount irq_context: 0 kn->active#71 &c->lock irq_context: 0 kn->active#72 &c->lock irq_context: 0 kn->active#72 &rq->__lock irq_context: 0 kn->active#72 &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)bond9#2 (work_completion)(&(&slave->notify_work)->work) rtnl_mutex net_rwsem irq_context: 0 (wq_completion)bond9#2 (work_completion)(&(&slave->notify_work)->work) rtnl_mutex net_rwsem &list->lock#2 irq_context: 0 (wq_completion)bond9#2 (work_completion)(&(&slave->notify_work)->work) rtnl_mutex &tn->lock irq_context: 0 (wq_completion)bond9#2 (work_completion)(&(&slave->notify_work)->work) rtnl_mutex.wait_lock irq_context: 0 (wq_completion)bond9#2 (work_completion)(&(&slave->notify_work)->work) &p->pi_lock irq_context: 0 (wq_completion)events (work_completion)(&rdev->wiphy_work) &rdev->wiphy.mtx &wdev->mtx &local->sta_mtx rcu_state.exp_mutex &rnp->exp_wq[2] irq_context: 0 sb_writers#11 &of->mutex &rq->__lock irq_context: 0 sb_writers#11 &of->mutex &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 kn->active#71 &____s->seqcount#2 irq_context: 0 kn->active#71 &rq->__lock irq_context: 0 kn->active#71 &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 kn->active#71 &____s->seqcount irq_context: 0 kn->active#72 &n->list_lock irq_context: 0 kn->active#72 &n->list_lock &c->lock irq_context: 0 cb_lock genl_mutex rtnl_mutex team->team_lock_key#4 irq_context: 0 cb_lock genl_mutex rtnl_mutex team->team_lock_key#4 &rq->__lock irq_context: 0 cb_lock genl_mutex rtnl_mutex team->team_lock_key#4 &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 cb_lock genl_mutex rtnl_mutex team->team_lock_key#4 fs_reclaim irq_context: 0 cb_lock genl_mutex rtnl_mutex team->team_lock_key#4 fs_reclaim mmu_notifier_invalidate_range_start irq_context: 0 cb_lock genl_mutex rtnl_mutex team->team_lock_key#4 &c->lock irq_context: 0 cb_lock genl_mutex rtnl_mutex team->team_lock_key#4 pool_lock#2 irq_context: 0 cb_lock genl_mutex rtnl_mutex team->team_lock_key#4 &obj_hash[i].lock irq_context: 0 cb_lock genl_mutex rtnl_mutex team->team_lock_key#4 nl_table_lock irq_context: 0 cb_lock genl_mutex rtnl_mutex team->team_lock_key#4 rcu_read_lock pool_lock#2 irq_context: 0 cb_lock genl_mutex rtnl_mutex team->team_lock_key#4 rcu_read_lock rcu_read_lock_bh &obj_hash[i].lock irq_context: 0 cb_lock genl_mutex rtnl_mutex team->team_lock_key#4 rcu_read_lock rcu_read_lock_bh pool_lock#2 irq_context: 0 cb_lock genl_mutex rtnl_mutex team->team_lock_key#4 rlock-AF_NETLINK irq_context: 0 cb_lock genl_mutex rtnl_mutex team->team_lock_key#4 nl_table_wait.lock irq_context: 0 cb_lock genl_mutex rtnl_mutex team->team_lock_key#4 (console_sem).lock irq_context: 0 cb_lock genl_mutex rtnl_mutex team->team_lock_key#4 console_lock console_srcu console_owner_lock irq_context: 0 cb_lock genl_mutex rtnl_mutex team->team_lock_key#4 console_lock console_srcu console_owner irq_context: 0 cb_lock genl_mutex rtnl_mutex team->team_lock_key#4 console_lock console_srcu console_owner &port_lock_key irq_context: 0 kn->active#70 &n->list_lock irq_context: 0 kn->active#70 &n->list_lock &c->lock irq_context: 0 cb_lock genl_mutex rtnl_mutex team->team_lock_key#4 console_lock console_srcu console_owner console_owner_lock irq_context: 0 cb_lock genl_mutex rtnl_mutex team->team_lock_key#4 rcu_read_lock &c->lock irq_context: 0 sb_writers#4 &iint->mutex fs_reclaim mmu_notifier_invalidate_range_start &rq->__lock irq_context: 0 sb_writers#4 &iint->mutex fs_reclaim mmu_notifier_invalidate_range_start &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 tracepoints_mutex &____s->seqcount irq_context: 0 sk_lock-AF_INET rcu_read_lock &n->lock &obj_hash[i].lock pool_lock irq_context: 0 rtnl_mutex &ul->lock#2 &____s->seqcount#2 irq_context: 0 rtnl_mutex &n->lock &obj_hash[i].lock irq_context: 0 sk_lock-AF_INET rcu_read_lock &n->lock quarantine_lock irq_context: 0 sk_lock-AF_INET rcu_read_lock &n->lock fill_pool_map-wait-type-override pool_lock#2 irq_context: 0 sk_lock-AF_INET rcu_read_lock &n->lock fill_pool_map-wait-type-override &c->lock irq_context: 0 sk_lock-AF_INET rcu_read_lock &n->lock fill_pool_map-wait-type-override &n->list_lock irq_context: 0 sk_lock-AF_INET rcu_read_lock &n->lock fill_pool_map-wait-type-override &n->list_lock &c->lock irq_context: 0 sk_lock-AF_INET rcu_read_lock &n->lock fill_pool_map-wait-type-override pool_lock irq_context: 0 sk_lock-AF_INET rcu_read_lock &n->lock &meta->lock irq_context: 0 sk_lock-AF_INET rcu_read_lock &n->lock kfence_freelist_lock irq_context: softirq (&n->timer) k-slock-AF_INET6 rcu_read_lock rcu_read_lock rcu_read_lock_bh rcu_read_lock &n->list_lock irq_context: softirq (&n->timer) k-slock-AF_INET6 rcu_read_lock rcu_read_lock rcu_read_lock_bh rcu_read_lock &n->list_lock &c->lock irq_context: 0 &sctp_ep_hashtable[i].lock irq_context: 0 sb_writers#11 &type->i_mutex_dir_key#7 &root->kernfs_rwsem &rq->__lock irq_context: 0 sb_writers#11 &type->i_mutex_dir_key#7 &root->kernfs_rwsem &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 kn->active#62 &n->list_lock irq_context: 0 kn->active#62 &n->list_lock &c->lock irq_context: 0 sb_writers#4 &sbi->s_writepages_rwsem rcu_read_lock &c->lock irq_context: 0 &vma->vm_lock->lock fs_reclaim mmu_notifier_invalidate_range_start &rq->__lock &cfs_rq->removed.lock irq_context: 0 (wq_completion)bat_events (work_completion)(&(&forw_packet_aggr->delayed_work)->work) &hard_iface->bat_iv.ogm_buff_mutex rcu_read_lock fill_pool_map-wait-type-override &____s->seqcount irq_context: 0 sb_writers#4 &sb->s_type->i_mutex_key#8 jbd2_handle &journal->j_state_lock &journal->j_wait_commit &p->pi_lock &cfs_rq->removed.lock irq_context: 0 sb_writers#4 &sbi->s_writepages_rwsem rcu_read_lock &rcu_state.gp_wq irq_context: 0 sb_writers#4 &sbi->s_writepages_rwsem rcu_read_lock &rcu_state.gp_wq &p->pi_lock irq_context: 0 sb_writers#4 &sbi->s_writepages_rwsem rcu_read_lock &rcu_state.gp_wq &p->pi_lock &rq->__lock irq_context: 0 sb_writers#4 &sbi->s_writepages_rwsem rcu_read_lock &rcu_state.gp_wq &p->pi_lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)ipv6_addrconf (work_completion)(&(&ifa->dad_work)->work) rtnl_mutex &idev->mc_lock rcu_read_lock rcu_read_lock rcu_read_lock_bh &sch->q.lock hrtimer_bases.lock tk_core.seq.seqcount irq_context: 0 &vma->vm_lock->lock &obj_hash[i].lock pool_lock irq_context: 0 &f->f_pos_lock &type->i_mutex_dir_key#3 rcu_read_lock &rcu_state.gp_wq irq_context: 0 &f->f_pos_lock &type->i_mutex_dir_key#3 rcu_read_lock &rcu_state.gp_wq &p->pi_lock irq_context: 0 &f->f_pos_lock &type->i_mutex_dir_key#3 rcu_read_lock &rcu_state.gp_wq &p->pi_lock &rq->__lock irq_context: 0 &f->f_pos_lock &type->i_mutex_dir_key#3 rcu_read_lock &rcu_state.gp_wq &p->pi_lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 sb_writers#4 &type->i_mutex_dir_key#3/1 &type->i_mutex_dir_key#3 &pcp->lock &zone->lock irq_context: 0 sb_writers#4 &type->i_mutex_dir_key#3/1 &type->i_mutex_dir_key#3 &pcp->lock &zone->lock &____s->seqcount irq_context: 0 __ip_vs_mutex &cfs_rq->removed.lock irq_context: 0 __ip_vs_mutex &obj_hash[i].lock irq_context: 0 __ip_vs_mutex pool_lock#2 irq_context: 0 (wq_completion)bond9#3 irq_context: 0 (wq_completion)bond9#3 (work_completion)(&(&slave->notify_work)->work) irq_context: 0 (wq_completion)bond9#3 (work_completion)(&(&slave->notify_work)->work) &obj_hash[i].lock irq_context: 0 (wq_completion)bond9#3 (work_completion)(&(&slave->notify_work)->work) &base->lock irq_context: 0 (wq_completion)bond9#3 (work_completion)(&(&slave->notify_work)->work) &base->lock &obj_hash[i].lock irq_context: 0 (wq_completion)events_unbound (work_completion)(&map->work) rcu_read_lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)events_unbound (work_completion)(&map->work) rcu_read_lock &cfs_rq->removed.lock irq_context: 0 (wq_completion)events_unbound (work_completion)(&map->work) rcu_read_lock &obj_hash[i].lock irq_context: 0 (wq_completion)gid-cache-wq (work_completion)(&ndev_work->work) devices_rwsem rcu_read_lock &rcu_state.gp_wq irq_context: 0 (wq_completion)gid-cache-wq (work_completion)(&ndev_work->work) devices_rwsem rcu_read_lock &rcu_state.gp_wq &p->pi_lock irq_context: 0 rtnl_mutex rcu_read_lock pgd_lock irq_context: 0 rtnl_mutex rcu_read_lock stock_lock irq_context: 0 rtnl_mutex rcu_read_lock key irq_context: 0 rtnl_mutex rcu_read_lock pcpu_lock irq_context: 0 rtnl_mutex rcu_read_lock percpu_counters_lock irq_context: 0 rtnl_mutex rcu_read_lock pcpu_lock stock_lock irq_context: 0 sk_lock-AF_INET6 elock-AF_INET6 irq_context: 0 rtnl_mutex &root->kernfs_rwsem &root->kernfs_iattr_rwsem &cfs_rq->removed.lock irq_context: 0 (wq_completion)bond9#3 (work_completion)(&(&slave->notify_work)->work) rtnl_mutex net_rwsem irq_context: 0 (wq_completion)bond9#3 (work_completion)(&(&slave->notify_work)->work) rtnl_mutex net_rwsem &list->lock#2 irq_context: 0 (wq_completion)bond9#3 (work_completion)(&(&slave->notify_work)->work) rtnl_mutex &tn->lock irq_context: 0 (wq_completion)bond9#3 (work_completion)(&(&slave->notify_work)->work) rtnl_mutex.wait_lock irq_context: 0 (wq_completion)bond9#3 (work_completion)(&(&slave->notify_work)->work) &p->pi_lock irq_context: 0 rtnl_mutex &root->kernfs_rwsem &root->kernfs_iattr_rwsem &obj_hash[i].lock irq_context: 0 cb_lock genl_mutex rtnl_mutex team->team_lock_key#10 irq_context: 0 cb_lock genl_mutex rtnl_mutex team->team_lock_key#10 &rq->__lock irq_context: 0 cb_lock genl_mutex rtnl_mutex team->team_lock_key#10 &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 cb_lock genl_mutex rtnl_mutex team->team_lock_key#10 mode_list_lock irq_context: 0 cb_lock genl_mutex rtnl_mutex team->team_lock_key#10 fs_reclaim irq_context: 0 cb_lock genl_mutex rtnl_mutex team->team_lock_key#10 fs_reclaim mmu_notifier_invalidate_range_start irq_context: 0 cb_lock genl_mutex rtnl_mutex team->team_lock_key#10 &____s->seqcount irq_context: 0 cb_lock genl_mutex rtnl_mutex team->team_lock_key#10 pool_lock#2 irq_context: 0 cb_lock genl_mutex rtnl_mutex team->team_lock_key#10 pcpu_alloc_mutex irq_context: 0 cb_lock genl_mutex rtnl_mutex team->team_lock_key#10 pcpu_alloc_mutex pcpu_lock irq_context: 0 cb_lock genl_mutex rtnl_mutex team->team_lock_key#10 &obj_hash[i].lock irq_context: 0 cb_lock genl_mutex rtnl_mutex team->team_lock_key#10 &c->lock irq_context: 0 rtnl_mutex &root->kernfs_rwsem rcu_read_lock &cfs_rq->removed.lock irq_context: 0 cb_lock genl_mutex rtnl_mutex team->team_lock_key#10 &____s->seqcount#2 irq_context: 0 cb_lock genl_mutex rtnl_mutex team->team_lock_key#10 &n->list_lock irq_context: 0 cb_lock genl_mutex rtnl_mutex team->team_lock_key#10 &n->list_lock &c->lock irq_context: 0 cb_lock genl_mutex rtnl_mutex team->team_lock_key#10 nl_table_lock irq_context: 0 cb_lock genl_mutex rtnl_mutex team->team_lock_key#10 rlock-AF_NETLINK irq_context: 0 cb_lock genl_mutex rtnl_mutex team->team_lock_key#10 nl_table_wait.lock irq_context: 0 cb_lock genl_mutex rtnl_mutex team->team_lock_key#10 (console_sem).lock irq_context: 0 cb_lock genl_mutex rtnl_mutex team->team_lock_key#10 console_lock console_srcu console_owner_lock irq_context: 0 cb_lock genl_mutex rtnl_mutex team->team_lock_key#10 console_lock console_srcu console_owner irq_context: 0 cb_lock genl_mutex rtnl_mutex team->team_lock_key#10 console_lock console_srcu console_owner &port_lock_key irq_context: 0 cb_lock genl_mutex rtnl_mutex team->team_lock_key#10 console_lock console_srcu console_owner console_owner_lock irq_context: 0 (wq_completion)wg-crypt-wg2#4 (work_completion)(&({ do { const void *__vpp_verify = (typeof((worker) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __asm__ ("" : "=r"(__ptr) : "0"((typeof(*((worker))) *)((worker)))); (typeof((typeof(*((worker))) *)((worker)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); })->work) &n->list_lock irq_context: 0 (wq_completion)wg-crypt-wg2#4 (work_completion)(&({ do { const void *__vpp_verify = (typeof((worker) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __asm__ ("" : "=r"(__ptr) : "0"((typeof(*((worker))) *)((worker)))); (typeof((typeof(*((worker))) *)((worker)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); })->work) &n->list_lock &c->lock irq_context: 0 (wq_completion)bond6 (work_completion)(&(&bond->mii_work)->work) &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)ipv6_addrconf (work_completion)(&(&ifa->dad_work)->work) rtnl_mutex rcu_read_lock rcu_read_lock rcu_read_lock_bh rcu_read_lock &meta->lock irq_context: 0 (wq_completion)ipv6_addrconf (work_completion)(&(&ifa->dad_work)->work) rtnl_mutex rcu_read_lock rcu_read_lock rcu_read_lock_bh rcu_read_lock kfence_freelist_lock irq_context: 0 pernet_ops_rwsem k-sk_lock-AF_RXRPC &rxnet->local_mutex batched_entropy_u8.lock irq_context: 0 pernet_ops_rwsem k-sk_lock-AF_RXRPC &rxnet->local_mutex kfence_freelist_lock irq_context: softirq (&timer) rcu_read_lock init_task.mems_allowed_seq.seqcount irq_context: 0 pernet_ops_rwsem batched_entropy_u8.lock crngs.lock irq_context: 0 (wq_completion)bat_events (work_completion)(&(&bat_priv->mcast.work)->work) &rcu_state.expedited_wq &p->pi_lock &rq->__lock irq_context: 0 sk_lock-AF_INET6 &sighand->siglock &c->lock irq_context: 0 sb_writers#4 &type->i_mutex_dir_key#3/1 &type->i_mutex_dir_key#3 jbd2_handle rcu_read_lock &cfs_rq->removed.lock irq_context: 0 sb_writers#4 &type->i_mutex_dir_key#3/1 &type->i_mutex_dir_key#3 jbd2_handle rcu_read_lock &obj_hash[i].lock irq_context: 0 cb_lock genl_mutex rtnl_mutex team->team_lock_key#11 irq_context: 0 cb_lock genl_mutex rtnl_mutex team->team_lock_key#11 mode_list_lock irq_context: 0 cb_lock genl_mutex rtnl_mutex team->team_lock_key#11 fs_reclaim irq_context: 0 cb_lock genl_mutex rtnl_mutex team->team_lock_key#11 fs_reclaim mmu_notifier_invalidate_range_start irq_context: 0 cb_lock genl_mutex rtnl_mutex team->team_lock_key#11 fs_reclaim mmu_notifier_invalidate_range_start &rq->__lock irq_context: 0 cb_lock genl_mutex rtnl_mutex team->team_lock_key#11 fs_reclaim mmu_notifier_invalidate_range_start &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 cb_lock genl_mutex rtnl_mutex team->team_lock_key#11 &____s->seqcount irq_context: 0 cb_lock genl_mutex rtnl_mutex team->team_lock_key#11 pcpu_alloc_mutex irq_context: 0 cb_lock genl_mutex rtnl_mutex team->team_lock_key#11 pcpu_alloc_mutex pcpu_lock irq_context: 0 cb_lock genl_mutex rtnl_mutex team->team_lock_key#11 &obj_hash[i].lock irq_context: 0 cb_lock genl_mutex rtnl_mutex team->team_lock_key#11 &c->lock irq_context: 0 cb_lock genl_mutex rtnl_mutex team->team_lock_key#11 &n->list_lock irq_context: 0 cb_lock genl_mutex rtnl_mutex team->team_lock_key#11 &n->list_lock &c->lock irq_context: 0 cb_lock genl_mutex rtnl_mutex team->team_lock_key#11 &rq->__lock irq_context: 0 cb_lock genl_mutex rtnl_mutex team->team_lock_key#11 &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 cb_lock genl_mutex rtnl_mutex team->team_lock_key#11 rcu_node_0 irq_context: 0 cb_lock genl_mutex rtnl_mutex team->team_lock_key#11 rcu_read_lock rcu_node_0 irq_context: 0 cb_lock genl_mutex rtnl_mutex team->team_lock_key#11 rcu_read_lock &rq->__lock irq_context: 0 cb_lock genl_mutex rtnl_mutex team->team_lock_key#11 rcu_read_lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 cb_lock genl_mutex rtnl_mutex team->team_lock_key#11 &____s->seqcount#2 irq_context: 0 cb_lock genl_mutex rtnl_mutex team->team_lock_key#11 nl_table_lock irq_context: 0 cb_lock genl_mutex rtnl_mutex team->team_lock_key#11 rlock-AF_NETLINK irq_context: 0 cb_lock genl_mutex rtnl_mutex team->team_lock_key#11 nl_table_wait.lock irq_context: 0 cb_lock genl_mutex rtnl_mutex team->team_lock_key#11 (console_sem).lock irq_context: 0 cb_lock genl_mutex rtnl_mutex team->team_lock_key#11 console_lock console_srcu console_owner_lock irq_context: 0 cb_lock genl_mutex rtnl_mutex team->team_lock_key#11 console_lock console_srcu console_owner irq_context: 0 cb_lock genl_mutex rtnl_mutex team->team_lock_key#11 console_lock console_srcu console_owner &port_lock_key irq_context: 0 cb_lock genl_mutex rtnl_mutex team->team_lock_key#11 console_lock console_srcu console_owner console_owner_lock irq_context: 0 &f->f_pos_lock sb_writers#4 &sb->s_type->i_mutex_key#8 rcu_read_lock fill_pool_map-wait-type-override &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 &f->f_pos_lock sb_writers#4 &sb->s_type->i_mutex_key#8 &sbi->s_writepages_rwsem jbd2_handle &ei->i_data_sem &ret->b_state_lock irq_context: 0 &f->f_pos_lock sb_writers#4 &sb->s_type->i_mutex_key#8 &sbi->s_writepages_rwsem jbd2_handle &ei->i_data_sem &ret->b_state_lock &journal->j_list_lock irq_context: 0 &f->f_pos_lock sb_writers#4 &sb->s_type->i_mutex_key#8 &sbi->s_writepages_rwsem jbd2_handle &journal->j_list_lock irq_context: 0 &pool->lock/1 &x->wait#10 &p->pi_lock &rq->__lock &cfs_rq->removed.lock irq_context: 0 dup_mmap_sem &mm->mmap_lock &mm->mmap_lock/1 &anon_vma->rwsem per_cpu_ptr(&cgroup_rstat_cpu_lock, cpu) irq_context: 0 &f->f_pos_lock sb_writers#4 &sb->s_type->i_mutex_key#8 mapping.invalidate_lock jbd2_handle &ei->i_data_sem &journal->j_revoke_lock irq_context: 0 &type->s_umount_key#23/1 list_lrus_mutex &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 &f->f_pos_lock sb_writers#4 &sb->s_type->i_mutex_key#8 &sbi->s_writepages_rwsem jbd2_handle &ei->i_data_sem &ei->i_es_lock &sbi->s_es_lock irq_context: 0 &f->f_pos_lock sb_writers#4 &sb->s_type->i_mutex_key#8 &sbi->s_writepages_rwsem &base->lock irq_context: 0 &f->f_pos_lock sb_writers#4 &sb->s_type->i_mutex_key#8 &sbi->s_writepages_rwsem &base->lock &obj_hash[i].lock irq_context: softirq &(&bond->alb_work)->timer rcu_read_lock &pool->lock/1 &p->pi_lock &rq->__lock &cfs_rq->removed.lock irq_context: 0 stock_lock per_cpu_ptr(&cgroup_rstat_cpu_lock, cpu) irq_context: 0 (wq_completion)gid-cache-wq (work_completion)(&ndev_work->work) quarantine_lock irq_context: 0 rtnl_mutex sk_lock-AF_UNSPEC rcu_read_lock &pool->lock irq_context: 0 rtnl_mutex sk_lock-AF_UNSPEC rcu_read_lock &pool->lock &obj_hash[i].lock irq_context: 0 rtnl_mutex sk_lock-AF_UNSPEC rcu_read_lock &pool->lock pool_lock#2 irq_context: 0 rtnl_mutex sk_lock-AF_UNSPEC rcu_read_lock &pool->lock &p->pi_lock irq_context: 0 rtnl_mutex sk_lock-AF_UNSPEC rcu_read_lock &pool->lock &p->pi_lock &rq->__lock irq_context: 0 rtnl_mutex sk_lock-AF_UNSPEC rcu_read_lock &pool->lock &p->pi_lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 rtnl_mutex sk_lock-AF_UNSPEC &rq->__lock irq_context: 0 rtnl_mutex sk_lock-AF_UNSPEC &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 rtnl_mutex sk_lock-AF_UNSPEC free_vmap_area_lock &obj_hash[i].lock irq_context: 0 rtnl_mutex sk_lock-AF_UNSPEC free_vmap_area_lock pool_lock#2 irq_context: 0 rtnl_mutex sk_lock-AF_UNSPEC &____s->seqcount#2 irq_context: 0 rtnl_mutex sk_lock-AF_UNSPEC rcu_read_lock rcu_node_0 irq_context: 0 rtnl_mutex sk_lock-AF_UNSPEC rcu_read_lock &rq->__lock irq_context: 0 rtnl_mutex sk_lock-AF_UNSPEC rcu_read_lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)wg-kex-wg0#7 (work_completion)(&peer->transmit_handshake_work) &peer->endpoint_lock rcu_read_lock_bh batched_entropy_u8.lock irq_context: 0 &pipe->mutex/1 nfnl_subsys_ctnetlink (console_sem).lock irq_context: 0 &pipe->mutex/1 nfnl_subsys_ctnetlink console_lock console_srcu console_owner_lock irq_context: 0 &pipe->mutex/1 nfnl_subsys_ctnetlink console_lock console_srcu console_owner irq_context: 0 &pipe->mutex/1 nfnl_subsys_ctnetlink console_lock console_srcu console_owner &port_lock_key irq_context: 0 &pipe->mutex/1 nfnl_subsys_ctnetlink console_lock console_srcu console_owner console_owner_lock irq_context: 0 &pipe->mutex/1 rcu_read_lock pgd_lock irq_context: 0 &pipe->mutex/1 rcu_read_lock key irq_context: 0 &pipe->mutex/1 rcu_read_lock pcpu_lock irq_context: 0 &pipe->mutex/1 rcu_read_lock percpu_counters_lock irq_context: 0 &pipe->mutex/1 rcu_read_lock &rcu_state.expedited_wq &p->pi_lock &rq->__lock &cfs_rq->removed.lock irq_context: softirq (&hc->tx_rtotimer) stock_lock irq_context: 0 &pipe->mutex/1 remove_cache_srcu pool_lock#2 irq_context: 0 &sb->s_type->i_mutex_key#10 (netlink_chain).rwsem &nft_net->commit_mutex &meta->lock irq_context: 0 &sb->s_type->i_mutex_key#10 (netlink_chain).rwsem &nft_net->commit_mutex kfence_freelist_lock irq_context: softirq rcu_read_lock rcu_read_lock slock-AF_INET6/1 sk_lock-AF_INET6 elock-AF_INET6 irq_context: softirq rcu_read_lock rcu_read_lock stock_lock irq_context: 0 sb_writers#4 &sb->s_type->i_mutex_key#8 mapping.invalidate_lock remove_cache_srcu rcu_node_0 irq_context: 0 &iint->mutex remove_cache_srcu rcu_read_lock rcu_node_0 irq_context: 0 &iint->mutex remove_cache_srcu rcu_read_lock &rq->__lock irq_context: 0 &iint->mutex remove_cache_srcu rcu_read_lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)bat_events (work_completion)(&(&bat_priv->dat.work)->work) &obj_hash[i].lock pool_lock irq_context: 0 sb_writers#4 &sb->s_type->i_mutex_key#8 ima_extend_list_mutex rcu_read_lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 &pipe->mutex/1 rcu_read_lock console_owner_lock irq_context: 0 &pipe->mutex/1 rcu_read_lock console_owner irq_context: 0 &sb->s_type->i_mutex_key#10 slock-AF_INET6 &dccp_hashinfo.bhash[i].lock stock_lock rcu_read_lock per_cpu_ptr(&cgroup_rstat_cpu_lock, cpu) irq_context: 0 pernet_ops_rwsem gdp_mutex &rq->__lock irq_context: softirq (&dp->dccps_xmit_timer) irq_context: softirq (&dp->dccps_xmit_timer) slock-AF_INET6 irq_context: softirq (&dp->dccps_xmit_timer) slock-AF_INET6 wlock-AF_INET6 irq_context: softirq (&dp->dccps_xmit_timer) slock-AF_INET6 pool_lock#2 irq_context: softirq (&dp->dccps_xmit_timer) slock-AF_INET6 rcu_read_lock rcu_read_lock &____s->seqcount#7 irq_context: softirq (&dp->dccps_xmit_timer) slock-AF_INET6 rcu_read_lock rcu_read_lock &ct->lock irq_context: softirq (&dp->dccps_xmit_timer) slock-AF_INET6 rcu_read_lock rcu_read_lock &____s->seqcount#10 irq_context: softirq (&dp->dccps_xmit_timer) slock-AF_INET6 rcu_read_lock rcu_read_lock rcu_read_lock_bh tk_core.seq.seqcount irq_context: softirq (&dp->dccps_xmit_timer) slock-AF_INET6 rcu_read_lock rcu_read_lock rcu_read_lock_bh &list->lock#5 irq_context: softirq (&dp->dccps_xmit_timer) slock-AF_INET6 &obj_hash[i].lock irq_context: softirq (&dp->dccps_xmit_timer) slock-AF_INET6 &base->lock irq_context: softirq (&dp->dccps_xmit_timer) slock-AF_INET6 &base->lock &obj_hash[i].lock irq_context: softirq (&hc->tx_rtotimer) &obj_hash[i].lock pool_lock irq_context: 0 sb_writers#4 &type->i_mutex_dir_key#3/1 sb_internal jbd2_handle &journal->j_wait_updates &p->pi_lock &cfs_rq->removed.lock irq_context: 0 (wq_completion)ext4-rsv-conversion (work_completion)(&ei->i_rsv_conversion_work) &____s->seqcount#2 irq_context: 0 (wq_completion)ext4-rsv-conversion (work_completion)(&ei->i_rsv_conversion_work) &____s->seqcount irq_context: 0 &pipe->mutex/1 sk_lock-AF_NETLINK rcu_read_lock rcu_node_0 irq_context: 0 &pipe->mutex/1 sk_lock-AF_NETLINK rcu_read_lock &rq->__lock irq_context: 0 &pipe->mutex/1 sk_lock-AF_NETLINK rcu_read_lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 &pipe->mutex/1 sk_lock-AF_NETLINK rcu_read_lock &cfs_rq->removed.lock irq_context: 0 &pipe->mutex/1 sk_lock-AF_NETLINK rcu_read_lock &rcu_state.gp_wq irq_context: 0 &pipe->mutex/1 sk_lock-AF_NETLINK rcu_read_lock &rcu_state.gp_wq &p->pi_lock irq_context: 0 &pipe->mutex/1 sk_lock-AF_NETLINK rcu_read_lock &rcu_state.gp_wq &p->pi_lock &rq->__lock irq_context: 0 &pipe->mutex/1 sk_lock-AF_NETLINK rcu_read_lock &rcu_state.gp_wq &p->pi_lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 sb_writers#4 &sb->s_type->i_mutex_key#8 &sbi->s_writepages_rwsem &journal->j_state_lock &journal->j_wait_transaction_locked irq_context: 0 sb_writers#4 &sb->s_type->i_mutex_key#8 &sbi->s_writepages_rwsem jbd2_handle irq_context: 0 &pipe->mutex/1 &sighand->siglock &p->pi_lock &rq->__lock &cfs_rq->removed.lock irq_context: 0 sb_writers#4 &type->i_mutex_dir_key#3/1 sb_internal jbd2_handle rcu_read_lock &cfs_rq->removed.lock irq_context: 0 sb_writers#4 &type->i_mutex_dir_key#3/1 sb_internal jbd2_handle rcu_read_lock &obj_hash[i].lock irq_context: 0 nl_table_lock &n->list_lock irq_context: 0 nl_table_lock &n->list_lock &c->lock irq_context: 0 kn->active#60 &kernfs_locks->open_file_mutex[count] fs_reclaim &rq->__lock irq_context: 0 kn->active#60 &kernfs_locks->open_file_mutex[count] fs_reclaim &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 sb_writers#4 sb_writers#4 remove_cache_srcu &pcp->lock &zone->lock irq_context: 0 sb_writers#4 sb_writers#4 remove_cache_srcu &pcp->lock &zone->lock &____s->seqcount irq_context: 0 cb_lock genl_mutex &nbd->config_lock uevent_sock_mutex rcu_read_lock &ei->socket.wq.wait &ep->lock &ep->wq &p->pi_lock &cfs_rq->removed.lock irq_context: 0 &disk->open_mutex &nbd->config_lock &obj_hash[i].lock pool_lock irq_context: 0 cb_lock genl_mutex &nbd->config_lock uevent_sock_mutex &n->list_lock irq_context: 0 cb_lock genl_mutex &nbd->config_lock uevent_sock_mutex &n->list_lock &c->lock irq_context: 0 &disk->open_mutex nbd_index_mutex nbd_index_mutex.wait_lock irq_context: 0 &disk->open_mutex nbd_index_mutex &rq->__lock irq_context: 0 &disk->open_mutex nbd_index_mutex &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 &disk->open_mutex nbd_index_mutex rcu_read_lock &rq->__lock irq_context: 0 &disk->open_mutex nbd_index_mutex rcu_read_lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 cb_lock genl_mutex nbd_index_mutex.wait_lock irq_context: 0 cb_lock genl_mutex &nbd->config_lock &lock->wait_lock irq_context: 0 &disk->open_mutex &lock->wait_lock irq_context: 0 sb_writers#4 jbd2_handle &ret->b_state_lock bit_wait_table + i irq_context: 0 &disk->open_mutex nbd_index_mutex &nbd->config_lock &n->list_lock irq_context: 0 &disk->open_mutex nbd_index_mutex &nbd->config_lock &n->list_lock &c->lock irq_context: 0 cb_lock genl_mutex &nbd->config_lock &____s->seqcount#2 irq_context: 0 cb_lock genl_mutex &nbd->config_lock &____s->seqcount irq_context: 0 &disk->open_mutex nbd_index_mutex &nbd->config_lock &____s->seqcount#2 irq_context: 0 &disk->open_mutex nbd_index_mutex &nbd->config_lock &____s->seqcount irq_context: 0 sb_writers#3 tomoyo_ss remove_cache_srcu irq_context: 0 sb_writers#3 tomoyo_ss remove_cache_srcu quarantine_lock irq_context: 0 sb_writers#3 tomoyo_ss remove_cache_srcu &c->lock irq_context: 0 sb_writers#3 tomoyo_ss remove_cache_srcu &n->list_lock irq_context: 0 sb_writers#3 tomoyo_ss remove_cache_srcu &obj_hash[i].lock irq_context: 0 sb_writers#3 tomoyo_ss remove_cache_srcu &rq->__lock irq_context: 0 sb_writers#3 tomoyo_ss remove_cache_srcu &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 &disk->open_mutex nbd_index_mutex &nbd->config_lock remove_cache_srcu irq_context: 0 &disk->open_mutex nbd_index_mutex &nbd->config_lock remove_cache_srcu quarantine_lock irq_context: 0 &disk->open_mutex nbd_index_mutex &nbd->config_lock remove_cache_srcu &c->lock irq_context: 0 &disk->open_mutex nbd_index_mutex &nbd->config_lock remove_cache_srcu &n->list_lock irq_context: 0 &disk->open_mutex nbd_index_mutex &nbd->config_lock remove_cache_srcu rcu_read_lock pool_lock#2 irq_context: 0 &disk->open_mutex nbd_index_mutex &nbd->config_lock remove_cache_srcu &obj_hash[i].lock irq_context: 0 &f->f_pos_lock sb_writers#4 &sb->s_type->i_mutex_key#8 &mm->mmap_lock &meta->lock irq_context: 0 sb_writers#4 &sbi->s_writepages_rwsem rcu_read_lock rcu_read_lock rcu_node_0 irq_context: 0 sb_writers#4 &sbi->s_writepages_rwsem rcu_read_lock rcu_read_lock &rq->__lock irq_context: 0 sb_writers#4 &sbi->s_writepages_rwsem rcu_read_lock rcu_read_lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 sb_writers#4 &sb->s_type->i_mutex_key#8 rcu_read_lock &xa->xa_lock#9 rcu_read_lock per_cpu_ptr(&cgroup_rstat_cpu_lock, cpu) irq_context: 0 &sb->s_type->i_mutex_key#10 sk_lock-AF_INET6 slock-AF_INET6 &n->list_lock irq_context: 0 &sb->s_type->i_mutex_key#10 sk_lock-AF_INET6 slock-AF_INET6 &n->list_lock &c->lock irq_context: 0 &mm->mmap_lock &vma->vm_lock->lock &sem->wait_lock irq_context: 0 rcu_read_lock &sem->wait_lock irq_context: softirq rcu_read_lock rcu_read_lock &dreq->dreq_lock slock-AF_INET6 &dccp_hashinfo.bhash[i].lock &dccp_hashinfo.bhash2[i].lock &c->lock irq_context: softirq rcu_read_lock rcu_read_lock &dreq->dreq_lock slock-AF_INET6 &dccp_hashinfo.bhash[i].lock &dccp_hashinfo.bhash2[i].lock &____s->seqcount#2 irq_context: softirq rcu_read_lock rcu_read_lock &dreq->dreq_lock slock-AF_INET6 &dccp_hashinfo.bhash[i].lock &dccp_hashinfo.bhash2[i].lock &____s->seqcount irq_context: 0 kn->active#60 &kernfs_locks->open_file_mutex[count] rcu_read_lock rcu_node_0 irq_context: 0 kn->active#60 &kernfs_locks->open_file_mutex[count] rcu_read_lock &rq->__lock irq_context: 0 kn->active#60 &kernfs_locks->open_file_mutex[count] rcu_read_lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 sk_lock-AF_INET6 rcu_read_lock rcu_read_lock rcu_read_lock_bh &sch->q.lock irq_context: 0 sk_lock-AF_INET6 rcu_read_lock rcu_read_lock rcu_read_lock_bh &sch->q.lock &c->lock irq_context: 0 sk_lock-AF_INET6 rcu_read_lock rcu_read_lock rcu_read_lock_bh &sch->q.lock pool_lock#2 irq_context: 0 sk_lock-AF_INET6 rcu_read_lock rcu_read_lock rcu_read_lock_bh &sch->q.lock tk_core.seq.seqcount irq_context: softirq rcu_read_lock rcu_read_lock slock-AF_INET6/1 sk_lock-AF_INET6 rcu_read_lock rcu_read_lock rcu_read_lock rcu_read_lock_bh &sch->q.lock irq_context: softirq rcu_read_lock rcu_read_lock slock-AF_INET6/1 sk_lock-AF_INET6 rcu_read_lock rcu_read_lock rcu_read_lock rcu_read_lock_bh &sch->q.lock pool_lock#2 irq_context: softirq rcu_read_lock rcu_read_lock slock-AF_INET6/1 sk_lock-AF_INET6 rcu_read_lock rcu_read_lock rcu_read_lock rcu_read_lock_bh &sch->q.lock tk_core.seq.seqcount irq_context: softirq rcu_read_lock rcu_read_lock slock-AF_INET6/1 sk_lock-AF_INET6 rcu_read_lock rcu_read_lock rcu_read_lock rcu_read_lock_bh &sch->q.lock hrtimer_bases.lock irq_context: softirq rcu_read_lock rcu_read_lock slock-AF_INET6/1 sk_lock-AF_INET6 rcu_read_lock rcu_read_lock rcu_read_lock rcu_read_lock_bh &sch->q.lock hrtimer_bases.lock &obj_hash[i].lock irq_context: softirq rcu_read_lock rcu_read_lock slock-AF_INET6/1 sk_lock-AF_INET6 rcu_read_lock rcu_read_lock rcu_read_lock rcu_read_lock_bh &sch->q.lock hrtimer_bases.lock tk_core.seq.seqcount irq_context: 0 sk_lock-AF_INET6 rcu_read_lock rcu_read_lock rcu_read_lock_bh &sch->q.lock rcu_read_lock &ei->socket.wq.wait irq_context: softirq rcu_read_lock rcu_read_lock slock-AF_INET6/1 sk_lock-AF_INET6 tk_core.seq.seqcount irq_context: 0 sk_lock-AF_INET6 rcu_read_lock rcu_read_lock rcu_read_lock_bh &sch->q.lock hrtimer_bases.lock irq_context: 0 sk_lock-AF_INET6 rcu_read_lock rcu_read_lock rcu_read_lock_bh &sch->q.lock hrtimer_bases.lock &obj_hash[i].lock irq_context: softirq rcu_read_lock rcu_read_lock slock-AF_INET6/1 sk_lock-AF_INET6 rcu_read_lock rcu_read_lock rcu_read_lock_bh &sch->q.lock irq_context: softirq rcu_read_lock rcu_read_lock slock-AF_INET6/1 sk_lock-AF_INET6 rcu_read_lock rcu_read_lock rcu_read_lock_bh &sch->q.lock &c->lock irq_context: softirq rcu_read_lock rcu_read_lock slock-AF_INET6/1 sk_lock-AF_INET6 rcu_read_lock rcu_read_lock rcu_read_lock_bh &sch->q.lock pool_lock#2 irq_context: softirq rcu_read_lock rcu_read_lock slock-AF_INET6/1 sk_lock-AF_INET6 rcu_read_lock rcu_read_lock rcu_read_lock_bh &sch->q.lock tk_core.seq.seqcount irq_context: softirq rcu_read_lock rcu_read_lock slock-AF_INET6/1 sk_lock-AF_INET6 rcu_read_lock rcu_read_lock rcu_read_lock_bh &sch->q.lock hrtimer_bases.lock irq_context: softirq rcu_read_lock rcu_read_lock slock-AF_INET6/1 sk_lock-AF_INET6 rcu_read_lock rcu_read_lock rcu_read_lock_bh &sch->q.lock hrtimer_bases.lock &obj_hash[i].lock irq_context: 0 &sb->s_type->i_mutex_key#10 sk_lock-AF_INET6 slock-AF_INET6 rcu_read_lock rcu_read_lock rcu_read_lock_bh &sch->q.lock irq_context: 0 &sb->s_type->i_mutex_key#10 sk_lock-AF_INET6 slock-AF_INET6 rcu_read_lock rcu_read_lock rcu_read_lock_bh &sch->q.lock pool_lock#2 irq_context: 0 &sb->s_type->i_mutex_key#10 sk_lock-AF_INET6 slock-AF_INET6 rcu_read_lock rcu_read_lock rcu_read_lock_bh &sch->q.lock tk_core.seq.seqcount irq_context: 0 &sb->s_type->i_mutex_key#10 sk_lock-AF_INET6 slock-AF_INET6 rcu_read_lock rcu_read_lock rcu_read_lock_bh &sch->q.lock hrtimer_bases.lock irq_context: 0 &sb->s_type->i_mutex_key#10 sk_lock-AF_INET6 slock-AF_INET6 rcu_read_lock rcu_read_lock rcu_read_lock_bh &sch->q.lock hrtimer_bases.lock &obj_hash[i].lock irq_context: 0 &sb->s_type->i_mutex_key#10 sk_lock-AF_INET6 slock-AF_INET6 rcu_read_lock rcu_read_lock rcu_read_lock_bh &sch->q.lock hrtimer_bases.lock tk_core.seq.seqcount irq_context: softirq (&dp->dccps_xmit_timer) slock-AF_INET6 rcu_read_lock rcu_read_lock rcu_read_lock_bh &sch->q.lock irq_context: softirq (&dp->dccps_xmit_timer) slock-AF_INET6 rcu_read_lock rcu_read_lock rcu_read_lock_bh &sch->q.lock pool_lock#2 irq_context: softirq (&dp->dccps_xmit_timer) slock-AF_INET6 rcu_read_lock rcu_read_lock rcu_read_lock_bh &sch->q.lock tk_core.seq.seqcount irq_context: softirq rcu_read_lock rcu_read_lock rcu_read_lock rcu_read_lock_bh &sch->q.lock irq_context: softirq rcu_read_lock rcu_read_lock rcu_read_lock rcu_read_lock_bh &sch->q.lock pool_lock#2 irq_context: softirq rcu_read_lock rcu_read_lock rcu_read_lock rcu_read_lock_bh &sch->q.lock tk_core.seq.seqcount irq_context: softirq rcu_read_lock rcu_read_lock rcu_read_lock rcu_read_lock_bh &sch->q.lock hrtimer_bases.lock irq_context: softirq rcu_read_lock rcu_read_lock rcu_read_lock rcu_read_lock_bh &sch->q.lock hrtimer_bases.lock &obj_hash[i].lock irq_context: softirq rcu_read_lock rcu_read_lock rcu_read_lock rcu_read_lock_bh &sch->q.lock hrtimer_bases.lock tk_core.seq.seqcount irq_context: 0 cb_lock genl_mutex &nbd->config_lock uevent_sock_mutex &____s->seqcount irq_context: 0 cb_lock genl_mutex &nbd->config_lock uevent_sock_mutex rcu_read_lock pool_lock#2 irq_context: 0 &disk->open_mutex nbd_index_mutex &rq->__lock &cfs_rq->removed.lock irq_context: 0 &disk->open_mutex &p->pi_lock &cfs_rq->removed.lock irq_context: 0 &disk->open_mutex &p->pi_lock &rq->__lock &cfs_rq->removed.lock irq_context: 0 sb_writers#3 &sb->s_type->i_mutex_key#9 tomoyo_ss rcu_read_lock rcu_node_0 irq_context: 0 sb_writers#3 &sb->s_type->i_mutex_key#9 tomoyo_ss rcu_read_lock &rq->__lock irq_context: 0 &disk->open_mutex &nbd->config_lock pgd_lock irq_context: 0 &disk->open_mutex &nbd->config_lock stock_lock irq_context: 0 &disk->open_mutex &nbd->config_lock rcu_read_lock pool_lock#2 irq_context: 0 &disk->open_mutex &nbd->config_lock key irq_context: 0 &disk->open_mutex &nbd->config_lock pcpu_lock irq_context: 0 &disk->open_mutex &nbd->config_lock percpu_counters_lock irq_context: 0 &disk->open_mutex &nbd->config_lock pcpu_lock stock_lock irq_context: 0 &disk->open_mutex &nbd->config_lock &cfs_rq->removed.lock irq_context: 0 pernet_ops_rwsem rtnl_mutex &root->kernfs_rwsem &root->kernfs_iattr_rwsem &cfs_rq->removed.lock irq_context: 0 pernet_ops_rwsem rtnl_mutex &root->kernfs_rwsem &root->kernfs_iattr_rwsem &obj_hash[i].lock irq_context: 0 cgroup_mutex &mm->mmap_lock irq_context: softirq (&ndev->rs_timer) rcu_read_lock rcu_read_lock rcu_read_lock_bh rcu_read_lock rcu_read_lock rcu_read_lock rcu_read_lock &n->list_lock irq_context: softirq (&ndev->rs_timer) rcu_read_lock rcu_read_lock rcu_read_lock_bh rcu_read_lock rcu_read_lock rcu_read_lock rcu_read_lock &n->list_lock &c->lock irq_context: 0 cb_lock genl_mutex net_dm_mutex irq_context: 0 cb_lock genl_mutex net_dm_mutex &rq->__lock irq_context: 0 cb_lock genl_mutex net_dm_mutex &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 cb_lock genl_mutex net_dm_mutex &obj_hash[i].lock irq_context: 0 cb_lock genl_mutex net_dm_mutex fs_reclaim irq_context: 0 cb_lock genl_mutex net_dm_mutex fs_reclaim mmu_notifier_invalidate_range_start irq_context: 0 cb_lock genl_mutex net_dm_mutex &c->lock irq_context: 0 cb_lock genl_mutex net_dm_mutex pool_lock#2 irq_context: 0 cb_lock genl_mutex net_dm_mutex &data->lock irq_context: 0 cb_lock genl_mutex net_dm_mutex tracepoints_mutex irq_context: 0 cb_lock genl_mutex net_dm_mutex tracepoints_mutex &rq->__lock irq_context: 0 cb_lock genl_mutex net_dm_mutex tracepoints_mutex &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 cb_lock genl_mutex net_dm_mutex tracepoints_mutex fs_reclaim irq_context: 0 cb_lock genl_mutex net_dm_mutex tracepoints_mutex fs_reclaim &rq->__lock irq_context: 0 cb_lock genl_mutex net_dm_mutex tracepoints_mutex fs_reclaim &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 cb_lock genl_mutex net_dm_mutex tracepoints_mutex fs_reclaim mmu_notifier_invalidate_range_start irq_context: 0 cb_lock genl_mutex net_dm_mutex tracepoints_mutex pool_lock#2 irq_context: 0 cb_lock genl_mutex net_dm_mutex tracepoints_mutex cpu_hotplug_lock irq_context: 0 cb_lock genl_mutex net_dm_mutex tracepoints_mutex cpu_hotplug_lock static_call_mutex irq_context: 0 cb_lock genl_mutex net_dm_mutex tracepoints_mutex cpu_hotplug_lock static_call_mutex text_mutex irq_context: 0 cb_lock genl_mutex net_dm_mutex tracepoints_mutex cpu_hotplug_lock static_call_mutex text_mutex ptlock_ptr(page)#2 irq_context: 0 cb_lock genl_mutex net_dm_mutex tracepoints_mutex cpu_hotplug_lock jump_label_mutex irq_context: 0 cb_lock genl_mutex net_dm_mutex tracepoints_mutex cpu_hotplug_lock jump_label_mutex text_mutex irq_context: 0 cb_lock genl_mutex net_dm_mutex tracepoints_mutex cpu_hotplug_lock jump_label_mutex text_mutex ptlock_ptr(page)#2 irq_context: 0 cb_lock genl_mutex net_dm_mutex tracepoints_mutex &c->lock irq_context: 0 cb_lock genl_mutex net_dm_mutex tracepoints_mutex cpu_hotplug_lock static_call_mutex text_mutex &rq->__lock irq_context: 0 cb_lock genl_mutex net_dm_mutex tracepoints_mutex cpu_hotplug_lock static_call_mutex text_mutex &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 cgroup_mutex cpu_hotplug_lock jump_label_mutex text_mutex text_mutex.wait_lock irq_context: 0 cb_lock genl_mutex net_dm_mutex tracepoints_mutex cpu_hotplug_lock static_call_mutex text_mutex.wait_lock irq_context: 0 cb_lock genl_mutex net_dm_mutex tracepoints_mutex cpu_hotplug_lock static_call_mutex &p->pi_lock irq_context: 0 cb_lock genl_mutex net_dm_mutex tracepoints_mutex cpu_hotplug_lock static_call_mutex &p->pi_lock &rq->__lock irq_context: 0 cb_lock genl_mutex net_dm_mutex tracepoints_mutex cpu_hotplug_lock static_call_mutex &p->pi_lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 cb_lock genl_mutex net_dm_mutex tracepoints_mutex cpu_hotplug_lock static_call_mutex &rq->__lock irq_context: 0 cb_lock genl_mutex net_dm_mutex tracepoints_mutex cpu_hotplug_lock static_call_mutex &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: softirq (&app->join_timer) rcu_read_lock_bh dev->qdisc_tx_busylock ?: &qdisc_tx_busylock &data->lock irq_context: softirq (&app->join_timer) rcu_read_lock_bh dev->qdisc_tx_busylock ?: &qdisc_tx_busylock &data->lock &obj_hash[i].lock irq_context: softirq (&app->join_timer) rcu_read_lock_bh dev->qdisc_tx_busylock ?: &qdisc_tx_busylock &data->lock &base->lock irq_context: softirq (&app->join_timer) rcu_read_lock_bh dev->qdisc_tx_busylock ?: &qdisc_tx_busylock &data->lock &base->lock &obj_hash[i].lock irq_context: softirq rcu_callback &data->lock irq_context: 0 (wq_completion)events (work_completion)(&rdev->wiphy_work) &rdev->wiphy.mtx &data->lock irq_context: 0 (wq_completion)events (work_completion)(&rdev->wiphy_work) &rdev->wiphy.mtx &data->lock &obj_hash[i].lock irq_context: 0 (wq_completion)events (work_completion)(&rdev->wiphy_work) &rdev->wiphy.mtx &data->lock &base->lock irq_context: 0 (wq_completion)events (work_completion)(&rdev->wiphy_work) &rdev->wiphy.mtx &data->lock &base->lock &obj_hash[i].lock irq_context: softirq rcu_read_lock rcu_read_lock &data->lock irq_context: 0 sk_lock-AF_INET6 &data->lock irq_context: softirq (&lapb->t1timer) &lapb->lock rcu_read_lock_bh &data->lock irq_context: softirq rcu_read_lock &data->lock irq_context: 0 &data->lock irq_context: 0 pernet_ops_rwsem devices_rwsem rdma_nets_rwsem &device->compat_devs_mutex remove_cache_srcu rcu_read_lock &cfs_rq->removed.lock irq_context: 0 pernet_ops_rwsem ipvs->est_mutex fs_reclaim mmu_notifier_invalidate_range_start &rq->__lock irq_context: 0 pernet_ops_rwsem ipvs->est_mutex fs_reclaim mmu_notifier_invalidate_range_start &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: softirq (&hsr->announce_timer) rcu_read_lock &hsr->seqnr_lock &data->lock irq_context: 0 pernet_ops_rwsem rtnl_mutex &data->lock irq_context: softirq (&app->join_timer) rcu_read_lock_bh &data->lock irq_context: 0 sb_writers#4 &type->i_mutex_dir_key#3/1 &type->i_mutex_dir_key#3 remove_cache_srcu rcu_read_lock pgd_lock irq_context: 0 sb_writers#4 &type->i_mutex_dir_key#3/1 &type->i_mutex_dir_key#3 remove_cache_srcu rcu_read_lock stock_lock irq_context: 0 &sb->s_type->i_mutex_key#10 slock-AF_INET6 quarantine_lock irq_context: 0 sb_writers#4 &type->i_mutex_dir_key#3/1 &type->i_mutex_dir_key#3 remove_cache_srcu rcu_read_lock &obj_hash[i].lock irq_context: 0 sb_writers#4 &type->i_mutex_dir_key#3/1 &type->i_mutex_dir_key#3 remove_cache_srcu rcu_read_lock key irq_context: 0 sb_writers#4 &type->i_mutex_dir_key#3/1 &type->i_mutex_dir_key#3 remove_cache_srcu rcu_read_lock pcpu_lock irq_context: 0 sb_writers#4 &type->i_mutex_dir_key#3/1 &type->i_mutex_dir_key#3 remove_cache_srcu rcu_read_lock percpu_counters_lock irq_context: 0 sb_writers#4 &type->i_mutex_dir_key#3/1 &type->i_mutex_dir_key#3 remove_cache_srcu rcu_read_lock pcpu_lock stock_lock irq_context: 0 sb_writers#4 &type->i_mutex_dir_key#3/1 &type->i_mutex_dir_key#3 remove_cache_srcu rcu_read_lock &cfs_rq->removed.lock irq_context: 0 pcpu_alloc_mutex rcu_read_lock &rq->__lock irq_context: 0 pcpu_alloc_mutex rcu_read_lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: softirq (&data->send_timer) irq_context: softirq (&data->send_timer) rcu_read_lock &pool->lock irq_context: softirq (&data->send_timer) rcu_read_lock &pool->lock &obj_hash[i].lock irq_context: softirq (&data->send_timer) rcu_read_lock &pool->lock &p->pi_lock irq_context: softirq (&data->send_timer) rcu_read_lock &pool->lock &p->pi_lock &rq->__lock irq_context: softirq (&data->send_timer) rcu_read_lock &pool->lock &p->pi_lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)events (work_completion)(&data->dm_alert_work) irq_context: 0 (wq_completion)events (work_completion)(&data->dm_alert_work) fs_reclaim irq_context: 0 (wq_completion)events (work_completion)(&data->dm_alert_work) fs_reclaim mmu_notifier_invalidate_range_start irq_context: 0 (wq_completion)events (work_completion)(&data->dm_alert_work) batched_entropy_u8.lock irq_context: 0 (wq_completion)events (work_completion)(&data->dm_alert_work) kfence_freelist_lock irq_context: 0 (wq_completion)events (work_completion)(&data->dm_alert_work) pool_lock#2 irq_context: 0 (wq_completion)events (work_completion)(&data->dm_alert_work) &data->lock irq_context: 0 (wq_completion)events (work_completion)(&data->dm_alert_work) &obj_hash[i].lock irq_context: 0 (wq_completion)events (work_completion)(&data->dm_alert_work) nl_table_lock irq_context: 0 (wq_completion)events (work_completion)(&data->dm_alert_work) nl_table_wait.lock irq_context: softirq rcu_read_lock rcu_read_lock &data->lock &obj_hash[i].lock irq_context: softirq rcu_read_lock rcu_read_lock &data->lock &base->lock irq_context: softirq rcu_read_lock rcu_read_lock &data->lock &base->lock &obj_hash[i].lock irq_context: softirq (&lapb->t1timer) &lapb->lock rcu_read_lock_bh &data->lock &obj_hash[i].lock irq_context: softirq (&lapb->t1timer) &lapb->lock rcu_read_lock_bh &data->lock &base->lock irq_context: softirq (&lapb->t1timer) &lapb->lock rcu_read_lock_bh &data->lock &base->lock &obj_hash[i].lock irq_context: softirq (&ndev->rs_timer) rcu_read_lock rcu_read_lock rcu_read_lock_bh rcu_read_lock &data->lock irq_context: 0 pernet_ops_rwsem &root->kernfs_rwsem kernfs_idr_lock fill_pool_map-wait-type-override pool_lock#2 irq_context: 0 pernet_ops_rwsem &root->kernfs_rwsem kernfs_idr_lock fill_pool_map-wait-type-override pool_lock irq_context: 0 (wq_completion)events (work_completion)(&data->dm_alert_work) &c->lock irq_context: 0 (wq_completion)events (work_completion)(&data->dm_alert_work) &n->list_lock irq_context: 0 (wq_completion)events (work_completion)(&data->dm_alert_work) &n->list_lock &c->lock irq_context: 0 (wq_completion)events (work_completion)(&data->dm_alert_work) rcu_node_0 irq_context: 0 (wq_completion)events (work_completion)(&data->dm_alert_work) &rcu_state.expedited_wq irq_context: 0 pernet_ops_rwsem k-sk_lock-AF_INET6 k-slock-AF_INET6 &tcp_hashinfo.bhash[i].lock per_cpu_ptr(&cgroup_rstat_cpu_lock, cpu) irq_context: 0 (wq_completion)events (work_completion)(&data->dm_alert_work) &rcu_state.expedited_wq &p->pi_lock irq_context: 0 (wq_completion)events (work_completion)(&data->dm_alert_work) &rcu_state.expedited_wq &p->pi_lock &rq->__lock irq_context: 0 (wq_completion)events (work_completion)(&data->dm_alert_work) &rcu_state.expedited_wq &p->pi_lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)events (work_completion)(&data->dm_alert_work) &rq->__lock irq_context: 0 (wq_completion)events (work_completion)(&data->dm_alert_work) &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 pernet_ops_rwsem &root->kernfs_rwsem &rcu_state.expedited_wq &p->pi_lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)events (work_completion)(&data->dm_alert_work) &meta->lock irq_context: 0 (wq_completion)wg-kex-wg0#7 (work_completion)(&peer->transmit_handshake_work) &peer->endpoint_lock rcu_read_lock_bh rcu_read_lock rcu_read_lock rcu_read_lock rcu_read_lock_bh &data->lock irq_context: 0 (wq_completion)bat_events (work_completion)(&(&bat_priv->bla.work)->work) rcu_read_lock &data->lock irq_context: softirq (&app->join_timer) &app->lock batched_entropy_u8.lock crngs.lock irq_context: softirq (&n->timer) &data->lock irq_context: 0 (wq_completion)wg-kex-wg2#9 (work_completion)(&peer->transmit_handshake_work) &peer->endpoint_lock rcu_read_lock_bh &data->lock irq_context: 0 (wq_completion)wg-kex-wg2#9 (work_completion)(&peer->transmit_handshake_work) &peer->endpoint_lock rcu_read_lock_bh &data->lock &obj_hash[i].lock irq_context: 0 (wq_completion)wg-kex-wg2#9 (work_completion)(&peer->transmit_handshake_work) &peer->endpoint_lock rcu_read_lock_bh &data->lock &base->lock irq_context: 0 (wq_completion)wg-kex-wg2#9 (work_completion)(&peer->transmit_handshake_work) &peer->endpoint_lock rcu_read_lock_bh &data->lock &base->lock &obj_hash[i].lock irq_context: 0 (wq_completion)events (work_completion)(&data->dm_alert_work) remove_cache_srcu irq_context: 0 (wq_completion)events (work_completion)(&data->dm_alert_work) remove_cache_srcu quarantine_lock irq_context: 0 (wq_completion)events (work_completion)(&data->dm_alert_work) remove_cache_srcu &c->lock irq_context: 0 (wq_completion)events (work_completion)(&data->dm_alert_work) remove_cache_srcu &n->list_lock irq_context: 0 (wq_completion)events (work_completion)(&data->dm_alert_work) remove_cache_srcu rcu_read_lock pool_lock#2 irq_context: 0 (wq_completion)events (work_completion)(&data->dm_alert_work) remove_cache_srcu &obj_hash[i].lock irq_context: softirq (&n->timer) rcu_read_lock rcu_read_lock rcu_read_lock_bh &data->lock irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem &net->ipv6.ip6addrlbl_table.lock fill_pool_map-wait-type-override &____s->seqcount#2 irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem &net->ipv6.ip6addrlbl_table.lock fill_pool_map-wait-type-override &____s->seqcount irq_context: 0 pernet_ops_rwsem kernfs_idr_lock fill_pool_map-wait-type-override pool_lock#2 irq_context: 0 pernet_ops_rwsem kernfs_idr_lock fill_pool_map-wait-type-override &c->lock irq_context: 0 pernet_ops_rwsem kernfs_idr_lock fill_pool_map-wait-type-override &____s->seqcount#2 irq_context: 0 pernet_ops_rwsem kernfs_idr_lock fill_pool_map-wait-type-override &____s->seqcount irq_context: 0 pernet_ops_rwsem kernfs_idr_lock fill_pool_map-wait-type-override pool_lock irq_context: 0 &data->lock &obj_hash[i].lock irq_context: 0 &data->lock &base->lock irq_context: 0 &data->lock &base->lock &obj_hash[i].lock irq_context: 0 (wq_completion)events (work_completion)(&data->dm_alert_work) remove_cache_srcu pool_lock#2 irq_context: 0 (wq_completion)bat_events (work_completion)(&(&bat_priv->nc.work)->work) rcu_read_lock &rq->__lock &obj_hash[i].lock irq_context: 0 (wq_completion)bat_events (work_completion)(&(&bat_priv->nc.work)->work) rcu_read_lock &rq->__lock &base->lock irq_context: 0 (wq_completion)bat_events (work_completion)(&(&bat_priv->nc.work)->work) rcu_read_lock &rq->__lock &base->lock &obj_hash[i].lock irq_context: 0 rtnl_mutex uevent_sock_mutex &data->lock irq_context: softirq (&in_dev->mr_ifc_timer) rcu_read_lock rcu_read_lock_bh rcu_read_lock &data->lock irq_context: 0 (wq_completion)events_power_efficient (work_completion)(&(&tbl->gc_work)->work) &tbl->lock rcu_read_lock &ei->socket.wq.wait &p->pi_lock &cfs_rq->removed.lock irq_context: 0 (wq_completion)events (work_completion)(&w->work)#2 &cfs_rq->removed.lock irq_context: 0 (wq_completion)mld (work_completion)(&(&idev->mc_ifc_work)->work) &idev->mc_lock rcu_read_lock rcu_read_lock rcu_read_lock_bh &data->lock irq_context: 0 (wq_completion)ipv6_addrconf (work_completion)(&(&ifa->dad_work)->work) rtnl_mutex &idev->mc_lock rcu_read_lock rcu_read_lock rcu_read_lock_bh rcu_read_lock &br->multicast_lock &n->list_lock irq_context: 0 (wq_completion)ipv6_addrconf (work_completion)(&(&ifa->dad_work)->work) rtnl_mutex &idev->mc_lock rcu_read_lock rcu_read_lock rcu_read_lock_bh rcu_read_lock &br->multicast_lock &n->list_lock &c->lock irq_context: 0 (wq_completion)ipv6_addrconf (work_completion)(&(&ifa->dad_work)->work) rtnl_mutex &idev->mc_lock rcu_read_lock rcu_read_lock rcu_read_lock_bh rcu_read_lock &data->lock irq_context: 0 (wq_completion)ipv6_addrconf (work_completion)(&(&ifa->dad_work)->work) rtnl_mutex rcu_read_lock rcu_read_lock rcu_read_lock_bh rcu_read_lock &data->lock irq_context: 0 (wq_completion)mld (work_completion)(&(&idev->mc_ifc_work)->work) &idev->mc_lock rcu_read_lock rcu_read_lock rcu_read_lock_bh rcu_read_lock &data->lock irq_context: 0 (wq_completion)mld (work_completion)(&(&idev->mc_ifc_work)->work) &idev->mc_lock rcu_read_lock rcu_read_lock rcu_read_lock_bh rcu_read_lock &base->lock irq_context: 0 (wq_completion)mld (work_completion)(&(&idev->mc_ifc_work)->work) &idev->mc_lock rcu_read_lock rcu_read_lock rcu_read_lock_bh rcu_read_lock &base->lock &obj_hash[i].lock irq_context: 0 (wq_completion)mld (work_completion)(&(&idev->mc_dad_work)->work) &idev->mc_lock remove_cache_srcu &pcp->lock &zone->lock irq_context: 0 (wq_completion)mld (work_completion)(&(&idev->mc_dad_work)->work) &idev->mc_lock remove_cache_srcu &pcp->lock &zone->lock &____s->seqcount irq_context: 0 (wq_completion)mld (work_completion)(&(&idev->mc_dad_work)->work) &idev->mc_lock rcu_read_lock rcu_read_lock rcu_read_lock_bh rcu_read_lock &data->lock irq_context: softirq rcu_read_lock &data->lock &obj_hash[i].lock irq_context: softirq rcu_read_lock &data->lock &base->lock irq_context: softirq rcu_read_lock &data->lock &base->lock &obj_hash[i].lock irq_context: softirq (&journal->j_commit_timer) &p->pi_lock &cfs_rq->removed.lock irq_context: softirq (&ndev->rs_timer) rcu_read_lock rcu_read_lock rcu_read_lock_bh rcu_read_lock &data->lock &obj_hash[i].lock irq_context: softirq (&ndev->rs_timer) rcu_read_lock rcu_read_lock rcu_read_lock_bh rcu_read_lock &data->lock &base->lock irq_context: softirq (&ndev->rs_timer) rcu_read_lock rcu_read_lock rcu_read_lock_bh rcu_read_lock &data->lock &base->lock &obj_hash[i].lock irq_context: 0 pernet_ops_rwsem rcu_read_lock &pool->lock/1 &p->pi_lock &rq->__lock &cfs_rq->removed.lock irq_context: 0 (wq_completion)wg-kex-wg1#4 (work_completion)(&({ do { const void *__vpp_verify = (typeof((worker) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __asm__ ("" : "=r"(__ptr) : "0"((typeof(*((worker))) *)((worker)))); (typeof((typeof(*((worker))) *)((worker)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); })->work) &wg->static_identity.lock &handshake->lock rcu_read_lock_bh batched_entropy_u32.lock crngs.lock irq_context: softirq (&peer->timer_retransmit_handshake) &list->lock#17 &data->lock irq_context: 0 (wq_completion)events (work_completion)(&data->dm_alert_work) remove_cache_srcu &rq->__lock irq_context: 0 (wq_completion)events (work_completion)(&data->dm_alert_work) remove_cache_srcu &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: softirq (&ndev->rs_timer) rcu_read_lock rcu_read_lock rcu_read_lock_bh &data->lock irq_context: 0 sk_lock-AF_INET6 &dccp_hashinfo.bhash[i].lock &dccp_hashinfo.bhash2[i].lock batched_entropy_u8.lock crngs.lock irq_context: 0 cb_lock genl_mutex rtnl_mutex &rdev->wiphy.mtx pcpu_alloc_mutex pcpu_alloc_mutex.wait_lock irq_context: 0 cb_lock genl_mutex rtnl_mutex &rdev->wiphy.mtx rcu_read_lock &ei->socket.wq.wait irq_context: 0 cb_lock genl_mutex rtnl_mutex &rdev->wiphy.mtx rcu_read_lock &ei->socket.wq.wait &p->pi_lock irq_context: 0 cb_lock genl_mutex rtnl_mutex &rdev->wiphy.mtx rcu_read_lock &ei->socket.wq.wait &p->pi_lock &rq->__lock irq_context: 0 cb_lock genl_mutex rtnl_mutex &rdev->wiphy.mtx rcu_read_lock &ei->socket.wq.wait &p->pi_lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 sb_writers#4 &type->i_mutex_dir_key#3/1 jbd2_handle &ei->i_data_sem remove_cache_srcu &____s->seqcount irq_context: 0 cb_lock genl_mutex rcu_read_lock &pool->lock/1 &p->pi_lock &cfs_rq->removed.lock irq_context: 0 cb_lock genl_mutex rfkill_global_mutex uevent_sock_mutex nl_table_wait.lock &p->pi_lock &cfs_rq->removed.lock irq_context: 0 kn->active#18 &kernfs_locks->open_file_mutex[count] &n->list_lock irq_context: 0 kn->active#18 &kernfs_locks->open_file_mutex[count] &n->list_lock &c->lock irq_context: 0 &p->lock &of->mutex kn->active#20 &rq->__lock irq_context: 0 &p->lock &of->mutex kn->active#20 &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 &sb->s_type->i_mutex_key#10 &data->lock irq_context: 0 (wq_completion)phy28 irq_context: 0 (wq_completion)phy28 (work_completion)(&local->reconfig_filter) irq_context: 0 (wq_completion)phy28 (work_completion)(&local->reconfig_filter) &local->filter_lock irq_context: 0 rtnl_mutex &data->lock irq_context: 0 sb_writers#3 tomoyo_ss quarantine_lock irq_context: softirq (&app->join_timer) rcu_read_lock_bh dev->qdisc_tx_busylock ?: &qdisc_tx_busylock rcu_read_lock &c->lock irq_context: softirq rcu_read_lock rcu_read_lock slock-AF_INET6/1 sk_lock-AF_INET6 rcu_read_lock &____s->seqcount#7 irq_context: softirq rcu_read_lock rcu_read_lock slock-AF_INET6/1 sk_lock-AF_INET6 rcu_read_lock &nf_conntrack_locks[i] irq_context: softirq rcu_read_lock rcu_read_lock slock-AF_INET6/1 sk_lock-AF_INET6 rcu_read_lock &nf_conntrack_locks[i] &nf_conntrack_locks[i]/1 irq_context: softirq rcu_read_lock rcu_read_lock slock-AF_INET6/1 sk_lock-AF_INET6 rcu_read_lock &nf_conntrack_locks[i]/1 irq_context: softirq rcu_read_lock rcu_read_lock slock-AF_INET6/1 sk_lock-AF_INET6 rcu_read_lock &____s->seqcount#10 irq_context: softirq rcu_read_lock rcu_read_lock slock-AF_INET6/1 sk_lock-AF_INET6 rcu_read_lock rcu_read_lock_bh rcu_read_lock pool_lock#2 irq_context: softirq rcu_read_lock rcu_read_lock slock-AF_INET6/1 sk_lock-AF_INET6 rcu_read_lock rcu_read_lock_bh rcu_read_lock tk_core.seq.seqcount irq_context: softirq rcu_read_lock rcu_read_lock slock-AF_INET6/1 sk_lock-AF_INET6 rcu_read_lock rcu_read_lock_bh rcu_read_lock &obj_hash[i].lock irq_context: softirq rcu_read_lock rcu_read_lock slock-AF_INET6/1 sk_lock-AF_INET6 rcu_read_lock rcu_read_lock_bh tk_core.seq.seqcount irq_context: softirq rcu_read_lock rcu_read_lock slock-AF_INET6/1 sk_lock-AF_INET6 rcu_read_lock rcu_read_lock_bh &list->lock#5 irq_context: softirq rcu_read_lock rcu_read_lock slock-AF_INET6/1 sk_lock-AF_INET6 &data->lock irq_context: 0 cb_lock nlk_cb_mutex-GENERIC genl_mutex fs_reclaim irq_context: 0 cb_lock nlk_cb_mutex-GENERIC genl_mutex fs_reclaim mmu_notifier_invalidate_range_start irq_context: 0 cb_lock nlk_cb_mutex-GENERIC genl_mutex pool_lock#2 irq_context: 0 cb_lock nlk_cb_mutex-GENERIC genl_mutex &c->lock irq_context: 0 cb_lock nlk_cb_mutex-GENERIC genl_mutex &obj_hash[i].lock irq_context: 0 cb_lock nlk_cb_mutex-GENERIC genl_mutex &n->list_lock irq_context: 0 cb_lock nlk_cb_mutex-GENERIC genl_mutex &n->list_lock &c->lock irq_context: 0 (wq_completion)events (work_completion)(&nlk->work) &data->lock irq_context: 0 (wq_completion)phy29 irq_context: 0 (wq_completion)phy29 (work_completion)(&local->reconfig_filter) irq_context: 0 (wq_completion)phy29 (work_completion)(&local->reconfig_filter) &local->filter_lock irq_context: 0 rtnl_mutex &lapb->lock rcu_read_lock_bh &data->lock irq_context: 0 (wq_completion)events (work_completion)(&data->dm_alert_work) &____s->seqcount#2 irq_context: 0 (wq_completion)events (work_completion)(&data->dm_alert_work) &____s->seqcount irq_context: 0 (wq_completion)events_unbound (work_completion)(&map->work) rcu_state.exp_mutex &rq->__lock &cfs_rq->removed.lock irq_context: 0 (wq_completion)wg-kex-wg2#3 (work_completion)(&peer->transmit_handshake_work) &peer->endpoint_lock rcu_read_lock_bh rcu_read_lock rcu_read_lock_bh rcu_read_lock tk_core.seq.seqcount irq_context: 0 (wq_completion)wg-kex-wg2#3 (work_completion)(&peer->transmit_handshake_work) &peer->endpoint_lock rcu_read_lock_bh rcu_read_lock rcu_read_lock_bh rcu_read_lock &obj_hash[i].lock irq_context: 0 (wq_completion)wg-kex-wg1#4 (work_completion)(&({ do { const void *__vpp_verify = (typeof((worker) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __asm__ ("" : "=r"(__ptr) : "0"((typeof(*((worker))) *)((worker)))); (typeof((typeof(*((worker))) *)((worker)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); })->work) &peer->endpoint_lock rcu_read_lock_bh rcu_read_lock rcu_read_lock_bh rcu_read_lock tk_core.seq.seqcount irq_context: 0 (wq_completion)wg-kex-wg1#4 (work_completion)(&({ do { const void *__vpp_verify = (typeof((worker) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __asm__ ("" : "=r"(__ptr) : "0"((typeof(*((worker))) *)((worker)))); (typeof((typeof(*((worker))) *)((worker)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); })->work) &peer->endpoint_lock rcu_read_lock_bh rcu_read_lock rcu_read_lock_bh rcu_read_lock &obj_hash[i].lock irq_context: 0 rtnl_mutex lapb_list_lock &n->list_lock irq_context: 0 rtnl_mutex lapb_list_lock &n->list_lock &c->lock irq_context: 0 &sig->cred_guard_mutex tomoyo_ss &rq->__lock &cfs_rq->removed.lock irq_context: 0 (wq_completion)wg-crypt-wg1#5 (work_completion)(&peer->transmit_packet_work) &peer->endpoint_lock rcu_read_lock_bh rcu_read_lock rcu_read_lock_bh rcu_read_lock batched_entropy_u8.lock irq_context: 0 (wq_completion)wg-crypt-wg1#5 (work_completion)(&peer->transmit_packet_work) &peer->endpoint_lock rcu_read_lock_bh rcu_read_lock rcu_read_lock_bh rcu_read_lock kfence_freelist_lock irq_context: 0 (wq_completion)wg-crypt-wg1#5 (work_completion)(&peer->transmit_packet_work) &peer->endpoint_lock rcu_read_lock_bh rcu_read_lock rcu_read_lock_bh rcu_read_lock &meta->lock irq_context: softirq (&app->join_timer) rcu_read_lock_bh &data->lock &obj_hash[i].lock irq_context: softirq (&app->join_timer) rcu_read_lock_bh &data->lock &base->lock irq_context: softirq (&app->join_timer) rcu_read_lock_bh &data->lock &base->lock &obj_hash[i].lock irq_context: 0 sb_writers#4 &sb->s_type->i_mutex_key#8 lock#4 &lruvec->lru_lock per_cpu_ptr(&cgroup_rstat_cpu_lock, cpu) irq_context: softirq (&hsr->announce_timer) rcu_read_lock &hsr->seqnr_lock &data->lock &obj_hash[i].lock irq_context: softirq (&hsr->announce_timer) rcu_read_lock &hsr->seqnr_lock &data->lock &base->lock irq_context: softirq (&hsr->announce_timer) rcu_read_lock &hsr->seqnr_lock &data->lock &base->lock &obj_hash[i].lock irq_context: 0 &f->f_pos_lock &type->i_mutex_dir_key#3 sb_writers#4 jbd2_handle rcu_read_lock &cfs_rq->removed.lock irq_context: 0 &f->f_pos_lock &type->i_mutex_dir_key#3 sb_writers#4 jbd2_handle rcu_read_lock &obj_hash[i].lock irq_context: 0 sk_lock-AF_INET6 &dccp_hashinfo.bhash[i].lock per_cpu_ptr(&cgroup_rstat_cpu_lock, cpu) irq_context: 0 (wq_completion)wg-kex-wg2#9 (work_completion)(&peer->transmit_handshake_work) &peer->endpoint_lock rcu_read_lock_bh rcu_read_lock rcu_read_lock rcu_read_lock rcu_read_lock_bh &data->lock irq_context: 0 (wq_completion)events (work_completion)(&data->dm_alert_work) remove_cache_srcu &pcp->lock &zone->lock irq_context: 0 (wq_completion)rcu_gp (work_completion)(&rew->rew_work) rcu_state.exp_wake_mutex &rnp->exp_wq[1] &p->pi_lock &cfs_rq->removed.lock irq_context: 0 pernet_ops_rwsem k-sk_lock-AF_RXRPC &rxnet->local_mutex fill_pool_map-wait-type-override &n->list_lock irq_context: 0 pernet_ops_rwsem k-sk_lock-AF_RXRPC &rxnet->local_mutex fill_pool_map-wait-type-override &n->list_lock &c->lock irq_context: 0 pernet_ops_rwsem k-sk_lock-AF_RXRPC &rxnet->local_mutex fill_pool_map-wait-type-override &rq->__lock irq_context: 0 pernet_ops_rwsem k-sk_lock-AF_RXRPC &rxnet->local_mutex fill_pool_map-wait-type-override &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)events (work_completion)(&data->dm_alert_work) remove_cache_srcu &pcp->lock &zone->lock &____s->seqcount irq_context: softirq (&ndev->rs_timer) rcu_read_lock rcu_read_lock &data->lock irq_context: 0 dup_mmap_sem &mm->mmap_lock &mm->mmap_lock/1 mmu_notifier_invalidate_range_start &cfs_rq->removed.lock irq_context: 0 dup_mmap_sem &mm->mmap_lock &mm->mmap_lock/1 mmu_notifier_invalidate_range_start &obj_hash[i].lock irq_context: 0 &net->xfrm.xfrm_cfg_mutex &data->lock irq_context: 0 &sb->s_type->i_mutex_key#10 &data->lock &obj_hash[i].lock irq_context: 0 &sb->s_type->i_mutex_key#10 &data->lock &base->lock irq_context: 0 &sb->s_type->i_mutex_key#10 &data->lock &base->lock &obj_hash[i].lock irq_context: 0 &mm->mmap_lock remove_cache_srcu rcu_read_lock &cfs_rq->removed.lock irq_context: 0 &mm->mmap_lock remove_cache_srcu rcu_read_lock &obj_hash[i].lock irq_context: 0 pernet_ops_rwsem rtnl_mutex rcu_state.exp_mutex pgd_lock irq_context: 0 pernet_ops_rwsem rtnl_mutex rcu_state.exp_mutex rcu_read_lock pool_lock#2 irq_context: 0 pernet_ops_rwsem rtnl_mutex rcu_state.exp_mutex key irq_context: 0 pernet_ops_rwsem rtnl_mutex rcu_state.exp_mutex pcpu_lock irq_context: 0 pernet_ops_rwsem rtnl_mutex rcu_state.exp_mutex percpu_counters_lock irq_context: 0 pernet_ops_rwsem rcu_read_lock &pool->lock fill_pool_map-wait-type-override pool_lock#2 irq_context: 0 pernet_ops_rwsem rcu_read_lock &pool->lock fill_pool_map-wait-type-override &c->lock irq_context: 0 pernet_ops_rwsem rcu_read_lock &pool->lock fill_pool_map-wait-type-override pool_lock irq_context: 0 &sb->s_type->i_mutex_key#10 sk_lock-AF_INET6 fill_pool_map-wait-type-override pool_lock#2 irq_context: 0 &sb->s_type->i_mutex_key#10 sk_lock-AF_INET6 fill_pool_map-wait-type-override pool_lock irq_context: 0 (wq_completion)events_unbound (work_completion)(&sub_info->work) fill_pool_map-wait-type-override &____s->seqcount#2 irq_context: 0 &u->iolock &mm->mmap_lock per_cpu_ptr(&cgroup_rstat_cpu_lock, cpu) irq_context: softirq (&n->timer) rcu_read_lock rcu_read_lock rcu_read_lock_bh &data->lock &obj_hash[i].lock irq_context: softirq (&n->timer) rcu_read_lock rcu_read_lock rcu_read_lock_bh &data->lock &base->lock irq_context: softirq (&n->timer) rcu_read_lock rcu_read_lock rcu_read_lock_bh &data->lock &base->lock &obj_hash[i].lock irq_context: 0 sk_lock-AF_INET6 &dccp_hashinfo.bhash[i].lock &dccp_hashinfo.bhash2[i].lock kfence_freelist_lock irq_context: 0 &sb->s_type->i_mutex_key#10 slock-AF_INET6 &dccp_hashinfo.bhash[i].lock &meta->lock irq_context: 0 &sb->s_type->i_mutex_key#10 slock-AF_INET6 &dccp_hashinfo.bhash[i].lock kfence_freelist_lock irq_context: softirq (&n->timer) &data->lock &obj_hash[i].lock irq_context: softirq (&n->timer) &data->lock &base->lock irq_context: softirq (&n->timer) &data->lock &base->lock &obj_hash[i].lock irq_context: 0 sk_lock-AF_INET6 rcu_read_lock rcu_read_lock &data->lock irq_context: 0 &sb->s_type->i_mutex_key#10 slock-AF_INET6 &data->lock irq_context: 0 sb_writers#7 &rq->__lock &cfs_rq->removed.lock irq_context: 0 cb_lock &rdev->wiphy.mtx &data->lock irq_context: softirq rcu_read_lock rcu_read_lock &dreq->dreq_lock slock-AF_INET6 batched_entropy_u8.lock irq_context: softirq rcu_read_lock rcu_read_lock &dreq->dreq_lock slock-AF_INET6 kfence_freelist_lock irq_context: 0 &sb->s_type->i_mutex_key#10 sk_lock-AF_INET6 slock-AF_INET6 &meta->lock irq_context: 0 &sb->s_type->i_mutex_key#10 sk_lock-AF_INET6 slock-AF_INET6 kfence_freelist_lock irq_context: softirq (&icsk->icsk_retransmit_timer) &obj_hash[i].lock pool_lock irq_context: 0 sk_lock-AF_INET6 rcu_read_lock rcu_read_lock rcu_read_lock_bh &sch->q.lock hrtimer_bases.lock tk_core.seq.seqcount irq_context: softirq rcu_read_lock rcu_read_lock slock-AF_INET6/1 sk_lock-AF_INET6 rcu_read_lock rcu_read_lock rcu_read_lock_bh &sch->q.lock hrtimer_bases.lock tk_core.seq.seqcount irq_context: 0 &sb->s_type->i_mutex_key#10 sk_lock-AF_INET6 slock-AF_INET6 rcu_read_lock rcu_read_lock rcu_read_lock_bh &sch->q.lock &c->lock irq_context: 0 &f->f_pos_lock sb_writers#4 &sb->s_type->i_mutex_key#8 mapping.invalidate_lock &mapping->private_lock rcu_read_lock &memcg->move_lock irq_context: 0 &f->f_pos_lock sb_writers#4 &sb->s_type->i_mutex_key#8 mapping.invalidate_lock &sb->s_type->i_lock_key#22 &xa->xa_lock#9 &obj_hash[i].lock irq_context: 0 &f->f_pos_lock sb_writers#4 &sb->s_type->i_mutex_key#8 mapping.invalidate_lock &sb->s_type->i_lock_key#22 &xa->xa_lock#9 pool_lock#2 irq_context: 0 &f->f_pos_lock sb_writers#4 &sb->s_type->i_mutex_key#8 mapping.invalidate_lock jbd2_handle &ei->i_data_sem &(ei->i_block_reservation_lock) irq_context: 0 epnested_mutex &ep->mtx wakeup_ida.xa_lock irq_context: 0 epnested_mutex &ep->mtx &x->wait#9 irq_context: 0 epnested_mutex &ep->mtx &obj_hash[i].lock irq_context: 0 epnested_mutex &ep->mtx &k->list_lock irq_context: 0 epnested_mutex &ep->mtx gdp_mutex irq_context: 0 epnested_mutex &ep->mtx gdp_mutex &k->list_lock irq_context: 0 epnested_mutex &ep->mtx gdp_mutex fs_reclaim irq_context: 0 epnested_mutex &ep->mtx gdp_mutex fs_reclaim mmu_notifier_invalidate_range_start irq_context: 0 epnested_mutex &ep->mtx gdp_mutex fs_reclaim mmu_notifier_invalidate_range_start &rq->__lock irq_context: 0 epnested_mutex &ep->mtx gdp_mutex fs_reclaim mmu_notifier_invalidate_range_start &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 &f->f_pos_lock sb_writers#4 &sb->s_type->i_mutex_key#8 bit_wait_table + i irq_context: 0 epnested_mutex &ep->mtx gdp_mutex lock irq_context: 0 epnested_mutex &ep->mtx gdp_mutex lock kernfs_idr_lock irq_context: 0 epnested_mutex &ep->mtx gdp_mutex &root->kernfs_rwsem irq_context: 0 epnested_mutex &ep->mtx gdp_mutex &root->kernfs_rwsem &root->kernfs_iattr_rwsem irq_context: 0 epnested_mutex &ep->mtx lock irq_context: 0 epnested_mutex &ep->mtx lock kernfs_idr_lock irq_context: 0 epnested_mutex &ep->mtx &root->kernfs_rwsem irq_context: 0 epnested_mutex &ep->mtx &root->kernfs_rwsem &root->kernfs_iattr_rwsem irq_context: 0 &f->f_pos_lock sb_writers#4 &sbi->s_writepages_rwsem rcu_read_lock key#10 irq_context: 0 epnested_mutex &ep->mtx bus_type_sem irq_context: 0 epnested_mutex &ep->mtx sysfs_symlink_target_lock irq_context: 0 epnested_mutex &ep->mtx &____s->seqcount#2 irq_context: 0 epnested_mutex &ep->mtx uevent_sock_mutex irq_context: 0 epnested_mutex &ep->mtx uevent_sock_mutex fs_reclaim irq_context: 0 epnested_mutex &ep->mtx uevent_sock_mutex fs_reclaim mmu_notifier_invalidate_range_start irq_context: 0 epnested_mutex &ep->mtx uevent_sock_mutex fs_reclaim mmu_notifier_invalidate_range_start &rq->__lock irq_context: 0 epnested_mutex &ep->mtx uevent_sock_mutex fs_reclaim mmu_notifier_invalidate_range_start &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 epnested_mutex &ep->mtx uevent_sock_mutex nl_table_lock irq_context: 0 epnested_mutex &ep->mtx uevent_sock_mutex rlock-AF_NETLINK irq_context: 0 epnested_mutex &ep->mtx uevent_sock_mutex rcu_read_lock &ei->socket.wq.wait irq_context: 0 epnested_mutex &ep->mtx uevent_sock_mutex rcu_read_lock &ei->socket.wq.wait &ep->lock irq_context: 0 epnested_mutex &ep->mtx uevent_sock_mutex rcu_read_lock &ei->socket.wq.wait &ep->lock &ep->wq irq_context: 0 epnested_mutex &ep->mtx uevent_sock_mutex rcu_read_lock &ei->socket.wq.wait &ep->lock &ep->wq &p->pi_lock irq_context: 0 epnested_mutex &ep->mtx uevent_sock_mutex rcu_read_lock &ei->socket.wq.wait &ep->lock &ep->wq &p->pi_lock &rq->__lock irq_context: 0 epnested_mutex &ep->mtx uevent_sock_mutex rcu_read_lock &ei->socket.wq.wait &ep->lock &ep->wq &p->pi_lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 epnested_mutex &ep->mtx uevent_sock_mutex rcu_read_lock rcu_node_0 irq_context: 0 epnested_mutex &ep->mtx uevent_sock_mutex rcu_read_lock &rq->__lock irq_context: 0 epnested_mutex &ep->mtx uevent_sock_mutex rcu_read_lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 epnested_mutex &ep->mtx uevent_sock_mutex nl_table_wait.lock irq_context: 0 epnested_mutex &ep->mtx uevent_sock_mutex &obj_hash[i].lock irq_context: 0 epnested_mutex &ep->mtx subsys mutex#15 irq_context: 0 epnested_mutex &ep->mtx subsys mutex#15 &k->k_lock irq_context: 0 epnested_mutex &ep->mtx events_lock irq_context: 0 epnested_mutex &ep->mtx &dentry->d_lock irq_context: 0 epnested_mutex &ep->mtx &rq->__lock irq_context: 0 epnested_mutex &ep->mtx &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 &f->f_pos_lock sb_writers#4 &mapping->private_lock irq_context: 0 &f->f_pos_lock sb_writers#4 stock_lock irq_context: 0 &f->f_pos_lock sb_writers#4 &sb->s_type->i_lock_key#22 &xa->xa_lock#9 irq_context: 0 &f->f_pos_lock sb_writers#4 &lruvec->lru_lock irq_context: 0 &f->f_pos_lock sb_writers#4 rcu_read_lock pool_lock#2 irq_context: 0 epnested_mutex &ep->mtx uevent_sock_mutex &c->lock irq_context: 0 &f->f_pos_lock sb_writers#4 rcu_read_lock rcu_node_0 irq_context: 0 &f->f_pos_lock sb_writers#4 &sb->s_type->i_lock_key#22 &xa->xa_lock#9 &obj_hash[i].lock irq_context: 0 &f->f_pos_lock sb_writers#4 &sb->s_type->i_lock_key#22 &xa->xa_lock#9 pool_lock#2 irq_context: 0 epnested_mutex &ep->mtx &n->list_lock irq_context: 0 epnested_mutex &ep->mtx &n->list_lock &c->lock irq_context: 0 epnested_mutex &ep->mtx rcu_node_0 irq_context: 0 epnested_mutex &ep->mtx &rcu_state.expedited_wq irq_context: 0 epnested_mutex &ep->mtx &rcu_state.expedited_wq &p->pi_lock irq_context: 0 epnested_mutex &ep->mtx &rcu_state.expedited_wq &p->pi_lock &rq->__lock irq_context: 0 epnested_mutex &ep->mtx &rcu_state.expedited_wq &p->pi_lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 epnested_mutex &ep->mtx lock kernfs_idr_lock pool_lock#2 irq_context: 0 &ep->mtx &root->kernfs_rwsem kernfs_idr_lock &obj_hash[i].lock irq_context: 0 &ep->mtx &root->kernfs_rwsem kernfs_idr_lock pool_lock#2 irq_context: 0 &ep->mtx kernfs_idr_lock irq_context: 0 dup_mmap_sem &mm->mmap_lock &mm->mmap_lock/1 fs_reclaim rcu_node_0 irq_context: 0 &f->f_pos_lock sb_writers#4 &sb->s_type->i_mutex_key#8 &mm->mmap_lock remove_cache_srcu rcu_read_lock &rcu_state.expedited_wq irq_context: 0 &f->f_pos_lock sb_writers#4 &sb->s_type->i_mutex_key#8 &mm->mmap_lock remove_cache_srcu rcu_read_lock &rcu_state.expedited_wq &p->pi_lock irq_context: 0 &f->f_pos_lock sb_writers#4 &sb->s_type->i_mutex_key#8 &mm->mmap_lock remove_cache_srcu rcu_read_lock &rcu_state.expedited_wq &p->pi_lock &rq->__lock irq_context: 0 &f->f_pos_lock sb_writers#4 &sb->s_type->i_mutex_key#8 &mm->mmap_lock remove_cache_srcu rcu_read_lock &rcu_state.expedited_wq &p->pi_lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 cgroup_threadgroup_rwsem rcu_read_lock &rcu_state.expedited_wq irq_context: 0 cgroup_threadgroup_rwsem rcu_read_lock &rcu_state.expedited_wq &p->pi_lock irq_context: 0 cgroup_threadgroup_rwsem rcu_read_lock &rcu_state.expedited_wq &p->pi_lock &rq->__lock irq_context: 0 cgroup_threadgroup_rwsem rcu_read_lock &rcu_state.expedited_wq &p->pi_lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 epnested_mutex &ep->mtx gdp_mutex &c->lock irq_context: 0 epnested_mutex &ep->mtx gdp_mutex &rq->__lock irq_context: 0 epnested_mutex &ep->mtx gdp_mutex &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 &ep->mtx kn->active#5 irq_context: 0 &ep->mtx kn->active#5 &root->deactivate_waitq irq_context: 0 &ep->mtx kn->active#5 &rq->__lock irq_context: 0 &ep->mtx kn->active#5 &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 &p->lock &of->mutex &root->deactivate_waitq irq_context: 0 &p->lock &of->mutex &root->deactivate_waitq &p->pi_lock irq_context: 0 &p->lock &of->mutex &root->deactivate_waitq &p->pi_lock &rq->__lock irq_context: 0 &p->lock &of->mutex &root->deactivate_waitq &p->pi_lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 &ep->mtx uevent_sock_mutex &n->list_lock irq_context: 0 &ep->mtx uevent_sock_mutex &n->list_lock &c->lock irq_context: 0 &sb->s_type->i_mutex_key#10 sk_lock-AF_INET6 slock-AF_INET6 &data->lock irq_context: 0 rcu_read_lock_bh &data->lock irq_context: softirq (&ndev->rs_timer) rcu_read_lock rcu_read_lock rcu_read_lock_bh rcu_read_lock_bh dev->qdisc_tx_busylock ?: &qdisc_tx_busylock &data->lock irq_context: 0 rcu_read_lock_bh dev->qdisc_tx_busylock ?: &qdisc_tx_busylock _xmit_NETROM &data->lock irq_context: 0 rcu_read_lock_bh &nr_netdev_xmit_lock_key &data->lock irq_context: 0 rcu_read_lock_bh _xmit_X25#2 &lapbeth->up_lock &data->lock irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem rtnl_mutex &nft_net->commit_mutex &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 pernet_ops_rwsem rtnl_mutex (work_completion)(&(&idev->mc_report_work)->work) &rq->__lock irq_context: 0 epnested_mutex &ep->mtx gdp_mutex lock kernfs_idr_lock &c->lock irq_context: 0 (wq_completion)wg-kex-wg0#4 (work_completion)(&({ do { const void *__vpp_verify = (typeof((worker) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __asm__ ("" : "=r"(__ptr) : "0"((typeof(*((worker))) *)((worker)))); (typeof((typeof(*((worker))) *)((worker)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); })->work) &n->list_lock irq_context: 0 (wq_completion)wg-kex-wg0#4 (work_completion)(&({ do { const void *__vpp_verify = (typeof((worker) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __asm__ ("" : "=r"(__ptr) : "0"((typeof(*((worker))) *)((worker)))); (typeof((typeof(*((worker))) *)((worker)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); })->work) &n->list_lock &c->lock irq_context: 0 kn->active#60 stock_lock rcu_read_lock per_cpu_ptr(&cgroup_rstat_cpu_lock, cpu) irq_context: softirq rcu_read_lock rcu_read_lock slock-AF_INET6/1 sk_lock-AF_INET6 rcu_read_lock &obj_hash[i].lock pool_lock irq_context: softirq rcu_read_lock rcu_read_lock slock-AF_INET6/1 sk_lock-AF_INET6 rcu_read_lock fill_pool_map-wait-type-override pool_lock#2 irq_context: softirq rcu_read_lock rcu_read_lock slock-AF_INET6/1 sk_lock-AF_INET6 rcu_read_lock fill_pool_map-wait-type-override pool_lock irq_context: 0 vlan_ioctl_mutex rtnl_mutex (console_sem).lock irq_context: 0 vlan_ioctl_mutex rtnl_mutex console_lock console_srcu console_owner_lock irq_context: 0 vlan_ioctl_mutex rtnl_mutex console_lock console_srcu console_owner irq_context: 0 vlan_ioctl_mutex rtnl_mutex console_lock console_srcu console_owner &port_lock_key irq_context: 0 vlan_ioctl_mutex rtnl_mutex console_lock console_srcu console_owner console_owner_lock irq_context: softirq (&app->join_timer)#2 &app->lock#2 &obj_hash[i].lock irq_context: softirq (&app->join_timer)#2 rcu_read_lock_bh dev->qdisc_tx_busylock ?: &qdisc_tx_busylock &data->lock irq_context: 0 vlan_ioctl_mutex rtnl_mutex pgd_lock irq_context: 0 vlan_ioctl_mutex rtnl_mutex key irq_context: 0 vlan_ioctl_mutex rtnl_mutex pcpu_lock irq_context: 0 vlan_ioctl_mutex rtnl_mutex percpu_counters_lock irq_context: 0 vlan_ioctl_mutex rtnl_mutex pcpu_lock stock_lock irq_context: 0 vlan_ioctl_mutex rtnl_mutex &cfs_rq->removed.lock irq_context: 0 vlan_ioctl_mutex stock_lock irq_context: softirq rcu_callback &data->lock &obj_hash[i].lock irq_context: softirq rcu_callback &data->lock &base->lock irq_context: softirq rcu_callback &data->lock &base->lock &obj_hash[i].lock irq_context: 0 sk_lock-AF_INET6 &dccp_hashinfo.bhash[i].lock batched_entropy_u8.lock irq_context: 0 sk_lock-AF_INET6 &dccp_hashinfo.bhash[i].lock kfence_freelist_lock irq_context: softirq rcu_read_lock rcu_read_lock &dreq->dreq_lock slock-AF_INET6 &c->lock irq_context: softirq rcu_read_lock rcu_read_lock &dreq->dreq_lock slock-AF_INET6 &____s->seqcount#2 irq_context: softirq rcu_read_lock rcu_read_lock slock-AF_INET6/1 sk_lock-AF_INET6 rcu_read_lock rcu_read_lock rcu_read_lock rcu_read_lock_bh &sch->q.lock &c->lock irq_context: 0 &sb->s_type->i_mutex_key#10 sk_lock-AF_INET6 rcu_read_lock rcu_read_lock rcu_read_lock_bh &sch->q.lock irq_context: 0 &sb->s_type->i_mutex_key#10 sk_lock-AF_INET6 rcu_read_lock rcu_read_lock rcu_read_lock_bh &sch->q.lock &c->lock irq_context: 0 &sb->s_type->i_mutex_key#10 sk_lock-AF_INET6 rcu_read_lock rcu_read_lock rcu_read_lock_bh &sch->q.lock pool_lock#2 irq_context: 0 &sb->s_type->i_mutex_key#10 sk_lock-AF_INET6 rcu_read_lock rcu_read_lock rcu_read_lock_bh &sch->q.lock tk_core.seq.seqcount irq_context: softirq rcu_read_lock &dir->lock irq_context: 0 &sb->s_type->i_mutex_key#10 sk_lock-AF_INET6 rcu_read_lock &rcu_state.gp_wq irq_context: 0 &sb->s_type->i_mutex_key#10 sk_lock-AF_INET6 rcu_read_lock &rcu_state.gp_wq &p->pi_lock irq_context: 0 &sb->s_type->i_mutex_key#10 sk_lock-AF_INET6 rcu_read_lock &rcu_state.gp_wq &p->pi_lock &rq->__lock irq_context: 0 &sb->s_type->i_mutex_key#10 sk_lock-AF_INET6 rcu_read_lock &rcu_state.gp_wq &p->pi_lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 &sb->s_type->i_mutex_key#10 sk_lock-AF_INET6 &data->lock irq_context: 0 rtnl_mutex rcu_read_lock rcu_read_lock &ndev->lock irq_context: 0 rtnl_mutex rcu_read_lock nl_table_lock irq_context: 0 rtnl_mutex rcu_read_lock nl_table_wait.lock irq_context: 0 rtnl_mutex rcu_read_lock &macvlan_netdev_addr_lock_key/1 _xmit_ETHER irq_context: 0 rtnl_mutex rcu_read_lock rcu_read_lock &bond->stats_lock/2 irq_context: 0 rtnl_mutex rcu_read_lock &dev_addr_list_lock_key/2 irq_context: 0 rtnl_mutex rcu_read_lock &dev_addr_list_lock_key/2 rcu_read_lock &macvlan_netdev_addr_lock_key/1 irq_context: 0 rtnl_mutex rcu_read_lock &dev_addr_list_lock_key/2 rcu_read_lock &macvlan_netdev_addr_lock_key/1 _xmit_ETHER irq_context: 0 rtnl_mutex &dev_addr_list_lock_key#2/3 irq_context: 0 rtnl_mutex _xmit_ETHER/4 irq_context: 0 rtnl_mutex nl_table_wait.lock &p->pi_lock &cfs_rq->removed.lock irq_context: 0 rtnl_mutex &idev->mc_lock _xmit_ETHER/4 irq_context: 0 rtnl_mutex &idev->mc_lock _xmit_ETHER/4 &obj_hash[i].lock irq_context: 0 rtnl_mutex &idev->mc_lock _xmit_ETHER/4 krc.lock irq_context: 0 sk_lock-AF_INET6 &data->lock &obj_hash[i].lock irq_context: 0 sk_lock-AF_INET6 &data->lock &base->lock irq_context: 0 sk_lock-AF_INET6 &data->lock &base->lock &obj_hash[i].lock irq_context: softirq (&ndev->rs_timer) rcu_read_lock rcu_read_lock rcu_read_lock_bh dev->qdisc_tx_busylock ?: &qdisc_tx_busylock &data->lock irq_context: softirq (&ndev->rs_timer) rcu_read_lock rcu_read_lock rcu_read_lock_bh &hsr->seqnr_lock &data->lock irq_context: softirq rcu_read_lock rcu_read_lock rcu_read_lock &data->lock irq_context: softirq rcu_read_lock rcu_read_lock rcu_read_lock rcu_read_lock rcu_read_lock &c->lock irq_context: 0 cb_lock genl_mutex &data->lock irq_context: 0 rtnl_mutex &sb->s_type->i_mutex_key#3 per_cpu_ptr(&cgroup_rstat_cpu_lock, cpu) irq_context: 0 rtnl_mutex rcu_read_lock quarantine_lock irq_context: 0 &sb->s_type->i_mutex_key#10 sk_lock-AF_INET6 slock-AF_INET6 quarantine_lock irq_context: 0 &sb->s_type->i_mutex_key#10 sk_lock-AF_INET6 slock-AF_INET6 &____s->seqcount#2 irq_context: 0 &sb->s_type->i_mutex_key#10 sk_lock-AF_INET6 slock-AF_INET6 &____s->seqcount irq_context: 0 &sb->s_type->i_mutex_key#10 &po->bind_lock &obj_hash[i].lock irq_context: 0 rtnl_mutex _xmit_TUNNEL6 pool_lock#2 irq_context: 0 rtnl_mutex _xmit_TUNNEL6 &obj_hash[i].lock irq_context: 0 rtnl_mutex _xmit_TUNNEL6 krc.lock irq_context: 0 sk_lock-AF_INET6 sk_lock-AF_INET6/1 &asoc->wait irq_context: 0 sk_lock-AF_INET6 sk_lock-AF_INET6/1 &asoc->wait &p->pi_lock irq_context: 0 sk_lock-AF_INET6 sk_lock-AF_INET6/1 &asoc->wait &p->pi_lock &rq->__lock irq_context: 0 sk_lock-AF_INET6 sk_lock-AF_INET6/1 &asoc->wait &p->pi_lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 sk_lock-AF_INET6 sk_lock-AF_INET6/1 &rq->__lock irq_context: 0 sk_lock-AF_INET6 sk_lock-AF_INET6/1 &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 &u->iolock &mm->mmap_lock fs_reclaim mmu_notifier_invalidate_range_start &rq->__lock irq_context: 0 &u->iolock &mm->mmap_lock fs_reclaim mmu_notifier_invalidate_range_start &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 &u->iolock &mm->mmap_lock mmu_notifier_invalidate_range_start &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 &pipe->mutex/1 batched_entropy_u32.lock irq_context: 0 &pipe->mutex/1 rtnl_mutex pcpu_alloc_mutex irq_context: 0 &pipe->mutex/1 rtnl_mutex pcpu_alloc_mutex pcpu_lock irq_context: 0 &pipe->mutex/1 rtnl_mutex &n->list_lock irq_context: 0 &pipe->mutex/1 rtnl_mutex &n->list_lock &c->lock irq_context: 0 &pipe->mutex/1 rtnl_mutex &obj_hash[i].lock pool_lock irq_context: 0 &pipe->mutex/1 rtnl_mutex &____s->seqcount#2 irq_context: 0 &pipe->mutex/1 rtnl_mutex &____s->seqcount irq_context: 0 &pipe->mutex/1 rtnl_mutex fill_pool_map-wait-type-override &rq->__lock irq_context: 0 &pipe->mutex/1 rtnl_mutex fill_pool_map-wait-type-override &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 &pipe->mutex/1 rtnl_mutex rlock-AF_NETLINK irq_context: 0 &mm->mmap_lock rcu_read_lock fill_pool_map-wait-type-override pool_lock#2 irq_context: 0 &mm->mmap_lock rcu_read_lock fill_pool_map-wait-type-override &c->lock irq_context: 0 &mm->mmap_lock rcu_read_lock fill_pool_map-wait-type-override &n->list_lock irq_context: 0 &mm->mmap_lock rcu_read_lock fill_pool_map-wait-type-override &n->list_lock &c->lock irq_context: 0 &mm->mmap_lock rcu_read_lock fill_pool_map-wait-type-override pool_lock irq_context: 0 dup_mmap_sem &mm->mmap_lock &mm->mmap_lock/1 &c->lock batched_entropy_u8.lock irq_context: 0 dup_mmap_sem &mm->mmap_lock &mm->mmap_lock/1 &c->lock kfence_freelist_lock irq_context: softirq (&tw->tw_timer) &c->lock irq_context: 0 &pipe->mutex/1 rtnl_mutex batched_entropy_u8.lock irq_context: 0 &pipe->mutex/1 rtnl_mutex kfence_freelist_lock irq_context: 0 &pipe->mutex/1 rtnl_mutex fs_reclaim mmu_notifier_invalidate_range_start &rq->__lock irq_context: 0 &pipe->mutex/1 rtnl_mutex rcu_read_lock pool_lock#2 irq_context: 0 &pipe->mutex/1 rtnl_mutex rcu_read_lock rcu_read_lock_bh &obj_hash[i].lock irq_context: 0 &pipe->mutex/1 rtnl_mutex rcu_read_lock rcu_read_lock_bh pool_lock#2 irq_context: 0 &u->iolock &base->lock irq_context: 0 &u->iolock &base->lock &obj_hash[i].lock irq_context: 0 &net->xfrm.xfrm_cfg_mutex crypto_alg_sem irq_context: 0 &net->xfrm.xfrm_cfg_mutex (kmod_concurrent_max).lock irq_context: 0 &net->xfrm.xfrm_cfg_mutex rcu_read_lock &pool->lock/1 irq_context: 0 &net->xfrm.xfrm_cfg_mutex rcu_read_lock &pool->lock/1 &obj_hash[i].lock irq_context: 0 &net->xfrm.xfrm_cfg_mutex rcu_read_lock &pool->lock/1 pool_lock#2 irq_context: 0 &net->xfrm.xfrm_cfg_mutex rcu_read_lock &pool->lock/1 &p->pi_lock irq_context: 0 &net->xfrm.xfrm_cfg_mutex rcu_read_lock &pool->lock/1 &p->pi_lock &rq->__lock irq_context: 0 &net->xfrm.xfrm_cfg_mutex rcu_read_lock &pool->lock/1 &p->pi_lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 &net->xfrm.xfrm_cfg_mutex &x->wait#17 irq_context: 0 &net->xfrm.xfrm_cfg_mutex running_helpers_waitq.lock irq_context: 0 &net->xfrm.xfrm_cfg_mutex rcu_read_lock &pool->lock/1 &p->pi_lock &cfs_rq->removed.lock irq_context: 0 &net->xfrm.xfrm_cfg_mutex crypto_alg_sem irq_context: 0 &net->xfrm.xfrm_cfg_mutex (crypto_chain).rwsem irq_context: 0 &net->xfrm.xfrm_cfg_mutex (crypto_chain).rwsem fs_reclaim irq_context: 0 &net->xfrm.xfrm_cfg_mutex (crypto_chain).rwsem fs_reclaim mmu_notifier_invalidate_range_start irq_context: 0 &net->xfrm.xfrm_cfg_mutex (crypto_chain).rwsem &c->lock irq_context: 0 &net->xfrm.xfrm_cfg_mutex (crypto_chain).rwsem pool_lock#2 irq_context: 0 &net->xfrm.xfrm_cfg_mutex (crypto_chain).rwsem kthread_create_lock irq_context: 0 &net->xfrm.xfrm_cfg_mutex (crypto_chain).rwsem &p->pi_lock irq_context: 0 &net->xfrm.xfrm_cfg_mutex (crypto_chain).rwsem &p->pi_lock &rq->__lock irq_context: 0 &net->xfrm.xfrm_cfg_mutex (crypto_chain).rwsem &p->pi_lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 &net->xfrm.xfrm_cfg_mutex (crypto_chain).rwsem &rq->__lock irq_context: 0 &net->xfrm.xfrm_cfg_mutex (crypto_chain).rwsem &x->wait irq_context: 0 &net->xfrm.xfrm_cfg_mutex (crypto_chain).rwsem &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 &net->xfrm.xfrm_cfg_mutex (crypto_chain).rwsem &obj_hash[i].lock irq_context: 0 &net->xfrm.xfrm_cfg_mutex &x->wait#21 irq_context: 0 &net->xfrm.xfrm_cfg_mutex &base->lock irq_context: 0 &net->xfrm.xfrm_cfg_mutex &base->lock &obj_hash[i].lock irq_context: 0 &net->xfrm.xfrm_cfg_mutex &cfs_rq->removed.lock irq_context: 0 &net->xfrm.xfrm_cfg_mutex (&timer.timer) irq_context: 0 &net->xfrm.xfrm_cfg_mutex crypto_default_null_skcipher_lock irq_context: 0 &net->xfrm.xfrm_cfg_mutex crypto_default_null_skcipher_lock crypto_alg_sem irq_context: 0 &net->xfrm.xfrm_cfg_mutex crypto_default_null_skcipher_lock fs_reclaim irq_context: 0 &net->xfrm.xfrm_cfg_mutex crypto_default_null_skcipher_lock fs_reclaim mmu_notifier_invalidate_range_start irq_context: 0 &net->xfrm.xfrm_cfg_mutex crypto_default_null_skcipher_lock pool_lock#2 irq_context: 0 &net->xfrm.xfrm_cfg_mutex &net->xfrm.xfrm_state_lock irq_context: 0 &net->xfrm.xfrm_cfg_mutex &net->xfrm.xfrm_state_lock hrtimer_bases.lock irq_context: 0 &net->xfrm.xfrm_cfg_mutex &net->xfrm.xfrm_state_lock hrtimer_bases.lock tk_core.seq.seqcount irq_context: 0 &net->xfrm.xfrm_cfg_mutex &net->xfrm.xfrm_state_lock hrtimer_bases.lock &obj_hash[i].lock irq_context: 0 &net->xfrm.xfrm_cfg_mutex &net->xfrm.xfrm_state_lock &obj_hash[i].lock irq_context: 0 &net->xfrm.xfrm_cfg_mutex &net->xfrm.xfrm_state_lock &base->lock irq_context: 0 &net->xfrm.xfrm_cfg_mutex &net->xfrm.xfrm_state_lock &base->lock &obj_hash[i].lock irq_context: 0 &net->xfrm.xfrm_cfg_mutex rcu_read_lock &data->lock irq_context: softirq &x->lock irq_context: softirq (&x->rtimer) irq_context: softirq (&x->rtimer) &x->lock irq_context: 0 &net->xfrm.xfrm_cfg_mutex &obj_hash[i].lock pool_lock irq_context: 0 (wq_completion)events (work_completion)(&(&nsim_dev->trap_data->trap_report_dw)->work) &devlink->lock_key#6 &obj_hash[i].lock pool_lock irq_context: 0 sb_writers#4 &type->i_mutex_dir_key#3/1 rcu_read_lock rcu_read_lock &obj_hash[i].lock irq_context: 0 &nft_net->commit_mutex fs_reclaim &rq->__lock irq_context: 0 sk_lock-AF_INET6 &dccp_hashinfo.bhash[i].lock &dccp_hashinfo.bhash2[i].lock &n->list_lock irq_context: 0 sk_lock-AF_INET6 &dccp_hashinfo.bhash[i].lock &dccp_hashinfo.bhash2[i].lock &n->list_lock &c->lock irq_context: 0 rtnl_mutex sk_lock-AF_INET &im->lock &c->lock irq_context: 0 rtnl_mutex sk_lock-AF_INET &im->lock &n->list_lock irq_context: 0 rtnl_mutex sk_lock-AF_INET &im->lock &n->list_lock &c->lock irq_context: 0 rtnl_mutex sk_lock-AF_INET &im->lock &base->lock irq_context: 0 rtnl_mutex sk_lock-AF_INET &im->lock &base->lock &obj_hash[i].lock irq_context: 0 &f->f_pos_lock &type->i_mutex_dir_key#3 &rcu_state.expedited_wq irq_context: 0 &f->f_pos_lock &type->i_mutex_dir_key#3 &rcu_state.expedited_wq &p->pi_lock irq_context: 0 &f->f_pos_lock &type->i_mutex_dir_key#3 &rcu_state.expedited_wq &p->pi_lock &rq->__lock irq_context: 0 &f->f_pos_lock &type->i_mutex_dir_key#3 &rcu_state.expedited_wq &p->pi_lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 lock pidmap_lock batched_entropy_u8.lock irq_context: 0 lock pidmap_lock kfence_freelist_lock irq_context: 0 (wq_completion)bat_events (work_completion)(&(&forw_packet_aggr->delayed_work)->work) &hard_iface->bat_iv.ogm_buff_mutex rcu_read_lock fill_pool_map-wait-type-override &____s->seqcount#2 irq_context: 0 pernet_ops_rwsem devices_rwsem rdma_nets_rwsem &device->compat_devs_mutex &rxe->usdev_lock rtnl_mutex &obj_hash[i].lock pool_lock irq_context: 0 pernet_ops_rwsem devices_rwsem rdma_nets_rwsem &device->compat_devs_mutex &rxe->usdev_lock rtnl_mutex &rq->__lock &cfs_rq->removed.lock irq_context: 0 pernet_ops_rwsem ovs_mutex nf_ct_proto_mutex defrag4_mutex fill_pool_map-wait-type-override &rq->__lock irq_context: 0 (wq_completion)bond8#2 (work_completion)(&(&bond->mii_work)->work) rcu_read_lock rcu_node_0 irq_context: 0 (wq_completion)bond8#2 (work_completion)(&(&bond->mii_work)->work) rcu_read_lock &rq->__lock irq_context: 0 rcu_read_lock rcu_read_lock icmp_global.lock irq_context: 0 rcu_read_lock rcu_read_lock icmp_global.lock batched_entropy_u8.lock irq_context: 0 rcu_read_lock rcu_read_lock k-slock-AF_INET6 rcu_read_lock rcu_read_lock rcu_read_lock stock_lock irq_context: 0 rcu_read_lock rcu_read_lock k-slock-AF_INET6 rcu_read_lock rcu_read_lock rcu_read_lock &c->lock irq_context: 0 rcu_read_lock rcu_read_lock k-slock-AF_INET6 rcu_read_lock rcu_read_lock rcu_read_lock pool_lock#2 irq_context: 0 rcu_read_lock rcu_read_lock k-slock-AF_INET6 rcu_read_lock rcu_read_lock rcu_read_lock &dir->lock#2 irq_context: 0 rcu_read_lock rcu_read_lock k-slock-AF_INET6 rcu_read_lock rcu_read_lock rcu_read_lock &____s->seqcount irq_context: 0 rcu_read_lock rcu_read_lock k-slock-AF_INET6 rcu_read_lock rcu_read_lock rcu_read_lock rcu_read_lock pool_lock#2 irq_context: 0 rcu_read_lock rcu_read_lock k-slock-AF_INET6 rcu_read_lock rcu_read_lock rcu_read_lock &obj_hash[i].lock irq_context: 0 rcu_read_lock rcu_read_lock k-slock-AF_INET6 rcu_read_lock &____s->seqcount#15 irq_context: 0 rcu_read_lock rcu_read_lock k-slock-AF_INET6 &(&bp->lock)->lock irq_context: 0 rcu_read_lock rcu_read_lock k-slock-AF_INET6 &(&bp->lock)->lock &____s->seqcount#15 irq_context: 0 rcu_read_lock rcu_read_lock k-slock-AF_INET6 &(&bp->lock)->lock &____s->seqcount#15 &____s->seqcount#2 irq_context: 0 rcu_read_lock rcu_read_lock k-slock-AF_INET6 &(&bp->lock)->lock &____s->seqcount#15 &____s->seqcount irq_context: 0 rcu_read_lock rcu_read_lock k-slock-AF_INET6 &(&bp->lock)->lock &____s->seqcount#15 pool_lock#2 irq_context: 0 rcu_read_lock rcu_read_lock k-slock-AF_INET6 &(&bp->lock)->lock &____s->seqcount#15 &c->lock irq_context: softirq rcu_read_lock rcu_read_lock slock-AF_INET6/1 sk_lock-AF_INET6 rcu_read_lock rcu_read_lock rcu_read_lock rcu_read_lock_bh rcu_read_lock &c->lock irq_context: softirq rcu_read_lock rcu_read_lock slock-AF_INET6/1 sk_lock-AF_INET6 rcu_read_lock rcu_read_lock rcu_read_lock rcu_read_lock_bh rcu_read_lock pool_lock#2 irq_context: softirq rcu_read_lock rcu_read_lock slock-AF_INET6/1 sk_lock-AF_INET6 rcu_read_lock rcu_read_lock rcu_read_lock rcu_read_lock_bh rcu_read_lock tk_core.seq.seqcount irq_context: softirq rcu_read_lock rcu_read_lock slock-AF_INET6/1 sk_lock-AF_INET6 rcu_read_lock rcu_read_lock rcu_read_lock rcu_read_lock_bh rcu_read_lock &obj_hash[i].lock irq_context: softirq rcu_read_lock rcu_read_lock slock-AF_INET6/1 sk_lock-AF_INET6 rcu_read_lock rcu_read_lock rcu_read_lock_bh rcu_read_lock pool_lock#2 irq_context: softirq rcu_read_lock rcu_read_lock slock-AF_INET6/1 sk_lock-AF_INET6 rcu_read_lock rcu_read_lock rcu_read_lock_bh rcu_read_lock tk_core.seq.seqcount irq_context: softirq rcu_read_lock rcu_read_lock slock-AF_INET6/1 sk_lock-AF_INET6 rcu_read_lock rcu_read_lock rcu_read_lock_bh rcu_read_lock &obj_hash[i].lock irq_context: 0 &sb->s_type->i_mutex_key#10 sk_lock-AF_INET6 slock-AF_INET6 rcu_read_lock rcu_read_lock rcu_read_lock_bh rcu_read_lock pool_lock#2 irq_context: 0 &sb->s_type->i_mutex_key#10 sk_lock-AF_INET6 slock-AF_INET6 rcu_read_lock rcu_read_lock rcu_read_lock_bh rcu_read_lock tk_core.seq.seqcount irq_context: 0 &sb->s_type->i_mutex_key#10 sk_lock-AF_INET6 slock-AF_INET6 rcu_read_lock rcu_read_lock rcu_read_lock_bh rcu_read_lock &obj_hash[i].lock irq_context: 0 rcu_read_lock rcu_read_lock_bh rcu_read_lock &data->lock irq_context: 0 &sb->s_type->i_mutex_key#8 &sbi->s_writepages_rwsem rcu_read_lock &rcu_state.gp_wq &p->pi_lock &cfs_rq->removed.lock irq_context: 0 &sb->s_type->i_mutex_key#8 &sbi->s_writepages_rwsem rcu_read_lock pgd_lock irq_context: 0 &sb->s_type->i_mutex_key#8 &sbi->s_writepages_rwsem rcu_read_lock stock_lock irq_context: 0 &sb->s_type->i_mutex_key#8 &sbi->s_writepages_rwsem rcu_read_lock rcu_read_lock pool_lock#2 irq_context: 0 &sb->s_type->i_mutex_key#8 &sbi->s_writepages_rwsem rcu_read_lock key irq_context: 0 &sb->s_type->i_mutex_key#8 &sbi->s_writepages_rwsem rcu_read_lock pcpu_lock irq_context: 0 &sb->s_type->i_mutex_key#8 &sbi->s_writepages_rwsem rcu_read_lock percpu_counters_lock irq_context: 0 &sb->s_type->i_mutex_key#8 &sbi->s_writepages_rwsem rcu_read_lock pcpu_lock stock_lock irq_context: 0 (wq_completion)bat_events (work_completion)(&(&forw_packet_aggr->delayed_work)->work) rcu_read_lock_bh rcu_read_lock &data->lock irq_context: softirq (&n->timer) k-slock-AF_INET6 rcu_read_lock rcu_read_lock rcu_read_lock_bh rcu_read_lock &data->lock irq_context: softirq (&hsr->announce_timer) rcu_read_lock &hsr->seqnr_lock rcu_read_lock rcu_read_lock_bh rcu_read_lock &data->lock irq_context: 0 sb_writers#4 &type->i_mutex_dir_key#3/1 jbd2_handle pgd_lock irq_context: 0 sb_writers#4 &type->i_mutex_dir_key#3/1 jbd2_handle key irq_context: 0 sb_writers#4 &type->i_mutex_dir_key#3/1 jbd2_handle pcpu_lock irq_context: 0 sb_writers#4 &type->i_mutex_dir_key#3/1 jbd2_handle percpu_counters_lock irq_context: 0 sb_writers#4 &type->i_mutex_dir_key#3/1 jbd2_handle pcpu_lock stock_lock irq_context: 0 bpf_stats_enabled_mutex cpu_hotplug_lock jump_label_mutex text_mutex rcu_node_0 irq_context: 0 &sb->s_type->i_mutex_key#10 sk_lock-AF_INET6 slock-AF_INET6 rcu_read_lock rcu_read_lock rcu_read_lock_bh rcu_read_lock &c->lock irq_context: 0 sk_lock-AF_INET6 rcu_read_lock rcu_read_lock rcu_read_lock rcu_read_lock rcu_read_lock_bh &data->lock irq_context: 0 sk_lock-AF_INET6 rcu_read_lock &ei->socket.wq.wait &p->pi_lock &rq->__lock &cfs_rq->removed.lock irq_context: 0 sb_writers#4 sb_internal jbd2_handle &p->pi_lock &rq->__lock &cfs_rq->removed.lock irq_context: 0 &xt[i].mutex rcu_read_lock stock_lock irq_context: 0 &xt[i].mutex rcu_read_lock pcpu_lock stock_lock irq_context: 0 (wq_completion)wg-kex-wg2#2 (work_completion)(&({ do { const void *__vpp_verify = (typeof((worker) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __asm__ ("" : "=r"(__ptr) : "0"((typeof(*((worker))) *)((worker)))); (typeof((typeof(*((worker))) *)((worker)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); })->work) &rq->__lock irq_context: 0 (wq_completion)wg-kex-wg2#2 (work_completion)(&({ do { const void *__vpp_verify = (typeof((worker) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __asm__ ("" : "=r"(__ptr) : "0"((typeof(*((worker))) *)((worker)))); (typeof((typeof(*((worker))) *)((worker)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); })->work) &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)wg-kex-wg2#2 (work_completion)(&({ do { const void *__vpp_verify = (typeof((worker) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __asm__ ("" : "=r"(__ptr) : "0"((typeof(*((worker))) *)((worker)))); (typeof((typeof(*((worker))) *)((worker)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); })->work) &n->list_lock irq_context: 0 (wq_completion)wg-kex-wg2#2 (work_completion)(&({ do { const void *__vpp_verify = (typeof((worker) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __asm__ ("" : "=r"(__ptr) : "0"((typeof(*((worker))) *)((worker)))); (typeof((typeof(*((worker))) *)((worker)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); })->work) &n->list_lock &c->lock irq_context: 0 (wq_completion)wg-kex-wg2#2 (work_completion)(&({ do { const void *__vpp_verify = (typeof((worker) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __asm__ ("" : "=r"(__ptr) : "0"((typeof(*((worker))) *)((worker)))); (typeof((typeof(*((worker))) *)((worker)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); })->work) &peer->endpoint_lock rcu_read_lock_bh batched_entropy_u32.lock irq_context: 0 nfnl_subsys_cthelper (console_sem).lock irq_context: 0 nfnl_subsys_cthelper console_lock console_srcu console_owner_lock irq_context: 0 nfnl_subsys_cthelper console_lock console_srcu console_owner irq_context: 0 nfnl_subsys_cthelper console_lock console_srcu console_owner &port_lock_key irq_context: 0 nfnl_subsys_cthelper console_lock console_srcu console_owner console_owner_lock irq_context: 0 nfnl_subsys_cthelper &rq->__lock irq_context: 0 nfnl_subsys_cthelper &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 nfnl_subsys_cthelper &lock->wait_lock irq_context: 0 sk_lock-AF_INET6 &rq->__lock &obj_hash[i].lock irq_context: 0 sk_lock-AF_INET6 &rq->__lock &base->lock irq_context: 0 sk_lock-AF_INET6 &rq->__lock &base->lock &obj_hash[i].lock irq_context: 0 tomoyo_ss remove_cache_srcu rcu_read_lock &cfs_rq->removed.lock irq_context: 0 tomoyo_ss remove_cache_srcu rcu_read_lock &obj_hash[i].lock irq_context: 0 pernet_ops_rwsem rtnl_mutex rcu_state.exp_mutex &rcu_state.expedited_wq &p->pi_lock &rq->__lock irq_context: 0 pernet_ops_rwsem rtnl_mutex rcu_state.exp_mutex &rcu_state.expedited_wq &p->pi_lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: softirq (&hc->tx_rtotimer) per_cpu_ptr(&cgroup_rstat_cpu_lock, cpu) irq_context: softirq (&hc->tx_rtotimer) &pcp->lock &zone->lock irq_context: softirq (&hc->tx_rtotimer) &pcp->lock &zone->lock &____s->seqcount irq_context: 0 (wq_completion)bat_events (work_completion)(&(&bat_priv->tt.work)->work) fill_pool_map-wait-type-override pool_lock#2 irq_context: 0 (wq_completion)bat_events (work_completion)(&(&bat_priv->tt.work)->work) fill_pool_map-wait-type-override pool_lock irq_context: softirq rcu_read_lock rcu_read_lock &dreq->dreq_lock slock-AF_INET6 fill_pool_map-wait-type-override pool_lock#2 irq_context: softirq rcu_read_lock rcu_read_lock &dreq->dreq_lock slock-AF_INET6 fill_pool_map-wait-type-override &c->lock irq_context: softirq rcu_read_lock rcu_read_lock &dreq->dreq_lock slock-AF_INET6 fill_pool_map-wait-type-override &n->list_lock irq_context: softirq rcu_read_lock rcu_read_lock &dreq->dreq_lock slock-AF_INET6 fill_pool_map-wait-type-override &n->list_lock &c->lock irq_context: softirq rcu_read_lock rcu_read_lock &dreq->dreq_lock slock-AF_INET6 fill_pool_map-wait-type-override pool_lock irq_context: 0 rcu_read_lock_bh dev->qdisc_tx_busylock ?: &qdisc_tx_busylock _xmit_NETROM rcu_read_lock &n->list_lock irq_context: 0 rcu_read_lock_bh dev->qdisc_tx_busylock ?: &qdisc_tx_busylock _xmit_NETROM rcu_read_lock &n->list_lock &c->lock irq_context: 0 rcu_read_lock_bh dev->qdisc_tx_busylock ?: &qdisc_tx_busylock _xmit_NETROM rcu_read_lock &____s->seqcount#2 irq_context: 0 rcu_read_lock_bh dev->qdisc_tx_busylock ?: &qdisc_tx_busylock _xmit_NETROM rcu_read_lock &____s->seqcount irq_context: 0 sb_writers#4 &sb->s_type->i_mutex_key#8 &sb->s_type->i_mutex_key#8/4 jbd2_handle tk_core.seq.seqcount irq_context: 0 sb_writers#4 &sb->s_type->i_mutex_key#8 &sb->s_type->i_mutex_key#8/4 jbd2_handle &dd->lock irq_context: 0 sb_writers#4 &sb->s_type->i_mutex_key#8 &sb->s_type->i_mutex_key#8/4 jbd2_handle rcu_read_lock &pool->lock irq_context: 0 sb_writers#4 &sb->s_type->i_mutex_key#8 &sb->s_type->i_mutex_key#8/4 jbd2_handle rcu_read_lock &pool->lock &obj_hash[i].lock irq_context: 0 sb_writers#4 &sb->s_type->i_mutex_key#8 &sb->s_type->i_mutex_key#8/4 jbd2_handle rcu_read_lock &pool->lock pool_lock#2 irq_context: 0 sb_writers#4 &sb->s_type->i_mutex_key#8 &sb->s_type->i_mutex_key#8/4 jbd2_handle rcu_read_lock &pool->lock &p->pi_lock irq_context: 0 sb_writers#4 &sb->s_type->i_mutex_key#8 &sb->s_type->i_mutex_key#8/4 jbd2_handle rcu_read_lock &pool->lock &p->pi_lock &rq->__lock irq_context: 0 sb_writers#4 &sb->s_type->i_mutex_key#8 &sb->s_type->i_mutex_key#8/4 jbd2_handle rcu_read_lock &pool->lock &p->pi_lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 sb_writers#4 &sb->s_type->i_mutex_key#8 &sb->s_type->i_mutex_key#8/4 jbd2_handle bit_wait_table + i irq_context: 0 sb_writers#4 &sb->s_type->i_mutex_key#8 &sb->s_type->i_mutex_key#8/4 rcu_read_lock rcu_node_0 irq_context: 0 pernet_ops_rwsem remove_cache_srcu rcu_read_lock &____s->seqcount irq_context: 0 pernet_ops_rwsem rtnl_mutex &caifn->caifdevs.lock quarantine_lock irq_context: 0 (wq_completion)wg-crypt-wg0#2 (work_completion)(&peer->transmit_packet_work) &peer->endpoint_lock rcu_read_lock_bh rcu_read_lock rcu_read_lock_bh rcu_read_lock &____s->seqcount#2 irq_context: 0 (wq_completion)events (work_completion)(&(&nsim_dev->trap_data->trap_report_dw)->work) rcu_node_0 irq_context: 0 (wq_completion)wg-crypt-wg0#2 (work_completion)(&peer->transmit_packet_work) &peer->endpoint_lock rcu_read_lock_bh rcu_read_lock rcu_read_lock_bh rcu_read_lock &____s->seqcount irq_context: 0 pernet_ops_rwsem pcpu_alloc_mutex pool_lock#2 irq_context: 0 sb_writers#4 &sb->s_type->i_mutex_key#8 &sb->s_type->i_mutex_key#8/4 &journal->j_state_lock &journal->j_wait_commit &p->pi_lock &cfs_rq->removed.lock irq_context: 0 sb_writers#4 &sb->s_type->i_mutex_key#8 &sb->s_type->i_mutex_key#8/4 &ei->i_data_sem &ei->i_data_sem/1 &n->list_lock irq_context: 0 sb_writers#4 &sb->s_type->i_mutex_key#8 &sb->s_type->i_mutex_key#8/4 &ei->i_data_sem &ei->i_data_sem/1 &n->list_lock &c->lock irq_context: softirq (&hc->tx_rtotimer) quarantine_lock irq_context: 0 (wq_completion)events_unbound (work_completion)(&pool->idle_cull_work) &rq->__lock irq_context: 0 (wq_completion)events_unbound (work_completion)(&pool->idle_cull_work) &cfs_rq->removed.lock irq_context: 0 (wq_completion)events_unbound (work_completion)(&pool->idle_cull_work) &obj_hash[i].lock irq_context: 0 (wq_completion)events_unbound (work_completion)(&pool->idle_cull_work) pool_lock#2 irq_context: 0 sb_writers#4 &sb->s_type->i_mutex_key#8 &sb->s_type->i_mutex_key#8/4 pgd_lock irq_context: 0 sb_writers#4 &sb->s_type->i_mutex_key#8 &sb->s_type->i_mutex_key#8/4 stock_lock irq_context: 0 sb_writers#4 &sb->s_type->i_mutex_key#8 &sb->s_type->i_mutex_key#8/4 rcu_read_lock pool_lock#2 irq_context: 0 sb_writers#4 &sb->s_type->i_mutex_key#8 &sb->s_type->i_mutex_key#8/4 key irq_context: 0 sb_writers#4 &sb->s_type->i_mutex_key#8 &sb->s_type->i_mutex_key#8/4 pcpu_lock irq_context: 0 sb_writers#4 &sb->s_type->i_mutex_key#8 &sb->s_type->i_mutex_key#8/4 percpu_counters_lock irq_context: 0 sb_writers#4 &sb->s_type->i_mutex_key#8 &sb->s_type->i_mutex_key#8/4 pcpu_lock stock_lock irq_context: 0 (wq_completion)bat_events (work_completion)(&(&forw_packet_aggr->delayed_work)->work) &hard_iface->bat_iv.ogm_buff_mutex rcu_read_lock &rcu_state.gp_wq &p->pi_lock &cfs_rq->removed.lock irq_context: softirq rcu_read_lock rcu_read_lock &dreq->dreq_lock slock-AF_INET6 &n->list_lock irq_context: softirq rcu_read_lock rcu_read_lock &dreq->dreq_lock slock-AF_INET6 &n->list_lock &c->lock irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem k-sk_lock-AF_TIPC &tn->nametbl_lock &service->lock &obj_hash[i].lock pool_lock irq_context: 0 bpf_stats_enabled_mutex batched_entropy_u8.lock irq_context: 0 bpf_stats_enabled_mutex kfence_freelist_lock irq_context: 0 bpf_stats_enabled_mutex fs_reclaim mmu_notifier_invalidate_range_start &rq->__lock irq_context: 0 bpf_stats_enabled_mutex fs_reclaim mmu_notifier_invalidate_range_start &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 &f->f_pos_lock sb_writers#4 &sb->s_type->i_mutex_key#8 rcu_read_lock rcu_read_lock rcu_read_lock &cfs_rq->removed.lock irq_context: 0 &f->f_pos_lock sb_writers#4 &sb->s_type->i_mutex_key#8 rcu_read_lock rcu_read_lock rcu_read_lock &obj_hash[i].lock irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem rcu_read_lock &rcu_state.gp_wq &p->pi_lock &cfs_rq->removed.lock irq_context: 0 sb_writers#4 &type->i_mutex_dir_key#3 fill_pool_map-wait-type-override &n->list_lock irq_context: 0 sb_writers#4 &type->i_mutex_dir_key#3 fill_pool_map-wait-type-override &n->list_lock &c->lock irq_context: 0 &f->f_pos_lock sb_writers#4 &sb->s_type->i_mutex_key#8 fill_pool_map-wait-type-override pool_lock#2 irq_context: 0 &f->f_pos_lock sb_writers#4 &sb->s_type->i_mutex_key#8 fill_pool_map-wait-type-override &c->lock irq_context: 0 &f->f_pos_lock sb_writers#4 &sb->s_type->i_mutex_key#8 fill_pool_map-wait-type-override pool_lock irq_context: 0 &iint->mutex sb_writers#4 jbd2_handle bit_wait_table + i irq_context: 0 &iint->mutex sb_writers#4 jbd2_handle &rq->__lock irq_context: 0 &iint->mutex sb_writers#4 jbd2_handle &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 &iint->mutex mapping.invalidate_lock &xa->xa_lock#9 &____s->seqcount#2 irq_context: 0 &iint->mutex sb_writers#4 &rq->__lock irq_context: 0 &iint->mutex sb_writers#4 &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 &tfile->socket.wq.wait irq_context: 0 sb_writers#4 &sb->s_type->i_mutex_key#8 mapping.invalidate_lock jbd2_handle stock_lock irq_context: 0 sb_writers#4 &sb->s_type->i_mutex_key#8 mapping.invalidate_lock jbd2_handle &xa->xa_lock#9 irq_context: 0 sb_writers#4 &sb->s_type->i_mutex_key#8 mapping.invalidate_lock jbd2_handle lock#4 irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem &ht->mutex &meta->lock irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem &ht->mutex kfence_freelist_lock irq_context: 0 sb_writers#4 &sb->s_type->i_mutex_key#8 mapping.invalidate_lock jbd2_handle &dd->lock irq_context: 0 sb_writers#4 &sb->s_type->i_mutex_key#8 mapping.invalidate_lock jbd2_handle rcu_read_lock &pool->lock irq_context: 0 sb_writers#4 &sb->s_type->i_mutex_key#8 mapping.invalidate_lock jbd2_handle rcu_read_lock &pool->lock &obj_hash[i].lock irq_context: 0 sb_writers#4 &sb->s_type->i_mutex_key#8 mapping.invalidate_lock jbd2_handle rcu_read_lock &pool->lock pool_lock#2 irq_context: 0 sb_writers#4 &sb->s_type->i_mutex_key#8 mapping.invalidate_lock jbd2_handle rcu_read_lock &pool->lock &p->pi_lock irq_context: 0 sb_writers#4 &sb->s_type->i_mutex_key#8 mapping.invalidate_lock jbd2_handle rcu_read_lock &pool->lock &p->pi_lock &rq->__lock irq_context: 0 sb_writers#4 &sb->s_type->i_mutex_key#8 mapping.invalidate_lock jbd2_handle rcu_read_lock &pool->lock &p->pi_lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 sb_writers#4 &sb->s_type->i_mutex_key#8 mapping.invalidate_lock jbd2_handle rcu_read_lock &memcg->move_lock irq_context: 0 sb_writers#4 &sb->s_type->i_mutex_key#8 mapping.invalidate_lock jbd2_handle rcu_read_lock &xa->xa_lock#9 irq_context: 0 (wq_completion)ipv6_addrconf (work_completion)(&(&ifa->dad_work)->work) rtnl_mutex &idev->mc_lock rcu_read_lock rcu_read_lock rcu_read_lock_bh dev->qdisc_tx_busylock ?: &qdisc_tx_busylock rcu_read_lock rcu_read_lock &tfile->socket.wq.wait irq_context: 0 (wq_completion)ipv6_addrconf (work_completion)(&(&ifa->dad_work)->work) rtnl_mutex &idev->mc_lock rcu_read_lock rcu_read_lock rcu_read_lock_bh dev->qdisc_tx_busylock ?: &qdisc_tx_busylock rcu_read_lock rcu_read_lock &tfile->socket.wq.wait &p->pi_lock irq_context: 0 (wq_completion)ipv6_addrconf (work_completion)(&(&ifa->dad_work)->work) rtnl_mutex &idev->mc_lock rcu_read_lock rcu_read_lock rcu_read_lock_bh dev->qdisc_tx_busylock ?: &qdisc_tx_busylock rcu_read_lock rcu_read_lock &tfile->socket.wq.wait &p->pi_lock &rq->__lock irq_context: 0 (wq_completion)ipv6_addrconf (work_completion)(&(&ifa->dad_work)->work) rtnl_mutex &idev->mc_lock rcu_read_lock rcu_read_lock rcu_read_lock_bh dev->qdisc_tx_busylock ?: &qdisc_tx_busylock rcu_read_lock rcu_read_lock &tfile->socket.wq.wait &p->pi_lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)ipv6_addrconf (work_completion)(&(&ifa->dad_work)->work) rtnl_mutex rcu_read_lock rcu_read_lock rcu_read_lock_bh dev->qdisc_tx_busylock ?: &qdisc_tx_busylock rcu_read_lock rcu_read_lock &tfile->socket.wq.wait irq_context: 0 (wq_completion)ipv6_addrconf (work_completion)(&(&ifa->dad_work)->work) rtnl_mutex rcu_read_lock rcu_read_lock rcu_read_lock_bh dev->qdisc_tx_busylock ?: &qdisc_tx_busylock rcu_read_lock rcu_read_lock &tfile->socket.wq.wait &p->pi_lock irq_context: 0 (wq_completion)mld (work_completion)(&(&idev->mc_ifc_work)->work) &idev->mc_lock rcu_read_lock rcu_read_lock rcu_read_lock_bh dev->qdisc_tx_busylock ?: &qdisc_tx_busylock rcu_read_lock &data->lock irq_context: 0 (wq_completion)mld (work_completion)(&(&idev->mc_ifc_work)->work) &idev->mc_lock rcu_read_lock rcu_read_lock rcu_read_lock_bh dev->qdisc_tx_busylock ?: &qdisc_tx_busylock rcu_read_lock &data->lock &obj_hash[i].lock irq_context: 0 (wq_completion)mld (work_completion)(&(&idev->mc_ifc_work)->work) &idev->mc_lock rcu_read_lock rcu_read_lock rcu_read_lock_bh dev->qdisc_tx_busylock ?: &qdisc_tx_busylock rcu_read_lock &data->lock &base->lock irq_context: 0 (wq_completion)mld (work_completion)(&(&idev->mc_ifc_work)->work) &idev->mc_lock rcu_read_lock rcu_read_lock rcu_read_lock_bh dev->qdisc_tx_busylock ?: &qdisc_tx_busylock rcu_read_lock &data->lock &base->lock &obj_hash[i].lock irq_context: 0 &f->f_pos_lock sb_writers#4 &sbi->s_writepages_rwsem jbd2_handle rcu_read_lock &cfs_rq->removed.lock irq_context: 0 rtnl_mutex pcpu_alloc_mutex &rq->__lock &cfs_rq->removed.lock irq_context: 0 pidmap_lock &obj_hash[i].lock irq_context: 0 pidmap_lock pool_lock#2 irq_context: 0 &f->f_pos_lock sb_writers#4 &sbi->s_writepages_rwsem jbd2_handle rcu_read_lock rcu_read_lock rcu_node_0 irq_context: 0 &f->f_pos_lock sb_writers#4 &sbi->s_writepages_rwsem jbd2_handle rcu_read_lock rcu_read_lock &rq->__lock irq_context: 0 &f->f_pos_lock sb_writers#4 &sbi->s_writepages_rwsem jbd2_handle rcu_read_lock rcu_read_lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 &f->f_pos_lock sb_writers#4 &sbi->s_writepages_rwsem jbd2_handle rcu_read_lock &rcu_state.gp_wq irq_context: 0 &f->f_pos_lock sb_writers#4 &sbi->s_writepages_rwsem jbd2_handle rcu_read_lock &rcu_state.gp_wq &p->pi_lock irq_context: 0 &f->f_pos_lock sb_writers#4 &sbi->s_writepages_rwsem jbd2_handle rcu_read_lock &rcu_state.gp_wq &p->pi_lock &rq->__lock irq_context: 0 &f->f_pos_lock sb_writers#4 &sbi->s_writepages_rwsem jbd2_handle rcu_read_lock &rcu_state.gp_wq &p->pi_lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 &f->f_pos_lock sb_writers#4 &sbi->s_writepages_rwsem rcu_read_lock &____s->seqcount#2 irq_context: 0 &f->f_pos_lock sb_writers#4 &sbi->s_writepages_rwsem rcu_read_lock &____s->seqcount irq_context: 0 lweventlist_lock &obj_hash[i].lock irq_context: 0 &f->f_pos_lock sb_writers#11 remove_cache_srcu irq_context: 0 &f->f_pos_lock sb_writers#11 remove_cache_srcu quarantine_lock irq_context: 0 &f->f_pos_lock sb_writers#11 remove_cache_srcu &c->lock irq_context: 0 &f->f_pos_lock sb_writers#11 remove_cache_srcu &n->list_lock irq_context: 0 &f->f_pos_lock sb_writers#11 remove_cache_srcu &obj_hash[i].lock irq_context: 0 &f->f_pos_lock sb_writers#11 remove_cache_srcu &rq->__lock irq_context: 0 &f->f_pos_lock sb_writers#11 remove_cache_srcu &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 sb_writers#4 &type->i_mutex_dir_key#3/1 &type->i_mutex_dir_key#3 jbd2_handle &sbi->s_orphan_lock &lock->wait_lock irq_context: 0 sb_writers#4 &type->i_mutex_dir_key#3/1 jbd2_handle &ei->i_data_sem remove_cache_srcu rcu_read_lock &rcu_state.gp_wq irq_context: 0 sb_writers#4 &type->i_mutex_dir_key#3/1 jbd2_handle &ei->i_data_sem remove_cache_srcu rcu_read_lock &rcu_state.gp_wq &p->pi_lock irq_context: 0 sb_writers#4 &type->i_mutex_dir_key#3/1 jbd2_handle &ei->i_data_sem remove_cache_srcu rcu_read_lock &rcu_state.gp_wq &p->pi_lock &rq->__lock irq_context: 0 sb_writers#4 &type->i_mutex_dir_key#3/1 jbd2_handle &ei->i_data_sem remove_cache_srcu rcu_read_lock &rcu_state.gp_wq &p->pi_lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: softirq rcu_read_lock rcu_read_lock slock-AF_INET6/1 sk_lock-AF_INET6 batched_entropy_u8.lock irq_context: softirq rcu_read_lock rcu_read_lock slock-AF_INET6/1 sk_lock-AF_INET6 kfence_freelist_lock irq_context: 0 &f->f_pos_lock sb_writers#4 &sbi->s_writepages_rwsem jbd2_handle rcu_read_lock &obj_hash[i].lock pool_lock irq_context: 0 sb_writers#11 &type->i_mutex_dir_key#7/1 rcu_node_0 irq_context: 0 sk_lock-AF_BLUETOOTH-BTPROTO_HCI &data->lock irq_context: softirq rcu_read_lock rcu_read_lock slock-AF_INET/1 &data->lock irq_context: softirq rcu_read_lock rcu_read_lock slock-AF_INET/1 &data->lock &obj_hash[i].lock irq_context: softirq rcu_read_lock rcu_read_lock slock-AF_INET/1 &data->lock &base->lock irq_context: softirq rcu_read_lock rcu_read_lock slock-AF_INET/1 &data->lock &base->lock &obj_hash[i].lock irq_context: 0 pernet_ops_rwsem key#25 irq_context: 0 rtnl_mutex &idev->mc_lock &macvlan_netdev_addr_lock_key/2 irq_context: 0 rtnl_mutex &idev->mc_lock &macvlan_netdev_addr_lock_key/2 &obj_hash[i].lock irq_context: 0 rtnl_mutex &idev->mc_lock &macvlan_netdev_addr_lock_key/2 krc.lock irq_context: 0 rtnl_mutex &macvlan_netdev_addr_lock_key/2 irq_context: softirq (&ndev->rs_timer) rcu_read_lock rcu_read_lock rcu_read_lock_bh dev->qdisc_tx_busylock ?: &qdisc_tx_busylock _xmit_ETHER#2 &data->lock irq_context: softirq (&hc->tx_rtotimer) &meta->lock irq_context: softirq (&hc->tx_rtotimer) kfence_freelist_lock irq_context: 0 &tfile->napi_mutex rcu_read_lock rcu_read_lock pcpu_lock irq_context: 0 &tfile->napi_mutex rcu_read_lock rcu_read_lock &data->lock irq_context: 0 &sb->s_type->i_mutex_key#10 &wq->mutex &rq->__lock irq_context: 0 pack_mutex pack_mutex.wait_lock irq_context: 0 (wq_completion)events (work_completion)(&aux->work) pack_mutex.wait_lock irq_context: 0 &sb->s_type->i_mutex_key#10 sk_lock-AF_INET6 slock-AF_INET6 &data->lock &obj_hash[i].lock irq_context: 0 &sb->s_type->i_mutex_key#10 sk_lock-AF_INET6 slock-AF_INET6 &data->lock &base->lock irq_context: 0 &sb->s_type->i_mutex_key#10 sk_lock-AF_INET6 slock-AF_INET6 &data->lock &base->lock &obj_hash[i].lock irq_context: 0 kn->active#59 &n->list_lock irq_context: 0 kn->active#59 &n->list_lock &c->lock irq_context: 0 kn->active#59 remove_cache_srcu irq_context: 0 kn->active#59 remove_cache_srcu quarantine_lock irq_context: 0 kn->active#59 remove_cache_srcu &c->lock irq_context: 0 kn->active#59 remove_cache_srcu &n->list_lock irq_context: 0 kn->active#59 remove_cache_srcu &obj_hash[i].lock irq_context: 0 kn->active#59 remove_cache_srcu &rq->__lock irq_context: 0 kn->active#59 remove_cache_srcu &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 kn->active#59 rcu_node_0 irq_context: 0 kn->active#59 rcu_read_lock rcu_node_0 irq_context: 0 remove_cache_srcu rcu_read_lock &rcu_state.expedited_wq &p->pi_lock &rq->__lock &cfs_rq->removed.lock irq_context: 0 (wq_completion)mld (work_completion)(&(&idev->mc_ifc_work)->work) &idev->mc_lock rcu_read_lock rcu_read_lock rcu_read_lock_bh &data->lock &obj_hash[i].lock irq_context: 0 (wq_completion)mld (work_completion)(&(&idev->mc_ifc_work)->work) &idev->mc_lock rcu_read_lock rcu_read_lock rcu_read_lock_bh &data->lock &base->lock irq_context: 0 (wq_completion)mld (work_completion)(&(&idev->mc_ifc_work)->work) &idev->mc_lock rcu_read_lock rcu_read_lock rcu_read_lock_bh &data->lock &base->lock &obj_hash[i].lock irq_context: softirq (&in_dev->mr_ifc_timer) rcu_read_lock rcu_read_lock_bh &data->lock irq_context: 0 (wq_completion)phy13 (work_completion)(&(&local->scan_work)->work) irq_context: 0 (wq_completion)phy13 (work_completion)(&(&local->scan_work)->work) &local->mtx irq_context: 0 (wq_completion)phy13 (work_completion)(&(&local->scan_work)->work) &local->mtx &local->iflist_mtx irq_context: 0 (wq_completion)phy13 (work_completion)(&(&local->scan_work)->work) &local->mtx &data->mutex irq_context: 0 (wq_completion)phy13 (work_completion)(&(&local->scan_work)->work) &local->mtx &obj_hash[i].lock irq_context: 0 (wq_completion)phy13 (work_completion)(&(&local->scan_work)->work) &local->mtx &base->lock irq_context: 0 (wq_completion)phy13 (work_completion)(&(&local->scan_work)->work) &local->mtx &base->lock &obj_hash[i].lock irq_context: 0 (wq_completion)phy13 (work_completion)(&(&local->scan_work)->work) &local->mtx &rq->__lock irq_context: 0 sb_writers#4 &sb->s_type->i_mutex_key#8 mapping.invalidate_lock jbd2_handle &sbi->s_orphan_lock pool_lock#2 irq_context: 0 sb_writers#4 &sb->s_type->i_mutex_key#8 mapping.invalidate_lock jbd2_handle &sbi->s_orphan_lock &rq->__lock &cfs_rq->removed.lock irq_context: softirq rcu_read_lock rcu_read_lock slock-AF_INET6/1 sk_lock-AF_INET6 rcu_read_lock batched_entropy_u8.lock irq_context: softirq rcu_read_lock rcu_read_lock slock-AF_INET6/1 sk_lock-AF_INET6 rcu_read_lock kfence_freelist_lock irq_context: 0 &sbi->s_writepages_rwsem rcu_read_lock rcu_read_lock &pool->lock irq_context: 0 &sbi->s_writepages_rwsem rcu_read_lock rcu_read_lock &pool->lock &obj_hash[i].lock irq_context: 0 &sbi->s_writepages_rwsem rcu_read_lock rcu_read_lock &pool->lock pool_lock#2 irq_context: 0 &sbi->s_writepages_rwsem rcu_read_lock rcu_read_lock &pool->lock &p->pi_lock irq_context: 0 &sbi->s_writepages_rwsem rcu_read_lock rcu_read_lock &pool->lock &p->pi_lock &rq->__lock irq_context: 0 &sbi->s_writepages_rwsem rcu_read_lock rcu_read_lock &pool->lock &p->pi_lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 sb_writers#4 &sb->s_type->i_mutex_key#8 mapping.invalidate_lock rcu_read_lock &rcu_state.gp_wq &p->pi_lock &cfs_rq->removed.lock irq_context: softirq rcu_read_lock &local->rx_path_lock &rdev->beacon_registrations_lock irq_context: softirq rcu_read_lock &local->rx_path_lock &data->lock irq_context: 0 sb_writers#4 &type->i_mutex_dir_key#3/1 rcu_read_lock &rcu_state.gp_wq &p->pi_lock &cfs_rq->removed.lock irq_context: softirq (&peer->timer_send_keepalive) rcu_read_lock_bh rcu_read_lock &pool->lock &p->pi_lock &rq->__lock &cfs_rq->removed.lock irq_context: 0 sk_lock-AF_INET6 rcu_read_lock rcu_read_lock batched_entropy_u8.lock irq_context: 0 sk_lock-AF_INET6 rcu_read_lock rcu_read_lock kfence_freelist_lock irq_context: 0 pernet_ops_rwsem devices_rwsem rdma_nets_rwsem &device->compat_devs_mutex rcu_read_lock &rcu_state.gp_wq &p->pi_lock &cfs_rq->removed.lock irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem &dir->lock#2 quarantine_lock irq_context: 0 pernet_ops_rwsem rtnl_mutex &data->lock &obj_hash[i].lock irq_context: 0 pernet_ops_rwsem rtnl_mutex &data->lock &base->lock irq_context: 0 pernet_ops_rwsem devices_rwsem rdma_nets_rwsem &device->compat_devs_mutex fs_reclaim rcu_node_0 irq_context: 0 pernet_ops_rwsem devices_rwsem rdma_nets_rwsem &device->compat_devs_mutex fs_reclaim &rcu_state.expedited_wq irq_context: 0 pernet_ops_rwsem rtnl_mutex &data->lock &base->lock &obj_hash[i].lock irq_context: 0 pernet_ops_rwsem devices_rwsem rdma_nets_rwsem &device->compat_devs_mutex fs_reclaim &rcu_state.expedited_wq &p->pi_lock irq_context: 0 pernet_ops_rwsem devices_rwsem rdma_nets_rwsem &device->compat_devs_mutex fs_reclaim &rcu_state.expedited_wq &p->pi_lock &rq->__lock irq_context: 0 pernet_ops_rwsem devices_rwsem rdma_nets_rwsem &device->compat_devs_mutex fs_reclaim &rcu_state.expedited_wq &p->pi_lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)phy13 (work_completion)(&(&local->scan_work)->work) &local->mtx rcu_state.exp_mutex rcu_node_0 irq_context: 0 (wq_completion)phy13 (work_completion)(&(&local->scan_work)->work) &local->mtx rcu_state.exp_mutex &obj_hash[i].lock irq_context: 0 (wq_completion)phy13 (work_completion)(&(&local->scan_work)->work) &local->mtx rcu_state.exp_mutex rcu_read_lock &pool->lock irq_context: 0 (wq_completion)phy13 (work_completion)(&(&local->scan_work)->work) &local->mtx rcu_state.exp_mutex rcu_read_lock &pool->lock &obj_hash[i].lock irq_context: 0 (wq_completion)phy13 (work_completion)(&(&local->scan_work)->work) &local->mtx rcu_state.exp_mutex rcu_read_lock &pool->lock &p->pi_lock irq_context: 0 (wq_completion)phy13 (work_completion)(&(&local->scan_work)->work) &local->mtx rcu_state.exp_mutex rcu_read_lock &pool->lock &p->pi_lock &rq->__lock irq_context: 0 (wq_completion)phy13 (work_completion)(&(&local->scan_work)->work) &local->mtx rcu_state.exp_mutex rcu_read_lock &pool->lock &p->pi_lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)phy13 (work_completion)(&(&local->scan_work)->work) &local->mtx rcu_state.exp_mutex &rnp->exp_wq[0] irq_context: 0 (wq_completion)phy13 (work_completion)(&(&local->scan_work)->work) &local->mtx rcu_state.exp_mutex &rq->__lock irq_context: 0 (wq_completion)phy13 (work_completion)(&(&local->scan_work)->work) &local->mtx rcu_state.exp_mutex &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)phy13 (work_completion)(&(&local->scan_work)->work) &local->mtx &rdev->wiphy_work_lock irq_context: 0 (wq_completion)phy13 (work_completion)(&(&local->scan_work)->work) &local->mtx rcu_read_lock &pool->lock irq_context: 0 (wq_completion)phy13 (work_completion)(&(&local->scan_work)->work) &local->mtx rcu_read_lock &pool->lock &obj_hash[i].lock irq_context: 0 (wq_completion)phy13 (work_completion)(&(&local->scan_work)->work) &local->mtx rcu_read_lock &pool->lock &p->pi_lock irq_context: 0 (wq_completion)phy13 (work_completion)(&(&local->scan_work)->work) &local->mtx rcu_read_lock &pool->lock &p->pi_lock &rq->__lock irq_context: 0 (wq_completion)phy13 (work_completion)(&(&local->scan_work)->work) &local->mtx rcu_read_lock &pool->lock &p->pi_lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)phy13 (work_completion)(&(&local->scan_work)->work) &local->mtx &local->filter_lock irq_context: 0 (wq_completion)phy13 (work_completion)(&(&local->scan_work)->work) &local->mtx &local->queue_stop_reason_lock irq_context: 0 (wq_completion)phy13 (work_completion)(&(&local->scan_work)->work) &local->mtx rcu_read_lock &fq->lock irq_context: 0 (wq_completion)phy13 (work_completion)(&(&local->scan_work)->work) &local->mtx rcu_read_lock &local->queue_stop_reason_lock irq_context: 0 (wq_completion)events (work_completion)(&rdev->wiphy_work) &rdev->wiphy.mtx &____s->seqcount#2 irq_context: 0 (wq_completion)events (work_completion)(&rdev->wiphy_work) &rdev->wiphy.mtx &____s->seqcount irq_context: 0 (wq_completion)events wireless_nlevent_work net_rwsem &data->lock irq_context: 0 pernet_ops_rwsem rtnl_mutex rcu_state.exp_mutex fill_pool_map-wait-type-override &____s->seqcount#2 irq_context: 0 pernet_ops_rwsem rtnl_mutex rcu_state.exp_mutex fill_pool_map-wait-type-override &____s->seqcount irq_context: 0 (wq_completion)wg-kex-wg2#9 (work_completion)(&peer->transmit_handshake_work) &peer->endpoint_lock rcu_read_lock_bh rcu_read_lock rcu_read_lock rcu_read_lock rcu_read_lock_bh &data->lock &obj_hash[i].lock irq_context: 0 (wq_completion)wg-kex-wg2#9 (work_completion)(&peer->transmit_handshake_work) &peer->endpoint_lock rcu_read_lock_bh rcu_read_lock rcu_read_lock rcu_read_lock rcu_read_lock_bh &data->lock &base->lock irq_context: 0 (wq_completion)wg-kex-wg2#9 (work_completion)(&peer->transmit_handshake_work) &peer->endpoint_lock rcu_read_lock_bh rcu_read_lock rcu_read_lock rcu_read_lock rcu_read_lock_bh &data->lock &base->lock &obj_hash[i].lock irq_context: 0 pernet_ops_rwsem &root->kernfs_rwsem &p->pi_lock &rq->__lock irq_context: 0 pernet_ops_rwsem &root->kernfs_rwsem &p->pi_lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: softirq (&hc->tx_rtotimer) stock_lock rcu_read_lock per_cpu_ptr(&cgroup_rstat_cpu_lock, cpu) irq_context: 0 rcu_read_lock_bh dev->qdisc_tx_busylock ?: &qdisc_tx_busylock _xmit_ETHER#2 rcu_read_lock &____s->seqcount#2 irq_context: 0 rcu_read_lock_bh dev->qdisc_tx_busylock ?: &qdisc_tx_busylock _xmit_ETHER#2 rcu_read_lock &____s->seqcount irq_context: 0 rcu_read_lock_bh quarantine_lock irq_context: 0 pernet_ops_rwsem rtnl_mutex &c->lock batched_entropy_u8.lock irq_context: 0 pernet_ops_rwsem rtnl_mutex &c->lock kfence_freelist_lock irq_context: 0 (wq_completion)wg-kex-wg0#7 (work_completion)(&peer->transmit_handshake_work) rcu_node_0 irq_context: 0 (wq_completion)events free_ipc_work per_cpu_ptr(&cgroup_rstat_cpu_lock, cpu) irq_context: 0 (wq_completion)wg-kex-wg0#4 (work_completion)(&({ do { const void *__vpp_verify = (typeof((worker) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __asm__ ("" : "=r"(__ptr) : "0"((typeof(*((worker))) *)((worker)))); (typeof((typeof(*((worker))) *)((worker)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); })->work) &handshake->lock &c->lock irq_context: softirq (&tw->tw_timer) &dccp_hashinfo.bhash[i].lock &dccp_hashinfo.bhash2[i].lock stock_lock rcu_read_lock per_cpu_ptr(&cgroup_rstat_cpu_lock, cpu) irq_context: softirq (&tw->tw_timer) &dccp_hashinfo.bhash[i].lock &dccp_hashinfo.bhash2[i].lock per_cpu_ptr(&cgroup_rstat_cpu_lock, cpu) irq_context: softirq rcu_read_lock rcu_read_lock slock-AF_INET6 &meta->lock irq_context: softirq rcu_read_lock rcu_read_lock slock-AF_INET6 kfence_freelist_lock irq_context: 0 rtnl_mutex &data->lock &obj_hash[i].lock irq_context: 0 rtnl_mutex &data->lock &base->lock irq_context: 0 rtnl_mutex &data->lock &base->lock &obj_hash[i].lock irq_context: softirq rcu_read_lock rcu_read_lock &dreq->dreq_lock slock-AF_INET6 &pcp->lock &zone->lock irq_context: softirq rcu_read_lock rcu_read_lock &dreq->dreq_lock slock-AF_INET6 &pcp->lock &zone->lock &____s->seqcount irq_context: 0 pernet_ops_rwsem devices_rwsem rdma_nets_rwsem &device->compat_devs_mutex &rxe->usdev_lock rtnl_mutex pgd_lock irq_context: 0 pernet_ops_rwsem devices_rwsem rdma_nets_rwsem &device->compat_devs_mutex &rxe->usdev_lock rtnl_mutex stock_lock irq_context: 0 pernet_ops_rwsem devices_rwsem rdma_nets_rwsem &device->compat_devs_mutex &rxe->usdev_lock rtnl_mutex rcu_read_lock pool_lock#2 irq_context: 0 pernet_ops_rwsem devices_rwsem rdma_nets_rwsem &device->compat_devs_mutex &rxe->usdev_lock rtnl_mutex key irq_context: 0 pernet_ops_rwsem devices_rwsem rdma_nets_rwsem &device->compat_devs_mutex &rxe->usdev_lock rtnl_mutex pcpu_lock irq_context: 0 pernet_ops_rwsem devices_rwsem rdma_nets_rwsem &device->compat_devs_mutex &rxe->usdev_lock rtnl_mutex percpu_counters_lock irq_context: 0 pernet_ops_rwsem devices_rwsem rdma_nets_rwsem &device->compat_devs_mutex &rxe->usdev_lock rtnl_mutex pcpu_lock stock_lock irq_context: 0 &mm->mmap_lock stock_lock per_cpu_ptr(&cgroup_rstat_cpu_lock, cpu) irq_context: 0 tracepoints_mutex cpu_hotplug_lock static_call_mutex text_mutex rcu_node_0 irq_context: 0 tracepoints_mutex cpu_hotplug_lock static_call_mutex text_mutex &rcu_state.expedited_wq irq_context: 0 tracepoints_mutex cpu_hotplug_lock static_call_mutex text_mutex &rcu_state.expedited_wq &p->pi_lock irq_context: 0 tracepoints_mutex cpu_hotplug_lock static_call_mutex text_mutex &rcu_state.expedited_wq &p->pi_lock &rq->__lock irq_context: 0 tracepoints_mutex cpu_hotplug_lock static_call_mutex text_mutex &rcu_state.expedited_wq &p->pi_lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 &pipe->mutex/1 &sem->wait_lock irq_context: 0 rtnl_mutex &c->lock batched_entropy_u8.lock irq_context: 0 rtnl_mutex &c->lock kfence_freelist_lock irq_context: 0 rtnl_mutex k-sk_lock-AF_INET &____s->seqcount#8 irq_context: 0 rtnl_mutex k-sk_lock-AF_INET batched_entropy_u32.lock irq_context: 0 cb_lock nlk_cb_mutex-GENERIC genl_mutex rtnl_mutex.wait_lock irq_context: 0 cb_lock nlk_cb_mutex-GENERIC genl_mutex &p->pi_lock irq_context: 0 cb_lock nlk_cb_mutex-GENERIC genl_mutex &p->pi_lock &rq->__lock irq_context: 0 cb_lock nlk_cb_mutex-GENERIC genl_mutex &p->pi_lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 cb_lock nlk_cb_mutex-GENERIC &p->pi_lock &rq->__lock &cfs_rq->removed.lock irq_context: 0 (wq_completion)mld (work_completion)(&(&idev->mc_ifc_work)->work) &idev->mc_lock rcu_read_lock rcu_read_lock rcu_read_lock_bh rcu_read_lock rcu_read_lock rcu_read_lock &____s->seqcount#7 irq_context: 0 (wq_completion)mld (work_completion)(&(&idev->mc_ifc_work)->work) &idev->mc_lock rcu_read_lock rcu_read_lock rcu_read_lock_bh rcu_read_lock rcu_read_lock &nf_nat_locks[i] irq_context: 0 (wq_completion)mld (work_completion)(&(&idev->mc_ifc_work)->work) &idev->mc_lock rcu_read_lock rcu_read_lock rcu_read_lock_bh rcu_read_lock rcu_read_lock rcu_read_lock_bh &data->lock irq_context: 0 &sb->s_type->i_mutex_key#8 &sbi->s_writepages_rwsem jbd2_handle remove_cache_srcu &rq->__lock irq_context: 0 &sb->s_type->i_mutex_key#8 &sbi->s_writepages_rwsem jbd2_handle remove_cache_srcu &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 &type->i_mutex_dir_key#3 &cfs_rq->removed.lock irq_context: 0 &type->i_mutex_dir_key#3 &rq->__lock &cfs_rq->removed.lock irq_context: 0 sb_writers#4 &type->i_mutex_dir_key#3/1 jbd2_handle &ei->i_data_sem remove_cache_srcu rcu_node_0 irq_context: 0 (wq_completion)ipv6_addrconf (work_completion)(&(&ifa->dad_work)->work) rtnl_mutex rcu_read_lock rcu_read_lock rcu_read_lock_bh rcu_read_lock rcu_read_lock rcu_read_lock_bh &data->lock irq_context: 0 &mm->mmap_lock sb_writers#4 rcu_read_lock &rcu_state.gp_wq irq_context: 0 &mm->mmap_lock sb_writers#4 rcu_read_lock &rcu_state.gp_wq &p->pi_lock irq_context: 0 &mm->mmap_lock sb_writers#4 rcu_read_lock &rcu_state.gp_wq &p->pi_lock &rq->__lock irq_context: 0 &mm->mmap_lock sb_writers#4 rcu_read_lock &rcu_state.gp_wq &p->pi_lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 sk_lock-AF_SMC k-sk_lock-AF_INET &tcp_hashinfo.bhash[i].lock k-clock-AF_INET irq_context: 0 sk_lock-AF_SMC k-sk_lock-AF_INET &tcp_hashinfo.bhash[i].lock &tcp_hashinfo.bhash2[i].lock k-clock-AF_INET irq_context: 0 &sb->s_type->i_mutex_key#10 &smc->clcsock_release_lock k-sk_lock-AF_INET k-slock-AF_INET &tcp_hashinfo.bhash[i].lock irq_context: 0 &sb->s_type->i_mutex_key#10 &smc->clcsock_release_lock k-sk_lock-AF_INET k-slock-AF_INET &tcp_hashinfo.bhash[i].lock stock_lock irq_context: 0 &sb->s_type->i_mutex_key#10 &smc->clcsock_release_lock k-sk_lock-AF_INET k-slock-AF_INET &tcp_hashinfo.bhash[i].lock &obj_hash[i].lock irq_context: 0 &sb->s_type->i_mutex_key#10 &smc->clcsock_release_lock k-sk_lock-AF_INET k-slock-AF_INET &tcp_hashinfo.bhash[i].lock pool_lock#2 irq_context: 0 &sb->s_type->i_mutex_key#10 &smc->clcsock_release_lock k-sk_lock-AF_INET k-slock-AF_INET &tcp_hashinfo.bhash[i].lock &tcp_hashinfo.bhash2[i].lock irq_context: 0 &sb->s_type->i_mutex_key#10 &smc->clcsock_release_lock k-sk_lock-AF_INET k-slock-AF_INET &tcp_hashinfo.bhash[i].lock &tcp_hashinfo.bhash2[i].lock stock_lock irq_context: 0 &sb->s_type->i_mutex_key#10 &smc->clcsock_release_lock k-sk_lock-AF_INET k-slock-AF_INET &tcp_hashinfo.bhash[i].lock &tcp_hashinfo.bhash2[i].lock &obj_hash[i].lock irq_context: 0 &sb->s_type->i_mutex_key#10 &smc->clcsock_release_lock k-sk_lock-AF_INET k-slock-AF_INET &tcp_hashinfo.bhash[i].lock &tcp_hashinfo.bhash2[i].lock pool_lock#2 irq_context: 0 &pipe->mutex/1 &f->f_lock irq_context: 0 &pipe->mutex/1 &f->f_lock fasync_lock irq_context: 0 rcu_read_lock &new->fa_lock irq_context: 0 rcu_read_lock &new->fa_lock &f->f_owner.lock irq_context: 0 sb_writers#4 &sb->s_type->i_mutex_key#8 jbd2_handle &ei->i_data_sem &pcp->lock &zone->lock &____s->seqcount irq_context: 0 &pipe->mutex/1 rcu_read_lock &new->fa_lock irq_context: 0 &pipe->mutex/1 rcu_read_lock &new->fa_lock &f->f_owner.lock irq_context: 0 &pipe->mutex/1 &f->f_lock fasync_lock &new->fa_lock irq_context: 0 &pipe->mutex/1 &f->f_lock fasync_lock &obj_hash[i].lock irq_context: 0 &pipe->mutex/1 &f->f_lock fasync_lock pool_lock#2 irq_context: 0 (wq_completion)ipv6_addrconf (work_completion)(&(&ifa->dad_work)->work) rtnl_mutex &idev->mc_lock rcu_read_lock rcu_read_lock rcu_read_lock_bh rcu_read_lock rcu_read_lock rcu_read_lock_bh noop_qdisc.q.lock irq_context: 0 (wq_completion)ipv6_addrconf (work_completion)(&(&ifa->dad_work)->work) rtnl_mutex &idev->mc_lock rcu_read_lock rcu_read_lock rcu_read_lock_bh rcu_read_lock rcu_read_lock rcu_read_lock_bh &data->lock irq_context: 0 (wq_completion)ipv6_addrconf (work_completion)(&(&ifa->dad_work)->work) rtnl_mutex &idev->mc_lock rcu_read_lock rcu_read_lock rcu_read_lock_bh rcu_read_lock rcu_read_lock rcu_read_lock_bh &obj_hash[i].lock irq_context: 0 (wq_completion)ipv6_addrconf (work_completion)(&(&ifa->dad_work)->work) rtnl_mutex &idev->mc_lock rcu_read_lock rcu_read_lock rcu_read_lock_bh rcu_read_lock rcu_read_lock rcu_read_lock_bh pool_lock#2 irq_context: 0 (wq_completion)mld (work_completion)(&(&idev->mc_dad_work)->work) &idev->mc_lock rcu_read_lock rcu_read_lock rcu_read_lock_bh rcu_read_lock rcu_read_lock rcu_read_lock_bh noop_qdisc.q.lock irq_context: 0 (wq_completion)mld (work_completion)(&(&idev->mc_dad_work)->work) &idev->mc_lock rcu_read_lock rcu_read_lock rcu_read_lock_bh rcu_read_lock rcu_read_lock rcu_read_lock_bh &data->lock irq_context: 0 (wq_completion)mld (work_completion)(&(&idev->mc_dad_work)->work) &idev->mc_lock rcu_read_lock rcu_read_lock rcu_read_lock_bh rcu_read_lock rcu_read_lock rcu_read_lock_bh &obj_hash[i].lock irq_context: 0 (wq_completion)mld (work_completion)(&(&idev->mc_dad_work)->work) &idev->mc_lock rcu_read_lock rcu_read_lock rcu_read_lock_bh rcu_read_lock rcu_read_lock rcu_read_lock_bh pool_lock#2 irq_context: 0 sb_writers#11 &type->i_mutex_dir_key#7/1 rcu_read_lock &rcu_state.gp_wq irq_context: 0 sb_writers#11 &type->i_mutex_dir_key#7/1 rcu_read_lock &rcu_state.gp_wq &p->pi_lock irq_context: 0 sb_writers#11 &type->i_mutex_dir_key#7/1 rcu_read_lock &rcu_state.gp_wq &p->pi_lock &rq->__lock irq_context: 0 sb_writers#11 &type->i_mutex_dir_key#7/1 rcu_read_lock &rcu_state.gp_wq &p->pi_lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 sb_writers#4 &type->i_mutex_dir_key#3/1 jbd2_handle &pcp->lock &zone->lock &____s->seqcount irq_context: 0 pernet_ops_rwsem __ip_vs_app_mutex rcu_read_lock rcu_node_0 irq_context: 0 pernet_ops_rwsem __ip_vs_app_mutex rcu_read_lock &rq->__lock irq_context: 0 pernet_ops_rwsem __ip_vs_app_mutex rcu_read_lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 pernet_ops_rwsem nf_hook_mutex &n->list_lock irq_context: 0 pernet_ops_rwsem nf_hook_mutex &n->list_lock &c->lock irq_context: 0 pernet_ops_rwsem rtnl_mutex &idev->mc_lock &pcp->lock &zone->lock irq_context: softirq (&ndev->rs_timer) rcu_read_lock rcu_read_lock rcu_read_lock_bh rcu_read_lock rcu_read_lock rcu_read_lock_bh noop_qdisc.q.lock irq_context: softirq (&ndev->rs_timer) rcu_read_lock rcu_read_lock rcu_read_lock_bh rcu_read_lock rcu_read_lock rcu_read_lock_bh &data->lock irq_context: softirq (&ndev->rs_timer) rcu_read_lock rcu_read_lock rcu_read_lock_bh rcu_read_lock rcu_read_lock rcu_read_lock_bh &obj_hash[i].lock irq_context: softirq (&ndev->rs_timer) rcu_read_lock rcu_read_lock rcu_read_lock_bh rcu_read_lock rcu_read_lock rcu_read_lock_bh pool_lock#2 irq_context: 0 pernet_ops_rwsem k-sk_lock-AF_RXRPC &rxnet->local_mutex &p->pi_lock &cfs_rq->removed.lock irq_context: 0 &mm->mmap_lock sb_writers#4 rcu_node_0 irq_context: 0 &mm->mmap_lock sb_writers#4 &rcu_state.expedited_wq irq_context: 0 &mm->mmap_lock sb_writers#4 &rcu_state.expedited_wq &p->pi_lock irq_context: 0 &mm->mmap_lock sb_writers#4 &rcu_state.expedited_wq &p->pi_lock &rq->__lock irq_context: 0 &mm->mmap_lock sb_writers#4 &rcu_state.expedited_wq &p->pi_lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)wg-kex-wg0#7 (work_completion)(&peer->transmit_handshake_work) &rcu_state.expedited_wq irq_context: 0 (wq_completion)wg-kex-wg0#7 (work_completion)(&peer->transmit_handshake_work) &rcu_state.expedited_wq &p->pi_lock irq_context: 0 (wq_completion)wg-kex-wg0#7 (work_completion)(&peer->transmit_handshake_work) &rcu_state.expedited_wq &p->pi_lock &rq->__lock irq_context: 0 (wq_completion)wg-kex-wg0#7 (work_completion)(&peer->transmit_handshake_work) &rcu_state.expedited_wq &p->pi_lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 pernet_ops_rwsem rcu_read_lock &rcu_state.gp_wq &p->pi_lock &cfs_rq->removed.lock irq_context: 0 pernet_ops_rwsem rcu_read_lock &rcu_state.expedited_wq &p->pi_lock &rq->__lock &cfs_rq->removed.lock irq_context: softirq (&tw->tw_timer) &dccp_hashinfo.bhash[i].lock &dccp_hashinfo.bhash2[i].lock &meta->lock irq_context: softirq (&tw->tw_timer) &dccp_hashinfo.bhash[i].lock &dccp_hashinfo.bhash2[i].lock kfence_freelist_lock irq_context: 0 &mm->mmap_lock &anon_vma->rwsem rcu_read_lock rcu_node_0 irq_context: 0 &mm->mmap_lock &anon_vma->rwsem rcu_read_lock &rq->__lock irq_context: 0 &mm->mmap_lock &anon_vma->rwsem rcu_read_lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 sk_lock-AF_INET6 rcu_read_lock &rcu_state.gp_wq &p->pi_lock &cfs_rq->removed.lock irq_context: 0 clock-AF_KEY irq_context: 0 wlock-AF_KEY irq_context: 0 pernet_ops_rwsem ovs_mutex rcu_read_lock rcu_node_0 irq_context: 0 pernet_ops_rwsem ovs_mutex rcu_read_lock &rq->__lock irq_context: softirq (&ndev->rs_timer) rcu_read_lock rcu_read_lock rcu_read_lock_bh rcu_read_lock rcu_read_lock rcu_read_lock_bh &meta->lock irq_context: softirq (&ndev->rs_timer) rcu_read_lock rcu_read_lock rcu_read_lock_bh rcu_read_lock rcu_read_lock rcu_read_lock_bh kfence_freelist_lock irq_context: 0 rtnl_mutex cpu_hotplug_lock xps_map_mutex &rq->__lock irq_context: 0 rtnl_mutex cpu_hotplug_lock xps_map_mutex &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: softirq &ret->b_uptodate_lock bit_wait_table + i irq_context: softirq &ret->b_uptodate_lock bit_wait_table + i &p->pi_lock irq_context: softirq &ret->b_uptodate_lock bit_wait_table + i &p->pi_lock &rq->__lock irq_context: softirq &ret->b_uptodate_lock bit_wait_table + i &p->pi_lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 sk_lock-AF_ALG remove_cache_srcu &pcp->lock &zone->lock irq_context: 0 sk_lock-AF_ALG remove_cache_srcu &pcp->lock &zone->lock &____s->seqcount irq_context: 0 sk_lock-AF_ALG remove_cache_srcu &cfs_rq->removed.lock irq_context: 0 &mm->mmap_lock &mapping->i_mmap_rwsem rcu_read_lock &rcu_state.gp_wq &p->pi_lock &cfs_rq->removed.lock irq_context: softirq rcu_read_lock rcu_read_lock slock-AF_INET6/1 sk_lock-AF_INET6 rcu_read_lock rcu_read_lock rcu_read_lock rcu_read_lock_bh &sch->q.lock &n->list_lock irq_context: softirq rcu_read_lock rcu_read_lock slock-AF_INET6/1 sk_lock-AF_INET6 rcu_read_lock rcu_read_lock rcu_read_lock rcu_read_lock_bh &sch->q.lock &n->list_lock &c->lock irq_context: 0 rtnl_mutex sk_lock-AF_INET6 &idev->mc_lock &bridge_netdev_addr_lock_key irq_context: 0 rtnl_mutex sk_lock-AF_INET6 &idev->mc_lock &bridge_netdev_addr_lock_key &c->lock irq_context: 0 rtnl_mutex &dev_addr_list_lock_key/2 rcu_read_lock &macvlan_netdev_addr_lock_key/1 irq_context: 0 rtnl_mutex &dev_addr_list_lock_key/2 rcu_read_lock &macvlan_netdev_addr_lock_key/1 _xmit_ETHER irq_context: 0 rtnl_mutex &dev_addr_list_lock_key/2 &obj_hash[i].lock irq_context: 0 rtnl_mutex &dev_addr_list_lock_key/2 krc.lock irq_context: 0 sb_writers#4 &iint->mutex ima_extend_list_mutex rcu_read_lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 sb_writers#4 &iint->mutex rcu_read_lock rcu_read_lock rcu_node_0 irq_context: 0 sb_writers#4 &iint->mutex rcu_read_lock rcu_read_lock &rq->__lock irq_context: 0 purge_vmap_area_lock rcu_read_lock &p->pi_lock irq_context: 0 purge_vmap_area_lock rcu_read_lock &p->pi_lock &rq->__lock irq_context: 0 purge_vmap_area_lock rcu_read_lock &p->pi_lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 sk_lock-AF_INET6 console_owner_lock irq_context: 0 sk_lock-AF_INET6 console_owner irq_context: softirq rcu_read_lock rcu_read_lock slock-AF_INET6 quarantine_lock irq_context: 0 (wq_completion)bond10 irq_context: 0 (wq_completion)bond10 (work_completion)(&(&bond->mii_work)->work) irq_context: 0 (wq_completion)bond10 (work_completion)(&(&bond->mii_work)->work) &obj_hash[i].lock irq_context: 0 (wq_completion)bond10 (work_completion)(&(&bond->mii_work)->work) &base->lock irq_context: 0 (wq_completion)bond10 (work_completion)(&(&bond->mii_work)->work) &base->lock &obj_hash[i].lock irq_context: 0 (wq_completion)bond10 (work_completion)(&(&bond->mii_work)->work) &rq->__lock irq_context: 0 (wq_completion)bond10 (work_completion)(&(&bond->ad_work)->work) irq_context: 0 (wq_completion)bond10 (work_completion)(&(&bond->ad_work)->work) &bond->mode_lock irq_context: 0 (wq_completion)bond10 (work_completion)(&(&bond->ad_work)->work) &obj_hash[i].lock irq_context: 0 (wq_completion)bond10 (work_completion)(&(&bond->ad_work)->work) &base->lock irq_context: 0 (wq_completion)bond10 (work_completion)(&(&bond->ad_work)->work) &base->lock &obj_hash[i].lock irq_context: 0 rtnl_mutex team->team_lock_key#4 &bond->mode_lock irq_context: 0 rtnl_mutex team->team_lock_key#4 &dev_addr_list_lock_key irq_context: 0 rtnl_mutex team->team_lock_key#4 rcu_read_lock &bond->stats_lock irq_context: 0 rtnl_mutex team->team_lock_key#4 &dev_addr_list_lock_key &c->lock irq_context: 0 rtnl_mutex team->team_lock_key#4 &dev_addr_list_lock_key#2/1 irq_context: 0 rtnl_mutex team->team_lock_key#4 &dev_addr_list_lock_key#2/1 &dev_addr_list_lock_key irq_context: 0 rtnl_mutex team->team_lock_key#4 &dev_addr_list_lock_key#2/1 &dev_addr_list_lock_key pool_lock#2 irq_context: 0 (wq_completion)events (linkwatch_work).work rtnl_mutex team->team_lock_key#4 rcu_read_lock &bond->stats_lock irq_context: 0 (wq_completion)bond10 (work_completion)(&(&slave->notify_work)->work) irq_context: 0 (wq_completion)bond10 (work_completion)(&(&slave->notify_work)->work) &obj_hash[i].lock irq_context: 0 (wq_completion)bond10 (work_completion)(&(&slave->notify_work)->work) &base->lock irq_context: 0 (wq_completion)bond10 (work_completion)(&(&slave->notify_work)->work) &base->lock &obj_hash[i].lock irq_context: 0 rtnl_mutex &dev_addr_list_lock_key/2 &macvlan_netdev_addr_lock_key/1 &c->lock irq_context: 0 rtnl_mutex &dev_addr_list_lock_key/2 &macvlan_netdev_addr_lock_key/1 pool_lock#2 irq_context: 0 rtnl_mutex &dev_addr_list_lock_key/2 &macvlan_netdev_addr_lock_key/1 _xmit_ETHER pool_lock#2 irq_context: 0 (wq_completion)bond10 (work_completion)(&(&slave->notify_work)->work) rtnl_mutex net_rwsem irq_context: 0 (wq_completion)bond10 (work_completion)(&(&slave->notify_work)->work) rtnl_mutex net_rwsem &list->lock#2 irq_context: 0 (wq_completion)bond10 (work_completion)(&(&slave->notify_work)->work) rtnl_mutex &tn->lock irq_context: 0 (wq_completion)bond10 (work_completion)(&(&bond->ad_work)->work) &bond->mode_lock rcu_read_lock rcu_read_lock lweventlist_lock irq_context: 0 (wq_completion)bond10 (work_completion)(&(&bond->ad_work)->work) &bond->mode_lock rcu_read_lock rcu_read_lock lweventlist_lock &c->lock irq_context: 0 (wq_completion)bond10 (work_completion)(&(&bond->ad_work)->work) &bond->mode_lock rcu_read_lock rcu_read_lock lweventlist_lock &dir->lock#2 irq_context: 0 (wq_completion)bond10 (work_completion)(&(&bond->ad_work)->work) &bond->mode_lock rcu_read_lock rcu_read_lock &obj_hash[i].lock irq_context: 0 (wq_completion)bond10 (work_completion)(&(&bond->ad_work)->work) &bond->mode_lock rcu_read_lock rcu_read_lock &base->lock irq_context: 0 (wq_completion)bond10 (work_completion)(&(&bond->ad_work)->work) &bond->mode_lock rcu_read_lock rcu_read_lock &base->lock &obj_hash[i].lock irq_context: 0 (wq_completion)bond10 (work_completion)(&(&bond->ad_work)->work) &bond->mode_lock rcu_read_lock rcu_read_lock rcu_read_lock &pool->lock irq_context: 0 (wq_completion)bond10 (work_completion)(&(&bond->ad_work)->work) &bond->mode_lock rcu_read_lock rcu_read_lock rcu_read_lock &pool->lock &obj_hash[i].lock irq_context: 0 (wq_completion)bond10 (work_completion)(&(&bond->ad_work)->work) &bond->mode_lock rcu_read_lock rcu_read_lock rcu_read_lock &pool->lock &p->pi_lock irq_context: 0 (wq_completion)bond10 (work_completion)(&(&bond->ad_work)->work) &rq->__lock irq_context: 0 (wq_completion)bond10 (work_completion)(&(&bond->ad_work)->work) rcu_read_lock &pool->lock/1 irq_context: 0 (wq_completion)bond10 (work_completion)(&(&bond->ad_work)->work) rcu_read_lock &pool->lock/1 &obj_hash[i].lock irq_context: 0 (wq_completion)bond10 (work_completion)(&(&bond->ad_work)->work) rcu_read_lock &pool->lock/1 &p->pi_lock irq_context: 0 (wq_completion)bond10 (work_completion)(&(&bond->ad_work)->work) rcu_read_lock &pool->lock/1 &p->pi_lock &rq->__lock irq_context: 0 (wq_completion)bond10 (work_completion)(&(&bond->ad_work)->work) rcu_read_lock &pool->lock/1 &p->pi_lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)bond10 (work_completion)(&(&bond->ad_work)->work) rtnl_mutex fs_reclaim irq_context: 0 (wq_completion)bond10 (work_completion)(&(&bond->ad_work)->work) rtnl_mutex fs_reclaim mmu_notifier_invalidate_range_start irq_context: 0 (wq_completion)bond10 (work_completion)(&(&bond->ad_work)->work) rtnl_mutex rcu_read_lock &ndev->lock irq_context: 0 (wq_completion)bond10 (work_completion)(&(&bond->ad_work)->work) rtnl_mutex &c->lock irq_context: 0 (wq_completion)bond10 (work_completion)(&(&bond->ad_work)->work) rtnl_mutex &obj_hash[i].lock irq_context: 0 (wq_completion)bond10 (work_completion)(&(&bond->ad_work)->work) rtnl_mutex nl_table_lock irq_context: 0 (wq_completion)bond10 (work_completion)(&(&bond->ad_work)->work) rtnl_mutex nl_table_wait.lock irq_context: 0 (wq_completion)bond10 (work_completion)(&(&bond->ad_work)->work) rtnl_mutex net_rwsem irq_context: 0 (wq_completion)bond10 (work_completion)(&(&bond->ad_work)->work) rtnl_mutex net_rwsem &list->lock#2 irq_context: 0 (wq_completion)bond10 (work_completion)(&(&bond->ad_work)->work) rtnl_mutex &tn->lock irq_context: 0 (wq_completion)bond10 (work_completion)(&(&bond->slave_arr_work)->work) irq_context: 0 (wq_completion)bond10 (work_completion)(&(&bond->slave_arr_work)->work) rtnl_mutex fs_reclaim irq_context: 0 (wq_completion)bond10 (work_completion)(&(&bond->slave_arr_work)->work) rtnl_mutex fs_reclaim mmu_notifier_invalidate_range_start irq_context: 0 (wq_completion)bond10 (work_completion)(&(&bond->slave_arr_work)->work) rtnl_mutex &c->lock irq_context: 0 (wq_completion)bond10 (work_completion)(&(&bond->slave_arr_work)->work) rtnl_mutex &n->list_lock irq_context: 0 (wq_completion)bond10 (work_completion)(&(&bond->slave_arr_work)->work) rtnl_mutex &n->list_lock &c->lock irq_context: 0 (wq_completion)bond10 (work_completion)(&(&bond->slave_arr_work)->work) rtnl_mutex &bond->mode_lock irq_context: 0 (wq_completion)events (linkwatch_work).work rtnl_mutex team->team_lock_key#4 &n->list_lock irq_context: 0 (wq_completion)events (linkwatch_work).work rtnl_mutex team->team_lock_key#4 &n->list_lock &c->lock irq_context: 0 (wq_completion)events (linkwatch_work).work rtnl_mutex team->team_lock_key#4 rcu_read_lock &bond->stats_lock/2 irq_context: 0 (wq_completion)bond10 (work_completion)(&(&bond->ad_work)->work) &bond->mode_lock rcu_read_lock pool_lock#2 irq_context: 0 (wq_completion)bond10 (work_completion)(&(&bond->ad_work)->work) &bond->mode_lock rcu_read_lock rcu_read_lock_bh rcu_read_lock pool_lock#2 irq_context: 0 (wq_completion)bond10 (work_completion)(&(&bond->ad_work)->work) &bond->mode_lock rcu_read_lock rcu_read_lock_bh rcu_read_lock tk_core.seq.seqcount irq_context: 0 (wq_completion)bond10 (work_completion)(&(&bond->ad_work)->work) &bond->mode_lock rcu_read_lock rcu_read_lock_bh rcu_read_lock rlock-AF_PACKET irq_context: 0 (wq_completion)bond10 (work_completion)(&(&bond->ad_work)->work) &bond->mode_lock rcu_read_lock rcu_read_lock_bh rcu_read_lock &obj_hash[i].lock irq_context: 0 (wq_completion)bond10 (work_completion)(&(&bond->ad_work)->work) &bond->mode_lock rcu_read_lock rcu_read_lock_bh rcu_read_lock_bh rcu_read_lock &c->lock irq_context: 0 (wq_completion)bond10 (work_completion)(&(&bond->ad_work)->work) &bond->mode_lock rcu_read_lock rcu_read_lock_bh rcu_read_lock_bh rcu_read_lock pool_lock#2 irq_context: 0 (wq_completion)bond10 (work_completion)(&(&bond->ad_work)->work) &bond->mode_lock rcu_read_lock rcu_read_lock_bh rcu_read_lock_bh rcu_read_lock tk_core.seq.seqcount irq_context: 0 (wq_completion)bond10 (work_completion)(&(&bond->ad_work)->work) &bond->mode_lock rcu_read_lock rcu_read_lock_bh rcu_read_lock_bh rcu_read_lock rlock-AF_PACKET irq_context: 0 (wq_completion)bond10 (work_completion)(&(&bond->ad_work)->work) &bond->mode_lock rcu_read_lock rcu_read_lock_bh rcu_read_lock_bh rcu_read_lock &obj_hash[i].lock irq_context: 0 (wq_completion)bond10 (work_completion)(&(&bond->ad_work)->work) &bond->mode_lock rcu_read_lock rcu_read_lock_bh rcu_read_lock_bh rcu_read_lock &____s->seqcount irq_context: 0 (wq_completion)bond10 (work_completion)(&(&bond->ad_work)->work) &bond->mode_lock rcu_read_lock rcu_read_lock_bh rcu_read_lock_bh rcu_read_lock &list->lock#5 irq_context: 0 rtnl_mutex team->team_lock_key#4 &obj_hash[i].lock pool_lock irq_context: 0 (wq_completion)bond11 irq_context: 0 (wq_completion)bond11 &rq->__lock irq_context: 0 (wq_completion)bond11 &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 rtnl_mutex team->team_lock_key#4 &cfs_rq->removed.lock irq_context: 0 rtnl_mutex team->team_lock_key#4 pool_lock#2 irq_context: 0 (wq_completion)bond11 (work_completion)(&(&bond->mii_work)->work) irq_context: 0 (wq_completion)bond11 (work_completion)(&(&bond->mii_work)->work) &rq->__lock irq_context: 0 (wq_completion)bond11 (work_completion)(&(&bond->mii_work)->work) &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)bond11 (work_completion)(&(&bond->mii_work)->work) &obj_hash[i].lock irq_context: 0 (wq_completion)bond11 (work_completion)(&(&bond->mii_work)->work) &base->lock irq_context: 0 (wq_completion)bond11 (work_completion)(&(&bond->mii_work)->work) &base->lock &obj_hash[i].lock irq_context: 0 (wq_completion)bond11 (work_completion)(&(&bond->ad_work)->work) irq_context: 0 (wq_completion)bond11 (work_completion)(&(&bond->ad_work)->work) &bond->mode_lock irq_context: 0 (wq_completion)bond11 (work_completion)(&(&bond->ad_work)->work) &obj_hash[i].lock irq_context: 0 (wq_completion)bond11 (work_completion)(&(&bond->ad_work)->work) &base->lock irq_context: 0 (wq_completion)bond11 (work_completion)(&(&bond->ad_work)->work) &base->lock &obj_hash[i].lock irq_context: 0 (wq_completion)bond11 (work_completion)(&(&bond->ad_work)->work) &rq->__lock irq_context: 0 (wq_completion)bond11 (work_completion)(&(&bond->ad_work)->work) &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 rtnl_mutex team->team_lock_key#4 &dev_addr_list_lock_key#2/3 irq_context: 0 rtnl_mutex team->team_lock_key#4 &dev_addr_list_lock_key#2/3 &dev_addr_list_lock_key irq_context: 0 rtnl_mutex team->team_lock_key#4 &dev_addr_list_lock_key#2/3 &dev_addr_list_lock_key &c->lock irq_context: 0 (wq_completion)events (work_completion)(&aux->work)#2 irq_context: 0 (wq_completion)events (work_completion)(&aux->work)#2 &aux->poke_mutex irq_context: 0 (wq_completion)events (work_completion)(&aux->work)#2 &aux->poke_mutex &rq->__lock irq_context: 0 (wq_completion)events (work_completion)(&aux->work)#2 &rq->__lock irq_context: 0 (wq_completion)events (work_completion)(&aux->work)#2 &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)events (work_completion)(&aux->work)#2 &aux->poke_mutex &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 sb_writers#4 &type->i_mutex_dir_key#3/1 &sb->s_type->i_mutex_key#8 jbd2_handle &p->pi_lock &cfs_rq->removed.lock irq_context: 0 sb_writers#4 &type->i_mutex_dir_key#3/1 &sb->s_type->i_mutex_key#8 jbd2_handle &p->pi_lock &rq->__lock &cfs_rq->removed.lock irq_context: 0 (wq_completion)events (work_completion)(&aux->work)#2 map_idr_lock irq_context: 0 (wq_completion)events (work_completion)(&aux->work)#2 map_idr_lock &obj_hash[i].lock irq_context: 0 (wq_completion)events (work_completion)(&aux->work)#2 map_idr_lock pool_lock#2 irq_context: 0 (wq_completion)events (work_completion)(&aux->work)#2 &obj_hash[i].lock irq_context: 0 (wq_completion)events (work_completion)(&aux->work)#2 rcu_read_lock &pool->lock/1 irq_context: 0 (wq_completion)events (work_completion)(&aux->work)#2 rcu_read_lock &pool->lock/1 &obj_hash[i].lock irq_context: 0 (wq_completion)events (work_completion)(&aux->work)#2 rcu_read_lock &pool->lock/1 &p->pi_lock irq_context: 0 (wq_completion)events (work_completion)(&aux->work)#2 rcu_read_lock &pool->lock/1 &p->pi_lock &rq->__lock irq_context: 0 (wq_completion)events (work_completion)(&aux->work)#2 rcu_read_lock &pool->lock/1 &p->pi_lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)events (work_completion)(&aux->work)#2 &aux->poke_mutex &cfs_rq->removed.lock irq_context: 0 (wq_completion)events (work_completion)(&aux->work)#2 &aux->poke_mutex &obj_hash[i].lock irq_context: 0 (wq_completion)events (work_completion)(&aux->work)#2 &aux->poke_mutex pool_lock#2 irq_context: softirq rcu_read_lock rcu_read_lock rcu_read_lock rcu_read_lock_bh &dir->lock irq_context: softirq rcu_read_lock rcu_read_lock rcu_read_lock rcu_read_lock_bh &sch->q.lock &c->lock irq_context: 0 &sb->s_type->i_mutex_key#10 sk_lock-AF_INET6 slock-AF_INET6 rcu_read_lock rcu_read_lock rcu_read_lock_bh rcu_read_lock &n->list_lock irq_context: 0 &sb->s_type->i_mutex_key#10 sk_lock-AF_INET6 slock-AF_INET6 rcu_read_lock rcu_read_lock rcu_read_lock_bh rcu_read_lock &n->list_lock &c->lock irq_context: 0 (wq_completion)bond13 irq_context: 0 (wq_completion)bond13 (work_completion)(&(&bond->mii_work)->work) irq_context: 0 (wq_completion)bond13 (work_completion)(&(&bond->mii_work)->work) &obj_hash[i].lock irq_context: 0 (wq_completion)bond13 (work_completion)(&(&bond->mii_work)->work) &base->lock irq_context: 0 (wq_completion)bond13 (work_completion)(&(&bond->mii_work)->work) &base->lock &obj_hash[i].lock irq_context: 0 (wq_completion)bond13 (work_completion)(&(&bond->mii_work)->work) &rq->__lock irq_context: 0 (wq_completion)bond13 (work_completion)(&(&bond->ad_work)->work) irq_context: 0 (wq_completion)bond13 (work_completion)(&(&bond->ad_work)->work) &bond->mode_lock irq_context: 0 (wq_completion)bond13 (work_completion)(&(&bond->ad_work)->work) &obj_hash[i].lock irq_context: 0 (wq_completion)bond13 (work_completion)(&(&bond->ad_work)->work) &base->lock irq_context: 0 (wq_completion)bond13 (work_completion)(&(&bond->ad_work)->work) &base->lock &obj_hash[i].lock irq_context: 0 rtnl_mutex team->team_lock_key#2 &dev_addr_list_lock_key &c->lock irq_context: 0 (wq_completion)bond10 (work_completion)(&(&bond->ad_work)->work) &bond->mode_lock rcu_read_lock rcu_read_lock_bh rcu_read_lock &c->lock irq_context: 0 (wq_completion)phy12 (work_completion)(&(&local->scan_work)->work) &local->mtx rcu_state.exp_mutex rcu_read_lock &pool->lock &p->pi_lock &rq->__lock &cfs_rq->removed.lock irq_context: 0 (wq_completion)bond13 (work_completion)(&(&slave->notify_work)->work) irq_context: 0 (wq_completion)bond13 (work_completion)(&(&slave->notify_work)->work) &obj_hash[i].lock irq_context: 0 (wq_completion)bond13 (work_completion)(&(&slave->notify_work)->work) &base->lock irq_context: 0 (wq_completion)bond13 (work_completion)(&(&slave->notify_work)->work) &base->lock &obj_hash[i].lock irq_context: 0 (wq_completion)bond13 (work_completion)(&(&slave->notify_work)->work) rtnl_mutex net_rwsem irq_context: 0 (wq_completion)bond13 (work_completion)(&(&slave->notify_work)->work) rtnl_mutex net_rwsem &list->lock#2 irq_context: 0 (wq_completion)bond13 (work_completion)(&(&slave->notify_work)->work) rtnl_mutex &tn->lock irq_context: 0 sk_lock-AF_UNIX &u->iolock irq_context: 0 rtnl_mutex remove_cache_srcu rcu_read_lock &cfs_rq->removed.lock irq_context: 0 rtnl_mutex remove_cache_srcu rcu_read_lock &obj_hash[i].lock irq_context: 0 fs_reclaim &obj_hash[i].lock pool_lock irq_context: softirq rcu_callback &zone->lock &____s->seqcount irq_context: 0 vlan_ioctl_mutex rtnl_mutex team->team_lock_key#4 irq_context: 0 vlan_ioctl_mutex rtnl_mutex team->team_lock_key#4 fs_reclaim irq_context: 0 vlan_ioctl_mutex rtnl_mutex team->team_lock_key#4 fs_reclaim mmu_notifier_invalidate_range_start irq_context: 0 vlan_ioctl_mutex rtnl_mutex team->team_lock_key#4 pool_lock#2 irq_context: 0 vlan_ioctl_mutex rtnl_mutex team->team_lock_key#4 &rq->__lock irq_context: 0 vlan_ioctl_mutex rtnl_mutex team->team_lock_key#4 &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 vlan_ioctl_mutex rtnl_mutex &dev_addr_list_lock_key#2/3 irq_context: 0 vlan_ioctl_mutex rtnl_mutex &dev_addr_list_lock_key#2/3 rcu_read_lock _xmit_ETHER irq_context: 0 vlan_ioctl_mutex rtnl_mutex &dev_addr_list_lock_key#2/3 rcu_read_lock &dev_addr_list_lock_key/2 irq_context: 0 vlan_ioctl_mutex rtnl_mutex &dev_addr_list_lock_key#2/3 rcu_read_lock &dev_addr_list_lock_key/2 rcu_read_lock &macvlan_netdev_addr_lock_key/1 irq_context: 0 vlan_ioctl_mutex rtnl_mutex &dev_addr_list_lock_key#2/3 rcu_read_lock &dev_addr_list_lock_key/2 rcu_read_lock &macvlan_netdev_addr_lock_key/1 _xmit_ETHER irq_context: 0 vlan_ioctl_mutex rtnl_mutex &dev_addr_list_lock_key#2/3 rcu_read_lock &dev_addr_list_lock_key/2 &c->lock irq_context: 0 vlan_ioctl_mutex rtnl_mutex &dev_addr_list_lock_key#2/3 rcu_read_lock &dev_addr_list_lock_key/2 rcu_read_lock &macvlan_netdev_addr_lock_key/1 pool_lock#2 irq_context: 0 vlan_ioctl_mutex rtnl_mutex &dev_addr_list_lock_key#2/3 rcu_read_lock &dev_addr_list_lock_key/2 rcu_read_lock &macvlan_netdev_addr_lock_key/1 _xmit_ETHER pool_lock#2 irq_context: 0 vlan_ioctl_mutex rtnl_mutex &dev_addr_list_lock_key#2/3 rcu_read_lock &dev_addr_list_lock_key irq_context: 0 vlan_ioctl_mutex rtnl_mutex &dev_addr_list_lock_key#2/3 rcu_read_lock &dev_addr_list_lock_key &c->lock irq_context: 0 vlan_ioctl_mutex rtnl_mutex &dev_addr_list_lock_key#2/3 rcu_read_lock &dev_addr_list_lock_key &____s->seqcount#2 irq_context: 0 vlan_ioctl_mutex rtnl_mutex &dev_addr_list_lock_key#2/3 rcu_read_lock &dev_addr_list_lock_key &____s->seqcount irq_context: 0 vlan_ioctl_mutex rtnl_mutex &dev_addr_list_lock_key#2/3 rcu_read_lock &dev_addr_list_lock_key pool_lock#2 irq_context: 0 vlan_ioctl_mutex rtnl_mutex pcpu_alloc_mutex &rq->__lock irq_context: 0 vlan_ioctl_mutex rtnl_mutex pcpu_alloc_mutex &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 kn->active#60 &kernfs_locks->open_file_mutex[count] &cfs_rq->removed.lock irq_context: 0 kn->active#60 &kernfs_locks->open_file_mutex[count] &obj_hash[i].lock irq_context: 0 kn->active#60 &kernfs_locks->open_file_mutex[count] pool_lock#2 irq_context: 0 &kernfs_locks->open_file_mutex[count] &cfs_rq->removed.lock irq_context: 0 (wq_completion)wg-kex-wg0#7 (work_completion)(&peer->transmit_handshake_work) &peer->endpoint_lock rcu_read_lock_bh rcu_read_lock rcu_read_lock rcu_read_lock &c->lock irq_context: 0 (wq_completion)wg-kex-wg0#7 (work_completion)(&peer->transmit_handshake_work) &peer->endpoint_lock rcu_read_lock_bh rcu_read_lock rcu_read_lock rcu_read_lock &n->list_lock irq_context: 0 (wq_completion)wg-kex-wg0#7 (work_completion)(&peer->transmit_handshake_work) &peer->endpoint_lock rcu_read_lock_bh rcu_read_lock rcu_read_lock rcu_read_lock &n->list_lock &c->lock irq_context: 0 (wq_completion)wg-kex-wg0#7 (work_completion)(&peer->transmit_handshake_work) &peer->endpoint_lock rcu_read_lock_bh rcu_read_lock &nf_conntrack_locks[i] &nf_conntrack_locks[i]/1 batched_entropy_u8.lock crngs.lock irq_context: 0 (wq_completion)wg-kex-wg0#7 (work_completion)(&peer->transmit_handshake_work) &peer->endpoint_lock rcu_read_lock_bh rcu_read_lock rcu_read_lock_bh &sch->q.lock irq_context: 0 (wq_completion)wg-kex-wg0#7 (work_completion)(&peer->transmit_handshake_work) &peer->endpoint_lock rcu_read_lock_bh rcu_read_lock rcu_read_lock_bh &sch->q.lock pool_lock#2 irq_context: 0 (wq_completion)wg-kex-wg0#7 (work_completion)(&peer->transmit_handshake_work) &peer->endpoint_lock rcu_read_lock_bh rcu_read_lock rcu_read_lock_bh &sch->q.lock tk_core.seq.seqcount irq_context: 0 (wq_completion)wg-kex-wg0#7 (work_completion)(&peer->transmit_handshake_work) &peer->endpoint_lock rcu_read_lock_bh rcu_read_lock rcu_read_lock_bh &sch->q.lock hrtimer_bases.lock irq_context: 0 (wq_completion)wg-kex-wg0#7 (work_completion)(&peer->transmit_handshake_work) &peer->endpoint_lock rcu_read_lock_bh rcu_read_lock rcu_read_lock_bh &sch->q.lock hrtimer_bases.lock &obj_hash[i].lock irq_context: 0 (wq_completion)wg-kex-wg0#7 (work_completion)(&peer->transmit_handshake_work) &peer->endpoint_lock rcu_read_lock_bh rcu_read_lock rcu_read_lock_bh &sch->q.lock hrtimer_bases.lock tk_core.seq.seqcount irq_context: 0 (wq_completion)wg-kex-wg2#10 (work_completion)(&({ do { const void *__vpp_verify = (typeof((worker) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __asm__ ("" : "=r"(__ptr) : "0"((typeof(*((worker))) *)((worker)))); (typeof((typeof(*((worker))) *)((worker)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); })->work) &wg->static_identity.lock &handshake->lock tk_core.seq.seqcount irq_context: 0 (wq_completion)wg-kex-wg2#10 (work_completion)(&({ do { const void *__vpp_verify = (typeof((worker) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __asm__ ("" : "=r"(__ptr) : "0"((typeof(*((worker))) *)((worker)))); (typeof((typeof(*((worker))) *)((worker)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); })->work) &wg->static_identity.lock &handshake->lock tk_core.seq.seqcount irq_context: 0 (wq_completion)wg-kex-wg2#10 (work_completion)(&({ do { const void *__vpp_verify = (typeof((worker) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __asm__ ("" : "=r"(__ptr) : "0"((typeof(*((worker))) *)((worker)))); (typeof((typeof(*((worker))) *)((worker)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); })->work) &wg->static_identity.lock &handshake->lock crngs.lock irq_context: 0 (wq_completion)wg-kex-wg2#10 (work_completion)(&({ do { const void *__vpp_verify = (typeof((worker) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __asm__ ("" : "=r"(__ptr) : "0"((typeof(*((worker))) *)((worker)))); (typeof((typeof(*((worker))) *)((worker)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); })->work) &wg->static_identity.lock &handshake->lock crngs.lock base_crng.lock irq_context: 0 (wq_completion)wg-kex-wg2#10 (work_completion)(&({ do { const void *__vpp_verify = (typeof((worker) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __asm__ ("" : "=r"(__ptr) : "0"((typeof(*((worker))) *)((worker)))); (typeof((typeof(*((worker))) *)((worker)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); })->work) &wg->static_identity.lock &handshake->lock &table->lock#2 irq_context: 0 (wq_completion)wg-kex-wg2#10 (work_completion)(&({ do { const void *__vpp_verify = (typeof((worker) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __asm__ ("" : "=r"(__ptr) : "0"((typeof(*((worker))) *)((worker)))); (typeof((typeof(*((worker))) *)((worker)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); })->work) &wg->static_identity.lock &handshake->lock rcu_read_lock_bh batched_entropy_u32.lock irq_context: 0 (wq_completion)wg-kex-wg2#10 (work_completion)(&({ do { const void *__vpp_verify = (typeof((worker) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __asm__ ("" : "=r"(__ptr) : "0"((typeof(*((worker))) *)((worker)))); (typeof((typeof(*((worker))) *)((worker)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); })->work) &wg->static_identity.lock &handshake->lock rcu_read_lock_bh batched_entropy_u32.lock crngs.lock irq_context: 0 (wq_completion)wg-kex-wg2#10 (work_completion)(&({ do { const void *__vpp_verify = (typeof((worker) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __asm__ ("" : "=r"(__ptr) : "0"((typeof(*((worker))) *)((worker)))); (typeof((typeof(*((worker))) *)((worker)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); })->work) &wg->static_identity.lock &handshake->lock rcu_read_lock_bh &table->lock#2 irq_context: 0 (wq_completion)wg-kex-wg2#10 (work_completion)(&({ do { const void *__vpp_verify = (typeof((worker) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __asm__ ("" : "=r"(__ptr) : "0"((typeof(*((worker))) *)((worker)))); (typeof((typeof(*((worker))) *)((worker)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); })->work) &cookie->lock irq_context: 0 (wq_completion)wg-kex-wg2#10 (work_completion)(&({ do { const void *__vpp_verify = (typeof((worker) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __asm__ ("" : "=r"(__ptr) : "0"((typeof(*((worker))) *)((worker)))); (typeof((typeof(*((worker))) *)((worker)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); })->work) &cookie->lock irq_context: 0 (wq_completion)wg-kex-wg2#10 (work_completion)(&({ do { const void *__vpp_verify = (typeof((worker) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __asm__ ("" : "=r"(__ptr) : "0"((typeof(*((worker))) *)((worker)))); (typeof((typeof(*((worker))) *)((worker)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); })->work) &peer->endpoint_lock irq_context: 0 (wq_completion)wg-kex-wg2#10 (work_completion)(&({ do { const void *__vpp_verify = (typeof((worker) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __asm__ ("" : "=r"(__ptr) : "0"((typeof(*((worker))) *)((worker)))); (typeof((typeof(*((worker))) *)((worker)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); })->work) &peer->endpoint_lock rcu_read_lock_bh rcu_read_lock rcu_read_lock rcu_read_lock &dir->lock#2 irq_context: 0 (wq_completion)wg-kex-wg2#10 (work_completion)(&({ do { const void *__vpp_verify = (typeof((worker) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __asm__ ("" : "=r"(__ptr) : "0"((typeof(*((worker))) *)((worker)))); (typeof((typeof(*((worker))) *)((worker)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); })->work) &peer->endpoint_lock rcu_read_lock_bh rcu_read_lock &____s->seqcount#7 irq_context: 0 (wq_completion)wg-kex-wg2#10 (work_completion)(&({ do { const void *__vpp_verify = (typeof((worker) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __asm__ ("" : "=r"(__ptr) : "0"((typeof(*((worker))) *)((worker)))); (typeof((typeof(*((worker))) *)((worker)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); })->work) &peer->endpoint_lock rcu_read_lock_bh rcu_read_lock &____s->seqcount#10 irq_context: 0 (wq_completion)wg-kex-wg2#10 (work_completion)(&({ do { const void *__vpp_verify = (typeof((worker) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __asm__ ("" : "=r"(__ptr) : "0"((typeof(*((worker))) *)((worker)))); (typeof((typeof(*((worker))) *)((worker)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); })->work) &peer->endpoint_lock rcu_read_lock_bh rcu_read_lock rcu_read_lock_bh &sch->q.lock irq_context: 0 (wq_completion)wg-kex-wg2#10 (work_completion)(&({ do { const void *__vpp_verify = (typeof((worker) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __asm__ ("" : "=r"(__ptr) : "0"((typeof(*((worker))) *)((worker)))); (typeof((typeof(*((worker))) *)((worker)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); })->work) &peer->endpoint_lock rcu_read_lock_bh rcu_read_lock rcu_read_lock_bh &sch->q.lock &c->lock irq_context: 0 (wq_completion)wg-kex-wg2#10 (work_completion)(&({ do { const void *__vpp_verify = (typeof((worker) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __asm__ ("" : "=r"(__ptr) : "0"((typeof(*((worker))) *)((worker)))); (typeof((typeof(*((worker))) *)((worker)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); })->work) &peer->endpoint_lock rcu_read_lock_bh rcu_read_lock rcu_read_lock_bh &sch->q.lock pool_lock#2 irq_context: 0 (wq_completion)wg-kex-wg2#10 (work_completion)(&({ do { const void *__vpp_verify = (typeof((worker) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __asm__ ("" : "=r"(__ptr) : "0"((typeof(*((worker))) *)((worker)))); (typeof((typeof(*((worker))) *)((worker)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); })->work) &peer->endpoint_lock rcu_read_lock_bh rcu_read_lock rcu_read_lock_bh &sch->q.lock tk_core.seq.seqcount irq_context: 0 (wq_completion)wg-kex-wg0#10 (work_completion)(&({ do { const void *__vpp_verify = (typeof((worker) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __asm__ ("" : "=r"(__ptr) : "0"((typeof(*((worker))) *)((worker)))); (typeof((typeof(*((worker))) *)((worker)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); })->work) &base->lock irq_context: 0 (wq_completion)wg-kex-wg0#10 (work_completion)(&({ do { const void *__vpp_verify = (typeof((worker) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __asm__ ("" : "=r"(__ptr) : "0"((typeof(*((worker))) *)((worker)))); (typeof((typeof(*((worker))) *)((worker)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); })->work) &base->lock &obj_hash[i].lock irq_context: 0 (wq_completion)wg-kex-wg0#10 (work_completion)(&({ do { const void *__vpp_verify = (typeof((worker) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __asm__ ("" : "=r"(__ptr) : "0"((typeof(*((worker))) *)((worker)))); (typeof((typeof(*((worker))) *)((worker)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); })->work) &list->lock#17 irq_context: 0 (wq_completion)wg-kex-wg0#10 (work_completion)(&({ do { const void *__vpp_verify = (typeof((worker) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __asm__ ("" : "=r"(__ptr) : "0"((typeof(*((worker))) *)((worker)))); (typeof((typeof(*((worker))) *)((worker)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); })->work) rcu_read_lock_bh &r->producer_lock#2 irq_context: 0 (wq_completion)wg-kex-wg0#10 (work_completion)(&({ do { const void *__vpp_verify = (typeof((worker) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __asm__ ("" : "=r"(__ptr) : "0"((typeof(*((worker))) *)((worker)))); (typeof((typeof(*((worker))) *)((worker)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); })->work) rcu_read_lock_bh rcu_read_lock &pool->lock irq_context: 0 (wq_completion)wg-kex-wg0#10 (work_completion)(&({ do { const void *__vpp_verify = (typeof((worker) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __asm__ ("" : "=r"(__ptr) : "0"((typeof(*((worker))) *)((worker)))); (typeof((typeof(*((worker))) *)((worker)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); })->work) rcu_read_lock_bh rcu_read_lock &pool->lock &obj_hash[i].lock irq_context: 0 (wq_completion)wg-kex-wg0#10 (work_completion)(&({ do { const void *__vpp_verify = (typeof((worker) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __asm__ ("" : "=r"(__ptr) : "0"((typeof(*((worker))) *)((worker)))); (typeof((typeof(*((worker))) *)((worker)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); })->work) rcu_read_lock_bh rcu_read_lock &pool->lock &p->pi_lock irq_context: 0 (wq_completion)wg-kex-wg0#10 (work_completion)(&({ do { const void *__vpp_verify = (typeof((worker) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __asm__ ("" : "=r"(__ptr) : "0"((typeof(*((worker))) *)((worker)))); (typeof((typeof(*((worker))) *)((worker)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); })->work) rcu_read_lock_bh rcu_read_lock &pool->lock &p->pi_lock &rq->__lock irq_context: 0 (wq_completion)wg-kex-wg0#10 (work_completion)(&({ do { const void *__vpp_verify = (typeof((worker) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __asm__ ("" : "=r"(__ptr) : "0"((typeof(*((worker))) *)((worker)))); (typeof((typeof(*((worker))) *)((worker)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); })->work) rcu_read_lock_bh rcu_read_lock &pool->lock &p->pi_lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)wg-crypt-wg0#4 (work_completion)(&peer->transmit_packet_work) &peer->endpoint_lock rcu_read_lock_bh rcu_read_lock rcu_read_lock_bh &sch->q.lock irq_context: 0 (wq_completion)wg-crypt-wg0#4 (work_completion)(&peer->transmit_packet_work) &peer->endpoint_lock rcu_read_lock_bh rcu_read_lock rcu_read_lock_bh &sch->q.lock pool_lock#2 irq_context: 0 (wq_completion)wg-crypt-wg0#4 (work_completion)(&peer->transmit_packet_work) &peer->endpoint_lock rcu_read_lock_bh rcu_read_lock rcu_read_lock_bh &sch->q.lock tk_core.seq.seqcount irq_context: 0 (wq_completion)wg-crypt-wg0#4 (work_completion)(&peer->transmit_packet_work) &peer->endpoint_lock rcu_read_lock_bh rcu_read_lock rcu_read_lock_bh &sch->q.lock hrtimer_bases.lock irq_context: 0 (wq_completion)wg-crypt-wg0#4 (work_completion)(&peer->transmit_packet_work) &peer->endpoint_lock rcu_read_lock_bh rcu_read_lock rcu_read_lock_bh &sch->q.lock hrtimer_bases.lock &obj_hash[i].lock irq_context: 0 (wq_completion)wg-crypt-wg0#4 (work_completion)(&peer->transmit_packet_work) &peer->endpoint_lock rcu_read_lock_bh rcu_read_lock rcu_read_lock_bh &sch->q.lock hrtimer_bases.lock tk_core.seq.seqcount irq_context: 0 (wq_completion)wg-crypt-wg0#4 (work_completion)(&peer->transmit_packet_work) &peer->endpoint_lock rcu_read_lock_bh rcu_read_lock rcu_read_lock_bh &sch->q.lock &c->lock irq_context: 0 (wq_completion)wg-crypt-wg0#4 (work_completion)(&peer->transmit_packet_work) batched_entropy_u8.lock crngs.lock irq_context: 0 (wq_completion)wg-crypt-wg0#4 (work_completion)(&peer->transmit_packet_work) rcu_read_lock_bh tk_core.seq.seqcount irq_context: 0 (wq_completion)wg-crypt-wg2#5 (work_completion)(&({ do { const void *__vpp_verify = (typeof((worker) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __asm__ ("" : "=r"(__ptr) : "0"((typeof(*((worker))) *)((worker)))); (typeof((typeof(*((worker))) *)((worker)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); })->work) &c->lock irq_context: 0 (wq_completion)wg-crypt-wg2#5 (work_completion)(&peer->transmit_packet_work) &peer->endpoint_lock rcu_read_lock_bh rcu_read_lock rcu_read_lock_bh &sch->q.lock irq_context: 0 (wq_completion)wg-crypt-wg2#5 (work_completion)(&peer->transmit_packet_work) &peer->endpoint_lock rcu_read_lock_bh rcu_read_lock rcu_read_lock_bh &sch->q.lock tk_core.seq.seqcount irq_context: 0 (wq_completion)wg-crypt-wg2#5 (work_completion)(&peer->transmit_packet_work) &peer->endpoint_lock rcu_read_lock_bh rcu_read_lock rcu_read_lock_bh &sch->q.lock hrtimer_bases.lock irq_context: 0 (wq_completion)wg-crypt-wg2#5 (work_completion)(&peer->transmit_packet_work) &peer->endpoint_lock rcu_read_lock_bh rcu_read_lock rcu_read_lock_bh &sch->q.lock hrtimer_bases.lock &obj_hash[i].lock irq_context: 0 (wq_completion)wg-crypt-wg2#5 (work_completion)(&peer->transmit_packet_work) &peer->endpoint_lock rcu_read_lock_bh rcu_read_lock rcu_read_lock_bh &sch->q.lock hrtimer_bases.lock tk_core.seq.seqcount irq_context: 0 pernet_ops_rwsem rtnl_mutex rcu_read_lock &pool->lock/1 fill_pool_map-wait-type-override &c->lock irq_context: 0 pernet_ops_rwsem rtnl_mutex rcu_read_lock &pool->lock/1 fill_pool_map-wait-type-override &n->list_lock irq_context: 0 pernet_ops_rwsem rtnl_mutex rcu_read_lock &pool->lock/1 fill_pool_map-wait-type-override &n->list_lock &c->lock irq_context: 0 (wq_completion)bat_events (work_completion)(&(&bat_priv->mcast.work)->work) fill_pool_map-wait-type-override &c->lock irq_context: 0 pernet_ops_rwsem k-sk_lock-AF_INET6 rcu_node_0 irq_context: 0 pernet_ops_rwsem k-sk_lock-AF_INET6 &rcu_state.expedited_wq irq_context: 0 pernet_ops_rwsem k-sk_lock-AF_INET6 &rcu_state.expedited_wq &p->pi_lock irq_context: 0 pernet_ops_rwsem k-sk_lock-AF_INET6 &rcu_state.expedited_wq &p->pi_lock &rq->__lock irq_context: 0 pernet_ops_rwsem k-sk_lock-AF_INET6 &rcu_state.expedited_wq &p->pi_lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: softirq (&ndev->rs_timer) rcu_read_lock rcu_read_lock rcu_read_lock_bh rcu_read_lock rcu_read_lock rcu_read_lock_bh &sch->q.lock irq_context: softirq (&ndev->rs_timer) rcu_read_lock rcu_read_lock rcu_read_lock_bh rcu_read_lock rcu_read_lock rcu_read_lock_bh &sch->q.lock pool_lock#2 irq_context: softirq (&ndev->rs_timer) rcu_read_lock rcu_read_lock rcu_read_lock_bh rcu_read_lock rcu_read_lock rcu_read_lock_bh &sch->q.lock tk_core.seq.seqcount irq_context: softirq (&ndev->rs_timer) rcu_read_lock rcu_read_lock rcu_read_lock_bh rcu_read_lock rcu_read_lock rcu_read_lock_bh &sch->q.lock hrtimer_bases.lock irq_context: softirq (&ndev->rs_timer) rcu_read_lock rcu_read_lock rcu_read_lock_bh rcu_read_lock rcu_read_lock rcu_read_lock_bh &sch->q.lock hrtimer_bases.lock &obj_hash[i].lock irq_context: softirq (&ndev->rs_timer) rcu_read_lock rcu_read_lock rcu_read_lock_bh rcu_read_lock rcu_read_lock rcu_read_lock_bh &sch->q.lock hrtimer_bases.lock tk_core.seq.seqcount irq_context: 0 (wq_completion)bond10 (work_completion)(&(&bond->slave_arr_work)->work) rtnl_mutex &obj_hash[i].lock irq_context: 0 (wq_completion)bond10 (work_completion)(&(&bond->slave_arr_work)->work) rtnl_mutex krc.lock irq_context: 0 (wq_completion)wg-kex-wg0#7 (work_completion)(&peer->transmit_handshake_work) &peer->endpoint_lock rcu_read_lock_bh rcu_read_lock &____s->seqcount#10 irq_context: softirq rcu_read_lock rcu_read_lock k-slock-AF_INET6 rcu_read_lock rcu_read_lock rcu_read_lock_bh &sch->q.lock irq_context: softirq rcu_read_lock rcu_read_lock k-slock-AF_INET6 rcu_read_lock rcu_read_lock rcu_read_lock_bh &sch->q.lock pool_lock#2 irq_context: softirq rcu_read_lock rcu_read_lock k-slock-AF_INET6 rcu_read_lock rcu_read_lock rcu_read_lock_bh &sch->q.lock tk_core.seq.seqcount irq_context: softirq rcu_read_lock rcu_read_lock k-slock-AF_INET6 rcu_read_lock rcu_read_lock rcu_read_lock_bh &sch->q.lock &c->lock irq_context: softirq rcu_read_lock rcu_read_lock k-slock-AF_INET6 rcu_read_lock rcu_read_lock rcu_read_lock_bh &sch->q.lock &n->list_lock irq_context: softirq rcu_read_lock rcu_read_lock k-slock-AF_INET6 rcu_read_lock rcu_read_lock rcu_read_lock_bh &sch->q.lock &n->list_lock &c->lock irq_context: softirq rcu_read_lock rcu_read_lock k-slock-AF_INET6 rcu_read_lock rcu_read_lock rcu_read_lock_bh &sch->q.lock hrtimer_bases.lock irq_context: softirq rcu_read_lock rcu_read_lock k-slock-AF_INET6 rcu_read_lock rcu_read_lock rcu_read_lock_bh &sch->q.lock hrtimer_bases.lock &obj_hash[i].lock irq_context: softirq rcu_read_lock rcu_read_lock k-slock-AF_INET6 rcu_read_lock rcu_read_lock rcu_read_lock_bh &sch->q.lock hrtimer_bases.lock tk_core.seq.seqcount irq_context: 0 pernet_ops_rwsem __ip_vs_app_mutex __ip_vs_app_mutex.wait_lock irq_context: 0 pernet_ops_rwsem __ip_vs_app_mutex &cfs_rq->removed.lock irq_context: 0 pernet_ops_rwsem __ip_vs_app_mutex.wait_lock irq_context: 0 &pipe->mutex/1 nfnl_subsys_ctnetlink rcu_read_lock nl_table_wait.lock &p->pi_lock irq_context: 0 &pipe->mutex/1 nfnl_subsys_ctnetlink rcu_read_lock rcu_read_lock &data->lock irq_context: 0 &pipe->mutex/1 nfnl_subsys_ctnetlink rcu_read_lock &data->lock irq_context: 0 &pipe->mutex/1 nfnl_subsys_ctnetlink rcu_read_lock &data->lock &obj_hash[i].lock irq_context: 0 &pipe->mutex/1 nfnl_subsys_ctnetlink rcu_read_lock &data->lock &base->lock irq_context: 0 &pipe->mutex/1 nfnl_subsys_ctnetlink rcu_read_lock &data->lock &base->lock &obj_hash[i].lock irq_context: 0 &mm->mmap_lock &mapping->i_mmap_rwsem fill_pool_map-wait-type-override rcu_read_lock rcu_node_0 irq_context: 0 &mm->mmap_lock &mapping->i_mmap_rwsem fill_pool_map-wait-type-override rcu_read_lock &rq->__lock irq_context: 0 &mm->mmap_lock &mapping->i_mmap_rwsem fill_pool_map-wait-type-override rcu_read_lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 &pipe->mutex/1 nfnl_subsys_ctnetlink rcu_read_lock &cfs_rq->removed.lock irq_context: 0 (wq_completion)bond13 (work_completion)(&(&bond->ad_work)->work) &rq->__lock irq_context: 0 (wq_completion)bond13 (work_completion)(&(&bond->ad_work)->work) &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 pernet_ops_rwsem cpu_hotplug_lock wq_pool_mutex fs_reclaim &rq->__lock irq_context: 0 pernet_ops_rwsem cpu_hotplug_lock wq_pool_mutex fs_reclaim &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem pool_lock irq_context: 0 (wq_completion)events (work_completion)(&(&cnet->ecache.dwork)->work) rcu_read_lock &pool->lock fill_pool_map-wait-type-override pool_lock#2 irq_context: 0 (wq_completion)events (work_completion)(&(&cnet->ecache.dwork)->work) rcu_read_lock &pool->lock fill_pool_map-wait-type-override &c->lock irq_context: 0 (wq_completion)events (work_completion)(&(&cnet->ecache.dwork)->work) rcu_read_lock &pool->lock fill_pool_map-wait-type-override pool_lock irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem rtnl_mutex bpf_devs_lock rcu_read_lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)wg-crypt-wg2#5 (work_completion)(&({ do { const void *__vpp_verify = (typeof((worker) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __asm__ ("" : "=r"(__ptr) : "0"((typeof(*((worker))) *)((worker)))); (typeof((typeof(*((worker))) *)((worker)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); })->work) &obj_hash[i].lock irq_context: 0 pernet_ops_rwsem rcu_state.exp_mutex fill_pool_map-wait-type-override pool_lock#2 irq_context: 0 pernet_ops_rwsem rcu_state.exp_mutex fill_pool_map-wait-type-override &c->lock irq_context: 0 pernet_ops_rwsem rcu_state.exp_mutex fill_pool_map-wait-type-override pool_lock irq_context: 0 sk_lock-AF_INET rcu_read_lock &ct->lock irq_context: 0 sk_lock-AF_INET rcu_read_lock rcu_read_lock_bh &data->lock irq_context: 0 sk_lock-AF_INET &data->lock irq_context: 0 (wq_completion)bat_events (work_completion)(&(&forw_packet_aggr->delayed_work)->work) &hard_iface->bat_iv.ogm_buff_mutex fill_pool_map-wait-type-override &n->list_lock irq_context: 0 (wq_completion)bat_events (work_completion)(&(&forw_packet_aggr->delayed_work)->work) &hard_iface->bat_iv.ogm_buff_mutex fill_pool_map-wait-type-override &n->list_lock &c->lock irq_context: 0 (wq_completion)bat_events (work_completion)(&(&forw_packet_aggr->delayed_work)->work) &hard_iface->bat_iv.ogm_buff_mutex fill_pool_map-wait-type-override &____s->seqcount#2 irq_context: 0 (wq_completion)bat_events (work_completion)(&(&forw_packet_aggr->delayed_work)->work) &hard_iface->bat_iv.ogm_buff_mutex fill_pool_map-wait-type-override &____s->seqcount irq_context: 0 (wq_completion)events (work_completion)(&pwq->unbound_release_work) rcu_state.exp_mutex rcu_read_lock &pool->lock &p->pi_lock irq_context: 0 (wq_completion)events (work_completion)(&pwq->unbound_release_work) rcu_state.exp_mutex rcu_read_lock &pool->lock &p->pi_lock &rq->__lock irq_context: 0 pernet_ops_rwsem cpu_hotplug_lock wq_pool_mutex rcu_read_lock &rcu_state.expedited_wq irq_context: 0 pernet_ops_rwsem cpu_hotplug_lock wq_pool_mutex rcu_read_lock &rcu_state.expedited_wq &p->pi_lock irq_context: 0 (wq_completion)events (work_completion)(&pwq->unbound_release_work) rcu_state.exp_mutex rcu_read_lock &pool->lock &p->pi_lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 pernet_ops_rwsem cpu_hotplug_lock wq_pool_mutex rcu_read_lock &rcu_state.expedited_wq &p->pi_lock &rq->__lock irq_context: 0 pernet_ops_rwsem cpu_hotplug_lock wq_pool_mutex rcu_read_lock &rcu_state.expedited_wq &p->pi_lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 sk_lock-AF_INET6 fill_pool_map-wait-type-override batched_entropy_u8.lock irq_context: 0 sk_lock-AF_INET6 fill_pool_map-wait-type-override kfence_freelist_lock irq_context: 0 sb_writers#4 mapping.invalidate_lock rcu_read_lock &cfs_rq->removed.lock irq_context: 0 &f->f_pos_lock sb_writers#4 &sb->s_type->i_mutex_key#8 jbd2_handle &ei->i_data_sem &n->list_lock irq_context: 0 &f->f_pos_lock sb_writers#4 &sb->s_type->i_mutex_key#8 jbd2_handle &ei->i_data_sem &n->list_lock &c->lock irq_context: 0 &sb->s_type->i_mutex_key#10 pcpu_lock stock_lock irq_context: softirq rcu_read_lock rcu_read_lock slock-AF_INET6/1 sk_lock-AF_INET6 rcu_read_lock &ei->socket.wq.wait irq_context: softirq rcu_read_lock rcu_read_lock slock-AF_INET6/1 sk_lock-AF_INET6 rcu_read_lock &ei->socket.wq.wait &p->pi_lock irq_context: softirq rcu_read_lock rcu_read_lock slock-AF_INET6/1 sk_lock-AF_INET6 rcu_read_lock &ei->socket.wq.wait &p->pi_lock &rq->__lock irq_context: softirq rcu_read_lock rcu_read_lock slock-AF_INET6/1 sk_lock-AF_INET6 rcu_read_lock &ei->socket.wq.wait &p->pi_lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: softirq rcu_read_lock rcu_read_lock slock-AF_INET6/1 sk_lock-AF_INET6 rcu_read_lock rcu_read_lock rcu_read_lock_bh &sch->q.lock rcu_read_lock &ei->socket.wq.wait irq_context: softirq rcu_read_lock rcu_read_lock slock-AF_INET6/1 sk_lock-AF_INET6 rcu_read_lock rcu_read_lock rcu_read_lock_bh &sch->q.lock rcu_read_lock &ei->socket.wq.wait &p->pi_lock irq_context: 0 sk_lock-AF_BLUETOOTH-BTPROTO_HCI mgmt_chan_list_lock &data->lock irq_context: 0 sk_lock-AF_BLUETOOTH-BTPROTO_HCI mgmt_chan_list_lock fs_reclaim irq_context: 0 sk_lock-AF_BLUETOOTH-BTPROTO_HCI mgmt_chan_list_lock fs_reclaim mmu_notifier_invalidate_range_start irq_context: 0 sk_lock-AF_BLUETOOTH-BTPROTO_HCI mgmt_chan_list_lock rlock-AF_BLUETOOTH irq_context: 0 &sb->s_type->i_mutex_key#10 sk_lock-AF_BLUETOOTH-BTPROTO_HCI &data->lock irq_context: 0 sb_writers#4 &sb->s_type->i_mutex_key#8 rcu_read_lock &pcp->lock &zone->lock irq_context: 0 (wq_completion)bond13 (work_completion)(&(&bond->mii_work)->work) rcu_read_lock rcu_read_lock rcu_node_0 irq_context: 0 (wq_completion)bond13 (work_completion)(&(&bond->mii_work)->work) rcu_read_lock rcu_read_lock &rq->__lock irq_context: 0 (wq_completion)bond13 (work_completion)(&(&bond->mii_work)->work) rcu_read_lock rcu_node_0 irq_context: 0 sk_lock-AF_BLUETOOTH-BTPROTO_HCI mgmt_chan_list_lock &____s->seqcount#2 irq_context: 0 sk_lock-AF_BLUETOOTH-BTPROTO_HCI mgmt_chan_list_lock &____s->seqcount irq_context: 0 rtnl_mutex &tn->idrinfo->lock#4 irq_context: 0 rtnl_mutex &tn->idrinfo->lock#4 fs_reclaim irq_context: 0 rtnl_mutex &tn->idrinfo->lock#4 fs_reclaim mmu_notifier_invalidate_range_start irq_context: 0 rtnl_mutex &tn->idrinfo->lock#4 pool_lock#2 irq_context: 0 rtnl_mutex &tn->idrinfo->lock#4 &obj_hash[i].lock irq_context: 0 rtnl_mutex &tn->idrinfo->lock#4 &c->lock irq_context: 0 rtnl_mutex &tn->idrinfo->lock#4 &n->list_lock irq_context: 0 rtnl_mutex &tn->idrinfo->lock#4 &n->list_lock &c->lock irq_context: 0 rtnl_mutex &tn->idrinfo->lock#4 &rq->__lock irq_context: 0 rtnl_mutex &tn->idrinfo->lock#4 &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 sk_lock-AF_RXRPC irq_context: 0 sk_lock-AF_RXRPC slock-AF_RXRPC irq_context: 0 sk_lock-AF_RXRPC &rxnet->local_mutex irq_context: 0 sk_lock-AF_RXRPC &rxnet->local_mutex fs_reclaim irq_context: 0 sk_lock-AF_RXRPC &rxnet->local_mutex fs_reclaim mmu_notifier_invalidate_range_start irq_context: 0 sk_lock-AF_RXRPC &rxnet->local_mutex fs_reclaim mmu_notifier_invalidate_range_start &rq->__lock irq_context: 0 sk_lock-AF_RXRPC &rxnet->local_mutex pool_lock#2 irq_context: 0 sk_lock-AF_RXRPC &rxnet->local_mutex &obj_hash[i].lock irq_context: 0 sk_lock-AF_RXRPC &rxnet->local_mutex crngs.lock irq_context: 0 sk_lock-AF_RXRPC &rxnet->local_mutex stock_lock irq_context: 0 sk_lock-AF_RXRPC &rxnet->local_mutex mmu_notifier_invalidate_range_start irq_context: 0 sk_lock-AF_RXRPC &rxnet->local_mutex &sb->s_type->i_lock_key#8 irq_context: 0 sk_lock-AF_RXRPC &rxnet->local_mutex &dir->lock irq_context: 0 sk_lock-AF_RXRPC &rxnet->local_mutex k-sk_lock-AF_INET irq_context: 0 sk_lock-AF_RXRPC &rxnet->local_mutex k-sk_lock-AF_INET k-slock-AF_INET irq_context: 0 sk_lock-AF_RXRPC &rxnet->local_mutex k-sk_lock-AF_INET &____s->seqcount#8 irq_context: 0 sk_lock-AF_RXRPC &rxnet->local_mutex k-sk_lock-AF_INET batched_entropy_u32.lock irq_context: 0 sk_lock-AF_RXRPC &rxnet->local_mutex k-sk_lock-AF_INET &table->hash[i].lock irq_context: 0 sk_lock-AF_RXRPC &rxnet->local_mutex k-sk_lock-AF_INET &table->hash[i].lock k-clock-AF_INET irq_context: 0 sk_lock-AF_RXRPC &rxnet->local_mutex k-sk_lock-AF_INET &table->hash[i].lock &table->hash2[i].lock irq_context: 0 sk_lock-AF_RXRPC &rxnet->local_mutex k-slock-AF_INET irq_context: 0 sk_lock-AF_RXRPC &rxnet->local_mutex cpu_hotplug_lock irq_context: 0 sk_lock-AF_RXRPC &rxnet->local_mutex &c->lock irq_context: 0 sk_lock-AF_RXRPC &rxnet->local_mutex kthread_create_lock irq_context: 0 sk_lock-AF_RXRPC &rxnet->local_mutex &p->pi_lock irq_context: 0 sk_lock-AF_RXRPC &rxnet->local_mutex &p->pi_lock &rq->__lock irq_context: 0 sk_lock-AF_RXRPC &rxnet->local_mutex &p->pi_lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 sk_lock-AF_RXRPC &rxnet->local_mutex &rq->__lock irq_context: 0 sk_lock-AF_RXRPC &rxnet->local_mutex &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 sk_lock-AF_RXRPC &rxnet->local_mutex &x->wait irq_context: 0 sk_lock-AF_RXRPC &rxnet->local_mutex &x->wait#22 irq_context: 0 slock-AF_RXRPC irq_context: 0 (wq_completion)wg-crypt-wg0#4 (work_completion)(&peer->transmit_packet_work) &peer->endpoint_lock rcu_read_lock_bh rcu_read_lock rcu_read_lock_bh &sch->q.lock &____s->seqcount#2 irq_context: 0 (wq_completion)wg-crypt-wg0#4 (work_completion)(&peer->transmit_packet_work) &peer->endpoint_lock rcu_read_lock_bh rcu_read_lock rcu_read_lock_bh &sch->q.lock &____s->seqcount irq_context: 0 sk_lock-AF_INET6 &dccp_hashinfo.bhash[i].lock stock_lock rcu_read_lock per_cpu_ptr(&cgroup_rstat_cpu_lock, cpu) irq_context: 0 sk_lock-AF_TIPC &data->lock irq_context: 0 &sb->s_type->i_mutex_key#8 &sbi->s_writepages_rwsem jbd2_handle &ei->i_data_sem rcu_read_lock &cfs_rq->removed.lock irq_context: 0 &sb->s_type->i_mutex_key#8 &sbi->s_writepages_rwsem jbd2_handle &ei->i_data_sem rcu_read_lock &obj_hash[i].lock irq_context: 0 k-clock-AF_INET irq_context: 0 &sb->s_type->i_mutex_key#10 &x->wait irq_context: 0 sk_lock-AF_RXRPC &rxnet->local_mutex remove_cache_srcu irq_context: 0 sk_lock-AF_RXRPC &rxnet->local_mutex remove_cache_srcu quarantine_lock irq_context: 0 sk_lock-AF_RXRPC &rxnet->local_mutex remove_cache_srcu &c->lock irq_context: 0 sk_lock-AF_RXRPC &rxnet->local_mutex remove_cache_srcu &n->list_lock irq_context: 0 sk_lock-AF_RXRPC &rxnet->local_mutex remove_cache_srcu &obj_hash[i].lock irq_context: 0 sk_lock-AF_RXRPC &rxnet->local_mutex remove_cache_srcu pool_lock#2 irq_context: 0 sk_lock-AF_RXRPC &rxnet->local_mutex remove_cache_srcu rcu_read_lock pool_lock#2 irq_context: 0 sk_lock-AF_RXRPC &rxnet->local_mutex &lock->wait_lock irq_context: 0 sk_lock-AF_RXRPC &rxnet->local_mutex &table->hash[i].lock irq_context: 0 sk_lock-AF_RXRPC &rxnet->local_mutex &table->hash[i].lock &table->hash2[i].lock irq_context: 0 sk_lock-AF_RXRPC &rxnet->local_mutex k-clock-AF_INET irq_context: 0 sk_lock-AF_RXRPC &rxnet->local_mutex &xa->xa_lock#9 irq_context: 0 sk_lock-AF_RXRPC &rxnet->local_mutex &fsnotify_mark_srcu irq_context: 0 sk_lock-AF_RXRPC &lock->wait_lock irq_context: 0 sk_lock-AF_RXRPC &p->pi_lock irq_context: 0 sk_lock-AF_RXRPC &p->pi_lock &rq->__lock irq_context: 0 sk_lock-AF_RXRPC &p->pi_lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 sk_lock-AF_RXRPC &rq->__lock irq_context: 0 sk_lock-AF_RXRPC &obj_hash[i].lock irq_context: 0 sk_lock-AF_RXRPC pool_lock#2 irq_context: 0 sk_lock-AF_TIPC &pcp->lock &zone->lock &____s->seqcount irq_context: 0 &sb->s_type->i_mutex_key#8 &sbi->s_writepages_rwsem rcu_read_lock &rq->__lock &cfs_rq->removed.lock irq_context: 0 &sbi->s_writepages_rwsem remove_cache_srcu rcu_read_lock rcu_node_0 irq_context: 0 &sbi->s_writepages_rwsem remove_cache_srcu rcu_read_lock &rq->__lock irq_context: 0 &f->f_pos_lock &mm->mmap_lock fs_reclaim irq_context: 0 &f->f_pos_lock &mm->mmap_lock fs_reclaim &rq->__lock irq_context: 0 &f->f_pos_lock &mm->mmap_lock fs_reclaim &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 &f->f_pos_lock &mm->mmap_lock fs_reclaim mmu_notifier_invalidate_range_start irq_context: 0 &f->f_pos_lock &mm->mmap_lock &____s->seqcount irq_context: 0 &f->f_pos_lock &mm->mmap_lock stock_lock irq_context: 0 &f->f_pos_lock &mm->mmap_lock &obj_hash[i].lock irq_context: 0 (wq_completion)hci1 (work_completion)(&hdev->cmd_sync_work) &hdev->req_lock &hdev->lock &hdev->unregister_lock irq_context: 0 (wq_completion)hci1 (work_completion)(&hdev->cmd_sync_work) &hdev->req_lock &hdev->lock &hdev->unregister_lock fs_reclaim irq_context: 0 (wq_completion)hci1 (work_completion)(&hdev->cmd_sync_work) &hdev->req_lock &hdev->lock &hdev->unregister_lock fs_reclaim mmu_notifier_invalidate_range_start irq_context: 0 (wq_completion)hci1 (work_completion)(&hdev->cmd_sync_work) &hdev->req_lock &hdev->lock &hdev->unregister_lock pool_lock#2 irq_context: 0 (wq_completion)hci1 (work_completion)(&hdev->cmd_sync_work) &hdev->req_lock &hdev->lock &hdev->unregister_lock &hdev->cmd_sync_work_lock irq_context: 0 (wq_completion)hci1 (work_completion)(&hdev->cmd_sync_work) &hdev->req_lock &hdev->lock &hdev->unregister_lock rcu_read_lock &pool->lock/1 irq_context: 0 (wq_completion)hci1 (work_completion)(&hdev->cmd_sync_work) &hdev->req_lock &hdev->lock &hdev->unregister_lock rcu_read_lock &pool->lock/1 &obj_hash[i].lock irq_context: 0 (wq_completion)hci1 (work_completion)(&hdev->cmd_sync_work) &hdev->req_lock &hdev->lock &hdev->unregister_lock rcu_read_lock &pool->lock/1 &p->pi_lock irq_context: 0 (wq_completion)hci1 (work_completion)(&hdev->cmd_sync_work) &hdev->req_lock &hdev->lock &hdev->unregister_lock rcu_read_lock &pool->lock/1 &p->pi_lock &rq->__lock irq_context: 0 (wq_completion)hci1 (work_completion)(&hdev->cmd_sync_work) &hdev->req_lock &hdev->lock &hdev->unregister_lock rcu_read_lock &pool->lock/1 &p->pi_lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)hci1 (work_completion)(&hdev->cmd_sync_work) &hdev->req_lock &hdev->lock &hdev->unregister_lock &rq->__lock irq_context: 0 (wq_completion)hci1 (work_completion)(&hdev->cmd_sync_work) &hdev->req_lock &hdev->lock pool_lock#2 irq_context: 0 (wq_completion)hci1 (work_completion)(&hdev->cmd_sync_work) &hdev->req_lock &hdev->lock rlock-AF_BLUETOOTH irq_context: 0 (wq_completion)hci1 (work_completion)(&hdev->cmd_sync_work) &hdev->req_lock &hdev->lock &data->lock irq_context: 0 &f->f_pos_lock &mm->mmap_lock &sem->wait_lock irq_context: 0 &f->f_pos_lock &sem->wait_lock irq_context: 0 &f->f_pos_lock &p->pi_lock irq_context: 0 &f->f_pos_lock &p->pi_lock &rq->__lock irq_context: 0 &f->f_pos_lock &p->pi_lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 &f->f_pos_lock &mm->mmap_lock &c->lock irq_context: 0 &f->f_pos_lock &mm->mmap_lock &____s->seqcount#2 irq_context: 0 cb_lock rtnl_mutex &rdev->wiphy.mtx &sb->s_type->i_mutex_key#3 rcu_read_lock &rcu_state.gp_wq irq_context: 0 cb_lock rtnl_mutex &rdev->wiphy.mtx &sb->s_type->i_mutex_key#3 rcu_read_lock &rcu_state.gp_wq &p->pi_lock irq_context: 0 cb_lock rtnl_mutex &rdev->wiphy.mtx &sb->s_type->i_mutex_key#3 rcu_read_lock &rcu_state.gp_wq &p->pi_lock &rq->__lock irq_context: 0 cb_lock rtnl_mutex &rdev->wiphy.mtx &sb->s_type->i_mutex_key#3 rcu_read_lock &rcu_state.gp_wq &p->pi_lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 cb_lock rtnl_mutex &rdev->wiphy.mtx &sb->s_type->i_mutex_key#3 &pcp->lock &zone->lock &____s->seqcount irq_context: 0 (wq_completion)hci1 (work_completion)(&hdev->cmd_sync_work) &hdev->req_lock &hdev->lock &lock->wait_lock irq_context: 0 (wq_completion)hci1 (work_completion)(&hdev->cmd_sync_work) &hdev->req_lock &hdev->lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 sk_lock-AF_BLUETOOTH-BTPROTO_HCI mgmt_chan_list_lock &lock->wait_lock irq_context: 0 sk_lock-AF_BLUETOOTH-BTPROTO_HCI mgmt_chan_list_lock &p->pi_lock irq_context: 0 sk_lock-AF_BLUETOOTH-BTPROTO_HCI mgmt_chan_list_lock &p->pi_lock &rq->__lock irq_context: 0 sk_lock-AF_BLUETOOTH-BTPROTO_HCI mgmt_chan_list_lock &p->pi_lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)wg-crypt-wg0#4 (work_completion)(&peer->transmit_packet_work) rcu_node_0 irq_context: 0 cb_lock genl_mutex rcu_read_lock &tipc_net(net)->bclock irq_context: 0 rtnl_mutex &br->lock lweventlist_lock &n->list_lock irq_context: 0 rtnl_mutex &br->lock lweventlist_lock &n->list_lock &c->lock irq_context: 0 cb_lock nlk_cb_mutex-GENERIC rcu_node_0 irq_context: 0 cb_lock nlk_cb_mutex-GENERIC &rcu_state.expedited_wq irq_context: 0 cb_lock nlk_cb_mutex-GENERIC &rcu_state.expedited_wq &p->pi_lock irq_context: 0 cb_lock nlk_cb_mutex-GENERIC &rcu_state.expedited_wq &p->pi_lock &rq->__lock irq_context: 0 cb_lock nlk_cb_mutex-GENERIC &rcu_state.expedited_wq &p->pi_lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)events deferred_process_work &rq->__lock irq_context: 0 nlk_cb_mutex-GENERIC &n->list_lock irq_context: 0 nlk_cb_mutex-GENERIC &n->list_lock &c->lock irq_context: 0 nlk_cb_mutex-GENERIC &____s->seqcount#2 irq_context: 0 nlk_cb_mutex-GENERIC &____s->seqcount irq_context: 0 ebt_mutex &cfs_rq->removed.lock irq_context: 0 ebt_mutex &obj_hash[i].lock irq_context: 0 (wq_completion)bat_events (work_completion)(&(&forw_packet_aggr->delayed_work)->work) &hard_iface->bat_iv.ogm_buff_mutex &bat_priv->tt.commit_lock &bat_priv->softif_vlan_list_lock &____s->seqcount#2 irq_context: 0 sk_lock-AF_NETLINK rcu_node_0 irq_context: 0 &wg->device_update_lock rcu_state.exp_mutex &cfs_rq->removed.lock irq_context: 0 &wg->device_update_lock rcu_state.exp_mutex pool_lock#2 irq_context: 0 &sb->s_type->i_mutex_key#10 &smc->clcsock_release_lock k-sk_lock-AF_INET &rq->__lock irq_context: 0 &sb->s_type->i_mutex_key#10 &smc->clcsock_release_lock k-sk_lock-AF_INET &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 sb_writers#4 &sb->s_type->i_mutex_key#8 &sbi->s_writepages_rwsem &rnp->exp_wq[2] irq_context: 0 &mm->mmap_lock fs_reclaim rcu_node_0 irq_context: 0 &mm->mmap_lock fs_reclaim &rcu_state.expedited_wq irq_context: 0 &mm->mmap_lock fs_reclaim &rcu_state.expedited_wq &p->pi_lock irq_context: 0 &mm->mmap_lock fs_reclaim &rcu_state.expedited_wq &p->pi_lock &rq->__lock irq_context: 0 &mm->mmap_lock sb_writers#4 jbd2_handle &ret->b_state_lock bit_wait_table + i irq_context: 0 &mm->mmap_lock fs_reclaim &rcu_state.expedited_wq &p->pi_lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 rtnl_mutex &idev->mc_lock &dev_addr_list_lock_key#3/1 irq_context: 0 rtnl_mutex &idev->mc_lock &dev_addr_list_lock_key#3/1 &obj_hash[i].lock irq_context: 0 rtnl_mutex &idev->mc_lock &dev_addr_list_lock_key#3/1 pool_lock#2 irq_context: 0 rtnl_mutex &idev->mc_lock &dev_addr_list_lock_key#3/1 krc.lock irq_context: 0 rtnl_mutex dev_pm_qos_sysfs_mtx &root->kernfs_rwsem &sem->wait_lock irq_context: 0 rtnl_mutex dev_pm_qos_sysfs_mtx &root->kernfs_rwsem &rq->__lock irq_context: 0 rtnl_mutex dev_pm_qos_sysfs_mtx &root->kernfs_rwsem &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 rtnl_mutex dev_pm_qos_sysfs_mtx &sem->wait_lock irq_context: 0 rtnl_mutex dev_pm_qos_sysfs_mtx &p->pi_lock irq_context: 0 rtnl_mutex dev_pm_qos_sysfs_mtx &p->pi_lock &rq->__lock irq_context: 0 rtnl_mutex dev_pm_qos_sysfs_mtx &p->pi_lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 rtnl_mutex nf_hook_mutex cpu_hotplug_lock irq_context: 0 sk_lock-AF_INET rcu_read_lock rcu_read_lock rcu_read_lock_bh &data->lock irq_context: 0 rtnl_mutex &nft_net->commit_mutex &rq->__lock irq_context: 0 rtnl_mutex &nft_net->commit_mutex &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 rtnl_mutex &idev->mc_lock &dev_addr_list_lock_key#3/2 pool_lock#2 irq_context: 0 (wq_completion)events (work_completion)(&data->dm_alert_work) quarantine_lock irq_context: 0 loop_validate_mutex loop_validate_mutex.wait_lock irq_context: 0 loop_validate_mutex.wait_lock irq_context: softirq rcu_read_lock rcu_read_lock &dreq->dreq_lock slock-AF_INET6 &dccp_hashinfo.bhash[i].lock &dccp_hashinfo.bhash2[i].lock &n->list_lock irq_context: 0 sb_writers#4 &sb->s_type->i_mutex_key#8 &sbi->s_writepages_rwsem jbd2_handle &ei->i_data_sem &lg->lg_mutex irq_context: 0 sb_writers#4 &sb->s_type->i_mutex_key#8 &sbi->s_writepages_rwsem jbd2_handle &ei->i_data_sem &lg->lg_mutex &ei->i_prealloc_lock irq_context: 0 sb_writers#4 &sb->s_type->i_mutex_key#8 &sbi->s_writepages_rwsem jbd2_handle &ei->i_data_sem &lg->lg_mutex rcu_read_lock &pa->pa_lock irq_context: 0 sb_writers#4 &sb->s_type->i_mutex_key#8 &sbi->s_writepages_rwsem jbd2_handle &ei->i_data_sem &lg->lg_mutex &mapping->private_lock irq_context: 0 sb_writers#4 &sb->s_type->i_mutex_key#8 &sbi->s_writepages_rwsem jbd2_handle &ei->i_data_sem &lg->lg_mutex &ret->b_state_lock irq_context: 0 sb_writers#4 &sb->s_type->i_mutex_key#8 &sbi->s_writepages_rwsem jbd2_handle &ei->i_data_sem &lg->lg_mutex &ret->b_state_lock &journal->j_list_lock irq_context: 0 sb_writers#4 &sb->s_type->i_mutex_key#8 &sbi->s_writepages_rwsem jbd2_handle &ei->i_data_sem &lg->lg_mutex &pa->pa_lock irq_context: 0 sb_writers#4 &sb->s_type->i_mutex_key#8 &sbi->s_writepages_rwsem jbd2_handle &ei->i_data_sem &lg->lg_mutex &lg->lg_prealloc_lock irq_context: 0 kn->active#60 &kernfs_locks->open_file_mutex[count] remove_cache_srcu irq_context: 0 kn->active#60 &kernfs_locks->open_file_mutex[count] remove_cache_srcu quarantine_lock irq_context: 0 loop_validate_mutex &cfs_rq->removed.lock irq_context: 0 loop_validate_mutex &obj_hash[i].lock irq_context: 0 loop_validate_mutex pool_lock#2 irq_context: 0 kn->active#60 rcu_read_lock &rcu_state.gp_wq irq_context: 0 kn->active#60 rcu_read_lock &rcu_state.gp_wq &p->pi_lock irq_context: 0 kn->active#60 rcu_read_lock &rcu_state.gp_wq &p->pi_lock &rq->__lock irq_context: 0 kn->active#60 rcu_read_lock &rcu_state.gp_wq &p->pi_lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 &sb->s_type->i_mutex_key#10 &smc->clcsock_release_lock &meta->lock irq_context: 0 &sb->s_type->i_mutex_key#10 &smc->clcsock_release_lock kfence_freelist_lock irq_context: 0 &sb->s_type->i_mutex_key#16 &rq->__lock irq_context: 0 &sb->s_type->i_mutex_key#16 &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 &xt[i].mutex batched_entropy_u8.lock crngs.lock irq_context: 0 sk_lock-AF_CAN &obj_hash[i].lock irq_context: 0 sk_lock-AF_CAN &____s->seqcount#2 irq_context: 0 sk_lock-AF_CAN &____s->seqcount irq_context: 0 sk_lock-AF_CAN &c->lock irq_context: 0 sk_lock-AF_CAN &net->can.rcvlists_lock irq_context: 0 sb_writers#4 &sb->s_type->i_mutex_key#8 &sbi->s_writepages_rwsem &ei->i_data_sem &rq->__lock irq_context: 0 sb_writers#4 &sb->s_type->i_mutex_key#8 &sbi->s_writepages_rwsem &ei->i_data_sem &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 &sb->s_type->i_mutex_key#10 sk_lock-AF_CAN &net->can.rcvlists_lock irq_context: 0 &type->i_mutex_dir_key#3 remove_cache_srcu irq_context: 0 &type->i_mutex_dir_key#3 remove_cache_srcu quarantine_lock irq_context: 0 &type->i_mutex_dir_key#3 remove_cache_srcu &c->lock irq_context: 0 &type->i_mutex_dir_key#3 remove_cache_srcu &n->list_lock irq_context: 0 &type->i_mutex_dir_key#3 remove_cache_srcu &obj_hash[i].lock irq_context: 0 &type->i_mutex_dir_key#3 remove_cache_srcu pool_lock#2 irq_context: 0 &type->i_mutex_dir_key#3 remove_cache_srcu rcu_read_lock pool_lock#2 irq_context: 0 sk_lock-AF_CAN &n->list_lock irq_context: 0 sk_lock-AF_CAN &n->list_lock &c->lock irq_context: 0 &f->f_pos_lock sb_writers#4 &sb->s_type->i_mutex_key#8 &sbi->s_writepages_rwsem jbd2_handle &ei->i_data_sem &ei->i_es_lock &c->lock irq_context: 0 &f->f_pos_lock sb_writers#4 &sb->s_type->i_mutex_key#8 mapping.invalidate_lock jbd2_handle rcu_read_lock rcu_node_0 irq_context: 0 &f->f_pos_lock sb_writers#4 &sb->s_type->i_mutex_key#8 mapping.invalidate_lock jbd2_handle rcu_read_lock &rq->__lock irq_context: 0 &f->f_pos_lock sb_writers#4 &sb->s_type->i_mutex_key#8 mapping.invalidate_lock jbd2_handle rcu_read_lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 &sb->s_type->i_mutex_key#9 &sb->s_type->i_lock_key#23 &dentry->d_lock &wq#2 &p->pi_lock irq_context: 0 &sb->s_type->i_mutex_key#9 &sb->s_type->i_lock_key#23 &dentry->d_lock &wq#2 &p->pi_lock &rq->__lock irq_context: 0 &sb->s_type->i_mutex_key#9 &sb->s_type->i_lock_key#23 &dentry->d_lock &wq#2 &p->pi_lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 &xt[i].mutex &mm->mmap_lock pgd_lock irq_context: 0 &xt[i].mutex &mm->mmap_lock stock_lock irq_context: 0 &xt[i].mutex &mm->mmap_lock key irq_context: 0 &xt[i].mutex &mm->mmap_lock pcpu_lock irq_context: 0 &xt[i].mutex &mm->mmap_lock percpu_counters_lock irq_context: 0 &xt[i].mutex &mm->mmap_lock pcpu_lock stock_lock irq_context: 0 sk_lock-AF_INET6 fill_pool_map-wait-type-override rcu_node_0 irq_context: 0 &sb->s_type->i_mutex_key#8 &sbi->s_writepages_rwsem rcu_read_lock fill_pool_map-wait-type-override &____s->seqcount#2 irq_context: 0 &sb->s_type->i_mutex_key#8 &sbi->s_writepages_rwsem rcu_read_lock fill_pool_map-wait-type-override &____s->seqcount irq_context: 0 &sb->s_type->i_mutex_key#10 sk_lock-AF_INET6/1 rcu_read_lock rcu_read_lock rcu_read_lock_bh &sch->q.lock irq_context: 0 &sb->s_type->i_mutex_key#10 sk_lock-AF_INET6/1 rcu_read_lock rcu_read_lock rcu_read_lock_bh &sch->q.lock pool_lock#2 irq_context: 0 &sb->s_type->i_mutex_key#10 sk_lock-AF_INET6/1 rcu_read_lock rcu_read_lock rcu_read_lock_bh &sch->q.lock tk_core.seq.seqcount irq_context: 0 &sb->s_type->i_mutex_key#10 sk_lock-AF_INET6/1 rcu_read_lock rcu_read_lock rcu_read_lock_bh &sch->q.lock hrtimer_bases.lock irq_context: 0 &sb->s_type->i_mutex_key#10 sk_lock-AF_INET6/1 rcu_read_lock rcu_read_lock rcu_read_lock_bh &sch->q.lock hrtimer_bases.lock &obj_hash[i].lock irq_context: 0 &sb->s_type->i_mutex_key#10 sk_lock-AF_INET6/1 rcu_read_lock rcu_read_lock rcu_read_lock_bh &sch->q.lock hrtimer_bases.lock tk_core.seq.seqcount irq_context: softirq rcu_read_lock rcu_read_lock slock-AF_INET6 &data->lock irq_context: softirq rcu_read_lock rcu_read_lock slock-AF_INET6 rcu_read_lock rcu_read_lock rcu_read_lock_bh &sch->q.lock irq_context: softirq rcu_read_lock rcu_read_lock slock-AF_INET6 rcu_read_lock rcu_read_lock rcu_read_lock_bh &sch->q.lock &c->lock irq_context: softirq rcu_read_lock rcu_read_lock slock-AF_INET6 rcu_read_lock rcu_read_lock rcu_read_lock_bh &sch->q.lock pool_lock#2 irq_context: softirq rcu_read_lock rcu_read_lock slock-AF_INET6 rcu_read_lock rcu_read_lock rcu_read_lock_bh &sch->q.lock tk_core.seq.seqcount irq_context: softirq rcu_read_lock rcu_read_lock slock-AF_INET6 &obj_hash[i].lock pool_lock irq_context: softirq rcu_read_lock rcu_read_lock slock-AF_INET6 fill_pool_map-wait-type-override pool_lock#2 irq_context: softirq rcu_read_lock rcu_read_lock slock-AF_INET6 fill_pool_map-wait-type-override pool_lock irq_context: softirq rcu_read_lock rcu_read_lock slock-AF_INET6 fill_pool_map-wait-type-override &c->lock irq_context: softirq rcu_read_lock rcu_read_lock slock-AF_INET6 fill_pool_map-wait-type-override &____s->seqcount#2 irq_context: softirq rcu_read_lock rcu_read_lock slock-AF_INET6 fill_pool_map-wait-type-override &____s->seqcount irq_context: softirq rcu_read_lock rcu_read_lock slock-AF_INET6 rcu_read_lock rcu_read_lock &pool->lock irq_context: softirq rcu_read_lock rcu_read_lock slock-AF_INET6 rcu_read_lock rcu_read_lock &pool->lock &obj_hash[i].lock irq_context: softirq rcu_read_lock rcu_read_lock slock-AF_INET6 rcu_read_lock rcu_read_lock &pool->lock &p->pi_lock irq_context: softirq rcu_read_lock rcu_read_lock slock-AF_INET6 crngs.lock irq_context: softirq rcu_read_lock rcu_read_lock slock-AF_INET6 rcu_read_lock &c->lock irq_context: softirq rcu_read_lock rcu_read_lock slock-AF_INET6 rcu_read_lock &____s->seqcount#2 irq_context: softirq rcu_read_lock rcu_read_lock slock-AF_INET6 rcu_read_lock &pcp->lock &zone->lock irq_context: softirq rcu_read_lock rcu_read_lock slock-AF_INET6 rcu_read_lock &dir->lock#2 irq_context: softirq rcu_read_lock rcu_read_lock slock-AF_INET6 rcu_read_lock &ul->lock irq_context: softirq rcu_read_lock rcu_read_lock slock-AF_INET6 rcu_read_lock &n->list_lock irq_context: softirq rcu_read_lock rcu_read_lock slock-AF_INET6 rcu_read_lock &n->list_lock &c->lock irq_context: softirq rcu_read_lock rcu_read_lock slock-AF_INET6 rcu_read_lock key#23 irq_context: softirq rcu_read_lock rcu_read_lock slock-AF_INET6 &____s->seqcount#2 irq_context: softirq rcu_read_lock rcu_read_lock slock-AF_INET6 rcu_read_lock rcu_read_lock rcu_read_lock_bh &sch->q.lock hrtimer_bases.lock irq_context: softirq rcu_read_lock rcu_read_lock slock-AF_INET6 rcu_read_lock rcu_read_lock rcu_read_lock_bh &sch->q.lock hrtimer_bases.lock &obj_hash[i].lock irq_context: softirq rcu_read_lock rcu_read_lock slock-AF_INET6 rcu_read_lock rcu_read_lock rcu_read_lock_bh &sch->q.lock hrtimer_bases.lock tk_core.seq.seqcount irq_context: softirq rcu_read_lock rcu_read_lock slock-AF_INET6 rcu_read_lock batched_entropy_u8.lock irq_context: softirq rcu_read_lock rcu_read_lock slock-AF_INET6 rcu_read_lock kfence_freelist_lock irq_context: softirq &(&hwstats->traffic_dw)->timer rcu_read_lock &pool->lock fill_pool_map-wait-type-override &c->lock irq_context: softirq &(&hwstats->traffic_dw)->timer rcu_read_lock &pool->lock fill_pool_map-wait-type-override &____s->seqcount#2 irq_context: softirq &(&hwstats->traffic_dw)->timer rcu_read_lock &pool->lock fill_pool_map-wait-type-override &____s->seqcount irq_context: softirq &(&hwstats->traffic_dw)->timer rcu_read_lock &pool->lock fill_pool_map-wait-type-override pool_lock irq_context: softirq rcu_read_lock rcu_read_lock slock-AF_INET6 krc.lock &obj_hash[i].lock irq_context: softirq rcu_read_lock rcu_read_lock slock-AF_INET6 krc.lock &base->lock irq_context: softirq rcu_read_lock rcu_read_lock slock-AF_INET6 krc.lock &base->lock &obj_hash[i].lock irq_context: softirq rcu_read_lock rcu_read_lock slock-AF_INET6 rcu_read_lock batched_entropy_u32.lock irq_context: softirq rcu_read_lock rcu_read_lock slock-AF_INET6 rcu_read_lock batched_entropy_u32.lock crngs.lock irq_context: softirq rcu_read_lock rcu_read_lock slock-AF_INET6 batched_entropy_u8.lock irq_context: softirq rcu_read_lock rcu_read_lock slock-AF_INET6 fill_pool_map-wait-type-override &pcp->lock &zone->lock irq_context: softirq rcu_read_lock rcu_read_lock slock-AF_INET6 fill_pool_map-wait-type-override &pcp->lock &zone->lock &____s->seqcount irq_context: softirq rcu_read_lock rcu_read_lock slock-AF_INET6 sctp_assocs_id_lock irq_context: softirq rcu_read_lock rcu_read_lock slock-AF_INET6 sctp_assocs_id_lock &c->lock irq_context: softirq rcu_read_lock rcu_read_lock slock-AF_INET6 sctp_assocs_id_lock pool_lock#2 irq_context: softirq rcu_read_lock rcu_read_lock slock-AF_INET6 &asoc->wait irq_context: softirq rcu_read_lock rcu_read_lock slock-AF_INET6 &asoc->wait &p->pi_lock irq_context: softirq rcu_read_lock rcu_read_lock slock-AF_INET6 &asoc->wait &p->pi_lock &rq->__lock irq_context: softirq rcu_read_lock rcu_read_lock slock-AF_INET6 &asoc->wait &p->pi_lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: softirq rcu_read_lock rcu_read_lock slock-AF_INET6 batched_entropy_u32.lock irq_context: softirq rcu_read_lock rcu_read_lock slock-AF_INET6 batched_entropy_u32.lock crngs.lock irq_context: softirq rcu_read_lock rcu_read_lock slock-AF_INET6 rcu_read_lock rcu_read_lock &pool->lock &p->pi_lock &rq->__lock irq_context: softirq rcu_read_lock rcu_read_lock slock-AF_INET6 rcu_read_lock rcu_read_lock &pool->lock &p->pi_lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)bat_events &cfs_rq->removed.lock irq_context: 0 sk_lock-AF_PPPOX clock-AF_PPPOX irq_context: 0 &pn->l2tp_tunnel_idr_lock &obj_hash[i].lock irq_context: 0 sk_lock-AF_INET6 rcu_read_lock rcu_read_lock rcu_read_lock_bh &data->lock irq_context: 0 &sb->s_type->i_mutex_key#10 sk_lock-AF_INET6/1 sctp_assocs_id_lock &obj_hash[i].lock pool_lock irq_context: 0 cb_lock &data->lock irq_context: softirq rcu_read_lock rcu_read_lock &dreq->dreq_lock slock-AF_INET6 &dccp_hashinfo.bhash[i].lock &dccp_hashinfo.bhash2[i].lock &n->list_lock &c->lock irq_context: 0 sk_lock-AF_INET6 rcu_read_lock rcu_read_lock &n->lock &data->lock irq_context: 0 sk_lock-AF_INET k-sk_lock-AF_INET rcu_read_lock rcu_read_lock rcu_read_lock_bh &sch->q.lock irq_context: 0 sk_lock-AF_INET k-sk_lock-AF_INET rcu_read_lock rcu_read_lock rcu_read_lock_bh &sch->q.lock &c->lock irq_context: 0 sk_lock-AF_INET k-sk_lock-AF_INET rcu_read_lock rcu_read_lock rcu_read_lock_bh &sch->q.lock &n->list_lock irq_context: 0 sk_lock-AF_INET k-sk_lock-AF_INET rcu_read_lock rcu_read_lock rcu_read_lock_bh &sch->q.lock &n->list_lock &c->lock irq_context: 0 sk_lock-AF_INET k-sk_lock-AF_INET rcu_read_lock rcu_read_lock rcu_read_lock_bh &sch->q.lock pool_lock#2 irq_context: 0 sk_lock-AF_INET k-sk_lock-AF_INET rcu_read_lock rcu_read_lock rcu_read_lock_bh &sch->q.lock tk_core.seq.seqcount irq_context: 0 sk_lock-AF_INET k-sk_lock-AF_INET rcu_read_lock rcu_read_lock rcu_read_lock_bh &sch->q.lock hrtimer_bases.lock irq_context: 0 sk_lock-AF_INET k-sk_lock-AF_INET rcu_read_lock rcu_read_lock rcu_read_lock_bh &sch->q.lock hrtimer_bases.lock &obj_hash[i].lock irq_context: 0 sk_lock-AF_INET k-sk_lock-AF_INET rcu_read_lock rcu_read_lock rcu_read_lock_bh &sch->q.lock hrtimer_bases.lock tk_core.seq.seqcount irq_context: softirq rcu_read_lock rcu_read_lock rcu_read_lock rcu_read_lock rcu_read_lock_bh &sch->q.lock irq_context: softirq rcu_read_lock rcu_read_lock rcu_read_lock rcu_read_lock rcu_read_lock_bh &sch->q.lock pool_lock#2 irq_context: softirq rcu_read_lock rcu_read_lock rcu_read_lock rcu_read_lock rcu_read_lock_bh &sch->q.lock tk_core.seq.seqcount irq_context: softirq rcu_read_lock rcu_read_lock rcu_read_lock rcu_read_lock rcu_read_lock_bh &sch->q.lock hrtimer_bases.lock irq_context: softirq rcu_read_lock rcu_read_lock rcu_read_lock rcu_read_lock rcu_read_lock_bh &sch->q.lock hrtimer_bases.lock &obj_hash[i].lock irq_context: softirq rcu_read_lock rcu_read_lock rcu_read_lock rcu_read_lock rcu_read_lock_bh &sch->q.lock hrtimer_bases.lock tk_core.seq.seqcount irq_context: 0 sk_lock-AF_INET k-sk_lock-AF_INET &data->lock irq_context: 0 rtnl_mutex &br->multicast_lock pool_lock#2 irq_context: 0 rtnl_mutex &br->multicast_lock &dir->lock#2 irq_context: 0 rtnl_mutex &br->multicast_lock deferred_lock irq_context: 0 rtnl_mutex &br->multicast_lock rcu_read_lock &pool->lock &p->pi_lock &rq->__lock irq_context: 0 rtnl_mutex &br->multicast_lock rcu_read_lock &pool->lock &p->pi_lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: softirq rcu_read_lock rcu_read_lock rcu_read_lock rcu_read_lock rcu_read_lock_bh &sch->q.lock &c->lock irq_context: softirq rcu_read_lock rcu_read_lock k-slock-AF_INET/1 pool_lock#2 irq_context: softirq rcu_read_lock rcu_read_lock k-slock-AF_INET/1 &hashinfo->ehash_locks[i] irq_context: softirq rcu_read_lock rcu_read_lock k-slock-AF_INET/1 &tcp_hashinfo.bhash[i].lock irq_context: softirq rcu_read_lock rcu_read_lock k-slock-AF_INET/1 &tcp_hashinfo.bhash[i].lock stock_lock irq_context: softirq rcu_read_lock rcu_read_lock k-slock-AF_INET/1 &tcp_hashinfo.bhash[i].lock &obj_hash[i].lock irq_context: softirq rcu_read_lock rcu_read_lock k-slock-AF_INET/1 &tcp_hashinfo.bhash[i].lock pool_lock#2 irq_context: softirq rcu_read_lock rcu_read_lock k-slock-AF_INET/1 &tcp_hashinfo.bhash[i].lock &tcp_hashinfo.bhash2[i].lock irq_context: softirq rcu_read_lock rcu_read_lock k-slock-AF_INET/1 &tcp_hashinfo.bhash[i].lock &tcp_hashinfo.bhash2[i].lock stock_lock irq_context: softirq rcu_read_lock rcu_read_lock k-slock-AF_INET/1 &tcp_hashinfo.bhash[i].lock &tcp_hashinfo.bhash2[i].lock &obj_hash[i].lock irq_context: softirq rcu_read_lock rcu_read_lock k-slock-AF_INET/1 &tcp_hashinfo.bhash[i].lock &tcp_hashinfo.bhash2[i].lock pool_lock#2 irq_context: softirq rcu_read_lock rcu_read_lock k-slock-AF_INET/1 &base->lock irq_context: softirq rcu_read_lock rcu_read_lock k-slock-AF_INET/1 &base->lock &obj_hash[i].lock irq_context: softirq rcu_read_lock rcu_read_lock k-slock-AF_INET/1 rcu_read_lock &ei->socket.wq.wait irq_context: softirq rcu_read_lock rcu_read_lock k-slock-AF_INET/1 rcu_read_lock &ei->socket.wq.wait &p->pi_lock irq_context: softirq rcu_read_lock rcu_read_lock k-slock-AF_INET/1 rcu_read_lock &ei->socket.wq.wait &p->pi_lock &rq->__lock irq_context: softirq rcu_read_lock rcu_read_lock k-slock-AF_INET/1 rcu_read_lock &ei->socket.wq.wait &p->pi_lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: softirq rcu_read_lock rcu_read_lock k-slock-AF_INET/1 rcu_read_lock &pool->lock irq_context: softirq rcu_read_lock rcu_read_lock k-slock-AF_INET/1 rcu_read_lock &pool->lock &obj_hash[i].lock irq_context: softirq rcu_read_lock rcu_read_lock k-slock-AF_INET/1 rcu_read_lock &pool->lock pool_lock#2 irq_context: softirq rcu_read_lock rcu_read_lock k-slock-AF_INET/1 slock-AF_INET rcu_read_lock &ei->socket.wq.wait irq_context: softirq rcu_read_lock rcu_read_lock k-slock-AF_INET/1 slock-AF_INET rcu_read_lock &ei->socket.wq.wait &p->pi_lock irq_context: 0 rtnl_mutex (switchdev_blocking_notif_chain).rwsem &rq->__lock irq_context: 0 rtnl_mutex (switchdev_blocking_notif_chain).rwsem &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 &pipe->mutex/1 sk_lock-AF_INET6 &____s->seqcount#8 irq_context: 0 &pipe->mutex/1 sk_lock-AF_INET6 batched_entropy_u32.lock irq_context: 0 &pipe->mutex/1 sk_lock-AF_INET6 &table->hash[i].lock irq_context: 0 &pipe->mutex/1 sk_lock-AF_INET6 &table->hash[i].lock clock-AF_INET6 irq_context: 0 &pipe->mutex/1 sk_lock-AF_INET6 &table->hash[i].lock &table->hash2[i].lock irq_context: 0 pernet_ops_rwsem nf_ct_proto_mutex nf_hook_mutex rcu_node_0 irq_context: 0 sb_writers#4 &type->i_mutex_dir_key#3/1 &sb->s_type->i_mutex_key#8 &rcu_state.expedited_wq irq_context: 0 &sb->s_type->i_mutex_key#10 &chan->lock/1 &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 sk_lock-AF_BLUETOOTH-BTPROTO_HCI &data->lock &obj_hash[i].lock irq_context: 0 sk_lock-AF_BLUETOOTH-BTPROTO_HCI &data->lock &base->lock irq_context: 0 sk_lock-AF_BLUETOOTH-BTPROTO_HCI &data->lock &base->lock &obj_hash[i].lock irq_context: 0 (work_completion)(&hdev->power_on) irq_context: 0 &hdev->req_lock (work_completion)(&(&hdev->interleave_scan)->work) irq_context: 0 &hdev->req_lock hci_dev_list_lock irq_context: 0 &hdev->req_lock (work_completion)(&hdev->tx_work) irq_context: 0 &hdev->req_lock (work_completion)(&hdev->rx_work) irq_context: 0 &hdev->req_lock (work_completion)(&(&hdev->rpa_expired)->work) irq_context: 0 &hdev->req_lock (work_completion)(&(&hdev->rpa_expired)->work) &rq->__lock irq_context: 0 &hdev->req_lock (work_completion)(&(&hdev->rpa_expired)->work) &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 &hdev->req_lock &wq->mutex irq_context: 0 &hdev->req_lock (wq_completion)hci0#2 irq_context: 0 &hdev->req_lock &wq->mutex &pool->lock/1 irq_context: 0 &hdev->req_lock &wq->mutex &x->wait#10 irq_context: 0 &hdev->req_lock &hdev->lock irq_context: 0 &hdev->req_lock &hdev->lock fs_reclaim irq_context: 0 &hdev->req_lock &hdev->lock fs_reclaim mmu_notifier_invalidate_range_start irq_context: 0 &hdev->req_lock &hdev->lock &c->lock irq_context: 0 &hdev->req_lock &hdev->lock pool_lock#2 irq_context: 0 &hdev->req_lock &hdev->lock tk_core.seq.seqcount irq_context: 0 &hdev->req_lock &hdev->lock hci_sk_list.lock irq_context: 0 &hdev->req_lock &hdev->lock &data->lock irq_context: 0 &hdev->req_lock &hdev->lock &obj_hash[i].lock irq_context: 0 &hdev->req_lock &hdev->lock hci_cb_list_lock irq_context: 0 &hdev->req_lock &hdev->lock hci_cb_list_lock &list->lock#10 irq_context: 0 &hdev->req_lock &hdev->lock hci_cb_list_lock &obj_hash[i].lock irq_context: 0 &hdev->req_lock &hdev->lock hci_cb_list_lock (work_completion)(&(&conn->id_addr_timer)->work) irq_context: 0 &hdev->req_lock &hdev->lock hci_cb_list_lock &conn->chan_lock irq_context: 0 &hdev->req_lock &hdev->lock hci_cb_list_lock rcu_state.exp_mutex rcu_node_0 irq_context: 0 &hdev->req_lock &hdev->lock hci_cb_list_lock rcu_state.exp_mutex &obj_hash[i].lock irq_context: 0 &hdev->req_lock &hdev->lock hci_cb_list_lock rcu_state.exp_mutex rcu_read_lock &pool->lock irq_context: 0 &hdev->req_lock &hdev->lock hci_cb_list_lock rcu_state.exp_mutex rcu_read_lock &pool->lock &obj_hash[i].lock irq_context: 0 &hdev->req_lock &hdev->lock hci_cb_list_lock rcu_state.exp_mutex rcu_read_lock &pool->lock &p->pi_lock irq_context: 0 &hdev->req_lock &hdev->lock hci_cb_list_lock rcu_state.exp_mutex rcu_read_lock &pool->lock &p->pi_lock &rq->__lock irq_context: 0 &hdev->req_lock &hdev->lock hci_cb_list_lock rcu_state.exp_mutex rcu_read_lock &pool->lock &p->pi_lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 &hdev->req_lock &hdev->lock hci_cb_list_lock rcu_state.exp_mutex &rnp->exp_wq[1] irq_context: 0 &hdev->req_lock &hdev->lock hci_cb_list_lock rcu_state.exp_mutex &rq->__lock irq_context: 0 &hdev->req_lock &hdev->lock hci_cb_list_lock rcu_state.exp_mutex &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 &hdev->req_lock &hdev->lock hci_cb_list_lock &list->lock#9 irq_context: 0 &hdev->req_lock &hdev->lock hci_cb_list_lock pool_lock#2 irq_context: 0 &hdev->req_lock &hdev->lock (work_completion)(&(&conn->disc_work)->work) irq_context: 0 &hdev->req_lock &hdev->lock (work_completion)(&(&conn->auto_accept_work)->work) irq_context: 0 &hdev->req_lock &hdev->lock (work_completion)(&(&conn->idle_work)->work) irq_context: 0 &hdev->req_lock &hdev->lock &list->lock#9 irq_context: 0 &hdev->req_lock &hdev->lock rcu_state.exp_mutex rcu_node_0 irq_context: 0 &hdev->req_lock &hdev->lock rcu_state.exp_mutex &obj_hash[i].lock irq_context: 0 &hdev->req_lock &hdev->lock rcu_state.exp_mutex rcu_read_lock &pool->lock irq_context: 0 &hdev->req_lock &hdev->lock rcu_state.exp_mutex rcu_read_lock &pool->lock &obj_hash[i].lock irq_context: 0 &hdev->req_lock &hdev->lock rcu_state.exp_mutex rcu_read_lock &pool->lock &p->pi_lock irq_context: 0 &hdev->req_lock &hdev->lock rcu_state.exp_mutex rcu_read_lock &pool->lock &p->pi_lock &rq->__lock irq_context: 0 &hdev->req_lock &hdev->lock rcu_state.exp_mutex rcu_read_lock &pool->lock &p->pi_lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 &hdev->req_lock &hdev->lock rcu_state.exp_mutex &rq->__lock irq_context: 0 &hdev->req_lock &hdev->lock rcu_state.exp_mutex &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 &hdev->req_lock &hdev->lock &k->k_lock irq_context: 0 &hdev->req_lock &hdev->lock &root->kernfs_rwsem irq_context: 0 &hdev->req_lock &hdev->lock &root->kernfs_rwsem irq_context: 0 &hdev->req_lock &hdev->lock dev_pm_qos_sysfs_mtx irq_context: 0 &hdev->req_lock &hdev->lock dev_pm_qos_sysfs_mtx &root->kernfs_rwsem irq_context: 0 &hdev->req_lock &hdev->lock dev_pm_qos_sysfs_mtx &root->kernfs_rwsem irq_context: 0 &hdev->req_lock &hdev->lock dev_pm_qos_sysfs_mtx dev_pm_qos_mtx irq_context: 0 &hdev->req_lock &hdev->lock &root->kernfs_rwsem &root->kernfs_iattr_rwsem irq_context: 0 &hdev->req_lock &hdev->lock &root->kernfs_rwsem &root->kernfs_iattr_rwsem &rq->__lock irq_context: 0 &hdev->req_lock &hdev->lock &root->kernfs_rwsem &root->kernfs_iattr_rwsem &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 &hdev->req_lock &hdev->lock &root->kernfs_rwsem kernfs_idr_lock irq_context: 0 &hdev->req_lock &hdev->lock &root->kernfs_rwsem &obj_hash[i].lock irq_context: 0 &hdev->req_lock &hdev->lock &root->kernfs_rwsem pool_lock#2 irq_context: 0 &hdev->req_lock &hdev->lock kernfs_idr_lock irq_context: 0 &hdev->req_lock &hdev->lock &k->k_lock klist_remove_lock irq_context: 0 &hdev->req_lock &hdev->lock &rq->__lock irq_context: 0 &hdev->req_lock &hdev->lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 &hdev->req_lock &hdev->lock &k->list_lock irq_context: 0 &hdev->req_lock &hdev->lock sysfs_symlink_target_lock irq_context: 0 &hdev->req_lock &hdev->lock subsys mutex#81 irq_context: 0 &hdev->req_lock &hdev->lock subsys mutex#81 &k->k_lock irq_context: 0 &hdev->req_lock &hdev->lock subsys mutex#81 &k->k_lock klist_remove_lock irq_context: 0 &hdev->req_lock &hdev->lock &x->wait#9 irq_context: 0 &hdev->req_lock &hdev->lock dpm_list_mtx irq_context: 0 &hdev->req_lock &hdev->lock &dev->power.lock irq_context: 0 &hdev->req_lock &hdev->lock deferred_probe_mutex irq_context: 0 &hdev->req_lock &hdev->lock device_links_lock irq_context: 0 &hdev->req_lock &hdev->lock mmu_notifier_invalidate_range_start irq_context: 0 &hdev->req_lock &hdev->lock uevent_sock_mutex irq_context: 0 &hdev->req_lock &hdev->lock uevent_sock_mutex mmu_notifier_invalidate_range_start irq_context: 0 &hdev->req_lock &hdev->lock uevent_sock_mutex pool_lock#2 irq_context: 0 &hdev->req_lock &hdev->lock uevent_sock_mutex &c->lock irq_context: 0 &hdev->req_lock &hdev->lock uevent_sock_mutex &n->list_lock irq_context: 0 &hdev->req_lock &hdev->lock uevent_sock_mutex &n->list_lock &c->lock irq_context: 0 &hdev->req_lock &hdev->lock uevent_sock_mutex &rq->__lock irq_context: 0 &hdev->req_lock &hdev->lock uevent_sock_mutex &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 &hdev->req_lock &hdev->lock uevent_sock_mutex nl_table_lock irq_context: 0 &hdev->req_lock &hdev->lock uevent_sock_mutex rlock-AF_NETLINK irq_context: 0 &hdev->req_lock &hdev->lock uevent_sock_mutex rcu_read_lock &ei->socket.wq.wait irq_context: 0 &hdev->req_lock &hdev->lock uevent_sock_mutex rcu_read_lock &ei->socket.wq.wait &ep->lock irq_context: 0 &hdev->req_lock &hdev->lock uevent_sock_mutex rcu_read_lock &ei->socket.wq.wait &ep->lock &ep->wq irq_context: 0 &hdev->req_lock &hdev->lock uevent_sock_mutex rcu_read_lock &ei->socket.wq.wait &ep->lock &ep->wq &p->pi_lock irq_context: 0 &hdev->req_lock &hdev->lock uevent_sock_mutex rcu_read_lock &ei->socket.wq.wait &ep->lock &ep->wq &p->pi_lock &rq->__lock irq_context: 0 &hdev->req_lock &hdev->lock uevent_sock_mutex rcu_read_lock &ei->socket.wq.wait &ep->lock &ep->wq &p->pi_lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 &hdev->req_lock &hdev->lock uevent_sock_mutex rcu_read_lock rcu_node_0 irq_context: 0 &hdev->req_lock &hdev->lock uevent_sock_mutex rcu_read_lock &rq->__lock irq_context: 0 &hdev->req_lock &hdev->lock uevent_sock_mutex rcu_read_lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 &hdev->req_lock &hdev->lock uevent_sock_mutex nl_table_wait.lock irq_context: 0 &hdev->req_lock &hdev->lock uevent_sock_mutex &obj_hash[i].lock irq_context: 0 &hdev->req_lock tk_core.seq.seqcount irq_context: 0 &hdev->req_lock hci_sk_list.lock irq_context: 0 &hdev->req_lock &data->lock irq_context: 0 &hdev->req_lock &cfs_rq->removed.lock irq_context: 0 &hdev->req_lock &msft->filter_lock irq_context: 0 &hdev->req_lock &list->lock#6 irq_context: 0 &hdev->req_lock (work_completion)(&hdev->cmd_work) irq_context: 0 &hdev->req_lock (work_completion)(&(&hdev->cmd_timer)->work) irq_context: 0 sb_writers#4 &type->i_mutex_dir_key#3/1 &sb->s_type->i_mutex_key#8 jbd2_handle &sbi->s_orphan_lock pgd_lock irq_context: 0 sb_writers#4 &type->i_mutex_dir_key#3/1 &sb->s_type->i_mutex_key#8 jbd2_handle &sbi->s_orphan_lock stock_lock irq_context: 0 sb_writers#4 &type->i_mutex_dir_key#3/1 &sb->s_type->i_mutex_key#8 jbd2_handle &sbi->s_orphan_lock rcu_read_lock pool_lock#2 irq_context: 0 sb_writers#4 &type->i_mutex_dir_key#3/1 &sb->s_type->i_mutex_key#8 jbd2_handle &sbi->s_orphan_lock key irq_context: 0 sb_writers#4 &type->i_mutex_dir_key#3/1 &sb->s_type->i_mutex_key#8 jbd2_handle &sbi->s_orphan_lock pcpu_lock irq_context: 0 sb_writers#4 &type->i_mutex_dir_key#3/1 &sb->s_type->i_mutex_key#8 jbd2_handle &sbi->s_orphan_lock percpu_counters_lock irq_context: 0 sb_writers#4 &type->i_mutex_dir_key#3/1 &sb->s_type->i_mutex_key#8 jbd2_handle &sbi->s_orphan_lock pcpu_lock stock_lock irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem &root->kernfs_rwsem &obj_hash[i].lock pool_lock irq_context: 0 pernet_ops_rwsem ovs_mutex nf_ct_proto_mutex nf_hook_mutex remove_cache_srcu irq_context: 0 pernet_ops_rwsem ovs_mutex nf_ct_proto_mutex nf_hook_mutex remove_cache_srcu quarantine_lock irq_context: 0 pernet_ops_rwsem ovs_mutex nf_ct_proto_mutex nf_hook_mutex remove_cache_srcu &c->lock irq_context: 0 pernet_ops_rwsem ovs_mutex nf_ct_proto_mutex nf_hook_mutex remove_cache_srcu &n->list_lock irq_context: 0 pernet_ops_rwsem ovs_mutex nf_ct_proto_mutex nf_hook_mutex remove_cache_srcu &obj_hash[i].lock irq_context: 0 pernet_ops_rwsem ovs_mutex nf_ct_proto_mutex nf_hook_mutex remove_cache_srcu pool_lock#2 irq_context: 0 (wq_completion)wg-kex-wg1#4 (work_completion)(&({ do { const void *__vpp_verify = (typeof((worker) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __asm__ ("" : "=r"(__ptr) : "0"((typeof(*((worker))) *)((worker)))); (typeof((typeof(*((worker))) *)((worker)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); })->work) &handshake->lock remove_cache_srcu irq_context: 0 (wq_completion)wg-kex-wg1#4 (work_completion)(&({ do { const void *__vpp_verify = (typeof((worker) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __asm__ ("" : "=r"(__ptr) : "0"((typeof(*((worker))) *)((worker)))); (typeof((typeof(*((worker))) *)((worker)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); })->work) &handshake->lock remove_cache_srcu quarantine_lock irq_context: 0 (wq_completion)wg-kex-wg1#4 (work_completion)(&({ do { const void *__vpp_verify = (typeof((worker) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __asm__ ("" : "=r"(__ptr) : "0"((typeof(*((worker))) *)((worker)))); (typeof((typeof(*((worker))) *)((worker)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); })->work) &handshake->lock remove_cache_srcu &c->lock irq_context: 0 (wq_completion)wg-kex-wg1#4 (work_completion)(&({ do { const void *__vpp_verify = (typeof((worker) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __asm__ ("" : "=r"(__ptr) : "0"((typeof(*((worker))) *)((worker)))); (typeof((typeof(*((worker))) *)((worker)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); })->work) &handshake->lock remove_cache_srcu &n->list_lock irq_context: 0 (wq_completion)wg-kex-wg1#4 (work_completion)(&({ do { const void *__vpp_verify = (typeof((worker) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __asm__ ("" : "=r"(__ptr) : "0"((typeof(*((worker))) *)((worker)))); (typeof((typeof(*((worker))) *)((worker)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); })->work) &handshake->lock remove_cache_srcu &obj_hash[i].lock irq_context: 0 (wq_completion)krxrpcd (work_completion)(&rxnet->peer_keepalive_work) &rq->__lock irq_context: 0 &ep->mtx fs_reclaim mmu_notifier_invalidate_range_start &rq->__lock irq_context: 0 &sb->s_type->i_mutex_key#10 &u->peer_wait &p->pi_lock &cfs_rq->removed.lock irq_context: 0 pernet_ops_rwsem &sn->gssp_lock &rq->__lock irq_context: 0 pernet_ops_rwsem &sn->gssp_lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 uevent_sock_mutex &____s->seqcount#2 irq_context: 0 uevent_sock_mutex &____s->seqcount irq_context: 0 &ep->mtx uevent_sock_mutex &____s->seqcount#2 irq_context: 0 &ep->mtx uevent_sock_mutex &____s->seqcount irq_context: 0 &ep->mtx quarantine_lock irq_context: 0 rtnl_mutex &tb->tb6_lock &base->lock irq_context: 0 rtnl_mutex &tb->tb6_lock &base->lock &obj_hash[i].lock irq_context: 0 sb_writers#4 tomoyo_ss remove_cache_srcu &cfs_rq->removed.lock irq_context: 0 pernet_ops_rwsem nf_ct_proto_mutex nf_hook_mutex fs_reclaim mmu_notifier_invalidate_range_start &rq->__lock irq_context: 0 pernet_ops_rwsem nf_ct_proto_mutex nf_hook_mutex fs_reclaim mmu_notifier_invalidate_range_start &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)phy12 (work_completion)(&(&local->scan_work)->work) &local->mtx &____s->seqcount#2 irq_context: 0 (wq_completion)phy12 (work_completion)(&(&local->scan_work)->work) &local->mtx &____s->seqcount irq_context: 0 (wq_completion)wg-crypt-wg0#4 (work_completion)(&peer->transmit_packet_work) &rcu_state.expedited_wq irq_context: 0 (wq_completion)wg-crypt-wg0#4 (work_completion)(&peer->transmit_packet_work) &rcu_state.expedited_wq &p->pi_lock irq_context: 0 (wq_completion)wg-crypt-wg0#4 (work_completion)(&peer->transmit_packet_work) &rcu_state.expedited_wq &p->pi_lock &rq->__lock irq_context: 0 (wq_completion)wg-crypt-wg0#4 (work_completion)(&peer->transmit_packet_work) &rcu_state.expedited_wq &p->pi_lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: softirq &(&local->scan_work)->timer rcu_read_lock &pool->lock/1 &p->pi_lock &rq->__lock &cfs_rq->removed.lock irq_context: 0 (wq_completion)mld (work_completion)(&(&idev->mc_ifc_work)->work) &idev->mc_lock rcu_read_lock rcu_read_lock rcu_read_lock_bh rcu_read_lock &data->lock &obj_hash[i].lock irq_context: 0 (wq_completion)mld (work_completion)(&(&idev->mc_ifc_work)->work) &idev->mc_lock rcu_read_lock rcu_read_lock rcu_read_lock_bh rcu_read_lock &data->lock &base->lock irq_context: 0 (wq_completion)mld (work_completion)(&(&idev->mc_ifc_work)->work) &idev->mc_lock rcu_read_lock rcu_read_lock rcu_read_lock_bh rcu_read_lock &data->lock &base->lock &obj_hash[i].lock irq_context: 0 sb_writers#4 &type->i_mutex_dir_key#3/1 remove_cache_srcu &pcp->lock &zone->lock irq_context: 0 sb_writers#4 &type->i_mutex_dir_key#3/1 remove_cache_srcu &pcp->lock &zone->lock &____s->seqcount irq_context: 0 (wq_completion)wg-crypt-wg1#5 (work_completion)(&peer->transmit_packet_work) &peer->endpoint_lock rcu_read_lock_bh batched_entropy_u32.lock crngs.lock base_crng.lock irq_context: 0 (wq_completion)wg-kex-wg2#7 (work_completion)(&peer->transmit_handshake_work) batched_entropy_u8.lock crngs.lock irq_context: 0 &net->xfrm.xfrm_cfg_mutex &net->xfrm.xfrm_policy_lock irq_context: 0 &net->xfrm.xfrm_cfg_mutex &net->xfrm.xfrm_policy_lock pool_lock#2 irq_context: 0 &net->xfrm.xfrm_cfg_mutex &net->xfrm.xfrm_policy_lock rcu_read_lock rhashtable_bucket irq_context: 0 &net->xfrm.xfrm_cfg_mutex &net->xfrm.xfrm_policy_lock rcu_read_lock &tb->tb6_lock irq_context: 0 &net->xfrm.xfrm_cfg_mutex &net->xfrm.xfrm_policy_lock rcu_read_lock &tb->tb6_lock &net->ipv6.fib6_walker_lock irq_context: 0 &net->xfrm.xfrm_cfg_mutex &net->xfrm.xfrm_policy_lock &obj_hash[i].lock irq_context: 0 &net->xfrm.xfrm_cfg_mutex &net->xfrm.xfrm_policy_lock &base->lock irq_context: 0 &net->xfrm.xfrm_cfg_mutex &net->xfrm.xfrm_policy_lock &base->lock &obj_hash[i].lock irq_context: 0 &net->xfrm.xfrm_cfg_mutex rcu_read_lock &____s->seqcount#2 irq_context: 0 &net->xfrm.xfrm_cfg_mutex rcu_read_lock &____s->seqcount irq_context: 0 &net->xfrm.xfrm_cfg_mutex rcu_read_lock rcu_read_lock pool_lock#2 irq_context: 0 &net->xfrm.xfrm_cfg_mutex rcu_read_lock rcu_read_lock rlock-AF_KEY irq_context: 0 sb_writers#4 &type->i_mutex_dir_key#3/1 sb_internal jbd2_handle &p->pi_lock &rq->__lock &cfs_rq->removed.lock irq_context: 0 &net->xfrm.xfrm_cfg_mutex &net->xfrm.xfrm_policy_lock &____s->seqcount#16 irq_context: 0 &net->xfrm.xfrm_cfg_mutex &policy->lock irq_context: 0 &net->xfrm.xfrm_cfg_mutex &list->lock#32 irq_context: 0 &net->xfrm.xfrm_cfg_mutex rcu_read_lock rcu_read_lock rcu_node_0 irq_context: 0 &net->xfrm.xfrm_cfg_mutex rcu_read_lock rcu_read_lock &rq->__lock irq_context: 0 &net->xfrm.xfrm_cfg_mutex rcu_read_lock rcu_read_lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 &net->xfrm.xfrm_cfg_mutex (console_sem).lock irq_context: 0 &net->xfrm.xfrm_cfg_mutex console_lock console_srcu console_owner_lock irq_context: 0 &net->xfrm.xfrm_cfg_mutex console_lock console_srcu console_owner irq_context: 0 &net->xfrm.xfrm_cfg_mutex console_lock console_srcu console_owner &port_lock_key irq_context: 0 &net->xfrm.xfrm_cfg_mutex console_lock console_srcu console_owner console_owner_lock irq_context: 0 &net->xfrm.xfrm_cfg_mutex &data->lock &obj_hash[i].lock irq_context: 0 &net->xfrm.xfrm_cfg_mutex &data->lock &base->lock irq_context: 0 &net->xfrm.xfrm_cfg_mutex &data->lock &base->lock &obj_hash[i].lock irq_context: 0 sk_lock-AF_INET6 k-sk_lock-AF_INET6/1 &obj_hash[i].lock irq_context: 0 (wq_completion)wg-kex-wg0#7 (work_completion)(&peer->transmit_handshake_work) &peer->endpoint_lock rcu_read_lock_bh rcu_read_lock &____s->seqcount#12 irq_context: 0 (wq_completion)wg-kex-wg0#7 (work_completion)(&peer->transmit_handshake_work) &peer->endpoint_lock rcu_read_lock_bh rcu_read_lock &____s->seqcount#16 irq_context: 0 (wq_completion)wg-kex-wg0#7 (work_completion)(&peer->transmit_handshake_work) &peer->endpoint_lock rcu_read_lock_bh &data->lock irq_context: softirq (&policy->timer) irq_context: softirq (&policy->timer) &policy->lock irq_context: 0 sb_writers#4 &type->i_mutex_dir_key#3/1 pgd_lock irq_context: 0 sb_writers#4 &type->i_mutex_dir_key#3/1 key irq_context: 0 sb_writers#4 &type->i_mutex_dir_key#3/1 pcpu_lock irq_context: 0 sb_writers#4 &type->i_mutex_dir_key#3/1 percpu_counters_lock irq_context: 0 sb_writers#4 &type->i_mutex_dir_key#3/1 pcpu_lock stock_lock irq_context: 0 (wq_completion)wg-crypt-wg0#4 (work_completion)(&peer->transmit_packet_work) &peer->endpoint_lock rcu_read_lock_bh rcu_read_lock &____s->seqcount#12 irq_context: 0 (wq_completion)wg-crypt-wg0#4 (work_completion)(&peer->transmit_packet_work) &peer->endpoint_lock rcu_read_lock_bh rcu_read_lock &____s->seqcount#16 irq_context: 0 (wq_completion)wg-crypt-wg0#4 (work_completion)(&peer->transmit_packet_work) &peer->endpoint_lock rcu_read_lock_bh &data->lock irq_context: 0 &sb->s_type->i_mutex_key#10 sk_lock-AF_INET6 slock-AF_INET6 &obj_hash[i].lock pool_lock irq_context: 0 dup_mmap_sem &mm->mmap_lock &mm->mmap_lock/1 rcu_read_lock pgd_lock irq_context: 0 dup_mmap_sem &mm->mmap_lock &mm->mmap_lock/1 rcu_read_lock stock_lock irq_context: 0 dup_mmap_sem &mm->mmap_lock &mm->mmap_lock/1 rcu_read_lock key irq_context: 0 dup_mmap_sem &mm->mmap_lock &mm->mmap_lock/1 rcu_read_lock pcpu_lock irq_context: 0 dup_mmap_sem &mm->mmap_lock &mm->mmap_lock/1 rcu_read_lock percpu_counters_lock irq_context: 0 dup_mmap_sem &mm->mmap_lock &mm->mmap_lock/1 rcu_read_lock pcpu_lock stock_lock irq_context: 0 &sb->s_type->i_mutex_key#10 sk_lock-AF_INET6 &data->lock &obj_hash[i].lock irq_context: 0 &sb->s_type->i_mutex_key#10 sk_lock-AF_INET6 &data->lock &base->lock irq_context: 0 &sb->s_type->i_mutex_key#10 sk_lock-AF_INET6 &data->lock &base->lock &obj_hash[i].lock irq_context: 0 rcu_read_lock &____s->seqcount#12 irq_context: 0 rcu_read_lock &____s->seqcount#16 irq_context: 0 rcu_read_lock rcu_read_lock_bh &data->lock irq_context: 0 rcu_read_lock rcu_read_lock rcu_read_lock rcu_read_lock_bh &data->lock irq_context: 0 &f->f_pos_lock sb_writers#11 rcu_read_lock rcu_node_0 irq_context: 0 &f->f_pos_lock sb_writers#11 rcu_read_lock &rq->__lock irq_context: 0 &f->f_pos_lock sb_writers#11 rcu_read_lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 &mm->mmap_lock mmu_notifier_invalidate_range_start &cfs_rq->removed.lock irq_context: 0 &mm->mmap_lock mmu_notifier_invalidate_range_start &obj_hash[i].lock irq_context: 0 isotp_notifier_lock irq_context: 0 &so->wait irq_context: 0 &sb->s_type->i_mutex_key#10 isotp_notifier_lock irq_context: 0 sk_lock-AF_INET rcu_read_lock rcu_read_lock &n->lock &data->lock irq_context: 0 sk_lock-AF_INET rcu_read_lock rcu_read_lock &n->lock &obj_hash[i].lock irq_context: 0 sk_lock-AF_INET rcu_read_lock rcu_read_lock &n->lock pool_lock#2 irq_context: 0 sk_lock-AF_INET6 k-sk_lock-AF_INET6/1 &rq->__lock irq_context: 0 sk_lock-AF_INET6 k-sk_lock-AF_INET6/1 &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 sk_lock-AF_INET clock-AF_INET &obj_hash[i].lock irq_context: 0 sk_lock-AF_INET &mux->lock irq_context: 0 sk_lock-AF_INET rcu_read_lock &pool->lock/1 irq_context: 0 sk_lock-AF_INET rcu_read_lock &pool->lock/1 &obj_hash[i].lock irq_context: 0 sk_lock-AF_INET rcu_read_lock &pool->lock/1 pool_lock#2 irq_context: 0 sk_lock-AF_INET rcu_read_lock &pool->lock/1 &p->pi_lock irq_context: 0 sk_lock-AF_INET rcu_read_lock &pool->lock/1 &p->pi_lock &rq->__lock irq_context: 0 sk_lock-AF_INET rcu_read_lock &pool->lock/1 &p->pi_lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)kstrp irq_context: 0 (wq_completion)kstrp (work_completion)(&strp->work)#2 irq_context: 0 (wq_completion)kstrp (work_completion)(&strp->work)#2 sk_lock-AF_INET irq_context: 0 (wq_completion)kstrp (work_completion)(&strp->work)#2 sk_lock-AF_INET slock-AF_INET irq_context: 0 (wq_completion)kstrp (work_completion)(&strp->work)#2 sk_lock-AF_INET slock-AF_INET &sk->sk_lock.wq irq_context: 0 (wq_completion)kstrp (work_completion)(&strp->work)#2 sk_lock-AF_INET &rq->__lock irq_context: 0 (wq_completion)kstrp (work_completion)(&strp->work)#2 sk_lock-AF_INET &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)kstrp (work_completion)(&strp->work)#2 sk_lock-AF_INET tk_core.seq.seqcount irq_context: 0 (wq_completion)kstrp (work_completion)(&strp->work)#2 slock-AF_INET irq_context: softirq (&hc->tx_rtotimer) &zone->lock irq_context: softirq (&hc->tx_rtotimer) &zone->lock &____s->seqcount irq_context: 0 sk_lock-AF_KCM rcu_read_lock &rcu_state.gp_wq irq_context: 0 sk_lock-AF_KCM rcu_read_lock &rcu_state.gp_wq &p->pi_lock irq_context: 0 sk_lock-AF_KCM rcu_read_lock &rcu_state.gp_wq &p->pi_lock &rq->__lock irq_context: 0 sk_lock-AF_KCM rcu_read_lock &rcu_state.gp_wq &p->pi_lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 sk_lock-AF_KCM sk_lock-AF_INET irq_context: 0 sk_lock-AF_KCM sk_lock-AF_INET slock-AF_INET irq_context: 0 sk_lock-AF_KCM slock-AF_INET irq_context: 0 &sb->s_type->i_mutex_key#10 sk_lock-AF_KCM &data->lock irq_context: 0 &sb->s_type->i_mutex_key#10 sk_lock-AF_KCM rcu_node_0 irq_context: 0 &sb->s_type->i_mutex_key#10 sk_lock-AF_KCM &mux->lock irq_context: 0 &sb->s_type->i_mutex_key#10 sk_lock-AF_KCM rcu_read_lock &ei->socket.wq.wait irq_context: 0 &sb->s_type->i_mutex_key#10 sk_lock-AF_KCM rcu_read_lock &ei->socket.wq.wait &p->pi_lock irq_context: 0 &sb->s_type->i_mutex_key#10 sk_lock-AF_KCM rcu_read_lock &ei->socket.wq.wait &p->pi_lock &cfs_rq->removed.lock irq_context: 0 &sb->s_type->i_mutex_key#10 sk_lock-AF_KCM rcu_read_lock &ei->socket.wq.wait &p->pi_lock &rq->__lock irq_context: 0 &sb->s_type->i_mutex_key#10 sk_lock-AF_KCM rcu_read_lock &ei->socket.wq.wait &p->pi_lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 &sb->s_type->i_mutex_key#10 sk_lock-AF_INET clock-AF_INET &mux->rx_lock irq_context: 0 &sb->s_type->i_mutex_key#10 (work_completion)(&(&strp->msg_timer_work)->work) irq_context: 0 &sb->s_type->i_mutex_key#10 (work_completion)(&strp->work)#2 irq_context: 0 &sb->s_type->i_mutex_key#10 sk_lock-AF_INET prog_idr_lock irq_context: 0 &sb->s_type->i_mutex_key#10 sk_lock-AF_INET prog_idr_lock &obj_hash[i].lock irq_context: 0 &sb->s_type->i_mutex_key#10 sk_lock-AF_INET prog_idr_lock pool_lock#2 irq_context: 0 &sb->s_type->i_mutex_key#10 sk_lock-AF_INET bpf_lock irq_context: 0 &sb->s_type->i_mutex_key#10 sk_lock-AF_INET &mux->lock irq_context: 0 sk_lock-AF_INET k-sk_lock-AF_INET &tcp_hashinfo.bhash[i].lock &c->lock irq_context: 0 &f->f_pos_lock sb_writers#15 irq_context: 0 (wq_completion)wg-kex-wg0#4 (work_completion)(&({ do { const void *__vpp_verify = (typeof((worker) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __asm__ ("" : "=r"(__ptr) : "0"((typeof(*((worker))) *)((worker)))); (typeof((typeof(*((worker))) *)((worker)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); })->work) &handshake->lock remove_cache_srcu rcu_read_lock rcu_node_0 irq_context: 0 (wq_completion)wg-kex-wg0#4 (work_completion)(&({ do { const void *__vpp_verify = (typeof((worker) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __asm__ ("" : "=r"(__ptr) : "0"((typeof(*((worker))) *)((worker)))); (typeof((typeof(*((worker))) *)((worker)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); })->work) &handshake->lock remove_cache_srcu rcu_read_lock &rq->__lock irq_context: 0 (wq_completion)wg-kex-wg0#4 (work_completion)(&({ do { const void *__vpp_verify = (typeof((worker) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __asm__ ("" : "=r"(__ptr) : "0"((typeof(*((worker))) *)((worker)))); (typeof((typeof(*((worker))) *)((worker)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); })->work) &handshake->lock remove_cache_srcu rcu_read_lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 &sb->s_type->i_mutex_key#10 sk_lock-AF_INET slock-AF_INET &tcp_hashinfo.bhash[i].lock stock_lock rcu_read_lock per_cpu_ptr(&cgroup_rstat_cpu_lock, cpu) irq_context: 0 sk_lock-AF_INET slock-AF_INET elock-AF_INET irq_context: softirq (&n->timer) k-slock-AF_INET6 rcu_read_lock rcu_read_lock rcu_read_lock_bh noop_qdisc.q.lock irq_context: softirq (&n->timer) k-slock-AF_INET6 rcu_read_lock rcu_read_lock rcu_read_lock_bh &data->lock irq_context: softirq (&n->timer) k-slock-AF_INET6 rcu_read_lock rcu_read_lock rcu_read_lock_bh &obj_hash[i].lock irq_context: softirq (&n->timer) k-slock-AF_INET6 rcu_read_lock rcu_read_lock rcu_read_lock_bh pool_lock#2 irq_context: 0 (wq_completion)bond10 (work_completion)(&(&bond->mii_work)->work) &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 cb_lock nlk_cb_mutex-GENERIC genl_mutex rtnl_mutex &obj_hash[i].lock irq_context: 0 sk_lock-AF_INET rcu_read_lock rcu_read_lock rcu_read_lock_bh &data->lock &obj_hash[i].lock irq_context: 0 sk_lock-AF_INET rcu_read_lock rcu_read_lock rcu_read_lock_bh &data->lock &base->lock irq_context: 0 sk_lock-AF_INET rcu_read_lock rcu_read_lock rcu_read_lock_bh &data->lock &base->lock &obj_hash[i].lock irq_context: 0 cb_lock genl_mutex rcu_state.exp_mutex rcu_node_0 irq_context: 0 cb_lock genl_mutex rcu_state.exp_mutex &obj_hash[i].lock irq_context: 0 cb_lock genl_mutex rcu_state.exp_mutex rcu_read_lock &pool->lock irq_context: 0 cb_lock genl_mutex rcu_state.exp_mutex rcu_read_lock &pool->lock &obj_hash[i].lock irq_context: 0 cb_lock genl_mutex rcu_state.exp_mutex rcu_read_lock &pool->lock &p->pi_lock irq_context: 0 cb_lock genl_mutex rcu_state.exp_mutex rcu_read_lock &pool->lock &p->pi_lock &rq->__lock irq_context: 0 cb_lock genl_mutex rcu_state.exp_mutex rcu_read_lock &pool->lock &p->pi_lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 cb_lock genl_mutex rcu_state.exp_mutex &rq->__lock irq_context: 0 cb_lock genl_mutex rcu_state.exp_mutex &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 cb_lock genl_mutex rcu_state.exp_mutex &rnp->exp_wq[0] irq_context: 0 (wq_completion)bond7 (work_completion)(&(&bond->mii_work)->work) &rq->__lock irq_context: 0 (wq_completion)bond7 (work_completion)(&(&bond->mii_work)->work) &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 cb_lock genl_mutex rcu_state.exp_mutex &rnp->exp_wq[1] irq_context: 0 sb_writers#4 &type->i_mutex_dir_key#3 &pcp->lock &zone->lock &____s->seqcount irq_context: 0 sk_lock-AF_ALG &mm->mmap_lock sb_pagefaults jbd2_handle &journal->j_state_lock irq_context: 0 sk_lock-AF_ALG &mm->mmap_lock sb_pagefaults jbd2_handle &journal->j_state_lock &journal->j_wait_commit irq_context: 0 sk_lock-AF_ALG &mm->mmap_lock sb_pagefaults jbd2_handle &journal->j_state_lock &journal->j_wait_commit &p->pi_lock irq_context: 0 sk_lock-AF_ALG &mm->mmap_lock sb_pagefaults jbd2_handle &journal->j_state_lock &journal->j_wait_commit &p->pi_lock &rq->__lock irq_context: 0 sk_lock-AF_ALG &mm->mmap_lock sb_pagefaults jbd2_handle &journal->j_state_lock &journal->j_wait_commit &p->pi_lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 &sb->s_type->i_mutex_key#10 sk_lock-AF_INET6 &rq->__lock &cfs_rq->removed.lock irq_context: 0 sk_lock-AF_ALG &mm->mmap_lock sb_pagefaults jbd2_handle &cfs_rq->removed.lock irq_context: 0 sk_lock-AF_ALG &mm->mmap_lock sb_pagefaults jbd2_handle &obj_hash[i].lock irq_context: 0 recent_mutex irq_context: 0 recent_mutex &rq->__lock irq_context: 0 recent_mutex &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 recent_mutex fs_reclaim irq_context: 0 recent_mutex fs_reclaim mmu_notifier_invalidate_range_start irq_context: 0 recent_mutex pool_lock#2 irq_context: 0 recent_mutex proc_subdir_lock irq_context: 0 recent_mutex proc_inum_ida.xa_lock irq_context: 0 recent_mutex proc_subdir_lock irq_context: 0 recent_mutex recent_lock irq_context: 0 recent_mutex &ent->pde_unload_lock irq_context: 0 recent_mutex &obj_hash[i].lock irq_context: 0 sk_lock-AF_INET rcu_read_lock &____s->seqcount#12 irq_context: 0 sk_lock-AF_INET rcu_read_lock rcu_read_lock rcu_read_lock_bh &sch->q.lock irq_context: 0 sk_lock-AF_INET rcu_read_lock rcu_read_lock rcu_read_lock_bh &sch->q.lock &c->lock irq_context: 0 sk_lock-AF_INET rcu_read_lock rcu_read_lock rcu_read_lock_bh &sch->q.lock pool_lock#2 irq_context: 0 sk_lock-AF_INET rcu_read_lock rcu_read_lock rcu_read_lock_bh &sch->q.lock tk_core.seq.seqcount irq_context: 0 sk_lock-AF_INET rcu_read_lock rcu_read_lock rcu_read_lock_bh &sch->q.lock hrtimer_bases.lock irq_context: 0 sk_lock-AF_INET rcu_read_lock rcu_read_lock rcu_read_lock_bh &sch->q.lock hrtimer_bases.lock &obj_hash[i].lock irq_context: 0 sk_lock-AF_INET rcu_read_lock rcu_read_lock rcu_read_lock_bh &sch->q.lock hrtimer_bases.lock tk_core.seq.seqcount irq_context: softirq rcu_read_lock rcu_read_lock slock-AF_INET/1 rcu_read_lock rcu_read_lock rcu_read_lock_bh &sch->q.lock irq_context: softirq rcu_read_lock rcu_read_lock slock-AF_INET/1 rcu_read_lock rcu_read_lock rcu_read_lock_bh &sch->q.lock pool_lock#2 irq_context: softirq rcu_read_lock rcu_read_lock slock-AF_INET/1 rcu_read_lock rcu_read_lock rcu_read_lock_bh &sch->q.lock tk_core.seq.seqcount irq_context: softirq rcu_read_lock rcu_read_lock slock-AF_INET/1 rcu_read_lock rcu_read_lock rcu_read_lock_bh &sch->q.lock hrtimer_bases.lock irq_context: softirq rcu_read_lock rcu_read_lock slock-AF_INET/1 rcu_read_lock rcu_read_lock rcu_read_lock_bh &sch->q.lock hrtimer_bases.lock &obj_hash[i].lock irq_context: softirq rcu_read_lock rcu_read_lock slock-AF_INET/1 rcu_read_lock rcu_read_lock rcu_read_lock_bh &sch->q.lock hrtimer_bases.lock tk_core.seq.seqcount irq_context: softirq slock-AF_INET rcu_read_lock rcu_read_lock rcu_read_lock_bh &sch->q.lock irq_context: softirq slock-AF_INET rcu_read_lock rcu_read_lock rcu_read_lock_bh &sch->q.lock pool_lock#2 irq_context: softirq slock-AF_INET rcu_read_lock rcu_read_lock rcu_read_lock_bh &sch->q.lock tk_core.seq.seqcount irq_context: softirq slock-AF_INET rcu_read_lock rcu_read_lock rcu_read_lock_bh &sch->q.lock hrtimer_bases.lock irq_context: softirq slock-AF_INET rcu_read_lock rcu_read_lock rcu_read_lock_bh &sch->q.lock hrtimer_bases.lock &obj_hash[i].lock irq_context: softirq slock-AF_INET rcu_read_lock rcu_read_lock rcu_read_lock_bh &sch->q.lock hrtimer_bases.lock tk_core.seq.seqcount irq_context: softirq slock-AF_INET rcu_read_lock rcu_read_lock rcu_read_lock_bh &sch->q.lock &c->lock irq_context: softirq slock-AF_INET &n->list_lock irq_context: softirq slock-AF_INET &n->list_lock &c->lock irq_context: softirq rcu_read_lock rcu_read_lock slock-AF_INET/1 rcu_read_lock rcu_read_lock rcu_read_lock_bh &sch->q.lock &c->lock irq_context: softirq rcu_read_lock rcu_read_lock slock-AF_INET/1 rcu_read_lock rcu_read_lock rcu_read_lock_bh &sch->q.lock &n->list_lock irq_context: softirq rcu_read_lock rcu_read_lock slock-AF_INET/1 rcu_read_lock rcu_read_lock rcu_read_lock_bh &sch->q.lock &n->list_lock &c->lock irq_context: softirq slock-AF_INET rcu_read_lock rcu_read_lock rcu_read_lock_bh &sch->q.lock batched_entropy_u8.lock irq_context: softirq slock-AF_INET rcu_read_lock rcu_read_lock rcu_read_lock_bh &sch->q.lock kfence_freelist_lock irq_context: 0 &sb->s_type->i_mutex_key#10 sk_lock-AF_INET rcu_read_lock rcu_read_lock rcu_read_lock_bh &sch->q.lock irq_context: 0 &sb->s_type->i_mutex_key#10 sk_lock-AF_INET rcu_read_lock rcu_read_lock rcu_read_lock_bh &sch->q.lock pool_lock#2 irq_context: 0 &sb->s_type->i_mutex_key#10 sk_lock-AF_INET rcu_read_lock rcu_read_lock rcu_read_lock_bh &sch->q.lock tk_core.seq.seqcount irq_context: 0 &sb->s_type->i_mutex_key#10 sk_lock-AF_INET &data->lock irq_context: 0 &sb->s_type->i_mutex_key#10 sk_lock-AF_INET slock-AF_INET &data->lock irq_context: softirq rcu_read_lock rcu_read_lock rcu_read_lock rcu_read_lock &____s->seqcount#12 irq_context: 0 &sb->s_type->i_mutex_key#10 sk_lock-AF_TIPC &data->lock irq_context: 0 (wq_completion)wg-kex-wg0#7 (work_completion)(&peer->transmit_handshake_work) &peer->endpoint_lock rcu_read_lock_bh &data->lock &obj_hash[i].lock irq_context: 0 (wq_completion)wg-kex-wg0#7 (work_completion)(&peer->transmit_handshake_work) &peer->endpoint_lock rcu_read_lock_bh &data->lock &base->lock irq_context: 0 (wq_completion)wg-kex-wg0#7 (work_completion)(&peer->transmit_handshake_work) &peer->endpoint_lock rcu_read_lock_bh &data->lock &base->lock &obj_hash[i].lock irq_context: 0 sk_lock-AF_INET6 rcu_read_lock &____s->seqcount#12 irq_context: 0 sk_lock-AF_INET6 rcu_read_lock &____s->seqcount#16 irq_context: 0 (wq_completion)events (work_completion)(&rdev->wiphy_work) &rdev->wiphy.mtx &cfs_rq->removed.lock irq_context: 0 (wq_completion)wg-crypt-wg2#5 (work_completion)(&peer->transmit_packet_work) &peer->endpoint_lock rcu_read_lock_bh rcu_read_lock &____s->seqcount#12 irq_context: 0 (wq_completion)wg-crypt-wg2#5 (work_completion)(&peer->transmit_packet_work) &peer->endpoint_lock rcu_read_lock_bh rcu_read_lock &____s->seqcount#16 irq_context: 0 (wq_completion)wg-crypt-wg2#5 (work_completion)(&peer->transmit_packet_work) &peer->endpoint_lock rcu_read_lock_bh &data->lock irq_context: 0 rcu_read_lock rcu_read_lock rcu_read_lock_bh &data->lock irq_context: 0 rcu_read_lock rcu_read_lock &____s->seqcount#10 irq_context: 0 rcu_read_lock &n->lock &data->lock irq_context: softirq (&tw->tw_timer) &obj_hash[i].lock pool_lock irq_context: 0 rtnl_mutex devnet_rename_sem &root->kernfs_rwsem &____s->seqcount irq_context: 0 rtnl_mutex devnet_rename_sem &root->kernfs_rwsem rcu_read_lock pool_lock#2 irq_context: 0 rcu_read_lock rcu_read_lock_bh &sch->q.lock irq_context: 0 rcu_read_lock rcu_read_lock_bh &sch->q.lock &c->lock irq_context: 0 rcu_read_lock rcu_read_lock_bh &sch->q.lock pool_lock#2 irq_context: 0 rcu_read_lock rcu_read_lock_bh &sch->q.lock tk_core.seq.seqcount irq_context: 0 rcu_read_lock rcu_read_lock_bh &sch->q.lock hrtimer_bases.lock irq_context: 0 rcu_read_lock rcu_read_lock_bh &sch->q.lock hrtimer_bases.lock &obj_hash[i].lock irq_context: softirq rcu_read_lock rcu_read_lock k-slock-AF_INET rcu_read_lock &____s->seqcount#12 irq_context: softirq rcu_read_lock rcu_read_lock k-slock-AF_INET rcu_read_lock rcu_read_lock_bh &sch->q.lock irq_context: softirq rcu_read_lock rcu_read_lock k-slock-AF_INET rcu_read_lock rcu_read_lock_bh &sch->q.lock pool_lock#2 irq_context: softirq rcu_read_lock rcu_read_lock k-slock-AF_INET rcu_read_lock rcu_read_lock_bh &sch->q.lock tk_core.seq.seqcount irq_context: softirq rcu_read_lock rcu_read_lock k-slock-AF_INET rcu_read_lock rcu_read_lock_bh &sch->q.lock hrtimer_bases.lock irq_context: softirq rcu_read_lock rcu_read_lock k-slock-AF_INET rcu_read_lock rcu_read_lock_bh &sch->q.lock hrtimer_bases.lock &obj_hash[i].lock irq_context: softirq rcu_read_lock rcu_read_lock k-slock-AF_INET rcu_read_lock rcu_read_lock_bh &sch->q.lock hrtimer_bases.lock tk_core.seq.seqcount irq_context: 0 dup_mmap_sem &mm->mmap_lock &mm->mmap_lock/1 rcu_read_lock per_cpu_ptr(&cgroup_rstat_cpu_lock, cpu) irq_context: softirq (&n->timer) k-slock-AF_INET6 rcu_read_lock rcu_read_lock rcu_read_lock_bh &data->lock &obj_hash[i].lock irq_context: softirq (&n->timer) k-slock-AF_INET6 rcu_read_lock rcu_read_lock rcu_read_lock_bh &data->lock &base->lock irq_context: softirq (&n->timer) k-slock-AF_INET6 rcu_read_lock rcu_read_lock rcu_read_lock_bh &data->lock &base->lock &obj_hash[i].lock irq_context: 0 (wq_completion)ext4-rsv-conversion (work_completion)(&ei->i_rsv_conversion_work) jbd2_handle rcu_read_lock &rcu_state.gp_wq irq_context: 0 (wq_completion)ext4-rsv-conversion (work_completion)(&ei->i_rsv_conversion_work) jbd2_handle rcu_read_lock &rcu_state.gp_wq &p->pi_lock irq_context: 0 (wq_completion)ext4-rsv-conversion (work_completion)(&ei->i_rsv_conversion_work) jbd2_handle rcu_read_lock &rcu_state.gp_wq &p->pi_lock &rq->__lock irq_context: 0 (wq_completion)ext4-rsv-conversion (work_completion)(&ei->i_rsv_conversion_work) jbd2_handle rcu_read_lock &rcu_state.gp_wq &p->pi_lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 dup_mmap_sem &mm->mmap_lock &mm->mmap_lock/1 &anon_vma->rwsem rcu_read_lock pool_lock#2 irq_context: 0 dup_mmap_sem &mm->mmap_lock &mm->mmap_lock/1 &anon_vma->rwsem pcpu_lock stock_lock irq_context: 0 sb_writers#4 &type->i_mutex_dir_key#3/1 sb_internal jbd2_handle rcu_read_lock pgd_lock irq_context: 0 sb_writers#4 &type->i_mutex_dir_key#3/1 sb_internal jbd2_handle rcu_read_lock stock_lock irq_context: 0 sb_writers#4 &type->i_mutex_dir_key#3/1 sb_internal jbd2_handle rcu_read_lock key irq_context: 0 sb_writers#4 &type->i_mutex_dir_key#3/1 sb_internal jbd2_handle rcu_read_lock pcpu_lock irq_context: 0 sb_writers#4 &type->i_mutex_dir_key#3/1 sb_internal jbd2_handle rcu_read_lock percpu_counters_lock irq_context: 0 sb_writers#4 &type->i_mutex_dir_key#3/1 sb_internal jbd2_handle rcu_read_lock pcpu_lock stock_lock irq_context: 0 sk_lock-AF_INET6 rcu_read_lock rcu_read_lock rcu_read_lock_bh rcu_read_lock &data->lock irq_context: 0 (wq_completion)events_long (work_completion)(&(&br->gc_work)->work) rcu_read_lock &br->hash_lock &c->lock irq_context: 0 sk_lock-AF_INET6 rcu_read_lock rcu_read_lock &____s->seqcount#12 irq_context: 0 sk_lock-AF_INET6 rcu_read_lock rcu_read_lock &____s->seqcount#16 irq_context: 0 sk_lock-AF_INET6 rcu_read_lock rcu_read_lock rcu_read_lock &rq->__lock &cfs_rq->removed.lock irq_context: 0 sk_lock-AF_INET k-sk_lock-AF_INET rcu_read_lock rcu_read_lock &n->lock irq_context: 0 sk_lock-AF_INET k-sk_lock-AF_INET rcu_read_lock rcu_read_lock &n->lock &data->lock irq_context: 0 sk_lock-AF_INET k-sk_lock-AF_INET rcu_read_lock rcu_read_lock &n->lock &obj_hash[i].lock irq_context: 0 sk_lock-AF_INET k-sk_lock-AF_INET rcu_read_lock rcu_read_lock &n->lock pool_lock#2 irq_context: 0 (wq_completion)wg-kex-wg1 (work_completion)(&({ do { const void *__vpp_verify = (typeof((worker) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __asm__ ("" : "=r"(__ptr) : "0"((typeof(*((worker))) *)((worker)))); (typeof((typeof(*((worker))) *)((worker)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); })->work) &handshake->lock &n->list_lock irq_context: 0 (wq_completion)wg-kex-wg1 (work_completion)(&({ do { const void *__vpp_verify = (typeof((worker) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __asm__ ("" : "=r"(__ptr) : "0"((typeof(*((worker))) *)((worker)))); (typeof((typeof(*((worker))) *)((worker)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); })->work) &handshake->lock &n->list_lock &c->lock irq_context: 0 sk_lock-AF_INET rcu_read_lock rcu_read_lock rcu_read_lock_bh &sch->q.lock &n->list_lock irq_context: 0 sk_lock-AF_INET rcu_read_lock rcu_read_lock rcu_read_lock_bh &sch->q.lock &n->list_lock &c->lock irq_context: 0 rtnl_mutex sk_lock-AF_CAN irq_context: 0 rtnl_mutex sk_lock-AF_CAN &rq->__lock irq_context: 0 rtnl_mutex sk_lock-AF_CAN &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 rtnl_mutex sk_lock-AF_CAN slock-AF_CAN irq_context: 0 rtnl_mutex slock-AF_CAN irq_context: 0 sk_lock-AF_INET rcu_read_lock rcu_read_lock rcu_read_lock_bh &sch->q.lock &____s->seqcount#2 irq_context: 0 sk_lock-AF_INET rcu_read_lock rcu_read_lock rcu_read_lock_bh &sch->q.lock &____s->seqcount irq_context: 0 sk_lock-AF_INET rcu_read_lock rcu_read_lock rcu_read_lock_bh &sch->q.lock rcu_read_lock pool_lock#2 irq_context: 0 sk_lock-AF_INET rcu_read_lock rcu_read_lock rcu_read_lock_bh &sch->q.lock &obj_hash[i].lock irq_context: softirq rcu_read_lock rcu_read_lock slock-AF_INET/1 rcu_read_lock rcu_read_lock rcu_read_lock_bh &sch->q.lock &____s->seqcount#2 irq_context: softirq rcu_read_lock rcu_read_lock slock-AF_INET/1 rcu_read_lock rcu_read_lock rcu_read_lock_bh &sch->q.lock &____s->seqcount irq_context: softirq slock-AF_INET rcu_read_lock pool_lock#2 irq_context: softirq slock-AF_INET rcu_read_lock rcu_read_lock rcu_read_lock_bh &sch->q.lock &____s->seqcount#2 irq_context: softirq slock-AF_INET rcu_read_lock rcu_read_lock rcu_read_lock_bh &sch->q.lock &____s->seqcount irq_context: softirq rcu_read_lock rcu_read_lock slock-AF_INET/1 batched_entropy_u32.lock irq_context: softirq (&net->ipv6.ip6_fib_timer) irq_context: softirq (&net->ipv6.ip6_fib_timer) &net->ipv6.fib6_gc_lock irq_context: softirq (&net->ipv6.ip6_fib_timer) &net->ipv6.fib6_gc_lock rcu_read_lock &tb->tb6_lock irq_context: softirq (&net->ipv6.ip6_fib_timer) &net->ipv6.fib6_gc_lock rcu_read_lock &tb->tb6_lock &net->ipv6.fib6_walker_lock irq_context: softirq (&net->ipv6.ip6_fib_timer) &net->ipv6.fib6_gc_lock rcu_read_lock &tb->tb6_lock rt6_exception_lock irq_context: softirq (&net->ipv6.ip6_fib_timer) &net->ipv6.fib6_gc_lock rcu_read_lock &tb->tb6_lock pool_lock#2 irq_context: softirq (&net->ipv6.ip6_fib_timer) &net->ipv6.fib6_gc_lock rcu_read_lock &tb->tb6_lock nl_table_lock irq_context: softirq (&net->ipv6.ip6_fib_timer) &net->ipv6.fib6_gc_lock rcu_read_lock &tb->tb6_lock &obj_hash[i].lock irq_context: softirq (&net->ipv6.ip6_fib_timer) &net->ipv6.fib6_gc_lock rcu_read_lock &tb->tb6_lock nl_table_wait.lock irq_context: softirq (&net->ipv6.ip6_fib_timer) &net->ipv6.fib6_gc_lock &obj_hash[i].lock irq_context: 0 sb_writers#4 &type->i_mutex_dir_key#3/1 &type->i_mutex_dir_key#3 remove_cache_srcu &pcp->lock &zone->lock irq_context: 0 sb_writers#4 &type->i_mutex_dir_key#3/1 &type->i_mutex_dir_key#3 remove_cache_srcu &pcp->lock &zone->lock &____s->seqcount irq_context: 0 &sb->s_type->i_mutex_key#10 sk_lock-AF_INET rcu_read_lock &rq->__lock irq_context: 0 &sb->s_type->i_mutex_key#10 sk_lock-AF_INET rcu_read_lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 &sb->s_type->i_mutex_key#10 sk_lock-AF_INET rcu_read_lock rcu_read_lock rcu_read_lock_bh &sch->q.lock hrtimer_bases.lock irq_context: 0 &sb->s_type->i_mutex_key#10 sk_lock-AF_INET rcu_read_lock rcu_read_lock rcu_read_lock_bh &sch->q.lock hrtimer_bases.lock &obj_hash[i].lock irq_context: 0 &sb->s_type->i_mutex_key#9 stock_lock rcu_read_lock per_cpu_ptr(&cgroup_rstat_cpu_lock, cpu) irq_context: 0 (wq_completion)wg-crypt-wg2#5 (work_completion)(&peer->transmit_packet_work) &peer->endpoint_lock rcu_read_lock_bh &data->lock &obj_hash[i].lock irq_context: 0 (wq_completion)wg-crypt-wg2#5 (work_completion)(&peer->transmit_packet_work) &peer->endpoint_lock rcu_read_lock_bh &data->lock &base->lock irq_context: 0 (wq_completion)wg-crypt-wg2#5 (work_completion)(&peer->transmit_packet_work) &peer->endpoint_lock rcu_read_lock_bh &data->lock &base->lock &obj_hash[i].lock irq_context: softirq slock-AF_INET rcu_read_lock rcu_read_lock rcu_read_lock_bh &sch->q.lock &n->list_lock irq_context: softirq slock-AF_INET rcu_read_lock rcu_read_lock rcu_read_lock_bh &sch->q.lock &n->list_lock &c->lock irq_context: 0 sk_lock-AF_INET slock-AF_INET rcu_read_lock rcu_read_lock rcu_read_lock_bh &sch->q.lock irq_context: 0 sk_lock-AF_INET slock-AF_INET rcu_read_lock rcu_read_lock rcu_read_lock_bh &sch->q.lock pool_lock#2 irq_context: 0 sk_lock-AF_INET slock-AF_INET rcu_read_lock rcu_read_lock rcu_read_lock_bh &sch->q.lock tk_core.seq.seqcount irq_context: 0 sk_lock-AF_INET slock-AF_INET rcu_read_lock rcu_read_lock rcu_read_lock_bh &sch->q.lock hrtimer_bases.lock irq_context: 0 sk_lock-AF_INET slock-AF_INET rcu_read_lock rcu_read_lock rcu_read_lock_bh &sch->q.lock hrtimer_bases.lock &obj_hash[i].lock irq_context: 0 sk_lock-AF_INET slock-AF_INET rcu_read_lock rcu_read_lock rcu_read_lock_bh &sch->q.lock &c->lock irq_context: softirq slock-AF_INET rcu_read_lock rcu_read_lock rcu_read_lock_bh dev->qdisc_tx_busylock ?: &qdisc_tx_busylock _xmit_ETHER#2 rcu_read_lock &____s->seqcount#2 irq_context: softirq (&icsk->icsk_retransmit_timer) slock-AF_INET rcu_read_lock rcu_read_lock rcu_read_lock_bh &sch->q.lock irq_context: softirq (&icsk->icsk_retransmit_timer) slock-AF_INET rcu_read_lock rcu_read_lock rcu_read_lock_bh &sch->q.lock pool_lock#2 irq_context: softirq (&icsk->icsk_retransmit_timer) slock-AF_INET rcu_read_lock rcu_read_lock rcu_read_lock_bh &sch->q.lock tk_core.seq.seqcount irq_context: softirq (&icsk->icsk_retransmit_timer) slock-AF_INET rcu_read_lock rcu_read_lock rcu_read_lock_bh &sch->q.lock hrtimer_bases.lock irq_context: softirq (&icsk->icsk_retransmit_timer) slock-AF_INET rcu_read_lock rcu_read_lock rcu_read_lock_bh &sch->q.lock hrtimer_bases.lock &obj_hash[i].lock irq_context: softirq (&icsk->icsk_retransmit_timer) slock-AF_INET rcu_read_lock rcu_read_lock rcu_read_lock_bh &sch->q.lock hrtimer_bases.lock tk_core.seq.seqcount irq_context: 0 &sb->s_type->i_mutex_key#10 sk_lock-AF_INET &____s->seqcount#2 irq_context: 0 &sb->s_type->i_mutex_key#10 sk_lock-AF_INET &____s->seqcount irq_context: 0 sk_lock-AF_INET slock-AF_INET &c->lock irq_context: 0 sk_lock-AF_INET slock-AF_INET &____s->seqcount#2 irq_context: 0 sk_lock-AF_INET slock-AF_INET &____s->seqcount irq_context: 0 sk_lock-AF_INET slock-AF_INET pool_lock#2 irq_context: 0 &sb->s_type->i_mutex_key#10 sk_lock-AF_INET rcu_read_lock rcu_read_lock rcu_read_lock_bh &sch->q.lock &c->lock irq_context: 0 &sb->s_type->i_mutex_key#10 sk_lock-AF_INET rcu_read_lock rcu_read_lock rcu_read_lock_bh &sch->q.lock &n->list_lock irq_context: 0 &sb->s_type->i_mutex_key#10 sk_lock-AF_INET rcu_read_lock rcu_read_lock rcu_read_lock_bh &sch->q.lock &n->list_lock &c->lock irq_context: 0 &sb->s_type->i_mutex_key#10 sk_lock-AF_INET rcu_read_lock rcu_read_lock rcu_read_lock_bh &sch->q.lock hrtimer_bases.lock tk_core.seq.seqcount irq_context: softirq (&icsk->icsk_retransmit_timer) slock-AF_INET rcu_read_lock rcu_read_lock rcu_read_lock_bh &sch->q.lock &c->lock irq_context: softirq (&icsk->icsk_retransmit_timer) slock-AF_INET hrtimer_bases.lock irq_context: softirq (&icsk->icsk_retransmit_timer) slock-AF_INET hrtimer_bases.lock &obj_hash[i].lock irq_context: softirq (&icsk->icsk_retransmit_timer) slock-AF_INET hrtimer_bases.lock tk_core.seq.seqcount irq_context: 0 sk_lock-AF_INET &mm->mmap_lock rcu_read_lock per_cpu_ptr(&cgroup_rstat_cpu_lock, cpu) irq_context: 0 &ep->mtx sk_lock-AF_VSOCK irq_context: 0 &ep->mtx sk_lock-AF_VSOCK slock-AF_VSOCK irq_context: 0 &ep->mtx slock-AF_VSOCK irq_context: 0 sk_lock-AF_INET rcu_read_lock rcu_read_lock rcu_read_lock_bh &sch->q.lock batched_entropy_u8.lock irq_context: 0 sk_lock-AF_INET rcu_read_lock rcu_read_lock rcu_read_lock_bh &sch->q.lock kfence_freelist_lock irq_context: softirq (&ndev->rs_timer) rcu_read_lock &____s->seqcount#12 irq_context: softirq (&ndev->rs_timer) rcu_read_lock &____s->seqcount#16 irq_context: softirq (&ndev->rs_timer) rcu_read_lock pool_lock#2 irq_context: softirq (&ndev->rs_timer) &obj_hash[i].lock irq_context: softirq (&ndev->rs_timer) &data->lock irq_context: 0 &ep->mtx rlock-AF_PACKET irq_context: 0 &ep->mtx wlock-AF_PACKET irq_context: 0 &ep->mtx rcu_state.exp_mutex rcu_node_0 irq_context: 0 &ep->mtx rcu_state.exp_mutex &obj_hash[i].lock irq_context: 0 &ep->mtx rcu_state.exp_mutex rcu_read_lock &pool->lock irq_context: 0 &ep->mtx rcu_state.exp_mutex rcu_read_lock &pool->lock &obj_hash[i].lock irq_context: 0 &ep->mtx rcu_state.exp_mutex rcu_read_lock &pool->lock &p->pi_lock irq_context: 0 &ep->mtx rcu_state.exp_mutex &rq->__lock irq_context: 0 &ep->mtx rcu_state.exp_mutex &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 &ep->mtx rcu_state.exp_mutex &rnp->exp_wq[1] irq_context: 0 uevent_sock_mutex nl_table_wait.lock &p->pi_lock irq_context: 0 gdp_mutex kernfs_idr_lock irq_context: 0 &mm->mmap_lock sb_pagefaults jbd2_handle rcu_read_lock &rcu_state.gp_wq irq_context: 0 &mm->mmap_lock sb_pagefaults jbd2_handle rcu_read_lock &rcu_state.gp_wq &p->pi_lock irq_context: 0 &mm->mmap_lock sb_pagefaults jbd2_handle rcu_read_lock &rcu_state.gp_wq &p->pi_lock &rq->__lock irq_context: 0 &mm->mmap_lock sb_pagefaults jbd2_handle rcu_read_lock &rcu_state.gp_wq &p->pi_lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 &ep->mtx rcu_state.exp_mutex rcu_read_lock &pool->lock &p->pi_lock &rq->__lock irq_context: 0 &ep->mtx rcu_state.exp_mutex rcu_read_lock &pool->lock &p->pi_lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 &ep->mtx rcu_state.exp_mutex &rnp->exp_wq[2] irq_context: 0 &ep->mtx rcu_state.exp_mutex &rnp->exp_wq[0] irq_context: 0 &sb->s_type->i_mutex_key#10 slock-AF_INET6 &dccp_hashinfo.bhash[i].lock &dccp_hashinfo.bhash2[i].lock quarantine_lock irq_context: 0 kn->active#73 &rq->__lock irq_context: 0 kn->active#73 fs_reclaim irq_context: 0 kn->active#73 fs_reclaim mmu_notifier_invalidate_range_start irq_context: 0 kn->active#73 stock_lock irq_context: 0 kn->active#73 &kernfs_locks->open_file_mutex[count] irq_context: 0 kn->active#73 &kernfs_locks->open_file_mutex[count] fs_reclaim irq_context: 0 kn->active#73 &kernfs_locks->open_file_mutex[count] fs_reclaim mmu_notifier_invalidate_range_start irq_context: 0 kn->active#73 &c->lock irq_context: 0 kn->active#73 &kernfs_locks->open_file_mutex[count] &c->lock irq_context: 0 kn->active#60 fs_reclaim &rq->__lock irq_context: 0 kn->active#60 fs_reclaim &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 misc_mtx rcu_node_0 irq_context: 0 &f->f_pos_lock sb_writers#4 &sb->s_type->i_mutex_key#8 &mm->mmap_lock fs_reclaim &rq->__lock irq_context: 0 &f->f_pos_lock sb_writers#4 &sb->s_type->i_mutex_key#8 &mm->mmap_lock fs_reclaim &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 sb_writers#4 &sb->s_type->i_mutex_key#8 &sb->s_type->i_mutex_key#8/4 jbd2_handle per_cpu_ptr(&cgroup_rstat_cpu_lock, cpu) irq_context: 0 &sb->s_type->i_mutex_key#8 &sbi->s_writepages_rwsem rcu_read_lock key#10 irq_context: 0 rtnl_mutex free_vmap_area_lock irq_context: 0 rtnl_mutex vmap_area_lock irq_context: 0 rtnl_mutex purge_vmap_area_lock irq_context: 0 rtnl_mutex purge_vmap_area_lock &obj_hash[i].lock irq_context: 0 rtnl_mutex purge_vmap_area_lock pool_lock#2 irq_context: 0 &sb->s_type->i_mutex_key#10 sk_lock-AF_INET6 slock-AF_INET6 &tcp_hashinfo.bhash[i].lock stock_lock rcu_read_lock per_cpu_ptr(&cgroup_rstat_cpu_lock, cpu) irq_context: 0 (wq_completion)gid-cache-wq (work_completion)(&work->work) &meta->lock irq_context: 0 (wq_completion)gid-cache-wq (work_completion)(&work->work) kfence_freelist_lock irq_context: 0 (wq_completion)mld (work_completion)(&(&idev->mc_ifc_work)->work) &idev->mc_lock rcu_read_lock rcu_read_lock &____s->seqcount#12 irq_context: 0 (wq_completion)mld (work_completion)(&(&idev->mc_ifc_work)->work) &idev->mc_lock rcu_read_lock rcu_read_lock &____s->seqcount#16 irq_context: 0 (wq_completion)mld (work_completion)(&(&idev->mc_ifc_work)->work) &idev->mc_lock rcu_read_lock &obj_hash[i].lock irq_context: 0 (wq_completion)mld (work_completion)(&(&idev->mc_ifc_work)->work) &idev->mc_lock rcu_read_lock &data->lock irq_context: 0 sk_lock-AF_AX25 &mm->mmap_lock irq_context: 0 sk_lock-AF_AX25 ax25_uid_lock irq_context: 0 cb_lock genl_mutex rcu_read_lock &ei->socket.wq.wait &p->pi_lock &cfs_rq->removed.lock irq_context: 0 ax25_uid_lock irq_context: 0 sk_lock-AF_AX25 ax25_dev_lock irq_context: 0 cb_lock genl_mutex rcu_read_lock &ei->socket.wq.wait &p->pi_lock &rq->__lock &cfs_rq->removed.lock irq_context: 0 sk_lock-AF_AX25 ax25_uid_lock irq_context: 0 sk_lock-AF_AX25 fs_reclaim irq_context: 0 sk_lock-AF_AX25 fs_reclaim mmu_notifier_invalidate_range_start irq_context: 0 sk_lock-AF_AX25 &c->lock irq_context: 0 sk_lock-AF_AX25 pool_lock#2 irq_context: softirq (&ndev->rs_timer) rcu_read_lock rcu_read_lock rcu_read_lock_bh rcu_read_lock rcu_read_lock rcu_read_lock_bh &____s->seqcount irq_context: softirq (&ndev->rs_timer) rcu_read_lock rcu_read_lock rcu_read_lock_bh rcu_read_lock_bh &data->lock irq_context: 0 sb_writers#4 &type->i_mutex_dir_key#3/1 &sb->s_type->i_lock_key#22 bit_wait_table + i irq_context: 0 rtnl_mutex dev_addr_sem irq_context: 0 rtnl_mutex dev_addr_sem &rq->__lock irq_context: 0 rtnl_mutex dev_addr_sem &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 sb_writers#4 &type->i_mutex_dir_key#3/1 jbd2_handle &ei->i_data_sem remove_cache_srcu &pcp->lock &zone->lock irq_context: 0 sb_writers#4 &type->i_mutex_dir_key#3/1 jbd2_handle &ei->i_data_sem remove_cache_srcu &pcp->lock &zone->lock &____s->seqcount irq_context: 0 sk_lock-AF_INET rcu_read_lock (console_sem).lock irq_context: 0 sk_lock-AF_INET rcu_read_lock console_lock console_srcu console_owner_lock irq_context: 0 sk_lock-AF_INET rcu_read_lock console_lock console_srcu console_owner irq_context: 0 sk_lock-AF_INET rcu_read_lock console_lock console_srcu console_owner &port_lock_key irq_context: 0 sk_lock-AF_INET rcu_read_lock console_lock console_srcu console_owner console_owner_lock irq_context: softirq rcu_read_lock rcu_read_lock slock-AF_INET/1 &tcp_hashinfo.bhash[i].lock stock_lock irq_context: softirq rcu_read_lock rcu_read_lock slock-AF_INET/1 &tcp_hashinfo.bhash[i].lock &obj_hash[i].lock irq_context: softirq rcu_read_lock rcu_read_lock slock-AF_INET/1 &tcp_hashinfo.bhash[i].lock &tcp_hashinfo.bhash2[i].lock stock_lock irq_context: 0 sb_writers &sb->s_type->i_mutex_key#4 rcu_read_lock &rq->__lock &cfs_rq->removed.lock irq_context: 0 sb_writers#4 &sb->s_type->i_mutex_key#8 &sbi->s_writepages_rwsem rcu_state.exp_mutex rcu_read_lock &pool->lock fill_pool_map-wait-type-override pool_lock#2 irq_context: 0 sb_writers#4 &sb->s_type->i_mutex_key#8 &sbi->s_writepages_rwsem rcu_state.exp_mutex rcu_read_lock &pool->lock fill_pool_map-wait-type-override &c->lock irq_context: 0 sb_writers#4 &sb->s_type->i_mutex_key#8 &sbi->s_writepages_rwsem rcu_state.exp_mutex rcu_read_lock &pool->lock fill_pool_map-wait-type-override pool_lock irq_context: softirq rcu_read_lock rcu_read_lock rcu_read_lock rcu_read_lock rcu_read_lock_bh &sch->q.lock &n->list_lock irq_context: softirq rcu_read_lock rcu_read_lock rcu_read_lock rcu_read_lock rcu_read_lock_bh &sch->q.lock &n->list_lock &c->lock irq_context: 0 &vma->vm_lock->lock remove_cache_srcu pool_lock#2 irq_context: 0 lock#3 rcu_read_lock &pool->lock &p->pi_lock &rq->__lock &cfs_rq->removed.lock irq_context: 0 sb_writers#4 &type->i_mutex_dir_key#3/1 sb_internal jbd2_handle &sbi->s_orphan_lock &rq->__lock &cfs_rq->removed.lock irq_context: 0 sk_lock-AF_INET k-sk_lock-AF_INET/1 &n->list_lock irq_context: 0 sk_lock-AF_INET k-sk_lock-AF_INET/1 &n->list_lock &c->lock irq_context: 0 (wq_completion)wg-crypt-wg0#4 (work_completion)(&peer->transmit_packet_work) &peer->endpoint_lock rcu_read_lock_bh rcu_read_lock batched_entropy_u8.lock irq_context: 0 (wq_completion)wg-crypt-wg0#4 (work_completion)(&peer->transmit_packet_work) &peer->endpoint_lock rcu_read_lock_bh rcu_read_lock kfence_freelist_lock irq_context: softirq rcu_read_lock rcu_read_lock rcu_read_lock rcu_read_lock rcu_read_lock_bh &dir->lock irq_context: softirq rcu_read_lock rcu_read_lock rcu_read_lock rcu_read_lock rcu_read_lock_bh stock_lock irq_context: softirq rcu_read_lock rcu_read_lock rcu_read_lock rcu_read_lock rcu_read_lock_bh rcu_read_lock &c->lock irq_context: 0 sb_writers#4 &sb->s_type->i_mutex_key#8 mapping.invalidate_lock jbd2_handle &ei->i_data_sem rcu_read_lock &cfs_rq->removed.lock irq_context: 0 (wq_completion)ipv6_addrconf (work_completion)(&(&ifa->dad_work)->work) rtnl_mutex &idev->mc_lock rcu_read_lock rcu_read_lock rcu_read_lock_bh rcu_read_lock rlock-AF_PACKET irq_context: 0 (wq_completion)ipv6_addrconf (work_completion)(&(&ifa->dad_work)->work) rtnl_mutex &idev->mc_lock rcu_read_lock rcu_read_lock rcu_read_lock_bh &data->lock irq_context: 0 sk_lock-AF_INET rcu_read_lock rcu_read_lock rcu_read_lock_bh rcu_read_lock &____s->seqcount irq_context: 0 sk_lock-AF_INET rcu_read_lock rcu_read_lock rcu_read_lock_bh rcu_read_lock rcu_read_lock pool_lock#2 irq_context: 0 sk_lock-AF_INET rcu_read_lock rcu_read_lock rcu_read_lock_bh rcu_read_lock &obj_hash[i].lock irq_context: 0 (wq_completion)mld (work_completion)(&(&idev->mc_dad_work)->work) &idev->mc_lock rcu_read_lock rcu_read_lock rcu_read_lock_bh &data->lock irq_context: softirq rcu_read_lock rcu_read_lock rcu_read_lock rcu_read_lock &____s->seqcount#2 irq_context: 0 sk_lock-AF_INET6 rcu_read_lock rcu_read_lock_bh rcu_read_lock &data->lock irq_context: 0 sb_writers#4 &sb->s_type->i_mutex_key#8 jbd2_handle &ei->i_data_sem &rq->__lock irq_context: 0 sk_lock-AF_INET6 batched_entropy_u8.lock crngs.lock irq_context: 0 sk_lock-AF_INET6 rcu_read_lock rcu_read_lock_bh rcu_read_lock &meta->lock irq_context: 0 sk_lock-AF_INET6 rcu_read_lock rcu_read_lock_bh rcu_read_lock kfence_freelist_lock irq_context: 0 &mm->mmap_lock remove_cache_srcu &rcu_state.expedited_wq irq_context: 0 sk_lock-AF_INET6 rcu_read_lock rcu_read_lock rcu_read_lock &nf_conntrack_locks[i] irq_context: 0 sk_lock-AF_INET6 rcu_read_lock rcu_read_lock rcu_read_lock &nf_conntrack_locks[i] &nf_conntrack_locks[i]/1 irq_context: 0 sk_lock-AF_INET6 rcu_read_lock rcu_read_lock rcu_read_lock &nf_conntrack_locks[i]/1 irq_context: 0 sk_lock-AF_INET6 rcu_read_lock rcu_read_lock rcu_read_lock rcu_read_lock &nf_nat_locks[i] irq_context: 0 &sb->s_type->i_mutex_key#10 sk_lock-AF_CAIF console_owner_lock irq_context: 0 &sb->s_type->i_mutex_key#10 sk_lock-AF_CAIF console_owner irq_context: 0 &sb->s_type->i_mutex_key#10 sk_lock-AF_CAIF console_lock console_srcu console_owner_lock irq_context: 0 &sb->s_type->i_mutex_key#10 sk_lock-AF_CAIF console_lock console_srcu console_owner irq_context: 0 &sb->s_type->i_mutex_key#10 sk_lock-AF_CAIF console_lock console_srcu console_owner console_owner_lock irq_context: 0 sk_lock-AF_INET &tcp_hashinfo.bhash[i].lock &tcp_hashinfo.bhash2[i].lock batched_entropy_u8.lock crngs.lock irq_context: 0 pernet_ops_rwsem nf_ct_proto_mutex defrag6_mutex nf_hook_mutex fs_reclaim &rq->__lock irq_context: 0 pernet_ops_rwsem k-sk_lock-AF_RXRPC fill_pool_map-wait-type-override &c->lock irq_context: 0 pernet_ops_rwsem k-sk_lock-AF_RXRPC fill_pool_map-wait-type-override &n->list_lock irq_context: 0 pernet_ops_rwsem k-sk_lock-AF_RXRPC fill_pool_map-wait-type-override &n->list_lock &c->lock irq_context: 0 pernet_ops_rwsem k-sk_lock-AF_RXRPC fill_pool_map-wait-type-override &rq->__lock irq_context: 0 pernet_ops_rwsem k-sk_lock-AF_RXRPC fill_pool_map-wait-type-override &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)wg-crypt-wg0#4 (work_completion)(&peer->transmit_packet_work) rcu_read_lock_bh rcu_read_lock &pool->lock/1 irq_context: 0 (wq_completion)wg-crypt-wg0#4 (work_completion)(&peer->transmit_packet_work) rcu_read_lock_bh rcu_read_lock &pool->lock/1 &obj_hash[i].lock irq_context: 0 (wq_completion)wg-crypt-wg0#4 (work_completion)(&peer->transmit_packet_work) rcu_read_lock_bh rcu_read_lock &pool->lock/1 &p->pi_lock irq_context: 0 pernet_ops_rwsem k-sk_lock-AF_INET6 &tcp_hashinfo.bhash[i].lock &____s->seqcount#2 irq_context: 0 pernet_ops_rwsem k-sk_lock-AF_INET6 &tcp_hashinfo.bhash[i].lock &____s->seqcount irq_context: 0 (wq_completion)bat_events (work_completion)(&(&forw_packet_aggr->delayed_work)->work) &hard_iface->bat_iv.ogm_buff_mutex &bat_priv->tvlv.container_list_lock quarantine_lock irq_context: softirq rcu_read_lock rcu_read_lock rcu_read_lock rcu_read_lock batched_entropy_u8.lock irq_context: softirq rcu_read_lock rcu_read_lock rcu_read_lock rcu_read_lock kfence_freelist_lock irq_context: 0 fill_pool_map-wait-type-override &rcu_state.expedited_wq &p->pi_lock irq_context: 0 fill_pool_map-wait-type-override &rcu_state.expedited_wq &p->pi_lock &rq->__lock irq_context: 0 fill_pool_map-wait-type-override &rcu_state.expedited_wq &p->pi_lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 sk_lock-AF_INET &tcp_hashinfo.bhash[i].lock &tcp_hashinfo.bhash2[i].lock &____s->seqcount#2 irq_context: 0 &f->f_pos_lock sb_writers#4 &sb->s_type->i_mutex_key#8 &ei->i_data_sem &ei->i_es_lock &pcp->lock &zone->lock irq_context: 0 &sb->s_type->i_mutex_key#8 &sbi->s_writepages_rwsem jbd2_handle &ei->i_data_sem &lg->lg_mutex &c->lock irq_context: softirq rcu_read_lock &sch->q.lock hrtimer_bases.lock fill_pool_map-wait-type-override pool_lock#2 irq_context: softirq rcu_read_lock &sch->q.lock hrtimer_bases.lock fill_pool_map-wait-type-override pool_lock irq_context: 0 rtnl_mutex lock link_idr_lock &c->lock irq_context: softirq (&data->send_timer) rcu_read_lock &pool->lock &p->pi_lock &rq->__lock &cfs_rq->removed.lock irq_context: 0 rtnl_mutex &root->kernfs_rwsem pgd_lock irq_context: 0 rtnl_mutex &root->kernfs_rwsem stock_lock irq_context: 0 rtnl_mutex &root->kernfs_rwsem key irq_context: 0 rtnl_mutex &root->kernfs_rwsem pcpu_lock irq_context: 0 rtnl_mutex &root->kernfs_rwsem percpu_counters_lock irq_context: 0 rtnl_mutex &root->kernfs_rwsem pcpu_lock stock_lock irq_context: 0 rtnl_mutex cpu_hotplug_lock jump_label_mutex text_mutex &rq->__lock irq_context: 0 rtnl_mutex cpu_hotplug_lock jump_label_mutex text_mutex &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 &sb->s_type->i_mutex_key#8 batched_entropy_u8.lock crngs.lock irq_context: 0 (wq_completion)ext4-rsv-conversion (work_completion)(&ei->i_rsv_conversion_work) jbd2_handle &ei->i_data_sem batched_entropy_u8.lock crngs.lock irq_context: 0 (wq_completion)bat_events (work_completion)(&(&bat_priv->bla.work)->work) rcu_read_lock &data->lock &obj_hash[i].lock irq_context: 0 (wq_completion)bat_events (work_completion)(&(&bat_priv->bla.work)->work) rcu_read_lock &data->lock &base->lock irq_context: 0 (wq_completion)bat_events (work_completion)(&(&bat_priv->bla.work)->work) rcu_read_lock &data->lock &base->lock &obj_hash[i].lock irq_context: 0 &xt[i].mutex remove_cache_srcu rcu_read_lock &rcu_state.expedited_wq irq_context: 0 &xt[i].mutex remove_cache_srcu rcu_read_lock &rcu_state.expedited_wq &p->pi_lock irq_context: 0 &xt[i].mutex remove_cache_srcu rcu_read_lock &rcu_state.expedited_wq &p->pi_lock &rq->__lock irq_context: 0 &xt[i].mutex remove_cache_srcu rcu_read_lock &rcu_state.expedited_wq &p->pi_lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)bond10 (work_completion)(&(&bond->ad_work)->work) &bond->mode_lock rcu_read_lock &c->lock irq_context: 0 cb_lock &data->lock &obj_hash[i].lock irq_context: 0 cb_lock &data->lock &base->lock irq_context: 0 cb_lock &data->lock &base->lock &obj_hash[i].lock irq_context: 0 cgroup_threadgroup_rwsem rcu_read_lock &rq->__lock &cfs_rq->removed.lock irq_context: 0 (wq_completion)wg-crypt-wg0#4 (work_completion)(&peer->transmit_packet_work) &peer->endpoint_lock rcu_read_lock_bh &data->lock &obj_hash[i].lock irq_context: 0 (wq_completion)wg-crypt-wg0#4 (work_completion)(&peer->transmit_packet_work) &peer->endpoint_lock rcu_read_lock_bh &data->lock &base->lock irq_context: 0 (wq_completion)wg-crypt-wg0#4 (work_completion)(&peer->transmit_packet_work) &peer->endpoint_lock rcu_read_lock_bh &data->lock &base->lock &obj_hash[i].lock irq_context: 0 sk_lock-AF_INET &tcp_hashinfo.bhash[i].lock &____s->seqcount#2 irq_context: 0 sk_lock-AF_VSOCK rcu_read_lock &data->lock irq_context: 0 rcu_read_lock rcu_read_lock pcpu_lock stock_lock irq_context: 0 &sb->s_type->i_mutex_key#10 &net->sctp.addr_wq_lock slock-AF_INET6/1 &zone->lock irq_context: 0 &sb->s_type->i_mutex_key#10 &net->sctp.addr_wq_lock slock-AF_INET6/1 &zone->lock &____s->seqcount irq_context: softirq (&ndev->rs_timer) rcu_read_lock &c->lock irq_context: 0 rtnl_mutex rcu_read_lock &dev_addr_list_lock_key#2/3 irq_context: 0 rtnl_mutex rcu_read_lock &dev_addr_list_lock_key#2/3 pool_lock#2 irq_context: 0 rtnl_mutex &br->hash_lock &dev_addr_list_lock_key#2/3 irq_context: 0 rtnl_mutex &br->hash_lock &dev_addr_list_lock_key#2/3 &c->lock irq_context: 0 rtnl_mutex &br->hash_lock &dev_addr_list_lock_key#2/3 pool_lock#2 irq_context: 0 rtnl_mutex &br->lock &br->hash_lock &dev_addr_list_lock_key#2/3 irq_context: 0 rtnl_mutex &br->lock &br->hash_lock &dev_addr_list_lock_key#2/3 &obj_hash[i].lock irq_context: 0 rtnl_mutex &br->lock &br->hash_lock &dev_addr_list_lock_key#2/3 pool_lock#2 irq_context: 0 rtnl_mutex &br->lock &br->hash_lock &dev_addr_list_lock_key#2/3 krc.lock irq_context: 0 sk_lock-AF_INET6 remove_cache_srcu &pcp->lock &zone->lock irq_context: 0 sk_lock-AF_INET6 remove_cache_srcu &pcp->lock &zone->lock &____s->seqcount irq_context: 0 rtnl_mutex rcu_read_lock &dev_addr_list_lock_key#2/1 irq_context: 0 rtnl_mutex rcu_read_lock &dev_addr_list_lock_key#2/1 &c->lock irq_context: 0 rtnl_mutex &br->hash_lock &dev_addr_list_lock_key#2/1 irq_context: 0 rtnl_mutex &br->hash_lock &dev_addr_list_lock_key#2/1 &c->lock irq_context: 0 rtnl_mutex &br->lock &br->hash_lock &dev_addr_list_lock_key#2/1 irq_context: 0 rtnl_mutex &br->lock &br->hash_lock &dev_addr_list_lock_key#2/1 &obj_hash[i].lock irq_context: 0 rtnl_mutex &br->lock &br->hash_lock &dev_addr_list_lock_key#2/1 krc.lock irq_context: 0 (wq_completion)bat_events (work_completion)(&(&bat_priv->mcast.work)->work) &bat_priv->mcast.mla_lock (console_sem).lock irq_context: 0 (wq_completion)bat_events (work_completion)(&(&bat_priv->mcast.work)->work) &bat_priv->mcast.mla_lock console_lock console_srcu console_owner_lock irq_context: 0 (wq_completion)bat_events (work_completion)(&(&bat_priv->mcast.work)->work) &bat_priv->mcast.mla_lock console_lock console_srcu console_owner irq_context: 0 (wq_completion)bat_events (work_completion)(&(&bat_priv->mcast.work)->work) &bat_priv->mcast.mla_lock console_lock console_srcu console_owner &port_lock_key irq_context: 0 (wq_completion)bat_events (work_completion)(&(&bat_priv->mcast.work)->work) &bat_priv->mcast.mla_lock console_lock console_srcu console_owner console_owner_lock irq_context: 0 (wq_completion)bat_events (work_completion)(&(&bat_priv->mcast.work)->work) &bat_priv->mcast.mla_lock &bat_priv->tvlv.container_list_lock &obj_hash[i].lock irq_context: 0 (wq_completion)bat_events (work_completion)(&(&bat_priv->mcast.work)->work) &bat_priv->mcast.mla_lock &bat_priv->tvlv.container_list_lock pool_lock#2 irq_context: 0 sk_lock-AF_INET rcu_read_lock rcu_read_lock &____s->seqcount#2 irq_context: 0 sb_writers#4 &type->i_mutex_dir_key#3 tomoyo_ss &rq->__lock &cfs_rq->removed.lock irq_context: 0 sb_writers#4 &pipe->mutex/1 &sb->s_type->i_mutex_key#8 jbd2_handle &ei->i_data_sem &lg->lg_mutex mmu_notifier_invalidate_range_start irq_context: 0 sb_writers#4 &pipe->mutex/1 &sb->s_type->i_mutex_key#8 jbd2_handle &ei->i_data_sem &lg->lg_mutex pool_lock#2 irq_context: 0 sk_lock-AF_INET6 &dccp_hashinfo.bhash[i].lock &n->list_lock irq_context: 0 sk_lock-AF_INET6 &dccp_hashinfo.bhash[i].lock &n->list_lock &c->lock irq_context: 0 &sb->s_type->i_mutex_key#10 (netlink_chain).rwsem &nft_net->commit_mutex rcu_state.exp_mutex &rq->__lock &cfs_rq->removed.lock irq_context: 0 &sb->s_type->i_mutex_key#10 (netlink_chain).rwsem &nft_net->commit_mutex rcu_state.exp_mutex rcu_read_lock &rq->__lock irq_context: 0 &sb->s_type->i_mutex_key#10 (netlink_chain).rwsem &nft_net->commit_mutex rcu_state.exp_mutex rcu_read_lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 sb_writers#4 &pipe->mutex/1 &sb->s_type->i_mutex_key#8 jbd2_handle &ei->i_data_sem &c->lock irq_context: 0 pernet_ops_rwsem devices_rwsem rdma_nets_rwsem &device->compat_devs_mutex fs_reclaim mmu_notifier_invalidate_range_start pool_lock#2 irq_context: 0 sb_writers#4 &pipe->mutex/1 &sb->s_type->i_mutex_key#8 jbd2_handle &ei->i_data_sem &lg->lg_mutex rcu_read_lock &rq->__lock irq_context: 0 sb_writers#4 &pipe->mutex/1 &sb->s_type->i_mutex_key#8 mapping.invalidate_lock jbd2_handle &ei->i_data_sem &bgl->locks[i].lock &obj_hash[i].lock irq_context: 0 sb_writers#4 &pipe->mutex/1 &sb->s_type->i_mutex_key#8 mapping.invalidate_lock jbd2_handle &ei->i_data_sem &bgl->locks[i].lock pool_lock#2 irq_context: 0 sk_lock-AF_TIPC k-slock-AF_TIPC &data->lock irq_context: 0 sk_lock-AF_INET6 &dccp_hashinfo.bhash[i].lock &hashinfo->ehash_locks[i] irq_context: 0 sk_lock-AF_TIPC &data->lock &obj_hash[i].lock irq_context: 0 sk_lock-AF_TIPC &data->lock &base->lock irq_context: 0 sk_lock-AF_TIPC &data->lock &base->lock &obj_hash[i].lock irq_context: 0 pernet_ops_rwsem rtnl_mutex &idev->mc_lock rcu_node_0 irq_context: 0 pernet_ops_rwsem rtnl_mutex &idev->mc_lock &rcu_state.expedited_wq irq_context: 0 pernet_ops_rwsem rtnl_mutex &idev->mc_lock &rcu_state.expedited_wq &p->pi_lock irq_context: 0 pernet_ops_rwsem rtnl_mutex &idev->mc_lock &rcu_state.expedited_wq &p->pi_lock &rq->__lock irq_context: 0 pernet_ops_rwsem rtnl_mutex &idev->mc_lock &rcu_state.expedited_wq &p->pi_lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 &sb->s_type->i_mutex_key#10 sk_lock-AF_TIPC rcu_read_lock rcu_node_0 irq_context: 0 (wq_completion)events (work_completion)(&data->dm_alert_work) remove_cache_srcu rcu_read_lock rcu_node_0 irq_context: 0 (wq_completion)events (work_completion)(&data->dm_alert_work) remove_cache_srcu rcu_read_lock &rq->__lock irq_context: 0 (wq_completion)bond14 irq_context: 0 (wq_completion)bond14 (work_completion)(&(&slave->notify_work)->work) irq_context: 0 (wq_completion)bond14 (work_completion)(&(&slave->notify_work)->work) &obj_hash[i].lock irq_context: 0 (wq_completion)bond14 (work_completion)(&(&slave->notify_work)->work) &base->lock irq_context: 0 (wq_completion)bond14 (work_completion)(&(&slave->notify_work)->work) &base->lock &obj_hash[i].lock irq_context: 0 (wq_completion)bond14 (work_completion)(&(&slave->notify_work)->work) &rq->__lock irq_context: 0 rtnl_mutex &dev_addr_list_lock_key/1 &bridge_netdev_addr_lock_key irq_context: 0 (wq_completion)bond14 (work_completion)(&(&bond->mcast_work)->work) irq_context: 0 (wq_completion)bond14 (work_completion)(&(&bond->mcast_work)->work) &obj_hash[i].lock irq_context: 0 (wq_completion)bond14 (work_completion)(&(&bond->mcast_work)->work) &base->lock irq_context: 0 (wq_completion)bond14 (work_completion)(&(&bond->mcast_work)->work) &base->lock &obj_hash[i].lock irq_context: 0 (wq_completion)ipv6_addrconf (work_completion)(&(&ifa->dad_work)->work) rtnl_mutex &idev->mc_lock &dev_addr_list_lock_key/1 rcu_read_lock &bridge_netdev_addr_lock_key irq_context: 0 (wq_completion)ipv6_addrconf (work_completion)(&(&ifa->dad_work)->work) rtnl_mutex &idev->mc_lock &dev_addr_list_lock_key/1 rcu_read_lock &bridge_netdev_addr_lock_key &c->lock irq_context: 0 (wq_completion)ipv6_addrconf (work_completion)(&(&ifa->dad_work)->work) rtnl_mutex &idev->mc_lock &dev_addr_list_lock_key/1 rcu_read_lock &bridge_netdev_addr_lock_key pool_lock#2 irq_context: 0 (wq_completion)bond14 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex fs_reclaim irq_context: 0 (wq_completion)bond14 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex fs_reclaim mmu_notifier_invalidate_range_start irq_context: 0 (wq_completion)bond14 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex fs_reclaim mmu_notifier_invalidate_range_start &rq->__lock irq_context: 0 (wq_completion)bond14 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex fs_reclaim mmu_notifier_invalidate_range_start &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)bond14 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex pool_lock#2 irq_context: 0 (wq_completion)bond14 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &c->lock irq_context: 0 (wq_completion)bond14 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &____s->seqcount#2 irq_context: 0 (wq_completion)bond14 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &____s->seqcount irq_context: 0 (wq_completion)bond14 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex rcu_read_lock &bond->stats_lock/1 irq_context: 0 (wq_completion)bond14 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex rcu_read_lock &ndev->lock irq_context: 0 (wq_completion)bond14 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex nl_table_lock irq_context: 0 (wq_completion)bond14 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &obj_hash[i].lock irq_context: 0 (wq_completion)bond14 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex nl_table_wait.lock irq_context: 0 (wq_completion)bond14 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex net_rwsem irq_context: 0 (wq_completion)bond14 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex net_rwsem &list->lock#2 irq_context: 0 (wq_completion)bond14 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &tn->lock irq_context: 0 (wq_completion)bond14 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &tn->lock &rq->__lock irq_context: 0 (wq_completion)bond14 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &tn->lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)bond14 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &idev->mc_lock irq_context: 0 (wq_completion)bond14 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &idev->mc_lock fs_reclaim irq_context: 0 (wq_completion)bond14 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &idev->mc_lock fs_reclaim mmu_notifier_invalidate_range_start irq_context: 0 (wq_completion)bond14 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &idev->mc_lock pool_lock#2 irq_context: 0 (wq_completion)bond14 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &idev->mc_lock &c->lock irq_context: 0 (wq_completion)bond14 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &idev->mc_lock rcu_read_lock &ndev->lock irq_context: 0 (wq_completion)bond14 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &idev->mc_lock rcu_read_lock pool_lock#2 irq_context: 0 (wq_completion)bond14 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &idev->mc_lock rcu_read_lock &c->lock irq_context: 0 (wq_completion)bond14 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &idev->mc_lock rcu_read_lock &dir->lock#2 irq_context: 0 (wq_completion)bond14 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &idev->mc_lock rcu_read_lock &ul->lock#2 irq_context: 0 (wq_completion)bond14 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &idev->mc_lock rcu_read_lock rcu_read_lock pool_lock#2 irq_context: 0 (wq_completion)bond14 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &idev->mc_lock rcu_read_lock rcu_read_lock &obj_hash[i].lock irq_context: 0 (wq_completion)bond14 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &idev->mc_lock rcu_read_lock rcu_read_lock &c->lock irq_context: 0 (wq_completion)bond14 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &idev->mc_lock rcu_read_lock rcu_read_lock &n->list_lock irq_context: 0 (wq_completion)bond14 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &idev->mc_lock rcu_read_lock rcu_read_lock &n->list_lock &c->lock irq_context: 0 (wq_completion)bond14 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &idev->mc_lock rcu_read_lock rcu_read_lock rcu_node_0 irq_context: 0 (wq_completion)bond14 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &idev->mc_lock rcu_read_lock rcu_read_lock &rq->__lock irq_context: 0 (wq_completion)bond14 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &idev->mc_lock rcu_read_lock rcu_read_lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)bond14 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &idev->mc_lock rcu_read_lock rcu_read_lock &dir->lock#2 irq_context: 0 (wq_completion)bond14 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &idev->mc_lock rcu_read_lock rcu_read_lock &tbl->lock irq_context: 0 (wq_completion)bond14 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &idev->mc_lock rcu_read_lock rcu_read_lock &n->lock irq_context: 0 (wq_completion)bond14 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &idev->mc_lock rcu_read_lock rcu_read_lock &____s->seqcount#9 irq_context: 0 (wq_completion)bond14 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &idev->mc_lock rcu_read_lock rcu_read_lock rcu_read_lock_bh rcu_read_lock rcu_read_lock_bh rcu_read_lock &br->multicast_lock irq_context: 0 (wq_completion)bond14 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &idev->mc_lock rcu_read_lock rcu_read_lock rcu_read_lock_bh rcu_read_lock rcu_read_lock_bh rcu_read_lock &br->multicast_lock pool_lock#2 irq_context: 0 (wq_completion)bond14 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &idev->mc_lock rcu_read_lock rcu_read_lock rcu_read_lock_bh rcu_read_lock rcu_read_lock_bh rcu_read_lock &br->multicast_lock &obj_hash[i].lock irq_context: 0 (wq_completion)bond14 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &idev->mc_lock rcu_read_lock rcu_read_lock rcu_read_lock_bh rcu_read_lock rcu_read_lock_bh rcu_read_lock &br->multicast_lock rcu_read_lock rhashtable_bucket irq_context: 0 (wq_completion)bond14 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &idev->mc_lock rcu_read_lock rcu_read_lock rcu_read_lock_bh rcu_read_lock rcu_read_lock_bh rcu_read_lock &br->multicast_lock nl_table_lock irq_context: 0 (wq_completion)bond14 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &idev->mc_lock rcu_read_lock rcu_read_lock rcu_read_lock_bh rcu_read_lock rcu_read_lock_bh rcu_read_lock &br->multicast_lock nl_table_wait.lock irq_context: 0 (wq_completion)bond14 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &idev->mc_lock rcu_read_lock rcu_read_lock rcu_read_lock_bh rcu_read_lock rcu_read_lock_bh rcu_read_lock &br->multicast_lock &base->lock irq_context: 0 (wq_completion)bond14 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &idev->mc_lock rcu_read_lock rcu_read_lock rcu_read_lock_bh rcu_read_lock rcu_read_lock_bh rcu_read_lock &br->multicast_lock &base->lock &obj_hash[i].lock irq_context: 0 (wq_completion)bond14 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &idev->mc_lock rcu_read_lock rcu_read_lock rcu_read_lock_bh rcu_read_lock rcu_read_lock_bh rcu_read_lock &data->lock irq_context: 0 (wq_completion)bond14 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &idev->mc_lock rcu_read_lock rcu_read_lock rcu_read_lock_bh rcu_read_lock rcu_read_lock_bh rcu_read_lock &obj_hash[i].lock irq_context: 0 (wq_completion)bond14 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &idev->mc_lock rcu_read_lock rcu_read_lock rcu_read_lock_bh rcu_read_lock rcu_read_lock_bh rcu_read_lock pool_lock#2 irq_context: 0 (wq_completion)bond14 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &idev->mc_lock rcu_read_lock rcu_node_0 irq_context: 0 (wq_completion)bond14 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &idev->mc_lock rcu_read_lock &rcu_state.gp_wq irq_context: 0 (wq_completion)bond14 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &idev->mc_lock rcu_read_lock &rcu_state.gp_wq &p->pi_lock irq_context: 0 (wq_completion)bond14 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &idev->mc_lock rcu_read_lock &rcu_state.gp_wq &p->pi_lock &rq->__lock irq_context: 0 (wq_completion)bond14 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &idev->mc_lock rcu_read_lock &rcu_state.gp_wq &p->pi_lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)bond14 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &rq->__lock irq_context: 0 (wq_completion)bond14 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)bond14 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex.wait_lock irq_context: 0 (wq_completion)bond14 (work_completion)(&(&bond->mcast_work)->work) &p->pi_lock irq_context: 0 (wq_completion)bond14 (work_completion)(&(&bond->mcast_work)->work) &p->pi_lock &rq->__lock irq_context: 0 (wq_completion)bond14 (work_completion)(&(&bond->mcast_work)->work) &p->pi_lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)mld (work_completion)(&(&idev->mc_ifc_work)->work) &idev->mc_lock rcu_read_lock rcu_read_lock rcu_read_lock_bh rcu_read_lock rcu_read_lock_bh rcu_read_lock &br->multicast_lock irq_context: 0 (wq_completion)mld (work_completion)(&(&idev->mc_ifc_work)->work) &idev->mc_lock rcu_read_lock rcu_read_lock rcu_read_lock_bh rcu_read_lock rcu_read_lock_bh rcu_read_lock &br->multicast_lock &obj_hash[i].lock irq_context: 0 (wq_completion)mld (work_completion)(&(&idev->mc_ifc_work)->work) &idev->mc_lock rcu_read_lock rcu_read_lock rcu_read_lock_bh rcu_read_lock rcu_read_lock_bh rcu_read_lock &br->multicast_lock &base->lock irq_context: 0 (wq_completion)mld (work_completion)(&(&idev->mc_ifc_work)->work) &idev->mc_lock rcu_read_lock rcu_read_lock rcu_read_lock_bh rcu_read_lock rcu_read_lock_bh rcu_read_lock &data->lock irq_context: 0 (wq_completion)bond14 (work_completion)(&(&slave->notify_work)->work) rtnl_mutex net_rwsem irq_context: 0 (wq_completion)bond14 (work_completion)(&(&slave->notify_work)->work) rtnl_mutex net_rwsem &list->lock#2 irq_context: 0 (wq_completion)bond14 (work_completion)(&(&slave->notify_work)->work) rtnl_mutex &tn->lock irq_context: 0 (wq_completion)bond14 (work_completion)(&(&slave->notify_work)->work) rtnl_mutex.wait_lock irq_context: 0 (wq_completion)bond14 (work_completion)(&(&slave->notify_work)->work) &p->pi_lock irq_context: softirq (&in_dev->mr_ifc_timer) rcu_read_lock rcu_read_lock_bh rcu_read_lock &data->lock &obj_hash[i].lock irq_context: softirq (&in_dev->mr_ifc_timer) rcu_read_lock rcu_read_lock_bh rcu_read_lock &data->lock &base->lock irq_context: softirq (&in_dev->mr_ifc_timer) rcu_read_lock rcu_read_lock_bh rcu_read_lock &data->lock &base->lock &obj_hash[i].lock irq_context: 0 (wq_completion)bond15 irq_context: 0 (wq_completion)bond15 (work_completion)(&(&slave->notify_work)->work) irq_context: 0 (wq_completion)bond15 (work_completion)(&(&slave->notify_work)->work) &obj_hash[i].lock irq_context: 0 (wq_completion)bond15 (work_completion)(&(&slave->notify_work)->work) &base->lock irq_context: 0 (wq_completion)bond15 (work_completion)(&(&slave->notify_work)->work) &base->lock &obj_hash[i].lock irq_context: 0 (wq_completion)bond15 (work_completion)(&(&slave->notify_work)->work) &rq->__lock irq_context: 0 (wq_completion)bond15 (work_completion)(&(&slave->notify_work)->work) &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)bond15 (work_completion)(&(&bond->mcast_work)->work) irq_context: 0 (wq_completion)bond15 (work_completion)(&(&bond->mcast_work)->work) &obj_hash[i].lock irq_context: 0 (wq_completion)bond15 (work_completion)(&(&bond->mcast_work)->work) &base->lock irq_context: 0 (wq_completion)bond15 (work_completion)(&(&bond->mcast_work)->work) &base->lock &obj_hash[i].lock irq_context: 0 (wq_completion)bond15 (work_completion)(&(&slave->notify_work)->work) rtnl_mutex net_rwsem irq_context: 0 (wq_completion)bond15 (work_completion)(&(&slave->notify_work)->work) rtnl_mutex net_rwsem &list->lock#2 irq_context: 0 (wq_completion)bond15 (work_completion)(&(&slave->notify_work)->work) rtnl_mutex net_rwsem &rq->__lock irq_context: 0 (wq_completion)bond15 (work_completion)(&(&slave->notify_work)->work) rtnl_mutex net_rwsem &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)bond15 (work_completion)(&(&slave->notify_work)->work) rtnl_mutex &tn->lock irq_context: 0 (wq_completion)bond15 (work_completion)(&(&slave->notify_work)->work) rtnl_mutex.wait_lock irq_context: 0 (wq_completion)bond15 (work_completion)(&(&slave->notify_work)->work) &p->pi_lock irq_context: 0 (wq_completion)bond15 (work_completion)(&(&slave->notify_work)->work) &p->pi_lock &rq->__lock irq_context: 0 (wq_completion)bond15 (work_completion)(&(&slave->notify_work)->work) &p->pi_lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)mld (work_completion)(&(&idev->mc_ifc_work)->work) &idev->mc_lock rcu_read_lock rcu_read_lock rcu_read_lock_bh rcu_read_lock rcu_read_lock_bh rcu_read_lock &br->multicast_lock pool_lock#2 irq_context: 0 (wq_completion)mld (work_completion)(&(&idev->mc_ifc_work)->work) &idev->mc_lock rcu_read_lock rcu_read_lock rcu_read_lock_bh rcu_read_lock rcu_read_lock_bh rcu_read_lock &br->multicast_lock rcu_read_lock rhashtable_bucket irq_context: 0 (wq_completion)mld (work_completion)(&(&idev->mc_ifc_work)->work) &idev->mc_lock rcu_read_lock rcu_read_lock rcu_read_lock_bh rcu_read_lock rcu_read_lock_bh rcu_read_lock &br->multicast_lock nl_table_lock irq_context: 0 (wq_completion)mld (work_completion)(&(&idev->mc_ifc_work)->work) &idev->mc_lock rcu_read_lock rcu_read_lock rcu_read_lock_bh rcu_read_lock rcu_read_lock_bh rcu_read_lock &br->multicast_lock nl_table_wait.lock irq_context: 0 (wq_completion)mld (work_completion)(&(&idev->mc_ifc_work)->work) &idev->mc_lock rcu_read_lock rcu_read_lock rcu_read_lock_bh rcu_read_lock rcu_read_lock_bh rcu_read_lock &br->multicast_lock &base->lock &obj_hash[i].lock irq_context: softirq &(&bond->mcast_work)->timer rcu_read_lock &pool->lock/1 &p->pi_lock &rq->__lock irq_context: softirq &(&bond->mcast_work)->timer rcu_read_lock &pool->lock/1 &p->pi_lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)ipv6_addrconf (work_completion)(&(&ifa->dad_work)->work) rtnl_mutex rcu_read_lock rcu_read_lock rcu_read_lock_bh rcu_read_lock rcu_read_lock_bh rcu_read_lock &data->lock irq_context: 0 (wq_completion)ipv6_addrconf (work_completion)(&(&ifa->dad_work)->work) rtnl_mutex rcu_read_lock rcu_read_lock rcu_read_lock_bh rcu_read_lock rcu_read_lock_bh rcu_read_lock &obj_hash[i].lock irq_context: 0 (wq_completion)ipv6_addrconf (work_completion)(&(&ifa->dad_work)->work) rtnl_mutex rcu_read_lock rcu_read_lock rcu_read_lock_bh rcu_read_lock rcu_read_lock_bh rcu_read_lock pool_lock#2 irq_context: 0 (wq_completion)bond15#2 irq_context: 0 (wq_completion)bond15#2 (work_completion)(&(&slave->notify_work)->work) irq_context: 0 (wq_completion)bond15#2 (work_completion)(&(&slave->notify_work)->work) &obj_hash[i].lock irq_context: 0 (wq_completion)bond15#2 (work_completion)(&(&slave->notify_work)->work) &base->lock irq_context: 0 (wq_completion)bond15#2 (work_completion)(&(&slave->notify_work)->work) &base->lock &obj_hash[i].lock irq_context: 0 (wq_completion)bond15#2 (work_completion)(&(&bond->mcast_work)->work) irq_context: 0 (wq_completion)bond15#2 (work_completion)(&(&bond->mcast_work)->work) &obj_hash[i].lock irq_context: 0 (wq_completion)bond15#2 (work_completion)(&(&bond->mcast_work)->work) &base->lock irq_context: 0 (wq_completion)bond15#2 (work_completion)(&(&bond->mcast_work)->work) &base->lock &obj_hash[i].lock irq_context: 0 (wq_completion)bond15 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex fs_reclaim irq_context: 0 (wq_completion)bond15 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex fs_reclaim mmu_notifier_invalidate_range_start irq_context: 0 (wq_completion)bond15 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &c->lock irq_context: 0 (wq_completion)bond15 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &n->list_lock irq_context: 0 (wq_completion)bond15 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &n->list_lock &c->lock irq_context: 0 (wq_completion)bond15 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex rcu_read_lock &bond->stats_lock/1 irq_context: 0 (wq_completion)bond15 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex rcu_read_lock &ndev->lock irq_context: 0 (wq_completion)bond15 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex nl_table_lock irq_context: 0 (wq_completion)bond15 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &obj_hash[i].lock irq_context: 0 (wq_completion)bond15 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex nl_table_wait.lock irq_context: 0 (wq_completion)bond15 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex net_rwsem irq_context: 0 (wq_completion)bond15 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex net_rwsem &list->lock#2 irq_context: 0 (wq_completion)bond15 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &tn->lock irq_context: 0 (wq_completion)bond15 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &idev->mc_lock irq_context: 0 (wq_completion)bond15 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &idev->mc_lock fs_reclaim irq_context: 0 (wq_completion)bond15 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &idev->mc_lock fs_reclaim mmu_notifier_invalidate_range_start irq_context: 0 (wq_completion)bond15 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &idev->mc_lock &c->lock irq_context: 0 (wq_completion)bond15 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &idev->mc_lock rcu_read_lock &ndev->lock irq_context: 0 (wq_completion)bond15 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &idev->mc_lock rcu_read_lock &dir->lock#2 irq_context: 0 (wq_completion)bond15 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &idev->mc_lock rcu_read_lock &c->lock irq_context: 0 (wq_completion)bond15 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &idev->mc_lock rcu_read_lock &ul->lock#2 irq_context: 0 (wq_completion)bond15 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &idev->mc_lock rcu_read_lock rcu_read_lock &____s->seqcount#10 irq_context: 0 (wq_completion)bond15 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &idev->mc_lock rcu_read_lock rcu_read_lock rcu_read_lock_bh rcu_read_lock rcu_read_lock_bh rcu_read_lock &br->multicast_lock irq_context: 0 (wq_completion)bond15 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &idev->mc_lock rcu_read_lock rcu_read_lock rcu_read_lock_bh rcu_read_lock rcu_read_lock_bh rcu_read_lock &br->multicast_lock &obj_hash[i].lock irq_context: 0 (wq_completion)bond15 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &idev->mc_lock rcu_read_lock rcu_read_lock rcu_read_lock_bh rcu_read_lock rcu_read_lock_bh rcu_read_lock &br->multicast_lock &base->lock irq_context: 0 (wq_completion)bond15 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &idev->mc_lock rcu_read_lock rcu_read_lock rcu_read_lock_bh rcu_read_lock rcu_read_lock_bh rcu_read_lock &data->lock irq_context: 0 (wq_completion)bond15 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &idev->mc_lock rcu_read_lock rcu_read_lock rcu_read_lock_bh rcu_read_lock rcu_read_lock_bh rcu_read_lock &obj_hash[i].lock irq_context: 0 (wq_completion)bond15 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &idev->mc_lock rcu_read_lock rcu_read_lock rcu_read_lock_bh rcu_read_lock rcu_read_lock_bh rcu_read_lock pool_lock#2 irq_context: 0 (wq_completion)bond15#2 (work_completion)(&(&slave->notify_work)->work) rtnl_mutex net_rwsem irq_context: 0 (wq_completion)bond15#2 (work_completion)(&(&slave->notify_work)->work) rtnl_mutex net_rwsem &list->lock#2 irq_context: 0 (wq_completion)bond15#2 (work_completion)(&(&slave->notify_work)->work) rtnl_mutex &tn->lock irq_context: 0 (wq_completion)bond15#2 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex fs_reclaim irq_context: 0 (wq_completion)bond15#2 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex fs_reclaim mmu_notifier_invalidate_range_start irq_context: 0 (wq_completion)bond15#2 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &c->lock irq_context: 0 (wq_completion)bond15#2 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex rcu_read_lock &bond->stats_lock/1 irq_context: 0 (wq_completion)bond15#2 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex rcu_read_lock &ndev->lock irq_context: 0 (wq_completion)bond15#2 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex nl_table_lock irq_context: 0 (wq_completion)bond15#2 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &obj_hash[i].lock irq_context: 0 (wq_completion)bond15#2 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex nl_table_wait.lock irq_context: 0 (wq_completion)bond15#2 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex net_rwsem irq_context: 0 (wq_completion)bond15#2 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex net_rwsem &list->lock#2 irq_context: 0 (wq_completion)bond15#2 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &tn->lock irq_context: 0 (wq_completion)bond15#2 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &idev->mc_lock irq_context: 0 (wq_completion)bond15#2 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &idev->mc_lock fs_reclaim irq_context: 0 (wq_completion)bond15#2 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &idev->mc_lock fs_reclaim mmu_notifier_invalidate_range_start irq_context: 0 (wq_completion)bond15#2 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &idev->mc_lock &c->lock irq_context: 0 (wq_completion)bond15#2 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &idev->mc_lock rcu_read_lock &ndev->lock irq_context: 0 (wq_completion)bond15#2 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &idev->mc_lock rcu_read_lock &dir->lock#2 irq_context: 0 (wq_completion)bond15#2 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &idev->mc_lock rcu_read_lock &ul->lock#2 irq_context: 0 (wq_completion)bond15#2 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &idev->mc_lock rcu_read_lock rcu_read_lock &obj_hash[i].lock irq_context: 0 (wq_completion)bond15#2 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &idev->mc_lock rcu_read_lock rcu_read_lock &dir->lock#2 irq_context: 0 (wq_completion)bond15#2 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &idev->mc_lock rcu_read_lock rcu_read_lock &tbl->lock irq_context: 0 (wq_completion)bond15#2 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &idev->mc_lock rcu_read_lock rcu_read_lock &n->lock irq_context: 0 (wq_completion)bond15#2 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &idev->mc_lock rcu_read_lock rcu_read_lock &____s->seqcount#9 irq_context: 0 (wq_completion)bond15#2 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &idev->mc_lock rcu_read_lock rcu_read_lock rcu_read_lock_bh rcu_read_lock rcu_read_lock_bh rcu_read_lock &br->multicast_lock irq_context: 0 (wq_completion)bond15#2 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &idev->mc_lock rcu_read_lock rcu_read_lock rcu_read_lock_bh rcu_read_lock rcu_read_lock_bh rcu_read_lock &br->multicast_lock &obj_hash[i].lock irq_context: 0 (wq_completion)bond15#2 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &idev->mc_lock rcu_read_lock rcu_read_lock rcu_read_lock_bh rcu_read_lock rcu_read_lock_bh rcu_read_lock &br->multicast_lock rcu_read_lock rhashtable_bucket irq_context: 0 (wq_completion)bond15#2 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &idev->mc_lock rcu_read_lock rcu_read_lock rcu_read_lock_bh rcu_read_lock rcu_read_lock_bh rcu_read_lock &br->multicast_lock nl_table_lock irq_context: 0 (wq_completion)bond15#2 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &idev->mc_lock rcu_read_lock rcu_read_lock rcu_read_lock_bh rcu_read_lock rcu_read_lock_bh rcu_read_lock &br->multicast_lock nl_table_wait.lock irq_context: 0 (wq_completion)bond15#2 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &idev->mc_lock rcu_read_lock rcu_read_lock rcu_read_lock_bh rcu_read_lock rcu_read_lock_bh rcu_read_lock &br->multicast_lock &base->lock irq_context: 0 (wq_completion)bond15#2 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &idev->mc_lock rcu_read_lock rcu_read_lock rcu_read_lock_bh rcu_read_lock rcu_read_lock_bh rcu_read_lock &br->multicast_lock &base->lock &obj_hash[i].lock irq_context: 0 (wq_completion)bond15#2 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &idev->mc_lock rcu_read_lock rcu_read_lock rcu_read_lock_bh rcu_read_lock rcu_read_lock_bh rcu_read_lock &data->lock irq_context: 0 (wq_completion)bond15#2 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &idev->mc_lock rcu_read_lock rcu_read_lock rcu_read_lock_bh rcu_read_lock rcu_read_lock_bh rcu_read_lock &obj_hash[i].lock irq_context: 0 (wq_completion)ipv6_addrconf (work_completion)(&(&ifa->dad_work)->work) rtnl_mutex &idev->mc_lock rcu_read_lock rcu_read_lock rcu_read_lock_bh rcu_read_lock rcu_read_lock_bh rcu_read_lock &br->multicast_lock irq_context: 0 (wq_completion)ipv6_addrconf (work_completion)(&(&ifa->dad_work)->work) rtnl_mutex &idev->mc_lock rcu_read_lock rcu_read_lock rcu_read_lock_bh rcu_read_lock rcu_read_lock_bh rcu_read_lock &br->multicast_lock &obj_hash[i].lock irq_context: 0 (wq_completion)ipv6_addrconf (work_completion)(&(&ifa->dad_work)->work) rtnl_mutex &idev->mc_lock rcu_read_lock rcu_read_lock rcu_read_lock_bh rcu_read_lock rcu_read_lock_bh rcu_read_lock &br->multicast_lock &base->lock irq_context: 0 (wq_completion)ipv6_addrconf (work_completion)(&(&ifa->dad_work)->work) rtnl_mutex &idev->mc_lock rcu_read_lock rcu_read_lock rcu_read_lock_bh rcu_read_lock rcu_read_lock_bh rcu_read_lock &br->multicast_lock &base->lock &obj_hash[i].lock irq_context: 0 (wq_completion)ipv6_addrconf (work_completion)(&(&ifa->dad_work)->work) rtnl_mutex &idev->mc_lock rcu_read_lock rcu_read_lock rcu_read_lock_bh rcu_read_lock rcu_read_lock_bh rcu_read_lock &data->lock irq_context: 0 (wq_completion)ipv6_addrconf (work_completion)(&(&ifa->dad_work)->work) rtnl_mutex &idev->mc_lock rcu_read_lock rcu_read_lock rcu_read_lock_bh rcu_read_lock rcu_read_lock_bh rcu_read_lock &obj_hash[i].lock irq_context: 0 (wq_completion)ipv6_addrconf (work_completion)(&(&ifa->dad_work)->work) rtnl_mutex &idev->mc_lock rcu_read_lock rcu_read_lock rcu_read_lock_bh rcu_read_lock rcu_read_lock_bh rcu_read_lock pool_lock#2 irq_context: 0 (wq_completion)mld (work_completion)(&(&idev->mc_dad_work)->work) &idev->mc_lock rcu_read_lock rcu_read_lock rcu_read_lock_bh rcu_read_lock rcu_read_lock_bh rcu_read_lock &br->multicast_lock irq_context: 0 (wq_completion)mld (work_completion)(&(&idev->mc_dad_work)->work) &idev->mc_lock rcu_read_lock rcu_read_lock rcu_read_lock_bh rcu_read_lock rcu_read_lock_bh rcu_read_lock &br->multicast_lock &obj_hash[i].lock irq_context: 0 (wq_completion)mld (work_completion)(&(&idev->mc_dad_work)->work) &idev->mc_lock rcu_read_lock rcu_read_lock rcu_read_lock_bh rcu_read_lock rcu_read_lock_bh rcu_read_lock &br->multicast_lock &base->lock irq_context: 0 (wq_completion)mld (work_completion)(&(&idev->mc_dad_work)->work) &idev->mc_lock rcu_read_lock rcu_read_lock rcu_read_lock_bh rcu_read_lock rcu_read_lock_bh rcu_read_lock &data->lock irq_context: 0 (wq_completion)bond16 irq_context: 0 (wq_completion)bond16 (work_completion)(&(&slave->notify_work)->work) irq_context: 0 (wq_completion)bond16 (work_completion)(&(&slave->notify_work)->work) &obj_hash[i].lock irq_context: 0 (wq_completion)bond16 (work_completion)(&(&slave->notify_work)->work) &base->lock irq_context: 0 (wq_completion)bond16 (work_completion)(&(&slave->notify_work)->work) &base->lock &obj_hash[i].lock irq_context: 0 (wq_completion)bond16 (work_completion)(&(&slave->notify_work)->work) &rq->__lock irq_context: 0 (wq_completion)bond16 (work_completion)(&(&slave->notify_work)->work) &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: softirq &(&bond->mcast_work)->timer rcu_read_lock &pool->lock/1 &p->pi_lock &cfs_rq->removed.lock irq_context: 0 (wq_completion)bond16 (work_completion)(&(&bond->mcast_work)->work) irq_context: 0 (wq_completion)bond16 (work_completion)(&(&bond->mcast_work)->work) &obj_hash[i].lock irq_context: 0 (wq_completion)bond16 (work_completion)(&(&bond->mcast_work)->work) &base->lock irq_context: 0 (wq_completion)bond16 (work_completion)(&(&bond->mcast_work)->work) &base->lock &obj_hash[i].lock irq_context: 0 (wq_completion)mld (work_completion)(&(&idev->mc_ifc_work)->work) &idev->mc_lock rcu_read_lock rcu_read_lock rcu_read_lock_bh rcu_read_lock rcu_read_lock_bh rcu_read_lock &meta->lock irq_context: 0 (wq_completion)mld (work_completion)(&(&idev->mc_ifc_work)->work) &idev->mc_lock rcu_read_lock rcu_read_lock rcu_read_lock_bh rcu_read_lock rcu_read_lock_bh rcu_read_lock kfence_freelist_lock irq_context: 0 (wq_completion)bond16 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex fs_reclaim irq_context: 0 (wq_completion)bond16 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex fs_reclaim mmu_notifier_invalidate_range_start irq_context: 0 (wq_completion)bond16 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &c->lock irq_context: 0 (wq_completion)bond16 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex rcu_read_lock &bond->stats_lock/1 irq_context: 0 (wq_completion)bond16 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex rcu_read_lock &ndev->lock irq_context: 0 (wq_completion)bond16 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex nl_table_lock irq_context: 0 (wq_completion)bond16 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &obj_hash[i].lock irq_context: 0 (wq_completion)bond16 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex nl_table_wait.lock irq_context: 0 (wq_completion)bond16 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex net_rwsem irq_context: 0 (wq_completion)bond16 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex net_rwsem &list->lock#2 irq_context: 0 (wq_completion)bond16 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &tn->lock irq_context: 0 (wq_completion)bond16 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &idev->mc_lock irq_context: 0 (wq_completion)bond16 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &idev->mc_lock fs_reclaim irq_context: 0 (wq_completion)bond16 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &idev->mc_lock fs_reclaim mmu_notifier_invalidate_range_start irq_context: 0 (wq_completion)bond16 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &idev->mc_lock rcu_read_lock &ndev->lock irq_context: 0 (wq_completion)bond16 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &idev->mc_lock rcu_read_lock &dir->lock#2 irq_context: 0 (wq_completion)bond16 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &idev->mc_lock rcu_read_lock &ul->lock#2 irq_context: 0 (wq_completion)bond16 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &idev->mc_lock rcu_read_lock rcu_read_lock &____s->seqcount#10 irq_context: 0 (wq_completion)bond16 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &idev->mc_lock rcu_read_lock rcu_read_lock rcu_read_lock_bh rcu_read_lock rcu_read_lock_bh rcu_read_lock &br->multicast_lock irq_context: 0 (wq_completion)bond16 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &idev->mc_lock rcu_read_lock rcu_read_lock rcu_read_lock_bh rcu_read_lock rcu_read_lock_bh rcu_read_lock &br->multicast_lock &obj_hash[i].lock irq_context: 0 (wq_completion)bond16 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &idev->mc_lock rcu_read_lock rcu_read_lock rcu_read_lock_bh rcu_read_lock rcu_read_lock_bh rcu_read_lock &br->multicast_lock &base->lock irq_context: 0 (wq_completion)bond16 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &idev->mc_lock rcu_read_lock rcu_read_lock rcu_read_lock_bh rcu_read_lock rcu_read_lock_bh rcu_read_lock &data->lock irq_context: 0 (wq_completion)bond16 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &idev->mc_lock rcu_read_lock rcu_read_lock rcu_read_lock_bh rcu_read_lock rcu_read_lock_bh rcu_read_lock &obj_hash[i].lock irq_context: 0 (wq_completion)bond16 (work_completion)(&(&slave->notify_work)->work) rtnl_mutex net_rwsem irq_context: 0 (wq_completion)bond16 (work_completion)(&(&slave->notify_work)->work) rtnl_mutex net_rwsem &list->lock#2 irq_context: 0 (wq_completion)bond16 (work_completion)(&(&slave->notify_work)->work) rtnl_mutex &tn->lock irq_context: 0 rtnl_mutex &xa->xa_lock#3 rcu_read_lock pool_lock#2 irq_context: 0 rtnl_mutex &xa->xa_lock#3 &obj_hash[i].lock irq_context: 0 (wq_completion)bond17 irq_context: 0 (wq_completion)bond17 (work_completion)(&(&slave->notify_work)->work) irq_context: 0 (wq_completion)bond17 (work_completion)(&(&slave->notify_work)->work) &obj_hash[i].lock irq_context: 0 (wq_completion)bond17 (work_completion)(&(&slave->notify_work)->work) &base->lock irq_context: 0 (wq_completion)bond17 (work_completion)(&(&slave->notify_work)->work) &base->lock &obj_hash[i].lock irq_context: 0 (wq_completion)bond17 (work_completion)(&(&bond->mcast_work)->work) irq_context: 0 (wq_completion)bond17 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex fs_reclaim irq_context: 0 (wq_completion)bond17 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex fs_reclaim mmu_notifier_invalidate_range_start irq_context: 0 (wq_completion)bond17 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &c->lock irq_context: 0 (wq_completion)bond17 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex rcu_read_lock &bond->stats_lock/1 irq_context: 0 (wq_completion)bond17 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex rcu_read_lock &ndev->lock irq_context: 0 (wq_completion)bond17 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex nl_table_lock irq_context: 0 (wq_completion)bond17 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &obj_hash[i].lock irq_context: 0 (wq_completion)bond17 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex nl_table_wait.lock irq_context: 0 (wq_completion)bond17 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex net_rwsem irq_context: 0 (wq_completion)bond17 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex net_rwsem &list->lock#2 irq_context: 0 (wq_completion)bond17 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &tn->lock irq_context: 0 (wq_completion)bond17 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &idev->mc_lock irq_context: 0 (wq_completion)bond17 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &idev->mc_lock fs_reclaim irq_context: 0 (wq_completion)bond17 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &idev->mc_lock fs_reclaim mmu_notifier_invalidate_range_start irq_context: 0 (wq_completion)bond17 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &idev->mc_lock &c->lock irq_context: 0 (wq_completion)bond17 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &idev->mc_lock &n->list_lock irq_context: 0 (wq_completion)bond17 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &idev->mc_lock &n->list_lock &c->lock irq_context: 0 (wq_completion)bond17 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &idev->mc_lock rcu_read_lock &ndev->lock irq_context: 0 (wq_completion)bond17 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &idev->mc_lock rcu_read_lock &c->lock irq_context: 0 (wq_completion)bond17 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &idev->mc_lock rcu_read_lock &dir->lock#2 irq_context: 0 (wq_completion)bond17 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &idev->mc_lock rcu_read_lock &ul->lock#2 irq_context: 0 (wq_completion)bond17 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &idev->mc_lock rcu_read_lock rcu_read_lock &obj_hash[i].lock irq_context: 0 (wq_completion)bond17 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &idev->mc_lock rcu_read_lock rcu_read_lock &dir->lock#2 irq_context: 0 (wq_completion)bond17 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &idev->mc_lock rcu_read_lock rcu_read_lock &tbl->lock irq_context: 0 (wq_completion)bond17 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &idev->mc_lock rcu_read_lock rcu_read_lock &n->lock irq_context: 0 (wq_completion)bond17 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &idev->mc_lock rcu_read_lock rcu_read_lock &____s->seqcount#9 irq_context: 0 (wq_completion)bond17 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &idev->mc_lock rcu_read_lock rcu_read_lock rcu_read_lock_bh rcu_read_lock rcu_read_lock_bh rcu_read_lock &br->multicast_lock irq_context: 0 (wq_completion)bond17 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &idev->mc_lock rcu_read_lock rcu_read_lock rcu_read_lock_bh rcu_read_lock rcu_read_lock_bh rcu_read_lock &br->multicast_lock &c->lock irq_context: 0 (wq_completion)bond17 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &idev->mc_lock rcu_read_lock rcu_read_lock rcu_read_lock_bh rcu_read_lock rcu_read_lock_bh rcu_read_lock &br->multicast_lock &obj_hash[i].lock irq_context: 0 (wq_completion)bond17 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &idev->mc_lock rcu_read_lock rcu_read_lock rcu_read_lock_bh rcu_read_lock rcu_read_lock_bh rcu_read_lock &br->multicast_lock rcu_read_lock rhashtable_bucket irq_context: 0 (wq_completion)bond17 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &idev->mc_lock rcu_read_lock rcu_read_lock rcu_read_lock_bh rcu_read_lock rcu_read_lock_bh rcu_read_lock &br->multicast_lock nl_table_lock irq_context: 0 (wq_completion)bond17 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &idev->mc_lock rcu_read_lock rcu_read_lock rcu_read_lock_bh rcu_read_lock rcu_read_lock_bh rcu_read_lock &br->multicast_lock nl_table_wait.lock irq_context: 0 (wq_completion)bond17 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &idev->mc_lock rcu_read_lock rcu_read_lock rcu_read_lock_bh rcu_read_lock rcu_read_lock_bh rcu_read_lock &br->multicast_lock &base->lock irq_context: 0 (wq_completion)bond17 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &idev->mc_lock rcu_read_lock rcu_read_lock rcu_read_lock_bh rcu_read_lock rcu_read_lock_bh rcu_read_lock &br->multicast_lock &base->lock &obj_hash[i].lock irq_context: 0 (wq_completion)bond17 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &idev->mc_lock rcu_read_lock rcu_read_lock rcu_read_lock_bh rcu_read_lock rcu_read_lock_bh rcu_read_lock &data->lock irq_context: 0 (wq_completion)bond17 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &idev->mc_lock rcu_read_lock rcu_read_lock rcu_read_lock_bh rcu_read_lock rcu_read_lock_bh rcu_read_lock &obj_hash[i].lock irq_context: 0 (wq_completion)bond17 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &idev->mc_lock rcu_read_lock rcu_read_lock rcu_read_lock_bh rcu_read_lock rcu_read_lock_bh rcu_read_lock &obj_hash[i].lock pool_lock irq_context: 0 (wq_completion)bond17 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex.wait_lock irq_context: 0 (wq_completion)bond17 (work_completion)(&(&bond->mcast_work)->work) &p->pi_lock irq_context: 0 (wq_completion)bond17 (work_completion)(&(&bond->mcast_work)->work) &p->pi_lock &rq->__lock irq_context: 0 (wq_completion)bond17 (work_completion)(&(&bond->mcast_work)->work) &p->pi_lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 rtnl_mutex rcu_read_lock &p->alloc_lock irq_context: 0 (wq_completion)bond17 (work_completion)(&(&slave->notify_work)->work) rtnl_mutex net_rwsem irq_context: 0 (wq_completion)bond17 (work_completion)(&(&slave->notify_work)->work) rtnl_mutex net_rwsem &list->lock#2 irq_context: 0 (wq_completion)bond17 (work_completion)(&(&slave->notify_work)->work) rtnl_mutex &tn->lock irq_context: 0 &mm->mmap_lock key#8 irq_context: 0 (wq_completion)bond18 irq_context: 0 (wq_completion)bond18 (work_completion)(&(&slave->notify_work)->work) irq_context: 0 (wq_completion)bond18 (work_completion)(&(&slave->notify_work)->work) &obj_hash[i].lock irq_context: 0 (wq_completion)bond18 (work_completion)(&(&slave->notify_work)->work) &base->lock irq_context: 0 (wq_completion)bond18 (work_completion)(&(&slave->notify_work)->work) &base->lock &obj_hash[i].lock irq_context: 0 (wq_completion)bond18 (work_completion)(&(&bond->mcast_work)->work) irq_context: 0 (wq_completion)bond18 (work_completion)(&(&bond->mcast_work)->work) &obj_hash[i].lock irq_context: 0 (wq_completion)bond18 (work_completion)(&(&bond->mcast_work)->work) &base->lock irq_context: 0 (wq_completion)bond18 (work_completion)(&(&bond->mcast_work)->work) &base->lock &obj_hash[i].lock irq_context: softirq (&ndev->rs_timer) rcu_read_lock rcu_read_lock rcu_read_lock_bh rcu_read_lock rcu_read_lock_bh rcu_read_lock &data->lock irq_context: 0 (wq_completion)bond18 (work_completion)(&(&slave->notify_work)->work) rtnl_mutex net_rwsem irq_context: 0 (wq_completion)bond18 (work_completion)(&(&slave->notify_work)->work) rtnl_mutex net_rwsem &list->lock#2 irq_context: 0 (wq_completion)bond18 (work_completion)(&(&slave->notify_work)->work) rtnl_mutex &tn->lock irq_context: 0 (wq_completion)bond18 (work_completion)(&(&slave->notify_work)->work) rtnl_mutex.wait_lock irq_context: 0 (wq_completion)bond18 (work_completion)(&(&slave->notify_work)->work) &p->pi_lock irq_context: 0 (wq_completion)bond18 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex fs_reclaim irq_context: 0 (wq_completion)bond18 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex fs_reclaim mmu_notifier_invalidate_range_start irq_context: 0 (wq_completion)bond18 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &c->lock irq_context: 0 (wq_completion)bond18 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex rcu_read_lock &bond->stats_lock/1 irq_context: 0 (wq_completion)bond18 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex rcu_read_lock &ndev->lock irq_context: 0 (wq_completion)bond18 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex nl_table_lock irq_context: 0 (wq_completion)bond18 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &obj_hash[i].lock irq_context: 0 (wq_completion)bond18 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex nl_table_wait.lock irq_context: 0 (wq_completion)bond18 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex net_rwsem irq_context: 0 (wq_completion)bond18 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex net_rwsem &list->lock#2 irq_context: 0 (wq_completion)bond18 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &tn->lock irq_context: 0 (wq_completion)bond18 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &idev->mc_lock irq_context: 0 (wq_completion)bond18 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex.wait_lock irq_context: 0 (wq_completion)bond18 (work_completion)(&(&bond->mcast_work)->work) &p->pi_lock irq_context: 0 (wq_completion)bond19 irq_context: 0 (wq_completion)bond19 (work_completion)(&(&slave->notify_work)->work) irq_context: 0 (wq_completion)bond19 (work_completion)(&(&slave->notify_work)->work) &obj_hash[i].lock irq_context: 0 (wq_completion)bond19 (work_completion)(&(&slave->notify_work)->work) &base->lock irq_context: 0 (wq_completion)bond19 (work_completion)(&(&slave->notify_work)->work) &base->lock &obj_hash[i].lock irq_context: 0 (wq_completion)bond19 (work_completion)(&(&bond->mcast_work)->work) irq_context: 0 (wq_completion)bond19 (work_completion)(&(&bond->mcast_work)->work) &obj_hash[i].lock irq_context: 0 (wq_completion)bond19 (work_completion)(&(&bond->mcast_work)->work) &base->lock irq_context: 0 (wq_completion)bond19 (work_completion)(&(&bond->mcast_work)->work) &base->lock &obj_hash[i].lock irq_context: 0 (wq_completion)events (linkwatch_work).work rtnl_mutex rcu_read_lock &net->sctp.local_addr_lock &net->sctp.addr_wq_lock &n->list_lock irq_context: 0 (wq_completion)events (linkwatch_work).work rtnl_mutex rcu_read_lock &net->sctp.local_addr_lock &net->sctp.addr_wq_lock &n->list_lock &c->lock irq_context: 0 (wq_completion)bond19 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex fs_reclaim irq_context: 0 (wq_completion)bond19 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex fs_reclaim mmu_notifier_invalidate_range_start irq_context: 0 (wq_completion)bond19 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &c->lock irq_context: 0 (wq_completion)bond19 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &____s->seqcount#2 irq_context: 0 (wq_completion)bond19 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &____s->seqcount irq_context: 0 (wq_completion)bond19 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex rcu_read_lock &bond->stats_lock/1 irq_context: 0 (wq_completion)bond19 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex rcu_read_lock &ndev->lock irq_context: 0 (wq_completion)bond19 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex nl_table_lock irq_context: 0 (wq_completion)bond19 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &obj_hash[i].lock irq_context: 0 (wq_completion)bond19 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex nl_table_wait.lock irq_context: 0 (wq_completion)bond19 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex net_rwsem irq_context: 0 (wq_completion)bond19 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex net_rwsem &list->lock#2 irq_context: 0 (wq_completion)bond19 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &tn->lock irq_context: 0 (wq_completion)bond19 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &idev->mc_lock irq_context: 0 (wq_completion)bond19 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &idev->mc_lock fs_reclaim irq_context: 0 (wq_completion)bond19 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &idev->mc_lock fs_reclaim mmu_notifier_invalidate_range_start irq_context: 0 (wq_completion)bond19 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &idev->mc_lock rcu_read_lock &ndev->lock irq_context: 0 (wq_completion)bond19 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &idev->mc_lock rcu_read_lock &dir->lock#2 irq_context: 0 (wq_completion)bond19 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &idev->mc_lock rcu_read_lock &ul->lock#2 irq_context: 0 (wq_completion)bond19 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &idev->mc_lock rcu_read_lock rcu_read_lock &obj_hash[i].lock irq_context: 0 (wq_completion)bond19 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &idev->mc_lock rcu_read_lock rcu_read_lock &dir->lock#2 irq_context: 0 (wq_completion)bond19 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &idev->mc_lock rcu_read_lock rcu_read_lock &tbl->lock irq_context: 0 (wq_completion)bond19 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &idev->mc_lock rcu_read_lock rcu_read_lock &n->lock irq_context: 0 (wq_completion)bond19 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &idev->mc_lock rcu_read_lock rcu_read_lock &____s->seqcount#9 irq_context: 0 (wq_completion)bond19 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &idev->mc_lock rcu_read_lock rcu_read_lock rcu_read_lock_bh rcu_read_lock rcu_read_lock_bh rcu_read_lock &br->multicast_lock irq_context: 0 (wq_completion)bond19 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &idev->mc_lock rcu_read_lock rcu_read_lock rcu_read_lock_bh rcu_read_lock rcu_read_lock_bh rcu_read_lock &br->multicast_lock &c->lock irq_context: 0 (wq_completion)bond19 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &idev->mc_lock rcu_read_lock rcu_read_lock rcu_read_lock_bh rcu_read_lock rcu_read_lock_bh rcu_read_lock &br->multicast_lock &obj_hash[i].lock irq_context: 0 (wq_completion)bond19 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &idev->mc_lock rcu_read_lock rcu_read_lock rcu_read_lock_bh rcu_read_lock rcu_read_lock_bh rcu_read_lock &br->multicast_lock rcu_read_lock rhashtable_bucket irq_context: 0 (wq_completion)bond19 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &idev->mc_lock rcu_read_lock rcu_read_lock rcu_read_lock_bh rcu_read_lock rcu_read_lock_bh rcu_read_lock &br->multicast_lock nl_table_lock irq_context: 0 (wq_completion)bond19 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &idev->mc_lock rcu_read_lock rcu_read_lock rcu_read_lock_bh rcu_read_lock rcu_read_lock_bh rcu_read_lock &br->multicast_lock nl_table_wait.lock irq_context: 0 (wq_completion)bond19 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &idev->mc_lock rcu_read_lock rcu_read_lock rcu_read_lock_bh rcu_read_lock rcu_read_lock_bh rcu_read_lock &br->multicast_lock &base->lock irq_context: 0 (wq_completion)bond19 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &idev->mc_lock rcu_read_lock rcu_read_lock rcu_read_lock_bh rcu_read_lock rcu_read_lock_bh rcu_read_lock &br->multicast_lock &base->lock &obj_hash[i].lock irq_context: 0 (wq_completion)bond19 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &idev->mc_lock rcu_read_lock rcu_read_lock rcu_read_lock_bh rcu_read_lock rcu_read_lock_bh rcu_read_lock &data->lock irq_context: 0 (wq_completion)bond19 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &idev->mc_lock rcu_read_lock rcu_read_lock rcu_read_lock_bh rcu_read_lock rcu_read_lock_bh rcu_read_lock &obj_hash[i].lock irq_context: 0 (wq_completion)bond19 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex.wait_lock irq_context: 0 (wq_completion)bond19 (work_completion)(&(&bond->mcast_work)->work) &p->pi_lock irq_context: 0 (wq_completion)bond19 (work_completion)(&(&slave->notify_work)->work) rtnl_mutex net_rwsem irq_context: 0 (wq_completion)bond19 (work_completion)(&(&slave->notify_work)->work) rtnl_mutex net_rwsem &list->lock#2 irq_context: 0 (wq_completion)bond19 (work_completion)(&(&slave->notify_work)->work) rtnl_mutex &tn->lock irq_context: 0 (wq_completion)ipv6_addrconf (work_completion)(&(&ifa->dad_work)->work) rtnl_mutex rcu_read_lock rcu_read_lock rcu_read_lock_bh rcu_read_lock rcu_read_lock_bh rcu_read_lock &meta->lock irq_context: 0 (wq_completion)ipv6_addrconf (work_completion)(&(&ifa->dad_work)->work) rtnl_mutex rcu_read_lock rcu_read_lock rcu_read_lock_bh rcu_read_lock rcu_read_lock_bh rcu_read_lock kfence_freelist_lock irq_context: 0 (wq_completion)wg-crypt-wg1#2 (work_completion)(&peer->transmit_packet_work) &peer->endpoint_lock rcu_read_lock_bh rcu_read_lock &n->list_lock irq_context: 0 (wq_completion)wg-crypt-wg1#2 (work_completion)(&peer->transmit_packet_work) &peer->endpoint_lock rcu_read_lock_bh rcu_read_lock &n->list_lock &c->lock irq_context: 0 pernet_ops_rwsem rtnl_mutex &root->kernfs_rwsem &rcu_state.expedited_wq irq_context: 0 pernet_ops_rwsem rtnl_mutex &root->kernfs_rwsem &rcu_state.expedited_wq &p->pi_lock irq_context: 0 pernet_ops_rwsem rtnl_mutex &root->kernfs_rwsem &rcu_state.expedited_wq &p->pi_lock &rq->__lock irq_context: 0 pernet_ops_rwsem rtnl_mutex &root->kernfs_rwsem &rcu_state.expedited_wq &p->pi_lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 &mm->mmap_lock fill_pool_map-wait-type-override &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 pernet_ops_rwsem &ht->mutex &pcp->lock &zone->lock irq_context: 0 pernet_ops_rwsem &ht->mutex &pcp->lock &zone->lock &____s->seqcount irq_context: 0 (wq_completion)bond10 (work_completion)(&(&bond->ad_work)->work) &bond->mode_lock rcu_read_lock &n->list_lock irq_context: 0 (wq_completion)bond10 (work_completion)(&(&bond->ad_work)->work) &bond->mode_lock rcu_read_lock &n->list_lock &c->lock irq_context: 0 (wq_completion)mld (work_completion)(&(&idev->mc_dad_work)->work) &idev->mc_lock rcu_read_lock rcu_read_lock rcu_read_lock_bh rcu_read_lock rcu_read_lock_bh rcu_read_lock &meta->lock irq_context: 0 (wq_completion)mld (work_completion)(&(&idev->mc_dad_work)->work) &idev->mc_lock rcu_read_lock rcu_read_lock rcu_read_lock_bh rcu_read_lock rcu_read_lock_bh rcu_read_lock kfence_freelist_lock irq_context: 0 (wq_completion)bond20 irq_context: 0 (wq_completion)bond20 (work_completion)(&(&slave->notify_work)->work) irq_context: 0 (wq_completion)bond20 (work_completion)(&(&slave->notify_work)->work) &obj_hash[i].lock irq_context: 0 (wq_completion)bond20 (work_completion)(&(&slave->notify_work)->work) &base->lock irq_context: 0 (wq_completion)bond20 (work_completion)(&(&slave->notify_work)->work) &base->lock &obj_hash[i].lock irq_context: 0 (wq_completion)bond20 (work_completion)(&(&slave->notify_work)->work) &rq->__lock irq_context: 0 (wq_completion)bond20 (work_completion)(&(&bond->mcast_work)->work) irq_context: 0 (wq_completion)bond20 (work_completion)(&(&bond->mcast_work)->work) &obj_hash[i].lock irq_context: 0 (wq_completion)bond20 (work_completion)(&(&bond->mcast_work)->work) &base->lock irq_context: 0 (wq_completion)bond20 (work_completion)(&(&bond->mcast_work)->work) &base->lock &obj_hash[i].lock irq_context: 0 (wq_completion)bond20 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex fs_reclaim irq_context: 0 (wq_completion)bond20 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex fs_reclaim mmu_notifier_invalidate_range_start irq_context: 0 (wq_completion)bond20 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex remove_cache_srcu irq_context: 0 (wq_completion)bond20 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex remove_cache_srcu quarantine_lock irq_context: 0 (wq_completion)bond20 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex remove_cache_srcu &c->lock irq_context: 0 (wq_completion)bond20 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex remove_cache_srcu &n->list_lock irq_context: 0 (wq_completion)bond20 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex remove_cache_srcu rcu_read_lock pool_lock#2 irq_context: 0 (wq_completion)bond20 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex remove_cache_srcu &obj_hash[i].lock irq_context: 0 (wq_completion)bond20 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex remove_cache_srcu pool_lock#2 irq_context: 0 (wq_completion)bond20 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex rcu_read_lock &bond->stats_lock/1 irq_context: 0 (wq_completion)bond20 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex rcu_read_lock &ndev->lock irq_context: 0 (wq_completion)bond20 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex nl_table_lock irq_context: 0 (wq_completion)bond20 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &obj_hash[i].lock irq_context: 0 (wq_completion)bond20 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex nl_table_wait.lock irq_context: 0 (wq_completion)bond20 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex net_rwsem irq_context: 0 (wq_completion)bond20 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex net_rwsem &list->lock#2 irq_context: 0 (wq_completion)bond20 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &tn->lock irq_context: 0 (wq_completion)bond20 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &idev->mc_lock irq_context: 0 (wq_completion)bond20 (work_completion)(&(&slave->notify_work)->work) rtnl_mutex net_rwsem irq_context: 0 (wq_completion)bond20 (work_completion)(&(&slave->notify_work)->work) rtnl_mutex net_rwsem &list->lock#2 irq_context: 0 (wq_completion)bond20 (work_completion)(&(&slave->notify_work)->work) rtnl_mutex &tn->lock irq_context: 0 (wq_completion)mld (work_completion)(&(&idev->mc_ifc_work)->work) &idev->mc_lock rcu_read_lock rcu_read_lock rcu_read_lock_bh rcu_read_lock rcu_read_lock_bh rcu_read_lock &br->multicast_lock &c->lock irq_context: 0 (wq_completion)bond21 irq_context: 0 (wq_completion)bond21 (work_completion)(&(&slave->notify_work)->work) irq_context: 0 (wq_completion)bond21 (work_completion)(&(&slave->notify_work)->work) &obj_hash[i].lock irq_context: 0 (wq_completion)bond21 (work_completion)(&(&slave->notify_work)->work) &base->lock irq_context: 0 (wq_completion)bond21 (work_completion)(&(&slave->notify_work)->work) &base->lock &obj_hash[i].lock irq_context: 0 (wq_completion)bond21 (work_completion)(&(&slave->notify_work)->work) &rq->__lock irq_context: 0 (wq_completion)bond21 (work_completion)(&(&slave->notify_work)->work) &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)bond21 (work_completion)(&(&bond->mcast_work)->work) irq_context: 0 (wq_completion)bond21 (work_completion)(&(&bond->mcast_work)->work) &obj_hash[i].lock irq_context: 0 (wq_completion)bond21 (work_completion)(&(&bond->mcast_work)->work) &base->lock irq_context: 0 (wq_completion)bond21 (work_completion)(&(&bond->mcast_work)->work) &base->lock &obj_hash[i].lock irq_context: 0 (wq_completion)bond21 (work_completion)(&(&slave->notify_work)->work) rtnl_mutex net_rwsem irq_context: 0 (wq_completion)bond21 (work_completion)(&(&slave->notify_work)->work) rtnl_mutex net_rwsem &list->lock#2 irq_context: 0 (wq_completion)bond21 (work_completion)(&(&slave->notify_work)->work) rtnl_mutex &tn->lock irq_context: 0 (wq_completion)bond21 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex fs_reclaim irq_context: 0 (wq_completion)bond21 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex fs_reclaim mmu_notifier_invalidate_range_start irq_context: 0 (wq_completion)bond21 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &c->lock irq_context: 0 (wq_completion)bond21 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex rcu_read_lock &bond->stats_lock/1 irq_context: 0 (wq_completion)bond21 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex rcu_read_lock &ndev->lock irq_context: 0 (wq_completion)bond21 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex nl_table_lock irq_context: 0 (wq_completion)bond21 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &obj_hash[i].lock irq_context: 0 (wq_completion)bond21 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex nl_table_wait.lock irq_context: 0 (wq_completion)bond21 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex net_rwsem irq_context: 0 (wq_completion)bond21 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex net_rwsem &list->lock#2 irq_context: 0 (wq_completion)bond21 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &tn->lock irq_context: 0 (wq_completion)bond21 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &idev->mc_lock irq_context: 0 (wq_completion)bond21 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &idev->mc_lock fs_reclaim irq_context: 0 (wq_completion)bond21 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &idev->mc_lock fs_reclaim mmu_notifier_invalidate_range_start irq_context: 0 (wq_completion)bond21 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &idev->mc_lock &c->lock irq_context: 0 (wq_completion)bond21 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &idev->mc_lock rcu_read_lock &ndev->lock irq_context: 0 (wq_completion)bond21 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &idev->mc_lock rcu_read_lock &dir->lock#2 irq_context: 0 (wq_completion)bond21 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &idev->mc_lock rcu_read_lock &ul->lock#2 irq_context: 0 (wq_completion)bond21 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &idev->mc_lock rcu_read_lock rcu_read_lock &____s->seqcount#10 irq_context: 0 (wq_completion)bond21 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &idev->mc_lock rcu_read_lock rcu_read_lock rcu_read_lock_bh rcu_read_lock rcu_read_lock_bh rcu_read_lock &br->multicast_lock irq_context: 0 (wq_completion)bond21 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &idev->mc_lock rcu_read_lock rcu_read_lock rcu_read_lock_bh rcu_read_lock rcu_read_lock_bh rcu_read_lock &br->multicast_lock &obj_hash[i].lock irq_context: 0 (wq_completion)bond21 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &idev->mc_lock rcu_read_lock rcu_read_lock rcu_read_lock_bh rcu_read_lock rcu_read_lock_bh rcu_read_lock &data->lock irq_context: 0 (wq_completion)bond21 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &idev->mc_lock rcu_read_lock rcu_read_lock rcu_read_lock_bh rcu_read_lock rcu_read_lock_bh rcu_read_lock &obj_hash[i].lock irq_context: 0 (wq_completion)mld (work_completion)(&(&idev->mc_dad_work)->work) &idev->mc_lock rcu_read_lock rcu_read_lock rcu_read_lock_bh rcu_read_lock rcu_read_lock_bh rcu_read_lock &data->lock &obj_hash[i].lock irq_context: 0 (wq_completion)mld (work_completion)(&(&idev->mc_dad_work)->work) &idev->mc_lock rcu_read_lock rcu_read_lock rcu_read_lock_bh rcu_read_lock rcu_read_lock_bh rcu_read_lock &data->lock &base->lock irq_context: 0 (wq_completion)mld (work_completion)(&(&idev->mc_dad_work)->work) &idev->mc_lock rcu_read_lock rcu_read_lock rcu_read_lock_bh rcu_read_lock rcu_read_lock_bh rcu_read_lock &data->lock &base->lock &obj_hash[i].lock irq_context: 0 (wq_completion)bond22 irq_context: 0 (wq_completion)bond22 (work_completion)(&(&slave->notify_work)->work) irq_context: 0 (wq_completion)bond22 (work_completion)(&(&slave->notify_work)->work) &obj_hash[i].lock irq_context: 0 (wq_completion)bond22 (work_completion)(&(&slave->notify_work)->work) &base->lock irq_context: 0 (wq_completion)bond22 (work_completion)(&(&slave->notify_work)->work) &base->lock &obj_hash[i].lock irq_context: 0 (wq_completion)bond22 (work_completion)(&(&bond->mcast_work)->work) irq_context: 0 (wq_completion)bond22 (work_completion)(&(&bond->mcast_work)->work) &obj_hash[i].lock irq_context: 0 (wq_completion)bond22 (work_completion)(&(&bond->mcast_work)->work) &base->lock irq_context: 0 (wq_completion)bond22 (work_completion)(&(&bond->mcast_work)->work) &base->lock &obj_hash[i].lock irq_context: 0 (wq_completion)bond22 (work_completion)(&(&bond->mcast_work)->work) &rq->__lock irq_context: 0 (wq_completion)bond22 (work_completion)(&(&bond->mcast_work)->work) &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)bond22 (work_completion)(&(&slave->notify_work)->work) rtnl_mutex net_rwsem irq_context: 0 (wq_completion)bond22 (work_completion)(&(&slave->notify_work)->work) rtnl_mutex net_rwsem &list->lock#2 irq_context: 0 (wq_completion)bond22 (work_completion)(&(&slave->notify_work)->work) rtnl_mutex &tn->lock irq_context: 0 (wq_completion)bond22 (work_completion)(&(&slave->notify_work)->work) rtnl_mutex.wait_lock irq_context: 0 (wq_completion)bond22 (work_completion)(&(&slave->notify_work)->work) &p->pi_lock irq_context: 0 (wq_completion)bond22 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex fs_reclaim irq_context: 0 (wq_completion)bond22 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex fs_reclaim mmu_notifier_invalidate_range_start irq_context: 0 (wq_completion)bond22 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &c->lock irq_context: 0 (wq_completion)bond22 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex rcu_read_lock &bond->stats_lock/1 irq_context: 0 (wq_completion)bond22 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex rcu_read_lock &ndev->lock irq_context: 0 (wq_completion)bond22 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex nl_table_lock irq_context: 0 (wq_completion)bond22 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &obj_hash[i].lock irq_context: 0 (wq_completion)bond22 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex nl_table_wait.lock irq_context: 0 (wq_completion)bond22 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex net_rwsem irq_context: 0 (wq_completion)bond22 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex net_rwsem &list->lock#2 irq_context: 0 (wq_completion)bond22 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &tn->lock irq_context: 0 (wq_completion)bond22 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &idev->mc_lock irq_context: 0 (wq_completion)bond22 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex.wait_lock irq_context: 0 (wq_completion)bond22 (work_completion)(&(&bond->mcast_work)->work) &p->pi_lock irq_context: 0 (wq_completion)bond23 irq_context: 0 (wq_completion)bond23 (work_completion)(&(&slave->notify_work)->work) irq_context: 0 (wq_completion)bond23 (work_completion)(&(&slave->notify_work)->work) &obj_hash[i].lock irq_context: 0 (wq_completion)bond23 (work_completion)(&(&slave->notify_work)->work) &base->lock irq_context: 0 (wq_completion)bond23 (work_completion)(&(&slave->notify_work)->work) &base->lock &obj_hash[i].lock irq_context: 0 (wq_completion)bond23 (work_completion)(&(&bond->mcast_work)->work) irq_context: 0 (wq_completion)bond23 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex fs_reclaim irq_context: 0 (wq_completion)bond23 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex fs_reclaim mmu_notifier_invalidate_range_start irq_context: 0 (wq_completion)bond23 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex rcu_read_lock &bond->stats_lock/1 irq_context: 0 (wq_completion)bond23 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex rcu_read_lock &ndev->lock irq_context: 0 (wq_completion)bond23 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex nl_table_lock irq_context: 0 (wq_completion)bond23 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &obj_hash[i].lock irq_context: 0 (wq_completion)bond23 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex nl_table_wait.lock irq_context: 0 (wq_completion)bond23 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex net_rwsem irq_context: 0 (wq_completion)bond23 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex net_rwsem &list->lock#2 irq_context: 0 (wq_completion)bond23 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &tn->lock irq_context: 0 (wq_completion)bond23 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &idev->mc_lock irq_context: 0 (wq_completion)bond23 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &idev->mc_lock fs_reclaim irq_context: 0 (wq_completion)bond23 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &idev->mc_lock fs_reclaim mmu_notifier_invalidate_range_start irq_context: 0 (wq_completion)bond23 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &idev->mc_lock &c->lock irq_context: 0 (wq_completion)bond23 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &idev->mc_lock rcu_read_lock &ndev->lock irq_context: 0 (wq_completion)bond23 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &idev->mc_lock rcu_read_lock &dir->lock#2 irq_context: 0 (wq_completion)bond23 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &idev->mc_lock rcu_read_lock &ul->lock#2 irq_context: 0 (wq_completion)bond23 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &idev->mc_lock rcu_read_lock rcu_read_lock &c->lock irq_context: 0 (wq_completion)bond23 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &idev->mc_lock rcu_read_lock rcu_read_lock &obj_hash[i].lock irq_context: 0 (wq_completion)bond23 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &idev->mc_lock rcu_read_lock rcu_read_lock &dir->lock#2 irq_context: 0 (wq_completion)bond23 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &idev->mc_lock rcu_read_lock rcu_read_lock &tbl->lock irq_context: 0 (wq_completion)bond23 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &idev->mc_lock rcu_read_lock rcu_read_lock &n->lock irq_context: 0 (wq_completion)bond23 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &idev->mc_lock rcu_read_lock rcu_read_lock &____s->seqcount#9 irq_context: 0 (wq_completion)bond23 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &idev->mc_lock rcu_read_lock rcu_read_lock rcu_read_lock_bh rcu_read_lock rcu_read_lock_bh rcu_read_lock &br->multicast_lock irq_context: 0 (wq_completion)bond23 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &idev->mc_lock rcu_read_lock rcu_read_lock rcu_read_lock_bh rcu_read_lock rcu_read_lock_bh rcu_read_lock &br->multicast_lock &c->lock irq_context: 0 (wq_completion)bond23 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &idev->mc_lock rcu_read_lock rcu_read_lock rcu_read_lock_bh rcu_read_lock rcu_read_lock_bh rcu_read_lock &br->multicast_lock &obj_hash[i].lock irq_context: 0 (wq_completion)bond23 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &idev->mc_lock rcu_read_lock rcu_read_lock rcu_read_lock_bh rcu_read_lock rcu_read_lock_bh rcu_read_lock &br->multicast_lock rcu_read_lock rhashtable_bucket irq_context: 0 (wq_completion)bond23 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &idev->mc_lock rcu_read_lock rcu_read_lock rcu_read_lock_bh rcu_read_lock rcu_read_lock_bh rcu_read_lock &br->multicast_lock nl_table_lock irq_context: 0 (wq_completion)bond23 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &idev->mc_lock rcu_read_lock rcu_read_lock rcu_read_lock_bh rcu_read_lock rcu_read_lock_bh rcu_read_lock &br->multicast_lock nl_table_wait.lock irq_context: 0 (wq_completion)bond23 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &idev->mc_lock rcu_read_lock rcu_read_lock rcu_read_lock_bh rcu_read_lock rcu_read_lock_bh rcu_read_lock &br->multicast_lock &base->lock irq_context: 0 (wq_completion)bond23 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &idev->mc_lock rcu_read_lock rcu_read_lock rcu_read_lock_bh rcu_read_lock rcu_read_lock_bh rcu_read_lock &br->multicast_lock &base->lock &obj_hash[i].lock irq_context: 0 (wq_completion)bond23 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &idev->mc_lock rcu_read_lock rcu_read_lock rcu_read_lock_bh rcu_read_lock rcu_read_lock_bh rcu_read_lock &data->lock irq_context: 0 (wq_completion)bond23 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &idev->mc_lock rcu_read_lock rcu_read_lock rcu_read_lock_bh rcu_read_lock rcu_read_lock_bh rcu_read_lock &obj_hash[i].lock irq_context: 0 (wq_completion)bond23 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex.wait_lock irq_context: 0 (wq_completion)bond23 (work_completion)(&(&bond->mcast_work)->work) &p->pi_lock irq_context: 0 (wq_completion)bond23 (work_completion)(&(&slave->notify_work)->work) rtnl_mutex net_rwsem irq_context: 0 (wq_completion)bond23 (work_completion)(&(&slave->notify_work)->work) rtnl_mutex net_rwsem &list->lock#2 irq_context: 0 (wq_completion)bond23 (work_completion)(&(&slave->notify_work)->work) rtnl_mutex &tn->lock irq_context: 0 (wq_completion)bond23 (work_completion)(&(&slave->notify_work)->work) rtnl_mutex.wait_lock irq_context: 0 (wq_completion)bond23 (work_completion)(&(&slave->notify_work)->work) &p->pi_lock irq_context: 0 pernet_ops_rwsem devices_rwsem rdma_nets_rwsem &device->compat_devs_mutex fs_reclaim &obj_hash[i].lock pool_lock irq_context: 0 (wq_completion)bond24 irq_context: 0 (wq_completion)bond24 (work_completion)(&(&slave->notify_work)->work) irq_context: 0 (wq_completion)bond24 (work_completion)(&(&slave->notify_work)->work) &obj_hash[i].lock irq_context: 0 (wq_completion)bond24 (work_completion)(&(&slave->notify_work)->work) &base->lock irq_context: 0 (wq_completion)bond24 (work_completion)(&(&slave->notify_work)->work) &base->lock &obj_hash[i].lock irq_context: 0 (wq_completion)bond24 (work_completion)(&(&bond->mcast_work)->work) irq_context: 0 (wq_completion)bond24 (work_completion)(&(&bond->mcast_work)->work) &obj_hash[i].lock irq_context: 0 (wq_completion)bond24 (work_completion)(&(&bond->mcast_work)->work) &base->lock irq_context: 0 (wq_completion)bond24 (work_completion)(&(&bond->mcast_work)->work) &base->lock &obj_hash[i].lock irq_context: 0 (wq_completion)bond24 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex fs_reclaim irq_context: 0 (wq_completion)bond24 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex fs_reclaim mmu_notifier_invalidate_range_start irq_context: 0 (wq_completion)bond24 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex rcu_read_lock &bond->stats_lock/1 irq_context: 0 (wq_completion)bond24 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex rcu_read_lock &ndev->lock irq_context: 0 (wq_completion)bond24 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex nl_table_lock irq_context: 0 (wq_completion)bond24 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &obj_hash[i].lock irq_context: 0 (wq_completion)bond24 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex nl_table_wait.lock irq_context: 0 (wq_completion)bond24 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex net_rwsem irq_context: 0 (wq_completion)bond24 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex net_rwsem &list->lock#2 irq_context: 0 (wq_completion)bond24 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &tn->lock irq_context: 0 (wq_completion)bond24 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &idev->mc_lock irq_context: 0 (wq_completion)bond24 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &idev->mc_lock fs_reclaim irq_context: 0 (wq_completion)bond24 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &idev->mc_lock fs_reclaim mmu_notifier_invalidate_range_start irq_context: 0 (wq_completion)bond24 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &idev->mc_lock rcu_read_lock &ndev->lock irq_context: 0 (wq_completion)bond24 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &idev->mc_lock rcu_read_lock &c->lock irq_context: 0 (wq_completion)bond24 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &idev->mc_lock rcu_read_lock &dir->lock#2 irq_context: 0 (wq_completion)bond24 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &idev->mc_lock rcu_read_lock &ul->lock#2 irq_context: 0 (wq_completion)bond24 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &idev->mc_lock rcu_read_lock rcu_read_lock &____s->seqcount#10 irq_context: 0 (wq_completion)bond24 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &idev->mc_lock rcu_read_lock rcu_read_lock rcu_read_lock_bh rcu_read_lock rcu_read_lock_bh rcu_read_lock &br->multicast_lock irq_context: 0 (wq_completion)bond24 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &idev->mc_lock rcu_read_lock rcu_read_lock rcu_read_lock_bh rcu_read_lock rcu_read_lock_bh rcu_read_lock &br->multicast_lock &obj_hash[i].lock irq_context: 0 (wq_completion)bond24 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &idev->mc_lock rcu_read_lock rcu_read_lock rcu_read_lock_bh rcu_read_lock rcu_read_lock_bh rcu_read_lock &br->multicast_lock &base->lock irq_context: 0 (wq_completion)bond24 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &idev->mc_lock rcu_read_lock rcu_read_lock rcu_read_lock_bh rcu_read_lock rcu_read_lock_bh rcu_read_lock &data->lock irq_context: 0 (wq_completion)bond24 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &idev->mc_lock rcu_read_lock rcu_read_lock rcu_read_lock_bh rcu_read_lock rcu_read_lock_bh rcu_read_lock &obj_hash[i].lock irq_context: 0 (wq_completion)bond24 (work_completion)(&(&slave->notify_work)->work) rtnl_mutex net_rwsem irq_context: 0 (wq_completion)bond24 (work_completion)(&(&slave->notify_work)->work) rtnl_mutex net_rwsem &list->lock#2 irq_context: 0 (wq_completion)bond24 (work_completion)(&(&slave->notify_work)->work) rtnl_mutex &tn->lock irq_context: softirq (&ndev->rs_timer) rcu_read_lock rcu_read_lock rcu_read_lock_bh rcu_read_lock rcu_read_lock_bh rcu_read_lock &obj_hash[i].lock pool_lock irq_context: 0 (wq_completion)bond25 irq_context: 0 (wq_completion)bond25 (work_completion)(&(&slave->notify_work)->work) irq_context: 0 (wq_completion)bond25 (work_completion)(&(&slave->notify_work)->work) &obj_hash[i].lock irq_context: 0 (wq_completion)bond25 (work_completion)(&(&slave->notify_work)->work) &base->lock irq_context: 0 (wq_completion)bond25 (work_completion)(&(&slave->notify_work)->work) &base->lock &obj_hash[i].lock irq_context: 0 (wq_completion)bond25 (work_completion)(&(&slave->notify_work)->work) &rq->__lock irq_context: 0 (wq_completion)bond25 (work_completion)(&(&slave->notify_work)->work) &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)bond25 (work_completion)(&(&bond->mcast_work)->work) irq_context: 0 (wq_completion)bond25 (work_completion)(&(&bond->mcast_work)->work) &obj_hash[i].lock irq_context: 0 (wq_completion)bond25 (work_completion)(&(&bond->mcast_work)->work) &base->lock irq_context: 0 (wq_completion)bond25 (work_completion)(&(&bond->mcast_work)->work) &base->lock &obj_hash[i].lock irq_context: 0 (wq_completion)bond25 (work_completion)(&(&bond->mcast_work)->work) &rq->__lock irq_context: 0 (wq_completion)bond25 (work_completion)(&(&bond->mcast_work)->work) &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 pernet_ops_rwsem devices_rwsem rcu_read_lock &rcu_state.expedited_wq irq_context: 0 pernet_ops_rwsem devices_rwsem rcu_read_lock &rcu_state.expedited_wq &p->pi_lock irq_context: 0 pernet_ops_rwsem devices_rwsem rcu_read_lock &rcu_state.expedited_wq &p->pi_lock &rq->__lock irq_context: 0 pernet_ops_rwsem devices_rwsem rcu_read_lock &rcu_state.expedited_wq &p->pi_lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)ipv6_addrconf (work_completion)(&(&ifa->dad_work)->work) rtnl_mutex &idev->mc_lock &dev_addr_list_lock_key/1 rcu_read_lock &bridge_netdev_addr_lock_key &n->list_lock irq_context: 0 (wq_completion)ipv6_addrconf (work_completion)(&(&ifa->dad_work)->work) rtnl_mutex &idev->mc_lock &dev_addr_list_lock_key/1 rcu_read_lock &bridge_netdev_addr_lock_key &n->list_lock &c->lock irq_context: 0 (wq_completion)bond25 (work_completion)(&(&slave->notify_work)->work) rtnl_mutex net_rwsem irq_context: 0 (wq_completion)bond25 (work_completion)(&(&slave->notify_work)->work) rtnl_mutex net_rwsem &list->lock#2 irq_context: 0 (wq_completion)bond25 (work_completion)(&(&slave->notify_work)->work) rtnl_mutex &tn->lock irq_context: 0 (wq_completion)bond25 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex fs_reclaim irq_context: 0 (wq_completion)bond25 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex fs_reclaim mmu_notifier_invalidate_range_start irq_context: 0 (wq_completion)bond25 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &c->lock irq_context: 0 (wq_completion)bond25 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex rcu_read_lock &bond->stats_lock/1 irq_context: 0 (wq_completion)bond25 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex rcu_read_lock &ndev->lock irq_context: 0 (wq_completion)bond25 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex nl_table_lock irq_context: 0 (wq_completion)bond25 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &obj_hash[i].lock irq_context: 0 (wq_completion)bond25 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex nl_table_wait.lock irq_context: 0 (wq_completion)bond25 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex net_rwsem irq_context: 0 (wq_completion)bond25 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex net_rwsem &list->lock#2 irq_context: 0 (wq_completion)bond25 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &tn->lock irq_context: 0 (wq_completion)bond25 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &idev->mc_lock irq_context: 0 (wq_completion)bond25 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &idev->mc_lock fs_reclaim irq_context: 0 (wq_completion)bond25 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &idev->mc_lock fs_reclaim mmu_notifier_invalidate_range_start irq_context: 0 (wq_completion)bond25 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &idev->mc_lock &c->lock irq_context: 0 (wq_completion)bond25 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &idev->mc_lock rcu_read_lock &ndev->lock irq_context: 0 (wq_completion)bond25 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &idev->mc_lock rcu_read_lock &c->lock irq_context: 0 (wq_completion)bond25 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &idev->mc_lock rcu_read_lock &dir->lock#2 irq_context: 0 (wq_completion)bond25 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &idev->mc_lock rcu_read_lock &ul->lock#2 irq_context: 0 (wq_completion)bond25 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &idev->mc_lock rcu_read_lock rcu_read_lock &c->lock irq_context: 0 (wq_completion)bond25 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &idev->mc_lock rcu_read_lock rcu_read_lock &obj_hash[i].lock irq_context: 0 (wq_completion)bond25 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &idev->mc_lock rcu_read_lock rcu_read_lock &dir->lock#2 irq_context: 0 (wq_completion)bond25 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &idev->mc_lock rcu_read_lock rcu_read_lock &tbl->lock irq_context: 0 (wq_completion)bond25 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &idev->mc_lock rcu_read_lock rcu_read_lock &n->lock irq_context: 0 (wq_completion)bond25 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &idev->mc_lock rcu_read_lock rcu_read_lock &____s->seqcount#9 irq_context: 0 (wq_completion)bond25 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &idev->mc_lock rcu_read_lock rcu_read_lock rcu_read_lock_bh rcu_read_lock rcu_read_lock_bh rcu_read_lock &br->multicast_lock irq_context: 0 (wq_completion)bond25 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &idev->mc_lock rcu_read_lock rcu_read_lock rcu_read_lock_bh rcu_read_lock rcu_read_lock_bh rcu_read_lock &br->multicast_lock &c->lock irq_context: 0 (wq_completion)bond25 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &idev->mc_lock rcu_read_lock rcu_read_lock rcu_read_lock_bh rcu_read_lock rcu_read_lock_bh rcu_read_lock &br->multicast_lock &____s->seqcount#2 irq_context: 0 (wq_completion)bond25 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &idev->mc_lock rcu_read_lock rcu_read_lock rcu_read_lock_bh rcu_read_lock rcu_read_lock_bh rcu_read_lock &br->multicast_lock &____s->seqcount irq_context: 0 (wq_completion)bond25 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &idev->mc_lock rcu_read_lock rcu_read_lock rcu_read_lock_bh rcu_read_lock rcu_read_lock_bh rcu_read_lock &br->multicast_lock &obj_hash[i].lock irq_context: 0 (wq_completion)bond25 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &idev->mc_lock rcu_read_lock rcu_read_lock rcu_read_lock_bh rcu_read_lock rcu_read_lock_bh rcu_read_lock &br->multicast_lock rcu_read_lock rhashtable_bucket irq_context: 0 (wq_completion)bond25 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &idev->mc_lock rcu_read_lock rcu_read_lock rcu_read_lock_bh rcu_read_lock rcu_read_lock_bh rcu_read_lock &br->multicast_lock nl_table_lock irq_context: 0 (wq_completion)bond25 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &idev->mc_lock rcu_read_lock rcu_read_lock rcu_read_lock_bh rcu_read_lock rcu_read_lock_bh rcu_read_lock &br->multicast_lock nl_table_wait.lock irq_context: 0 (wq_completion)bond25 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &idev->mc_lock rcu_read_lock rcu_read_lock rcu_read_lock_bh rcu_read_lock rcu_read_lock_bh rcu_read_lock &br->multicast_lock &base->lock irq_context: 0 (wq_completion)bond25 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &idev->mc_lock rcu_read_lock rcu_read_lock rcu_read_lock_bh rcu_read_lock rcu_read_lock_bh rcu_read_lock &br->multicast_lock &base->lock &obj_hash[i].lock irq_context: 0 (wq_completion)bond25 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &idev->mc_lock rcu_read_lock rcu_read_lock rcu_read_lock_bh rcu_read_lock rcu_read_lock_bh rcu_read_lock &data->lock irq_context: 0 (wq_completion)bond25 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &idev->mc_lock rcu_read_lock rcu_read_lock rcu_read_lock_bh rcu_read_lock rcu_read_lock_bh rcu_read_lock &obj_hash[i].lock irq_context: 0 (wq_completion)ipv6_addrconf (work_completion)(&(&ifa->dad_work)->work) rtnl_mutex rcu_read_lock rcu_read_lock fill_pool_map-wait-type-override &rq->__lock irq_context: 0 (wq_completion)wg-crypt-wg1 (work_completion)(&peer->transmit_packet_work) &peer->endpoint_lock rcu_read_lock_bh rcu_read_lock &tbl->lock irq_context: 0 (wq_completion)wg-crypt-wg1 (work_completion)(&peer->transmit_packet_work) &peer->endpoint_lock rcu_read_lock_bh rcu_read_lock &n->lock irq_context: 0 (wq_completion)wg-crypt-wg1 (work_completion)(&peer->transmit_packet_work) &peer->endpoint_lock rcu_read_lock_bh rcu_read_lock &____s->seqcount#9 irq_context: 0 (wq_completion)events (work_completion)(&(&nsim_dev->trap_data->trap_report_dw)->work) &devlink->lock_key#3 &nsim_trap_data->trap_lock rcu_read_lock &p->pi_lock irq_context: 0 (wq_completion)events (work_completion)(&(&nsim_dev->trap_data->trap_report_dw)->work) &devlink->lock_key#3 &nsim_trap_data->trap_lock rcu_read_lock &p->pi_lock &rq->__lock irq_context: 0 (wq_completion)events (work_completion)(&(&nsim_dev->trap_data->trap_report_dw)->work) &devlink->lock_key#3 &nsim_trap_data->trap_lock rcu_read_lock &p->pi_lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)events (work_completion)(&(&nsim_dev->trap_data->trap_report_dw)->work) &devlink->lock_key#3 &rcu_state.expedited_wq &p->pi_lock &rq->__lock &cfs_rq->removed.lock irq_context: 0 (wq_completion)mld (work_completion)(&(&idev->mc_dad_work)->work) &idev->mc_lock rcu_read_lock rcu_read_lock rcu_read_lock_bh rcu_read_lock rcu_read_lock_bh rcu_read_lock &br->multicast_lock &base->lock &obj_hash[i].lock irq_context: 0 (wq_completion)bond26 irq_context: 0 (wq_completion)bond26 (work_completion)(&(&slave->notify_work)->work) irq_context: 0 (wq_completion)bond26 (work_completion)(&(&slave->notify_work)->work) &obj_hash[i].lock irq_context: 0 (wq_completion)bond26 (work_completion)(&(&slave->notify_work)->work) &base->lock irq_context: 0 (wq_completion)bond26 (work_completion)(&(&slave->notify_work)->work) &base->lock &obj_hash[i].lock irq_context: 0 (wq_completion)bond26 (work_completion)(&(&bond->mcast_work)->work) irq_context: 0 (wq_completion)bond26 (work_completion)(&(&bond->mcast_work)->work) &obj_hash[i].lock irq_context: 0 (wq_completion)bond26 (work_completion)(&(&bond->mcast_work)->work) &base->lock irq_context: 0 (wq_completion)bond26 (work_completion)(&(&bond->mcast_work)->work) &base->lock &obj_hash[i].lock irq_context: 0 (wq_completion)bond26 (work_completion)(&(&slave->notify_work)->work) rtnl_mutex net_rwsem irq_context: 0 (wq_completion)bond26 (work_completion)(&(&slave->notify_work)->work) rtnl_mutex net_rwsem &list->lock#2 irq_context: 0 (wq_completion)bond26 (work_completion)(&(&slave->notify_work)->work) rtnl_mutex &tn->lock irq_context: 0 (wq_completion)bond26 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex fs_reclaim irq_context: 0 (wq_completion)bond26 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex fs_reclaim mmu_notifier_invalidate_range_start irq_context: 0 (wq_completion)bond26 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex remove_cache_srcu irq_context: 0 (wq_completion)bond26 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex remove_cache_srcu quarantine_lock irq_context: 0 (wq_completion)bond26 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex remove_cache_srcu &c->lock irq_context: 0 (wq_completion)bond26 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex remove_cache_srcu &n->list_lock irq_context: 0 (wq_completion)bond26 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex remove_cache_srcu &obj_hash[i].lock irq_context: 0 (wq_completion)bond26 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex rcu_read_lock &bond->stats_lock/1 irq_context: 0 (wq_completion)bond26 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex rcu_read_lock &ndev->lock irq_context: 0 (wq_completion)bond26 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex nl_table_lock irq_context: 0 (wq_completion)bond26 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &obj_hash[i].lock irq_context: 0 (wq_completion)bond26 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex nl_table_wait.lock irq_context: 0 (wq_completion)bond26 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex net_rwsem irq_context: 0 (wq_completion)bond26 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex net_rwsem &list->lock#2 irq_context: 0 (wq_completion)bond26 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &tn->lock irq_context: 0 (wq_completion)bond26 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &idev->mc_lock irq_context: 0 (wq_completion)bond26 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &idev->mc_lock fs_reclaim irq_context: 0 (wq_completion)bond26 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &idev->mc_lock fs_reclaim mmu_notifier_invalidate_range_start irq_context: 0 (wq_completion)bond26 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &idev->mc_lock &c->lock irq_context: 0 (wq_completion)bond26 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &idev->mc_lock rcu_read_lock &ndev->lock irq_context: 0 (wq_completion)bond26 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &idev->mc_lock rcu_read_lock &dir->lock#2 irq_context: 0 (wq_completion)bond26 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &idev->mc_lock rcu_read_lock &ul->lock#2 irq_context: 0 (wq_completion)bond26 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &idev->mc_lock rcu_read_lock rcu_read_lock &obj_hash[i].lock irq_context: 0 (wq_completion)bond26 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &idev->mc_lock rcu_read_lock rcu_read_lock &dir->lock#2 irq_context: 0 (wq_completion)bond26 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &idev->mc_lock rcu_read_lock rcu_read_lock &tbl->lock irq_context: 0 (wq_completion)bond26 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &idev->mc_lock rcu_read_lock rcu_read_lock &n->lock irq_context: 0 (wq_completion)bond26 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &idev->mc_lock rcu_read_lock rcu_read_lock &____s->seqcount#9 irq_context: 0 (wq_completion)bond26 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &idev->mc_lock rcu_read_lock rcu_read_lock rcu_read_lock_bh rcu_read_lock rcu_read_lock_bh rcu_read_lock &br->multicast_lock irq_context: 0 (wq_completion)bond26 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &idev->mc_lock rcu_read_lock rcu_read_lock rcu_read_lock_bh rcu_read_lock rcu_read_lock_bh rcu_read_lock &br->multicast_lock &obj_hash[i].lock irq_context: 0 (wq_completion)bond26 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &idev->mc_lock rcu_read_lock rcu_read_lock rcu_read_lock_bh rcu_read_lock rcu_read_lock_bh rcu_read_lock &br->multicast_lock rcu_read_lock rhashtable_bucket irq_context: 0 (wq_completion)bond26 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &idev->mc_lock rcu_read_lock rcu_read_lock rcu_read_lock_bh rcu_read_lock rcu_read_lock_bh rcu_read_lock &br->multicast_lock nl_table_lock irq_context: 0 (wq_completion)bond26 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &idev->mc_lock rcu_read_lock rcu_read_lock rcu_read_lock_bh rcu_read_lock rcu_read_lock_bh rcu_read_lock &br->multicast_lock nl_table_wait.lock irq_context: 0 (wq_completion)bond26 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &idev->mc_lock rcu_read_lock rcu_read_lock rcu_read_lock_bh rcu_read_lock rcu_read_lock_bh rcu_read_lock &br->multicast_lock &base->lock irq_context: 0 (wq_completion)bond26 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &idev->mc_lock rcu_read_lock rcu_read_lock rcu_read_lock_bh rcu_read_lock rcu_read_lock_bh rcu_read_lock &br->multicast_lock &base->lock &obj_hash[i].lock irq_context: 0 (wq_completion)bond26 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &idev->mc_lock rcu_read_lock rcu_read_lock rcu_read_lock_bh rcu_read_lock rcu_read_lock_bh rcu_read_lock &data->lock irq_context: 0 (wq_completion)bond26 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &idev->mc_lock rcu_read_lock rcu_read_lock rcu_read_lock_bh rcu_read_lock rcu_read_lock_bh rcu_read_lock &obj_hash[i].lock irq_context: 0 (wq_completion)bond10 (work_completion)(&(&bond->ad_work)->work) &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 rtnl_mutex cpu_hotplug_lock wq_pool_mutex &____s->seqcount#2 irq_context: 0 sk_lock-AF_INET pcpu_alloc_mutex &rq->__lock irq_context: 0 sk_lock-AF_INET pcpu_alloc_mutex &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 sk_lock-AF_INET pcpu_alloc_mutex.wait_lock irq_context: 0 (wq_completion)bond16#2 irq_context: 0 (wq_completion)bond16#2 (work_completion)(&(&slave->notify_work)->work) irq_context: 0 (wq_completion)bond16#2 (work_completion)(&(&slave->notify_work)->work) &obj_hash[i].lock irq_context: 0 (wq_completion)bond16#2 (work_completion)(&(&slave->notify_work)->work) &base->lock irq_context: 0 (wq_completion)bond16#2 (work_completion)(&(&slave->notify_work)->work) &base->lock &obj_hash[i].lock irq_context: 0 (wq_completion)bond16#2 (work_completion)(&(&slave->notify_work)->work) &rq->__lock irq_context: 0 (wq_completion)bond16#2 (work_completion)(&(&slave->notify_work)->work) &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 &sb->s_type->i_mutex_key#10 sk_lock-AF_INET6 slock-AF_INET6 batched_entropy_u8.lock irq_context: 0 &sb->s_type->i_mutex_key#10 sk_lock-AF_INET6 &meta->lock irq_context: 0 &sb->s_type->i_mutex_key#10 sk_lock-AF_INET6 kfence_freelist_lock irq_context: 0 (wq_completion)bond16#2 (work_completion)(&(&bond->mcast_work)->work) irq_context: 0 (wq_completion)bond16#2 (work_completion)(&(&bond->mcast_work)->work) &obj_hash[i].lock irq_context: 0 (wq_completion)bond16#2 (work_completion)(&(&bond->mcast_work)->work) &base->lock irq_context: 0 (wq_completion)bond16#2 (work_completion)(&(&bond->mcast_work)->work) &base->lock &obj_hash[i].lock irq_context: 0 (wq_completion)bond16#2 (work_completion)(&(&slave->notify_work)->work) rtnl_mutex net_rwsem irq_context: 0 (wq_completion)bond16#2 (work_completion)(&(&slave->notify_work)->work) rtnl_mutex net_rwsem &list->lock#2 irq_context: 0 (wq_completion)bond16#2 (work_completion)(&(&slave->notify_work)->work) rtnl_mutex &tn->lock irq_context: 0 (wq_completion)bond16#2 (work_completion)(&(&slave->notify_work)->work) rtnl_mutex.wait_lock irq_context: 0 (wq_completion)bond16#2 (work_completion)(&(&slave->notify_work)->work) &p->pi_lock irq_context: 0 (wq_completion)bond27 irq_context: 0 (wq_completion)bond27 (work_completion)(&(&slave->notify_work)->work) irq_context: 0 (wq_completion)bond27 (work_completion)(&(&slave->notify_work)->work) &obj_hash[i].lock irq_context: 0 (wq_completion)bond27 (work_completion)(&(&slave->notify_work)->work) &base->lock irq_context: 0 (wq_completion)bond27 (work_completion)(&(&slave->notify_work)->work) &base->lock &obj_hash[i].lock irq_context: 0 (wq_completion)bond27 (work_completion)(&(&bond->mcast_work)->work) irq_context: 0 (wq_completion)bond27 (work_completion)(&(&bond->mcast_work)->work) &obj_hash[i].lock irq_context: 0 (wq_completion)bond27 (work_completion)(&(&bond->mcast_work)->work) &base->lock irq_context: 0 (wq_completion)bond27 (work_completion)(&(&bond->mcast_work)->work) &base->lock &obj_hash[i].lock irq_context: 0 (wq_completion)bond27 (work_completion)(&(&bond->mcast_work)->work) &rq->__lock irq_context: 0 (wq_completion)bond27 (work_completion)(&(&bond->mcast_work)->work) &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)bond27 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex fs_reclaim irq_context: 0 (wq_completion)bond27 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex fs_reclaim mmu_notifier_invalidate_range_start irq_context: 0 (wq_completion)bond27 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &c->lock irq_context: 0 (wq_completion)bond27 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &n->list_lock irq_context: 0 (wq_completion)bond27 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &n->list_lock &c->lock irq_context: 0 (wq_completion)bond27 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex rcu_read_lock &bond->stats_lock/1 irq_context: 0 (wq_completion)bond27 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex rcu_read_lock &ndev->lock irq_context: 0 (wq_completion)bond27 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex nl_table_lock irq_context: 0 (wq_completion)bond27 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &obj_hash[i].lock irq_context: 0 (wq_completion)bond27 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex nl_table_wait.lock irq_context: 0 (wq_completion)bond27 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex net_rwsem irq_context: 0 (wq_completion)bond27 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex net_rwsem &list->lock#2 irq_context: 0 (wq_completion)bond27 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &tn->lock irq_context: 0 (wq_completion)bond27 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &idev->mc_lock irq_context: 0 (wq_completion)bond27 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &idev->mc_lock fs_reclaim irq_context: 0 (wq_completion)bond27 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &idev->mc_lock fs_reclaim mmu_notifier_invalidate_range_start irq_context: 0 (wq_completion)bond27 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &idev->mc_lock &c->lock irq_context: 0 (wq_completion)bond27 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &idev->mc_lock rcu_read_lock &ndev->lock irq_context: 0 (wq_completion)bond27 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &idev->mc_lock rcu_read_lock &dir->lock#2 irq_context: 0 (wq_completion)bond27 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &idev->mc_lock rcu_read_lock &ul->lock#2 irq_context: 0 (wq_completion)bond27 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &idev->mc_lock rcu_read_lock rcu_read_lock &obj_hash[i].lock irq_context: 0 (wq_completion)bond27 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &idev->mc_lock rcu_read_lock rcu_read_lock &dir->lock#2 irq_context: 0 (wq_completion)bond27 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &idev->mc_lock rcu_read_lock rcu_read_lock &tbl->lock irq_context: 0 (wq_completion)bond27 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &idev->mc_lock rcu_read_lock rcu_read_lock &n->lock irq_context: 0 (wq_completion)bond27 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &idev->mc_lock rcu_read_lock rcu_read_lock &____s->seqcount#9 irq_context: 0 (wq_completion)bond27 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &idev->mc_lock rcu_read_lock rcu_read_lock rcu_read_lock_bh rcu_read_lock rcu_read_lock_bh rcu_read_lock &br->multicast_lock irq_context: 0 (wq_completion)bond27 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &idev->mc_lock rcu_read_lock rcu_read_lock rcu_read_lock_bh rcu_read_lock rcu_read_lock_bh rcu_read_lock &br->multicast_lock &obj_hash[i].lock irq_context: 0 (wq_completion)bond27 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &idev->mc_lock rcu_read_lock rcu_read_lock rcu_read_lock_bh rcu_read_lock rcu_read_lock_bh rcu_read_lock &br->multicast_lock rcu_read_lock rhashtable_bucket irq_context: 0 (wq_completion)bond27 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &idev->mc_lock rcu_read_lock rcu_read_lock rcu_read_lock_bh rcu_read_lock rcu_read_lock_bh rcu_read_lock &br->multicast_lock nl_table_lock irq_context: 0 (wq_completion)bond27 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &idev->mc_lock rcu_read_lock rcu_read_lock rcu_read_lock_bh rcu_read_lock rcu_read_lock_bh rcu_read_lock &br->multicast_lock nl_table_wait.lock irq_context: 0 (wq_completion)bond27 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &idev->mc_lock rcu_read_lock rcu_read_lock rcu_read_lock_bh rcu_read_lock rcu_read_lock_bh rcu_read_lock &br->multicast_lock &base->lock irq_context: 0 (wq_completion)bond27 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &idev->mc_lock rcu_read_lock rcu_read_lock rcu_read_lock_bh rcu_read_lock rcu_read_lock_bh rcu_read_lock &br->multicast_lock &base->lock &obj_hash[i].lock irq_context: 0 (wq_completion)bond27 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &idev->mc_lock rcu_read_lock rcu_read_lock rcu_read_lock_bh rcu_read_lock rcu_read_lock_bh rcu_read_lock &data->lock irq_context: 0 (wq_completion)bond27 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &idev->mc_lock rcu_read_lock rcu_read_lock rcu_read_lock_bh rcu_read_lock rcu_read_lock_bh rcu_read_lock &obj_hash[i].lock irq_context: 0 (wq_completion)bond16#2 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex fs_reclaim irq_context: 0 (wq_completion)bond16#2 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex fs_reclaim mmu_notifier_invalidate_range_start irq_context: 0 (wq_completion)bond16#2 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &c->lock irq_context: 0 (wq_completion)bond16#2 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex rcu_read_lock &bond->stats_lock/1 irq_context: 0 (wq_completion)bond16#2 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex rcu_read_lock &ndev->lock irq_context: 0 (wq_completion)bond16#2 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex nl_table_lock irq_context: 0 (wq_completion)bond16#2 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &obj_hash[i].lock irq_context: 0 (wq_completion)bond16#2 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex nl_table_wait.lock irq_context: 0 (wq_completion)bond16#2 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex net_rwsem irq_context: 0 (wq_completion)bond16#2 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex net_rwsem &list->lock#2 irq_context: 0 (wq_completion)bond16#2 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &tn->lock irq_context: 0 (wq_completion)bond16#2 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &idev->mc_lock irq_context: 0 (wq_completion)bond16#2 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &idev->mc_lock fs_reclaim irq_context: 0 (wq_completion)bond16#2 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &idev->mc_lock fs_reclaim mmu_notifier_invalidate_range_start irq_context: 0 (wq_completion)bond16#2 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &idev->mc_lock rcu_read_lock &ndev->lock irq_context: 0 (wq_completion)bond16#2 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &idev->mc_lock rcu_read_lock &dir->lock#2 irq_context: 0 (wq_completion)bond16#2 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &idev->mc_lock rcu_read_lock &ul->lock#2 irq_context: 0 (wq_completion)bond16#2 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &idev->mc_lock rcu_read_lock rcu_read_lock &obj_hash[i].lock irq_context: 0 (wq_completion)bond16#2 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &idev->mc_lock rcu_read_lock rcu_read_lock &dir->lock#2 irq_context: 0 (wq_completion)bond16#2 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &idev->mc_lock rcu_read_lock rcu_read_lock &tbl->lock irq_context: 0 (wq_completion)bond16#2 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &idev->mc_lock rcu_read_lock rcu_read_lock &n->lock irq_context: 0 (wq_completion)bond16#2 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &idev->mc_lock rcu_read_lock rcu_read_lock &____s->seqcount#9 irq_context: 0 (wq_completion)bond16#2 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &idev->mc_lock rcu_read_lock rcu_read_lock rcu_read_lock_bh rcu_read_lock rcu_read_lock_bh rcu_read_lock &br->multicast_lock irq_context: 0 (wq_completion)bond16#2 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &idev->mc_lock rcu_read_lock rcu_read_lock rcu_read_lock_bh rcu_read_lock rcu_read_lock_bh rcu_read_lock &br->multicast_lock &obj_hash[i].lock irq_context: 0 (wq_completion)bond16#2 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &idev->mc_lock rcu_read_lock rcu_read_lock rcu_read_lock_bh rcu_read_lock rcu_read_lock_bh rcu_read_lock &br->multicast_lock rcu_read_lock rhashtable_bucket irq_context: 0 (wq_completion)bond16#2 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &idev->mc_lock rcu_read_lock rcu_read_lock rcu_read_lock_bh rcu_read_lock rcu_read_lock_bh rcu_read_lock &br->multicast_lock &c->lock irq_context: 0 (wq_completion)bond16#2 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &idev->mc_lock rcu_read_lock rcu_read_lock rcu_read_lock_bh rcu_read_lock rcu_read_lock_bh rcu_read_lock &br->multicast_lock &n->list_lock irq_context: 0 (wq_completion)bond16#2 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &idev->mc_lock rcu_read_lock rcu_read_lock rcu_read_lock_bh rcu_read_lock rcu_read_lock_bh rcu_read_lock &br->multicast_lock &n->list_lock &c->lock irq_context: 0 (wq_completion)bond16#2 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &idev->mc_lock rcu_read_lock rcu_read_lock rcu_read_lock_bh rcu_read_lock rcu_read_lock_bh rcu_read_lock &br->multicast_lock nl_table_lock irq_context: 0 (wq_completion)bond16#2 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &idev->mc_lock rcu_read_lock rcu_read_lock rcu_read_lock_bh rcu_read_lock rcu_read_lock_bh rcu_read_lock &br->multicast_lock nl_table_wait.lock irq_context: 0 (wq_completion)bond16#2 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &idev->mc_lock rcu_read_lock rcu_read_lock rcu_read_lock_bh rcu_read_lock rcu_read_lock_bh rcu_read_lock &br->multicast_lock &base->lock irq_context: 0 (wq_completion)bond16#2 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &idev->mc_lock rcu_read_lock rcu_read_lock rcu_read_lock_bh rcu_read_lock rcu_read_lock_bh rcu_read_lock &br->multicast_lock &base->lock &obj_hash[i].lock irq_context: 0 (wq_completion)bond16#2 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &idev->mc_lock rcu_read_lock rcu_read_lock rcu_read_lock_bh rcu_read_lock rcu_read_lock_bh rcu_read_lock &data->lock irq_context: 0 (wq_completion)bond16#2 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &idev->mc_lock rcu_read_lock rcu_read_lock rcu_read_lock_bh rcu_read_lock rcu_read_lock_bh rcu_read_lock &obj_hash[i].lock irq_context: 0 (wq_completion)bond27 (work_completion)(&(&slave->notify_work)->work) rtnl_mutex net_rwsem irq_context: 0 (wq_completion)bond27 (work_completion)(&(&slave->notify_work)->work) rtnl_mutex net_rwsem &list->lock#2 irq_context: 0 (wq_completion)bond27 (work_completion)(&(&slave->notify_work)->work) rtnl_mutex &tn->lock irq_context: 0 sb_writers#4 &type->i_mutex_dir_key#3 jbd2_handle &rcu_state.expedited_wq irq_context: 0 sb_writers#4 &type->i_mutex_dir_key#3 jbd2_handle &rcu_state.expedited_wq &p->pi_lock irq_context: 0 sb_writers#4 &type->i_mutex_dir_key#3 jbd2_handle &rcu_state.expedited_wq &p->pi_lock &rq->__lock irq_context: 0 sb_writers#4 &type->i_mutex_dir_key#3 jbd2_handle &rcu_state.expedited_wq &p->pi_lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 &sb->s_type->i_mutex_key#10 slock-AF_INET6 &meta->lock irq_context: 0 &sb->s_type->i_mutex_key#10 slock-AF_INET6 kfence_freelist_lock irq_context: 0 &vma->vm_lock->lock remove_cache_srcu rcu_read_lock rcu_node_0 irq_context: 0 &vma->vm_lock->lock remove_cache_srcu rcu_read_lock &rq->__lock irq_context: 0 &vma->vm_lock->lock remove_cache_srcu rcu_read_lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)bond28 irq_context: 0 (wq_completion)bond28 (work_completion)(&(&slave->notify_work)->work) irq_context: 0 (wq_completion)bond28 (work_completion)(&(&slave->notify_work)->work) &obj_hash[i].lock irq_context: 0 (wq_completion)bond28 (work_completion)(&(&slave->notify_work)->work) &base->lock irq_context: 0 (wq_completion)bond28 (work_completion)(&(&slave->notify_work)->work) &base->lock &obj_hash[i].lock irq_context: 0 (wq_completion)bond28 (work_completion)(&(&slave->notify_work)->work) &rq->__lock irq_context: 0 (wq_completion)bond28 (work_completion)(&(&slave->notify_work)->work) &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)mld (work_completion)(&(&idev->mc_ifc_work)->work) &idev->mc_lock fs_reclaim mmu_notifier_invalidate_range_start &rq->__lock irq_context: softirq (&tw->tw_timer) &n->list_lock irq_context: 0 (wq_completion)bond28 (work_completion)(&(&bond->mcast_work)->work) irq_context: 0 (wq_completion)bond28 (work_completion)(&(&bond->mcast_work)->work) &obj_hash[i].lock irq_context: 0 (wq_completion)bond28 (work_completion)(&(&bond->mcast_work)->work) &base->lock irq_context: 0 (wq_completion)bond28 (work_completion)(&(&bond->mcast_work)->work) &base->lock &obj_hash[i].lock irq_context: 0 (wq_completion)bond28 (work_completion)(&(&slave->notify_work)->work) rtnl_mutex net_rwsem irq_context: 0 (wq_completion)bond28 (work_completion)(&(&slave->notify_work)->work) rtnl_mutex net_rwsem &list->lock#2 irq_context: 0 (wq_completion)bond28 (work_completion)(&(&slave->notify_work)->work) rtnl_mutex &tn->lock irq_context: 0 (wq_completion)bond28 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex fs_reclaim irq_context: 0 (wq_completion)bond28 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex fs_reclaim mmu_notifier_invalidate_range_start irq_context: 0 (wq_completion)bond28 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &c->lock irq_context: 0 (wq_completion)bond28 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &rq->__lock irq_context: 0 (wq_completion)bond28 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)bond28 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex rcu_read_lock &bond->stats_lock/1 irq_context: 0 (wq_completion)bond28 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex rcu_read_lock &ndev->lock irq_context: 0 (wq_completion)bond28 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex nl_table_lock irq_context: 0 (wq_completion)bond28 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &obj_hash[i].lock irq_context: 0 (wq_completion)bond28 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex nl_table_wait.lock irq_context: 0 (wq_completion)bond28 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex net_rwsem irq_context: 0 (wq_completion)bond28 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex net_rwsem &list->lock#2 irq_context: 0 (wq_completion)bond28 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &tn->lock irq_context: 0 (wq_completion)bond28 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &idev->mc_lock irq_context: 0 (wq_completion)bond29 irq_context: 0 (wq_completion)bond29 (work_completion)(&(&slave->notify_work)->work) irq_context: 0 (wq_completion)bond29 (work_completion)(&(&slave->notify_work)->work) &obj_hash[i].lock irq_context: 0 (wq_completion)bond29 (work_completion)(&(&slave->notify_work)->work) &base->lock irq_context: 0 (wq_completion)bond29 (work_completion)(&(&slave->notify_work)->work) &base->lock &obj_hash[i].lock irq_context: 0 (wq_completion)bond29 (work_completion)(&(&bond->mcast_work)->work) irq_context: 0 (wq_completion)bond29 (work_completion)(&(&bond->mcast_work)->work) &obj_hash[i].lock irq_context: 0 (wq_completion)bond29 (work_completion)(&(&bond->mcast_work)->work) &base->lock irq_context: 0 (wq_completion)bond29 (work_completion)(&(&bond->mcast_work)->work) &base->lock &obj_hash[i].lock irq_context: 0 (wq_completion)bond29 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex fs_reclaim irq_context: 0 (wq_completion)bond29 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex fs_reclaim mmu_notifier_invalidate_range_start irq_context: 0 (wq_completion)bond29 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &c->lock irq_context: 0 (wq_completion)bond29 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex rcu_read_lock &bond->stats_lock/1 irq_context: 0 (wq_completion)bond29 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex rcu_read_lock &ndev->lock irq_context: 0 (wq_completion)bond29 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex nl_table_lock irq_context: 0 (wq_completion)bond29 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &obj_hash[i].lock irq_context: 0 (wq_completion)bond29 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex nl_table_wait.lock irq_context: 0 (wq_completion)bond29 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex net_rwsem irq_context: 0 (wq_completion)bond29 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex net_rwsem &list->lock#2 irq_context: 0 (wq_completion)bond29 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &tn->lock irq_context: 0 (wq_completion)bond29 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &idev->mc_lock irq_context: 0 (wq_completion)bond29 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &idev->mc_lock fs_reclaim irq_context: 0 (wq_completion)bond29 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &idev->mc_lock fs_reclaim mmu_notifier_invalidate_range_start irq_context: 0 (wq_completion)bond29 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &idev->mc_lock &c->lock irq_context: 0 (wq_completion)bond29 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &idev->mc_lock &n->list_lock irq_context: 0 (wq_completion)bond29 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &idev->mc_lock &n->list_lock &c->lock irq_context: 0 (wq_completion)bond29 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &idev->mc_lock rcu_read_lock &ndev->lock irq_context: 0 (wq_completion)bond29 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &idev->mc_lock rcu_read_lock &dir->lock#2 irq_context: 0 (wq_completion)bond29 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &idev->mc_lock rcu_read_lock &ul->lock#2 irq_context: 0 (wq_completion)bond29 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &idev->mc_lock rcu_read_lock rcu_read_lock &obj_hash[i].lock irq_context: 0 (wq_completion)bond29 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &idev->mc_lock rcu_read_lock rcu_read_lock &dir->lock#2 irq_context: 0 (wq_completion)bond29 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &idev->mc_lock rcu_read_lock rcu_read_lock &tbl->lock irq_context: 0 (wq_completion)bond29 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &idev->mc_lock rcu_read_lock rcu_read_lock &n->lock irq_context: 0 (wq_completion)bond29 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &idev->mc_lock rcu_read_lock rcu_read_lock &____s->seqcount#9 irq_context: 0 (wq_completion)bond29 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &idev->mc_lock rcu_read_lock rcu_read_lock rcu_read_lock_bh rcu_read_lock rcu_read_lock_bh rcu_read_lock &br->multicast_lock irq_context: 0 (wq_completion)bond29 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &idev->mc_lock rcu_read_lock rcu_read_lock rcu_read_lock_bh rcu_read_lock rcu_read_lock_bh rcu_read_lock &br->multicast_lock &obj_hash[i].lock irq_context: 0 (wq_completion)bond29 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &idev->mc_lock rcu_read_lock rcu_read_lock rcu_read_lock_bh rcu_read_lock rcu_read_lock_bh rcu_read_lock &br->multicast_lock rcu_read_lock rhashtable_bucket irq_context: 0 (wq_completion)bond29 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &idev->mc_lock rcu_read_lock rcu_read_lock rcu_read_lock_bh rcu_read_lock rcu_read_lock_bh rcu_read_lock &br->multicast_lock nl_table_lock irq_context: 0 (wq_completion)bond29 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &idev->mc_lock rcu_read_lock rcu_read_lock rcu_read_lock_bh rcu_read_lock rcu_read_lock_bh rcu_read_lock &br->multicast_lock nl_table_wait.lock irq_context: 0 (wq_completion)bond29 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &idev->mc_lock rcu_read_lock rcu_read_lock rcu_read_lock_bh rcu_read_lock rcu_read_lock_bh rcu_read_lock &br->multicast_lock &base->lock irq_context: 0 (wq_completion)bond29 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &idev->mc_lock rcu_read_lock rcu_read_lock rcu_read_lock_bh rcu_read_lock rcu_read_lock_bh rcu_read_lock &br->multicast_lock &base->lock &obj_hash[i].lock irq_context: 0 (wq_completion)bond29 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &idev->mc_lock rcu_read_lock rcu_read_lock rcu_read_lock_bh rcu_read_lock rcu_read_lock_bh rcu_read_lock &data->lock irq_context: 0 (wq_completion)bond29 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &idev->mc_lock rcu_read_lock rcu_read_lock rcu_read_lock_bh rcu_read_lock rcu_read_lock_bh rcu_read_lock &obj_hash[i].lock irq_context: 0 (wq_completion)bond29 (work_completion)(&(&slave->notify_work)->work) rtnl_mutex net_rwsem irq_context: 0 (wq_completion)bond29 (work_completion)(&(&slave->notify_work)->work) rtnl_mutex net_rwsem &list->lock#2 irq_context: 0 (wq_completion)bond29 (work_completion)(&(&slave->notify_work)->work) rtnl_mutex &tn->lock irq_context: 0 (wq_completion)mld (work_completion)(&(&idev->mc_ifc_work)->work) &idev->mc_lock rcu_read_lock rcu_read_lock rcu_read_lock_bh rcu_read_lock rcu_read_lock_bh rcu_read_lock &data->lock &obj_hash[i].lock irq_context: 0 (wq_completion)mld (work_completion)(&(&idev->mc_ifc_work)->work) &idev->mc_lock rcu_read_lock rcu_read_lock rcu_read_lock_bh rcu_read_lock rcu_read_lock_bh rcu_read_lock &data->lock &base->lock irq_context: 0 (wq_completion)mld (work_completion)(&(&idev->mc_ifc_work)->work) &idev->mc_lock rcu_read_lock rcu_read_lock rcu_read_lock_bh rcu_read_lock rcu_read_lock_bh rcu_read_lock &data->lock &base->lock &obj_hash[i].lock irq_context: 0 rcu_read_lock rcu_read_lock fill_pool_map-wait-type-override pool_lock irq_context: 0 (wq_completion)bond30 irq_context: 0 (wq_completion)bond30 (work_completion)(&(&slave->notify_work)->work) irq_context: 0 (wq_completion)bond30 (work_completion)(&(&slave->notify_work)->work) &obj_hash[i].lock irq_context: 0 (wq_completion)bond30 (work_completion)(&(&slave->notify_work)->work) &base->lock irq_context: 0 (wq_completion)bond30 (work_completion)(&(&slave->notify_work)->work) &base->lock &obj_hash[i].lock irq_context: 0 (wq_completion)bond30 (work_completion)(&(&bond->mcast_work)->work) irq_context: 0 (wq_completion)bond30 (work_completion)(&(&bond->mcast_work)->work) &obj_hash[i].lock irq_context: 0 (wq_completion)bond30 (work_completion)(&(&bond->mcast_work)->work) &base->lock irq_context: 0 (wq_completion)bond30 (work_completion)(&(&bond->mcast_work)->work) &base->lock &obj_hash[i].lock irq_context: 0 (wq_completion)bond30 (work_completion)(&(&slave->notify_work)->work) rtnl_mutex net_rwsem irq_context: 0 (wq_completion)bond30 (work_completion)(&(&slave->notify_work)->work) rtnl_mutex net_rwsem &list->lock#2 irq_context: 0 (wq_completion)bond30 (work_completion)(&(&slave->notify_work)->work) rtnl_mutex &tn->lock irq_context: 0 (wq_completion)bond30 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex fs_reclaim irq_context: 0 (wq_completion)bond30 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex fs_reclaim mmu_notifier_invalidate_range_start irq_context: 0 (wq_completion)bond30 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex rcu_read_lock &bond->stats_lock/1 irq_context: 0 (wq_completion)bond30 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex rcu_read_lock &ndev->lock irq_context: 0 (wq_completion)bond30 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex nl_table_lock irq_context: 0 (wq_completion)bond30 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &obj_hash[i].lock irq_context: 0 (wq_completion)bond30 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex nl_table_wait.lock irq_context: 0 (wq_completion)bond30 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex net_rwsem irq_context: 0 (wq_completion)bond30 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex net_rwsem &list->lock#2 irq_context: 0 (wq_completion)bond30 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &tn->lock irq_context: 0 (wq_completion)bond30 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &idev->mc_lock irq_context: 0 (wq_completion)bond30 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &idev->mc_lock fs_reclaim irq_context: 0 (wq_completion)bond30 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &idev->mc_lock fs_reclaim mmu_notifier_invalidate_range_start irq_context: 0 (wq_completion)bond30 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &idev->mc_lock rcu_read_lock &ndev->lock irq_context: 0 (wq_completion)bond30 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &idev->mc_lock rcu_read_lock &dir->lock#2 irq_context: 0 (wq_completion)bond30 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &idev->mc_lock rcu_read_lock &ul->lock#2 irq_context: 0 (wq_completion)bond30 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &idev->mc_lock rcu_read_lock rcu_read_lock &c->lock irq_context: 0 (wq_completion)bond30 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &idev->mc_lock rcu_read_lock rcu_read_lock &n->list_lock irq_context: 0 (wq_completion)bond30 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &idev->mc_lock rcu_read_lock rcu_read_lock &n->list_lock &c->lock irq_context: 0 (wq_completion)bond30 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &idev->mc_lock rcu_read_lock rcu_read_lock &obj_hash[i].lock irq_context: 0 (wq_completion)bond30 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &idev->mc_lock rcu_read_lock rcu_read_lock &dir->lock#2 irq_context: 0 (wq_completion)bond30 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &idev->mc_lock rcu_read_lock rcu_read_lock &tbl->lock irq_context: 0 (wq_completion)bond30 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &idev->mc_lock rcu_read_lock rcu_read_lock &n->lock irq_context: 0 (wq_completion)bond30 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &idev->mc_lock rcu_read_lock rcu_read_lock &____s->seqcount#9 irq_context: 0 (wq_completion)bond30 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &idev->mc_lock rcu_read_lock rcu_read_lock rcu_read_lock_bh rcu_read_lock rcu_read_lock_bh rcu_read_lock &br->multicast_lock irq_context: 0 (wq_completion)bond30 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &idev->mc_lock rcu_read_lock rcu_read_lock rcu_read_lock_bh rcu_read_lock rcu_read_lock_bh rcu_read_lock &br->multicast_lock &obj_hash[i].lock irq_context: 0 (wq_completion)bond30 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &idev->mc_lock rcu_read_lock rcu_read_lock rcu_read_lock_bh rcu_read_lock rcu_read_lock_bh rcu_read_lock &br->multicast_lock rcu_read_lock rhashtable_bucket irq_context: 0 (wq_completion)bond30 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &idev->mc_lock rcu_read_lock rcu_read_lock rcu_read_lock_bh rcu_read_lock rcu_read_lock_bh rcu_read_lock &br->multicast_lock &c->lock irq_context: 0 (wq_completion)bond30 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &idev->mc_lock rcu_read_lock rcu_read_lock rcu_read_lock_bh rcu_read_lock rcu_read_lock_bh rcu_read_lock &br->multicast_lock nl_table_lock irq_context: 0 (wq_completion)bond30 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &idev->mc_lock rcu_read_lock rcu_read_lock rcu_read_lock_bh rcu_read_lock rcu_read_lock_bh rcu_read_lock &br->multicast_lock nl_table_wait.lock irq_context: 0 (wq_completion)bond30 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &idev->mc_lock rcu_read_lock rcu_read_lock rcu_read_lock_bh rcu_read_lock rcu_read_lock_bh rcu_read_lock &br->multicast_lock &base->lock irq_context: 0 (wq_completion)bond30 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &idev->mc_lock rcu_read_lock rcu_read_lock rcu_read_lock_bh rcu_read_lock rcu_read_lock_bh rcu_read_lock &br->multicast_lock &base->lock &obj_hash[i].lock irq_context: 0 (wq_completion)bond30 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &idev->mc_lock rcu_read_lock rcu_read_lock rcu_read_lock_bh rcu_read_lock rcu_read_lock_bh rcu_read_lock &data->lock irq_context: 0 (wq_completion)bond30 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &idev->mc_lock rcu_read_lock rcu_read_lock rcu_read_lock_bh rcu_read_lock rcu_read_lock_bh rcu_read_lock &obj_hash[i].lock irq_context: 0 (wq_completion)bond31 irq_context: 0 (wq_completion)bond31 (work_completion)(&(&slave->notify_work)->work) irq_context: 0 (wq_completion)bond31 (work_completion)(&(&slave->notify_work)->work) &obj_hash[i].lock irq_context: 0 (wq_completion)bond31 (work_completion)(&(&slave->notify_work)->work) &base->lock irq_context: 0 (wq_completion)bond31 (work_completion)(&(&slave->notify_work)->work) &base->lock &obj_hash[i].lock irq_context: 0 (wq_completion)events (work_completion)(&p->wq) rcu_read_lock &pool->lock &p->pi_lock &rq->__lock irq_context: 0 (wq_completion)events (work_completion)(&p->wq) rcu_read_lock &pool->lock &p->pi_lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)bond31 (work_completion)(&(&bond->mcast_work)->work) irq_context: 0 (wq_completion)bond31 (work_completion)(&(&bond->mcast_work)->work) &obj_hash[i].lock irq_context: 0 (wq_completion)bond31 (work_completion)(&(&bond->mcast_work)->work) &base->lock irq_context: 0 (wq_completion)bond31 (work_completion)(&(&bond->mcast_work)->work) &base->lock &obj_hash[i].lock irq_context: 0 (wq_completion)bond31 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex fs_reclaim irq_context: 0 (wq_completion)bond31 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex fs_reclaim mmu_notifier_invalidate_range_start irq_context: 0 (wq_completion)bond31 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex rcu_read_lock &bond->stats_lock/1 irq_context: 0 (wq_completion)bond31 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex rcu_read_lock &ndev->lock irq_context: 0 (wq_completion)bond31 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex nl_table_lock irq_context: 0 (wq_completion)bond31 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &obj_hash[i].lock irq_context: 0 (wq_completion)bond31 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex nl_table_wait.lock irq_context: 0 (wq_completion)bond31 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex net_rwsem irq_context: 0 (wq_completion)bond31 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex net_rwsem &list->lock#2 irq_context: 0 (wq_completion)bond31 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &tn->lock irq_context: 0 (wq_completion)bond31 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &idev->mc_lock irq_context: 0 (wq_completion)bond31 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex.wait_lock irq_context: 0 (wq_completion)bond31 (work_completion)(&(&bond->mcast_work)->work) &p->pi_lock irq_context: 0 (wq_completion)bond31 (work_completion)(&(&slave->notify_work)->work) rtnl_mutex net_rwsem irq_context: 0 (wq_completion)bond31 (work_completion)(&(&slave->notify_work)->work) rtnl_mutex net_rwsem &list->lock#2 irq_context: 0 (wq_completion)bond31 (work_completion)(&(&slave->notify_work)->work) rtnl_mutex &tn->lock irq_context: 0 (wq_completion)bond31 (work_completion)(&(&slave->notify_work)->work) rtnl_mutex.wait_lock irq_context: 0 (wq_completion)bond31 (work_completion)(&(&slave->notify_work)->work) &p->pi_lock irq_context: 0 (wq_completion)events (work_completion)(&rdev->wiphy_work) &rdev->wiphy.mtx &wdev->mtx console_owner_lock irq_context: 0 (wq_completion)events (work_completion)(&rdev->wiphy_work) &rdev->wiphy.mtx &wdev->mtx console_owner irq_context: 0 (wq_completion)bond17#2 irq_context: 0 (wq_completion)bond17#2 (work_completion)(&(&slave->notify_work)->work) irq_context: 0 (wq_completion)bond17#2 (work_completion)(&(&slave->notify_work)->work) &rq->__lock irq_context: 0 (wq_completion)bond17#2 (work_completion)(&(&slave->notify_work)->work) &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)bond17#2 (work_completion)(&(&slave->notify_work)->work) &obj_hash[i].lock irq_context: 0 (wq_completion)bond17#2 (work_completion)(&(&slave->notify_work)->work) &base->lock irq_context: 0 (wq_completion)bond17#2 (work_completion)(&(&slave->notify_work)->work) &base->lock &obj_hash[i].lock irq_context: 0 (wq_completion)bond17#2 (work_completion)(&(&bond->mcast_work)->work) irq_context: 0 (wq_completion)bond17#2 (work_completion)(&(&bond->mcast_work)->work) &obj_hash[i].lock irq_context: 0 (wq_completion)bond17#2 (work_completion)(&(&bond->mcast_work)->work) &base->lock irq_context: 0 (wq_completion)bond17#2 (work_completion)(&(&bond->mcast_work)->work) &base->lock &obj_hash[i].lock irq_context: 0 (wq_completion)ipv6_addrconf (work_completion)(&(&ifa->dad_work)->work) rtnl_mutex &idev->mc_lock rcu_read_lock rcu_read_lock rcu_read_lock_bh rcu_read_lock rcu_read_lock_bh rcu_read_lock &data->lock &obj_hash[i].lock irq_context: 0 (wq_completion)ipv6_addrconf (work_completion)(&(&ifa->dad_work)->work) rtnl_mutex &idev->mc_lock rcu_read_lock rcu_read_lock rcu_read_lock_bh rcu_read_lock rcu_read_lock_bh rcu_read_lock &data->lock &base->lock irq_context: 0 (wq_completion)ipv6_addrconf (work_completion)(&(&ifa->dad_work)->work) rtnl_mutex &idev->mc_lock rcu_read_lock rcu_read_lock rcu_read_lock_bh rcu_read_lock rcu_read_lock_bh rcu_read_lock &data->lock &base->lock &obj_hash[i].lock irq_context: 0 (wq_completion)bond17#2 (work_completion)(&(&slave->notify_work)->work) rtnl_mutex net_rwsem irq_context: 0 (wq_completion)bond17#2 (work_completion)(&(&slave->notify_work)->work) rtnl_mutex net_rwsem &list->lock#2 irq_context: 0 (wq_completion)bond17#2 (work_completion)(&(&slave->notify_work)->work) rtnl_mutex &tn->lock irq_context: 0 (wq_completion)bond17#2 (work_completion)(&(&slave->notify_work)->work) rtnl_mutex.wait_lock irq_context: 0 (wq_completion)bond17#2 (work_completion)(&(&slave->notify_work)->work) &p->pi_lock irq_context: 0 (wq_completion)bond17#2 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex fs_reclaim irq_context: 0 (wq_completion)bond17#2 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex fs_reclaim mmu_notifier_invalidate_range_start irq_context: 0 (wq_completion)bond17#2 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &c->lock irq_context: 0 (wq_completion)bond17#2 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &____s->seqcount#2 irq_context: 0 (wq_completion)bond17#2 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &____s->seqcount irq_context: 0 (wq_completion)bond17#2 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex rcu_read_lock &bond->stats_lock/1 irq_context: 0 (wq_completion)bond17#2 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex rcu_read_lock &ndev->lock irq_context: 0 (wq_completion)bond17#2 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex nl_table_lock irq_context: 0 (wq_completion)bond17#2 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &obj_hash[i].lock irq_context: 0 (wq_completion)bond17#2 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex quarantine_lock irq_context: 0 (wq_completion)bond17#2 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex nl_table_wait.lock irq_context: 0 (wq_completion)bond17#2 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex net_rwsem irq_context: 0 (wq_completion)bond17#2 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex net_rwsem &list->lock#2 irq_context: 0 (wq_completion)bond17#2 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &tn->lock irq_context: 0 (wq_completion)bond17#2 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &idev->mc_lock irq_context: 0 (wq_completion)bond17#2 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex.wait_lock irq_context: 0 (wq_completion)bond17#2 (work_completion)(&(&bond->mcast_work)->work) &p->pi_lock irq_context: 0 (wq_completion)wg-crypt-wg1#5 (work_completion)(&peer->transmit_packet_work) &peer->endpoint_lock rcu_read_lock_bh rcu_read_lock &tbl->lock irq_context: 0 (wq_completion)wg-crypt-wg1#5 (work_completion)(&peer->transmit_packet_work) &peer->endpoint_lock rcu_read_lock_bh rcu_read_lock &n->lock irq_context: 0 (wq_completion)wg-crypt-wg1#5 (work_completion)(&peer->transmit_packet_work) &peer->endpoint_lock rcu_read_lock_bh rcu_read_lock &____s->seqcount#9 irq_context: 0 (wq_completion)events_power_efficient (work_completion)(&barr->work) &x->wait#10 &p->pi_lock &cfs_rq->removed.lock irq_context: 0 (wq_completion)bond32 irq_context: 0 (wq_completion)bond32 (work_completion)(&(&slave->notify_work)->work) irq_context: 0 (wq_completion)bond32 (work_completion)(&(&slave->notify_work)->work) &obj_hash[i].lock irq_context: 0 (wq_completion)bond32 (work_completion)(&(&slave->notify_work)->work) &base->lock irq_context: 0 (wq_completion)bond32 (work_completion)(&(&slave->notify_work)->work) &base->lock &obj_hash[i].lock irq_context: 0 (wq_completion)bond32 (work_completion)(&(&slave->notify_work)->work) &rq->__lock irq_context: 0 (wq_completion)bond32 (work_completion)(&(&slave->notify_work)->work) &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)bond32 (work_completion)(&(&bond->mcast_work)->work) irq_context: 0 (wq_completion)bond32 (work_completion)(&(&bond->mcast_work)->work) &obj_hash[i].lock irq_context: 0 (wq_completion)bond32 (work_completion)(&(&bond->mcast_work)->work) &base->lock irq_context: 0 (wq_completion)bond32 (work_completion)(&(&bond->mcast_work)->work) &base->lock &obj_hash[i].lock irq_context: 0 (wq_completion)bond32 (work_completion)(&(&bond->mcast_work)->work) &rq->__lock irq_context: 0 (wq_completion)bond32 (work_completion)(&(&bond->mcast_work)->work) &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)bond18#2 irq_context: 0 (wq_completion)bond18#2 (work_completion)(&(&slave->notify_work)->work) irq_context: 0 (wq_completion)bond18#2 (work_completion)(&(&slave->notify_work)->work) &obj_hash[i].lock irq_context: 0 (wq_completion)bond18#2 (work_completion)(&(&slave->notify_work)->work) &base->lock irq_context: 0 (wq_completion)bond18#2 (work_completion)(&(&slave->notify_work)->work) &base->lock &obj_hash[i].lock irq_context: 0 (wq_completion)bond18#2 (work_completion)(&(&slave->notify_work)->work) &rq->__lock irq_context: 0 (wq_completion)bond18#2 (work_completion)(&(&slave->notify_work)->work) &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)bond18#2 (work_completion)(&(&bond->mcast_work)->work) irq_context: 0 (wq_completion)bond18#2 (work_completion)(&(&bond->mcast_work)->work) &obj_hash[i].lock irq_context: 0 (wq_completion)bond18#2 (work_completion)(&(&bond->mcast_work)->work) &base->lock irq_context: 0 (wq_completion)bond18#2 (work_completion)(&(&bond->mcast_work)->work) &base->lock &obj_hash[i].lock irq_context: 0 (wq_completion)bond18#2 (work_completion)(&(&bond->mcast_work)->work) &rq->__lock irq_context: 0 (wq_completion)bond18#2 (work_completion)(&(&bond->mcast_work)->work) &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)bond18#2 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex fs_reclaim irq_context: 0 (wq_completion)bond18#2 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex fs_reclaim mmu_notifier_invalidate_range_start irq_context: 0 (wq_completion)bond18#2 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &c->lock irq_context: 0 (wq_completion)bond18#2 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex rcu_read_lock &bond->stats_lock/1 irq_context: 0 (wq_completion)bond18#2 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex rcu_read_lock &ndev->lock irq_context: 0 (wq_completion)bond18#2 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex nl_table_lock irq_context: 0 (wq_completion)bond18#2 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &obj_hash[i].lock irq_context: 0 (wq_completion)bond18#2 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex nl_table_wait.lock irq_context: 0 (wq_completion)bond18#2 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex net_rwsem irq_context: 0 (wq_completion)bond18#2 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex net_rwsem &list->lock#2 irq_context: 0 (wq_completion)bond18#2 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &tn->lock irq_context: 0 (wq_completion)bond18#2 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &idev->mc_lock irq_context: 0 (wq_completion)bond18#2 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &idev->mc_lock fs_reclaim irq_context: 0 (wq_completion)bond18#2 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &idev->mc_lock fs_reclaim mmu_notifier_invalidate_range_start irq_context: 0 (wq_completion)bond18#2 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &idev->mc_lock rcu_read_lock &ndev->lock irq_context: 0 (wq_completion)bond18#2 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &idev->mc_lock rcu_read_lock &dir->lock#2 irq_context: 0 (wq_completion)bond18#2 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &idev->mc_lock rcu_read_lock &ul->lock#2 irq_context: 0 (wq_completion)bond18#2 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &idev->mc_lock rcu_read_lock rcu_read_lock &obj_hash[i].lock irq_context: 0 (wq_completion)bond18#2 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &idev->mc_lock rcu_read_lock rcu_read_lock &c->lock irq_context: 0 (wq_completion)bond18#2 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &idev->mc_lock rcu_read_lock rcu_read_lock &dir->lock#2 irq_context: 0 (wq_completion)bond18#2 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &idev->mc_lock rcu_read_lock rcu_read_lock &tbl->lock irq_context: 0 (wq_completion)bond18#2 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &idev->mc_lock rcu_read_lock rcu_read_lock &n->lock irq_context: 0 (wq_completion)bond18#2 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &idev->mc_lock rcu_read_lock rcu_read_lock &____s->seqcount#9 irq_context: 0 (wq_completion)bond18#2 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &idev->mc_lock rcu_read_lock rcu_read_lock rcu_read_lock_bh rcu_read_lock rcu_read_lock_bh rcu_read_lock &br->multicast_lock irq_context: 0 (wq_completion)bond18#2 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &idev->mc_lock rcu_read_lock rcu_read_lock rcu_read_lock_bh rcu_read_lock rcu_read_lock_bh rcu_read_lock &br->multicast_lock &obj_hash[i].lock irq_context: 0 (wq_completion)bond18#2 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &idev->mc_lock rcu_read_lock rcu_read_lock rcu_read_lock_bh rcu_read_lock rcu_read_lock_bh rcu_read_lock &br->multicast_lock rcu_read_lock rhashtable_bucket irq_context: 0 (wq_completion)bond18#2 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &idev->mc_lock rcu_read_lock rcu_read_lock rcu_read_lock_bh rcu_read_lock rcu_read_lock_bh rcu_read_lock &br->multicast_lock &c->lock irq_context: 0 (wq_completion)bond18#2 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &idev->mc_lock rcu_read_lock rcu_read_lock rcu_read_lock_bh rcu_read_lock rcu_read_lock_bh rcu_read_lock &br->multicast_lock nl_table_lock irq_context: 0 (wq_completion)bond18#2 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &idev->mc_lock rcu_read_lock rcu_read_lock rcu_read_lock_bh rcu_read_lock rcu_read_lock_bh rcu_read_lock &br->multicast_lock nl_table_wait.lock irq_context: 0 (wq_completion)bond18#2 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &idev->mc_lock rcu_read_lock rcu_read_lock rcu_read_lock_bh rcu_read_lock rcu_read_lock_bh rcu_read_lock &br->multicast_lock &base->lock irq_context: 0 (wq_completion)bond18#2 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &idev->mc_lock rcu_read_lock rcu_read_lock rcu_read_lock_bh rcu_read_lock rcu_read_lock_bh rcu_read_lock &br->multicast_lock &base->lock &obj_hash[i].lock irq_context: 0 (wq_completion)bond18#2 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &idev->mc_lock rcu_read_lock rcu_read_lock rcu_read_lock_bh rcu_read_lock rcu_read_lock_bh rcu_read_lock &data->lock irq_context: 0 (wq_completion)bond18#2 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &idev->mc_lock rcu_read_lock rcu_read_lock rcu_read_lock_bh rcu_read_lock rcu_read_lock_bh rcu_read_lock &obj_hash[i].lock irq_context: 0 (wq_completion)bond18#2 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &idev->mc_lock rcu_read_lock rcu_read_lock rcu_read_lock_bh rcu_read_lock rcu_read_lock_bh rcu_read_lock &obj_hash[i].lock pool_lock irq_context: 0 (wq_completion)bond32 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex fs_reclaim irq_context: 0 (wq_completion)bond32 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex fs_reclaim mmu_notifier_invalidate_range_start irq_context: 0 (wq_completion)bond32 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex rcu_read_lock &bond->stats_lock/1 irq_context: 0 (wq_completion)bond32 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex rcu_read_lock &ndev->lock irq_context: 0 (wq_completion)bond32 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex nl_table_lock irq_context: 0 (wq_completion)bond32 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &obj_hash[i].lock irq_context: 0 (wq_completion)bond32 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex nl_table_wait.lock irq_context: 0 (wq_completion)bond32 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex net_rwsem irq_context: 0 (wq_completion)bond32 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex net_rwsem &list->lock#2 irq_context: 0 (wq_completion)bond32 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &tn->lock irq_context: 0 (wq_completion)bond32 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &idev->mc_lock irq_context: 0 (wq_completion)bond32 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &idev->mc_lock fs_reclaim irq_context: 0 (wq_completion)bond32 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &idev->mc_lock fs_reclaim mmu_notifier_invalidate_range_start irq_context: 0 (wq_completion)bond32 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &idev->mc_lock rcu_read_lock &ndev->lock irq_context: 0 (wq_completion)bond32 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &idev->mc_lock rcu_read_lock &dir->lock#2 irq_context: 0 (wq_completion)bond32 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &idev->mc_lock rcu_read_lock &ul->lock#2 irq_context: 0 (wq_completion)bond32 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &idev->mc_lock rcu_read_lock rcu_read_lock &obj_hash[i].lock irq_context: 0 (wq_completion)bond32 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &idev->mc_lock rcu_read_lock rcu_read_lock &c->lock irq_context: 0 (wq_completion)bond32 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &idev->mc_lock rcu_read_lock rcu_read_lock &dir->lock#2 irq_context: 0 (wq_completion)bond32 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &idev->mc_lock rcu_read_lock rcu_read_lock &tbl->lock irq_context: 0 (wq_completion)bond32 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &idev->mc_lock rcu_read_lock rcu_read_lock &n->lock irq_context: 0 (wq_completion)bond32 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &idev->mc_lock rcu_read_lock rcu_read_lock &____s->seqcount#9 irq_context: 0 (wq_completion)bond32 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &idev->mc_lock rcu_read_lock rcu_read_lock rcu_read_lock_bh rcu_read_lock rcu_read_lock_bh rcu_read_lock &br->multicast_lock irq_context: 0 (wq_completion)bond32 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &idev->mc_lock rcu_read_lock rcu_read_lock rcu_read_lock_bh rcu_read_lock rcu_read_lock_bh rcu_read_lock &br->multicast_lock &c->lock irq_context: 0 (wq_completion)bond32 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &idev->mc_lock rcu_read_lock rcu_read_lock rcu_read_lock_bh rcu_read_lock rcu_read_lock_bh rcu_read_lock &br->multicast_lock &obj_hash[i].lock irq_context: 0 (wq_completion)bond32 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &idev->mc_lock rcu_read_lock rcu_read_lock rcu_read_lock_bh rcu_read_lock rcu_read_lock_bh rcu_read_lock &br->multicast_lock rcu_read_lock rhashtable_bucket irq_context: 0 (wq_completion)bond32 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &idev->mc_lock rcu_read_lock rcu_read_lock rcu_read_lock_bh rcu_read_lock rcu_read_lock_bh rcu_read_lock &br->multicast_lock nl_table_lock irq_context: 0 (wq_completion)bond32 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &idev->mc_lock rcu_read_lock rcu_read_lock rcu_read_lock_bh rcu_read_lock rcu_read_lock_bh rcu_read_lock &br->multicast_lock nl_table_wait.lock irq_context: 0 (wq_completion)bond32 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &idev->mc_lock rcu_read_lock rcu_read_lock rcu_read_lock_bh rcu_read_lock rcu_read_lock_bh rcu_read_lock &br->multicast_lock &base->lock irq_context: 0 (wq_completion)bond32 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &idev->mc_lock rcu_read_lock rcu_read_lock rcu_read_lock_bh rcu_read_lock rcu_read_lock_bh rcu_read_lock &br->multicast_lock &base->lock &obj_hash[i].lock irq_context: 0 (wq_completion)bond32 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &idev->mc_lock rcu_read_lock rcu_read_lock rcu_read_lock_bh rcu_read_lock rcu_read_lock_bh rcu_read_lock &data->lock irq_context: 0 (wq_completion)bond32 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &idev->mc_lock rcu_read_lock rcu_read_lock rcu_read_lock_bh rcu_read_lock rcu_read_lock_bh rcu_read_lock &obj_hash[i].lock irq_context: 0 (wq_completion)bond18#2 (work_completion)(&(&slave->notify_work)->work) rtnl_mutex net_rwsem irq_context: 0 (wq_completion)bond18#2 (work_completion)(&(&slave->notify_work)->work) rtnl_mutex net_rwsem &list->lock#2 irq_context: 0 (wq_completion)bond18#2 (work_completion)(&(&slave->notify_work)->work) rtnl_mutex &tn->lock irq_context: 0 (wq_completion)bond32 (work_completion)(&(&slave->notify_work)->work) rtnl_mutex net_rwsem irq_context: 0 (wq_completion)bond32 (work_completion)(&(&slave->notify_work)->work) rtnl_mutex net_rwsem &list->lock#2 irq_context: 0 (wq_completion)bond32 (work_completion)(&(&slave->notify_work)->work) rtnl_mutex &tn->lock irq_context: 0 (wq_completion)wg-crypt-wg1#2 (work_completion)(&peer->transmit_packet_work) &peer->endpoint_lock rcu_read_lock_bh batched_entropy_u32.lock crngs.lock base_crng.lock irq_context: 0 (wq_completion)bond19#2 irq_context: 0 (wq_completion)bond19#2 (work_completion)(&(&slave->notify_work)->work) irq_context: 0 (wq_completion)bond19#2 (work_completion)(&(&slave->notify_work)->work) &obj_hash[i].lock irq_context: 0 (wq_completion)bond19#2 (work_completion)(&(&slave->notify_work)->work) &base->lock irq_context: 0 (wq_completion)bond19#2 (work_completion)(&(&slave->notify_work)->work) &base->lock &obj_hash[i].lock irq_context: 0 (wq_completion)bond19#2 (work_completion)(&(&bond->mcast_work)->work) irq_context: 0 (wq_completion)bond19#2 (work_completion)(&(&bond->mcast_work)->work) &obj_hash[i].lock irq_context: 0 (wq_completion)bond19#2 (work_completion)(&(&bond->mcast_work)->work) &base->lock irq_context: 0 (wq_completion)bond19#2 (work_completion)(&(&bond->mcast_work)->work) &base->lock &obj_hash[i].lock irq_context: 0 (wq_completion)bond19#2 (work_completion)(&(&bond->mcast_work)->work) &rq->__lock irq_context: 0 (wq_completion)bond19#2 &rq->__lock irq_context: 0 (wq_completion)bond19#2 &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)bond19#2 (work_completion)(&(&bond->mcast_work)->work) &base->lock fill_pool_map-wait-type-override pool_lock#2 irq_context: 0 (wq_completion)bond19#2 (work_completion)(&(&bond->mcast_work)->work) &base->lock fill_pool_map-wait-type-override pool_lock irq_context: 0 (wq_completion)bond19#2 (work_completion)(&(&bond->mcast_work)->work) &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)bond19#2 (work_completion)(&(&slave->notify_work)->work) rtnl_mutex net_rwsem irq_context: 0 (wq_completion)bond19#2 (work_completion)(&(&slave->notify_work)->work) rtnl_mutex net_rwsem &list->lock#2 irq_context: 0 (wq_completion)bond19#2 (work_completion)(&(&slave->notify_work)->work) rtnl_mutex &tn->lock irq_context: 0 (wq_completion)bond19#2 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex fs_reclaim irq_context: 0 (wq_completion)bond19#2 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex fs_reclaim mmu_notifier_invalidate_range_start irq_context: 0 (wq_completion)bond19#2 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex rcu_read_lock &bond->stats_lock/1 irq_context: 0 (wq_completion)bond19#2 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex rcu_read_lock &ndev->lock irq_context: 0 (wq_completion)bond19#2 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex nl_table_lock irq_context: 0 (wq_completion)bond19#2 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &obj_hash[i].lock irq_context: 0 (wq_completion)bond19#2 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex nl_table_wait.lock irq_context: 0 (wq_completion)bond19#2 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex net_rwsem irq_context: 0 (wq_completion)bond19#2 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex net_rwsem &list->lock#2 irq_context: 0 (wq_completion)bond19#2 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &tn->lock irq_context: 0 (wq_completion)bond19#2 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &idev->mc_lock irq_context: 0 (wq_completion)bond19#2 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &idev->mc_lock fs_reclaim irq_context: 0 (wq_completion)bond19#2 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &idev->mc_lock fs_reclaim mmu_notifier_invalidate_range_start irq_context: 0 (wq_completion)bond19#2 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &idev->mc_lock &c->lock irq_context: 0 (wq_completion)bond19#2 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &idev->mc_lock rcu_read_lock &ndev->lock irq_context: 0 (wq_completion)bond19#2 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &idev->mc_lock rcu_read_lock &c->lock irq_context: 0 (wq_completion)bond19#2 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &idev->mc_lock rcu_read_lock &dir->lock#2 irq_context: 0 (wq_completion)bond19#2 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &idev->mc_lock rcu_read_lock &ul->lock#2 irq_context: 0 (wq_completion)bond19#2 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &idev->mc_lock rcu_read_lock rcu_read_lock &c->lock irq_context: 0 (wq_completion)bond19#2 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &idev->mc_lock rcu_read_lock rcu_read_lock &n->list_lock irq_context: 0 (wq_completion)bond19#2 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &idev->mc_lock rcu_read_lock rcu_read_lock &n->list_lock &c->lock irq_context: 0 (wq_completion)bond19#2 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &idev->mc_lock rcu_read_lock rcu_read_lock &obj_hash[i].lock irq_context: 0 (wq_completion)bond19#2 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &idev->mc_lock rcu_read_lock rcu_read_lock &dir->lock#2 irq_context: 0 (wq_completion)bond19#2 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &idev->mc_lock rcu_read_lock rcu_read_lock &tbl->lock irq_context: 0 (wq_completion)bond19#2 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &idev->mc_lock rcu_read_lock rcu_read_lock &n->lock irq_context: 0 (wq_completion)bond19#2 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &idev->mc_lock rcu_read_lock rcu_read_lock &____s->seqcount#9 irq_context: 0 (wq_completion)bond19#2 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &idev->mc_lock rcu_read_lock rcu_read_lock rcu_read_lock_bh rcu_read_lock rcu_read_lock_bh rcu_read_lock &br->multicast_lock irq_context: 0 (wq_completion)bond19#2 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &idev->mc_lock rcu_read_lock rcu_read_lock rcu_read_lock_bh rcu_read_lock rcu_read_lock_bh rcu_read_lock &br->multicast_lock &obj_hash[i].lock irq_context: 0 (wq_completion)bond19#2 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &idev->mc_lock rcu_read_lock rcu_read_lock rcu_read_lock_bh rcu_read_lock rcu_read_lock_bh rcu_read_lock &br->multicast_lock rcu_read_lock rhashtable_bucket irq_context: 0 (wq_completion)bond19#2 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &idev->mc_lock rcu_read_lock rcu_read_lock rcu_read_lock_bh rcu_read_lock rcu_read_lock_bh rcu_read_lock &br->multicast_lock nl_table_lock irq_context: 0 (wq_completion)bond19#2 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &idev->mc_lock rcu_read_lock rcu_read_lock rcu_read_lock_bh rcu_read_lock rcu_read_lock_bh rcu_read_lock &br->multicast_lock nl_table_wait.lock irq_context: 0 (wq_completion)bond19#2 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &idev->mc_lock rcu_read_lock rcu_read_lock rcu_read_lock_bh rcu_read_lock rcu_read_lock_bh rcu_read_lock &br->multicast_lock &base->lock irq_context: 0 (wq_completion)bond19#2 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &idev->mc_lock rcu_read_lock rcu_read_lock rcu_read_lock_bh rcu_read_lock rcu_read_lock_bh rcu_read_lock &br->multicast_lock &base->lock &obj_hash[i].lock irq_context: 0 (wq_completion)bond19#2 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &idev->mc_lock rcu_read_lock rcu_read_lock rcu_read_lock_bh rcu_read_lock rcu_read_lock_bh rcu_read_lock &data->lock irq_context: 0 (wq_completion)bond19#2 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &idev->mc_lock rcu_read_lock rcu_read_lock rcu_read_lock_bh rcu_read_lock rcu_read_lock_bh rcu_read_lock &obj_hash[i].lock irq_context: softirq rcu_callback fill_pool_map-wait-type-override pool_lock#2 irq_context: softirq rcu_callback fill_pool_map-wait-type-override pool_lock irq_context: 0 rtnl_mutex rcu_state.exp_mutex fill_pool_map-wait-type-override &n->list_lock irq_context: 0 rtnl_mutex rcu_state.exp_mutex fill_pool_map-wait-type-override &n->list_lock &c->lock irq_context: 0 rtnl_mutex rcu_state.barrier_mutex rcu_read_lock &rq->__lock irq_context: 0 rtnl_mutex rcu_state.barrier_mutex rcu_read_lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)bond20#2 irq_context: 0 (wq_completion)bond20#2 (work_completion)(&(&slave->notify_work)->work) irq_context: 0 (wq_completion)bond20#2 (work_completion)(&(&slave->notify_work)->work) &obj_hash[i].lock irq_context: 0 (wq_completion)bond20#2 (work_completion)(&(&slave->notify_work)->work) &base->lock irq_context: 0 (wq_completion)bond20#2 (work_completion)(&(&slave->notify_work)->work) &base->lock &obj_hash[i].lock irq_context: 0 (wq_completion)bond20#2 (work_completion)(&(&slave->notify_work)->work) &rq->__lock irq_context: softirq (&in_dev->mr_ifc_timer) rcu_read_lock rcu_read_lock_bh rcu_read_lock quarantine_lock irq_context: 0 (wq_completion)bond20#2 (work_completion)(&(&slave->notify_work)->work) rtnl_mutex net_rwsem irq_context: 0 (wq_completion)bond20#2 (work_completion)(&(&slave->notify_work)->work) rtnl_mutex net_rwsem &list->lock#2 irq_context: 0 (wq_completion)bond20#2 (work_completion)(&(&slave->notify_work)->work) rtnl_mutex &tn->lock irq_context: 0 (wq_completion)bond20#2 (work_completion)(&(&slave->notify_work)->work) rtnl_mutex.wait_lock irq_context: 0 (wq_completion)bond20#2 (work_completion)(&(&slave->notify_work)->work) &p->pi_lock irq_context: 0 (wq_completion)bond20#2 (work_completion)(&(&bond->mcast_work)->work) irq_context: 0 (wq_completion)bond20#2 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex fs_reclaim irq_context: 0 (wq_completion)bond20#2 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex fs_reclaim mmu_notifier_invalidate_range_start irq_context: 0 (wq_completion)bond20#2 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &c->lock irq_context: 0 (wq_completion)bond20#2 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex rcu_read_lock &bond->stats_lock/1 irq_context: 0 (wq_completion)bond20#2 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex rcu_read_lock &ndev->lock irq_context: 0 (wq_completion)bond20#2 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex nl_table_lock irq_context: 0 (wq_completion)bond20#2 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &obj_hash[i].lock irq_context: 0 (wq_completion)bond20#2 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &rq->__lock irq_context: 0 (wq_completion)bond20#2 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)bond20#2 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex nl_table_wait.lock irq_context: 0 (wq_completion)bond20#2 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex net_rwsem irq_context: 0 (wq_completion)bond20#2 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex net_rwsem &list->lock#2 irq_context: 0 (wq_completion)bond20#2 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &tn->lock irq_context: 0 (wq_completion)bond20#2 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &idev->mc_lock irq_context: 0 (wq_completion)bond20#2 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex.wait_lock irq_context: 0 (wq_completion)bond20#2 (work_completion)(&(&bond->mcast_work)->work) &p->pi_lock irq_context: 0 (wq_completion)mld (work_completion)(&(&idev->mc_dad_work)->work) &idev->mc_lock rcu_read_lock rcu_read_lock rcu_read_lock_bh rcu_read_lock rcu_read_lock_bh rcu_read_lock &obj_hash[i].lock pool_lock irq_context: 0 (wq_completion)bond21#2 irq_context: 0 (wq_completion)bond21#2 (work_completion)(&(&slave->notify_work)->work) irq_context: 0 (wq_completion)bond21#2 (work_completion)(&(&slave->notify_work)->work) &obj_hash[i].lock irq_context: 0 (wq_completion)bond21#2 (work_completion)(&(&slave->notify_work)->work) &base->lock irq_context: 0 (wq_completion)bond21#2 (work_completion)(&(&slave->notify_work)->work) &base->lock &obj_hash[i].lock irq_context: 0 (wq_completion)bond21#2 (work_completion)(&(&slave->notify_work)->work) &rq->__lock irq_context: 0 (wq_completion)bond21#2 (work_completion)(&(&slave->notify_work)->work) &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)bond21#2 (work_completion)(&(&bond->mcast_work)->work) irq_context: 0 (wq_completion)bond21#2 (work_completion)(&(&bond->mcast_work)->work) &obj_hash[i].lock irq_context: 0 (wq_completion)bond21#2 (work_completion)(&(&bond->mcast_work)->work) &base->lock irq_context: 0 (wq_completion)bond21#2 (work_completion)(&(&bond->mcast_work)->work) &base->lock &obj_hash[i].lock irq_context: 0 (wq_completion)bond21#2 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex fs_reclaim irq_context: 0 (wq_completion)bond21#2 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex fs_reclaim mmu_notifier_invalidate_range_start irq_context: 0 (wq_completion)mld (work_completion)(&(&idev->mc_ifc_work)->work) &idev->mc_lock rcu_read_lock rcu_read_lock rcu_read_lock_bh rcu_read_lock rcu_read_lock_bh rcu_read_lock quarantine_lock irq_context: 0 (wq_completion)bond21#2 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &c->lock irq_context: 0 (wq_completion)bond21#2 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex remove_cache_srcu irq_context: 0 (wq_completion)bond21#2 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex remove_cache_srcu quarantine_lock irq_context: 0 (wq_completion)bond21#2 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex remove_cache_srcu &c->lock irq_context: 0 (wq_completion)bond21#2 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex remove_cache_srcu &n->list_lock irq_context: 0 (wq_completion)bond21#2 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex remove_cache_srcu rcu_read_lock pool_lock#2 irq_context: 0 (wq_completion)bond21#2 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex remove_cache_srcu rcu_read_lock rcu_node_0 irq_context: 0 (wq_completion)bond21#2 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex remove_cache_srcu rcu_read_lock &rq->__lock irq_context: 0 (wq_completion)bond21#2 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex remove_cache_srcu rcu_read_lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)bond21#2 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex remove_cache_srcu &obj_hash[i].lock irq_context: 0 (wq_completion)bond21#2 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex rcu_read_lock &bond->stats_lock/1 irq_context: 0 (wq_completion)bond21#2 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex rcu_read_lock rcu_node_0 irq_context: 0 (wq_completion)bond21#2 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex rcu_read_lock &rq->__lock irq_context: 0 (wq_completion)bond21#2 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex rcu_read_lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)bond21#2 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex rcu_read_lock &ndev->lock irq_context: 0 (wq_completion)bond21#2 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex nl_table_lock irq_context: 0 (wq_completion)bond21#2 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &rq->__lock irq_context: 0 (wq_completion)bond21#2 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)bond21#2 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &cfs_rq->removed.lock irq_context: 0 (wq_completion)bond21#2 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &obj_hash[i].lock irq_context: 0 (wq_completion)bond21#2 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex nl_table_wait.lock irq_context: 0 (wq_completion)bond21#2 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex nl_table_wait.lock &p->pi_lock irq_context: 0 (wq_completion)bond21#2 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex nl_table_wait.lock &p->pi_lock &rq->__lock irq_context: 0 (wq_completion)bond21#2 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex nl_table_wait.lock &p->pi_lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)bond21#2 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex pool_lock#2 irq_context: 0 (wq_completion)bond21#2 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex net_rwsem irq_context: 0 (wq_completion)bond21#2 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex net_rwsem &list->lock#2 irq_context: 0 (wq_completion)bond21#2 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &tn->lock irq_context: 0 (wq_completion)bond21#2 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &idev->mc_lock irq_context: 0 (wq_completion)bond21#2 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &idev->mc_lock fs_reclaim irq_context: 0 (wq_completion)bond21#2 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &idev->mc_lock fs_reclaim mmu_notifier_invalidate_range_start irq_context: 0 (wq_completion)bond21#2 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &idev->mc_lock &rq->__lock irq_context: 0 (wq_completion)bond21#2 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &idev->mc_lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)bond21#2 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &idev->mc_lock &c->lock irq_context: 0 (wq_completion)bond21#2 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &idev->mc_lock rcu_read_lock &ndev->lock irq_context: 0 (wq_completion)bond21#2 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &idev->mc_lock rcu_read_lock &c->lock irq_context: 0 (wq_completion)bond21#2 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &idev->mc_lock rcu_read_lock &dir->lock#2 irq_context: 0 (wq_completion)bond21#2 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &idev->mc_lock rcu_read_lock &ul->lock#2 irq_context: 0 (wq_completion)bond21#2 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &idev->mc_lock rcu_read_lock rcu_read_lock &____s->seqcount#10 irq_context: 0 (wq_completion)bond21#2 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &idev->mc_lock rcu_read_lock rcu_read_lock rcu_read_lock_bh rcu_read_lock rcu_read_lock_bh rcu_read_lock &br->multicast_lock irq_context: 0 (wq_completion)bond21#2 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &idev->mc_lock rcu_read_lock rcu_read_lock rcu_read_lock_bh rcu_read_lock rcu_read_lock_bh rcu_read_lock &br->multicast_lock &obj_hash[i].lock irq_context: 0 (wq_completion)bond21#2 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &idev->mc_lock rcu_read_lock rcu_read_lock rcu_read_lock_bh rcu_read_lock rcu_read_lock_bh rcu_read_lock &br->multicast_lock &base->lock irq_context: 0 (wq_completion)bond21#2 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &idev->mc_lock rcu_read_lock rcu_read_lock rcu_read_lock_bh rcu_read_lock rcu_read_lock_bh rcu_read_lock &data->lock irq_context: 0 (wq_completion)bond21#2 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &idev->mc_lock rcu_read_lock rcu_read_lock rcu_read_lock_bh rcu_read_lock rcu_read_lock_bh rcu_read_lock &obj_hash[i].lock irq_context: 0 (wq_completion)bond21#2 (work_completion)(&(&slave->notify_work)->work) rtnl_mutex net_rwsem irq_context: 0 (wq_completion)bond21#2 (work_completion)(&(&slave->notify_work)->work) rtnl_mutex net_rwsem &rq->__lock irq_context: 0 (wq_completion)bond21#2 (work_completion)(&(&slave->notify_work)->work) rtnl_mutex net_rwsem &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)bond21#2 (work_completion)(&(&slave->notify_work)->work) rtnl_mutex net_rwsem &list->lock#2 irq_context: 0 (wq_completion)bond21#2 (work_completion)(&(&slave->notify_work)->work) rtnl_mutex &tn->lock irq_context: 0 (wq_completion)bond22#2 irq_context: 0 (wq_completion)bond22#2 (work_completion)(&(&slave->notify_work)->work) irq_context: 0 (wq_completion)bond22#2 (work_completion)(&(&slave->notify_work)->work) &obj_hash[i].lock irq_context: 0 (wq_completion)bond22#2 (work_completion)(&(&slave->notify_work)->work) &base->lock irq_context: 0 (wq_completion)bond22#2 (work_completion)(&(&slave->notify_work)->work) &base->lock &obj_hash[i].lock irq_context: 0 (wq_completion)bond22#2 (work_completion)(&(&bond->mcast_work)->work) irq_context: 0 (wq_completion)bond22#2 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex fs_reclaim irq_context: 0 (wq_completion)bond22#2 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex fs_reclaim mmu_notifier_invalidate_range_start irq_context: 0 (wq_completion)bond22#2 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex rcu_read_lock &bond->stats_lock/1 irq_context: 0 (wq_completion)bond22#2 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex rcu_read_lock &ndev->lock irq_context: 0 (wq_completion)bond22#2 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex nl_table_lock irq_context: 0 (wq_completion)bond22#2 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &obj_hash[i].lock irq_context: 0 (wq_completion)bond22#2 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex nl_table_wait.lock irq_context: 0 (wq_completion)bond22#2 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex net_rwsem irq_context: 0 (wq_completion)bond22#2 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex net_rwsem &list->lock#2 irq_context: 0 (wq_completion)bond22#2 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &tn->lock irq_context: 0 (wq_completion)bond22#2 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &idev->mc_lock irq_context: 0 (wq_completion)bond22#2 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &idev->mc_lock fs_reclaim irq_context: 0 (wq_completion)bond22#2 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &idev->mc_lock fs_reclaim mmu_notifier_invalidate_range_start irq_context: 0 (wq_completion)bond22#2 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &idev->mc_lock &c->lock irq_context: 0 (wq_completion)bond22#2 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &idev->mc_lock rcu_read_lock &ndev->lock irq_context: 0 (wq_completion)bond22#2 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &idev->mc_lock rcu_read_lock &dir->lock#2 irq_context: 0 (wq_completion)bond22#2 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &idev->mc_lock rcu_read_lock &ul->lock#2 irq_context: 0 (wq_completion)bond22#2 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &idev->mc_lock rcu_read_lock rcu_read_lock &c->lock irq_context: 0 (wq_completion)bond22#2 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &idev->mc_lock rcu_read_lock rcu_read_lock &obj_hash[i].lock irq_context: 0 (wq_completion)bond22#2 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &idev->mc_lock rcu_read_lock rcu_read_lock &dir->lock#2 irq_context: 0 (wq_completion)bond22#2 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &idev->mc_lock rcu_read_lock rcu_read_lock &tbl->lock irq_context: 0 (wq_completion)bond22#2 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &idev->mc_lock rcu_read_lock rcu_read_lock &n->lock irq_context: 0 (wq_completion)bond22#2 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &idev->mc_lock rcu_read_lock rcu_read_lock &____s->seqcount#9 irq_context: 0 (wq_completion)bond22#2 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &idev->mc_lock rcu_read_lock rcu_read_lock rcu_read_lock_bh rcu_read_lock rcu_read_lock_bh rcu_read_lock &br->multicast_lock irq_context: 0 (wq_completion)bond22#2 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &idev->mc_lock rcu_read_lock rcu_read_lock rcu_read_lock_bh rcu_read_lock rcu_read_lock_bh rcu_read_lock &br->multicast_lock &obj_hash[i].lock irq_context: 0 (wq_completion)bond22#2 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &idev->mc_lock rcu_read_lock rcu_read_lock rcu_read_lock_bh rcu_read_lock rcu_read_lock_bh rcu_read_lock &br->multicast_lock rcu_read_lock rhashtable_bucket irq_context: 0 (wq_completion)bond22#2 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &idev->mc_lock rcu_read_lock rcu_read_lock rcu_read_lock_bh rcu_read_lock rcu_read_lock_bh rcu_read_lock &br->multicast_lock &c->lock irq_context: 0 (wq_completion)bond22#2 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &idev->mc_lock rcu_read_lock rcu_read_lock rcu_read_lock_bh rcu_read_lock rcu_read_lock_bh rcu_read_lock &br->multicast_lock &____s->seqcount#2 irq_context: 0 (wq_completion)bond22#2 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &idev->mc_lock rcu_read_lock rcu_read_lock rcu_read_lock_bh rcu_read_lock rcu_read_lock_bh rcu_read_lock &br->multicast_lock &____s->seqcount irq_context: 0 (wq_completion)bond22#2 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &idev->mc_lock rcu_read_lock rcu_read_lock rcu_read_lock_bh rcu_read_lock rcu_read_lock_bh rcu_read_lock &br->multicast_lock nl_table_lock irq_context: 0 (wq_completion)bond22#2 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &idev->mc_lock rcu_read_lock rcu_read_lock rcu_read_lock_bh rcu_read_lock rcu_read_lock_bh rcu_read_lock &br->multicast_lock nl_table_wait.lock irq_context: 0 (wq_completion)bond22#2 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &idev->mc_lock rcu_read_lock rcu_read_lock rcu_read_lock_bh rcu_read_lock rcu_read_lock_bh rcu_read_lock &br->multicast_lock &base->lock irq_context: 0 (wq_completion)bond22#2 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &idev->mc_lock rcu_read_lock rcu_read_lock rcu_read_lock_bh rcu_read_lock rcu_read_lock_bh rcu_read_lock &br->multicast_lock &base->lock &obj_hash[i].lock irq_context: 0 (wq_completion)bond22#2 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &idev->mc_lock rcu_read_lock rcu_read_lock rcu_read_lock_bh rcu_read_lock rcu_read_lock_bh rcu_read_lock &data->lock irq_context: 0 (wq_completion)bond22#2 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &idev->mc_lock rcu_read_lock rcu_read_lock rcu_read_lock_bh rcu_read_lock rcu_read_lock_bh rcu_read_lock &obj_hash[i].lock irq_context: 0 (wq_completion)bond22#2 (work_completion)(&(&slave->notify_work)->work) rtnl_mutex net_rwsem irq_context: 0 (wq_completion)bond22#2 (work_completion)(&(&slave->notify_work)->work) rtnl_mutex net_rwsem &list->lock#2 irq_context: 0 (wq_completion)bond22#2 (work_completion)(&(&slave->notify_work)->work) rtnl_mutex &tn->lock irq_context: 0 (wq_completion)wg-kex-wg2#9 (work_completion)(&peer->transmit_handshake_work) &peer->endpoint_lock rcu_read_lock_bh rcu_read_lock &____s->seqcount#12 irq_context: 0 (wq_completion)wg-kex-wg2#9 (work_completion)(&peer->transmit_handshake_work) &peer->endpoint_lock rcu_read_lock_bh rcu_read_lock &____s->seqcount#16 irq_context: 0 rtnl_mutex &root->kernfs_rwsem rcu_read_lock &obj_hash[i].lock irq_context: 0 (wq_completion)events (work_completion)(&(&hwstats->traffic_dw)->work) fill_pool_map-wait-type-override &c->lock irq_context: 0 sb_writers#4 &type->i_mutex_dir_key#3/1 fill_pool_map-wait-type-override rcu_read_lock rcu_node_0 irq_context: 0 sb_writers#4 &type->i_mutex_dir_key#3/1 fill_pool_map-wait-type-override rcu_read_lock &rq->__lock irq_context: 0 sb_writers#4 &type->i_mutex_dir_key#3/1 fill_pool_map-wait-type-override rcu_read_lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 &mm->mmap_lock &folio_wait_table[i] &p->pi_lock &cfs_rq->removed.lock irq_context: 0 (wq_completion)bond36 irq_context: 0 (wq_completion)bond36 (work_completion)(&(&slave->notify_work)->work) irq_context: 0 (wq_completion)bond36 (work_completion)(&(&slave->notify_work)->work) &obj_hash[i].lock irq_context: 0 (wq_completion)bond36 (work_completion)(&(&slave->notify_work)->work) &base->lock irq_context: 0 (wq_completion)bond36 (work_completion)(&(&slave->notify_work)->work) &base->lock &obj_hash[i].lock irq_context: 0 (wq_completion)bond36 (work_completion)(&(&slave->notify_work)->work) &rq->__lock irq_context: 0 (wq_completion)bond36 (work_completion)(&(&bond->mcast_work)->work) irq_context: 0 (wq_completion)bond36 (work_completion)(&(&bond->mcast_work)->work) &obj_hash[i].lock irq_context: 0 (wq_completion)bond36 (work_completion)(&(&bond->mcast_work)->work) &base->lock irq_context: 0 (wq_completion)bond36 (work_completion)(&(&bond->mcast_work)->work) &base->lock &obj_hash[i].lock irq_context: 0 (wq_completion)bond36 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex fs_reclaim irq_context: 0 (wq_completion)bond36 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex fs_reclaim mmu_notifier_invalidate_range_start irq_context: 0 (wq_completion)bond36 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex rcu_read_lock &bond->stats_lock/1 irq_context: 0 (wq_completion)bond36 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex rcu_read_lock &ndev->lock irq_context: 0 (wq_completion)bond36 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex nl_table_lock irq_context: 0 (wq_completion)bond36 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &obj_hash[i].lock irq_context: 0 (wq_completion)bond36 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex nl_table_wait.lock irq_context: 0 (wq_completion)bond36 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex net_rwsem irq_context: 0 (wq_completion)bond36 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex net_rwsem &list->lock#2 irq_context: 0 (wq_completion)bond36 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &tn->lock irq_context: 0 (wq_completion)bond36 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &idev->mc_lock irq_context: 0 (wq_completion)bond36 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &idev->mc_lock fs_reclaim irq_context: 0 (wq_completion)bond36 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &idev->mc_lock fs_reclaim mmu_notifier_invalidate_range_start irq_context: 0 (wq_completion)bond36 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &idev->mc_lock rcu_read_lock &ndev->lock irq_context: 0 (wq_completion)bond36 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &idev->mc_lock rcu_read_lock &dir->lock#2 irq_context: 0 (wq_completion)bond36 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &idev->mc_lock rcu_read_lock &ul->lock#2 irq_context: 0 (wq_completion)bond36 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &idev->mc_lock rcu_read_lock rcu_read_lock &c->lock irq_context: 0 (wq_completion)bond36 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &idev->mc_lock rcu_read_lock rcu_read_lock &obj_hash[i].lock irq_context: 0 (wq_completion)bond36 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &idev->mc_lock rcu_read_lock rcu_read_lock &dir->lock#2 irq_context: 0 (wq_completion)bond36 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &idev->mc_lock rcu_read_lock rcu_read_lock &tbl->lock irq_context: 0 (wq_completion)bond36 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &idev->mc_lock rcu_read_lock rcu_read_lock &n->lock irq_context: 0 (wq_completion)bond36 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &idev->mc_lock rcu_read_lock rcu_read_lock &____s->seqcount#9 irq_context: 0 (wq_completion)bond36 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &idev->mc_lock rcu_read_lock rcu_read_lock rcu_read_lock_bh rcu_read_lock rcu_read_lock_bh rcu_read_lock &br->multicast_lock irq_context: 0 (wq_completion)bond36 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &idev->mc_lock rcu_read_lock rcu_read_lock rcu_read_lock_bh rcu_read_lock rcu_read_lock_bh rcu_read_lock &br->multicast_lock &obj_hash[i].lock irq_context: 0 (wq_completion)bond36 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &idev->mc_lock rcu_read_lock rcu_read_lock rcu_read_lock_bh rcu_read_lock rcu_read_lock_bh rcu_read_lock &br->multicast_lock &obj_hash[i].lock pool_lock irq_context: 0 (wq_completion)bond36 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &idev->mc_lock rcu_read_lock rcu_read_lock rcu_read_lock_bh rcu_read_lock rcu_read_lock_bh rcu_read_lock &br->multicast_lock rcu_read_lock rhashtable_bucket irq_context: 0 (wq_completion)bond36 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &idev->mc_lock rcu_read_lock rcu_read_lock rcu_read_lock_bh rcu_read_lock rcu_read_lock_bh rcu_read_lock &br->multicast_lock &c->lock irq_context: 0 (wq_completion)bond36 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &idev->mc_lock rcu_read_lock rcu_read_lock rcu_read_lock_bh rcu_read_lock rcu_read_lock_bh rcu_read_lock &br->multicast_lock nl_table_lock irq_context: 0 (wq_completion)bond36 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &idev->mc_lock rcu_read_lock rcu_read_lock rcu_read_lock_bh rcu_read_lock rcu_read_lock_bh rcu_read_lock &br->multicast_lock nl_table_wait.lock irq_context: 0 (wq_completion)bond36 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &idev->mc_lock rcu_read_lock rcu_read_lock rcu_read_lock_bh rcu_read_lock rcu_read_lock_bh rcu_read_lock &br->multicast_lock &base->lock irq_context: 0 (wq_completion)bond36 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &idev->mc_lock rcu_read_lock rcu_read_lock rcu_read_lock_bh rcu_read_lock rcu_read_lock_bh rcu_read_lock &br->multicast_lock &base->lock &obj_hash[i].lock irq_context: 0 (wq_completion)bond36 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &idev->mc_lock rcu_read_lock rcu_read_lock rcu_read_lock_bh rcu_read_lock rcu_read_lock_bh rcu_read_lock &data->lock irq_context: 0 (wq_completion)bond36 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &idev->mc_lock rcu_read_lock rcu_read_lock rcu_read_lock_bh rcu_read_lock rcu_read_lock_bh rcu_read_lock &obj_hash[i].lock irq_context: 0 (wq_completion)bond36 (work_completion)(&(&slave->notify_work)->work) rtnl_mutex net_rwsem irq_context: 0 (wq_completion)bond36 (work_completion)(&(&slave->notify_work)->work) rtnl_mutex net_rwsem &list->lock#2 irq_context: 0 (wq_completion)bond36 (work_completion)(&(&slave->notify_work)->work) rtnl_mutex &tn->lock irq_context: 0 sk_lock-AF_INET &sighand->siglock &p->pi_lock &cfs_rq->removed.lock irq_context: 0 (wq_completion)bond37 irq_context: 0 (wq_completion)bond37 (work_completion)(&(&slave->notify_work)->work) irq_context: 0 (wq_completion)bond37 (work_completion)(&(&slave->notify_work)->work) &obj_hash[i].lock irq_context: 0 (wq_completion)bond37 (work_completion)(&(&slave->notify_work)->work) &base->lock irq_context: 0 (wq_completion)bond37 (work_completion)(&(&slave->notify_work)->work) &base->lock &obj_hash[i].lock irq_context: 0 (wq_completion)bond37 (work_completion)(&(&slave->notify_work)->work) &rq->__lock irq_context: 0 (wq_completion)bond37 (work_completion)(&(&bond->mcast_work)->work) irq_context: 0 (wq_completion)bond37 (work_completion)(&(&bond->mcast_work)->work) &obj_hash[i].lock irq_context: 0 (wq_completion)bond37 (work_completion)(&(&bond->mcast_work)->work) &base->lock irq_context: 0 (wq_completion)bond37 (work_completion)(&(&bond->mcast_work)->work) &base->lock &obj_hash[i].lock irq_context: 0 (wq_completion)bond37 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex fs_reclaim irq_context: 0 (wq_completion)bond37 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex fs_reclaim mmu_notifier_invalidate_range_start irq_context: 0 (wq_completion)bond37 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex rcu_read_lock &bond->stats_lock/1 irq_context: 0 (wq_completion)bond37 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex rcu_read_lock &ndev->lock irq_context: 0 (wq_completion)bond37 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex nl_table_lock irq_context: 0 (wq_completion)bond37 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &obj_hash[i].lock irq_context: 0 (wq_completion)bond37 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex nl_table_wait.lock irq_context: 0 (wq_completion)bond37 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex net_rwsem irq_context: 0 (wq_completion)bond37 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex net_rwsem &list->lock#2 irq_context: 0 (wq_completion)bond37 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &tn->lock irq_context: 0 (wq_completion)bond37 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &idev->mc_lock irq_context: 0 (wq_completion)bond37 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &idev->mc_lock fs_reclaim irq_context: 0 (wq_completion)bond37 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &idev->mc_lock fs_reclaim mmu_notifier_invalidate_range_start irq_context: 0 (wq_completion)bond37 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &idev->mc_lock rcu_read_lock &ndev->lock irq_context: 0 (wq_completion)bond37 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &idev->mc_lock rcu_read_lock &dir->lock#2 irq_context: 0 (wq_completion)bond37 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &idev->mc_lock rcu_read_lock &ul->lock#2 irq_context: 0 (wq_completion)bond37 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &idev->mc_lock rcu_read_lock rcu_read_lock &c->lock irq_context: 0 (wq_completion)bond37 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &idev->mc_lock rcu_read_lock rcu_read_lock &n->list_lock irq_context: 0 (wq_completion)bond37 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &idev->mc_lock rcu_read_lock rcu_read_lock &n->list_lock &c->lock irq_context: 0 (wq_completion)bond37 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &idev->mc_lock rcu_read_lock rcu_read_lock &obj_hash[i].lock irq_context: 0 (wq_completion)bond37 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &idev->mc_lock rcu_read_lock rcu_read_lock &dir->lock#2 irq_context: 0 (wq_completion)bond37 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &idev->mc_lock rcu_read_lock rcu_read_lock &tbl->lock irq_context: 0 (wq_completion)bond37 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &idev->mc_lock rcu_read_lock rcu_read_lock &n->lock irq_context: 0 (wq_completion)bond37 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &idev->mc_lock rcu_read_lock rcu_read_lock &____s->seqcount#9 irq_context: 0 (wq_completion)bond37 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &idev->mc_lock rcu_read_lock rcu_read_lock rcu_read_lock_bh rcu_read_lock rcu_read_lock_bh rcu_read_lock &br->multicast_lock irq_context: 0 (wq_completion)bond37 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &idev->mc_lock rcu_read_lock rcu_read_lock rcu_read_lock_bh rcu_read_lock rcu_read_lock_bh rcu_read_lock &br->multicast_lock &obj_hash[i].lock irq_context: 0 (wq_completion)bond37 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &idev->mc_lock rcu_read_lock rcu_read_lock rcu_read_lock_bh rcu_read_lock rcu_read_lock_bh rcu_read_lock &br->multicast_lock rcu_read_lock rhashtable_bucket irq_context: 0 (wq_completion)bond37 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &idev->mc_lock rcu_read_lock rcu_read_lock rcu_read_lock_bh rcu_read_lock rcu_read_lock_bh rcu_read_lock &br->multicast_lock nl_table_lock irq_context: 0 (wq_completion)bond37 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &idev->mc_lock rcu_read_lock rcu_read_lock rcu_read_lock_bh rcu_read_lock rcu_read_lock_bh rcu_read_lock &br->multicast_lock nl_table_wait.lock irq_context: 0 (wq_completion)bond37 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &idev->mc_lock rcu_read_lock rcu_read_lock rcu_read_lock_bh rcu_read_lock rcu_read_lock_bh rcu_read_lock &br->multicast_lock &base->lock irq_context: 0 (wq_completion)bond37 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &idev->mc_lock rcu_read_lock rcu_read_lock rcu_read_lock_bh rcu_read_lock rcu_read_lock_bh rcu_read_lock &br->multicast_lock &base->lock &obj_hash[i].lock irq_context: 0 (wq_completion)bond37 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &idev->mc_lock rcu_read_lock rcu_read_lock rcu_read_lock_bh rcu_read_lock rcu_read_lock_bh rcu_read_lock &data->lock irq_context: 0 (wq_completion)bond37 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &idev->mc_lock rcu_read_lock rcu_read_lock rcu_read_lock_bh rcu_read_lock rcu_read_lock_bh rcu_read_lock &obj_hash[i].lock irq_context: 0 (wq_completion)bond37 (work_completion)(&(&slave->notify_work)->work) rtnl_mutex net_rwsem irq_context: 0 (wq_completion)bond37 (work_completion)(&(&slave->notify_work)->work) rtnl_mutex net_rwsem &list->lock#2 irq_context: 0 (wq_completion)bond37 (work_completion)(&(&slave->notify_work)->work) rtnl_mutex &tn->lock irq_context: 0 (wq_completion)bond38 irq_context: 0 (wq_completion)bond38 (work_completion)(&(&slave->notify_work)->work) irq_context: 0 (wq_completion)bond38 (work_completion)(&(&slave->notify_work)->work) &obj_hash[i].lock irq_context: 0 (wq_completion)bond38 (work_completion)(&(&slave->notify_work)->work) &base->lock irq_context: 0 (wq_completion)bond38 (work_completion)(&(&slave->notify_work)->work) &base->lock &obj_hash[i].lock irq_context: 0 (wq_completion)bond38 (work_completion)(&(&bond->mcast_work)->work) irq_context: 0 (wq_completion)bond38 (work_completion)(&(&bond->mcast_work)->work) &obj_hash[i].lock irq_context: 0 (wq_completion)bond38 (work_completion)(&(&bond->mcast_work)->work) &base->lock irq_context: 0 (wq_completion)bond38 (work_completion)(&(&bond->mcast_work)->work) &base->lock &obj_hash[i].lock irq_context: 0 (wq_completion)bond38 (work_completion)(&(&slave->notify_work)->work) rtnl_mutex net_rwsem irq_context: 0 (wq_completion)bond38 (work_completion)(&(&slave->notify_work)->work) rtnl_mutex net_rwsem &list->lock#2 irq_context: 0 (wq_completion)bond38 (work_completion)(&(&slave->notify_work)->work) rtnl_mutex &tn->lock irq_context: 0 (wq_completion)bond38 (work_completion)(&(&slave->notify_work)->work) rtnl_mutex.wait_lock irq_context: 0 (wq_completion)bond38 (work_completion)(&(&slave->notify_work)->work) &p->pi_lock irq_context: 0 (wq_completion)bond38 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex fs_reclaim irq_context: 0 (wq_completion)bond38 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex fs_reclaim mmu_notifier_invalidate_range_start irq_context: 0 (wq_completion)bond38 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &c->lock irq_context: 0 (wq_completion)bond38 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &n->list_lock irq_context: 0 (wq_completion)bond38 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &n->list_lock &c->lock irq_context: 0 (wq_completion)bond38 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex rcu_read_lock &bond->stats_lock/1 irq_context: 0 (wq_completion)bond38 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex rcu_read_lock &ndev->lock irq_context: 0 (wq_completion)bond38 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex nl_table_lock irq_context: 0 (wq_completion)bond38 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &obj_hash[i].lock irq_context: 0 (wq_completion)bond38 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex nl_table_wait.lock irq_context: 0 (wq_completion)bond38 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex net_rwsem irq_context: 0 (wq_completion)bond38 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex net_rwsem &list->lock#2 irq_context: 0 (wq_completion)bond38 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &tn->lock irq_context: 0 (wq_completion)bond38 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &idev->mc_lock irq_context: 0 (wq_completion)bond38 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex.wait_lock irq_context: 0 (wq_completion)bond38 (work_completion)(&(&bond->mcast_work)->work) &p->pi_lock irq_context: 0 (wq_completion)bond2#2 irq_context: 0 (wq_completion)bond2#2 (work_completion)(&(&slave->notify_work)->work) irq_context: 0 (wq_completion)bond2#2 (work_completion)(&(&slave->notify_work)->work) &rq->__lock irq_context: 0 (wq_completion)bond2#2 (work_completion)(&(&slave->notify_work)->work) &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)bond2#2 (work_completion)(&(&slave->notify_work)->work) &obj_hash[i].lock irq_context: 0 (wq_completion)bond2#2 (work_completion)(&(&slave->notify_work)->work) &base->lock irq_context: 0 (wq_completion)bond2#2 (work_completion)(&(&slave->notify_work)->work) &base->lock &obj_hash[i].lock irq_context: 0 (wq_completion)bond2#2 (work_completion)(&(&bond->mcast_work)->work) irq_context: 0 (wq_completion)bond2#2 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex fs_reclaim irq_context: 0 (wq_completion)bond2#2 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex fs_reclaim mmu_notifier_invalidate_range_start irq_context: 0 (wq_completion)bond2#2 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex rcu_read_lock &bond->stats_lock/1 irq_context: 0 (wq_completion)bond2#2 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex rcu_read_lock &ndev->lock irq_context: 0 (wq_completion)bond2#2 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex nl_table_lock irq_context: 0 (wq_completion)bond2#2 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &obj_hash[i].lock irq_context: 0 (wq_completion)bond2#2 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex nl_table_wait.lock irq_context: 0 (wq_completion)bond2#2 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex net_rwsem irq_context: 0 (wq_completion)bond2#2 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex net_rwsem &list->lock#2 irq_context: 0 (wq_completion)bond2#2 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &tn->lock irq_context: 0 (wq_completion)bond2#2 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &idev->mc_lock irq_context: 0 (wq_completion)bond2#2 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex.wait_lock irq_context: 0 (wq_completion)bond2#2 (work_completion)(&(&bond->mcast_work)->work) &p->pi_lock irq_context: 0 (wq_completion)bond2#2 (work_completion)(&(&slave->notify_work)->work) rtnl_mutex net_rwsem irq_context: 0 (wq_completion)bond2#2 (work_completion)(&(&slave->notify_work)->work) rtnl_mutex net_rwsem &list->lock#2 irq_context: 0 (wq_completion)bond2#2 (work_completion)(&(&slave->notify_work)->work) rtnl_mutex &tn->lock irq_context: 0 &mm->mmap_lock fs_reclaim mmu_notifier_invalidate_range_start pgd_lock irq_context: 0 &mm->mmap_lock fs_reclaim mmu_notifier_invalidate_range_start stock_lock irq_context: 0 &mm->mmap_lock fs_reclaim mmu_notifier_invalidate_range_start key irq_context: 0 &mm->mmap_lock fs_reclaim mmu_notifier_invalidate_range_start pcpu_lock irq_context: 0 &mm->mmap_lock fs_reclaim mmu_notifier_invalidate_range_start percpu_counters_lock irq_context: 0 &mm->mmap_lock fs_reclaim mmu_notifier_invalidate_range_start pcpu_lock stock_lock irq_context: 0 (wq_completion)mld (work_completion)(&(&idev->mc_ifc_work)->work) &idev->mc_lock rcu_read_lock rcu_read_lock rcu_read_lock_bh rcu_read_lock rcu_read_lock_bh rcu_read_lock &br->multicast_lock &n->list_lock irq_context: 0 (wq_completion)mld (work_completion)(&(&idev->mc_ifc_work)->work) &idev->mc_lock rcu_read_lock rcu_read_lock rcu_read_lock_bh rcu_read_lock rcu_read_lock_bh rcu_read_lock &br->multicast_lock &n->list_lock &c->lock irq_context: 0 rtnl_mutex &dev_addr_list_lock_key &____s->seqcount#2 irq_context: 0 rtnl_mutex &dev_addr_list_lock_key &____s->seqcount irq_context: 0 &f->f_pos_lock &type->i_mutex_dir_key#3 fs_reclaim mmu_notifier_invalidate_range_start &rq->__lock &cfs_rq->removed.lock irq_context: 0 (wq_completion)bond3#3 irq_context: 0 (wq_completion)bond3#3 (work_completion)(&(&slave->notify_work)->work) irq_context: 0 (wq_completion)bond3#3 (work_completion)(&(&slave->notify_work)->work) &obj_hash[i].lock irq_context: 0 (wq_completion)bond3#3 (work_completion)(&(&slave->notify_work)->work) &base->lock irq_context: 0 (wq_completion)bond3#3 (work_completion)(&(&slave->notify_work)->work) &base->lock &obj_hash[i].lock irq_context: 0 (wq_completion)bond3#3 (work_completion)(&(&slave->notify_work)->work) &rq->__lock irq_context: 0 (wq_completion)bond3#3 (work_completion)(&(&bond->mcast_work)->work) irq_context: 0 (wq_completion)bond3#3 (work_completion)(&(&bond->mcast_work)->work) &rq->__lock irq_context: 0 (wq_completion)bond3#3 (work_completion)(&(&bond->mcast_work)->work) &obj_hash[i].lock irq_context: 0 (wq_completion)bond3#3 (work_completion)(&(&bond->mcast_work)->work) &base->lock irq_context: 0 (wq_completion)bond3#3 (work_completion)(&(&bond->mcast_work)->work) &base->lock &obj_hash[i].lock irq_context: 0 (wq_completion)bond26#2 irq_context: 0 (wq_completion)bond26#2 (work_completion)(&(&slave->notify_work)->work) irq_context: 0 (wq_completion)bond26#2 (work_completion)(&(&slave->notify_work)->work) &obj_hash[i].lock irq_context: 0 (wq_completion)bond26#2 (work_completion)(&(&slave->notify_work)->work) &base->lock irq_context: 0 (wq_completion)bond26#2 (work_completion)(&(&slave->notify_work)->work) &base->lock &obj_hash[i].lock irq_context: 0 (wq_completion)bond26#2 (work_completion)(&(&slave->notify_work)->work) &rq->__lock irq_context: 0 (wq_completion)bond26#2 (work_completion)(&(&slave->notify_work)->work) &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)bond26#2 (work_completion)(&(&bond->mcast_work)->work) irq_context: 0 (wq_completion)bond26#2 (work_completion)(&(&bond->mcast_work)->work) &obj_hash[i].lock irq_context: 0 (wq_completion)bond26#2 (work_completion)(&(&bond->mcast_work)->work) &base->lock irq_context: 0 (wq_completion)bond26#2 (work_completion)(&(&bond->mcast_work)->work) &base->lock &obj_hash[i].lock irq_context: 0 (wq_completion)bond26#2 (work_completion)(&(&bond->mcast_work)->work) &rq->__lock irq_context: 0 (wq_completion)bond26#2 (work_completion)(&(&bond->mcast_work)->work) &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)bond4#2 irq_context: 0 (wq_completion)bond4#2 (work_completion)(&(&slave->notify_work)->work) irq_context: 0 (wq_completion)bond4#2 (work_completion)(&(&slave->notify_work)->work) &obj_hash[i].lock irq_context: 0 (wq_completion)bond4#2 (work_completion)(&(&slave->notify_work)->work) &base->lock irq_context: 0 (wq_completion)bond4#2 (work_completion)(&(&slave->notify_work)->work) &base->lock &obj_hash[i].lock irq_context: 0 (wq_completion)bond4#2 (work_completion)(&(&slave->notify_work)->work) &rq->__lock irq_context: 0 (wq_completion)bond4#2 (work_completion)(&(&slave->notify_work)->work) &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)bond26#2 (work_completion)(&(&bond->mcast_work)->work) fill_pool_map-wait-type-override pool_lock#2 irq_context: softirq (&in_dev->mr_ifc_timer) rcu_read_lock rcu_read_lock_bh rcu_read_lock fill_pool_map-wait-type-override pool_lock#2 irq_context: 0 (wq_completion)bond26#2 (work_completion)(&(&bond->mcast_work)->work) fill_pool_map-wait-type-override pool_lock irq_context: softirq (&in_dev->mr_ifc_timer) rcu_read_lock rcu_read_lock_bh rcu_read_lock fill_pool_map-wait-type-override pool_lock irq_context: 0 (wq_completion)bond4#2 (work_completion)(&(&bond->mcast_work)->work) irq_context: 0 (wq_completion)bond4#2 (work_completion)(&(&bond->mcast_work)->work) &obj_hash[i].lock irq_context: 0 (wq_completion)bond4#2 (work_completion)(&(&bond->mcast_work)->work) &base->lock irq_context: 0 (wq_completion)bond4#2 (work_completion)(&(&bond->mcast_work)->work) &base->lock &obj_hash[i].lock irq_context: 0 sb_writers#4 &type->i_mutex_dir_key#3/1 sb_internal jbd2_handle rcu_read_lock &rcu_state.expedited_wq irq_context: 0 sb_writers#4 &type->i_mutex_dir_key#3/1 sb_internal jbd2_handle rcu_read_lock &rcu_state.expedited_wq &p->pi_lock irq_context: 0 sb_writers#4 &type->i_mutex_dir_key#3/1 sb_internal jbd2_handle rcu_read_lock &rcu_state.expedited_wq &p->pi_lock &rq->__lock irq_context: 0 sb_writers#4 &type->i_mutex_dir_key#3/1 sb_internal jbd2_handle rcu_read_lock &rcu_state.expedited_wq &p->pi_lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)bond27#2 irq_context: 0 (wq_completion)bond27#2 (work_completion)(&(&slave->notify_work)->work) irq_context: 0 (wq_completion)bond27#2 (work_completion)(&(&slave->notify_work)->work) &obj_hash[i].lock irq_context: 0 (wq_completion)bond27#2 (work_completion)(&(&slave->notify_work)->work) &base->lock irq_context: 0 (wq_completion)bond27#2 (work_completion)(&(&slave->notify_work)->work) &base->lock &obj_hash[i].lock irq_context: 0 (wq_completion)bond27#2 (work_completion)(&(&slave->notify_work)->work) &rq->__lock irq_context: 0 (wq_completion)bond27#2 (work_completion)(&(&bond->mcast_work)->work) irq_context: 0 (wq_completion)bond27#2 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex fs_reclaim irq_context: 0 (wq_completion)bond27#2 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex fs_reclaim mmu_notifier_invalidate_range_start irq_context: 0 (wq_completion)bond27#2 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex batched_entropy_u8.lock irq_context: 0 (wq_completion)bond27#2 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex kfence_freelist_lock irq_context: 0 (wq_completion)bond27#2 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex rcu_read_lock &bond->stats_lock/1 irq_context: 0 (wq_completion)bond27#2 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex rcu_read_lock &ndev->lock irq_context: 0 (wq_completion)bond27#2 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex nl_table_lock irq_context: 0 (wq_completion)bond27#2 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &obj_hash[i].lock irq_context: 0 (wq_completion)bond27#2 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &meta->lock irq_context: 0 (wq_completion)bond27#2 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex nl_table_wait.lock irq_context: 0 (wq_completion)bond27#2 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex net_rwsem irq_context: 0 (wq_completion)bond27#2 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex net_rwsem &list->lock#2 irq_context: 0 (wq_completion)bond27#2 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &tn->lock irq_context: 0 (wq_completion)bond27#2 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &idev->mc_lock irq_context: 0 (wq_completion)bond27#2 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &idev->mc_lock fs_reclaim irq_context: 0 (wq_completion)bond27#2 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &idev->mc_lock fs_reclaim mmu_notifier_invalidate_range_start irq_context: 0 (wq_completion)bond27#2 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &idev->mc_lock rcu_read_lock &ndev->lock irq_context: 0 (wq_completion)bond27#2 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &idev->mc_lock rcu_read_lock &dir->lock#2 irq_context: 0 (wq_completion)bond27#2 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &idev->mc_lock rcu_read_lock &ul->lock#2 irq_context: 0 (wq_completion)bond27#2 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &idev->mc_lock rcu_read_lock rcu_read_lock &c->lock irq_context: 0 (wq_completion)bond27#2 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &idev->mc_lock rcu_read_lock rcu_read_lock &obj_hash[i].lock irq_context: 0 (wq_completion)bond27#2 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &idev->mc_lock rcu_read_lock rcu_read_lock &dir->lock#2 irq_context: 0 (wq_completion)bond27#2 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &idev->mc_lock rcu_read_lock rcu_read_lock &tbl->lock irq_context: 0 (wq_completion)bond27#2 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &idev->mc_lock rcu_read_lock rcu_read_lock &n->lock irq_context: 0 (wq_completion)bond27#2 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &idev->mc_lock rcu_read_lock rcu_read_lock &____s->seqcount#9 irq_context: 0 (wq_completion)bond27#2 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &idev->mc_lock rcu_read_lock rcu_read_lock rcu_read_lock_bh rcu_read_lock rcu_read_lock_bh rcu_read_lock &br->multicast_lock irq_context: 0 (wq_completion)bond27#2 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &idev->mc_lock rcu_read_lock rcu_read_lock rcu_read_lock_bh rcu_read_lock rcu_read_lock_bh rcu_read_lock &br->multicast_lock &obj_hash[i].lock irq_context: 0 (wq_completion)bond27#2 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &idev->mc_lock rcu_read_lock rcu_read_lock rcu_read_lock_bh rcu_read_lock rcu_read_lock_bh rcu_read_lock &br->multicast_lock rcu_read_lock rhashtable_bucket irq_context: 0 (wq_completion)bond27#2 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &idev->mc_lock rcu_read_lock rcu_read_lock rcu_read_lock_bh rcu_read_lock rcu_read_lock_bh rcu_read_lock &br->multicast_lock &c->lock irq_context: 0 (wq_completion)bond27#2 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &idev->mc_lock rcu_read_lock rcu_read_lock rcu_read_lock_bh rcu_read_lock rcu_read_lock_bh rcu_read_lock &br->multicast_lock nl_table_lock irq_context: 0 (wq_completion)bond27#2 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &idev->mc_lock rcu_read_lock rcu_read_lock rcu_read_lock_bh rcu_read_lock rcu_read_lock_bh rcu_read_lock &br->multicast_lock nl_table_wait.lock irq_context: 0 (wq_completion)bond27#2 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &idev->mc_lock rcu_read_lock rcu_read_lock rcu_read_lock_bh rcu_read_lock rcu_read_lock_bh rcu_read_lock &br->multicast_lock &base->lock irq_context: 0 (wq_completion)bond27#2 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &idev->mc_lock rcu_read_lock rcu_read_lock rcu_read_lock_bh rcu_read_lock rcu_read_lock_bh rcu_read_lock &br->multicast_lock &base->lock &obj_hash[i].lock irq_context: 0 (wq_completion)bond27#2 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &idev->mc_lock rcu_read_lock rcu_read_lock rcu_read_lock_bh rcu_read_lock rcu_read_lock_bh rcu_read_lock &data->lock irq_context: 0 (wq_completion)bond27#2 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &idev->mc_lock rcu_read_lock rcu_read_lock rcu_read_lock_bh rcu_read_lock rcu_read_lock_bh rcu_read_lock &obj_hash[i].lock irq_context: 0 (wq_completion)bond27#2 (work_completion)(&(&slave->notify_work)->work) rtnl_mutex net_rwsem irq_context: 0 (wq_completion)bond27#2 (work_completion)(&(&slave->notify_work)->work) rtnl_mutex net_rwsem &list->lock#2 irq_context: 0 (wq_completion)bond27#2 (work_completion)(&(&slave->notify_work)->work) rtnl_mutex &tn->lock irq_context: 0 (wq_completion)bond4#2 (work_completion)(&(&slave->notify_work)->work) rtnl_mutex net_rwsem irq_context: 0 (wq_completion)bond4#2 (work_completion)(&(&slave->notify_work)->work) rtnl_mutex net_rwsem &list->lock#2 irq_context: 0 (wq_completion)bond4#2 (work_completion)(&(&slave->notify_work)->work) rtnl_mutex &tn->lock irq_context: 0 (wq_completion)bond26#2 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex fs_reclaim irq_context: 0 (wq_completion)bond26#2 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex fs_reclaim mmu_notifier_invalidate_range_start irq_context: 0 (wq_completion)bond26#2 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex fs_reclaim mmu_notifier_invalidate_range_start &rq->__lock irq_context: 0 (wq_completion)bond26#2 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex fs_reclaim mmu_notifier_invalidate_range_start &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)bond26#2 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex rcu_read_lock &bond->stats_lock/1 irq_context: 0 (wq_completion)bond26#2 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex rcu_read_lock &ndev->lock irq_context: 0 (wq_completion)bond26#2 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex nl_table_lock irq_context: 0 (wq_completion)bond26#2 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &obj_hash[i].lock irq_context: 0 (wq_completion)bond26#2 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex nl_table_wait.lock irq_context: 0 (wq_completion)bond26#2 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex net_rwsem irq_context: 0 (wq_completion)bond26#2 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex net_rwsem &list->lock#2 irq_context: 0 (wq_completion)bond26#2 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &tn->lock irq_context: 0 (wq_completion)bond26#2 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &idev->mc_lock irq_context: 0 (wq_completion)bond26#2 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &idev->mc_lock fs_reclaim irq_context: 0 (wq_completion)bond26#2 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &idev->mc_lock fs_reclaim mmu_notifier_invalidate_range_start irq_context: 0 (wq_completion)bond26#2 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &idev->mc_lock fs_reclaim mmu_notifier_invalidate_range_start &rq->__lock irq_context: 0 (wq_completion)bond26#2 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &idev->mc_lock fs_reclaim mmu_notifier_invalidate_range_start &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)bond26#2 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &idev->mc_lock rcu_read_lock &ndev->lock irq_context: 0 (wq_completion)bond26#2 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &idev->mc_lock rcu_read_lock &c->lock irq_context: 0 (wq_completion)bond26#2 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &idev->mc_lock rcu_read_lock &dir->lock#2 irq_context: 0 (wq_completion)bond26#2 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &idev->mc_lock rcu_read_lock &ul->lock#2 irq_context: 0 (wq_completion)bond26#2 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &idev->mc_lock rcu_read_lock rcu_read_lock &____s->seqcount#10 irq_context: 0 (wq_completion)bond26#2 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &idev->mc_lock rcu_read_lock rcu_read_lock rcu_read_lock_bh rcu_read_lock rcu_read_lock_bh rcu_read_lock &br->multicast_lock irq_context: 0 (wq_completion)bond26#2 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &idev->mc_lock rcu_read_lock rcu_read_lock rcu_read_lock_bh rcu_read_lock rcu_read_lock_bh rcu_read_lock &br->multicast_lock &obj_hash[i].lock irq_context: 0 (wq_completion)bond26#2 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &idev->mc_lock rcu_read_lock rcu_read_lock rcu_read_lock_bh rcu_read_lock rcu_read_lock_bh rcu_read_lock &br->multicast_lock &base->lock irq_context: 0 (wq_completion)bond26#2 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &idev->mc_lock rcu_read_lock rcu_read_lock rcu_read_lock_bh rcu_read_lock rcu_read_lock_bh rcu_read_lock &br->multicast_lock &base->lock &obj_hash[i].lock irq_context: 0 (wq_completion)bond26#2 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &idev->mc_lock rcu_read_lock rcu_read_lock rcu_read_lock_bh rcu_read_lock rcu_read_lock_bh rcu_read_lock &data->lock irq_context: 0 (wq_completion)bond26#2 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &idev->mc_lock rcu_read_lock rcu_read_lock rcu_read_lock_bh rcu_read_lock rcu_read_lock_bh rcu_read_lock &obj_hash[i].lock irq_context: 0 (wq_completion)bond26#2 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &idev->mc_lock rcu_read_lock rcu_read_lock rcu_node_0 irq_context: 0 (wq_completion)bond26#2 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &idev->mc_lock rcu_read_lock rcu_read_lock &rq->__lock irq_context: 0 (wq_completion)bond26#2 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &idev->mc_lock rcu_read_lock rcu_read_lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)bond26#2 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &idev->mc_lock rcu_read_lock rcu_node_0 irq_context: 0 (wq_completion)bond26#2 (work_completion)(&(&slave->notify_work)->work) rtnl_mutex net_rwsem irq_context: 0 (wq_completion)bond26#2 (work_completion)(&(&slave->notify_work)->work) rtnl_mutex net_rwsem &list->lock#2 irq_context: 0 (wq_completion)bond26#2 (work_completion)(&(&slave->notify_work)->work) rtnl_mutex &tn->lock irq_context: 0 (wq_completion)bond3#3 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex fs_reclaim irq_context: 0 (wq_completion)bond3#3 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex fs_reclaim mmu_notifier_invalidate_range_start irq_context: 0 (wq_completion)bond3#3 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &c->lock irq_context: 0 (wq_completion)bond3#3 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex rcu_read_lock &bond->stats_lock/1 irq_context: 0 (wq_completion)bond3#3 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex rcu_read_lock &ndev->lock irq_context: 0 (wq_completion)bond3#3 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex nl_table_lock irq_context: 0 (wq_completion)bond3#3 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &obj_hash[i].lock irq_context: 0 (wq_completion)bond3#3 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex nl_table_wait.lock irq_context: 0 (wq_completion)bond3#3 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex net_rwsem irq_context: 0 (wq_completion)bond3#3 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex net_rwsem &rq->__lock irq_context: 0 (wq_completion)bond3#3 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex net_rwsem &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)bond3#3 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex net_rwsem &list->lock#2 irq_context: 0 (wq_completion)bond3#3 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &tn->lock irq_context: 0 (wq_completion)bond3#3 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &idev->mc_lock irq_context: 0 (wq_completion)bond3#3 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &idev->mc_lock fs_reclaim irq_context: 0 (wq_completion)bond3#3 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &idev->mc_lock fs_reclaim mmu_notifier_invalidate_range_start irq_context: 0 (wq_completion)bond3#3 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &idev->mc_lock rcu_read_lock &ndev->lock irq_context: 0 (wq_completion)bond3#3 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &idev->mc_lock rcu_read_lock &dir->lock#2 irq_context: 0 (wq_completion)bond3#3 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &idev->mc_lock rcu_read_lock &ul->lock#2 irq_context: 0 (wq_completion)bond3#3 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &idev->mc_lock rcu_read_lock rcu_read_lock &____s->seqcount#10 irq_context: 0 (wq_completion)bond3#3 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &idev->mc_lock rcu_read_lock rcu_read_lock rcu_read_lock_bh rcu_read_lock rcu_read_lock_bh rcu_read_lock &br->multicast_lock irq_context: 0 (wq_completion)bond3#3 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &idev->mc_lock rcu_read_lock rcu_read_lock rcu_read_lock_bh rcu_read_lock rcu_read_lock_bh rcu_read_lock &br->multicast_lock &obj_hash[i].lock irq_context: 0 (wq_completion)bond3#3 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &idev->mc_lock rcu_read_lock rcu_read_lock rcu_read_lock_bh rcu_read_lock rcu_read_lock_bh rcu_read_lock &br->multicast_lock &base->lock irq_context: 0 (wq_completion)bond3#3 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &idev->mc_lock rcu_read_lock rcu_read_lock rcu_read_lock_bh rcu_read_lock rcu_read_lock_bh rcu_read_lock &br->multicast_lock &base->lock &obj_hash[i].lock irq_context: 0 (wq_completion)bond3#3 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &idev->mc_lock rcu_read_lock rcu_read_lock rcu_read_lock_bh rcu_read_lock rcu_read_lock_bh rcu_read_lock &data->lock irq_context: 0 (wq_completion)bond3#3 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &idev->mc_lock rcu_read_lock rcu_read_lock rcu_read_lock_bh rcu_read_lock rcu_read_lock_bh rcu_read_lock &obj_hash[i].lock irq_context: 0 (wq_completion)bond3#3 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &idev->mc_lock rcu_read_lock rcu_read_lock rcu_node_0 irq_context: 0 (wq_completion)bond3#3 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &idev->mc_lock rcu_read_lock rcu_read_lock &rq->__lock irq_context: 0 (wq_completion)bond3#3 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &idev->mc_lock rcu_read_lock rcu_node_0 irq_context: 0 (wq_completion)bond3#3 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex.wait_lock irq_context: 0 (wq_completion)bond3#3 (work_completion)(&(&bond->mcast_work)->work) &p->pi_lock irq_context: 0 (wq_completion)bond3#3 (work_completion)(&(&bond->mcast_work)->work) &p->pi_lock &rq->__lock irq_context: 0 (wq_completion)bond3#3 (work_completion)(&(&bond->mcast_work)->work) &p->pi_lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)bond3#3 (work_completion)(&(&slave->notify_work)->work) rtnl_mutex net_rwsem irq_context: 0 (wq_completion)bond3#3 (work_completion)(&(&slave->notify_work)->work) rtnl_mutex net_rwsem &list->lock#2 irq_context: 0 (wq_completion)bond3#3 (work_completion)(&(&slave->notify_work)->work) rtnl_mutex &tn->lock irq_context: 0 (wq_completion)bond3#3 (work_completion)(&(&slave->notify_work)->work) rtnl_mutex.wait_lock irq_context: 0 (wq_completion)bond3#3 (work_completion)(&(&slave->notify_work)->work) &p->pi_lock irq_context: 0 (wq_completion)bond4#2 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex fs_reclaim irq_context: 0 (wq_completion)bond4#2 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex fs_reclaim mmu_notifier_invalidate_range_start irq_context: 0 (wq_completion)bond4#2 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex rcu_read_lock &bond->stats_lock/1 irq_context: 0 (wq_completion)bond4#2 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex rcu_read_lock &ndev->lock irq_context: 0 (wq_completion)bond4#2 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex nl_table_lock irq_context: 0 (wq_completion)bond4#2 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &obj_hash[i].lock irq_context: 0 (wq_completion)bond4#2 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex nl_table_wait.lock irq_context: 0 (wq_completion)bond4#2 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex net_rwsem irq_context: 0 (wq_completion)bond4#2 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex net_rwsem &list->lock#2 irq_context: 0 (wq_completion)bond4#2 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &tn->lock irq_context: 0 (wq_completion)bond4#2 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &idev->mc_lock irq_context: 0 (wq_completion)bond4#2 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &idev->mc_lock fs_reclaim irq_context: 0 (wq_completion)bond4#2 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &idev->mc_lock fs_reclaim mmu_notifier_invalidate_range_start irq_context: 0 (wq_completion)bond4#2 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &idev->mc_lock rcu_read_lock &ndev->lock irq_context: 0 (wq_completion)bond4#2 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &idev->mc_lock rcu_read_lock &dir->lock#2 irq_context: 0 (wq_completion)bond4#2 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &idev->mc_lock rcu_read_lock &c->lock irq_context: 0 (wq_completion)bond4#2 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &idev->mc_lock rcu_read_lock &ul->lock#2 irq_context: 0 (wq_completion)bond4#2 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &idev->mc_lock rcu_read_lock rcu_read_lock &____s->seqcount#10 irq_context: 0 (wq_completion)bond4#2 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &idev->mc_lock rcu_read_lock rcu_read_lock rcu_read_lock_bh rcu_read_lock rcu_read_lock_bh rcu_read_lock &br->multicast_lock irq_context: 0 (wq_completion)bond4#2 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &idev->mc_lock rcu_read_lock rcu_read_lock rcu_read_lock_bh rcu_read_lock rcu_read_lock_bh rcu_read_lock &br->multicast_lock &obj_hash[i].lock irq_context: 0 (wq_completion)bond4#2 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &idev->mc_lock rcu_read_lock rcu_read_lock rcu_read_lock_bh rcu_read_lock rcu_read_lock_bh rcu_read_lock &br->multicast_lock &base->lock irq_context: 0 (wq_completion)bond4#2 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &idev->mc_lock rcu_read_lock rcu_read_lock rcu_read_lock_bh rcu_read_lock rcu_read_lock_bh rcu_read_lock &data->lock irq_context: 0 (wq_completion)bond4#2 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &idev->mc_lock rcu_read_lock rcu_read_lock rcu_read_lock_bh rcu_read_lock rcu_read_lock_bh rcu_read_lock &obj_hash[i].lock irq_context: 0 rtnl_mutex &br->hash_lock &obj_hash[i].lock pool_lock irq_context: 0 sb_writers#4 &type->i_mutex_dir_key#3/1 jbd2_handle &ei->i_data_sem &ei->i_es_lock &pcp->lock &zone->lock irq_context: 0 (wq_completion)bond28#2 irq_context: 0 (wq_completion)bond28#2 (work_completion)(&(&slave->notify_work)->work) irq_context: 0 (wq_completion)bond28#2 (work_completion)(&(&slave->notify_work)->work) &obj_hash[i].lock irq_context: 0 (wq_completion)bond28#2 (work_completion)(&(&slave->notify_work)->work) &base->lock irq_context: 0 (wq_completion)bond28#2 (work_completion)(&(&slave->notify_work)->work) &base->lock &obj_hash[i].lock irq_context: 0 (wq_completion)bond28#2 (work_completion)(&(&slave->notify_work)->work) &rq->__lock irq_context: 0 (wq_completion)bond28#2 (work_completion)(&(&slave->notify_work)->work) &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)bond28#2 (work_completion)(&(&bond->mcast_work)->work) irq_context: 0 (wq_completion)bond28#2 (work_completion)(&(&bond->mcast_work)->work) &obj_hash[i].lock irq_context: 0 (wq_completion)bond28#2 (work_completion)(&(&bond->mcast_work)->work) &base->lock irq_context: 0 (wq_completion)bond28#2 (work_completion)(&(&bond->mcast_work)->work) &base->lock &obj_hash[i].lock irq_context: 0 (wq_completion)ipv6_addrconf (work_completion)(&(&ifa->dad_work)->work) rtnl_mutex &idev->mc_lock &dev_addr_list_lock_key/1 rcu_read_lock &bridge_netdev_addr_lock_key &____s->seqcount#2 irq_context: 0 (wq_completion)bond28#2 (work_completion)(&(&slave->notify_work)->work) rtnl_mutex net_rwsem irq_context: 0 (wq_completion)bond28#2 (work_completion)(&(&slave->notify_work)->work) rtnl_mutex net_rwsem &list->lock#2 irq_context: 0 (wq_completion)bond28#2 (work_completion)(&(&slave->notify_work)->work) rtnl_mutex &tn->lock irq_context: 0 (wq_completion)bond28#2 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex fs_reclaim irq_context: 0 (wq_completion)bond28#2 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex fs_reclaim mmu_notifier_invalidate_range_start irq_context: 0 (wq_completion)bond28#2 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex rcu_read_lock &bond->stats_lock/1 irq_context: 0 (wq_completion)bond28#2 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex rcu_read_lock &ndev->lock irq_context: 0 (wq_completion)bond28#2 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex nl_table_lock irq_context: 0 (wq_completion)bond28#2 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &obj_hash[i].lock irq_context: 0 (wq_completion)bond28#2 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex nl_table_wait.lock irq_context: 0 (wq_completion)bond28#2 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex net_rwsem irq_context: 0 (wq_completion)bond28#2 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex net_rwsem &list->lock#2 irq_context: 0 (wq_completion)bond28#2 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &tn->lock irq_context: 0 (wq_completion)bond28#2 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &idev->mc_lock irq_context: 0 (wq_completion)bond28#2 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &idev->mc_lock fs_reclaim irq_context: 0 (wq_completion)bond28#2 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &idev->mc_lock fs_reclaim mmu_notifier_invalidate_range_start irq_context: 0 (wq_completion)bond28#2 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &idev->mc_lock rcu_read_lock &ndev->lock irq_context: 0 (wq_completion)bond28#2 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &idev->mc_lock rcu_read_lock &dir->lock#2 irq_context: 0 (wq_completion)bond28#2 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &idev->mc_lock rcu_read_lock &ul->lock#2 irq_context: 0 (wq_completion)bond28#2 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &idev->mc_lock rcu_read_lock rcu_read_lock &obj_hash[i].lock irq_context: 0 (wq_completion)bond28#2 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &idev->mc_lock rcu_read_lock rcu_read_lock &c->lock irq_context: 0 (wq_completion)bond28#2 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &idev->mc_lock rcu_read_lock rcu_read_lock &dir->lock#2 irq_context: 0 (wq_completion)bond28#2 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &idev->mc_lock rcu_read_lock rcu_read_lock &tbl->lock irq_context: 0 (wq_completion)bond28#2 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &idev->mc_lock rcu_read_lock rcu_read_lock &n->lock irq_context: 0 (wq_completion)bond28#2 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &idev->mc_lock rcu_read_lock rcu_read_lock &____s->seqcount#9 irq_context: 0 (wq_completion)bond28#2 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &idev->mc_lock rcu_read_lock rcu_read_lock rcu_read_lock_bh rcu_read_lock rcu_read_lock_bh rcu_read_lock &br->multicast_lock irq_context: 0 (wq_completion)bond28#2 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &idev->mc_lock rcu_read_lock rcu_read_lock rcu_read_lock_bh rcu_read_lock rcu_read_lock_bh rcu_read_lock &br->multicast_lock &obj_hash[i].lock irq_context: 0 (wq_completion)bond28#2 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &idev->mc_lock rcu_read_lock rcu_read_lock rcu_read_lock_bh rcu_read_lock rcu_read_lock_bh rcu_read_lock &br->multicast_lock rcu_read_lock rhashtable_bucket irq_context: 0 (wq_completion)bond28#2 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &idev->mc_lock rcu_read_lock rcu_read_lock rcu_read_lock_bh rcu_read_lock rcu_read_lock_bh rcu_read_lock &br->multicast_lock nl_table_lock irq_context: 0 (wq_completion)bond28#2 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &idev->mc_lock rcu_read_lock rcu_read_lock rcu_read_lock_bh rcu_read_lock rcu_read_lock_bh rcu_read_lock &br->multicast_lock nl_table_wait.lock irq_context: 0 (wq_completion)bond28#2 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &idev->mc_lock rcu_read_lock rcu_read_lock rcu_read_lock_bh rcu_read_lock rcu_read_lock_bh rcu_read_lock &br->multicast_lock &base->lock irq_context: 0 (wq_completion)bond28#2 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &idev->mc_lock rcu_read_lock rcu_read_lock rcu_read_lock_bh rcu_read_lock rcu_read_lock_bh rcu_read_lock &br->multicast_lock &base->lock &obj_hash[i].lock irq_context: 0 (wq_completion)bond28#2 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &idev->mc_lock rcu_read_lock rcu_read_lock rcu_read_lock_bh rcu_read_lock rcu_read_lock_bh rcu_read_lock &data->lock irq_context: 0 (wq_completion)bond28#2 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &idev->mc_lock rcu_read_lock rcu_read_lock rcu_read_lock_bh rcu_read_lock rcu_read_lock_bh rcu_read_lock &obj_hash[i].lock irq_context: 0 (wq_completion)bond29#2 irq_context: 0 (wq_completion)bond29#2 (work_completion)(&(&slave->notify_work)->work) irq_context: 0 (wq_completion)bond29#2 (work_completion)(&(&slave->notify_work)->work) &obj_hash[i].lock irq_context: 0 (wq_completion)bond29#2 (work_completion)(&(&slave->notify_work)->work) &base->lock irq_context: 0 (wq_completion)bond29#2 (work_completion)(&(&slave->notify_work)->work) &base->lock &obj_hash[i].lock irq_context: 0 (wq_completion)bond29#2 (work_completion)(&(&slave->notify_work)->work) &rq->__lock irq_context: 0 (wq_completion)bond29#2 (work_completion)(&(&slave->notify_work)->work) &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)bond29#2 (work_completion)(&(&bond->mcast_work)->work) irq_context: 0 (wq_completion)bond29#2 (work_completion)(&(&bond->mcast_work)->work) &obj_hash[i].lock irq_context: 0 (wq_completion)bond29#2 (work_completion)(&(&bond->mcast_work)->work) &base->lock irq_context: 0 (wq_completion)bond29#2 (work_completion)(&(&bond->mcast_work)->work) &base->lock &obj_hash[i].lock irq_context: 0 (wq_completion)bond29#2 (work_completion)(&(&bond->mcast_work)->work) &rq->__lock irq_context: 0 (wq_completion)bond29#2 (work_completion)(&(&bond->mcast_work)->work) &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)bond29#2 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex fs_reclaim irq_context: 0 (wq_completion)bond29#2 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex fs_reclaim mmu_notifier_invalidate_range_start irq_context: 0 (wq_completion)bond29#2 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &c->lock irq_context: 0 (wq_completion)bond29#2 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &n->list_lock irq_context: 0 (wq_completion)bond29#2 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex rcu_read_lock &bond->stats_lock/1 irq_context: 0 (wq_completion)bond29#2 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex rcu_read_lock &ndev->lock irq_context: 0 (wq_completion)bond29#2 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex nl_table_lock irq_context: 0 (wq_completion)bond29#2 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &obj_hash[i].lock irq_context: 0 (wq_completion)bond29#2 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex nl_table_wait.lock irq_context: 0 (wq_completion)bond29#2 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex net_rwsem irq_context: 0 (wq_completion)bond29#2 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex net_rwsem &list->lock#2 irq_context: 0 (wq_completion)bond29#2 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &tn->lock irq_context: 0 (wq_completion)bond29#2 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &idev->mc_lock irq_context: 0 (wq_completion)bond29#2 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex.wait_lock irq_context: 0 (wq_completion)bond29#2 (work_completion)(&(&bond->mcast_work)->work) &p->pi_lock irq_context: 0 (wq_completion)bond29#2 (work_completion)(&(&slave->notify_work)->work) rtnl_mutex net_rwsem irq_context: 0 (wq_completion)bond29#2 (work_completion)(&(&slave->notify_work)->work) rtnl_mutex net_rwsem &list->lock#2 irq_context: 0 (wq_completion)bond29#2 (work_completion)(&(&slave->notify_work)->work) rtnl_mutex &tn->lock irq_context: 0 (wq_completion)bond29#2 (work_completion)(&(&slave->notify_work)->work) rtnl_mutex.wait_lock irq_context: 0 (wq_completion)bond29#2 (work_completion)(&(&slave->notify_work)->work) &p->pi_lock irq_context: softirq (&ndev->rs_timer) rcu_read_lock rcu_read_lock rcu_read_lock_bh rcu_read_lock rcu_read_lock_bh rcu_read_lock quarantine_lock irq_context: 0 (wq_completion)bond30#2 irq_context: 0 (wq_completion)bond30#2 &rq->__lock irq_context: 0 (wq_completion)bond30#2 &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)bond30#2 (work_completion)(&(&slave->notify_work)->work) irq_context: 0 (wq_completion)bond30#2 (work_completion)(&(&slave->notify_work)->work) &obj_hash[i].lock irq_context: 0 (wq_completion)bond30#2 (work_completion)(&(&slave->notify_work)->work) &base->lock irq_context: 0 (wq_completion)bond30#2 (work_completion)(&(&slave->notify_work)->work) &base->lock &obj_hash[i].lock irq_context: 0 (wq_completion)gid-cache-wq (work_completion)(&ndev_work->work) devices_rwsem &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)bond30#2 (work_completion)(&(&bond->mcast_work)->work) irq_context: 0 (wq_completion)bond30#2 (work_completion)(&(&bond->mcast_work)->work) &obj_hash[i].lock irq_context: 0 (wq_completion)bond30#2 (work_completion)(&(&bond->mcast_work)->work) &base->lock irq_context: 0 (wq_completion)bond30#2 (work_completion)(&(&bond->mcast_work)->work) &base->lock &obj_hash[i].lock irq_context: 0 (wq_completion)bond30#2 (work_completion)(&(&bond->mcast_work)->work) &rq->__lock irq_context: 0 (wq_completion)bond30#2 (work_completion)(&(&bond->mcast_work)->work) &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)bond30#2 (work_completion)(&(&slave->notify_work)->work) rtnl_mutex net_rwsem irq_context: 0 (wq_completion)bond30#2 (work_completion)(&(&slave->notify_work)->work) rtnl_mutex net_rwsem &list->lock#2 irq_context: 0 (wq_completion)bond30#2 (work_completion)(&(&slave->notify_work)->work) rtnl_mutex &tn->lock irq_context: 0 (wq_completion)bond30#2 (work_completion)(&(&slave->notify_work)->work) rtnl_mutex &tn->lock &rq->__lock irq_context: 0 (wq_completion)bond30#2 (work_completion)(&(&slave->notify_work)->work) rtnl_mutex &tn->lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)bond30#2 (work_completion)(&(&slave->notify_work)->work) rtnl_mutex.wait_lock irq_context: 0 (wq_completion)bond30#2 (work_completion)(&(&slave->notify_work)->work) &p->pi_lock irq_context: 0 (wq_completion)bond30#2 (work_completion)(&(&slave->notify_work)->work) &p->pi_lock &rq->__lock irq_context: 0 (wq_completion)bond30#2 (work_completion)(&(&slave->notify_work)->work) &p->pi_lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)bond30#2 (work_completion)(&(&slave->notify_work)->work) &rq->__lock irq_context: 0 (wq_completion)bond30#2 (work_completion)(&(&slave->notify_work)->work) &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)events (linkwatch_work).work rtnl_mutex rcu_state.exp_mutex fill_pool_map-wait-type-override pool_lock#2 irq_context: 0 (wq_completion)events (linkwatch_work).work rtnl_mutex rcu_state.exp_mutex fill_pool_map-wait-type-override pool_lock irq_context: 0 (wq_completion)events (linkwatch_work).work rtnl_mutex rcu_state.exp_mutex &obj_hash[i].lock pool_lock irq_context: 0 (wq_completion)events (linkwatch_work).work rtnl_mutex rcu_state.exp_mutex rcu_read_lock &pool->lock fill_pool_map-wait-type-override pool_lock#2 irq_context: 0 (wq_completion)events (linkwatch_work).work rtnl_mutex rcu_state.exp_mutex rcu_read_lock &pool->lock fill_pool_map-wait-type-override pool_lock irq_context: 0 (wq_completion)bond30#2 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex fs_reclaim irq_context: 0 (wq_completion)bond30#2 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex fs_reclaim mmu_notifier_invalidate_range_start irq_context: 0 (wq_completion)bond30#2 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &c->lock irq_context: 0 (wq_completion)bond30#2 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex rcu_read_lock &bond->stats_lock/1 irq_context: 0 (wq_completion)bond30#2 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex rcu_read_lock &ndev->lock irq_context: 0 (wq_completion)bond30#2 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex nl_table_lock irq_context: 0 (wq_completion)bond30#2 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &obj_hash[i].lock irq_context: 0 (wq_completion)bond30#2 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex nl_table_wait.lock irq_context: 0 (wq_completion)bond30#2 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex net_rwsem irq_context: 0 (wq_completion)bond30#2 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex net_rwsem &list->lock#2 irq_context: 0 (wq_completion)bond30#2 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &tn->lock irq_context: 0 (wq_completion)bond30#2 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &idev->mc_lock irq_context: 0 (wq_completion)bond30#2 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex.wait_lock irq_context: 0 (wq_completion)bond30#2 (work_completion)(&(&bond->mcast_work)->work) &p->pi_lock irq_context: 0 (wq_completion)ipv6_addrconf (work_completion)(&(&ifa->dad_work)->work) rtnl_mutex &idev->mc_lock &dev_addr_list_lock_key/1 &n->list_lock irq_context: 0 (wq_completion)ipv6_addrconf (work_completion)(&(&ifa->dad_work)->work) rtnl_mutex &idev->mc_lock &dev_addr_list_lock_key/1 &n->list_lock &c->lock irq_context: 0 (wq_completion)wg-kex-wg1#4 (work_completion)(&({ do { const void *__vpp_verify = (typeof((worker) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __asm__ ("" : "=r"(__ptr) : "0"((typeof(*((worker))) *)((worker)))); (typeof((typeof(*((worker))) *)((worker)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); })->work) &peer->endpoint_lock rcu_read_lock_bh rcu_read_lock &dir->lock#2 irq_context: 0 (wq_completion)wg-kex-wg1#4 (work_completion)(&({ do { const void *__vpp_verify = (typeof((worker) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __asm__ ("" : "=r"(__ptr) : "0"((typeof(*((worker))) *)((worker)))); (typeof((typeof(*((worker))) *)((worker)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); })->work) &peer->endpoint_lock rcu_read_lock_bh rcu_read_lock &ul->lock irq_context: 0 (wq_completion)wg-kex-wg1#4 (work_completion)(&({ do { const void *__vpp_verify = (typeof((worker) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __asm__ ("" : "=r"(__ptr) : "0"((typeof(*((worker))) *)((worker)))); (typeof((typeof(*((worker))) *)((worker)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); })->work) &peer->endpoint_lock rcu_read_lock_bh rcu_read_lock &obj_hash[i].lock irq_context: 0 (wq_completion)mld (work_completion)(&(&idev->mc_dad_work)->work) &idev->mc_lock rcu_read_lock rcu_read_lock rcu_read_lock_bh rcu_read_lock rcu_read_lock_bh rcu_read_lock quarantine_lock irq_context: 0 sk_lock-AF_NFC irq_context: 0 sk_lock-AF_NFC slock-AF_NFC irq_context: 0 slock-AF_NFC irq_context: 0 (wq_completion)bond8#4 irq_context: 0 (wq_completion)bond8#4 (work_completion)(&(&slave->notify_work)->work) irq_context: 0 (wq_completion)bond8#4 (work_completion)(&(&slave->notify_work)->work) &obj_hash[i].lock irq_context: 0 (wq_completion)bond8#4 (work_completion)(&(&slave->notify_work)->work) &base->lock irq_context: 0 (wq_completion)bond8#4 (work_completion)(&(&slave->notify_work)->work) &base->lock &obj_hash[i].lock irq_context: 0 (wq_completion)bond8#4 (work_completion)(&(&slave->notify_work)->work) &rq->__lock irq_context: 0 (wq_completion)bond8#4 (work_completion)(&(&slave->notify_work)->work) &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)bond8#4 (work_completion)(&(&bond->mcast_work)->work) irq_context: 0 (wq_completion)bond8#4 (work_completion)(&(&bond->mcast_work)->work) &obj_hash[i].lock irq_context: 0 (wq_completion)bond8#4 (work_completion)(&(&bond->mcast_work)->work) &base->lock irq_context: 0 (wq_completion)bond8#4 (work_completion)(&(&bond->mcast_work)->work) &base->lock &obj_hash[i].lock irq_context: 0 (wq_completion)bond8#4 (work_completion)(&(&slave->notify_work)->work) rtnl_mutex net_rwsem irq_context: 0 (wq_completion)bond8#4 (work_completion)(&(&slave->notify_work)->work) rtnl_mutex net_rwsem &list->lock#2 irq_context: 0 (wq_completion)bond8#4 (work_completion)(&(&slave->notify_work)->work) rtnl_mutex &tn->lock irq_context: 0 (wq_completion)bond8#4 (work_completion)(&(&slave->notify_work)->work) rtnl_mutex.wait_lock irq_context: 0 (wq_completion)bond8#4 (work_completion)(&(&slave->notify_work)->work) &p->pi_lock irq_context: 0 (wq_completion)bond8#4 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex fs_reclaim irq_context: 0 (wq_completion)bond8#4 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex fs_reclaim mmu_notifier_invalidate_range_start irq_context: 0 (wq_completion)bond8#4 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &c->lock irq_context: 0 (wq_completion)bond8#4 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &____s->seqcount#2 irq_context: 0 (wq_completion)bond8#4 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &____s->seqcount irq_context: 0 (wq_completion)bond8#4 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex rcu_read_lock &bond->stats_lock/1 irq_context: 0 (wq_completion)bond8#4 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex rcu_read_lock &ndev->lock irq_context: 0 (wq_completion)bond8#4 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex nl_table_lock irq_context: 0 (wq_completion)bond8#4 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &obj_hash[i].lock irq_context: 0 (wq_completion)bond8#4 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex nl_table_wait.lock irq_context: 0 (wq_completion)bond8#4 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex net_rwsem irq_context: 0 (wq_completion)bond8#4 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex net_rwsem &list->lock#2 irq_context: 0 (wq_completion)bond8#4 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &tn->lock irq_context: 0 (wq_completion)bond8#4 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &idev->mc_lock irq_context: 0 (wq_completion)bond8#4 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &idev->mc_lock fs_reclaim irq_context: 0 (wq_completion)bond8#4 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &idev->mc_lock fs_reclaim mmu_notifier_invalidate_range_start irq_context: 0 (wq_completion)bond8#4 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &idev->mc_lock &c->lock irq_context: 0 (wq_completion)bond8#4 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &idev->mc_lock rcu_read_lock &ndev->lock irq_context: 0 (wq_completion)bond8#4 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &idev->mc_lock rcu_read_lock &c->lock irq_context: 0 (wq_completion)bond8#4 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &idev->mc_lock rcu_read_lock &dir->lock#2 irq_context: 0 (wq_completion)bond8#4 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &idev->mc_lock rcu_read_lock &ul->lock#2 irq_context: 0 (wq_completion)bond8#4 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &idev->mc_lock rcu_read_lock rcu_read_lock &c->lock irq_context: 0 (wq_completion)bond8#4 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &idev->mc_lock rcu_read_lock rcu_read_lock &obj_hash[i].lock irq_context: 0 (wq_completion)bond8#4 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &idev->mc_lock rcu_read_lock rcu_read_lock &dir->lock#2 irq_context: 0 (wq_completion)bond8#4 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &idev->mc_lock rcu_read_lock rcu_read_lock &tbl->lock irq_context: 0 (wq_completion)bond8#4 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &idev->mc_lock rcu_read_lock rcu_read_lock &n->lock irq_context: 0 (wq_completion)bond8#4 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &idev->mc_lock rcu_read_lock rcu_read_lock &____s->seqcount#9 irq_context: 0 (wq_completion)bond8#4 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &idev->mc_lock rcu_read_lock rcu_read_lock rcu_read_lock_bh rcu_read_lock rcu_read_lock_bh rcu_read_lock &br->multicast_lock irq_context: 0 (wq_completion)bond8#4 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &idev->mc_lock rcu_read_lock rcu_read_lock rcu_read_lock_bh rcu_read_lock rcu_read_lock_bh rcu_read_lock &br->multicast_lock &obj_hash[i].lock irq_context: 0 (wq_completion)bond8#4 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &idev->mc_lock rcu_read_lock rcu_read_lock rcu_read_lock_bh rcu_read_lock rcu_read_lock_bh rcu_read_lock &br->multicast_lock rcu_read_lock rhashtable_bucket irq_context: 0 (wq_completion)bond8#4 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &idev->mc_lock rcu_read_lock rcu_read_lock rcu_read_lock_bh rcu_read_lock rcu_read_lock_bh rcu_read_lock &br->multicast_lock nl_table_lock irq_context: 0 (wq_completion)bond8#4 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &idev->mc_lock rcu_read_lock rcu_read_lock rcu_read_lock_bh rcu_read_lock rcu_read_lock_bh rcu_read_lock &br->multicast_lock nl_table_wait.lock irq_context: 0 (wq_completion)bond8#4 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &idev->mc_lock rcu_read_lock rcu_read_lock rcu_read_lock_bh rcu_read_lock rcu_read_lock_bh rcu_read_lock &br->multicast_lock &base->lock irq_context: 0 (wq_completion)bond8#4 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &idev->mc_lock rcu_read_lock rcu_read_lock rcu_read_lock_bh rcu_read_lock rcu_read_lock_bh rcu_read_lock &br->multicast_lock &base->lock &obj_hash[i].lock irq_context: 0 (wq_completion)bond8#4 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &idev->mc_lock rcu_read_lock rcu_read_lock rcu_read_lock_bh rcu_read_lock rcu_read_lock_bh rcu_read_lock &data->lock irq_context: 0 (wq_completion)bond8#4 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &idev->mc_lock rcu_read_lock rcu_read_lock rcu_read_lock_bh rcu_read_lock rcu_read_lock_bh rcu_read_lock &obj_hash[i].lock irq_context: 0 sk_lock-AF_INET rcu_read_lock rcu_read_lock_bh &sch->q.lock irq_context: 0 sk_lock-AF_INET rcu_read_lock rcu_read_lock_bh &sch->q.lock pool_lock#2 irq_context: 0 sk_lock-AF_INET rcu_read_lock rcu_read_lock_bh &sch->q.lock tk_core.seq.seqcount irq_context: 0 (wq_completion)bond9#4 irq_context: 0 (wq_completion)bond9#4 (work_completion)(&(&slave->notify_work)->work) irq_context: 0 (wq_completion)bond9#4 (work_completion)(&(&slave->notify_work)->work) &obj_hash[i].lock irq_context: 0 (wq_completion)bond9#4 (work_completion)(&(&slave->notify_work)->work) &base->lock irq_context: 0 (wq_completion)bond9#4 (work_completion)(&(&slave->notify_work)->work) &base->lock &obj_hash[i].lock irq_context: 0 (wq_completion)bond9#4 (work_completion)(&(&slave->notify_work)->work) &rq->__lock irq_context: 0 (wq_completion)bond9#4 (work_completion)(&(&slave->notify_work)->work) &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)bond9#4 (work_completion)(&(&bond->mcast_work)->work) irq_context: 0 (wq_completion)bond9#4 (work_completion)(&(&bond->mcast_work)->work) &obj_hash[i].lock irq_context: 0 (wq_completion)bond9#4 (work_completion)(&(&bond->mcast_work)->work) &base->lock irq_context: 0 (wq_completion)bond9#4 (work_completion)(&(&bond->mcast_work)->work) &base->lock &obj_hash[i].lock irq_context: 0 &tsk->futex_exit_mutex &obj_hash[i].lock pool_lock irq_context: 0 (wq_completion)bond9#4 (work_completion)(&(&slave->notify_work)->work) rtnl_mutex net_rwsem irq_context: 0 (wq_completion)bond9#4 (work_completion)(&(&slave->notify_work)->work) rtnl_mutex net_rwsem &list->lock#2 irq_context: 0 (wq_completion)bond9#4 (work_completion)(&(&slave->notify_work)->work) rtnl_mutex &tn->lock irq_context: 0 (wq_completion)bond9#4 (work_completion)(&(&slave->notify_work)->work) rtnl_mutex.wait_lock irq_context: 0 (wq_completion)bond9#4 (work_completion)(&(&slave->notify_work)->work) &p->pi_lock irq_context: 0 (wq_completion)bond9#4 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex fs_reclaim irq_context: 0 (wq_completion)bond9#4 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex fs_reclaim mmu_notifier_invalidate_range_start irq_context: 0 (wq_completion)bond9#4 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &c->lock irq_context: 0 (wq_completion)bond9#4 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &n->list_lock irq_context: 0 (wq_completion)bond9#4 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &n->list_lock &c->lock irq_context: 0 (wq_completion)bond9#4 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &rq->__lock irq_context: 0 (wq_completion)bond9#4 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)bond9#4 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex rcu_read_lock &bond->stats_lock/1 irq_context: 0 (wq_completion)bond9#4 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex rcu_read_lock &ndev->lock irq_context: 0 (wq_completion)bond9#4 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex nl_table_lock irq_context: 0 (wq_completion)bond9#4 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &obj_hash[i].lock irq_context: 0 (wq_completion)bond9#4 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex nl_table_wait.lock irq_context: 0 (wq_completion)bond9#4 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex net_rwsem irq_context: 0 (wq_completion)bond9#4 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex net_rwsem &list->lock#2 irq_context: 0 (wq_completion)bond9#4 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &tn->lock irq_context: 0 (wq_completion)bond9#4 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &idev->mc_lock irq_context: 0 (wq_completion)bond9#4 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex.wait_lock irq_context: 0 (wq_completion)bond9#4 (work_completion)(&(&bond->mcast_work)->work) &p->pi_lock irq_context: 0 nf_sockopt_mutex rcu_read_lock &rq->__lock irq_context: 0 nf_sockopt_mutex rcu_read_lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 sk_lock-AF_INET rcu_read_lock rcu_read_lock_bh &sch->q.lock hrtimer_bases.lock irq_context: 0 sk_lock-AF_INET rcu_read_lock rcu_read_lock_bh &sch->q.lock hrtimer_bases.lock &obj_hash[i].lock irq_context: 0 sk_lock-AF_INET rcu_read_lock rcu_read_lock_bh &sch->q.lock hrtimer_bases.lock tk_core.seq.seqcount irq_context: softirq rcu_read_lock rcu_read_lock rcu_read_lock batched_entropy_u8.lock irq_context: 0 sk_lock-AF_INET rcu_read_lock rcu_read_lock_bh &sch->q.lock &c->lock irq_context: 0 sk_lock-AF_INET rcu_read_lock rcu_read_lock_bh &sch->q.lock &____s->seqcount#2 irq_context: 0 sk_lock-AF_INET rcu_read_lock rcu_read_lock_bh &sch->q.lock &____s->seqcount irq_context: 0 &xt[i].mutex &mm->mmap_lock &rq->__lock &obj_hash[i].lock irq_context: 0 &xt[i].mutex &mm->mmap_lock &rq->__lock &base->lock irq_context: 0 &xt[i].mutex &mm->mmap_lock &rq->__lock &base->lock &obj_hash[i].lock irq_context: 0 &mm->mmap_lock mmu_notifier_invalidate_range_start &cfs_rq->removed.lock irq_context: 0 rtnl_mutex &idev->mc_lock &dev_addr_list_lock_key &____s->seqcount#2 irq_context: 0 rtnl_mutex &idev->mc_lock &dev_addr_list_lock_key &____s->seqcount irq_context: 0 rtnl_mutex cpu_hotplug_lock wq_pool_mutex fs_reclaim mmu_notifier_invalidate_range_start &rq->__lock irq_context: 0 rtnl_mutex cpu_hotplug_lock wq_pool_mutex fs_reclaim mmu_notifier_invalidate_range_start &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 rtnl_mutex &br->hash_lock rcu_read_lock rcu_read_lock &pool->lock &p->pi_lock &rq->__lock &cfs_rq->removed.lock irq_context: 0 &mm->mmap_lock &mapping->i_mmap_rwsem &anon_vma->rwsem &rcu_state.expedited_wq &p->pi_lock &rq->__lock irq_context: 0 &mm->mmap_lock &mapping->i_mmap_rwsem &anon_vma->rwsem &rcu_state.expedited_wq &p->pi_lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 &mm->mmap_lock &mapping->i_mmap_rwsem &anon_vma->rwsem fill_pool_map-wait-type-override &____s->seqcount#2 irq_context: 0 sk_lock-AF_INET &dccp_hashinfo.bhash[i].lock &dccp_hashinfo.bhash2[i].lock &c->lock irq_context: 0 sk_lock-AF_INET &dccp_hashinfo.bhash[i].lock &dccp_hashinfo.bhash2[i].lock clock-AF_INET irq_context: 0 &sb->s_type->i_mutex_key#10 slock-AF_INET &dccp_hashinfo.bhash[i].lock irq_context: 0 &sb->s_type->i_mutex_key#10 slock-AF_INET &dccp_hashinfo.bhash[i].lock stock_lock irq_context: 0 &sb->s_type->i_mutex_key#10 slock-AF_INET &dccp_hashinfo.bhash[i].lock &obj_hash[i].lock irq_context: 0 &sb->s_type->i_mutex_key#10 slock-AF_INET &dccp_hashinfo.bhash[i].lock pool_lock#2 irq_context: 0 &sb->s_type->i_mutex_key#10 slock-AF_INET &dccp_hashinfo.bhash[i].lock &dccp_hashinfo.bhash2[i].lock irq_context: 0 &sb->s_type->i_mutex_key#10 slock-AF_INET &dccp_hashinfo.bhash[i].lock &dccp_hashinfo.bhash2[i].lock stock_lock irq_context: 0 &sb->s_type->i_mutex_key#10 slock-AF_INET &dccp_hashinfo.bhash[i].lock &dccp_hashinfo.bhash2[i].lock &obj_hash[i].lock irq_context: 0 &sb->s_type->i_mutex_key#10 slock-AF_INET &dccp_hashinfo.bhash[i].lock &dccp_hashinfo.bhash2[i].lock pool_lock#2 irq_context: 0 delayed_uprobe_lock rcu_read_lock &rq->__lock irq_context: 0 delayed_uprobe_lock rcu_read_lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)events_unbound (work_completion)(&(&kfence_timer)->work) fill_pool_map-wait-type-override &c->lock irq_context: 0 (wq_completion)events_unbound (work_completion)(&(&kfence_timer)->work) fill_pool_map-wait-type-override &____s->seqcount#2 irq_context: 0 (wq_completion)events_unbound (work_completion)(&(&kfence_timer)->work) fill_pool_map-wait-type-override &____s->seqcount irq_context: 0 rtnl_mutex fill_pool_map-wait-type-override batched_entropy_u8.lock crngs.lock irq_context: 0 &sb->s_type->i_mutex_key#10 sk_lock-AF_INET slock-AF_INET fill_pool_map-wait-type-override &____s->seqcount#2 irq_context: 0 &sb->s_type->i_mutex_key#10 sk_lock-AF_INET slock-AF_INET fill_pool_map-wait-type-override &____s->seqcount irq_context: 0 (wq_completion)events (linkwatch_work).work rtnl_mutex rcu_state.exp_mutex &cfs_rq->removed.lock irq_context: 0 (wq_completion)events (linkwatch_work).work rtnl_mutex rcu_state.exp_mutex &pool->lock &p->pi_lock &rq->__lock &cfs_rq->removed.lock irq_context: 0 hashlimit_mutex remove_cache_srcu irq_context: 0 hashlimit_mutex remove_cache_srcu quarantine_lock irq_context: 0 (wq_completion)bond13 (work_completion)(&(&bond->ad_work)->work) rcu_node_0 irq_context: 0 sb_writers#4 &type->i_mutex_dir_key#3/1 &type->i_mutex_dir_key#3 jbd2_handle &journal->j_state_lock &journal->j_wait_commit irq_context: 0 sb_writers#4 &type->i_mutex_dir_key#3/1 &type->i_mutex_dir_key#3 jbd2_handle &journal->j_state_lock &journal->j_wait_commit &p->pi_lock irq_context: 0 sb_writers#4 &type->i_mutex_dir_key#3/1 &type->i_mutex_dir_key#3 jbd2_handle &journal->j_state_lock &journal->j_wait_commit &p->pi_lock &rq->__lock irq_context: 0 sb_writers#4 &type->i_mutex_dir_key#3/1 &type->i_mutex_dir_key#3 jbd2_handle &journal->j_state_lock &journal->j_wait_commit &p->pi_lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 sb_writers#4 &type->i_mutex_dir_key#3/1 sb_internal jbd2_handle &sbi->s_orphan_lock mmu_notifier_invalidate_range_start irq_context: 0 sb_writers#4 &type->i_mutex_dir_key#3/1 sb_internal jbd2_handle &sbi->s_orphan_lock &____s->seqcount irq_context: 0 sb_writers#4 &type->i_mutex_dir_key#3/1 sb_internal jbd2_handle &sbi->s_orphan_lock pool_lock#2 irq_context: 0 sb_writers#4 &type->i_mutex_dir_key#3/1 sb_internal jbd2_handle &sbi->s_orphan_lock rcu_read_lock &____s->seqcount irq_context: 0 sb_writers#4 &type->i_mutex_dir_key#3/1 sb_internal jbd2_handle &sbi->s_orphan_lock rcu_read_lock pool_lock#2 irq_context: 0 sb_writers#4 &type->i_mutex_dir_key#3/1 sb_internal jbd2_handle &sbi->s_orphan_lock &obj_hash[i].lock irq_context: 0 sk_lock-AF_INET fill_pool_map-wait-type-override &____s->seqcount#2 irq_context: 0 sk_lock-AF_INET fill_pool_map-wait-type-override &____s->seqcount irq_context: 0 &vma->vm_lock->lock pcpu_lock stock_lock irq_context: 0 sk_lock-AF_INET rcu_read_lock rcu_read_lock_bh &data->lock &obj_hash[i].lock irq_context: 0 sk_lock-AF_INET rcu_read_lock rcu_read_lock_bh &data->lock &base->lock irq_context: 0 sk_lock-AF_INET rcu_read_lock rcu_read_lock_bh &data->lock &base->lock &obj_hash[i].lock irq_context: softirq rcu_read_lock &local->rx_path_lock rcu_read_lock &pool->lock fill_pool_map-wait-type-override &c->lock irq_context: softirq rcu_read_lock &local->rx_path_lock rcu_read_lock &pool->lock fill_pool_map-wait-type-override &____s->seqcount#2 irq_context: softirq rcu_read_lock &local->rx_path_lock rcu_read_lock &pool->lock fill_pool_map-wait-type-override &____s->seqcount irq_context: 0 (wq_completion)events_power_efficient (work_completion)(&(&gc_work->dwork)->work) rcu_read_lock &meta->lock irq_context: 0 (wq_completion)events_power_efficient (work_completion)(&(&gc_work->dwork)->work) rcu_read_lock kfence_freelist_lock irq_context: 0 rtnl_mutex &idev->mc_lock fill_pool_map-wait-type-override &____s->seqcount#2 irq_context: 0 dup_mmap_sem &mm->mmap_lock &mm->mmap_lock/1 fs_reclaim &rcu_state.expedited_wq irq_context: 0 dup_mmap_sem &mm->mmap_lock &mm->mmap_lock/1 fs_reclaim &rcu_state.expedited_wq &p->pi_lock irq_context: 0 dup_mmap_sem &mm->mmap_lock &mm->mmap_lock/1 fs_reclaim &rcu_state.expedited_wq &p->pi_lock &rq->__lock irq_context: 0 dup_mmap_sem &mm->mmap_lock &mm->mmap_lock/1 fs_reclaim &rcu_state.expedited_wq &p->pi_lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 sb_writers#4 sb_internal rcu_read_lock &rcu_state.expedited_wq &p->pi_lock irq_context: 0 sb_writers#4 sb_internal rcu_read_lock &rcu_state.expedited_wq &p->pi_lock &rq->__lock irq_context: 0 sb_writers#4 sb_internal rcu_read_lock &rcu_state.expedited_wq &p->pi_lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 rtnl_mutex bpf_devs_lock rcu_read_lock rcu_node_0 irq_context: 0 rtnl_mutex bpf_devs_lock rcu_read_lock &rq->__lock irq_context: 0 rtnl_mutex bpf_devs_lock rcu_read_lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)bond51 irq_context: 0 (wq_completion)bond51 (work_completion)(&(&slave->notify_work)->work) irq_context: 0 (wq_completion)bond51 (work_completion)(&(&slave->notify_work)->work) &obj_hash[i].lock irq_context: 0 (wq_completion)bond51 (work_completion)(&(&slave->notify_work)->work) &base->lock irq_context: 0 (wq_completion)bond51 (work_completion)(&(&slave->notify_work)->work) &base->lock &obj_hash[i].lock irq_context: 0 (wq_completion)bond51 (work_completion)(&(&slave->notify_work)->work) rtnl_mutex net_rwsem irq_context: 0 (wq_completion)bond51 (work_completion)(&(&slave->notify_work)->work) rtnl_mutex net_rwsem &list->lock#2 irq_context: 0 (wq_completion)bond51 (work_completion)(&(&slave->notify_work)->work) rtnl_mutex &tn->lock irq_context: softirq (&ndev->rs_timer) rcu_read_lock rcu_read_lock rcu_read_lock_bh rcu_read_lock rcu_read_lock_bh rcu_read_lock fill_pool_map-wait-type-override pool_lock#2 irq_context: softirq (&ndev->rs_timer) &ndev->lock fill_pool_map-wait-type-override &____s->seqcount#2 irq_context: softirq (&ndev->rs_timer) &ndev->lock fill_pool_map-wait-type-override &____s->seqcount irq_context: softirq (&ndev->rs_timer) rcu_read_lock rcu_read_lock rcu_read_lock_bh rcu_read_lock rcu_read_lock_bh rcu_read_lock fill_pool_map-wait-type-override pool_lock irq_context: 0 (wq_completion)bat_events (work_completion)(&(&bat_priv->bla.work)->work) rcu_read_lock &rcu_state.expedited_wq &p->pi_lock irq_context: 0 (wq_completion)bat_events (work_completion)(&(&bat_priv->bla.work)->work) rcu_read_lock &rcu_state.expedited_wq &p->pi_lock &rq->__lock irq_context: 0 (wq_completion)bat_events (work_completion)(&(&bat_priv->bla.work)->work) rcu_read_lock &rcu_state.expedited_wq &p->pi_lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)bond52 irq_context: 0 (wq_completion)bond52 (work_completion)(&(&slave->notify_work)->work) irq_context: 0 (wq_completion)bond52 (work_completion)(&(&slave->notify_work)->work) &obj_hash[i].lock irq_context: 0 (wq_completion)bond52 (work_completion)(&(&slave->notify_work)->work) &base->lock irq_context: 0 (wq_completion)bond52 (work_completion)(&(&slave->notify_work)->work) &base->lock &obj_hash[i].lock irq_context: 0 (wq_completion)bond52 (work_completion)(&(&slave->notify_work)->work) rtnl_mutex net_rwsem irq_context: 0 (wq_completion)bond52 (work_completion)(&(&slave->notify_work)->work) rtnl_mutex net_rwsem &list->lock#2 irq_context: 0 (wq_completion)bond52 (work_completion)(&(&slave->notify_work)->work) rtnl_mutex &tn->lock irq_context: softirq rcu_read_lock rcu_read_lock slock-AF_INET6/1 sk_lock-AF_INET6 fill_pool_map-wait-type-override pool_lock#2 irq_context: softirq rcu_read_lock rcu_read_lock slock-AF_INET6/1 sk_lock-AF_INET6 fill_pool_map-wait-type-override pool_lock irq_context: 0 (wq_completion)bond53 irq_context: 0 (wq_completion)bond53 (work_completion)(&(&slave->notify_work)->work) irq_context: 0 (wq_completion)bond53 (work_completion)(&(&slave->notify_work)->work) &obj_hash[i].lock irq_context: 0 (wq_completion)bond53 (work_completion)(&(&slave->notify_work)->work) &base->lock irq_context: 0 (wq_completion)bond53 (work_completion)(&(&slave->notify_work)->work) &base->lock &obj_hash[i].lock irq_context: 0 (wq_completion)bond53 (work_completion)(&(&slave->notify_work)->work) &rq->__lock irq_context: 0 (wq_completion)bond53 (work_completion)(&(&slave->notify_work)->work) &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)bond54 irq_context: 0 (wq_completion)bond54 (work_completion)(&(&slave->notify_work)->work) irq_context: 0 (wq_completion)bond54 (work_completion)(&(&slave->notify_work)->work) &obj_hash[i].lock irq_context: 0 (wq_completion)bond54 (work_completion)(&(&slave->notify_work)->work) &base->lock irq_context: 0 (wq_completion)bond54 (work_completion)(&(&slave->notify_work)->work) &base->lock &obj_hash[i].lock irq_context: 0 (wq_completion)bond54 (work_completion)(&(&slave->notify_work)->work) &rq->__lock irq_context: 0 (wq_completion)bond54 (work_completion)(&(&slave->notify_work)->work) &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)bond54 (work_completion)(&(&slave->notify_work)->work) rtnl_mutex net_rwsem irq_context: 0 (wq_completion)bond54 (work_completion)(&(&slave->notify_work)->work) rtnl_mutex net_rwsem &list->lock#2 irq_context: 0 (wq_completion)bond54 (work_completion)(&(&slave->notify_work)->work) rtnl_mutex &tn->lock irq_context: 0 (wq_completion)bond54 (work_completion)(&(&slave->notify_work)->work) rtnl_mutex.wait_lock irq_context: 0 (wq_completion)bond54 (work_completion)(&(&slave->notify_work)->work) &p->pi_lock irq_context: 0 (wq_completion)bond53 (work_completion)(&(&slave->notify_work)->work) rtnl_mutex net_rwsem irq_context: 0 (wq_completion)bond53 (work_completion)(&(&slave->notify_work)->work) rtnl_mutex net_rwsem &list->lock#2 irq_context: 0 (wq_completion)bond53 (work_completion)(&(&slave->notify_work)->work) rtnl_mutex &tn->lock irq_context: 0 (wq_completion)bond53 (work_completion)(&(&slave->notify_work)->work) rtnl_mutex.wait_lock irq_context: 0 (wq_completion)bond53 (work_completion)(&(&slave->notify_work)->work) &p->pi_lock irq_context: 0 sk_lock-AF_INET rcu_read_lock &ei->socket.wq.wait &p->pi_lock &cfs_rq->removed.lock irq_context: 0 (wq_completion)bond55 irq_context: 0 (wq_completion)bond55 (work_completion)(&(&slave->notify_work)->work) irq_context: 0 (wq_completion)bond55 (work_completion)(&(&slave->notify_work)->work) &obj_hash[i].lock irq_context: 0 (wq_completion)bond55 (work_completion)(&(&slave->notify_work)->work) &base->lock irq_context: 0 (wq_completion)bond55 (work_completion)(&(&slave->notify_work)->work) &base->lock &obj_hash[i].lock irq_context: 0 (wq_completion)bond55 (work_completion)(&(&slave->notify_work)->work) &rq->__lock irq_context: 0 (wq_completion)bond55 (work_completion)(&(&slave->notify_work)->work) &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)bond55 (work_completion)(&(&slave->notify_work)->work) rtnl_mutex net_rwsem irq_context: 0 (wq_completion)bond55 (work_completion)(&(&slave->notify_work)->work) rtnl_mutex net_rwsem &list->lock#2 irq_context: 0 (wq_completion)bond55 (work_completion)(&(&slave->notify_work)->work) rtnl_mutex &tn->lock irq_context: softirq rcu_read_lock rcu_read_lock slock-AF_INET/1 rcu_read_lock rcu_read_lock &____s->seqcount#12 irq_context: softirq rcu_read_lock rcu_read_lock rcu_read_lock &ul->lock irq_context: softirq rcu_read_lock rcu_read_lock rcu_read_lock &____s->seqcount#12 irq_context: softirq rcu_read_lock rcu_read_lock rcu_read_lock rcu_read_lock_bh &sch->q.lock &n->list_lock irq_context: softirq rcu_read_lock rcu_read_lock rcu_read_lock rcu_read_lock_bh &sch->q.lock &n->list_lock &c->lock irq_context: 0 (wq_completion)bond56 irq_context: 0 (wq_completion)bond56 (work_completion)(&(&slave->notify_work)->work) irq_context: 0 (wq_completion)bond56 (work_completion)(&(&slave->notify_work)->work) &obj_hash[i].lock irq_context: 0 (wq_completion)bond56 (work_completion)(&(&slave->notify_work)->work) &base->lock irq_context: 0 (wq_completion)bond56 (work_completion)(&(&slave->notify_work)->work) &base->lock &obj_hash[i].lock irq_context: 0 (wq_completion)bond56 (work_completion)(&(&slave->notify_work)->work) &rq->__lock irq_context: 0 (wq_completion)bond56 (work_completion)(&(&slave->notify_work)->work) &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)wg-kex-wg2#9 (work_completion)(&peer->transmit_handshake_work) &peer->endpoint_lock rcu_read_lock_bh rcu_read_lock batched_entropy_u8.lock irq_context: 0 (wq_completion)wg-kex-wg2#9 (work_completion)(&peer->transmit_handshake_work) &peer->endpoint_lock rcu_read_lock_bh rcu_read_lock kfence_freelist_lock irq_context: 0 (wq_completion)bond57 irq_context: 0 (wq_completion)bond57 (work_completion)(&(&slave->notify_work)->work) irq_context: 0 (wq_completion)bond57 (work_completion)(&(&slave->notify_work)->work) &obj_hash[i].lock irq_context: 0 (wq_completion)bond57 (work_completion)(&(&slave->notify_work)->work) &base->lock irq_context: 0 (wq_completion)bond57 (work_completion)(&(&slave->notify_work)->work) &base->lock &obj_hash[i].lock irq_context: 0 &f->f_pos_lock &type->i_mutex_dir_key#3 sb_writers#4 &wb->work_lock irq_context: 0 &f->f_pos_lock &type->i_mutex_dir_key#3 sb_writers#4 &wb->work_lock &obj_hash[i].lock irq_context: 0 &f->f_pos_lock &type->i_mutex_dir_key#3 sb_writers#4 &wb->work_lock &base->lock irq_context: 0 &f->f_pos_lock &type->i_mutex_dir_key#3 sb_writers#4 &wb->work_lock &base->lock &obj_hash[i].lock irq_context: 0 (wq_completion)bond57 (work_completion)(&(&slave->notify_work)->work) rtnl_mutex net_rwsem irq_context: 0 (wq_completion)bond57 (work_completion)(&(&slave->notify_work)->work) rtnl_mutex net_rwsem &list->lock#2 irq_context: 0 (wq_completion)bond57 (work_completion)(&(&slave->notify_work)->work) rtnl_mutex &tn->lock irq_context: 0 (wq_completion)bond56 (work_completion)(&(&slave->notify_work)->work) rtnl_mutex net_rwsem irq_context: 0 (wq_completion)bond56 (work_completion)(&(&slave->notify_work)->work) rtnl_mutex net_rwsem &list->lock#2 irq_context: 0 (wq_completion)bond56 (work_completion)(&(&slave->notify_work)->work) rtnl_mutex &tn->lock irq_context: softirq rcu_read_lock rcu_read_lock rcu_read_lock rcu_read_lock rcu_read_lock_bh &sch->q.lock &____s->seqcount#2 irq_context: softirq rcu_read_lock rcu_read_lock rcu_read_lock rcu_read_lock rcu_read_lock_bh &sch->q.lock &____s->seqcount irq_context: 0 (wq_completion)bond58 irq_context: 0 (wq_completion)bond58 (work_completion)(&(&slave->notify_work)->work) irq_context: 0 (wq_completion)bond58 (work_completion)(&(&slave->notify_work)->work) &obj_hash[i].lock irq_context: 0 (wq_completion)bond58 (work_completion)(&(&slave->notify_work)->work) &base->lock irq_context: 0 (wq_completion)bond58 (work_completion)(&(&slave->notify_work)->work) &base->lock &obj_hash[i].lock irq_context: 0 (wq_completion)bond58 (work_completion)(&(&slave->notify_work)->work) &rq->__lock irq_context: 0 (wq_completion)bond58 (work_completion)(&(&slave->notify_work)->work) &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)bond58 (work_completion)(&(&slave->notify_work)->work) rtnl_mutex net_rwsem irq_context: 0 (wq_completion)bond58 (work_completion)(&(&slave->notify_work)->work) rtnl_mutex net_rwsem &list->lock#2 irq_context: 0 (wq_completion)bond58 (work_completion)(&(&slave->notify_work)->work) rtnl_mutex &tn->lock irq_context: 0 (wq_completion)bond58 (work_completion)(&(&slave->notify_work)->work) rtnl_mutex.wait_lock irq_context: 0 (wq_completion)bond58 (work_completion)(&(&slave->notify_work)->work) &p->pi_lock irq_context: 0 (wq_completion)mld (work_completion)(&(&idev->mc_ifc_work)->work) &idev->mc_lock rcu_read_lock rcu_read_lock rcu_read_lock_bh rcu_read_lock rcu_read_lock_bh rcu_read_lock &br->multicast_lock &____s->seqcount#2 irq_context: 0 (wq_completion)mld (work_completion)(&(&idev->mc_ifc_work)->work) &idev->mc_lock rcu_read_lock rcu_read_lock rcu_read_lock_bh rcu_read_lock rcu_read_lock_bh rcu_read_lock &br->multicast_lock &____s->seqcount irq_context: 0 (wq_completion)bond59 irq_context: 0 (wq_completion)bond59 (work_completion)(&(&slave->notify_work)->work) irq_context: 0 (wq_completion)bond59 (work_completion)(&(&slave->notify_work)->work) &obj_hash[i].lock irq_context: 0 (wq_completion)bond59 (work_completion)(&(&slave->notify_work)->work) &base->lock irq_context: 0 (wq_completion)bond59 (work_completion)(&(&slave->notify_work)->work) &base->lock &obj_hash[i].lock irq_context: 0 (wq_completion)bond59 (work_completion)(&(&slave->notify_work)->work) &rq->__lock irq_context: 0 (wq_completion)bond59 (work_completion)(&(&slave->notify_work)->work) &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)bond59 (work_completion)(&(&slave->notify_work)->work) rtnl_mutex net_rwsem irq_context: 0 (wq_completion)bond59 (work_completion)(&(&slave->notify_work)->work) rtnl_mutex net_rwsem &list->lock#2 irq_context: 0 (wq_completion)bond59 (work_completion)(&(&slave->notify_work)->work) rtnl_mutex &tn->lock irq_context: 0 (wq_completion)bond60 irq_context: 0 (wq_completion)bond60 (work_completion)(&(&slave->notify_work)->work) irq_context: 0 (wq_completion)bond60 (work_completion)(&(&slave->notify_work)->work) &obj_hash[i].lock irq_context: 0 (wq_completion)bond60 (work_completion)(&(&slave->notify_work)->work) &base->lock irq_context: 0 (wq_completion)bond60 (work_completion)(&(&slave->notify_work)->work) &base->lock &obj_hash[i].lock irq_context: 0 (wq_completion)bond60 (work_completion)(&(&slave->notify_work)->work) &rq->__lock irq_context: 0 (wq_completion)bond60 (work_completion)(&(&slave->notify_work)->work) &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)bond43 irq_context: 0 (wq_completion)bond43 (work_completion)(&(&slave->notify_work)->work) irq_context: 0 (wq_completion)bond43 (work_completion)(&(&slave->notify_work)->work) &obj_hash[i].lock irq_context: 0 (wq_completion)bond43 (work_completion)(&(&slave->notify_work)->work) &base->lock irq_context: 0 (wq_completion)bond43 (work_completion)(&(&slave->notify_work)->work) &base->lock &obj_hash[i].lock irq_context: 0 (wq_completion)bond43 (work_completion)(&(&slave->notify_work)->work) &rq->__lock irq_context: 0 (wq_completion)bond43 (work_completion)(&(&slave->notify_work)->work) &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)bond60 (work_completion)(&(&slave->notify_work)->work) rtnl_mutex net_rwsem irq_context: 0 (wq_completion)bond60 (work_completion)(&(&slave->notify_work)->work) rtnl_mutex net_rwsem &list->lock#2 irq_context: 0 (wq_completion)bond60 (work_completion)(&(&slave->notify_work)->work) rtnl_mutex &tn->lock irq_context: 0 (wq_completion)bond60 (work_completion)(&(&slave->notify_work)->work) rtnl_mutex.wait_lock irq_context: 0 (wq_completion)bond60 (work_completion)(&(&slave->notify_work)->work) &p->pi_lock irq_context: 0 (wq_completion)bond43 (work_completion)(&(&slave->notify_work)->work) rtnl_mutex net_rwsem irq_context: 0 (wq_completion)bond43 (work_completion)(&(&slave->notify_work)->work) rtnl_mutex net_rwsem &list->lock#2 irq_context: 0 (wq_completion)bond43 (work_completion)(&(&slave->notify_work)->work) rtnl_mutex &tn->lock irq_context: 0 (wq_completion)bond43 (work_completion)(&(&slave->notify_work)->work) rtnl_mutex.wait_lock irq_context: 0 (wq_completion)bond43 (work_completion)(&(&slave->notify_work)->work) &p->pi_lock irq_context: 0 (wq_completion)bond44 irq_context: 0 (wq_completion)bond44 (work_completion)(&(&slave->notify_work)->work) irq_context: 0 (wq_completion)bond44 (work_completion)(&(&slave->notify_work)->work) &obj_hash[i].lock irq_context: 0 (wq_completion)bond44 (work_completion)(&(&slave->notify_work)->work) &base->lock irq_context: 0 (wq_completion)bond44 (work_completion)(&(&slave->notify_work)->work) &base->lock &obj_hash[i].lock irq_context: 0 (wq_completion)bond44 (work_completion)(&(&slave->notify_work)->work) rtnl_mutex net_rwsem irq_context: 0 (wq_completion)bond44 (work_completion)(&(&slave->notify_work)->work) rtnl_mutex net_rwsem &list->lock#2 irq_context: 0 (wq_completion)bond44 (work_completion)(&(&slave->notify_work)->work) rtnl_mutex &tn->lock irq_context: 0 (wq_completion)bond44 (work_completion)(&(&slave->notify_work)->work) rtnl_mutex.wait_lock irq_context: 0 (wq_completion)bond44 (work_completion)(&(&slave->notify_work)->work) &p->pi_lock irq_context: 0 ebt_mutex pgd_lock irq_context: 0 ebt_mutex stock_lock irq_context: 0 ebt_mutex rcu_read_lock pool_lock#2 irq_context: 0 ebt_mutex key irq_context: 0 ebt_mutex pcpu_lock irq_context: 0 ebt_mutex percpu_counters_lock irq_context: 0 ebt_mutex pcpu_lock stock_lock irq_context: 0 (wq_completion)events (linkwatch_work).work rtnl_mutex rcu_state.exp_mutex pool_lock#2 irq_context: 0 (wq_completion)bond62 irq_context: 0 (wq_completion)bond62 (work_completion)(&(&slave->notify_work)->work) irq_context: 0 (wq_completion)bond62 (work_completion)(&(&slave->notify_work)->work) &obj_hash[i].lock irq_context: 0 (wq_completion)bond62 (work_completion)(&(&slave->notify_work)->work) &base->lock irq_context: 0 (wq_completion)bond62 (work_completion)(&(&slave->notify_work)->work) &base->lock &obj_hash[i].lock irq_context: 0 (wq_completion)bond45 irq_context: 0 (wq_completion)bond45 (work_completion)(&(&slave->notify_work)->work) irq_context: 0 (wq_completion)bond45 (work_completion)(&(&slave->notify_work)->work) &obj_hash[i].lock irq_context: 0 (wq_completion)bond45 (work_completion)(&(&slave->notify_work)->work) &base->lock irq_context: 0 (wq_completion)bond45 (work_completion)(&(&slave->notify_work)->work) &base->lock &obj_hash[i].lock irq_context: 0 (wq_completion)bond45 (work_completion)(&(&slave->notify_work)->work) &rq->__lock irq_context: 0 (wq_completion)bond45 (work_completion)(&(&slave->notify_work)->work) &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)bond46 irq_context: 0 (wq_completion)bond46 (work_completion)(&(&slave->notify_work)->work) irq_context: 0 (wq_completion)bond46 (work_completion)(&(&slave->notify_work)->work) &obj_hash[i].lock irq_context: 0 (wq_completion)bond46 (work_completion)(&(&slave->notify_work)->work) &base->lock irq_context: 0 (wq_completion)bond46 (work_completion)(&(&slave->notify_work)->work) &base->lock &obj_hash[i].lock irq_context: 0 (wq_completion)bond46 (work_completion)(&(&slave->notify_work)->work) rtnl_mutex net_rwsem irq_context: 0 (wq_completion)bond46 (work_completion)(&(&slave->notify_work)->work) rtnl_mutex net_rwsem &list->lock#2 irq_context: 0 (wq_completion)bond46 (work_completion)(&(&slave->notify_work)->work) rtnl_mutex &tn->lock irq_context: 0 (wq_completion)bond46 (work_completion)(&(&slave->notify_work)->work) rtnl_mutex.wait_lock irq_context: 0 (wq_completion)bond46 (work_completion)(&(&slave->notify_work)->work) &p->pi_lock irq_context: 0 rtnl_mutex &br->hash_lock nl_table_wait.lock &p->pi_lock &rq->__lock irq_context: 0 rtnl_mutex &br->hash_lock nl_table_wait.lock &p->pi_lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)bond62 (work_completion)(&(&slave->notify_work)->work) rtnl_mutex net_rwsem irq_context: 0 (wq_completion)bond62 (work_completion)(&(&slave->notify_work)->work) rtnl_mutex net_rwsem &list->lock#2 irq_context: 0 (wq_completion)bond62 (work_completion)(&(&slave->notify_work)->work) rtnl_mutex &tn->lock irq_context: 0 (wq_completion)bond62 (work_completion)(&(&slave->notify_work)->work) rtnl_mutex.wait_lock irq_context: 0 (wq_completion)bond62 (work_completion)(&(&slave->notify_work)->work) &p->pi_lock irq_context: 0 (wq_completion)bond45 (work_completion)(&(&slave->notify_work)->work) rtnl_mutex net_rwsem irq_context: 0 (wq_completion)bond45 (work_completion)(&(&slave->notify_work)->work) rtnl_mutex net_rwsem &list->lock#2 irq_context: 0 (wq_completion)bond45 (work_completion)(&(&slave->notify_work)->work) rtnl_mutex &tn->lock irq_context: 0 (wq_completion)bond45 (work_completion)(&(&slave->notify_work)->work) rtnl_mutex.wait_lock irq_context: 0 (wq_completion)bond45 (work_completion)(&(&slave->notify_work)->work) &p->pi_lock irq_context: 0 (wq_completion)bond47 irq_context: 0 (wq_completion)bond47 (work_completion)(&(&slave->notify_work)->work) irq_context: 0 (wq_completion)bond47 (work_completion)(&(&slave->notify_work)->work) &obj_hash[i].lock irq_context: 0 (wq_completion)bond47 (work_completion)(&(&slave->notify_work)->work) &base->lock irq_context: 0 (wq_completion)bond47 (work_completion)(&(&slave->notify_work)->work) &base->lock &obj_hash[i].lock irq_context: 0 (wq_completion)bond47 (work_completion)(&(&slave->notify_work)->work) &rq->__lock irq_context: 0 (wq_completion)bond47 (work_completion)(&(&slave->notify_work)->work) &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)bat_events (work_completion)(&(&bat_priv->nc.work)->work) rcu_read_lock fill_pool_map-wait-type-override pool_lock#2 irq_context: 0 (wq_completion)bat_events (work_completion)(&(&bat_priv->nc.work)->work) rcu_read_lock fill_pool_map-wait-type-override pool_lock irq_context: 0 rtnl_mutex rcu_state.exp_mutex rcu_read_lock &pool->lock fill_pool_map-wait-type-override &____s->seqcount#2 irq_context: 0 rtnl_mutex rcu_state.exp_mutex rcu_read_lock &pool->lock fill_pool_map-wait-type-override &____s->seqcount irq_context: 0 (wq_completion)bond47 (work_completion)(&(&slave->notify_work)->work) rtnl_mutex net_rwsem irq_context: 0 (wq_completion)bond47 (work_completion)(&(&slave->notify_work)->work) rtnl_mutex net_rwsem &list->lock#2 irq_context: 0 (wq_completion)bond47 (work_completion)(&(&slave->notify_work)->work) rtnl_mutex &tn->lock irq_context: 0 &mm->mmap_lock &c->lock batched_entropy_u8.lock crngs.lock irq_context: 0 (wq_completion)bond48 irq_context: 0 (wq_completion)bond48 (work_completion)(&(&slave->notify_work)->work) irq_context: 0 (wq_completion)bond48 (work_completion)(&(&slave->notify_work)->work) &obj_hash[i].lock irq_context: 0 (wq_completion)bond48 (work_completion)(&(&slave->notify_work)->work) &base->lock irq_context: 0 (wq_completion)bond48 (work_completion)(&(&slave->notify_work)->work) &base->lock &obj_hash[i].lock irq_context: 0 (wq_completion)bond48 (work_completion)(&(&slave->notify_work)->work) &rq->__lock irq_context: 0 (wq_completion)bond48 (work_completion)(&(&slave->notify_work)->work) &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)bond48 (work_completion)(&(&slave->notify_work)->work) rtnl_mutex net_rwsem irq_context: 0 (wq_completion)bond48 (work_completion)(&(&slave->notify_work)->work) rtnl_mutex net_rwsem &list->lock#2 irq_context: 0 (wq_completion)bond48 (work_completion)(&(&slave->notify_work)->work) rtnl_mutex &tn->lock irq_context: 0 (wq_completion)bond49 irq_context: 0 (wq_completion)bond49 (work_completion)(&(&slave->notify_work)->work) irq_context: 0 (wq_completion)bond49 (work_completion)(&(&slave->notify_work)->work) &obj_hash[i].lock irq_context: 0 (wq_completion)bond49 (work_completion)(&(&slave->notify_work)->work) &base->lock irq_context: 0 (wq_completion)bond49 (work_completion)(&(&slave->notify_work)->work) &base->lock &obj_hash[i].lock irq_context: 0 (wq_completion)bond49 (work_completion)(&(&slave->notify_work)->work) &rq->__lock irq_context: 0 (wq_completion)bond49 (work_completion)(&(&slave->notify_work)->work) &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)bond22#3 irq_context: 0 (wq_completion)bond22#3 (work_completion)(&(&slave->notify_work)->work) irq_context: 0 (wq_completion)bond22#3 (work_completion)(&(&slave->notify_work)->work) &obj_hash[i].lock irq_context: 0 (wq_completion)bond22#3 (work_completion)(&(&slave->notify_work)->work) &base->lock irq_context: 0 (wq_completion)bond22#3 (work_completion)(&(&slave->notify_work)->work) &base->lock &obj_hash[i].lock irq_context: 0 (wq_completion)bond22#3 (work_completion)(&(&slave->notify_work)->work) &rq->__lock irq_context: 0 (wq_completion)bond22#3 (work_completion)(&(&slave->notify_work)->work) &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)bond22#3 (work_completion)(&(&slave->notify_work)->work) rtnl_mutex net_rwsem irq_context: 0 (wq_completion)bond22#3 (work_completion)(&(&slave->notify_work)->work) rtnl_mutex net_rwsem &rq->__lock irq_context: 0 (wq_completion)bond22#3 (work_completion)(&(&slave->notify_work)->work) rtnl_mutex net_rwsem &list->lock#2 irq_context: 0 (wq_completion)bond22#3 (work_completion)(&(&slave->notify_work)->work) rtnl_mutex net_rwsem &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)bond22#3 (work_completion)(&(&slave->notify_work)->work) rtnl_mutex &tn->lock irq_context: 0 (wq_completion)bond22#3 (work_completion)(&(&slave->notify_work)->work) rtnl_mutex &tn->lock &rq->__lock irq_context: 0 (wq_completion)bond22#3 (work_completion)(&(&slave->notify_work)->work) rtnl_mutex rcu_read_lock rcu_node_0 irq_context: 0 (wq_completion)bond22#3 (work_completion)(&(&slave->notify_work)->work) rtnl_mutex rcu_read_lock &rq->__lock irq_context: 0 (wq_completion)bond49 (work_completion)(&(&slave->notify_work)->work) rtnl_mutex net_rwsem irq_context: 0 (wq_completion)bond49 (work_completion)(&(&slave->notify_work)->work) rtnl_mutex net_rwsem &list->lock#2 irq_context: 0 (wq_completion)bond49 (work_completion)(&(&slave->notify_work)->work) rtnl_mutex &tn->lock irq_context: 0 (wq_completion)events (linkwatch_work).work rtnl_mutex rcu_node_0 irq_context: 0 (wq_completion)bond50 irq_context: 0 (wq_completion)bond50 (work_completion)(&(&slave->notify_work)->work) irq_context: 0 (wq_completion)bond50 (work_completion)(&(&slave->notify_work)->work) &obj_hash[i].lock irq_context: 0 (wq_completion)bond50 (work_completion)(&(&slave->notify_work)->work) &base->lock irq_context: 0 (wq_completion)bond50 (work_completion)(&(&slave->notify_work)->work) &base->lock &obj_hash[i].lock irq_context: 0 (wq_completion)bond23#2 irq_context: 0 (wq_completion)bond23#2 (work_completion)(&(&slave->notify_work)->work) irq_context: 0 (wq_completion)bond23#2 (work_completion)(&(&slave->notify_work)->work) &obj_hash[i].lock irq_context: 0 (wq_completion)bond23#2 (work_completion)(&(&slave->notify_work)->work) &base->lock irq_context: 0 (wq_completion)bond23#2 (work_completion)(&(&slave->notify_work)->work) &base->lock &obj_hash[i].lock irq_context: 0 (wq_completion)bond23#2 (work_completion)(&(&slave->notify_work)->work) &rq->__lock irq_context: 0 (wq_completion)bond50 (work_completion)(&(&slave->notify_work)->work) rtnl_mutex net_rwsem irq_context: 0 (wq_completion)bond50 (work_completion)(&(&slave->notify_work)->work) rtnl_mutex net_rwsem &list->lock#2 irq_context: 0 (wq_completion)bond50 (work_completion)(&(&slave->notify_work)->work) rtnl_mutex &tn->lock irq_context: 0 (wq_completion)bond23#2 (work_completion)(&(&slave->notify_work)->work) rtnl_mutex net_rwsem irq_context: 0 (wq_completion)bond23#2 (work_completion)(&(&slave->notify_work)->work) rtnl_mutex net_rwsem &list->lock#2 irq_context: 0 (wq_completion)bond23#2 (work_completion)(&(&slave->notify_work)->work) rtnl_mutex &tn->lock irq_context: 0 (wq_completion)bond24#2 irq_context: 0 (wq_completion)bond24#2 (work_completion)(&(&slave->notify_work)->work) irq_context: 0 (wq_completion)bond24#2 (work_completion)(&(&slave->notify_work)->work) &obj_hash[i].lock irq_context: 0 (wq_completion)bond24#2 (work_completion)(&(&slave->notify_work)->work) &base->lock irq_context: 0 (wq_completion)bond24#2 (work_completion)(&(&slave->notify_work)->work) &base->lock &obj_hash[i].lock irq_context: 0 (wq_completion)bond24#2 (work_completion)(&(&slave->notify_work)->work) &rq->__lock irq_context: 0 (wq_completion)bond24#2 (work_completion)(&(&slave->notify_work)->work) &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)bond51#2 irq_context: 0 (wq_completion)bond51#2 (work_completion)(&(&slave->notify_work)->work) irq_context: 0 (wq_completion)bond51#2 (work_completion)(&(&slave->notify_work)->work) &obj_hash[i].lock irq_context: 0 (wq_completion)bond51#2 (work_completion)(&(&slave->notify_work)->work) &base->lock irq_context: 0 (wq_completion)bond51#2 (work_completion)(&(&slave->notify_work)->work) &base->lock &obj_hash[i].lock irq_context: 0 (wq_completion)bond51#2 (work_completion)(&(&slave->notify_work)->work) &rq->__lock irq_context: 0 (wq_completion)bond51#2 (work_completion)(&(&slave->notify_work)->work) rtnl_mutex net_rwsem irq_context: 0 (wq_completion)bond51#2 (work_completion)(&(&slave->notify_work)->work) rtnl_mutex net_rwsem &list->lock#2 irq_context: 0 (wq_completion)bond51#2 (work_completion)(&(&slave->notify_work)->work) rtnl_mutex &tn->lock irq_context: 0 (wq_completion)bond51#2 (work_completion)(&(&slave->notify_work)->work) rtnl_mutex.wait_lock irq_context: 0 (wq_completion)bond51#2 (work_completion)(&(&slave->notify_work)->work) &p->pi_lock irq_context: 0 (wq_completion)bond24#2 (work_completion)(&(&slave->notify_work)->work) rtnl_mutex net_rwsem irq_context: 0 (wq_completion)bond24#2 (work_completion)(&(&slave->notify_work)->work) rtnl_mutex net_rwsem &list->lock#2 irq_context: 0 (wq_completion)bond24#2 (work_completion)(&(&slave->notify_work)->work) rtnl_mutex &tn->lock irq_context: 0 (wq_completion)bond24#2 (work_completion)(&(&slave->notify_work)->work) rtnl_mutex.wait_lock irq_context: 0 (wq_completion)bond24#2 (work_completion)(&(&slave->notify_work)->work) &p->pi_lock irq_context: 0 (wq_completion)ipv6_addrconf (work_completion)(&(&ifa->dad_work)->work) rtnl_mutex rcu_read_lock rcu_read_lock rcu_read_lock_bh rcu_read_lock rcu_read_lock_bh rcu_read_lock &obj_hash[i].lock pool_lock irq_context: 0 (wq_completion)ipv6_addrconf (work_completion)(&(&ifa->dad_work)->work) rtnl_mutex &idev->mc_lock &dev_addr_list_lock_key/1 rcu_read_lock &bridge_netdev_addr_lock_key &____s->seqcount irq_context: 0 (wq_completion)events (work_completion)(&(&krcp->monitor_work)->work) rcu_callback rcu_read_lock rcu_read_lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)bond25#2 irq_context: 0 (wq_completion)bond25#2 (work_completion)(&(&slave->notify_work)->work) irq_context: 0 (wq_completion)bond25#2 (work_completion)(&(&slave->notify_work)->work) &obj_hash[i].lock irq_context: 0 (wq_completion)bond25#2 (work_completion)(&(&slave->notify_work)->work) &base->lock irq_context: 0 (wq_completion)bond25#2 (work_completion)(&(&slave->notify_work)->work) &base->lock &obj_hash[i].lock irq_context: 0 (wq_completion)bond25#2 (work_completion)(&(&slave->notify_work)->work) &rq->__lock irq_context: 0 (wq_completion)bond25#2 (work_completion)(&(&slave->notify_work)->work) &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)bond52#2 irq_context: 0 (wq_completion)bond52#2 &rq->__lock irq_context: 0 (wq_completion)bond52#2 &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)bond52#2 (work_completion)(&(&slave->notify_work)->work) irq_context: 0 (wq_completion)bond52#2 (work_completion)(&(&slave->notify_work)->work) &obj_hash[i].lock irq_context: 0 (wq_completion)bond52#2 (work_completion)(&(&slave->notify_work)->work) &base->lock irq_context: 0 (wq_completion)bond52#2 (work_completion)(&(&slave->notify_work)->work) &base->lock &obj_hash[i].lock irq_context: 0 (wq_completion)bond52#2 (work_completion)(&(&slave->notify_work)->work) rtnl_mutex net_rwsem irq_context: 0 (wq_completion)bond52#2 (work_completion)(&(&slave->notify_work)->work) rtnl_mutex net_rwsem &list->lock#2 irq_context: 0 (wq_completion)bond52#2 (work_completion)(&(&slave->notify_work)->work) rtnl_mutex &tn->lock irq_context: 0 (wq_completion)bond52#2 (work_completion)(&(&slave->notify_work)->work) rtnl_mutex.wait_lock irq_context: 0 (wq_completion)bond52#2 (work_completion)(&(&slave->notify_work)->work) &p->pi_lock irq_context: 0 (wq_completion)bond25#2 (work_completion)(&(&slave->notify_work)->work) rtnl_mutex net_rwsem irq_context: 0 (wq_completion)bond25#2 (work_completion)(&(&slave->notify_work)->work) rtnl_mutex net_rwsem &list->lock#2 irq_context: 0 (wq_completion)bond25#2 (work_completion)(&(&slave->notify_work)->work) rtnl_mutex &tn->lock irq_context: 0 (wq_completion)mld (work_completion)(&(&idev->mc_ifc_work)->work) &idev->mc_lock rcu_read_lock rcu_read_lock rcu_read_lock_bh rcu_read_lock rcu_read_lock_bh rcu_read_lock &br->multicast_lock &pcp->lock &zone->lock irq_context: 0 (wq_completion)bond26#3 irq_context: 0 (wq_completion)bond26#3 (work_completion)(&(&slave->notify_work)->work) irq_context: 0 (wq_completion)bond26#3 (work_completion)(&(&slave->notify_work)->work) &obj_hash[i].lock irq_context: 0 (wq_completion)bond26#3 (work_completion)(&(&slave->notify_work)->work) &base->lock irq_context: 0 (wq_completion)bond26#3 (work_completion)(&(&slave->notify_work)->work) &base->lock &obj_hash[i].lock irq_context: 0 (wq_completion)bond26#3 (work_completion)(&(&slave->notify_work)->work) &rq->__lock irq_context: 0 (wq_completion)bond53#2 irq_context: 0 (wq_completion)bond53#2 (work_completion)(&(&slave->notify_work)->work) irq_context: 0 (wq_completion)bond53#2 (work_completion)(&(&slave->notify_work)->work) &obj_hash[i].lock irq_context: 0 (wq_completion)bond53#2 (work_completion)(&(&slave->notify_work)->work) &base->lock irq_context: 0 (wq_completion)bond53#2 (work_completion)(&(&slave->notify_work)->work) &base->lock &obj_hash[i].lock irq_context: 0 (wq_completion)bond26#3 (work_completion)(&(&slave->notify_work)->work) rtnl_mutex net_rwsem irq_context: 0 (wq_completion)bond26#3 (work_completion)(&(&slave->notify_work)->work) rtnl_mutex net_rwsem &list->lock#2 irq_context: 0 (wq_completion)bond26#3 (work_completion)(&(&slave->notify_work)->work) rtnl_mutex &tn->lock irq_context: 0 (wq_completion)bond26#3 (work_completion)(&(&slave->notify_work)->work) rtnl_mutex.wait_lock irq_context: 0 (wq_completion)bond26#3 (work_completion)(&(&slave->notify_work)->work) &p->pi_lock irq_context: 0 (wq_completion)events (linkwatch_work).work rtnl_mutex remove_cache_srcu pool_lock#2 irq_context: 0 (wq_completion)bond53#2 (work_completion)(&(&slave->notify_work)->work) rtnl_mutex net_rwsem irq_context: 0 (wq_completion)bond53#2 (work_completion)(&(&slave->notify_work)->work) rtnl_mutex net_rwsem &list->lock#2 irq_context: 0 (wq_completion)bond53#2 (work_completion)(&(&slave->notify_work)->work) rtnl_mutex &tn->lock irq_context: 0 (wq_completion)bond54#2 irq_context: 0 (wq_completion)bond54#2 (work_completion)(&(&slave->notify_work)->work) irq_context: 0 (wq_completion)bond54#2 (work_completion)(&(&slave->notify_work)->work) &obj_hash[i].lock irq_context: 0 (wq_completion)bond54#2 (work_completion)(&(&slave->notify_work)->work) &base->lock irq_context: 0 (wq_completion)bond54#2 (work_completion)(&(&slave->notify_work)->work) &base->lock &obj_hash[i].lock irq_context: 0 (wq_completion)bond54#2 (work_completion)(&(&slave->notify_work)->work) &rq->__lock irq_context: 0 (wq_completion)bond54#2 (work_completion)(&(&slave->notify_work)->work) &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)bond27#3 irq_context: 0 (wq_completion)bond27#3 (work_completion)(&(&slave->notify_work)->work) irq_context: 0 (wq_completion)bond27#3 (work_completion)(&(&slave->notify_work)->work) &obj_hash[i].lock irq_context: 0 (wq_completion)bond27#3 (work_completion)(&(&slave->notify_work)->work) &base->lock irq_context: 0 (wq_completion)bond27#3 (work_completion)(&(&slave->notify_work)->work) &base->lock &obj_hash[i].lock irq_context: 0 (wq_completion)bond27#3 (work_completion)(&(&slave->notify_work)->work) &rq->__lock irq_context: 0 (wq_completion)bond27#3 (work_completion)(&(&slave->notify_work)->work) &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)bond54#2 (work_completion)(&(&slave->notify_work)->work) rtnl_mutex net_rwsem irq_context: 0 (wq_completion)bond54#2 (work_completion)(&(&slave->notify_work)->work) rtnl_mutex net_rwsem &list->lock#2 irq_context: 0 (wq_completion)bond54#2 (work_completion)(&(&slave->notify_work)->work) rtnl_mutex &tn->lock irq_context: 0 (wq_completion)bond54#2 (work_completion)(&(&slave->notify_work)->work) rtnl_mutex.wait_lock irq_context: 0 (wq_completion)bond54#2 (work_completion)(&(&slave->notify_work)->work) &p->pi_lock irq_context: 0 (wq_completion)bond27#3 (work_completion)(&(&slave->notify_work)->work) rtnl_mutex net_rwsem irq_context: 0 (wq_completion)bond27#3 (work_completion)(&(&slave->notify_work)->work) rtnl_mutex net_rwsem &list->lock#2 irq_context: 0 (wq_completion)bond27#3 (work_completion)(&(&slave->notify_work)->work) rtnl_mutex &tn->lock irq_context: 0 (wq_completion)bond27#3 (work_completion)(&(&slave->notify_work)->work) rtnl_mutex.wait_lock irq_context: 0 (wq_completion)bond27#3 (work_completion)(&(&slave->notify_work)->work) &p->pi_lock irq_context: 0 (wq_completion)mld (work_completion)(&(&idev->mc_ifc_work)->work) &idev->mc_lock rcu_read_lock rcu_read_lock rcu_read_lock_bh rcu_read_lock rcu_read_lock_bh rcu_read_lock &obj_hash[i].lock pool_lock irq_context: 0 (wq_completion)bond28#3 irq_context: 0 (wq_completion)bond28#3 (work_completion)(&(&slave->notify_work)->work) irq_context: 0 (wq_completion)bond28#3 (work_completion)(&(&slave->notify_work)->work) &obj_hash[i].lock irq_context: 0 (wq_completion)bond28#3 (work_completion)(&(&slave->notify_work)->work) &base->lock irq_context: 0 (wq_completion)bond28#3 (work_completion)(&(&slave->notify_work)->work) &base->lock &obj_hash[i].lock irq_context: 0 &f->f_pos_lock &type->i_mutex_dir_key#3 sb_writers#4 jbd2_handle rcu_read_lock &rq->__lock &cfs_rq->removed.lock irq_context: 0 (wq_completion)bond28#3 (work_completion)(&(&slave->notify_work)->work) rtnl_mutex net_rwsem irq_context: 0 (wq_completion)bond28#3 (work_completion)(&(&slave->notify_work)->work) rtnl_mutex net_rwsem &list->lock#2 irq_context: 0 (wq_completion)bond28#3 (work_completion)(&(&slave->notify_work)->work) rtnl_mutex &tn->lock irq_context: 0 (wq_completion)bond28#3 (work_completion)(&(&slave->notify_work)->work) rtnl_mutex.wait_lock irq_context: 0 (wq_completion)bond28#3 (work_completion)(&(&slave->notify_work)->work) &p->pi_lock irq_context: 0 (wq_completion)bond29#3 irq_context: 0 (wq_completion)bond29#3 (work_completion)(&(&slave->notify_work)->work) irq_context: 0 (wq_completion)bond29#3 (work_completion)(&(&slave->notify_work)->work) &obj_hash[i].lock irq_context: 0 (wq_completion)bond29#3 (work_completion)(&(&slave->notify_work)->work) &base->lock irq_context: 0 (wq_completion)bond29#3 (work_completion)(&(&slave->notify_work)->work) &base->lock &obj_hash[i].lock irq_context: 0 (wq_completion)bond29#3 (work_completion)(&(&slave->notify_work)->work) &rq->__lock irq_context: 0 (wq_completion)bond29#3 (work_completion)(&(&slave->notify_work)->work) &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)bond30#3 irq_context: 0 (wq_completion)bond30#3 (work_completion)(&(&slave->notify_work)->work) irq_context: 0 (wq_completion)bond30#3 (work_completion)(&(&slave->notify_work)->work) &obj_hash[i].lock irq_context: 0 (wq_completion)bond30#3 (work_completion)(&(&slave->notify_work)->work) &base->lock irq_context: 0 (wq_completion)bond30#3 (work_completion)(&(&slave->notify_work)->work) &base->lock &obj_hash[i].lock irq_context: 0 (wq_completion)bond30#3 (work_completion)(&(&slave->notify_work)->work) &rq->__lock irq_context: 0 (wq_completion)bond30#3 (work_completion)(&(&slave->notify_work)->work) &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)events (work_completion)(&data->fib_event_work) &data->fib_lock fill_pool_map-wait-type-override &c->lock irq_context: 0 (wq_completion)events (work_completion)(&data->fib_event_work) &data->fib_lock fill_pool_map-wait-type-override &____s->seqcount#2 irq_context: 0 (wq_completion)events (work_completion)(&data->fib_event_work) &data->fib_lock fill_pool_map-wait-type-override &____s->seqcount irq_context: 0 (wq_completion)bond30#3 (work_completion)(&(&slave->notify_work)->work) rtnl_mutex net_rwsem irq_context: 0 (wq_completion)bond30#3 (work_completion)(&(&slave->notify_work)->work) rtnl_mutex net_rwsem &list->lock#2 irq_context: 0 (wq_completion)bond30#3 (work_completion)(&(&slave->notify_work)->work) rtnl_mutex &tn->lock irq_context: 0 (wq_completion)bond30#3 (work_completion)(&(&slave->notify_work)->work) rtnl_mutex.wait_lock irq_context: 0 (wq_completion)bond30#3 (work_completion)(&(&slave->notify_work)->work) &p->pi_lock irq_context: 0 (wq_completion)bond29#3 (work_completion)(&(&slave->notify_work)->work) rtnl_mutex net_rwsem irq_context: 0 (wq_completion)bond29#3 (work_completion)(&(&slave->notify_work)->work) rtnl_mutex net_rwsem &list->lock#2 irq_context: 0 (wq_completion)bond29#3 (work_completion)(&(&slave->notify_work)->work) rtnl_mutex &tn->lock irq_context: 0 (wq_completion)bond29#3 (work_completion)(&(&slave->notify_work)->work) rtnl_mutex.wait_lock irq_context: 0 (wq_completion)bond29#3 (work_completion)(&(&slave->notify_work)->work) &p->pi_lock irq_context: 0 (wq_completion)bond31#2 irq_context: 0 (wq_completion)bond31#2 (work_completion)(&(&slave->notify_work)->work) irq_context: 0 (wq_completion)bond31#2 (work_completion)(&(&slave->notify_work)->work) &obj_hash[i].lock irq_context: 0 (wq_completion)bond31#2 (work_completion)(&(&slave->notify_work)->work) &base->lock irq_context: 0 (wq_completion)bond31#2 (work_completion)(&(&slave->notify_work)->work) &base->lock &obj_hash[i].lock irq_context: 0 rtnl_mutex rcu_state.barrier_mutex &rq->__lock &cfs_rq->removed.lock irq_context: 0 (wq_completion)bond32#2 irq_context: 0 (wq_completion)bond32#2 (work_completion)(&(&slave->notify_work)->work) irq_context: 0 (wq_completion)bond32#2 (work_completion)(&(&slave->notify_work)->work) &obj_hash[i].lock irq_context: 0 (wq_completion)bond32#2 (work_completion)(&(&slave->notify_work)->work) &base->lock irq_context: 0 (wq_completion)bond32#2 (work_completion)(&(&slave->notify_work)->work) &base->lock &obj_hash[i].lock irq_context: 0 (wq_completion)bond31#2 (work_completion)(&(&slave->notify_work)->work) rtnl_mutex net_rwsem irq_context: 0 (wq_completion)bond31#2 (work_completion)(&(&slave->notify_work)->work) rtnl_mutex net_rwsem &list->lock#2 irq_context: 0 (wq_completion)bond31#2 (work_completion)(&(&slave->notify_work)->work) rtnl_mutex &tn->lock irq_context: 0 (wq_completion)bond31#2 (work_completion)(&(&slave->notify_work)->work) rtnl_mutex.wait_lock irq_context: 0 (wq_completion)bond31#2 (work_completion)(&(&slave->notify_work)->work) &p->pi_lock irq_context: 0 (wq_completion)bond33 irq_context: 0 (wq_completion)bond33 (work_completion)(&(&slave->notify_work)->work) irq_context: 0 (wq_completion)bond33 (work_completion)(&(&slave->notify_work)->work) &obj_hash[i].lock irq_context: 0 (wq_completion)bond33 (work_completion)(&(&slave->notify_work)->work) &base->lock irq_context: 0 (wq_completion)bond33 (work_completion)(&(&slave->notify_work)->work) &base->lock &obj_hash[i].lock irq_context: 0 (wq_completion)bond33 (work_completion)(&(&slave->notify_work)->work) &rq->__lock irq_context: 0 (wq_completion)ipv6_addrconf (work_completion)(&(&ifa->dad_work)->work) rtnl_mutex rcu_read_lock rcu_read_lock rcu_read_lock_bh rcu_read_lock rcu_read_lock_bh rcu_read_lock &data->lock &obj_hash[i].lock irq_context: 0 (wq_completion)ipv6_addrconf (work_completion)(&(&ifa->dad_work)->work) rtnl_mutex rcu_read_lock rcu_read_lock rcu_read_lock_bh rcu_read_lock rcu_read_lock_bh rcu_read_lock &data->lock &base->lock irq_context: 0 (wq_completion)ipv6_addrconf (work_completion)(&(&ifa->dad_work)->work) rtnl_mutex rcu_read_lock rcu_read_lock rcu_read_lock_bh rcu_read_lock rcu_read_lock_bh rcu_read_lock &data->lock &base->lock &obj_hash[i].lock irq_context: 0 sk_lock-AF_INET k-sk_lock-AF_INET rcu_read_lock &____s->seqcount#12 irq_context: 0 (wq_completion)bond33 (work_completion)(&(&slave->notify_work)->work) rtnl_mutex net_rwsem irq_context: 0 (wq_completion)bond33 (work_completion)(&(&slave->notify_work)->work) rtnl_mutex net_rwsem &list->lock#2 irq_context: 0 (wq_completion)bond33 (work_completion)(&(&slave->notify_work)->work) rtnl_mutex &tn->lock irq_context: 0 (wq_completion)bond33 (work_completion)(&(&slave->notify_work)->work) rtnl_mutex.wait_lock irq_context: 0 (wq_completion)bond33 (work_completion)(&(&slave->notify_work)->work) &p->pi_lock irq_context: 0 sk_lock-AF_INET k-sk_lock-AF_INET rcu_read_lock rcu_read_lock &cfs_rq->removed.lock irq_context: 0 sk_lock-AF_INET k-sk_lock-AF_INET &tcp_hashinfo.bhash[i].lock &tcp_hashinfo.bhash2[i].lock &c->lock irq_context: 0 sk_lock-AF_INET k-sk_lock-AF_INET batched_entropy_u32.lock crngs.lock irq_context: softirq rcu_read_lock rcu_read_lock k-slock-AF_INET/1 rcu_read_lock &pool->lock &p->pi_lock irq_context: softirq rcu_read_lock rcu_read_lock k-slock-AF_INET/1 rcu_read_lock &pool->lock &p->pi_lock &rq->__lock irq_context: softirq rcu_read_lock rcu_read_lock k-slock-AF_INET/1 rcu_read_lock &pool->lock &p->pi_lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)bond32#2 (work_completion)(&(&slave->notify_work)->work) rtnl_mutex net_rwsem irq_context: 0 (wq_completion)bond32#2 (work_completion)(&(&slave->notify_work)->work) rtnl_mutex net_rwsem &list->lock#2 irq_context: 0 (wq_completion)bond32#2 (work_completion)(&(&slave->notify_work)->work) rtnl_mutex &tn->lock irq_context: 0 (wq_completion)bond32#2 (work_completion)(&(&slave->notify_work)->work) rtnl_mutex.wait_lock irq_context: 0 (wq_completion)bond32#2 (work_completion)(&(&slave->notify_work)->work) &p->pi_lock irq_context: 0 (wq_completion)mld (work_completion)(&(&idev->mc_dad_work)->work) &idev->mc_lock rcu_read_lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 &dir->lock#2 &meta->lock irq_context: 0 &dir->lock#2 kfence_freelist_lock irq_context: 0 rtnl_mutex &idev->mc_lock &pcp->lock &zone->lock irq_context: softirq (&app->join_timer) &app->lock &zone->lock irq_context: 0 (wq_completion)writeback (work_completion)(&(&wb->dwork)->work) &type->s_umount_key#43 rcu_read_lock &xa->xa_lock#9 &obj_hash[i].lock irq_context: 0 (wq_completion)writeback (work_completion)(&(&wb->dwork)->work) &type->s_umount_key#43 rcu_read_lock &xa->xa_lock#9 &base->lock irq_context: 0 (wq_completion)writeback (work_completion)(&(&wb->dwork)->work) &type->s_umount_key#43 rcu_read_lock &xa->xa_lock#9 &base->lock &obj_hash[i].lock irq_context: 0 rtnl_mutex &hwstats->hwsdev_list_lock &rq->__lock irq_context: 0 rtnl_mutex &hwstats->hwsdev_list_lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 &type->i_mutex_dir_key#3 fs_reclaim &rq->__lock irq_context: 0 &type->i_mutex_dir_key#3 fs_reclaim &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 rtnl_mutex &xa->xa_lock#3 &pcp->lock &zone->lock irq_context: 0 rtnl_mutex &xa->xa_lock#3 &pcp->lock &zone->lock &____s->seqcount irq_context: 0 sb_writers#4 fill_pool_map-wait-type-override &rq->__lock &cfs_rq->removed.lock irq_context: 0 (wq_completion)bat_events (work_completion)(&(&bat_priv->dat.work)->work) fill_pool_map-wait-type-override pool_lock irq_context: softirq &(&nsim_dev->trap_data->trap_report_dw)->timer rcu_read_lock &pool->lock fill_pool_map-wait-type-override pool_lock irq_context: 0 fs_reclaim rcu_node_0 irq_context: 0 fs_reclaim &rcu_state.expedited_wq irq_context: 0 fs_reclaim &rcu_state.expedited_wq &p->pi_lock irq_context: 0 fs_reclaim &rcu_state.expedited_wq &p->pi_lock &rq->__lock irq_context: 0 fs_reclaim &rcu_state.expedited_wq &p->pi_lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: softirq (&ndev->rs_timer) rcu_read_lock rcu_read_lock &zone->lock irq_context: 0 fs_reclaim pool_lock#2 irq_context: 0 sk_lock-AF_PACKET &po->bind_lock rcu_read_lock &____s->seqcount irq_context: 0 &mm->mmap_lock fill_pool_map-wait-type-override &____s->seqcount#2 irq_context: 0 &mm->mmap_lock fill_pool_map-wait-type-override &____s->seqcount irq_context: 0 (wq_completion)events (linkwatch_work).work rtnl_mutex rcu_state.exp_mutex rcu_read_lock &rq->__lock irq_context: 0 (wq_completion)events (linkwatch_work).work rtnl_mutex rcu_state.exp_mutex rcu_read_lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 rtnl_mutex uevent_sock_mutex fs_reclaim &rq->__lock &cfs_rq->removed.lock irq_context: 0 rtnl_mutex remove_cache_srcu &____s->seqcount irq_context: 0 cgroup_threadgroup_rwsem &p->pi_lock &cfs_rq->removed.lock irq_context: 0 sk_lock-AF_INET rcu_read_lock batched_entropy_u8.lock irq_context: 0 sk_lock-AF_INET rcu_read_lock kfence_freelist_lock irq_context: 0 sk_lock-AF_INET rcu_read_lock rcu_read_lock_bh quarantine_lock irq_context: 0 tasklist_lock &sighand->siglock &(&sig->stats_lock)->lock &____s->seqcount#5 pidmap_lock fill_pool_map-wait-type-override batched_entropy_u8.lock irq_context: 0 tasklist_lock &sighand->siglock &(&sig->stats_lock)->lock &____s->seqcount#5 pidmap_lock fill_pool_map-wait-type-override kfence_freelist_lock irq_context: 0 (wq_completion)bat_events (work_completion)(&(&bat_priv->orig_work)->work) fill_pool_map-wait-type-override &____s->seqcount#2 irq_context: 0 (wq_completion)bat_events (work_completion)(&(&bat_priv->orig_work)->work) fill_pool_map-wait-type-override &____s->seqcount irq_context: softirq rcu_read_lock &rdev->bss_lock &____s->seqcount#2 irq_context: softirq rcu_read_lock &rdev->bss_lock &____s->seqcount irq_context: 0 &f->f_pos_lock &type->i_mutex_dir_key#3 batched_entropy_u8.lock crngs.lock irq_context: 0 sk_lock-AF_INET rcu_read_lock rcu_read_lock &cfs_rq->removed.lock irq_context: 0 &sb->s_type->i_mutex_key#10 &net->sctp.addr_wq_lock slock-AF_INET/1 fill_pool_map-wait-type-override pool_lock#2 irq_context: 0 &sb->s_type->i_mutex_key#10 &net->sctp.addr_wq_lock slock-AF_INET/1 fill_pool_map-wait-type-override pool_lock irq_context: 0 sk_lock-AF_INET6 free_vmap_area_lock &obj_hash[i].lock irq_context: 0 sk_lock-AF_INET6 free_vmap_area_lock pool_lock#2 irq_context: 0 rtnl_mutex lock kernfs_idr_lock &pcp->lock &zone->lock &____s->seqcount irq_context: 0 (wq_completion)bond79 irq_context: 0 (wq_completion)bond79 (work_completion)(&(&slave->notify_work)->work) irq_context: 0 (wq_completion)bond79 (work_completion)(&(&slave->notify_work)->work) &obj_hash[i].lock irq_context: 0 (wq_completion)bond79 (work_completion)(&(&slave->notify_work)->work) &base->lock irq_context: 0 (wq_completion)bond79 (work_completion)(&(&slave->notify_work)->work) &base->lock &obj_hash[i].lock irq_context: 0 (wq_completion)bond79 (work_completion)(&(&slave->notify_work)->work) &rq->__lock irq_context: 0 (wq_completion)bond79 (work_completion)(&(&slave->notify_work)->work) &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)bond79 (work_completion)(&(&slave->notify_work)->work) rtnl_mutex net_rwsem irq_context: 0 (wq_completion)bond79 (work_completion)(&(&slave->notify_work)->work) rtnl_mutex net_rwsem &list->lock#2 irq_context: 0 (wq_completion)bond79 (work_completion)(&(&slave->notify_work)->work) rtnl_mutex &tn->lock irq_context: 0 (wq_completion)bond79 (work_completion)(&(&slave->notify_work)->work) rtnl_mutex.wait_lock irq_context: 0 (wq_completion)bond79 (work_completion)(&(&slave->notify_work)->work) &p->pi_lock irq_context: 0 (wq_completion)bond80 irq_context: 0 (wq_completion)bond80 (work_completion)(&(&slave->notify_work)->work) irq_context: 0 (wq_completion)bond80 (work_completion)(&(&slave->notify_work)->work) &obj_hash[i].lock irq_context: 0 (wq_completion)bond80 (work_completion)(&(&slave->notify_work)->work) &base->lock irq_context: 0 (wq_completion)bond80 (work_completion)(&(&slave->notify_work)->work) &base->lock &obj_hash[i].lock irq_context: 0 (wq_completion)bond80 (work_completion)(&(&slave->notify_work)->work) rtnl_mutex net_rwsem irq_context: 0 (wq_completion)bond80 (work_completion)(&(&slave->notify_work)->work) rtnl_mutex net_rwsem &list->lock#2 irq_context: 0 (wq_completion)bond80 (work_completion)(&(&slave->notify_work)->work) rtnl_mutex &tn->lock irq_context: 0 sk_lock-AF_INET6 &dccp_hashinfo.bhash[i].lock &dccp_hashinfo.bhash2[i].lock per_cpu_ptr(&cgroup_rstat_cpu_lock, cpu) irq_context: 0 (wq_completion)wg-kex-wg0#12 (work_completion)(&({ do { const void *__vpp_verify = (typeof((worker) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __asm__ ("" : "=r"(__ptr) : "0"((typeof(*((worker))) *)((worker)))); (typeof((typeof(*((worker))) *)((worker)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); })->work) &wg->static_identity.lock &handshake->lock &rq->__lock irq_context: 0 (wq_completion)wg-kex-wg0#12 (work_completion)(&({ do { const void *__vpp_verify = (typeof((worker) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __asm__ ("" : "=r"(__ptr) : "0"((typeof(*((worker))) *)((worker)))); (typeof((typeof(*((worker))) *)((worker)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); })->work) &wg->static_identity.lock &handshake->lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)wg-kex-wg0#12 (work_completion)(&({ do { const void *__vpp_verify = (typeof((worker) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __asm__ ("" : "=r"(__ptr) : "0"((typeof(*((worker))) *)((worker)))); (typeof((typeof(*((worker))) *)((worker)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); })->work) &handshake->lock remove_cache_srcu &rq->__lock irq_context: 0 (wq_completion)wg-kex-wg0#12 (work_completion)(&({ do { const void *__vpp_verify = (typeof((worker) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __asm__ ("" : "=r"(__ptr) : "0"((typeof(*((worker))) *)((worker)))); (typeof((typeof(*((worker))) *)((worker)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); })->work) &handshake->lock remove_cache_srcu &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 sk_lock-AF_INET rcu_read_lock rcu_read_lock &pcp->lock &zone->lock irq_context: 0 &sb->s_type->i_mutex_key#10 sk_lock-AF_INET6 slock-AF_INET6 fill_pool_map-wait-type-override &____s->seqcount#2 irq_context: 0 &sb->s_type->i_mutex_key#10 sk_lock-AF_INET6 slock-AF_INET6 fill_pool_map-wait-type-override &____s->seqcount irq_context: 0 &f->f_pos_lock &type->i_mutex_dir_key#3 remove_cache_srcu rcu_node_0 irq_context: 0 (wq_completion)bond81 irq_context: 0 (wq_completion)bond81 (work_completion)(&(&slave->notify_work)->work) irq_context: 0 (wq_completion)bond81 (work_completion)(&(&slave->notify_work)->work) &obj_hash[i].lock irq_context: 0 (wq_completion)bond81 (work_completion)(&(&slave->notify_work)->work) &base->lock irq_context: 0 (wq_completion)bond81 (work_completion)(&(&slave->notify_work)->work) &base->lock &obj_hash[i].lock irq_context: 0 (wq_completion)bond81 (work_completion)(&(&slave->notify_work)->work) &rq->__lock irq_context: 0 (wq_completion)bond81 (work_completion)(&(&slave->notify_work)->work) &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)bond81 (work_completion)(&(&bond->mcast_work)->work) irq_context: 0 (wq_completion)bond81 (work_completion)(&(&bond->mcast_work)->work) &obj_hash[i].lock irq_context: 0 (wq_completion)bond81 (work_completion)(&(&bond->mcast_work)->work) &base->lock irq_context: 0 (wq_completion)bond81 (work_completion)(&(&bond->mcast_work)->work) &base->lock &obj_hash[i].lock irq_context: 0 (wq_completion)bond81 (work_completion)(&(&bond->mcast_work)->work) &rq->__lock irq_context: 0 (wq_completion)bond81 (work_completion)(&(&bond->mcast_work)->work) &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)bond81 (work_completion)(&(&slave->notify_work)->work) rtnl_mutex net_rwsem irq_context: 0 (wq_completion)bond81 (work_completion)(&(&slave->notify_work)->work) rtnl_mutex net_rwsem &list->lock#2 irq_context: 0 (wq_completion)bond81 (work_completion)(&(&slave->notify_work)->work) rtnl_mutex &tn->lock irq_context: 0 (wq_completion)bond81 (work_completion)(&(&slave->notify_work)->work) rtnl_mutex.wait_lock irq_context: 0 (wq_completion)bond81 (work_completion)(&(&slave->notify_work)->work) &p->pi_lock irq_context: 0 (wq_completion)bond81 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex fs_reclaim irq_context: 0 (wq_completion)bond81 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex fs_reclaim mmu_notifier_invalidate_range_start irq_context: 0 (wq_completion)bond81 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &c->lock irq_context: 0 (wq_completion)bond81 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &n->list_lock irq_context: 0 (wq_completion)bond81 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &n->list_lock &c->lock irq_context: 0 (wq_completion)bond81 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex rcu_read_lock &bond->stats_lock/1 irq_context: 0 (wq_completion)bond81 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex rcu_read_lock &ndev->lock irq_context: 0 (wq_completion)bond81 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex nl_table_lock irq_context: 0 (wq_completion)bond81 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &obj_hash[i].lock irq_context: 0 (wq_completion)bond81 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex nl_table_wait.lock irq_context: 0 (wq_completion)bond81 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex net_rwsem irq_context: 0 (wq_completion)bond81 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex net_rwsem &list->lock#2 irq_context: 0 (wq_completion)bond81 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &tn->lock irq_context: 0 (wq_completion)bond81 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &idev->mc_lock irq_context: 0 (wq_completion)bond81 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex.wait_lock irq_context: 0 (wq_completion)bond81 (work_completion)(&(&bond->mcast_work)->work) &p->pi_lock irq_context: 0 (wq_completion)bond62#2 irq_context: 0 (wq_completion)bond62#2 &rq->__lock irq_context: 0 (wq_completion)bond62#2 &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)bond62#2 (work_completion)(&(&slave->notify_work)->work) irq_context: 0 (wq_completion)bond62#2 (work_completion)(&(&slave->notify_work)->work) &obj_hash[i].lock irq_context: 0 (wq_completion)bond62#2 (work_completion)(&(&slave->notify_work)->work) &base->lock irq_context: 0 (wq_completion)bond62#2 (work_completion)(&(&slave->notify_work)->work) &base->lock &obj_hash[i].lock irq_context: 0 (wq_completion)bond62#2 (work_completion)(&(&slave->notify_work)->work) &rq->__lock irq_context: 0 (wq_completion)bond62#2 (work_completion)(&(&slave->notify_work)->work) &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)bond62#2 (work_completion)(&(&bond->mcast_work)->work) irq_context: 0 (wq_completion)bond62#2 (work_completion)(&(&bond->mcast_work)->work) &obj_hash[i].lock irq_context: 0 (wq_completion)bond62#2 (work_completion)(&(&bond->mcast_work)->work) &base->lock irq_context: 0 (wq_completion)bond62#2 (work_completion)(&(&bond->mcast_work)->work) &base->lock &obj_hash[i].lock irq_context: 0 (wq_completion)bond62#2 (work_completion)(&(&bond->mcast_work)->work) &rq->__lock irq_context: 0 (wq_completion)bond62#2 (work_completion)(&(&bond->mcast_work)->work) &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)bond62#2 (work_completion)(&(&slave->notify_work)->work) rtnl_mutex net_rwsem irq_context: 0 (wq_completion)bond62#2 (work_completion)(&(&slave->notify_work)->work) rtnl_mutex net_rwsem &list->lock#2 irq_context: 0 (wq_completion)bond62#2 (work_completion)(&(&slave->notify_work)->work) rtnl_mutex &tn->lock irq_context: 0 (wq_completion)bond62#2 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex fs_reclaim irq_context: 0 (wq_completion)bond62#2 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex fs_reclaim mmu_notifier_invalidate_range_start irq_context: 0 (wq_completion)bond62#2 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex rcu_read_lock &bond->stats_lock/1 irq_context: 0 (wq_completion)bond62#2 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex rcu_read_lock &ndev->lock irq_context: 0 (wq_completion)bond62#2 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex nl_table_lock irq_context: 0 (wq_completion)bond62#2 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &obj_hash[i].lock irq_context: 0 (wq_completion)bond62#2 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex nl_table_wait.lock irq_context: 0 (wq_completion)bond62#2 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex net_rwsem irq_context: 0 (wq_completion)bond62#2 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex net_rwsem &list->lock#2 irq_context: 0 (wq_completion)bond62#2 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &tn->lock irq_context: 0 (wq_completion)bond62#2 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &idev->mc_lock irq_context: 0 (wq_completion)bond62#2 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &idev->mc_lock fs_reclaim irq_context: 0 (wq_completion)bond62#2 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &idev->mc_lock fs_reclaim mmu_notifier_invalidate_range_start irq_context: 0 (wq_completion)bond62#2 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &idev->mc_lock &c->lock irq_context: 0 (wq_completion)bond62#2 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &idev->mc_lock &n->list_lock irq_context: 0 (wq_completion)bond62#2 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &idev->mc_lock &n->list_lock &c->lock irq_context: 0 (wq_completion)bond62#2 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &idev->mc_lock rcu_read_lock &ndev->lock irq_context: 0 (wq_completion)bond62#2 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &idev->mc_lock rcu_read_lock &dir->lock#2 irq_context: 0 (wq_completion)bond62#2 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &idev->mc_lock rcu_read_lock &ul->lock#2 irq_context: 0 (wq_completion)bond62#2 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &idev->mc_lock rcu_read_lock rcu_read_lock &obj_hash[i].lock irq_context: 0 (wq_completion)bond62#2 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &idev->mc_lock rcu_read_lock rcu_read_lock &dir->lock#2 irq_context: 0 (wq_completion)bond62#2 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &idev->mc_lock rcu_read_lock rcu_read_lock &tbl->lock irq_context: 0 (wq_completion)bond62#2 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &idev->mc_lock rcu_read_lock rcu_read_lock &n->lock irq_context: 0 (wq_completion)bond62#2 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &idev->mc_lock rcu_read_lock rcu_read_lock &____s->seqcount#9 irq_context: 0 (wq_completion)bond62#2 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &idev->mc_lock rcu_read_lock rcu_read_lock rcu_read_lock_bh rcu_read_lock rcu_read_lock_bh rcu_read_lock &br->multicast_lock irq_context: 0 (wq_completion)bond62#2 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &idev->mc_lock rcu_read_lock rcu_read_lock rcu_read_lock_bh rcu_read_lock rcu_read_lock_bh rcu_read_lock &br->multicast_lock &obj_hash[i].lock irq_context: 0 (wq_completion)bond62#2 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &idev->mc_lock rcu_read_lock rcu_read_lock rcu_read_lock_bh rcu_read_lock rcu_read_lock_bh rcu_read_lock &br->multicast_lock rcu_read_lock rhashtable_bucket irq_context: 0 (wq_completion)bond62#2 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &idev->mc_lock rcu_read_lock rcu_read_lock rcu_read_lock_bh rcu_read_lock rcu_read_lock_bh rcu_read_lock &br->multicast_lock nl_table_lock irq_context: 0 (wq_completion)bond62#2 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &idev->mc_lock rcu_read_lock rcu_read_lock rcu_read_lock_bh rcu_read_lock rcu_read_lock_bh rcu_read_lock &br->multicast_lock nl_table_wait.lock irq_context: 0 (wq_completion)bond62#2 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &idev->mc_lock rcu_read_lock rcu_read_lock rcu_read_lock_bh rcu_read_lock rcu_read_lock_bh rcu_read_lock &br->multicast_lock &base->lock irq_context: 0 (wq_completion)bond62#2 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &idev->mc_lock rcu_read_lock rcu_read_lock rcu_read_lock_bh rcu_read_lock rcu_read_lock_bh rcu_read_lock &br->multicast_lock &base->lock &obj_hash[i].lock irq_context: 0 (wq_completion)bond62#2 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &idev->mc_lock rcu_read_lock rcu_read_lock rcu_read_lock_bh rcu_read_lock rcu_read_lock_bh rcu_read_lock &data->lock irq_context: 0 (wq_completion)bond62#2 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &idev->mc_lock rcu_read_lock rcu_read_lock rcu_read_lock_bh rcu_read_lock rcu_read_lock_bh rcu_read_lock &obj_hash[i].lock irq_context: 0 rtnl_mutex pcpu_alloc_mutex rcu_read_lock &rq->__lock irq_context: 0 rtnl_mutex pcpu_alloc_mutex rcu_read_lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 sk_lock-AF_INET rcu_read_lock rcu_read_lock rcu_read_lock_bh _xmit_NONE#2 irq_context: 0 sk_lock-AF_INET rcu_read_lock rcu_read_lock rcu_read_lock_bh _xmit_NONE#2 &data->lock irq_context: softirq rcu_read_lock rcu_read_lock &dreq->dreq_lock slock-AF_INET6 fill_pool_map-wait-type-override &____s->seqcount#2 irq_context: softirq rcu_read_lock rcu_read_lock &dreq->dreq_lock slock-AF_INET6 fill_pool_map-wait-type-override &____s->seqcount irq_context: softirq (&icsk->icsk_retransmit_timer) slock-AF_INET rcu_read_lock rcu_read_lock rcu_read_lock_bh _xmit_NONE#2 irq_context: softirq (&icsk->icsk_retransmit_timer) slock-AF_INET rcu_read_lock rcu_read_lock rcu_read_lock_bh _xmit_NONE#2 &data->lock irq_context: 0 rcu_read_lock console_owner_lock irq_context: 0 rcu_read_lock console_owner irq_context: 0 (wq_completion)bond96 irq_context: 0 (wq_completion)bond96 (work_completion)(&(&slave->notify_work)->work) irq_context: 0 (wq_completion)bond96 (work_completion)(&(&slave->notify_work)->work) &obj_hash[i].lock irq_context: 0 (wq_completion)bond96 (work_completion)(&(&slave->notify_work)->work) &base->lock irq_context: 0 (wq_completion)bond96 (work_completion)(&(&slave->notify_work)->work) &base->lock &obj_hash[i].lock irq_context: 0 (wq_completion)bond96 (work_completion)(&(&bond->mcast_work)->work) irq_context: 0 (wq_completion)bond96 (work_completion)(&(&bond->mcast_work)->work) &obj_hash[i].lock irq_context: 0 (wq_completion)bond96 (work_completion)(&(&bond->mcast_work)->work) &base->lock irq_context: 0 (wq_completion)bond96 (work_completion)(&(&bond->mcast_work)->work) &base->lock &obj_hash[i].lock irq_context: 0 (wq_completion)bond67 irq_context: 0 (wq_completion)bond67 (work_completion)(&(&slave->notify_work)->work) irq_context: 0 (wq_completion)bond67 (work_completion)(&(&slave->notify_work)->work) &obj_hash[i].lock irq_context: 0 (wq_completion)bond67 (work_completion)(&(&slave->notify_work)->work) &base->lock irq_context: 0 (wq_completion)bond67 (work_completion)(&(&slave->notify_work)->work) &base->lock &obj_hash[i].lock irq_context: 0 (wq_completion)bond96 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex fs_reclaim irq_context: 0 (wq_completion)bond96 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex fs_reclaim mmu_notifier_invalidate_range_start irq_context: 0 (wq_completion)bond96 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex batched_entropy_u8.lock irq_context: 0 (wq_completion)bond96 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex kfence_freelist_lock irq_context: 0 (wq_completion)bond96 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &c->lock irq_context: 0 (wq_completion)bond67 (work_completion)(&(&bond->mcast_work)->work) irq_context: 0 (wq_completion)bond96 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex rcu_read_lock &bond->stats_lock/1 irq_context: 0 (wq_completion)bond67 (work_completion)(&(&bond->mcast_work)->work) &obj_hash[i].lock irq_context: 0 (wq_completion)bond96 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex rcu_read_lock &ndev->lock irq_context: 0 (wq_completion)bond67 (work_completion)(&(&bond->mcast_work)->work) &base->lock irq_context: 0 (wq_completion)bond96 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex nl_table_lock irq_context: 0 (wq_completion)bond67 (work_completion)(&(&bond->mcast_work)->work) &base->lock &obj_hash[i].lock irq_context: 0 (wq_completion)bond96 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &obj_hash[i].lock irq_context: 0 (wq_completion)bond96 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &meta->lock irq_context: 0 (wq_completion)bond96 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex nl_table_wait.lock irq_context: 0 (wq_completion)bond96 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex net_rwsem irq_context: 0 (wq_completion)bond96 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex net_rwsem &list->lock#2 irq_context: 0 (wq_completion)bond96 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &tn->lock irq_context: 0 (wq_completion)bond96 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &idev->mc_lock irq_context: 0 (wq_completion)bond96 (work_completion)(&(&slave->notify_work)->work) rtnl_mutex net_rwsem irq_context: 0 (wq_completion)bond96 (work_completion)(&(&slave->notify_work)->work) rtnl_mutex net_rwsem &list->lock#2 irq_context: 0 (wq_completion)bond96 (work_completion)(&(&slave->notify_work)->work) rtnl_mutex &tn->lock irq_context: 0 (wq_completion)bond67 (work_completion)(&(&slave->notify_work)->work) rtnl_mutex net_rwsem irq_context: 0 (wq_completion)bond67 (work_completion)(&(&slave->notify_work)->work) rtnl_mutex net_rwsem &list->lock#2 irq_context: 0 (wq_completion)bond67 (work_completion)(&(&slave->notify_work)->work) rtnl_mutex &tn->lock irq_context: 0 (wq_completion)bond67 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex fs_reclaim irq_context: 0 (wq_completion)bond67 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex fs_reclaim mmu_notifier_invalidate_range_start irq_context: 0 (wq_completion)bond67 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex rcu_read_lock &bond->stats_lock/1 irq_context: 0 (wq_completion)bond67 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex rcu_read_lock &ndev->lock irq_context: 0 (wq_completion)bond67 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex nl_table_lock irq_context: 0 (wq_completion)bond67 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &obj_hash[i].lock irq_context: 0 (wq_completion)bond67 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex nl_table_wait.lock irq_context: 0 (wq_completion)bond67 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex net_rwsem irq_context: 0 (wq_completion)bond67 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex net_rwsem &list->lock#2 irq_context: 0 (wq_completion)bond67 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &tn->lock irq_context: 0 (wq_completion)bond67 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &idev->mc_lock irq_context: 0 (wq_completion)bond67 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &idev->mc_lock fs_reclaim irq_context: 0 (wq_completion)bond67 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &idev->mc_lock fs_reclaim mmu_notifier_invalidate_range_start irq_context: 0 (wq_completion)bond67 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &idev->mc_lock &c->lock irq_context: 0 (wq_completion)bond67 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &idev->mc_lock rcu_read_lock &ndev->lock irq_context: 0 (wq_completion)bond67 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &idev->mc_lock rcu_read_lock &c->lock irq_context: 0 (wq_completion)bond67 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &idev->mc_lock rcu_read_lock &dir->lock#2 irq_context: 0 (wq_completion)bond67 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &idev->mc_lock rcu_read_lock &ul->lock#2 irq_context: 0 (wq_completion)bond67 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &idev->mc_lock rcu_read_lock rcu_read_lock &obj_hash[i].lock irq_context: 0 (wq_completion)bond67 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &idev->mc_lock rcu_read_lock rcu_read_lock &c->lock irq_context: 0 (wq_completion)bond67 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &idev->mc_lock rcu_read_lock rcu_read_lock &dir->lock#2 irq_context: 0 (wq_completion)bond67 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &idev->mc_lock rcu_read_lock rcu_read_lock &tbl->lock irq_context: 0 (wq_completion)bond67 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &idev->mc_lock rcu_read_lock rcu_read_lock &n->lock irq_context: 0 (wq_completion)bond67 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &idev->mc_lock rcu_read_lock rcu_read_lock &____s->seqcount#9 irq_context: 0 (wq_completion)bond67 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &idev->mc_lock rcu_read_lock rcu_read_lock rcu_read_lock_bh rcu_read_lock rcu_read_lock_bh rcu_read_lock &br->multicast_lock irq_context: 0 (wq_completion)bond67 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &idev->mc_lock rcu_read_lock rcu_read_lock rcu_read_lock_bh rcu_read_lock rcu_read_lock_bh rcu_read_lock &br->multicast_lock &obj_hash[i].lock irq_context: 0 (wq_completion)bond67 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &idev->mc_lock rcu_read_lock rcu_read_lock rcu_read_lock_bh rcu_read_lock rcu_read_lock_bh rcu_read_lock &br->multicast_lock rcu_read_lock rhashtable_bucket irq_context: 0 (wq_completion)bond67 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &idev->mc_lock rcu_read_lock rcu_read_lock rcu_read_lock_bh rcu_read_lock rcu_read_lock_bh rcu_read_lock &br->multicast_lock &c->lock irq_context: 0 (wq_completion)bond67 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &idev->mc_lock rcu_read_lock rcu_read_lock rcu_read_lock_bh rcu_read_lock rcu_read_lock_bh rcu_read_lock &br->multicast_lock nl_table_lock irq_context: 0 (wq_completion)bond67 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &idev->mc_lock rcu_read_lock rcu_read_lock rcu_read_lock_bh rcu_read_lock rcu_read_lock_bh rcu_read_lock &br->multicast_lock nl_table_wait.lock irq_context: 0 (wq_completion)bond67 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &idev->mc_lock rcu_read_lock rcu_read_lock rcu_read_lock_bh rcu_read_lock rcu_read_lock_bh rcu_read_lock &br->multicast_lock &base->lock irq_context: 0 (wq_completion)bond67 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &idev->mc_lock rcu_read_lock rcu_read_lock rcu_read_lock_bh rcu_read_lock rcu_read_lock_bh rcu_read_lock &br->multicast_lock &base->lock &obj_hash[i].lock irq_context: 0 (wq_completion)bond67 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &idev->mc_lock rcu_read_lock rcu_read_lock rcu_read_lock_bh rcu_read_lock rcu_read_lock_bh rcu_read_lock &data->lock irq_context: 0 (wq_completion)bond67 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &idev->mc_lock rcu_read_lock rcu_read_lock rcu_read_lock_bh rcu_read_lock rcu_read_lock_bh rcu_read_lock &obj_hash[i].lock irq_context: 0 (wq_completion)bond67 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &idev->mc_lock rcu_read_lock rcu_read_lock rcu_read_lock_bh rcu_read_lock rcu_read_lock_bh rcu_read_lock quarantine_lock irq_context: 0 (wq_completion)bond13 &rq->__lock irq_context: 0 (wq_completion)bond97 irq_context: 0 (wq_completion)bond97 (work_completion)(&(&slave->notify_work)->work) irq_context: 0 (wq_completion)bond97 (work_completion)(&(&slave->notify_work)->work) &obj_hash[i].lock irq_context: 0 (wq_completion)bond97 (work_completion)(&(&slave->notify_work)->work) &base->lock irq_context: 0 (wq_completion)bond97 (work_completion)(&(&slave->notify_work)->work) &base->lock &obj_hash[i].lock irq_context: 0 (wq_completion)bond97 (work_completion)(&(&slave->notify_work)->work) &rq->__lock irq_context: 0 (wq_completion)bond97 (work_completion)(&(&slave->notify_work)->work) &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 rtnl_mutex rcu_read_lock &pool->lock/1 fill_pool_map-wait-type-override &____s->seqcount#2 irq_context: 0 (wq_completion)bond97 (work_completion)(&(&bond->mcast_work)->work) irq_context: 0 (wq_completion)bond97 (work_completion)(&(&bond->mcast_work)->work) &obj_hash[i].lock irq_context: 0 (wq_completion)bond97 (work_completion)(&(&bond->mcast_work)->work) &base->lock irq_context: 0 (wq_completion)bond97 (work_completion)(&(&bond->mcast_work)->work) &base->lock &obj_hash[i].lock irq_context: 0 (wq_completion)bond97 (work_completion)(&(&bond->mcast_work)->work) &rq->__lock irq_context: 0 (wq_completion)bond97 (work_completion)(&(&bond->mcast_work)->work) &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: softirq rcu_read_lock_bh rcu_read_lock &pool->lock/1 &p->pi_lock &rq->__lock &cfs_rq->removed.lock irq_context: 0 (wq_completion)wg-kex-wg2#2 (work_completion)(&({ do { const void *__vpp_verify = (typeof((worker) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __asm__ ("" : "=r"(__ptr) : "0"((typeof(*((worker))) *)((worker)))); (typeof((typeof(*((worker))) *)((worker)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); })->work) &handshake->lock &n->list_lock irq_context: 0 (wq_completion)wg-kex-wg2#2 (work_completion)(&({ do { const void *__vpp_verify = (typeof((worker) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __asm__ ("" : "=r"(__ptr) : "0"((typeof(*((worker))) *)((worker)))); (typeof((typeof(*((worker))) *)((worker)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); })->work) &handshake->lock &n->list_lock &c->lock irq_context: 0 rtnl_mutex free_vmap_area_lock &obj_hash[i].lock irq_context: 0 rtnl_mutex free_vmap_area_lock pool_lock#2 irq_context: 0 (wq_completion)bond85 irq_context: 0 (wq_completion)bond85 (work_completion)(&(&slave->notify_work)->work) irq_context: 0 (wq_completion)bond85 (work_completion)(&(&slave->notify_work)->work) &obj_hash[i].lock irq_context: 0 (wq_completion)bond85 (work_completion)(&(&slave->notify_work)->work) &base->lock irq_context: 0 (wq_completion)bond85 (work_completion)(&(&slave->notify_work)->work) &base->lock &obj_hash[i].lock irq_context: 0 (wq_completion)bond85 (work_completion)(&(&slave->notify_work)->work) &rq->__lock irq_context: 0 (wq_completion)bond85 (work_completion)(&(&slave->notify_work)->work) &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)bond85 (work_completion)(&(&bond->mcast_work)->work) irq_context: 0 (wq_completion)bond85 (work_completion)(&(&bond->mcast_work)->work) &obj_hash[i].lock irq_context: 0 (wq_completion)bond85 (work_completion)(&(&bond->mcast_work)->work) &base->lock irq_context: 0 (wq_completion)bond85 (work_completion)(&(&bond->mcast_work)->work) &base->lock &obj_hash[i].lock irq_context: 0 (wq_completion)bond85 (work_completion)(&(&bond->mcast_work)->work) &rq->__lock irq_context: 0 (wq_completion)bond68 irq_context: 0 (wq_completion)bond68 (work_completion)(&(&slave->notify_work)->work) irq_context: 0 (wq_completion)bond68 (work_completion)(&(&slave->notify_work)->work) &obj_hash[i].lock irq_context: 0 (wq_completion)bond68 (work_completion)(&(&slave->notify_work)->work) &base->lock irq_context: 0 (wq_completion)bond68 (work_completion)(&(&slave->notify_work)->work) &base->lock &obj_hash[i].lock irq_context: 0 (wq_completion)bond68 (work_completion)(&(&slave->notify_work)->work) &rq->__lock irq_context: 0 (wq_completion)bond68 (work_completion)(&(&slave->notify_work)->work) &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)bond13 (work_completion)(&(&bond->mii_work)->work) &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)bond68 (work_completion)(&(&bond->mcast_work)->work) irq_context: 0 (wq_completion)bond97 (work_completion)(&(&slave->notify_work)->work) rtnl_mutex net_rwsem irq_context: 0 (wq_completion)bond68 (work_completion)(&(&bond->mcast_work)->work) &obj_hash[i].lock irq_context: 0 (wq_completion)bond97 (work_completion)(&(&slave->notify_work)->work) rtnl_mutex net_rwsem &list->lock#2 irq_context: 0 (wq_completion)bond68 (work_completion)(&(&bond->mcast_work)->work) &base->lock irq_context: 0 (wq_completion)bond97 (work_completion)(&(&slave->notify_work)->work) rtnl_mutex &tn->lock irq_context: 0 (wq_completion)bond68 (work_completion)(&(&bond->mcast_work)->work) &base->lock &obj_hash[i].lock irq_context: 0 (wq_completion)bond97 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex fs_reclaim irq_context: 0 (wq_completion)bond68 (work_completion)(&(&bond->mcast_work)->work) &rq->__lock irq_context: 0 (wq_completion)bond97 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex fs_reclaim mmu_notifier_invalidate_range_start irq_context: 0 (wq_completion)bond68 (work_completion)(&(&bond->mcast_work)->work) &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)bond97 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &c->lock irq_context: 0 (wq_completion)bond97 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex rcu_read_lock &bond->stats_lock/1 irq_context: 0 (wq_completion)bond97 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex rcu_read_lock &ndev->lock irq_context: 0 (wq_completion)bond97 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex nl_table_lock irq_context: 0 (wq_completion)bond97 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &obj_hash[i].lock irq_context: 0 (wq_completion)bond97 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex nl_table_wait.lock irq_context: 0 (wq_completion)bond97 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex net_rwsem irq_context: 0 (wq_completion)bond97 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex net_rwsem &list->lock#2 irq_context: 0 (wq_completion)bond97 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &tn->lock irq_context: 0 (wq_completion)bond97 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &idev->mc_lock irq_context: 0 (wq_completion)bond97 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &idev->mc_lock fs_reclaim irq_context: 0 (wq_completion)bond97 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &idev->mc_lock fs_reclaim mmu_notifier_invalidate_range_start irq_context: 0 (wq_completion)bond97 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &idev->mc_lock &c->lock irq_context: 0 (wq_completion)bond97 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &idev->mc_lock &n->list_lock irq_context: 0 (wq_completion)bond97 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &idev->mc_lock &n->list_lock &c->lock irq_context: 0 (wq_completion)bond97 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &idev->mc_lock rcu_read_lock &ndev->lock irq_context: 0 (wq_completion)bond97 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &idev->mc_lock rcu_read_lock &dir->lock#2 irq_context: 0 (wq_completion)bond97 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &idev->mc_lock rcu_read_lock &ul->lock#2 irq_context: 0 (wq_completion)bond97 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &idev->mc_lock rcu_read_lock rcu_read_lock &obj_hash[i].lock irq_context: 0 (wq_completion)bond97 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &idev->mc_lock rcu_read_lock rcu_read_lock &dir->lock#2 irq_context: 0 (wq_completion)bond97 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &idev->mc_lock rcu_read_lock rcu_read_lock &tbl->lock irq_context: 0 (wq_completion)bond97 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &idev->mc_lock rcu_read_lock rcu_read_lock &n->lock irq_context: 0 (wq_completion)bond97 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &idev->mc_lock rcu_read_lock rcu_read_lock &____s->seqcount#9 irq_context: 0 (wq_completion)bond97 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &idev->mc_lock rcu_read_lock rcu_read_lock rcu_read_lock_bh rcu_read_lock rcu_read_lock_bh rcu_read_lock &br->multicast_lock irq_context: 0 (wq_completion)bond97 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &idev->mc_lock rcu_read_lock rcu_read_lock rcu_read_lock_bh rcu_read_lock rcu_read_lock_bh rcu_read_lock &br->multicast_lock &obj_hash[i].lock irq_context: 0 (wq_completion)bond97 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &idev->mc_lock rcu_read_lock rcu_read_lock rcu_read_lock_bh rcu_read_lock rcu_read_lock_bh rcu_read_lock &br->multicast_lock rcu_read_lock rhashtable_bucket irq_context: 0 (wq_completion)bond97 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &idev->mc_lock rcu_read_lock rcu_read_lock rcu_read_lock_bh rcu_read_lock rcu_read_lock_bh rcu_read_lock &br->multicast_lock &c->lock irq_context: 0 (wq_completion)bond97 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &idev->mc_lock rcu_read_lock rcu_read_lock rcu_read_lock_bh rcu_read_lock rcu_read_lock_bh rcu_read_lock &br->multicast_lock nl_table_lock irq_context: 0 (wq_completion)bond97 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &idev->mc_lock rcu_read_lock rcu_read_lock rcu_read_lock_bh rcu_read_lock rcu_read_lock_bh rcu_read_lock &br->multicast_lock nl_table_wait.lock irq_context: 0 (wq_completion)bond97 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &idev->mc_lock rcu_read_lock rcu_read_lock rcu_read_lock_bh rcu_read_lock rcu_read_lock_bh rcu_read_lock &br->multicast_lock &base->lock irq_context: 0 (wq_completion)bond97 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &idev->mc_lock rcu_read_lock rcu_read_lock rcu_read_lock_bh rcu_read_lock rcu_read_lock_bh rcu_read_lock &br->multicast_lock &base->lock &obj_hash[i].lock irq_context: 0 (wq_completion)bond97 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &idev->mc_lock rcu_read_lock rcu_read_lock rcu_read_lock_bh rcu_read_lock rcu_read_lock_bh rcu_read_lock &data->lock irq_context: 0 (wq_completion)bond97 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &idev->mc_lock rcu_read_lock rcu_read_lock rcu_read_lock_bh rcu_read_lock rcu_read_lock_bh rcu_read_lock &obj_hash[i].lock irq_context: 0 (wq_completion)bond85 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex fs_reclaim irq_context: 0 (wq_completion)bond85 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex fs_reclaim mmu_notifier_invalidate_range_start irq_context: 0 (wq_completion)bond85 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex fs_reclaim mmu_notifier_invalidate_range_start &rq->__lock irq_context: 0 (wq_completion)bond85 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &c->lock irq_context: 0 (wq_completion)bond85 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &n->list_lock irq_context: 0 (wq_completion)bond85 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &n->list_lock &c->lock irq_context: 0 (wq_completion)bond85 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex rcu_read_lock &bond->stats_lock/1 irq_context: 0 (wq_completion)bond85 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex rcu_read_lock &ndev->lock irq_context: 0 (wq_completion)bond85 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex nl_table_lock irq_context: 0 (wq_completion)bond85 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &obj_hash[i].lock irq_context: 0 (wq_completion)bond85 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex nl_table_wait.lock irq_context: 0 (wq_completion)bond85 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex net_rwsem irq_context: 0 (wq_completion)bond85 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex net_rwsem &list->lock#2 irq_context: 0 (wq_completion)bond85 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &tn->lock irq_context: 0 (wq_completion)bond85 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &tn->lock &rq->__lock irq_context: 0 (wq_completion)bond85 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &tn->lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)bond85 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &idev->mc_lock irq_context: 0 (wq_completion)bond85 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &idev->mc_lock fs_reclaim irq_context: 0 (wq_completion)bond85 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &idev->mc_lock fs_reclaim mmu_notifier_invalidate_range_start irq_context: 0 (wq_completion)bond85 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &idev->mc_lock &c->lock irq_context: 0 (wq_completion)bond85 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &idev->mc_lock rcu_read_lock &ndev->lock irq_context: 0 (wq_completion)bond85 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &idev->mc_lock rcu_read_lock &dir->lock#2 irq_context: 0 (wq_completion)bond85 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &idev->mc_lock rcu_read_lock &ul->lock#2 irq_context: 0 (wq_completion)bond85 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &idev->mc_lock rcu_read_lock rcu_read_lock &obj_hash[i].lock irq_context: 0 (wq_completion)bond85 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &idev->mc_lock rcu_read_lock rcu_read_lock &c->lock irq_context: 0 (wq_completion)bond85 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &idev->mc_lock rcu_read_lock rcu_read_lock &n->list_lock irq_context: 0 (wq_completion)bond85 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &idev->mc_lock rcu_read_lock rcu_read_lock &n->list_lock &c->lock irq_context: 0 (wq_completion)bond85 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &idev->mc_lock rcu_read_lock rcu_read_lock &dir->lock#2 irq_context: 0 (wq_completion)bond85 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &idev->mc_lock rcu_read_lock rcu_read_lock &tbl->lock irq_context: 0 (wq_completion)bond85 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &idev->mc_lock rcu_read_lock rcu_read_lock rcu_node_0 irq_context: 0 (wq_completion)bond85 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &idev->mc_lock rcu_read_lock rcu_read_lock &rq->__lock irq_context: 0 (wq_completion)bond85 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &idev->mc_lock rcu_read_lock rcu_read_lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)bond85 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &idev->mc_lock rcu_read_lock rcu_read_lock &n->lock irq_context: 0 (wq_completion)bond85 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &idev->mc_lock rcu_read_lock rcu_read_lock &____s->seqcount#9 irq_context: 0 (wq_completion)bond85 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &idev->mc_lock rcu_read_lock rcu_read_lock rcu_read_lock_bh rcu_read_lock rcu_read_lock_bh rcu_read_lock &br->multicast_lock irq_context: 0 (wq_completion)bond85 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &idev->mc_lock rcu_read_lock rcu_read_lock rcu_read_lock_bh rcu_read_lock rcu_read_lock_bh rcu_read_lock &br->multicast_lock &obj_hash[i].lock irq_context: 0 (wq_completion)bond85 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &idev->mc_lock rcu_read_lock rcu_read_lock rcu_read_lock_bh rcu_read_lock rcu_read_lock_bh rcu_read_lock &br->multicast_lock rcu_read_lock rhashtable_bucket irq_context: 0 (wq_completion)bond85 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &idev->mc_lock rcu_read_lock rcu_read_lock rcu_read_lock_bh rcu_read_lock rcu_read_lock_bh rcu_read_lock &br->multicast_lock &c->lock irq_context: 0 (wq_completion)bond85 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &idev->mc_lock rcu_read_lock rcu_read_lock rcu_read_lock_bh rcu_read_lock rcu_read_lock_bh rcu_read_lock &br->multicast_lock nl_table_lock irq_context: 0 (wq_completion)bond85 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &idev->mc_lock rcu_read_lock rcu_read_lock rcu_read_lock_bh rcu_read_lock rcu_read_lock_bh rcu_read_lock &br->multicast_lock nl_table_wait.lock irq_context: 0 (wq_completion)bond85 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &idev->mc_lock rcu_read_lock rcu_read_lock rcu_read_lock_bh rcu_read_lock rcu_read_lock_bh rcu_read_lock &br->multicast_lock &base->lock irq_context: 0 (wq_completion)bond85 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &idev->mc_lock rcu_read_lock rcu_read_lock rcu_read_lock_bh rcu_read_lock rcu_read_lock_bh rcu_read_lock &br->multicast_lock &base->lock &obj_hash[i].lock irq_context: 0 (wq_completion)bond85 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &idev->mc_lock rcu_read_lock rcu_read_lock rcu_read_lock_bh rcu_read_lock rcu_read_lock_bh rcu_read_lock &data->lock irq_context: 0 (wq_completion)bond85 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &idev->mc_lock rcu_read_lock rcu_read_lock rcu_read_lock_bh rcu_read_lock rcu_read_lock_bh rcu_read_lock &obj_hash[i].lock irq_context: 0 (wq_completion)bond85 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &idev->mc_lock rcu_read_lock rcu_node_0 irq_context: 0 (wq_completion)bond85 (work_completion)(&(&slave->notify_work)->work) rtnl_mutex net_rwsem irq_context: 0 (wq_completion)bond85 (work_completion)(&(&slave->notify_work)->work) rtnl_mutex net_rwsem &list->lock#2 irq_context: 0 (wq_completion)bond85 (work_completion)(&(&slave->notify_work)->work) rtnl_mutex &tn->lock irq_context: 0 (wq_completion)bond68 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex fs_reclaim irq_context: 0 (wq_completion)bond68 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex fs_reclaim mmu_notifier_invalidate_range_start irq_context: 0 (wq_completion)bond68 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex rcu_read_lock &bond->stats_lock/1 irq_context: 0 (wq_completion)bond68 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex rcu_read_lock &ndev->lock irq_context: 0 (wq_completion)bond68 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex nl_table_lock irq_context: 0 (wq_completion)bond68 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &obj_hash[i].lock irq_context: 0 (wq_completion)bond68 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex nl_table_wait.lock irq_context: 0 (wq_completion)bond68 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex net_rwsem irq_context: 0 (wq_completion)bond68 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex net_rwsem &list->lock#2 irq_context: 0 (wq_completion)bond68 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &tn->lock irq_context: 0 (wq_completion)bond68 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &idev->mc_lock irq_context: 0 (wq_completion)bond68 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &idev->mc_lock fs_reclaim irq_context: 0 (wq_completion)bond68 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &idev->mc_lock fs_reclaim &rq->__lock irq_context: 0 (wq_completion)bond68 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &idev->mc_lock fs_reclaim &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)bond68 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &idev->mc_lock fs_reclaim mmu_notifier_invalidate_range_start irq_context: 0 (wq_completion)bond68 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &idev->mc_lock &c->lock irq_context: 0 (wq_completion)bond68 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &idev->mc_lock rcu_read_lock &ndev->lock irq_context: 0 (wq_completion)bond68 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &idev->mc_lock rcu_read_lock &c->lock irq_context: 0 (wq_completion)bond68 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &idev->mc_lock rcu_read_lock &dir->lock#2 irq_context: 0 (wq_completion)bond68 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &idev->mc_lock rcu_read_lock &ul->lock#2 irq_context: 0 (wq_completion)bond68 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &idev->mc_lock rcu_read_lock rcu_read_lock &obj_hash[i].lock irq_context: 0 (wq_completion)bond68 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &idev->mc_lock rcu_read_lock rcu_read_lock &dir->lock#2 irq_context: 0 (wq_completion)bond68 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &idev->mc_lock rcu_read_lock rcu_read_lock &tbl->lock irq_context: 0 (wq_completion)bond68 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &idev->mc_lock rcu_read_lock rcu_read_lock &n->lock irq_context: 0 (wq_completion)bond68 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &idev->mc_lock rcu_read_lock rcu_read_lock &____s->seqcount#9 irq_context: 0 (wq_completion)bond68 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &idev->mc_lock rcu_read_lock rcu_read_lock rcu_read_lock_bh rcu_read_lock rcu_read_lock_bh rcu_read_lock &br->multicast_lock irq_context: 0 (wq_completion)bond68 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &idev->mc_lock rcu_read_lock rcu_read_lock rcu_read_lock_bh rcu_read_lock rcu_read_lock_bh rcu_read_lock &br->multicast_lock &obj_hash[i].lock irq_context: 0 (wq_completion)bond68 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &idev->mc_lock rcu_read_lock rcu_read_lock rcu_read_lock_bh rcu_read_lock rcu_read_lock_bh rcu_read_lock &br->multicast_lock rcu_read_lock rhashtable_bucket irq_context: 0 (wq_completion)bond68 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &idev->mc_lock rcu_read_lock rcu_read_lock rcu_read_lock_bh rcu_read_lock rcu_read_lock_bh rcu_read_lock &br->multicast_lock &c->lock irq_context: 0 (wq_completion)bond68 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &idev->mc_lock rcu_read_lock rcu_read_lock rcu_read_lock_bh rcu_read_lock rcu_read_lock_bh rcu_read_lock &br->multicast_lock nl_table_lock irq_context: 0 (wq_completion)bond68 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &idev->mc_lock rcu_read_lock rcu_read_lock rcu_read_lock_bh rcu_read_lock rcu_read_lock_bh rcu_read_lock &br->multicast_lock nl_table_wait.lock irq_context: 0 (wq_completion)bond68 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &idev->mc_lock rcu_read_lock rcu_read_lock rcu_read_lock_bh rcu_read_lock rcu_read_lock_bh rcu_read_lock &br->multicast_lock &base->lock irq_context: 0 (wq_completion)bond68 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &idev->mc_lock rcu_read_lock rcu_read_lock rcu_read_lock_bh rcu_read_lock rcu_read_lock_bh rcu_read_lock &br->multicast_lock &base->lock &obj_hash[i].lock irq_context: 0 (wq_completion)bond68 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &idev->mc_lock rcu_read_lock rcu_read_lock rcu_read_lock_bh rcu_read_lock rcu_read_lock_bh rcu_read_lock &data->lock irq_context: 0 (wq_completion)bond68 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &idev->mc_lock rcu_read_lock rcu_read_lock rcu_read_lock_bh rcu_read_lock rcu_read_lock_bh rcu_read_lock &obj_hash[i].lock irq_context: 0 (wq_completion)bond68 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &idev->mc_lock rcu_read_lock rcu_read_lock rcu_node_0 irq_context: 0 (wq_completion)bond68 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &idev->mc_lock rcu_read_lock rcu_read_lock &rq->__lock irq_context: 0 (wq_completion)bond68 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &idev->mc_lock rcu_read_lock rcu_read_lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)mld (work_completion)(&(&idev->mc_ifc_work)->work) &idev->mc_lock &lock->wait_lock irq_context: 0 (wq_completion)mld (work_completion)(&(&idev->mc_ifc_work)->work) &idev->mc_lock &pool->lock irq_context: 0 (wq_completion)mld (work_completion)(&(&idev->mc_ifc_work)->work) &idev->mc_lock &pool->lock &p->pi_lock irq_context: 0 (wq_completion)mld (work_completion)(&(&idev->mc_ifc_work)->work) &idev->mc_lock &pool->lock &p->pi_lock &rq->__lock irq_context: 0 (wq_completion)mld (work_completion)(&(&idev->mc_ifc_work)->work) &idev->mc_lock &pool->lock &p->pi_lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)bond68 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &idev->mc_lock rcu_read_lock rcu_node_0 irq_context: 0 (wq_completion)bond68 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &lock->wait_lock irq_context: 0 (wq_completion)bond68 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &p->pi_lock irq_context: 0 (wq_completion)bond68 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &p->pi_lock &rq->__lock irq_context: 0 (wq_completion)bond68 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &p->pi_lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)bond68 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &rq->__lock irq_context: 0 (wq_completion)bond68 (work_completion)(&(&slave->notify_work)->work) rtnl_mutex net_rwsem irq_context: 0 (wq_completion)bond68 (work_completion)(&(&slave->notify_work)->work) rtnl_mutex net_rwsem &list->lock#2 irq_context: 0 (wq_completion)bond68 (work_completion)(&(&slave->notify_work)->work) rtnl_mutex &tn->lock irq_context: 0 sb_writers#4 &type->i_mutex_dir_key#3/1 &sb->s_type->i_mutex_key#8 jbd2_handle rcu_read_lock &obj_hash[i].lock pool_lock irq_context: softirq (&hc->tx_rtotimer) &c->lock irq_context: 0 (wq_completion)bond98 irq_context: 0 (wq_completion)bond98 (work_completion)(&(&slave->notify_work)->work) irq_context: 0 (wq_completion)bond98 (work_completion)(&(&slave->notify_work)->work) &obj_hash[i].lock irq_context: 0 (wq_completion)bond98 (work_completion)(&(&slave->notify_work)->work) &base->lock irq_context: 0 (wq_completion)bond98 (work_completion)(&(&slave->notify_work)->work) &base->lock &obj_hash[i].lock irq_context: 0 (wq_completion)bond98 (work_completion)(&(&slave->notify_work)->work) &rq->__lock irq_context: 0 (wq_completion)bond98 (work_completion)(&(&slave->notify_work)->work) &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)bond98 (work_completion)(&(&bond->mcast_work)->work) irq_context: 0 (wq_completion)bond98 (work_completion)(&(&bond->mcast_work)->work) &obj_hash[i].lock irq_context: 0 (wq_completion)bond98 (work_completion)(&(&bond->mcast_work)->work) &base->lock irq_context: 0 (wq_completion)bond98 (work_completion)(&(&bond->mcast_work)->work) &base->lock &obj_hash[i].lock irq_context: 0 (wq_completion)bond98 (work_completion)(&(&slave->notify_work)->work) rtnl_mutex net_rwsem irq_context: 0 (wq_completion)bond98 (work_completion)(&(&slave->notify_work)->work) rtnl_mutex net_rwsem &list->lock#2 irq_context: 0 (wq_completion)bond98 (work_completion)(&(&slave->notify_work)->work) rtnl_mutex &tn->lock irq_context: 0 (wq_completion)bond98 (work_completion)(&(&slave->notify_work)->work) rtnl_mutex.wait_lock irq_context: 0 (wq_completion)bond98 (work_completion)(&(&slave->notify_work)->work) &p->pi_lock irq_context: 0 (wq_completion)bond98 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex fs_reclaim irq_context: 0 (wq_completion)bond98 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex fs_reclaim mmu_notifier_invalidate_range_start irq_context: 0 (wq_completion)bond98 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &c->lock irq_context: 0 (wq_completion)bond98 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex rcu_read_lock &bond->stats_lock/1 irq_context: 0 (wq_completion)bond98 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex rcu_read_lock &ndev->lock irq_context: 0 (wq_completion)bond98 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex nl_table_lock irq_context: 0 (wq_completion)bond98 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &obj_hash[i].lock irq_context: 0 (wq_completion)bond98 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex quarantine_lock irq_context: 0 (wq_completion)bond98 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex nl_table_wait.lock irq_context: 0 (wq_completion)bond98 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex net_rwsem irq_context: 0 (wq_completion)bond98 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex net_rwsem &list->lock#2 irq_context: 0 (wq_completion)bond98 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &tn->lock irq_context: 0 (wq_completion)bond98 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &idev->mc_lock irq_context: 0 (wq_completion)bond98 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex.wait_lock irq_context: 0 (wq_completion)bond98 (work_completion)(&(&bond->mcast_work)->work) &p->pi_lock irq_context: 0 (wq_completion)bond69 irq_context: 0 (wq_completion)bond69 (work_completion)(&(&slave->notify_work)->work) irq_context: 0 (wq_completion)bond69 (work_completion)(&(&slave->notify_work)->work) &obj_hash[i].lock irq_context: 0 (wq_completion)bond69 (work_completion)(&(&slave->notify_work)->work) &base->lock irq_context: 0 (wq_completion)bond69 (work_completion)(&(&slave->notify_work)->work) &base->lock &obj_hash[i].lock irq_context: 0 (wq_completion)bond69 (work_completion)(&(&slave->notify_work)->work) &rq->__lock irq_context: 0 (wq_completion)bond69 (work_completion)(&(&slave->notify_work)->work) &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)bond69 (work_completion)(&(&bond->mcast_work)->work) irq_context: 0 (wq_completion)bond69 (work_completion)(&(&bond->mcast_work)->work) &obj_hash[i].lock irq_context: 0 (wq_completion)bond69 (work_completion)(&(&bond->mcast_work)->work) &base->lock irq_context: 0 (wq_completion)bond69 (work_completion)(&(&bond->mcast_work)->work) &base->lock &obj_hash[i].lock irq_context: 0 (wq_completion)bond99 irq_context: 0 (wq_completion)bond99 (work_completion)(&(&slave->notify_work)->work) irq_context: 0 (wq_completion)bond99 (work_completion)(&(&slave->notify_work)->work) &obj_hash[i].lock irq_context: 0 (wq_completion)bond99 (work_completion)(&(&slave->notify_work)->work) &base->lock irq_context: 0 (wq_completion)bond99 (work_completion)(&(&slave->notify_work)->work) &base->lock &obj_hash[i].lock irq_context: 0 (wq_completion)bond99 (work_completion)(&(&bond->mcast_work)->work) irq_context: 0 (wq_completion)bond99 (work_completion)(&(&bond->mcast_work)->work) &obj_hash[i].lock irq_context: 0 (wq_completion)bond99 (work_completion)(&(&bond->mcast_work)->work) &base->lock irq_context: 0 (wq_completion)bond99 (work_completion)(&(&bond->mcast_work)->work) &base->lock &obj_hash[i].lock irq_context: 0 (wq_completion)bond99 (work_completion)(&(&bond->mcast_work)->work) &rq->__lock irq_context: 0 (wq_completion)bond99 (work_completion)(&(&bond->mcast_work)->work) &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)bond69 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex fs_reclaim irq_context: 0 (wq_completion)bond69 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex fs_reclaim mmu_notifier_invalidate_range_start irq_context: 0 (wq_completion)bond69 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &c->lock irq_context: 0 (wq_completion)bond69 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &n->list_lock irq_context: 0 (wq_completion)bond69 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &n->list_lock &c->lock irq_context: 0 (wq_completion)bond69 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex rcu_read_lock &bond->stats_lock/1 irq_context: 0 (wq_completion)bond69 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex rcu_read_lock &ndev->lock irq_context: 0 (wq_completion)bond69 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex nl_table_lock irq_context: 0 (wq_completion)bond69 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &obj_hash[i].lock irq_context: 0 (wq_completion)bond69 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex nl_table_wait.lock irq_context: 0 (wq_completion)bond69 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex net_rwsem irq_context: 0 (wq_completion)bond69 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex net_rwsem &list->lock#2 irq_context: 0 (wq_completion)bond69 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &tn->lock irq_context: 0 (wq_completion)bond69 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &idev->mc_lock irq_context: 0 (wq_completion)bond69 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &idev->mc_lock fs_reclaim irq_context: 0 (wq_completion)bond69 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &idev->mc_lock fs_reclaim mmu_notifier_invalidate_range_start irq_context: 0 (wq_completion)bond69 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &idev->mc_lock &c->lock irq_context: 0 (wq_completion)bond69 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &idev->mc_lock rcu_read_lock &ndev->lock irq_context: 0 (wq_completion)bond69 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &idev->mc_lock rcu_read_lock &dir->lock#2 irq_context: 0 (wq_completion)bond69 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &idev->mc_lock rcu_read_lock &ul->lock#2 irq_context: 0 (wq_completion)bond69 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &idev->mc_lock rcu_read_lock rcu_read_lock &____s->seqcount#10 irq_context: 0 (wq_completion)bond69 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &idev->mc_lock rcu_read_lock rcu_read_lock rcu_read_lock_bh rcu_read_lock rcu_read_lock_bh rcu_read_lock &br->multicast_lock irq_context: 0 (wq_completion)bond69 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &idev->mc_lock rcu_read_lock rcu_read_lock rcu_read_lock_bh rcu_read_lock rcu_read_lock_bh rcu_read_lock &br->multicast_lock &obj_hash[i].lock irq_context: 0 (wq_completion)bond69 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &idev->mc_lock rcu_read_lock rcu_read_lock rcu_read_lock_bh rcu_read_lock rcu_read_lock_bh rcu_read_lock &br->multicast_lock &base->lock irq_context: 0 (wq_completion)bond69 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &idev->mc_lock rcu_read_lock rcu_read_lock rcu_read_lock_bh rcu_read_lock rcu_read_lock_bh rcu_read_lock &data->lock irq_context: 0 (wq_completion)bond69 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &idev->mc_lock rcu_read_lock rcu_read_lock rcu_read_lock_bh rcu_read_lock rcu_read_lock_bh rcu_read_lock &obj_hash[i].lock irq_context: 0 (wq_completion)bond69 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex.wait_lock irq_context: 0 (wq_completion)bond69 (work_completion)(&(&bond->mcast_work)->work) &p->pi_lock irq_context: 0 (wq_completion)bond99 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex fs_reclaim irq_context: 0 (wq_completion)bond99 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex fs_reclaim mmu_notifier_invalidate_range_start irq_context: 0 (wq_completion)bond99 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex fs_reclaim mmu_notifier_invalidate_range_start &rq->__lock irq_context: 0 (wq_completion)bond99 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex rcu_read_lock &bond->stats_lock/1 irq_context: 0 (wq_completion)bond99 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex rcu_read_lock &ndev->lock irq_context: 0 (wq_completion)bond99 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex nl_table_lock irq_context: 0 (wq_completion)bond99 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &obj_hash[i].lock irq_context: 0 (wq_completion)bond99 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex nl_table_wait.lock irq_context: 0 (wq_completion)bond99 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex net_rwsem irq_context: 0 (wq_completion)bond99 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex net_rwsem &list->lock#2 irq_context: 0 (wq_completion)bond99 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &tn->lock irq_context: 0 (wq_completion)bond99 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &idev->mc_lock irq_context: 0 (wq_completion)bond99 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex.wait_lock irq_context: 0 (wq_completion)bond99 (work_completion)(&(&bond->mcast_work)->work) &p->pi_lock irq_context: 0 (wq_completion)bond69 (work_completion)(&(&slave->notify_work)->work) rtnl_mutex net_rwsem irq_context: 0 (wq_completion)bond69 (work_completion)(&(&slave->notify_work)->work) rtnl_mutex net_rwsem &list->lock#2 irq_context: 0 (wq_completion)bond69 (work_completion)(&(&slave->notify_work)->work) rtnl_mutex &tn->lock irq_context: 0 (wq_completion)bond69 (work_completion)(&(&slave->notify_work)->work) rtnl_mutex.wait_lock irq_context: 0 (wq_completion)bond69 (work_completion)(&(&slave->notify_work)->work) &p->pi_lock irq_context: 0 (wq_completion)bond99 (work_completion)(&(&slave->notify_work)->work) rtnl_mutex net_rwsem irq_context: 0 (wq_completion)bond99 (work_completion)(&(&slave->notify_work)->work) rtnl_mutex net_rwsem &list->lock#2 irq_context: 0 (wq_completion)bond99 (work_completion)(&(&slave->notify_work)->work) rtnl_mutex net_rwsem &rq->__lock irq_context: 0 (wq_completion)bond99 (work_completion)(&(&slave->notify_work)->work) rtnl_mutex net_rwsem &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)bond99 (work_completion)(&(&slave->notify_work)->work) rtnl_mutex &tn->lock irq_context: 0 (wq_completion)bond99 (work_completion)(&(&slave->notify_work)->work) rtnl_mutex.wait_lock irq_context: 0 (wq_completion)bond99 (work_completion)(&(&slave->notify_work)->work) &p->pi_lock irq_context: softirq (&ndev->rs_timer) &pcp->lock &zone->lock &____s->seqcount irq_context: 0 (wq_completion)bond70 irq_context: 0 (wq_completion)bond70 (work_completion)(&(&slave->notify_work)->work) irq_context: 0 (wq_completion)bond70 (work_completion)(&(&slave->notify_work)->work) &obj_hash[i].lock irq_context: 0 (wq_completion)bond70 (work_completion)(&(&slave->notify_work)->work) &base->lock irq_context: 0 (wq_completion)bond70 (work_completion)(&(&slave->notify_work)->work) &base->lock &obj_hash[i].lock irq_context: 0 (wq_completion)bond70 (work_completion)(&(&slave->notify_work)->work) &rq->__lock irq_context: 0 (wq_completion)bond70 (work_completion)(&(&slave->notify_work)->work) &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)bond70 (work_completion)(&(&bond->mcast_work)->work) irq_context: 0 (wq_completion)bond70 (work_completion)(&(&bond->mcast_work)->work) &obj_hash[i].lock irq_context: 0 (wq_completion)bond70 (work_completion)(&(&bond->mcast_work)->work) &base->lock irq_context: 0 (wq_completion)bond70 (work_completion)(&(&bond->mcast_work)->work) &base->lock &obj_hash[i].lock irq_context: 0 (wq_completion)bond70 (work_completion)(&(&bond->mcast_work)->work) &rq->__lock irq_context: 0 (wq_completion)bond70 (work_completion)(&(&bond->mcast_work)->work) &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)bond100 irq_context: 0 (wq_completion)bond100 (work_completion)(&(&slave->notify_work)->work) irq_context: 0 (wq_completion)bond100 (work_completion)(&(&slave->notify_work)->work) &obj_hash[i].lock irq_context: 0 (wq_completion)bond100 (work_completion)(&(&slave->notify_work)->work) &base->lock irq_context: 0 (wq_completion)bond100 (work_completion)(&(&slave->notify_work)->work) &base->lock &obj_hash[i].lock irq_context: 0 (wq_completion)bond100 (work_completion)(&(&slave->notify_work)->work) &rq->__lock irq_context: 0 (wq_completion)bond100 (work_completion)(&(&slave->notify_work)->work) &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)bond100 (work_completion)(&(&bond->mcast_work)->work) irq_context: 0 (wq_completion)bond100 (work_completion)(&(&bond->mcast_work)->work) &obj_hash[i].lock irq_context: 0 (wq_completion)bond100 (work_completion)(&(&bond->mcast_work)->work) &base->lock irq_context: 0 (wq_completion)bond100 (work_completion)(&(&bond->mcast_work)->work) &base->lock &obj_hash[i].lock irq_context: 0 (wq_completion)bond88 irq_context: 0 (wq_completion)bond88 (work_completion)(&(&slave->notify_work)->work) irq_context: 0 (wq_completion)bond88 (work_completion)(&(&slave->notify_work)->work) &obj_hash[i].lock irq_context: 0 (wq_completion)bond88 (work_completion)(&(&slave->notify_work)->work) &base->lock irq_context: 0 (wq_completion)bond88 (work_completion)(&(&slave->notify_work)->work) &base->lock &obj_hash[i].lock irq_context: 0 (wq_completion)bond88 (work_completion)(&(&slave->notify_work)->work) &rq->__lock irq_context: 0 (wq_completion)bond88 (work_completion)(&(&bond->mcast_work)->work) irq_context: 0 (wq_completion)bond88 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex fs_reclaim irq_context: 0 (wq_completion)bond88 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex fs_reclaim mmu_notifier_invalidate_range_start irq_context: 0 (wq_completion)bond88 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex rcu_read_lock &bond->stats_lock/1 irq_context: 0 (wq_completion)bond88 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex rcu_read_lock &ndev->lock irq_context: 0 (wq_completion)bond88 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex nl_table_lock irq_context: 0 (wq_completion)bond88 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &obj_hash[i].lock irq_context: 0 (wq_completion)bond88 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &rq->__lock irq_context: 0 (wq_completion)bond88 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)bond88 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex nl_table_wait.lock irq_context: 0 (wq_completion)bond88 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex net_rwsem irq_context: 0 (wq_completion)bond88 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex net_rwsem &list->lock#2 irq_context: 0 (wq_completion)bond88 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &tn->lock irq_context: 0 (wq_completion)bond88 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &idev->mc_lock irq_context: 0 (wq_completion)bond88 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex.wait_lock irq_context: 0 (wq_completion)bond88 (work_completion)(&(&bond->mcast_work)->work) &p->pi_lock irq_context: 0 (wq_completion)bond71 irq_context: 0 (wq_completion)bond71 (work_completion)(&(&slave->notify_work)->work) irq_context: 0 (wq_completion)bond71 (work_completion)(&(&slave->notify_work)->work) &obj_hash[i].lock irq_context: 0 (wq_completion)bond71 (work_completion)(&(&slave->notify_work)->work) &base->lock irq_context: 0 (wq_completion)bond71 (work_completion)(&(&slave->notify_work)->work) &base->lock &obj_hash[i].lock irq_context: 0 (wq_completion)bond71 (work_completion)(&(&slave->notify_work)->work) &rq->__lock irq_context: 0 (wq_completion)bond71 (work_completion)(&(&slave->notify_work)->work) &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)bond71 (work_completion)(&(&bond->mcast_work)->work) irq_context: 0 (wq_completion)bond71 (work_completion)(&(&bond->mcast_work)->work) &obj_hash[i].lock irq_context: 0 (wq_completion)bond71 (work_completion)(&(&bond->mcast_work)->work) &base->lock irq_context: 0 (wq_completion)bond71 (work_completion)(&(&bond->mcast_work)->work) &base->lock &obj_hash[i].lock irq_context: 0 (wq_completion)bond71 (work_completion)(&(&bond->mcast_work)->work) &rq->__lock irq_context: 0 (wq_completion)bond71 (work_completion)(&(&bond->mcast_work)->work) &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)bond101 irq_context: 0 (wq_completion)bond101 (work_completion)(&(&slave->notify_work)->work) irq_context: 0 (wq_completion)bond101 (work_completion)(&(&slave->notify_work)->work) &obj_hash[i].lock irq_context: 0 (wq_completion)bond101 (work_completion)(&(&slave->notify_work)->work) &base->lock irq_context: 0 (wq_completion)bond101 (work_completion)(&(&slave->notify_work)->work) &base->lock &obj_hash[i].lock irq_context: 0 (wq_completion)bond101 (work_completion)(&(&slave->notify_work)->work) &rq->__lock irq_context: 0 (wq_completion)bond101 (work_completion)(&(&slave->notify_work)->work) &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)ipv6_addrconf (work_completion)(&(&ifa->dad_work)->work) rtnl_mutex &idev->mc_lock rcu_read_lock rcu_read_lock rcu_read_lock_bh rcu_read_lock rcu_read_lock_bh rcu_read_lock &obj_hash[i].lock pool_lock irq_context: 0 (wq_completion)bond101 (work_completion)(&(&bond->mcast_work)->work) irq_context: 0 (wq_completion)bond101 (work_completion)(&(&bond->mcast_work)->work) &obj_hash[i].lock irq_context: 0 (wq_completion)bond101 (work_completion)(&(&bond->mcast_work)->work) &base->lock irq_context: 0 (wq_completion)bond101 (work_completion)(&(&bond->mcast_work)->work) &base->lock &obj_hash[i].lock irq_context: 0 (wq_completion)bond89 irq_context: 0 (wq_completion)bond89 (work_completion)(&(&slave->notify_work)->work) irq_context: 0 (wq_completion)bond89 (work_completion)(&(&slave->notify_work)->work) &obj_hash[i].lock irq_context: 0 (wq_completion)bond89 (work_completion)(&(&slave->notify_work)->work) &base->lock irq_context: 0 (wq_completion)bond89 (work_completion)(&(&slave->notify_work)->work) &base->lock &obj_hash[i].lock irq_context: 0 (wq_completion)bond89 (work_completion)(&(&slave->notify_work)->work) &rq->__lock irq_context: 0 (wq_completion)bond89 (work_completion)(&(&slave->notify_work)->work) &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: softirq (&in_dev->mr_ifc_timer) rcu_read_lock rcu_read_lock_bh rcu_read_lock fill_pool_map-wait-type-override &c->lock irq_context: softirq (&in_dev->mr_ifc_timer) rcu_read_lock rcu_read_lock_bh rcu_read_lock fill_pool_map-wait-type-override &____s->seqcount#2 irq_context: softirq (&in_dev->mr_ifc_timer) rcu_read_lock rcu_read_lock_bh rcu_read_lock fill_pool_map-wait-type-override &____s->seqcount irq_context: 0 (wq_completion)bond89 (work_completion)(&(&bond->mcast_work)->work) irq_context: 0 (wq_completion)bond89 (work_completion)(&(&bond->mcast_work)->work) &obj_hash[i].lock irq_context: 0 (wq_completion)bond89 (work_completion)(&(&bond->mcast_work)->work) &base->lock irq_context: 0 (wq_completion)bond89 (work_completion)(&(&bond->mcast_work)->work) &base->lock &obj_hash[i].lock irq_context: 0 (wq_completion)bond89 (work_completion)(&(&bond->mcast_work)->work) &rq->__lock irq_context: 0 (wq_completion)bond89 (work_completion)(&(&bond->mcast_work)->work) &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)bond72 irq_context: 0 (wq_completion)bond72 &rq->__lock irq_context: 0 (wq_completion)bond72 (work_completion)(&(&slave->notify_work)->work) irq_context: 0 (wq_completion)bond72 (work_completion)(&(&slave->notify_work)->work) &obj_hash[i].lock irq_context: 0 (wq_completion)bond72 (work_completion)(&(&slave->notify_work)->work) &base->lock irq_context: 0 (wq_completion)bond72 (work_completion)(&(&slave->notify_work)->work) &base->lock &obj_hash[i].lock irq_context: 0 (wq_completion)bond72 (work_completion)(&(&slave->notify_work)->work) &rq->__lock irq_context: 0 (wq_completion)bond72 (work_completion)(&(&bond->mcast_work)->work) irq_context: 0 (wq_completion)bond72 (work_completion)(&(&bond->mcast_work)->work) &obj_hash[i].lock irq_context: 0 (wq_completion)bond72 (work_completion)(&(&bond->mcast_work)->work) &base->lock irq_context: 0 (wq_completion)bond72 (work_completion)(&(&bond->mcast_work)->work) &base->lock &obj_hash[i].lock irq_context: 0 (wq_completion)ipv6_addrconf (work_completion)(&(&ifa->dad_work)->work) rtnl_mutex rcu_read_lock &rcu_state.gp_wq &p->pi_lock &cfs_rq->removed.lock irq_context: 0 (wq_completion)bond72 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex fs_reclaim irq_context: 0 (wq_completion)bond72 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex fs_reclaim mmu_notifier_invalidate_range_start irq_context: 0 (wq_completion)bond72 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex rcu_read_lock &bond->stats_lock/1 irq_context: 0 (wq_completion)bond72 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex rcu_read_lock &ndev->lock irq_context: 0 (wq_completion)bond72 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex nl_table_lock irq_context: 0 (wq_completion)bond72 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &obj_hash[i].lock irq_context: 0 (wq_completion)bond72 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex nl_table_wait.lock irq_context: 0 (wq_completion)bond72 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex net_rwsem irq_context: 0 (wq_completion)bond72 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex net_rwsem &list->lock#2 irq_context: 0 (wq_completion)bond72 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &tn->lock irq_context: 0 (wq_completion)bond72 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &idev->mc_lock irq_context: 0 (wq_completion)bond72 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &idev->mc_lock fs_reclaim irq_context: 0 (wq_completion)bond72 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &idev->mc_lock fs_reclaim mmu_notifier_invalidate_range_start irq_context: 0 (wq_completion)bond72 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &idev->mc_lock rcu_read_lock &ndev->lock irq_context: 0 (wq_completion)bond72 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &idev->mc_lock rcu_read_lock &dir->lock#2 irq_context: 0 (wq_completion)bond72 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &idev->mc_lock rcu_read_lock &ul->lock#2 irq_context: 0 (wq_completion)bond72 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &idev->mc_lock rcu_read_lock rcu_read_lock &c->lock irq_context: 0 (wq_completion)bond72 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &idev->mc_lock rcu_read_lock rcu_read_lock &____s->seqcount#2 irq_context: 0 (wq_completion)bond72 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &idev->mc_lock rcu_read_lock rcu_read_lock &____s->seqcount irq_context: 0 (wq_completion)bond72 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &idev->mc_lock rcu_read_lock rcu_read_lock pool_lock#2 irq_context: 0 (wq_completion)bond72 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &idev->mc_lock rcu_read_lock rcu_read_lock &obj_hash[i].lock irq_context: 0 (wq_completion)bond72 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &idev->mc_lock rcu_read_lock rcu_read_lock &dir->lock#2 irq_context: 0 (wq_completion)bond72 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &idev->mc_lock rcu_read_lock rcu_read_lock &tbl->lock irq_context: 0 (wq_completion)bond72 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &idev->mc_lock rcu_read_lock rcu_read_lock &n->lock irq_context: 0 (wq_completion)bond72 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &idev->mc_lock rcu_read_lock rcu_read_lock &____s->seqcount#9 irq_context: 0 (wq_completion)bond72 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &idev->mc_lock rcu_read_lock rcu_read_lock rcu_read_lock_bh rcu_read_lock rcu_read_lock_bh rcu_read_lock &br->multicast_lock irq_context: 0 (wq_completion)bond72 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &idev->mc_lock rcu_read_lock rcu_read_lock rcu_read_lock_bh rcu_read_lock rcu_read_lock_bh rcu_read_lock &br->multicast_lock &obj_hash[i].lock irq_context: 0 (wq_completion)bond72 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &idev->mc_lock rcu_read_lock rcu_read_lock rcu_read_lock_bh rcu_read_lock rcu_read_lock_bh rcu_read_lock &br->multicast_lock rcu_read_lock rhashtable_bucket irq_context: 0 (wq_completion)bond72 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &idev->mc_lock rcu_read_lock rcu_read_lock rcu_read_lock_bh rcu_read_lock rcu_read_lock_bh rcu_read_lock &br->multicast_lock nl_table_lock irq_context: 0 (wq_completion)bond72 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &idev->mc_lock rcu_read_lock rcu_read_lock rcu_read_lock_bh rcu_read_lock rcu_read_lock_bh rcu_read_lock &br->multicast_lock nl_table_wait.lock irq_context: 0 (wq_completion)bond72 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &idev->mc_lock rcu_read_lock rcu_read_lock rcu_read_lock_bh rcu_read_lock rcu_read_lock_bh rcu_read_lock &br->multicast_lock &base->lock irq_context: 0 (wq_completion)bond72 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &idev->mc_lock rcu_read_lock rcu_read_lock rcu_read_lock_bh rcu_read_lock rcu_read_lock_bh rcu_read_lock &br->multicast_lock &base->lock &obj_hash[i].lock irq_context: 0 (wq_completion)bond72 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &idev->mc_lock rcu_read_lock rcu_read_lock rcu_read_lock_bh rcu_read_lock rcu_read_lock_bh rcu_read_lock &data->lock irq_context: 0 (wq_completion)bond72 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &idev->mc_lock rcu_read_lock rcu_read_lock rcu_read_lock_bh rcu_read_lock rcu_read_lock_bh rcu_read_lock &obj_hash[i].lock irq_context: 0 (wq_completion)bond72 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex.wait_lock irq_context: 0 (wq_completion)bond72 (work_completion)(&(&bond->mcast_work)->work) &p->pi_lock irq_context: 0 (wq_completion)bond72 (work_completion)(&(&slave->notify_work)->work) rtnl_mutex net_rwsem irq_context: 0 (wq_completion)bond72 (work_completion)(&(&slave->notify_work)->work) rtnl_mutex net_rwsem &list->lock#2 irq_context: 0 (wq_completion)bond72 (work_completion)(&(&slave->notify_work)->work) rtnl_mutex &tn->lock irq_context: 0 (wq_completion)bond72 (work_completion)(&(&slave->notify_work)->work) rtnl_mutex.wait_lock irq_context: 0 (wq_completion)bond72 (work_completion)(&(&slave->notify_work)->work) &p->pi_lock irq_context: 0 (wq_completion)ipv6_addrconf (work_completion)(&(&ifa->dad_work)->work) rtnl_mutex &idev->mc_lock &dev_addr_list_lock_key/1 &____s->seqcount#2 irq_context: 0 (wq_completion)mld (work_completion)(&(&idev->mc_ifc_work)->work) &idev->mc_lock rcu_read_lock rcu_read_lock rcu_read_lock_bh rcu_read_lock rcu_read_lock_bh rcu_read_lock &br->multicast_lock fill_pool_map-wait-type-override pool_lock#2 irq_context: 0 (wq_completion)mld (work_completion)(&(&idev->mc_ifc_work)->work) &idev->mc_lock rcu_read_lock rcu_read_lock rcu_read_lock_bh rcu_read_lock rcu_read_lock_bh rcu_read_lock &br->multicast_lock fill_pool_map-wait-type-override &c->lock irq_context: 0 (wq_completion)mld (work_completion)(&(&idev->mc_ifc_work)->work) &idev->mc_lock rcu_read_lock rcu_read_lock rcu_read_lock_bh rcu_read_lock rcu_read_lock_bh rcu_read_lock &br->multicast_lock fill_pool_map-wait-type-override &____s->seqcount#2 irq_context: 0 (wq_completion)mld (work_completion)(&(&idev->mc_ifc_work)->work) &idev->mc_lock rcu_read_lock rcu_read_lock rcu_read_lock_bh rcu_read_lock rcu_read_lock_bh rcu_read_lock &br->multicast_lock fill_pool_map-wait-type-override &____s->seqcount irq_context: 0 (wq_completion)mld (work_completion)(&(&idev->mc_ifc_work)->work) &idev->mc_lock rcu_read_lock rcu_read_lock rcu_read_lock_bh rcu_read_lock rcu_read_lock_bh rcu_read_lock &br->multicast_lock fill_pool_map-wait-type-override pool_lock irq_context: 0 (wq_completion)bond102 irq_context: 0 (wq_completion)bond102 &rq->__lock irq_context: 0 (wq_completion)bond102 &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)bond102 (work_completion)(&(&slave->notify_work)->work) irq_context: 0 (wq_completion)bond102 (work_completion)(&(&slave->notify_work)->work) &obj_hash[i].lock irq_context: 0 (wq_completion)bond102 (work_completion)(&(&slave->notify_work)->work) &base->lock irq_context: 0 (wq_completion)bond102 (work_completion)(&(&slave->notify_work)->work) &base->lock &obj_hash[i].lock irq_context: 0 (wq_completion)bond102 (work_completion)(&(&bond->mcast_work)->work) irq_context: 0 (wq_completion)bond102 (work_completion)(&(&bond->mcast_work)->work) &obj_hash[i].lock irq_context: 0 (wq_completion)bond102 (work_completion)(&(&bond->mcast_work)->work) &base->lock irq_context: 0 (wq_completion)bond102 (work_completion)(&(&bond->mcast_work)->work) &base->lock &obj_hash[i].lock irq_context: 0 (wq_completion)bond89 (work_completion)(&(&slave->notify_work)->work) rtnl_mutex net_rwsem irq_context: 0 (wq_completion)bond89 (work_completion)(&(&slave->notify_work)->work) rtnl_mutex net_rwsem &list->lock#2 irq_context: 0 (wq_completion)bond89 (work_completion)(&(&slave->notify_work)->work) rtnl_mutex &tn->lock irq_context: 0 (wq_completion)bond89 (work_completion)(&(&slave->notify_work)->work) rtnl_mutex.wait_lock irq_context: 0 (wq_completion)bond89 (work_completion)(&(&slave->notify_work)->work) &p->pi_lock irq_context: 0 (wq_completion)bond73 irq_context: 0 (wq_completion)bond73 (work_completion)(&(&slave->notify_work)->work) irq_context: 0 (wq_completion)bond73 (work_completion)(&(&slave->notify_work)->work) &obj_hash[i].lock irq_context: 0 (wq_completion)bond73 (work_completion)(&(&slave->notify_work)->work) &base->lock irq_context: 0 (wq_completion)bond73 (work_completion)(&(&slave->notify_work)->work) &base->lock &obj_hash[i].lock irq_context: 0 (wq_completion)bond73 (work_completion)(&(&slave->notify_work)->work) &rq->__lock irq_context: 0 (wq_completion)bond73 (work_completion)(&(&slave->notify_work)->work) &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)bond73 (work_completion)(&(&bond->mcast_work)->work) irq_context: 0 (wq_completion)bond73 (work_completion)(&(&bond->mcast_work)->work) &obj_hash[i].lock irq_context: 0 (wq_completion)bond73 (work_completion)(&(&bond->mcast_work)->work) &base->lock irq_context: 0 (wq_completion)bond73 (work_completion)(&(&bond->mcast_work)->work) &base->lock &obj_hash[i].lock irq_context: 0 (wq_completion)bond101 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex fs_reclaim irq_context: 0 (wq_completion)bond101 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex fs_reclaim mmu_notifier_invalidate_range_start irq_context: 0 (wq_completion)bond101 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex fs_reclaim mmu_notifier_invalidate_range_start &rq->__lock irq_context: 0 (wq_completion)bond101 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex fs_reclaim mmu_notifier_invalidate_range_start &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)bond101 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &c->lock irq_context: 0 (wq_completion)bond101 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex rcu_read_lock &bond->stats_lock/1 irq_context: 0 (wq_completion)bond101 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex rcu_read_lock &ndev->lock irq_context: 0 (wq_completion)bond101 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex nl_table_lock irq_context: 0 (wq_completion)bond101 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &obj_hash[i].lock irq_context: 0 (wq_completion)bond101 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex nl_table_wait.lock irq_context: 0 (wq_completion)bond101 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex net_rwsem irq_context: 0 (wq_completion)bond101 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex net_rwsem &list->lock#2 irq_context: 0 (wq_completion)bond101 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &tn->lock irq_context: 0 (wq_completion)bond101 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &idev->mc_lock irq_context: 0 (wq_completion)bond101 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &idev->mc_lock &rq->__lock irq_context: 0 (wq_completion)bond101 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &idev->mc_lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)bond101 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &idev->mc_lock fs_reclaim irq_context: 0 (wq_completion)bond101 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &idev->mc_lock fs_reclaim mmu_notifier_invalidate_range_start irq_context: 0 (wq_completion)bond101 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &idev->mc_lock &c->lock irq_context: 0 (wq_completion)bond101 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &idev->mc_lock rcu_read_lock &ndev->lock irq_context: 0 (wq_completion)bond101 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &idev->mc_lock rcu_read_lock &dir->lock#2 irq_context: 0 (wq_completion)bond101 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &idev->mc_lock rcu_read_lock &c->lock irq_context: 0 (wq_completion)bond101 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &idev->mc_lock rcu_read_lock &ul->lock#2 irq_context: 0 (wq_completion)bond101 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &idev->mc_lock rcu_read_lock rcu_read_lock &____s->seqcount#10 irq_context: 0 (wq_completion)bond101 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &idev->mc_lock rcu_read_lock rcu_read_lock rcu_read_lock_bh rcu_read_lock rcu_read_lock_bh rcu_read_lock &br->multicast_lock irq_context: 0 (wq_completion)bond101 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &idev->mc_lock rcu_read_lock rcu_read_lock rcu_read_lock_bh rcu_read_lock rcu_read_lock_bh rcu_read_lock &br->multicast_lock &obj_hash[i].lock irq_context: 0 (wq_completion)bond101 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &idev->mc_lock rcu_read_lock rcu_read_lock rcu_read_lock_bh rcu_read_lock rcu_read_lock_bh rcu_read_lock &br->multicast_lock &base->lock irq_context: 0 (wq_completion)bond101 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &idev->mc_lock rcu_read_lock rcu_read_lock rcu_read_lock_bh rcu_read_lock rcu_read_lock_bh rcu_read_lock &data->lock irq_context: 0 (wq_completion)bond101 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &idev->mc_lock rcu_read_lock rcu_read_lock rcu_read_lock_bh rcu_read_lock rcu_read_lock_bh rcu_read_lock fill_pool_map-wait-type-override pool_lock#2 irq_context: 0 (wq_completion)bond101 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &idev->mc_lock rcu_read_lock rcu_read_lock rcu_read_lock_bh rcu_read_lock rcu_read_lock_bh rcu_read_lock fill_pool_map-wait-type-override pool_lock irq_context: 0 (wq_completion)bond101 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &idev->mc_lock rcu_read_lock rcu_read_lock rcu_read_lock_bh rcu_read_lock rcu_read_lock_bh rcu_read_lock &obj_hash[i].lock irq_context: 0 (wq_completion)bond101 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex.wait_lock irq_context: 0 (wq_completion)bond101 (work_completion)(&(&bond->mcast_work)->work) &p->pi_lock irq_context: 0 (wq_completion)bond101 (work_completion)(&(&bond->mcast_work)->work) &p->pi_lock &rq->__lock irq_context: 0 (wq_completion)bond101 (work_completion)(&(&bond->mcast_work)->work) &p->pi_lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)bond101 (work_completion)(&(&bond->mcast_work)->work) &rq->__lock irq_context: 0 (wq_completion)bond101 (work_completion)(&(&bond->mcast_work)->work) &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)bond88 (work_completion)(&(&slave->notify_work)->work) rtnl_mutex net_rwsem irq_context: 0 (wq_completion)bond88 (work_completion)(&(&slave->notify_work)->work) rtnl_mutex net_rwsem &list->lock#2 irq_context: 0 (wq_completion)bond88 (work_completion)(&(&slave->notify_work)->work) rtnl_mutex &tn->lock irq_context: 0 (wq_completion)bond88 (work_completion)(&(&slave->notify_work)->work) rtnl_mutex.wait_lock irq_context: 0 (wq_completion)bond88 (work_completion)(&(&slave->notify_work)->work) &p->pi_lock irq_context: 0 (wq_completion)bond89 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex fs_reclaim irq_context: 0 (wq_completion)bond89 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex fs_reclaim mmu_notifier_invalidate_range_start irq_context: 0 (wq_completion)bond89 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &c->lock irq_context: 0 (wq_completion)bond89 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex rcu_read_lock &bond->stats_lock/1 irq_context: 0 (wq_completion)bond89 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex rcu_read_lock &ndev->lock irq_context: 0 (wq_completion)bond89 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex nl_table_lock irq_context: 0 (wq_completion)bond89 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &obj_hash[i].lock irq_context: 0 (wq_completion)bond89 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex nl_table_wait.lock irq_context: 0 (wq_completion)bond89 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex net_rwsem irq_context: 0 (wq_completion)bond89 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex net_rwsem &list->lock#2 irq_context: 0 (wq_completion)bond89 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &tn->lock irq_context: 0 (wq_completion)bond89 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &idev->mc_lock irq_context: 0 (wq_completion)bond89 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &idev->mc_lock fs_reclaim irq_context: 0 (wq_completion)bond89 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &idev->mc_lock fs_reclaim mmu_notifier_invalidate_range_start irq_context: 0 (wq_completion)bond89 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &idev->mc_lock rcu_read_lock &ndev->lock irq_context: 0 (wq_completion)bond89 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &idev->mc_lock rcu_read_lock &dir->lock#2 irq_context: 0 (wq_completion)bond89 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &idev->mc_lock rcu_read_lock &ul->lock#2 irq_context: 0 (wq_completion)bond89 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &idev->mc_lock rcu_read_lock rcu_read_lock &____s->seqcount#10 irq_context: 0 (wq_completion)bond89 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &idev->mc_lock rcu_read_lock rcu_read_lock rcu_read_lock_bh rcu_read_lock rcu_read_lock_bh rcu_read_lock &br->multicast_lock irq_context: 0 (wq_completion)bond89 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &idev->mc_lock rcu_read_lock rcu_read_lock rcu_read_lock_bh rcu_read_lock rcu_read_lock_bh rcu_read_lock &br->multicast_lock &obj_hash[i].lock irq_context: 0 (wq_completion)bond89 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &idev->mc_lock rcu_read_lock rcu_read_lock rcu_read_lock_bh rcu_read_lock rcu_read_lock_bh rcu_read_lock &br->multicast_lock &base->lock irq_context: 0 (wq_completion)bond89 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &idev->mc_lock rcu_read_lock rcu_read_lock rcu_read_lock_bh rcu_read_lock rcu_read_lock_bh rcu_read_lock &br->multicast_lock &base->lock &obj_hash[i].lock irq_context: 0 (wq_completion)bond89 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &idev->mc_lock rcu_read_lock rcu_read_lock rcu_read_lock_bh rcu_read_lock rcu_read_lock_bh rcu_read_lock &data->lock irq_context: 0 (wq_completion)bond89 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &idev->mc_lock rcu_read_lock rcu_read_lock rcu_read_lock_bh rcu_read_lock rcu_read_lock_bh rcu_read_lock &obj_hash[i].lock irq_context: 0 (wq_completion)bond89 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex.wait_lock irq_context: 0 (wq_completion)bond89 (work_completion)(&(&bond->mcast_work)->work) &p->pi_lock irq_context: 0 (wq_completion)bond71 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex fs_reclaim irq_context: 0 (wq_completion)bond71 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex fs_reclaim mmu_notifier_invalidate_range_start irq_context: 0 (wq_completion)bond71 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &c->lock irq_context: 0 (wq_completion)bond71 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &n->list_lock irq_context: 0 (wq_completion)bond71 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &n->list_lock &c->lock irq_context: 0 (wq_completion)bond71 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex rcu_read_lock &bond->stats_lock/1 irq_context: 0 (wq_completion)bond71 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex rcu_read_lock &ndev->lock irq_context: 0 (wq_completion)bond71 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex nl_table_lock irq_context: 0 (wq_completion)bond71 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &obj_hash[i].lock irq_context: 0 (wq_completion)bond71 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex nl_table_wait.lock irq_context: 0 (wq_completion)bond71 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex net_rwsem irq_context: 0 (wq_completion)bond71 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex net_rwsem &list->lock#2 irq_context: 0 (wq_completion)bond71 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &tn->lock irq_context: 0 (wq_completion)bond71 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &idev->mc_lock irq_context: 0 (wq_completion)bond71 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &idev->mc_lock fs_reclaim irq_context: 0 (wq_completion)bond71 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &idev->mc_lock fs_reclaim mmu_notifier_invalidate_range_start irq_context: 0 (wq_completion)bond71 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &idev->mc_lock &c->lock irq_context: 0 (wq_completion)bond71 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &idev->mc_lock rcu_read_lock &ndev->lock irq_context: 0 (wq_completion)bond71 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &idev->mc_lock rcu_read_lock &dir->lock#2 irq_context: 0 (wq_completion)bond71 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &idev->mc_lock rcu_read_lock &c->lock irq_context: 0 (wq_completion)bond71 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &idev->mc_lock rcu_read_lock &ul->lock#2 irq_context: 0 (wq_completion)bond71 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &idev->mc_lock rcu_read_lock rcu_read_lock &____s->seqcount#10 irq_context: 0 (wq_completion)bond71 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &idev->mc_lock rcu_read_lock rcu_read_lock rcu_read_lock_bh rcu_read_lock rcu_read_lock_bh rcu_read_lock &br->multicast_lock irq_context: 0 (wq_completion)bond71 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &idev->mc_lock rcu_read_lock rcu_read_lock rcu_read_lock_bh rcu_read_lock rcu_read_lock_bh rcu_read_lock &br->multicast_lock &obj_hash[i].lock irq_context: 0 (wq_completion)bond71 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &idev->mc_lock rcu_read_lock rcu_read_lock rcu_read_lock_bh rcu_read_lock rcu_read_lock_bh rcu_read_lock &br->multicast_lock &base->lock irq_context: 0 (wq_completion)bond71 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &idev->mc_lock rcu_read_lock rcu_read_lock rcu_read_lock_bh rcu_read_lock rcu_read_lock_bh rcu_read_lock &data->lock irq_context: 0 (wq_completion)bond71 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &idev->mc_lock rcu_read_lock rcu_read_lock rcu_read_lock_bh rcu_read_lock rcu_read_lock_bh rcu_read_lock &obj_hash[i].lock irq_context: 0 (wq_completion)bond71 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex.wait_lock irq_context: 0 (wq_completion)bond71 (work_completion)(&(&bond->mcast_work)->work) &p->pi_lock irq_context: 0 (wq_completion)bond73 (work_completion)(&(&slave->notify_work)->work) rtnl_mutex net_rwsem irq_context: 0 (wq_completion)bond73 (work_completion)(&(&slave->notify_work)->work) rtnl_mutex net_rwsem &list->lock#2 irq_context: 0 (wq_completion)bond73 (work_completion)(&(&slave->notify_work)->work) rtnl_mutex &tn->lock irq_context: 0 (wq_completion)bond73 (work_completion)(&(&slave->notify_work)->work) rtnl_mutex.wait_lock irq_context: 0 (wq_completion)bond73 (work_completion)(&(&slave->notify_work)->work) &p->pi_lock irq_context: 0 (wq_completion)bond102 (work_completion)(&(&slave->notify_work)->work) rtnl_mutex net_rwsem irq_context: 0 (wq_completion)bond102 (work_completion)(&(&slave->notify_work)->work) rtnl_mutex net_rwsem &list->lock#2 irq_context: 0 (wq_completion)bond102 (work_completion)(&(&slave->notify_work)->work) rtnl_mutex &tn->lock irq_context: 0 (wq_completion)bond102 (work_completion)(&(&slave->notify_work)->work) rtnl_mutex &rq->__lock irq_context: 0 (wq_completion)bond102 (work_completion)(&(&slave->notify_work)->work) rtnl_mutex &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)bond102 (work_completion)(&(&slave->notify_work)->work) rtnl_mutex.wait_lock irq_context: 0 (wq_completion)bond102 (work_completion)(&(&slave->notify_work)->work) &p->pi_lock irq_context: 0 (wq_completion)bond101 (work_completion)(&(&slave->notify_work)->work) rtnl_mutex net_rwsem irq_context: 0 (wq_completion)bond101 (work_completion)(&(&slave->notify_work)->work) rtnl_mutex net_rwsem &list->lock#2 irq_context: 0 (wq_completion)bond101 (work_completion)(&(&slave->notify_work)->work) rtnl_mutex &tn->lock irq_context: 0 (wq_completion)bond101 (work_completion)(&(&slave->notify_work)->work) rtnl_mutex.wait_lock irq_context: 0 (wq_completion)bond101 (work_completion)(&(&slave->notify_work)->work) &p->pi_lock irq_context: 0 (wq_completion)bond103 irq_context: 0 (wq_completion)bond103 &rq->__lock irq_context: 0 (wq_completion)bond103 &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)bond103 (work_completion)(&(&slave->notify_work)->work) irq_context: 0 (wq_completion)bond103 (work_completion)(&(&slave->notify_work)->work) &obj_hash[i].lock irq_context: 0 (wq_completion)bond103 (work_completion)(&(&slave->notify_work)->work) &base->lock irq_context: 0 (wq_completion)bond103 (work_completion)(&(&slave->notify_work)->work) &base->lock &obj_hash[i].lock irq_context: 0 (wq_completion)bond103 (work_completion)(&(&bond->mcast_work)->work) irq_context: 0 (wq_completion)bond103 (work_completion)(&(&bond->mcast_work)->work) &obj_hash[i].lock irq_context: 0 (wq_completion)bond103 (work_completion)(&(&bond->mcast_work)->work) &base->lock irq_context: 0 (wq_completion)bond103 (work_completion)(&(&bond->mcast_work)->work) &base->lock &obj_hash[i].lock irq_context: 0 (wq_completion)bond103 (work_completion)(&(&bond->mcast_work)->work) &rq->__lock irq_context: 0 (wq_completion)bond103 (work_completion)(&(&bond->mcast_work)->work) &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 rtnl_mutex cpu_hotplug_lock wq_pool_mutex &cfs_rq->removed.lock irq_context: 0 (wq_completion)bond74 irq_context: 0 (wq_completion)bond74 (work_completion)(&(&slave->notify_work)->work) irq_context: 0 (wq_completion)bond74 (work_completion)(&(&slave->notify_work)->work) &obj_hash[i].lock irq_context: 0 (wq_completion)bond74 (work_completion)(&(&slave->notify_work)->work) &base->lock irq_context: 0 (wq_completion)bond74 (work_completion)(&(&slave->notify_work)->work) &base->lock &obj_hash[i].lock irq_context: 0 (wq_completion)bond74 (work_completion)(&(&slave->notify_work)->work) &rq->__lock irq_context: 0 (wq_completion)bond74 (work_completion)(&(&slave->notify_work)->work) &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)bond74 (work_completion)(&(&bond->mcast_work)->work) irq_context: 0 (wq_completion)bond74 (work_completion)(&(&bond->mcast_work)->work) &obj_hash[i].lock irq_context: 0 (wq_completion)bond74 (work_completion)(&(&bond->mcast_work)->work) &base->lock irq_context: 0 (wq_completion)bond74 (work_completion)(&(&bond->mcast_work)->work) &base->lock &obj_hash[i].lock irq_context: 0 (wq_completion)gid-cache-wq (work_completion)(&work->work) devices_rwsem rcu_node_0 irq_context: 0 (wq_completion)gid-cache-wq (work_completion)(&work->work) devices_rwsem &rq->__lock irq_context: 0 (wq_completion)gid-cache-wq (work_completion)(&work->work) devices_rwsem &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)gid-cache-wq (work_completion)(&work->work) devices_rwsem rcu_read_lock rcu_node_0 irq_context: 0 (wq_completion)bond70 (work_completion)(&(&slave->notify_work)->work) rtnl_mutex net_rwsem irq_context: 0 (wq_completion)bond70 (work_completion)(&(&slave->notify_work)->work) rtnl_mutex net_rwsem &list->lock#2 irq_context: 0 (wq_completion)bond70 (work_completion)(&(&slave->notify_work)->work) rtnl_mutex &tn->lock irq_context: 0 (wq_completion)bond70 (work_completion)(&(&slave->notify_work)->work) rtnl_mutex.wait_lock irq_context: 0 (wq_completion)bond70 (work_completion)(&(&slave->notify_work)->work) &p->pi_lock irq_context: 0 (wq_completion)bond103 (work_completion)(&(&slave->notify_work)->work) rtnl_mutex net_rwsem irq_context: 0 (wq_completion)bond103 (work_completion)(&(&slave->notify_work)->work) rtnl_mutex net_rwsem &list->lock#2 irq_context: 0 (wq_completion)bond103 (work_completion)(&(&slave->notify_work)->work) rtnl_mutex &tn->lock irq_context: 0 (wq_completion)bond103 (work_completion)(&(&slave->notify_work)->work) rtnl_mutex.wait_lock irq_context: 0 (wq_completion)bond103 (work_completion)(&(&slave->notify_work)->work) &p->pi_lock irq_context: 0 (wq_completion)bond74 (work_completion)(&(&slave->notify_work)->work) rtnl_mutex net_rwsem irq_context: 0 (wq_completion)bond74 (work_completion)(&(&slave->notify_work)->work) rtnl_mutex net_rwsem &list->lock#2 irq_context: 0 (wq_completion)bond74 (work_completion)(&(&slave->notify_work)->work) rtnl_mutex &tn->lock irq_context: 0 (wq_completion)bond74 (work_completion)(&(&slave->notify_work)->work) rtnl_mutex.wait_lock irq_context: 0 (wq_completion)bond74 (work_completion)(&(&slave->notify_work)->work) &p->pi_lock irq_context: 0 (wq_completion)bond100 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex fs_reclaim irq_context: 0 (wq_completion)bond100 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex fs_reclaim mmu_notifier_invalidate_range_start irq_context: 0 (wq_completion)bond100 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &c->lock irq_context: 0 (wq_completion)bond100 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex rcu_read_lock &bond->stats_lock/1 irq_context: 0 (wq_completion)bond100 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex rcu_read_lock &ndev->lock irq_context: 0 (wq_completion)bond100 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex nl_table_lock irq_context: 0 (wq_completion)bond100 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &obj_hash[i].lock irq_context: 0 (wq_completion)bond100 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex nl_table_wait.lock irq_context: 0 (wq_completion)bond100 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex net_rwsem irq_context: 0 (wq_completion)bond100 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex net_rwsem &list->lock#2 irq_context: 0 (wq_completion)bond100 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &tn->lock irq_context: 0 (wq_completion)bond100 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &idev->mc_lock irq_context: 0 (wq_completion)bond100 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &idev->mc_lock fs_reclaim irq_context: 0 (wq_completion)bond100 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &idev->mc_lock fs_reclaim mmu_notifier_invalidate_range_start irq_context: 0 (wq_completion)bond100 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &idev->mc_lock rcu_read_lock &ndev->lock irq_context: 0 (wq_completion)bond100 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &idev->mc_lock rcu_read_lock &dir->lock#2 irq_context: 0 (wq_completion)bond100 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &idev->mc_lock rcu_read_lock &c->lock irq_context: 0 (wq_completion)bond100 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &idev->mc_lock rcu_read_lock &ul->lock#2 irq_context: 0 (wq_completion)bond100 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &idev->mc_lock rcu_read_lock rcu_read_lock &____s->seqcount#10 irq_context: 0 (wq_completion)bond100 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &idev->mc_lock rcu_read_lock rcu_read_lock rcu_read_lock_bh rcu_read_lock rcu_read_lock_bh rcu_read_lock &br->multicast_lock irq_context: 0 (wq_completion)bond100 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &idev->mc_lock rcu_read_lock rcu_read_lock rcu_read_lock_bh rcu_read_lock rcu_read_lock_bh rcu_read_lock &br->multicast_lock &obj_hash[i].lock irq_context: 0 (wq_completion)bond100 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &idev->mc_lock rcu_read_lock rcu_read_lock rcu_read_lock_bh rcu_read_lock rcu_read_lock_bh rcu_read_lock &br->multicast_lock &base->lock irq_context: 0 (wq_completion)bond100 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &idev->mc_lock rcu_read_lock rcu_read_lock rcu_read_lock_bh rcu_read_lock rcu_read_lock_bh rcu_read_lock &data->lock irq_context: 0 (wq_completion)bond100 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &idev->mc_lock rcu_read_lock rcu_read_lock rcu_read_lock_bh rcu_read_lock rcu_read_lock_bh rcu_read_lock &obj_hash[i].lock irq_context: 0 (wq_completion)bond100 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex.wait_lock irq_context: 0 (wq_completion)bond100 (work_completion)(&(&bond->mcast_work)->work) &p->pi_lock irq_context: 0 (wq_completion)bond102 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex fs_reclaim irq_context: 0 (wq_completion)bond102 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex fs_reclaim mmu_notifier_invalidate_range_start irq_context: 0 (wq_completion)bond102 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &c->lock irq_context: 0 (wq_completion)bond102 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &n->list_lock irq_context: 0 (wq_completion)bond102 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &n->list_lock &c->lock irq_context: 0 (wq_completion)bond102 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex rcu_read_lock &bond->stats_lock/1 irq_context: 0 (wq_completion)bond102 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex rcu_read_lock &ndev->lock irq_context: 0 (wq_completion)bond102 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex nl_table_lock irq_context: 0 (wq_completion)bond102 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &obj_hash[i].lock irq_context: 0 (wq_completion)bond102 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex nl_table_wait.lock irq_context: 0 (wq_completion)bond102 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex net_rwsem irq_context: 0 (wq_completion)bond102 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex net_rwsem &list->lock#2 irq_context: 0 (wq_completion)bond102 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &tn->lock irq_context: 0 (wq_completion)bond102 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &idev->mc_lock irq_context: 0 (wq_completion)bond102 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &idev->mc_lock fs_reclaim irq_context: 0 (wq_completion)bond102 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &idev->mc_lock fs_reclaim mmu_notifier_invalidate_range_start irq_context: 0 (wq_completion)bond102 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &idev->mc_lock &c->lock irq_context: 0 (wq_completion)bond102 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &idev->mc_lock &n->list_lock irq_context: 0 (wq_completion)bond102 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &idev->mc_lock &n->list_lock &c->lock irq_context: 0 (wq_completion)bond102 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &idev->mc_lock rcu_read_lock &ndev->lock irq_context: 0 (wq_completion)bond102 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &idev->mc_lock rcu_read_lock &c->lock irq_context: 0 (wq_completion)bond102 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &idev->mc_lock rcu_read_lock &dir->lock#2 irq_context: 0 (wq_completion)bond102 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &idev->mc_lock rcu_read_lock &ul->lock#2 irq_context: 0 (wq_completion)bond102 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &idev->mc_lock rcu_read_lock rcu_read_lock &____s->seqcount#10 irq_context: 0 (wq_completion)bond102 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &idev->mc_lock rcu_read_lock rcu_read_lock rcu_read_lock_bh rcu_read_lock rcu_read_lock_bh rcu_read_lock &br->multicast_lock irq_context: 0 (wq_completion)bond102 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &idev->mc_lock rcu_read_lock rcu_read_lock rcu_read_lock_bh rcu_read_lock rcu_read_lock_bh rcu_read_lock &br->multicast_lock &obj_hash[i].lock irq_context: 0 (wq_completion)bond102 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &idev->mc_lock rcu_read_lock rcu_read_lock rcu_read_lock_bh rcu_read_lock rcu_read_lock_bh rcu_read_lock &br->multicast_lock &base->lock irq_context: 0 (wq_completion)bond102 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &idev->mc_lock rcu_read_lock rcu_read_lock rcu_read_lock_bh rcu_read_lock rcu_read_lock_bh rcu_read_lock &data->lock irq_context: 0 (wq_completion)bond102 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &idev->mc_lock rcu_read_lock rcu_read_lock rcu_read_lock_bh rcu_read_lock rcu_read_lock_bh rcu_read_lock &obj_hash[i].lock irq_context: 0 (wq_completion)bond102 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex.wait_lock irq_context: 0 (wq_completion)bond102 (work_completion)(&(&bond->mcast_work)->work) &p->pi_lock irq_context: 0 (wq_completion)bond73 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex fs_reclaim irq_context: 0 (wq_completion)bond73 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex fs_reclaim mmu_notifier_invalidate_range_start irq_context: 0 (wq_completion)bond73 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &c->lock irq_context: 0 (wq_completion)bond73 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex rcu_read_lock &bond->stats_lock/1 irq_context: 0 (wq_completion)bond73 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex rcu_read_lock &ndev->lock irq_context: 0 (wq_completion)bond73 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex nl_table_lock irq_context: 0 (wq_completion)bond73 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &obj_hash[i].lock irq_context: 0 (wq_completion)bond73 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex nl_table_wait.lock irq_context: 0 (wq_completion)bond73 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex net_rwsem irq_context: 0 (wq_completion)bond73 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex net_rwsem &list->lock#2 irq_context: 0 (wq_completion)bond73 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &tn->lock irq_context: 0 (wq_completion)bond73 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &idev->mc_lock irq_context: 0 (wq_completion)bond73 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &idev->mc_lock fs_reclaim irq_context: 0 (wq_completion)bond73 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &idev->mc_lock fs_reclaim &rq->__lock irq_context: 0 (wq_completion)bond73 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &idev->mc_lock fs_reclaim mmu_notifier_invalidate_range_start irq_context: 0 (wq_completion)bond73 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &idev->mc_lock &c->lock irq_context: 0 (wq_completion)bond73 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &idev->mc_lock rcu_read_lock &ndev->lock irq_context: 0 (wq_completion)bond73 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &idev->mc_lock rcu_read_lock &dir->lock#2 irq_context: 0 (wq_completion)bond73 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &idev->mc_lock rcu_read_lock &ul->lock#2 irq_context: 0 (wq_completion)bond73 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &idev->mc_lock rcu_read_lock rcu_read_lock &____s->seqcount#10 irq_context: 0 (wq_completion)bond73 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &idev->mc_lock rcu_read_lock rcu_read_lock rcu_read_lock_bh rcu_read_lock rcu_read_lock_bh rcu_read_lock &br->multicast_lock irq_context: 0 (wq_completion)bond73 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &idev->mc_lock rcu_read_lock rcu_read_lock rcu_read_lock_bh rcu_read_lock rcu_read_lock_bh rcu_read_lock &br->multicast_lock &obj_hash[i].lock irq_context: 0 (wq_completion)bond73 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &idev->mc_lock rcu_read_lock rcu_read_lock rcu_read_lock_bh rcu_read_lock rcu_read_lock_bh rcu_read_lock &br->multicast_lock &base->lock irq_context: 0 (wq_completion)bond73 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &idev->mc_lock rcu_read_lock rcu_read_lock rcu_read_lock_bh rcu_read_lock rcu_read_lock_bh rcu_read_lock &data->lock irq_context: 0 (wq_completion)bond73 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &idev->mc_lock rcu_read_lock rcu_read_lock rcu_read_lock_bh rcu_read_lock rcu_read_lock_bh rcu_read_lock &obj_hash[i].lock irq_context: 0 (wq_completion)bond73 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &idev->mc_lock rcu_read_lock rcu_read_lock rcu_node_0 irq_context: 0 (wq_completion)bond73 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &idev->mc_lock rcu_read_lock rcu_read_lock &rq->__lock irq_context: 0 (wq_completion)bond73 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &idev->mc_lock rcu_read_lock rcu_node_0 irq_context: 0 (wq_completion)bond73 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex.wait_lock irq_context: 0 (wq_completion)bond73 (work_completion)(&(&bond->mcast_work)->work) &p->pi_lock irq_context: 0 (wq_completion)bond73 (work_completion)(&(&bond->mcast_work)->work) &p->pi_lock &rq->__lock irq_context: 0 (wq_completion)bond73 (work_completion)(&(&bond->mcast_work)->work) &p->pi_lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)bond73 (work_completion)(&(&bond->mcast_work)->work) &rq->__lock irq_context: 0 (wq_completion)bond73 (work_completion)(&(&bond->mcast_work)->work) &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)bat_events (work_completion)(&(&forw_packet_aggr->delayed_work)->work) &pcp->lock &zone->lock irq_context: 0 (wq_completion)bat_events (work_completion)(&(&forw_packet_aggr->delayed_work)->work) &pcp->lock &zone->lock &____s->seqcount irq_context: 0 (wq_completion)bond104 irq_context: 0 (wq_completion)bond104 (work_completion)(&(&slave->notify_work)->work) irq_context: 0 (wq_completion)bond104 (work_completion)(&(&slave->notify_work)->work) &obj_hash[i].lock irq_context: 0 (wq_completion)bond104 (work_completion)(&(&slave->notify_work)->work) &base->lock irq_context: 0 (wq_completion)bond104 (work_completion)(&(&slave->notify_work)->work) &base->lock &obj_hash[i].lock irq_context: 0 (wq_completion)bond104 (work_completion)(&(&bond->mcast_work)->work) irq_context: 0 (wq_completion)bond104 (work_completion)(&(&bond->mcast_work)->work) &obj_hash[i].lock irq_context: 0 (wq_completion)bond104 (work_completion)(&(&bond->mcast_work)->work) &base->lock irq_context: 0 (wq_completion)bond104 (work_completion)(&(&bond->mcast_work)->work) &base->lock &obj_hash[i].lock irq_context: 0 (wq_completion)bond92 irq_context: 0 (wq_completion)bond92 (work_completion)(&(&slave->notify_work)->work) irq_context: 0 (wq_completion)bond92 (work_completion)(&(&slave->notify_work)->work) &obj_hash[i].lock irq_context: 0 (wq_completion)bond92 (work_completion)(&(&slave->notify_work)->work) &base->lock irq_context: 0 (wq_completion)bond92 (work_completion)(&(&slave->notify_work)->work) &base->lock &obj_hash[i].lock irq_context: 0 (wq_completion)bond92 (work_completion)(&(&bond->mcast_work)->work) irq_context: 0 (wq_completion)bond92 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex fs_reclaim irq_context: 0 (wq_completion)bond92 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex fs_reclaim mmu_notifier_invalidate_range_start irq_context: 0 (wq_completion)bond92 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &c->lock irq_context: 0 (wq_completion)bond92 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex rcu_read_lock &bond->stats_lock/1 irq_context: 0 (wq_completion)bond92 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex rcu_read_lock &ndev->lock irq_context: 0 (wq_completion)bond92 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex nl_table_lock irq_context: 0 (wq_completion)bond92 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &obj_hash[i].lock irq_context: 0 (wq_completion)bond92 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex nl_table_wait.lock irq_context: 0 (wq_completion)bond92 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex net_rwsem irq_context: 0 (wq_completion)bond92 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex net_rwsem &list->lock#2 irq_context: 0 (wq_completion)bond92 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &tn->lock irq_context: 0 (wq_completion)bond92 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &idev->mc_lock irq_context: 0 (wq_completion)bond92 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex.wait_lock irq_context: 0 (wq_completion)bond92 (work_completion)(&(&bond->mcast_work)->work) &p->pi_lock irq_context: 0 (wq_completion)bond92 (work_completion)(&(&slave->notify_work)->work) rtnl_mutex net_rwsem irq_context: 0 (wq_completion)bond92 (work_completion)(&(&slave->notify_work)->work) rtnl_mutex net_rwsem &list->lock#2 irq_context: 0 (wq_completion)bond92 (work_completion)(&(&slave->notify_work)->work) rtnl_mutex &tn->lock irq_context: 0 (wq_completion)bond92 (work_completion)(&(&slave->notify_work)->work) rtnl_mutex.wait_lock irq_context: 0 (wq_completion)bond92 (work_completion)(&(&slave->notify_work)->work) &p->pi_lock irq_context: 0 (wq_completion)bond104 (work_completion)(&(&slave->notify_work)->work) rtnl_mutex net_rwsem irq_context: 0 (wq_completion)bond104 (work_completion)(&(&slave->notify_work)->work) rtnl_mutex net_rwsem &list->lock#2 irq_context: 0 (wq_completion)bond104 (work_completion)(&(&slave->notify_work)->work) rtnl_mutex &tn->lock irq_context: 0 (wq_completion)bond104 (work_completion)(&(&slave->notify_work)->work) rtnl_mutex.wait_lock irq_context: 0 (wq_completion)bond104 (work_completion)(&(&slave->notify_work)->work) &p->pi_lock irq_context: 0 (wq_completion)bond100 (work_completion)(&(&slave->notify_work)->work) rtnl_mutex net_rwsem irq_context: 0 (wq_completion)bond100 (work_completion)(&(&slave->notify_work)->work) rtnl_mutex net_rwsem &list->lock#2 irq_context: 0 (wq_completion)bond100 (work_completion)(&(&slave->notify_work)->work) rtnl_mutex &tn->lock irq_context: 0 (wq_completion)bond100 (work_completion)(&(&slave->notify_work)->work) rtnl_mutex.wait_lock irq_context: 0 (wq_completion)bond100 (work_completion)(&(&slave->notify_work)->work) &p->pi_lock irq_context: 0 (wq_completion)bond71 (work_completion)(&(&slave->notify_work)->work) rtnl_mutex net_rwsem irq_context: 0 (wq_completion)bond71 (work_completion)(&(&slave->notify_work)->work) rtnl_mutex net_rwsem &list->lock#2 irq_context: 0 (wq_completion)bond71 (work_completion)(&(&slave->notify_work)->work) rtnl_mutex &tn->lock irq_context: 0 (wq_completion)bond71 (work_completion)(&(&slave->notify_work)->work) rtnl_mutex.wait_lock irq_context: 0 (wq_completion)bond71 (work_completion)(&(&slave->notify_work)->work) &p->pi_lock irq_context: 0 (wq_completion)bond103 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex fs_reclaim irq_context: 0 (wq_completion)bond103 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex fs_reclaim mmu_notifier_invalidate_range_start irq_context: 0 (wq_completion)bond103 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &c->lock irq_context: 0 (wq_completion)bond103 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex rcu_read_lock &bond->stats_lock/1 irq_context: 0 (wq_completion)bond103 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex rcu_read_lock &ndev->lock irq_context: 0 (wq_completion)bond103 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex nl_table_lock irq_context: 0 (wq_completion)bond103 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &obj_hash[i].lock irq_context: 0 (wq_completion)bond103 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &rq->__lock irq_context: 0 (wq_completion)bond103 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)bond103 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex nl_table_wait.lock irq_context: 0 (wq_completion)bond103 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex net_rwsem irq_context: 0 (wq_completion)bond103 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex net_rwsem &list->lock#2 irq_context: 0 (wq_completion)bond103 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &tn->lock irq_context: 0 (wq_completion)bond103 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &idev->mc_lock irq_context: 0 (wq_completion)bond103 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &idev->mc_lock fs_reclaim irq_context: 0 (wq_completion)bond103 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &idev->mc_lock fs_reclaim mmu_notifier_invalidate_range_start irq_context: 0 (wq_completion)bond103 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &idev->mc_lock rcu_read_lock &ndev->lock irq_context: 0 (wq_completion)bond103 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &idev->mc_lock rcu_read_lock &dir->lock#2 irq_context: 0 (wq_completion)bond103 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &idev->mc_lock rcu_read_lock &ul->lock#2 irq_context: 0 (wq_completion)bond103 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &idev->mc_lock rcu_read_lock rcu_read_lock &____s->seqcount#10 irq_context: 0 (wq_completion)bond103 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &idev->mc_lock rcu_read_lock rcu_read_lock rcu_read_lock_bh rcu_read_lock rcu_read_lock_bh rcu_read_lock &br->multicast_lock irq_context: 0 (wq_completion)bond103 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &idev->mc_lock rcu_read_lock rcu_read_lock rcu_read_lock_bh rcu_read_lock rcu_read_lock_bh rcu_read_lock &br->multicast_lock &obj_hash[i].lock irq_context: 0 (wq_completion)bond103 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &idev->mc_lock rcu_read_lock rcu_read_lock rcu_read_lock_bh rcu_read_lock rcu_read_lock_bh rcu_read_lock &br->multicast_lock &base->lock irq_context: 0 (wq_completion)bond103 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &idev->mc_lock rcu_read_lock rcu_read_lock rcu_read_lock_bh rcu_read_lock rcu_read_lock_bh rcu_read_lock &data->lock irq_context: 0 (wq_completion)bond103 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &idev->mc_lock rcu_read_lock rcu_read_lock rcu_read_lock_bh rcu_read_lock rcu_read_lock_bh rcu_read_lock &obj_hash[i].lock irq_context: 0 (wq_completion)bond103 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex.wait_lock irq_context: 0 (wq_completion)bond103 (work_completion)(&(&bond->mcast_work)->work) &p->pi_lock irq_context: 0 (wq_completion)bond70 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex fs_reclaim irq_context: 0 (wq_completion)bond70 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex fs_reclaim mmu_notifier_invalidate_range_start irq_context: 0 (wq_completion)bond70 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &c->lock irq_context: 0 (wq_completion)bond70 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex rcu_read_lock &bond->stats_lock/1 irq_context: 0 (wq_completion)bond70 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex rcu_read_lock &ndev->lock irq_context: 0 (wq_completion)bond70 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex nl_table_lock irq_context: 0 (wq_completion)bond70 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &obj_hash[i].lock irq_context: 0 (wq_completion)bond70 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex nl_table_wait.lock irq_context: 0 (wq_completion)bond70 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex net_rwsem irq_context: 0 (wq_completion)bond70 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex net_rwsem &list->lock#2 irq_context: 0 (wq_completion)bond70 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &tn->lock irq_context: 0 (wq_completion)bond70 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &idev->mc_lock irq_context: 0 (wq_completion)bond70 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &idev->mc_lock fs_reclaim irq_context: 0 (wq_completion)bond70 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &idev->mc_lock fs_reclaim mmu_notifier_invalidate_range_start irq_context: 0 (wq_completion)bond70 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &idev->mc_lock rcu_read_lock &ndev->lock irq_context: 0 (wq_completion)bond70 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &idev->mc_lock rcu_read_lock &c->lock irq_context: 0 (wq_completion)bond70 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &idev->mc_lock rcu_read_lock &dir->lock#2 irq_context: 0 (wq_completion)bond70 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &idev->mc_lock rcu_read_lock &ul->lock#2 irq_context: 0 (wq_completion)bond70 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &idev->mc_lock rcu_read_lock rcu_read_lock &____s->seqcount#10 irq_context: 0 (wq_completion)bond70 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &idev->mc_lock rcu_read_lock rcu_read_lock rcu_read_lock_bh rcu_read_lock rcu_read_lock_bh rcu_read_lock &br->multicast_lock irq_context: 0 (wq_completion)bond70 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &idev->mc_lock rcu_read_lock rcu_read_lock rcu_read_lock_bh rcu_read_lock rcu_read_lock_bh rcu_read_lock &br->multicast_lock &obj_hash[i].lock irq_context: 0 (wq_completion)bond70 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &idev->mc_lock rcu_read_lock rcu_read_lock rcu_read_lock_bh rcu_read_lock rcu_read_lock_bh rcu_read_lock &br->multicast_lock &base->lock irq_context: 0 (wq_completion)bond70 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &idev->mc_lock rcu_read_lock rcu_read_lock rcu_read_lock_bh rcu_read_lock rcu_read_lock_bh rcu_read_lock &data->lock irq_context: 0 (wq_completion)bond70 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &idev->mc_lock rcu_read_lock rcu_read_lock rcu_read_lock_bh rcu_read_lock rcu_read_lock_bh rcu_read_lock &obj_hash[i].lock irq_context: 0 (wq_completion)bond70 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex.wait_lock irq_context: 0 (wq_completion)bond70 (work_completion)(&(&bond->mcast_work)->work) &p->pi_lock irq_context: 0 (wq_completion)bond74 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex fs_reclaim irq_context: 0 (wq_completion)bond74 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex fs_reclaim mmu_notifier_invalidate_range_start irq_context: 0 (wq_completion)bond74 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &c->lock irq_context: 0 (wq_completion)bond74 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &n->list_lock irq_context: 0 (wq_completion)bond74 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &n->list_lock &c->lock irq_context: 0 (wq_completion)bond74 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex rcu_read_lock &bond->stats_lock/1 irq_context: 0 (wq_completion)bond74 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex rcu_read_lock &ndev->lock irq_context: 0 (wq_completion)bond74 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex nl_table_lock irq_context: 0 (wq_completion)bond74 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &obj_hash[i].lock irq_context: 0 (wq_completion)bond74 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex nl_table_wait.lock irq_context: 0 (wq_completion)bond74 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex net_rwsem irq_context: 0 (wq_completion)bond74 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex net_rwsem &list->lock#2 irq_context: 0 (wq_completion)bond74 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &tn->lock irq_context: 0 (wq_completion)bond74 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &idev->mc_lock irq_context: 0 (wq_completion)bond74 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex.wait_lock irq_context: 0 (wq_completion)bond74 (work_completion)(&(&bond->mcast_work)->work) &p->pi_lock irq_context: 0 (wq_completion)bond104 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex fs_reclaim irq_context: 0 (wq_completion)bond104 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex fs_reclaim mmu_notifier_invalidate_range_start irq_context: 0 (wq_completion)bond104 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex rcu_read_lock &bond->stats_lock/1 irq_context: 0 (wq_completion)bond104 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex rcu_read_lock &ndev->lock irq_context: 0 (wq_completion)bond104 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex nl_table_lock irq_context: 0 (wq_completion)bond104 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &obj_hash[i].lock irq_context: 0 (wq_completion)bond104 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex nl_table_wait.lock irq_context: 0 (wq_completion)bond104 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex net_rwsem irq_context: 0 (wq_completion)bond104 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex net_rwsem &list->lock#2 irq_context: 0 (wq_completion)bond104 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &tn->lock irq_context: 0 (wq_completion)bond104 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &idev->mc_lock irq_context: 0 (wq_completion)bond104 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &rq->__lock irq_context: 0 (wq_completion)bond104 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex.wait_lock irq_context: 0 (wq_completion)bond104 (work_completion)(&(&bond->mcast_work)->work) &p->pi_lock irq_context: 0 sb_writers#4 &type->i_mutex_dir_key#3/1 fill_pool_map-wait-type-override rcu_read_lock &rcu_state.gp_wq irq_context: 0 sb_writers#4 &type->i_mutex_dir_key#3/1 fill_pool_map-wait-type-override rcu_read_lock &rcu_state.gp_wq &p->pi_lock irq_context: 0 sb_writers#4 &type->i_mutex_dir_key#3/1 fill_pool_map-wait-type-override rcu_read_lock &rcu_state.gp_wq &p->pi_lock &rq->__lock irq_context: 0 sb_writers#4 &type->i_mutex_dir_key#3/1 fill_pool_map-wait-type-override rcu_read_lock &rcu_state.gp_wq &p->pi_lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)bond75 irq_context: 0 (wq_completion)bond75 (work_completion)(&(&slave->notify_work)->work) irq_context: 0 (wq_completion)bond75 (work_completion)(&(&slave->notify_work)->work) &rq->__lock irq_context: 0 (wq_completion)bond75 (work_completion)(&(&slave->notify_work)->work) &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)bond75 (work_completion)(&(&slave->notify_work)->work) &obj_hash[i].lock irq_context: 0 (wq_completion)bond75 (work_completion)(&(&slave->notify_work)->work) &base->lock irq_context: 0 (wq_completion)bond75 (work_completion)(&(&slave->notify_work)->work) &base->lock &obj_hash[i].lock irq_context: 0 (wq_completion)bond75 (work_completion)(&(&bond->mcast_work)->work) irq_context: 0 (wq_completion)bond75 (work_completion)(&(&bond->mcast_work)->work) &obj_hash[i].lock irq_context: 0 (wq_completion)bond75 (work_completion)(&(&bond->mcast_work)->work) &base->lock irq_context: 0 (wq_completion)bond75 (work_completion)(&(&bond->mcast_work)->work) &base->lock &obj_hash[i].lock irq_context: 0 (wq_completion)bond105 irq_context: 0 (wq_completion)bond105 (work_completion)(&(&slave->notify_work)->work) irq_context: 0 (wq_completion)bond105 (work_completion)(&(&slave->notify_work)->work) &obj_hash[i].lock irq_context: 0 (wq_completion)bond105 (work_completion)(&(&slave->notify_work)->work) &base->lock irq_context: 0 (wq_completion)bond105 (work_completion)(&(&slave->notify_work)->work) &base->lock &obj_hash[i].lock irq_context: 0 (wq_completion)bond105 (work_completion)(&(&slave->notify_work)->work) &rq->__lock irq_context: 0 (wq_completion)bond105 (work_completion)(&(&slave->notify_work)->work) &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)bond105 (work_completion)(&(&bond->mcast_work)->work) irq_context: 0 (wq_completion)bond105 (work_completion)(&(&bond->mcast_work)->work) &rq->__lock irq_context: 0 (wq_completion)bond105 (work_completion)(&(&bond->mcast_work)->work) &obj_hash[i].lock irq_context: 0 (wq_completion)bond105 (work_completion)(&(&bond->mcast_work)->work) &base->lock irq_context: 0 (wq_completion)bond105 (work_completion)(&(&bond->mcast_work)->work) &base->lock &obj_hash[i].lock irq_context: 0 (wq_completion)bond105 (work_completion)(&(&bond->mcast_work)->work) &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)bond105 (work_completion)(&(&slave->notify_work)->work) rtnl_mutex net_rwsem irq_context: 0 (wq_completion)bond105 (work_completion)(&(&slave->notify_work)->work) rtnl_mutex net_rwsem &list->lock#2 irq_context: 0 (wq_completion)bond105 (work_completion)(&(&slave->notify_work)->work) rtnl_mutex &tn->lock irq_context: 0 (wq_completion)bond105 (work_completion)(&(&slave->notify_work)->work) rtnl_mutex.wait_lock irq_context: 0 (wq_completion)bond105 (work_completion)(&(&slave->notify_work)->work) &p->pi_lock irq_context: 0 (wq_completion)bond75 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex fs_reclaim irq_context: 0 (wq_completion)bond75 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex fs_reclaim mmu_notifier_invalidate_range_start irq_context: 0 (wq_completion)bond75 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &c->lock irq_context: 0 (wq_completion)bond75 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex rcu_read_lock &bond->stats_lock/1 irq_context: 0 (wq_completion)bond75 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex rcu_read_lock &ndev->lock irq_context: 0 (wq_completion)bond75 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex nl_table_lock irq_context: 0 (wq_completion)bond75 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &obj_hash[i].lock irq_context: 0 (wq_completion)bond75 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex nl_table_wait.lock irq_context: 0 (wq_completion)bond75 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex net_rwsem irq_context: 0 (wq_completion)bond75 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex net_rwsem &list->lock#2 irq_context: 0 (wq_completion)bond75 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &tn->lock irq_context: 0 (wq_completion)bond75 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &idev->mc_lock irq_context: 0 (wq_completion)bond75 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex.wait_lock irq_context: 0 (wq_completion)bond75 (work_completion)(&(&bond->mcast_work)->work) &p->pi_lock irq_context: 0 (wq_completion)bond93 irq_context: 0 (wq_completion)bond93 (work_completion)(&(&slave->notify_work)->work) irq_context: 0 (wq_completion)bond93 (work_completion)(&(&slave->notify_work)->work) &obj_hash[i].lock irq_context: 0 (wq_completion)bond93 (work_completion)(&(&slave->notify_work)->work) &base->lock irq_context: 0 (wq_completion)bond93 (work_completion)(&(&slave->notify_work)->work) &base->lock &obj_hash[i].lock irq_context: 0 (wq_completion)bond93 (work_completion)(&(&slave->notify_work)->work) &rq->__lock irq_context: 0 (wq_completion)bond93 (work_completion)(&(&slave->notify_work)->work) &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)bond93 (work_completion)(&(&bond->mcast_work)->work) irq_context: 0 (wq_completion)bond93 (work_completion)(&(&bond->mcast_work)->work) &obj_hash[i].lock irq_context: 0 (wq_completion)bond93 (work_completion)(&(&bond->mcast_work)->work) &base->lock irq_context: 0 (wq_completion)bond93 (work_completion)(&(&bond->mcast_work)->work) &base->lock &obj_hash[i].lock irq_context: 0 (wq_completion)bond76 irq_context: 0 (wq_completion)bond76 &rq->__lock irq_context: 0 (wq_completion)bond76 &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)bond76 (work_completion)(&(&slave->notify_work)->work) irq_context: 0 (wq_completion)bond76 (work_completion)(&(&slave->notify_work)->work) &obj_hash[i].lock irq_context: 0 (wq_completion)bond76 (work_completion)(&(&slave->notify_work)->work) &base->lock irq_context: 0 (wq_completion)bond76 (work_completion)(&(&slave->notify_work)->work) &base->lock &obj_hash[i].lock irq_context: 0 (wq_completion)bond76 (work_completion)(&(&slave->notify_work)->work) &rq->__lock irq_context: 0 (wq_completion)bond76 (work_completion)(&(&slave->notify_work)->work) &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)bond76 (work_completion)(&(&bond->mcast_work)->work) irq_context: 0 (wq_completion)bond76 (work_completion)(&(&bond->mcast_work)->work) &obj_hash[i].lock irq_context: 0 (wq_completion)bond76 (work_completion)(&(&bond->mcast_work)->work) &base->lock irq_context: 0 (wq_completion)bond76 (work_completion)(&(&bond->mcast_work)->work) &base->lock &obj_hash[i].lock irq_context: 0 (wq_completion)bond106 irq_context: 0 (wq_completion)bond106 (work_completion)(&(&slave->notify_work)->work) irq_context: 0 (wq_completion)bond106 (work_completion)(&(&slave->notify_work)->work) &obj_hash[i].lock irq_context: 0 (wq_completion)bond106 (work_completion)(&(&slave->notify_work)->work) &base->lock irq_context: 0 (wq_completion)bond106 (work_completion)(&(&slave->notify_work)->work) &base->lock &obj_hash[i].lock irq_context: 0 (wq_completion)bond106 (work_completion)(&(&bond->mcast_work)->work) irq_context: 0 (wq_completion)bond106 (work_completion)(&(&bond->mcast_work)->work) &obj_hash[i].lock irq_context: 0 (wq_completion)bond106 (work_completion)(&(&bond->mcast_work)->work) &base->lock irq_context: 0 (wq_completion)bond106 (work_completion)(&(&bond->mcast_work)->work) &base->lock &obj_hash[i].lock irq_context: 0 (wq_completion)bond106 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex fs_reclaim irq_context: 0 (wq_completion)bond106 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex fs_reclaim mmu_notifier_invalidate_range_start irq_context: 0 (wq_completion)bond106 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &c->lock irq_context: 0 (wq_completion)bond106 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex rcu_read_lock &bond->stats_lock/1 irq_context: 0 (wq_completion)bond106 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex rcu_read_lock &ndev->lock irq_context: 0 (wq_completion)bond106 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex nl_table_lock irq_context: 0 (wq_completion)bond106 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &obj_hash[i].lock irq_context: 0 (wq_completion)bond106 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex nl_table_wait.lock irq_context: 0 (wq_completion)bond106 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex net_rwsem irq_context: 0 (wq_completion)bond106 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex net_rwsem &list->lock#2 irq_context: 0 (wq_completion)bond106 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &tn->lock irq_context: 0 (wq_completion)bond106 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &idev->mc_lock irq_context: 0 (wq_completion)bond106 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &idev->mc_lock fs_reclaim irq_context: 0 (wq_completion)bond106 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &idev->mc_lock fs_reclaim mmu_notifier_invalidate_range_start irq_context: 0 (wq_completion)bond106 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &idev->mc_lock rcu_read_lock &ndev->lock irq_context: 0 (wq_completion)bond106 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &idev->mc_lock rcu_read_lock &c->lock irq_context: 0 (wq_completion)bond106 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &idev->mc_lock rcu_read_lock &dir->lock#2 irq_context: 0 (wq_completion)bond106 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &idev->mc_lock rcu_read_lock &ul->lock#2 irq_context: 0 (wq_completion)bond106 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &idev->mc_lock rcu_read_lock rcu_read_lock &____s->seqcount#10 irq_context: 0 (wq_completion)bond106 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &idev->mc_lock rcu_read_lock rcu_read_lock rcu_read_lock_bh rcu_read_lock rcu_read_lock_bh rcu_read_lock &br->multicast_lock irq_context: 0 (wq_completion)bond106 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &idev->mc_lock rcu_read_lock rcu_read_lock rcu_read_lock_bh rcu_read_lock rcu_read_lock_bh rcu_read_lock &br->multicast_lock &obj_hash[i].lock irq_context: 0 (wq_completion)bond106 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &idev->mc_lock rcu_read_lock rcu_read_lock rcu_read_lock_bh rcu_read_lock rcu_read_lock_bh rcu_read_lock &br->multicast_lock &base->lock irq_context: 0 (wq_completion)bond106 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &idev->mc_lock rcu_read_lock rcu_read_lock rcu_read_lock_bh rcu_read_lock rcu_read_lock_bh rcu_read_lock &data->lock irq_context: 0 (wq_completion)bond106 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &idev->mc_lock rcu_read_lock rcu_read_lock rcu_read_lock_bh rcu_read_lock rcu_read_lock_bh rcu_read_lock &obj_hash[i].lock irq_context: 0 (wq_completion)bond106 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex.wait_lock irq_context: 0 (wq_completion)bond106 (work_completion)(&(&bond->mcast_work)->work) &p->pi_lock irq_context: 0 (wq_completion)bond77 irq_context: 0 (wq_completion)bond77 (work_completion)(&(&slave->notify_work)->work) irq_context: 0 (wq_completion)bond77 (work_completion)(&(&slave->notify_work)->work) &obj_hash[i].lock irq_context: 0 (wq_completion)bond77 (work_completion)(&(&slave->notify_work)->work) &base->lock irq_context: 0 (wq_completion)bond77 (work_completion)(&(&slave->notify_work)->work) &base->lock &obj_hash[i].lock irq_context: 0 sk_lock-AF_INET rcu_read_lock rcu_read_lock rcu_read_lock &nf_conntrack_locks[i] irq_context: 0 sk_lock-AF_INET rcu_read_lock rcu_read_lock rcu_read_lock &nf_conntrack_locks[i] &nf_conntrack_locks[i]/1 irq_context: 0 sk_lock-AF_INET rcu_read_lock rcu_read_lock rcu_read_lock rcu_read_lock &nf_nat_locks[i] irq_context: 0 sk_lock-AF_INET rcu_read_lock rcu_read_lock rcu_read_lock &obj_hash[i].lock irq_context: 0 sk_lock-AF_INET rcu_read_lock rcu_read_lock rcu_read_lock pool_lock#2 irq_context: 0 sk_lock-AF_INET rcu_read_lock rcu_read_lock rcu_read_lock &c->lock irq_context: 0 (wq_completion)bond77 (work_completion)(&(&bond->mcast_work)->work) irq_context: 0 (wq_completion)bond77 (work_completion)(&(&bond->mcast_work)->work) &obj_hash[i].lock irq_context: 0 (wq_completion)bond77 (work_completion)(&(&bond->mcast_work)->work) &base->lock irq_context: 0 (wq_completion)bond77 (work_completion)(&(&bond->mcast_work)->work) &base->lock &obj_hash[i].lock irq_context: 0 (wq_completion)bond105 (work_completion)(&(&bond->mcast_work)->work) fill_pool_map-wait-type-override pool_lock irq_context: 0 (wq_completion)bond94 irq_context: 0 (wq_completion)bond94 (work_completion)(&(&slave->notify_work)->work) irq_context: 0 (wq_completion)bond94 (work_completion)(&(&slave->notify_work)->work) &obj_hash[i].lock irq_context: 0 (wq_completion)bond94 (work_completion)(&(&slave->notify_work)->work) &base->lock irq_context: 0 (wq_completion)bond94 (work_completion)(&(&slave->notify_work)->work) &base->lock &obj_hash[i].lock irq_context: 0 (wq_completion)bond94 (work_completion)(&(&slave->notify_work)->work) &rq->__lock irq_context: 0 (wq_completion)bond94 (work_completion)(&(&slave->notify_work)->work) &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)bond94 (work_completion)(&(&bond->mcast_work)->work) irq_context: 0 (wq_completion)bond94 (work_completion)(&(&bond->mcast_work)->work) &obj_hash[i].lock irq_context: 0 (wq_completion)bond94 (work_completion)(&(&bond->mcast_work)->work) &base->lock irq_context: 0 (wq_completion)bond94 (work_completion)(&(&bond->mcast_work)->work) &base->lock &obj_hash[i].lock irq_context: 0 (wq_completion)bond93 (work_completion)(&(&bond->mcast_work)->work) fill_pool_map-wait-type-override pool_lock irq_context: 0 (wq_completion)bond107 irq_context: 0 (wq_completion)bond107 (work_completion)(&(&slave->notify_work)->work) irq_context: 0 (wq_completion)bond107 (work_completion)(&(&slave->notify_work)->work) &obj_hash[i].lock irq_context: 0 (wq_completion)bond107 (work_completion)(&(&slave->notify_work)->work) &base->lock irq_context: 0 (wq_completion)bond107 (work_completion)(&(&slave->notify_work)->work) &base->lock &obj_hash[i].lock irq_context: 0 (wq_completion)bond107 (work_completion)(&(&slave->notify_work)->work) &rq->__lock irq_context: 0 (wq_completion)bond94 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex fs_reclaim irq_context: 0 (wq_completion)bond94 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex fs_reclaim mmu_notifier_invalidate_range_start irq_context: 0 (wq_completion)bond94 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex rcu_read_lock &bond->stats_lock/1 irq_context: 0 (wq_completion)bond94 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex rcu_read_lock &ndev->lock irq_context: 0 (wq_completion)bond94 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex nl_table_lock irq_context: 0 (wq_completion)bond94 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &obj_hash[i].lock irq_context: 0 (wq_completion)bond94 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex nl_table_wait.lock irq_context: 0 (wq_completion)bond94 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex net_rwsem irq_context: 0 (wq_completion)bond94 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex net_rwsem &list->lock#2 irq_context: 0 (wq_completion)bond94 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &tn->lock irq_context: 0 (wq_completion)bond94 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &idev->mc_lock irq_context: 0 (wq_completion)bond94 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &idev->mc_lock fs_reclaim irq_context: 0 (wq_completion)bond94 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &idev->mc_lock fs_reclaim mmu_notifier_invalidate_range_start irq_context: 0 (wq_completion)bond94 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &idev->mc_lock rcu_read_lock &ndev->lock irq_context: 0 (wq_completion)bond94 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &idev->mc_lock rcu_read_lock &dir->lock#2 irq_context: 0 (wq_completion)bond94 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &idev->mc_lock rcu_read_lock &ul->lock#2 irq_context: 0 (wq_completion)bond94 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &idev->mc_lock rcu_read_lock rcu_read_lock &____s->seqcount#10 irq_context: 0 (wq_completion)bond94 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &idev->mc_lock rcu_read_lock rcu_read_lock rcu_read_lock_bh rcu_read_lock rcu_read_lock_bh rcu_read_lock &br->multicast_lock irq_context: 0 (wq_completion)bond94 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &idev->mc_lock rcu_read_lock rcu_read_lock rcu_read_lock_bh rcu_read_lock rcu_read_lock_bh rcu_read_lock &br->multicast_lock &obj_hash[i].lock irq_context: 0 (wq_completion)bond94 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &idev->mc_lock rcu_read_lock rcu_read_lock rcu_read_lock_bh rcu_read_lock rcu_read_lock_bh rcu_read_lock &br->multicast_lock &base->lock irq_context: 0 (wq_completion)bond94 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &idev->mc_lock rcu_read_lock rcu_read_lock rcu_read_lock_bh rcu_read_lock rcu_read_lock_bh rcu_read_lock &data->lock irq_context: 0 (wq_completion)bond94 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &idev->mc_lock rcu_read_lock rcu_read_lock rcu_read_lock_bh rcu_read_lock rcu_read_lock_bh rcu_read_lock &obj_hash[i].lock irq_context: 0 (wq_completion)bond94 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex.wait_lock irq_context: 0 (wq_completion)bond94 (work_completion)(&(&bond->mcast_work)->work) &p->pi_lock irq_context: 0 (wq_completion)bond107 (work_completion)(&(&bond->mcast_work)->work) irq_context: 0 (wq_completion)bond107 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex fs_reclaim irq_context: 0 (wq_completion)bond107 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex fs_reclaim mmu_notifier_invalidate_range_start irq_context: 0 (wq_completion)bond107 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &c->lock irq_context: 0 (wq_completion)bond107 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex rcu_read_lock &bond->stats_lock/1 irq_context: 0 (wq_completion)bond107 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex rcu_read_lock &ndev->lock irq_context: 0 (wq_completion)bond107 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex nl_table_lock irq_context: 0 (wq_completion)bond107 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &obj_hash[i].lock irq_context: 0 (wq_completion)bond107 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex nl_table_wait.lock irq_context: 0 (wq_completion)bond107 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex net_rwsem irq_context: 0 (wq_completion)bond107 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex net_rwsem &list->lock#2 irq_context: 0 (wq_completion)bond107 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &tn->lock irq_context: 0 (wq_completion)bond107 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &idev->mc_lock irq_context: 0 (wq_completion)bond107 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex.wait_lock irq_context: 0 (wq_completion)bond107 (work_completion)(&(&bond->mcast_work)->work) &p->pi_lock irq_context: 0 (wq_completion)bond77 (work_completion)(&(&slave->notify_work)->work) rtnl_mutex net_rwsem irq_context: 0 (wq_completion)bond77 (work_completion)(&(&slave->notify_work)->work) rtnl_mutex net_rwsem &list->lock#2 irq_context: 0 (wq_completion)bond77 (work_completion)(&(&slave->notify_work)->work) rtnl_mutex &tn->lock irq_context: 0 (wq_completion)bond77 (work_completion)(&(&slave->notify_work)->work) rtnl_mutex.wait_lock irq_context: 0 (wq_completion)bond77 (work_completion)(&(&slave->notify_work)->work) &p->pi_lock irq_context: 0 (wq_completion)bond93 (work_completion)(&(&slave->notify_work)->work) rtnl_mutex net_rwsem irq_context: 0 (wq_completion)bond93 (work_completion)(&(&slave->notify_work)->work) rtnl_mutex net_rwsem &list->lock#2 irq_context: 0 (wq_completion)bond93 (work_completion)(&(&slave->notify_work)->work) rtnl_mutex &tn->lock irq_context: 0 (wq_completion)bond93 (work_completion)(&(&slave->notify_work)->work) rtnl_mutex.wait_lock irq_context: 0 (wq_completion)bond93 (work_completion)(&(&slave->notify_work)->work) &p->pi_lock irq_context: 0 (wq_completion)bond106 (work_completion)(&(&slave->notify_work)->work) rtnl_mutex net_rwsem irq_context: 0 (wq_completion)bond106 (work_completion)(&(&slave->notify_work)->work) rtnl_mutex net_rwsem &list->lock#2 irq_context: 0 (wq_completion)bond106 (work_completion)(&(&slave->notify_work)->work) rtnl_mutex &tn->lock irq_context: 0 (wq_completion)bond105 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex fs_reclaim irq_context: 0 (wq_completion)bond105 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex fs_reclaim mmu_notifier_invalidate_range_start irq_context: 0 (wq_completion)bond105 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &c->lock irq_context: 0 (wq_completion)bond105 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &n->list_lock irq_context: 0 (wq_completion)bond105 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &n->list_lock &c->lock irq_context: 0 (wq_completion)bond105 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex rcu_read_lock &bond->stats_lock/1 irq_context: 0 (wq_completion)bond105 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex rcu_read_lock &ndev->lock irq_context: 0 (wq_completion)bond105 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex nl_table_lock irq_context: 0 (wq_completion)bond105 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &obj_hash[i].lock irq_context: 0 (wq_completion)bond105 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex nl_table_wait.lock irq_context: 0 (wq_completion)bond105 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex net_rwsem irq_context: 0 (wq_completion)bond105 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex net_rwsem &list->lock#2 irq_context: 0 (wq_completion)bond105 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &tn->lock irq_context: 0 (wq_completion)bond105 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &idev->mc_lock irq_context: 0 (wq_completion)bond105 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &idev->mc_lock fs_reclaim irq_context: 0 (wq_completion)bond105 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &idev->mc_lock fs_reclaim mmu_notifier_invalidate_range_start irq_context: 0 (wq_completion)bond105 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &idev->mc_lock rcu_read_lock &ndev->lock irq_context: 0 (wq_completion)bond105 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &idev->mc_lock rcu_read_lock &c->lock irq_context: 0 (wq_completion)bond105 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &idev->mc_lock rcu_read_lock &dir->lock#2 irq_context: 0 (wq_completion)bond105 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &idev->mc_lock rcu_read_lock &ul->lock#2 irq_context: 0 (wq_completion)bond105 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &idev->mc_lock rcu_read_lock rcu_read_lock &____s->seqcount#10 irq_context: 0 (wq_completion)bond105 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &idev->mc_lock rcu_read_lock rcu_read_lock rcu_read_lock_bh rcu_read_lock rcu_read_lock_bh rcu_read_lock &br->multicast_lock irq_context: 0 (wq_completion)bond105 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &idev->mc_lock rcu_read_lock rcu_read_lock rcu_read_lock_bh rcu_read_lock rcu_read_lock_bh rcu_read_lock &br->multicast_lock &obj_hash[i].lock irq_context: 0 (wq_completion)bond105 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &idev->mc_lock rcu_read_lock rcu_read_lock rcu_read_lock_bh rcu_read_lock rcu_read_lock_bh rcu_read_lock &br->multicast_lock &base->lock irq_context: 0 (wq_completion)bond105 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &idev->mc_lock rcu_read_lock rcu_read_lock rcu_read_lock_bh rcu_read_lock rcu_read_lock_bh rcu_read_lock &data->lock irq_context: 0 (wq_completion)bond105 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &idev->mc_lock rcu_read_lock rcu_read_lock rcu_read_lock_bh rcu_read_lock rcu_read_lock_bh rcu_read_lock &obj_hash[i].lock irq_context: 0 (wq_completion)bond75 (work_completion)(&(&slave->notify_work)->work) rtnl_mutex net_rwsem irq_context: 0 (wq_completion)bond75 (work_completion)(&(&slave->notify_work)->work) rtnl_mutex net_rwsem &list->lock#2 irq_context: 0 (wq_completion)bond75 (work_completion)(&(&slave->notify_work)->work) rtnl_mutex &tn->lock irq_context: 0 (wq_completion)bond94 (work_completion)(&(&slave->notify_work)->work) rtnl_mutex net_rwsem irq_context: 0 (wq_completion)bond94 (work_completion)(&(&slave->notify_work)->work) rtnl_mutex net_rwsem &list->lock#2 irq_context: 0 (wq_completion)bond94 (work_completion)(&(&slave->notify_work)->work) rtnl_mutex &tn->lock irq_context: 0 (wq_completion)bond77 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex fs_reclaim irq_context: 0 (wq_completion)bond77 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex fs_reclaim mmu_notifier_invalidate_range_start irq_context: 0 (wq_completion)bond77 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &c->lock irq_context: 0 (wq_completion)bond77 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex rcu_read_lock &bond->stats_lock/1 irq_context: 0 (wq_completion)bond77 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex rcu_read_lock &ndev->lock irq_context: 0 (wq_completion)bond77 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex nl_table_lock irq_context: 0 (wq_completion)bond77 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &obj_hash[i].lock irq_context: 0 (wq_completion)bond77 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex nl_table_wait.lock irq_context: 0 (wq_completion)bond77 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex net_rwsem irq_context: 0 (wq_completion)bond77 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex net_rwsem &list->lock#2 irq_context: 0 (wq_completion)bond77 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &tn->lock irq_context: 0 (wq_completion)bond77 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &idev->mc_lock irq_context: 0 (wq_completion)bond77 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &idev->mc_lock fs_reclaim irq_context: 0 (wq_completion)bond77 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &idev->mc_lock fs_reclaim mmu_notifier_invalidate_range_start irq_context: 0 (wq_completion)bond77 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &idev->mc_lock &c->lock irq_context: 0 (wq_completion)bond77 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &idev->mc_lock &____s->seqcount#2 irq_context: 0 (wq_completion)bond77 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &idev->mc_lock &____s->seqcount irq_context: 0 (wq_completion)bond77 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &idev->mc_lock rcu_read_lock &ndev->lock irq_context: 0 (wq_completion)bond77 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &idev->mc_lock rcu_read_lock &c->lock irq_context: 0 (wq_completion)bond77 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &idev->mc_lock rcu_read_lock &dir->lock#2 irq_context: 0 (wq_completion)bond77 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &idev->mc_lock rcu_read_lock &ul->lock#2 irq_context: 0 (wq_completion)bond77 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &idev->mc_lock rcu_read_lock rcu_read_lock &____s->seqcount#10 irq_context: 0 (wq_completion)bond77 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &idev->mc_lock rcu_read_lock rcu_read_lock rcu_read_lock_bh rcu_read_lock rcu_read_lock_bh rcu_read_lock &br->multicast_lock irq_context: 0 (wq_completion)bond77 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &idev->mc_lock rcu_read_lock rcu_read_lock rcu_read_lock_bh rcu_read_lock rcu_read_lock_bh rcu_read_lock &br->multicast_lock &obj_hash[i].lock irq_context: 0 (wq_completion)bond77 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &idev->mc_lock rcu_read_lock rcu_read_lock rcu_read_lock_bh rcu_read_lock rcu_read_lock_bh rcu_read_lock &br->multicast_lock &base->lock irq_context: 0 (wq_completion)bond77 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &idev->mc_lock rcu_read_lock rcu_read_lock rcu_read_lock_bh rcu_read_lock rcu_read_lock_bh rcu_read_lock &data->lock irq_context: 0 (wq_completion)bond77 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &idev->mc_lock rcu_read_lock rcu_read_lock rcu_read_lock_bh rcu_read_lock rcu_read_lock_bh rcu_read_lock &obj_hash[i].lock irq_context: 0 (wq_completion)bond93 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex fs_reclaim irq_context: 0 (wq_completion)bond93 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex fs_reclaim mmu_notifier_invalidate_range_start irq_context: 0 (wq_completion)bond93 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &c->lock irq_context: 0 (wq_completion)bond93 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex rcu_read_lock &bond->stats_lock/1 irq_context: 0 (wq_completion)bond93 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex rcu_read_lock &ndev->lock irq_context: 0 (wq_completion)bond93 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex nl_table_lock irq_context: 0 (wq_completion)bond93 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &obj_hash[i].lock irq_context: 0 (wq_completion)bond93 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex nl_table_wait.lock irq_context: 0 (wq_completion)bond93 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex net_rwsem irq_context: 0 (wq_completion)bond93 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex net_rwsem &list->lock#2 irq_context: 0 (wq_completion)bond93 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &tn->lock irq_context: 0 (wq_completion)bond93 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &idev->mc_lock irq_context: 0 (wq_completion)bond93 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &idev->mc_lock fs_reclaim irq_context: 0 (wq_completion)bond93 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &idev->mc_lock fs_reclaim mmu_notifier_invalidate_range_start irq_context: 0 (wq_completion)bond93 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &idev->mc_lock rcu_read_lock &ndev->lock irq_context: 0 (wq_completion)bond93 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &idev->mc_lock rcu_read_lock &c->lock irq_context: 0 (wq_completion)bond93 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &idev->mc_lock rcu_read_lock &____s->seqcount#2 irq_context: 0 (wq_completion)bond93 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &idev->mc_lock rcu_read_lock &____s->seqcount irq_context: 0 (wq_completion)bond93 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &idev->mc_lock rcu_read_lock pool_lock#2 irq_context: 0 (wq_completion)bond93 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &idev->mc_lock rcu_read_lock &dir->lock#2 irq_context: 0 (wq_completion)bond93 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &idev->mc_lock rcu_read_lock &ul->lock#2 irq_context: 0 (wq_completion)bond93 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &idev->mc_lock rcu_read_lock rcu_read_lock &____s->seqcount#10 irq_context: 0 (wq_completion)bond93 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &idev->mc_lock rcu_read_lock rcu_read_lock rcu_read_lock_bh rcu_read_lock rcu_read_lock_bh rcu_read_lock &br->multicast_lock irq_context: 0 (wq_completion)bond93 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &idev->mc_lock rcu_read_lock rcu_read_lock rcu_read_lock_bh rcu_read_lock rcu_read_lock_bh rcu_read_lock &br->multicast_lock &obj_hash[i].lock irq_context: 0 (wq_completion)bond93 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &idev->mc_lock rcu_read_lock rcu_read_lock rcu_read_lock_bh rcu_read_lock rcu_read_lock_bh rcu_read_lock &br->multicast_lock &base->lock irq_context: 0 (wq_completion)bond93 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &idev->mc_lock rcu_read_lock rcu_read_lock rcu_read_lock_bh rcu_read_lock rcu_read_lock_bh rcu_read_lock &data->lock irq_context: 0 (wq_completion)bond93 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &idev->mc_lock rcu_read_lock rcu_read_lock rcu_read_lock_bh rcu_read_lock rcu_read_lock_bh rcu_read_lock &obj_hash[i].lock irq_context: 0 (wq_completion)bond76 (work_completion)(&(&slave->notify_work)->work) rtnl_mutex net_rwsem irq_context: 0 (wq_completion)bond76 (work_completion)(&(&slave->notify_work)->work) rtnl_mutex net_rwsem &list->lock#2 irq_context: 0 (wq_completion)bond76 (work_completion)(&(&slave->notify_work)->work) rtnl_mutex &tn->lock irq_context: 0 (wq_completion)bond107 (work_completion)(&(&slave->notify_work)->work) rtnl_mutex net_rwsem irq_context: 0 (wq_completion)bond107 (work_completion)(&(&slave->notify_work)->work) rtnl_mutex net_rwsem &list->lock#2 irq_context: 0 (wq_completion)bond107 (work_completion)(&(&slave->notify_work)->work) rtnl_mutex &tn->lock irq_context: 0 (wq_completion)bond76 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex fs_reclaim irq_context: 0 (wq_completion)bond76 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex fs_reclaim mmu_notifier_invalidate_range_start irq_context: 0 (wq_completion)bond76 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &c->lock irq_context: 0 (wq_completion)bond76 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &n->list_lock irq_context: 0 (wq_completion)bond76 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &n->list_lock &c->lock irq_context: 0 (wq_completion)bond76 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex rcu_read_lock &bond->stats_lock/1 irq_context: 0 (wq_completion)bond76 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex rcu_read_lock &ndev->lock irq_context: 0 (wq_completion)bond76 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex nl_table_lock irq_context: 0 (wq_completion)bond76 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &obj_hash[i].lock irq_context: 0 (wq_completion)bond76 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex nl_table_wait.lock irq_context: 0 (wq_completion)bond76 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex net_rwsem irq_context: 0 (wq_completion)bond76 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex net_rwsem &list->lock#2 irq_context: 0 (wq_completion)bond76 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &tn->lock irq_context: 0 (wq_completion)bond76 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &idev->mc_lock irq_context: 0 (wq_completion)bond76 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &idev->mc_lock fs_reclaim irq_context: 0 (wq_completion)bond76 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &idev->mc_lock fs_reclaim mmu_notifier_invalidate_range_start irq_context: 0 (wq_completion)bond76 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &idev->mc_lock fs_reclaim mmu_notifier_invalidate_range_start &rq->__lock irq_context: 0 (wq_completion)bond76 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &idev->mc_lock fs_reclaim mmu_notifier_invalidate_range_start &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)bond76 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &idev->mc_lock rcu_read_lock &ndev->lock irq_context: 0 (wq_completion)bond76 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &idev->mc_lock rcu_read_lock &dir->lock#2 irq_context: 0 (wq_completion)bond76 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &idev->mc_lock rcu_read_lock &ul->lock#2 irq_context: 0 (wq_completion)bond76 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &idev->mc_lock rcu_read_lock rcu_read_lock &____s->seqcount#10 irq_context: 0 (wq_completion)bond76 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &idev->mc_lock rcu_read_lock rcu_read_lock rcu_read_lock_bh rcu_read_lock rcu_read_lock_bh rcu_read_lock &br->multicast_lock irq_context: 0 (wq_completion)bond76 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &idev->mc_lock rcu_read_lock rcu_read_lock rcu_read_lock_bh rcu_read_lock rcu_read_lock_bh rcu_read_lock &br->multicast_lock &obj_hash[i].lock irq_context: 0 (wq_completion)bond76 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &idev->mc_lock rcu_read_lock rcu_read_lock rcu_read_lock_bh rcu_read_lock rcu_read_lock_bh rcu_read_lock &br->multicast_lock &base->lock irq_context: 0 (wq_completion)bond76 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &idev->mc_lock rcu_read_lock rcu_read_lock rcu_read_lock_bh rcu_read_lock rcu_read_lock_bh rcu_read_lock &data->lock irq_context: 0 (wq_completion)bond76 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &idev->mc_lock rcu_read_lock rcu_read_lock rcu_read_lock_bh rcu_read_lock rcu_read_lock_bh rcu_read_lock &obj_hash[i].lock irq_context: 0 (wq_completion)bond76 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &lock->wait_lock irq_context: 0 (wq_completion)bond76 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &p->pi_lock irq_context: 0 (wq_completion)bond76 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &p->pi_lock &rq->__lock irq_context: 0 (wq_completion)bond76 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &p->pi_lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)bond76 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &rq->__lock irq_context: 0 (wq_completion)bond76 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)bond76 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex.wait_lock irq_context: 0 (wq_completion)bond76 (work_completion)(&(&bond->mcast_work)->work) &p->pi_lock irq_context: 0 (wq_completion)bond76 (work_completion)(&(&bond->mcast_work)->work) &p->pi_lock &rq->__lock irq_context: 0 (wq_completion)bond76 (work_completion)(&(&bond->mcast_work)->work) &p->pi_lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)bond76 (work_completion)(&(&bond->mcast_work)->work) &rq->__lock irq_context: 0 (wq_completion)bond76 (work_completion)(&(&bond->mcast_work)->work) &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)bond78 irq_context: 0 (wq_completion)bond78 (work_completion)(&(&slave->notify_work)->work) irq_context: 0 (wq_completion)bond78 (work_completion)(&(&slave->notify_work)->work) &obj_hash[i].lock irq_context: 0 (wq_completion)bond78 (work_completion)(&(&slave->notify_work)->work) &base->lock irq_context: 0 (wq_completion)bond78 (work_completion)(&(&slave->notify_work)->work) &base->lock &obj_hash[i].lock irq_context: 0 (wq_completion)bond78 (work_completion)(&(&slave->notify_work)->work) &rq->__lock irq_context: 0 (wq_completion)bond78 (work_completion)(&(&slave->notify_work)->work) &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)bond78 (work_completion)(&(&bond->mcast_work)->work) irq_context: 0 (wq_completion)bond78 (work_completion)(&(&bond->mcast_work)->work) &obj_hash[i].lock irq_context: 0 (wq_completion)bond78 (work_completion)(&(&bond->mcast_work)->work) &base->lock irq_context: 0 (wq_completion)bond78 (work_completion)(&(&bond->mcast_work)->work) &base->lock &obj_hash[i].lock irq_context: 0 (wq_completion)bond95 irq_context: 0 (wq_completion)bond95 &rq->__lock irq_context: 0 (wq_completion)bond95 &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)bond95 (work_completion)(&(&slave->notify_work)->work) irq_context: 0 (wq_completion)bond95 (work_completion)(&(&slave->notify_work)->work) &obj_hash[i].lock irq_context: 0 (wq_completion)bond95 (work_completion)(&(&slave->notify_work)->work) &base->lock irq_context: 0 (wq_completion)bond95 (work_completion)(&(&slave->notify_work)->work) &base->lock &obj_hash[i].lock irq_context: 0 (wq_completion)bond95 (work_completion)(&(&bond->mcast_work)->work) irq_context: 0 (wq_completion)bond95 (work_completion)(&(&bond->mcast_work)->work) &obj_hash[i].lock irq_context: 0 (wq_completion)bond95 (work_completion)(&(&bond->mcast_work)->work) &base->lock irq_context: 0 (wq_completion)bond95 (work_completion)(&(&bond->mcast_work)->work) &base->lock &obj_hash[i].lock irq_context: 0 (wq_completion)bond95 (work_completion)(&(&bond->mcast_work)->work) &rq->__lock irq_context: 0 (wq_completion)bond95 (work_completion)(&(&bond->mcast_work)->work) &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)bond108 irq_context: 0 (wq_completion)bond108 (work_completion)(&(&slave->notify_work)->work) irq_context: 0 (wq_completion)bond108 (work_completion)(&(&slave->notify_work)->work) &obj_hash[i].lock irq_context: 0 (wq_completion)bond108 (work_completion)(&(&slave->notify_work)->work) &base->lock irq_context: 0 (wq_completion)bond108 (work_completion)(&(&slave->notify_work)->work) &base->lock &obj_hash[i].lock irq_context: 0 (wq_completion)bond108 (work_completion)(&(&bond->mcast_work)->work) irq_context: 0 (wq_completion)bond108 (work_completion)(&(&bond->mcast_work)->work) &obj_hash[i].lock irq_context: 0 (wq_completion)bond108 (work_completion)(&(&bond->mcast_work)->work) &base->lock irq_context: 0 (wq_completion)bond108 (work_completion)(&(&bond->mcast_work)->work) &base->lock &obj_hash[i].lock irq_context: 0 (wq_completion)bond108 (work_completion)(&(&bond->mcast_work)->work) &rq->__lock irq_context: 0 (wq_completion)bond108 (work_completion)(&(&bond->mcast_work)->work) &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)bond79#2 irq_context: 0 (wq_completion)bond79#2 (work_completion)(&(&slave->notify_work)->work) irq_context: 0 (wq_completion)bond79#2 (work_completion)(&(&slave->notify_work)->work) &obj_hash[i].lock irq_context: 0 (wq_completion)bond79#2 (work_completion)(&(&slave->notify_work)->work) &base->lock irq_context: 0 (wq_completion)bond79#2 (work_completion)(&(&slave->notify_work)->work) &base->lock &obj_hash[i].lock irq_context: 0 (wq_completion)bond79#2 (work_completion)(&(&slave->notify_work)->work) &rq->__lock irq_context: 0 (wq_completion)bond79#2 (work_completion)(&(&slave->notify_work)->work) &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)bond79#2 (work_completion)(&(&bond->mcast_work)->work) irq_context: 0 (wq_completion)bond79#2 (work_completion)(&(&bond->mcast_work)->work) &rq->__lock irq_context: 0 (wq_completion)bond79#2 (work_completion)(&(&bond->mcast_work)->work) &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)bond79#2 (work_completion)(&(&bond->mcast_work)->work) &obj_hash[i].lock irq_context: 0 (wq_completion)bond79#2 (work_completion)(&(&bond->mcast_work)->work) &base->lock irq_context: 0 (wq_completion)bond79#2 (work_completion)(&(&bond->mcast_work)->work) &base->lock &obj_hash[i].lock irq_context: 0 (wq_completion)bond96#2 irq_context: 0 (wq_completion)bond96#2 (work_completion)(&(&slave->notify_work)->work) irq_context: 0 (wq_completion)bond96#2 (work_completion)(&(&slave->notify_work)->work) &obj_hash[i].lock irq_context: 0 (wq_completion)bond96#2 (work_completion)(&(&slave->notify_work)->work) &base->lock irq_context: 0 (wq_completion)bond96#2 (work_completion)(&(&slave->notify_work)->work) &base->lock &obj_hash[i].lock irq_context: 0 (wq_completion)bond96#2 (work_completion)(&(&slave->notify_work)->work) &rq->__lock irq_context: 0 (wq_completion)bond96#2 (work_completion)(&(&slave->notify_work)->work) &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)bond96#2 (work_completion)(&(&bond->mcast_work)->work) irq_context: 0 (wq_completion)bond96#2 (work_completion)(&(&bond->mcast_work)->work) &obj_hash[i].lock irq_context: 0 (wq_completion)bond96#2 (work_completion)(&(&bond->mcast_work)->work) &base->lock irq_context: 0 (wq_completion)bond96#2 (work_completion)(&(&bond->mcast_work)->work) &base->lock &obj_hash[i].lock irq_context: 0 (wq_completion)bond96#2 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex fs_reclaim irq_context: 0 (wq_completion)bond96#2 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex fs_reclaim mmu_notifier_invalidate_range_start irq_context: 0 (wq_completion)bond96#2 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &c->lock irq_context: 0 (wq_completion)bond96#2 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &n->list_lock irq_context: 0 (wq_completion)bond96#2 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &n->list_lock &c->lock irq_context: 0 (wq_completion)bond96#2 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex rcu_read_lock &bond->stats_lock/1 irq_context: 0 (wq_completion)bond96#2 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex rcu_read_lock &ndev->lock irq_context: 0 (wq_completion)bond96#2 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex nl_table_lock irq_context: 0 (wq_completion)bond96#2 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &obj_hash[i].lock irq_context: 0 (wq_completion)bond96#2 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex nl_table_wait.lock irq_context: 0 (wq_completion)bond96#2 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex net_rwsem irq_context: 0 (wq_completion)bond96#2 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex net_rwsem &list->lock#2 irq_context: 0 (wq_completion)bond96#2 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &tn->lock irq_context: 0 (wq_completion)bond96#2 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &idev->mc_lock irq_context: 0 (wq_completion)bond96#2 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex.wait_lock irq_context: 0 (wq_completion)bond96#2 (work_completion)(&(&bond->mcast_work)->work) &p->pi_lock irq_context: 0 (wq_completion)bond96#2 (work_completion)(&(&slave->notify_work)->work) rtnl_mutex net_rwsem irq_context: 0 (wq_completion)bond96#2 (work_completion)(&(&slave->notify_work)->work) rtnl_mutex net_rwsem &list->lock#2 irq_context: 0 (wq_completion)bond96#2 (work_completion)(&(&slave->notify_work)->work) rtnl_mutex &tn->lock irq_context: 0 (wq_completion)bond96#2 (work_completion)(&(&slave->notify_work)->work) rtnl_mutex.wait_lock irq_context: 0 (wq_completion)bond96#2 (work_completion)(&(&slave->notify_work)->work) &p->pi_lock irq_context: 0 (wq_completion)bond109 irq_context: 0 (wq_completion)bond109 (work_completion)(&(&slave->notify_work)->work) irq_context: 0 (wq_completion)bond109 (work_completion)(&(&slave->notify_work)->work) &obj_hash[i].lock irq_context: 0 (wq_completion)bond109 (work_completion)(&(&slave->notify_work)->work) &base->lock irq_context: 0 (wq_completion)bond109 (work_completion)(&(&slave->notify_work)->work) &base->lock &obj_hash[i].lock irq_context: 0 (wq_completion)bond109 (work_completion)(&(&slave->notify_work)->work) &rq->__lock irq_context: 0 (wq_completion)bond109 (work_completion)(&(&slave->notify_work)->work) &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)bond109 (work_completion)(&(&bond->mcast_work)->work) irq_context: 0 (wq_completion)bond109 (work_completion)(&(&bond->mcast_work)->work) &obj_hash[i].lock irq_context: 0 (wq_completion)bond109 (work_completion)(&(&bond->mcast_work)->work) &base->lock irq_context: 0 (wq_completion)bond109 (work_completion)(&(&bond->mcast_work)->work) &base->lock &obj_hash[i].lock irq_context: 0 (wq_completion)bond80#2 irq_context: 0 (wq_completion)bond80#2 (work_completion)(&(&slave->notify_work)->work) irq_context: 0 (wq_completion)bond80#2 (work_completion)(&(&slave->notify_work)->work) &obj_hash[i].lock irq_context: 0 (wq_completion)bond80#2 (work_completion)(&(&slave->notify_work)->work) &base->lock irq_context: 0 (wq_completion)bond80#2 (work_completion)(&(&slave->notify_work)->work) &base->lock &obj_hash[i].lock irq_context: 0 (wq_completion)bond80#2 (work_completion)(&(&slave->notify_work)->work) &rq->__lock irq_context: 0 (wq_completion)bond80#2 (work_completion)(&(&slave->notify_work)->work) &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)bond80#2 (work_completion)(&(&bond->mcast_work)->work) irq_context: 0 (wq_completion)bond80#2 (work_completion)(&(&bond->mcast_work)->work) &obj_hash[i].lock irq_context: 0 (wq_completion)bond80#2 (work_completion)(&(&bond->mcast_work)->work) &base->lock irq_context: 0 (wq_completion)bond80#2 (work_completion)(&(&bond->mcast_work)->work) &base->lock &obj_hash[i].lock irq_context: 0 (wq_completion)bond97#2 irq_context: 0 (wq_completion)bond97#2 (work_completion)(&(&slave->notify_work)->work) irq_context: 0 (wq_completion)bond97#2 (work_completion)(&(&slave->notify_work)->work) &obj_hash[i].lock irq_context: 0 (wq_completion)bond97#2 (work_completion)(&(&slave->notify_work)->work) &base->lock irq_context: 0 (wq_completion)bond97#2 (work_completion)(&(&slave->notify_work)->work) &base->lock &obj_hash[i].lock irq_context: 0 (wq_completion)bond97#2 (work_completion)(&(&bond->mcast_work)->work) irq_context: 0 (wq_completion)bond97#2 (work_completion)(&(&bond->mcast_work)->work) &obj_hash[i].lock irq_context: 0 (wq_completion)bond97#2 (work_completion)(&(&bond->mcast_work)->work) &base->lock irq_context: 0 (wq_completion)bond97#2 (work_completion)(&(&bond->mcast_work)->work) &base->lock &obj_hash[i].lock irq_context: 0 (wq_completion)bond108 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex fs_reclaim irq_context: 0 (wq_completion)bond108 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex fs_reclaim mmu_notifier_invalidate_range_start irq_context: 0 (wq_completion)bond108 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &c->lock irq_context: 0 (wq_completion)bond108 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &n->list_lock irq_context: 0 (wq_completion)bond108 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &n->list_lock &c->lock irq_context: 0 (wq_completion)bond108 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &rq->__lock irq_context: 0 (wq_completion)bond108 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)bond108 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex rcu_read_lock &bond->stats_lock/1 irq_context: 0 (wq_completion)bond108 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex rcu_read_lock &ndev->lock irq_context: 0 (wq_completion)bond108 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex nl_table_lock irq_context: 0 (wq_completion)bond108 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &obj_hash[i].lock irq_context: 0 (wq_completion)bond108 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex nl_table_wait.lock irq_context: 0 (wq_completion)bond108 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex net_rwsem irq_context: 0 (wq_completion)bond108 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex net_rwsem &list->lock#2 irq_context: 0 (wq_completion)bond108 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &tn->lock irq_context: 0 (wq_completion)bond108 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &idev->mc_lock irq_context: 0 (wq_completion)bond108 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &idev->mc_lock fs_reclaim irq_context: 0 (wq_completion)bond108 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &idev->mc_lock fs_reclaim mmu_notifier_invalidate_range_start irq_context: 0 (wq_completion)bond108 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &idev->mc_lock &c->lock irq_context: 0 (wq_completion)bond108 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &idev->mc_lock rcu_read_lock &ndev->lock irq_context: 0 (wq_completion)bond108 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &idev->mc_lock rcu_read_lock rcu_node_0 irq_context: 0 (wq_completion)bond108 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &idev->mc_lock rcu_read_lock &rq->__lock irq_context: 0 (wq_completion)bond108 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &idev->mc_lock rcu_read_lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)bond108 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &idev->mc_lock rcu_read_lock &dir->lock#2 irq_context: 0 (wq_completion)bond108 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &idev->mc_lock rcu_read_lock &ul->lock#2 irq_context: 0 (wq_completion)bond108 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &idev->mc_lock rcu_read_lock rcu_read_lock &____s->seqcount#10 irq_context: 0 (wq_completion)bond108 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &idev->mc_lock rcu_read_lock rcu_read_lock rcu_read_lock_bh rcu_read_lock rcu_read_lock_bh rcu_read_lock &br->multicast_lock irq_context: 0 (wq_completion)bond108 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &idev->mc_lock rcu_read_lock rcu_read_lock rcu_read_lock_bh rcu_read_lock rcu_read_lock_bh rcu_read_lock &br->multicast_lock &obj_hash[i].lock irq_context: 0 (wq_completion)bond108 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &idev->mc_lock rcu_read_lock rcu_read_lock rcu_read_lock_bh rcu_read_lock rcu_read_lock_bh rcu_read_lock &br->multicast_lock &base->lock irq_context: 0 (wq_completion)bond108 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &idev->mc_lock rcu_read_lock rcu_read_lock rcu_read_lock_bh rcu_read_lock rcu_read_lock_bh rcu_read_lock &data->lock irq_context: 0 (wq_completion)bond108 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &idev->mc_lock rcu_read_lock rcu_read_lock rcu_read_lock_bh rcu_read_lock rcu_read_lock_bh rcu_read_lock &obj_hash[i].lock irq_context: 0 (wq_completion)bond108 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex.wait_lock irq_context: 0 (wq_completion)bond108 (work_completion)(&(&bond->mcast_work)->work) &p->pi_lock irq_context: 0 (wq_completion)bond108 (work_completion)(&(&slave->notify_work)->work) rtnl_mutex net_rwsem irq_context: 0 (wq_completion)bond108 (work_completion)(&(&slave->notify_work)->work) rtnl_mutex net_rwsem &list->lock#2 irq_context: 0 (wq_completion)bond108 (work_completion)(&(&slave->notify_work)->work) rtnl_mutex &tn->lock irq_context: 0 (wq_completion)bond108 (work_completion)(&(&slave->notify_work)->work) rtnl_mutex.wait_lock irq_context: 0 (wq_completion)bond108 (work_completion)(&(&slave->notify_work)->work) &p->pi_lock irq_context: 0 (wq_completion)bond110 irq_context: 0 (wq_completion)bond110 (work_completion)(&(&slave->notify_work)->work) irq_context: 0 (wq_completion)bond110 (work_completion)(&(&slave->notify_work)->work) &obj_hash[i].lock irq_context: 0 (wq_completion)bond110 (work_completion)(&(&slave->notify_work)->work) &base->lock irq_context: 0 (wq_completion)bond110 (work_completion)(&(&slave->notify_work)->work) &base->lock &obj_hash[i].lock irq_context: 0 (wq_completion)bond110 (work_completion)(&(&bond->mcast_work)->work) irq_context: 0 (wq_completion)bond110 (work_completion)(&(&bond->mcast_work)->work) &obj_hash[i].lock irq_context: 0 (wq_completion)bond110 (work_completion)(&(&bond->mcast_work)->work) &base->lock irq_context: 0 (wq_completion)bond110 (work_completion)(&(&bond->mcast_work)->work) &base->lock &obj_hash[i].lock irq_context: 0 (wq_completion)bond78 &rq->__lock irq_context: 0 rcu_read_lock key#28 irq_context: softirq rcu_callback key#28 irq_context: 0 (wq_completion)bond81#2 irq_context: 0 (wq_completion)bond81#2 (work_completion)(&(&slave->notify_work)->work) irq_context: 0 (wq_completion)bond81#2 (work_completion)(&(&slave->notify_work)->work) &obj_hash[i].lock irq_context: 0 (wq_completion)bond81#2 (work_completion)(&(&slave->notify_work)->work) &base->lock irq_context: 0 (wq_completion)bond81#2 (work_completion)(&(&slave->notify_work)->work) &base->lock &obj_hash[i].lock irq_context: 0 (wq_completion)bond81#2 (work_completion)(&(&slave->notify_work)->work) &rq->__lock irq_context: 0 (wq_completion)bond81#2 (work_completion)(&(&slave->notify_work)->work) &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)bond109 (work_completion)(&(&bond->mcast_work)->work) &rq->__lock irq_context: 0 (wq_completion)bond109 (work_completion)(&(&bond->mcast_work)->work) &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)bond81#2 (work_completion)(&(&bond->mcast_work)->work) irq_context: 0 (wq_completion)bond81#2 (work_completion)(&(&bond->mcast_work)->work) &obj_hash[i].lock irq_context: 0 (wq_completion)bond81#2 (work_completion)(&(&bond->mcast_work)->work) &base->lock irq_context: 0 (wq_completion)bond81#2 (work_completion)(&(&bond->mcast_work)->work) &base->lock &obj_hash[i].lock irq_context: 0 (wq_completion)bond81#2 (work_completion)(&(&bond->mcast_work)->work) &rq->__lock irq_context: 0 (wq_completion)bond81#2 (work_completion)(&(&bond->mcast_work)->work) &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)bond97#2 (work_completion)(&(&slave->notify_work)->work) &rq->__lock irq_context: 0 (wq_completion)bond97#2 (work_completion)(&(&slave->notify_work)->work) &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)bond98#2 irq_context: 0 (wq_completion)bond98#2 &rq->__lock irq_context: 0 (wq_completion)bond98#2 &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)bond98#2 (work_completion)(&(&slave->notify_work)->work) irq_context: 0 (wq_completion)bond98#2 (work_completion)(&(&slave->notify_work)->work) &obj_hash[i].lock irq_context: 0 (wq_completion)bond98#2 (work_completion)(&(&slave->notify_work)->work) &base->lock irq_context: 0 (wq_completion)bond98#2 (work_completion)(&(&slave->notify_work)->work) &base->lock &obj_hash[i].lock irq_context: 0 (wq_completion)bond98#2 (work_completion)(&(&slave->notify_work)->work) &rq->__lock irq_context: 0 (wq_completion)bond98#2 (work_completion)(&(&slave->notify_work)->work) &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)bond98#2 (work_completion)(&(&bond->mcast_work)->work) irq_context: 0 (wq_completion)bond98#2 (work_completion)(&(&bond->mcast_work)->work) &rq->__lock irq_context: 0 (wq_completion)bond98#2 (work_completion)(&(&bond->mcast_work)->work) &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)bond98#2 (work_completion)(&(&bond->mcast_work)->work) &obj_hash[i].lock irq_context: 0 (wq_completion)bond98#2 (work_completion)(&(&bond->mcast_work)->work) &base->lock irq_context: 0 (wq_completion)bond98#2 (work_completion)(&(&bond->mcast_work)->work) &base->lock &obj_hash[i].lock irq_context: 0 (wq_completion)bond110 (work_completion)(&(&slave->notify_work)->work) &rq->__lock irq_context: 0 (wq_completion)bond110 (work_completion)(&(&slave->notify_work)->work) &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)bond111 irq_context: 0 (wq_completion)bond111 &rq->__lock irq_context: 0 (wq_completion)bond111 &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)bond111 (work_completion)(&(&slave->notify_work)->work) irq_context: 0 (wq_completion)bond111 (work_completion)(&(&slave->notify_work)->work) &obj_hash[i].lock irq_context: 0 (wq_completion)bond111 (work_completion)(&(&slave->notify_work)->work) &base->lock irq_context: 0 (wq_completion)bond111 (work_completion)(&(&slave->notify_work)->work) &base->lock &obj_hash[i].lock irq_context: 0 (wq_completion)bond111 (work_completion)(&(&bond->mcast_work)->work) irq_context: 0 (wq_completion)bond111 (work_completion)(&(&bond->mcast_work)->work) &obj_hash[i].lock irq_context: 0 (wq_completion)bond111 (work_completion)(&(&bond->mcast_work)->work) &base->lock irq_context: 0 (wq_completion)bond111 (work_completion)(&(&bond->mcast_work)->work) &base->lock &obj_hash[i].lock irq_context: 0 (wq_completion)bond111 (work_completion)(&(&bond->mcast_work)->work) &rq->__lock irq_context: 0 (wq_completion)bond111 (work_completion)(&(&bond->mcast_work)->work) &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 rtnl_mutex &idev->mc_lock &pcp->lock &zone->lock &____s->seqcount irq_context: 0 (wq_completion)bond82 irq_context: 0 (wq_completion)bond82 (work_completion)(&(&slave->notify_work)->work) irq_context: 0 (wq_completion)bond82 (work_completion)(&(&slave->notify_work)->work) &rq->__lock irq_context: 0 (wq_completion)bond82 (work_completion)(&(&slave->notify_work)->work) &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)bond82 (work_completion)(&(&slave->notify_work)->work) &obj_hash[i].lock irq_context: 0 (wq_completion)bond82 (work_completion)(&(&slave->notify_work)->work) &base->lock irq_context: 0 (wq_completion)bond82 (work_completion)(&(&slave->notify_work)->work) &base->lock &obj_hash[i].lock irq_context: 0 (wq_completion)bond82 (work_completion)(&(&bond->mcast_work)->work) irq_context: 0 (wq_completion)bond82 (work_completion)(&(&bond->mcast_work)->work) &obj_hash[i].lock irq_context: 0 (wq_completion)bond82 (work_completion)(&(&bond->mcast_work)->work) &base->lock irq_context: 0 (wq_completion)bond82 (work_completion)(&(&bond->mcast_work)->work) &base->lock &obj_hash[i].lock irq_context: 0 (wq_completion)bond99#2 irq_context: 0 (wq_completion)bond99#2 (work_completion)(&(&slave->notify_work)->work) irq_context: 0 (wq_completion)bond99#2 (work_completion)(&(&slave->notify_work)->work) &obj_hash[i].lock irq_context: 0 (wq_completion)bond99#2 (work_completion)(&(&slave->notify_work)->work) &base->lock irq_context: 0 (wq_completion)bond99#2 (work_completion)(&(&slave->notify_work)->work) &base->lock &obj_hash[i].lock irq_context: 0 (wq_completion)bond99#2 (work_completion)(&(&slave->notify_work)->work) &rq->__lock irq_context: 0 (wq_completion)bond110 (work_completion)(&(&slave->notify_work)->work) rtnl_mutex net_rwsem irq_context: 0 (wq_completion)bond110 (work_completion)(&(&slave->notify_work)->work) rtnl_mutex net_rwsem &list->lock#2 irq_context: 0 (wq_completion)bond110 (work_completion)(&(&slave->notify_work)->work) rtnl_mutex &tn->lock irq_context: 0 (wq_completion)bond110 (work_completion)(&(&slave->notify_work)->work) rtnl_mutex.wait_lock irq_context: 0 (wq_completion)bond110 (work_completion)(&(&slave->notify_work)->work) &p->pi_lock irq_context: 0 (wq_completion)bond99#2 (work_completion)(&(&bond->mcast_work)->work) irq_context: 0 (wq_completion)bond99#2 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex fs_reclaim irq_context: 0 (wq_completion)bond99#2 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex fs_reclaim mmu_notifier_invalidate_range_start irq_context: 0 (wq_completion)bond99#2 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex rcu_read_lock &bond->stats_lock/1 irq_context: 0 (wq_completion)bond99#2 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex rcu_read_lock &ndev->lock irq_context: 0 (wq_completion)bond99#2 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex nl_table_lock irq_context: 0 (wq_completion)bond99#2 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &obj_hash[i].lock irq_context: 0 (wq_completion)bond99#2 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex nl_table_wait.lock irq_context: 0 (wq_completion)bond99#2 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex net_rwsem irq_context: 0 (wq_completion)bond99#2 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex net_rwsem &list->lock#2 irq_context: 0 (wq_completion)bond99#2 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &tn->lock irq_context: 0 (wq_completion)bond99#2 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &idev->mc_lock irq_context: 0 (wq_completion)bond99#2 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex.wait_lock irq_context: 0 (wq_completion)bond99#2 (work_completion)(&(&bond->mcast_work)->work) &p->pi_lock irq_context: 0 (wq_completion)bond97#2 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex fs_reclaim irq_context: 0 (wq_completion)bond97#2 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex fs_reclaim mmu_notifier_invalidate_range_start irq_context: 0 (wq_completion)bond97#2 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &c->lock irq_context: 0 (wq_completion)bond97#2 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex rcu_read_lock &bond->stats_lock/1 irq_context: 0 (wq_completion)bond97#2 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex rcu_read_lock &ndev->lock irq_context: 0 (wq_completion)bond97#2 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex nl_table_lock irq_context: 0 (wq_completion)bond97#2 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &obj_hash[i].lock irq_context: 0 (wq_completion)bond97#2 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex nl_table_wait.lock irq_context: 0 (wq_completion)bond97#2 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex net_rwsem irq_context: 0 (wq_completion)bond97#2 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex net_rwsem &list->lock#2 irq_context: 0 (wq_completion)bond97#2 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &tn->lock irq_context: 0 (wq_completion)bond97#2 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &idev->mc_lock irq_context: 0 (wq_completion)bond97#2 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &idev->mc_lock fs_reclaim irq_context: 0 (wq_completion)bond97#2 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &idev->mc_lock fs_reclaim mmu_notifier_invalidate_range_start irq_context: 0 (wq_completion)bond97#2 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &idev->mc_lock &c->lock irq_context: 0 (wq_completion)bond97#2 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &idev->mc_lock rcu_read_lock &ndev->lock irq_context: 0 (wq_completion)bond97#2 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &idev->mc_lock rcu_read_lock &dir->lock#2 irq_context: 0 (wq_completion)bond97#2 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &idev->mc_lock rcu_read_lock &ul->lock#2 irq_context: 0 (wq_completion)bond97#2 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &idev->mc_lock rcu_read_lock rcu_read_lock &____s->seqcount#10 irq_context: 0 (wq_completion)bond97#2 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &idev->mc_lock rcu_read_lock rcu_read_lock rcu_read_lock_bh rcu_read_lock rcu_read_lock_bh rcu_read_lock &br->multicast_lock irq_context: 0 (wq_completion)bond97#2 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &idev->mc_lock rcu_read_lock rcu_read_lock rcu_read_lock_bh rcu_read_lock rcu_read_lock_bh rcu_read_lock &br->multicast_lock &obj_hash[i].lock irq_context: 0 (wq_completion)bond97#2 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &idev->mc_lock rcu_read_lock rcu_read_lock rcu_read_lock_bh rcu_read_lock rcu_read_lock_bh rcu_read_lock &br->multicast_lock &base->lock irq_context: 0 (wq_completion)bond97#2 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &idev->mc_lock rcu_read_lock rcu_read_lock rcu_read_lock_bh rcu_read_lock rcu_read_lock_bh rcu_read_lock &data->lock irq_context: 0 (wq_completion)bond97#2 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &idev->mc_lock rcu_read_lock rcu_read_lock rcu_read_lock_bh rcu_read_lock rcu_read_lock_bh rcu_read_lock &obj_hash[i].lock irq_context: 0 (wq_completion)bond97#2 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex.wait_lock irq_context: 0 (wq_completion)bond97#2 (work_completion)(&(&bond->mcast_work)->work) &p->pi_lock irq_context: 0 (wq_completion)bond82 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex fs_reclaim irq_context: 0 (wq_completion)bond82 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex fs_reclaim mmu_notifier_invalidate_range_start irq_context: 0 (wq_completion)bond82 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &c->lock irq_context: 0 (wq_completion)bond82 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &n->list_lock irq_context: 0 (wq_completion)bond82 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &n->list_lock &c->lock irq_context: 0 (wq_completion)bond82 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex rcu_read_lock &bond->stats_lock/1 irq_context: 0 (wq_completion)bond82 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex rcu_read_lock &ndev->lock irq_context: 0 (wq_completion)bond82 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex nl_table_lock irq_context: 0 (wq_completion)bond82 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &obj_hash[i].lock irq_context: 0 (wq_completion)bond82 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex nl_table_wait.lock irq_context: 0 (wq_completion)bond82 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex net_rwsem irq_context: 0 (wq_completion)bond82 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex net_rwsem &list->lock#2 irq_context: 0 (wq_completion)bond82 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &tn->lock irq_context: 0 (wq_completion)bond82 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &idev->mc_lock irq_context: 0 (wq_completion)bond82 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex.wait_lock irq_context: 0 (wq_completion)bond82 (work_completion)(&(&bond->mcast_work)->work) &p->pi_lock irq_context: 0 (wq_completion)bond78 (work_completion)(&(&slave->notify_work)->work) rtnl_mutex net_rwsem irq_context: 0 (wq_completion)bond78 (work_completion)(&(&slave->notify_work)->work) rtnl_mutex net_rwsem &list->lock#2 irq_context: 0 (wq_completion)bond78 (work_completion)(&(&slave->notify_work)->work) rtnl_mutex &tn->lock irq_context: 0 (wq_completion)bond78 (work_completion)(&(&slave->notify_work)->work) rtnl_mutex.wait_lock irq_context: 0 (wq_completion)bond78 (work_completion)(&(&slave->notify_work)->work) &p->pi_lock irq_context: 0 (wq_completion)bond81#2 (work_completion)(&(&slave->notify_work)->work) rtnl_mutex net_rwsem irq_context: 0 (wq_completion)bond81#2 (work_completion)(&(&slave->notify_work)->work) rtnl_mutex net_rwsem &list->lock#2 irq_context: 0 (wq_completion)bond81#2 (work_completion)(&(&slave->notify_work)->work) rtnl_mutex &tn->lock irq_context: 0 (wq_completion)bond81#2 (work_completion)(&(&slave->notify_work)->work) rtnl_mutex.wait_lock irq_context: 0 (wq_completion)bond81#2 (work_completion)(&(&slave->notify_work)->work) &p->pi_lock irq_context: 0 (wq_completion)bond98#2 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex fs_reclaim irq_context: 0 (wq_completion)bond98#2 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex fs_reclaim mmu_notifier_invalidate_range_start irq_context: 0 (wq_completion)bond98#2 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &c->lock irq_context: 0 (wq_completion)bond98#2 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex rcu_read_lock &bond->stats_lock/1 irq_context: 0 (wq_completion)bond98#2 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex rcu_read_lock &ndev->lock irq_context: 0 (wq_completion)bond98#2 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex nl_table_lock irq_context: 0 (wq_completion)bond98#2 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &obj_hash[i].lock irq_context: 0 (wq_completion)bond98#2 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex nl_table_wait.lock irq_context: 0 (wq_completion)bond98#2 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex net_rwsem irq_context: 0 (wq_completion)bond98#2 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex net_rwsem &list->lock#2 irq_context: 0 (wq_completion)bond98#2 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &tn->lock irq_context: 0 (wq_completion)ipv6_addrconf (work_completion)(&(&ifa->dad_work)->work) rtnl_mutex &obj_hash[i].lock pool_lock irq_context: 0 (wq_completion)bond98#2 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &idev->mc_lock irq_context: 0 (wq_completion)bond98#2 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &rq->__lock irq_context: 0 (wq_completion)bond98#2 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &cfs_rq->removed.lock irq_context: 0 (wq_completion)bond98#2 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex.wait_lock irq_context: 0 (wq_completion)bond98#2 (work_completion)(&(&bond->mcast_work)->work) &p->pi_lock irq_context: 0 (wq_completion)bond98#2 (work_completion)(&(&bond->mcast_work)->work) &p->pi_lock &rq->__lock irq_context: 0 (wq_completion)bond98#2 (work_completion)(&(&bond->mcast_work)->work) &p->pi_lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)bond98#2 (work_completion)(&(&slave->notify_work)->work) rtnl_mutex net_rwsem irq_context: 0 (wq_completion)bond98#2 (work_completion)(&(&slave->notify_work)->work) rtnl_mutex net_rwsem &list->lock#2 irq_context: 0 (wq_completion)bond98#2 (work_completion)(&(&slave->notify_work)->work) rtnl_mutex &tn->lock irq_context: 0 (wq_completion)bond111 (work_completion)(&(&slave->notify_work)->work) rtnl_mutex net_rwsem irq_context: 0 (wq_completion)bond111 (work_completion)(&(&slave->notify_work)->work) rtnl_mutex net_rwsem &list->lock#2 irq_context: 0 (wq_completion)bond111 (work_completion)(&(&slave->notify_work)->work) rtnl_mutex &tn->lock irq_context: 0 (wq_completion)bond95 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex fs_reclaim irq_context: 0 (wq_completion)bond95 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex fs_reclaim mmu_notifier_invalidate_range_start irq_context: 0 (wq_completion)bond95 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &c->lock irq_context: 0 (wq_completion)bond95 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &n->list_lock irq_context: 0 (wq_completion)bond95 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &n->list_lock &c->lock irq_context: 0 (wq_completion)bond95 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &rq->__lock irq_context: 0 (wq_completion)bond95 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)bond95 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex rcu_read_lock &bond->stats_lock/1 irq_context: 0 (wq_completion)bond95 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex rcu_read_lock &ndev->lock irq_context: 0 (wq_completion)bond95 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex nl_table_lock irq_context: 0 (wq_completion)bond95 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &obj_hash[i].lock irq_context: 0 (wq_completion)bond95 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex nl_table_wait.lock irq_context: 0 (wq_completion)bond95 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex net_rwsem irq_context: 0 (wq_completion)bond95 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex net_rwsem &list->lock#2 irq_context: 0 (wq_completion)bond95 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &tn->lock irq_context: 0 (wq_completion)bond95 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &idev->mc_lock irq_context: 0 (wq_completion)bond95 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &idev->mc_lock fs_reclaim irq_context: 0 (wq_completion)bond95 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &idev->mc_lock fs_reclaim mmu_notifier_invalidate_range_start irq_context: 0 (wq_completion)bond95 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &idev->mc_lock rcu_read_lock &ndev->lock irq_context: 0 (wq_completion)bond95 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &idev->mc_lock rcu_read_lock &c->lock irq_context: 0 (wq_completion)bond95 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &idev->mc_lock rcu_read_lock rcu_node_0 irq_context: 0 (wq_completion)bond95 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &idev->mc_lock rcu_read_lock &rq->__lock irq_context: 0 (wq_completion)bond95 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &idev->mc_lock rcu_read_lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)bond95 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &idev->mc_lock rcu_read_lock &dir->lock#2 irq_context: 0 (wq_completion)bond95 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &idev->mc_lock rcu_read_lock &ul->lock#2 irq_context: 0 (wq_completion)bond95 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &idev->mc_lock rcu_read_lock rcu_read_lock &____s->seqcount#10 irq_context: 0 (wq_completion)bond95 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &idev->mc_lock rcu_read_lock rcu_read_lock rcu_read_lock_bh rcu_read_lock rcu_read_lock_bh rcu_read_lock &br->multicast_lock irq_context: 0 (wq_completion)bond95 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &idev->mc_lock rcu_read_lock rcu_read_lock rcu_read_lock_bh rcu_read_lock rcu_read_lock_bh rcu_read_lock &br->multicast_lock &obj_hash[i].lock irq_context: 0 (wq_completion)bond95 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &idev->mc_lock rcu_read_lock rcu_read_lock rcu_read_lock_bh rcu_read_lock rcu_read_lock_bh rcu_read_lock &br->multicast_lock &base->lock irq_context: 0 (wq_completion)bond95 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &idev->mc_lock rcu_read_lock rcu_read_lock rcu_read_lock_bh rcu_read_lock rcu_read_lock_bh rcu_read_lock &data->lock irq_context: 0 (wq_completion)bond95 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &idev->mc_lock rcu_read_lock rcu_read_lock rcu_read_lock_bh rcu_read_lock rcu_read_lock_bh rcu_read_lock &obj_hash[i].lock irq_context: 0 (wq_completion)bond111 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex fs_reclaim irq_context: 0 (wq_completion)bond111 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex fs_reclaim mmu_notifier_invalidate_range_start irq_context: 0 (wq_completion)bond111 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &c->lock irq_context: 0 (wq_completion)bond111 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex rcu_read_lock &bond->stats_lock/1 irq_context: 0 (wq_completion)bond111 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex rcu_read_lock &ndev->lock irq_context: 0 (wq_completion)bond111 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex nl_table_lock irq_context: 0 (wq_completion)bond111 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &obj_hash[i].lock irq_context: 0 (wq_completion)bond111 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex nl_table_wait.lock irq_context: 0 (wq_completion)bond111 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex net_rwsem irq_context: 0 (wq_completion)bond111 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex net_rwsem &list->lock#2 irq_context: 0 (wq_completion)bond111 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &tn->lock irq_context: 0 (wq_completion)bond111 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &idev->mc_lock irq_context: 0 (wq_completion)bond111 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &idev->mc_lock fs_reclaim irq_context: 0 (wq_completion)bond111 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &idev->mc_lock fs_reclaim mmu_notifier_invalidate_range_start irq_context: 0 (wq_completion)bond111 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &idev->mc_lock rcu_read_lock &ndev->lock irq_context: 0 (wq_completion)bond111 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &idev->mc_lock rcu_read_lock &dir->lock#2 irq_context: 0 (wq_completion)bond111 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &idev->mc_lock rcu_read_lock &ul->lock#2 irq_context: 0 (wq_completion)bond111 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &idev->mc_lock rcu_read_lock rcu_read_lock &c->lock irq_context: 0 (wq_completion)bond111 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &idev->mc_lock rcu_read_lock rcu_read_lock &obj_hash[i].lock irq_context: 0 (wq_completion)bond111 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &idev->mc_lock rcu_read_lock rcu_read_lock &dir->lock#2 irq_context: 0 (wq_completion)bond111 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &idev->mc_lock rcu_read_lock rcu_read_lock &tbl->lock irq_context: 0 (wq_completion)bond111 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &idev->mc_lock rcu_read_lock rcu_read_lock &n->lock irq_context: 0 (wq_completion)bond111 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &idev->mc_lock rcu_read_lock rcu_read_lock &____s->seqcount#9 irq_context: 0 (wq_completion)bond111 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &idev->mc_lock rcu_read_lock rcu_read_lock rcu_read_lock_bh rcu_read_lock rcu_read_lock_bh rcu_read_lock &br->multicast_lock irq_context: 0 (wq_completion)bond111 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &idev->mc_lock rcu_read_lock rcu_read_lock rcu_read_lock_bh rcu_read_lock rcu_read_lock_bh rcu_read_lock &br->multicast_lock &obj_hash[i].lock irq_context: 0 (wq_completion)bond111 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &idev->mc_lock rcu_read_lock rcu_read_lock rcu_read_lock_bh rcu_read_lock rcu_read_lock_bh rcu_read_lock &br->multicast_lock rcu_read_lock rhashtable_bucket irq_context: 0 (wq_completion)bond111 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &idev->mc_lock rcu_read_lock rcu_read_lock rcu_read_lock_bh rcu_read_lock rcu_read_lock_bh rcu_read_lock &br->multicast_lock &c->lock irq_context: 0 (wq_completion)bond111 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &idev->mc_lock rcu_read_lock rcu_read_lock rcu_read_lock_bh rcu_read_lock rcu_read_lock_bh rcu_read_lock &br->multicast_lock nl_table_lock irq_context: 0 (wq_completion)bond111 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &idev->mc_lock rcu_read_lock rcu_read_lock rcu_read_lock_bh rcu_read_lock rcu_read_lock_bh rcu_read_lock &br->multicast_lock nl_table_wait.lock irq_context: 0 (wq_completion)bond111 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &idev->mc_lock rcu_read_lock rcu_read_lock rcu_read_lock_bh rcu_read_lock rcu_read_lock_bh rcu_read_lock &br->multicast_lock &base->lock irq_context: 0 (wq_completion)bond111 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &idev->mc_lock rcu_read_lock rcu_read_lock rcu_read_lock_bh rcu_read_lock rcu_read_lock_bh rcu_read_lock &br->multicast_lock &base->lock &obj_hash[i].lock irq_context: 0 (wq_completion)bond111 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &idev->mc_lock rcu_read_lock rcu_read_lock rcu_read_lock_bh rcu_read_lock rcu_read_lock_bh rcu_read_lock &data->lock irq_context: 0 (wq_completion)bond111 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &idev->mc_lock rcu_read_lock rcu_read_lock rcu_read_lock_bh rcu_read_lock rcu_read_lock_bh rcu_read_lock &obj_hash[i].lock irq_context: 0 (wq_completion)bond111 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &idev->mc_lock rcu_read_lock rcu_read_lock rcu_node_0 irq_context: 0 (wq_completion)bond111 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &idev->mc_lock rcu_read_lock rcu_read_lock &rq->__lock irq_context: 0 (wq_completion)bond111 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &idev->mc_lock rcu_read_lock rcu_read_lock &cfs_rq->removed.lock irq_context: 0 (wq_completion)bond111 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &idev->mc_lock rcu_read_lock rcu_read_lock pool_lock#2 irq_context: 0 (wq_completion)bond111 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &idev->mc_lock rcu_read_lock rcu_node_0 irq_context: 0 (wq_completion)bond111 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &idev->mc_lock rcu_read_lock &rcu_state.gp_wq irq_context: 0 (wq_completion)bond111 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &idev->mc_lock rcu_read_lock &rcu_state.gp_wq &p->pi_lock irq_context: 0 (wq_completion)bond111 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &idev->mc_lock rcu_read_lock &rcu_state.gp_wq &p->pi_lock &rq->__lock irq_context: 0 (wq_completion)bond111 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &idev->mc_lock rcu_read_lock &rcu_state.gp_wq &p->pi_lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)bond111 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &lock->wait_lock irq_context: 0 (wq_completion)bond111 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &p->pi_lock irq_context: 0 (wq_completion)bond111 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &p->pi_lock &rq->__lock irq_context: 0 (wq_completion)bond111 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &p->pi_lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 rtnl_mutex uevent_sock_mutex remove_cache_srcu &rq->__lock irq_context: 0 rtnl_mutex uevent_sock_mutex remove_cache_srcu &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 rtnl_mutex uevent_sock_mutex remove_cache_srcu pool_lock#2 irq_context: 0 (wq_completion)bond13#2 irq_context: 0 (wq_completion)bond13#2 (work_completion)(&(&slave->notify_work)->work) irq_context: 0 (wq_completion)bond13#2 (work_completion)(&(&slave->notify_work)->work) &obj_hash[i].lock irq_context: 0 (wq_completion)bond13#2 (work_completion)(&(&slave->notify_work)->work) &base->lock irq_context: 0 (wq_completion)bond13#2 (work_completion)(&(&slave->notify_work)->work) &base->lock &obj_hash[i].lock irq_context: 0 (wq_completion)bond13#2 (work_completion)(&(&slave->notify_work)->work) &rq->__lock irq_context: softirq (&in_dev->mr_ifc_timer) rcu_read_lock &im->lock rcu_read_lock &____s->seqcount#12 irq_context: 0 (wq_completion)bond13#2 (work_completion)(&(&bond->mcast_work)->work) irq_context: 0 (wq_completion)bond13#2 (work_completion)(&(&bond->mcast_work)->work) &obj_hash[i].lock irq_context: 0 (wq_completion)bond13#2 (work_completion)(&(&bond->mcast_work)->work) &base->lock irq_context: 0 (wq_completion)bond13#2 (work_completion)(&(&bond->mcast_work)->work) &base->lock &obj_hash[i].lock irq_context: 0 (wq_completion)bond13#2 (work_completion)(&(&bond->mcast_work)->work) &rq->__lock irq_context: 0 (wq_completion)bond13#2 (work_completion)(&(&bond->mcast_work)->work) &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)bond112 irq_context: 0 (wq_completion)bond112 (work_completion)(&(&slave->notify_work)->work) irq_context: 0 (wq_completion)bond112 (work_completion)(&(&slave->notify_work)->work) &obj_hash[i].lock irq_context: 0 (wq_completion)bond112 (work_completion)(&(&slave->notify_work)->work) &base->lock irq_context: 0 (wq_completion)bond112 (work_completion)(&(&slave->notify_work)->work) &base->lock &obj_hash[i].lock irq_context: 0 (wq_completion)bond112 (work_completion)(&(&slave->notify_work)->work) &rq->__lock irq_context: 0 (wq_completion)bond112 (work_completion)(&(&slave->notify_work)->work) &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)bond112 (work_completion)(&(&bond->mcast_work)->work) irq_context: 0 (wq_completion)bond112 (work_completion)(&(&bond->mcast_work)->work) &obj_hash[i].lock irq_context: 0 (wq_completion)bond112 (work_completion)(&(&bond->mcast_work)->work) &base->lock irq_context: 0 (wq_completion)bond112 (work_completion)(&(&bond->mcast_work)->work) &base->lock &obj_hash[i].lock irq_context: 0 (wq_completion)gid-cache-wq &rq->__lock irq_context: 0 (wq_completion)gid-cache-wq &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)bond83 irq_context: 0 (wq_completion)bond83 (work_completion)(&(&slave->notify_work)->work) irq_context: 0 (wq_completion)bond83 (work_completion)(&(&slave->notify_work)->work) &obj_hash[i].lock irq_context: 0 (wq_completion)bond83 (work_completion)(&(&slave->notify_work)->work) &base->lock irq_context: 0 (wq_completion)bond83 (work_completion)(&(&slave->notify_work)->work) &base->lock &obj_hash[i].lock irq_context: 0 (wq_completion)bond83 (work_completion)(&(&slave->notify_work)->work) &rq->__lock irq_context: 0 (wq_completion)bond83 (work_completion)(&(&bond->mcast_work)->work) irq_context: 0 (wq_completion)bond83 (work_completion)(&(&bond->mcast_work)->work) &obj_hash[i].lock irq_context: 0 (wq_completion)bond83 (work_completion)(&(&bond->mcast_work)->work) &base->lock irq_context: 0 (wq_completion)bond83 (work_completion)(&(&bond->mcast_work)->work) &base->lock &obj_hash[i].lock irq_context: 0 (wq_completion)bond100#2 irq_context: 0 (wq_completion)bond100#2 (work_completion)(&(&slave->notify_work)->work) irq_context: 0 (wq_completion)bond100#2 (work_completion)(&(&slave->notify_work)->work) &obj_hash[i].lock irq_context: 0 (wq_completion)bond100#2 (work_completion)(&(&slave->notify_work)->work) &base->lock irq_context: 0 (wq_completion)bond100#2 (work_completion)(&(&slave->notify_work)->work) &base->lock &obj_hash[i].lock irq_context: 0 (wq_completion)bond100#2 (work_completion)(&(&bond->mcast_work)->work) irq_context: 0 (wq_completion)bond100#2 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex fs_reclaim irq_context: 0 (wq_completion)bond100#2 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex fs_reclaim mmu_notifier_invalidate_range_start irq_context: 0 (wq_completion)bond100#2 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &c->lock irq_context: 0 (wq_completion)bond100#2 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex rcu_read_lock &bond->stats_lock/1 irq_context: 0 (wq_completion)bond100#2 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex rcu_read_lock &ndev->lock irq_context: 0 (wq_completion)bond100#2 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex nl_table_lock irq_context: 0 (wq_completion)bond100#2 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &obj_hash[i].lock irq_context: 0 (wq_completion)bond100#2 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex nl_table_wait.lock irq_context: 0 (wq_completion)bond100#2 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex net_rwsem irq_context: 0 (wq_completion)bond100#2 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex net_rwsem &list->lock#2 irq_context: 0 (wq_completion)bond100#2 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &tn->lock irq_context: 0 (wq_completion)bond100#2 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &idev->mc_lock irq_context: 0 (wq_completion)bond100#2 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex.wait_lock irq_context: 0 (wq_completion)bond100#2 (work_completion)(&(&bond->mcast_work)->work) &p->pi_lock irq_context: 0 (wq_completion)bond100#2 (work_completion)(&(&slave->notify_work)->work) rtnl_mutex net_rwsem irq_context: 0 (wq_completion)bond100#2 (work_completion)(&(&slave->notify_work)->work) rtnl_mutex net_rwsem &list->lock#2 irq_context: 0 (wq_completion)bond100#2 (work_completion)(&(&slave->notify_work)->work) rtnl_mutex &tn->lock irq_context: 0 (wq_completion)bond100#2 (work_completion)(&(&slave->notify_work)->work) rtnl_mutex.wait_lock irq_context: 0 (wq_completion)bond100#2 (work_completion)(&(&slave->notify_work)->work) &p->pi_lock irq_context: 0 (wq_completion)bond113 irq_context: 0 (wq_completion)bond113 (work_completion)(&(&slave->notify_work)->work) irq_context: 0 (wq_completion)bond113 (work_completion)(&(&slave->notify_work)->work) &obj_hash[i].lock irq_context: 0 (wq_completion)bond113 (work_completion)(&(&slave->notify_work)->work) &base->lock irq_context: 0 (wq_completion)bond113 (work_completion)(&(&slave->notify_work)->work) &base->lock &obj_hash[i].lock irq_context: 0 (wq_completion)bond113 (work_completion)(&(&bond->mcast_work)->work) irq_context: 0 (wq_completion)bond113 (work_completion)(&(&bond->mcast_work)->work) &obj_hash[i].lock irq_context: 0 (wq_completion)bond113 (work_completion)(&(&bond->mcast_work)->work) &base->lock irq_context: 0 (wq_completion)bond113 (work_completion)(&(&bond->mcast_work)->work) &base->lock &obj_hash[i].lock irq_context: 0 (wq_completion)mld (work_completion)(&(&idev->mc_ifc_work)->work) &idev->mc_lock rcu_read_lock rcu_read_lock rcu_read_lock_bh rcu_read_lock rcu_read_lock_bh rcu_read_lock &br->multicast_lock &obj_hash[i].lock pool_lock irq_context: 0 (wq_completion)bond84 irq_context: 0 (wq_completion)bond84 (work_completion)(&(&slave->notify_work)->work) irq_context: 0 (wq_completion)bond84 (work_completion)(&(&slave->notify_work)->work) &obj_hash[i].lock irq_context: 0 (wq_completion)bond84 (work_completion)(&(&slave->notify_work)->work) &base->lock irq_context: 0 (wq_completion)bond84 (work_completion)(&(&slave->notify_work)->work) &base->lock &obj_hash[i].lock irq_context: 0 (wq_completion)bond84 (work_completion)(&(&slave->notify_work)->work) &rq->__lock irq_context: 0 (wq_completion)bond113 (work_completion)(&(&bond->mcast_work)->work) &rq->__lock irq_context: 0 (wq_completion)bond113 (work_completion)(&(&bond->mcast_work)->work) &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)bond84 (work_completion)(&(&bond->mcast_work)->work) irq_context: 0 (wq_completion)bond84 (work_completion)(&(&bond->mcast_work)->work) &obj_hash[i].lock irq_context: 0 (wq_completion)bond84 (work_completion)(&(&bond->mcast_work)->work) &base->lock irq_context: 0 (wq_completion)bond84 (work_completion)(&(&bond->mcast_work)->work) &base->lock &obj_hash[i].lock irq_context: 0 (wq_completion)bond84 (work_completion)(&(&bond->mcast_work)->work) &rq->__lock irq_context: 0 (wq_completion)bond84 (work_completion)(&(&bond->mcast_work)->work) &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)bond101#2 irq_context: 0 (wq_completion)bond101#2 (work_completion)(&(&slave->notify_work)->work) irq_context: 0 (wq_completion)bond101#2 (work_completion)(&(&slave->notify_work)->work) &obj_hash[i].lock irq_context: 0 (wq_completion)bond101#2 (work_completion)(&(&slave->notify_work)->work) &base->lock irq_context: 0 (wq_completion)bond101#2 (work_completion)(&(&slave->notify_work)->work) &base->lock &obj_hash[i].lock irq_context: 0 (wq_completion)bond101#2 (work_completion)(&(&slave->notify_work)->work) &rq->__lock irq_context: 0 (wq_completion)bond101#2 (work_completion)(&(&slave->notify_work)->work) &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)bond101#2 (work_completion)(&(&bond->mcast_work)->work) irq_context: 0 (wq_completion)bond101#2 (work_completion)(&(&bond->mcast_work)->work) &obj_hash[i].lock irq_context: 0 (wq_completion)bond101#2 (work_completion)(&(&bond->mcast_work)->work) &base->lock irq_context: 0 (wq_completion)bond101#2 (work_completion)(&(&bond->mcast_work)->work) &base->lock &obj_hash[i].lock irq_context: 0 (wq_completion)bond101#2 (work_completion)(&(&bond->mcast_work)->work) &rq->__lock irq_context: 0 (wq_completion)bond14#2 irq_context: 0 (wq_completion)bond14#2 (work_completion)(&(&slave->notify_work)->work) irq_context: 0 (wq_completion)bond14#2 (work_completion)(&(&slave->notify_work)->work) &obj_hash[i].lock irq_context: 0 (wq_completion)bond14#2 (work_completion)(&(&slave->notify_work)->work) &base->lock irq_context: 0 (wq_completion)bond14#2 (work_completion)(&(&slave->notify_work)->work) &base->lock &obj_hash[i].lock irq_context: 0 (wq_completion)bond14#2 (work_completion)(&(&slave->notify_work)->work) &rq->__lock irq_context: 0 (wq_completion)bond14#2 (work_completion)(&(&slave->notify_work)->work) &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)bond14#2 (work_completion)(&(&bond->mcast_work)->work) irq_context: 0 (wq_completion)bond14#2 (work_completion)(&(&bond->mcast_work)->work) &obj_hash[i].lock irq_context: 0 (wq_completion)bond14#2 (work_completion)(&(&bond->mcast_work)->work) &base->lock irq_context: 0 (wq_completion)bond14#2 (work_completion)(&(&bond->mcast_work)->work) &base->lock &obj_hash[i].lock irq_context: 0 (wq_completion)bond81#2 &rq->__lock irq_context: 0 (wq_completion)bond81#2 &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)bond110 (work_completion)(&(&bond->mcast_work)->work) &rq->__lock irq_context: 0 (wq_completion)bond110 (work_completion)(&(&bond->mcast_work)->work) &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)events (linkwatch_work).work rtnl_mutex rcu_read_lock rcu_read_lock &pool->lock/1 fill_pool_map-wait-type-override &c->lock irq_context: 0 (wq_completion)events (linkwatch_work).work rtnl_mutex rcu_read_lock rcu_read_lock &pool->lock/1 fill_pool_map-wait-type-override &____s->seqcount#2 irq_context: 0 (wq_completion)events (linkwatch_work).work rtnl_mutex rcu_read_lock rcu_read_lock &pool->lock/1 fill_pool_map-wait-type-override &____s->seqcount irq_context: 0 (wq_completion)bond114 irq_context: 0 (wq_completion)bond114 (work_completion)(&(&slave->notify_work)->work) irq_context: 0 (wq_completion)bond114 (work_completion)(&(&slave->notify_work)->work) &obj_hash[i].lock irq_context: 0 (wq_completion)bond114 (work_completion)(&(&slave->notify_work)->work) &base->lock irq_context: 0 (wq_completion)bond114 (work_completion)(&(&slave->notify_work)->work) &base->lock &obj_hash[i].lock irq_context: 0 (wq_completion)bond114 (work_completion)(&(&slave->notify_work)->work) &rq->__lock irq_context: 0 (wq_completion)bond114 (work_completion)(&(&slave->notify_work)->work) &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)bond114 (work_completion)(&(&bond->mcast_work)->work) irq_context: 0 (wq_completion)bond114 (work_completion)(&(&bond->mcast_work)->work) &obj_hash[i].lock irq_context: 0 (wq_completion)bond114 (work_completion)(&(&bond->mcast_work)->work) &base->lock irq_context: 0 (wq_completion)bond114 (work_completion)(&(&bond->mcast_work)->work) &base->lock &obj_hash[i].lock irq_context: 0 (wq_completion)bond85#2 irq_context: 0 (wq_completion)bond85#2 (work_completion)(&(&slave->notify_work)->work) irq_context: 0 (wq_completion)bond85#2 (work_completion)(&(&slave->notify_work)->work) &obj_hash[i].lock irq_context: 0 (wq_completion)bond85#2 (work_completion)(&(&slave->notify_work)->work) &base->lock irq_context: 0 (wq_completion)bond85#2 (work_completion)(&(&slave->notify_work)->work) &base->lock &obj_hash[i].lock irq_context: 0 (wq_completion)bond85#2 (work_completion)(&(&slave->notify_work)->work) &rq->__lock irq_context: 0 (wq_completion)bond85#2 (work_completion)(&(&bond->mcast_work)->work) irq_context: 0 (wq_completion)bond85#2 (work_completion)(&(&bond->mcast_work)->work) &obj_hash[i].lock irq_context: 0 (wq_completion)bond85#2 (work_completion)(&(&bond->mcast_work)->work) &base->lock irq_context: 0 (wq_completion)bond85#2 (work_completion)(&(&bond->mcast_work)->work) &base->lock &obj_hash[i].lock irq_context: 0 (wq_completion)bond112 (work_completion)(&(&slave->notify_work)->work) rtnl_mutex net_rwsem irq_context: 0 (wq_completion)bond112 (work_completion)(&(&slave->notify_work)->work) rtnl_mutex net_rwsem &list->lock#2 irq_context: 0 (wq_completion)bond112 (work_completion)(&(&slave->notify_work)->work) rtnl_mutex &tn->lock irq_context: 0 (wq_completion)bond109 &rq->__lock irq_context: 0 (wq_completion)bond109 &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)bond112 (work_completion)(&(&slave->notify_work)->work) rtnl_mutex.wait_lock irq_context: 0 (wq_completion)bond112 (work_completion)(&(&slave->notify_work)->work) &p->pi_lock irq_context: 0 (wq_completion)bond109 (work_completion)(&(&slave->notify_work)->work) rtnl_mutex net_rwsem irq_context: 0 (wq_completion)bond109 (work_completion)(&(&slave->notify_work)->work) rtnl_mutex net_rwsem &list->lock#2 irq_context: 0 (wq_completion)bond109 (work_completion)(&(&slave->notify_work)->work) rtnl_mutex &tn->lock irq_context: 0 (wq_completion)bond109 (work_completion)(&(&slave->notify_work)->work) rtnl_mutex.wait_lock irq_context: 0 (wq_completion)bond109 (work_completion)(&(&slave->notify_work)->work) &p->pi_lock irq_context: 0 (wq_completion)bond84 (work_completion)(&(&slave->notify_work)->work) rtnl_mutex net_rwsem irq_context: 0 (wq_completion)bond84 (work_completion)(&(&slave->notify_work)->work) rtnl_mutex net_rwsem &list->lock#2 irq_context: 0 (wq_completion)bond84 (work_completion)(&(&slave->notify_work)->work) rtnl_mutex &tn->lock irq_context: 0 (wq_completion)bond84 (work_completion)(&(&slave->notify_work)->work) rtnl_mutex.wait_lock irq_context: 0 (wq_completion)bond84 (work_completion)(&(&slave->notify_work)->work) &p->pi_lock irq_context: 0 (wq_completion)bond13#2 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex fs_reclaim irq_context: 0 (wq_completion)bond13#2 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex fs_reclaim mmu_notifier_invalidate_range_start irq_context: 0 (wq_completion)bond13#2 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex fs_reclaim mmu_notifier_invalidate_range_start &rq->__lock irq_context: 0 (wq_completion)bond13#2 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex fs_reclaim mmu_notifier_invalidate_range_start &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)bond13#2 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex rcu_read_lock &bond->stats_lock/1 irq_context: 0 (wq_completion)bond13#2 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex rcu_read_lock &ndev->lock irq_context: 0 (wq_completion)bond13#2 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex nl_table_lock irq_context: 0 (wq_completion)bond13#2 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &obj_hash[i].lock irq_context: 0 (wq_completion)bond13#2 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex nl_table_wait.lock irq_context: 0 (wq_completion)bond13#2 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex net_rwsem irq_context: 0 (wq_completion)bond13#2 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex net_rwsem &list->lock#2 irq_context: 0 (wq_completion)bond13#2 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &tn->lock irq_context: 0 (wq_completion)bond13#2 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &idev->mc_lock irq_context: 0 (wq_completion)bond13#2 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &idev->mc_lock fs_reclaim irq_context: 0 (wq_completion)bond13#2 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &idev->mc_lock fs_reclaim mmu_notifier_invalidate_range_start irq_context: 0 (wq_completion)bond13#2 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &idev->mc_lock fs_reclaim mmu_notifier_invalidate_range_start &rq->__lock irq_context: 0 (wq_completion)bond13#2 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &idev->mc_lock fs_reclaim mmu_notifier_invalidate_range_start &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)bond13#2 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &idev->mc_lock rcu_read_lock &ndev->lock irq_context: 0 (wq_completion)bond13#2 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &idev->mc_lock rcu_read_lock &dir->lock#2 irq_context: 0 (wq_completion)bond13#2 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &idev->mc_lock rcu_read_lock &ul->lock#2 irq_context: 0 (wq_completion)bond13#2 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &idev->mc_lock rcu_read_lock rcu_read_lock &____s->seqcount#12 irq_context: 0 (wq_completion)bond13#2 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &idev->mc_lock rcu_read_lock rcu_read_lock &____s->seqcount#16 irq_context: 0 (wq_completion)bond13#2 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &idev->mc_lock rcu_read_lock rcu_read_lock pool_lock#2 irq_context: 0 (wq_completion)bond13#2 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &idev->mc_lock rcu_read_lock &obj_hash[i].lock irq_context: 0 (wq_completion)bond13#2 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &idev->mc_lock rcu_read_lock pool_lock#2 irq_context: 0 (wq_completion)bond13#2 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &idev->mc_lock rcu_read_lock &data->lock irq_context: 0 (wq_completion)bond13#2 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex.wait_lock irq_context: 0 (wq_completion)bond13#2 (work_completion)(&(&bond->mcast_work)->work) &p->pi_lock irq_context: 0 (wq_completion)bond102#2 irq_context: 0 (wq_completion)bond102#2 (work_completion)(&(&slave->notify_work)->work) irq_context: 0 (wq_completion)bond102#2 (work_completion)(&(&slave->notify_work)->work) &obj_hash[i].lock irq_context: 0 (wq_completion)bond102#2 (work_completion)(&(&slave->notify_work)->work) &base->lock irq_context: 0 (wq_completion)bond102#2 (work_completion)(&(&slave->notify_work)->work) &base->lock &obj_hash[i].lock irq_context: 0 (wq_completion)bond102#2 (work_completion)(&(&bond->mcast_work)->work) irq_context: 0 (wq_completion)bond102#2 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex fs_reclaim irq_context: 0 (wq_completion)bond102#2 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex fs_reclaim mmu_notifier_invalidate_range_start irq_context: 0 (wq_completion)bond102#2 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex rcu_read_lock &bond->stats_lock/1 irq_context: 0 (wq_completion)bond102#2 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex rcu_read_lock &ndev->lock irq_context: 0 (wq_completion)bond102#2 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex nl_table_lock irq_context: 0 (wq_completion)bond102#2 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &obj_hash[i].lock irq_context: 0 (wq_completion)bond102#2 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex nl_table_wait.lock irq_context: 0 (wq_completion)bond102#2 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &rq->__lock irq_context: 0 (wq_completion)bond102#2 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)bond102#2 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex net_rwsem irq_context: 0 (wq_completion)bond102#2 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex net_rwsem &list->lock#2 irq_context: 0 (wq_completion)bond102#2 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &tn->lock irq_context: 0 (wq_completion)bond102#2 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &idev->mc_lock irq_context: 0 (wq_completion)bond102#2 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex.wait_lock irq_context: 0 (wq_completion)bond102#2 (work_completion)(&(&bond->mcast_work)->work) &p->pi_lock irq_context: 0 (wq_completion)ipv6_addrconf (work_completion)(&(&ifa->dad_work)->work) rtnl_mutex rcu_read_lock rcu_read_lock fill_pool_map-wait-type-override pool_lock#2 irq_context: 0 (wq_completion)ipv6_addrconf (work_completion)(&(&ifa->dad_work)->work) rtnl_mutex rcu_read_lock rcu_read_lock fill_pool_map-wait-type-override pool_lock irq_context: 0 (wq_completion)ipv6_addrconf (work_completion)(&(&ifa->dad_work)->work) rtnl_mutex rcu_read_lock &____s->seqcount#12 irq_context: 0 (wq_completion)ipv6_addrconf (work_completion)(&(&ifa->dad_work)->work) rtnl_mutex rcu_read_lock &____s->seqcount#16 irq_context: 0 (wq_completion)ipv6_addrconf (work_completion)(&(&ifa->dad_work)->work) rtnl_mutex &data->lock irq_context: 0 (wq_completion)bond85#2 (work_completion)(&(&slave->notify_work)->work) rtnl_mutex net_rwsem irq_context: 0 (wq_completion)bond85#2 (work_completion)(&(&slave->notify_work)->work) rtnl_mutex net_rwsem &list->lock#2 irq_context: 0 (wq_completion)bond85#2 (work_completion)(&(&slave->notify_work)->work) rtnl_mutex &tn->lock irq_context: 0 (wq_completion)bond85#2 (work_completion)(&(&slave->notify_work)->work) rtnl_mutex.wait_lock irq_context: 0 (wq_completion)bond85#2 (work_completion)(&(&slave->notify_work)->work) &p->pi_lock irq_context: 0 (wq_completion)bond15#3 irq_context: 0 (wq_completion)bond15#3 &rq->__lock irq_context: 0 (wq_completion)bond15#3 &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)bond15#3 (work_completion)(&(&slave->notify_work)->work) irq_context: 0 (wq_completion)bond15#3 (work_completion)(&(&slave->notify_work)->work) &obj_hash[i].lock irq_context: 0 (wq_completion)bond15#3 (work_completion)(&(&slave->notify_work)->work) &base->lock irq_context: 0 (wq_completion)bond15#3 (work_completion)(&(&slave->notify_work)->work) &base->lock &obj_hash[i].lock irq_context: 0 (wq_completion)bond15#3 (work_completion)(&(&bond->mcast_work)->work) irq_context: 0 (wq_completion)bond15#3 (work_completion)(&(&bond->mcast_work)->work) &obj_hash[i].lock irq_context: 0 (wq_completion)bond15#3 (work_completion)(&(&bond->mcast_work)->work) &base->lock irq_context: 0 (wq_completion)bond15#3 (work_completion)(&(&bond->mcast_work)->work) &base->lock &obj_hash[i].lock irq_context: 0 (wq_completion)bond115 irq_context: 0 (wq_completion)bond115 (work_completion)(&(&slave->notify_work)->work) irq_context: 0 (wq_completion)bond115 (work_completion)(&(&slave->notify_work)->work) &obj_hash[i].lock irq_context: 0 (wq_completion)bond115 (work_completion)(&(&slave->notify_work)->work) &base->lock irq_context: 0 (wq_completion)bond115 (work_completion)(&(&slave->notify_work)->work) &base->lock &obj_hash[i].lock irq_context: 0 (wq_completion)bond115 (work_completion)(&(&slave->notify_work)->work) &rq->__lock irq_context: 0 (wq_completion)bond115 (work_completion)(&(&slave->notify_work)->work) &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)bond115 (work_completion)(&(&bond->mcast_work)->work) irq_context: 0 (wq_completion)bond115 (work_completion)(&(&bond->mcast_work)->work) &obj_hash[i].lock irq_context: 0 (wq_completion)bond115 (work_completion)(&(&bond->mcast_work)->work) &base->lock irq_context: 0 (wq_completion)bond115 (work_completion)(&(&bond->mcast_work)->work) &base->lock &obj_hash[i].lock irq_context: 0 (wq_completion)bond114 (work_completion)(&(&bond->mcast_work)->work) &rq->__lock irq_context: 0 (wq_completion)bond114 (work_completion)(&(&bond->mcast_work)->work) &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)bond86 irq_context: 0 (wq_completion)bond86 (work_completion)(&(&slave->notify_work)->work) irq_context: 0 (wq_completion)bond86 (work_completion)(&(&slave->notify_work)->work) &obj_hash[i].lock irq_context: 0 (wq_completion)bond86 (work_completion)(&(&slave->notify_work)->work) &base->lock irq_context: 0 (wq_completion)bond86 (work_completion)(&(&slave->notify_work)->work) &base->lock &obj_hash[i].lock irq_context: 0 (wq_completion)bond86 (work_completion)(&(&slave->notify_work)->work) &rq->__lock irq_context: 0 (wq_completion)bond86 (work_completion)(&(&bond->mcast_work)->work) irq_context: 0 (wq_completion)bond86 (work_completion)(&(&bond->mcast_work)->work) &obj_hash[i].lock irq_context: 0 (wq_completion)bond86 (work_completion)(&(&bond->mcast_work)->work) &base->lock irq_context: 0 (wq_completion)bond86 (work_completion)(&(&bond->mcast_work)->work) &base->lock &obj_hash[i].lock irq_context: 0 (wq_completion)ipv6_addrconf (work_completion)(&(&ifa->dad_work)->work) rtnl_mutex &ndev->lock &ifa->lock fill_pool_map-wait-type-override pool_lock#2 irq_context: 0 (wq_completion)ipv6_addrconf (work_completion)(&(&ifa->dad_work)->work) rtnl_mutex &ndev->lock &ifa->lock fill_pool_map-wait-type-override pool_lock irq_context: 0 (wq_completion)bond86 (work_completion)(&(&slave->notify_work)->work) rtnl_mutex net_rwsem irq_context: 0 (wq_completion)bond86 (work_completion)(&(&slave->notify_work)->work) rtnl_mutex net_rwsem &list->lock#2 irq_context: 0 (wq_completion)bond86 (work_completion)(&(&slave->notify_work)->work) rtnl_mutex &tn->lock irq_context: 0 (wq_completion)bond86 (work_completion)(&(&slave->notify_work)->work) rtnl_mutex.wait_lock irq_context: 0 (wq_completion)bond86 (work_completion)(&(&slave->notify_work)->work) &p->pi_lock irq_context: 0 (wq_completion)bond86 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex fs_reclaim irq_context: 0 (wq_completion)bond86 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex fs_reclaim mmu_notifier_invalidate_range_start irq_context: 0 (wq_completion)bond86 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &c->lock irq_context: 0 (wq_completion)bond86 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex rcu_read_lock &bond->stats_lock/1 irq_context: 0 (wq_completion)bond86 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex rcu_read_lock &ndev->lock irq_context: 0 (wq_completion)bond86 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex nl_table_lock irq_context: 0 (wq_completion)bond86 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &obj_hash[i].lock irq_context: 0 (wq_completion)bond86 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex nl_table_wait.lock irq_context: 0 (wq_completion)bond86 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex net_rwsem irq_context: 0 (wq_completion)bond86 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex net_rwsem &list->lock#2 irq_context: 0 (wq_completion)bond86 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &tn->lock irq_context: 0 (wq_completion)bond86 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &idev->mc_lock irq_context: 0 (wq_completion)bond86 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &idev->mc_lock fs_reclaim irq_context: 0 (wq_completion)bond86 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &idev->mc_lock fs_reclaim mmu_notifier_invalidate_range_start irq_context: 0 (wq_completion)bond86 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &idev->mc_lock fs_reclaim mmu_notifier_invalidate_range_start &rq->__lock irq_context: 0 (wq_completion)bond86 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &idev->mc_lock rcu_read_lock &ndev->lock irq_context: 0 (wq_completion)bond86 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &idev->mc_lock rcu_read_lock &dir->lock#2 irq_context: 0 (wq_completion)bond86 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &idev->mc_lock rcu_read_lock &ul->lock#2 irq_context: 0 (wq_completion)bond86 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &idev->mc_lock rcu_read_lock rcu_read_lock &obj_hash[i].lock irq_context: 0 (wq_completion)bond86 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &idev->mc_lock rcu_read_lock rcu_read_lock &dir->lock#2 irq_context: 0 (wq_completion)bond86 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &idev->mc_lock rcu_read_lock rcu_read_lock &tbl->lock irq_context: 0 (wq_completion)bond86 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &idev->mc_lock rcu_read_lock rcu_read_lock &n->lock irq_context: 0 (wq_completion)bond86 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &idev->mc_lock rcu_read_lock rcu_read_lock &____s->seqcount#9 irq_context: 0 (wq_completion)bond86 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &idev->mc_lock rcu_read_lock rcu_read_lock rcu_read_lock_bh rcu_read_lock rcu_read_lock_bh rcu_read_lock &br->multicast_lock irq_context: 0 (wq_completion)bond86 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &idev->mc_lock rcu_read_lock rcu_read_lock rcu_read_lock_bh rcu_read_lock rcu_read_lock_bh rcu_read_lock &br->multicast_lock &obj_hash[i].lock irq_context: 0 (wq_completion)bond86 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &idev->mc_lock rcu_read_lock rcu_read_lock rcu_read_lock_bh rcu_read_lock rcu_read_lock_bh rcu_read_lock &br->multicast_lock rcu_read_lock rhashtable_bucket irq_context: 0 (wq_completion)bond86 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &idev->mc_lock rcu_read_lock rcu_read_lock rcu_read_lock_bh rcu_read_lock rcu_read_lock_bh rcu_read_lock &br->multicast_lock &c->lock irq_context: 0 (wq_completion)bond86 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &idev->mc_lock rcu_read_lock rcu_read_lock rcu_read_lock_bh rcu_read_lock rcu_read_lock_bh rcu_read_lock &br->multicast_lock nl_table_lock irq_context: 0 (wq_completion)bond86 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &idev->mc_lock rcu_read_lock rcu_read_lock rcu_read_lock_bh rcu_read_lock rcu_read_lock_bh rcu_read_lock &br->multicast_lock nl_table_wait.lock irq_context: 0 (wq_completion)bond86 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &idev->mc_lock rcu_read_lock rcu_read_lock rcu_read_lock_bh rcu_read_lock rcu_read_lock_bh rcu_read_lock &br->multicast_lock &base->lock irq_context: 0 (wq_completion)bond86 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &idev->mc_lock rcu_read_lock rcu_read_lock rcu_read_lock_bh rcu_read_lock rcu_read_lock_bh rcu_read_lock &br->multicast_lock &base->lock &obj_hash[i].lock irq_context: 0 (wq_completion)bond86 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &idev->mc_lock rcu_read_lock rcu_read_lock rcu_read_lock_bh rcu_read_lock rcu_read_lock_bh rcu_read_lock &data->lock irq_context: 0 (wq_completion)bond86 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &idev->mc_lock rcu_read_lock rcu_read_lock rcu_read_lock_bh rcu_read_lock rcu_read_lock_bh rcu_read_lock &obj_hash[i].lock irq_context: 0 (wq_completion)bond86 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex.wait_lock irq_context: 0 (wq_completion)bond86 (work_completion)(&(&bond->mcast_work)->work) &p->pi_lock irq_context: 0 (wq_completion)bond86 (work_completion)(&(&bond->mcast_work)->work) &p->pi_lock &rq->__lock irq_context: 0 (wq_completion)bond86 (work_completion)(&(&bond->mcast_work)->work) &p->pi_lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)bond86 (work_completion)(&(&bond->mcast_work)->work) &rq->__lock irq_context: 0 rtnl_mutex &br->hash_lock quarantine_lock irq_context: 0 (wq_completion)bond103#2 irq_context: 0 (wq_completion)bond103#2 (work_completion)(&(&slave->notify_work)->work) irq_context: 0 (wq_completion)bond103#2 (work_completion)(&(&slave->notify_work)->work) &obj_hash[i].lock irq_context: 0 (wq_completion)bond103#2 (work_completion)(&(&slave->notify_work)->work) &base->lock irq_context: 0 (wq_completion)bond103#2 (work_completion)(&(&slave->notify_work)->work) &base->lock &obj_hash[i].lock irq_context: 0 (wq_completion)bond103#2 (work_completion)(&(&bond->mcast_work)->work) irq_context: 0 (wq_completion)bond103#2 (work_completion)(&(&bond->mcast_work)->work) &obj_hash[i].lock irq_context: 0 (wq_completion)bond103#2 (work_completion)(&(&bond->mcast_work)->work) &base->lock irq_context: 0 (wq_completion)bond103#2 (work_completion)(&(&bond->mcast_work)->work) &base->lock &obj_hash[i].lock irq_context: 0 (wq_completion)bond103#2 (work_completion)(&(&bond->mcast_work)->work) &rq->__lock irq_context: 0 (wq_completion)bond103#2 (work_completion)(&(&bond->mcast_work)->work) &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)bond116 irq_context: 0 (wq_completion)bond116 (work_completion)(&(&slave->notify_work)->work) irq_context: 0 (wq_completion)bond116 (work_completion)(&(&slave->notify_work)->work) &obj_hash[i].lock irq_context: 0 (wq_completion)bond116 (work_completion)(&(&slave->notify_work)->work) &base->lock irq_context: 0 (wq_completion)bond116 (work_completion)(&(&slave->notify_work)->work) &base->lock &obj_hash[i].lock irq_context: 0 (wq_completion)bond116 (work_completion)(&(&slave->notify_work)->work) &rq->__lock irq_context: 0 (wq_completion)bond116 (work_completion)(&(&slave->notify_work)->work) &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)bond116 (work_completion)(&(&bond->mcast_work)->work) irq_context: 0 (wq_completion)bond116 (work_completion)(&(&bond->mcast_work)->work) &obj_hash[i].lock irq_context: 0 (wq_completion)bond116 (work_completion)(&(&bond->mcast_work)->work) &base->lock irq_context: 0 (wq_completion)bond116 (work_completion)(&(&bond->mcast_work)->work) &base->lock &obj_hash[i].lock irq_context: 0 (wq_completion)bond80#2 &rq->__lock irq_context: 0 (wq_completion)bond80#2 &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: softirq &(&bond->mcast_work)->timer rcu_read_lock &pool->lock/1 &p->pi_lock &rq->__lock &cfs_rq->removed.lock irq_context: 0 (wq_completion)bond15#3 (work_completion)(&(&bond->mcast_work)->work) &rq->__lock irq_context: 0 (wq_completion)bond15#3 (work_completion)(&(&bond->mcast_work)->work) &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: softirq rcu_read_lock rcu_read_lock slock-AF_INET6/1 sk_lock-AF_INET6 rcu_read_lock fill_pool_map-wait-type-override &c->lock irq_context: softirq rcu_read_lock rcu_read_lock slock-AF_INET6/1 sk_lock-AF_INET6 rcu_read_lock fill_pool_map-wait-type-override &____s->seqcount#2 irq_context: softirq rcu_read_lock rcu_read_lock slock-AF_INET6/1 sk_lock-AF_INET6 rcu_read_lock fill_pool_map-wait-type-override &____s->seqcount irq_context: 0 (wq_completion)bond87 irq_context: 0 (wq_completion)bond87 (work_completion)(&(&slave->notify_work)->work) irq_context: 0 (wq_completion)bond87 (work_completion)(&(&slave->notify_work)->work) &obj_hash[i].lock irq_context: 0 (wq_completion)bond87 (work_completion)(&(&slave->notify_work)->work) &base->lock irq_context: 0 (wq_completion)bond87 (work_completion)(&(&slave->notify_work)->work) &base->lock &obj_hash[i].lock irq_context: 0 (wq_completion)bond87 (work_completion)(&(&slave->notify_work)->work) &rq->__lock irq_context: 0 (wq_completion)ipv6_addrconf (work_completion)(&(&ifa->dad_work)->work) rtnl_mutex &data->lock &obj_hash[i].lock irq_context: 0 (wq_completion)ipv6_addrconf (work_completion)(&(&ifa->dad_work)->work) rtnl_mutex &data->lock &base->lock irq_context: 0 (wq_completion)ipv6_addrconf (work_completion)(&(&ifa->dad_work)->work) rtnl_mutex &data->lock &base->lock &obj_hash[i].lock irq_context: 0 (wq_completion)bond87 (work_completion)(&(&bond->mcast_work)->work) irq_context: 0 (wq_completion)bond87 (work_completion)(&(&bond->mcast_work)->work) &obj_hash[i].lock irq_context: 0 (wq_completion)bond87 (work_completion)(&(&bond->mcast_work)->work) &base->lock irq_context: 0 (wq_completion)bond87 (work_completion)(&(&bond->mcast_work)->work) &base->lock &obj_hash[i].lock irq_context: 0 (wq_completion)ipv6_addrconf (work_completion)(&(&net->ipv6.addr_chk_work)->work) &rq->__lock irq_context: 0 rtnl_mutex fs_reclaim mmu_notifier_invalidate_range_start &cfs_rq->removed.lock irq_context: 0 (wq_completion)bond104#2 irq_context: 0 (wq_completion)bond104#2 &rq->__lock irq_context: 0 (wq_completion)bond104#2 &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)bond104#2 (work_completion)(&(&slave->notify_work)->work) irq_context: 0 (wq_completion)bond104#2 (work_completion)(&(&slave->notify_work)->work) &obj_hash[i].lock irq_context: 0 (wq_completion)bond104#2 (work_completion)(&(&slave->notify_work)->work) &base->lock irq_context: 0 (wq_completion)bond104#2 (work_completion)(&(&slave->notify_work)->work) &base->lock &obj_hash[i].lock irq_context: 0 (wq_completion)bond104#2 (work_completion)(&(&slave->notify_work)->work) &rq->__lock irq_context: 0 (wq_completion)bond104#2 (work_completion)(&(&slave->notify_work)->work) &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)bond104#2 (work_completion)(&(&bond->mcast_work)->work) irq_context: 0 (wq_completion)bond104#2 (work_completion)(&(&bond->mcast_work)->work) &obj_hash[i].lock irq_context: 0 (wq_completion)bond104#2 (work_completion)(&(&bond->mcast_work)->work) &base->lock irq_context: 0 (wq_completion)bond104#2 (work_completion)(&(&bond->mcast_work)->work) &base->lock &obj_hash[i].lock irq_context: 0 (wq_completion)bond117 irq_context: 0 (wq_completion)bond117 (work_completion)(&(&slave->notify_work)->work) irq_context: 0 (wq_completion)bond117 (work_completion)(&(&slave->notify_work)->work) &obj_hash[i].lock irq_context: 0 (wq_completion)bond117 (work_completion)(&(&slave->notify_work)->work) &base->lock irq_context: 0 (wq_completion)bond117 (work_completion)(&(&slave->notify_work)->work) &base->lock &obj_hash[i].lock irq_context: 0 (wq_completion)bond117 (work_completion)(&(&slave->notify_work)->work) &rq->__lock irq_context: 0 (wq_completion)bond117 (work_completion)(&(&slave->notify_work)->work) &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)bond104#2 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex fs_reclaim irq_context: 0 (wq_completion)bond104#2 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex fs_reclaim mmu_notifier_invalidate_range_start irq_context: 0 (wq_completion)bond104#2 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex rcu_read_lock &bond->stats_lock/1 irq_context: 0 (wq_completion)bond104#2 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex rcu_read_lock &ndev->lock irq_context: 0 (wq_completion)bond104#2 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex nl_table_lock irq_context: 0 (wq_completion)bond104#2 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &obj_hash[i].lock irq_context: 0 (wq_completion)bond104#2 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex nl_table_wait.lock irq_context: 0 (wq_completion)bond104#2 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex net_rwsem irq_context: 0 (wq_completion)bond104#2 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex net_rwsem &list->lock#2 irq_context: 0 (wq_completion)bond104#2 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &tn->lock irq_context: 0 (wq_completion)bond104#2 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &idev->mc_lock irq_context: 0 (wq_completion)bond104#2 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex.wait_lock irq_context: 0 (wq_completion)bond104#2 (work_completion)(&(&bond->mcast_work)->work) &p->pi_lock irq_context: 0 (wq_completion)bond83 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex fs_reclaim irq_context: 0 (wq_completion)bond83 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex fs_reclaim mmu_notifier_invalidate_range_start irq_context: 0 (wq_completion)bond83 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &c->lock irq_context: 0 (wq_completion)bond83 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex rcu_read_lock &bond->stats_lock/1 irq_context: 0 (wq_completion)bond83 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex rcu_read_lock &ndev->lock irq_context: 0 (wq_completion)bond83 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex nl_table_lock irq_context: 0 (wq_completion)bond83 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &rq->__lock irq_context: 0 (wq_completion)bond83 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)bond117 (work_completion)(&(&bond->mcast_work)->work) irq_context: 0 (wq_completion)bond117 (work_completion)(&(&bond->mcast_work)->work) &obj_hash[i].lock irq_context: 0 (wq_completion)bond117 (work_completion)(&(&bond->mcast_work)->work) &base->lock irq_context: 0 (wq_completion)bond117 (work_completion)(&(&bond->mcast_work)->work) &base->lock &obj_hash[i].lock irq_context: 0 (wq_completion)bond83 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &obj_hash[i].lock irq_context: 0 (wq_completion)bond83 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex nl_table_wait.lock irq_context: 0 (wq_completion)bond83 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex nl_table_wait.lock &p->pi_lock irq_context: 0 (wq_completion)bond83 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex nl_table_wait.lock &p->pi_lock &cfs_rq->removed.lock irq_context: 0 (wq_completion)bond83 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex nl_table_wait.lock &p->pi_lock &rq->__lock irq_context: 0 (wq_completion)bond83 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex nl_table_wait.lock &p->pi_lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)bond83 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex net_rwsem irq_context: 0 (wq_completion)bond83 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex net_rwsem &list->lock#2 irq_context: 0 (wq_completion)bond83 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &tn->lock irq_context: 0 (wq_completion)bond83 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &idev->mc_lock irq_context: 0 (wq_completion)bond83 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &idev->mc_lock fs_reclaim irq_context: 0 (wq_completion)bond83 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &idev->mc_lock fs_reclaim mmu_notifier_invalidate_range_start irq_context: 0 (wq_completion)bond83 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &idev->mc_lock &c->lock irq_context: 0 (wq_completion)bond83 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &idev->mc_lock rcu_read_lock &ndev->lock irq_context: 0 (wq_completion)bond83 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &idev->mc_lock rcu_read_lock &dir->lock#2 irq_context: 0 (wq_completion)bond83 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &idev->mc_lock rcu_read_lock &ul->lock#2 irq_context: 0 (wq_completion)bond83 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &idev->mc_lock rcu_read_lock rcu_read_lock &____s->seqcount#10 irq_context: 0 (wq_completion)bond83 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &idev->mc_lock rcu_read_lock rcu_read_lock rcu_read_lock_bh rcu_read_lock rcu_read_lock_bh rcu_read_lock &br->multicast_lock irq_context: 0 (wq_completion)bond83 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &idev->mc_lock rcu_read_lock rcu_read_lock rcu_read_lock_bh rcu_read_lock rcu_read_lock_bh rcu_read_lock &br->multicast_lock &obj_hash[i].lock irq_context: 0 (wq_completion)bond83 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &idev->mc_lock rcu_read_lock rcu_read_lock rcu_read_lock_bh rcu_read_lock rcu_read_lock_bh rcu_read_lock &br->multicast_lock &base->lock irq_context: 0 (wq_completion)bond83 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &idev->mc_lock rcu_read_lock rcu_read_lock rcu_read_lock_bh rcu_read_lock rcu_read_lock_bh rcu_read_lock &br->multicast_lock &base->lock &obj_hash[i].lock irq_context: 0 (wq_completion)bond83 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &idev->mc_lock rcu_read_lock rcu_read_lock rcu_read_lock_bh rcu_read_lock rcu_read_lock_bh rcu_read_lock &data->lock irq_context: 0 (wq_completion)bond83 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &idev->mc_lock rcu_read_lock rcu_read_lock rcu_read_lock_bh rcu_read_lock rcu_read_lock_bh rcu_read_lock &obj_hash[i].lock irq_context: 0 (wq_completion)bond83 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &idev->mc_lock rcu_read_lock rcu_read_lock rcu_node_0 irq_context: 0 (wq_completion)bond83 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &idev->mc_lock rcu_read_lock rcu_read_lock &rq->__lock irq_context: 0 (wq_completion)bond83 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &idev->mc_lock rcu_read_lock rcu_read_lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)bond83 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &idev->mc_lock rcu_read_lock rcu_node_0 irq_context: 0 (wq_completion)bond83 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex.wait_lock irq_context: 0 (wq_completion)bond83 (work_completion)(&(&bond->mcast_work)->work) &p->pi_lock irq_context: 0 (wq_completion)bond83 (work_completion)(&(&bond->mcast_work)->work) &p->pi_lock &rq->__lock irq_context: 0 (wq_completion)bond83 (work_completion)(&(&bond->mcast_work)->work) &p->pi_lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)bond83 (work_completion)(&(&bond->mcast_work)->work) &rq->__lock irq_context: 0 (wq_completion)bond83 (work_completion)(&(&bond->mcast_work)->work) &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)ipv6_addrconf (work_completion)(&(&ifa->dad_work)->work) rtnl_mutex rcu_read_lock rcu_read_lock rcu_read_lock_bh rcu_read_lock rcu_read_lock_bh rcu_read_lock fill_pool_map-wait-type-override pool_lock#2 irq_context: 0 (wq_completion)ipv6_addrconf (work_completion)(&(&ifa->dad_work)->work) rtnl_mutex rcu_read_lock rcu_read_lock rcu_read_lock_bh rcu_read_lock rcu_read_lock_bh rcu_read_lock fill_pool_map-wait-type-override pool_lock irq_context: 0 (wq_completion)bond83 (work_completion)(&(&bond->mcast_work)->work) &cfs_rq->removed.lock irq_context: 0 (wq_completion)bond83 (work_completion)(&(&bond->mcast_work)->work) pool_lock#2 irq_context: 0 (wq_completion)bond88#2 irq_context: 0 (wq_completion)bond88#2 (work_completion)(&(&slave->notify_work)->work) irq_context: 0 (wq_completion)bond88#2 (work_completion)(&(&slave->notify_work)->work) &obj_hash[i].lock irq_context: 0 (wq_completion)bond88#2 (work_completion)(&(&slave->notify_work)->work) &base->lock irq_context: 0 (wq_completion)bond88#2 (work_completion)(&(&slave->notify_work)->work) &base->lock &obj_hash[i].lock irq_context: 0 (wq_completion)bond88#2 (work_completion)(&(&slave->notify_work)->work) &rq->__lock irq_context: 0 (wq_completion)bond88#2 (work_completion)(&(&slave->notify_work)->work) &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)bond102#2 &rq->__lock irq_context: 0 (wq_completion)bond102#2 &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)bond88#2 (work_completion)(&(&bond->mcast_work)->work) irq_context: 0 (wq_completion)bond88#2 (work_completion)(&(&bond->mcast_work)->work) &rq->__lock irq_context: 0 (wq_completion)bond88#2 (work_completion)(&(&bond->mcast_work)->work) &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)bond88#2 (work_completion)(&(&bond->mcast_work)->work) &obj_hash[i].lock irq_context: 0 (wq_completion)bond88#2 (work_completion)(&(&bond->mcast_work)->work) &base->lock irq_context: 0 (wq_completion)bond88#2 (work_completion)(&(&bond->mcast_work)->work) &base->lock &obj_hash[i].lock irq_context: 0 (wq_completion)bond87 (work_completion)(&(&slave->notify_work)->work) rtnl_mutex net_rwsem irq_context: 0 (wq_completion)bond87 (work_completion)(&(&slave->notify_work)->work) rtnl_mutex net_rwsem &rq->__lock irq_context: 0 (wq_completion)bond87 (work_completion)(&(&slave->notify_work)->work) rtnl_mutex net_rwsem &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)bond87 (work_completion)(&(&slave->notify_work)->work) rtnl_mutex net_rwsem &list->lock#2 irq_context: 0 (wq_completion)bond87 (work_completion)(&(&slave->notify_work)->work) rtnl_mutex &tn->lock irq_context: 0 (wq_completion)bond87 (work_completion)(&(&slave->notify_work)->work) rtnl_mutex.wait_lock irq_context: 0 (wq_completion)bond87 (work_completion)(&(&slave->notify_work)->work) &p->pi_lock irq_context: 0 (wq_completion)bond87 (work_completion)(&(&slave->notify_work)->work) &p->pi_lock &rq->__lock irq_context: 0 (wq_completion)bond87 (work_completion)(&(&slave->notify_work)->work) &p->pi_lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)bond87 (work_completion)(&(&slave->notify_work)->work) &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)bond105#2 irq_context: 0 (wq_completion)bond105#2 (work_completion)(&(&slave->notify_work)->work) irq_context: 0 (wq_completion)bond105#2 (work_completion)(&(&slave->notify_work)->work) &obj_hash[i].lock irq_context: 0 (wq_completion)bond105#2 (work_completion)(&(&slave->notify_work)->work) &base->lock irq_context: 0 (wq_completion)bond105#2 (work_completion)(&(&slave->notify_work)->work) &base->lock &obj_hash[i].lock irq_context: 0 (wq_completion)bond105#2 (work_completion)(&(&slave->notify_work)->work) &rq->__lock irq_context: 0 (wq_completion)bond105#2 (work_completion)(&(&slave->notify_work)->work) &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)bond105#2 (work_completion)(&(&bond->mcast_work)->work) irq_context: 0 (wq_completion)bond105#2 (work_completion)(&(&bond->mcast_work)->work) &obj_hash[i].lock irq_context: 0 (wq_completion)bond105#2 (work_completion)(&(&bond->mcast_work)->work) &base->lock irq_context: 0 (wq_completion)bond105#2 (work_completion)(&(&bond->mcast_work)->work) &base->lock &obj_hash[i].lock irq_context: 0 (wq_completion)bond105#2 (work_completion)(&(&bond->mcast_work)->work) &rq->__lock irq_context: 0 (wq_completion)bond105#2 (work_completion)(&(&bond->mcast_work)->work) &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)ipv6_addrconf (work_completion)(&(&ifa->dad_work)->work) rtnl_mutex &idev->mc_lock rcu_read_lock rcu_read_lock &____s->seqcount#12 irq_context: 0 (wq_completion)ipv6_addrconf (work_completion)(&(&ifa->dad_work)->work) rtnl_mutex &idev->mc_lock rcu_read_lock rcu_read_lock &____s->seqcount#16 irq_context: 0 (wq_completion)ipv6_addrconf (work_completion)(&(&ifa->dad_work)->work) rtnl_mutex &idev->mc_lock rcu_read_lock &obj_hash[i].lock irq_context: 0 (wq_completion)ipv6_addrconf (work_completion)(&(&ifa->dad_work)->work) rtnl_mutex &idev->mc_lock rcu_read_lock &data->lock irq_context: 0 (wq_completion)bond83 (work_completion)(&(&slave->notify_work)->work) &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)bond17#3 irq_context: 0 (wq_completion)bond17#3 (work_completion)(&(&slave->notify_work)->work) irq_context: 0 (wq_completion)bond17#3 (work_completion)(&(&slave->notify_work)->work) &obj_hash[i].lock irq_context: 0 (wq_completion)bond17#3 (work_completion)(&(&slave->notify_work)->work) &base->lock irq_context: 0 (wq_completion)bond17#3 (work_completion)(&(&slave->notify_work)->work) &base->lock &obj_hash[i].lock irq_context: 0 (wq_completion)bond17#3 (work_completion)(&(&slave->notify_work)->work) &rq->__lock irq_context: 0 (wq_completion)bond17#3 (work_completion)(&(&slave->notify_work)->work) &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)bond17#3 (work_completion)(&(&bond->mcast_work)->work) irq_context: 0 (wq_completion)bond17#3 (work_completion)(&(&bond->mcast_work)->work) &obj_hash[i].lock irq_context: 0 (wq_completion)bond17#3 (work_completion)(&(&bond->mcast_work)->work) &base->lock irq_context: 0 (wq_completion)bond17#3 (work_completion)(&(&bond->mcast_work)->work) &base->lock &obj_hash[i].lock irq_context: 0 (wq_completion)bond95 (work_completion)(&(&slave->notify_work)->work) &rq->__lock irq_context: 0 (wq_completion)bond95 (work_completion)(&(&slave->notify_work)->work) &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)bond118 irq_context: 0 (wq_completion)bond118 (work_completion)(&(&slave->notify_work)->work) irq_context: 0 (wq_completion)bond118 (work_completion)(&(&slave->notify_work)->work) &rq->__lock irq_context: 0 (wq_completion)bond118 (work_completion)(&(&slave->notify_work)->work) &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)bond118 (work_completion)(&(&slave->notify_work)->work) &obj_hash[i].lock irq_context: 0 (wq_completion)bond118 (work_completion)(&(&slave->notify_work)->work) &base->lock irq_context: 0 (wq_completion)bond118 (work_completion)(&(&slave->notify_work)->work) &base->lock &obj_hash[i].lock irq_context: 0 (wq_completion)bond118 (work_completion)(&(&bond->mcast_work)->work) irq_context: 0 (wq_completion)bond118 (work_completion)(&(&bond->mcast_work)->work) &obj_hash[i].lock irq_context: 0 (wq_completion)bond118 (work_completion)(&(&bond->mcast_work)->work) &base->lock irq_context: 0 (wq_completion)bond118 (work_completion)(&(&bond->mcast_work)->work) &base->lock &obj_hash[i].lock irq_context: 0 (wq_completion)events (linkwatch_work).work rtnl_mutex rcu_read_lock fill_pool_map-wait-type-override rcu_node_0 irq_context: 0 (wq_completion)events (linkwatch_work).work rtnl_mutex rcu_read_lock fill_pool_map-wait-type-override &rq->__lock irq_context: 0 (wq_completion)events (linkwatch_work).work rtnl_mutex rcu_read_lock fill_pool_map-wait-type-override &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)mld (work_completion)(&(&idev->mc_dad_work)->work) &idev->mc_lock rcu_read_lock rcu_read_lock &____s->seqcount#12 irq_context: 0 (wq_completion)mld (work_completion)(&(&idev->mc_dad_work)->work) &idev->mc_lock rcu_read_lock rcu_read_lock &____s->seqcount#16 irq_context: 0 (wq_completion)mld (work_completion)(&(&idev->mc_dad_work)->work) &idev->mc_lock rcu_read_lock &obj_hash[i].lock irq_context: 0 (wq_completion)mld (work_completion)(&(&idev->mc_dad_work)->work) &idev->mc_lock rcu_read_lock &data->lock irq_context: 0 (wq_completion)bond89#2 irq_context: 0 (wq_completion)bond89#2 (work_completion)(&(&slave->notify_work)->work) irq_context: 0 (wq_completion)bond89#2 (work_completion)(&(&slave->notify_work)->work) &obj_hash[i].lock irq_context: 0 (wq_completion)bond89#2 (work_completion)(&(&slave->notify_work)->work) &base->lock irq_context: 0 (wq_completion)bond89#2 (work_completion)(&(&slave->notify_work)->work) &base->lock &obj_hash[i].lock irq_context: 0 (wq_completion)bond89#2 (work_completion)(&(&slave->notify_work)->work) &rq->__lock irq_context: 0 (wq_completion)bond89#2 (work_completion)(&(&bond->mcast_work)->work) irq_context: 0 (wq_completion)bond89#2 (work_completion)(&(&bond->mcast_work)->work) &obj_hash[i].lock irq_context: 0 (wq_completion)bond89#2 (work_completion)(&(&bond->mcast_work)->work) &base->lock irq_context: 0 (wq_completion)bond89#2 (work_completion)(&(&bond->mcast_work)->work) &base->lock &obj_hash[i].lock irq_context: 0 (wq_completion)bond113 (work_completion)(&(&slave->notify_work)->work) rtnl_mutex net_rwsem irq_context: 0 (wq_completion)bond113 (work_completion)(&(&slave->notify_work)->work) rtnl_mutex net_rwsem &list->lock#2 irq_context: 0 (wq_completion)bond113 (work_completion)(&(&slave->notify_work)->work) rtnl_mutex &tn->lock irq_context: 0 (wq_completion)bond113 (work_completion)(&(&slave->notify_work)->work) rtnl_mutex.wait_lock irq_context: 0 (wq_completion)bond113 (work_completion)(&(&slave->notify_work)->work) &p->pi_lock irq_context: 0 (wq_completion)bond114 (work_completion)(&(&slave->notify_work)->work) rtnl_mutex net_rwsem irq_context: 0 (wq_completion)bond114 (work_completion)(&(&slave->notify_work)->work) rtnl_mutex net_rwsem &list->lock#2 irq_context: 0 (wq_completion)bond114 (work_completion)(&(&slave->notify_work)->work) rtnl_mutex &tn->lock irq_context: 0 (wq_completion)bond114 (work_completion)(&(&slave->notify_work)->work) rtnl_mutex.wait_lock irq_context: 0 (wq_completion)bond114 (work_completion)(&(&slave->notify_work)->work) &p->pi_lock irq_context: 0 (wq_completion)bond115 (work_completion)(&(&slave->notify_work)->work) rtnl_mutex net_rwsem irq_context: 0 (wq_completion)bond115 (work_completion)(&(&slave->notify_work)->work) rtnl_mutex net_rwsem &list->lock#2 irq_context: 0 (wq_completion)bond115 (work_completion)(&(&slave->notify_work)->work) rtnl_mutex &tn->lock irq_context: 0 (wq_completion)bond115 (work_completion)(&(&slave->notify_work)->work) rtnl_mutex.wait_lock irq_context: 0 (wq_completion)bond115 (work_completion)(&(&slave->notify_work)->work) &p->pi_lock irq_context: 0 (wq_completion)bond80#2 (work_completion)(&(&slave->notify_work)->work) rtnl_mutex net_rwsem irq_context: 0 (wq_completion)bond80#2 (work_completion)(&(&slave->notify_work)->work) rtnl_mutex net_rwsem &list->lock#2 irq_context: 0 (wq_completion)bond80#2 (work_completion)(&(&slave->notify_work)->work) rtnl_mutex &tn->lock irq_context: 0 (wq_completion)bond80#2 (work_completion)(&(&slave->notify_work)->work) rtnl_mutex.wait_lock irq_context: 0 (wq_completion)bond80#2 (work_completion)(&(&slave->notify_work)->work) &p->pi_lock irq_context: 0 (wq_completion)bond105#2 (work_completion)(&(&slave->notify_work)->work) rtnl_mutex net_rwsem irq_context: 0 (wq_completion)bond105#2 (work_completion)(&(&slave->notify_work)->work) rtnl_mutex net_rwsem &list->lock#2 irq_context: 0 (wq_completion)bond105#2 (work_completion)(&(&slave->notify_work)->work) rtnl_mutex &tn->lock irq_context: 0 (wq_completion)bond105#2 (work_completion)(&(&slave->notify_work)->work) rtnl_mutex.wait_lock irq_context: 0 (wq_completion)bond105#2 (work_completion)(&(&slave->notify_work)->work) &p->pi_lock irq_context: 0 (wq_completion)bond14#2 (work_completion)(&(&slave->notify_work)->work) rtnl_mutex net_rwsem irq_context: 0 (wq_completion)bond14#2 (work_completion)(&(&slave->notify_work)->work) rtnl_mutex net_rwsem &list->lock#2 irq_context: 0 (wq_completion)bond14#2 (work_completion)(&(&slave->notify_work)->work) rtnl_mutex &tn->lock irq_context: 0 (wq_completion)bond14#2 (work_completion)(&(&slave->notify_work)->work) rtnl_mutex.wait_lock irq_context: 0 (wq_completion)bond14#2 (work_completion)(&(&slave->notify_work)->work) &p->pi_lock irq_context: 0 (wq_completion)bond103#2 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex fs_reclaim irq_context: 0 (wq_completion)bond103#2 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex fs_reclaim mmu_notifier_invalidate_range_start irq_context: 0 (wq_completion)bond103#2 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &c->lock irq_context: 0 (wq_completion)bond103#2 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex rcu_read_lock &bond->stats_lock/1 irq_context: 0 (wq_completion)bond103#2 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex rcu_read_lock &ndev->lock irq_context: 0 (wq_completion)bond103#2 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex nl_table_lock irq_context: 0 (wq_completion)bond103#2 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &obj_hash[i].lock irq_context: 0 (wq_completion)bond103#2 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex nl_table_wait.lock irq_context: 0 (wq_completion)bond103#2 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex net_rwsem irq_context: 0 (wq_completion)bond103#2 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex net_rwsem &list->lock#2 irq_context: 0 (wq_completion)bond103#2 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &tn->lock irq_context: 0 (wq_completion)bond103#2 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &idev->mc_lock irq_context: 0 (wq_completion)bond103#2 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &idev->mc_lock fs_reclaim irq_context: 0 (wq_completion)bond103#2 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &idev->mc_lock fs_reclaim mmu_notifier_invalidate_range_start irq_context: 0 (wq_completion)bond103#2 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &idev->mc_lock rcu_read_lock &ndev->lock irq_context: 0 (wq_completion)bond103#2 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &idev->mc_lock rcu_read_lock &dir->lock#2 irq_context: 0 (wq_completion)bond103#2 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &idev->mc_lock rcu_read_lock &ul->lock#2 irq_context: 0 (wq_completion)bond103#2 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &idev->mc_lock rcu_read_lock rcu_read_lock &____s->seqcount#10 irq_context: 0 (wq_completion)bond103#2 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &idev->mc_lock rcu_read_lock rcu_read_lock rcu_read_lock_bh rcu_read_lock rcu_read_lock_bh rcu_read_lock &br->multicast_lock irq_context: 0 (wq_completion)bond103#2 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &idev->mc_lock rcu_read_lock rcu_read_lock rcu_read_lock_bh rcu_read_lock rcu_read_lock_bh rcu_read_lock &br->multicast_lock &obj_hash[i].lock irq_context: 0 (wq_completion)bond103#2 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &idev->mc_lock rcu_read_lock rcu_read_lock rcu_read_lock_bh rcu_read_lock rcu_read_lock_bh rcu_read_lock &br->multicast_lock &base->lock irq_context: 0 (wq_completion)bond103#2 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &idev->mc_lock rcu_read_lock rcu_read_lock rcu_read_lock_bh rcu_read_lock rcu_read_lock_bh rcu_read_lock &data->lock irq_context: 0 (wq_completion)bond103#2 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &idev->mc_lock rcu_read_lock rcu_read_lock rcu_read_lock_bh rcu_read_lock rcu_read_lock_bh rcu_read_lock &obj_hash[i].lock irq_context: 0 (wq_completion)bond103#2 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex.wait_lock irq_context: 0 (wq_completion)bond103#2 (work_completion)(&(&bond->mcast_work)->work) &p->pi_lock irq_context: 0 (wq_completion)bond116 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex fs_reclaim irq_context: 0 (wq_completion)bond116 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex fs_reclaim mmu_notifier_invalidate_range_start irq_context: 0 (wq_completion)bond116 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &c->lock irq_context: 0 (wq_completion)bond116 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex rcu_read_lock &bond->stats_lock/1 irq_context: 0 (wq_completion)bond116 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex rcu_read_lock &ndev->lock irq_context: 0 (wq_completion)bond116 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex nl_table_lock irq_context: 0 (wq_completion)bond116 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &obj_hash[i].lock irq_context: 0 (wq_completion)bond116 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex nl_table_wait.lock irq_context: 0 (wq_completion)bond116 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex net_rwsem irq_context: 0 (wq_completion)bond116 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex net_rwsem &list->lock#2 irq_context: 0 (wq_completion)bond116 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &tn->lock irq_context: 0 (wq_completion)bond116 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &idev->mc_lock irq_context: 0 (wq_completion)bond116 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &idev->mc_lock &rq->__lock irq_context: 0 (wq_completion)bond116 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &idev->mc_lock fs_reclaim irq_context: 0 (wq_completion)bond116 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &idev->mc_lock fs_reclaim mmu_notifier_invalidate_range_start irq_context: 0 (wq_completion)bond116 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &idev->mc_lock rcu_read_lock &ndev->lock irq_context: 0 (wq_completion)bond116 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &idev->mc_lock rcu_read_lock &dir->lock#2 irq_context: 0 (wq_completion)bond116 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &idev->mc_lock rcu_read_lock &ul->lock#2 irq_context: 0 (wq_completion)bond116 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &idev->mc_lock rcu_read_lock rcu_read_lock &____s->seqcount#10 irq_context: 0 (wq_completion)bond116 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &idev->mc_lock rcu_read_lock rcu_read_lock rcu_read_lock_bh rcu_read_lock rcu_read_lock_bh rcu_read_lock &br->multicast_lock irq_context: 0 (wq_completion)bond116 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &idev->mc_lock rcu_read_lock rcu_read_lock rcu_read_lock_bh rcu_read_lock rcu_read_lock_bh rcu_read_lock &br->multicast_lock &obj_hash[i].lock irq_context: 0 (wq_completion)bond116 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &idev->mc_lock rcu_read_lock rcu_read_lock rcu_read_lock_bh rcu_read_lock rcu_read_lock_bh rcu_read_lock &br->multicast_lock &base->lock irq_context: 0 (wq_completion)bond116 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &idev->mc_lock rcu_read_lock rcu_read_lock rcu_read_lock_bh rcu_read_lock rcu_read_lock_bh rcu_read_lock &data->lock irq_context: 0 (wq_completion)bond116 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &idev->mc_lock rcu_read_lock rcu_read_lock rcu_read_lock_bh rcu_read_lock rcu_read_lock_bh rcu_read_lock &obj_hash[i].lock irq_context: 0 (wq_completion)bond116 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex.wait_lock irq_context: 0 (wq_completion)bond116 (work_completion)(&(&bond->mcast_work)->work) &p->pi_lock irq_context: 0 (wq_completion)bond116 (work_completion)(&(&bond->mcast_work)->work) &p->pi_lock &rq->__lock irq_context: 0 (wq_completion)bond116 (work_completion)(&(&bond->mcast_work)->work) &p->pi_lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)bond116 (work_completion)(&(&bond->mcast_work)->work) &rq->__lock irq_context: 0 (wq_completion)bond106#2 irq_context: 0 (wq_completion)bond106#2 (work_completion)(&(&slave->notify_work)->work) irq_context: 0 (wq_completion)bond106#2 (work_completion)(&(&slave->notify_work)->work) &obj_hash[i].lock irq_context: 0 (wq_completion)bond106#2 (work_completion)(&(&slave->notify_work)->work) &base->lock irq_context: 0 (wq_completion)bond106#2 (work_completion)(&(&slave->notify_work)->work) &base->lock &obj_hash[i].lock irq_context: 0 (wq_completion)bond106#2 (work_completion)(&(&slave->notify_work)->work) &rq->__lock irq_context: 0 (wq_completion)bond106#2 (work_completion)(&(&slave->notify_work)->work) &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)bond106#2 (work_completion)(&(&bond->mcast_work)->work) irq_context: 0 (wq_completion)bond106#2 (work_completion)(&(&bond->mcast_work)->work) &obj_hash[i].lock irq_context: 0 (wq_completion)bond106#2 (work_completion)(&(&bond->mcast_work)->work) &base->lock irq_context: 0 (wq_completion)bond106#2 (work_completion)(&(&bond->mcast_work)->work) &base->lock &obj_hash[i].lock irq_context: 0 (wq_completion)bond106#2 (work_completion)(&(&slave->notify_work)->work) fill_pool_map-wait-type-override pool_lock#2 irq_context: 0 (wq_completion)bond106#2 (work_completion)(&(&slave->notify_work)->work) fill_pool_map-wait-type-override pool_lock irq_context: 0 (wq_completion)bond119 irq_context: 0 (wq_completion)bond119 (work_completion)(&(&slave->notify_work)->work) irq_context: 0 (wq_completion)bond119 (work_completion)(&(&slave->notify_work)->work) &obj_hash[i].lock irq_context: 0 (wq_completion)bond119 (work_completion)(&(&slave->notify_work)->work) &base->lock irq_context: 0 (wq_completion)bond119 (work_completion)(&(&slave->notify_work)->work) &base->lock &obj_hash[i].lock irq_context: 0 (wq_completion)bond119 (work_completion)(&(&slave->notify_work)->work) &rq->__lock irq_context: 0 (wq_completion)bond119 (work_completion)(&(&bond->mcast_work)->work) irq_context: 0 (wq_completion)bond119 (work_completion)(&(&bond->mcast_work)->work) &obj_hash[i].lock irq_context: 0 (wq_completion)bond119 (work_completion)(&(&bond->mcast_work)->work) &base->lock irq_context: 0 (wq_completion)bond119 (work_completion)(&(&bond->mcast_work)->work) &base->lock &obj_hash[i].lock irq_context: 0 (wq_completion)bond88#2 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex fs_reclaim irq_context: 0 (wq_completion)bond88#2 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex fs_reclaim mmu_notifier_invalidate_range_start irq_context: 0 (wq_completion)bond88#2 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &c->lock irq_context: 0 (wq_completion)bond88#2 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &n->list_lock irq_context: 0 (wq_completion)bond88#2 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &n->list_lock &c->lock irq_context: 0 (wq_completion)bond88#2 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex rcu_read_lock &bond->stats_lock/1 irq_context: 0 (wq_completion)bond88#2 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex rcu_read_lock &ndev->lock irq_context: 0 (wq_completion)bond88#2 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex nl_table_lock irq_context: 0 (wq_completion)bond88#2 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &obj_hash[i].lock irq_context: 0 (wq_completion)bond88#2 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex nl_table_wait.lock irq_context: 0 (wq_completion)bond88#2 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex net_rwsem irq_context: 0 (wq_completion)bond88#2 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex net_rwsem &list->lock#2 irq_context: 0 (wq_completion)bond88#2 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &tn->lock irq_context: 0 (wq_completion)bond88#2 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &idev->mc_lock irq_context: 0 (wq_completion)bond88#2 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &idev->mc_lock fs_reclaim irq_context: 0 (wq_completion)bond88#2 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &idev->mc_lock fs_reclaim mmu_notifier_invalidate_range_start irq_context: 0 (wq_completion)bond88#2 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &idev->mc_lock rcu_read_lock &ndev->lock irq_context: 0 (wq_completion)bond88#2 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &idev->mc_lock rcu_read_lock &dir->lock#2 irq_context: 0 (wq_completion)bond88#2 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &idev->mc_lock rcu_read_lock &c->lock irq_context: 0 (wq_completion)bond88#2 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &idev->mc_lock rcu_read_lock &ul->lock#2 irq_context: 0 (wq_completion)bond88#2 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &idev->mc_lock rcu_read_lock rcu_read_lock &____s->seqcount#10 irq_context: 0 (wq_completion)bond88#2 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &idev->mc_lock rcu_read_lock rcu_read_lock rcu_read_lock_bh rcu_read_lock rcu_read_lock_bh rcu_read_lock &br->multicast_lock irq_context: 0 (wq_completion)bond88#2 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &idev->mc_lock rcu_read_lock rcu_read_lock rcu_read_lock_bh rcu_read_lock rcu_read_lock_bh rcu_read_lock &br->multicast_lock &obj_hash[i].lock irq_context: 0 (wq_completion)bond88#2 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &idev->mc_lock rcu_read_lock rcu_read_lock rcu_read_lock_bh rcu_read_lock rcu_read_lock_bh rcu_read_lock &br->multicast_lock &base->lock irq_context: 0 (wq_completion)bond88#2 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &idev->mc_lock rcu_read_lock rcu_read_lock rcu_read_lock_bh rcu_read_lock rcu_read_lock_bh rcu_read_lock &data->lock irq_context: 0 (wq_completion)bond88#2 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &idev->mc_lock rcu_read_lock rcu_read_lock rcu_read_lock_bh rcu_read_lock rcu_read_lock_bh rcu_read_lock &obj_hash[i].lock irq_context: 0 (wq_completion)bond88#2 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex.wait_lock irq_context: 0 (wq_completion)bond88#2 (work_completion)(&(&bond->mcast_work)->work) &p->pi_lock irq_context: 0 (wq_completion)bond15#3 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex fs_reclaim irq_context: 0 (wq_completion)bond15#3 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex fs_reclaim mmu_notifier_invalidate_range_start irq_context: 0 (wq_completion)bond15#3 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &c->lock irq_context: 0 (wq_completion)bond15#3 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex rcu_read_lock &bond->stats_lock/1 irq_context: 0 (wq_completion)bond15#3 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex rcu_read_lock &ndev->lock irq_context: 0 (wq_completion)bond15#3 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex nl_table_lock irq_context: 0 (wq_completion)bond15#3 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &obj_hash[i].lock irq_context: 0 (wq_completion)bond15#3 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex nl_table_wait.lock irq_context: 0 (wq_completion)bond15#3 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex net_rwsem irq_context: 0 (wq_completion)bond15#3 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex net_rwsem &list->lock#2 irq_context: 0 (wq_completion)bond15#3 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &tn->lock irq_context: 0 (wq_completion)bond15#3 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &idev->mc_lock irq_context: 0 (wq_completion)bond15#3 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &idev->mc_lock fs_reclaim irq_context: 0 (wq_completion)bond15#3 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &idev->mc_lock fs_reclaim mmu_notifier_invalidate_range_start irq_context: 0 (wq_completion)bond15#3 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &idev->mc_lock &c->lock irq_context: 0 (wq_completion)bond15#3 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &idev->mc_lock rcu_read_lock &ndev->lock irq_context: 0 (wq_completion)bond15#3 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &idev->mc_lock rcu_read_lock &dir->lock#2 irq_context: 0 (wq_completion)bond15#3 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &idev->mc_lock rcu_read_lock &c->lock irq_context: 0 (wq_completion)bond15#3 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &idev->mc_lock rcu_read_lock &ul->lock#2 irq_context: 0 (wq_completion)bond15#3 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &idev->mc_lock rcu_read_lock rcu_read_lock &____s->seqcount#12 irq_context: 0 (wq_completion)bond15#3 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &idev->mc_lock rcu_read_lock rcu_read_lock &____s->seqcount#16 irq_context: 0 (wq_completion)bond15#3 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &idev->mc_lock rcu_read_lock &obj_hash[i].lock irq_context: 0 (wq_completion)bond15#3 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &idev->mc_lock rcu_read_lock &data->lock irq_context: 0 (wq_completion)bond15#3 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex.wait_lock irq_context: 0 (wq_completion)bond15#3 (work_completion)(&(&bond->mcast_work)->work) &p->pi_lock irq_context: 0 (wq_completion)bond90 irq_context: 0 (wq_completion)bond90 (work_completion)(&(&slave->notify_work)->work) irq_context: 0 (wq_completion)bond90 (work_completion)(&(&slave->notify_work)->work) &obj_hash[i].lock irq_context: 0 (wq_completion)bond90 (work_completion)(&(&slave->notify_work)->work) &base->lock irq_context: 0 (wq_completion)bond90 (work_completion)(&(&slave->notify_work)->work) &base->lock &obj_hash[i].lock irq_context: 0 (wq_completion)bond90 (work_completion)(&(&slave->notify_work)->work) &rq->__lock irq_context: 0 (wq_completion)bond90 (work_completion)(&(&bond->mcast_work)->work) irq_context: 0 (wq_completion)bond90 (work_completion)(&(&bond->mcast_work)->work) &obj_hash[i].lock irq_context: 0 (wq_completion)bond90 (work_completion)(&(&bond->mcast_work)->work) &base->lock irq_context: 0 (wq_completion)bond90 (work_completion)(&(&bond->mcast_work)->work) &base->lock &obj_hash[i].lock irq_context: 0 (wq_completion)events (linkwatch_work).work rtnl_mutex remove_cache_srcu &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)bond107#2 irq_context: 0 (wq_completion)bond107#2 (work_completion)(&(&slave->notify_work)->work) irq_context: 0 (wq_completion)bond107#2 (work_completion)(&(&slave->notify_work)->work) &obj_hash[i].lock irq_context: 0 (wq_completion)bond107#2 (work_completion)(&(&slave->notify_work)->work) &base->lock irq_context: 0 (wq_completion)bond107#2 (work_completion)(&(&slave->notify_work)->work) &base->lock &obj_hash[i].lock irq_context: 0 (wq_completion)bond107#2 (work_completion)(&(&slave->notify_work)->work) &rq->__lock irq_context: 0 (wq_completion)bond107#2 (work_completion)(&(&slave->notify_work)->work) &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)bond107#2 (work_completion)(&(&bond->mcast_work)->work) irq_context: 0 (wq_completion)bond107#2 (work_completion)(&(&bond->mcast_work)->work) &obj_hash[i].lock irq_context: 0 (wq_completion)bond107#2 (work_completion)(&(&bond->mcast_work)->work) &base->lock irq_context: 0 (wq_completion)bond107#2 (work_completion)(&(&bond->mcast_work)->work) &base->lock &obj_hash[i].lock irq_context: 0 (wq_completion)bond120 irq_context: 0 (wq_completion)bond120 (work_completion)(&(&slave->notify_work)->work) irq_context: 0 (wq_completion)bond120 (work_completion)(&(&slave->notify_work)->work) &obj_hash[i].lock irq_context: 0 (wq_completion)bond120 (work_completion)(&(&slave->notify_work)->work) &base->lock irq_context: 0 (wq_completion)bond120 (work_completion)(&(&slave->notify_work)->work) &base->lock &obj_hash[i].lock irq_context: 0 (wq_completion)bond120 (work_completion)(&(&bond->mcast_work)->work) irq_context: 0 (wq_completion)bond120 (work_completion)(&(&bond->mcast_work)->work) &obj_hash[i].lock irq_context: 0 (wq_completion)bond120 (work_completion)(&(&bond->mcast_work)->work) &base->lock irq_context: 0 (wq_completion)bond120 (work_completion)(&(&bond->mcast_work)->work) &base->lock &obj_hash[i].lock irq_context: 0 (wq_completion)bond115 (work_completion)(&(&bond->mcast_work)->work) &rq->__lock irq_context: 0 (wq_completion)bond115 (work_completion)(&(&bond->mcast_work)->work) &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)bond91 irq_context: 0 (wq_completion)bond91 (work_completion)(&(&slave->notify_work)->work) irq_context: 0 (wq_completion)bond91 (work_completion)(&(&slave->notify_work)->work) &obj_hash[i].lock irq_context: 0 (wq_completion)bond91 (work_completion)(&(&slave->notify_work)->work) &base->lock irq_context: 0 (wq_completion)bond91 (work_completion)(&(&slave->notify_work)->work) &base->lock &obj_hash[i].lock irq_context: 0 (wq_completion)bond91 (work_completion)(&(&slave->notify_work)->work) &rq->__lock irq_context: 0 (wq_completion)bond91 (work_completion)(&(&slave->notify_work)->work) &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)bond78 (work_completion)(&(&bond->mcast_work)->work) &rq->__lock irq_context: 0 (wq_completion)bond78 (work_completion)(&(&bond->mcast_work)->work) &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)bond120 (work_completion)(&(&bond->mcast_work)->work) &rq->__lock irq_context: 0 (wq_completion)bond120 (work_completion)(&(&bond->mcast_work)->work) &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)bond91 (work_completion)(&(&bond->mcast_work)->work) irq_context: 0 (wq_completion)bond91 (work_completion)(&(&bond->mcast_work)->work) &obj_hash[i].lock irq_context: 0 (wq_completion)bond91 (work_completion)(&(&bond->mcast_work)->work) &base->lock irq_context: 0 (wq_completion)bond91 (work_completion)(&(&bond->mcast_work)->work) &base->lock &obj_hash[i].lock irq_context: 0 (wq_completion)bond18#3 irq_context: 0 (wq_completion)bond18#3 (work_completion)(&(&slave->notify_work)->work) irq_context: 0 (wq_completion)bond18#3 (work_completion)(&(&slave->notify_work)->work) &obj_hash[i].lock irq_context: 0 (wq_completion)bond18#3 (work_completion)(&(&slave->notify_work)->work) &base->lock irq_context: 0 (wq_completion)bond18#3 (work_completion)(&(&slave->notify_work)->work) &base->lock &obj_hash[i].lock irq_context: 0 (wq_completion)ipv6_addrconf (work_completion)(&(&ifa->dad_work)->work) rtnl_mutex rcu_read_lock &c->lock irq_context: 0 (wq_completion)ipv6_addrconf (work_completion)(&(&ifa->dad_work)->work) rtnl_mutex rcu_read_lock &____s->seqcount#2 irq_context: 0 (wq_completion)ipv6_addrconf (work_completion)(&(&ifa->dad_work)->work) rtnl_mutex rcu_read_lock &____s->seqcount irq_context: 0 (wq_completion)bond18#3 (work_completion)(&(&bond->mcast_work)->work) irq_context: 0 (wq_completion)bond18#3 (work_completion)(&(&bond->mcast_work)->work) &obj_hash[i].lock irq_context: 0 (wq_completion)bond18#3 (work_completion)(&(&bond->mcast_work)->work) &base->lock irq_context: 0 (wq_completion)bond18#3 (work_completion)(&(&bond->mcast_work)->work) &base->lock &obj_hash[i].lock irq_context: 0 (wq_completion)bond18#3 (work_completion)(&(&bond->mcast_work)->work) &rq->__lock irq_context: 0 (wq_completion)bond18#3 (work_completion)(&(&bond->mcast_work)->work) &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)mld (work_completion)(&(&idev->mc_ifc_work)->work) &idev->mc_lock rcu_read_lock &meta->lock irq_context: 0 (wq_completion)bond90 (work_completion)(&(&slave->notify_work)->work) rtnl_mutex net_rwsem irq_context: 0 (wq_completion)bond90 (work_completion)(&(&slave->notify_work)->work) rtnl_mutex net_rwsem &list->lock#2 irq_context: 0 (wq_completion)bond90 (work_completion)(&(&slave->notify_work)->work) rtnl_mutex &tn->lock irq_context: 0 (wq_completion)bond90 (work_completion)(&(&slave->notify_work)->work) rtnl_mutex.wait_lock irq_context: 0 (wq_completion)bond114 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex fs_reclaim irq_context: 0 (wq_completion)bond90 (work_completion)(&(&slave->notify_work)->work) &p->pi_lock irq_context: 0 (wq_completion)bond114 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex fs_reclaim mmu_notifier_invalidate_range_start irq_context: 0 (wq_completion)bond114 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &c->lock irq_context: 0 (wq_completion)bond114 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &rq->__lock irq_context: 0 (wq_completion)bond114 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)bond114 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex rcu_read_lock &bond->stats_lock/1 irq_context: 0 (wq_completion)bond114 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex rcu_read_lock &ndev->lock irq_context: 0 (wq_completion)bond114 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex nl_table_lock irq_context: 0 (wq_completion)bond114 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &obj_hash[i].lock irq_context: 0 (wq_completion)bond114 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex nl_table_wait.lock irq_context: 0 (wq_completion)bond114 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex net_rwsem irq_context: 0 (wq_completion)bond114 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex net_rwsem &list->lock#2 irq_context: 0 (wq_completion)bond114 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex net_rwsem &rq->__lock irq_context: 0 (wq_completion)bond114 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex net_rwsem &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)bond114 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &tn->lock irq_context: 0 (wq_completion)bond114 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &idev->mc_lock irq_context: 0 (wq_completion)bond114 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &idev->mc_lock fs_reclaim irq_context: 0 (wq_completion)bond114 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &idev->mc_lock fs_reclaim mmu_notifier_invalidate_range_start irq_context: 0 (wq_completion)bond114 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &idev->mc_lock &c->lock irq_context: 0 (wq_completion)bond114 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &idev->mc_lock rcu_read_lock &ndev->lock irq_context: 0 (wq_completion)bond114 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &idev->mc_lock rcu_read_lock &dir->lock#2 irq_context: 0 (wq_completion)bond114 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &idev->mc_lock rcu_read_lock &c->lock irq_context: 0 (wq_completion)bond114 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &idev->mc_lock rcu_read_lock &ul->lock#2 irq_context: 0 (wq_completion)bond114 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &idev->mc_lock rcu_read_lock rcu_read_lock &____s->seqcount#10 irq_context: 0 (wq_completion)bond114 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &idev->mc_lock rcu_read_lock rcu_read_lock rcu_read_lock_bh rcu_read_lock rcu_read_lock_bh rcu_read_lock &br->multicast_lock irq_context: 0 (wq_completion)bond114 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &idev->mc_lock rcu_read_lock rcu_read_lock rcu_read_lock_bh rcu_read_lock rcu_read_lock_bh rcu_read_lock &br->multicast_lock &obj_hash[i].lock irq_context: 0 (wq_completion)bond114 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &idev->mc_lock rcu_read_lock rcu_read_lock rcu_read_lock_bh rcu_read_lock rcu_read_lock_bh rcu_read_lock &br->multicast_lock &base->lock irq_context: 0 (wq_completion)bond114 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &idev->mc_lock rcu_read_lock rcu_read_lock rcu_read_lock_bh rcu_read_lock rcu_read_lock_bh rcu_read_lock &data->lock irq_context: 0 (wq_completion)bond114 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &idev->mc_lock rcu_read_lock rcu_read_lock rcu_read_lock_bh rcu_read_lock rcu_read_lock_bh rcu_read_lock &obj_hash[i].lock irq_context: 0 (wq_completion)bond114 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &idev->mc_lock rcu_read_lock rcu_read_lock rcu_node_0 irq_context: 0 (wq_completion)bond114 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &idev->mc_lock rcu_read_lock rcu_read_lock &rq->__lock irq_context: 0 (wq_completion)bond114 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &idev->mc_lock rcu_read_lock rcu_node_0 irq_context: 0 (wq_completion)bond114 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex.wait_lock irq_context: 0 (wq_completion)bond114 (work_completion)(&(&bond->mcast_work)->work) &p->pi_lock irq_context: 0 (wq_completion)bond90 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex fs_reclaim irq_context: 0 (wq_completion)bond90 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex fs_reclaim mmu_notifier_invalidate_range_start irq_context: 0 (wq_completion)bond90 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &c->lock irq_context: 0 (wq_completion)bond90 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex rcu_read_lock &bond->stats_lock/1 irq_context: 0 (wq_completion)bond90 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex rcu_read_lock &ndev->lock irq_context: 0 (wq_completion)bond90 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex nl_table_lock irq_context: 0 (wq_completion)bond90 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &obj_hash[i].lock irq_context: 0 (wq_completion)bond90 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex nl_table_wait.lock irq_context: 0 (wq_completion)bond90 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex net_rwsem irq_context: 0 (wq_completion)bond90 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex net_rwsem &list->lock#2 irq_context: 0 (wq_completion)bond90 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &tn->lock irq_context: 0 (wq_completion)bond90 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &idev->mc_lock irq_context: 0 (wq_completion)bond90 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex.wait_lock irq_context: 0 (wq_completion)bond90 (work_completion)(&(&bond->mcast_work)->work) &p->pi_lock irq_context: 0 (wq_completion)bond79#2 (work_completion)(&(&slave->notify_work)->work) rtnl_mutex net_rwsem irq_context: 0 (wq_completion)bond79#2 (work_completion)(&(&slave->notify_work)->work) rtnl_mutex net_rwsem &list->lock#2 irq_context: 0 (wq_completion)bond79#2 (work_completion)(&(&slave->notify_work)->work) rtnl_mutex &tn->lock irq_context: 0 (wq_completion)bond79#2 (work_completion)(&(&slave->notify_work)->work) rtnl_mutex.wait_lock irq_context: 0 (wq_completion)bond79#2 (work_completion)(&(&slave->notify_work)->work) &p->pi_lock irq_context: 0 (wq_completion)bond89#2 (work_completion)(&(&slave->notify_work)->work) rtnl_mutex net_rwsem irq_context: 0 (wq_completion)bond89#2 (work_completion)(&(&slave->notify_work)->work) rtnl_mutex net_rwsem &list->lock#2 irq_context: 0 (wq_completion)bond89#2 (work_completion)(&(&slave->notify_work)->work) rtnl_mutex &tn->lock irq_context: 0 (wq_completion)bond89#2 (work_completion)(&(&slave->notify_work)->work) rtnl_mutex.wait_lock irq_context: 0 (wq_completion)bond89#2 (work_completion)(&(&slave->notify_work)->work) &p->pi_lock irq_context: 0 (wq_completion)bond101#2 (work_completion)(&(&slave->notify_work)->work) rtnl_mutex net_rwsem irq_context: 0 (wq_completion)bond101#2 (work_completion)(&(&slave->notify_work)->work) rtnl_mutex net_rwsem &list->lock#2 irq_context: 0 (wq_completion)bond101#2 (work_completion)(&(&slave->notify_work)->work) rtnl_mutex &tn->lock irq_context: 0 (wq_completion)bond101#2 (work_completion)(&(&slave->notify_work)->work) rtnl_mutex.wait_lock irq_context: 0 (wq_completion)bond101#2 (work_completion)(&(&slave->notify_work)->work) &p->pi_lock irq_context: 0 (wq_completion)bond119 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex fs_reclaim irq_context: 0 (wq_completion)bond119 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex fs_reclaim mmu_notifier_invalidate_range_start irq_context: 0 (wq_completion)bond119 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex rcu_read_lock &bond->stats_lock/1 irq_context: 0 (wq_completion)bond119 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex rcu_read_lock &ndev->lock irq_context: 0 (wq_completion)bond119 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex nl_table_lock irq_context: 0 (wq_completion)bond119 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &obj_hash[i].lock irq_context: 0 (wq_completion)bond119 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex nl_table_wait.lock irq_context: 0 (wq_completion)bond119 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex net_rwsem irq_context: 0 (wq_completion)bond119 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex net_rwsem &list->lock#2 irq_context: 0 (wq_completion)bond119 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &tn->lock irq_context: 0 (wq_completion)bond119 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &idev->mc_lock irq_context: 0 (wq_completion)bond119 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex.wait_lock irq_context: 0 (wq_completion)bond119 (work_completion)(&(&bond->mcast_work)->work) &p->pi_lock irq_context: 0 (wq_completion)bond120 (work_completion)(&(&slave->notify_work)->work) rtnl_mutex net_rwsem irq_context: 0 (wq_completion)bond120 (work_completion)(&(&slave->notify_work)->work) rtnl_mutex net_rwsem &list->lock#2 irq_context: 0 (wq_completion)bond120 (work_completion)(&(&slave->notify_work)->work) rtnl_mutex &tn->lock irq_context: 0 (wq_completion)bond120 (work_completion)(&(&slave->notify_work)->work) rtnl_mutex.wait_lock irq_context: 0 (wq_completion)bond120 (work_completion)(&(&slave->notify_work)->work) &p->pi_lock irq_context: 0 (wq_completion)bond18#3 (work_completion)(&(&slave->notify_work)->work) rtnl_mutex net_rwsem irq_context: 0 (wq_completion)bond18#3 (work_completion)(&(&slave->notify_work)->work) rtnl_mutex net_rwsem &list->lock#2 irq_context: 0 (wq_completion)bond18#3 (work_completion)(&(&slave->notify_work)->work) rtnl_mutex &tn->lock irq_context: 0 (wq_completion)bond18#3 (work_completion)(&(&slave->notify_work)->work) rtnl_mutex.wait_lock irq_context: 0 (wq_completion)bond18#3 (work_completion)(&(&slave->notify_work)->work) &p->pi_lock irq_context: 0 (wq_completion)bond91 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex fs_reclaim irq_context: 0 (wq_completion)bond91 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex fs_reclaim mmu_notifier_invalidate_range_start irq_context: 0 (wq_completion)bond91 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex rcu_read_lock &bond->stats_lock/1 irq_context: 0 (wq_completion)bond91 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex rcu_read_lock &ndev->lock irq_context: 0 (wq_completion)bond91 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex nl_table_lock irq_context: 0 (wq_completion)bond91 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &obj_hash[i].lock irq_context: 0 (wq_completion)bond91 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex nl_table_wait.lock irq_context: 0 (wq_completion)bond91 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex net_rwsem irq_context: 0 (wq_completion)bond91 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex net_rwsem &list->lock#2 irq_context: 0 (wq_completion)bond91 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &tn->lock irq_context: 0 (wq_completion)bond91 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &idev->mc_lock irq_context: 0 (wq_completion)bond91 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex.wait_lock irq_context: 0 (wq_completion)bond91 (work_completion)(&(&bond->mcast_work)->work) &p->pi_lock irq_context: 0 (wq_completion)bond91 (work_completion)(&(&slave->notify_work)->work) rtnl_mutex net_rwsem irq_context: 0 (wq_completion)bond91 (work_completion)(&(&slave->notify_work)->work) rtnl_mutex net_rwsem &list->lock#2 irq_context: 0 (wq_completion)bond91 (work_completion)(&(&slave->notify_work)->work) rtnl_mutex &tn->lock irq_context: 0 (wq_completion)bond91 (work_completion)(&(&slave->notify_work)->work) rtnl_mutex.wait_lock irq_context: 0 (wq_completion)bond91 (work_completion)(&(&slave->notify_work)->work) &p->pi_lock irq_context: 0 (wq_completion)bond108#2 irq_context: 0 (wq_completion)bond108#2 (work_completion)(&(&slave->notify_work)->work) irq_context: 0 (wq_completion)bond108#2 (work_completion)(&(&slave->notify_work)->work) &rq->__lock irq_context: 0 (wq_completion)bond108#2 (work_completion)(&(&slave->notify_work)->work) &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)bond108#2 (work_completion)(&(&slave->notify_work)->work) &obj_hash[i].lock irq_context: 0 (wq_completion)bond108#2 (work_completion)(&(&slave->notify_work)->work) &base->lock irq_context: 0 (wq_completion)bond108#2 (work_completion)(&(&slave->notify_work)->work) &base->lock &obj_hash[i].lock irq_context: 0 (wq_completion)bond108#2 (work_completion)(&(&bond->mcast_work)->work) irq_context: 0 (wq_completion)bond108#2 (work_completion)(&(&bond->mcast_work)->work) &obj_hash[i].lock irq_context: 0 (wq_completion)bond108#2 (work_completion)(&(&bond->mcast_work)->work) &base->lock irq_context: 0 (wq_completion)bond108#2 (work_completion)(&(&bond->mcast_work)->work) &base->lock &obj_hash[i].lock irq_context: 0 (wq_completion)bond121 irq_context: 0 (wq_completion)bond121 (work_completion)(&(&slave->notify_work)->work) irq_context: 0 (wq_completion)bond121 (work_completion)(&(&slave->notify_work)->work) &obj_hash[i].lock irq_context: 0 (wq_completion)bond121 (work_completion)(&(&slave->notify_work)->work) &base->lock irq_context: 0 (wq_completion)bond121 (work_completion)(&(&slave->notify_work)->work) &base->lock &obj_hash[i].lock irq_context: 0 (wq_completion)bond121 (work_completion)(&(&slave->notify_work)->work) &rq->__lock irq_context: 0 (wq_completion)bond121 (work_completion)(&(&slave->notify_work)->work) &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)bond121 (work_completion)(&(&bond->mcast_work)->work) irq_context: 0 (wq_completion)bond121 (work_completion)(&(&bond->mcast_work)->work) &obj_hash[i].lock irq_context: 0 (wq_completion)bond121 (work_completion)(&(&bond->mcast_work)->work) &base->lock irq_context: 0 (wq_completion)bond121 (work_completion)(&(&bond->mcast_work)->work) &base->lock &obj_hash[i].lock irq_context: 0 (wq_completion)bond106#2 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex fs_reclaim irq_context: 0 (wq_completion)bond106#2 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex fs_reclaim mmu_notifier_invalidate_range_start irq_context: 0 (wq_completion)bond106#2 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &c->lock irq_context: 0 (wq_completion)bond106#2 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex rcu_read_lock &bond->stats_lock/1 irq_context: 0 (wq_completion)bond106#2 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex rcu_read_lock &ndev->lock irq_context: 0 (wq_completion)bond106#2 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex nl_table_lock irq_context: 0 (wq_completion)bond106#2 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &obj_hash[i].lock irq_context: 0 (wq_completion)bond106#2 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex nl_table_wait.lock irq_context: 0 (wq_completion)bond106#2 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex net_rwsem irq_context: 0 (wq_completion)bond106#2 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex net_rwsem &list->lock#2 irq_context: 0 (wq_completion)bond106#2 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &tn->lock irq_context: 0 (wq_completion)bond106#2 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &idev->mc_lock irq_context: 0 (wq_completion)bond106#2 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &idev->mc_lock fs_reclaim irq_context: 0 (wq_completion)bond106#2 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &idev->mc_lock fs_reclaim mmu_notifier_invalidate_range_start irq_context: 0 (wq_completion)bond106#2 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &idev->mc_lock &c->lock irq_context: 0 (wq_completion)bond106#2 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &idev->mc_lock &n->list_lock irq_context: 0 (wq_completion)bond106#2 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &idev->mc_lock &n->list_lock &c->lock irq_context: 0 (wq_completion)bond106#2 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &idev->mc_lock rcu_read_lock &ndev->lock irq_context: 0 (wq_completion)bond106#2 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &idev->mc_lock rcu_read_lock &dir->lock#2 irq_context: 0 (wq_completion)bond106#2 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &idev->mc_lock rcu_read_lock &ul->lock#2 irq_context: 0 (wq_completion)bond106#2 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &idev->mc_lock rcu_read_lock rcu_read_lock &____s->seqcount#10 irq_context: 0 (wq_completion)bond106#2 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &idev->mc_lock rcu_read_lock rcu_read_lock rcu_read_lock_bh rcu_read_lock rcu_read_lock_bh rcu_read_lock &br->multicast_lock irq_context: 0 (wq_completion)bond106#2 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &idev->mc_lock rcu_read_lock rcu_read_lock rcu_read_lock_bh rcu_read_lock rcu_read_lock_bh rcu_read_lock &br->multicast_lock &obj_hash[i].lock irq_context: 0 (wq_completion)bond106#2 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &idev->mc_lock rcu_read_lock rcu_read_lock rcu_read_lock_bh rcu_read_lock rcu_read_lock_bh rcu_read_lock &br->multicast_lock &base->lock irq_context: 0 (wq_completion)bond106#2 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &idev->mc_lock rcu_read_lock rcu_read_lock rcu_read_lock_bh rcu_read_lock rcu_read_lock_bh rcu_read_lock &data->lock irq_context: 0 (wq_completion)bond106#2 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &idev->mc_lock rcu_read_lock rcu_read_lock rcu_read_lock_bh rcu_read_lock rcu_read_lock_bh rcu_read_lock &obj_hash[i].lock irq_context: 0 (wq_completion)bond106#2 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex.wait_lock irq_context: 0 (wq_completion)bond106#2 (work_completion)(&(&bond->mcast_work)->work) &p->pi_lock irq_context: 0 (wq_completion)bond117 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex fs_reclaim irq_context: 0 (wq_completion)bond117 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex fs_reclaim mmu_notifier_invalidate_range_start irq_context: 0 (wq_completion)bond117 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &c->lock irq_context: 0 (wq_completion)bond117 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &n->list_lock irq_context: 0 (wq_completion)bond117 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &n->list_lock &c->lock irq_context: 0 (wq_completion)bond117 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex rcu_read_lock &bond->stats_lock/1 irq_context: 0 (wq_completion)bond117 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex rcu_read_lock &ndev->lock irq_context: 0 (wq_completion)bond117 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex nl_table_lock irq_context: 0 (wq_completion)bond117 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &obj_hash[i].lock irq_context: 0 (wq_completion)bond117 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex nl_table_wait.lock irq_context: 0 (wq_completion)bond117 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex net_rwsem irq_context: 0 (wq_completion)bond117 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex net_rwsem &list->lock#2 irq_context: 0 (wq_completion)bond117 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &tn->lock irq_context: 0 (wq_completion)bond117 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &idev->mc_lock irq_context: 0 (wq_completion)bond117 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &idev->mc_lock fs_reclaim irq_context: 0 (wq_completion)bond117 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &idev->mc_lock fs_reclaim mmu_notifier_invalidate_range_start irq_context: 0 (wq_completion)bond117 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &idev->mc_lock rcu_read_lock &ndev->lock irq_context: 0 (wq_completion)bond117 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &idev->mc_lock rcu_read_lock &c->lock irq_context: 0 (wq_completion)bond117 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &idev->mc_lock rcu_read_lock &dir->lock#2 irq_context: 0 (wq_completion)bond117 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &idev->mc_lock rcu_read_lock &ul->lock#2 irq_context: 0 (wq_completion)bond117 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &idev->mc_lock rcu_read_lock rcu_read_lock &____s->seqcount#10 irq_context: 0 (wq_completion)bond117 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &idev->mc_lock rcu_read_lock rcu_read_lock rcu_read_lock_bh rcu_read_lock rcu_read_lock_bh rcu_read_lock &br->multicast_lock irq_context: 0 (wq_completion)bond117 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &idev->mc_lock rcu_read_lock rcu_read_lock rcu_read_lock_bh rcu_read_lock rcu_read_lock_bh rcu_read_lock &br->multicast_lock &obj_hash[i].lock irq_context: 0 (wq_completion)bond117 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &idev->mc_lock rcu_read_lock rcu_read_lock rcu_read_lock_bh rcu_read_lock rcu_read_lock_bh rcu_read_lock &br->multicast_lock &base->lock irq_context: 0 (wq_completion)bond117 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &idev->mc_lock rcu_read_lock rcu_read_lock rcu_read_lock_bh rcu_read_lock rcu_read_lock_bh rcu_read_lock &data->lock irq_context: 0 (wq_completion)bond117 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &idev->mc_lock rcu_read_lock rcu_read_lock rcu_read_lock_bh rcu_read_lock rcu_read_lock_bh rcu_read_lock &obj_hash[i].lock irq_context: 0 (wq_completion)bond117 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex.wait_lock irq_context: 0 (wq_completion)bond117 (work_completion)(&(&bond->mcast_work)->work) &p->pi_lock irq_context: 0 (wq_completion)bond108#2 (work_completion)(&(&slave->notify_work)->work) rtnl_mutex net_rwsem irq_context: 0 (wq_completion)bond108#2 (work_completion)(&(&slave->notify_work)->work) rtnl_mutex net_rwsem &list->lock#2 irq_context: 0 (wq_completion)bond108#2 (work_completion)(&(&slave->notify_work)->work) rtnl_mutex &tn->lock irq_context: 0 (wq_completion)bond108#2 (work_completion)(&(&slave->notify_work)->work) rtnl_mutex.wait_lock irq_context: 0 (wq_completion)bond108#2 (work_completion)(&(&slave->notify_work)->work) &p->pi_lock irq_context: 0 (wq_completion)bond121 (work_completion)(&(&slave->notify_work)->work) rtnl_mutex net_rwsem irq_context: 0 (wq_completion)bond121 (work_completion)(&(&slave->notify_work)->work) rtnl_mutex net_rwsem &list->lock#2 irq_context: 0 (wq_completion)bond121 (work_completion)(&(&slave->notify_work)->work) rtnl_mutex &tn->lock irq_context: 0 (wq_completion)bond121 (work_completion)(&(&slave->notify_work)->work) rtnl_mutex.wait_lock irq_context: 0 (wq_completion)bond121 (work_completion)(&(&slave->notify_work)->work) &p->pi_lock irq_context: 0 (wq_completion)bond118 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex fs_reclaim irq_context: 0 (wq_completion)bond118 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex fs_reclaim mmu_notifier_invalidate_range_start irq_context: 0 (wq_completion)bond118 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex rcu_read_lock &bond->stats_lock/1 irq_context: 0 (wq_completion)bond118 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex rcu_read_lock &ndev->lock irq_context: 0 (wq_completion)bond118 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex nl_table_lock irq_context: 0 (wq_completion)bond118 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &obj_hash[i].lock irq_context: 0 (wq_completion)bond118 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex nl_table_wait.lock irq_context: 0 (wq_completion)bond118 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &rq->__lock irq_context: 0 (wq_completion)bond118 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)bond118 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex net_rwsem irq_context: 0 (wq_completion)bond118 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex net_rwsem &list->lock#2 irq_context: 0 (wq_completion)bond118 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &tn->lock irq_context: 0 (wq_completion)bond118 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &idev->mc_lock irq_context: 0 (wq_completion)bond118 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &idev->mc_lock fs_reclaim irq_context: 0 (wq_completion)bond118 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &idev->mc_lock fs_reclaim mmu_notifier_invalidate_range_start irq_context: 0 (wq_completion)bond118 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &idev->mc_lock rcu_read_lock &ndev->lock irq_context: 0 (wq_completion)bond118 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &idev->mc_lock rcu_read_lock &dir->lock#2 irq_context: 0 (wq_completion)bond118 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &idev->mc_lock rcu_read_lock &ul->lock#2 irq_context: 0 (wq_completion)bond118 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &idev->mc_lock rcu_read_lock rcu_read_lock &____s->seqcount#10 irq_context: 0 (wq_completion)bond118 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &idev->mc_lock rcu_read_lock rcu_read_lock rcu_read_lock_bh rcu_read_lock rcu_read_lock_bh rcu_read_lock &br->multicast_lock irq_context: 0 (wq_completion)bond118 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &idev->mc_lock rcu_read_lock rcu_read_lock rcu_read_lock_bh rcu_read_lock rcu_read_lock_bh rcu_read_lock &br->multicast_lock &obj_hash[i].lock irq_context: 0 (wq_completion)bond118 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &idev->mc_lock rcu_read_lock rcu_read_lock rcu_read_lock_bh rcu_read_lock rcu_read_lock_bh rcu_read_lock &br->multicast_lock &base->lock irq_context: 0 (wq_completion)bond118 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &idev->mc_lock rcu_read_lock rcu_read_lock rcu_read_lock_bh rcu_read_lock rcu_read_lock_bh rcu_read_lock &data->lock irq_context: 0 (wq_completion)bond118 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &idev->mc_lock rcu_read_lock rcu_read_lock rcu_read_lock_bh rcu_read_lock rcu_read_lock_bh rcu_read_lock &obj_hash[i].lock irq_context: 0 (wq_completion)bond118 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex.wait_lock irq_context: 0 (wq_completion)bond118 (work_completion)(&(&bond->mcast_work)->work) &p->pi_lock irq_context: 0 (wq_completion)bond107#2 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex fs_reclaim irq_context: 0 (wq_completion)bond107#2 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex fs_reclaim mmu_notifier_invalidate_range_start irq_context: 0 (wq_completion)bond107#2 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex rcu_read_lock &bond->stats_lock/1 irq_context: 0 (wq_completion)bond107#2 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex rcu_read_lock &ndev->lock irq_context: 0 (wq_completion)bond107#2 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex nl_table_lock irq_context: 0 (wq_completion)bond107#2 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &obj_hash[i].lock irq_context: 0 (wq_completion)bond107#2 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex nl_table_wait.lock irq_context: 0 (wq_completion)bond107#2 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex net_rwsem irq_context: 0 (wq_completion)bond107#2 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex net_rwsem &list->lock#2 irq_context: 0 (wq_completion)bond107#2 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &tn->lock irq_context: 0 (wq_completion)bond107#2 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &tn->lock &rq->__lock irq_context: 0 (wq_completion)bond107#2 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &tn->lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)bond107#2 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &idev->mc_lock irq_context: 0 (wq_completion)bond107#2 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &idev->mc_lock fs_reclaim irq_context: 0 (wq_completion)bond107#2 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &idev->mc_lock fs_reclaim mmu_notifier_invalidate_range_start irq_context: 0 (wq_completion)bond107#2 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &idev->mc_lock &c->lock irq_context: 0 (wq_completion)bond107#2 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &idev->mc_lock rcu_read_lock &ndev->lock irq_context: 0 (wq_completion)bond107#2 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &idev->mc_lock rcu_read_lock &c->lock irq_context: 0 (wq_completion)bond107#2 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &idev->mc_lock rcu_read_lock &dir->lock#2 irq_context: 0 (wq_completion)bond107#2 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &idev->mc_lock rcu_read_lock &ul->lock#2 irq_context: 0 (wq_completion)bond107#2 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &idev->mc_lock rcu_read_lock rcu_read_lock &____s->seqcount#10 irq_context: 0 (wq_completion)bond107#2 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &idev->mc_lock rcu_read_lock rcu_read_lock rcu_read_lock_bh rcu_read_lock rcu_read_lock_bh rcu_read_lock &br->multicast_lock irq_context: 0 (wq_completion)bond107#2 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &idev->mc_lock rcu_read_lock rcu_read_lock rcu_read_lock_bh rcu_read_lock rcu_read_lock_bh rcu_read_lock &br->multicast_lock &obj_hash[i].lock irq_context: 0 (wq_completion)bond107#2 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &idev->mc_lock rcu_read_lock rcu_read_lock rcu_read_lock_bh rcu_read_lock rcu_read_lock_bh rcu_read_lock &br->multicast_lock &base->lock irq_context: 0 (wq_completion)bond107#2 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &idev->mc_lock rcu_read_lock rcu_read_lock rcu_read_lock_bh rcu_read_lock rcu_read_lock_bh rcu_read_lock &data->lock irq_context: 0 (wq_completion)bond107#2 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &idev->mc_lock rcu_read_lock rcu_read_lock rcu_read_lock_bh rcu_read_lock rcu_read_lock_bh rcu_read_lock &obj_hash[i].lock irq_context: 0 (wq_completion)bond107#2 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &idev->mc_lock rcu_read_lock rcu_read_lock rcu_node_0 irq_context: 0 (wq_completion)bond107#2 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &idev->mc_lock rcu_read_lock rcu_read_lock &rq->__lock irq_context: 0 (wq_completion)bond107#2 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &idev->mc_lock rcu_read_lock rcu_read_lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)bond107#2 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &idev->mc_lock rcu_read_lock rcu_node_0 irq_context: 0 (wq_completion)bond107#2 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex.wait_lock irq_context: 0 (wq_completion)bond107#2 (work_completion)(&(&bond->mcast_work)->work) &p->pi_lock irq_context: 0 (wq_completion)bond121 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex fs_reclaim irq_context: 0 (wq_completion)bond121 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex fs_reclaim mmu_notifier_invalidate_range_start irq_context: 0 (wq_completion)bond121 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &c->lock irq_context: 0 (wq_completion)bond121 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex rcu_read_lock &bond->stats_lock/1 irq_context: 0 (wq_completion)bond121 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex rcu_read_lock &ndev->lock irq_context: 0 (wq_completion)bond121 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex nl_table_lock irq_context: 0 (wq_completion)bond121 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &obj_hash[i].lock irq_context: 0 (wq_completion)bond121 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex nl_table_wait.lock irq_context: 0 (wq_completion)bond121 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex net_rwsem irq_context: 0 (wq_completion)bond121 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex net_rwsem &list->lock#2 irq_context: 0 (wq_completion)bond121 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &tn->lock irq_context: 0 (wq_completion)bond121 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &tn->lock &rq->__lock irq_context: 0 (wq_completion)bond121 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &tn->lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)bond121 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &idev->mc_lock irq_context: 0 (wq_completion)bond121 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex.wait_lock irq_context: 0 (wq_completion)bond121 (work_completion)(&(&bond->mcast_work)->work) &p->pi_lock irq_context: 0 (wq_completion)bond92#2 irq_context: 0 (wq_completion)bond92#2 (work_completion)(&(&slave->notify_work)->work) irq_context: 0 (wq_completion)bond92#2 (work_completion)(&(&slave->notify_work)->work) &obj_hash[i].lock irq_context: 0 (wq_completion)bond92#2 (work_completion)(&(&slave->notify_work)->work) &base->lock irq_context: 0 (wq_completion)bond92#2 (work_completion)(&(&slave->notify_work)->work) &base->lock &obj_hash[i].lock irq_context: 0 (wq_completion)bond92#2 (work_completion)(&(&bond->mcast_work)->work) irq_context: 0 (wq_completion)bond92#2 (work_completion)(&(&bond->mcast_work)->work) &obj_hash[i].lock irq_context: 0 (wq_completion)bond92#2 (work_completion)(&(&bond->mcast_work)->work) &base->lock irq_context: 0 (wq_completion)bond92#2 (work_completion)(&(&bond->mcast_work)->work) &base->lock &obj_hash[i].lock irq_context: 0 (wq_completion)bond92#2 (work_completion)(&(&bond->mcast_work)->work) &rq->__lock irq_context: 0 (wq_completion)bond92#2 (work_completion)(&(&bond->mcast_work)->work) &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)bond13#2 (work_completion)(&(&slave->notify_work)->work) rtnl_mutex net_rwsem irq_context: 0 (wq_completion)bond13#2 (work_completion)(&(&slave->notify_work)->work) rtnl_mutex net_rwsem &list->lock#2 irq_context: 0 (wq_completion)bond13#2 (work_completion)(&(&slave->notify_work)->work) rtnl_mutex &tn->lock irq_context: 0 (wq_completion)bond13#2 (work_completion)(&(&slave->notify_work)->work) rtnl_mutex.wait_lock irq_context: 0 (wq_completion)bond13#2 (work_completion)(&(&slave->notify_work)->work) &p->pi_lock irq_context: 0 (wq_completion)bond103#2 (work_completion)(&(&slave->notify_work)->work) rtnl_mutex net_rwsem irq_context: 0 (wq_completion)bond103#2 (work_completion)(&(&slave->notify_work)->work) rtnl_mutex net_rwsem &list->lock#2 irq_context: 0 (wq_completion)bond103#2 (work_completion)(&(&slave->notify_work)->work) rtnl_mutex &tn->lock irq_context: 0 (wq_completion)bond103#2 (work_completion)(&(&slave->notify_work)->work) rtnl_mutex.wait_lock irq_context: 0 (wq_completion)bond103#2 (work_completion)(&(&slave->notify_work)->work) &p->pi_lock irq_context: 0 (wq_completion)bond102#2 (work_completion)(&(&slave->notify_work)->work) rtnl_mutex net_rwsem irq_context: 0 (wq_completion)bond102#2 (work_completion)(&(&slave->notify_work)->work) rtnl_mutex net_rwsem &list->lock#2 irq_context: 0 (wq_completion)bond102#2 (work_completion)(&(&slave->notify_work)->work) rtnl_mutex &tn->lock irq_context: 0 (wq_completion)bond102#2 (work_completion)(&(&slave->notify_work)->work) rtnl_mutex.wait_lock irq_context: 0 (wq_completion)bond102#2 (work_completion)(&(&slave->notify_work)->work) &p->pi_lock irq_context: 0 (wq_completion)bond97#2 (work_completion)(&(&slave->notify_work)->work) rtnl_mutex net_rwsem irq_context: 0 (wq_completion)bond97#2 (work_completion)(&(&slave->notify_work)->work) rtnl_mutex net_rwsem &list->lock#2 irq_context: 0 (wq_completion)bond97#2 (work_completion)(&(&slave->notify_work)->work) rtnl_mutex &tn->lock irq_context: 0 (wq_completion)bond97#2 (work_completion)(&(&slave->notify_work)->work) rtnl_mutex.wait_lock irq_context: 0 (wq_completion)bond97#2 (work_completion)(&(&slave->notify_work)->work) &p->pi_lock irq_context: 0 (wq_completion)bond82 (work_completion)(&(&slave->notify_work)->work) rtnl_mutex net_rwsem irq_context: 0 (wq_completion)bond82 (work_completion)(&(&slave->notify_work)->work) rtnl_mutex net_rwsem &list->lock#2 irq_context: 0 (wq_completion)bond82 (work_completion)(&(&slave->notify_work)->work) rtnl_mutex &tn->lock irq_context: 0 (wq_completion)bond82 (work_completion)(&(&slave->notify_work)->work) rtnl_mutex.wait_lock irq_context: 0 (wq_completion)bond82 (work_completion)(&(&slave->notify_work)->work) &p->pi_lock irq_context: 0 (wq_completion)bond78 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex fs_reclaim irq_context: 0 (wq_completion)bond78 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex fs_reclaim mmu_notifier_invalidate_range_start irq_context: 0 (wq_completion)bond78 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &c->lock irq_context: 0 (wq_completion)bond78 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &n->list_lock irq_context: 0 (wq_completion)bond78 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &n->list_lock &c->lock irq_context: 0 (wq_completion)bond78 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex rcu_read_lock &bond->stats_lock/1 irq_context: 0 (wq_completion)bond78 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex rcu_read_lock &ndev->lock irq_context: 0 (wq_completion)bond78 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex nl_table_lock irq_context: 0 (wq_completion)bond78 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &obj_hash[i].lock irq_context: 0 (wq_completion)bond78 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex nl_table_wait.lock irq_context: 0 (wq_completion)bond78 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex net_rwsem irq_context: 0 (wq_completion)bond78 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex net_rwsem &list->lock#2 irq_context: 0 (wq_completion)bond78 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &tn->lock irq_context: 0 (wq_completion)bond78 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &idev->mc_lock irq_context: 0 (wq_completion)bond78 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &idev->mc_lock fs_reclaim irq_context: 0 (wq_completion)bond78 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &idev->mc_lock fs_reclaim mmu_notifier_invalidate_range_start irq_context: 0 (wq_completion)bond78 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &idev->mc_lock fs_reclaim mmu_notifier_invalidate_range_start &rq->__lock irq_context: 0 (wq_completion)bond78 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &idev->mc_lock fs_reclaim mmu_notifier_invalidate_range_start &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)bond78 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &idev->mc_lock &c->lock irq_context: 0 (wq_completion)bond78 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &idev->mc_lock &n->list_lock irq_context: 0 (wq_completion)bond78 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &idev->mc_lock &n->list_lock &c->lock irq_context: 0 (wq_completion)bond78 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &idev->mc_lock rcu_read_lock &ndev->lock irq_context: 0 (wq_completion)bond78 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &idev->mc_lock rcu_read_lock &dir->lock#2 irq_context: 0 (wq_completion)bond78 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &idev->mc_lock rcu_read_lock &ul->lock#2 irq_context: 0 (wq_completion)bond78 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &idev->mc_lock rcu_read_lock rcu_read_lock &____s->seqcount#10 irq_context: 0 (wq_completion)bond78 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &idev->mc_lock rcu_read_lock rcu_read_lock rcu_read_lock_bh rcu_read_lock rcu_read_lock_bh rcu_read_lock &br->multicast_lock irq_context: 0 (wq_completion)bond78 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &idev->mc_lock rcu_read_lock rcu_read_lock rcu_read_lock_bh rcu_read_lock rcu_read_lock_bh rcu_read_lock &br->multicast_lock &obj_hash[i].lock irq_context: 0 (wq_completion)bond78 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &idev->mc_lock rcu_read_lock rcu_read_lock rcu_read_lock_bh rcu_read_lock rcu_read_lock_bh rcu_read_lock &br->multicast_lock &base->lock irq_context: 0 (wq_completion)bond78 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &idev->mc_lock rcu_read_lock rcu_read_lock rcu_read_lock_bh rcu_read_lock rcu_read_lock_bh rcu_read_lock &br->multicast_lock &base->lock &obj_hash[i].lock irq_context: 0 (wq_completion)bond78 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &idev->mc_lock rcu_read_lock rcu_read_lock rcu_read_lock_bh rcu_read_lock rcu_read_lock_bh rcu_read_lock &data->lock irq_context: 0 (wq_completion)bond78 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &idev->mc_lock rcu_read_lock rcu_read_lock rcu_read_lock_bh rcu_read_lock rcu_read_lock_bh rcu_read_lock &obj_hash[i].lock irq_context: 0 (wq_completion)bond78 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &idev->mc_lock rcu_read_lock rcu_read_lock rcu_node_0 irq_context: 0 (wq_completion)bond78 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &idev->mc_lock rcu_read_lock rcu_read_lock &rq->__lock irq_context: 0 (wq_completion)bond78 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &idev->mc_lock rcu_read_lock rcu_node_0 irq_context: 0 (wq_completion)bond78 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex.wait_lock irq_context: 0 (wq_completion)bond78 (work_completion)(&(&bond->mcast_work)->work) &p->pi_lock irq_context: 0 (wq_completion)bond109#2 irq_context: 0 (wq_completion)bond109#2 &rq->__lock irq_context: 0 (wq_completion)bond109#2 &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)bond109#2 (work_completion)(&(&slave->notify_work)->work) irq_context: 0 (wq_completion)bond109#2 (work_completion)(&(&slave->notify_work)->work) &obj_hash[i].lock irq_context: 0 (wq_completion)bond109#2 (work_completion)(&(&slave->notify_work)->work) &base->lock irq_context: 0 (wq_completion)bond109#2 (work_completion)(&(&slave->notify_work)->work) &base->lock &obj_hash[i].lock irq_context: 0 (wq_completion)bond14#2 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex fs_reclaim irq_context: 0 (wq_completion)bond14#2 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex fs_reclaim mmu_notifier_invalidate_range_start irq_context: 0 (wq_completion)bond14#2 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &c->lock irq_context: 0 (wq_completion)bond14#2 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex rcu_read_lock &bond->stats_lock/1 irq_context: 0 (wq_completion)bond14#2 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex rcu_read_lock &ndev->lock irq_context: 0 (wq_completion)bond14#2 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex rcu_read_lock rcu_node_0 irq_context: 0 (wq_completion)bond14#2 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex rcu_read_lock &rq->__lock irq_context: 0 (wq_completion)bond14#2 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex rcu_read_lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)bond14#2 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex nl_table_lock irq_context: 0 (wq_completion)bond14#2 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &obj_hash[i].lock irq_context: 0 (wq_completion)bond14#2 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex nl_table_wait.lock irq_context: 0 (wq_completion)bond14#2 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex net_rwsem irq_context: 0 (wq_completion)bond14#2 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex net_rwsem &list->lock#2 irq_context: 0 (wq_completion)bond14#2 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &tn->lock irq_context: 0 (wq_completion)bond14#2 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &idev->mc_lock irq_context: 0 (wq_completion)bond14#2 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &idev->mc_lock fs_reclaim irq_context: 0 (wq_completion)bond14#2 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &idev->mc_lock fs_reclaim mmu_notifier_invalidate_range_start irq_context: 0 (wq_completion)bond14#2 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &idev->mc_lock fs_reclaim mmu_notifier_invalidate_range_start &rq->__lock irq_context: 0 (wq_completion)bond14#2 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &idev->mc_lock fs_reclaim mmu_notifier_invalidate_range_start &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)bond109#2 (work_completion)(&(&bond->mcast_work)->work) irq_context: 0 (wq_completion)bond109#2 (work_completion)(&(&bond->mcast_work)->work) &obj_hash[i].lock irq_context: 0 (wq_completion)bond109#2 (work_completion)(&(&bond->mcast_work)->work) &base->lock irq_context: 0 (wq_completion)bond109#2 (work_completion)(&(&bond->mcast_work)->work) &base->lock &obj_hash[i].lock irq_context: 0 (wq_completion)bond14#2 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &idev->mc_lock rcu_read_lock &ndev->lock irq_context: 0 (wq_completion)bond14#2 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &idev->mc_lock rcu_read_lock &dir->lock#2 irq_context: 0 (wq_completion)bond14#2 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &idev->mc_lock rcu_read_lock &ul->lock#2 irq_context: 0 (wq_completion)bond14#2 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &idev->mc_lock rcu_read_lock rcu_read_lock &____s->seqcount#12 irq_context: 0 (wq_completion)bond14#2 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &idev->mc_lock rcu_read_lock rcu_read_lock &____s->seqcount#16 irq_context: 0 (wq_completion)bond14#2 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &idev->mc_lock rcu_read_lock &obj_hash[i].lock irq_context: 0 (wq_completion)bond14#2 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &idev->mc_lock rcu_read_lock &data->lock irq_context: 0 (wq_completion)bond14#2 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &rq->__lock irq_context: 0 (wq_completion)bond14#2 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)bond14#2 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex.wait_lock irq_context: 0 (wq_completion)bond14#2 (work_completion)(&(&bond->mcast_work)->work) &p->pi_lock irq_context: 0 (wq_completion)bond14#2 (work_completion)(&(&bond->mcast_work)->work) &p->pi_lock &cfs_rq->removed.lock irq_context: 0 (wq_completion)bond14#2 (work_completion)(&(&bond->mcast_work)->work) &p->pi_lock &rq->__lock irq_context: 0 (wq_completion)bond14#2 (work_completion)(&(&bond->mcast_work)->work) &p->pi_lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)bond14#2 (work_completion)(&(&bond->mcast_work)->work) &rq->__lock irq_context: 0 (wq_completion)bond81#2 (work_completion)(&(&bond->mcast_work)->work) fill_pool_map-wait-type-override pool_lock irq_context: 0 (wq_completion)events_power_efficient (work_completion)(&(&tbl->gc_work)->work) &tbl->lock fill_pool_map-wait-type-override &____s->seqcount#2 irq_context: 0 (wq_completion)events_power_efficient (work_completion)(&(&tbl->gc_work)->work) &tbl->lock fill_pool_map-wait-type-override &____s->seqcount irq_context: 0 (wq_completion)bond122 irq_context: 0 (wq_completion)bond122 (work_completion)(&(&slave->notify_work)->work) irq_context: 0 (wq_completion)bond122 (work_completion)(&(&slave->notify_work)->work) &obj_hash[i].lock irq_context: 0 (wq_completion)bond122 (work_completion)(&(&slave->notify_work)->work) &base->lock irq_context: 0 (wq_completion)bond122 (work_completion)(&(&slave->notify_work)->work) &base->lock &obj_hash[i].lock irq_context: 0 (wq_completion)bond122 (work_completion)(&(&bond->mcast_work)->work) irq_context: 0 (wq_completion)bond122 (work_completion)(&(&bond->mcast_work)->work) &obj_hash[i].lock irq_context: 0 (wq_completion)bond122 (work_completion)(&(&bond->mcast_work)->work) &base->lock irq_context: 0 (wq_completion)bond122 (work_completion)(&(&bond->mcast_work)->work) &base->lock &obj_hash[i].lock irq_context: 0 (wq_completion)events (linkwatch_work).work rtnl_mutex net_rwsem &rq->__lock irq_context: 0 (wq_completion)events (linkwatch_work).work rtnl_mutex net_rwsem &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)bond17#3 (work_completion)(&(&slave->notify_work)->work) rtnl_mutex net_rwsem irq_context: 0 (wq_completion)bond17#3 (work_completion)(&(&slave->notify_work)->work) rtnl_mutex net_rwsem &list->lock#2 irq_context: 0 (wq_completion)bond17#3 (work_completion)(&(&slave->notify_work)->work) rtnl_mutex &tn->lock irq_context: 0 (wq_completion)bond17#3 (work_completion)(&(&slave->notify_work)->work) rtnl_mutex.wait_lock irq_context: 0 (wq_completion)bond17#3 (work_completion)(&(&slave->notify_work)->work) &p->pi_lock irq_context: 0 (wq_completion)bond109#2 (work_completion)(&(&slave->notify_work)->work) rtnl_mutex net_rwsem irq_context: 0 (wq_completion)bond109#2 (work_completion)(&(&slave->notify_work)->work) rtnl_mutex net_rwsem &list->lock#2 irq_context: 0 (wq_completion)bond109#2 (work_completion)(&(&slave->notify_work)->work) rtnl_mutex &tn->lock irq_context: 0 (wq_completion)bond109#2 (work_completion)(&(&slave->notify_work)->work) rtnl_mutex.wait_lock irq_context: 0 (wq_completion)bond109#2 (work_completion)(&(&slave->notify_work)->work) &p->pi_lock irq_context: 0 (wq_completion)bond92#2 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex fs_reclaim irq_context: 0 (wq_completion)bond92#2 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex fs_reclaim mmu_notifier_invalidate_range_start irq_context: 0 (wq_completion)bond92#2 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &c->lock irq_context: 0 (wq_completion)bond92#2 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &n->list_lock irq_context: 0 (wq_completion)bond92#2 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &n->list_lock &c->lock irq_context: 0 (wq_completion)bond92#2 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex rcu_read_lock &bond->stats_lock/1 irq_context: 0 (wq_completion)bond92#2 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex rcu_read_lock &ndev->lock irq_context: 0 (wq_completion)bond92#2 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex nl_table_lock irq_context: 0 (wq_completion)bond92#2 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &obj_hash[i].lock irq_context: 0 (wq_completion)bond92#2 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex nl_table_wait.lock irq_context: 0 (wq_completion)bond92#2 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex net_rwsem irq_context: 0 (wq_completion)bond92#2 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex net_rwsem &list->lock#2 irq_context: 0 (wq_completion)bond92#2 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &tn->lock irq_context: 0 (wq_completion)bond92#2 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &idev->mc_lock irq_context: 0 (wq_completion)bond92#2 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex.wait_lock irq_context: 0 (wq_completion)bond92#2 (work_completion)(&(&bond->mcast_work)->work) &p->pi_lock irq_context: 0 (wq_completion)bond118 (work_completion)(&(&slave->notify_work)->work) rtnl_mutex net_rwsem irq_context: 0 (wq_completion)bond118 (work_completion)(&(&slave->notify_work)->work) rtnl_mutex net_rwsem &list->lock#2 irq_context: 0 (wq_completion)bond118 (work_completion)(&(&slave->notify_work)->work) rtnl_mutex &tn->lock irq_context: 0 (wq_completion)bond118 (work_completion)(&(&slave->notify_work)->work) rtnl_mutex.wait_lock irq_context: 0 (wq_completion)bond118 (work_completion)(&(&slave->notify_work)->work) &p->pi_lock irq_context: 0 (wq_completion)bond85#2 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex fs_reclaim irq_context: 0 (wq_completion)bond85#2 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex fs_reclaim mmu_notifier_invalidate_range_start irq_context: 0 (wq_completion)bond85#2 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &c->lock irq_context: 0 (wq_completion)bond85#2 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex rcu_read_lock &bond->stats_lock/1 irq_context: 0 (wq_completion)bond85#2 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex rcu_read_lock &ndev->lock irq_context: 0 (wq_completion)bond85#2 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex nl_table_lock irq_context: 0 (wq_completion)bond85#2 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &obj_hash[i].lock irq_context: 0 (wq_completion)bond85#2 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex nl_table_wait.lock irq_context: 0 (wq_completion)bond85#2 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex net_rwsem irq_context: 0 (wq_completion)bond85#2 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex net_rwsem &list->lock#2 irq_context: 0 (wq_completion)bond85#2 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &tn->lock irq_context: 0 (wq_completion)bond85#2 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &idev->mc_lock irq_context: 0 (wq_completion)bond85#2 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &idev->mc_lock fs_reclaim irq_context: 0 (wq_completion)bond85#2 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &idev->mc_lock fs_reclaim mmu_notifier_invalidate_range_start irq_context: 0 (wq_completion)bond85#2 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &idev->mc_lock rcu_read_lock &ndev->lock irq_context: 0 (wq_completion)bond85#2 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &idev->mc_lock rcu_read_lock &c->lock irq_context: 0 (wq_completion)bond85#2 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &idev->mc_lock rcu_read_lock &dir->lock#2 irq_context: 0 (wq_completion)bond85#2 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &idev->mc_lock rcu_read_lock &ul->lock#2 irq_context: 0 (wq_completion)bond85#2 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &idev->mc_lock rcu_read_lock rcu_read_lock &____s->seqcount#10 irq_context: 0 (wq_completion)bond85#2 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &idev->mc_lock rcu_read_lock rcu_read_lock rcu_read_lock_bh rcu_read_lock rcu_read_lock_bh rcu_read_lock &br->multicast_lock irq_context: 0 (wq_completion)bond85#2 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &idev->mc_lock rcu_read_lock rcu_read_lock rcu_read_lock_bh rcu_read_lock rcu_read_lock_bh rcu_read_lock &br->multicast_lock &obj_hash[i].lock irq_context: 0 (wq_completion)bond85#2 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &idev->mc_lock rcu_read_lock rcu_read_lock rcu_read_lock_bh rcu_read_lock rcu_read_lock_bh rcu_read_lock &br->multicast_lock &base->lock irq_context: 0 (wq_completion)bond85#2 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &idev->mc_lock rcu_read_lock rcu_read_lock rcu_read_lock_bh rcu_read_lock rcu_read_lock_bh rcu_read_lock &data->lock irq_context: 0 (wq_completion)bond85#2 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &idev->mc_lock rcu_read_lock rcu_read_lock rcu_read_lock_bh rcu_read_lock rcu_read_lock_bh rcu_read_lock &obj_hash[i].lock irq_context: 0 (wq_completion)bond85#2 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex.wait_lock irq_context: 0 (wq_completion)bond85#2 (work_completion)(&(&bond->mcast_work)->work) &p->pi_lock irq_context: 0 (wq_completion)bond110 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex fs_reclaim irq_context: 0 (wq_completion)bond110 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex fs_reclaim mmu_notifier_invalidate_range_start irq_context: 0 (wq_completion)bond110 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &c->lock irq_context: 0 (wq_completion)bond110 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex rcu_read_lock &bond->stats_lock/1 irq_context: 0 (wq_completion)bond110 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex rcu_read_lock &ndev->lock irq_context: 0 (wq_completion)bond110 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex nl_table_lock irq_context: 0 (wq_completion)bond110 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &obj_hash[i].lock irq_context: 0 (wq_completion)bond110 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex nl_table_wait.lock irq_context: 0 (wq_completion)bond110 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex net_rwsem irq_context: 0 (wq_completion)bond110 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex net_rwsem &list->lock#2 irq_context: 0 (wq_completion)bond110 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &tn->lock irq_context: 0 (wq_completion)bond110 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &idev->mc_lock irq_context: 0 (wq_completion)bond110 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &idev->mc_lock fs_reclaim irq_context: 0 (wq_completion)bond110 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &idev->mc_lock fs_reclaim mmu_notifier_invalidate_range_start irq_context: 0 (wq_completion)bond110 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &idev->mc_lock &c->lock irq_context: 0 (wq_completion)bond110 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &idev->mc_lock &n->list_lock irq_context: 0 (wq_completion)bond110 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &idev->mc_lock &n->list_lock &c->lock irq_context: 0 (wq_completion)bond110 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &idev->mc_lock rcu_read_lock &ndev->lock irq_context: 0 (wq_completion)bond110 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &idev->mc_lock rcu_read_lock &dir->lock#2 irq_context: 0 (wq_completion)bond110 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &idev->mc_lock rcu_read_lock &ul->lock#2 irq_context: 0 (wq_completion)bond110 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &idev->mc_lock rcu_read_lock rcu_read_lock &____s->seqcount#10 irq_context: 0 (wq_completion)bond110 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &idev->mc_lock rcu_read_lock rcu_read_lock rcu_read_lock_bh rcu_read_lock rcu_read_lock_bh rcu_read_lock &br->multicast_lock irq_context: 0 (wq_completion)bond110 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &idev->mc_lock rcu_read_lock rcu_read_lock rcu_read_lock_bh rcu_read_lock rcu_read_lock_bh rcu_read_lock &br->multicast_lock &obj_hash[i].lock irq_context: 0 (wq_completion)bond110 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &idev->mc_lock rcu_read_lock rcu_read_lock rcu_read_lock_bh rcu_read_lock rcu_read_lock_bh rcu_read_lock &br->multicast_lock &base->lock irq_context: 0 (wq_completion)bond110 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &idev->mc_lock rcu_read_lock rcu_read_lock rcu_read_lock_bh rcu_read_lock rcu_read_lock_bh rcu_read_lock &br->multicast_lock &base->lock &obj_hash[i].lock irq_context: 0 (wq_completion)bond110 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &idev->mc_lock rcu_read_lock rcu_read_lock rcu_read_lock_bh rcu_read_lock rcu_read_lock_bh rcu_read_lock &data->lock irq_context: 0 (wq_completion)bond110 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &idev->mc_lock rcu_read_lock rcu_read_lock rcu_read_lock_bh rcu_read_lock rcu_read_lock_bh rcu_read_lock &obj_hash[i].lock irq_context: 0 (wq_completion)bond110 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &idev->mc_lock rcu_read_lock rcu_read_lock rcu_node_0 irq_context: 0 (wq_completion)bond110 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &idev->mc_lock rcu_read_lock rcu_read_lock &rq->__lock irq_context: 0 (wq_completion)bond110 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &idev->mc_lock rcu_read_lock rcu_read_lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)bond110 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &idev->mc_lock rcu_read_lock rcu_node_0 irq_context: 0 (wq_completion)bond110 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex.wait_lock irq_context: 0 (wq_completion)bond110 (work_completion)(&(&bond->mcast_work)->work) &p->pi_lock irq_context: 0 (wq_completion)bond110 (work_completion)(&(&bond->mcast_work)->work) &p->pi_lock &rq->__lock irq_context: 0 (wq_completion)bond110 (work_completion)(&(&bond->mcast_work)->work) &p->pi_lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)bond93#2 irq_context: 0 (wq_completion)bond93#2 (work_completion)(&(&slave->notify_work)->work) irq_context: 0 (wq_completion)bond93#2 (work_completion)(&(&slave->notify_work)->work) &rq->__lock irq_context: 0 (wq_completion)bond93#2 (work_completion)(&(&slave->notify_work)->work) &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)bond93#2 (work_completion)(&(&slave->notify_work)->work) &obj_hash[i].lock irq_context: 0 (wq_completion)bond93#2 (work_completion)(&(&slave->notify_work)->work) &base->lock irq_context: 0 (wq_completion)bond93#2 (work_completion)(&(&slave->notify_work)->work) &base->lock &obj_hash[i].lock irq_context: 0 (wq_completion)bond80#2 (work_completion)(&(&bond->mcast_work)->work) &rq->__lock irq_context: 0 (wq_completion)bond80#2 (work_completion)(&(&bond->mcast_work)->work) &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)bond93#2 (work_completion)(&(&bond->mcast_work)->work) irq_context: 0 (wq_completion)bond93#2 (work_completion)(&(&bond->mcast_work)->work) &obj_hash[i].lock irq_context: 0 (wq_completion)bond93#2 (work_completion)(&(&bond->mcast_work)->work) &base->lock irq_context: 0 (wq_completion)bond93#2 (work_completion)(&(&bond->mcast_work)->work) &base->lock &obj_hash[i].lock irq_context: 0 (wq_completion)bond93#2 (work_completion)(&(&bond->mcast_work)->work) &rq->__lock irq_context: 0 (wq_completion)bond93#2 (work_completion)(&(&bond->mcast_work)->work) &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)bond110#2 irq_context: 0 (wq_completion)bond110#2 (work_completion)(&(&slave->notify_work)->work) irq_context: 0 (wq_completion)bond110#2 (work_completion)(&(&slave->notify_work)->work) &obj_hash[i].lock irq_context: 0 (wq_completion)bond110#2 (work_completion)(&(&slave->notify_work)->work) &base->lock irq_context: 0 (wq_completion)bond110#2 (work_completion)(&(&slave->notify_work)->work) &base->lock &obj_hash[i].lock irq_context: 0 (wq_completion)bond110#2 (work_completion)(&(&bond->mcast_work)->work) irq_context: 0 (wq_completion)bond110#2 (work_completion)(&(&bond->mcast_work)->work) &obj_hash[i].lock irq_context: 0 (wq_completion)bond110#2 (work_completion)(&(&bond->mcast_work)->work) &base->lock irq_context: 0 (wq_completion)bond110#2 (work_completion)(&(&bond->mcast_work)->work) &base->lock &obj_hash[i].lock irq_context: 0 (wq_completion)bond123 irq_context: 0 (wq_completion)bond123 (work_completion)(&(&slave->notify_work)->work) irq_context: 0 (wq_completion)bond123 (work_completion)(&(&slave->notify_work)->work) &obj_hash[i].lock irq_context: 0 (wq_completion)bond123 (work_completion)(&(&slave->notify_work)->work) &base->lock irq_context: 0 (wq_completion)bond123 (work_completion)(&(&slave->notify_work)->work) &base->lock &obj_hash[i].lock irq_context: 0 (wq_completion)bond123 (work_completion)(&(&slave->notify_work)->work) &rq->__lock irq_context: 0 (wq_completion)bond123 (work_completion)(&(&slave->notify_work)->work) &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: softirq (&ndev->rs_timer) rcu_read_lock rcu_read_lock rcu_read_lock_bh rcu_read_lock rcu_read_lock_bh rcu_read_lock &data->lock &obj_hash[i].lock irq_context: softirq (&ndev->rs_timer) rcu_read_lock rcu_read_lock rcu_read_lock_bh rcu_read_lock rcu_read_lock_bh rcu_read_lock &data->lock &base->lock irq_context: softirq (&ndev->rs_timer) rcu_read_lock rcu_read_lock rcu_read_lock_bh rcu_read_lock rcu_read_lock_bh rcu_read_lock &data->lock &base->lock &obj_hash[i].lock irq_context: softirq &(&bond->mcast_work)->timer rcu_read_lock &pool->lock/1 fill_pool_map-wait-type-override pool_lock irq_context: 0 (wq_completion)bond123 (work_completion)(&(&bond->mcast_work)->work) irq_context: 0 (wq_completion)bond123 (work_completion)(&(&bond->mcast_work)->work) &obj_hash[i].lock irq_context: 0 (wq_completion)bond123 (work_completion)(&(&bond->mcast_work)->work) &base->lock irq_context: 0 (wq_completion)bond123 (work_completion)(&(&bond->mcast_work)->work) &base->lock &obj_hash[i].lock irq_context: 0 sb_writers#4 &type->i_mutex_dir_key#3/1 &type->i_mutex_dir_key#3 remove_cache_srcu rcu_read_lock rcu_read_lock rcu_node_0 irq_context: 0 sb_writers#4 &type->i_mutex_dir_key#3/1 &type->i_mutex_dir_key#3 remove_cache_srcu rcu_read_lock rcu_read_lock &rq->__lock irq_context: 0 sb_writers#4 &type->i_mutex_dir_key#3/1 &type->i_mutex_dir_key#3 remove_cache_srcu rcu_read_lock rcu_read_lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)bond94#2 irq_context: 0 (wq_completion)bond94#2 (work_completion)(&(&slave->notify_work)->work) irq_context: 0 (wq_completion)bond94#2 (work_completion)(&(&slave->notify_work)->work) &obj_hash[i].lock irq_context: 0 (wq_completion)bond94#2 (work_completion)(&(&slave->notify_work)->work) &base->lock irq_context: 0 (wq_completion)bond94#2 (work_completion)(&(&slave->notify_work)->work) &base->lock &obj_hash[i].lock irq_context: 0 (wq_completion)bond94#2 (work_completion)(&(&slave->notify_work)->work) &rq->__lock irq_context: 0 (wq_completion)bond94#2 (work_completion)(&(&bond->mcast_work)->work) irq_context: 0 (wq_completion)bond94#2 (work_completion)(&(&bond->mcast_work)->work) &obj_hash[i].lock irq_context: 0 (wq_completion)bond94#2 (work_completion)(&(&bond->mcast_work)->work) &base->lock irq_context: 0 (wq_completion)bond94#2 (work_completion)(&(&bond->mcast_work)->work) &base->lock &obj_hash[i].lock irq_context: 0 (wq_completion)bond94#2 (work_completion)(&(&bond->mcast_work)->work) &rq->__lock irq_context: 0 (wq_completion)bond94#2 (work_completion)(&(&bond->mcast_work)->work) &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)bond123 (work_completion)(&(&slave->notify_work)->work) rtnl_mutex net_rwsem irq_context: 0 (wq_completion)bond123 (work_completion)(&(&slave->notify_work)->work) rtnl_mutex net_rwsem &list->lock#2 irq_context: 0 (wq_completion)bond123 (work_completion)(&(&slave->notify_work)->work) rtnl_mutex &tn->lock irq_context: 0 (wq_completion)bond123 (work_completion)(&(&slave->notify_work)->work) rtnl_mutex.wait_lock irq_context: 0 (wq_completion)bond123 (work_completion)(&(&slave->notify_work)->work) &p->pi_lock irq_context: 0 (wq_completion)bond93#2 (work_completion)(&(&slave->notify_work)->work) rtnl_mutex net_rwsem irq_context: 0 (wq_completion)bond93#2 (work_completion)(&(&slave->notify_work)->work) rtnl_mutex net_rwsem &list->lock#2 irq_context: 0 (wq_completion)bond93#2 (work_completion)(&(&slave->notify_work)->work) rtnl_mutex &tn->lock irq_context: 0 (wq_completion)bond93#2 (work_completion)(&(&slave->notify_work)->work) rtnl_mutex.wait_lock irq_context: 0 (wq_completion)bond93#2 (work_completion)(&(&slave->notify_work)->work) &p->pi_lock irq_context: 0 (wq_completion)bond94#2 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex fs_reclaim irq_context: 0 (wq_completion)bond94#2 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex fs_reclaim mmu_notifier_invalidate_range_start irq_context: 0 (wq_completion)bond94#2 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &c->lock irq_context: 0 (wq_completion)bond94#2 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &n->list_lock irq_context: 0 (wq_completion)bond94#2 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &n->list_lock &c->lock irq_context: 0 (wq_completion)bond94#2 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex rcu_read_lock &bond->stats_lock/1 irq_context: 0 (wq_completion)bond94#2 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex rcu_read_lock &ndev->lock irq_context: 0 (wq_completion)bond94#2 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex nl_table_lock irq_context: 0 (wq_completion)bond94#2 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &obj_hash[i].lock irq_context: 0 (wq_completion)bond94#2 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex nl_table_wait.lock irq_context: 0 (wq_completion)bond94#2 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex net_rwsem irq_context: 0 (wq_completion)bond94#2 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex net_rwsem &list->lock#2 irq_context: 0 (wq_completion)bond94#2 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &tn->lock irq_context: 0 (wq_completion)bond94#2 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &idev->mc_lock irq_context: 0 (wq_completion)bond94#2 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex.wait_lock irq_context: 0 (wq_completion)bond94#2 (work_completion)(&(&bond->mcast_work)->work) &p->pi_lock irq_context: 0 (wq_completion)bond89#2 (work_completion)(&(&bond->mcast_work)->work) &rq->__lock irq_context: 0 (wq_completion)bond111#2 irq_context: 0 (wq_completion)bond111#2 (work_completion)(&(&slave->notify_work)->work) irq_context: 0 (wq_completion)bond111#2 (work_completion)(&(&slave->notify_work)->work) &obj_hash[i].lock irq_context: 0 (wq_completion)bond111#2 (work_completion)(&(&slave->notify_work)->work) &base->lock irq_context: 0 (wq_completion)bond111#2 (work_completion)(&(&slave->notify_work)->work) &base->lock &obj_hash[i].lock irq_context: 0 (wq_completion)bond111#2 (work_completion)(&(&slave->notify_work)->work) &rq->__lock irq_context: 0 (wq_completion)bond111#2 (work_completion)(&(&slave->notify_work)->work) &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)bond122 (work_completion)(&(&bond->mcast_work)->work) &rq->__lock irq_context: 0 (wq_completion)bond122 (work_completion)(&(&bond->mcast_work)->work) &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)bond111#2 (work_completion)(&(&bond->mcast_work)->work) irq_context: 0 (wq_completion)bond111#2 (work_completion)(&(&bond->mcast_work)->work) &obj_hash[i].lock irq_context: 0 (wq_completion)bond111#2 (work_completion)(&(&bond->mcast_work)->work) &base->lock irq_context: 0 (wq_completion)bond111#2 (work_completion)(&(&bond->mcast_work)->work) &base->lock &obj_hash[i].lock irq_context: 0 rtnl_mutex cpu_hotplug_lock wq_pool_mutex fs_reclaim &rq->__lock irq_context: 0 rtnl_mutex cpu_hotplug_lock wq_pool_mutex fs_reclaim &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)bond124 irq_context: 0 (wq_completion)bond124 (work_completion)(&(&slave->notify_work)->work) irq_context: 0 (wq_completion)bond124 (work_completion)(&(&slave->notify_work)->work) &obj_hash[i].lock irq_context: 0 (wq_completion)bond124 (work_completion)(&(&slave->notify_work)->work) &base->lock irq_context: 0 (wq_completion)bond124 (work_completion)(&(&slave->notify_work)->work) &base->lock &obj_hash[i].lock irq_context: 0 (wq_completion)bond124 (work_completion)(&(&slave->notify_work)->work) &rq->__lock irq_context: 0 (wq_completion)bond124 (work_completion)(&(&slave->notify_work)->work) &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)bond124 (work_completion)(&(&bond->mcast_work)->work) irq_context: 0 (wq_completion)bond124 (work_completion)(&(&bond->mcast_work)->work) &obj_hash[i].lock irq_context: 0 (wq_completion)bond124 (work_completion)(&(&bond->mcast_work)->work) &base->lock irq_context: 0 (wq_completion)bond124 (work_completion)(&(&bond->mcast_work)->work) &base->lock &obj_hash[i].lock irq_context: 0 (wq_completion)mld (work_completion)(&(&idev->mc_ifc_work)->work) &rq->__lock irq_context: 0 (wq_completion)mld (work_completion)(&(&idev->mc_ifc_work)->work) &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)wg-crypt-wg1#2 (work_completion)(&peer->transmit_packet_work) &peer->endpoint_lock rcu_read_lock_bh rcu_read_lock &tbl->lock irq_context: 0 (wq_completion)wg-crypt-wg1#2 (work_completion)(&peer->transmit_packet_work) &peer->endpoint_lock rcu_read_lock_bh rcu_read_lock &n->lock irq_context: 0 (wq_completion)wg-crypt-wg1#2 (work_completion)(&peer->transmit_packet_work) &peer->endpoint_lock rcu_read_lock_bh rcu_read_lock &____s->seqcount#9 irq_context: 0 (wq_completion)ipv6_addrconf (work_completion)(&(&ifa->dad_work)->work) rtnl_mutex &idev->mc_lock &dev_addr_list_lock_key/1 &pcp->lock &zone->lock irq_context: 0 (wq_completion)ipv6_addrconf (work_completion)(&(&ifa->dad_work)->work) rtnl_mutex &idev->mc_lock &dev_addr_list_lock_key/1 &pcp->lock &zone->lock &____s->seqcount irq_context: 0 (wq_completion)bond95#2 irq_context: 0 (wq_completion)bond95#2 (work_completion)(&(&slave->notify_work)->work) irq_context: 0 (wq_completion)bond95#2 (work_completion)(&(&slave->notify_work)->work) &obj_hash[i].lock irq_context: 0 (wq_completion)bond95#2 (work_completion)(&(&slave->notify_work)->work) &base->lock irq_context: 0 (wq_completion)bond95#2 (work_completion)(&(&slave->notify_work)->work) &base->lock &obj_hash[i].lock irq_context: 0 (wq_completion)bond95#2 (work_completion)(&(&slave->notify_work)->work) &rq->__lock irq_context: 0 (wq_completion)bond95#2 (work_completion)(&(&slave->notify_work)->work) &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)bond95#2 (work_completion)(&(&bond->mcast_work)->work) irq_context: 0 (wq_completion)bond95#2 (work_completion)(&(&bond->mcast_work)->work) &obj_hash[i].lock irq_context: 0 (wq_completion)bond95#2 (work_completion)(&(&bond->mcast_work)->work) &base->lock irq_context: 0 (wq_completion)bond95#2 (work_completion)(&(&bond->mcast_work)->work) &base->lock &obj_hash[i].lock irq_context: 0 (wq_completion)bond95#2 (work_completion)(&(&bond->mcast_work)->work) &rq->__lock irq_context: 0 (wq_completion)bond95#2 (work_completion)(&(&bond->mcast_work)->work) &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)bond20#3 irq_context: 0 (wq_completion)bond20#3 (work_completion)(&(&slave->notify_work)->work) irq_context: 0 (wq_completion)bond20#3 (work_completion)(&(&slave->notify_work)->work) &obj_hash[i].lock irq_context: 0 (wq_completion)bond20#3 (work_completion)(&(&slave->notify_work)->work) &base->lock irq_context: 0 (wq_completion)bond20#3 (work_completion)(&(&slave->notify_work)->work) &base->lock &obj_hash[i].lock irq_context: 0 (wq_completion)bond20#3 (work_completion)(&(&slave->notify_work)->work) &rq->__lock irq_context: 0 (wq_completion)bond20#3 (work_completion)(&(&slave->notify_work)->work) &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)bond95#2 (work_completion)(&(&slave->notify_work)->work) fill_pool_map-wait-type-override pool_lock irq_context: 0 (wq_completion)bond20#3 (work_completion)(&(&bond->mcast_work)->work) irq_context: 0 (wq_completion)bond20#3 (work_completion)(&(&bond->mcast_work)->work) &obj_hash[i].lock irq_context: 0 (wq_completion)bond20#3 (work_completion)(&(&bond->mcast_work)->work) &base->lock irq_context: 0 (wq_completion)bond20#3 (work_completion)(&(&bond->mcast_work)->work) &base->lock &obj_hash[i].lock irq_context: 0 (wq_completion)bond20#3 (work_completion)(&(&bond->mcast_work)->work) &rq->__lock irq_context: 0 (wq_completion)bond20#3 (work_completion)(&(&bond->mcast_work)->work) &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)bond112#2 irq_context: 0 (wq_completion)bond112#2 (work_completion)(&(&slave->notify_work)->work) irq_context: 0 (wq_completion)bond112#2 (work_completion)(&(&slave->notify_work)->work) &obj_hash[i].lock irq_context: 0 (wq_completion)bond112#2 (work_completion)(&(&slave->notify_work)->work) &base->lock irq_context: 0 (wq_completion)bond112#2 (work_completion)(&(&slave->notify_work)->work) &base->lock &obj_hash[i].lock irq_context: 0 (wq_completion)bond112#2 (work_completion)(&(&slave->notify_work)->work) &rq->__lock irq_context: 0 (wq_completion)bond112#2 (work_completion)(&(&slave->notify_work)->work) &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)bond122 (work_completion)(&(&slave->notify_work)->work) &rq->__lock irq_context: 0 (wq_completion)bond122 (work_completion)(&(&slave->notify_work)->work) &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)bond112#2 (work_completion)(&(&bond->mcast_work)->work) irq_context: 0 (wq_completion)bond112#2 (work_completion)(&(&bond->mcast_work)->work) &obj_hash[i].lock irq_context: 0 (wq_completion)bond112#2 (work_completion)(&(&bond->mcast_work)->work) &base->lock irq_context: 0 (wq_completion)bond112#2 (work_completion)(&(&bond->mcast_work)->work) &base->lock &obj_hash[i].lock irq_context: 0 (wq_completion)bond125 irq_context: 0 (wq_completion)bond125 (work_completion)(&(&slave->notify_work)->work) irq_context: 0 (wq_completion)bond125 (work_completion)(&(&slave->notify_work)->work) &obj_hash[i].lock irq_context: 0 (wq_completion)bond125 (work_completion)(&(&slave->notify_work)->work) &base->lock irq_context: 0 (wq_completion)bond125 (work_completion)(&(&slave->notify_work)->work) &base->lock &obj_hash[i].lock irq_context: 0 (wq_completion)bond125 (work_completion)(&(&bond->mcast_work)->work) irq_context: 0 (wq_completion)bond125 (work_completion)(&(&bond->mcast_work)->work) &obj_hash[i].lock irq_context: 0 (wq_completion)bond125 (work_completion)(&(&bond->mcast_work)->work) &base->lock irq_context: 0 (wq_completion)bond125 (work_completion)(&(&bond->mcast_work)->work) &base->lock &obj_hash[i].lock irq_context: 0 (wq_completion)bond96#3 irq_context: 0 (wq_completion)bond96#3 (work_completion)(&(&slave->notify_work)->work) irq_context: 0 (wq_completion)bond96#3 (work_completion)(&(&slave->notify_work)->work) &obj_hash[i].lock irq_context: 0 (wq_completion)bond96#3 (work_completion)(&(&slave->notify_work)->work) &base->lock irq_context: 0 (wq_completion)bond96#3 (work_completion)(&(&slave->notify_work)->work) &base->lock &obj_hash[i].lock irq_context: 0 (wq_completion)bond96#3 (work_completion)(&(&slave->notify_work)->work) &rq->__lock irq_context: 0 (wq_completion)bond96#3 (work_completion)(&(&slave->notify_work)->work) &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)bond96#3 (work_completion)(&(&bond->mcast_work)->work) irq_context: 0 (wq_completion)bond96#3 (work_completion)(&(&bond->mcast_work)->work) &obj_hash[i].lock irq_context: 0 (wq_completion)bond96#3 (work_completion)(&(&bond->mcast_work)->work) &base->lock irq_context: 0 (wq_completion)bond96#3 (work_completion)(&(&bond->mcast_work)->work) &base->lock &obj_hash[i].lock irq_context: 0 (wq_completion)bond125 (work_completion)(&(&bond->mcast_work)->work) fill_pool_map-wait-type-override pool_lock irq_context: 0 (wq_completion)bond113#2 irq_context: 0 (wq_completion)bond113#2 (work_completion)(&(&slave->notify_work)->work) irq_context: 0 (wq_completion)bond113#2 (work_completion)(&(&slave->notify_work)->work) &obj_hash[i].lock irq_context: 0 (wq_completion)bond113#2 (work_completion)(&(&slave->notify_work)->work) &base->lock irq_context: 0 (wq_completion)bond113#2 (work_completion)(&(&slave->notify_work)->work) &base->lock &obj_hash[i].lock irq_context: 0 (wq_completion)bond113#2 (work_completion)(&(&slave->notify_work)->work) &rq->__lock irq_context: 0 (wq_completion)bond113#2 (work_completion)(&(&slave->notify_work)->work) &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)bond113#2 (work_completion)(&(&bond->mcast_work)->work) irq_context: 0 (wq_completion)bond113#2 (work_completion)(&(&bond->mcast_work)->work) &obj_hash[i].lock irq_context: 0 (wq_completion)bond113#2 (work_completion)(&(&bond->mcast_work)->work) &base->lock irq_context: 0 (wq_completion)bond113#2 (work_completion)(&(&bond->mcast_work)->work) &base->lock &obj_hash[i].lock irq_context: 0 (wq_completion)bond21#3 irq_context: 0 (wq_completion)bond21#3 (work_completion)(&(&slave->notify_work)->work) irq_context: 0 (wq_completion)bond21#3 (work_completion)(&(&slave->notify_work)->work) &obj_hash[i].lock irq_context: 0 (wq_completion)bond21#3 (work_completion)(&(&slave->notify_work)->work) &base->lock irq_context: 0 (wq_completion)bond21#3 (work_completion)(&(&slave->notify_work)->work) &base->lock &obj_hash[i].lock irq_context: 0 (wq_completion)bond21#3 (work_completion)(&(&bond->mcast_work)->work) irq_context: 0 (wq_completion)bond21#3 (work_completion)(&(&bond->mcast_work)->work) &obj_hash[i].lock irq_context: 0 (wq_completion)bond21#3 (work_completion)(&(&bond->mcast_work)->work) &base->lock irq_context: 0 (wq_completion)bond21#3 (work_completion)(&(&bond->mcast_work)->work) &base->lock &obj_hash[i].lock irq_context: 0 (wq_completion)bond21#3 (work_completion)(&(&bond->mcast_work)->work) &rq->__lock irq_context: 0 (wq_completion)bond21#3 (work_completion)(&(&bond->mcast_work)->work) &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)bond113#2 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex fs_reclaim irq_context: 0 (wq_completion)bond113#2 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex fs_reclaim mmu_notifier_invalidate_range_start irq_context: 0 (wq_completion)bond113#2 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &c->lock irq_context: 0 (wq_completion)bond113#2 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &n->list_lock irq_context: 0 (wq_completion)bond113#2 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &n->list_lock &c->lock irq_context: 0 (wq_completion)bond113#2 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &rq->__lock irq_context: 0 (wq_completion)bond113#2 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)bond113#2 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex rcu_read_lock &bond->stats_lock/1 irq_context: 0 (wq_completion)bond113#2 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex rcu_read_lock &ndev->lock irq_context: 0 (wq_completion)bond113#2 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex nl_table_lock irq_context: 0 (wq_completion)bond113#2 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &obj_hash[i].lock irq_context: 0 (wq_completion)bond113#2 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex nl_table_wait.lock irq_context: 0 (wq_completion)bond113#2 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex net_rwsem irq_context: 0 (wq_completion)bond113#2 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex net_rwsem &list->lock#2 irq_context: 0 (wq_completion)bond113#2 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &tn->lock irq_context: 0 (wq_completion)bond113#2 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &idev->mc_lock irq_context: 0 (wq_completion)bond113#2 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex.wait_lock irq_context: 0 (wq_completion)bond113#2 (work_completion)(&(&bond->mcast_work)->work) &p->pi_lock irq_context: 0 (wq_completion)bond113#2 (work_completion)(&(&bond->mcast_work)->work) &p->pi_lock &rq->__lock irq_context: 0 (wq_completion)bond113#2 (work_completion)(&(&bond->mcast_work)->work) &p->pi_lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)bond113#2 (work_completion)(&(&bond->mcast_work)->work) &rq->__lock irq_context: 0 (wq_completion)bond113#2 (work_completion)(&(&bond->mcast_work)->work) &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)bond113#2 (work_completion)(&(&slave->notify_work)->work) rtnl_mutex net_rwsem irq_context: 0 (wq_completion)bond113#2 (work_completion)(&(&slave->notify_work)->work) rtnl_mutex net_rwsem &list->lock#2 irq_context: 0 (wq_completion)bond113#2 (work_completion)(&(&slave->notify_work)->work) rtnl_mutex &tn->lock irq_context: 0 (wq_completion)bond113#2 (work_completion)(&(&slave->notify_work)->work) rtnl_mutex.wait_lock irq_context: 0 (wq_completion)bond113#2 (work_completion)(&(&slave->notify_work)->work) &p->pi_lock irq_context: 0 (wq_completion)bond126 irq_context: 0 (wq_completion)bond126 &rq->__lock irq_context: 0 (wq_completion)bond126 &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)bond126 (work_completion)(&(&slave->notify_work)->work) irq_context: 0 (wq_completion)bond126 (work_completion)(&(&slave->notify_work)->work) &obj_hash[i].lock irq_context: 0 (wq_completion)bond126 (work_completion)(&(&slave->notify_work)->work) &base->lock irq_context: 0 (wq_completion)bond126 (work_completion)(&(&slave->notify_work)->work) &base->lock &obj_hash[i].lock irq_context: 0 (wq_completion)bond126 (work_completion)(&(&slave->notify_work)->work) &rq->__lock irq_context: 0 (wq_completion)bond126 (work_completion)(&(&slave->notify_work)->work) &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)bond126 (work_completion)(&(&bond->mcast_work)->work) irq_context: 0 (wq_completion)bond126 (work_completion)(&(&bond->mcast_work)->work) &obj_hash[i].lock irq_context: 0 (wq_completion)bond126 (work_completion)(&(&bond->mcast_work)->work) &base->lock irq_context: 0 (wq_completion)bond126 (work_completion)(&(&bond->mcast_work)->work) &base->lock &obj_hash[i].lock irq_context: 0 (wq_completion)bond126 (work_completion)(&(&bond->mcast_work)->work) &rq->__lock irq_context: 0 (wq_completion)bond126 (work_completion)(&(&bond->mcast_work)->work) &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)bond126 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex fs_reclaim irq_context: 0 (wq_completion)bond126 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex fs_reclaim mmu_notifier_invalidate_range_start irq_context: 0 (wq_completion)bond126 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex rcu_read_lock &bond->stats_lock/1 irq_context: 0 (wq_completion)bond126 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex rcu_read_lock &ndev->lock irq_context: 0 (wq_completion)bond126 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex nl_table_lock irq_context: 0 (wq_completion)bond126 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &obj_hash[i].lock irq_context: 0 (wq_completion)bond126 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex nl_table_wait.lock irq_context: 0 (wq_completion)bond126 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex net_rwsem irq_context: 0 (wq_completion)bond126 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex net_rwsem &list->lock#2 irq_context: 0 (wq_completion)bond126 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &tn->lock irq_context: 0 (wq_completion)bond126 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &idev->mc_lock irq_context: 0 (wq_completion)bond126 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex.wait_lock irq_context: 0 (wq_completion)bond126 (work_completion)(&(&bond->mcast_work)->work) &p->pi_lock irq_context: 0 (wq_completion)bond111#2 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex fs_reclaim irq_context: 0 (wq_completion)bond111#2 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex fs_reclaim mmu_notifier_invalidate_range_start irq_context: 0 (wq_completion)bond111#2 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &c->lock irq_context: 0 (wq_completion)bond111#2 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex rcu_read_lock &bond->stats_lock/1 irq_context: 0 (wq_completion)bond111#2 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex rcu_read_lock &ndev->lock irq_context: 0 (wq_completion)bond111#2 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex nl_table_lock irq_context: 0 (wq_completion)bond111#2 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &obj_hash[i].lock irq_context: 0 (wq_completion)bond111#2 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex nl_table_wait.lock irq_context: 0 (wq_completion)bond111#2 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex net_rwsem irq_context: 0 (wq_completion)bond111#2 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex net_rwsem &list->lock#2 irq_context: 0 (wq_completion)bond111#2 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &tn->lock irq_context: 0 (wq_completion)bond111#2 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &idev->mc_lock irq_context: 0 (wq_completion)bond111#2 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &idev->mc_lock fs_reclaim irq_context: 0 (wq_completion)bond111#2 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &idev->mc_lock fs_reclaim mmu_notifier_invalidate_range_start irq_context: 0 (wq_completion)bond111#2 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &idev->mc_lock &c->lock irq_context: 0 (wq_completion)bond111#2 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &idev->mc_lock rcu_read_lock &ndev->lock irq_context: 0 (wq_completion)bond111#2 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &idev->mc_lock rcu_read_lock &dir->lock#2 irq_context: 0 (wq_completion)bond111#2 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &idev->mc_lock rcu_read_lock &ul->lock#2 irq_context: 0 (wq_completion)bond111#2 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &idev->mc_lock rcu_read_lock rcu_read_lock &____s->seqcount#10 irq_context: 0 (wq_completion)bond111#2 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &idev->mc_lock rcu_read_lock rcu_read_lock rcu_read_lock_bh rcu_read_lock rcu_read_lock_bh rcu_read_lock &br->multicast_lock irq_context: 0 (wq_completion)bond111#2 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &idev->mc_lock rcu_read_lock rcu_read_lock rcu_read_lock_bh rcu_read_lock rcu_read_lock_bh rcu_read_lock &br->multicast_lock &obj_hash[i].lock irq_context: 0 (wq_completion)bond111#2 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &idev->mc_lock rcu_read_lock rcu_read_lock rcu_read_lock_bh rcu_read_lock rcu_read_lock_bh rcu_read_lock &br->multicast_lock &base->lock irq_context: 0 (wq_completion)bond111#2 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &idev->mc_lock rcu_read_lock rcu_read_lock rcu_read_lock_bh rcu_read_lock rcu_read_lock_bh rcu_read_lock &data->lock irq_context: 0 (wq_completion)bond111#2 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &idev->mc_lock rcu_read_lock rcu_read_lock rcu_read_lock_bh rcu_read_lock rcu_read_lock_bh rcu_read_lock &obj_hash[i].lock irq_context: 0 (wq_completion)bond111#2 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex.wait_lock irq_context: 0 (wq_completion)bond111#2 (work_completion)(&(&bond->mcast_work)->work) &p->pi_lock irq_context: 0 (wq_completion)bond95#2 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex fs_reclaim irq_context: 0 (wq_completion)bond95#2 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex fs_reclaim mmu_notifier_invalidate_range_start irq_context: 0 (wq_completion)bond95#2 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex rcu_read_lock &bond->stats_lock/1 irq_context: 0 (wq_completion)bond95#2 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex rcu_read_lock &ndev->lock irq_context: 0 (wq_completion)bond95#2 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex nl_table_lock irq_context: 0 (wq_completion)bond95#2 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &obj_hash[i].lock irq_context: 0 (wq_completion)bond95#2 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex nl_table_wait.lock irq_context: 0 (wq_completion)bond95#2 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &rq->__lock irq_context: 0 (wq_completion)bond95#2 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)bond95#2 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex net_rwsem irq_context: 0 (wq_completion)bond95#2 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex net_rwsem &list->lock#2 irq_context: 0 (wq_completion)bond95#2 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &tn->lock irq_context: 0 (wq_completion)bond95#2 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &idev->mc_lock irq_context: 0 (wq_completion)bond95#2 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex.wait_lock irq_context: 0 (wq_completion)bond95#2 (work_completion)(&(&bond->mcast_work)->work) &p->pi_lock irq_context: 0 (wq_completion)bond97#3 irq_context: 0 (wq_completion)bond97#3 (work_completion)(&(&slave->notify_work)->work) irq_context: 0 (wq_completion)bond97#3 (work_completion)(&(&slave->notify_work)->work) &obj_hash[i].lock irq_context: 0 (wq_completion)bond97#3 (work_completion)(&(&slave->notify_work)->work) &base->lock irq_context: 0 (wq_completion)bond97#3 (work_completion)(&(&slave->notify_work)->work) &base->lock &obj_hash[i].lock irq_context: 0 (wq_completion)bond97#3 (work_completion)(&(&slave->notify_work)->work) &rq->__lock irq_context: 0 (wq_completion)bond97#3 (work_completion)(&(&slave->notify_work)->work) &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)bond97#3 (work_completion)(&(&bond->mcast_work)->work) irq_context: 0 (wq_completion)bond97#3 (work_completion)(&(&bond->mcast_work)->work) &obj_hash[i].lock irq_context: 0 (wq_completion)bond97#3 (work_completion)(&(&bond->mcast_work)->work) &base->lock irq_context: 0 (wq_completion)bond97#3 (work_completion)(&(&bond->mcast_work)->work) &base->lock &obj_hash[i].lock irq_context: 0 (wq_completion)bond114#2 irq_context: 0 (wq_completion)bond114#2 (work_completion)(&(&slave->notify_work)->work) irq_context: 0 (wq_completion)bond114#2 (work_completion)(&(&slave->notify_work)->work) &obj_hash[i].lock irq_context: 0 (wq_completion)bond114#2 (work_completion)(&(&slave->notify_work)->work) &base->lock irq_context: 0 (wq_completion)bond114#2 (work_completion)(&(&slave->notify_work)->work) &base->lock &obj_hash[i].lock irq_context: 0 (wq_completion)bond114#2 (work_completion)(&(&slave->notify_work)->work) &rq->__lock irq_context: 0 (wq_completion)bond114#2 (work_completion)(&(&slave->notify_work)->work) &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)bond114#2 (work_completion)(&(&bond->mcast_work)->work) irq_context: 0 (wq_completion)bond114#2 (work_completion)(&(&bond->mcast_work)->work) &obj_hash[i].lock irq_context: 0 (wq_completion)bond114#2 (work_completion)(&(&bond->mcast_work)->work) &base->lock irq_context: 0 (wq_completion)bond114#2 (work_completion)(&(&bond->mcast_work)->work) &base->lock &obj_hash[i].lock irq_context: 0 (wq_completion)bond114#2 (work_completion)(&(&bond->mcast_work)->work) &rq->__lock irq_context: 0 (wq_completion)bond114#2 (work_completion)(&(&bond->mcast_work)->work) &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)bond97#3 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex fs_reclaim irq_context: 0 (wq_completion)bond97#3 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex fs_reclaim mmu_notifier_invalidate_range_start irq_context: 0 (wq_completion)bond97#3 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex rcu_read_lock &bond->stats_lock/1 irq_context: 0 (wq_completion)bond97#3 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex rcu_read_lock &ndev->lock irq_context: 0 (wq_completion)bond97#3 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex nl_table_lock irq_context: 0 (wq_completion)bond97#3 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &obj_hash[i].lock irq_context: 0 (wq_completion)bond97#3 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex nl_table_wait.lock irq_context: 0 (wq_completion)bond97#3 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex net_rwsem irq_context: 0 (wq_completion)bond97#3 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex net_rwsem &rq->__lock irq_context: 0 (wq_completion)bond97#3 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex net_rwsem &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)bond97#3 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex net_rwsem &list->lock#2 irq_context: 0 (wq_completion)bond97#3 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &tn->lock irq_context: 0 (wq_completion)bond97#3 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &idev->mc_lock irq_context: 0 (wq_completion)bond97#3 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex.wait_lock irq_context: 0 (wq_completion)bond97#3 (work_completion)(&(&bond->mcast_work)->work) &p->pi_lock irq_context: 0 (wq_completion)bond114#2 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex fs_reclaim irq_context: 0 (wq_completion)bond114#2 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex fs_reclaim mmu_notifier_invalidate_range_start irq_context: 0 (wq_completion)bond114#2 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex fs_reclaim mmu_notifier_invalidate_range_start &rq->__lock irq_context: 0 (wq_completion)bond114#2 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex fs_reclaim mmu_notifier_invalidate_range_start &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)bond114#2 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex rcu_read_lock &bond->stats_lock/1 irq_context: 0 (wq_completion)bond114#2 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex rcu_read_lock &ndev->lock irq_context: 0 (wq_completion)bond114#2 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex nl_table_lock irq_context: 0 (wq_completion)bond114#2 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &obj_hash[i].lock irq_context: 0 (wq_completion)bond114#2 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex nl_table_wait.lock irq_context: 0 (wq_completion)bond114#2 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex net_rwsem irq_context: 0 (wq_completion)bond114#2 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex net_rwsem &list->lock#2 irq_context: 0 (wq_completion)bond114#2 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &tn->lock irq_context: 0 (wq_completion)bond114#2 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &idev->mc_lock irq_context: 0 (wq_completion)bond114#2 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex rcu_read_lock rcu_node_0 irq_context: 0 (wq_completion)bond114#2 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex rcu_read_lock &rq->__lock irq_context: 0 (wq_completion)bond114#2 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex rcu_read_lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)bond114#2 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex.wait_lock irq_context: 0 (wq_completion)bond114#2 (work_completion)(&(&bond->mcast_work)->work) &p->pi_lock irq_context: 0 (wq_completion)bond114#2 (work_completion)(&(&bond->mcast_work)->work) &p->pi_lock &rq->__lock irq_context: 0 (wq_completion)bond114#2 (work_completion)(&(&bond->mcast_work)->work) &p->pi_lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)bond22#4 irq_context: 0 (wq_completion)bond22#4 (work_completion)(&(&slave->notify_work)->work) irq_context: 0 (wq_completion)bond22#4 (work_completion)(&(&slave->notify_work)->work) &obj_hash[i].lock irq_context: 0 (wq_completion)bond22#4 (work_completion)(&(&slave->notify_work)->work) &base->lock irq_context: 0 (wq_completion)bond22#4 (work_completion)(&(&slave->notify_work)->work) &base->lock &obj_hash[i].lock irq_context: 0 (wq_completion)bond22#4 (work_completion)(&(&bond->mcast_work)->work) irq_context: 0 (wq_completion)bond22#4 (work_completion)(&(&bond->mcast_work)->work) &rq->__lock irq_context: 0 (wq_completion)bond22#4 (work_completion)(&(&bond->mcast_work)->work) &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)bond22#4 (work_completion)(&(&bond->mcast_work)->work) &obj_hash[i].lock irq_context: 0 (wq_completion)bond22#4 (work_completion)(&(&bond->mcast_work)->work) &base->lock irq_context: 0 (wq_completion)bond22#4 (work_completion)(&(&bond->mcast_work)->work) &base->lock &obj_hash[i].lock irq_context: 0 (wq_completion)bond125 (work_completion)(&(&slave->notify_work)->work) &base->lock fill_pool_map-wait-type-override pool_lock#2 irq_context: 0 (wq_completion)bond125 (work_completion)(&(&slave->notify_work)->work) &base->lock fill_pool_map-wait-type-override pool_lock irq_context: 0 (wq_completion)bond125 (work_completion)(&(&slave->notify_work)->work) &rq->__lock irq_context: 0 dup_mmap_sem &mm->mmap_lock &mm->mmap_lock/1 fs_reclaim pool_lock#2 irq_context: 0 (wq_completion)bond127 irq_context: 0 (wq_completion)bond127 (work_completion)(&(&slave->notify_work)->work) irq_context: 0 (wq_completion)bond127 (work_completion)(&(&slave->notify_work)->work) &obj_hash[i].lock irq_context: 0 (wq_completion)bond127 (work_completion)(&(&slave->notify_work)->work) &base->lock irq_context: 0 (wq_completion)bond127 (work_completion)(&(&slave->notify_work)->work) &base->lock &obj_hash[i].lock irq_context: 0 (wq_completion)bond127 (work_completion)(&(&slave->notify_work)->work) &rq->__lock irq_context: 0 (wq_completion)bond127 (work_completion)(&(&bond->mcast_work)->work) irq_context: 0 (wq_completion)bond127 (work_completion)(&(&bond->mcast_work)->work) &obj_hash[i].lock irq_context: 0 (wq_completion)bond127 (work_completion)(&(&bond->mcast_work)->work) &base->lock irq_context: 0 (wq_completion)bond127 (work_completion)(&(&bond->mcast_work)->work) &base->lock &obj_hash[i].lock irq_context: 0 (wq_completion)bond127 (work_completion)(&(&bond->mcast_work)->work) &rq->__lock irq_context: 0 (wq_completion)mld (work_completion)(&(&idev->mc_ifc_work)->work) &idev->mc_lock rcu_read_lock &obj_hash[i].lock pool_lock irq_context: 0 (wq_completion)bond21#3 (work_completion)(&(&slave->notify_work)->work) &rq->__lock irq_context: 0 (wq_completion)bond21#3 (work_completion)(&(&slave->notify_work)->work) &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)bond98#3 irq_context: 0 (wq_completion)bond98#3 (work_completion)(&(&slave->notify_work)->work) irq_context: 0 (wq_completion)bond98#3 (work_completion)(&(&slave->notify_work)->work) &obj_hash[i].lock irq_context: 0 (wq_completion)bond98#3 (work_completion)(&(&slave->notify_work)->work) &base->lock irq_context: 0 (wq_completion)bond98#3 (work_completion)(&(&slave->notify_work)->work) &base->lock &obj_hash[i].lock irq_context: 0 (wq_completion)bond98#3 (work_completion)(&(&slave->notify_work)->work) &rq->__lock irq_context: 0 (wq_completion)bond98#3 (work_completion)(&(&slave->notify_work)->work) &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)bond98#3 (work_completion)(&(&bond->mcast_work)->work) irq_context: 0 (wq_completion)bond98#3 (work_completion)(&(&bond->mcast_work)->work) &obj_hash[i].lock irq_context: 0 (wq_completion)bond98#3 (work_completion)(&(&bond->mcast_work)->work) &base->lock irq_context: 0 (wq_completion)bond98#3 (work_completion)(&(&bond->mcast_work)->work) &base->lock &obj_hash[i].lock irq_context: softirq (&ndev->rs_timer) &obj_hash[i].lock pool_lock irq_context: 0 (wq_completion)bond115#2 irq_context: 0 (wq_completion)bond115#2 (work_completion)(&(&slave->notify_work)->work) irq_context: 0 (wq_completion)bond115#2 (work_completion)(&(&slave->notify_work)->work) &obj_hash[i].lock irq_context: 0 (wq_completion)bond115#2 (work_completion)(&(&slave->notify_work)->work) &base->lock irq_context: 0 (wq_completion)bond115#2 (work_completion)(&(&slave->notify_work)->work) &base->lock &obj_hash[i].lock irq_context: 0 (wq_completion)bond115#2 (work_completion)(&(&bond->mcast_work)->work) irq_context: 0 (wq_completion)bond115#2 (work_completion)(&(&bond->mcast_work)->work) &obj_hash[i].lock irq_context: 0 (wq_completion)bond115#2 (work_completion)(&(&bond->mcast_work)->work) &base->lock irq_context: 0 (wq_completion)bond115#2 (work_completion)(&(&bond->mcast_work)->work) &base->lock &obj_hash[i].lock irq_context: 0 (wq_completion)bond115#2 (work_completion)(&(&bond->mcast_work)->work) &rq->__lock irq_context: 0 (wq_completion)bond115#2 (work_completion)(&(&bond->mcast_work)->work) &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)bond128 irq_context: 0 (wq_completion)bond128 (work_completion)(&(&slave->notify_work)->work) irq_context: 0 (wq_completion)bond128 (work_completion)(&(&slave->notify_work)->work) &obj_hash[i].lock irq_context: 0 (wq_completion)bond128 (work_completion)(&(&slave->notify_work)->work) &base->lock irq_context: 0 (wq_completion)bond128 (work_completion)(&(&slave->notify_work)->work) &base->lock &obj_hash[i].lock irq_context: 0 (wq_completion)bond128 (work_completion)(&(&bond->mcast_work)->work) irq_context: 0 (wq_completion)bond128 (work_completion)(&(&bond->mcast_work)->work) &obj_hash[i].lock irq_context: 0 (wq_completion)bond128 (work_completion)(&(&bond->mcast_work)->work) &base->lock irq_context: 0 (wq_completion)bond128 (work_completion)(&(&bond->mcast_work)->work) &base->lock &obj_hash[i].lock irq_context: 0 (wq_completion)bond99#3 irq_context: 0 (wq_completion)bond99#3 &rq->__lock irq_context: 0 (wq_completion)bond99#3 &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)bond99#3 (work_completion)(&(&slave->notify_work)->work) irq_context: 0 (wq_completion)bond99#3 (work_completion)(&(&slave->notify_work)->work) &obj_hash[i].lock irq_context: 0 (wq_completion)bond99#3 (work_completion)(&(&slave->notify_work)->work) &base->lock irq_context: 0 (wq_completion)bond99#3 (work_completion)(&(&slave->notify_work)->work) &base->lock &obj_hash[i].lock irq_context: 0 (wq_completion)mld (work_completion)(&(&idev->mc_ifc_work)->work) &idev->mc_lock rcu_read_lock &data->lock &obj_hash[i].lock irq_context: 0 (wq_completion)mld (work_completion)(&(&idev->mc_ifc_work)->work) &idev->mc_lock rcu_read_lock &data->lock &base->lock irq_context: 0 (wq_completion)mld (work_completion)(&(&idev->mc_ifc_work)->work) &idev->mc_lock rcu_read_lock &data->lock &base->lock &obj_hash[i].lock irq_context: 0 (wq_completion)bond99#3 (work_completion)(&(&bond->mcast_work)->work) irq_context: 0 (wq_completion)bond99#3 (work_completion)(&(&bond->mcast_work)->work) &obj_hash[i].lock irq_context: 0 (wq_completion)bond99#3 (work_completion)(&(&bond->mcast_work)->work) &base->lock irq_context: 0 (wq_completion)bond99#3 (work_completion)(&(&bond->mcast_work)->work) &base->lock &obj_hash[i].lock irq_context: 0 (wq_completion)bond99#3 (work_completion)(&(&bond->mcast_work)->work) &rq->__lock irq_context: 0 (wq_completion)ipv6_addrconf (work_completion)(&(&ifa->dad_work)->work) rtnl_mutex &idev->mc_lock rcu_read_lock rcu_read_lock rcu_read_lock_bh rcu_read_lock rcu_read_lock_bh rcu_read_lock quarantine_lock irq_context: 0 (wq_completion)bond23#3 irq_context: 0 (wq_completion)bond23#3 (work_completion)(&(&slave->notify_work)->work) irq_context: 0 (wq_completion)bond23#3 (work_completion)(&(&slave->notify_work)->work) &obj_hash[i].lock irq_context: 0 (wq_completion)bond23#3 (work_completion)(&(&slave->notify_work)->work) &base->lock irq_context: 0 (wq_completion)bond23#3 (work_completion)(&(&slave->notify_work)->work) &base->lock &obj_hash[i].lock irq_context: 0 (wq_completion)bond23#3 (work_completion)(&(&bond->mcast_work)->work) irq_context: 0 (wq_completion)bond23#3 (work_completion)(&(&bond->mcast_work)->work) &obj_hash[i].lock irq_context: 0 (wq_completion)bond23#3 (work_completion)(&(&bond->mcast_work)->work) &base->lock irq_context: 0 (wq_completion)bond23#3 (work_completion)(&(&bond->mcast_work)->work) &base->lock &obj_hash[i].lock irq_context: 0 (wq_completion)bond23#3 (work_completion)(&(&bond->mcast_work)->work) &rq->__lock irq_context: 0 (wq_completion)bond23#3 (work_completion)(&(&bond->mcast_work)->work) &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)bond116#2 irq_context: 0 (wq_completion)bond116#2 &rq->__lock irq_context: 0 (wq_completion)bond116#2 &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)bond116#2 (work_completion)(&(&slave->notify_work)->work) irq_context: 0 (wq_completion)bond116#2 (work_completion)(&(&slave->notify_work)->work) &obj_hash[i].lock irq_context: 0 (wq_completion)bond116#2 (work_completion)(&(&slave->notify_work)->work) &base->lock irq_context: 0 (wq_completion)bond116#2 (work_completion)(&(&slave->notify_work)->work) &base->lock &obj_hash[i].lock irq_context: 0 (wq_completion)bond116#2 (work_completion)(&(&slave->notify_work)->work) &rq->__lock irq_context: 0 (wq_completion)bond116#2 (work_completion)(&(&slave->notify_work)->work) &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)bond116#2 (work_completion)(&(&bond->mcast_work)->work) irq_context: 0 (wq_completion)bond116#2 (work_completion)(&(&bond->mcast_work)->work) &obj_hash[i].lock irq_context: 0 (wq_completion)bond116#2 (work_completion)(&(&bond->mcast_work)->work) &base->lock irq_context: 0 (wq_completion)bond116#2 (work_completion)(&(&bond->mcast_work)->work) &base->lock &obj_hash[i].lock irq_context: 0 (wq_completion)bond129 irq_context: 0 (wq_completion)bond129 (work_completion)(&(&slave->notify_work)->work) irq_context: 0 (wq_completion)bond129 (work_completion)(&(&slave->notify_work)->work) &obj_hash[i].lock irq_context: 0 (wq_completion)bond129 (work_completion)(&(&slave->notify_work)->work) &base->lock irq_context: 0 (wq_completion)bond129 (work_completion)(&(&slave->notify_work)->work) &base->lock &obj_hash[i].lock irq_context: 0 (wq_completion)bond129 (work_completion)(&(&bond->mcast_work)->work) irq_context: 0 (wq_completion)bond129 (work_completion)(&(&bond->mcast_work)->work) &obj_hash[i].lock irq_context: 0 (wq_completion)bond129 (work_completion)(&(&bond->mcast_work)->work) &base->lock irq_context: 0 (wq_completion)bond129 (work_completion)(&(&bond->mcast_work)->work) &base->lock &obj_hash[i].lock irq_context: 0 (wq_completion)bond100#3 irq_context: 0 (wq_completion)bond100#3 (work_completion)(&(&slave->notify_work)->work) irq_context: 0 (wq_completion)bond100#3 (work_completion)(&(&slave->notify_work)->work) &obj_hash[i].lock irq_context: 0 (wq_completion)bond100#3 (work_completion)(&(&slave->notify_work)->work) &base->lock irq_context: 0 (wq_completion)bond100#3 (work_completion)(&(&slave->notify_work)->work) &base->lock &obj_hash[i].lock irq_context: 0 (wq_completion)bond100#3 (work_completion)(&(&slave->notify_work)->work) &rq->__lock irq_context: 0 (wq_completion)bond100#3 (work_completion)(&(&slave->notify_work)->work) &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)bond100#3 (work_completion)(&(&bond->mcast_work)->work) irq_context: 0 (wq_completion)bond100#3 (work_completion)(&(&bond->mcast_work)->work) &obj_hash[i].lock irq_context: 0 (wq_completion)bond100#3 (work_completion)(&(&bond->mcast_work)->work) &base->lock irq_context: 0 (wq_completion)bond100#3 (work_completion)(&(&bond->mcast_work)->work) &base->lock &obj_hash[i].lock irq_context: 0 (wq_completion)events (linkwatch_work).work rtnl_mutex nl_table_wait.lock &p->pi_lock irq_context: 0 (wq_completion)events (linkwatch_work).work rtnl_mutex nl_table_wait.lock &p->pi_lock &rq->__lock irq_context: 0 (wq_completion)events (linkwatch_work).work rtnl_mutex nl_table_wait.lock &p->pi_lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)bond96#3 (work_completion)(&(&bond->mcast_work)->work) &rq->__lock irq_context: 0 (wq_completion)bond96#3 (work_completion)(&(&bond->mcast_work)->work) &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)bond117#2 irq_context: 0 (wq_completion)bond117#2 &rq->__lock irq_context: 0 (wq_completion)bond117#2 &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)bond117#2 (work_completion)(&(&slave->notify_work)->work) irq_context: 0 (wq_completion)bond117#2 (work_completion)(&(&slave->notify_work)->work) &obj_hash[i].lock irq_context: 0 (wq_completion)bond117#2 (work_completion)(&(&slave->notify_work)->work) &base->lock irq_context: 0 (wq_completion)bond117#2 (work_completion)(&(&slave->notify_work)->work) &base->lock &obj_hash[i].lock irq_context: 0 (wq_completion)bond117#2 (work_completion)(&(&slave->notify_work)->work) &rq->__lock irq_context: 0 (wq_completion)bond117#2 (work_completion)(&(&slave->notify_work)->work) &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)bond117#2 (work_completion)(&(&bond->mcast_work)->work) irq_context: 0 (wq_completion)bond117#2 (work_completion)(&(&bond->mcast_work)->work) &obj_hash[i].lock irq_context: 0 (wq_completion)bond117#2 (work_completion)(&(&bond->mcast_work)->work) &base->lock irq_context: 0 (wq_completion)bond117#2 (work_completion)(&(&bond->mcast_work)->work) &base->lock &obj_hash[i].lock irq_context: 0 (wq_completion)bond129 (work_completion)(&(&slave->notify_work)->work) &rq->__lock irq_context: 0 (wq_completion)bond23#3 &rq->__lock irq_context: 0 (wq_completion)bond23#3 &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)bond130 irq_context: 0 (wq_completion)bond130 (work_completion)(&(&slave->notify_work)->work) irq_context: 0 (wq_completion)bond130 (work_completion)(&(&slave->notify_work)->work) &obj_hash[i].lock irq_context: 0 (wq_completion)bond130 (work_completion)(&(&slave->notify_work)->work) &base->lock irq_context: 0 (wq_completion)bond130 (work_completion)(&(&slave->notify_work)->work) &base->lock &obj_hash[i].lock irq_context: 0 (wq_completion)bond130 (work_completion)(&(&slave->notify_work)->work) &rq->__lock irq_context: 0 (wq_completion)bond130 (work_completion)(&(&slave->notify_work)->work) &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)bond130 (work_completion)(&(&bond->mcast_work)->work) irq_context: 0 (wq_completion)bond130 (work_completion)(&(&bond->mcast_work)->work) &obj_hash[i].lock irq_context: 0 (wq_completion)bond130 (work_completion)(&(&bond->mcast_work)->work) &base->lock irq_context: 0 (wq_completion)bond130 (work_completion)(&(&bond->mcast_work)->work) &base->lock &obj_hash[i].lock irq_context: 0 (wq_completion)bond130 (work_completion)(&(&bond->mcast_work)->work) &rq->__lock irq_context: 0 (wq_completion)bond130 (work_completion)(&(&bond->mcast_work)->work) &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)bond95#2 (work_completion)(&(&slave->notify_work)->work) rtnl_mutex net_rwsem irq_context: 0 (wq_completion)bond95#2 (work_completion)(&(&slave->notify_work)->work) rtnl_mutex net_rwsem &list->lock#2 irq_context: 0 (wq_completion)bond95#2 (work_completion)(&(&slave->notify_work)->work) rtnl_mutex &tn->lock irq_context: 0 (wq_completion)bond95#2 (work_completion)(&(&slave->notify_work)->work) rtnl_mutex rcu_read_lock rcu_node_0 irq_context: 0 (wq_completion)bond95#2 (work_completion)(&(&slave->notify_work)->work) rtnl_mutex rcu_read_lock &rq->__lock irq_context: 0 (wq_completion)bond95#2 (work_completion)(&(&slave->notify_work)->work) rtnl_mutex rcu_read_lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)bond95#2 (work_completion)(&(&slave->notify_work)->work) rtnl_mutex.wait_lock irq_context: 0 (wq_completion)bond95#2 (work_completion)(&(&slave->notify_work)->work) &p->pi_lock irq_context: 0 (wq_completion)bond95#2 (work_completion)(&(&slave->notify_work)->work) &p->pi_lock &rq->__lock irq_context: 0 (wq_completion)bond95#2 (work_completion)(&(&slave->notify_work)->work) &p->pi_lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)bond24#3 irq_context: 0 (wq_completion)bond24#3 &rq->__lock irq_context: 0 (wq_completion)bond24#3 (work_completion)(&(&slave->notify_work)->work) irq_context: 0 (wq_completion)bond24#3 (work_completion)(&(&slave->notify_work)->work) &rq->__lock irq_context: 0 (wq_completion)bond24#3 (work_completion)(&(&slave->notify_work)->work) &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)bond24#3 (work_completion)(&(&slave->notify_work)->work) &obj_hash[i].lock irq_context: 0 (wq_completion)bond24#3 (work_completion)(&(&slave->notify_work)->work) &base->lock irq_context: 0 (wq_completion)bond24#3 (work_completion)(&(&slave->notify_work)->work) &base->lock &obj_hash[i].lock irq_context: 0 (wq_completion)bond24#3 (work_completion)(&(&bond->mcast_work)->work) irq_context: 0 (wq_completion)bond24#3 (work_completion)(&(&bond->mcast_work)->work) &obj_hash[i].lock irq_context: 0 (wq_completion)bond24#3 (work_completion)(&(&bond->mcast_work)->work) &base->lock irq_context: 0 (wq_completion)bond24#3 (work_completion)(&(&bond->mcast_work)->work) &base->lock &obj_hash[i].lock irq_context: 0 (wq_completion)bond101#3 irq_context: 0 (wq_completion)bond101#3 (work_completion)(&(&slave->notify_work)->work) irq_context: 0 (wq_completion)bond101#3 (work_completion)(&(&slave->notify_work)->work) &obj_hash[i].lock irq_context: 0 (wq_completion)bond101#3 (work_completion)(&(&slave->notify_work)->work) &base->lock irq_context: 0 (wq_completion)bond101#3 (work_completion)(&(&slave->notify_work)->work) &base->lock &obj_hash[i].lock irq_context: 0 (wq_completion)bond101#3 (work_completion)(&(&slave->notify_work)->work) &rq->__lock irq_context: 0 (wq_completion)bond101#3 (work_completion)(&(&slave->notify_work)->work) &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)bond101#3 (work_completion)(&(&bond->mcast_work)->work) irq_context: 0 (wq_completion)bond101#3 (work_completion)(&(&bond->mcast_work)->work) &obj_hash[i].lock irq_context: 0 (wq_completion)bond101#3 (work_completion)(&(&bond->mcast_work)->work) &base->lock irq_context: 0 (wq_completion)bond101#3 (work_completion)(&(&bond->mcast_work)->work) &base->lock &obj_hash[i].lock irq_context: 0 (wq_completion)bond101#3 (work_completion)(&(&bond->mcast_work)->work) &rq->__lock irq_context: 0 (wq_completion)bond101#3 (work_completion)(&(&bond->mcast_work)->work) &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)bond118#2 irq_context: 0 (wq_completion)bond118#2 (work_completion)(&(&slave->notify_work)->work) irq_context: 0 (wq_completion)bond118#2 (work_completion)(&(&slave->notify_work)->work) &obj_hash[i].lock irq_context: 0 (wq_completion)bond118#2 (work_completion)(&(&slave->notify_work)->work) &base->lock irq_context: 0 (wq_completion)bond118#2 (work_completion)(&(&slave->notify_work)->work) &base->lock &obj_hash[i].lock irq_context: 0 (wq_completion)bond118#2 (work_completion)(&(&bond->mcast_work)->work) irq_context: 0 (wq_completion)bond118#2 (work_completion)(&(&bond->mcast_work)->work) &obj_hash[i].lock irq_context: 0 (wq_completion)bond118#2 (work_completion)(&(&bond->mcast_work)->work) &base->lock irq_context: 0 (wq_completion)bond118#2 (work_completion)(&(&bond->mcast_work)->work) &base->lock &obj_hash[i].lock irq_context: 0 (wq_completion)bond118#2 &rq->__lock irq_context: 0 (wq_completion)bond118#2 &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)bond131 irq_context: 0 (wq_completion)bond131 (work_completion)(&(&slave->notify_work)->work) irq_context: 0 (wq_completion)bond131 (work_completion)(&(&slave->notify_work)->work) &obj_hash[i].lock irq_context: 0 (wq_completion)bond131 (work_completion)(&(&slave->notify_work)->work) &base->lock irq_context: 0 (wq_completion)bond131 (work_completion)(&(&slave->notify_work)->work) &base->lock &obj_hash[i].lock irq_context: 0 (wq_completion)bond131 (work_completion)(&(&slave->notify_work)->work) &rq->__lock irq_context: 0 (wq_completion)bond131 (work_completion)(&(&slave->notify_work)->work) &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)events (linkwatch_work).work rtnl_mutex fill_pool_map-wait-type-override &c->lock irq_context: 0 (wq_completion)events (linkwatch_work).work rtnl_mutex fill_pool_map-wait-type-override &____s->seqcount#2 irq_context: 0 (wq_completion)events (linkwatch_work).work rtnl_mutex fill_pool_map-wait-type-override &____s->seqcount irq_context: 0 (wq_completion)bond131 (work_completion)(&(&bond->mcast_work)->work) irq_context: 0 (wq_completion)bond131 (work_completion)(&(&bond->mcast_work)->work) &obj_hash[i].lock irq_context: 0 (wq_completion)bond131 (work_completion)(&(&bond->mcast_work)->work) &base->lock irq_context: 0 (wq_completion)bond131 (work_completion)(&(&bond->mcast_work)->work) &base->lock &obj_hash[i].lock irq_context: 0 rtnl_mutex cpu_hotplug_lock wq_pool_mutex &wq->mutex &rq->__lock irq_context: 0 rtnl_mutex cpu_hotplug_lock wq_pool_mutex &wq->mutex &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)bond102#3 irq_context: 0 (wq_completion)bond102#3 (work_completion)(&(&slave->notify_work)->work) irq_context: 0 (wq_completion)bond102#3 (work_completion)(&(&slave->notify_work)->work) &obj_hash[i].lock irq_context: 0 (wq_completion)bond102#3 (work_completion)(&(&slave->notify_work)->work) &base->lock irq_context: 0 (wq_completion)bond102#3 (work_completion)(&(&slave->notify_work)->work) &base->lock &obj_hash[i].lock irq_context: 0 (wq_completion)bond102#3 (work_completion)(&(&bond->mcast_work)->work) irq_context: 0 (wq_completion)bond102#3 (work_completion)(&(&bond->mcast_work)->work) &rq->__lock irq_context: 0 (wq_completion)bond102#3 (work_completion)(&(&bond->mcast_work)->work) &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)bond102#3 (work_completion)(&(&bond->mcast_work)->work) &obj_hash[i].lock irq_context: 0 (wq_completion)bond102#3 (work_completion)(&(&bond->mcast_work)->work) &base->lock irq_context: 0 (wq_completion)bond102#3 (work_completion)(&(&bond->mcast_work)->work) &base->lock &obj_hash[i].lock irq_context: 0 (wq_completion)bond115#2 (work_completion)(&(&slave->notify_work)->work) &rq->__lock irq_context: 0 (wq_completion)bond115#2 (work_completion)(&(&slave->notify_work)->work) &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)bond119#2 irq_context: 0 (wq_completion)bond119#2 (work_completion)(&(&slave->notify_work)->work) irq_context: 0 (wq_completion)bond119#2 (work_completion)(&(&slave->notify_work)->work) &obj_hash[i].lock irq_context: 0 (wq_completion)bond119#2 (work_completion)(&(&slave->notify_work)->work) &base->lock irq_context: 0 (wq_completion)bond119#2 (work_completion)(&(&slave->notify_work)->work) &base->lock &obj_hash[i].lock irq_context: 0 (wq_completion)bond119#2 (work_completion)(&(&slave->notify_work)->work) &rq->__lock irq_context: 0 (wq_completion)bond119#2 (work_completion)(&(&slave->notify_work)->work) &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)bond119#2 (work_completion)(&(&bond->mcast_work)->work) irq_context: 0 (wq_completion)bond119#2 (work_completion)(&(&bond->mcast_work)->work) &obj_hash[i].lock irq_context: 0 (wq_completion)bond119#2 (work_completion)(&(&bond->mcast_work)->work) &base->lock irq_context: 0 (wq_completion)bond119#2 (work_completion)(&(&bond->mcast_work)->work) &base->lock &obj_hash[i].lock irq_context: 0 (wq_completion)bond21#3 (work_completion)(&(&slave->notify_work)->work) rtnl_mutex net_rwsem irq_context: 0 (wq_completion)bond21#3 (work_completion)(&(&slave->notify_work)->work) rtnl_mutex net_rwsem &list->lock#2 irq_context: 0 (wq_completion)bond21#3 (work_completion)(&(&slave->notify_work)->work) rtnl_mutex &tn->lock irq_context: 0 (wq_completion)bond21#3 (work_completion)(&(&slave->notify_work)->work) rtnl_mutex.wait_lock irq_context: 0 (wq_completion)bond21#3 (work_completion)(&(&slave->notify_work)->work) &p->pi_lock irq_context: 0 (wq_completion)bond115#2 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex fs_reclaim irq_context: 0 (wq_completion)bond115#2 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex fs_reclaim mmu_notifier_invalidate_range_start irq_context: 0 (wq_completion)bond115#2 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &c->lock irq_context: 0 (wq_completion)bond115#2 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &____s->seqcount#2 irq_context: 0 (wq_completion)bond115#2 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &____s->seqcount irq_context: 0 (wq_completion)bond115#2 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex pool_lock#2 irq_context: 0 (wq_completion)bond115#2 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex rcu_read_lock &bond->stats_lock/1 irq_context: 0 (wq_completion)bond115#2 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex rcu_read_lock &ndev->lock irq_context: 0 (wq_completion)bond115#2 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex nl_table_lock irq_context: 0 (wq_completion)bond115#2 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &obj_hash[i].lock irq_context: 0 (wq_completion)bond115#2 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex nl_table_wait.lock irq_context: 0 (wq_completion)bond115#2 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex net_rwsem irq_context: 0 (wq_completion)bond115#2 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex net_rwsem &list->lock#2 irq_context: 0 (wq_completion)bond115#2 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &tn->lock irq_context: 0 (wq_completion)bond115#2 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &idev->mc_lock irq_context: 0 (wq_completion)bond115#2 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &idev->mc_lock fs_reclaim irq_context: 0 (wq_completion)bond115#2 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &idev->mc_lock fs_reclaim mmu_notifier_invalidate_range_start irq_context: 0 (wq_completion)bond115#2 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &idev->mc_lock &c->lock irq_context: 0 (wq_completion)bond115#2 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &idev->mc_lock &n->list_lock irq_context: 0 (wq_completion)bond115#2 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &idev->mc_lock &n->list_lock &c->lock irq_context: 0 (wq_completion)bond115#2 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &idev->mc_lock rcu_read_lock &ndev->lock irq_context: 0 (wq_completion)bond115#2 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &idev->mc_lock rcu_read_lock &c->lock irq_context: 0 (wq_completion)bond115#2 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &idev->mc_lock rcu_read_lock &dir->lock#2 irq_context: 0 (wq_completion)bond115#2 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &idev->mc_lock rcu_read_lock rcu_node_0 irq_context: 0 (wq_completion)bond115#2 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &idev->mc_lock rcu_read_lock &rq->__lock irq_context: 0 (wq_completion)bond115#2 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &idev->mc_lock rcu_read_lock &ul->lock#2 irq_context: 0 (wq_completion)bond115#2 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &idev->mc_lock rcu_read_lock rcu_read_lock &____s->seqcount#10 irq_context: 0 (wq_completion)bond115#2 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &idev->mc_lock rcu_read_lock rcu_read_lock rcu_read_lock_bh rcu_read_lock rcu_read_lock_bh rcu_read_lock &br->multicast_lock irq_context: 0 (wq_completion)bond115#2 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &idev->mc_lock rcu_read_lock rcu_read_lock rcu_read_lock_bh rcu_read_lock rcu_read_lock_bh rcu_read_lock &br->multicast_lock &obj_hash[i].lock irq_context: 0 (wq_completion)bond115#2 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &idev->mc_lock rcu_read_lock rcu_read_lock rcu_read_lock_bh rcu_read_lock rcu_read_lock_bh rcu_read_lock &br->multicast_lock &base->lock irq_context: 0 (wq_completion)bond115#2 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &idev->mc_lock rcu_read_lock rcu_read_lock rcu_read_lock_bh rcu_read_lock rcu_read_lock_bh rcu_read_lock &data->lock irq_context: 0 (wq_completion)bond115#2 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &idev->mc_lock rcu_read_lock rcu_read_lock rcu_read_lock_bh rcu_read_lock rcu_read_lock_bh rcu_read_lock &obj_hash[i].lock irq_context: 0 (wq_completion)bond115#2 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &idev->mc_lock rcu_read_lock &rcu_state.gp_wq irq_context: 0 (wq_completion)bond115#2 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &idev->mc_lock rcu_read_lock &rcu_state.gp_wq &p->pi_lock irq_context: 0 (wq_completion)bond115#2 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &idev->mc_lock rcu_read_lock &rcu_state.gp_wq &p->pi_lock &rq->__lock irq_context: 0 (wq_completion)bond115#2 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &idev->mc_lock rcu_read_lock &rcu_state.gp_wq &p->pi_lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)bond115#2 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex.wait_lock irq_context: 0 (wq_completion)bond115#2 (work_completion)(&(&bond->mcast_work)->work) &p->pi_lock irq_context: 0 (wq_completion)bond115#2 (work_completion)(&(&bond->mcast_work)->work) &p->pi_lock &rq->__lock irq_context: 0 (wq_completion)bond115#2 (work_completion)(&(&bond->mcast_work)->work) &p->pi_lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)bond132 irq_context: 0 (wq_completion)bond132 (work_completion)(&(&slave->notify_work)->work) irq_context: 0 (wq_completion)bond132 (work_completion)(&(&slave->notify_work)->work) &obj_hash[i].lock irq_context: 0 (wq_completion)bond132 (work_completion)(&(&slave->notify_work)->work) &base->lock irq_context: 0 (wq_completion)bond132 (work_completion)(&(&slave->notify_work)->work) &base->lock &obj_hash[i].lock irq_context: 0 (wq_completion)bond124 (work_completion)(&(&bond->mcast_work)->work) &rq->__lock irq_context: 0 (wq_completion)bond124 (work_completion)(&(&bond->mcast_work)->work) &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)bond132 (work_completion)(&(&bond->mcast_work)->work) irq_context: 0 (wq_completion)bond132 (work_completion)(&(&bond->mcast_work)->work) &obj_hash[i].lock irq_context: 0 (wq_completion)bond132 (work_completion)(&(&bond->mcast_work)->work) &base->lock irq_context: 0 (wq_completion)bond132 (work_completion)(&(&bond->mcast_work)->work) &base->lock &obj_hash[i].lock irq_context: 0 (wq_completion)bond15#3 (work_completion)(&(&slave->notify_work)->work) &rq->__lock irq_context: 0 (wq_completion)bond15#3 (work_completion)(&(&slave->notify_work)->work) &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)bond103#3 irq_context: 0 (wq_completion)bond103#3 (work_completion)(&(&slave->notify_work)->work) irq_context: 0 (wq_completion)bond103#3 (work_completion)(&(&slave->notify_work)->work) &obj_hash[i].lock irq_context: 0 (wq_completion)bond103#3 (work_completion)(&(&slave->notify_work)->work) &base->lock irq_context: 0 (wq_completion)bond103#3 (work_completion)(&(&slave->notify_work)->work) &base->lock &obj_hash[i].lock irq_context: 0 (wq_completion)bond103#3 (work_completion)(&(&slave->notify_work)->work) &rq->__lock irq_context: 0 (wq_completion)bond103#3 (work_completion)(&(&slave->notify_work)->work) &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)bond103#3 (work_completion)(&(&bond->mcast_work)->work) irq_context: 0 (wq_completion)bond103#3 (work_completion)(&(&bond->mcast_work)->work) &obj_hash[i].lock irq_context: 0 (wq_completion)bond103#3 (work_completion)(&(&bond->mcast_work)->work) &base->lock irq_context: 0 (wq_completion)bond103#3 (work_completion)(&(&bond->mcast_work)->work) &base->lock &obj_hash[i].lock irq_context: 0 (wq_completion)bond103#3 (work_completion)(&(&bond->mcast_work)->work) &rq->__lock irq_context: 0 (wq_completion)bond103#3 (work_completion)(&(&bond->mcast_work)->work) &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)bond110#2 (work_completion)(&(&slave->notify_work)->work) &rq->__lock irq_context: 0 (wq_completion)bond110#2 (work_completion)(&(&slave->notify_work)->work) &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)bond120#2 irq_context: 0 (wq_completion)bond120#2 (work_completion)(&(&slave->notify_work)->work) irq_context: 0 (wq_completion)bond120#2 (work_completion)(&(&slave->notify_work)->work) &obj_hash[i].lock irq_context: 0 (wq_completion)bond120#2 (work_completion)(&(&slave->notify_work)->work) &base->lock irq_context: 0 (wq_completion)bond120#2 (work_completion)(&(&slave->notify_work)->work) &base->lock &obj_hash[i].lock irq_context: 0 (wq_completion)bond120#2 (work_completion)(&(&slave->notify_work)->work) &rq->__lock irq_context: 0 (wq_completion)bond128 &rq->__lock irq_context: 0 (wq_completion)bond128 &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)bond120#2 (work_completion)(&(&bond->mcast_work)->work) irq_context: 0 (wq_completion)bond120#2 (work_completion)(&(&bond->mcast_work)->work) &rq->__lock irq_context: 0 (wq_completion)bond120#2 (work_completion)(&(&bond->mcast_work)->work) &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)bond120#2 (work_completion)(&(&bond->mcast_work)->work) &obj_hash[i].lock irq_context: 0 (wq_completion)bond120#2 (work_completion)(&(&bond->mcast_work)->work) &base->lock irq_context: 0 (wq_completion)bond120#2 (work_completion)(&(&bond->mcast_work)->work) &base->lock &obj_hash[i].lock irq_context: 0 (wq_completion)bond87 (work_completion)(&(&bond->mcast_work)->work) &base->lock fill_pool_map-wait-type-override pool_lock irq_context: 0 (wq_completion)bond25#3 irq_context: 0 (wq_completion)bond25#3 (work_completion)(&(&slave->notify_work)->work) irq_context: 0 (wq_completion)bond25#3 (work_completion)(&(&slave->notify_work)->work) &obj_hash[i].lock irq_context: 0 (wq_completion)bond25#3 (work_completion)(&(&slave->notify_work)->work) &base->lock irq_context: 0 (wq_completion)bond25#3 (work_completion)(&(&slave->notify_work)->work) &base->lock &obj_hash[i].lock irq_context: 0 (wq_completion)bond100#3 (work_completion)(&(&bond->mcast_work)->work) &rq->__lock irq_context: 0 (wq_completion)bond100#3 (work_completion)(&(&bond->mcast_work)->work) &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)bond25#3 (work_completion)(&(&bond->mcast_work)->work) irq_context: 0 (wq_completion)bond25#3 (work_completion)(&(&bond->mcast_work)->work) &obj_hash[i].lock irq_context: 0 (wq_completion)bond25#3 (work_completion)(&(&bond->mcast_work)->work) &base->lock irq_context: 0 (wq_completion)bond25#3 (work_completion)(&(&bond->mcast_work)->work) &base->lock &obj_hash[i].lock irq_context: 0 (wq_completion)bond133 irq_context: 0 (wq_completion)bond133 &rq->__lock irq_context: 0 (wq_completion)bond133 &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)bond133 (work_completion)(&(&slave->notify_work)->work) irq_context: 0 (wq_completion)bond133 (work_completion)(&(&slave->notify_work)->work) &obj_hash[i].lock irq_context: 0 (wq_completion)bond133 (work_completion)(&(&slave->notify_work)->work) &base->lock irq_context: 0 (wq_completion)bond133 (work_completion)(&(&slave->notify_work)->work) &base->lock &obj_hash[i].lock irq_context: 0 (wq_completion)bond133 (work_completion)(&(&bond->mcast_work)->work) irq_context: 0 (wq_completion)bond133 (work_completion)(&(&bond->mcast_work)->work) &obj_hash[i].lock irq_context: 0 (wq_completion)bond133 (work_completion)(&(&bond->mcast_work)->work) &base->lock irq_context: 0 (wq_completion)bond133 (work_completion)(&(&bond->mcast_work)->work) &base->lock &obj_hash[i].lock irq_context: 0 (wq_completion)bond104#3 irq_context: 0 (wq_completion)bond104#3 (work_completion)(&(&slave->notify_work)->work) irq_context: 0 (wq_completion)bond104#3 (work_completion)(&(&slave->notify_work)->work) &obj_hash[i].lock irq_context: 0 (wq_completion)bond104#3 (work_completion)(&(&slave->notify_work)->work) &base->lock irq_context: 0 (wq_completion)bond104#3 (work_completion)(&(&slave->notify_work)->work) &base->lock &obj_hash[i].lock irq_context: 0 (wq_completion)bond104#3 (work_completion)(&(&slave->notify_work)->work) &rq->__lock irq_context: 0 (wq_completion)bond104#3 (work_completion)(&(&slave->notify_work)->work) &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)bond104#3 (work_completion)(&(&bond->mcast_work)->work) irq_context: 0 (wq_completion)bond104#3 (work_completion)(&(&bond->mcast_work)->work) &obj_hash[i].lock irq_context: 0 (wq_completion)bond104#3 (work_completion)(&(&bond->mcast_work)->work) &base->lock irq_context: 0 (wq_completion)bond104#3 (work_completion)(&(&bond->mcast_work)->work) &base->lock &obj_hash[i].lock irq_context: 0 (wq_completion)bond104#3 (work_completion)(&(&bond->mcast_work)->work) &rq->__lock irq_context: 0 (wq_completion)bond104#3 (work_completion)(&(&bond->mcast_work)->work) &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)bond98#3 (work_completion)(&(&bond->mcast_work)->work) fill_pool_map-wait-type-override pool_lock irq_context: 0 (wq_completion)bond121#2 irq_context: 0 (wq_completion)bond121#2 (work_completion)(&(&slave->notify_work)->work) irq_context: 0 (wq_completion)bond121#2 (work_completion)(&(&slave->notify_work)->work) &obj_hash[i].lock irq_context: 0 (wq_completion)bond121#2 (work_completion)(&(&slave->notify_work)->work) &base->lock irq_context: 0 (wq_completion)bond121#2 (work_completion)(&(&slave->notify_work)->work) &base->lock &obj_hash[i].lock irq_context: 0 (wq_completion)bond121#2 (work_completion)(&(&bond->mcast_work)->work) irq_context: 0 (wq_completion)bond121#2 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex fs_reclaim irq_context: 0 (wq_completion)bond121#2 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex fs_reclaim mmu_notifier_invalidate_range_start irq_context: 0 (wq_completion)bond121#2 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &c->lock irq_context: 0 (wq_completion)bond121#2 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &____s->seqcount#2 irq_context: 0 (wq_completion)bond121#2 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &____s->seqcount irq_context: 0 (wq_completion)bond121#2 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex rcu_read_lock &bond->stats_lock/1 irq_context: 0 (wq_completion)bond121#2 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex rcu_read_lock &ndev->lock irq_context: 0 (wq_completion)bond121#2 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex nl_table_lock irq_context: 0 (wq_completion)bond121#2 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &obj_hash[i].lock irq_context: 0 (wq_completion)bond121#2 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex nl_table_wait.lock irq_context: 0 (wq_completion)bond121#2 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex net_rwsem irq_context: 0 (wq_completion)bond121#2 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex net_rwsem &list->lock#2 irq_context: 0 (wq_completion)bond121#2 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &tn->lock irq_context: 0 (wq_completion)bond121#2 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &idev->mc_lock irq_context: 0 (wq_completion)bond121#2 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex.wait_lock irq_context: 0 (wq_completion)bond121#2 (work_completion)(&(&bond->mcast_work)->work) &p->pi_lock irq_context: 0 (wq_completion)bond124 (work_completion)(&(&slave->notify_work)->work) rtnl_mutex net_rwsem irq_context: 0 (wq_completion)bond124 (work_completion)(&(&slave->notify_work)->work) rtnl_mutex net_rwsem &list->lock#2 irq_context: 0 (wq_completion)bond124 (work_completion)(&(&slave->notify_work)->work) rtnl_mutex &tn->lock irq_context: 0 (wq_completion)bond124 (work_completion)(&(&slave->notify_work)->work) rtnl_mutex.wait_lock irq_context: 0 (wq_completion)bond124 (work_completion)(&(&slave->notify_work)->work) &p->pi_lock irq_context: 0 (wq_completion)bond103#3 (work_completion)(&(&slave->notify_work)->work) rtnl_mutex net_rwsem irq_context: 0 (wq_completion)bond103#3 (work_completion)(&(&slave->notify_work)->work) rtnl_mutex net_rwsem &list->lock#2 irq_context: 0 (wq_completion)bond103#3 (work_completion)(&(&slave->notify_work)->work) rtnl_mutex &tn->lock irq_context: 0 (wq_completion)bond103#3 (work_completion)(&(&slave->notify_work)->work) rtnl_mutex.wait_lock irq_context: 0 (wq_completion)bond103#3 (work_completion)(&(&slave->notify_work)->work) &p->pi_lock irq_context: 0 (wq_completion)bond130 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex fs_reclaim irq_context: 0 (wq_completion)bond130 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex fs_reclaim mmu_notifier_invalidate_range_start irq_context: 0 (wq_completion)bond130 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &c->lock irq_context: 0 (wq_completion)bond130 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex rcu_read_lock &bond->stats_lock/1 irq_context: 0 (wq_completion)bond130 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex rcu_read_lock &ndev->lock irq_context: 0 (wq_completion)bond130 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex nl_table_lock irq_context: 0 (wq_completion)bond130 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &obj_hash[i].lock irq_context: 0 (wq_completion)bond130 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex quarantine_lock irq_context: 0 (wq_completion)bond130 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex nl_table_wait.lock irq_context: 0 (wq_completion)bond130 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex net_rwsem irq_context: 0 (wq_completion)bond130 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex net_rwsem &list->lock#2 irq_context: 0 (wq_completion)bond130 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &tn->lock irq_context: 0 (wq_completion)bond130 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &idev->mc_lock irq_context: 0 (wq_completion)bond130 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &idev->mc_lock fs_reclaim irq_context: 0 (wq_completion)bond130 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &idev->mc_lock fs_reclaim mmu_notifier_invalidate_range_start irq_context: 0 (wq_completion)bond130 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &idev->mc_lock &c->lock irq_context: 0 (wq_completion)bond130 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &idev->mc_lock rcu_read_lock &ndev->lock irq_context: 0 (wq_completion)bond130 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &idev->mc_lock rcu_read_lock &c->lock irq_context: 0 (wq_completion)bond130 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &idev->mc_lock rcu_read_lock &n->list_lock irq_context: 0 (wq_completion)bond130 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &idev->mc_lock rcu_read_lock &n->list_lock &c->lock irq_context: 0 (wq_completion)bond130 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &idev->mc_lock rcu_read_lock &dir->lock#2 irq_context: 0 (wq_completion)bond130 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &idev->mc_lock rcu_read_lock &ul->lock#2 irq_context: 0 (wq_completion)bond130 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &idev->mc_lock rcu_read_lock rcu_read_lock &____s->seqcount#10 irq_context: 0 (wq_completion)bond130 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &idev->mc_lock rcu_read_lock rcu_read_lock rcu_read_lock_bh rcu_read_lock rcu_read_lock_bh rcu_read_lock &br->multicast_lock irq_context: 0 (wq_completion)bond130 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &idev->mc_lock rcu_read_lock rcu_read_lock rcu_read_lock_bh rcu_read_lock rcu_read_lock_bh rcu_read_lock &br->multicast_lock &obj_hash[i].lock irq_context: 0 (wq_completion)bond130 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &idev->mc_lock rcu_read_lock rcu_read_lock rcu_read_lock_bh rcu_read_lock rcu_read_lock_bh rcu_read_lock &br->multicast_lock &base->lock irq_context: 0 (wq_completion)bond130 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &idev->mc_lock rcu_read_lock rcu_read_lock rcu_read_lock_bh rcu_read_lock rcu_read_lock_bh rcu_read_lock &br->multicast_lock &base->lock &obj_hash[i].lock irq_context: 0 (wq_completion)bond130 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &idev->mc_lock rcu_read_lock rcu_read_lock rcu_read_lock_bh rcu_read_lock rcu_read_lock_bh rcu_read_lock &data->lock irq_context: 0 (wq_completion)bond130 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &idev->mc_lock rcu_read_lock rcu_read_lock rcu_read_lock_bh rcu_read_lock rcu_read_lock_bh rcu_read_lock &obj_hash[i].lock irq_context: 0 (wq_completion)bond130 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex.wait_lock irq_context: 0 (wq_completion)bond130 (work_completion)(&(&bond->mcast_work)->work) &p->pi_lock irq_context: 0 (wq_completion)bond100#3 (work_completion)(&(&slave->notify_work)->work) rtnl_mutex net_rwsem irq_context: 0 (wq_completion)bond100#3 (work_completion)(&(&slave->notify_work)->work) rtnl_mutex net_rwsem &list->lock#2 irq_context: 0 (wq_completion)bond100#3 (work_completion)(&(&slave->notify_work)->work) rtnl_mutex &tn->lock irq_context: 0 (wq_completion)bond100#3 (work_completion)(&(&slave->notify_work)->work) rtnl_mutex.wait_lock irq_context: 0 (wq_completion)bond100#3 (work_completion)(&(&slave->notify_work)->work) &p->pi_lock irq_context: 0 (wq_completion)bond133 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex fs_reclaim irq_context: 0 (wq_completion)bond133 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex fs_reclaim mmu_notifier_invalidate_range_start irq_context: 0 (wq_completion)bond133 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex rcu_read_lock &bond->stats_lock/1 irq_context: 0 (wq_completion)bond133 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex rcu_read_lock &ndev->lock irq_context: 0 (wq_completion)bond133 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex nl_table_lock irq_context: 0 (wq_completion)bond133 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &obj_hash[i].lock irq_context: 0 (wq_completion)bond133 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex nl_table_wait.lock irq_context: 0 (wq_completion)bond133 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex net_rwsem irq_context: 0 (wq_completion)bond133 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex net_rwsem &list->lock#2 irq_context: 0 (wq_completion)bond133 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &tn->lock irq_context: 0 (wq_completion)bond133 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &idev->mc_lock irq_context: 0 (wq_completion)bond133 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &rq->__lock irq_context: 0 (wq_completion)bond133 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex.wait_lock irq_context: 0 (wq_completion)bond133 (work_completion)(&(&bond->mcast_work)->work) &p->pi_lock irq_context: 0 (wq_completion)bond133 (work_completion)(&(&bond->mcast_work)->work) &p->pi_lock &rq->__lock irq_context: 0 (wq_completion)bond133 (work_completion)(&(&bond->mcast_work)->work) &p->pi_lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)bond133 (work_completion)(&(&bond->mcast_work)->work) &rq->__lock irq_context: 0 (wq_completion)bond133 (work_completion)(&(&bond->mcast_work)->work) &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)bond134 irq_context: 0 (wq_completion)bond134 (work_completion)(&(&slave->notify_work)->work) irq_context: 0 (wq_completion)bond134 (work_completion)(&(&slave->notify_work)->work) &obj_hash[i].lock irq_context: 0 (wq_completion)bond134 (work_completion)(&(&slave->notify_work)->work) &base->lock irq_context: 0 (wq_completion)bond134 (work_completion)(&(&slave->notify_work)->work) &base->lock &obj_hash[i].lock irq_context: 0 (wq_completion)bond134 (work_completion)(&(&slave->notify_work)->work) &rq->__lock irq_context: 0 (wq_completion)bond134 (work_completion)(&(&slave->notify_work)->work) &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)bond92#2 (work_completion)(&(&slave->notify_work)->work) &rq->__lock irq_context: 0 (wq_completion)bond92#2 (work_completion)(&(&slave->notify_work)->work) &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)bond134 (work_completion)(&(&bond->mcast_work)->work) irq_context: 0 (wq_completion)bond134 (work_completion)(&(&bond->mcast_work)->work) &rq->__lock irq_context: 0 (wq_completion)bond134 (work_completion)(&(&bond->mcast_work)->work) &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)bond134 (work_completion)(&(&bond->mcast_work)->work) &obj_hash[i].lock irq_context: 0 (wq_completion)bond134 (work_completion)(&(&bond->mcast_work)->work) &base->lock irq_context: 0 (wq_completion)bond134 (work_completion)(&(&bond->mcast_work)->work) &base->lock &obj_hash[i].lock irq_context: 0 (wq_completion)bond128 (work_completion)(&(&bond->mcast_work)->work) &rq->__lock irq_context: 0 (wq_completion)bond128 (work_completion)(&(&bond->mcast_work)->work) &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)bond25#3 (work_completion)(&(&bond->mcast_work)->work) &rq->__lock irq_context: 0 (wq_completion)bond25#3 (work_completion)(&(&bond->mcast_work)->work) &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)bond98#3 &rq->__lock irq_context: 0 (wq_completion)bond98#3 &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)bond105#3 irq_context: 0 (wq_completion)bond105#3 (work_completion)(&(&slave->notify_work)->work) irq_context: 0 (wq_completion)bond105#3 (work_completion)(&(&slave->notify_work)->work) &obj_hash[i].lock irq_context: 0 (wq_completion)bond105#3 (work_completion)(&(&slave->notify_work)->work) &base->lock irq_context: 0 (wq_completion)bond105#3 (work_completion)(&(&slave->notify_work)->work) &base->lock &obj_hash[i].lock irq_context: 0 (wq_completion)bond105#3 (work_completion)(&(&slave->notify_work)->work) &rq->__lock irq_context: 0 (wq_completion)bond105#3 (work_completion)(&(&slave->notify_work)->work) &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)bond131 (work_completion)(&(&bond->mcast_work)->work) &rq->__lock irq_context: 0 (wq_completion)bond131 (work_completion)(&(&bond->mcast_work)->work) &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)bond105#3 (work_completion)(&(&bond->mcast_work)->work) irq_context: 0 (wq_completion)bond105#3 (work_completion)(&(&bond->mcast_work)->work) &obj_hash[i].lock irq_context: 0 (wq_completion)bond105#3 (work_completion)(&(&bond->mcast_work)->work) &base->lock irq_context: 0 (wq_completion)bond105#3 (work_completion)(&(&bond->mcast_work)->work) &base->lock &obj_hash[i].lock irq_context: 0 (wq_completion)bond105#3 (work_completion)(&(&bond->mcast_work)->work) &rq->__lock irq_context: 0 (wq_completion)bond105#3 (work_completion)(&(&bond->mcast_work)->work) &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)bond133 (work_completion)(&(&slave->notify_work)->work) rtnl_mutex net_rwsem irq_context: 0 (wq_completion)bond133 (work_completion)(&(&slave->notify_work)->work) rtnl_mutex net_rwsem &list->lock#2 irq_context: 0 (wq_completion)bond133 (work_completion)(&(&slave->notify_work)->work) rtnl_mutex &tn->lock irq_context: 0 (wq_completion)bond133 (work_completion)(&(&slave->notify_work)->work) rtnl_mutex.wait_lock irq_context: 0 (wq_completion)bond133 (work_completion)(&(&slave->notify_work)->work) &p->pi_lock irq_context: 0 (wq_completion)bond121#2 (work_completion)(&(&slave->notify_work)->work) rtnl_mutex net_rwsem irq_context: 0 (wq_completion)bond121#2 (work_completion)(&(&slave->notify_work)->work) rtnl_mutex net_rwsem &list->lock#2 irq_context: 0 (wq_completion)bond121#2 (work_completion)(&(&slave->notify_work)->work) rtnl_mutex &tn->lock irq_context: 0 (wq_completion)bond121#2 (work_completion)(&(&slave->notify_work)->work) rtnl_mutex.wait_lock irq_context: 0 (wq_completion)bond121#2 (work_completion)(&(&slave->notify_work)->work) &p->pi_lock irq_context: 0 (wq_completion)bond103#3 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex fs_reclaim irq_context: 0 (wq_completion)bond103#3 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex fs_reclaim mmu_notifier_invalidate_range_start irq_context: 0 (wq_completion)bond103#3 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex rcu_read_lock &bond->stats_lock/1 irq_context: 0 (wq_completion)bond103#3 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex rcu_read_lock &ndev->lock irq_context: 0 (wq_completion)bond103#3 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex nl_table_lock irq_context: 0 (wq_completion)bond103#3 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &obj_hash[i].lock irq_context: 0 (wq_completion)bond103#3 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex nl_table_wait.lock irq_context: 0 (wq_completion)bond103#3 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex net_rwsem irq_context: 0 (wq_completion)bond103#3 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex net_rwsem &list->lock#2 irq_context: 0 (wq_completion)bond103#3 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &tn->lock irq_context: 0 (wq_completion)bond103#3 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &idev->mc_lock irq_context: 0 (wq_completion)bond103#3 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &idev->mc_lock fs_reclaim irq_context: 0 (wq_completion)bond103#3 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &idev->mc_lock fs_reclaim mmu_notifier_invalidate_range_start irq_context: 0 (wq_completion)bond103#3 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &idev->mc_lock rcu_read_lock &ndev->lock irq_context: 0 (wq_completion)bond103#3 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &idev->mc_lock rcu_read_lock &dir->lock#2 irq_context: 0 (wq_completion)bond103#3 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &idev->mc_lock rcu_read_lock &ul->lock#2 irq_context: 0 (wq_completion)bond103#3 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &idev->mc_lock rcu_read_lock rcu_read_lock &____s->seqcount#10 irq_context: 0 (wq_completion)bond103#3 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &idev->mc_lock rcu_read_lock rcu_read_lock rcu_read_lock_bh rcu_read_lock rcu_read_lock_bh rcu_read_lock &br->multicast_lock irq_context: 0 (wq_completion)bond103#3 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &idev->mc_lock rcu_read_lock rcu_read_lock rcu_read_lock_bh rcu_read_lock rcu_read_lock_bh rcu_read_lock &br->multicast_lock &obj_hash[i].lock irq_context: 0 (wq_completion)bond103#3 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &idev->mc_lock rcu_read_lock rcu_read_lock rcu_read_lock_bh rcu_read_lock rcu_read_lock_bh rcu_read_lock &br->multicast_lock &base->lock irq_context: 0 (wq_completion)bond103#3 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &idev->mc_lock rcu_read_lock rcu_read_lock rcu_read_lock_bh rcu_read_lock rcu_read_lock_bh rcu_read_lock &br->multicast_lock &base->lock &obj_hash[i].lock irq_context: 0 (wq_completion)bond103#3 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &idev->mc_lock rcu_read_lock rcu_read_lock rcu_read_lock_bh rcu_read_lock rcu_read_lock_bh rcu_read_lock &data->lock irq_context: 0 (wq_completion)bond103#3 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &idev->mc_lock rcu_read_lock rcu_read_lock rcu_read_lock_bh rcu_read_lock rcu_read_lock_bh rcu_read_lock &obj_hash[i].lock irq_context: 0 (wq_completion)bond103#3 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex.wait_lock irq_context: 0 (wq_completion)bond103#3 (work_completion)(&(&bond->mcast_work)->work) &p->pi_lock irq_context: 0 (wq_completion)bond92#2 (work_completion)(&(&slave->notify_work)->work) rtnl_mutex net_rwsem irq_context: 0 (wq_completion)bond92#2 (work_completion)(&(&slave->notify_work)->work) rtnl_mutex net_rwsem &list->lock#2 irq_context: 0 (wq_completion)bond92#2 (work_completion)(&(&slave->notify_work)->work) rtnl_mutex &tn->lock irq_context: 0 (wq_completion)bond92#2 (work_completion)(&(&slave->notify_work)->work) rtnl_mutex.wait_lock irq_context: 0 (wq_completion)bond92#2 (work_completion)(&(&slave->notify_work)->work) &p->pi_lock irq_context: 0 (wq_completion)bond79#2 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex fs_reclaim irq_context: 0 (wq_completion)bond79#2 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex fs_reclaim mmu_notifier_invalidate_range_start irq_context: 0 (wq_completion)bond79#2 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &c->lock irq_context: 0 (wq_completion)bond79#2 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &n->list_lock irq_context: 0 (wq_completion)bond79#2 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &n->list_lock &c->lock irq_context: 0 (wq_completion)bond79#2 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex rcu_read_lock &bond->stats_lock/1 irq_context: 0 (wq_completion)bond79#2 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex rcu_read_lock &ndev->lock irq_context: 0 (wq_completion)bond79#2 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex nl_table_lock irq_context: 0 (wq_completion)bond79#2 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &obj_hash[i].lock irq_context: 0 (wq_completion)bond79#2 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex nl_table_wait.lock irq_context: 0 (wq_completion)bond79#2 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex net_rwsem irq_context: 0 (wq_completion)bond79#2 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex net_rwsem &list->lock#2 irq_context: 0 (wq_completion)bond79#2 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &tn->lock irq_context: 0 (wq_completion)bond79#2 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &idev->mc_lock irq_context: 0 (wq_completion)bond79#2 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &idev->mc_lock fs_reclaim irq_context: 0 (wq_completion)bond79#2 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &idev->mc_lock fs_reclaim mmu_notifier_invalidate_range_start irq_context: 0 (wq_completion)bond79#2 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &idev->mc_lock rcu_read_lock &ndev->lock irq_context: 0 (wq_completion)bond79#2 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &idev->mc_lock rcu_read_lock &dir->lock#2 irq_context: 0 (wq_completion)bond79#2 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &idev->mc_lock rcu_read_lock &ul->lock#2 irq_context: 0 (wq_completion)bond79#2 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &idev->mc_lock rcu_read_lock rcu_read_lock &____s->seqcount#10 irq_context: 0 (wq_completion)bond79#2 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &idev->mc_lock rcu_read_lock rcu_read_lock rcu_read_lock_bh rcu_read_lock rcu_read_lock_bh rcu_read_lock &br->multicast_lock irq_context: 0 (wq_completion)bond79#2 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &idev->mc_lock rcu_read_lock rcu_read_lock rcu_read_lock_bh rcu_read_lock rcu_read_lock_bh rcu_read_lock &br->multicast_lock &obj_hash[i].lock irq_context: 0 (wq_completion)bond79#2 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &idev->mc_lock rcu_read_lock rcu_read_lock rcu_read_lock_bh rcu_read_lock rcu_read_lock_bh rcu_read_lock &br->multicast_lock &base->lock irq_context: 0 (wq_completion)bond79#2 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &idev->mc_lock rcu_read_lock rcu_read_lock rcu_read_lock_bh rcu_read_lock rcu_read_lock_bh rcu_read_lock &br->multicast_lock &base->lock &obj_hash[i].lock irq_context: 0 (wq_completion)bond79#2 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &idev->mc_lock rcu_read_lock rcu_read_lock rcu_read_lock_bh rcu_read_lock rcu_read_lock_bh rcu_read_lock &data->lock irq_context: 0 (wq_completion)bond79#2 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &idev->mc_lock rcu_read_lock rcu_read_lock rcu_read_lock_bh rcu_read_lock rcu_read_lock_bh rcu_read_lock &obj_hash[i].lock irq_context: 0 (wq_completion)bond79#2 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &idev->mc_lock rcu_read_lock rcu_read_lock rcu_node_0 irq_context: 0 (wq_completion)bond79#2 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &idev->mc_lock rcu_read_lock rcu_read_lock &rq->__lock irq_context: 0 (wq_completion)bond79#2 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &idev->mc_lock rcu_read_lock rcu_read_lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)bond79#2 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &idev->mc_lock rcu_read_lock rcu_node_0 irq_context: 0 (wq_completion)bond79#2 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &idev->mc_lock rcu_read_lock &rcu_state.gp_wq irq_context: 0 (wq_completion)bond79#2 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &idev->mc_lock rcu_read_lock &rcu_state.gp_wq &p->pi_lock irq_context: 0 (wq_completion)bond79#2 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &idev->mc_lock rcu_read_lock &rcu_state.gp_wq &p->pi_lock &rq->__lock irq_context: 0 (wq_completion)bond79#2 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &idev->mc_lock rcu_read_lock &rcu_state.gp_wq &p->pi_lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)bond79#2 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &idev->mc_lock rcu_read_lock &rq->__lock irq_context: 0 (wq_completion)bond79#2 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &idev->mc_lock rcu_read_lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)bond79#2 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex.wait_lock irq_context: 0 (wq_completion)bond79#2 (work_completion)(&(&bond->mcast_work)->work) &p->pi_lock irq_context: 0 (wq_completion)bond79#2 (work_completion)(&(&bond->mcast_work)->work) &p->pi_lock &rq->__lock irq_context: 0 (wq_completion)bond79#2 (work_completion)(&(&bond->mcast_work)->work) &p->pi_lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)bond26#4 irq_context: 0 (wq_completion)bond26#4 (work_completion)(&(&slave->notify_work)->work) irq_context: 0 (wq_completion)bond26#4 (work_completion)(&(&slave->notify_work)->work) &obj_hash[i].lock irq_context: 0 (wq_completion)bond26#4 (work_completion)(&(&slave->notify_work)->work) &base->lock irq_context: 0 (wq_completion)bond26#4 (work_completion)(&(&slave->notify_work)->work) &base->lock &obj_hash[i].lock irq_context: 0 (wq_completion)bond26#4 (work_completion)(&(&slave->notify_work)->work) &rq->__lock irq_context: 0 (wq_completion)bond26#4 (work_completion)(&(&slave->notify_work)->work) &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)bond122 &rq->__lock irq_context: 0 (wq_completion)bond122 &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)bond26#4 (work_completion)(&(&bond->mcast_work)->work) irq_context: 0 (wq_completion)bond26#4 (work_completion)(&(&bond->mcast_work)->work) &obj_hash[i].lock irq_context: 0 (wq_completion)bond26#4 (work_completion)(&(&bond->mcast_work)->work) &base->lock irq_context: 0 (wq_completion)bond26#4 (work_completion)(&(&bond->mcast_work)->work) &base->lock &obj_hash[i].lock irq_context: 0 (wq_completion)bond26#4 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex fs_reclaim irq_context: 0 (wq_completion)bond26#4 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex fs_reclaim mmu_notifier_invalidate_range_start irq_context: 0 (wq_completion)bond26#4 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex fs_reclaim mmu_notifier_invalidate_range_start &rq->__lock irq_context: 0 (wq_completion)bond26#4 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex fs_reclaim mmu_notifier_invalidate_range_start &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)bond26#4 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &c->lock irq_context: 0 (wq_completion)bond26#4 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex rcu_read_lock &bond->stats_lock/1 irq_context: 0 (wq_completion)bond26#4 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex rcu_read_lock &ndev->lock irq_context: 0 (wq_completion)bond26#4 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex nl_table_lock irq_context: 0 (wq_completion)bond26#4 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &obj_hash[i].lock irq_context: 0 (wq_completion)bond26#4 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex nl_table_wait.lock irq_context: 0 (wq_completion)bond26#4 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex net_rwsem irq_context: 0 (wq_completion)bond26#4 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex net_rwsem &list->lock#2 irq_context: 0 (wq_completion)bond26#4 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &tn->lock irq_context: 0 (wq_completion)bond26#4 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &idev->mc_lock irq_context: 0 (wq_completion)bond26#4 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &rq->__lock irq_context: 0 (wq_completion)bond26#4 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)bond26#4 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex.wait_lock irq_context: 0 (wq_completion)bond26#4 (work_completion)(&(&bond->mcast_work)->work) &p->pi_lock irq_context: 0 (wq_completion)bond122#2 irq_context: 0 (wq_completion)bond122#2 (work_completion)(&(&slave->notify_work)->work) irq_context: 0 (wq_completion)bond122#2 (work_completion)(&(&slave->notify_work)->work) &obj_hash[i].lock irq_context: 0 (wq_completion)bond122#2 (work_completion)(&(&slave->notify_work)->work) &base->lock irq_context: 0 (wq_completion)bond122#2 (work_completion)(&(&slave->notify_work)->work) &base->lock &obj_hash[i].lock irq_context: 0 (wq_completion)bond122#2 (work_completion)(&(&bond->mcast_work)->work) irq_context: 0 (wq_completion)bond122#2 (work_completion)(&(&bond->mcast_work)->work) &rq->__lock irq_context: 0 (wq_completion)bond122#2 (work_completion)(&(&bond->mcast_work)->work) &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)bond122#2 (work_completion)(&(&bond->mcast_work)->work) &obj_hash[i].lock irq_context: 0 (wq_completion)bond122#2 (work_completion)(&(&bond->mcast_work)->work) &base->lock irq_context: 0 (wq_completion)bond122#2 (work_completion)(&(&bond->mcast_work)->work) &base->lock &obj_hash[i].lock irq_context: 0 (wq_completion)bond129 (work_completion)(&(&bond->mcast_work)->work) &rq->__lock irq_context: 0 (wq_completion)bond129 (work_completion)(&(&bond->mcast_work)->work) &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)bond135 irq_context: 0 (wq_completion)bond135 (work_completion)(&(&slave->notify_work)->work) irq_context: 0 (wq_completion)bond135 (work_completion)(&(&slave->notify_work)->work) &obj_hash[i].lock irq_context: 0 (wq_completion)bond135 (work_completion)(&(&slave->notify_work)->work) &base->lock irq_context: 0 (wq_completion)bond135 (work_completion)(&(&slave->notify_work)->work) &base->lock &obj_hash[i].lock irq_context: 0 (wq_completion)bond135 (work_completion)(&(&slave->notify_work)->work) &rq->__lock irq_context: 0 (wq_completion)bond135 (work_completion)(&(&slave->notify_work)->work) &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)bond135 (work_completion)(&(&bond->mcast_work)->work) irq_context: 0 (wq_completion)bond135 (work_completion)(&(&bond->mcast_work)->work) &obj_hash[i].lock irq_context: 0 (wq_completion)bond135 (work_completion)(&(&bond->mcast_work)->work) &base->lock irq_context: 0 (wq_completion)bond135 (work_completion)(&(&bond->mcast_work)->work) &base->lock &obj_hash[i].lock irq_context: 0 (wq_completion)bond111#2 &rq->__lock irq_context: 0 (wq_completion)bond111#2 &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)bond111#2 (work_completion)(&(&slave->notify_work)->work) rtnl_mutex net_rwsem irq_context: 0 (wq_completion)bond111#2 (work_completion)(&(&slave->notify_work)->work) rtnl_mutex net_rwsem &list->lock#2 irq_context: 0 (wq_completion)bond111#2 (work_completion)(&(&slave->notify_work)->work) rtnl_mutex &tn->lock irq_context: 0 (wq_completion)bond111#2 (work_completion)(&(&slave->notify_work)->work) rtnl_mutex.wait_lock irq_context: 0 (wq_completion)bond125 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex fs_reclaim irq_context: 0 (wq_completion)bond111#2 (work_completion)(&(&slave->notify_work)->work) &p->pi_lock irq_context: 0 (wq_completion)bond125 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex fs_reclaim mmu_notifier_invalidate_range_start irq_context: 0 (wq_completion)bond125 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &c->lock irq_context: 0 (wq_completion)bond125 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex rcu_read_lock &bond->stats_lock/1 irq_context: 0 (wq_completion)bond125 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex rcu_read_lock &ndev->lock irq_context: 0 (wq_completion)bond125 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex nl_table_lock irq_context: 0 (wq_completion)bond125 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &obj_hash[i].lock irq_context: 0 (wq_completion)bond125 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex nl_table_wait.lock irq_context: 0 (wq_completion)bond125 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex net_rwsem irq_context: 0 (wq_completion)bond125 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex net_rwsem &list->lock#2 irq_context: 0 (wq_completion)bond125 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &tn->lock irq_context: 0 (wq_completion)bond125 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &idev->mc_lock irq_context: 0 (wq_completion)bond125 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &idev->mc_lock fs_reclaim irq_context: 0 (wq_completion)bond125 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &idev->mc_lock fs_reclaim mmu_notifier_invalidate_range_start irq_context: 0 (wq_completion)bond125 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &idev->mc_lock rcu_read_lock &ndev->lock irq_context: 0 (wq_completion)bond125 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &idev->mc_lock rcu_read_lock &dir->lock#2 irq_context: 0 (wq_completion)bond125 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &idev->mc_lock rcu_read_lock &ul->lock#2 irq_context: 0 (wq_completion)bond125 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &idev->mc_lock rcu_read_lock rcu_read_lock &____s->seqcount#10 irq_context: 0 (wq_completion)bond125 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &idev->mc_lock rcu_read_lock rcu_read_lock rcu_read_lock_bh rcu_read_lock rcu_read_lock_bh rcu_read_lock &br->multicast_lock irq_context: 0 (wq_completion)bond125 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &idev->mc_lock rcu_read_lock rcu_read_lock rcu_read_lock_bh rcu_read_lock rcu_read_lock_bh rcu_read_lock &br->multicast_lock &obj_hash[i].lock irq_context: 0 (wq_completion)bond125 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &idev->mc_lock rcu_read_lock rcu_read_lock rcu_read_lock_bh rcu_read_lock rcu_read_lock_bh rcu_read_lock &br->multicast_lock &base->lock irq_context: 0 (wq_completion)bond125 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &idev->mc_lock rcu_read_lock rcu_read_lock rcu_read_lock_bh rcu_read_lock rcu_read_lock_bh rcu_read_lock &data->lock irq_context: 0 (wq_completion)bond125 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &idev->mc_lock rcu_read_lock rcu_read_lock rcu_read_lock_bh rcu_read_lock rcu_read_lock_bh rcu_read_lock &obj_hash[i].lock irq_context: 0 (wq_completion)bond125 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex.wait_lock irq_context: 0 (wq_completion)bond125 (work_completion)(&(&bond->mcast_work)->work) &p->pi_lock irq_context: 0 (wq_completion)bond135 (work_completion)(&(&bond->mcast_work)->work) &rq->__lock irq_context: 0 (wq_completion)bond135 (work_completion)(&(&bond->mcast_work)->work) &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)bond106#3 irq_context: 0 (wq_completion)bond106#3 (work_completion)(&(&slave->notify_work)->work) irq_context: 0 (wq_completion)bond106#3 (work_completion)(&(&slave->notify_work)->work) &obj_hash[i].lock irq_context: 0 (wq_completion)bond106#3 (work_completion)(&(&slave->notify_work)->work) &base->lock irq_context: 0 (wq_completion)bond106#3 (work_completion)(&(&slave->notify_work)->work) &base->lock &obj_hash[i].lock irq_context: 0 (wq_completion)bond106#3 (work_completion)(&(&bond->mcast_work)->work) irq_context: 0 (wq_completion)bond116#2 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex fs_reclaim irq_context: 0 (wq_completion)bond106#3 (work_completion)(&(&bond->mcast_work)->work) &obj_hash[i].lock irq_context: 0 (wq_completion)bond116#2 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex fs_reclaim mmu_notifier_invalidate_range_start irq_context: 0 (wq_completion)bond106#3 (work_completion)(&(&bond->mcast_work)->work) &base->lock irq_context: 0 (wq_completion)bond116#2 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &c->lock irq_context: 0 (wq_completion)bond106#3 (work_completion)(&(&bond->mcast_work)->work) &base->lock &obj_hash[i].lock irq_context: 0 (wq_completion)bond116#2 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &____s->seqcount#2 irq_context: 0 (wq_completion)bond116#2 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &____s->seqcount irq_context: 0 (wq_completion)bond116#2 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex rcu_read_lock &bond->stats_lock/1 irq_context: 0 (wq_completion)bond116#2 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex rcu_read_lock &ndev->lock irq_context: 0 (wq_completion)bond116#2 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex nl_table_lock irq_context: 0 (wq_completion)bond116#2 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &obj_hash[i].lock irq_context: 0 (wq_completion)bond116#2 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex nl_table_wait.lock irq_context: 0 (wq_completion)bond116#2 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex net_rwsem irq_context: 0 (wq_completion)bond116#2 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex net_rwsem &list->lock#2 irq_context: 0 (wq_completion)bond116#2 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &tn->lock irq_context: 0 (wq_completion)bond116#2 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &idev->mc_lock irq_context: 0 (wq_completion)bond116#2 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &idev->mc_lock fs_reclaim irq_context: 0 (wq_completion)bond116#2 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &idev->mc_lock fs_reclaim mmu_notifier_invalidate_range_start irq_context: 0 (wq_completion)bond116#2 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &idev->mc_lock &rq->__lock irq_context: 0 (wq_completion)bond116#2 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &idev->mc_lock rcu_read_lock &ndev->lock irq_context: 0 (wq_completion)bond116#2 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &idev->mc_lock rcu_read_lock &dir->lock#2 irq_context: 0 (wq_completion)bond116#2 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &idev->mc_lock rcu_read_lock &ul->lock#2 irq_context: 0 (wq_completion)bond116#2 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &idev->mc_lock rcu_read_lock rcu_read_lock &____s->seqcount#10 irq_context: 0 (wq_completion)bond116#2 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &idev->mc_lock rcu_read_lock rcu_read_lock rcu_read_lock_bh rcu_read_lock rcu_read_lock_bh rcu_read_lock &br->multicast_lock irq_context: 0 (wq_completion)bond116#2 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &idev->mc_lock rcu_read_lock rcu_read_lock rcu_read_lock_bh rcu_read_lock rcu_read_lock_bh rcu_read_lock &br->multicast_lock &obj_hash[i].lock irq_context: 0 (wq_completion)bond116#2 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &idev->mc_lock rcu_read_lock rcu_read_lock rcu_read_lock_bh rcu_read_lock rcu_read_lock_bh rcu_read_lock &br->multicast_lock &base->lock irq_context: 0 (wq_completion)bond116#2 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &idev->mc_lock rcu_read_lock rcu_read_lock rcu_read_lock_bh rcu_read_lock rcu_read_lock_bh rcu_read_lock &data->lock irq_context: 0 (wq_completion)bond116#2 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &idev->mc_lock rcu_read_lock rcu_read_lock rcu_read_lock_bh rcu_read_lock rcu_read_lock_bh rcu_read_lock &data->lock &obj_hash[i].lock irq_context: 0 (wq_completion)bond116#2 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &idev->mc_lock rcu_read_lock rcu_read_lock rcu_read_lock_bh rcu_read_lock rcu_read_lock_bh rcu_read_lock &data->lock &base->lock irq_context: 0 (wq_completion)bond116#2 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &idev->mc_lock rcu_read_lock rcu_read_lock rcu_read_lock_bh rcu_read_lock rcu_read_lock_bh rcu_read_lock &data->lock &base->lock &obj_hash[i].lock irq_context: 0 (wq_completion)bond116#2 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &idev->mc_lock rcu_read_lock rcu_read_lock rcu_read_lock_bh rcu_read_lock rcu_read_lock_bh rcu_read_lock &obj_hash[i].lock irq_context: 0 (wq_completion)bond116#2 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &idev->mc_lock rcu_read_lock rcu_read_lock rcu_node_0 irq_context: 0 (wq_completion)bond116#2 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &idev->mc_lock rcu_read_lock rcu_read_lock &rq->__lock irq_context: 0 (wq_completion)bond116#2 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &idev->mc_lock rcu_read_lock rcu_read_lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)bond116#2 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &idev->mc_lock rcu_read_lock rcu_node_0 irq_context: 0 (wq_completion)bond116#2 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex.wait_lock irq_context: 0 (wq_completion)bond116#2 (work_completion)(&(&bond->mcast_work)->work) &p->pi_lock irq_context: 0 (wq_completion)bond116#2 (work_completion)(&(&bond->mcast_work)->work) &p->pi_lock &rq->__lock irq_context: 0 (wq_completion)bond116#2 (work_completion)(&(&bond->mcast_work)->work) &p->pi_lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)bond123#2 irq_context: 0 (wq_completion)bond123#2 (work_completion)(&(&slave->notify_work)->work) irq_context: 0 (wq_completion)bond123#2 (work_completion)(&(&slave->notify_work)->work) &obj_hash[i].lock irq_context: 0 (wq_completion)bond123#2 (work_completion)(&(&slave->notify_work)->work) &base->lock irq_context: 0 (wq_completion)bond123#2 (work_completion)(&(&slave->notify_work)->work) &base->lock &obj_hash[i].lock irq_context: 0 (wq_completion)bond123#2 (work_completion)(&(&bond->mcast_work)->work) irq_context: 0 (wq_completion)bond123#2 (work_completion)(&(&bond->mcast_work)->work) &rq->__lock irq_context: 0 (wq_completion)bond123#2 (work_completion)(&(&bond->mcast_work)->work) &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)bond123#2 (work_completion)(&(&bond->mcast_work)->work) &obj_hash[i].lock irq_context: 0 (wq_completion)bond123#2 (work_completion)(&(&bond->mcast_work)->work) &base->lock irq_context: 0 (wq_completion)bond123#2 (work_completion)(&(&bond->mcast_work)->work) &base->lock &obj_hash[i].lock irq_context: 0 (wq_completion)bond27#4 irq_context: 0 (wq_completion)bond27#4 (work_completion)(&(&slave->notify_work)->work) irq_context: 0 (wq_completion)bond27#4 (work_completion)(&(&slave->notify_work)->work) &obj_hash[i].lock irq_context: 0 (wq_completion)bond27#4 (work_completion)(&(&slave->notify_work)->work) &base->lock irq_context: 0 (wq_completion)bond27#4 (work_completion)(&(&slave->notify_work)->work) &base->lock &obj_hash[i].lock irq_context: 0 (wq_completion)bond27#4 (work_completion)(&(&slave->notify_work)->work) &rq->__lock irq_context: 0 (wq_completion)bond27#4 (work_completion)(&(&slave->notify_work)->work) &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)events (linkwatch_work).work rtnl_mutex pcpu_alloc_mutex &rq->__lock irq_context: 0 (wq_completion)events (linkwatch_work).work rtnl_mutex pcpu_alloc_mutex &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)bond27#4 (work_completion)(&(&bond->mcast_work)->work) irq_context: 0 (wq_completion)bond27#4 (work_completion)(&(&bond->mcast_work)->work) &obj_hash[i].lock irq_context: 0 (wq_completion)bond27#4 (work_completion)(&(&bond->mcast_work)->work) &base->lock irq_context: 0 (wq_completion)bond27#4 (work_completion)(&(&bond->mcast_work)->work) &base->lock &obj_hash[i].lock irq_context: 0 (wq_completion)events (linkwatch_work).work rtnl_mutex pcpu_alloc_mutex.wait_lock irq_context: 0 (wq_completion)bond27#4 (work_completion)(&(&bond->mcast_work)->work) &rq->__lock irq_context: 0 (wq_completion)bond136 irq_context: 0 (wq_completion)bond136 (work_completion)(&(&slave->notify_work)->work) irq_context: 0 (wq_completion)bond136 (work_completion)(&(&slave->notify_work)->work) &obj_hash[i].lock irq_context: 0 (wq_completion)bond136 (work_completion)(&(&slave->notify_work)->work) &base->lock irq_context: 0 (wq_completion)bond136 (work_completion)(&(&slave->notify_work)->work) &base->lock &obj_hash[i].lock irq_context: 0 (wq_completion)bond136 (work_completion)(&(&slave->notify_work)->work) &rq->__lock irq_context: 0 (wq_completion)bond136 (work_completion)(&(&slave->notify_work)->work) &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)bond136 (work_completion)(&(&bond->mcast_work)->work) irq_context: 0 (wq_completion)bond136 (work_completion)(&(&bond->mcast_work)->work) &obj_hash[i].lock irq_context: 0 (wq_completion)bond136 (work_completion)(&(&bond->mcast_work)->work) &base->lock irq_context: 0 (wq_completion)bond136 (work_completion)(&(&bond->mcast_work)->work) &base->lock &obj_hash[i].lock irq_context: 0 (wq_completion)bond109#2 (work_completion)(&(&bond->mcast_work)->work) fill_pool_map-wait-type-override pool_lock irq_context: 0 (wq_completion)bond107#3 irq_context: 0 (wq_completion)bond107#3 (work_completion)(&(&slave->notify_work)->work) irq_context: 0 (wq_completion)bond107#3 (work_completion)(&(&slave->notify_work)->work) &obj_hash[i].lock irq_context: 0 (wq_completion)bond107#3 (work_completion)(&(&slave->notify_work)->work) &base->lock irq_context: 0 (wq_completion)bond107#3 (work_completion)(&(&slave->notify_work)->work) &base->lock &obj_hash[i].lock irq_context: 0 (wq_completion)bond107#3 (work_completion)(&(&slave->notify_work)->work) &rq->__lock irq_context: 0 (wq_completion)bond107#3 (work_completion)(&(&slave->notify_work)->work) &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)bond107#3 (work_completion)(&(&bond->mcast_work)->work) irq_context: 0 (wq_completion)bond107#3 (work_completion)(&(&bond->mcast_work)->work) &obj_hash[i].lock irq_context: 0 (wq_completion)bond107#3 (work_completion)(&(&bond->mcast_work)->work) &base->lock irq_context: 0 (wq_completion)bond107#3 (work_completion)(&(&bond->mcast_work)->work) &base->lock &obj_hash[i].lock irq_context: 0 (wq_completion)bond107#3 (work_completion)(&(&bond->mcast_work)->work) &rq->__lock irq_context: 0 (wq_completion)bond107#3 (work_completion)(&(&bond->mcast_work)->work) &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)bond124#2 irq_context: 0 (wq_completion)bond124#2 (work_completion)(&(&slave->notify_work)->work) irq_context: 0 (wq_completion)bond124#2 (work_completion)(&(&slave->notify_work)->work) &obj_hash[i].lock irq_context: 0 (wq_completion)bond124#2 (work_completion)(&(&slave->notify_work)->work) &base->lock irq_context: 0 (wq_completion)bond124#2 (work_completion)(&(&slave->notify_work)->work) &base->lock &obj_hash[i].lock irq_context: 0 (wq_completion)bond124#2 (work_completion)(&(&slave->notify_work)->work) &rq->__lock irq_context: 0 (wq_completion)bond119#2 (work_completion)(&(&bond->mcast_work)->work) &rq->__lock irq_context: 0 (wq_completion)bond119#2 (work_completion)(&(&bond->mcast_work)->work) &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)bond124#2 (work_completion)(&(&bond->mcast_work)->work) irq_context: 0 (wq_completion)bond124#2 (work_completion)(&(&bond->mcast_work)->work) &obj_hash[i].lock irq_context: 0 (wq_completion)bond124#2 (work_completion)(&(&bond->mcast_work)->work) &base->lock irq_context: 0 (wq_completion)bond124#2 (work_completion)(&(&bond->mcast_work)->work) &base->lock &obj_hash[i].lock irq_context: 0 (wq_completion)bond124#2 (work_completion)(&(&bond->mcast_work)->work) &rq->__lock irq_context: softirq rcu_read_lock rcu_read_lock &dreq->dreq_lock batched_entropy_u8.lock irq_context: softirq rcu_read_lock rcu_read_lock &dreq->dreq_lock kfence_freelist_lock irq_context: 0 (wq_completion)bond21#3 (work_completion)(&(&bond->mcast_work)->work) fill_pool_map-wait-type-override pool_lock irq_context: 0 (wq_completion)bond28#4 irq_context: 0 (wq_completion)bond28#4 (work_completion)(&(&slave->notify_work)->work) irq_context: 0 (wq_completion)bond28#4 (work_completion)(&(&slave->notify_work)->work) &obj_hash[i].lock irq_context: 0 (wq_completion)bond28#4 (work_completion)(&(&slave->notify_work)->work) &base->lock irq_context: 0 (wq_completion)bond28#4 (work_completion)(&(&slave->notify_work)->work) &base->lock &obj_hash[i].lock irq_context: 0 (wq_completion)bond28#4 (work_completion)(&(&slave->notify_work)->work) &rq->__lock irq_context: 0 (wq_completion)bond28#4 (work_completion)(&(&slave->notify_work)->work) &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)bond28#4 (work_completion)(&(&bond->mcast_work)->work) irq_context: 0 (wq_completion)bond28#4 (work_completion)(&(&bond->mcast_work)->work) &obj_hash[i].lock irq_context: 0 (wq_completion)bond28#4 (work_completion)(&(&bond->mcast_work)->work) &base->lock irq_context: 0 (wq_completion)bond28#4 (work_completion)(&(&bond->mcast_work)->work) &base->lock &obj_hash[i].lock irq_context: 0 (wq_completion)bond28#4 (work_completion)(&(&bond->mcast_work)->work) &rq->__lock irq_context: 0 (wq_completion)bond28#4 (work_completion)(&(&bond->mcast_work)->work) &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)bond137 irq_context: 0 (wq_completion)bond137 (work_completion)(&(&slave->notify_work)->work) irq_context: 0 (wq_completion)bond137 (work_completion)(&(&slave->notify_work)->work) &obj_hash[i].lock irq_context: 0 (wq_completion)bond137 (work_completion)(&(&slave->notify_work)->work) &base->lock irq_context: 0 (wq_completion)bond137 (work_completion)(&(&slave->notify_work)->work) &base->lock &obj_hash[i].lock irq_context: 0 (wq_completion)bond137 (work_completion)(&(&slave->notify_work)->work) &rq->__lock irq_context: 0 (wq_completion)bond137 (work_completion)(&(&slave->notify_work)->work) &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)bond137 (work_completion)(&(&bond->mcast_work)->work) irq_context: 0 (wq_completion)bond137 (work_completion)(&(&bond->mcast_work)->work) &obj_hash[i].lock irq_context: 0 (wq_completion)bond137 (work_completion)(&(&bond->mcast_work)->work) &base->lock irq_context: 0 (wq_completion)bond137 (work_completion)(&(&bond->mcast_work)->work) &base->lock &obj_hash[i].lock irq_context: 0 (wq_completion)bond108#3 irq_context: 0 (wq_completion)bond108#3 &rq->__lock irq_context: 0 (wq_completion)bond108#3 &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)bond108#3 (work_completion)(&(&slave->notify_work)->work) irq_context: 0 (wq_completion)bond108#3 (work_completion)(&(&slave->notify_work)->work) &obj_hash[i].lock irq_context: 0 (wq_completion)bond108#3 (work_completion)(&(&slave->notify_work)->work) &base->lock irq_context: 0 (wq_completion)bond108#3 (work_completion)(&(&slave->notify_work)->work) &base->lock &obj_hash[i].lock irq_context: 0 (wq_completion)bond108#3 (work_completion)(&(&bond->mcast_work)->work) irq_context: 0 (wq_completion)bond108#3 (work_completion)(&(&bond->mcast_work)->work) &obj_hash[i].lock irq_context: 0 (wq_completion)bond108#3 (work_completion)(&(&bond->mcast_work)->work) &base->lock irq_context: 0 (wq_completion)bond108#3 (work_completion)(&(&bond->mcast_work)->work) &base->lock &obj_hash[i].lock irq_context: 0 (wq_completion)bond99#3 (work_completion)(&(&slave->notify_work)->work) &rq->__lock irq_context: 0 (wq_completion)bond99#3 (work_completion)(&(&slave->notify_work)->work) &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)bond118#2 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex fs_reclaim irq_context: 0 (wq_completion)bond118#2 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex fs_reclaim mmu_notifier_invalidate_range_start irq_context: 0 (wq_completion)bond118#2 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &c->lock irq_context: 0 (wq_completion)bond118#2 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex rcu_read_lock &bond->stats_lock/1 irq_context: 0 (wq_completion)bond118#2 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex rcu_read_lock rcu_node_0 irq_context: 0 (wq_completion)bond118#2 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex rcu_read_lock &rq->__lock irq_context: 0 (wq_completion)bond118#2 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex rcu_read_lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)bond127 (work_completion)(&(&bond->mcast_work)->work) &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)bond118#2 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex rcu_read_lock &ndev->lock irq_context: 0 (wq_completion)bond118#2 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex nl_table_lock irq_context: 0 (wq_completion)bond118#2 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &obj_hash[i].lock irq_context: 0 (wq_completion)bond118#2 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex nl_table_wait.lock irq_context: 0 (wq_completion)bond118#2 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex net_rwsem irq_context: 0 (wq_completion)bond118#2 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex net_rwsem &list->lock#2 irq_context: 0 (wq_completion)bond118#2 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex net_rwsem &rq->__lock irq_context: 0 (wq_completion)bond118#2 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex net_rwsem &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)bond118#2 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &tn->lock irq_context: 0 (wq_completion)bond118#2 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &idev->mc_lock irq_context: 0 (wq_completion)bond118#2 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &idev->mc_lock fs_reclaim irq_context: 0 (wq_completion)bond118#2 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &idev->mc_lock fs_reclaim mmu_notifier_invalidate_range_start irq_context: 0 (wq_completion)bond118#2 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &idev->mc_lock rcu_read_lock &ndev->lock irq_context: 0 (wq_completion)bond118#2 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &idev->mc_lock rcu_read_lock &c->lock irq_context: 0 (wq_completion)bond118#2 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &idev->mc_lock rcu_read_lock &dir->lock#2 irq_context: 0 (wq_completion)bond118#2 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &idev->mc_lock rcu_read_lock &ul->lock#2 irq_context: 0 (wq_completion)bond118#2 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &idev->mc_lock rcu_read_lock rcu_read_lock rcu_node_0 irq_context: 0 (wq_completion)bond118#2 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &idev->mc_lock rcu_read_lock rcu_read_lock &rq->__lock irq_context: 0 (wq_completion)bond118#2 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &idev->mc_lock rcu_read_lock rcu_read_lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)bond118#2 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &idev->mc_lock rcu_read_lock rcu_read_lock &____s->seqcount#10 irq_context: 0 (wq_completion)bond118#2 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &idev->mc_lock rcu_read_lock rcu_read_lock rcu_read_lock_bh rcu_read_lock rcu_read_lock_bh rcu_read_lock &br->multicast_lock irq_context: 0 (wq_completion)bond118#2 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &idev->mc_lock rcu_read_lock rcu_read_lock rcu_read_lock_bh rcu_read_lock rcu_read_lock_bh rcu_read_lock &br->multicast_lock &obj_hash[i].lock irq_context: 0 (wq_completion)bond118#2 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &idev->mc_lock rcu_read_lock rcu_read_lock rcu_read_lock_bh rcu_read_lock rcu_read_lock_bh rcu_read_lock &br->multicast_lock &base->lock irq_context: 0 (wq_completion)bond118#2 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &idev->mc_lock rcu_read_lock rcu_read_lock rcu_read_lock_bh rcu_read_lock rcu_read_lock_bh rcu_read_lock &data->lock irq_context: 0 (wq_completion)bond118#2 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &idev->mc_lock rcu_read_lock rcu_read_lock rcu_read_lock_bh rcu_read_lock rcu_read_lock_bh rcu_read_lock &obj_hash[i].lock irq_context: 0 (wq_completion)bond118#2 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &idev->mc_lock rcu_read_lock rcu_node_0 irq_context: 0 (wq_completion)bond118#2 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex.wait_lock irq_context: 0 (wq_completion)bond118#2 (work_completion)(&(&bond->mcast_work)->work) &rq->__lock irq_context: 0 (wq_completion)bond118#2 (work_completion)(&(&bond->mcast_work)->work) &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)bond118#2 (work_completion)(&(&bond->mcast_work)->work) &p->pi_lock irq_context: 0 (wq_completion)bond118#2 (work_completion)(&(&bond->mcast_work)->work) &p->pi_lock &cfs_rq->removed.lock irq_context: 0 (wq_completion)bond118#2 (work_completion)(&(&bond->mcast_work)->work) &p->pi_lock &rq->__lock irq_context: 0 (wq_completion)bond118#2 (work_completion)(&(&bond->mcast_work)->work) &p->pi_lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)bond125#2 irq_context: 0 (wq_completion)bond125#2 &rq->__lock irq_context: 0 (wq_completion)bond125#2 &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)bond125#2 (work_completion)(&(&slave->notify_work)->work) irq_context: 0 (wq_completion)bond125#2 (work_completion)(&(&slave->notify_work)->work) &obj_hash[i].lock irq_context: 0 (wq_completion)bond125#2 (work_completion)(&(&slave->notify_work)->work) &base->lock irq_context: 0 (wq_completion)bond125#2 (work_completion)(&(&slave->notify_work)->work) &base->lock &obj_hash[i].lock irq_context: 0 (wq_completion)bond20#3 &rq->__lock irq_context: 0 (wq_completion)bond102#3 (work_completion)(&(&slave->notify_work)->work) &rq->__lock irq_context: 0 (wq_completion)bond125#2 (work_completion)(&(&bond->mcast_work)->work) irq_context: 0 (wq_completion)bond125#2 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex fs_reclaim irq_context: 0 (wq_completion)bond125#2 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex fs_reclaim mmu_notifier_invalidate_range_start irq_context: 0 (wq_completion)bond125#2 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex fs_reclaim mmu_notifier_invalidate_range_start &rq->__lock irq_context: 0 (wq_completion)bond125#2 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex fs_reclaim mmu_notifier_invalidate_range_start &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)bond125#2 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &c->lock irq_context: 0 (wq_completion)bond125#2 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &n->list_lock irq_context: 0 (wq_completion)bond125#2 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &n->list_lock &c->lock irq_context: 0 (wq_completion)bond125#2 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &rq->__lock irq_context: 0 (wq_completion)bond125#2 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex rcu_read_lock &bond->stats_lock/1 irq_context: 0 (wq_completion)bond125#2 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex rcu_read_lock &ndev->lock irq_context: 0 (wq_completion)bond125#2 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex nl_table_lock irq_context: 0 (wq_completion)bond125#2 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &obj_hash[i].lock irq_context: 0 (wq_completion)bond125#2 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)bond125#2 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex nl_table_wait.lock irq_context: 0 (wq_completion)bond125#2 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex net_rwsem irq_context: 0 (wq_completion)bond125#2 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex net_rwsem &list->lock#2 irq_context: 0 (wq_completion)bond125#2 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &tn->lock irq_context: 0 (wq_completion)bond125#2 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &idev->mc_lock irq_context: 0 (wq_completion)bond125#2 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex.wait_lock irq_context: 0 (wq_completion)bond125#2 (work_completion)(&(&bond->mcast_work)->work) &p->pi_lock irq_context: 0 (wq_completion)bond125#2 (work_completion)(&(&bond->mcast_work)->work) &p->pi_lock &rq->__lock irq_context: 0 (wq_completion)bond125#2 (work_completion)(&(&bond->mcast_work)->work) &p->pi_lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)bond125#2 (work_completion)(&(&bond->mcast_work)->work) &rq->__lock irq_context: 0 (wq_completion)bond125#2 (work_completion)(&(&bond->mcast_work)->work) &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)ipv6_addrconf (work_completion)(&(&ifa->dad_work)->work) rtnl_mutex nl_table_wait.lock &p->pi_lock &rq->__lock irq_context: 0 (wq_completion)ipv6_addrconf (work_completion)(&(&ifa->dad_work)->work) rtnl_mutex nl_table_wait.lock &p->pi_lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 sk_lock-AF_INET6 rcu_read_lock rcu_read_lock pgd_lock irq_context: 0 sk_lock-AF_INET6 rcu_read_lock rcu_read_lock key irq_context: 0 sk_lock-AF_INET6 rcu_read_lock rcu_read_lock pcpu_lock irq_context: 0 sk_lock-AF_INET6 rcu_read_lock rcu_read_lock percpu_counters_lock irq_context: 0 sk_lock-AF_INET6 rcu_read_lock rcu_read_lock pcpu_lock stock_lock irq_context: 0 (wq_completion)bond28#4 (work_completion)(&(&bond->mcast_work)->work) fill_pool_map-wait-type-override &c->lock irq_context: 0 (wq_completion)bond28#4 (work_completion)(&(&bond->mcast_work)->work) fill_pool_map-wait-type-override &____s->seqcount#2 irq_context: 0 (wq_completion)bond28#4 (work_completion)(&(&bond->mcast_work)->work) fill_pool_map-wait-type-override &____s->seqcount irq_context: 0 (wq_completion)bond28#4 (work_completion)(&(&bond->mcast_work)->work) fill_pool_map-wait-type-override pool_lock#2 irq_context: 0 (wq_completion)bond28#4 (work_completion)(&(&bond->mcast_work)->work) fill_pool_map-wait-type-override pool_lock irq_context: 0 (wq_completion)bond109#3 irq_context: 0 (wq_completion)bond109#3 (work_completion)(&(&slave->notify_work)->work) irq_context: 0 (wq_completion)bond109#3 (work_completion)(&(&slave->notify_work)->work) &obj_hash[i].lock irq_context: 0 (wq_completion)bond109#3 (work_completion)(&(&slave->notify_work)->work) &base->lock irq_context: 0 (wq_completion)bond109#3 (work_completion)(&(&slave->notify_work)->work) &base->lock &obj_hash[i].lock irq_context: 0 (wq_completion)bond109#3 (work_completion)(&(&slave->notify_work)->work) &rq->__lock irq_context: 0 (wq_completion)bond109#3 (work_completion)(&(&slave->notify_work)->work) &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)bond109#3 (work_completion)(&(&bond->mcast_work)->work) irq_context: 0 (wq_completion)bond109#3 (work_completion)(&(&bond->mcast_work)->work) &obj_hash[i].lock irq_context: 0 (wq_completion)bond109#3 (work_completion)(&(&bond->mcast_work)->work) &base->lock irq_context: 0 (wq_completion)bond109#3 (work_completion)(&(&bond->mcast_work)->work) &base->lock &obj_hash[i].lock irq_context: 0 (wq_completion)bond125 (work_completion)(&(&slave->notify_work)->work) rtnl_mutex net_rwsem irq_context: 0 (wq_completion)bond125 (work_completion)(&(&slave->notify_work)->work) rtnl_mutex net_rwsem &list->lock#2 irq_context: 0 (wq_completion)bond125 (work_completion)(&(&slave->notify_work)->work) rtnl_mutex &tn->lock irq_context: 0 (wq_completion)bond125 (work_completion)(&(&slave->notify_work)->work) rtnl_mutex.wait_lock irq_context: 0 (wq_completion)bond125 (work_completion)(&(&slave->notify_work)->work) &p->pi_lock irq_context: 0 (wq_completion)bond22#4 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex fs_reclaim irq_context: 0 (wq_completion)bond22#4 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex fs_reclaim mmu_notifier_invalidate_range_start irq_context: 0 (wq_completion)bond22#4 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &c->lock irq_context: 0 (wq_completion)bond22#4 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex rcu_read_lock &bond->stats_lock/1 irq_context: 0 (wq_completion)bond22#4 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex rcu_read_lock &ndev->lock irq_context: 0 (wq_completion)bond22#4 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex nl_table_lock irq_context: 0 (wq_completion)bond22#4 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &obj_hash[i].lock irq_context: 0 (wq_completion)bond22#4 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex nl_table_wait.lock irq_context: 0 (wq_completion)bond22#4 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex net_rwsem irq_context: 0 (wq_completion)bond22#4 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex net_rwsem &list->lock#2 irq_context: 0 (wq_completion)bond22#4 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &tn->lock irq_context: 0 (wq_completion)bond22#4 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &idev->mc_lock irq_context: 0 (wq_completion)bond22#4 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &idev->mc_lock fs_reclaim irq_context: 0 (wq_completion)bond22#4 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &idev->mc_lock fs_reclaim mmu_notifier_invalidate_range_start irq_context: 0 (wq_completion)bond22#4 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &idev->mc_lock rcu_read_lock &ndev->lock irq_context: 0 (wq_completion)bond22#4 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &idev->mc_lock rcu_read_lock &c->lock irq_context: 0 (wq_completion)bond22#4 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &idev->mc_lock rcu_read_lock &dir->lock#2 irq_context: 0 (wq_completion)bond22#4 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &idev->mc_lock rcu_read_lock &ul->lock#2 irq_context: 0 (wq_completion)bond22#4 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &idev->mc_lock rcu_read_lock rcu_read_lock &____s->seqcount#12 irq_context: 0 (wq_completion)bond22#4 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &idev->mc_lock rcu_read_lock rcu_read_lock &____s->seqcount#16 irq_context: 0 (wq_completion)bond22#4 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &idev->mc_lock rcu_read_lock &obj_hash[i].lock irq_context: 0 (wq_completion)bond22#4 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &idev->mc_lock rcu_read_lock &data->lock irq_context: 0 (wq_completion)bond22#4 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex.wait_lock irq_context: 0 (wq_completion)bond22#4 (work_completion)(&(&bond->mcast_work)->work) &p->pi_lock irq_context: 0 (wq_completion)bond106#3 (work_completion)(&(&slave->notify_work)->work) &rq->__lock irq_context: 0 (wq_completion)bond126#2 irq_context: 0 (wq_completion)bond126#2 &rq->__lock irq_context: 0 (wq_completion)bond126#2 (work_completion)(&(&slave->notify_work)->work) irq_context: 0 (wq_completion)bond126#2 (work_completion)(&(&slave->notify_work)->work) &obj_hash[i].lock irq_context: 0 (wq_completion)bond126#2 (work_completion)(&(&slave->notify_work)->work) &base->lock irq_context: 0 (wq_completion)bond126#2 (work_completion)(&(&slave->notify_work)->work) &base->lock &obj_hash[i].lock irq_context: 0 (wq_completion)bond126#2 (work_completion)(&(&slave->notify_work)->work) &rq->__lock irq_context: 0 (wq_completion)bond126#2 (work_completion)(&(&slave->notify_work)->work) &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)bond126#2 (work_completion)(&(&bond->mcast_work)->work) irq_context: 0 (wq_completion)bond126#2 (work_completion)(&(&bond->mcast_work)->work) &obj_hash[i].lock irq_context: 0 (wq_completion)bond126#2 (work_completion)(&(&bond->mcast_work)->work) &base->lock irq_context: 0 (wq_completion)bond126#2 (work_completion)(&(&bond->mcast_work)->work) &base->lock &obj_hash[i].lock irq_context: 0 (wq_completion)bond126#2 (work_completion)(&(&bond->mcast_work)->work) &rq->__lock irq_context: 0 (wq_completion)bond126#2 (work_completion)(&(&bond->mcast_work)->work) &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)bond123 (work_completion)(&(&bond->mcast_work)->work) &rq->__lock irq_context: 0 (wq_completion)bond123 (work_completion)(&(&bond->mcast_work)->work) &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)bond23#3 (work_completion)(&(&slave->notify_work)->work) &rq->__lock irq_context: 0 (wq_completion)bond117#2 (work_completion)(&(&bond->mcast_work)->work) &rq->__lock irq_context: 0 (wq_completion)bond117#2 (work_completion)(&(&bond->mcast_work)->work) &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)bond139 irq_context: 0 (wq_completion)bond139 (work_completion)(&(&slave->notify_work)->work) irq_context: 0 (wq_completion)bond139 (work_completion)(&(&slave->notify_work)->work) &obj_hash[i].lock irq_context: 0 (wq_completion)bond139 (work_completion)(&(&slave->notify_work)->work) &base->lock irq_context: 0 (wq_completion)bond139 (work_completion)(&(&slave->notify_work)->work) &base->lock &obj_hash[i].lock irq_context: 0 (wq_completion)bond139 (work_completion)(&(&slave->notify_work)->work) &rq->__lock irq_context: 0 (wq_completion)bond109#2 (work_completion)(&(&bond->mcast_work)->work) &rq->__lock irq_context: 0 (wq_completion)bond109#2 (work_completion)(&(&bond->mcast_work)->work) &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 rtnl_mutex &idev->mc_lock fill_pool_map-wait-type-override rcu_read_lock rcu_node_0 irq_context: 0 rtnl_mutex &idev->mc_lock fill_pool_map-wait-type-override rcu_read_lock &rq->__lock irq_context: 0 rtnl_mutex &idev->mc_lock fill_pool_map-wait-type-override rcu_read_lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)bond139 (work_completion)(&(&bond->mcast_work)->work) irq_context: 0 (wq_completion)bond139 (work_completion)(&(&bond->mcast_work)->work) &obj_hash[i].lock irq_context: 0 (wq_completion)bond139 (work_completion)(&(&bond->mcast_work)->work) &base->lock irq_context: 0 (wq_completion)bond139 (work_completion)(&(&bond->mcast_work)->work) &base->lock &obj_hash[i].lock irq_context: 0 (wq_completion)bond139 (work_completion)(&(&bond->mcast_work)->work) &rq->__lock irq_context: 0 (wq_completion)bond139 (work_completion)(&(&bond->mcast_work)->work) &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)bond110#3 irq_context: 0 (wq_completion)bond110#3 (work_completion)(&(&slave->notify_work)->work) irq_context: 0 (wq_completion)bond110#3 (work_completion)(&(&slave->notify_work)->work) &obj_hash[i].lock irq_context: 0 (wq_completion)bond110#3 (work_completion)(&(&slave->notify_work)->work) &base->lock irq_context: 0 (wq_completion)bond110#3 (work_completion)(&(&slave->notify_work)->work) &base->lock &obj_hash[i].lock irq_context: 0 (wq_completion)bond110#3 (work_completion)(&(&slave->notify_work)->work) &rq->__lock irq_context: 0 (wq_completion)bond110#3 (work_completion)(&(&slave->notify_work)->work) &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)bond110#3 (work_completion)(&(&bond->mcast_work)->work) irq_context: 0 (wq_completion)bond110#3 (work_completion)(&(&bond->mcast_work)->work) &obj_hash[i].lock irq_context: 0 (wq_completion)bond110#3 (work_completion)(&(&bond->mcast_work)->work) &base->lock irq_context: 0 (wq_completion)bond110#3 (work_completion)(&(&bond->mcast_work)->work) &base->lock &obj_hash[i].lock irq_context: 0 (wq_completion)bond110#3 (work_completion)(&(&bond->mcast_work)->work) fill_pool_map-wait-type-override pool_lock irq_context: softirq (&ndev->rs_timer) fill_pool_map-wait-type-override pool_lock#2 irq_context: softirq (&ndev->rs_timer) fill_pool_map-wait-type-override &c->lock irq_context: softirq (&ndev->rs_timer) fill_pool_map-wait-type-override &____s->seqcount#2 irq_context: softirq (&ndev->rs_timer) fill_pool_map-wait-type-override &____s->seqcount irq_context: softirq (&ndev->rs_timer) fill_pool_map-wait-type-override pool_lock irq_context: 0 (wq_completion)bond29#4 irq_context: 0 (wq_completion)bond29#4 (work_completion)(&(&slave->notify_work)->work) irq_context: 0 (wq_completion)bond29#4 (work_completion)(&(&slave->notify_work)->work) &obj_hash[i].lock irq_context: 0 (wq_completion)bond29#4 (work_completion)(&(&slave->notify_work)->work) &base->lock irq_context: 0 (wq_completion)bond29#4 (work_completion)(&(&slave->notify_work)->work) &base->lock &obj_hash[i].lock irq_context: 0 (wq_completion)bond29#4 (work_completion)(&(&slave->notify_work)->work) &rq->__lock irq_context: 0 (wq_completion)bond29#4 (work_completion)(&(&slave->notify_work)->work) &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 &sb->s_type->i_mutex_key#9 &rcu_state.gp_wq irq_context: 0 &sb->s_type->i_mutex_key#9 &rcu_state.gp_wq &p->pi_lock irq_context: 0 &sb->s_type->i_mutex_key#9 &rcu_state.gp_wq &p->pi_lock &rq->__lock irq_context: 0 &sb->s_type->i_mutex_key#9 &rcu_state.gp_wq &p->pi_lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)bond29#4 (work_completion)(&(&bond->mcast_work)->work) irq_context: 0 (wq_completion)bond29#4 (work_completion)(&(&bond->mcast_work)->work) &obj_hash[i].lock irq_context: 0 (wq_completion)bond29#4 (work_completion)(&(&bond->mcast_work)->work) &base->lock irq_context: 0 (wq_completion)bond29#4 (work_completion)(&(&bond->mcast_work)->work) &base->lock &obj_hash[i].lock irq_context: 0 (wq_completion)bond136 (work_completion)(&(&bond->mcast_work)->work) &rq->__lock irq_context: 0 (wq_completion)bond136 (work_completion)(&(&bond->mcast_work)->work) &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)bond27#4 &rq->__lock irq_context: 0 (wq_completion)bond27#4 &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)bond123#2 &rq->__lock irq_context: 0 (wq_completion)bond123#2 &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)bond110#3 (work_completion)(&(&bond->mcast_work)->work) &rq->__lock irq_context: 0 (wq_completion)bond110#3 (work_completion)(&(&bond->mcast_work)->work) &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)bond111#3 irq_context: 0 (wq_completion)bond111#3 (work_completion)(&(&slave->notify_work)->work) irq_context: 0 (wq_completion)bond111#3 (work_completion)(&(&slave->notify_work)->work) &obj_hash[i].lock irq_context: 0 (wq_completion)bond111#3 (work_completion)(&(&slave->notify_work)->work) &base->lock irq_context: 0 (wq_completion)bond111#3 (work_completion)(&(&slave->notify_work)->work) &base->lock &obj_hash[i].lock irq_context: 0 (wq_completion)bond111#3 (work_completion)(&(&slave->notify_work)->work) &rq->__lock irq_context: 0 (wq_completion)bond111#3 (work_completion)(&(&slave->notify_work)->work) &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)bond111#3 (work_completion)(&(&bond->mcast_work)->work) irq_context: 0 (wq_completion)bond111#3 (work_completion)(&(&bond->mcast_work)->work) &obj_hash[i].lock irq_context: 0 (wq_completion)bond111#3 (work_completion)(&(&bond->mcast_work)->work) &base->lock irq_context: 0 (wq_completion)bond111#3 (work_completion)(&(&bond->mcast_work)->work) &base->lock &obj_hash[i].lock irq_context: 0 (wq_completion)bond111#3 (work_completion)(&(&bond->mcast_work)->work) &rq->__lock irq_context: 0 (wq_completion)bond111#3 (work_completion)(&(&bond->mcast_work)->work) &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)bond123#2 (work_completion)(&(&slave->notify_work)->work) &rq->__lock irq_context: 0 (wq_completion)bond123#2 (work_completion)(&(&slave->notify_work)->work) &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)bond127#2 irq_context: 0 (wq_completion)bond127#2 (work_completion)(&(&slave->notify_work)->work) irq_context: 0 (wq_completion)bond127#2 (work_completion)(&(&slave->notify_work)->work) &obj_hash[i].lock irq_context: 0 (wq_completion)bond127#2 (work_completion)(&(&slave->notify_work)->work) &base->lock irq_context: 0 (wq_completion)bond127#2 (work_completion)(&(&slave->notify_work)->work) &base->lock &obj_hash[i].lock irq_context: 0 (wq_completion)bond127#2 (work_completion)(&(&slave->notify_work)->work) &rq->__lock irq_context: 0 (wq_completion)bond112#2 (work_completion)(&(&bond->mcast_work)->work) &rq->__lock irq_context: 0 (wq_completion)bond112#2 (work_completion)(&(&bond->mcast_work)->work) &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)bond127#2 (work_completion)(&(&slave->notify_work)->work) rtnl_mutex net_rwsem irq_context: 0 (wq_completion)bond127#2 (work_completion)(&(&slave->notify_work)->work) rtnl_mutex net_rwsem &list->lock#2 irq_context: 0 (wq_completion)bond127#2 (work_completion)(&(&slave->notify_work)->work) rtnl_mutex &tn->lock irq_context: 0 (wq_completion)bond127#2 (work_completion)(&(&slave->notify_work)->work) rtnl_mutex.wait_lock irq_context: 0 (wq_completion)bond127#2 (work_completion)(&(&slave->notify_work)->work) &p->pi_lock irq_context: 0 (wq_completion)bond27#4 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex fs_reclaim irq_context: 0 (wq_completion)bond27#4 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex fs_reclaim mmu_notifier_invalidate_range_start irq_context: 0 (wq_completion)bond27#4 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex fs_reclaim mmu_notifier_invalidate_range_start &rq->__lock irq_context: 0 (wq_completion)bond27#4 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex fs_reclaim mmu_notifier_invalidate_range_start &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)bond27#4 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &c->lock irq_context: 0 (wq_completion)bond27#4 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &____s->seqcount#2 irq_context: 0 (wq_completion)bond27#4 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &____s->seqcount irq_context: 0 (wq_completion)bond27#4 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex rcu_read_lock &bond->stats_lock/1 irq_context: 0 (wq_completion)bond27#4 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex rcu_read_lock &ndev->lock irq_context: 0 (wq_completion)bond27#4 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex nl_table_lock irq_context: 0 (wq_completion)bond27#4 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &obj_hash[i].lock irq_context: 0 (wq_completion)bond27#4 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex nl_table_wait.lock irq_context: 0 (wq_completion)bond27#4 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex net_rwsem irq_context: 0 (wq_completion)bond27#4 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex net_rwsem &list->lock#2 irq_context: 0 (wq_completion)bond27#4 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &tn->lock irq_context: 0 (wq_completion)bond27#4 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &idev->mc_lock irq_context: 0 (wq_completion)bond27#4 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &idev->mc_lock &rq->__lock irq_context: 0 (wq_completion)bond27#4 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &idev->mc_lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)bond27#4 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &idev->mc_lock fs_reclaim irq_context: 0 (wq_completion)bond27#4 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &idev->mc_lock fs_reclaim mmu_notifier_invalidate_range_start irq_context: 0 (wq_completion)bond27#4 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &idev->mc_lock &c->lock irq_context: 0 (wq_completion)bond27#4 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &idev->mc_lock rcu_read_lock &ndev->lock irq_context: 0 (wq_completion)bond27#4 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &idev->mc_lock rcu_read_lock &dir->lock#2 irq_context: 0 (wq_completion)bond27#4 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &idev->mc_lock rcu_read_lock &c->lock irq_context: 0 (wq_completion)bond27#4 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &idev->mc_lock rcu_read_lock &ul->lock#2 irq_context: 0 (wq_completion)bond27#4 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &idev->mc_lock rcu_read_lock rcu_read_lock &____s->seqcount#12 irq_context: 0 (wq_completion)bond27#4 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &idev->mc_lock rcu_read_lock rcu_read_lock &____s->seqcount#16 irq_context: 0 (wq_completion)bond27#4 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &idev->mc_lock rcu_read_lock &obj_hash[i].lock irq_context: 0 (wq_completion)bond27#4 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &idev->mc_lock rcu_read_lock rcu_node_0 irq_context: 0 (wq_completion)bond27#4 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &idev->mc_lock rcu_read_lock &rq->__lock irq_context: 0 (wq_completion)bond27#4 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &idev->mc_lock rcu_read_lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)bond27#4 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &idev->mc_lock rcu_read_lock &data->lock irq_context: 0 (wq_completion)bond27#4 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex.wait_lock irq_context: 0 (wq_completion)bond27#4 (work_completion)(&(&bond->mcast_work)->work) &p->pi_lock irq_context: 0 (wq_completion)bond27#4 (work_completion)(&(&bond->mcast_work)->work) &p->pi_lock &rq->__lock irq_context: 0 (wq_completion)bond27#4 (work_completion)(&(&bond->mcast_work)->work) &p->pi_lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)bond27#4 (work_completion)(&(&bond->mcast_work)->work) &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)bond127#2 (work_completion)(&(&bond->mcast_work)->work) irq_context: 0 (wq_completion)bond127#2 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex fs_reclaim irq_context: 0 (wq_completion)bond127#2 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex fs_reclaim &rq->__lock irq_context: 0 (wq_completion)bond127#2 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex fs_reclaim &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)bond127#2 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex fs_reclaim mmu_notifier_invalidate_range_start irq_context: 0 (wq_completion)bond127#2 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex rcu_read_lock &bond->stats_lock/1 irq_context: 0 (wq_completion)bond127#2 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex rcu_read_lock &ndev->lock irq_context: 0 (wq_completion)bond127#2 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex nl_table_lock irq_context: 0 (wq_completion)bond127#2 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &obj_hash[i].lock irq_context: 0 (wq_completion)bond127#2 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex nl_table_wait.lock irq_context: 0 (wq_completion)bond127#2 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex net_rwsem irq_context: 0 (wq_completion)bond127#2 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex net_rwsem &list->lock#2 irq_context: 0 (wq_completion)bond127#2 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &tn->lock irq_context: 0 (wq_completion)bond127#2 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &idev->mc_lock irq_context: 0 (wq_completion)bond127#2 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &rq->__lock irq_context: 0 (wq_completion)bond127#2 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)bond127#2 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex.wait_lock irq_context: 0 (wq_completion)bond127#2 (work_completion)(&(&bond->mcast_work)->work) &p->pi_lock irq_context: 0 (wq_completion)bond102#3 &rq->__lock irq_context: 0 (wq_completion)bond102#3 &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)bond120#2 (work_completion)(&(&slave->notify_work)->work) &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)bond140 irq_context: 0 (wq_completion)bond140 (work_completion)(&(&slave->notify_work)->work) irq_context: 0 (wq_completion)bond140 (work_completion)(&(&slave->notify_work)->work) &obj_hash[i].lock irq_context: 0 (wq_completion)bond140 (work_completion)(&(&slave->notify_work)->work) &base->lock irq_context: 0 (wq_completion)bond140 (work_completion)(&(&slave->notify_work)->work) &base->lock &obj_hash[i].lock irq_context: 0 (wq_completion)bond140 (work_completion)(&(&slave->notify_work)->work) &rq->__lock irq_context: 0 (wq_completion)bond140 (work_completion)(&(&slave->notify_work)->work) &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)bond94#2 (work_completion)(&(&slave->notify_work)->work) &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)bond136 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex fs_reclaim irq_context: 0 (wq_completion)bond136 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex fs_reclaim mmu_notifier_invalidate_range_start irq_context: 0 (wq_completion)bond136 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex rcu_read_lock &bond->stats_lock/1 irq_context: 0 (wq_completion)bond136 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex rcu_read_lock &ndev->lock irq_context: 0 (wq_completion)bond136 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex nl_table_lock irq_context: 0 (wq_completion)bond136 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &obj_hash[i].lock irq_context: 0 (wq_completion)bond136 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex nl_table_wait.lock irq_context: 0 (wq_completion)bond136 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex net_rwsem irq_context: 0 (wq_completion)bond136 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex net_rwsem &list->lock#2 irq_context: 0 (wq_completion)bond136 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &tn->lock irq_context: 0 (wq_completion)bond136 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &idev->mc_lock irq_context: 0 (wq_completion)bond136 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &idev->mc_lock fs_reclaim irq_context: 0 (wq_completion)bond136 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &idev->mc_lock fs_reclaim mmu_notifier_invalidate_range_start irq_context: 0 (wq_completion)bond136 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &idev->mc_lock &c->lock irq_context: 0 (wq_completion)bond136 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &idev->mc_lock &n->list_lock irq_context: 0 (wq_completion)bond136 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &idev->mc_lock &n->list_lock &c->lock irq_context: 0 (wq_completion)bond136 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &idev->mc_lock rcu_read_lock &ndev->lock irq_context: 0 (wq_completion)bond136 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &idev->mc_lock rcu_read_lock &c->lock irq_context: 0 (wq_completion)bond136 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &idev->mc_lock rcu_read_lock &dir->lock#2 irq_context: 0 (wq_completion)bond136 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &idev->mc_lock rcu_read_lock &ul->lock#2 irq_context: 0 (wq_completion)bond136 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &idev->mc_lock rcu_read_lock rcu_read_lock &____s->seqcount#10 irq_context: 0 (wq_completion)bond136 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &idev->mc_lock rcu_read_lock rcu_read_lock rcu_read_lock_bh rcu_read_lock rcu_read_lock_bh rcu_read_lock &br->multicast_lock irq_context: 0 (wq_completion)bond136 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &idev->mc_lock rcu_read_lock rcu_read_lock rcu_read_lock_bh rcu_read_lock rcu_read_lock_bh rcu_read_lock &br->multicast_lock &obj_hash[i].lock irq_context: 0 (wq_completion)bond136 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &idev->mc_lock rcu_read_lock rcu_read_lock rcu_read_lock_bh rcu_read_lock rcu_read_lock_bh rcu_read_lock &br->multicast_lock &base->lock irq_context: 0 (wq_completion)bond136 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &idev->mc_lock rcu_read_lock rcu_read_lock rcu_read_lock_bh rcu_read_lock rcu_read_lock_bh rcu_read_lock &data->lock irq_context: 0 (wq_completion)bond136 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &idev->mc_lock rcu_read_lock rcu_read_lock rcu_read_lock_bh rcu_read_lock rcu_read_lock_bh rcu_read_lock &obj_hash[i].lock irq_context: 0 (wq_completion)bond136 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex.wait_lock irq_context: 0 (wq_completion)bond136 (work_completion)(&(&bond->mcast_work)->work) &p->pi_lock irq_context: 0 (wq_completion)bond140 (work_completion)(&(&bond->mcast_work)->work) irq_context: 0 (wq_completion)bond140 (work_completion)(&(&bond->mcast_work)->work) &obj_hash[i].lock irq_context: 0 (wq_completion)bond140 (work_completion)(&(&bond->mcast_work)->work) &base->lock irq_context: 0 (wq_completion)bond140 (work_completion)(&(&bond->mcast_work)->work) &base->lock &obj_hash[i].lock irq_context: 0 (wq_completion)bond122 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex fs_reclaim irq_context: 0 (wq_completion)bond122 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex fs_reclaim mmu_notifier_invalidate_range_start irq_context: 0 (wq_completion)bond122 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &c->lock irq_context: 0 (wq_completion)bond122 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &n->list_lock irq_context: 0 (wq_completion)bond122 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &n->list_lock &c->lock irq_context: 0 (wq_completion)bond122 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex rcu_read_lock &bond->stats_lock/1 irq_context: 0 (wq_completion)bond122 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex rcu_read_lock &ndev->lock irq_context: 0 (wq_completion)bond122 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex nl_table_lock irq_context: 0 (wq_completion)bond122 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &obj_hash[i].lock irq_context: 0 (wq_completion)bond122 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex nl_table_wait.lock irq_context: 0 (wq_completion)bond122 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex net_rwsem irq_context: 0 (wq_completion)bond122 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex net_rwsem &list->lock#2 irq_context: 0 (wq_completion)bond122 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &tn->lock irq_context: 0 (wq_completion)bond122 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &idev->mc_lock irq_context: 0 (wq_completion)bond122 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &idev->mc_lock fs_reclaim irq_context: 0 (wq_completion)bond122 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &idev->mc_lock fs_reclaim mmu_notifier_invalidate_range_start irq_context: 0 (wq_completion)bond122 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &idev->mc_lock &c->lock irq_context: 0 (wq_completion)bond122 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &idev->mc_lock rcu_read_lock &ndev->lock irq_context: 0 (wq_completion)bond122 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &idev->mc_lock rcu_read_lock &dir->lock#2 irq_context: 0 (wq_completion)bond122 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &idev->mc_lock rcu_read_lock &ul->lock#2 irq_context: 0 (wq_completion)bond122 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &idev->mc_lock rcu_read_lock rcu_read_lock &____s->seqcount#10 irq_context: 0 (wq_completion)bond122 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &idev->mc_lock rcu_read_lock rcu_read_lock rcu_read_lock_bh rcu_read_lock rcu_read_lock_bh rcu_read_lock &br->multicast_lock irq_context: 0 (wq_completion)bond122 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &idev->mc_lock rcu_read_lock rcu_read_lock rcu_read_lock_bh rcu_read_lock rcu_read_lock_bh rcu_read_lock &br->multicast_lock &obj_hash[i].lock irq_context: 0 (wq_completion)bond122 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &idev->mc_lock rcu_read_lock rcu_read_lock rcu_read_lock_bh rcu_read_lock rcu_read_lock_bh rcu_read_lock &br->multicast_lock &base->lock irq_context: 0 (wq_completion)bond122 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &idev->mc_lock rcu_read_lock rcu_read_lock rcu_read_lock_bh rcu_read_lock rcu_read_lock_bh rcu_read_lock &br->multicast_lock &base->lock &obj_hash[i].lock irq_context: 0 (wq_completion)bond122 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &idev->mc_lock rcu_read_lock rcu_read_lock rcu_read_lock_bh rcu_read_lock rcu_read_lock_bh rcu_read_lock &data->lock irq_context: 0 (wq_completion)bond122 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &idev->mc_lock rcu_read_lock rcu_read_lock rcu_read_lock_bh rcu_read_lock rcu_read_lock_bh rcu_read_lock &obj_hash[i].lock irq_context: 0 (wq_completion)bond122 (work_completion)(&(&slave->notify_work)->work) rtnl_mutex net_rwsem irq_context: 0 (wq_completion)bond122 (work_completion)(&(&slave->notify_work)->work) rtnl_mutex net_rwsem &list->lock#2 irq_context: 0 (wq_completion)bond122 (work_completion)(&(&slave->notify_work)->work) rtnl_mutex &tn->lock irq_context: 0 (wq_completion)bond105#3 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex fs_reclaim irq_context: 0 (wq_completion)bond105#3 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex fs_reclaim mmu_notifier_invalidate_range_start irq_context: 0 (wq_completion)bond105#3 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex rcu_read_lock &bond->stats_lock/1 irq_context: 0 (wq_completion)bond105#3 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex rcu_read_lock &ndev->lock irq_context: 0 (wq_completion)bond105#3 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex nl_table_lock irq_context: 0 (wq_completion)bond105#3 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &obj_hash[i].lock irq_context: 0 (wq_completion)bond105#3 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex nl_table_wait.lock irq_context: 0 (wq_completion)bond105#3 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex net_rwsem irq_context: 0 (wq_completion)bond105#3 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex net_rwsem &list->lock#2 irq_context: 0 (wq_completion)bond105#3 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &tn->lock irq_context: 0 (wq_completion)bond105#3 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &idev->mc_lock irq_context: 0 (wq_completion)bond105#3 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &idev->mc_lock fs_reclaim irq_context: 0 (wq_completion)bond105#3 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &idev->mc_lock fs_reclaim mmu_notifier_invalidate_range_start irq_context: 0 (wq_completion)bond105#3 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &idev->mc_lock &c->lock irq_context: 0 (wq_completion)bond105#3 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &idev->mc_lock rcu_read_lock &ndev->lock irq_context: 0 (wq_completion)bond105#3 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &idev->mc_lock rcu_read_lock &dir->lock#2 irq_context: 0 (wq_completion)bond105#3 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &idev->mc_lock rcu_read_lock &ul->lock#2 irq_context: 0 (wq_completion)bond105#3 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &idev->mc_lock rcu_read_lock rcu_read_lock &____s->seqcount#10 irq_context: 0 (wq_completion)bond105#3 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &idev->mc_lock rcu_read_lock rcu_read_lock rcu_read_lock_bh rcu_read_lock rcu_read_lock_bh rcu_read_lock &br->multicast_lock irq_context: 0 (wq_completion)bond105#3 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &idev->mc_lock rcu_read_lock rcu_read_lock rcu_read_lock_bh rcu_read_lock rcu_read_lock_bh rcu_read_lock &br->multicast_lock &obj_hash[i].lock irq_context: 0 (wq_completion)bond105#3 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &idev->mc_lock rcu_read_lock rcu_read_lock rcu_read_lock_bh rcu_read_lock rcu_read_lock_bh rcu_read_lock &br->multicast_lock &base->lock irq_context: 0 (wq_completion)bond105#3 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &idev->mc_lock rcu_read_lock rcu_read_lock rcu_read_lock_bh rcu_read_lock rcu_read_lock_bh rcu_read_lock &data->lock irq_context: 0 (wq_completion)bond105#3 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &idev->mc_lock rcu_read_lock rcu_read_lock rcu_read_lock_bh rcu_read_lock rcu_read_lock_bh rcu_read_lock &obj_hash[i].lock irq_context: 0 (wq_completion)bond112#2 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex fs_reclaim irq_context: 0 (wq_completion)bond112#2 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex fs_reclaim mmu_notifier_invalidate_range_start irq_context: 0 (wq_completion)bond112#2 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &c->lock irq_context: 0 (wq_completion)bond112#2 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &n->list_lock irq_context: 0 (wq_completion)bond112#2 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &n->list_lock &c->lock irq_context: 0 (wq_completion)bond112#2 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex rcu_read_lock &bond->stats_lock/1 irq_context: 0 (wq_completion)bond112#2 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex rcu_read_lock &ndev->lock irq_context: 0 (wq_completion)bond112#2 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex nl_table_lock irq_context: 0 (wq_completion)bond112#2 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &obj_hash[i].lock irq_context: 0 (wq_completion)bond112#2 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex nl_table_wait.lock irq_context: 0 (wq_completion)bond112#2 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex net_rwsem irq_context: 0 (wq_completion)bond112#2 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex net_rwsem &list->lock#2 irq_context: 0 (wq_completion)bond112#2 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &tn->lock irq_context: 0 (wq_completion)bond112#2 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &idev->mc_lock irq_context: 0 (wq_completion)bond112#2 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &idev->mc_lock fs_reclaim irq_context: 0 (wq_completion)bond112#2 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &idev->mc_lock fs_reclaim mmu_notifier_invalidate_range_start irq_context: 0 (wq_completion)bond112#2 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &idev->mc_lock rcu_read_lock &ndev->lock irq_context: 0 (wq_completion)bond112#2 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &idev->mc_lock rcu_read_lock &dir->lock#2 irq_context: 0 (wq_completion)bond112#2 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &idev->mc_lock rcu_read_lock &ul->lock#2 irq_context: 0 (wq_completion)bond112#2 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &idev->mc_lock rcu_read_lock rcu_read_lock &____s->seqcount#10 irq_context: 0 (wq_completion)bond112#2 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &idev->mc_lock rcu_read_lock rcu_read_lock rcu_read_lock_bh rcu_read_lock rcu_read_lock_bh rcu_read_lock &br->multicast_lock irq_context: 0 (wq_completion)bond112#2 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &idev->mc_lock rcu_read_lock rcu_read_lock rcu_read_lock_bh rcu_read_lock rcu_read_lock_bh rcu_read_lock &br->multicast_lock &obj_hash[i].lock irq_context: 0 (wq_completion)bond112#2 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &idev->mc_lock rcu_read_lock rcu_read_lock rcu_read_lock_bh rcu_read_lock rcu_read_lock_bh rcu_read_lock &br->multicast_lock &base->lock irq_context: 0 (wq_completion)bond112#2 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &idev->mc_lock rcu_read_lock rcu_read_lock rcu_read_lock_bh rcu_read_lock rcu_read_lock_bh rcu_read_lock &br->multicast_lock &base->lock &obj_hash[i].lock irq_context: 0 (wq_completion)bond112#2 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &idev->mc_lock rcu_read_lock rcu_read_lock rcu_read_lock_bh rcu_read_lock rcu_read_lock_bh rcu_read_lock &data->lock irq_context: 0 (wq_completion)bond112#2 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &idev->mc_lock rcu_read_lock rcu_read_lock rcu_read_lock_bh rcu_read_lock rcu_read_lock_bh rcu_read_lock &obj_hash[i].lock irq_context: 0 (wq_completion)bond104#3 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex fs_reclaim irq_context: 0 (wq_completion)bond104#3 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex fs_reclaim mmu_notifier_invalidate_range_start irq_context: 0 (wq_completion)bond104#3 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &c->lock irq_context: 0 (wq_completion)bond104#3 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex rcu_read_lock &bond->stats_lock/1 irq_context: 0 (wq_completion)bond104#3 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex rcu_read_lock &ndev->lock irq_context: 0 (wq_completion)bond104#3 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex nl_table_lock irq_context: 0 (wq_completion)bond104#3 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &obj_hash[i].lock irq_context: 0 (wq_completion)bond104#3 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex nl_table_wait.lock irq_context: 0 (wq_completion)bond104#3 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex net_rwsem irq_context: 0 (wq_completion)bond104#3 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex net_rwsem &list->lock#2 irq_context: 0 (wq_completion)bond104#3 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &tn->lock irq_context: 0 (wq_completion)bond104#3 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &idev->mc_lock irq_context: 0 (wq_completion)bond104#3 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &idev->mc_lock fs_reclaim irq_context: 0 (wq_completion)bond104#3 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &idev->mc_lock fs_reclaim mmu_notifier_invalidate_range_start irq_context: 0 (wq_completion)bond104#3 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &idev->mc_lock &c->lock irq_context: 0 (wq_completion)bond104#3 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &idev->mc_lock rcu_read_lock &ndev->lock irq_context: 0 (wq_completion)bond104#3 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &idev->mc_lock rcu_read_lock &dir->lock#2 irq_context: 0 (wq_completion)bond104#3 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &idev->mc_lock rcu_read_lock &ul->lock#2 irq_context: 0 (wq_completion)bond104#3 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &idev->mc_lock rcu_read_lock rcu_read_lock &____s->seqcount#10 irq_context: 0 (wq_completion)bond104#3 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &idev->mc_lock rcu_read_lock rcu_read_lock rcu_read_lock_bh rcu_read_lock rcu_read_lock_bh rcu_read_lock &br->multicast_lock irq_context: 0 (wq_completion)bond104#3 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &idev->mc_lock rcu_read_lock rcu_read_lock rcu_read_lock_bh rcu_read_lock rcu_read_lock_bh rcu_read_lock &br->multicast_lock &obj_hash[i].lock irq_context: 0 (wq_completion)bond104#3 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &idev->mc_lock rcu_read_lock rcu_read_lock rcu_read_lock_bh rcu_read_lock rcu_read_lock_bh rcu_read_lock &br->multicast_lock &base->lock irq_context: 0 (wq_completion)bond104#3 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &idev->mc_lock rcu_read_lock rcu_read_lock rcu_read_lock_bh rcu_read_lock rcu_read_lock_bh rcu_read_lock &data->lock irq_context: 0 (wq_completion)bond104#3 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &idev->mc_lock rcu_read_lock rcu_read_lock rcu_read_lock_bh rcu_read_lock rcu_read_lock_bh rcu_read_lock &obj_hash[i].lock irq_context: 0 (wq_completion)bond127 (work_completion)(&(&slave->notify_work)->work) rtnl_mutex net_rwsem irq_context: 0 (wq_completion)bond127 (work_completion)(&(&slave->notify_work)->work) rtnl_mutex net_rwsem &list->lock#2 irq_context: 0 (wq_completion)bond127 (work_completion)(&(&slave->notify_work)->work) rtnl_mutex &tn->lock irq_context: 0 (wq_completion)bond109#3 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex fs_reclaim irq_context: 0 (wq_completion)bond109#3 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex fs_reclaim mmu_notifier_invalidate_range_start irq_context: 0 (wq_completion)bond109#3 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex rcu_read_lock &bond->stats_lock/1 irq_context: 0 (wq_completion)bond109#3 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex rcu_read_lock &ndev->lock irq_context: 0 (wq_completion)bond109#3 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex nl_table_lock irq_context: 0 (wq_completion)bond109#3 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &obj_hash[i].lock irq_context: 0 (wq_completion)bond109#3 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex nl_table_wait.lock irq_context: 0 (wq_completion)bond109#3 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex net_rwsem irq_context: 0 (wq_completion)bond109#3 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex net_rwsem &list->lock#2 irq_context: 0 (wq_completion)bond109#3 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &tn->lock irq_context: 0 (wq_completion)bond109#3 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &idev->mc_lock irq_context: 0 (wq_completion)bond109#3 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &idev->mc_lock &rq->__lock irq_context: 0 (wq_completion)bond109#3 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &idev->mc_lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)bond109#3 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &idev->mc_lock fs_reclaim irq_context: 0 (wq_completion)bond109#3 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &idev->mc_lock fs_reclaim mmu_notifier_invalidate_range_start irq_context: 0 (wq_completion)bond109#3 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &idev->mc_lock &c->lock irq_context: 0 (wq_completion)bond109#3 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &idev->mc_lock rcu_read_lock &ndev->lock irq_context: 0 (wq_completion)bond109#3 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &idev->mc_lock rcu_read_lock &dir->lock#2 irq_context: 0 (wq_completion)bond109#3 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &idev->mc_lock rcu_read_lock &ul->lock#2 irq_context: 0 (wq_completion)bond109#3 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &idev->mc_lock rcu_read_lock rcu_read_lock &____s->seqcount#10 irq_context: 0 (wq_completion)bond109#3 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &idev->mc_lock rcu_read_lock rcu_read_lock rcu_read_lock_bh rcu_read_lock rcu_read_lock_bh rcu_read_lock &br->multicast_lock irq_context: 0 (wq_completion)bond109#3 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &idev->mc_lock rcu_read_lock rcu_read_lock rcu_read_lock_bh rcu_read_lock rcu_read_lock_bh rcu_read_lock &br->multicast_lock &obj_hash[i].lock irq_context: 0 (wq_completion)bond109#3 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &idev->mc_lock rcu_read_lock rcu_read_lock rcu_read_lock_bh rcu_read_lock rcu_read_lock_bh rcu_read_lock &br->multicast_lock &base->lock irq_context: 0 (wq_completion)bond109#3 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &idev->mc_lock rcu_read_lock rcu_read_lock rcu_read_lock_bh rcu_read_lock rcu_read_lock_bh rcu_read_lock &data->lock irq_context: 0 (wq_completion)bond109#3 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &idev->mc_lock rcu_read_lock rcu_read_lock rcu_read_lock_bh rcu_read_lock rcu_read_lock_bh rcu_read_lock &obj_hash[i].lock irq_context: 0 (wq_completion)bond109#3 (work_completion)(&(&slave->notify_work)->work) rtnl_mutex net_rwsem irq_context: 0 (wq_completion)bond109#3 (work_completion)(&(&slave->notify_work)->work) rtnl_mutex net_rwsem &list->lock#2 irq_context: 0 (wq_completion)bond109#3 (work_completion)(&(&slave->notify_work)->work) rtnl_mutex &tn->lock irq_context: 0 (wq_completion)bond120#2 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex fs_reclaim irq_context: 0 (wq_completion)bond120#2 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex fs_reclaim mmu_notifier_invalidate_range_start irq_context: 0 (wq_completion)bond120#2 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &c->lock irq_context: 0 (wq_completion)bond120#2 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex rcu_read_lock &bond->stats_lock/1 irq_context: 0 (wq_completion)bond120#2 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex rcu_read_lock &ndev->lock irq_context: 0 (wq_completion)bond120#2 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex nl_table_lock irq_context: 0 (wq_completion)bond120#2 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &obj_hash[i].lock irq_context: 0 (wq_completion)bond120#2 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex nl_table_wait.lock irq_context: 0 (wq_completion)bond120#2 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex net_rwsem irq_context: 0 (wq_completion)bond120#2 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex net_rwsem &list->lock#2 irq_context: 0 (wq_completion)bond120#2 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &tn->lock irq_context: 0 (wq_completion)bond120#2 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &idev->mc_lock irq_context: 0 (wq_completion)bond120#2 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &idev->mc_lock fs_reclaim irq_context: 0 (wq_completion)bond120#2 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &idev->mc_lock fs_reclaim mmu_notifier_invalidate_range_start irq_context: 0 (wq_completion)bond120#2 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &idev->mc_lock &c->lock irq_context: 0 (wq_completion)bond120#2 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &idev->mc_lock &n->list_lock irq_context: 0 (wq_completion)bond120#2 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &idev->mc_lock &n->list_lock &c->lock irq_context: 0 (wq_completion)bond120#2 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &idev->mc_lock rcu_read_lock &ndev->lock irq_context: 0 (wq_completion)bond120#2 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &idev->mc_lock rcu_read_lock &dir->lock#2 irq_context: 0 (wq_completion)bond120#2 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &idev->mc_lock rcu_read_lock &ul->lock#2 irq_context: 0 (wq_completion)bond120#2 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &idev->mc_lock rcu_read_lock rcu_read_lock &____s->seqcount#10 irq_context: 0 (wq_completion)bond120#2 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &idev->mc_lock rcu_read_lock rcu_read_lock rcu_read_lock_bh rcu_read_lock rcu_read_lock_bh rcu_read_lock &br->multicast_lock irq_context: 0 (wq_completion)bond120#2 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &idev->mc_lock rcu_read_lock rcu_read_lock rcu_read_lock_bh rcu_read_lock rcu_read_lock_bh rcu_read_lock &br->multicast_lock &obj_hash[i].lock irq_context: 0 (wq_completion)bond120#2 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &idev->mc_lock rcu_read_lock rcu_read_lock rcu_read_lock_bh rcu_read_lock rcu_read_lock_bh rcu_read_lock &br->multicast_lock &base->lock irq_context: 0 (wq_completion)bond120#2 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &idev->mc_lock rcu_read_lock rcu_read_lock rcu_read_lock_bh rcu_read_lock rcu_read_lock_bh rcu_read_lock &data->lock irq_context: 0 (wq_completion)bond120#2 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &idev->mc_lock rcu_read_lock rcu_read_lock rcu_read_lock_bh rcu_read_lock rcu_read_lock_bh rcu_read_lock &obj_hash[i].lock irq_context: 0 (wq_completion)bond110#2 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex fs_reclaim irq_context: 0 (wq_completion)bond110#2 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex fs_reclaim mmu_notifier_invalidate_range_start irq_context: 0 (wq_completion)bond110#2 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex rcu_read_lock &bond->stats_lock/1 irq_context: 0 (wq_completion)bond110#2 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex rcu_read_lock &ndev->lock irq_context: 0 (wq_completion)bond110#2 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex nl_table_lock irq_context: 0 (wq_completion)bond110#2 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &obj_hash[i].lock irq_context: 0 (wq_completion)bond110#2 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex nl_table_wait.lock irq_context: 0 (wq_completion)bond110#2 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex net_rwsem irq_context: 0 (wq_completion)bond110#2 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex net_rwsem &list->lock#2 irq_context: 0 (wq_completion)bond110#2 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &tn->lock irq_context: 0 (wq_completion)bond110#2 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &idev->mc_lock irq_context: 0 (wq_completion)bond110#2 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &idev->mc_lock fs_reclaim irq_context: 0 (wq_completion)bond110#2 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &idev->mc_lock fs_reclaim mmu_notifier_invalidate_range_start irq_context: 0 (wq_completion)bond110#2 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &idev->mc_lock &c->lock irq_context: 0 (wq_completion)bond110#2 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &idev->mc_lock rcu_read_lock &ndev->lock irq_context: 0 (wq_completion)bond110#2 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &idev->mc_lock rcu_read_lock &c->lock irq_context: 0 (wq_completion)bond110#2 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &idev->mc_lock rcu_read_lock &____s->seqcount#2 irq_context: 0 (wq_completion)bond110#2 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &idev->mc_lock rcu_read_lock &____s->seqcount irq_context: 0 (wq_completion)bond110#2 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &idev->mc_lock rcu_read_lock &dir->lock#2 irq_context: 0 (wq_completion)bond110#2 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &idev->mc_lock rcu_read_lock &ul->lock#2 irq_context: 0 (wq_completion)bond110#2 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &idev->mc_lock rcu_read_lock rcu_read_lock &____s->seqcount#10 irq_context: 0 (wq_completion)bond110#2 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &idev->mc_lock rcu_read_lock rcu_read_lock rcu_read_lock_bh rcu_read_lock rcu_read_lock_bh rcu_read_lock &br->multicast_lock irq_context: 0 (wq_completion)bond110#2 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &idev->mc_lock rcu_read_lock rcu_read_lock rcu_read_lock_bh rcu_read_lock rcu_read_lock_bh rcu_read_lock &br->multicast_lock &obj_hash[i].lock irq_context: 0 (wq_completion)bond110#2 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &idev->mc_lock rcu_read_lock rcu_read_lock rcu_read_lock_bh rcu_read_lock rcu_read_lock_bh rcu_read_lock &br->multicast_lock &base->lock irq_context: 0 (wq_completion)bond110#2 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &idev->mc_lock rcu_read_lock rcu_read_lock rcu_read_lock_bh rcu_read_lock rcu_read_lock_bh rcu_read_lock &br->multicast_lock &base->lock &obj_hash[i].lock irq_context: 0 (wq_completion)bond110#2 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &idev->mc_lock rcu_read_lock rcu_read_lock rcu_read_lock_bh rcu_read_lock rcu_read_lock_bh rcu_read_lock &data->lock irq_context: 0 (wq_completion)bond110#2 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &idev->mc_lock rcu_read_lock rcu_read_lock rcu_read_lock_bh rcu_read_lock rcu_read_lock_bh rcu_read_lock &obj_hash[i].lock irq_context: 0 (wq_completion)bond137 (work_completion)(&(&bond->mcast_work)->work) &rq->__lock irq_context: 0 (wq_completion)bond137 (work_completion)(&(&bond->mcast_work)->work) &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)wg-kex-wg2#9 (work_completion)(&peer->transmit_handshake_work) &peer->endpoint_lock rcu_read_lock_bh &obj_hash[i].lock pool_lock irq_context: 0 rtnl_mutex &net->ipv6.fib6_gc_lock fill_pool_map-wait-type-override pool_lock#2 irq_context: 0 rtnl_mutex &net->ipv6.fib6_gc_lock fill_pool_map-wait-type-override pool_lock irq_context: 0 (wq_completion)bond112#3 irq_context: 0 (wq_completion)bond112#3 &rq->__lock irq_context: 0 (wq_completion)bond112#3 &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)bond112#3 (work_completion)(&(&slave->notify_work)->work) irq_context: 0 (wq_completion)bond112#3 (work_completion)(&(&slave->notify_work)->work) &obj_hash[i].lock irq_context: 0 (wq_completion)bond112#3 (work_completion)(&(&slave->notify_work)->work) &base->lock irq_context: 0 (wq_completion)bond112#3 (work_completion)(&(&slave->notify_work)->work) &base->lock &obj_hash[i].lock irq_context: 0 (wq_completion)bond112#3 (work_completion)(&(&slave->notify_work)->work) &rq->__lock irq_context: 0 (wq_completion)bond112#3 (work_completion)(&(&bond->mcast_work)->work) irq_context: 0 (wq_completion)bond112#3 (work_completion)(&(&bond->mcast_work)->work) &obj_hash[i].lock irq_context: 0 (wq_completion)bond112#3 (work_completion)(&(&bond->mcast_work)->work) &base->lock irq_context: 0 (wq_completion)bond112#3 (work_completion)(&(&bond->mcast_work)->work) &base->lock &obj_hash[i].lock irq_context: 0 (wq_completion)bond112#3 (work_completion)(&(&bond->mcast_work)->work) &rq->__lock irq_context: 0 (wq_completion)bond112#3 (work_completion)(&(&bond->mcast_work)->work) &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)bond24#3 &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)bond128#2 irq_context: 0 (wq_completion)bond128#2 (work_completion)(&(&slave->notify_work)->work) irq_context: 0 (wq_completion)bond128#2 (work_completion)(&(&slave->notify_work)->work) &obj_hash[i].lock irq_context: 0 (wq_completion)bond128#2 (work_completion)(&(&slave->notify_work)->work) &base->lock irq_context: 0 (wq_completion)bond128#2 (work_completion)(&(&slave->notify_work)->work) &base->lock &obj_hash[i].lock irq_context: 0 (wq_completion)bond128#2 (work_completion)(&(&slave->notify_work)->work) &rq->__lock irq_context: 0 (wq_completion)bond128#2 (work_completion)(&(&slave->notify_work)->work) &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)bond128#2 (work_completion)(&(&bond->mcast_work)->work) irq_context: 0 (wq_completion)bond128#2 (work_completion)(&(&bond->mcast_work)->work) &obj_hash[i].lock irq_context: 0 (wq_completion)bond128#2 (work_completion)(&(&bond->mcast_work)->work) &base->lock irq_context: 0 (wq_completion)bond128#2 (work_completion)(&(&bond->mcast_work)->work) &base->lock &obj_hash[i].lock irq_context: 0 (wq_completion)bond30#4 irq_context: 0 (wq_completion)bond30#4 (work_completion)(&(&slave->notify_work)->work) irq_context: 0 (wq_completion)bond30#4 (work_completion)(&(&slave->notify_work)->work) &obj_hash[i].lock irq_context: 0 (wq_completion)bond30#4 (work_completion)(&(&slave->notify_work)->work) &base->lock irq_context: 0 (wq_completion)bond30#4 (work_completion)(&(&slave->notify_work)->work) &base->lock &obj_hash[i].lock irq_context: 0 (wq_completion)bond30#4 (work_completion)(&(&bond->mcast_work)->work) irq_context: 0 (wq_completion)bond30#4 (work_completion)(&(&bond->mcast_work)->work) &obj_hash[i].lock irq_context: 0 (wq_completion)bond30#4 (work_completion)(&(&bond->mcast_work)->work) &base->lock irq_context: 0 (wq_completion)bond30#4 (work_completion)(&(&bond->mcast_work)->work) &base->lock &obj_hash[i].lock irq_context: 0 (wq_completion)bond141 irq_context: 0 (wq_completion)bond141 (work_completion)(&(&slave->notify_work)->work) irq_context: 0 (wq_completion)bond141 (work_completion)(&(&slave->notify_work)->work) &obj_hash[i].lock irq_context: 0 (wq_completion)bond141 (work_completion)(&(&slave->notify_work)->work) &base->lock irq_context: 0 (wq_completion)bond141 (work_completion)(&(&slave->notify_work)->work) &base->lock &obj_hash[i].lock irq_context: 0 (wq_completion)bond141 (work_completion)(&(&slave->notify_work)->work) &rq->__lock irq_context: 0 (wq_completion)bond141 (work_completion)(&(&slave->notify_work)->work) &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)bond141 (work_completion)(&(&bond->mcast_work)->work) irq_context: 0 (wq_completion)bond126 (work_completion)(&(&slave->notify_work)->work) rtnl_mutex net_rwsem irq_context: 0 (wq_completion)bond141 (work_completion)(&(&bond->mcast_work)->work) &obj_hash[i].lock irq_context: 0 (wq_completion)bond126 (work_completion)(&(&slave->notify_work)->work) rtnl_mutex net_rwsem &list->lock#2 irq_context: 0 (wq_completion)bond141 (work_completion)(&(&bond->mcast_work)->work) &base->lock irq_context: 0 (wq_completion)bond126 (work_completion)(&(&slave->notify_work)->work) rtnl_mutex &tn->lock irq_context: 0 (wq_completion)bond141 (work_completion)(&(&bond->mcast_work)->work) &base->lock &obj_hash[i].lock irq_context: 0 (wq_completion)bond88#2 (work_completion)(&(&slave->notify_work)->work) rtnl_mutex net_rwsem irq_context: 0 (wq_completion)bond141 (work_completion)(&(&bond->mcast_work)->work) &rq->__lock irq_context: 0 (wq_completion)bond88#2 (work_completion)(&(&slave->notify_work)->work) rtnl_mutex net_rwsem &list->lock#2 irq_context: 0 (wq_completion)bond141 (work_completion)(&(&bond->mcast_work)->work) &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)bond88#2 (work_completion)(&(&slave->notify_work)->work) rtnl_mutex &tn->lock irq_context: 0 (wq_completion)bond115#2 (work_completion)(&(&slave->notify_work)->work) rtnl_mutex net_rwsem irq_context: 0 (wq_completion)bond115#2 (work_completion)(&(&slave->notify_work)->work) rtnl_mutex net_rwsem &list->lock#2 irq_context: 0 (wq_completion)bond115#2 (work_completion)(&(&slave->notify_work)->work) rtnl_mutex &tn->lock irq_context: 0 (wq_completion)bond109#2 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex fs_reclaim irq_context: 0 (wq_completion)bond109#2 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex fs_reclaim mmu_notifier_invalidate_range_start irq_context: 0 (wq_completion)bond109#2 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &c->lock irq_context: 0 (wq_completion)bond109#2 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex rcu_read_lock &bond->stats_lock/1 irq_context: 0 (wq_completion)bond109#2 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex rcu_read_lock &ndev->lock irq_context: 0 (wq_completion)bond109#2 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex nl_table_lock irq_context: 0 (wq_completion)bond109#2 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &obj_hash[i].lock irq_context: 0 (wq_completion)bond109#2 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex nl_table_wait.lock irq_context: 0 (wq_completion)bond109#2 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex net_rwsem irq_context: 0 (wq_completion)bond109#2 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex net_rwsem &list->lock#2 irq_context: 0 (wq_completion)bond109#2 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &tn->lock irq_context: 0 (wq_completion)bond109#2 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &idev->mc_lock irq_context: 0 (wq_completion)bond109#2 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &idev->mc_lock fs_reclaim irq_context: 0 (wq_completion)bond109#2 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &idev->mc_lock fs_reclaim mmu_notifier_invalidate_range_start irq_context: 0 (wq_completion)bond109#2 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &idev->mc_lock &c->lock irq_context: 0 (wq_completion)bond109#2 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &idev->mc_lock rcu_read_lock &ndev->lock irq_context: 0 (wq_completion)bond109#2 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &idev->mc_lock rcu_read_lock &c->lock irq_context: 0 (wq_completion)bond109#2 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &idev->mc_lock rcu_read_lock &____s->seqcount#2 irq_context: 0 (wq_completion)bond109#2 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &idev->mc_lock rcu_read_lock &____s->seqcount irq_context: 0 (wq_completion)bond109#2 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &idev->mc_lock rcu_read_lock &dir->lock#2 irq_context: 0 (wq_completion)bond109#2 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &idev->mc_lock rcu_read_lock &ul->lock#2 irq_context: 0 (wq_completion)bond109#2 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &idev->mc_lock rcu_read_lock rcu_read_lock &____s->seqcount#10 irq_context: 0 (wq_completion)bond109#2 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &idev->mc_lock rcu_read_lock rcu_read_lock rcu_read_lock_bh rcu_read_lock rcu_read_lock_bh rcu_read_lock &br->multicast_lock irq_context: 0 (wq_completion)bond109#2 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &idev->mc_lock rcu_read_lock rcu_read_lock rcu_read_lock_bh rcu_read_lock rcu_read_lock_bh rcu_read_lock &br->multicast_lock &obj_hash[i].lock irq_context: 0 (wq_completion)bond109#2 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &idev->mc_lock rcu_read_lock rcu_read_lock rcu_read_lock_bh rcu_read_lock rcu_read_lock_bh rcu_read_lock &br->multicast_lock &base->lock irq_context: 0 (wq_completion)bond109#2 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &idev->mc_lock rcu_read_lock rcu_read_lock rcu_read_lock_bh rcu_read_lock rcu_read_lock_bh rcu_read_lock &br->multicast_lock &base->lock &obj_hash[i].lock irq_context: 0 (wq_completion)bond109#2 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &idev->mc_lock rcu_read_lock rcu_read_lock rcu_read_lock_bh rcu_read_lock rcu_read_lock_bh rcu_read_lock &data->lock irq_context: 0 (wq_completion)bond109#2 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &idev->mc_lock rcu_read_lock rcu_read_lock rcu_read_lock_bh rcu_read_lock rcu_read_lock_bh rcu_read_lock &obj_hash[i].lock irq_context: 0 (wq_completion)bond141 (work_completion)(&(&slave->notify_work)->work) rtnl_mutex net_rwsem irq_context: 0 (wq_completion)bond141 (work_completion)(&(&slave->notify_work)->work) rtnl_mutex net_rwsem &list->lock#2 irq_context: 0 (wq_completion)bond141 (work_completion)(&(&slave->notify_work)->work) rtnl_mutex &tn->lock irq_context: 0 (wq_completion)bond30#4 (work_completion)(&(&slave->notify_work)->work) rtnl_mutex net_rwsem irq_context: 0 (wq_completion)bond30#4 (work_completion)(&(&slave->notify_work)->work) rtnl_mutex net_rwsem &list->lock#2 irq_context: 0 (wq_completion)bond30#4 (work_completion)(&(&slave->notify_work)->work) rtnl_mutex &tn->lock irq_context: 0 (wq_completion)bond112#3 (work_completion)(&(&slave->notify_work)->work) rtnl_mutex net_rwsem irq_context: 0 (wq_completion)bond112#3 (work_completion)(&(&slave->notify_work)->work) rtnl_mutex net_rwsem &list->lock#2 irq_context: 0 (wq_completion)bond112#3 (work_completion)(&(&slave->notify_work)->work) rtnl_mutex &tn->lock irq_context: 0 (wq_completion)bond98#3 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex fs_reclaim irq_context: 0 (wq_completion)bond98#3 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex fs_reclaim mmu_notifier_invalidate_range_start irq_context: 0 (wq_completion)bond98#3 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex rcu_read_lock &bond->stats_lock/1 irq_context: 0 (wq_completion)bond98#3 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex rcu_read_lock &ndev->lock irq_context: 0 (wq_completion)bond98#3 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex rcu_read_lock rcu_node_0 irq_context: 0 (wq_completion)bond98#3 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex rcu_read_lock &rq->__lock irq_context: 0 (wq_completion)bond98#3 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex rcu_read_lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)bond98#3 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex nl_table_lock irq_context: 0 (wq_completion)bond98#3 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &obj_hash[i].lock irq_context: 0 (wq_completion)bond98#3 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex nl_table_wait.lock irq_context: 0 (wq_completion)bond98#3 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex net_rwsem irq_context: 0 (wq_completion)bond98#3 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex net_rwsem &list->lock#2 irq_context: 0 (wq_completion)bond98#3 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &tn->lock irq_context: 0 (wq_completion)bond98#3 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &idev->mc_lock irq_context: 0 (wq_completion)bond98#3 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &idev->mc_lock fs_reclaim irq_context: 0 (wq_completion)bond98#3 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &idev->mc_lock fs_reclaim mmu_notifier_invalidate_range_start irq_context: 0 (wq_completion)bond98#3 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &idev->mc_lock &c->lock irq_context: 0 (wq_completion)bond98#3 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &idev->mc_lock rcu_read_lock &ndev->lock irq_context: 0 (wq_completion)bond98#3 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &idev->mc_lock rcu_read_lock &c->lock irq_context: 0 (wq_completion)bond98#3 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &idev->mc_lock rcu_read_lock &dir->lock#2 irq_context: 0 (wq_completion)bond98#3 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &idev->mc_lock rcu_read_lock rcu_node_0 irq_context: 0 (wq_completion)bond98#3 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &idev->mc_lock rcu_read_lock &rq->__lock irq_context: 0 (wq_completion)bond98#3 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &idev->mc_lock rcu_read_lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)bond98#3 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &idev->mc_lock rcu_read_lock &ul->lock#2 irq_context: 0 (wq_completion)bond98#3 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &idev->mc_lock rcu_read_lock rcu_read_lock &____s->seqcount#10 irq_context: 0 (wq_completion)bond98#3 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &idev->mc_lock rcu_read_lock rcu_read_lock rcu_read_lock_bh rcu_read_lock rcu_read_lock_bh rcu_read_lock &br->multicast_lock irq_context: 0 (wq_completion)bond98#3 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &idev->mc_lock rcu_read_lock rcu_read_lock rcu_read_lock_bh rcu_read_lock rcu_read_lock_bh rcu_read_lock &br->multicast_lock &obj_hash[i].lock irq_context: 0 (wq_completion)bond98#3 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &idev->mc_lock rcu_read_lock rcu_read_lock rcu_read_lock_bh rcu_read_lock rcu_read_lock_bh rcu_read_lock &br->multicast_lock &base->lock irq_context: 0 (wq_completion)bond98#3 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &idev->mc_lock rcu_read_lock rcu_read_lock rcu_read_lock_bh rcu_read_lock rcu_read_lock_bh rcu_read_lock &data->lock irq_context: 0 (wq_completion)bond98#3 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &idev->mc_lock rcu_read_lock rcu_read_lock rcu_read_lock_bh rcu_read_lock rcu_read_lock_bh rcu_read_lock &obj_hash[i].lock irq_context: 0 (wq_completion)bond98#3 (work_completion)(&(&slave->notify_work)->work) rtnl_mutex net_rwsem irq_context: 0 (wq_completion)bond98#3 (work_completion)(&(&slave->notify_work)->work) rtnl_mutex net_rwsem &list->lock#2 irq_context: 0 (wq_completion)bond98#3 (work_completion)(&(&slave->notify_work)->work) rtnl_mutex &tn->lock irq_context: 0 (wq_completion)bond98#3 (work_completion)(&(&slave->notify_work)->work) rtnl_mutex rcu_read_lock rcu_node_0 irq_context: 0 (wq_completion)bond98#3 (work_completion)(&(&slave->notify_work)->work) rtnl_mutex rcu_read_lock &rq->__lock irq_context: 0 (wq_completion)bond98#3 (work_completion)(&(&slave->notify_work)->work) rtnl_mutex rcu_read_lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)bond128#2 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex fs_reclaim irq_context: 0 (wq_completion)bond128#2 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex fs_reclaim mmu_notifier_invalidate_range_start irq_context: 0 (wq_completion)bond128#2 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex rcu_read_lock &bond->stats_lock/1 irq_context: 0 (wq_completion)bond128#2 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex rcu_read_lock &ndev->lock irq_context: 0 (wq_completion)bond128#2 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex rcu_read_lock rcu_node_0 irq_context: 0 (wq_completion)bond128#2 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex rcu_read_lock &rq->__lock irq_context: 0 (wq_completion)bond128#2 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex nl_table_lock irq_context: 0 (wq_completion)bond128#2 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &obj_hash[i].lock irq_context: 0 (wq_completion)bond128#2 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex nl_table_wait.lock irq_context: 0 (wq_completion)bond128#2 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex net_rwsem irq_context: 0 (wq_completion)bond128#2 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex net_rwsem &list->lock#2 irq_context: 0 (wq_completion)bond128#2 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &tn->lock irq_context: 0 (wq_completion)bond128#2 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &idev->mc_lock irq_context: 0 (wq_completion)bond128#2 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &idev->mc_lock fs_reclaim irq_context: 0 (wq_completion)bond128#2 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &idev->mc_lock fs_reclaim mmu_notifier_invalidate_range_start irq_context: 0 (wq_completion)bond128#2 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &idev->mc_lock rcu_read_lock &ndev->lock irq_context: 0 (wq_completion)bond128#2 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &idev->mc_lock rcu_read_lock &c->lock irq_context: 0 (wq_completion)bond128#2 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &idev->mc_lock rcu_read_lock &dir->lock#2 irq_context: 0 (wq_completion)bond128#2 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &idev->mc_lock rcu_read_lock &ul->lock#2 irq_context: 0 (wq_completion)bond128#2 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &idev->mc_lock rcu_read_lock rcu_node_0 irq_context: 0 (wq_completion)bond128#2 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &idev->mc_lock rcu_read_lock &rq->__lock irq_context: 0 (wq_completion)bond128#2 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &idev->mc_lock rcu_read_lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)bond128#2 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &idev->mc_lock rcu_read_lock rcu_read_lock &____s->seqcount#10 irq_context: 0 (wq_completion)bond128#2 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &idev->mc_lock rcu_read_lock rcu_read_lock rcu_read_lock_bh rcu_read_lock rcu_read_lock_bh rcu_read_lock &br->multicast_lock irq_context: 0 (wq_completion)bond128#2 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &idev->mc_lock rcu_read_lock rcu_read_lock rcu_read_lock_bh rcu_read_lock rcu_read_lock_bh rcu_read_lock &br->multicast_lock &obj_hash[i].lock irq_context: 0 (wq_completion)bond128#2 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &idev->mc_lock rcu_read_lock rcu_read_lock rcu_read_lock_bh rcu_read_lock rcu_read_lock_bh rcu_read_lock &data->lock irq_context: 0 (wq_completion)bond128#2 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &idev->mc_lock rcu_read_lock rcu_read_lock rcu_read_lock_bh rcu_read_lock rcu_read_lock_bh rcu_read_lock &obj_hash[i].lock irq_context: 0 (wq_completion)bond124#2 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex fs_reclaim irq_context: 0 (wq_completion)bond124#2 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex fs_reclaim &rq->__lock irq_context: 0 (wq_completion)bond124#2 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex fs_reclaim mmu_notifier_invalidate_range_start irq_context: 0 (wq_completion)bond124#2 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex rcu_read_lock &bond->stats_lock/1 irq_context: 0 (wq_completion)bond124#2 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex rcu_read_lock &ndev->lock irq_context: 0 (wq_completion)bond124#2 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex nl_table_lock irq_context: 0 (wq_completion)bond124#2 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &obj_hash[i].lock irq_context: 0 (wq_completion)bond124#2 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex nl_table_wait.lock irq_context: 0 (wq_completion)bond124#2 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex net_rwsem irq_context: 0 (wq_completion)bond124#2 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex net_rwsem &list->lock#2 irq_context: 0 (wq_completion)bond124#2 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &tn->lock irq_context: 0 (wq_completion)bond124#2 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &idev->mc_lock irq_context: 0 (wq_completion)bond124#2 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &idev->mc_lock fs_reclaim irq_context: 0 (wq_completion)bond124#2 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &idev->mc_lock fs_reclaim mmu_notifier_invalidate_range_start irq_context: 0 (wq_completion)bond124#2 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &idev->mc_lock fs_reclaim mmu_notifier_invalidate_range_start &rq->__lock irq_context: 0 (wq_completion)bond124#2 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &idev->mc_lock fs_reclaim mmu_notifier_invalidate_range_start &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)bond124#2 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &idev->mc_lock fs_reclaim mmu_notifier_invalidate_range_start &cfs_rq->removed.lock irq_context: 0 (wq_completion)bond124#2 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &idev->mc_lock fs_reclaim mmu_notifier_invalidate_range_start &obj_hash[i].lock irq_context: 0 (wq_completion)bond124#2 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &idev->mc_lock rcu_read_lock &ndev->lock irq_context: 0 (wq_completion)bond124#2 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &idev->mc_lock rcu_read_lock &c->lock irq_context: 0 (wq_completion)bond124#2 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &idev->mc_lock rcu_read_lock &dir->lock#2 irq_context: 0 (wq_completion)bond124#2 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &idev->mc_lock rcu_read_lock &ul->lock#2 irq_context: 0 (wq_completion)bond124#2 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &idev->mc_lock rcu_read_lock rcu_node_0 irq_context: 0 (wq_completion)bond124#2 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &idev->mc_lock rcu_read_lock &rq->__lock irq_context: 0 (wq_completion)bond124#2 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &idev->mc_lock rcu_read_lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)bond124#2 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &idev->mc_lock rcu_read_lock rcu_read_lock &____s->seqcount#10 irq_context: 0 (wq_completion)bond124#2 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &idev->mc_lock rcu_read_lock rcu_read_lock rcu_read_lock_bh rcu_read_lock rcu_read_lock_bh rcu_read_lock &br->multicast_lock irq_context: 0 (wq_completion)bond124#2 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &idev->mc_lock rcu_read_lock rcu_read_lock rcu_read_lock_bh rcu_read_lock rcu_read_lock_bh rcu_read_lock &br->multicast_lock &obj_hash[i].lock irq_context: 0 (wq_completion)bond124#2 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &idev->mc_lock rcu_read_lock rcu_read_lock rcu_read_lock_bh rcu_read_lock rcu_read_lock_bh rcu_read_lock &br->multicast_lock &base->lock irq_context: 0 (wq_completion)bond124#2 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &idev->mc_lock rcu_read_lock rcu_read_lock rcu_read_lock_bh rcu_read_lock rcu_read_lock_bh rcu_read_lock &data->lock irq_context: 0 (wq_completion)bond124#2 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &idev->mc_lock rcu_read_lock rcu_read_lock rcu_read_lock_bh rcu_read_lock rcu_read_lock_bh rcu_read_lock &obj_hash[i].lock irq_context: 0 (wq_completion)bond124#2 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &idev->mc_lock rcu_read_lock &rcu_state.gp_wq irq_context: 0 (wq_completion)bond124#2 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &idev->mc_lock rcu_read_lock &rcu_state.gp_wq &p->pi_lock irq_context: 0 (wq_completion)bond124#2 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &idev->mc_lock rcu_read_lock &rcu_state.gp_wq &p->pi_lock &rq->__lock irq_context: 0 (wq_completion)bond124#2 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &idev->mc_lock rcu_read_lock &rcu_state.gp_wq &p->pi_lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)bond124#2 (work_completion)(&(&slave->notify_work)->work) rtnl_mutex net_rwsem irq_context: 0 (wq_completion)bond124#2 (work_completion)(&(&slave->notify_work)->work) rtnl_mutex net_rwsem &list->lock#2 irq_context: 0 (wq_completion)bond124#2 (work_completion)(&(&slave->notify_work)->work) rtnl_mutex &tn->lock irq_context: 0 (wq_completion)bond139 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex fs_reclaim irq_context: 0 (wq_completion)bond139 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex fs_reclaim &rq->__lock irq_context: 0 (wq_completion)bond139 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex fs_reclaim mmu_notifier_invalidate_range_start irq_context: 0 (wq_completion)bond139 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex batched_entropy_u8.lock irq_context: 0 (wq_completion)bond139 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex kfence_freelist_lock irq_context: 0 (wq_completion)bond139 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &c->lock irq_context: 0 (wq_completion)bond139 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &n->list_lock irq_context: 0 (wq_completion)bond139 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &n->list_lock &c->lock irq_context: 0 (wq_completion)bond139 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex rcu_read_lock &bond->stats_lock/1 irq_context: 0 (wq_completion)bond139 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex rcu_read_lock &ndev->lock irq_context: 0 (wq_completion)bond139 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex nl_table_lock irq_context: 0 (wq_completion)bond139 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &obj_hash[i].lock irq_context: 0 (wq_completion)bond139 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &meta->lock irq_context: 0 (wq_completion)bond139 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex nl_table_wait.lock irq_context: 0 (wq_completion)bond139 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex net_rwsem irq_context: 0 (wq_completion)bond139 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex net_rwsem &list->lock#2 irq_context: 0 (wq_completion)bond139 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &tn->lock irq_context: 0 (wq_completion)bond139 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &idev->mc_lock irq_context: 0 (wq_completion)bond139 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &idev->mc_lock fs_reclaim irq_context: 0 (wq_completion)bond139 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &idev->mc_lock fs_reclaim mmu_notifier_invalidate_range_start irq_context: 0 (wq_completion)bond139 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &idev->mc_lock rcu_read_lock &ndev->lock irq_context: 0 (wq_completion)bond139 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &idev->mc_lock rcu_read_lock &c->lock irq_context: 0 (wq_completion)bond139 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &idev->mc_lock rcu_read_lock &____s->seqcount#2 irq_context: 0 (wq_completion)bond139 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &idev->mc_lock rcu_read_lock &____s->seqcount irq_context: 0 (wq_completion)bond139 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &idev->mc_lock rcu_read_lock pool_lock#2 irq_context: 0 (wq_completion)bond139 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &idev->mc_lock rcu_read_lock &dir->lock#2 irq_context: 0 (wq_completion)bond139 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &idev->mc_lock rcu_read_lock &ul->lock#2 irq_context: 0 (wq_completion)bond139 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &idev->mc_lock rcu_read_lock rcu_read_lock &____s->seqcount#10 irq_context: 0 (wq_completion)bond139 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &idev->mc_lock rcu_read_lock rcu_read_lock rcu_read_lock_bh rcu_read_lock rcu_read_lock_bh rcu_read_lock &br->multicast_lock irq_context: 0 (wq_completion)bond139 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &idev->mc_lock rcu_read_lock rcu_read_lock rcu_read_lock_bh rcu_read_lock rcu_read_lock_bh rcu_read_lock &br->multicast_lock &obj_hash[i].lock irq_context: 0 (wq_completion)bond139 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &idev->mc_lock rcu_read_lock rcu_read_lock rcu_read_lock_bh rcu_read_lock rcu_read_lock_bh rcu_read_lock &br->multicast_lock &base->lock irq_context: 0 (wq_completion)bond139 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &idev->mc_lock rcu_read_lock rcu_read_lock rcu_read_lock_bh rcu_read_lock rcu_read_lock_bh rcu_read_lock &data->lock irq_context: 0 (wq_completion)bond139 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &idev->mc_lock rcu_read_lock rcu_read_lock rcu_read_lock_bh rcu_read_lock rcu_read_lock_bh rcu_read_lock &obj_hash[i].lock irq_context: 0 (wq_completion)bond139 (work_completion)(&(&slave->notify_work)->work) rtnl_mutex net_rwsem irq_context: 0 (wq_completion)bond139 (work_completion)(&(&slave->notify_work)->work) rtnl_mutex net_rwsem &list->lock#2 irq_context: 0 (wq_completion)bond139 (work_completion)(&(&slave->notify_work)->work) rtnl_mutex &tn->lock irq_context: 0 (wq_completion)bond139 (work_completion)(&(&slave->notify_work)->work) rtnl_mutex rcu_read_lock rcu_node_0 irq_context: 0 (wq_completion)bond139 (work_completion)(&(&slave->notify_work)->work) rtnl_mutex rcu_read_lock &rq->__lock irq_context: 0 (wq_completion)bond129 (work_completion)(&(&slave->notify_work)->work) rtnl_mutex net_rwsem irq_context: 0 (wq_completion)bond129 (work_completion)(&(&slave->notify_work)->work) rtnl_mutex net_rwsem &list->lock#2 irq_context: 0 (wq_completion)bond129 (work_completion)(&(&slave->notify_work)->work) rtnl_mutex &tn->lock irq_context: 0 (wq_completion)bond23#3 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex fs_reclaim irq_context: 0 (wq_completion)bond23#3 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex fs_reclaim mmu_notifier_invalidate_range_start irq_context: 0 (wq_completion)bond23#3 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex rcu_read_lock &bond->stats_lock/1 irq_context: 0 (wq_completion)bond23#3 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex rcu_read_lock &ndev->lock irq_context: 0 (wq_completion)bond23#3 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex nl_table_lock irq_context: 0 (wq_completion)bond23#3 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &obj_hash[i].lock irq_context: 0 (wq_completion)bond23#3 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex nl_table_wait.lock irq_context: 0 (wq_completion)bond23#3 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex net_rwsem irq_context: 0 (wq_completion)bond23#3 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex net_rwsem &list->lock#2 irq_context: 0 (wq_completion)bond23#3 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &tn->lock irq_context: 0 (wq_completion)bond23#3 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &idev->mc_lock irq_context: 0 (wq_completion)bond23#3 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &idev->mc_lock fs_reclaim irq_context: 0 (wq_completion)bond23#3 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &idev->mc_lock fs_reclaim mmu_notifier_invalidate_range_start irq_context: 0 (wq_completion)bond23#3 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &idev->mc_lock &c->lock irq_context: 0 (wq_completion)bond23#3 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &idev->mc_lock &n->list_lock irq_context: 0 (wq_completion)bond23#3 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &idev->mc_lock &n->list_lock &c->lock irq_context: 0 (wq_completion)bond23#3 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &idev->mc_lock rcu_read_lock &ndev->lock irq_context: 0 (wq_completion)bond23#3 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &idev->mc_lock rcu_read_lock &dir->lock#2 irq_context: 0 (wq_completion)bond23#3 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &idev->mc_lock rcu_read_lock &ul->lock#2 irq_context: 0 (wq_completion)bond23#3 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &idev->mc_lock rcu_read_lock rcu_read_lock &____s->seqcount#12 irq_context: 0 (wq_completion)bond23#3 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &idev->mc_lock rcu_read_lock rcu_read_lock &____s->seqcount#16 irq_context: 0 (wq_completion)bond23#3 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &idev->mc_lock rcu_read_lock &obj_hash[i].lock irq_context: 0 (wq_completion)bond23#3 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &idev->mc_lock rcu_read_lock &data->lock irq_context: 0 (wq_completion)bond107#3 (work_completion)(&(&slave->notify_work)->work) rtnl_mutex net_rwsem irq_context: 0 (wq_completion)bond107#3 (work_completion)(&(&slave->notify_work)->work) rtnl_mutex net_rwsem &list->lock#2 irq_context: 0 (wq_completion)bond107#3 (work_completion)(&(&slave->notify_work)->work) rtnl_mutex &tn->lock irq_context: 0 (wq_completion)bond122#2 (work_completion)(&(&slave->notify_work)->work) rtnl_mutex net_rwsem irq_context: 0 (wq_completion)bond122#2 (work_completion)(&(&slave->notify_work)->work) rtnl_mutex net_rwsem &list->lock#2 irq_context: 0 (wq_completion)bond122#2 (work_completion)(&(&slave->notify_work)->work) rtnl_mutex &tn->lock irq_context: 0 (wq_completion)bond112#3 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex fs_reclaim irq_context: 0 (wq_completion)bond112#3 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex fs_reclaim mmu_notifier_invalidate_range_start irq_context: 0 (wq_completion)bond112#3 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &c->lock irq_context: 0 (wq_completion)bond112#3 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex rcu_read_lock &bond->stats_lock/1 irq_context: 0 (wq_completion)bond112#3 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex rcu_read_lock &ndev->lock irq_context: 0 (wq_completion)bond112#3 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex nl_table_lock irq_context: 0 (wq_completion)bond112#3 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &obj_hash[i].lock irq_context: 0 (wq_completion)bond112#3 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex nl_table_wait.lock irq_context: 0 (wq_completion)bond112#3 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex net_rwsem irq_context: 0 (wq_completion)bond112#3 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex net_rwsem &list->lock#2 irq_context: 0 (wq_completion)bond112#3 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &tn->lock irq_context: 0 (wq_completion)bond112#3 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &idev->mc_lock irq_context: 0 (wq_completion)bond112#3 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &idev->mc_lock fs_reclaim irq_context: 0 (wq_completion)bond112#3 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &idev->mc_lock fs_reclaim mmu_notifier_invalidate_range_start irq_context: 0 (wq_completion)bond112#3 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &idev->mc_lock &c->lock irq_context: 0 (wq_completion)bond112#3 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &idev->mc_lock rcu_read_lock &ndev->lock irq_context: 0 (wq_completion)bond112#3 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &idev->mc_lock rcu_read_lock &dir->lock#2 irq_context: 0 (wq_completion)bond112#3 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &idev->mc_lock rcu_read_lock &ul->lock#2 irq_context: 0 (wq_completion)bond112#3 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &idev->mc_lock rcu_read_lock rcu_read_lock &____s->seqcount#10 irq_context: 0 (wq_completion)bond112#3 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &idev->mc_lock rcu_read_lock rcu_read_lock rcu_read_lock_bh rcu_read_lock rcu_read_lock_bh rcu_read_lock &br->multicast_lock irq_context: 0 (wq_completion)bond112#3 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &idev->mc_lock rcu_read_lock rcu_read_lock rcu_read_lock_bh rcu_read_lock rcu_read_lock_bh rcu_read_lock &br->multicast_lock &obj_hash[i].lock irq_context: 0 (wq_completion)bond112#3 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &idev->mc_lock rcu_read_lock rcu_read_lock rcu_read_lock_bh rcu_read_lock rcu_read_lock_bh rcu_read_lock &br->multicast_lock &base->lock irq_context: 0 (wq_completion)bond112#3 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &idev->mc_lock rcu_read_lock rcu_read_lock rcu_read_lock_bh rcu_read_lock rcu_read_lock_bh rcu_read_lock &data->lock irq_context: 0 (wq_completion)bond112#3 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &idev->mc_lock rcu_read_lock rcu_read_lock rcu_read_lock_bh rcu_read_lock rcu_read_lock_bh rcu_read_lock &obj_hash[i].lock irq_context: 0 (wq_completion)bond126#2 (work_completion)(&(&slave->notify_work)->work) rtnl_mutex net_rwsem irq_context: 0 (wq_completion)bond126#2 (work_completion)(&(&slave->notify_work)->work) rtnl_mutex net_rwsem &list->lock#2 irq_context: 0 (wq_completion)bond126#2 (work_completion)(&(&slave->notify_work)->work) rtnl_mutex &tn->lock irq_context: 0 (wq_completion)bond117 (work_completion)(&(&slave->notify_work)->work) rtnl_mutex net_rwsem irq_context: 0 (wq_completion)bond117 (work_completion)(&(&slave->notify_work)->work) rtnl_mutex net_rwsem &list->lock#2 irq_context: 0 (wq_completion)bond117 (work_completion)(&(&slave->notify_work)->work) rtnl_mutex &tn->lock irq_context: 0 (wq_completion)bond29#4 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex fs_reclaim irq_context: 0 (wq_completion)bond29#4 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex fs_reclaim mmu_notifier_invalidate_range_start irq_context: 0 (wq_completion)bond29#4 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &c->lock irq_context: 0 (wq_completion)bond29#4 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &n->list_lock irq_context: 0 (wq_completion)bond29#4 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &n->list_lock &c->lock irq_context: 0 (wq_completion)bond29#4 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex rcu_read_lock &bond->stats_lock/1 irq_context: 0 (wq_completion)bond29#4 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex rcu_read_lock &ndev->lock irq_context: 0 (wq_completion)bond29#4 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex nl_table_lock irq_context: 0 (wq_completion)bond29#4 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &obj_hash[i].lock irq_context: 0 (wq_completion)bond29#4 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex nl_table_wait.lock irq_context: 0 (wq_completion)bond29#4 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex net_rwsem irq_context: 0 (wq_completion)bond29#4 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex net_rwsem &list->lock#2 irq_context: 0 (wq_completion)bond29#4 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &tn->lock irq_context: 0 (wq_completion)bond29#4 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &idev->mc_lock irq_context: 0 (wq_completion)bond29#4 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &idev->mc_lock fs_reclaim irq_context: 0 (wq_completion)bond29#4 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &idev->mc_lock fs_reclaim mmu_notifier_invalidate_range_start irq_context: 0 (wq_completion)bond29#4 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &idev->mc_lock &c->lock irq_context: 0 (wq_completion)bond29#4 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &idev->mc_lock rcu_read_lock &ndev->lock irq_context: 0 (wq_completion)bond29#4 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &idev->mc_lock rcu_read_lock &dir->lock#2 irq_context: 0 (wq_completion)bond29#4 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &idev->mc_lock rcu_read_lock &c->lock irq_context: 0 (wq_completion)bond29#4 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &idev->mc_lock rcu_read_lock &ul->lock#2 irq_context: 0 (wq_completion)bond29#4 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &idev->mc_lock rcu_read_lock rcu_read_lock &____s->seqcount#12 irq_context: 0 (wq_completion)bond29#4 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &idev->mc_lock rcu_read_lock rcu_read_lock &____s->seqcount#16 irq_context: 0 (wq_completion)bond29#4 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &idev->mc_lock rcu_read_lock &obj_hash[i].lock irq_context: 0 (wq_completion)bond29#4 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &idev->mc_lock rcu_read_lock &data->lock irq_context: 0 (wq_completion)bond29#4 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &rq->__lock irq_context: 0 (wq_completion)bond29#4 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex.wait_lock irq_context: 0 (wq_completion)bond29#4 (work_completion)(&(&bond->mcast_work)->work) &p->pi_lock irq_context: 0 (wq_completion)bond29#4 (work_completion)(&(&bond->mcast_work)->work) &p->pi_lock &rq->__lock irq_context: 0 (wq_completion)bond29#4 (work_completion)(&(&bond->mcast_work)->work) &p->pi_lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)bond29#4 (work_completion)(&(&bond->mcast_work)->work) &rq->__lock irq_context: 0 (wq_completion)bond29#4 (work_completion)(&(&bond->mcast_work)->work) &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)bond122#2 &rq->__lock irq_context: 0 (wq_completion)bond122#2 &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)events (work_completion)(&rdev->wiphy_work) &rdev->wiphy.mtx &wdev->mtx &rdev->bss_lock fill_pool_map-wait-type-override &____s->seqcount#2 irq_context: 0 (wq_completion)events (work_completion)(&rdev->wiphy_work) &rdev->wiphy.mtx &wdev->mtx &rdev->bss_lock fill_pool_map-wait-type-override &____s->seqcount irq_context: 0 (wq_completion)bond101#2 (work_completion)(&(&bond->mcast_work)->work) &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)bond113#3 irq_context: 0 (wq_completion)bond113#3 (work_completion)(&(&slave->notify_work)->work) irq_context: 0 (wq_completion)bond113#3 (work_completion)(&(&slave->notify_work)->work) &obj_hash[i].lock irq_context: 0 (wq_completion)bond113#3 (work_completion)(&(&slave->notify_work)->work) &base->lock irq_context: 0 (wq_completion)bond113#3 (work_completion)(&(&slave->notify_work)->work) &base->lock &obj_hash[i].lock irq_context: 0 (wq_completion)bond113#3 (work_completion)(&(&slave->notify_work)->work) &rq->__lock irq_context: 0 (wq_completion)bond113#3 (work_completion)(&(&slave->notify_work)->work) &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)bond113#3 (work_completion)(&(&bond->mcast_work)->work) irq_context: 0 (wq_completion)bond113#3 (work_completion)(&(&bond->mcast_work)->work) &obj_hash[i].lock irq_context: 0 (wq_completion)bond113#3 (work_completion)(&(&bond->mcast_work)->work) &base->lock irq_context: 0 (wq_completion)bond113#3 (work_completion)(&(&bond->mcast_work)->work) &base->lock &obj_hash[i].lock irq_context: 0 (wq_completion)bond113#3 (work_completion)(&(&bond->mcast_work)->work) &rq->__lock irq_context: 0 (wq_completion)bond113#3 (work_completion)(&(&bond->mcast_work)->work) &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)wg-kex-wg1#10 (work_completion)(&peer->transmit_handshake_work) &wg->static_identity.lock &handshake->lock &rq->__lock irq_context: 0 (wq_completion)wg-kex-wg1#10 (work_completion)(&peer->transmit_handshake_work) &wg->static_identity.lock &handshake->lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)bond129#2 irq_context: 0 (wq_completion)bond129#2 (work_completion)(&(&slave->notify_work)->work) irq_context: 0 (wq_completion)bond129#2 (work_completion)(&(&slave->notify_work)->work) &obj_hash[i].lock irq_context: 0 (wq_completion)bond129#2 (work_completion)(&(&slave->notify_work)->work) &base->lock irq_context: 0 (wq_completion)bond129#2 (work_completion)(&(&slave->notify_work)->work) &base->lock &obj_hash[i].lock irq_context: 0 (wq_completion)bond129#2 (work_completion)(&(&slave->notify_work)->work) &rq->__lock irq_context: 0 (wq_completion)bond129#2 (work_completion)(&(&slave->notify_work)->work) &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)bond113#3 (work_completion)(&(&slave->notify_work)->work) fill_pool_map-wait-type-override &c->lock irq_context: 0 (wq_completion)bond113#3 (work_completion)(&(&slave->notify_work)->work) fill_pool_map-wait-type-override &____s->seqcount#2 irq_context: 0 (wq_completion)bond113#3 (work_completion)(&(&slave->notify_work)->work) fill_pool_map-wait-type-override &____s->seqcount irq_context: 0 (wq_completion)bond113#3 (work_completion)(&(&slave->notify_work)->work) fill_pool_map-wait-type-override pool_lock#2 irq_context: 0 (wq_completion)bond113#3 (work_completion)(&(&slave->notify_work)->work) fill_pool_map-wait-type-override pool_lock irq_context: 0 (wq_completion)mld (work_completion)(&(&idev->mc_ifc_work)->work) &idev->mc_lock rcu_read_lock rcu_read_lock &tbl->lock &n->list_lock irq_context: 0 (wq_completion)mld (work_completion)(&(&idev->mc_ifc_work)->work) &idev->mc_lock rcu_read_lock rcu_read_lock &tbl->lock &n->list_lock &c->lock irq_context: 0 (wq_completion)mld (work_completion)(&(&idev->mc_ifc_work)->work) &idev->mc_lock rcu_read_lock rcu_read_lock &tbl->lock fill_pool_map-wait-type-override pool_lock#2 irq_context: 0 (wq_completion)mld (work_completion)(&(&idev->mc_ifc_work)->work) &idev->mc_lock rcu_read_lock rcu_read_lock &tbl->lock fill_pool_map-wait-type-override pool_lock irq_context: 0 (wq_completion)mld (work_completion)(&(&idev->mc_ifc_work)->work) &idev->mc_lock rcu_read_lock rcu_read_lock &tbl->lock &____s->seqcount#2 irq_context: 0 (wq_completion)mld (work_completion)(&(&idev->mc_ifc_work)->work) &idev->mc_lock rcu_read_lock rcu_read_lock &tbl->lock &____s->seqcount irq_context: 0 (wq_completion)mld (work_completion)(&(&idev->mc_ifc_work)->work) &idev->mc_lock rcu_read_lock rcu_read_lock &tbl->lock fill_pool_map-wait-type-override &c->lock irq_context: 0 (wq_completion)mld (work_completion)(&(&idev->mc_ifc_work)->work) &idev->mc_lock rcu_read_lock rcu_read_lock &tbl->lock fill_pool_map-wait-type-override &____s->seqcount#2 irq_context: 0 (wq_completion)mld (work_completion)(&(&idev->mc_ifc_work)->work) &idev->mc_lock rcu_read_lock rcu_read_lock &tbl->lock fill_pool_map-wait-type-override &____s->seqcount irq_context: 0 (wq_completion)bond129#2 (work_completion)(&(&bond->mcast_work)->work) irq_context: 0 (wq_completion)bond129#2 (work_completion)(&(&bond->mcast_work)->work) &obj_hash[i].lock irq_context: 0 (wq_completion)bond129#2 (work_completion)(&(&bond->mcast_work)->work) &base->lock irq_context: 0 (wq_completion)bond129#2 (work_completion)(&(&bond->mcast_work)->work) &base->lock &obj_hash[i].lock irq_context: 0 (wq_completion)bond129#2 (work_completion)(&(&bond->mcast_work)->work) &rq->__lock irq_context: 0 (wq_completion)bond129#2 (work_completion)(&(&bond->mcast_work)->work) &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)bond106#3 (work_completion)(&(&bond->mcast_work)->work) fill_pool_map-wait-type-override &c->lock irq_context: 0 (wq_completion)bond106#3 (work_completion)(&(&bond->mcast_work)->work) fill_pool_map-wait-type-override &____s->seqcount#2 irq_context: 0 (wq_completion)bond106#3 (work_completion)(&(&bond->mcast_work)->work) fill_pool_map-wait-type-override &____s->seqcount irq_context: 0 (wq_completion)bond106#3 (work_completion)(&(&bond->mcast_work)->work) fill_pool_map-wait-type-override pool_lock irq_context: 0 (wq_completion)bond31#3 irq_context: 0 (wq_completion)bond31#3 &rq->__lock irq_context: 0 (wq_completion)bond31#3 &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)bond31#3 (work_completion)(&(&slave->notify_work)->work) irq_context: 0 (wq_completion)bond31#3 (work_completion)(&(&slave->notify_work)->work) &obj_hash[i].lock irq_context: 0 (wq_completion)bond31#3 (work_completion)(&(&slave->notify_work)->work) &base->lock irq_context: 0 (wq_completion)bond31#3 (work_completion)(&(&slave->notify_work)->work) &base->lock &obj_hash[i].lock irq_context: 0 (wq_completion)bond142 irq_context: 0 (wq_completion)bond142 (work_completion)(&(&slave->notify_work)->work) irq_context: 0 (wq_completion)bond142 (work_completion)(&(&slave->notify_work)->work) &obj_hash[i].lock irq_context: 0 (wq_completion)bond142 (work_completion)(&(&slave->notify_work)->work) &base->lock irq_context: 0 (wq_completion)bond142 (work_completion)(&(&slave->notify_work)->work) &base->lock &obj_hash[i].lock irq_context: 0 (wq_completion)bond142 (work_completion)(&(&slave->notify_work)->work) &rq->__lock irq_context: 0 (wq_completion)bond31#3 (work_completion)(&(&bond->mcast_work)->work) irq_context: 0 (wq_completion)bond31#3 (work_completion)(&(&bond->mcast_work)->work) &rq->__lock irq_context: 0 (wq_completion)bond31#3 (work_completion)(&(&bond->mcast_work)->work) &obj_hash[i].lock irq_context: 0 (wq_completion)bond31#3 (work_completion)(&(&bond->mcast_work)->work) &base->lock irq_context: 0 (wq_completion)bond31#3 (work_completion)(&(&bond->mcast_work)->work) &base->lock &obj_hash[i].lock irq_context: 0 (wq_completion)bond142 (work_completion)(&(&bond->mcast_work)->work) irq_context: 0 (wq_completion)bond142 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex fs_reclaim irq_context: 0 (wq_completion)bond142 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex fs_reclaim mmu_notifier_invalidate_range_start irq_context: 0 (wq_completion)bond142 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &c->lock irq_context: 0 (wq_completion)bond142 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex rcu_read_lock &bond->stats_lock/1 irq_context: 0 (wq_completion)bond142 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex rcu_read_lock &ndev->lock irq_context: 0 (wq_completion)bond142 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex nl_table_lock irq_context: 0 (wq_completion)bond142 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &obj_hash[i].lock irq_context: 0 (wq_completion)bond142 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex nl_table_wait.lock irq_context: 0 (wq_completion)bond142 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex nl_table_wait.lock &p->pi_lock irq_context: 0 (wq_completion)bond142 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex nl_table_wait.lock &p->pi_lock &cfs_rq->removed.lock irq_context: 0 (wq_completion)bond142 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex nl_table_wait.lock &p->pi_lock &rq->__lock irq_context: 0 (wq_completion)bond142 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex nl_table_wait.lock &p->pi_lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)bond142 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &rq->__lock irq_context: 0 (wq_completion)bond142 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex net_rwsem irq_context: 0 (wq_completion)bond142 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex net_rwsem &list->lock#2 irq_context: 0 (wq_completion)bond142 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &tn->lock irq_context: 0 (wq_completion)bond142 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &idev->mc_lock irq_context: 0 (wq_completion)bond142 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex.wait_lock irq_context: 0 (wq_completion)bond142 (work_completion)(&(&bond->mcast_work)->work) &p->pi_lock irq_context: 0 (wq_completion)bond142 (work_completion)(&(&slave->notify_work)->work) rtnl_mutex net_rwsem irq_context: 0 (wq_completion)bond142 (work_completion)(&(&slave->notify_work)->work) rtnl_mutex net_rwsem &list->lock#2 irq_context: 0 (wq_completion)bond142 (work_completion)(&(&slave->notify_work)->work) rtnl_mutex net_rwsem &rq->__lock irq_context: 0 (wq_completion)bond142 (work_completion)(&(&slave->notify_work)->work) rtnl_mutex net_rwsem &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)bond142 (work_completion)(&(&slave->notify_work)->work) rtnl_mutex &tn->lock irq_context: 0 (wq_completion)bond142 (work_completion)(&(&slave->notify_work)->work) rtnl_mutex.wait_lock irq_context: 0 (wq_completion)bond142 (work_completion)(&(&slave->notify_work)->work) &p->pi_lock irq_context: 0 (wq_completion)bond114#3 irq_context: 0 (wq_completion)bond114#3 (work_completion)(&(&slave->notify_work)->work) irq_context: 0 (wq_completion)bond114#3 (work_completion)(&(&slave->notify_work)->work) &obj_hash[i].lock irq_context: 0 (wq_completion)bond114#3 (work_completion)(&(&slave->notify_work)->work) &base->lock irq_context: 0 (wq_completion)bond114#3 (work_completion)(&(&slave->notify_work)->work) &base->lock &obj_hash[i].lock irq_context: 0 (wq_completion)bond114#3 (work_completion)(&(&slave->notify_work)->work) &rq->__lock irq_context: 0 (wq_completion)bond114#3 (work_completion)(&(&slave->notify_work)->work) &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)bond114#3 (work_completion)(&(&bond->mcast_work)->work) irq_context: 0 (wq_completion)bond114#3 (work_completion)(&(&bond->mcast_work)->work) &obj_hash[i].lock irq_context: 0 (wq_completion)bond114#3 (work_completion)(&(&bond->mcast_work)->work) &base->lock irq_context: 0 (wq_completion)bond114#3 (work_completion)(&(&bond->mcast_work)->work) &base->lock &obj_hash[i].lock irq_context: 0 (wq_completion)bond94#2 &rq->__lock irq_context: 0 (wq_completion)bond94#2 &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)bond130#2 irq_context: 0 (wq_completion)bond130#2 (work_completion)(&(&slave->notify_work)->work) irq_context: 0 (wq_completion)bond130#2 (work_completion)(&(&slave->notify_work)->work) &obj_hash[i].lock irq_context: 0 (wq_completion)bond130#2 (work_completion)(&(&slave->notify_work)->work) &base->lock irq_context: 0 (wq_completion)bond130#2 (work_completion)(&(&slave->notify_work)->work) &base->lock &obj_hash[i].lock irq_context: 0 (wq_completion)bond130#2 (work_completion)(&(&slave->notify_work)->work) &rq->__lock irq_context: 0 (wq_completion)bond130#2 (work_completion)(&(&slave->notify_work)->work) &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)bond130#2 (work_completion)(&(&bond->mcast_work)->work) irq_context: 0 (wq_completion)bond130#2 (work_completion)(&(&bond->mcast_work)->work) &obj_hash[i].lock irq_context: 0 (wq_completion)bond130#2 (work_completion)(&(&bond->mcast_work)->work) &base->lock irq_context: 0 (wq_completion)bond130#2 (work_completion)(&(&bond->mcast_work)->work) &base->lock &obj_hash[i].lock irq_context: 0 (wq_completion)bond130#2 (work_completion)(&(&bond->mcast_work)->work) &rq->__lock irq_context: 0 (wq_completion)bond130#2 (work_completion)(&(&bond->mcast_work)->work) &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)bond143 irq_context: 0 (wq_completion)bond143 (work_completion)(&(&slave->notify_work)->work) irq_context: 0 (wq_completion)bond143 (work_completion)(&(&slave->notify_work)->work) &obj_hash[i].lock irq_context: 0 (wq_completion)bond143 (work_completion)(&(&slave->notify_work)->work) &base->lock irq_context: 0 (wq_completion)bond143 (work_completion)(&(&slave->notify_work)->work) &base->lock &obj_hash[i].lock irq_context: 0 (wq_completion)bond143 (work_completion)(&(&bond->mcast_work)->work) irq_context: 0 (wq_completion)bond143 (work_completion)(&(&bond->mcast_work)->work) &obj_hash[i].lock irq_context: 0 (wq_completion)bond143 (work_completion)(&(&bond->mcast_work)->work) &base->lock irq_context: 0 (wq_completion)bond143 (work_completion)(&(&bond->mcast_work)->work) &base->lock &obj_hash[i].lock irq_context: 0 (wq_completion)bond143 (work_completion)(&(&bond->mcast_work)->work) &rq->__lock irq_context: 0 (wq_completion)bond143 (work_completion)(&(&bond->mcast_work)->work) &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)bond115#3 irq_context: 0 (wq_completion)bond115#3 (work_completion)(&(&slave->notify_work)->work) irq_context: 0 (wq_completion)bond115#3 (work_completion)(&(&slave->notify_work)->work) &obj_hash[i].lock irq_context: 0 (wq_completion)bond115#3 (work_completion)(&(&slave->notify_work)->work) &base->lock irq_context: 0 (wq_completion)bond115#3 (work_completion)(&(&slave->notify_work)->work) &base->lock &obj_hash[i].lock irq_context: 0 (wq_completion)bond115#3 (work_completion)(&(&slave->notify_work)->work) &rq->__lock irq_context: 0 (wq_completion)bond115#3 (work_completion)(&(&slave->notify_work)->work) &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)bond115#3 (work_completion)(&(&bond->mcast_work)->work) irq_context: 0 (wq_completion)bond115#3 (work_completion)(&(&bond->mcast_work)->work) &obj_hash[i].lock irq_context: 0 (wq_completion)bond115#3 (work_completion)(&(&bond->mcast_work)->work) &base->lock irq_context: 0 (wq_completion)bond115#3 (work_completion)(&(&bond->mcast_work)->work) &base->lock &obj_hash[i].lock irq_context: 0 (wq_completion)bond117#2 (work_completion)(&(&bond->mcast_work)->work) fill_pool_map-wait-type-override pool_lock irq_context: 0 (wq_completion)bond25#3 (work_completion)(&(&slave->notify_work)->work) &rq->__lock irq_context: 0 (wq_completion)bond25#3 (work_completion)(&(&slave->notify_work)->work) &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)bond115#3 &rq->__lock irq_context: 0 (wq_completion)bond115#3 &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)bond32#3 irq_context: 0 (wq_completion)bond32#3 (work_completion)(&(&slave->notify_work)->work) irq_context: 0 (wq_completion)bond32#3 (work_completion)(&(&slave->notify_work)->work) &obj_hash[i].lock irq_context: 0 (wq_completion)bond32#3 (work_completion)(&(&slave->notify_work)->work) &base->lock irq_context: 0 (wq_completion)bond32#3 (work_completion)(&(&slave->notify_work)->work) &base->lock &obj_hash[i].lock irq_context: 0 (wq_completion)bond32#3 (work_completion)(&(&slave->notify_work)->work) &rq->__lock irq_context: 0 (wq_completion)bond32#3 (work_completion)(&(&slave->notify_work)->work) &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)bond136 &rq->__lock irq_context: 0 (wq_completion)bond136 &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)bond119 (work_completion)(&(&slave->notify_work)->work) &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)bond32#3 (work_completion)(&(&bond->mcast_work)->work) irq_context: 0 (wq_completion)bond32#3 (work_completion)(&(&bond->mcast_work)->work) &obj_hash[i].lock irq_context: 0 (wq_completion)bond32#3 (work_completion)(&(&bond->mcast_work)->work) &base->lock irq_context: 0 (wq_completion)bond32#3 (work_completion)(&(&bond->mcast_work)->work) &base->lock &obj_hash[i].lock irq_context: 0 (wq_completion)bond110#3 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex fs_reclaim irq_context: 0 (wq_completion)bond110#3 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex fs_reclaim mmu_notifier_invalidate_range_start irq_context: 0 (wq_completion)bond110#3 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex rcu_read_lock &bond->stats_lock/1 irq_context: 0 (wq_completion)bond110#3 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex rcu_read_lock &ndev->lock irq_context: 0 (wq_completion)bond110#3 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex nl_table_lock irq_context: 0 (wq_completion)bond110#3 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &obj_hash[i].lock irq_context: 0 (wq_completion)bond110#3 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex nl_table_wait.lock irq_context: 0 (wq_completion)bond110#3 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex net_rwsem irq_context: 0 (wq_completion)bond110#3 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex net_rwsem &rq->__lock irq_context: 0 (wq_completion)bond110#3 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex net_rwsem &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)bond110#3 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex net_rwsem &list->lock#2 irq_context: 0 (wq_completion)bond110#3 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &tn->lock irq_context: 0 (wq_completion)bond110#3 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &idev->mc_lock irq_context: 0 (wq_completion)bond110#3 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &idev->mc_lock fs_reclaim irq_context: 0 (wq_completion)bond110#3 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &idev->mc_lock fs_reclaim mmu_notifier_invalidate_range_start irq_context: 0 (wq_completion)bond110#3 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &idev->mc_lock rcu_read_lock &ndev->lock irq_context: 0 (wq_completion)bond110#3 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &idev->mc_lock rcu_read_lock &dir->lock#2 irq_context: 0 (wq_completion)bond110#3 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &idev->mc_lock rcu_read_lock &ul->lock#2 irq_context: 0 (wq_completion)bond110#3 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &idev->mc_lock rcu_read_lock rcu_read_lock &____s->seqcount#10 irq_context: 0 (wq_completion)bond110#3 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &idev->mc_lock rcu_read_lock rcu_read_lock rcu_read_lock_bh rcu_read_lock rcu_read_lock_bh rcu_read_lock &br->multicast_lock irq_context: 0 (wq_completion)bond110#3 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &idev->mc_lock rcu_read_lock rcu_read_lock rcu_read_lock_bh rcu_read_lock rcu_read_lock_bh rcu_read_lock &br->multicast_lock &obj_hash[i].lock irq_context: 0 (wq_completion)bond110#3 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &idev->mc_lock rcu_read_lock rcu_read_lock rcu_read_lock_bh rcu_read_lock rcu_read_lock_bh rcu_read_lock &br->multicast_lock &base->lock irq_context: 0 (wq_completion)bond110#3 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &idev->mc_lock rcu_read_lock rcu_read_lock rcu_read_lock_bh rcu_read_lock rcu_read_lock_bh rcu_read_lock &data->lock irq_context: 0 (wq_completion)bond110#3 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &idev->mc_lock rcu_read_lock rcu_read_lock rcu_read_lock_bh rcu_read_lock rcu_read_lock_bh rcu_read_lock &obj_hash[i].lock irq_context: 0 (wq_completion)bond110#3 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &idev->mc_lock rcu_read_lock rcu_read_lock rcu_node_0 irq_context: 0 (wq_completion)bond110#3 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &idev->mc_lock rcu_read_lock rcu_read_lock &rq->__lock irq_context: 0 (wq_completion)bond110#3 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &idev->mc_lock rcu_read_lock rcu_read_lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)bond110#3 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &idev->mc_lock rcu_read_lock rcu_node_0 irq_context: 0 (wq_completion)bond110#3 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex.wait_lock irq_context: 0 (wq_completion)bond110#3 (work_completion)(&(&bond->mcast_work)->work) &p->pi_lock irq_context: 0 (wq_completion)bond110#3 (work_completion)(&(&bond->mcast_work)->work) &p->pi_lock &rq->__lock irq_context: 0 (wq_completion)bond110#3 (work_completion)(&(&bond->mcast_work)->work) &p->pi_lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)bond144 irq_context: 0 (wq_completion)bond144 (work_completion)(&(&slave->notify_work)->work) irq_context: 0 (wq_completion)bond144 (work_completion)(&(&slave->notify_work)->work) &obj_hash[i].lock irq_context: 0 (wq_completion)bond144 (work_completion)(&(&slave->notify_work)->work) &base->lock irq_context: 0 (wq_completion)bond144 (work_completion)(&(&slave->notify_work)->work) &base->lock &obj_hash[i].lock irq_context: 0 (wq_completion)bond144 (work_completion)(&(&slave->notify_work)->work) &rq->__lock irq_context: 0 (wq_completion)bond144 (work_completion)(&(&slave->notify_work)->work) &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)bond144 (work_completion)(&(&bond->mcast_work)->work) irq_context: 0 (wq_completion)bond144 (work_completion)(&(&bond->mcast_work)->work) &obj_hash[i].lock irq_context: 0 (wq_completion)bond144 (work_completion)(&(&bond->mcast_work)->work) &base->lock irq_context: 0 (wq_completion)bond144 (work_completion)(&(&bond->mcast_work)->work) &base->lock &obj_hash[i].lock irq_context: 0 (wq_completion)bond144 (work_completion)(&(&bond->mcast_work)->work) &rq->__lock irq_context: 0 (wq_completion)bond144 (work_completion)(&(&bond->mcast_work)->work) &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)bond132 (work_completion)(&(&bond->mcast_work)->work) fill_pool_map-wait-type-override pool_lock irq_context: 0 (wq_completion)bond116#3 irq_context: 0 (wq_completion)bond116#3 (work_completion)(&(&slave->notify_work)->work) irq_context: 0 (wq_completion)bond116#3 (work_completion)(&(&slave->notify_work)->work) &obj_hash[i].lock irq_context: 0 (wq_completion)bond116#3 (work_completion)(&(&slave->notify_work)->work) &base->lock irq_context: 0 (wq_completion)bond116#3 (work_completion)(&(&slave->notify_work)->work) &base->lock &obj_hash[i].lock irq_context: 0 (wq_completion)bond116#3 (work_completion)(&(&bond->mcast_work)->work) irq_context: 0 (wq_completion)bond116#3 (work_completion)(&(&bond->mcast_work)->work) &obj_hash[i].lock irq_context: 0 (wq_completion)bond116#3 (work_completion)(&(&bond->mcast_work)->work) &base->lock irq_context: 0 (wq_completion)bond116#3 (work_completion)(&(&bond->mcast_work)->work) &base->lock &obj_hash[i].lock irq_context: 0 (wq_completion)bond143 (work_completion)(&(&slave->notify_work)->work) &rq->__lock irq_context: 0 (wq_completion)bond108#3 (work_completion)(&(&slave->notify_work)->work) &rq->__lock irq_context: softirq (&in_dev->mr_ifc_timer) fill_pool_map-wait-type-override pool_lock#2 irq_context: softirq (&in_dev->mr_ifc_timer) fill_pool_map-wait-type-override pool_lock irq_context: 0 (wq_completion)events (linkwatch_work).work rtnl_mutex pcpu_alloc_mutex pcpu_alloc_mutex.wait_lock irq_context: 0 (wq_completion)events (linkwatch_work).work rtnl_mutex pcpu_alloc_mutex &pool->lock irq_context: 0 (wq_completion)bond123#2 (work_completion)(&(&slave->notify_work)->work) rtnl_mutex net_rwsem irq_context: 0 (wq_completion)bond123#2 (work_completion)(&(&slave->notify_work)->work) rtnl_mutex net_rwsem &list->lock#2 irq_context: 0 (wq_completion)bond123#2 (work_completion)(&(&slave->notify_work)->work) rtnl_mutex &tn->lock irq_context: 0 (wq_completion)bond123#2 (work_completion)(&(&slave->notify_work)->work) rtnl_mutex.wait_lock irq_context: 0 (wq_completion)bond123#2 (work_completion)(&(&slave->notify_work)->work) &p->pi_lock irq_context: 0 (wq_completion)bond28#4 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex fs_reclaim irq_context: 0 (wq_completion)bond28#4 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex fs_reclaim mmu_notifier_invalidate_range_start irq_context: 0 (wq_completion)bond28#4 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &c->lock irq_context: 0 (wq_completion)bond28#4 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &n->list_lock irq_context: 0 (wq_completion)bond28#4 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &n->list_lock &c->lock irq_context: 0 (wq_completion)bond28#4 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex rcu_read_lock &bond->stats_lock/1 irq_context: 0 (wq_completion)bond28#4 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex rcu_read_lock &ndev->lock irq_context: 0 (wq_completion)bond28#4 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex nl_table_lock irq_context: 0 (wq_completion)bond28#4 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &obj_hash[i].lock irq_context: 0 (wq_completion)bond28#4 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex nl_table_wait.lock irq_context: 0 (wq_completion)bond28#4 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex net_rwsem irq_context: 0 (wq_completion)bond28#4 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex net_rwsem &list->lock#2 irq_context: 0 (wq_completion)bond28#4 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &tn->lock irq_context: 0 (wq_completion)bond28#4 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &idev->mc_lock irq_context: 0 (wq_completion)bond28#4 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &idev->mc_lock fs_reclaim irq_context: 0 (wq_completion)bond28#4 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &idev->mc_lock fs_reclaim mmu_notifier_invalidate_range_start irq_context: 0 (wq_completion)bond28#4 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &idev->mc_lock &c->lock irq_context: 0 (wq_completion)bond28#4 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &idev->mc_lock rcu_read_lock &ndev->lock irq_context: 0 (wq_completion)bond28#4 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &idev->mc_lock rcu_read_lock &c->lock irq_context: 0 (wq_completion)bond28#4 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &idev->mc_lock rcu_read_lock &dir->lock#2 irq_context: 0 (wq_completion)bond28#4 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &idev->mc_lock rcu_read_lock &ul->lock#2 irq_context: 0 (wq_completion)bond28#4 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &idev->mc_lock rcu_read_lock rcu_read_lock &____s->seqcount#12 irq_context: 0 (wq_completion)bond28#4 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &idev->mc_lock rcu_read_lock rcu_read_lock &____s->seqcount#16 irq_context: 0 (wq_completion)bond28#4 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &idev->mc_lock rcu_read_lock &obj_hash[i].lock irq_context: 0 (wq_completion)bond28#4 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &idev->mc_lock rcu_read_lock &data->lock irq_context: 0 (wq_completion)bond28#4 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex.wait_lock irq_context: 0 (wq_completion)bond28#4 (work_completion)(&(&bond->mcast_work)->work) &p->pi_lock irq_context: 0 (wq_completion)bond130#2 (work_completion)(&(&slave->notify_work)->work) rtnl_mutex net_rwsem irq_context: 0 (wq_completion)bond130#2 (work_completion)(&(&slave->notify_work)->work) rtnl_mutex net_rwsem &list->lock#2 irq_context: 0 (wq_completion)bond130#2 (work_completion)(&(&slave->notify_work)->work) rtnl_mutex &tn->lock irq_context: 0 (wq_completion)bond130#2 (work_completion)(&(&slave->notify_work)->work) rtnl_mutex.wait_lock irq_context: 0 (wq_completion)bond130#2 (work_completion)(&(&slave->notify_work)->work) &p->pi_lock irq_context: 0 (wq_completion)bond129#2 (work_completion)(&(&slave->notify_work)->work) rtnl_mutex net_rwsem irq_context: 0 (wq_completion)bond129#2 (work_completion)(&(&slave->notify_work)->work) rtnl_mutex net_rwsem &list->lock#2 irq_context: 0 (wq_completion)bond129#2 (work_completion)(&(&slave->notify_work)->work) rtnl_mutex &tn->lock irq_context: 0 (wq_completion)bond129#2 (work_completion)(&(&slave->notify_work)->work) rtnl_mutex.wait_lock irq_context: 0 (wq_completion)bond129#2 (work_completion)(&(&slave->notify_work)->work) &p->pi_lock irq_context: 0 (wq_completion)bond137 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex fs_reclaim irq_context: 0 (wq_completion)bond137 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex fs_reclaim mmu_notifier_invalidate_range_start irq_context: 0 (wq_completion)bond137 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex rcu_read_lock &bond->stats_lock/1 irq_context: 0 (wq_completion)bond137 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex rcu_read_lock rcu_node_0 irq_context: 0 (wq_completion)bond137 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex rcu_read_lock &rq->__lock irq_context: 0 (wq_completion)bond137 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex rcu_read_lock &ndev->lock irq_context: 0 (wq_completion)bond137 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex nl_table_lock irq_context: 0 (wq_completion)bond137 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &obj_hash[i].lock irq_context: 0 (wq_completion)bond137 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex nl_table_wait.lock irq_context: 0 (wq_completion)bond137 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex net_rwsem irq_context: 0 (wq_completion)bond137 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex net_rwsem &list->lock#2 irq_context: 0 (wq_completion)bond137 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &tn->lock irq_context: 0 (wq_completion)bond137 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &idev->mc_lock irq_context: 0 (wq_completion)bond137 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &idev->mc_lock fs_reclaim irq_context: 0 (wq_completion)bond137 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &idev->mc_lock fs_reclaim mmu_notifier_invalidate_range_start irq_context: 0 (wq_completion)bond137 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &idev->mc_lock rcu_read_lock &ndev->lock irq_context: 0 (wq_completion)bond137 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &idev->mc_lock rcu_read_lock &c->lock irq_context: 0 (wq_completion)bond137 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &idev->mc_lock rcu_read_lock &n->list_lock irq_context: 0 (wq_completion)bond137 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &idev->mc_lock rcu_read_lock &n->list_lock &c->lock irq_context: 0 (wq_completion)bond137 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &idev->mc_lock rcu_read_lock &dir->lock#2 irq_context: 0 (wq_completion)bond137 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &idev->mc_lock rcu_read_lock &ul->lock#2 irq_context: 0 (wq_completion)bond137 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &idev->mc_lock rcu_read_lock rcu_read_lock &____s->seqcount#10 irq_context: 0 (wq_completion)bond137 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &idev->mc_lock rcu_read_lock rcu_read_lock rcu_read_lock_bh rcu_read_lock rcu_read_lock_bh rcu_read_lock &br->multicast_lock irq_context: 0 (wq_completion)bond137 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &idev->mc_lock rcu_read_lock rcu_read_lock rcu_read_lock_bh rcu_read_lock rcu_read_lock_bh rcu_read_lock &br->multicast_lock &obj_hash[i].lock irq_context: 0 (wq_completion)bond137 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &idev->mc_lock rcu_read_lock rcu_read_lock rcu_read_lock_bh rcu_read_lock rcu_read_lock_bh rcu_read_lock &br->multicast_lock &base->lock irq_context: 0 (wq_completion)bond137 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &idev->mc_lock rcu_read_lock rcu_read_lock rcu_read_lock_bh rcu_read_lock rcu_read_lock_bh rcu_read_lock &data->lock irq_context: 0 (wq_completion)bond137 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &idev->mc_lock rcu_read_lock rcu_read_lock rcu_read_lock_bh rcu_read_lock rcu_read_lock_bh rcu_read_lock &obj_hash[i].lock irq_context: 0 (wq_completion)bond137 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &idev->mc_lock rcu_read_lock rcu_read_lock rcu_node_0 irq_context: 0 (wq_completion)bond137 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &idev->mc_lock rcu_read_lock rcu_read_lock &rq->__lock irq_context: 0 (wq_completion)bond137 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &idev->mc_lock rcu_read_lock rcu_read_lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)bond137 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &idev->mc_lock rcu_read_lock rcu_node_0 irq_context: 0 (wq_completion)bond137 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex.wait_lock irq_context: 0 (wq_completion)bond137 (work_completion)(&(&bond->mcast_work)->work) &p->pi_lock irq_context: 0 (wq_completion)bond137 (work_completion)(&(&bond->mcast_work)->work) &p->pi_lock &rq->__lock irq_context: 0 (wq_completion)bond137 (work_completion)(&(&bond->mcast_work)->work) &p->pi_lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)bond137 (work_completion)(&(&slave->notify_work)->work) rtnl_mutex net_rwsem irq_context: 0 (wq_completion)bond137 (work_completion)(&(&slave->notify_work)->work) rtnl_mutex net_rwsem &list->lock#2 irq_context: 0 (wq_completion)bond137 (work_completion)(&(&slave->notify_work)->work) rtnl_mutex &tn->lock irq_context: 0 (wq_completion)bond137 (work_completion)(&(&slave->notify_work)->work) rtnl_mutex.wait_lock irq_context: 0 (wq_completion)bond137 (work_completion)(&(&slave->notify_work)->work) &p->pi_lock irq_context: 0 (wq_completion)bond106#3 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex fs_reclaim irq_context: 0 (wq_completion)bond106#3 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex fs_reclaim mmu_notifier_invalidate_range_start irq_context: 0 (wq_completion)bond106#3 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex rcu_read_lock &bond->stats_lock/1 irq_context: 0 (wq_completion)bond106#3 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex rcu_read_lock &ndev->lock irq_context: 0 (wq_completion)bond106#3 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex nl_table_lock irq_context: 0 (wq_completion)bond106#3 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &obj_hash[i].lock irq_context: 0 (wq_completion)bond106#3 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex nl_table_wait.lock irq_context: 0 (wq_completion)bond106#3 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex net_rwsem irq_context: 0 (wq_completion)bond106#3 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex net_rwsem &list->lock#2 irq_context: 0 (wq_completion)bond106#3 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &tn->lock irq_context: 0 (wq_completion)bond106#3 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &idev->mc_lock irq_context: 0 (wq_completion)bond106#3 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &idev->mc_lock fs_reclaim irq_context: 0 (wq_completion)bond106#3 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &idev->mc_lock fs_reclaim mmu_notifier_invalidate_range_start irq_context: 0 (wq_completion)bond106#3 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &idev->mc_lock rcu_read_lock &ndev->lock irq_context: 0 (wq_completion)bond106#3 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &idev->mc_lock rcu_read_lock &dir->lock#2 irq_context: 0 (wq_completion)bond106#3 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &idev->mc_lock rcu_read_lock &c->lock irq_context: 0 (wq_completion)bond106#3 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &idev->mc_lock rcu_read_lock &ul->lock#2 irq_context: 0 (wq_completion)bond106#3 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &idev->mc_lock rcu_read_lock rcu_read_lock &____s->seqcount#10 irq_context: 0 (wq_completion)bond106#3 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &idev->mc_lock rcu_read_lock rcu_read_lock rcu_read_lock_bh rcu_read_lock rcu_read_lock_bh rcu_read_lock &br->multicast_lock irq_context: 0 (wq_completion)bond106#3 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &idev->mc_lock rcu_read_lock rcu_read_lock rcu_read_lock_bh rcu_read_lock rcu_read_lock_bh rcu_read_lock &br->multicast_lock &obj_hash[i].lock irq_context: 0 (wq_completion)bond106#3 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &idev->mc_lock rcu_read_lock rcu_read_lock rcu_read_lock_bh rcu_read_lock rcu_read_lock_bh rcu_read_lock &br->multicast_lock &base->lock irq_context: 0 (wq_completion)bond106#3 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &idev->mc_lock rcu_read_lock rcu_read_lock rcu_read_lock_bh rcu_read_lock rcu_read_lock_bh rcu_read_lock &br->multicast_lock &base->lock &obj_hash[i].lock irq_context: 0 (wq_completion)bond106#3 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &idev->mc_lock rcu_read_lock rcu_read_lock rcu_read_lock_bh rcu_read_lock rcu_read_lock_bh rcu_read_lock &data->lock irq_context: 0 (wq_completion)bond106#3 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &idev->mc_lock rcu_read_lock rcu_read_lock rcu_read_lock_bh rcu_read_lock rcu_read_lock_bh rcu_read_lock &obj_hash[i].lock irq_context: 0 (wq_completion)bond106#3 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex.wait_lock irq_context: 0 (wq_completion)bond106#3 (work_completion)(&(&bond->mcast_work)->work) &p->pi_lock irq_context: 0 (wq_completion)bond114#3 (work_completion)(&(&slave->notify_work)->work) rtnl_mutex net_rwsem irq_context: 0 (wq_completion)bond114#3 (work_completion)(&(&slave->notify_work)->work) rtnl_mutex net_rwsem &list->lock#2 irq_context: 0 (wq_completion)bond114#3 (work_completion)(&(&slave->notify_work)->work) rtnl_mutex &tn->lock irq_context: 0 (wq_completion)bond114#3 (work_completion)(&(&slave->notify_work)->work) rtnl_mutex.wait_lock irq_context: 0 (wq_completion)bond114#3 (work_completion)(&(&slave->notify_work)->work) &p->pi_lock irq_context: 0 (wq_completion)bond25#3 (work_completion)(&(&slave->notify_work)->work) rtnl_mutex net_rwsem irq_context: 0 (wq_completion)bond25#3 (work_completion)(&(&slave->notify_work)->work) rtnl_mutex net_rwsem &list->lock#2 irq_context: 0 (wq_completion)bond25#3 (work_completion)(&(&slave->notify_work)->work) rtnl_mutex &tn->lock irq_context: 0 (wq_completion)bond25#3 (work_completion)(&(&slave->notify_work)->work) rtnl_mutex.wait_lock irq_context: 0 (wq_completion)bond25#3 (work_completion)(&(&slave->notify_work)->work) &p->pi_lock irq_context: 0 (wq_completion)bond117#2 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex fs_reclaim irq_context: 0 (wq_completion)bond117#2 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex fs_reclaim mmu_notifier_invalidate_range_start irq_context: 0 (wq_completion)bond117#2 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &c->lock irq_context: 0 (wq_completion)bond117#2 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &n->list_lock irq_context: 0 (wq_completion)bond117#2 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &n->list_lock &c->lock irq_context: 0 (wq_completion)bond117#2 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex rcu_read_lock &bond->stats_lock/1 irq_context: 0 (wq_completion)bond117#2 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex rcu_read_lock &ndev->lock irq_context: 0 (wq_completion)bond117#2 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex nl_table_lock irq_context: 0 (wq_completion)bond117#2 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &obj_hash[i].lock irq_context: 0 (wq_completion)bond117#2 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex nl_table_wait.lock irq_context: 0 (wq_completion)bond117#2 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex net_rwsem irq_context: 0 (wq_completion)bond117#2 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex net_rwsem &list->lock#2 irq_context: 0 (wq_completion)bond117#2 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &tn->lock irq_context: 0 (wq_completion)bond117#2 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &idev->mc_lock irq_context: 0 (wq_completion)bond117#2 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &idev->mc_lock fs_reclaim irq_context: 0 (wq_completion)bond117#2 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &idev->mc_lock fs_reclaim mmu_notifier_invalidate_range_start irq_context: 0 (wq_completion)bond117#2 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &idev->mc_lock rcu_read_lock &ndev->lock irq_context: 0 (wq_completion)bond117#2 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &idev->mc_lock rcu_read_lock &c->lock irq_context: 0 (wq_completion)bond117#2 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &idev->mc_lock rcu_read_lock &dir->lock#2 irq_context: 0 (wq_completion)bond117#2 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &idev->mc_lock rcu_read_lock &ul->lock#2 irq_context: 0 (wq_completion)bond117#2 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &idev->mc_lock rcu_read_lock rcu_read_lock &____s->seqcount#10 irq_context: 0 (wq_completion)bond117#2 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &idev->mc_lock rcu_read_lock rcu_read_lock rcu_read_lock_bh rcu_read_lock rcu_read_lock_bh rcu_read_lock &br->multicast_lock irq_context: 0 (wq_completion)bond117#2 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &idev->mc_lock rcu_read_lock rcu_read_lock rcu_read_lock_bh rcu_read_lock rcu_read_lock_bh rcu_read_lock &br->multicast_lock &obj_hash[i].lock irq_context: 0 (wq_completion)bond117#2 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &idev->mc_lock rcu_read_lock rcu_read_lock rcu_read_lock_bh rcu_read_lock rcu_read_lock_bh rcu_read_lock &br->multicast_lock &base->lock irq_context: 0 (wq_completion)bond117#2 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &idev->mc_lock rcu_read_lock rcu_read_lock rcu_read_lock_bh rcu_read_lock rcu_read_lock_bh rcu_read_lock &br->multicast_lock &base->lock &obj_hash[i].lock irq_context: 0 (wq_completion)bond117#2 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &idev->mc_lock rcu_read_lock rcu_read_lock rcu_read_lock_bh rcu_read_lock rcu_read_lock_bh rcu_read_lock &data->lock irq_context: 0 (wq_completion)bond117#2 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &idev->mc_lock rcu_read_lock rcu_read_lock rcu_read_lock_bh rcu_read_lock rcu_read_lock_bh rcu_read_lock &obj_hash[i].lock irq_context: 0 (wq_completion)bond117#2 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &idev->mc_lock rcu_read_lock rcu_read_lock rcu_node_0 irq_context: 0 (wq_completion)bond117#2 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &idev->mc_lock rcu_read_lock rcu_read_lock &rq->__lock irq_context: 0 (wq_completion)bond117#2 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &idev->mc_lock rcu_read_lock rcu_node_0 irq_context: 0 (wq_completion)bond117#2 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &rq->__lock irq_context: 0 (wq_completion)bond117#2 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)bond117#2 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex.wait_lock irq_context: 0 (wq_completion)bond117#2 (work_completion)(&(&bond->mcast_work)->work) &p->pi_lock irq_context: 0 (wq_completion)bond117#2 (work_completion)(&(&slave->notify_work)->work) rtnl_mutex net_rwsem irq_context: 0 (wq_completion)bond117#2 (work_completion)(&(&slave->notify_work)->work) rtnl_mutex net_rwsem &list->lock#2 irq_context: 0 (wq_completion)bond117#2 (work_completion)(&(&slave->notify_work)->work) rtnl_mutex &tn->lock irq_context: 0 (wq_completion)bond117#2 (work_completion)(&(&slave->notify_work)->work) rtnl_mutex.wait_lock irq_context: 0 (wq_completion)bond117#2 (work_completion)(&(&slave->notify_work)->work) &p->pi_lock irq_context: 0 (wq_completion)bond128 (work_completion)(&(&slave->notify_work)->work) &rq->__lock irq_context: 0 (wq_completion)bond128 (work_completion)(&(&slave->notify_work)->work) &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)bond132#2 irq_context: 0 (wq_completion)bond132#2 (work_completion)(&(&slave->notify_work)->work) irq_context: 0 (wq_completion)bond132#2 (work_completion)(&(&slave->notify_work)->work) &obj_hash[i].lock irq_context: 0 (wq_completion)bond132#2 (work_completion)(&(&slave->notify_work)->work) &base->lock irq_context: 0 (wq_completion)bond132#2 (work_completion)(&(&slave->notify_work)->work) &base->lock &obj_hash[i].lock irq_context: 0 (wq_completion)bond132#2 (work_completion)(&(&slave->notify_work)->work) &rq->__lock irq_context: 0 (wq_completion)bond132#2 (work_completion)(&(&slave->notify_work)->work) &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)bond132#2 (work_completion)(&(&bond->mcast_work)->work) irq_context: 0 (wq_completion)bond132#2 (work_completion)(&(&bond->mcast_work)->work) &obj_hash[i].lock irq_context: 0 (wq_completion)bond132#2 (work_completion)(&(&bond->mcast_work)->work) &base->lock irq_context: 0 (wq_completion)bond132#2 (work_completion)(&(&bond->mcast_work)->work) &base->lock &obj_hash[i].lock irq_context: 0 (wq_completion)bond112 (work_completion)(&(&bond->mcast_work)->work) &rq->__lock irq_context: 0 (wq_completion)bond112 (work_completion)(&(&bond->mcast_work)->work) &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)bond145 irq_context: 0 (wq_completion)bond145 (work_completion)(&(&slave->notify_work)->work) irq_context: 0 (wq_completion)bond145 (work_completion)(&(&slave->notify_work)->work) &obj_hash[i].lock irq_context: 0 (wq_completion)bond145 (work_completion)(&(&slave->notify_work)->work) &base->lock irq_context: 0 (wq_completion)bond145 (work_completion)(&(&slave->notify_work)->work) &base->lock &obj_hash[i].lock irq_context: 0 (wq_completion)bond145 (work_completion)(&(&slave->notify_work)->work) &rq->__lock irq_context: 0 (wq_completion)bond145 (work_completion)(&(&slave->notify_work)->work) &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)bond145 (work_completion)(&(&bond->mcast_work)->work) irq_context: 0 (wq_completion)bond145 (work_completion)(&(&bond->mcast_work)->work) &obj_hash[i].lock irq_context: 0 (wq_completion)bond145 (work_completion)(&(&bond->mcast_work)->work) &base->lock irq_context: 0 (wq_completion)bond145 (work_completion)(&(&bond->mcast_work)->work) &base->lock &obj_hash[i].lock irq_context: 0 (wq_completion)bond145 (work_completion)(&(&bond->mcast_work)->work) &rq->__lock irq_context: 0 (wq_completion)bond117#3 irq_context: 0 (wq_completion)bond117#3 (work_completion)(&(&slave->notify_work)->work) irq_context: 0 (wq_completion)bond117#3 (work_completion)(&(&slave->notify_work)->work) &obj_hash[i].lock irq_context: 0 (wq_completion)bond117#3 (work_completion)(&(&slave->notify_work)->work) &base->lock irq_context: 0 (wq_completion)bond117#3 (work_completion)(&(&slave->notify_work)->work) &base->lock &obj_hash[i].lock irq_context: 0 (wq_completion)bond117#3 (work_completion)(&(&bond->mcast_work)->work) irq_context: 0 (wq_completion)bond117#3 (work_completion)(&(&bond->mcast_work)->work) &obj_hash[i].lock irq_context: 0 (wq_completion)bond117#3 (work_completion)(&(&bond->mcast_work)->work) &base->lock irq_context: 0 (wq_completion)bond117#3 (work_completion)(&(&bond->mcast_work)->work) &base->lock &obj_hash[i].lock irq_context: 0 (wq_completion)bond89#2 (work_completion)(&(&bond->mcast_work)->work) &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)bond97#3 &rq->__lock irq_context: 0 (wq_completion)bond97#3 &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)bond101#2 &rq->__lock irq_context: 0 (wq_completion)bond101#2 &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)bond33#2 irq_context: 0 (wq_completion)bond33#2 &rq->__lock irq_context: 0 (wq_completion)bond33#2 &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)bond33#2 (work_completion)(&(&slave->notify_work)->work) irq_context: 0 (wq_completion)bond33#2 (work_completion)(&(&slave->notify_work)->work) &obj_hash[i].lock irq_context: 0 (wq_completion)bond33#2 (work_completion)(&(&slave->notify_work)->work) &base->lock irq_context: 0 (wq_completion)bond33#2 (work_completion)(&(&slave->notify_work)->work) &base->lock &obj_hash[i].lock irq_context: 0 (wq_completion)bond33#2 (work_completion)(&(&slave->notify_work)->work) &rq->__lock irq_context: 0 (wq_completion)bond33#2 (work_completion)(&(&slave->notify_work)->work) &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)bond33#2 (work_completion)(&(&bond->mcast_work)->work) irq_context: 0 (wq_completion)bond33#2 (work_completion)(&(&bond->mcast_work)->work) &obj_hash[i].lock irq_context: 0 (wq_completion)bond33#2 (work_completion)(&(&bond->mcast_work)->work) &base->lock irq_context: 0 (wq_completion)bond33#2 (work_completion)(&(&bond->mcast_work)->work) &base->lock &obj_hash[i].lock irq_context: 0 (wq_completion)bond135 (work_completion)(&(&slave->notify_work)->work) rtnl_mutex net_rwsem irq_context: 0 (wq_completion)bond135 (work_completion)(&(&slave->notify_work)->work) rtnl_mutex net_rwsem &rq->__lock irq_context: 0 (wq_completion)bond135 (work_completion)(&(&slave->notify_work)->work) rtnl_mutex net_rwsem &list->lock#2 irq_context: 0 (wq_completion)bond135 (work_completion)(&(&slave->notify_work)->work) rtnl_mutex &tn->lock irq_context: 0 (wq_completion)bond135 (work_completion)(&(&slave->notify_work)->work) rtnl_mutex.wait_lock irq_context: 0 (wq_completion)bond135 (work_completion)(&(&slave->notify_work)->work) &p->pi_lock irq_context: 0 (wq_completion)bond135 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex fs_reclaim irq_context: 0 (wq_completion)bond135 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex fs_reclaim mmu_notifier_invalidate_range_start irq_context: 0 (wq_completion)bond135 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &c->lock irq_context: 0 (wq_completion)bond135 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &rq->__lock irq_context: 0 (wq_completion)bond135 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)bond135 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex rcu_read_lock &bond->stats_lock/1 irq_context: 0 (wq_completion)bond135 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex rcu_read_lock &ndev->lock irq_context: 0 (wq_completion)bond135 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex nl_table_lock irq_context: 0 (wq_completion)bond135 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &obj_hash[i].lock irq_context: 0 (wq_completion)bond135 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex nl_table_wait.lock irq_context: 0 (wq_completion)bond135 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex net_rwsem irq_context: 0 (wq_completion)bond135 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex net_rwsem &list->lock#2 irq_context: 0 (wq_completion)bond135 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &tn->lock irq_context: 0 (wq_completion)bond135 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &idev->mc_lock irq_context: 0 (wq_completion)bond135 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &idev->mc_lock fs_reclaim irq_context: 0 (wq_completion)bond135 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &idev->mc_lock fs_reclaim mmu_notifier_invalidate_range_start irq_context: 0 (wq_completion)bond135 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &idev->mc_lock &c->lock irq_context: 0 (wq_completion)bond135 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &idev->mc_lock rcu_read_lock &ndev->lock irq_context: 0 (wq_completion)bond135 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &idev->mc_lock rcu_read_lock &dir->lock#2 irq_context: 0 (wq_completion)bond135 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &idev->mc_lock rcu_read_lock rcu_node_0 irq_context: 0 (wq_completion)bond135 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &idev->mc_lock rcu_read_lock &rq->__lock irq_context: 0 (wq_completion)bond135 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &idev->mc_lock rcu_read_lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)bond135 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &idev->mc_lock rcu_read_lock &ul->lock#2 irq_context: 0 (wq_completion)bond135 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &idev->mc_lock rcu_read_lock rcu_read_lock &____s->seqcount#10 irq_context: 0 (wq_completion)bond135 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &idev->mc_lock rcu_read_lock rcu_read_lock rcu_read_lock_bh rcu_read_lock rcu_read_lock_bh rcu_read_lock &br->multicast_lock irq_context: 0 (wq_completion)bond135 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &idev->mc_lock rcu_read_lock rcu_read_lock rcu_read_lock_bh rcu_read_lock rcu_read_lock_bh rcu_read_lock &br->multicast_lock &obj_hash[i].lock irq_context: 0 (wq_completion)bond135 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &idev->mc_lock rcu_read_lock rcu_read_lock rcu_read_lock_bh rcu_read_lock rcu_read_lock_bh rcu_read_lock &br->multicast_lock &base->lock irq_context: 0 (wq_completion)bond135 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &idev->mc_lock rcu_read_lock rcu_read_lock rcu_read_lock_bh rcu_read_lock rcu_read_lock_bh rcu_read_lock &br->multicast_lock &base->lock &obj_hash[i].lock irq_context: 0 (wq_completion)bond135 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &idev->mc_lock rcu_read_lock rcu_read_lock rcu_read_lock_bh rcu_read_lock rcu_read_lock_bh rcu_read_lock &data->lock irq_context: 0 (wq_completion)bond135 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &idev->mc_lock rcu_read_lock rcu_read_lock rcu_read_lock_bh rcu_read_lock rcu_read_lock_bh rcu_read_lock &obj_hash[i].lock irq_context: 0 (wq_completion)bond135 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &idev->mc_lock rcu_read_lock rcu_read_lock &rq->__lock irq_context: 0 (wq_completion)bond135 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &idev->mc_lock rcu_read_lock rcu_read_lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)bond135 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex.wait_lock irq_context: 0 (wq_completion)bond135 (work_completion)(&(&bond->mcast_work)->work) &p->pi_lock irq_context: 0 (wq_completion)bond133#2 irq_context: 0 (wq_completion)bond133#2 (work_completion)(&(&slave->notify_work)->work) irq_context: 0 (wq_completion)bond133#2 (work_completion)(&(&slave->notify_work)->work) &obj_hash[i].lock irq_context: 0 (wq_completion)bond133#2 (work_completion)(&(&slave->notify_work)->work) &base->lock irq_context: 0 (wq_completion)bond133#2 (work_completion)(&(&slave->notify_work)->work) &base->lock &obj_hash[i].lock irq_context: 0 (wq_completion)bond133#2 (work_completion)(&(&slave->notify_work)->work) &rq->__lock irq_context: 0 (wq_completion)bond111#3 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex fs_reclaim irq_context: 0 (wq_completion)bond111#3 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex fs_reclaim mmu_notifier_invalidate_range_start irq_context: 0 (wq_completion)bond111#3 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &rq->__lock irq_context: 0 (wq_completion)bond111#3 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &c->lock irq_context: 0 (wq_completion)bond111#3 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex rcu_read_lock &bond->stats_lock/1 irq_context: 0 (wq_completion)bond111#3 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex rcu_read_lock &ndev->lock irq_context: 0 (wq_completion)bond111#3 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex nl_table_lock irq_context: 0 (wq_completion)bond111#3 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &obj_hash[i].lock irq_context: 0 (wq_completion)bond111#3 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex nl_table_wait.lock irq_context: 0 (wq_completion)bond111#3 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex net_rwsem irq_context: 0 (wq_completion)bond111#3 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex net_rwsem &list->lock#2 irq_context: 0 (wq_completion)bond111#3 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &tn->lock irq_context: 0 (wq_completion)bond111#3 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &idev->mc_lock irq_context: 0 (wq_completion)bond111#3 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &idev->mc_lock fs_reclaim irq_context: 0 (wq_completion)bond111#3 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &idev->mc_lock fs_reclaim mmu_notifier_invalidate_range_start irq_context: 0 (wq_completion)bond111#3 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &idev->mc_lock &c->lock irq_context: 0 (wq_completion)bond111#3 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &idev->mc_lock &n->list_lock irq_context: 0 (wq_completion)bond111#3 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &idev->mc_lock &n->list_lock &c->lock irq_context: 0 (wq_completion)bond111#3 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &idev->mc_lock rcu_read_lock &ndev->lock irq_context: 0 (wq_completion)bond111#3 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &idev->mc_lock rcu_read_lock &dir->lock#2 irq_context: 0 (wq_completion)bond111#3 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &idev->mc_lock rcu_read_lock &ul->lock#2 irq_context: 0 (wq_completion)bond111#3 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &idev->mc_lock rcu_read_lock rcu_read_lock &____s->seqcount#10 irq_context: 0 (wq_completion)bond111#3 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &idev->mc_lock rcu_read_lock rcu_read_lock rcu_read_lock_bh rcu_read_lock rcu_read_lock_bh rcu_read_lock &br->multicast_lock irq_context: 0 (wq_completion)bond111#3 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &idev->mc_lock rcu_read_lock rcu_read_lock rcu_read_lock_bh rcu_read_lock rcu_read_lock_bh rcu_read_lock &br->multicast_lock &obj_hash[i].lock irq_context: 0 (wq_completion)bond111#3 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &idev->mc_lock rcu_read_lock rcu_read_lock rcu_read_lock_bh rcu_read_lock rcu_read_lock_bh rcu_read_lock &br->multicast_lock &base->lock irq_context: 0 (wq_completion)bond111#3 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &idev->mc_lock rcu_read_lock rcu_read_lock rcu_read_lock_bh rcu_read_lock rcu_read_lock_bh rcu_read_lock &data->lock irq_context: 0 (wq_completion)bond111#3 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &idev->mc_lock rcu_read_lock rcu_read_lock rcu_read_lock_bh rcu_read_lock rcu_read_lock_bh rcu_read_lock &obj_hash[i].lock irq_context: 0 (wq_completion)bond111#3 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex.wait_lock irq_context: 0 (wq_completion)bond111#3 (work_completion)(&(&bond->mcast_work)->work) &p->pi_lock irq_context: 0 (wq_completion)bond133#2 (work_completion)(&(&bond->mcast_work)->work) irq_context: 0 (wq_completion)bond133#2 (work_completion)(&(&bond->mcast_work)->work) &obj_hash[i].lock irq_context: 0 (wq_completion)bond133#2 (work_completion)(&(&bond->mcast_work)->work) &base->lock irq_context: 0 (wq_completion)bond133#2 (work_completion)(&(&bond->mcast_work)->work) &base->lock &obj_hash[i].lock irq_context: 0 (wq_completion)bond133#2 (work_completion)(&(&bond->mcast_work)->work) &rq->__lock irq_context: 0 (wq_completion)bond133#2 (work_completion)(&(&bond->mcast_work)->work) &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)bond133#2 (work_completion)(&(&bond->mcast_work)->work) &cfs_rq->removed.lock irq_context: 0 (wq_completion)bond146 irq_context: 0 (wq_completion)bond146 &rq->__lock irq_context: 0 (wq_completion)bond146 &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)bond146 (work_completion)(&(&slave->notify_work)->work) irq_context: 0 (wq_completion)bond146 (work_completion)(&(&slave->notify_work)->work) &obj_hash[i].lock irq_context: 0 (wq_completion)bond146 (work_completion)(&(&slave->notify_work)->work) &base->lock irq_context: 0 (wq_completion)bond146 (work_completion)(&(&slave->notify_work)->work) &base->lock &obj_hash[i].lock irq_context: 0 (wq_completion)bond146 (work_completion)(&(&bond->mcast_work)->work) irq_context: 0 (wq_completion)bond146 (work_completion)(&(&bond->mcast_work)->work) &obj_hash[i].lock irq_context: 0 (wq_completion)bond146 (work_completion)(&(&bond->mcast_work)->work) &base->lock irq_context: 0 (wq_completion)bond146 (work_completion)(&(&bond->mcast_work)->work) &base->lock &obj_hash[i].lock irq_context: 0 (wq_completion)bond146 (work_completion)(&(&bond->mcast_work)->work) &rq->__lock irq_context: 0 (wq_completion)bond146 (work_completion)(&(&bond->mcast_work)->work) &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)bond118#3 irq_context: 0 (wq_completion)bond118#3 (work_completion)(&(&slave->notify_work)->work) irq_context: 0 (wq_completion)bond118#3 (work_completion)(&(&slave->notify_work)->work) &obj_hash[i].lock irq_context: 0 (wq_completion)bond118#3 (work_completion)(&(&slave->notify_work)->work) &base->lock irq_context: 0 (wq_completion)bond118#3 (work_completion)(&(&slave->notify_work)->work) &base->lock &obj_hash[i].lock irq_context: 0 (wq_completion)bond118#3 (work_completion)(&(&bond->mcast_work)->work) irq_context: 0 (wq_completion)bond118#3 (work_completion)(&(&bond->mcast_work)->work) &obj_hash[i].lock irq_context: 0 (wq_completion)bond118#3 (work_completion)(&(&bond->mcast_work)->work) &base->lock irq_context: 0 (wq_completion)bond118#3 (work_completion)(&(&bond->mcast_work)->work) &base->lock &obj_hash[i].lock irq_context: 0 (wq_completion)bond118#3 (work_completion)(&(&bond->mcast_work)->work) &rq->__lock irq_context: 0 (wq_completion)bond118#3 (work_completion)(&(&bond->mcast_work)->work) &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)bond140 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex fs_reclaim irq_context: 0 (wq_completion)bond140 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex fs_reclaim mmu_notifier_invalidate_range_start irq_context: 0 (wq_completion)bond140 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &c->lock irq_context: 0 (wq_completion)bond140 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex rcu_read_lock &bond->stats_lock/1 irq_context: 0 (wq_completion)bond140 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex rcu_read_lock &ndev->lock irq_context: 0 (wq_completion)bond140 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex nl_table_lock irq_context: 0 (wq_completion)bond140 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &obj_hash[i].lock irq_context: 0 (wq_completion)bond140 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex quarantine_lock irq_context: 0 (wq_completion)bond140 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex nl_table_wait.lock irq_context: 0 (wq_completion)bond140 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex net_rwsem irq_context: 0 (wq_completion)bond140 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex net_rwsem &list->lock#2 irq_context: 0 (wq_completion)bond140 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &tn->lock irq_context: 0 (wq_completion)bond140 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &idev->mc_lock irq_context: 0 (wq_completion)bond140 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &idev->mc_lock fs_reclaim irq_context: 0 (wq_completion)bond140 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &idev->mc_lock fs_reclaim mmu_notifier_invalidate_range_start irq_context: 0 (wq_completion)bond140 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &idev->mc_lock &c->lock irq_context: 0 (wq_completion)bond140 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &idev->mc_lock remove_cache_srcu irq_context: 0 (wq_completion)bond140 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &idev->mc_lock remove_cache_srcu quarantine_lock irq_context: 0 (wq_completion)bond140 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &idev->mc_lock remove_cache_srcu &c->lock irq_context: 0 (wq_completion)bond140 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &idev->mc_lock remove_cache_srcu &n->list_lock irq_context: 0 (wq_completion)bond140 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &idev->mc_lock remove_cache_srcu rcu_read_lock pool_lock#2 irq_context: 0 (wq_completion)bond140 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &idev->mc_lock remove_cache_srcu &obj_hash[i].lock irq_context: 0 (wq_completion)bond140 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &idev->mc_lock rcu_read_lock &ndev->lock irq_context: 0 (wq_completion)bond140 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &idev->mc_lock rcu_read_lock &dir->lock#2 irq_context: 0 (wq_completion)bond140 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &idev->mc_lock rcu_read_lock &ul->lock#2 irq_context: 0 (wq_completion)bond140 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &idev->mc_lock rcu_read_lock rcu_read_lock &____s->seqcount#10 irq_context: 0 (wq_completion)bond140 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &idev->mc_lock rcu_read_lock rcu_read_lock rcu_read_lock_bh rcu_read_lock rcu_read_lock_bh rcu_read_lock &br->multicast_lock irq_context: 0 (wq_completion)bond140 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &idev->mc_lock rcu_read_lock rcu_read_lock rcu_read_lock_bh rcu_read_lock rcu_read_lock_bh rcu_read_lock &br->multicast_lock &obj_hash[i].lock irq_context: 0 (wq_completion)bond140 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &idev->mc_lock rcu_read_lock rcu_read_lock rcu_read_lock_bh rcu_read_lock rcu_read_lock_bh rcu_read_lock &br->multicast_lock &base->lock irq_context: 0 (wq_completion)bond140 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &idev->mc_lock rcu_read_lock rcu_read_lock rcu_read_lock_bh rcu_read_lock rcu_read_lock_bh rcu_read_lock &data->lock irq_context: 0 (wq_completion)bond140 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &idev->mc_lock rcu_read_lock rcu_read_lock rcu_read_lock_bh rcu_read_lock rcu_read_lock_bh rcu_read_lock &obj_hash[i].lock irq_context: 0 (wq_completion)bond140 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex.wait_lock irq_context: 0 (wq_completion)bond140 (work_completion)(&(&bond->mcast_work)->work) &p->pi_lock irq_context: 0 (wq_completion)bond118#3 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex fs_reclaim irq_context: 0 (wq_completion)bond118#3 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex fs_reclaim mmu_notifier_invalidate_range_start irq_context: 0 (wq_completion)bond118#3 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex rcu_read_lock &bond->stats_lock/1 irq_context: 0 (wq_completion)bond118#3 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex rcu_read_lock &ndev->lock irq_context: 0 (wq_completion)bond118#3 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex nl_table_lock irq_context: 0 (wq_completion)bond118#3 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &obj_hash[i].lock irq_context: 0 (wq_completion)bond118#3 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex nl_table_wait.lock irq_context: 0 (wq_completion)bond118#3 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex net_rwsem irq_context: 0 (wq_completion)bond118#3 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex net_rwsem &list->lock#2 irq_context: 0 (wq_completion)bond118#3 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &tn->lock irq_context: 0 (wq_completion)bond118#3 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &idev->mc_lock irq_context: 0 (wq_completion)bond118#3 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex.wait_lock irq_context: 0 (wq_completion)bond118#3 (work_completion)(&(&bond->mcast_work)->work) &p->pi_lock irq_context: 0 (wq_completion)bond113#3 (work_completion)(&(&slave->notify_work)->work) rtnl_mutex net_rwsem irq_context: 0 (wq_completion)bond113#3 (work_completion)(&(&slave->notify_work)->work) rtnl_mutex net_rwsem &list->lock#2 irq_context: 0 (wq_completion)bond113#3 (work_completion)(&(&slave->notify_work)->work) rtnl_mutex &tn->lock irq_context: 0 (wq_completion)bond113#3 (work_completion)(&(&slave->notify_work)->work) rtnl_mutex.wait_lock irq_context: 0 (wq_completion)bond113#3 (work_completion)(&(&slave->notify_work)->work) &p->pi_lock irq_context: 0 (wq_completion)bond24#3 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex fs_reclaim irq_context: 0 (wq_completion)bond24#3 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex fs_reclaim mmu_notifier_invalidate_range_start irq_context: 0 (wq_completion)bond24#3 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &c->lock irq_context: 0 (wq_completion)bond24#3 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex rcu_read_lock &bond->stats_lock/1 irq_context: 0 (wq_completion)bond24#3 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex rcu_read_lock &ndev->lock irq_context: 0 (wq_completion)bond24#3 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex nl_table_lock irq_context: 0 (wq_completion)bond24#3 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &obj_hash[i].lock irq_context: 0 (wq_completion)bond24#3 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex nl_table_wait.lock irq_context: 0 (wq_completion)bond24#3 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex net_rwsem irq_context: 0 (wq_completion)bond24#3 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex net_rwsem &list->lock#2 irq_context: 0 (wq_completion)bond24#3 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &tn->lock irq_context: 0 (wq_completion)bond24#3 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &idev->mc_lock irq_context: 0 (wq_completion)bond24#3 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &idev->mc_lock fs_reclaim irq_context: 0 (wq_completion)bond24#3 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &idev->mc_lock fs_reclaim mmu_notifier_invalidate_range_start irq_context: 0 (wq_completion)bond24#3 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &idev->mc_lock rcu_read_lock &ndev->lock irq_context: 0 (wq_completion)bond24#3 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &idev->mc_lock rcu_read_lock &dir->lock#2 irq_context: 0 (wq_completion)bond24#3 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &idev->mc_lock rcu_read_lock &ul->lock#2 irq_context: 0 (wq_completion)bond24#3 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &idev->mc_lock rcu_read_lock rcu_read_lock &____s->seqcount#12 irq_context: 0 (wq_completion)bond24#3 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &idev->mc_lock rcu_read_lock rcu_read_lock &____s->seqcount#16 irq_context: 0 (wq_completion)bond24#3 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &idev->mc_lock rcu_read_lock &obj_hash[i].lock irq_context: 0 (wq_completion)bond24#3 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &idev->mc_lock rcu_read_lock &data->lock irq_context: 0 (wq_completion)bond24#3 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex.wait_lock irq_context: 0 (wq_completion)bond24#3 (work_completion)(&(&bond->mcast_work)->work) &p->pi_lock irq_context: 0 (wq_completion)bond134 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex fs_reclaim irq_context: 0 (wq_completion)bond134 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex fs_reclaim mmu_notifier_invalidate_range_start irq_context: 0 (wq_completion)bond134 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &c->lock irq_context: 0 (wq_completion)bond134 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &n->list_lock irq_context: 0 (wq_completion)bond134 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &n->list_lock &c->lock irq_context: 0 (wq_completion)bond134 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &rq->__lock irq_context: 0 (wq_completion)bond134 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)bond134 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex rcu_read_lock &bond->stats_lock/1 irq_context: 0 (wq_completion)bond134 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex rcu_read_lock &ndev->lock irq_context: 0 (wq_completion)bond134 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex nl_table_lock irq_context: 0 (wq_completion)bond134 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &obj_hash[i].lock irq_context: 0 (wq_completion)bond134 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex nl_table_wait.lock irq_context: 0 (wq_completion)bond134 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex net_rwsem irq_context: 0 (wq_completion)bond134 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex net_rwsem &rq->__lock irq_context: 0 (wq_completion)bond134 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex net_rwsem &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)bond134 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex net_rwsem &list->lock#2 irq_context: 0 (wq_completion)bond134 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &tn->lock irq_context: 0 (wq_completion)bond134 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &idev->mc_lock irq_context: 0 (wq_completion)bond134 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &idev->mc_lock fs_reclaim irq_context: 0 (wq_completion)bond134 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &idev->mc_lock fs_reclaim mmu_notifier_invalidate_range_start irq_context: 0 (wq_completion)bond134 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &idev->mc_lock &c->lock irq_context: 0 (wq_completion)bond134 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &idev->mc_lock rcu_read_lock &ndev->lock irq_context: 0 (wq_completion)bond134 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &idev->mc_lock rcu_read_lock &dir->lock#2 irq_context: 0 (wq_completion)bond134 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &idev->mc_lock rcu_read_lock &ul->lock#2 irq_context: 0 (wq_completion)bond134 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &idev->mc_lock rcu_read_lock rcu_read_lock &____s->seqcount#10 irq_context: 0 (wq_completion)bond134 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &idev->mc_lock rcu_read_lock rcu_read_lock rcu_read_lock_bh rcu_read_lock rcu_read_lock_bh rcu_read_lock &br->multicast_lock irq_context: 0 (wq_completion)bond134 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &idev->mc_lock rcu_read_lock rcu_read_lock rcu_read_lock_bh rcu_read_lock rcu_read_lock_bh rcu_read_lock &br->multicast_lock &obj_hash[i].lock irq_context: 0 (wq_completion)bond134 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &idev->mc_lock rcu_read_lock rcu_read_lock rcu_read_lock_bh rcu_read_lock rcu_read_lock_bh rcu_read_lock &br->multicast_lock &base->lock irq_context: 0 (wq_completion)bond134 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &idev->mc_lock rcu_read_lock rcu_read_lock rcu_read_lock_bh rcu_read_lock rcu_read_lock_bh rcu_read_lock &br->multicast_lock &base->lock &obj_hash[i].lock irq_context: 0 (wq_completion)bond134 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &idev->mc_lock rcu_read_lock rcu_read_lock rcu_read_lock_bh rcu_read_lock rcu_read_lock_bh rcu_read_lock &data->lock irq_context: 0 (wq_completion)bond134 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &idev->mc_lock rcu_read_lock rcu_read_lock rcu_read_lock_bh rcu_read_lock rcu_read_lock_bh rcu_read_lock &obj_hash[i].lock irq_context: 0 (wq_completion)bond134 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex.wait_lock irq_context: 0 (wq_completion)bond134 (work_completion)(&(&bond->mcast_work)->work) &p->pi_lock irq_context: 0 (wq_completion)bond134 (work_completion)(&(&bond->mcast_work)->work) &p->pi_lock &rq->__lock irq_context: 0 (wq_completion)bond134 (work_completion)(&(&bond->mcast_work)->work) &p->pi_lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)bond134#2 irq_context: 0 (wq_completion)bond134#2 (work_completion)(&(&slave->notify_work)->work) irq_context: 0 (wq_completion)bond134#2 (work_completion)(&(&slave->notify_work)->work) &obj_hash[i].lock irq_context: 0 (wq_completion)bond134#2 (work_completion)(&(&slave->notify_work)->work) &base->lock irq_context: 0 (wq_completion)bond134#2 (work_completion)(&(&slave->notify_work)->work) &base->lock &obj_hash[i].lock irq_context: 0 (wq_completion)bond134#2 (work_completion)(&(&slave->notify_work)->work) &rq->__lock irq_context: 0 (wq_completion)bond134#2 (work_completion)(&(&slave->notify_work)->work) &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)bond134#2 (work_completion)(&(&bond->mcast_work)->work) irq_context: 0 (wq_completion)bond134#2 (work_completion)(&(&bond->mcast_work)->work) &obj_hash[i].lock irq_context: 0 (wq_completion)bond134#2 (work_completion)(&(&bond->mcast_work)->work) &base->lock irq_context: 0 (wq_completion)bond134#2 (work_completion)(&(&bond->mcast_work)->work) &base->lock &obj_hash[i].lock irq_context: 0 (wq_completion)bond134#2 (work_completion)(&(&bond->mcast_work)->work) &rq->__lock irq_context: 0 (wq_completion)bond134#2 (work_completion)(&(&bond->mcast_work)->work) &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)bond22#4 (work_completion)(&(&slave->notify_work)->work) &rq->__lock irq_context: 0 (wq_completion)bond22#4 (work_completion)(&(&slave->notify_work)->work) &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)bond83 (work_completion)(&(&slave->notify_work)->work) rtnl_mutex net_rwsem irq_context: 0 (wq_completion)bond83 (work_completion)(&(&slave->notify_work)->work) rtnl_mutex net_rwsem &list->lock#2 irq_context: 0 (wq_completion)bond83 (work_completion)(&(&slave->notify_work)->work) rtnl_mutex &tn->lock irq_context: 0 (wq_completion)bond83 (work_completion)(&(&slave->notify_work)->work) rtnl_mutex.wait_lock irq_context: 0 (wq_completion)bond83 (work_completion)(&(&slave->notify_work)->work) &p->pi_lock irq_context: 0 (wq_completion)bond130#2 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex fs_reclaim irq_context: 0 (wq_completion)bond130#2 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex fs_reclaim mmu_notifier_invalidate_range_start irq_context: 0 (wq_completion)bond130#2 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &c->lock irq_context: 0 (wq_completion)bond130#2 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex rcu_read_lock &bond->stats_lock/1 irq_context: 0 (wq_completion)bond130#2 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex rcu_read_lock &ndev->lock irq_context: 0 (wq_completion)bond130#2 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex nl_table_lock irq_context: 0 (wq_completion)bond130#2 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &obj_hash[i].lock irq_context: 0 (wq_completion)bond130#2 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex nl_table_wait.lock irq_context: 0 (wq_completion)bond130#2 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex net_rwsem irq_context: 0 (wq_completion)bond130#2 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex net_rwsem &list->lock#2 irq_context: 0 (wq_completion)bond130#2 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &tn->lock irq_context: 0 (wq_completion)bond130#2 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &idev->mc_lock irq_context: 0 (wq_completion)bond130#2 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &idev->mc_lock fs_reclaim irq_context: 0 (wq_completion)bond130#2 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &idev->mc_lock fs_reclaim mmu_notifier_invalidate_range_start irq_context: 0 (wq_completion)bond130#2 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &idev->mc_lock rcu_read_lock &ndev->lock irq_context: 0 (wq_completion)bond130#2 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &idev->mc_lock rcu_read_lock &dir->lock#2 irq_context: 0 (wq_completion)bond130#2 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &idev->mc_lock rcu_read_lock &ul->lock#2 irq_context: 0 (wq_completion)bond130#2 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &idev->mc_lock rcu_read_lock rcu_read_lock &____s->seqcount#10 irq_context: 0 (wq_completion)bond130#2 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &idev->mc_lock rcu_read_lock rcu_read_lock rcu_read_lock_bh rcu_read_lock rcu_read_lock_bh rcu_read_lock &br->multicast_lock irq_context: 0 (wq_completion)bond130#2 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &idev->mc_lock rcu_read_lock rcu_read_lock rcu_read_lock_bh rcu_read_lock rcu_read_lock_bh rcu_read_lock &br->multicast_lock &obj_hash[i].lock irq_context: 0 (wq_completion)bond130#2 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &idev->mc_lock rcu_read_lock rcu_read_lock rcu_read_lock_bh rcu_read_lock rcu_read_lock_bh rcu_read_lock &br->multicast_lock &base->lock irq_context: 0 (wq_completion)bond130#2 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &idev->mc_lock rcu_read_lock rcu_read_lock rcu_read_lock_bh rcu_read_lock rcu_read_lock_bh rcu_read_lock &data->lock irq_context: 0 (wq_completion)bond130#2 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &idev->mc_lock rcu_read_lock rcu_read_lock rcu_read_lock_bh rcu_read_lock rcu_read_lock_bh rcu_read_lock &obj_hash[i].lock irq_context: 0 (wq_completion)bond130#2 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex.wait_lock irq_context: 0 (wq_completion)bond130#2 (work_completion)(&(&bond->mcast_work)->work) &p->pi_lock irq_context: 0 (wq_completion)bond104#2 (work_completion)(&(&slave->notify_work)->work) rtnl_mutex net_rwsem irq_context: 0 (wq_completion)bond104#2 (work_completion)(&(&slave->notify_work)->work) rtnl_mutex net_rwsem &list->lock#2 irq_context: 0 (wq_completion)bond104#2 (work_completion)(&(&slave->notify_work)->work) rtnl_mutex net_rwsem &rq->__lock irq_context: 0 (wq_completion)bond104#2 (work_completion)(&(&slave->notify_work)->work) rtnl_mutex net_rwsem &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)bond104#2 (work_completion)(&(&slave->notify_work)->work) rtnl_mutex &tn->lock irq_context: 0 (wq_completion)bond104#2 (work_completion)(&(&slave->notify_work)->work) rtnl_mutex.wait_lock irq_context: 0 (wq_completion)bond104#2 (work_completion)(&(&slave->notify_work)->work) &p->pi_lock irq_context: 0 (wq_completion)bond104#2 (work_completion)(&(&slave->notify_work)->work) &p->pi_lock &rq->__lock irq_context: 0 (wq_completion)bond104#2 (work_completion)(&(&slave->notify_work)->work) &p->pi_lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)mld (work_completion)(&(&idev->mc_ifc_work)->work) &idev->mc_lock fs_reclaim &rq->__lock irq_context: 0 (wq_completion)mld (work_completion)(&(&idev->mc_ifc_work)->work) &idev->mc_lock fs_reclaim &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)bond119#3 irq_context: 0 (wq_completion)bond119#3 (work_completion)(&(&slave->notify_work)->work) irq_context: 0 (wq_completion)bond119#3 (work_completion)(&(&slave->notify_work)->work) &obj_hash[i].lock irq_context: 0 (wq_completion)bond119#3 (work_completion)(&(&slave->notify_work)->work) &base->lock irq_context: 0 (wq_completion)bond119#3 (work_completion)(&(&slave->notify_work)->work) &base->lock &obj_hash[i].lock irq_context: 0 (wq_completion)bond119#3 (work_completion)(&(&bond->mcast_work)->work) irq_context: 0 (wq_completion)bond119#3 (work_completion)(&(&bond->mcast_work)->work) &obj_hash[i].lock irq_context: 0 (wq_completion)bond119#3 (work_completion)(&(&bond->mcast_work)->work) &base->lock irq_context: 0 (wq_completion)bond119#3 (work_completion)(&(&bond->mcast_work)->work) &base->lock &obj_hash[i].lock irq_context: 0 (wq_completion)bond147 irq_context: 0 (wq_completion)bond147 (work_completion)(&(&slave->notify_work)->work) irq_context: 0 (wq_completion)bond147 (work_completion)(&(&slave->notify_work)->work) &obj_hash[i].lock irq_context: 0 (wq_completion)bond147 (work_completion)(&(&slave->notify_work)->work) &base->lock irq_context: 0 (wq_completion)bond147 (work_completion)(&(&slave->notify_work)->work) &base->lock &obj_hash[i].lock irq_context: 0 (wq_completion)bond147 (work_completion)(&(&bond->mcast_work)->work) irq_context: 0 (wq_completion)bond147 (work_completion)(&(&bond->mcast_work)->work) &obj_hash[i].lock irq_context: 0 (wq_completion)bond147 (work_completion)(&(&bond->mcast_work)->work) &base->lock irq_context: 0 (wq_completion)bond147 (work_completion)(&(&bond->mcast_work)->work) &base->lock &obj_hash[i].lock irq_context: 0 (wq_completion)bond147 (work_completion)(&(&bond->mcast_work)->work) &rq->__lock irq_context: 0 (wq_completion)bond147 (work_completion)(&(&bond->mcast_work)->work) &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)bond147 (work_completion)(&(&bond->mcast_work)->work) &cfs_rq->removed.lock irq_context: 0 (wq_completion)bond144 &rq->__lock irq_context: 0 (wq_completion)bond144 &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)mld (work_completion)(&(&idev->mc_dad_work)->work) &idev->mc_lock rcu_read_lock &obj_hash[i].lock pool_lock irq_context: 0 (wq_completion)bond135#2 irq_context: 0 (wq_completion)bond135#2 (work_completion)(&(&slave->notify_work)->work) irq_context: 0 (wq_completion)bond135#2 (work_completion)(&(&slave->notify_work)->work) &obj_hash[i].lock irq_context: 0 (wq_completion)bond135#2 (work_completion)(&(&slave->notify_work)->work) &base->lock irq_context: 0 (wq_completion)bond135#2 (work_completion)(&(&slave->notify_work)->work) &base->lock &obj_hash[i].lock irq_context: 0 (wq_completion)bond135#2 (work_completion)(&(&slave->notify_work)->work) &rq->__lock irq_context: 0 (wq_completion)bond135#2 (work_completion)(&(&slave->notify_work)->work) &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)bond135#2 (work_completion)(&(&bond->mcast_work)->work) irq_context: 0 (wq_completion)bond135#2 (work_completion)(&(&bond->mcast_work)->work) &rq->__lock irq_context: 0 (wq_completion)bond135#2 (work_completion)(&(&bond->mcast_work)->work) &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)bond135#2 (work_completion)(&(&bond->mcast_work)->work) &obj_hash[i].lock irq_context: 0 (wq_completion)bond135#2 (work_completion)(&(&bond->mcast_work)->work) &base->lock irq_context: 0 (wq_completion)bond135#2 (work_completion)(&(&bond->mcast_work)->work) &base->lock &obj_hash[i].lock irq_context: 0 (wq_completion)bond34 irq_context: 0 (wq_completion)bond34 &rq->__lock irq_context: 0 (wq_completion)bond34 &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)bond34 (work_completion)(&(&slave->notify_work)->work) irq_context: 0 (wq_completion)bond34 (work_completion)(&(&slave->notify_work)->work) &obj_hash[i].lock irq_context: 0 (wq_completion)bond34 (work_completion)(&(&slave->notify_work)->work) &base->lock irq_context: 0 (wq_completion)bond34 (work_completion)(&(&slave->notify_work)->work) &base->lock &obj_hash[i].lock irq_context: 0 (wq_completion)bond34 (work_completion)(&(&slave->notify_work)->work) &rq->__lock irq_context: 0 (wq_completion)bond34 (work_completion)(&(&slave->notify_work)->work) &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)bond114#3 (work_completion)(&(&bond->mcast_work)->work) &rq->__lock irq_context: 0 (wq_completion)bond114#3 (work_completion)(&(&bond->mcast_work)->work) &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)bond34 (work_completion)(&(&bond->mcast_work)->work) irq_context: 0 (wq_completion)bond34 (work_completion)(&(&bond->mcast_work)->work) &obj_hash[i].lock irq_context: 0 (wq_completion)bond34 (work_completion)(&(&bond->mcast_work)->work) &base->lock irq_context: 0 (wq_completion)bond34 (work_completion)(&(&bond->mcast_work)->work) &base->lock &obj_hash[i].lock irq_context: 0 (wq_completion)bond34 (work_completion)(&(&bond->mcast_work)->work) &rq->__lock irq_context: 0 (wq_completion)bond34 (work_completion)(&(&bond->mcast_work)->work) &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)bond120#3 irq_context: 0 (wq_completion)bond120#3 (work_completion)(&(&slave->notify_work)->work) irq_context: 0 (wq_completion)bond120#3 (work_completion)(&(&slave->notify_work)->work) &obj_hash[i].lock irq_context: 0 (wq_completion)bond120#3 (work_completion)(&(&slave->notify_work)->work) &base->lock irq_context: 0 (wq_completion)bond120#3 (work_completion)(&(&slave->notify_work)->work) &base->lock &obj_hash[i].lock irq_context: 0 (wq_completion)bond120#3 (work_completion)(&(&slave->notify_work)->work) &rq->__lock irq_context: 0 (wq_completion)bond120#3 (work_completion)(&(&slave->notify_work)->work) &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)bond120#3 (work_completion)(&(&bond->mcast_work)->work) irq_context: 0 (wq_completion)bond120#3 (work_completion)(&(&bond->mcast_work)->work) &obj_hash[i].lock irq_context: 0 (wq_completion)bond120#3 (work_completion)(&(&bond->mcast_work)->work) &base->lock irq_context: 0 (wq_completion)bond120#3 (work_completion)(&(&bond->mcast_work)->work) &base->lock &obj_hash[i].lock irq_context: 0 (wq_completion)bond31#3 (work_completion)(&(&bond->mcast_work)->work) &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)bond148 irq_context: 0 (wq_completion)bond148 (work_completion)(&(&slave->notify_work)->work) irq_context: 0 (wq_completion)bond148 (work_completion)(&(&slave->notify_work)->work) &obj_hash[i].lock irq_context: 0 (wq_completion)bond148 (work_completion)(&(&slave->notify_work)->work) &base->lock irq_context: 0 (wq_completion)bond148 (work_completion)(&(&slave->notify_work)->work) &base->lock &obj_hash[i].lock irq_context: 0 (wq_completion)bond148 (work_completion)(&(&slave->notify_work)->work) &rq->__lock irq_context: 0 (wq_completion)bond148 (work_completion)(&(&slave->notify_work)->work) &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)bond148 (work_completion)(&(&bond->mcast_work)->work) irq_context: 0 (wq_completion)bond148 (work_completion)(&(&bond->mcast_work)->work) &obj_hash[i].lock irq_context: 0 (wq_completion)bond148 (work_completion)(&(&bond->mcast_work)->work) &base->lock irq_context: 0 (wq_completion)bond148 (work_completion)(&(&bond->mcast_work)->work) &base->lock &obj_hash[i].lock irq_context: 0 (wq_completion)bond148 (work_completion)(&(&bond->mcast_work)->work) &rq->__lock irq_context: 0 (wq_completion)bond148 (work_completion)(&(&bond->mcast_work)->work) &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)bond136#2 irq_context: 0 (wq_completion)bond136#2 (work_completion)(&(&slave->notify_work)->work) irq_context: 0 (wq_completion)bond136#2 (work_completion)(&(&slave->notify_work)->work) &obj_hash[i].lock irq_context: 0 (wq_completion)bond136#2 (work_completion)(&(&slave->notify_work)->work) &base->lock irq_context: 0 (wq_completion)bond136#2 (work_completion)(&(&slave->notify_work)->work) &base->lock fill_pool_map-wait-type-override pool_lock irq_context: 0 (wq_completion)bond136#2 (work_completion)(&(&slave->notify_work)->work) &base->lock &obj_hash[i].lock irq_context: 0 (wq_completion)bond136#2 (work_completion)(&(&slave->notify_work)->work) &rq->__lock irq_context: 0 (wq_completion)bond136#2 (work_completion)(&(&bond->mcast_work)->work) irq_context: 0 (wq_completion)bond136#2 (work_completion)(&(&bond->mcast_work)->work) &obj_hash[i].lock irq_context: 0 (wq_completion)bond136#2 (work_completion)(&(&bond->mcast_work)->work) &base->lock irq_context: 0 (wq_completion)bond136#2 (work_completion)(&(&bond->mcast_work)->work) &base->lock &obj_hash[i].lock irq_context: 0 (wq_completion)bond106#2 &rq->__lock irq_context: 0 (wq_completion)bond106#2 &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)bond148 &rq->__lock irq_context: 0 (wq_completion)bond148 &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)bond116#3 &rq->__lock irq_context: 0 (wq_completion)bond116#3 &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)bond149 irq_context: 0 (wq_completion)bond149 (work_completion)(&(&slave->notify_work)->work) irq_context: 0 (wq_completion)bond149 (work_completion)(&(&slave->notify_work)->work) &obj_hash[i].lock irq_context: 0 (wq_completion)bond149 (work_completion)(&(&slave->notify_work)->work) &base->lock irq_context: 0 (wq_completion)bond149 (work_completion)(&(&slave->notify_work)->work) &base->lock &obj_hash[i].lock irq_context: 0 (wq_completion)bond149 (work_completion)(&(&slave->notify_work)->work) &rq->__lock irq_context: 0 (wq_completion)bond149 (work_completion)(&(&slave->notify_work)->work) &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)bond149 (work_completion)(&(&bond->mcast_work)->work) irq_context: 0 (wq_completion)bond149 (work_completion)(&(&bond->mcast_work)->work) &obj_hash[i].lock irq_context: 0 (wq_completion)bond149 (work_completion)(&(&bond->mcast_work)->work) &base->lock irq_context: 0 (wq_completion)bond149 (work_completion)(&(&bond->mcast_work)->work) &base->lock &obj_hash[i].lock irq_context: 0 (wq_completion)bond149 (work_completion)(&(&bond->mcast_work)->work) &rq->__lock irq_context: 0 (wq_completion)bond149 (work_completion)(&(&bond->mcast_work)->work) &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)bond111#3 (work_completion)(&(&slave->notify_work)->work) &base->lock fill_pool_map-wait-type-override pool_lock irq_context: 0 rtnl_mutex fill_pool_map-wait-type-override &cfs_rq->removed.lock irq_context: 0 (wq_completion)bond121#3 irq_context: 0 (wq_completion)bond121#3 (work_completion)(&(&slave->notify_work)->work) irq_context: 0 (wq_completion)bond121#3 (work_completion)(&(&slave->notify_work)->work) &obj_hash[i].lock irq_context: 0 (wq_completion)bond121#3 (work_completion)(&(&slave->notify_work)->work) &base->lock irq_context: 0 (wq_completion)bond121#3 (work_completion)(&(&slave->notify_work)->work) &base->lock &obj_hash[i].lock irq_context: 0 (wq_completion)bond121#3 (work_completion)(&(&slave->notify_work)->work) &rq->__lock irq_context: 0 (wq_completion)bond121#3 (work_completion)(&(&slave->notify_work)->work) &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)bond121#3 (work_completion)(&(&bond->mcast_work)->work) irq_context: 0 (wq_completion)bond121#3 (work_completion)(&(&bond->mcast_work)->work) &obj_hash[i].lock irq_context: 0 (wq_completion)bond121#3 (work_completion)(&(&bond->mcast_work)->work) &base->lock irq_context: 0 (wq_completion)bond121#3 (work_completion)(&(&bond->mcast_work)->work) &base->lock &obj_hash[i].lock irq_context: 0 (wq_completion)bond121#3 (work_completion)(&(&bond->mcast_work)->work) &rq->__lock irq_context: 0 (wq_completion)bond121#3 (work_completion)(&(&bond->mcast_work)->work) &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)bond32#3 (work_completion)(&(&bond->mcast_work)->work) &rq->__lock irq_context: 0 (wq_completion)bond32#3 (work_completion)(&(&bond->mcast_work)->work) &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)bond137#2 irq_context: 0 (wq_completion)bond137#2 (work_completion)(&(&slave->notify_work)->work) irq_context: 0 (wq_completion)bond137#2 (work_completion)(&(&slave->notify_work)->work) &obj_hash[i].lock irq_context: 0 (wq_completion)bond137#2 (work_completion)(&(&slave->notify_work)->work) &base->lock irq_context: 0 (wq_completion)bond137#2 (work_completion)(&(&slave->notify_work)->work) &base->lock &obj_hash[i].lock irq_context: 0 (wq_completion)bond137#2 (work_completion)(&(&slave->notify_work)->work) &rq->__lock irq_context: 0 (wq_completion)bond137#2 (work_completion)(&(&slave->notify_work)->work) &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)bond137#2 (work_completion)(&(&bond->mcast_work)->work) irq_context: 0 (wq_completion)bond137#2 (work_completion)(&(&bond->mcast_work)->work) &obj_hash[i].lock irq_context: 0 (wq_completion)bond137#2 (work_completion)(&(&bond->mcast_work)->work) &base->lock irq_context: 0 (wq_completion)bond137#2 (work_completion)(&(&bond->mcast_work)->work) &base->lock &obj_hash[i].lock irq_context: 0 (wq_completion)bond150 irq_context: 0 (wq_completion)bond150 (work_completion)(&(&slave->notify_work)->work) irq_context: 0 (wq_completion)bond150 (work_completion)(&(&slave->notify_work)->work) &obj_hash[i].lock irq_context: 0 (wq_completion)bond150 (work_completion)(&(&slave->notify_work)->work) &base->lock irq_context: 0 (wq_completion)bond150 (work_completion)(&(&slave->notify_work)->work) &base->lock &obj_hash[i].lock irq_context: 0 (wq_completion)bond117#3 (work_completion)(&(&slave->notify_work)->work) &rq->__lock irq_context: 0 (wq_completion)bond117#3 (work_completion)(&(&slave->notify_work)->work) &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)bond150 (work_completion)(&(&bond->mcast_work)->work) irq_context: 0 (wq_completion)bond150 (work_completion)(&(&bond->mcast_work)->work) &obj_hash[i].lock irq_context: 0 (wq_completion)bond150 (work_completion)(&(&bond->mcast_work)->work) &base->lock irq_context: 0 (wq_completion)bond150 (work_completion)(&(&bond->mcast_work)->work) &base->lock &obj_hash[i].lock irq_context: 0 (wq_completion)bond150 (work_completion)(&(&slave->notify_work)->work) &rq->__lock irq_context: 0 (wq_completion)bond150 (work_completion)(&(&slave->notify_work)->work) &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)bond122#3 irq_context: 0 (wq_completion)bond122#3 (work_completion)(&(&slave->notify_work)->work) irq_context: 0 (wq_completion)bond122#3 (work_completion)(&(&slave->notify_work)->work) &obj_hash[i].lock irq_context: 0 (wq_completion)bond122#3 (work_completion)(&(&slave->notify_work)->work) &base->lock irq_context: 0 (wq_completion)bond122#3 (work_completion)(&(&slave->notify_work)->work) &base->lock &obj_hash[i].lock irq_context: 0 (wq_completion)bond122#3 (work_completion)(&(&bond->mcast_work)->work) irq_context: 0 (wq_completion)bond122#3 (work_completion)(&(&bond->mcast_work)->work) &obj_hash[i].lock irq_context: 0 (wq_completion)bond122#3 (work_completion)(&(&bond->mcast_work)->work) &base->lock irq_context: 0 (wq_completion)bond122#3 (work_completion)(&(&bond->mcast_work)->work) &base->lock &obj_hash[i].lock irq_context: 0 (wq_completion)bond138 irq_context: 0 (wq_completion)bond138 (work_completion)(&(&slave->notify_work)->work) irq_context: 0 (wq_completion)bond138 (work_completion)(&(&slave->notify_work)->work) &obj_hash[i].lock irq_context: 0 (wq_completion)bond138 (work_completion)(&(&slave->notify_work)->work) &base->lock irq_context: 0 (wq_completion)bond138 (work_completion)(&(&slave->notify_work)->work) &base->lock &obj_hash[i].lock irq_context: 0 (wq_completion)bond138 (work_completion)(&(&slave->notify_work)->work) &rq->__lock irq_context: 0 (wq_completion)bond138 (work_completion)(&(&slave->notify_work)->work) &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)bond108#3 (work_completion)(&(&slave->notify_work)->work) &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)bond138 (work_completion)(&(&bond->mcast_work)->work) irq_context: 0 (wq_completion)bond138 (work_completion)(&(&bond->mcast_work)->work) &obj_hash[i].lock irq_context: 0 (wq_completion)bond138 (work_completion)(&(&bond->mcast_work)->work) &base->lock irq_context: 0 (wq_completion)bond138 (work_completion)(&(&bond->mcast_work)->work) &base->lock &obj_hash[i].lock irq_context: 0 (wq_completion)bond138 (work_completion)(&(&bond->mcast_work)->work) &rq->__lock irq_context: 0 (wq_completion)bond138 (work_completion)(&(&bond->mcast_work)->work) &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)bond118#2 (work_completion)(&(&slave->notify_work)->work) &rq->__lock irq_context: 0 (wq_completion)bond118#2 (work_completion)(&(&slave->notify_work)->work) &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 sb_writers#3 &p->pi_lock &cfs_rq->removed.lock irq_context: 0 (wq_completion)bond30#4 &rq->__lock irq_context: 0 (wq_completion)bond30#4 &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)bond123#3 irq_context: 0 (wq_completion)bond123#3 (work_completion)(&(&slave->notify_work)->work) irq_context: 0 (wq_completion)bond123#3 (work_completion)(&(&slave->notify_work)->work) &obj_hash[i].lock irq_context: 0 (wq_completion)bond123#3 (work_completion)(&(&slave->notify_work)->work) &base->lock irq_context: 0 (wq_completion)bond123#3 (work_completion)(&(&slave->notify_work)->work) &base->lock &obj_hash[i].lock irq_context: 0 (wq_completion)bond123#3 (work_completion)(&(&slave->notify_work)->work) &rq->__lock irq_context: 0 (wq_completion)bond123#3 (work_completion)(&(&slave->notify_work)->work) &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)bond123#3 (work_completion)(&(&bond->mcast_work)->work) irq_context: 0 (wq_completion)bond123#3 (work_completion)(&(&bond->mcast_work)->work) &obj_hash[i].lock irq_context: 0 (wq_completion)bond123#3 (work_completion)(&(&bond->mcast_work)->work) &base->lock irq_context: 0 (wq_completion)bond123#3 (work_completion)(&(&bond->mcast_work)->work) &base->lock &obj_hash[i].lock irq_context: 0 (wq_completion)mld (work_completion)(&(&idev->mc_dad_work)->work) &idev->mc_lock rcu_read_lock &data->lock &obj_hash[i].lock irq_context: 0 (wq_completion)mld (work_completion)(&(&idev->mc_dad_work)->work) &idev->mc_lock rcu_read_lock &data->lock &base->lock irq_context: 0 (wq_completion)mld (work_completion)(&(&idev->mc_dad_work)->work) &idev->mc_lock rcu_read_lock &data->lock &base->lock &obj_hash[i].lock irq_context: 0 (wq_completion)bond139#2 irq_context: 0 (wq_completion)bond139#2 (work_completion)(&(&slave->notify_work)->work) irq_context: 0 (wq_completion)bond139#2 (work_completion)(&(&slave->notify_work)->work) &obj_hash[i].lock irq_context: 0 (wq_completion)bond139#2 (work_completion)(&(&slave->notify_work)->work) &base->lock irq_context: 0 (wq_completion)bond139#2 (work_completion)(&(&slave->notify_work)->work) &base->lock &obj_hash[i].lock irq_context: 0 (wq_completion)bond139#2 (work_completion)(&(&slave->notify_work)->work) &rq->__lock irq_context: 0 (wq_completion)bond139#2 (work_completion)(&(&slave->notify_work)->work) &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)bond139#2 (work_completion)(&(&bond->mcast_work)->work) irq_context: 0 (wq_completion)bond139#2 (work_completion)(&(&bond->mcast_work)->work) &obj_hash[i].lock irq_context: 0 (wq_completion)bond139#2 (work_completion)(&(&bond->mcast_work)->work) &base->lock irq_context: 0 (wq_completion)bond139#2 (work_completion)(&(&bond->mcast_work)->work) &base->lock &obj_hash[i].lock irq_context: 0 (wq_completion)bond100#3 &rq->__lock irq_context: 0 (wq_completion)bond100#3 &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)bond34 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex fs_reclaim irq_context: 0 (wq_completion)bond34 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex fs_reclaim &rq->__lock irq_context: 0 (wq_completion)bond34 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex fs_reclaim &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)bond34 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex fs_reclaim mmu_notifier_invalidate_range_start irq_context: 0 (wq_completion)bond34 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex rcu_read_lock &bond->stats_lock/1 irq_context: 0 (wq_completion)bond34 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex rcu_read_lock &ndev->lock irq_context: 0 (wq_completion)bond34 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex nl_table_lock irq_context: 0 (wq_completion)bond34 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &obj_hash[i].lock irq_context: 0 (wq_completion)bond34 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex nl_table_wait.lock irq_context: 0 (wq_completion)bond34 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex net_rwsem irq_context: 0 (wq_completion)bond34 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex net_rwsem &list->lock#2 irq_context: 0 (wq_completion)bond34 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &tn->lock irq_context: 0 (wq_completion)bond34 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &idev->mc_lock irq_context: 0 (wq_completion)bond34 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &idev->mc_lock fs_reclaim irq_context: 0 (wq_completion)bond34 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &idev->mc_lock fs_reclaim mmu_notifier_invalidate_range_start irq_context: 0 (wq_completion)bond34 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &idev->mc_lock fs_reclaim mmu_notifier_invalidate_range_start &rq->__lock irq_context: 0 (wq_completion)bond34 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &idev->mc_lock fs_reclaim mmu_notifier_invalidate_range_start &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)bond34 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &idev->mc_lock &c->lock irq_context: 0 (wq_completion)bond34 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &idev->mc_lock rcu_read_lock &ndev->lock irq_context: 0 (wq_completion)bond34 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &idev->mc_lock rcu_read_lock &dir->lock#2 irq_context: 0 (wq_completion)bond34 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &idev->mc_lock rcu_read_lock &c->lock irq_context: 0 (wq_completion)bond34 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &idev->mc_lock rcu_read_lock &ul->lock#2 irq_context: 0 (wq_completion)bond34 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &idev->mc_lock rcu_read_lock rcu_read_lock &____s->seqcount#12 irq_context: 0 (wq_completion)bond34 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &idev->mc_lock rcu_read_lock rcu_read_lock &____s->seqcount#16 irq_context: 0 (wq_completion)bond34 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &idev->mc_lock rcu_read_lock &obj_hash[i].lock irq_context: 0 (wq_completion)bond34 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &idev->mc_lock rcu_read_lock &data->lock irq_context: 0 (wq_completion)bond34 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex.wait_lock irq_context: 0 (wq_completion)bond34 (work_completion)(&(&bond->mcast_work)->work) &p->pi_lock irq_context: 0 (wq_completion)bond34 (work_completion)(&(&bond->mcast_work)->work) &p->pi_lock &rq->__lock irq_context: 0 (wq_completion)bond34 (work_completion)(&(&bond->mcast_work)->work) &p->pi_lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)bond31#3 (work_completion)(&(&slave->notify_work)->work) &rq->__lock irq_context: 0 (wq_completion)bond31#3 (work_completion)(&(&slave->notify_work)->work) &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)bond152 irq_context: 0 (wq_completion)bond152 (work_completion)(&(&slave->notify_work)->work) irq_context: 0 (wq_completion)bond152 (work_completion)(&(&slave->notify_work)->work) &obj_hash[i].lock irq_context: 0 (wq_completion)bond152 (work_completion)(&(&slave->notify_work)->work) &base->lock irq_context: 0 (wq_completion)bond152 (work_completion)(&(&slave->notify_work)->work) &base->lock &obj_hash[i].lock irq_context: 0 (wq_completion)bond152 (work_completion)(&(&slave->notify_work)->work) &rq->__lock irq_context: 0 (wq_completion)bond152 (work_completion)(&(&slave->notify_work)->work) &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)bond146 (work_completion)(&(&slave->notify_work)->work) rtnl_mutex net_rwsem irq_context: 0 (wq_completion)bond146 (work_completion)(&(&slave->notify_work)->work) rtnl_mutex net_rwsem &list->lock#2 irq_context: 0 (wq_completion)bond146 (work_completion)(&(&slave->notify_work)->work) rtnl_mutex &tn->lock irq_context: 0 (wq_completion)bond146 (work_completion)(&(&slave->notify_work)->work) rtnl_mutex.wait_lock irq_context: 0 (wq_completion)bond146 (work_completion)(&(&slave->notify_work)->work) &p->pi_lock irq_context: 0 (wq_completion)bond152 (work_completion)(&(&bond->mcast_work)->work) irq_context: 0 (wq_completion)bond152 (work_completion)(&(&bond->mcast_work)->work) &obj_hash[i].lock irq_context: 0 (wq_completion)bond152 (work_completion)(&(&bond->mcast_work)->work) &base->lock irq_context: 0 (wq_completion)bond152 (work_completion)(&(&bond->mcast_work)->work) &base->lock &obj_hash[i].lock irq_context: 0 (wq_completion)bond152 (work_completion)(&(&bond->mcast_work)->work) &rq->__lock irq_context: 0 (wq_completion)bond152 (work_completion)(&(&bond->mcast_work)->work) &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)bond124#3 irq_context: 0 (wq_completion)bond124#3 (work_completion)(&(&slave->notify_work)->work) irq_context: 0 (wq_completion)bond124#3 (work_completion)(&(&slave->notify_work)->work) &obj_hash[i].lock irq_context: 0 (wq_completion)bond124#3 (work_completion)(&(&slave->notify_work)->work) &base->lock irq_context: 0 (wq_completion)bond124#3 (work_completion)(&(&slave->notify_work)->work) &base->lock &obj_hash[i].lock irq_context: 0 (wq_completion)bond124#3 (work_completion)(&(&slave->notify_work)->work) &rq->__lock irq_context: 0 (wq_completion)bond124#3 (work_completion)(&(&slave->notify_work)->work) &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)bond124#3 (work_completion)(&(&bond->mcast_work)->work) irq_context: 0 (wq_completion)bond124#3 (work_completion)(&(&bond->mcast_work)->work) &obj_hash[i].lock irq_context: 0 (wq_completion)bond124#3 (work_completion)(&(&bond->mcast_work)->work) &base->lock irq_context: 0 (wq_completion)bond124#3 (work_completion)(&(&bond->mcast_work)->work) &base->lock &obj_hash[i].lock irq_context: 0 (wq_completion)bond35 irq_context: 0 (wq_completion)bond35 (work_completion)(&(&slave->notify_work)->work) irq_context: 0 (wq_completion)bond35 (work_completion)(&(&slave->notify_work)->work) &obj_hash[i].lock irq_context: 0 (wq_completion)bond35 (work_completion)(&(&slave->notify_work)->work) &base->lock irq_context: 0 (wq_completion)bond35 (work_completion)(&(&slave->notify_work)->work) &base->lock &obj_hash[i].lock irq_context: 0 (wq_completion)bond35 (work_completion)(&(&slave->notify_work)->work) &rq->__lock irq_context: 0 (wq_completion)bond35 (work_completion)(&(&slave->notify_work)->work) &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)bond35 (work_completion)(&(&bond->mcast_work)->work) irq_context: 0 (wq_completion)bond35 (work_completion)(&(&bond->mcast_work)->work) &obj_hash[i].lock irq_context: 0 (wq_completion)bond35 (work_completion)(&(&bond->mcast_work)->work) &base->lock irq_context: 0 (wq_completion)bond35 (work_completion)(&(&bond->mcast_work)->work) &base->lock &obj_hash[i].lock irq_context: 0 (wq_completion)bond140#2 irq_context: 0 (wq_completion)bond140#2 (work_completion)(&(&slave->notify_work)->work) irq_context: 0 (wq_completion)bond140#2 (work_completion)(&(&slave->notify_work)->work) &obj_hash[i].lock irq_context: 0 (wq_completion)bond140#2 (work_completion)(&(&slave->notify_work)->work) &base->lock irq_context: 0 (wq_completion)bond140#2 (work_completion)(&(&slave->notify_work)->work) &base->lock &obj_hash[i].lock irq_context: 0 (wq_completion)bond140#2 (work_completion)(&(&bond->mcast_work)->work) irq_context: 0 (wq_completion)bond140#2 (work_completion)(&(&bond->mcast_work)->work) &obj_hash[i].lock irq_context: 0 (wq_completion)bond140#2 (work_completion)(&(&bond->mcast_work)->work) &base->lock irq_context: 0 (wq_completion)bond140#2 (work_completion)(&(&bond->mcast_work)->work) &base->lock &obj_hash[i].lock irq_context: 0 (wq_completion)bond140#2 (work_completion)(&(&bond->mcast_work)->work) &rq->__lock irq_context: 0 (wq_completion)bond140#2 (work_completion)(&(&bond->mcast_work)->work) &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)bond140#2 (work_completion)(&(&slave->notify_work)->work) &rq->__lock irq_context: 0 (wq_completion)bond140#2 (work_completion)(&(&slave->notify_work)->work) &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)bond140#2 &rq->__lock irq_context: 0 (wq_completion)bond140#2 &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)bond112#2 (work_completion)(&(&slave->notify_work)->work) rtnl_mutex net_rwsem irq_context: 0 (wq_completion)bond112#2 (work_completion)(&(&slave->notify_work)->work) rtnl_mutex net_rwsem &list->lock#2 irq_context: 0 (wq_completion)bond112#2 (work_completion)(&(&slave->notify_work)->work) rtnl_mutex &tn->lock irq_context: 0 (wq_completion)bond112#2 (work_completion)(&(&slave->notify_work)->work) rtnl_mutex.wait_lock irq_context: 0 (wq_completion)bond112#2 (work_completion)(&(&slave->notify_work)->work) &p->pi_lock irq_context: 0 (wq_completion)bond153 irq_context: 0 (wq_completion)bond153 (work_completion)(&(&slave->notify_work)->work) irq_context: 0 (wq_completion)bond153 (work_completion)(&(&slave->notify_work)->work) &obj_hash[i].lock irq_context: 0 (wq_completion)bond153 (work_completion)(&(&slave->notify_work)->work) &base->lock irq_context: 0 (wq_completion)bond153 (work_completion)(&(&slave->notify_work)->work) &base->lock &obj_hash[i].lock irq_context: 0 (wq_completion)bond153 (work_completion)(&(&slave->notify_work)->work) &rq->__lock irq_context: 0 (wq_completion)bond153 (work_completion)(&(&slave->notify_work)->work) &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)bond153 (work_completion)(&(&bond->mcast_work)->work) irq_context: 0 (wq_completion)bond153 (work_completion)(&(&bond->mcast_work)->work) &obj_hash[i].lock irq_context: 0 (wq_completion)bond153 (work_completion)(&(&bond->mcast_work)->work) &base->lock irq_context: 0 (wq_completion)bond153 (work_completion)(&(&bond->mcast_work)->work) &base->lock &obj_hash[i].lock irq_context: 0 (wq_completion)bond26#4 &rq->__lock irq_context: 0 (wq_completion)bond26#4 &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)bond125#3 irq_context: 0 (wq_completion)bond125#3 (work_completion)(&(&slave->notify_work)->work) irq_context: 0 (wq_completion)bond125#3 (work_completion)(&(&slave->notify_work)->work) &obj_hash[i].lock irq_context: 0 (wq_completion)bond125#3 (work_completion)(&(&slave->notify_work)->work) &base->lock irq_context: 0 (wq_completion)bond125#3 (work_completion)(&(&slave->notify_work)->work) &base->lock &obj_hash[i].lock irq_context: 0 (wq_completion)bond125#3 (work_completion)(&(&slave->notify_work)->work) &rq->__lock irq_context: 0 (wq_completion)bond125#3 (work_completion)(&(&slave->notify_work)->work) &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)bond125#3 (work_completion)(&(&bond->mcast_work)->work) irq_context: 0 (wq_completion)bond125#3 (work_completion)(&(&bond->mcast_work)->work) &obj_hash[i].lock irq_context: 0 (wq_completion)bond125#3 (work_completion)(&(&bond->mcast_work)->work) &base->lock irq_context: 0 (wq_completion)bond125#3 (work_completion)(&(&bond->mcast_work)->work) &base->lock &obj_hash[i].lock irq_context: 0 (wq_completion)bond140#2 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex fs_reclaim irq_context: 0 (wq_completion)bond140#2 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex fs_reclaim mmu_notifier_invalidate_range_start irq_context: 0 (wq_completion)bond140#2 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &c->lock irq_context: 0 (wq_completion)bond140#2 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex rcu_read_lock &bond->stats_lock/1 irq_context: 0 (wq_completion)bond140#2 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex rcu_read_lock &ndev->lock irq_context: 0 (wq_completion)bond140#2 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex nl_table_lock irq_context: 0 (wq_completion)bond140#2 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &obj_hash[i].lock irq_context: 0 (wq_completion)bond140#2 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex nl_table_wait.lock irq_context: 0 (wq_completion)bond140#2 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex net_rwsem irq_context: 0 (wq_completion)bond140#2 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex net_rwsem &list->lock#2 irq_context: 0 (wq_completion)bond140#2 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &tn->lock irq_context: 0 (wq_completion)bond140#2 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &idev->mc_lock irq_context: 0 (wq_completion)bond140#2 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex.wait_lock irq_context: 0 (wq_completion)bond140#2 (work_completion)(&(&bond->mcast_work)->work) &p->pi_lock irq_context: 0 (wq_completion)bond141#2 irq_context: 0 (wq_completion)bond141#2 &rq->__lock irq_context: 0 (wq_completion)bond141#2 &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)bond141#2 (work_completion)(&(&slave->notify_work)->work) irq_context: 0 (wq_completion)bond141#2 (work_completion)(&(&slave->notify_work)->work) &obj_hash[i].lock irq_context: 0 (wq_completion)bond141#2 (work_completion)(&(&slave->notify_work)->work) &base->lock irq_context: 0 (wq_completion)bond141#2 (work_completion)(&(&slave->notify_work)->work) &base->lock &obj_hash[i].lock irq_context: 0 (wq_completion)bond141#2 (work_completion)(&(&bond->mcast_work)->work) irq_context: 0 (wq_completion)bond141#2 (work_completion)(&(&bond->mcast_work)->work) &obj_hash[i].lock irq_context: 0 (wq_completion)bond141#2 (work_completion)(&(&bond->mcast_work)->work) &base->lock irq_context: 0 (wq_completion)bond141#2 (work_completion)(&(&bond->mcast_work)->work) &base->lock &obj_hash[i].lock irq_context: 0 (wq_completion)bond89#2 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex fs_reclaim irq_context: 0 (wq_completion)bond89#2 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex fs_reclaim mmu_notifier_invalidate_range_start irq_context: 0 (wq_completion)bond89#2 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &c->lock irq_context: 0 (wq_completion)bond89#2 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &n->list_lock irq_context: 0 (wq_completion)bond89#2 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &n->list_lock &c->lock irq_context: 0 (wq_completion)bond89#2 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex rcu_read_lock &bond->stats_lock/1 irq_context: 0 (wq_completion)bond89#2 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex rcu_read_lock &ndev->lock irq_context: 0 (wq_completion)bond89#2 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex nl_table_lock irq_context: 0 (wq_completion)bond89#2 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &obj_hash[i].lock irq_context: 0 (wq_completion)bond89#2 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex nl_table_wait.lock irq_context: 0 (wq_completion)bond89#2 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex net_rwsem irq_context: 0 (wq_completion)bond89#2 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex net_rwsem &list->lock#2 irq_context: 0 (wq_completion)bond89#2 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &tn->lock irq_context: 0 (wq_completion)bond89#2 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &idev->mc_lock irq_context: 0 (wq_completion)bond89#2 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &idev->mc_lock fs_reclaim irq_context: 0 (wq_completion)bond89#2 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &idev->mc_lock fs_reclaim mmu_notifier_invalidate_range_start irq_context: 0 (wq_completion)bond89#2 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &idev->mc_lock rcu_read_lock &ndev->lock irq_context: 0 (wq_completion)bond89#2 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &idev->mc_lock rcu_read_lock &dir->lock#2 irq_context: 0 (wq_completion)bond89#2 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &idev->mc_lock rcu_read_lock &ul->lock#2 irq_context: 0 (wq_completion)bond89#2 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &idev->mc_lock rcu_read_lock rcu_read_lock &____s->seqcount#10 irq_context: 0 (wq_completion)bond89#2 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &idev->mc_lock rcu_read_lock rcu_read_lock rcu_read_lock_bh rcu_read_lock rcu_read_lock_bh rcu_read_lock &br->multicast_lock irq_context: 0 (wq_completion)bond89#2 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &idev->mc_lock rcu_read_lock rcu_read_lock rcu_read_lock_bh rcu_read_lock rcu_read_lock_bh rcu_read_lock &br->multicast_lock &obj_hash[i].lock irq_context: 0 (wq_completion)bond89#2 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &idev->mc_lock rcu_read_lock rcu_read_lock rcu_read_lock_bh rcu_read_lock rcu_read_lock_bh rcu_read_lock &br->multicast_lock &base->lock irq_context: 0 (wq_completion)bond89#2 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &idev->mc_lock rcu_read_lock rcu_read_lock rcu_read_lock_bh rcu_read_lock rcu_read_lock_bh rcu_read_lock &br->multicast_lock &base->lock &obj_hash[i].lock irq_context: 0 (wq_completion)bond89#2 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &idev->mc_lock rcu_read_lock rcu_read_lock rcu_read_lock_bh rcu_read_lock rcu_read_lock_bh rcu_read_lock &data->lock irq_context: 0 (wq_completion)bond89#2 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &idev->mc_lock rcu_read_lock rcu_read_lock rcu_read_lock_bh rcu_read_lock rcu_read_lock_bh rcu_read_lock &obj_hash[i].lock irq_context: 0 (wq_completion)bond89#2 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex.wait_lock irq_context: 0 (wq_completion)bond89#2 (work_completion)(&(&bond->mcast_work)->work) &p->pi_lock irq_context: 0 (wq_completion)bond116 (work_completion)(&(&slave->notify_work)->work) rtnl_mutex net_rwsem irq_context: 0 (wq_completion)bond116 (work_completion)(&(&slave->notify_work)->work) rtnl_mutex net_rwsem &list->lock#2 irq_context: 0 (wq_completion)bond116 (work_completion)(&(&slave->notify_work)->work) rtnl_mutex &tn->lock irq_context: 0 (wq_completion)bond116 (work_completion)(&(&slave->notify_work)->work) rtnl_mutex.wait_lock irq_context: 0 (wq_completion)bond116 (work_completion)(&(&slave->notify_work)->work) &p->pi_lock irq_context: 0 (wq_completion)bond17#3 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex fs_reclaim irq_context: 0 (wq_completion)bond17#3 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex fs_reclaim mmu_notifier_invalidate_range_start irq_context: 0 (wq_completion)bond17#3 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex rcu_read_lock &bond->stats_lock/1 irq_context: 0 (wq_completion)bond17#3 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex rcu_read_lock &ndev->lock irq_context: 0 (wq_completion)bond17#3 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex nl_table_lock irq_context: 0 (wq_completion)bond17#3 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &obj_hash[i].lock irq_context: 0 (wq_completion)bond17#3 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex nl_table_wait.lock irq_context: 0 (wq_completion)bond17#3 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &rq->__lock irq_context: 0 (wq_completion)bond17#3 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)bond17#3 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex net_rwsem irq_context: 0 (wq_completion)bond17#3 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex net_rwsem &list->lock#2 irq_context: 0 (wq_completion)bond17#3 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &tn->lock irq_context: 0 (wq_completion)bond17#3 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &idev->mc_lock irq_context: 0 (wq_completion)bond17#3 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &idev->mc_lock fs_reclaim irq_context: 0 (wq_completion)bond17#3 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &idev->mc_lock fs_reclaim mmu_notifier_invalidate_range_start irq_context: 0 (wq_completion)bond17#3 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &idev->mc_lock &c->lock irq_context: 0 (wq_completion)bond17#3 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &idev->mc_lock &n->list_lock irq_context: 0 (wq_completion)bond17#3 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &idev->mc_lock &n->list_lock &c->lock irq_context: 0 (wq_completion)bond17#3 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &idev->mc_lock &rq->__lock irq_context: 0 (wq_completion)bond17#3 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &idev->mc_lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)bond17#3 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &idev->mc_lock rcu_read_lock &ndev->lock irq_context: 0 (wq_completion)bond17#3 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &idev->mc_lock rcu_read_lock &c->lock irq_context: 0 (wq_completion)bond17#3 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &idev->mc_lock rcu_read_lock &dir->lock#2 irq_context: 0 (wq_completion)bond17#3 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &idev->mc_lock rcu_read_lock &ul->lock#2 irq_context: 0 (wq_completion)bond17#3 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &idev->mc_lock rcu_read_lock rcu_read_lock &____s->seqcount#12 irq_context: 0 (wq_completion)bond17#3 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &idev->mc_lock rcu_read_lock rcu_read_lock &____s->seqcount#16 irq_context: 0 (wq_completion)bond17#3 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &idev->mc_lock rcu_read_lock &obj_hash[i].lock irq_context: 0 (wq_completion)bond17#3 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &idev->mc_lock rcu_read_lock &data->lock irq_context: 0 (wq_completion)bond17#3 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex.wait_lock irq_context: 0 (wq_completion)bond17#3 (work_completion)(&(&bond->mcast_work)->work) &p->pi_lock irq_context: 0 (wq_completion)bond17#3 (work_completion)(&(&bond->mcast_work)->work) &p->pi_lock &rq->__lock irq_context: 0 (wq_completion)bond17#3 (work_completion)(&(&bond->mcast_work)->work) &p->pi_lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)bond17#3 (work_completion)(&(&bond->mcast_work)->work) &rq->__lock irq_context: 0 (wq_completion)bond17#3 (work_completion)(&(&bond->mcast_work)->work) &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)bond154 irq_context: 0 (wq_completion)bond154 (work_completion)(&(&slave->notify_work)->work) irq_context: 0 (wq_completion)bond154 (work_completion)(&(&slave->notify_work)->work) &obj_hash[i].lock irq_context: 0 (wq_completion)bond154 (work_completion)(&(&slave->notify_work)->work) &base->lock irq_context: 0 (wq_completion)bond154 (work_completion)(&(&slave->notify_work)->work) &base->lock &obj_hash[i].lock irq_context: 0 (wq_completion)bond154 (work_completion)(&(&slave->notify_work)->work) &rq->__lock irq_context: 0 (wq_completion)bond154 (work_completion)(&(&slave->notify_work)->work) &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)bond154 (work_completion)(&(&bond->mcast_work)->work) irq_context: 0 (wq_completion)bond154 (work_completion)(&(&bond->mcast_work)->work) &obj_hash[i].lock irq_context: 0 (wq_completion)bond154 (work_completion)(&(&bond->mcast_work)->work) &base->lock irq_context: 0 (wq_completion)bond154 (work_completion)(&(&bond->mcast_work)->work) &base->lock &obj_hash[i].lock irq_context: 0 (wq_completion)bond126#3 irq_context: 0 (wq_completion)bond126#3 &rq->__lock irq_context: 0 (wq_completion)bond126#3 &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)bond126#3 (work_completion)(&(&slave->notify_work)->work) irq_context: 0 (wq_completion)bond126#3 (work_completion)(&(&slave->notify_work)->work) &obj_hash[i].lock irq_context: 0 (wq_completion)bond126#3 (work_completion)(&(&slave->notify_work)->work) &base->lock irq_context: 0 (wq_completion)bond126#3 (work_completion)(&(&slave->notify_work)->work) &base->lock &obj_hash[i].lock irq_context: 0 (wq_completion)bond126#3 (work_completion)(&(&slave->notify_work)->work) &rq->__lock irq_context: 0 (wq_completion)bond126#3 (work_completion)(&(&bond->mcast_work)->work) irq_context: 0 (wq_completion)bond126#3 (work_completion)(&(&bond->mcast_work)->work) &obj_hash[i].lock irq_context: 0 (wq_completion)bond126#3 (work_completion)(&(&bond->mcast_work)->work) &base->lock irq_context: 0 (wq_completion)bond126#3 (work_completion)(&(&bond->mcast_work)->work) &base->lock &obj_hash[i].lock irq_context: 0 (wq_completion)bond126#3 (work_completion)(&(&bond->mcast_work)->work) &rq->__lock irq_context: 0 (wq_completion)bond126#3 (work_completion)(&(&bond->mcast_work)->work) &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)bond36#2 irq_context: 0 (wq_completion)bond36#2 (work_completion)(&(&slave->notify_work)->work) irq_context: 0 (wq_completion)bond36#2 (work_completion)(&(&slave->notify_work)->work) &obj_hash[i].lock irq_context: 0 (wq_completion)bond36#2 (work_completion)(&(&slave->notify_work)->work) &base->lock irq_context: 0 (wq_completion)bond36#2 (work_completion)(&(&slave->notify_work)->work) &base->lock &obj_hash[i].lock irq_context: 0 sk_lock-AF_INET6 &dccp_hashinfo.bhash[i].lock &dccp_hashinfo.bhash2[i].lock &pcp->lock &zone->lock &____s->seqcount irq_context: 0 (wq_completion)bond102#3 (work_completion)(&(&slave->notify_work)->work) &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)bond123#3 (work_completion)(&(&bond->mcast_work)->work) &rq->__lock irq_context: 0 (wq_completion)bond123#3 (work_completion)(&(&bond->mcast_work)->work) &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)bond36#2 (work_completion)(&(&bond->mcast_work)->work) irq_context: 0 (wq_completion)bond36#2 (work_completion)(&(&bond->mcast_work)->work) &obj_hash[i].lock irq_context: 0 (wq_completion)bond36#2 (work_completion)(&(&bond->mcast_work)->work) &base->lock irq_context: 0 (wq_completion)bond36#2 (work_completion)(&(&bond->mcast_work)->work) &base->lock &obj_hash[i].lock irq_context: 0 (wq_completion)bond36#2 (work_completion)(&(&bond->mcast_work)->work) &rq->__lock irq_context: 0 (wq_completion)bond36#2 (work_completion)(&(&bond->mcast_work)->work) &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)bond142#2 irq_context: 0 (wq_completion)bond142#2 (work_completion)(&(&slave->notify_work)->work) irq_context: 0 (wq_completion)bond142#2 (work_completion)(&(&slave->notify_work)->work) &obj_hash[i].lock irq_context: 0 (wq_completion)bond142#2 (work_completion)(&(&slave->notify_work)->work) &base->lock irq_context: 0 (wq_completion)bond142#2 (work_completion)(&(&slave->notify_work)->work) &base->lock &obj_hash[i].lock irq_context: 0 (wq_completion)bond147 (work_completion)(&(&slave->notify_work)->work) &rq->__lock irq_context: 0 (wq_completion)bond147 (work_completion)(&(&slave->notify_work)->work) &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)bond142#2 (work_completion)(&(&bond->mcast_work)->work) irq_context: 0 (wq_completion)bond142#2 (work_completion)(&(&bond->mcast_work)->work) &obj_hash[i].lock irq_context: 0 (wq_completion)bond142#2 (work_completion)(&(&bond->mcast_work)->work) &base->lock irq_context: 0 (wq_completion)bond142#2 (work_completion)(&(&bond->mcast_work)->work) &base->lock &obj_hash[i].lock irq_context: 0 (wq_completion)bond139#2 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex fs_reclaim irq_context: 0 (wq_completion)bond139#2 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex fs_reclaim mmu_notifier_invalidate_range_start irq_context: 0 (wq_completion)bond139#2 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &c->lock irq_context: 0 (wq_completion)bond139#2 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex rcu_read_lock &bond->stats_lock/1 irq_context: 0 (wq_completion)bond139#2 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex rcu_read_lock &ndev->lock irq_context: 0 (wq_completion)bond139#2 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex nl_table_lock irq_context: 0 (wq_completion)bond139#2 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &obj_hash[i].lock irq_context: 0 (wq_completion)bond139#2 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex nl_table_wait.lock irq_context: 0 (wq_completion)bond139#2 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex net_rwsem irq_context: 0 (wq_completion)bond139#2 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex net_rwsem &list->lock#2 irq_context: 0 (wq_completion)bond139#2 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &tn->lock irq_context: 0 (wq_completion)bond139#2 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &idev->mc_lock irq_context: 0 (wq_completion)bond139#2 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &idev->mc_lock fs_reclaim irq_context: 0 (wq_completion)bond139#2 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &idev->mc_lock fs_reclaim mmu_notifier_invalidate_range_start irq_context: 0 (wq_completion)bond139#2 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &idev->mc_lock &c->lock irq_context: 0 (wq_completion)bond139#2 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &idev->mc_lock rcu_read_lock &ndev->lock irq_context: 0 (wq_completion)bond139#2 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &idev->mc_lock rcu_read_lock &dir->lock#2 irq_context: 0 (wq_completion)bond139#2 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &idev->mc_lock rcu_read_lock &ul->lock#2 irq_context: 0 (wq_completion)bond139#2 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &idev->mc_lock rcu_read_lock rcu_read_lock &____s->seqcount#10 irq_context: 0 (wq_completion)bond139#2 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &idev->mc_lock rcu_read_lock rcu_read_lock rcu_read_lock_bh rcu_read_lock rcu_read_lock_bh rcu_read_lock &br->multicast_lock irq_context: 0 (wq_completion)bond139#2 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &idev->mc_lock rcu_read_lock rcu_read_lock rcu_read_lock_bh rcu_read_lock rcu_read_lock_bh rcu_read_lock &br->multicast_lock &obj_hash[i].lock irq_context: 0 (wq_completion)bond139#2 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &idev->mc_lock rcu_read_lock rcu_read_lock rcu_read_lock_bh rcu_read_lock rcu_read_lock_bh rcu_read_lock &br->multicast_lock &base->lock irq_context: 0 (wq_completion)bond139#2 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &idev->mc_lock rcu_read_lock rcu_read_lock rcu_read_lock_bh rcu_read_lock rcu_read_lock_bh rcu_read_lock &data->lock irq_context: 0 (wq_completion)bond139#2 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &idev->mc_lock rcu_read_lock rcu_read_lock rcu_read_lock_bh rcu_read_lock rcu_read_lock_bh rcu_read_lock &data->lock &obj_hash[i].lock irq_context: 0 (wq_completion)bond139#2 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &idev->mc_lock rcu_read_lock rcu_read_lock rcu_read_lock_bh rcu_read_lock rcu_read_lock_bh rcu_read_lock &data->lock &base->lock irq_context: 0 (wq_completion)bond139#2 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &idev->mc_lock rcu_read_lock rcu_read_lock rcu_read_lock_bh rcu_read_lock rcu_read_lock_bh rcu_read_lock &data->lock &base->lock &obj_hash[i].lock irq_context: 0 (wq_completion)bond139#2 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &idev->mc_lock rcu_read_lock rcu_read_lock rcu_read_lock_bh rcu_read_lock rcu_read_lock_bh rcu_read_lock &obj_hash[i].lock irq_context: 0 (wq_completion)bond139#2 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex.wait_lock irq_context: 0 (wq_completion)bond139#2 (work_completion)(&(&bond->mcast_work)->work) &p->pi_lock irq_context: 0 (wq_completion)bond117#3 (work_completion)(&(&slave->notify_work)->work) rtnl_mutex net_rwsem irq_context: 0 (wq_completion)bond117#3 (work_completion)(&(&slave->notify_work)->work) rtnl_mutex net_rwsem &list->lock#2 irq_context: 0 (wq_completion)bond117#3 (work_completion)(&(&slave->notify_work)->work) rtnl_mutex &tn->lock irq_context: 0 (wq_completion)bond117#3 (work_completion)(&(&slave->notify_work)->work) rtnl_mutex.wait_lock irq_context: 0 (wq_completion)bond117#3 (work_completion)(&(&slave->notify_work)->work) &p->pi_lock irq_context: 0 (wq_completion)bond124#3 (work_completion)(&(&slave->notify_work)->work) rtnl_mutex net_rwsem irq_context: 0 (wq_completion)bond124#3 (work_completion)(&(&slave->notify_work)->work) rtnl_mutex net_rwsem &list->lock#2 irq_context: 0 (wq_completion)bond124#3 (work_completion)(&(&slave->notify_work)->work) rtnl_mutex &tn->lock irq_context: 0 (wq_completion)bond124#3 (work_completion)(&(&slave->notify_work)->work) rtnl_mutex.wait_lock irq_context: 0 (wq_completion)bond124#3 (work_completion)(&(&slave->notify_work)->work) &p->pi_lock irq_context: 0 (wq_completion)bond141#2 (work_completion)(&(&slave->notify_work)->work) rtnl_mutex net_rwsem irq_context: 0 (wq_completion)bond141#2 (work_completion)(&(&slave->notify_work)->work) rtnl_mutex net_rwsem &list->lock#2 irq_context: 0 (wq_completion)bond141#2 (work_completion)(&(&slave->notify_work)->work) rtnl_mutex &tn->lock irq_context: 0 (wq_completion)bond141#2 (work_completion)(&(&slave->notify_work)->work) rtnl_mutex.wait_lock irq_context: 0 (wq_completion)bond141#2 (work_completion)(&(&slave->notify_work)->work) &p->pi_lock irq_context: 0 (wq_completion)bond144 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex fs_reclaim irq_context: 0 (wq_completion)bond144 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex fs_reclaim &rq->__lock irq_context: 0 (wq_completion)bond144 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex fs_reclaim &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)bond144 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex fs_reclaim mmu_notifier_invalidate_range_start irq_context: 0 (wq_completion)bond144 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex fs_reclaim mmu_notifier_invalidate_range_start &rq->__lock irq_context: 0 (wq_completion)bond144 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &c->lock irq_context: 0 (wq_completion)bond144 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &____s->seqcount#2 irq_context: 0 (wq_completion)bond144 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &rq->__lock irq_context: 0 (wq_completion)bond144 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)bond144 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &____s->seqcount irq_context: 0 (wq_completion)bond144 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex rcu_read_lock &bond->stats_lock/1 irq_context: 0 (wq_completion)bond144 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex rcu_read_lock rcu_node_0 irq_context: 0 (wq_completion)bond144 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex rcu_read_lock &rq->__lock irq_context: 0 (wq_completion)bond144 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex rcu_read_lock &ndev->lock irq_context: 0 (wq_completion)bond144 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex nl_table_lock irq_context: 0 (wq_completion)bond144 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &obj_hash[i].lock irq_context: 0 (wq_completion)bond144 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex nl_table_wait.lock irq_context: 0 (wq_completion)bond144 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex net_rwsem irq_context: 0 (wq_completion)bond144 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex net_rwsem &rq->__lock irq_context: 0 (wq_completion)bond144 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex net_rwsem &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)bond144 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex net_rwsem &list->lock#2 irq_context: 0 (wq_completion)bond144 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &tn->lock irq_context: 0 (wq_completion)bond144 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &idev->mc_lock irq_context: 0 (wq_completion)bond144 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &idev->mc_lock fs_reclaim irq_context: 0 (wq_completion)bond144 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &idev->mc_lock fs_reclaim mmu_notifier_invalidate_range_start irq_context: 0 (wq_completion)bond144 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &idev->mc_lock rcu_read_lock &ndev->lock irq_context: 0 (wq_completion)bond144 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &idev->mc_lock rcu_read_lock &dir->lock#2 irq_context: 0 (wq_completion)bond144 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &idev->mc_lock rcu_read_lock &ul->lock#2 irq_context: 0 (wq_completion)bond144 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &idev->mc_lock rcu_read_lock rcu_read_lock &____s->seqcount#10 irq_context: 0 (wq_completion)bond144 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &idev->mc_lock rcu_read_lock rcu_read_lock rcu_node_0 irq_context: 0 (wq_completion)bond144 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &idev->mc_lock rcu_read_lock rcu_read_lock &rq->__lock irq_context: 0 (wq_completion)bond144 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &idev->mc_lock rcu_read_lock rcu_read_lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)bond144 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &idev->mc_lock rcu_read_lock rcu_read_lock rcu_read_lock_bh rcu_read_lock rcu_read_lock_bh rcu_read_lock &br->multicast_lock irq_context: 0 (wq_completion)bond144 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &idev->mc_lock rcu_read_lock rcu_read_lock rcu_read_lock_bh rcu_read_lock rcu_read_lock_bh rcu_read_lock &br->multicast_lock &obj_hash[i].lock irq_context: 0 (wq_completion)bond144 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &idev->mc_lock rcu_read_lock rcu_read_lock rcu_read_lock_bh rcu_read_lock rcu_read_lock_bh rcu_read_lock &br->multicast_lock &base->lock irq_context: 0 (wq_completion)bond144 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &idev->mc_lock rcu_read_lock rcu_read_lock rcu_read_lock_bh rcu_read_lock rcu_read_lock_bh rcu_read_lock &br->multicast_lock &base->lock &obj_hash[i].lock irq_context: 0 (wq_completion)bond144 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &idev->mc_lock rcu_read_lock rcu_read_lock rcu_read_lock_bh rcu_read_lock rcu_read_lock_bh rcu_read_lock &data->lock irq_context: 0 (wq_completion)bond144 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &idev->mc_lock rcu_read_lock rcu_read_lock rcu_read_lock_bh rcu_read_lock rcu_read_lock_bh rcu_read_lock &obj_hash[i].lock irq_context: 0 (wq_completion)bond144 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &idev->mc_lock rcu_read_lock rcu_node_0 irq_context: 0 (wq_completion)bond144 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex.wait_lock irq_context: 0 (wq_completion)bond144 (work_completion)(&(&bond->mcast_work)->work) &p->pi_lock irq_context: 0 (wq_completion)bond144 (work_completion)(&(&bond->mcast_work)->work) &p->pi_lock &rq->__lock irq_context: 0 (wq_completion)bond144 (work_completion)(&(&bond->mcast_work)->work) &p->pi_lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)bond155 irq_context: 0 (wq_completion)bond155 (work_completion)(&(&slave->notify_work)->work) irq_context: 0 (wq_completion)bond155 (work_completion)(&(&slave->notify_work)->work) &rq->__lock irq_context: 0 (wq_completion)bond155 (work_completion)(&(&slave->notify_work)->work) &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)bond155 (work_completion)(&(&slave->notify_work)->work) &obj_hash[i].lock irq_context: 0 (wq_completion)bond155 (work_completion)(&(&slave->notify_work)->work) &base->lock irq_context: 0 (wq_completion)bond155 (work_completion)(&(&slave->notify_work)->work) &base->lock &obj_hash[i].lock irq_context: 0 (wq_completion)bond132 (work_completion)(&(&bond->mcast_work)->work) &rq->__lock irq_context: 0 (wq_completion)bond132 (work_completion)(&(&bond->mcast_work)->work) &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)mld (work_completion)(&(&idev->mc_ifc_work)->work) &idev->mc_lock rcu_read_lock rcu_read_lock fill_pool_map-wait-type-override rcu_node_0 irq_context: 0 (wq_completion)mld (work_completion)(&(&idev->mc_ifc_work)->work) &idev->mc_lock rcu_read_lock rcu_read_lock fill_pool_map-wait-type-override &rq->__lock irq_context: 0 (wq_completion)mld (work_completion)(&(&idev->mc_ifc_work)->work) &idev->mc_lock rcu_read_lock rcu_read_lock fill_pool_map-wait-type-override &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)bond138 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex fs_reclaim irq_context: 0 (wq_completion)bond138 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex fs_reclaim mmu_notifier_invalidate_range_start irq_context: 0 (wq_completion)bond138 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex batched_entropy_u8.lock irq_context: 0 (wq_completion)bond138 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &rq->__lock irq_context: 0 (wq_completion)bond138 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)bond155 (work_completion)(&(&bond->mcast_work)->work) irq_context: 0 (wq_completion)bond155 (work_completion)(&(&bond->mcast_work)->work) &obj_hash[i].lock irq_context: 0 (wq_completion)bond155 (work_completion)(&(&bond->mcast_work)->work) &base->lock irq_context: 0 (wq_completion)bond155 (work_completion)(&(&bond->mcast_work)->work) &base->lock &obj_hash[i].lock irq_context: 0 (wq_completion)bond155 (work_completion)(&(&bond->mcast_work)->work) &rq->__lock irq_context: 0 (wq_completion)bond155 (work_completion)(&(&bond->mcast_work)->work) &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)bond138 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex kfence_freelist_lock irq_context: 0 (wq_completion)bond138 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex rcu_read_lock &bond->stats_lock/1 irq_context: 0 (wq_completion)bond138 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex rcu_read_lock &ndev->lock irq_context: 0 (wq_completion)bond138 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex nl_table_lock irq_context: 0 (wq_completion)bond138 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &obj_hash[i].lock irq_context: 0 (wq_completion)bond138 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &meta->lock irq_context: 0 (wq_completion)bond138 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex nl_table_wait.lock irq_context: 0 (wq_completion)bond138 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex net_rwsem irq_context: 0 (wq_completion)bond138 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex net_rwsem &list->lock#2 irq_context: 0 (wq_completion)bond138 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &tn->lock irq_context: 0 (wq_completion)bond138 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &idev->mc_lock irq_context: 0 (wq_completion)bond138 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &idev->mc_lock fs_reclaim irq_context: 0 (wq_completion)bond138 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &idev->mc_lock fs_reclaim mmu_notifier_invalidate_range_start irq_context: 0 (wq_completion)bond138 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &idev->mc_lock rcu_read_lock &ndev->lock irq_context: 0 (wq_completion)bond138 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &idev->mc_lock rcu_read_lock &dir->lock#2 irq_context: 0 (wq_completion)bond138 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &idev->mc_lock rcu_read_lock &c->lock irq_context: 0 (wq_completion)bond138 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &idev->mc_lock rcu_read_lock &ul->lock#2 irq_context: 0 (wq_completion)bond138 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &idev->mc_lock rcu_read_lock rcu_read_lock &____s->seqcount#10 irq_context: 0 (wq_completion)bond138 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &idev->mc_lock rcu_read_lock rcu_read_lock rcu_read_lock_bh rcu_read_lock rcu_read_lock_bh rcu_read_lock &br->multicast_lock irq_context: 0 (wq_completion)bond138 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &idev->mc_lock rcu_read_lock rcu_read_lock rcu_read_lock_bh rcu_read_lock rcu_read_lock_bh rcu_read_lock &br->multicast_lock &obj_hash[i].lock irq_context: 0 (wq_completion)bond138 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &idev->mc_lock rcu_read_lock rcu_read_lock rcu_read_lock_bh rcu_read_lock rcu_read_lock_bh rcu_read_lock &br->multicast_lock &base->lock irq_context: 0 (wq_completion)bond138 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &idev->mc_lock rcu_read_lock rcu_read_lock rcu_read_lock_bh rcu_read_lock rcu_read_lock_bh rcu_read_lock &data->lock irq_context: 0 (wq_completion)bond138 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &idev->mc_lock rcu_read_lock rcu_read_lock rcu_read_lock_bh rcu_read_lock rcu_read_lock_bh rcu_read_lock &obj_hash[i].lock irq_context: 0 (wq_completion)bond138 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &idev->mc_lock rcu_read_lock rcu_read_lock rcu_node_0 irq_context: 0 (wq_completion)bond138 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &idev->mc_lock rcu_read_lock rcu_read_lock &rq->__lock irq_context: 0 (wq_completion)bond138 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &idev->mc_lock rcu_read_lock rcu_read_lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)bond138 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &idev->mc_lock rcu_read_lock rcu_node_0 irq_context: 0 (wq_completion)bond138 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex.wait_lock irq_context: 0 (wq_completion)bond138 (work_completion)(&(&bond->mcast_work)->work) &p->pi_lock irq_context: 0 (wq_completion)bond138 (work_completion)(&(&bond->mcast_work)->work) &p->pi_lock &rq->__lock irq_context: 0 (wq_completion)bond138 (work_completion)(&(&bond->mcast_work)->work) &p->pi_lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)ipv6_addrconf (work_completion)(&(&ifa->dad_work)->work) rtnl_mutex &tb->tb6_lock nl_table_wait.lock &p->pi_lock &rq->__lock irq_context: 0 (wq_completion)ipv6_addrconf (work_completion)(&(&ifa->dad_work)->work) rtnl_mutex &tb->tb6_lock nl_table_wait.lock &p->pi_lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)bond115#3 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex fs_reclaim irq_context: 0 (wq_completion)bond115#3 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex fs_reclaim mmu_notifier_invalidate_range_start irq_context: 0 (wq_completion)bond115#3 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex rcu_read_lock &bond->stats_lock/1 irq_context: 0 (wq_completion)bond115#3 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex rcu_read_lock &ndev->lock irq_context: 0 (wq_completion)bond115#3 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex nl_table_lock irq_context: 0 (wq_completion)bond115#3 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &obj_hash[i].lock irq_context: 0 (wq_completion)bond115#3 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex nl_table_wait.lock irq_context: 0 (wq_completion)bond115#3 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex net_rwsem irq_context: 0 (wq_completion)bond115#3 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex net_rwsem &list->lock#2 irq_context: 0 (wq_completion)bond115#3 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &tn->lock irq_context: 0 (wq_completion)bond115#3 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &idev->mc_lock irq_context: 0 (wq_completion)bond115#3 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &idev->mc_lock &rq->__lock irq_context: 0 (wq_completion)bond115#3 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &idev->mc_lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)bond115#3 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &idev->mc_lock fs_reclaim irq_context: 0 (wq_completion)bond115#3 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &idev->mc_lock fs_reclaim mmu_notifier_invalidate_range_start irq_context: 0 (wq_completion)bond115#3 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &idev->mc_lock rcu_read_lock &ndev->lock irq_context: 0 (wq_completion)bond115#3 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &idev->mc_lock rcu_read_lock &dir->lock#2 irq_context: 0 (wq_completion)bond115#3 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &idev->mc_lock rcu_read_lock &ul->lock#2 irq_context: 0 (wq_completion)bond115#3 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &idev->mc_lock rcu_read_lock rcu_read_lock &____s->seqcount#10 irq_context: 0 (wq_completion)bond115#3 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &idev->mc_lock rcu_read_lock rcu_read_lock rcu_read_lock_bh rcu_read_lock rcu_read_lock_bh rcu_read_lock &br->multicast_lock irq_context: 0 (wq_completion)bond115#3 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &idev->mc_lock rcu_read_lock rcu_read_lock rcu_read_lock_bh rcu_read_lock rcu_read_lock_bh rcu_read_lock &br->multicast_lock &obj_hash[i].lock irq_context: 0 (wq_completion)bond115#3 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &idev->mc_lock rcu_read_lock rcu_read_lock rcu_read_lock_bh rcu_read_lock rcu_read_lock_bh rcu_read_lock &br->multicast_lock &base->lock irq_context: 0 (wq_completion)bond115#3 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &idev->mc_lock rcu_read_lock rcu_read_lock rcu_read_lock_bh rcu_read_lock rcu_read_lock_bh rcu_read_lock &br->multicast_lock &base->lock &obj_hash[i].lock irq_context: 0 (wq_completion)bond115#3 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &idev->mc_lock rcu_read_lock rcu_read_lock rcu_read_lock_bh rcu_read_lock rcu_read_lock_bh rcu_read_lock &data->lock irq_context: 0 (wq_completion)bond115#3 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &idev->mc_lock rcu_read_lock rcu_read_lock rcu_read_lock_bh rcu_read_lock rcu_read_lock_bh rcu_read_lock &obj_hash[i].lock irq_context: 0 (wq_completion)bond115#3 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &idev->mc_lock rcu_read_lock rcu_read_lock rcu_node_0 irq_context: 0 (wq_completion)bond115#3 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &idev->mc_lock rcu_read_lock rcu_read_lock &rq->__lock irq_context: 0 (wq_completion)bond115#3 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &idev->mc_lock rcu_read_lock rcu_read_lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)bond115#3 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &idev->mc_lock rcu_read_lock rcu_node_0 irq_context: 0 (wq_completion)bond115#3 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex.wait_lock irq_context: 0 (wq_completion)bond115#3 (work_completion)(&(&bond->mcast_work)->work) &p->pi_lock irq_context: 0 (wq_completion)bond115#3 (work_completion)(&(&slave->notify_work)->work) rtnl_mutex net_rwsem irq_context: 0 (wq_completion)bond115#3 (work_completion)(&(&slave->notify_work)->work) rtnl_mutex net_rwsem &list->lock#2 irq_context: 0 (wq_completion)bond115#3 (work_completion)(&(&slave->notify_work)->work) rtnl_mutex &tn->lock irq_context: 0 (wq_completion)bond115#3 (work_completion)(&(&slave->notify_work)->work) rtnl_mutex.wait_lock irq_context: 0 (wq_completion)bond115#3 (work_completion)(&(&slave->notify_work)->work) &p->pi_lock irq_context: 0 (wq_completion)bond32#3 (work_completion)(&(&bond->mcast_work)->work) fill_pool_map-wait-type-override &c->lock irq_context: 0 (wq_completion)bond32#3 (work_completion)(&(&bond->mcast_work)->work) fill_pool_map-wait-type-override &____s->seqcount#2 irq_context: 0 (wq_completion)bond32#3 (work_completion)(&(&bond->mcast_work)->work) fill_pool_map-wait-type-override &____s->seqcount irq_context: 0 (wq_completion)bond32#3 (work_completion)(&(&bond->mcast_work)->work) fill_pool_map-wait-type-override pool_lock irq_context: 0 (wq_completion)bond127#3 irq_context: 0 (wq_completion)bond127#3 &rq->__lock irq_context: 0 (wq_completion)bond127#3 (work_completion)(&(&slave->notify_work)->work) irq_context: 0 (wq_completion)bond127#3 (work_completion)(&(&slave->notify_work)->work) &obj_hash[i].lock irq_context: 0 (wq_completion)bond127#3 (work_completion)(&(&slave->notify_work)->work) &base->lock irq_context: 0 (wq_completion)bond127#3 (work_completion)(&(&slave->notify_work)->work) &base->lock &obj_hash[i].lock irq_context: 0 (wq_completion)bond127#3 (work_completion)(&(&slave->notify_work)->work) &rq->__lock irq_context: 0 (wq_completion)bond127#3 (work_completion)(&(&slave->notify_work)->work) &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)bond127#3 (work_completion)(&(&bond->mcast_work)->work) irq_context: 0 (wq_completion)bond127#3 (work_completion)(&(&bond->mcast_work)->work) &rq->__lock irq_context: 0 (wq_completion)bond127#3 (work_completion)(&(&bond->mcast_work)->work) &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)bond127#3 (work_completion)(&(&bond->mcast_work)->work) &obj_hash[i].lock irq_context: 0 (wq_completion)bond127#3 (work_completion)(&(&bond->mcast_work)->work) &base->lock irq_context: 0 (wq_completion)bond127#3 (work_completion)(&(&bond->mcast_work)->work) &base->lock &obj_hash[i].lock irq_context: 0 (wq_completion)bond123 &rq->__lock irq_context: 0 (wq_completion)bond123 &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)bond26#4 (work_completion)(&(&slave->notify_work)->work) rtnl_mutex net_rwsem irq_context: 0 (wq_completion)bond26#4 (work_completion)(&(&slave->notify_work)->work) rtnl_mutex net_rwsem &list->lock#2 irq_context: 0 (wq_completion)bond26#4 (work_completion)(&(&slave->notify_work)->work) rtnl_mutex &tn->lock irq_context: 0 (wq_completion)bond26#4 (work_completion)(&(&slave->notify_work)->work) rtnl_mutex.wait_lock irq_context: 0 (wq_completion)bond26#4 (work_completion)(&(&slave->notify_work)->work) &p->pi_lock irq_context: 0 (wq_completion)bond37#2 irq_context: 0 (wq_completion)bond37#2 &rq->__lock irq_context: 0 (wq_completion)bond37#2 &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)bond37#2 (work_completion)(&(&slave->notify_work)->work) irq_context: 0 (wq_completion)bond37#2 (work_completion)(&(&slave->notify_work)->work) &obj_hash[i].lock irq_context: 0 (wq_completion)bond37#2 (work_completion)(&(&slave->notify_work)->work) &base->lock irq_context: 0 (wq_completion)bond37#2 (work_completion)(&(&slave->notify_work)->work) &base->lock &obj_hash[i].lock irq_context: 0 (wq_completion)bond37#2 (work_completion)(&(&bond->mcast_work)->work) irq_context: 0 (wq_completion)bond37#2 (work_completion)(&(&bond->mcast_work)->work) &obj_hash[i].lock irq_context: 0 (wq_completion)bond37#2 (work_completion)(&(&bond->mcast_work)->work) &base->lock irq_context: 0 (wq_completion)bond37#2 (work_completion)(&(&bond->mcast_work)->work) &base->lock &obj_hash[i].lock irq_context: 0 (wq_completion)bond37#2 (work_completion)(&(&bond->mcast_work)->work) &rq->__lock irq_context: 0 (wq_completion)bond119#3 (work_completion)(&(&slave->notify_work)->work) &rq->__lock irq_context: 0 (wq_completion)bond119#3 (work_completion)(&(&slave->notify_work)->work) &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)bond124#3 &rq->__lock irq_context: 0 (wq_completion)bond124#3 &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: softirq (&ndev->rs_timer) rcu_read_lock rcu_read_lock &pcp->lock &zone->lock &____s->seqcount irq_context: 0 (wq_completion)bond143#2 irq_context: 0 (wq_completion)bond143#2 (work_completion)(&(&slave->notify_work)->work) irq_context: 0 (wq_completion)bond143#2 (work_completion)(&(&slave->notify_work)->work) &obj_hash[i].lock irq_context: 0 (wq_completion)bond143#2 (work_completion)(&(&slave->notify_work)->work) &base->lock irq_context: 0 (wq_completion)bond143#2 (work_completion)(&(&slave->notify_work)->work) &base->lock &obj_hash[i].lock irq_context: 0 (wq_completion)bond143#2 (work_completion)(&(&bond->mcast_work)->work) irq_context: 0 (wq_completion)bond143#2 (work_completion)(&(&bond->mcast_work)->work) &obj_hash[i].lock irq_context: 0 (wq_completion)bond143#2 (work_completion)(&(&bond->mcast_work)->work) &base->lock irq_context: 0 (wq_completion)bond143#2 (work_completion)(&(&bond->mcast_work)->work) &base->lock &obj_hash[i].lock irq_context: 0 (wq_completion)bond143#2 (work_completion)(&(&bond->mcast_work)->work) &rq->__lock irq_context: 0 (wq_completion)bond156 irq_context: 0 (wq_completion)bond156 (work_completion)(&(&slave->notify_work)->work) irq_context: 0 (wq_completion)bond156 (work_completion)(&(&slave->notify_work)->work) &obj_hash[i].lock irq_context: 0 (wq_completion)bond156 (work_completion)(&(&slave->notify_work)->work) &base->lock irq_context: 0 (wq_completion)bond156 (work_completion)(&(&slave->notify_work)->work) &base->lock &obj_hash[i].lock irq_context: 0 (wq_completion)bond156 (work_completion)(&(&bond->mcast_work)->work) irq_context: 0 (wq_completion)bond156 (work_completion)(&(&bond->mcast_work)->work) &obj_hash[i].lock irq_context: 0 (wq_completion)bond156 (work_completion)(&(&bond->mcast_work)->work) &base->lock irq_context: 0 (wq_completion)bond156 (work_completion)(&(&bond->mcast_work)->work) &base->lock &obj_hash[i].lock irq_context: 0 (wq_completion)bond156 (work_completion)(&(&bond->mcast_work)->work) &rq->__lock irq_context: 0 (wq_completion)bond156 (work_completion)(&(&bond->mcast_work)->work) &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)bond154 (work_completion)(&(&bond->mcast_work)->work) &rq->__lock irq_context: 0 (wq_completion)bond154 (work_completion)(&(&bond->mcast_work)->work) &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)bond108#3 (work_completion)(&(&bond->mcast_work)->work) &rq->__lock irq_context: 0 (wq_completion)bond108#3 (work_completion)(&(&bond->mcast_work)->work) &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)bond128#3 irq_context: softirq &(&bond->mcast_work)->timer rcu_read_lock &pool->lock/1 fill_pool_map-wait-type-override &c->lock irq_context: softirq &(&bond->mcast_work)->timer rcu_read_lock &pool->lock/1 fill_pool_map-wait-type-override &____s->seqcount#2 irq_context: softirq &(&bond->mcast_work)->timer rcu_read_lock &pool->lock/1 fill_pool_map-wait-type-override &____s->seqcount irq_context: 0 (wq_completion)bond128#3 &rq->__lock irq_context: 0 (wq_completion)bond128#3 &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)bond105#3 &rq->__lock irq_context: 0 (wq_completion)bond105#3 &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)bond128#3 (work_completion)(&(&slave->notify_work)->work) irq_context: 0 (wq_completion)bond128#3 (work_completion)(&(&slave->notify_work)->work) &obj_hash[i].lock irq_context: 0 (wq_completion)bond128#3 (work_completion)(&(&slave->notify_work)->work) &base->lock irq_context: 0 (wq_completion)bond128#3 (work_completion)(&(&slave->notify_work)->work) &base->lock &obj_hash[i].lock irq_context: 0 (wq_completion)bond128#3 (work_completion)(&(&slave->notify_work)->work) &rq->__lock irq_context: 0 (wq_completion)bond128#3 (work_completion)(&(&slave->notify_work)->work) &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)bond128#3 (work_completion)(&(&bond->mcast_work)->work) irq_context: 0 (wq_completion)bond128#3 (work_completion)(&(&bond->mcast_work)->work) &obj_hash[i].lock irq_context: 0 (wq_completion)bond128#3 (work_completion)(&(&bond->mcast_work)->work) &base->lock irq_context: 0 (wq_completion)bond128#3 (work_completion)(&(&bond->mcast_work)->work) &base->lock &obj_hash[i].lock irq_context: 0 (wq_completion)bond135#2 (work_completion)(&(&slave->notify_work)->work) fill_pool_map-wait-type-override pool_lock irq_context: 0 (wq_completion)bond148 (work_completion)(&(&bond->mcast_work)->work) fill_pool_map-wait-type-override pool_lock irq_context: 0 (wq_completion)bond128#3 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex fs_reclaim irq_context: 0 (wq_completion)bond128#3 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex fs_reclaim mmu_notifier_invalidate_range_start irq_context: 0 (wq_completion)bond128#3 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex rcu_read_lock &bond->stats_lock/1 irq_context: 0 (wq_completion)bond128#3 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex rcu_read_lock &ndev->lock irq_context: 0 (wq_completion)bond128#3 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex nl_table_lock irq_context: 0 (wq_completion)bond128#3 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &obj_hash[i].lock irq_context: 0 (wq_completion)bond128#3 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex nl_table_wait.lock irq_context: 0 (wq_completion)bond128#3 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex net_rwsem irq_context: 0 (wq_completion)bond128#3 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex net_rwsem &list->lock#2 irq_context: 0 (wq_completion)bond128#3 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &tn->lock irq_context: 0 (wq_completion)bond128#3 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &idev->mc_lock irq_context: 0 (wq_completion)bond128#3 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &idev->mc_lock fs_reclaim irq_context: 0 (wq_completion)bond128#3 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &idev->mc_lock fs_reclaim mmu_notifier_invalidate_range_start irq_context: 0 (wq_completion)bond128#3 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &idev->mc_lock &c->lock irq_context: 0 (wq_completion)bond128#3 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &idev->mc_lock rcu_read_lock &ndev->lock irq_context: 0 (wq_completion)bond128#3 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &idev->mc_lock rcu_read_lock &dir->lock#2 irq_context: 0 (wq_completion)bond128#3 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &idev->mc_lock rcu_read_lock &ul->lock#2 irq_context: 0 (wq_completion)bond128#3 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &idev->mc_lock rcu_read_lock rcu_read_lock &obj_hash[i].lock irq_context: 0 (wq_completion)bond128#3 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &idev->mc_lock rcu_read_lock rcu_read_lock &dir->lock#2 irq_context: 0 (wq_completion)bond128#3 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &idev->mc_lock rcu_read_lock rcu_read_lock &tbl->lock irq_context: 0 (wq_completion)bond128#3 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &idev->mc_lock rcu_read_lock rcu_read_lock &n->lock irq_context: 0 (wq_completion)bond128#3 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &idev->mc_lock rcu_read_lock rcu_read_lock &____s->seqcount#9 irq_context: 0 (wq_completion)bond128#3 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &idev->mc_lock rcu_read_lock rcu_read_lock rcu_read_lock_bh rcu_read_lock rcu_read_lock_bh rcu_read_lock &br->multicast_lock irq_context: 0 (wq_completion)bond128#3 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &idev->mc_lock rcu_read_lock rcu_read_lock rcu_read_lock_bh rcu_read_lock rcu_read_lock_bh rcu_read_lock &br->multicast_lock &obj_hash[i].lock irq_context: 0 (wq_completion)bond128#3 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &idev->mc_lock rcu_read_lock rcu_read_lock rcu_read_lock_bh rcu_read_lock rcu_read_lock_bh rcu_read_lock &br->multicast_lock &obj_hash[i].lock pool_lock irq_context: 0 (wq_completion)bond128#3 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &idev->mc_lock rcu_read_lock rcu_read_lock rcu_read_lock_bh rcu_read_lock rcu_read_lock_bh rcu_read_lock &br->multicast_lock rcu_read_lock rhashtable_bucket irq_context: 0 (wq_completion)bond128#3 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &idev->mc_lock rcu_read_lock rcu_read_lock rcu_read_lock_bh rcu_read_lock rcu_read_lock_bh rcu_read_lock &br->multicast_lock nl_table_lock irq_context: 0 (wq_completion)bond128#3 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &idev->mc_lock rcu_read_lock rcu_read_lock rcu_read_lock_bh rcu_read_lock rcu_read_lock_bh rcu_read_lock &br->multicast_lock nl_table_wait.lock irq_context: 0 (wq_completion)bond128#3 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &idev->mc_lock rcu_read_lock rcu_read_lock rcu_read_lock_bh rcu_read_lock rcu_read_lock_bh rcu_read_lock &br->multicast_lock &base->lock irq_context: 0 (wq_completion)bond128#3 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &idev->mc_lock rcu_read_lock rcu_read_lock rcu_read_lock_bh rcu_read_lock rcu_read_lock_bh rcu_read_lock &br->multicast_lock &base->lock &obj_hash[i].lock irq_context: 0 (wq_completion)bond128#3 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &idev->mc_lock rcu_read_lock rcu_read_lock rcu_read_lock_bh rcu_read_lock rcu_read_lock_bh rcu_read_lock &data->lock irq_context: 0 (wq_completion)bond128#3 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &idev->mc_lock rcu_read_lock rcu_read_lock rcu_read_lock_bh rcu_read_lock rcu_read_lock_bh rcu_read_lock &obj_hash[i].lock irq_context: 0 (wq_completion)bond128#3 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex.wait_lock irq_context: 0 (wq_completion)bond128#3 (work_completion)(&(&bond->mcast_work)->work) &p->pi_lock irq_context: 0 (wq_completion)bond128#3 (work_completion)(&(&slave->notify_work)->work) rtnl_mutex net_rwsem irq_context: 0 (wq_completion)bond128#3 (work_completion)(&(&slave->notify_work)->work) rtnl_mutex net_rwsem &list->lock#2 irq_context: 0 (wq_completion)bond128#3 (work_completion)(&(&slave->notify_work)->work) rtnl_mutex &tn->lock irq_context: 0 (wq_completion)bond128#3 (work_completion)(&(&slave->notify_work)->work) rtnl_mutex.wait_lock irq_context: 0 (wq_completion)bond128#3 (work_completion)(&(&slave->notify_work)->work) &p->pi_lock irq_context: 0 (wq_completion)bond144#2 irq_context: 0 (wq_completion)bond144#2 (work_completion)(&(&slave->notify_work)->work) irq_context: 0 (wq_completion)bond144#2 (work_completion)(&(&slave->notify_work)->work) &obj_hash[i].lock irq_context: 0 (wq_completion)bond144#2 (work_completion)(&(&slave->notify_work)->work) &base->lock irq_context: 0 (wq_completion)bond144#2 (work_completion)(&(&slave->notify_work)->work) &base->lock &obj_hash[i].lock irq_context: 0 (wq_completion)bond144#2 (work_completion)(&(&slave->notify_work)->work) &rq->__lock irq_context: 0 (wq_completion)bond144#2 (work_completion)(&(&bond->mcast_work)->work) irq_context: 0 (wq_completion)bond144#2 (work_completion)(&(&bond->mcast_work)->work) &rq->__lock irq_context: 0 (wq_completion)bond144#2 (work_completion)(&(&bond->mcast_work)->work) &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)bond144#2 (work_completion)(&(&bond->mcast_work)->work) &obj_hash[i].lock irq_context: 0 (wq_completion)bond144#2 (work_completion)(&(&bond->mcast_work)->work) &base->lock irq_context: 0 (wq_completion)bond144#2 (work_completion)(&(&bond->mcast_work)->work) &base->lock &obj_hash[i].lock irq_context: 0 (wq_completion)bond18#3 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex fs_reclaim irq_context: 0 (wq_completion)bond18#3 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex fs_reclaim mmu_notifier_invalidate_range_start irq_context: 0 (wq_completion)bond18#3 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex rcu_read_lock &bond->stats_lock/1 irq_context: 0 (wq_completion)bond18#3 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex rcu_read_lock &ndev->lock irq_context: 0 (wq_completion)bond18#3 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex nl_table_lock irq_context: 0 (wq_completion)bond18#3 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &obj_hash[i].lock irq_context: 0 (wq_completion)bond18#3 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex nl_table_wait.lock irq_context: 0 (wq_completion)bond18#3 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex net_rwsem irq_context: 0 (wq_completion)bond18#3 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex net_rwsem &list->lock#2 irq_context: 0 (wq_completion)bond18#3 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &tn->lock irq_context: 0 (wq_completion)bond18#3 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &idev->mc_lock irq_context: 0 (wq_completion)bond18#3 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &idev->mc_lock fs_reclaim irq_context: 0 (wq_completion)bond18#3 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &idev->mc_lock fs_reclaim mmu_notifier_invalidate_range_start irq_context: 0 (wq_completion)bond18#3 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &idev->mc_lock &c->lock irq_context: 0 (wq_completion)bond18#3 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &idev->mc_lock &n->list_lock irq_context: 0 (wq_completion)bond18#3 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &idev->mc_lock &n->list_lock &c->lock irq_context: 0 (wq_completion)bond18#3 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &idev->mc_lock rcu_read_lock &ndev->lock irq_context: 0 (wq_completion)bond18#3 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &idev->mc_lock rcu_read_lock &dir->lock#2 irq_context: 0 (wq_completion)bond18#3 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &idev->mc_lock rcu_read_lock &c->lock irq_context: 0 (wq_completion)bond18#3 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &idev->mc_lock rcu_read_lock &ul->lock#2 irq_context: 0 (wq_completion)bond18#3 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &idev->mc_lock rcu_read_lock rcu_read_lock &____s->seqcount#12 irq_context: 0 (wq_completion)bond18#3 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &idev->mc_lock rcu_read_lock rcu_read_lock &____s->seqcount#16 irq_context: 0 (wq_completion)bond18#3 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &idev->mc_lock rcu_read_lock &obj_hash[i].lock irq_context: 0 (wq_completion)bond18#3 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &idev->mc_lock rcu_read_lock &data->lock irq_context: 0 (wq_completion)bond18#3 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex.wait_lock irq_context: 0 (wq_completion)bond18#3 (work_completion)(&(&bond->mcast_work)->work) &p->pi_lock irq_context: 0 (wq_completion)bond100#3 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex fs_reclaim irq_context: 0 (wq_completion)bond100#3 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex fs_reclaim mmu_notifier_invalidate_range_start irq_context: 0 (wq_completion)bond100#3 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex rcu_read_lock &bond->stats_lock/1 irq_context: 0 (wq_completion)bond100#3 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex rcu_read_lock &ndev->lock irq_context: 0 (wq_completion)bond100#3 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex nl_table_lock irq_context: 0 (wq_completion)bond100#3 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &obj_hash[i].lock irq_context: 0 (wq_completion)bond100#3 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex nl_table_wait.lock irq_context: 0 (wq_completion)bond100#3 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex net_rwsem irq_context: 0 (wq_completion)bond100#3 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex net_rwsem &list->lock#2 irq_context: 0 (wq_completion)bond100#3 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &tn->lock irq_context: 0 (wq_completion)bond100#3 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &idev->mc_lock irq_context: 0 (wq_completion)bond100#3 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &idev->mc_lock fs_reclaim irq_context: 0 (wq_completion)bond100#3 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &idev->mc_lock fs_reclaim mmu_notifier_invalidate_range_start irq_context: 0 (wq_completion)bond100#3 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &idev->mc_lock &c->lock irq_context: 0 (wq_completion)bond100#3 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &idev->mc_lock rcu_read_lock &ndev->lock irq_context: 0 (wq_completion)bond100#3 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &idev->mc_lock rcu_read_lock &c->lock irq_context: 0 (wq_completion)bond100#3 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &idev->mc_lock rcu_read_lock &dir->lock#2 irq_context: 0 (wq_completion)bond100#3 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &idev->mc_lock rcu_read_lock &ul->lock#2 irq_context: 0 (wq_completion)bond100#3 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &idev->mc_lock rcu_read_lock rcu_read_lock &____s->seqcount#10 irq_context: 0 (wq_completion)bond100#3 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &idev->mc_lock rcu_read_lock rcu_read_lock rcu_read_lock_bh rcu_read_lock rcu_read_lock_bh rcu_read_lock &br->multicast_lock irq_context: 0 (wq_completion)bond100#3 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &idev->mc_lock rcu_read_lock rcu_read_lock rcu_read_lock_bh rcu_read_lock rcu_read_lock_bh rcu_read_lock &br->multicast_lock &obj_hash[i].lock irq_context: 0 (wq_completion)bond100#3 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &idev->mc_lock rcu_read_lock rcu_read_lock rcu_read_lock_bh rcu_read_lock rcu_read_lock_bh rcu_read_lock &br->multicast_lock &base->lock irq_context: 0 (wq_completion)bond100#3 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &idev->mc_lock rcu_read_lock rcu_read_lock rcu_read_lock_bh rcu_read_lock rcu_read_lock_bh rcu_read_lock &br->multicast_lock &base->lock &obj_hash[i].lock irq_context: 0 (wq_completion)bond100#3 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &idev->mc_lock rcu_read_lock rcu_read_lock rcu_read_lock_bh rcu_read_lock rcu_read_lock_bh rcu_read_lock &data->lock irq_context: 0 (wq_completion)bond100#3 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &idev->mc_lock rcu_read_lock rcu_read_lock rcu_read_lock_bh rcu_read_lock rcu_read_lock_bh rcu_read_lock &obj_hash[i].lock irq_context: 0 (wq_completion)bond100#3 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &idev->mc_lock rcu_read_lock rcu_read_lock rcu_node_0 irq_context: 0 (wq_completion)bond100#3 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &idev->mc_lock rcu_read_lock rcu_read_lock &rq->__lock irq_context: 0 (wq_completion)bond100#3 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &idev->mc_lock rcu_read_lock rcu_read_lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)bond100#3 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &idev->mc_lock rcu_read_lock rcu_node_0 irq_context: 0 (wq_completion)bond100#3 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex.wait_lock irq_context: 0 (wq_completion)bond100#3 (work_completion)(&(&bond->mcast_work)->work) &p->pi_lock irq_context: 0 (wq_completion)bond100#3 (work_completion)(&(&bond->mcast_work)->work) &p->pi_lock &rq->__lock irq_context: 0 (wq_completion)bond100#3 (work_completion)(&(&bond->mcast_work)->work) &p->pi_lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)bond157 irq_context: 0 (wq_completion)bond157 (work_completion)(&(&slave->notify_work)->work) irq_context: 0 (wq_completion)bond157 (work_completion)(&(&slave->notify_work)->work) &obj_hash[i].lock irq_context: 0 (wq_completion)bond157 (work_completion)(&(&slave->notify_work)->work) &base->lock irq_context: 0 (wq_completion)bond157 (work_completion)(&(&slave->notify_work)->work) &base->lock &obj_hash[i].lock irq_context: 0 (wq_completion)bond157 (work_completion)(&(&slave->notify_work)->work) &rq->__lock irq_context: 0 (wq_completion)bond157 (work_completion)(&(&slave->notify_work)->work) &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)bond157 (work_completion)(&(&bond->mcast_work)->work) irq_context: 0 (wq_completion)bond157 (work_completion)(&(&bond->mcast_work)->work) &obj_hash[i].lock irq_context: 0 (wq_completion)bond157 (work_completion)(&(&bond->mcast_work)->work) &base->lock irq_context: 0 (wq_completion)bond157 (work_completion)(&(&bond->mcast_work)->work) &base->lock &obj_hash[i].lock irq_context: 0 (wq_completion)bond99#2 &rq->__lock irq_context: 0 (wq_completion)bond99#2 &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)bond144#2 (work_completion)(&(&slave->notify_work)->work) &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)bond99#2 (work_completion)(&(&slave->notify_work)->work) &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)bond117#3 (work_completion)(&(&bond->mcast_work)->work) &rq->__lock irq_context: 0 (wq_completion)bond117#3 (work_completion)(&(&bond->mcast_work)->work) &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)bond38#2 irq_context: 0 (wq_completion)bond38#2 &rq->__lock irq_context: 0 (wq_completion)bond38#2 &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)bond38#2 (work_completion)(&(&slave->notify_work)->work) irq_context: 0 (wq_completion)bond38#2 (work_completion)(&(&slave->notify_work)->work) &obj_hash[i].lock irq_context: 0 (wq_completion)bond38#2 (work_completion)(&(&slave->notify_work)->work) &base->lock irq_context: 0 (wq_completion)bond38#2 (work_completion)(&(&slave->notify_work)->work) &base->lock &obj_hash[i].lock irq_context: 0 (wq_completion)bond38#2 (work_completion)(&(&slave->notify_work)->work) &rq->__lock irq_context: 0 (wq_completion)bond143 (work_completion)(&(&slave->notify_work)->work) &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)bond38#2 (work_completion)(&(&bond->mcast_work)->work) irq_context: 0 (wq_completion)bond38#2 (work_completion)(&(&bond->mcast_work)->work) &obj_hash[i].lock irq_context: 0 (wq_completion)bond38#2 (work_completion)(&(&bond->mcast_work)->work) &base->lock irq_context: 0 (wq_completion)bond38#2 (work_completion)(&(&bond->mcast_work)->work) &base->lock &obj_hash[i].lock irq_context: 0 (wq_completion)bond38#2 (work_completion)(&(&bond->mcast_work)->work) &rq->__lock irq_context: 0 (wq_completion)bond38#2 (work_completion)(&(&bond->mcast_work)->work) &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)bond127#3 (work_completion)(&(&bond->mcast_work)->work) fill_pool_map-wait-type-override pool_lock irq_context: 0 (wq_completion)bond129#3 irq_context: 0 (wq_completion)bond129#3 (work_completion)(&(&slave->notify_work)->work) irq_context: 0 (wq_completion)bond129#3 (work_completion)(&(&slave->notify_work)->work) &obj_hash[i].lock irq_context: 0 (wq_completion)bond129#3 (work_completion)(&(&slave->notify_work)->work) &base->lock irq_context: 0 (wq_completion)bond129#3 (work_completion)(&(&slave->notify_work)->work) &base->lock &obj_hash[i].lock irq_context: 0 (wq_completion)bond129#3 (work_completion)(&(&bond->mcast_work)->work) irq_context: 0 (wq_completion)bond129#3 (work_completion)(&(&bond->mcast_work)->work) &obj_hash[i].lock irq_context: 0 (wq_completion)bond129#3 (work_completion)(&(&bond->mcast_work)->work) &base->lock irq_context: 0 (wq_completion)bond129#3 (work_completion)(&(&bond->mcast_work)->work) &base->lock &obj_hash[i].lock irq_context: 0 (wq_completion)bond129#3 (work_completion)(&(&bond->mcast_work)->work) &rq->__lock irq_context: 0 (wq_completion)bond129#3 (work_completion)(&(&bond->mcast_work)->work) &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)bond129#3 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex fs_reclaim irq_context: 0 (wq_completion)bond129#3 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex fs_reclaim mmu_notifier_invalidate_range_start irq_context: 0 (wq_completion)bond129#3 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex fs_reclaim mmu_notifier_invalidate_range_start &rq->__lock irq_context: 0 (wq_completion)bond129#3 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex fs_reclaim mmu_notifier_invalidate_range_start &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)bond129#3 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &c->lock irq_context: 0 (wq_completion)bond129#3 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex rcu_read_lock &bond->stats_lock/1 irq_context: 0 (wq_completion)bond129#3 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex rcu_read_lock &ndev->lock irq_context: 0 (wq_completion)bond129#3 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex nl_table_lock irq_context: 0 (wq_completion)bond129#3 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &obj_hash[i].lock irq_context: 0 (wq_completion)bond129#3 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex nl_table_wait.lock irq_context: 0 (wq_completion)bond129#3 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex net_rwsem irq_context: 0 (wq_completion)bond129#3 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex net_rwsem &list->lock#2 irq_context: 0 (wq_completion)bond129#3 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &tn->lock irq_context: 0 (wq_completion)bond129#3 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &idev->mc_lock irq_context: 0 (wq_completion)bond129#3 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &rq->__lock irq_context: 0 (wq_completion)bond129#3 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)bond129#3 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex.wait_lock irq_context: 0 (wq_completion)bond129#3 (work_completion)(&(&bond->mcast_work)->work) &p->pi_lock irq_context: 0 (wq_completion)bond129#3 (work_completion)(&(&slave->notify_work)->work) rtnl_mutex net_rwsem irq_context: 0 (wq_completion)bond129#3 (work_completion)(&(&slave->notify_work)->work) rtnl_mutex net_rwsem &list->lock#2 irq_context: 0 (wq_completion)bond129#3 (work_completion)(&(&slave->notify_work)->work) rtnl_mutex &tn->lock irq_context: 0 (wq_completion)bond129#3 (work_completion)(&(&slave->notify_work)->work) rtnl_mutex.wait_lock irq_context: 0 (wq_completion)bond129#3 (work_completion)(&(&slave->notify_work)->work) &rq->__lock irq_context: 0 (wq_completion)bond129#3 (work_completion)(&(&slave->notify_work)->work) &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)bond129#3 (work_completion)(&(&slave->notify_work)->work) &p->pi_lock irq_context: 0 (wq_completion)bond145#2 irq_context: 0 (wq_completion)bond145#2 (work_completion)(&(&slave->notify_work)->work) irq_context: 0 (wq_completion)bond145#2 (work_completion)(&(&slave->notify_work)->work) &obj_hash[i].lock irq_context: 0 (wq_completion)bond145#2 (work_completion)(&(&slave->notify_work)->work) &base->lock irq_context: 0 (wq_completion)bond145#2 (work_completion)(&(&slave->notify_work)->work) &base->lock &obj_hash[i].lock irq_context: 0 (wq_completion)bond145#2 (work_completion)(&(&slave->notify_work)->work) &rq->__lock irq_context: 0 (wq_completion)bond133#2 &rq->__lock irq_context: 0 (wq_completion)bond133#2 &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)bond145#2 (work_completion)(&(&bond->mcast_work)->work) irq_context: 0 (wq_completion)bond145#2 (work_completion)(&(&bond->mcast_work)->work) &obj_hash[i].lock irq_context: 0 (wq_completion)bond145#2 (work_completion)(&(&bond->mcast_work)->work) &base->lock irq_context: 0 (wq_completion)bond145#2 (work_completion)(&(&bond->mcast_work)->work) &base->lock &obj_hash[i].lock irq_context: 0 (wq_completion)bond23#3 (work_completion)(&(&slave->notify_work)->work) &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)bond158 irq_context: 0 (wq_completion)bond158 &rq->__lock irq_context: 0 (wq_completion)bond158 &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)bond158 (work_completion)(&(&slave->notify_work)->work) irq_context: 0 (wq_completion)bond158 (work_completion)(&(&slave->notify_work)->work) &obj_hash[i].lock irq_context: 0 (wq_completion)bond158 (work_completion)(&(&slave->notify_work)->work) &base->lock irq_context: 0 (wq_completion)bond158 (work_completion)(&(&slave->notify_work)->work) &base->lock &obj_hash[i].lock irq_context: 0 (wq_completion)mld (work_completion)(&(&idev->mc_ifc_work)->work) &idev->mc_lock rcu_read_lock rcu_read_lock &tbl->lock &pcp->lock &zone->lock irq_context: 0 (wq_completion)mld (work_completion)(&(&idev->mc_ifc_work)->work) &idev->mc_lock rcu_read_lock rcu_read_lock &tbl->lock &pcp->lock &zone->lock &____s->seqcount irq_context: 0 (wq_completion)bond158 (work_completion)(&(&bond->mcast_work)->work) irq_context: 0 (wq_completion)bond158 (work_completion)(&(&bond->mcast_work)->work) &obj_hash[i].lock irq_context: 0 (wq_completion)bond158 (work_completion)(&(&bond->mcast_work)->work) &base->lock irq_context: 0 (wq_completion)bond158 (work_completion)(&(&bond->mcast_work)->work) &base->lock fill_pool_map-wait-type-override pool_lock irq_context: 0 (wq_completion)bond158 (work_completion)(&(&bond->mcast_work)->work) &base->lock &obj_hash[i].lock irq_context: 0 (wq_completion)bond158 (work_completion)(&(&bond->mcast_work)->work) &rq->__lock irq_context: 0 (wq_completion)bond158 (work_completion)(&(&bond->mcast_work)->work) &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)mld (work_completion)(&(&idev->mc_ifc_work)->work) &idev->mc_lock rcu_read_lock rcu_read_lock &tbl->lock batched_entropy_u8.lock irq_context: 0 (wq_completion)mld (work_completion)(&(&idev->mc_ifc_work)->work) &idev->mc_lock rcu_read_lock rcu_read_lock &tbl->lock kfence_freelist_lock irq_context: 0 (wq_completion)mld (work_completion)(&(&idev->mc_ifc_work)->work) &idev->mc_lock rcu_read_lock rcu_read_lock &tbl->lock &meta->lock irq_context: 0 (wq_completion)bond128#2 (work_completion)(&(&slave->notify_work)->work) fill_pool_map-wait-type-override &c->lock irq_context: 0 (wq_completion)bond128#2 (work_completion)(&(&slave->notify_work)->work) fill_pool_map-wait-type-override &____s->seqcount#2 irq_context: 0 (wq_completion)bond128#2 (work_completion)(&(&slave->notify_work)->work) fill_pool_map-wait-type-override &____s->seqcount irq_context: 0 (wq_completion)bond128#2 (work_completion)(&(&slave->notify_work)->work) fill_pool_map-wait-type-override pool_lock irq_context: 0 (wq_completion)bond108#2 (work_completion)(&(&bond->mcast_work)->work) &rq->__lock irq_context: 0 (wq_completion)bond108#2 (work_completion)(&(&bond->mcast_work)->work) &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)bond150 (work_completion)(&(&bond->mcast_work)->work) &rq->__lock irq_context: 0 (wq_completion)bond150 (work_completion)(&(&bond->mcast_work)->work) &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)bond39 irq_context: 0 (wq_completion)bond39 (work_completion)(&(&slave->notify_work)->work) irq_context: 0 (wq_completion)bond39 (work_completion)(&(&slave->notify_work)->work) &obj_hash[i].lock irq_context: 0 (wq_completion)bond39 (work_completion)(&(&slave->notify_work)->work) &base->lock irq_context: 0 (wq_completion)bond39 (work_completion)(&(&slave->notify_work)->work) &base->lock &obj_hash[i].lock irq_context: 0 (wq_completion)bond39 (work_completion)(&(&slave->notify_work)->work) &rq->__lock irq_context: 0 (wq_completion)bond39 (work_completion)(&(&slave->notify_work)->work) &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)bond39 (work_completion)(&(&bond->mcast_work)->work) irq_context: 0 (wq_completion)bond39 (work_completion)(&(&bond->mcast_work)->work) &obj_hash[i].lock irq_context: 0 (wq_completion)bond39 (work_completion)(&(&bond->mcast_work)->work) &base->lock irq_context: 0 (wq_completion)bond39 (work_completion)(&(&bond->mcast_work)->work) &base->lock &obj_hash[i].lock irq_context: 0 (wq_completion)bond39 (work_completion)(&(&bond->mcast_work)->work) &rq->__lock irq_context: 0 (wq_completion)bond39 (work_completion)(&(&bond->mcast_work)->work) &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)bond145#2 &rq->__lock irq_context: 0 (wq_completion)bond130#3 irq_context: 0 (wq_completion)bond130#3 (work_completion)(&(&slave->notify_work)->work) irq_context: 0 (wq_completion)bond130#3 (work_completion)(&(&slave->notify_work)->work) &obj_hash[i].lock irq_context: 0 (wq_completion)bond130#3 (work_completion)(&(&slave->notify_work)->work) &base->lock irq_context: 0 (wq_completion)bond130#3 (work_completion)(&(&slave->notify_work)->work) &base->lock &obj_hash[i].lock irq_context: 0 (wq_completion)bond130#3 (work_completion)(&(&slave->notify_work)->work) &rq->__lock irq_context: 0 (wq_completion)bond130#3 (work_completion)(&(&slave->notify_work)->work) &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)bond130#3 (work_completion)(&(&bond->mcast_work)->work) irq_context: 0 (wq_completion)bond130#3 (work_completion)(&(&bond->mcast_work)->work) &obj_hash[i].lock irq_context: 0 (wq_completion)bond130#3 (work_completion)(&(&bond->mcast_work)->work) &base->lock irq_context: 0 (wq_completion)bond130#3 (work_completion)(&(&bond->mcast_work)->work) &base->lock &obj_hash[i].lock irq_context: 0 (wq_completion)bond157 &rq->__lock irq_context: 0 (wq_completion)bond157 &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)bond146#2 irq_context: 0 (wq_completion)bond146#2 &rq->__lock irq_context: 0 (wq_completion)bond146#2 (work_completion)(&(&slave->notify_work)->work) irq_context: 0 (wq_completion)bond146#2 (work_completion)(&(&slave->notify_work)->work) &obj_hash[i].lock irq_context: 0 (wq_completion)bond146#2 (work_completion)(&(&slave->notify_work)->work) &base->lock irq_context: 0 (wq_completion)bond146#2 (work_completion)(&(&slave->notify_work)->work) &base->lock &obj_hash[i].lock irq_context: 0 (wq_completion)bond146#2 (work_completion)(&(&slave->notify_work)->work) &rq->__lock irq_context: 0 (wq_completion)bond146#2 (work_completion)(&(&slave->notify_work)->work) &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)bond146#2 (work_completion)(&(&bond->mcast_work)->work) irq_context: 0 (wq_completion)bond146#2 (work_completion)(&(&bond->mcast_work)->work) &obj_hash[i].lock irq_context: 0 (wq_completion)bond146#2 (work_completion)(&(&bond->mcast_work)->work) &base->lock irq_context: 0 (wq_completion)bond146#2 (work_completion)(&(&bond->mcast_work)->work) &base->lock &obj_hash[i].lock irq_context: 0 (wq_completion)bond146#2 (work_completion)(&(&bond->mcast_work)->work) &rq->__lock irq_context: 0 (wq_completion)bond146#2 (work_completion)(&(&bond->mcast_work)->work) &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)bond33#2 (work_completion)(&(&bond->mcast_work)->work) &rq->__lock irq_context: 0 (wq_completion)bond33#2 (work_completion)(&(&bond->mcast_work)->work) &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)bond159 irq_context: 0 (wq_completion)bond159 &rq->__lock irq_context: 0 (wq_completion)bond159 &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)bond159 (work_completion)(&(&slave->notify_work)->work) irq_context: 0 (wq_completion)bond159 (work_completion)(&(&slave->notify_work)->work) &obj_hash[i].lock irq_context: 0 (wq_completion)bond159 (work_completion)(&(&slave->notify_work)->work) &base->lock irq_context: 0 (wq_completion)bond159 (work_completion)(&(&slave->notify_work)->work) &base->lock &obj_hash[i].lock irq_context: 0 (wq_completion)bond159 (work_completion)(&(&bond->mcast_work)->work) irq_context: 0 (wq_completion)bond159 (work_completion)(&(&bond->mcast_work)->work) &obj_hash[i].lock irq_context: 0 (wq_completion)bond159 (work_completion)(&(&bond->mcast_work)->work) &base->lock irq_context: 0 (wq_completion)bond159 (work_completion)(&(&bond->mcast_work)->work) &base->lock &obj_hash[i].lock irq_context: 0 (wq_completion)bond39 (work_completion)(&(&slave->notify_work)->work) rtnl_mutex net_rwsem irq_context: 0 (wq_completion)bond39 (work_completion)(&(&slave->notify_work)->work) rtnl_mutex net_rwsem &list->lock#2 irq_context: 0 (wq_completion)bond39 (work_completion)(&(&slave->notify_work)->work) rtnl_mutex &tn->lock irq_context: 0 (wq_completion)bond39 (work_completion)(&(&slave->notify_work)->work) rtnl_mutex.wait_lock irq_context: 0 (wq_completion)bond39 (work_completion)(&(&slave->notify_work)->work) &p->pi_lock irq_context: 0 (wq_completion)bond126#3 (work_completion)(&(&slave->notify_work)->work) rtnl_mutex net_rwsem irq_context: 0 (wq_completion)bond126#3 (work_completion)(&(&slave->notify_work)->work) rtnl_mutex net_rwsem &list->lock#2 irq_context: 0 (wq_completion)bond126#3 (work_completion)(&(&slave->notify_work)->work) rtnl_mutex &tn->lock irq_context: 0 (wq_completion)bond126#3 (work_completion)(&(&slave->notify_work)->work) rtnl_mutex.wait_lock irq_context: 0 (wq_completion)bond126#3 (work_completion)(&(&slave->notify_work)->work) &p->pi_lock irq_context: 0 (wq_completion)bond145#2 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex fs_reclaim irq_context: 0 (wq_completion)bond145#2 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex fs_reclaim mmu_notifier_invalidate_range_start irq_context: 0 (wq_completion)bond145#2 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &c->lock irq_context: 0 (wq_completion)bond145#2 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex rcu_read_lock &bond->stats_lock/1 irq_context: 0 (wq_completion)bond145#2 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex rcu_read_lock &ndev->lock irq_context: 0 (wq_completion)bond145#2 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex nl_table_lock irq_context: 0 (wq_completion)bond145#2 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &obj_hash[i].lock irq_context: 0 (wq_completion)bond145#2 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex nl_table_wait.lock irq_context: 0 (wq_completion)bond145#2 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex net_rwsem irq_context: 0 (wq_completion)bond145#2 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex net_rwsem &list->lock#2 irq_context: 0 (wq_completion)bond145#2 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &tn->lock irq_context: 0 (wq_completion)bond145#2 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &idev->mc_lock irq_context: 0 (wq_completion)bond145#2 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &idev->mc_lock fs_reclaim irq_context: 0 (wq_completion)bond145#2 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &idev->mc_lock fs_reclaim mmu_notifier_invalidate_range_start irq_context: 0 (wq_completion)bond145#2 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &idev->mc_lock &c->lock irq_context: 0 (wq_completion)bond145#2 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &idev->mc_lock rcu_read_lock &ndev->lock irq_context: 0 (wq_completion)bond145#2 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &idev->mc_lock rcu_read_lock &dir->lock#2 irq_context: 0 (wq_completion)bond145#2 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &idev->mc_lock rcu_read_lock &ul->lock#2 irq_context: 0 (wq_completion)bond145#2 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &idev->mc_lock rcu_read_lock rcu_read_lock &____s->seqcount#10 irq_context: 0 (wq_completion)bond145#2 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &idev->mc_lock rcu_read_lock rcu_read_lock rcu_read_lock_bh rcu_read_lock rcu_read_lock_bh rcu_read_lock &br->multicast_lock irq_context: 0 (wq_completion)bond145#2 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &idev->mc_lock rcu_read_lock rcu_read_lock rcu_read_lock_bh rcu_read_lock rcu_read_lock_bh rcu_read_lock &br->multicast_lock &obj_hash[i].lock irq_context: 0 (wq_completion)bond145#2 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &idev->mc_lock rcu_read_lock rcu_read_lock rcu_read_lock_bh rcu_read_lock rcu_read_lock_bh rcu_read_lock &br->multicast_lock &base->lock irq_context: 0 (wq_completion)bond145#2 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &idev->mc_lock rcu_read_lock rcu_read_lock rcu_read_lock_bh rcu_read_lock rcu_read_lock_bh rcu_read_lock &data->lock irq_context: 0 (wq_completion)bond145#2 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &idev->mc_lock rcu_read_lock rcu_read_lock rcu_read_lock_bh rcu_read_lock rcu_read_lock_bh rcu_read_lock &obj_hash[i].lock irq_context: 0 (wq_completion)bond145#2 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex.wait_lock irq_context: 0 (wq_completion)bond145#2 (work_completion)(&(&bond->mcast_work)->work) &p->pi_lock irq_context: 0 (wq_completion)bond116#3 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex fs_reclaim irq_context: 0 (wq_completion)bond116#3 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex fs_reclaim mmu_notifier_invalidate_range_start irq_context: 0 (wq_completion)bond116#3 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &c->lock irq_context: 0 (wq_completion)bond116#3 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &n->list_lock irq_context: 0 (wq_completion)bond116#3 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &n->list_lock &c->lock irq_context: 0 (wq_completion)bond116#3 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex rcu_read_lock &bond->stats_lock/1 irq_context: 0 (wq_completion)bond116#3 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex rcu_read_lock &ndev->lock irq_context: 0 (wq_completion)bond116#3 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex nl_table_lock irq_context: 0 (wq_completion)bond116#3 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &obj_hash[i].lock irq_context: 0 (wq_completion)bond116#3 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex nl_table_wait.lock irq_context: 0 (wq_completion)bond116#3 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex net_rwsem irq_context: 0 (wq_completion)bond116#3 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex net_rwsem &rq->__lock irq_context: 0 (wq_completion)bond116#3 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex net_rwsem &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)bond116#3 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex net_rwsem &list->lock#2 irq_context: 0 (wq_completion)bond116#3 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &tn->lock irq_context: 0 (wq_completion)bond116#3 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &idev->mc_lock irq_context: 0 (wq_completion)bond116#3 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &idev->mc_lock fs_reclaim irq_context: 0 (wq_completion)bond116#3 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &idev->mc_lock fs_reclaim mmu_notifier_invalidate_range_start irq_context: 0 (wq_completion)bond116#3 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &idev->mc_lock rcu_read_lock &ndev->lock irq_context: 0 (wq_completion)bond116#3 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &idev->mc_lock rcu_read_lock &c->lock irq_context: 0 (wq_completion)bond116#3 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &idev->mc_lock rcu_read_lock &____s->seqcount#2 irq_context: 0 (wq_completion)bond116#3 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &idev->mc_lock rcu_read_lock rcu_node_0 irq_context: 0 (wq_completion)bond116#3 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &idev->mc_lock rcu_read_lock &rq->__lock irq_context: 0 (wq_completion)bond116#3 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &idev->mc_lock rcu_read_lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)bond116#3 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &idev->mc_lock rcu_read_lock &____s->seqcount irq_context: 0 (wq_completion)bond116#3 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &idev->mc_lock rcu_read_lock &dir->lock#2 irq_context: 0 (wq_completion)bond116#3 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &idev->mc_lock rcu_read_lock &ul->lock#2 irq_context: 0 (wq_completion)bond116#3 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &idev->mc_lock rcu_read_lock rcu_read_lock &____s->seqcount#10 irq_context: 0 (wq_completion)bond116#3 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &idev->mc_lock rcu_read_lock rcu_read_lock rcu_read_lock_bh rcu_read_lock rcu_read_lock_bh rcu_read_lock &br->multicast_lock irq_context: 0 (wq_completion)bond116#3 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &idev->mc_lock rcu_read_lock rcu_read_lock rcu_read_lock_bh rcu_read_lock rcu_read_lock_bh rcu_read_lock &br->multicast_lock &obj_hash[i].lock irq_context: 0 (wq_completion)bond116#3 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &idev->mc_lock rcu_read_lock rcu_read_lock rcu_read_lock_bh rcu_read_lock rcu_read_lock_bh rcu_read_lock &br->multicast_lock &base->lock irq_context: 0 (wq_completion)bond116#3 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &idev->mc_lock rcu_read_lock rcu_read_lock rcu_read_lock_bh rcu_read_lock rcu_read_lock_bh rcu_read_lock &br->multicast_lock &base->lock &obj_hash[i].lock irq_context: 0 (wq_completion)bond116#3 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &idev->mc_lock rcu_read_lock rcu_read_lock rcu_read_lock_bh rcu_read_lock rcu_read_lock_bh rcu_read_lock &data->lock irq_context: 0 (wq_completion)bond116#3 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &idev->mc_lock rcu_read_lock rcu_read_lock rcu_read_lock_bh rcu_read_lock rcu_read_lock_bh rcu_read_lock &obj_hash[i].lock irq_context: 0 (wq_completion)bond116#3 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &idev->mc_lock rcu_read_lock rcu_read_lock &rq->__lock irq_context: 0 (wq_completion)bond116#3 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &idev->mc_lock rcu_read_lock rcu_read_lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)bond116#3 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex.wait_lock irq_context: 0 (wq_completion)bond116#3 (work_completion)(&(&bond->mcast_work)->work) &p->pi_lock irq_context: 0 (wq_completion)bond116#3 (work_completion)(&(&bond->mcast_work)->work) &p->pi_lock &rq->__lock irq_context: 0 (wq_completion)bond116#3 (work_completion)(&(&bond->mcast_work)->work) &p->pi_lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)bond116#3 (work_completion)(&(&bond->mcast_work)->work) &rq->__lock irq_context: 0 (wq_completion)bond116#3 (work_completion)(&(&bond->mcast_work)->work) &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)bond106#3 (work_completion)(&(&slave->notify_work)->work) &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)bond131#2 irq_context: 0 (wq_completion)bond131#2 (work_completion)(&(&slave->notify_work)->work) irq_context: 0 (wq_completion)bond131#2 (work_completion)(&(&slave->notify_work)->work) &obj_hash[i].lock irq_context: 0 (wq_completion)bond131#2 (work_completion)(&(&slave->notify_work)->work) &base->lock irq_context: 0 (wq_completion)bond131#2 (work_completion)(&(&slave->notify_work)->work) &base->lock &obj_hash[i].lock irq_context: 0 (wq_completion)bond131#2 (work_completion)(&(&slave->notify_work)->work) &rq->__lock irq_context: 0 (wq_completion)bond131#2 (work_completion)(&(&slave->notify_work)->work) &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)bond131#2 (work_completion)(&(&bond->mcast_work)->work) irq_context: 0 (wq_completion)bond131#2 (work_completion)(&(&bond->mcast_work)->work) &obj_hash[i].lock irq_context: 0 (wq_completion)bond131#2 (work_completion)(&(&bond->mcast_work)->work) &base->lock irq_context: 0 (wq_completion)bond131#2 (work_completion)(&(&bond->mcast_work)->work) &base->lock &obj_hash[i].lock irq_context: 0 (wq_completion)bond143#2 (work_completion)(&(&slave->notify_work)->work) rtnl_mutex net_rwsem irq_context: 0 (wq_completion)bond143#2 (work_completion)(&(&slave->notify_work)->work) rtnl_mutex net_rwsem &list->lock#2 irq_context: 0 (wq_completion)bond143#2 (work_completion)(&(&slave->notify_work)->work) rtnl_mutex &tn->lock irq_context: 0 (wq_completion)bond143#2 (work_completion)(&(&slave->notify_work)->work) rtnl_mutex rcu_read_lock rcu_node_0 irq_context: 0 (wq_completion)bond143#2 (work_completion)(&(&slave->notify_work)->work) rtnl_mutex rcu_read_lock &rq->__lock irq_context: 0 (wq_completion)bond143#2 (work_completion)(&(&slave->notify_work)->work) rtnl_mutex rcu_read_lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)bond143#2 (work_completion)(&(&slave->notify_work)->work) rtnl_mutex.wait_lock irq_context: 0 (wq_completion)bond143#2 (work_completion)(&(&slave->notify_work)->work) &p->pi_lock irq_context: 0 (wq_completion)bond143#2 (work_completion)(&(&slave->notify_work)->work) &p->pi_lock &rq->__lock irq_context: 0 (wq_completion)bond143#2 (work_completion)(&(&slave->notify_work)->work) &p->pi_lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)bond13 &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)bond143#2 (work_completion)(&(&slave->notify_work)->work) &rq->__lock irq_context: 0 (wq_completion)bond36#2 &rq->__lock irq_context: 0 (wq_completion)bond36#2 &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)bond107#2 &rq->__lock irq_context: 0 (wq_completion)bond107#2 &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)bond147#2 irq_context: 0 (wq_completion)bond147#2 (work_completion)(&(&slave->notify_work)->work) irq_context: 0 (wq_completion)bond147#2 (work_completion)(&(&slave->notify_work)->work) &obj_hash[i].lock irq_context: 0 (wq_completion)bond147#2 (work_completion)(&(&slave->notify_work)->work) &base->lock irq_context: 0 (wq_completion)bond147#2 (work_completion)(&(&slave->notify_work)->work) &base->lock &obj_hash[i].lock irq_context: 0 (wq_completion)bond147#2 (work_completion)(&(&slave->notify_work)->work) &rq->__lock irq_context: 0 (wq_completion)bond147#2 (work_completion)(&(&slave->notify_work)->work) &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)bond147#2 (work_completion)(&(&bond->mcast_work)->work) irq_context: 0 (wq_completion)bond147#2 (work_completion)(&(&bond->mcast_work)->work) &obj_hash[i].lock irq_context: 0 (wq_completion)bond147#2 (work_completion)(&(&bond->mcast_work)->work) &base->lock irq_context: 0 (wq_completion)bond147#2 (work_completion)(&(&bond->mcast_work)->work) &base->lock &obj_hash[i].lock irq_context: 0 (wq_completion)bond147#2 (work_completion)(&(&bond->mcast_work)->work) &rq->__lock irq_context: 0 (wq_completion)bond147#2 (work_completion)(&(&bond->mcast_work)->work) &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)bond147#2 (work_completion)(&(&bond->mcast_work)->work) &cfs_rq->removed.lock irq_context: 0 (wq_completion)bond122#3 (work_completion)(&(&bond->mcast_work)->work) &rq->__lock irq_context: 0 (wq_completion)bond122#3 (work_completion)(&(&bond->mcast_work)->work) &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)bond145 (work_completion)(&(&bond->mcast_work)->work) &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)bond160 irq_context: 0 (wq_completion)bond160 (work_completion)(&(&slave->notify_work)->work) irq_context: 0 (wq_completion)bond160 (work_completion)(&(&slave->notify_work)->work) &obj_hash[i].lock irq_context: 0 (wq_completion)bond160 (work_completion)(&(&slave->notify_work)->work) &base->lock irq_context: 0 (wq_completion)bond160 (work_completion)(&(&slave->notify_work)->work) &base->lock &obj_hash[i].lock irq_context: 0 (wq_completion)bond160 (work_completion)(&(&slave->notify_work)->work) &rq->__lock irq_context: 0 (wq_completion)bond160 (work_completion)(&(&slave->notify_work)->work) &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)bond125#3 (work_completion)(&(&bond->mcast_work)->work) &rq->__lock irq_context: 0 (wq_completion)bond160 (work_completion)(&(&bond->mcast_work)->work) irq_context: 0 (wq_completion)bond160 (work_completion)(&(&bond->mcast_work)->work) &obj_hash[i].lock irq_context: 0 (wq_completion)bond160 (work_completion)(&(&bond->mcast_work)->work) &base->lock irq_context: 0 (wq_completion)bond160 (work_completion)(&(&bond->mcast_work)->work) &base->lock &obj_hash[i].lock irq_context: 0 (wq_completion)bond136#2 (work_completion)(&(&bond->mcast_work)->work) &rq->__lock irq_context: 0 (wq_completion)bond136#2 (work_completion)(&(&bond->mcast_work)->work) &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)bond132 (work_completion)(&(&bond->mcast_work)->work) &base->lock fill_pool_map-wait-type-override pool_lock irq_context: 0 (wq_completion)bond122#3 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex fs_reclaim irq_context: 0 (wq_completion)bond122#3 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex fs_reclaim mmu_notifier_invalidate_range_start irq_context: 0 (wq_completion)bond122#3 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &c->lock irq_context: 0 (wq_completion)bond122#3 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex rcu_read_lock &bond->stats_lock/1 irq_context: 0 (wq_completion)bond122#3 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex rcu_read_lock &ndev->lock irq_context: 0 (wq_completion)bond122#3 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex nl_table_lock irq_context: 0 (wq_completion)bond122#3 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &obj_hash[i].lock irq_context: 0 (wq_completion)bond122#3 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex nl_table_wait.lock irq_context: 0 (wq_completion)bond122#3 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex net_rwsem irq_context: 0 (wq_completion)bond122#3 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex net_rwsem &list->lock#2 irq_context: 0 (wq_completion)bond122#3 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &tn->lock irq_context: 0 (wq_completion)bond122#3 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &idev->mc_lock irq_context: 0 (wq_completion)bond122#3 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &idev->mc_lock fs_reclaim irq_context: 0 (wq_completion)bond122#3 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &idev->mc_lock fs_reclaim mmu_notifier_invalidate_range_start irq_context: 0 (wq_completion)bond122#3 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &idev->mc_lock &c->lock irq_context: 0 (wq_completion)bond122#3 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &idev->mc_lock rcu_read_lock &ndev->lock irq_context: 0 (wq_completion)bond122#3 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &idev->mc_lock rcu_read_lock &dir->lock#2 irq_context: 0 (wq_completion)bond122#3 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &idev->mc_lock rcu_read_lock &ul->lock#2 irq_context: 0 (wq_completion)bond122#3 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &idev->mc_lock rcu_read_lock rcu_read_lock &____s->seqcount#10 irq_context: 0 (wq_completion)bond122#3 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &idev->mc_lock rcu_read_lock rcu_read_lock rcu_read_lock_bh rcu_read_lock rcu_read_lock_bh rcu_read_lock &br->multicast_lock irq_context: 0 (wq_completion)bond122#3 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &idev->mc_lock rcu_read_lock rcu_read_lock rcu_read_lock_bh rcu_read_lock rcu_read_lock_bh rcu_read_lock &br->multicast_lock &obj_hash[i].lock irq_context: 0 (wq_completion)bond122#3 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &idev->mc_lock rcu_read_lock rcu_read_lock rcu_read_lock_bh rcu_read_lock rcu_read_lock_bh rcu_read_lock &br->multicast_lock &base->lock irq_context: 0 (wq_completion)bond122#3 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &idev->mc_lock rcu_read_lock rcu_read_lock rcu_read_lock_bh rcu_read_lock rcu_read_lock_bh rcu_read_lock &br->multicast_lock &base->lock &obj_hash[i].lock irq_context: 0 (wq_completion)bond122#3 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &idev->mc_lock rcu_read_lock rcu_read_lock rcu_read_lock_bh rcu_read_lock rcu_read_lock_bh rcu_read_lock &data->lock irq_context: 0 (wq_completion)bond122#3 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &idev->mc_lock rcu_read_lock rcu_read_lock rcu_read_lock_bh rcu_read_lock rcu_read_lock_bh rcu_read_lock &obj_hash[i].lock irq_context: 0 (wq_completion)bond122#3 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex.wait_lock irq_context: 0 (wq_completion)bond122#3 (work_completion)(&(&bond->mcast_work)->work) &p->pi_lock irq_context: 0 (wq_completion)bond121#3 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex fs_reclaim irq_context: 0 (wq_completion)bond121#3 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex fs_reclaim mmu_notifier_invalidate_range_start irq_context: 0 (wq_completion)bond121#3 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &c->lock irq_context: 0 (wq_completion)bond121#3 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex rcu_read_lock &bond->stats_lock/1 irq_context: 0 (wq_completion)bond121#3 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex rcu_read_lock &ndev->lock irq_context: 0 (wq_completion)bond121#3 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex nl_table_lock irq_context: 0 (wq_completion)bond121#3 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &obj_hash[i].lock irq_context: 0 (wq_completion)bond121#3 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex nl_table_wait.lock irq_context: 0 (wq_completion)bond121#3 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex net_rwsem irq_context: 0 (wq_completion)bond121#3 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex net_rwsem &list->lock#2 irq_context: 0 (wq_completion)bond121#3 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &tn->lock irq_context: 0 (wq_completion)bond121#3 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &idev->mc_lock irq_context: 0 (wq_completion)bond121#3 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &idev->mc_lock fs_reclaim irq_context: 0 (wq_completion)bond121#3 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &idev->mc_lock fs_reclaim mmu_notifier_invalidate_range_start irq_context: 0 (wq_completion)bond121#3 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &idev->mc_lock &c->lock irq_context: 0 (wq_completion)bond121#3 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &idev->mc_lock rcu_read_lock &ndev->lock irq_context: 0 (wq_completion)bond121#3 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &idev->mc_lock rcu_read_lock &dir->lock#2 irq_context: 0 (wq_completion)bond121#3 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &idev->mc_lock rcu_read_lock &ul->lock#2 irq_context: 0 (wq_completion)bond121#3 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &idev->mc_lock rcu_read_lock rcu_read_lock &____s->seqcount#10 irq_context: 0 (wq_completion)bond121#3 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &idev->mc_lock rcu_read_lock rcu_read_lock rcu_read_lock_bh rcu_read_lock rcu_read_lock_bh rcu_read_lock &br->multicast_lock irq_context: 0 (wq_completion)bond121#3 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &idev->mc_lock rcu_read_lock rcu_read_lock rcu_read_lock_bh rcu_read_lock rcu_read_lock_bh rcu_read_lock &br->multicast_lock &obj_hash[i].lock irq_context: 0 (wq_completion)bond121#3 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &idev->mc_lock rcu_read_lock rcu_read_lock rcu_read_lock_bh rcu_read_lock rcu_read_lock_bh rcu_read_lock &br->multicast_lock &base->lock irq_context: 0 (wq_completion)bond121#3 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &idev->mc_lock rcu_read_lock rcu_read_lock rcu_read_lock_bh rcu_read_lock rcu_read_lock_bh rcu_read_lock &br->multicast_lock &base->lock &obj_hash[i].lock irq_context: 0 (wq_completion)bond121#3 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &idev->mc_lock rcu_read_lock rcu_read_lock rcu_read_lock_bh rcu_read_lock rcu_read_lock_bh rcu_read_lock &data->lock irq_context: 0 (wq_completion)bond121#3 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &idev->mc_lock rcu_read_lock rcu_read_lock rcu_read_lock_bh rcu_read_lock rcu_read_lock_bh rcu_read_lock &obj_hash[i].lock irq_context: 0 (wq_completion)bond121#3 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex.wait_lock irq_context: 0 (wq_completion)bond121#3 (work_completion)(&(&bond->mcast_work)->work) &p->pi_lock irq_context: 0 (wq_completion)bond146#2 (work_completion)(&(&slave->notify_work)->work) rtnl_mutex net_rwsem irq_context: 0 (wq_completion)bond146#2 (work_completion)(&(&slave->notify_work)->work) rtnl_mutex net_rwsem &list->lock#2 irq_context: 0 (wq_completion)bond146#2 (work_completion)(&(&slave->notify_work)->work) rtnl_mutex &tn->lock irq_context: 0 (wq_completion)bond146#2 (work_completion)(&(&slave->notify_work)->work) rtnl_mutex rcu_read_lock rcu_node_0 irq_context: 0 (wq_completion)bond146#2 (work_completion)(&(&slave->notify_work)->work) rtnl_mutex rcu_read_lock &rq->__lock irq_context: 0 (wq_completion)bond146#2 (work_completion)(&(&slave->notify_work)->work) rtnl_mutex rcu_read_lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)bond116#3 (work_completion)(&(&slave->notify_work)->work) &rq->__lock irq_context: 0 (wq_completion)bond116#3 (work_completion)(&(&slave->notify_work)->work) &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)bond146#2 (work_completion)(&(&slave->notify_work)->work) rtnl_mutex.wait_lock irq_context: 0 (wq_completion)bond146#2 (work_completion)(&(&slave->notify_work)->work) &p->pi_lock irq_context: 0 (wq_completion)bond146#2 (work_completion)(&(&slave->notify_work)->work) &p->pi_lock &rq->__lock irq_context: 0 (wq_completion)bond146#2 (work_completion)(&(&slave->notify_work)->work) &p->pi_lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)bond115 &rq->__lock irq_context: 0 (wq_completion)bond115 &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)bond154 &rq->__lock irq_context: 0 (wq_completion)bond154 &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)bond132#3 irq_context: 0 (wq_completion)bond132#3 &rq->__lock irq_context: 0 (wq_completion)bond132#3 &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)bond132#3 (work_completion)(&(&slave->notify_work)->work) irq_context: 0 (wq_completion)bond132#3 (work_completion)(&(&slave->notify_work)->work) &obj_hash[i].lock irq_context: 0 (wq_completion)bond132#3 (work_completion)(&(&slave->notify_work)->work) &base->lock irq_context: 0 (wq_completion)bond132#3 (work_completion)(&(&slave->notify_work)->work) &base->lock &obj_hash[i].lock irq_context: 0 (wq_completion)bond132#3 (work_completion)(&(&slave->notify_work)->work) &rq->__lock irq_context: 0 (wq_completion)bond132#3 (work_completion)(&(&slave->notify_work)->work) &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)bond132#3 (work_completion)(&(&bond->mcast_work)->work) irq_context: 0 (wq_completion)bond132#3 (work_completion)(&(&bond->mcast_work)->work) &obj_hash[i].lock irq_context: 0 (wq_completion)bond132#3 (work_completion)(&(&bond->mcast_work)->work) &base->lock irq_context: 0 (wq_completion)bond132#3 (work_completion)(&(&bond->mcast_work)->work) &base->lock &obj_hash[i].lock irq_context: 0 (wq_completion)bond148#2 irq_context: 0 (wq_completion)bond148#2 (work_completion)(&(&slave->notify_work)->work) irq_context: 0 (wq_completion)bond148#2 (work_completion)(&(&slave->notify_work)->work) &obj_hash[i].lock irq_context: 0 (wq_completion)bond148#2 (work_completion)(&(&slave->notify_work)->work) &base->lock irq_context: 0 (wq_completion)bond148#2 (work_completion)(&(&slave->notify_work)->work) &base->lock &obj_hash[i].lock irq_context: 0 (wq_completion)bond148#2 (work_completion)(&(&bond->mcast_work)->work) irq_context: 0 (wq_completion)bond148#2 (work_completion)(&(&bond->mcast_work)->work) &obj_hash[i].lock irq_context: 0 (wq_completion)bond148#2 (work_completion)(&(&bond->mcast_work)->work) &base->lock irq_context: 0 (wq_completion)bond148#2 (work_completion)(&(&bond->mcast_work)->work) &base->lock &obj_hash[i].lock irq_context: 0 (wq_completion)bond31#3 (work_completion)(&(&bond->mcast_work)->work) fill_pool_map-wait-type-override &c->lock irq_context: 0 (wq_completion)bond31#3 (work_completion)(&(&bond->mcast_work)->work) fill_pool_map-wait-type-override &____s->seqcount#2 irq_context: 0 (wq_completion)bond31#3 (work_completion)(&(&bond->mcast_work)->work) fill_pool_map-wait-type-override &____s->seqcount irq_context: 0 (wq_completion)bond31#3 (work_completion)(&(&bond->mcast_work)->work) fill_pool_map-wait-type-override pool_lock irq_context: 0 (wq_completion)bond133#2 (work_completion)(&(&slave->notify_work)->work) &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)bond143#2 (work_completion)(&(&bond->mcast_work)->work) &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)bond143 (work_completion)(&(&slave->notify_work)->work) rtnl_mutex net_rwsem irq_context: 0 (wq_completion)bond143 (work_completion)(&(&slave->notify_work)->work) rtnl_mutex net_rwsem &list->lock#2 irq_context: 0 (wq_completion)bond143 (work_completion)(&(&slave->notify_work)->work) rtnl_mutex &tn->lock irq_context: 0 (wq_completion)bond143 (work_completion)(&(&slave->notify_work)->work) rtnl_mutex.wait_lock irq_context: 0 (wq_completion)bond143 (work_completion)(&(&slave->notify_work)->work) &p->pi_lock irq_context: 0 (wq_completion)bond159 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex fs_reclaim irq_context: 0 (wq_completion)bond159 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex fs_reclaim &rq->__lock irq_context: 0 (wq_completion)bond159 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex fs_reclaim &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)bond159 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex fs_reclaim mmu_notifier_invalidate_range_start irq_context: 0 (wq_completion)bond159 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex rcu_read_lock &bond->stats_lock/1 irq_context: 0 (wq_completion)bond159 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex rcu_read_lock &ndev->lock irq_context: 0 (wq_completion)bond159 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex nl_table_lock irq_context: 0 (wq_completion)bond159 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &obj_hash[i].lock irq_context: 0 (wq_completion)bond159 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex nl_table_wait.lock irq_context: 0 (wq_completion)bond159 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex net_rwsem irq_context: 0 (wq_completion)bond159 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex net_rwsem &list->lock#2 irq_context: 0 (wq_completion)bond159 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &rq->__lock irq_context: 0 (wq_completion)bond159 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)bond159 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &tn->lock irq_context: 0 (wq_completion)bond159 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &idev->mc_lock irq_context: 0 (wq_completion)bond159 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &idev->mc_lock fs_reclaim irq_context: 0 (wq_completion)bond159 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &idev->mc_lock fs_reclaim mmu_notifier_invalidate_range_start irq_context: 0 (wq_completion)bond159 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &idev->mc_lock rcu_read_lock &ndev->lock irq_context: 0 (wq_completion)bond159 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &idev->mc_lock rcu_read_lock &dir->lock#2 irq_context: 0 (wq_completion)bond159 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &idev->mc_lock rcu_read_lock &ul->lock#2 irq_context: 0 (wq_completion)bond159 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &idev->mc_lock rcu_read_lock rcu_read_lock &____s->seqcount#10 irq_context: 0 (wq_completion)bond159 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &idev->mc_lock rcu_read_lock rcu_read_lock rcu_read_lock_bh rcu_read_lock rcu_read_lock_bh rcu_read_lock &br->multicast_lock irq_context: 0 (wq_completion)bond159 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &idev->mc_lock rcu_read_lock rcu_read_lock rcu_read_lock_bh rcu_read_lock rcu_read_lock_bh rcu_read_lock &br->multicast_lock &obj_hash[i].lock irq_context: 0 (wq_completion)bond159 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &idev->mc_lock rcu_read_lock rcu_read_lock rcu_read_lock_bh rcu_read_lock rcu_read_lock_bh rcu_read_lock &br->multicast_lock &base->lock irq_context: 0 (wq_completion)bond159 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &idev->mc_lock rcu_read_lock rcu_read_lock rcu_read_lock_bh rcu_read_lock rcu_read_lock_bh rcu_read_lock &data->lock irq_context: 0 (wq_completion)bond159 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &idev->mc_lock rcu_read_lock rcu_read_lock rcu_read_lock_bh rcu_read_lock rcu_read_lock_bh rcu_read_lock &obj_hash[i].lock irq_context: 0 (wq_completion)bond159 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &idev->mc_lock rcu_read_lock rcu_read_lock rcu_node_0 irq_context: 0 (wq_completion)bond159 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &idev->mc_lock rcu_read_lock rcu_read_lock &rq->__lock irq_context: 0 (wq_completion)bond159 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &idev->mc_lock rcu_read_lock rcu_read_lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)bond159 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &idev->mc_lock rcu_read_lock rcu_node_0 irq_context: 0 (wq_completion)bond159 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex.wait_lock irq_context: 0 (wq_completion)bond159 (work_completion)(&(&bond->mcast_work)->work) &p->pi_lock irq_context: 0 (wq_completion)bond150 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex fs_reclaim irq_context: 0 (wq_completion)bond150 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex fs_reclaim &rq->__lock irq_context: 0 (wq_completion)bond150 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex fs_reclaim mmu_notifier_invalidate_range_start irq_context: 0 (wq_completion)bond150 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &c->lock irq_context: 0 (wq_completion)bond150 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &n->list_lock irq_context: 0 (wq_completion)bond150 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &n->list_lock &c->lock irq_context: 0 (wq_completion)bond150 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex remove_cache_srcu irq_context: 0 (wq_completion)bond150 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex remove_cache_srcu &rq->__lock irq_context: 0 (wq_completion)bond150 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex remove_cache_srcu &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)bond150 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex remove_cache_srcu quarantine_lock irq_context: 0 (wq_completion)bond150 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex rcu_read_lock &bond->stats_lock/1 irq_context: 0 (wq_completion)bond150 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex rcu_read_lock &ndev->lock irq_context: 0 (wq_completion)bond150 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex nl_table_lock irq_context: 0 (wq_completion)bond150 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &obj_hash[i].lock irq_context: 0 (wq_completion)bond150 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex nl_table_wait.lock irq_context: 0 (wq_completion)bond150 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex net_rwsem irq_context: 0 (wq_completion)bond150 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex net_rwsem &rq->__lock irq_context: 0 (wq_completion)bond150 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex net_rwsem &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)bond150 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex net_rwsem &list->lock#2 irq_context: 0 (wq_completion)bond150 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &tn->lock irq_context: 0 (wq_completion)bond150 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &idev->mc_lock irq_context: 0 (wq_completion)bond150 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &idev->mc_lock fs_reclaim irq_context: 0 (wq_completion)bond150 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &idev->mc_lock fs_reclaim mmu_notifier_invalidate_range_start irq_context: 0 (wq_completion)bond150 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &idev->mc_lock &c->lock irq_context: 0 (wq_completion)bond150 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &idev->mc_lock rcu_read_lock &ndev->lock irq_context: 0 (wq_completion)bond150 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &idev->mc_lock rcu_read_lock &dir->lock#2 irq_context: 0 (wq_completion)bond150 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &idev->mc_lock rcu_read_lock rcu_node_0 irq_context: 0 (wq_completion)bond150 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &idev->mc_lock rcu_read_lock &rq->__lock irq_context: 0 (wq_completion)bond150 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &idev->mc_lock rcu_read_lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)bond150 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &idev->mc_lock rcu_read_lock &ul->lock#2 irq_context: 0 (wq_completion)bond150 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &idev->mc_lock rcu_read_lock rcu_read_lock &____s->seqcount#10 irq_context: 0 (wq_completion)bond150 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &idev->mc_lock rcu_read_lock rcu_read_lock rcu_read_lock_bh rcu_read_lock rcu_read_lock_bh rcu_read_lock &br->multicast_lock irq_context: 0 (wq_completion)bond150 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &idev->mc_lock rcu_read_lock rcu_read_lock rcu_read_lock_bh rcu_read_lock rcu_read_lock_bh rcu_read_lock &br->multicast_lock &obj_hash[i].lock irq_context: 0 (wq_completion)bond150 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &idev->mc_lock rcu_read_lock rcu_read_lock rcu_read_lock_bh rcu_read_lock rcu_read_lock_bh rcu_read_lock &br->multicast_lock &base->lock irq_context: 0 (wq_completion)bond150 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &idev->mc_lock rcu_read_lock rcu_read_lock rcu_read_lock_bh rcu_read_lock rcu_read_lock_bh rcu_read_lock &br->multicast_lock &base->lock &obj_hash[i].lock irq_context: 0 (wq_completion)bond150 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &idev->mc_lock rcu_read_lock rcu_read_lock rcu_read_lock_bh rcu_read_lock rcu_read_lock_bh rcu_read_lock &data->lock irq_context: 0 (wq_completion)bond150 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &idev->mc_lock rcu_read_lock rcu_read_lock rcu_read_lock_bh rcu_read_lock rcu_read_lock_bh rcu_read_lock &obj_hash[i].lock irq_context: 0 (wq_completion)bond150 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &idev->mc_lock rcu_read_lock rcu_read_lock &rq->__lock irq_context: 0 (wq_completion)bond150 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &idev->mc_lock rcu_read_lock rcu_read_lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)bond150 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex.wait_lock irq_context: 0 (wq_completion)bond150 (work_completion)(&(&bond->mcast_work)->work) &p->pi_lock irq_context: 0 (wq_completion)bond150 (work_completion)(&(&bond->mcast_work)->work) &p->pi_lock &cfs_rq->removed.lock irq_context: 0 (wq_completion)bond150 (work_completion)(&(&bond->mcast_work)->work) &p->pi_lock &rq->__lock irq_context: 0 (wq_completion)bond150 (work_completion)(&(&bond->mcast_work)->work) &p->pi_lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)bond38#2 (work_completion)(&(&slave->notify_work)->work) &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)bond40 irq_context: 0 (wq_completion)bond40 (work_completion)(&(&slave->notify_work)->work) irq_context: 0 (wq_completion)bond40 (work_completion)(&(&slave->notify_work)->work) &rq->__lock irq_context: 0 (wq_completion)bond40 (work_completion)(&(&slave->notify_work)->work) &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)bond40 (work_completion)(&(&slave->notify_work)->work) &obj_hash[i].lock irq_context: 0 (wq_completion)bond40 (work_completion)(&(&slave->notify_work)->work) &base->lock irq_context: 0 (wq_completion)bond40 (work_completion)(&(&slave->notify_work)->work) &base->lock &obj_hash[i].lock irq_context: 0 (wq_completion)bond40 (work_completion)(&(&bond->mcast_work)->work) irq_context: 0 (wq_completion)bond40 (work_completion)(&(&bond->mcast_work)->work) &rq->__lock irq_context: 0 (wq_completion)bond40 (work_completion)(&(&bond->mcast_work)->work) &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)bond40 (work_completion)(&(&bond->mcast_work)->work) &obj_hash[i].lock irq_context: 0 (wq_completion)bond40 (work_completion)(&(&bond->mcast_work)->work) &base->lock irq_context: 0 (wq_completion)bond40 (work_completion)(&(&bond->mcast_work)->work) &base->lock &obj_hash[i].lock irq_context: 0 (wq_completion)bond157 (work_completion)(&(&bond->mcast_work)->work) &rq->__lock irq_context: 0 (wq_completion)bond157 (work_completion)(&(&bond->mcast_work)->work) &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)bond129 &rq->__lock irq_context: 0 (wq_completion)bond129 &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)bond161 irq_context: 0 (wq_completion)bond161 (work_completion)(&(&slave->notify_work)->work) irq_context: 0 (wq_completion)bond161 (work_completion)(&(&slave->notify_work)->work) &obj_hash[i].lock irq_context: 0 (wq_completion)bond161 (work_completion)(&(&slave->notify_work)->work) &base->lock irq_context: 0 (wq_completion)bond161 (work_completion)(&(&slave->notify_work)->work) &base->lock &obj_hash[i].lock irq_context: 0 (wq_completion)bond161 (work_completion)(&(&slave->notify_work)->work) &rq->__lock irq_context: 0 (wq_completion)bond161 (work_completion)(&(&slave->notify_work)->work) &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)bond131 &rq->__lock irq_context: 0 (wq_completion)bond131 &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)bond161 (work_completion)(&(&bond->mcast_work)->work) irq_context: 0 (wq_completion)bond161 (work_completion)(&(&bond->mcast_work)->work) &obj_hash[i].lock irq_context: 0 (wq_completion)bond161 (work_completion)(&(&bond->mcast_work)->work) &base->lock irq_context: 0 (wq_completion)bond161 (work_completion)(&(&bond->mcast_work)->work) &base->lock &obj_hash[i].lock irq_context: 0 (wq_completion)bond161 (work_completion)(&(&bond->mcast_work)->work) &rq->__lock irq_context: 0 (wq_completion)bond161 (work_completion)(&(&bond->mcast_work)->work) &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)bond148#2 (work_completion)(&(&slave->notify_work)->work) rtnl_mutex net_rwsem irq_context: 0 (wq_completion)bond148#2 (work_completion)(&(&slave->notify_work)->work) rtnl_mutex net_rwsem &list->lock#2 irq_context: 0 (wq_completion)bond148#2 (work_completion)(&(&slave->notify_work)->work) rtnl_mutex &tn->lock irq_context: 0 (wq_completion)bond148#2 (work_completion)(&(&slave->notify_work)->work) rtnl_mutex.wait_lock irq_context: 0 (wq_completion)bond148#2 (work_completion)(&(&slave->notify_work)->work) &p->pi_lock irq_context: 0 (wq_completion)bond31#3 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex fs_reclaim irq_context: 0 (wq_completion)bond31#3 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex fs_reclaim mmu_notifier_invalidate_range_start irq_context: 0 (wq_completion)bond31#3 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex rcu_read_lock &bond->stats_lock/1 irq_context: 0 (wq_completion)bond31#3 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex rcu_read_lock &ndev->lock irq_context: 0 (wq_completion)bond31#3 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex nl_table_lock irq_context: 0 (wq_completion)bond31#3 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &obj_hash[i].lock irq_context: 0 (wq_completion)bond31#3 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex nl_table_wait.lock irq_context: 0 (wq_completion)bond31#3 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex net_rwsem irq_context: 0 (wq_completion)bond31#3 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex net_rwsem &list->lock#2 irq_context: 0 (wq_completion)bond31#3 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &tn->lock irq_context: 0 (wq_completion)bond31#3 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &idev->mc_lock irq_context: 0 (wq_completion)bond31#3 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &idev->mc_lock fs_reclaim irq_context: 0 (wq_completion)bond31#3 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &idev->mc_lock fs_reclaim mmu_notifier_invalidate_range_start irq_context: 0 (wq_completion)bond31#3 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &idev->mc_lock rcu_read_lock &ndev->lock irq_context: 0 (wq_completion)bond31#3 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &idev->mc_lock rcu_read_lock &dir->lock#2 irq_context: 0 (wq_completion)bond31#3 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &idev->mc_lock rcu_read_lock &ul->lock#2 irq_context: 0 (wq_completion)bond31#3 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &idev->mc_lock rcu_read_lock rcu_read_lock &____s->seqcount#12 irq_context: 0 (wq_completion)bond31#3 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &idev->mc_lock rcu_read_lock rcu_read_lock &____s->seqcount#16 irq_context: 0 (wq_completion)bond31#3 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &idev->mc_lock rcu_read_lock &obj_hash[i].lock irq_context: 0 (wq_completion)bond31#3 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &idev->mc_lock rcu_read_lock &data->lock irq_context: 0 (wq_completion)bond31#3 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex.wait_lock irq_context: 0 (wq_completion)bond31#3 (work_completion)(&(&bond->mcast_work)->work) &p->pi_lock irq_context: 0 (wq_completion)bond143#2 &rq->__lock irq_context: 0 (wq_completion)bond143#2 &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)bond133#3 irq_context: 0 (wq_completion)bond133#3 (work_completion)(&(&slave->notify_work)->work) irq_context: 0 (wq_completion)bond133#3 (work_completion)(&(&slave->notify_work)->work) &obj_hash[i].lock irq_context: 0 (wq_completion)bond133#3 (work_completion)(&(&slave->notify_work)->work) &base->lock irq_context: 0 (wq_completion)bond133#3 (work_completion)(&(&slave->notify_work)->work) &base->lock &obj_hash[i].lock irq_context: 0 (wq_completion)bond133#3 (work_completion)(&(&slave->notify_work)->work) &rq->__lock irq_context: 0 (wq_completion)bond133#3 (work_completion)(&(&slave->notify_work)->work) &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)bond22#4 &rq->__lock irq_context: 0 (wq_completion)bond22#4 &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)bond154 (work_completion)(&(&bond->mcast_work)->work) fill_pool_map-wait-type-override pool_lock irq_context: 0 (wq_completion)bond133#3 (work_completion)(&(&bond->mcast_work)->work) irq_context: 0 (wq_completion)bond133#3 (work_completion)(&(&bond->mcast_work)->work) &obj_hash[i].lock irq_context: 0 (wq_completion)bond133#3 (work_completion)(&(&bond->mcast_work)->work) &base->lock irq_context: 0 (wq_completion)bond133#3 (work_completion)(&(&bond->mcast_work)->work) &base->lock &obj_hash[i].lock irq_context: 0 (wq_completion)bond133#3 (work_completion)(&(&bond->mcast_work)->work) &rq->__lock irq_context: 0 (wq_completion)bond133#3 (work_completion)(&(&bond->mcast_work)->work) &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)bond149#2 irq_context: 0 (wq_completion)bond149#2 (work_completion)(&(&slave->notify_work)->work) irq_context: 0 (wq_completion)bond149#2 (work_completion)(&(&slave->notify_work)->work) &obj_hash[i].lock irq_context: 0 (wq_completion)bond149#2 (work_completion)(&(&slave->notify_work)->work) &base->lock irq_context: 0 (wq_completion)bond149#2 (work_completion)(&(&slave->notify_work)->work) &base->lock &obj_hash[i].lock irq_context: 0 (wq_completion)bond148#2 (work_completion)(&(&bond->mcast_work)->work) &rq->__lock irq_context: 0 (wq_completion)bond148#2 (work_completion)(&(&bond->mcast_work)->work) &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)bond145#2 (work_completion)(&(&slave->notify_work)->work) &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)bond149#2 (work_completion)(&(&bond->mcast_work)->work) irq_context: 0 (wq_completion)bond149#2 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex fs_reclaim irq_context: 0 (wq_completion)bond149#2 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex fs_reclaim mmu_notifier_invalidate_range_start irq_context: 0 (wq_completion)bond149#2 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex rcu_read_lock &bond->stats_lock/1 irq_context: 0 (wq_completion)bond149#2 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex rcu_read_lock &ndev->lock irq_context: 0 (wq_completion)bond149#2 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex nl_table_lock irq_context: 0 (wq_completion)bond149#2 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &obj_hash[i].lock irq_context: 0 (wq_completion)bond149#2 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex nl_table_wait.lock irq_context: 0 (wq_completion)bond149#2 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex net_rwsem irq_context: 0 (wq_completion)bond149#2 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex net_rwsem &list->lock#2 irq_context: 0 (wq_completion)bond149#2 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &tn->lock irq_context: 0 (wq_completion)bond149#2 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &idev->mc_lock irq_context: 0 (wq_completion)bond149#2 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex.wait_lock irq_context: 0 (wq_completion)bond149#2 (work_completion)(&(&bond->mcast_work)->work) &p->pi_lock irq_context: 0 (wq_completion)bond132 (work_completion)(&(&slave->notify_work)->work) rtnl_mutex net_rwsem irq_context: 0 (wq_completion)bond132 (work_completion)(&(&slave->notify_work)->work) rtnl_mutex net_rwsem &list->lock#2 irq_context: 0 (wq_completion)bond132 (work_completion)(&(&slave->notify_work)->work) rtnl_mutex &tn->lock irq_context: 0 (wq_completion)bond132 (work_completion)(&(&slave->notify_work)->work) rtnl_mutex.wait_lock irq_context: 0 (wq_completion)bond132 (work_completion)(&(&slave->notify_work)->work) &p->pi_lock irq_context: 0 (wq_completion)bond130#3 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex fs_reclaim irq_context: 0 (wq_completion)bond130#3 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex fs_reclaim mmu_notifier_invalidate_range_start irq_context: 0 (wq_completion)bond130#3 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex fs_reclaim mmu_notifier_invalidate_range_start &rq->__lock irq_context: 0 (wq_completion)bond130#3 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex fs_reclaim mmu_notifier_invalidate_range_start &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)bond130#3 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &c->lock irq_context: 0 (wq_completion)bond130#3 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex rcu_read_lock &bond->stats_lock/1 irq_context: 0 (wq_completion)bond130#3 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex rcu_read_lock &ndev->lock irq_context: 0 (wq_completion)bond130#3 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex nl_table_lock irq_context: 0 (wq_completion)bond130#3 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &obj_hash[i].lock irq_context: 0 (wq_completion)bond130#3 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex nl_table_wait.lock irq_context: 0 (wq_completion)bond130#3 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex net_rwsem irq_context: 0 (wq_completion)bond130#3 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex net_rwsem &list->lock#2 irq_context: 0 (wq_completion)bond130#3 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &tn->lock irq_context: 0 (wq_completion)bond130#3 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &idev->mc_lock irq_context: 0 (wq_completion)bond130#3 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &idev->mc_lock &rq->__lock irq_context: 0 (wq_completion)bond130#3 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &idev->mc_lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)bond130#3 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &idev->mc_lock fs_reclaim irq_context: 0 (wq_completion)bond130#3 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &idev->mc_lock fs_reclaim mmu_notifier_invalidate_range_start irq_context: 0 (wq_completion)bond130#3 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &idev->mc_lock rcu_read_lock &ndev->lock irq_context: 0 (wq_completion)bond130#3 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &idev->mc_lock rcu_read_lock &dir->lock#2 irq_context: 0 (wq_completion)bond130#3 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &idev->mc_lock rcu_read_lock &ul->lock#2 irq_context: 0 (wq_completion)bond130#3 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &idev->mc_lock rcu_read_lock rcu_read_lock &____s->seqcount#10 irq_context: 0 (wq_completion)bond130#3 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &idev->mc_lock rcu_read_lock rcu_read_lock rcu_read_lock_bh rcu_read_lock rcu_read_lock_bh rcu_read_lock &br->multicast_lock irq_context: 0 (wq_completion)bond130#3 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &idev->mc_lock rcu_read_lock rcu_read_lock rcu_read_lock_bh rcu_read_lock rcu_read_lock_bh rcu_read_lock &br->multicast_lock &obj_hash[i].lock irq_context: 0 (wq_completion)bond130#3 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &idev->mc_lock rcu_read_lock rcu_read_lock rcu_read_lock_bh rcu_read_lock rcu_read_lock_bh rcu_read_lock &br->multicast_lock &base->lock irq_context: 0 (wq_completion)bond130#3 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &idev->mc_lock rcu_read_lock rcu_read_lock rcu_read_lock_bh rcu_read_lock rcu_read_lock_bh rcu_read_lock &data->lock irq_context: 0 (wq_completion)bond130#3 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &idev->mc_lock rcu_read_lock rcu_read_lock rcu_read_lock_bh rcu_read_lock rcu_read_lock_bh rcu_read_lock &obj_hash[i].lock irq_context: 0 (wq_completion)bond130#3 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &rq->__lock irq_context: 0 (wq_completion)bond130#3 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)bond130#3 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex.wait_lock irq_context: 0 (wq_completion)bond130#3 (work_completion)(&(&bond->mcast_work)->work) &p->pi_lock irq_context: 0 (wq_completion)bond130#3 (work_completion)(&(&bond->mcast_work)->work) &p->pi_lock &rq->__lock irq_context: 0 (wq_completion)bond130#3 (work_completion)(&(&bond->mcast_work)->work) &p->pi_lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)bond130#3 (work_completion)(&(&bond->mcast_work)->work) &rq->__lock irq_context: 0 (wq_completion)bond130#3 (work_completion)(&(&bond->mcast_work)->work) &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)bond148 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex fs_reclaim irq_context: 0 (wq_completion)bond148 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex fs_reclaim mmu_notifier_invalidate_range_start irq_context: 0 (wq_completion)bond148 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &c->lock irq_context: 0 (wq_completion)bond148 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex rcu_read_lock &bond->stats_lock/1 irq_context: 0 (wq_completion)bond148 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex rcu_read_lock &ndev->lock irq_context: 0 (wq_completion)bond148 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex nl_table_lock irq_context: 0 (wq_completion)bond148 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &obj_hash[i].lock irq_context: 0 (wq_completion)bond148 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex nl_table_wait.lock irq_context: 0 (wq_completion)bond148 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex net_rwsem irq_context: 0 (wq_completion)bond148 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex net_rwsem &list->lock#2 irq_context: 0 (wq_completion)bond148 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &tn->lock irq_context: 0 (wq_completion)bond148 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &idev->mc_lock irq_context: 0 (wq_completion)bond148 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &idev->mc_lock fs_reclaim irq_context: 0 (wq_completion)bond148 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &idev->mc_lock fs_reclaim mmu_notifier_invalidate_range_start irq_context: 0 (wq_completion)bond148 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &idev->mc_lock &c->lock irq_context: 0 (wq_completion)bond148 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &idev->mc_lock rcu_read_lock &ndev->lock irq_context: 0 (wq_completion)bond148 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &idev->mc_lock rcu_read_lock &dir->lock#2 irq_context: 0 (wq_completion)bond148 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &idev->mc_lock rcu_read_lock &ul->lock#2 irq_context: 0 (wq_completion)bond148 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &idev->mc_lock rcu_read_lock rcu_read_lock &____s->seqcount#10 irq_context: 0 (wq_completion)bond148 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &idev->mc_lock rcu_read_lock rcu_read_lock rcu_read_lock_bh rcu_read_lock rcu_read_lock_bh rcu_read_lock &br->multicast_lock irq_context: 0 (wq_completion)bond148 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &idev->mc_lock rcu_read_lock rcu_read_lock rcu_read_lock_bh rcu_read_lock rcu_read_lock_bh rcu_read_lock &br->multicast_lock &obj_hash[i].lock irq_context: 0 (wq_completion)bond148 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &idev->mc_lock rcu_read_lock rcu_read_lock rcu_read_lock_bh rcu_read_lock rcu_read_lock_bh rcu_read_lock &br->multicast_lock &base->lock irq_context: 0 (wq_completion)bond148 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &idev->mc_lock rcu_read_lock rcu_read_lock rcu_read_lock_bh rcu_read_lock rcu_read_lock_bh rcu_read_lock &br->multicast_lock &base->lock &obj_hash[i].lock irq_context: 0 (wq_completion)bond148 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &idev->mc_lock rcu_read_lock rcu_read_lock rcu_read_lock_bh rcu_read_lock rcu_read_lock_bh rcu_read_lock &data->lock irq_context: 0 (wq_completion)bond148 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &idev->mc_lock rcu_read_lock rcu_read_lock rcu_read_lock_bh rcu_read_lock rcu_read_lock_bh rcu_read_lock &obj_hash[i].lock irq_context: 0 (wq_completion)bond148 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex.wait_lock irq_context: 0 (wq_completion)bond148 (work_completion)(&(&bond->mcast_work)->work) &p->pi_lock irq_context: 0 (wq_completion)bond37#2 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex fs_reclaim irq_context: 0 (wq_completion)bond37#2 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex fs_reclaim mmu_notifier_invalidate_range_start irq_context: 0 (wq_completion)bond37#2 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &c->lock irq_context: 0 (wq_completion)bond37#2 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex rcu_read_lock &bond->stats_lock/1 irq_context: 0 (wq_completion)bond37#2 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex rcu_read_lock &ndev->lock irq_context: 0 (wq_completion)bond37#2 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex nl_table_lock irq_context: 0 (wq_completion)bond37#2 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &obj_hash[i].lock irq_context: 0 (wq_completion)bond37#2 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex nl_table_wait.lock irq_context: 0 (wq_completion)bond37#2 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex net_rwsem irq_context: 0 (wq_completion)bond37#2 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex net_rwsem &list->lock#2 irq_context: 0 (wq_completion)bond37#2 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &tn->lock irq_context: 0 (wq_completion)bond37#2 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &idev->mc_lock irq_context: 0 (wq_completion)bond37#2 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &idev->mc_lock fs_reclaim irq_context: 0 (wq_completion)bond37#2 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &idev->mc_lock fs_reclaim mmu_notifier_invalidate_range_start irq_context: 0 (wq_completion)bond37#2 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &idev->mc_lock &c->lock irq_context: 0 (wq_completion)bond37#2 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &idev->mc_lock &n->list_lock irq_context: 0 (wq_completion)bond37#2 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &idev->mc_lock &n->list_lock &c->lock irq_context: 0 (wq_completion)bond37#2 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &idev->mc_lock rcu_read_lock &ndev->lock irq_context: 0 (wq_completion)bond37#2 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &idev->mc_lock rcu_read_lock rcu_node_0 irq_context: 0 (wq_completion)bond37#2 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &idev->mc_lock rcu_read_lock &rq->__lock irq_context: 0 (wq_completion)bond37#2 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &idev->mc_lock rcu_read_lock &dir->lock#2 irq_context: 0 (wq_completion)bond37#2 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &idev->mc_lock rcu_read_lock &ul->lock#2 irq_context: 0 (wq_completion)bond37#2 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &idev->mc_lock rcu_read_lock rcu_read_lock &____s->seqcount#12 irq_context: 0 (wq_completion)bond37#2 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &idev->mc_lock rcu_read_lock rcu_read_lock &____s->seqcount#16 irq_context: 0 (wq_completion)bond37#2 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &idev->mc_lock rcu_read_lock rcu_read_lock &c->lock irq_context: 0 (wq_completion)bond37#2 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &idev->mc_lock rcu_read_lock &obj_hash[i].lock irq_context: 0 (wq_completion)bond37#2 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &idev->mc_lock rcu_read_lock &data->lock irq_context: 0 (wq_completion)bond37#2 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex.wait_lock irq_context: 0 (wq_completion)bond37#2 (work_completion)(&(&bond->mcast_work)->work) &p->pi_lock irq_context: 0 (wq_completion)bond128 (work_completion)(&(&slave->notify_work)->work) rtnl_mutex net_rwsem irq_context: 0 (wq_completion)bond128 (work_completion)(&(&slave->notify_work)->work) rtnl_mutex net_rwsem &list->lock#2 irq_context: 0 (wq_completion)bond128 (work_completion)(&(&slave->notify_work)->work) rtnl_mutex &tn->lock irq_context: 0 (wq_completion)bond128 (work_completion)(&(&slave->notify_work)->work) rtnl_mutex.wait_lock irq_context: 0 (wq_completion)bond128 (work_completion)(&(&slave->notify_work)->work) &p->pi_lock irq_context: 0 (wq_completion)bond96#3 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex fs_reclaim irq_context: 0 (wq_completion)bond96#3 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex fs_reclaim mmu_notifier_invalidate_range_start irq_context: 0 (wq_completion)bond96#3 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex rcu_read_lock &bond->stats_lock/1 irq_context: 0 (wq_completion)bond96#3 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex rcu_read_lock &ndev->lock irq_context: 0 (wq_completion)bond96#3 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex nl_table_lock irq_context: 0 (wq_completion)bond96#3 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &obj_hash[i].lock irq_context: 0 (wq_completion)bond96#3 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex nl_table_wait.lock irq_context: 0 (wq_completion)bond96#3 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex net_rwsem irq_context: 0 (wq_completion)bond96#3 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex net_rwsem &list->lock#2 irq_context: 0 (wq_completion)bond96#3 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &tn->lock irq_context: 0 (wq_completion)bond96#3 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &idev->mc_lock irq_context: 0 (wq_completion)bond96#3 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &idev->mc_lock fs_reclaim irq_context: 0 (wq_completion)bond96#3 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &idev->mc_lock fs_reclaim mmu_notifier_invalidate_range_start irq_context: 0 (wq_completion)bond96#3 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &idev->mc_lock &c->lock irq_context: 0 (wq_completion)bond96#3 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &idev->mc_lock rcu_read_lock &ndev->lock irq_context: 0 (wq_completion)bond96#3 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &idev->mc_lock rcu_read_lock &dir->lock#2 irq_context: 0 (wq_completion)bond96#3 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &idev->mc_lock rcu_read_lock &ul->lock#2 irq_context: 0 (wq_completion)bond96#3 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &idev->mc_lock rcu_read_lock rcu_read_lock &____s->seqcount#10 irq_context: 0 (wq_completion)bond96#3 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &idev->mc_lock rcu_read_lock rcu_read_lock rcu_read_lock_bh rcu_read_lock rcu_read_lock_bh rcu_read_lock &br->multicast_lock irq_context: 0 (wq_completion)bond96#3 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &idev->mc_lock rcu_read_lock rcu_read_lock rcu_read_lock_bh rcu_read_lock rcu_read_lock_bh rcu_read_lock &br->multicast_lock &obj_hash[i].lock irq_context: 0 (wq_completion)bond96#3 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &idev->mc_lock rcu_read_lock rcu_read_lock rcu_read_lock_bh rcu_read_lock rcu_read_lock_bh rcu_read_lock &br->multicast_lock &base->lock irq_context: 0 (wq_completion)bond96#3 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &idev->mc_lock rcu_read_lock rcu_read_lock rcu_read_lock_bh rcu_read_lock rcu_read_lock_bh rcu_read_lock &br->multicast_lock &base->lock &obj_hash[i].lock irq_context: 0 (wq_completion)bond96#3 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &idev->mc_lock rcu_read_lock rcu_read_lock rcu_read_lock_bh rcu_read_lock rcu_read_lock_bh rcu_read_lock &data->lock irq_context: 0 (wq_completion)bond96#3 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &idev->mc_lock rcu_read_lock rcu_read_lock rcu_read_lock_bh rcu_read_lock rcu_read_lock_bh rcu_read_lock &obj_hash[i].lock irq_context: 0 (wq_completion)bond96#3 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex.wait_lock irq_context: 0 (wq_completion)bond96#3 (work_completion)(&(&bond->mcast_work)->work) &p->pi_lock irq_context: 0 (wq_completion)bond108#3 (work_completion)(&(&slave->notify_work)->work) rtnl_mutex net_rwsem irq_context: 0 (wq_completion)bond108#3 (work_completion)(&(&slave->notify_work)->work) rtnl_mutex net_rwsem &list->lock#2 irq_context: 0 (wq_completion)bond108#3 (work_completion)(&(&slave->notify_work)->work) rtnl_mutex &tn->lock irq_context: 0 (wq_completion)bond108#3 (work_completion)(&(&slave->notify_work)->work) rtnl_mutex.wait_lock irq_context: 0 (wq_completion)bond108#3 (work_completion)(&(&slave->notify_work)->work) &p->pi_lock irq_context: 0 (wq_completion)bond38#2 (work_completion)(&(&slave->notify_work)->work) rtnl_mutex net_rwsem irq_context: 0 (wq_completion)bond38#2 (work_completion)(&(&slave->notify_work)->work) rtnl_mutex net_rwsem &list->lock#2 irq_context: 0 (wq_completion)bond38#2 (work_completion)(&(&slave->notify_work)->work) rtnl_mutex &tn->lock irq_context: 0 (wq_completion)bond38#2 (work_completion)(&(&slave->notify_work)->work) rtnl_mutex.wait_lock irq_context: 0 (wq_completion)bond38#2 (work_completion)(&(&slave->notify_work)->work) &p->pi_lock irq_context: 0 (wq_completion)bond147 (work_completion)(&(&slave->notify_work)->work) rtnl_mutex net_rwsem irq_context: 0 (wq_completion)bond147 (work_completion)(&(&slave->notify_work)->work) rtnl_mutex net_rwsem &list->lock#2 irq_context: 0 (wq_completion)bond147 (work_completion)(&(&slave->notify_work)->work) rtnl_mutex &tn->lock irq_context: 0 (wq_completion)bond147 (work_completion)(&(&slave->notify_work)->work) rtnl_mutex.wait_lock irq_context: 0 (wq_completion)bond147 (work_completion)(&(&slave->notify_work)->work) &p->pi_lock irq_context: 0 (wq_completion)bond125#3 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex fs_reclaim irq_context: 0 (wq_completion)bond125#3 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex fs_reclaim mmu_notifier_invalidate_range_start irq_context: 0 (wq_completion)bond125#3 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &c->lock irq_context: 0 (wq_completion)bond125#3 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &n->list_lock irq_context: 0 (wq_completion)bond125#3 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &n->list_lock &c->lock irq_context: 0 (wq_completion)bond125#3 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex rcu_read_lock &bond->stats_lock/1 irq_context: 0 (wq_completion)bond125#3 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex rcu_read_lock &ndev->lock irq_context: 0 (wq_completion)bond125#3 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex nl_table_lock irq_context: 0 (wq_completion)bond125#3 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &obj_hash[i].lock irq_context: 0 (wq_completion)bond125#3 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex nl_table_wait.lock irq_context: 0 (wq_completion)bond125#3 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex net_rwsem irq_context: 0 (wq_completion)bond125#3 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex net_rwsem &list->lock#2 irq_context: 0 (wq_completion)bond125#3 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &tn->lock irq_context: 0 (wq_completion)bond125#3 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &idev->mc_lock irq_context: 0 (wq_completion)bond125#3 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &idev->mc_lock fs_reclaim irq_context: 0 (wq_completion)bond125#3 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &idev->mc_lock fs_reclaim mmu_notifier_invalidate_range_start irq_context: 0 (wq_completion)bond125#3 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &idev->mc_lock &c->lock irq_context: 0 (wq_completion)bond125#3 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &idev->mc_lock rcu_read_lock &ndev->lock irq_context: 0 (wq_completion)bond125#3 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &idev->mc_lock rcu_read_lock &c->lock irq_context: 0 (wq_completion)bond125#3 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &idev->mc_lock rcu_read_lock &dir->lock#2 irq_context: 0 (wq_completion)bond125#3 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &idev->mc_lock rcu_read_lock &ul->lock#2 irq_context: 0 (wq_completion)bond125#3 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &idev->mc_lock rcu_read_lock rcu_read_lock &____s->seqcount#10 irq_context: 0 (wq_completion)bond125#3 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &idev->mc_lock rcu_read_lock rcu_read_lock rcu_read_lock_bh rcu_read_lock rcu_read_lock_bh rcu_read_lock &br->multicast_lock irq_context: 0 (wq_completion)bond125#3 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &idev->mc_lock rcu_read_lock rcu_read_lock rcu_read_lock_bh rcu_read_lock rcu_read_lock_bh rcu_read_lock &br->multicast_lock &obj_hash[i].lock irq_context: 0 (wq_completion)bond125#3 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &idev->mc_lock rcu_read_lock rcu_read_lock rcu_read_lock_bh rcu_read_lock rcu_read_lock_bh rcu_read_lock &br->multicast_lock &base->lock irq_context: 0 (wq_completion)bond125#3 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &idev->mc_lock rcu_read_lock rcu_read_lock rcu_read_lock_bh rcu_read_lock rcu_read_lock_bh rcu_read_lock &data->lock irq_context: 0 (wq_completion)bond125#3 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &idev->mc_lock rcu_read_lock rcu_read_lock rcu_read_lock_bh rcu_read_lock rcu_read_lock_bh rcu_read_lock &obj_hash[i].lock irq_context: 0 (wq_completion)bond125#3 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex.wait_lock irq_context: 0 (wq_completion)bond84 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex fs_reclaim irq_context: 0 (wq_completion)bond125#3 (work_completion)(&(&bond->mcast_work)->work) &p->pi_lock irq_context: 0 (wq_completion)bond84 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex fs_reclaim mmu_notifier_invalidate_range_start irq_context: 0 (wq_completion)bond84 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &c->lock irq_context: 0 (wq_completion)bond84 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex rcu_read_lock &bond->stats_lock/1 irq_context: 0 (wq_completion)bond84 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex rcu_read_lock &ndev->lock irq_context: 0 (wq_completion)bond84 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex nl_table_lock irq_context: 0 (wq_completion)bond84 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &obj_hash[i].lock irq_context: 0 (wq_completion)bond84 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex nl_table_wait.lock irq_context: 0 (wq_completion)bond84 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex net_rwsem irq_context: 0 (wq_completion)bond84 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex net_rwsem &list->lock#2 irq_context: 0 (wq_completion)bond84 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &tn->lock irq_context: 0 (wq_completion)bond84 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &idev->mc_lock irq_context: 0 (wq_completion)bond84 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &idev->mc_lock fs_reclaim irq_context: 0 (wq_completion)bond84 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &idev->mc_lock fs_reclaim mmu_notifier_invalidate_range_start irq_context: 0 (wq_completion)bond84 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &idev->mc_lock fs_reclaim mmu_notifier_invalidate_range_start &rq->__lock irq_context: 0 (wq_completion)bond84 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &idev->mc_lock &c->lock irq_context: 0 (wq_completion)bond84 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &idev->mc_lock rcu_read_lock &ndev->lock irq_context: 0 (wq_completion)bond84 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &idev->mc_lock rcu_read_lock &c->lock irq_context: 0 (wq_completion)bond84 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &idev->mc_lock rcu_read_lock &dir->lock#2 irq_context: 0 (wq_completion)bond84 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &idev->mc_lock rcu_read_lock &ul->lock#2 irq_context: 0 (wq_completion)bond84 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &idev->mc_lock rcu_read_lock rcu_read_lock &obj_hash[i].lock irq_context: 0 (wq_completion)bond84 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &idev->mc_lock rcu_read_lock rcu_read_lock &dir->lock#2 irq_context: 0 (wq_completion)bond84 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &idev->mc_lock rcu_read_lock rcu_read_lock &tbl->lock irq_context: 0 (wq_completion)bond84 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &idev->mc_lock rcu_read_lock rcu_read_lock &n->lock irq_context: 0 (wq_completion)bond84 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &idev->mc_lock rcu_read_lock rcu_read_lock &____s->seqcount#9 irq_context: 0 (wq_completion)bond84 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &idev->mc_lock rcu_read_lock rcu_read_lock rcu_read_lock_bh rcu_read_lock rcu_read_lock_bh rcu_read_lock &br->multicast_lock irq_context: 0 (wq_completion)bond84 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &idev->mc_lock rcu_read_lock rcu_read_lock rcu_read_lock_bh rcu_read_lock rcu_read_lock_bh rcu_read_lock &br->multicast_lock &obj_hash[i].lock irq_context: 0 (wq_completion)bond84 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &idev->mc_lock rcu_read_lock rcu_read_lock rcu_read_lock_bh rcu_read_lock rcu_read_lock_bh rcu_read_lock &br->multicast_lock &base->lock irq_context: 0 (wq_completion)bond84 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &idev->mc_lock rcu_read_lock rcu_read_lock rcu_read_lock_bh rcu_read_lock rcu_read_lock_bh rcu_read_lock &br->multicast_lock &base->lock &obj_hash[i].lock irq_context: 0 (wq_completion)bond84 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &idev->mc_lock rcu_read_lock rcu_read_lock rcu_read_lock_bh rcu_read_lock rcu_read_lock_bh rcu_read_lock &data->lock irq_context: 0 (wq_completion)bond84 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &idev->mc_lock rcu_read_lock rcu_read_lock rcu_read_lock_bh rcu_read_lock rcu_read_lock_bh rcu_read_lock &obj_hash[i].lock irq_context: 0 (wq_completion)bond84 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex.wait_lock irq_context: 0 (wq_completion)bond84 (work_completion)(&(&bond->mcast_work)->work) &p->pi_lock irq_context: 0 (wq_completion)bond162 irq_context: 0 (wq_completion)bond162 (work_completion)(&(&slave->notify_work)->work) irq_context: 0 (wq_completion)bond162 (work_completion)(&(&slave->notify_work)->work) &obj_hash[i].lock irq_context: 0 (wq_completion)bond162 (work_completion)(&(&slave->notify_work)->work) &base->lock irq_context: 0 (wq_completion)bond162 (work_completion)(&(&slave->notify_work)->work) &base->lock &obj_hash[i].lock irq_context: 0 (wq_completion)bond162 (work_completion)(&(&slave->notify_work)->work) &rq->__lock irq_context: 0 (wq_completion)bond162 (work_completion)(&(&bond->mcast_work)->work) irq_context: 0 (wq_completion)bond162 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex fs_reclaim irq_context: 0 (wq_completion)bond162 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex fs_reclaim mmu_notifier_invalidate_range_start irq_context: 0 (wq_completion)bond162 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &c->lock irq_context: 0 (wq_completion)bond162 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex rcu_read_lock &bond->stats_lock/1 irq_context: 0 (wq_completion)bond162 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex rcu_read_lock &ndev->lock irq_context: 0 (wq_completion)bond162 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex nl_table_lock irq_context: 0 (wq_completion)bond162 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &obj_hash[i].lock irq_context: 0 (wq_completion)bond162 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex nl_table_wait.lock irq_context: 0 (wq_completion)bond162 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex net_rwsem irq_context: 0 (wq_completion)bond162 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex net_rwsem &list->lock#2 irq_context: 0 (wq_completion)bond162 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &tn->lock irq_context: 0 (wq_completion)bond162 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &idev->mc_lock irq_context: 0 (wq_completion)bond162 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex.wait_lock irq_context: 0 (wq_completion)bond162 (work_completion)(&(&bond->mcast_work)->work) &p->pi_lock irq_context: 0 (wq_completion)bond101#3 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex fs_reclaim irq_context: 0 (wq_completion)bond101#3 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex fs_reclaim mmu_notifier_invalidate_range_start irq_context: 0 (wq_completion)bond101#3 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &c->lock irq_context: 0 (wq_completion)bond101#3 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex rcu_read_lock &bond->stats_lock/1 irq_context: 0 (wq_completion)bond101#3 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex rcu_read_lock &ndev->lock irq_context: 0 (wq_completion)bond101#3 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex nl_table_lock irq_context: 0 (wq_completion)bond101#3 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &obj_hash[i].lock irq_context: 0 (wq_completion)bond101#3 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex nl_table_wait.lock irq_context: 0 (wq_completion)bond101#3 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex net_rwsem irq_context: 0 (wq_completion)bond101#3 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex net_rwsem &list->lock#2 irq_context: 0 (wq_completion)bond101#3 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &tn->lock irq_context: 0 (wq_completion)bond101#3 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &idev->mc_lock irq_context: 0 (wq_completion)bond101#3 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &idev->mc_lock fs_reclaim irq_context: 0 (wq_completion)bond101#3 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &idev->mc_lock fs_reclaim mmu_notifier_invalidate_range_start irq_context: 0 (wq_completion)bond101#3 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &idev->mc_lock &c->lock irq_context: 0 (wq_completion)bond101#3 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &idev->mc_lock rcu_read_lock &ndev->lock irq_context: 0 (wq_completion)bond101#3 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &idev->mc_lock rcu_read_lock &c->lock irq_context: 0 (wq_completion)bond101#3 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &idev->mc_lock rcu_read_lock &dir->lock#2 irq_context: 0 (wq_completion)bond101#3 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &idev->mc_lock rcu_read_lock &ul->lock#2 irq_context: 0 (wq_completion)bond101#3 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &idev->mc_lock rcu_read_lock rcu_read_lock &____s->seqcount#10 irq_context: 0 (wq_completion)bond101#3 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &idev->mc_lock rcu_read_lock rcu_read_lock rcu_read_lock_bh rcu_read_lock rcu_read_lock_bh rcu_read_lock &br->multicast_lock irq_context: 0 (wq_completion)bond101#3 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &idev->mc_lock rcu_read_lock rcu_read_lock rcu_read_lock_bh rcu_read_lock rcu_read_lock_bh rcu_read_lock &br->multicast_lock &obj_hash[i].lock irq_context: 0 (wq_completion)bond101#3 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &idev->mc_lock rcu_read_lock rcu_read_lock rcu_read_lock_bh rcu_read_lock rcu_read_lock_bh rcu_read_lock &br->multicast_lock &base->lock irq_context: 0 (wq_completion)bond101#3 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &idev->mc_lock rcu_read_lock rcu_read_lock rcu_read_lock_bh rcu_read_lock rcu_read_lock_bh rcu_read_lock &br->multicast_lock &base->lock &obj_hash[i].lock irq_context: 0 (wq_completion)bond101#3 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &idev->mc_lock rcu_read_lock rcu_read_lock rcu_read_lock_bh rcu_read_lock rcu_read_lock_bh rcu_read_lock &data->lock irq_context: 0 (wq_completion)bond101#3 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &idev->mc_lock rcu_read_lock rcu_read_lock rcu_read_lock_bh rcu_read_lock rcu_read_lock_bh rcu_read_lock &obj_hash[i].lock irq_context: 0 (wq_completion)bond101#3 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex.wait_lock irq_context: 0 (wq_completion)bond101#3 (work_completion)(&(&bond->mcast_work)->work) &p->pi_lock irq_context: 0 (wq_completion)bond147#2 &rq->__lock irq_context: 0 (wq_completion)bond147#2 &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)bond134#3 irq_context: 0 (wq_completion)bond134#3 (work_completion)(&(&slave->notify_work)->work) irq_context: 0 (wq_completion)bond134#3 (work_completion)(&(&slave->notify_work)->work) &obj_hash[i].lock irq_context: 0 (wq_completion)bond134#3 (work_completion)(&(&slave->notify_work)->work) &base->lock irq_context: 0 (wq_completion)bond134#3 (work_completion)(&(&slave->notify_work)->work) &base->lock &obj_hash[i].lock irq_context: 0 (wq_completion)bond134#3 (work_completion)(&(&bond->mcast_work)->work) irq_context: 0 (wq_completion)bond134#3 (work_completion)(&(&bond->mcast_work)->work) &obj_hash[i].lock irq_context: 0 (wq_completion)bond134#3 (work_completion)(&(&bond->mcast_work)->work) &base->lock irq_context: 0 (wq_completion)bond134#3 (work_completion)(&(&bond->mcast_work)->work) &base->lock &obj_hash[i].lock irq_context: 0 (wq_completion)bond134#3 (work_completion)(&(&bond->mcast_work)->work) &rq->__lock irq_context: 0 (wq_completion)bond134#3 (work_completion)(&(&bond->mcast_work)->work) &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)bond150#2 irq_context: 0 (wq_completion)bond150#2 (work_completion)(&(&slave->notify_work)->work) irq_context: 0 (wq_completion)bond150#2 (work_completion)(&(&slave->notify_work)->work) &obj_hash[i].lock irq_context: 0 (wq_completion)bond150#2 (work_completion)(&(&slave->notify_work)->work) &base->lock irq_context: 0 (wq_completion)bond150#2 (work_completion)(&(&slave->notify_work)->work) &base->lock &obj_hash[i].lock irq_context: 0 (wq_completion)bond150#2 (work_completion)(&(&slave->notify_work)->work) &rq->__lock irq_context: 0 (wq_completion)bond150#2 (work_completion)(&(&slave->notify_work)->work) &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)bond111#3 (work_completion)(&(&slave->notify_work)->work) rtnl_mutex net_rwsem irq_context: 0 (wq_completion)bond111#3 (work_completion)(&(&slave->notify_work)->work) rtnl_mutex net_rwsem &list->lock#2 irq_context: 0 (wq_completion)bond111#3 (work_completion)(&(&slave->notify_work)->work) rtnl_mutex &tn->lock irq_context: 0 (wq_completion)bond111#3 (work_completion)(&(&slave->notify_work)->work) rtnl_mutex.wait_lock irq_context: 0 (wq_completion)bond111#3 (work_completion)(&(&slave->notify_work)->work) &p->pi_lock irq_context: 0 (wq_completion)bond80#2 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex fs_reclaim irq_context: 0 (wq_completion)bond80#2 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex fs_reclaim mmu_notifier_invalidate_range_start irq_context: 0 (wq_completion)bond80#2 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex rcu_read_lock &bond->stats_lock/1 irq_context: 0 (wq_completion)bond80#2 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex rcu_read_lock &ndev->lock irq_context: 0 (wq_completion)bond80#2 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex nl_table_lock irq_context: 0 (wq_completion)bond80#2 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &obj_hash[i].lock irq_context: 0 (wq_completion)bond80#2 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex nl_table_wait.lock irq_context: 0 (wq_completion)bond80#2 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex net_rwsem irq_context: 0 (wq_completion)bond80#2 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex net_rwsem &list->lock#2 irq_context: 0 (wq_completion)bond80#2 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &tn->lock irq_context: 0 (wq_completion)bond80#2 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &idev->mc_lock irq_context: 0 (wq_completion)bond80#2 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &idev->mc_lock fs_reclaim irq_context: 0 (wq_completion)bond80#2 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &idev->mc_lock fs_reclaim mmu_notifier_invalidate_range_start irq_context: 0 (wq_completion)bond80#2 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &idev->mc_lock &c->lock irq_context: 0 (wq_completion)bond80#2 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &idev->mc_lock rcu_read_lock &ndev->lock irq_context: 0 (wq_completion)bond80#2 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &idev->mc_lock rcu_read_lock &c->lock irq_context: 0 (wq_completion)bond80#2 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &idev->mc_lock rcu_read_lock &dir->lock#2 irq_context: 0 (wq_completion)bond80#2 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &idev->mc_lock rcu_read_lock &ul->lock#2 irq_context: 0 (wq_completion)bond80#2 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &idev->mc_lock rcu_read_lock rcu_read_lock &obj_hash[i].lock irq_context: 0 (wq_completion)bond80#2 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &idev->mc_lock rcu_read_lock rcu_read_lock &dir->lock#2 irq_context: 0 (wq_completion)bond80#2 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &idev->mc_lock rcu_read_lock rcu_read_lock &tbl->lock irq_context: 0 (wq_completion)bond80#2 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &idev->mc_lock rcu_read_lock rcu_read_lock &n->lock irq_context: 0 (wq_completion)bond80#2 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &idev->mc_lock rcu_read_lock rcu_read_lock &____s->seqcount#9 irq_context: 0 (wq_completion)bond80#2 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &idev->mc_lock rcu_read_lock rcu_read_lock rcu_read_lock_bh rcu_read_lock rcu_read_lock_bh rcu_read_lock &br->multicast_lock irq_context: 0 (wq_completion)bond80#2 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &idev->mc_lock rcu_read_lock rcu_read_lock rcu_read_lock_bh rcu_read_lock rcu_read_lock_bh rcu_read_lock &br->multicast_lock &obj_hash[i].lock irq_context: 0 (wq_completion)bond80#2 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &idev->mc_lock rcu_read_lock rcu_read_lock rcu_read_lock_bh rcu_read_lock rcu_read_lock_bh rcu_read_lock &br->multicast_lock &base->lock irq_context: 0 (wq_completion)bond80#2 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &idev->mc_lock rcu_read_lock rcu_read_lock rcu_read_lock_bh rcu_read_lock rcu_read_lock_bh rcu_read_lock &br->multicast_lock &base->lock &obj_hash[i].lock irq_context: 0 (wq_completion)bond80#2 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &idev->mc_lock rcu_read_lock rcu_read_lock rcu_read_lock_bh rcu_read_lock rcu_read_lock_bh rcu_read_lock &data->lock irq_context: 0 (wq_completion)bond80#2 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &idev->mc_lock rcu_read_lock rcu_read_lock rcu_read_lock_bh rcu_read_lock rcu_read_lock_bh rcu_read_lock &obj_hash[i].lock irq_context: 0 (wq_completion)bond80#2 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex.wait_lock irq_context: 0 (wq_completion)bond80#2 (work_completion)(&(&bond->mcast_work)->work) &p->pi_lock irq_context: 0 (wq_completion)bond150#2 (work_completion)(&(&bond->mcast_work)->work) irq_context: 0 (wq_completion)bond150#2 (work_completion)(&(&bond->mcast_work)->work) &obj_hash[i].lock irq_context: 0 (wq_completion)bond150#2 (work_completion)(&(&bond->mcast_work)->work) &base->lock irq_context: 0 (wq_completion)bond150#2 (work_completion)(&(&bond->mcast_work)->work) &base->lock &obj_hash[i].lock irq_context: 0 (wq_completion)bond13 (work_completion)(&(&bond->mii_work)->work) rcu_read_lock rcu_read_lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)bond117#3 &rq->__lock irq_context: 0 (wq_completion)bond117#3 &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)bond163 irq_context: 0 (wq_completion)bond163 (work_completion)(&(&slave->notify_work)->work) irq_context: 0 (wq_completion)bond163 (work_completion)(&(&slave->notify_work)->work) &obj_hash[i].lock irq_context: 0 (wq_completion)bond163 (work_completion)(&(&slave->notify_work)->work) &base->lock irq_context: 0 (wq_completion)bond163 (work_completion)(&(&slave->notify_work)->work) &base->lock &obj_hash[i].lock irq_context: 0 (wq_completion)bond163 (work_completion)(&(&slave->notify_work)->work) &rq->__lock irq_context: 0 (wq_completion)bond163 (work_completion)(&(&slave->notify_work)->work) &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)bond127#3 &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)bond163 (work_completion)(&(&bond->mcast_work)->work) irq_context: 0 (wq_completion)bond163 (work_completion)(&(&bond->mcast_work)->work) &obj_hash[i].lock irq_context: 0 (wq_completion)bond163 (work_completion)(&(&bond->mcast_work)->work) &base->lock irq_context: 0 (wq_completion)bond163 (work_completion)(&(&bond->mcast_work)->work) &base->lock &obj_hash[i].lock irq_context: 0 (wq_completion)bond41 irq_context: 0 (wq_completion)bond41 (work_completion)(&(&slave->notify_work)->work) irq_context: 0 (wq_completion)bond41 (work_completion)(&(&slave->notify_work)->work) &obj_hash[i].lock irq_context: 0 (wq_completion)bond41 (work_completion)(&(&slave->notify_work)->work) &base->lock irq_context: 0 (wq_completion)bond41 (work_completion)(&(&slave->notify_work)->work) &base->lock &obj_hash[i].lock irq_context: 0 (wq_completion)bond41 (work_completion)(&(&slave->notify_work)->work) &rq->__lock irq_context: 0 (wq_completion)bond41 (work_completion)(&(&bond->mcast_work)->work) irq_context: 0 (wq_completion)bond41 (work_completion)(&(&bond->mcast_work)->work) &obj_hash[i].lock irq_context: 0 (wq_completion)bond41 (work_completion)(&(&bond->mcast_work)->work) &base->lock irq_context: 0 (wq_completion)bond41 (work_completion)(&(&bond->mcast_work)->work) &base->lock &obj_hash[i].lock irq_context: 0 (wq_completion)bond41 (work_completion)(&(&bond->mcast_work)->work) &rq->__lock irq_context: 0 (wq_completion)bond151 irq_context: 0 (wq_completion)bond151 (work_completion)(&(&slave->notify_work)->work) irq_context: 0 (wq_completion)bond151 (work_completion)(&(&slave->notify_work)->work) &obj_hash[i].lock irq_context: 0 (wq_completion)bond151 (work_completion)(&(&slave->notify_work)->work) &base->lock irq_context: 0 (wq_completion)bond151 (work_completion)(&(&slave->notify_work)->work) &base->lock &obj_hash[i].lock irq_context: 0 (wq_completion)bond151 (work_completion)(&(&bond->mcast_work)->work) irq_context: 0 (wq_completion)bond151 (work_completion)(&(&bond->mcast_work)->work) &obj_hash[i].lock irq_context: 0 (wq_completion)bond151 (work_completion)(&(&bond->mcast_work)->work) &base->lock irq_context: 0 (wq_completion)bond151 (work_completion)(&(&bond->mcast_work)->work) &base->lock &obj_hash[i].lock irq_context: 0 (wq_completion)bond135#3 irq_context: 0 (wq_completion)bond135#3 (work_completion)(&(&slave->notify_work)->work) irq_context: 0 (wq_completion)bond135#3 (work_completion)(&(&slave->notify_work)->work) &obj_hash[i].lock irq_context: 0 (wq_completion)bond135#3 (work_completion)(&(&slave->notify_work)->work) &base->lock irq_context: 0 (wq_completion)bond135#3 (work_completion)(&(&slave->notify_work)->work) &base->lock &obj_hash[i].lock irq_context: 0 (wq_completion)bond135#3 (work_completion)(&(&slave->notify_work)->work) &rq->__lock irq_context: 0 (wq_completion)bond135#3 (work_completion)(&(&slave->notify_work)->work) &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)bond134#3 &rq->__lock irq_context: 0 (wq_completion)bond135#3 (work_completion)(&(&bond->mcast_work)->work) irq_context: 0 (wq_completion)bond135#3 (work_completion)(&(&bond->mcast_work)->work) &obj_hash[i].lock irq_context: 0 (wq_completion)bond135#3 (work_completion)(&(&bond->mcast_work)->work) &base->lock irq_context: 0 (wq_completion)bond135#3 (work_completion)(&(&bond->mcast_work)->work) &base->lock &obj_hash[i].lock irq_context: 0 (wq_completion)bond124#3 (work_completion)(&(&bond->mcast_work)->work) &rq->__lock irq_context: 0 (wq_completion)bond99#3 (work_completion)(&(&bond->mcast_work)->work) &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)bond150#2 (work_completion)(&(&bond->mcast_work)->work) &rq->__lock irq_context: 0 (wq_completion)bond150#2 (work_completion)(&(&bond->mcast_work)->work) &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)bond131 (work_completion)(&(&slave->notify_work)->work) fill_pool_map-wait-type-override pool_lock irq_context: 0 (wq_completion)bond164 irq_context: 0 (wq_completion)bond164 (work_completion)(&(&slave->notify_work)->work) irq_context: 0 (wq_completion)bond164 (work_completion)(&(&slave->notify_work)->work) &obj_hash[i].lock irq_context: 0 (wq_completion)bond164 (work_completion)(&(&slave->notify_work)->work) &base->lock irq_context: 0 (wq_completion)bond164 (work_completion)(&(&slave->notify_work)->work) &base->lock &obj_hash[i].lock irq_context: 0 (wq_completion)bond164 (work_completion)(&(&slave->notify_work)->work) &rq->__lock irq_context: 0 (wq_completion)bond164 (work_completion)(&(&slave->notify_work)->work) &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)bond110#3 (work_completion)(&(&slave->notify_work)->work) rtnl_mutex net_rwsem irq_context: 0 (wq_completion)bond110#3 (work_completion)(&(&slave->notify_work)->work) rtnl_mutex net_rwsem &list->lock#2 irq_context: 0 (wq_completion)bond110#3 (work_completion)(&(&slave->notify_work)->work) rtnl_mutex &tn->lock irq_context: 0 (wq_completion)bond110#3 (work_completion)(&(&slave->notify_work)->work) rtnl_mutex.wait_lock irq_context: 0 (wq_completion)bond110#3 (work_completion)(&(&slave->notify_work)->work) &p->pi_lock irq_context: 0 (wq_completion)bond97#3 (work_completion)(&(&slave->notify_work)->work) rtnl_mutex net_rwsem irq_context: 0 (wq_completion)bond97#3 (work_completion)(&(&slave->notify_work)->work) rtnl_mutex net_rwsem &list->lock#2 irq_context: 0 (wq_completion)bond97#3 (work_completion)(&(&slave->notify_work)->work) rtnl_mutex &tn->lock irq_context: 0 (wq_completion)bond97#3 (work_completion)(&(&slave->notify_work)->work) rtnl_mutex.wait_lock irq_context: 0 (wq_completion)bond97#3 (work_completion)(&(&slave->notify_work)->work) &p->pi_lock irq_context: 0 (wq_completion)bond24#3 (work_completion)(&(&slave->notify_work)->work) rtnl_mutex net_rwsem irq_context: 0 (wq_completion)bond24#3 (work_completion)(&(&slave->notify_work)->work) rtnl_mutex net_rwsem &list->lock#2 irq_context: 0 (wq_completion)bond24#3 (work_completion)(&(&slave->notify_work)->work) rtnl_mutex &tn->lock irq_context: 0 (wq_completion)bond24#3 (work_completion)(&(&slave->notify_work)->work) rtnl_mutex.wait_lock irq_context: 0 (wq_completion)bond24#3 (work_completion)(&(&slave->notify_work)->work) &p->pi_lock irq_context: 0 (wq_completion)bond41 (work_completion)(&(&slave->notify_work)->work) rtnl_mutex net_rwsem irq_context: 0 (wq_completion)bond41 (work_completion)(&(&slave->notify_work)->work) rtnl_mutex net_rwsem &list->lock#2 irq_context: 0 (wq_completion)bond41 (work_completion)(&(&slave->notify_work)->work) rtnl_mutex &tn->lock irq_context: 0 (wq_completion)bond41 (work_completion)(&(&slave->notify_work)->work) rtnl_mutex.wait_lock irq_context: 0 (wq_completion)bond41 (work_completion)(&(&slave->notify_work)->work) &p->pi_lock irq_context: 0 (wq_completion)bond134#2 (work_completion)(&(&slave->notify_work)->work) rtnl_mutex net_rwsem irq_context: 0 (wq_completion)bond134#2 (work_completion)(&(&slave->notify_work)->work) rtnl_mutex net_rwsem &list->lock#2 irq_context: 0 (wq_completion)bond134#2 (work_completion)(&(&slave->notify_work)->work) rtnl_mutex &tn->lock irq_context: 0 (wq_completion)bond134#2 (work_completion)(&(&slave->notify_work)->work) rtnl_mutex.wait_lock irq_context: 0 (wq_completion)bond134#2 (work_completion)(&(&slave->notify_work)->work) &p->pi_lock irq_context: 0 (wq_completion)bond152 (work_completion)(&(&slave->notify_work)->work) rtnl_mutex net_rwsem irq_context: 0 (wq_completion)bond152 (work_completion)(&(&slave->notify_work)->work) rtnl_mutex net_rwsem &list->lock#2 irq_context: 0 (wq_completion)bond152 (work_completion)(&(&slave->notify_work)->work) rtnl_mutex &tn->lock irq_context: 0 (wq_completion)bond152 (work_completion)(&(&slave->notify_work)->work) rtnl_mutex.wait_lock irq_context: 0 (wq_completion)bond152 (work_completion)(&(&slave->notify_work)->work) &p->pi_lock irq_context: 0 (wq_completion)bond95 (work_completion)(&(&slave->notify_work)->work) rtnl_mutex net_rwsem irq_context: 0 (wq_completion)bond95 (work_completion)(&(&slave->notify_work)->work) rtnl_mutex net_rwsem &list->lock#2 irq_context: 0 (wq_completion)bond95 (work_completion)(&(&slave->notify_work)->work) rtnl_mutex &tn->lock irq_context: 0 (wq_completion)bond95 (work_completion)(&(&slave->notify_work)->work) rtnl_mutex.wait_lock irq_context: 0 (wq_completion)bond95 (work_completion)(&(&slave->notify_work)->work) &p->pi_lock irq_context: 0 (wq_completion)bond35 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex fs_reclaim irq_context: 0 (wq_completion)bond35 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex fs_reclaim &rq->__lock irq_context: 0 (wq_completion)bond135#2 (work_completion)(&(&bond->mcast_work)->work) fill_pool_map-wait-type-override pool_lock irq_context: 0 (wq_completion)bond104#3 &rq->__lock irq_context: 0 (wq_completion)bond104#3 &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)bond164 (work_completion)(&(&bond->mcast_work)->work) irq_context: 0 (wq_completion)bond164 (work_completion)(&(&bond->mcast_work)->work) &obj_hash[i].lock irq_context: 0 (wq_completion)bond164 (work_completion)(&(&bond->mcast_work)->work) &base->lock irq_context: 0 (wq_completion)bond164 (work_completion)(&(&bond->mcast_work)->work) &base->lock &obj_hash[i].lock irq_context: 0 (wq_completion)bond164 (work_completion)(&(&bond->mcast_work)->work) &rq->__lock irq_context: 0 (wq_completion)bond164 (work_completion)(&(&bond->mcast_work)->work) &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)bond35 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex fs_reclaim mmu_notifier_invalidate_range_start irq_context: 0 (wq_completion)bond35 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex batched_entropy_u8.lock irq_context: 0 (wq_completion)bond35 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex kfence_freelist_lock irq_context: 0 (wq_completion)bond35 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex rcu_read_lock &bond->stats_lock/1 irq_context: 0 (wq_completion)bond35 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex rcu_read_lock &ndev->lock irq_context: 0 (wq_completion)bond35 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex nl_table_lock irq_context: 0 (wq_completion)bond35 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &obj_hash[i].lock irq_context: 0 (wq_completion)bond35 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &rq->__lock irq_context: 0 (wq_completion)bond35 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)bond35 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &meta->lock irq_context: 0 (wq_completion)bond35 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex nl_table_wait.lock irq_context: 0 (wq_completion)bond35 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex net_rwsem irq_context: 0 (wq_completion)bond35 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex net_rwsem &list->lock#2 irq_context: 0 (wq_completion)bond35 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &tn->lock irq_context: 0 (wq_completion)bond35 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &idev->mc_lock irq_context: 0 (wq_completion)bond35 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &idev->mc_lock fs_reclaim irq_context: 0 (wq_completion)bond35 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &idev->mc_lock fs_reclaim mmu_notifier_invalidate_range_start irq_context: 0 (wq_completion)bond35 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &idev->mc_lock &c->lock irq_context: 0 (wq_completion)bond35 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &idev->mc_lock &n->list_lock irq_context: 0 (wq_completion)bond35 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &idev->mc_lock &n->list_lock &c->lock irq_context: 0 (wq_completion)bond35 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &idev->mc_lock &rq->__lock irq_context: 0 (wq_completion)bond35 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &idev->mc_lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)bond35 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &idev->mc_lock rcu_read_lock &ndev->lock irq_context: 0 (wq_completion)bond35 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &idev->mc_lock rcu_read_lock rcu_node_0 irq_context: 0 (wq_completion)bond35 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &idev->mc_lock rcu_read_lock &rq->__lock irq_context: 0 (wq_completion)bond35 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &idev->mc_lock rcu_read_lock &c->lock irq_context: 0 (wq_completion)bond35 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &idev->mc_lock rcu_read_lock &dir->lock#2 irq_context: 0 (wq_completion)bond35 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &idev->mc_lock rcu_read_lock &ul->lock#2 irq_context: 0 (wq_completion)bond35 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &idev->mc_lock rcu_read_lock rcu_read_lock &____s->seqcount#12 irq_context: 0 (wq_completion)bond35 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &idev->mc_lock rcu_read_lock rcu_read_lock &rq->__lock irq_context: 0 (wq_completion)bond35 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &idev->mc_lock rcu_read_lock rcu_read_lock &____s->seqcount#16 irq_context: 0 (wq_completion)bond35 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &idev->mc_lock rcu_read_lock &obj_hash[i].lock irq_context: 0 (wq_completion)bond35 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &idev->mc_lock rcu_read_lock &data->lock irq_context: 0 (wq_completion)bond35 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex rcu_read_lock rcu_node_0 irq_context: 0 (wq_completion)bond35 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex rcu_read_lock &rq->__lock irq_context: 0 (wq_completion)bond35 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex rcu_read_lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)bond35 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex.wait_lock irq_context: 0 (wq_completion)bond35 (work_completion)(&(&bond->mcast_work)->work) &p->pi_lock irq_context: 0 (wq_completion)bond35 (work_completion)(&(&bond->mcast_work)->work) &p->pi_lock &rq->__lock irq_context: 0 (wq_completion)bond35 (work_completion)(&(&bond->mcast_work)->work) &p->pi_lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)bond35 (work_completion)(&(&bond->mcast_work)->work) &rq->__lock irq_context: 0 (wq_completion)bond35 (work_completion)(&(&bond->mcast_work)->work) &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 rtnl_mutex batched_entropy_u32.lock crngs.lock base_crng.lock irq_context: 0 (wq_completion)bond159 (work_completion)(&(&slave->notify_work)->work) &rq->__lock irq_context: 0 (wq_completion)bond159 (work_completion)(&(&slave->notify_work)->work) &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)bond42 irq_context: 0 (wq_completion)bond42 &rq->__lock irq_context: 0 (wq_completion)bond42 &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)bond42 (work_completion)(&(&slave->notify_work)->work) irq_context: 0 (wq_completion)bond42 (work_completion)(&(&slave->notify_work)->work) &obj_hash[i].lock irq_context: 0 (wq_completion)bond42 (work_completion)(&(&slave->notify_work)->work) &base->lock irq_context: 0 (wq_completion)bond42 (work_completion)(&(&slave->notify_work)->work) &base->lock &obj_hash[i].lock irq_context: 0 (wq_completion)bond133#3 &rq->__lock irq_context: 0 (wq_completion)bond133#3 &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)bond42 (work_completion)(&(&bond->mcast_work)->work) irq_context: 0 (wq_completion)bond42 (work_completion)(&(&bond->mcast_work)->work) &obj_hash[i].lock irq_context: 0 (wq_completion)bond42 (work_completion)(&(&bond->mcast_work)->work) &base->lock irq_context: 0 (wq_completion)bond42 (work_completion)(&(&bond->mcast_work)->work) &base->lock &obj_hash[i].lock irq_context: 0 (wq_completion)bond42 (work_completion)(&(&bond->mcast_work)->work) &rq->__lock irq_context: 0 (wq_completion)bond42 (work_completion)(&(&bond->mcast_work)->work) &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)bond152#2 irq_context: 0 (wq_completion)bond152#2 &rq->__lock irq_context: 0 (wq_completion)bond152#2 (work_completion)(&(&slave->notify_work)->work) irq_context: 0 (wq_completion)bond152#2 (work_completion)(&(&slave->notify_work)->work) &obj_hash[i].lock irq_context: 0 (wq_completion)bond152#2 (work_completion)(&(&slave->notify_work)->work) &base->lock irq_context: 0 (wq_completion)bond152#2 (work_completion)(&(&slave->notify_work)->work) &base->lock &obj_hash[i].lock irq_context: 0 (wq_completion)bond142#2 (work_completion)(&(&slave->notify_work)->work) &rq->__lock irq_context: 0 (wq_completion)bond142#2 (work_completion)(&(&slave->notify_work)->work) &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)bond152#2 (work_completion)(&(&bond->mcast_work)->work) irq_context: 0 (wq_completion)bond152#2 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex fs_reclaim irq_context: 0 (wq_completion)bond152#2 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex fs_reclaim mmu_notifier_invalidate_range_start irq_context: 0 (wq_completion)bond152#2 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &c->lock irq_context: 0 (wq_completion)bond152#2 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex rcu_read_lock &bond->stats_lock/1 irq_context: 0 (wq_completion)bond152#2 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex rcu_read_lock &ndev->lock irq_context: 0 (wq_completion)bond152#2 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex rcu_read_lock rcu_node_0 irq_context: 0 (wq_completion)bond152#2 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex rcu_read_lock &rq->__lock irq_context: 0 (wq_completion)bond152#2 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex rcu_read_lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)bond152#2 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex rcu_read_lock &cfs_rq->removed.lock irq_context: 0 (wq_completion)bond152#2 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex rcu_read_lock &obj_hash[i].lock irq_context: 0 (wq_completion)bond152#2 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex rcu_read_lock pool_lock#2 irq_context: 0 (wq_completion)bond152#2 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex nl_table_lock irq_context: 0 (wq_completion)bond152#2 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &obj_hash[i].lock irq_context: 0 (wq_completion)bond152#2 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex quarantine_lock irq_context: 0 (wq_completion)bond152#2 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &rq->__lock irq_context: 0 (wq_completion)bond152#2 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)bond152#2 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex nl_table_wait.lock irq_context: 0 (wq_completion)bond152#2 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex net_rwsem irq_context: 0 (wq_completion)bond152#2 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex net_rwsem &list->lock#2 irq_context: 0 (wq_completion)bond152#2 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &tn->lock irq_context: 0 (wq_completion)bond152#2 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &idev->mc_lock irq_context: 0 (wq_completion)bond152#2 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex.wait_lock irq_context: 0 (wq_completion)bond152#2 (work_completion)(&(&bond->mcast_work)->work) &p->pi_lock irq_context: 0 (wq_completion)bond152#2 (work_completion)(&(&slave->notify_work)->work) rtnl_mutex net_rwsem irq_context: 0 (wq_completion)bond152#2 (work_completion)(&(&slave->notify_work)->work) rtnl_mutex net_rwsem &list->lock#2 irq_context: 0 (wq_completion)bond152#2 (work_completion)(&(&slave->notify_work)->work) rtnl_mutex &tn->lock irq_context: 0 (wq_completion)bond152#2 (work_completion)(&(&slave->notify_work)->work) rtnl_mutex.wait_lock irq_context: 0 (wq_completion)bond152#2 (work_completion)(&(&slave->notify_work)->work) &p->pi_lock irq_context: 0 (wq_completion)bond134#3 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex fs_reclaim irq_context: 0 (wq_completion)bond134#3 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex fs_reclaim mmu_notifier_invalidate_range_start irq_context: 0 (wq_completion)bond134#3 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &c->lock irq_context: 0 (wq_completion)bond134#3 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex rcu_read_lock &bond->stats_lock/1 irq_context: 0 (wq_completion)bond134#3 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex rcu_read_lock &ndev->lock irq_context: 0 (wq_completion)bond134#3 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex nl_table_lock irq_context: 0 (wq_completion)bond134#3 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &obj_hash[i].lock irq_context: 0 (wq_completion)bond134#3 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex nl_table_wait.lock irq_context: 0 (wq_completion)bond134#3 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex net_rwsem irq_context: 0 (wq_completion)bond134#3 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex net_rwsem &list->lock#2 irq_context: 0 (wq_completion)bond134#3 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &tn->lock irq_context: 0 (wq_completion)bond134#3 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &idev->mc_lock irq_context: 0 (wq_completion)bond134#3 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &idev->mc_lock fs_reclaim irq_context: 0 (wq_completion)bond134#3 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &idev->mc_lock fs_reclaim mmu_notifier_invalidate_range_start irq_context: 0 (wq_completion)bond134#3 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &idev->mc_lock &c->lock irq_context: 0 (wq_completion)bond134#3 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &idev->mc_lock rcu_read_lock &ndev->lock irq_context: 0 (wq_completion)bond134#3 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &idev->mc_lock rcu_read_lock &dir->lock#2 irq_context: 0 (wq_completion)bond134#3 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &idev->mc_lock rcu_read_lock &ul->lock#2 irq_context: 0 (wq_completion)bond134#3 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &idev->mc_lock rcu_read_lock rcu_read_lock &____s->seqcount#10 irq_context: 0 (wq_completion)bond134#3 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &idev->mc_lock rcu_read_lock rcu_read_lock rcu_read_lock_bh rcu_read_lock rcu_read_lock_bh rcu_read_lock &br->multicast_lock irq_context: 0 (wq_completion)bond134#3 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &idev->mc_lock rcu_read_lock rcu_read_lock rcu_read_lock_bh rcu_read_lock rcu_read_lock_bh rcu_read_lock &br->multicast_lock &obj_hash[i].lock irq_context: 0 (wq_completion)bond134#3 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &idev->mc_lock rcu_read_lock rcu_read_lock rcu_read_lock_bh rcu_read_lock rcu_read_lock_bh rcu_read_lock &br->multicast_lock &base->lock irq_context: 0 (wq_completion)bond134#3 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &idev->mc_lock rcu_read_lock rcu_read_lock rcu_read_lock_bh rcu_read_lock rcu_read_lock_bh rcu_read_lock &data->lock irq_context: 0 (wq_completion)bond134#3 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &idev->mc_lock rcu_read_lock rcu_read_lock rcu_read_lock_bh rcu_read_lock rcu_read_lock_bh rcu_read_lock &obj_hash[i].lock irq_context: 0 (wq_completion)bond134#3 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex.wait_lock irq_context: 0 (wq_completion)bond134#3 (work_completion)(&(&bond->mcast_work)->work) &p->pi_lock irq_context: 0 (wq_completion)bond136#3 irq_context: 0 (wq_completion)bond136#3 (work_completion)(&(&slave->notify_work)->work) irq_context: 0 (wq_completion)bond136#3 (work_completion)(&(&slave->notify_work)->work) &obj_hash[i].lock irq_context: 0 (wq_completion)bond136#3 (work_completion)(&(&slave->notify_work)->work) &base->lock irq_context: 0 (wq_completion)bond136#3 (work_completion)(&(&slave->notify_work)->work) &base->lock &obj_hash[i].lock irq_context: 0 (wq_completion)bond136#3 (work_completion)(&(&bond->mcast_work)->work) irq_context: 0 (wq_completion)bond136#3 (work_completion)(&(&bond->mcast_work)->work) &obj_hash[i].lock irq_context: 0 (wq_completion)bond136#3 (work_completion)(&(&bond->mcast_work)->work) &base->lock irq_context: 0 (wq_completion)bond136#3 (work_completion)(&(&bond->mcast_work)->work) &base->lock &obj_hash[i].lock irq_context: 0 (wq_completion)bond136#3 (work_completion)(&(&bond->mcast_work)->work) &rq->__lock irq_context: 0 (wq_completion)bond136#3 (work_completion)(&(&bond->mcast_work)->work) &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)bond161 (work_completion)(&(&slave->notify_work)->work) rtnl_mutex net_rwsem irq_context: 0 (wq_completion)bond161 (work_completion)(&(&slave->notify_work)->work) rtnl_mutex net_rwsem &list->lock#2 irq_context: 0 (wq_completion)bond161 (work_completion)(&(&slave->notify_work)->work) rtnl_mutex &tn->lock irq_context: 0 (wq_completion)bond161 (work_completion)(&(&slave->notify_work)->work) rtnl_mutex.wait_lock irq_context: 0 (wq_completion)bond161 (work_completion)(&(&slave->notify_work)->work) &p->pi_lock irq_context: 0 (wq_completion)bond151 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex fs_reclaim irq_context: 0 (wq_completion)bond151 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex fs_reclaim mmu_notifier_invalidate_range_start irq_context: 0 (wq_completion)bond151 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &c->lock irq_context: 0 (wq_completion)bond151 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &n->list_lock irq_context: 0 (wq_completion)bond151 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &n->list_lock &c->lock irq_context: 0 (wq_completion)bond151 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex rcu_read_lock &bond->stats_lock/1 irq_context: 0 (wq_completion)bond151 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex rcu_read_lock &ndev->lock irq_context: 0 (wq_completion)bond151 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex nl_table_lock irq_context: 0 (wq_completion)bond151 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &obj_hash[i].lock irq_context: 0 (wq_completion)bond151 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex nl_table_wait.lock irq_context: 0 (wq_completion)bond151 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &rq->__lock irq_context: 0 (wq_completion)bond151 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)bond151 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex net_rwsem irq_context: 0 (wq_completion)bond151 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex net_rwsem &list->lock#2 irq_context: 0 (wq_completion)bond151 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &tn->lock irq_context: 0 (wq_completion)bond151 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &idev->mc_lock irq_context: 0 (wq_completion)bond151 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &idev->mc_lock fs_reclaim irq_context: 0 (wq_completion)bond151 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &idev->mc_lock fs_reclaim mmu_notifier_invalidate_range_start irq_context: 0 (wq_completion)bond151 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &idev->mc_lock rcu_read_lock &ndev->lock irq_context: 0 (wq_completion)bond151 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &idev->mc_lock rcu_read_lock &dir->lock#2 irq_context: 0 (wq_completion)bond151 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &idev->mc_lock rcu_read_lock &ul->lock#2 irq_context: 0 (wq_completion)bond151 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &idev->mc_lock rcu_read_lock rcu_read_lock &____s->seqcount#10 irq_context: 0 (wq_completion)bond151 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &idev->mc_lock rcu_read_lock rcu_read_lock rcu_read_lock_bh rcu_read_lock rcu_read_lock_bh rcu_read_lock &br->multicast_lock irq_context: 0 (wq_completion)bond151 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &idev->mc_lock rcu_read_lock rcu_read_lock rcu_read_lock_bh rcu_read_lock rcu_read_lock_bh rcu_read_lock &br->multicast_lock &obj_hash[i].lock irq_context: 0 (wq_completion)bond151 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &idev->mc_lock rcu_read_lock rcu_read_lock rcu_read_lock_bh rcu_read_lock rcu_read_lock_bh rcu_read_lock &br->multicast_lock &base->lock irq_context: 0 (wq_completion)bond151 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &idev->mc_lock rcu_read_lock rcu_read_lock rcu_read_lock_bh rcu_read_lock rcu_read_lock_bh rcu_read_lock &data->lock irq_context: 0 (wq_completion)bond151 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &idev->mc_lock rcu_read_lock rcu_read_lock rcu_read_lock_bh rcu_read_lock rcu_read_lock_bh rcu_read_lock &obj_hash[i].lock irq_context: 0 (wq_completion)bond151 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex.wait_lock irq_context: 0 (wq_completion)bond151 (work_completion)(&(&bond->mcast_work)->work) &p->pi_lock irq_context: 0 (wq_completion)bond151 (work_completion)(&(&bond->mcast_work)->work) &p->pi_lock &rq->__lock irq_context: 0 (wq_completion)bond151 (work_completion)(&(&bond->mcast_work)->work) &p->pi_lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)bond151 (work_completion)(&(&bond->mcast_work)->work) &rq->__lock irq_context: 0 (wq_completion)bond151 (work_completion)(&(&bond->mcast_work)->work) &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)bond134#3 (work_completion)(&(&slave->notify_work)->work) &rq->__lock irq_context: 0 (wq_completion)bond165 irq_context: 0 (wq_completion)bond165 (work_completion)(&(&slave->notify_work)->work) irq_context: 0 (wq_completion)bond165 (work_completion)(&(&slave->notify_work)->work) &obj_hash[i].lock irq_context: 0 (wq_completion)bond165 (work_completion)(&(&slave->notify_work)->work) &base->lock irq_context: 0 (wq_completion)bond165 (work_completion)(&(&slave->notify_work)->work) &base->lock &obj_hash[i].lock irq_context: 0 (wq_completion)bond165 (work_completion)(&(&slave->notify_work)->work) &rq->__lock irq_context: 0 (wq_completion)bond165 (work_completion)(&(&slave->notify_work)->work) &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)bond165 (work_completion)(&(&bond->mcast_work)->work) irq_context: 0 (wq_completion)bond165 (work_completion)(&(&bond->mcast_work)->work) &obj_hash[i].lock irq_context: 0 (wq_completion)bond165 (work_completion)(&(&bond->mcast_work)->work) &base->lock irq_context: 0 (wq_completion)bond165 (work_completion)(&(&bond->mcast_work)->work) &base->lock &obj_hash[i].lock irq_context: 0 (wq_completion)bond165 (work_completion)(&(&bond->mcast_work)->work) &rq->__lock irq_context: 0 (wq_completion)bond105#2 (work_completion)(&(&bond->mcast_work)->work) fill_pool_map-wait-type-override pool_lock irq_context: 0 (wq_completion)bond149#2 (work_completion)(&(&slave->notify_work)->work) &rq->__lock irq_context: 0 (wq_completion)bond149#2 (work_completion)(&(&slave->notify_work)->work) &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)bond153#2 irq_context: 0 (wq_completion)bond153#2 (work_completion)(&(&slave->notify_work)->work) irq_context: 0 (wq_completion)bond153#2 (work_completion)(&(&slave->notify_work)->work) &obj_hash[i].lock irq_context: 0 (wq_completion)bond153#2 (work_completion)(&(&slave->notify_work)->work) &base->lock irq_context: 0 (wq_completion)bond153#2 (work_completion)(&(&slave->notify_work)->work) &base->lock &obj_hash[i].lock irq_context: 0 (wq_completion)bond122#3 (work_completion)(&(&slave->notify_work)->work) &rq->__lock irq_context: 0 (wq_completion)bond122#3 (work_completion)(&(&slave->notify_work)->work) &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)bond153#2 (work_completion)(&(&bond->mcast_work)->work) irq_context: 0 (wq_completion)bond153#2 (work_completion)(&(&bond->mcast_work)->work) &obj_hash[i].lock irq_context: 0 (wq_completion)bond153#2 (work_completion)(&(&bond->mcast_work)->work) &base->lock irq_context: 0 (wq_completion)bond153#2 (work_completion)(&(&bond->mcast_work)->work) &base->lock &obj_hash[i].lock irq_context: 0 (wq_completion)bond153#2 (work_completion)(&(&bond->mcast_work)->work) &rq->__lock irq_context: 0 (wq_completion)bond153#2 (work_completion)(&(&bond->mcast_work)->work) &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)bond137#3 irq_context: 0 (wq_completion)bond137#3 (work_completion)(&(&slave->notify_work)->work) irq_context: 0 (wq_completion)bond137#3 (work_completion)(&(&slave->notify_work)->work) &obj_hash[i].lock irq_context: 0 (wq_completion)bond137#3 (work_completion)(&(&slave->notify_work)->work) &base->lock irq_context: 0 (wq_completion)bond137#3 (work_completion)(&(&slave->notify_work)->work) &base->lock &obj_hash[i].lock irq_context: 0 (wq_completion)bond137#3 (work_completion)(&(&slave->notify_work)->work) &rq->__lock irq_context: 0 (wq_completion)bond137#3 (work_completion)(&(&bond->mcast_work)->work) irq_context: 0 (wq_completion)bond137#3 (work_completion)(&(&bond->mcast_work)->work) &obj_hash[i].lock irq_context: 0 (wq_completion)bond137#3 (work_completion)(&(&bond->mcast_work)->work) &base->lock irq_context: 0 (wq_completion)bond137#3 (work_completion)(&(&bond->mcast_work)->work) &base->lock &obj_hash[i].lock irq_context: 0 (wq_completion)bond137#3 (work_completion)(&(&bond->mcast_work)->work) &rq->__lock irq_context: 0 (wq_completion)bond137#3 (work_completion)(&(&bond->mcast_work)->work) &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)bond162 (work_completion)(&(&slave->notify_work)->work) rtnl_mutex net_rwsem irq_context: 0 (wq_completion)bond162 (work_completion)(&(&slave->notify_work)->work) rtnl_mutex net_rwsem &list->lock#2 irq_context: 0 (wq_completion)bond162 (work_completion)(&(&slave->notify_work)->work) rtnl_mutex &tn->lock irq_context: 0 (wq_completion)bond162 (work_completion)(&(&slave->notify_work)->work) rtnl_mutex rcu_read_lock rcu_node_0 irq_context: 0 (wq_completion)bond162 (work_completion)(&(&slave->notify_work)->work) rtnl_mutex rcu_read_lock &rq->__lock irq_context: 0 (wq_completion)bond162 (work_completion)(&(&slave->notify_work)->work) rtnl_mutex rcu_read_lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)bond162 (work_completion)(&(&slave->notify_work)->work) rtnl_mutex.wait_lock irq_context: 0 (wq_completion)bond162 (work_completion)(&(&slave->notify_work)->work) &p->pi_lock irq_context: 0 (wq_completion)bond162 (work_completion)(&(&slave->notify_work)->work) &p->pi_lock &rq->__lock irq_context: 0 (wq_completion)bond162 (work_completion)(&(&slave->notify_work)->work) &p->pi_lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)bond162 (work_completion)(&(&slave->notify_work)->work) &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)bond135#3 (work_completion)(&(&bond->mcast_work)->work) &rq->__lock irq_context: 0 (wq_completion)bond135#3 (work_completion)(&(&bond->mcast_work)->work) &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)bond43#2 irq_context: 0 (wq_completion)bond43#2 &rq->__lock irq_context: 0 (wq_completion)bond43#2 (work_completion)(&(&slave->notify_work)->work) irq_context: 0 (wq_completion)bond43#2 (work_completion)(&(&slave->notify_work)->work) &obj_hash[i].lock irq_context: 0 (wq_completion)bond43#2 (work_completion)(&(&slave->notify_work)->work) &base->lock irq_context: 0 (wq_completion)bond43#2 (work_completion)(&(&slave->notify_work)->work) &base->lock &obj_hash[i].lock irq_context: 0 (wq_completion)bond43#2 (work_completion)(&(&slave->notify_work)->work) &rq->__lock irq_context: 0 (wq_completion)bond43#2 (work_completion)(&(&slave->notify_work)->work) &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)bond43#2 (work_completion)(&(&bond->mcast_work)->work) irq_context: 0 (wq_completion)bond43#2 (work_completion)(&(&bond->mcast_work)->work) &obj_hash[i].lock irq_context: 0 (wq_completion)bond43#2 (work_completion)(&(&bond->mcast_work)->work) &base->lock irq_context: 0 (wq_completion)bond43#2 (work_completion)(&(&bond->mcast_work)->work) &base->lock &obj_hash[i].lock irq_context: 0 (wq_completion)bond43#2 (work_completion)(&(&bond->mcast_work)->work) &rq->__lock irq_context: 0 (wq_completion)bond43#2 (work_completion)(&(&bond->mcast_work)->work) &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)bond166 irq_context: 0 (wq_completion)bond166 (work_completion)(&(&slave->notify_work)->work) irq_context: 0 (wq_completion)bond166 (work_completion)(&(&slave->notify_work)->work) &obj_hash[i].lock irq_context: 0 (wq_completion)bond166 (work_completion)(&(&slave->notify_work)->work) &base->lock irq_context: 0 (wq_completion)bond166 (work_completion)(&(&slave->notify_work)->work) &base->lock &obj_hash[i].lock irq_context: 0 (wq_completion)mld (work_completion)(&(&idev->mc_ifc_work)->work) &idev->mc_lock rcu_read_lock rcu_read_lock &tbl->lock quarantine_lock irq_context: 0 (wq_completion)bond131#2 (work_completion)(&(&slave->notify_work)->work) rtnl_mutex net_rwsem irq_context: 0 (wq_completion)bond131#2 (work_completion)(&(&slave->notify_work)->work) rtnl_mutex net_rwsem &list->lock#2 irq_context: 0 (wq_completion)bond131#2 (work_completion)(&(&slave->notify_work)->work) rtnl_mutex &tn->lock irq_context: 0 (wq_completion)bond131#2 (work_completion)(&(&slave->notify_work)->work) rtnl_mutex.wait_lock irq_context: 0 (wq_completion)bond131#2 (work_completion)(&(&slave->notify_work)->work) &p->pi_lock irq_context: 0 (wq_completion)bond137#2 (work_completion)(&(&bond->mcast_work)->work) &rq->__lock irq_context: 0 (wq_completion)bond137#2 (work_completion)(&(&bond->mcast_work)->work) &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)bond166 (work_completion)(&(&bond->mcast_work)->work) irq_context: 0 (wq_completion)bond166 (work_completion)(&(&bond->mcast_work)->work) &obj_hash[i].lock irq_context: 0 (wq_completion)bond166 (work_completion)(&(&bond->mcast_work)->work) &base->lock irq_context: 0 (wq_completion)bond166 (work_completion)(&(&bond->mcast_work)->work) &base->lock &obj_hash[i].lock irq_context: 0 (wq_completion)bond30#4 (work_completion)(&(&bond->mcast_work)->work) &rq->__lock irq_context: 0 (wq_completion)bond30#4 (work_completion)(&(&bond->mcast_work)->work) &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)bond123#3 (work_completion)(&(&slave->notify_work)->work) rtnl_mutex net_rwsem irq_context: 0 (wq_completion)bond123#3 (work_completion)(&(&slave->notify_work)->work) rtnl_mutex net_rwsem &list->lock#2 irq_context: 0 (wq_completion)bond123#3 (work_completion)(&(&slave->notify_work)->work) rtnl_mutex &tn->lock irq_context: 0 (wq_completion)bond123#3 (work_completion)(&(&slave->notify_work)->work) rtnl_mutex.wait_lock irq_context: 0 (wq_completion)bond123#3 (work_completion)(&(&slave->notify_work)->work) &p->pi_lock irq_context: 0 (wq_completion)bond43#2 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex fs_reclaim irq_context: 0 (wq_completion)bond43#2 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex fs_reclaim mmu_notifier_invalidate_range_start irq_context: 0 (wq_completion)bond43#2 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &c->lock irq_context: 0 (wq_completion)bond43#2 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex rcu_read_lock &bond->stats_lock/1 irq_context: 0 (wq_completion)bond43#2 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex rcu_read_lock &ndev->lock irq_context: 0 (wq_completion)bond43#2 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex nl_table_lock irq_context: 0 (wq_completion)bond43#2 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &obj_hash[i].lock irq_context: 0 (wq_completion)bond43#2 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex nl_table_wait.lock irq_context: 0 (wq_completion)bond43#2 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex net_rwsem irq_context: 0 (wq_completion)bond43#2 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex net_rwsem &list->lock#2 irq_context: 0 (wq_completion)bond43#2 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &tn->lock irq_context: 0 (wq_completion)bond43#2 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &idev->mc_lock irq_context: 0 (wq_completion)bond43#2 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &idev->mc_lock fs_reclaim irq_context: 0 (wq_completion)bond43#2 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &idev->mc_lock fs_reclaim mmu_notifier_invalidate_range_start irq_context: 0 (wq_completion)bond43#2 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &idev->mc_lock &c->lock irq_context: 0 (wq_completion)bond43#2 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &idev->mc_lock rcu_read_lock &ndev->lock irq_context: 0 (wq_completion)bond43#2 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &idev->mc_lock rcu_read_lock &dir->lock#2 irq_context: 0 (wq_completion)bond43#2 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &idev->mc_lock rcu_read_lock &ul->lock#2 irq_context: 0 (wq_completion)bond43#2 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &idev->mc_lock rcu_read_lock rcu_read_lock &____s->seqcount#12 irq_context: 0 (wq_completion)bond43#2 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &idev->mc_lock rcu_read_lock rcu_read_lock &____s->seqcount#16 irq_context: 0 (wq_completion)bond43#2 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &idev->mc_lock rcu_read_lock &obj_hash[i].lock irq_context: 0 (wq_completion)bond43#2 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &idev->mc_lock rcu_read_lock &data->lock irq_context: 0 (wq_completion)bond43#2 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex.wait_lock irq_context: 0 (wq_completion)bond43#2 (work_completion)(&(&bond->mcast_work)->work) &p->pi_lock irq_context: 0 (wq_completion)bond119#3 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex fs_reclaim irq_context: 0 (wq_completion)bond119#3 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex fs_reclaim mmu_notifier_invalidate_range_start irq_context: 0 (wq_completion)bond119#3 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex rcu_read_lock &bond->stats_lock/1 irq_context: 0 (wq_completion)bond119#3 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex rcu_read_lock &ndev->lock irq_context: 0 (wq_completion)bond119#3 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex nl_table_lock irq_context: 0 (wq_completion)bond119#3 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &obj_hash[i].lock irq_context: 0 (wq_completion)bond119#3 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex nl_table_wait.lock irq_context: 0 (wq_completion)bond119#3 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex net_rwsem irq_context: 0 (wq_completion)bond119#3 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex net_rwsem &list->lock#2 irq_context: 0 (wq_completion)bond119#3 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &tn->lock irq_context: 0 (wq_completion)bond119#3 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &idev->mc_lock irq_context: 0 (wq_completion)bond119#3 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &idev->mc_lock fs_reclaim irq_context: 0 (wq_completion)bond119#3 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &idev->mc_lock fs_reclaim mmu_notifier_invalidate_range_start irq_context: 0 (wq_completion)bond119#3 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &idev->mc_lock &c->lock irq_context: 0 (wq_completion)bond119#3 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &idev->mc_lock &n->list_lock irq_context: 0 (wq_completion)bond119#3 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &idev->mc_lock &n->list_lock &c->lock irq_context: 0 (wq_completion)bond119#3 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &idev->mc_lock rcu_read_lock &ndev->lock irq_context: 0 (wq_completion)bond119#3 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &idev->mc_lock rcu_read_lock &dir->lock#2 irq_context: 0 (wq_completion)bond119#3 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &idev->mc_lock rcu_read_lock &ul->lock#2 irq_context: 0 (wq_completion)bond119#3 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &idev->mc_lock rcu_read_lock rcu_read_lock &____s->seqcount#10 irq_context: 0 (wq_completion)bond119#3 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &idev->mc_lock rcu_read_lock rcu_read_lock rcu_read_lock_bh rcu_read_lock rcu_read_lock_bh rcu_read_lock &br->multicast_lock irq_context: 0 (wq_completion)bond119#3 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &idev->mc_lock rcu_read_lock rcu_read_lock rcu_read_lock_bh rcu_read_lock rcu_read_lock_bh rcu_read_lock &br->multicast_lock &obj_hash[i].lock irq_context: 0 (wq_completion)bond119#3 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &idev->mc_lock rcu_read_lock rcu_read_lock rcu_read_lock_bh rcu_read_lock rcu_read_lock_bh rcu_read_lock &br->multicast_lock &base->lock irq_context: 0 (wq_completion)bond119#3 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &idev->mc_lock rcu_read_lock rcu_read_lock rcu_read_lock_bh rcu_read_lock rcu_read_lock_bh rcu_read_lock &br->multicast_lock &base->lock &obj_hash[i].lock irq_context: 0 (wq_completion)bond119#3 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &idev->mc_lock rcu_read_lock rcu_read_lock rcu_read_lock_bh rcu_read_lock rcu_read_lock_bh rcu_read_lock &data->lock irq_context: 0 (wq_completion)bond119#3 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &idev->mc_lock rcu_read_lock rcu_read_lock rcu_read_lock_bh rcu_read_lock rcu_read_lock_bh rcu_read_lock &obj_hash[i].lock irq_context: 0 (wq_completion)bond119#3 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &idev->mc_lock rcu_read_lock rcu_read_lock rcu_node_0 irq_context: 0 (wq_completion)bond119#3 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &idev->mc_lock rcu_read_lock rcu_read_lock &rq->__lock irq_context: 0 (wq_completion)bond119#3 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &idev->mc_lock rcu_read_lock rcu_read_lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)bond119#3 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &idev->mc_lock rcu_read_lock rcu_node_0 irq_context: 0 (wq_completion)bond119#3 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex.wait_lock irq_context: 0 (wq_completion)bond119#3 (work_completion)(&(&bond->mcast_work)->work) &p->pi_lock irq_context: 0 (wq_completion)bond119#3 (work_completion)(&(&slave->notify_work)->work) rtnl_mutex net_rwsem irq_context: 0 (wq_completion)bond119#3 (work_completion)(&(&slave->notify_work)->work) rtnl_mutex net_rwsem &list->lock#2 irq_context: 0 (wq_completion)bond119#3 (work_completion)(&(&slave->notify_work)->work) rtnl_mutex &tn->lock irq_context: 0 (wq_completion)bond119#3 (work_completion)(&(&slave->notify_work)->work) rtnl_mutex rcu_read_lock rcu_node_0 irq_context: 0 (wq_completion)bond119#3 (work_completion)(&(&slave->notify_work)->work) rtnl_mutex rcu_read_lock &rq->__lock irq_context: 0 (wq_completion)bond119#3 (work_completion)(&(&slave->notify_work)->work) rtnl_mutex rcu_read_lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)bond119#3 (work_completion)(&(&slave->notify_work)->work) rtnl_mutex.wait_lock irq_context: 0 (wq_completion)bond119#3 (work_completion)(&(&slave->notify_work)->work) &p->pi_lock irq_context: 0 (wq_completion)bond154#2 irq_context: 0 (wq_completion)bond154#2 (work_completion)(&(&slave->notify_work)->work) irq_context: 0 (wq_completion)bond154#2 (work_completion)(&(&slave->notify_work)->work) &obj_hash[i].lock irq_context: 0 (wq_completion)bond154#2 (work_completion)(&(&slave->notify_work)->work) &base->lock irq_context: 0 (wq_completion)bond154#2 (work_completion)(&(&slave->notify_work)->work) &base->lock &obj_hash[i].lock irq_context: 0 (wq_completion)bond154#2 (work_completion)(&(&slave->notify_work)->work) &rq->__lock irq_context: 0 (wq_completion)bond154#2 (work_completion)(&(&slave->notify_work)->work) &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)bond165 (work_completion)(&(&bond->mcast_work)->work) &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)bond154#2 (work_completion)(&(&bond->mcast_work)->work) irq_context: 0 (wq_completion)bond154#2 (work_completion)(&(&bond->mcast_work)->work) &obj_hash[i].lock irq_context: 0 (wq_completion)bond154#2 (work_completion)(&(&bond->mcast_work)->work) &base->lock irq_context: 0 (wq_completion)bond154#2 (work_completion)(&(&bond->mcast_work)->work) &base->lock &obj_hash[i].lock irq_context: 0 (wq_completion)bond154#2 (work_completion)(&(&bond->mcast_work)->work) &rq->__lock irq_context: 0 (wq_completion)bond154#2 (work_completion)(&(&bond->mcast_work)->work) &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)bond137#2 (work_completion)(&(&slave->notify_work)->work) rtnl_mutex net_rwsem irq_context: 0 (wq_completion)bond137#2 (work_completion)(&(&slave->notify_work)->work) rtnl_mutex net_rwsem &list->lock#2 irq_context: 0 (wq_completion)bond137#2 (work_completion)(&(&slave->notify_work)->work) rtnl_mutex &tn->lock irq_context: 0 (wq_completion)bond137#2 (work_completion)(&(&slave->notify_work)->work) rtnl_mutex.wait_lock irq_context: 0 (wq_completion)bond137#2 (work_completion)(&(&slave->notify_work)->work) &p->pi_lock irq_context: 0 (wq_completion)bond119#2 (work_completion)(&(&slave->notify_work)->work) rtnl_mutex net_rwsem irq_context: 0 (wq_completion)bond119#2 (work_completion)(&(&slave->notify_work)->work) rtnl_mutex net_rwsem &list->lock#2 irq_context: 0 (wq_completion)bond119#2 (work_completion)(&(&slave->notify_work)->work) rtnl_mutex &tn->lock irq_context: 0 (wq_completion)bond119#2 (work_completion)(&(&slave->notify_work)->work) rtnl_mutex.wait_lock irq_context: 0 (wq_completion)bond99#3 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex fs_reclaim irq_context: 0 (wq_completion)bond119#2 (work_completion)(&(&slave->notify_work)->work) &p->pi_lock irq_context: 0 (wq_completion)bond99#3 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex fs_reclaim &rq->__lock irq_context: 0 (wq_completion)bond99#3 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex fs_reclaim &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)bond99#3 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex fs_reclaim mmu_notifier_invalidate_range_start irq_context: 0 (wq_completion)bond99#3 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &c->lock irq_context: 0 (wq_completion)bond99#3 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &n->list_lock irq_context: 0 (wq_completion)bond99#3 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &n->list_lock &c->lock irq_context: 0 (wq_completion)bond99#3 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex rcu_read_lock &bond->stats_lock/1 irq_context: 0 (wq_completion)bond99#3 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex rcu_read_lock &ndev->lock irq_context: 0 (wq_completion)bond99#3 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex rcu_read_lock rcu_node_0 irq_context: 0 (wq_completion)bond99#3 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex rcu_read_lock &rq->__lock irq_context: 0 (wq_completion)bond99#3 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex rcu_read_lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)bond99#3 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex nl_table_lock irq_context: 0 (wq_completion)bond99#3 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &obj_hash[i].lock irq_context: 0 (wq_completion)bond99#3 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex nl_table_wait.lock irq_context: 0 (wq_completion)bond99#3 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex net_rwsem irq_context: 0 (wq_completion)bond99#3 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex net_rwsem &list->lock#2 irq_context: 0 (wq_completion)bond99#3 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &tn->lock irq_context: 0 (wq_completion)bond99#3 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &idev->mc_lock irq_context: 0 (wq_completion)bond99#3 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &idev->mc_lock fs_reclaim irq_context: 0 (wq_completion)bond99#3 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &idev->mc_lock fs_reclaim mmu_notifier_invalidate_range_start irq_context: 0 (wq_completion)bond99#3 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &idev->mc_lock rcu_read_lock &ndev->lock irq_context: 0 (wq_completion)bond99#3 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &idev->mc_lock rcu_read_lock rcu_node_0 irq_context: 0 (wq_completion)bond99#3 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &idev->mc_lock rcu_read_lock &rq->__lock irq_context: 0 (wq_completion)bond99#3 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &idev->mc_lock rcu_read_lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)bond99#3 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &idev->mc_lock rcu_read_lock &dir->lock#2 irq_context: 0 (wq_completion)bond99#3 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &idev->mc_lock rcu_read_lock &ul->lock#2 irq_context: 0 (wq_completion)bond99#3 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &idev->mc_lock rcu_read_lock rcu_read_lock &____s->seqcount#10 irq_context: 0 (wq_completion)bond99#3 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &idev->mc_lock rcu_read_lock rcu_read_lock rcu_read_lock_bh rcu_read_lock rcu_read_lock_bh rcu_read_lock &br->multicast_lock irq_context: 0 (wq_completion)bond99#3 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &idev->mc_lock rcu_read_lock rcu_read_lock rcu_read_lock_bh rcu_read_lock rcu_read_lock_bh rcu_read_lock &br->multicast_lock &obj_hash[i].lock irq_context: 0 (wq_completion)bond99#3 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &idev->mc_lock rcu_read_lock rcu_read_lock rcu_read_lock_bh rcu_read_lock rcu_read_lock_bh rcu_read_lock &br->multicast_lock &base->lock irq_context: 0 (wq_completion)bond99#3 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &idev->mc_lock rcu_read_lock rcu_read_lock rcu_read_lock_bh rcu_read_lock rcu_read_lock_bh rcu_read_lock &br->multicast_lock &base->lock &obj_hash[i].lock irq_context: 0 (wq_completion)bond99#3 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &idev->mc_lock rcu_read_lock rcu_read_lock rcu_read_lock_bh rcu_read_lock rcu_read_lock_bh rcu_read_lock &data->lock irq_context: 0 (wq_completion)bond99#3 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &idev->mc_lock rcu_read_lock rcu_read_lock rcu_read_lock_bh rcu_read_lock rcu_read_lock_bh rcu_read_lock &obj_hash[i].lock irq_context: 0 (wq_completion)bond99#3 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &idev->mc_lock rcu_read_lock rcu_read_lock rcu_node_0 irq_context: 0 (wq_completion)bond99#3 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &idev->mc_lock rcu_read_lock rcu_read_lock &rq->__lock irq_context: 0 (wq_completion)bond99#3 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex.wait_lock irq_context: 0 (wq_completion)bond99#3 (work_completion)(&(&bond->mcast_work)->work) &p->pi_lock irq_context: 0 (wq_completion)bond99#3 (work_completion)(&(&slave->notify_work)->work) rtnl_mutex net_rwsem irq_context: 0 (wq_completion)bond99#3 (work_completion)(&(&slave->notify_work)->work) rtnl_mutex net_rwsem &list->lock#2 irq_context: 0 (wq_completion)bond99#3 (work_completion)(&(&slave->notify_work)->work) rtnl_mutex &tn->lock irq_context: 0 (wq_completion)bond99#3 (work_completion)(&(&slave->notify_work)->work) rtnl_mutex.wait_lock irq_context: 0 (wq_completion)bond99#3 (work_completion)(&(&slave->notify_work)->work) &p->pi_lock irq_context: 0 (wq_completion)bond138#2 irq_context: 0 (wq_completion)bond138#2 (work_completion)(&(&slave->notify_work)->work) irq_context: 0 (wq_completion)bond138#2 (work_completion)(&(&slave->notify_work)->work) &obj_hash[i].lock irq_context: 0 (wq_completion)bond138#2 (work_completion)(&(&slave->notify_work)->work) &base->lock irq_context: 0 (wq_completion)bond138#2 (work_completion)(&(&slave->notify_work)->work) &base->lock &obj_hash[i].lock irq_context: 0 (wq_completion)bond138#2 (work_completion)(&(&bond->mcast_work)->work) irq_context: 0 (wq_completion)bond138#2 (work_completion)(&(&bond->mcast_work)->work) &obj_hash[i].lock irq_context: 0 (wq_completion)bond138#2 (work_completion)(&(&bond->mcast_work)->work) &base->lock irq_context: 0 (wq_completion)bond138#2 (work_completion)(&(&bond->mcast_work)->work) &base->lock &obj_hash[i].lock irq_context: 0 (wq_completion)bond167 irq_context: 0 (wq_completion)bond167 (work_completion)(&(&slave->notify_work)->work) irq_context: 0 (wq_completion)bond167 (work_completion)(&(&slave->notify_work)->work) &obj_hash[i].lock irq_context: 0 (wq_completion)bond167 (work_completion)(&(&slave->notify_work)->work) &base->lock irq_context: 0 (wq_completion)bond167 (work_completion)(&(&slave->notify_work)->work) &base->lock &obj_hash[i].lock irq_context: 0 (wq_completion)bond167 (work_completion)(&(&slave->notify_work)->work) &rq->__lock irq_context: 0 (wq_completion)bond167 (work_completion)(&(&bond->mcast_work)->work) irq_context: 0 (wq_completion)bond167 (work_completion)(&(&bond->mcast_work)->work) &obj_hash[i].lock irq_context: 0 (wq_completion)bond167 (work_completion)(&(&bond->mcast_work)->work) &base->lock irq_context: 0 (wq_completion)bond167 (work_completion)(&(&bond->mcast_work)->work) &base->lock &obj_hash[i].lock irq_context: 0 (wq_completion)bond167 (work_completion)(&(&bond->mcast_work)->work) &rq->__lock irq_context: 0 (wq_completion)bond167 (work_completion)(&(&bond->mcast_work)->work) &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)bond166 (work_completion)(&(&bond->mcast_work)->work) &rq->__lock irq_context: 0 (wq_completion)bond166 (work_completion)(&(&bond->mcast_work)->work) &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)bond125#2 (work_completion)(&(&slave->notify_work)->work) &rq->__lock irq_context: 0 (wq_completion)bond125#2 (work_completion)(&(&slave->notify_work)->work) &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)bond155#2 irq_context: 0 (wq_completion)bond155#2 (work_completion)(&(&slave->notify_work)->work) irq_context: 0 (wq_completion)bond155#2 (work_completion)(&(&slave->notify_work)->work) &obj_hash[i].lock irq_context: 0 (wq_completion)bond155#2 (work_completion)(&(&slave->notify_work)->work) &base->lock irq_context: 0 (wq_completion)bond155#2 (work_completion)(&(&slave->notify_work)->work) &base->lock &obj_hash[i].lock irq_context: 0 (wq_completion)bond155#2 (work_completion)(&(&slave->notify_work)->work) &rq->__lock irq_context: 0 (wq_completion)bond155#2 (work_completion)(&(&slave->notify_work)->work) &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)bond155#2 (work_completion)(&(&bond->mcast_work)->work) irq_context: 0 (wq_completion)bond155#2 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex fs_reclaim irq_context: 0 (wq_completion)bond155#2 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex fs_reclaim mmu_notifier_invalidate_range_start irq_context: 0 (wq_completion)bond155#2 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &c->lock irq_context: 0 (wq_completion)bond155#2 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex rcu_read_lock &bond->stats_lock/1 irq_context: 0 (wq_completion)bond155#2 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex rcu_read_lock &ndev->lock irq_context: 0 (wq_completion)bond155#2 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex nl_table_lock irq_context: 0 (wq_completion)bond155#2 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &obj_hash[i].lock irq_context: 0 (wq_completion)bond155#2 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex nl_table_wait.lock irq_context: 0 (wq_completion)bond155#2 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex net_rwsem irq_context: 0 (wq_completion)bond155#2 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex net_rwsem &list->lock#2 irq_context: 0 (wq_completion)bond155#2 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &tn->lock irq_context: 0 (wq_completion)bond155#2 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &idev->mc_lock irq_context: 0 (wq_completion)bond155#2 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex.wait_lock irq_context: 0 (wq_completion)bond155#2 (work_completion)(&(&bond->mcast_work)->work) &p->pi_lock irq_context: 0 (wq_completion)bond160 (work_completion)(&(&slave->notify_work)->work) rtnl_mutex net_rwsem irq_context: 0 (wq_completion)bond160 (work_completion)(&(&slave->notify_work)->work) rtnl_mutex net_rwsem &list->lock#2 irq_context: 0 (wq_completion)bond160 (work_completion)(&(&slave->notify_work)->work) rtnl_mutex &tn->lock irq_context: 0 (wq_completion)bond160 (work_completion)(&(&slave->notify_work)->work) rtnl_mutex.wait_lock irq_context: 0 (wq_completion)bond160 (work_completion)(&(&slave->notify_work)->work) &p->pi_lock irq_context: 0 (wq_completion)bond149 (work_completion)(&(&slave->notify_work)->work) rtnl_mutex net_rwsem irq_context: 0 (wq_completion)bond149 (work_completion)(&(&slave->notify_work)->work) rtnl_mutex net_rwsem &list->lock#2 irq_context: 0 (wq_completion)bond149 (work_completion)(&(&slave->notify_work)->work) rtnl_mutex &tn->lock irq_context: 0 (wq_completion)bond149 (work_completion)(&(&slave->notify_work)->work) rtnl_mutex.wait_lock irq_context: 0 (wq_completion)bond149 (work_completion)(&(&slave->notify_work)->work) &p->pi_lock irq_context: 0 (wq_completion)bond42 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex fs_reclaim irq_context: 0 (wq_completion)bond42 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex fs_reclaim mmu_notifier_invalidate_range_start irq_context: 0 (wq_completion)bond42 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex remove_cache_srcu irq_context: 0 (wq_completion)bond42 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex remove_cache_srcu quarantine_lock irq_context: 0 (wq_completion)bond42 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex remove_cache_srcu &c->lock irq_context: 0 (wq_completion)bond42 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex remove_cache_srcu &n->list_lock irq_context: 0 (wq_completion)bond42 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex remove_cache_srcu &obj_hash[i].lock irq_context: 0 (wq_completion)bond42 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex rcu_read_lock &bond->stats_lock/1 irq_context: 0 (wq_completion)bond42 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex rcu_read_lock &ndev->lock irq_context: 0 (wq_completion)bond42 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex nl_table_lock irq_context: 0 (wq_completion)bond42 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &obj_hash[i].lock irq_context: 0 (wq_completion)bond42 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex nl_table_wait.lock irq_context: 0 (wq_completion)bond42 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex nl_table_wait.lock &p->pi_lock irq_context: 0 (wq_completion)bond42 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex nl_table_wait.lock &p->pi_lock &cfs_rq->removed.lock irq_context: 0 (wq_completion)bond42 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex nl_table_wait.lock &p->pi_lock &rq->__lock irq_context: 0 (wq_completion)bond42 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex nl_table_wait.lock &p->pi_lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)bond42 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &rq->__lock irq_context: 0 (wq_completion)bond42 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)bond42 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex net_rwsem irq_context: 0 (wq_completion)bond42 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex net_rwsem &list->lock#2 irq_context: 0 (wq_completion)bond42 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &tn->lock irq_context: 0 (wq_completion)bond42 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &idev->mc_lock irq_context: 0 (wq_completion)bond42 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &idev->mc_lock fs_reclaim irq_context: 0 (wq_completion)bond42 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &idev->mc_lock fs_reclaim mmu_notifier_invalidate_range_start irq_context: 0 (wq_completion)bond42 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &idev->mc_lock &c->lock irq_context: 0 (wq_completion)bond42 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &idev->mc_lock rcu_read_lock &ndev->lock irq_context: 0 (wq_completion)bond42 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &idev->mc_lock rcu_read_lock &dir->lock#2 irq_context: 0 (wq_completion)bond42 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &idev->mc_lock rcu_read_lock &ul->lock#2 irq_context: 0 (wq_completion)bond42 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &idev->mc_lock rcu_read_lock rcu_node_0 irq_context: 0 (wq_completion)bond42 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &idev->mc_lock rcu_read_lock &rq->__lock irq_context: 0 (wq_completion)bond42 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &idev->mc_lock rcu_read_lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)bond42 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &idev->mc_lock rcu_read_lock rcu_read_lock &____s->seqcount#12 irq_context: 0 (wq_completion)bond42 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &idev->mc_lock rcu_read_lock rcu_read_lock &____s->seqcount#16 irq_context: 0 (wq_completion)bond42 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &idev->mc_lock rcu_read_lock &obj_hash[i].lock irq_context: 0 (wq_completion)bond42 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &idev->mc_lock rcu_read_lock &data->lock irq_context: 0 (wq_completion)bond42 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex.wait_lock irq_context: 0 (wq_completion)bond42 (work_completion)(&(&bond->mcast_work)->work) &p->pi_lock irq_context: 0 (wq_completion)bond42 (work_completion)(&(&bond->mcast_work)->work) &p->pi_lock &rq->__lock irq_context: 0 (wq_completion)bond42 (work_completion)(&(&bond->mcast_work)->work) &p->pi_lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)bond135#3 &rq->__lock irq_context: 0 (wq_completion)bond135#3 &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)bond163 (work_completion)(&(&slave->notify_work)->work) fill_pool_map-wait-type-override pool_lock irq_context: 0 (wq_completion)bond150#2 &rq->__lock irq_context: 0 (wq_completion)bond150#2 &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)bond151 (work_completion)(&(&slave->notify_work)->work) &rq->__lock irq_context: 0 (wq_completion)bond151 (work_completion)(&(&slave->notify_work)->work) &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)bond11 (work_completion)(&(&bond->mii_work)->work) fill_pool_map-wait-type-override pool_lock#2 irq_context: 0 (wq_completion)bond11 (work_completion)(&(&bond->mii_work)->work) fill_pool_map-wait-type-override pool_lock irq_context: 0 (wq_completion)bond139#3 irq_context: 0 (wq_completion)bond139#3 (work_completion)(&(&slave->notify_work)->work) irq_context: 0 (wq_completion)bond139#3 (work_completion)(&(&slave->notify_work)->work) &obj_hash[i].lock irq_context: 0 (wq_completion)bond139#3 (work_completion)(&(&slave->notify_work)->work) &base->lock irq_context: 0 (wq_completion)bond139#3 (work_completion)(&(&slave->notify_work)->work) &base->lock &obj_hash[i].lock irq_context: 0 (wq_completion)bond139#3 (work_completion)(&(&bond->mcast_work)->work) irq_context: 0 (wq_completion)bond139#3 (work_completion)(&(&bond->mcast_work)->work) &obj_hash[i].lock irq_context: 0 (wq_completion)bond139#3 (work_completion)(&(&bond->mcast_work)->work) &base->lock irq_context: 0 (wq_completion)bond139#3 (work_completion)(&(&bond->mcast_work)->work) &base->lock &obj_hash[i].lock irq_context: 0 (wq_completion)bond139#3 (work_completion)(&(&bond->mcast_work)->work) &rq->__lock irq_context: 0 (wq_completion)bond139#3 (work_completion)(&(&bond->mcast_work)->work) &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)bond137#2 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex fs_reclaim irq_context: 0 (wq_completion)bond137#2 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex fs_reclaim mmu_notifier_invalidate_range_start irq_context: 0 (wq_completion)bond137#2 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex fs_reclaim mmu_notifier_invalidate_range_start &rq->__lock irq_context: 0 (wq_completion)bond137#2 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex fs_reclaim mmu_notifier_invalidate_range_start &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)bond137#2 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex remove_cache_srcu irq_context: 0 (wq_completion)bond137#2 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex remove_cache_srcu quarantine_lock irq_context: 0 (wq_completion)bond137#2 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex remove_cache_srcu &c->lock irq_context: 0 (wq_completion)bond137#2 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex remove_cache_srcu &n->list_lock irq_context: 0 (wq_completion)bond137#2 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex remove_cache_srcu &obj_hash[i].lock irq_context: 0 (wq_completion)bond137#2 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex rcu_read_lock rcu_node_0 irq_context: 0 (wq_completion)bond137#2 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex rcu_read_lock &rq->__lock irq_context: 0 (wq_completion)bond137#2 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex rcu_read_lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)bond137#2 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex rcu_read_lock &bond->stats_lock/1 irq_context: 0 (wq_completion)bond137#2 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex rcu_read_lock &ndev->lock irq_context: 0 (wq_completion)bond137#2 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex nl_table_lock irq_context: 0 (wq_completion)bond137#2 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &obj_hash[i].lock irq_context: 0 (wq_completion)bond137#2 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex nl_table_wait.lock irq_context: 0 (wq_completion)bond137#2 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex net_rwsem irq_context: 0 (wq_completion)bond137#2 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex net_rwsem &list->lock#2 irq_context: 0 (wq_completion)bond137#2 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &tn->lock irq_context: 0 (wq_completion)bond137#2 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &idev->mc_lock irq_context: 0 (wq_completion)bond137#2 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &idev->mc_lock fs_reclaim irq_context: 0 (wq_completion)bond137#2 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &idev->mc_lock fs_reclaim &rq->__lock irq_context: 0 (wq_completion)bond137#2 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &idev->mc_lock fs_reclaim &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)bond137#2 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &idev->mc_lock fs_reclaim mmu_notifier_invalidate_range_start irq_context: 0 (wq_completion)bond137#2 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &idev->mc_lock rcu_read_lock &ndev->lock irq_context: 0 (wq_completion)bond137#2 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &idev->mc_lock rcu_read_lock rcu_node_0 irq_context: 0 (wq_completion)bond137#2 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &idev->mc_lock rcu_read_lock &rq->__lock irq_context: 0 (wq_completion)bond137#2 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &idev->mc_lock rcu_read_lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)bond137#2 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &idev->mc_lock rcu_read_lock &dir->lock#2 irq_context: 0 (wq_completion)bond137#2 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &idev->mc_lock rcu_read_lock &ul->lock#2 irq_context: 0 (wq_completion)bond137#2 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &idev->mc_lock rcu_read_lock rcu_read_lock &____s->seqcount#10 irq_context: 0 (wq_completion)bond137#2 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &idev->mc_lock rcu_read_lock rcu_read_lock rcu_read_lock_bh rcu_read_lock rcu_read_lock_bh rcu_read_lock &br->multicast_lock irq_context: 0 (wq_completion)bond137#2 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &idev->mc_lock rcu_read_lock rcu_read_lock rcu_read_lock_bh rcu_read_lock rcu_read_lock_bh rcu_read_lock &br->multicast_lock &obj_hash[i].lock irq_context: 0 (wq_completion)bond137#2 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &idev->mc_lock rcu_read_lock rcu_read_lock rcu_read_lock_bh rcu_read_lock rcu_read_lock_bh rcu_read_lock &br->multicast_lock &base->lock irq_context: 0 (wq_completion)bond137#2 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &idev->mc_lock rcu_read_lock rcu_read_lock rcu_read_lock_bh rcu_read_lock rcu_read_lock_bh rcu_read_lock &br->multicast_lock &base->lock &obj_hash[i].lock irq_context: 0 (wq_completion)bond137#2 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &idev->mc_lock rcu_read_lock rcu_read_lock rcu_read_lock_bh rcu_read_lock rcu_read_lock_bh rcu_read_lock &data->lock irq_context: 0 (wq_completion)bond137#2 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &idev->mc_lock rcu_read_lock rcu_read_lock rcu_read_lock_bh rcu_read_lock rcu_read_lock_bh rcu_read_lock &obj_hash[i].lock irq_context: 0 (wq_completion)bond137#2 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &idev->mc_lock rcu_read_lock rcu_read_lock &rq->__lock irq_context: 0 (wq_completion)bond137#2 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &idev->mc_lock rcu_read_lock rcu_read_lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)bond137#2 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex.wait_lock irq_context: 0 (wq_completion)bond137#2 (work_completion)(&(&bond->mcast_work)->work) &p->pi_lock irq_context: 0 (wq_completion)bond137#2 (work_completion)(&(&bond->mcast_work)->work) &p->pi_lock &rq->__lock irq_context: 0 (wq_completion)bond137#2 (work_completion)(&(&bond->mcast_work)->work) &p->pi_lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)bond44#2 irq_context: 0 (wq_completion)bond44#2 &rq->__lock irq_context: 0 (wq_completion)bond44#2 &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)bond44#2 (work_completion)(&(&slave->notify_work)->work) irq_context: 0 (wq_completion)bond44#2 (work_completion)(&(&slave->notify_work)->work) &obj_hash[i].lock irq_context: 0 (wq_completion)bond44#2 (work_completion)(&(&slave->notify_work)->work) &base->lock irq_context: 0 (wq_completion)bond44#2 (work_completion)(&(&slave->notify_work)->work) &base->lock &obj_hash[i].lock irq_context: 0 (wq_completion)bond44#2 (work_completion)(&(&bond->mcast_work)->work) irq_context: 0 (wq_completion)bond44#2 (work_completion)(&(&bond->mcast_work)->work) &obj_hash[i].lock irq_context: 0 (wq_completion)bond44#2 (work_completion)(&(&bond->mcast_work)->work) &base->lock irq_context: 0 (wq_completion)bond44#2 (work_completion)(&(&bond->mcast_work)->work) &base->lock &obj_hash[i].lock irq_context: 0 (wq_completion)bond44#2 (work_completion)(&(&bond->mcast_work)->work) &rq->__lock irq_context: 0 (wq_completion)bond44#2 (work_completion)(&(&bond->mcast_work)->work) &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)bond132#2 (work_completion)(&(&bond->mcast_work)->work) &rq->__lock irq_context: 0 (wq_completion)bond132#2 (work_completion)(&(&bond->mcast_work)->work) &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)bond23#3 (work_completion)(&(&slave->notify_work)->work) fill_pool_map-wait-type-override pool_lock irq_context: 0 (wq_completion)bond153 (work_completion)(&(&bond->mcast_work)->work) &rq->__lock irq_context: 0 (wq_completion)bond153 (work_completion)(&(&bond->mcast_work)->work) &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)bond168 irq_context: 0 (wq_completion)bond168 &rq->__lock irq_context: 0 (wq_completion)bond168 &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)bond168 (work_completion)(&(&slave->notify_work)->work) irq_context: 0 (wq_completion)bond168 (work_completion)(&(&slave->notify_work)->work) &obj_hash[i].lock irq_context: 0 (wq_completion)bond168 (work_completion)(&(&slave->notify_work)->work) &base->lock irq_context: 0 (wq_completion)bond168 (work_completion)(&(&slave->notify_work)->work) &base->lock &obj_hash[i].lock irq_context: 0 (wq_completion)bond168 (work_completion)(&(&bond->mcast_work)->work) irq_context: 0 (wq_completion)bond168 (work_completion)(&(&bond->mcast_work)->work) &obj_hash[i].lock irq_context: 0 (wq_completion)bond168 (work_completion)(&(&bond->mcast_work)->work) &base->lock irq_context: 0 (wq_completion)bond168 (work_completion)(&(&bond->mcast_work)->work) &base->lock &obj_hash[i].lock irq_context: 0 (wq_completion)bond168 (work_completion)(&(&bond->mcast_work)->work) &rq->__lock irq_context: 0 (wq_completion)bond168 (work_completion)(&(&bond->mcast_work)->work) &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)bond132#3 (work_completion)(&(&bond->mcast_work)->work) fill_pool_map-wait-type-override &c->lock irq_context: 0 (wq_completion)bond132#3 (work_completion)(&(&bond->mcast_work)->work) fill_pool_map-wait-type-override &____s->seqcount#2 irq_context: 0 (wq_completion)bond132#3 (work_completion)(&(&bond->mcast_work)->work) fill_pool_map-wait-type-override &____s->seqcount irq_context: 0 (wq_completion)bond132#3 (work_completion)(&(&bond->mcast_work)->work) fill_pool_map-wait-type-override pool_lock irq_context: 0 (wq_completion)bond156#2 irq_context: 0 (wq_completion)bond156#2 (work_completion)(&(&slave->notify_work)->work) irq_context: 0 (wq_completion)bond156#2 (work_completion)(&(&slave->notify_work)->work) &obj_hash[i].lock irq_context: 0 (wq_completion)bond156#2 (work_completion)(&(&slave->notify_work)->work) &base->lock irq_context: 0 (wq_completion)bond156#2 (work_completion)(&(&slave->notify_work)->work) &base->lock &obj_hash[i].lock irq_context: 0 (wq_completion)bond156#2 (work_completion)(&(&bond->mcast_work)->work) irq_context: 0 (wq_completion)bond156#2 (work_completion)(&(&bond->mcast_work)->work) &obj_hash[i].lock irq_context: 0 (wq_completion)bond156#2 (work_completion)(&(&bond->mcast_work)->work) &base->lock irq_context: 0 (wq_completion)bond156#2 (work_completion)(&(&bond->mcast_work)->work) &base->lock &obj_hash[i].lock irq_context: 0 (wq_completion)bond156#2 (work_completion)(&(&bond->mcast_work)->work) &rq->__lock irq_context: 0 (wq_completion)bond156#2 (work_completion)(&(&bond->mcast_work)->work) &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)bond163 (work_completion)(&(&slave->notify_work)->work) rtnl_mutex net_rwsem irq_context: 0 (wq_completion)bond163 (work_completion)(&(&slave->notify_work)->work) rtnl_mutex net_rwsem &list->lock#2 irq_context: 0 (wq_completion)bond163 (work_completion)(&(&slave->notify_work)->work) rtnl_mutex &tn->lock irq_context: 0 (wq_completion)bond163 (work_completion)(&(&slave->notify_work)->work) rtnl_mutex.wait_lock irq_context: 0 (wq_completion)bond163 (work_completion)(&(&slave->notify_work)->work) &p->pi_lock irq_context: 0 (wq_completion)bond130 (work_completion)(&(&slave->notify_work)->work) rtnl_mutex net_rwsem irq_context: 0 (wq_completion)bond130 (work_completion)(&(&slave->notify_work)->work) rtnl_mutex net_rwsem &list->lock#2 irq_context: 0 (wq_completion)bond130 (work_completion)(&(&slave->notify_work)->work) rtnl_mutex &tn->lock irq_context: 0 (wq_completion)bond130 (work_completion)(&(&slave->notify_work)->work) rtnl_mutex.wait_lock irq_context: 0 (wq_completion)bond130 (work_completion)(&(&slave->notify_work)->work) &p->pi_lock irq_context: 0 (wq_completion)bond102#3 (work_completion)(&(&slave->notify_work)->work) rtnl_mutex net_rwsem irq_context: 0 (wq_completion)bond102#3 (work_completion)(&(&slave->notify_work)->work) rtnl_mutex net_rwsem &list->lock#2 irq_context: 0 (wq_completion)bond102#3 (work_completion)(&(&slave->notify_work)->work) rtnl_mutex &tn->lock irq_context: 0 (wq_completion)bond102#3 (work_completion)(&(&slave->notify_work)->work) rtnl_mutex.wait_lock irq_context: 0 (wq_completion)bond102#3 (work_completion)(&(&slave->notify_work)->work) &p->pi_lock irq_context: 0 (wq_completion)bond121#3 (work_completion)(&(&slave->notify_work)->work) rtnl_mutex net_rwsem irq_context: 0 (wq_completion)bond121#3 (work_completion)(&(&slave->notify_work)->work) rtnl_mutex net_rwsem &list->lock#2 irq_context: 0 (wq_completion)bond121#3 (work_completion)(&(&slave->notify_work)->work) rtnl_mutex &tn->lock irq_context: 0 (wq_completion)bond121#3 (work_completion)(&(&slave->notify_work)->work) rtnl_mutex.wait_lock irq_context: 0 (wq_completion)bond121#3 (work_completion)(&(&slave->notify_work)->work) &p->pi_lock irq_context: 0 (wq_completion)bond44#2 (work_completion)(&(&slave->notify_work)->work) rtnl_mutex net_rwsem irq_context: 0 (wq_completion)bond44#2 (work_completion)(&(&slave->notify_work)->work) rtnl_mutex net_rwsem &list->lock#2 irq_context: 0 (wq_completion)bond44#2 (work_completion)(&(&slave->notify_work)->work) rtnl_mutex &tn->lock irq_context: 0 (wq_completion)bond44#2 (work_completion)(&(&slave->notify_work)->work) rtnl_mutex.wait_lock irq_context: 0 (wq_completion)bond44#2 (work_completion)(&(&slave->notify_work)->work) &p->pi_lock irq_context: 0 (wq_completion)bond125#2 (work_completion)(&(&slave->notify_work)->work) rtnl_mutex net_rwsem irq_context: 0 (wq_completion)bond125#2 (work_completion)(&(&slave->notify_work)->work) rtnl_mutex net_rwsem &rq->__lock irq_context: 0 (wq_completion)bond125#2 (work_completion)(&(&slave->notify_work)->work) rtnl_mutex net_rwsem &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)bond125#2 (work_completion)(&(&slave->notify_work)->work) rtnl_mutex net_rwsem &list->lock#2 irq_context: 0 (wq_completion)bond125#2 (work_completion)(&(&slave->notify_work)->work) rtnl_mutex &tn->lock irq_context: 0 (wq_completion)bond125#2 (work_completion)(&(&slave->notify_work)->work) rtnl_mutex.wait_lock irq_context: 0 (wq_completion)bond125#2 (work_completion)(&(&slave->notify_work)->work) &p->pi_lock irq_context: 0 (wq_completion)bond125#2 (work_completion)(&(&slave->notify_work)->work) &p->pi_lock &rq->__lock irq_context: 0 (wq_completion)bond125#2 (work_completion)(&(&slave->notify_work)->work) &p->pi_lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)bond140#3 irq_context: 0 (wq_completion)bond140#3 (work_completion)(&(&slave->notify_work)->work) irq_context: 0 (wq_completion)bond140#3 (work_completion)(&(&slave->notify_work)->work) &obj_hash[i].lock irq_context: 0 (wq_completion)bond140#3 (work_completion)(&(&slave->notify_work)->work) &base->lock irq_context: 0 (wq_completion)bond140#3 (work_completion)(&(&slave->notify_work)->work) &base->lock &obj_hash[i].lock irq_context: 0 (wq_completion)bond108#2 &rq->__lock irq_context: 0 (wq_completion)bond108#2 &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)bond140#3 (work_completion)(&(&bond->mcast_work)->work) irq_context: 0 (wq_completion)bond140#3 (work_completion)(&(&bond->mcast_work)->work) &obj_hash[i].lock irq_context: 0 (wq_completion)bond140#3 (work_completion)(&(&bond->mcast_work)->work) &base->lock irq_context: 0 (wq_completion)bond140#3 (work_completion)(&(&bond->mcast_work)->work) &base->lock &obj_hash[i].lock irq_context: 0 (wq_completion)bond140#3 (work_completion)(&(&bond->mcast_work)->work) &rq->__lock irq_context: 0 (wq_completion)bond140#3 (work_completion)(&(&bond->mcast_work)->work) &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)bond149 (work_completion)(&(&bond->mcast_work)->work) fill_pool_map-wait-type-override pool_lock irq_context: 0 (wq_completion)bond169 irq_context: 0 (wq_completion)bond169 &rq->__lock irq_context: 0 (wq_completion)bond169 &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)bond169 (work_completion)(&(&slave->notify_work)->work) irq_context: 0 (wq_completion)bond169 (work_completion)(&(&slave->notify_work)->work) &obj_hash[i].lock irq_context: 0 (wq_completion)bond169 (work_completion)(&(&slave->notify_work)->work) &base->lock irq_context: 0 (wq_completion)bond169 (work_completion)(&(&slave->notify_work)->work) &base->lock &obj_hash[i].lock irq_context: 0 (wq_completion)bond136#3 (work_completion)(&(&slave->notify_work)->work) &rq->__lock irq_context: 0 (wq_completion)bond136#3 (work_completion)(&(&slave->notify_work)->work) &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)bond169 (work_completion)(&(&bond->mcast_work)->work) irq_context: 0 (wq_completion)bond169 (work_completion)(&(&bond->mcast_work)->work) &obj_hash[i].lock irq_context: 0 (wq_completion)bond169 (work_completion)(&(&bond->mcast_work)->work) &base->lock irq_context: 0 (wq_completion)bond169 (work_completion)(&(&bond->mcast_work)->work) &base->lock &obj_hash[i].lock irq_context: 0 (wq_completion)bond169 (work_completion)(&(&bond->mcast_work)->work) &rq->__lock irq_context: 0 (wq_completion)bond169 (work_completion)(&(&bond->mcast_work)->work) &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)bond155 (work_completion)(&(&slave->notify_work)->work) fill_pool_map-wait-type-override &c->lock irq_context: 0 (wq_completion)bond155 (work_completion)(&(&slave->notify_work)->work) fill_pool_map-wait-type-override &____s->seqcount#2 irq_context: 0 (wq_completion)bond155 (work_completion)(&(&slave->notify_work)->work) fill_pool_map-wait-type-override &____s->seqcount irq_context: 0 (wq_completion)bond155 (work_completion)(&(&slave->notify_work)->work) fill_pool_map-wait-type-override pool_lock irq_context: 0 (wq_completion)bond157#2 irq_context: 0 (wq_completion)bond157#2 (work_completion)(&(&slave->notify_work)->work) irq_context: 0 (wq_completion)bond157#2 (work_completion)(&(&slave->notify_work)->work) &obj_hash[i].lock irq_context: 0 (wq_completion)bond157#2 (work_completion)(&(&slave->notify_work)->work) &base->lock irq_context: 0 (wq_completion)bond157#2 (work_completion)(&(&slave->notify_work)->work) &base->lock &obj_hash[i].lock irq_context: 0 (wq_completion)bond157#2 (work_completion)(&(&slave->notify_work)->work) &rq->__lock irq_context: 0 (wq_completion)bond157#2 (work_completion)(&(&slave->notify_work)->work) &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)events pcpu_balance_work pcpu_alloc_mutex &pool->lock &p->pi_lock irq_context: 0 (wq_completion)events pcpu_balance_work pcpu_alloc_mutex &pool->lock &p->pi_lock &rq->__lock irq_context: 0 (wq_completion)events pcpu_balance_work pcpu_alloc_mutex &pool->lock &p->pi_lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)bond157#2 (work_completion)(&(&bond->mcast_work)->work) irq_context: 0 (wq_completion)bond157#2 (work_completion)(&(&bond->mcast_work)->work) &obj_hash[i].lock irq_context: 0 (wq_completion)bond157#2 (work_completion)(&(&bond->mcast_work)->work) &base->lock irq_context: 0 (wq_completion)bond157#2 (work_completion)(&(&bond->mcast_work)->work) &base->lock &obj_hash[i].lock irq_context: 0 (wq_completion)bond157#2 (work_completion)(&(&bond->mcast_work)->work) &rq->__lock irq_context: 0 (wq_completion)bond157#2 (work_completion)(&(&bond->mcast_work)->work) &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)bond157#2 (work_completion)(&(&bond->mcast_work)->work) &cfs_rq->removed.lock irq_context: 0 (wq_completion)bond141#3 irq_context: 0 (wq_completion)bond141#3 (work_completion)(&(&slave->notify_work)->work) irq_context: 0 (wq_completion)bond141#3 (work_completion)(&(&slave->notify_work)->work) &obj_hash[i].lock irq_context: 0 (wq_completion)bond141#3 (work_completion)(&(&slave->notify_work)->work) &base->lock irq_context: 0 (wq_completion)bond141#3 (work_completion)(&(&slave->notify_work)->work) &base->lock &obj_hash[i].lock irq_context: 0 (wq_completion)bond141#3 (work_completion)(&(&slave->notify_work)->work) &rq->__lock irq_context: 0 (wq_completion)bond141#3 (work_completion)(&(&slave->notify_work)->work) &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: softirq (&in_dev->mr_ifc_timer) rcu_read_lock &tbl->lock batched_entropy_u32.lock crngs.lock irq_context: 0 (wq_completion)bond129 (work_completion)(&(&bond->mcast_work)->work) fill_pool_map-wait-type-override pool_lock irq_context: 0 (wq_completion)bond141#3 (work_completion)(&(&bond->mcast_work)->work) irq_context: 0 (wq_completion)bond141#3 (work_completion)(&(&bond->mcast_work)->work) &obj_hash[i].lock irq_context: 0 (wq_completion)bond141#3 (work_completion)(&(&bond->mcast_work)->work) &base->lock irq_context: 0 (wq_completion)bond141#3 (work_completion)(&(&bond->mcast_work)->work) &base->lock &obj_hash[i].lock irq_context: 0 (wq_completion)bond151 (work_completion)(&(&slave->notify_work)->work) rtnl_mutex net_rwsem irq_context: 0 (wq_completion)bond151 (work_completion)(&(&slave->notify_work)->work) rtnl_mutex net_rwsem &list->lock#2 irq_context: 0 (wq_completion)bond151 (work_completion)(&(&slave->notify_work)->work) rtnl_mutex &tn->lock irq_context: 0 (wq_completion)bond151 (work_completion)(&(&slave->notify_work)->work) rtnl_mutex.wait_lock irq_context: 0 (wq_completion)bond151 (work_completion)(&(&slave->notify_work)->work) &p->pi_lock irq_context: 0 (wq_completion)bond106#3 (work_completion)(&(&slave->notify_work)->work) rtnl_mutex net_rwsem irq_context: 0 (wq_completion)bond106#3 (work_completion)(&(&slave->notify_work)->work) rtnl_mutex net_rwsem &list->lock#2 irq_context: 0 (wq_completion)bond106#3 (work_completion)(&(&slave->notify_work)->work) rtnl_mutex &tn->lock irq_context: 0 (wq_completion)bond106#3 (work_completion)(&(&slave->notify_work)->work) rtnl_mutex.wait_lock irq_context: 0 (wq_completion)bond106#3 (work_completion)(&(&slave->notify_work)->work) &p->pi_lock irq_context: 0 (wq_completion)bond93#2 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex fs_reclaim irq_context: 0 (wq_completion)bond93#2 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex fs_reclaim mmu_notifier_invalidate_range_start irq_context: 0 (wq_completion)bond93#2 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &c->lock irq_context: 0 (wq_completion)bond93#2 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex rcu_read_lock &bond->stats_lock/1 irq_context: 0 (wq_completion)bond93#2 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex rcu_read_lock &ndev->lock irq_context: 0 (wq_completion)bond93#2 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex nl_table_lock irq_context: 0 (wq_completion)bond93#2 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &obj_hash[i].lock irq_context: 0 (wq_completion)bond93#2 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex nl_table_wait.lock irq_context: 0 (wq_completion)bond93#2 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex net_rwsem irq_context: 0 (wq_completion)bond93#2 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex net_rwsem &list->lock#2 irq_context: 0 (wq_completion)bond93#2 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &tn->lock irq_context: 0 (wq_completion)bond93#2 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &idev->mc_lock irq_context: 0 (wq_completion)bond93#2 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &idev->mc_lock fs_reclaim irq_context: 0 (wq_completion)bond93#2 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &idev->mc_lock fs_reclaim mmu_notifier_invalidate_range_start irq_context: 0 (wq_completion)bond93#2 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &idev->mc_lock rcu_read_lock &ndev->lock irq_context: 0 (wq_completion)bond93#2 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &idev->mc_lock rcu_read_lock rcu_node_0 irq_context: 0 (wq_completion)bond93#2 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &idev->mc_lock rcu_read_lock &rq->__lock irq_context: 0 (wq_completion)bond93#2 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &idev->mc_lock rcu_read_lock &dir->lock#2 irq_context: 0 (wq_completion)bond93#2 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &idev->mc_lock rcu_read_lock &c->lock irq_context: 0 (wq_completion)bond93#2 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &idev->mc_lock rcu_read_lock &ul->lock#2 irq_context: 0 (wq_completion)bond93#2 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &idev->mc_lock rcu_read_lock rcu_read_lock &obj_hash[i].lock irq_context: 0 (wq_completion)bond93#2 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &idev->mc_lock rcu_read_lock rcu_read_lock &dir->lock#2 irq_context: 0 (wq_completion)bond93#2 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &idev->mc_lock rcu_read_lock rcu_read_lock &tbl->lock irq_context: 0 (wq_completion)bond93#2 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &idev->mc_lock rcu_read_lock rcu_read_lock &n->lock irq_context: 0 (wq_completion)bond93#2 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &idev->mc_lock rcu_read_lock rcu_read_lock &____s->seqcount#9 irq_context: 0 (wq_completion)bond93#2 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &idev->mc_lock rcu_read_lock rcu_read_lock rcu_read_lock_bh rcu_read_lock rcu_read_lock_bh rcu_read_lock &br->multicast_lock irq_context: 0 (wq_completion)bond93#2 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &idev->mc_lock rcu_read_lock rcu_read_lock rcu_read_lock_bh rcu_read_lock rcu_read_lock_bh rcu_read_lock &br->multicast_lock &obj_hash[i].lock irq_context: 0 (wq_completion)bond93#2 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &idev->mc_lock rcu_read_lock rcu_read_lock rcu_read_lock_bh rcu_read_lock rcu_read_lock_bh rcu_read_lock &br->multicast_lock &base->lock irq_context: 0 (wq_completion)bond93#2 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &idev->mc_lock rcu_read_lock rcu_read_lock rcu_read_lock_bh rcu_read_lock rcu_read_lock_bh rcu_read_lock &br->multicast_lock &base->lock &obj_hash[i].lock irq_context: 0 (wq_completion)bond93#2 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &idev->mc_lock rcu_read_lock rcu_read_lock rcu_read_lock_bh rcu_read_lock rcu_read_lock_bh rcu_read_lock &data->lock irq_context: 0 (wq_completion)bond93#2 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &idev->mc_lock rcu_read_lock rcu_read_lock rcu_read_lock_bh rcu_read_lock rcu_read_lock_bh rcu_read_lock &obj_hash[i].lock irq_context: 0 (wq_completion)bond93#2 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex.wait_lock irq_context: 0 (wq_completion)bond93#2 (work_completion)(&(&bond->mcast_work)->work) &p->pi_lock irq_context: 0 (wq_completion)bond25#3 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex fs_reclaim irq_context: 0 (wq_completion)bond25#3 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex fs_reclaim mmu_notifier_invalidate_range_start irq_context: 0 (wq_completion)bond25#3 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &c->lock irq_context: 0 (wq_completion)bond25#3 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex rcu_read_lock &bond->stats_lock/1 irq_context: 0 (wq_completion)bond25#3 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex rcu_read_lock &ndev->lock irq_context: 0 (wq_completion)bond25#3 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex nl_table_lock irq_context: 0 (wq_completion)bond25#3 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &obj_hash[i].lock irq_context: 0 (wq_completion)bond25#3 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex nl_table_wait.lock irq_context: 0 (wq_completion)bond25#3 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex net_rwsem irq_context: 0 (wq_completion)bond25#3 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex net_rwsem &list->lock#2 irq_context: 0 (wq_completion)bond25#3 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &tn->lock irq_context: 0 (wq_completion)bond25#3 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &idev->mc_lock irq_context: 0 (wq_completion)bond25#3 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &idev->mc_lock fs_reclaim irq_context: 0 (wq_completion)bond25#3 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &idev->mc_lock fs_reclaim mmu_notifier_invalidate_range_start irq_context: 0 (wq_completion)bond25#3 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &idev->mc_lock rcu_read_lock &ndev->lock irq_context: 0 (wq_completion)bond25#3 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &idev->mc_lock rcu_read_lock &dir->lock#2 irq_context: 0 (wq_completion)bond25#3 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &idev->mc_lock rcu_read_lock &ul->lock#2 irq_context: 0 (wq_completion)bond25#3 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &idev->mc_lock rcu_read_lock rcu_read_lock &____s->seqcount#12 irq_context: 0 (wq_completion)bond25#3 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &idev->mc_lock rcu_read_lock rcu_read_lock &____s->seqcount#16 irq_context: 0 (wq_completion)bond25#3 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &idev->mc_lock rcu_read_lock &obj_hash[i].lock irq_context: 0 (wq_completion)bond25#3 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &idev->mc_lock rcu_read_lock &data->lock irq_context: 0 (wq_completion)bond25#3 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &idev->mc_lock rcu_read_lock rcu_node_0 irq_context: 0 (wq_completion)bond25#3 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &idev->mc_lock rcu_read_lock &rq->__lock irq_context: 0 (wq_completion)bond25#3 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &idev->mc_lock rcu_read_lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)bond25#3 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex.wait_lock irq_context: 0 (wq_completion)bond25#3 (work_completion)(&(&bond->mcast_work)->work) &p->pi_lock irq_context: 0 (wq_completion)bond25#3 (work_completion)(&(&bond->mcast_work)->work) &p->pi_lock &rq->__lock irq_context: 0 (wq_completion)bond25#3 (work_completion)(&(&bond->mcast_work)->work) &p->pi_lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)bond170 irq_context: 0 (wq_completion)bond170 (work_completion)(&(&slave->notify_work)->work) irq_context: 0 (wq_completion)bond170 (work_completion)(&(&slave->notify_work)->work) &obj_hash[i].lock irq_context: 0 (wq_completion)bond170 (work_completion)(&(&slave->notify_work)->work) &base->lock irq_context: 0 (wq_completion)bond170 (work_completion)(&(&slave->notify_work)->work) &base->lock &obj_hash[i].lock irq_context: 0 (wq_completion)bond170 (work_completion)(&(&slave->notify_work)->work) &rq->__lock irq_context: 0 (wq_completion)bond170 (work_completion)(&(&slave->notify_work)->work) &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: softirq (&in_dev->mr_ifc_timer) rcu_read_lock &pcp->lock &zone->lock irq_context: softirq (&in_dev->mr_ifc_timer) rcu_read_lock &pcp->lock &zone->lock &____s->seqcount irq_context: 0 (wq_completion)bond170 (work_completion)(&(&bond->mcast_work)->work) irq_context: 0 (wq_completion)bond170 (work_completion)(&(&bond->mcast_work)->work) &obj_hash[i].lock irq_context: 0 (wq_completion)bond170 (work_completion)(&(&bond->mcast_work)->work) &base->lock irq_context: 0 (wq_completion)bond170 (work_completion)(&(&bond->mcast_work)->work) &base->lock &obj_hash[i].lock irq_context: 0 (wq_completion)bond36#2 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex fs_reclaim irq_context: 0 (wq_completion)bond36#2 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex fs_reclaim mmu_notifier_invalidate_range_start irq_context: 0 (wq_completion)bond36#2 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &c->lock irq_context: 0 (wq_completion)bond36#2 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex rcu_read_lock &bond->stats_lock/1 irq_context: 0 (wq_completion)bond36#2 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex rcu_read_lock &ndev->lock irq_context: 0 (wq_completion)bond36#2 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex nl_table_lock irq_context: 0 (wq_completion)bond36#2 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &obj_hash[i].lock irq_context: 0 (wq_completion)bond36#2 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex nl_table_wait.lock irq_context: 0 (wq_completion)bond36#2 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex net_rwsem irq_context: 0 (wq_completion)bond36#2 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex net_rwsem &list->lock#2 irq_context: 0 (wq_completion)bond36#2 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &tn->lock irq_context: 0 (wq_completion)bond36#2 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &idev->mc_lock irq_context: 0 (wq_completion)bond36#2 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &idev->mc_lock fs_reclaim irq_context: 0 (wq_completion)bond36#2 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &idev->mc_lock fs_reclaim mmu_notifier_invalidate_range_start irq_context: 0 (wq_completion)bond36#2 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &idev->mc_lock &c->lock irq_context: 0 (wq_completion)bond36#2 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &idev->mc_lock &n->list_lock irq_context: 0 (wq_completion)bond36#2 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &idev->mc_lock &n->list_lock &c->lock irq_context: 0 (wq_completion)bond36#2 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &idev->mc_lock rcu_read_lock &ndev->lock irq_context: 0 (wq_completion)bond36#2 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &idev->mc_lock rcu_read_lock &dir->lock#2 irq_context: 0 (wq_completion)bond36#2 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &idev->mc_lock rcu_read_lock &c->lock irq_context: 0 (wq_completion)bond36#2 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &idev->mc_lock rcu_read_lock &ul->lock#2 irq_context: 0 (wq_completion)bond36#2 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &idev->mc_lock rcu_read_lock rcu_read_lock &____s->seqcount#12 irq_context: 0 (wq_completion)bond36#2 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &idev->mc_lock rcu_read_lock rcu_read_lock &____s->seqcount#16 irq_context: 0 (wq_completion)bond36#2 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &idev->mc_lock rcu_read_lock &obj_hash[i].lock irq_context: 0 (wq_completion)bond36#2 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &idev->mc_lock rcu_read_lock &data->lock irq_context: 0 (wq_completion)bond36#2 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex.wait_lock irq_context: 0 (wq_completion)bond36#2 (work_completion)(&(&bond->mcast_work)->work) &p->pi_lock irq_context: 0 (wq_completion)bond116#3 (work_completion)(&(&slave->notify_work)->work) rtnl_mutex net_rwsem irq_context: 0 (wq_completion)bond116#3 (work_completion)(&(&slave->notify_work)->work) rtnl_mutex net_rwsem &list->lock#2 irq_context: 0 (wq_completion)bond116#3 (work_completion)(&(&slave->notify_work)->work) rtnl_mutex &tn->lock irq_context: 0 (wq_completion)bond116#3 (work_completion)(&(&slave->notify_work)->work) rtnl_mutex.wait_lock irq_context: 0 (wq_completion)bond42 (work_completion)(&(&slave->notify_work)->work) rtnl_mutex net_rwsem irq_context: 0 (wq_completion)bond116#3 (work_completion)(&(&slave->notify_work)->work) &p->pi_lock irq_context: 0 (wq_completion)bond42 (work_completion)(&(&slave->notify_work)->work) rtnl_mutex net_rwsem &list->lock#2 irq_context: 0 (wq_completion)bond42 (work_completion)(&(&slave->notify_work)->work) rtnl_mutex &tn->lock irq_context: 0 (wq_completion)bond42 (work_completion)(&(&slave->notify_work)->work) rtnl_mutex.wait_lock irq_context: 0 (wq_completion)bond32#3 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex fs_reclaim irq_context: 0 (wq_completion)bond42 (work_completion)(&(&slave->notify_work)->work) &p->pi_lock irq_context: 0 (wq_completion)bond32#3 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex fs_reclaim mmu_notifier_invalidate_range_start irq_context: 0 (wq_completion)bond32#3 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &c->lock irq_context: 0 (wq_completion)bond32#3 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex rcu_read_lock &bond->stats_lock/1 irq_context: 0 (wq_completion)bond32#3 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex rcu_read_lock &ndev->lock irq_context: 0 (wq_completion)bond32#3 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex nl_table_lock irq_context: 0 (wq_completion)bond32#3 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &obj_hash[i].lock irq_context: 0 (wq_completion)bond32#3 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex nl_table_wait.lock irq_context: 0 (wq_completion)bond32#3 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex net_rwsem irq_context: 0 (wq_completion)bond32#3 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex net_rwsem &list->lock#2 irq_context: 0 (wq_completion)bond32#3 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &tn->lock irq_context: 0 (wq_completion)bond32#3 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &idev->mc_lock irq_context: 0 (wq_completion)bond32#3 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &idev->mc_lock fs_reclaim irq_context: 0 (wq_completion)bond32#3 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &idev->mc_lock fs_reclaim mmu_notifier_invalidate_range_start irq_context: 0 (wq_completion)bond32#3 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &idev->mc_lock &c->lock irq_context: 0 (wq_completion)bond32#3 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &idev->mc_lock rcu_read_lock &ndev->lock irq_context: 0 (wq_completion)bond32#3 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &idev->mc_lock rcu_read_lock &dir->lock#2 irq_context: 0 (wq_completion)bond32#3 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &idev->mc_lock rcu_read_lock &ul->lock#2 irq_context: 0 (wq_completion)bond32#3 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &idev->mc_lock rcu_read_lock rcu_read_lock &____s->seqcount#12 irq_context: 0 (wq_completion)bond32#3 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &idev->mc_lock rcu_read_lock rcu_read_lock &____s->seqcount#16 irq_context: 0 (wq_completion)bond32#3 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &idev->mc_lock rcu_read_lock &obj_hash[i].lock irq_context: 0 (wq_completion)bond32#3 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &idev->mc_lock rcu_read_lock &data->lock irq_context: 0 (wq_completion)bond32#3 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex.wait_lock irq_context: 0 (wq_completion)bond32#3 (work_completion)(&(&bond->mcast_work)->work) &p->pi_lock irq_context: 0 (wq_completion)bond156 (work_completion)(&(&slave->notify_work)->work) rtnl_mutex net_rwsem irq_context: 0 (wq_completion)bond156 (work_completion)(&(&slave->notify_work)->work) rtnl_mutex net_rwsem &list->lock#2 irq_context: 0 (wq_completion)bond156 (work_completion)(&(&slave->notify_work)->work) rtnl_mutex &tn->lock irq_context: 0 (wq_completion)bond156 (work_completion)(&(&slave->notify_work)->work) rtnl_mutex.wait_lock irq_context: 0 (wq_completion)bond156 (work_completion)(&(&slave->notify_work)->work) &p->pi_lock irq_context: 0 (wq_completion)bond120#3 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex fs_reclaim irq_context: 0 (wq_completion)bond120#3 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex fs_reclaim mmu_notifier_invalidate_range_start irq_context: 0 (wq_completion)bond120#3 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex fs_reclaim mmu_notifier_invalidate_range_start &rq->__lock irq_context: 0 (wq_completion)bond120#3 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex fs_reclaim mmu_notifier_invalidate_range_start &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)bond120#3 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &c->lock irq_context: 0 (wq_completion)bond120#3 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &n->list_lock irq_context: 0 (wq_completion)bond120#3 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &n->list_lock &c->lock irq_context: 0 (wq_completion)bond120#3 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex rcu_read_lock &bond->stats_lock/1 irq_context: 0 (wq_completion)bond120#3 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex rcu_read_lock &ndev->lock irq_context: 0 (wq_completion)bond120#3 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex nl_table_lock irq_context: 0 (wq_completion)bond120#3 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &obj_hash[i].lock irq_context: 0 (wq_completion)bond120#3 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex nl_table_wait.lock irq_context: 0 (wq_completion)bond120#3 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex net_rwsem irq_context: 0 (wq_completion)bond120#3 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex net_rwsem &list->lock#2 irq_context: 0 (wq_completion)bond120#3 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &tn->lock irq_context: 0 (wq_completion)bond120#3 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &idev->mc_lock irq_context: 0 (wq_completion)bond120#3 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &idev->mc_lock fs_reclaim irq_context: 0 (wq_completion)bond120#3 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &idev->mc_lock fs_reclaim mmu_notifier_invalidate_range_start irq_context: 0 (wq_completion)bond120#3 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &idev->mc_lock rcu_read_lock &ndev->lock irq_context: 0 (wq_completion)bond120#3 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &idev->mc_lock rcu_read_lock &dir->lock#2 irq_context: 0 (wq_completion)bond120#3 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &idev->mc_lock rcu_read_lock &ul->lock#2 irq_context: 0 (wq_completion)bond120#3 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &idev->mc_lock rcu_read_lock rcu_read_lock &____s->seqcount#10 irq_context: 0 (wq_completion)bond120#3 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &idev->mc_lock rcu_read_lock rcu_read_lock rcu_read_lock_bh rcu_read_lock rcu_read_lock_bh rcu_read_lock &br->multicast_lock irq_context: 0 (wq_completion)bond120#3 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &idev->mc_lock rcu_read_lock rcu_read_lock rcu_read_lock_bh rcu_read_lock rcu_read_lock_bh rcu_read_lock &br->multicast_lock &obj_hash[i].lock irq_context: 0 (wq_completion)bond120#3 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &idev->mc_lock rcu_read_lock rcu_read_lock rcu_read_lock_bh rcu_read_lock rcu_read_lock_bh rcu_read_lock &br->multicast_lock &base->lock irq_context: 0 (wq_completion)bond120#3 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &idev->mc_lock rcu_read_lock rcu_read_lock rcu_read_lock_bh rcu_read_lock rcu_read_lock_bh rcu_read_lock &br->multicast_lock &base->lock &obj_hash[i].lock irq_context: 0 (wq_completion)bond120#3 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &idev->mc_lock rcu_read_lock rcu_read_lock rcu_read_lock_bh rcu_read_lock rcu_read_lock_bh rcu_read_lock &data->lock irq_context: 0 (wq_completion)bond120#3 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &idev->mc_lock rcu_read_lock rcu_read_lock rcu_read_lock_bh rcu_read_lock rcu_read_lock_bh rcu_read_lock &obj_hash[i].lock irq_context: 0 (wq_completion)bond120#3 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex.wait_lock irq_context: 0 (wq_completion)bond120#3 (work_completion)(&(&bond->mcast_work)->work) &p->pi_lock irq_context: 0 (wq_completion)bond141#3 (work_completion)(&(&slave->notify_work)->work) rtnl_mutex net_rwsem irq_context: 0 (wq_completion)bond141#3 (work_completion)(&(&slave->notify_work)->work) rtnl_mutex net_rwsem &rq->__lock irq_context: 0 (wq_completion)bond141#3 (work_completion)(&(&slave->notify_work)->work) rtnl_mutex net_rwsem &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)bond141#3 (work_completion)(&(&slave->notify_work)->work) rtnl_mutex net_rwsem &list->lock#2 irq_context: 0 (wq_completion)bond141#3 (work_completion)(&(&slave->notify_work)->work) rtnl_mutex &tn->lock irq_context: 0 (wq_completion)bond141#3 (work_completion)(&(&slave->notify_work)->work) rtnl_mutex.wait_lock irq_context: 0 (wq_completion)bond105#2 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex fs_reclaim irq_context: 0 (wq_completion)bond141#3 (work_completion)(&(&slave->notify_work)->work) &p->pi_lock irq_context: 0 (wq_completion)bond105#2 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex fs_reclaim mmu_notifier_invalidate_range_start irq_context: 0 (wq_completion)bond105#2 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &c->lock irq_context: 0 (wq_completion)bond105#2 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex rcu_read_lock &bond->stats_lock/1 irq_context: 0 (wq_completion)bond105#2 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex rcu_read_lock &ndev->lock irq_context: 0 (wq_completion)bond105#2 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex nl_table_lock irq_context: 0 (wq_completion)bond105#2 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &obj_hash[i].lock irq_context: 0 (wq_completion)bond105#2 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex nl_table_wait.lock irq_context: 0 (wq_completion)bond105#2 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex net_rwsem irq_context: 0 (wq_completion)bond105#2 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex net_rwsem &list->lock#2 irq_context: 0 (wq_completion)bond105#2 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &tn->lock irq_context: 0 (wq_completion)bond105#2 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &idev->mc_lock irq_context: 0 (wq_completion)bond105#2 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &idev->mc_lock &rq->__lock irq_context: 0 (wq_completion)bond105#2 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &idev->mc_lock fs_reclaim irq_context: 0 (wq_completion)bond105#2 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &idev->mc_lock fs_reclaim mmu_notifier_invalidate_range_start irq_context: 0 (wq_completion)bond105#2 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &idev->mc_lock &c->lock irq_context: 0 (wq_completion)bond105#2 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &idev->mc_lock rcu_read_lock &ndev->lock irq_context: 0 (wq_completion)bond105#2 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &idev->mc_lock rcu_read_lock &dir->lock#2 irq_context: 0 (wq_completion)bond105#2 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &idev->mc_lock rcu_read_lock &ul->lock#2 irq_context: 0 (wq_completion)bond105#2 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &idev->mc_lock rcu_read_lock rcu_read_lock &obj_hash[i].lock irq_context: 0 (wq_completion)bond105#2 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &idev->mc_lock rcu_read_lock rcu_read_lock &dir->lock#2 irq_context: 0 (wq_completion)bond105#2 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &idev->mc_lock rcu_read_lock rcu_read_lock &tbl->lock irq_context: 0 (wq_completion)bond105#2 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &idev->mc_lock rcu_read_lock rcu_read_lock &n->lock irq_context: 0 (wq_completion)bond105#2 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &idev->mc_lock rcu_read_lock rcu_read_lock &____s->seqcount#9 irq_context: 0 (wq_completion)bond105#2 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &idev->mc_lock rcu_read_lock rcu_read_lock rcu_read_lock_bh rcu_read_lock rcu_read_lock_bh rcu_read_lock &br->multicast_lock irq_context: 0 (wq_completion)bond105#2 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &idev->mc_lock rcu_read_lock rcu_read_lock rcu_read_lock_bh rcu_read_lock rcu_read_lock_bh rcu_read_lock &br->multicast_lock &obj_hash[i].lock irq_context: 0 (wq_completion)bond105#2 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &idev->mc_lock rcu_read_lock rcu_read_lock rcu_read_lock_bh rcu_read_lock rcu_read_lock_bh rcu_read_lock &br->multicast_lock &base->lock irq_context: 0 (wq_completion)bond105#2 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &idev->mc_lock rcu_read_lock rcu_read_lock rcu_read_lock_bh rcu_read_lock rcu_read_lock_bh rcu_read_lock &br->multicast_lock &base->lock &obj_hash[i].lock irq_context: 0 (wq_completion)bond105#2 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &idev->mc_lock rcu_read_lock rcu_read_lock rcu_read_lock_bh rcu_read_lock rcu_read_lock_bh rcu_read_lock &data->lock irq_context: 0 (wq_completion)bond105#2 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &idev->mc_lock rcu_read_lock rcu_read_lock rcu_read_lock_bh rcu_read_lock rcu_read_lock_bh rcu_read_lock &obj_hash[i].lock irq_context: 0 (wq_completion)bond105#2 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex.wait_lock irq_context: 0 (wq_completion)bond105#2 (work_completion)(&(&bond->mcast_work)->work) &p->pi_lock irq_context: 0 (wq_completion)bond105#2 (work_completion)(&(&bond->mcast_work)->work) &p->pi_lock &rq->__lock irq_context: 0 (wq_completion)bond105#2 (work_completion)(&(&bond->mcast_work)->work) &p->pi_lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)bond36#2 (work_completion)(&(&slave->notify_work)->work) rtnl_mutex net_rwsem irq_context: 0 (wq_completion)bond36#2 (work_completion)(&(&slave->notify_work)->work) rtnl_mutex net_rwsem &list->lock#2 irq_context: 0 (wq_completion)bond36#2 (work_completion)(&(&slave->notify_work)->work) rtnl_mutex &tn->lock irq_context: 0 (wq_completion)bond36#2 (work_completion)(&(&slave->notify_work)->work) rtnl_mutex.wait_lock irq_context: 0 (wq_completion)bond36#2 (work_completion)(&(&slave->notify_work)->work) &p->pi_lock irq_context: 0 (wq_completion)bond33#2 (work_completion)(&(&slave->notify_work)->work) rtnl_mutex net_rwsem irq_context: 0 (wq_completion)bond33#2 (work_completion)(&(&slave->notify_work)->work) rtnl_mutex net_rwsem &list->lock#2 irq_context: 0 (wq_completion)bond33#2 (work_completion)(&(&slave->notify_work)->work) rtnl_mutex &tn->lock irq_context: 0 (wq_completion)bond33#2 (work_completion)(&(&slave->notify_work)->work) rtnl_mutex.wait_lock irq_context: 0 (wq_completion)bond33#2 (work_completion)(&(&slave->notify_work)->work) &p->pi_lock irq_context: 0 (wq_completion)bond141#3 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex fs_reclaim irq_context: 0 (wq_completion)bond141#3 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex fs_reclaim mmu_notifier_invalidate_range_start irq_context: 0 (wq_completion)bond141#3 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex rcu_read_lock &bond->stats_lock/1 irq_context: 0 (wq_completion)bond141#3 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex rcu_read_lock &ndev->lock irq_context: 0 (wq_completion)bond141#3 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex nl_table_lock irq_context: 0 (wq_completion)bond141#3 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &obj_hash[i].lock irq_context: 0 (wq_completion)bond141#3 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex nl_table_wait.lock irq_context: 0 (wq_completion)bond141#3 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex net_rwsem irq_context: 0 (wq_completion)bond141#3 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex net_rwsem &list->lock#2 irq_context: 0 (wq_completion)bond141#3 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &tn->lock irq_context: 0 (wq_completion)bond141#3 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &idev->mc_lock irq_context: 0 (wq_completion)bond141#3 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &idev->mc_lock fs_reclaim irq_context: 0 (wq_completion)bond141#3 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &idev->mc_lock fs_reclaim mmu_notifier_invalidate_range_start irq_context: 0 (wq_completion)bond141#3 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &idev->mc_lock &c->lock irq_context: 0 (wq_completion)bond141#3 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &idev->mc_lock rcu_read_lock &ndev->lock irq_context: 0 (wq_completion)bond141#3 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &idev->mc_lock rcu_read_lock &dir->lock#2 irq_context: 0 (wq_completion)bond141#3 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &idev->mc_lock rcu_read_lock &ul->lock#2 irq_context: 0 (wq_completion)bond141#3 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &idev->mc_lock rcu_read_lock rcu_read_lock &____s->seqcount#10 irq_context: 0 (wq_completion)bond141#3 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &idev->mc_lock rcu_read_lock rcu_read_lock rcu_read_lock_bh rcu_read_lock rcu_read_lock_bh rcu_read_lock &br->multicast_lock irq_context: 0 (wq_completion)bond141#3 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &idev->mc_lock rcu_read_lock rcu_read_lock rcu_read_lock_bh rcu_read_lock rcu_read_lock_bh rcu_read_lock &br->multicast_lock &obj_hash[i].lock irq_context: 0 (wq_completion)bond141#3 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &idev->mc_lock rcu_read_lock rcu_read_lock rcu_read_lock_bh rcu_read_lock rcu_read_lock_bh rcu_read_lock &data->lock irq_context: 0 (wq_completion)bond141#3 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &idev->mc_lock rcu_read_lock rcu_read_lock rcu_read_lock_bh rcu_read_lock rcu_read_lock_bh rcu_read_lock &obj_hash[i].lock irq_context: 0 (wq_completion)bond141#3 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex.wait_lock irq_context: 0 (wq_completion)bond141#3 (work_completion)(&(&bond->mcast_work)->work) &p->pi_lock irq_context: 0 (wq_completion)bond127#3 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex fs_reclaim irq_context: 0 (wq_completion)bond127#3 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex fs_reclaim mmu_notifier_invalidate_range_start irq_context: 0 (wq_completion)bond127#3 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex rcu_read_lock &bond->stats_lock/1 irq_context: 0 (wq_completion)bond127#3 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex rcu_read_lock &ndev->lock irq_context: 0 (wq_completion)bond127#3 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex nl_table_lock irq_context: 0 (wq_completion)bond127#3 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &obj_hash[i].lock irq_context: 0 (wq_completion)bond127#3 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex nl_table_wait.lock irq_context: 0 (wq_completion)bond127#3 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex net_rwsem irq_context: 0 (wq_completion)bond127#3 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex net_rwsem &list->lock#2 irq_context: 0 (wq_completion)bond127#3 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &tn->lock irq_context: 0 (wq_completion)bond127#3 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &idev->mc_lock irq_context: 0 (wq_completion)bond127#3 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &idev->mc_lock fs_reclaim irq_context: 0 (wq_completion)bond127#3 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &idev->mc_lock fs_reclaim mmu_notifier_invalidate_range_start irq_context: 0 (wq_completion)bond127#3 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &idev->mc_lock &c->lock irq_context: 0 (wq_completion)bond127#3 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &idev->mc_lock rcu_read_lock &ndev->lock irq_context: 0 (wq_completion)bond127#3 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &idev->mc_lock rcu_read_lock &c->lock irq_context: 0 (wq_completion)bond127#3 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &idev->mc_lock rcu_read_lock &dir->lock#2 irq_context: 0 (wq_completion)bond127#3 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &idev->mc_lock rcu_read_lock &ul->lock#2 irq_context: 0 (wq_completion)bond127#3 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &idev->mc_lock rcu_read_lock rcu_read_lock &____s->seqcount#10 irq_context: 0 (wq_completion)bond127#3 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &idev->mc_lock rcu_read_lock rcu_read_lock rcu_read_lock_bh rcu_read_lock rcu_read_lock_bh rcu_read_lock &br->multicast_lock irq_context: 0 (wq_completion)bond127#3 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &idev->mc_lock rcu_read_lock rcu_read_lock rcu_read_lock_bh rcu_read_lock rcu_read_lock_bh rcu_read_lock &br->multicast_lock &obj_hash[i].lock irq_context: 0 (wq_completion)bond127#3 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &idev->mc_lock rcu_read_lock rcu_read_lock rcu_read_lock_bh rcu_read_lock rcu_read_lock_bh rcu_read_lock &br->multicast_lock &base->lock irq_context: 0 (wq_completion)bond127#3 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &idev->mc_lock rcu_read_lock rcu_read_lock rcu_read_lock_bh rcu_read_lock rcu_read_lock_bh rcu_read_lock &br->multicast_lock &base->lock &obj_hash[i].lock irq_context: 0 (wq_completion)bond127#3 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &idev->mc_lock rcu_read_lock rcu_read_lock rcu_read_lock_bh rcu_read_lock rcu_read_lock_bh rcu_read_lock &data->lock irq_context: 0 (wq_completion)bond127#3 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &idev->mc_lock rcu_read_lock rcu_read_lock rcu_read_lock_bh rcu_read_lock rcu_read_lock_bh rcu_read_lock &obj_hash[i].lock irq_context: 0 (wq_completion)bond127#3 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex.wait_lock irq_context: 0 (wq_completion)bond127#3 (work_completion)(&(&bond->mcast_work)->work) &p->pi_lock irq_context: 0 (wq_completion)bond45#2 irq_context: 0 (wq_completion)bond45#2 (work_completion)(&(&slave->notify_work)->work) irq_context: 0 (wq_completion)bond45#2 (work_completion)(&(&slave->notify_work)->work) &obj_hash[i].lock irq_context: 0 (wq_completion)bond45#2 (work_completion)(&(&slave->notify_work)->work) &base->lock irq_context: 0 (wq_completion)bond45#2 (work_completion)(&(&slave->notify_work)->work) &base->lock &obj_hash[i].lock irq_context: 0 (wq_completion)bond45#2 (work_completion)(&(&slave->notify_work)->work) &rq->__lock irq_context: 0 (wq_completion)bond45#2 (work_completion)(&(&slave->notify_work)->work) &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)bond45#2 (work_completion)(&(&bond->mcast_work)->work) irq_context: 0 (wq_completion)bond45#2 (work_completion)(&(&bond->mcast_work)->work) &obj_hash[i].lock irq_context: 0 (wq_completion)bond45#2 (work_completion)(&(&bond->mcast_work)->work) &base->lock irq_context: 0 (wq_completion)bond45#2 (work_completion)(&(&bond->mcast_work)->work) &base->lock &obj_hash[i].lock irq_context: 0 (wq_completion)bond45#2 (work_completion)(&(&bond->mcast_work)->work) &rq->__lock irq_context: 0 (wq_completion)bond45#2 (work_completion)(&(&bond->mcast_work)->work) &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)bond5 (work_completion)(&(&bond->mii_work)->work) fill_pool_map-wait-type-override pool_lock irq_context: 0 (wq_completion)bond158#2 irq_context: 0 (wq_completion)bond158#2 (work_completion)(&(&slave->notify_work)->work) irq_context: 0 (wq_completion)bond158#2 (work_completion)(&(&slave->notify_work)->work) &obj_hash[i].lock irq_context: 0 (wq_completion)bond158#2 (work_completion)(&(&slave->notify_work)->work) &base->lock irq_context: 0 (wq_completion)bond158#2 (work_completion)(&(&slave->notify_work)->work) &base->lock &obj_hash[i].lock irq_context: 0 (wq_completion)bond158#2 (work_completion)(&(&slave->notify_work)->work) &rq->__lock irq_context: 0 (wq_completion)bond158#2 (work_completion)(&(&slave->notify_work)->work) &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)bond43#2 (work_completion)(&(&slave->notify_work)->work) rtnl_mutex net_rwsem irq_context: 0 (wq_completion)bond43#2 (work_completion)(&(&slave->notify_work)->work) rtnl_mutex net_rwsem &list->lock#2 irq_context: 0 (wq_completion)bond43#2 (work_completion)(&(&slave->notify_work)->work) rtnl_mutex &tn->lock irq_context: 0 (wq_completion)bond43#2 (work_completion)(&(&slave->notify_work)->work) rtnl_mutex.wait_lock irq_context: 0 (wq_completion)bond43#2 (work_completion)(&(&slave->notify_work)->work) &p->pi_lock irq_context: 0 (wq_completion)bond158#2 (work_completion)(&(&bond->mcast_work)->work) irq_context: 0 (wq_completion)bond158#2 (work_completion)(&(&bond->mcast_work)->work) &obj_hash[i].lock irq_context: 0 (wq_completion)bond158#2 (work_completion)(&(&bond->mcast_work)->work) &base->lock irq_context: 0 (wq_completion)bond158#2 (work_completion)(&(&bond->mcast_work)->work) &base->lock &obj_hash[i].lock irq_context: 0 (wq_completion)bat_events (work_completion)(&(&bat_priv->mcast.work)->work) &base->lock fill_pool_map-wait-type-override pool_lock#2 irq_context: 0 (wq_completion)bat_events (work_completion)(&(&bat_priv->mcast.work)->work) &base->lock fill_pool_map-wait-type-override pool_lock irq_context: 0 (wq_completion)bond140#3 (work_completion)(&(&slave->notify_work)->work) &rq->__lock irq_context: 0 (wq_completion)bond140#3 (work_completion)(&(&slave->notify_work)->work) &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)bond142#3 irq_context: 0 (wq_completion)bond142#3 (work_completion)(&(&slave->notify_work)->work) irq_context: 0 (wq_completion)bond142#3 (work_completion)(&(&slave->notify_work)->work) &obj_hash[i].lock irq_context: 0 (wq_completion)bond142#3 (work_completion)(&(&slave->notify_work)->work) &base->lock irq_context: 0 (wq_completion)bond142#3 (work_completion)(&(&slave->notify_work)->work) &base->lock &obj_hash[i].lock irq_context: 0 (wq_completion)bond142#3 (work_completion)(&(&slave->notify_work)->work) &rq->__lock irq_context: 0 (wq_completion)bond142#3 (work_completion)(&(&bond->mcast_work)->work) irq_context: 0 (wq_completion)bond142#3 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex fs_reclaim irq_context: 0 (wq_completion)bond142#3 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex fs_reclaim mmu_notifier_invalidate_range_start irq_context: 0 (wq_completion)bond142#3 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &c->lock irq_context: 0 (wq_completion)bond142#3 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &rq->__lock irq_context: 0 (wq_completion)bond142#3 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)bond142#3 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex rcu_read_lock &bond->stats_lock/1 irq_context: 0 (wq_completion)bond142#3 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex rcu_read_lock &ndev->lock irq_context: 0 (wq_completion)bond142#3 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex nl_table_lock irq_context: 0 (wq_completion)bond142#3 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &obj_hash[i].lock irq_context: 0 (wq_completion)bond142#3 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex nl_table_wait.lock irq_context: 0 (wq_completion)bond142#3 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex net_rwsem irq_context: 0 (wq_completion)bond142#3 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex net_rwsem &list->lock#2 irq_context: 0 (wq_completion)bond142#3 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &tn->lock irq_context: 0 (wq_completion)bond142#3 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &idev->mc_lock irq_context: 0 (wq_completion)bond142#3 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex.wait_lock irq_context: 0 (wq_completion)bond142#3 (work_completion)(&(&bond->mcast_work)->work) &rq->__lock irq_context: 0 (wq_completion)bond142#3 (work_completion)(&(&bond->mcast_work)->work) &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)bond142#3 (work_completion)(&(&bond->mcast_work)->work) &p->pi_lock irq_context: 0 (wq_completion)bond142#3 (work_completion)(&(&bond->mcast_work)->work) &p->pi_lock &rq->__lock irq_context: 0 (wq_completion)bond142#3 (work_completion)(&(&bond->mcast_work)->work) &p->pi_lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)bond142#3 &rq->__lock irq_context: 0 (wq_completion)bond142#3 &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)bond139#3 (work_completion)(&(&slave->notify_work)->work) &rq->__lock irq_context: 0 (wq_completion)bond139#3 (work_completion)(&(&slave->notify_work)->work) &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)bond171 irq_context: 0 (wq_completion)bond171 (work_completion)(&(&slave->notify_work)->work) irq_context: 0 (wq_completion)bond171 (work_completion)(&(&slave->notify_work)->work) &obj_hash[i].lock irq_context: 0 (wq_completion)bond171 (work_completion)(&(&slave->notify_work)->work) &base->lock irq_context: 0 (wq_completion)bond171 (work_completion)(&(&slave->notify_work)->work) &base->lock &obj_hash[i].lock irq_context: 0 (wq_completion)bond171 (work_completion)(&(&slave->notify_work)->work) &rq->__lock irq_context: 0 (wq_completion)bond171 (work_completion)(&(&slave->notify_work)->work) &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)bond171 (work_completion)(&(&bond->mcast_work)->work) irq_context: 0 (wq_completion)bond171 (work_completion)(&(&bond->mcast_work)->work) &obj_hash[i].lock irq_context: 0 (wq_completion)bond171 (work_completion)(&(&bond->mcast_work)->work) &base->lock irq_context: 0 (wq_completion)bond171 (work_completion)(&(&bond->mcast_work)->work) &base->lock &obj_hash[i].lock irq_context: 0 (wq_completion)bond171 (work_completion)(&(&bond->mcast_work)->work) &rq->__lock irq_context: 0 (wq_completion)bond171 (work_completion)(&(&bond->mcast_work)->work) &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)bond46#2 irq_context: 0 (wq_completion)bond46#2 (work_completion)(&(&slave->notify_work)->work) irq_context: 0 (wq_completion)bond46#2 (work_completion)(&(&slave->notify_work)->work) &obj_hash[i].lock irq_context: 0 (wq_completion)bond46#2 (work_completion)(&(&slave->notify_work)->work) &base->lock irq_context: 0 (wq_completion)bond46#2 (work_completion)(&(&slave->notify_work)->work) &base->lock &obj_hash[i].lock irq_context: 0 (wq_completion)bond46#2 (work_completion)(&(&slave->notify_work)->work) &rq->__lock irq_context: 0 (wq_completion)bond46#2 (work_completion)(&(&slave->notify_work)->work) &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)bond46#2 (work_completion)(&(&bond->mcast_work)->work) irq_context: 0 (wq_completion)bond46#2 (work_completion)(&(&bond->mcast_work)->work) &obj_hash[i].lock irq_context: 0 (wq_completion)bond46#2 (work_completion)(&(&bond->mcast_work)->work) &base->lock irq_context: 0 (wq_completion)bond46#2 (work_completion)(&(&bond->mcast_work)->work) &base->lock &obj_hash[i].lock irq_context: 0 (wq_completion)bond46#2 (work_completion)(&(&bond->mcast_work)->work) &rq->__lock irq_context: 0 (wq_completion)bond46#2 (work_completion)(&(&bond->mcast_work)->work) &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)bond159#2 irq_context: 0 (wq_completion)bond159#2 (work_completion)(&(&slave->notify_work)->work) irq_context: 0 (wq_completion)bond159#2 (work_completion)(&(&slave->notify_work)->work) &obj_hash[i].lock irq_context: 0 (wq_completion)bond159#2 (work_completion)(&(&slave->notify_work)->work) &base->lock irq_context: 0 (wq_completion)bond159#2 (work_completion)(&(&slave->notify_work)->work) &base->lock &obj_hash[i].lock irq_context: 0 (wq_completion)bond159#2 (work_completion)(&(&slave->notify_work)->work) &rq->__lock irq_context: 0 (wq_completion)bond159#2 (work_completion)(&(&slave->notify_work)->work) &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 rtnl_mutex net_rwsem &cfs_rq->removed.lock irq_context: 0 rtnl_mutex net_rwsem &obj_hash[i].lock irq_context: 0 rtnl_mutex net_rwsem pool_lock#2 irq_context: 0 (wq_completion)bond159#2 (work_completion)(&(&bond->mcast_work)->work) irq_context: 0 (wq_completion)bond159#2 (work_completion)(&(&bond->mcast_work)->work) &rq->__lock irq_context: 0 (wq_completion)bond159#2 (work_completion)(&(&bond->mcast_work)->work) &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)bond138#2 (work_completion)(&(&slave->notify_work)->work) &rq->__lock irq_context: 0 (wq_completion)bond138#2 (work_completion)(&(&slave->notify_work)->work) &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)bond45#2 (work_completion)(&(&slave->notify_work)->work) fill_pool_map-wait-type-override pool_lock irq_context: 0 (wq_completion)bond159#2 (work_completion)(&(&bond->mcast_work)->work) &obj_hash[i].lock irq_context: 0 (wq_completion)bond159#2 (work_completion)(&(&bond->mcast_work)->work) &base->lock irq_context: 0 (wq_completion)bond159#2 (work_completion)(&(&bond->mcast_work)->work) &base->lock &obj_hash[i].lock irq_context: 0 (wq_completion)bond20#3 (work_completion)(&(&slave->notify_work)->work) rtnl_mutex net_rwsem irq_context: 0 (wq_completion)bond20#3 (work_completion)(&(&slave->notify_work)->work) rtnl_mutex net_rwsem &list->lock#2 irq_context: 0 (wq_completion)bond20#3 (work_completion)(&(&slave->notify_work)->work) rtnl_mutex &tn->lock irq_context: 0 (wq_completion)bond20#3 (work_completion)(&(&slave->notify_work)->work) rtnl_mutex.wait_lock irq_context: 0 (wq_completion)bond20#3 (work_completion)(&(&slave->notify_work)->work) &p->pi_lock irq_context: 0 (wq_completion)bond140#3 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex fs_reclaim irq_context: 0 (wq_completion)bond140#3 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex fs_reclaim mmu_notifier_invalidate_range_start irq_context: 0 (wq_completion)bond140#3 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &c->lock irq_context: 0 (wq_completion)bond140#3 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &n->list_lock irq_context: 0 (wq_completion)bond140#3 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &n->list_lock &c->lock irq_context: 0 (wq_completion)bond140#3 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex rcu_read_lock &bond->stats_lock/1 irq_context: 0 (wq_completion)bond140#3 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex rcu_read_lock &ndev->lock irq_context: 0 (wq_completion)bond140#3 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex nl_table_lock irq_context: 0 (wq_completion)bond140#3 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &obj_hash[i].lock irq_context: 0 (wq_completion)bond140#3 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex nl_table_wait.lock irq_context: 0 (wq_completion)bond140#3 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex net_rwsem irq_context: 0 (wq_completion)bond140#3 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex net_rwsem &list->lock#2 irq_context: 0 (wq_completion)bond140#3 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &tn->lock irq_context: 0 (wq_completion)bond140#3 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &idev->mc_lock irq_context: 0 (wq_completion)bond140#3 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &idev->mc_lock fs_reclaim irq_context: 0 (wq_completion)bond140#3 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &idev->mc_lock fs_reclaim mmu_notifier_invalidate_range_start irq_context: 0 (wq_completion)bond140#3 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &idev->mc_lock rcu_read_lock &ndev->lock irq_context: 0 (wq_completion)bond140#3 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &idev->mc_lock rcu_read_lock &dir->lock#2 irq_context: 0 (wq_completion)bond140#3 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &idev->mc_lock rcu_read_lock &ul->lock#2 irq_context: 0 (wq_completion)bond140#3 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &idev->mc_lock rcu_read_lock rcu_read_lock &____s->seqcount#10 irq_context: 0 (wq_completion)bond140#3 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &idev->mc_lock rcu_read_lock rcu_read_lock rcu_read_lock_bh rcu_read_lock rcu_read_lock_bh rcu_read_lock &br->multicast_lock irq_context: 0 (wq_completion)bond140#3 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &idev->mc_lock rcu_read_lock rcu_read_lock rcu_read_lock_bh rcu_read_lock rcu_read_lock_bh rcu_read_lock &br->multicast_lock &obj_hash[i].lock irq_context: 0 (wq_completion)bond140#3 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &idev->mc_lock rcu_read_lock rcu_read_lock rcu_read_lock_bh rcu_read_lock rcu_read_lock_bh rcu_read_lock &br->multicast_lock &base->lock irq_context: 0 (wq_completion)bond140#3 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &idev->mc_lock rcu_read_lock rcu_read_lock rcu_read_lock_bh rcu_read_lock rcu_read_lock_bh rcu_read_lock &data->lock irq_context: 0 (wq_completion)bond140#3 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &idev->mc_lock rcu_read_lock rcu_read_lock rcu_read_lock_bh rcu_read_lock rcu_read_lock_bh rcu_read_lock &obj_hash[i].lock irq_context: 0 (wq_completion)bond140#3 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex.wait_lock irq_context: 0 (wq_completion)bond140#3 (work_completion)(&(&bond->mcast_work)->work) &p->pi_lock irq_context: 0 (wq_completion)bond164 (work_completion)(&(&slave->notify_work)->work) rtnl_mutex net_rwsem irq_context: 0 (wq_completion)bond164 (work_completion)(&(&slave->notify_work)->work) rtnl_mutex net_rwsem &list->lock#2 irq_context: 0 (wq_completion)bond164 (work_completion)(&(&slave->notify_work)->work) rtnl_mutex &tn->lock irq_context: 0 (wq_completion)bond164 (work_completion)(&(&slave->notify_work)->work) rtnl_mutex.wait_lock irq_context: 0 (wq_completion)bond164 (work_completion)(&(&slave->notify_work)->work) &p->pi_lock irq_context: 0 (wq_completion)bond143#3 irq_context: 0 (wq_completion)bond143#3 (work_completion)(&(&slave->notify_work)->work) irq_context: 0 (wq_completion)bond143#3 (work_completion)(&(&slave->notify_work)->work) &obj_hash[i].lock irq_context: 0 (wq_completion)bond143#3 (work_completion)(&(&slave->notify_work)->work) &base->lock irq_context: 0 (wq_completion)bond143#3 (work_completion)(&(&slave->notify_work)->work) &base->lock &obj_hash[i].lock irq_context: 0 (wq_completion)bond143#3 (work_completion)(&(&slave->notify_work)->work) &rq->__lock irq_context: 0 (wq_completion)bond143#3 (work_completion)(&(&slave->notify_work)->work) &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)bond143#3 (work_completion)(&(&bond->mcast_work)->work) irq_context: 0 (wq_completion)bond143#3 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex fs_reclaim irq_context: 0 (wq_completion)bond143#3 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex fs_reclaim mmu_notifier_invalidate_range_start irq_context: 0 (wq_completion)bond143#3 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &c->lock irq_context: 0 (wq_completion)bond143#3 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &rq->__lock irq_context: 0 (wq_completion)bond143#3 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)bond143#3 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex rcu_read_lock &bond->stats_lock/1 irq_context: 0 (wq_completion)bond143#3 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex rcu_read_lock &ndev->lock irq_context: 0 (wq_completion)bond143#3 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex nl_table_lock irq_context: 0 (wq_completion)bond143#3 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &obj_hash[i].lock irq_context: 0 (wq_completion)bond143#3 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex nl_table_wait.lock irq_context: 0 (wq_completion)bond143#3 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex net_rwsem irq_context: 0 (wq_completion)bond143#3 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex net_rwsem &list->lock#2 irq_context: 0 (wq_completion)bond143#3 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &tn->lock irq_context: 0 (wq_completion)bond143#3 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &idev->mc_lock irq_context: 0 (wq_completion)bond143#3 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex rcu_read_lock rcu_node_0 irq_context: 0 (wq_completion)bond143#3 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex rcu_read_lock &rq->__lock irq_context: 0 (wq_completion)bond143#3 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex rcu_read_lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)bond143#3 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex.wait_lock irq_context: 0 (wq_completion)bond143#3 (work_completion)(&(&bond->mcast_work)->work) &p->pi_lock irq_context: 0 (wq_completion)bond165 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex fs_reclaim irq_context: 0 (wq_completion)bond165 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex fs_reclaim mmu_notifier_invalidate_range_start irq_context: 0 (wq_completion)bond165 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex rcu_read_lock &bond->stats_lock/1 irq_context: 0 (wq_completion)bond165 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex rcu_read_lock &ndev->lock irq_context: 0 (wq_completion)bond165 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex nl_table_lock irq_context: 0 (wq_completion)bond165 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &obj_hash[i].lock irq_context: 0 (wq_completion)bond165 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex nl_table_wait.lock irq_context: 0 (wq_completion)bond165 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex net_rwsem irq_context: 0 (wq_completion)bond165 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex net_rwsem &list->lock#2 irq_context: 0 (wq_completion)bond165 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex net_rwsem &rq->__lock irq_context: 0 (wq_completion)bond165 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex net_rwsem &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)bond165 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &tn->lock irq_context: 0 (wq_completion)bond165 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &idev->mc_lock irq_context: 0 (wq_completion)bond165 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &idev->mc_lock fs_reclaim irq_context: 0 (wq_completion)bond165 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &idev->mc_lock fs_reclaim mmu_notifier_invalidate_range_start irq_context: 0 (wq_completion)bond165 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &idev->mc_lock rcu_read_lock &ndev->lock irq_context: 0 (wq_completion)bond165 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &idev->mc_lock rcu_read_lock &c->lock irq_context: 0 (wq_completion)bond165 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &idev->mc_lock rcu_read_lock &dir->lock#2 irq_context: 0 (wq_completion)bond165 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &idev->mc_lock rcu_read_lock &ul->lock#2 irq_context: 0 (wq_completion)bond165 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &idev->mc_lock rcu_read_lock rcu_read_lock &____s->seqcount#10 irq_context: 0 (wq_completion)bond165 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &idev->mc_lock rcu_read_lock rcu_read_lock rcu_read_lock_bh rcu_read_lock rcu_read_lock_bh rcu_read_lock &br->multicast_lock irq_context: 0 (wq_completion)bond165 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &idev->mc_lock rcu_read_lock rcu_read_lock rcu_read_lock_bh rcu_read_lock rcu_read_lock_bh rcu_read_lock &br->multicast_lock &obj_hash[i].lock irq_context: 0 (wq_completion)bond165 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &idev->mc_lock rcu_read_lock rcu_read_lock rcu_read_lock_bh rcu_read_lock rcu_read_lock_bh rcu_read_lock &br->multicast_lock &base->lock irq_context: 0 (wq_completion)bond165 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &idev->mc_lock rcu_read_lock rcu_read_lock rcu_read_lock_bh rcu_read_lock rcu_read_lock_bh rcu_read_lock &data->lock irq_context: 0 (wq_completion)bond165 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &idev->mc_lock rcu_read_lock rcu_read_lock rcu_read_lock_bh rcu_read_lock rcu_read_lock_bh rcu_read_lock &obj_hash[i].lock irq_context: 0 (wq_completion)bond165 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &idev->mc_lock rcu_read_lock rcu_read_lock rcu_node_0 irq_context: 0 (wq_completion)bond165 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &idev->mc_lock rcu_read_lock rcu_read_lock &rq->__lock irq_context: 0 (wq_completion)bond165 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &idev->mc_lock rcu_read_lock rcu_read_lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)bond165 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &idev->mc_lock rcu_read_lock rcu_node_0 irq_context: 0 (wq_completion)bond165 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex.wait_lock irq_context: 0 (wq_completion)bond165 (work_completion)(&(&bond->mcast_work)->work) &p->pi_lock irq_context: 0 (wq_completion)bond165 (work_completion)(&(&slave->notify_work)->work) rtnl_mutex net_rwsem irq_context: 0 (wq_completion)bond165 (work_completion)(&(&slave->notify_work)->work) rtnl_mutex net_rwsem &rq->__lock irq_context: 0 (wq_completion)bond165 (work_completion)(&(&slave->notify_work)->work) rtnl_mutex net_rwsem &list->lock#2 irq_context: 0 (wq_completion)bond165 (work_completion)(&(&slave->notify_work)->work) rtnl_mutex &tn->lock irq_context: 0 (wq_completion)bond165 (work_completion)(&(&slave->notify_work)->work) rtnl_mutex.wait_lock irq_context: 0 (wq_completion)bond165 (work_completion)(&(&slave->notify_work)->work) &p->pi_lock irq_context: 0 (wq_completion)bond165 (work_completion)(&(&slave->notify_work)->work) &p->pi_lock &rq->__lock irq_context: 0 (wq_completion)bond165 (work_completion)(&(&slave->notify_work)->work) &p->pi_lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)bond168 (work_completion)(&(&slave->notify_work)->work) &rq->__lock irq_context: 0 (wq_completion)bond168 (work_completion)(&(&slave->notify_work)->work) &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)bond132 &rq->__lock irq_context: 0 (wq_completion)bond132 &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)bond138#2 (work_completion)(&(&bond->mcast_work)->work) &rq->__lock irq_context: 0 (wq_completion)bond138#2 (work_completion)(&(&bond->mcast_work)->work) &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)bond154#2 (work_completion)(&(&slave->notify_work)->work) rtnl_mutex net_rwsem irq_context: 0 (wq_completion)bond154#2 (work_completion)(&(&slave->notify_work)->work) rtnl_mutex net_rwsem &list->lock#2 irq_context: 0 (wq_completion)bond154#2 (work_completion)(&(&slave->notify_work)->work) rtnl_mutex &tn->lock irq_context: 0 (wq_completion)bond154#2 (work_completion)(&(&slave->notify_work)->work) rtnl_mutex.wait_lock irq_context: 0 (wq_completion)bond154#2 (work_completion)(&(&slave->notify_work)->work) &p->pi_lock irq_context: 0 (wq_completion)bond154 (work_completion)(&(&slave->notify_work)->work) rtnl_mutex net_rwsem irq_context: 0 (wq_completion)bond154 (work_completion)(&(&slave->notify_work)->work) rtnl_mutex net_rwsem &list->lock#2 irq_context: 0 (wq_completion)bond154 (work_completion)(&(&slave->notify_work)->work) rtnl_mutex &tn->lock irq_context: 0 (wq_completion)bond154 (work_completion)(&(&slave->notify_work)->work) rtnl_mutex.wait_lock irq_context: 0 (wq_completion)bond154 (work_completion)(&(&slave->notify_work)->work) &p->pi_lock irq_context: 0 (wq_completion)bond147#2 (work_completion)(&(&slave->notify_work)->work) rtnl_mutex net_rwsem irq_context: 0 (wq_completion)bond147#2 (work_completion)(&(&slave->notify_work)->work) rtnl_mutex net_rwsem &list->lock#2 irq_context: 0 (wq_completion)bond147#2 (work_completion)(&(&slave->notify_work)->work) rtnl_mutex &tn->lock irq_context: 0 (wq_completion)bond147#2 (work_completion)(&(&slave->notify_work)->work) rtnl_mutex.wait_lock irq_context: 0 (wq_completion)bond147#2 (work_completion)(&(&slave->notify_work)->work) &p->pi_lock irq_context: 0 (wq_completion)bond158#2 (work_completion)(&(&slave->notify_work)->work) rtnl_mutex net_rwsem irq_context: 0 (wq_completion)bond158#2 (work_completion)(&(&slave->notify_work)->work) rtnl_mutex net_rwsem &list->lock#2 irq_context: 0 (wq_completion)bond158#2 (work_completion)(&(&slave->notify_work)->work) rtnl_mutex &tn->lock irq_context: 0 (wq_completion)bond158#2 (work_completion)(&(&slave->notify_work)->work) rtnl_mutex.wait_lock irq_context: 0 (wq_completion)bond158#2 (work_completion)(&(&slave->notify_work)->work) &p->pi_lock irq_context: 0 (wq_completion)bond157 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex fs_reclaim irq_context: 0 (wq_completion)bond157 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex fs_reclaim mmu_notifier_invalidate_range_start irq_context: 0 (wq_completion)bond157 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &c->lock irq_context: 0 (wq_completion)bond157 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex rcu_read_lock &bond->stats_lock/1 irq_context: 0 (wq_completion)bond157 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex rcu_read_lock &ndev->lock irq_context: 0 (wq_completion)bond157 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex nl_table_lock irq_context: 0 (wq_completion)bond157 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &obj_hash[i].lock irq_context: 0 (wq_completion)bond157 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex nl_table_wait.lock irq_context: 0 (wq_completion)bond157 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex net_rwsem irq_context: 0 (wq_completion)bond157 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex net_rwsem &list->lock#2 irq_context: 0 (wq_completion)bond157 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &tn->lock irq_context: 0 (wq_completion)bond157 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &idev->mc_lock irq_context: 0 (wq_completion)bond157 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &idev->mc_lock fs_reclaim irq_context: 0 (wq_completion)bond157 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &idev->mc_lock fs_reclaim mmu_notifier_invalidate_range_start irq_context: 0 (wq_completion)bond157 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &idev->mc_lock &c->lock irq_context: 0 (wq_completion)bond157 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &idev->mc_lock rcu_read_lock &ndev->lock irq_context: 0 (wq_completion)bond157 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &idev->mc_lock rcu_read_lock &dir->lock#2 irq_context: 0 (wq_completion)bond157 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &idev->mc_lock rcu_read_lock &ul->lock#2 irq_context: 0 (wq_completion)bond157 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &idev->mc_lock rcu_read_lock rcu_read_lock &____s->seqcount#10 irq_context: 0 (wq_completion)bond157 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &idev->mc_lock rcu_read_lock rcu_read_lock rcu_read_lock_bh rcu_read_lock rcu_read_lock_bh rcu_read_lock &br->multicast_lock irq_context: 0 (wq_completion)bond157 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &idev->mc_lock rcu_read_lock rcu_read_lock rcu_read_lock_bh rcu_read_lock rcu_read_lock_bh rcu_read_lock &br->multicast_lock &obj_hash[i].lock irq_context: 0 (wq_completion)bond157 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &idev->mc_lock rcu_read_lock rcu_read_lock rcu_read_lock_bh rcu_read_lock rcu_read_lock_bh rcu_read_lock &br->multicast_lock &base->lock irq_context: 0 (wq_completion)bond157 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &idev->mc_lock rcu_read_lock rcu_read_lock rcu_read_lock_bh rcu_read_lock rcu_read_lock_bh rcu_read_lock &br->multicast_lock &base->lock &obj_hash[i].lock irq_context: 0 (wq_completion)bond157 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &idev->mc_lock rcu_read_lock rcu_read_lock rcu_read_lock_bh rcu_read_lock rcu_read_lock_bh rcu_read_lock &data->lock irq_context: 0 (wq_completion)bond157 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &idev->mc_lock rcu_read_lock rcu_read_lock rcu_read_lock_bh rcu_read_lock rcu_read_lock_bh rcu_read_lock &obj_hash[i].lock irq_context: 0 (wq_completion)bond157 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &idev->mc_lock rcu_read_lock rcu_read_lock rcu_node_0 irq_context: 0 (wq_completion)bond157 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &idev->mc_lock rcu_read_lock rcu_read_lock &rq->__lock irq_context: 0 (wq_completion)bond157 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &idev->mc_lock rcu_read_lock rcu_read_lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)bond87 (work_completion)(&(&bond->mcast_work)->work) &rq->__lock irq_context: 0 (wq_completion)bond87 (work_completion)(&(&bond->mcast_work)->work) &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)bond157 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &idev->mc_lock rcu_read_lock rcu_node_0 irq_context: 0 (wq_completion)bond157 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &idev->mc_lock rcu_read_lock &rq->__lock irq_context: 0 (wq_completion)bond157 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &idev->mc_lock rcu_read_lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)bond157 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex.wait_lock irq_context: 0 (wq_completion)bond157 (work_completion)(&(&bond->mcast_work)->work) &p->pi_lock irq_context: 0 (wq_completion)bond157 (work_completion)(&(&bond->mcast_work)->work) &p->pi_lock &cfs_rq->removed.lock irq_context: 0 (wq_completion)bond157 (work_completion)(&(&bond->mcast_work)->work) &p->pi_lock &rq->__lock irq_context: 0 (wq_completion)bond157 (work_completion)(&(&bond->mcast_work)->work) &p->pi_lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)bond172 irq_context: 0 (wq_completion)bond172 &rq->__lock irq_context: 0 (wq_completion)bond172 &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)bond172 (work_completion)(&(&slave->notify_work)->work) irq_context: 0 (wq_completion)bond172 (work_completion)(&(&slave->notify_work)->work) &obj_hash[i].lock irq_context: 0 (wq_completion)bond172 (work_completion)(&(&slave->notify_work)->work) &base->lock irq_context: 0 (wq_completion)bond172 (work_completion)(&(&slave->notify_work)->work) &base->lock &obj_hash[i].lock irq_context: 0 (wq_completion)bond172 (work_completion)(&(&slave->notify_work)->work) &rq->__lock irq_context: 0 (wq_completion)bond172 (work_completion)(&(&slave->notify_work)->work) &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)bond172 (work_completion)(&(&bond->mcast_work)->work) irq_context: 0 (wq_completion)bond172 (work_completion)(&(&bond->mcast_work)->work) &obj_hash[i].lock irq_context: 0 (wq_completion)bond172 (work_completion)(&(&bond->mcast_work)->work) &base->lock irq_context: 0 (wq_completion)bond172 (work_completion)(&(&bond->mcast_work)->work) &base->lock &obj_hash[i].lock irq_context: 0 (wq_completion)bond152 &rq->__lock irq_context: 0 (wq_completion)bond152 &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)bond169 (work_completion)(&(&slave->notify_work)->work) &rq->__lock irq_context: 0 (wq_completion)bond169 (work_completion)(&(&slave->notify_work)->work) &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)bond160#2 irq_context: 0 (wq_completion)bond160#2 (work_completion)(&(&slave->notify_work)->work) irq_context: 0 (wq_completion)bond160#2 (work_completion)(&(&slave->notify_work)->work) &obj_hash[i].lock irq_context: 0 (wq_completion)bond160#2 (work_completion)(&(&slave->notify_work)->work) &base->lock irq_context: 0 (wq_completion)bond160#2 (work_completion)(&(&slave->notify_work)->work) &base->lock &obj_hash[i].lock irq_context: 0 (wq_completion)bond128 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex fs_reclaim irq_context: 0 (wq_completion)bond128 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex fs_reclaim mmu_notifier_invalidate_range_start irq_context: 0 (wq_completion)bond128 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex fs_reclaim mmu_notifier_invalidate_range_start &rq->__lock irq_context: 0 (wq_completion)bond128 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex fs_reclaim mmu_notifier_invalidate_range_start &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)bond128 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &c->lock irq_context: 0 (wq_completion)bond128 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex rcu_read_lock &bond->stats_lock/1 irq_context: 0 (wq_completion)bond128 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex rcu_read_lock &ndev->lock irq_context: 0 (wq_completion)bond128 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex nl_table_lock irq_context: 0 (wq_completion)bond128 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &obj_hash[i].lock irq_context: 0 (wq_completion)bond128 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex nl_table_wait.lock irq_context: 0 (wq_completion)bond128 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex net_rwsem irq_context: 0 (wq_completion)bond128 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex net_rwsem &list->lock#2 irq_context: 0 (wq_completion)bond128 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &tn->lock irq_context: 0 (wq_completion)bond128 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &idev->mc_lock irq_context: 0 (wq_completion)bond128 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &idev->mc_lock &rq->__lock irq_context: 0 (wq_completion)bond128 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &idev->mc_lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)bond128 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &idev->mc_lock fs_reclaim irq_context: 0 (wq_completion)bond128 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &idev->mc_lock fs_reclaim mmu_notifier_invalidate_range_start irq_context: 0 (wq_completion)bond128 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &idev->mc_lock &c->lock irq_context: 0 (wq_completion)bond128 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &idev->mc_lock rcu_read_lock &ndev->lock irq_context: 0 (wq_completion)bond128 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &idev->mc_lock rcu_read_lock &dir->lock#2 irq_context: 0 (wq_completion)bond128 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &idev->mc_lock rcu_read_lock &ul->lock#2 irq_context: 0 (wq_completion)bond128 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &idev->mc_lock rcu_read_lock rcu_read_lock &____s->seqcount#10 irq_context: 0 (wq_completion)bond128 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &idev->mc_lock rcu_read_lock rcu_read_lock rcu_read_lock_bh rcu_read_lock rcu_read_lock_bh rcu_read_lock &br->multicast_lock irq_context: 0 (wq_completion)bond128 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &idev->mc_lock rcu_read_lock rcu_read_lock rcu_read_lock_bh rcu_read_lock rcu_read_lock_bh rcu_read_lock &br->multicast_lock &obj_hash[i].lock irq_context: 0 (wq_completion)bond128 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &idev->mc_lock rcu_read_lock rcu_read_lock rcu_read_lock_bh rcu_read_lock rcu_read_lock_bh rcu_read_lock &br->multicast_lock &base->lock irq_context: 0 (wq_completion)bond128 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &idev->mc_lock rcu_read_lock rcu_read_lock rcu_read_lock_bh rcu_read_lock rcu_read_lock_bh rcu_read_lock &br->multicast_lock &base->lock &obj_hash[i].lock irq_context: 0 (wq_completion)bond128 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &idev->mc_lock rcu_read_lock rcu_read_lock rcu_read_lock_bh rcu_read_lock rcu_read_lock_bh rcu_read_lock &data->lock irq_context: 0 (wq_completion)bond128 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &idev->mc_lock rcu_read_lock rcu_read_lock rcu_read_lock_bh rcu_read_lock rcu_read_lock_bh rcu_read_lock &obj_hash[i].lock irq_context: 0 (wq_completion)bond128 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &idev->mc_lock rcu_read_lock rcu_read_lock rcu_node_0 irq_context: 0 (wq_completion)bond128 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &idev->mc_lock rcu_read_lock rcu_read_lock &rq->__lock irq_context: 0 (wq_completion)bond128 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &idev->mc_lock rcu_read_lock rcu_read_lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)bond128 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &idev->mc_lock rcu_read_lock rcu_node_0 irq_context: 0 (wq_completion)bond128 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex.wait_lock irq_context: 0 (wq_completion)bond128 (work_completion)(&(&bond->mcast_work)->work) &p->pi_lock irq_context: 0 (wq_completion)bond128 (work_completion)(&(&bond->mcast_work)->work) &p->pi_lock &rq->__lock irq_context: 0 (wq_completion)bond128 (work_completion)(&(&bond->mcast_work)->work) &p->pi_lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)bond160#2 (work_completion)(&(&bond->mcast_work)->work) irq_context: 0 (wq_completion)bond160#2 (work_completion)(&(&bond->mcast_work)->work) &obj_hash[i].lock irq_context: 0 (wq_completion)bond160#2 (work_completion)(&(&bond->mcast_work)->work) &base->lock irq_context: 0 (wq_completion)bond160#2 (work_completion)(&(&bond->mcast_work)->work) &base->lock &obj_hash[i].lock irq_context: 0 (wq_completion)bond160#2 (work_completion)(&(&bond->mcast_work)->work) &rq->__lock irq_context: 0 (wq_completion)bond160#2 (work_completion)(&(&bond->mcast_work)->work) &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)bond144#3 irq_context: 0 (wq_completion)bond144#3 (work_completion)(&(&slave->notify_work)->work) irq_context: 0 (wq_completion)bond144#3 (work_completion)(&(&slave->notify_work)->work) &obj_hash[i].lock irq_context: 0 (wq_completion)bond144#3 (work_completion)(&(&slave->notify_work)->work) &base->lock irq_context: 0 (wq_completion)bond144#3 (work_completion)(&(&slave->notify_work)->work) &base->lock &obj_hash[i].lock irq_context: 0 (wq_completion)bond144#3 (work_completion)(&(&slave->notify_work)->work) &rq->__lock irq_context: 0 (wq_completion)bond144#3 (work_completion)(&(&slave->notify_work)->work) &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)bond134 &rq->__lock irq_context: 0 (wq_completion)bond134 &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)bond144#3 (work_completion)(&(&bond->mcast_work)->work) irq_context: 0 (wq_completion)bond144#3 (work_completion)(&(&bond->mcast_work)->work) &obj_hash[i].lock irq_context: 0 (wq_completion)bond144#3 (work_completion)(&(&bond->mcast_work)->work) &base->lock irq_context: 0 (wq_completion)bond144#3 (work_completion)(&(&bond->mcast_work)->work) &base->lock &obj_hash[i].lock irq_context: 0 sk_lock-AF_INET6 &base->lock fill_pool_map-wait-type-override pool_lock#2 irq_context: 0 sk_lock-AF_INET6 &base->lock fill_pool_map-wait-type-override pool_lock irq_context: 0 (wq_completion)bond173 irq_context: 0 (wq_completion)bond173 &rq->__lock irq_context: 0 (wq_completion)bond173 &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)bond173 (work_completion)(&(&slave->notify_work)->work) irq_context: 0 (wq_completion)bond173 (work_completion)(&(&slave->notify_work)->work) &obj_hash[i].lock irq_context: 0 (wq_completion)bond173 (work_completion)(&(&slave->notify_work)->work) &base->lock irq_context: 0 (wq_completion)bond173 (work_completion)(&(&slave->notify_work)->work) &base->lock &obj_hash[i].lock irq_context: 0 (wq_completion)bond173 (work_completion)(&(&bond->mcast_work)->work) irq_context: 0 (wq_completion)bond173 (work_completion)(&(&bond->mcast_work)->work) &obj_hash[i].lock irq_context: 0 (wq_completion)bond173 (work_completion)(&(&bond->mcast_work)->work) &base->lock irq_context: 0 (wq_completion)bond173 (work_completion)(&(&bond->mcast_work)->work) &base->lock &obj_hash[i].lock irq_context: 0 (wq_completion)bond173 (work_completion)(&(&bond->mcast_work)->work) &rq->__lock irq_context: 0 (wq_completion)bond173 (work_completion)(&(&bond->mcast_work)->work) &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)bond170 (work_completion)(&(&bond->mcast_work)->work) &rq->__lock irq_context: 0 (wq_completion)bond170 (work_completion)(&(&bond->mcast_work)->work) &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)bond47#2 irq_context: 0 (wq_completion)bond47#2 (work_completion)(&(&slave->notify_work)->work) irq_context: 0 (wq_completion)bond47#2 (work_completion)(&(&slave->notify_work)->work) &obj_hash[i].lock irq_context: 0 (wq_completion)bond47#2 (work_completion)(&(&slave->notify_work)->work) &base->lock irq_context: 0 (wq_completion)bond47#2 (work_completion)(&(&slave->notify_work)->work) &base->lock &obj_hash[i].lock irq_context: 0 (wq_completion)bond47#2 (work_completion)(&(&slave->notify_work)->work) &rq->__lock irq_context: 0 (wq_completion)bond47#2 (work_completion)(&(&slave->notify_work)->work) &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)bond47#2 (work_completion)(&(&bond->mcast_work)->work) irq_context: 0 (wq_completion)bond47#2 (work_completion)(&(&bond->mcast_work)->work) &obj_hash[i].lock irq_context: 0 (wq_completion)bond47#2 (work_completion)(&(&bond->mcast_work)->work) &base->lock irq_context: 0 (wq_completion)bond47#2 (work_completion)(&(&bond->mcast_work)->work) &base->lock &obj_hash[i].lock irq_context: 0 (wq_completion)bond47#2 (work_completion)(&(&bond->mcast_work)->work) &rq->__lock irq_context: 0 (wq_completion)bond47#2 (work_completion)(&(&bond->mcast_work)->work) &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)bond124#3 (work_completion)(&(&bond->mcast_work)->work) &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)bond154#2 &rq->__lock irq_context: 0 (wq_completion)bond161#2 irq_context: 0 (wq_completion)bond161#2 (work_completion)(&(&slave->notify_work)->work) irq_context: 0 (wq_completion)bond161#2 (work_completion)(&(&slave->notify_work)->work) &rq->__lock irq_context: 0 (wq_completion)bond161#2 (work_completion)(&(&slave->notify_work)->work) &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)bond161#2 (work_completion)(&(&slave->notify_work)->work) &obj_hash[i].lock irq_context: 0 (wq_completion)bond161#2 (work_completion)(&(&slave->notify_work)->work) &base->lock irq_context: 0 (wq_completion)bond161#2 (work_completion)(&(&slave->notify_work)->work) &base->lock &obj_hash[i].lock irq_context: 0 (wq_completion)bond161#2 (work_completion)(&(&bond->mcast_work)->work) irq_context: 0 (wq_completion)bond161#2 (work_completion)(&(&bond->mcast_work)->work) &obj_hash[i].lock irq_context: 0 (wq_completion)bond161#2 (work_completion)(&(&bond->mcast_work)->work) &base->lock irq_context: 0 (wq_completion)bond161#2 (work_completion)(&(&bond->mcast_work)->work) &base->lock &obj_hash[i].lock irq_context: 0 (wq_completion)bond161#2 (work_completion)(&(&bond->mcast_work)->work) &rq->__lock irq_context: 0 (wq_completion)bond161#2 (work_completion)(&(&bond->mcast_work)->work) &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)bond145#3 irq_context: 0 (wq_completion)bond145#3 (work_completion)(&(&slave->notify_work)->work) irq_context: 0 (wq_completion)bond145#3 (work_completion)(&(&slave->notify_work)->work) &obj_hash[i].lock irq_context: 0 (wq_completion)bond145#3 (work_completion)(&(&slave->notify_work)->work) &base->lock irq_context: 0 (wq_completion)bond145#3 (work_completion)(&(&slave->notify_work)->work) &base->lock &obj_hash[i].lock irq_context: 0 (wq_completion)bond145#3 (work_completion)(&(&slave->notify_work)->work) &rq->__lock irq_context: 0 (wq_completion)bond145#3 (work_completion)(&(&slave->notify_work)->work) &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)bond153#2 &rq->__lock irq_context: 0 (wq_completion)bond153#2 &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)bond145#3 (work_completion)(&(&bond->mcast_work)->work) irq_context: 0 (wq_completion)bond145#3 (work_completion)(&(&bond->mcast_work)->work) &obj_hash[i].lock irq_context: 0 (wq_completion)bond145#3 (work_completion)(&(&bond->mcast_work)->work) &base->lock irq_context: 0 (wq_completion)bond145#3 (work_completion)(&(&bond->mcast_work)->work) &base->lock &obj_hash[i].lock irq_context: 0 (wq_completion)bond132#3 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex fs_reclaim irq_context: 0 (wq_completion)bond132#3 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex fs_reclaim &rq->__lock irq_context: 0 (wq_completion)bond132#3 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex fs_reclaim &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)bond132#3 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex fs_reclaim mmu_notifier_invalidate_range_start irq_context: 0 (wq_completion)bond132#3 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &c->lock irq_context: 0 (wq_completion)bond132#3 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex rcu_read_lock &bond->stats_lock/1 irq_context: 0 (wq_completion)bond132#3 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex rcu_read_lock &ndev->lock irq_context: 0 (wq_completion)bond132#3 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex nl_table_lock irq_context: 0 (wq_completion)bond132#3 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &obj_hash[i].lock irq_context: 0 (wq_completion)bond132#3 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex nl_table_wait.lock irq_context: 0 (wq_completion)bond132#3 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex net_rwsem irq_context: 0 (wq_completion)bond132#3 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex net_rwsem &list->lock#2 irq_context: 0 (wq_completion)bond132#3 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &tn->lock irq_context: 0 (wq_completion)bond132#3 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &idev->mc_lock irq_context: 0 (wq_completion)bond132#3 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &idev->mc_lock &rq->__lock irq_context: 0 (wq_completion)bond132#3 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &idev->mc_lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)bond132#3 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &idev->mc_lock fs_reclaim irq_context: 0 (wq_completion)bond132#3 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &idev->mc_lock fs_reclaim mmu_notifier_invalidate_range_start irq_context: 0 (wq_completion)bond132#3 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &idev->mc_lock rcu_read_lock &ndev->lock irq_context: 0 (wq_completion)bond132#3 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &idev->mc_lock rcu_read_lock &c->lock irq_context: 0 (wq_completion)bond132#3 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &idev->mc_lock rcu_read_lock &dir->lock#2 irq_context: 0 (wq_completion)bond132#3 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &idev->mc_lock rcu_read_lock &ul->lock#2 irq_context: 0 (wq_completion)bond132#3 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &idev->mc_lock rcu_read_lock rcu_read_lock &____s->seqcount#10 irq_context: 0 (wq_completion)bond132#3 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &idev->mc_lock rcu_read_lock rcu_read_lock rcu_read_lock_bh rcu_read_lock rcu_read_lock_bh rcu_read_lock &br->multicast_lock irq_context: 0 (wq_completion)bond132#3 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &idev->mc_lock rcu_read_lock rcu_read_lock rcu_read_lock_bh rcu_read_lock rcu_read_lock_bh rcu_read_lock &br->multicast_lock &obj_hash[i].lock irq_context: 0 (wq_completion)bond132#3 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &idev->mc_lock rcu_read_lock rcu_read_lock rcu_read_lock_bh rcu_read_lock rcu_read_lock_bh rcu_read_lock &br->multicast_lock &base->lock irq_context: 0 (wq_completion)bond132#3 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &idev->mc_lock rcu_read_lock rcu_read_lock rcu_read_lock_bh rcu_read_lock rcu_read_lock_bh rcu_read_lock &br->multicast_lock &base->lock &obj_hash[i].lock irq_context: 0 (wq_completion)bond132#3 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &idev->mc_lock rcu_read_lock rcu_read_lock rcu_read_lock_bh rcu_read_lock rcu_read_lock_bh rcu_read_lock &data->lock irq_context: 0 (wq_completion)bond132#3 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &idev->mc_lock rcu_read_lock rcu_read_lock rcu_read_lock_bh rcu_read_lock rcu_read_lock_bh rcu_read_lock &obj_hash[i].lock irq_context: 0 (wq_completion)bond132#3 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &idev->mc_lock rcu_read_lock rcu_read_lock rcu_node_0 irq_context: 0 (wq_completion)bond132#3 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &idev->mc_lock rcu_read_lock rcu_read_lock &rq->__lock irq_context: 0 (wq_completion)bond132#3 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &idev->mc_lock rcu_read_lock rcu_read_lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)bond132#3 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &idev->mc_lock rcu_read_lock rcu_node_0 irq_context: 0 (wq_completion)bond132#3 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex.wait_lock irq_context: 0 (wq_completion)bond132#3 (work_completion)(&(&bond->mcast_work)->work) &p->pi_lock irq_context: 0 (wq_completion)bond132#3 (work_completion)(&(&slave->notify_work)->work) rtnl_mutex net_rwsem irq_context: 0 (wq_completion)bond132#3 (work_completion)(&(&slave->notify_work)->work) rtnl_mutex net_rwsem &list->lock#2 irq_context: 0 (wq_completion)bond132#3 (work_completion)(&(&slave->notify_work)->work) rtnl_mutex &tn->lock irq_context: 0 (wq_completion)bond132#3 (work_completion)(&(&slave->notify_work)->work) rtnl_mutex.wait_lock irq_context: 0 (wq_completion)bond132#3 (work_completion)(&(&slave->notify_work)->work) &p->pi_lock irq_context: 0 (wq_completion)bond174 irq_context: 0 (wq_completion)bond174 (work_completion)(&(&slave->notify_work)->work) irq_context: 0 (wq_completion)bond174 (work_completion)(&(&slave->notify_work)->work) &obj_hash[i].lock irq_context: 0 (wq_completion)bond174 (work_completion)(&(&slave->notify_work)->work) &base->lock irq_context: 0 (wq_completion)bond174 (work_completion)(&(&slave->notify_work)->work) &base->lock &obj_hash[i].lock irq_context: 0 (wq_completion)bond174 (work_completion)(&(&slave->notify_work)->work) &rq->__lock irq_context: 0 (wq_completion)bond174 (work_completion)(&(&slave->notify_work)->work) &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)bond174 (work_completion)(&(&bond->mcast_work)->work) irq_context: 0 (wq_completion)bond174 (work_completion)(&(&bond->mcast_work)->work) &obj_hash[i].lock irq_context: 0 (wq_completion)bond174 (work_completion)(&(&bond->mcast_work)->work) &base->lock irq_context: 0 (wq_completion)bond174 (work_completion)(&(&bond->mcast_work)->work) &base->lock &obj_hash[i].lock irq_context: 0 (wq_completion)bond174 (work_completion)(&(&bond->mcast_work)->work) &rq->__lock irq_context: 0 (wq_completion)bond174 (work_completion)(&(&bond->mcast_work)->work) &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)bond118#2 (work_completion)(&(&slave->notify_work)->work) &cfs_rq->removed.lock irq_context: 0 (wq_completion)bond174 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex fs_reclaim irq_context: 0 (wq_completion)bond174 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex fs_reclaim mmu_notifier_invalidate_range_start irq_context: 0 (wq_completion)bond174 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex rcu_read_lock &bond->stats_lock/1 irq_context: 0 (wq_completion)bond174 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex rcu_read_lock &ndev->lock irq_context: 0 (wq_completion)bond174 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex nl_table_lock irq_context: 0 (wq_completion)bond174 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &obj_hash[i].lock irq_context: 0 (wq_completion)bond174 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex nl_table_wait.lock irq_context: 0 (wq_completion)bond174 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex net_rwsem irq_context: 0 (wq_completion)bond174 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex net_rwsem &list->lock#2 irq_context: 0 (wq_completion)bond174 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &tn->lock irq_context: 0 (wq_completion)bond174 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &idev->mc_lock irq_context: 0 (wq_completion)bond174 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &idev->mc_lock fs_reclaim irq_context: 0 (wq_completion)bond174 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &idev->mc_lock fs_reclaim mmu_notifier_invalidate_range_start irq_context: 0 (wq_completion)bond174 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &idev->mc_lock rcu_read_lock &ndev->lock irq_context: 0 (wq_completion)bond174 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &idev->mc_lock rcu_read_lock &c->lock irq_context: 0 (wq_completion)bond174 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &idev->mc_lock rcu_read_lock &dir->lock#2 irq_context: 0 (wq_completion)bond174 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &idev->mc_lock rcu_read_lock &ul->lock#2 irq_context: 0 (wq_completion)bond174 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &idev->mc_lock rcu_read_lock rcu_read_lock &obj_hash[i].lock irq_context: 0 (wq_completion)bond174 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &idev->mc_lock rcu_read_lock rcu_read_lock &c->lock irq_context: 0 (wq_completion)bond174 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &idev->mc_lock rcu_read_lock rcu_read_lock &dir->lock#2 irq_context: 0 (wq_completion)bond174 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &idev->mc_lock rcu_read_lock rcu_read_lock &tbl->lock irq_context: 0 (wq_completion)bond174 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &idev->mc_lock rcu_read_lock rcu_read_lock &n->lock irq_context: 0 (wq_completion)bond174 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &idev->mc_lock rcu_read_lock rcu_read_lock &____s->seqcount#9 irq_context: 0 (wq_completion)bond174 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &idev->mc_lock rcu_read_lock rcu_read_lock rcu_read_lock_bh rcu_read_lock rcu_read_lock_bh rcu_read_lock &br->multicast_lock irq_context: 0 (wq_completion)bond174 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &idev->mc_lock rcu_read_lock rcu_read_lock rcu_read_lock_bh rcu_read_lock rcu_read_lock_bh rcu_read_lock &br->multicast_lock &obj_hash[i].lock irq_context: 0 (wq_completion)bond174 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &idev->mc_lock rcu_read_lock rcu_read_lock rcu_read_lock_bh rcu_read_lock rcu_read_lock_bh rcu_read_lock &br->multicast_lock rcu_read_lock rhashtable_bucket irq_context: 0 (wq_completion)bond174 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &idev->mc_lock rcu_read_lock rcu_read_lock rcu_read_lock_bh rcu_read_lock rcu_read_lock_bh rcu_read_lock &br->multicast_lock nl_table_lock irq_context: 0 (wq_completion)bond174 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &idev->mc_lock rcu_read_lock rcu_read_lock rcu_read_lock_bh rcu_read_lock rcu_read_lock_bh rcu_read_lock &br->multicast_lock nl_table_wait.lock irq_context: 0 (wq_completion)bond174 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &idev->mc_lock rcu_read_lock rcu_read_lock rcu_read_lock_bh rcu_read_lock rcu_read_lock_bh rcu_read_lock &br->multicast_lock &base->lock irq_context: 0 (wq_completion)bond174 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &idev->mc_lock rcu_read_lock rcu_read_lock rcu_read_lock_bh rcu_read_lock rcu_read_lock_bh rcu_read_lock &br->multicast_lock &base->lock &obj_hash[i].lock irq_context: 0 (wq_completion)bond174 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &idev->mc_lock rcu_read_lock rcu_read_lock rcu_read_lock_bh rcu_read_lock rcu_read_lock_bh rcu_read_lock &data->lock irq_context: 0 (wq_completion)bond174 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &idev->mc_lock rcu_read_lock rcu_read_lock rcu_read_lock_bh rcu_read_lock rcu_read_lock_bh rcu_read_lock &obj_hash[i].lock irq_context: 0 (wq_completion)bond174 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex.wait_lock irq_context: 0 (wq_completion)bond174 (work_completion)(&(&bond->mcast_work)->work) &p->pi_lock irq_context: 0 (wq_completion)bond131#2 (work_completion)(&(&bond->mcast_work)->work) &rq->__lock irq_context: 0 (wq_completion)bond131#2 (work_completion)(&(&bond->mcast_work)->work) &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)bond160#2 &rq->__lock irq_context: 0 (wq_completion)bond162#2 irq_context: 0 (wq_completion)bond162#2 &rq->__lock irq_context: 0 (wq_completion)bond162#2 &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)bond162#2 (work_completion)(&(&slave->notify_work)->work) irq_context: 0 (wq_completion)bond162#2 (work_completion)(&(&slave->notify_work)->work) &obj_hash[i].lock irq_context: 0 (wq_completion)bond162#2 (work_completion)(&(&slave->notify_work)->work) &base->lock irq_context: 0 (wq_completion)bond162#2 (work_completion)(&(&slave->notify_work)->work) &base->lock &obj_hash[i].lock irq_context: 0 (wq_completion)bond162#2 (work_completion)(&(&slave->notify_work)->work) &rq->__lock irq_context: 0 (wq_completion)bond172 (work_completion)(&(&bond->mcast_work)->work) &rq->__lock irq_context: 0 (wq_completion)bond172 (work_completion)(&(&bond->mcast_work)->work) &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)bond162#2 (work_completion)(&(&bond->mcast_work)->work) irq_context: 0 (wq_completion)bond162#2 (work_completion)(&(&bond->mcast_work)->work) &obj_hash[i].lock irq_context: 0 (wq_completion)bond162#2 (work_completion)(&(&bond->mcast_work)->work) &base->lock irq_context: 0 (wq_completion)bond162#2 (work_completion)(&(&bond->mcast_work)->work) &base->lock &obj_hash[i].lock irq_context: 0 (wq_completion)bond162#2 (work_completion)(&(&bond->mcast_work)->work) &rq->__lock irq_context: 0 (wq_completion)bond162#2 (work_completion)(&(&bond->mcast_work)->work) &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)bond166 (work_completion)(&(&slave->notify_work)->work) &rq->__lock irq_context: 0 (wq_completion)bond166 (work_completion)(&(&slave->notify_work)->work) &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)bond172 (work_completion)(&(&slave->notify_work)->work) rtnl_mutex &rq->__lock irq_context: 0 (wq_completion)bond172 (work_completion)(&(&slave->notify_work)->work) rtnl_mutex net_rwsem irq_context: 0 (wq_completion)bond172 (work_completion)(&(&slave->notify_work)->work) rtnl_mutex net_rwsem &rq->__lock irq_context: 0 (wq_completion)bond172 (work_completion)(&(&slave->notify_work)->work) rtnl_mutex net_rwsem &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)bond172 (work_completion)(&(&slave->notify_work)->work) rtnl_mutex net_rwsem &list->lock#2 irq_context: 0 (wq_completion)bond172 (work_completion)(&(&slave->notify_work)->work) rtnl_mutex &tn->lock irq_context: 0 (wq_completion)bond172 (work_completion)(&(&slave->notify_work)->work) rtnl_mutex.wait_lock irq_context: 0 (wq_completion)bond172 (work_completion)(&(&slave->notify_work)->work) &p->pi_lock irq_context: 0 (wq_completion)bond172 (work_completion)(&(&slave->notify_work)->work) &p->pi_lock &rq->__lock irq_context: 0 (wq_completion)bond172 (work_completion)(&(&slave->notify_work)->work) &p->pi_lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)bond158#2 (work_completion)(&(&bond->mcast_work)->work) &rq->__lock irq_context: 0 (wq_completion)bond158#2 (work_completion)(&(&bond->mcast_work)->work) &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)bond146#3 irq_context: 0 (wq_completion)bond146#3 &rq->__lock irq_context: 0 (wq_completion)bond146#3 &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)bond146#3 (work_completion)(&(&slave->notify_work)->work) irq_context: 0 (wq_completion)bond146#3 (work_completion)(&(&slave->notify_work)->work) &obj_hash[i].lock irq_context: 0 (wq_completion)bond146#3 (work_completion)(&(&slave->notify_work)->work) &base->lock irq_context: 0 (wq_completion)bond146#3 (work_completion)(&(&slave->notify_work)->work) &base->lock &obj_hash[i].lock irq_context: 0 (wq_completion)bond146#3 (work_completion)(&(&slave->notify_work)->work) &rq->__lock irq_context: 0 (wq_completion)bond146#3 (work_completion)(&(&slave->notify_work)->work) &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)bond146#3 (work_completion)(&(&bond->mcast_work)->work) irq_context: 0 (wq_completion)bond146#3 (work_completion)(&(&bond->mcast_work)->work) &obj_hash[i].lock irq_context: 0 (wq_completion)bond146#3 (work_completion)(&(&bond->mcast_work)->work) &base->lock irq_context: 0 (wq_completion)bond146#3 (work_completion)(&(&bond->mcast_work)->work) &base->lock &obj_hash[i].lock irq_context: 0 (wq_completion)bond146#2 &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)bond48#2 irq_context: 0 (wq_completion)bond48#2 (work_completion)(&(&slave->notify_work)->work) irq_context: 0 (wq_completion)bond48#2 (work_completion)(&(&slave->notify_work)->work) &obj_hash[i].lock irq_context: 0 (wq_completion)bond48#2 (work_completion)(&(&slave->notify_work)->work) &base->lock irq_context: 0 (wq_completion)bond48#2 (work_completion)(&(&slave->notify_work)->work) &base->lock &obj_hash[i].lock irq_context: 0 (wq_completion)bond48#2 (work_completion)(&(&slave->notify_work)->work) &rq->__lock irq_context: 0 (wq_completion)bond48#2 (work_completion)(&(&slave->notify_work)->work) &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)bond48#2 (work_completion)(&(&bond->mcast_work)->work) irq_context: 0 (wq_completion)bond48#2 (work_completion)(&(&bond->mcast_work)->work) &obj_hash[i].lock irq_context: 0 (wq_completion)bond48#2 (work_completion)(&(&bond->mcast_work)->work) &base->lock irq_context: 0 (wq_completion)bond48#2 (work_completion)(&(&bond->mcast_work)->work) &base->lock &obj_hash[i].lock irq_context: 0 (wq_completion)bond48#2 (work_completion)(&(&bond->mcast_work)->work) &rq->__lock irq_context: 0 (wq_completion)bond48#2 (work_completion)(&(&bond->mcast_work)->work) &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)bond175 irq_context: 0 (wq_completion)bond175 &rq->__lock irq_context: 0 (wq_completion)bond175 &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)bond175 (work_completion)(&(&slave->notify_work)->work) irq_context: 0 (wq_completion)bond175 (work_completion)(&(&slave->notify_work)->work) &obj_hash[i].lock irq_context: 0 (wq_completion)bond175 (work_completion)(&(&slave->notify_work)->work) &base->lock irq_context: 0 (wq_completion)bond175 (work_completion)(&(&slave->notify_work)->work) &base->lock &obj_hash[i].lock irq_context: 0 (wq_completion)bond175 (work_completion)(&(&slave->notify_work)->work) &rq->__lock irq_context: 0 (wq_completion)bond175 (work_completion)(&(&slave->notify_work)->work) &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)bond175 (work_completion)(&(&bond->mcast_work)->work) irq_context: 0 (wq_completion)bond175 (work_completion)(&(&bond->mcast_work)->work) &obj_hash[i].lock irq_context: 0 (wq_completion)bond175 (work_completion)(&(&bond->mcast_work)->work) &base->lock irq_context: 0 (wq_completion)bond175 (work_completion)(&(&bond->mcast_work)->work) &base->lock &obj_hash[i].lock irq_context: 0 (wq_completion)bond163#2 irq_context: 0 (wq_completion)bond163#2 (work_completion)(&(&slave->notify_work)->work) irq_context: 0 (wq_completion)bond163#2 (work_completion)(&(&slave->notify_work)->work) &rq->__lock irq_context: 0 (wq_completion)bond163#2 (work_completion)(&(&slave->notify_work)->work) &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)bond163#2 (work_completion)(&(&slave->notify_work)->work) &obj_hash[i].lock irq_context: 0 (wq_completion)bond163#2 (work_completion)(&(&slave->notify_work)->work) &base->lock irq_context: 0 (wq_completion)bond163#2 (work_completion)(&(&slave->notify_work)->work) &base->lock &obj_hash[i].lock irq_context: 0 (wq_completion)bond163#2 (work_completion)(&(&bond->mcast_work)->work) irq_context: 0 (wq_completion)bond163#2 (work_completion)(&(&bond->mcast_work)->work) &obj_hash[i].lock irq_context: 0 (wq_completion)bond163#2 (work_completion)(&(&bond->mcast_work)->work) &base->lock irq_context: 0 (wq_completion)bond163#2 (work_completion)(&(&bond->mcast_work)->work) &base->lock &obj_hash[i].lock irq_context: 0 (wq_completion)bond163#2 (work_completion)(&(&bond->mcast_work)->work) &rq->__lock irq_context: 0 (wq_completion)bond163#2 (work_completion)(&(&bond->mcast_work)->work) &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)bond143#2 (work_completion)(&(&bond->mcast_work)->work) fill_pool_map-wait-type-override &c->lock irq_context: 0 (wq_completion)bond143#2 (work_completion)(&(&bond->mcast_work)->work) fill_pool_map-wait-type-override &____s->seqcount#2 irq_context: 0 (wq_completion)bond143#2 (work_completion)(&(&bond->mcast_work)->work) fill_pool_map-wait-type-override &____s->seqcount irq_context: 0 (wq_completion)bond143#2 (work_completion)(&(&bond->mcast_work)->work) fill_pool_map-wait-type-override pool_lock irq_context: 0 (wq_completion)bond147#3 irq_context: 0 (wq_completion)bond147#3 (work_completion)(&(&slave->notify_work)->work) irq_context: 0 (wq_completion)bond147#3 (work_completion)(&(&slave->notify_work)->work) &obj_hash[i].lock irq_context: 0 (wq_completion)bond147#3 (work_completion)(&(&slave->notify_work)->work) &base->lock irq_context: 0 (wq_completion)bond147#3 (work_completion)(&(&slave->notify_work)->work) &base->lock &obj_hash[i].lock irq_context: 0 (wq_completion)bond147#3 (work_completion)(&(&slave->notify_work)->work) &rq->__lock irq_context: 0 (wq_completion)bond147#3 (work_completion)(&(&bond->mcast_work)->work) irq_context: 0 (wq_completion)bond147#3 (work_completion)(&(&bond->mcast_work)->work) &obj_hash[i].lock irq_context: 0 (wq_completion)bond147#3 (work_completion)(&(&bond->mcast_work)->work) &base->lock irq_context: 0 (wq_completion)bond147#3 (work_completion)(&(&bond->mcast_work)->work) &base->lock &obj_hash[i].lock irq_context: 0 (wq_completion)bond147#3 (work_completion)(&(&bond->mcast_work)->work) &rq->__lock irq_context: 0 (wq_completion)bond147#3 (work_completion)(&(&bond->mcast_work)->work) &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)bond31#3 (work_completion)(&(&slave->notify_work)->work) fill_pool_map-wait-type-override pool_lock irq_context: 0 (wq_completion)bond145#3 (work_completion)(&(&bond->mcast_work)->work) &rq->__lock irq_context: 0 (wq_completion)bond145#3 (work_completion)(&(&bond->mcast_work)->work) &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)bond49#2 irq_context: 0 (wq_completion)bond49#2 (work_completion)(&(&slave->notify_work)->work) irq_context: 0 (wq_completion)bond49#2 (work_completion)(&(&slave->notify_work)->work) &rq->__lock irq_context: 0 (wq_completion)bond49#2 (work_completion)(&(&slave->notify_work)->work) &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)bond49#2 (work_completion)(&(&slave->notify_work)->work) &obj_hash[i].lock irq_context: 0 (wq_completion)bond49#2 (work_completion)(&(&slave->notify_work)->work) &base->lock irq_context: 0 (wq_completion)bond49#2 (work_completion)(&(&slave->notify_work)->work) &base->lock &obj_hash[i].lock irq_context: 0 (wq_completion)bond49#2 (work_completion)(&(&bond->mcast_work)->work) irq_context: 0 (wq_completion)bond49#2 (work_completion)(&(&bond->mcast_work)->work) &obj_hash[i].lock irq_context: 0 (wq_completion)bond49#2 (work_completion)(&(&bond->mcast_work)->work) &base->lock irq_context: 0 (wq_completion)bond49#2 (work_completion)(&(&bond->mcast_work)->work) &base->lock &obj_hash[i].lock irq_context: 0 (wq_completion)mld (work_completion)(&(&idev->mc_dad_work)->work) &idev->mc_lock rcu_read_lock quarantine_lock irq_context: 0 (wq_completion)gid-cache-wq (work_completion)(&ndev_work->work) devices_rwsem rcu_node_0 irq_context: 0 (wq_completion)bond176 irq_context: 0 (wq_completion)bond176 (work_completion)(&(&slave->notify_work)->work) irq_context: 0 (wq_completion)bond176 (work_completion)(&(&slave->notify_work)->work) &obj_hash[i].lock irq_context: 0 (wq_completion)bond176 (work_completion)(&(&slave->notify_work)->work) &base->lock irq_context: 0 (wq_completion)bond176 (work_completion)(&(&slave->notify_work)->work) &base->lock &obj_hash[i].lock irq_context: 0 (wq_completion)bond176 (work_completion)(&(&slave->notify_work)->work) &rq->__lock irq_context: 0 (wq_completion)bond176 (work_completion)(&(&slave->notify_work)->work) &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)bond176 (work_completion)(&(&bond->mcast_work)->work) irq_context: 0 (wq_completion)bond176 (work_completion)(&(&bond->mcast_work)->work) &obj_hash[i].lock irq_context: 0 (wq_completion)bond176 (work_completion)(&(&bond->mcast_work)->work) &base->lock irq_context: 0 (wq_completion)bond176 (work_completion)(&(&bond->mcast_work)->work) &base->lock &obj_hash[i].lock irq_context: 0 (wq_completion)bond150#2 (work_completion)(&(&slave->notify_work)->work) rtnl_mutex net_rwsem irq_context: 0 (wq_completion)bond150#2 (work_completion)(&(&slave->notify_work)->work) rtnl_mutex net_rwsem &rq->__lock irq_context: 0 (wq_completion)bond150#2 (work_completion)(&(&slave->notify_work)->work) rtnl_mutex net_rwsem &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)bond150#2 (work_completion)(&(&slave->notify_work)->work) rtnl_mutex net_rwsem &list->lock#2 irq_context: 0 (wq_completion)bond150#2 (work_completion)(&(&slave->notify_work)->work) rtnl_mutex &tn->lock irq_context: 0 (wq_completion)bond150#2 (work_completion)(&(&slave->notify_work)->work) rtnl_mutex.wait_lock irq_context: 0 (wq_completion)bond150#2 (work_completion)(&(&slave->notify_work)->work) &p->pi_lock irq_context: 0 (wq_completion)bond176 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex fs_reclaim irq_context: 0 (wq_completion)bond176 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex fs_reclaim mmu_notifier_invalidate_range_start irq_context: 0 (wq_completion)bond176 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &c->lock irq_context: 0 (wq_completion)bond176 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex rcu_read_lock &bond->stats_lock/1 irq_context: 0 (wq_completion)bond176 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex rcu_read_lock &ndev->lock irq_context: 0 (wq_completion)bond176 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex rcu_read_lock rcu_node_0 irq_context: 0 (wq_completion)bond176 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex rcu_read_lock &rq->__lock irq_context: 0 (wq_completion)bond176 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex rcu_read_lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)bond176 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex nl_table_lock irq_context: 0 (wq_completion)bond176 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &obj_hash[i].lock irq_context: 0 (wq_completion)bond176 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex nl_table_wait.lock irq_context: 0 (wq_completion)bond176 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex net_rwsem irq_context: 0 (wq_completion)bond176 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex net_rwsem &list->lock#2 irq_context: 0 (wq_completion)bond176 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &tn->lock irq_context: 0 (wq_completion)bond176 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &idev->mc_lock irq_context: 0 (wq_completion)bond176 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex.wait_lock irq_context: 0 (wq_completion)bond134#3 (work_completion)(&(&slave->notify_work)->work) rtnl_mutex net_rwsem irq_context: 0 (wq_completion)bond176 (work_completion)(&(&bond->mcast_work)->work) &rq->__lock irq_context: 0 (wq_completion)bond134#3 (work_completion)(&(&slave->notify_work)->work) rtnl_mutex net_rwsem &list->lock#2 irq_context: 0 (wq_completion)bond176 (work_completion)(&(&bond->mcast_work)->work) &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)bond134#3 (work_completion)(&(&slave->notify_work)->work) rtnl_mutex &tn->lock irq_context: 0 (wq_completion)bond176 (work_completion)(&(&bond->mcast_work)->work) &p->pi_lock irq_context: 0 (wq_completion)bond134#3 (work_completion)(&(&slave->notify_work)->work) rtnl_mutex.wait_lock irq_context: 0 (wq_completion)bond102#3 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex fs_reclaim irq_context: 0 (wq_completion)bond134#3 (work_completion)(&(&slave->notify_work)->work) &p->pi_lock irq_context: 0 (wq_completion)bond102#3 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex fs_reclaim mmu_notifier_invalidate_range_start irq_context: 0 (wq_completion)bond102#3 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &c->lock irq_context: 0 (wq_completion)bond102#3 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &n->list_lock irq_context: 0 (wq_completion)bond102#3 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &n->list_lock &c->lock irq_context: 0 (wq_completion)bond102#3 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex rcu_read_lock &bond->stats_lock/1 irq_context: 0 (wq_completion)bond102#3 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex rcu_read_lock rcu_node_0 irq_context: 0 (wq_completion)bond102#3 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex rcu_read_lock &rq->__lock irq_context: 0 (wq_completion)bond102#3 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex rcu_read_lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)bond102#3 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex rcu_read_lock &ndev->lock irq_context: 0 (wq_completion)bond102#3 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex nl_table_lock irq_context: 0 (wq_completion)bond102#3 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &obj_hash[i].lock irq_context: 0 (wq_completion)bond102#3 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex nl_table_wait.lock irq_context: 0 (wq_completion)bond102#3 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex net_rwsem irq_context: 0 (wq_completion)bond102#3 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex net_rwsem &list->lock#2 irq_context: 0 (wq_completion)bond102#3 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &tn->lock irq_context: 0 (wq_completion)bond102#3 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &idev->mc_lock irq_context: 0 (wq_completion)bond102#3 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &idev->mc_lock fs_reclaim irq_context: 0 (wq_completion)bond102#3 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &idev->mc_lock fs_reclaim mmu_notifier_invalidate_range_start irq_context: 0 (wq_completion)bond102#3 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &idev->mc_lock &c->lock irq_context: 0 (wq_completion)bond102#3 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &idev->mc_lock &rq->__lock irq_context: 0 (wq_completion)bond102#3 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &idev->mc_lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)bond102#3 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &idev->mc_lock rcu_read_lock &ndev->lock irq_context: 0 (wq_completion)bond102#3 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &idev->mc_lock rcu_read_lock &dir->lock#2 irq_context: 0 (wq_completion)bond102#3 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &idev->mc_lock rcu_read_lock &ul->lock#2 irq_context: 0 (wq_completion)bond102#3 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &idev->mc_lock rcu_read_lock rcu_read_lock &obj_hash[i].lock irq_context: 0 (wq_completion)bond102#3 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &idev->mc_lock rcu_read_lock rcu_read_lock &dir->lock#2 irq_context: 0 (wq_completion)bond102#3 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &idev->mc_lock rcu_read_lock rcu_read_lock &tbl->lock irq_context: 0 (wq_completion)bond102#3 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &idev->mc_lock rcu_read_lock rcu_read_lock &n->lock irq_context: 0 (wq_completion)bond102#3 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &idev->mc_lock rcu_read_lock rcu_read_lock &____s->seqcount#9 irq_context: 0 (wq_completion)bond102#3 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &idev->mc_lock rcu_read_lock rcu_read_lock rcu_read_lock_bh rcu_read_lock rcu_read_lock_bh rcu_read_lock &br->multicast_lock irq_context: 0 (wq_completion)bond102#3 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &idev->mc_lock rcu_read_lock rcu_read_lock rcu_read_lock_bh rcu_read_lock rcu_read_lock_bh rcu_read_lock &br->multicast_lock &obj_hash[i].lock irq_context: 0 (wq_completion)bond102#3 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &idev->mc_lock rcu_read_lock rcu_read_lock rcu_read_lock_bh rcu_read_lock rcu_read_lock_bh rcu_read_lock &br->multicast_lock &base->lock irq_context: 0 (wq_completion)bond102#3 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &idev->mc_lock rcu_read_lock rcu_read_lock rcu_read_lock_bh rcu_read_lock rcu_read_lock_bh rcu_read_lock &br->multicast_lock &base->lock &obj_hash[i].lock irq_context: 0 (wq_completion)bond102#3 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &idev->mc_lock rcu_read_lock rcu_read_lock rcu_read_lock_bh rcu_read_lock rcu_read_lock_bh rcu_read_lock &data->lock irq_context: 0 (wq_completion)bond102#3 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &idev->mc_lock rcu_read_lock rcu_read_lock rcu_read_lock_bh rcu_read_lock rcu_read_lock_bh rcu_read_lock &obj_hash[i].lock irq_context: 0 (wq_completion)bond102#3 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex.wait_lock irq_context: 0 (wq_completion)bond102#3 (work_completion)(&(&bond->mcast_work)->work) &p->pi_lock irq_context: 0 (wq_completion)bond164#2 irq_context: 0 (wq_completion)bond164#2 (work_completion)(&(&slave->notify_work)->work) irq_context: 0 (wq_completion)bond164#2 (work_completion)(&(&slave->notify_work)->work) &obj_hash[i].lock irq_context: 0 (wq_completion)bond164#2 (work_completion)(&(&slave->notify_work)->work) &base->lock irq_context: 0 (wq_completion)bond164#2 (work_completion)(&(&slave->notify_work)->work) &base->lock &obj_hash[i].lock irq_context: 0 (wq_completion)bond164#2 (work_completion)(&(&slave->notify_work)->work) &rq->__lock irq_context: 0 (wq_completion)bond164#2 (work_completion)(&(&slave->notify_work)->work) &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)bond160#2 (work_completion)(&(&slave->notify_work)->work) &rq->__lock irq_context: 0 (wq_completion)bond160#2 (work_completion)(&(&slave->notify_work)->work) &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)bond137#3 (work_completion)(&(&bond->mcast_work)->work) fill_pool_map-wait-type-override pool_lock irq_context: 0 (wq_completion)bond164#2 (work_completion)(&(&bond->mcast_work)->work) irq_context: 0 (wq_completion)bond164#2 (work_completion)(&(&bond->mcast_work)->work) &obj_hash[i].lock irq_context: 0 (wq_completion)bond164#2 (work_completion)(&(&bond->mcast_work)->work) &base->lock irq_context: 0 (wq_completion)bond164#2 (work_completion)(&(&bond->mcast_work)->work) &base->lock &obj_hash[i].lock irq_context: 0 (wq_completion)bond164#2 (work_completion)(&(&bond->mcast_work)->work) &rq->__lock irq_context: 0 (wq_completion)bond129#2 &rq->__lock irq_context: 0 (wq_completion)bond129#2 &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)bond148#3 irq_context: 0 (wq_completion)bond148#3 &rq->__lock irq_context: 0 (wq_completion)bond148#3 &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)bond148#3 (work_completion)(&(&slave->notify_work)->work) irq_context: 0 (wq_completion)bond148#3 (work_completion)(&(&slave->notify_work)->work) &obj_hash[i].lock irq_context: 0 (wq_completion)bond148#3 (work_completion)(&(&slave->notify_work)->work) &base->lock irq_context: 0 (wq_completion)bond148#3 (work_completion)(&(&slave->notify_work)->work) &base->lock &obj_hash[i].lock irq_context: 0 (wq_completion)bond148#3 (work_completion)(&(&slave->notify_work)->work) &rq->__lock irq_context: 0 (wq_completion)bond148#3 (work_completion)(&(&slave->notify_work)->work) &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)bond148#3 (work_completion)(&(&bond->mcast_work)->work) irq_context: 0 (wq_completion)bond148#3 (work_completion)(&(&bond->mcast_work)->work) &obj_hash[i].lock irq_context: 0 (wq_completion)bond148#3 (work_completion)(&(&bond->mcast_work)->work) &base->lock irq_context: 0 (wq_completion)bond148#3 (work_completion)(&(&bond->mcast_work)->work) &base->lock &obj_hash[i].lock irq_context: 0 (wq_completion)bond177 irq_context: 0 (wq_completion)bond177 (work_completion)(&(&slave->notify_work)->work) irq_context: 0 (wq_completion)bond177 (work_completion)(&(&slave->notify_work)->work) &obj_hash[i].lock irq_context: 0 (wq_completion)bond177 (work_completion)(&(&slave->notify_work)->work) &base->lock irq_context: 0 (wq_completion)bond177 (work_completion)(&(&slave->notify_work)->work) &base->lock &obj_hash[i].lock irq_context: 0 (wq_completion)bond177 (work_completion)(&(&slave->notify_work)->work) &rq->__lock irq_context: 0 (wq_completion)bond177 (work_completion)(&(&slave->notify_work)->work) &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)bond154 (work_completion)(&(&bond->mcast_work)->work) fill_pool_map-wait-type-override &c->lock irq_context: 0 (wq_completion)bond154 (work_completion)(&(&bond->mcast_work)->work) fill_pool_map-wait-type-override &____s->seqcount#2 irq_context: 0 (wq_completion)bond154 (work_completion)(&(&bond->mcast_work)->work) fill_pool_map-wait-type-override &____s->seqcount irq_context: 0 (wq_completion)bond177 (work_completion)(&(&bond->mcast_work)->work) irq_context: 0 (wq_completion)bond177 (work_completion)(&(&bond->mcast_work)->work) &obj_hash[i].lock irq_context: 0 (wq_completion)bond177 (work_completion)(&(&bond->mcast_work)->work) &base->lock irq_context: 0 (wq_completion)bond177 (work_completion)(&(&bond->mcast_work)->work) &base->lock &obj_hash[i].lock irq_context: 0 (wq_completion)bond50#2 irq_context: 0 (wq_completion)bond50#2 (work_completion)(&(&slave->notify_work)->work) irq_context: 0 (wq_completion)bond50#2 (work_completion)(&(&slave->notify_work)->work) &obj_hash[i].lock irq_context: 0 (wq_completion)bond50#2 (work_completion)(&(&slave->notify_work)->work) &base->lock irq_context: 0 (wq_completion)bond50#2 (work_completion)(&(&slave->notify_work)->work) &base->lock &obj_hash[i].lock irq_context: 0 (wq_completion)bond50#2 (work_completion)(&(&slave->notify_work)->work) &rq->__lock irq_context: 0 (wq_completion)bond50#2 (work_completion)(&(&slave->notify_work)->work) &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)bond150#2 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex fs_reclaim irq_context: 0 (wq_completion)bond150#2 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex fs_reclaim mmu_notifier_invalidate_range_start irq_context: 0 (wq_completion)bond150#2 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &c->lock irq_context: 0 (wq_completion)bond150#2 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex rcu_read_lock &bond->stats_lock/1 irq_context: 0 (wq_completion)bond150#2 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex rcu_read_lock &ndev->lock irq_context: 0 (wq_completion)bond150#2 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex nl_table_lock irq_context: 0 (wq_completion)bond150#2 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &obj_hash[i].lock irq_context: 0 (wq_completion)bond150#2 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &rq->__lock irq_context: 0 (wq_completion)bond150#2 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)bond150#2 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex nl_table_wait.lock irq_context: 0 (wq_completion)bond150#2 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex net_rwsem irq_context: 0 (wq_completion)bond150#2 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex net_rwsem &list->lock#2 irq_context: 0 (wq_completion)bond150#2 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &tn->lock irq_context: 0 (wq_completion)bond150#2 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &idev->mc_lock irq_context: 0 (wq_completion)bond150#2 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &idev->mc_lock fs_reclaim irq_context: 0 (wq_completion)bond150#2 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &idev->mc_lock fs_reclaim mmu_notifier_invalidate_range_start irq_context: 0 (wq_completion)bond150#2 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &idev->mc_lock fs_reclaim mmu_notifier_invalidate_range_start &rq->__lock irq_context: 0 (wq_completion)bond150#2 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &idev->mc_lock fs_reclaim mmu_notifier_invalidate_range_start &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)bond150#2 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &idev->mc_lock &c->lock irq_context: 0 (wq_completion)bond150#2 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &idev->mc_lock &n->list_lock irq_context: 0 (wq_completion)bond150#2 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &idev->mc_lock &n->list_lock &c->lock irq_context: 0 (wq_completion)bond150#2 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &idev->mc_lock rcu_read_lock &ndev->lock irq_context: 0 (wq_completion)bond150#2 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &idev->mc_lock rcu_read_lock &dir->lock#2 irq_context: 0 (wq_completion)bond150#2 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &idev->mc_lock rcu_read_lock &ul->lock#2 irq_context: 0 (wq_completion)bond150#2 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &idev->mc_lock rcu_read_lock rcu_read_lock &____s->seqcount#10 irq_context: 0 (wq_completion)bond150#2 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &idev->mc_lock rcu_read_lock rcu_read_lock rcu_read_lock_bh rcu_read_lock rcu_read_lock_bh rcu_read_lock &br->multicast_lock irq_context: 0 (wq_completion)bond150#2 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &idev->mc_lock rcu_read_lock rcu_read_lock rcu_read_lock_bh rcu_read_lock rcu_read_lock_bh rcu_read_lock &br->multicast_lock &obj_hash[i].lock irq_context: 0 (wq_completion)bond150#2 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &idev->mc_lock rcu_read_lock rcu_read_lock rcu_read_lock_bh rcu_read_lock rcu_read_lock_bh rcu_read_lock &br->multicast_lock &base->lock irq_context: 0 (wq_completion)bond150#2 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &idev->mc_lock rcu_read_lock rcu_read_lock rcu_read_lock_bh rcu_read_lock rcu_read_lock_bh rcu_read_lock &br->multicast_lock &base->lock &obj_hash[i].lock irq_context: 0 (wq_completion)bond150#2 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &idev->mc_lock rcu_read_lock rcu_read_lock rcu_read_lock_bh rcu_read_lock rcu_read_lock_bh rcu_read_lock &data->lock irq_context: 0 (wq_completion)bond150#2 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &idev->mc_lock rcu_read_lock rcu_read_lock rcu_read_lock_bh rcu_read_lock rcu_read_lock_bh rcu_read_lock &obj_hash[i].lock irq_context: 0 (wq_completion)bond150#2 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &idev->mc_lock rcu_read_lock rcu_read_lock rcu_node_0 irq_context: 0 (wq_completion)bond150#2 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &idev->mc_lock rcu_read_lock rcu_read_lock &rq->__lock irq_context: 0 (wq_completion)bond150#2 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &idev->mc_lock rcu_read_lock rcu_read_lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)bond50#2 (work_completion)(&(&bond->mcast_work)->work) irq_context: 0 (wq_completion)bond50#2 (work_completion)(&(&bond->mcast_work)->work) &obj_hash[i].lock irq_context: 0 (wq_completion)bond50#2 (work_completion)(&(&bond->mcast_work)->work) &base->lock irq_context: 0 (wq_completion)bond50#2 (work_completion)(&(&bond->mcast_work)->work) &base->lock &obj_hash[i].lock irq_context: 0 (wq_completion)bond150#2 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &idev->mc_lock rcu_read_lock rcu_node_0 irq_context: 0 (wq_completion)bond150#2 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex.wait_lock irq_context: 0 (wq_completion)bond150#2 (work_completion)(&(&bond->mcast_work)->work) &p->pi_lock irq_context: 0 (wq_completion)bond150#2 (work_completion)(&(&bond->mcast_work)->work) &p->pi_lock &rq->__lock irq_context: 0 (wq_completion)bond150#2 (work_completion)(&(&bond->mcast_work)->work) &p->pi_lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)bond139#2 (work_completion)(&(&slave->notify_work)->work) rtnl_mutex net_rwsem irq_context: 0 (wq_completion)bond139#2 (work_completion)(&(&slave->notify_work)->work) rtnl_mutex net_rwsem &list->lock#2 irq_context: 0 (wq_completion)bond139#2 (work_completion)(&(&slave->notify_work)->work) rtnl_mutex &tn->lock irq_context: 0 (wq_completion)bond139#2 (work_completion)(&(&slave->notify_work)->work) rtnl_mutex.wait_lock irq_context: 0 (wq_completion)bond139#2 (work_completion)(&(&slave->notify_work)->work) &p->pi_lock irq_context: 0 (wq_completion)bond109 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex fs_reclaim irq_context: 0 (wq_completion)bond109 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex fs_reclaim mmu_notifier_invalidate_range_start irq_context: 0 (wq_completion)bond109 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex fs_reclaim mmu_notifier_invalidate_range_start &rq->__lock irq_context: 0 (wq_completion)bond109 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex fs_reclaim mmu_notifier_invalidate_range_start &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)bond109 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &c->lock irq_context: 0 (wq_completion)bond109 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex rcu_read_lock &bond->stats_lock/1 irq_context: 0 (wq_completion)bond109 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex rcu_read_lock &ndev->lock irq_context: 0 (wq_completion)bond109 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex nl_table_lock irq_context: 0 (wq_completion)bond109 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &obj_hash[i].lock irq_context: 0 (wq_completion)bond109 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex nl_table_wait.lock irq_context: 0 (wq_completion)bond109 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex net_rwsem irq_context: 0 (wq_completion)bond109 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex net_rwsem &list->lock#2 irq_context: 0 (wq_completion)bond109 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &tn->lock irq_context: 0 (wq_completion)bond109 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &idev->mc_lock irq_context: 0 (wq_completion)bond109 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &idev->mc_lock fs_reclaim irq_context: 0 (wq_completion)bond109 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &idev->mc_lock fs_reclaim &rq->__lock irq_context: 0 (wq_completion)bond109 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &idev->mc_lock fs_reclaim &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)bond109 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &idev->mc_lock fs_reclaim mmu_notifier_invalidate_range_start irq_context: 0 (wq_completion)bond109 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &idev->mc_lock &c->lock irq_context: 0 (wq_completion)bond109 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &idev->mc_lock rcu_read_lock &ndev->lock irq_context: 0 (wq_completion)bond109 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &idev->mc_lock rcu_read_lock &dir->lock#2 irq_context: 0 (wq_completion)bond109 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &idev->mc_lock rcu_read_lock &ul->lock#2 irq_context: 0 (wq_completion)bond109 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &idev->mc_lock rcu_read_lock rcu_read_lock &obj_hash[i].lock irq_context: 0 (wq_completion)bond109 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &idev->mc_lock rcu_read_lock rcu_read_lock &dir->lock#2 irq_context: 0 (wq_completion)bond109 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &idev->mc_lock rcu_read_lock rcu_read_lock &tbl->lock irq_context: 0 (wq_completion)bond109 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &idev->mc_lock rcu_read_lock rcu_read_lock &n->lock irq_context: 0 (wq_completion)bond109 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &idev->mc_lock rcu_read_lock rcu_read_lock &____s->seqcount#9 irq_context: 0 (wq_completion)bond109 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &idev->mc_lock rcu_read_lock rcu_read_lock rcu_read_lock_bh rcu_read_lock rcu_read_lock_bh rcu_read_lock &br->multicast_lock irq_context: 0 (wq_completion)bond109 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &idev->mc_lock rcu_read_lock rcu_read_lock rcu_read_lock_bh rcu_read_lock rcu_read_lock_bh rcu_read_lock &br->multicast_lock &obj_hash[i].lock irq_context: 0 (wq_completion)bond109 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &idev->mc_lock rcu_read_lock rcu_read_lock rcu_read_lock_bh rcu_read_lock rcu_read_lock_bh rcu_read_lock &br->multicast_lock &base->lock irq_context: 0 (wq_completion)bond109 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &idev->mc_lock rcu_read_lock rcu_read_lock rcu_read_lock_bh rcu_read_lock rcu_read_lock_bh rcu_read_lock &br->multicast_lock &base->lock &obj_hash[i].lock irq_context: 0 (wq_completion)bond109 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &idev->mc_lock rcu_read_lock rcu_read_lock rcu_read_lock_bh rcu_read_lock rcu_read_lock_bh rcu_read_lock &data->lock irq_context: 0 (wq_completion)bond109 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &idev->mc_lock rcu_read_lock rcu_read_lock rcu_read_lock_bh rcu_read_lock rcu_read_lock_bh rcu_read_lock &obj_hash[i].lock irq_context: 0 (wq_completion)bond109 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &idev->mc_lock rcu_read_lock rcu_read_lock rcu_node_0 irq_context: 0 (wq_completion)bond109 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &idev->mc_lock rcu_read_lock rcu_read_lock &rq->__lock irq_context: 0 (wq_completion)bond109 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &idev->mc_lock rcu_read_lock rcu_node_0 irq_context: 0 (wq_completion)bond109 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex.wait_lock irq_context: 0 (wq_completion)bond109 (work_completion)(&(&bond->mcast_work)->work) &p->pi_lock irq_context: 0 (wq_completion)bond110#2 (work_completion)(&(&slave->notify_work)->work) rtnl_mutex net_rwsem irq_context: 0 (wq_completion)bond110#2 (work_completion)(&(&slave->notify_work)->work) rtnl_mutex net_rwsem &list->lock#2 irq_context: 0 (wq_completion)bond110#2 (work_completion)(&(&slave->notify_work)->work) rtnl_mutex &tn->lock irq_context: 0 (wq_completion)bond110#2 (work_completion)(&(&slave->notify_work)->work) rtnl_mutex.wait_lock irq_context: 0 (wq_completion)bond110#2 (work_completion)(&(&slave->notify_work)->work) &p->pi_lock irq_context: 0 (wq_completion)bond154 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex fs_reclaim irq_context: 0 (wq_completion)bond154 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex fs_reclaim mmu_notifier_invalidate_range_start irq_context: 0 (wq_completion)bond154 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &c->lock irq_context: 0 (wq_completion)bond154 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &n->list_lock irq_context: 0 (wq_completion)bond154 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &n->list_lock &c->lock irq_context: 0 (wq_completion)bond154 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex rcu_read_lock &bond->stats_lock/1 irq_context: 0 (wq_completion)bond154 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex rcu_read_lock &ndev->lock irq_context: 0 (wq_completion)bond154 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex rcu_read_lock rcu_node_0 irq_context: 0 (wq_completion)bond154 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex rcu_read_lock &rq->__lock irq_context: 0 (wq_completion)bond154 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex rcu_read_lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)bond154 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex nl_table_lock irq_context: 0 (wq_completion)bond154 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &obj_hash[i].lock irq_context: 0 (wq_completion)bond154 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex nl_table_wait.lock irq_context: 0 (wq_completion)bond154 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex net_rwsem irq_context: 0 (wq_completion)bond154 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex net_rwsem &list->lock#2 irq_context: 0 (wq_completion)bond154 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &tn->lock irq_context: 0 (wq_completion)bond154 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &idev->mc_lock irq_context: 0 (wq_completion)bond154 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &idev->mc_lock fs_reclaim irq_context: 0 (wq_completion)bond154 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &idev->mc_lock fs_reclaim mmu_notifier_invalidate_range_start irq_context: 0 (wq_completion)bond154 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &idev->mc_lock &c->lock irq_context: 0 (wq_completion)bond154 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &idev->mc_lock &rq->__lock irq_context: 0 (wq_completion)bond154 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &idev->mc_lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)bond154 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &idev->mc_lock rcu_read_lock &ndev->lock irq_context: 0 (wq_completion)bond154 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &idev->mc_lock rcu_read_lock &dir->lock#2 irq_context: 0 (wq_completion)bond154 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &idev->mc_lock rcu_read_lock &ul->lock#2 irq_context: 0 (wq_completion)bond154 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &idev->mc_lock rcu_read_lock rcu_read_lock &____s->seqcount#10 irq_context: 0 (wq_completion)bond154 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &idev->mc_lock rcu_read_lock rcu_read_lock rcu_read_lock_bh rcu_read_lock rcu_read_lock_bh rcu_read_lock &br->multicast_lock irq_context: 0 (wq_completion)bond154 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &idev->mc_lock rcu_read_lock rcu_read_lock rcu_read_lock_bh rcu_read_lock rcu_read_lock_bh rcu_read_lock &br->multicast_lock &obj_hash[i].lock irq_context: 0 (wq_completion)bond154 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &idev->mc_lock rcu_read_lock rcu_read_lock rcu_read_lock_bh rcu_read_lock rcu_read_lock_bh rcu_read_lock &br->multicast_lock &base->lock irq_context: 0 (wq_completion)bond154 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &idev->mc_lock rcu_read_lock rcu_read_lock rcu_read_lock_bh rcu_read_lock rcu_read_lock_bh rcu_read_lock &br->multicast_lock &base->lock &obj_hash[i].lock irq_context: 0 (wq_completion)bond154 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &idev->mc_lock rcu_read_lock rcu_read_lock rcu_read_lock_bh rcu_read_lock rcu_read_lock_bh rcu_read_lock &data->lock irq_context: 0 (wq_completion)bond154 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &idev->mc_lock rcu_read_lock rcu_read_lock rcu_read_lock_bh rcu_read_lock rcu_read_lock_bh rcu_read_lock &obj_hash[i].lock irq_context: 0 (wq_completion)bond154 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &idev->mc_lock rcu_read_lock rcu_read_lock rcu_node_0 irq_context: 0 (wq_completion)bond154 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &idev->mc_lock rcu_read_lock rcu_read_lock &rq->__lock irq_context: 0 (wq_completion)bond154 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &idev->mc_lock rcu_read_lock rcu_node_0 irq_context: 0 (wq_completion)bond154 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &idev->mc_lock rcu_read_lock &rcu_state.gp_wq irq_context: 0 (wq_completion)bond154 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &idev->mc_lock rcu_read_lock &rcu_state.gp_wq &p->pi_lock irq_context: 0 (wq_completion)bond154 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &idev->mc_lock rcu_read_lock &rcu_state.gp_wq &p->pi_lock &rq->__lock irq_context: 0 (wq_completion)bond154 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &idev->mc_lock rcu_read_lock &rcu_state.gp_wq &p->pi_lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)bond154 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &idev->mc_lock rcu_read_lock &rq->__lock irq_context: 0 (wq_completion)bond154 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &idev->mc_lock rcu_read_lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)bond154 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &rq->__lock irq_context: 0 (wq_completion)bond154 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)bond154 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex.wait_lock irq_context: 0 (wq_completion)bond154 (work_completion)(&(&bond->mcast_work)->work) &p->pi_lock irq_context: 0 (wq_completion)bond154 (work_completion)(&(&bond->mcast_work)->work) &p->pi_lock &rq->__lock irq_context: 0 (wq_completion)bond154 (work_completion)(&(&bond->mcast_work)->work) &p->pi_lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)bond169 (work_completion)(&(&slave->notify_work)->work) rtnl_mutex net_rwsem irq_context: 0 (wq_completion)bond169 (work_completion)(&(&slave->notify_work)->work) rtnl_mutex net_rwsem &list->lock#2 irq_context: 0 (wq_completion)bond169 (work_completion)(&(&slave->notify_work)->work) rtnl_mutex &tn->lock irq_context: 0 (wq_completion)bond169 (work_completion)(&(&slave->notify_work)->work) rtnl_mutex rcu_read_lock rcu_node_0 irq_context: 0 (wq_completion)bond169 (work_completion)(&(&slave->notify_work)->work) rtnl_mutex rcu_read_lock &rq->__lock irq_context: 0 (wq_completion)bond169 (work_completion)(&(&slave->notify_work)->work) rtnl_mutex rcu_read_lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)bond169 (work_completion)(&(&slave->notify_work)->work) rtnl_mutex.wait_lock irq_context: 0 (wq_completion)bond169 (work_completion)(&(&slave->notify_work)->work) &p->pi_lock irq_context: 0 (wq_completion)bond160 (work_completion)(&(&bond->mcast_work)->work) &rq->__lock irq_context: 0 (wq_completion)bond160 (work_completion)(&(&bond->mcast_work)->work) &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)bond165#2 irq_context: 0 (wq_completion)bond165#2 (work_completion)(&(&slave->notify_work)->work) irq_context: 0 (wq_completion)bond165#2 (work_completion)(&(&slave->notify_work)->work) &rq->__lock irq_context: 0 (wq_completion)bond165#2 (work_completion)(&(&slave->notify_work)->work) &obj_hash[i].lock irq_context: 0 (wq_completion)bond165#2 (work_completion)(&(&slave->notify_work)->work) &base->lock irq_context: 0 (wq_completion)bond165#2 (work_completion)(&(&slave->notify_work)->work) &base->lock &obj_hash[i].lock irq_context: 0 (wq_completion)bond165#2 (work_completion)(&(&slave->notify_work)->work) &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)bond165#2 (work_completion)(&(&bond->mcast_work)->work) irq_context: 0 (wq_completion)bond165#2 (work_completion)(&(&bond->mcast_work)->work) &obj_hash[i].lock irq_context: 0 (wq_completion)bond165#2 (work_completion)(&(&bond->mcast_work)->work) &base->lock irq_context: 0 (wq_completion)bond165#2 (work_completion)(&(&bond->mcast_work)->work) &base->lock &obj_hash[i].lock irq_context: 0 (wq_completion)bond165#2 (work_completion)(&(&bond->mcast_work)->work) &rq->__lock irq_context: 0 (wq_completion)bond165#2 (work_completion)(&(&bond->mcast_work)->work) &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)bond136#2 (work_completion)(&(&slave->notify_work)->work) &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)bond149#3 irq_context: 0 (wq_completion)bond149#3 &rq->__lock irq_context: 0 (wq_completion)bond149#3 &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)bond149#3 (work_completion)(&(&slave->notify_work)->work) irq_context: 0 (wq_completion)bond149#3 (work_completion)(&(&slave->notify_work)->work) &obj_hash[i].lock irq_context: 0 (wq_completion)bond149#3 (work_completion)(&(&slave->notify_work)->work) &base->lock irq_context: 0 (wq_completion)bond149#3 (work_completion)(&(&slave->notify_work)->work) &base->lock &obj_hash[i].lock irq_context: 0 (wq_completion)bond149#3 (work_completion)(&(&slave->notify_work)->work) &rq->__lock irq_context: 0 (wq_completion)bond149#3 (work_completion)(&(&slave->notify_work)->work) &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)bond149#3 (work_completion)(&(&bond->mcast_work)->work) irq_context: 0 (wq_completion)bond149#3 (work_completion)(&(&bond->mcast_work)->work) &obj_hash[i].lock irq_context: 0 (wq_completion)bond149#3 (work_completion)(&(&bond->mcast_work)->work) &base->lock irq_context: 0 (wq_completion)bond149#3 (work_completion)(&(&bond->mcast_work)->work) &base->lock &obj_hash[i].lock irq_context: 0 (wq_completion)bond149#3 (work_completion)(&(&bond->mcast_work)->work) &rq->__lock irq_context: 0 (wq_completion)bond149#3 (work_completion)(&(&bond->mcast_work)->work) &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)bond167 &rq->__lock irq_context: 0 (wq_completion)bond167 &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)bond141#2 (work_completion)(&(&bond->mcast_work)->work) &rq->__lock irq_context: 0 (wq_completion)bond141#2 (work_completion)(&(&bond->mcast_work)->work) &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)bond145#2 (work_completion)(&(&slave->notify_work)->work) rtnl_mutex net_rwsem irq_context: 0 (wq_completion)bond145#2 (work_completion)(&(&slave->notify_work)->work) rtnl_mutex net_rwsem &list->lock#2 irq_context: 0 (wq_completion)bond145#2 (work_completion)(&(&slave->notify_work)->work) rtnl_mutex &tn->lock irq_context: 0 (wq_completion)bond145#2 (work_completion)(&(&slave->notify_work)->work) rtnl_mutex.wait_lock irq_context: 0 (wq_completion)bond145#2 (work_completion)(&(&slave->notify_work)->work) &p->pi_lock irq_context: 0 (wq_completion)bond126#2 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex fs_reclaim irq_context: 0 (wq_completion)bond126#2 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex fs_reclaim mmu_notifier_invalidate_range_start irq_context: 0 (wq_completion)bond126#2 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &c->lock irq_context: 0 (wq_completion)bond126#2 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex rcu_read_lock &bond->stats_lock/1 irq_context: 0 (wq_completion)bond126#2 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex rcu_read_lock &ndev->lock irq_context: 0 (wq_completion)bond126#2 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex nl_table_lock irq_context: 0 (wq_completion)bond126#2 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &obj_hash[i].lock irq_context: 0 (wq_completion)bond126#2 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex nl_table_wait.lock irq_context: 0 (wq_completion)bond126#2 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex net_rwsem irq_context: 0 (wq_completion)bond126#2 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex net_rwsem &list->lock#2 irq_context: 0 (wq_completion)bond126#2 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &tn->lock irq_context: 0 (wq_completion)bond126#2 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &idev->mc_lock irq_context: 0 (wq_completion)bond126#2 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &idev->mc_lock fs_reclaim irq_context: 0 (wq_completion)bond126#2 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &idev->mc_lock fs_reclaim mmu_notifier_invalidate_range_start irq_context: 0 (wq_completion)bond126#2 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &idev->mc_lock &c->lock irq_context: 0 (wq_completion)bond126#2 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &idev->mc_lock rcu_read_lock &ndev->lock irq_context: 0 (wq_completion)bond126#2 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &idev->mc_lock rcu_read_lock &c->lock irq_context: 0 (wq_completion)bond126#2 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &idev->mc_lock rcu_read_lock &dir->lock#2 irq_context: 0 (wq_completion)bond126#2 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &idev->mc_lock rcu_read_lock &ul->lock#2 irq_context: 0 (wq_completion)bond126#2 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &idev->mc_lock rcu_read_lock rcu_read_lock &____s->seqcount#10 irq_context: 0 (wq_completion)bond126#2 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &idev->mc_lock rcu_read_lock rcu_read_lock rcu_read_lock_bh rcu_read_lock rcu_read_lock_bh rcu_read_lock &br->multicast_lock irq_context: 0 (wq_completion)bond126#2 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &idev->mc_lock rcu_read_lock rcu_read_lock rcu_read_lock_bh rcu_read_lock rcu_read_lock_bh rcu_read_lock &br->multicast_lock &obj_hash[i].lock irq_context: 0 (wq_completion)bond126#2 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &idev->mc_lock rcu_read_lock rcu_read_lock rcu_read_lock_bh rcu_read_lock rcu_read_lock_bh rcu_read_lock &br->multicast_lock &base->lock irq_context: 0 (wq_completion)bond126#2 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &idev->mc_lock rcu_read_lock rcu_read_lock rcu_read_lock_bh rcu_read_lock rcu_read_lock_bh rcu_read_lock &br->multicast_lock &base->lock &obj_hash[i].lock irq_context: 0 (wq_completion)bond126#2 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &idev->mc_lock rcu_read_lock rcu_read_lock rcu_read_lock_bh rcu_read_lock rcu_read_lock_bh rcu_read_lock &data->lock irq_context: 0 (wq_completion)bond126#2 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &idev->mc_lock rcu_read_lock rcu_read_lock rcu_read_lock_bh rcu_read_lock rcu_read_lock_bh rcu_read_lock &obj_hash[i].lock irq_context: 0 (wq_completion)bond126#2 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex.wait_lock irq_context: 0 (wq_completion)bond126#2 (work_completion)(&(&bond->mcast_work)->work) &p->pi_lock irq_context: 0 (wq_completion)bond114#3 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex fs_reclaim irq_context: 0 (wq_completion)bond114#3 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex fs_reclaim mmu_notifier_invalidate_range_start irq_context: 0 (wq_completion)bond114#3 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &c->lock irq_context: 0 (wq_completion)bond114#3 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &n->list_lock irq_context: 0 (wq_completion)bond114#3 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &n->list_lock &c->lock irq_context: 0 (wq_completion)bond114#3 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex rcu_read_lock &bond->stats_lock/1 irq_context: 0 (wq_completion)bond114#3 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex rcu_read_lock &ndev->lock irq_context: 0 (wq_completion)bond114#3 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex nl_table_lock irq_context: 0 (wq_completion)bond114#3 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &obj_hash[i].lock irq_context: 0 (wq_completion)bond114#3 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex nl_table_wait.lock irq_context: 0 (wq_completion)bond114#3 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex net_rwsem irq_context: 0 (wq_completion)bond114#3 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex net_rwsem &list->lock#2 irq_context: 0 (wq_completion)bond114#3 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &tn->lock irq_context: 0 (wq_completion)bond114#3 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &idev->mc_lock irq_context: 0 (wq_completion)bond114#3 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &idev->mc_lock fs_reclaim irq_context: 0 (wq_completion)bond114#3 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &idev->mc_lock fs_reclaim mmu_notifier_invalidate_range_start irq_context: 0 (wq_completion)bond114#3 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &idev->mc_lock &c->lock irq_context: 0 (wq_completion)bond114#3 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &idev->mc_lock &n->list_lock irq_context: 0 (wq_completion)bond114#3 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &idev->mc_lock &n->list_lock &c->lock irq_context: 0 (wq_completion)bond114#3 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &idev->mc_lock rcu_read_lock &ndev->lock irq_context: 0 (wq_completion)bond114#3 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &idev->mc_lock rcu_read_lock &c->lock irq_context: 0 (wq_completion)bond114#3 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &idev->mc_lock rcu_read_lock &dir->lock#2 irq_context: 0 (wq_completion)bond114#3 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &idev->mc_lock rcu_read_lock &ul->lock#2 irq_context: 0 (wq_completion)bond114#3 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &idev->mc_lock rcu_read_lock rcu_read_lock &____s->seqcount#10 irq_context: 0 (wq_completion)bond114#3 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &idev->mc_lock rcu_read_lock rcu_read_lock rcu_read_lock_bh rcu_read_lock rcu_read_lock_bh rcu_read_lock &br->multicast_lock irq_context: 0 (wq_completion)bond114#3 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &idev->mc_lock rcu_read_lock rcu_read_lock rcu_read_lock_bh rcu_read_lock rcu_read_lock_bh rcu_read_lock &br->multicast_lock &obj_hash[i].lock irq_context: 0 (wq_completion)bond114#3 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &idev->mc_lock rcu_read_lock rcu_read_lock rcu_read_lock_bh rcu_read_lock rcu_read_lock_bh rcu_read_lock &br->multicast_lock &base->lock irq_context: 0 (wq_completion)bond114#3 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &idev->mc_lock rcu_read_lock rcu_read_lock rcu_read_lock_bh rcu_read_lock rcu_read_lock_bh rcu_read_lock &br->multicast_lock &base->lock &obj_hash[i].lock irq_context: 0 (wq_completion)bond114#3 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &idev->mc_lock rcu_read_lock rcu_read_lock rcu_read_lock_bh rcu_read_lock rcu_read_lock_bh rcu_read_lock &data->lock irq_context: 0 (wq_completion)bond114#3 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &idev->mc_lock rcu_read_lock rcu_read_lock rcu_read_lock_bh rcu_read_lock rcu_read_lock_bh rcu_read_lock &obj_hash[i].lock irq_context: 0 (wq_completion)bond114#3 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &rq->__lock irq_context: 0 (wq_completion)bond114#3 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)bond114#3 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex.wait_lock irq_context: 0 (wq_completion)bond114#3 (work_completion)(&(&bond->mcast_work)->work) &p->pi_lock irq_context: 0 (wq_completion)bond114#3 (work_completion)(&(&bond->mcast_work)->work) &p->pi_lock &cfs_rq->removed.lock irq_context: 0 (wq_completion)bond114#3 (work_completion)(&(&bond->mcast_work)->work) &p->pi_lock &rq->__lock irq_context: 0 (wq_completion)bond114#3 (work_completion)(&(&bond->mcast_work)->work) &p->pi_lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)bond178 irq_context: 0 (wq_completion)bond178 &rq->__lock irq_context: 0 (wq_completion)bond178 &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)bond178 (work_completion)(&(&slave->notify_work)->work) irq_context: 0 (wq_completion)bond178 (work_completion)(&(&slave->notify_work)->work) &obj_hash[i].lock irq_context: 0 (wq_completion)bond178 (work_completion)(&(&slave->notify_work)->work) &base->lock irq_context: 0 (wq_completion)bond178 (work_completion)(&(&slave->notify_work)->work) &base->lock &obj_hash[i].lock irq_context: 0 (wq_completion)bond162#2 (work_completion)(&(&slave->notify_work)->work) &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)bond158#2 &rq->__lock irq_context: 0 (wq_completion)bond158#2 &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)bond178 (work_completion)(&(&bond->mcast_work)->work) irq_context: 0 (wq_completion)bond178 (work_completion)(&(&bond->mcast_work)->work) &obj_hash[i].lock irq_context: 0 (wq_completion)bond178 (work_completion)(&(&bond->mcast_work)->work) &base->lock irq_context: 0 (wq_completion)bond178 (work_completion)(&(&bond->mcast_work)->work) &base->lock &obj_hash[i].lock irq_context: 0 (wq_completion)bond178 (work_completion)(&(&bond->mcast_work)->work) &rq->__lock irq_context: 0 (wq_completion)bond178 (work_completion)(&(&bond->mcast_work)->work) &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)bond51#3 irq_context: 0 (wq_completion)bond51#3 &rq->__lock irq_context: 0 (wq_completion)bond51#3 &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)bond51#3 (work_completion)(&(&slave->notify_work)->work) irq_context: 0 (wq_completion)bond51#3 (work_completion)(&(&slave->notify_work)->work) &obj_hash[i].lock irq_context: 0 (wq_completion)bond51#3 (work_completion)(&(&slave->notify_work)->work) &base->lock irq_context: 0 (wq_completion)bond51#3 (work_completion)(&(&slave->notify_work)->work) &base->lock &obj_hash[i].lock irq_context: 0 (wq_completion)bond51#3 (work_completion)(&(&slave->notify_work)->work) &rq->__lock irq_context: 0 (wq_completion)bond51#3 (work_completion)(&(&slave->notify_work)->work) &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)bond147#3 &rq->__lock irq_context: 0 (wq_completion)bond147#3 &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)bond122#3 (work_completion)(&(&slave->notify_work)->work) fill_pool_map-wait-type-override pool_lock irq_context: 0 (wq_completion)bond51#3 (work_completion)(&(&bond->mcast_work)->work) irq_context: 0 (wq_completion)bond51#3 (work_completion)(&(&bond->mcast_work)->work) &obj_hash[i].lock irq_context: 0 (wq_completion)bond51#3 (work_completion)(&(&bond->mcast_work)->work) &base->lock irq_context: 0 (wq_completion)bond51#3 (work_completion)(&(&bond->mcast_work)->work) &base->lock &obj_hash[i].lock irq_context: 0 (wq_completion)bond51#3 (work_completion)(&(&bond->mcast_work)->work) &rq->__lock irq_context: 0 (wq_completion)bond51#3 (work_completion)(&(&bond->mcast_work)->work) &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)bond107#3 &rq->__lock irq_context: 0 (wq_completion)bond107#3 &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)bond166#2 irq_context: 0 (wq_completion)bond166#2 &rq->__lock irq_context: 0 (wq_completion)bond166#2 (work_completion)(&(&slave->notify_work)->work) irq_context: 0 (wq_completion)bond166#2 (work_completion)(&(&slave->notify_work)->work) &obj_hash[i].lock irq_context: 0 (wq_completion)bond166#2 (work_completion)(&(&slave->notify_work)->work) &base->lock irq_context: 0 (wq_completion)bond166#2 (work_completion)(&(&slave->notify_work)->work) &base->lock &obj_hash[i].lock irq_context: 0 (wq_completion)bond166#2 (work_completion)(&(&slave->notify_work)->work) &rq->__lock irq_context: 0 (wq_completion)bond166#2 (work_completion)(&(&slave->notify_work)->work) &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)bond166#2 (work_completion)(&(&bond->mcast_work)->work) irq_context: 0 (wq_completion)bond166#2 (work_completion)(&(&bond->mcast_work)->work) &obj_hash[i].lock irq_context: 0 (wq_completion)bond166#2 (work_completion)(&(&bond->mcast_work)->work) &base->lock irq_context: 0 (wq_completion)bond166#2 (work_completion)(&(&bond->mcast_work)->work) &base->lock &obj_hash[i].lock irq_context: 0 (wq_completion)bond166#2 (work_completion)(&(&bond->mcast_work)->work) &rq->__lock irq_context: 0 (wq_completion)bond166#2 (work_completion)(&(&bond->mcast_work)->work) &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 rtnl_mutex fs_reclaim mmu_notifier_invalidate_range_start &obj_hash[i].lock irq_context: 0 (wq_completion)bond137#3 &rq->__lock irq_context: 0 (wq_completion)bond137#3 &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)bond144#3 (work_completion)(&(&bond->mcast_work)->work) &rq->__lock irq_context: 0 (wq_completion)bond144#3 (work_completion)(&(&bond->mcast_work)->work) &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)bond123#3 &rq->__lock irq_context: 0 (wq_completion)bond123#3 &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)bond150#3 irq_context: 0 (wq_completion)bond150#3 &rq->__lock irq_context: 0 (wq_completion)bond150#3 &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)bond150#3 (work_completion)(&(&slave->notify_work)->work) irq_context: 0 (wq_completion)bond150#3 (work_completion)(&(&slave->notify_work)->work) &obj_hash[i].lock irq_context: 0 (wq_completion)bond150#3 (work_completion)(&(&slave->notify_work)->work) &base->lock irq_context: 0 (wq_completion)bond150#3 (work_completion)(&(&slave->notify_work)->work) &base->lock &obj_hash[i].lock irq_context: 0 (wq_completion)bond146#3 (work_completion)(&(&bond->mcast_work)->work) &rq->__lock irq_context: 0 (wq_completion)bond146#3 (work_completion)(&(&bond->mcast_work)->work) &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)bond150#3 (work_completion)(&(&bond->mcast_work)->work) irq_context: 0 (wq_completion)bond150#3 (work_completion)(&(&bond->mcast_work)->work) &obj_hash[i].lock irq_context: 0 (wq_completion)bond150#3 (work_completion)(&(&bond->mcast_work)->work) &base->lock irq_context: 0 (wq_completion)bond150#3 (work_completion)(&(&bond->mcast_work)->work) &base->lock &obj_hash[i].lock irq_context: 0 (wq_completion)bond150#3 (work_completion)(&(&bond->mcast_work)->work) &rq->__lock irq_context: 0 (wq_completion)bond131 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex fs_reclaim irq_context: 0 (wq_completion)bond131 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex fs_reclaim mmu_notifier_invalidate_range_start irq_context: 0 (wq_completion)bond131 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &c->lock irq_context: 0 (wq_completion)bond131 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &n->list_lock irq_context: 0 (wq_completion)bond131 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &n->list_lock &c->lock irq_context: 0 (wq_completion)bond131 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex rcu_read_lock &bond->stats_lock/1 irq_context: 0 (wq_completion)bond131 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex rcu_read_lock &ndev->lock irq_context: 0 (wq_completion)bond131 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex nl_table_lock irq_context: 0 (wq_completion)bond131 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &obj_hash[i].lock irq_context: 0 (wq_completion)bond131 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex nl_table_wait.lock irq_context: 0 (wq_completion)bond131 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex net_rwsem irq_context: 0 (wq_completion)bond131 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex net_rwsem &list->lock#2 irq_context: 0 (wq_completion)bond131 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &tn->lock irq_context: 0 (wq_completion)bond131 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &idev->mc_lock irq_context: 0 (wq_completion)bond131 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &idev->mc_lock fs_reclaim irq_context: 0 (wq_completion)bond131 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &idev->mc_lock fs_reclaim mmu_notifier_invalidate_range_start irq_context: 0 (wq_completion)bond131 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &idev->mc_lock rcu_read_lock &ndev->lock irq_context: 0 (wq_completion)bond131 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &idev->mc_lock rcu_read_lock &dir->lock#2 irq_context: 0 (wq_completion)bond131 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &idev->mc_lock rcu_read_lock &ul->lock#2 irq_context: 0 (wq_completion)bond131 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &idev->mc_lock rcu_read_lock rcu_read_lock &obj_hash[i].lock irq_context: 0 (wq_completion)bond131 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &idev->mc_lock rcu_read_lock rcu_read_lock &dir->lock#2 irq_context: 0 (wq_completion)bond131 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &idev->mc_lock rcu_read_lock rcu_read_lock &tbl->lock irq_context: 0 (wq_completion)bond131 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &idev->mc_lock rcu_read_lock rcu_read_lock &n->lock irq_context: 0 (wq_completion)bond131 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &idev->mc_lock rcu_read_lock rcu_read_lock &____s->seqcount#9 irq_context: 0 (wq_completion)bond131 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &idev->mc_lock rcu_read_lock rcu_read_lock rcu_read_lock_bh rcu_read_lock rcu_read_lock_bh rcu_read_lock &br->multicast_lock irq_context: 0 (wq_completion)bond131 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &idev->mc_lock rcu_read_lock rcu_read_lock rcu_read_lock_bh rcu_read_lock rcu_read_lock_bh rcu_read_lock &br->multicast_lock &obj_hash[i].lock irq_context: 0 (wq_completion)bond131 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &idev->mc_lock rcu_read_lock rcu_read_lock rcu_read_lock_bh rcu_read_lock rcu_read_lock_bh rcu_read_lock &br->multicast_lock &base->lock irq_context: 0 (wq_completion)bond131 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &idev->mc_lock rcu_read_lock rcu_read_lock rcu_read_lock_bh rcu_read_lock rcu_read_lock_bh rcu_read_lock &br->multicast_lock &base->lock &obj_hash[i].lock irq_context: 0 (wq_completion)bond131 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &idev->mc_lock rcu_read_lock rcu_read_lock rcu_read_lock_bh rcu_read_lock rcu_read_lock_bh rcu_read_lock &data->lock irq_context: 0 (wq_completion)bond131 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &idev->mc_lock rcu_read_lock rcu_read_lock rcu_read_lock_bh rcu_read_lock rcu_read_lock_bh rcu_read_lock &obj_hash[i].lock irq_context: 0 (wq_completion)bond131 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex.wait_lock irq_context: 0 (wq_completion)bond131 (work_completion)(&(&bond->mcast_work)->work) &p->pi_lock irq_context: 0 (wq_completion)bond170 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex fs_reclaim irq_context: 0 (wq_completion)bond170 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex fs_reclaim mmu_notifier_invalidate_range_start irq_context: 0 (wq_completion)bond170 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex rcu_read_lock &bond->stats_lock/1 irq_context: 0 (wq_completion)bond170 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex rcu_read_lock &ndev->lock irq_context: 0 (wq_completion)bond170 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex nl_table_lock irq_context: 0 (wq_completion)bond170 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &obj_hash[i].lock irq_context: 0 (wq_completion)bond170 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex nl_table_wait.lock irq_context: 0 (wq_completion)bond170 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex net_rwsem irq_context: 0 (wq_completion)bond170 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex net_rwsem &list->lock#2 irq_context: 0 (wq_completion)bond170 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &tn->lock irq_context: 0 (wq_completion)bond170 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &idev->mc_lock irq_context: 0 (wq_completion)bond170 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &idev->mc_lock fs_reclaim irq_context: 0 (wq_completion)bond170 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &idev->mc_lock fs_reclaim mmu_notifier_invalidate_range_start irq_context: 0 (wq_completion)bond170 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &idev->mc_lock rcu_read_lock &ndev->lock irq_context: 0 (wq_completion)bond170 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &idev->mc_lock rcu_read_lock &dir->lock#2 irq_context: 0 (wq_completion)bond170 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &idev->mc_lock rcu_read_lock rcu_node_0 irq_context: 0 (wq_completion)bond170 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &idev->mc_lock rcu_read_lock &rq->__lock irq_context: 0 (wq_completion)bond170 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &idev->mc_lock rcu_read_lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)bond140#3 (work_completion)(&(&slave->notify_work)->work) &base->lock fill_pool_map-wait-type-override pool_lock irq_context: 0 (wq_completion)bond170 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &idev->mc_lock rcu_read_lock &ul->lock#2 irq_context: 0 (wq_completion)bond170 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &idev->mc_lock rcu_read_lock rcu_read_lock &____s->seqcount#10 irq_context: 0 (wq_completion)bond170 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &idev->mc_lock rcu_read_lock rcu_read_lock rcu_read_lock_bh rcu_read_lock rcu_read_lock_bh rcu_read_lock &br->multicast_lock irq_context: 0 (wq_completion)bond170 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &idev->mc_lock rcu_read_lock rcu_read_lock rcu_read_lock_bh rcu_read_lock rcu_read_lock_bh rcu_read_lock &br->multicast_lock &obj_hash[i].lock irq_context: 0 (wq_completion)bond170 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &idev->mc_lock rcu_read_lock rcu_read_lock rcu_read_lock_bh rcu_read_lock rcu_read_lock_bh rcu_read_lock &br->multicast_lock &base->lock irq_context: 0 (wq_completion)bond170 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &idev->mc_lock rcu_read_lock rcu_read_lock rcu_read_lock_bh rcu_read_lock rcu_read_lock_bh rcu_read_lock &data->lock irq_context: 0 (wq_completion)bond170 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &idev->mc_lock rcu_read_lock rcu_read_lock rcu_read_lock_bh rcu_read_lock rcu_read_lock_bh rcu_read_lock &obj_hash[i].lock irq_context: 0 (wq_completion)bond170 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex.wait_lock irq_context: 0 (wq_completion)bond170 (work_completion)(&(&bond->mcast_work)->work) &p->pi_lock irq_context: 0 (wq_completion)bond170 (work_completion)(&(&bond->mcast_work)->work) &p->pi_lock &rq->__lock irq_context: 0 (wq_completion)bond170 (work_completion)(&(&bond->mcast_work)->work) &p->pi_lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)bond52#3 irq_context: 0 (wq_completion)bond52#3 (work_completion)(&(&slave->notify_work)->work) irq_context: 0 (wq_completion)bond52#3 (work_completion)(&(&slave->notify_work)->work) &obj_hash[i].lock irq_context: 0 (wq_completion)bond52#3 (work_completion)(&(&slave->notify_work)->work) &base->lock irq_context: 0 (wq_completion)bond52#3 (work_completion)(&(&slave->notify_work)->work) &base->lock &obj_hash[i].lock irq_context: 0 (wq_completion)bond52#3 (work_completion)(&(&slave->notify_work)->work) &rq->__lock irq_context: 0 (wq_completion)bond52#3 (work_completion)(&(&slave->notify_work)->work) &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)bond52#3 (work_completion)(&(&bond->mcast_work)->work) irq_context: 0 (wq_completion)bond52#3 (work_completion)(&(&bond->mcast_work)->work) &obj_hash[i].lock irq_context: 0 (wq_completion)bond52#3 (work_completion)(&(&bond->mcast_work)->work) &base->lock irq_context: 0 (wq_completion)bond52#3 (work_completion)(&(&bond->mcast_work)->work) &base->lock &obj_hash[i].lock irq_context: 0 (wq_completion)bond52#3 (work_completion)(&(&bond->mcast_work)->work) &rq->__lock irq_context: 0 (wq_completion)bond52#3 (work_completion)(&(&bond->mcast_work)->work) &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)bond153#2 (work_completion)(&(&slave->notify_work)->work) &rq->__lock irq_context: 0 (wq_completion)bond153#2 (work_completion)(&(&slave->notify_work)->work) &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)bond179 irq_context: 0 (wq_completion)bond179 (work_completion)(&(&slave->notify_work)->work) irq_context: 0 (wq_completion)bond179 (work_completion)(&(&slave->notify_work)->work) &obj_hash[i].lock irq_context: 0 (wq_completion)bond179 (work_completion)(&(&slave->notify_work)->work) &base->lock irq_context: 0 (wq_completion)bond179 (work_completion)(&(&slave->notify_work)->work) &base->lock &obj_hash[i].lock irq_context: 0 (wq_completion)bond179 (work_completion)(&(&slave->notify_work)->work) &rq->__lock irq_context: 0 (wq_completion)bond179 (work_completion)(&(&slave->notify_work)->work) &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)bond179 (work_completion)(&(&bond->mcast_work)->work) irq_context: 0 (wq_completion)bond179 (work_completion)(&(&bond->mcast_work)->work) &obj_hash[i].lock irq_context: 0 (wq_completion)bond179 (work_completion)(&(&bond->mcast_work)->work) &base->lock irq_context: 0 (wq_completion)bond179 (work_completion)(&(&bond->mcast_work)->work) &base->lock &obj_hash[i].lock irq_context: 0 (wq_completion)bond138#2 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex fs_reclaim irq_context: 0 (wq_completion)bond179 (work_completion)(&(&bond->mcast_work)->work) &rq->__lock irq_context: 0 (wq_completion)bond138#2 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex fs_reclaim mmu_notifier_invalidate_range_start irq_context: 0 (wq_completion)bond179 (work_completion)(&(&bond->mcast_work)->work) &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)bond138#2 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &c->lock irq_context: 0 (wq_completion)bond138#2 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &rq->__lock irq_context: 0 (wq_completion)bond138#2 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)bond138#2 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex rcu_read_lock &bond->stats_lock/1 irq_context: 0 (wq_completion)bond138#2 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex rcu_read_lock &ndev->lock irq_context: 0 (wq_completion)bond138#2 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex rcu_read_lock rcu_node_0 irq_context: 0 (wq_completion)bond138#2 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex rcu_read_lock &rq->__lock irq_context: 0 (wq_completion)bond138#2 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex rcu_read_lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)bond138#2 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex nl_table_lock irq_context: 0 (wq_completion)bond138#2 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &obj_hash[i].lock irq_context: 0 (wq_completion)bond138#2 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex nl_table_wait.lock irq_context: 0 (wq_completion)bond138#2 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex net_rwsem irq_context: 0 (wq_completion)bond138#2 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex net_rwsem &list->lock#2 irq_context: 0 (wq_completion)bond138#2 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &tn->lock irq_context: 0 (wq_completion)bond138#2 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &idev->mc_lock irq_context: 0 (wq_completion)bond138#2 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &idev->mc_lock fs_reclaim irq_context: 0 (wq_completion)bond138#2 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &idev->mc_lock fs_reclaim mmu_notifier_invalidate_range_start irq_context: 0 (wq_completion)bond138#2 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &idev->mc_lock fs_reclaim mmu_notifier_invalidate_range_start &rq->__lock irq_context: 0 (wq_completion)bond138#2 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &idev->mc_lock fs_reclaim mmu_notifier_invalidate_range_start &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)bond138#2 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &idev->mc_lock &c->lock irq_context: 0 (wq_completion)bond138#2 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &idev->mc_lock rcu_read_lock &ndev->lock irq_context: 0 (wq_completion)bond138#2 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &idev->mc_lock rcu_read_lock rcu_node_0 irq_context: 0 (wq_completion)bond138#2 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &idev->mc_lock rcu_read_lock &rq->__lock irq_context: 0 (wq_completion)bond138#2 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &idev->mc_lock rcu_read_lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)bond138#2 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &idev->mc_lock rcu_read_lock &dir->lock#2 irq_context: 0 (wq_completion)bond138#2 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &idev->mc_lock rcu_read_lock &ul->lock#2 irq_context: 0 (wq_completion)bond138#2 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &idev->mc_lock rcu_read_lock rcu_read_lock &____s->seqcount#10 irq_context: 0 (wq_completion)bond138#2 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &idev->mc_lock rcu_read_lock rcu_read_lock rcu_read_lock_bh rcu_read_lock rcu_read_lock_bh rcu_read_lock &br->multicast_lock irq_context: 0 (wq_completion)bond138#2 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &idev->mc_lock rcu_read_lock rcu_read_lock rcu_read_lock_bh rcu_read_lock rcu_read_lock_bh rcu_read_lock &br->multicast_lock &obj_hash[i].lock irq_context: 0 (wq_completion)bond138#2 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &idev->mc_lock rcu_read_lock rcu_read_lock rcu_read_lock_bh rcu_read_lock rcu_read_lock_bh rcu_read_lock &br->multicast_lock &base->lock irq_context: 0 (wq_completion)bond138#2 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &idev->mc_lock rcu_read_lock rcu_read_lock rcu_read_lock_bh rcu_read_lock rcu_read_lock_bh rcu_read_lock &data->lock irq_context: 0 (wq_completion)bond138#2 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &idev->mc_lock rcu_read_lock rcu_read_lock rcu_read_lock_bh rcu_read_lock rcu_read_lock_bh rcu_read_lock &obj_hash[i].lock irq_context: 0 (wq_completion)bond138#2 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex.wait_lock irq_context: 0 (wq_completion)bond138#2 (work_completion)(&(&bond->mcast_work)->work) &p->pi_lock irq_context: 0 (wq_completion)bond138#2 (work_completion)(&(&bond->mcast_work)->work) &p->pi_lock &rq->__lock irq_context: 0 (wq_completion)bond138#2 (work_completion)(&(&bond->mcast_work)->work) &p->pi_lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)bond144#3 &rq->__lock irq_context: 0 (wq_completion)bond167#2 irq_context: 0 (wq_completion)bond167#2 (work_completion)(&(&slave->notify_work)->work) irq_context: 0 (wq_completion)bond167#2 (work_completion)(&(&slave->notify_work)->work) &obj_hash[i].lock irq_context: 0 (wq_completion)bond167#2 (work_completion)(&(&slave->notify_work)->work) &base->lock irq_context: 0 (wq_completion)bond167#2 (work_completion)(&(&slave->notify_work)->work) &base->lock &obj_hash[i].lock irq_context: 0 (wq_completion)bond167#2 (work_completion)(&(&bond->mcast_work)->work) irq_context: 0 (wq_completion)bond167#2 (work_completion)(&(&bond->mcast_work)->work) &rq->__lock irq_context: 0 (wq_completion)bond167#2 (work_completion)(&(&bond->mcast_work)->work) &obj_hash[i].lock irq_context: 0 (wq_completion)bond167#2 (work_completion)(&(&bond->mcast_work)->work) &base->lock irq_context: 0 (wq_completion)bond167#2 (work_completion)(&(&bond->mcast_work)->work) &base->lock &obj_hash[i].lock irq_context: 0 (wq_completion)bond152 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex fs_reclaim irq_context: 0 (wq_completion)bond152 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex fs_reclaim mmu_notifier_invalidate_range_start irq_context: 0 (wq_completion)bond152 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex rcu_read_lock &bond->stats_lock/1 irq_context: 0 (wq_completion)bond152 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex rcu_node_0 irq_context: 0 (wq_completion)bond152 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &rq->__lock irq_context: 0 (wq_completion)bond152 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)bond152 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex rcu_read_lock rcu_node_0 irq_context: 0 (wq_completion)bond152 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex rcu_read_lock &ndev->lock irq_context: 0 (wq_completion)bond152 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex nl_table_lock irq_context: 0 (wq_completion)bond152 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &obj_hash[i].lock irq_context: 0 (wq_completion)bond152 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex nl_table_wait.lock irq_context: 0 (wq_completion)bond152 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex net_rwsem irq_context: 0 (wq_completion)bond152 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex net_rwsem &list->lock#2 irq_context: 0 (wq_completion)bond152 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &tn->lock irq_context: 0 (wq_completion)bond152 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &idev->mc_lock irq_context: 0 (wq_completion)bond152 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &idev->mc_lock fs_reclaim irq_context: 0 (wq_completion)bond152 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &idev->mc_lock fs_reclaim &rq->__lock irq_context: 0 (wq_completion)bond152 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &idev->mc_lock fs_reclaim &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)bond152 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &idev->mc_lock fs_reclaim mmu_notifier_invalidate_range_start irq_context: 0 (wq_completion)bond152 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &idev->mc_lock rcu_read_lock &ndev->lock irq_context: 0 (wq_completion)bond152 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &idev->mc_lock rcu_read_lock &dir->lock#2 irq_context: 0 (wq_completion)bond152 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &idev->mc_lock rcu_read_lock &c->lock irq_context: 0 (wq_completion)bond152 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &idev->mc_lock rcu_read_lock &ul->lock#2 irq_context: 0 (wq_completion)bond152 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &idev->mc_lock rcu_read_lock rcu_read_lock &____s->seqcount#10 irq_context: 0 (wq_completion)bond152 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &idev->mc_lock rcu_read_lock rcu_read_lock rcu_read_lock_bh rcu_read_lock rcu_read_lock_bh rcu_read_lock &br->multicast_lock irq_context: 0 (wq_completion)bond152 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &idev->mc_lock rcu_read_lock rcu_read_lock rcu_read_lock_bh rcu_read_lock rcu_read_lock_bh rcu_read_lock &br->multicast_lock &obj_hash[i].lock irq_context: 0 (wq_completion)bond152 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &idev->mc_lock rcu_read_lock rcu_read_lock rcu_read_lock_bh rcu_read_lock rcu_read_lock_bh rcu_read_lock &br->multicast_lock &base->lock irq_context: 0 (wq_completion)bond152 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &idev->mc_lock rcu_read_lock rcu_read_lock rcu_read_lock_bh rcu_read_lock rcu_read_lock_bh rcu_read_lock &br->multicast_lock &base->lock &obj_hash[i].lock irq_context: 0 (wq_completion)bond152 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &idev->mc_lock rcu_read_lock rcu_read_lock rcu_read_lock_bh rcu_read_lock rcu_read_lock_bh rcu_read_lock &data->lock irq_context: 0 (wq_completion)bond152 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &idev->mc_lock rcu_read_lock rcu_read_lock rcu_read_lock_bh rcu_read_lock rcu_read_lock_bh rcu_read_lock &obj_hash[i].lock irq_context: 0 (wq_completion)bond152 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &idev->mc_lock rcu_read_lock rcu_read_lock rcu_node_0 irq_context: 0 (wq_completion)bond152 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &idev->mc_lock rcu_read_lock rcu_read_lock &rq->__lock irq_context: 0 (wq_completion)bond152 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &idev->mc_lock rcu_read_lock rcu_read_lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)bond152 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &idev->mc_lock rcu_read_lock rcu_node_0 irq_context: 0 (wq_completion)bond152 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &idev->mc_lock rcu_read_lock &rq->__lock irq_context: 0 (wq_completion)bond152 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex rcu_read_lock &rq->__lock irq_context: 0 (wq_completion)bond152 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex.wait_lock irq_context: 0 (wq_completion)bond152 (work_completion)(&(&bond->mcast_work)->work) &p->pi_lock irq_context: 0 (wq_completion)bond152 (work_completion)(&(&bond->mcast_work)->work) &p->pi_lock &rq->__lock irq_context: 0 (wq_completion)bond152 (work_completion)(&(&bond->mcast_work)->work) &p->pi_lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)mld (work_completion)(&(&idev->mc_ifc_work)->work) &idev->mc_lock rcu_read_lock fill_pool_map-wait-type-override pool_lock#2 irq_context: 0 (wq_completion)mld (work_completion)(&(&idev->mc_ifc_work)->work) &idev->mc_lock rcu_read_lock fill_pool_map-wait-type-override pool_lock irq_context: 0 (wq_completion)bond151#2 irq_context: 0 (wq_completion)bond151#2 &rq->__lock irq_context: 0 (wq_completion)bond151#2 &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)bond151#2 (work_completion)(&(&slave->notify_work)->work) irq_context: 0 (wq_completion)bond151#2 (work_completion)(&(&slave->notify_work)->work) &obj_hash[i].lock irq_context: 0 (wq_completion)bond151#2 (work_completion)(&(&slave->notify_work)->work) &base->lock irq_context: 0 (wq_completion)bond151#2 (work_completion)(&(&slave->notify_work)->work) &base->lock &obj_hash[i].lock irq_context: 0 (wq_completion)bond144 (work_completion)(&(&slave->notify_work)->work) rtnl_mutex net_rwsem irq_context: 0 (wq_completion)bond144 (work_completion)(&(&slave->notify_work)->work) rtnl_mutex net_rwsem &list->lock#2 irq_context: 0 (wq_completion)bond144 (work_completion)(&(&slave->notify_work)->work) rtnl_mutex &tn->lock irq_context: 0 (wq_completion)bond144 (work_completion)(&(&slave->notify_work)->work) rtnl_mutex.wait_lock irq_context: 0 (wq_completion)bond144 (work_completion)(&(&slave->notify_work)->work) &p->pi_lock irq_context: 0 (wq_completion)bond151#2 (work_completion)(&(&bond->mcast_work)->work) irq_context: 0 (wq_completion)bond151#2 (work_completion)(&(&bond->mcast_work)->work) &obj_hash[i].lock irq_context: 0 (wq_completion)bond151#2 (work_completion)(&(&bond->mcast_work)->work) &base->lock irq_context: 0 (wq_completion)bond151#2 (work_completion)(&(&bond->mcast_work)->work) &base->lock &obj_hash[i].lock irq_context: 0 (wq_completion)bond151#2 (work_completion)(&(&bond->mcast_work)->work) &rq->__lock irq_context: 0 (wq_completion)bond151#2 (work_completion)(&(&bond->mcast_work)->work) &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)bond180 irq_context: 0 (wq_completion)bond180 &rq->__lock irq_context: 0 (wq_completion)bond180 &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)bond180 (work_completion)(&(&slave->notify_work)->work) irq_context: 0 (wq_completion)bond180 (work_completion)(&(&slave->notify_work)->work) &obj_hash[i].lock irq_context: 0 (wq_completion)bond180 (work_completion)(&(&slave->notify_work)->work) &base->lock irq_context: 0 (wq_completion)bond180 (work_completion)(&(&slave->notify_work)->work) &base->lock &obj_hash[i].lock irq_context: 0 (wq_completion)bond180 (work_completion)(&(&bond->mcast_work)->work) irq_context: 0 (wq_completion)bond180 (work_completion)(&(&bond->mcast_work)->work) &obj_hash[i].lock irq_context: 0 (wq_completion)bond180 (work_completion)(&(&bond->mcast_work)->work) &base->lock irq_context: 0 (wq_completion)bond180 (work_completion)(&(&bond->mcast_work)->work) &base->lock &obj_hash[i].lock irq_context: 0 (wq_completion)bond180 (work_completion)(&(&bond->mcast_work)->work) &rq->__lock irq_context: 0 (wq_completion)bond49#2 (work_completion)(&(&bond->mcast_work)->work) &rq->__lock irq_context: 0 (wq_completion)bond49#2 (work_completion)(&(&bond->mcast_work)->work) &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)bond53#3 irq_context: 0 (wq_completion)bond53#3 (work_completion)(&(&slave->notify_work)->work) irq_context: 0 (wq_completion)bond53#3 (work_completion)(&(&slave->notify_work)->work) &obj_hash[i].lock irq_context: 0 (wq_completion)bond53#3 (work_completion)(&(&slave->notify_work)->work) &base->lock irq_context: 0 (wq_completion)bond53#3 (work_completion)(&(&slave->notify_work)->work) &base->lock &obj_hash[i].lock irq_context: 0 (wq_completion)bond53#3 (work_completion)(&(&slave->notify_work)->work) &rq->__lock irq_context: 0 (wq_completion)bond53#3 (work_completion)(&(&slave->notify_work)->work) &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)bond15#3 (work_completion)(&(&slave->notify_work)->work) fill_pool_map-wait-type-override pool_lock irq_context: 0 (wq_completion)bond53#3 (work_completion)(&(&bond->mcast_work)->work) irq_context: 0 (wq_completion)bond53#3 (work_completion)(&(&bond->mcast_work)->work) &obj_hash[i].lock irq_context: 0 (wq_completion)bond53#3 (work_completion)(&(&bond->mcast_work)->work) &base->lock irq_context: 0 (wq_completion)bond53#3 (work_completion)(&(&bond->mcast_work)->work) &base->lock &obj_hash[i].lock irq_context: 0 &f->f_pos_lock &type->i_mutex_dir_key#3 sb_writers#4 mmu_notifier_invalidate_range_start &rq->__lock &cfs_rq->removed.lock irq_context: 0 (wq_completion)bond168#2 irq_context: 0 (wq_completion)bond168#2 (work_completion)(&(&slave->notify_work)->work) irq_context: 0 (wq_completion)bond168#2 (work_completion)(&(&slave->notify_work)->work) &obj_hash[i].lock irq_context: 0 (wq_completion)bond168#2 (work_completion)(&(&slave->notify_work)->work) &base->lock irq_context: 0 (wq_completion)bond168#2 (work_completion)(&(&slave->notify_work)->work) &base->lock &obj_hash[i].lock irq_context: 0 (wq_completion)bond168#2 (work_completion)(&(&slave->notify_work)->work) &rq->__lock irq_context: 0 (wq_completion)bond168#2 (work_completion)(&(&slave->notify_work)->work) &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)bond168#2 (work_completion)(&(&bond->mcast_work)->work) irq_context: 0 (wq_completion)bond168#2 (work_completion)(&(&bond->mcast_work)->work) &obj_hash[i].lock irq_context: 0 (wq_completion)bond168#2 (work_completion)(&(&bond->mcast_work)->work) &base->lock irq_context: 0 (wq_completion)bond168#2 (work_completion)(&(&bond->mcast_work)->work) &base->lock &obj_hash[i].lock irq_context: 0 (wq_completion)bond168#2 (work_completion)(&(&bond->mcast_work)->work) &rq->__lock irq_context: 0 (wq_completion)bond168#2 (work_completion)(&(&bond->mcast_work)->work) &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)bond152#3 irq_context: 0 (wq_completion)bond152#3 (work_completion)(&(&slave->notify_work)->work) irq_context: 0 (wq_completion)bond152#3 (work_completion)(&(&slave->notify_work)->work) &obj_hash[i].lock irq_context: 0 (wq_completion)bond152#3 (work_completion)(&(&slave->notify_work)->work) &base->lock irq_context: 0 (wq_completion)bond152#3 (work_completion)(&(&slave->notify_work)->work) &base->lock &obj_hash[i].lock irq_context: 0 (wq_completion)bond152#3 (work_completion)(&(&slave->notify_work)->work) &rq->__lock irq_context: 0 (wq_completion)bond152#3 (work_completion)(&(&slave->notify_work)->work) &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)bond152#3 (work_completion)(&(&bond->mcast_work)->work) irq_context: 0 (wq_completion)bond152#3 (work_completion)(&(&bond->mcast_work)->work) &obj_hash[i].lock irq_context: 0 (wq_completion)bond152#3 (work_completion)(&(&bond->mcast_work)->work) &base->lock irq_context: 0 (wq_completion)bond152#3 (work_completion)(&(&bond->mcast_work)->work) &base->lock &obj_hash[i].lock irq_context: softirq (&ndev->rs_timer) rcu_read_lock rcu_read_lock &tbl->lock fill_pool_map-wait-type-override pool_lock#2 irq_context: softirq (&ndev->rs_timer) rcu_read_lock rcu_read_lock &tbl->lock fill_pool_map-wait-type-override pool_lock irq_context: softirq (&ndev->rs_timer) rcu_read_lock rcu_read_lock &tbl->lock fill_pool_map-wait-type-override &c->lock irq_context: softirq (&ndev->rs_timer) rcu_read_lock rcu_read_lock &tbl->lock fill_pool_map-wait-type-override &____s->seqcount#2 irq_context: softirq (&ndev->rs_timer) rcu_read_lock rcu_read_lock &tbl->lock fill_pool_map-wait-type-override &____s->seqcount irq_context: 0 rtnl_mutex wq_pool_mutex &wq->mutex &rq->__lock irq_context: 0 (wq_completion)bond41 (work_completion)(&(&bond->mcast_work)->work) &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)bond181 irq_context: 0 (wq_completion)bond181 (work_completion)(&(&slave->notify_work)->work) irq_context: 0 (wq_completion)bond181 (work_completion)(&(&slave->notify_work)->work) &obj_hash[i].lock irq_context: 0 (wq_completion)bond181 (work_completion)(&(&slave->notify_work)->work) &base->lock irq_context: 0 (wq_completion)bond181 (work_completion)(&(&slave->notify_work)->work) &base->lock &obj_hash[i].lock irq_context: 0 (wq_completion)bond181 (work_completion)(&(&slave->notify_work)->work) &rq->__lock irq_context: 0 (wq_completion)bond181 (work_completion)(&(&slave->notify_work)->work) &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)bond181 (work_completion)(&(&bond->mcast_work)->work) irq_context: 0 (wq_completion)bond181 (work_completion)(&(&bond->mcast_work)->work) &obj_hash[i].lock irq_context: 0 (wq_completion)bond181 (work_completion)(&(&bond->mcast_work)->work) &base->lock irq_context: 0 (wq_completion)bond181 (work_completion)(&(&bond->mcast_work)->work) &base->lock &obj_hash[i].lock irq_context: 0 (wq_completion)bond181 (work_completion)(&(&bond->mcast_work)->work) &rq->__lock irq_context: 0 (wq_completion)bond181 (work_completion)(&(&bond->mcast_work)->work) &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)bond169#2 irq_context: 0 (wq_completion)bond169#2 &rq->__lock irq_context: 0 (wq_completion)bond169#2 &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)bond169#2 (work_completion)(&(&slave->notify_work)->work) irq_context: 0 (wq_completion)bond169#2 (work_completion)(&(&slave->notify_work)->work) &obj_hash[i].lock irq_context: 0 (wq_completion)bond169#2 (work_completion)(&(&slave->notify_work)->work) &base->lock irq_context: 0 (wq_completion)bond169#2 (work_completion)(&(&slave->notify_work)->work) &base->lock &obj_hash[i].lock irq_context: 0 (wq_completion)bond169#2 (work_completion)(&(&slave->notify_work)->work) &rq->__lock irq_context: 0 (wq_completion)bond169#2 (work_completion)(&(&slave->notify_work)->work) &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)bond169#2 (work_completion)(&(&bond->mcast_work)->work) irq_context: 0 (wq_completion)bond169#2 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex fs_reclaim irq_context: 0 (wq_completion)bond169#2 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex fs_reclaim mmu_notifier_invalidate_range_start irq_context: 0 (wq_completion)bond169#2 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &c->lock irq_context: 0 (wq_completion)bond169#2 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &rq->__lock irq_context: 0 (wq_completion)bond169#2 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)bond169#2 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex rcu_read_lock &bond->stats_lock/1 irq_context: 0 (wq_completion)bond169#2 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex rcu_read_lock &ndev->lock irq_context: 0 (wq_completion)bond169#2 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex nl_table_lock irq_context: 0 (wq_completion)bond169#2 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &obj_hash[i].lock irq_context: 0 (wq_completion)bond169#2 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex nl_table_wait.lock irq_context: 0 (wq_completion)bond169#2 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex net_rwsem irq_context: 0 (wq_completion)bond169#2 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex net_rwsem &list->lock#2 irq_context: 0 (wq_completion)bond169#2 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &tn->lock irq_context: 0 (wq_completion)bond169#2 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &idev->mc_lock irq_context: 0 (wq_completion)bond169#2 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex.wait_lock irq_context: 0 (wq_completion)bond169#2 (work_completion)(&(&bond->mcast_work)->work) &p->pi_lock irq_context: 0 (wq_completion)bond169#2 (work_completion)(&(&bond->mcast_work)->work) &p->pi_lock &rq->__lock irq_context: 0 (wq_completion)bond169#2 (work_completion)(&(&bond->mcast_work)->work) &p->pi_lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)bond169#2 (work_completion)(&(&bond->mcast_work)->work) &rq->__lock irq_context: 0 (wq_completion)bond169#2 (work_completion)(&(&bond->mcast_work)->work) &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)bond153#3 irq_context: 0 (wq_completion)bond153#3 &rq->__lock irq_context: 0 (wq_completion)bond153#3 &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)bond153#3 (work_completion)(&(&slave->notify_work)->work) irq_context: 0 (wq_completion)bond153#3 (work_completion)(&(&slave->notify_work)->work) &obj_hash[i].lock irq_context: 0 (wq_completion)bond153#3 (work_completion)(&(&slave->notify_work)->work) &base->lock irq_context: 0 (wq_completion)bond153#3 (work_completion)(&(&slave->notify_work)->work) &base->lock &obj_hash[i].lock irq_context: 0 (wq_completion)bond153#3 (work_completion)(&(&slave->notify_work)->work) &rq->__lock irq_context: 0 (wq_completion)bond153#3 (work_completion)(&(&slave->notify_work)->work) &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)bond153#3 (work_completion)(&(&bond->mcast_work)->work) irq_context: 0 (wq_completion)bond153#3 (work_completion)(&(&bond->mcast_work)->work) &obj_hash[i].lock irq_context: 0 (wq_completion)bond153#3 (work_completion)(&(&bond->mcast_work)->work) &base->lock irq_context: 0 (wq_completion)bond153#3 (work_completion)(&(&bond->mcast_work)->work) &base->lock &obj_hash[i].lock irq_context: 0 (wq_completion)bond54#3 irq_context: 0 (wq_completion)bond54#3 (work_completion)(&(&slave->notify_work)->work) irq_context: 0 (wq_completion)bond54#3 (work_completion)(&(&slave->notify_work)->work) &rq->__lock irq_context: 0 (wq_completion)bond54#3 (work_completion)(&(&slave->notify_work)->work) &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)bond54#3 (work_completion)(&(&slave->notify_work)->work) &obj_hash[i].lock irq_context: 0 (wq_completion)bond54#3 (work_completion)(&(&slave->notify_work)->work) &base->lock irq_context: 0 (wq_completion)bond54#3 (work_completion)(&(&slave->notify_work)->work) &base->lock &obj_hash[i].lock irq_context: 0 (wq_completion)bond54#3 (work_completion)(&(&bond->mcast_work)->work) irq_context: 0 (wq_completion)bond54#3 (work_completion)(&(&bond->mcast_work)->work) &obj_hash[i].lock irq_context: 0 (wq_completion)bond54#3 (work_completion)(&(&bond->mcast_work)->work) &base->lock irq_context: 0 (wq_completion)bond54#3 (work_completion)(&(&bond->mcast_work)->work) &base->lock &obj_hash[i].lock irq_context: 0 (wq_completion)bond54#3 (work_completion)(&(&bond->mcast_work)->work) &rq->__lock irq_context: 0 (wq_completion)bond54#3 (work_completion)(&(&bond->mcast_work)->work) &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)bond152#3 (work_completion)(&(&bond->mcast_work)->work) &rq->__lock irq_context: 0 (wq_completion)bond152#3 (work_completion)(&(&bond->mcast_work)->work) &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)bond182 irq_context: 0 (wq_completion)bond182 (work_completion)(&(&slave->notify_work)->work) irq_context: 0 (wq_completion)bond182 (work_completion)(&(&slave->notify_work)->work) &obj_hash[i].lock irq_context: 0 (wq_completion)bond182 (work_completion)(&(&slave->notify_work)->work) &base->lock irq_context: 0 (wq_completion)bond182 (work_completion)(&(&slave->notify_work)->work) &base->lock &obj_hash[i].lock irq_context: 0 (wq_completion)bond182 (work_completion)(&(&slave->notify_work)->work) &rq->__lock irq_context: 0 (wq_completion)bond123#3 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex fs_reclaim irq_context: 0 (wq_completion)bond123#3 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex fs_reclaim mmu_notifier_invalidate_range_start irq_context: 0 (wq_completion)bond123#3 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &c->lock irq_context: 0 (wq_completion)bond123#3 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex rcu_read_lock &bond->stats_lock/1 irq_context: 0 (wq_completion)bond123#3 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex rcu_read_lock &ndev->lock irq_context: 0 (wq_completion)bond123#3 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex nl_table_lock irq_context: 0 (wq_completion)bond123#3 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &obj_hash[i].lock irq_context: 0 (wq_completion)bond123#3 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex nl_table_wait.lock irq_context: 0 (wq_completion)bond123#3 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex net_rwsem irq_context: 0 (wq_completion)bond123#3 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex net_rwsem &list->lock#2 irq_context: 0 (wq_completion)bond123#3 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &tn->lock irq_context: 0 (wq_completion)bond123#3 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &idev->mc_lock irq_context: 0 (wq_completion)bond123#3 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &idev->mc_lock fs_reclaim irq_context: 0 (wq_completion)bond123#3 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &idev->mc_lock fs_reclaim mmu_notifier_invalidate_range_start irq_context: 0 (wq_completion)bond123#3 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &idev->mc_lock &c->lock irq_context: 0 (wq_completion)bond123#3 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &idev->mc_lock rcu_read_lock &ndev->lock irq_context: 0 (wq_completion)bond123#3 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &idev->mc_lock rcu_read_lock &dir->lock#2 irq_context: 0 (wq_completion)bond123#3 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &idev->mc_lock rcu_read_lock &ul->lock#2 irq_context: 0 (wq_completion)bond123#3 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &idev->mc_lock rcu_read_lock rcu_read_lock &____s->seqcount#10 irq_context: 0 (wq_completion)bond123#3 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &idev->mc_lock rcu_read_lock rcu_read_lock rcu_read_lock_bh rcu_read_lock rcu_read_lock_bh rcu_read_lock &br->multicast_lock irq_context: 0 (wq_completion)bond123#3 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &idev->mc_lock rcu_read_lock rcu_read_lock rcu_read_lock_bh rcu_read_lock rcu_read_lock_bh rcu_read_lock &br->multicast_lock &obj_hash[i].lock irq_context: 0 (wq_completion)bond123#3 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &idev->mc_lock rcu_read_lock rcu_read_lock rcu_read_lock_bh rcu_read_lock rcu_read_lock_bh rcu_read_lock &br->multicast_lock &base->lock irq_context: 0 (wq_completion)bond123#3 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &idev->mc_lock rcu_read_lock rcu_read_lock rcu_read_lock_bh rcu_read_lock rcu_read_lock_bh rcu_read_lock &br->multicast_lock &base->lock &obj_hash[i].lock irq_context: 0 (wq_completion)bond123#3 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &idev->mc_lock rcu_read_lock rcu_read_lock rcu_read_lock_bh rcu_read_lock rcu_read_lock_bh rcu_read_lock &data->lock irq_context: 0 (wq_completion)bond123#3 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &idev->mc_lock rcu_read_lock rcu_read_lock rcu_read_lock_bh rcu_read_lock rcu_read_lock_bh rcu_read_lock &obj_hash[i].lock irq_context: 0 (wq_completion)bond123#3 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex.wait_lock irq_context: 0 (wq_completion)bond123#3 (work_completion)(&(&bond->mcast_work)->work) &p->pi_lock irq_context: 0 (wq_completion)bond154#2 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex fs_reclaim irq_context: 0 (wq_completion)bond154#2 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex fs_reclaim mmu_notifier_invalidate_range_start irq_context: 0 (wq_completion)bond154#2 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &c->lock irq_context: 0 (wq_completion)bond154#2 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex rcu_read_lock &bond->stats_lock/1 irq_context: 0 (wq_completion)bond154#2 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex rcu_read_lock &ndev->lock irq_context: 0 (wq_completion)bond154#2 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex nl_table_lock irq_context: 0 (wq_completion)bond154#2 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &obj_hash[i].lock irq_context: 0 (wq_completion)bond154#2 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex nl_table_wait.lock irq_context: 0 (wq_completion)bond154#2 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex net_rwsem irq_context: 0 (wq_completion)bond154#2 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex net_rwsem &list->lock#2 irq_context: 0 (wq_completion)bond154#2 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &tn->lock irq_context: 0 (wq_completion)bond154#2 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &idev->mc_lock irq_context: 0 (wq_completion)bond154#2 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &idev->mc_lock fs_reclaim irq_context: 0 (wq_completion)bond154#2 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &idev->mc_lock fs_reclaim mmu_notifier_invalidate_range_start irq_context: 0 (wq_completion)bond154#2 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &idev->mc_lock &c->lock irq_context: 0 (wq_completion)bond154#2 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &idev->mc_lock rcu_read_lock &ndev->lock irq_context: 0 (wq_completion)bond154#2 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &idev->mc_lock rcu_read_lock &c->lock irq_context: 0 (wq_completion)bond154#2 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &idev->mc_lock rcu_read_lock rcu_node_0 irq_context: 0 (wq_completion)bond154#2 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &idev->mc_lock rcu_read_lock &rq->__lock irq_context: 0 (wq_completion)bond154#2 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &idev->mc_lock rcu_read_lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)bond154#2 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &idev->mc_lock rcu_read_lock &dir->lock#2 irq_context: 0 (wq_completion)bond154#2 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &idev->mc_lock rcu_read_lock &ul->lock#2 irq_context: 0 (wq_completion)bond154#2 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &idev->mc_lock rcu_read_lock rcu_read_lock &____s->seqcount#10 irq_context: 0 (wq_completion)bond154#2 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &idev->mc_lock rcu_read_lock rcu_read_lock rcu_read_lock_bh rcu_read_lock rcu_read_lock_bh rcu_read_lock &br->multicast_lock irq_context: 0 (wq_completion)bond154#2 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &idev->mc_lock rcu_read_lock rcu_read_lock rcu_read_lock_bh rcu_read_lock rcu_read_lock_bh rcu_read_lock &br->multicast_lock &obj_hash[i].lock irq_context: 0 (wq_completion)bond154#2 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &idev->mc_lock rcu_read_lock rcu_read_lock rcu_read_lock_bh rcu_read_lock rcu_read_lock_bh rcu_read_lock &br->multicast_lock &base->lock irq_context: 0 (wq_completion)bond154#2 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &idev->mc_lock rcu_read_lock rcu_read_lock rcu_read_lock_bh rcu_read_lock rcu_read_lock_bh rcu_read_lock &br->multicast_lock &base->lock &obj_hash[i].lock irq_context: 0 (wq_completion)bond154#2 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &idev->mc_lock rcu_read_lock rcu_read_lock rcu_read_lock_bh rcu_read_lock rcu_read_lock_bh rcu_read_lock &data->lock irq_context: 0 (wq_completion)bond154#2 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &idev->mc_lock rcu_read_lock rcu_read_lock rcu_read_lock_bh rcu_read_lock rcu_read_lock_bh rcu_read_lock &obj_hash[i].lock irq_context: 0 (wq_completion)bond154#2 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &idev->mc_lock rcu_read_lock rcu_read_lock &rq->__lock irq_context: 0 (wq_completion)bond154#2 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &idev->mc_lock rcu_read_lock rcu_read_lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)bond154#2 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex.wait_lock irq_context: 0 (wq_completion)bond154#2 (work_completion)(&(&bond->mcast_work)->work) &p->pi_lock irq_context: 0 (wq_completion)bond154#2 (work_completion)(&(&bond->mcast_work)->work) &p->pi_lock &rq->__lock irq_context: 0 (wq_completion)bond154#2 (work_completion)(&(&bond->mcast_work)->work) &p->pi_lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)bond182 (work_completion)(&(&bond->mcast_work)->work) irq_context: 0 (wq_completion)bond182 (work_completion)(&(&bond->mcast_work)->work) &obj_hash[i].lock irq_context: 0 (wq_completion)bond182 (work_completion)(&(&bond->mcast_work)->work) &base->lock irq_context: 0 (wq_completion)bond182 (work_completion)(&(&bond->mcast_work)->work) &base->lock &obj_hash[i].lock irq_context: 0 (wq_completion)bond182 (work_completion)(&(&bond->mcast_work)->work) &rq->__lock irq_context: 0 (wq_completion)bond182 (work_completion)(&(&bond->mcast_work)->work) &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)bond152#3 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex fs_reclaim irq_context: 0 (wq_completion)bond152#3 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex fs_reclaim mmu_notifier_invalidate_range_start irq_context: 0 (wq_completion)bond152#3 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &c->lock irq_context: 0 (wq_completion)bond152#3 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &n->list_lock irq_context: 0 (wq_completion)bond152#3 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &n->list_lock &c->lock irq_context: 0 (wq_completion)bond152#3 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex rcu_read_lock &bond->stats_lock/1 irq_context: 0 (wq_completion)bond152#3 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex rcu_read_lock &ndev->lock irq_context: 0 (wq_completion)bond152#3 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex nl_table_lock irq_context: 0 (wq_completion)bond152#3 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &obj_hash[i].lock irq_context: 0 (wq_completion)bond152#3 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex nl_table_wait.lock irq_context: 0 (wq_completion)bond152#3 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex net_rwsem irq_context: 0 (wq_completion)bond152#3 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex net_rwsem &list->lock#2 irq_context: 0 (wq_completion)bond152#3 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &tn->lock irq_context: 0 (wq_completion)bond152#3 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &idev->mc_lock irq_context: 0 (wq_completion)bond152#3 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &idev->mc_lock fs_reclaim irq_context: 0 (wq_completion)bond152#3 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &idev->mc_lock fs_reclaim mmu_notifier_invalidate_range_start irq_context: 0 (wq_completion)bond152#3 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &idev->mc_lock rcu_read_lock &ndev->lock irq_context: 0 (wq_completion)bond152#3 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &idev->mc_lock rcu_read_lock &c->lock irq_context: 0 (wq_completion)bond152#3 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &idev->mc_lock rcu_read_lock &dir->lock#2 irq_context: 0 (wq_completion)bond152#3 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &idev->mc_lock rcu_read_lock &ul->lock#2 irq_context: 0 (wq_completion)bond152#3 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &idev->mc_lock rcu_read_lock rcu_read_lock &obj_hash[i].lock irq_context: 0 (wq_completion)bond152#3 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &idev->mc_lock rcu_read_lock rcu_read_lock &c->lock irq_context: 0 (wq_completion)bond152#3 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &idev->mc_lock rcu_read_lock rcu_read_lock &n->list_lock irq_context: 0 (wq_completion)bond152#3 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &idev->mc_lock rcu_read_lock rcu_read_lock &n->list_lock &c->lock irq_context: 0 (wq_completion)bond152#3 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &idev->mc_lock rcu_read_lock rcu_read_lock &dir->lock#2 irq_context: 0 (wq_completion)bond152#3 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &idev->mc_lock rcu_read_lock rcu_read_lock &tbl->lock irq_context: 0 (wq_completion)bond152#3 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &idev->mc_lock rcu_read_lock rcu_read_lock &n->lock irq_context: 0 (wq_completion)bond152#3 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &idev->mc_lock rcu_read_lock rcu_read_lock &____s->seqcount#9 irq_context: 0 (wq_completion)bond152#3 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &idev->mc_lock rcu_read_lock rcu_read_lock rcu_read_lock_bh rcu_read_lock rcu_read_lock_bh rcu_read_lock &br->multicast_lock irq_context: 0 (wq_completion)bond152#3 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &idev->mc_lock rcu_read_lock rcu_read_lock rcu_read_lock_bh rcu_read_lock rcu_read_lock_bh rcu_read_lock &br->multicast_lock &obj_hash[i].lock irq_context: 0 (wq_completion)bond152#3 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &idev->mc_lock rcu_read_lock rcu_read_lock rcu_read_lock_bh rcu_read_lock rcu_read_lock_bh rcu_read_lock &br->multicast_lock rcu_read_lock rhashtable_bucket irq_context: 0 (wq_completion)bond152#3 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &idev->mc_lock rcu_read_lock rcu_read_lock rcu_read_lock_bh rcu_read_lock rcu_read_lock_bh rcu_read_lock &br->multicast_lock &c->lock irq_context: 0 (wq_completion)bond152#3 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &idev->mc_lock rcu_read_lock rcu_read_lock rcu_read_lock_bh rcu_read_lock rcu_read_lock_bh rcu_read_lock &br->multicast_lock nl_table_lock irq_context: 0 (wq_completion)bond152#3 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &idev->mc_lock rcu_read_lock rcu_read_lock rcu_read_lock_bh rcu_read_lock rcu_read_lock_bh rcu_read_lock &br->multicast_lock nl_table_wait.lock irq_context: 0 (wq_completion)bond152#3 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &idev->mc_lock rcu_read_lock rcu_read_lock rcu_read_lock_bh rcu_read_lock rcu_read_lock_bh rcu_read_lock &br->multicast_lock &base->lock irq_context: 0 (wq_completion)bond152#3 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &idev->mc_lock rcu_read_lock rcu_read_lock rcu_read_lock_bh rcu_read_lock rcu_read_lock_bh rcu_read_lock &br->multicast_lock &base->lock &obj_hash[i].lock irq_context: 0 (wq_completion)bond152#3 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &idev->mc_lock rcu_read_lock rcu_read_lock rcu_read_lock_bh rcu_read_lock rcu_read_lock_bh rcu_read_lock &data->lock irq_context: 0 (wq_completion)bond152#3 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &idev->mc_lock rcu_read_lock rcu_read_lock rcu_read_lock_bh rcu_read_lock rcu_read_lock_bh rcu_read_lock &obj_hash[i].lock irq_context: 0 (wq_completion)bond152#3 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex.wait_lock irq_context: 0 (wq_completion)bond152#3 (work_completion)(&(&bond->mcast_work)->work) &p->pi_lock irq_context: 0 (wq_completion)bond144#3 (work_completion)(&(&slave->notify_work)->work) rtnl_mutex net_rwsem irq_context: 0 (wq_completion)bond144#3 (work_completion)(&(&slave->notify_work)->work) rtnl_mutex net_rwsem &list->lock#2 irq_context: 0 (wq_completion)bond144#3 (work_completion)(&(&slave->notify_work)->work) rtnl_mutex &tn->lock irq_context: 0 (wq_completion)bond144#3 (work_completion)(&(&slave->notify_work)->work) rtnl_mutex.wait_lock irq_context: 0 (wq_completion)bond144#3 (work_completion)(&(&slave->notify_work)->work) &p->pi_lock irq_context: 0 (wq_completion)bond49#2 (work_completion)(&(&slave->notify_work)->work) rtnl_mutex net_rwsem irq_context: 0 (wq_completion)bond49#2 (work_completion)(&(&slave->notify_work)->work) rtnl_mutex net_rwsem &list->lock#2 irq_context: 0 (wq_completion)bond49#2 (work_completion)(&(&slave->notify_work)->work) rtnl_mutex &tn->lock irq_context: 0 (wq_completion)bond49#2 (work_completion)(&(&slave->notify_work)->work) rtnl_mutex.wait_lock irq_context: 0 (wq_completion)bond49#2 (work_completion)(&(&slave->notify_work)->work) &p->pi_lock irq_context: 0 (wq_completion)bond167#2 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex fs_reclaim irq_context: 0 (wq_completion)bond167#2 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex fs_reclaim mmu_notifier_invalidate_range_start irq_context: 0 (wq_completion)bond167#2 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex rcu_read_lock &bond->stats_lock/1 irq_context: 0 (wq_completion)bond167#2 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex rcu_read_lock &ndev->lock irq_context: 0 (wq_completion)bond167#2 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex nl_table_lock irq_context: 0 (wq_completion)bond167#2 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &obj_hash[i].lock irq_context: 0 (wq_completion)bond167#2 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex nl_table_wait.lock irq_context: 0 (wq_completion)bond167#2 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex net_rwsem irq_context: 0 (wq_completion)bond167#2 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex net_rwsem &list->lock#2 irq_context: 0 (wq_completion)bond167#2 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &tn->lock irq_context: 0 (wq_completion)bond167#2 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &idev->mc_lock irq_context: 0 (wq_completion)bond167#2 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &idev->mc_lock fs_reclaim irq_context: 0 (wq_completion)bond167#2 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &idev->mc_lock fs_reclaim mmu_notifier_invalidate_range_start irq_context: 0 (wq_completion)bond167#2 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &idev->mc_lock rcu_read_lock &ndev->lock irq_context: 0 (wq_completion)bond167#2 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &idev->mc_lock rcu_read_lock &dir->lock#2 irq_context: 0 (wq_completion)bond167#2 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &idev->mc_lock rcu_read_lock &ul->lock#2 irq_context: 0 (wq_completion)bond167#2 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &idev->mc_lock rcu_read_lock rcu_node_0 irq_context: 0 (wq_completion)bond167#2 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &idev->mc_lock rcu_read_lock &rq->__lock irq_context: 0 (wq_completion)bond167#2 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &idev->mc_lock rcu_read_lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)bond167#2 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &idev->mc_lock rcu_read_lock rcu_read_lock &____s->seqcount#10 irq_context: 0 (wq_completion)bond167#2 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &idev->mc_lock rcu_read_lock rcu_read_lock rcu_read_lock_bh rcu_read_lock rcu_read_lock_bh rcu_read_lock &br->multicast_lock irq_context: 0 (wq_completion)bond167#2 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &idev->mc_lock rcu_read_lock rcu_read_lock rcu_read_lock_bh rcu_read_lock rcu_read_lock_bh rcu_read_lock &br->multicast_lock &obj_hash[i].lock irq_context: 0 (wq_completion)bond167#2 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &idev->mc_lock rcu_read_lock rcu_read_lock rcu_read_lock_bh rcu_read_lock rcu_read_lock_bh rcu_read_lock &br->multicast_lock &base->lock irq_context: 0 (wq_completion)bond167#2 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &idev->mc_lock rcu_read_lock rcu_read_lock rcu_read_lock_bh rcu_read_lock rcu_read_lock_bh rcu_read_lock &br->multicast_lock &base->lock &obj_hash[i].lock irq_context: 0 (wq_completion)bond167#2 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &idev->mc_lock rcu_read_lock rcu_read_lock rcu_read_lock_bh rcu_read_lock rcu_read_lock_bh rcu_read_lock &data->lock irq_context: 0 (wq_completion)bond167#2 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &idev->mc_lock rcu_read_lock rcu_read_lock rcu_read_lock_bh rcu_read_lock rcu_read_lock_bh rcu_read_lock &obj_hash[i].lock irq_context: 0 (wq_completion)bond37#2 (work_completion)(&(&slave->notify_work)->work) &rq->__lock irq_context: 0 (wq_completion)bond167#2 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &idev->mc_lock rcu_read_lock &rcu_state.gp_wq irq_context: 0 (wq_completion)bond37#2 (work_completion)(&(&slave->notify_work)->work) &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)bond167#2 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &idev->mc_lock rcu_read_lock &rcu_state.gp_wq &p->pi_lock irq_context: 0 (wq_completion)bond167#2 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &idev->mc_lock rcu_read_lock &rcu_state.gp_wq &p->pi_lock &rq->__lock irq_context: 0 (wq_completion)bond167#2 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &idev->mc_lock rcu_read_lock &rcu_state.gp_wq &p->pi_lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)bond167#2 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex.wait_lock irq_context: 0 (wq_completion)bond167#2 (work_completion)(&(&bond->mcast_work)->work) &p->pi_lock irq_context: 0 (wq_completion)bond141 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex fs_reclaim irq_context: 0 (wq_completion)bond141 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex fs_reclaim mmu_notifier_invalidate_range_start irq_context: 0 (wq_completion)bond141 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &c->lock irq_context: 0 (wq_completion)bond141 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &n->list_lock irq_context: 0 (wq_completion)bond141 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &n->list_lock &c->lock irq_context: 0 (wq_completion)bond141 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex rcu_read_lock &bond->stats_lock/1 irq_context: 0 (wq_completion)bond141 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex rcu_read_lock &ndev->lock irq_context: 0 (wq_completion)bond141 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex nl_table_lock irq_context: 0 (wq_completion)bond141 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &obj_hash[i].lock irq_context: 0 (wq_completion)bond141 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex nl_table_wait.lock irq_context: 0 (wq_completion)bond141 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex net_rwsem irq_context: 0 (wq_completion)bond141 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex net_rwsem &list->lock#2 irq_context: 0 (wq_completion)bond141 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &tn->lock irq_context: 0 (wq_completion)bond141 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &idev->mc_lock irq_context: 0 (wq_completion)bond141 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &idev->mc_lock fs_reclaim irq_context: 0 (wq_completion)bond141 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &idev->mc_lock fs_reclaim mmu_notifier_invalidate_range_start irq_context: 0 (wq_completion)bond141 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &idev->mc_lock rcu_read_lock &ndev->lock irq_context: 0 (wq_completion)bond141 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &idev->mc_lock rcu_read_lock &c->lock irq_context: 0 (wq_completion)bond141 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &idev->mc_lock rcu_read_lock &dir->lock#2 irq_context: 0 (wq_completion)bond141 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &idev->mc_lock rcu_read_lock &ul->lock#2 irq_context: 0 (wq_completion)bond141 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &idev->mc_lock rcu_read_lock rcu_read_lock &obj_hash[i].lock irq_context: 0 (wq_completion)bond141 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &idev->mc_lock rcu_read_lock rcu_read_lock &c->lock irq_context: 0 (wq_completion)bond141 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &idev->mc_lock rcu_read_lock rcu_read_lock &n->list_lock irq_context: 0 (wq_completion)bond141 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &idev->mc_lock rcu_read_lock rcu_read_lock &n->list_lock &c->lock irq_context: 0 (wq_completion)bond141 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &idev->mc_lock rcu_read_lock rcu_read_lock &dir->lock#2 irq_context: 0 (wq_completion)bond141 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &idev->mc_lock rcu_read_lock rcu_read_lock &tbl->lock irq_context: 0 (wq_completion)bond141 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &idev->mc_lock rcu_read_lock rcu_read_lock &n->lock irq_context: 0 (wq_completion)bond141 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &idev->mc_lock rcu_read_lock rcu_read_lock &____s->seqcount#9 irq_context: 0 (wq_completion)bond141 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &idev->mc_lock rcu_read_lock rcu_read_lock rcu_read_lock_bh rcu_read_lock rcu_read_lock_bh rcu_read_lock &br->multicast_lock irq_context: 0 (wq_completion)bond141 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &idev->mc_lock rcu_read_lock rcu_read_lock rcu_read_lock_bh rcu_read_lock rcu_read_lock_bh rcu_read_lock &br->multicast_lock &obj_hash[i].lock irq_context: 0 (wq_completion)bond141 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &idev->mc_lock rcu_read_lock rcu_read_lock rcu_read_lock_bh rcu_read_lock rcu_read_lock_bh rcu_read_lock &br->multicast_lock &base->lock irq_context: 0 (wq_completion)bond141 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &idev->mc_lock rcu_read_lock rcu_read_lock rcu_read_lock_bh rcu_read_lock rcu_read_lock_bh rcu_read_lock &br->multicast_lock &base->lock &obj_hash[i].lock irq_context: 0 (wq_completion)bond141 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &idev->mc_lock rcu_read_lock rcu_read_lock rcu_read_lock_bh rcu_read_lock rcu_read_lock_bh rcu_read_lock &data->lock irq_context: 0 (wq_completion)bond141 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &idev->mc_lock rcu_read_lock rcu_read_lock rcu_read_lock_bh rcu_read_lock rcu_read_lock_bh rcu_read_lock &obj_hash[i].lock irq_context: 0 (wq_completion)bond141 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &rq->__lock irq_context: 0 (wq_completion)bond141 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)bond141 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex.wait_lock irq_context: 0 (wq_completion)bond141 (work_completion)(&(&bond->mcast_work)->work) &p->pi_lock irq_context: 0 (wq_completion)bond141 (work_completion)(&(&bond->mcast_work)->work) &p->pi_lock &rq->__lock irq_context: 0 (wq_completion)bond141 (work_completion)(&(&bond->mcast_work)->work) &p->pi_lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)bond52#3 &rq->__lock irq_context: 0 (wq_completion)bond52#3 &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)bond170#2 irq_context: 0 (wq_completion)bond170#2 &rq->__lock irq_context: 0 (wq_completion)bond170#2 &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)bond170#2 (work_completion)(&(&slave->notify_work)->work) irq_context: 0 (wq_completion)bond170#2 (work_completion)(&(&slave->notify_work)->work) &obj_hash[i].lock irq_context: 0 (wq_completion)bond170#2 (work_completion)(&(&slave->notify_work)->work) &base->lock irq_context: 0 (wq_completion)bond170#2 (work_completion)(&(&slave->notify_work)->work) &base->lock &obj_hash[i].lock irq_context: 0 (wq_completion)bond170#2 (work_completion)(&(&bond->mcast_work)->work) irq_context: 0 (wq_completion)bond170#2 (work_completion)(&(&bond->mcast_work)->work) &obj_hash[i].lock irq_context: 0 (wq_completion)bond170#2 (work_completion)(&(&bond->mcast_work)->work) &base->lock irq_context: 0 (wq_completion)bond170#2 (work_completion)(&(&bond->mcast_work)->work) &base->lock &obj_hash[i].lock irq_context: 0 (wq_completion)bond167 (work_completion)(&(&slave->notify_work)->work) rtnl_mutex net_rwsem irq_context: 0 (wq_completion)bond167 (work_completion)(&(&slave->notify_work)->work) rtnl_mutex net_rwsem &list->lock#2 irq_context: 0 (wq_completion)bond167 (work_completion)(&(&slave->notify_work)->work) rtnl_mutex &tn->lock irq_context: 0 (wq_completion)bond167 (work_completion)(&(&slave->notify_work)->work) rtnl_mutex.wait_lock irq_context: 0 (wq_completion)bond167 (work_completion)(&(&slave->notify_work)->work) &p->pi_lock irq_context: 0 (wq_completion)bond47#2 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex fs_reclaim irq_context: 0 (wq_completion)bond47#2 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex fs_reclaim mmu_notifier_invalidate_range_start irq_context: 0 (wq_completion)bond47#2 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex fs_reclaim mmu_notifier_invalidate_range_start &rq->__lock irq_context: 0 (wq_completion)bond47#2 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex fs_reclaim mmu_notifier_invalidate_range_start &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)bond47#2 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &c->lock irq_context: 0 (wq_completion)bond47#2 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex rcu_read_lock &bond->stats_lock/1 irq_context: 0 (wq_completion)bond47#2 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex rcu_read_lock &ndev->lock irq_context: 0 (wq_completion)bond47#2 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex nl_table_lock irq_context: 0 (wq_completion)bond47#2 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &obj_hash[i].lock irq_context: 0 (wq_completion)bond47#2 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex nl_table_wait.lock irq_context: 0 (wq_completion)bond47#2 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex net_rwsem irq_context: 0 (wq_completion)bond47#2 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex net_rwsem &list->lock#2 irq_context: 0 (wq_completion)bond47#2 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &tn->lock irq_context: 0 (wq_completion)bond47#2 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &tn->lock &rq->__lock irq_context: 0 (wq_completion)bond47#2 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &tn->lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)bond47#2 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &idev->mc_lock irq_context: 0 (wq_completion)bond47#2 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &idev->mc_lock fs_reclaim irq_context: 0 (wq_completion)bond47#2 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &idev->mc_lock fs_reclaim mmu_notifier_invalidate_range_start irq_context: 0 (wq_completion)bond47#2 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &idev->mc_lock rcu_read_lock &ndev->lock irq_context: 0 (wq_completion)bond47#2 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &idev->mc_lock rcu_read_lock &c->lock irq_context: 0 (wq_completion)bond47#2 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &idev->mc_lock rcu_read_lock &dir->lock#2 irq_context: 0 (wq_completion)bond47#2 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &idev->mc_lock rcu_read_lock &ul->lock#2 irq_context: 0 (wq_completion)bond47#2 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &idev->mc_lock rcu_read_lock rcu_read_lock &____s->seqcount#12 irq_context: 0 (wq_completion)bond47#2 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &idev->mc_lock rcu_read_lock rcu_read_lock rcu_node_0 irq_context: 0 (wq_completion)bond47#2 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &idev->mc_lock rcu_read_lock rcu_read_lock &rq->__lock irq_context: 0 (wq_completion)bond47#2 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &idev->mc_lock rcu_read_lock rcu_read_lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)bond47#2 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &idev->mc_lock rcu_read_lock rcu_read_lock &____s->seqcount#16 irq_context: 0 (wq_completion)bond47#2 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &idev->mc_lock rcu_read_lock &obj_hash[i].lock irq_context: 0 (wq_completion)bond47#2 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &idev->mc_lock rcu_read_lock &data->lock irq_context: 0 (wq_completion)bond47#2 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &idev->mc_lock rcu_read_lock rcu_node_0 irq_context: 0 (wq_completion)bond47#2 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex.wait_lock irq_context: 0 (wq_completion)bond47#2 (work_completion)(&(&bond->mcast_work)->work) &p->pi_lock irq_context: 0 (wq_completion)bond47#2 (work_completion)(&(&bond->mcast_work)->work) &p->pi_lock &rq->__lock irq_context: 0 (wq_completion)bond47#2 (work_completion)(&(&bond->mcast_work)->work) &p->pi_lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)bond154#3 irq_context: 0 (wq_completion)bond154#3 (work_completion)(&(&slave->notify_work)->work) irq_context: 0 (wq_completion)bond154#3 (work_completion)(&(&slave->notify_work)->work) &obj_hash[i].lock irq_context: 0 (wq_completion)bond154#3 (work_completion)(&(&slave->notify_work)->work) &base->lock irq_context: 0 (wq_completion)bond154#3 (work_completion)(&(&slave->notify_work)->work) &base->lock &obj_hash[i].lock irq_context: 0 (wq_completion)bond154#3 (work_completion)(&(&slave->notify_work)->work) &rq->__lock irq_context: 0 (wq_completion)bond154#3 (work_completion)(&(&slave->notify_work)->work) &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)bond123 &cfs_rq->removed.lock irq_context: 0 (wq_completion)bond123 &obj_hash[i].lock irq_context: 0 (wq_completion)bond154#3 (work_completion)(&(&bond->mcast_work)->work) irq_context: 0 (wq_completion)bond154#3 (work_completion)(&(&bond->mcast_work)->work) &obj_hash[i].lock irq_context: 0 (wq_completion)bond154#3 (work_completion)(&(&bond->mcast_work)->work) &base->lock irq_context: 0 (wq_completion)bond154#3 (work_completion)(&(&bond->mcast_work)->work) &base->lock &obj_hash[i].lock irq_context: 0 (wq_completion)bond154#3 (work_completion)(&(&bond->mcast_work)->work) &rq->__lock irq_context: 0 (wq_completion)bond154#3 (work_completion)(&(&bond->mcast_work)->work) &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)bond180 (work_completion)(&(&slave->notify_work)->work) rtnl_mutex net_rwsem irq_context: 0 (wq_completion)bond180 (work_completion)(&(&slave->notify_work)->work) rtnl_mutex net_rwsem &rq->__lock irq_context: 0 (wq_completion)bond180 (work_completion)(&(&slave->notify_work)->work) rtnl_mutex net_rwsem &list->lock#2 irq_context: 0 (wq_completion)bond180 (work_completion)(&(&slave->notify_work)->work) rtnl_mutex &tn->lock irq_context: 0 (wq_completion)bond180 (work_completion)(&(&slave->notify_work)->work) rtnl_mutex.wait_lock irq_context: 0 (wq_completion)bond180 (work_completion)(&(&slave->notify_work)->work) &p->pi_lock irq_context: 0 (wq_completion)bond132#2 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex fs_reclaim irq_context: 0 (wq_completion)bond132#2 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex fs_reclaim mmu_notifier_invalidate_range_start irq_context: 0 (wq_completion)bond132#2 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &c->lock irq_context: 0 (wq_completion)bond132#2 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &n->list_lock irq_context: 0 (wq_completion)bond132#2 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &n->list_lock &c->lock irq_context: 0 (wq_completion)bond132#2 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex rcu_read_lock &bond->stats_lock/1 irq_context: 0 (wq_completion)bond132#2 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex rcu_read_lock &ndev->lock irq_context: 0 (wq_completion)bond132#2 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex nl_table_lock irq_context: 0 (wq_completion)bond132#2 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &obj_hash[i].lock irq_context: 0 (wq_completion)bond132#2 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex nl_table_wait.lock irq_context: 0 (wq_completion)bond132#2 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex net_rwsem irq_context: 0 (wq_completion)bond132#2 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex net_rwsem &rq->__lock irq_context: 0 (wq_completion)bond132#2 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex net_rwsem &list->lock#2 irq_context: 0 (wq_completion)bond132#2 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &tn->lock irq_context: 0 (wq_completion)bond132#2 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &idev->mc_lock irq_context: 0 (wq_completion)bond132#2 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &idev->mc_lock fs_reclaim irq_context: 0 (wq_completion)bond132#2 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &idev->mc_lock fs_reclaim mmu_notifier_invalidate_range_start irq_context: 0 (wq_completion)bond132#2 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &idev->mc_lock &c->lock irq_context: 0 (wq_completion)bond132#2 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &idev->mc_lock rcu_read_lock &ndev->lock irq_context: 0 (wq_completion)bond132#2 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &idev->mc_lock rcu_read_lock &dir->lock#2 irq_context: 0 (wq_completion)bond132#2 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &idev->mc_lock rcu_read_lock &ul->lock#2 irq_context: 0 (wq_completion)bond132#2 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &idev->mc_lock rcu_read_lock rcu_read_lock &obj_hash[i].lock irq_context: 0 (wq_completion)bond132#2 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &idev->mc_lock rcu_read_lock rcu_read_lock &dir->lock#2 irq_context: 0 (wq_completion)bond132#2 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &idev->mc_lock rcu_read_lock rcu_read_lock &tbl->lock irq_context: 0 (wq_completion)bond132#2 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &idev->mc_lock rcu_read_lock rcu_read_lock &n->lock irq_context: 0 (wq_completion)bond132#2 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &idev->mc_lock rcu_read_lock rcu_read_lock &____s->seqcount#9 irq_context: 0 (wq_completion)bond132#2 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &idev->mc_lock rcu_read_lock rcu_read_lock rcu_read_lock_bh rcu_read_lock rcu_read_lock_bh rcu_read_lock &br->multicast_lock irq_context: 0 (wq_completion)bond132#2 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &idev->mc_lock rcu_read_lock rcu_read_lock rcu_read_lock_bh rcu_read_lock rcu_read_lock_bh rcu_read_lock &br->multicast_lock &obj_hash[i].lock irq_context: 0 (wq_completion)bond132#2 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &idev->mc_lock rcu_read_lock rcu_read_lock rcu_read_lock_bh rcu_read_lock rcu_read_lock_bh rcu_read_lock &br->multicast_lock &base->lock irq_context: 0 (wq_completion)bond132#2 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &idev->mc_lock rcu_read_lock rcu_read_lock rcu_read_lock_bh rcu_read_lock rcu_read_lock_bh rcu_read_lock &br->multicast_lock &base->lock &obj_hash[i].lock irq_context: 0 (wq_completion)bond132#2 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &idev->mc_lock rcu_read_lock rcu_read_lock rcu_read_lock_bh rcu_read_lock rcu_read_lock_bh rcu_read_lock &data->lock irq_context: 0 (wq_completion)bond132#2 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &idev->mc_lock rcu_read_lock rcu_read_lock rcu_read_lock_bh rcu_read_lock rcu_read_lock_bh rcu_read_lock &obj_hash[i].lock irq_context: 0 (wq_completion)bond132#2 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex.wait_lock irq_context: 0 (wq_completion)bond132#2 (work_completion)(&(&bond->mcast_work)->work) &p->pi_lock irq_context: 0 (wq_completion)bond132#2 (work_completion)(&(&bond->mcast_work)->work) &p->pi_lock &rq->__lock irq_context: 0 (wq_completion)bond132#2 (work_completion)(&(&bond->mcast_work)->work) &p->pi_lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)bond140#3 &rq->__lock irq_context: 0 (wq_completion)bond140#3 &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)bond183 irq_context: 0 (wq_completion)bond183 (work_completion)(&(&slave->notify_work)->work) irq_context: 0 (wq_completion)bond183 (work_completion)(&(&slave->notify_work)->work) &obj_hash[i].lock irq_context: 0 (wq_completion)bond183 (work_completion)(&(&slave->notify_work)->work) &base->lock irq_context: 0 (wq_completion)bond183 (work_completion)(&(&slave->notify_work)->work) &base->lock &obj_hash[i].lock irq_context: 0 (wq_completion)bond183 (work_completion)(&(&slave->notify_work)->work) &rq->__lock irq_context: 0 (wq_completion)bond183 (work_completion)(&(&slave->notify_work)->work) &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)bond183 (work_completion)(&(&bond->mcast_work)->work) irq_context: 0 (wq_completion)bond183 (work_completion)(&(&bond->mcast_work)->work) &obj_hash[i].lock irq_context: 0 (wq_completion)bond183 (work_completion)(&(&bond->mcast_work)->work) &base->lock irq_context: 0 (wq_completion)bond183 (work_completion)(&(&bond->mcast_work)->work) &base->lock &obj_hash[i].lock irq_context: 0 (wq_completion)bond118#3 (work_completion)(&(&slave->notify_work)->work) &rq->__lock irq_context: 0 (wq_completion)bond118#3 (work_completion)(&(&slave->notify_work)->work) &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)bond171#2 irq_context: 0 (wq_completion)bond171#2 (work_completion)(&(&slave->notify_work)->work) irq_context: 0 (wq_completion)bond171#2 (work_completion)(&(&slave->notify_work)->work) &obj_hash[i].lock irq_context: 0 (wq_completion)bond171#2 (work_completion)(&(&slave->notify_work)->work) &base->lock irq_context: 0 (wq_completion)bond171#2 (work_completion)(&(&slave->notify_work)->work) &base->lock &obj_hash[i].lock irq_context: 0 (wq_completion)bond171#2 (work_completion)(&(&bond->mcast_work)->work) irq_context: 0 (wq_completion)bond171#2 (work_completion)(&(&bond->mcast_work)->work) &obj_hash[i].lock irq_context: 0 (wq_completion)bond171#2 (work_completion)(&(&bond->mcast_work)->work) &base->lock irq_context: 0 (wq_completion)bond171#2 (work_completion)(&(&bond->mcast_work)->work) &base->lock &obj_hash[i].lock irq_context: 0 (wq_completion)bond155#3 irq_context: 0 (wq_completion)bond155#3 (work_completion)(&(&slave->notify_work)->work) irq_context: 0 (wq_completion)bond155#3 (work_completion)(&(&slave->notify_work)->work) &obj_hash[i].lock irq_context: 0 (wq_completion)bond155#3 (work_completion)(&(&slave->notify_work)->work) &base->lock irq_context: 0 (wq_completion)bond155#3 (work_completion)(&(&slave->notify_work)->work) &base->lock &obj_hash[i].lock irq_context: 0 (wq_completion)bond155#3 (work_completion)(&(&slave->notify_work)->work) &rq->__lock irq_context: 0 (wq_completion)bond155#3 (work_completion)(&(&slave->notify_work)->work) &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)bond155#3 (work_completion)(&(&bond->mcast_work)->work) irq_context: 0 (wq_completion)bond155#3 (work_completion)(&(&bond->mcast_work)->work) &obj_hash[i].lock irq_context: 0 (wq_completion)bond155#3 (work_completion)(&(&bond->mcast_work)->work) &base->lock irq_context: 0 (wq_completion)bond155#3 (work_completion)(&(&bond->mcast_work)->work) &base->lock &obj_hash[i].lock irq_context: 0 (wq_completion)bond166#2 (work_completion)(&(&bond->mcast_work)->work) fill_pool_map-wait-type-override pool_lock irq_context: 0 (wq_completion)bond175 (work_completion)(&(&bond->mcast_work)->work) &rq->__lock irq_context: 0 (wq_completion)bond175 (work_completion)(&(&bond->mcast_work)->work) &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)bond184 irq_context: 0 (wq_completion)bond184 (work_completion)(&(&slave->notify_work)->work) irq_context: 0 (wq_completion)bond184 (work_completion)(&(&slave->notify_work)->work) &obj_hash[i].lock irq_context: 0 (wq_completion)bond184 (work_completion)(&(&slave->notify_work)->work) &base->lock irq_context: 0 (wq_completion)bond184 (work_completion)(&(&slave->notify_work)->work) &base->lock &obj_hash[i].lock irq_context: 0 (wq_completion)bond184 (work_completion)(&(&bond->mcast_work)->work) irq_context: 0 (wq_completion)bond184 (work_completion)(&(&bond->mcast_work)->work) &obj_hash[i].lock irq_context: 0 (wq_completion)bond184 (work_completion)(&(&bond->mcast_work)->work) &base->lock irq_context: 0 (wq_completion)bond184 (work_completion)(&(&bond->mcast_work)->work) &base->lock &obj_hash[i].lock irq_context: 0 (wq_completion)bond184 (work_completion)(&(&bond->mcast_work)->work) &rq->__lock irq_context: 0 (wq_completion)bond184 (work_completion)(&(&bond->mcast_work)->work) &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)bond173 (work_completion)(&(&slave->notify_work)->work) &rq->__lock irq_context: 0 (wq_completion)bond173 (work_completion)(&(&slave->notify_work)->work) &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)bond136#2 (work_completion)(&(&slave->notify_work)->work) rtnl_mutex net_rwsem irq_context: 0 (wq_completion)bond136#2 (work_completion)(&(&slave->notify_work)->work) rtnl_mutex net_rwsem &list->lock#2 irq_context: 0 (wq_completion)bond136#2 (work_completion)(&(&slave->notify_work)->work) rtnl_mutex &tn->lock irq_context: 0 (wq_completion)bond136#2 (work_completion)(&(&slave->notify_work)->work) rtnl_mutex.wait_lock irq_context: 0 (wq_completion)bond136#2 (work_completion)(&(&slave->notify_work)->work) &p->pi_lock irq_context: 0 (wq_completion)bond168#2 (work_completion)(&(&slave->notify_work)->work) rtnl_mutex net_rwsem irq_context: 0 (wq_completion)bond168#2 (work_completion)(&(&slave->notify_work)->work) rtnl_mutex net_rwsem &list->lock#2 irq_context: 0 (wq_completion)bond168#2 (work_completion)(&(&slave->notify_work)->work) rtnl_mutex &tn->lock irq_context: 0 (wq_completion)bond168#2 (work_completion)(&(&slave->notify_work)->work) rtnl_mutex rcu_read_lock rcu_node_0 irq_context: 0 (wq_completion)bond168#2 (work_completion)(&(&slave->notify_work)->work) rtnl_mutex rcu_read_lock &rq->__lock irq_context: 0 (wq_completion)bond168#2 (work_completion)(&(&slave->notify_work)->work) rtnl_mutex rcu_read_lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)bond168#2 (work_completion)(&(&slave->notify_work)->work) rtnl_mutex.wait_lock irq_context: 0 (wq_completion)bond168#2 (work_completion)(&(&slave->notify_work)->work) &p->pi_lock irq_context: 0 (wq_completion)bond136#2 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex fs_reclaim irq_context: 0 (wq_completion)bond136#2 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex fs_reclaim mmu_notifier_invalidate_range_start irq_context: 0 (wq_completion)bond136#2 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &c->lock irq_context: 0 (wq_completion)bond136#2 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex rcu_read_lock &bond->stats_lock/1 irq_context: 0 (wq_completion)bond136#2 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex rcu_read_lock &ndev->lock irq_context: 0 (wq_completion)bond136#2 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex nl_table_lock irq_context: 0 (wq_completion)bond136#2 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &obj_hash[i].lock irq_context: 0 (wq_completion)bond136#2 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex nl_table_wait.lock irq_context: 0 (wq_completion)bond136#2 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &rq->__lock irq_context: 0 (wq_completion)bond136#2 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)bond136#2 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex net_rwsem irq_context: 0 (wq_completion)bond136#2 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex net_rwsem &list->lock#2 irq_context: 0 (wq_completion)bond136#2 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &tn->lock irq_context: 0 (wq_completion)bond136#2 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &idev->mc_lock irq_context: 0 (wq_completion)bond136#2 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &idev->mc_lock fs_reclaim irq_context: 0 (wq_completion)bond136#2 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &idev->mc_lock fs_reclaim mmu_notifier_invalidate_range_start irq_context: 0 (wq_completion)bond136#2 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &idev->mc_lock &c->lock irq_context: 0 (wq_completion)bond136#2 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &idev->mc_lock rcu_read_lock &ndev->lock irq_context: 0 (wq_completion)bond136#2 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &idev->mc_lock rcu_read_lock &dir->lock#2 irq_context: 0 (wq_completion)bond136#2 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &idev->mc_lock rcu_read_lock &ul->lock#2 irq_context: 0 (wq_completion)bond136#2 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &idev->mc_lock rcu_read_lock rcu_read_lock &____s->seqcount#10 irq_context: 0 (wq_completion)bond136#2 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &idev->mc_lock rcu_read_lock rcu_read_lock rcu_read_lock_bh rcu_read_lock rcu_read_lock_bh rcu_read_lock &br->multicast_lock irq_context: 0 (wq_completion)bond136#2 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &idev->mc_lock rcu_read_lock rcu_read_lock rcu_read_lock_bh rcu_read_lock rcu_read_lock_bh rcu_read_lock &br->multicast_lock &obj_hash[i].lock irq_context: 0 (wq_completion)bond136#2 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &idev->mc_lock rcu_read_lock rcu_read_lock rcu_read_lock_bh rcu_read_lock rcu_read_lock_bh rcu_read_lock &br->multicast_lock &base->lock irq_context: 0 (wq_completion)bond136#2 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &idev->mc_lock rcu_read_lock rcu_read_lock rcu_read_lock_bh rcu_read_lock rcu_read_lock_bh rcu_read_lock &br->multicast_lock &base->lock &obj_hash[i].lock irq_context: 0 (wq_completion)bond136#2 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &idev->mc_lock rcu_read_lock rcu_read_lock rcu_read_lock_bh rcu_read_lock rcu_read_lock_bh rcu_read_lock &data->lock irq_context: 0 (wq_completion)bond136#2 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &idev->mc_lock rcu_read_lock rcu_read_lock rcu_read_lock_bh rcu_read_lock rcu_read_lock_bh rcu_read_lock &obj_hash[i].lock irq_context: 0 (wq_completion)bond136#2 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &idev->mc_lock rcu_read_lock rcu_read_lock rcu_node_0 irq_context: 0 (wq_completion)bond136#2 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &idev->mc_lock rcu_read_lock rcu_read_lock &rq->__lock irq_context: 0 (wq_completion)bond136#2 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &idev->mc_lock rcu_read_lock rcu_read_lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)bond136#2 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &idev->mc_lock rcu_read_lock rcu_node_0 irq_context: 0 (wq_completion)bond136#2 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex.wait_lock irq_context: 0 (wq_completion)bond136#2 (work_completion)(&(&bond->mcast_work)->work) &p->pi_lock irq_context: 0 (wq_completion)bond172#2 irq_context: 0 (wq_completion)bond172#2 (work_completion)(&(&slave->notify_work)->work) irq_context: 0 (wq_completion)bond172#2 (work_completion)(&(&slave->notify_work)->work) &obj_hash[i].lock irq_context: 0 (wq_completion)bond172#2 (work_completion)(&(&slave->notify_work)->work) &base->lock irq_context: 0 (wq_completion)bond172#2 (work_completion)(&(&slave->notify_work)->work) &base->lock &obj_hash[i].lock irq_context: 0 (wq_completion)bond172#2 (work_completion)(&(&slave->notify_work)->work) &rq->__lock irq_context: 0 (wq_completion)bond172#2 (work_completion)(&(&slave->notify_work)->work) &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)bond172#2 (work_completion)(&(&bond->mcast_work)->work) irq_context: 0 (wq_completion)bond172#2 (work_completion)(&(&bond->mcast_work)->work) &obj_hash[i].lock irq_context: 0 (wq_completion)bond172#2 (work_completion)(&(&bond->mcast_work)->work) &base->lock irq_context: 0 (wq_completion)bond172#2 (work_completion)(&(&bond->mcast_work)->work) &base->lock &obj_hash[i].lock irq_context: 0 (wq_completion)bond172#2 (work_completion)(&(&bond->mcast_work)->work) &rq->__lock irq_context: 0 (wq_completion)bond172#2 (work_completion)(&(&bond->mcast_work)->work) &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)bond113 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex fs_reclaim irq_context: 0 (wq_completion)bond113 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex fs_reclaim mmu_notifier_invalidate_range_start irq_context: 0 (wq_completion)bond113 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex fs_reclaim mmu_notifier_invalidate_range_start &rq->__lock irq_context: 0 (wq_completion)bond113 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &c->lock irq_context: 0 (wq_completion)bond113 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &rq->__lock irq_context: 0 (wq_completion)bond113 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex rcu_read_lock &bond->stats_lock/1 irq_context: 0 (wq_completion)bond113 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex rcu_read_lock rcu_node_0 irq_context: 0 (wq_completion)bond113 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex rcu_read_lock &rq->__lock irq_context: 0 (wq_completion)bond113 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex rcu_read_lock &ndev->lock irq_context: 0 (wq_completion)bond113 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex nl_table_lock irq_context: 0 (wq_completion)bond113 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)bond113 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &obj_hash[i].lock irq_context: 0 (wq_completion)bond113 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex nl_table_wait.lock irq_context: 0 (wq_completion)bond113 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex net_rwsem irq_context: 0 (wq_completion)bond113 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex net_rwsem &rq->__lock irq_context: 0 (wq_completion)bond113 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex net_rwsem &list->lock#2 irq_context: 0 (wq_completion)bond113 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &tn->lock irq_context: 0 (wq_completion)bond113 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &idev->mc_lock irq_context: 0 (wq_completion)bond113 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &idev->mc_lock fs_reclaim irq_context: 0 (wq_completion)bond113 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &idev->mc_lock fs_reclaim mmu_notifier_invalidate_range_start irq_context: 0 (wq_completion)bond113 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &idev->mc_lock &c->lock irq_context: 0 (wq_completion)bond113 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &idev->mc_lock &n->list_lock irq_context: 0 (wq_completion)bond113 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &idev->mc_lock &n->list_lock &c->lock irq_context: 0 (wq_completion)bond113 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &idev->mc_lock &rq->__lock irq_context: 0 (wq_completion)bond113 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &idev->mc_lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)bond113 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &idev->mc_lock rcu_read_lock &ndev->lock irq_context: 0 (wq_completion)bond113 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &idev->mc_lock rcu_read_lock &c->lock irq_context: 0 (wq_completion)bond113 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &idev->mc_lock rcu_read_lock &dir->lock#2 irq_context: 0 (wq_completion)bond113 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &idev->mc_lock rcu_read_lock &ul->lock#2 irq_context: 0 (wq_completion)bond113 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &idev->mc_lock rcu_read_lock rcu_read_lock &obj_hash[i].lock irq_context: 0 (wq_completion)bond113 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &idev->mc_lock rcu_read_lock rcu_read_lock &dir->lock#2 irq_context: 0 (wq_completion)bond113 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &idev->mc_lock rcu_read_lock rcu_read_lock &tbl->lock irq_context: 0 (wq_completion)bond113 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &idev->mc_lock rcu_read_lock rcu_read_lock &n->lock irq_context: 0 (wq_completion)bond113 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &idev->mc_lock rcu_read_lock rcu_read_lock &____s->seqcount#9 irq_context: 0 (wq_completion)bond113 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &idev->mc_lock rcu_read_lock rcu_read_lock rcu_read_lock_bh rcu_read_lock rcu_read_lock_bh rcu_read_lock &br->multicast_lock irq_context: 0 (wq_completion)bond113 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &idev->mc_lock rcu_read_lock rcu_read_lock rcu_read_lock_bh rcu_read_lock rcu_read_lock_bh rcu_read_lock &br->multicast_lock &obj_hash[i].lock irq_context: 0 (wq_completion)bond113 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &idev->mc_lock rcu_read_lock rcu_read_lock rcu_read_lock_bh rcu_read_lock rcu_read_lock_bh rcu_read_lock &br->multicast_lock &base->lock irq_context: 0 (wq_completion)bond113 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &idev->mc_lock rcu_read_lock rcu_read_lock rcu_read_lock_bh rcu_read_lock rcu_read_lock_bh rcu_read_lock &br->multicast_lock &base->lock &obj_hash[i].lock irq_context: 0 (wq_completion)bond113 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &idev->mc_lock rcu_read_lock rcu_read_lock rcu_read_lock_bh rcu_read_lock rcu_read_lock_bh rcu_read_lock &data->lock irq_context: 0 (wq_completion)bond113 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &idev->mc_lock rcu_read_lock rcu_read_lock rcu_read_lock_bh rcu_read_lock rcu_read_lock_bh rcu_read_lock &obj_hash[i].lock irq_context: 0 (wq_completion)bond113 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &idev->mc_lock rcu_read_lock rcu_read_lock rcu_node_0 irq_context: 0 (wq_completion)bond113 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &idev->mc_lock rcu_read_lock rcu_read_lock &rq->__lock irq_context: 0 (wq_completion)bond113 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &idev->mc_lock rcu_read_lock rcu_read_lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)bond113 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &idev->mc_lock rcu_read_lock rcu_node_0 irq_context: 0 (wq_completion)bond113 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex.wait_lock irq_context: 0 (wq_completion)bond113 (work_completion)(&(&bond->mcast_work)->work) &p->pi_lock irq_context: 0 (wq_completion)bond113 (work_completion)(&(&bond->mcast_work)->work) &p->pi_lock &rq->__lock irq_context: 0 (wq_completion)bond113 (work_completion)(&(&bond->mcast_work)->work) &p->pi_lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)bond156#3 irq_context: 0 (wq_completion)bond156#3 (work_completion)(&(&slave->notify_work)->work) irq_context: 0 (wq_completion)bond156#3 (work_completion)(&(&slave->notify_work)->work) &obj_hash[i].lock irq_context: 0 (wq_completion)bond156#3 (work_completion)(&(&slave->notify_work)->work) &base->lock irq_context: 0 (wq_completion)bond156#3 (work_completion)(&(&slave->notify_work)->work) &base->lock &obj_hash[i].lock irq_context: 0 (wq_completion)bond156#3 (work_completion)(&(&bond->mcast_work)->work) irq_context: 0 (wq_completion)bond156#3 (work_completion)(&(&bond->mcast_work)->work) &obj_hash[i].lock irq_context: 0 (wq_completion)bond156#3 (work_completion)(&(&bond->mcast_work)->work) &base->lock irq_context: 0 (wq_completion)bond156#3 (work_completion)(&(&bond->mcast_work)->work) &base->lock &obj_hash[i].lock irq_context: 0 (wq_completion)bond156#3 (work_completion)(&(&bond->mcast_work)->work) &rq->__lock irq_context: 0 (wq_completion)bond156#3 (work_completion)(&(&bond->mcast_work)->work) &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)bond133#3 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex fs_reclaim irq_context: 0 (wq_completion)bond133#3 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex fs_reclaim mmu_notifier_invalidate_range_start irq_context: 0 (wq_completion)bond133#3 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex rcu_read_lock &bond->stats_lock/1 irq_context: 0 (wq_completion)bond133#3 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex rcu_read_lock &ndev->lock irq_context: 0 (wq_completion)bond133#3 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex nl_table_lock irq_context: 0 (wq_completion)bond133#3 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &obj_hash[i].lock irq_context: 0 (wq_completion)bond133#3 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex nl_table_wait.lock irq_context: 0 (wq_completion)bond133#3 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex nl_table_wait.lock &p->pi_lock irq_context: 0 (wq_completion)bond133#3 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex net_rwsem irq_context: 0 (wq_completion)bond133#3 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex net_rwsem &list->lock#2 irq_context: 0 (wq_completion)bond133#3 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &tn->lock irq_context: 0 (wq_completion)bond133#3 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &idev->mc_lock irq_context: 0 (wq_completion)bond133#3 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &idev->mc_lock fs_reclaim irq_context: 0 (wq_completion)bond133#3 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &idev->mc_lock fs_reclaim mmu_notifier_invalidate_range_start irq_context: 0 (wq_completion)bond133#3 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &idev->mc_lock &c->lock irq_context: 0 (wq_completion)bond133#3 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &idev->mc_lock rcu_read_lock &ndev->lock irq_context: 0 (wq_completion)bond133#3 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &idev->mc_lock rcu_read_lock &dir->lock#2 irq_context: 0 (wq_completion)bond133#3 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &idev->mc_lock rcu_read_lock &ul->lock#2 irq_context: 0 (wq_completion)bond133#3 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &idev->mc_lock rcu_read_lock rcu_read_lock &____s->seqcount#10 irq_context: 0 (wq_completion)bond133#3 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &idev->mc_lock rcu_read_lock rcu_read_lock rcu_read_lock_bh rcu_read_lock rcu_read_lock_bh rcu_read_lock &br->multicast_lock irq_context: 0 (wq_completion)bond133#3 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &idev->mc_lock rcu_read_lock rcu_read_lock rcu_read_lock_bh rcu_read_lock rcu_read_lock_bh rcu_read_lock &br->multicast_lock &obj_hash[i].lock irq_context: 0 (wq_completion)bond133#3 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &idev->mc_lock rcu_read_lock rcu_read_lock rcu_read_lock_bh rcu_read_lock rcu_read_lock_bh rcu_read_lock &br->multicast_lock &base->lock irq_context: 0 (wq_completion)bond133#3 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &idev->mc_lock rcu_read_lock rcu_read_lock rcu_read_lock_bh rcu_read_lock rcu_read_lock_bh rcu_read_lock &br->multicast_lock &base->lock &obj_hash[i].lock irq_context: 0 (wq_completion)bond133#3 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &idev->mc_lock rcu_read_lock rcu_read_lock rcu_read_lock_bh rcu_read_lock rcu_read_lock_bh rcu_read_lock &data->lock irq_context: 0 (wq_completion)bond133#3 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &idev->mc_lock rcu_read_lock rcu_read_lock rcu_read_lock_bh rcu_read_lock rcu_read_lock_bh rcu_read_lock &obj_hash[i].lock irq_context: 0 (wq_completion)bond133#3 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex.wait_lock irq_context: 0 (wq_completion)bond133#3 (work_completion)(&(&bond->mcast_work)->work) &p->pi_lock irq_context: 0 (wq_completion)bond145#3 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex fs_reclaim irq_context: 0 (wq_completion)bond145#3 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex fs_reclaim mmu_notifier_invalidate_range_start irq_context: 0 (wq_completion)bond145#3 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex rcu_read_lock &bond->stats_lock/1 irq_context: 0 (wq_completion)bond145#3 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex rcu_read_lock &ndev->lock irq_context: 0 (wq_completion)bond145#3 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex nl_table_lock irq_context: 0 (wq_completion)bond145#3 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &obj_hash[i].lock irq_context: 0 (wq_completion)bond145#3 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex nl_table_wait.lock irq_context: 0 (wq_completion)bond145#3 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex net_rwsem irq_context: 0 (wq_completion)bond145#3 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex net_rwsem &list->lock#2 irq_context: 0 (wq_completion)bond145#3 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &tn->lock irq_context: 0 (wq_completion)bond145#3 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &idev->mc_lock irq_context: 0 (wq_completion)bond145#3 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &idev->mc_lock fs_reclaim irq_context: 0 (wq_completion)bond145#3 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &idev->mc_lock fs_reclaim mmu_notifier_invalidate_range_start irq_context: 0 (wq_completion)bond145#3 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &idev->mc_lock &c->lock irq_context: 0 (wq_completion)bond145#3 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &idev->mc_lock rcu_read_lock &ndev->lock irq_context: 0 (wq_completion)bond145#3 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &idev->mc_lock rcu_read_lock &c->lock irq_context: 0 (wq_completion)bond145#3 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &idev->mc_lock rcu_read_lock &dir->lock#2 irq_context: 0 (wq_completion)bond145#3 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &idev->mc_lock rcu_read_lock &ul->lock#2 irq_context: 0 (wq_completion)bond145#3 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &idev->mc_lock rcu_read_lock rcu_read_lock &____s->seqcount#10 irq_context: 0 (wq_completion)bond145#3 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &idev->mc_lock rcu_read_lock rcu_read_lock rcu_read_lock_bh rcu_read_lock rcu_read_lock_bh rcu_read_lock &br->multicast_lock irq_context: 0 (wq_completion)bond145#3 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &idev->mc_lock rcu_read_lock rcu_read_lock rcu_read_lock_bh rcu_read_lock rcu_read_lock_bh rcu_read_lock &br->multicast_lock &obj_hash[i].lock irq_context: 0 (wq_completion)bond145#3 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &idev->mc_lock rcu_read_lock rcu_read_lock rcu_read_lock_bh rcu_read_lock rcu_read_lock_bh rcu_read_lock &br->multicast_lock &base->lock irq_context: 0 (wq_completion)bond145#3 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &idev->mc_lock rcu_read_lock rcu_read_lock rcu_read_lock_bh rcu_read_lock rcu_read_lock_bh rcu_read_lock &br->multicast_lock &base->lock &obj_hash[i].lock irq_context: 0 (wq_completion)bond145#3 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &idev->mc_lock rcu_read_lock rcu_read_lock rcu_read_lock_bh rcu_read_lock rcu_read_lock_bh rcu_read_lock &data->lock irq_context: 0 (wq_completion)bond145#3 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &idev->mc_lock rcu_read_lock rcu_read_lock rcu_read_lock_bh rcu_read_lock rcu_read_lock_bh rcu_read_lock &obj_hash[i].lock irq_context: 0 (wq_completion)bond145#3 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &idev->mc_lock rcu_read_lock rcu_read_lock rcu_node_0 irq_context: 0 (wq_completion)bond145#3 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &idev->mc_lock rcu_read_lock rcu_read_lock &rq->__lock irq_context: 0 (wq_completion)bond145#3 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &idev->mc_lock rcu_read_lock rcu_node_0 irq_context: 0 (wq_completion)bond145#3 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &idev->mc_lock rcu_read_lock &rcu_state.gp_wq irq_context: 0 (wq_completion)bond145#3 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &idev->mc_lock rcu_read_lock &rcu_state.gp_wq &p->pi_lock irq_context: 0 (wq_completion)bond145#3 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &idev->mc_lock rcu_read_lock &rcu_state.gp_wq &p->pi_lock &rq->__lock irq_context: 0 (wq_completion)bond145#3 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &idev->mc_lock rcu_read_lock &rcu_state.gp_wq &p->pi_lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)bond145#3 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &idev->mc_lock rcu_read_lock &rq->__lock irq_context: 0 (wq_completion)bond145#3 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &idev->mc_lock rcu_read_lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)bond145#3 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex.wait_lock irq_context: 0 (wq_completion)bond145#3 (work_completion)(&(&bond->mcast_work)->work) &p->pi_lock irq_context: 0 (wq_completion)bond145#3 (work_completion)(&(&bond->mcast_work)->work) &p->pi_lock &rq->__lock irq_context: 0 (wq_completion)bond145#3 (work_completion)(&(&bond->mcast_work)->work) &p->pi_lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: softirq &(&ipvs->defense_work)->timer rcu_read_lock &pool->lock fill_pool_map-wait-type-override pool_lock irq_context: 0 (wq_completion)bond148#3 (work_completion)(&(&bond->mcast_work)->work) &rq->__lock irq_context: 0 (wq_completion)bond148#3 (work_completion)(&(&bond->mcast_work)->work) &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)bond185 irq_context: 0 (wq_completion)bond185 (work_completion)(&(&slave->notify_work)->work) irq_context: 0 (wq_completion)bond185 (work_completion)(&(&slave->notify_work)->work) &obj_hash[i].lock irq_context: 0 (wq_completion)bond185 (work_completion)(&(&slave->notify_work)->work) &base->lock irq_context: 0 (wq_completion)bond185 (work_completion)(&(&slave->notify_work)->work) &base->lock &obj_hash[i].lock irq_context: 0 (wq_completion)bond185 (work_completion)(&(&slave->notify_work)->work) &rq->__lock irq_context: 0 (wq_completion)bond185 (work_completion)(&(&slave->notify_work)->work) &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)bond148#3 (work_completion)(&(&slave->notify_work)->work) rtnl_mutex net_rwsem irq_context: 0 (wq_completion)bond148#3 (work_completion)(&(&slave->notify_work)->work) rtnl_mutex net_rwsem &list->lock#2 irq_context: 0 (wq_completion)bond148#3 (work_completion)(&(&slave->notify_work)->work) rtnl_mutex &tn->lock irq_context: 0 (wq_completion)bond148#3 (work_completion)(&(&slave->notify_work)->work) rtnl_mutex.wait_lock irq_context: 0 (wq_completion)bond148#3 (work_completion)(&(&slave->notify_work)->work) &p->pi_lock irq_context: 0 (wq_completion)bond156#3 (work_completion)(&(&slave->notify_work)->work) rtnl_mutex net_rwsem irq_context: 0 (wq_completion)bond156#3 (work_completion)(&(&slave->notify_work)->work) rtnl_mutex net_rwsem &list->lock#2 irq_context: 0 (wq_completion)bond156#3 (work_completion)(&(&slave->notify_work)->work) rtnl_mutex &tn->lock irq_context: 0 (wq_completion)bond156#3 (work_completion)(&(&slave->notify_work)->work) rtnl_mutex.wait_lock irq_context: 0 (wq_completion)bond156#3 (work_completion)(&(&slave->notify_work)->work) &p->pi_lock irq_context: 0 (wq_completion)bond153#3 (work_completion)(&(&slave->notify_work)->work) rtnl_mutex net_rwsem irq_context: 0 (wq_completion)bond153#3 (work_completion)(&(&slave->notify_work)->work) rtnl_mutex net_rwsem &list->lock#2 irq_context: 0 (wq_completion)bond153#3 (work_completion)(&(&slave->notify_work)->work) rtnl_mutex &tn->lock irq_context: 0 (wq_completion)bond153#3 (work_completion)(&(&slave->notify_work)->work) rtnl_mutex.wait_lock irq_context: 0 (wq_completion)bond153#3 (work_completion)(&(&slave->notify_work)->work) &p->pi_lock irq_context: 0 (wq_completion)bond52#3 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex fs_reclaim irq_context: 0 (wq_completion)bond52#3 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex fs_reclaim mmu_notifier_invalidate_range_start irq_context: 0 (wq_completion)bond52#3 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &c->lock irq_context: 0 (wq_completion)bond52#3 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex rcu_read_lock &bond->stats_lock/1 irq_context: 0 (wq_completion)bond52#3 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex rcu_read_lock &ndev->lock irq_context: 0 (wq_completion)bond52#3 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex nl_table_lock irq_context: 0 (wq_completion)bond52#3 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &obj_hash[i].lock irq_context: 0 (wq_completion)bond52#3 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex nl_table_wait.lock irq_context: 0 (wq_completion)bond52#3 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex net_rwsem irq_context: 0 (wq_completion)bond52#3 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex net_rwsem &list->lock#2 irq_context: 0 (wq_completion)bond52#3 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &tn->lock irq_context: 0 (wq_completion)bond52#3 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &idev->mc_lock irq_context: 0 (wq_completion)bond52#3 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &idev->mc_lock &rq->__lock irq_context: 0 (wq_completion)bond52#3 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &idev->mc_lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)bond185 (work_completion)(&(&bond->mcast_work)->work) irq_context: 0 (wq_completion)bond185 (work_completion)(&(&bond->mcast_work)->work) &obj_hash[i].lock irq_context: 0 (wq_completion)bond185 (work_completion)(&(&bond->mcast_work)->work) &base->lock irq_context: 0 (wq_completion)bond185 (work_completion)(&(&bond->mcast_work)->work) &base->lock &obj_hash[i].lock irq_context: 0 (wq_completion)bond185 (work_completion)(&(&bond->mcast_work)->work) &rq->__lock irq_context: 0 (wq_completion)bond185 (work_completion)(&(&bond->mcast_work)->work) &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)bond52#3 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &idev->mc_lock fs_reclaim irq_context: 0 (wq_completion)bond52#3 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &idev->mc_lock fs_reclaim mmu_notifier_invalidate_range_start irq_context: 0 (wq_completion)bond52#3 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &idev->mc_lock rcu_read_lock &ndev->lock irq_context: 0 (wq_completion)bond52#3 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &idev->mc_lock rcu_read_lock rcu_node_0 irq_context: 0 (wq_completion)bond52#3 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &idev->mc_lock rcu_read_lock &rq->__lock irq_context: 0 (wq_completion)bond52#3 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &idev->mc_lock rcu_read_lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)bond52#3 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &idev->mc_lock rcu_read_lock &dir->lock#2 irq_context: 0 (wq_completion)bond52#3 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &idev->mc_lock rcu_read_lock &ul->lock#2 irq_context: 0 (wq_completion)bond52#3 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &idev->mc_lock rcu_read_lock rcu_read_lock &____s->seqcount#12 irq_context: 0 (wq_completion)bond52#3 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &idev->mc_lock rcu_read_lock rcu_read_lock &____s->seqcount#16 irq_context: 0 (wq_completion)bond52#3 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &idev->mc_lock rcu_read_lock &obj_hash[i].lock irq_context: 0 (wq_completion)bond52#3 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &idev->mc_lock rcu_read_lock &data->lock irq_context: 0 (wq_completion)bond52#3 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex.wait_lock irq_context: 0 (wq_completion)bond52#3 (work_completion)(&(&bond->mcast_work)->work) &p->pi_lock irq_context: 0 (wq_completion)bond52#3 (work_completion)(&(&bond->mcast_work)->work) &p->pi_lock &rq->__lock irq_context: 0 (wq_completion)bond52#3 (work_completion)(&(&bond->mcast_work)->work) &p->pi_lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)bond55#2 irq_context: 0 (wq_completion)bond55#2 (work_completion)(&(&slave->notify_work)->work) irq_context: 0 (wq_completion)bond55#2 (work_completion)(&(&slave->notify_work)->work) &obj_hash[i].lock irq_context: 0 (wq_completion)bond55#2 (work_completion)(&(&slave->notify_work)->work) &base->lock irq_context: 0 (wq_completion)bond55#2 (work_completion)(&(&slave->notify_work)->work) &base->lock &obj_hash[i].lock irq_context: 0 (wq_completion)bond55#2 (work_completion)(&(&slave->notify_work)->work) &rq->__lock irq_context: 0 (wq_completion)bond55#2 (work_completion)(&(&bond->mcast_work)->work) irq_context: 0 (wq_completion)bond55#2 (work_completion)(&(&bond->mcast_work)->work) &obj_hash[i].lock irq_context: 0 (wq_completion)bond55#2 (work_completion)(&(&bond->mcast_work)->work) &base->lock irq_context: 0 (wq_completion)bond55#2 (work_completion)(&(&bond->mcast_work)->work) &base->lock &obj_hash[i].lock irq_context: 0 (wq_completion)bond55#2 (work_completion)(&(&bond->mcast_work)->work) &rq->__lock irq_context: 0 (wq_completion)bond55#2 (work_completion)(&(&bond->mcast_work)->work) &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)bond173#2 irq_context: 0 (wq_completion)bond173#2 (work_completion)(&(&slave->notify_work)->work) irq_context: 0 (wq_completion)bond173#2 (work_completion)(&(&slave->notify_work)->work) &obj_hash[i].lock irq_context: 0 (wq_completion)bond173#2 (work_completion)(&(&slave->notify_work)->work) &base->lock irq_context: 0 (wq_completion)bond173#2 (work_completion)(&(&slave->notify_work)->work) &base->lock &obj_hash[i].lock irq_context: 0 (wq_completion)bond173#2 (work_completion)(&(&slave->notify_work)->work) &rq->__lock irq_context: 0 (wq_completion)bond173#2 (work_completion)(&(&slave->notify_work)->work) &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)bond163 (work_completion)(&(&bond->mcast_work)->work) &rq->__lock irq_context: 0 (wq_completion)bond163 (work_completion)(&(&bond->mcast_work)->work) &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)bond173#2 (work_completion)(&(&bond->mcast_work)->work) irq_context: 0 (wq_completion)bond173#2 (work_completion)(&(&bond->mcast_work)->work) &obj_hash[i].lock irq_context: 0 (wq_completion)bond173#2 (work_completion)(&(&bond->mcast_work)->work) &base->lock irq_context: 0 (wq_completion)bond173#2 (work_completion)(&(&bond->mcast_work)->work) &base->lock &obj_hash[i].lock irq_context: 0 (wq_completion)bond173#2 (work_completion)(&(&bond->mcast_work)->work) &rq->__lock irq_context: 0 (wq_completion)bond173#2 (work_completion)(&(&bond->mcast_work)->work) &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)bond182 (work_completion)(&(&slave->notify_work)->work) &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)bond171#2 (work_completion)(&(&bond->mcast_work)->work) &rq->__lock irq_context: 0 (wq_completion)bond171#2 (work_completion)(&(&bond->mcast_work)->work) &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)bond173#2 (work_completion)(&(&slave->notify_work)->work) rtnl_mutex net_rwsem irq_context: 0 (wq_completion)bond173#2 (work_completion)(&(&slave->notify_work)->work) rtnl_mutex net_rwsem &list->lock#2 irq_context: 0 (wq_completion)bond173#2 (work_completion)(&(&slave->notify_work)->work) rtnl_mutex &tn->lock irq_context: 0 (wq_completion)bond173#2 (work_completion)(&(&slave->notify_work)->work) rtnl_mutex.wait_lock irq_context: 0 (wq_completion)bond173#2 (work_completion)(&(&slave->notify_work)->work) &p->pi_lock irq_context: 0 (wq_completion)bond149#3 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex fs_reclaim irq_context: 0 (wq_completion)bond149#3 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex fs_reclaim mmu_notifier_invalidate_range_start irq_context: 0 (wq_completion)bond149#3 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &c->lock irq_context: 0 (wq_completion)bond149#3 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex rcu_read_lock &bond->stats_lock/1 irq_context: 0 (wq_completion)bond149#3 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex rcu_read_lock &ndev->lock irq_context: 0 (wq_completion)bond149#3 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex nl_table_lock irq_context: 0 (wq_completion)bond149#3 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &obj_hash[i].lock irq_context: 0 (wq_completion)bond149#3 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex nl_table_wait.lock irq_context: 0 (wq_completion)bond149#3 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &rq->__lock irq_context: 0 (wq_completion)bond149#3 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)bond149#3 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex net_rwsem irq_context: 0 (wq_completion)bond149#3 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex net_rwsem &list->lock#2 irq_context: 0 (wq_completion)bond149#3 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &tn->lock irq_context: 0 (wq_completion)bond149#3 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &idev->mc_lock irq_context: 0 (wq_completion)bond149#3 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &idev->mc_lock fs_reclaim irq_context: 0 (wq_completion)bond149#3 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &idev->mc_lock fs_reclaim mmu_notifier_invalidate_range_start irq_context: 0 (wq_completion)bond149#3 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &idev->mc_lock rcu_read_lock &ndev->lock irq_context: 0 (wq_completion)bond149#3 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &idev->mc_lock rcu_read_lock &c->lock irq_context: 0 (wq_completion)bond149#3 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &idev->mc_lock rcu_read_lock &dir->lock#2 irq_context: 0 (wq_completion)bond149#3 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &idev->mc_lock rcu_read_lock &ul->lock#2 irq_context: 0 (wq_completion)bond149#3 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &idev->mc_lock rcu_read_lock rcu_read_lock &____s->seqcount#10 irq_context: 0 (wq_completion)bond149#3 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &idev->mc_lock rcu_read_lock rcu_read_lock rcu_node_0 irq_context: 0 (wq_completion)bond149#3 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &idev->mc_lock rcu_read_lock rcu_read_lock &rq->__lock irq_context: 0 (wq_completion)bond149#3 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &idev->mc_lock rcu_read_lock rcu_read_lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)bond149#3 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &idev->mc_lock rcu_read_lock rcu_read_lock rcu_read_lock_bh rcu_read_lock rcu_read_lock_bh rcu_read_lock &br->multicast_lock irq_context: 0 (wq_completion)bond149#3 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &idev->mc_lock rcu_read_lock rcu_read_lock rcu_read_lock_bh rcu_read_lock rcu_read_lock_bh rcu_read_lock &br->multicast_lock &obj_hash[i].lock irq_context: 0 (wq_completion)bond149#3 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &idev->mc_lock rcu_read_lock rcu_read_lock rcu_read_lock_bh rcu_read_lock rcu_read_lock_bh rcu_read_lock &br->multicast_lock &base->lock irq_context: 0 (wq_completion)bond149#3 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &idev->mc_lock rcu_read_lock rcu_read_lock rcu_read_lock_bh rcu_read_lock rcu_read_lock_bh rcu_read_lock &data->lock irq_context: 0 (wq_completion)bond149#3 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &idev->mc_lock rcu_read_lock rcu_read_lock rcu_read_lock_bh rcu_read_lock rcu_read_lock_bh rcu_read_lock &obj_hash[i].lock irq_context: 0 (wq_completion)bond149#3 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &idev->mc_lock rcu_read_lock rcu_node_0 irq_context: 0 (wq_completion)bond149#3 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex.wait_lock irq_context: 0 (wq_completion)bond149#3 (work_completion)(&(&bond->mcast_work)->work) &p->pi_lock irq_context: 0 (wq_completion)bond154#3 (work_completion)(&(&slave->notify_work)->work) rtnl_mutex net_rwsem irq_context: 0 (wq_completion)bond154#3 (work_completion)(&(&slave->notify_work)->work) rtnl_mutex net_rwsem &rq->__lock irq_context: 0 (wq_completion)bond154#3 (work_completion)(&(&slave->notify_work)->work) rtnl_mutex net_rwsem &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)bond154#3 (work_completion)(&(&slave->notify_work)->work) rtnl_mutex net_rwsem &list->lock#2 irq_context: 0 (wq_completion)bond154#3 (work_completion)(&(&slave->notify_work)->work) rtnl_mutex &tn->lock irq_context: 0 (wq_completion)bond154#3 (work_completion)(&(&slave->notify_work)->work) rtnl_mutex.wait_lock irq_context: 0 (wq_completion)bond154#3 (work_completion)(&(&slave->notify_work)->work) &p->pi_lock irq_context: 0 (wq_completion)bond165#2 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex fs_reclaim irq_context: 0 (wq_completion)bond165#2 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex fs_reclaim mmu_notifier_invalidate_range_start irq_context: 0 (wq_completion)bond165#2 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &c->lock irq_context: 0 (wq_completion)bond165#2 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex rcu_read_lock &bond->stats_lock/1 irq_context: 0 (wq_completion)bond165#2 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex rcu_read_lock &ndev->lock irq_context: 0 (wq_completion)bond165#2 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex nl_table_lock irq_context: 0 (wq_completion)bond165#2 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &obj_hash[i].lock irq_context: 0 (wq_completion)bond165#2 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &rq->__lock irq_context: 0 (wq_completion)bond165#2 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)bond165#2 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex nl_table_wait.lock irq_context: 0 (wq_completion)bond165#2 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex net_rwsem irq_context: 0 (wq_completion)bond165#2 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex net_rwsem &list->lock#2 irq_context: 0 (wq_completion)bond165#2 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &tn->lock irq_context: 0 (wq_completion)bond165#2 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &idev->mc_lock irq_context: 0 (wq_completion)bond165#2 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &idev->mc_lock fs_reclaim irq_context: 0 (wq_completion)bond165#2 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &idev->mc_lock fs_reclaim mmu_notifier_invalidate_range_start irq_context: 0 (wq_completion)bond165#2 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &idev->mc_lock &c->lock irq_context: 0 (wq_completion)bond165#2 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &idev->mc_lock rcu_read_lock &ndev->lock irq_context: 0 (wq_completion)bond165#2 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &idev->mc_lock rcu_read_lock &dir->lock#2 irq_context: 0 (wq_completion)bond165#2 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &idev->mc_lock rcu_read_lock &ul->lock#2 irq_context: 0 (wq_completion)bond165#2 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &idev->mc_lock rcu_read_lock rcu_read_lock rcu_node_0 irq_context: 0 (wq_completion)bond165#2 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &idev->mc_lock rcu_read_lock rcu_read_lock &rq->__lock irq_context: 0 (wq_completion)bond165#2 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &idev->mc_lock rcu_read_lock rcu_read_lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)bond165#2 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &idev->mc_lock rcu_read_lock rcu_read_lock &____s->seqcount#10 irq_context: 0 (wq_completion)bond165#2 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &idev->mc_lock rcu_read_lock rcu_read_lock rcu_read_lock_bh rcu_read_lock rcu_read_lock_bh rcu_read_lock &br->multicast_lock irq_context: 0 (wq_completion)bond165#2 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &idev->mc_lock rcu_read_lock rcu_read_lock rcu_read_lock_bh rcu_read_lock rcu_read_lock_bh rcu_read_lock &br->multicast_lock &obj_hash[i].lock irq_context: 0 (wq_completion)bond165#2 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &idev->mc_lock rcu_read_lock rcu_read_lock rcu_read_lock_bh rcu_read_lock rcu_read_lock_bh rcu_read_lock &br->multicast_lock &base->lock irq_context: 0 (wq_completion)bond165#2 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &idev->mc_lock rcu_read_lock rcu_read_lock rcu_read_lock_bh rcu_read_lock rcu_read_lock_bh rcu_read_lock &data->lock irq_context: 0 (wq_completion)bond165#2 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &idev->mc_lock rcu_read_lock rcu_read_lock rcu_read_lock_bh rcu_read_lock rcu_read_lock_bh rcu_read_lock &obj_hash[i].lock irq_context: 0 (wq_completion)bond165#2 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &idev->mc_lock rcu_read_lock rcu_node_0 irq_context: 0 (wq_completion)bond165#2 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex.wait_lock irq_context: 0 (wq_completion)bond142#2 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex fs_reclaim irq_context: 0 (wq_completion)bond165#2 (work_completion)(&(&bond->mcast_work)->work) &p->pi_lock irq_context: 0 (wq_completion)bond142#2 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex fs_reclaim &rq->__lock irq_context: 0 (wq_completion)bond142#2 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex fs_reclaim &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)bond142#2 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex fs_reclaim mmu_notifier_invalidate_range_start irq_context: 0 (wq_completion)bond142#2 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex rcu_read_lock &bond->stats_lock/1 irq_context: 0 (wq_completion)bond142#2 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex rcu_read_lock rcu_node_0 irq_context: 0 (wq_completion)bond142#2 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex rcu_read_lock &rq->__lock irq_context: 0 (wq_completion)bond142#2 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex rcu_read_lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)writeback (work_completion)(&(&wb->dwork)->work) &type->s_umount_key#32 &sbi->s_writepages_rwsem jbd2_handle &ei->i_data_sem &ei->i_prealloc_lock &pa->pa_lock#2 irq_context: 0 (wq_completion)writeback (work_completion)(&(&wb->dwork)->work) &type->s_umount_key#32 &sbi->s_writepages_rwsem jbd2_handle &ei->i_data_sem &pa->pa_lock#2 irq_context: 0 (wq_completion)writeback (work_completion)(&(&wb->dwork)->work) &type->s_umount_key#32 &sbi->s_writepages_rwsem jbd2_handle &mapping->i_mmap_rwsem mmu_notifier_invalidate_range_start irq_context: 0 (wq_completion)writeback (work_completion)(&(&wb->dwork)->work) &type->s_umount_key#32 &sbi->s_writepages_rwsem jbd2_handle &mapping->i_mmap_rwsem ptlock_ptr(page)#2 irq_context: 0 (wq_completion)bond142#2 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex rcu_read_lock &ndev->lock irq_context: 0 (wq_completion)bond142#2 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex nl_table_lock irq_context: 0 (wq_completion)bond142#2 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &obj_hash[i].lock irq_context: 0 (wq_completion)bond142#2 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex nl_table_wait.lock irq_context: 0 (wq_completion)bond142#2 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex net_rwsem irq_context: 0 (wq_completion)bond142#2 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex net_rwsem &list->lock#2 irq_context: 0 (wq_completion)bond142#2 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &tn->lock irq_context: 0 (wq_completion)bond142#2 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &idev->mc_lock irq_context: 0 (wq_completion)bond142#2 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &idev->mc_lock fs_reclaim irq_context: 0 (wq_completion)bond142#2 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &idev->mc_lock fs_reclaim &rq->__lock irq_context: 0 (wq_completion)bond142#2 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &idev->mc_lock fs_reclaim mmu_notifier_invalidate_range_start irq_context: 0 (wq_completion)bond142#2 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &idev->mc_lock rcu_read_lock &ndev->lock irq_context: 0 (wq_completion)bond142#2 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &idev->mc_lock rcu_read_lock &c->lock irq_context: 0 (wq_completion)bond142#2 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &idev->mc_lock rcu_read_lock &dir->lock#2 irq_context: 0 (wq_completion)bond142#2 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &idev->mc_lock rcu_read_lock &ul->lock#2 irq_context: 0 (wq_completion)bond142#2 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &idev->mc_lock rcu_read_lock rcu_read_lock &____s->seqcount#10 irq_context: 0 (wq_completion)bond142#2 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &idev->mc_lock rcu_read_lock rcu_read_lock rcu_read_lock_bh rcu_read_lock rcu_read_lock_bh rcu_read_lock &br->multicast_lock irq_context: 0 (wq_completion)bond142#2 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &idev->mc_lock rcu_read_lock rcu_read_lock rcu_read_lock_bh rcu_read_lock rcu_read_lock_bh rcu_read_lock &br->multicast_lock &obj_hash[i].lock irq_context: 0 (wq_completion)bond142#2 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &idev->mc_lock rcu_read_lock rcu_read_lock rcu_read_lock_bh rcu_read_lock rcu_read_lock_bh rcu_read_lock &br->multicast_lock &base->lock irq_context: 0 (wq_completion)bond142#2 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &idev->mc_lock rcu_read_lock rcu_read_lock rcu_read_lock_bh rcu_read_lock rcu_read_lock_bh rcu_read_lock &br->multicast_lock &base->lock &obj_hash[i].lock irq_context: 0 (wq_completion)bond142#2 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &idev->mc_lock rcu_read_lock rcu_read_lock rcu_read_lock_bh rcu_read_lock rcu_read_lock_bh rcu_read_lock &data->lock irq_context: 0 (wq_completion)bond142#2 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &idev->mc_lock rcu_read_lock rcu_read_lock rcu_read_lock_bh rcu_read_lock rcu_read_lock_bh rcu_read_lock &obj_hash[i].lock irq_context: 0 (wq_completion)bond142#2 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &idev->mc_lock rcu_read_lock rcu_read_lock rcu_node_0 irq_context: 0 (wq_completion)bond142#2 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &idev->mc_lock rcu_read_lock rcu_read_lock &rq->__lock irq_context: 0 (wq_completion)bond142#2 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &idev->mc_lock rcu_read_lock rcu_read_lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)writeback (work_completion)(&(&wb->dwork)->work) &type->s_umount_key#32 &sbi->s_writepages_rwsem jbd2_handle rcu_read_lock &rcu_state.gp_wq irq_context: 0 (wq_completion)writeback (work_completion)(&(&wb->dwork)->work) &type->s_umount_key#32 &sbi->s_writepages_rwsem jbd2_handle rcu_read_lock &rcu_state.gp_wq &p->pi_lock irq_context: 0 (wq_completion)writeback (work_completion)(&(&wb->dwork)->work) &type->s_umount_key#32 &sbi->s_writepages_rwsem jbd2_handle rcu_read_lock &rcu_state.gp_wq &p->pi_lock &rq->__lock irq_context: 0 (wq_completion)writeback (work_completion)(&(&wb->dwork)->work) &type->s_umount_key#32 &sbi->s_writepages_rwsem jbd2_handle rcu_read_lock &rcu_state.gp_wq &p->pi_lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)bond142#2 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &idev->mc_lock rcu_read_lock rcu_node_0 irq_context: 0 (wq_completion)bond142#2 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex.wait_lock irq_context: 0 (wq_completion)bond142#2 (work_completion)(&(&bond->mcast_work)->work) &p->pi_lock irq_context: 0 (wq_completion)bond142#2 (work_completion)(&(&bond->mcast_work)->work) &p->pi_lock &rq->__lock irq_context: 0 (wq_completion)bond142#2 (work_completion)(&(&bond->mcast_work)->work) &p->pi_lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)bond142#2 (work_completion)(&(&bond->mcast_work)->work) &rq->__lock irq_context: 0 (wq_completion)bond142#2 (work_completion)(&(&bond->mcast_work)->work) &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)bond157#3 irq_context: 0 (wq_completion)bond157#3 (work_completion)(&(&slave->notify_work)->work) irq_context: 0 (wq_completion)bond157#3 (work_completion)(&(&slave->notify_work)->work) &obj_hash[i].lock irq_context: 0 (wq_completion)bond157#3 (work_completion)(&(&slave->notify_work)->work) &base->lock irq_context: 0 (wq_completion)bond157#3 (work_completion)(&(&slave->notify_work)->work) &base->lock &obj_hash[i].lock irq_context: 0 (wq_completion)bond35 &rq->__lock irq_context: 0 (wq_completion)bond35 &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)bond157#3 (work_completion)(&(&bond->mcast_work)->work) irq_context: 0 (wq_completion)bond157#3 (work_completion)(&(&bond->mcast_work)->work) &obj_hash[i].lock irq_context: 0 (wq_completion)bond157#3 (work_completion)(&(&bond->mcast_work)->work) &base->lock irq_context: 0 (wq_completion)bond157#3 (work_completion)(&(&bond->mcast_work)->work) &base->lock &obj_hash[i].lock irq_context: 0 (wq_completion)bond50#2 (work_completion)(&(&slave->notify_work)->work) rtnl_mutex net_rwsem irq_context: 0 (wq_completion)bond50#2 (work_completion)(&(&slave->notify_work)->work) rtnl_mutex net_rwsem &list->lock#2 irq_context: 0 (wq_completion)bond50#2 (work_completion)(&(&slave->notify_work)->work) rtnl_mutex &tn->lock irq_context: 0 (wq_completion)bond50#2 (work_completion)(&(&slave->notify_work)->work) rtnl_mutex.wait_lock irq_context: 0 (wq_completion)bond50#2 (work_completion)(&(&slave->notify_work)->work) &p->pi_lock irq_context: 0 (wq_completion)ipv6_addrconf (work_completion)(&(&ifa->dad_work)->work) rtnl_mutex rcu_read_lock rcu_read_lock rcu_read_lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)bond186 irq_context: 0 (wq_completion)bond186 (work_completion)(&(&slave->notify_work)->work) irq_context: 0 (wq_completion)bond186 (work_completion)(&(&slave->notify_work)->work) &obj_hash[i].lock irq_context: 0 (wq_completion)bond186 (work_completion)(&(&slave->notify_work)->work) &base->lock irq_context: 0 (wq_completion)bond186 (work_completion)(&(&slave->notify_work)->work) &base->lock &obj_hash[i].lock irq_context: 0 (wq_completion)bond186 (work_completion)(&(&slave->notify_work)->work) rtnl_mutex net_rwsem irq_context: 0 (wq_completion)bond186 (work_completion)(&(&slave->notify_work)->work) rtnl_mutex net_rwsem &list->lock#2 irq_context: 0 (wq_completion)bond186 (work_completion)(&(&slave->notify_work)->work) rtnl_mutex &tn->lock irq_context: 0 (wq_completion)bond186 (work_completion)(&(&slave->notify_work)->work) rtnl_mutex.wait_lock irq_context: 0 (wq_completion)bond186 (work_completion)(&(&slave->notify_work)->work) &p->pi_lock irq_context: 0 (wq_completion)bond157#3 (work_completion)(&(&slave->notify_work)->work) rtnl_mutex net_rwsem irq_context: 0 (wq_completion)bond157#3 (work_completion)(&(&slave->notify_work)->work) rtnl_mutex net_rwsem &list->lock#2 irq_context: 0 (wq_completion)bond157#3 (work_completion)(&(&slave->notify_work)->work) rtnl_mutex &tn->lock irq_context: 0 (wq_completion)bond157#3 (work_completion)(&(&slave->notify_work)->work) rtnl_mutex.wait_lock irq_context: 0 (wq_completion)bond157#3 (work_completion)(&(&slave->notify_work)->work) &p->pi_lock irq_context: 0 (wq_completion)bond129#2 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex fs_reclaim irq_context: 0 (wq_completion)bond129#2 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex fs_reclaim mmu_notifier_invalidate_range_start irq_context: 0 (wq_completion)bond129#2 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex rcu_read_lock &bond->stats_lock/1 irq_context: 0 (wq_completion)bond129#2 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex rcu_read_lock &ndev->lock irq_context: 0 (wq_completion)bond129#2 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex nl_table_lock irq_context: 0 (wq_completion)bond129#2 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &obj_hash[i].lock irq_context: 0 (wq_completion)bond129#2 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex nl_table_wait.lock irq_context: 0 (wq_completion)bond129#2 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex net_rwsem irq_context: 0 (wq_completion)bond129#2 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex net_rwsem &list->lock#2 irq_context: 0 (wq_completion)bond129#2 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &tn->lock irq_context: 0 (wq_completion)bond129#2 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &idev->mc_lock irq_context: 0 (wq_completion)bond129#2 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &idev->mc_lock fs_reclaim irq_context: 0 (wq_completion)bond129#2 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &idev->mc_lock fs_reclaim mmu_notifier_invalidate_range_start irq_context: 0 (wq_completion)bond129#2 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &idev->mc_lock &c->lock irq_context: 0 (wq_completion)bond129#2 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &idev->mc_lock rcu_read_lock &ndev->lock irq_context: 0 (wq_completion)bond129#2 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &idev->mc_lock rcu_read_lock &c->lock irq_context: 0 (wq_completion)bond129#2 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &idev->mc_lock rcu_read_lock &n->list_lock irq_context: 0 (wq_completion)bond129#2 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &idev->mc_lock rcu_read_lock &n->list_lock &c->lock irq_context: 0 (wq_completion)bond129#2 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &idev->mc_lock rcu_read_lock rcu_node_0 irq_context: 0 (wq_completion)bond129#2 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &idev->mc_lock rcu_read_lock &rq->__lock irq_context: 0 (wq_completion)bond186 (work_completion)(&(&bond->mcast_work)->work) irq_context: 0 (wq_completion)bond186 (work_completion)(&(&bond->mcast_work)->work) &rq->__lock irq_context: 0 (wq_completion)bond186 (work_completion)(&(&bond->mcast_work)->work) &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)bond186 (work_completion)(&(&bond->mcast_work)->work) &obj_hash[i].lock irq_context: 0 (wq_completion)bond186 (work_completion)(&(&bond->mcast_work)->work) &base->lock irq_context: 0 (wq_completion)bond186 (work_completion)(&(&bond->mcast_work)->work) &base->lock &obj_hash[i].lock irq_context: 0 (wq_completion)bond129#2 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &idev->mc_lock rcu_read_lock &dir->lock#2 irq_context: 0 (wq_completion)bond129#2 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &idev->mc_lock rcu_read_lock &ul->lock#2 irq_context: 0 (wq_completion)bond129#2 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &idev->mc_lock rcu_read_lock rcu_read_lock &obj_hash[i].lock irq_context: 0 (wq_completion)bond129#2 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &idev->mc_lock rcu_read_lock rcu_read_lock &dir->lock#2 irq_context: 0 (wq_completion)bond129#2 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &idev->mc_lock rcu_read_lock rcu_read_lock &tbl->lock irq_context: 0 (wq_completion)bond129#2 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &idev->mc_lock rcu_read_lock rcu_read_lock &n->lock irq_context: 0 (wq_completion)bond129#2 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &idev->mc_lock rcu_read_lock rcu_read_lock &rq->__lock irq_context: 0 (wq_completion)bond129#2 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &idev->mc_lock rcu_read_lock rcu_read_lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)bond129#2 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &idev->mc_lock rcu_read_lock rcu_read_lock &____s->seqcount#9 irq_context: 0 (wq_completion)bond129#2 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &idev->mc_lock rcu_read_lock rcu_read_lock rcu_read_lock_bh rcu_read_lock rcu_read_lock_bh rcu_read_lock &br->multicast_lock irq_context: 0 (wq_completion)bond129#2 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &idev->mc_lock rcu_read_lock rcu_read_lock rcu_read_lock_bh rcu_read_lock rcu_read_lock_bh rcu_read_lock &br->multicast_lock &obj_hash[i].lock irq_context: 0 (wq_completion)bond129#2 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &idev->mc_lock rcu_read_lock rcu_read_lock rcu_read_lock_bh rcu_read_lock rcu_read_lock_bh rcu_read_lock &br->multicast_lock &base->lock irq_context: 0 (wq_completion)bond129#2 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &idev->mc_lock rcu_read_lock rcu_read_lock rcu_read_lock_bh rcu_read_lock rcu_read_lock_bh rcu_read_lock &br->multicast_lock &base->lock &obj_hash[i].lock irq_context: 0 (wq_completion)bond129#2 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &idev->mc_lock rcu_read_lock rcu_read_lock rcu_read_lock_bh rcu_read_lock rcu_read_lock_bh rcu_read_lock &data->lock irq_context: 0 (wq_completion)bond129#2 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &idev->mc_lock rcu_read_lock rcu_read_lock rcu_read_lock_bh rcu_read_lock rcu_read_lock_bh rcu_read_lock &obj_hash[i].lock irq_context: 0 (wq_completion)bond129#2 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex.wait_lock irq_context: 0 (wq_completion)bond129#2 (work_completion)(&(&bond->mcast_work)->work) &p->pi_lock irq_context: 0 (wq_completion)bond129#2 (work_completion)(&(&bond->mcast_work)->work) &p->pi_lock &rq->__lock irq_context: 0 (wq_completion)bond129#2 (work_completion)(&(&bond->mcast_work)->work) &p->pi_lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)bond174#2 irq_context: 0 (wq_completion)bond174#2 (work_completion)(&(&slave->notify_work)->work) irq_context: 0 (wq_completion)bond174#2 (work_completion)(&(&slave->notify_work)->work) &obj_hash[i].lock irq_context: 0 (wq_completion)bond174#2 (work_completion)(&(&slave->notify_work)->work) &base->lock irq_context: 0 (wq_completion)bond174#2 (work_completion)(&(&slave->notify_work)->work) &base->lock &obj_hash[i].lock irq_context: 0 (wq_completion)bond184 (work_completion)(&(&slave->notify_work)->work) &rq->__lock irq_context: 0 (wq_completion)bond184 (work_completion)(&(&slave->notify_work)->work) &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)bond174#2 (work_completion)(&(&bond->mcast_work)->work) irq_context: 0 (wq_completion)bond174#2 (work_completion)(&(&bond->mcast_work)->work) &obj_hash[i].lock irq_context: 0 (wq_completion)bond174#2 (work_completion)(&(&bond->mcast_work)->work) &base->lock irq_context: 0 (wq_completion)bond174#2 (work_completion)(&(&bond->mcast_work)->work) &base->lock &obj_hash[i].lock irq_context: 0 (wq_completion)bond174#2 (work_completion)(&(&bond->mcast_work)->work) &rq->__lock irq_context: 0 (wq_completion)bond174#2 (work_completion)(&(&bond->mcast_work)->work) &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)bond55#2 (work_completion)(&(&slave->notify_work)->work) &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)bond183 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex fs_reclaim irq_context: 0 (wq_completion)bond183 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex fs_reclaim mmu_notifier_invalidate_range_start irq_context: 0 (wq_completion)bond183 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex fs_reclaim mmu_notifier_invalidate_range_start &rq->__lock irq_context: 0 (wq_completion)bond183 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex fs_reclaim mmu_notifier_invalidate_range_start &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)bond183 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &c->lock irq_context: 0 (wq_completion)bond183 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex rcu_read_lock &bond->stats_lock/1 irq_context: 0 (wq_completion)bond183 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex rcu_read_lock &ndev->lock irq_context: 0 (wq_completion)bond183 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex nl_table_lock irq_context: 0 (wq_completion)bond183 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &rq->__lock irq_context: 0 (wq_completion)bond183 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)bond183 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &obj_hash[i].lock irq_context: 0 (wq_completion)bond183 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex nl_table_wait.lock irq_context: 0 (wq_completion)bond183 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex net_rwsem irq_context: 0 (wq_completion)bond183 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex net_rwsem &list->lock#2 irq_context: 0 (wq_completion)bond183 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &tn->lock irq_context: 0 (wq_completion)bond183 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &idev->mc_lock irq_context: 0 (wq_completion)bond183 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &idev->mc_lock fs_reclaim irq_context: 0 (wq_completion)bond183 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &idev->mc_lock fs_reclaim mmu_notifier_invalidate_range_start irq_context: 0 (wq_completion)bond183 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &idev->mc_lock &c->lock irq_context: 0 (wq_completion)bond183 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &idev->mc_lock rcu_read_lock &ndev->lock irq_context: 0 (wq_completion)bond183 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &idev->mc_lock rcu_read_lock &dir->lock#2 irq_context: 0 (wq_completion)bond183 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &idev->mc_lock rcu_read_lock rcu_node_0 irq_context: 0 (wq_completion)bond183 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &idev->mc_lock rcu_read_lock &rq->__lock irq_context: 0 (wq_completion)bond183 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &idev->mc_lock rcu_read_lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)bond183 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &idev->mc_lock rcu_read_lock &ul->lock#2 irq_context: 0 (wq_completion)bond183 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &idev->mc_lock rcu_read_lock rcu_read_lock &____s->seqcount#10 irq_context: 0 (wq_completion)bond183 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &idev->mc_lock rcu_read_lock rcu_read_lock rcu_read_lock_bh rcu_read_lock rcu_read_lock_bh rcu_read_lock &br->multicast_lock irq_context: 0 (wq_completion)bond183 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &idev->mc_lock rcu_read_lock rcu_read_lock rcu_read_lock_bh rcu_read_lock rcu_read_lock_bh rcu_read_lock &br->multicast_lock &obj_hash[i].lock irq_context: 0 (wq_completion)bond183 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &idev->mc_lock rcu_read_lock rcu_read_lock rcu_read_lock_bh rcu_read_lock rcu_read_lock_bh rcu_read_lock &br->multicast_lock &base->lock irq_context: 0 (wq_completion)bond183 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &idev->mc_lock rcu_read_lock rcu_read_lock rcu_read_lock_bh rcu_read_lock rcu_read_lock_bh rcu_read_lock &data->lock irq_context: 0 (wq_completion)bond183 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &idev->mc_lock rcu_read_lock rcu_read_lock rcu_read_lock_bh rcu_read_lock rcu_read_lock_bh rcu_read_lock &obj_hash[i].lock irq_context: 0 (wq_completion)bond183 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex.wait_lock irq_context: 0 (wq_completion)bond183 (work_completion)(&(&bond->mcast_work)->work) &p->pi_lock irq_context: 0 (wq_completion)bond183 (work_completion)(&(&bond->mcast_work)->work) &p->pi_lock &rq->__lock irq_context: 0 (wq_completion)bond183 (work_completion)(&(&bond->mcast_work)->work) &p->pi_lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)bond183 (work_completion)(&(&bond->mcast_work)->work) &rq->__lock irq_context: 0 (wq_completion)bond183 (work_completion)(&(&bond->mcast_work)->work) &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)bond184 (work_completion)(&(&slave->notify_work)->work) rtnl_mutex net_rwsem irq_context: 0 (wq_completion)bond184 (work_completion)(&(&slave->notify_work)->work) rtnl_mutex net_rwsem &list->lock#2 irq_context: 0 (wq_completion)bond184 (work_completion)(&(&slave->notify_work)->work) rtnl_mutex &tn->lock irq_context: 0 (wq_completion)bond184 (work_completion)(&(&slave->notify_work)->work) rtnl_mutex.wait_lock irq_context: 0 (wq_completion)bond184 (work_completion)(&(&slave->notify_work)->work) &p->pi_lock irq_context: 0 (wq_completion)bond174#2 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex fs_reclaim irq_context: 0 (wq_completion)bond174#2 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex fs_reclaim mmu_notifier_invalidate_range_start irq_context: 0 (wq_completion)bond174#2 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &c->lock irq_context: 0 (wq_completion)bond174#2 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &n->list_lock irq_context: 0 (wq_completion)bond174#2 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &n->list_lock &c->lock irq_context: 0 (wq_completion)bond174#2 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &rq->__lock irq_context: 0 (wq_completion)bond174#2 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)bond174#2 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex rcu_read_lock &bond->stats_lock/1 irq_context: 0 (wq_completion)bond174#2 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex rcu_read_lock &ndev->lock irq_context: 0 (wq_completion)bond174#2 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex nl_table_lock irq_context: 0 (wq_completion)bond174#2 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &obj_hash[i].lock irq_context: 0 (wq_completion)bond174#2 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex nl_table_wait.lock irq_context: 0 (wq_completion)bond174#2 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex net_rwsem irq_context: 0 (wq_completion)bond174#2 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex net_rwsem &list->lock#2 irq_context: 0 (wq_completion)bond174#2 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &tn->lock irq_context: 0 (wq_completion)bond174#2 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &idev->mc_lock irq_context: 0 (wq_completion)bond174#2 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex rcu_read_lock rcu_node_0 irq_context: 0 (wq_completion)bond174#2 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex rcu_read_lock &rq->__lock irq_context: 0 (wq_completion)bond174#2 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex rcu_read_lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)bond174#2 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex.wait_lock irq_context: 0 (wq_completion)bond174#2 (work_completion)(&(&bond->mcast_work)->work) &p->pi_lock irq_context: 0 (wq_completion)bond174#2 (work_completion)(&(&bond->mcast_work)->work) &p->pi_lock &rq->__lock irq_context: 0 (wq_completion)bond174#2 (work_completion)(&(&bond->mcast_work)->work) &p->pi_lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)bond136#3 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex fs_reclaim irq_context: 0 (wq_completion)bond136#3 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex fs_reclaim mmu_notifier_invalidate_range_start irq_context: 0 (wq_completion)bond136#3 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &c->lock irq_context: 0 (wq_completion)bond136#3 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex rcu_read_lock &bond->stats_lock/1 irq_context: 0 (wq_completion)bond136#3 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex rcu_read_lock &ndev->lock irq_context: 0 (wq_completion)bond136#3 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex nl_table_lock irq_context: 0 (wq_completion)bond136#3 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &obj_hash[i].lock irq_context: 0 (wq_completion)bond136#3 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex nl_table_wait.lock irq_context: 0 (wq_completion)bond136#3 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex net_rwsem irq_context: 0 (wq_completion)bond136#3 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex net_rwsem &list->lock#2 irq_context: 0 (wq_completion)bond136#3 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &tn->lock irq_context: 0 (wq_completion)bond136#3 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &idev->mc_lock irq_context: 0 (wq_completion)bond136#3 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &idev->mc_lock fs_reclaim irq_context: 0 (wq_completion)bond136#3 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &idev->mc_lock fs_reclaim mmu_notifier_invalidate_range_start irq_context: 0 (wq_completion)bond136#3 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &idev->mc_lock rcu_read_lock &ndev->lock irq_context: 0 (wq_completion)bond136#3 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &idev->mc_lock rcu_read_lock &dir->lock#2 irq_context: 0 (wq_completion)bond136#3 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &idev->mc_lock rcu_read_lock &c->lock irq_context: 0 (wq_completion)bond136#3 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &idev->mc_lock rcu_read_lock &ul->lock#2 irq_context: 0 (wq_completion)bond136#3 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &idev->mc_lock rcu_read_lock rcu_read_lock &____s->seqcount#10 irq_context: 0 (wq_completion)bond136#3 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &idev->mc_lock rcu_read_lock rcu_read_lock rcu_read_lock_bh rcu_read_lock rcu_read_lock_bh rcu_read_lock &br->multicast_lock irq_context: 0 (wq_completion)bond136#3 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &idev->mc_lock rcu_read_lock rcu_read_lock rcu_read_lock_bh rcu_read_lock rcu_read_lock_bh rcu_read_lock &br->multicast_lock &obj_hash[i].lock irq_context: 0 (wq_completion)bond136#3 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &idev->mc_lock rcu_read_lock rcu_read_lock rcu_read_lock_bh rcu_read_lock rcu_read_lock_bh rcu_read_lock &br->multicast_lock &base->lock irq_context: 0 (wq_completion)bond136#3 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &idev->mc_lock rcu_read_lock rcu_read_lock rcu_read_lock_bh rcu_read_lock rcu_read_lock_bh rcu_read_lock &br->multicast_lock &base->lock &obj_hash[i].lock irq_context: 0 (wq_completion)bond136#3 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &idev->mc_lock rcu_read_lock rcu_read_lock rcu_read_lock_bh rcu_read_lock rcu_read_lock_bh rcu_read_lock &data->lock irq_context: 0 (wq_completion)bond136#3 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &idev->mc_lock rcu_read_lock rcu_read_lock rcu_read_lock_bh rcu_read_lock rcu_read_lock_bh rcu_read_lock &obj_hash[i].lock irq_context: 0 (wq_completion)bond136#3 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex.wait_lock irq_context: 0 (wq_completion)bond141#2 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex fs_reclaim irq_context: 0 (wq_completion)bond136#3 (work_completion)(&(&bond->mcast_work)->work) &p->pi_lock irq_context: 0 (wq_completion)bond141#2 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex fs_reclaim mmu_notifier_invalidate_range_start irq_context: 0 (wq_completion)bond141#2 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &c->lock irq_context: 0 (wq_completion)bond141#2 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &n->list_lock irq_context: 0 (wq_completion)bond141#2 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &n->list_lock &c->lock irq_context: 0 (wq_completion)bond141#2 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &rq->__lock irq_context: 0 (wq_completion)bond141#2 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)bond141#2 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex rcu_read_lock &bond->stats_lock/1 irq_context: 0 (wq_completion)bond141#2 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex rcu_read_lock &ndev->lock irq_context: 0 (wq_completion)bond141#2 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex nl_table_lock irq_context: 0 (wq_completion)bond141#2 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &obj_hash[i].lock irq_context: 0 (wq_completion)bond141#2 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex nl_table_wait.lock irq_context: 0 (wq_completion)bond141#2 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex net_rwsem irq_context: 0 (wq_completion)bond141#2 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex net_rwsem &list->lock#2 irq_context: 0 (wq_completion)bond141#2 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &tn->lock irq_context: 0 (wq_completion)bond141#2 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &idev->mc_lock irq_context: 0 (wq_completion)bond141#2 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &idev->mc_lock fs_reclaim irq_context: 0 (wq_completion)bond141#2 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &idev->mc_lock fs_reclaim &rq->__lock irq_context: 0 (wq_completion)bond141#2 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &idev->mc_lock fs_reclaim &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)bond141#2 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &idev->mc_lock fs_reclaim mmu_notifier_invalidate_range_start irq_context: 0 (wq_completion)bond141#2 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &idev->mc_lock &c->lock irq_context: 0 (wq_completion)bond141#2 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &idev->mc_lock rcu_read_lock &ndev->lock irq_context: 0 (wq_completion)bond141#2 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &idev->mc_lock rcu_read_lock &dir->lock#2 irq_context: 0 (wq_completion)bond141#2 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &idev->mc_lock rcu_read_lock &ul->lock#2 irq_context: 0 (wq_completion)bond141#2 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &idev->mc_lock rcu_read_lock rcu_read_lock &____s->seqcount#10 irq_context: 0 (wq_completion)bond141#2 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &idev->mc_lock rcu_read_lock rcu_read_lock rcu_read_lock_bh rcu_read_lock rcu_read_lock_bh rcu_read_lock &br->multicast_lock irq_context: 0 (wq_completion)bond141#2 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &idev->mc_lock rcu_read_lock rcu_read_lock rcu_read_lock_bh rcu_read_lock rcu_read_lock_bh rcu_read_lock &br->multicast_lock &obj_hash[i].lock irq_context: 0 (wq_completion)bond141#2 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &idev->mc_lock rcu_read_lock rcu_read_lock rcu_read_lock_bh rcu_read_lock rcu_read_lock_bh rcu_read_lock &br->multicast_lock &base->lock irq_context: 0 (wq_completion)bond141#2 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &idev->mc_lock rcu_read_lock rcu_read_lock rcu_read_lock_bh rcu_read_lock rcu_read_lock_bh rcu_read_lock &br->multicast_lock &base->lock &obj_hash[i].lock irq_context: 0 (wq_completion)bond141#2 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &idev->mc_lock rcu_read_lock rcu_read_lock rcu_read_lock_bh rcu_read_lock rcu_read_lock_bh rcu_read_lock &data->lock irq_context: 0 (wq_completion)bond141#2 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &idev->mc_lock rcu_read_lock rcu_read_lock rcu_read_lock_bh rcu_read_lock rcu_read_lock_bh rcu_read_lock &obj_hash[i].lock irq_context: 0 (wq_completion)bond141#2 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &idev->mc_lock rcu_read_lock rcu_read_lock rcu_node_0 irq_context: 0 (wq_completion)bond141#2 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &idev->mc_lock rcu_read_lock rcu_read_lock &rq->__lock irq_context: 0 (wq_completion)bond141#2 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &idev->mc_lock rcu_read_lock rcu_read_lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)bond141#2 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &idev->mc_lock rcu_read_lock rcu_node_0 irq_context: 0 (wq_completion)bond141#2 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex.wait_lock irq_context: 0 (wq_completion)bond141#2 (work_completion)(&(&bond->mcast_work)->work) &p->pi_lock irq_context: 0 (wq_completion)bond151#2 (work_completion)(&(&slave->notify_work)->work) rtnl_mutex net_rwsem irq_context: 0 (wq_completion)bond151#2 (work_completion)(&(&slave->notify_work)->work) rtnl_mutex net_rwsem &list->lock#2 irq_context: 0 (wq_completion)bond151#2 (work_completion)(&(&slave->notify_work)->work) rtnl_mutex &tn->lock irq_context: 0 (wq_completion)bond151#2 (work_completion)(&(&slave->notify_work)->work) rtnl_mutex.wait_lock irq_context: 0 (wq_completion)bond54#3 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex fs_reclaim irq_context: 0 (wq_completion)bond151#2 (work_completion)(&(&slave->notify_work)->work) &rq->__lock irq_context: 0 (wq_completion)bond54#3 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex fs_reclaim mmu_notifier_invalidate_range_start irq_context: 0 (wq_completion)bond151#2 (work_completion)(&(&slave->notify_work)->work) &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)bond54#3 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex rcu_read_lock &bond->stats_lock/1 irq_context: 0 (wq_completion)bond54#3 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex rcu_read_lock &ndev->lock irq_context: 0 (wq_completion)bond54#3 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex nl_table_lock irq_context: 0 (wq_completion)bond54#3 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &obj_hash[i].lock irq_context: 0 (wq_completion)bond54#3 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex nl_table_wait.lock irq_context: 0 (wq_completion)bond54#3 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex net_rwsem irq_context: 0 (wq_completion)bond54#3 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex net_rwsem &list->lock#2 irq_context: 0 (wq_completion)bond151#2 (work_completion)(&(&slave->notify_work)->work) &p->pi_lock irq_context: 0 (wq_completion)bond54#3 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &tn->lock irq_context: 0 (wq_completion)bond54#3 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &idev->mc_lock irq_context: 0 (wq_completion)bond54#3 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &idev->mc_lock fs_reclaim irq_context: 0 (wq_completion)bond54#3 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &idev->mc_lock fs_reclaim mmu_notifier_invalidate_range_start irq_context: 0 (wq_completion)bond54#3 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &idev->mc_lock fs_reclaim mmu_notifier_invalidate_range_start &rq->__lock irq_context: 0 (wq_completion)bond54#3 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &idev->mc_lock fs_reclaim mmu_notifier_invalidate_range_start &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)bond54#3 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &idev->mc_lock rcu_read_lock &ndev->lock irq_context: 0 (wq_completion)bond54#3 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &idev->mc_lock rcu_read_lock &c->lock irq_context: 0 (wq_completion)bond54#3 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &idev->mc_lock rcu_read_lock &dir->lock#2 irq_context: 0 (wq_completion)bond54#3 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &idev->mc_lock rcu_read_lock &ul->lock#2 irq_context: 0 (wq_completion)bond54#3 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &idev->mc_lock rcu_read_lock rcu_read_lock &____s->seqcount#12 irq_context: 0 (wq_completion)bond54#3 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &idev->mc_lock rcu_read_lock rcu_read_lock &____s->seqcount#16 irq_context: 0 (wq_completion)bond54#3 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &idev->mc_lock rcu_read_lock &obj_hash[i].lock irq_context: 0 (wq_completion)bond54#3 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &idev->mc_lock rcu_read_lock &data->lock irq_context: 0 (wq_completion)bond54#3 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex.wait_lock irq_context: 0 (wq_completion)bond54#3 (work_completion)(&(&bond->mcast_work)->work) &p->pi_lock irq_context: 0 (wq_completion)bond158#3 irq_context: 0 (wq_completion)bond158#3 (work_completion)(&(&slave->notify_work)->work) irq_context: 0 (wq_completion)bond158#3 (work_completion)(&(&slave->notify_work)->work) &obj_hash[i].lock irq_context: 0 (wq_completion)bond158#3 (work_completion)(&(&slave->notify_work)->work) &base->lock irq_context: 0 (wq_completion)bond158#3 (work_completion)(&(&slave->notify_work)->work) &base->lock &obj_hash[i].lock irq_context: 0 (wq_completion)bond158#3 (work_completion)(&(&slave->notify_work)->work) &rq->__lock irq_context: 0 (wq_completion)bond158#3 (work_completion)(&(&slave->notify_work)->work) &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)bond142#3 (work_completion)(&(&slave->notify_work)->work) &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)bond158#3 (work_completion)(&(&bond->mcast_work)->work) irq_context: 0 (wq_completion)bond158#3 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex fs_reclaim irq_context: 0 (wq_completion)bond158#3 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex fs_reclaim &rq->__lock irq_context: 0 (wq_completion)bond158#3 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex fs_reclaim &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)bond158#3 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex fs_reclaim mmu_notifier_invalidate_range_start irq_context: 0 (wq_completion)bond158#3 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &c->lock irq_context: 0 (wq_completion)bond158#3 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex rcu_read_lock &bond->stats_lock/1 irq_context: 0 (wq_completion)bond158#3 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex rcu_read_lock &ndev->lock irq_context: 0 (wq_completion)bond158#3 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex nl_table_lock irq_context: 0 (wq_completion)bond158#3 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &obj_hash[i].lock irq_context: 0 (wq_completion)bond158#3 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex nl_table_wait.lock irq_context: 0 (wq_completion)bond158#3 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex net_rwsem irq_context: 0 (wq_completion)bond158#3 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex net_rwsem &rq->__lock irq_context: 0 (wq_completion)bond158#3 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex net_rwsem &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)bond158#3 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex net_rwsem &list->lock#2 irq_context: 0 (wq_completion)bond158#3 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &tn->lock irq_context: 0 (wq_completion)bond158#3 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &idev->mc_lock irq_context: 0 (wq_completion)bond158#3 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex.wait_lock irq_context: 0 (wq_completion)bond158#3 (work_completion)(&(&bond->mcast_work)->work) &p->pi_lock irq_context: 0 (wq_completion)bond158#3 (work_completion)(&(&bond->mcast_work)->work) &p->pi_lock &rq->__lock irq_context: 0 (wq_completion)bond158#3 (work_completion)(&(&bond->mcast_work)->work) &p->pi_lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)bond158#3 (work_completion)(&(&bond->mcast_work)->work) &rq->__lock irq_context: 0 (wq_completion)bond158#3 (work_completion)(&(&bond->mcast_work)->work) &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)bond187 irq_context: 0 (wq_completion)bond187 (work_completion)(&(&slave->notify_work)->work) irq_context: 0 (wq_completion)bond187 (work_completion)(&(&slave->notify_work)->work) &rq->__lock irq_context: 0 (wq_completion)bond187 (work_completion)(&(&slave->notify_work)->work) &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)bond187 (work_completion)(&(&slave->notify_work)->work) &obj_hash[i].lock irq_context: 0 (wq_completion)bond187 (work_completion)(&(&slave->notify_work)->work) &base->lock irq_context: 0 (wq_completion)bond187 (work_completion)(&(&slave->notify_work)->work) &base->lock &obj_hash[i].lock irq_context: 0 (wq_completion)bond187 (work_completion)(&(&bond->mcast_work)->work) irq_context: 0 (wq_completion)bond187 (work_completion)(&(&bond->mcast_work)->work) &obj_hash[i].lock irq_context: 0 (wq_completion)bond187 (work_completion)(&(&bond->mcast_work)->work) &base->lock irq_context: 0 (wq_completion)bond187 (work_completion)(&(&bond->mcast_work)->work) &base->lock &obj_hash[i].lock irq_context: 0 (wq_completion)bond187 (work_completion)(&(&bond->mcast_work)->work) &rq->__lock irq_context: 0 (wq_completion)bond187 (work_completion)(&(&bond->mcast_work)->work) &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)events_unbound (work_completion)(&(&kfence_timer)->work) cpu_hotplug_lock jump_label_mutex &cfs_rq->removed.lock irq_context: 0 (wq_completion)events_unbound (work_completion)(&(&kfence_timer)->work) cpu_hotplug_lock jump_label_mutex &obj_hash[i].lock irq_context: 0 (wq_completion)events_unbound (work_completion)(&(&kfence_timer)->work) cpu_hotplug_lock jump_label_mutex pool_lock#2 irq_context: 0 (wq_completion)bond153 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex fs_reclaim irq_context: 0 (wq_completion)bond153 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex fs_reclaim mmu_notifier_invalidate_range_start irq_context: 0 (wq_completion)bond153 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &c->lock irq_context: 0 (wq_completion)bond153 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex rcu_read_lock &bond->stats_lock/1 irq_context: 0 (wq_completion)bond153 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex rcu_read_lock &ndev->lock irq_context: 0 (wq_completion)bond153 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex nl_table_lock irq_context: 0 (wq_completion)bond153 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &obj_hash[i].lock irq_context: 0 (wq_completion)bond153 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex nl_table_wait.lock irq_context: 0 (wq_completion)bond153 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex net_rwsem irq_context: 0 (wq_completion)bond153 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex net_rwsem &list->lock#2 irq_context: 0 (wq_completion)bond153 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &tn->lock irq_context: 0 (wq_completion)bond153 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &idev->mc_lock irq_context: 0 (wq_completion)bond153 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &idev->mc_lock fs_reclaim irq_context: 0 (wq_completion)bond153 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &idev->mc_lock fs_reclaim mmu_notifier_invalidate_range_start irq_context: 0 (wq_completion)bond153 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &idev->mc_lock &c->lock irq_context: 0 (wq_completion)bond153 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &idev->mc_lock rcu_read_lock &ndev->lock irq_context: 0 (wq_completion)bond153 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &idev->mc_lock rcu_read_lock &dir->lock#2 irq_context: 0 (wq_completion)bond153 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &idev->mc_lock rcu_read_lock &ul->lock#2 irq_context: 0 (wq_completion)bond153 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &idev->mc_lock rcu_read_lock rcu_read_lock &____s->seqcount#10 irq_context: 0 (wq_completion)bond153 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &idev->mc_lock rcu_read_lock rcu_read_lock rcu_read_lock_bh rcu_read_lock rcu_read_lock_bh rcu_read_lock &br->multicast_lock irq_context: 0 (wq_completion)bond153 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &idev->mc_lock rcu_read_lock rcu_read_lock rcu_read_lock_bh rcu_read_lock rcu_read_lock_bh rcu_read_lock &br->multicast_lock &obj_hash[i].lock irq_context: 0 (wq_completion)bond153 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &idev->mc_lock rcu_read_lock rcu_read_lock rcu_read_lock_bh rcu_read_lock rcu_read_lock_bh rcu_read_lock &br->multicast_lock &base->lock irq_context: 0 (wq_completion)bond153 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &idev->mc_lock rcu_read_lock rcu_read_lock rcu_read_lock_bh rcu_read_lock rcu_read_lock_bh rcu_read_lock &br->multicast_lock &base->lock &obj_hash[i].lock irq_context: 0 (wq_completion)bond153 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &idev->mc_lock rcu_read_lock rcu_read_lock rcu_read_lock_bh rcu_read_lock rcu_read_lock_bh rcu_read_lock &data->lock irq_context: 0 (wq_completion)bond153 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &idev->mc_lock rcu_read_lock rcu_read_lock rcu_read_lock_bh rcu_read_lock rcu_read_lock_bh rcu_read_lock &obj_hash[i].lock irq_context: 0 (wq_completion)bond153 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex.wait_lock irq_context: 0 (wq_completion)bond153 (work_completion)(&(&bond->mcast_work)->work) &p->pi_lock irq_context: softirq &(&slave->notify_work)->timer rcu_read_lock &pool->lock/1 fill_pool_map-wait-type-override &____s->seqcount#2 irq_context: 0 (wq_completion)bond159#3 irq_context: 0 (wq_completion)bond159#3 (work_completion)(&(&slave->notify_work)->work) irq_context: 0 (wq_completion)bond159#3 (work_completion)(&(&slave->notify_work)->work) &obj_hash[i].lock irq_context: 0 (wq_completion)bond159#3 (work_completion)(&(&slave->notify_work)->work) &base->lock irq_context: 0 (wq_completion)bond159#3 (work_completion)(&(&slave->notify_work)->work) &base->lock &obj_hash[i].lock irq_context: 0 (wq_completion)bond159#3 (work_completion)(&(&slave->notify_work)->work) &rq->__lock irq_context: 0 (wq_completion)bond159#3 (work_completion)(&(&slave->notify_work)->work) &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)bond159#3 (work_completion)(&(&bond->mcast_work)->work) irq_context: 0 (wq_completion)bond159#3 (work_completion)(&(&bond->mcast_work)->work) &obj_hash[i].lock irq_context: 0 (wq_completion)bond159#3 (work_completion)(&(&bond->mcast_work)->work) &base->lock irq_context: 0 (wq_completion)bond159#3 (work_completion)(&(&bond->mcast_work)->work) &base->lock &obj_hash[i].lock irq_context: 0 (wq_completion)bond159#3 (work_completion)(&(&bond->mcast_work)->work) &rq->__lock irq_context: 0 (wq_completion)bond159#3 (work_completion)(&(&bond->mcast_work)->work) &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)bond50#2 (work_completion)(&(&bond->mcast_work)->work) &rq->__lock irq_context: 0 (wq_completion)bond50#2 (work_completion)(&(&bond->mcast_work)->work) &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)bond150#3 (work_completion)(&(&bond->mcast_work)->work) &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)bond96#3 (work_completion)(&(&slave->notify_work)->work) fill_pool_map-wait-type-override &c->lock irq_context: 0 (wq_completion)bond96#3 (work_completion)(&(&slave->notify_work)->work) fill_pool_map-wait-type-override &____s->seqcount#2 irq_context: 0 (wq_completion)bond96#3 (work_completion)(&(&slave->notify_work)->work) fill_pool_map-wait-type-override &____s->seqcount irq_context: 0 (wq_completion)bond96#3 (work_completion)(&(&slave->notify_work)->work) fill_pool_map-wait-type-override pool_lock irq_context: 0 (wq_completion)bond188 irq_context: 0 (wq_completion)bond188 (work_completion)(&(&slave->notify_work)->work) irq_context: 0 (wq_completion)bond188 (work_completion)(&(&slave->notify_work)->work) &obj_hash[i].lock irq_context: 0 (wq_completion)bond188 (work_completion)(&(&slave->notify_work)->work) &base->lock irq_context: 0 (wq_completion)bond188 (work_completion)(&(&slave->notify_work)->work) &base->lock &obj_hash[i].lock irq_context: 0 (wq_completion)bond188 (work_completion)(&(&slave->notify_work)->work) &rq->__lock irq_context: 0 (wq_completion)bond188 (work_completion)(&(&bond->mcast_work)->work) irq_context: 0 (wq_completion)bond188 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex fs_reclaim irq_context: 0 (wq_completion)bond188 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex fs_reclaim mmu_notifier_invalidate_range_start irq_context: 0 (wq_completion)bond188 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &c->lock irq_context: 0 (wq_completion)bond188 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &n->list_lock irq_context: 0 (wq_completion)bond188 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &n->list_lock &c->lock irq_context: 0 (wq_completion)bond188 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &rq->__lock irq_context: 0 (wq_completion)bond188 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex rcu_read_lock &bond->stats_lock/1 irq_context: 0 (wq_completion)bond188 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex rcu_read_lock &ndev->lock irq_context: 0 (wq_completion)bond188 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex nl_table_lock irq_context: 0 (wq_completion)bond188 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &obj_hash[i].lock irq_context: 0 (wq_completion)bond188 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex nl_table_wait.lock irq_context: 0 (wq_completion)bond188 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex net_rwsem irq_context: 0 (wq_completion)bond188 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex net_rwsem &list->lock#2 irq_context: 0 (wq_completion)bond188 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &tn->lock irq_context: 0 (wq_completion)bond188 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &idev->mc_lock irq_context: 0 (wq_completion)bond188 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)bond188 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex.wait_lock irq_context: 0 (wq_completion)bond188 (work_completion)(&(&bond->mcast_work)->work) &p->pi_lock irq_context: 0 (wq_completion)bond188 (work_completion)(&(&bond->mcast_work)->work) &p->pi_lock &rq->__lock irq_context: 0 (wq_completion)bond188 (work_completion)(&(&bond->mcast_work)->work) &p->pi_lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)bond188 (work_completion)(&(&bond->mcast_work)->work) &rq->__lock irq_context: 0 (wq_completion)bond188 (work_completion)(&(&bond->mcast_work)->work) &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)bond159#3 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex fs_reclaim irq_context: 0 (wq_completion)bond159#3 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex fs_reclaim mmu_notifier_invalidate_range_start irq_context: 0 (wq_completion)bond159#3 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &c->lock irq_context: 0 (wq_completion)bond159#3 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex rcu_read_lock &bond->stats_lock/1 irq_context: 0 (wq_completion)bond159#3 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex rcu_read_lock &ndev->lock irq_context: 0 (wq_completion)bond159#3 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex nl_table_lock irq_context: 0 (wq_completion)bond159#3 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &obj_hash[i].lock irq_context: 0 (wq_completion)bond159#3 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex nl_table_wait.lock irq_context: 0 (wq_completion)bond159#3 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex net_rwsem irq_context: 0 (wq_completion)bond159#3 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex net_rwsem &list->lock#2 irq_context: 0 (wq_completion)bond159#3 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &tn->lock irq_context: 0 (wq_completion)bond159#3 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &idev->mc_lock irq_context: 0 (wq_completion)bond159#3 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex.wait_lock irq_context: 0 (wq_completion)bond159#3 (work_completion)(&(&bond->mcast_work)->work) &p->pi_lock irq_context: 0 (wq_completion)bond161#2 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex fs_reclaim irq_context: 0 (wq_completion)bond161#2 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex fs_reclaim mmu_notifier_invalidate_range_start irq_context: 0 (wq_completion)bond161#2 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex rcu_read_lock &bond->stats_lock/1 irq_context: 0 (wq_completion)bond161#2 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex rcu_read_lock &ndev->lock irq_context: 0 (wq_completion)bond161#2 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex nl_table_lock irq_context: 0 (wq_completion)bond161#2 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &obj_hash[i].lock irq_context: 0 (wq_completion)bond161#2 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex nl_table_wait.lock irq_context: 0 (wq_completion)bond161#2 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex net_rwsem irq_context: 0 (wq_completion)bond161#2 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex net_rwsem &list->lock#2 irq_context: 0 (wq_completion)bond161#2 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &tn->lock irq_context: 0 (wq_completion)bond161#2 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &idev->mc_lock irq_context: 0 (wq_completion)bond161#2 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &idev->mc_lock fs_reclaim irq_context: 0 (wq_completion)bond161#2 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &idev->mc_lock fs_reclaim mmu_notifier_invalidate_range_start irq_context: 0 (wq_completion)bond161#2 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &idev->mc_lock &c->lock irq_context: 0 (wq_completion)bond161#2 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &idev->mc_lock rcu_read_lock &ndev->lock irq_context: 0 (wq_completion)bond161#2 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &idev->mc_lock rcu_read_lock &dir->lock#2 irq_context: 0 (wq_completion)bond161#2 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &idev->mc_lock rcu_read_lock &ul->lock#2 irq_context: 0 (wq_completion)bond161#2 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &idev->mc_lock rcu_read_lock rcu_read_lock &____s->seqcount#10 irq_context: 0 (wq_completion)bond161#2 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &idev->mc_lock rcu_read_lock rcu_read_lock rcu_read_lock_bh rcu_read_lock rcu_read_lock_bh rcu_read_lock &br->multicast_lock irq_context: 0 (wq_completion)bond161#2 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &idev->mc_lock rcu_read_lock rcu_read_lock rcu_read_lock_bh rcu_read_lock rcu_read_lock_bh rcu_read_lock &br->multicast_lock &obj_hash[i].lock irq_context: 0 (wq_completion)bond161#2 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &idev->mc_lock rcu_read_lock rcu_read_lock rcu_read_lock_bh rcu_read_lock rcu_read_lock_bh rcu_read_lock &br->multicast_lock &base->lock irq_context: 0 (wq_completion)bond161#2 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &idev->mc_lock rcu_read_lock rcu_read_lock rcu_read_lock_bh rcu_read_lock rcu_read_lock_bh rcu_read_lock &br->multicast_lock &base->lock &obj_hash[i].lock irq_context: 0 (wq_completion)bond161#2 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &idev->mc_lock rcu_read_lock rcu_read_lock rcu_read_lock_bh rcu_read_lock rcu_read_lock_bh rcu_read_lock &data->lock irq_context: 0 (wq_completion)bond161#2 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &idev->mc_lock rcu_read_lock rcu_read_lock rcu_read_lock_bh rcu_read_lock rcu_read_lock_bh rcu_read_lock &obj_hash[i].lock irq_context: 0 (wq_completion)bond161#2 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex.wait_lock irq_context: 0 (wq_completion)bond161#2 (work_completion)(&(&bond->mcast_work)->work) &p->pi_lock irq_context: 0 (wq_completion)bond145 (work_completion)(&(&slave->notify_work)->work) rtnl_mutex net_rwsem irq_context: 0 (wq_completion)bond145 (work_completion)(&(&slave->notify_work)->work) rtnl_mutex net_rwsem &list->lock#2 irq_context: 0 (wq_completion)bond145 (work_completion)(&(&slave->notify_work)->work) rtnl_mutex &tn->lock irq_context: 0 (wq_completion)bond145 (work_completion)(&(&slave->notify_work)->work) rtnl_mutex.wait_lock irq_context: 0 (wq_completion)bond145 (work_completion)(&(&slave->notify_work)->work) &p->pi_lock irq_context: 0 (wq_completion)bond135#3 (work_completion)(&(&slave->notify_work)->work) rtnl_mutex net_rwsem irq_context: 0 (wq_completion)bond135#3 (work_completion)(&(&slave->notify_work)->work) rtnl_mutex net_rwsem &list->lock#2 irq_context: 0 (wq_completion)bond135#3 (work_completion)(&(&slave->notify_work)->work) rtnl_mutex &tn->lock irq_context: 0 (wq_completion)bond135#3 (work_completion)(&(&slave->notify_work)->work) rtnl_mutex.wait_lock irq_context: 0 (wq_completion)bond135#3 (work_completion)(&(&slave->notify_work)->work) &p->pi_lock irq_context: 0 (wq_completion)bond40 (work_completion)(&(&slave->notify_work)->work) rtnl_mutex net_rwsem irq_context: 0 (wq_completion)bond40 (work_completion)(&(&slave->notify_work)->work) rtnl_mutex net_rwsem &list->lock#2 irq_context: 0 (wq_completion)bond40 (work_completion)(&(&slave->notify_work)->work) rtnl_mutex &tn->lock irq_context: 0 (wq_completion)bond40 (work_completion)(&(&slave->notify_work)->work) rtnl_mutex.wait_lock irq_context: 0 (wq_completion)bond40 (work_completion)(&(&slave->notify_work)->work) &p->pi_lock irq_context: 0 (wq_completion)bond170#2 (work_completion)(&(&slave->notify_work)->work) rtnl_mutex net_rwsem irq_context: 0 (wq_completion)bond170#2 (work_completion)(&(&slave->notify_work)->work) rtnl_mutex net_rwsem &list->lock#2 irq_context: 0 (wq_completion)bond170#2 (work_completion)(&(&slave->notify_work)->work) rtnl_mutex &tn->lock irq_context: 0 (wq_completion)bond170#2 (work_completion)(&(&slave->notify_work)->work) rtnl_mutex.wait_lock irq_context: 0 (wq_completion)bond170#2 (work_completion)(&(&slave->notify_work)->work) &p->pi_lock irq_context: 0 (wq_completion)bond55#2 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex fs_reclaim irq_context: 0 (wq_completion)bond55#2 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex fs_reclaim mmu_notifier_invalidate_range_start irq_context: 0 (wq_completion)bond55#2 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &c->lock irq_context: 0 (wq_completion)bond55#2 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex rcu_read_lock &bond->stats_lock/1 irq_context: 0 (wq_completion)bond55#2 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex rcu_read_lock &ndev->lock irq_context: 0 (wq_completion)bond55#2 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex nl_table_lock irq_context: 0 (wq_completion)bond55#2 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &obj_hash[i].lock irq_context: 0 (wq_completion)bond55#2 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex nl_table_wait.lock irq_context: 0 (wq_completion)bond55#2 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex net_rwsem irq_context: 0 (wq_completion)bond55#2 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex net_rwsem &list->lock#2 irq_context: 0 (wq_completion)bond55#2 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &tn->lock irq_context: 0 (wq_completion)bond55#2 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &idev->mc_lock irq_context: 0 (wq_completion)bond55#2 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &idev->mc_lock fs_reclaim irq_context: 0 (wq_completion)bond55#2 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &idev->mc_lock fs_reclaim mmu_notifier_invalidate_range_start irq_context: 0 (wq_completion)bond55#2 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &idev->mc_lock &c->lock irq_context: 0 (wq_completion)bond55#2 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &idev->mc_lock &rq->__lock irq_context: 0 (wq_completion)bond55#2 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &idev->mc_lock rcu_read_lock &ndev->lock irq_context: 0 (wq_completion)bond55#2 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &idev->mc_lock rcu_read_lock &dir->lock#2 irq_context: 0 (wq_completion)bond55#2 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &idev->mc_lock rcu_read_lock &ul->lock#2 irq_context: 0 (wq_completion)bond55#2 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &idev->mc_lock rcu_read_lock rcu_read_lock &____s->seqcount#12 irq_context: 0 (wq_completion)bond55#2 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &idev->mc_lock rcu_read_lock rcu_read_lock &____s->seqcount#16 irq_context: 0 (wq_completion)bond55#2 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &idev->mc_lock rcu_read_lock &obj_hash[i].lock irq_context: 0 (wq_completion)bond55#2 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &idev->mc_lock rcu_read_lock &data->lock irq_context: 0 (wq_completion)bond55#2 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex.wait_lock irq_context: 0 (wq_completion)bond55#2 (work_completion)(&(&bond->mcast_work)->work) &p->pi_lock irq_context: 0 (wq_completion)bond55#2 (work_completion)(&(&bond->mcast_work)->work) &p->pi_lock &rq->__lock irq_context: 0 (wq_completion)bond55#2 (work_completion)(&(&bond->mcast_work)->work) &p->pi_lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)bond170 (work_completion)(&(&slave->notify_work)->work) fill_pool_map-wait-type-override pool_lock irq_context: 0 (wq_completion)bond176#2 irq_context: 0 (wq_completion)bond176#2 (work_completion)(&(&slave->notify_work)->work) irq_context: 0 (wq_completion)bond176#2 (work_completion)(&(&slave->notify_work)->work) &obj_hash[i].lock irq_context: 0 (wq_completion)bond176#2 (work_completion)(&(&slave->notify_work)->work) &base->lock irq_context: 0 (wq_completion)bond176#2 (work_completion)(&(&slave->notify_work)->work) &base->lock &obj_hash[i].lock irq_context: 0 (wq_completion)bond155#3 (work_completion)(&(&bond->mcast_work)->work) &rq->__lock irq_context: 0 (wq_completion)bond155#3 (work_completion)(&(&bond->mcast_work)->work) &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)bond158#3 (work_completion)(&(&slave->notify_work)->work) &base->lock fill_pool_map-wait-type-override &c->lock irq_context: 0 (wq_completion)bond158#3 (work_completion)(&(&slave->notify_work)->work) &base->lock fill_pool_map-wait-type-override &____s->seqcount#2 irq_context: 0 (wq_completion)bond158#3 (work_completion)(&(&slave->notify_work)->work) &base->lock fill_pool_map-wait-type-override &____s->seqcount irq_context: 0 (wq_completion)bond158#3 (work_completion)(&(&slave->notify_work)->work) &base->lock fill_pool_map-wait-type-override pool_lock#2 irq_context: 0 (wq_completion)bond158#3 (work_completion)(&(&slave->notify_work)->work) &base->lock fill_pool_map-wait-type-override pool_lock irq_context: 0 (wq_completion)bond176#2 (work_completion)(&(&slave->notify_work)->work) fill_pool_map-wait-type-override pool_lock irq_context: 0 (wq_completion)bond176#2 (work_completion)(&(&slave->notify_work)->work) &rq->__lock irq_context: 0 (wq_completion)bond176#2 (work_completion)(&(&slave->notify_work)->work) &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)mld (work_completion)(&(&idev->mc_ifc_work)->work) &idev->mc_lock rcu_read_lock rcu_read_lock &rq->__lock &cfs_rq->removed.lock irq_context: 0 (wq_completion)bond115 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex fs_reclaim irq_context: 0 (wq_completion)bond115 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex fs_reclaim mmu_notifier_invalidate_range_start irq_context: 0 (wq_completion)bond115 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex rcu_read_lock &bond->stats_lock/1 irq_context: 0 (wq_completion)bond115 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex rcu_read_lock rcu_node_0 irq_context: 0 (wq_completion)bond115 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex rcu_read_lock &rq->__lock irq_context: 0 (wq_completion)bond115 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex rcu_read_lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)bond115 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex rcu_read_lock &ndev->lock irq_context: 0 (wq_completion)bond115 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex nl_table_lock irq_context: 0 (wq_completion)bond115 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &obj_hash[i].lock irq_context: 0 (wq_completion)bond115 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex nl_table_wait.lock irq_context: 0 (wq_completion)bond115 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex net_rwsem irq_context: 0 (wq_completion)bond115 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex net_rwsem &list->lock#2 irq_context: 0 (wq_completion)bond115 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &tn->lock irq_context: 0 (wq_completion)bond115 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &idev->mc_lock irq_context: 0 (wq_completion)bond115 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &idev->mc_lock fs_reclaim irq_context: 0 (wq_completion)bond115 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &idev->mc_lock fs_reclaim mmu_notifier_invalidate_range_start irq_context: 0 (wq_completion)bond115 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &idev->mc_lock fs_reclaim mmu_notifier_invalidate_range_start &rq->__lock irq_context: 0 (wq_completion)bond115 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &idev->mc_lock fs_reclaim mmu_notifier_invalidate_range_start &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)bond115 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &idev->mc_lock &c->lock irq_context: 0 (wq_completion)bond115 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &idev->mc_lock &n->list_lock irq_context: 0 (wq_completion)bond115 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &idev->mc_lock &n->list_lock &c->lock irq_context: 0 (wq_completion)bond115 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &idev->mc_lock rcu_read_lock &ndev->lock irq_context: 0 (wq_completion)bond115 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &idev->mc_lock rcu_read_lock &c->lock irq_context: 0 (wq_completion)bond115 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &idev->mc_lock rcu_read_lock &dir->lock#2 irq_context: 0 (wq_completion)bond115 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &idev->mc_lock rcu_read_lock &ul->lock#2 irq_context: 0 (wq_completion)bond115 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &idev->mc_lock rcu_read_lock rcu_read_lock &obj_hash[i].lock irq_context: 0 (wq_completion)bond115 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &idev->mc_lock rcu_read_lock rcu_read_lock &obj_hash[i].lock pool_lock irq_context: 0 (wq_completion)bond115 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &idev->mc_lock rcu_read_lock rcu_read_lock rcu_node_0 irq_context: 0 (wq_completion)bond115 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &idev->mc_lock rcu_read_lock rcu_read_lock &rq->__lock irq_context: 0 (wq_completion)bond115 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &idev->mc_lock rcu_read_lock rcu_read_lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)bond115 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &idev->mc_lock rcu_read_lock rcu_read_lock &dir->lock#2 irq_context: 0 (wq_completion)bond115 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &idev->mc_lock rcu_read_lock rcu_read_lock &tbl->lock irq_context: 0 (wq_completion)bond115 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &idev->mc_lock rcu_read_lock rcu_read_lock &n->lock irq_context: 0 (wq_completion)bond115 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &idev->mc_lock rcu_read_lock rcu_read_lock &____s->seqcount#9 irq_context: 0 (wq_completion)bond115 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &idev->mc_lock rcu_read_lock rcu_read_lock rcu_read_lock_bh rcu_read_lock rcu_read_lock_bh rcu_read_lock &br->multicast_lock irq_context: 0 (wq_completion)bond115 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &idev->mc_lock rcu_read_lock rcu_read_lock rcu_read_lock_bh rcu_read_lock rcu_read_lock_bh rcu_read_lock &br->multicast_lock &obj_hash[i].lock irq_context: 0 (wq_completion)bond115 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &idev->mc_lock rcu_read_lock rcu_read_lock rcu_read_lock_bh rcu_read_lock rcu_read_lock_bh rcu_read_lock &br->multicast_lock &base->lock irq_context: 0 (wq_completion)bond115 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &idev->mc_lock rcu_read_lock rcu_read_lock rcu_read_lock_bh rcu_read_lock rcu_read_lock_bh rcu_read_lock &br->multicast_lock &base->lock &obj_hash[i].lock irq_context: 0 (wq_completion)bond115 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &idev->mc_lock rcu_read_lock rcu_read_lock rcu_read_lock_bh rcu_read_lock rcu_read_lock_bh rcu_read_lock &data->lock irq_context: 0 (wq_completion)bond115 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &idev->mc_lock rcu_read_lock rcu_read_lock rcu_read_lock_bh rcu_read_lock rcu_read_lock_bh rcu_read_lock &obj_hash[i].lock irq_context: 0 (wq_completion)bond115 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &idev->mc_lock rcu_read_lock rcu_node_0 irq_context: 0 (wq_completion)bond115 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &rq->__lock irq_context: 0 (wq_completion)bond115 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex.wait_lock irq_context: 0 (wq_completion)bond115 (work_completion)(&(&bond->mcast_work)->work) &p->pi_lock irq_context: 0 (wq_completion)bond118#3 (work_completion)(&(&slave->notify_work)->work) rtnl_mutex net_rwsem irq_context: 0 (wq_completion)bond118#3 (work_completion)(&(&slave->notify_work)->work) rtnl_mutex net_rwsem &list->lock#2 irq_context: 0 (wq_completion)bond118#3 (work_completion)(&(&slave->notify_work)->work) rtnl_mutex &tn->lock irq_context: 0 (wq_completion)bond118#3 (work_completion)(&(&slave->notify_work)->work) rtnl_mutex &rq->__lock irq_context: 0 (wq_completion)bond118#3 (work_completion)(&(&slave->notify_work)->work) rtnl_mutex &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)bond118#3 (work_completion)(&(&slave->notify_work)->work) rtnl_mutex.wait_lock irq_context: 0 (wq_completion)bond118#3 (work_completion)(&(&slave->notify_work)->work) &p->pi_lock irq_context: 0 (wq_completion)bond55#2 (work_completion)(&(&slave->notify_work)->work) rtnl_mutex net_rwsem irq_context: 0 (wq_completion)bond55#2 (work_completion)(&(&slave->notify_work)->work) rtnl_mutex net_rwsem &list->lock#2 irq_context: 0 (wq_completion)bond55#2 (work_completion)(&(&slave->notify_work)->work) rtnl_mutex &tn->lock irq_context: 0 (wq_completion)bond55#2 (work_completion)(&(&slave->notify_work)->work) rtnl_mutex.wait_lock irq_context: 0 (wq_completion)bond55#2 (work_completion)(&(&slave->notify_work)->work) &p->pi_lock irq_context: 0 (wq_completion)bond170#2 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex fs_reclaim irq_context: 0 (wq_completion)bond170#2 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex fs_reclaim mmu_notifier_invalidate_range_start irq_context: 0 (wq_completion)bond170#2 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex fs_reclaim mmu_notifier_invalidate_range_start &rq->__lock irq_context: 0 (wq_completion)bond170#2 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex fs_reclaim mmu_notifier_invalidate_range_start &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)bond170#2 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &c->lock irq_context: 0 (wq_completion)bond170#2 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex rcu_read_lock &bond->stats_lock/1 irq_context: 0 (wq_completion)bond170#2 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex rcu_read_lock &ndev->lock irq_context: 0 (wq_completion)bond170#2 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex nl_table_lock irq_context: 0 (wq_completion)bond170#2 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &obj_hash[i].lock irq_context: 0 (wq_completion)bond170#2 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex nl_table_wait.lock irq_context: 0 (wq_completion)bond170#2 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex net_rwsem irq_context: 0 (wq_completion)bond170#2 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex net_rwsem &list->lock#2 irq_context: 0 (wq_completion)bond170#2 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &tn->lock irq_context: 0 (wq_completion)bond170#2 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &idev->mc_lock irq_context: 0 (wq_completion)bond170#2 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &idev->mc_lock fs_reclaim irq_context: 0 (wq_completion)bond170#2 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &idev->mc_lock fs_reclaim &rq->__lock irq_context: 0 (wq_completion)bond170#2 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &idev->mc_lock fs_reclaim &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)bond170#2 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &idev->mc_lock fs_reclaim mmu_notifier_invalidate_range_start irq_context: 0 (wq_completion)bond170#2 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &idev->mc_lock rcu_read_lock &ndev->lock irq_context: 0 (wq_completion)bond170#2 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &idev->mc_lock rcu_read_lock &dir->lock#2 irq_context: 0 (wq_completion)bond170#2 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &idev->mc_lock rcu_read_lock &ul->lock#2 irq_context: 0 (wq_completion)bond170#2 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &idev->mc_lock rcu_read_lock rcu_read_lock &____s->seqcount#10 irq_context: 0 (wq_completion)bond170#2 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &idev->mc_lock rcu_read_lock rcu_read_lock rcu_read_lock_bh rcu_read_lock rcu_read_lock_bh rcu_read_lock &br->multicast_lock irq_context: 0 (wq_completion)bond170#2 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &idev->mc_lock rcu_read_lock rcu_read_lock rcu_read_lock_bh rcu_read_lock rcu_read_lock_bh rcu_read_lock &br->multicast_lock &obj_hash[i].lock irq_context: 0 (wq_completion)bond170#2 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &idev->mc_lock rcu_read_lock rcu_read_lock rcu_read_lock_bh rcu_read_lock rcu_read_lock_bh rcu_read_lock &br->multicast_lock &base->lock irq_context: 0 (wq_completion)bond170#2 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &idev->mc_lock rcu_read_lock rcu_read_lock rcu_read_lock_bh rcu_read_lock rcu_read_lock_bh rcu_read_lock &br->multicast_lock &base->lock &obj_hash[i].lock irq_context: 0 (wq_completion)bond170#2 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &idev->mc_lock rcu_read_lock rcu_read_lock rcu_read_lock_bh rcu_read_lock rcu_read_lock_bh rcu_read_lock &data->lock irq_context: 0 (wq_completion)bond170#2 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &idev->mc_lock rcu_read_lock rcu_read_lock rcu_read_lock_bh rcu_read_lock rcu_read_lock_bh rcu_read_lock &obj_hash[i].lock irq_context: 0 (wq_completion)bond170#2 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &idev->mc_lock rcu_read_lock rcu_read_lock rcu_node_0 irq_context: 0 (wq_completion)bond170#2 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &idev->mc_lock rcu_read_lock rcu_read_lock &rq->__lock irq_context: 0 (wq_completion)bond170#2 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &idev->mc_lock rcu_read_lock rcu_read_lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)bond170#2 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &idev->mc_lock rcu_read_lock rcu_node_0 irq_context: 0 (wq_completion)bond170#2 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex.wait_lock irq_context: 0 (wq_completion)bond170#2 (work_completion)(&(&bond->mcast_work)->work) &p->pi_lock irq_context: 0 (wq_completion)bond170#2 (work_completion)(&(&bond->mcast_work)->work) &p->pi_lock &rq->__lock irq_context: 0 (wq_completion)bond176#2 (work_completion)(&(&bond->mcast_work)->work) irq_context: 0 (wq_completion)bond170#2 (work_completion)(&(&bond->mcast_work)->work) &p->pi_lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)bond176#2 (work_completion)(&(&bond->mcast_work)->work) &obj_hash[i].lock irq_context: 0 (wq_completion)bond170#2 (work_completion)(&(&bond->mcast_work)->work) &rq->__lock irq_context: 0 (wq_completion)bond176#2 (work_completion)(&(&bond->mcast_work)->work) &base->lock irq_context: 0 (wq_completion)bond170#2 (work_completion)(&(&bond->mcast_work)->work) &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)bond176#2 (work_completion)(&(&bond->mcast_work)->work) &base->lock &obj_hash[i].lock irq_context: 0 (wq_completion)bond176#2 (work_completion)(&(&bond->mcast_work)->work) &rq->__lock irq_context: 0 (wq_completion)bond176#2 (work_completion)(&(&bond->mcast_work)->work) &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)bond157#2 &rq->__lock irq_context: 0 (wq_completion)bond157#2 &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 sb_writers#4 &type->i_mutex_dir_key#3/1 sb_internal mmu_notifier_invalidate_range_start &cfs_rq->removed.lock irq_context: 0 sb_writers#4 &type->i_mutex_dir_key#3/1 sb_internal mmu_notifier_invalidate_range_start &obj_hash[i].lock irq_context: 0 (wq_completion)bond171#2 &rq->__lock irq_context: 0 (wq_completion)bond188 (work_completion)(&(&slave->notify_work)->work) &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)bond160#3 irq_context: 0 (wq_completion)bond160#3 (work_completion)(&(&slave->notify_work)->work) irq_context: 0 (wq_completion)bond160#3 (work_completion)(&(&slave->notify_work)->work) &obj_hash[i].lock irq_context: 0 (wq_completion)bond160#3 (work_completion)(&(&slave->notify_work)->work) &base->lock irq_context: 0 (wq_completion)bond160#3 (work_completion)(&(&slave->notify_work)->work) &base->lock &obj_hash[i].lock irq_context: 0 (wq_completion)bond160#3 (work_completion)(&(&slave->notify_work)->work) &rq->__lock irq_context: 0 (wq_completion)bond160#3 (work_completion)(&(&slave->notify_work)->work) &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)bond189 irq_context: 0 (wq_completion)bond189 (work_completion)(&(&slave->notify_work)->work) irq_context: 0 (wq_completion)bond189 (work_completion)(&(&slave->notify_work)->work) &obj_hash[i].lock irq_context: 0 (wq_completion)bond189 (work_completion)(&(&slave->notify_work)->work) &base->lock irq_context: 0 (wq_completion)bond189 (work_completion)(&(&slave->notify_work)->work) &base->lock &obj_hash[i].lock irq_context: 0 (wq_completion)bond189 (work_completion)(&(&slave->notify_work)->work) &rq->__lock irq_context: 0 (wq_completion)bond160#3 (work_completion)(&(&bond->mcast_work)->work) irq_context: 0 (wq_completion)bond160#3 (work_completion)(&(&bond->mcast_work)->work) &obj_hash[i].lock irq_context: 0 (wq_completion)bond160#3 (work_completion)(&(&bond->mcast_work)->work) &base->lock irq_context: 0 (wq_completion)bond160#3 (work_completion)(&(&bond->mcast_work)->work) &base->lock &obj_hash[i].lock irq_context: 0 (wq_completion)bond189 (work_completion)(&(&bond->mcast_work)->work) irq_context: 0 (wq_completion)bond189 (work_completion)(&(&bond->mcast_work)->work) &obj_hash[i].lock irq_context: 0 (wq_completion)bond189 (work_completion)(&(&bond->mcast_work)->work) &base->lock irq_context: 0 (wq_completion)bond189 (work_completion)(&(&bond->mcast_work)->work) &base->lock &obj_hash[i].lock irq_context: 0 (wq_completion)bond189 (work_completion)(&(&bond->mcast_work)->work) &rq->__lock irq_context: 0 (wq_completion)bond189 (work_completion)(&(&bond->mcast_work)->work) &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 &f->f_pos_lock &type->i_mutex_dir_key#3 rcu_read_lock &obj_hash[i].lock pool_lock irq_context: 0 &f->f_pos_lock &type->i_mutex_dir_key#3 rcu_read_lock &rq->__lock &cfs_rq->removed.lock irq_context: 0 (wq_completion)bond155#3 &rq->__lock irq_context: 0 (wq_completion)bond155#3 &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)bond158 (work_completion)(&(&slave->notify_work)->work) &rq->__lock irq_context: 0 (wq_completion)bond158 (work_completion)(&(&slave->notify_work)->work) &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)bond56#2 irq_context: 0 (wq_completion)bond56#2 (work_completion)(&(&slave->notify_work)->work) irq_context: 0 (wq_completion)bond56#2 (work_completion)(&(&slave->notify_work)->work) &obj_hash[i].lock irq_context: 0 (wq_completion)bond56#2 (work_completion)(&(&slave->notify_work)->work) &base->lock irq_context: 0 (wq_completion)bond56#2 (work_completion)(&(&slave->notify_work)->work) &base->lock &obj_hash[i].lock irq_context: 0 (wq_completion)bond56#2 (work_completion)(&(&slave->notify_work)->work) &rq->__lock irq_context: 0 (wq_completion)bond56#2 (work_completion)(&(&slave->notify_work)->work) &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)bond56#2 (work_completion)(&(&bond->mcast_work)->work) irq_context: 0 (wq_completion)bond56#2 (work_completion)(&(&bond->mcast_work)->work) &obj_hash[i].lock irq_context: 0 (wq_completion)bond56#2 (work_completion)(&(&bond->mcast_work)->work) &base->lock irq_context: 0 (wq_completion)bond56#2 (work_completion)(&(&bond->mcast_work)->work) &base->lock &obj_hash[i].lock irq_context: 0 (wq_completion)bond56#2 (work_completion)(&(&bond->mcast_work)->work) &rq->__lock irq_context: 0 (wq_completion)bond56#2 (work_completion)(&(&bond->mcast_work)->work) &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)bond177#2 irq_context: 0 (wq_completion)bond177#2 (work_completion)(&(&slave->notify_work)->work) irq_context: 0 (wq_completion)bond177#2 (work_completion)(&(&slave->notify_work)->work) &obj_hash[i].lock irq_context: 0 (wq_completion)bond177#2 (work_completion)(&(&slave->notify_work)->work) &base->lock irq_context: 0 (wq_completion)bond177#2 (work_completion)(&(&slave->notify_work)->work) &base->lock &obj_hash[i].lock irq_context: 0 (wq_completion)mld (work_completion)(&(&idev->mc_ifc_work)->work) &idev->mc_lock rcu_read_lock rcu_read_lock rcu_read_lock_bh rcu_read_lock &br->multicast_lock fill_pool_map-wait-type-override &____s->seqcount#2 irq_context: 0 (wq_completion)mld (work_completion)(&(&idev->mc_ifc_work)->work) &idev->mc_lock rcu_read_lock rcu_read_lock rcu_read_lock_bh rcu_read_lock &br->multicast_lock fill_pool_map-wait-type-override &____s->seqcount irq_context: 0 (wq_completion)bond177#2 (work_completion)(&(&bond->mcast_work)->work) irq_context: 0 (wq_completion)bond177#2 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex fs_reclaim irq_context: 0 (wq_completion)bond177#2 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex fs_reclaim mmu_notifier_invalidate_range_start irq_context: 0 (wq_completion)bond177#2 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &c->lock irq_context: 0 (wq_completion)bond177#2 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &____s->seqcount#2 irq_context: 0 (wq_completion)bond177#2 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &____s->seqcount irq_context: 0 (wq_completion)bond177#2 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex rcu_read_lock &bond->stats_lock/1 irq_context: 0 (wq_completion)bond177#2 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex rcu_read_lock &ndev->lock irq_context: 0 (wq_completion)bond177#2 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex nl_table_lock irq_context: 0 (wq_completion)bond177#2 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &obj_hash[i].lock irq_context: 0 (wq_completion)bond177#2 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex nl_table_wait.lock irq_context: 0 (wq_completion)bond177#2 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex net_rwsem irq_context: 0 (wq_completion)bond177#2 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex net_rwsem &list->lock#2 irq_context: 0 (wq_completion)bond177#2 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &tn->lock irq_context: 0 (wq_completion)bond177#2 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &idev->mc_lock irq_context: 0 (wq_completion)bond177#2 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex.wait_lock irq_context: 0 (wq_completion)bond177#2 (work_completion)(&(&bond->mcast_work)->work) &p->pi_lock irq_context: 0 (wq_completion)bond107#3 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex fs_reclaim irq_context: 0 (wq_completion)bond107#3 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex fs_reclaim mmu_notifier_invalidate_range_start irq_context: 0 (wq_completion)bond107#3 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex rcu_read_lock &bond->stats_lock/1 irq_context: 0 (wq_completion)bond107#3 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex rcu_read_lock &ndev->lock irq_context: 0 (wq_completion)bond107#3 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex nl_table_lock irq_context: 0 (wq_completion)bond107#3 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &obj_hash[i].lock irq_context: 0 (wq_completion)bond107#3 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex nl_table_wait.lock irq_context: 0 (wq_completion)bond107#3 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex net_rwsem irq_context: 0 (wq_completion)bond107#3 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex net_rwsem &list->lock#2 irq_context: 0 (wq_completion)bond107#3 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &tn->lock irq_context: 0 (wq_completion)bond107#3 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &idev->mc_lock irq_context: 0 (wq_completion)bond107#3 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &idev->mc_lock fs_reclaim irq_context: 0 (wq_completion)bond107#3 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &idev->mc_lock fs_reclaim mmu_notifier_invalidate_range_start irq_context: 0 (wq_completion)bond107#3 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &idev->mc_lock &c->lock irq_context: 0 (wq_completion)bond107#3 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &idev->mc_lock rcu_read_lock &ndev->lock irq_context: 0 (wq_completion)bond107#3 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &idev->mc_lock rcu_read_lock &dir->lock#2 irq_context: 0 (wq_completion)bond107#3 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &idev->mc_lock rcu_read_lock &ul->lock#2 irq_context: 0 (wq_completion)bond107#3 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &idev->mc_lock rcu_read_lock rcu_read_lock &obj_hash[i].lock irq_context: 0 (wq_completion)bond107#3 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &idev->mc_lock rcu_read_lock rcu_read_lock &dir->lock#2 irq_context: 0 (wq_completion)bond107#3 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &idev->mc_lock rcu_read_lock rcu_read_lock &tbl->lock irq_context: 0 (wq_completion)bond107#3 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &idev->mc_lock rcu_read_lock rcu_read_lock &n->lock irq_context: 0 (wq_completion)bond107#3 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &idev->mc_lock rcu_read_lock rcu_read_lock &____s->seqcount#9 irq_context: 0 (wq_completion)bond107#3 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &idev->mc_lock rcu_read_lock rcu_read_lock rcu_read_lock_bh rcu_read_lock rcu_read_lock_bh rcu_read_lock &br->multicast_lock irq_context: 0 (wq_completion)bond107#3 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &idev->mc_lock rcu_read_lock rcu_read_lock rcu_read_lock_bh rcu_read_lock rcu_read_lock_bh rcu_read_lock &br->multicast_lock &obj_hash[i].lock irq_context: 0 (wq_completion)bond107#3 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &idev->mc_lock rcu_read_lock rcu_read_lock rcu_read_lock_bh rcu_read_lock rcu_read_lock_bh rcu_read_lock &br->multicast_lock &base->lock irq_context: 0 (wq_completion)bond107#3 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &idev->mc_lock rcu_read_lock rcu_read_lock rcu_read_lock_bh rcu_read_lock rcu_read_lock_bh rcu_read_lock &br->multicast_lock &base->lock &obj_hash[i].lock irq_context: 0 (wq_completion)bond107#3 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &idev->mc_lock rcu_read_lock rcu_read_lock rcu_read_lock_bh rcu_read_lock rcu_read_lock_bh rcu_read_lock &data->lock irq_context: 0 (wq_completion)bond107#3 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &idev->mc_lock rcu_read_lock rcu_read_lock rcu_read_lock_bh rcu_read_lock rcu_read_lock_bh rcu_read_lock &obj_hash[i].lock irq_context: 0 (wq_completion)bond107#3 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex.wait_lock irq_context: 0 (wq_completion)bond107#3 (work_completion)(&(&bond->mcast_work)->work) &p->pi_lock irq_context: 0 (wq_completion)bond160#2 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex fs_reclaim irq_context: 0 (wq_completion)bond160#2 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex fs_reclaim mmu_notifier_invalidate_range_start irq_context: 0 (wq_completion)bond160#2 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex rcu_read_lock &bond->stats_lock/1 irq_context: 0 (wq_completion)bond160#2 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex rcu_read_lock &ndev->lock irq_context: 0 (wq_completion)bond160#2 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex nl_table_lock irq_context: 0 (wq_completion)bond160#2 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &obj_hash[i].lock irq_context: 0 (wq_completion)bond160#2 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex nl_table_wait.lock irq_context: 0 (wq_completion)bond160#2 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex net_rwsem irq_context: 0 (wq_completion)bond160#2 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex net_rwsem &list->lock#2 irq_context: 0 (wq_completion)bond160#2 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &tn->lock irq_context: 0 (wq_completion)bond160#2 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &idev->mc_lock irq_context: 0 (wq_completion)bond160#2 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &idev->mc_lock fs_reclaim irq_context: 0 (wq_completion)bond160#2 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &idev->mc_lock fs_reclaim mmu_notifier_invalidate_range_start irq_context: 0 (wq_completion)bond160#2 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &idev->mc_lock rcu_read_lock &ndev->lock irq_context: 0 (wq_completion)bond160#2 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &idev->mc_lock rcu_read_lock &c->lock irq_context: 0 (wq_completion)bond160#2 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &idev->mc_lock rcu_read_lock &dir->lock#2 irq_context: 0 (wq_completion)bond160#2 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &idev->mc_lock rcu_read_lock &ul->lock#2 irq_context: 0 (wq_completion)bond160#2 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &idev->mc_lock rcu_read_lock rcu_node_0 irq_context: 0 (wq_completion)bond160#2 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &idev->mc_lock rcu_read_lock &rq->__lock irq_context: 0 (wq_completion)bond160#2 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &idev->mc_lock rcu_read_lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)bond160#2 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &idev->mc_lock rcu_read_lock rcu_read_lock &____s->seqcount#10 irq_context: 0 (wq_completion)bond160#2 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &idev->mc_lock rcu_read_lock rcu_read_lock rcu_read_lock_bh rcu_read_lock rcu_read_lock_bh rcu_read_lock &br->multicast_lock irq_context: 0 (wq_completion)bond160#2 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &idev->mc_lock rcu_read_lock rcu_read_lock rcu_read_lock_bh rcu_read_lock rcu_read_lock_bh rcu_read_lock &br->multicast_lock &obj_hash[i].lock irq_context: 0 (wq_completion)bond160#2 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &idev->mc_lock rcu_read_lock rcu_read_lock rcu_read_lock_bh rcu_read_lock rcu_read_lock_bh rcu_read_lock &br->multicast_lock &base->lock irq_context: 0 (wq_completion)bond160#2 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &idev->mc_lock rcu_read_lock rcu_read_lock rcu_read_lock_bh rcu_read_lock rcu_read_lock_bh rcu_read_lock &br->multicast_lock &base->lock &obj_hash[i].lock irq_context: 0 (wq_completion)bond160#2 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &idev->mc_lock rcu_read_lock rcu_read_lock rcu_read_lock_bh rcu_read_lock rcu_read_lock_bh rcu_read_lock &data->lock irq_context: 0 (wq_completion)bond160#2 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &idev->mc_lock rcu_read_lock rcu_read_lock rcu_read_lock_bh rcu_read_lock rcu_read_lock_bh rcu_read_lock &obj_hash[i].lock irq_context: 0 (wq_completion)bond160#2 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &idev->mc_lock rcu_read_lock rcu_read_lock &rq->__lock irq_context: 0 (wq_completion)bond160#2 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &idev->mc_lock rcu_read_lock rcu_read_lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)bond160#2 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex.wait_lock irq_context: 0 (wq_completion)bond160#2 (work_completion)(&(&bond->mcast_work)->work) &p->pi_lock irq_context: 0 (wq_completion)bond160#2 (work_completion)(&(&slave->notify_work)->work) rtnl_mutex net_rwsem irq_context: 0 (wq_completion)bond160#2 (work_completion)(&(&slave->notify_work)->work) rtnl_mutex net_rwsem &list->lock#2 irq_context: 0 (wq_completion)bond160#2 (work_completion)(&(&slave->notify_work)->work) rtnl_mutex &tn->lock irq_context: 0 (wq_completion)bond160#2 (work_completion)(&(&slave->notify_work)->work) rtnl_mutex rcu_read_lock rcu_node_0 irq_context: 0 (wq_completion)bond160#2 (work_completion)(&(&slave->notify_work)->work) rtnl_mutex rcu_read_lock &rq->__lock irq_context: 0 (wq_completion)bond160#2 (work_completion)(&(&slave->notify_work)->work) rtnl_mutex rcu_read_lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)bond160#2 (work_completion)(&(&slave->notify_work)->work) rtnl_mutex.wait_lock irq_context: 0 (wq_completion)bond160#2 (work_completion)(&(&slave->notify_work)->work) &p->pi_lock irq_context: 0 (wq_completion)bond160#2 (work_completion)(&(&slave->notify_work)->work) &p->pi_lock &rq->__lock irq_context: 0 (wq_completion)bond160#2 (work_completion)(&(&slave->notify_work)->work) &p->pi_lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)bond171#2 (work_completion)(&(&slave->notify_work)->work) &rq->__lock irq_context: 0 (wq_completion)bond171#2 (work_completion)(&(&slave->notify_work)->work) &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)bond161#3 irq_context: 0 (wq_completion)bond161#3 (work_completion)(&(&slave->notify_work)->work) irq_context: 0 (wq_completion)bond161#3 (work_completion)(&(&slave->notify_work)->work) &rq->__lock irq_context: 0 (wq_completion)bond161#3 (work_completion)(&(&slave->notify_work)->work) &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)bond161#3 (work_completion)(&(&slave->notify_work)->work) &obj_hash[i].lock irq_context: 0 (wq_completion)bond161#3 (work_completion)(&(&slave->notify_work)->work) &base->lock irq_context: 0 (wq_completion)bond161#3 (work_completion)(&(&slave->notify_work)->work) &base->lock &obj_hash[i].lock irq_context: 0 (wq_completion)bond161#3 (work_completion)(&(&bond->mcast_work)->work) irq_context: 0 (wq_completion)bond161#3 (work_completion)(&(&bond->mcast_work)->work) &obj_hash[i].lock irq_context: 0 (wq_completion)bond161#3 (work_completion)(&(&bond->mcast_work)->work) &base->lock irq_context: 0 (wq_completion)bond161#3 (work_completion)(&(&bond->mcast_work)->work) &base->lock &obj_hash[i].lock irq_context: 0 (wq_completion)bond190 irq_context: 0 (wq_completion)bond190 (work_completion)(&(&slave->notify_work)->work) irq_context: 0 (wq_completion)bond190 (work_completion)(&(&slave->notify_work)->work) &obj_hash[i].lock irq_context: 0 (wq_completion)bond190 (work_completion)(&(&slave->notify_work)->work) &base->lock irq_context: 0 (wq_completion)bond190 (work_completion)(&(&slave->notify_work)->work) &base->lock &obj_hash[i].lock irq_context: 0 (wq_completion)bond190 (work_completion)(&(&slave->notify_work)->work) &rq->__lock irq_context: 0 (wq_completion)bond190 (work_completion)(&(&slave->notify_work)->work) &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)bond190 (work_completion)(&(&bond->mcast_work)->work) irq_context: 0 (wq_completion)bond190 (work_completion)(&(&bond->mcast_work)->work) &obj_hash[i].lock irq_context: 0 (wq_completion)bond190 (work_completion)(&(&bond->mcast_work)->work) &base->lock irq_context: 0 (wq_completion)bond190 (work_completion)(&(&bond->mcast_work)->work) &base->lock &obj_hash[i].lock irq_context: 0 (wq_completion)bond190 (work_completion)(&(&bond->mcast_work)->work) &rq->__lock irq_context: 0 (wq_completion)bond190 (work_completion)(&(&bond->mcast_work)->work) &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)bond158 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex fs_reclaim irq_context: 0 (wq_completion)bond158 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex fs_reclaim mmu_notifier_invalidate_range_start irq_context: 0 (wq_completion)bond158 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &c->lock irq_context: 0 (wq_completion)bond158 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex rcu_read_lock &bond->stats_lock/1 irq_context: 0 (wq_completion)bond158 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex rcu_read_lock &ndev->lock irq_context: 0 (wq_completion)bond158 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex nl_table_lock irq_context: 0 (wq_completion)bond158 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &obj_hash[i].lock irq_context: 0 (wq_completion)bond158 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex nl_table_wait.lock irq_context: 0 (wq_completion)bond158 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex net_rwsem irq_context: 0 (wq_completion)bond158 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex net_rwsem &list->lock#2 irq_context: 0 (wq_completion)bond158 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &tn->lock irq_context: 0 (wq_completion)bond158 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &idev->mc_lock irq_context: 0 (wq_completion)bond158 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &idev->mc_lock fs_reclaim irq_context: 0 (wq_completion)bond158 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &idev->mc_lock fs_reclaim mmu_notifier_invalidate_range_start irq_context: 0 (wq_completion)bond158 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &idev->mc_lock rcu_read_lock &ndev->lock irq_context: 0 (wq_completion)bond158 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &idev->mc_lock rcu_read_lock &dir->lock#2 irq_context: 0 (wq_completion)bond158 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &idev->mc_lock rcu_read_lock &ul->lock#2 irq_context: 0 (wq_completion)bond158 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &idev->mc_lock rcu_read_lock rcu_read_lock &____s->seqcount#10 irq_context: 0 (wq_completion)bond158 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &idev->mc_lock rcu_read_lock rcu_read_lock rcu_read_lock_bh rcu_read_lock rcu_read_lock_bh rcu_read_lock &br->multicast_lock irq_context: 0 (wq_completion)bond158 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &idev->mc_lock rcu_read_lock rcu_read_lock rcu_read_lock_bh rcu_read_lock rcu_read_lock_bh rcu_read_lock &br->multicast_lock &obj_hash[i].lock irq_context: 0 (wq_completion)bond158 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &idev->mc_lock rcu_read_lock rcu_read_lock rcu_read_lock_bh rcu_read_lock rcu_read_lock_bh rcu_read_lock &br->multicast_lock &base->lock irq_context: 0 (wq_completion)bond158 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &idev->mc_lock rcu_read_lock rcu_read_lock rcu_read_lock_bh rcu_read_lock rcu_read_lock_bh rcu_read_lock &br->multicast_lock &base->lock &obj_hash[i].lock irq_context: 0 (wq_completion)bond158 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &idev->mc_lock rcu_read_lock rcu_read_lock rcu_read_lock_bh rcu_read_lock rcu_read_lock_bh rcu_read_lock &data->lock irq_context: 0 (wq_completion)bond158 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &idev->mc_lock rcu_read_lock rcu_read_lock rcu_read_lock_bh rcu_read_lock rcu_read_lock_bh rcu_read_lock &obj_hash[i].lock irq_context: 0 (wq_completion)bond158 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex.wait_lock irq_context: 0 (wq_completion)bond158 (work_completion)(&(&bond->mcast_work)->work) &p->pi_lock irq_context: 0 (wq_completion)bond182 (work_completion)(&(&slave->notify_work)->work) rtnl_mutex net_rwsem irq_context: 0 (wq_completion)bond182 (work_completion)(&(&slave->notify_work)->work) rtnl_mutex net_rwsem &list->lock#2 irq_context: 0 (wq_completion)bond182 (work_completion)(&(&slave->notify_work)->work) rtnl_mutex &tn->lock irq_context: 0 (wq_completion)bond182 (work_completion)(&(&slave->notify_work)->work) rtnl_mutex.wait_lock irq_context: 0 (wq_completion)bond182 (work_completion)(&(&slave->notify_work)->work) &p->pi_lock irq_context: 0 (wq_completion)bond146#3 (work_completion)(&(&slave->notify_work)->work) rtnl_mutex net_rwsem irq_context: 0 (wq_completion)bond146#3 (work_completion)(&(&slave->notify_work)->work) rtnl_mutex net_rwsem &list->lock#2 irq_context: 0 (wq_completion)bond146#3 (work_completion)(&(&slave->notify_work)->work) rtnl_mutex &tn->lock irq_context: 0 (wq_completion)bond146#3 (work_completion)(&(&slave->notify_work)->work) rtnl_mutex.wait_lock irq_context: 0 (wq_completion)bond146#3 (work_completion)(&(&slave->notify_work)->work) &p->pi_lock irq_context: 0 (wq_completion)bond56#2 (work_completion)(&(&slave->notify_work)->work) rtnl_mutex net_rwsem irq_context: 0 (wq_completion)bond56#2 (work_completion)(&(&slave->notify_work)->work) rtnl_mutex net_rwsem &list->lock#2 irq_context: 0 (wq_completion)bond56#2 (work_completion)(&(&slave->notify_work)->work) rtnl_mutex &tn->lock irq_context: 0 (wq_completion)bond56#2 (work_completion)(&(&slave->notify_work)->work) rtnl_mutex.wait_lock irq_context: 0 (wq_completion)bond56#2 (work_completion)(&(&slave->notify_work)->work) &p->pi_lock irq_context: 0 (wq_completion)bond51#3 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex fs_reclaim irq_context: 0 (wq_completion)bond51#3 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex fs_reclaim mmu_notifier_invalidate_range_start irq_context: 0 (wq_completion)bond51#3 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &c->lock irq_context: 0 (wq_completion)bond51#3 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &n->list_lock irq_context: 0 (wq_completion)bond51#3 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &n->list_lock &c->lock irq_context: 0 (wq_completion)bond51#3 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex rcu_read_lock &bond->stats_lock/1 irq_context: 0 (wq_completion)bond51#3 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex rcu_read_lock &ndev->lock irq_context: 0 (wq_completion)bond51#3 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex nl_table_lock irq_context: 0 (wq_completion)bond51#3 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &obj_hash[i].lock irq_context: 0 (wq_completion)bond51#3 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex nl_table_wait.lock irq_context: 0 (wq_completion)bond51#3 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex net_rwsem irq_context: 0 (wq_completion)bond51#3 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex net_rwsem &list->lock#2 irq_context: 0 (wq_completion)bond51#3 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &tn->lock irq_context: 0 (wq_completion)bond51#3 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &idev->mc_lock irq_context: 0 (wq_completion)bond51#3 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &idev->mc_lock fs_reclaim irq_context: 0 (wq_completion)bond51#3 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &idev->mc_lock fs_reclaim mmu_notifier_invalidate_range_start irq_context: 0 (wq_completion)bond51#3 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &idev->mc_lock rcu_read_lock &ndev->lock irq_context: 0 (wq_completion)bond51#3 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &idev->mc_lock rcu_read_lock &dir->lock#2 irq_context: 0 (wq_completion)bond51#3 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &idev->mc_lock rcu_read_lock &c->lock irq_context: 0 (wq_completion)bond51#3 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &idev->mc_lock rcu_read_lock &ul->lock#2 irq_context: 0 (wq_completion)bond51#3 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &idev->mc_lock rcu_read_lock rcu_node_0 irq_context: 0 (wq_completion)bond51#3 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &idev->mc_lock rcu_read_lock &rq->__lock irq_context: 0 (wq_completion)bond51#3 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &idev->mc_lock rcu_read_lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)bond51#3 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &idev->mc_lock rcu_read_lock rcu_read_lock &____s->seqcount#12 irq_context: 0 (wq_completion)bond51#3 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &idev->mc_lock rcu_read_lock rcu_read_lock &____s->seqcount#16 irq_context: 0 (wq_completion)bond51#3 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &idev->mc_lock rcu_read_lock rcu_read_lock &c->lock irq_context: 0 (wq_completion)bond51#3 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &idev->mc_lock rcu_read_lock &obj_hash[i].lock irq_context: 0 (wq_completion)bond51#3 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &idev->mc_lock rcu_read_lock &data->lock irq_context: 0 (wq_completion)bond51#3 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex.wait_lock irq_context: 0 (wq_completion)bond51#3 (work_completion)(&(&bond->mcast_work)->work) &p->pi_lock irq_context: 0 (wq_completion)bond51#3 (work_completion)(&(&bond->mcast_work)->work) &p->pi_lock &rq->__lock irq_context: 0 (wq_completion)bond51#3 (work_completion)(&(&bond->mcast_work)->work) &p->pi_lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)bond177#2 (work_completion)(&(&slave->notify_work)->work) &rq->__lock irq_context: 0 (wq_completion)bond177#2 (work_completion)(&(&slave->notify_work)->work) &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)bond164#2 (work_completion)(&(&bond->mcast_work)->work) &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)bond168 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex fs_reclaim irq_context: 0 (wq_completion)bond168 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex fs_reclaim &rq->__lock irq_context: 0 (wq_completion)bond168 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex fs_reclaim mmu_notifier_invalidate_range_start irq_context: 0 (wq_completion)bond168 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &c->lock irq_context: 0 (wq_completion)bond168 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &n->list_lock irq_context: 0 (wq_completion)bond168 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &n->list_lock &c->lock irq_context: 0 (wq_completion)bond168 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &rq->__lock irq_context: 0 (wq_completion)bond168 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)bond168 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex rcu_read_lock &bond->stats_lock/1 irq_context: 0 (wq_completion)bond168 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex rcu_read_lock &ndev->lock irq_context: 0 (wq_completion)bond168 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex nl_table_lock irq_context: 0 (wq_completion)bond168 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &obj_hash[i].lock irq_context: 0 (wq_completion)bond168 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex nl_table_wait.lock irq_context: 0 (wq_completion)bond168 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex net_rwsem irq_context: 0 (wq_completion)bond168 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex net_rwsem &list->lock#2 irq_context: 0 (wq_completion)bond168 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &tn->lock irq_context: 0 (wq_completion)bond168 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &idev->mc_lock irq_context: 0 (wq_completion)bond168 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &idev->mc_lock fs_reclaim irq_context: 0 (wq_completion)bond168 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &idev->mc_lock fs_reclaim mmu_notifier_invalidate_range_start irq_context: 0 (wq_completion)bond168 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &idev->mc_lock fs_reclaim mmu_notifier_invalidate_range_start &rq->__lock irq_context: 0 (wq_completion)bond168 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &idev->mc_lock fs_reclaim mmu_notifier_invalidate_range_start &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)bond168 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &idev->mc_lock &c->lock irq_context: 0 (wq_completion)bond168 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &idev->mc_lock rcu_read_lock &ndev->lock irq_context: 0 (wq_completion)bond168 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &idev->mc_lock rcu_read_lock &c->lock irq_context: 0 (wq_completion)bond168 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &idev->mc_lock rcu_read_lock &dir->lock#2 irq_context: 0 (wq_completion)bond168 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &idev->mc_lock rcu_read_lock &ul->lock#2 irq_context: 0 (wq_completion)bond168 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &idev->mc_lock rcu_read_lock rcu_read_lock &____s->seqcount#10 irq_context: 0 (wq_completion)bond168 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &idev->mc_lock rcu_read_lock rcu_read_lock rcu_read_lock_bh rcu_read_lock rcu_read_lock_bh rcu_read_lock &br->multicast_lock irq_context: 0 (wq_completion)bond168 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &idev->mc_lock rcu_read_lock rcu_read_lock rcu_read_lock_bh rcu_read_lock rcu_read_lock_bh rcu_read_lock &br->multicast_lock &obj_hash[i].lock irq_context: 0 (wq_completion)bond168 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &idev->mc_lock rcu_read_lock rcu_read_lock rcu_read_lock_bh rcu_read_lock rcu_read_lock_bh rcu_read_lock &br->multicast_lock &base->lock irq_context: 0 (wq_completion)bond168 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &idev->mc_lock rcu_read_lock rcu_read_lock rcu_read_lock_bh rcu_read_lock rcu_read_lock_bh rcu_read_lock &br->multicast_lock &base->lock &obj_hash[i].lock irq_context: 0 (wq_completion)bond168 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &idev->mc_lock rcu_read_lock rcu_read_lock rcu_read_lock_bh rcu_read_lock rcu_read_lock_bh rcu_read_lock &data->lock irq_context: 0 (wq_completion)bond168 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &idev->mc_lock rcu_read_lock rcu_read_lock rcu_read_lock_bh rcu_read_lock rcu_read_lock_bh rcu_read_lock &obj_hash[i].lock irq_context: 0 (wq_completion)bond168 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &idev->mc_lock rcu_read_lock rcu_read_lock rcu_node_0 irq_context: 0 (wq_completion)bond168 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &idev->mc_lock rcu_read_lock rcu_read_lock &rq->__lock irq_context: 0 (wq_completion)bond168 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &idev->mc_lock rcu_read_lock rcu_read_lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)bond168 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &idev->mc_lock rcu_read_lock rcu_node_0 irq_context: 0 (wq_completion)bond168 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex.wait_lock irq_context: 0 (wq_completion)bond168 (work_completion)(&(&bond->mcast_work)->work) &p->pi_lock irq_context: 0 (wq_completion)bond168 (work_completion)(&(&slave->notify_work)->work) rtnl_mutex net_rwsem irq_context: 0 (wq_completion)bond168 (work_completion)(&(&slave->notify_work)->work) rtnl_mutex net_rwsem &list->lock#2 irq_context: 0 (wq_completion)bond168 (work_completion)(&(&slave->notify_work)->work) rtnl_mutex &tn->lock irq_context: 0 (wq_completion)bond168 (work_completion)(&(&slave->notify_work)->work) rtnl_mutex.wait_lock irq_context: 0 (wq_completion)bond168 (work_completion)(&(&slave->notify_work)->work) &p->pi_lock irq_context: 0 (wq_completion)bond154#3 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex fs_reclaim irq_context: 0 (wq_completion)bond154#3 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex fs_reclaim mmu_notifier_invalidate_range_start irq_context: 0 (wq_completion)bond154#3 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &c->lock irq_context: 0 (wq_completion)bond154#3 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex rcu_read_lock &bond->stats_lock/1 irq_context: 0 (wq_completion)bond154#3 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex rcu_read_lock &ndev->lock irq_context: 0 (wq_completion)bond154#3 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex nl_table_lock irq_context: 0 (wq_completion)bond154#3 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &obj_hash[i].lock irq_context: 0 (wq_completion)bond154#3 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex nl_table_wait.lock irq_context: 0 (wq_completion)bond154#3 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &rq->__lock irq_context: 0 (wq_completion)bond154#3 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)bond154#3 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex net_rwsem irq_context: 0 (wq_completion)bond154#3 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex net_rwsem &list->lock#2 irq_context: 0 (wq_completion)bond154#3 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &tn->lock irq_context: 0 (wq_completion)bond154#3 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &idev->mc_lock irq_context: 0 (wq_completion)bond154#3 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &idev->mc_lock fs_reclaim irq_context: 0 (wq_completion)bond154#3 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &idev->mc_lock fs_reclaim mmu_notifier_invalidate_range_start irq_context: 0 (wq_completion)bond154#3 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &idev->mc_lock rcu_read_lock &ndev->lock irq_context: 0 (wq_completion)bond154#3 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &idev->mc_lock rcu_read_lock &c->lock irq_context: 0 (wq_completion)bond154#3 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &idev->mc_lock rcu_read_lock &dir->lock#2 irq_context: 0 (wq_completion)bond154#3 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &idev->mc_lock rcu_read_lock &ul->lock#2 irq_context: 0 (wq_completion)bond154#3 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &idev->mc_lock rcu_read_lock rcu_read_lock &____s->seqcount#10 irq_context: 0 (wq_completion)bond154#3 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &idev->mc_lock rcu_read_lock rcu_read_lock rcu_node_0 irq_context: 0 (wq_completion)bond154#3 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &idev->mc_lock rcu_read_lock rcu_read_lock &rq->__lock irq_context: 0 (wq_completion)bond154#3 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &idev->mc_lock rcu_read_lock rcu_read_lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)bond154#3 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &idev->mc_lock rcu_read_lock rcu_read_lock rcu_read_lock_bh rcu_read_lock rcu_read_lock_bh rcu_read_lock &br->multicast_lock irq_context: 0 (wq_completion)bond154#3 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &idev->mc_lock rcu_read_lock rcu_read_lock rcu_read_lock_bh rcu_read_lock rcu_read_lock_bh rcu_read_lock &br->multicast_lock &obj_hash[i].lock irq_context: 0 (wq_completion)bond154#3 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &idev->mc_lock rcu_read_lock rcu_read_lock rcu_read_lock_bh rcu_read_lock rcu_read_lock_bh rcu_read_lock &br->multicast_lock &base->lock irq_context: 0 (wq_completion)bond154#3 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &idev->mc_lock rcu_read_lock rcu_read_lock rcu_read_lock_bh rcu_read_lock rcu_read_lock_bh rcu_read_lock &data->lock irq_context: 0 (wq_completion)bond154#3 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &idev->mc_lock rcu_read_lock rcu_read_lock rcu_read_lock_bh rcu_read_lock rcu_read_lock_bh rcu_read_lock &obj_hash[i].lock irq_context: 0 (wq_completion)bond154#3 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &idev->mc_lock rcu_read_lock rcu_node_0 irq_context: 0 (wq_completion)bond154#3 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex.wait_lock irq_context: 0 (wq_completion)bond154#3 (work_completion)(&(&bond->mcast_work)->work) &p->pi_lock irq_context: 0 (wq_completion)bond94#2 (work_completion)(&(&slave->notify_work)->work) rtnl_mutex net_rwsem irq_context: 0 (wq_completion)bond94#2 (work_completion)(&(&slave->notify_work)->work) rtnl_mutex net_rwsem &list->lock#2 irq_context: 0 (wq_completion)bond94#2 (work_completion)(&(&slave->notify_work)->work) rtnl_mutex net_rwsem &rq->__lock irq_context: 0 (wq_completion)bond94#2 (work_completion)(&(&slave->notify_work)->work) rtnl_mutex net_rwsem &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)bond94#2 (work_completion)(&(&slave->notify_work)->work) rtnl_mutex &tn->lock irq_context: 0 (wq_completion)bond94#2 (work_completion)(&(&slave->notify_work)->work) rtnl_mutex.wait_lock irq_context: 0 (wq_completion)bond94#2 (work_completion)(&(&slave->notify_work)->work) &p->pi_lock irq_context: 0 (wq_completion)bond108#2 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex fs_reclaim irq_context: 0 (wq_completion)bond108#2 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex fs_reclaim mmu_notifier_invalidate_range_start irq_context: 0 (wq_completion)bond108#2 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &c->lock irq_context: 0 (wq_completion)bond108#2 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex rcu_read_lock &bond->stats_lock/1 irq_context: 0 (wq_completion)bond108#2 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex rcu_read_lock &ndev->lock irq_context: 0 (wq_completion)bond108#2 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex nl_table_lock irq_context: 0 (wq_completion)bond108#2 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &obj_hash[i].lock irq_context: 0 (wq_completion)bond108#2 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &rq->__lock irq_context: 0 (wq_completion)bond108#2 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)bond108#2 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex nl_table_wait.lock irq_context: 0 (wq_completion)bond108#2 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex net_rwsem irq_context: 0 (wq_completion)bond108#2 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex net_rwsem &list->lock#2 irq_context: 0 (wq_completion)bond108#2 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &tn->lock irq_context: 0 (wq_completion)bond108#2 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &idev->mc_lock irq_context: 0 (wq_completion)bond108#2 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &idev->mc_lock fs_reclaim irq_context: 0 (wq_completion)bond108#2 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &idev->mc_lock fs_reclaim mmu_notifier_invalidate_range_start irq_context: 0 (wq_completion)bond108#2 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &idev->mc_lock rcu_read_lock &ndev->lock irq_context: 0 (wq_completion)bond108#2 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &idev->mc_lock rcu_read_lock &c->lock irq_context: 0 (wq_completion)bond108#2 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &idev->mc_lock rcu_read_lock &n->list_lock irq_context: 0 (wq_completion)bond108#2 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &idev->mc_lock rcu_read_lock &n->list_lock &c->lock irq_context: 0 (wq_completion)bond108#2 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &idev->mc_lock rcu_read_lock rcu_node_0 irq_context: 0 (wq_completion)bond108#2 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &idev->mc_lock rcu_read_lock &rq->__lock irq_context: 0 (wq_completion)bond108#2 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &idev->mc_lock rcu_read_lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)bond108#2 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &idev->mc_lock rcu_read_lock &dir->lock#2 irq_context: 0 (wq_completion)bond108#2 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &idev->mc_lock rcu_read_lock &ul->lock#2 irq_context: 0 (wq_completion)bond108#2 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &idev->mc_lock rcu_read_lock rcu_read_lock &obj_hash[i].lock irq_context: 0 (wq_completion)bond108#2 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &idev->mc_lock rcu_read_lock rcu_read_lock &dir->lock#2 irq_context: 0 (wq_completion)bond108#2 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &idev->mc_lock rcu_read_lock rcu_read_lock &tbl->lock irq_context: 0 (wq_completion)bond108#2 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &idev->mc_lock rcu_read_lock rcu_read_lock &n->lock irq_context: 0 (wq_completion)bond108#2 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &idev->mc_lock rcu_read_lock rcu_read_lock &____s->seqcount#9 irq_context: 0 (wq_completion)bond108#2 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &idev->mc_lock rcu_read_lock rcu_read_lock rcu_read_lock_bh rcu_read_lock rcu_read_lock_bh rcu_read_lock &br->multicast_lock irq_context: 0 (wq_completion)bond108#2 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &idev->mc_lock rcu_read_lock rcu_read_lock rcu_read_lock_bh rcu_read_lock rcu_read_lock_bh rcu_read_lock &br->multicast_lock &obj_hash[i].lock irq_context: 0 (wq_completion)bond108#2 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &idev->mc_lock rcu_read_lock rcu_read_lock rcu_read_lock_bh rcu_read_lock rcu_read_lock_bh rcu_read_lock &br->multicast_lock &base->lock irq_context: 0 (wq_completion)bond108#2 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &idev->mc_lock rcu_read_lock rcu_read_lock rcu_read_lock_bh rcu_read_lock rcu_read_lock_bh rcu_read_lock &br->multicast_lock &base->lock &obj_hash[i].lock irq_context: 0 (wq_completion)bond108#2 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &idev->mc_lock rcu_read_lock rcu_read_lock rcu_read_lock_bh rcu_read_lock rcu_read_lock_bh rcu_read_lock &data->lock irq_context: 0 (wq_completion)bond108#2 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &idev->mc_lock rcu_read_lock rcu_read_lock rcu_read_lock_bh rcu_read_lock rcu_read_lock_bh rcu_read_lock &obj_hash[i].lock irq_context: 0 (wq_completion)bond108#2 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &idev->mc_lock rcu_read_lock rcu_read_lock &rq->__lock irq_context: 0 (wq_completion)bond108#2 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex.wait_lock irq_context: 0 (wq_completion)bond108#2 (work_completion)(&(&bond->mcast_work)->work) &p->pi_lock irq_context: 0 (wq_completion)bond149#3 (work_completion)(&(&slave->notify_work)->work) rtnl_mutex net_rwsem irq_context: 0 (wq_completion)bond149#3 (work_completion)(&(&slave->notify_work)->work) rtnl_mutex net_rwsem &list->lock#2 irq_context: 0 (wq_completion)bond149#3 (work_completion)(&(&slave->notify_work)->work) rtnl_mutex &tn->lock irq_context: 0 (wq_completion)bond149#3 (work_completion)(&(&slave->notify_work)->work) rtnl_mutex.wait_lock irq_context: 0 (wq_completion)bond149#3 (work_completion)(&(&slave->notify_work)->work) &p->pi_lock irq_context: 0 (wq_completion)bond49#2 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex fs_reclaim irq_context: 0 (wq_completion)bond49#2 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex fs_reclaim &rq->__lock irq_context: 0 (wq_completion)bond49#2 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex fs_reclaim &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)bond49#2 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex fs_reclaim mmu_notifier_invalidate_range_start irq_context: 0 (wq_completion)bond49#2 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &c->lock irq_context: 0 (wq_completion)bond49#2 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex rcu_read_lock &bond->stats_lock/1 irq_context: 0 (wq_completion)bond49#2 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex rcu_read_lock &ndev->lock irq_context: 0 (wq_completion)bond49#2 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex nl_table_lock irq_context: 0 (wq_completion)bond49#2 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &obj_hash[i].lock irq_context: 0 (wq_completion)bond49#2 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex nl_table_wait.lock irq_context: 0 (wq_completion)bond49#2 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex net_rwsem irq_context: 0 (wq_completion)bond49#2 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex net_rwsem &list->lock#2 irq_context: 0 (wq_completion)bond49#2 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &tn->lock irq_context: 0 (wq_completion)bond49#2 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &idev->mc_lock irq_context: 0 (wq_completion)bond49#2 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &idev->mc_lock &rq->__lock irq_context: 0 (wq_completion)bond49#2 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &idev->mc_lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)bond49#2 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &idev->mc_lock fs_reclaim irq_context: 0 (wq_completion)bond49#2 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &idev->mc_lock fs_reclaim mmu_notifier_invalidate_range_start irq_context: 0 (wq_completion)bond49#2 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &idev->mc_lock &c->lock irq_context: 0 (wq_completion)bond49#2 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &idev->mc_lock rcu_read_lock &ndev->lock irq_context: 0 (wq_completion)bond49#2 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &idev->mc_lock rcu_read_lock rcu_node_0 irq_context: 0 (wq_completion)bond49#2 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &idev->mc_lock rcu_read_lock &rq->__lock irq_context: 0 (wq_completion)bond49#2 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &idev->mc_lock rcu_read_lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)bond49#2 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &idev->mc_lock rcu_read_lock &c->lock irq_context: 0 (wq_completion)bond49#2 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &idev->mc_lock rcu_read_lock &dir->lock#2 irq_context: 0 (wq_completion)bond49#2 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &idev->mc_lock rcu_read_lock &ul->lock#2 irq_context: 0 (wq_completion)bond49#2 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &idev->mc_lock rcu_read_lock rcu_read_lock &____s->seqcount#12 irq_context: 0 (wq_completion)bond49#2 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &idev->mc_lock rcu_read_lock rcu_read_lock &____s->seqcount#16 irq_context: 0 (wq_completion)bond49#2 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &idev->mc_lock rcu_read_lock &obj_hash[i].lock irq_context: 0 (wq_completion)bond49#2 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &idev->mc_lock rcu_read_lock &data->lock irq_context: 0 (wq_completion)bond49#2 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex.wait_lock irq_context: 0 (wq_completion)bond49#2 (work_completion)(&(&bond->mcast_work)->work) &p->pi_lock irq_context: 0 (wq_completion)bond49#2 (work_completion)(&(&bond->mcast_work)->work) &p->pi_lock &rq->__lock irq_context: 0 (wq_completion)bond49#2 (work_completion)(&(&bond->mcast_work)->work) &p->pi_lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)bond167#2 (work_completion)(&(&slave->notify_work)->work) &rq->__lock irq_context: 0 (wq_completion)bond167#2 (work_completion)(&(&slave->notify_work)->work) &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)bond164#2 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex fs_reclaim irq_context: 0 (wq_completion)bond164#2 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex fs_reclaim &rq->__lock irq_context: 0 (wq_completion)bond164#2 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex fs_reclaim &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)bond164#2 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex fs_reclaim mmu_notifier_invalidate_range_start irq_context: 0 (wq_completion)bond164#2 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &c->lock irq_context: 0 (wq_completion)bond164#2 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex rcu_read_lock &bond->stats_lock/1 irq_context: 0 (wq_completion)bond164#2 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex rcu_read_lock &ndev->lock irq_context: 0 (wq_completion)bond164#2 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex nl_table_lock irq_context: 0 (wq_completion)bond164#2 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &obj_hash[i].lock irq_context: 0 (wq_completion)bond164#2 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex nl_table_wait.lock irq_context: 0 (wq_completion)bond164#2 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex net_rwsem irq_context: 0 (wq_completion)bond164#2 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex net_rwsem &list->lock#2 irq_context: 0 (wq_completion)bond164#2 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex net_rwsem &rq->__lock irq_context: 0 (wq_completion)bond164#2 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex net_rwsem &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)bond164#2 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &tn->lock irq_context: 0 (wq_completion)bond164#2 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &idev->mc_lock irq_context: 0 (wq_completion)bond164#2 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &idev->mc_lock fs_reclaim irq_context: 0 (wq_completion)bond164#2 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &idev->mc_lock fs_reclaim mmu_notifier_invalidate_range_start irq_context: 0 (wq_completion)bond164#2 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &idev->mc_lock &c->lock irq_context: 0 (wq_completion)bond164#2 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &idev->mc_lock rcu_read_lock &ndev->lock irq_context: 0 (wq_completion)bond164#2 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &idev->mc_lock rcu_read_lock &dir->lock#2 irq_context: 0 (wq_completion)bond164#2 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &idev->mc_lock rcu_read_lock &ul->lock#2 irq_context: 0 (wq_completion)bond164#2 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &idev->mc_lock rcu_read_lock rcu_read_lock &____s->seqcount#10 irq_context: 0 (wq_completion)bond164#2 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &idev->mc_lock rcu_read_lock rcu_read_lock rcu_read_lock_bh rcu_read_lock rcu_read_lock_bh rcu_read_lock &br->multicast_lock irq_context: 0 (wq_completion)bond164#2 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &idev->mc_lock rcu_read_lock rcu_read_lock rcu_read_lock_bh rcu_read_lock rcu_read_lock_bh rcu_read_lock &br->multicast_lock &obj_hash[i].lock irq_context: 0 (wq_completion)bond164#2 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &idev->mc_lock rcu_read_lock rcu_read_lock rcu_read_lock_bh rcu_read_lock rcu_read_lock_bh rcu_read_lock &br->multicast_lock &base->lock irq_context: 0 (wq_completion)bond164#2 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &idev->mc_lock rcu_read_lock rcu_read_lock rcu_read_lock_bh rcu_read_lock rcu_read_lock_bh rcu_read_lock &br->multicast_lock &base->lock &obj_hash[i].lock irq_context: 0 (wq_completion)bond164#2 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &idev->mc_lock rcu_read_lock rcu_read_lock rcu_read_lock_bh rcu_read_lock rcu_read_lock_bh rcu_read_lock &data->lock irq_context: 0 (wq_completion)bond164#2 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &idev->mc_lock rcu_read_lock rcu_read_lock rcu_read_lock_bh rcu_read_lock rcu_read_lock_bh rcu_read_lock &data->lock &obj_hash[i].lock irq_context: 0 (wq_completion)bond164#2 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &idev->mc_lock rcu_read_lock rcu_read_lock rcu_read_lock_bh rcu_read_lock rcu_read_lock_bh rcu_read_lock &data->lock &base->lock irq_context: 0 (wq_completion)bond164#2 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &idev->mc_lock rcu_read_lock rcu_read_lock rcu_read_lock_bh rcu_read_lock rcu_read_lock_bh rcu_read_lock &data->lock &base->lock &obj_hash[i].lock irq_context: 0 (wq_completion)bond164#2 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &idev->mc_lock rcu_read_lock rcu_read_lock rcu_read_lock_bh rcu_read_lock rcu_read_lock_bh rcu_read_lock &obj_hash[i].lock irq_context: 0 (wq_completion)bond164#2 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &idev->mc_lock rcu_read_lock rcu_read_lock rcu_node_0 irq_context: 0 (wq_completion)bond164#2 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &idev->mc_lock rcu_read_lock rcu_read_lock &rq->__lock irq_context: 0 (wq_completion)bond164#2 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &idev->mc_lock rcu_read_lock rcu_read_lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)bond164#2 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &idev->mc_lock rcu_read_lock rcu_node_0 irq_context: 0 (wq_completion)bond164#2 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &rq->__lock irq_context: 0 (wq_completion)bond164#2 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)bond164#2 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex.wait_lock irq_context: 0 (wq_completion)bond164#2 (work_completion)(&(&bond->mcast_work)->work) &p->pi_lock irq_context: 0 (wq_completion)bond164#2 (work_completion)(&(&bond->mcast_work)->work) &p->pi_lock &cfs_rq->removed.lock irq_context: 0 (wq_completion)bond164#2 (work_completion)(&(&bond->mcast_work)->work) &p->pi_lock &rq->__lock irq_context: 0 (wq_completion)bond164#2 (work_completion)(&(&bond->mcast_work)->work) &p->pi_lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)bond125#3 &rq->__lock irq_context: 0 (wq_completion)bond125#3 &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)bond159#2 &rq->__lock irq_context: 0 (wq_completion)bond159#2 &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)bond57#2 irq_context: 0 (wq_completion)bond57#2 &rq->__lock irq_context: 0 (wq_completion)bond57#2 &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)bond57#2 (work_completion)(&(&slave->notify_work)->work) irq_context: 0 (wq_completion)bond57#2 (work_completion)(&(&slave->notify_work)->work) &obj_hash[i].lock irq_context: 0 (wq_completion)bond57#2 (work_completion)(&(&slave->notify_work)->work) &base->lock irq_context: 0 (wq_completion)bond57#2 (work_completion)(&(&slave->notify_work)->work) &base->lock &obj_hash[i].lock irq_context: 0 (wq_completion)bond57#2 (work_completion)(&(&bond->mcast_work)->work) irq_context: 0 (wq_completion)bond57#2 (work_completion)(&(&bond->mcast_work)->work) &obj_hash[i].lock irq_context: 0 (wq_completion)bond57#2 (work_completion)(&(&bond->mcast_work)->work) &base->lock irq_context: 0 (wq_completion)bond57#2 (work_completion)(&(&bond->mcast_work)->work) &base->lock &obj_hash[i].lock irq_context: 0 (wq_completion)bond57#2 (work_completion)(&(&bond->mcast_work)->work) &rq->__lock irq_context: 0 (wq_completion)bond57#2 (work_completion)(&(&bond->mcast_work)->work) &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)bond120#2 (work_completion)(&(&slave->notify_work)->work) &base->lock fill_pool_map-wait-type-override &c->lock irq_context: 0 (wq_completion)bond120#2 (work_completion)(&(&slave->notify_work)->work) &base->lock fill_pool_map-wait-type-override &____s->seqcount#2 irq_context: 0 (wq_completion)bond120#2 (work_completion)(&(&slave->notify_work)->work) &base->lock fill_pool_map-wait-type-override &____s->seqcount irq_context: 0 (wq_completion)bond120#2 (work_completion)(&(&slave->notify_work)->work) &base->lock fill_pool_map-wait-type-override pool_lock irq_context: 0 (wq_completion)bond120#2 (work_completion)(&(&slave->notify_work)->work) &cfs_rq->removed.lock irq_context: 0 (wq_completion)bond178#2 irq_context: 0 (wq_completion)bond178#2 (work_completion)(&(&slave->notify_work)->work) irq_context: 0 (wq_completion)bond178#2 (work_completion)(&(&slave->notify_work)->work) &obj_hash[i].lock irq_context: 0 (wq_completion)bond178#2 (work_completion)(&(&slave->notify_work)->work) &base->lock irq_context: 0 (wq_completion)bond178#2 (work_completion)(&(&slave->notify_work)->work) &base->lock &obj_hash[i].lock irq_context: 0 (wq_completion)bond178#2 (work_completion)(&(&slave->notify_work)->work) &rq->__lock irq_context: 0 (wq_completion)bond178#2 (work_completion)(&(&slave->notify_work)->work) &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)bond41 &rq->__lock irq_context: 0 (wq_completion)bond178#2 (work_completion)(&(&bond->mcast_work)->work) irq_context: 0 (wq_completion)bond178#2 (work_completion)(&(&bond->mcast_work)->work) &obj_hash[i].lock irq_context: 0 (wq_completion)bond178#2 (work_completion)(&(&bond->mcast_work)->work) &base->lock irq_context: 0 (wq_completion)bond178#2 (work_completion)(&(&bond->mcast_work)->work) &base->lock &obj_hash[i].lock irq_context: 0 (wq_completion)bond178#2 (work_completion)(&(&bond->mcast_work)->work) &rq->__lock irq_context: 0 (wq_completion)bond178#2 (work_completion)(&(&bond->mcast_work)->work) &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)bond129 (work_completion)(&(&bond->mcast_work)->work) &base->lock fill_pool_map-wait-type-override pool_lock irq_context: 0 (wq_completion)bond162#3 irq_context: 0 (wq_completion)bond162#3 (work_completion)(&(&slave->notify_work)->work) irq_context: 0 (wq_completion)bond162#3 (work_completion)(&(&slave->notify_work)->work) &obj_hash[i].lock irq_context: 0 (wq_completion)bond162#3 (work_completion)(&(&slave->notify_work)->work) &base->lock irq_context: 0 (wq_completion)bond162#3 (work_completion)(&(&slave->notify_work)->work) &base->lock &obj_hash[i].lock irq_context: 0 (wq_completion)bond162#3 (work_completion)(&(&slave->notify_work)->work) &rq->__lock irq_context: 0 (wq_completion)bond162#3 (work_completion)(&(&slave->notify_work)->work) &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)bond161#3 (work_completion)(&(&bond->mcast_work)->work) &rq->__lock irq_context: 0 (wq_completion)bond161#3 (work_completion)(&(&bond->mcast_work)->work) &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)bond162#3 (work_completion)(&(&bond->mcast_work)->work) irq_context: 0 (wq_completion)bond162#3 (work_completion)(&(&bond->mcast_work)->work) &obj_hash[i].lock irq_context: 0 (wq_completion)bond162#3 (work_completion)(&(&bond->mcast_work)->work) &base->lock irq_context: 0 (wq_completion)bond162#3 (work_completion)(&(&bond->mcast_work)->work) &base->lock &obj_hash[i].lock irq_context: 0 (wq_completion)bond191 irq_context: 0 (wq_completion)bond191 (work_completion)(&(&slave->notify_work)->work) irq_context: 0 (wq_completion)bond191 (work_completion)(&(&slave->notify_work)->work) &obj_hash[i].lock irq_context: 0 (wq_completion)bond191 (work_completion)(&(&slave->notify_work)->work) &base->lock irq_context: 0 (wq_completion)bond191 (work_completion)(&(&slave->notify_work)->work) &base->lock &obj_hash[i].lock irq_context: 0 (wq_completion)bond191 (work_completion)(&(&slave->notify_work)->work) &rq->__lock irq_context: 0 (wq_completion)bond191 (work_completion)(&(&slave->notify_work)->work) &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)bond191 (work_completion)(&(&bond->mcast_work)->work) irq_context: 0 (wq_completion)bond191 (work_completion)(&(&bond->mcast_work)->work) &obj_hash[i].lock irq_context: 0 (wq_completion)bond191 (work_completion)(&(&bond->mcast_work)->work) &base->lock irq_context: 0 (wq_completion)bond191 (work_completion)(&(&bond->mcast_work)->work) &base->lock &obj_hash[i].lock irq_context: 0 (wq_completion)bond151#2 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex fs_reclaim irq_context: 0 (wq_completion)bond151#2 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex fs_reclaim mmu_notifier_invalidate_range_start irq_context: 0 (wq_completion)bond151#2 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex rcu_read_lock &bond->stats_lock/1 irq_context: 0 (wq_completion)bond151#2 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex rcu_read_lock &ndev->lock irq_context: 0 (wq_completion)bond151#2 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex nl_table_lock irq_context: 0 (wq_completion)bond151#2 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &obj_hash[i].lock irq_context: 0 (wq_completion)bond151#2 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex nl_table_wait.lock irq_context: 0 (wq_completion)bond151#2 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex net_rwsem irq_context: 0 (wq_completion)bond151#2 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex net_rwsem &list->lock#2 irq_context: 0 (wq_completion)bond151#2 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &tn->lock irq_context: 0 (wq_completion)bond151#2 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &idev->mc_lock irq_context: 0 (wq_completion)bond151#2 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &idev->mc_lock fs_reclaim irq_context: 0 (wq_completion)bond151#2 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &idev->mc_lock fs_reclaim mmu_notifier_invalidate_range_start irq_context: 0 (wq_completion)bond151#2 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &idev->mc_lock rcu_read_lock &ndev->lock irq_context: 0 (wq_completion)bond151#2 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &idev->mc_lock rcu_read_lock &dir->lock#2 irq_context: 0 (wq_completion)bond151#2 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &idev->mc_lock rcu_read_lock &c->lock irq_context: 0 (wq_completion)bond151#2 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &idev->mc_lock rcu_read_lock &ul->lock#2 irq_context: 0 (wq_completion)bond151#2 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &idev->mc_lock rcu_read_lock rcu_read_lock &____s->seqcount#10 irq_context: 0 (wq_completion)bond151#2 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &idev->mc_lock rcu_read_lock rcu_read_lock rcu_read_lock_bh rcu_read_lock rcu_read_lock_bh rcu_read_lock &br->multicast_lock irq_context: 0 (wq_completion)bond151#2 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &idev->mc_lock rcu_read_lock rcu_read_lock rcu_read_lock_bh rcu_read_lock rcu_read_lock_bh rcu_read_lock &br->multicast_lock &obj_hash[i].lock irq_context: 0 (wq_completion)bond151#2 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &idev->mc_lock rcu_read_lock rcu_read_lock rcu_read_lock_bh rcu_read_lock rcu_read_lock_bh rcu_read_lock &br->multicast_lock &base->lock irq_context: 0 (wq_completion)bond151#2 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &idev->mc_lock rcu_read_lock rcu_read_lock rcu_read_lock_bh rcu_read_lock rcu_read_lock_bh rcu_read_lock &br->multicast_lock &base->lock &obj_hash[i].lock irq_context: 0 (wq_completion)bond151#2 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &idev->mc_lock rcu_read_lock rcu_read_lock rcu_read_lock_bh rcu_read_lock rcu_read_lock_bh rcu_read_lock &data->lock irq_context: 0 (wq_completion)bond151#2 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &idev->mc_lock rcu_read_lock rcu_read_lock rcu_read_lock_bh rcu_read_lock rcu_read_lock_bh rcu_read_lock &obj_hash[i].lock irq_context: 0 (wq_completion)bond151#2 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex.wait_lock irq_context: 0 (wq_completion)bond151#2 (work_completion)(&(&bond->mcast_work)->work) &p->pi_lock irq_context: 0 (wq_completion)bond181 (work_completion)(&(&slave->notify_work)->work) rtnl_mutex net_rwsem irq_context: 0 (wq_completion)bond181 (work_completion)(&(&slave->notify_work)->work) rtnl_mutex net_rwsem &list->lock#2 irq_context: 0 (wq_completion)bond181 (work_completion)(&(&slave->notify_work)->work) rtnl_mutex &tn->lock irq_context: 0 (wq_completion)bond181 (work_completion)(&(&slave->notify_work)->work) rtnl_mutex.wait_lock irq_context: 0 (wq_completion)bond181 (work_completion)(&(&slave->notify_work)->work) &p->pi_lock irq_context: 0 (wq_completion)bond185 (work_completion)(&(&slave->notify_work)->work) rtnl_mutex net_rwsem irq_context: 0 (wq_completion)bond185 (work_completion)(&(&slave->notify_work)->work) rtnl_mutex net_rwsem &list->lock#2 irq_context: 0 (wq_completion)bond185 (work_completion)(&(&slave->notify_work)->work) rtnl_mutex &tn->lock irq_context: 0 (wq_completion)bond185 (work_completion)(&(&slave->notify_work)->work) rtnl_mutex.wait_lock irq_context: 0 (wq_completion)bond185 (work_completion)(&(&slave->notify_work)->work) &p->pi_lock irq_context: 0 (wq_completion)bond166#2 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex fs_reclaim irq_context: 0 (wq_completion)bond166#2 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex fs_reclaim mmu_notifier_invalidate_range_start irq_context: 0 (wq_completion)bond166#2 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &c->lock irq_context: 0 (wq_completion)bond166#2 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex rcu_read_lock &bond->stats_lock/1 irq_context: 0 (wq_completion)bond166#2 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex rcu_read_lock &ndev->lock irq_context: 0 (wq_completion)bond166#2 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex nl_table_lock irq_context: 0 (wq_completion)bond166#2 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &obj_hash[i].lock irq_context: 0 (wq_completion)bond166#2 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex nl_table_wait.lock irq_context: 0 (wq_completion)bond166#2 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex net_rwsem irq_context: 0 (wq_completion)bond166#2 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex net_rwsem &list->lock#2 irq_context: 0 (wq_completion)bond166#2 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &tn->lock irq_context: 0 (wq_completion)bond166#2 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &idev->mc_lock irq_context: 0 (wq_completion)bond166#2 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &idev->mc_lock fs_reclaim irq_context: 0 (wq_completion)bond166#2 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &idev->mc_lock fs_reclaim mmu_notifier_invalidate_range_start irq_context: 0 (wq_completion)bond166#2 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &idev->mc_lock &c->lock irq_context: 0 (wq_completion)bond166#2 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &idev->mc_lock rcu_read_lock &ndev->lock irq_context: 0 (wq_completion)bond166#2 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &idev->mc_lock rcu_read_lock &dir->lock#2 irq_context: 0 (wq_completion)bond166#2 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &idev->mc_lock rcu_read_lock &ul->lock#2 irq_context: 0 (wq_completion)bond166#2 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &idev->mc_lock rcu_read_lock rcu_read_lock &____s->seqcount#10 irq_context: 0 (wq_completion)bond166#2 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &idev->mc_lock rcu_read_lock rcu_read_lock rcu_read_lock_bh rcu_read_lock rcu_read_lock_bh rcu_read_lock &br->multicast_lock irq_context: 0 (wq_completion)bond166#2 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &idev->mc_lock rcu_read_lock rcu_read_lock rcu_read_lock_bh rcu_read_lock rcu_read_lock_bh rcu_read_lock &br->multicast_lock &obj_hash[i].lock irq_context: 0 (wq_completion)bond166#2 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &idev->mc_lock rcu_read_lock rcu_read_lock rcu_read_lock_bh rcu_read_lock rcu_read_lock_bh rcu_read_lock &br->multicast_lock &base->lock irq_context: 0 (wq_completion)bond166#2 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &idev->mc_lock rcu_read_lock rcu_read_lock rcu_read_lock_bh rcu_read_lock rcu_read_lock_bh rcu_read_lock &br->multicast_lock &base->lock &obj_hash[i].lock irq_context: 0 (wq_completion)bond166#2 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &idev->mc_lock rcu_read_lock rcu_read_lock rcu_read_lock_bh rcu_read_lock rcu_read_lock_bh rcu_read_lock &data->lock irq_context: 0 (wq_completion)bond166#2 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &idev->mc_lock rcu_read_lock rcu_read_lock rcu_read_lock_bh rcu_read_lock rcu_read_lock_bh rcu_read_lock &obj_hash[i].lock irq_context: 0 (wq_completion)bond166#2 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex.wait_lock irq_context: 0 (wq_completion)bond166#2 (work_completion)(&(&bond->mcast_work)->work) &p->pi_lock irq_context: 0 (wq_completion)bond46#2 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex fs_reclaim irq_context: 0 (wq_completion)bond46#2 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex fs_reclaim mmu_notifier_invalidate_range_start irq_context: 0 (wq_completion)bond46#2 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex rcu_read_lock &bond->stats_lock/1 irq_context: 0 (wq_completion)bond46#2 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex rcu_read_lock &ndev->lock irq_context: 0 (wq_completion)bond46#2 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex nl_table_lock irq_context: 0 (wq_completion)bond46#2 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &obj_hash[i].lock irq_context: 0 (wq_completion)bond46#2 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex nl_table_wait.lock irq_context: 0 (wq_completion)bond46#2 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex net_rwsem irq_context: 0 (wq_completion)bond46#2 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex net_rwsem &rq->__lock irq_context: 0 (wq_completion)bond46#2 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex net_rwsem &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)bond46#2 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex net_rwsem &list->lock#2 irq_context: 0 (wq_completion)bond46#2 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &tn->lock irq_context: 0 (wq_completion)bond46#2 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &tn->lock &rq->__lock irq_context: 0 (wq_completion)bond46#2 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &tn->lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)bond46#2 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &idev->mc_lock irq_context: 0 (wq_completion)bond46#2 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &idev->mc_lock fs_reclaim irq_context: 0 (wq_completion)bond46#2 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &idev->mc_lock fs_reclaim mmu_notifier_invalidate_range_start irq_context: 0 (wq_completion)bond46#2 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &idev->mc_lock rcu_read_lock &ndev->lock irq_context: 0 (wq_completion)bond46#2 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &idev->mc_lock rcu_read_lock &dir->lock#2 irq_context: 0 (wq_completion)bond46#2 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &idev->mc_lock rcu_read_lock &c->lock irq_context: 0 (wq_completion)bond46#2 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &idev->mc_lock rcu_read_lock &ul->lock#2 irq_context: 0 (wq_completion)bond46#2 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &idev->mc_lock rcu_read_lock rcu_read_lock &____s->seqcount#12 irq_context: 0 (wq_completion)bond46#2 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &idev->mc_lock rcu_read_lock rcu_read_lock &____s->seqcount#16 irq_context: 0 (wq_completion)bond46#2 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &idev->mc_lock rcu_read_lock &obj_hash[i].lock irq_context: 0 (wq_completion)bond46#2 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &idev->mc_lock rcu_read_lock &data->lock irq_context: 0 (wq_completion)bond46#2 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex.wait_lock irq_context: 0 (wq_completion)bond46#2 (work_completion)(&(&bond->mcast_work)->work) &p->pi_lock irq_context: 0 (wq_completion)bond46#2 (work_completion)(&(&bond->mcast_work)->work) &p->pi_lock &rq->__lock irq_context: 0 (wq_completion)bond46#2 (work_completion)(&(&bond->mcast_work)->work) &p->pi_lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)bond179#2 irq_context: 0 (wq_completion)bond179#2 &rq->__lock irq_context: 0 (wq_completion)bond179#2 &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)bond179#2 (work_completion)(&(&slave->notify_work)->work) irq_context: 0 (wq_completion)bond179#2 (work_completion)(&(&slave->notify_work)->work) &obj_hash[i].lock irq_context: 0 (wq_completion)bond179#2 (work_completion)(&(&slave->notify_work)->work) &base->lock irq_context: 0 (wq_completion)bond179#2 (work_completion)(&(&slave->notify_work)->work) &base->lock &obj_hash[i].lock irq_context: 0 (wq_completion)bond45#2 &rq->__lock irq_context: 0 (wq_completion)bond45#2 &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)bond179#2 (work_completion)(&(&bond->mcast_work)->work) irq_context: 0 (wq_completion)bond179#2 (work_completion)(&(&bond->mcast_work)->work) &obj_hash[i].lock irq_context: 0 (wq_completion)bond179#2 (work_completion)(&(&bond->mcast_work)->work) &base->lock irq_context: 0 (wq_completion)bond179#2 (work_completion)(&(&bond->mcast_work)->work) &base->lock &obj_hash[i].lock irq_context: 0 (wq_completion)bond179#2 (work_completion)(&(&bond->mcast_work)->work) &rq->__lock irq_context: 0 (wq_completion)bond81#2 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex fs_reclaim irq_context: 0 (wq_completion)bond81#2 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex fs_reclaim mmu_notifier_invalidate_range_start irq_context: 0 (wq_completion)bond81#2 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex rcu_read_lock &bond->stats_lock/1 irq_context: 0 (wq_completion)bond81#2 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex rcu_read_lock &ndev->lock irq_context: 0 (wq_completion)bond81#2 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex nl_table_lock irq_context: 0 (wq_completion)bond81#2 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &obj_hash[i].lock irq_context: 0 (wq_completion)bond81#2 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex nl_table_wait.lock irq_context: 0 (wq_completion)bond81#2 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex net_rwsem irq_context: 0 (wq_completion)bond81#2 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex net_rwsem &list->lock#2 irq_context: 0 (wq_completion)bond81#2 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &tn->lock irq_context: 0 (wq_completion)bond81#2 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &idev->mc_lock irq_context: 0 (wq_completion)bond81#2 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &idev->mc_lock fs_reclaim irq_context: 0 (wq_completion)bond81#2 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &idev->mc_lock fs_reclaim mmu_notifier_invalidate_range_start irq_context: 0 (wq_completion)bond81#2 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &idev->mc_lock &c->lock irq_context: 0 (wq_completion)bond81#2 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &idev->mc_lock rcu_read_lock &ndev->lock irq_context: 0 (wq_completion)bond81#2 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &idev->mc_lock rcu_read_lock &c->lock irq_context: 0 (wq_completion)bond81#2 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &idev->mc_lock rcu_read_lock &dir->lock#2 irq_context: 0 (wq_completion)bond81#2 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &idev->mc_lock rcu_read_lock &ul->lock#2 irq_context: 0 (wq_completion)bond81#2 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &idev->mc_lock rcu_read_lock rcu_read_lock &obj_hash[i].lock irq_context: 0 (wq_completion)bond81#2 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &idev->mc_lock rcu_read_lock rcu_read_lock &c->lock irq_context: 0 (wq_completion)bond81#2 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &idev->mc_lock rcu_read_lock rcu_read_lock &dir->lock#2 irq_context: 0 (wq_completion)bond81#2 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &idev->mc_lock rcu_read_lock rcu_read_lock &tbl->lock irq_context: 0 (wq_completion)bond81#2 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &idev->mc_lock rcu_read_lock rcu_read_lock &n->lock irq_context: 0 (wq_completion)bond81#2 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &idev->mc_lock rcu_read_lock rcu_read_lock &____s->seqcount#9 irq_context: 0 (wq_completion)bond81#2 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &idev->mc_lock rcu_read_lock rcu_read_lock rcu_read_lock_bh rcu_read_lock rcu_read_lock_bh rcu_read_lock &br->multicast_lock irq_context: 0 (wq_completion)bond81#2 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &idev->mc_lock rcu_read_lock rcu_read_lock rcu_read_lock_bh rcu_read_lock rcu_read_lock_bh rcu_read_lock &br->multicast_lock &obj_hash[i].lock irq_context: 0 (wq_completion)bond81#2 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &idev->mc_lock rcu_read_lock rcu_read_lock rcu_read_lock_bh rcu_read_lock rcu_read_lock_bh rcu_read_lock &br->multicast_lock &base->lock irq_context: 0 (wq_completion)bond81#2 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &idev->mc_lock rcu_read_lock rcu_read_lock rcu_read_lock_bh rcu_read_lock rcu_read_lock_bh rcu_read_lock &br->multicast_lock &base->lock &obj_hash[i].lock irq_context: 0 (wq_completion)bond81#2 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &idev->mc_lock rcu_read_lock rcu_read_lock rcu_read_lock_bh rcu_read_lock rcu_read_lock_bh rcu_read_lock &data->lock irq_context: 0 (wq_completion)bond81#2 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &idev->mc_lock rcu_read_lock rcu_read_lock rcu_read_lock_bh rcu_read_lock rcu_read_lock_bh rcu_read_lock &obj_hash[i].lock irq_context: 0 (wq_completion)bond81#2 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex.wait_lock irq_context: 0 (wq_completion)bond81#2 (work_completion)(&(&bond->mcast_work)->work) &p->pi_lock irq_context: 0 (wq_completion)bond163#3 irq_context: 0 (wq_completion)bond163#3 (work_completion)(&(&slave->notify_work)->work) irq_context: 0 (wq_completion)bond163#3 (work_completion)(&(&slave->notify_work)->work) &obj_hash[i].lock irq_context: 0 (wq_completion)bond163#3 (work_completion)(&(&slave->notify_work)->work) &base->lock irq_context: 0 (wq_completion)bond163#3 (work_completion)(&(&slave->notify_work)->work) &base->lock &obj_hash[i].lock irq_context: 0 (wq_completion)bond163#3 (work_completion)(&(&slave->notify_work)->work) &rq->__lock irq_context: 0 (wq_completion)bond163#3 (work_completion)(&(&slave->notify_work)->work) &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)bond163#3 (work_completion)(&(&bond->mcast_work)->work) irq_context: 0 (wq_completion)bond163#3 (work_completion)(&(&bond->mcast_work)->work) &obj_hash[i].lock irq_context: 0 (wq_completion)bond163#3 (work_completion)(&(&bond->mcast_work)->work) &base->lock irq_context: 0 (wq_completion)bond163#3 (work_completion)(&(&bond->mcast_work)->work) &base->lock &obj_hash[i].lock irq_context: 0 (wq_completion)bond163#3 (work_completion)(&(&bond->mcast_work)->work) &rq->__lock irq_context: 0 (wq_completion)bond163#3 (work_completion)(&(&bond->mcast_work)->work) &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)bond157#2 (work_completion)(&(&slave->notify_work)->work) fill_pool_map-wait-type-override pool_lock irq_context: 0 (wq_completion)bond191 (work_completion)(&(&bond->mcast_work)->work) fill_pool_map-wait-type-override pool_lock irq_context: 0 (wq_completion)bond192 irq_context: 0 (wq_completion)bond192 (work_completion)(&(&slave->notify_work)->work) irq_context: 0 (wq_completion)bond192 (work_completion)(&(&slave->notify_work)->work) &obj_hash[i].lock irq_context: 0 (wq_completion)bond192 (work_completion)(&(&slave->notify_work)->work) &base->lock irq_context: 0 (wq_completion)bond192 (work_completion)(&(&slave->notify_work)->work) &base->lock &obj_hash[i].lock irq_context: 0 (wq_completion)bond192 (work_completion)(&(&bond->mcast_work)->work) irq_context: 0 (wq_completion)bond192 (work_completion)(&(&bond->mcast_work)->work) &obj_hash[i].lock irq_context: 0 (wq_completion)bond192 (work_completion)(&(&bond->mcast_work)->work) &base->lock irq_context: 0 (wq_completion)bond192 (work_completion)(&(&bond->mcast_work)->work) &base->lock &obj_hash[i].lock irq_context: 0 (wq_completion)bond192 (work_completion)(&(&bond->mcast_work)->work) &rq->__lock irq_context: 0 (wq_completion)bond192 (work_completion)(&(&bond->mcast_work)->work) &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)bond58#2 irq_context: 0 (wq_completion)bond58#2 (work_completion)(&(&slave->notify_work)->work) irq_context: 0 (wq_completion)bond58#2 (work_completion)(&(&slave->notify_work)->work) &rq->__lock irq_context: 0 (wq_completion)bond58#2 (work_completion)(&(&slave->notify_work)->work) &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)bond58#2 (work_completion)(&(&slave->notify_work)->work) &obj_hash[i].lock irq_context: 0 (wq_completion)bond58#2 (work_completion)(&(&slave->notify_work)->work) &base->lock irq_context: 0 (wq_completion)bond58#2 (work_completion)(&(&slave->notify_work)->work) &base->lock &obj_hash[i].lock irq_context: 0 (wq_completion)bond58#2 (work_completion)(&(&bond->mcast_work)->work) irq_context: 0 (wq_completion)bond58#2 (work_completion)(&(&bond->mcast_work)->work) &rq->__lock irq_context: 0 (wq_completion)bond58#2 (work_completion)(&(&bond->mcast_work)->work) &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)bond58#2 (work_completion)(&(&bond->mcast_work)->work) &obj_hash[i].lock irq_context: 0 (wq_completion)bond58#2 (work_completion)(&(&bond->mcast_work)->work) &base->lock irq_context: 0 (wq_completion)bond58#2 (work_completion)(&(&bond->mcast_work)->work) &base->lock &obj_hash[i].lock irq_context: 0 (wq_completion)bond179#2 (work_completion)(&(&slave->notify_work)->work) fill_pool_map-wait-type-override pool_lock irq_context: 0 (wq_completion)bond57#2 (work_completion)(&(&slave->notify_work)->work) rtnl_mutex net_rwsem irq_context: 0 (wq_completion)bond57#2 (work_completion)(&(&slave->notify_work)->work) rtnl_mutex net_rwsem &list->lock#2 irq_context: 0 (wq_completion)bond57#2 (work_completion)(&(&slave->notify_work)->work) rtnl_mutex &tn->lock irq_context: 0 (wq_completion)bond57#2 (work_completion)(&(&slave->notify_work)->work) rtnl_mutex.wait_lock irq_context: 0 (wq_completion)bond57#2 (work_completion)(&(&slave->notify_work)->work) &p->pi_lock irq_context: 0 (wq_completion)bond180#2 irq_context: 0 (wq_completion)bond180#2 (work_completion)(&(&slave->notify_work)->work) irq_context: 0 (wq_completion)bond180#2 (work_completion)(&(&slave->notify_work)->work) &obj_hash[i].lock irq_context: 0 (wq_completion)bond180#2 (work_completion)(&(&slave->notify_work)->work) &base->lock irq_context: 0 (wq_completion)bond180#2 (work_completion)(&(&slave->notify_work)->work) &base->lock &obj_hash[i].lock irq_context: 0 (wq_completion)bond180#2 (work_completion)(&(&slave->notify_work)->work) &rq->__lock irq_context: 0 (wq_completion)bond180#2 (work_completion)(&(&slave->notify_work)->work) &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)bond180#2 (work_completion)(&(&slave->notify_work)->work) &cfs_rq->removed.lock irq_context: 0 (wq_completion)bond180#2 (work_completion)(&(&bond->mcast_work)->work) irq_context: 0 (wq_completion)bond180#2 (work_completion)(&(&bond->mcast_work)->work) &obj_hash[i].lock irq_context: 0 (wq_completion)bond180#2 (work_completion)(&(&bond->mcast_work)->work) &base->lock irq_context: 0 (wq_completion)bond180#2 (work_completion)(&(&bond->mcast_work)->work) &base->lock &obj_hash[i].lock irq_context: 0 (wq_completion)bond180#2 (work_completion)(&(&bond->mcast_work)->work) &rq->__lock irq_context: 0 (wq_completion)bond180#2 (work_completion)(&(&bond->mcast_work)->work) &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)bond155#2 &rq->__lock irq_context: 0 (wq_completion)bond155#2 &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)bond164#3 irq_context: 0 (wq_completion)bond164#3 &rq->__lock irq_context: 0 (wq_completion)bond164#3 &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)bond164#3 (work_completion)(&(&slave->notify_work)->work) irq_context: 0 (wq_completion)bond164#3 (work_completion)(&(&slave->notify_work)->work) &obj_hash[i].lock irq_context: 0 (wq_completion)bond164#3 (work_completion)(&(&slave->notify_work)->work) &base->lock irq_context: 0 (wq_completion)bond164#3 (work_completion)(&(&slave->notify_work)->work) &base->lock &obj_hash[i].lock irq_context: 0 (wq_completion)bond164#3 (work_completion)(&(&bond->mcast_work)->work) irq_context: 0 (wq_completion)bond164#3 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex fs_reclaim irq_context: 0 (wq_completion)bond164#3 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex fs_reclaim mmu_notifier_invalidate_range_start irq_context: 0 (wq_completion)bond164#3 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &c->lock irq_context: 0 (wq_completion)bond164#3 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex rcu_read_lock &bond->stats_lock/1 irq_context: 0 (wq_completion)bond164#3 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex rcu_read_lock &ndev->lock irq_context: 0 (wq_completion)bond164#3 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex nl_table_lock irq_context: 0 (wq_completion)bond164#3 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &obj_hash[i].lock irq_context: 0 (wq_completion)bond164#3 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex nl_table_wait.lock irq_context: 0 (wq_completion)bond164#3 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex net_rwsem irq_context: 0 (wq_completion)bond164#3 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex net_rwsem &list->lock#2 irq_context: 0 (wq_completion)bond164#3 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &tn->lock irq_context: 0 (wq_completion)bond164#3 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &idev->mc_lock irq_context: 0 (wq_completion)bond164#3 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex.wait_lock irq_context: 0 (wq_completion)bond140 (work_completion)(&(&slave->notify_work)->work) rtnl_mutex net_rwsem irq_context: 0 (wq_completion)bond164#3 (work_completion)(&(&bond->mcast_work)->work) &p->pi_lock irq_context: 0 (wq_completion)bond140 (work_completion)(&(&slave->notify_work)->work) rtnl_mutex net_rwsem &list->lock#2 irq_context: 0 (wq_completion)bond140 (work_completion)(&(&slave->notify_work)->work) rtnl_mutex &tn->lock irq_context: 0 (wq_completion)bond140 (work_completion)(&(&slave->notify_work)->work) rtnl_mutex.wait_lock irq_context: 0 (wq_completion)bond162#2 (work_completion)(&(&slave->notify_work)->work) rtnl_mutex net_rwsem irq_context: 0 (wq_completion)bond140 (work_completion)(&(&slave->notify_work)->work) &p->pi_lock irq_context: 0 (wq_completion)bond162#2 (work_completion)(&(&slave->notify_work)->work) rtnl_mutex net_rwsem &list->lock#2 irq_context: 0 (wq_completion)bond162#2 (work_completion)(&(&slave->notify_work)->work) rtnl_mutex &tn->lock irq_context: 0 (wq_completion)bond162#2 (work_completion)(&(&slave->notify_work)->work) rtnl_mutex.wait_lock irq_context: 0 (wq_completion)bond162#2 (work_completion)(&(&slave->notify_work)->work) &p->pi_lock irq_context: 0 (wq_completion)bond193 irq_context: 0 (wq_completion)bond193 (work_completion)(&(&slave->notify_work)->work) irq_context: 0 (wq_completion)bond193 (work_completion)(&(&slave->notify_work)->work) &obj_hash[i].lock irq_context: 0 (wq_completion)bond193 (work_completion)(&(&slave->notify_work)->work) &base->lock irq_context: 0 (wq_completion)bond193 (work_completion)(&(&slave->notify_work)->work) &base->lock &obj_hash[i].lock irq_context: 0 (wq_completion)bond193 (work_completion)(&(&slave->notify_work)->work) &rq->__lock irq_context: 0 (wq_completion)bond193 (work_completion)(&(&slave->notify_work)->work) &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)bond189 (work_completion)(&(&slave->notify_work)->work) &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)bond172#2 (work_completion)(&(&slave->notify_work)->work) rtnl_mutex net_rwsem irq_context: 0 (wq_completion)bond172#2 (work_completion)(&(&slave->notify_work)->work) rtnl_mutex net_rwsem &list->lock#2 irq_context: 0 (wq_completion)bond172#2 (work_completion)(&(&slave->notify_work)->work) rtnl_mutex &tn->lock irq_context: 0 (wq_completion)bond172#2 (work_completion)(&(&slave->notify_work)->work) rtnl_mutex.wait_lock irq_context: 0 (wq_completion)bond172#2 (work_completion)(&(&slave->notify_work)->work) &p->pi_lock irq_context: 0 (wq_completion)bond135#2 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex fs_reclaim irq_context: 0 (wq_completion)bond135#2 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex fs_reclaim mmu_notifier_invalidate_range_start irq_context: 0 (wq_completion)bond135#2 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &c->lock irq_context: 0 (wq_completion)bond135#2 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex rcu_read_lock &bond->stats_lock/1 irq_context: 0 (wq_completion)bond135#2 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex rcu_read_lock &ndev->lock irq_context: 0 (wq_completion)bond135#2 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex nl_table_lock irq_context: 0 (wq_completion)bond135#2 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &obj_hash[i].lock irq_context: 0 (wq_completion)bond135#2 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex nl_table_wait.lock irq_context: 0 (wq_completion)bond135#2 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex net_rwsem irq_context: 0 (wq_completion)bond135#2 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex net_rwsem &list->lock#2 irq_context: 0 (wq_completion)bond135#2 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &tn->lock irq_context: 0 (wq_completion)bond135#2 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &idev->mc_lock irq_context: 0 (wq_completion)bond135#2 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &idev->mc_lock &rq->__lock irq_context: 0 (wq_completion)bond135#2 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &idev->mc_lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)bond135#2 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &idev->mc_lock fs_reclaim irq_context: 0 (wq_completion)bond135#2 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &idev->mc_lock fs_reclaim mmu_notifier_invalidate_range_start irq_context: 0 (wq_completion)bond135#2 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &idev->mc_lock &c->lock irq_context: 0 (wq_completion)bond135#2 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &idev->mc_lock rcu_read_lock &ndev->lock irq_context: 0 (wq_completion)bond135#2 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &idev->mc_lock rcu_read_lock &c->lock irq_context: 0 (wq_completion)bond135#2 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &idev->mc_lock rcu_read_lock &dir->lock#2 irq_context: 0 (wq_completion)bond135#2 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &idev->mc_lock rcu_read_lock &ul->lock#2 irq_context: 0 (wq_completion)bond135#2 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &idev->mc_lock rcu_read_lock rcu_read_lock &obj_hash[i].lock irq_context: 0 (wq_completion)bond135#2 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &idev->mc_lock rcu_read_lock rcu_read_lock rcu_node_0 irq_context: 0 (wq_completion)bond135#2 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &idev->mc_lock rcu_read_lock rcu_read_lock &rq->__lock irq_context: 0 (wq_completion)bond135#2 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &idev->mc_lock rcu_read_lock rcu_read_lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)bond135#2 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &idev->mc_lock rcu_read_lock rcu_read_lock &c->lock irq_context: 0 (wq_completion)bond135#2 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &idev->mc_lock rcu_read_lock rcu_read_lock &n->list_lock irq_context: 0 (wq_completion)bond135#2 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &idev->mc_lock rcu_read_lock rcu_read_lock &n->list_lock &c->lock irq_context: 0 (wq_completion)bond135#2 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &idev->mc_lock rcu_read_lock rcu_read_lock &dir->lock#2 irq_context: 0 (wq_completion)bond135#2 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &idev->mc_lock rcu_read_lock rcu_read_lock &tbl->lock irq_context: 0 (wq_completion)bond135#2 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &idev->mc_lock rcu_read_lock rcu_read_lock &n->lock irq_context: 0 (wq_completion)bond135#2 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &idev->mc_lock rcu_read_lock rcu_read_lock &____s->seqcount#9 irq_context: 0 (wq_completion)bond135#2 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &idev->mc_lock rcu_read_lock rcu_read_lock rcu_read_lock_bh rcu_read_lock rcu_read_lock_bh rcu_read_lock &br->multicast_lock irq_context: 0 (wq_completion)bond135#2 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &idev->mc_lock rcu_read_lock rcu_read_lock rcu_read_lock_bh rcu_read_lock rcu_read_lock_bh rcu_read_lock &br->multicast_lock &obj_hash[i].lock irq_context: 0 (wq_completion)bond135#2 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &idev->mc_lock rcu_read_lock rcu_read_lock rcu_read_lock_bh rcu_read_lock rcu_read_lock_bh rcu_read_lock &br->multicast_lock &base->lock irq_context: 0 (wq_completion)bond135#2 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &idev->mc_lock rcu_read_lock rcu_read_lock rcu_read_lock_bh rcu_read_lock rcu_read_lock_bh rcu_read_lock &br->multicast_lock &base->lock &obj_hash[i].lock irq_context: 0 (wq_completion)bond135#2 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &idev->mc_lock rcu_read_lock rcu_read_lock rcu_read_lock_bh rcu_read_lock rcu_read_lock_bh rcu_read_lock &data->lock irq_context: 0 (wq_completion)bond135#2 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &idev->mc_lock rcu_read_lock rcu_read_lock rcu_read_lock_bh rcu_read_lock rcu_read_lock_bh rcu_read_lock &obj_hash[i].lock irq_context: 0 (wq_completion)bond193 (work_completion)(&(&bond->mcast_work)->work) irq_context: 0 (wq_completion)bond193 (work_completion)(&(&bond->mcast_work)->work) &obj_hash[i].lock irq_context: 0 (wq_completion)bond193 (work_completion)(&(&bond->mcast_work)->work) &base->lock irq_context: 0 (wq_completion)bond193 (work_completion)(&(&bond->mcast_work)->work) &base->lock &obj_hash[i].lock irq_context: 0 (wq_completion)bond135#2 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &idev->mc_lock rcu_read_lock rcu_node_0 irq_context: 0 (wq_completion)bond135#2 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &idev->mc_lock rcu_read_lock &rq->__lock irq_context: 0 (wq_completion)bond135#2 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &rq->__lock irq_context: 0 (wq_completion)bond135#2 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)bond135#2 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex rcu_read_lock rcu_node_0 irq_context: 0 (wq_completion)bond135#2 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex rcu_read_lock &rq->__lock irq_context: 0 (wq_completion)bond135#2 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex.wait_lock irq_context: 0 (wq_completion)bond135#2 (work_completion)(&(&bond->mcast_work)->work) &p->pi_lock irq_context: 0 (wq_completion)bond135#2 (work_completion)(&(&bond->mcast_work)->work) &p->pi_lock &rq->__lock irq_context: 0 (wq_completion)bond135#2 (work_completion)(&(&bond->mcast_work)->work) &p->pi_lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 sk_lock-AF_CAN batched_entropy_u8.lock irq_context: 0 sk_lock-AF_CAN kfence_freelist_lock irq_context: 0 &sb->s_type->i_mutex_key#10 sk_lock-AF_CAN &meta->lock irq_context: 0 &sb->s_type->i_mutex_key#10 sk_lock-AF_CAN kfence_freelist_lock irq_context: 0 (wq_completion)bond157#3 (work_completion)(&(&bond->mcast_work)->work) &rq->__lock irq_context: 0 (wq_completion)bond157#3 (work_completion)(&(&bond->mcast_work)->work) &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)bond158#3 (work_completion)(&(&slave->notify_work)->work) rtnl_mutex net_rwsem irq_context: 0 (wq_completion)bond158#3 (work_completion)(&(&slave->notify_work)->work) rtnl_mutex net_rwsem &list->lock#2 irq_context: 0 (wq_completion)bond158#3 (work_completion)(&(&slave->notify_work)->work) rtnl_mutex &tn->lock irq_context: 0 (wq_completion)bond158#3 (work_completion)(&(&slave->notify_work)->work) rtnl_mutex.wait_lock irq_context: 0 (wq_completion)bond158#3 (work_completion)(&(&slave->notify_work)->work) &p->pi_lock irq_context: 0 (wq_completion)bond132 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex fs_reclaim irq_context: 0 (wq_completion)bond132 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex fs_reclaim mmu_notifier_invalidate_range_start irq_context: 0 (wq_completion)bond132 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &c->lock irq_context: 0 (wq_completion)bond132 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex rcu_read_lock &bond->stats_lock/1 irq_context: 0 (wq_completion)bond132 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex rcu_read_lock &ndev->lock irq_context: 0 (wq_completion)bond132 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex nl_table_lock irq_context: 0 (wq_completion)bond132 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &obj_hash[i].lock irq_context: 0 (wq_completion)bond132 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex nl_table_wait.lock irq_context: 0 (wq_completion)bond132 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex net_rwsem irq_context: 0 (wq_completion)bond132 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex net_rwsem &list->lock#2 irq_context: 0 (wq_completion)bond132 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &tn->lock irq_context: 0 (wq_completion)bond132 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &idev->mc_lock irq_context: 0 (wq_completion)bond132 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &idev->mc_lock fs_reclaim irq_context: 0 (wq_completion)bond132 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &idev->mc_lock fs_reclaim mmu_notifier_invalidate_range_start irq_context: 0 (wq_completion)bond132 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &idev->mc_lock &c->lock irq_context: 0 (wq_completion)bond132 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &idev->mc_lock &n->list_lock irq_context: 0 (wq_completion)bond132 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &idev->mc_lock &n->list_lock &c->lock irq_context: 0 (wq_completion)bond132 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &idev->mc_lock rcu_read_lock &ndev->lock irq_context: 0 (wq_completion)bond132 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &idev->mc_lock rcu_read_lock &c->lock irq_context: 0 (wq_completion)bond132 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &idev->mc_lock rcu_read_lock &n->list_lock irq_context: 0 (wq_completion)bond132 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &idev->mc_lock rcu_read_lock &n->list_lock &c->lock irq_context: 0 (wq_completion)bond132 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &idev->mc_lock rcu_read_lock &dir->lock#2 irq_context: 0 (wq_completion)bond132 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &idev->mc_lock rcu_read_lock &ul->lock#2 irq_context: 0 (wq_completion)bond132 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &idev->mc_lock rcu_read_lock rcu_read_lock &obj_hash[i].lock irq_context: 0 (wq_completion)bond132 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &idev->mc_lock rcu_read_lock rcu_read_lock &c->lock irq_context: 0 (wq_completion)bond132 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &idev->mc_lock rcu_read_lock rcu_read_lock &dir->lock#2 irq_context: 0 (wq_completion)bond132 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &idev->mc_lock rcu_read_lock rcu_read_lock &tbl->lock irq_context: 0 (wq_completion)bond132 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &idev->mc_lock rcu_read_lock rcu_read_lock &n->lock irq_context: 0 (wq_completion)bond132 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &idev->mc_lock rcu_read_lock rcu_read_lock &____s->seqcount#9 irq_context: 0 (wq_completion)bond132 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &idev->mc_lock rcu_read_lock rcu_read_lock rcu_read_lock_bh rcu_read_lock rcu_read_lock_bh rcu_read_lock &br->multicast_lock irq_context: 0 (wq_completion)bond132 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &idev->mc_lock rcu_read_lock rcu_read_lock rcu_read_lock_bh rcu_read_lock rcu_read_lock_bh rcu_read_lock &br->multicast_lock &obj_hash[i].lock irq_context: 0 (wq_completion)bond132 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &idev->mc_lock rcu_read_lock rcu_read_lock rcu_read_lock_bh rcu_read_lock rcu_read_lock_bh rcu_read_lock &br->multicast_lock &base->lock irq_context: 0 (wq_completion)bond132 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &idev->mc_lock rcu_read_lock rcu_read_lock rcu_read_lock_bh rcu_read_lock rcu_read_lock_bh rcu_read_lock &br->multicast_lock &base->lock &obj_hash[i].lock irq_context: 0 (wq_completion)bond132 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &idev->mc_lock rcu_read_lock rcu_read_lock rcu_read_lock_bh rcu_read_lock rcu_read_lock_bh rcu_read_lock &data->lock irq_context: 0 (wq_completion)bond132 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &idev->mc_lock rcu_read_lock rcu_read_lock rcu_read_lock_bh rcu_read_lock rcu_read_lock_bh rcu_read_lock &obj_hash[i].lock irq_context: 0 (wq_completion)bond132 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex.wait_lock irq_context: 0 (wq_completion)bond132 (work_completion)(&(&bond->mcast_work)->work) &p->pi_lock irq_context: 0 (wq_completion)bond181#2 irq_context: 0 (wq_completion)bond181#2 (work_completion)(&(&slave->notify_work)->work) irq_context: 0 (wq_completion)bond181#2 (work_completion)(&(&slave->notify_work)->work) &obj_hash[i].lock irq_context: 0 (wq_completion)bond181#2 (work_completion)(&(&slave->notify_work)->work) &base->lock irq_context: 0 (wq_completion)bond181#2 (work_completion)(&(&slave->notify_work)->work) &base->lock &obj_hash[i].lock irq_context: 0 (wq_completion)bond181#2 (work_completion)(&(&slave->notify_work)->work) &rq->__lock irq_context: 0 (wq_completion)bond181#2 (work_completion)(&(&slave->notify_work)->work) &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)bond173 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex fs_reclaim irq_context: 0 (wq_completion)bond173 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex fs_reclaim mmu_notifier_invalidate_range_start irq_context: 0 (wq_completion)bond173 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex rcu_read_lock &bond->stats_lock/1 irq_context: 0 (wq_completion)bond173 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex rcu_read_lock &ndev->lock irq_context: 0 (wq_completion)bond173 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex nl_table_lock irq_context: 0 (wq_completion)bond173 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &obj_hash[i].lock irq_context: 0 (wq_completion)bond173 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex nl_table_wait.lock irq_context: 0 (wq_completion)bond173 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex net_rwsem irq_context: 0 (wq_completion)bond173 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex net_rwsem &list->lock#2 irq_context: 0 (wq_completion)bond173 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &tn->lock irq_context: 0 (wq_completion)bond173 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &idev->mc_lock irq_context: 0 (wq_completion)bond173 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &idev->mc_lock fs_reclaim irq_context: 0 (wq_completion)bond173 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &idev->mc_lock fs_reclaim mmu_notifier_invalidate_range_start irq_context: 0 (wq_completion)bond173 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &idev->mc_lock rcu_read_lock &ndev->lock irq_context: 0 (wq_completion)bond173 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &idev->mc_lock rcu_read_lock &c->lock irq_context: 0 (wq_completion)bond173 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &idev->mc_lock rcu_read_lock &dir->lock#2 irq_context: 0 (wq_completion)bond173 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &idev->mc_lock rcu_read_lock &ul->lock#2 irq_context: 0 (wq_completion)bond173 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &idev->mc_lock rcu_read_lock rcu_read_lock &____s->seqcount#10 irq_context: 0 (wq_completion)bond173 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &idev->mc_lock rcu_read_lock rcu_read_lock rcu_read_lock_bh rcu_read_lock rcu_read_lock_bh rcu_read_lock &br->multicast_lock irq_context: 0 (wq_completion)bond173 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &idev->mc_lock rcu_read_lock rcu_read_lock rcu_read_lock_bh rcu_read_lock rcu_read_lock_bh rcu_read_lock &br->multicast_lock &obj_hash[i].lock irq_context: 0 (wq_completion)bond173 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &idev->mc_lock rcu_read_lock rcu_read_lock rcu_read_lock_bh rcu_read_lock rcu_read_lock_bh rcu_read_lock &br->multicast_lock &base->lock irq_context: 0 (wq_completion)bond173 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &idev->mc_lock rcu_read_lock rcu_read_lock rcu_read_lock_bh rcu_read_lock rcu_read_lock_bh rcu_read_lock &br->multicast_lock &base->lock &obj_hash[i].lock irq_context: 0 (wq_completion)bond173 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &idev->mc_lock rcu_read_lock rcu_read_lock rcu_read_lock_bh rcu_read_lock rcu_read_lock_bh rcu_read_lock &data->lock irq_context: 0 (wq_completion)bond173 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &idev->mc_lock rcu_read_lock rcu_read_lock rcu_read_lock_bh rcu_read_lock rcu_read_lock_bh rcu_read_lock &obj_hash[i].lock irq_context: 0 (wq_completion)bond173 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex.wait_lock irq_context: 0 (wq_completion)bond173 (work_completion)(&(&bond->mcast_work)->work) &p->pi_lock irq_context: 0 (wq_completion)bond179 (work_completion)(&(&slave->notify_work)->work) rtnl_mutex net_rwsem irq_context: 0 (wq_completion)bond179 (work_completion)(&(&slave->notify_work)->work) rtnl_mutex net_rwsem &list->lock#2 irq_context: 0 (wq_completion)bond179 (work_completion)(&(&slave->notify_work)->work) rtnl_mutex &tn->lock irq_context: 0 (wq_completion)bond179 (work_completion)(&(&slave->notify_work)->work) rtnl_mutex.wait_lock irq_context: 0 (wq_completion)bond179 (work_completion)(&(&slave->notify_work)->work) &p->pi_lock irq_context: 0 (wq_completion)bond139#3 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex fs_reclaim irq_context: 0 (wq_completion)bond139#3 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex fs_reclaim mmu_notifier_invalidate_range_start irq_context: 0 (wq_completion)bond139#3 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &c->lock irq_context: 0 (wq_completion)bond139#3 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &n->list_lock irq_context: 0 (wq_completion)bond139#3 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &n->list_lock &c->lock irq_context: 0 (wq_completion)bond139#3 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex rcu_read_lock &bond->stats_lock/1 irq_context: 0 (wq_completion)bond139#3 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex rcu_read_lock &ndev->lock irq_context: 0 (wq_completion)bond139#3 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex nl_table_lock irq_context: 0 (wq_completion)bond139#3 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &obj_hash[i].lock irq_context: 0 (wq_completion)bond139#3 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex nl_table_wait.lock irq_context: 0 (wq_completion)bond139#3 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex net_rwsem irq_context: 0 (wq_completion)bond139#3 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex net_rwsem &list->lock#2 irq_context: 0 (wq_completion)bond139#3 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &tn->lock irq_context: 0 (wq_completion)bond139#3 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &idev->mc_lock irq_context: 0 (wq_completion)bond139#3 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &idev->mc_lock fs_reclaim irq_context: 0 (wq_completion)bond139#3 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &idev->mc_lock fs_reclaim mmu_notifier_invalidate_range_start irq_context: 0 (wq_completion)bond139#3 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &idev->mc_lock &c->lock irq_context: 0 (wq_completion)bond139#3 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &idev->mc_lock &n->list_lock irq_context: 0 (wq_completion)bond139#3 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &idev->mc_lock &n->list_lock &c->lock irq_context: 0 (wq_completion)bond139#3 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &idev->mc_lock rcu_read_lock &ndev->lock irq_context: 0 (wq_completion)bond139#3 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &idev->mc_lock rcu_read_lock &dir->lock#2 irq_context: 0 (wq_completion)bond139#3 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &idev->mc_lock rcu_read_lock &ul->lock#2 irq_context: 0 (wq_completion)bond139#3 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &idev->mc_lock rcu_read_lock rcu_node_0 irq_context: 0 (wq_completion)bond139#3 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &idev->mc_lock rcu_read_lock &rq->__lock irq_context: 0 (wq_completion)bond139#3 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &idev->mc_lock rcu_read_lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)bond181#2 (work_completion)(&(&bond->mcast_work)->work) irq_context: 0 (wq_completion)bond181#2 (work_completion)(&(&bond->mcast_work)->work) &obj_hash[i].lock irq_context: 0 (wq_completion)bond181#2 (work_completion)(&(&bond->mcast_work)->work) &base->lock irq_context: 0 (wq_completion)bond181#2 (work_completion)(&(&bond->mcast_work)->work) &base->lock &obj_hash[i].lock irq_context: 0 (wq_completion)bond139#3 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &idev->mc_lock rcu_read_lock rcu_read_lock &____s->seqcount#10 irq_context: 0 (wq_completion)bond139#3 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &idev->mc_lock rcu_read_lock rcu_read_lock rcu_read_lock_bh rcu_read_lock rcu_read_lock_bh rcu_read_lock &br->multicast_lock irq_context: 0 (wq_completion)bond139#3 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &idev->mc_lock rcu_read_lock rcu_read_lock rcu_read_lock_bh rcu_read_lock rcu_read_lock_bh rcu_read_lock &br->multicast_lock &obj_hash[i].lock irq_context: 0 (wq_completion)bond139#3 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &idev->mc_lock rcu_read_lock rcu_read_lock rcu_read_lock_bh rcu_read_lock rcu_read_lock_bh rcu_read_lock &br->multicast_lock &base->lock irq_context: 0 (wq_completion)bond139#3 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &idev->mc_lock rcu_read_lock rcu_read_lock rcu_read_lock_bh rcu_read_lock rcu_read_lock_bh rcu_read_lock &br->multicast_lock &base->lock &obj_hash[i].lock irq_context: 0 (wq_completion)bond139#3 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &idev->mc_lock rcu_read_lock rcu_read_lock rcu_read_lock_bh rcu_read_lock rcu_read_lock_bh rcu_read_lock &data->lock irq_context: 0 (wq_completion)bond139#3 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &idev->mc_lock rcu_read_lock rcu_read_lock rcu_read_lock_bh rcu_read_lock rcu_read_lock_bh rcu_read_lock &obj_hash[i].lock irq_context: 0 (wq_completion)bond139#3 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &idev->mc_lock rcu_read_lock rcu_read_lock &rq->__lock irq_context: 0 (wq_completion)bond139#3 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &idev->mc_lock rcu_read_lock rcu_read_lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)bond139#3 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex.wait_lock irq_context: 0 (wq_completion)bond139#3 (work_completion)(&(&bond->mcast_work)->work) &p->pi_lock irq_context: 0 (wq_completion)bond139#3 (work_completion)(&(&bond->mcast_work)->work) &p->pi_lock &rq->__lock irq_context: 0 (wq_completion)bond139#3 (work_completion)(&(&bond->mcast_work)->work) &p->pi_lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)bond122#2 (work_completion)(&(&bond->mcast_work)->work) fill_pool_map-wait-type-override pool_lock irq_context: 0 (wq_completion)bond165#3 irq_context: 0 (wq_completion)bond165#3 (work_completion)(&(&slave->notify_work)->work) irq_context: 0 (wq_completion)bond165#3 (work_completion)(&(&slave->notify_work)->work) &obj_hash[i].lock irq_context: 0 (wq_completion)bond165#3 (work_completion)(&(&slave->notify_work)->work) &base->lock irq_context: 0 (wq_completion)bond165#3 (work_completion)(&(&slave->notify_work)->work) &base->lock &obj_hash[i].lock irq_context: 0 (wq_completion)bond160 &rq->__lock irq_context: 0 (wq_completion)bond160 &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)bond134#2 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex fs_reclaim irq_context: 0 (wq_completion)bond134#2 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex fs_reclaim mmu_notifier_invalidate_range_start irq_context: 0 (wq_completion)bond134#2 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex rcu_read_lock &bond->stats_lock/1 irq_context: 0 (wq_completion)bond134#2 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex rcu_read_lock &ndev->lock irq_context: 0 (wq_completion)bond134#2 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex nl_table_lock irq_context: 0 (wq_completion)bond134#2 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &obj_hash[i].lock irq_context: 0 (wq_completion)bond134#2 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex nl_table_wait.lock irq_context: 0 (wq_completion)bond134#2 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex net_rwsem irq_context: 0 (wq_completion)bond134#2 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex net_rwsem &list->lock#2 irq_context: 0 (wq_completion)bond134#2 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &tn->lock irq_context: 0 (wq_completion)bond134#2 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &idev->mc_lock irq_context: 0 (wq_completion)bond134#2 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &idev->mc_lock fs_reclaim irq_context: 0 (wq_completion)bond134#2 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &idev->mc_lock fs_reclaim mmu_notifier_invalidate_range_start irq_context: 0 (wq_completion)bond134#2 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &idev->mc_lock rcu_read_lock &ndev->lock irq_context: 0 (wq_completion)bond134#2 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &idev->mc_lock rcu_read_lock &c->lock irq_context: 0 (wq_completion)bond134#2 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &idev->mc_lock rcu_read_lock &dir->lock#2 irq_context: 0 (wq_completion)bond134#2 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &idev->mc_lock rcu_read_lock &ul->lock#2 irq_context: 0 (wq_completion)bond134#2 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &idev->mc_lock rcu_read_lock rcu_read_lock &obj_hash[i].lock irq_context: 0 (wq_completion)bond134#2 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &idev->mc_lock rcu_read_lock rcu_read_lock &dir->lock#2 irq_context: 0 (wq_completion)bond134#2 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &idev->mc_lock rcu_read_lock rcu_read_lock &tbl->lock irq_context: 0 (wq_completion)bond134#2 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &idev->mc_lock rcu_read_lock rcu_read_lock &n->lock irq_context: 0 (wq_completion)bond134#2 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &idev->mc_lock rcu_read_lock rcu_read_lock &____s->seqcount#9 irq_context: 0 (wq_completion)bond134#2 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &idev->mc_lock rcu_read_lock rcu_read_lock rcu_read_lock_bh rcu_read_lock rcu_read_lock_bh rcu_read_lock &br->multicast_lock irq_context: 0 (wq_completion)bond134#2 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &idev->mc_lock rcu_read_lock rcu_read_lock rcu_read_lock_bh rcu_read_lock rcu_read_lock_bh rcu_read_lock &br->multicast_lock &obj_hash[i].lock irq_context: 0 (wq_completion)bond134#2 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &idev->mc_lock rcu_read_lock rcu_read_lock rcu_read_lock_bh rcu_read_lock rcu_read_lock_bh rcu_read_lock &br->multicast_lock &base->lock irq_context: 0 (wq_completion)bond134#2 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &idev->mc_lock rcu_read_lock rcu_read_lock rcu_read_lock_bh rcu_read_lock rcu_read_lock_bh rcu_read_lock &br->multicast_lock &base->lock &obj_hash[i].lock irq_context: 0 (wq_completion)bond134#2 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &idev->mc_lock rcu_read_lock rcu_read_lock rcu_read_lock_bh rcu_read_lock rcu_read_lock_bh rcu_read_lock &data->lock irq_context: 0 (wq_completion)bond134#2 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &idev->mc_lock rcu_read_lock rcu_read_lock rcu_read_lock_bh rcu_read_lock rcu_read_lock_bh rcu_read_lock &obj_hash[i].lock irq_context: 0 (wq_completion)bond134#2 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex.wait_lock irq_context: 0 (wq_completion)bond134#2 (work_completion)(&(&bond->mcast_work)->work) &p->pi_lock irq_context: 0 (wq_completion)bond153#2 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex fs_reclaim irq_context: 0 (wq_completion)bond153#2 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex fs_reclaim &rq->__lock irq_context: 0 (wq_completion)bond153#2 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex fs_reclaim &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)bond153#2 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex fs_reclaim mmu_notifier_invalidate_range_start irq_context: 0 (wq_completion)bond153#2 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex rcu_read_lock &bond->stats_lock/1 irq_context: 0 (wq_completion)bond153#2 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex rcu_read_lock &ndev->lock irq_context: 0 (wq_completion)bond153#2 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex nl_table_lock irq_context: 0 (wq_completion)bond153#2 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &obj_hash[i].lock irq_context: 0 (wq_completion)bond153#2 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex nl_table_wait.lock irq_context: 0 (wq_completion)bond153#2 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex net_rwsem irq_context: 0 (wq_completion)bond153#2 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex net_rwsem &list->lock#2 irq_context: 0 (wq_completion)bond153#2 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &tn->lock irq_context: 0 (wq_completion)bond153#2 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &idev->mc_lock irq_context: 0 (wq_completion)bond153#2 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &idev->mc_lock &rq->__lock irq_context: 0 (wq_completion)bond153#2 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &idev->mc_lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)bond153#2 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &idev->mc_lock fs_reclaim irq_context: 0 (wq_completion)bond153#2 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &idev->mc_lock fs_reclaim mmu_notifier_invalidate_range_start irq_context: 0 (wq_completion)bond153#2 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &idev->mc_lock &c->lock irq_context: 0 (wq_completion)bond153#2 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &idev->mc_lock rcu_read_lock &ndev->lock irq_context: 0 (wq_completion)bond153#2 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &idev->mc_lock rcu_read_lock &dir->lock#2 irq_context: 0 (wq_completion)bond153#2 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &idev->mc_lock rcu_read_lock &ul->lock#2 irq_context: 0 (wq_completion)bond153#2 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &idev->mc_lock rcu_read_lock rcu_read_lock &____s->seqcount#10 irq_context: 0 (wq_completion)bond153#2 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &idev->mc_lock rcu_read_lock rcu_read_lock rcu_read_lock_bh rcu_read_lock rcu_read_lock_bh rcu_read_lock &br->multicast_lock irq_context: 0 (wq_completion)bond153#2 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &idev->mc_lock rcu_read_lock rcu_read_lock rcu_read_lock_bh rcu_read_lock rcu_read_lock_bh rcu_read_lock &br->multicast_lock &obj_hash[i].lock irq_context: 0 (wq_completion)bond153#2 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &idev->mc_lock rcu_read_lock rcu_read_lock rcu_read_lock_bh rcu_read_lock rcu_read_lock_bh rcu_read_lock &br->multicast_lock &base->lock irq_context: 0 (wq_completion)bond153#2 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &idev->mc_lock rcu_read_lock rcu_read_lock rcu_read_lock_bh rcu_read_lock rcu_read_lock_bh rcu_read_lock &br->multicast_lock &base->lock &obj_hash[i].lock irq_context: 0 (wq_completion)bond153#2 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &idev->mc_lock rcu_read_lock rcu_read_lock rcu_read_lock_bh rcu_read_lock rcu_read_lock_bh rcu_read_lock &data->lock irq_context: 0 (wq_completion)bond153#2 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &idev->mc_lock rcu_read_lock rcu_read_lock rcu_read_lock_bh rcu_read_lock rcu_read_lock_bh rcu_read_lock &obj_hash[i].lock irq_context: 0 (wq_completion)bond153#2 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &idev->mc_lock rcu_read_lock rcu_read_lock rcu_node_0 irq_context: 0 (wq_completion)bond153#2 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &idev->mc_lock rcu_read_lock rcu_read_lock &rq->__lock irq_context: 0 (wq_completion)bond153#2 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &idev->mc_lock rcu_read_lock rcu_read_lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)bond153#2 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &idev->mc_lock rcu_read_lock rcu_node_0 irq_context: 0 (wq_completion)bond153#2 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex.wait_lock irq_context: 0 (wq_completion)bond153#2 (work_completion)(&(&bond->mcast_work)->work) &p->pi_lock irq_context: 0 (wq_completion)bond153#2 (work_completion)(&(&bond->mcast_work)->work) &p->pi_lock &rq->__lock irq_context: 0 (wq_completion)bond153#2 (work_completion)(&(&bond->mcast_work)->work) &p->pi_lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)bond165#3 (work_completion)(&(&bond->mcast_work)->work) irq_context: 0 (wq_completion)bond165#3 (work_completion)(&(&bond->mcast_work)->work) &rq->__lock irq_context: 0 (wq_completion)bond165#3 (work_completion)(&(&bond->mcast_work)->work) &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)bond165#3 (work_completion)(&(&bond->mcast_work)->work) &obj_hash[i].lock irq_context: 0 (wq_completion)bond165#3 (work_completion)(&(&bond->mcast_work)->work) &base->lock irq_context: 0 (wq_completion)bond165#3 (work_completion)(&(&bond->mcast_work)->work) &base->lock &obj_hash[i].lock irq_context: 0 rtnl_mutex cpu_hotplug_lock wq_pool_mutex rcu_node_0 irq_context: 0 (wq_completion)bond194 irq_context: 0 (wq_completion)bond194 (work_completion)(&(&slave->notify_work)->work) irq_context: 0 (wq_completion)bond194 (work_completion)(&(&slave->notify_work)->work) &obj_hash[i].lock irq_context: 0 (wq_completion)bond194 (work_completion)(&(&slave->notify_work)->work) &base->lock irq_context: 0 (wq_completion)bond194 (work_completion)(&(&slave->notify_work)->work) &base->lock &obj_hash[i].lock irq_context: 0 (wq_completion)bond194 (work_completion)(&(&slave->notify_work)->work) &rq->__lock irq_context: 0 rcu_read_lock_bh dev->qdisc_tx_busylock ?: &qdisc_tx_busylock _xmit_NETROM console_owner_lock irq_context: 0 (wq_completion)bond194 (work_completion)(&(&bond->mcast_work)->work) irq_context: 0 (wq_completion)bond194 (work_completion)(&(&bond->mcast_work)->work) &obj_hash[i].lock irq_context: 0 (wq_completion)bond194 (work_completion)(&(&bond->mcast_work)->work) &base->lock irq_context: 0 (wq_completion)bond194 (work_completion)(&(&bond->mcast_work)->work) &base->lock &obj_hash[i].lock irq_context: 0 (wq_completion)bond194 (work_completion)(&(&bond->mcast_work)->work) &rq->__lock irq_context: 0 (wq_completion)bond194 (work_completion)(&(&bond->mcast_work)->work) &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)bond182#2 irq_context: 0 (wq_completion)bond182#2 (work_completion)(&(&slave->notify_work)->work) irq_context: 0 (wq_completion)bond182#2 (work_completion)(&(&slave->notify_work)->work) &rq->__lock irq_context: 0 (wq_completion)bond182#2 (work_completion)(&(&slave->notify_work)->work) &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)bond182#2 (work_completion)(&(&slave->notify_work)->work) &obj_hash[i].lock irq_context: 0 (wq_completion)bond182#2 (work_completion)(&(&slave->notify_work)->work) &base->lock irq_context: 0 (wq_completion)bond182#2 (work_completion)(&(&slave->notify_work)->work) &base->lock &obj_hash[i].lock irq_context: 0 (wq_completion)bond182#2 (work_completion)(&(&bond->mcast_work)->work) irq_context: 0 (wq_completion)bond182#2 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex fs_reclaim irq_context: 0 (wq_completion)bond182#2 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex fs_reclaim mmu_notifier_invalidate_range_start irq_context: 0 (wq_completion)bond182#2 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex rcu_read_lock &bond->stats_lock/1 irq_context: 0 (wq_completion)bond182#2 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex rcu_read_lock &ndev->lock irq_context: 0 (wq_completion)bond182#2 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex nl_table_lock irq_context: 0 (wq_completion)bond182#2 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &obj_hash[i].lock irq_context: 0 (wq_completion)bond182#2 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex nl_table_wait.lock irq_context: 0 (wq_completion)bond182#2 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex net_rwsem irq_context: 0 (wq_completion)bond182#2 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex net_rwsem &list->lock#2 irq_context: 0 (wq_completion)bond182#2 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &tn->lock irq_context: 0 (wq_completion)bond182#2 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &idev->mc_lock irq_context: 0 (wq_completion)bond182#2 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex.wait_lock irq_context: 0 (wq_completion)bond182#2 (work_completion)(&(&bond->mcast_work)->work) &p->pi_lock irq_context: 0 (wq_completion)bond165#3 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex fs_reclaim irq_context: 0 (wq_completion)bond165#3 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex fs_reclaim mmu_notifier_invalidate_range_start irq_context: 0 (wq_completion)bond165#3 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &c->lock irq_context: 0 (wq_completion)bond165#3 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex rcu_read_lock &bond->stats_lock/1 irq_context: 0 (wq_completion)bond165#3 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex rcu_read_lock &ndev->lock irq_context: 0 (wq_completion)bond165#3 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex nl_table_lock irq_context: 0 (wq_completion)bond165#3 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &obj_hash[i].lock irq_context: 0 (wq_completion)bond165#3 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex nl_table_wait.lock irq_context: 0 (wq_completion)bond165#3 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex net_rwsem irq_context: 0 (wq_completion)bond165#3 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex net_rwsem &list->lock#2 irq_context: 0 (wq_completion)bond165#3 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &tn->lock irq_context: 0 (wq_completion)bond165#3 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &idev->mc_lock irq_context: 0 (wq_completion)bond165#3 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex.wait_lock irq_context: 0 (wq_completion)bond165#3 (work_completion)(&(&bond->mcast_work)->work) &p->pi_lock irq_context: 0 (wq_completion)bond155 (work_completion)(&(&slave->notify_work)->work) rtnl_mutex net_rwsem irq_context: 0 (wq_completion)bond155 (work_completion)(&(&slave->notify_work)->work) rtnl_mutex net_rwsem &rq->__lock irq_context: 0 (wq_completion)bond155 (work_completion)(&(&slave->notify_work)->work) rtnl_mutex net_rwsem &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)bond155 (work_completion)(&(&slave->notify_work)->work) rtnl_mutex net_rwsem &list->lock#2 irq_context: 0 (wq_completion)bond155 (work_completion)(&(&slave->notify_work)->work) rtnl_mutex &tn->lock irq_context: 0 (wq_completion)bond155 (work_completion)(&(&slave->notify_work)->work) rtnl_mutex.wait_lock irq_context: 0 (wq_completion)bond155 (work_completion)(&(&slave->notify_work)->work) &p->pi_lock irq_context: 0 (wq_completion)bond166#3 irq_context: 0 (wq_completion)bond166#3 (work_completion)(&(&slave->notify_work)->work) irq_context: 0 (wq_completion)bond166#3 (work_completion)(&(&slave->notify_work)->work) &obj_hash[i].lock irq_context: 0 (wq_completion)bond166#3 (work_completion)(&(&slave->notify_work)->work) &base->lock irq_context: 0 (wq_completion)bond166#3 (work_completion)(&(&slave->notify_work)->work) &base->lock &obj_hash[i].lock irq_context: 0 (wq_completion)bond166#3 (work_completion)(&(&bond->mcast_work)->work) irq_context: 0 (wq_completion)bond166#3 (work_completion)(&(&bond->mcast_work)->work) &obj_hash[i].lock irq_context: 0 (wq_completion)bond166#3 (work_completion)(&(&bond->mcast_work)->work) &base->lock irq_context: 0 (wq_completion)bond166#3 (work_completion)(&(&bond->mcast_work)->work) &base->lock &obj_hash[i].lock irq_context: 0 (wq_completion)bond177 (work_completion)(&(&bond->mcast_work)->work) &rq->__lock irq_context: 0 (wq_completion)bond177 (work_completion)(&(&bond->mcast_work)->work) &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)bond167#2 &rq->__lock irq_context: 0 (wq_completion)bond167#2 &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)bond183#2 irq_context: 0 (wq_completion)bond183#2 &rq->__lock irq_context: 0 (wq_completion)bond183#2 &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)bond183#2 (work_completion)(&(&slave->notify_work)->work) irq_context: 0 (wq_completion)bond183#2 (work_completion)(&(&slave->notify_work)->work) &obj_hash[i].lock irq_context: 0 (wq_completion)bond183#2 (work_completion)(&(&slave->notify_work)->work) &base->lock irq_context: 0 (wq_completion)bond183#2 (work_completion)(&(&slave->notify_work)->work) &base->lock &obj_hash[i].lock irq_context: 0 (wq_completion)bond183#2 (work_completion)(&(&slave->notify_work)->work) &rq->__lock irq_context: 0 (wq_completion)bond183#2 (work_completion)(&(&bond->mcast_work)->work) irq_context: 0 (wq_completion)bond183#2 (work_completion)(&(&bond->mcast_work)->work) &obj_hash[i].lock irq_context: 0 (wq_completion)bond183#2 (work_completion)(&(&bond->mcast_work)->work) &base->lock irq_context: 0 (wq_completion)bond183#2 (work_completion)(&(&bond->mcast_work)->work) &base->lock &obj_hash[i].lock irq_context: 0 (wq_completion)bond183#2 (work_completion)(&(&bond->mcast_work)->work) &rq->__lock irq_context: 0 (wq_completion)bond183#2 (work_completion)(&(&bond->mcast_work)->work) &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)bond195 irq_context: 0 (wq_completion)bond195 (work_completion)(&(&slave->notify_work)->work) irq_context: 0 (wq_completion)bond195 (work_completion)(&(&slave->notify_work)->work) &obj_hash[i].lock irq_context: 0 (wq_completion)bond195 (work_completion)(&(&slave->notify_work)->work) &base->lock irq_context: 0 (wq_completion)bond195 (work_completion)(&(&slave->notify_work)->work) &base->lock &obj_hash[i].lock irq_context: 0 (wq_completion)bond195 (work_completion)(&(&slave->notify_work)->work) &rq->__lock irq_context: 0 (wq_completion)bond195 (work_completion)(&(&slave->notify_work)->work) &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)bond195 (work_completion)(&(&bond->mcast_work)->work) irq_context: 0 (wq_completion)bond195 (work_completion)(&(&bond->mcast_work)->work) &obj_hash[i].lock irq_context: 0 (wq_completion)bond195 (work_completion)(&(&bond->mcast_work)->work) &base->lock irq_context: 0 (wq_completion)bond195 (work_completion)(&(&bond->mcast_work)->work) &base->lock &obj_hash[i].lock irq_context: 0 (wq_completion)bond167#3 irq_context: 0 (wq_completion)bond167#3 (work_completion)(&(&slave->notify_work)->work) irq_context: 0 (wq_completion)bond167#3 (work_completion)(&(&slave->notify_work)->work) &obj_hash[i].lock irq_context: 0 (wq_completion)bond167#3 (work_completion)(&(&slave->notify_work)->work) &base->lock irq_context: 0 (wq_completion)bond167#3 (work_completion)(&(&slave->notify_work)->work) &base->lock &obj_hash[i].lock irq_context: 0 (wq_completion)bond167#3 (work_completion)(&(&bond->mcast_work)->work) irq_context: 0 (wq_completion)bond167#3 (work_completion)(&(&bond->mcast_work)->work) &obj_hash[i].lock irq_context: 0 (wq_completion)bond167#3 (work_completion)(&(&bond->mcast_work)->work) &base->lock irq_context: 0 (wq_completion)bond167#3 (work_completion)(&(&bond->mcast_work)->work) &base->lock &obj_hash[i].lock irq_context: 0 (wq_completion)bond137#3 (work_completion)(&(&slave->notify_work)->work) &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)bond185 &rq->__lock irq_context: 0 (wq_completion)bond185 &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)events (linkwatch_work).work rtnl_mutex rcu_read_lock &cfs_rq->removed.lock irq_context: 0 (wq_completion)bond45#2 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex fs_reclaim irq_context: 0 (wq_completion)bond45#2 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex fs_reclaim mmu_notifier_invalidate_range_start irq_context: 0 (wq_completion)bond45#2 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex rcu_read_lock &bond->stats_lock/1 irq_context: 0 (wq_completion)bond45#2 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex rcu_read_lock &ndev->lock irq_context: 0 (wq_completion)bond45#2 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex nl_table_lock irq_context: 0 (wq_completion)bond45#2 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &obj_hash[i].lock irq_context: 0 (wq_completion)bond45#2 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex quarantine_lock irq_context: 0 (wq_completion)bond45#2 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex nl_table_wait.lock irq_context: 0 (wq_completion)bond45#2 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex net_rwsem irq_context: 0 (wq_completion)bond45#2 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex net_rwsem &list->lock#2 irq_context: 0 (wq_completion)bond45#2 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &tn->lock irq_context: 0 (wq_completion)bond45#2 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &idev->mc_lock irq_context: 0 (wq_completion)bond45#2 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &idev->mc_lock fs_reclaim irq_context: 0 (wq_completion)bond45#2 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &idev->mc_lock fs_reclaim mmu_notifier_invalidate_range_start irq_context: 0 (wq_completion)bond45#2 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &idev->mc_lock remove_cache_srcu irq_context: 0 (wq_completion)bond45#2 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &idev->mc_lock remove_cache_srcu quarantine_lock irq_context: 0 (wq_completion)bond45#2 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &idev->mc_lock remove_cache_srcu &c->lock irq_context: 0 (wq_completion)bond45#2 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &idev->mc_lock remove_cache_srcu &n->list_lock irq_context: 0 (wq_completion)bond45#2 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &idev->mc_lock remove_cache_srcu &obj_hash[i].lock irq_context: 0 (wq_completion)bond45#2 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &idev->mc_lock remove_cache_srcu pool_lock#2 irq_context: 0 (wq_completion)bond45#2 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &idev->mc_lock &c->lock irq_context: 0 (wq_completion)bond45#2 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &idev->mc_lock rcu_read_lock &ndev->lock irq_context: 0 (wq_completion)bond45#2 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &idev->mc_lock rcu_read_lock &dir->lock#2 irq_context: 0 (wq_completion)bond45#2 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &idev->mc_lock rcu_read_lock &ul->lock#2 irq_context: 0 (wq_completion)bond45#2 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &idev->mc_lock rcu_read_lock rcu_read_lock &____s->seqcount#12 irq_context: 0 (wq_completion)bond45#2 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &idev->mc_lock rcu_read_lock rcu_read_lock &____s->seqcount#16 irq_context: 0 (wq_completion)bond45#2 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &idev->mc_lock rcu_read_lock rcu_read_lock &c->lock irq_context: 0 (wq_completion)bond45#2 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &idev->mc_lock rcu_read_lock &obj_hash[i].lock irq_context: 0 (wq_completion)bond45#2 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &idev->mc_lock rcu_read_lock &data->lock irq_context: 0 (wq_completion)bond45#2 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex.wait_lock irq_context: 0 (wq_completion)bond45#2 (work_completion)(&(&bond->mcast_work)->work) &p->pi_lock irq_context: 0 (wq_completion)ipv6_addrconf (work_completion)(&(&ifa->dad_work)->work) rtnl_mutex rcu_read_lock rcu_read_lock &rq->__lock &cfs_rq->removed.lock irq_context: 0 (wq_completion)bond196 irq_context: 0 (wq_completion)bond196 (work_completion)(&(&slave->notify_work)->work) irq_context: 0 (wq_completion)bond196 (work_completion)(&(&slave->notify_work)->work) &obj_hash[i].lock irq_context: 0 (wq_completion)bond196 (work_completion)(&(&slave->notify_work)->work) &base->lock irq_context: 0 (wq_completion)bond196 (work_completion)(&(&slave->notify_work)->work) &base->lock &obj_hash[i].lock irq_context: 0 (wq_completion)bond196 (work_completion)(&(&bond->mcast_work)->work) irq_context: 0 (wq_completion)bond196 (work_completion)(&(&bond->mcast_work)->work) &obj_hash[i].lock irq_context: 0 (wq_completion)bond196 (work_completion)(&(&bond->mcast_work)->work) &base->lock irq_context: 0 (wq_completion)bond196 (work_completion)(&(&bond->mcast_work)->work) &base->lock &obj_hash[i].lock irq_context: 0 (wq_completion)bond196 (work_completion)(&(&bond->mcast_work)->work) &rq->__lock irq_context: 0 (wq_completion)bond196 (work_completion)(&(&bond->mcast_work)->work) &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)bond184#2 irq_context: 0 (wq_completion)bond184#2 (work_completion)(&(&slave->notify_work)->work) irq_context: 0 (wq_completion)bond184#2 (work_completion)(&(&slave->notify_work)->work) &obj_hash[i].lock irq_context: 0 (wq_completion)bond184#2 (work_completion)(&(&slave->notify_work)->work) &base->lock irq_context: 0 (wq_completion)bond184#2 (work_completion)(&(&slave->notify_work)->work) &base->lock &obj_hash[i].lock irq_context: 0 (wq_completion)bond184#2 (work_completion)(&(&slave->notify_work)->work) &rq->__lock irq_context: 0 (wq_completion)bond184#2 (work_completion)(&(&slave->notify_work)->work) &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)bond184#2 (work_completion)(&(&bond->mcast_work)->work) irq_context: 0 (wq_completion)bond184#2 (work_completion)(&(&bond->mcast_work)->work) &obj_hash[i].lock irq_context: 0 (wq_completion)bond184#2 (work_completion)(&(&bond->mcast_work)->work) &base->lock irq_context: 0 (wq_completion)bond184#2 (work_completion)(&(&bond->mcast_work)->work) &base->lock &obj_hash[i].lock irq_context: 0 (wq_completion)bond168#3 irq_context: 0 (wq_completion)bond168#3 (work_completion)(&(&slave->notify_work)->work) irq_context: 0 (wq_completion)bond168#3 (work_completion)(&(&slave->notify_work)->work) &obj_hash[i].lock irq_context: 0 (wq_completion)bond168#3 (work_completion)(&(&slave->notify_work)->work) &base->lock irq_context: 0 (wq_completion)bond168#3 (work_completion)(&(&slave->notify_work)->work) &base->lock &obj_hash[i].lock irq_context: 0 (wq_completion)bond138 (work_completion)(&(&slave->notify_work)->work) rtnl_mutex net_rwsem irq_context: 0 (wq_completion)bond138 (work_completion)(&(&slave->notify_work)->work) rtnl_mutex net_rwsem &list->lock#2 irq_context: 0 (wq_completion)bond138 (work_completion)(&(&slave->notify_work)->work) rtnl_mutex &tn->lock irq_context: 0 (wq_completion)bond138 (work_completion)(&(&slave->notify_work)->work) rtnl_mutex.wait_lock irq_context: 0 (wq_completion)bond138 (work_completion)(&(&slave->notify_work)->work) &p->pi_lock irq_context: 0 (wq_completion)bond171#2 (work_completion)(&(&slave->notify_work)->work) rtnl_mutex net_rwsem irq_context: 0 (wq_completion)bond171#2 (work_completion)(&(&slave->notify_work)->work) rtnl_mutex net_rwsem &list->lock#2 irq_context: 0 (wq_completion)bond171#2 (work_completion)(&(&slave->notify_work)->work) rtnl_mutex &tn->lock irq_context: 0 (wq_completion)bond171#2 (work_completion)(&(&slave->notify_work)->work) rtnl_mutex.wait_lock irq_context: 0 (wq_completion)bond171#2 (work_completion)(&(&slave->notify_work)->work) &p->pi_lock irq_context: 0 (wq_completion)bond178 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex fs_reclaim irq_context: 0 (wq_completion)bond178 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex fs_reclaim mmu_notifier_invalidate_range_start irq_context: 0 (wq_completion)bond178 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &c->lock irq_context: 0 (wq_completion)bond178 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex rcu_read_lock &bond->stats_lock/1 irq_context: 0 (wq_completion)bond178 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex rcu_read_lock &ndev->lock irq_context: 0 (wq_completion)bond178 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex nl_table_lock irq_context: 0 (wq_completion)bond178 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &obj_hash[i].lock irq_context: 0 (wq_completion)bond178 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex nl_table_wait.lock irq_context: 0 (wq_completion)bond178 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex net_rwsem irq_context: 0 (wq_completion)bond178 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex net_rwsem &list->lock#2 irq_context: 0 (wq_completion)bond178 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &tn->lock irq_context: 0 (wq_completion)bond178 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &idev->mc_lock irq_context: 0 (wq_completion)bond178 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &idev->mc_lock fs_reclaim irq_context: 0 (wq_completion)bond178 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &idev->mc_lock fs_reclaim mmu_notifier_invalidate_range_start irq_context: 0 (wq_completion)bond178 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &idev->mc_lock rcu_read_lock &ndev->lock irq_context: 0 (wq_completion)bond178 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &idev->mc_lock rcu_read_lock &dir->lock#2 irq_context: 0 (wq_completion)bond178 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &idev->mc_lock rcu_read_lock &ul->lock#2 irq_context: 0 (wq_completion)bond178 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &idev->mc_lock rcu_read_lock rcu_read_lock &____s->seqcount#10 irq_context: 0 (wq_completion)bond178 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &idev->mc_lock rcu_read_lock rcu_read_lock rcu_read_lock_bh rcu_read_lock rcu_read_lock_bh rcu_read_lock &br->multicast_lock irq_context: 0 (wq_completion)bond178 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &idev->mc_lock rcu_read_lock rcu_read_lock rcu_read_lock_bh rcu_read_lock rcu_read_lock_bh rcu_read_lock &br->multicast_lock &obj_hash[i].lock irq_context: 0 (wq_completion)bond178 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &idev->mc_lock rcu_read_lock rcu_read_lock rcu_read_lock_bh rcu_read_lock rcu_read_lock_bh rcu_read_lock &br->multicast_lock &base->lock irq_context: 0 (wq_completion)bond178 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &idev->mc_lock rcu_read_lock rcu_read_lock rcu_read_lock_bh rcu_read_lock rcu_read_lock_bh rcu_read_lock &br->multicast_lock &base->lock &obj_hash[i].lock irq_context: 0 (wq_completion)bond178 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &idev->mc_lock rcu_read_lock rcu_read_lock rcu_read_lock_bh rcu_read_lock rcu_read_lock_bh rcu_read_lock &data->lock irq_context: 0 (wq_completion)bond178 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &idev->mc_lock rcu_read_lock rcu_read_lock rcu_read_lock_bh rcu_read_lock rcu_read_lock_bh rcu_read_lock &obj_hash[i].lock irq_context: 0 (wq_completion)bond178 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex.wait_lock irq_context: 0 (wq_completion)bond178 (work_completion)(&(&bond->mcast_work)->work) &p->pi_lock irq_context: 0 (wq_completion)bond53#3 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex fs_reclaim irq_context: 0 (wq_completion)bond53#3 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex fs_reclaim mmu_notifier_invalidate_range_start irq_context: 0 (wq_completion)bond53#3 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &c->lock irq_context: 0 (wq_completion)bond53#3 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex rcu_read_lock &bond->stats_lock/1 irq_context: 0 (wq_completion)bond53#3 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex rcu_read_lock &ndev->lock irq_context: 0 (wq_completion)bond53#3 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex nl_table_lock irq_context: 0 (wq_completion)bond53#3 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &obj_hash[i].lock irq_context: 0 (wq_completion)bond53#3 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex nl_table_wait.lock irq_context: 0 (wq_completion)bond53#3 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex net_rwsem irq_context: 0 (wq_completion)bond53#3 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex net_rwsem &list->lock#2 irq_context: 0 (wq_completion)bond53#3 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &tn->lock irq_context: 0 (wq_completion)bond53#3 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &idev->mc_lock irq_context: 0 (wq_completion)bond53#3 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &idev->mc_lock fs_reclaim irq_context: 0 (wq_completion)bond53#3 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &idev->mc_lock fs_reclaim mmu_notifier_invalidate_range_start irq_context: 0 (wq_completion)bond53#3 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &idev->mc_lock remove_cache_srcu irq_context: 0 (wq_completion)bond53#3 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &idev->mc_lock remove_cache_srcu quarantine_lock irq_context: 0 (wq_completion)bond53#3 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &idev->mc_lock remove_cache_srcu &c->lock irq_context: 0 (wq_completion)bond53#3 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &idev->mc_lock remove_cache_srcu &n->list_lock irq_context: 0 (wq_completion)bond53#3 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &idev->mc_lock remove_cache_srcu &obj_hash[i].lock irq_context: 0 (wq_completion)bond53#3 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &idev->mc_lock rcu_read_lock &ndev->lock irq_context: 0 (wq_completion)bond53#3 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &idev->mc_lock rcu_read_lock &dir->lock#2 irq_context: 0 (wq_completion)bond53#3 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &idev->mc_lock rcu_read_lock &ul->lock#2 irq_context: 0 (wq_completion)bond53#3 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &idev->mc_lock rcu_read_lock rcu_read_lock &____s->seqcount#12 irq_context: 0 (wq_completion)bond53#3 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &idev->mc_lock rcu_read_lock rcu_read_lock &____s->seqcount#16 irq_context: 0 (wq_completion)bond53#3 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &idev->mc_lock rcu_read_lock &obj_hash[i].lock irq_context: 0 (wq_completion)bond53#3 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &idev->mc_lock rcu_read_lock &data->lock irq_context: 0 (wq_completion)bond53#3 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex.wait_lock irq_context: 0 (wq_completion)bond53#3 (work_completion)(&(&bond->mcast_work)->work) &p->pi_lock irq_context: 0 (wq_completion)bond190 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex fs_reclaim irq_context: 0 (wq_completion)bond190 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex fs_reclaim &rq->__lock irq_context: 0 (wq_completion)bond190 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex fs_reclaim mmu_notifier_invalidate_range_start irq_context: 0 (wq_completion)bond190 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &c->lock irq_context: 0 (wq_completion)bond190 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex rcu_read_lock &bond->stats_lock/1 irq_context: 0 (wq_completion)bond190 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex rcu_read_lock &ndev->lock irq_context: 0 (wq_completion)bond190 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex nl_table_lock irq_context: 0 (wq_completion)bond190 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &obj_hash[i].lock irq_context: 0 (wq_completion)bond190 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex nl_table_wait.lock irq_context: 0 (wq_completion)bond190 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex net_rwsem irq_context: 0 (wq_completion)bond190 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex net_rwsem &list->lock#2 irq_context: 0 (wq_completion)bond190 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &tn->lock irq_context: 0 (wq_completion)bond190 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &idev->mc_lock irq_context: 0 (wq_completion)bond190 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &idev->mc_lock fs_reclaim irq_context: 0 (wq_completion)bond190 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &idev->mc_lock fs_reclaim mmu_notifier_invalidate_range_start irq_context: 0 (wq_completion)bond190 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &idev->mc_lock &c->lock irq_context: 0 (wq_completion)bond190 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &idev->mc_lock rcu_read_lock &ndev->lock irq_context: 0 (wq_completion)bond190 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &idev->mc_lock rcu_read_lock &dir->lock#2 irq_context: 0 (wq_completion)bond190 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &idev->mc_lock rcu_read_lock &ul->lock#2 irq_context: 0 (wq_completion)bond190 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &idev->mc_lock rcu_read_lock rcu_read_lock &____s->seqcount#10 irq_context: 0 (wq_completion)bond190 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &idev->mc_lock rcu_read_lock rcu_read_lock rcu_read_lock_bh rcu_read_lock rcu_read_lock_bh rcu_read_lock &br->multicast_lock irq_context: 0 (wq_completion)bond190 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &idev->mc_lock rcu_read_lock rcu_read_lock rcu_read_lock_bh rcu_read_lock rcu_read_lock_bh rcu_read_lock &br->multicast_lock &obj_hash[i].lock irq_context: 0 (wq_completion)bond190 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &idev->mc_lock rcu_read_lock rcu_read_lock rcu_read_lock_bh rcu_read_lock rcu_read_lock_bh rcu_read_lock &br->multicast_lock &base->lock irq_context: 0 (wq_completion)bond190 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &idev->mc_lock rcu_read_lock rcu_read_lock rcu_read_lock_bh rcu_read_lock rcu_read_lock_bh rcu_read_lock &data->lock irq_context: 0 (wq_completion)bond190 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &idev->mc_lock rcu_read_lock rcu_read_lock rcu_read_lock_bh rcu_read_lock rcu_read_lock_bh rcu_read_lock &obj_hash[i].lock irq_context: 0 (wq_completion)bond190 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex.wait_lock irq_context: 0 (wq_completion)bond190 (work_completion)(&(&bond->mcast_work)->work) &p->pi_lock irq_context: 0 (wq_completion)bond190 (work_completion)(&(&bond->mcast_work)->work) &p->pi_lock &rq->__lock irq_context: 0 (wq_completion)bond190 (work_completion)(&(&bond->mcast_work)->work) &p->pi_lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)bond168#3 (work_completion)(&(&bond->mcast_work)->work) irq_context: 0 (wq_completion)bond168#3 (work_completion)(&(&bond->mcast_work)->work) &obj_hash[i].lock irq_context: 0 (wq_completion)bond168#3 (work_completion)(&(&bond->mcast_work)->work) &base->lock irq_context: 0 (wq_completion)bond184#2 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex fs_reclaim irq_context: 0 (wq_completion)bond168#3 (work_completion)(&(&bond->mcast_work)->work) &base->lock &obj_hash[i].lock irq_context: 0 (wq_completion)bond184#2 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex fs_reclaim mmu_notifier_invalidate_range_start irq_context: 0 (wq_completion)bond168#3 (work_completion)(&(&bond->mcast_work)->work) &rq->__lock irq_context: 0 (wq_completion)bond184#2 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex rcu_read_lock &bond->stats_lock/1 irq_context: 0 (wq_completion)bond184#2 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex rcu_read_lock &ndev->lock irq_context: 0 (wq_completion)bond184#2 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex nl_table_lock irq_context: 0 (wq_completion)bond184#2 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &obj_hash[i].lock irq_context: 0 (wq_completion)bond184#2 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex nl_table_wait.lock irq_context: 0 (wq_completion)bond184#2 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex net_rwsem irq_context: 0 (wq_completion)bond184#2 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex net_rwsem &list->lock#2 irq_context: 0 (wq_completion)bond184#2 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &tn->lock irq_context: 0 (wq_completion)bond184#2 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &idev->mc_lock irq_context: 0 (wq_completion)bond184#2 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex.wait_lock irq_context: 0 (wq_completion)bond184#2 (work_completion)(&(&bond->mcast_work)->work) &p->pi_lock irq_context: 0 (wq_completion)bond181#2 (work_completion)(&(&bond->mcast_work)->work) &rq->__lock irq_context: 0 (wq_completion)bond181#2 (work_completion)(&(&bond->mcast_work)->work) &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)bond197 irq_context: 0 (wq_completion)bond197 &rq->__lock irq_context: 0 (wq_completion)bond197 &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)bond197 (work_completion)(&(&slave->notify_work)->work) irq_context: 0 (wq_completion)bond197 (work_completion)(&(&slave->notify_work)->work) &obj_hash[i].lock irq_context: 0 (wq_completion)bond197 (work_completion)(&(&slave->notify_work)->work) &base->lock irq_context: 0 (wq_completion)bond197 (work_completion)(&(&slave->notify_work)->work) &base->lock &obj_hash[i].lock irq_context: 0 (wq_completion)bond197 (work_completion)(&(&bond->mcast_work)->work) irq_context: 0 (wq_completion)bond197 (work_completion)(&(&bond->mcast_work)->work) &rq->__lock irq_context: 0 (wq_completion)bond197 (work_completion)(&(&bond->mcast_work)->work) &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)bond197 (work_completion)(&(&bond->mcast_work)->work) &obj_hash[i].lock irq_context: 0 (wq_completion)bond197 (work_completion)(&(&bond->mcast_work)->work) &base->lock irq_context: 0 (wq_completion)bond197 (work_completion)(&(&bond->mcast_work)->work) &base->lock &obj_hash[i].lock irq_context: 0 (wq_completion)bond185#2 irq_context: 0 (wq_completion)bond185#2 (work_completion)(&(&slave->notify_work)->work) irq_context: 0 (wq_completion)bond185#2 (work_completion)(&(&slave->notify_work)->work) &obj_hash[i].lock irq_context: 0 (wq_completion)bond185#2 (work_completion)(&(&slave->notify_work)->work) &base->lock irq_context: 0 (wq_completion)bond185#2 (work_completion)(&(&slave->notify_work)->work) &base->lock &obj_hash[i].lock irq_context: 0 (wq_completion)bond185#2 (work_completion)(&(&slave->notify_work)->work) &rq->__lock irq_context: 0 (wq_completion)bond185#2 (work_completion)(&(&slave->notify_work)->work) &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)bond185#2 (work_completion)(&(&slave->notify_work)->work) rtnl_mutex net_rwsem irq_context: 0 (wq_completion)bond185#2 (work_completion)(&(&slave->notify_work)->work) rtnl_mutex net_rwsem &list->lock#2 irq_context: 0 (wq_completion)bond185#2 (work_completion)(&(&slave->notify_work)->work) rtnl_mutex &tn->lock irq_context: 0 (wq_completion)bond185#2 (work_completion)(&(&slave->notify_work)->work) rtnl_mutex.wait_lock irq_context: 0 (wq_completion)bond185#2 (work_completion)(&(&slave->notify_work)->work) &p->pi_lock irq_context: 0 (wq_completion)bond183#2 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &rq->__lock irq_context: 0 (wq_completion)bond183#2 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)bond183#2 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &cfs_rq->removed.lock irq_context: 0 (wq_completion)bond183#2 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &obj_hash[i].lock irq_context: 0 (wq_completion)bond183#2 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex fs_reclaim irq_context: 0 (wq_completion)bond183#2 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex fs_reclaim mmu_notifier_invalidate_range_start irq_context: 0 (wq_completion)bond183#2 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &c->lock irq_context: 0 (wq_completion)bond183#2 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &n->list_lock irq_context: 0 (wq_completion)bond183#2 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &n->list_lock &c->lock irq_context: 0 (wq_completion)bond183#2 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex rcu_read_lock &bond->stats_lock/1 irq_context: 0 (wq_completion)bond183#2 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex rcu_read_lock &ndev->lock irq_context: 0 (wq_completion)bond183#2 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex nl_table_lock irq_context: 0 (wq_completion)bond183#2 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex nl_table_wait.lock irq_context: 0 (wq_completion)bond183#2 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex net_rwsem irq_context: 0 (wq_completion)bond183#2 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex net_rwsem &list->lock#2 irq_context: 0 (wq_completion)bond183#2 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &tn->lock irq_context: 0 (wq_completion)bond183#2 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &idev->mc_lock irq_context: 0 (wq_completion)bond183#2 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex rcu_read_lock rcu_node_0 irq_context: 0 (wq_completion)bond183#2 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex rcu_read_lock &rq->__lock irq_context: 0 (wq_completion)bond183#2 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex rcu_read_lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)bond183#2 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex.wait_lock irq_context: 0 (wq_completion)bond183#2 (work_completion)(&(&bond->mcast_work)->work) &p->pi_lock irq_context: 0 (wq_completion)bond183#2 (work_completion)(&(&bond->mcast_work)->work) &p->pi_lock &rq->__lock irq_context: 0 (wq_completion)bond183#2 (work_completion)(&(&bond->mcast_work)->work) &p->pi_lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)bond185#2 (work_completion)(&(&bond->mcast_work)->work) irq_context: 0 (wq_completion)bond185#2 (work_completion)(&(&bond->mcast_work)->work) &obj_hash[i].lock irq_context: 0 (wq_completion)bond185#2 (work_completion)(&(&bond->mcast_work)->work) &base->lock irq_context: 0 (wq_completion)bond185#2 (work_completion)(&(&bond->mcast_work)->work) &base->lock &obj_hash[i].lock irq_context: 0 (wq_completion)bond185#2 (work_completion)(&(&bond->mcast_work)->work) &rq->__lock irq_context: 0 (wq_completion)bond195 (work_completion)(&(&bond->mcast_work)->work) fill_pool_map-wait-type-override &c->lock irq_context: 0 (wq_completion)bond195 (work_completion)(&(&bond->mcast_work)->work) fill_pool_map-wait-type-override &____s->seqcount#2 irq_context: 0 (wq_completion)bond195 (work_completion)(&(&bond->mcast_work)->work) fill_pool_map-wait-type-override &____s->seqcount irq_context: 0 (wq_completion)bond195 (work_completion)(&(&bond->mcast_work)->work) fill_pool_map-wait-type-override pool_lock irq_context: 0 (wq_completion)bond169#3 irq_context: 0 (wq_completion)bond169#3 (work_completion)(&(&slave->notify_work)->work) irq_context: 0 (wq_completion)bond169#3 (work_completion)(&(&slave->notify_work)->work) &obj_hash[i].lock irq_context: 0 (wq_completion)bond169#3 (work_completion)(&(&slave->notify_work)->work) &base->lock irq_context: 0 (wq_completion)bond169#3 (work_completion)(&(&slave->notify_work)->work) &base->lock &obj_hash[i].lock irq_context: 0 (wq_completion)bond169#3 (work_completion)(&(&bond->mcast_work)->work) irq_context: 0 (wq_completion)bond169#3 (work_completion)(&(&bond->mcast_work)->work) &obj_hash[i].lock irq_context: 0 (wq_completion)bond169#3 (work_completion)(&(&bond->mcast_work)->work) &base->lock irq_context: 0 (wq_completion)bond169#3 (work_completion)(&(&bond->mcast_work)->work) &base->lock &obj_hash[i].lock irq_context: 0 (wq_completion)bond169#3 (work_completion)(&(&bond->mcast_work)->work) &rq->__lock irq_context: 0 (wq_completion)bond169#3 (work_completion)(&(&slave->notify_work)->work) &rq->__lock irq_context: 0 (wq_completion)bond169#3 (work_completion)(&(&slave->notify_work)->work) &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)bond198 irq_context: 0 (wq_completion)bond198 (work_completion)(&(&slave->notify_work)->work) irq_context: 0 (wq_completion)bond198 (work_completion)(&(&slave->notify_work)->work) &obj_hash[i].lock irq_context: 0 (wq_completion)bond198 (work_completion)(&(&slave->notify_work)->work) &base->lock irq_context: 0 (wq_completion)bond198 (work_completion)(&(&slave->notify_work)->work) &base->lock &obj_hash[i].lock irq_context: softirq (&in_dev->mr_ifc_timer) fill_pool_map-wait-type-override &c->lock irq_context: softirq (&in_dev->mr_ifc_timer) fill_pool_map-wait-type-override &____s->seqcount#2 irq_context: softirq (&in_dev->mr_ifc_timer) fill_pool_map-wait-type-override &____s->seqcount irq_context: 0 (wq_completion)bond198 (work_completion)(&(&bond->mcast_work)->work) irq_context: 0 (wq_completion)bond198 (work_completion)(&(&bond->mcast_work)->work) &obj_hash[i].lock irq_context: 0 (wq_completion)bond198 (work_completion)(&(&bond->mcast_work)->work) &base->lock irq_context: 0 (wq_completion)bond198 (work_completion)(&(&bond->mcast_work)->work) &base->lock &obj_hash[i].lock irq_context: 0 (wq_completion)bond179#2 (work_completion)(&(&slave->notify_work)->work) &rq->__lock irq_context: 0 (wq_completion)bond179#2 (work_completion)(&(&slave->notify_work)->work) &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)bond186#2 irq_context: 0 (wq_completion)bond186#2 (work_completion)(&(&slave->notify_work)->work) irq_context: 0 (wq_completion)bond186#2 (work_completion)(&(&slave->notify_work)->work) &obj_hash[i].lock irq_context: 0 (wq_completion)bond186#2 (work_completion)(&(&slave->notify_work)->work) &base->lock irq_context: 0 (wq_completion)bond186#2 (work_completion)(&(&slave->notify_work)->work) &base->lock &obj_hash[i].lock irq_context: 0 (wq_completion)bond186#2 (work_completion)(&(&slave->notify_work)->work) &rq->__lock irq_context: 0 (wq_completion)bond186#2 (work_completion)(&(&slave->notify_work)->work) &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)bond178#2 &rq->__lock irq_context: 0 (wq_completion)bond178#2 &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)bond186#2 (work_completion)(&(&bond->mcast_work)->work) irq_context: 0 (wq_completion)bond186#2 (work_completion)(&(&bond->mcast_work)->work) &obj_hash[i].lock irq_context: 0 (wq_completion)bond186#2 (work_completion)(&(&bond->mcast_work)->work) &base->lock irq_context: 0 (wq_completion)bond186#2 (work_completion)(&(&bond->mcast_work)->work) &base->lock &obj_hash[i].lock irq_context: 0 (wq_completion)bond186#2 (work_completion)(&(&bond->mcast_work)->work) &rq->__lock irq_context: 0 (wq_completion)bond186#2 (work_completion)(&(&bond->mcast_work)->work) &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)bond170#3 irq_context: 0 (wq_completion)bond170#3 (work_completion)(&(&slave->notify_work)->work) irq_context: 0 (wq_completion)bond170#3 (work_completion)(&(&slave->notify_work)->work) &obj_hash[i].lock irq_context: 0 (wq_completion)bond170#3 (work_completion)(&(&slave->notify_work)->work) &base->lock irq_context: 0 (wq_completion)bond170#3 (work_completion)(&(&slave->notify_work)->work) &base->lock &obj_hash[i].lock irq_context: 0 (wq_completion)bond170#3 (work_completion)(&(&slave->notify_work)->work) &rq->__lock irq_context: 0 (wq_completion)bond170#3 (work_completion)(&(&slave->notify_work)->work) &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)bond170#3 (work_completion)(&(&bond->mcast_work)->work) irq_context: 0 (wq_completion)bond170#3 (work_completion)(&(&bond->mcast_work)->work) &obj_hash[i].lock irq_context: 0 (wq_completion)bond170#3 (work_completion)(&(&bond->mcast_work)->work) &base->lock irq_context: 0 (wq_completion)bond170#3 (work_completion)(&(&bond->mcast_work)->work) &base->lock &obj_hash[i].lock irq_context: 0 (wq_completion)bond164#3 (work_completion)(&(&slave->notify_work)->work) &rq->__lock irq_context: 0 (wq_completion)bond164#3 (work_completion)(&(&slave->notify_work)->work) &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)bond168#3 (work_completion)(&(&bond->mcast_work)->work) &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)bond199 irq_context: 0 (wq_completion)bond199 &rq->__lock irq_context: 0 (wq_completion)bond199 &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)bond199 (work_completion)(&(&slave->notify_work)->work) irq_context: 0 (wq_completion)bond199 (work_completion)(&(&slave->notify_work)->work) &obj_hash[i].lock irq_context: 0 (wq_completion)bond199 (work_completion)(&(&slave->notify_work)->work) &base->lock irq_context: 0 (wq_completion)bond199 (work_completion)(&(&slave->notify_work)->work) &base->lock &obj_hash[i].lock irq_context: 0 (wq_completion)bond199 (work_completion)(&(&slave->notify_work)->work) &rq->__lock irq_context: 0 (wq_completion)bond199 (work_completion)(&(&slave->notify_work)->work) &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)bond199 (work_completion)(&(&bond->mcast_work)->work) irq_context: 0 (wq_completion)bond199 (work_completion)(&(&bond->mcast_work)->work) &obj_hash[i].lock irq_context: 0 (wq_completion)bond199 (work_completion)(&(&bond->mcast_work)->work) &base->lock irq_context: 0 (wq_completion)bond199 (work_completion)(&(&bond->mcast_work)->work) &base->lock &obj_hash[i].lock irq_context: 0 (wq_completion)bond199 (work_completion)(&(&bond->mcast_work)->work) &rq->__lock irq_context: 0 (wq_completion)bond199 (work_completion)(&(&bond->mcast_work)->work) &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)bond59#2 irq_context: 0 (wq_completion)bond59#2 (work_completion)(&(&slave->notify_work)->work) irq_context: 0 (wq_completion)bond59#2 (work_completion)(&(&slave->notify_work)->work) &obj_hash[i].lock irq_context: 0 (wq_completion)bond59#2 (work_completion)(&(&slave->notify_work)->work) &base->lock irq_context: 0 (wq_completion)bond59#2 (work_completion)(&(&slave->notify_work)->work) &base->lock &obj_hash[i].lock irq_context: 0 (wq_completion)bond59#2 (work_completion)(&(&slave->notify_work)->work) &rq->__lock irq_context: 0 (wq_completion)bond59#2 (work_completion)(&(&slave->notify_work)->work) &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)bond59#2 (work_completion)(&(&bond->mcast_work)->work) irq_context: 0 (wq_completion)bond59#2 (work_completion)(&(&bond->mcast_work)->work) &obj_hash[i].lock irq_context: 0 (wq_completion)bond59#2 (work_completion)(&(&bond->mcast_work)->work) &base->lock irq_context: 0 (wq_completion)bond59#2 (work_completion)(&(&bond->mcast_work)->work) &base->lock &obj_hash[i].lock irq_context: 0 (wq_completion)bond59#2 (work_completion)(&(&bond->mcast_work)->work) &rq->__lock irq_context: 0 (wq_completion)bond59#2 (work_completion)(&(&bond->mcast_work)->work) &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)bond132#2 &rq->__lock irq_context: 0 (wq_completion)bond153 (work_completion)(&(&slave->notify_work)->work) rtnl_mutex net_rwsem irq_context: 0 (wq_completion)bond153 (work_completion)(&(&slave->notify_work)->work) rtnl_mutex net_rwsem &list->lock#2 irq_context: 0 (wq_completion)bond153 (work_completion)(&(&slave->notify_work)->work) rtnl_mutex &tn->lock irq_context: 0 (wq_completion)bond153 (work_completion)(&(&slave->notify_work)->work) rtnl_mutex.wait_lock irq_context: 0 (wq_completion)bond153 (work_completion)(&(&slave->notify_work)->work) &p->pi_lock irq_context: 0 (wq_completion)bond146#3 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex fs_reclaim irq_context: 0 (wq_completion)bond146#3 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex fs_reclaim mmu_notifier_invalidate_range_start irq_context: 0 (wq_completion)bond146#3 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex rcu_read_lock &bond->stats_lock/1 irq_context: 0 (wq_completion)bond146#3 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex rcu_read_lock &ndev->lock irq_context: 0 (wq_completion)bond146#3 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex nl_table_lock irq_context: 0 (wq_completion)bond146#3 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &obj_hash[i].lock irq_context: 0 (wq_completion)bond146#3 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex nl_table_wait.lock irq_context: 0 (wq_completion)bond146#3 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex net_rwsem irq_context: 0 (wq_completion)bond146#3 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex net_rwsem &list->lock#2 irq_context: 0 (wq_completion)bond146#3 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &tn->lock irq_context: 0 (wq_completion)bond146#3 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &idev->mc_lock irq_context: 0 (wq_completion)bond146#3 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &idev->mc_lock fs_reclaim irq_context: 0 (wq_completion)bond146#3 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &idev->mc_lock fs_reclaim mmu_notifier_invalidate_range_start irq_context: 0 (wq_completion)bond146#3 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &idev->mc_lock rcu_read_lock &ndev->lock irq_context: 0 (wq_completion)bond146#3 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &idev->mc_lock rcu_read_lock &dir->lock#2 irq_context: 0 (wq_completion)bond146#3 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &idev->mc_lock rcu_read_lock &ul->lock#2 irq_context: 0 (wq_completion)bond146#3 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &idev->mc_lock rcu_read_lock rcu_read_lock &____s->seqcount#10 irq_context: 0 (wq_completion)bond146#3 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &idev->mc_lock rcu_read_lock rcu_read_lock rcu_read_lock_bh rcu_read_lock rcu_read_lock_bh rcu_read_lock &br->multicast_lock irq_context: 0 (wq_completion)bond146#3 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &idev->mc_lock rcu_read_lock rcu_read_lock rcu_read_lock_bh rcu_read_lock rcu_read_lock_bh rcu_read_lock &br->multicast_lock &obj_hash[i].lock irq_context: 0 (wq_completion)bond146#3 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &idev->mc_lock rcu_read_lock rcu_read_lock rcu_read_lock_bh rcu_read_lock rcu_read_lock_bh rcu_read_lock &br->multicast_lock &base->lock irq_context: 0 (wq_completion)bond146#3 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &idev->mc_lock rcu_read_lock rcu_read_lock rcu_read_lock_bh rcu_read_lock rcu_read_lock_bh rcu_read_lock &br->multicast_lock &base->lock &obj_hash[i].lock irq_context: 0 (wq_completion)bond146#3 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &idev->mc_lock rcu_read_lock rcu_read_lock rcu_read_lock_bh rcu_read_lock rcu_read_lock_bh rcu_read_lock &data->lock irq_context: 0 (wq_completion)bond146#3 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &idev->mc_lock rcu_read_lock rcu_read_lock rcu_read_lock_bh rcu_read_lock rcu_read_lock_bh rcu_read_lock &obj_hash[i].lock irq_context: 0 (wq_completion)bond146#3 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex.wait_lock irq_context: 0 (wq_completion)bond146#3 (work_completion)(&(&bond->mcast_work)->work) &p->pi_lock irq_context: 0 (wq_completion)bond157 (work_completion)(&(&slave->notify_work)->work) rtnl_mutex net_rwsem irq_context: 0 (wq_completion)bond157 (work_completion)(&(&slave->notify_work)->work) rtnl_mutex net_rwsem &list->lock#2 irq_context: 0 (wq_completion)bond157 (work_completion)(&(&slave->notify_work)->work) rtnl_mutex &tn->lock irq_context: 0 (wq_completion)bond157 (work_completion)(&(&slave->notify_work)->work) rtnl_mutex.wait_lock irq_context: 0 (wq_completion)bond157 (work_completion)(&(&slave->notify_work)->work) &p->pi_lock irq_context: 0 (wq_completion)bond34 (work_completion)(&(&slave->notify_work)->work) rtnl_mutex net_rwsem irq_context: 0 (wq_completion)bond34 (work_completion)(&(&slave->notify_work)->work) rtnl_mutex net_rwsem &list->lock#2 irq_context: 0 (wq_completion)bond34 (work_completion)(&(&slave->notify_work)->work) rtnl_mutex &tn->lock irq_context: 0 (wq_completion)bond34 (work_completion)(&(&slave->notify_work)->work) rtnl_mutex.wait_lock irq_context: 0 (wq_completion)bond34 (work_completion)(&(&slave->notify_work)->work) &p->pi_lock irq_context: 0 (wq_completion)bond161 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex fs_reclaim irq_context: 0 (wq_completion)bond161 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex fs_reclaim mmu_notifier_invalidate_range_start irq_context: 0 (wq_completion)bond161 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &c->lock irq_context: 0 (wq_completion)bond161 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex rcu_read_lock &bond->stats_lock/1 irq_context: 0 (wq_completion)bond161 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex rcu_read_lock &ndev->lock irq_context: 0 (wq_completion)bond161 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex nl_table_lock irq_context: 0 (wq_completion)bond161 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &obj_hash[i].lock irq_context: 0 (wq_completion)bond161 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex nl_table_wait.lock irq_context: 0 (wq_completion)bond161 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex net_rwsem irq_context: 0 (wq_completion)bond161 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex net_rwsem &list->lock#2 irq_context: 0 (wq_completion)bond161 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &tn->lock irq_context: 0 (wq_completion)bond161 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &idev->mc_lock irq_context: 0 (wq_completion)bond161 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &idev->mc_lock fs_reclaim irq_context: 0 (wq_completion)bond161 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &idev->mc_lock fs_reclaim mmu_notifier_invalidate_range_start irq_context: 0 (wq_completion)bond161 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &idev->mc_lock &c->lock irq_context: 0 (wq_completion)bond161 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &idev->mc_lock rcu_read_lock &ndev->lock irq_context: 0 (wq_completion)bond161 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &idev->mc_lock rcu_read_lock &dir->lock#2 irq_context: 0 (wq_completion)bond161 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &idev->mc_lock rcu_read_lock &ul->lock#2 irq_context: 0 (wq_completion)bond161 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &idev->mc_lock rcu_read_lock rcu_read_lock rcu_node_0 irq_context: 0 (wq_completion)bond161 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &idev->mc_lock rcu_read_lock rcu_read_lock &rq->__lock irq_context: 0 (wq_completion)bond161 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &idev->mc_lock rcu_read_lock rcu_read_lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)bond156#2 (work_completion)(&(&slave->notify_work)->work) &rq->__lock irq_context: 0 (wq_completion)bond156#2 (work_completion)(&(&slave->notify_work)->work) &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)bond32#3 &rq->__lock irq_context: 0 (wq_completion)bond32#3 &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)bond161 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &idev->mc_lock rcu_read_lock rcu_read_lock &obj_hash[i].lock irq_context: 0 (wq_completion)bond161 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &idev->mc_lock rcu_read_lock rcu_read_lock &dir->lock#2 irq_context: 0 (wq_completion)bond161 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &idev->mc_lock rcu_read_lock rcu_read_lock &tbl->lock irq_context: 0 (wq_completion)bond161 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &idev->mc_lock rcu_read_lock rcu_read_lock &n->lock irq_context: 0 (wq_completion)bond161 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &idev->mc_lock rcu_read_lock rcu_read_lock &____s->seqcount#9 irq_context: 0 (wq_completion)bond161 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &idev->mc_lock rcu_read_lock rcu_read_lock rcu_read_lock_bh rcu_read_lock rcu_read_lock_bh rcu_read_lock &br->multicast_lock irq_context: 0 (wq_completion)bond161 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &idev->mc_lock rcu_read_lock rcu_read_lock rcu_read_lock_bh rcu_read_lock rcu_read_lock_bh rcu_read_lock &br->multicast_lock &obj_hash[i].lock irq_context: 0 (wq_completion)bond161 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &idev->mc_lock rcu_read_lock rcu_read_lock rcu_read_lock_bh rcu_read_lock rcu_read_lock_bh rcu_read_lock &br->multicast_lock &base->lock irq_context: 0 (wq_completion)bond161 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &idev->mc_lock rcu_read_lock rcu_read_lock rcu_read_lock_bh rcu_read_lock rcu_read_lock_bh rcu_read_lock &br->multicast_lock &base->lock &obj_hash[i].lock irq_context: 0 (wq_completion)bond161 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &idev->mc_lock rcu_read_lock rcu_read_lock rcu_read_lock_bh rcu_read_lock rcu_read_lock_bh rcu_read_lock &data->lock irq_context: 0 (wq_completion)bond161 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &idev->mc_lock rcu_read_lock rcu_read_lock rcu_read_lock_bh rcu_read_lock rcu_read_lock_bh rcu_read_lock &obj_hash[i].lock irq_context: 0 (wq_completion)bond161 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &idev->mc_lock rcu_read_lock rcu_node_0 irq_context: 0 (wq_completion)bond161 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &idev->mc_lock rcu_read_lock &rcu_state.gp_wq irq_context: 0 (wq_completion)bond161 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &idev->mc_lock rcu_read_lock &rcu_state.gp_wq &p->pi_lock irq_context: 0 (wq_completion)bond161 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &idev->mc_lock rcu_read_lock &rcu_state.gp_wq &p->pi_lock &rq->__lock irq_context: 0 (wq_completion)bond161 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &idev->mc_lock rcu_read_lock &rcu_state.gp_wq &p->pi_lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)bond161 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &idev->mc_lock rcu_read_lock &rq->__lock irq_context: 0 (wq_completion)bond161 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &idev->mc_lock rcu_read_lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)bond161 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex.wait_lock irq_context: 0 (wq_completion)bond161 (work_completion)(&(&bond->mcast_work)->work) &p->pi_lock irq_context: 0 (wq_completion)bond161 (work_completion)(&(&bond->mcast_work)->work) &p->pi_lock &rq->__lock irq_context: 0 (wq_completion)bond161 (work_completion)(&(&bond->mcast_work)->work) &p->pi_lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)bond7 (work_completion)(&(&bond->alb_work)->work) fill_pool_map-wait-type-override pool_lock irq_context: 0 (wq_completion)bond187#2 irq_context: 0 (wq_completion)bond187#2 (work_completion)(&(&slave->notify_work)->work) irq_context: 0 (wq_completion)bond187#2 (work_completion)(&(&slave->notify_work)->work) &rq->__lock irq_context: 0 (wq_completion)bond187#2 (work_completion)(&(&slave->notify_work)->work) &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)bond187#2 (work_completion)(&(&slave->notify_work)->work) &obj_hash[i].lock irq_context: 0 (wq_completion)bond187#2 (work_completion)(&(&slave->notify_work)->work) &base->lock irq_context: 0 (wq_completion)bond187#2 (work_completion)(&(&slave->notify_work)->work) &base->lock &obj_hash[i].lock irq_context: 0 (wq_completion)bond187#2 (work_completion)(&(&bond->mcast_work)->work) irq_context: 0 (wq_completion)bond187#2 (work_completion)(&(&bond->mcast_work)->work) &obj_hash[i].lock irq_context: 0 (wq_completion)bond187#2 (work_completion)(&(&bond->mcast_work)->work) &base->lock irq_context: 0 (wq_completion)bond187#2 (work_completion)(&(&bond->mcast_work)->work) &base->lock &obj_hash[i].lock irq_context: 0 (wq_completion)bond187#2 (work_completion)(&(&bond->mcast_work)->work) &rq->__lock irq_context: 0 (wq_completion)bond187#2 (work_completion)(&(&bond->mcast_work)->work) &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)bond171#3 irq_context: 0 (wq_completion)bond171#3 (work_completion)(&(&slave->notify_work)->work) irq_context: 0 (wq_completion)bond171#3 (work_completion)(&(&slave->notify_work)->work) &obj_hash[i].lock irq_context: 0 (wq_completion)bond171#3 (work_completion)(&(&slave->notify_work)->work) &base->lock irq_context: 0 (wq_completion)bond171#3 (work_completion)(&(&slave->notify_work)->work) &base->lock &obj_hash[i].lock irq_context: 0 (wq_completion)bond171#3 (work_completion)(&(&slave->notify_work)->work) &rq->__lock irq_context: 0 (wq_completion)bond171#3 (work_completion)(&(&slave->notify_work)->work) &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)bond136#3 (work_completion)(&(&slave->notify_work)->work) rtnl_mutex net_rwsem irq_context: 0 (wq_completion)bond171#3 (work_completion)(&(&bond->mcast_work)->work) irq_context: 0 (wq_completion)bond136#3 (work_completion)(&(&slave->notify_work)->work) rtnl_mutex net_rwsem &list->lock#2 irq_context: 0 (wq_completion)bond171#3 (work_completion)(&(&bond->mcast_work)->work) &obj_hash[i].lock irq_context: 0 (wq_completion)bond136#3 (work_completion)(&(&slave->notify_work)->work) rtnl_mutex &tn->lock irq_context: 0 (wq_completion)bond171#3 (work_completion)(&(&bond->mcast_work)->work) &base->lock irq_context: 0 (wq_completion)bond171#3 (work_completion)(&(&bond->mcast_work)->work) &base->lock &obj_hash[i].lock irq_context: 0 (wq_completion)bond136#3 (work_completion)(&(&slave->notify_work)->work) rtnl_mutex.wait_lock irq_context: 0 (wq_completion)bond119 (work_completion)(&(&slave->notify_work)->work) rtnl_mutex net_rwsem irq_context: 0 (wq_completion)bond136#3 (work_completion)(&(&slave->notify_work)->work) &p->pi_lock irq_context: 0 (wq_completion)bond119 (work_completion)(&(&slave->notify_work)->work) rtnl_mutex net_rwsem &list->lock#2 irq_context: 0 (wq_completion)bond119 (work_completion)(&(&slave->notify_work)->work) rtnl_mutex &tn->lock irq_context: 0 (wq_completion)bond119 (work_completion)(&(&slave->notify_work)->work) rtnl_mutex.wait_lock irq_context: 0 (wq_completion)bond186 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex fs_reclaim irq_context: 0 (wq_completion)bond119 (work_completion)(&(&slave->notify_work)->work) &p->pi_lock irq_context: 0 (wq_completion)bond186 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex fs_reclaim mmu_notifier_invalidate_range_start irq_context: 0 (wq_completion)bond186 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &c->lock irq_context: 0 (wq_completion)bond186 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex rcu_read_lock &bond->stats_lock/1 irq_context: 0 (wq_completion)bond186 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex rcu_read_lock &ndev->lock irq_context: 0 (wq_completion)bond180 (work_completion)(&(&bond->mcast_work)->work) &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)bond186 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex rcu_read_lock rcu_node_0 irq_context: 0 (wq_completion)bond186 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex rcu_read_lock &rq->__lock irq_context: 0 (wq_completion)bond186 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex nl_table_lock irq_context: 0 (wq_completion)bond186 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &obj_hash[i].lock irq_context: 0 (wq_completion)bond186 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex nl_table_wait.lock irq_context: 0 (wq_completion)bond186 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex net_rwsem irq_context: 0 (wq_completion)bond186 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex net_rwsem &list->lock#2 irq_context: 0 (wq_completion)bond186 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &tn->lock irq_context: 0 (wq_completion)bond186 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &idev->mc_lock irq_context: 0 (wq_completion)bond186 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &idev->mc_lock fs_reclaim irq_context: 0 (wq_completion)bond186 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &idev->mc_lock fs_reclaim mmu_notifier_invalidate_range_start irq_context: 0 (wq_completion)bond186 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &idev->mc_lock &c->lock irq_context: 0 (wq_completion)bond186 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &idev->mc_lock rcu_read_lock &ndev->lock irq_context: 0 (wq_completion)bond186 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &idev->mc_lock rcu_read_lock &c->lock irq_context: 0 (wq_completion)bond186 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &idev->mc_lock rcu_read_lock &dir->lock#2 irq_context: 0 (wq_completion)bond186 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &idev->mc_lock rcu_read_lock &n->list_lock irq_context: 0 (wq_completion)bond186 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &idev->mc_lock rcu_read_lock &n->list_lock &c->lock irq_context: 0 (wq_completion)bond186 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &idev->mc_lock rcu_read_lock &ul->lock#2 irq_context: 0 (wq_completion)bond186 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &idev->mc_lock rcu_read_lock rcu_read_lock &____s->seqcount#10 irq_context: 0 (wq_completion)bond186 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &idev->mc_lock rcu_read_lock rcu_read_lock rcu_read_lock_bh rcu_read_lock rcu_read_lock_bh rcu_read_lock &br->multicast_lock irq_context: 0 (wq_completion)bond186 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &idev->mc_lock rcu_read_lock rcu_read_lock rcu_read_lock_bh rcu_read_lock rcu_read_lock_bh rcu_read_lock &br->multicast_lock &obj_hash[i].lock irq_context: 0 (wq_completion)bond186 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &idev->mc_lock rcu_read_lock rcu_read_lock rcu_read_lock_bh rcu_read_lock rcu_read_lock_bh rcu_read_lock &br->multicast_lock &base->lock irq_context: 0 (wq_completion)bond186 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &idev->mc_lock rcu_read_lock rcu_read_lock rcu_read_lock_bh rcu_read_lock rcu_read_lock_bh rcu_read_lock &br->multicast_lock &base->lock &obj_hash[i].lock irq_context: 0 (wq_completion)bond186 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &idev->mc_lock rcu_read_lock rcu_read_lock rcu_read_lock_bh rcu_read_lock rcu_read_lock_bh rcu_read_lock &data->lock irq_context: 0 (wq_completion)bond186 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &idev->mc_lock rcu_read_lock rcu_read_lock rcu_read_lock_bh rcu_read_lock rcu_read_lock_bh rcu_read_lock &obj_hash[i].lock irq_context: 0 (wq_completion)bond186 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex.wait_lock irq_context: 0 (wq_completion)bond186 (work_completion)(&(&bond->mcast_work)->work) &p->pi_lock irq_context: 0 (wq_completion)bond186 (work_completion)(&(&bond->mcast_work)->work) &p->pi_lock &rq->__lock irq_context: 0 (wq_completion)bond186 (work_completion)(&(&bond->mcast_work)->work) &p->pi_lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)bond200 irq_context: 0 (wq_completion)bond200 &rq->__lock irq_context: 0 (wq_completion)bond200 &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)bond200 (work_completion)(&(&slave->notify_work)->work) irq_context: 0 (wq_completion)bond200 (work_completion)(&(&slave->notify_work)->work) &obj_hash[i].lock irq_context: 0 (wq_completion)bond200 (work_completion)(&(&slave->notify_work)->work) &base->lock irq_context: 0 (wq_completion)bond200 (work_completion)(&(&slave->notify_work)->work) &base->lock &obj_hash[i].lock irq_context: 0 (wq_completion)bond200 (work_completion)(&(&slave->notify_work)->work) &rq->__lock irq_context: 0 (wq_completion)bond200 (work_completion)(&(&slave->notify_work)->work) &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)bond160#3 &rq->__lock irq_context: 0 (wq_completion)bond160#3 &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)bond200 (work_completion)(&(&bond->mcast_work)->work) irq_context: 0 (wq_completion)bond200 (work_completion)(&(&bond->mcast_work)->work) &obj_hash[i].lock irq_context: 0 (wq_completion)bond200 (work_completion)(&(&bond->mcast_work)->work) &base->lock irq_context: 0 (wq_completion)bond200 (work_completion)(&(&bond->mcast_work)->work) &base->lock &obj_hash[i].lock irq_context: 0 (wq_completion)bond200 (work_completion)(&(&bond->mcast_work)->work) &rq->__lock irq_context: 0 (wq_completion)bond200 (work_completion)(&(&bond->mcast_work)->work) &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)events (work_completion)(&(&hwstats->traffic_dw)->work) &base->lock fill_pool_map-wait-type-override pool_lock#2 irq_context: 0 (wq_completion)events (work_completion)(&(&hwstats->traffic_dw)->work) &base->lock fill_pool_map-wait-type-override pool_lock irq_context: 0 (wq_completion)bond184#2 &rq->__lock irq_context: 0 (wq_completion)bond184#2 &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)bond188#2 irq_context: 0 (wq_completion)bond188#2 (work_completion)(&(&slave->notify_work)->work) irq_context: 0 (wq_completion)bond188#2 (work_completion)(&(&slave->notify_work)->work) &rq->__lock irq_context: 0 (wq_completion)bond188#2 (work_completion)(&(&slave->notify_work)->work) &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)bond188#2 (work_completion)(&(&slave->notify_work)->work) &obj_hash[i].lock irq_context: 0 (wq_completion)bond188#2 (work_completion)(&(&slave->notify_work)->work) &base->lock irq_context: 0 (wq_completion)bond188#2 (work_completion)(&(&slave->notify_work)->work) &base->lock &obj_hash[i].lock irq_context: 0 (wq_completion)bond197 (work_completion)(&(&slave->notify_work)->work) &rq->__lock irq_context: 0 (wq_completion)bond197 (work_completion)(&(&slave->notify_work)->work) &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)bond188#2 (work_completion)(&(&bond->mcast_work)->work) irq_context: 0 (wq_completion)bond188#2 (work_completion)(&(&bond->mcast_work)->work) &obj_hash[i].lock irq_context: 0 (wq_completion)bond188#2 (work_completion)(&(&bond->mcast_work)->work) &base->lock irq_context: 0 (wq_completion)bond188#2 (work_completion)(&(&bond->mcast_work)->work) &base->lock &obj_hash[i].lock irq_context: 0 (wq_completion)bond188#2 (work_completion)(&(&bond->mcast_work)->work) &rq->__lock irq_context: 0 (wq_completion)bond188#2 (work_completion)(&(&bond->mcast_work)->work) &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)bond175 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex fs_reclaim irq_context: 0 (wq_completion)bond175 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex fs_reclaim mmu_notifier_invalidate_range_start irq_context: 0 (wq_completion)bond175 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &c->lock irq_context: 0 (wq_completion)bond175 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex rcu_read_lock &bond->stats_lock/1 irq_context: 0 (wq_completion)bond175 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex rcu_read_lock &ndev->lock irq_context: 0 (wq_completion)bond175 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex nl_table_lock irq_context: 0 (wq_completion)bond175 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &obj_hash[i].lock irq_context: 0 (wq_completion)bond175 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &rq->__lock irq_context: 0 (wq_completion)bond175 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)bond175 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex nl_table_wait.lock irq_context: 0 (wq_completion)bond175 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex net_rwsem irq_context: 0 (wq_completion)bond175 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex net_rwsem &list->lock#2 irq_context: 0 (wq_completion)bond175 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &tn->lock irq_context: 0 (wq_completion)bond175 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &idev->mc_lock irq_context: 0 (wq_completion)bond175 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &idev->mc_lock fs_reclaim irq_context: 0 (wq_completion)bond175 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &idev->mc_lock fs_reclaim mmu_notifier_invalidate_range_start irq_context: 0 (wq_completion)bond175 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &idev->mc_lock &c->lock irq_context: 0 (wq_completion)bond175 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &idev->mc_lock rcu_read_lock &ndev->lock irq_context: 0 (wq_completion)bond175 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &idev->mc_lock rcu_read_lock &dir->lock#2 irq_context: 0 (wq_completion)bond175 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &idev->mc_lock rcu_read_lock rcu_node_0 irq_context: 0 (wq_completion)bond175 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &idev->mc_lock rcu_read_lock &rq->__lock irq_context: 0 (wq_completion)bond175 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &idev->mc_lock rcu_read_lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)bond175 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &idev->mc_lock rcu_read_lock &ul->lock#2 irq_context: 0 (wq_completion)bond175 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &idev->mc_lock rcu_read_lock rcu_read_lock &____s->seqcount#10 irq_context: 0 (wq_completion)bond175 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &idev->mc_lock rcu_read_lock rcu_read_lock rcu_read_lock_bh rcu_read_lock &c->lock irq_context: 0 (wq_completion)bond175 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &idev->mc_lock rcu_read_lock rcu_read_lock rcu_read_lock_bh rcu_read_lock pool_lock#2 irq_context: 0 (wq_completion)bond175 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &idev->mc_lock rcu_read_lock rcu_read_lock rcu_read_lock_bh rcu_read_lock tk_core.seq.seqcount irq_context: 0 (wq_completion)bond175 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &idev->mc_lock rcu_read_lock rcu_read_lock rcu_read_lock_bh rcu_read_lock &obj_hash[i].lock irq_context: 0 (wq_completion)bond175 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &idev->mc_lock rcu_read_lock rcu_read_lock rcu_read_lock_bh rcu_read_lock rcu_read_lock_bh rcu_read_lock &c->lock irq_context: 0 (wq_completion)bond175 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &idev->mc_lock rcu_read_lock rcu_read_lock rcu_read_lock_bh rcu_read_lock rcu_read_lock_bh rcu_read_lock pool_lock#2 irq_context: 0 (wq_completion)bond175 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &idev->mc_lock rcu_read_lock rcu_read_lock rcu_read_lock_bh rcu_read_lock rcu_read_lock_bh rcu_read_lock tk_core.seq.seqcount irq_context: 0 (wq_completion)bond175 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &idev->mc_lock rcu_read_lock rcu_read_lock rcu_read_lock_bh rcu_read_lock rcu_read_lock_bh rcu_read_lock &obj_hash[i].lock irq_context: 0 (wq_completion)bond175 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &idev->mc_lock rcu_read_lock rcu_read_lock rcu_read_lock_bh rcu_read_lock rcu_read_lock_bh rcu_read_lock &br->multicast_lock irq_context: 0 (wq_completion)bond175 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &idev->mc_lock rcu_read_lock rcu_read_lock rcu_read_lock_bh rcu_read_lock rcu_read_lock_bh rcu_read_lock &br->multicast_lock &obj_hash[i].lock irq_context: 0 (wq_completion)bond175 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &idev->mc_lock rcu_read_lock rcu_read_lock rcu_read_lock_bh rcu_read_lock rcu_read_lock_bh rcu_read_lock &br->multicast_lock &base->lock irq_context: 0 (wq_completion)bond175 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &idev->mc_lock rcu_read_lock rcu_read_lock rcu_read_lock_bh rcu_read_lock rcu_read_lock_bh rcu_read_lock &br->multicast_lock &base->lock &obj_hash[i].lock irq_context: 0 (wq_completion)bond175 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &idev->mc_lock rcu_read_lock rcu_read_lock rcu_read_lock_bh rcu_read_lock rcu_read_lock_bh rcu_read_lock &data->lock irq_context: 0 (wq_completion)bond175 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &idev->mc_lock rcu_read_lock rcu_read_lock &rq->__lock irq_context: 0 (wq_completion)bond175 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &idev->mc_lock rcu_read_lock rcu_read_lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)bond175 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &idev->mc_lock rcu_read_lock &rcu_state.gp_wq irq_context: 0 (wq_completion)bond175 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &idev->mc_lock rcu_read_lock &rcu_state.gp_wq &p->pi_lock irq_context: 0 (wq_completion)bond175 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &idev->mc_lock rcu_read_lock &rcu_state.gp_wq &p->pi_lock &rq->__lock irq_context: 0 (wq_completion)bond175 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &idev->mc_lock rcu_read_lock &rcu_state.gp_wq &p->pi_lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)bond175 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex.wait_lock irq_context: 0 (wq_completion)bond175 (work_completion)(&(&bond->mcast_work)->work) &p->pi_lock irq_context: 0 (wq_completion)bond175 (work_completion)(&(&bond->mcast_work)->work) &p->pi_lock &rq->__lock irq_context: 0 (wq_completion)bond175 (work_completion)(&(&bond->mcast_work)->work) &p->pi_lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)bond172#3 irq_context: 0 (wq_completion)bond172#3 (work_completion)(&(&slave->notify_work)->work) irq_context: 0 (wq_completion)bond172#3 (work_completion)(&(&slave->notify_work)->work) &obj_hash[i].lock irq_context: 0 (wq_completion)bond172#3 (work_completion)(&(&slave->notify_work)->work) &base->lock irq_context: 0 (wq_completion)bond172#3 (work_completion)(&(&slave->notify_work)->work) &base->lock &obj_hash[i].lock irq_context: 0 (wq_completion)bond178 (work_completion)(&(&slave->notify_work)->work) &rq->__lock irq_context: 0 (wq_completion)bond178 (work_completion)(&(&slave->notify_work)->work) &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)ipv6_addrconf (work_completion)(&(&ifa->dad_work)->work) rtnl_mutex rcu_read_lock rcu_read_lock rcu_read_lock_bh rcu_read_lock rcu_read_lock_bh rcu_read_lock &c->lock irq_context: 0 (wq_completion)bond172#3 (work_completion)(&(&bond->mcast_work)->work) irq_context: 0 (wq_completion)bond172#3 (work_completion)(&(&bond->mcast_work)->work) &obj_hash[i].lock irq_context: 0 (wq_completion)bond172#3 (work_completion)(&(&bond->mcast_work)->work) &base->lock irq_context: 0 (wq_completion)bond172#3 (work_completion)(&(&bond->mcast_work)->work) &base->lock &obj_hash[i].lock irq_context: 0 (wq_completion)bond172#3 (work_completion)(&(&bond->mcast_work)->work) &rq->__lock irq_context: 0 (wq_completion)bond172#3 (work_completion)(&(&bond->mcast_work)->work) &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: softirq rcu_read_lock rcu_read_lock slock-AF_INET6/1 sk_lock-AF_INET6 rcu_read_lock rcu_read_lock rcu_read_lock rcu_read_lock_bh rcu_read_lock &n->list_lock irq_context: softirq rcu_read_lock rcu_read_lock slock-AF_INET6/1 sk_lock-AF_INET6 rcu_read_lock rcu_read_lock rcu_read_lock rcu_read_lock_bh rcu_read_lock &n->list_lock &c->lock irq_context: 0 (wq_completion)bond54#3 &rq->__lock irq_context: 0 (wq_completion)bond54#3 &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)mld (work_completion)(&(&idev->mc_ifc_work)->work) &idev->mc_lock rcu_read_lock rcu_read_lock rcu_read_lock_bh rcu_read_lock rcu_read_lock_bh rcu_read_lock &n->list_lock irq_context: 0 (wq_completion)mld (work_completion)(&(&idev->mc_ifc_work)->work) &idev->mc_lock rcu_read_lock rcu_read_lock rcu_read_lock_bh rcu_read_lock rcu_read_lock_bh rcu_read_lock &n->list_lock &c->lock irq_context: 0 (wq_completion)bond201 irq_context: 0 (wq_completion)bond201 (work_completion)(&(&slave->notify_work)->work) irq_context: 0 (wq_completion)bond201 (work_completion)(&(&slave->notify_work)->work) &obj_hash[i].lock irq_context: 0 (wq_completion)bond201 (work_completion)(&(&slave->notify_work)->work) &base->lock irq_context: 0 (wq_completion)bond201 (work_completion)(&(&slave->notify_work)->work) &base->lock &obj_hash[i].lock irq_context: 0 (wq_completion)bond201 (work_completion)(&(&slave->notify_work)->work) &rq->__lock irq_context: 0 (wq_completion)bond201 (work_completion)(&(&bond->mcast_work)->work) irq_context: 0 (wq_completion)bond201 (work_completion)(&(&bond->mcast_work)->work) &obj_hash[i].lock irq_context: 0 (wq_completion)bond201 (work_completion)(&(&bond->mcast_work)->work) &base->lock irq_context: 0 (wq_completion)bond201 (work_completion)(&(&bond->mcast_work)->work) &base->lock &obj_hash[i].lock irq_context: 0 (wq_completion)bond201 (work_completion)(&(&bond->mcast_work)->work) &rq->__lock irq_context: 0 (wq_completion)bond201 (work_completion)(&(&bond->mcast_work)->work) &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)bond189#2 irq_context: 0 (wq_completion)bond189#2 &rq->__lock irq_context: 0 (wq_completion)bond189#2 &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)bond189#2 (work_completion)(&(&slave->notify_work)->work) irq_context: 0 (wq_completion)bond189#2 (work_completion)(&(&slave->notify_work)->work) &obj_hash[i].lock irq_context: 0 (wq_completion)bond189#2 (work_completion)(&(&slave->notify_work)->work) &base->lock irq_context: 0 (wq_completion)bond189#2 (work_completion)(&(&slave->notify_work)->work) &base->lock &obj_hash[i].lock irq_context: 0 (wq_completion)bond189#2 (work_completion)(&(&slave->notify_work)->work) &rq->__lock irq_context: 0 (wq_completion)ipv6_addrconf (work_completion)(&(&ifa->dad_work)->work) rtnl_mutex rcu_read_lock rcu_read_lock rcu_read_lock_bh rcu_read_lock rcu_read_lock_bh rcu_read_lock &n->list_lock irq_context: 0 (wq_completion)ipv6_addrconf (work_completion)(&(&ifa->dad_work)->work) rtnl_mutex rcu_read_lock rcu_read_lock rcu_read_lock_bh rcu_read_lock rcu_read_lock_bh rcu_read_lock &n->list_lock &c->lock irq_context: 0 (wq_completion)bond189#2 (work_completion)(&(&bond->mcast_work)->work) irq_context: 0 (wq_completion)bond189#2 (work_completion)(&(&bond->mcast_work)->work) &rq->__lock irq_context: 0 (wq_completion)bond189#2 (work_completion)(&(&bond->mcast_work)->work) &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)bond189#2 (work_completion)(&(&bond->mcast_work)->work) &obj_hash[i].lock irq_context: 0 (wq_completion)bond189#2 (work_completion)(&(&bond->mcast_work)->work) &base->lock irq_context: 0 (wq_completion)bond189#2 (work_completion)(&(&bond->mcast_work)->work) &base->lock &obj_hash[i].lock irq_context: 0 (wq_completion)bond173#3 irq_context: 0 (wq_completion)bond173#3 (work_completion)(&(&slave->notify_work)->work) irq_context: 0 (wq_completion)bond173#3 (work_completion)(&(&slave->notify_work)->work) &obj_hash[i].lock irq_context: 0 (wq_completion)bond173#3 (work_completion)(&(&slave->notify_work)->work) &base->lock irq_context: 0 (wq_completion)bond173#3 (work_completion)(&(&slave->notify_work)->work) &base->lock &obj_hash[i].lock irq_context: 0 (wq_completion)bond173#3 (work_completion)(&(&slave->notify_work)->work) &rq->__lock irq_context: 0 (wq_completion)bond173#3 (work_completion)(&(&slave->notify_work)->work) &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)bond146#2 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex fs_reclaim irq_context: 0 (wq_completion)bond146#2 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex fs_reclaim &rq->__lock irq_context: 0 (wq_completion)bond146#2 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex fs_reclaim &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)bond146#2 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex fs_reclaim mmu_notifier_invalidate_range_start irq_context: 0 (wq_completion)bond146#2 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &c->lock irq_context: 0 (wq_completion)bond146#2 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex rcu_read_lock &bond->stats_lock/1 irq_context: 0 (wq_completion)bond146#2 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex rcu_read_lock &ndev->lock irq_context: 0 (wq_completion)bond146#2 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex nl_table_lock irq_context: 0 (wq_completion)bond146#2 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &obj_hash[i].lock irq_context: 0 (wq_completion)bond146#2 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex nl_table_wait.lock irq_context: 0 (wq_completion)bond146#2 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex net_rwsem irq_context: 0 (wq_completion)bond146#2 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex net_rwsem &list->lock#2 irq_context: 0 (wq_completion)bond146#2 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex net_rwsem &rq->__lock irq_context: 0 (wq_completion)bond146#2 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex net_rwsem &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)bond146#2 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &tn->lock irq_context: 0 (wq_completion)bond146#2 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &idev->mc_lock irq_context: 0 (wq_completion)bond146#2 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &idev->mc_lock fs_reclaim irq_context: 0 (wq_completion)bond146#2 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &idev->mc_lock fs_reclaim mmu_notifier_invalidate_range_start irq_context: 0 (wq_completion)bond146#2 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &idev->mc_lock rcu_read_lock &ndev->lock irq_context: 0 (wq_completion)bond146#2 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &idev->mc_lock rcu_read_lock &c->lock irq_context: 0 (wq_completion)bond146#2 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &idev->mc_lock rcu_read_lock &dir->lock#2 irq_context: 0 (wq_completion)bond146#2 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &idev->mc_lock rcu_read_lock &ul->lock#2 irq_context: 0 (wq_completion)bond146#2 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &idev->mc_lock rcu_read_lock rcu_read_lock &obj_hash[i].lock irq_context: 0 (wq_completion)bond146#2 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &idev->mc_lock rcu_read_lock rcu_read_lock rcu_node_0 irq_context: 0 (wq_completion)bond146#2 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &idev->mc_lock rcu_read_lock rcu_read_lock &rq->__lock irq_context: 0 (wq_completion)bond146#2 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &idev->mc_lock rcu_read_lock rcu_read_lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)bond146#2 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &idev->mc_lock rcu_read_lock rcu_read_lock &c->lock irq_context: 0 (wq_completion)bond146#2 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &idev->mc_lock rcu_read_lock rcu_read_lock &dir->lock#2 irq_context: 0 (wq_completion)bond146#2 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &idev->mc_lock rcu_read_lock rcu_read_lock &tbl->lock irq_context: 0 (wq_completion)bond146#2 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &idev->mc_lock rcu_read_lock rcu_read_lock &n->lock irq_context: 0 (wq_completion)bond146#2 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &idev->mc_lock rcu_read_lock rcu_read_lock &____s->seqcount#9 irq_context: 0 (wq_completion)bond146#2 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &idev->mc_lock rcu_read_lock rcu_read_lock rcu_read_lock_bh rcu_read_lock &c->lock irq_context: 0 (wq_completion)bond146#2 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &idev->mc_lock rcu_read_lock rcu_read_lock rcu_read_lock_bh rcu_read_lock &n->list_lock irq_context: 0 (wq_completion)bond146#2 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &idev->mc_lock rcu_read_lock rcu_read_lock rcu_read_lock_bh rcu_read_lock &n->list_lock &c->lock irq_context: 0 (wq_completion)bond146#2 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &idev->mc_lock rcu_read_lock rcu_read_lock rcu_read_lock_bh rcu_read_lock pool_lock#2 irq_context: 0 (wq_completion)bond146#2 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &idev->mc_lock rcu_read_lock rcu_read_lock rcu_read_lock_bh rcu_read_lock tk_core.seq.seqcount irq_context: 0 (wq_completion)bond146#2 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &idev->mc_lock rcu_read_lock rcu_read_lock rcu_read_lock_bh rcu_read_lock &obj_hash[i].lock irq_context: 0 (wq_completion)bond146#2 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &idev->mc_lock rcu_read_lock rcu_read_lock rcu_read_lock_bh rcu_read_lock rcu_read_lock_bh rcu_read_lock pool_lock#2 irq_context: 0 (wq_completion)bond146#2 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &idev->mc_lock rcu_read_lock rcu_read_lock rcu_read_lock_bh rcu_read_lock rcu_read_lock_bh rcu_read_lock tk_core.seq.seqcount irq_context: 0 (wq_completion)bond146#2 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &idev->mc_lock rcu_read_lock rcu_read_lock rcu_read_lock_bh rcu_read_lock rcu_read_lock_bh rcu_read_lock &obj_hash[i].lock irq_context: 0 (wq_completion)bond146#2 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &idev->mc_lock rcu_read_lock rcu_read_lock rcu_read_lock_bh rcu_read_lock rcu_read_lock_bh rcu_read_lock &br->multicast_lock irq_context: 0 (wq_completion)bond146#2 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &idev->mc_lock rcu_read_lock rcu_read_lock rcu_read_lock_bh rcu_read_lock rcu_read_lock_bh rcu_read_lock &br->multicast_lock &obj_hash[i].lock irq_context: 0 (wq_completion)bond146#2 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &idev->mc_lock rcu_read_lock rcu_read_lock rcu_read_lock_bh rcu_read_lock rcu_read_lock_bh rcu_read_lock &br->multicast_lock &base->lock irq_context: 0 (wq_completion)bond146#2 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &idev->mc_lock rcu_read_lock rcu_read_lock rcu_read_lock_bh rcu_read_lock rcu_read_lock_bh rcu_read_lock &br->multicast_lock &base->lock &obj_hash[i].lock irq_context: 0 (wq_completion)bond146#2 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &idev->mc_lock rcu_read_lock rcu_read_lock rcu_read_lock_bh rcu_read_lock rcu_read_lock_bh rcu_read_lock &data->lock irq_context: 0 (wq_completion)bond146#2 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &idev->mc_lock rcu_read_lock rcu_node_0 irq_context: 0 (wq_completion)bond146#2 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &rq->__lock irq_context: 0 (wq_completion)bond146#2 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)bond173#3 (work_completion)(&(&bond->mcast_work)->work) irq_context: 0 (wq_completion)bond173#3 (work_completion)(&(&bond->mcast_work)->work) &obj_hash[i].lock irq_context: 0 (wq_completion)bond173#3 (work_completion)(&(&bond->mcast_work)->work) &base->lock irq_context: 0 (wq_completion)bond173#3 (work_completion)(&(&bond->mcast_work)->work) &base->lock &obj_hash[i].lock irq_context: 0 (wq_completion)bond146#2 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex.wait_lock irq_context: 0 (wq_completion)bond146#2 (work_completion)(&(&bond->mcast_work)->work) &p->pi_lock irq_context: 0 (wq_completion)bond146#2 (work_completion)(&(&bond->mcast_work)->work) &p->pi_lock &cfs_rq->removed.lock irq_context: 0 (wq_completion)bond146#2 (work_completion)(&(&bond->mcast_work)->work) &p->pi_lock &rq->__lock irq_context: 0 (wq_completion)bond146#2 (work_completion)(&(&bond->mcast_work)->work) &p->pi_lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)bond176#2 (work_completion)(&(&slave->notify_work)->work) rtnl_mutex net_rwsem irq_context: 0 (wq_completion)bond176#2 (work_completion)(&(&slave->notify_work)->work) rtnl_mutex net_rwsem &list->lock#2 irq_context: 0 (wq_completion)bond176#2 (work_completion)(&(&slave->notify_work)->work) rtnl_mutex &tn->lock irq_context: 0 (wq_completion)bond176#2 (work_completion)(&(&slave->notify_work)->work) rtnl_mutex &tn->lock &rq->__lock irq_context: 0 (wq_completion)bond176#2 (work_completion)(&(&slave->notify_work)->work) rtnl_mutex &tn->lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)bond176#2 (work_completion)(&(&slave->notify_work)->work) rtnl_mutex.wait_lock irq_context: 0 (wq_completion)bond176#2 (work_completion)(&(&slave->notify_work)->work) &p->pi_lock irq_context: 0 (wq_completion)bond176#2 (work_completion)(&(&slave->notify_work)->work) &p->pi_lock &rq->__lock irq_context: 0 (wq_completion)bond176#2 (work_completion)(&(&slave->notify_work)->work) &p->pi_lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)bond166#3 (work_completion)(&(&bond->mcast_work)->work) &rq->__lock irq_context: 0 (wq_completion)bond166#3 (work_completion)(&(&bond->mcast_work)->work) &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)bond202 irq_context: 0 (wq_completion)bond202 (work_completion)(&(&slave->notify_work)->work) irq_context: 0 (wq_completion)bond202 (work_completion)(&(&slave->notify_work)->work) &obj_hash[i].lock irq_context: 0 (wq_completion)bond202 (work_completion)(&(&slave->notify_work)->work) &base->lock irq_context: 0 (wq_completion)bond202 (work_completion)(&(&slave->notify_work)->work) &base->lock &obj_hash[i].lock irq_context: 0 (wq_completion)bond202 (work_completion)(&(&slave->notify_work)->work) &rq->__lock irq_context: 0 (wq_completion)bond202 (work_completion)(&(&slave->notify_work)->work) &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)bond202 (work_completion)(&(&bond->mcast_work)->work) irq_context: 0 (wq_completion)bond202 (work_completion)(&(&bond->mcast_work)->work) &obj_hash[i].lock irq_context: 0 (wq_completion)bond202 (work_completion)(&(&bond->mcast_work)->work) &base->lock irq_context: 0 (wq_completion)bond202 (work_completion)(&(&bond->mcast_work)->work) &base->lock &obj_hash[i].lock irq_context: 0 (wq_completion)bond202 (work_completion)(&(&bond->mcast_work)->work) &rq->__lock irq_context: 0 (wq_completion)bond202 (work_completion)(&(&bond->mcast_work)->work) &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)bond190#2 irq_context: 0 (wq_completion)bond190#2 &rq->__lock irq_context: 0 (wq_completion)bond190#2 (work_completion)(&(&slave->notify_work)->work) irq_context: 0 (wq_completion)bond190#2 (work_completion)(&(&slave->notify_work)->work) &obj_hash[i].lock irq_context: 0 (wq_completion)bond190#2 (work_completion)(&(&slave->notify_work)->work) &base->lock irq_context: 0 (wq_completion)bond190#2 (work_completion)(&(&slave->notify_work)->work) &base->lock &obj_hash[i].lock irq_context: 0 (wq_completion)bond190#2 (work_completion)(&(&slave->notify_work)->work) &rq->__lock irq_context: 0 (wq_completion)bond156#2 &rq->__lock irq_context: 0 (wq_completion)bond156#2 &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)bond181 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex fs_reclaim irq_context: 0 (wq_completion)bond181 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex fs_reclaim mmu_notifier_invalidate_range_start irq_context: 0 (wq_completion)bond181 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex fs_reclaim mmu_notifier_invalidate_range_start &rq->__lock irq_context: 0 (wq_completion)bond181 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex fs_reclaim mmu_notifier_invalidate_range_start &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)bond181 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex rcu_read_lock &bond->stats_lock/1 irq_context: 0 (wq_completion)bond181 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex rcu_read_lock &ndev->lock irq_context: 0 (wq_completion)bond181 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex nl_table_lock irq_context: 0 (wq_completion)bond181 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &obj_hash[i].lock irq_context: 0 (wq_completion)bond181 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex nl_table_wait.lock irq_context: 0 (wq_completion)bond181 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex net_rwsem irq_context: 0 (wq_completion)bond181 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex net_rwsem &list->lock#2 irq_context: 0 (wq_completion)bond181 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &tn->lock irq_context: 0 (wq_completion)bond181 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &idev->mc_lock irq_context: 0 (wq_completion)bond181 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &idev->mc_lock fs_reclaim irq_context: 0 (wq_completion)bond181 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &idev->mc_lock fs_reclaim &rq->__lock irq_context: 0 (wq_completion)bond181 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &idev->mc_lock fs_reclaim &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)bond190#2 (work_completion)(&(&bond->mcast_work)->work) irq_context: 0 (wq_completion)bond190#2 (work_completion)(&(&bond->mcast_work)->work) &obj_hash[i].lock irq_context: 0 (wq_completion)bond190#2 (work_completion)(&(&bond->mcast_work)->work) &base->lock irq_context: 0 (wq_completion)bond190#2 (work_completion)(&(&bond->mcast_work)->work) &base->lock &obj_hash[i].lock irq_context: 0 (wq_completion)bond181 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &idev->mc_lock fs_reclaim mmu_notifier_invalidate_range_start irq_context: 0 (wq_completion)bond181 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &idev->mc_lock rcu_read_lock &ndev->lock irq_context: 0 (wq_completion)bond181 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &idev->mc_lock rcu_read_lock &c->lock irq_context: 0 (wq_completion)bond181 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &idev->mc_lock rcu_read_lock &dir->lock#2 irq_context: 0 (wq_completion)bond181 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &idev->mc_lock rcu_read_lock &ul->lock#2 irq_context: 0 (wq_completion)bond181 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &idev->mc_lock rcu_read_lock rcu_read_lock &____s->seqcount#10 irq_context: 0 (wq_completion)bond181 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &idev->mc_lock rcu_read_lock rcu_read_lock rcu_read_lock_bh rcu_read_lock rcu_read_lock_bh rcu_read_lock &br->multicast_lock irq_context: 0 (wq_completion)bond181 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &idev->mc_lock rcu_read_lock rcu_read_lock rcu_read_lock_bh rcu_read_lock rcu_read_lock_bh rcu_read_lock &br->multicast_lock &obj_hash[i].lock irq_context: 0 (wq_completion)bond181 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &idev->mc_lock rcu_read_lock rcu_read_lock rcu_read_lock_bh rcu_read_lock rcu_read_lock_bh rcu_read_lock &br->multicast_lock &base->lock irq_context: 0 (wq_completion)bond181 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &idev->mc_lock rcu_read_lock rcu_read_lock rcu_read_lock_bh rcu_read_lock rcu_read_lock_bh rcu_read_lock &br->multicast_lock &base->lock &obj_hash[i].lock irq_context: 0 (wq_completion)bond181 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &idev->mc_lock rcu_read_lock rcu_read_lock rcu_read_lock_bh rcu_read_lock rcu_read_lock_bh rcu_read_lock &data->lock irq_context: 0 (wq_completion)bond181 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &idev->mc_lock rcu_read_lock rcu_read_lock rcu_read_lock_bh rcu_read_lock rcu_read_lock_bh rcu_read_lock &obj_hash[i].lock irq_context: 0 (wq_completion)bond181 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &idev->mc_lock rcu_read_lock rcu_read_lock rcu_node_0 irq_context: 0 (wq_completion)bond181 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &idev->mc_lock rcu_read_lock rcu_read_lock &rq->__lock irq_context: 0 (wq_completion)bond181 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &idev->mc_lock rcu_read_lock rcu_read_lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)bond181 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &idev->mc_lock rcu_read_lock rcu_node_0 irq_context: 0 (wq_completion)bond181 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex.wait_lock irq_context: 0 (wq_completion)bond181 (work_completion)(&(&bond->mcast_work)->work) &p->pi_lock irq_context: 0 (wq_completion)bond181 (work_completion)(&(&bond->mcast_work)->work) &p->pi_lock &rq->__lock irq_context: 0 (wq_completion)bond181 (work_completion)(&(&bond->mcast_work)->work) &p->pi_lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)bond174#3 irq_context: 0 (wq_completion)bond174#3 (work_completion)(&(&slave->notify_work)->work) irq_context: 0 (wq_completion)bond174#3 (work_completion)(&(&slave->notify_work)->work) &obj_hash[i].lock irq_context: 0 (wq_completion)bond174#3 (work_completion)(&(&slave->notify_work)->work) &base->lock irq_context: 0 (wq_completion)bond174#3 (work_completion)(&(&slave->notify_work)->work) &base->lock &obj_hash[i].lock irq_context: 0 (wq_completion)bond174#3 (work_completion)(&(&slave->notify_work)->work) &rq->__lock irq_context: 0 (wq_completion)bond174#3 (work_completion)(&(&slave->notify_work)->work) &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)bond174#3 (work_completion)(&(&bond->mcast_work)->work) irq_context: 0 (wq_completion)bond174#3 (work_completion)(&(&bond->mcast_work)->work) &obj_hash[i].lock irq_context: 0 (wq_completion)bond174#3 (work_completion)(&(&bond->mcast_work)->work) &base->lock irq_context: 0 (wq_completion)bond174#3 (work_completion)(&(&bond->mcast_work)->work) &base->lock &obj_hash[i].lock irq_context: 0 (wq_completion)bond179#2 (work_completion)(&(&bond->mcast_work)->work) &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)bond116#2 (work_completion)(&(&slave->notify_work)->work) rtnl_mutex net_rwsem irq_context: 0 (wq_completion)bond116#2 (work_completion)(&(&slave->notify_work)->work) rtnl_mutex net_rwsem &list->lock#2 irq_context: 0 (wq_completion)bond116#2 (work_completion)(&(&slave->notify_work)->work) rtnl_mutex &tn->lock irq_context: 0 (wq_completion)bond116#2 (work_completion)(&(&slave->notify_work)->work) rtnl_mutex.wait_lock irq_context: 0 (wq_completion)bond116#2 (work_completion)(&(&slave->notify_work)->work) &p->pi_lock irq_context: 0 (wq_completion)bond166#2 (work_completion)(&(&slave->notify_work)->work) rtnl_mutex net_rwsem irq_context: 0 (wq_completion)bond166#2 (work_completion)(&(&slave->notify_work)->work) rtnl_mutex net_rwsem &list->lock#2 irq_context: 0 (wq_completion)bond166#2 (work_completion)(&(&slave->notify_work)->work) rtnl_mutex &tn->lock irq_context: 0 (wq_completion)bond166#2 (work_completion)(&(&slave->notify_work)->work) rtnl_mutex rcu_read_lock rcu_node_0 irq_context: 0 (wq_completion)bond166#2 (work_completion)(&(&slave->notify_work)->work) rtnl_mutex rcu_read_lock &rq->__lock irq_context: 0 (wq_completion)bond166#2 (work_completion)(&(&slave->notify_work)->work) rtnl_mutex rcu_read_lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)bond166#2 (work_completion)(&(&slave->notify_work)->work) rtnl_mutex.wait_lock irq_context: 0 (wq_completion)bond44#2 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex fs_reclaim irq_context: 0 (wq_completion)bond166#2 (work_completion)(&(&slave->notify_work)->work) &p->pi_lock irq_context: 0 (wq_completion)bond44#2 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex fs_reclaim mmu_notifier_invalidate_range_start irq_context: 0 (wq_completion)bond44#2 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex remove_cache_srcu irq_context: 0 (wq_completion)bond44#2 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex remove_cache_srcu quarantine_lock irq_context: 0 (wq_completion)bond44#2 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex remove_cache_srcu &c->lock irq_context: 0 (wq_completion)bond44#2 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex remove_cache_srcu &n->list_lock irq_context: 0 (wq_completion)bond44#2 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex remove_cache_srcu &obj_hash[i].lock irq_context: 0 (wq_completion)bond44#2 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex rcu_read_lock &bond->stats_lock/1 irq_context: 0 (wq_completion)bond44#2 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex rcu_read_lock &ndev->lock irq_context: 0 (wq_completion)bond44#2 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex nl_table_lock irq_context: 0 (wq_completion)bond44#2 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &obj_hash[i].lock irq_context: 0 (wq_completion)bond44#2 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex nl_table_wait.lock irq_context: 0 (wq_completion)bond44#2 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex net_rwsem irq_context: 0 (wq_completion)bond44#2 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex net_rwsem &list->lock#2 irq_context: 0 (wq_completion)bond44#2 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &tn->lock irq_context: 0 (wq_completion)bond44#2 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &idev->mc_lock irq_context: 0 (wq_completion)bond44#2 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &idev->mc_lock fs_reclaim irq_context: 0 (wq_completion)bond44#2 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &idev->mc_lock fs_reclaim mmu_notifier_invalidate_range_start irq_context: 0 (wq_completion)bond44#2 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &idev->mc_lock rcu_read_lock &ndev->lock irq_context: 0 (wq_completion)bond44#2 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &idev->mc_lock rcu_read_lock &c->lock irq_context: 0 (wq_completion)bond44#2 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &idev->mc_lock rcu_read_lock &dir->lock#2 irq_context: 0 (wq_completion)bond44#2 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &idev->mc_lock rcu_read_lock &ul->lock#2 irq_context: 0 (wq_completion)bond44#2 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &idev->mc_lock rcu_read_lock rcu_read_lock &____s->seqcount#12 irq_context: 0 (wq_completion)bond44#2 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &idev->mc_lock rcu_read_lock rcu_read_lock &____s->seqcount#16 irq_context: 0 (wq_completion)bond44#2 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &idev->mc_lock rcu_read_lock &obj_hash[i].lock irq_context: 0 (wq_completion)bond44#2 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &idev->mc_lock rcu_read_lock &data->lock irq_context: 0 (wq_completion)bond44#2 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex.wait_lock irq_context: 0 (wq_completion)bond15#3 (work_completion)(&(&slave->notify_work)->work) rtnl_mutex net_rwsem irq_context: 0 (wq_completion)bond15#3 (work_completion)(&(&slave->notify_work)->work) rtnl_mutex net_rwsem &list->lock#2 irq_context: 0 (wq_completion)bond15#3 (work_completion)(&(&slave->notify_work)->work) rtnl_mutex &tn->lock irq_context: 0 (wq_completion)bond15#3 (work_completion)(&(&slave->notify_work)->work) rtnl_mutex &tn->lock &rq->__lock irq_context: 0 (wq_completion)bond15#3 (work_completion)(&(&slave->notify_work)->work) rtnl_mutex &tn->lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)bond15#3 (work_completion)(&(&slave->notify_work)->work) rtnl_mutex.wait_lock irq_context: 0 (wq_completion)bond44#2 (work_completion)(&(&bond->mcast_work)->work) &p->pi_lock irq_context: 0 (wq_completion)bond119#2 (work_completion)(&(&bond->mcast_work)->work) &base->lock fill_pool_map-wait-type-override pool_lock irq_context: 0 (wq_completion)bond203 irq_context: 0 (wq_completion)bond203 (work_completion)(&(&slave->notify_work)->work) irq_context: 0 (wq_completion)bond203 (work_completion)(&(&slave->notify_work)->work) &obj_hash[i].lock irq_context: 0 (wq_completion)bond203 (work_completion)(&(&slave->notify_work)->work) &base->lock irq_context: 0 (wq_completion)bond203 (work_completion)(&(&slave->notify_work)->work) &base->lock &obj_hash[i].lock irq_context: 0 (wq_completion)bond165#3 (work_completion)(&(&slave->notify_work)->work) &rq->__lock irq_context: 0 (wq_completion)bond165#3 (work_completion)(&(&slave->notify_work)->work) &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)bond203 (work_completion)(&(&bond->mcast_work)->work) irq_context: 0 (wq_completion)bond203 (work_completion)(&(&bond->mcast_work)->work) &obj_hash[i].lock irq_context: 0 (wq_completion)bond203 (work_completion)(&(&bond->mcast_work)->work) &base->lock irq_context: 0 (wq_completion)bond203 (work_completion)(&(&bond->mcast_work)->work) &base->lock &obj_hash[i].lock irq_context: 0 (wq_completion)bond191#2 irq_context: 0 (wq_completion)bond191#2 (work_completion)(&(&slave->notify_work)->work) irq_context: 0 (wq_completion)bond191#2 (work_completion)(&(&slave->notify_work)->work) &obj_hash[i].lock irq_context: 0 (wq_completion)bond191#2 (work_completion)(&(&slave->notify_work)->work) &base->lock irq_context: 0 (wq_completion)bond191#2 (work_completion)(&(&slave->notify_work)->work) &base->lock &obj_hash[i].lock irq_context: 0 (wq_completion)bond191#2 (work_completion)(&(&bond->mcast_work)->work) irq_context: 0 (wq_completion)bond191#2 (work_completion)(&(&bond->mcast_work)->work) &obj_hash[i].lock irq_context: 0 (wq_completion)bond191#2 (work_completion)(&(&bond->mcast_work)->work) &base->lock irq_context: 0 (wq_completion)bond191#2 (work_completion)(&(&bond->mcast_work)->work) &base->lock &obj_hash[i].lock irq_context: 0 (wq_completion)bond191 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex fs_reclaim irq_context: 0 (wq_completion)bond191 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex fs_reclaim mmu_notifier_invalidate_range_start irq_context: 0 (wq_completion)bond191 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &c->lock irq_context: 0 (wq_completion)bond191 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex rcu_read_lock &bond->stats_lock/1 irq_context: 0 (wq_completion)bond191 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex rcu_read_lock &ndev->lock irq_context: 0 (wq_completion)bond191 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex nl_table_lock irq_context: 0 (wq_completion)bond191 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &obj_hash[i].lock irq_context: 0 (wq_completion)bond191 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex nl_table_wait.lock irq_context: 0 (wq_completion)bond191 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex net_rwsem irq_context: 0 (wq_completion)bond191 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex net_rwsem &list->lock#2 irq_context: 0 (wq_completion)bond191 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &tn->lock irq_context: 0 (wq_completion)bond191 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &idev->mc_lock irq_context: 0 (wq_completion)bond191 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &idev->mc_lock fs_reclaim irq_context: 0 (wq_completion)bond191 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &idev->mc_lock fs_reclaim mmu_notifier_invalidate_range_start irq_context: 0 (wq_completion)bond191 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &idev->mc_lock &c->lock irq_context: 0 (wq_completion)bond191 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &idev->mc_lock rcu_read_lock &ndev->lock irq_context: 0 (wq_completion)bond191 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &idev->mc_lock rcu_read_lock &dir->lock#2 irq_context: 0 (wq_completion)bond191 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &idev->mc_lock rcu_read_lock &ul->lock#2 irq_context: 0 (wq_completion)bond191 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &idev->mc_lock rcu_read_lock rcu_read_lock &____s->seqcount#10 irq_context: 0 (wq_completion)bond191 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &idev->mc_lock rcu_read_lock rcu_read_lock rcu_read_lock_bh rcu_read_lock rcu_read_lock_bh rcu_read_lock &br->multicast_lock irq_context: 0 (wq_completion)bond191 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &idev->mc_lock rcu_read_lock rcu_read_lock rcu_read_lock_bh rcu_read_lock rcu_read_lock_bh rcu_read_lock &br->multicast_lock &obj_hash[i].lock irq_context: 0 (wq_completion)bond191 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &idev->mc_lock rcu_read_lock rcu_read_lock rcu_read_lock_bh rcu_read_lock rcu_read_lock_bh rcu_read_lock &br->multicast_lock &base->lock irq_context: 0 (wq_completion)bond191 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &idev->mc_lock rcu_read_lock rcu_read_lock rcu_read_lock_bh rcu_read_lock rcu_read_lock_bh rcu_read_lock &data->lock irq_context: 0 (wq_completion)bond191 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &idev->mc_lock rcu_read_lock rcu_read_lock rcu_read_lock_bh rcu_read_lock rcu_read_lock_bh rcu_read_lock &obj_hash[i].lock irq_context: 0 (wq_completion)bond150#3 (work_completion)(&(&slave->notify_work)->work) rtnl_mutex net_rwsem irq_context: 0 (wq_completion)bond150#3 (work_completion)(&(&slave->notify_work)->work) rtnl_mutex net_rwsem &list->lock#2 irq_context: 0 (wq_completion)bond150#3 (work_completion)(&(&slave->notify_work)->work) rtnl_mutex &tn->lock irq_context: 0 (wq_completion)bond199 (work_completion)(&(&slave->notify_work)->work) rtnl_mutex net_rwsem irq_context: 0 (wq_completion)bond199 (work_completion)(&(&slave->notify_work)->work) rtnl_mutex net_rwsem &list->lock#2 irq_context: 0 (wq_completion)bond199 (work_completion)(&(&slave->notify_work)->work) rtnl_mutex &tn->lock irq_context: 0 (wq_completion)bond170#3 (work_completion)(&(&slave->notify_work)->work) rtnl_mutex net_rwsem irq_context: 0 (wq_completion)bond170#3 (work_completion)(&(&slave->notify_work)->work) rtnl_mutex net_rwsem &list->lock#2 irq_context: 0 (wq_completion)bond170#3 (work_completion)(&(&slave->notify_work)->work) rtnl_mutex &tn->lock irq_context: 0 (wq_completion)bond202 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex fs_reclaim irq_context: 0 (wq_completion)bond202 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex fs_reclaim mmu_notifier_invalidate_range_start irq_context: 0 (wq_completion)bond202 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &c->lock irq_context: 0 (wq_completion)bond202 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &n->list_lock irq_context: 0 (wq_completion)bond202 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &n->list_lock &c->lock irq_context: 0 (wq_completion)bond202 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex rcu_read_lock &bond->stats_lock/1 irq_context: 0 (wq_completion)bond202 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex rcu_read_lock &ndev->lock irq_context: 0 (wq_completion)bond202 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex nl_table_lock irq_context: 0 (wq_completion)bond202 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &obj_hash[i].lock irq_context: 0 (wq_completion)bond202 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex nl_table_wait.lock irq_context: 0 (wq_completion)bond202 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex net_rwsem irq_context: 0 (wq_completion)bond202 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex net_rwsem &list->lock#2 irq_context: 0 (wq_completion)bond202 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &tn->lock irq_context: 0 (wq_completion)bond202 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &idev->mc_lock irq_context: 0 (wq_completion)bond193 (work_completion)(&(&slave->notify_work)->work) rtnl_mutex net_rwsem irq_context: 0 (wq_completion)bond193 (work_completion)(&(&slave->notify_work)->work) rtnl_mutex net_rwsem &list->lock#2 irq_context: 0 (wq_completion)bond193 (work_completion)(&(&slave->notify_work)->work) rtnl_mutex &tn->lock irq_context: 0 (wq_completion)bond171 (work_completion)(&(&slave->notify_work)->work) rtnl_mutex net_rwsem irq_context: 0 (wq_completion)bond171 (work_completion)(&(&slave->notify_work)->work) rtnl_mutex net_rwsem &list->lock#2 irq_context: 0 (wq_completion)bond171 (work_completion)(&(&slave->notify_work)->work) rtnl_mutex &tn->lock irq_context: 0 (wq_completion)bond166 (work_completion)(&(&slave->notify_work)->work) rtnl_mutex net_rwsem irq_context: 0 (wq_completion)bond166 (work_completion)(&(&slave->notify_work)->work) rtnl_mutex net_rwsem &list->lock#2 irq_context: 0 (wq_completion)bond166 (work_completion)(&(&slave->notify_work)->work) rtnl_mutex &tn->lock irq_context: 0 (wq_completion)bond174#3 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex fs_reclaim irq_context: 0 (wq_completion)bond174#3 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex fs_reclaim mmu_notifier_invalidate_range_start irq_context: 0 (wq_completion)bond174#3 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &c->lock irq_context: 0 (wq_completion)bond174#3 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex rcu_read_lock &bond->stats_lock/1 irq_context: 0 (wq_completion)bond174#3 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex rcu_read_lock &ndev->lock irq_context: 0 (wq_completion)bond174#3 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex nl_table_lock irq_context: 0 (wq_completion)bond174#3 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &obj_hash[i].lock irq_context: 0 (wq_completion)bond174#3 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex nl_table_wait.lock irq_context: 0 (wq_completion)bond174#3 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex net_rwsem irq_context: 0 (wq_completion)bond174#3 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex net_rwsem &list->lock#2 irq_context: 0 (wq_completion)bond174#3 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &tn->lock irq_context: 0 (wq_completion)bond174#3 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &idev->mc_lock irq_context: 0 (wq_completion)bond174#3 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &idev->mc_lock fs_reclaim irq_context: 0 (wq_completion)bond174#3 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &idev->mc_lock fs_reclaim mmu_notifier_invalidate_range_start irq_context: 0 (wq_completion)bond174#3 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &idev->mc_lock rcu_read_lock &ndev->lock irq_context: 0 (wq_completion)bond174#3 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &idev->mc_lock rcu_read_lock &dir->lock#2 irq_context: 0 (wq_completion)bond174#3 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &idev->mc_lock rcu_read_lock &ul->lock#2 irq_context: 0 (wq_completion)bond174#3 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &idev->mc_lock rcu_read_lock rcu_read_lock &____s->seqcount#10 irq_context: 0 (wq_completion)bond174#3 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &idev->mc_lock rcu_read_lock rcu_read_lock rcu_read_lock_bh rcu_read_lock rcu_read_lock_bh rcu_read_lock &br->multicast_lock irq_context: 0 (wq_completion)bond174#3 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &idev->mc_lock rcu_read_lock rcu_read_lock rcu_read_lock_bh rcu_read_lock rcu_read_lock_bh rcu_read_lock &br->multicast_lock &obj_hash[i].lock irq_context: 0 (wq_completion)bond174#3 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &idev->mc_lock rcu_read_lock rcu_read_lock rcu_read_lock_bh rcu_read_lock rcu_read_lock_bh rcu_read_lock &br->multicast_lock &base->lock irq_context: 0 (wq_completion)bond174#3 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &idev->mc_lock rcu_read_lock rcu_read_lock rcu_read_lock_bh rcu_read_lock rcu_read_lock_bh rcu_read_lock &data->lock irq_context: 0 (wq_completion)bond174#3 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &idev->mc_lock rcu_read_lock rcu_read_lock rcu_read_lock_bh rcu_read_lock rcu_read_lock_bh rcu_read_lock &obj_hash[i].lock irq_context: 0 (wq_completion)bond174#3 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &rq->__lock irq_context: 0 (wq_completion)bond174#3 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex.wait_lock irq_context: 0 (wq_completion)bond174#3 (work_completion)(&(&bond->mcast_work)->work) &p->pi_lock irq_context: 0 (wq_completion)bond174#3 (work_completion)(&(&bond->mcast_work)->work) &p->pi_lock &rq->__lock irq_context: 0 (wq_completion)bond174#3 (work_completion)(&(&bond->mcast_work)->work) &p->pi_lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)bond174#3 (work_completion)(&(&bond->mcast_work)->work) &rq->__lock irq_context: 0 (wq_completion)bond203 (work_completion)(&(&slave->notify_work)->work) rtnl_mutex net_rwsem irq_context: 0 (wq_completion)bond203 (work_completion)(&(&slave->notify_work)->work) rtnl_mutex net_rwsem &list->lock#2 irq_context: 0 (wq_completion)bond203 (work_completion)(&(&slave->notify_work)->work) rtnl_mutex &tn->lock irq_context: 0 (wq_completion)bond203 (work_completion)(&(&slave->notify_work)->work) rtnl_mutex.wait_lock irq_context: 0 (wq_completion)bond203 (work_completion)(&(&slave->notify_work)->work) &p->pi_lock irq_context: 0 (wq_completion)bond191#2 (work_completion)(&(&slave->notify_work)->work) rtnl_mutex net_rwsem irq_context: 0 (wq_completion)bond191#2 (work_completion)(&(&slave->notify_work)->work) rtnl_mutex net_rwsem &list->lock#2 irq_context: 0 (wq_completion)bond191#2 (work_completion)(&(&slave->notify_work)->work) rtnl_mutex &tn->lock irq_context: 0 (wq_completion)bond191#2 (work_completion)(&(&slave->notify_work)->work) rtnl_mutex.wait_lock irq_context: 0 (wq_completion)bond191#2 (work_completion)(&(&slave->notify_work)->work) &p->pi_lock irq_context: 0 (wq_completion)bond167#3 (work_completion)(&(&slave->notify_work)->work) rtnl_mutex net_rwsem irq_context: 0 (wq_completion)bond167#3 (work_completion)(&(&slave->notify_work)->work) rtnl_mutex net_rwsem &list->lock#2 irq_context: 0 (wq_completion)bond167#3 (work_completion)(&(&slave->notify_work)->work) rtnl_mutex &tn->lock irq_context: 0 (wq_completion)bond167#3 (work_completion)(&(&slave->notify_work)->work) rtnl_mutex.wait_lock irq_context: 0 (wq_completion)bond167#3 (work_completion)(&(&slave->notify_work)->work) &p->pi_lock irq_context: 0 (wq_completion)bond159#2 (work_completion)(&(&slave->notify_work)->work) rtnl_mutex net_rwsem irq_context: 0 (wq_completion)bond159#2 (work_completion)(&(&slave->notify_work)->work) rtnl_mutex net_rwsem &list->lock#2 irq_context: 0 (wq_completion)bond159#2 (work_completion)(&(&slave->notify_work)->work) rtnl_mutex &tn->lock irq_context: 0 (wq_completion)bond159#2 (work_completion)(&(&slave->notify_work)->work) rtnl_mutex.wait_lock irq_context: 0 (wq_completion)bond159#2 (work_completion)(&(&slave->notify_work)->work) &p->pi_lock irq_context: 0 (wq_completion)bond186#2 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex fs_reclaim irq_context: 0 (wq_completion)bond186#2 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex fs_reclaim mmu_notifier_invalidate_range_start irq_context: 0 (wq_completion)bond186#2 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex rcu_read_lock &bond->stats_lock/1 irq_context: 0 (wq_completion)bond186#2 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex rcu_read_lock &ndev->lock irq_context: 0 (wq_completion)bond186#2 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex nl_table_lock irq_context: 0 (wq_completion)bond186#2 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &obj_hash[i].lock irq_context: 0 (wq_completion)bond186#2 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex nl_table_wait.lock irq_context: 0 (wq_completion)bond186#2 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex net_rwsem irq_context: 0 (wq_completion)bond186#2 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex net_rwsem &list->lock#2 irq_context: 0 (wq_completion)bond186#2 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &tn->lock irq_context: 0 (wq_completion)bond186#2 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &idev->mc_lock irq_context: 0 (wq_completion)bond186#2 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &idev->mc_lock fs_reclaim irq_context: 0 (wq_completion)bond186#2 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &idev->mc_lock fs_reclaim mmu_notifier_invalidate_range_start irq_context: 0 (wq_completion)bond186#2 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &idev->mc_lock &c->lock irq_context: 0 (wq_completion)bond186#2 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &idev->mc_lock rcu_read_lock &ndev->lock irq_context: 0 (wq_completion)bond186#2 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &idev->mc_lock rcu_read_lock &c->lock irq_context: 0 (wq_completion)bond186#2 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &idev->mc_lock rcu_read_lock &dir->lock#2 irq_context: 0 (wq_completion)bond186#2 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &idev->mc_lock rcu_read_lock &ul->lock#2 irq_context: 0 (wq_completion)bond186#2 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &idev->mc_lock rcu_read_lock rcu_read_lock &____s->seqcount#10 irq_context: 0 (wq_completion)bond186#2 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &idev->mc_lock rcu_read_lock rcu_read_lock rcu_read_lock_bh rcu_read_lock rcu_read_lock_bh rcu_read_lock &br->multicast_lock irq_context: 0 (wq_completion)bond186#2 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &idev->mc_lock rcu_read_lock rcu_read_lock rcu_read_lock_bh rcu_read_lock rcu_read_lock_bh rcu_read_lock &br->multicast_lock &obj_hash[i].lock irq_context: 0 (wq_completion)bond186#2 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &idev->mc_lock rcu_read_lock rcu_read_lock rcu_read_lock_bh rcu_read_lock rcu_read_lock_bh rcu_read_lock &br->multicast_lock &base->lock irq_context: 0 (wq_completion)bond186#2 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &idev->mc_lock rcu_read_lock rcu_read_lock rcu_read_lock_bh rcu_read_lock rcu_read_lock_bh rcu_read_lock &data->lock irq_context: 0 (wq_completion)bond186#2 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &idev->mc_lock rcu_read_lock rcu_read_lock rcu_read_lock_bh rcu_read_lock rcu_read_lock_bh rcu_read_lock &obj_hash[i].lock irq_context: 0 (wq_completion)bond186#2 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &idev->mc_lock rcu_read_lock rcu_read_lock rcu_node_0 irq_context: 0 (wq_completion)bond186#2 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &idev->mc_lock rcu_read_lock rcu_read_lock &rq->__lock irq_context: 0 (wq_completion)bond186#2 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &idev->mc_lock rcu_read_lock rcu_read_lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)bond186#2 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &idev->mc_lock rcu_read_lock rcu_read_lock &cfs_rq->removed.lock irq_context: 0 (wq_completion)bond186#2 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &idev->mc_lock rcu_read_lock rcu_read_lock &obj_hash[i].lock irq_context: 0 (wq_completion)bond186#2 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &idev->mc_lock rcu_read_lock rcu_read_lock pool_lock#2 irq_context: 0 (wq_completion)bond186#2 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &idev->mc_lock rcu_read_lock rcu_node_0 irq_context: 0 (wq_completion)bond186#2 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &rq->__lock irq_context: 0 (wq_completion)bond186#2 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)bond186#2 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex.wait_lock irq_context: 0 (wq_completion)bond186#2 (work_completion)(&(&bond->mcast_work)->work) &p->pi_lock irq_context: 0 (wq_completion)bond186#2 (work_completion)(&(&bond->mcast_work)->work) &p->pi_lock &rq->__lock irq_context: 0 (wq_completion)bond186#2 (work_completion)(&(&bond->mcast_work)->work) &p->pi_lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)bond147#3 (work_completion)(&(&slave->notify_work)->work) fill_pool_map-wait-type-override pool_lock irq_context: 0 (wq_completion)bond57#2 (work_completion)(&(&bond->mcast_work)->work) fill_pool_map-wait-type-override &c->lock irq_context: 0 (wq_completion)bond57#2 (work_completion)(&(&bond->mcast_work)->work) fill_pool_map-wait-type-override &____s->seqcount#2 irq_context: 0 (wq_completion)bond57#2 (work_completion)(&(&bond->mcast_work)->work) fill_pool_map-wait-type-override &____s->seqcount irq_context: 0 (wq_completion)bond57#2 (work_completion)(&(&bond->mcast_work)->work) fill_pool_map-wait-type-override pool_lock irq_context: 0 (wq_completion)bond177 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex fs_reclaim irq_context: 0 (wq_completion)bond177 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex fs_reclaim mmu_notifier_invalidate_range_start irq_context: 0 (wq_completion)bond177 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &c->lock irq_context: 0 (wq_completion)bond177 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex rcu_read_lock &bond->stats_lock/1 irq_context: 0 (wq_completion)bond177 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex rcu_read_lock &ndev->lock irq_context: 0 (wq_completion)bond177 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex nl_table_lock irq_context: 0 (wq_completion)bond177 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &obj_hash[i].lock irq_context: 0 (wq_completion)bond177 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex nl_table_wait.lock irq_context: 0 (wq_completion)bond177 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex net_rwsem irq_context: 0 (wq_completion)bond177 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex net_rwsem &list->lock#2 irq_context: 0 (wq_completion)bond177 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &tn->lock irq_context: 0 (wq_completion)bond177 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &idev->mc_lock irq_context: 0 (wq_completion)bond177 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &idev->mc_lock fs_reclaim irq_context: 0 (wq_completion)bond177 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &idev->mc_lock fs_reclaim mmu_notifier_invalidate_range_start irq_context: 0 (wq_completion)bond177 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &idev->mc_lock &c->lock irq_context: 0 (wq_completion)bond177 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &idev->mc_lock &n->list_lock irq_context: 0 (wq_completion)bond177 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &idev->mc_lock &n->list_lock &c->lock irq_context: 0 (wq_completion)bond177 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &idev->mc_lock rcu_read_lock &ndev->lock irq_context: 0 (wq_completion)bond177 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &idev->mc_lock rcu_read_lock &dir->lock#2 irq_context: 0 (wq_completion)bond177 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &idev->mc_lock rcu_read_lock &c->lock irq_context: 0 (wq_completion)bond177 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &idev->mc_lock rcu_read_lock &ul->lock#2 irq_context: 0 (wq_completion)bond177 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &idev->mc_lock rcu_read_lock rcu_read_lock &____s->seqcount#10 irq_context: 0 (wq_completion)bond177 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &idev->mc_lock rcu_read_lock rcu_read_lock rcu_read_lock_bh rcu_read_lock rcu_read_lock_bh rcu_read_lock &br->multicast_lock irq_context: 0 (wq_completion)bond177 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &idev->mc_lock rcu_read_lock rcu_read_lock rcu_read_lock_bh rcu_read_lock rcu_read_lock_bh rcu_read_lock &br->multicast_lock &obj_hash[i].lock irq_context: 0 (wq_completion)bond177 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &idev->mc_lock rcu_read_lock rcu_read_lock rcu_read_lock_bh rcu_read_lock rcu_read_lock_bh rcu_read_lock &br->multicast_lock &base->lock irq_context: 0 (wq_completion)bond177 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &idev->mc_lock rcu_read_lock rcu_read_lock rcu_read_lock_bh rcu_read_lock rcu_read_lock_bh rcu_read_lock &br->multicast_lock &base->lock &obj_hash[i].lock irq_context: 0 (wq_completion)bond177 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &idev->mc_lock rcu_read_lock rcu_read_lock rcu_read_lock_bh rcu_read_lock rcu_read_lock_bh rcu_read_lock &data->lock irq_context: 0 (wq_completion)bond177 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &idev->mc_lock rcu_read_lock rcu_read_lock rcu_read_lock_bh rcu_read_lock rcu_read_lock_bh rcu_read_lock &obj_hash[i].lock irq_context: 0 (wq_completion)bond177 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex.wait_lock irq_context: 0 (wq_completion)bond177 (work_completion)(&(&bond->mcast_work)->work) &p->pi_lock irq_context: 0 (wq_completion)bond197 (work_completion)(&(&slave->notify_work)->work) rtnl_mutex net_rwsem irq_context: 0 (wq_completion)bond197 (work_completion)(&(&slave->notify_work)->work) rtnl_mutex net_rwsem &list->lock#2 irq_context: 0 (wq_completion)bond197 (work_completion)(&(&slave->notify_work)->work) rtnl_mutex &tn->lock irq_context: 0 (wq_completion)bond197 (work_completion)(&(&slave->notify_work)->work) rtnl_mutex.wait_lock irq_context: 0 (wq_completion)bond197 (work_completion)(&(&slave->notify_work)->work) &p->pi_lock irq_context: 0 (wq_completion)bond133#2 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex fs_reclaim irq_context: 0 (wq_completion)bond133#2 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex fs_reclaim mmu_notifier_invalidate_range_start irq_context: 0 (wq_completion)bond133#2 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &c->lock irq_context: 0 (wq_completion)bond133#2 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex rcu_read_lock &bond->stats_lock/1 irq_context: 0 (wq_completion)bond133#2 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex rcu_read_lock &ndev->lock irq_context: 0 (wq_completion)bond133#2 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex nl_table_lock irq_context: 0 (wq_completion)bond133#2 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &obj_hash[i].lock irq_context: 0 (wq_completion)bond133#2 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &rq->__lock irq_context: 0 (wq_completion)bond133#2 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex nl_table_wait.lock irq_context: 0 (wq_completion)bond133#2 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex net_rwsem irq_context: 0 (wq_completion)bond133#2 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex net_rwsem &list->lock#2 irq_context: 0 (wq_completion)bond133#2 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &tn->lock irq_context: 0 (wq_completion)bond133#2 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &idev->mc_lock irq_context: 0 (wq_completion)bond133#2 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &idev->mc_lock fs_reclaim irq_context: 0 (wq_completion)bond133#2 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &idev->mc_lock fs_reclaim mmu_notifier_invalidate_range_start irq_context: 0 (wq_completion)bond133#2 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &idev->mc_lock rcu_read_lock &ndev->lock irq_context: 0 (wq_completion)bond133#2 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &idev->mc_lock rcu_read_lock rcu_node_0 irq_context: 0 (wq_completion)bond133#2 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &idev->mc_lock rcu_read_lock &rq->__lock irq_context: 0 (wq_completion)bond133#2 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &idev->mc_lock rcu_read_lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)bond133#2 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &idev->mc_lock rcu_read_lock &dir->lock#2 irq_context: 0 (wq_completion)bond133#2 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &idev->mc_lock rcu_read_lock &ul->lock#2 irq_context: 0 (wq_completion)bond133#2 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &idev->mc_lock rcu_read_lock rcu_read_lock &obj_hash[i].lock irq_context: 0 (wq_completion)bond133#2 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &idev->mc_lock rcu_read_lock rcu_read_lock &dir->lock#2 irq_context: 0 (wq_completion)bond133#2 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &idev->mc_lock rcu_read_lock rcu_read_lock &tbl->lock irq_context: 0 (wq_completion)bond133#2 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &idev->mc_lock rcu_read_lock rcu_read_lock &n->lock irq_context: 0 (wq_completion)bond133#2 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &idev->mc_lock rcu_read_lock rcu_read_lock &____s->seqcount#9 irq_context: 0 (wq_completion)bond133#2 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &idev->mc_lock rcu_read_lock rcu_read_lock rcu_read_lock_bh rcu_read_lock rcu_read_lock_bh rcu_read_lock &br->multicast_lock irq_context: 0 (wq_completion)bond133#2 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &idev->mc_lock rcu_read_lock rcu_read_lock rcu_read_lock_bh rcu_read_lock rcu_read_lock_bh rcu_read_lock &br->multicast_lock &obj_hash[i].lock irq_context: 0 (wq_completion)bond133#2 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &idev->mc_lock rcu_read_lock rcu_read_lock rcu_read_lock_bh rcu_read_lock rcu_read_lock_bh rcu_read_lock &br->multicast_lock &base->lock irq_context: 0 (wq_completion)bond133#2 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &idev->mc_lock rcu_read_lock rcu_read_lock rcu_read_lock_bh rcu_read_lock rcu_read_lock_bh rcu_read_lock &br->multicast_lock &base->lock &obj_hash[i].lock irq_context: 0 (wq_completion)bond133#2 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &idev->mc_lock rcu_read_lock rcu_read_lock rcu_read_lock_bh rcu_read_lock rcu_read_lock_bh rcu_read_lock &data->lock irq_context: 0 (wq_completion)bond133#2 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &idev->mc_lock rcu_read_lock rcu_read_lock rcu_read_lock_bh rcu_read_lock rcu_read_lock_bh rcu_read_lock &obj_hash[i].lock irq_context: 0 (wq_completion)bond133#2 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &idev->mc_lock rcu_read_lock rcu_read_lock rcu_node_0 irq_context: 0 (wq_completion)bond133#2 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &idev->mc_lock rcu_read_lock rcu_read_lock &rq->__lock irq_context: 0 (wq_completion)bond133#2 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &idev->mc_lock rcu_read_lock rcu_read_lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)bond133#2 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex.wait_lock irq_context: 0 (wq_completion)bond133#2 (work_completion)(&(&bond->mcast_work)->work) &p->pi_lock irq_context: 0 (wq_completion)bond133#2 (work_completion)(&(&bond->mcast_work)->work) &p->pi_lock &rq->__lock irq_context: 0 (wq_completion)bond133#2 (work_completion)(&(&bond->mcast_work)->work) &p->pi_lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)bond133#2 (work_completion)(&(&slave->notify_work)->work) rtnl_mutex net_rwsem irq_context: 0 (wq_completion)bond133#2 (work_completion)(&(&slave->notify_work)->work) rtnl_mutex net_rwsem &rq->__lock irq_context: 0 (wq_completion)bond133#2 (work_completion)(&(&slave->notify_work)->work) rtnl_mutex net_rwsem &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)bond133#2 (work_completion)(&(&slave->notify_work)->work) rtnl_mutex net_rwsem &list->lock#2 irq_context: 0 (wq_completion)bond133#2 (work_completion)(&(&slave->notify_work)->work) rtnl_mutex &tn->lock irq_context: 0 (wq_completion)bond133#2 (work_completion)(&(&slave->notify_work)->work) rtnl_mutex.wait_lock irq_context: 0 (wq_completion)bond133#2 (work_completion)(&(&slave->notify_work)->work) &p->pi_lock irq_context: 0 (wq_completion)bond162#3 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex fs_reclaim irq_context: 0 (wq_completion)bond162#3 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex fs_reclaim mmu_notifier_invalidate_range_start irq_context: 0 (wq_completion)bond162#3 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &c->lock irq_context: 0 (wq_completion)bond162#3 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &rq->__lock irq_context: 0 (wq_completion)bond162#3 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)bond162#3 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex batched_entropy_u8.lock irq_context: 0 (wq_completion)bond162#3 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex kfence_freelist_lock irq_context: 0 (wq_completion)bond162#3 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex rcu_read_lock &bond->stats_lock/1 irq_context: 0 (wq_completion)bond162#3 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex rcu_read_lock &ndev->lock irq_context: 0 (wq_completion)bond162#3 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex nl_table_lock irq_context: 0 (wq_completion)bond162#3 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &obj_hash[i].lock irq_context: 0 (wq_completion)bond162#3 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &meta->lock irq_context: 0 (wq_completion)bond162#3 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex nl_table_wait.lock irq_context: 0 (wq_completion)bond162#3 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex net_rwsem irq_context: 0 (wq_completion)bond162#3 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex net_rwsem &list->lock#2 irq_context: 0 (wq_completion)bond162#3 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &tn->lock irq_context: 0 (wq_completion)bond162#3 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &idev->mc_lock irq_context: 0 (wq_completion)bond162#3 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &idev->mc_lock fs_reclaim irq_context: 0 (wq_completion)bond162#3 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &idev->mc_lock fs_reclaim mmu_notifier_invalidate_range_start irq_context: 0 (wq_completion)bond162#3 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &idev->mc_lock fs_reclaim mmu_notifier_invalidate_range_start &rq->__lock irq_context: 0 (wq_completion)bond162#3 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &idev->mc_lock fs_reclaim mmu_notifier_invalidate_range_start &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)bond162#3 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &idev->mc_lock rcu_read_lock &ndev->lock irq_context: 0 (wq_completion)bond162#3 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &idev->mc_lock rcu_read_lock rcu_node_0 irq_context: 0 (wq_completion)bond162#3 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &idev->mc_lock rcu_read_lock &rq->__lock irq_context: 0 (wq_completion)bond162#3 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &idev->mc_lock rcu_read_lock &dir->lock#2 irq_context: 0 (wq_completion)bond162#3 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &idev->mc_lock rcu_read_lock &ul->lock#2 irq_context: 0 (wq_completion)bond162#3 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &idev->mc_lock rcu_read_lock rcu_read_lock &____s->seqcount#10 irq_context: 0 (wq_completion)bond162#3 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &idev->mc_lock rcu_read_lock rcu_read_lock rcu_read_lock_bh rcu_read_lock rcu_read_lock_bh rcu_read_lock &br->multicast_lock irq_context: 0 (wq_completion)bond162#3 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &idev->mc_lock rcu_read_lock rcu_read_lock rcu_read_lock_bh rcu_read_lock rcu_read_lock_bh rcu_read_lock &br->multicast_lock &obj_hash[i].lock irq_context: 0 (wq_completion)bond162#3 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &idev->mc_lock rcu_read_lock rcu_read_lock rcu_read_lock_bh rcu_read_lock rcu_read_lock_bh rcu_read_lock &br->multicast_lock &base->lock irq_context: 0 (wq_completion)bond162#3 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &idev->mc_lock rcu_read_lock rcu_read_lock rcu_read_lock_bh rcu_read_lock rcu_read_lock_bh rcu_read_lock &br->multicast_lock &base->lock &obj_hash[i].lock irq_context: 0 (wq_completion)bond162#3 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &idev->mc_lock rcu_read_lock rcu_read_lock rcu_read_lock_bh rcu_read_lock rcu_read_lock_bh rcu_read_lock &data->lock irq_context: 0 (wq_completion)bond162#3 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &idev->mc_lock rcu_read_lock rcu_read_lock rcu_read_lock_bh rcu_read_lock rcu_read_lock_bh rcu_read_lock &obj_hash[i].lock irq_context: 0 (wq_completion)bond162#3 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &idev->mc_lock rcu_read_lock rcu_read_lock rcu_node_0 irq_context: 0 (wq_completion)bond162#3 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &idev->mc_lock rcu_read_lock rcu_read_lock &rq->__lock irq_context: 0 (wq_completion)bond162#3 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &idev->mc_lock rcu_read_lock rcu_read_lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)bond162#3 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex.wait_lock irq_context: 0 (wq_completion)bond162#3 (work_completion)(&(&bond->mcast_work)->work) &p->pi_lock irq_context: 0 (wq_completion)bond162#3 (work_completion)(&(&bond->mcast_work)->work) &p->pi_lock &rq->__lock irq_context: 0 (wq_completion)bond162#3 (work_completion)(&(&bond->mcast_work)->work) &p->pi_lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)bond162#3 (work_completion)(&(&bond->mcast_work)->work) &rq->__lock irq_context: 0 (wq_completion)bond162#3 (work_completion)(&(&bond->mcast_work)->work) &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)bond175#2 irq_context: 0 (wq_completion)bond175#2 &rq->__lock irq_context: 0 (wq_completion)bond175#2 &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)bond175#2 (work_completion)(&(&slave->notify_work)->work) irq_context: 0 (wq_completion)bond175#2 (work_completion)(&(&slave->notify_work)->work) &obj_hash[i].lock irq_context: 0 (wq_completion)bond175#2 (work_completion)(&(&slave->notify_work)->work) &base->lock irq_context: 0 (wq_completion)bond175#2 (work_completion)(&(&slave->notify_work)->work) &base->lock &obj_hash[i].lock irq_context: 0 (wq_completion)bond196 (work_completion)(&(&slave->notify_work)->work) &rq->__lock irq_context: 0 (wq_completion)bond196 (work_completion)(&(&slave->notify_work)->work) &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)bond189#2 (work_completion)(&(&slave->notify_work)->work) &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)bond175#2 (work_completion)(&(&bond->mcast_work)->work) irq_context: 0 (wq_completion)bond175#2 (work_completion)(&(&bond->mcast_work)->work) &obj_hash[i].lock irq_context: 0 (wq_completion)bond175#2 (work_completion)(&(&bond->mcast_work)->work) &base->lock irq_context: 0 (wq_completion)bond175#2 (work_completion)(&(&bond->mcast_work)->work) &base->lock fill_pool_map-wait-type-override &c->lock irq_context: 0 (wq_completion)bond175#2 (work_completion)(&(&bond->mcast_work)->work) &base->lock fill_pool_map-wait-type-override &____s->seqcount#2 irq_context: 0 (wq_completion)bond175#2 (work_completion)(&(&bond->mcast_work)->work) &base->lock fill_pool_map-wait-type-override &____s->seqcount irq_context: 0 (wq_completion)bond175#2 (work_completion)(&(&bond->mcast_work)->work) &base->lock fill_pool_map-wait-type-override pool_lock#2 irq_context: 0 (wq_completion)bond175#2 (work_completion)(&(&bond->mcast_work)->work) &base->lock fill_pool_map-wait-type-override pool_lock irq_context: 0 (wq_completion)bond175#2 (work_completion)(&(&bond->mcast_work)->work) &base->lock &obj_hash[i].lock irq_context: 0 (wq_completion)bond175#2 (work_completion)(&(&bond->mcast_work)->work) &rq->__lock irq_context: 0 (wq_completion)bond175#2 (work_completion)(&(&bond->mcast_work)->work) &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)bond171#3 (work_completion)(&(&bond->mcast_work)->work) &rq->__lock irq_context: 0 (wq_completion)bond171#3 (work_completion)(&(&bond->mcast_work)->work) &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)bond204 irq_context: 0 (wq_completion)bond204 (work_completion)(&(&slave->notify_work)->work) irq_context: 0 (wq_completion)bond204 (work_completion)(&(&slave->notify_work)->work) &obj_hash[i].lock irq_context: 0 (wq_completion)bond204 (work_completion)(&(&slave->notify_work)->work) &base->lock irq_context: 0 (wq_completion)bond204 (work_completion)(&(&slave->notify_work)->work) &base->lock &obj_hash[i].lock irq_context: 0 (wq_completion)bond160 (work_completion)(&(&bond->mcast_work)->work) fill_pool_map-wait-type-override pool_lock irq_context: 0 (wq_completion)bond204 (work_completion)(&(&bond->mcast_work)->work) irq_context: 0 (wq_completion)bond204 (work_completion)(&(&bond->mcast_work)->work) &obj_hash[i].lock irq_context: 0 (wq_completion)bond204 (work_completion)(&(&bond->mcast_work)->work) &base->lock irq_context: 0 (wq_completion)bond204 (work_completion)(&(&bond->mcast_work)->work) &base->lock &obj_hash[i].lock irq_context: 0 (wq_completion)bond192#2 irq_context: 0 (wq_completion)bond192#2 (work_completion)(&(&slave->notify_work)->work) irq_context: 0 (wq_completion)bond192#2 (work_completion)(&(&slave->notify_work)->work) &obj_hash[i].lock irq_context: 0 (wq_completion)bond192#2 (work_completion)(&(&slave->notify_work)->work) &base->lock irq_context: 0 (wq_completion)bond192#2 (work_completion)(&(&slave->notify_work)->work) &base->lock &obj_hash[i].lock irq_context: 0 (wq_completion)bond192#2 (work_completion)(&(&slave->notify_work)->work) &rq->__lock irq_context: 0 (wq_completion)bond137#3 (work_completion)(&(&slave->notify_work)->work) rtnl_mutex net_rwsem irq_context: 0 (wq_completion)bond137#3 (work_completion)(&(&slave->notify_work)->work) rtnl_mutex net_rwsem &list->lock#2 irq_context: 0 (wq_completion)bond137#3 (work_completion)(&(&slave->notify_work)->work) rtnl_mutex &tn->lock irq_context: 0 (wq_completion)bond137#3 (work_completion)(&(&slave->notify_work)->work) rtnl_mutex.wait_lock irq_context: 0 (wq_completion)bond137#3 (work_completion)(&(&slave->notify_work)->work) &p->pi_lock irq_context: 0 (wq_completion)bond48#2 (work_completion)(&(&slave->notify_work)->work) rtnl_mutex &rq->__lock irq_context: 0 (wq_completion)bond48#2 (work_completion)(&(&slave->notify_work)->work) rtnl_mutex net_rwsem irq_context: 0 (wq_completion)bond192#2 (work_completion)(&(&bond->mcast_work)->work) irq_context: 0 (wq_completion)bond48#2 (work_completion)(&(&slave->notify_work)->work) rtnl_mutex net_rwsem &list->lock#2 irq_context: 0 (wq_completion)bond192#2 (work_completion)(&(&bond->mcast_work)->work) &obj_hash[i].lock irq_context: 0 (wq_completion)bond48#2 (work_completion)(&(&slave->notify_work)->work) rtnl_mutex &tn->lock irq_context: 0 (wq_completion)bond192#2 (work_completion)(&(&bond->mcast_work)->work) &base->lock irq_context: 0 (wq_completion)bond192#2 (work_completion)(&(&bond->mcast_work)->work) &base->lock &obj_hash[i].lock irq_context: 0 (wq_completion)bond48#2 (work_completion)(&(&slave->notify_work)->work) rtnl_mutex.wait_lock irq_context: 0 (wq_completion)bond54#3 (work_completion)(&(&slave->notify_work)->work) rtnl_mutex net_rwsem irq_context: 0 (wq_completion)bond54#3 (work_completion)(&(&slave->notify_work)->work) rtnl_mutex net_rwsem &list->lock#2 irq_context: 0 (wq_completion)bond54#3 (work_completion)(&(&slave->notify_work)->work) rtnl_mutex net_rwsem &rq->__lock irq_context: 0 (wq_completion)bond48#2 (work_completion)(&(&slave->notify_work)->work) &p->pi_lock irq_context: 0 (wq_completion)bond54#3 (work_completion)(&(&slave->notify_work)->work) rtnl_mutex &tn->lock irq_context: 0 (wq_completion)bond54#3 (work_completion)(&(&slave->notify_work)->work) rtnl_mutex.wait_lock irq_context: 0 (wq_completion)bond54#3 (work_completion)(&(&slave->notify_work)->work) &p->pi_lock irq_context: 0 (wq_completion)bond54#3 (work_completion)(&(&slave->notify_work)->work) &p->pi_lock &rq->__lock irq_context: 0 (wq_completion)bond54#3 (work_completion)(&(&slave->notify_work)->work) &p->pi_lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)bond175#2 (work_completion)(&(&slave->notify_work)->work) &rq->__lock irq_context: 0 (wq_completion)bond175#2 (work_completion)(&(&slave->notify_work)->work) &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)bond163#3 &rq->__lock irq_context: 0 (wq_completion)bond163#3 &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)bond176#3 irq_context: 0 (wq_completion)bond176#3 &rq->__lock irq_context: 0 (wq_completion)bond176#3 &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)bond176#3 (work_completion)(&(&slave->notify_work)->work) irq_context: 0 (wq_completion)bond176#3 (work_completion)(&(&slave->notify_work)->work) &obj_hash[i].lock irq_context: 0 (wq_completion)bond176#3 (work_completion)(&(&slave->notify_work)->work) &base->lock irq_context: 0 (wq_completion)bond176#3 (work_completion)(&(&slave->notify_work)->work) &base->lock &obj_hash[i].lock irq_context: 0 (wq_completion)bond190#2 (work_completion)(&(&bond->mcast_work)->work) &rq->__lock irq_context: 0 (wq_completion)bond190#2 (work_completion)(&(&bond->mcast_work)->work) &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)mld (work_completion)(&(&idev->mc_ifc_work)->work) &idev->mc_lock rcu_read_lock rcu_read_lock rcu_read_lock rcu_node_0 irq_context: 0 (wq_completion)mld (work_completion)(&(&idev->mc_ifc_work)->work) &idev->mc_lock rcu_read_lock rcu_read_lock rcu_read_lock &rq->__lock irq_context: 0 (wq_completion)bond190#2 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex fs_reclaim irq_context: 0 (wq_completion)bond190#2 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex fs_reclaim mmu_notifier_invalidate_range_start irq_context: 0 (wq_completion)bond190#2 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &c->lock irq_context: 0 (wq_completion)bond190#2 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &____s->seqcount#2 irq_context: 0 (wq_completion)bond190#2 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &____s->seqcount irq_context: 0 (wq_completion)bond190#2 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex rcu_read_lock &bond->stats_lock/1 irq_context: 0 (wq_completion)bond190#2 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex rcu_read_lock &ndev->lock irq_context: 0 (wq_completion)bond190#2 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex nl_table_lock irq_context: 0 (wq_completion)bond190#2 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &obj_hash[i].lock irq_context: 0 (wq_completion)bond190#2 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex nl_table_wait.lock irq_context: 0 (wq_completion)bond190#2 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex net_rwsem irq_context: 0 (wq_completion)bond190#2 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex net_rwsem &list->lock#2 irq_context: 0 (wq_completion)bond190#2 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &tn->lock irq_context: 0 (wq_completion)bond190#2 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &idev->mc_lock irq_context: 0 (wq_completion)bond190#2 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &idev->mc_lock fs_reclaim irq_context: 0 (wq_completion)bond190#2 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &idev->mc_lock fs_reclaim mmu_notifier_invalidate_range_start irq_context: 0 (wq_completion)bond190#2 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &idev->mc_lock rcu_read_lock &ndev->lock irq_context: 0 (wq_completion)bond190#2 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &idev->mc_lock rcu_read_lock &dir->lock#2 irq_context: 0 (wq_completion)bond190#2 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &idev->mc_lock rcu_read_lock &ul->lock#2 irq_context: 0 (wq_completion)bond190#2 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &idev->mc_lock rcu_read_lock rcu_read_lock &____s->seqcount#10 irq_context: 0 (wq_completion)bond190#2 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &idev->mc_lock rcu_read_lock rcu_read_lock rcu_read_lock_bh rcu_read_lock rcu_read_lock_bh rcu_read_lock &br->multicast_lock irq_context: 0 (wq_completion)bond190#2 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &idev->mc_lock rcu_read_lock rcu_read_lock rcu_read_lock_bh rcu_read_lock rcu_read_lock_bh rcu_read_lock &br->multicast_lock &obj_hash[i].lock irq_context: 0 (wq_completion)bond190#2 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &idev->mc_lock rcu_read_lock rcu_read_lock rcu_read_lock_bh rcu_read_lock rcu_read_lock_bh rcu_read_lock &br->multicast_lock &base->lock irq_context: 0 (wq_completion)bond190#2 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &idev->mc_lock rcu_read_lock rcu_read_lock rcu_read_lock_bh rcu_read_lock rcu_read_lock_bh rcu_read_lock &data->lock irq_context: 0 (wq_completion)bond190#2 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &idev->mc_lock rcu_read_lock rcu_read_lock rcu_read_lock_bh rcu_read_lock rcu_read_lock_bh rcu_read_lock &obj_hash[i].lock irq_context: 0 (wq_completion)bond190#2 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex.wait_lock irq_context: 0 (wq_completion)bond155 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex fs_reclaim irq_context: 0 (wq_completion)bond190#2 (work_completion)(&(&bond->mcast_work)->work) &p->pi_lock irq_context: 0 (wq_completion)bond155 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex fs_reclaim mmu_notifier_invalidate_range_start irq_context: 0 (wq_completion)bond155 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &c->lock irq_context: 0 (wq_completion)bond155 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex rcu_read_lock &bond->stats_lock/1 irq_context: 0 (wq_completion)bond155 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex rcu_read_lock &ndev->lock irq_context: 0 (wq_completion)bond155 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex nl_table_lock irq_context: 0 (wq_completion)bond155 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &obj_hash[i].lock irq_context: 0 (wq_completion)bond155 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex nl_table_wait.lock irq_context: 0 (wq_completion)bond155 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex net_rwsem irq_context: 0 (wq_completion)bond155 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex net_rwsem &list->lock#2 irq_context: 0 (wq_completion)bond155 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &tn->lock irq_context: 0 (wq_completion)bond155 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &idev->mc_lock irq_context: 0 (wq_completion)bond155 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &idev->mc_lock fs_reclaim irq_context: 0 (wq_completion)bond155 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &idev->mc_lock fs_reclaim mmu_notifier_invalidate_range_start irq_context: 0 (wq_completion)bond155 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &idev->mc_lock rcu_read_lock &ndev->lock irq_context: 0 (wq_completion)bond155 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &idev->mc_lock rcu_read_lock &dir->lock#2 irq_context: 0 (wq_completion)bond155 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &idev->mc_lock rcu_read_lock &c->lock irq_context: 0 (wq_completion)bond155 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &idev->mc_lock rcu_read_lock &ul->lock#2 irq_context: 0 (wq_completion)bond155 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &idev->mc_lock rcu_read_lock rcu_read_lock &obj_hash[i].lock irq_context: 0 (wq_completion)bond155 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &idev->mc_lock rcu_read_lock rcu_read_lock &dir->lock#2 irq_context: 0 (wq_completion)bond155 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &idev->mc_lock rcu_read_lock rcu_read_lock &tbl->lock irq_context: 0 (wq_completion)bond155 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &idev->mc_lock rcu_read_lock rcu_read_lock &n->lock irq_context: 0 (wq_completion)bond155 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &idev->mc_lock rcu_read_lock rcu_read_lock &____s->seqcount#9 irq_context: 0 (wq_completion)bond155 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &idev->mc_lock rcu_read_lock rcu_read_lock rcu_read_lock_bh rcu_read_lock rcu_read_lock_bh rcu_read_lock &br->multicast_lock irq_context: 0 (wq_completion)bond155 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &idev->mc_lock rcu_read_lock rcu_read_lock rcu_read_lock_bh rcu_read_lock rcu_read_lock_bh rcu_read_lock &br->multicast_lock &obj_hash[i].lock irq_context: 0 (wq_completion)bond155 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &idev->mc_lock rcu_read_lock rcu_read_lock rcu_read_lock_bh rcu_read_lock rcu_read_lock_bh rcu_read_lock &br->multicast_lock &base->lock irq_context: 0 (wq_completion)bond155 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &idev->mc_lock rcu_read_lock rcu_read_lock rcu_read_lock_bh rcu_read_lock rcu_read_lock_bh rcu_read_lock &br->multicast_lock &base->lock &obj_hash[i].lock irq_context: 0 (wq_completion)bond155 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &idev->mc_lock rcu_read_lock rcu_read_lock rcu_read_lock_bh rcu_read_lock rcu_read_lock_bh rcu_read_lock &data->lock irq_context: 0 (wq_completion)bond155 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &idev->mc_lock rcu_read_lock rcu_read_lock rcu_read_lock_bh rcu_read_lock rcu_read_lock_bh rcu_read_lock &obj_hash[i].lock irq_context: 0 (wq_completion)bond155 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &idev->mc_lock rcu_read_lock rcu_read_lock rcu_node_0 irq_context: 0 (wq_completion)bond155 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &idev->mc_lock rcu_read_lock rcu_read_lock &rq->__lock irq_context: 0 (wq_completion)bond155 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &idev->mc_lock rcu_read_lock rcu_node_0 irq_context: 0 (wq_completion)bond155 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex.wait_lock irq_context: 0 (wq_completion)bond155 (work_completion)(&(&bond->mcast_work)->work) &p->pi_lock irq_context: 0 (wq_completion)bond176#3 (work_completion)(&(&bond->mcast_work)->work) irq_context: 0 (wq_completion)bond176#3 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex fs_reclaim irq_context: 0 (wq_completion)bond176#3 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex fs_reclaim mmu_notifier_invalidate_range_start irq_context: 0 (wq_completion)bond176#3 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &c->lock irq_context: 0 (wq_completion)bond176#3 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex rcu_read_lock &bond->stats_lock/1 irq_context: 0 (wq_completion)bond176#3 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex rcu_read_lock &ndev->lock irq_context: 0 (wq_completion)bond176#3 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex nl_table_lock irq_context: 0 (wq_completion)bond176#3 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &obj_hash[i].lock irq_context: 0 (wq_completion)bond176#3 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex nl_table_wait.lock irq_context: 0 (wq_completion)bond176#3 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex net_rwsem irq_context: 0 (wq_completion)bond176#3 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex net_rwsem &list->lock#2 irq_context: 0 (wq_completion)bond176#3 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &tn->lock irq_context: 0 (wq_completion)bond176#3 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &idev->mc_lock irq_context: 0 (wq_completion)bond176#3 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex.wait_lock irq_context: 0 (wq_completion)bond176#3 (work_completion)(&(&bond->mcast_work)->work) &p->pi_lock irq_context: 0 (wq_completion)bond163#2 (work_completion)(&(&slave->notify_work)->work) rtnl_mutex net_rwsem irq_context: 0 (wq_completion)bond163#2 (work_completion)(&(&slave->notify_work)->work) rtnl_mutex net_rwsem &list->lock#2 irq_context: 0 (wq_completion)bond163#2 (work_completion)(&(&slave->notify_work)->work) rtnl_mutex &tn->lock irq_context: 0 (wq_completion)bond163#2 (work_completion)(&(&slave->notify_work)->work) rtnl_mutex.wait_lock irq_context: 0 (wq_completion)bond163#2 (work_completion)(&(&slave->notify_work)->work) &p->pi_lock irq_context: 0 (wq_completion)bond187 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex fs_reclaim irq_context: 0 (wq_completion)bond187 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex fs_reclaim mmu_notifier_invalidate_range_start irq_context: 0 (wq_completion)bond187 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex rcu_read_lock &bond->stats_lock/1 irq_context: 0 (wq_completion)bond187 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex rcu_read_lock &ndev->lock irq_context: 0 (wq_completion)bond187 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex nl_table_lock irq_context: 0 (wq_completion)bond187 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &obj_hash[i].lock irq_context: 0 (wq_completion)bond187 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex nl_table_wait.lock irq_context: 0 (wq_completion)bond187 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex net_rwsem irq_context: 0 (wq_completion)bond187 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex net_rwsem &list->lock#2 irq_context: 0 (wq_completion)bond187 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &tn->lock irq_context: 0 (wq_completion)bond187 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &idev->mc_lock irq_context: 0 (wq_completion)bond187 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &idev->mc_lock fs_reclaim irq_context: 0 (wq_completion)bond187 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &idev->mc_lock fs_reclaim mmu_notifier_invalidate_range_start irq_context: 0 (wq_completion)bond187 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &idev->mc_lock &c->lock irq_context: 0 (wq_completion)bond187 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &idev->mc_lock rcu_read_lock &ndev->lock irq_context: 0 (wq_completion)bond187 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &idev->mc_lock rcu_read_lock &dir->lock#2 irq_context: 0 (wq_completion)bond187 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &idev->mc_lock rcu_read_lock &ul->lock#2 irq_context: 0 (wq_completion)bond187 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &idev->mc_lock rcu_read_lock rcu_read_lock &____s->seqcount#10 irq_context: 0 (wq_completion)bond187 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &idev->mc_lock rcu_read_lock rcu_read_lock rcu_read_lock_bh rcu_read_lock rcu_read_lock_bh rcu_read_lock &br->multicast_lock irq_context: 0 (wq_completion)bond187 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &idev->mc_lock rcu_read_lock rcu_read_lock rcu_read_lock_bh rcu_read_lock rcu_read_lock_bh rcu_read_lock &br->multicast_lock &obj_hash[i].lock irq_context: 0 (wq_completion)bond187 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &idev->mc_lock rcu_read_lock rcu_read_lock rcu_read_lock_bh rcu_read_lock rcu_read_lock_bh rcu_read_lock &br->multicast_lock &base->lock irq_context: 0 (wq_completion)bond187 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &idev->mc_lock rcu_read_lock rcu_read_lock rcu_read_lock_bh rcu_read_lock rcu_read_lock_bh rcu_read_lock &br->multicast_lock &base->lock &obj_hash[i].lock irq_context: 0 (wq_completion)bond187 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &idev->mc_lock rcu_read_lock rcu_read_lock rcu_read_lock_bh rcu_read_lock rcu_read_lock_bh rcu_read_lock &data->lock irq_context: 0 (wq_completion)bond187 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &idev->mc_lock rcu_read_lock rcu_read_lock rcu_read_lock_bh rcu_read_lock rcu_read_lock_bh rcu_read_lock &obj_hash[i].lock irq_context: 0 (wq_completion)bond187 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &idev->mc_lock rcu_read_lock rcu_read_lock rcu_node_0 irq_context: 0 (wq_completion)bond187 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &idev->mc_lock rcu_read_lock rcu_read_lock &rq->__lock irq_context: 0 (wq_completion)bond187 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &idev->mc_lock rcu_read_lock rcu_node_0 irq_context: 0 (wq_completion)bond187 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &idev->mc_lock rcu_read_lock &rcu_state.gp_wq irq_context: 0 (wq_completion)bond187 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &idev->mc_lock rcu_read_lock &rcu_state.gp_wq &p->pi_lock irq_context: 0 (wq_completion)bond187 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &idev->mc_lock rcu_read_lock &rcu_state.gp_wq &p->pi_lock &rq->__lock irq_context: 0 (wq_completion)bond187 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &idev->mc_lock rcu_read_lock &rcu_state.gp_wq &p->pi_lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)bond187 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &idev->mc_lock rcu_read_lock &rq->__lock irq_context: 0 (wq_completion)bond187 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex.wait_lock irq_context: 0 (wq_completion)bond187 (work_completion)(&(&bond->mcast_work)->work) &p->pi_lock irq_context: 0 (wq_completion)bond187 (work_completion)(&(&bond->mcast_work)->work) &p->pi_lock &rq->__lock irq_context: 0 (wq_completion)bond187 (work_completion)(&(&bond->mcast_work)->work) &p->pi_lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)bond179#2 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex fs_reclaim irq_context: 0 (wq_completion)bond179#2 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex fs_reclaim mmu_notifier_invalidate_range_start irq_context: 0 (wq_completion)bond179#2 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &c->lock irq_context: 0 (wq_completion)bond179#2 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex rcu_read_lock &bond->stats_lock/1 irq_context: 0 (wq_completion)bond179#2 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex rcu_read_lock &ndev->lock irq_context: 0 (wq_completion)bond179#2 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex nl_table_lock irq_context: 0 (wq_completion)bond179#2 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &obj_hash[i].lock irq_context: 0 (wq_completion)bond179#2 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex nl_table_wait.lock irq_context: 0 (wq_completion)bond179#2 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex net_rwsem irq_context: 0 (wq_completion)bond179#2 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex net_rwsem &list->lock#2 irq_context: 0 (wq_completion)bond179#2 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &tn->lock irq_context: 0 (wq_completion)bond179#2 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &idev->mc_lock irq_context: 0 (wq_completion)bond179#2 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &idev->mc_lock fs_reclaim irq_context: 0 (wq_completion)bond179#2 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &idev->mc_lock fs_reclaim mmu_notifier_invalidate_range_start irq_context: 0 (wq_completion)bond179#2 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &idev->mc_lock &c->lock irq_context: 0 (wq_completion)bond179#2 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &idev->mc_lock &n->list_lock irq_context: 0 (wq_completion)bond179#2 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &idev->mc_lock &n->list_lock &c->lock irq_context: 0 (wq_completion)bond179#2 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &idev->mc_lock rcu_read_lock &ndev->lock irq_context: 0 (wq_completion)bond179#2 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &idev->mc_lock rcu_read_lock &dir->lock#2 irq_context: 0 (wq_completion)bond179#2 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &idev->mc_lock rcu_read_lock &ul->lock#2 irq_context: 0 (wq_completion)bond179#2 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &idev->mc_lock rcu_read_lock rcu_read_lock &____s->seqcount#10 irq_context: 0 (wq_completion)bond179#2 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &idev->mc_lock rcu_read_lock rcu_read_lock rcu_read_lock_bh rcu_read_lock rcu_read_lock_bh rcu_read_lock &br->multicast_lock irq_context: 0 (wq_completion)bond179#2 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &idev->mc_lock rcu_read_lock rcu_read_lock rcu_read_lock_bh rcu_read_lock rcu_read_lock_bh rcu_read_lock &br->multicast_lock &obj_hash[i].lock irq_context: 0 (wq_completion)bond179#2 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &idev->mc_lock rcu_read_lock rcu_read_lock rcu_read_lock_bh rcu_read_lock rcu_read_lock_bh rcu_read_lock &br->multicast_lock &base->lock irq_context: 0 (wq_completion)bond179#2 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &idev->mc_lock rcu_read_lock rcu_read_lock rcu_read_lock_bh rcu_read_lock rcu_read_lock_bh rcu_read_lock &br->multicast_lock &base->lock &obj_hash[i].lock irq_context: 0 (wq_completion)bond179#2 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &idev->mc_lock rcu_read_lock rcu_read_lock rcu_read_lock_bh rcu_read_lock rcu_read_lock_bh rcu_read_lock &data->lock irq_context: 0 (wq_completion)bond179#2 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &idev->mc_lock rcu_read_lock rcu_read_lock rcu_read_lock_bh rcu_read_lock rcu_read_lock_bh rcu_read_lock &obj_hash[i].lock irq_context: 0 (wq_completion)bond179#2 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex.wait_lock irq_context: 0 (wq_completion)bond179#2 (work_completion)(&(&bond->mcast_work)->work) &p->pi_lock irq_context: 0 (wq_completion)bond171#3 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex fs_reclaim irq_context: 0 (wq_completion)bond171#3 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex fs_reclaim mmu_notifier_invalidate_range_start irq_context: 0 (wq_completion)bond171#3 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex rcu_read_lock &bond->stats_lock/1 irq_context: 0 (wq_completion)bond171#3 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex rcu_read_lock &ndev->lock irq_context: 0 (wq_completion)bond171#3 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex nl_table_lock irq_context: 0 (wq_completion)bond171#3 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &obj_hash[i].lock irq_context: 0 (wq_completion)bond171#3 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex nl_table_wait.lock irq_context: 0 (wq_completion)bond171#3 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex net_rwsem irq_context: 0 (wq_completion)bond171#3 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex net_rwsem &list->lock#2 irq_context: 0 (wq_completion)bond171#3 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &tn->lock irq_context: 0 (wq_completion)bond171#3 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &idev->mc_lock irq_context: 0 (wq_completion)bond171#3 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &idev->mc_lock fs_reclaim irq_context: 0 (wq_completion)bond171#3 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &idev->mc_lock fs_reclaim mmu_notifier_invalidate_range_start irq_context: 0 (wq_completion)bond171#3 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &idev->mc_lock rcu_read_lock &ndev->lock irq_context: 0 (wq_completion)bond171#3 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &idev->mc_lock rcu_read_lock &c->lock irq_context: 0 (wq_completion)bond171#3 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &idev->mc_lock rcu_read_lock &dir->lock#2 irq_context: 0 (wq_completion)bond171#3 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &idev->mc_lock rcu_read_lock &ul->lock#2 irq_context: 0 (wq_completion)bond171#3 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &idev->mc_lock rcu_read_lock rcu_read_lock &____s->seqcount#10 irq_context: 0 (wq_completion)bond171#3 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &idev->mc_lock rcu_read_lock rcu_read_lock rcu_read_lock_bh rcu_read_lock rcu_read_lock_bh rcu_read_lock &br->multicast_lock irq_context: 0 (wq_completion)bond171#3 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &idev->mc_lock rcu_read_lock rcu_read_lock rcu_read_lock_bh rcu_read_lock rcu_read_lock_bh rcu_read_lock &br->multicast_lock &obj_hash[i].lock irq_context: 0 (wq_completion)bond171#3 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &idev->mc_lock rcu_read_lock rcu_read_lock rcu_read_lock_bh rcu_read_lock rcu_read_lock_bh rcu_read_lock &br->multicast_lock &base->lock irq_context: 0 (wq_completion)bond171#3 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &idev->mc_lock rcu_read_lock rcu_read_lock rcu_read_lock_bh rcu_read_lock rcu_read_lock_bh rcu_read_lock &data->lock irq_context: 0 (wq_completion)bond171#3 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &idev->mc_lock rcu_read_lock rcu_read_lock rcu_read_lock_bh rcu_read_lock rcu_read_lock_bh rcu_read_lock &obj_hash[i].lock irq_context: 0 (wq_completion)bond171#3 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &idev->mc_lock rcu_read_lock rcu_read_lock rcu_node_0 irq_context: 0 (wq_completion)bond171#3 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &idev->mc_lock rcu_read_lock rcu_read_lock &rq->__lock irq_context: 0 (wq_completion)bond171#3 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &idev->mc_lock rcu_read_lock rcu_node_0 irq_context: 0 (wq_completion)bond171#3 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex.wait_lock irq_context: 0 (wq_completion)bond171#3 (work_completion)(&(&bond->mcast_work)->work) &p->pi_lock irq_context: 0 (wq_completion)bond171#3 (work_completion)(&(&bond->mcast_work)->work) &p->pi_lock &cfs_rq->removed.lock irq_context: 0 (wq_completion)bond171#3 (work_completion)(&(&bond->mcast_work)->work) &p->pi_lock &rq->__lock irq_context: 0 (wq_completion)bond171#3 (work_completion)(&(&bond->mcast_work)->work) &p->pi_lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)bond192#2 (work_completion)(&(&slave->notify_work)->work) &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)bond177#3 irq_context: 0 (wq_completion)bond177#3 &rq->__lock irq_context: 0 (wq_completion)bond177#3 &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)bond177#3 (work_completion)(&(&slave->notify_work)->work) irq_context: 0 (wq_completion)bond177#3 (work_completion)(&(&slave->notify_work)->work) &obj_hash[i].lock irq_context: 0 (wq_completion)bond177#3 (work_completion)(&(&slave->notify_work)->work) &base->lock irq_context: 0 (wq_completion)bond177#3 (work_completion)(&(&slave->notify_work)->work) &base->lock &obj_hash[i].lock irq_context: 0 (wq_completion)bond150#3 (work_completion)(&(&bond->mcast_work)->work) fill_pool_map-wait-type-override &c->lock irq_context: 0 (wq_completion)bond150#3 (work_completion)(&(&bond->mcast_work)->work) fill_pool_map-wait-type-override &____s->seqcount#2 irq_context: 0 (wq_completion)bond150#3 (work_completion)(&(&bond->mcast_work)->work) fill_pool_map-wait-type-override &____s->seqcount irq_context: 0 (wq_completion)bond150#3 (work_completion)(&(&bond->mcast_work)->work) fill_pool_map-wait-type-override pool_lock irq_context: 0 (wq_completion)bond177#3 (work_completion)(&(&bond->mcast_work)->work) irq_context: 0 (wq_completion)bond177#3 (work_completion)(&(&bond->mcast_work)->work) &obj_hash[i].lock irq_context: 0 (wq_completion)bond177#3 (work_completion)(&(&bond->mcast_work)->work) &base->lock irq_context: 0 (wq_completion)bond177#3 (work_completion)(&(&bond->mcast_work)->work) &base->lock &obj_hash[i].lock irq_context: 0 (wq_completion)bond193#2 irq_context: 0 (wq_completion)bond193#2 (work_completion)(&(&slave->notify_work)->work) irq_context: 0 (wq_completion)bond193#2 (work_completion)(&(&slave->notify_work)->work) &obj_hash[i].lock irq_context: 0 (wq_completion)bond193#2 (work_completion)(&(&slave->notify_work)->work) &base->lock irq_context: 0 (wq_completion)bond193#2 (work_completion)(&(&slave->notify_work)->work) &base->lock &obj_hash[i].lock irq_context: 0 (wq_completion)bond193#2 (work_completion)(&(&slave->notify_work)->work) &rq->__lock irq_context: 0 (wq_completion)bond193#2 (work_completion)(&(&slave->notify_work)->work) &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)bond193#2 (work_completion)(&(&bond->mcast_work)->work) irq_context: 0 (wq_completion)bond193#2 (work_completion)(&(&bond->mcast_work)->work) &rq->__lock irq_context: 0 (wq_completion)bond193#2 (work_completion)(&(&bond->mcast_work)->work) &obj_hash[i].lock irq_context: 0 (wq_completion)bond193#2 (work_completion)(&(&bond->mcast_work)->work) &base->lock irq_context: 0 (wq_completion)bond193#2 (work_completion)(&(&bond->mcast_work)->work) &base->lock &obj_hash[i].lock irq_context: 0 (wq_completion)bond193#2 (work_completion)(&(&bond->mcast_work)->work) &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)bond205 irq_context: 0 (wq_completion)bond205 (work_completion)(&(&slave->notify_work)->work) irq_context: 0 (wq_completion)bond205 (work_completion)(&(&slave->notify_work)->work) &obj_hash[i].lock irq_context: 0 (wq_completion)bond205 (work_completion)(&(&slave->notify_work)->work) &base->lock irq_context: 0 (wq_completion)bond205 (work_completion)(&(&slave->notify_work)->work) &base->lock &obj_hash[i].lock irq_context: 0 (wq_completion)bond205 (work_completion)(&(&slave->notify_work)->work) &rq->__lock irq_context: 0 (wq_completion)bond205 (work_completion)(&(&slave->notify_work)->work) &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)bond163#2 &rq->__lock irq_context: 0 (wq_completion)bond205 (work_completion)(&(&bond->mcast_work)->work) irq_context: 0 (wq_completion)bond205 (work_completion)(&(&bond->mcast_work)->work) &obj_hash[i].lock irq_context: 0 (wq_completion)bond205 (work_completion)(&(&bond->mcast_work)->work) &base->lock irq_context: 0 (wq_completion)bond205 (work_completion)(&(&bond->mcast_work)->work) &base->lock &obj_hash[i].lock irq_context: 0 (wq_completion)bond205 (work_completion)(&(&bond->mcast_work)->work) &rq->__lock irq_context: 0 (wq_completion)bond205 (work_completion)(&(&bond->mcast_work)->work) &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)bond205 (work_completion)(&(&bond->mcast_work)->work) &cfs_rq->removed.lock irq_context: 0 (wq_completion)bond205 (work_completion)(&(&slave->notify_work)->work) rtnl_mutex net_rwsem irq_context: 0 (wq_completion)bond205 (work_completion)(&(&slave->notify_work)->work) rtnl_mutex net_rwsem &list->lock#2 irq_context: 0 (wq_completion)bond205 (work_completion)(&(&slave->notify_work)->work) rtnl_mutex net_rwsem &rq->__lock irq_context: 0 (wq_completion)bond205 (work_completion)(&(&slave->notify_work)->work) rtnl_mutex net_rwsem &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)bond205 (work_completion)(&(&slave->notify_work)->work) rtnl_mutex &tn->lock irq_context: 0 (wq_completion)bond160#3 (work_completion)(&(&slave->notify_work)->work) rtnl_mutex net_rwsem irq_context: 0 (wq_completion)bond160#3 (work_completion)(&(&slave->notify_work)->work) rtnl_mutex net_rwsem &list->lock#2 irq_context: 0 (wq_completion)bond160#3 (work_completion)(&(&slave->notify_work)->work) rtnl_mutex &tn->lock irq_context: 0 (wq_completion)bond189#2 (work_completion)(&(&slave->notify_work)->work) rtnl_mutex net_rwsem irq_context: 0 (wq_completion)bond189#2 (work_completion)(&(&slave->notify_work)->work) rtnl_mutex net_rwsem &list->lock#2 irq_context: 0 (wq_completion)bond189#2 (work_completion)(&(&slave->notify_work)->work) rtnl_mutex &tn->lock irq_context: 0 (wq_completion)bond147#3 (work_completion)(&(&slave->notify_work)->work) &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)bond147#3 (work_completion)(&(&slave->notify_work)->work) rtnl_mutex net_rwsem irq_context: 0 (wq_completion)bond147#3 (work_completion)(&(&slave->notify_work)->work) rtnl_mutex net_rwsem &list->lock#2 irq_context: 0 (wq_completion)bond147#3 (work_completion)(&(&slave->notify_work)->work) rtnl_mutex &tn->lock irq_context: 0 (wq_completion)bond187#2 (work_completion)(&(&slave->notify_work)->work) rtnl_mutex net_rwsem irq_context: 0 (wq_completion)bond187#2 (work_completion)(&(&slave->notify_work)->work) rtnl_mutex net_rwsem &list->lock#2 irq_context: 0 (wq_completion)bond187#2 (work_completion)(&(&slave->notify_work)->work) rtnl_mutex &tn->lock irq_context: 0 (wq_completion)bond198 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex fs_reclaim irq_context: 0 (wq_completion)bond198 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex fs_reclaim mmu_notifier_invalidate_range_start irq_context: 0 (wq_completion)bond198 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &rq->__lock irq_context: 0 (wq_completion)bond198 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)bond198 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex rcu_read_lock &bond->stats_lock/1 irq_context: 0 (wq_completion)bond198 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex rcu_read_lock &ndev->lock irq_context: 0 (wq_completion)bond198 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex nl_table_lock irq_context: 0 (wq_completion)bond198 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &obj_hash[i].lock irq_context: 0 (wq_completion)bond198 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex nl_table_wait.lock irq_context: 0 (wq_completion)bond198 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex net_rwsem irq_context: 0 (wq_completion)bond198 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex net_rwsem &list->lock#2 irq_context: 0 (wq_completion)bond198 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &tn->lock irq_context: 0 (wq_completion)bond198 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &idev->mc_lock irq_context: 0 (wq_completion)bond198 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &idev->mc_lock fs_reclaim irq_context: 0 (wq_completion)bond198 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &idev->mc_lock fs_reclaim &rq->__lock irq_context: 0 (wq_completion)bond198 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &idev->mc_lock fs_reclaim &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)bond198 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &idev->mc_lock fs_reclaim mmu_notifier_invalidate_range_start irq_context: 0 (wq_completion)bond198 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &idev->mc_lock rcu_read_lock &ndev->lock irq_context: 0 (wq_completion)bond198 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &idev->mc_lock rcu_read_lock &dir->lock#2 irq_context: 0 (wq_completion)bond198 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &idev->mc_lock rcu_read_lock &ul->lock#2 irq_context: 0 (wq_completion)bond198 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &idev->mc_lock rcu_read_lock rcu_read_lock &____s->seqcount#10 irq_context: 0 (wq_completion)bond198 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &idev->mc_lock rcu_read_lock rcu_read_lock rcu_node_0 irq_context: 0 (wq_completion)bond198 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &idev->mc_lock rcu_read_lock rcu_read_lock &rq->__lock irq_context: 0 (wq_completion)bond198 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &idev->mc_lock rcu_read_lock rcu_read_lock rcu_read_lock_bh rcu_read_lock rcu_read_lock_bh rcu_read_lock &br->multicast_lock irq_context: 0 (wq_completion)bond198 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &idev->mc_lock rcu_read_lock rcu_read_lock rcu_read_lock_bh rcu_read_lock rcu_read_lock_bh rcu_read_lock &br->multicast_lock &obj_hash[i].lock irq_context: 0 (wq_completion)bond198 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &idev->mc_lock rcu_read_lock rcu_read_lock rcu_read_lock_bh rcu_read_lock rcu_read_lock_bh rcu_read_lock &br->multicast_lock &base->lock irq_context: 0 (wq_completion)bond198 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &idev->mc_lock rcu_read_lock rcu_read_lock rcu_read_lock_bh rcu_read_lock rcu_read_lock_bh rcu_read_lock &data->lock irq_context: 0 (wq_completion)bond198 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &idev->mc_lock rcu_read_lock rcu_read_lock rcu_read_lock_bh rcu_read_lock rcu_read_lock_bh rcu_read_lock &obj_hash[i].lock irq_context: 0 (wq_completion)bond198 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &idev->mc_lock rcu_read_lock rcu_node_0 irq_context: 0 (wq_completion)bond198 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex rcu_read_lock rcu_node_0 irq_context: 0 (wq_completion)bond198 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex rcu_read_lock &rq->__lock irq_context: 0 (wq_completion)bond198 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex.wait_lock irq_context: 0 (wq_completion)bond198 (work_completion)(&(&bond->mcast_work)->work) &p->pi_lock irq_context: 0 (wq_completion)bond198 (work_completion)(&(&bond->mcast_work)->work) &p->pi_lock &rq->__lock irq_context: 0 (wq_completion)bond198 (work_completion)(&(&bond->mcast_work)->work) &p->pi_lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)bond198 (work_completion)(&(&bond->mcast_work)->work) &rq->__lock irq_context: 0 (wq_completion)bond198 (work_completion)(&(&slave->notify_work)->work) rtnl_mutex net_rwsem irq_context: 0 (wq_completion)bond198 (work_completion)(&(&slave->notify_work)->work) rtnl_mutex net_rwsem &list->lock#2 irq_context: 0 (wq_completion)bond198 (work_completion)(&(&slave->notify_work)->work) rtnl_mutex &tn->lock irq_context: 0 (wq_completion)bond198 (work_completion)(&(&slave->notify_work)->work) rtnl_mutex.wait_lock irq_context: 0 (wq_completion)bond198 (work_completion)(&(&slave->notify_work)->work) &p->pi_lock irq_context: 0 (wq_completion)bond21#3 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex fs_reclaim irq_context: 0 (wq_completion)bond21#3 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex fs_reclaim mmu_notifier_invalidate_range_start irq_context: 0 (wq_completion)bond21#3 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &c->lock irq_context: 0 (wq_completion)bond21#3 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex rcu_read_lock &bond->stats_lock/1 irq_context: 0 (wq_completion)bond21#3 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex rcu_read_lock &ndev->lock irq_context: 0 (wq_completion)bond21#3 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex nl_table_lock irq_context: 0 (wq_completion)bond21#3 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &obj_hash[i].lock irq_context: 0 (wq_completion)bond21#3 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex nl_table_wait.lock irq_context: 0 (wq_completion)bond21#3 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex net_rwsem irq_context: 0 (wq_completion)bond21#3 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex net_rwsem &list->lock#2 irq_context: 0 (wq_completion)bond21#3 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &tn->lock irq_context: 0 (wq_completion)bond21#3 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &idev->mc_lock irq_context: 0 (wq_completion)bond21#3 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &idev->mc_lock fs_reclaim irq_context: 0 (wq_completion)bond21#3 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &idev->mc_lock fs_reclaim mmu_notifier_invalidate_range_start irq_context: 0 (wq_completion)bond21#3 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &idev->mc_lock rcu_read_lock &ndev->lock irq_context: 0 (wq_completion)bond21#3 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &idev->mc_lock rcu_read_lock &dir->lock#2 irq_context: 0 (wq_completion)bond21#3 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &idev->mc_lock rcu_read_lock &ul->lock#2 irq_context: 0 (wq_completion)bond21#3 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &idev->mc_lock rcu_read_lock rcu_read_lock &____s->seqcount#12 irq_context: 0 (wq_completion)bond21#3 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &idev->mc_lock rcu_read_lock rcu_read_lock &____s->seqcount#16 irq_context: 0 (wq_completion)bond21#3 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &idev->mc_lock rcu_read_lock &obj_hash[i].lock irq_context: 0 (wq_completion)bond21#3 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &idev->mc_lock rcu_read_lock &data->lock irq_context: 0 (wq_completion)bond175#2 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex fs_reclaim irq_context: 0 (wq_completion)bond175#2 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex fs_reclaim mmu_notifier_invalidate_range_start irq_context: 0 (wq_completion)bond175#2 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &c->lock irq_context: 0 (wq_completion)bond175#2 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex rcu_read_lock &bond->stats_lock/1 irq_context: 0 (wq_completion)bond175#2 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex rcu_read_lock &ndev->lock irq_context: 0 (wq_completion)bond175#2 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex nl_table_lock irq_context: 0 (wq_completion)bond175#2 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &obj_hash[i].lock irq_context: 0 (wq_completion)bond175#2 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex nl_table_wait.lock irq_context: 0 (wq_completion)bond175#2 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex net_rwsem irq_context: 0 (wq_completion)bond175#2 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex net_rwsem &list->lock#2 irq_context: 0 (wq_completion)bond175#2 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &tn->lock irq_context: 0 (wq_completion)bond175#2 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &idev->mc_lock irq_context: 0 (wq_completion)bond175#2 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &idev->mc_lock fs_reclaim irq_context: 0 (wq_completion)bond175#2 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &idev->mc_lock fs_reclaim mmu_notifier_invalidate_range_start irq_context: 0 (wq_completion)bond175#2 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &idev->mc_lock &c->lock irq_context: 0 (wq_completion)bond175#2 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &idev->mc_lock rcu_read_lock &ndev->lock irq_context: 0 (wq_completion)bond175#2 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &idev->mc_lock rcu_read_lock &c->lock irq_context: 0 (wq_completion)bond175#2 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &idev->mc_lock rcu_read_lock &dir->lock#2 irq_context: 0 (wq_completion)bond175#2 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &idev->mc_lock rcu_read_lock &ul->lock#2 irq_context: 0 (wq_completion)bond175#2 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &idev->mc_lock rcu_read_lock rcu_read_lock &____s->seqcount#10 irq_context: 0 (wq_completion)bond175#2 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &idev->mc_lock rcu_read_lock rcu_read_lock rcu_read_lock_bh rcu_read_lock rcu_read_lock_bh rcu_read_lock &br->multicast_lock irq_context: 0 (wq_completion)bond175#2 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &idev->mc_lock rcu_read_lock rcu_read_lock rcu_read_lock_bh rcu_read_lock rcu_read_lock_bh rcu_read_lock &br->multicast_lock &obj_hash[i].lock irq_context: 0 (wq_completion)bond175#2 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &idev->mc_lock rcu_read_lock rcu_read_lock rcu_read_lock_bh rcu_read_lock rcu_read_lock_bh rcu_read_lock &br->multicast_lock &base->lock irq_context: 0 (wq_completion)bond175#2 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &idev->mc_lock rcu_read_lock rcu_read_lock rcu_read_lock_bh rcu_read_lock rcu_read_lock_bh rcu_read_lock &data->lock irq_context: 0 (wq_completion)bond175#2 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &idev->mc_lock rcu_read_lock rcu_read_lock rcu_read_lock_bh rcu_read_lock rcu_read_lock_bh rcu_read_lock &obj_hash[i].lock irq_context: 0 (wq_completion)bond150#3 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex fs_reclaim irq_context: 0 (wq_completion)bond150#3 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex fs_reclaim mmu_notifier_invalidate_range_start irq_context: 0 (wq_completion)bond150#3 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex rcu_read_lock &bond->stats_lock/1 irq_context: 0 (wq_completion)bond150#3 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex rcu_read_lock &ndev->lock irq_context: 0 (wq_completion)bond150#3 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex nl_table_lock irq_context: 0 (wq_completion)bond150#3 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &obj_hash[i].lock irq_context: 0 (wq_completion)bond150#3 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex nl_table_wait.lock irq_context: 0 (wq_completion)bond150#3 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex net_rwsem irq_context: 0 (wq_completion)bond150#3 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex net_rwsem &list->lock#2 irq_context: 0 (wq_completion)bond150#3 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &tn->lock irq_context: 0 (wq_completion)bond150#3 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &idev->mc_lock irq_context: 0 (wq_completion)bond150#3 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &idev->mc_lock fs_reclaim irq_context: 0 (wq_completion)bond150#3 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &idev->mc_lock fs_reclaim mmu_notifier_invalidate_range_start irq_context: 0 (wq_completion)bond150#3 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &idev->mc_lock &c->lock irq_context: 0 (wq_completion)bond150#3 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &idev->mc_lock rcu_read_lock &ndev->lock irq_context: 0 (wq_completion)bond150#3 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &idev->mc_lock rcu_read_lock &dir->lock#2 irq_context: 0 (wq_completion)bond150#3 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &idev->mc_lock rcu_read_lock &c->lock irq_context: 0 (wq_completion)bond150#3 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &idev->mc_lock rcu_read_lock &ul->lock#2 irq_context: 0 (wq_completion)bond150#3 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &idev->mc_lock rcu_read_lock rcu_read_lock &____s->seqcount#10 irq_context: 0 (wq_completion)bond150#3 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &idev->mc_lock rcu_read_lock rcu_read_lock rcu_read_lock_bh rcu_read_lock rcu_read_lock_bh rcu_read_lock &br->multicast_lock irq_context: 0 (wq_completion)bond150#3 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &idev->mc_lock rcu_read_lock rcu_read_lock rcu_read_lock_bh rcu_read_lock rcu_read_lock_bh rcu_read_lock &br->multicast_lock &obj_hash[i].lock irq_context: 0 (wq_completion)bond150#3 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &idev->mc_lock rcu_read_lock rcu_read_lock rcu_read_lock_bh rcu_read_lock rcu_read_lock_bh rcu_read_lock &br->multicast_lock &base->lock irq_context: 0 (wq_completion)bond150#3 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &idev->mc_lock rcu_read_lock rcu_read_lock rcu_read_lock_bh rcu_read_lock rcu_read_lock_bh rcu_read_lock &br->multicast_lock &base->lock &obj_hash[i].lock irq_context: 0 (wq_completion)bond150#3 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &idev->mc_lock rcu_read_lock rcu_read_lock rcu_read_lock_bh rcu_read_lock rcu_read_lock_bh rcu_read_lock &data->lock irq_context: 0 (wq_completion)bond150#3 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &idev->mc_lock rcu_read_lock rcu_read_lock rcu_read_lock_bh rcu_read_lock rcu_read_lock_bh rcu_read_lock &obj_hash[i].lock irq_context: 0 (wq_completion)bond177#3 (work_completion)(&(&slave->notify_work)->work) rtnl_mutex net_rwsem irq_context: 0 (wq_completion)bond177#3 (work_completion)(&(&slave->notify_work)->work) rtnl_mutex net_rwsem &list->lock#2 irq_context: 0 (wq_completion)bond177#3 (work_completion)(&(&slave->notify_work)->work) rtnl_mutex &tn->lock irq_context: 0 (wq_completion)bond192#2 (work_completion)(&(&slave->notify_work)->work) rtnl_mutex net_rwsem irq_context: 0 (wq_completion)bond192#2 (work_completion)(&(&slave->notify_work)->work) rtnl_mutex net_rwsem &list->lock#2 irq_context: 0 (wq_completion)bond192#2 (work_completion)(&(&slave->notify_work)->work) rtnl_mutex &tn->lock irq_context: 0 (wq_completion)bond163#3 (work_completion)(&(&slave->notify_work)->work) rtnl_mutex net_rwsem irq_context: 0 (wq_completion)bond163#3 (work_completion)(&(&slave->notify_work)->work) rtnl_mutex net_rwsem &list->lock#2 irq_context: 0 (wq_completion)bond163#3 (work_completion)(&(&slave->notify_work)->work) rtnl_mutex &tn->lock irq_context: 0 (wq_completion)bond196 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex fs_reclaim irq_context: 0 (wq_completion)bond196 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex fs_reclaim mmu_notifier_invalidate_range_start irq_context: 0 (wq_completion)bond196 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex batched_entropy_u8.lock irq_context: 0 (wq_completion)bond196 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex kfence_freelist_lock irq_context: 0 (wq_completion)bond196 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &c->lock irq_context: 0 (wq_completion)bond196 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex rcu_read_lock &bond->stats_lock/1 irq_context: 0 (wq_completion)bond196 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex rcu_read_lock &ndev->lock irq_context: 0 (wq_completion)bond196 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex nl_table_lock irq_context: 0 (wq_completion)bond196 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &obj_hash[i].lock irq_context: 0 (wq_completion)bond196 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &meta->lock irq_context: 0 (wq_completion)bond196 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex nl_table_wait.lock irq_context: 0 (wq_completion)bond196 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex net_rwsem irq_context: 0 (wq_completion)bond196 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex net_rwsem &list->lock#2 irq_context: 0 (wq_completion)bond196 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &tn->lock irq_context: 0 (wq_completion)bond196 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &idev->mc_lock irq_context: 0 (wq_completion)bond196 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &idev->mc_lock fs_reclaim irq_context: 0 (wq_completion)bond196 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &idev->mc_lock fs_reclaim mmu_notifier_invalidate_range_start irq_context: 0 (wq_completion)bond196 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &idev->mc_lock rcu_read_lock &ndev->lock irq_context: 0 (wq_completion)bond196 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &idev->mc_lock rcu_read_lock rcu_node_0 irq_context: 0 (wq_completion)bond196 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &idev->mc_lock rcu_read_lock &rq->__lock irq_context: 0 (wq_completion)bond196 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &idev->mc_lock rcu_read_lock &dir->lock#2 irq_context: 0 (wq_completion)bond196 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &idev->mc_lock rcu_read_lock &ul->lock#2 irq_context: 0 (wq_completion)bond196 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &idev->mc_lock rcu_read_lock rcu_read_lock &____s->seqcount#10 irq_context: 0 (wq_completion)bond196 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &idev->mc_lock rcu_read_lock rcu_read_lock rcu_read_lock_bh rcu_read_lock rcu_read_lock_bh rcu_read_lock &br->multicast_lock irq_context: 0 (wq_completion)bond196 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &idev->mc_lock rcu_read_lock rcu_read_lock rcu_read_lock_bh rcu_read_lock rcu_read_lock_bh rcu_read_lock &br->multicast_lock &obj_hash[i].lock irq_context: 0 (wq_completion)bond196 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &idev->mc_lock rcu_read_lock rcu_read_lock rcu_read_lock_bh rcu_read_lock rcu_read_lock_bh rcu_read_lock &br->multicast_lock &base->lock irq_context: 0 (wq_completion)bond196 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &idev->mc_lock rcu_read_lock rcu_read_lock rcu_read_lock_bh rcu_read_lock rcu_read_lock_bh rcu_read_lock &data->lock irq_context: 0 (wq_completion)bond196 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &idev->mc_lock rcu_read_lock rcu_read_lock rcu_read_lock_bh rcu_read_lock rcu_read_lock_bh rcu_read_lock &obj_hash[i].lock irq_context: 0 (wq_completion)bond196 (work_completion)(&(&slave->notify_work)->work) rtnl_mutex net_rwsem irq_context: 0 (wq_completion)bond196 (work_completion)(&(&slave->notify_work)->work) rtnl_mutex net_rwsem &list->lock#2 irq_context: 0 (wq_completion)bond196 (work_completion)(&(&slave->notify_work)->work) rtnl_mutex &tn->lock irq_context: 0 (wq_completion)bond53#3 &rq->__lock irq_context: 0 (wq_completion)bond53#3 &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)bond182#2 (work_completion)(&(&slave->notify_work)->work) rtnl_mutex net_rwsem irq_context: 0 (wq_completion)bond182#2 (work_completion)(&(&slave->notify_work)->work) rtnl_mutex net_rwsem &list->lock#2 irq_context: 0 (wq_completion)bond182#2 (work_completion)(&(&slave->notify_work)->work) rtnl_mutex &tn->lock irq_context: 0 (wq_completion)bond182#2 (work_completion)(&(&slave->notify_work)->work) rtnl_mutex.wait_lock irq_context: 0 (wq_completion)bond182#2 (work_completion)(&(&slave->notify_work)->work) &p->pi_lock irq_context: 0 (wq_completion)ipv6_addrconf (work_completion)(&(&ifa->dad_work)->work) rtnl_mutex &idev->mc_lock rcu_read_lock rcu_read_lock rcu_read_lock_bh rcu_read_lock rcu_read_lock_bh rcu_read_lock &c->lock irq_context: 0 (wq_completion)ipv6_addrconf (work_completion)(&(&ifa->dad_work)->work) rtnl_mutex &idev->mc_lock rcu_read_lock rcu_read_lock rcu_read_lock_bh rcu_read_lock rcu_read_lock_bh rcu_read_lock &n->list_lock irq_context: 0 (wq_completion)ipv6_addrconf (work_completion)(&(&ifa->dad_work)->work) rtnl_mutex &idev->mc_lock rcu_read_lock rcu_read_lock rcu_read_lock_bh rcu_read_lock rcu_read_lock_bh rcu_read_lock &n->list_lock &c->lock irq_context: 0 (wq_completion)bond169#3 (work_completion)(&(&bond->mcast_work)->work) &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)bond191 &rq->__lock irq_context: 0 (wq_completion)bond191 &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)bond150 &rq->__lock irq_context: 0 (wq_completion)bond150 &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)bond120 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex fs_reclaim irq_context: 0 (wq_completion)bond120 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex fs_reclaim mmu_notifier_invalidate_range_start irq_context: 0 (wq_completion)bond120 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex rcu_read_lock &bond->stats_lock/1 irq_context: 0 (wq_completion)bond120 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex rcu_read_lock &ndev->lock irq_context: 0 (wq_completion)bond120 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex nl_table_lock irq_context: 0 (wq_completion)bond120 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &obj_hash[i].lock irq_context: 0 (wq_completion)bond120 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex nl_table_wait.lock irq_context: 0 (wq_completion)bond120 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex net_rwsem irq_context: 0 (wq_completion)bond120 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex net_rwsem &list->lock#2 irq_context: 0 (wq_completion)bond120 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &tn->lock irq_context: 0 (wq_completion)bond120 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &idev->mc_lock irq_context: 0 (wq_completion)bond120 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &idev->mc_lock fs_reclaim irq_context: 0 (wq_completion)bond120 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &idev->mc_lock fs_reclaim mmu_notifier_invalidate_range_start irq_context: 0 (wq_completion)bond120 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &idev->mc_lock &c->lock irq_context: 0 (wq_completion)bond120 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &idev->mc_lock rcu_read_lock &ndev->lock irq_context: 0 (wq_completion)bond120 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &idev->mc_lock rcu_read_lock &c->lock irq_context: 0 (wq_completion)bond120 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &idev->mc_lock rcu_read_lock &dir->lock#2 irq_context: 0 (wq_completion)bond120 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &idev->mc_lock rcu_read_lock &ul->lock#2 irq_context: 0 (wq_completion)bond120 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &idev->mc_lock rcu_read_lock rcu_read_lock &obj_hash[i].lock irq_context: 0 (wq_completion)bond120 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &idev->mc_lock rcu_read_lock rcu_read_lock &dir->lock#2 irq_context: 0 (wq_completion)bond120 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &idev->mc_lock rcu_read_lock rcu_read_lock &tbl->lock irq_context: 0 (wq_completion)bond120 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &idev->mc_lock rcu_read_lock rcu_read_lock &n->lock irq_context: 0 (wq_completion)bond120 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &idev->mc_lock rcu_read_lock rcu_read_lock &____s->seqcount#9 irq_context: 0 (wq_completion)bond120 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &idev->mc_lock rcu_read_lock rcu_read_lock rcu_read_lock_bh rcu_read_lock tk_core.seq.seqcount irq_context: 0 (wq_completion)bond120 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &idev->mc_lock rcu_read_lock rcu_read_lock rcu_read_lock_bh rcu_read_lock &obj_hash[i].lock irq_context: 0 (wq_completion)bond120 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &idev->mc_lock rcu_read_lock rcu_read_lock rcu_read_lock_bh rcu_read_lock rcu_read_lock_bh rcu_read_lock tk_core.seq.seqcount irq_context: 0 (wq_completion)bond120 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &idev->mc_lock rcu_read_lock rcu_read_lock rcu_read_lock_bh rcu_read_lock rcu_read_lock_bh rcu_read_lock &obj_hash[i].lock irq_context: 0 (wq_completion)bond120 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &idev->mc_lock rcu_read_lock rcu_read_lock rcu_read_lock_bh rcu_read_lock rcu_read_lock_bh rcu_read_lock &br->multicast_lock irq_context: 0 (wq_completion)bond120 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &idev->mc_lock rcu_read_lock rcu_read_lock rcu_read_lock_bh rcu_read_lock rcu_read_lock_bh rcu_read_lock &br->multicast_lock &obj_hash[i].lock irq_context: 0 (wq_completion)bond120 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &idev->mc_lock rcu_read_lock rcu_read_lock rcu_read_lock_bh rcu_read_lock rcu_read_lock_bh rcu_read_lock &br->multicast_lock &base->lock irq_context: 0 (wq_completion)bond120 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &idev->mc_lock rcu_read_lock rcu_read_lock rcu_read_lock_bh rcu_read_lock rcu_read_lock_bh rcu_read_lock &br->multicast_lock &base->lock &obj_hash[i].lock irq_context: 0 (wq_completion)bond120 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &idev->mc_lock rcu_read_lock rcu_read_lock rcu_read_lock_bh rcu_read_lock rcu_read_lock_bh rcu_read_lock &data->lock irq_context: 0 (wq_completion)bond120 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex.wait_lock irq_context: 0 (wq_completion)bond120 (work_completion)(&(&bond->mcast_work)->work) &p->pi_lock irq_context: 0 (wq_completion)bond187 (work_completion)(&(&slave->notify_work)->work) rtnl_mutex net_rwsem irq_context: 0 (wq_completion)bond187 (work_completion)(&(&slave->notify_work)->work) rtnl_mutex net_rwsem &list->lock#2 irq_context: 0 (wq_completion)bond187 (work_completion)(&(&slave->notify_work)->work) rtnl_mutex &tn->lock irq_context: 0 (wq_completion)bond187 (work_completion)(&(&slave->notify_work)->work) rtnl_mutex.wait_lock irq_context: 0 (wq_completion)bond187 (work_completion)(&(&slave->notify_work)->work) &p->pi_lock irq_context: 0 (wq_completion)bond112 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex fs_reclaim irq_context: 0 (wq_completion)bond112 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex fs_reclaim mmu_notifier_invalidate_range_start irq_context: 0 (wq_completion)bond112 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex rcu_read_lock &bond->stats_lock/1 irq_context: 0 (wq_completion)bond112 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex rcu_read_lock &ndev->lock irq_context: 0 (wq_completion)bond112 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex nl_table_lock irq_context: 0 (wq_completion)bond112 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &obj_hash[i].lock irq_context: 0 (wq_completion)bond112 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex nl_table_wait.lock irq_context: 0 (wq_completion)bond112 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex net_rwsem irq_context: 0 (wq_completion)bond112 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex net_rwsem &list->lock#2 irq_context: 0 (wq_completion)bond112 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &tn->lock irq_context: 0 (wq_completion)bond112 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &idev->mc_lock irq_context: 0 (wq_completion)bond112 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &idev->mc_lock fs_reclaim irq_context: 0 (wq_completion)bond112 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &idev->mc_lock fs_reclaim mmu_notifier_invalidate_range_start irq_context: 0 (wq_completion)bond112 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &idev->mc_lock &c->lock irq_context: 0 (wq_completion)bond112 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &idev->mc_lock &n->list_lock irq_context: 0 (wq_completion)bond112 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &idev->mc_lock &n->list_lock &c->lock irq_context: 0 (wq_completion)bond112 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &idev->mc_lock rcu_read_lock &ndev->lock irq_context: 0 (wq_completion)bond112 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &idev->mc_lock rcu_read_lock &dir->lock#2 irq_context: 0 (wq_completion)bond112 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &idev->mc_lock rcu_read_lock &c->lock irq_context: 0 (wq_completion)bond112 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &idev->mc_lock rcu_read_lock &ul->lock#2 irq_context: 0 (wq_completion)bond112 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &idev->mc_lock rcu_read_lock rcu_read_lock &obj_hash[i].lock irq_context: 0 (wq_completion)bond112 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &idev->mc_lock rcu_read_lock rcu_read_lock &dir->lock#2 irq_context: 0 (wq_completion)bond112 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &idev->mc_lock rcu_read_lock rcu_read_lock &tbl->lock irq_context: 0 (wq_completion)bond112 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &idev->mc_lock rcu_read_lock rcu_read_lock &n->lock irq_context: 0 (wq_completion)bond112 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &idev->mc_lock rcu_read_lock rcu_read_lock &____s->seqcount#9 irq_context: 0 (wq_completion)bond112 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &idev->mc_lock rcu_read_lock rcu_read_lock rcu_read_lock_bh rcu_read_lock tk_core.seq.seqcount irq_context: 0 (wq_completion)bond112 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &idev->mc_lock rcu_read_lock rcu_read_lock rcu_read_lock_bh rcu_read_lock &obj_hash[i].lock irq_context: 0 (wq_completion)bond112 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &idev->mc_lock rcu_read_lock rcu_read_lock rcu_read_lock_bh rcu_read_lock rcu_read_lock_bh rcu_read_lock tk_core.seq.seqcount irq_context: 0 (wq_completion)bond112 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &idev->mc_lock rcu_read_lock rcu_read_lock rcu_read_lock_bh rcu_read_lock rcu_read_lock_bh rcu_read_lock &obj_hash[i].lock irq_context: 0 (wq_completion)bond112 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &idev->mc_lock rcu_read_lock rcu_read_lock rcu_read_lock_bh rcu_read_lock rcu_read_lock_bh rcu_read_lock &br->multicast_lock irq_context: 0 (wq_completion)bond112 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &idev->mc_lock rcu_read_lock rcu_read_lock rcu_read_lock_bh rcu_read_lock rcu_read_lock_bh rcu_read_lock &br->multicast_lock &obj_hash[i].lock irq_context: 0 (wq_completion)bond112 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &idev->mc_lock rcu_read_lock rcu_read_lock rcu_read_lock_bh rcu_read_lock rcu_read_lock_bh rcu_read_lock &br->multicast_lock &base->lock irq_context: 0 (wq_completion)bond112 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &idev->mc_lock rcu_read_lock rcu_read_lock rcu_read_lock_bh rcu_read_lock rcu_read_lock_bh rcu_read_lock &br->multicast_lock &base->lock &obj_hash[i].lock irq_context: 0 (wq_completion)bond112 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &idev->mc_lock rcu_read_lock rcu_read_lock rcu_read_lock_bh rcu_read_lock rcu_read_lock_bh rcu_read_lock &data->lock irq_context: 0 (wq_completion)bond112 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &idev->mc_lock rcu_read_lock rcu_read_lock rcu_node_0 irq_context: 0 (wq_completion)bond112 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &idev->mc_lock rcu_read_lock rcu_read_lock &rq->__lock irq_context: 0 (wq_completion)bond112 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &idev->mc_lock rcu_read_lock rcu_read_lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)bond112 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &idev->mc_lock rcu_read_lock rcu_node_0 irq_context: 0 (wq_completion)bond112 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &idev->mc_lock rcu_read_lock &rq->__lock irq_context: 0 (wq_completion)bond112 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &idev->mc_lock rcu_read_lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)bond112 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex.wait_lock irq_context: 0 (wq_completion)bond112 (work_completion)(&(&bond->mcast_work)->work) &p->pi_lock irq_context: 0 (wq_completion)bond112 (work_completion)(&(&bond->mcast_work)->work) &p->pi_lock &rq->__lock irq_context: 0 (wq_completion)bond112 (work_completion)(&(&bond->mcast_work)->work) &p->pi_lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)bond194#2 irq_context: 0 (wq_completion)bond194#2 &rq->__lock irq_context: 0 (wq_completion)bond194#2 &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)bond194#2 (work_completion)(&(&slave->notify_work)->work) irq_context: 0 (wq_completion)bond194#2 (work_completion)(&(&slave->notify_work)->work) &obj_hash[i].lock irq_context: 0 (wq_completion)bond194#2 (work_completion)(&(&slave->notify_work)->work) &base->lock irq_context: 0 (wq_completion)bond194#2 (work_completion)(&(&slave->notify_work)->work) &base->lock &obj_hash[i].lock irq_context: 0 (wq_completion)bond194#2 (work_completion)(&(&slave->notify_work)->work) &rq->__lock irq_context: 0 (wq_completion)bond194#2 (work_completion)(&(&slave->notify_work)->work) &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)bond194#2 (work_completion)(&(&bond->mcast_work)->work) irq_context: 0 (wq_completion)bond194#2 (work_completion)(&(&bond->mcast_work)->work) &obj_hash[i].lock irq_context: 0 (wq_completion)bond194#2 (work_completion)(&(&bond->mcast_work)->work) &base->lock irq_context: 0 (wq_completion)bond194#2 (work_completion)(&(&bond->mcast_work)->work) &base->lock &obj_hash[i].lock irq_context: 0 (wq_completion)bond178#3 irq_context: 0 (wq_completion)bond178#3 (work_completion)(&(&slave->notify_work)->work) irq_context: 0 (wq_completion)bond178#3 (work_completion)(&(&slave->notify_work)->work) &obj_hash[i].lock irq_context: 0 (wq_completion)bond178#3 (work_completion)(&(&slave->notify_work)->work) &base->lock irq_context: 0 (wq_completion)bond178#3 (work_completion)(&(&slave->notify_work)->work) &base->lock &obj_hash[i].lock irq_context: 0 (wq_completion)bond176#2 &rq->__lock irq_context: 0 (wq_completion)bond176#2 &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)bond178#3 (work_completion)(&(&bond->mcast_work)->work) irq_context: 0 (wq_completion)bond178#3 (work_completion)(&(&bond->mcast_work)->work) &obj_hash[i].lock irq_context: 0 (wq_completion)bond178#3 (work_completion)(&(&bond->mcast_work)->work) &base->lock irq_context: 0 (wq_completion)bond178#3 (work_completion)(&(&bond->mcast_work)->work) &base->lock &obj_hash[i].lock irq_context: 0 (wq_completion)bond178#3 (work_completion)(&(&bond->mcast_work)->work) &rq->__lock irq_context: 0 (wq_completion)bond178#3 (work_completion)(&(&bond->mcast_work)->work) &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)bond206 irq_context: 0 (wq_completion)bond206 (work_completion)(&(&slave->notify_work)->work) irq_context: 0 (wq_completion)bond206 (work_completion)(&(&slave->notify_work)->work) &obj_hash[i].lock irq_context: 0 (wq_completion)bond206 (work_completion)(&(&slave->notify_work)->work) &base->lock irq_context: 0 (wq_completion)bond206 (work_completion)(&(&slave->notify_work)->work) &base->lock &obj_hash[i].lock irq_context: 0 (wq_completion)bond206 (work_completion)(&(&bond->mcast_work)->work) irq_context: 0 (wq_completion)bond206 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex fs_reclaim irq_context: 0 (wq_completion)bond206 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex fs_reclaim mmu_notifier_invalidate_range_start irq_context: 0 (wq_completion)bond206 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &c->lock irq_context: 0 (wq_completion)bond206 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex rcu_read_lock &bond->stats_lock/1 irq_context: 0 (wq_completion)bond206 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex rcu_read_lock &ndev->lock irq_context: 0 (wq_completion)bond206 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex nl_table_lock irq_context: 0 (wq_completion)bond206 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &obj_hash[i].lock irq_context: 0 (wq_completion)bond206 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex nl_table_wait.lock irq_context: 0 (wq_completion)bond206 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex net_rwsem irq_context: 0 (wq_completion)bond206 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex net_rwsem &list->lock#2 irq_context: 0 (wq_completion)bond206 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &tn->lock irq_context: 0 (wq_completion)bond206 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &idev->mc_lock irq_context: 0 (wq_completion)bond206 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex.wait_lock irq_context: 0 (wq_completion)bond206 (work_completion)(&(&bond->mcast_work)->work) &p->pi_lock irq_context: 0 (wq_completion)events (linkwatch_work).work rtnl_mutex &idev->mc_lock _xmit_ETHER &c->lock irq_context: 0 (wq_completion)events (linkwatch_work).work rtnl_mutex &idev->mc_lock _xmit_ETHER &____s->seqcount#2 irq_context: 0 (wq_completion)events (linkwatch_work).work rtnl_mutex &idev->mc_lock _xmit_ETHER &____s->seqcount irq_context: 0 (wq_completion)events (linkwatch_work).work rtnl_mutex lweventlist_lock &n->list_lock irq_context: 0 (wq_completion)events (linkwatch_work).work rtnl_mutex lweventlist_lock &n->list_lock &c->lock irq_context: 0 (wq_completion)bond153#3 (work_completion)(&(&bond->mcast_work)->work) &rq->__lock irq_context: 0 (wq_completion)mld (work_completion)(&(&idev->mc_dad_work)->work) &idev->mc_lock rcu_read_lock rcu_read_lock rcu_read_lock_bh rcu_read_lock rcu_read_lock_bh rcu_read_lock &c->lock irq_context: 0 (wq_completion)bond195#2 irq_context: 0 (wq_completion)bond195#2 &rq->__lock irq_context: 0 (wq_completion)bond195#2 &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)bond195#2 (work_completion)(&(&slave->notify_work)->work) irq_context: 0 (wq_completion)bond195#2 (work_completion)(&(&slave->notify_work)->work) &obj_hash[i].lock irq_context: 0 (wq_completion)bond195#2 (work_completion)(&(&slave->notify_work)->work) &base->lock irq_context: 0 (wq_completion)bond195#2 (work_completion)(&(&slave->notify_work)->work) &base->lock &obj_hash[i].lock irq_context: 0 (wq_completion)bond189 &rq->__lock irq_context: 0 (wq_completion)bond189 &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)bond195#2 (work_completion)(&(&bond->mcast_work)->work) irq_context: 0 (wq_completion)bond195#2 (work_completion)(&(&bond->mcast_work)->work) &obj_hash[i].lock irq_context: 0 (wq_completion)bond195#2 (work_completion)(&(&bond->mcast_work)->work) &base->lock irq_context: 0 (wq_completion)bond195#2 (work_completion)(&(&bond->mcast_work)->work) &base->lock &obj_hash[i].lock irq_context: 0 (wq_completion)bond204 &rq->__lock irq_context: 0 (wq_completion)bond166 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex fs_reclaim irq_context: 0 (wq_completion)bond166 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex fs_reclaim mmu_notifier_invalidate_range_start irq_context: 0 (wq_completion)bond166 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex batched_entropy_u8.lock irq_context: 0 (wq_completion)bond166 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex kfence_freelist_lock irq_context: 0 (wq_completion)bond166 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &c->lock irq_context: 0 (wq_completion)bond166 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &rq->__lock irq_context: 0 (wq_completion)bond166 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)bond166 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex rcu_read_lock &bond->stats_lock/1 irq_context: 0 (wq_completion)bond166 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex rcu_read_lock &ndev->lock irq_context: 0 (wq_completion)bond166 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex nl_table_lock irq_context: 0 (wq_completion)bond166 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &obj_hash[i].lock irq_context: 0 (wq_completion)bond166 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &meta->lock irq_context: 0 (wq_completion)bond166 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex nl_table_wait.lock irq_context: 0 (wq_completion)bond166 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex net_rwsem irq_context: 0 (wq_completion)bond166 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex net_rwsem &list->lock#2 irq_context: 0 (wq_completion)bond166 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &tn->lock irq_context: 0 (wq_completion)bond166 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &idev->mc_lock irq_context: 0 (wq_completion)bond166 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &idev->mc_lock &rq->__lock irq_context: 0 (wq_completion)bond166 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &idev->mc_lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)bond166 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &idev->mc_lock fs_reclaim irq_context: 0 (wq_completion)bond166 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &idev->mc_lock fs_reclaim mmu_notifier_invalidate_range_start irq_context: 0 (wq_completion)bond166 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &idev->mc_lock fs_reclaim mmu_notifier_invalidate_range_start &rq->__lock irq_context: 0 (wq_completion)bond166 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &idev->mc_lock fs_reclaim mmu_notifier_invalidate_range_start &rq->__lock &cfs_rq->removed.lock irq_context: 0 (wq_completion)bond166 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &idev->mc_lock rcu_read_lock &ndev->lock irq_context: 0 (wq_completion)bond166 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &idev->mc_lock rcu_read_lock &c->lock irq_context: 0 (wq_completion)bond166 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &idev->mc_lock rcu_read_lock &____s->seqcount#2 irq_context: 0 (wq_completion)bond166 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &idev->mc_lock rcu_read_lock &____s->seqcount irq_context: 0 (wq_completion)bond166 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &idev->mc_lock rcu_read_lock &dir->lock#2 irq_context: 0 (wq_completion)bond166 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &idev->mc_lock rcu_read_lock &ul->lock#2 irq_context: 0 (wq_completion)bond166 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &idev->mc_lock rcu_read_lock rcu_read_lock &obj_hash[i].lock irq_context: 0 (wq_completion)bond166 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &idev->mc_lock rcu_read_lock rcu_read_lock &dir->lock#2 irq_context: 0 (wq_completion)bond166 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &idev->mc_lock rcu_read_lock rcu_read_lock &tbl->lock irq_context: 0 (wq_completion)bond166 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &idev->mc_lock rcu_read_lock rcu_read_lock &n->lock irq_context: 0 (wq_completion)bond166 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &idev->mc_lock rcu_read_lock rcu_read_lock &____s->seqcount#9 irq_context: 0 (wq_completion)bond166 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &idev->mc_lock rcu_read_lock rcu_read_lock rcu_node_0 irq_context: 0 (wq_completion)bond166 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &idev->mc_lock rcu_read_lock rcu_read_lock &rq->__lock irq_context: 0 (wq_completion)bond166 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &idev->mc_lock rcu_read_lock rcu_read_lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)bond166 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &idev->mc_lock rcu_read_lock rcu_read_lock rcu_read_lock_bh rcu_read_lock tk_core.seq.seqcount irq_context: 0 (wq_completion)bond166 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &idev->mc_lock rcu_read_lock rcu_read_lock rcu_read_lock_bh rcu_read_lock &obj_hash[i].lock irq_context: 0 (wq_completion)bond166 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &idev->mc_lock rcu_read_lock rcu_read_lock rcu_read_lock_bh rcu_read_lock rcu_read_lock_bh rcu_read_lock &c->lock irq_context: 0 (wq_completion)bond166 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &idev->mc_lock rcu_read_lock rcu_read_lock rcu_read_lock_bh rcu_read_lock rcu_read_lock_bh rcu_read_lock tk_core.seq.seqcount irq_context: 0 (wq_completion)bond166 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &idev->mc_lock rcu_read_lock rcu_read_lock rcu_read_lock_bh rcu_read_lock rcu_read_lock_bh rcu_read_lock &obj_hash[i].lock irq_context: 0 (wq_completion)bond166 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &idev->mc_lock rcu_read_lock rcu_read_lock rcu_read_lock_bh rcu_read_lock rcu_read_lock_bh rcu_read_lock &br->multicast_lock irq_context: 0 (wq_completion)bond166 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &idev->mc_lock rcu_read_lock rcu_read_lock rcu_read_lock_bh rcu_read_lock rcu_read_lock_bh rcu_read_lock &br->multicast_lock &obj_hash[i].lock irq_context: 0 (wq_completion)bond166 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &idev->mc_lock rcu_read_lock rcu_read_lock rcu_read_lock_bh rcu_read_lock rcu_read_lock_bh rcu_read_lock &br->multicast_lock &base->lock irq_context: 0 (wq_completion)bond166 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &idev->mc_lock rcu_read_lock rcu_read_lock rcu_read_lock_bh rcu_read_lock rcu_read_lock_bh rcu_read_lock &br->multicast_lock &base->lock &obj_hash[i].lock irq_context: 0 (wq_completion)bond166 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &idev->mc_lock rcu_read_lock rcu_read_lock rcu_read_lock_bh rcu_read_lock rcu_read_lock_bh rcu_read_lock &data->lock irq_context: 0 (wq_completion)bond166 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &idev->mc_lock rcu_read_lock rcu_node_0 irq_context: 0 (wq_completion)bond166 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex rcu_read_lock rcu_node_0 irq_context: 0 (wq_completion)bond166 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex rcu_read_lock &rq->__lock irq_context: 0 (wq_completion)bond166 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex rcu_read_lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)bond166 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex.wait_lock irq_context: 0 (wq_completion)bond166 (work_completion)(&(&bond->mcast_work)->work) &p->pi_lock irq_context: 0 (wq_completion)bond166 (work_completion)(&(&bond->mcast_work)->work) &p->pi_lock &rq->__lock irq_context: 0 (wq_completion)bond166 (work_completion)(&(&bond->mcast_work)->work) &p->pi_lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 rtnl_mutex &net->ipv6.addrconf_hash_lock rcu_read_lock &pool->lock irq_context: 0 rtnl_mutex &net->ipv6.addrconf_hash_lock rcu_read_lock &pool->lock &obj_hash[i].lock irq_context: 0 (wq_completion)events (work_completion)(&w->work)#2 nf_conntrack_mutex pool_lock#2 irq_context: 0 (wq_completion)bond123#2 (work_completion)(&(&bond->mcast_work)->work) fill_pool_map-wait-type-override &c->lock irq_context: 0 (wq_completion)bond123#2 (work_completion)(&(&bond->mcast_work)->work) fill_pool_map-wait-type-override &____s->seqcount#2 irq_context: 0 (wq_completion)bond123#2 (work_completion)(&(&bond->mcast_work)->work) fill_pool_map-wait-type-override &____s->seqcount irq_context: 0 (wq_completion)bond123#2 (work_completion)(&(&bond->mcast_work)->work) fill_pool_map-wait-type-override pool_lock irq_context: 0 (wq_completion)bond192 (work_completion)(&(&slave->notify_work)->work) &rq->__lock irq_context: 0 (wq_completion)bond168#3 (work_completion)(&(&bond->mcast_work)->work) &cfs_rq->removed.lock irq_context: 0 (wq_completion)bond194#2 (work_completion)(&(&bond->mcast_work)->work) &rq->__lock irq_context: 0 (wq_completion)bond194#2 (work_completion)(&(&bond->mcast_work)->work) &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)bond153#3 (work_completion)(&(&bond->mcast_work)->work) &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)bond179#3 irq_context: 0 (wq_completion)bond179#3 &rq->__lock irq_context: 0 (wq_completion)bond179#3 &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)bond179#3 (work_completion)(&(&slave->notify_work)->work) irq_context: 0 (wq_completion)bond179#3 (work_completion)(&(&slave->notify_work)->work) &obj_hash[i].lock irq_context: 0 (wq_completion)bond179#3 (work_completion)(&(&slave->notify_work)->work) &base->lock irq_context: 0 (wq_completion)bond179#3 (work_completion)(&(&slave->notify_work)->work) &base->lock &obj_hash[i].lock irq_context: 0 (wq_completion)bond179#3 (work_completion)(&(&slave->notify_work)->work) &rq->__lock irq_context: 0 (wq_completion)bond179#3 (work_completion)(&(&slave->notify_work)->work) &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)bond179#3 (work_completion)(&(&bond->mcast_work)->work) irq_context: 0 (wq_completion)bond179#3 (work_completion)(&(&bond->mcast_work)->work) &obj_hash[i].lock irq_context: 0 (wq_completion)bond179#3 (work_completion)(&(&bond->mcast_work)->work) &base->lock irq_context: 0 (wq_completion)bond179#3 (work_completion)(&(&bond->mcast_work)->work) &base->lock &obj_hash[i].lock irq_context: 0 (wq_completion)bond179#3 (work_completion)(&(&bond->mcast_work)->work) &rq->__lock irq_context: 0 (wq_completion)bond179#3 (work_completion)(&(&bond->mcast_work)->work) &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)bond167#3 (work_completion)(&(&bond->mcast_work)->work) &rq->__lock irq_context: 0 (wq_completion)bond167#3 (work_completion)(&(&bond->mcast_work)->work) &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)bond139#3 (work_completion)(&(&slave->notify_work)->work) rtnl_mutex net_rwsem irq_context: 0 (wq_completion)bond139#3 (work_completion)(&(&slave->notify_work)->work) rtnl_mutex net_rwsem &list->lock#2 irq_context: 0 (wq_completion)bond139#3 (work_completion)(&(&slave->notify_work)->work) rtnl_mutex &tn->lock irq_context: 0 (wq_completion)bond139#3 (work_completion)(&(&slave->notify_work)->work) rtnl_mutex.wait_lock irq_context: 0 (wq_completion)bond139#3 (work_completion)(&(&slave->notify_work)->work) &p->pi_lock irq_context: 0 (wq_completion)bond171 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex fs_reclaim irq_context: 0 (wq_completion)bond171 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex fs_reclaim mmu_notifier_invalidate_range_start irq_context: 0 (wq_completion)bond171 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &c->lock irq_context: 0 (wq_completion)bond171 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex rcu_read_lock &bond->stats_lock/1 irq_context: 0 (wq_completion)bond171 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex rcu_read_lock &ndev->lock irq_context: 0 (wq_completion)bond171 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex nl_table_lock irq_context: 0 (wq_completion)bond171 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &obj_hash[i].lock irq_context: 0 (wq_completion)bond171 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex nl_table_wait.lock irq_context: 0 (wq_completion)bond171 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex net_rwsem irq_context: 0 (wq_completion)bond171 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex net_rwsem &list->lock#2 irq_context: 0 (wq_completion)bond171 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &tn->lock irq_context: 0 (wq_completion)bond171 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &idev->mc_lock irq_context: 0 (wq_completion)bond171 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &idev->mc_lock fs_reclaim irq_context: 0 (wq_completion)bond171 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &idev->mc_lock fs_reclaim mmu_notifier_invalidate_range_start irq_context: 0 (wq_completion)bond171 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &idev->mc_lock &c->lock irq_context: 0 (wq_completion)bond171 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &idev->mc_lock rcu_read_lock &ndev->lock irq_context: 0 (wq_completion)bond171 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &idev->mc_lock rcu_read_lock &c->lock irq_context: 0 (wq_completion)bond171 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &idev->mc_lock rcu_read_lock &dir->lock#2 irq_context: 0 (wq_completion)bond171 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &idev->mc_lock rcu_read_lock &ul->lock#2 irq_context: 0 (wq_completion)bond171 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &idev->mc_lock rcu_read_lock rcu_read_lock &obj_hash[i].lock irq_context: 0 (wq_completion)bond171 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &idev->mc_lock rcu_read_lock rcu_read_lock &c->lock irq_context: 0 (wq_completion)bond171 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &idev->mc_lock rcu_read_lock rcu_read_lock &dir->lock#2 irq_context: 0 (wq_completion)bond171 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &idev->mc_lock rcu_read_lock rcu_read_lock &tbl->lock irq_context: 0 (wq_completion)bond171 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &idev->mc_lock rcu_read_lock rcu_read_lock &n->lock irq_context: 0 (wq_completion)bond171 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &idev->mc_lock rcu_read_lock rcu_read_lock &____s->seqcount#9 irq_context: 0 (wq_completion)bond171 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &idev->mc_lock rcu_read_lock rcu_read_lock rcu_read_lock_bh rcu_read_lock tk_core.seq.seqcount irq_context: 0 (wq_completion)bond171 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &idev->mc_lock rcu_read_lock rcu_read_lock rcu_read_lock_bh rcu_read_lock &obj_hash[i].lock irq_context: 0 (wq_completion)bond171 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &idev->mc_lock rcu_read_lock rcu_read_lock rcu_read_lock_bh rcu_read_lock rcu_read_lock_bh rcu_read_lock &c->lock irq_context: 0 (wq_completion)bond171 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &idev->mc_lock rcu_read_lock rcu_read_lock rcu_read_lock_bh rcu_read_lock rcu_read_lock_bh rcu_read_lock &n->list_lock irq_context: 0 (wq_completion)bond171 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &idev->mc_lock rcu_read_lock rcu_read_lock rcu_read_lock_bh rcu_read_lock rcu_read_lock_bh rcu_read_lock &n->list_lock &c->lock irq_context: 0 (wq_completion)bond171 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &idev->mc_lock rcu_read_lock rcu_read_lock rcu_read_lock_bh rcu_read_lock rcu_read_lock_bh rcu_read_lock tk_core.seq.seqcount irq_context: 0 (wq_completion)bond171 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &idev->mc_lock rcu_read_lock rcu_read_lock rcu_read_lock_bh rcu_read_lock rcu_read_lock_bh rcu_read_lock &obj_hash[i].lock irq_context: 0 (wq_completion)bond171 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &idev->mc_lock rcu_read_lock rcu_read_lock rcu_read_lock_bh rcu_read_lock rcu_read_lock_bh rcu_read_lock &br->multicast_lock irq_context: 0 (wq_completion)bond171 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &idev->mc_lock rcu_read_lock rcu_read_lock rcu_read_lock_bh rcu_read_lock rcu_read_lock_bh rcu_read_lock &br->multicast_lock &obj_hash[i].lock irq_context: 0 (wq_completion)bond171 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &idev->mc_lock rcu_read_lock rcu_read_lock rcu_read_lock_bh rcu_read_lock rcu_read_lock_bh rcu_read_lock &br->multicast_lock &base->lock irq_context: 0 (wq_completion)bond171 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &idev->mc_lock rcu_read_lock rcu_read_lock rcu_read_lock_bh rcu_read_lock rcu_read_lock_bh rcu_read_lock &br->multicast_lock &base->lock &obj_hash[i].lock irq_context: 0 (wq_completion)bond171 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &idev->mc_lock rcu_read_lock rcu_read_lock rcu_read_lock_bh rcu_read_lock rcu_read_lock_bh rcu_read_lock &data->lock irq_context: 0 (wq_completion)bond171 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex.wait_lock irq_context: 0 (wq_completion)bond171 (work_completion)(&(&bond->mcast_work)->work) &p->pi_lock irq_context: 0 (wq_completion)bond207 irq_context: 0 (wq_completion)bond207 (work_completion)(&(&slave->notify_work)->work) irq_context: 0 (wq_completion)bond207 (work_completion)(&(&slave->notify_work)->work) &obj_hash[i].lock irq_context: 0 (wq_completion)bond207 (work_completion)(&(&slave->notify_work)->work) &base->lock irq_context: 0 (wq_completion)bond207 (work_completion)(&(&slave->notify_work)->work) &base->lock &obj_hash[i].lock irq_context: 0 (wq_completion)bond207 (work_completion)(&(&slave->notify_work)->work) &rq->__lock irq_context: 0 (wq_completion)events (linkwatch_work).work rtnl_mutex fill_pool_map-wait-type-override &rq->__lock irq_context: 0 (wq_completion)events (linkwatch_work).work rtnl_mutex fill_pool_map-wait-type-override &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)bond207 (work_completion)(&(&bond->mcast_work)->work) irq_context: 0 (wq_completion)bond207 (work_completion)(&(&bond->mcast_work)->work) &obj_hash[i].lock irq_context: 0 (wq_completion)bond207 (work_completion)(&(&bond->mcast_work)->work) &base->lock irq_context: 0 (wq_completion)bond207 (work_completion)(&(&bond->mcast_work)->work) &base->lock &obj_hash[i].lock irq_context: 0 (wq_completion)ipv6_addrconf (work_completion)(&(&ifa->dad_work)->work) rtnl_mutex &idev->mc_lock rcu_read_lock rcu_read_lock rcu_read_lock_bh rcu_read_lock &n->list_lock irq_context: 0 (wq_completion)ipv6_addrconf (work_completion)(&(&ifa->dad_work)->work) rtnl_mutex &idev->mc_lock rcu_read_lock rcu_read_lock rcu_read_lock_bh rcu_read_lock &n->list_lock &c->lock irq_context: 0 (wq_completion)bond180#3 irq_context: 0 (wq_completion)bond180#3 &rq->__lock irq_context: 0 (wq_completion)bond180#3 &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)bond180#3 (work_completion)(&(&slave->notify_work)->work) irq_context: 0 (wq_completion)bond180#3 (work_completion)(&(&slave->notify_work)->work) &obj_hash[i].lock irq_context: 0 (wq_completion)bond180#3 (work_completion)(&(&slave->notify_work)->work) &base->lock irq_context: 0 (wq_completion)bond180#3 (work_completion)(&(&slave->notify_work)->work) &base->lock &obj_hash[i].lock irq_context: 0 (wq_completion)gid-cache-wq (work_completion)(&ndev_work->work) devices_rwsem rcu_read_lock &rcu_state.expedited_wq irq_context: 0 (wq_completion)gid-cache-wq (work_completion)(&ndev_work->work) devices_rwsem rcu_read_lock &rcu_state.expedited_wq &p->pi_lock irq_context: 0 (wq_completion)gid-cache-wq (work_completion)(&ndev_work->work) devices_rwsem rcu_read_lock &rcu_state.expedited_wq &p->pi_lock &rq->__lock irq_context: 0 (wq_completion)gid-cache-wq (work_completion)(&ndev_work->work) devices_rwsem rcu_read_lock &rcu_state.expedited_wq &p->pi_lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)bond180#3 (work_completion)(&(&bond->mcast_work)->work) irq_context: 0 (wq_completion)bond180#3 (work_completion)(&(&bond->mcast_work)->work) &obj_hash[i].lock irq_context: 0 (wq_completion)bond180#3 (work_completion)(&(&bond->mcast_work)->work) &base->lock irq_context: 0 (wq_completion)bond180#3 (work_completion)(&(&bond->mcast_work)->work) &base->lock &obj_hash[i].lock irq_context: 0 (wq_completion)bond173#3 (work_completion)(&(&slave->notify_work)->work) rtnl_mutex net_rwsem irq_context: 0 (wq_completion)bond173#3 (work_completion)(&(&slave->notify_work)->work) rtnl_mutex net_rwsem &list->lock#2 irq_context: 0 (wq_completion)bond173#3 (work_completion)(&(&slave->notify_work)->work) rtnl_mutex &tn->lock irq_context: 0 (wq_completion)bond173#3 (work_completion)(&(&slave->notify_work)->work) rtnl_mutex.wait_lock irq_context: 0 (wq_completion)bond173#3 (work_completion)(&(&slave->notify_work)->work) &p->pi_lock irq_context: 0 (wq_completion)bond50#2 (work_completion)(&(&bond->mcast_work)->work) fill_pool_map-wait-type-override pool_lock irq_context: 0 (wq_completion)bond56#2 (work_completion)(&(&bond->mcast_work)->work) fill_pool_map-wait-type-override pool_lock irq_context: 0 (wq_completion)bond185#2 (work_completion)(&(&bond->mcast_work)->work) &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)bond163 (work_completion)(&(&bond->mcast_work)->work) rcu_node_0 irq_context: 0 (wq_completion)bond163 (work_completion)(&(&bond->mcast_work)->work) &rcu_state.expedited_wq irq_context: 0 (wq_completion)bond163 (work_completion)(&(&bond->mcast_work)->work) &rcu_state.expedited_wq &p->pi_lock irq_context: 0 (wq_completion)bond163 (work_completion)(&(&bond->mcast_work)->work) &rcu_state.expedited_wq &p->pi_lock &rq->__lock irq_context: 0 (wq_completion)bond163 (work_completion)(&(&bond->mcast_work)->work) &rcu_state.expedited_wq &p->pi_lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)bond208 irq_context: 0 (wq_completion)bond208 &rq->__lock irq_context: 0 (wq_completion)bond208 (work_completion)(&(&slave->notify_work)->work) irq_context: 0 (wq_completion)bond208 (work_completion)(&(&slave->notify_work)->work) &obj_hash[i].lock irq_context: 0 (wq_completion)bond208 (work_completion)(&(&slave->notify_work)->work) &base->lock irq_context: 0 (wq_completion)bond208 (work_completion)(&(&slave->notify_work)->work) &base->lock &obj_hash[i].lock irq_context: 0 (wq_completion)bond208 (work_completion)(&(&slave->notify_work)->work) &rq->__lock irq_context: 0 (wq_completion)bond208 (work_completion)(&(&slave->notify_work)->work) &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)bond208 (work_completion)(&(&bond->mcast_work)->work) irq_context: 0 (wq_completion)bond208 (work_completion)(&(&bond->mcast_work)->work) &rq->__lock irq_context: 0 (wq_completion)bond208 (work_completion)(&(&bond->mcast_work)->work) &obj_hash[i].lock irq_context: 0 (wq_completion)bond208 (work_completion)(&(&bond->mcast_work)->work) &base->lock irq_context: 0 (wq_completion)bond208 (work_completion)(&(&bond->mcast_work)->work) &base->lock &obj_hash[i].lock irq_context: 0 (wq_completion)bond208 (work_completion)(&(&bond->mcast_work)->work) &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)bond200 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex fs_reclaim irq_context: 0 (wq_completion)bond200 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex fs_reclaim mmu_notifier_invalidate_range_start irq_context: 0 (wq_completion)bond200 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex rcu_read_lock &bond->stats_lock/1 irq_context: 0 (wq_completion)bond200 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex rcu_read_lock &ndev->lock irq_context: 0 (wq_completion)bond200 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex nl_table_lock irq_context: 0 (wq_completion)bond200 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &obj_hash[i].lock irq_context: 0 (wq_completion)bond200 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex nl_table_wait.lock irq_context: 0 (wq_completion)bond200 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex net_rwsem irq_context: 0 (wq_completion)bond200 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex net_rwsem &list->lock#2 irq_context: 0 (wq_completion)bond200 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &tn->lock irq_context: 0 (wq_completion)bond200 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &idev->mc_lock irq_context: 0 (wq_completion)bond200 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &idev->mc_lock fs_reclaim irq_context: 0 (wq_completion)bond200 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &idev->mc_lock fs_reclaim mmu_notifier_invalidate_range_start irq_context: 0 (wq_completion)bond200 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &idev->mc_lock &c->lock irq_context: 0 (wq_completion)bond200 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &idev->mc_lock rcu_read_lock &ndev->lock irq_context: 0 (wq_completion)bond200 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &idev->mc_lock rcu_read_lock &dir->lock#2 irq_context: 0 (wq_completion)bond200 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &idev->mc_lock rcu_read_lock &ul->lock#2 irq_context: 0 (wq_completion)bond200 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &idev->mc_lock rcu_read_lock rcu_read_lock &____s->seqcount#10 irq_context: 0 (wq_completion)bond200 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &idev->mc_lock rcu_read_lock rcu_read_lock rcu_read_lock_bh rcu_read_lock rcu_read_lock_bh rcu_read_lock &br->multicast_lock irq_context: 0 (wq_completion)bond200 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &idev->mc_lock rcu_read_lock rcu_read_lock rcu_read_lock_bh rcu_read_lock rcu_read_lock_bh rcu_read_lock &br->multicast_lock &obj_hash[i].lock irq_context: 0 (wq_completion)bond200 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &idev->mc_lock rcu_read_lock rcu_read_lock rcu_read_lock_bh rcu_read_lock rcu_read_lock_bh rcu_read_lock &br->multicast_lock &base->lock irq_context: 0 (wq_completion)bond200 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &idev->mc_lock rcu_read_lock rcu_read_lock rcu_read_lock_bh rcu_read_lock rcu_read_lock_bh rcu_read_lock &data->lock irq_context: 0 (wq_completion)bond200 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &idev->mc_lock rcu_read_lock rcu_read_lock rcu_read_lock_bh rcu_read_lock rcu_read_lock_bh rcu_read_lock &obj_hash[i].lock irq_context: 0 (wq_completion)bond200 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex.wait_lock irq_context: 0 (wq_completion)bond123#2 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex fs_reclaim irq_context: 0 (wq_completion)bond200 (work_completion)(&(&bond->mcast_work)->work) &p->pi_lock irq_context: 0 (wq_completion)bond123#2 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex fs_reclaim &rq->__lock irq_context: 0 (wq_completion)bond123#2 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex fs_reclaim &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)bond123#2 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex fs_reclaim mmu_notifier_invalidate_range_start irq_context: 0 (wq_completion)bond123#2 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &c->lock irq_context: 0 (wq_completion)bond123#2 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex rcu_read_lock &bond->stats_lock/1 irq_context: 0 (wq_completion)bond123#2 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex rcu_read_lock &ndev->lock irq_context: 0 (wq_completion)bond123#2 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex nl_table_lock irq_context: 0 (wq_completion)bond123#2 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &obj_hash[i].lock irq_context: 0 (wq_completion)bond123#2 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex nl_table_wait.lock irq_context: 0 (wq_completion)bond123#2 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex net_rwsem irq_context: 0 (wq_completion)bond123#2 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex net_rwsem &list->lock#2 irq_context: 0 (wq_completion)bond123#2 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &tn->lock irq_context: 0 (wq_completion)bond123#2 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &tn->lock &rq->__lock irq_context: 0 (wq_completion)bond123#2 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &tn->lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)bond123#2 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &idev->mc_lock irq_context: 0 (wq_completion)bond123#2 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &idev->mc_lock fs_reclaim irq_context: 0 (wq_completion)bond123#2 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &idev->mc_lock fs_reclaim mmu_notifier_invalidate_range_start irq_context: 0 (wq_completion)bond123#2 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &idev->mc_lock rcu_read_lock &ndev->lock irq_context: 0 (wq_completion)bond123#2 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &idev->mc_lock rcu_read_lock &dir->lock#2 irq_context: 0 (wq_completion)bond123#2 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &idev->mc_lock rcu_read_lock &ul->lock#2 irq_context: 0 (wq_completion)bond123#2 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &idev->mc_lock rcu_read_lock rcu_read_lock &obj_hash[i].lock irq_context: 0 (wq_completion)bond123#2 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &idev->mc_lock rcu_read_lock rcu_read_lock &dir->lock#2 irq_context: 0 (wq_completion)bond123#2 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &idev->mc_lock rcu_read_lock rcu_read_lock &tbl->lock irq_context: 0 (wq_completion)bond123#2 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &idev->mc_lock rcu_read_lock rcu_read_lock rcu_node_0 irq_context: 0 (wq_completion)bond123#2 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &idev->mc_lock rcu_read_lock rcu_read_lock &rq->__lock irq_context: 0 (wq_completion)bond123#2 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &idev->mc_lock rcu_read_lock rcu_read_lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)bond123#2 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &idev->mc_lock rcu_read_lock rcu_read_lock &n->lock irq_context: 0 (wq_completion)bond123#2 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &idev->mc_lock rcu_read_lock rcu_read_lock &____s->seqcount#9 irq_context: 0 (wq_completion)bond123#2 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &idev->mc_lock rcu_read_lock rcu_read_lock rcu_read_lock_bh rcu_read_lock rcu_read_lock_bh rcu_read_lock &br->multicast_lock irq_context: 0 (wq_completion)bond123#2 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &idev->mc_lock rcu_read_lock rcu_read_lock rcu_read_lock_bh rcu_read_lock rcu_read_lock_bh rcu_read_lock &br->multicast_lock &obj_hash[i].lock irq_context: 0 (wq_completion)bond123#2 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &idev->mc_lock rcu_read_lock rcu_read_lock rcu_read_lock_bh rcu_read_lock rcu_read_lock_bh rcu_read_lock &br->multicast_lock &base->lock irq_context: 0 (wq_completion)bond123#2 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &idev->mc_lock rcu_read_lock rcu_read_lock rcu_read_lock_bh rcu_read_lock rcu_read_lock_bh rcu_read_lock &br->multicast_lock &base->lock &obj_hash[i].lock irq_context: 0 (wq_completion)bond123#2 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &idev->mc_lock rcu_read_lock rcu_read_lock rcu_read_lock_bh rcu_read_lock rcu_read_lock_bh rcu_read_lock &data->lock irq_context: 0 (wq_completion)bond123#2 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &idev->mc_lock rcu_read_lock rcu_read_lock rcu_read_lock_bh rcu_read_lock rcu_read_lock_bh rcu_read_lock &obj_hash[i].lock irq_context: 0 (wq_completion)bond123#2 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &idev->mc_lock rcu_read_lock rcu_node_0 irq_context: 0 (wq_completion)bond123#2 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex.wait_lock irq_context: 0 (wq_completion)bond123#2 (work_completion)(&(&bond->mcast_work)->work) &p->pi_lock irq_context: 0 (wq_completion)bond197#2 irq_context: 0 (wq_completion)bond197#2 &rq->__lock irq_context: 0 (wq_completion)bond197#2 &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)bond197#2 (work_completion)(&(&slave->notify_work)->work) irq_context: 0 (wq_completion)bond197#2 (work_completion)(&(&slave->notify_work)->work) &obj_hash[i].lock irq_context: 0 (wq_completion)bond197#2 (work_completion)(&(&slave->notify_work)->work) &base->lock irq_context: 0 (wq_completion)bond197#2 (work_completion)(&(&slave->notify_work)->work) &base->lock &obj_hash[i].lock irq_context: 0 (wq_completion)bond197#2 (work_completion)(&(&slave->notify_work)->work) &rq->__lock irq_context: 0 (wq_completion)bond197#2 (work_completion)(&(&slave->notify_work)->work) &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)bond197#2 (work_completion)(&(&bond->mcast_work)->work) irq_context: 0 (wq_completion)bond197#2 (work_completion)(&(&bond->mcast_work)->work) &obj_hash[i].lock irq_context: 0 (wq_completion)bond197#2 (work_completion)(&(&bond->mcast_work)->work) &base->lock irq_context: 0 (wq_completion)bond197#2 (work_completion)(&(&bond->mcast_work)->work) &base->lock &obj_hash[i].lock irq_context: 0 (wq_completion)bond197#2 (work_completion)(&(&bond->mcast_work)->work) &rq->__lock irq_context: 0 (wq_completion)bond201 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex fs_reclaim irq_context: 0 (wq_completion)bond201 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex fs_reclaim mmu_notifier_invalidate_range_start irq_context: 0 (wq_completion)bond201 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &c->lock irq_context: 0 (wq_completion)bond201 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex rcu_read_lock &bond->stats_lock/1 irq_context: 0 (wq_completion)bond201 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex rcu_read_lock &ndev->lock irq_context: 0 (wq_completion)bond201 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex nl_table_lock irq_context: 0 (wq_completion)bond201 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &obj_hash[i].lock irq_context: 0 (wq_completion)bond201 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex nl_table_wait.lock irq_context: 0 (wq_completion)bond201 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex net_rwsem irq_context: 0 (wq_completion)bond201 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex net_rwsem &list->lock#2 irq_context: 0 (wq_completion)bond201 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &tn->lock irq_context: 0 (wq_completion)bond201 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &idev->mc_lock irq_context: 0 (wq_completion)bond201 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &idev->mc_lock fs_reclaim irq_context: 0 (wq_completion)bond201 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &idev->mc_lock fs_reclaim mmu_notifier_invalidate_range_start irq_context: 0 (wq_completion)bond201 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &idev->mc_lock &c->lock irq_context: 0 (wq_completion)bond201 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &idev->mc_lock rcu_read_lock &ndev->lock irq_context: 0 (wq_completion)bond201 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &idev->mc_lock rcu_read_lock &dir->lock#2 irq_context: 0 (wq_completion)bond201 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &idev->mc_lock rcu_read_lock &ul->lock#2 irq_context: 0 (wq_completion)bond201 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &idev->mc_lock rcu_read_lock rcu_read_lock &____s->seqcount#10 irq_context: 0 (wq_completion)bond201 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &idev->mc_lock rcu_read_lock rcu_read_lock rcu_read_lock_bh rcu_read_lock rcu_read_lock_bh rcu_read_lock &br->multicast_lock irq_context: 0 (wq_completion)bond201 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &idev->mc_lock rcu_read_lock rcu_read_lock rcu_read_lock_bh rcu_read_lock rcu_read_lock_bh rcu_read_lock &br->multicast_lock &obj_hash[i].lock irq_context: 0 (wq_completion)bond201 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &idev->mc_lock rcu_read_lock rcu_read_lock rcu_read_lock_bh rcu_read_lock rcu_read_lock_bh rcu_read_lock &br->multicast_lock &base->lock irq_context: 0 (wq_completion)bond201 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &idev->mc_lock rcu_read_lock rcu_read_lock rcu_read_lock_bh rcu_read_lock rcu_read_lock_bh rcu_read_lock &data->lock irq_context: 0 (wq_completion)bond201 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &idev->mc_lock rcu_read_lock rcu_read_lock rcu_read_lock_bh rcu_read_lock rcu_read_lock_bh rcu_read_lock &obj_hash[i].lock irq_context: 0 (wq_completion)bond201 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex.wait_lock irq_context: 0 (wq_completion)bond204 (work_completion)(&(&bond->mcast_work)->work) &rq->__lock irq_context: 0 (wq_completion)bond203 (work_completion)(&(&bond->mcast_work)->work) &rq->__lock irq_context: 0 (wq_completion)bond203 (work_completion)(&(&bond->mcast_work)->work) &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)bond201 (work_completion)(&(&bond->mcast_work)->work) &p->pi_lock irq_context: 0 (wq_completion)bond204 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex fs_reclaim irq_context: 0 (wq_completion)bond204 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex fs_reclaim mmu_notifier_invalidate_range_start irq_context: 0 (wq_completion)bond204 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &c->lock irq_context: 0 (wq_completion)bond204 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &rq->__lock irq_context: 0 (wq_completion)bond204 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex remove_cache_srcu irq_context: 0 (wq_completion)bond204 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex remove_cache_srcu &rq->__lock irq_context: 0 (wq_completion)bond204 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex remove_cache_srcu quarantine_lock irq_context: 0 (wq_completion)bond204 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &n->list_lock irq_context: 0 (wq_completion)bond204 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &n->list_lock &c->lock irq_context: 0 (wq_completion)bond204 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex rcu_read_lock &bond->stats_lock/1 irq_context: 0 (wq_completion)bond204 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex rcu_read_lock rcu_node_0 irq_context: 0 (wq_completion)bond204 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex rcu_read_lock &rq->__lock irq_context: 0 (wq_completion)bond204 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex rcu_read_lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)bond204 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex rcu_read_lock &ndev->lock irq_context: 0 (wq_completion)bond204 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex nl_table_lock irq_context: 0 (wq_completion)bond204 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &obj_hash[i].lock irq_context: 0 (wq_completion)bond204 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex nl_table_wait.lock irq_context: 0 (wq_completion)bond204 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex net_rwsem irq_context: 0 (wq_completion)bond204 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex net_rwsem &list->lock#2 irq_context: 0 (wq_completion)bond204 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &tn->lock irq_context: 0 (wq_completion)bond204 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &idev->mc_lock irq_context: 0 (wq_completion)bond204 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &idev->mc_lock fs_reclaim irq_context: 0 (wq_completion)bond204 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &idev->mc_lock fs_reclaim mmu_notifier_invalidate_range_start irq_context: 0 (wq_completion)bond204 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &idev->mc_lock &c->lock irq_context: 0 (wq_completion)bond204 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &idev->mc_lock rcu_read_lock &ndev->lock irq_context: 0 (wq_completion)bond204 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &idev->mc_lock rcu_read_lock &dir->lock#2 irq_context: 0 (wq_completion)bond204 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &idev->mc_lock rcu_read_lock &ul->lock#2 irq_context: 0 (wq_completion)bond204 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &idev->mc_lock rcu_read_lock rcu_read_lock &____s->seqcount#10 irq_context: 0 (wq_completion)bond204 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &idev->mc_lock rcu_read_lock rcu_read_lock rcu_read_lock_bh rcu_read_lock rcu_read_lock_bh rcu_read_lock &br->multicast_lock irq_context: 0 (wq_completion)bond204 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &idev->mc_lock rcu_read_lock rcu_read_lock rcu_read_lock_bh rcu_read_lock rcu_read_lock_bh rcu_read_lock &br->multicast_lock &obj_hash[i].lock irq_context: 0 (wq_completion)bond204 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &idev->mc_lock rcu_read_lock rcu_read_lock rcu_read_lock_bh rcu_read_lock rcu_read_lock_bh rcu_read_lock &br->multicast_lock &base->lock irq_context: 0 (wq_completion)bond204 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &idev->mc_lock rcu_read_lock rcu_read_lock rcu_read_lock_bh rcu_read_lock rcu_read_lock_bh rcu_read_lock &data->lock irq_context: 0 (wq_completion)bond204 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &idev->mc_lock rcu_read_lock rcu_read_lock rcu_read_lock_bh rcu_read_lock rcu_read_lock_bh rcu_read_lock &obj_hash[i].lock irq_context: 0 (wq_completion)bond204 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex.wait_lock irq_context: 0 (wq_completion)bond204 (work_completion)(&(&bond->mcast_work)->work) &p->pi_lock irq_context: 0 (wq_completion)bond204 (work_completion)(&(&bond->mcast_work)->work) &p->pi_lock &rq->__lock irq_context: 0 (wq_completion)bond204 (work_completion)(&(&bond->mcast_work)->work) &p->pi_lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 rtnl_mutex fill_pool_map-wait-type-override &pcp->lock &zone->lock irq_context: 0 rtnl_mutex fill_pool_map-wait-type-override &pcp->lock &zone->lock &____s->seqcount irq_context: 0 (wq_completion)bond209 irq_context: 0 (wq_completion)bond209 &rq->__lock irq_context: 0 (wq_completion)bond209 &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)bond209 (work_completion)(&(&slave->notify_work)->work) irq_context: 0 (wq_completion)bond209 (work_completion)(&(&slave->notify_work)->work) &obj_hash[i].lock irq_context: 0 (wq_completion)bond209 (work_completion)(&(&slave->notify_work)->work) &base->lock irq_context: 0 (wq_completion)bond209 (work_completion)(&(&slave->notify_work)->work) &base->lock &obj_hash[i].lock irq_context: 0 (wq_completion)bond209 (work_completion)(&(&slave->notify_work)->work) &rq->__lock irq_context: 0 (wq_completion)bond209 (work_completion)(&(&slave->notify_work)->work) &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)bond208 &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)bond209 (work_completion)(&(&bond->mcast_work)->work) irq_context: 0 (wq_completion)bond209 (work_completion)(&(&bond->mcast_work)->work) &obj_hash[i].lock irq_context: 0 (wq_completion)bond209 (work_completion)(&(&bond->mcast_work)->work) &base->lock irq_context: 0 (wq_completion)bond209 (work_completion)(&(&bond->mcast_work)->work) &base->lock &obj_hash[i].lock irq_context: 0 (wq_completion)bond170#3 (work_completion)(&(&bond->mcast_work)->work) &rq->__lock irq_context: 0 (wq_completion)bond170#3 (work_completion)(&(&bond->mcast_work)->work) &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)bond198#2 irq_context: 0 (wq_completion)bond198#2 (work_completion)(&(&slave->notify_work)->work) irq_context: 0 (wq_completion)bond198#2 (work_completion)(&(&slave->notify_work)->work) &obj_hash[i].lock irq_context: 0 (wq_completion)bond198#2 (work_completion)(&(&slave->notify_work)->work) &base->lock irq_context: 0 (wq_completion)bond198#2 (work_completion)(&(&slave->notify_work)->work) &base->lock &obj_hash[i].lock irq_context: 0 (wq_completion)bond198#2 (work_completion)(&(&bond->mcast_work)->work) irq_context: 0 (wq_completion)bond198#2 (work_completion)(&(&bond->mcast_work)->work) &obj_hash[i].lock irq_context: 0 (wq_completion)bond198#2 (work_completion)(&(&bond->mcast_work)->work) &base->lock irq_context: 0 (wq_completion)bond198#2 (work_completion)(&(&bond->mcast_work)->work) &base->lock &obj_hash[i].lock irq_context: 0 (wq_completion)bond198#2 (work_completion)(&(&slave->notify_work)->work) rtnl_mutex net_rwsem irq_context: 0 (wq_completion)bond198#2 (work_completion)(&(&slave->notify_work)->work) rtnl_mutex net_rwsem &list->lock#2 irq_context: 0 (wq_completion)bond198#2 (work_completion)(&(&slave->notify_work)->work) rtnl_mutex &tn->lock irq_context: 0 (wq_completion)bond198#2 (work_completion)(&(&slave->notify_work)->work) rtnl_mutex.wait_lock irq_context: 0 (wq_completion)bond198#2 (work_completion)(&(&slave->notify_work)->work) &p->pi_lock irq_context: 0 (wq_completion)bond179#3 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex fs_reclaim irq_context: 0 (wq_completion)bond179#3 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex fs_reclaim mmu_notifier_invalidate_range_start irq_context: 0 (wq_completion)bond179#3 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &c->lock irq_context: 0 (wq_completion)bond179#3 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &n->list_lock irq_context: 0 (wq_completion)bond179#3 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &n->list_lock &c->lock irq_context: 0 (wq_completion)bond179#3 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex rcu_read_lock &bond->stats_lock/1 irq_context: 0 (wq_completion)bond179#3 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex rcu_read_lock &ndev->lock irq_context: 0 (wq_completion)bond179#3 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex nl_table_lock irq_context: 0 (wq_completion)bond179#3 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &obj_hash[i].lock irq_context: 0 (wq_completion)bond179#3 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex nl_table_wait.lock irq_context: 0 (wq_completion)bond179#3 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex net_rwsem irq_context: 0 (wq_completion)bond179#3 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex net_rwsem &list->lock#2 irq_context: 0 (wq_completion)bond179#3 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &tn->lock irq_context: 0 (wq_completion)bond179#3 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &idev->mc_lock irq_context: 0 (wq_completion)bond179#3 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &idev->mc_lock fs_reclaim irq_context: 0 (wq_completion)bond179#3 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &idev->mc_lock fs_reclaim mmu_notifier_invalidate_range_start irq_context: 0 (wq_completion)bond179#3 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &idev->mc_lock &c->lock irq_context: 0 (wq_completion)bond179#3 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &idev->mc_lock &n->list_lock irq_context: 0 (wq_completion)bond179#3 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &idev->mc_lock rcu_read_lock &ndev->lock irq_context: 0 (wq_completion)bond179#3 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &idev->mc_lock rcu_read_lock &dir->lock#2 irq_context: 0 (wq_completion)bond179#3 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &idev->mc_lock rcu_read_lock &ul->lock#2 irq_context: 0 (wq_completion)bond179#3 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &idev->mc_lock rcu_read_lock rcu_read_lock &obj_hash[i].lock irq_context: 0 (wq_completion)bond179#3 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &idev->mc_lock rcu_read_lock rcu_read_lock &dir->lock#2 irq_context: 0 (wq_completion)bond179#3 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &idev->mc_lock rcu_read_lock rcu_read_lock &tbl->lock irq_context: 0 (wq_completion)bond179#3 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &idev->mc_lock rcu_read_lock rcu_read_lock &n->lock irq_context: 0 (wq_completion)bond179#3 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &idev->mc_lock rcu_read_lock rcu_read_lock &____s->seqcount#9 irq_context: 0 (wq_completion)bond179#3 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &idev->mc_lock rcu_read_lock rcu_read_lock rcu_read_lock_bh rcu_read_lock rcu_read_lock_bh rcu_read_lock &br->multicast_lock irq_context: 0 (wq_completion)bond179#3 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &idev->mc_lock rcu_read_lock rcu_read_lock rcu_read_lock_bh rcu_read_lock rcu_read_lock_bh rcu_read_lock &br->multicast_lock &obj_hash[i].lock irq_context: 0 (wq_completion)bond179#3 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &idev->mc_lock rcu_read_lock rcu_read_lock rcu_read_lock_bh rcu_read_lock rcu_read_lock_bh rcu_read_lock &br->multicast_lock rcu_read_lock rhashtable_bucket irq_context: 0 (wq_completion)bond179#3 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &idev->mc_lock rcu_read_lock rcu_read_lock rcu_read_lock_bh rcu_read_lock rcu_read_lock_bh rcu_read_lock &br->multicast_lock nl_table_lock irq_context: 0 (wq_completion)bond179#3 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &idev->mc_lock rcu_read_lock rcu_read_lock rcu_read_lock_bh rcu_read_lock rcu_read_lock_bh rcu_read_lock &br->multicast_lock nl_table_wait.lock irq_context: 0 (wq_completion)bond179#3 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &idev->mc_lock rcu_read_lock rcu_read_lock rcu_read_lock_bh rcu_read_lock rcu_read_lock_bh rcu_read_lock &br->multicast_lock &base->lock irq_context: 0 (wq_completion)bond179#3 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &idev->mc_lock rcu_read_lock rcu_read_lock rcu_read_lock_bh rcu_read_lock rcu_read_lock_bh rcu_read_lock &br->multicast_lock &base->lock &obj_hash[i].lock irq_context: 0 (wq_completion)bond179#3 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &idev->mc_lock rcu_read_lock rcu_read_lock rcu_read_lock_bh rcu_read_lock rcu_read_lock_bh rcu_read_lock &data->lock irq_context: 0 (wq_completion)bond179#3 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &idev->mc_lock rcu_read_lock rcu_read_lock rcu_read_lock_bh rcu_read_lock rcu_read_lock_bh rcu_read_lock &obj_hash[i].lock irq_context: 0 (wq_completion)bond179#3 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex rcu_read_lock rcu_node_0 irq_context: 0 (wq_completion)bond179#3 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex rcu_read_lock &rq->__lock irq_context: 0 (wq_completion)bond179#3 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex.wait_lock irq_context: 0 (wq_completion)bond179#3 (work_completion)(&(&bond->mcast_work)->work) &p->pi_lock irq_context: 0 (wq_completion)bond179#3 (work_completion)(&(&slave->notify_work)->work) rtnl_mutex net_rwsem irq_context: 0 (wq_completion)bond179#3 (work_completion)(&(&slave->notify_work)->work) rtnl_mutex net_rwsem &list->lock#2 irq_context: 0 (wq_completion)bond179#3 (work_completion)(&(&slave->notify_work)->work) rtnl_mutex &tn->lock irq_context: 0 (wq_completion)bond179#3 (work_completion)(&(&slave->notify_work)->work) rtnl_mutex.wait_lock irq_context: 0 (wq_completion)bond145 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex fs_reclaim irq_context: 0 (wq_completion)bond179#3 (work_completion)(&(&slave->notify_work)->work) &p->pi_lock irq_context: 0 (wq_completion)bond145 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex fs_reclaim mmu_notifier_invalidate_range_start irq_context: 0 (wq_completion)bond145 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex rcu_read_lock &bond->stats_lock/1 irq_context: 0 (wq_completion)bond145 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex rcu_read_lock &ndev->lock irq_context: 0 (wq_completion)bond145 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex nl_table_lock irq_context: 0 (wq_completion)bond145 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &obj_hash[i].lock irq_context: 0 (wq_completion)bond145 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &rq->__lock irq_context: 0 (wq_completion)bond145 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)bond145 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex nl_table_wait.lock irq_context: 0 (wq_completion)bond145 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex net_rwsem irq_context: 0 (wq_completion)bond145 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex net_rwsem &list->lock#2 irq_context: 0 (wq_completion)bond145 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &tn->lock irq_context: 0 (wq_completion)bond145 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &idev->mc_lock irq_context: 0 (wq_completion)bond145 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &idev->mc_lock fs_reclaim irq_context: 0 (wq_completion)bond145 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &idev->mc_lock fs_reclaim mmu_notifier_invalidate_range_start irq_context: 0 (wq_completion)bond145 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &idev->mc_lock &c->lock irq_context: 0 (wq_completion)bond145 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &idev->mc_lock rcu_read_lock &ndev->lock irq_context: 0 (wq_completion)bond145 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &idev->mc_lock rcu_read_lock &c->lock irq_context: 0 (wq_completion)bond145 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &idev->mc_lock rcu_read_lock rcu_node_0 irq_context: 0 (wq_completion)bond145 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &idev->mc_lock rcu_read_lock &rq->__lock irq_context: 0 (wq_completion)bond145 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &idev->mc_lock rcu_read_lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)bond145 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &idev->mc_lock rcu_read_lock &dir->lock#2 irq_context: 0 (wq_completion)bond145 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &idev->mc_lock rcu_read_lock &n->list_lock irq_context: 0 (wq_completion)bond145 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &idev->mc_lock rcu_read_lock &n->list_lock &c->lock irq_context: 0 (wq_completion)bond145 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &idev->mc_lock rcu_read_lock &ul->lock#2 irq_context: 0 (wq_completion)bond145 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &idev->mc_lock rcu_read_lock rcu_read_lock &obj_hash[i].lock irq_context: 0 (wq_completion)bond145 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &idev->mc_lock rcu_read_lock rcu_read_lock &dir->lock#2 irq_context: 0 (wq_completion)bond145 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &idev->mc_lock rcu_read_lock rcu_read_lock &tbl->lock irq_context: 0 (wq_completion)bond145 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &idev->mc_lock rcu_read_lock rcu_read_lock &n->lock irq_context: 0 (wq_completion)bond145 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &idev->mc_lock rcu_read_lock rcu_read_lock &____s->seqcount#9 irq_context: 0 (wq_completion)bond145 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &idev->mc_lock rcu_read_lock rcu_read_lock rcu_read_lock_bh rcu_read_lock rcu_read_lock_bh rcu_read_lock &br->multicast_lock irq_context: 0 (wq_completion)bond145 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &idev->mc_lock rcu_read_lock rcu_read_lock rcu_read_lock_bh rcu_read_lock rcu_read_lock_bh rcu_read_lock &br->multicast_lock &obj_hash[i].lock irq_context: 0 (wq_completion)bond145 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &idev->mc_lock rcu_read_lock rcu_read_lock rcu_read_lock_bh rcu_read_lock rcu_read_lock_bh rcu_read_lock &br->multicast_lock &base->lock irq_context: 0 (wq_completion)bond145 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &idev->mc_lock rcu_read_lock rcu_read_lock rcu_read_lock_bh rcu_read_lock rcu_read_lock_bh rcu_read_lock &br->multicast_lock &base->lock &obj_hash[i].lock irq_context: 0 (wq_completion)bond145 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &idev->mc_lock rcu_read_lock rcu_read_lock rcu_read_lock_bh rcu_read_lock rcu_read_lock_bh rcu_read_lock &data->lock irq_context: 0 (wq_completion)bond145 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &idev->mc_lock rcu_read_lock rcu_read_lock rcu_read_lock_bh rcu_read_lock rcu_read_lock_bh rcu_read_lock &obj_hash[i].lock irq_context: 0 (wq_completion)bond145 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &idev->mc_lock rcu_read_lock rcu_read_lock &rq->__lock irq_context: 0 (wq_completion)bond145 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &idev->mc_lock rcu_read_lock rcu_read_lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)bond145 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex.wait_lock irq_context: 0 (wq_completion)bond145 (work_completion)(&(&bond->mcast_work)->work) &p->pi_lock irq_context: 0 (wq_completion)bond182#3 irq_context: 0 (wq_completion)bond182#3 (work_completion)(&(&slave->notify_work)->work) irq_context: 0 (wq_completion)bond182#3 (work_completion)(&(&slave->notify_work)->work) &obj_hash[i].lock irq_context: 0 (wq_completion)bond182#3 (work_completion)(&(&slave->notify_work)->work) &base->lock irq_context: 0 (wq_completion)bond182#3 (work_completion)(&(&slave->notify_work)->work) &base->lock &obj_hash[i].lock irq_context: 0 (wq_completion)bond182#3 (work_completion)(&(&slave->notify_work)->work) &rq->__lock irq_context: 0 (wq_completion)bond182#3 (work_completion)(&(&slave->notify_work)->work) &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)bond182#3 (work_completion)(&(&bond->mcast_work)->work) irq_context: 0 (wq_completion)bond182#3 (work_completion)(&(&bond->mcast_work)->work) &obj_hash[i].lock irq_context: 0 (wq_completion)bond182#3 (work_completion)(&(&bond->mcast_work)->work) &base->lock irq_context: 0 (wq_completion)bond182#3 (work_completion)(&(&bond->mcast_work)->work) &base->lock &obj_hash[i].lock irq_context: 0 (wq_completion)bond207 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex fs_reclaim irq_context: 0 (wq_completion)bond207 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex fs_reclaim mmu_notifier_invalidate_range_start irq_context: 0 (wq_completion)bond207 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &c->lock irq_context: 0 (wq_completion)bond207 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex rcu_read_lock &bond->stats_lock/1 irq_context: 0 (wq_completion)bond207 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex rcu_read_lock &ndev->lock irq_context: 0 (wq_completion)bond207 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex nl_table_lock irq_context: 0 (wq_completion)bond207 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &obj_hash[i].lock irq_context: 0 (wq_completion)bond207 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex nl_table_wait.lock irq_context: 0 (wq_completion)bond207 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex net_rwsem irq_context: 0 (wq_completion)bond207 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex net_rwsem &list->lock#2 irq_context: 0 (wq_completion)bond207 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &tn->lock irq_context: 0 (wq_completion)bond207 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &idev->mc_lock irq_context: 0 (wq_completion)bond207 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &idev->mc_lock fs_reclaim irq_context: 0 (wq_completion)bond207 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &idev->mc_lock fs_reclaim mmu_notifier_invalidate_range_start irq_context: 0 (wq_completion)bond207 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &idev->mc_lock &c->lock irq_context: 0 (wq_completion)bond207 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &idev->mc_lock rcu_read_lock &ndev->lock irq_context: 0 (wq_completion)bond207 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &idev->mc_lock rcu_read_lock &dir->lock#2 irq_context: 0 (wq_completion)bond207 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &idev->mc_lock rcu_read_lock &ul->lock#2 irq_context: 0 (wq_completion)bond207 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &idev->mc_lock rcu_read_lock rcu_read_lock &____s->seqcount#10 irq_context: 0 (wq_completion)bond207 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &idev->mc_lock rcu_read_lock rcu_read_lock rcu_read_lock_bh rcu_read_lock rcu_read_lock_bh rcu_read_lock &br->multicast_lock irq_context: 0 (wq_completion)bond207 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &idev->mc_lock rcu_read_lock rcu_read_lock rcu_read_lock_bh rcu_read_lock rcu_read_lock_bh rcu_read_lock &br->multicast_lock &obj_hash[i].lock irq_context: 0 (wq_completion)bond207 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &idev->mc_lock rcu_read_lock rcu_read_lock rcu_read_lock_bh rcu_read_lock rcu_read_lock_bh rcu_read_lock &br->multicast_lock &base->lock irq_context: 0 (wq_completion)bond207 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &idev->mc_lock rcu_read_lock rcu_read_lock rcu_read_lock_bh rcu_read_lock rcu_read_lock_bh rcu_read_lock &data->lock irq_context: 0 (wq_completion)bond207 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &idev->mc_lock rcu_read_lock rcu_read_lock rcu_read_lock_bh rcu_read_lock rcu_read_lock_bh rcu_read_lock &obj_hash[i].lock irq_context: 0 (wq_completion)bond207 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &idev->mc_lock rcu_read_lock rcu_read_lock rcu_read_lock_bh rcu_read_lock rcu_read_lock_bh rcu_read_lock quarantine_lock irq_context: 0 (wq_completion)bond207 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex.wait_lock irq_context: 0 (wq_completion)bond207 (work_completion)(&(&bond->mcast_work)->work) &p->pi_lock irq_context: 0 (wq_completion)ipv6_addrconf (work_completion)(&(&ifa->dad_work)->work) rtnl_mutex rcu_read_lock rcu_read_lock &tbl->lock quarantine_lock irq_context: 0 (wq_completion)mld (work_completion)(&(&idev->mc_ifc_work)->work) &idev->mc_lock rcu_read_lock rcu_read_lock rcu_read_lock_bh rcu_read_lock &br->multicast_lock &base->lock fill_pool_map-wait-type-override pool_lock#2 irq_context: 0 (wq_completion)mld (work_completion)(&(&idev->mc_ifc_work)->work) &idev->mc_lock rcu_read_lock rcu_read_lock rcu_read_lock_bh rcu_read_lock &br->multicast_lock &base->lock fill_pool_map-wait-type-override pool_lock irq_context: 0 (wq_completion)bond210 irq_context: 0 (wq_completion)bond210 (work_completion)(&(&slave->notify_work)->work) irq_context: 0 (wq_completion)bond210 (work_completion)(&(&slave->notify_work)->work) &rq->__lock irq_context: 0 (wq_completion)bond210 (work_completion)(&(&slave->notify_work)->work) &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)bond210 (work_completion)(&(&slave->notify_work)->work) &obj_hash[i].lock irq_context: 0 (wq_completion)bond210 (work_completion)(&(&slave->notify_work)->work) &base->lock irq_context: 0 (wq_completion)bond210 (work_completion)(&(&slave->notify_work)->work) &base->lock &obj_hash[i].lock irq_context: 0 (wq_completion)bond166#3 (work_completion)(&(&slave->notify_work)->work) &rq->__lock irq_context: 0 (wq_completion)bond166#3 (work_completion)(&(&slave->notify_work)->work) &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)bond210 (work_completion)(&(&bond->mcast_work)->work) irq_context: 0 (wq_completion)bond210 (work_completion)(&(&bond->mcast_work)->work) &rq->__lock irq_context: 0 (wq_completion)bond210 (work_completion)(&(&bond->mcast_work)->work) &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)bond210 (work_completion)(&(&bond->mcast_work)->work) &obj_hash[i].lock irq_context: 0 (wq_completion)bond210 (work_completion)(&(&bond->mcast_work)->work) &base->lock irq_context: 0 (wq_completion)bond210 (work_completion)(&(&bond->mcast_work)->work) &base->lock &obj_hash[i].lock irq_context: 0 (wq_completion)bond199#2 irq_context: 0 (wq_completion)bond199#2 (work_completion)(&(&slave->notify_work)->work) irq_context: 0 (wq_completion)bond199#2 (work_completion)(&(&slave->notify_work)->work) &obj_hash[i].lock irq_context: 0 (wq_completion)bond199#2 (work_completion)(&(&slave->notify_work)->work) &base->lock irq_context: 0 (wq_completion)bond199#2 (work_completion)(&(&slave->notify_work)->work) &base->lock &obj_hash[i].lock irq_context: 0 (wq_completion)bond148#3 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex fs_reclaim irq_context: 0 (wq_completion)bond148#3 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex fs_reclaim &rq->__lock irq_context: 0 (wq_completion)bond148#3 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex fs_reclaim &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)bond148#3 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex fs_reclaim mmu_notifier_invalidate_range_start irq_context: 0 (wq_completion)bond148#3 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &c->lock irq_context: 0 (wq_completion)bond148#3 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex rcu_read_lock &bond->stats_lock/1 irq_context: 0 (wq_completion)bond148#3 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex rcu_read_lock &ndev->lock irq_context: 0 (wq_completion)bond148#3 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex nl_table_lock irq_context: 0 (wq_completion)bond148#3 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &rq->__lock irq_context: 0 (wq_completion)bond148#3 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)bond199#2 (work_completion)(&(&bond->mcast_work)->work) irq_context: 0 (wq_completion)bond148#3 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &obj_hash[i].lock irq_context: 0 (wq_completion)bond199#2 (work_completion)(&(&bond->mcast_work)->work) &obj_hash[i].lock irq_context: 0 (wq_completion)bond148#3 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex nl_table_wait.lock irq_context: 0 (wq_completion)bond199#2 (work_completion)(&(&bond->mcast_work)->work) &base->lock irq_context: 0 (wq_completion)bond148#3 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex net_rwsem irq_context: 0 (wq_completion)bond199#2 (work_completion)(&(&bond->mcast_work)->work) &base->lock &obj_hash[i].lock irq_context: 0 (wq_completion)bond148#3 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex net_rwsem &list->lock#2 irq_context: 0 (wq_completion)bond148#3 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &tn->lock irq_context: 0 (wq_completion)bond148#3 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &idev->mc_lock irq_context: 0 (wq_completion)bond148#3 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &idev->mc_lock fs_reclaim irq_context: 0 (wq_completion)bond148#3 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &idev->mc_lock fs_reclaim mmu_notifier_invalidate_range_start irq_context: 0 (wq_completion)bond148#3 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &idev->mc_lock rcu_read_lock &ndev->lock irq_context: 0 (wq_completion)bond148#3 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &idev->mc_lock rcu_read_lock &dir->lock#2 irq_context: 0 (wq_completion)bond148#3 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &idev->mc_lock rcu_read_lock &ul->lock#2 irq_context: 0 (wq_completion)bond148#3 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &idev->mc_lock rcu_read_lock rcu_read_lock &obj_hash[i].lock irq_context: 0 (wq_completion)bond148#3 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &idev->mc_lock rcu_read_lock rcu_read_lock &c->lock irq_context: 0 (wq_completion)bond148#3 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &idev->mc_lock rcu_read_lock rcu_read_lock &n->list_lock irq_context: 0 (wq_completion)bond148#3 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &idev->mc_lock rcu_read_lock rcu_read_lock &n->list_lock &c->lock irq_context: 0 (wq_completion)bond148#3 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &idev->mc_lock rcu_read_lock rcu_read_lock rcu_node_0 irq_context: 0 (wq_completion)bond148#3 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &idev->mc_lock rcu_read_lock rcu_read_lock &rq->__lock irq_context: 0 (wq_completion)bond148#3 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &idev->mc_lock rcu_read_lock rcu_read_lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)bond148#3 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &idev->mc_lock rcu_read_lock rcu_read_lock &dir->lock#2 irq_context: 0 (wq_completion)bond148#3 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &idev->mc_lock rcu_read_lock rcu_read_lock &tbl->lock irq_context: 0 (wq_completion)bond148#3 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &idev->mc_lock rcu_read_lock rcu_read_lock &n->lock irq_context: 0 (wq_completion)bond148#3 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &idev->mc_lock rcu_read_lock rcu_read_lock &____s->seqcount#9 irq_context: 0 (wq_completion)bond148#3 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &idev->mc_lock rcu_read_lock rcu_read_lock rcu_read_lock_bh rcu_read_lock rcu_read_lock_bh rcu_read_lock &br->multicast_lock irq_context: 0 (wq_completion)bond148#3 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &idev->mc_lock rcu_read_lock rcu_read_lock rcu_read_lock_bh rcu_read_lock rcu_read_lock_bh rcu_read_lock &br->multicast_lock &obj_hash[i].lock irq_context: 0 (wq_completion)bond148#3 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &idev->mc_lock rcu_read_lock rcu_read_lock rcu_read_lock_bh rcu_read_lock rcu_read_lock_bh rcu_read_lock &br->multicast_lock &base->lock irq_context: 0 (wq_completion)bond148#3 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &idev->mc_lock rcu_read_lock rcu_read_lock rcu_read_lock_bh rcu_read_lock rcu_read_lock_bh rcu_read_lock &br->multicast_lock &base->lock &obj_hash[i].lock irq_context: 0 (wq_completion)bond148#3 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &idev->mc_lock rcu_read_lock rcu_read_lock rcu_read_lock_bh rcu_read_lock rcu_read_lock_bh rcu_read_lock &data->lock irq_context: 0 (wq_completion)bond148#3 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &idev->mc_lock rcu_read_lock rcu_read_lock rcu_read_lock_bh rcu_read_lock rcu_read_lock_bh rcu_read_lock &obj_hash[i].lock irq_context: 0 (wq_completion)bond148#3 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &idev->mc_lock rcu_read_lock rcu_read_lock rcu_read_lock_bh rcu_read_lock rcu_read_lock_bh rcu_read_lock &obj_hash[i].lock pool_lock irq_context: 0 (wq_completion)bond148#3 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &idev->mc_lock rcu_read_lock rcu_node_0 irq_context: 0 (wq_completion)bond148#3 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex.wait_lock irq_context: 0 (wq_completion)bond148#3 (work_completion)(&(&bond->mcast_work)->work) &p->pi_lock irq_context: 0 (wq_completion)bond148#3 (work_completion)(&(&bond->mcast_work)->work) &p->pi_lock &rq->__lock irq_context: 0 (wq_completion)bond148#3 (work_completion)(&(&bond->mcast_work)->work) &p->pi_lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: softirq &(&bond->mii_work)->timer rcu_read_lock &pool->lock/1 fill_pool_map-wait-type-override pool_lock irq_context: 0 (wq_completion)bond211 irq_context: 0 (wq_completion)bond211 (work_completion)(&(&slave->notify_work)->work) irq_context: 0 (wq_completion)bond211 (work_completion)(&(&slave->notify_work)->work) &obj_hash[i].lock irq_context: 0 (wq_completion)bond211 (work_completion)(&(&slave->notify_work)->work) &base->lock irq_context: 0 (wq_completion)bond211 (work_completion)(&(&slave->notify_work)->work) &base->lock &obj_hash[i].lock irq_context: 0 (wq_completion)bond211 (work_completion)(&(&bond->mcast_work)->work) irq_context: 0 (wq_completion)bond211 (work_completion)(&(&bond->mcast_work)->work) &obj_hash[i].lock irq_context: 0 (wq_completion)bond211 (work_completion)(&(&bond->mcast_work)->work) &base->lock irq_context: 0 (wq_completion)bond211 (work_completion)(&(&bond->mcast_work)->work) &base->lock &obj_hash[i].lock irq_context: 0 (wq_completion)phy12 (work_completion)(&(&local->scan_work)->work) &local->mtx rcu_read_lock hwsim_radio_lock batched_entropy_u8.lock irq_context: 0 (wq_completion)phy12 (work_completion)(&(&local->scan_work)->work) &local->mtx rcu_read_lock hwsim_radio_lock kfence_freelist_lock irq_context: 0 (wq_completion)bond183#3 irq_context: 0 (wq_completion)bond183#3 (work_completion)(&(&slave->notify_work)->work) irq_context: 0 (wq_completion)bond183#3 (work_completion)(&(&slave->notify_work)->work) &obj_hash[i].lock irq_context: 0 (wq_completion)bond183#3 (work_completion)(&(&slave->notify_work)->work) &base->lock irq_context: 0 (wq_completion)bond183#3 (work_completion)(&(&slave->notify_work)->work) &base->lock &obj_hash[i].lock irq_context: 0 (wq_completion)bond183#3 (work_completion)(&(&slave->notify_work)->work) &rq->__lock irq_context: 0 (wq_completion)bond183#3 (work_completion)(&(&slave->notify_work)->work) &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)bond183#3 (work_completion)(&(&bond->mcast_work)->work) irq_context: 0 (wq_completion)bond183#3 (work_completion)(&(&bond->mcast_work)->work) &obj_hash[i].lock irq_context: 0 (wq_completion)bond183#3 (work_completion)(&(&bond->mcast_work)->work) &base->lock irq_context: 0 (wq_completion)bond183#3 (work_completion)(&(&bond->mcast_work)->work) &base->lock &obj_hash[i].lock irq_context: 0 (wq_completion)bond183#3 (work_completion)(&(&bond->mcast_work)->work) &rq->__lock irq_context: 0 (wq_completion)bond183#3 (work_completion)(&(&bond->mcast_work)->work) &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)bond41 &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)bond58#2 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex fs_reclaim irq_context: 0 (wq_completion)bond58#2 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex fs_reclaim mmu_notifier_invalidate_range_start irq_context: 0 (wq_completion)bond58#2 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &c->lock irq_context: 0 (wq_completion)bond58#2 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &n->list_lock irq_context: 0 (wq_completion)bond58#2 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &n->list_lock &c->lock irq_context: 0 (wq_completion)bond58#2 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &rq->__lock irq_context: 0 (wq_completion)bond58#2 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)bond58#2 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex rcu_read_lock &bond->stats_lock/1 irq_context: 0 (wq_completion)bond58#2 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex rcu_read_lock &ndev->lock irq_context: 0 (wq_completion)bond58#2 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex nl_table_lock irq_context: 0 (wq_completion)bond58#2 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &obj_hash[i].lock irq_context: 0 (wq_completion)bond58#2 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex nl_table_wait.lock irq_context: 0 (wq_completion)bond58#2 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex net_rwsem irq_context: 0 (wq_completion)bond58#2 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex net_rwsem &list->lock#2 irq_context: 0 (wq_completion)bond58#2 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &tn->lock irq_context: 0 (wq_completion)bond58#2 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &idev->mc_lock irq_context: 0 (wq_completion)bond58#2 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &idev->mc_lock fs_reclaim irq_context: 0 (wq_completion)bond58#2 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &idev->mc_lock fs_reclaim &rq->__lock irq_context: 0 (wq_completion)bond58#2 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &idev->mc_lock fs_reclaim &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)bond58#2 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &idev->mc_lock fs_reclaim mmu_notifier_invalidate_range_start irq_context: 0 (wq_completion)bond58#2 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &idev->mc_lock &c->lock irq_context: 0 (wq_completion)bond58#2 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &idev->mc_lock &n->list_lock irq_context: 0 (wq_completion)bond58#2 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &idev->mc_lock &n->list_lock &c->lock irq_context: 0 (wq_completion)bond58#2 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &idev->mc_lock rcu_read_lock &ndev->lock irq_context: 0 (wq_completion)bond58#2 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &idev->mc_lock rcu_read_lock &dir->lock#2 irq_context: 0 (wq_completion)bond58#2 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &idev->mc_lock rcu_read_lock &c->lock irq_context: 0 (wq_completion)bond58#2 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &idev->mc_lock rcu_read_lock &ul->lock#2 irq_context: 0 (wq_completion)bond58#2 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &idev->mc_lock rcu_read_lock rcu_read_lock &____s->seqcount#12 irq_context: 0 (wq_completion)bond58#2 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &idev->mc_lock rcu_read_lock rcu_read_lock rcu_node_0 irq_context: 0 (wq_completion)bond58#2 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &idev->mc_lock rcu_read_lock rcu_read_lock &rq->__lock irq_context: 0 (wq_completion)bond58#2 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &idev->mc_lock rcu_read_lock rcu_read_lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)bond58#2 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &idev->mc_lock rcu_read_lock rcu_read_lock &____s->seqcount#16 irq_context: 0 (wq_completion)bond58#2 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &idev->mc_lock rcu_read_lock &obj_hash[i].lock irq_context: 0 (wq_completion)bond58#2 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &idev->mc_lock rcu_read_lock &data->lock irq_context: 0 (wq_completion)bond58#2 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &idev->mc_lock rcu_read_lock rcu_node_0 irq_context: 0 (wq_completion)bond58#2 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex.wait_lock irq_context: 0 (wq_completion)bond58#2 (work_completion)(&(&bond->mcast_work)->work) &p->pi_lock irq_context: 0 (wq_completion)bond39 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex fs_reclaim irq_context: 0 (wq_completion)bond39 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex fs_reclaim mmu_notifier_invalidate_range_start irq_context: 0 (wq_completion)bond39 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex fs_reclaim mmu_notifier_invalidate_range_start &rq->__lock irq_context: 0 (wq_completion)bond39 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex fs_reclaim mmu_notifier_invalidate_range_start &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)bond39 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &c->lock irq_context: 0 (wq_completion)bond39 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex rcu_read_lock &bond->stats_lock/1 irq_context: 0 (wq_completion)bond39 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex rcu_read_lock &ndev->lock irq_context: 0 (wq_completion)bond39 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex nl_table_lock irq_context: 0 (wq_completion)bond39 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &obj_hash[i].lock irq_context: 0 (wq_completion)bond39 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex nl_table_wait.lock irq_context: 0 (wq_completion)bond39 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex net_rwsem irq_context: 0 (wq_completion)bond39 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex net_rwsem &list->lock#2 irq_context: 0 (wq_completion)bond39 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &tn->lock irq_context: 0 (wq_completion)bond39 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &idev->mc_lock irq_context: 0 (wq_completion)bond39 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &idev->mc_lock fs_reclaim irq_context: 0 (wq_completion)bond39 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &idev->mc_lock fs_reclaim mmu_notifier_invalidate_range_start irq_context: 0 (wq_completion)bond39 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &idev->mc_lock &c->lock irq_context: 0 (wq_completion)bond39 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &idev->mc_lock rcu_read_lock &ndev->lock irq_context: 0 (wq_completion)bond39 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &idev->mc_lock rcu_read_lock &c->lock irq_context: 0 (wq_completion)bond39 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &idev->mc_lock rcu_read_lock &dir->lock#2 irq_context: 0 (wq_completion)bond39 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &idev->mc_lock rcu_read_lock &ul->lock#2 irq_context: 0 (wq_completion)bond39 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &idev->mc_lock rcu_read_lock rcu_read_lock &____s->seqcount#12 irq_context: 0 (wq_completion)bond39 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &idev->mc_lock rcu_read_lock rcu_read_lock &____s->seqcount#16 irq_context: 0 (wq_completion)bond39 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &idev->mc_lock rcu_read_lock &obj_hash[i].lock irq_context: 0 (wq_completion)bond39 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &idev->mc_lock rcu_read_lock &obj_hash[i].lock pool_lock irq_context: 0 (wq_completion)bond39 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &idev->mc_lock rcu_read_lock &data->lock irq_context: 0 (wq_completion)bond39 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex.wait_lock irq_context: 0 (wq_completion)bond39 (work_completion)(&(&bond->mcast_work)->work) &p->pi_lock irq_context: 0 (wq_completion)bond200#2 irq_context: 0 (wq_completion)bond200#2 &rq->__lock irq_context: 0 (wq_completion)bond200#2 &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)bond200#2 (work_completion)(&(&slave->notify_work)->work) irq_context: 0 (wq_completion)bond200#2 (work_completion)(&(&slave->notify_work)->work) &obj_hash[i].lock irq_context: 0 (wq_completion)bond200#2 (work_completion)(&(&slave->notify_work)->work) &base->lock irq_context: 0 (wq_completion)bond200#2 (work_completion)(&(&slave->notify_work)->work) &base->lock &obj_hash[i].lock irq_context: 0 (wq_completion)bond199#2 (work_completion)(&(&bond->mcast_work)->work) &rq->__lock irq_context: 0 (wq_completion)bond199#2 (work_completion)(&(&bond->mcast_work)->work) &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)bond200#2 (work_completion)(&(&bond->mcast_work)->work) irq_context: 0 (wq_completion)bond200#2 (work_completion)(&(&bond->mcast_work)->work) &obj_hash[i].lock irq_context: 0 (wq_completion)bond200#2 (work_completion)(&(&bond->mcast_work)->work) &base->lock irq_context: 0 (wq_completion)bond200#2 (work_completion)(&(&bond->mcast_work)->work) &base->lock &obj_hash[i].lock irq_context: 0 (wq_completion)bond178#3 (work_completion)(&(&slave->notify_work)->work) &rq->__lock irq_context: 0 (wq_completion)bond178#3 (work_completion)(&(&slave->notify_work)->work) &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)bond212 irq_context: 0 (wq_completion)bond212 (work_completion)(&(&slave->notify_work)->work) irq_context: 0 (wq_completion)bond212 (work_completion)(&(&slave->notify_work)->work) &obj_hash[i].lock irq_context: 0 (wq_completion)bond212 (work_completion)(&(&slave->notify_work)->work) &base->lock irq_context: 0 (wq_completion)bond212 (work_completion)(&(&slave->notify_work)->work) &base->lock &obj_hash[i].lock irq_context: 0 (wq_completion)bond212 (work_completion)(&(&bond->mcast_work)->work) irq_context: 0 (wq_completion)bond212 (work_completion)(&(&bond->mcast_work)->work) &obj_hash[i].lock irq_context: 0 (wq_completion)bond212 (work_completion)(&(&bond->mcast_work)->work) &base->lock irq_context: 0 (wq_completion)bond212 (work_completion)(&(&bond->mcast_work)->work) &base->lock &obj_hash[i].lock irq_context: 0 (wq_completion)bond212 (work_completion)(&(&bond->mcast_work)->work) &rq->__lock irq_context: 0 (wq_completion)bond212 (work_completion)(&(&bond->mcast_work)->work) &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)bond60#2 irq_context: 0 (wq_completion)bond60#2 &rq->__lock irq_context: 0 (wq_completion)bond60#2 &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)bond60#2 (work_completion)(&(&slave->notify_work)->work) irq_context: 0 (wq_completion)bond60#2 (work_completion)(&(&slave->notify_work)->work) &obj_hash[i].lock irq_context: 0 (wq_completion)bond60#2 (work_completion)(&(&slave->notify_work)->work) &base->lock irq_context: 0 (wq_completion)bond60#2 (work_completion)(&(&slave->notify_work)->work) &base->lock &obj_hash[i].lock irq_context: 0 (wq_completion)bond60#2 (work_completion)(&(&bond->mcast_work)->work) irq_context: 0 (wq_completion)bond60#2 (work_completion)(&(&bond->mcast_work)->work) &obj_hash[i].lock irq_context: 0 (wq_completion)bond60#2 (work_completion)(&(&bond->mcast_work)->work) &base->lock irq_context: 0 (wq_completion)bond60#2 (work_completion)(&(&bond->mcast_work)->work) &base->lock &obj_hash[i].lock irq_context: 0 (wq_completion)bond60#2 (work_completion)(&(&bond->mcast_work)->work) &rq->__lock irq_context: 0 (wq_completion)bond60#2 (work_completion)(&(&bond->mcast_work)->work) &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)bond208 (work_completion)(&(&slave->notify_work)->work) fill_pool_map-wait-type-override pool_lock irq_context: 0 (wq_completion)bond184#3 irq_context: 0 (wq_completion)bond184#3 (work_completion)(&(&slave->notify_work)->work) irq_context: 0 (wq_completion)bond184#3 (work_completion)(&(&slave->notify_work)->work) &obj_hash[i].lock irq_context: 0 (wq_completion)bond184#3 (work_completion)(&(&slave->notify_work)->work) &base->lock irq_context: 0 (wq_completion)bond184#3 (work_completion)(&(&slave->notify_work)->work) &base->lock &obj_hash[i].lock irq_context: 0 (wq_completion)bond184#3 (work_completion)(&(&slave->notify_work)->work) &rq->__lock irq_context: 0 (wq_completion)bond184#3 (work_completion)(&(&slave->notify_work)->work) &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)bond168#2 &rq->__lock irq_context: 0 (wq_completion)bond168#2 &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)bond184#3 (work_completion)(&(&bond->mcast_work)->work) irq_context: 0 (wq_completion)bond184#3 (work_completion)(&(&bond->mcast_work)->work) &obj_hash[i].lock irq_context: 0 (wq_completion)bond184#3 (work_completion)(&(&bond->mcast_work)->work) &base->lock irq_context: 0 (wq_completion)bond184#3 (work_completion)(&(&bond->mcast_work)->work) &base->lock &obj_hash[i].lock irq_context: 0 (wq_completion)bond184#3 (work_completion)(&(&bond->mcast_work)->work) &rq->__lock irq_context: 0 (wq_completion)bond184#3 (work_completion)(&(&bond->mcast_work)->work) &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)bond188 &rq->__lock irq_context: 0 (wq_completion)bond188 &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)bond195#2 (work_completion)(&(&slave->notify_work)->work) &rq->__lock irq_context: 0 (wq_completion)bond195#2 (work_completion)(&(&slave->notify_work)->work) &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)bond195#2 (work_completion)(&(&slave->notify_work)->work) &cfs_rq->removed.lock irq_context: 0 (wq_completion)bond135#2 &rq->__lock irq_context: 0 (wq_completion)bond135#2 &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)bond173#3 (work_completion)(&(&bond->mcast_work)->work) &rq->__lock irq_context: 0 (wq_completion)bond173#3 (work_completion)(&(&bond->mcast_work)->work) &rq->__lock &cfs_rq->removed.lock irq_context: 0 (wq_completion)bond173#3 (work_completion)(&(&bond->mcast_work)->work) &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)bond182#3 (work_completion)(&(&bond->mcast_work)->work) &rq->__lock irq_context: 0 (wq_completion)bond157#3 &rq->__lock irq_context: 0 (wq_completion)bond157#3 &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)bond169 (work_completion)(&(&bond->mcast_work)->work) &base->lock fill_pool_map-wait-type-override pool_lock irq_context: 0 (wq_completion)bond201#2 irq_context: 0 (wq_completion)bond201#2 (work_completion)(&(&slave->notify_work)->work) irq_context: 0 (wq_completion)bond201#2 (work_completion)(&(&slave->notify_work)->work) &obj_hash[i].lock irq_context: 0 (wq_completion)bond201#2 (work_completion)(&(&slave->notify_work)->work) &base->lock irq_context: 0 (wq_completion)bond201#2 (work_completion)(&(&slave->notify_work)->work) &base->lock &obj_hash[i].lock irq_context: 0 (wq_completion)bond201#2 (work_completion)(&(&slave->notify_work)->work) &rq->__lock irq_context: 0 (wq_completion)bond201#2 (work_completion)(&(&slave->notify_work)->work) &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)bond201#2 (work_completion)(&(&bond->mcast_work)->work) irq_context: 0 (wq_completion)bond201#2 (work_completion)(&(&bond->mcast_work)->work) &obj_hash[i].lock irq_context: 0 (wq_completion)bond201#2 (work_completion)(&(&bond->mcast_work)->work) &base->lock irq_context: 0 (wq_completion)bond201#2 (work_completion)(&(&bond->mcast_work)->work) &base->lock &obj_hash[i].lock irq_context: 0 (wq_completion)bond201#2 (work_completion)(&(&bond->mcast_work)->work) &rq->__lock irq_context: 0 (wq_completion)bond201#2 (work_completion)(&(&bond->mcast_work)->work) &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)bond195 (work_completion)(&(&bond->mcast_work)->work) &rq->__lock irq_context: 0 (wq_completion)bond195 (work_completion)(&(&bond->mcast_work)->work) &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)bond160#3 (work_completion)(&(&bond->mcast_work)->work) &rq->__lock irq_context: 0 (wq_completion)bond160#3 (work_completion)(&(&bond->mcast_work)->work) &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)bond213 irq_context: 0 (wq_completion)bond213 (work_completion)(&(&slave->notify_work)->work) irq_context: 0 (wq_completion)bond213 (work_completion)(&(&slave->notify_work)->work) &obj_hash[i].lock irq_context: 0 (wq_completion)bond213 (work_completion)(&(&slave->notify_work)->work) &base->lock irq_context: 0 (wq_completion)bond213 (work_completion)(&(&slave->notify_work)->work) &base->lock &obj_hash[i].lock irq_context: 0 (wq_completion)bond213 (work_completion)(&(&slave->notify_work)->work) &rq->__lock irq_context: 0 (wq_completion)bond213 (work_completion)(&(&slave->notify_work)->work) &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: softirq rcu_read_lock rcu_read_lock slock-AF_INET6/1 sk_lock-AF_INET6 rcu_read_lock &pcp->lock &zone->lock irq_context: softirq rcu_read_lock rcu_read_lock slock-AF_INET6/1 sk_lock-AF_INET6 rcu_read_lock &pcp->lock &zone->lock &____s->seqcount irq_context: 0 (wq_completion)bond213 (work_completion)(&(&bond->mcast_work)->work) irq_context: 0 (wq_completion)bond213 (work_completion)(&(&bond->mcast_work)->work) &obj_hash[i].lock irq_context: 0 (wq_completion)bond213 (work_completion)(&(&bond->mcast_work)->work) &base->lock irq_context: 0 (wq_completion)bond213 (work_completion)(&(&bond->mcast_work)->work) &base->lock &obj_hash[i].lock irq_context: 0 (wq_completion)bond213 (work_completion)(&(&bond->mcast_work)->work) &rq->__lock irq_context: 0 (wq_completion)bond213 (work_completion)(&(&bond->mcast_work)->work) &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)bond187#2 (work_completion)(&(&bond->mcast_work)->work) fill_pool_map-wait-type-override pool_lock irq_context: 0 (wq_completion)bond164 (work_completion)(&(&bond->mcast_work)->work) &cfs_rq->removed.lock irq_context: 0 (wq_completion)bond185#3 irq_context: 0 (wq_completion)bond185#3 (work_completion)(&(&slave->notify_work)->work) irq_context: 0 (wq_completion)bond185#3 (work_completion)(&(&slave->notify_work)->work) &obj_hash[i].lock irq_context: 0 (wq_completion)bond185#3 (work_completion)(&(&slave->notify_work)->work) &base->lock irq_context: 0 (wq_completion)bond185#3 (work_completion)(&(&slave->notify_work)->work) &base->lock &obj_hash[i].lock irq_context: 0 (wq_completion)bond144#3 &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)bond185#3 (work_completion)(&(&bond->mcast_work)->work) irq_context: 0 (wq_completion)bond185#3 (work_completion)(&(&bond->mcast_work)->work) &obj_hash[i].lock irq_context: 0 (wq_completion)bond185#3 (work_completion)(&(&bond->mcast_work)->work) &base->lock irq_context: 0 (wq_completion)bond185#3 (work_completion)(&(&bond->mcast_work)->work) &base->lock &obj_hash[i].lock irq_context: 0 (wq_completion)bond185#3 (work_completion)(&(&bond->mcast_work)->work) &rq->__lock irq_context: 0 (wq_completion)bond185#3 (work_completion)(&(&bond->mcast_work)->work) &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: softirq &(&bat_priv->tt.work)->timer rcu_read_lock &pool->lock/1 &p->pi_lock &rq->__lock &obj_hash[i].lock irq_context: softirq &(&bat_priv->tt.work)->timer rcu_read_lock &pool->lock/1 &p->pi_lock &rq->__lock &base->lock irq_context: softirq &(&bat_priv->tt.work)->timer rcu_read_lock &pool->lock/1 &p->pi_lock &rq->__lock &base->lock &obj_hash[i].lock irq_context: 0 rtnl_mutex uevent_sock_mutex fs_reclaim mmu_notifier_invalidate_range_start &cfs_rq->removed.lock irq_context: 0 rtnl_mutex uevent_sock_mutex fs_reclaim mmu_notifier_invalidate_range_start &obj_hash[i].lock irq_context: 0 (wq_completion)bond162#2 (work_completion)(&(&bond->mcast_work)->work) fill_pool_map-wait-type-override pool_lock irq_context: 0 (wq_completion)bond202#2 irq_context: 0 (wq_completion)bond202#2 (work_completion)(&(&slave->notify_work)->work) irq_context: 0 (wq_completion)bond202#2 (work_completion)(&(&slave->notify_work)->work) &obj_hash[i].lock irq_context: 0 (wq_completion)bond202#2 (work_completion)(&(&slave->notify_work)->work) &base->lock irq_context: 0 (wq_completion)bond202#2 (work_completion)(&(&slave->notify_work)->work) &base->lock &obj_hash[i].lock irq_context: 0 (wq_completion)bond202#2 (work_completion)(&(&slave->notify_work)->work) &rq->__lock irq_context: 0 (wq_completion)bond202#2 (work_completion)(&(&slave->notify_work)->work) &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)bond194 (work_completion)(&(&slave->notify_work)->work) rtnl_mutex net_rwsem irq_context: 0 (wq_completion)bond194 (work_completion)(&(&slave->notify_work)->work) rtnl_mutex net_rwsem &list->lock#2 irq_context: 0 (wq_completion)bond194 (work_completion)(&(&slave->notify_work)->work) rtnl_mutex &tn->lock irq_context: 0 (wq_completion)bond194 (work_completion)(&(&slave->notify_work)->work) rtnl_mutex.wait_lock irq_context: 0 (wq_completion)bond194 (work_completion)(&(&slave->notify_work)->work) &p->pi_lock irq_context: 0 (wq_completion)bond193#2 (work_completion)(&(&slave->notify_work)->work) rtnl_mutex net_rwsem irq_context: 0 (wq_completion)bond193#2 (work_completion)(&(&slave->notify_work)->work) rtnl_mutex net_rwsem &list->lock#2 irq_context: 0 (wq_completion)bond193#2 (work_completion)(&(&slave->notify_work)->work) rtnl_mutex &tn->lock irq_context: 0 (wq_completion)bond193#2 (work_completion)(&(&slave->notify_work)->work) rtnl_mutex rcu_read_lock rcu_node_0 irq_context: 0 (wq_completion)bond193#2 (work_completion)(&(&slave->notify_work)->work) rtnl_mutex rcu_read_lock &rq->__lock irq_context: 0 (wq_completion)bond193#2 (work_completion)(&(&slave->notify_work)->work) rtnl_mutex rcu_read_lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)bond193#2 (work_completion)(&(&slave->notify_work)->work) rtnl_mutex.wait_lock irq_context: 0 (wq_completion)bond193#2 (work_completion)(&(&slave->notify_work)->work) &p->pi_lock irq_context: 0 (wq_completion)bond200#2 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex fs_reclaim irq_context: 0 (wq_completion)bond200#2 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex fs_reclaim mmu_notifier_invalidate_range_start irq_context: 0 (wq_completion)bond200#2 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &c->lock irq_context: 0 (wq_completion)bond200#2 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex rcu_read_lock &bond->stats_lock/1 irq_context: 0 (wq_completion)bond200#2 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex rcu_read_lock &ndev->lock irq_context: 0 (wq_completion)bond200#2 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex nl_table_lock irq_context: 0 (wq_completion)bond200#2 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &obj_hash[i].lock irq_context: 0 (wq_completion)bond200#2 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &rq->__lock irq_context: 0 (wq_completion)bond200#2 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)bond200#2 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex nl_table_wait.lock irq_context: 0 (wq_completion)bond200#2 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex net_rwsem irq_context: 0 (wq_completion)bond200#2 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex net_rwsem &list->lock#2 irq_context: 0 (wq_completion)bond200#2 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &tn->lock irq_context: 0 (wq_completion)bond200#2 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &idev->mc_lock irq_context: 0 (wq_completion)bond200#2 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &idev->mc_lock fs_reclaim irq_context: 0 (wq_completion)bond200#2 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &idev->mc_lock fs_reclaim mmu_notifier_invalidate_range_start irq_context: 0 (wq_completion)bond200#2 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &idev->mc_lock &c->lock irq_context: 0 (wq_completion)bond200#2 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &idev->mc_lock &n->list_lock irq_context: 0 (wq_completion)bond200#2 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &idev->mc_lock &n->list_lock &c->lock irq_context: 0 (wq_completion)bond200#2 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &idev->mc_lock rcu_read_lock &ndev->lock irq_context: 0 (wq_completion)bond200#2 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &idev->mc_lock rcu_read_lock &dir->lock#2 irq_context: 0 (wq_completion)bond200#2 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &idev->mc_lock rcu_read_lock &ul->lock#2 irq_context: 0 (wq_completion)bond200#2 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &idev->mc_lock rcu_read_lock rcu_read_lock &____s->seqcount#10 irq_context: 0 (wq_completion)bond200#2 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &idev->mc_lock rcu_read_lock rcu_read_lock rcu_read_lock_bh rcu_read_lock rcu_read_lock_bh rcu_read_lock &br->multicast_lock irq_context: 0 (wq_completion)bond200#2 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &idev->mc_lock rcu_read_lock rcu_read_lock rcu_read_lock_bh rcu_read_lock rcu_read_lock_bh rcu_read_lock &br->multicast_lock &obj_hash[i].lock irq_context: 0 (wq_completion)bond200#2 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &idev->mc_lock rcu_read_lock rcu_read_lock rcu_read_lock_bh rcu_read_lock rcu_read_lock_bh rcu_read_lock &br->multicast_lock &base->lock irq_context: 0 (wq_completion)bond200#2 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &idev->mc_lock rcu_read_lock rcu_read_lock rcu_read_lock_bh rcu_read_lock rcu_read_lock_bh rcu_read_lock &data->lock irq_context: 0 (wq_completion)bond200#2 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &idev->mc_lock rcu_read_lock rcu_read_lock rcu_read_lock_bh rcu_read_lock rcu_read_lock_bh rcu_read_lock &obj_hash[i].lock irq_context: 0 (wq_completion)bond143#3 &rq->__lock irq_context: 0 (wq_completion)bond143#3 &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)bond200#2 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex.wait_lock irq_context: 0 (wq_completion)bond200#2 (work_completion)(&(&bond->mcast_work)->work) &p->pi_lock irq_context: 0 (wq_completion)bond200#2 (work_completion)(&(&bond->mcast_work)->work) &p->pi_lock &rq->__lock irq_context: 0 (wq_completion)bond200#2 (work_completion)(&(&bond->mcast_work)->work) &p->pi_lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)bond202#2 (work_completion)(&(&bond->mcast_work)->work) irq_context: 0 (wq_completion)bond202#2 (work_completion)(&(&bond->mcast_work)->work) &rq->__lock irq_context: 0 (wq_completion)bond202#2 (work_completion)(&(&bond->mcast_work)->work) &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)bond202#2 (work_completion)(&(&bond->mcast_work)->work) &obj_hash[i].lock irq_context: 0 (wq_completion)bond202#2 (work_completion)(&(&bond->mcast_work)->work) &base->lock irq_context: 0 (wq_completion)bond202#2 (work_completion)(&(&bond->mcast_work)->work) &base->lock &obj_hash[i].lock irq_context: 0 (wq_completion)bond209 (work_completion)(&(&bond->mcast_work)->work) &rq->__lock irq_context: 0 (wq_completion)bond209 (work_completion)(&(&bond->mcast_work)->work) &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)bond142#2 (work_completion)(&(&slave->notify_work)->work) fill_pool_map-wait-type-override pool_lock irq_context: 0 (wq_completion)bond214 irq_context: 0 (wq_completion)bond214 (work_completion)(&(&slave->notify_work)->work) irq_context: 0 (wq_completion)bond214 (work_completion)(&(&slave->notify_work)->work) &obj_hash[i].lock irq_context: 0 (wq_completion)bond214 (work_completion)(&(&slave->notify_work)->work) &base->lock irq_context: 0 (wq_completion)bond214 (work_completion)(&(&slave->notify_work)->work) &base->lock &obj_hash[i].lock irq_context: 0 (wq_completion)bond214 (work_completion)(&(&slave->notify_work)->work) &rq->__lock irq_context: 0 (wq_completion)bond214 (work_completion)(&(&bond->mcast_work)->work) irq_context: 0 (wq_completion)bond214 (work_completion)(&(&bond->mcast_work)->work) &obj_hash[i].lock irq_context: 0 (wq_completion)bond214 (work_completion)(&(&bond->mcast_work)->work) &base->lock irq_context: 0 (wq_completion)bond214 (work_completion)(&(&bond->mcast_work)->work) &base->lock &obj_hash[i].lock irq_context: 0 (wq_completion)bond191#2 (work_completion)(&(&bond->mcast_work)->work) &rq->__lock irq_context: 0 (wq_completion)bond191#2 (work_completion)(&(&bond->mcast_work)->work) &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)bond203#2 irq_context: 0 (wq_completion)bond203#2 (work_completion)(&(&slave->notify_work)->work) irq_context: 0 (wq_completion)bond203#2 (work_completion)(&(&slave->notify_work)->work) &obj_hash[i].lock irq_context: 0 (wq_completion)bond203#2 (work_completion)(&(&slave->notify_work)->work) &base->lock irq_context: 0 (wq_completion)bond203#2 (work_completion)(&(&slave->notify_work)->work) &base->lock &obj_hash[i].lock irq_context: 0 (wq_completion)bond203#2 (work_completion)(&(&slave->notify_work)->work) &rq->__lock irq_context: 0 (wq_completion)bond203#2 (work_completion)(&(&slave->notify_work)->work) &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)bond129 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex fs_reclaim irq_context: 0 (wq_completion)bond129 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex fs_reclaim mmu_notifier_invalidate_range_start irq_context: 0 (wq_completion)bond129 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &c->lock irq_context: 0 (wq_completion)bond129 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &n->list_lock irq_context: 0 (wq_completion)bond129 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &n->list_lock &c->lock irq_context: 0 (wq_completion)bond129 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &rq->__lock irq_context: 0 (wq_completion)bond203#2 (work_completion)(&(&bond->mcast_work)->work) irq_context: 0 (wq_completion)bond203#2 (work_completion)(&(&bond->mcast_work)->work) &obj_hash[i].lock irq_context: 0 (wq_completion)bond203#2 (work_completion)(&(&bond->mcast_work)->work) &base->lock irq_context: 0 (wq_completion)bond203#2 (work_completion)(&(&bond->mcast_work)->work) &base->lock &obj_hash[i].lock irq_context: 0 (wq_completion)bond129 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex rcu_read_lock &bond->stats_lock/1 irq_context: 0 (wq_completion)bond129 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex rcu_read_lock &ndev->lock irq_context: 0 (wq_completion)bond129 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex nl_table_lock irq_context: 0 (wq_completion)bond129 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &obj_hash[i].lock irq_context: 0 (wq_completion)bond129 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex nl_table_wait.lock irq_context: 0 (wq_completion)bond129 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex net_rwsem irq_context: 0 (wq_completion)bond129 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex net_rwsem &list->lock#2 irq_context: 0 (wq_completion)bond129 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &tn->lock irq_context: 0 (wq_completion)bond129 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &idev->mc_lock irq_context: 0 (wq_completion)bond129 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &idev->mc_lock fs_reclaim irq_context: 0 (wq_completion)bond129 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &idev->mc_lock fs_reclaim mmu_notifier_invalidate_range_start irq_context: 0 (wq_completion)bond129 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &idev->mc_lock &c->lock irq_context: 0 (wq_completion)bond129 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &idev->mc_lock rcu_read_lock &ndev->lock irq_context: 0 (wq_completion)bond129 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &idev->mc_lock rcu_read_lock &dir->lock#2 irq_context: 0 (wq_completion)bond129 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &idev->mc_lock rcu_read_lock &ul->lock#2 irq_context: 0 (wq_completion)bond129 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &idev->mc_lock rcu_read_lock rcu_read_lock &obj_hash[i].lock irq_context: 0 (wq_completion)bond129 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &idev->mc_lock rcu_read_lock rcu_read_lock &dir->lock#2 irq_context: 0 (wq_completion)bond129 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &idev->mc_lock rcu_read_lock rcu_read_lock &tbl->lock irq_context: 0 (wq_completion)bond129 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &idev->mc_lock rcu_read_lock rcu_read_lock &n->lock irq_context: 0 (wq_completion)bond129 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &idev->mc_lock rcu_read_lock rcu_read_lock &____s->seqcount#9 irq_context: 0 (wq_completion)bond129 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &idev->mc_lock rcu_read_lock rcu_read_lock rcu_read_lock_bh rcu_read_lock rcu_read_lock_bh rcu_read_lock &br->multicast_lock irq_context: 0 (wq_completion)bond129 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &idev->mc_lock rcu_read_lock rcu_read_lock rcu_read_lock_bh rcu_read_lock rcu_read_lock_bh rcu_read_lock &br->multicast_lock &obj_hash[i].lock irq_context: 0 (wq_completion)bond129 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &idev->mc_lock rcu_read_lock rcu_read_lock rcu_read_lock_bh rcu_read_lock rcu_read_lock_bh rcu_read_lock &br->multicast_lock &base->lock irq_context: 0 (wq_completion)bond129 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &idev->mc_lock rcu_read_lock rcu_read_lock rcu_read_lock_bh rcu_read_lock rcu_read_lock_bh rcu_read_lock &br->multicast_lock &base->lock &obj_hash[i].lock irq_context: 0 (wq_completion)bond129 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &idev->mc_lock rcu_read_lock rcu_read_lock rcu_read_lock_bh rcu_read_lock rcu_read_lock_bh rcu_read_lock &data->lock irq_context: 0 (wq_completion)bond129 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &idev->mc_lock rcu_read_lock rcu_read_lock rcu_read_lock_bh rcu_read_lock rcu_read_lock_bh rcu_read_lock &obj_hash[i].lock irq_context: 0 (wq_completion)bond129 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &idev->mc_lock rcu_read_lock rcu_read_lock rcu_node_0 irq_context: 0 (wq_completion)bond129 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &idev->mc_lock rcu_read_lock rcu_read_lock &rq->__lock irq_context: 0 (wq_completion)bond129 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &idev->mc_lock rcu_read_lock rcu_node_0 irq_context: 0 (wq_completion)bond129 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &idev->mc_lock rcu_read_lock &rcu_state.gp_wq irq_context: 0 (wq_completion)bond129 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &idev->mc_lock rcu_read_lock &rcu_state.gp_wq &p->pi_lock irq_context: 0 (wq_completion)bond129 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &idev->mc_lock rcu_read_lock &rcu_state.gp_wq &p->pi_lock &rq->__lock irq_context: 0 (wq_completion)bond129 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &idev->mc_lock rcu_read_lock &rcu_state.gp_wq &p->pi_lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)bond129 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &idev->mc_lock rcu_read_lock &rq->__lock irq_context: 0 (wq_completion)bond129 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &idev->mc_lock rcu_read_lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)bond129 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)bond129 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex.wait_lock irq_context: 0 (wq_completion)bond129 (work_completion)(&(&bond->mcast_work)->work) &p->pi_lock irq_context: 0 (wq_completion)bond129 (work_completion)(&(&bond->mcast_work)->work) &p->pi_lock &rq->__lock irq_context: 0 (wq_completion)bond129 (work_completion)(&(&bond->mcast_work)->work) &p->pi_lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)bond214 (work_completion)(&(&bond->mcast_work)->work) &rq->__lock irq_context: 0 (wq_completion)bond214 (work_completion)(&(&bond->mcast_work)->work) &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)bond187#3 irq_context: 0 (wq_completion)bond187#3 (work_completion)(&(&slave->notify_work)->work) irq_context: 0 (wq_completion)bond187#3 (work_completion)(&(&slave->notify_work)->work) &obj_hash[i].lock irq_context: 0 (wq_completion)bond187#3 (work_completion)(&(&slave->notify_work)->work) &base->lock irq_context: 0 (wq_completion)bond187#3 (work_completion)(&(&slave->notify_work)->work) &base->lock &obj_hash[i].lock irq_context: 0 (wq_completion)bond187#3 (work_completion)(&(&bond->mcast_work)->work) irq_context: 0 (wq_completion)bond187#3 (work_completion)(&(&bond->mcast_work)->work) &obj_hash[i].lock irq_context: 0 (wq_completion)bond187#3 (work_completion)(&(&bond->mcast_work)->work) &base->lock irq_context: 0 (wq_completion)bond187#3 (work_completion)(&(&bond->mcast_work)->work) &base->lock &obj_hash[i].lock irq_context: 0 (wq_completion)bond187#3 (work_completion)(&(&bond->mcast_work)->work) &rq->__lock irq_context: 0 (wq_completion)bond187#3 (work_completion)(&(&bond->mcast_work)->work) &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)bond201 (work_completion)(&(&slave->notify_work)->work) &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)bond204#2 irq_context: 0 (wq_completion)bond204#2 (work_completion)(&(&slave->notify_work)->work) irq_context: 0 (wq_completion)bond204#2 (work_completion)(&(&slave->notify_work)->work) &obj_hash[i].lock irq_context: 0 (wq_completion)bond204#2 (work_completion)(&(&slave->notify_work)->work) &base->lock irq_context: 0 (wq_completion)bond204#2 (work_completion)(&(&slave->notify_work)->work) &base->lock &obj_hash[i].lock irq_context: 0 (wq_completion)bond204#2 (work_completion)(&(&slave->notify_work)->work) &rq->__lock irq_context: 0 (wq_completion)bond204#2 (work_completion)(&(&bond->mcast_work)->work) irq_context: 0 (wq_completion)bond204#2 (work_completion)(&(&bond->mcast_work)->work) &obj_hash[i].lock irq_context: 0 (wq_completion)bond204#2 (work_completion)(&(&bond->mcast_work)->work) &base->lock irq_context: 0 (wq_completion)bond204#2 (work_completion)(&(&bond->mcast_work)->work) &base->lock &obj_hash[i].lock irq_context: 0 (wq_completion)bond204#2 (work_completion)(&(&bond->mcast_work)->work) &rq->__lock irq_context: 0 (wq_completion)bond204#2 (work_completion)(&(&bond->mcast_work)->work) &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)bond215 irq_context: 0 (wq_completion)bond215 &rq->__lock irq_context: 0 (wq_completion)bond215 &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)bond215 (work_completion)(&(&slave->notify_work)->work) irq_context: 0 (wq_completion)bond215 (work_completion)(&(&slave->notify_work)->work) &obj_hash[i].lock irq_context: 0 (wq_completion)bond215 (work_completion)(&(&slave->notify_work)->work) &base->lock irq_context: 0 (wq_completion)bond215 (work_completion)(&(&slave->notify_work)->work) &base->lock &obj_hash[i].lock irq_context: 0 (wq_completion)bond215 (work_completion)(&(&slave->notify_work)->work) &rq->__lock irq_context: 0 (wq_completion)bond215 (work_completion)(&(&slave->notify_work)->work) &cfs_rq->removed.lock irq_context: 0 (wq_completion)bond215 (work_completion)(&(&bond->mcast_work)->work) irq_context: 0 (wq_completion)bond215 (work_completion)(&(&bond->mcast_work)->work) &obj_hash[i].lock irq_context: 0 (wq_completion)bond215 (work_completion)(&(&bond->mcast_work)->work) &base->lock irq_context: 0 (wq_completion)bond215 (work_completion)(&(&bond->mcast_work)->work) &base->lock &obj_hash[i].lock irq_context: 0 (wq_completion)bond108#3 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex fs_reclaim irq_context: 0 (wq_completion)bond108#3 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex fs_reclaim &rq->__lock irq_context: 0 (wq_completion)bond108#3 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex fs_reclaim mmu_notifier_invalidate_range_start irq_context: 0 (wq_completion)bond108#3 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex fs_reclaim mmu_notifier_invalidate_range_start &rq->__lock irq_context: 0 (wq_completion)bond108#3 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &rq->__lock irq_context: 0 (wq_completion)bond108#3 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex rcu_read_lock &bond->stats_lock/1 irq_context: 0 (wq_completion)bond108#3 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex rcu_read_lock rcu_node_0 irq_context: 0 (wq_completion)bond108#3 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex rcu_read_lock &rq->__lock irq_context: 0 (wq_completion)bond108#3 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex rcu_read_lock &ndev->lock irq_context: 0 (wq_completion)bond108#3 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex nl_table_lock irq_context: 0 (wq_completion)bond108#3 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &obj_hash[i].lock irq_context: 0 (wq_completion)bond108#3 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex nl_table_wait.lock irq_context: 0 (wq_completion)bond108#3 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex net_rwsem irq_context: 0 (wq_completion)bond108#3 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex net_rwsem &rq->__lock irq_context: 0 (wq_completion)bond108#3 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex net_rwsem &list->lock#2 irq_context: 0 (wq_completion)bond108#3 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &tn->lock irq_context: 0 (wq_completion)bond108#3 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &tn->lock &rq->__lock irq_context: 0 (wq_completion)bond108#3 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &tn->lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)bond108#3 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &idev->mc_lock irq_context: 0 (wq_completion)bond108#3 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &idev->mc_lock &rq->__lock irq_context: 0 (wq_completion)bond108#3 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &idev->mc_lock fs_reclaim irq_context: 0 (wq_completion)bond108#3 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &idev->mc_lock fs_reclaim mmu_notifier_invalidate_range_start irq_context: 0 (wq_completion)bond108#3 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &idev->mc_lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)bond108#3 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &idev->mc_lock rcu_read_lock &ndev->lock irq_context: 0 (wq_completion)bond108#3 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &idev->mc_lock rcu_read_lock &dir->lock#2 irq_context: 0 (wq_completion)bond108#3 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &idev->mc_lock rcu_read_lock &ul->lock#2 irq_context: 0 (wq_completion)bond108#3 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &idev->mc_lock rcu_read_lock rcu_node_0 irq_context: 0 (wq_completion)bond108#3 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &idev->mc_lock rcu_read_lock &rq->__lock irq_context: 0 (wq_completion)bond108#3 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &idev->mc_lock rcu_read_lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)bond108#3 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &idev->mc_lock rcu_read_lock rcu_read_lock &obj_hash[i].lock irq_context: 0 (wq_completion)bond108#3 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &idev->mc_lock rcu_read_lock rcu_read_lock &rq->__lock irq_context: 0 (wq_completion)bond108#3 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &idev->mc_lock rcu_read_lock rcu_read_lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)bond108#3 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &idev->mc_lock rcu_read_lock rcu_read_lock &dir->lock#2 irq_context: 0 (wq_completion)bond108#3 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &idev->mc_lock rcu_read_lock rcu_read_lock &tbl->lock irq_context: 0 (wq_completion)bond108#3 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &idev->mc_lock rcu_read_lock rcu_read_lock &n->lock irq_context: 0 (wq_completion)bond108#3 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &idev->mc_lock rcu_read_lock rcu_read_lock &____s->seqcount#9 irq_context: 0 (wq_completion)bond108#3 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &idev->mc_lock rcu_read_lock rcu_read_lock rcu_read_lock_bh rcu_read_lock rcu_read_lock_bh rcu_read_lock &br->multicast_lock irq_context: 0 (wq_completion)bond108#3 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &idev->mc_lock rcu_read_lock rcu_read_lock rcu_read_lock_bh rcu_read_lock rcu_read_lock_bh rcu_read_lock &br->multicast_lock &obj_hash[i].lock irq_context: 0 (wq_completion)bond108#3 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &idev->mc_lock rcu_read_lock rcu_read_lock rcu_read_lock_bh rcu_read_lock rcu_read_lock_bh rcu_read_lock &br->multicast_lock &base->lock irq_context: 0 (wq_completion)bond108#3 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &idev->mc_lock rcu_read_lock rcu_read_lock rcu_read_lock_bh rcu_read_lock rcu_read_lock_bh rcu_read_lock &br->multicast_lock &base->lock &obj_hash[i].lock irq_context: 0 (wq_completion)bond108#3 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &idev->mc_lock rcu_read_lock rcu_read_lock rcu_read_lock_bh rcu_read_lock rcu_read_lock_bh rcu_read_lock &data->lock irq_context: 0 (wq_completion)bond108#3 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &idev->mc_lock rcu_read_lock rcu_read_lock rcu_read_lock_bh rcu_read_lock rcu_read_lock_bh rcu_read_lock &obj_hash[i].lock irq_context: 0 (wq_completion)bond108#3 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex.wait_lock irq_context: 0 (wq_completion)bond108#3 (work_completion)(&(&bond->mcast_work)->work) &p->pi_lock irq_context: 0 (wq_completion)bond108#3 (work_completion)(&(&bond->mcast_work)->work) &p->pi_lock &rq->__lock irq_context: 0 (wq_completion)bond108#3 (work_completion)(&(&bond->mcast_work)->work) &p->pi_lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)bond206 (work_completion)(&(&slave->notify_work)->work) &rq->__lock irq_context: 0 (wq_completion)bond206 (work_completion)(&(&slave->notify_work)->work) &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)bond174 &rq->__lock irq_context: 0 (wq_completion)bond188#3 irq_context: 0 (wq_completion)bond188#3 (work_completion)(&(&slave->notify_work)->work) irq_context: 0 (wq_completion)bond188#3 (work_completion)(&(&slave->notify_work)->work) &obj_hash[i].lock irq_context: 0 (wq_completion)bond188#3 (work_completion)(&(&slave->notify_work)->work) &base->lock irq_context: 0 (wq_completion)bond188#3 (work_completion)(&(&slave->notify_work)->work) &base->lock &obj_hash[i].lock irq_context: 0 (wq_completion)bond188#3 (work_completion)(&(&slave->notify_work)->work) &rq->__lock irq_context: 0 (wq_completion)bond187#3 (work_completion)(&(&slave->notify_work)->work) &rq->__lock irq_context: 0 (wq_completion)bond187#3 (work_completion)(&(&slave->notify_work)->work) &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)bond188#3 (work_completion)(&(&bond->mcast_work)->work) irq_context: 0 (wq_completion)bond188#3 (work_completion)(&(&bond->mcast_work)->work) &obj_hash[i].lock irq_context: 0 (wq_completion)bond188#3 (work_completion)(&(&bond->mcast_work)->work) &base->lock irq_context: 0 (wq_completion)bond188#3 (work_completion)(&(&bond->mcast_work)->work) &base->lock &obj_hash[i].lock irq_context: 0 (wq_completion)bond168#3 (work_completion)(&(&slave->notify_work)->work) &rq->__lock irq_context: 0 (wq_completion)bond168#3 (work_completion)(&(&slave->notify_work)->work) &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)bond147#3 (work_completion)(&(&bond->mcast_work)->work) fill_pool_map-wait-type-override &c->lock irq_context: 0 (wq_completion)bond147#3 (work_completion)(&(&bond->mcast_work)->work) fill_pool_map-wait-type-override &____s->seqcount#2 irq_context: 0 (wq_completion)bond147#3 (work_completion)(&(&bond->mcast_work)->work) fill_pool_map-wait-type-override &____s->seqcount irq_context: 0 (wq_completion)bond147#3 (work_completion)(&(&bond->mcast_work)->work) fill_pool_map-wait-type-override pool_lock irq_context: 0 (wq_completion)bond205#2 irq_context: 0 (wq_completion)bond205#2 (work_completion)(&(&slave->notify_work)->work) irq_context: 0 (wq_completion)bond205#2 (work_completion)(&(&slave->notify_work)->work) &obj_hash[i].lock irq_context: 0 (wq_completion)bond205#2 (work_completion)(&(&slave->notify_work)->work) &base->lock irq_context: 0 (wq_completion)bond205#2 (work_completion)(&(&slave->notify_work)->work) &base->lock &obj_hash[i].lock irq_context: 0 (wq_completion)bond205#2 (work_completion)(&(&slave->notify_work)->work) &rq->__lock irq_context: 0 (wq_completion)events (linkwatch_work).work rtnl_mutex fs_reclaim mmu_notifier_invalidate_range_start &rq->__lock irq_context: 0 (wq_completion)bond205#2 (work_completion)(&(&bond->mcast_work)->work) irq_context: 0 (wq_completion)bond178#3 (work_completion)(&(&slave->notify_work)->work) rtnl_mutex net_rwsem irq_context: 0 (wq_completion)bond205#2 (work_completion)(&(&bond->mcast_work)->work) &obj_hash[i].lock irq_context: 0 (wq_completion)bond178#3 (work_completion)(&(&slave->notify_work)->work) rtnl_mutex net_rwsem &list->lock#2 irq_context: 0 (wq_completion)bond205#2 (work_completion)(&(&bond->mcast_work)->work) &base->lock irq_context: 0 (wq_completion)bond178#3 (work_completion)(&(&slave->notify_work)->work) rtnl_mutex &tn->lock irq_context: 0 (wq_completion)bond205#2 (work_completion)(&(&bond->mcast_work)->work) &base->lock &obj_hash[i].lock irq_context: 0 (wq_completion)bond178#3 (work_completion)(&(&slave->notify_work)->work) rtnl_mutex.wait_lock irq_context: 0 (wq_completion)bond205#2 (work_completion)(&(&bond->mcast_work)->work) &rq->__lock irq_context: 0 (wq_completion)bond205#2 (work_completion)(&(&bond->mcast_work)->work) &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)bond204#2 (work_completion)(&(&slave->notify_work)->work) rtnl_mutex net_rwsem irq_context: 0 (wq_completion)bond204#2 (work_completion)(&(&slave->notify_work)->work) rtnl_mutex net_rwsem &list->lock#2 irq_context: 0 (wq_completion)bond204#2 (work_completion)(&(&slave->notify_work)->work) rtnl_mutex &tn->lock irq_context: 0 (wq_completion)bond204#2 (work_completion)(&(&slave->notify_work)->work) rtnl_mutex.wait_lock irq_context: 0 (wq_completion)bond197 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex fs_reclaim irq_context: 0 (wq_completion)bond197 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex fs_reclaim mmu_notifier_invalidate_range_start irq_context: 0 (wq_completion)bond178#3 (work_completion)(&(&slave->notify_work)->work) &p->pi_lock irq_context: 0 (wq_completion)bond197 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &c->lock irq_context: 0 (wq_completion)bond197 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &n->list_lock irq_context: 0 (wq_completion)bond197 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &n->list_lock &c->lock irq_context: 0 (wq_completion)bond197 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &rq->__lock irq_context: 0 (wq_completion)bond197 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)bond197 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex rcu_read_lock &bond->stats_lock/1 irq_context: 0 (wq_completion)bond197 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex rcu_read_lock &ndev->lock irq_context: 0 (wq_completion)bond197 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex nl_table_lock irq_context: 0 (wq_completion)bond197 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &obj_hash[i].lock irq_context: 0 (wq_completion)bond197 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex nl_table_wait.lock irq_context: 0 (wq_completion)bond197 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex net_rwsem irq_context: 0 (wq_completion)bond197 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex net_rwsem &list->lock#2 irq_context: 0 (wq_completion)bond197 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &tn->lock irq_context: 0 (wq_completion)bond197 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &idev->mc_lock irq_context: 0 (wq_completion)bond197 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &idev->mc_lock fs_reclaim irq_context: 0 (wq_completion)bond197 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &idev->mc_lock fs_reclaim &rq->__lock irq_context: 0 (wq_completion)bond197 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &idev->mc_lock fs_reclaim &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)bond197 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &idev->mc_lock fs_reclaim mmu_notifier_invalidate_range_start irq_context: 0 (wq_completion)bond197 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &idev->mc_lock rcu_read_lock &ndev->lock irq_context: 0 (wq_completion)bond197 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &idev->mc_lock rcu_read_lock &dir->lock#2 irq_context: 0 (wq_completion)bond197 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &idev->mc_lock rcu_read_lock &ul->lock#2 irq_context: 0 (wq_completion)bond197 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &idev->mc_lock rcu_read_lock rcu_read_lock &____s->seqcount#10 irq_context: 0 (wq_completion)bond197 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &idev->mc_lock rcu_read_lock rcu_read_lock rcu_read_lock_bh rcu_read_lock rcu_read_lock_bh rcu_read_lock &br->multicast_lock irq_context: 0 (wq_completion)bond197 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &idev->mc_lock rcu_read_lock rcu_read_lock rcu_read_lock_bh rcu_read_lock rcu_read_lock_bh rcu_read_lock &br->multicast_lock &obj_hash[i].lock irq_context: 0 (wq_completion)bond197 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &idev->mc_lock rcu_read_lock rcu_read_lock rcu_read_lock_bh rcu_read_lock rcu_read_lock_bh rcu_read_lock &br->multicast_lock &base->lock irq_context: 0 (wq_completion)bond197 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &idev->mc_lock rcu_read_lock rcu_read_lock rcu_read_lock_bh rcu_read_lock rcu_read_lock_bh rcu_read_lock &br->multicast_lock &base->lock &obj_hash[i].lock irq_context: 0 (wq_completion)bond197 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &idev->mc_lock rcu_read_lock rcu_read_lock rcu_read_lock_bh rcu_read_lock rcu_read_lock_bh rcu_read_lock &data->lock irq_context: 0 (wq_completion)bond197 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &idev->mc_lock rcu_read_lock rcu_read_lock rcu_read_lock_bh rcu_read_lock rcu_read_lock_bh rcu_read_lock &obj_hash[i].lock irq_context: 0 (wq_completion)bond197 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &idev->mc_lock rcu_read_lock rcu_read_lock rcu_node_0 irq_context: 0 (wq_completion)bond197 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &idev->mc_lock rcu_read_lock rcu_read_lock &rq->__lock irq_context: 0 (wq_completion)bond197 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &idev->mc_lock rcu_read_lock rcu_read_lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)bond197 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &idev->mc_lock rcu_read_lock rcu_node_0 irq_context: 0 (wq_completion)bond197 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex.wait_lock irq_context: 0 (wq_completion)bond197 (work_completion)(&(&bond->mcast_work)->work) &p->pi_lock irq_context: 0 (wq_completion)bond197 (work_completion)(&(&bond->mcast_work)->work) &p->pi_lock &rq->__lock irq_context: 0 (wq_completion)bond197 (work_completion)(&(&bond->mcast_work)->work) &p->pi_lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)bond216 irq_context: 0 (wq_completion)bond216 &rq->__lock irq_context: 0 (wq_completion)bond216 &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)bond216 (work_completion)(&(&slave->notify_work)->work) irq_context: 0 (wq_completion)bond216 (work_completion)(&(&slave->notify_work)->work) &obj_hash[i].lock irq_context: 0 (wq_completion)bond216 (work_completion)(&(&slave->notify_work)->work) &base->lock irq_context: 0 (wq_completion)bond216 (work_completion)(&(&slave->notify_work)->work) &base->lock &obj_hash[i].lock irq_context: 0 (wq_completion)bond216 (work_completion)(&(&slave->notify_work)->work) &rq->__lock irq_context: 0 (wq_completion)bond216 (work_completion)(&(&slave->notify_work)->work) &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)bond216 (work_completion)(&(&bond->mcast_work)->work) irq_context: 0 (wq_completion)bond216 (work_completion)(&(&bond->mcast_work)->work) &obj_hash[i].lock irq_context: 0 (wq_completion)bond216 (work_completion)(&(&bond->mcast_work)->work) &base->lock irq_context: 0 (wq_completion)bond216 (work_completion)(&(&bond->mcast_work)->work) &base->lock &obj_hash[i].lock irq_context: 0 (wq_completion)bond61 irq_context: 0 (wq_completion)bond61 (work_completion)(&(&slave->notify_work)->work) irq_context: 0 (wq_completion)bond61 (work_completion)(&(&slave->notify_work)->work) &obj_hash[i].lock irq_context: 0 (wq_completion)bond61 (work_completion)(&(&slave->notify_work)->work) &base->lock irq_context: 0 (wq_completion)bond61 (work_completion)(&(&slave->notify_work)->work) &base->lock &obj_hash[i].lock irq_context: 0 (wq_completion)bond61 (work_completion)(&(&slave->notify_work)->work) &rq->__lock irq_context: 0 (wq_completion)bond61 (work_completion)(&(&slave->notify_work)->work) &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)bond61 (work_completion)(&(&bond->mcast_work)->work) irq_context: 0 (wq_completion)bond61 (work_completion)(&(&bond->mcast_work)->work) &obj_hash[i].lock irq_context: 0 (wq_completion)bond61 (work_completion)(&(&bond->mcast_work)->work) &base->lock irq_context: 0 (wq_completion)bond61 (work_completion)(&(&bond->mcast_work)->work) &base->lock fill_pool_map-wait-type-override pool_lock irq_context: 0 (wq_completion)bond61 (work_completion)(&(&bond->mcast_work)->work) &base->lock &obj_hash[i].lock irq_context: 0 (wq_completion)bond61 (work_completion)(&(&bond->mcast_work)->work) &rq->__lock irq_context: 0 (wq_completion)bond61 (work_completion)(&(&bond->mcast_work)->work) &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)bond189#3 irq_context: 0 (wq_completion)bond189#3 (work_completion)(&(&slave->notify_work)->work) irq_context: 0 (wq_completion)bond189#3 (work_completion)(&(&slave->notify_work)->work) &obj_hash[i].lock irq_context: 0 (wq_completion)bond189#3 (work_completion)(&(&slave->notify_work)->work) &base->lock irq_context: 0 (wq_completion)bond189#3 (work_completion)(&(&slave->notify_work)->work) &base->lock &obj_hash[i].lock irq_context: 0 (wq_completion)bond189#3 (work_completion)(&(&slave->notify_work)->work) &rq->__lock irq_context: 0 (wq_completion)bond189#3 (work_completion)(&(&slave->notify_work)->work) &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)bond189#3 (work_completion)(&(&bond->mcast_work)->work) irq_context: 0 (wq_completion)bond189#3 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex fs_reclaim irq_context: 0 (wq_completion)bond189#3 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex fs_reclaim mmu_notifier_invalidate_range_start irq_context: 0 (wq_completion)bond189#3 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex fs_reclaim mmu_notifier_invalidate_range_start &rq->__lock irq_context: 0 (wq_completion)bond189#3 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex fs_reclaim mmu_notifier_invalidate_range_start &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)bond189#3 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex rcu_read_lock &bond->stats_lock/1 irq_context: 0 (wq_completion)bond189#3 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex rcu_read_lock &ndev->lock irq_context: 0 (wq_completion)bond189#3 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex nl_table_lock irq_context: 0 (wq_completion)bond189#3 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &obj_hash[i].lock irq_context: 0 (wq_completion)bond189#3 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex nl_table_wait.lock irq_context: 0 (wq_completion)bond189#3 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex net_rwsem irq_context: 0 (wq_completion)bond189#3 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex net_rwsem &list->lock#2 irq_context: 0 (wq_completion)bond189#3 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &tn->lock irq_context: 0 (wq_completion)bond189#3 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &idev->mc_lock irq_context: 0 (wq_completion)bond189#3 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &idev->mc_lock &rq->__lock irq_context: 0 (wq_completion)bond189#3 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &idev->mc_lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)bond189#3 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex.wait_lock irq_context: 0 (wq_completion)bond189#3 (work_completion)(&(&bond->mcast_work)->work) &p->pi_lock irq_context: 0 (wq_completion)bond189#3 (work_completion)(&(&bond->mcast_work)->work) &p->pi_lock &rq->__lock irq_context: 0 (wq_completion)bond189#3 (work_completion)(&(&bond->mcast_work)->work) &p->pi_lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)bond189#3 (work_completion)(&(&bond->mcast_work)->work) &rq->__lock irq_context: 0 (wq_completion)bond189#3 (work_completion)(&(&bond->mcast_work)->work) &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)bond61 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex fs_reclaim irq_context: 0 (wq_completion)bond61 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex fs_reclaim mmu_notifier_invalidate_range_start irq_context: 0 (wq_completion)bond61 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &c->lock irq_context: 0 (wq_completion)bond61 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &n->list_lock irq_context: 0 (wq_completion)bond61 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &n->list_lock &c->lock irq_context: 0 (wq_completion)bond61 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex rcu_read_lock &bond->stats_lock/1 irq_context: 0 (wq_completion)bond61 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex rcu_read_lock rcu_node_0 irq_context: 0 (wq_completion)bond61 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex rcu_read_lock &rq->__lock irq_context: 0 (wq_completion)bond61 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex rcu_read_lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)bond61 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex rcu_read_lock &ndev->lock irq_context: 0 (wq_completion)bond61 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex nl_table_lock irq_context: 0 (wq_completion)bond61 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &obj_hash[i].lock irq_context: 0 (wq_completion)bond61 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &rq->__lock irq_context: 0 (wq_completion)bond61 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex nl_table_wait.lock irq_context: 0 (wq_completion)bond61 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex nl_table_wait.lock &p->pi_lock irq_context: 0 (wq_completion)bond61 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex nl_table_wait.lock &p->pi_lock &rq->__lock irq_context: 0 (wq_completion)bond61 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex nl_table_wait.lock &p->pi_lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)bond61 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)bond61 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex net_rwsem irq_context: 0 (wq_completion)bond61 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex net_rwsem &list->lock#2 irq_context: 0 (wq_completion)bond61 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &tn->lock irq_context: 0 (wq_completion)bond61 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &idev->mc_lock irq_context: 0 (wq_completion)bond61 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex.wait_lock irq_context: 0 (wq_completion)bond61 (work_completion)(&(&bond->mcast_work)->work) &p->pi_lock irq_context: 0 (wq_completion)bond61 (work_completion)(&(&bond->mcast_work)->work) &p->pi_lock &rq->__lock irq_context: 0 (wq_completion)bond61 (work_completion)(&(&bond->mcast_work)->work) &p->pi_lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)bond203#2 (work_completion)(&(&bond->mcast_work)->work) &rq->__lock irq_context: 0 (wq_completion)bond156 &rq->__lock irq_context: 0 (wq_completion)bond192#2 (work_completion)(&(&bond->mcast_work)->work) &rq->__lock irq_context: 0 (wq_completion)bond192#2 (work_completion)(&(&bond->mcast_work)->work) &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)bond206#2 irq_context: 0 (wq_completion)bond206#2 (work_completion)(&(&slave->notify_work)->work) irq_context: 0 (wq_completion)bond206#2 (work_completion)(&(&slave->notify_work)->work) &obj_hash[i].lock irq_context: 0 (wq_completion)bond206#2 (work_completion)(&(&slave->notify_work)->work) &base->lock irq_context: 0 (wq_completion)bond206#2 (work_completion)(&(&slave->notify_work)->work) &base->lock &obj_hash[i].lock irq_context: 0 (wq_completion)bond206#2 (work_completion)(&(&slave->notify_work)->work) &rq->__lock irq_context: 0 (wq_completion)bond206#2 (work_completion)(&(&slave->notify_work)->work) &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)bond206#2 (work_completion)(&(&bond->mcast_work)->work) irq_context: 0 (wq_completion)bond206#2 (work_completion)(&(&bond->mcast_work)->work) &obj_hash[i].lock irq_context: 0 (wq_completion)bond206#2 (work_completion)(&(&bond->mcast_work)->work) &base->lock irq_context: 0 (wq_completion)bond206#2 (work_completion)(&(&bond->mcast_work)->work) &base->lock &obj_hash[i].lock irq_context: 0 (wq_completion)bond206#2 (work_completion)(&(&bond->mcast_work)->work) &rq->__lock irq_context: 0 (wq_completion)bond217 irq_context: 0 (wq_completion)bond217 (work_completion)(&(&slave->notify_work)->work) irq_context: 0 (wq_completion)bond217 (work_completion)(&(&slave->notify_work)->work) &obj_hash[i].lock irq_context: 0 (wq_completion)bond217 (work_completion)(&(&slave->notify_work)->work) &base->lock irq_context: 0 (wq_completion)bond217 (work_completion)(&(&slave->notify_work)->work) &base->lock &obj_hash[i].lock irq_context: softirq (&app->join_timer)#2 &base->lock fill_pool_map-wait-type-override pool_lock#2 irq_context: softirq (&app->join_timer)#2 &base->lock fill_pool_map-wait-type-override pool_lock irq_context: 0 (wq_completion)bond217 (work_completion)(&(&bond->mcast_work)->work) irq_context: 0 (wq_completion)bond217 (work_completion)(&(&bond->mcast_work)->work) &obj_hash[i].lock irq_context: 0 (wq_completion)bond217 (work_completion)(&(&bond->mcast_work)->work) &base->lock irq_context: 0 (wq_completion)bond217 (work_completion)(&(&bond->mcast_work)->work) &base->lock &obj_hash[i].lock irq_context: 0 (wq_completion)bond217 (work_completion)(&(&bond->mcast_work)->work) &rq->__lock irq_context: 0 (wq_completion)bond217 (work_completion)(&(&bond->mcast_work)->work) &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)bond62#3 irq_context: 0 (wq_completion)bond62#3 (work_completion)(&(&slave->notify_work)->work) irq_context: 0 (wq_completion)bond62#3 (work_completion)(&(&slave->notify_work)->work) &obj_hash[i].lock irq_context: 0 (wq_completion)bond62#3 (work_completion)(&(&slave->notify_work)->work) &base->lock irq_context: 0 (wq_completion)bond62#3 (work_completion)(&(&slave->notify_work)->work) &base->lock &obj_hash[i].lock irq_context: 0 (wq_completion)bond62#3 (work_completion)(&(&bond->mcast_work)->work) irq_context: 0 (wq_completion)bond62#3 (work_completion)(&(&bond->mcast_work)->work) &rq->__lock irq_context: 0 (wq_completion)bond62#3 (work_completion)(&(&bond->mcast_work)->work) &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)bond62#3 (work_completion)(&(&bond->mcast_work)->work) &obj_hash[i].lock irq_context: 0 (wq_completion)bond62#3 (work_completion)(&(&bond->mcast_work)->work) &base->lock irq_context: 0 (wq_completion)bond62#3 (work_completion)(&(&bond->mcast_work)->work) &base->lock &obj_hash[i].lock irq_context: 0 (wq_completion)bond188#2 &rq->__lock irq_context: 0 (wq_completion)bond188#2 &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)bond190#3 irq_context: 0 (wq_completion)bond190#3 (work_completion)(&(&slave->notify_work)->work) irq_context: 0 (wq_completion)bond190#3 (work_completion)(&(&slave->notify_work)->work) &obj_hash[i].lock irq_context: 0 (wq_completion)bond190#3 (work_completion)(&(&slave->notify_work)->work) &base->lock irq_context: 0 (wq_completion)bond190#3 (work_completion)(&(&slave->notify_work)->work) &base->lock &obj_hash[i].lock irq_context: 0 (wq_completion)bond190#3 (work_completion)(&(&bond->mcast_work)->work) irq_context: 0 (wq_completion)bond190#3 (work_completion)(&(&bond->mcast_work)->work) &obj_hash[i].lock irq_context: 0 (wq_completion)bond190#3 (work_completion)(&(&bond->mcast_work)->work) &base->lock irq_context: 0 (wq_completion)bond190#3 (work_completion)(&(&bond->mcast_work)->work) &base->lock &obj_hash[i].lock irq_context: 0 (wq_completion)bond212 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex fs_reclaim irq_context: 0 (wq_completion)bond212 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex fs_reclaim mmu_notifier_invalidate_range_start irq_context: 0 (wq_completion)bond212 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex rcu_read_lock &bond->stats_lock/1 irq_context: 0 (wq_completion)bond212 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex rcu_read_lock &ndev->lock irq_context: 0 (wq_completion)bond212 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex nl_table_lock irq_context: 0 (wq_completion)bond212 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &obj_hash[i].lock irq_context: 0 (wq_completion)bond212 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex nl_table_wait.lock irq_context: 0 (wq_completion)bond212 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex net_rwsem irq_context: 0 (wq_completion)bond212 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex net_rwsem &list->lock#2 irq_context: 0 (wq_completion)bond212 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &tn->lock irq_context: 0 (wq_completion)bond212 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &idev->mc_lock irq_context: 0 (wq_completion)bond212 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &idev->mc_lock fs_reclaim irq_context: 0 (wq_completion)bond212 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &idev->mc_lock fs_reclaim mmu_notifier_invalidate_range_start irq_context: 0 (wq_completion)bond212 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &idev->mc_lock &c->lock irq_context: 0 (wq_completion)bond212 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &idev->mc_lock rcu_read_lock &ndev->lock irq_context: 0 (wq_completion)bond212 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &idev->mc_lock rcu_read_lock &c->lock irq_context: 0 (wq_completion)bond212 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &idev->mc_lock rcu_read_lock &dir->lock#2 irq_context: 0 (wq_completion)bond212 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &idev->mc_lock rcu_read_lock &ul->lock#2 irq_context: 0 (wq_completion)bond212 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &idev->mc_lock rcu_read_lock rcu_read_lock &____s->seqcount#10 irq_context: 0 (wq_completion)bond212 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &idev->mc_lock rcu_read_lock rcu_read_lock rcu_read_lock_bh rcu_read_lock rcu_read_lock_bh rcu_read_lock &br->multicast_lock irq_context: 0 (wq_completion)bond212 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &idev->mc_lock rcu_read_lock rcu_read_lock rcu_read_lock_bh rcu_read_lock rcu_read_lock_bh rcu_read_lock &br->multicast_lock &obj_hash[i].lock irq_context: 0 (wq_completion)bond212 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &idev->mc_lock rcu_read_lock rcu_read_lock rcu_read_lock_bh rcu_read_lock rcu_read_lock_bh rcu_read_lock &br->multicast_lock &base->lock irq_context: 0 (wq_completion)bond212 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &idev->mc_lock rcu_read_lock rcu_read_lock rcu_read_lock_bh rcu_read_lock rcu_read_lock_bh rcu_read_lock &br->multicast_lock &base->lock &obj_hash[i].lock irq_context: 0 (wq_completion)bond212 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &idev->mc_lock rcu_read_lock rcu_read_lock rcu_read_lock_bh rcu_read_lock rcu_read_lock_bh rcu_read_lock &data->lock irq_context: 0 (wq_completion)bond212 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &idev->mc_lock rcu_read_lock rcu_read_lock rcu_read_lock_bh rcu_read_lock rcu_read_lock_bh rcu_read_lock &obj_hash[i].lock irq_context: 0 (wq_completion)bond212 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex.wait_lock irq_context: 0 (wq_completion)bond212 (work_completion)(&(&bond->mcast_work)->work) &p->pi_lock irq_context: 0 (wq_completion)bond210 (work_completion)(&(&slave->notify_work)->work) rtnl_mutex net_rwsem irq_context: 0 (wq_completion)bond210 (work_completion)(&(&slave->notify_work)->work) rtnl_mutex net_rwsem &list->lock#2 irq_context: 0 (wq_completion)bond210 (work_completion)(&(&slave->notify_work)->work) rtnl_mutex &tn->lock irq_context: 0 (wq_completion)bond210 (work_completion)(&(&slave->notify_work)->work) rtnl_mutex.wait_lock irq_context: 0 (wq_completion)bond210 (work_completion)(&(&slave->notify_work)->work) &p->pi_lock irq_context: 0 (wq_completion)bond60#2 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex fs_reclaim irq_context: 0 (wq_completion)bond60#2 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex fs_reclaim mmu_notifier_invalidate_range_start irq_context: 0 (wq_completion)bond60#2 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &c->lock irq_context: 0 (wq_completion)bond60#2 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex rcu_read_lock &bond->stats_lock/1 irq_context: 0 (wq_completion)bond60#2 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex rcu_read_lock &ndev->lock irq_context: 0 (wq_completion)bond60#2 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex nl_table_lock irq_context: 0 (wq_completion)bond60#2 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &obj_hash[i].lock irq_context: 0 (wq_completion)bond60#2 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex nl_table_wait.lock irq_context: 0 (wq_completion)bond60#2 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex net_rwsem irq_context: 0 (wq_completion)bond60#2 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex net_rwsem &list->lock#2 irq_context: 0 (wq_completion)bond60#2 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &tn->lock irq_context: 0 (wq_completion)bond60#2 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &idev->mc_lock irq_context: 0 (wq_completion)bond60#2 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &idev->mc_lock fs_reclaim irq_context: 0 (wq_completion)bond60#2 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &idev->mc_lock fs_reclaim mmu_notifier_invalidate_range_start irq_context: 0 (wq_completion)bond60#2 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &idev->mc_lock &c->lock irq_context: 0 (wq_completion)bond60#2 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &idev->mc_lock rcu_read_lock &ndev->lock irq_context: 0 (wq_completion)bond60#2 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &idev->mc_lock rcu_read_lock &dir->lock#2 irq_context: 0 (wq_completion)bond60#2 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &idev->mc_lock rcu_read_lock &ul->lock#2 irq_context: 0 (wq_completion)bond60#2 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &idev->mc_lock rcu_read_lock rcu_node_0 irq_context: 0 (wq_completion)bond60#2 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &idev->mc_lock rcu_read_lock &rq->__lock irq_context: 0 (wq_completion)bond60#2 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &idev->mc_lock rcu_read_lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)bond120#3 &rq->__lock irq_context: 0 (wq_completion)bond120#3 &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)bond60#2 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &idev->mc_lock rcu_read_lock rcu_read_lock &____s->seqcount#12 irq_context: 0 (wq_completion)bond60#2 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &idev->mc_lock rcu_read_lock rcu_read_lock &____s->seqcount#16 irq_context: 0 (wq_completion)bond60#2 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &idev->mc_lock rcu_read_lock &obj_hash[i].lock irq_context: 0 (wq_completion)bond60#2 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &idev->mc_lock rcu_read_lock &data->lock irq_context: 0 (wq_completion)bond60#2 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex.wait_lock irq_context: 0 (wq_completion)bond60#2 (work_completion)(&(&bond->mcast_work)->work) &p->pi_lock irq_context: 0 (wq_completion)bond60#2 (work_completion)(&(&bond->mcast_work)->work) &p->pi_lock &rq->__lock irq_context: 0 (wq_completion)bond60#2 (work_completion)(&(&bond->mcast_work)->work) &p->pi_lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)bond203#2 (work_completion)(&(&bond->mcast_work)->work) &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 &f->f_pos_lock &type->i_mutex_dir_key#3 fs_reclaim mmu_notifier_invalidate_range_start &cfs_rq->removed.lock irq_context: 0 &f->f_pos_lock &type->i_mutex_dir_key#3 fs_reclaim mmu_notifier_invalidate_range_start &obj_hash[i].lock irq_context: 0 (wq_completion)bond149 &rq->__lock irq_context: 0 (wq_completion)bond149 &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)bond207#2 irq_context: 0 (wq_completion)bond207#2 &rq->__lock irq_context: 0 (wq_completion)bond207#2 &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)bond207#2 (work_completion)(&(&slave->notify_work)->work) irq_context: 0 (wq_completion)bond207#2 (work_completion)(&(&slave->notify_work)->work) &obj_hash[i].lock irq_context: 0 (wq_completion)bond207#2 (work_completion)(&(&slave->notify_work)->work) &base->lock irq_context: 0 (wq_completion)bond207#2 (work_completion)(&(&slave->notify_work)->work) &base->lock &obj_hash[i].lock irq_context: 0 (wq_completion)bond218 irq_context: 0 (wq_completion)bond218 (work_completion)(&(&slave->notify_work)->work) irq_context: 0 (wq_completion)bond218 (work_completion)(&(&slave->notify_work)->work) &obj_hash[i].lock irq_context: 0 (wq_completion)bond218 (work_completion)(&(&slave->notify_work)->work) &base->lock irq_context: 0 (wq_completion)bond218 (work_completion)(&(&slave->notify_work)->work) &base->lock &obj_hash[i].lock irq_context: 0 (wq_completion)bond207#2 (work_completion)(&(&bond->mcast_work)->work) irq_context: 0 (wq_completion)bond207#2 (work_completion)(&(&bond->mcast_work)->work) &obj_hash[i].lock irq_context: 0 (wq_completion)bond207#2 (work_completion)(&(&bond->mcast_work)->work) &base->lock irq_context: 0 (wq_completion)bond207#2 (work_completion)(&(&bond->mcast_work)->work) &base->lock &obj_hash[i].lock irq_context: 0 (wq_completion)bond207#2 (work_completion)(&(&bond->mcast_work)->work) &rq->__lock irq_context: 0 (wq_completion)bond207#2 (work_completion)(&(&bond->mcast_work)->work) &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)bond218 (work_completion)(&(&bond->mcast_work)->work) irq_context: 0 (wq_completion)bond218 (work_completion)(&(&bond->mcast_work)->work) &obj_hash[i].lock irq_context: 0 (wq_completion)bond218 (work_completion)(&(&bond->mcast_work)->work) &base->lock irq_context: 0 (wq_completion)bond218 (work_completion)(&(&bond->mcast_work)->work) &base->lock &obj_hash[i].lock irq_context: 0 (wq_completion)bond218 (work_completion)(&(&bond->mcast_work)->work) &rq->__lock irq_context: 0 (wq_completion)bond218 (work_completion)(&(&bond->mcast_work)->work) &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)events (linkwatch_work).work rtnl_mutex &tb->tb6_lock nl_table_wait.lock &p->pi_lock irq_context: 0 (wq_completion)events (linkwatch_work).work rtnl_mutex &tb->tb6_lock nl_table_wait.lock &p->pi_lock &cfs_rq->removed.lock irq_context: 0 (wq_completion)events (linkwatch_work).work rtnl_mutex &tb->tb6_lock nl_table_wait.lock &p->pi_lock &rq->__lock irq_context: 0 (wq_completion)events (linkwatch_work).work rtnl_mutex &tb->tb6_lock nl_table_wait.lock &p->pi_lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)bond191#2 &rq->__lock irq_context: 0 (wq_completion)bond191#2 &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)bond191#3 irq_context: 0 (wq_completion)bond191#3 (work_completion)(&(&slave->notify_work)->work) irq_context: 0 (wq_completion)bond191#3 (work_completion)(&(&slave->notify_work)->work) &obj_hash[i].lock irq_context: 0 (wq_completion)bond191#3 (work_completion)(&(&slave->notify_work)->work) &base->lock irq_context: 0 (wq_completion)bond191#3 (work_completion)(&(&slave->notify_work)->work) &base->lock &obj_hash[i].lock irq_context: 0 (wq_completion)bond191#3 (work_completion)(&(&bond->mcast_work)->work) irq_context: 0 (wq_completion)bond191#3 (work_completion)(&(&bond->mcast_work)->work) &obj_hash[i].lock irq_context: 0 (wq_completion)bond191#3 (work_completion)(&(&bond->mcast_work)->work) &base->lock irq_context: 0 (wq_completion)bond191#3 (work_completion)(&(&bond->mcast_work)->work) &base->lock &obj_hash[i].lock irq_context: 0 (wq_completion)bond191#3 (work_completion)(&(&bond->mcast_work)->work) &rq->__lock irq_context: 0 (wq_completion)bond191#3 (work_completion)(&(&bond->mcast_work)->work) &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)bond195#2 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex fs_reclaim irq_context: 0 (wq_completion)bond195#2 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex fs_reclaim &rq->__lock irq_context: 0 (wq_completion)bond195#2 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex fs_reclaim &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)bond195#2 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex fs_reclaim mmu_notifier_invalidate_range_start irq_context: 0 (wq_completion)bond195#2 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &c->lock irq_context: 0 (wq_completion)bond195#2 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex rcu_read_lock &bond->stats_lock/1 irq_context: 0 (wq_completion)bond195#2 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex rcu_read_lock &ndev->lock irq_context: 0 (wq_completion)bond195#2 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex nl_table_lock irq_context: 0 (wq_completion)bond195#2 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &obj_hash[i].lock irq_context: 0 (wq_completion)bond195#2 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex nl_table_wait.lock irq_context: 0 (wq_completion)bond195#2 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex net_rwsem irq_context: 0 (wq_completion)bond195#2 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex net_rwsem &list->lock#2 irq_context: 0 (wq_completion)bond195#2 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex net_rwsem &rq->__lock irq_context: 0 (wq_completion)bond180#3 (work_completion)(&(&slave->notify_work)->work) &rq->__lock irq_context: 0 (wq_completion)bond195#2 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex net_rwsem &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)bond180#3 (work_completion)(&(&slave->notify_work)->work) &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)bond195#2 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &tn->lock irq_context: 0 (wq_completion)bond195#2 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &idev->mc_lock irq_context: 0 (wq_completion)bond195#2 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &idev->mc_lock fs_reclaim irq_context: 0 (wq_completion)bond195#2 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &idev->mc_lock fs_reclaim mmu_notifier_invalidate_range_start irq_context: 0 (wq_completion)bond195#2 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &idev->mc_lock fs_reclaim mmu_notifier_invalidate_range_start &rq->__lock irq_context: 0 (wq_completion)bond195#2 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &idev->mc_lock fs_reclaim mmu_notifier_invalidate_range_start &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)bond195#2 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &idev->mc_lock &c->lock irq_context: 0 (wq_completion)bond195#2 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &idev->mc_lock rcu_read_lock &ndev->lock irq_context: 0 (wq_completion)bond195#2 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &idev->mc_lock rcu_read_lock &dir->lock#2 irq_context: 0 (wq_completion)bond195#2 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &idev->mc_lock rcu_read_lock &ul->lock#2 irq_context: 0 (wq_completion)bond195#2 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &idev->mc_lock rcu_read_lock rcu_read_lock &____s->seqcount#10 irq_context: 0 (wq_completion)bond195#2 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &idev->mc_lock rcu_read_lock rcu_read_lock rcu_read_lock_bh rcu_read_lock rcu_read_lock_bh rcu_read_lock &br->multicast_lock irq_context: 0 (wq_completion)bond195#2 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &idev->mc_lock rcu_read_lock rcu_read_lock rcu_read_lock_bh rcu_read_lock rcu_read_lock_bh rcu_read_lock &br->multicast_lock &obj_hash[i].lock irq_context: 0 (wq_completion)bond195#2 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &idev->mc_lock rcu_read_lock rcu_read_lock rcu_read_lock_bh rcu_read_lock rcu_read_lock_bh rcu_read_lock &br->multicast_lock &base->lock irq_context: 0 (wq_completion)bond195#2 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &idev->mc_lock rcu_read_lock rcu_read_lock rcu_read_lock_bh rcu_read_lock rcu_read_lock_bh rcu_read_lock &br->multicast_lock &base->lock &obj_hash[i].lock irq_context: 0 (wq_completion)bond195#2 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &idev->mc_lock rcu_read_lock rcu_read_lock rcu_read_lock_bh rcu_read_lock rcu_read_lock_bh rcu_read_lock &data->lock irq_context: 0 (wq_completion)bond195#2 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &idev->mc_lock rcu_read_lock rcu_read_lock rcu_read_lock_bh rcu_read_lock rcu_read_lock_bh rcu_read_lock &obj_hash[i].lock irq_context: 0 (wq_completion)bond195#2 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &idev->mc_lock rcu_read_lock rcu_read_lock rcu_node_0 irq_context: 0 (wq_completion)bond195#2 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &idev->mc_lock rcu_read_lock rcu_read_lock &rq->__lock irq_context: 0 (wq_completion)bond195#2 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &idev->mc_lock rcu_read_lock rcu_node_0 irq_context: 0 (wq_completion)bond195#2 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex.wait_lock irq_context: 0 (wq_completion)bond195#2 (work_completion)(&(&bond->mcast_work)->work) &p->pi_lock irq_context: 0 (wq_completion)bond195#2 (work_completion)(&(&bond->mcast_work)->work) &p->pi_lock &rq->__lock irq_context: 0 (wq_completion)bond195#2 (work_completion)(&(&bond->mcast_work)->work) &p->pi_lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)bond195#2 (work_completion)(&(&bond->mcast_work)->work) &rq->__lock irq_context: 0 (wq_completion)bond195#2 (work_completion)(&(&bond->mcast_work)->work) &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)bond142#2 &rq->__lock irq_context: 0 (wq_completion)bond142#2 &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)bond180#3 (work_completion)(&(&bond->mcast_work)->work) &rq->__lock irq_context: 0 (wq_completion)bond180#3 (work_completion)(&(&bond->mcast_work)->work) &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)bond198#2 (work_completion)(&(&bond->mcast_work)->work) &rq->__lock irq_context: 0 (wq_completion)bond198#2 (work_completion)(&(&bond->mcast_work)->work) &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)bond207#2 (work_completion)(&(&slave->notify_work)->work) rtnl_mutex net_rwsem irq_context: 0 (wq_completion)bond207#2 (work_completion)(&(&slave->notify_work)->work) rtnl_mutex net_rwsem &list->lock#2 irq_context: 0 (wq_completion)bond207#2 (work_completion)(&(&slave->notify_work)->work) rtnl_mutex &tn->lock irq_context: 0 (wq_completion)bond207#2 (work_completion)(&(&slave->notify_work)->work) rtnl_mutex.wait_lock irq_context: 0 (wq_completion)bond207#2 (work_completion)(&(&slave->notify_work)->work) &p->pi_lock irq_context: 0 (wq_completion)bond203#2 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex fs_reclaim irq_context: 0 (wq_completion)bond203#2 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex fs_reclaim mmu_notifier_invalidate_range_start irq_context: 0 (wq_completion)bond203#2 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &c->lock irq_context: 0 (wq_completion)bond203#2 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex rcu_read_lock &bond->stats_lock/1 irq_context: 0 (wq_completion)bond203#2 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex rcu_read_lock &ndev->lock irq_context: 0 (wq_completion)bond203#2 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex nl_table_lock irq_context: 0 (wq_completion)bond203#2 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &obj_hash[i].lock irq_context: 0 (wq_completion)bond203#2 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex nl_table_wait.lock irq_context: 0 (wq_completion)bond203#2 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex net_rwsem irq_context: 0 (wq_completion)bond203#2 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex net_rwsem &list->lock#2 irq_context: 0 (wq_completion)bond203#2 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &tn->lock irq_context: 0 (wq_completion)bond203#2 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &idev->mc_lock irq_context: 0 (wq_completion)bond203#2 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &idev->mc_lock fs_reclaim irq_context: 0 (wq_completion)bond203#2 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &idev->mc_lock fs_reclaim mmu_notifier_invalidate_range_start irq_context: 0 (wq_completion)bond203#2 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &idev->mc_lock &c->lock irq_context: 0 (wq_completion)bond203#2 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &idev->mc_lock rcu_read_lock &ndev->lock irq_context: 0 (wq_completion)bond203#2 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &idev->mc_lock rcu_read_lock &c->lock irq_context: 0 (wq_completion)bond203#2 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &idev->mc_lock rcu_read_lock &dir->lock#2 irq_context: 0 (wq_completion)bond203#2 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &idev->mc_lock rcu_read_lock &ul->lock#2 irq_context: 0 (wq_completion)bond203#2 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &idev->mc_lock rcu_read_lock rcu_read_lock &____s->seqcount#10 irq_context: 0 (wq_completion)bond203#2 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &idev->mc_lock rcu_read_lock rcu_read_lock rcu_read_lock_bh rcu_read_lock rcu_read_lock_bh rcu_read_lock &br->multicast_lock irq_context: 0 (wq_completion)bond203#2 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &idev->mc_lock rcu_read_lock rcu_read_lock rcu_read_lock_bh rcu_read_lock rcu_read_lock_bh rcu_read_lock &br->multicast_lock &obj_hash[i].lock irq_context: 0 (wq_completion)bond203#2 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &idev->mc_lock rcu_read_lock rcu_read_lock rcu_read_lock_bh rcu_read_lock rcu_read_lock_bh rcu_read_lock &br->multicast_lock &base->lock irq_context: 0 (wq_completion)bond203#2 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &idev->mc_lock rcu_read_lock rcu_read_lock rcu_read_lock_bh rcu_read_lock rcu_read_lock_bh rcu_read_lock &data->lock irq_context: 0 (wq_completion)bond203#2 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &idev->mc_lock rcu_read_lock rcu_read_lock rcu_read_lock_bh rcu_read_lock rcu_read_lock_bh rcu_read_lock &obj_hash[i].lock irq_context: 0 (wq_completion)bond203#2 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex.wait_lock irq_context: 0 (wq_completion)bond203#2 (work_completion)(&(&bond->mcast_work)->work) &p->pi_lock irq_context: 0 (wq_completion)bond171#2 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex fs_reclaim irq_context: 0 (wq_completion)bond171#2 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex fs_reclaim mmu_notifier_invalidate_range_start irq_context: 0 (wq_completion)bond171#2 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &c->lock irq_context: 0 (wq_completion)bond171#2 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &n->list_lock irq_context: 0 (wq_completion)bond171#2 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &n->list_lock &c->lock irq_context: 0 (wq_completion)bond171#2 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &rq->__lock irq_context: 0 (wq_completion)bond171#2 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)bond171#2 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex rcu_read_lock &bond->stats_lock/1 irq_context: 0 (wq_completion)bond171#2 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex rcu_read_lock &ndev->lock irq_context: 0 (wq_completion)bond171#2 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex nl_table_lock irq_context: 0 (wq_completion)bond171#2 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &obj_hash[i].lock irq_context: 0 (wq_completion)bond171#2 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex nl_table_wait.lock irq_context: 0 (wq_completion)bond171#2 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex net_rwsem irq_context: 0 (wq_completion)bond171#2 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex net_rwsem &list->lock#2 irq_context: 0 (wq_completion)bond171#2 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &tn->lock irq_context: 0 (wq_completion)bond171#2 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &idev->mc_lock irq_context: 0 (wq_completion)bond171#2 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &idev->mc_lock &rq->__lock irq_context: 0 (wq_completion)bond171#2 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &idev->mc_lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)bond171#2 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &idev->mc_lock fs_reclaim irq_context: 0 (wq_completion)bond171#2 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &idev->mc_lock fs_reclaim mmu_notifier_invalidate_range_start irq_context: 0 (wq_completion)bond171#2 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &idev->mc_lock &c->lock irq_context: 0 (wq_completion)bond171#2 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &idev->mc_lock rcu_read_lock &ndev->lock irq_context: 0 (wq_completion)bond171#2 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &idev->mc_lock rcu_read_lock &dir->lock#2 irq_context: 0 (wq_completion)bond171#2 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &idev->mc_lock rcu_read_lock &ul->lock#2 irq_context: 0 (wq_completion)bond171#2 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &idev->mc_lock rcu_read_lock rcu_read_lock &obj_hash[i].lock irq_context: 0 (wq_completion)bond171#2 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &idev->mc_lock rcu_read_lock rcu_read_lock &dir->lock#2 irq_context: 0 (wq_completion)bond171#2 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &idev->mc_lock rcu_read_lock rcu_read_lock &tbl->lock irq_context: 0 (wq_completion)bond171#2 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &idev->mc_lock rcu_read_lock rcu_read_lock &n->lock irq_context: 0 (wq_completion)bond171#2 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &idev->mc_lock rcu_read_lock rcu_read_lock rcu_node_0 irq_context: 0 (wq_completion)bond171#2 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &idev->mc_lock rcu_read_lock rcu_read_lock &rq->__lock irq_context: 0 (wq_completion)bond171#2 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &idev->mc_lock rcu_read_lock rcu_read_lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)bond171#2 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &idev->mc_lock rcu_read_lock rcu_read_lock &____s->seqcount#9 irq_context: 0 (wq_completion)bond171#2 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &idev->mc_lock rcu_read_lock rcu_read_lock rcu_read_lock_bh rcu_read_lock rcu_read_lock_bh rcu_read_lock &br->multicast_lock irq_context: 0 (wq_completion)bond171#2 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &idev->mc_lock rcu_read_lock rcu_read_lock rcu_read_lock_bh rcu_read_lock rcu_read_lock_bh rcu_read_lock &br->multicast_lock &obj_hash[i].lock irq_context: 0 (wq_completion)bond171#2 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &idev->mc_lock rcu_read_lock rcu_read_lock rcu_read_lock_bh rcu_read_lock rcu_read_lock_bh rcu_read_lock &br->multicast_lock &base->lock irq_context: 0 (wq_completion)bond171#2 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &idev->mc_lock rcu_read_lock rcu_read_lock rcu_read_lock_bh rcu_read_lock rcu_read_lock_bh rcu_read_lock &br->multicast_lock &base->lock &obj_hash[i].lock irq_context: 0 (wq_completion)bond171#2 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &idev->mc_lock rcu_read_lock rcu_read_lock rcu_read_lock_bh rcu_read_lock rcu_read_lock_bh rcu_read_lock &data->lock irq_context: 0 (wq_completion)bond171#2 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &idev->mc_lock rcu_read_lock rcu_read_lock rcu_read_lock_bh rcu_read_lock rcu_read_lock_bh rcu_read_lock &obj_hash[i].lock irq_context: 0 (wq_completion)bond171#2 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &idev->mc_lock rcu_read_lock rcu_node_0 irq_context: 0 (wq_completion)bond171#2 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex.wait_lock irq_context: 0 (wq_completion)bond171#2 (work_completion)(&(&bond->mcast_work)->work) &p->pi_lock irq_context: 0 (wq_completion)bond219 irq_context: 0 (wq_completion)bond219 (work_completion)(&(&slave->notify_work)->work) irq_context: 0 (wq_completion)bond219 (work_completion)(&(&slave->notify_work)->work) &obj_hash[i].lock irq_context: 0 (wq_completion)bond219 (work_completion)(&(&slave->notify_work)->work) &base->lock irq_context: 0 (wq_completion)bond219 (work_completion)(&(&slave->notify_work)->work) &base->lock &obj_hash[i].lock irq_context: 0 (wq_completion)bond219 (work_completion)(&(&slave->notify_work)->work) &rq->__lock irq_context: 0 (wq_completion)bond219 (work_completion)(&(&slave->notify_work)->work) &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)bond219 (work_completion)(&(&bond->mcast_work)->work) irq_context: 0 (wq_completion)bond219 (work_completion)(&(&bond->mcast_work)->work) &obj_hash[i].lock irq_context: 0 (wq_completion)bond219 (work_completion)(&(&bond->mcast_work)->work) &base->lock irq_context: 0 (wq_completion)bond219 (work_completion)(&(&bond->mcast_work)->work) &base->lock &obj_hash[i].lock irq_context: 0 (wq_completion)bond219 (work_completion)(&(&bond->mcast_work)->work) &rq->__lock irq_context: 0 (wq_completion)bond219 (work_completion)(&(&bond->mcast_work)->work) &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)bond214 (work_completion)(&(&slave->notify_work)->work) &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)bond63 irq_context: 0 (wq_completion)bond63 (work_completion)(&(&slave->notify_work)->work) irq_context: 0 (wq_completion)bond63 (work_completion)(&(&slave->notify_work)->work) &obj_hash[i].lock irq_context: 0 (wq_completion)bond63 (work_completion)(&(&slave->notify_work)->work) &base->lock irq_context: 0 (wq_completion)bond63 (work_completion)(&(&slave->notify_work)->work) &base->lock &obj_hash[i].lock irq_context: 0 (wq_completion)bond63 (work_completion)(&(&slave->notify_work)->work) &rq->__lock irq_context: 0 (wq_completion)bond63 (work_completion)(&(&slave->notify_work)->work) &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)bond63 (work_completion)(&(&bond->mcast_work)->work) irq_context: 0 (wq_completion)bond63 (work_completion)(&(&bond->mcast_work)->work) &obj_hash[i].lock irq_context: 0 (wq_completion)bond63 (work_completion)(&(&bond->mcast_work)->work) &base->lock irq_context: 0 (wq_completion)bond63 (work_completion)(&(&bond->mcast_work)->work) &base->lock &obj_hash[i].lock irq_context: 0 (wq_completion)bond63 (work_completion)(&(&bond->mcast_work)->work) &rq->__lock irq_context: 0 (wq_completion)bond63 (work_completion)(&(&bond->mcast_work)->work) &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)bond208#2 irq_context: 0 (wq_completion)bond208#2 (work_completion)(&(&slave->notify_work)->work) irq_context: 0 (wq_completion)bond208#2 (work_completion)(&(&slave->notify_work)->work) &obj_hash[i].lock irq_context: 0 (wq_completion)bond208#2 (work_completion)(&(&slave->notify_work)->work) &base->lock irq_context: 0 (wq_completion)bond208#2 (work_completion)(&(&slave->notify_work)->work) &base->lock &obj_hash[i].lock irq_context: 0 (wq_completion)bond208#2 (work_completion)(&(&bond->mcast_work)->work) irq_context: 0 (wq_completion)bond208#2 (work_completion)(&(&bond->mcast_work)->work) &obj_hash[i].lock irq_context: 0 (wq_completion)bond208#2 (work_completion)(&(&bond->mcast_work)->work) &base->lock irq_context: 0 (wq_completion)bond208#2 (work_completion)(&(&bond->mcast_work)->work) &base->lock &obj_hash[i].lock irq_context: 0 (wq_completion)bond165#3 (work_completion)(&(&slave->notify_work)->work) fill_pool_map-wait-type-override &c->lock irq_context: 0 (wq_completion)bond165#3 (work_completion)(&(&slave->notify_work)->work) fill_pool_map-wait-type-override &____s->seqcount#2 irq_context: 0 (wq_completion)bond165#3 (work_completion)(&(&slave->notify_work)->work) fill_pool_map-wait-type-override &____s->seqcount irq_context: 0 (wq_completion)bond165#3 (work_completion)(&(&slave->notify_work)->work) fill_pool_map-wait-type-override pool_lock irq_context: 0 (wq_completion)bond214 &rq->__lock irq_context: 0 (wq_completion)bond214 &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 sb_writers#4 &type->i_mutex_dir_key#3/1 jbd2_handle &ei->i_data_sem &ei->i_es_lock batched_entropy_u8.lock irq_context: 0 sb_writers#4 &type->i_mutex_dir_key#3/1 jbd2_handle &ei->i_data_sem &ei->i_es_lock kfence_freelist_lock irq_context: 0 (wq_completion)bond192#3 irq_context: 0 (wq_completion)bond192#3 (work_completion)(&(&slave->notify_work)->work) irq_context: 0 (wq_completion)bond192#3 (work_completion)(&(&slave->notify_work)->work) &obj_hash[i].lock irq_context: 0 (wq_completion)bond192#3 (work_completion)(&(&slave->notify_work)->work) &base->lock irq_context: 0 (wq_completion)bond192#3 (work_completion)(&(&slave->notify_work)->work) &base->lock &obj_hash[i].lock irq_context: 0 (wq_completion)bond192#3 (work_completion)(&(&slave->notify_work)->work) &rq->__lock irq_context: 0 (wq_completion)bond192#3 (work_completion)(&(&slave->notify_work)->work) &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)bond192#3 (work_completion)(&(&bond->mcast_work)->work) irq_context: 0 (wq_completion)bond192#3 (work_completion)(&(&bond->mcast_work)->work) &obj_hash[i].lock irq_context: 0 (wq_completion)bond192#3 (work_completion)(&(&bond->mcast_work)->work) &base->lock irq_context: 0 (wq_completion)bond192#3 (work_completion)(&(&bond->mcast_work)->work) &base->lock &obj_hash[i].lock irq_context: 0 (wq_completion)bond192#3 (work_completion)(&(&bond->mcast_work)->work) &rq->__lock irq_context: 0 rtnl_mutex &idev->mc_lock &cfs_rq->removed.lock irq_context: 0 (wq_completion)bond220 irq_context: 0 (wq_completion)bond220 (work_completion)(&(&slave->notify_work)->work) irq_context: 0 (wq_completion)bond220 (work_completion)(&(&slave->notify_work)->work) &obj_hash[i].lock irq_context: 0 (wq_completion)bond220 (work_completion)(&(&slave->notify_work)->work) &base->lock irq_context: 0 (wq_completion)bond220 (work_completion)(&(&slave->notify_work)->work) &base->lock &obj_hash[i].lock irq_context: 0 (wq_completion)bond220 (work_completion)(&(&bond->mcast_work)->work) irq_context: 0 (wq_completion)bond220 (work_completion)(&(&bond->mcast_work)->work) &obj_hash[i].lock irq_context: 0 (wq_completion)bond220 (work_completion)(&(&bond->mcast_work)->work) &base->lock irq_context: 0 (wq_completion)bond220 (work_completion)(&(&bond->mcast_work)->work) &base->lock &obj_hash[i].lock irq_context: 0 (wq_completion)bond220 (work_completion)(&(&bond->mcast_work)->work) &rq->__lock irq_context: 0 (wq_completion)bond220 (work_completion)(&(&bond->mcast_work)->work) &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)bond192#3 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex fs_reclaim irq_context: 0 (wq_completion)bond192#3 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex fs_reclaim mmu_notifier_invalidate_range_start irq_context: 0 (wq_completion)bond192#3 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &c->lock irq_context: 0 (wq_completion)bond192#3 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &n->list_lock irq_context: 0 (wq_completion)bond192#3 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &n->list_lock &c->lock irq_context: 0 (wq_completion)bond192#3 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &rq->__lock irq_context: 0 (wq_completion)bond192#3 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)bond192#3 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex rcu_read_lock &bond->stats_lock/1 irq_context: 0 (wq_completion)bond192#3 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex rcu_read_lock &ndev->lock irq_context: 0 (wq_completion)bond192#3 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex nl_table_lock irq_context: 0 (wq_completion)bond192#3 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &obj_hash[i].lock irq_context: 0 (wq_completion)bond192#3 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex nl_table_wait.lock irq_context: 0 (wq_completion)bond192#3 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex net_rwsem irq_context: 0 (wq_completion)bond192#3 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex net_rwsem &list->lock#2 irq_context: 0 (wq_completion)bond192#3 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &tn->lock irq_context: 0 (wq_completion)bond192#3 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &idev->mc_lock irq_context: 0 (wq_completion)bond192#3 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex.wait_lock irq_context: 0 (wq_completion)bond192#3 (work_completion)(&(&bond->mcast_work)->work) &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)bond192#3 (work_completion)(&(&bond->mcast_work)->work) &p->pi_lock irq_context: 0 (wq_completion)bond192#3 (work_completion)(&(&bond->mcast_work)->work) &p->pi_lock &rq->__lock irq_context: 0 (wq_completion)bond192#3 (work_completion)(&(&bond->mcast_work)->work) &p->pi_lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)bond209#2 irq_context: 0 (wq_completion)bond209#2 (work_completion)(&(&slave->notify_work)->work) irq_context: 0 (wq_completion)bond209#2 (work_completion)(&(&slave->notify_work)->work) &obj_hash[i].lock irq_context: 0 (wq_completion)bond209#2 (work_completion)(&(&slave->notify_work)->work) &base->lock irq_context: 0 (wq_completion)bond209#2 (work_completion)(&(&slave->notify_work)->work) &base->lock &obj_hash[i].lock irq_context: 0 (wq_completion)bond209#2 (work_completion)(&(&slave->notify_work)->work) &rq->__lock irq_context: 0 (wq_completion)bond209#2 (work_completion)(&(&slave->notify_work)->work) &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)bond209#2 (work_completion)(&(&bond->mcast_work)->work) irq_context: 0 (wq_completion)bond209#2 (work_completion)(&(&bond->mcast_work)->work) &obj_hash[i].lock irq_context: 0 (wq_completion)bond209#2 (work_completion)(&(&bond->mcast_work)->work) &base->lock irq_context: 0 (wq_completion)bond209#2 (work_completion)(&(&bond->mcast_work)->work) &base->lock &obj_hash[i].lock irq_context: 0 (wq_completion)bond186#2 &rq->__lock irq_context: 0 (wq_completion)bond186#2 &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)bond143#2 (work_completion)(&(&bond->mcast_work)->work) &base->lock fill_pool_map-wait-type-override &c->lock irq_context: 0 (wq_completion)bond143#2 (work_completion)(&(&bond->mcast_work)->work) &base->lock fill_pool_map-wait-type-override &____s->seqcount#2 irq_context: 0 (wq_completion)bond143#2 (work_completion)(&(&bond->mcast_work)->work) &base->lock fill_pool_map-wait-type-override &____s->seqcount irq_context: 0 (wq_completion)bond143#2 (work_completion)(&(&bond->mcast_work)->work) &base->lock fill_pool_map-wait-type-override pool_lock irq_context: 0 (wq_completion)bond64 irq_context: 0 (wq_completion)bond64 &rq->__lock irq_context: 0 (wq_completion)bond64 &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)bond64 (work_completion)(&(&slave->notify_work)->work) irq_context: 0 (wq_completion)bond64 (work_completion)(&(&slave->notify_work)->work) &obj_hash[i].lock irq_context: 0 (wq_completion)bond64 (work_completion)(&(&slave->notify_work)->work) &base->lock irq_context: 0 (wq_completion)bond64 (work_completion)(&(&slave->notify_work)->work) &base->lock &obj_hash[i].lock irq_context: 0 (wq_completion)bond143#2 (work_completion)(&(&bond->mcast_work)->work) &cfs_rq->removed.lock irq_context: 0 (wq_completion)bond64 (work_completion)(&(&bond->mcast_work)->work) irq_context: 0 (wq_completion)bond64 (work_completion)(&(&bond->mcast_work)->work) &obj_hash[i].lock irq_context: 0 (wq_completion)bond64 (work_completion)(&(&bond->mcast_work)->work) &base->lock irq_context: 0 (wq_completion)bond64 (work_completion)(&(&bond->mcast_work)->work) &base->lock &obj_hash[i].lock irq_context: 0 (wq_completion)bond200#2 (work_completion)(&(&slave->notify_work)->work) &rq->__lock irq_context: 0 (wq_completion)bond200#2 (work_completion)(&(&slave->notify_work)->work) &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)bond193#3 irq_context: 0 (wq_completion)bond193#3 (work_completion)(&(&slave->notify_work)->work) irq_context: 0 (wq_completion)bond193#3 (work_completion)(&(&slave->notify_work)->work) &obj_hash[i].lock irq_context: 0 (wq_completion)bond193#3 (work_completion)(&(&slave->notify_work)->work) &base->lock irq_context: 0 (wq_completion)bond193#3 (work_completion)(&(&slave->notify_work)->work) &base->lock &obj_hash[i].lock irq_context: 0 (wq_completion)bond193#3 (work_completion)(&(&bond->mcast_work)->work) irq_context: 0 (wq_completion)bond193#3 (work_completion)(&(&bond->mcast_work)->work) &obj_hash[i].lock irq_context: 0 (wq_completion)bond193#3 (work_completion)(&(&bond->mcast_work)->work) &base->lock irq_context: 0 (wq_completion)bond193#3 (work_completion)(&(&bond->mcast_work)->work) &base->lock &obj_hash[i].lock irq_context: 0 (wq_completion)bond193#3 (work_completion)(&(&bond->mcast_work)->work) &rq->__lock irq_context: 0 (wq_completion)bond193#3 (work_completion)(&(&bond->mcast_work)->work) &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)bond221 irq_context: 0 (wq_completion)bond221 &rq->__lock irq_context: 0 (wq_completion)bond221 &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)bond221 (work_completion)(&(&slave->notify_work)->work) irq_context: 0 (wq_completion)bond221 (work_completion)(&(&slave->notify_work)->work) &obj_hash[i].lock irq_context: 0 (wq_completion)bond221 (work_completion)(&(&slave->notify_work)->work) &base->lock irq_context: 0 (wq_completion)bond221 (work_completion)(&(&slave->notify_work)->work) &base->lock &obj_hash[i].lock irq_context: 0 (wq_completion)bond221 (work_completion)(&(&slave->notify_work)->work) &rq->__lock irq_context: 0 (wq_completion)bond221 (work_completion)(&(&slave->notify_work)->work) &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)bond221 (work_completion)(&(&bond->mcast_work)->work) irq_context: 0 (wq_completion)bond221 (work_completion)(&(&bond->mcast_work)->work) &obj_hash[i].lock irq_context: 0 (wq_completion)bond221 (work_completion)(&(&bond->mcast_work)->work) &base->lock irq_context: 0 (wq_completion)bond221 (work_completion)(&(&bond->mcast_work)->work) &base->lock &obj_hash[i].lock irq_context: 0 (wq_completion)bond221 (work_completion)(&(&bond->mcast_work)->work) &rq->__lock irq_context: 0 (wq_completion)bond221 (work_completion)(&(&bond->mcast_work)->work) &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)bond221 (work_completion)(&(&slave->notify_work)->work) rtnl_mutex net_rwsem irq_context: 0 (wq_completion)bond221 (work_completion)(&(&slave->notify_work)->work) rtnl_mutex net_rwsem &list->lock#2 irq_context: 0 (wq_completion)bond221 (work_completion)(&(&slave->notify_work)->work) rtnl_mutex &tn->lock irq_context: 0 (wq_completion)bond221 (work_completion)(&(&slave->notify_work)->work) rtnl_mutex.wait_lock irq_context: 0 (wq_completion)bond221 (work_completion)(&(&slave->notify_work)->work) &p->pi_lock irq_context: 0 (wq_completion)bond204 &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)bond210#2 irq_context: 0 (wq_completion)bond210#2 (work_completion)(&(&slave->notify_work)->work) irq_context: 0 (wq_completion)bond210#2 (work_completion)(&(&slave->notify_work)->work) &obj_hash[i].lock irq_context: 0 (wq_completion)bond210#2 (work_completion)(&(&slave->notify_work)->work) &base->lock irq_context: 0 (wq_completion)bond210#2 (work_completion)(&(&slave->notify_work)->work) &base->lock &obj_hash[i].lock irq_context: 0 (wq_completion)bond210#2 (work_completion)(&(&slave->notify_work)->work) &rq->__lock irq_context: 0 (wq_completion)bond210#2 (work_completion)(&(&slave->notify_work)->work) &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)bond62#3 (work_completion)(&(&slave->notify_work)->work) &rq->__lock irq_context: 0 (wq_completion)bond62#3 (work_completion)(&(&slave->notify_work)->work) &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)bond210#2 (work_completion)(&(&bond->mcast_work)->work) irq_context: 0 (wq_completion)bond210#2 (work_completion)(&(&bond->mcast_work)->work) &obj_hash[i].lock irq_context: 0 (wq_completion)bond210#2 (work_completion)(&(&bond->mcast_work)->work) &base->lock irq_context: 0 (wq_completion)bond210#2 (work_completion)(&(&bond->mcast_work)->work) &base->lock &obj_hash[i].lock irq_context: 0 (wq_completion)bond210#2 (work_completion)(&(&bond->mcast_work)->work) &rq->__lock irq_context: 0 (wq_completion)bond210#2 (work_completion)(&(&bond->mcast_work)->work) &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)bond191#3 (work_completion)(&(&slave->notify_work)->work) &rq->__lock irq_context: 0 (wq_completion)bond191#3 (work_completion)(&(&slave->notify_work)->work) &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)bond164#3 (work_completion)(&(&slave->notify_work)->work) fill_pool_map-wait-type-override pool_lock irq_context: 0 (wq_completion)bond65 irq_context: 0 (wq_completion)bond65 (work_completion)(&(&slave->notify_work)->work) irq_context: 0 (wq_completion)bond65 (work_completion)(&(&slave->notify_work)->work) &obj_hash[i].lock irq_context: 0 (wq_completion)bond65 (work_completion)(&(&slave->notify_work)->work) &base->lock irq_context: 0 (wq_completion)bond65 (work_completion)(&(&slave->notify_work)->work) &base->lock &obj_hash[i].lock irq_context: 0 (wq_completion)bond65 (work_completion)(&(&slave->notify_work)->work) &rq->__lock irq_context: 0 (wq_completion)bond65 (work_completion)(&(&slave->notify_work)->work) &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)bond174#2 &rq->__lock irq_context: 0 (wq_completion)bond174#2 &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)bond65 (work_completion)(&(&bond->mcast_work)->work) irq_context: 0 (wq_completion)bond65 (work_completion)(&(&bond->mcast_work)->work) &obj_hash[i].lock irq_context: 0 (wq_completion)bond65 (work_completion)(&(&bond->mcast_work)->work) &base->lock irq_context: 0 (wq_completion)bond65 (work_completion)(&(&bond->mcast_work)->work) &base->lock &obj_hash[i].lock irq_context: 0 (wq_completion)bond220 (work_completion)(&(&slave->notify_work)->work) &rq->__lock irq_context: 0 (wq_completion)bond220 (work_completion)(&(&slave->notify_work)->work) &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)bond218 &rq->__lock irq_context: 0 (wq_completion)bond218 &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)bond194#3 irq_context: 0 (wq_completion)bond194#3 (work_completion)(&(&slave->notify_work)->work) irq_context: 0 (wq_completion)bond194#3 (work_completion)(&(&slave->notify_work)->work) &obj_hash[i].lock irq_context: 0 (wq_completion)bond194#3 (work_completion)(&(&slave->notify_work)->work) &base->lock irq_context: 0 (wq_completion)bond194#3 (work_completion)(&(&slave->notify_work)->work) &base->lock &obj_hash[i].lock irq_context: 0 (wq_completion)bond194#3 (work_completion)(&(&slave->notify_work)->work) &rq->__lock irq_context: 0 (wq_completion)bond194#3 (work_completion)(&(&slave->notify_work)->work) &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)bond176#3 (work_completion)(&(&slave->notify_work)->work) &rq->__lock irq_context: 0 (wq_completion)bond194#3 (work_completion)(&(&bond->mcast_work)->work) irq_context: 0 (wq_completion)bond194#3 (work_completion)(&(&bond->mcast_work)->work) &obj_hash[i].lock irq_context: 0 (wq_completion)bond194#3 (work_completion)(&(&bond->mcast_work)->work) &base->lock irq_context: 0 (wq_completion)bond194#3 (work_completion)(&(&bond->mcast_work)->work) &base->lock &obj_hash[i].lock irq_context: 0 (wq_completion)bond217 (work_completion)(&(&slave->notify_work)->work) &rq->__lock irq_context: 0 (wq_completion)bond217 (work_completion)(&(&slave->notify_work)->work) &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)bond222 irq_context: 0 (wq_completion)bond222 (work_completion)(&(&slave->notify_work)->work) irq_context: 0 (wq_completion)bond222 (work_completion)(&(&slave->notify_work)->work) &obj_hash[i].lock irq_context: 0 (wq_completion)bond222 (work_completion)(&(&slave->notify_work)->work) &base->lock irq_context: 0 (wq_completion)bond222 (work_completion)(&(&slave->notify_work)->work) &base->lock &obj_hash[i].lock irq_context: 0 (wq_completion)bond222 (work_completion)(&(&bond->mcast_work)->work) irq_context: 0 (wq_completion)bond222 (work_completion)(&(&bond->mcast_work)->work) &obj_hash[i].lock irq_context: 0 (wq_completion)bond222 (work_completion)(&(&bond->mcast_work)->work) &base->lock irq_context: 0 (wq_completion)bond222 (work_completion)(&(&bond->mcast_work)->work) &base->lock &obj_hash[i].lock irq_context: 0 (wq_completion)bond222 (work_completion)(&(&bond->mcast_work)->work) &rq->__lock irq_context: 0 (wq_completion)bond222 (work_completion)(&(&bond->mcast_work)->work) &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)bond211#2 irq_context: 0 (wq_completion)bond211#2 &rq->__lock irq_context: 0 (wq_completion)bond211#2 &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)bond211#2 (work_completion)(&(&slave->notify_work)->work) irq_context: 0 (wq_completion)bond211#2 (work_completion)(&(&slave->notify_work)->work) &obj_hash[i].lock irq_context: 0 (wq_completion)bond211#2 (work_completion)(&(&slave->notify_work)->work) &base->lock irq_context: 0 (wq_completion)bond211#2 (work_completion)(&(&slave->notify_work)->work) &base->lock &obj_hash[i].lock irq_context: 0 (wq_completion)bond172#3 (work_completion)(&(&slave->notify_work)->work) &rq->__lock irq_context: 0 (wq_completion)bond172#3 (work_completion)(&(&slave->notify_work)->work) &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)bond211#2 (work_completion)(&(&bond->mcast_work)->work) irq_context: 0 (wq_completion)bond211#2 (work_completion)(&(&bond->mcast_work)->work) &obj_hash[i].lock irq_context: 0 (wq_completion)bond211#2 (work_completion)(&(&bond->mcast_work)->work) &base->lock irq_context: 0 (wq_completion)bond211#2 (work_completion)(&(&bond->mcast_work)->work) &base->lock &obj_hash[i].lock irq_context: 0 (wq_completion)bond211#2 (work_completion)(&(&bond->mcast_work)->work) &rq->__lock irq_context: 0 (wq_completion)bond211#2 (work_completion)(&(&bond->mcast_work)->work) &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)bond147 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex fs_reclaim irq_context: 0 (wq_completion)bond147 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex fs_reclaim mmu_notifier_invalidate_range_start irq_context: 0 (wq_completion)bond147 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex fs_reclaim mmu_notifier_invalidate_range_start &rq->__lock irq_context: 0 (wq_completion)bond147 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex fs_reclaim mmu_notifier_invalidate_range_start &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)bond147 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &c->lock irq_context: 0 (wq_completion)bond190#3 (work_completion)(&(&slave->notify_work)->work) &rq->__lock irq_context: 0 (wq_completion)bond147 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex rcu_read_lock &bond->stats_lock/1 irq_context: 0 (wq_completion)bond147 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex rcu_read_lock &ndev->lock irq_context: 0 (wq_completion)bond147 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex nl_table_lock irq_context: 0 (wq_completion)bond147 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &obj_hash[i].lock irq_context: 0 (wq_completion)bond147 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex nl_table_wait.lock irq_context: 0 (wq_completion)bond193#3 &rq->__lock irq_context: 0 (wq_completion)bond147 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex net_rwsem irq_context: 0 (wq_completion)bond147 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex net_rwsem &list->lock#2 irq_context: 0 (wq_completion)bond147 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex net_rwsem &rq->__lock irq_context: 0 (wq_completion)bond147 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex net_rwsem &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)bond147 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &tn->lock irq_context: 0 (wq_completion)bond147 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &idev->mc_lock irq_context: 0 (wq_completion)bond147 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &idev->mc_lock fs_reclaim irq_context: 0 (wq_completion)bond147 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &idev->mc_lock fs_reclaim mmu_notifier_invalidate_range_start irq_context: 0 (wq_completion)bond147 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &idev->mc_lock rcu_read_lock &ndev->lock irq_context: 0 (wq_completion)bond147 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &idev->mc_lock rcu_read_lock &dir->lock#2 irq_context: 0 (wq_completion)bond147 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &idev->mc_lock rcu_read_lock &ul->lock#2 irq_context: 0 (wq_completion)bond147 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &idev->mc_lock rcu_read_lock rcu_read_lock &obj_hash[i].lock irq_context: 0 (wq_completion)bond147 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &idev->mc_lock rcu_read_lock rcu_read_lock rcu_node_0 irq_context: 0 (wq_completion)bond147 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &idev->mc_lock rcu_read_lock rcu_read_lock &rq->__lock irq_context: 0 (wq_completion)bond147 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &idev->mc_lock rcu_read_lock rcu_read_lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)bond147 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &idev->mc_lock rcu_read_lock rcu_read_lock &dir->lock#2 irq_context: 0 (wq_completion)bond147 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &idev->mc_lock rcu_read_lock rcu_read_lock &tbl->lock irq_context: 0 (wq_completion)bond147 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &idev->mc_lock rcu_read_lock rcu_read_lock &n->lock irq_context: 0 (wq_completion)bond147 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &idev->mc_lock rcu_read_lock rcu_read_lock &____s->seqcount#9 irq_context: 0 (wq_completion)bond147 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &idev->mc_lock rcu_read_lock rcu_read_lock rcu_read_lock_bh rcu_read_lock rcu_read_lock_bh rcu_read_lock &br->multicast_lock irq_context: 0 (wq_completion)bond147 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &idev->mc_lock rcu_read_lock rcu_read_lock rcu_read_lock_bh rcu_read_lock rcu_read_lock_bh rcu_read_lock &br->multicast_lock &obj_hash[i].lock irq_context: 0 (wq_completion)bond147 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &idev->mc_lock rcu_read_lock rcu_read_lock rcu_read_lock_bh rcu_read_lock rcu_read_lock_bh rcu_read_lock &br->multicast_lock &base->lock irq_context: 0 (wq_completion)bond147 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &idev->mc_lock rcu_read_lock rcu_read_lock rcu_read_lock_bh rcu_read_lock rcu_read_lock_bh rcu_read_lock &br->multicast_lock &base->lock &obj_hash[i].lock irq_context: 0 (wq_completion)bond147 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &idev->mc_lock rcu_read_lock rcu_read_lock rcu_read_lock_bh rcu_read_lock rcu_read_lock_bh rcu_read_lock &data->lock irq_context: 0 (wq_completion)bond147 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &idev->mc_lock rcu_read_lock rcu_read_lock rcu_read_lock_bh rcu_read_lock rcu_read_lock_bh rcu_read_lock &obj_hash[i].lock irq_context: 0 (wq_completion)bond147 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &idev->mc_lock rcu_read_lock rcu_node_0 irq_context: 0 (wq_completion)bond147 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex.wait_lock irq_context: 0 (wq_completion)bond147 (work_completion)(&(&bond->mcast_work)->work) &p->pi_lock irq_context: 0 (wq_completion)bond147 (work_completion)(&(&bond->mcast_work)->work) &p->pi_lock &rq->__lock irq_context: 0 (wq_completion)bond147 (work_completion)(&(&bond->mcast_work)->work) &p->pi_lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)bond144#2 &rq->__lock irq_context: 0 (wq_completion)bond193 (work_completion)(&(&bond->mcast_work)->work) &rq->__lock irq_context: 0 (wq_completion)bond193 (work_completion)(&(&bond->mcast_work)->work) &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)bond174#2 (work_completion)(&(&slave->notify_work)->work) &rq->__lock irq_context: 0 (wq_completion)bond174#2 (work_completion)(&(&slave->notify_work)->work) &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)bond167 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex fs_reclaim irq_context: 0 (wq_completion)bond167 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex fs_reclaim mmu_notifier_invalidate_range_start irq_context: 0 (wq_completion)bond167 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &c->lock irq_context: 0 (wq_completion)bond167 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &rq->__lock irq_context: 0 (wq_completion)bond167 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)bond167 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex rcu_read_lock &bond->stats_lock/1 irq_context: 0 (wq_completion)bond167 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex rcu_read_lock &ndev->lock irq_context: 0 (wq_completion)bond167 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex nl_table_lock irq_context: 0 (wq_completion)bond167 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &obj_hash[i].lock irq_context: 0 (wq_completion)bond167 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex nl_table_wait.lock irq_context: 0 (wq_completion)bond167 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex net_rwsem irq_context: 0 (wq_completion)bond167 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex net_rwsem &list->lock#2 irq_context: 0 (wq_completion)bond167 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &tn->lock irq_context: 0 (wq_completion)bond167 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &idev->mc_lock irq_context: 0 (wq_completion)bond167 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &idev->mc_lock fs_reclaim irq_context: 0 (wq_completion)bond167 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &idev->mc_lock fs_reclaim &rq->__lock irq_context: 0 (wq_completion)bond167 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &idev->mc_lock fs_reclaim &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)bond167 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &idev->mc_lock fs_reclaim mmu_notifier_invalidate_range_start irq_context: 0 (wq_completion)bond167 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &idev->mc_lock &c->lock irq_context: 0 (wq_completion)bond167 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &idev->mc_lock &n->list_lock irq_context: 0 (wq_completion)bond167 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &idev->mc_lock &n->list_lock &c->lock irq_context: 0 (wq_completion)bond167 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &idev->mc_lock rcu_read_lock &ndev->lock irq_context: 0 (wq_completion)bond167 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &idev->mc_lock rcu_read_lock &c->lock irq_context: 0 (wq_completion)bond167 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &idev->mc_lock rcu_read_lock &dir->lock#2 irq_context: 0 (wq_completion)bond167 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &idev->mc_lock rcu_read_lock rcu_node_0 irq_context: 0 (wq_completion)bond167 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &idev->mc_lock rcu_read_lock &rq->__lock irq_context: 0 (wq_completion)bond167 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &idev->mc_lock rcu_read_lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)bond167 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &idev->mc_lock rcu_read_lock &ul->lock#2 irq_context: 0 (wq_completion)bond167 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &idev->mc_lock rcu_read_lock rcu_read_lock &c->lock irq_context: 0 (wq_completion)bond167 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &idev->mc_lock rcu_read_lock rcu_read_lock &obj_hash[i].lock irq_context: 0 (wq_completion)bond167 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &idev->mc_lock rcu_read_lock rcu_read_lock &dir->lock#2 irq_context: 0 (wq_completion)bond167 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &idev->mc_lock rcu_read_lock rcu_read_lock &tbl->lock irq_context: 0 (wq_completion)bond167 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &idev->mc_lock rcu_read_lock rcu_read_lock &n->lock irq_context: 0 (wq_completion)bond167 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &idev->mc_lock rcu_read_lock rcu_read_lock &____s->seqcount#9 irq_context: 0 (wq_completion)bond167 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &idev->mc_lock rcu_read_lock rcu_read_lock rcu_read_lock_bh rcu_read_lock rcu_read_lock_bh rcu_read_lock &br->multicast_lock irq_context: 0 (wq_completion)bond167 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &idev->mc_lock rcu_read_lock rcu_read_lock rcu_read_lock_bh rcu_read_lock rcu_read_lock_bh rcu_read_lock &br->multicast_lock &obj_hash[i].lock irq_context: 0 (wq_completion)bond167 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &idev->mc_lock rcu_read_lock rcu_read_lock rcu_read_lock_bh rcu_read_lock rcu_read_lock_bh rcu_read_lock &br->multicast_lock &base->lock irq_context: 0 (wq_completion)bond167 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &idev->mc_lock rcu_read_lock rcu_read_lock rcu_read_lock_bh rcu_read_lock rcu_read_lock_bh rcu_read_lock &br->multicast_lock &base->lock &obj_hash[i].lock irq_context: 0 (wq_completion)bond167 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &idev->mc_lock rcu_read_lock rcu_read_lock rcu_read_lock_bh rcu_read_lock rcu_read_lock_bh rcu_read_lock &data->lock irq_context: 0 (wq_completion)bond167 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &idev->mc_lock rcu_read_lock rcu_read_lock rcu_read_lock_bh rcu_read_lock rcu_read_lock_bh rcu_read_lock &obj_hash[i].lock irq_context: 0 (wq_completion)bond167 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &idev->mc_lock rcu_read_lock rcu_read_lock &rq->__lock irq_context: 0 (wq_completion)bond167 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &idev->mc_lock rcu_read_lock rcu_read_lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)bond167 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex.wait_lock irq_context: 0 (wq_completion)bond167 (work_completion)(&(&bond->mcast_work)->work) &p->pi_lock irq_context: 0 (wq_completion)bond167 (work_completion)(&(&bond->mcast_work)->work) &p->pi_lock &rq->__lock irq_context: 0 (wq_completion)bond167 (work_completion)(&(&bond->mcast_work)->work) &p->pi_lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)bond201 &rq->__lock irq_context: 0 (wq_completion)bond201 &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)bond208#2 &rq->__lock irq_context: 0 (wq_completion)bond208#2 &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)bond188#3 (work_completion)(&(&bond->mcast_work)->work) &rq->__lock irq_context: 0 (wq_completion)bond188#3 (work_completion)(&(&bond->mcast_work)->work) &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)bond66 irq_context: 0 (wq_completion)bond66 &rq->__lock irq_context: 0 (wq_completion)bond66 &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)bond66 (work_completion)(&(&slave->notify_work)->work) irq_context: 0 (wq_completion)bond66 (work_completion)(&(&slave->notify_work)->work) &obj_hash[i].lock irq_context: 0 (wq_completion)bond66 (work_completion)(&(&slave->notify_work)->work) &base->lock irq_context: 0 (wq_completion)bond66 (work_completion)(&(&slave->notify_work)->work) &base->lock &obj_hash[i].lock irq_context: 0 (wq_completion)bond65 (work_completion)(&(&bond->mcast_work)->work) &rq->__lock irq_context: 0 (wq_completion)bond65 (work_completion)(&(&bond->mcast_work)->work) &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)bond170 &rq->__lock irq_context: 0 (wq_completion)bond170 &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)bond170 &cfs_rq->removed.lock irq_context: 0 (wq_completion)bond66 (work_completion)(&(&bond->mcast_work)->work) irq_context: 0 (wq_completion)bond66 (work_completion)(&(&bond->mcast_work)->work) &obj_hash[i].lock irq_context: 0 (wq_completion)bond66 (work_completion)(&(&bond->mcast_work)->work) &base->lock irq_context: 0 (wq_completion)bond66 (work_completion)(&(&bond->mcast_work)->work) &base->lock &obj_hash[i].lock irq_context: 0 (wq_completion)bond66 (work_completion)(&(&bond->mcast_work)->work) &rq->__lock irq_context: 0 (wq_completion)bond66 (work_completion)(&(&bond->mcast_work)->work) &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)bond208#2 (work_completion)(&(&slave->notify_work)->work) &rq->__lock irq_context: 0 (wq_completion)bond208#2 (work_completion)(&(&slave->notify_work)->work) &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)bond182#3 (work_completion)(&(&slave->notify_work)->work) rtnl_mutex net_rwsem irq_context: 0 (wq_completion)bond182#3 (work_completion)(&(&slave->notify_work)->work) rtnl_mutex net_rwsem &list->lock#2 irq_context: 0 (wq_completion)bond182#3 (work_completion)(&(&slave->notify_work)->work) rtnl_mutex &tn->lock irq_context: 0 (wq_completion)bond182#3 (work_completion)(&(&slave->notify_work)->work) rtnl_mutex &rq->__lock irq_context: 0 (wq_completion)bond182#3 (work_completion)(&(&slave->notify_work)->work) rtnl_mutex rcu_read_lock rcu_node_0 irq_context: 0 (wq_completion)bond182#3 (work_completion)(&(&slave->notify_work)->work) rtnl_mutex rcu_read_lock &rq->__lock irq_context: 0 (wq_completion)bond182#3 (work_completion)(&(&slave->notify_work)->work) rtnl_mutex rcu_read_lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)bond182#3 (work_completion)(&(&slave->notify_work)->work) rtnl_mutex.wait_lock irq_context: 0 (wq_completion)bond182#3 (work_completion)(&(&slave->notify_work)->work) &p->pi_lock irq_context: 0 (wq_completion)bond182#3 (work_completion)(&(&slave->notify_work)->work) &p->pi_lock &cfs_rq->removed.lock irq_context: 0 (wq_completion)bond182#3 (work_completion)(&(&slave->notify_work)->work) &p->pi_lock &rq->__lock irq_context: 0 (wq_completion)bond182#3 (work_completion)(&(&slave->notify_work)->work) &p->pi_lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)bond204 (work_completion)(&(&slave->notify_work)->work) &rq->__lock irq_context: 0 (wq_completion)bond204 (work_completion)(&(&slave->notify_work)->work) &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)bond212 (work_completion)(&(&slave->notify_work)->work) &rq->__lock irq_context: 0 (wq_completion)bond212 (work_completion)(&(&slave->notify_work)->work) &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)bond195#3 irq_context: 0 (wq_completion)bond195#3 &rq->__lock irq_context: 0 (wq_completion)bond195#3 &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)bond195#3 (work_completion)(&(&slave->notify_work)->work) irq_context: 0 (wq_completion)bond195#3 (work_completion)(&(&slave->notify_work)->work) &rq->__lock irq_context: 0 (wq_completion)bond195#3 (work_completion)(&(&slave->notify_work)->work) &obj_hash[i].lock irq_context: 0 (wq_completion)bond195#3 (work_completion)(&(&slave->notify_work)->work) &base->lock irq_context: 0 (wq_completion)bond195#3 (work_completion)(&(&slave->notify_work)->work) &base->lock &obj_hash[i].lock irq_context: 0 (wq_completion)bond195#3 (work_completion)(&(&slave->notify_work)->work) &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)bond194#3 (work_completion)(&(&bond->mcast_work)->work) &rq->__lock irq_context: 0 (wq_completion)bond194#3 (work_completion)(&(&bond->mcast_work)->work) &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)bond195#3 (work_completion)(&(&bond->mcast_work)->work) irq_context: 0 (wq_completion)bond195#3 (work_completion)(&(&bond->mcast_work)->work) &rq->__lock irq_context: 0 (wq_completion)bond195#3 (work_completion)(&(&bond->mcast_work)->work) &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)bond195#3 (work_completion)(&(&bond->mcast_work)->work) &obj_hash[i].lock irq_context: 0 (wq_completion)bond195#3 (work_completion)(&(&bond->mcast_work)->work) &base->lock irq_context: 0 (wq_completion)bond195#3 (work_completion)(&(&bond->mcast_work)->work) &base->lock &obj_hash[i].lock irq_context: 0 (wq_completion)bond66 (work_completion)(&(&slave->notify_work)->work) &rq->__lock irq_context: 0 (wq_completion)bond66 (work_completion)(&(&slave->notify_work)->work) &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)bond223 irq_context: 0 (wq_completion)bond223 (work_completion)(&(&slave->notify_work)->work) irq_context: 0 (wq_completion)bond223 (work_completion)(&(&slave->notify_work)->work) &obj_hash[i].lock irq_context: 0 (wq_completion)bond223 (work_completion)(&(&slave->notify_work)->work) &base->lock irq_context: 0 (wq_completion)bond223 (work_completion)(&(&slave->notify_work)->work) &base->lock &obj_hash[i].lock irq_context: 0 (wq_completion)bond223 (work_completion)(&(&slave->notify_work)->work) &rq->__lock irq_context: 0 (wq_completion)bond223 (work_completion)(&(&slave->notify_work)->work) &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)bond223 (work_completion)(&(&bond->mcast_work)->work) irq_context: 0 (wq_completion)bond223 (work_completion)(&(&bond->mcast_work)->work) &obj_hash[i].lock irq_context: 0 (wq_completion)bond223 (work_completion)(&(&bond->mcast_work)->work) &base->lock irq_context: 0 (wq_completion)bond223 (work_completion)(&(&bond->mcast_work)->work) &base->lock &obj_hash[i].lock irq_context: 0 (wq_completion)bond223 (work_completion)(&(&bond->mcast_work)->work) &rq->__lock irq_context: 0 (wq_completion)bond223 (work_completion)(&(&bond->mcast_work)->work) &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)bond127 &rq->__lock irq_context: 0 (wq_completion)bond127 &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)bond212#2 irq_context: 0 (wq_completion)bond212#2 &rq->__lock irq_context: 0 (wq_completion)bond212#2 &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)bond64 (work_completion)(&(&bond->mcast_work)->work) &rq->__lock irq_context: 0 (wq_completion)bond212#2 (work_completion)(&(&slave->notify_work)->work) irq_context: 0 (wq_completion)bond212#2 (work_completion)(&(&slave->notify_work)->work) &rq->__lock irq_context: 0 (wq_completion)bond212#2 (work_completion)(&(&slave->notify_work)->work) &obj_hash[i].lock irq_context: 0 (wq_completion)bond212#2 (work_completion)(&(&slave->notify_work)->work) &base->lock irq_context: 0 (wq_completion)bond212#2 (work_completion)(&(&slave->notify_work)->work) &base->lock &obj_hash[i].lock irq_context: 0 (wq_completion)bond212#2 (work_completion)(&(&slave->notify_work)->work) &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)bond64 (work_completion)(&(&slave->notify_work)->work) rtnl_mutex net_rwsem irq_context: 0 (wq_completion)bond64 (work_completion)(&(&slave->notify_work)->work) rtnl_mutex net_rwsem &list->lock#2 irq_context: 0 (wq_completion)bond64 (work_completion)(&(&slave->notify_work)->work) rtnl_mutex net_rwsem &rq->__lock irq_context: 0 (wq_completion)bond64 (work_completion)(&(&slave->notify_work)->work) rtnl_mutex &tn->lock irq_context: 0 (wq_completion)bond64 (work_completion)(&(&slave->notify_work)->work) rtnl_mutex.wait_lock irq_context: 0 (wq_completion)bond64 (work_completion)(&(&slave->notify_work)->work) &rq->__lock irq_context: 0 (wq_completion)bond64 (work_completion)(&(&slave->notify_work)->work) &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)bond64 (work_completion)(&(&slave->notify_work)->work) &p->pi_lock irq_context: 0 (wq_completion)bond168#3 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex fs_reclaim irq_context: 0 (wq_completion)bond168#3 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex fs_reclaim mmu_notifier_invalidate_range_start irq_context: 0 (wq_completion)bond168#3 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &c->lock irq_context: 0 (wq_completion)bond168#3 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &n->list_lock irq_context: 0 (wq_completion)bond168#3 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &n->list_lock &c->lock irq_context: 0 (wq_completion)bond168#3 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &rq->__lock irq_context: 0 (wq_completion)bond168#3 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)bond168#3 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex rcu_read_lock &bond->stats_lock/1 irq_context: 0 (wq_completion)bond168#3 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex rcu_read_lock &ndev->lock irq_context: 0 (wq_completion)bond168#3 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex nl_table_lock irq_context: 0 (wq_completion)bond168#3 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &obj_hash[i].lock irq_context: 0 (wq_completion)bond168#3 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex nl_table_wait.lock irq_context: 0 (wq_completion)bond168#3 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex nl_table_wait.lock &p->pi_lock irq_context: 0 (wq_completion)bond168#3 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex nl_table_wait.lock &p->pi_lock &rq->__lock irq_context: 0 (wq_completion)bond168#3 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex nl_table_wait.lock &p->pi_lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)bond168#3 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex net_rwsem irq_context: 0 (wq_completion)bond168#3 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex net_rwsem &list->lock#2 irq_context: 0 (wq_completion)bond168#3 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &tn->lock irq_context: 0 (wq_completion)bond168#3 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &idev->mc_lock irq_context: 0 (wq_completion)bond168#3 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &idev->mc_lock fs_reclaim irq_context: 0 (wq_completion)bond168#3 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &idev->mc_lock fs_reclaim mmu_notifier_invalidate_range_start irq_context: 0 (wq_completion)bond168#3 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &idev->mc_lock &c->lock irq_context: 0 (wq_completion)bond168#3 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &idev->mc_lock rcu_read_lock &ndev->lock irq_context: 0 (wq_completion)bond168#3 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &idev->mc_lock rcu_read_lock &c->lock irq_context: 0 (wq_completion)bond168#3 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &idev->mc_lock rcu_read_lock &____s->seqcount#2 irq_context: 0 (wq_completion)bond168#3 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &idev->mc_lock rcu_read_lock &____s->seqcount irq_context: 0 (wq_completion)bond168#3 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &idev->mc_lock rcu_read_lock rcu_node_0 irq_context: 0 (wq_completion)bond168#3 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &idev->mc_lock rcu_read_lock &rq->__lock irq_context: 0 (wq_completion)bond168#3 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &idev->mc_lock rcu_read_lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)bond168#3 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &idev->mc_lock rcu_read_lock &dir->lock#2 irq_context: 0 (wq_completion)bond168#3 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &idev->mc_lock rcu_read_lock &ul->lock#2 irq_context: 0 (wq_completion)bond168#3 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &idev->mc_lock rcu_read_lock rcu_read_lock &____s->seqcount#10 irq_context: 0 (wq_completion)bond168#3 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &idev->mc_lock rcu_read_lock rcu_read_lock rcu_read_lock_bh rcu_read_lock rcu_read_lock_bh rcu_read_lock &br->multicast_lock irq_context: 0 (wq_completion)bond168#3 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &idev->mc_lock rcu_read_lock rcu_read_lock rcu_read_lock_bh rcu_read_lock rcu_read_lock_bh rcu_read_lock &br->multicast_lock &obj_hash[i].lock irq_context: 0 (wq_completion)bond168#3 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &idev->mc_lock rcu_read_lock rcu_read_lock rcu_read_lock_bh rcu_read_lock rcu_read_lock_bh rcu_read_lock &br->multicast_lock &base->lock irq_context: 0 (wq_completion)bond168#3 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &idev->mc_lock rcu_read_lock rcu_read_lock rcu_read_lock_bh rcu_read_lock rcu_read_lock_bh rcu_read_lock &br->multicast_lock &base->lock &obj_hash[i].lock irq_context: 0 (wq_completion)bond168#3 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &idev->mc_lock rcu_read_lock rcu_read_lock rcu_read_lock_bh rcu_read_lock rcu_read_lock_bh rcu_read_lock &data->lock irq_context: 0 (wq_completion)bond168#3 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &idev->mc_lock rcu_read_lock rcu_read_lock rcu_read_lock_bh rcu_read_lock rcu_read_lock_bh rcu_read_lock &obj_hash[i].lock irq_context: 0 (wq_completion)bond168#3 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &idev->mc_lock rcu_read_lock rcu_read_lock &rq->__lock irq_context: 0 (wq_completion)bond168#3 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &idev->mc_lock rcu_read_lock rcu_read_lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)bond168#3 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex.wait_lock irq_context: 0 (wq_completion)bond168#3 (work_completion)(&(&bond->mcast_work)->work) &p->pi_lock irq_context: 0 (wq_completion)bond168#3 (work_completion)(&(&bond->mcast_work)->work) &p->pi_lock &rq->__lock irq_context: 0 (wq_completion)bond168#3 (work_completion)(&(&bond->mcast_work)->work) &p->pi_lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)bond153#3 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex fs_reclaim irq_context: 0 (wq_completion)bond153#3 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex fs_reclaim mmu_notifier_invalidate_range_start irq_context: 0 (wq_completion)bond153#3 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &c->lock irq_context: 0 (wq_completion)bond153#3 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &____s->seqcount#2 irq_context: 0 (wq_completion)bond153#3 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &____s->seqcount irq_context: 0 (wq_completion)bond153#3 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex rcu_read_lock &bond->stats_lock/1 irq_context: 0 (wq_completion)bond153#3 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex rcu_read_lock &ndev->lock irq_context: 0 (wq_completion)bond153#3 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex rcu_read_lock rcu_node_0 irq_context: 0 (wq_completion)bond153#3 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex rcu_read_lock &rq->__lock irq_context: 0 (wq_completion)bond153#3 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex rcu_read_lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)bond153#3 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex nl_table_lock irq_context: 0 (wq_completion)bond153#3 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &obj_hash[i].lock irq_context: 0 (wq_completion)bond153#3 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex nl_table_wait.lock irq_context: 0 (wq_completion)bond153#3 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex net_rwsem irq_context: 0 (wq_completion)bond153#3 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex net_rwsem &list->lock#2 irq_context: 0 (wq_completion)bond153#3 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &tn->lock irq_context: 0 (wq_completion)bond153#3 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &idev->mc_lock irq_context: 0 (wq_completion)bond153#3 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &idev->mc_lock fs_reclaim irq_context: 0 (wq_completion)bond153#3 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &idev->mc_lock fs_reclaim mmu_notifier_invalidate_range_start irq_context: 0 (wq_completion)bond153#3 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &idev->mc_lock fs_reclaim mmu_notifier_invalidate_range_start &rq->__lock irq_context: 0 (wq_completion)bond153#3 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &idev->mc_lock &c->lock irq_context: 0 (wq_completion)bond153#3 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &idev->mc_lock rcu_read_lock &ndev->lock irq_context: 0 (wq_completion)bond153#3 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &idev->mc_lock rcu_read_lock &dir->lock#2 irq_context: 0 (wq_completion)bond153#3 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &idev->mc_lock rcu_read_lock &ul->lock#2 irq_context: 0 (wq_completion)bond153#3 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &idev->mc_lock rcu_read_lock rcu_read_lock &obj_hash[i].lock irq_context: 0 (wq_completion)bond153#3 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &idev->mc_lock rcu_read_lock rcu_read_lock rcu_node_0 irq_context: 0 (wq_completion)bond153#3 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &idev->mc_lock rcu_read_lock rcu_read_lock &rq->__lock irq_context: 0 (wq_completion)bond153#3 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &idev->mc_lock rcu_read_lock rcu_read_lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)bond153#3 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &idev->mc_lock rcu_read_lock rcu_read_lock &dir->lock#2 irq_context: 0 (wq_completion)bond153#3 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &idev->mc_lock rcu_read_lock rcu_read_lock &tbl->lock irq_context: 0 (wq_completion)bond153#3 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &idev->mc_lock rcu_read_lock rcu_read_lock &n->lock irq_context: 0 (wq_completion)bond153#3 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &idev->mc_lock rcu_read_lock rcu_read_lock &____s->seqcount#9 irq_context: 0 (wq_completion)bond153#3 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &idev->mc_lock rcu_read_lock rcu_read_lock rcu_read_lock_bh rcu_read_lock rcu_read_lock_bh rcu_read_lock &br->multicast_lock irq_context: 0 (wq_completion)bond153#3 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &idev->mc_lock rcu_read_lock rcu_read_lock rcu_read_lock_bh rcu_read_lock rcu_read_lock_bh rcu_read_lock &br->multicast_lock &obj_hash[i].lock irq_context: 0 (wq_completion)bond153#3 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &idev->mc_lock rcu_read_lock rcu_read_lock rcu_read_lock_bh rcu_read_lock rcu_read_lock_bh rcu_read_lock &br->multicast_lock &base->lock irq_context: 0 (wq_completion)bond153#3 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &idev->mc_lock rcu_read_lock rcu_read_lock rcu_read_lock_bh rcu_read_lock rcu_read_lock_bh rcu_read_lock &br->multicast_lock &base->lock &obj_hash[i].lock irq_context: 0 (wq_completion)bond153#3 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &idev->mc_lock rcu_read_lock rcu_read_lock rcu_read_lock_bh rcu_read_lock rcu_read_lock_bh rcu_read_lock &data->lock irq_context: 0 (wq_completion)bond153#3 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &idev->mc_lock rcu_read_lock rcu_read_lock rcu_read_lock_bh rcu_read_lock rcu_read_lock_bh rcu_read_lock &obj_hash[i].lock irq_context: 0 (wq_completion)bond153#3 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &idev->mc_lock rcu_read_lock rcu_node_0 irq_context: 0 (wq_completion)bond153#3 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex.wait_lock irq_context: 0 (wq_completion)bond153#3 (work_completion)(&(&bond->mcast_work)->work) &p->pi_lock irq_context: 0 (wq_completion)bond153#3 (work_completion)(&(&bond->mcast_work)->work) &p->pi_lock &rq->__lock irq_context: 0 (wq_completion)bond153#3 (work_completion)(&(&bond->mcast_work)->work) &p->pi_lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)bond212#2 (work_completion)(&(&bond->mcast_work)->work) irq_context: 0 (wq_completion)bond212#2 (work_completion)(&(&bond->mcast_work)->work) &obj_hash[i].lock irq_context: 0 (wq_completion)bond212#2 (work_completion)(&(&bond->mcast_work)->work) &base->lock irq_context: 0 (wq_completion)bond212#2 (work_completion)(&(&bond->mcast_work)->work) &base->lock fill_pool_map-wait-type-override pool_lock irq_context: 0 (wq_completion)bond212#2 (work_completion)(&(&bond->mcast_work)->work) &base->lock &obj_hash[i].lock irq_context: 0 (wq_completion)bond212#2 (work_completion)(&(&bond->mcast_work)->work) &rq->__lock irq_context: 0 (wq_completion)bond185#3 (work_completion)(&(&slave->notify_work)->work) &rq->__lock irq_context: 0 (wq_completion)bond185#3 (work_completion)(&(&slave->notify_work)->work) &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)bond48#2 &rq->__lock irq_context: 0 (wq_completion)bond48#2 &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)bond182#3 (work_completion)(&(&bond->mcast_work)->work) &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)bond148#2 &rq->__lock irq_context: 0 (wq_completion)bond196#2 irq_context: 0 (wq_completion)bond196#2 (work_completion)(&(&slave->notify_work)->work) irq_context: 0 (wq_completion)bond196#2 (work_completion)(&(&slave->notify_work)->work) &obj_hash[i].lock irq_context: 0 (wq_completion)bond196#2 (work_completion)(&(&slave->notify_work)->work) &base->lock irq_context: 0 (wq_completion)bond196#2 (work_completion)(&(&slave->notify_work)->work) &base->lock &obj_hash[i].lock irq_context: 0 (wq_completion)bond196#2 (work_completion)(&(&slave->notify_work)->work) &rq->__lock irq_context: 0 (wq_completion)bond196#2 (work_completion)(&(&slave->notify_work)->work) &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)bond196#2 (work_completion)(&(&bond->mcast_work)->work) irq_context: 0 (wq_completion)bond196#2 (work_completion)(&(&bond->mcast_work)->work) &obj_hash[i].lock irq_context: 0 (wq_completion)bond196#2 (work_completion)(&(&bond->mcast_work)->work) &base->lock irq_context: 0 (wq_completion)bond196#2 (work_completion)(&(&bond->mcast_work)->work) &base->lock &obj_hash[i].lock irq_context: 0 (wq_completion)bond196#2 (work_completion)(&(&bond->mcast_work)->work) &rq->__lock irq_context: 0 (wq_completion)bond196#2 (work_completion)(&(&bond->mcast_work)->work) &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)bond114#2 &rq->__lock irq_context: 0 rtnl_mutex &xa->xa_lock#3 batched_entropy_u8.lock irq_context: 0 rtnl_mutex &xa->xa_lock#3 kfence_freelist_lock irq_context: 0 (wq_completion)bond176#3 (work_completion)(&(&slave->notify_work)->work) &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)bond224 irq_context: 0 (wq_completion)bond224 &rq->__lock irq_context: 0 (wq_completion)bond224 &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)bond224 (work_completion)(&(&slave->notify_work)->work) irq_context: 0 (wq_completion)bond224 (work_completion)(&(&slave->notify_work)->work) &obj_hash[i].lock irq_context: 0 (wq_completion)bond224 (work_completion)(&(&slave->notify_work)->work) &base->lock irq_context: 0 (wq_completion)bond224 (work_completion)(&(&slave->notify_work)->work) &base->lock &obj_hash[i].lock irq_context: 0 (wq_completion)bond224 (work_completion)(&(&slave->notify_work)->work) &rq->__lock irq_context: 0 (wq_completion)bond224 (work_completion)(&(&slave->notify_work)->work) &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)bond206 &rq->__lock irq_context: 0 (wq_completion)bond206 &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)mld (work_completion)(&(&idev->mc_ifc_work)->work) &idev->mc_lock rcu_read_lock rcu_read_lock &tbl->lock nl_table_wait.lock &p->pi_lock irq_context: 0 (wq_completion)bond224 (work_completion)(&(&bond->mcast_work)->work) irq_context: 0 (wq_completion)mld (work_completion)(&(&idev->mc_ifc_work)->work) &idev->mc_lock rcu_read_lock rcu_read_lock &tbl->lock nl_table_wait.lock &p->pi_lock &rq->__lock irq_context: 0 (wq_completion)bond224 (work_completion)(&(&bond->mcast_work)->work) &obj_hash[i].lock irq_context: 0 (wq_completion)mld (work_completion)(&(&idev->mc_ifc_work)->work) &idev->mc_lock rcu_read_lock rcu_read_lock &tbl->lock nl_table_wait.lock &p->pi_lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)bond224 (work_completion)(&(&bond->mcast_work)->work) &base->lock irq_context: 0 (wq_completion)bond224 (work_completion)(&(&bond->mcast_work)->work) &base->lock &obj_hash[i].lock irq_context: 0 (wq_completion)bond224 (work_completion)(&(&bond->mcast_work)->work) &rq->__lock irq_context: 0 (wq_completion)bond224 (work_completion)(&(&bond->mcast_work)->work) &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)mld (work_completion)(&(&idev->mc_ifc_work)->work) &idev->mc_lock rcu_read_lock rcu_read_lock &cfs_rq->removed.lock irq_context: 0 (wq_completion)bond67#2 irq_context: 0 (wq_completion)bond67#2 (work_completion)(&(&slave->notify_work)->work) irq_context: 0 (wq_completion)bond67#2 (work_completion)(&(&slave->notify_work)->work) &obj_hash[i].lock irq_context: 0 (wq_completion)bond67#2 (work_completion)(&(&slave->notify_work)->work) &base->lock irq_context: 0 (wq_completion)bond67#2 (work_completion)(&(&slave->notify_work)->work) &base->lock &obj_hash[i].lock irq_context: 0 (wq_completion)bond67#2 (work_completion)(&(&slave->notify_work)->work) &rq->__lock irq_context: 0 (wq_completion)bond188 (work_completion)(&(&slave->notify_work)->work) fill_pool_map-wait-type-override pool_lock irq_context: 0 (wq_completion)bond67#2 (work_completion)(&(&bond->mcast_work)->work) irq_context: 0 (wq_completion)bond67#2 (work_completion)(&(&bond->mcast_work)->work) &obj_hash[i].lock irq_context: 0 (wq_completion)bond67#2 (work_completion)(&(&bond->mcast_work)->work) &base->lock irq_context: 0 (wq_completion)bond67#2 (work_completion)(&(&bond->mcast_work)->work) &base->lock &obj_hash[i].lock irq_context: 0 (wq_completion)bond67#2 (work_completion)(&(&bond->mcast_work)->work) &rq->__lock irq_context: 0 (wq_completion)bond212 (work_completion)(&(&slave->notify_work)->work) fill_pool_map-wait-type-override pool_lock irq_context: 0 (wq_completion)bond140#3 (work_completion)(&(&slave->notify_work)->work) rtnl_mutex net_rwsem irq_context: 0 (wq_completion)bond140#3 (work_completion)(&(&slave->notify_work)->work) rtnl_mutex net_rwsem &list->lock#2 irq_context: 0 (wq_completion)bond140#3 (work_completion)(&(&slave->notify_work)->work) rtnl_mutex &tn->lock irq_context: 0 (wq_completion)bond140#3 (work_completion)(&(&slave->notify_work)->work) rtnl_mutex.wait_lock irq_context: 0 (wq_completion)bond140#3 (work_completion)(&(&slave->notify_work)->work) &p->pi_lock irq_context: 0 (wq_completion)bond101#2 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex fs_reclaim irq_context: 0 (wq_completion)bond101#2 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex fs_reclaim mmu_notifier_invalidate_range_start irq_context: 0 (wq_completion)bond101#2 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &c->lock irq_context: 0 (wq_completion)bond101#2 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex rcu_read_lock &bond->stats_lock/1 irq_context: 0 (wq_completion)bond101#2 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex rcu_read_lock &ndev->lock irq_context: 0 (wq_completion)bond101#2 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex nl_table_lock irq_context: 0 (wq_completion)bond101#2 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &obj_hash[i].lock irq_context: 0 (wq_completion)bond101#2 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex nl_table_wait.lock irq_context: 0 (wq_completion)bond101#2 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex net_rwsem irq_context: 0 (wq_completion)bond101#2 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex net_rwsem &list->lock#2 irq_context: 0 (wq_completion)bond101#2 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &tn->lock irq_context: 0 (wq_completion)bond101#2 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &idev->mc_lock irq_context: 0 (wq_completion)bond101#2 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &idev->mc_lock fs_reclaim irq_context: 0 (wq_completion)bond101#2 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &idev->mc_lock fs_reclaim mmu_notifier_invalidate_range_start irq_context: 0 (wq_completion)bond101#2 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &idev->mc_lock fs_reclaim mmu_notifier_invalidate_range_start &rq->__lock irq_context: 0 (wq_completion)bond101#2 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &idev->mc_lock fs_reclaim mmu_notifier_invalidate_range_start &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)bond101#2 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &idev->mc_lock &c->lock irq_context: 0 (wq_completion)bond101#2 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &idev->mc_lock rcu_read_lock &ndev->lock irq_context: 0 (wq_completion)bond101#2 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &idev->mc_lock rcu_read_lock &dir->lock#2 irq_context: 0 (wq_completion)bond101#2 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &idev->mc_lock rcu_read_lock &c->lock irq_context: 0 (wq_completion)bond101#2 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &idev->mc_lock rcu_read_lock &n->list_lock irq_context: 0 (wq_completion)bond101#2 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &idev->mc_lock rcu_read_lock &n->list_lock &c->lock irq_context: 0 (wq_completion)bond101#2 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &idev->mc_lock rcu_read_lock rcu_node_0 irq_context: 0 (wq_completion)bond101#2 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &idev->mc_lock rcu_read_lock &rq->__lock irq_context: 0 (wq_completion)bond101#2 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &idev->mc_lock rcu_read_lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)bond101#2 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &idev->mc_lock rcu_read_lock &ul->lock#2 irq_context: 0 (wq_completion)bond101#2 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &idev->mc_lock rcu_read_lock rcu_read_lock &obj_hash[i].lock irq_context: 0 (wq_completion)bond101#2 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &idev->mc_lock rcu_read_lock rcu_read_lock &c->lock irq_context: 0 (wq_completion)bond101#2 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &idev->mc_lock rcu_read_lock rcu_read_lock &dir->lock#2 irq_context: 0 (wq_completion)bond101#2 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &idev->mc_lock rcu_read_lock rcu_read_lock &tbl->lock irq_context: 0 (wq_completion)bond101#2 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &idev->mc_lock rcu_read_lock rcu_read_lock &n->lock irq_context: 0 (wq_completion)bond101#2 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &idev->mc_lock rcu_read_lock rcu_read_lock &____s->seqcount#9 irq_context: 0 (wq_completion)bond101#2 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &idev->mc_lock rcu_read_lock rcu_read_lock rcu_read_lock_bh rcu_read_lock rcu_read_lock_bh rcu_read_lock &br->multicast_lock irq_context: 0 (wq_completion)bond101#2 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &idev->mc_lock rcu_read_lock rcu_read_lock rcu_read_lock_bh rcu_read_lock rcu_read_lock_bh rcu_read_lock &br->multicast_lock &obj_hash[i].lock irq_context: 0 (wq_completion)bond101#2 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &idev->mc_lock rcu_read_lock rcu_read_lock rcu_read_lock_bh rcu_read_lock rcu_read_lock_bh rcu_read_lock &br->multicast_lock &base->lock irq_context: 0 (wq_completion)bond101#2 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &idev->mc_lock rcu_read_lock rcu_read_lock rcu_read_lock_bh rcu_read_lock rcu_read_lock_bh rcu_read_lock &br->multicast_lock &base->lock &obj_hash[i].lock irq_context: 0 (wq_completion)bond101#2 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &idev->mc_lock rcu_read_lock rcu_read_lock rcu_read_lock_bh rcu_read_lock rcu_read_lock_bh rcu_read_lock &data->lock irq_context: 0 (wq_completion)bond101#2 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &idev->mc_lock rcu_read_lock rcu_read_lock rcu_read_lock_bh rcu_read_lock rcu_read_lock_bh rcu_read_lock &obj_hash[i].lock irq_context: 0 (wq_completion)bond101#2 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &idev->mc_lock rcu_read_lock rcu_read_lock &rq->__lock irq_context: 0 (wq_completion)bond101#2 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &idev->mc_lock rcu_read_lock rcu_read_lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)bond101#2 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex.wait_lock irq_context: 0 (wq_completion)bond101#2 (work_completion)(&(&bond->mcast_work)->work) &p->pi_lock irq_context: 0 (wq_completion)bond197#3 irq_context: 0 (wq_completion)bond197#3 (work_completion)(&(&slave->notify_work)->work) irq_context: 0 (wq_completion)bond197#3 (work_completion)(&(&slave->notify_work)->work) &obj_hash[i].lock irq_context: 0 (wq_completion)bond197#3 (work_completion)(&(&slave->notify_work)->work) &base->lock irq_context: 0 (wq_completion)bond197#3 (work_completion)(&(&slave->notify_work)->work) &base->lock &obj_hash[i].lock irq_context: 0 (wq_completion)bond197#3 (work_completion)(&(&slave->notify_work)->work) &rq->__lock irq_context: 0 (wq_completion)bond197#3 (work_completion)(&(&bond->mcast_work)->work) irq_context: 0 (wq_completion)bond197#3 (work_completion)(&(&bond->mcast_work)->work) &obj_hash[i].lock irq_context: 0 (wq_completion)bond197#3 (work_completion)(&(&bond->mcast_work)->work) &base->lock irq_context: 0 (wq_completion)bond197#3 (work_completion)(&(&bond->mcast_work)->work) &base->lock &obj_hash[i].lock irq_context: 0 (wq_completion)bond207 (work_completion)(&(&slave->notify_work)->work) &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)bond172#3 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex fs_reclaim irq_context: 0 (wq_completion)bond172#3 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex fs_reclaim mmu_notifier_invalidate_range_start irq_context: 0 (wq_completion)bond172#3 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex rcu_read_lock &bond->stats_lock/1 irq_context: 0 (wq_completion)bond172#3 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex rcu_read_lock &ndev->lock irq_context: 0 (wq_completion)bond172#3 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex nl_table_lock irq_context: 0 (wq_completion)bond172#3 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &obj_hash[i].lock irq_context: 0 (wq_completion)bond172#3 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex nl_table_wait.lock irq_context: 0 (wq_completion)bond172#3 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex net_rwsem irq_context: 0 (wq_completion)bond172#3 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex net_rwsem &list->lock#2 irq_context: 0 (wq_completion)bond172#3 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &tn->lock irq_context: 0 (wq_completion)bond172#3 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &idev->mc_lock irq_context: 0 (wq_completion)bond172#3 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &idev->mc_lock fs_reclaim irq_context: 0 (wq_completion)bond172#3 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &idev->mc_lock fs_reclaim mmu_notifier_invalidate_range_start irq_context: 0 (wq_completion)bond172#3 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &idev->mc_lock &c->lock irq_context: 0 (wq_completion)bond172#3 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &idev->mc_lock rcu_read_lock &ndev->lock irq_context: 0 (wq_completion)bond172#3 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &idev->mc_lock rcu_read_lock &c->lock irq_context: 0 (wq_completion)bond172#3 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &idev->mc_lock rcu_read_lock &dir->lock#2 irq_context: 0 (wq_completion)bond172#3 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &idev->mc_lock rcu_read_lock &ul->lock#2 irq_context: 0 (wq_completion)bond172#3 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &idev->mc_lock rcu_read_lock rcu_read_lock &____s->seqcount#10 irq_context: 0 (wq_completion)bond172#3 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &idev->mc_lock rcu_read_lock rcu_read_lock rcu_read_lock_bh rcu_read_lock rcu_read_lock_bh rcu_read_lock &br->multicast_lock irq_context: 0 (wq_completion)bond172#3 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &idev->mc_lock rcu_read_lock rcu_read_lock rcu_read_lock_bh rcu_read_lock rcu_read_lock_bh rcu_read_lock &br->multicast_lock &obj_hash[i].lock irq_context: 0 (wq_completion)bond172#3 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &idev->mc_lock rcu_read_lock rcu_read_lock rcu_read_lock_bh rcu_read_lock rcu_read_lock_bh rcu_read_lock &br->multicast_lock &base->lock irq_context: 0 (wq_completion)bond172#3 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &idev->mc_lock rcu_read_lock rcu_read_lock rcu_read_lock_bh rcu_read_lock rcu_read_lock_bh rcu_read_lock &br->multicast_lock &base->lock &obj_hash[i].lock irq_context: 0 (wq_completion)bond172#3 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &idev->mc_lock rcu_read_lock rcu_read_lock rcu_read_lock_bh rcu_read_lock rcu_read_lock_bh rcu_read_lock &data->lock irq_context: 0 (wq_completion)bond172#3 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &idev->mc_lock rcu_read_lock rcu_read_lock rcu_read_lock_bh rcu_read_lock rcu_read_lock_bh rcu_read_lock &obj_hash[i].lock irq_context: 0 (wq_completion)bond172#3 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex.wait_lock irq_context: 0 (wq_completion)bond172#3 (work_completion)(&(&bond->mcast_work)->work) &p->pi_lock irq_context: 0 (wq_completion)bond149#2 (work_completion)(&(&slave->notify_work)->work) rtnl_mutex net_rwsem irq_context: 0 (wq_completion)bond149#2 (work_completion)(&(&slave->notify_work)->work) rtnl_mutex net_rwsem &list->lock#2 irq_context: 0 (wq_completion)bond149#2 (work_completion)(&(&slave->notify_work)->work) rtnl_mutex &tn->lock irq_context: 0 (wq_completion)bond149#2 (work_completion)(&(&slave->notify_work)->work) rtnl_mutex.wait_lock irq_context: 0 (wq_completion)bond149#2 (work_completion)(&(&slave->notify_work)->work) &p->pi_lock irq_context: 0 (wq_completion)bond197#3 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex fs_reclaim irq_context: 0 (wq_completion)bond197#3 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex fs_reclaim mmu_notifier_invalidate_range_start irq_context: 0 (wq_completion)bond197#3 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &c->lock irq_context: 0 (wq_completion)bond197#3 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &____s->seqcount#2 irq_context: 0 (wq_completion)bond197#3 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &____s->seqcount irq_context: 0 (wq_completion)bond197#3 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &rq->__lock irq_context: 0 (wq_completion)bond197#3 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex rcu_read_lock &bond->stats_lock/1 irq_context: 0 (wq_completion)bond197#3 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex rcu_read_lock &ndev->lock irq_context: 0 (wq_completion)bond197#3 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex rcu_read_lock rcu_node_0 irq_context: 0 (wq_completion)bond197#3 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex rcu_read_lock &rq->__lock irq_context: 0 (wq_completion)bond197#3 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex rcu_read_lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)bond197#3 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex nl_table_lock irq_context: 0 (wq_completion)bond197#3 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &obj_hash[i].lock irq_context: 0 (wq_completion)bond197#3 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex nl_table_wait.lock irq_context: 0 (wq_completion)bond197#3 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex net_rwsem irq_context: 0 (wq_completion)bond197#3 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex net_rwsem &rq->__lock irq_context: 0 (wq_completion)bond197#3 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex net_rwsem &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)bond197#3 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex net_rwsem &list->lock#2 irq_context: 0 (wq_completion)bond197#3 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &tn->lock irq_context: 0 (wq_completion)bond197#3 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &idev->mc_lock irq_context: 0 (wq_completion)bond197#3 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex.wait_lock irq_context: 0 (wq_completion)bond197#3 (work_completion)(&(&bond->mcast_work)->work) &p->pi_lock irq_context: 0 (wq_completion)bond197#3 (work_completion)(&(&bond->mcast_work)->work) &p->pi_lock &rq->__lock irq_context: 0 (wq_completion)bond197#3 (work_completion)(&(&bond->mcast_work)->work) &p->pi_lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)bond197#3 (work_completion)(&(&bond->mcast_work)->work) &rq->__lock irq_context: 0 (wq_completion)bond197#3 (work_completion)(&(&bond->mcast_work)->work) &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)bond214#2 irq_context: 0 (wq_completion)bond214#2 (work_completion)(&(&slave->notify_work)->work) irq_context: 0 (wq_completion)bond214#2 (work_completion)(&(&slave->notify_work)->work) &obj_hash[i].lock irq_context: 0 (wq_completion)bond214#2 (work_completion)(&(&slave->notify_work)->work) &base->lock irq_context: 0 (wq_completion)bond214#2 (work_completion)(&(&slave->notify_work)->work) &base->lock &obj_hash[i].lock irq_context: 0 (wq_completion)bond214#2 (work_completion)(&(&bond->mcast_work)->work) irq_context: 0 (wq_completion)bond214#2 (work_completion)(&(&bond->mcast_work)->work) &obj_hash[i].lock irq_context: 0 (wq_completion)bond214#2 (work_completion)(&(&bond->mcast_work)->work) &base->lock irq_context: 0 (wq_completion)bond214#2 (work_completion)(&(&bond->mcast_work)->work) &base->lock &obj_hash[i].lock irq_context: 0 (wq_completion)bond214#2 (work_completion)(&(&bond->mcast_work)->work) &rq->__lock irq_context: 0 (wq_completion)bond214#2 (work_completion)(&(&bond->mcast_work)->work) &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)bond211 (work_completion)(&(&slave->notify_work)->work) &rq->__lock irq_context: 0 (wq_completion)bond225 irq_context: 0 (wq_completion)bond225 (work_completion)(&(&slave->notify_work)->work) irq_context: 0 (wq_completion)bond225 (work_completion)(&(&slave->notify_work)->work) &obj_hash[i].lock irq_context: 0 (wq_completion)bond225 (work_completion)(&(&slave->notify_work)->work) &base->lock irq_context: 0 (wq_completion)bond225 (work_completion)(&(&slave->notify_work)->work) &base->lock &obj_hash[i].lock irq_context: 0 (wq_completion)bond225 (work_completion)(&(&slave->notify_work)->work) &rq->__lock irq_context: 0 (wq_completion)bond225 (work_completion)(&(&slave->notify_work)->work) &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)bond173#3 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex fs_reclaim irq_context: 0 (wq_completion)bond173#3 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex fs_reclaim &rq->__lock irq_context: 0 (wq_completion)bond173#3 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex fs_reclaim &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)bond173#3 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex fs_reclaim mmu_notifier_invalidate_range_start irq_context: 0 (wq_completion)bond173#3 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &c->lock irq_context: 0 (wq_completion)bond173#3 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex rcu_read_lock &bond->stats_lock/1 irq_context: 0 (wq_completion)bond173#3 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex rcu_read_lock &ndev->lock irq_context: 0 (wq_completion)bond173#3 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex nl_table_lock irq_context: 0 (wq_completion)bond173#3 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &obj_hash[i].lock irq_context: 0 (wq_completion)bond173#3 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex nl_table_wait.lock irq_context: 0 (wq_completion)bond173#3 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex net_rwsem irq_context: 0 (wq_completion)bond173#3 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex net_rwsem &list->lock#2 irq_context: 0 (wq_completion)bond173#3 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex net_rwsem &rq->__lock irq_context: 0 (wq_completion)bond173#3 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex net_rwsem &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)bond173#3 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &tn->lock irq_context: 0 (wq_completion)bond173#3 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &idev->mc_lock irq_context: 0 (wq_completion)bond173#3 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &idev->mc_lock fs_reclaim irq_context: 0 (wq_completion)bond173#3 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &idev->mc_lock fs_reclaim mmu_notifier_invalidate_range_start irq_context: 0 (wq_completion)bond173#3 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &idev->mc_lock rcu_read_lock &ndev->lock irq_context: 0 (wq_completion)bond173#3 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &idev->mc_lock rcu_read_lock &c->lock irq_context: 0 (wq_completion)bond173#3 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &idev->mc_lock rcu_read_lock &dir->lock#2 irq_context: 0 (wq_completion)bond173#3 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &idev->mc_lock rcu_read_lock &ul->lock#2 irq_context: 0 (wq_completion)bond173#3 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &idev->mc_lock rcu_read_lock rcu_read_lock &____s->seqcount#10 irq_context: 0 (wq_completion)bond173#3 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &idev->mc_lock rcu_read_lock rcu_read_lock rcu_read_lock_bh rcu_read_lock rcu_read_lock_bh rcu_read_lock &br->multicast_lock irq_context: 0 (wq_completion)bond173#3 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &idev->mc_lock rcu_read_lock rcu_read_lock rcu_read_lock_bh rcu_read_lock rcu_read_lock_bh rcu_read_lock &br->multicast_lock &obj_hash[i].lock irq_context: 0 (wq_completion)bond173#3 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &idev->mc_lock rcu_read_lock rcu_read_lock rcu_read_lock_bh rcu_read_lock rcu_read_lock_bh rcu_read_lock &br->multicast_lock &base->lock irq_context: 0 (wq_completion)bond173#3 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &idev->mc_lock rcu_read_lock rcu_read_lock rcu_read_lock_bh rcu_read_lock rcu_read_lock_bh rcu_read_lock &br->multicast_lock &base->lock &obj_hash[i].lock irq_context: 0 (wq_completion)bond173#3 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &idev->mc_lock rcu_read_lock rcu_read_lock rcu_read_lock_bh rcu_read_lock rcu_read_lock_bh rcu_read_lock &data->lock irq_context: 0 (wq_completion)bond173#3 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &idev->mc_lock rcu_read_lock rcu_read_lock rcu_read_lock_bh rcu_read_lock rcu_read_lock_bh rcu_read_lock &obj_hash[i].lock irq_context: 0 (wq_completion)bond173#3 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &idev->mc_lock rcu_read_lock rcu_read_lock rcu_node_0 irq_context: 0 (wq_completion)bond173#3 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &idev->mc_lock rcu_read_lock rcu_read_lock &rq->__lock irq_context: 0 (wq_completion)bond173#3 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &idev->mc_lock rcu_read_lock rcu_node_0 irq_context: 0 (wq_completion)bond173#3 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex.wait_lock irq_context: 0 (wq_completion)bond173#3 (work_completion)(&(&bond->mcast_work)->work) &p->pi_lock irq_context: 0 (wq_completion)bond173#3 (work_completion)(&(&bond->mcast_work)->work) &p->pi_lock &rq->__lock irq_context: 0 (wq_completion)bond173#3 (work_completion)(&(&bond->mcast_work)->work) &p->pi_lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)events (linkwatch_work).work &cfs_rq->removed.lock irq_context: 0 (wq_completion)events (linkwatch_work).work &obj_hash[i].lock irq_context: 0 (wq_completion)events (linkwatch_work).work pool_lock#2 irq_context: 0 (wq_completion)bond225 (work_completion)(&(&bond->mcast_work)->work) irq_context: 0 (wq_completion)bond225 (work_completion)(&(&bond->mcast_work)->work) &obj_hash[i].lock irq_context: 0 (wq_completion)bond225 (work_completion)(&(&bond->mcast_work)->work) &base->lock irq_context: 0 (wq_completion)bond225 (work_completion)(&(&bond->mcast_work)->work) &base->lock &obj_hash[i].lock irq_context: 0 (wq_completion)bond194 &rq->__lock irq_context: 0 (wq_completion)bond194 &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)bond193#3 (work_completion)(&(&slave->notify_work)->work) &rq->__lock irq_context: 0 (wq_completion)bond193#3 (work_completion)(&(&slave->notify_work)->work) &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)bond198#3 irq_context: 0 (wq_completion)bond198#3 &rq->__lock irq_context: 0 (wq_completion)bond198#3 &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)bond198#3 (work_completion)(&(&slave->notify_work)->work) irq_context: 0 (wq_completion)bond198#3 (work_completion)(&(&slave->notify_work)->work) &obj_hash[i].lock irq_context: 0 (wq_completion)bond198#3 (work_completion)(&(&slave->notify_work)->work) &base->lock irq_context: 0 (wq_completion)bond198#3 (work_completion)(&(&slave->notify_work)->work) &base->lock &obj_hash[i].lock irq_context: 0 (wq_completion)bond198#3 (work_completion)(&(&slave->notify_work)->work) &rq->__lock irq_context: 0 (wq_completion)bond198#3 (work_completion)(&(&slave->notify_work)->work) &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)bond210#2 &rq->__lock irq_context: 0 (wq_completion)bond210#2 &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)bond165#2 &rq->__lock irq_context: 0 (wq_completion)bond165#2 &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)bond67#2 (work_completion)(&(&slave->notify_work)->work) &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 ipvs->sync_mutex irq_context: 0 ipvs->sync_mutex &rq->__lock irq_context: 0 ipvs->sync_mutex &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 ipvs->sync_mutex &mm->mmap_lock irq_context: 0 (wq_completion)bond198#3 (work_completion)(&(&bond->mcast_work)->work) irq_context: 0 (wq_completion)bond198#3 (work_completion)(&(&bond->mcast_work)->work) &obj_hash[i].lock irq_context: 0 (wq_completion)bond198#3 (work_completion)(&(&bond->mcast_work)->work) &base->lock irq_context: 0 (wq_completion)bond198#3 (work_completion)(&(&bond->mcast_work)->work) &base->lock &obj_hash[i].lock irq_context: 0 (wq_completion)bond198#3 (work_completion)(&(&bond->mcast_work)->work) &rq->__lock irq_context: 0 (wq_completion)bond198#3 (work_completion)(&(&bond->mcast_work)->work) &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)bond143 &rq->__lock irq_context: 0 (wq_completion)bond143 &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)bond183#2 (work_completion)(&(&slave->notify_work)->work) &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)bond226 irq_context: 0 (wq_completion)bond226 (work_completion)(&(&slave->notify_work)->work) irq_context: 0 (wq_completion)bond226 (work_completion)(&(&slave->notify_work)->work) &obj_hash[i].lock irq_context: 0 (wq_completion)bond226 (work_completion)(&(&slave->notify_work)->work) &base->lock irq_context: 0 (wq_completion)bond226 (work_completion)(&(&slave->notify_work)->work) &base->lock &obj_hash[i].lock irq_context: 0 (wq_completion)bond226 (work_completion)(&(&slave->notify_work)->work) &rq->__lock irq_context: 0 (wq_completion)bond226 (work_completion)(&(&slave->notify_work)->work) &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)bond215 (work_completion)(&(&bond->mcast_work)->work) &rq->__lock irq_context: 0 (wq_completion)bond215 (work_completion)(&(&bond->mcast_work)->work) &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)bond188 (work_completion)(&(&slave->notify_work)->work) rtnl_mutex net_rwsem irq_context: 0 (wq_completion)bond188 (work_completion)(&(&slave->notify_work)->work) rtnl_mutex net_rwsem &list->lock#2 irq_context: 0 (wq_completion)bond188 (work_completion)(&(&slave->notify_work)->work) rtnl_mutex &tn->lock irq_context: 0 (wq_completion)bond188 (work_completion)(&(&slave->notify_work)->work) rtnl_mutex.wait_lock irq_context: 0 (wq_completion)bond188 (work_completion)(&(&slave->notify_work)->work) &p->pi_lock irq_context: 0 (wq_completion)bond177#3 (work_completion)(&(&bond->mcast_work)->work) &rq->__lock irq_context: 0 (wq_completion)bond177#3 (work_completion)(&(&bond->mcast_work)->work) &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)bond226 (work_completion)(&(&bond->mcast_work)->work) irq_context: 0 (wq_completion)bond226 (work_completion)(&(&bond->mcast_work)->work) &obj_hash[i].lock irq_context: 0 (wq_completion)bond226 (work_completion)(&(&bond->mcast_work)->work) &base->lock irq_context: 0 (wq_completion)bond226 (work_completion)(&(&bond->mcast_work)->work) &base->lock &obj_hash[i].lock irq_context: 0 (wq_completion)bond226 (work_completion)(&(&bond->mcast_work)->work) &rq->__lock irq_context: 0 (wq_completion)bond226 (work_completion)(&(&bond->mcast_work)->work) &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)bond215#2 irq_context: 0 (wq_completion)bond215#2 (work_completion)(&(&slave->notify_work)->work) irq_context: 0 (wq_completion)bond215#2 (work_completion)(&(&slave->notify_work)->work) &obj_hash[i].lock irq_context: 0 (wq_completion)bond215#2 (work_completion)(&(&slave->notify_work)->work) &base->lock irq_context: 0 (wq_completion)bond215#2 (work_completion)(&(&slave->notify_work)->work) &base->lock &obj_hash[i].lock irq_context: 0 (wq_completion)bond215#2 (work_completion)(&(&slave->notify_work)->work) &rq->__lock irq_context: 0 (wq_completion)bond215#2 (work_completion)(&(&slave->notify_work)->work) &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)bond215#2 (work_completion)(&(&bond->mcast_work)->work) irq_context: 0 (wq_completion)bond215#2 (work_completion)(&(&bond->mcast_work)->work) &obj_hash[i].lock irq_context: 0 (wq_completion)bond215#2 (work_completion)(&(&bond->mcast_work)->work) &base->lock irq_context: 0 (wq_completion)bond215#2 (work_completion)(&(&bond->mcast_work)->work) &base->lock &obj_hash[i].lock irq_context: 0 (wq_completion)bond209#2 &rq->__lock irq_context: 0 (wq_completion)bond209#2 &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)bond192#3 &rq->__lock irq_context: 0 (wq_completion)bond192#3 &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)bond199#3 irq_context: 0 (wq_completion)bond199#3 (work_completion)(&(&slave->notify_work)->work) irq_context: 0 (wq_completion)bond199#3 (work_completion)(&(&slave->notify_work)->work) &obj_hash[i].lock irq_context: 0 (wq_completion)bond199#3 (work_completion)(&(&slave->notify_work)->work) &base->lock irq_context: 0 (wq_completion)bond199#3 (work_completion)(&(&slave->notify_work)->work) &base->lock &obj_hash[i].lock irq_context: 0 (wq_completion)bond199#3 (work_completion)(&(&slave->notify_work)->work) &rq->__lock irq_context: 0 (wq_completion)bond199#3 (work_completion)(&(&slave->notify_work)->work) &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)bond96#3 (work_completion)(&(&slave->notify_work)->work) &base->lock fill_pool_map-wait-type-override pool_lock irq_context: 0 (wq_completion)bond199#3 (work_completion)(&(&bond->mcast_work)->work) irq_context: 0 (wq_completion)bond199#3 (work_completion)(&(&bond->mcast_work)->work) &obj_hash[i].lock irq_context: 0 (wq_completion)bond199#3 (work_completion)(&(&bond->mcast_work)->work) &base->lock irq_context: 0 (wq_completion)bond199#3 (work_completion)(&(&bond->mcast_work)->work) &base->lock &obj_hash[i].lock irq_context: 0 (wq_completion)bond212#2 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex fs_reclaim irq_context: 0 (wq_completion)bond212#2 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex fs_reclaim &rq->__lock irq_context: 0 (wq_completion)bond212#2 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex fs_reclaim &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)bond212#2 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex fs_reclaim mmu_notifier_invalidate_range_start irq_context: 0 (wq_completion)bond212#2 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex batched_entropy_u8.lock irq_context: 0 (wq_completion)bond212#2 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex kfence_freelist_lock irq_context: 0 (wq_completion)bond212#2 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex rcu_read_lock &bond->stats_lock/1 irq_context: 0 (wq_completion)bond212#2 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex rcu_read_lock &ndev->lock irq_context: 0 (wq_completion)bond212#2 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex nl_table_lock irq_context: 0 (wq_completion)bond212#2 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &obj_hash[i].lock irq_context: 0 (wq_completion)bond212#2 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex quarantine_lock irq_context: 0 (wq_completion)bond212#2 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &rq->__lock irq_context: 0 (wq_completion)bond212#2 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)bond212#2 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &meta->lock irq_context: 0 (wq_completion)bond212#2 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex nl_table_wait.lock irq_context: 0 (wq_completion)bond212#2 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex net_rwsem irq_context: 0 (wq_completion)bond212#2 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex net_rwsem &list->lock#2 irq_context: 0 (wq_completion)bond212#2 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &tn->lock irq_context: 0 (wq_completion)bond212#2 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &idev->mc_lock irq_context: 0 (wq_completion)bond212#2 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &idev->mc_lock fs_reclaim irq_context: 0 (wq_completion)bond212#2 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &idev->mc_lock fs_reclaim mmu_notifier_invalidate_range_start irq_context: 0 (wq_completion)bond212#2 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &idev->mc_lock remove_cache_srcu irq_context: 0 (wq_completion)bond212#2 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &idev->mc_lock remove_cache_srcu quarantine_lock irq_context: 0 (wq_completion)bond212#2 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &idev->mc_lock remove_cache_srcu &c->lock irq_context: 0 (wq_completion)bond212#2 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &idev->mc_lock remove_cache_srcu &rq->__lock irq_context: 0 (wq_completion)bond212#2 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &idev->mc_lock remove_cache_srcu &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)bond212#2 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &idev->mc_lock remove_cache_srcu &n->list_lock irq_context: 0 (wq_completion)bond212#2 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &idev->mc_lock &c->lock irq_context: 0 (wq_completion)bond212#2 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &idev->mc_lock rcu_read_lock &ndev->lock irq_context: 0 (wq_completion)bond212#2 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &idev->mc_lock rcu_read_lock &dir->lock#2 irq_context: 0 (wq_completion)bond212#2 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &idev->mc_lock rcu_read_lock &ul->lock#2 irq_context: 0 (wq_completion)bond212#2 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &idev->mc_lock rcu_read_lock rcu_read_lock &____s->seqcount#10 irq_context: 0 (wq_completion)bond212#2 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &idev->mc_lock rcu_read_lock rcu_read_lock rcu_node_0 irq_context: 0 (wq_completion)bond212#2 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &idev->mc_lock rcu_read_lock rcu_read_lock &rq->__lock irq_context: 0 (wq_completion)bond212#2 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &idev->mc_lock rcu_read_lock rcu_read_lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)bond212#2 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &idev->mc_lock rcu_read_lock rcu_read_lock rcu_read_lock_bh rcu_read_lock rcu_read_lock_bh rcu_read_lock &br->multicast_lock irq_context: 0 (wq_completion)bond212#2 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &idev->mc_lock rcu_read_lock rcu_read_lock rcu_read_lock_bh rcu_read_lock rcu_read_lock_bh rcu_read_lock &br->multicast_lock &obj_hash[i].lock irq_context: 0 (wq_completion)bond212#2 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &idev->mc_lock rcu_read_lock rcu_read_lock rcu_read_lock_bh rcu_read_lock rcu_read_lock_bh rcu_read_lock &br->multicast_lock &base->lock irq_context: 0 (wq_completion)bond212#2 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &idev->mc_lock rcu_read_lock rcu_read_lock rcu_read_lock_bh rcu_read_lock rcu_read_lock_bh rcu_read_lock &data->lock irq_context: 0 (wq_completion)bond212#2 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &idev->mc_lock rcu_read_lock rcu_read_lock rcu_read_lock_bh rcu_read_lock rcu_read_lock_bh rcu_read_lock &obj_hash[i].lock irq_context: 0 (wq_completion)bond212#2 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &idev->mc_lock rcu_read_lock rcu_node_0 irq_context: 0 (wq_completion)bond212#2 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex.wait_lock irq_context: 0 (wq_completion)bond212#2 (work_completion)(&(&bond->mcast_work)->work) &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)bond212#2 (work_completion)(&(&bond->mcast_work)->work) &p->pi_lock irq_context: 0 (wq_completion)bond212#2 (work_completion)(&(&slave->notify_work)->work) rtnl_mutex net_rwsem irq_context: 0 (wq_completion)bond212#2 (work_completion)(&(&slave->notify_work)->work) rtnl_mutex net_rwsem &list->lock#2 irq_context: 0 (wq_completion)bond212#2 (work_completion)(&(&slave->notify_work)->work) rtnl_mutex &tn->lock irq_context: 0 (wq_completion)bond212#2 (work_completion)(&(&slave->notify_work)->work) rtnl_mutex.wait_lock irq_context: 0 (wq_completion)bond212#2 (work_completion)(&(&slave->notify_work)->work) &p->pi_lock irq_context: 0 (wq_completion)bond227 irq_context: 0 (wq_completion)bond227 (work_completion)(&(&slave->notify_work)->work) irq_context: 0 (wq_completion)bond227 (work_completion)(&(&slave->notify_work)->work) &obj_hash[i].lock irq_context: 0 (wq_completion)bond227 (work_completion)(&(&slave->notify_work)->work) &base->lock irq_context: 0 (wq_completion)bond227 (work_completion)(&(&slave->notify_work)->work) &base->lock &obj_hash[i].lock irq_context: 0 (wq_completion)bond227 (work_completion)(&(&bond->mcast_work)->work) irq_context: 0 (wq_completion)bond227 (work_completion)(&(&bond->mcast_work)->work) &obj_hash[i].lock irq_context: 0 (wq_completion)bond227 (work_completion)(&(&bond->mcast_work)->work) &base->lock irq_context: 0 (wq_completion)bond227 (work_completion)(&(&bond->mcast_work)->work) &base->lock &obj_hash[i].lock irq_context: 0 (wq_completion)bond227 (work_completion)(&(&bond->mcast_work)->work) &rq->__lock irq_context: 0 (wq_completion)bond227 (work_completion)(&(&bond->mcast_work)->work) &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)bond216#2 irq_context: 0 (wq_completion)bond216#2 &rq->__lock irq_context: 0 (wq_completion)bond216#2 (work_completion)(&(&slave->notify_work)->work) irq_context: 0 (wq_completion)bond216#2 (work_completion)(&(&slave->notify_work)->work) &obj_hash[i].lock irq_context: 0 (wq_completion)bond216#2 (work_completion)(&(&slave->notify_work)->work) &base->lock irq_context: 0 (wq_completion)bond216#2 (work_completion)(&(&slave->notify_work)->work) &base->lock &obj_hash[i].lock irq_context: 0 (wq_completion)bond216#2 (work_completion)(&(&slave->notify_work)->work) &rq->__lock irq_context: 0 (wq_completion)bond216#2 (work_completion)(&(&slave->notify_work)->work) &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)bond216#2 (work_completion)(&(&bond->mcast_work)->work) irq_context: 0 (wq_completion)bond216#2 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex fs_reclaim irq_context: 0 (wq_completion)bond216#2 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex fs_reclaim mmu_notifier_invalidate_range_start irq_context: 0 (wq_completion)bond216#2 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &c->lock irq_context: 0 (wq_completion)bond216#2 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &rq->__lock irq_context: 0 (wq_completion)bond216#2 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)bond216#2 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex rcu_read_lock &bond->stats_lock/1 irq_context: 0 (wq_completion)bond216#2 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex rcu_read_lock &ndev->lock irq_context: 0 (wq_completion)bond216#2 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex nl_table_lock irq_context: 0 (wq_completion)bond216#2 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &obj_hash[i].lock irq_context: 0 (wq_completion)bond216#2 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex nl_table_wait.lock irq_context: 0 (wq_completion)bond216#2 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex net_rwsem irq_context: 0 (wq_completion)bond216#2 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex net_rwsem &list->lock#2 irq_context: 0 (wq_completion)bond216#2 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &tn->lock irq_context: 0 (wq_completion)bond216#2 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &idev->mc_lock irq_context: 0 (wq_completion)bond63 (work_completion)(&(&slave->notify_work)->work) rtnl_mutex net_rwsem irq_context: 0 (wq_completion)bond63 (work_completion)(&(&slave->notify_work)->work) rtnl_mutex net_rwsem &list->lock#2 irq_context: 0 (wq_completion)bond63 (work_completion)(&(&slave->notify_work)->work) rtnl_mutex &tn->lock irq_context: 0 (wq_completion)bond63 (work_completion)(&(&slave->notify_work)->work) rtnl_mutex.wait_lock irq_context: 0 (wq_completion)bond63 (work_completion)(&(&slave->notify_work)->work) &p->pi_lock irq_context: 0 (wq_completion)bond191#3 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex fs_reclaim irq_context: 0 (wq_completion)bond191#3 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex fs_reclaim mmu_notifier_invalidate_range_start irq_context: 0 (wq_completion)bond191#3 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex rcu_read_lock &bond->stats_lock/1 irq_context: 0 (wq_completion)bond191#3 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex rcu_read_lock &ndev->lock irq_context: 0 (wq_completion)bond191#3 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex nl_table_lock irq_context: 0 (wq_completion)bond191#3 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &obj_hash[i].lock irq_context: 0 (wq_completion)bond191#3 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex nl_table_wait.lock irq_context: 0 (wq_completion)bond191#3 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex net_rwsem irq_context: 0 (wq_completion)bond191#3 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex net_rwsem &list->lock#2 irq_context: 0 (wq_completion)bond191#3 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &tn->lock irq_context: 0 (wq_completion)bond191#3 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &idev->mc_lock irq_context: 0 (wq_completion)bond191#3 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &idev->mc_lock fs_reclaim irq_context: 0 (wq_completion)bond191#3 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &idev->mc_lock fs_reclaim mmu_notifier_invalidate_range_start irq_context: 0 (wq_completion)bond191#3 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &idev->mc_lock &c->lock irq_context: 0 (wq_completion)bond191#3 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &idev->mc_lock rcu_read_lock &ndev->lock irq_context: 0 (wq_completion)bond191#3 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &idev->mc_lock rcu_read_lock &dir->lock#2 irq_context: 0 (wq_completion)bond191#3 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &idev->mc_lock rcu_read_lock &c->lock irq_context: 0 (wq_completion)bond191#3 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &idev->mc_lock rcu_read_lock &ul->lock#2 irq_context: 0 (wq_completion)bond191#3 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &idev->mc_lock rcu_read_lock rcu_read_lock &____s->seqcount#10 irq_context: 0 (wq_completion)bond191#3 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &idev->mc_lock rcu_read_lock rcu_read_lock rcu_read_lock_bh rcu_read_lock rcu_read_lock_bh rcu_read_lock &br->multicast_lock irq_context: 0 (wq_completion)bond191#3 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &idev->mc_lock rcu_read_lock rcu_read_lock rcu_read_lock_bh rcu_read_lock rcu_read_lock_bh rcu_read_lock &br->multicast_lock &obj_hash[i].lock irq_context: 0 (wq_completion)bond191#3 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &idev->mc_lock rcu_read_lock rcu_read_lock rcu_read_lock_bh rcu_read_lock rcu_read_lock_bh rcu_read_lock &br->multicast_lock &base->lock irq_context: 0 (wq_completion)bond191#3 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &idev->mc_lock rcu_read_lock rcu_read_lock rcu_read_lock_bh rcu_read_lock rcu_read_lock_bh rcu_read_lock &br->multicast_lock &base->lock &obj_hash[i].lock irq_context: 0 (wq_completion)bond191#3 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &idev->mc_lock rcu_read_lock rcu_read_lock rcu_read_lock_bh rcu_read_lock rcu_read_lock_bh rcu_read_lock &data->lock irq_context: 0 (wq_completion)bond191#3 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &idev->mc_lock rcu_read_lock rcu_read_lock rcu_read_lock_bh rcu_read_lock rcu_read_lock_bh rcu_read_lock &obj_hash[i].lock irq_context: 0 (wq_completion)bond191#3 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex.wait_lock irq_context: 0 (wq_completion)bond191#3 (work_completion)(&(&bond->mcast_work)->work) &p->pi_lock irq_context: 0 (wq_completion)bond180#3 (work_completion)(&(&slave->notify_work)->work) rtnl_mutex net_rwsem irq_context: 0 (wq_completion)bond180#3 (work_completion)(&(&slave->notify_work)->work) rtnl_mutex net_rwsem &list->lock#2 irq_context: 0 (wq_completion)bond180#3 (work_completion)(&(&slave->notify_work)->work) rtnl_mutex &tn->lock irq_context: 0 (wq_completion)bond180#3 (work_completion)(&(&slave->notify_work)->work) rtnl_mutex.wait_lock irq_context: 0 (wq_completion)bond180#3 (work_completion)(&(&slave->notify_work)->work) &p->pi_lock irq_context: 0 (wq_completion)bond181#2 (work_completion)(&(&slave->notify_work)->work) rtnl_mutex net_rwsem irq_context: 0 (wq_completion)bond181#2 (work_completion)(&(&slave->notify_work)->work) rtnl_mutex net_rwsem &list->lock#2 irq_context: 0 (wq_completion)bond181#2 (work_completion)(&(&slave->notify_work)->work) rtnl_mutex &tn->lock irq_context: 0 (wq_completion)bond181#2 (work_completion)(&(&slave->notify_work)->work) rtnl_mutex.wait_lock irq_context: 0 (wq_completion)bond181#2 (work_completion)(&(&slave->notify_work)->work) &p->pi_lock irq_context: 0 (wq_completion)bond161#3 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex fs_reclaim irq_context: 0 (wq_completion)bond161#3 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex fs_reclaim mmu_notifier_invalidate_range_start irq_context: 0 (wq_completion)bond161#3 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &c->lock irq_context: 0 (wq_completion)bond161#3 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &n->list_lock irq_context: 0 (wq_completion)bond161#3 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &n->list_lock &c->lock irq_context: 0 (wq_completion)bond161#3 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex rcu_read_lock &bond->stats_lock/1 irq_context: 0 (wq_completion)bond161#3 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex rcu_read_lock &ndev->lock irq_context: 0 (wq_completion)bond161#3 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex nl_table_lock irq_context: 0 (wq_completion)bond161#3 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &obj_hash[i].lock irq_context: 0 (wq_completion)bond161#3 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex nl_table_wait.lock irq_context: 0 (wq_completion)bond161#3 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex net_rwsem irq_context: 0 (wq_completion)bond161#3 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex net_rwsem &list->lock#2 irq_context: 0 (wq_completion)bond161#3 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &tn->lock irq_context: 0 (wq_completion)bond161#3 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &idev->mc_lock irq_context: 0 (wq_completion)bond161#3 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &idev->mc_lock fs_reclaim irq_context: 0 (wq_completion)bond161#3 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &idev->mc_lock fs_reclaim mmu_notifier_invalidate_range_start irq_context: 0 (wq_completion)bond161#3 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &idev->mc_lock fs_reclaim mmu_notifier_invalidate_range_start &rq->__lock irq_context: 0 (wq_completion)bond161#3 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &idev->mc_lock fs_reclaim mmu_notifier_invalidate_range_start &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)bond161#3 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &idev->mc_lock &c->lock irq_context: 0 (wq_completion)bond161#3 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &idev->mc_lock rcu_read_lock &ndev->lock irq_context: 0 (wq_completion)bond161#3 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &idev->mc_lock rcu_read_lock rcu_node_0 irq_context: 0 (wq_completion)bond161#3 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &idev->mc_lock rcu_read_lock &rq->__lock irq_context: 0 (wq_completion)bond161#3 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &idev->mc_lock rcu_read_lock &dir->lock#2 irq_context: 0 (wq_completion)bond161#3 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &idev->mc_lock rcu_read_lock &ul->lock#2 irq_context: 0 (wq_completion)bond161#3 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &idev->mc_lock rcu_read_lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)bond161#3 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &idev->mc_lock rcu_read_lock rcu_read_lock &obj_hash[i].lock irq_context: 0 (wq_completion)bond161#3 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &idev->mc_lock rcu_read_lock rcu_read_lock &dir->lock#2 irq_context: 0 (wq_completion)bond161#3 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &idev->mc_lock rcu_read_lock rcu_read_lock &tbl->lock irq_context: 0 (wq_completion)bond161#3 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &idev->mc_lock rcu_read_lock rcu_read_lock &n->lock irq_context: 0 (wq_completion)bond161#3 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &idev->mc_lock rcu_read_lock rcu_read_lock &____s->seqcount#9 irq_context: 0 (wq_completion)bond161#3 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &idev->mc_lock rcu_read_lock rcu_read_lock rcu_read_lock_bh rcu_read_lock rcu_read_lock_bh rcu_read_lock &br->multicast_lock irq_context: 0 (wq_completion)bond161#3 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &idev->mc_lock rcu_read_lock rcu_read_lock rcu_read_lock_bh rcu_read_lock rcu_read_lock_bh rcu_read_lock &br->multicast_lock &obj_hash[i].lock irq_context: 0 (wq_completion)bond161#3 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &idev->mc_lock rcu_read_lock rcu_read_lock rcu_read_lock_bh rcu_read_lock rcu_read_lock_bh rcu_read_lock &br->multicast_lock &base->lock irq_context: 0 (wq_completion)bond161#3 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &idev->mc_lock rcu_read_lock rcu_read_lock rcu_read_lock_bh rcu_read_lock rcu_read_lock_bh rcu_read_lock &br->multicast_lock &base->lock &obj_hash[i].lock irq_context: 0 (wq_completion)bond161#3 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &idev->mc_lock rcu_read_lock rcu_read_lock rcu_read_lock_bh rcu_read_lock rcu_read_lock_bh rcu_read_lock &data->lock irq_context: 0 (wq_completion)bond161#3 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &idev->mc_lock rcu_read_lock rcu_read_lock rcu_read_lock_bh rcu_read_lock rcu_read_lock_bh rcu_read_lock &obj_hash[i].lock irq_context: 0 (wq_completion)bond161#3 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &idev->mc_lock rcu_read_lock rcu_read_lock &rq->__lock irq_context: 0 (wq_completion)bond161#3 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &idev->mc_lock rcu_read_lock rcu_read_lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)bond161#3 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &idev->mc_lock rcu_read_lock &rcu_state.gp_wq irq_context: 0 (wq_completion)bond161#3 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &idev->mc_lock rcu_read_lock &rcu_state.gp_wq &p->pi_lock irq_context: 0 (wq_completion)bond161#3 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &idev->mc_lock rcu_read_lock &rcu_state.gp_wq &p->pi_lock &rq->__lock irq_context: 0 (wq_completion)bond161#3 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &idev->mc_lock rcu_read_lock &rcu_state.gp_wq &p->pi_lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)bond161#3 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex.wait_lock irq_context: 0 (wq_completion)bond161#3 (work_completion)(&(&bond->mcast_work)->work) &p->pi_lock irq_context: 0 (wq_completion)bond161#3 (work_completion)(&(&bond->mcast_work)->work) &p->pi_lock &rq->__lock irq_context: 0 (wq_completion)bond161#3 (work_completion)(&(&bond->mcast_work)->work) &p->pi_lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)bond200#3 irq_context: 0 (wq_completion)bond200#3 (work_completion)(&(&slave->notify_work)->work) irq_context: 0 (wq_completion)bond200#3 (work_completion)(&(&slave->notify_work)->work) &obj_hash[i].lock irq_context: 0 (wq_completion)bond200#3 (work_completion)(&(&slave->notify_work)->work) &base->lock irq_context: 0 (wq_completion)bond200#3 (work_completion)(&(&slave->notify_work)->work) &base->lock &obj_hash[i].lock irq_context: 0 (wq_completion)bond200#3 (work_completion)(&(&slave->notify_work)->work) &rq->__lock irq_context: 0 (wq_completion)bond200#3 (work_completion)(&(&bond->mcast_work)->work) irq_context: 0 (wq_completion)bond200#3 (work_completion)(&(&bond->mcast_work)->work) &obj_hash[i].lock irq_context: 0 (wq_completion)bond200#3 (work_completion)(&(&bond->mcast_work)->work) &base->lock irq_context: 0 (wq_completion)bond200#3 (work_completion)(&(&bond->mcast_work)->work) &base->lock &obj_hash[i].lock irq_context: 0 (wq_completion)bond200#3 (work_completion)(&(&bond->mcast_work)->work) &rq->__lock irq_context: 0 (wq_completion)bond200#3 (work_completion)(&(&bond->mcast_work)->work) &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)bond228 irq_context: 0 (wq_completion)bond228 (work_completion)(&(&slave->notify_work)->work) irq_context: 0 (wq_completion)bond228 (work_completion)(&(&slave->notify_work)->work) &obj_hash[i].lock irq_context: 0 (wq_completion)bond228 (work_completion)(&(&slave->notify_work)->work) &base->lock irq_context: 0 (wq_completion)bond228 (work_completion)(&(&slave->notify_work)->work) &base->lock &obj_hash[i].lock irq_context: 0 (wq_completion)bond228 (work_completion)(&(&slave->notify_work)->work) &rq->__lock irq_context: 0 (wq_completion)bond228 (work_completion)(&(&slave->notify_work)->work) &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)bond122#3 &rq->__lock irq_context: 0 (wq_completion)bond122#3 &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)bond190 &rq->__lock irq_context: 0 (wq_completion)bond190 &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)bond203 &rq->__lock irq_context: 0 (wq_completion)bond203 &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)bond228 (work_completion)(&(&bond->mcast_work)->work) irq_context: 0 (wq_completion)bond228 (work_completion)(&(&bond->mcast_work)->work) &obj_hash[i].lock irq_context: 0 (wq_completion)bond228 (work_completion)(&(&bond->mcast_work)->work) &base->lock irq_context: 0 (wq_completion)bond228 (work_completion)(&(&bond->mcast_work)->work) &base->lock &obj_hash[i].lock irq_context: 0 (wq_completion)bond228 (work_completion)(&(&bond->mcast_work)->work) &rq->__lock irq_context: 0 (wq_completion)bond228 (work_completion)(&(&bond->mcast_work)->work) &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)bond225 (work_completion)(&(&slave->notify_work)->work) rtnl_mutex net_rwsem irq_context: 0 (wq_completion)bond225 (work_completion)(&(&slave->notify_work)->work) rtnl_mutex net_rwsem &list->lock#2 irq_context: 0 (wq_completion)bond225 (work_completion)(&(&slave->notify_work)->work) rtnl_mutex &tn->lock irq_context: 0 (wq_completion)bond225 (work_completion)(&(&slave->notify_work)->work) rtnl_mutex.wait_lock irq_context: 0 (wq_completion)bond225 (work_completion)(&(&slave->notify_work)->work) &p->pi_lock irq_context: 0 (wq_completion)bond214#2 (work_completion)(&(&slave->notify_work)->work) rtnl_mutex net_rwsem irq_context: 0 (wq_completion)bond214#2 (work_completion)(&(&slave->notify_work)->work) rtnl_mutex net_rwsem &list->lock#2 irq_context: 0 (wq_completion)bond214#2 (work_completion)(&(&slave->notify_work)->work) rtnl_mutex &tn->lock irq_context: 0 (wq_completion)bond214#2 (work_completion)(&(&slave->notify_work)->work) rtnl_mutex.wait_lock irq_context: 0 (wq_completion)bond214#2 (work_completion)(&(&slave->notify_work)->work) &p->pi_lock irq_context: 0 (wq_completion)bond200#3 (work_completion)(&(&slave->notify_work)->work) rtnl_mutex net_rwsem irq_context: 0 (wq_completion)bond200#3 (work_completion)(&(&slave->notify_work)->work) rtnl_mutex net_rwsem &list->lock#2 irq_context: 0 (wq_completion)bond200#3 (work_completion)(&(&slave->notify_work)->work) rtnl_mutex &tn->lock irq_context: 0 (wq_completion)bond200#3 (work_completion)(&(&slave->notify_work)->work) rtnl_mutex.wait_lock irq_context: 0 (wq_completion)bond200#3 (work_completion)(&(&slave->notify_work)->work) &p->pi_lock irq_context: 0 (wq_completion)bond183#3 (work_completion)(&(&slave->notify_work)->work) rtnl_mutex net_rwsem irq_context: 0 (wq_completion)bond183#3 (work_completion)(&(&slave->notify_work)->work) rtnl_mutex net_rwsem &list->lock#2 irq_context: 0 (wq_completion)bond183#3 (work_completion)(&(&slave->notify_work)->work) rtnl_mutex &tn->lock irq_context: 0 (wq_completion)bond183#3 (work_completion)(&(&slave->notify_work)->work) rtnl_mutex.wait_lock irq_context: 0 (wq_completion)bond183#3 (work_completion)(&(&slave->notify_work)->work) &p->pi_lock irq_context: 0 (wq_completion)bond156#2 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex fs_reclaim irq_context: 0 (wq_completion)bond156#2 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex fs_reclaim mmu_notifier_invalidate_range_start irq_context: 0 (wq_completion)bond156#2 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex fs_reclaim mmu_notifier_invalidate_range_start &rq->__lock irq_context: 0 (wq_completion)bond156#2 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex fs_reclaim mmu_notifier_invalidate_range_start &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)bond156#2 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &c->lock irq_context: 0 (wq_completion)bond156#2 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &____s->seqcount#2 irq_context: 0 (wq_completion)bond156#2 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &____s->seqcount irq_context: 0 (wq_completion)bond156#2 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex rcu_read_lock &bond->stats_lock/1 irq_context: 0 (wq_completion)bond156#2 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex rcu_read_lock &ndev->lock irq_context: 0 (wq_completion)bond156#2 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex rcu_read_lock rcu_node_0 irq_context: 0 (wq_completion)bond156#2 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex rcu_read_lock &rq->__lock irq_context: 0 (wq_completion)bond156#2 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex rcu_read_lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)bond156#2 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex nl_table_lock irq_context: 0 (wq_completion)bond156#2 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &obj_hash[i].lock irq_context: 0 (wq_completion)bond156#2 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex nl_table_wait.lock irq_context: 0 (wq_completion)bond156#2 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex net_rwsem irq_context: 0 (wq_completion)bond156#2 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex net_rwsem &rq->__lock irq_context: 0 (wq_completion)bond156#2 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex net_rwsem &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)bond156#2 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex net_rwsem &list->lock#2 irq_context: 0 (wq_completion)bond156#2 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &tn->lock irq_context: 0 (wq_completion)bond156#2 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &idev->mc_lock irq_context: 0 (wq_completion)bond156#2 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &idev->mc_lock &rq->__lock irq_context: 0 (wq_completion)bond156#2 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &idev->mc_lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)bond156#2 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &idev->mc_lock fs_reclaim irq_context: 0 (wq_completion)bond156#2 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &idev->mc_lock fs_reclaim mmu_notifier_invalidate_range_start irq_context: 0 (wq_completion)bond156#2 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &idev->mc_lock &c->lock irq_context: 0 (wq_completion)bond156#2 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &idev->mc_lock &n->list_lock irq_context: 0 (wq_completion)bond156#2 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &idev->mc_lock &n->list_lock &c->lock irq_context: 0 (wq_completion)bond156#2 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &idev->mc_lock rcu_read_lock &ndev->lock irq_context: 0 (wq_completion)bond156#2 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &idev->mc_lock rcu_read_lock &dir->lock#2 irq_context: 0 (wq_completion)bond156#2 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &idev->mc_lock rcu_read_lock &ul->lock#2 irq_context: 0 (wq_completion)bond156#2 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &idev->mc_lock rcu_read_lock rcu_read_lock &obj_hash[i].lock irq_context: 0 (wq_completion)bond156#2 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &idev->mc_lock rcu_read_lock rcu_read_lock &dir->lock#2 irq_context: 0 (wq_completion)bond156#2 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &idev->mc_lock rcu_read_lock rcu_read_lock &tbl->lock irq_context: 0 (wq_completion)bond156#2 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &idev->mc_lock rcu_read_lock rcu_read_lock &n->lock irq_context: 0 (wq_completion)bond156#2 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &idev->mc_lock rcu_read_lock rcu_read_lock &____s->seqcount#9 irq_context: 0 (wq_completion)bond156#2 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &idev->mc_lock rcu_read_lock rcu_read_lock rcu_read_lock_bh rcu_read_lock rcu_read_lock_bh rcu_read_lock &br->multicast_lock irq_context: 0 (wq_completion)bond156#2 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &idev->mc_lock rcu_read_lock rcu_read_lock rcu_read_lock_bh rcu_read_lock rcu_read_lock_bh rcu_read_lock &br->multicast_lock &obj_hash[i].lock irq_context: 0 (wq_completion)bond156#2 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &idev->mc_lock rcu_read_lock rcu_read_lock rcu_read_lock_bh rcu_read_lock rcu_read_lock_bh rcu_read_lock &br->multicast_lock &base->lock irq_context: 0 (wq_completion)bond156#2 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &idev->mc_lock rcu_read_lock rcu_read_lock rcu_read_lock_bh rcu_read_lock rcu_read_lock_bh rcu_read_lock &br->multicast_lock &base->lock &obj_hash[i].lock irq_context: 0 (wq_completion)bond156#2 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &idev->mc_lock rcu_read_lock rcu_read_lock rcu_read_lock_bh rcu_read_lock rcu_read_lock_bh rcu_read_lock &data->lock irq_context: 0 (wq_completion)bond156#2 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &idev->mc_lock rcu_read_lock rcu_read_lock rcu_read_lock_bh rcu_read_lock rcu_read_lock_bh rcu_read_lock &obj_hash[i].lock irq_context: 0 (wq_completion)bond156#2 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &idev->mc_lock rcu_read_lock rcu_read_lock rcu_node_0 irq_context: 0 (wq_completion)bond156#2 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &idev->mc_lock rcu_read_lock rcu_read_lock &rq->__lock irq_context: 0 (wq_completion)bond156#2 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &idev->mc_lock rcu_read_lock rcu_read_lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)bond156#2 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &idev->mc_lock rcu_read_lock rcu_node_0 irq_context: 0 (wq_completion)bond156#2 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &idev->mc_lock rcu_read_lock &rcu_state.gp_wq irq_context: 0 (wq_completion)bond156#2 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &idev->mc_lock rcu_read_lock &rcu_state.gp_wq &p->pi_lock irq_context: 0 (wq_completion)bond156#2 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &idev->mc_lock rcu_read_lock &rcu_state.gp_wq &p->pi_lock &rq->__lock irq_context: 0 (wq_completion)bond156#2 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &idev->mc_lock rcu_read_lock &rcu_state.gp_wq &p->pi_lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)bond156#2 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &idev->mc_lock rcu_read_lock &rq->__lock irq_context: 0 (wq_completion)bond156#2 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &idev->mc_lock rcu_read_lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)bond156#2 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex.wait_lock irq_context: 0 (wq_completion)bond156#2 (work_completion)(&(&bond->mcast_work)->work) &p->pi_lock irq_context: 0 (wq_completion)bond156#2 (work_completion)(&(&bond->mcast_work)->work) &p->pi_lock &rq->__lock irq_context: 0 (wq_completion)bond156#2 (work_completion)(&(&bond->mcast_work)->work) &p->pi_lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)bond190#3 (work_completion)(&(&slave->notify_work)->work) &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)bond217#2 irq_context: 0 (wq_completion)bond217#2 &rq->__lock irq_context: 0 (wq_completion)bond217#2 &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)bond217#2 (work_completion)(&(&slave->notify_work)->work) irq_context: 0 (wq_completion)bond217#2 (work_completion)(&(&slave->notify_work)->work) &obj_hash[i].lock irq_context: 0 (wq_completion)bond217#2 (work_completion)(&(&slave->notify_work)->work) &base->lock irq_context: 0 (wq_completion)bond217#2 (work_completion)(&(&slave->notify_work)->work) &base->lock &obj_hash[i].lock irq_context: 0 (wq_completion)bond195 &rq->__lock irq_context: 0 (wq_completion)bond195 &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)bond217#2 (work_completion)(&(&bond->mcast_work)->work) irq_context: 0 (wq_completion)bond217#2 (work_completion)(&(&bond->mcast_work)->work) &obj_hash[i].lock irq_context: 0 (wq_completion)bond217#2 (work_completion)(&(&bond->mcast_work)->work) &base->lock irq_context: 0 (wq_completion)bond217#2 (work_completion)(&(&bond->mcast_work)->work) &base->lock &obj_hash[i].lock irq_context: 0 (wq_completion)bond217#2 (work_completion)(&(&bond->mcast_work)->work) &rq->__lock irq_context: 0 (wq_completion)bond217#2 (work_completion)(&(&bond->mcast_work)->work) &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)bond68#2 irq_context: 0 (wq_completion)bond68#2 (work_completion)(&(&slave->notify_work)->work) irq_context: 0 (wq_completion)bond68#2 (work_completion)(&(&slave->notify_work)->work) &obj_hash[i].lock irq_context: 0 (wq_completion)bond68#2 (work_completion)(&(&slave->notify_work)->work) &base->lock irq_context: 0 (wq_completion)bond68#2 (work_completion)(&(&slave->notify_work)->work) &base->lock &obj_hash[i].lock irq_context: 0 (wq_completion)bond68#2 (work_completion)(&(&bond->mcast_work)->work) irq_context: 0 (wq_completion)bond68#2 (work_completion)(&(&bond->mcast_work)->work) &obj_hash[i].lock irq_context: 0 (wq_completion)bond68#2 (work_completion)(&(&bond->mcast_work)->work) &base->lock irq_context: 0 (wq_completion)bond68#2 (work_completion)(&(&bond->mcast_work)->work) &base->lock &obj_hash[i].lock irq_context: 0 (wq_completion)bond68#2 (work_completion)(&(&bond->mcast_work)->work) &rq->__lock irq_context: 0 (wq_completion)events_long (work_completion)(&(&br->gc_work)->work) rcu_read_lock rcu_node_0 irq_context: 0 (wq_completion)events_long (work_completion)(&(&br->gc_work)->work) rcu_read_lock &rq->__lock irq_context: 0 (wq_completion)events_long (work_completion)(&(&br->gc_work)->work) rcu_read_lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)bond201#3 irq_context: 0 (wq_completion)bond201#3 (work_completion)(&(&slave->notify_work)->work) irq_context: 0 (wq_completion)bond201#3 (work_completion)(&(&slave->notify_work)->work) &obj_hash[i].lock irq_context: 0 (wq_completion)bond201#3 (work_completion)(&(&slave->notify_work)->work) &base->lock irq_context: 0 (wq_completion)bond201#3 (work_completion)(&(&slave->notify_work)->work) &base->lock &obj_hash[i].lock irq_context: 0 (wq_completion)bond201#3 (work_completion)(&(&slave->notify_work)->work) &rq->__lock irq_context: 0 (wq_completion)bond201#3 (work_completion)(&(&slave->notify_work)->work) &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)bond201#3 (work_completion)(&(&slave->notify_work)->work) &cfs_rq->removed.lock irq_context: 0 (wq_completion)bond201#3 (work_completion)(&(&bond->mcast_work)->work) irq_context: 0 (wq_completion)bond201#3 (work_completion)(&(&bond->mcast_work)->work) &obj_hash[i].lock irq_context: 0 (wq_completion)bond201#3 (work_completion)(&(&bond->mcast_work)->work) &base->lock irq_context: 0 (wq_completion)bond201#3 (work_completion)(&(&bond->mcast_work)->work) &base->lock &obj_hash[i].lock irq_context: 0 (wq_completion)bond193#3 &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)bond229 irq_context: 0 (wq_completion)bond229 (work_completion)(&(&slave->notify_work)->work) irq_context: 0 (wq_completion)bond229 (work_completion)(&(&slave->notify_work)->work) &obj_hash[i].lock irq_context: 0 (wq_completion)bond229 (work_completion)(&(&slave->notify_work)->work) &base->lock irq_context: 0 (wq_completion)bond229 (work_completion)(&(&slave->notify_work)->work) &base->lock &obj_hash[i].lock irq_context: 0 (wq_completion)bond229 (work_completion)(&(&bond->mcast_work)->work) irq_context: 0 (wq_completion)bond229 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex fs_reclaim irq_context: 0 (wq_completion)bond229 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex fs_reclaim mmu_notifier_invalidate_range_start irq_context: 0 (wq_completion)bond229 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex rcu_read_lock &bond->stats_lock/1 irq_context: 0 (wq_completion)bond229 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex rcu_read_lock &ndev->lock irq_context: 0 (wq_completion)bond229 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex nl_table_lock irq_context: 0 (wq_completion)bond229 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &obj_hash[i].lock irq_context: 0 (wq_completion)bond229 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex nl_table_wait.lock irq_context: 0 (wq_completion)bond229 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex net_rwsem irq_context: 0 (wq_completion)bond229 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex net_rwsem &list->lock#2 irq_context: 0 (wq_completion)bond229 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &tn->lock irq_context: 0 (wq_completion)bond229 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &idev->mc_lock irq_context: 0 (wq_completion)bond229 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex.wait_lock irq_context: 0 (wq_completion)bond229 (work_completion)(&(&bond->mcast_work)->work) &p->pi_lock irq_context: 0 (wq_completion)bond199#2 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex fs_reclaim irq_context: 0 (wq_completion)bond199#2 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex fs_reclaim mmu_notifier_invalidate_range_start irq_context: 0 (wq_completion)bond199#2 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &c->lock irq_context: 0 (wq_completion)bond199#2 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &n->list_lock irq_context: 0 (wq_completion)bond199#2 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &n->list_lock &c->lock irq_context: 0 (wq_completion)bond199#2 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex rcu_read_lock &bond->stats_lock/1 irq_context: 0 (wq_completion)bond199#2 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex rcu_read_lock &ndev->lock irq_context: 0 (wq_completion)bond199#2 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex nl_table_lock irq_context: 0 (wq_completion)bond199#2 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &obj_hash[i].lock irq_context: 0 (wq_completion)bond199#2 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex nl_table_wait.lock irq_context: 0 (wq_completion)bond199#2 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex net_rwsem irq_context: 0 (wq_completion)bond199#2 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex net_rwsem &list->lock#2 irq_context: 0 (wq_completion)bond199#2 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &tn->lock irq_context: 0 (wq_completion)bond199#2 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &idev->mc_lock irq_context: 0 (wq_completion)bond199#2 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &idev->mc_lock fs_reclaim irq_context: 0 (wq_completion)bond199#2 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &idev->mc_lock fs_reclaim mmu_notifier_invalidate_range_start irq_context: 0 (wq_completion)bond199#2 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &idev->mc_lock rcu_read_lock &ndev->lock irq_context: 0 (wq_completion)bond199#2 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &idev->mc_lock rcu_read_lock &dir->lock#2 irq_context: 0 (wq_completion)bond199#2 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &idev->mc_lock rcu_read_lock &ul->lock#2 irq_context: 0 (wq_completion)bond199#2 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &idev->mc_lock rcu_read_lock rcu_read_lock &____s->seqcount#10 irq_context: 0 (wq_completion)bond199#2 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &idev->mc_lock rcu_read_lock rcu_read_lock rcu_read_lock_bh rcu_read_lock rcu_read_lock_bh rcu_read_lock &br->multicast_lock irq_context: 0 (wq_completion)bond199#2 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &idev->mc_lock rcu_read_lock rcu_read_lock rcu_read_lock_bh rcu_read_lock rcu_read_lock_bh rcu_read_lock &br->multicast_lock &obj_hash[i].lock irq_context: 0 (wq_completion)bond199#2 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &idev->mc_lock rcu_read_lock rcu_read_lock rcu_read_lock_bh rcu_read_lock rcu_read_lock_bh rcu_read_lock &br->multicast_lock &base->lock irq_context: 0 (wq_completion)bond199#2 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &idev->mc_lock rcu_read_lock rcu_read_lock rcu_read_lock_bh rcu_read_lock rcu_read_lock_bh rcu_read_lock &br->multicast_lock &base->lock &obj_hash[i].lock irq_context: 0 (wq_completion)bond199#2 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &idev->mc_lock rcu_read_lock rcu_read_lock rcu_read_lock_bh rcu_read_lock rcu_read_lock_bh rcu_read_lock &data->lock irq_context: 0 (wq_completion)bond199#2 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &idev->mc_lock rcu_read_lock rcu_read_lock rcu_read_lock_bh rcu_read_lock rcu_read_lock_bh rcu_read_lock &obj_hash[i].lock irq_context: 0 (wq_completion)bond199#2 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex.wait_lock irq_context: 0 (wq_completion)bond199#2 (work_completion)(&(&bond->mcast_work)->work) &p->pi_lock irq_context: 0 (wq_completion)bond213 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex rcu_read_lock rcu_node_0 irq_context: 0 (wq_completion)bond213 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex rcu_read_lock &rq->__lock irq_context: 0 (wq_completion)bond213 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex fs_reclaim irq_context: 0 (wq_completion)bond213 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex fs_reclaim mmu_notifier_invalidate_range_start irq_context: 0 (wq_completion)bond213 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex rcu_read_lock &bond->stats_lock/1 irq_context: 0 (wq_completion)bond213 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex rcu_read_lock &ndev->lock irq_context: 0 (wq_completion)bond213 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex nl_table_lock irq_context: 0 (wq_completion)bond213 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &obj_hash[i].lock irq_context: 0 (wq_completion)bond213 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex nl_table_wait.lock irq_context: 0 (wq_completion)bond213 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex net_rwsem irq_context: 0 (wq_completion)bond213 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex net_rwsem &list->lock#2 irq_context: 0 (wq_completion)bond213 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &tn->lock irq_context: 0 (wq_completion)bond213 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &idev->mc_lock irq_context: 0 (wq_completion)bond213 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &idev->mc_lock fs_reclaim irq_context: 0 (wq_completion)bond213 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &idev->mc_lock fs_reclaim mmu_notifier_invalidate_range_start irq_context: 0 (wq_completion)bond213 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &idev->mc_lock rcu_read_lock &ndev->lock irq_context: 0 (wq_completion)bond213 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &idev->mc_lock rcu_read_lock &c->lock irq_context: 0 (wq_completion)bond213 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &idev->mc_lock rcu_read_lock &dir->lock#2 irq_context: 0 (wq_completion)bond213 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &idev->mc_lock rcu_read_lock &ul->lock#2 irq_context: 0 (wq_completion)bond213 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &idev->mc_lock rcu_read_lock rcu_read_lock &____s->seqcount#10 irq_context: 0 (wq_completion)bond213 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &idev->mc_lock rcu_read_lock rcu_read_lock rcu_read_lock_bh rcu_read_lock rcu_read_lock_bh rcu_read_lock &br->multicast_lock irq_context: 0 (wq_completion)bond213 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &idev->mc_lock rcu_read_lock rcu_read_lock rcu_read_lock_bh rcu_read_lock rcu_read_lock_bh rcu_read_lock &br->multicast_lock &obj_hash[i].lock irq_context: 0 (wq_completion)bond213 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &idev->mc_lock rcu_read_lock rcu_read_lock rcu_read_lock_bh rcu_read_lock rcu_read_lock_bh rcu_read_lock &br->multicast_lock &base->lock irq_context: 0 (wq_completion)bond213 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &idev->mc_lock rcu_read_lock rcu_read_lock rcu_read_lock_bh rcu_read_lock rcu_read_lock_bh rcu_read_lock &br->multicast_lock &base->lock &obj_hash[i].lock irq_context: 0 (wq_completion)bond213 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &idev->mc_lock rcu_read_lock rcu_read_lock rcu_read_lock_bh rcu_read_lock rcu_read_lock_bh rcu_read_lock &data->lock irq_context: 0 (wq_completion)bond213 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &idev->mc_lock rcu_read_lock rcu_read_lock rcu_read_lock_bh rcu_read_lock rcu_read_lock_bh rcu_read_lock &obj_hash[i].lock irq_context: 0 (wq_completion)bond213 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &idev->mc_lock rcu_read_lock rcu_read_lock rcu_node_0 irq_context: 0 (wq_completion)bond213 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &idev->mc_lock rcu_read_lock rcu_read_lock &rq->__lock irq_context: 0 (wq_completion)bond213 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &idev->mc_lock rcu_read_lock rcu_node_0 irq_context: 0 (wq_completion)bond213 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex.wait_lock irq_context: 0 (wq_completion)bond213 (work_completion)(&(&bond->mcast_work)->work) &p->pi_lock irq_context: 0 (wq_completion)bond213 (work_completion)(&(&bond->mcast_work)->work) &p->pi_lock &rq->__lock irq_context: 0 (wq_completion)bond213 (work_completion)(&(&bond->mcast_work)->work) &p->pi_lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)bond218#2 irq_context: 0 (wq_completion)bond218#2 &rq->__lock irq_context: 0 (wq_completion)bond218#2 (work_completion)(&(&slave->notify_work)->work) irq_context: 0 (wq_completion)bond218#2 (work_completion)(&(&slave->notify_work)->work) &obj_hash[i].lock irq_context: 0 (wq_completion)bond218#2 (work_completion)(&(&slave->notify_work)->work) &base->lock irq_context: 0 (wq_completion)bond218#2 (work_completion)(&(&slave->notify_work)->work) &base->lock &obj_hash[i].lock irq_context: 0 (wq_completion)bond218#2 (work_completion)(&(&bond->mcast_work)->work) irq_context: 0 (wq_completion)bond218#2 (work_completion)(&(&bond->mcast_work)->work) &obj_hash[i].lock irq_context: 0 (wq_completion)bond218#2 (work_completion)(&(&bond->mcast_work)->work) &base->lock irq_context: 0 (wq_completion)bond218#2 (work_completion)(&(&bond->mcast_work)->work) &base->lock &obj_hash[i].lock irq_context: 0 (wq_completion)bond176#2 (work_completion)(&(&bond->mcast_work)->work) fill_pool_map-wait-type-override pool_lock irq_context: 0 (wq_completion)bond202#3 irq_context: 0 (wq_completion)bond202#3 (work_completion)(&(&slave->notify_work)->work) irq_context: 0 (wq_completion)bond202#3 (work_completion)(&(&slave->notify_work)->work) &obj_hash[i].lock irq_context: 0 (wq_completion)bond202#3 (work_completion)(&(&slave->notify_work)->work) &base->lock irq_context: 0 (wq_completion)bond202#3 (work_completion)(&(&slave->notify_work)->work) &base->lock &obj_hash[i].lock irq_context: 0 (wq_completion)bond202#3 (work_completion)(&(&slave->notify_work)->work) &rq->__lock irq_context: 0 (wq_completion)bond202#3 (work_completion)(&(&bond->mcast_work)->work) irq_context: 0 (wq_completion)bond202#3 (work_completion)(&(&bond->mcast_work)->work) &obj_hash[i].lock irq_context: 0 (wq_completion)bond202#3 (work_completion)(&(&bond->mcast_work)->work) &base->lock irq_context: 0 (wq_completion)bond202#3 (work_completion)(&(&bond->mcast_work)->work) &base->lock &obj_hash[i].lock irq_context: 0 (wq_completion)bond202#3 (work_completion)(&(&bond->mcast_work)->work) &rq->__lock irq_context: 0 (wq_completion)bond202#3 (work_completion)(&(&bond->mcast_work)->work) &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)bond198#2 &rq->__lock irq_context: 0 (wq_completion)bond198#2 &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)bond211 (work_completion)(&(&bond->mcast_work)->work) &rq->__lock irq_context: 0 (wq_completion)bond211 (work_completion)(&(&bond->mcast_work)->work) &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)bond211 (work_completion)(&(&bond->mcast_work)->work) &cfs_rq->removed.lock irq_context: 0 (wq_completion)bond189#2 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex fs_reclaim irq_context: 0 (wq_completion)bond189#2 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex fs_reclaim mmu_notifier_invalidate_range_start irq_context: 0 (wq_completion)bond189#2 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &c->lock irq_context: 0 (wq_completion)bond189#2 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex rcu_read_lock &bond->stats_lock/1 irq_context: 0 (wq_completion)bond189#2 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex rcu_read_lock &ndev->lock irq_context: 0 (wq_completion)bond189#2 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex nl_table_lock irq_context: 0 (wq_completion)bond189#2 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &obj_hash[i].lock irq_context: 0 (wq_completion)bond189#2 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex nl_table_wait.lock irq_context: 0 (wq_completion)bond189#2 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex net_rwsem irq_context: 0 (wq_completion)bond189#2 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex net_rwsem &list->lock#2 irq_context: 0 (wq_completion)bond189#2 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &tn->lock irq_context: 0 (wq_completion)bond189#2 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &idev->mc_lock irq_context: 0 (wq_completion)bond189#2 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &idev->mc_lock fs_reclaim irq_context: 0 (wq_completion)bond189#2 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &idev->mc_lock fs_reclaim mmu_notifier_invalidate_range_start irq_context: 0 (wq_completion)bond189#2 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &idev->mc_lock &c->lock irq_context: 0 (wq_completion)bond189#2 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &idev->mc_lock rcu_read_lock &ndev->lock irq_context: 0 (wq_completion)bond189#2 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &idev->mc_lock rcu_read_lock &dir->lock#2 irq_context: 0 (wq_completion)bond189#2 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &idev->mc_lock rcu_read_lock &c->lock irq_context: 0 (wq_completion)bond189#2 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &idev->mc_lock rcu_read_lock &ul->lock#2 irq_context: 0 (wq_completion)bond189#2 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &idev->mc_lock rcu_read_lock rcu_read_lock &____s->seqcount#10 irq_context: 0 (wq_completion)bond189#2 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &idev->mc_lock rcu_read_lock rcu_read_lock rcu_read_lock_bh rcu_read_lock rcu_read_lock_bh rcu_read_lock &br->multicast_lock irq_context: 0 (wq_completion)bond189#2 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &idev->mc_lock rcu_read_lock rcu_read_lock rcu_read_lock_bh rcu_read_lock rcu_read_lock_bh rcu_read_lock &br->multicast_lock &obj_hash[i].lock irq_context: 0 (wq_completion)bond189#2 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &idev->mc_lock rcu_read_lock rcu_read_lock rcu_read_lock_bh rcu_read_lock rcu_read_lock_bh rcu_read_lock &br->multicast_lock &base->lock irq_context: 0 (wq_completion)bond189#2 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &idev->mc_lock rcu_read_lock rcu_read_lock rcu_read_lock_bh rcu_read_lock rcu_read_lock_bh rcu_read_lock &br->multicast_lock &base->lock &obj_hash[i].lock irq_context: 0 (wq_completion)bond189#2 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &idev->mc_lock rcu_read_lock rcu_read_lock rcu_read_lock_bh rcu_read_lock rcu_read_lock_bh rcu_read_lock &data->lock irq_context: 0 (wq_completion)bond189#2 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &idev->mc_lock rcu_read_lock rcu_read_lock rcu_read_lock_bh rcu_read_lock rcu_read_lock_bh rcu_read_lock &obj_hash[i].lock irq_context: 0 (wq_completion)bond189#2 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex.wait_lock irq_context: 0 (wq_completion)bond189#2 (work_completion)(&(&bond->mcast_work)->work) &p->pi_lock irq_context: 0 (wq_completion)bond124#3 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex fs_reclaim irq_context: 0 (wq_completion)bond124#3 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex fs_reclaim mmu_notifier_invalidate_range_start irq_context: 0 (wq_completion)bond124#3 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex rcu_read_lock &bond->stats_lock/1 irq_context: 0 (wq_completion)bond124#3 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex rcu_read_lock &ndev->lock irq_context: 0 (wq_completion)bond124#3 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex nl_table_lock irq_context: 0 (wq_completion)bond124#3 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &obj_hash[i].lock irq_context: 0 (wq_completion)bond124#3 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex nl_table_wait.lock irq_context: 0 (wq_completion)bond124#3 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex net_rwsem irq_context: 0 (wq_completion)bond124#3 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex net_rwsem &list->lock#2 irq_context: 0 (wq_completion)bond124#3 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &tn->lock irq_context: 0 (wq_completion)bond124#3 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &idev->mc_lock irq_context: 0 (wq_completion)bond124#3 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &idev->mc_lock fs_reclaim irq_context: 0 (wq_completion)bond124#3 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &idev->mc_lock fs_reclaim mmu_notifier_invalidate_range_start irq_context: 0 (wq_completion)bond124#3 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &idev->mc_lock &c->lock irq_context: 0 (wq_completion)bond124#3 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &idev->mc_lock rcu_read_lock &ndev->lock irq_context: 0 (wq_completion)bond124#3 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &idev->mc_lock rcu_read_lock &dir->lock#2 irq_context: 0 (wq_completion)bond124#3 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &idev->mc_lock rcu_read_lock &ul->lock#2 irq_context: 0 (wq_completion)bond124#3 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &idev->mc_lock rcu_read_lock rcu_read_lock &obj_hash[i].lock irq_context: 0 (wq_completion)bond124#3 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &idev->mc_lock rcu_read_lock rcu_read_lock &dir->lock#2 irq_context: 0 (wq_completion)bond124#3 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &idev->mc_lock rcu_read_lock rcu_read_lock &tbl->lock irq_context: 0 (wq_completion)bond124#3 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &idev->mc_lock rcu_read_lock rcu_read_lock &n->lock irq_context: 0 (wq_completion)bond124#3 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &idev->mc_lock rcu_read_lock rcu_read_lock &____s->seqcount#9 irq_context: 0 (wq_completion)bond124#3 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &idev->mc_lock rcu_read_lock rcu_read_lock rcu_read_lock_bh rcu_read_lock rcu_read_lock_bh rcu_read_lock &br->multicast_lock irq_context: 0 (wq_completion)bond124#3 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &idev->mc_lock rcu_read_lock rcu_read_lock rcu_read_lock_bh rcu_read_lock rcu_read_lock_bh rcu_read_lock &br->multicast_lock &obj_hash[i].lock irq_context: 0 (wq_completion)bond124#3 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &idev->mc_lock rcu_read_lock rcu_read_lock rcu_read_lock_bh rcu_read_lock rcu_read_lock_bh rcu_read_lock &br->multicast_lock &base->lock irq_context: 0 (wq_completion)bond124#3 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &idev->mc_lock rcu_read_lock rcu_read_lock rcu_read_lock_bh rcu_read_lock rcu_read_lock_bh rcu_read_lock &br->multicast_lock &base->lock &obj_hash[i].lock irq_context: 0 (wq_completion)bond124#3 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &idev->mc_lock rcu_read_lock rcu_read_lock rcu_read_lock_bh rcu_read_lock rcu_read_lock_bh rcu_read_lock &data->lock irq_context: 0 (wq_completion)bond124#3 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &idev->mc_lock rcu_read_lock rcu_read_lock rcu_read_lock_bh rcu_read_lock rcu_read_lock_bh rcu_read_lock &obj_hash[i].lock irq_context: 0 (wq_completion)bond124#3 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &rq->__lock irq_context: 0 (wq_completion)bond124#3 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)bond124#3 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex.wait_lock irq_context: 0 (wq_completion)bond124#3 (work_completion)(&(&bond->mcast_work)->work) &p->pi_lock irq_context: 0 (wq_completion)bond230 irq_context: 0 (wq_completion)bond230 (work_completion)(&(&slave->notify_work)->work) irq_context: 0 (wq_completion)bond230 (work_completion)(&(&slave->notify_work)->work) &obj_hash[i].lock irq_context: 0 (wq_completion)bond230 (work_completion)(&(&slave->notify_work)->work) &base->lock irq_context: 0 (wq_completion)bond230 (work_completion)(&(&slave->notify_work)->work) &base->lock &obj_hash[i].lock irq_context: 0 (wq_completion)bond230 (work_completion)(&(&slave->notify_work)->work) &rq->__lock irq_context: 0 (wq_completion)bond230 (work_completion)(&(&slave->notify_work)->work) &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)bond230 (work_completion)(&(&bond->mcast_work)->work) irq_context: 0 (wq_completion)bond230 (work_completion)(&(&bond->mcast_work)->work) &obj_hash[i].lock irq_context: 0 (wq_completion)bond230 (work_completion)(&(&bond->mcast_work)->work) &base->lock irq_context: 0 (wq_completion)bond230 (work_completion)(&(&bond->mcast_work)->work) &base->lock &obj_hash[i].lock irq_context: 0 (wq_completion)bond230 (work_completion)(&(&bond->mcast_work)->work) &rq->__lock irq_context: 0 (wq_completion)bond230 (work_completion)(&(&bond->mcast_work)->work) &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: softirq (&ndev->rs_timer) rcu_read_lock rcu_read_lock rcu_read_lock_bh &sch->q.lock batched_entropy_u16.lock crngs.lock irq_context: 0 (wq_completion)bond204#2 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex fs_reclaim irq_context: 0 (wq_completion)bond204#2 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex fs_reclaim mmu_notifier_invalidate_range_start irq_context: 0 (wq_completion)bond204#2 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex rcu_read_lock &bond->stats_lock/1 irq_context: 0 (wq_completion)bond204#2 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex rcu_read_lock &ndev->lock irq_context: 0 (wq_completion)bond204#2 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex nl_table_lock irq_context: 0 (wq_completion)bond204#2 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &obj_hash[i].lock irq_context: 0 (wq_completion)bond204#2 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &rq->__lock irq_context: 0 (wq_completion)bond204#2 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)bond204#2 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex nl_table_wait.lock irq_context: 0 (wq_completion)bond204#2 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex net_rwsem irq_context: 0 (wq_completion)bond204#2 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex net_rwsem &list->lock#2 irq_context: 0 (wq_completion)bond204#2 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &tn->lock irq_context: 0 (wq_completion)bond204#2 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &idev->mc_lock irq_context: 0 (wq_completion)bond204#2 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &idev->mc_lock fs_reclaim irq_context: 0 (wq_completion)bond204#2 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &idev->mc_lock fs_reclaim mmu_notifier_invalidate_range_start irq_context: 0 (wq_completion)bond204#2 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &idev->mc_lock rcu_read_lock &ndev->lock irq_context: 0 (wq_completion)bond204#2 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &idev->mc_lock rcu_read_lock rcu_node_0 irq_context: 0 (wq_completion)bond204#2 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &idev->mc_lock rcu_read_lock &rq->__lock irq_context: 0 (wq_completion)bond204#2 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &idev->mc_lock rcu_read_lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)bond204#2 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &idev->mc_lock rcu_read_lock &dir->lock#2 irq_context: 0 (wq_completion)bond204#2 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &idev->mc_lock rcu_read_lock &ul->lock#2 irq_context: 0 (wq_completion)bond204#2 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &idev->mc_lock rcu_read_lock rcu_read_lock &____s->seqcount#10 irq_context: 0 (wq_completion)bond204#2 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &idev->mc_lock rcu_read_lock rcu_read_lock rcu_read_lock_bh rcu_read_lock rcu_read_lock_bh rcu_read_lock &br->multicast_lock irq_context: 0 (wq_completion)bond204#2 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &idev->mc_lock rcu_read_lock rcu_read_lock rcu_read_lock_bh rcu_read_lock rcu_read_lock_bh rcu_read_lock &br->multicast_lock &obj_hash[i].lock irq_context: 0 (wq_completion)bond204#2 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &idev->mc_lock rcu_read_lock rcu_read_lock rcu_read_lock_bh rcu_read_lock rcu_read_lock_bh rcu_read_lock &br->multicast_lock &base->lock irq_context: 0 (wq_completion)bond204#2 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &idev->mc_lock rcu_read_lock rcu_read_lock rcu_read_lock_bh rcu_read_lock rcu_read_lock_bh rcu_read_lock &br->multicast_lock &base->lock &obj_hash[i].lock irq_context: 0 (wq_completion)bond204#2 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &idev->mc_lock rcu_read_lock rcu_read_lock rcu_read_lock_bh rcu_read_lock rcu_read_lock_bh rcu_read_lock &data->lock irq_context: 0 (wq_completion)bond204#2 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &idev->mc_lock rcu_read_lock rcu_read_lock rcu_read_lock_bh rcu_read_lock rcu_read_lock_bh rcu_read_lock &obj_hash[i].lock irq_context: 0 (wq_completion)bond204#2 (work_completion)(&(&bond->mcast_work)->work) rtnl_mutex &idev->mc_lock rcu_read_lock rcu_read_lock &rq->__lock irq_context: 0 (wq_completi